Update README.md
Browse files
README.md
CHANGED
|
@@ -34,7 +34,7 @@ The model is a result of fine-tuning Mistral-7B-v0.1 on a down stream task, in l
|
|
| 34 |
### Model Description
|
| 35 |
|
| 36 |
<!-- Provide a longer summary of what this model is. -->
|
| 37 |
-
dsfsi/
|
| 38 |
for large language model in regard to low resourced morphologically rich African languages using custom prompt.
|
| 39 |
|
| 40 |
|
|
@@ -64,7 +64,7 @@ be used in the Education industry to teach core subjects in native South African
|
|
| 64 |
### Direct Use
|
| 65 |
|
| 66 |
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
|
| 67 |
-
You can download the model, dsfsi/
|
| 68 |
|
| 69 |
[More Information Needed]
|
| 70 |
|
|
@@ -135,7 +135,8 @@ translator("Translate to Zulu: The cow is eating grass.")
|
|
| 135 |
|
| 136 |
#### Training Hyperparameters
|
| 137 |
|
| 138 |
-
- **Training regime:**
|
|
|
|
| 139 |
- peft_config = LoraConfig(
|
| 140 |
lora_alpha=16,
|
| 141 |
lora_dropout=0.05,
|
|
@@ -157,7 +158,7 @@ translator("Translate to Zulu: The cow is eating grass.")
|
|
| 157 |
warmup_steps=100,
|
| 158 |
lr_scheduler_type="linear",
|
| 159 |
)
|
| 160 |
-
|
| 161 |
|
| 162 |
#### Speeds, Sizes, Times [optional]
|
| 163 |
|
|
|
|
| 34 |
### Model Description
|
| 35 |
|
| 36 |
<!-- Provide a longer summary of what this model is. -->
|
| 37 |
+
dsfsi/OMT-LR-Mistral7b, model, was fine-tuned for 31 GPU days from base model mistralai/Mistral-7B-v0.1. The model was fine-tuned in efforts to improve translation task
|
| 38 |
for large language model in regard to low resourced morphologically rich African languages using custom prompt.
|
| 39 |
|
| 40 |
|
|
|
|
| 64 |
### Direct Use
|
| 65 |
|
| 66 |
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
|
| 67 |
+
You can download the model, dsfsi/OMT-LR-Mistral7b, and prompt it to translate English sentences to Zulu and Xhosa sentences.
|
| 68 |
|
| 69 |
[More Information Needed]
|
| 70 |
|
|
|
|
| 135 |
|
| 136 |
#### Training Hyperparameters
|
| 137 |
|
| 138 |
+
- **Training regime:**
|
| 139 |
+
```python
|
| 140 |
- peft_config = LoraConfig(
|
| 141 |
lora_alpha=16,
|
| 142 |
lora_dropout=0.05,
|
|
|
|
| 158 |
warmup_steps=100,
|
| 159 |
lr_scheduler_type="linear",
|
| 160 |
)
|
| 161 |
+
```
|
| 162 |
|
| 163 |
#### Speeds, Sizes, Times [optional]
|
| 164 |
|