Update README.md
Browse files
README.md
CHANGED
|
@@ -21,7 +21,7 @@ base_model:
|
|
| 21 |
# Model Card for Model ID
|
| 22 |
|
| 23 |
<!-- Provide a quick summary of what the model is/does. -->
|
| 24 |
-
The model is a result of fine-tuning Mistral-7B-v0.1 on a down stream task, in low resourced setting. It is able to translate English sentences to Zulu and Xhosa
|
| 25 |
|
| 26 |
|
| 27 |
## Model Details
|
|
@@ -57,8 +57,8 @@ for large language model in regard to low resourced morphologically rich African
|
|
| 57 |
## Uses
|
| 58 |
|
| 59 |
The model can be used to translate Engslih to Zulu and Xhosa. With further improvement it can be used to translate domain specific infromation from English to Zulu and Xhosa,
|
| 60 |
-
thus it can be used to
|
| 61 |
-
be used in the Education industry to teach core subjects in native South African langauges thus can improve pupils' performance in
|
| 62 |
|
| 63 |
|
| 64 |
### Direct Use
|
|
@@ -66,13 +66,6 @@ be used in the Education industry to teach core subjects in native South African
|
|
| 66 |
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
|
| 67 |
You can download the model, dsfsi/OMT-LR-Mistral7b, and prompt it to translate English sentences to Zulu and Xhosa sentences.
|
| 68 |
|
| 69 |
-
[More Information Needed]
|
| 70 |
-
|
| 71 |
-
### Downstream Use [optional]
|
| 72 |
-
|
| 73 |
-
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
|
| 74 |
-
|
| 75 |
-
[More Information Needed]
|
| 76 |
|
| 77 |
### Out-of-Scope Use
|
| 78 |
|
|
@@ -130,7 +123,7 @@ translator("Translate to Zulu: The cow is eating grass.")
|
|
| 130 |
|
| 131 |
#### Preprocessing [optional]
|
| 132 |
|
| 133 |
-
|
| 134 |
|
| 135 |
|
| 136 |
#### Training Hyperparameters
|
|
@@ -160,12 +153,6 @@ translator("Translate to Zulu: The cow is eating grass.")
|
|
| 160 |
)
|
| 161 |
```
|
| 162 |
|
| 163 |
-
#### Speeds, Sizes, Times [optional]
|
| 164 |
-
|
| 165 |
-
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
|
| 166 |
-
|
| 167 |
-
[More Information Needed]
|
| 168 |
-
|
| 169 |
## Evaluation
|
| 170 |
|
| 171 |
<!-- This section describes the evaluation protocols and provides the results. -->
|
|
@@ -174,15 +161,11 @@ translator("Translate to Zulu: The cow is eating grass.")
|
|
| 174 |
|
| 175 |
#### Testing Data
|
| 176 |
|
| 177 |
-
|
| 178 |
-
|
| 179 |
-
[
|
| 180 |
-
|
| 181 |
-
#### Factors
|
| 182 |
-
|
| 183 |
-
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
|
| 184 |
|
| 185 |
-
[More Information Needed]
|
| 186 |
|
| 187 |
#### Metrics
|
| 188 |
|
|
@@ -259,15 +242,6 @@ Carbon emissions can be estimated using the [Machine Learning Impact calculator]
|
|
| 259 |
|
| 260 |
Khoboko, P. W., Marivate, V., & Sefara, J. (2025). Optimizing translation for low-resource languages: Efficient fine-tuning with custom prompt engineering in large language models. Machine Learning with Applications, 20, 100649.
|
| 261 |
|
| 262 |
-
## Glossary [optional]
|
| 263 |
-
|
| 264 |
-
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
|
| 265 |
-
|
| 266 |
-
[More Information Needed]
|
| 267 |
-
|
| 268 |
-
## More Information [optional]
|
| 269 |
-
|
| 270 |
-
[More Information Needed]
|
| 271 |
|
| 272 |
## Model Card Authors [optional]
|
| 273 |
|
|
|
|
| 21 |
# Model Card for Model ID
|
| 22 |
|
| 23 |
<!-- Provide a quick summary of what the model is/does. -->
|
| 24 |
+
The model is a result of fine-tuning Mistral-7B-v0.1 on a down stream task, in low resourced setting. It is able to translate English sentences to Zulu and Xhosa sentences.
|
| 25 |
|
| 26 |
|
| 27 |
## Model Details
|
|
|
|
| 57 |
## Uses
|
| 58 |
|
| 59 |
The model can be used to translate Engslih to Zulu and Xhosa. With further improvement it can be used to translate domain specific infromation from English to Zulu and Xhosa,
|
| 60 |
+
thus it can be used to translate research information that was written in English, in the agriculture industry, to small scale farmers that speak Zulu and Xhosa. Further, it can
|
| 61 |
+
be used in the Education industry to teach core subjects in native South African langauges thus can improve pupils' performance in these subjects.
|
| 62 |
|
| 63 |
|
| 64 |
### Direct Use
|
|
|
|
| 66 |
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
|
| 67 |
You can download the model, dsfsi/OMT-LR-Mistral7b, and prompt it to translate English sentences to Zulu and Xhosa sentences.
|
| 68 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 69 |
|
| 70 |
### Out-of-Scope Use
|
| 71 |
|
|
|
|
| 123 |
|
| 124 |
#### Preprocessing [optional]
|
| 125 |
|
| 126 |
+
Look at the repo to find out how the dataset clean up and preparation code in python.
|
| 127 |
|
| 128 |
|
| 129 |
#### Training Hyperparameters
|
|
|
|
| 153 |
)
|
| 154 |
```
|
| 155 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 156 |
## Evaluation
|
| 157 |
|
| 158 |
<!-- This section describes the evaluation protocols and provides the results. -->
|
|
|
|
| 161 |
|
| 162 |
#### Testing Data
|
| 163 |
|
| 164 |
+
- [nwu-ctext/autshumato](https://huggingface.co/datasets/nwu-ctext/autshumato)
|
| 165 |
+
- [Helsinki-NLP/opus-100](https://huggingface.co/datasets/Helsinki-NLP/opus-100)
|
| 166 |
+
- [WMT22](https://huggingface.co/datasets/wmt22)
|
| 167 |
+
- [gsarti/flores_101](https://huggingface.co/datasets/gsarti/flores_101)
|
|
|
|
|
|
|
|
|
|
| 168 |
|
|
|
|
| 169 |
|
| 170 |
#### Metrics
|
| 171 |
|
|
|
|
| 242 |
|
| 243 |
Khoboko, P. W., Marivate, V., & Sefara, J. (2025). Optimizing translation for low-resource languages: Efficient fine-tuning with custom prompt engineering in large language models. Machine Learning with Applications, 20, 100649.
|
| 244 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 245 |
|
| 246 |
## Model Card Authors [optional]
|
| 247 |
|