Update README.md
Browse files
README.md
CHANGED
@@ -34,7 +34,7 @@ The model is a result of fine-tuning Mistral-7B-v0.1 on a down stream task, in l
|
|
34 |
### Model Description
|
35 |
|
36 |
<!-- Provide a longer summary of what this model is. -->
|
37 |
-
dsfsi/
|
38 |
for large language model in regard to low resourced morphologically rich African languages using custom prompt.
|
39 |
|
40 |
|
@@ -64,7 +64,7 @@ be used in the Education industry to teach core subjects in native South African
|
|
64 |
### Direct Use
|
65 |
|
66 |
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
|
67 |
-
You can download the model, dsfsi/
|
68 |
|
69 |
[More Information Needed]
|
70 |
|
@@ -135,7 +135,8 @@ translator("Translate to Zulu: The cow is eating grass.")
|
|
135 |
|
136 |
#### Training Hyperparameters
|
137 |
|
138 |
-
- **Training regime:**
|
|
|
139 |
- peft_config = LoraConfig(
|
140 |
lora_alpha=16,
|
141 |
lora_dropout=0.05,
|
@@ -157,7 +158,7 @@ translator("Translate to Zulu: The cow is eating grass.")
|
|
157 |
warmup_steps=100,
|
158 |
lr_scheduler_type="linear",
|
159 |
)
|
160 |
-
|
161 |
|
162 |
#### Speeds, Sizes, Times [optional]
|
163 |
|
|
|
34 |
### Model Description
|
35 |
|
36 |
<!-- Provide a longer summary of what this model is. -->
|
37 |
+
dsfsi/OMT-LR-Mistral7b, model, was fine-tuned for 31 GPU days from base model mistralai/Mistral-7B-v0.1. The model was fine-tuned in efforts to improve translation task
|
38 |
for large language model in regard to low resourced morphologically rich African languages using custom prompt.
|
39 |
|
40 |
|
|
|
64 |
### Direct Use
|
65 |
|
66 |
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
|
67 |
+
You can download the model, dsfsi/OMT-LR-Mistral7b, and prompt it to translate English sentences to Zulu and Xhosa sentences.
|
68 |
|
69 |
[More Information Needed]
|
70 |
|
|
|
135 |
|
136 |
#### Training Hyperparameters
|
137 |
|
138 |
+
- **Training regime:**
|
139 |
+
```python
|
140 |
- peft_config = LoraConfig(
|
141 |
lora_alpha=16,
|
142 |
lora_dropout=0.05,
|
|
|
158 |
warmup_steps=100,
|
159 |
lr_scheduler_type="linear",
|
160 |
)
|
161 |
+
```
|
162 |
|
163 |
#### Speeds, Sizes, Times [optional]
|
164 |
|