PKhoboko commited on
Commit
60c69b9
·
verified ·
1 Parent(s): 4417feb

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -4
README.md CHANGED
@@ -34,7 +34,7 @@ The model is a result of fine-tuning Mistral-7B-v0.1 on a down stream task, in l
34
  ### Model Description
35
 
36
  <!-- Provide a longer summary of what this model is. -->
37
- dsfsi/mistral-7b-custom_prompt_long_short_31_gpu_days, model, was fine-tuned for 31 GPU days from base model mistralai/Mistral-7B-v0.1. The model was fine-tuned in efforts to improve translation task
38
  for large language model in regard to low resourced morphologically rich African languages using custom prompt.
39
 
40
 
@@ -64,7 +64,7 @@ be used in the Education industry to teach core subjects in native South African
64
  ### Direct Use
65
 
66
  <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
67
- You can download the model, dsfsi/mistral-7b-custom_prompt_long_short_31_gpu_days, and prompt it to translate English sentences to Zulu and Xhosa sentences.
68
 
69
  [More Information Needed]
70
 
@@ -135,7 +135,8 @@ translator("Translate to Zulu: The cow is eating grass.")
135
 
136
  #### Training Hyperparameters
137
 
138
- - **Training regime:**
 
139
  - peft_config = LoraConfig(
140
  lora_alpha=16,
141
  lora_dropout=0.05,
@@ -157,7 +158,7 @@ translator("Translate to Zulu: The cow is eating grass.")
157
  warmup_steps=100,
158
  lr_scheduler_type="linear",
159
  )
160
-
161
 
162
  #### Speeds, Sizes, Times [optional]
163
 
 
34
  ### Model Description
35
 
36
  <!-- Provide a longer summary of what this model is. -->
37
+ dsfsi/OMT-LR-Mistral7b, model, was fine-tuned for 31 GPU days from base model mistralai/Mistral-7B-v0.1. The model was fine-tuned in efforts to improve translation task
38
  for large language model in regard to low resourced morphologically rich African languages using custom prompt.
39
 
40
 
 
64
  ### Direct Use
65
 
66
  <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
67
+ You can download the model, dsfsi/OMT-LR-Mistral7b, and prompt it to translate English sentences to Zulu and Xhosa sentences.
68
 
69
  [More Information Needed]
70
 
 
135
 
136
  #### Training Hyperparameters
137
 
138
+ - **Training regime:**
139
+ ```python
140
  - peft_config = LoraConfig(
141
  lora_alpha=16,
142
  lora_dropout=0.05,
 
158
  warmup_steps=100,
159
  lr_scheduler_type="linear",
160
  )
161
+ ```
162
 
163
  #### Speeds, Sizes, Times [optional]
164