Update README.md (#2)
Browse files- Update README.md (69dc0ee36de8192234467e965eb72f20cde3770e)
Co-authored-by: Pengzhi Gao <[email protected]>
README.md
CHANGED
@@ -44,15 +44,9 @@ GemmaX2-28-2B-Pretrain is a language model developed through continual pretraini
|
|
44 |
|
45 |
- **Developed by:** Xiaomi
|
46 |
- **Model type:** GemmaX2-28-2B-Pretrain is obtained by continually pretraining Gemma2-2B on a large amount of monolingual and parallel data. Subsequently, GemmaX2-28-2B-v0.1 is derived through supervised finetuning on a small set of high-quality translation instruction data.
|
47 |
-
- **Languages:** Arabic, Bengali, Czech, German, English, Spanish, Persian, French, Hebrew, Hindi, Indonesian, Italian, Japanese, Khmer, Korean, Lao, Malay, Burmese, Dutch, polish, Portuguese, Russian, Thai, Tagalog, Turkish, Urdu, Vietnamese, Chinese.
|
48 |
|
49 |
-
|
50 |
-
|
51 |
-
- paper: [Multilingual Machine Translation with Open Large Language Models at Practical Scale: An Empirical Study](https://arxiv.org/pdf/2502.02481)
|
52 |
-
|
53 |
-
## Model Performance
|
54 |
-
|
55 |
-

|
56 |
|
57 |
|
58 |
## Training Data
|
|
|
44 |
|
45 |
- **Developed by:** Xiaomi
|
46 |
- **Model type:** GemmaX2-28-2B-Pretrain is obtained by continually pretraining Gemma2-2B on a large amount of monolingual and parallel data. Subsequently, GemmaX2-28-2B-v0.1 is derived through supervised finetuning on a small set of high-quality translation instruction data.
|
47 |
+
- **Languages:** Arabic, Bengali, Czech, German, English, Spanish, Persian, French, Hebrew, Hindi, Indonesian, Italian, Japanese, Khmer, Korean, Lao, Malay, Burmese, Dutch, polish, Portuguese, Russian, Thai, Tagalog, Turkish, Urdu, Vietnamese, Chinese.
|
48 |
|
49 |
+
**Note that GemmaX2-28-2B-Pretrain is NOT translation model.**
|
|
|
|
|
|
|
|
|
|
|
|
|
50 |
|
51 |
|
52 |
## Training Data
|