alirezamsh commited on
Commit
a9023d6
1 Parent(s): f2a2fc5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -117,7 +117,7 @@ datasets:
117
 
118
  SMaLL-100 is a compact and fast massively multilingual machine translation model covering more than 10K language pairs, that achieves competitive results with M2M-100 while being much smaller and faster. It is introduced in [this paper](https://arxiv.org/abs/2210.11621), and initially released in [this repository](https://github.com/alirezamshi/small100).
119
 
120
- The model architecture and config are the same as [M2M-100](https://huggingface.co/facebook/m2m100_418M/tree/main) implementation, but the tokenizer is modified to adjust language codes. So, you should load the tokenizer locally from tokenization_small100.py file for the moment.
121
 
122
  ```
123
  from transformers import M2M100ForConditionalGeneration
 
117
 
118
  SMaLL-100 is a compact and fast massively multilingual machine translation model covering more than 10K language pairs, that achieves competitive results with M2M-100 while being much smaller and faster. It is introduced in [this paper](https://arxiv.org/abs/2210.11621), and initially released in [this repository](https://github.com/alirezamshi/small100).
119
 
120
+ The model architecture and config are the same as [M2M-100](https://huggingface.co/facebook/m2m100_418M/tree/main) implementation, but the tokenizer is modified to adjust language codes. So, you should load the tokenizer locally from (tokenization_small100.py)[https://huggingface.co/alirezamsh/small100/blob/main/tokenization_small100.py] file for the moment.
121
 
122
  ```
123
  from transformers import M2M100ForConditionalGeneration