add tokenizer
Browse files- merges.txt +0 -0
- tokenizer.json +0 -0
- tokenizer_config.json +1 -8
- vocab.json +0 -0
merges.txt
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer.json
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
CHANGED
@@ -58,14 +58,7 @@
|
|
58 |
"single_word": false
|
59 |
},
|
60 |
"sep": "<sep>",
|
61 |
-
"sep_token":
|
62 |
-
"__type": "AddedToken",
|
63 |
-
"content": "<sep>",
|
64 |
-
"lstrip": false,
|
65 |
-
"normalized": true,
|
66 |
-
"rstrip": false,
|
67 |
-
"single_word": false
|
68 |
-
},
|
69 |
"special_tokens_map_file": "/home/ashraf.haddad/.cache/huggingface/transformers/4b7a3619321a39f6780e0c775802e3523e52c1efd3a46e6d1baac9e1e8e234e6.898eb95aac9bb57440c2f57caa963ae18b9b10ba4731cc81020283286b0391fc",
|
70 |
"tokenizer_class": "BartTokenizer",
|
71 |
"trim_offsets": true,
|
|
|
58 |
"single_word": false
|
59 |
},
|
60 |
"sep": "<sep>",
|
61 |
+
"sep_token": "<sep>",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
62 |
"special_tokens_map_file": "/home/ashraf.haddad/.cache/huggingface/transformers/4b7a3619321a39f6780e0c775802e3523e52c1efd3a46e6d1baac9e1e8e234e6.898eb95aac9bb57440c2f57caa963ae18b9b10ba4731cc81020283286b0391fc",
|
63 |
"tokenizer_class": "BartTokenizer",
|
64 |
"trim_offsets": true,
|
vocab.json
CHANGED
The diff for this file is too large to render.
See raw diff
|
|