seastar105 commited on
Commit
9abebda
·
verified ·
1 Parent(s): 5839bc8

Upload tokenizer

Browse files
special_tokens_map.json CHANGED
@@ -1,3 +1,4 @@
1
  {
2
- "eos_token": "<|endoftext|>"
 
3
  }
 
1
  {
2
+ "eos_token": "<|endoftext|>",
3
+ "pad_token": "<|endoftext|>"
4
  }
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -9,9 +9,11 @@
9
  "special": true
10
  }
11
  },
 
12
  "clean_up_tokenization_spaces": false,
13
  "eos_token": "<|endoftext|>",
14
  "extra_special_tokens": {},
15
  "model_max_length": 1000000000000000019884624838656,
 
16
  "tokenizer_class": "PreTrainedTokenizer"
17
  }
 
9
  "special": true
10
  }
11
  },
12
+ "bos_token": null,
13
  "clean_up_tokenization_spaces": false,
14
  "eos_token": "<|endoftext|>",
15
  "extra_special_tokens": {},
16
  "model_max_length": 1000000000000000019884624838656,
17
+ "pad_token": "<|endoftext|>",
18
  "tokenizer_class": "PreTrainedTokenizer"
19
  }