{ "added_tokens_decoder": {}, "additional_special_tokens": [ "<|startoftext|>", "<|extra_0|>", "<|extra_4|>", "<|extra_5|>", "<|eos|>" ], "architectures": [ "GPT2LMHeadModel" ], "auto_map": { "AutoTokenizer": [ "tokenization_hy.HYTokenizer", null ] }, "clean_up_tokenization_spaces": false, "eos_token": "<|eos|>", "extra_special_tokens": {}, "model_max_length": 1048576, "model_type": "gpt2", "pad_token": "<|pad|>", "tokenizer_class": "HYTokenizer" }