danielhanchen commited on
Commit
692a9a5
·
verified ·
1 Parent(s): 0412dc4

Add files using upload-large-folder tool

Browse files
config.json CHANGED
@@ -3,24 +3,26 @@
3
  "Qwen2ForCausalLM"
4
  ],
5
  "attention_dropout": 0.0,
6
- "bos_token_id": 151643,
7
  "eos_token_id": 151645,
8
  "hidden_act": "silu",
9
  "hidden_size": 5120,
10
  "initializer_range": 0.02,
11
  "intermediate_size": 27648,
12
- "max_position_embeddings": 40960,
13
  "max_window_layers": 64,
14
  "model_type": "qwen2",
15
  "num_attention_heads": 40,
16
  "num_hidden_layers": 64,
17
  "num_key_value_heads": 8,
 
18
  "rms_norm_eps": 1e-05,
 
19
  "rope_theta": 1000000.0,
20
  "sliding_window": 32768,
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "bfloat16",
23
- "transformers_version": "4.43.1",
 
24
  "use_cache": true,
25
  "use_sliding_window": false,
26
  "vocab_size": 152064
 
3
  "Qwen2ForCausalLM"
4
  ],
5
  "attention_dropout": 0.0,
 
6
  "eos_token_id": 151645,
7
  "hidden_act": "silu",
8
  "hidden_size": 5120,
9
  "initializer_range": 0.02,
10
  "intermediate_size": 27648,
11
+ "max_position_embeddings": 32768,
12
  "max_window_layers": 64,
13
  "model_type": "qwen2",
14
  "num_attention_heads": 40,
15
  "num_hidden_layers": 64,
16
  "num_key_value_heads": 8,
17
+ "pad_token_id": 151654,
18
  "rms_norm_eps": 1e-05,
19
+ "rope_scaling": null,
20
  "rope_theta": 1000000.0,
21
  "sliding_window": 32768,
22
  "tie_word_embeddings": false,
23
  "torch_dtype": "bfloat16",
24
+ "transformers_version": "4.51.3",
25
+ "unsloth_fixed": true,
26
  "use_cache": true,
27
  "use_sliding_window": false,
28
  "vocab_size": 152064
generation_config.json CHANGED
@@ -5,10 +5,10 @@
5
  151645,
6
  151643
7
  ],
8
- "pad_token_id": 151643,
9
- "repetition_penalty": 1.0,
10
  "temperature": 0.6,
11
  "top_k": 40,
12
  "top_p": 0.95,
13
- "transformers_version": "4.45.2"
14
  }
 
5
  151645,
6
  151643
7
  ],
8
+ "max_length": 32768,
9
+ "pad_token_id": 151654,
10
  "temperature": 0.6,
11
  "top_k": 40,
12
  "top_p": 0.95,
13
+ "transformers_version": "4.51.3"
14
  }
special_tokens_map.json CHANGED
@@ -22,7 +22,7 @@
22
  "single_word": false
23
  },
24
  "pad_token": {
25
- "content": "<|endoftext|>",
26
  "lstrip": false,
27
  "normalized": false,
28
  "rstrip": false,
 
22
  "single_word": false
23
  },
24
  "pad_token": {
25
+ "content": "<|vision_pad|>",
26
  "lstrip": false,
27
  "normalized": false,
28
  "rstrip": false,
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:19564a48c4f71a2a1b937cce34c737a1e662b171c5f5d7edf641a15cd896f07d
3
- size 7032399
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aeb13307a71acd8fe81861d94ad54ab689df773318809eed3cbe794b4492dae4
3
+ size 11422654
tokenizer_config.json CHANGED
@@ -231,8 +231,10 @@
231
  "clean_up_tokenization_spaces": false,
232
  "eos_token": "<|im_end|>",
233
  "errors": "replace",
234
- "model_max_length": 131072,
235
- "pad_token": "<|endoftext|>",
 
 
236
  "split_special_tokens": false,
237
  "tokenizer_class": "Qwen2Tokenizer",
238
  "unk_token": null
 
231
  "clean_up_tokenization_spaces": false,
232
  "eos_token": "<|im_end|>",
233
  "errors": "replace",
234
+ "extra_special_tokens": {},
235
+ "model_max_length": 32768,
236
+ "pad_token": "<|vision_pad|>",
237
+ "padding_side": "left",
238
  "split_special_tokens": false,
239
  "tokenizer_class": "Qwen2Tokenizer",
240
  "unk_token": null