danielhanchen commited on
Commit
e94a99e
·
verified ·
1 Parent(s): 9f4b82c

Add files using upload-large-folder tool

Browse files
config.json CHANGED
@@ -4,7 +4,6 @@
4
  ],
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
7
- "bos_token_id": 151643,
8
  "decoder_sparse_step": 1,
9
  "eos_token_id": 151645,
10
  "head_dim": 128,
@@ -24,6 +23,7 @@
24
  "num_hidden_layers": 48,
25
  "num_key_value_heads": 4,
26
  "output_router_logits": false,
 
27
  "rms_norm_eps": 1e-06,
28
  "rope_scaling": null,
29
  "rope_theta": 1000000.0,
@@ -31,8 +31,9 @@
31
  "sliding_window": null,
32
  "tie_word_embeddings": false,
33
  "torch_dtype": "bfloat16",
34
- "transformers_version": "4.51.0",
 
35
  "use_cache": true,
36
  "use_sliding_window": false,
37
  "vocab_size": 151936
38
- }
 
4
  ],
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
 
7
  "decoder_sparse_step": 1,
8
  "eos_token_id": 151645,
9
  "head_dim": 128,
 
23
  "num_hidden_layers": 48,
24
  "num_key_value_heads": 4,
25
  "output_router_logits": false,
26
+ "pad_token_id": 151654,
27
  "rms_norm_eps": 1e-06,
28
  "rope_scaling": null,
29
  "rope_theta": 1000000.0,
 
31
  "sliding_window": null,
32
  "tie_word_embeddings": false,
33
  "torch_dtype": "bfloat16",
34
+ "transformers_version": "4.52.0.dev0",
35
+ "unsloth_fixed": true,
36
  "use_cache": true,
37
  "use_sliding_window": false,
38
  "vocab_size": 151936
39
+ }
generation_config.json CHANGED
@@ -1,13 +1,14 @@
1
  {
2
- "bos_token_id": 151643,
3
- "do_sample": true,
4
- "eos_token_id": [
5
- 151645,
6
- 151643
7
- ],
8
- "pad_token_id": 151643,
9
- "temperature": 0.6,
10
- "top_k": 20,
11
- "top_p": 0.95,
12
- "transformers_version": "4.51.0"
13
- }
 
 
1
  {
2
+ "bos_token_id": 151643,
3
+ "do_sample": true,
4
+ "eos_token_id": [
5
+ 151645,
6
+ 151643
7
+ ],
8
+ "max_length": 40960,
9
+ "pad_token_id": 151654,
10
+ "temperature": 0.6,
11
+ "top_k": 20,
12
+ "top_p": 0.95,
13
+ "transformers_version": "4.52.0.dev0"
14
+ }
model-00001-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e80e793f3403fbc12b7360cea3b7488dd9e13bcfe53732e0d2f85700ab6a344
3
+ size 4997184968
model-00002-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:50754c62e928b29e2eaa2ade1e61624b27d24877ca37eda85a51de8a3e74e469
3
+ size 4997741608
model-00003-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26dea45e135f339e82140cfa1e7c06a8888873a36ef78405df92a5af3fd29695
3
+ size 4997742208
model-00004-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4b69db4edcad5460dfd30dbcce27478aeccb37d58c59ed58f7bb91cb5062c69
3
+ size 4997743184
model-00005-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:23cf0c76fe660de4143e0c9865c5ec98aa4d5d57298f89a831ceaa08a3def7b7
3
+ size 4997743184
model-00006-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1eb7cf589b67bac2cb47e9fb836bf55272aa694528447c5993c1b37c0df0d648
3
+ size 4997743184
model-00007-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d561d6abce11ff85a914ea00e3c8befada0070325a8e5c01d935fea0b85aafe1
3
+ size 4997743184
model-00008-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a3828fb7caa650fb58ada0d94702c4d54d8f120002d9d86926f4b82608ff249
3
+ size 4997743184
model-00009-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:732a117f266a266432e19fdb4e3d1a1b77965b90be98e3c4e4d054b94727944f
3
+ size 4997743184
model-00010-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed69f078d52a6cef3075d7b4c9547a0abcd62b1483a55d7711a594f7bb7ebefa
3
+ size 4997743184
model-00011-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:625947b70501b3f3e45c4196ef37947cbdde01fd2d4d5009be14173362125314
3
+ size 4997743184
model-00012-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:08bf9d9d3eb50aebd667033a2d18bde73f0558c1552e92a1ed77a8563279fafa
3
+ size 4997743184
model-00013-of-00013.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4c8b6778746d3bea52fe78db8e7709d277e80ea48e1c8fe8940d928e6b7168b4
3
+ size 1094220288
model.safetensors.index.json CHANGED
The diff for this file is too large to render. See raw diff
 
special_tokens_map.json CHANGED
@@ -22,7 +22,7 @@
22
  "single_word": false
23
  },
24
  "pad_token": {
25
- "content": "<|endoftext|>",
26
  "lstrip": false,
27
  "normalized": false,
28
  "rstrip": false,
 
22
  "single_word": false
23
  },
24
  "pad_token": {
25
+ "content": "<|vision_pad|>",
26
  "lstrip": false,
27
  "normalized": false,
28
  "rstrip": false,
tokenizer_config.json CHANGED
@@ -231,8 +231,9 @@
231
  "eos_token": "<|im_end|>",
232
  "errors": "replace",
233
  "extra_special_tokens": {},
234
- "model_max_length": 131072,
235
- "pad_token": "<|endoftext|>",
 
236
  "split_special_tokens": false,
237
  "tokenizer_class": "Qwen2Tokenizer",
238
  "unk_token": null,
 
231
  "eos_token": "<|im_end|>",
232
  "errors": "replace",
233
  "extra_special_tokens": {},
234
+ "model_max_length": 40960,
235
+ "pad_token": "<|vision_pad|>",
236
+ "padding_side": "left",
237
  "split_special_tokens": false,
238
  "tokenizer_class": "Qwen2Tokenizer",
239
  "unk_token": null,