danielhanchen commited on
Commit
1352d51
·
verified ·
1 Parent(s): aa1928f

Add files using upload-large-folder tool

Browse files
.gitattributes CHANGED
@@ -33,4 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
- tokenizer.json filter=lfs diff=lfs merge=lfs -text
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
LICENSE ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ MIT License
2
+
3
+ Copyright (c) 2025 Zhipu AI
4
+
5
+ Permission is hereby granted, free of charge, to any person obtaining a copy
6
+ of this software and associated documentation files (the "Software"), to deal
7
+ in the Software without restriction, including without limitation the rights
8
+ to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
9
+ copies of the Software, and to permit persons to whom the Software is
10
+ furnished to do so, subject to the following conditions:
11
+
12
+ The above copyright notice and this permission notice shall be included in all
13
+ copies or substantial portions of the Software.
14
+
15
+ THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16
+ IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17
+ FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
18
+ AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
19
+ LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
20
+ OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
21
+ SOFTWARE.
config.json CHANGED
@@ -4,7 +4,11 @@
4
  ],
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
7
- "eos_token_id": 151336,
 
 
 
 
8
  "head_dim": 128,
9
  "hidden_act": "silu",
10
  "hidden_size": 6144,
@@ -15,14 +19,13 @@
15
  "num_attention_heads": 48,
16
  "num_hidden_layers": 61,
17
  "num_key_value_heads": 2,
18
- "pad_token_id": 151330,
19
  "partial_rotary_factor": 0.5,
20
  "rms_norm_eps": 1e-05,
21
  "rope_theta": 10000.0,
22
  "tie_word_embeddings": false,
23
  "torch_dtype": "bfloat16",
24
  "transformers_version": "4.52.0.dev0",
25
- "unsloth_fixed": true,
26
  "use_cache": true,
27
  "vocab_size": 151552
28
  }
 
4
  ],
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
7
+ "eos_token_id": [
8
+ 151329,
9
+ 151336,
10
+ 151338
11
+ ],
12
  "head_dim": 128,
13
  "hidden_act": "silu",
14
  "hidden_size": 6144,
 
19
  "num_attention_heads": 48,
20
  "num_hidden_layers": 61,
21
  "num_key_value_heads": 2,
22
+ "pad_token_id": 151329,
23
  "partial_rotary_factor": 0.5,
24
  "rms_norm_eps": 1e-05,
25
  "rope_theta": 10000.0,
26
  "tie_word_embeddings": false,
27
  "torch_dtype": "bfloat16",
28
  "transformers_version": "4.52.0.dev0",
 
29
  "use_cache": true,
30
  "vocab_size": 151552
31
  }
generation_config.json CHANGED
@@ -5,7 +5,6 @@
5
  151336,
6
  151338
7
  ],
8
- "max_length": 32768,
9
- "pad_token_id": 151330,
10
  "transformers_version": "4.52.0.dev0"
11
  }
 
5
  151336,
6
  151338
7
  ],
8
+ "pad_token_id": 151329,
 
9
  "transformers_version": "4.52.0.dev0"
10
  }
special_tokens_map.json CHANGED
@@ -16,14 +16,14 @@
16
  "<|end_of_video|>"
17
  ],
18
  "eos_token": {
19
- "content": "<|user|>",
20
  "lstrip": false,
21
  "normalized": false,
22
  "rstrip": false,
23
  "single_word": false
24
  },
25
  "pad_token": {
26
- "content": "[MASK]",
27
  "lstrip": false,
28
  "normalized": false,
29
  "rstrip": false,
 
16
  "<|end_of_video|>"
17
  ],
18
  "eos_token": {
19
+ "content": "<|endoftext|>",
20
  "lstrip": false,
21
  "normalized": false,
22
  "rstrip": false,
23
  "single_word": false
24
  },
25
  "pad_token": {
26
+ "content": "<|endoftext|>",
27
  "lstrip": false,
28
  "normalized": false,
29
  "rstrip": false,
tokenizer_config.json CHANGED
@@ -129,7 +129,7 @@
129
  "<|begin_of_video|>",
130
  "<|end_of_video|>"
131
  ],
132
- "bos_token": null,
133
  "clean_up_tokenization_spaces": false,
134
  "do_lower_case": false,
135
  "eos_token": "<|user|>",
@@ -138,11 +138,9 @@
138
  "input_ids",
139
  "attention_mask"
140
  ],
141
- "model_max_length": 32768,
142
- "pad_token": "[MASK]",
143
  "padding_side": "left",
144
  "remove_space": false,
145
- "tokenizer_class": "PreTrainedTokenizer",
146
- "unk_token": null,
147
- "chat_template": "[gMASK]<sop>\n{%- if tools -%}\n<|system|>\n# \u53ef\u7528\u5de5\u5177\n{% for tool in tools %}\n {%- set function = tool.function if tool.get(\"function\") else tool %}\n\n## {{ function.name }}\n\n{{ function | tojson(indent=4, ensure_ascii=False) }}\n\u5728\u8c03\u7528\u4e0a\u8ff0\u51fd\u6570\u65f6\uff0c\u8bf7\u4f7f\u7528 Json \u683c\u5f0f\u8868\u793a\u8c03\u7528\u7684\u53c2\u6570\u3002\n{%- endfor %}\n{%- endif -%}\n\n{%- for msg in messages %}\n {%- if msg.role == 'system' %}\n<|system|>\n{{ msg.content }}\n {%- endif %}\n{%- endfor %}\n\n{%- for message in messages if message.role != 'system' %}\n {%- set role = message['role'] %}\n {%- set content = message['content'] %}\n {%- set meta = message.get(\"metadata\", \"\") %}\n\n {%- if role == 'user' %}\n<|user|>\n{{ content }}\n {%- elif role == 'assistant' and not meta %}\n<|assistant|>\n{{ content }}\n {%- elif role == 'assistant' and meta %}\n<|assistant|>{{ meta }}\n{{ content }}\n {%- elif role == 'observation' %}\n<|observation|>\n{{ content }}\n {%- endif %}\n{%- endfor %}\n{% if add_generation_prompt %}<|assistant|>{% endif %}"
148
- }
 
129
  "<|begin_of_video|>",
130
  "<|end_of_video|>"
131
  ],
132
+ "chat_template": "[gMASK]<sop>{%- if tools -%}<|system|>\n# 可用工具\n{% for tool in tools %}{%- set function = tool.function if tool.get(\"function\") else tool %}\n\n## {{ function.name }}\n\n{{ function | tojson(indent=4, ensure_ascii=False) }}\n在调用上述函数时,请使用 Json 格式表示调用的参数。{%- endfor %}{%- endif -%}{%- for msg in messages %}{%- if msg.role == 'system' %}<|system|>\n{{ msg.content }}{%- endif %}{%- endfor %}{%- for message in messages if message.role != 'system' %}{%- set role = message['role'] %}{%- set content = message['content'] %}{%- set meta = message.get(\"metadata\", \"\") %}{%- if role == 'user' %}<|user|>\n{{ content }}{%- elif role == 'assistant' and not meta %}<|assistant|>\n{{ content }}{%- elif role == 'assistant' and meta %}<|assistant|>{{ meta }} \n{{ content }}{%- elif role == 'observation' %}<|observation|>\n{{ content }}{%- endif %}{%- endfor %}{% if add_generation_prompt %}<|assistant|>{% endif %}",
133
  "clean_up_tokenization_spaces": false,
134
  "do_lower_case": false,
135
  "eos_token": "<|user|>",
 
138
  "input_ids",
139
  "attention_mask"
140
  ],
141
+ "model_max_length": 128000,
142
+ "pad_token": "<|endoftext|>",
143
  "padding_side": "left",
144
  "remove_space": false,
145
+ "tokenizer_class": "PreTrainedTokenizer"
146
+ }