Training in progress, step 301
Browse files- chat_template.jinja +1 -10
- model.safetensors +1 -1
- special_tokens_map.json +2 -8
- tokenizer.json +1 -1
- tokenizer_config.json +6 -2
- training_args.bin +1 -1
chat_template.jinja
CHANGED
@@ -1,10 +1 @@
|
|
1 |
-
|
2 |
-
{%- for message in messages %}
|
3 |
-
{%- if message['role'] == 'system' %}
|
4 |
-
{{ '<|system|>' + ' ' + message['content'].strip() }}
|
5 |
-
{%- elif message['role'] == 'user' %}
|
6 |
-
{{ '<|instruction|>' + ' ' + message['content'].strip() }}
|
7 |
-
{%- elif message['role'] == 'assistant' %}
|
8 |
-
{{ '<|response|>' + ' ' + message['content'].strip() }}
|
9 |
-
{%- endif %}
|
10 |
-
{%- endfor %}
|
|
|
1 |
+
{{ bos_token }}{% for message in messages %}{% if message['role'] == 'user' %}{{ '<|start_header_id|>user<|end_header_id|> ' + message['content'] + ' <|eot_id|>' }}{% elif message['role'] == 'assistant' %}{{ '<|start_header_id|>assistant<|end_header_id|> ' + message['content'] + ' <|eot_id|>' }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|> ' }}{% endif %}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 2013432032
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dff691ae8c26f29915c826edf637c3238a0409808304a965811c84b6d215c76e
|
3 |
size 2013432032
|
special_tokens_map.json
CHANGED
@@ -6,14 +6,8 @@
|
|
6 |
"rstrip": false,
|
7 |
"single_word": false
|
8 |
},
|
9 |
-
"eos_token":
|
10 |
-
|
11 |
-
"lstrip": false,
|
12 |
-
"normalized": false,
|
13 |
-
"rstrip": false,
|
14 |
-
"single_word": false
|
15 |
-
},
|
16 |
-
"pad_token": "<|endoftext|>",
|
17 |
"unk_token": {
|
18 |
"content": "<unk>",
|
19 |
"lstrip": false,
|
|
|
6 |
"rstrip": false,
|
7 |
"single_word": false
|
8 |
},
|
9 |
+
"eos_token": "<|eot_id|>",
|
10 |
+
"pad_token": "<|eot_id|>",
|
|
|
|
|
|
|
|
|
|
|
|
|
11 |
"unk_token": {
|
12 |
"content": "<unk>",
|
13 |
"lstrip": false,
|
tokenizer.json
CHANGED
@@ -2,7 +2,7 @@
|
|
2 |
"version": "1.0",
|
3 |
"truncation": {
|
4 |
"direction": "Right",
|
5 |
-
"max_length":
|
6 |
"strategy": "LongestFirst",
|
7 |
"stride": 0
|
8 |
},
|
|
|
2 |
"version": "1.0",
|
3 |
"truncation": {
|
4 |
"direction": "Right",
|
5 |
+
"max_length": 256,
|
6 |
"strategy": "LongestFirst",
|
7 |
"stride": 0
|
8 |
},
|
tokenizer_config.json
CHANGED
@@ -83,11 +83,15 @@
|
|
83 |
},
|
84 |
"bos_token": "<sos>",
|
85 |
"clean_up_tokenization_spaces": true,
|
86 |
-
"eos_token": "<|
|
87 |
"extra_special_tokens": {},
|
|
|
88 |
"model_max_length": 1000000000000000019884624838656,
|
89 |
-
"pad_token": "<|
|
|
|
90 |
"tokenizer_class": "PreTrainedTokenizer",
|
91 |
"trim_offsets": true,
|
|
|
|
|
92 |
"unk_token": "<unk>"
|
93 |
}
|
|
|
83 |
},
|
84 |
"bos_token": "<sos>",
|
85 |
"clean_up_tokenization_spaces": true,
|
86 |
+
"eos_token": "<|eot_id|>",
|
87 |
"extra_special_tokens": {},
|
88 |
+
"max_length": 512,
|
89 |
"model_max_length": 1000000000000000019884624838656,
|
90 |
+
"pad_token": "<|eot_id|>",
|
91 |
+
"stride": 0,
|
92 |
"tokenizer_class": "PreTrainedTokenizer",
|
93 |
"trim_offsets": true,
|
94 |
+
"truncation_side": "right",
|
95 |
+
"truncation_strategy": "longest_first",
|
96 |
"unk_token": "<unk>"
|
97 |
}
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 5240
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2916ab9e1755803b1193e934ee58981fcce2b041c1bb51d121df465f7d8fd9b2
|
3 |
size 5240
|