Thomas-Chou commited on
Commit
020eb83
·
verified ·
1 Parent(s): 1328365

Training in progress, epoch 1

Browse files
chat_template.jinja CHANGED
@@ -3,7 +3,7 @@
3
  {%- if messages[0]['role'] == 'system' %}
4
  {{- messages[0]['content'] }}
5
  {%- else %}
6
- {{- 'Please reason step by step, and put your final answer within \\boxed{}.' }}
7
  {%- endif %}
8
  {{- "\n\n# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>" }}
9
  {%- for tool in tools %}
@@ -15,7 +15,7 @@
15
  {%- if messages[0]['role'] == 'system' %}
16
  {{- '<|im_start|>system\n' + messages[0]['content'] + '<|im_end|>\n' }}
17
  {%- else %}
18
- {{- '<|im_start|>system\nPlease reason step by step, and put your final answer within \\boxed{}.<|im_end|>\n' }}
19
  {%- endif %}
20
  {%- endif %}
21
  {%- for message in messages %}
 
3
  {%- if messages[0]['role'] == 'system' %}
4
  {{- messages[0]['content'] }}
5
  {%- else %}
6
+ {{- 'You are Open-R1, a language model trained by Hugging Face to help users. Your role as an assistant involves thoroughly exploring questions through a systematic thinking process before providing the final precise and accurate solutions. This requires engaging in a comprehensive cycle of analysis, summarizing, exploration, reassessment, reflection, backtracing, and iteration to develop well-considered thinking process. Please structure your response into two main sections: Thought and Solution using the specified format: <think> Thought section </think> Solution section. In the Thought section, detail your reasoning process in steps. Each step should include detailed considerations such as analysing questions, summarizing relevant findings, brainstorming new ideas, verifying the accuracy of the current steps, refining any errors, and revisiting previous steps. In the Solution section, based on various attempts, explorations, and reflections from the Thought section, systematically present the final solution that you deem correct. The Solution section should be logical, accurate, and concise and detail necessary steps needed to reach the conclusion. Now, try to solve the following question through the above guidelines.' }}
7
  {%- endif %}
8
  {{- "\n\n# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>" }}
9
  {%- for tool in tools %}
 
15
  {%- if messages[0]['role'] == 'system' %}
16
  {{- '<|im_start|>system\n' + messages[0]['content'] + '<|im_end|>\n' }}
17
  {%- else %}
18
+ {{- '<|im_start|>system\nYou are Open-R1, a language model trained by Hugging Face to help users. Your role as an assistant involves thoroughly exploring questions through a systematic thinking process before providing the final precise and accurate solutions. This requires engaging in a comprehensive cycle of analysis, summarizing, exploration, reassessment, reflection, backtracing, and iteration to develop well-considered thinking process. Please structure your response into two main sections: Thought and Solution using the specified format: <think> Thought section </think> Solution section. In the Thought section, detail your reasoning process in steps. Each step should include detailed considerations such as analysing questions, summarizing relevant findings, brainstorming new ideas, verifying the accuracy of the current steps, refining any errors, and revisiting previous steps. In the Solution section, based on various attempts, explorations, and reflections from the Thought section, systematically present the final solution that you deem correct. The Solution section should be logical, accurate, and concise and detail necessary steps needed to reach the conclusion. Now, try to solve the following question through the above guidelines.<|im_end|>\n' }}
19
  {%- endif %}
20
  {%- endif %}
21
  {%- for message in messages %}
config.json CHANGED
@@ -22,7 +22,7 @@
22
  "tie_word_embeddings": false,
23
  "torch_dtype": "bfloat16",
24
  "transformers_version": "4.52.3",
25
- "use_cache": true,
26
  "use_sliding_window": false,
27
  "vocab_size": 152064
28
  }
 
22
  "tie_word_embeddings": false,
23
  "torch_dtype": "bfloat16",
24
  "transformers_version": "4.52.3",
25
+ "use_cache": false,
26
  "use_sliding_window": false,
27
  "vocab_size": 152064
28
  }
model-00001-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:22dede961cd25c99d99dba67bf0e052c031e90105c4524912a6f02e8ff665a4e
3
  size 4877660776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:17f4b4acd013c1a049ddb0daa84c37b071ab95da0eb1161d8b4811a12094eb64
3
  size 4877660776
model-00002-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2ea67a0b8864386be73477bc08a531c9e6843267b63466a21e2c50229fe51bc5
3
  size 4932751008
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f6557d447c8e243f4862d8ce20fe2a398096514883f9864bf4932fe0490a98d
3
  size 4932751008
model-00003-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fde9238b2b2a63557ce6a674cd1a0e0629d2276e83f6a857c39364a71ddc3ffe
3
  size 4330865200
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c07af1346d5464c4a7c1c572a032f5a32b6cca0883d3bb356d6ecdee6401a7d8
3
  size 4330865200
model-00004-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:97ebe44b7924488148b753f3d5fc49fe2050746aaf098bb26f9c3effa5d2541c
3
  size 1089994880
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a0c2e4bfd9a940c9c6b192a4b3d582d65947ccdf124b8e8c4dc4e11d26c1ad0
3
  size 1089994880
special_tokens_map.json CHANGED
@@ -14,13 +14,7 @@
14
  "<|image_pad|>",
15
  "<|video_pad|>"
16
  ],
17
- "eos_token": {
18
- "content": "<|endoftext|>",
19
- "lstrip": false,
20
- "normalized": false,
21
- "rstrip": false,
22
- "single_word": false
23
- },
24
  "pad_token": {
25
  "content": "<|endoftext|>",
26
  "lstrip": false,
 
14
  "<|image_pad|>",
15
  "<|video_pad|>"
16
  ],
17
+ "eos_token": "<|im_end|>",
 
 
 
 
 
 
18
  "pad_token": {
19
  "content": "<|endoftext|>",
20
  "lstrip": false,
tokenizer_config.json CHANGED
@@ -196,7 +196,7 @@
196
  ],
197
  "bos_token": null,
198
  "clean_up_tokenization_spaces": false,
199
- "eos_token": "<|endoftext|>",
200
  "errors": "replace",
201
  "extra_special_tokens": {},
202
  "model_max_length": 131072,
 
196
  ],
197
  "bos_token": null,
198
  "clean_up_tokenization_spaces": false,
199
+ "eos_token": "<|im_end|>",
200
  "errors": "replace",
201
  "extra_special_tokens": {},
202
  "model_max_length": 131072,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0597b01b59f78b533a8c8785d7f32d7d65f4e1fcbc6c853dd5201ba8ca8e0392
3
- size 7288
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c65576908344f84010c4d94640c5332fe4a0f21bfdd4c02b3a54743d30df66b
3
+ size 12152