yuta0x89 commited on
Commit
f48d77c
·
verified ·
1 Parent(s): f06c677

Training in progress, step 50

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "/work/bmb/hf_models/openr1/DeepSeek-R1-Distill-Qwen-1.5B-GRPO-beta-0.04",
3
  "architectures": [
4
  "Qwen2ForCausalLM"
5
  ],
 
1
  {
2
+ "_name_or_path": "/work/bmb/hf_models/Qwen/Qwen2.5-Math-1.5B-Instruct",
3
  "architectures": [
4
  "Qwen2ForCausalLM"
5
  ],
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5e7475ef2edb68e70e8d698526ec225e4c63e6e5311be8a7aac2806061c8983f
3
  size 3554214752
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:79ed37c0e75a8f083d29ba3cf67cdca706b821bffb075d26c132e1f75f3a7207
3
  size 3554214752
tokenizer_config.json CHANGED
@@ -200,12 +200,8 @@
200
  "eos_token": "<|im_end|>",
201
  "errors": "replace",
202
  "extra_special_tokens": {},
203
- "max_length": null,
204
  "model_max_length": 131072,
205
- "pad_to_multiple_of": null,
206
  "pad_token": "<|endoftext|>",
207
- "pad_token_type_id": 0,
208
- "padding_side": "left",
209
  "split_special_tokens": false,
210
  "tokenizer_class": "Qwen2Tokenizer",
211
  "unk_token": null
 
200
  "eos_token": "<|im_end|>",
201
  "errors": "replace",
202
  "extra_special_tokens": {},
 
203
  "model_max_length": 131072,
 
204
  "pad_token": "<|endoftext|>",
 
 
205
  "split_special_tokens": false,
206
  "tokenizer_class": "Qwen2Tokenizer",
207
  "unk_token": null
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cfde4e89983697b6286a46e88ca7b2e4c2863ec5823bb968abb668e1b166b2ef
3
  size 8184
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f642d0d845f7ab297fb4cfb3b81476d48a3417bc468255b997008ec5958d9f3
3
  size 8184