RyanYr commited on
Commit
751a686
·
verified ·
1 Parent(s): dd8ac24

Save model at global step 5

Browse files
config.yaml ADDED
@@ -0,0 +1,224 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ data:
2
+ tokenizer: null
3
+ train_files: dapo_ds_train_sample.parquet
4
+ val_files: matheval.parquet
5
+ prompt_key: prompt
6
+ reward_fn_key: data_source
7
+ max_prompt_length: 768
8
+ max_response_length: 13312
9
+ train_batch_size: 1024
10
+ val_batch_size: null
11
+ return_raw_input_ids: false
12
+ return_raw_chat: false
13
+ shuffle: true
14
+ filter_overlong_prompts: true
15
+ filter_overlong_prompts_workers: 1
16
+ truncation: error
17
+ image_key: images
18
+ custom_cls:
19
+ path: null
20
+ name: null
21
+ actor_rollout_ref:
22
+ hybrid_engine: true
23
+ model:
24
+ path: deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
25
+ external_lib: null
26
+ override_config: {}
27
+ enable_gradient_checkpointing: true
28
+ use_remove_padding: true
29
+ use_liger: false
30
+ save_hf_repo_id: RyanYr/ppo-dapo-r1qwen1.5B-base-lr-mbs256-gpu8_actor
31
+ tokenizer_chat_template: null
32
+ actor:
33
+ strategy: fsdp
34
+ ppo_mini_batch_size: 256
35
+ ppo_micro_batch_size: null
36
+ ppo_micro_batch_size_per_gpu: 1
37
+ use_dynamic_bsz: false
38
+ ppo_max_token_len_per_gpu: 16384
39
+ grad_clip: 1.0
40
+ clip_ratio: 0.2
41
+ clip_ratio_low: 0.2
42
+ clip_ratio_high: 0.2
43
+ clip_ratio_c: 3.0
44
+ loss_agg_mode: token-mean
45
+ entropy_coeff: 0.001
46
+ use_kl_loss: false
47
+ use_torch_compile: true
48
+ kl_loss_coef: 0.001
49
+ kl_loss_type: low_var_kl
50
+ ppo_epochs: 1
51
+ shuffle: false
52
+ ulysses_sequence_parallel_size: 1
53
+ checkpoint:
54
+ contents:
55
+ - model
56
+ - optimizer
57
+ - extra
58
+ optim:
59
+ lr: 1.0e-06
60
+ lr_warmup_steps: -1
61
+ lr_warmup_steps_ratio: 0.0
62
+ min_lr_ratio: null
63
+ warmup_style: constant
64
+ total_training_steps: 100
65
+ weight_decay: 0.01
66
+ fsdp_config:
67
+ wrap_policy:
68
+ min_num_params: 0
69
+ param_offload: false
70
+ optimizer_offload: false
71
+ fsdp_size: -1
72
+ ref:
73
+ strategy: fsdp
74
+ fsdp_config:
75
+ param_offload: false
76
+ wrap_policy:
77
+ min_num_params: 0
78
+ log_prob_micro_batch_size: null
79
+ log_prob_micro_batch_size_per_gpu: 1
80
+ log_prob_use_dynamic_bsz: false
81
+ log_prob_max_token_len_per_gpu: 16384
82
+ ulysses_sequence_parallel_size: 1
83
+ rollout:
84
+ name: vllm
85
+ temperature: 1.0
86
+ top_k: -1
87
+ top_p: 1
88
+ use_fire_sampling: false
89
+ prompt_length: 768
90
+ response_length: 13312
91
+ dtype: bfloat16
92
+ gpu_memory_utilization: 0.7
93
+ ignore_eos: false
94
+ enforce_eager: false
95
+ free_cache_engine: false
96
+ load_format: dummy_dtensor
97
+ tensor_model_parallel_size: 4
98
+ max_num_batched_tokens: 14080
99
+ max_model_len: null
100
+ max_num_seqs: 1024
101
+ log_prob_micro_batch_size: null
102
+ log_prob_micro_batch_size_per_gpu: 1
103
+ log_prob_use_dynamic_bsz: false
104
+ log_prob_max_token_len_per_gpu: 16384
105
+ disable_log_stats: true
106
+ enable_chunked_prefill: true
107
+ do_sample: true
108
+ 'n': 1
109
+ engine_kwargs:
110
+ swap_space: null
111
+ val_kwargs:
112
+ top_k: -1
113
+ top_p: 1.0
114
+ temperature: 0
115
+ 'n': 1
116
+ do_sample: false
117
+ critic:
118
+ rollout_n: 1
119
+ strategy: fsdp
120
+ optim:
121
+ lr: 1.0e-05
122
+ lr_warmup_steps_ratio: 0.0
123
+ min_lr_ratio: null
124
+ warmup_style: constant
125
+ total_training_steps: 100
126
+ weight_decay: 0.01
127
+ model:
128
+ path: deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
129
+ tokenizer_path: deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
130
+ override_config: {}
131
+ external_lib: null
132
+ enable_gradient_checkpointing: false
133
+ use_remove_padding: true
134
+ fsdp_config:
135
+ param_offload: false
136
+ optimizer_offload: false
137
+ wrap_policy:
138
+ min_num_params: 0
139
+ fsdp_size: -1
140
+ save_hf_repo_id: RyanYr/ppo-dapo-r1qwen1.5B-base-lr-mbs256-gpu8_critic
141
+ ppo_mini_batch_size: 256
142
+ ppo_micro_batch_size: null
143
+ ppo_micro_batch_size_per_gpu: 1
144
+ forward_micro_batch_size: null
145
+ forward_micro_batch_size_per_gpu: 1
146
+ use_dynamic_bsz: false
147
+ ppo_max_token_len_per_gpu: 32768
148
+ forward_max_token_len_per_gpu: 32768
149
+ ulysses_sequence_parallel_size: 1
150
+ ppo_epochs: 1
151
+ shuffle: false
152
+ grad_clip: 1.0
153
+ cliprange_value: 0.5
154
+ checkpoint:
155
+ contents:
156
+ - model
157
+ - optimizer
158
+ - extra
159
+ reward_model:
160
+ enable: false
161
+ strategy: fsdp
162
+ model:
163
+ input_tokenizer: deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
164
+ path: ~/models/FsfairX-LLaMA3-RM-v0.1
165
+ external_lib: null
166
+ use_remove_padding: false
167
+ fsdp_config:
168
+ wrap_policy:
169
+ min_num_params: 0
170
+ param_offload: false
171
+ fsdp_size: -1
172
+ micro_batch_size: null
173
+ micro_batch_size_per_gpu: null
174
+ max_length: null
175
+ ulysses_sequence_parallel_size: 1
176
+ use_dynamic_bsz: false
177
+ forward_max_token_len_per_gpu: 32768
178
+ reward_manager: prime
179
+ custom_reward_function:
180
+ path: null
181
+ name: compute_score
182
+ algorithm:
183
+ gamma: 1.0
184
+ lam: 1.0
185
+ adv_estimator: gae
186
+ use_kl_in_reward: true
187
+ kl_penalty: kl
188
+ kl_ctrl:
189
+ type: fixed
190
+ kl_coef: 0.001
191
+ horizon: 10000
192
+ target_kl: 0.1
193
+ trainer:
194
+ balance_batch: true
195
+ total_epochs: 100
196
+ total_training_steps: 100
197
+ project_name: value-LLM
198
+ experiment_name: ppo-dapo-r1qwen1.5B-base_lr-mbs256-gpu8
199
+ logger:
200
+ - console
201
+ - wandb
202
+ log_val_generations: 0
203
+ nnodes: 1
204
+ n_gpus_per_node: 8
205
+ save_freq: 5
206
+ resume_mode: auto
207
+ resume_from_path: null
208
+ val_before_train: false
209
+ test_freq: -1
210
+ critic_warmup: 0
211
+ default_hdfs_dir: null
212
+ del_local_ckpt_after_load: false
213
+ default_local_dir: checkpoints/value-LLM/ppo-dapo-r1qwen1.5B-base_lr-mbs256-gpu8
214
+ max_actor_ckpt_to_keep: 1
215
+ max_critic_ckpt_to_keep: 1
216
+ ray_wait_register_center_timeout: 300
217
+ hf_token: null
218
+ resume_from_hf:
219
+ enable: false
220
+ actor_hf_repo_id: null
221
+ actor_revision: main
222
+ critic_hf_repo_id: null
223
+ critic_revision: main
224
+ hf_token: null
data.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d652b8fb542c50c9e3da045c04599fc560c10edb7716089c3b430f1f3362e34b
3
+ size 1492
extra_state_world_size_8_rank_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d26a59f27bd732e4c612d8d181bdaf8e2602741fd9168b63b083900414555da6
3
+ size 14632
extra_state_world_size_8_rank_1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:79ecdb326599a4e54468a177285efcb643176d661ce2ac2e596ee040050c9658
3
+ size 14632
extra_state_world_size_8_rank_2.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d857c457b5764c683c29ceb5624a113f371f9fc5349661055d4c9e5c13e01dd3
3
+ size 14632
extra_state_world_size_8_rank_3.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6d9eaf0ea4780cb43049910655bd7ed15f8b5eeaa276bcc5f338206b31929425
3
+ size 14632
extra_state_world_size_8_rank_4.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02e33b60788c672801ab2d3a3532acca8eea49b3d85e814de689602b0156bfa4
3
+ size 14632
extra_state_world_size_8_rank_5.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2877d1139c733db5a39689cf05cda32bd5ab88e45fb65d8adcb19242c2807926
3
+ size 14632
extra_state_world_size_8_rank_6.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:75048502776790ad78c23cb22300d7c6ee395ccadc2d9d7ecc40c26972e8ffe6
3
+ size 14632
extra_state_world_size_8_rank_7.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44cb2df6c1bef3e6f59919fed64aa5fb97cad895d2aaa0a5e8136e04bc9e4692
3
+ size 14632
model_world_size_8_rank_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:23c6641e2e3597b8c44a87570271db2cf2992706461657ae1cded60cc93ac647
3
+ size 888732218
model_world_size_8_rank_1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b58da916a79788367114c5e73d4526ec953f64a8cf94d7044b2e2e3c861f1933
3
+ size 888732218
model_world_size_8_rank_2.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cbe831687b13b9b33ba2b7b2cf10ea76ddcd06c849ca4ae6854f311852e31293
3
+ size 888732218
model_world_size_8_rank_3.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b2c2b37efe3989f2d00a052b97ee6c3e2ce3354ff9583e7d4164093f7f1f3724
3
+ size 888732218
model_world_size_8_rank_4.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f5ae5a32bb7a0c03215bf9e617c1984a4c5aeb3e22537158890212a745515df0
3
+ size 888732218
model_world_size_8_rank_5.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5fe66fbab0c90cffd625840082537b5512d67a9294c783716f40913b8509e724
3
+ size 888732218
model_world_size_8_rank_6.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b594a7dc92622e3607e82aa2b1d9dfeeb5c56bf48143b37d40200a7d6147568
3
+ size 888732218
model_world_size_8_rank_7.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:469709046db106cc227115d79bb9f396e1c5d0b45c94a2058508e03db44ce155
3
+ size 888732218
optim_world_size_8_rank_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce5212d4ab187c6f84c68dbdf93638e25f36dcf2cc74648279f8d467f8a433b4
3
+ size 1777114167
optim_world_size_8_rank_1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6702ad776eb9baa73e6ebb6e3cb11fdbc5ccec2203cca48e4069d88e2caabc53
3
+ size 1777114167
optim_world_size_8_rank_2.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3527368d8b5356f802f65668638e54cbf7b14141456e03a3665f7862bd7c18b9
3
+ size 1777114167
optim_world_size_8_rank_3.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:00e8806c2fe37aea2df5db541edda391410fe765f1153a01705b97bb9a50e6c8
3
+ size 1777114167
optim_world_size_8_rank_4.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:94b56b3a5e919c865020bffbe339fca0d077801b16bf326bb3672ec6d44341f4
3
+ size 1777114167
optim_world_size_8_rank_5.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:143df3afc3ed5375791695dfd88a555221cea4dcc27039452b10405b1f9fa757
3
+ size 1777114167
optim_world_size_8_rank_6.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:80f1e05927b6a2e2f75b76a8836819fa773b643b2c0972f3017d730a20a1beaf
3
+ size 1777114167
optim_world_size_8_rank_7.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:524a66d4c035541d58d994f10b12874cfbf0f397e4c6fc0d6c4072db72fb6ba1
3
+ size 1777114167