RyanYr commited on
Commit
ce28ebc
·
verified ·
1 Parent(s): 0f22b0e

Save model at global step 40

Browse files
config.yaml ADDED
@@ -0,0 +1,176 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ data:
2
+ tokenizer: null
3
+ train_files: dapo_ds_train_sample.parquet
4
+ val_files: matheval.parquet
5
+ prompt_key: prompt
6
+ reward_fn_key: data_source
7
+ max_prompt_length: 1024
8
+ max_response_length: 2048
9
+ train_batch_size: 128
10
+ val_batch_size: 640
11
+ return_raw_input_ids: false
12
+ return_raw_chat: false
13
+ shuffle: true
14
+ filter_overlong_prompts: false
15
+ filter_overlong_prompts_workers: 1
16
+ truncation: left
17
+ image_key: images
18
+ custom_cls:
19
+ path: null
20
+ name: null
21
+ actor_rollout_ref:
22
+ hybrid_engine: true
23
+ model:
24
+ path: Qwen/Qwen2.5-Math-1.5B
25
+ external_lib: null
26
+ override_config: {}
27
+ enable_gradient_checkpointing: true
28
+ use_remove_padding: false
29
+ use_liger: false
30
+ save_hf_repo_id: RyanYr/brm-dapo-qwen2.5math-1.5B-base-lr2.5e-6-beta0.002
31
+ tokenizer_chat_template: null
32
+ actor:
33
+ brm:
34
+ norm_factor: value
35
+ value_constant: null
36
+ buffer:
37
+ buffer_type: null
38
+ offline_dataset_buffer:
39
+ train_files: null
40
+ response_key: response
41
+ response_truncation: right
42
+ shuffle: true
43
+ update_size: 128
44
+ strategy: fsdp
45
+ ppo_mini_batch_size: 128
46
+ ppo_micro_batch_size: null
47
+ ppo_micro_batch_size_per_gpu: 4
48
+ use_dynamic_bsz: false
49
+ ppo_max_token_len_per_gpu: 16384
50
+ grad_clip: 1.0
51
+ use_torch_compile: true
52
+ ppo_epochs: 1
53
+ shuffle: false
54
+ ulysses_sequence_parallel_size: 1
55
+ checkpoint:
56
+ contents:
57
+ - model
58
+ - optimizer
59
+ - extra
60
+ optim:
61
+ lr: 2.5e-06
62
+ lr_warmup_steps: -1
63
+ lr_warmup_steps_ratio: 0
64
+ min_lr_ratio: null
65
+ warmup_style: constant
66
+ total_training_steps: 810
67
+ weight_decay: 0.01
68
+ fsdp_config:
69
+ wrap_policy:
70
+ min_num_params: 0
71
+ param_offload: false
72
+ optimizer_offload: false
73
+ fsdp_size: -1
74
+ ref:
75
+ ref_model_path: Qwen/Qwen2.5-Math-1.5B
76
+ strategy: fsdp
77
+ fsdp_config:
78
+ param_offload: false
79
+ wrap_policy:
80
+ min_num_params: 0
81
+ log_prob_micro_batch_size: null
82
+ log_prob_micro_batch_size_per_gpu: 16
83
+ log_prob_use_dynamic_bsz: false
84
+ log_prob_max_token_len_per_gpu: 16384
85
+ ulysses_sequence_parallel_size: 1
86
+ rollout:
87
+ name: vllm
88
+ temperature: 1.0
89
+ top_k: -1
90
+ top_p: 1
91
+ use_fire_sampling: false
92
+ prompt_length: 1024
93
+ response_length: 2048
94
+ dtype: bfloat16
95
+ gpu_memory_utilization: 0.7
96
+ ignore_eos: false
97
+ enforce_eager: false
98
+ free_cache_engine: false
99
+ load_format: dummy_dtensor
100
+ tensor_model_parallel_size: 4
101
+ max_num_batched_tokens: 4072
102
+ max_model_len: null
103
+ max_num_seqs: 1024
104
+ log_prob_micro_batch_size: null
105
+ log_prob_micro_batch_size_per_gpu: 16
106
+ log_prob_use_dynamic_bsz: false
107
+ log_prob_max_token_len_per_gpu: 16384
108
+ disable_log_stats: true
109
+ enable_chunked_prefill: true
110
+ do_sample: true
111
+ 'n': 1
112
+ engine_kwargs:
113
+ swap_space: null
114
+ val_kwargs:
115
+ top_k: -1
116
+ top_p: 1.0
117
+ temperature: 0.6
118
+ 'n': 32
119
+ do_sample: true
120
+ reward_model:
121
+ enable: false
122
+ strategy: fsdp
123
+ model:
124
+ input_tokenizer: Qwen/Qwen2.5-Math-1.5B
125
+ path: ~/models/FsfairX-LLaMA3-RM-v0.1
126
+ external_lib: null
127
+ use_remove_padding: false
128
+ fsdp_config:
129
+ wrap_policy:
130
+ min_num_params: 0
131
+ param_offload: false
132
+ fsdp_size: -1
133
+ micro_batch_size: null
134
+ micro_batch_size_per_gpu: null
135
+ max_length: null
136
+ ulysses_sequence_parallel_size: 1
137
+ use_dynamic_bsz: false
138
+ forward_max_token_len_per_gpu: 16384
139
+ reward_manager: prime
140
+ reward_kwargs:
141
+ format_reward: 0.0
142
+ format_type: null
143
+ custom_reward_function:
144
+ path: null
145
+ name: compute_score
146
+ trainer:
147
+ balance_batch: true
148
+ total_epochs: 6
149
+ total_training_steps: null
150
+ project_name: value-LLM
151
+ experiment_name: brm-dapo-qwen2.5math-1.5B-base_lr2.5e-6-beta0.002
152
+ logger:
153
+ - console
154
+ - wandb
155
+ log_val_generations: 0
156
+ nnodes: 1
157
+ n_gpus_per_node: 4
158
+ save_freq: 40
159
+ resume_mode: auto
160
+ resume_from_path: null
161
+ val_before_train: false
162
+ test_freq: -1
163
+ default_hdfs_dir: null
164
+ del_local_ckpt_after_load: false
165
+ default_local_dir: ./BRM
166
+ max_actor_ckpt_to_keep: 1
167
+ ray_wait_register_center_timeout: 300
168
+ hf_token: null
169
+ resume_from_hf:
170
+ enable: false
171
+ hf_repo_id: null
172
+ hf_token: null
173
+ revision: main
174
+ algorithm:
175
+ kl_ctrl:
176
+ kl_coef: 0.002
data.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:972a5213f98845a35d97265f0f211d784de8b77d1d8f876c32d83a5f01de0bda
3
+ size 1492
extra_state_world_size_4_rank_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bac478195e64c5adc1589f984ee9504b31ee50d7781d99cdf22ecfe513641e38
3
+ size 14632
extra_state_world_size_4_rank_1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5bd9eb602b937ae3abfdc710b6d8ee83926d7c65b5de1c6c28b9ae4f7e9b142f
3
+ size 14632
extra_state_world_size_4_rank_2.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:80a362d156678550afad0428cd91309dea2ab23369a46041c12e2ed80ad62ff7
3
+ size 14632
extra_state_world_size_4_rank_3.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a53baac35ae90ef7df9bc4c96819f371f332549cb2b330d17683f70d526a426
3
+ size 14632
model_world_size_4_rank_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86251f89de498ee633c346e25b3f2960784397b63c13f10781dcb1d81a26b2d5
3
+ size 1777276538
model_world_size_4_rank_1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e604abd3287779308deabdba0b53571da05d37f7a755885a95f35f6864fd2ab
3
+ size 1777276538
model_world_size_4_rank_2.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:08940ed45aa2c3fe769b6a93237a2dfd7d8d150842c4487fa3d3045d6646a3d6
3
+ size 1777276538
model_world_size_4_rank_3.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f0ea4d52c6d12df821bb1426e78a8398a8b375f038a5a21f549b31c5d56d6d4
3
+ size 1777276538
optim_world_size_4_rank_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d3d908e0d8ce6376de568c16933f0fb14a12b723d7836198c4daad383b0b247
3
+ size 3087454775
optim_world_size_4_rank_1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1dc2ee8e96825a91b3aa645f9063421767c3bdfc466ac4e3cd1fbab8d83a7436
3
+ size 3087454775
optim_world_size_4_rank_2.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:273b5aebe47cf5198ba6b1cab4e3d5e274aa2497edf13a0c5d02904198b4659e
3
+ size 3087454775
optim_world_size_4_rank_3.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13add9ec1501920902f6a93cdcc2516715946a2721d0234e1765c973c88013b0
3
+ size 3087454775