RyanYr commited on
Commit
790c3d2
·
verified ·
1 Parent(s): af3651e

Save model at global step 40

Browse files
config.yaml ADDED
@@ -0,0 +1,177 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ data:
2
+ tokenizer: null
3
+ train_files: dapo_ds_train_sample.parquet
4
+ val_files: matheval.parquet
5
+ prompt_key: prompt
6
+ reward_fn_key: data_source
7
+ max_prompt_length: 768
8
+ max_response_length: 4096
9
+ train_batch_size: 128
10
+ val_batch_size: 640
11
+ return_raw_input_ids: false
12
+ return_raw_chat: false
13
+ shuffle: true
14
+ filter_overlong_prompts: false
15
+ filter_overlong_prompts_workers: 1
16
+ truncation: left
17
+ image_key: images
18
+ custom_cls:
19
+ path: null
20
+ name: null
21
+ actor_rollout_ref:
22
+ hybrid_engine: true
23
+ model:
24
+ path: meta-llama/Llama-3.2-3B-Instruct
25
+ external_lib: null
26
+ override_config: {}
27
+ enable_gradient_checkpointing: true
28
+ use_remove_padding: false
29
+ use_liger: false
30
+ save_hf_repo_id: RyanYr/brm-dapo-llama3.2-3B-it_lr1.2e-6_rspL4096-beta0.002
31
+ tokenizer_chat_template: null
32
+ actor:
33
+ brm:
34
+ norm_factor: value
35
+ value_constant: null
36
+ buffer:
37
+ buffer_type: null
38
+ offline_dataset_buffer:
39
+ train_files: null
40
+ response_key: response
41
+ response_truncation: right
42
+ shuffle: true
43
+ update_size: 128
44
+ strategy: fsdp
45
+ ppo_mini_batch_size: 128
46
+ ppo_micro_batch_size: null
47
+ ppo_micro_batch_size_per_gpu: 1
48
+ use_dynamic_bsz: false
49
+ ppo_max_token_len_per_gpu: 16384
50
+ grad_clip: 1.0
51
+ use_torch_compile: true
52
+ ppo_epochs: 1
53
+ shuffle: false
54
+ ulysses_sequence_parallel_size: 1
55
+ checkpoint:
56
+ contents:
57
+ - model
58
+ - optimizer
59
+ - extra
60
+ optim:
61
+ lr: 1.2e-06
62
+ lr_warmup_steps: -1
63
+ lr_warmup_steps_ratio: 0
64
+ min_lr_ratio: null
65
+ warmup_style: constant
66
+ total_training_steps: 400
67
+ weight_decay: 0.01
68
+ fsdp_config:
69
+ wrap_policy:
70
+ min_num_params: 0
71
+ param_offload: false
72
+ optimizer_offload: false
73
+ fsdp_size: -1
74
+ report_entropy: true
75
+ ref:
76
+ ref_model_path: meta-llama/Llama-3.2-3B-Instruct
77
+ strategy: fsdp
78
+ fsdp_config:
79
+ param_offload: false
80
+ wrap_policy:
81
+ min_num_params: 0
82
+ log_prob_micro_batch_size: null
83
+ log_prob_micro_batch_size_per_gpu: 8
84
+ log_prob_use_dynamic_bsz: false
85
+ log_prob_max_token_len_per_gpu: 16384
86
+ ulysses_sequence_parallel_size: 1
87
+ rollout:
88
+ name: vllm
89
+ temperature: 1.0
90
+ top_k: -1
91
+ top_p: 1
92
+ use_fire_sampling: false
93
+ prompt_length: 768
94
+ response_length: 4096
95
+ dtype: bfloat16
96
+ gpu_memory_utilization: 0.6
97
+ ignore_eos: false
98
+ enforce_eager: false
99
+ free_cache_engine: false
100
+ load_format: dummy_dtensor
101
+ tensor_model_parallel_size: 4
102
+ max_num_batched_tokens: 5864
103
+ max_model_len: null
104
+ max_num_seqs: 1024
105
+ log_prob_micro_batch_size: null
106
+ log_prob_micro_batch_size_per_gpu: 8
107
+ log_prob_use_dynamic_bsz: false
108
+ log_prob_max_token_len_per_gpu: 16384
109
+ disable_log_stats: true
110
+ enable_chunked_prefill: true
111
+ do_sample: true
112
+ 'n': 1
113
+ engine_kwargs:
114
+ swap_space: null
115
+ val_kwargs:
116
+ top_k: -1
117
+ top_p: 1.0
118
+ temperature: 0.6
119
+ 'n': 32
120
+ do_sample: true
121
+ reward_model:
122
+ enable: false
123
+ strategy: fsdp
124
+ model:
125
+ input_tokenizer: meta-llama/Llama-3.2-3B-Instruct
126
+ path: ~/models/FsfairX-LLaMA3-RM-v0.1
127
+ external_lib: null
128
+ use_remove_padding: false
129
+ fsdp_config:
130
+ wrap_policy:
131
+ min_num_params: 0
132
+ param_offload: false
133
+ fsdp_size: -1
134
+ micro_batch_size: null
135
+ micro_batch_size_per_gpu: null
136
+ max_length: null
137
+ ulysses_sequence_parallel_size: 1
138
+ use_dynamic_bsz: false
139
+ forward_max_token_len_per_gpu: 16384
140
+ reward_manager: prime
141
+ reward_kwargs:
142
+ format_reward: 0.0
143
+ format_type: null
144
+ custom_reward_function:
145
+ path: null
146
+ name: compute_score
147
+ trainer:
148
+ balance_batch: true
149
+ total_epochs: 100
150
+ total_training_steps: 400
151
+ project_name: value-LLM
152
+ experiment_name: brm-dapo-llama3.2-3B-it_lr1.2e-6_rspL4096-beta0.002
153
+ logger:
154
+ - console
155
+ - wandb
156
+ log_val_generations: 0
157
+ nnodes: 1
158
+ n_gpus_per_node: 4
159
+ save_freq: 40
160
+ resume_mode: auto
161
+ resume_from_path: null
162
+ val_before_train: false
163
+ test_freq: -1
164
+ default_hdfs_dir: null
165
+ del_local_ckpt_after_load: false
166
+ default_local_dir: ./BRM
167
+ max_actor_ckpt_to_keep: 1
168
+ ray_wait_register_center_timeout: 300
169
+ hf_token: null
170
+ resume_from_hf:
171
+ enable: false
172
+ hf_repo_id: null
173
+ hf_token: null
174
+ revision: main
175
+ algorithm:
176
+ kl_ctrl:
177
+ kl_coef: 0.002
data.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:abb75c3ac8d43d5d7e8492e245f16591382108d8cd44cc4c3ceefaced2d979bc
3
+ size 1492
extra_state_world_size_4_rank_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d097a47fb7461da5d27460785c7433b2a97e1215dde8f79f9905108150c14b2
3
+ size 14632
extra_state_world_size_4_rank_1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d4781f9e12c4faa98139099a21d9101e55271dad50e9dc4a1978c324d946bd4a
3
+ size 14632
extra_state_world_size_4_rank_2.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c6fd74afcb4ec9c9da0c26634975d3ce792a3b5c6de14ff8a94b26a0b498c8ab
3
+ size 14632
extra_state_world_size_4_rank_3.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4e050958b4f2a0a92d3b95efe6615acd3d54029cc29d368dc5261ebab3027c0
3
+ size 14632
model_world_size_4_rank_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f92ead5ead02701ffe64f6a6e827d21c3c63fb644b9a4aaf9ea63ede113d32ee
3
+ size 3606904378
model_world_size_4_rank_1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:72284fd5d26ac03da4d3d5af26fa94ec439718d91fc60938692b4cd89721f0b1
3
+ size 3606904378
model_world_size_4_rank_2.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c5f06b9aa96c62657c7c2d1fe729ad5c22784468932d9bee2c5c32bd6b30b69
3
+ size 3606904378
model_world_size_4_rank_3.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b30ea09f06f23f1cb685b39ea0f94fd66f7174884b901fa60fe6869090af3f13
3
+ size 3606904378
optim_world_size_4_rank_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24fce421015ca1afbb9a28c4a26b78224cc0e08656545b1848a1409ce90548e0
3
+ size 6425526231
optim_world_size_4_rank_1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2990ad8b971e390fd46d19517b6b75bbb18a744f50f81cf5affece74211de969
3
+ size 6425526231
optim_world_size_4_rank_2.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ae74814c5aacce3c8c1b5dda75e301610e4f79907b9f2f6176f05fc3a7a6254
3
+ size 6425526231
optim_world_size_4_rank_3.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:94e744a990ef773a63e6353d2131f99e53078528a19d5a881f210f6eff0dae50
3
+ size 6425526231