CharlieJi commited on
Commit
bde9343
·
verified ·
1 Parent(s): def6d93

End of training

Browse files
Files changed (5) hide show
  1. README.md +2 -1
  2. all_results.json +8 -0
  3. train_results.json +8 -0
  4. trainer_state.json +288 -0
  5. training_loss.png +0 -0
README.md CHANGED
@@ -4,6 +4,7 @@ license: apache-2.0
4
  base_model: Qwen/Qwen3-1.7B
5
  tags:
6
  - llama-factory
 
7
  - generated_from_trainer
8
  model-index:
9
  - name: qwen3-1.7b-dabstep-reasoning-108-fixed-reasoning-sharegpt-sft
@@ -15,7 +16,7 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  # qwen3-1.7b-dabstep-reasoning-108-fixed-reasoning-sharegpt-sft
17
 
18
- This model is a fine-tuned version of [Qwen/Qwen3-1.7B](https://huggingface.co/Qwen/Qwen3-1.7B) on an unknown dataset.
19
 
20
  ## Model description
21
 
 
4
  base_model: Qwen/Qwen3-1.7B
5
  tags:
6
  - llama-factory
7
+ - full
8
  - generated_from_trainer
9
  model-index:
10
  - name: qwen3-1.7b-dabstep-reasoning-108-fixed-reasoning-sharegpt-sft
 
16
 
17
  # qwen3-1.7b-dabstep-reasoning-108-fixed-reasoning-sharegpt-sft
18
 
19
+ This model is a fine-tuned version of [Qwen/Qwen3-1.7B](https://huggingface.co/Qwen/Qwen3-1.7B) on the eval-ds-dabstep-reasoning-108-fixed-reasoning-sharegpt dataset.
20
 
21
  ## Model description
22
 
all_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 5.0,
3
+ "total_flos": 852618117120.0,
4
+ "train_loss": 0.9887651102883476,
5
+ "train_runtime": 133.1517,
6
+ "train_samples_per_second": 3.793,
7
+ "train_steps_per_second": 0.263
8
+ }
train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 5.0,
3
+ "total_flos": 852618117120.0,
4
+ "train_loss": 0.9887651102883476,
5
+ "train_runtime": 133.1517,
6
+ "train_samples_per_second": 3.793,
7
+ "train_steps_per_second": 0.263
8
+ }
trainer_state.json ADDED
@@ -0,0 +1,288 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 5.0,
6
+ "eval_steps": 500,
7
+ "global_step": 35,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.14285714285714285,
14
+ "grad_norm": 20.565131124519954,
15
+ "learning_rate": 0.0,
16
+ "loss": 1.6904,
17
+ "step": 1
18
+ },
19
+ {
20
+ "epoch": 0.2857142857142857,
21
+ "grad_norm": 20.126905785368916,
22
+ "learning_rate": 1e-05,
23
+ "loss": 1.6039,
24
+ "step": 2
25
+ },
26
+ {
27
+ "epoch": 0.42857142857142855,
28
+ "grad_norm": 11.090236436649034,
29
+ "learning_rate": 9.978670881475173e-06,
30
+ "loss": 1.5826,
31
+ "step": 3
32
+ },
33
+ {
34
+ "epoch": 0.5714285714285714,
35
+ "grad_norm": 5.614097530062799,
36
+ "learning_rate": 9.91486549841951e-06,
37
+ "loss": 1.4068,
38
+ "step": 4
39
+ },
40
+ {
41
+ "epoch": 0.7142857142857143,
42
+ "grad_norm": 5.234708126178992,
43
+ "learning_rate": 9.809128215864096e-06,
44
+ "loss": 1.4604,
45
+ "step": 5
46
+ },
47
+ {
48
+ "epoch": 0.8571428571428571,
49
+ "grad_norm": 4.306326905631283,
50
+ "learning_rate": 9.66236114702178e-06,
51
+ "loss": 1.3071,
52
+ "step": 6
53
+ },
54
+ {
55
+ "epoch": 1.0,
56
+ "grad_norm": 3.8203719300545993,
57
+ "learning_rate": 9.475816456775313e-06,
58
+ "loss": 1.2656,
59
+ "step": 7
60
+ },
61
+ {
62
+ "epoch": 1.1428571428571428,
63
+ "grad_norm": 3.336385963547391,
64
+ "learning_rate": 9.251085678648072e-06,
65
+ "loss": 1.1997,
66
+ "step": 8
67
+ },
68
+ {
69
+ "epoch": 1.2857142857142856,
70
+ "grad_norm": 3.0415666992391386,
71
+ "learning_rate": 8.990086136401199e-06,
72
+ "loss": 1.1369,
73
+ "step": 9
74
+ },
75
+ {
76
+ "epoch": 1.4285714285714286,
77
+ "grad_norm": 3.0682485455683,
78
+ "learning_rate": 8.695044586103297e-06,
79
+ "loss": 1.153,
80
+ "step": 10
81
+ },
82
+ {
83
+ "epoch": 1.5714285714285714,
84
+ "grad_norm": 2.660915607465715,
85
+ "learning_rate": 8.368478218232787e-06,
86
+ "loss": 1.0615,
87
+ "step": 11
88
+ },
89
+ {
90
+ "epoch": 1.7142857142857144,
91
+ "grad_norm": 2.3104573446692664,
92
+ "learning_rate": 8.013173181896283e-06,
93
+ "loss": 1.015,
94
+ "step": 12
95
+ },
96
+ {
97
+ "epoch": 1.8571428571428572,
98
+ "grad_norm": 2.1383910273570086,
99
+ "learning_rate": 7.63216081438678e-06,
100
+ "loss": 1.06,
101
+ "step": 13
102
+ },
103
+ {
104
+ "epoch": 2.0,
105
+ "grad_norm": 1.9283231920822732,
106
+ "learning_rate": 7.2286917788826926e-06,
107
+ "loss": 0.9481,
108
+ "step": 14
109
+ },
110
+ {
111
+ "epoch": 2.142857142857143,
112
+ "grad_norm": 1.8886170821861419,
113
+ "learning_rate": 6.806208330935766e-06,
114
+ "loss": 0.9303,
115
+ "step": 15
116
+ },
117
+ {
118
+ "epoch": 2.2857142857142856,
119
+ "grad_norm": 1.8907890024162215,
120
+ "learning_rate": 6.368314950360416e-06,
121
+ "loss": 0.9709,
122
+ "step": 16
123
+ },
124
+ {
125
+ "epoch": 2.4285714285714284,
126
+ "grad_norm": 1.9093053411125347,
127
+ "learning_rate": 5.918747589082853e-06,
128
+ "loss": 0.9213,
129
+ "step": 17
130
+ },
131
+ {
132
+ "epoch": 2.571428571428571,
133
+ "grad_norm": 1.866649612150217,
134
+ "learning_rate": 5.46134179731651e-06,
135
+ "loss": 0.8741,
136
+ "step": 18
137
+ },
138
+ {
139
+ "epoch": 2.7142857142857144,
140
+ "grad_norm": 1.7099290243677971,
141
+ "learning_rate": 5e-06,
142
+ "loss": 0.8563,
143
+ "step": 19
144
+ },
145
+ {
146
+ "epoch": 2.857142857142857,
147
+ "grad_norm": 1.7446978143600087,
148
+ "learning_rate": 4.53865820268349e-06,
149
+ "loss": 0.8587,
150
+ "step": 20
151
+ },
152
+ {
153
+ "epoch": 3.0,
154
+ "grad_norm": 1.6658354125290393,
155
+ "learning_rate": 4.081252410917148e-06,
156
+ "loss": 0.7558,
157
+ "step": 21
158
+ },
159
+ {
160
+ "epoch": 3.142857142857143,
161
+ "grad_norm": 1.6588731170974642,
162
+ "learning_rate": 3.6316850496395863e-06,
163
+ "loss": 0.7808,
164
+ "step": 22
165
+ },
166
+ {
167
+ "epoch": 3.2857142857142856,
168
+ "grad_norm": 1.7054713909510084,
169
+ "learning_rate": 3.1937916690642356e-06,
170
+ "loss": 0.8018,
171
+ "step": 23
172
+ },
173
+ {
174
+ "epoch": 3.4285714285714284,
175
+ "grad_norm": 1.5044309582959277,
176
+ "learning_rate": 2.771308221117309e-06,
177
+ "loss": 0.785,
178
+ "step": 24
179
+ },
180
+ {
181
+ "epoch": 3.571428571428571,
182
+ "grad_norm": 1.6109753709228194,
183
+ "learning_rate": 2.3678391856132203e-06,
184
+ "loss": 0.7427,
185
+ "step": 25
186
+ },
187
+ {
188
+ "epoch": 3.7142857142857144,
189
+ "grad_norm": 1.6141990688430101,
190
+ "learning_rate": 1.9868268181037186e-06,
191
+ "loss": 0.7829,
192
+ "step": 26
193
+ },
194
+ {
195
+ "epoch": 3.857142857142857,
196
+ "grad_norm": 1.6811314196010585,
197
+ "learning_rate": 1.6315217817672142e-06,
198
+ "loss": 0.785,
199
+ "step": 27
200
+ },
201
+ {
202
+ "epoch": 4.0,
203
+ "grad_norm": 1.6017235236105927,
204
+ "learning_rate": 1.3049554138967052e-06,
205
+ "loss": 0.7273,
206
+ "step": 28
207
+ },
208
+ {
209
+ "epoch": 4.142857142857143,
210
+ "grad_norm": 1.5419350535809524,
211
+ "learning_rate": 1.0099138635988026e-06,
212
+ "loss": 0.7737,
213
+ "step": 29
214
+ },
215
+ {
216
+ "epoch": 4.285714285714286,
217
+ "grad_norm": 1.5992532932151025,
218
+ "learning_rate": 7.489143213519301e-07,
219
+ "loss": 0.6784,
220
+ "step": 30
221
+ },
222
+ {
223
+ "epoch": 4.428571428571429,
224
+ "grad_norm": 1.6017284844116721,
225
+ "learning_rate": 5.241835432246888e-07,
226
+ "loss": 0.7267,
227
+ "step": 31
228
+ },
229
+ {
230
+ "epoch": 4.571428571428571,
231
+ "grad_norm": 1.5403007797347406,
232
+ "learning_rate": 3.3763885297822153e-07,
233
+ "loss": 0.767,
234
+ "step": 32
235
+ },
236
+ {
237
+ "epoch": 4.714285714285714,
238
+ "grad_norm": 1.515361106622691,
239
+ "learning_rate": 1.908717841359048e-07,
240
+ "loss": 0.6971,
241
+ "step": 33
242
+ },
243
+ {
244
+ "epoch": 4.857142857142857,
245
+ "grad_norm": 1.5420726170851473,
246
+ "learning_rate": 8.513450158049109e-08,
247
+ "loss": 0.7647,
248
+ "step": 34
249
+ },
250
+ {
251
+ "epoch": 5.0,
252
+ "grad_norm": 1.5561245777833046,
253
+ "learning_rate": 2.1329118524827662e-08,
254
+ "loss": 0.7352,
255
+ "step": 35
256
+ },
257
+ {
258
+ "epoch": 5.0,
259
+ "step": 35,
260
+ "total_flos": 852618117120.0,
261
+ "train_loss": 0.9887651102883476,
262
+ "train_runtime": 133.1517,
263
+ "train_samples_per_second": 3.793,
264
+ "train_steps_per_second": 0.263
265
+ }
266
+ ],
267
+ "logging_steps": 1,
268
+ "max_steps": 35,
269
+ "num_input_tokens_seen": 0,
270
+ "num_train_epochs": 5,
271
+ "save_steps": 42,
272
+ "stateful_callbacks": {
273
+ "TrainerControl": {
274
+ "args": {
275
+ "should_epoch_stop": false,
276
+ "should_evaluate": false,
277
+ "should_log": false,
278
+ "should_save": true,
279
+ "should_training_stop": true
280
+ },
281
+ "attributes": {}
282
+ }
283
+ },
284
+ "total_flos": 852618117120.0,
285
+ "train_batch_size": 2,
286
+ "trial_name": null,
287
+ "trial_params": null
288
+ }
training_loss.png ADDED