li-muyang commited on
Commit
188f635
·
verified ·
1 Parent(s): a86b718

Model save

Browse files
Files changed (5) hide show
  1. README.md +15 -15
  2. all_results.json +4 -4
  3. generation_config.json +3 -2
  4. train_results.json +4 -4
  5. trainer_state.json +134 -134
README.md CHANGED
@@ -16,15 +16,15 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  This model was trained from scratch on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
- - Loss: 0.4802
20
- - Rewards/chosen: 0.1868
21
- - Rewards/rejected: -0.8036
22
- - Rewards/accuracies: 0.75
23
- - Rewards/margins: 0.9904
24
- - Logps/rejected: -347.3108
25
- - Logps/chosen: -331.7999
26
- - Logits/rejected: -3.0382
27
- - Logits/chosen: -3.0187
28
 
29
  ## Model description
30
 
@@ -44,14 +44,14 @@ More information needed
44
 
45
  The following hyperparameters were used during training:
46
  - learning_rate: 5e-07
47
- - train_batch_size: 4
48
- - eval_batch_size: 4
49
  - seed: 42
50
  - distributed_type: multi-GPU
51
  - num_devices: 8
52
- - gradient_accumulation_steps: 4
53
  - total_train_batch_size: 128
54
- - total_eval_batch_size: 32
55
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
56
  - lr_scheduler_type: cosine
57
  - lr_scheduler_warmup_ratio: 0.1
@@ -61,8 +61,8 @@ The following hyperparameters were used during training:
61
 
62
  | Training Loss | Epoch | Step | Validation Loss | Rewards/chosen | Rewards/rejected | Rewards/accuracies | Rewards/margins | Logps/rejected | Logps/chosen | Logits/rejected | Logits/chosen |
63
  |:-------------:|:------:|:----:|:---------------:|:--------------:|:----------------:|:------------------:|:---------------:|:--------------:|:------------:|:---------------:|:-------------:|
64
- | 0.493 | 0.9858 | 52 | 0.5064 | 0.4437 | -0.3824 | 0.7812 | 0.8261 | -338.8869 | -326.6608 | -3.1043 | -3.0877 |
65
- | 0.1356 | 1.9716 | 104 | 0.4802 | 0.1868 | -0.8036 | 0.75 | 0.9904 | -347.3108 | -331.7999 | -3.0382 | -3.0187 |
66
 
67
 
68
  ### Framework versions
 
16
 
17
  This model was trained from scratch on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
+ - Loss: 0.5490
20
+ - Rewards/chosen: -0.3865
21
+ - Rewards/rejected: -1.0064
22
+ - Rewards/accuracies: 0.6915
23
+ - Rewards/margins: 0.6199
24
+ - Logps/rejected: -415.0506
25
+ - Logps/chosen: -355.7762
26
+ - Logits/rejected: 326.4959
27
+ - Logits/chosen: 326.0051
28
 
29
  ## Model description
30
 
 
44
 
45
  The following hyperparameters were used during training:
46
  - learning_rate: 5e-07
47
+ - train_batch_size: 1
48
+ - eval_batch_size: 2
49
  - seed: 42
50
  - distributed_type: multi-GPU
51
  - num_devices: 8
52
+ - gradient_accumulation_steps: 16
53
  - total_train_batch_size: 128
54
+ - total_eval_batch_size: 16
55
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
56
  - lr_scheduler_type: cosine
57
  - lr_scheduler_warmup_ratio: 0.1
 
61
 
62
  | Training Loss | Epoch | Step | Validation Loss | Rewards/chosen | Rewards/rejected | Rewards/accuracies | Rewards/margins | Logps/rejected | Logps/chosen | Logits/rejected | Logits/chosen |
63
  |:-------------:|:------:|:----:|:---------------:|:--------------:|:----------------:|:------------------:|:---------------:|:--------------:|:------------:|:---------------:|:-------------:|
64
+ | 0.5334 | 0.9858 | 52 | 0.5680 | -0.2471 | -0.7713 | 0.6702 | 0.5242 | -410.3489 | -352.9877 | 326.9832 | 326.6964 |
65
+ | 0.2495 | 1.9716 | 104 | 0.5490 | -0.3865 | -1.0064 | 0.6915 | 0.6199 | -415.0506 | -355.7762 | 326.4959 | 326.0051 |
66
 
67
 
68
  ### Framework versions
all_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "epoch": 1.971563981042654,
3
  "total_flos": 0.0,
4
- "train_loss": 0.373156875371933,
5
- "train_runtime": 3298.3137,
6
  "train_samples": 6750,
7
- "train_samples_per_second": 4.093,
8
- "train_steps_per_second": 0.032
9
  }
 
1
  {
2
  "epoch": 1.971563981042654,
3
  "total_flos": 0.0,
4
+ "train_loss": 0.44703544962864655,
5
+ "train_runtime": 4801.7476,
6
  "train_samples": 6750,
7
+ "train_samples_per_second": 2.811,
8
+ "train_steps_per_second": 0.022
9
  }
generation_config.json CHANGED
@@ -1,6 +1,7 @@
1
  {
2
  "_from_model_config": true,
3
- "bos_token_id": 1,
4
- "eos_token_id": 2,
 
5
  "transformers_version": "4.45.2"
6
  }
 
1
  {
2
  "_from_model_config": true,
3
+ "bos_token_id": 2,
4
+ "eos_token_id": 1,
5
+ "pad_token_id": 0,
6
  "transformers_version": "4.45.2"
7
  }
train_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "epoch": 1.971563981042654,
3
  "total_flos": 0.0,
4
- "train_loss": 0.373156875371933,
5
- "train_runtime": 3298.3137,
6
  "train_samples": 6750,
7
- "train_samples_per_second": 4.093,
8
- "train_steps_per_second": 0.032
9
  }
 
1
  {
2
  "epoch": 1.971563981042654,
3
  "total_flos": 0.0,
4
+ "train_loss": 0.44703544962864655,
5
+ "train_runtime": 4801.7476,
6
  "train_samples": 6750,
7
+ "train_samples_per_second": 2.811,
8
+ "train_steps_per_second": 0.022
9
  }
trainer_state.json CHANGED
@@ -10,12 +10,12 @@
10
  "log_history": [
11
  {
12
  "epoch": 0.018957345971563982,
13
- "grad_norm": 83.90713111595433,
14
  "learning_rate": 4.545454545454545e-08,
15
- "logits/chosen": -2.873765468597412,
16
- "logits/rejected": -2.883460521697998,
17
- "logps/chosen": -422.4774169921875,
18
- "logps/rejected": -480.524169921875,
19
  "loss": 0.6931,
20
  "rewards/accuracies": 0.0,
21
  "rewards/chosen": 0.0,
@@ -25,194 +25,194 @@
25
  },
26
  {
27
  "epoch": 0.1895734597156398,
28
- "grad_norm": 77.5065062501301,
29
  "learning_rate": 4.545454545454545e-07,
30
- "logits/chosen": -2.859947681427002,
31
- "logits/rejected": -2.877683162689209,
32
- "logps/chosen": -346.8866882324219,
33
- "logps/rejected": -398.614013671875,
34
- "loss": 0.6895,
35
- "rewards/accuracies": 0.4930555522441864,
36
- "rewards/chosen": 0.007129705511033535,
37
- "rewards/margins": 0.008572477847337723,
38
- "rewards/rejected": -0.001442772219888866,
39
  "step": 10
40
  },
41
  {
42
  "epoch": 0.3791469194312796,
43
- "grad_norm": 68.52599234402585,
44
  "learning_rate": 4.885348141000122e-07,
45
- "logits/chosen": -2.9137072563171387,
46
- "logits/rejected": -2.92988920211792,
47
- "logps/chosen": -389.4097595214844,
48
- "logps/rejected": -417.3075256347656,
49
- "loss": 0.643,
50
- "rewards/accuracies": 0.6875,
51
- "rewards/chosen": 0.13587789237499237,
52
- "rewards/margins": 0.10481055080890656,
53
- "rewards/rejected": 0.031067365780472755,
54
  "step": 20
55
  },
56
  {
57
  "epoch": 0.5687203791469194,
58
- "grad_norm": 61.72689387884387,
59
  "learning_rate": 4.5025027361734613e-07,
60
- "logits/chosen": -2.9870388507843018,
61
- "logits/rejected": -2.9750781059265137,
62
- "logps/chosen": -334.49029541015625,
63
- "logps/rejected": -352.12322998046875,
64
- "loss": 0.5547,
65
- "rewards/accuracies": 0.75,
66
- "rewards/chosen": 0.3548569083213806,
67
- "rewards/margins": 0.3839985430240631,
68
- "rewards/rejected": -0.02914164587855339,
69
  "step": 30
70
  },
71
  {
72
  "epoch": 0.7582938388625592,
73
- "grad_norm": 64.8854081169712,
74
  "learning_rate": 3.893311157806091e-07,
75
- "logits/chosen": -3.0583784580230713,
76
- "logits/rejected": -3.054187536239624,
77
- "logps/chosen": -364.53973388671875,
78
- "logps/rejected": -349.81390380859375,
79
- "loss": 0.5569,
80
- "rewards/accuracies": 0.706250011920929,
81
- "rewards/chosen": 0.515569269657135,
82
- "rewards/margins": 0.7298178672790527,
83
- "rewards/rejected": -0.2142486274242401,
84
  "step": 40
85
  },
86
  {
87
  "epoch": 0.9478672985781991,
88
- "grad_norm": 64.27580122584432,
89
  "learning_rate": 3.126631330646801e-07,
90
- "logits/chosen": -3.055851459503174,
91
- "logits/rejected": -3.055706739425659,
92
- "logps/chosen": -368.67901611328125,
93
- "logps/rejected": -415.5660095214844,
94
- "loss": 0.493,
95
- "rewards/accuracies": 0.7749999761581421,
96
- "rewards/chosen": 0.567669153213501,
97
- "rewards/margins": 0.7747016549110413,
98
- "rewards/rejected": -0.2070324867963791,
99
  "step": 50
100
  },
101
  {
102
  "epoch": 0.985781990521327,
103
- "eval_logits/chosen": -3.087723970413208,
104
- "eval_logits/rejected": -3.104304313659668,
105
- "eval_logps/chosen": -326.6607666015625,
106
- "eval_logps/rejected": -338.8869323730469,
107
- "eval_loss": 0.5064198970794678,
108
- "eval_rewards/accuracies": 0.78125,
109
- "eval_rewards/chosen": 0.44371843338012695,
110
- "eval_rewards/margins": 0.8261151313781738,
111
- "eval_rewards/rejected": -0.3823966085910797,
112
- "eval_runtime": 69.4367,
113
- "eval_samples_per_second": 10.801,
114
- "eval_steps_per_second": 0.346,
115
  "step": 52
116
  },
117
  {
118
  "epoch": 1.1374407582938388,
119
- "grad_norm": 23.801523175944517,
120
  "learning_rate": 2.2891223348923882e-07,
121
- "logits/chosen": -3.084404945373535,
122
- "logits/rejected": -3.054302930831909,
123
- "logps/chosen": -351.47882080078125,
124
- "logps/rejected": -398.0810852050781,
125
- "loss": 0.2739,
126
- "rewards/accuracies": 0.887499988079071,
127
- "rewards/chosen": 0.9971694946289062,
128
- "rewards/margins": 1.8351186513900757,
129
- "rewards/rejected": -0.8379490971565247,
130
  "step": 60
131
  },
132
  {
133
  "epoch": 1.3270142180094786,
134
- "grad_norm": 23.77088500982387,
135
  "learning_rate": 1.4754491880085317e-07,
136
- "logits/chosen": -3.0304384231567383,
137
- "logits/rejected": -3.0326294898986816,
138
- "logps/chosen": -318.4073791503906,
139
- "logps/rejected": -392.42071533203125,
140
- "loss": 0.1625,
141
- "rewards/accuracies": 0.9937499761581421,
142
- "rewards/chosen": 1.2244815826416016,
143
- "rewards/margins": 2.5673928260803223,
144
- "rewards/rejected": -1.3429110050201416,
145
  "step": 70
146
  },
147
  {
148
  "epoch": 1.5165876777251186,
149
- "grad_norm": 19.870466912542945,
150
  "learning_rate": 7.775827023107834e-08,
151
- "logits/chosen": -3.0366415977478027,
152
- "logits/rejected": -3.0298948287963867,
153
- "logps/chosen": -358.39984130859375,
154
- "logps/rejected": -446.78857421875,
155
- "loss": 0.1491,
156
- "rewards/accuracies": 0.987500011920929,
157
- "rewards/chosen": 1.2719032764434814,
158
- "rewards/margins": 2.9869067668914795,
159
- "rewards/rejected": -1.7150036096572876,
160
  "step": 80
161
  },
162
  {
163
  "epoch": 1.7061611374407581,
164
- "grad_norm": 19.556816629836018,
165
  "learning_rate": 2.7440387297912122e-08,
166
- "logits/chosen": -3.0141148567199707,
167
- "logits/rejected": -3.015047073364258,
168
- "logps/chosen": -346.9141540527344,
169
- "logps/rejected": -437.4671325683594,
170
- "loss": 0.1358,
171
- "rewards/accuracies": 0.987500011920929,
172
- "rewards/chosen": 1.1179364919662476,
173
- "rewards/margins": 2.981201410293579,
174
- "rewards/rejected": -1.863265037536621,
175
  "step": 90
176
  },
177
  {
178
  "epoch": 1.8957345971563981,
179
- "grad_norm": 18.965988787956388,
180
  "learning_rate": 2.27878296044029e-09,
181
- "logits/chosen": -2.9989800453186035,
182
- "logits/rejected": -3.0166327953338623,
183
- "logps/chosen": -332.5897216796875,
184
- "logps/rejected": -401.76165771484375,
185
- "loss": 0.1356,
186
- "rewards/accuracies": 0.9937499761581421,
187
- "rewards/chosen": 1.0875240564346313,
188
- "rewards/margins": 2.662875175476074,
189
- "rewards/rejected": -1.5753511190414429,
190
  "step": 100
191
  },
192
  {
193
  "epoch": 1.971563981042654,
194
- "eval_logits/chosen": -3.0187489986419678,
195
- "eval_logits/rejected": -3.038198232650757,
196
- "eval_logps/chosen": -331.7999267578125,
197
- "eval_logps/rejected": -347.3108215332031,
198
- "eval_loss": 0.48019182682037354,
199
- "eval_rewards/accuracies": 0.75,
200
- "eval_rewards/chosen": 0.18675994873046875,
201
- "eval_rewards/margins": 0.9903542399406433,
202
- "eval_rewards/rejected": -0.803594172000885,
203
- "eval_runtime": 69.326,
204
- "eval_samples_per_second": 10.818,
205
- "eval_steps_per_second": 0.346,
206
  "step": 104
207
  },
208
  {
209
  "epoch": 1.971563981042654,
210
  "step": 104,
211
  "total_flos": 0.0,
212
- "train_loss": 0.373156875371933,
213
- "train_runtime": 3298.3137,
214
- "train_samples_per_second": 4.093,
215
- "train_steps_per_second": 0.032
216
  }
217
  ],
218
  "logging_steps": 10,
@@ -233,7 +233,7 @@
233
  }
234
  },
235
  "total_flos": 0.0,
236
- "train_batch_size": 4,
237
  "trial_name": null,
238
  "trial_params": null
239
  }
 
10
  "log_history": [
11
  {
12
  "epoch": 0.018957345971563982,
13
+ "grad_norm": 64.29333218701703,
14
  "learning_rate": 4.545454545454545e-08,
15
+ "logits/chosen": 260.044921875,
16
+ "logits/rejected": 273.92498779296875,
17
+ "logps/chosen": -421.27789306640625,
18
+ "logps/rejected": -408.871337890625,
19
  "loss": 0.6931,
20
  "rewards/accuracies": 0.0,
21
  "rewards/chosen": 0.0,
 
25
  },
26
  {
27
  "epoch": 0.1895734597156398,
28
+ "grad_norm": 66.1506989600867,
29
  "learning_rate": 4.545454545454545e-07,
30
+ "logits/chosen": 282.578857421875,
31
+ "logits/rejected": 277.1462097167969,
32
+ "logps/chosen": -362.7806091308594,
33
+ "logps/rejected": -423.041015625,
34
+ "loss": 0.7032,
35
+ "rewards/accuracies": 0.5347222089767456,
36
+ "rewards/chosen": 0.04673011228442192,
37
+ "rewards/margins": 0.09045842289924622,
38
+ "rewards/rejected": -0.043728310614824295,
39
  "step": 10
40
  },
41
  {
42
  "epoch": 0.3791469194312796,
43
+ "grad_norm": 69.27857567048012,
44
  "learning_rate": 4.885348141000122e-07,
45
+ "logits/chosen": 276.78228759765625,
46
+ "logits/rejected": 275.58184814453125,
47
+ "logps/chosen": -364.314208984375,
48
+ "logps/rejected": -391.9237365722656,
49
+ "loss": 0.6778,
50
+ "rewards/accuracies": 0.668749988079071,
51
+ "rewards/chosen": 0.04571449011564255,
52
+ "rewards/margins": 0.1900663673877716,
53
+ "rewards/rejected": -0.14435191452503204,
54
  "step": 20
55
  },
56
  {
57
  "epoch": 0.5687203791469194,
58
+ "grad_norm": 56.71129503742715,
59
  "learning_rate": 4.5025027361734613e-07,
60
+ "logits/chosen": 276.2879333496094,
61
+ "logits/rejected": 274.1214904785156,
62
+ "logps/chosen": -326.874755859375,
63
+ "logps/rejected": -380.1121520996094,
64
+ "loss": 0.6173,
65
+ "rewards/accuracies": 0.6812499761581421,
66
+ "rewards/chosen": -0.22750525176525116,
67
+ "rewards/margins": 0.4899619221687317,
68
+ "rewards/rejected": -0.7174672484397888,
69
  "step": 30
70
  },
71
  {
72
  "epoch": 0.7582938388625592,
73
+ "grad_norm": 57.871031614943476,
74
  "learning_rate": 3.893311157806091e-07,
75
+ "logits/chosen": 277.44049072265625,
76
+ "logits/rejected": 286.52374267578125,
77
+ "logps/chosen": -351.580078125,
78
+ "logps/rejected": -366.3487854003906,
79
+ "loss": 0.5947,
80
+ "rewards/accuracies": 0.6937500238418579,
81
+ "rewards/chosen": -0.05702618509531021,
82
+ "rewards/margins": 0.7331444025039673,
83
+ "rewards/rejected": -0.790170431137085,
84
  "step": 40
85
  },
86
  {
87
  "epoch": 0.9478672985781991,
88
+ "grad_norm": 73.77852490676803,
89
  "learning_rate": 3.126631330646801e-07,
90
+ "logits/chosen": 275.7063293457031,
91
+ "logits/rejected": 279.91253662109375,
92
+ "logps/chosen": -378.2023010253906,
93
+ "logps/rejected": -397.1396179199219,
94
+ "loss": 0.5334,
95
+ "rewards/accuracies": 0.731249988079071,
96
+ "rewards/chosen": -0.07171835005283356,
97
+ "rewards/margins": 0.9270683526992798,
98
+ "rewards/rejected": -0.9987867474555969,
99
  "step": 50
100
  },
101
  {
102
  "epoch": 0.985781990521327,
103
+ "eval_logits/chosen": 326.6963806152344,
104
+ "eval_logits/rejected": 326.98321533203125,
105
+ "eval_logps/chosen": -352.98773193359375,
106
+ "eval_logps/rejected": -410.34893798828125,
107
+ "eval_loss": 0.5679783821105957,
108
+ "eval_rewards/accuracies": 0.6702127456665039,
109
+ "eval_rewards/chosen": -0.2470909059047699,
110
+ "eval_rewards/margins": 0.52419114112854,
111
+ "eval_rewards/rejected": -0.7712819576263428,
112
+ "eval_runtime": 90.9704,
113
+ "eval_samples_per_second": 8.244,
114
+ "eval_steps_per_second": 0.517,
115
  "step": 52
116
  },
117
  {
118
  "epoch": 1.1374407582938388,
119
+ "grad_norm": 31.05292858682528,
120
  "learning_rate": 2.2891223348923882e-07,
121
+ "logits/chosen": 274.8268737792969,
122
+ "logits/rejected": 274.29949951171875,
123
+ "logps/chosen": -326.9039611816406,
124
+ "logps/rejected": -397.5994567871094,
125
+ "loss": 0.3923,
126
+ "rewards/accuracies": 0.824999988079071,
127
+ "rewards/chosen": 0.19279229640960693,
128
+ "rewards/margins": 1.5808745622634888,
129
+ "rewards/rejected": -1.3880822658538818,
130
  "step": 60
131
  },
132
  {
133
  "epoch": 1.3270142180094786,
134
+ "grad_norm": 33.93405312466621,
135
  "learning_rate": 1.4754491880085317e-07,
136
+ "logits/chosen": 279.77435302734375,
137
+ "logits/rejected": 278.4494934082031,
138
+ "logps/chosen": -331.9073486328125,
139
+ "logps/rejected": -370.1907653808594,
140
+ "loss": 0.2876,
141
+ "rewards/accuracies": 0.925000011920929,
142
+ "rewards/chosen": 0.4298977851867676,
143
+ "rewards/margins": 2.020857334136963,
144
+ "rewards/rejected": -1.5909594297409058,
145
  "step": 70
146
  },
147
  {
148
  "epoch": 1.5165876777251186,
149
+ "grad_norm": 23.235180840728248,
150
  "learning_rate": 7.775827023107834e-08,
151
+ "logits/chosen": 281.94732666015625,
152
+ "logits/rejected": 273.55474853515625,
153
+ "logps/chosen": -365.9495849609375,
154
+ "logps/rejected": -429.3880310058594,
155
+ "loss": 0.2574,
156
+ "rewards/accuracies": 0.9375,
157
+ "rewards/chosen": 0.3758580982685089,
158
+ "rewards/margins": 2.054001569747925,
159
+ "rewards/rejected": -1.6781437397003174,
160
  "step": 80
161
  },
162
  {
163
  "epoch": 1.7061611374407581,
164
+ "grad_norm": 26.571064677269206,
165
  "learning_rate": 2.7440387297912122e-08,
166
+ "logits/chosen": 276.1026916503906,
167
+ "logits/rejected": 270.5816345214844,
168
+ "logps/chosen": -347.1163330078125,
169
+ "logps/rejected": -437.5128479003906,
170
+ "loss": 0.2439,
171
+ "rewards/accuracies": 0.949999988079071,
172
+ "rewards/chosen": 0.39976122975349426,
173
+ "rewards/margins": 2.205505847930908,
174
+ "rewards/rejected": -1.8057447671890259,
175
  "step": 90
176
  },
177
  {
178
  "epoch": 1.8957345971563981,
179
+ "grad_norm": 25.094637702590482,
180
  "learning_rate": 2.27878296044029e-09,
181
+ "logits/chosen": 278.7438659667969,
182
+ "logits/rejected": 279.1758728027344,
183
+ "logps/chosen": -350.7715148925781,
184
+ "logps/rejected": -397.90484619140625,
185
+ "loss": 0.2495,
186
+ "rewards/accuracies": 0.9375,
187
+ "rewards/chosen": 0.19362537562847137,
188
+ "rewards/margins": 2.171940565109253,
189
+ "rewards/rejected": -1.9783153533935547,
190
  "step": 100
191
  },
192
  {
193
  "epoch": 1.971563981042654,
194
+ "eval_logits/chosen": 326.0050964355469,
195
+ "eval_logits/rejected": 326.4959411621094,
196
+ "eval_logps/chosen": -355.77618408203125,
197
+ "eval_logps/rejected": -415.0506286621094,
198
+ "eval_loss": 0.548955500125885,
199
+ "eval_rewards/accuracies": 0.6914893388748169,
200
+ "eval_rewards/chosen": -0.38651418685913086,
201
+ "eval_rewards/margins": 0.6198524832725525,
202
+ "eval_rewards/rejected": -1.0063666105270386,
203
+ "eval_runtime": 87.0456,
204
+ "eval_samples_per_second": 8.616,
205
+ "eval_steps_per_second": 0.54,
206
  "step": 104
207
  },
208
  {
209
  "epoch": 1.971563981042654,
210
  "step": 104,
211
  "total_flos": 0.0,
212
+ "train_loss": 0.44703544962864655,
213
+ "train_runtime": 4801.7476,
214
+ "train_samples_per_second": 2.811,
215
+ "train_steps_per_second": 0.022
216
  }
217
  ],
218
  "logging_steps": 10,
 
233
  }
234
  },
235
  "total_flos": 0.0,
236
+ "train_batch_size": 1,
237
  "trial_name": null,
238
  "trial_params": null
239
  }