eeeebbb2 commited on
Commit
723a5dc
·
verified ·
1 Parent(s): 21023a4

Training in progress, step 32, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8c59218aeed30c8b4e4d689a63794cee1afc92cbe005871cb267aa12eddebe28
3
  size 335604696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f64f44255fc6a71075712992b10e7db97e69c91efa5e69776ec2fc144a4efb3b
3
  size 335604696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a2a0f559bdd44b5b313e2769f89c5a87aeb83788402322c137889ceed21eef3f
3
  size 671466706
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c4901ff228d37f80b8593785a4d03ec9a436ef383d867132df3f3fed6f5f8db7
3
  size 671466706
last-checkpoint/rng_state_0.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5b7c72f98dacab526d7f08eec0014de02c536c465ede32e975e11056d306de6b
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2b40f8c21535d5cb3040ebd6098de18bc4b54d3919e39854677564dd2248fca8
3
  size 15024
last-checkpoint/rng_state_1.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b149256f5213147bb7dda2c43f9b7eba4882d860494af34e39bbc6c6d7e0e1c0
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a99fbc57599ab28bd8c12c1c1421120e46a75f8dbf99fa457773e219d0648ec
3
  size 15024
last-checkpoint/rng_state_2.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:89e31681cbdbe7be2a44f51e44e9fc4bea2061ee88a5c23a0ca17909c99e6639
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c5c819c7818a631c0cb290b3977e95c7f54598a017022454f25e1aeda1c08663
3
  size 15024
last-checkpoint/rng_state_3.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:483761894f386ce5c613b15225f8f41e70cee47fb6633ab8084f8e3faa2fffff
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eedcb0330f97ca4e8d8dc31801d05424a6942b71102d32b41ed9c1b09503809c
3
  size 15024
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e52d696b576c80447b09be7d8794036033efac5da11af94320214f843d978cbe
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:07c71024bc4e99ca36ceb2f728e261e6c0a3853b01e7c038fbdeb0dde69b327f
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 1.1290745735168457,
3
  "best_model_checkpoint": "miner_id_24/checkpoint-25",
4
- "epoch": 2.4188790560471976,
5
  "eval_steps": 25,
6
- "global_step": 25,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -198,6 +198,55 @@
198
  "eval_samples_per_second": 11.465,
199
  "eval_steps_per_second": 2.981,
200
  "step": 25
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
201
  }
202
  ],
203
  "logging_steps": 1,
@@ -221,12 +270,12 @@
221
  "should_evaluate": false,
222
  "should_log": false,
223
  "should_save": true,
224
- "should_training_stop": false
225
  },
226
  "attributes": {}
227
  }
228
  },
229
- "total_flos": 2.984041808658432e+17,
230
  "train_batch_size": 1,
231
  "trial_name": null,
232
  "trial_params": null
 
1
  {
2
  "best_metric": 1.1290745735168457,
3
  "best_model_checkpoint": "miner_id_24/checkpoint-25",
4
+ "epoch": 3.1091445427728615,
5
  "eval_steps": 25,
6
+ "global_step": 32,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
198
  "eval_samples_per_second": 11.465,
199
  "eval_steps_per_second": 2.981,
200
  "step": 25
201
+ },
202
+ {
203
+ "epoch": 2.5132743362831858,
204
+ "grad_norm": 0.819108784198761,
205
+ "learning_rate": 1.8594235253127375e-05,
206
+ "loss": 0.9056,
207
+ "step": 26
208
+ },
209
+ {
210
+ "epoch": 2.607669616519174,
211
+ "grad_norm": 0.766586184501648,
212
+ "learning_rate": 1.602885682970026e-05,
213
+ "loss": 0.8938,
214
+ "step": 27
215
+ },
216
+ {
217
+ "epoch": 2.702064896755162,
218
+ "grad_norm": 0.7301772832870483,
219
+ "learning_rate": 1.389045440608296e-05,
220
+ "loss": 0.6849,
221
+ "step": 28
222
+ },
223
+ {
224
+ "epoch": 2.7964601769911503,
225
+ "grad_norm": 0.8863487839698792,
226
+ "learning_rate": 1.2202456766718093e-05,
227
+ "loss": 1.1958,
228
+ "step": 29
229
+ },
230
+ {
231
+ "epoch": 2.8908554572271385,
232
+ "grad_norm": 0.8773151636123657,
233
+ "learning_rate": 1.0983357966978745e-05,
234
+ "loss": 0.8598,
235
+ "step": 30
236
+ },
237
+ {
238
+ "epoch": 3.0147492625368733,
239
+ "grad_norm": 1.5916945934295654,
240
+ "learning_rate": 1.0246514708427702e-05,
241
+ "loss": 1.5241,
242
+ "step": 31
243
+ },
244
+ {
245
+ "epoch": 3.1091445427728615,
246
+ "grad_norm": 0.6731727123260498,
247
+ "learning_rate": 1e-05,
248
+ "loss": 0.8907,
249
+ "step": 32
250
  }
251
  ],
252
  "logging_steps": 1,
 
270
  "should_evaluate": false,
271
  "should_log": false,
272
  "should_save": true,
273
+ "should_training_stop": true
274
  },
275
  "attributes": {}
276
  }
277
  },
278
+ "total_flos": 3.819573515082793e+17,
279
  "train_batch_size": 1,
280
  "trial_name": null,
281
  "trial_params": null