alchemist69 commited on
Commit
b7a6568
·
verified ·
1 Parent(s): e098f7a

Training in progress, step 2000, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a316a602ffe058f14451b1db7142b2a19d5828f2c0844203b31a8f747716b165
3
  size 18064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a56c8dc710d9f29e16868b51bd961a7de1a36f29310a1637e2e82ed3fc47b85
3
  size 18064
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e2fbb1c81ab2509a4dc949c5fb9cc762278a199fe5a04aca2c5ece9b290b8153
3
  size 43578
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d11ddfef92c94b87f0213162a6e3bdbc1fa097b9bc1c417dc5d323d7570ffed
3
  size 43578
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b144f7cb70aafc2e7b54f7a16659cc872d3061fb4e845c6cb04f1164d9a6048d
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:553dcd59c778ca530dcc80ab66bd21953efec38b342a78fa029e7bb7063620bb
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:41a6da4479e4d6e904a1d1657795824ceae911abb9c58383db394b5c0579911a
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a466a28b4a8a9c102025ce436fa94b7198b461eca0d4d66f86eed6792700b45
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 11.5,
3
  "best_model_checkpoint": "miner_id_24/checkpoint-500",
4
- "epoch": 0.014698821154543405,
5
  "eval_steps": 500,
6
- "global_step": 1500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -249,6 +249,84 @@
249
  "eval_samples_per_second": 176.878,
250
  "eval_steps_per_second": 88.439,
251
  "step": 1500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
252
  }
253
  ],
254
  "logging_steps": 50,
@@ -263,7 +341,7 @@
263
  "early_stopping_threshold": 0.0
264
  },
265
  "attributes": {
266
- "early_stopping_patience_counter": 2
267
  }
268
  },
269
  "TrainerControl": {
@@ -272,12 +350,12 @@
272
  "should_evaluate": false,
273
  "should_log": false,
274
  "should_save": true,
275
- "should_training_stop": false
276
  },
277
  "attributes": {}
278
  }
279
  },
280
- "total_flos": 30067558219776.0,
281
  "train_batch_size": 2,
282
  "trial_name": null,
283
  "trial_params": null
 
1
  {
2
  "best_metric": 11.5,
3
  "best_model_checkpoint": "miner_id_24/checkpoint-500",
4
+ "epoch": 0.019598428206057873,
5
  "eval_steps": 500,
6
+ "global_step": 2000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
249
  "eval_samples_per_second": 176.878,
250
  "eval_steps_per_second": 88.439,
251
  "step": 1500
252
+ },
253
+ {
254
+ "epoch": 0.015188781859694852,
255
+ "grad_norm": 0.0018345484277233481,
256
+ "learning_rate": 0.00015800569095711982,
257
+ "loss": 46.0,
258
+ "step": 1550
259
+ },
260
+ {
261
+ "epoch": 0.0156787425648463,
262
+ "grad_norm": 0.00938267633318901,
263
+ "learning_rate": 0.00015539200638661104,
264
+ "loss": 46.0,
265
+ "step": 1600
266
+ },
267
+ {
268
+ "epoch": 0.016168703269997746,
269
+ "grad_norm": 0.013782349415123463,
270
+ "learning_rate": 0.00015272254676105025,
271
+ "loss": 46.0,
272
+ "step": 1650
273
+ },
274
+ {
275
+ "epoch": 0.016658663975149195,
276
+ "grad_norm": 0.005259850528091192,
277
+ "learning_rate": 0.00015000000000000001,
278
+ "loss": 46.0,
279
+ "step": 1700
280
+ },
281
+ {
282
+ "epoch": 0.01714862468030064,
283
+ "grad_norm": 0.007433369290083647,
284
+ "learning_rate": 0.0001472271074772683,
285
+ "loss": 46.0,
286
+ "step": 1750
287
+ },
288
+ {
289
+ "epoch": 0.017638585385452085,
290
+ "grad_norm": 0.003847072832286358,
291
+ "learning_rate": 0.00014440666126057744,
292
+ "loss": 46.0,
293
+ "step": 1800
294
+ },
295
+ {
296
+ "epoch": 0.018128546090603534,
297
+ "grad_norm": 0.0026508287992328405,
298
+ "learning_rate": 0.00014154150130018866,
299
+ "loss": 46.0,
300
+ "step": 1850
301
+ },
302
+ {
303
+ "epoch": 0.01861850679575498,
304
+ "grad_norm": 0.005416387226432562,
305
+ "learning_rate": 0.00013863451256931287,
306
+ "loss": 46.0,
307
+ "step": 1900
308
+ },
309
+ {
310
+ "epoch": 0.019108467500906428,
311
+ "grad_norm": 0.0011210152879357338,
312
+ "learning_rate": 0.00013568862215918717,
313
+ "loss": 46.0,
314
+ "step": 1950
315
+ },
316
+ {
317
+ "epoch": 0.019598428206057873,
318
+ "grad_norm": 0.004934302996844053,
319
+ "learning_rate": 0.00013270679633174218,
320
+ "loss": 46.0,
321
+ "step": 2000
322
+ },
323
+ {
324
+ "epoch": 0.019598428206057873,
325
+ "eval_loss": 11.5,
326
+ "eval_runtime": 239.7758,
327
+ "eval_samples_per_second": 179.201,
328
+ "eval_steps_per_second": 89.6,
329
+ "step": 2000
330
  }
331
  ],
332
  "logging_steps": 50,
 
341
  "early_stopping_threshold": 0.0
342
  },
343
  "attributes": {
344
+ "early_stopping_patience_counter": 3
345
  }
346
  },
347
  "TrainerControl": {
 
350
  "should_evaluate": false,
351
  "should_log": false,
352
  "should_save": true,
353
+ "should_training_stop": true
354
  },
355
  "attributes": {}
356
  }
357
  },
358
+ "total_flos": 40091744157696.0,
359
  "train_batch_size": 2,
360
  "trial_name": null,
361
  "trial_params": null