crowbarmassage commited on
Commit
0fe9888
·
verified ·
1 Parent(s): 23ba597

Upload DeepCoder14B_DSPy LoRA adapter checkpoint

Browse files
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:229486ce1d4253a54d0cd47c5005c0eacd850b87a95cd3216a73e019d10945c6
3
  size 6799256256
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5dc489c7ab4ab92650304b6f75c652bde544d2c53f17b9cc633f1cf12789c71d
3
  size 6799256256
optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5062a362186c064ec8701c9f44214104908f82eb3da4e29d6a4dcd1e277acdcb
3
  size 6519149878
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:18dfdd7067e9a4837aa29ad3b34d650c432e5374748c5580051b51610e4e9175
3
  size 6519149878
rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7642d4d8677c028eb706576fdd39c921ffb1007ec52044603d519b96b28ba821
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c59efed336e3af92ec9b9410099a59c2927ff1388b85211c0161b5fff7836a2
3
  size 14244
scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:49f276f737c8546acf52b47aa0a82341e11a9aa193441ce56e73d93c1135b634
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:54b8ad803d85e44a0c1ade1fcda24008ebe595e19cb0f6fef0c012f10d4f5f35
3
  size 1064
trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": 275,
3
  "best_metric": 0.9421520233154297,
4
  "best_model_checkpoint": "./dspy-8bit-lora-deepcoder-safe/checkpoint-275",
5
- "epoch": 1.9415929203539823,
6
  "eval_steps": 25,
7
- "global_step": 275,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -217,6 +217,120 @@
217
  "eval_samples_per_second": 2.059,
218
  "eval_steps_per_second": 2.059,
219
  "step": 275
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
220
  }
221
  ],
222
  "logging_steps": 25,
@@ -236,7 +350,7 @@
236
  "attributes": {}
237
  }
238
  },
239
- "total_flos": 5.329544702974157e+16,
240
  "train_batch_size": 1,
241
  "trial_name": null,
242
  "trial_params": null
 
2
  "best_global_step": 275,
3
  "best_metric": 0.9421520233154297,
4
  "best_model_checkpoint": "./dspy-8bit-lora-deepcoder-safe/checkpoint-275",
5
+ "epoch": 2.9982300884955753,
6
  "eval_steps": 25,
7
+ "global_step": 425,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
217
  "eval_samples_per_second": 2.059,
218
  "eval_steps_per_second": 2.059,
219
  "step": 275
220
+ },
221
+ {
222
+ "epoch": 2.113274336283186,
223
+ "grad_norm": 1.1352859735488892,
224
+ "learning_rate": 7.791411042944787e-05,
225
+ "loss": 0.5105,
226
+ "mean_token_accuracy": 0.8593325974400511,
227
+ "num_tokens": 680687.0,
228
+ "step": 300
229
+ },
230
+ {
231
+ "epoch": 2.113274336283186,
232
+ "eval_loss": 0.9843020439147949,
233
+ "eval_mean_token_accuracy": 0.7742766647040844,
234
+ "eval_num_tokens": 680687.0,
235
+ "eval_runtime": 98.2979,
236
+ "eval_samples_per_second": 2.035,
237
+ "eval_steps_per_second": 2.035,
238
+ "step": 300
239
+ },
240
+ {
241
+ "epoch": 2.2902654867256635,
242
+ "grad_norm": 0.8646391034126282,
243
+ "learning_rate": 6.257668711656443e-05,
244
+ "loss": 0.4312,
245
+ "mean_token_accuracy": 0.878243299126625,
246
+ "num_tokens": 733979.0,
247
+ "step": 325
248
+ },
249
+ {
250
+ "epoch": 2.2902654867256635,
251
+ "eval_loss": 0.993569552898407,
252
+ "eval_mean_token_accuracy": 0.775199833214283,
253
+ "eval_num_tokens": 733979.0,
254
+ "eval_runtime": 96.9581,
255
+ "eval_samples_per_second": 2.063,
256
+ "eval_steps_per_second": 2.063,
257
+ "step": 325
258
+ },
259
+ {
260
+ "epoch": 2.4672566371681417,
261
+ "grad_norm": 1.0304445028305054,
262
+ "learning_rate": 4.723926380368098e-05,
263
+ "loss": 0.3952,
264
+ "mean_token_accuracy": 0.8835302656888961,
265
+ "num_tokens": 789649.0,
266
+ "step": 350
267
+ },
268
+ {
269
+ "epoch": 2.4672566371681417,
270
+ "eval_loss": 0.9702684283256531,
271
+ "eval_mean_token_accuracy": 0.77936582878232,
272
+ "eval_num_tokens": 789649.0,
273
+ "eval_runtime": 98.4219,
274
+ "eval_samples_per_second": 2.032,
275
+ "eval_steps_per_second": 2.032,
276
+ "step": 350
277
+ },
278
+ {
279
+ "epoch": 2.6442477876106194,
280
+ "grad_norm": 0.8441785573959351,
281
+ "learning_rate": 3.1901840490797544e-05,
282
+ "loss": 0.3966,
283
+ "mean_token_accuracy": 0.8844721549749375,
284
+ "num_tokens": 845295.0,
285
+ "step": 375
286
+ },
287
+ {
288
+ "epoch": 2.6442477876106194,
289
+ "eval_loss": 0.9620086550712585,
290
+ "eval_mean_token_accuracy": 0.7812778241932392,
291
+ "eval_num_tokens": 845295.0,
292
+ "eval_runtime": 97.2727,
293
+ "eval_samples_per_second": 2.056,
294
+ "eval_steps_per_second": 2.056,
295
+ "step": 375
296
+ },
297
+ {
298
+ "epoch": 2.8212389380530976,
299
+ "grad_norm": 0.8250017762184143,
300
+ "learning_rate": 1.656441717791411e-05,
301
+ "loss": 0.3743,
302
+ "mean_token_accuracy": 0.8831119546294213,
303
+ "num_tokens": 908151.0,
304
+ "step": 400
305
+ },
306
+ {
307
+ "epoch": 2.8212389380530976,
308
+ "eval_loss": 0.9527075290679932,
309
+ "eval_mean_token_accuracy": 0.7827157293260097,
310
+ "eval_num_tokens": 908151.0,
311
+ "eval_runtime": 98.1465,
312
+ "eval_samples_per_second": 2.038,
313
+ "eval_steps_per_second": 2.038,
314
+ "step": 400
315
+ },
316
+ {
317
+ "epoch": 2.9982300884955753,
318
+ "grad_norm": 0.8464552164077759,
319
+ "learning_rate": 1.226993865030675e-06,
320
+ "loss": 0.4002,
321
+ "mean_token_accuracy": 0.883520859181881,
322
+ "num_tokens": 967441.0,
323
+ "step": 425
324
+ },
325
+ {
326
+ "epoch": 2.9982300884955753,
327
+ "eval_loss": 0.9509771466255188,
328
+ "eval_mean_token_accuracy": 0.7838174617290496,
329
+ "eval_num_tokens": 967441.0,
330
+ "eval_runtime": 98.6966,
331
+ "eval_samples_per_second": 2.026,
332
+ "eval_steps_per_second": 2.026,
333
+ "step": 425
334
  }
335
  ],
336
  "logging_steps": 25,
 
350
  "attributes": {}
351
  }
352
  },
353
+ "total_flos": 8.204357166595891e+16,
354
  "train_batch_size": 1,
355
  "trial_name": null,
356
  "trial_params": null