|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.99972616910579, |
|
"eval_steps": 500, |
|
"global_step": 24648, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 17.19582176208496, |
|
"learning_rate": 1.0101419878296147e-05, |
|
"loss": 3.88, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 22.53339385986328, |
|
"learning_rate": 2.0223123732251522e-05, |
|
"loss": 1.9753, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 20.645118713378906, |
|
"learning_rate": 3.0344827586206897e-05, |
|
"loss": 1.4051, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 11.275988578796387, |
|
"learning_rate": 4.0486815415821506e-05, |
|
"loss": 1.2663, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 9.870928764343262, |
|
"learning_rate": 4.993012667357887e-05, |
|
"loss": 1.2106, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 13.5437650680542, |
|
"learning_rate": 4.880313753775414e-05, |
|
"loss": 1.179, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 14.121590614318848, |
|
"learning_rate": 4.767614840192941e-05, |
|
"loss": 1.1044, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 11.260817527770996, |
|
"learning_rate": 4.6549159266104676e-05, |
|
"loss": 1.043, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 15.443611145019531, |
|
"learning_rate": 4.542217013027995e-05, |
|
"loss": 1.0212, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 14.637772560119629, |
|
"learning_rate": 4.429518099445522e-05, |
|
"loss": 0.9764, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 18.14031410217285, |
|
"learning_rate": 4.316819185863049e-05, |
|
"loss": 0.9848, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 10.293859481811523, |
|
"learning_rate": 4.204120272280576e-05, |
|
"loss": 0.9378, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 10.198644638061523, |
|
"learning_rate": 4.0916467565252675e-05, |
|
"loss": 0.9363, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 9.705594062805176, |
|
"learning_rate": 3.978947842942794e-05, |
|
"loss": 0.9482, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 9.531152725219727, |
|
"learning_rate": 3.866248929360321e-05, |
|
"loss": 0.9033, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 11.644086837768555, |
|
"learning_rate": 3.753550015777848e-05, |
|
"loss": 0.9042, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"grad_norm": 9.761592864990234, |
|
"learning_rate": 3.640851102195375e-05, |
|
"loss": 0.7736, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"grad_norm": 8.461353302001953, |
|
"learning_rate": 3.528152188612902e-05, |
|
"loss": 0.6728, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 7.326571941375732, |
|
"learning_rate": 3.415453275030429e-05, |
|
"loss": 0.6811, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"grad_norm": 7.815227031707764, |
|
"learning_rate": 3.3027543614479556e-05, |
|
"loss": 0.6769, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 11.3690824508667, |
|
"learning_rate": 3.190055447865483e-05, |
|
"loss": 0.6725, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"grad_norm": 16.308076858520508, |
|
"learning_rate": 3.077581932110174e-05, |
|
"loss": 0.6596, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 8.466190338134766, |
|
"learning_rate": 2.9651084163548663e-05, |
|
"loss": 0.6683, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"grad_norm": 15.437102317810059, |
|
"learning_rate": 2.8524095027723934e-05, |
|
"loss": 0.6908, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"grad_norm": 7.6999406814575195, |
|
"learning_rate": 2.7397105891899204e-05, |
|
"loss": 0.6875, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"grad_norm": 9.09485912322998, |
|
"learning_rate": 2.6270116756074475e-05, |
|
"loss": 0.6514, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"grad_norm": 9.5618314743042, |
|
"learning_rate": 2.514312762024974e-05, |
|
"loss": 0.6452, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"grad_norm": 11.317076683044434, |
|
"learning_rate": 2.401613848442501e-05, |
|
"loss": 0.6474, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"grad_norm": 9.310576438903809, |
|
"learning_rate": 2.288914934860028e-05, |
|
"loss": 0.645, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"grad_norm": 10.67317008972168, |
|
"learning_rate": 2.176216021277555e-05, |
|
"loss": 0.6259, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"grad_norm": 9.449134826660156, |
|
"learning_rate": 2.0635171076950818e-05, |
|
"loss": 0.6342, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"grad_norm": 5.870083808898926, |
|
"learning_rate": 1.9510435919397738e-05, |
|
"loss": 0.6219, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"grad_norm": 12.512371063232422, |
|
"learning_rate": 1.838344678357301e-05, |
|
"loss": 0.6211, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"grad_norm": 7.509356498718262, |
|
"learning_rate": 1.7256457647748276e-05, |
|
"loss": 0.416, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"grad_norm": 28.715145111083984, |
|
"learning_rate": 1.6129468511923546e-05, |
|
"loss": 0.4208, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"grad_norm": 5.760591983795166, |
|
"learning_rate": 1.5002479376098815e-05, |
|
"loss": 0.4239, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"grad_norm": 7.888522624969482, |
|
"learning_rate": 1.3877744218545735e-05, |
|
"loss": 0.4135, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"grad_norm": 16.289827346801758, |
|
"learning_rate": 1.2750755082721002e-05, |
|
"loss": 0.4387, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"grad_norm": 12.066337585449219, |
|
"learning_rate": 1.1623765946896273e-05, |
|
"loss": 0.4299, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"grad_norm": 12.49575424194336, |
|
"learning_rate": 1.0496776811071542e-05, |
|
"loss": 0.4255, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"grad_norm": 9.139410972595215, |
|
"learning_rate": 9.369787675246811e-06, |
|
"loss": 0.4237, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"grad_norm": 18.495920181274414, |
|
"learning_rate": 8.24505251769373e-06, |
|
"loss": 0.4293, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"grad_norm": 22.488752365112305, |
|
"learning_rate": 7.118063381868999e-06, |
|
"loss": 0.4201, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"grad_norm": 3.9392380714416504, |
|
"learning_rate": 5.9910742460442686e-06, |
|
"loss": 0.3984, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"grad_norm": 5.216261386871338, |
|
"learning_rate": 4.8640851102195375e-06, |
|
"loss": 0.4006, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"grad_norm": 8.506802558898926, |
|
"learning_rate": 3.7393499526664564e-06, |
|
"loss": 0.4085, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"grad_norm": 15.280343055725098, |
|
"learning_rate": 2.6123608168417258e-06, |
|
"loss": 0.4029, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"grad_norm": 22.39061164855957, |
|
"learning_rate": 1.485371681016995e-06, |
|
"loss": 0.4135, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"grad_norm": 4.789835453033447, |
|
"learning_rate": 3.583825451922644e-07, |
|
"loss": 0.4149, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"step": 24648, |
|
"total_flos": 3.86452500250624e+16, |
|
"train_loss": 0.7860646006427233, |
|
"train_runtime": 13844.0651, |
|
"train_samples_per_second": 28.489, |
|
"train_steps_per_second": 1.78 |
|
} |
|
], |
|
"logging_steps": 500, |
|
"max_steps": 24648, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 5000, |
|
"total_flos": 3.86452500250624e+16, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|