|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.9795004547680682, |
|
"eval_steps": 500, |
|
"global_step": 21000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 1.301074743270874, |
|
"learning_rate": 0.00046641791044776124, |
|
"loss": 7.611, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"eval_loss": 5.238980293273926, |
|
"eval_runtime": 567.2757, |
|
"eval_samples_per_second": 99.477, |
|
"eval_steps_per_second": 3.11, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 1.1419352293014526, |
|
"learning_rate": 0.0009328358208955225, |
|
"loss": 4.4792, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"eval_loss": 4.0363240242004395, |
|
"eval_runtime": 568.283, |
|
"eval_samples_per_second": 99.301, |
|
"eval_steps_per_second": 3.104, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 1.1277365684509277, |
|
"learning_rate": 0.0009789877154220063, |
|
"loss": 3.6503, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"eval_loss": 3.468844175338745, |
|
"eval_runtime": 568.7397, |
|
"eval_samples_per_second": 99.221, |
|
"eval_steps_per_second": 3.102, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 1.1394764184951782, |
|
"learning_rate": 0.0009544406539991162, |
|
"loss": 3.2321, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"eval_loss": 3.1853878498077393, |
|
"eval_runtime": 568.7397, |
|
"eval_samples_per_second": 99.221, |
|
"eval_steps_per_second": 3.102, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 1.0943962335586548, |
|
"learning_rate": 0.0009298935925762261, |
|
"loss": 3.0165, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"eval_loss": 3.0076937675476074, |
|
"eval_runtime": 569.7751, |
|
"eval_samples_per_second": 99.041, |
|
"eval_steps_per_second": 3.096, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.9250391721725464, |
|
"learning_rate": 0.0009053465311533362, |
|
"loss": 2.8757, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"eval_loss": 2.8989241123199463, |
|
"eval_runtime": 569.8016, |
|
"eval_samples_per_second": 99.036, |
|
"eval_steps_per_second": 3.096, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.9219993948936462, |
|
"learning_rate": 0.0008807994697304462, |
|
"loss": 2.7831, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"eval_loss": 2.815277099609375, |
|
"eval_runtime": 569.7431, |
|
"eval_samples_per_second": 99.046, |
|
"eval_steps_per_second": 3.096, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 1.0010333061218262, |
|
"learning_rate": 0.0008562524083075562, |
|
"loss": 2.7098, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"eval_loss": 2.7535383701324463, |
|
"eval_runtime": 564.7218, |
|
"eval_samples_per_second": 99.927, |
|
"eval_steps_per_second": 3.124, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.9716792702674866, |
|
"learning_rate": 0.0008317053468846664, |
|
"loss": 2.6551, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"eval_loss": 2.7090485095977783, |
|
"eval_runtime": 566.8661, |
|
"eval_samples_per_second": 99.549, |
|
"eval_steps_per_second": 3.112, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.9213451743125916, |
|
"learning_rate": 0.0008071582854617764, |
|
"loss": 2.6152, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"eval_loss": 2.666491746902466, |
|
"eval_runtime": 565.311, |
|
"eval_samples_per_second": 99.823, |
|
"eval_steps_per_second": 3.12, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.8985408544540405, |
|
"learning_rate": 0.0007826112240388863, |
|
"loss": 2.575, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"eval_loss": 2.6350765228271484, |
|
"eval_runtime": 564.212, |
|
"eval_samples_per_second": 100.017, |
|
"eval_steps_per_second": 3.126, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.8632990717887878, |
|
"learning_rate": 0.0007580641626159963, |
|
"loss": 2.5464, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"eval_loss": 2.6097421646118164, |
|
"eval_runtime": 568.485, |
|
"eval_samples_per_second": 99.266, |
|
"eval_steps_per_second": 3.103, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.8819186687469482, |
|
"learning_rate": 0.0007335171011931065, |
|
"loss": 2.5183, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"eval_loss": 2.580737590789795, |
|
"eval_runtime": 568.2279, |
|
"eval_samples_per_second": 99.311, |
|
"eval_steps_per_second": 3.104, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.9012317657470703, |
|
"learning_rate": 0.0007089700397702165, |
|
"loss": 2.4977, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"eval_loss": 2.5541210174560547, |
|
"eval_runtime": 568.3683, |
|
"eval_samples_per_second": 99.286, |
|
"eval_steps_per_second": 3.104, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.9294862151145935, |
|
"learning_rate": 0.0006844229783473265, |
|
"loss": 2.4723, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"eval_loss": 2.535729169845581, |
|
"eval_runtime": 569.135, |
|
"eval_samples_per_second": 99.152, |
|
"eval_steps_per_second": 3.099, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.9035699367523193, |
|
"learning_rate": 0.0006598759169244364, |
|
"loss": 2.4549, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"eval_loss": 2.5194778442382812, |
|
"eval_runtime": 569.6481, |
|
"eval_samples_per_second": 99.063, |
|
"eval_steps_per_second": 3.097, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.9459491968154907, |
|
"learning_rate": 0.0006353779496243923, |
|
"loss": 2.4428, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"eval_loss": 2.502213716506958, |
|
"eval_runtime": 565.3877, |
|
"eval_samples_per_second": 99.809, |
|
"eval_steps_per_second": 3.12, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.8880784511566162, |
|
"learning_rate": 0.0006108308882015024, |
|
"loss": 2.4255, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"eval_loss": 2.4834067821502686, |
|
"eval_runtime": 567.6005, |
|
"eval_samples_per_second": 99.42, |
|
"eval_steps_per_second": 3.108, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.9531754851341248, |
|
"learning_rate": 0.0005862838267786123, |
|
"loss": 2.4092, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"eval_loss": 2.473543643951416, |
|
"eval_runtime": 561.7685, |
|
"eval_samples_per_second": 100.452, |
|
"eval_steps_per_second": 3.14, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.9443451762199402, |
|
"learning_rate": 0.0005617367653557224, |
|
"loss": 2.3859, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"eval_loss": 2.456810712814331, |
|
"eval_runtime": 563.5685, |
|
"eval_samples_per_second": 100.132, |
|
"eval_steps_per_second": 3.13, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.9646140933036804, |
|
"learning_rate": 0.0005372387980556782, |
|
"loss": 2.3753, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"eval_loss": 2.44191837310791, |
|
"eval_runtime": 563.9978, |
|
"eval_samples_per_second": 100.055, |
|
"eval_steps_per_second": 3.128, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 0.9498624801635742, |
|
"learning_rate": 0.0005126917366327883, |
|
"loss": 2.3625, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"eval_loss": 2.428806781768799, |
|
"eval_runtime": 563.9618, |
|
"eval_samples_per_second": 100.062, |
|
"eval_steps_per_second": 3.128, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.9920567870140076, |
|
"learning_rate": 0.0004881446752098984, |
|
"loss": 2.3547, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"eval_loss": 2.4165408611297607, |
|
"eval_runtime": 563.8141, |
|
"eval_samples_per_second": 100.088, |
|
"eval_steps_per_second": 3.129, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.9328219294548035, |
|
"learning_rate": 0.0004635976137870084, |
|
"loss": 2.3387, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"eval_loss": 2.40574312210083, |
|
"eval_runtime": 563.9451, |
|
"eval_samples_per_second": 100.065, |
|
"eval_steps_per_second": 3.128, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 0.9478664994239807, |
|
"learning_rate": 0.00043909964648696415, |
|
"loss": 2.3277, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"eval_loss": 2.390106201171875, |
|
"eval_runtime": 563.4302, |
|
"eval_samples_per_second": 100.156, |
|
"eval_steps_per_second": 3.131, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 0.9822357296943665, |
|
"learning_rate": 0.0004145525850640742, |
|
"loss": 2.3071, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"eval_loss": 2.380479335784912, |
|
"eval_runtime": 563.2217, |
|
"eval_samples_per_second": 100.193, |
|
"eval_steps_per_second": 3.132, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 0.9824458956718445, |
|
"learning_rate": 0.00039000552364118425, |
|
"loss": 2.3087, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"eval_loss": 2.366305112838745, |
|
"eval_runtime": 563.1953, |
|
"eval_samples_per_second": 100.198, |
|
"eval_steps_per_second": 3.132, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 1.0211139917373657, |
|
"learning_rate": 0.00036545846221829433, |
|
"loss": 2.283, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"eval_loss": 2.3575291633605957, |
|
"eval_runtime": 563.2425, |
|
"eval_samples_per_second": 100.19, |
|
"eval_steps_per_second": 3.132, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 1.0012933015823364, |
|
"learning_rate": 0.00034091140079540435, |
|
"loss": 2.2716, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"eval_loss": 2.3457164764404297, |
|
"eval_runtime": 562.9008, |
|
"eval_samples_per_second": 100.25, |
|
"eval_steps_per_second": 3.134, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 1.0424264669418335, |
|
"learning_rate": 0.0003164134334953601, |
|
"loss": 2.2709, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"eval_loss": 2.3343236446380615, |
|
"eval_runtime": 562.6281, |
|
"eval_samples_per_second": 100.299, |
|
"eval_steps_per_second": 3.135, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 1.1129311323165894, |
|
"learning_rate": 0.00029186637207247017, |
|
"loss": 2.2584, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"eval_loss": 2.3277084827423096, |
|
"eval_runtime": 574.0103, |
|
"eval_samples_per_second": 98.31, |
|
"eval_steps_per_second": 3.073, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 1.0602174997329712, |
|
"learning_rate": 0.0002673193106495802, |
|
"loss": 2.2509, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"eval_loss": 2.3144421577453613, |
|
"eval_runtime": 566.0389, |
|
"eval_samples_per_second": 99.695, |
|
"eval_steps_per_second": 3.116, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 1.1678231954574585, |
|
"learning_rate": 0.00024277224922669027, |
|
"loss": 2.237, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"eval_loss": 2.305784225463867, |
|
"eval_runtime": 563.0281, |
|
"eval_samples_per_second": 100.228, |
|
"eval_steps_per_second": 3.133, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 1.1629687547683716, |
|
"learning_rate": 0.000218274281926646, |
|
"loss": 2.2249, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"eval_loss": 2.2936198711395264, |
|
"eval_runtime": 562.9504, |
|
"eval_samples_per_second": 100.242, |
|
"eval_steps_per_second": 3.133, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 1.0790985822677612, |
|
"learning_rate": 0.00019372722050375605, |
|
"loss": 2.219, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"eval_loss": 2.2830076217651367, |
|
"eval_runtime": 562.6154, |
|
"eval_samples_per_second": 100.301, |
|
"eval_steps_per_second": 3.135, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 1.157651424407959, |
|
"learning_rate": 0.0001691801590808661, |
|
"loss": 2.1999, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"eval_loss": 2.2722585201263428, |
|
"eval_runtime": 562.9469, |
|
"eval_samples_per_second": 100.242, |
|
"eval_steps_per_second": 3.134, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 1.0704936981201172, |
|
"learning_rate": 0.00014463309765797616, |
|
"loss": 2.1999, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"eval_loss": 2.2624754905700684, |
|
"eval_runtime": 562.9358, |
|
"eval_samples_per_second": 100.244, |
|
"eval_steps_per_second": 3.134, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 1.1210389137268066, |
|
"learning_rate": 0.00012008603623508621, |
|
"loss": 2.1898, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"eval_loss": 2.254652976989746, |
|
"eval_runtime": 563.3356, |
|
"eval_samples_per_second": 100.173, |
|
"eval_steps_per_second": 3.131, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 1.0510447025299072, |
|
"learning_rate": 9.558806893504194e-05, |
|
"loss": 2.1796, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"eval_loss": 2.2447702884674072, |
|
"eval_runtime": 563.3055, |
|
"eval_samples_per_second": 100.178, |
|
"eval_steps_per_second": 3.132, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 1.0447243452072144, |
|
"learning_rate": 7.1041007512152e-05, |
|
"loss": 2.1649, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"eval_loss": 2.236150026321411, |
|
"eval_runtime": 563.3424, |
|
"eval_samples_per_second": 100.172, |
|
"eval_steps_per_second": 3.131, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 1.039281964302063, |
|
"learning_rate": 4.649394608926206e-05, |
|
"loss": 2.1606, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"eval_loss": 2.228179693222046, |
|
"eval_runtime": 563.3883, |
|
"eval_samples_per_second": 100.164, |
|
"eval_steps_per_second": 3.131, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 1.0719681978225708, |
|
"learning_rate": 2.1946884666372103e-05, |
|
"loss": 2.1521, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"eval_loss": 2.221137285232544, |
|
"eval_runtime": 563.1866, |
|
"eval_samples_per_second": 100.199, |
|
"eval_steps_per_second": 3.132, |
|
"step": 21000 |
|
} |
|
], |
|
"logging_steps": 500, |
|
"max_steps": 21439, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 7000, |
|
"total_flos": 4102529679360000.0, |
|
"train_batch_size": 32, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|