|
{ |
|
"best_global_step": null, |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.9961880559085133, |
|
"eval_steps": 49, |
|
"global_step": 196, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.005082592121982211, |
|
"grad_norm": 2.417446798898839, |
|
"learning_rate": 6.666666666666667e-07, |
|
"loss": 1.3227, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.005082592121982211, |
|
"eval_loss": NaN, |
|
"eval_runtime": 462.2137, |
|
"eval_samples_per_second": 3.014, |
|
"eval_steps_per_second": 0.379, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.010165184243964422, |
|
"grad_norm": 2.928692071056059, |
|
"learning_rate": 1.3333333333333334e-06, |
|
"loss": 1.3358, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.015247776365946633, |
|
"grad_norm": 2.295922498777242, |
|
"learning_rate": 2.0000000000000003e-06, |
|
"loss": 1.3668, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.020330368487928845, |
|
"grad_norm": 2.165639635427681, |
|
"learning_rate": 2.666666666666667e-06, |
|
"loss": 1.2876, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.025412960609911054, |
|
"grad_norm": 1.4335801076375077, |
|
"learning_rate": 3.3333333333333333e-06, |
|
"loss": 1.2307, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.030495552731893267, |
|
"grad_norm": 1.3021209574622363, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 1.2824, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.035578144853875476, |
|
"grad_norm": 1.233144475780061, |
|
"learning_rate": 4.666666666666667e-06, |
|
"loss": 1.2609, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.04066073697585769, |
|
"grad_norm": 1.8649443938694577, |
|
"learning_rate": 5.333333333333334e-06, |
|
"loss": 1.2592, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.045743329097839895, |
|
"grad_norm": 2.344947953156482, |
|
"learning_rate": 6e-06, |
|
"loss": 1.1416, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.05082592121982211, |
|
"grad_norm": 4.601477812008778, |
|
"learning_rate": 6.666666666666667e-06, |
|
"loss": 1.1554, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.05590851334180432, |
|
"grad_norm": 0.9781052406224989, |
|
"learning_rate": 7.333333333333334e-06, |
|
"loss": 1.1799, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.060991105463786534, |
|
"grad_norm": 1.5150887808544182, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 1.1938, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.06607369758576874, |
|
"grad_norm": 1.195420841394076, |
|
"learning_rate": 8.666666666666668e-06, |
|
"loss": 1.2875, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.07115628970775095, |
|
"grad_norm": 1.446747842249035, |
|
"learning_rate": 9.333333333333334e-06, |
|
"loss": 1.2113, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.07623888182973317, |
|
"grad_norm": 1.0206440434231885, |
|
"learning_rate": 1e-05, |
|
"loss": 1.1023, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.08132147395171538, |
|
"grad_norm": 0.7766830838305955, |
|
"learning_rate": 1.0666666666666667e-05, |
|
"loss": 1.1893, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.08640406607369759, |
|
"grad_norm": 0.8562502894880828, |
|
"learning_rate": 1.1333333333333334e-05, |
|
"loss": 1.099, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.09148665819567979, |
|
"grad_norm": 1.1790195599158821, |
|
"learning_rate": 1.2e-05, |
|
"loss": 1.1266, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.096569250317662, |
|
"grad_norm": 1.1536312340591073, |
|
"learning_rate": 1.2666666666666667e-05, |
|
"loss": 1.1868, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.10165184243964422, |
|
"grad_norm": 0.7170693721812672, |
|
"learning_rate": 1.3333333333333333e-05, |
|
"loss": 1.1192, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.10673443456162643, |
|
"grad_norm": 0.960585656847019, |
|
"learning_rate": 1.4e-05, |
|
"loss": 1.1264, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.11181702668360864, |
|
"grad_norm": 0.8933902995648031, |
|
"learning_rate": 1.4666666666666668e-05, |
|
"loss": 1.0475, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.11689961880559085, |
|
"grad_norm": 0.9526763881562114, |
|
"learning_rate": 1.5333333333333334e-05, |
|
"loss": 1.0869, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.12198221092757307, |
|
"grad_norm": 0.8261066261177119, |
|
"learning_rate": 1.6000000000000003e-05, |
|
"loss": 1.1273, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.12706480304955528, |
|
"grad_norm": 1.0138316404290697, |
|
"learning_rate": 1.6666666666666667e-05, |
|
"loss": 1.0167, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.13214739517153748, |
|
"grad_norm": 0.6244135166489783, |
|
"learning_rate": 1.7333333333333336e-05, |
|
"loss": 1.0642, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.1372299872935197, |
|
"grad_norm": 0.6860413953391165, |
|
"learning_rate": 1.8e-05, |
|
"loss": 1.0598, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.1423125794155019, |
|
"grad_norm": 0.7123933488497386, |
|
"learning_rate": 1.866666666666667e-05, |
|
"loss": 1.1359, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.1473951715374841, |
|
"grad_norm": 0.8194224963685599, |
|
"learning_rate": 1.9333333333333333e-05, |
|
"loss": 1.0469, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.15247776365946633, |
|
"grad_norm": 0.7428981466577943, |
|
"learning_rate": 2e-05, |
|
"loss": 1.0621, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.15756035578144853, |
|
"grad_norm": 0.5978655906350417, |
|
"learning_rate": 1.9994461368042097e-05, |
|
"loss": 1.0809, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.16264294790343076, |
|
"grad_norm": 0.5523263409594675, |
|
"learning_rate": 1.9988895058300946e-05, |
|
"loss": 1.1207, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.16772554002541296, |
|
"grad_norm": 0.6386088808060907, |
|
"learning_rate": 1.998330086278876e-05, |
|
"loss": 1.1099, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.17280813214739518, |
|
"grad_norm": 0.5889449105044371, |
|
"learning_rate": 1.9977678571428572e-05, |
|
"loss": 1.1882, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.17789072426937738, |
|
"grad_norm": 0.6171122595004415, |
|
"learning_rate": 1.9972027972027975e-05, |
|
"loss": 1.1434, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.18297331639135958, |
|
"grad_norm": 0.660722632169665, |
|
"learning_rate": 1.9966348850252384e-05, |
|
"loss": 1.1197, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.1880559085133418, |
|
"grad_norm": 0.8384340606450145, |
|
"learning_rate": 1.9960640989597977e-05, |
|
"loss": 1.0796, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.193138500635324, |
|
"grad_norm": 0.6854031551884068, |
|
"learning_rate": 1.995490417136415e-05, |
|
"loss": 1.081, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.19822109275730623, |
|
"grad_norm": 0.6791992732368597, |
|
"learning_rate": 1.9949138174625602e-05, |
|
"loss": 1.1199, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.20330368487928843, |
|
"grad_norm": 0.5653035313976759, |
|
"learning_rate": 1.9943342776203965e-05, |
|
"loss": 1.1191, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.20838627700127066, |
|
"grad_norm": 0.7529491506419821, |
|
"learning_rate": 1.9937517750639022e-05, |
|
"loss": 1.1309, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.21346886912325286, |
|
"grad_norm": 2.739538155317824, |
|
"learning_rate": 1.9931662870159458e-05, |
|
"loss": 1.1405, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.21855146124523506, |
|
"grad_norm": 0.8235383414154238, |
|
"learning_rate": 1.9925777904653155e-05, |
|
"loss": 1.0874, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.22363405336721728, |
|
"grad_norm": 0.636431907641577, |
|
"learning_rate": 1.9919862621637095e-05, |
|
"loss": 1.1171, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.22871664548919948, |
|
"grad_norm": 0.5849746433127464, |
|
"learning_rate": 1.9913916786226688e-05, |
|
"loss": 1.0112, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.2337992376111817, |
|
"grad_norm": 0.771258264938101, |
|
"learning_rate": 1.990794016110472e-05, |
|
"loss": 1.0638, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.2388818297331639, |
|
"grad_norm": 0.7496683574824505, |
|
"learning_rate": 1.9901932506489762e-05, |
|
"loss": 0.9786, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.24396442185514614, |
|
"grad_norm": 0.8193989912254472, |
|
"learning_rate": 1.989589358010411e-05, |
|
"loss": 1.1322, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.24904701397712833, |
|
"grad_norm": 0.5316839323339139, |
|
"learning_rate": 1.98898231371412e-05, |
|
"loss": 1.0685, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.24904701397712833, |
|
"eval_loss": NaN, |
|
"eval_runtime": 454.8058, |
|
"eval_samples_per_second": 3.063, |
|
"eval_steps_per_second": 0.385, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.25412960609911056, |
|
"grad_norm": 0.5814548864442379, |
|
"learning_rate": 1.988372093023256e-05, |
|
"loss": 0.9883, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.25921219822109276, |
|
"grad_norm": 0.6816224250940385, |
|
"learning_rate": 1.9877586709414165e-05, |
|
"loss": 1.0678, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.26429479034307496, |
|
"grad_norm": 0.6659302304663588, |
|
"learning_rate": 1.9871420222092346e-05, |
|
"loss": 1.1103, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.26937738246505716, |
|
"grad_norm": 0.6557010282469508, |
|
"learning_rate": 1.9865221213009086e-05, |
|
"loss": 1.1027, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.2744599745870394, |
|
"grad_norm": 0.5102660591956714, |
|
"learning_rate": 1.9858989424206816e-05, |
|
"loss": 1.0463, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.2795425667090216, |
|
"grad_norm": 0.9092403386769214, |
|
"learning_rate": 1.985272459499264e-05, |
|
"loss": 1.1451, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.2846251588310038, |
|
"grad_norm": 0.6067477876422573, |
|
"learning_rate": 1.984642646190195e-05, |
|
"loss": 1.0269, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.289707750952986, |
|
"grad_norm": 0.9539627940086233, |
|
"learning_rate": 1.9840094758661536e-05, |
|
"loss": 1.035, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.2947903430749682, |
|
"grad_norm": 0.5662189783414849, |
|
"learning_rate": 1.983372921615202e-05, |
|
"loss": 1.0151, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.29987293519695046, |
|
"grad_norm": 0.657165599335774, |
|
"learning_rate": 1.9827329562369756e-05, |
|
"loss": 1.0643, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.30495552731893266, |
|
"grad_norm": 0.6751154617452281, |
|
"learning_rate": 1.9820895522388063e-05, |
|
"loss": 1.054, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.31003811944091486, |
|
"grad_norm": 0.6619429396351022, |
|
"learning_rate": 1.9814426818317872e-05, |
|
"loss": 1.078, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.31512071156289706, |
|
"grad_norm": 0.5939366111849416, |
|
"learning_rate": 1.980792316926771e-05, |
|
"loss": 1.0312, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.3202033036848793, |
|
"grad_norm": 0.6050698536675223, |
|
"learning_rate": 1.980138429130304e-05, |
|
"loss": 0.9909, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.3252858958068615, |
|
"grad_norm": 0.709448530717843, |
|
"learning_rate": 1.9794809897404953e-05, |
|
"loss": 1.0087, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.3303684879288437, |
|
"grad_norm": 0.669736059058813, |
|
"learning_rate": 1.978819969742814e-05, |
|
"loss": 1.0107, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.3354510800508259, |
|
"grad_norm": 0.6986784735462407, |
|
"learning_rate": 1.9781553398058257e-05, |
|
"loss": 0.9947, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.3405336721728081, |
|
"grad_norm": 1.1264664572165972, |
|
"learning_rate": 1.977487070276848e-05, |
|
"loss": 0.9921, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.34561626429479037, |
|
"grad_norm": 0.6700835793434858, |
|
"learning_rate": 1.9768151311775475e-05, |
|
"loss": 1.111, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.35069885641677256, |
|
"grad_norm": 0.6822724145331737, |
|
"learning_rate": 1.9761394921994493e-05, |
|
"loss": 1.0596, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.35578144853875476, |
|
"grad_norm": 0.6040759384639588, |
|
"learning_rate": 1.9754601226993868e-05, |
|
"loss": 1.1039, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.36086404066073696, |
|
"grad_norm": 0.6449973639426745, |
|
"learning_rate": 1.9747769916948632e-05, |
|
"loss": 1.0643, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.36594663278271916, |
|
"grad_norm": 1.3707597910260259, |
|
"learning_rate": 1.9740900678593462e-05, |
|
"loss": 1.0051, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.3710292249047014, |
|
"grad_norm": 0.8317766567975909, |
|
"learning_rate": 1.973399319517476e-05, |
|
"loss": 1.0358, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.3761118170266836, |
|
"grad_norm": 0.6836349542770239, |
|
"learning_rate": 1.972704714640199e-05, |
|
"loss": 0.9989, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.3811944091486658, |
|
"grad_norm": 0.672285267907023, |
|
"learning_rate": 1.9720062208398136e-05, |
|
"loss": 1.1016, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.386277001270648, |
|
"grad_norm": 0.6008642636107718, |
|
"learning_rate": 1.971303805364941e-05, |
|
"loss": 1.0148, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.39135959339263027, |
|
"grad_norm": 0.6137126435530088, |
|
"learning_rate": 1.9705974350954024e-05, |
|
"loss": 0.946, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.39644218551461247, |
|
"grad_norm": 0.6511125540817548, |
|
"learning_rate": 1.9698870765370138e-05, |
|
"loss": 1.0868, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.40152477763659467, |
|
"grad_norm": 0.65753027445619, |
|
"learning_rate": 1.9691726958162946e-05, |
|
"loss": 1.0311, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.40660736975857686, |
|
"grad_norm": 0.5461817721258532, |
|
"learning_rate": 1.968454258675079e-05, |
|
"loss": 0.9735, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.41168996188055906, |
|
"grad_norm": 4.0148449609783015, |
|
"learning_rate": 1.967731730465043e-05, |
|
"loss": 1.087, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.4167725540025413, |
|
"grad_norm": 0.5687951661793367, |
|
"learning_rate": 1.967005076142132e-05, |
|
"loss": 1.0163, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.4218551461245235, |
|
"grad_norm": 0.5981525816475669, |
|
"learning_rate": 1.9662742602608975e-05, |
|
"loss": 0.953, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.4269377382465057, |
|
"grad_norm": 0.6527915836371785, |
|
"learning_rate": 1.96553924696873e-05, |
|
"loss": 1.0516, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.4320203303684879, |
|
"grad_norm": 0.5963214365816606, |
|
"learning_rate": 1.9648000000000002e-05, |
|
"loss": 1.1112, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.4371029224904701, |
|
"grad_norm": 0.8494180495407153, |
|
"learning_rate": 1.96405648267009e-05, |
|
"loss": 1.0262, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.44218551461245237, |
|
"grad_norm": 0.5619218395689203, |
|
"learning_rate": 1.9633086578693274e-05, |
|
"loss": 0.9716, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.44726810673443457, |
|
"grad_norm": 0.5453585209280232, |
|
"learning_rate": 1.9625564880568112e-05, |
|
"loss": 1.0342, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.45235069885641677, |
|
"grad_norm": 0.8362630225169261, |
|
"learning_rate": 1.9617999352541277e-05, |
|
"loss": 0.9356, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.45743329097839897, |
|
"grad_norm": 0.5377879009778299, |
|
"learning_rate": 1.9610389610389612e-05, |
|
"loss": 0.9989, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.4625158831003812, |
|
"grad_norm": 0.6137834508091341, |
|
"learning_rate": 1.9602735265385868e-05, |
|
"loss": 0.9171, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.4675984752223634, |
|
"grad_norm": 1.769683080350329, |
|
"learning_rate": 1.959503592423253e-05, |
|
"loss": 1.0589, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.4726810673443456, |
|
"grad_norm": 0.6001041891666766, |
|
"learning_rate": 1.9587291188994433e-05, |
|
"loss": 1.0287, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.4777636594663278, |
|
"grad_norm": 0.6289491210658303, |
|
"learning_rate": 1.9579500657030224e-05, |
|
"loss": 1.0286, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.48284625158831, |
|
"grad_norm": 0.7114332846322945, |
|
"learning_rate": 1.957166392092257e-05, |
|
"loss": 0.9367, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.48792884371029227, |
|
"grad_norm": 0.5391940496295085, |
|
"learning_rate": 1.956378056840714e-05, |
|
"loss": 0.9247, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.49301143583227447, |
|
"grad_norm": 0.589674085065515, |
|
"learning_rate": 1.9555850182300298e-05, |
|
"loss": 0.9389, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.49809402795425667, |
|
"grad_norm": 2.5768068703243765, |
|
"learning_rate": 1.9547872340425533e-05, |
|
"loss": 1.0362, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.49809402795425667, |
|
"eval_loss": NaN, |
|
"eval_runtime": 454.7897, |
|
"eval_samples_per_second": 3.063, |
|
"eval_steps_per_second": 0.385, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.5031766200762389, |
|
"grad_norm": 0.5898947886720844, |
|
"learning_rate": 1.9539846615538515e-05, |
|
"loss": 1.0852, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.5082592121982211, |
|
"grad_norm": 1.170247791808104, |
|
"learning_rate": 1.9531772575250837e-05, |
|
"loss": 0.9584, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.5133418043202033, |
|
"grad_norm": 0.5296707610138294, |
|
"learning_rate": 1.9523649781952368e-05, |
|
"loss": 0.9995, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.5184243964421855, |
|
"grad_norm": 1.005560288522664, |
|
"learning_rate": 1.9515477792732166e-05, |
|
"loss": 0.9777, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.5235069885641678, |
|
"grad_norm": 1.0428938939017154, |
|
"learning_rate": 1.9507256159298012e-05, |
|
"loss": 1.0735, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.5285895806861499, |
|
"grad_norm": 0.7851181913743503, |
|
"learning_rate": 1.949898442789438e-05, |
|
"loss": 0.991, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.5336721728081322, |
|
"grad_norm": 1.2550922763496395, |
|
"learning_rate": 1.9490662139219015e-05, |
|
"loss": 0.983, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.5387547649301143, |
|
"grad_norm": 0.622273113232527, |
|
"learning_rate": 1.9482288828337875e-05, |
|
"loss": 0.9351, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.5438373570520966, |
|
"grad_norm": 1.3022101837507225, |
|
"learning_rate": 1.9473864024598567e-05, |
|
"loss": 0.9618, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.5489199491740788, |
|
"grad_norm": 0.5578554807566163, |
|
"learning_rate": 1.946538725154215e-05, |
|
"loss": 1.0337, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.554002541296061, |
|
"grad_norm": 0.7324460383858523, |
|
"learning_rate": 1.9456858026813338e-05, |
|
"loss": 1.0602, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.5590851334180432, |
|
"grad_norm": 0.5524354794155449, |
|
"learning_rate": 1.9448275862068968e-05, |
|
"loss": 0.9662, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.5641677255400254, |
|
"grad_norm": 0.668823878032025, |
|
"learning_rate": 1.9439640262884816e-05, |
|
"loss": 1.0083, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.5692503176620076, |
|
"grad_norm": 0.6100277575552286, |
|
"learning_rate": 1.9430950728660656e-05, |
|
"loss": 1.0001, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.5743329097839899, |
|
"grad_norm": 0.5908486995002686, |
|
"learning_rate": 1.9422206752523496e-05, |
|
"loss": 1.094, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.579415501905972, |
|
"grad_norm": 0.5679655400443153, |
|
"learning_rate": 1.9413407821229053e-05, |
|
"loss": 0.9961, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.5844980940279543, |
|
"grad_norm": 0.6224828274554025, |
|
"learning_rate": 1.9404553415061298e-05, |
|
"loss": 1.0172, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.5895806861499364, |
|
"grad_norm": 0.5913419268978121, |
|
"learning_rate": 1.939564300773015e-05, |
|
"loss": 1.0175, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.5946632782719187, |
|
"grad_norm": 0.6162736480008056, |
|
"learning_rate": 1.9386676066267185e-05, |
|
"loss": 1.0338, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.5997458703939009, |
|
"grad_norm": 0.6128921125188392, |
|
"learning_rate": 1.937765205091938e-05, |
|
"loss": 1.0118, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.6048284625158831, |
|
"grad_norm": 0.5935201407467202, |
|
"learning_rate": 1.9368570415040795e-05, |
|
"loss": 1.0772, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.6099110546378653, |
|
"grad_norm": 2.2955297203218574, |
|
"learning_rate": 1.935943060498221e-05, |
|
"loss": 1.0561, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.6149936467598475, |
|
"grad_norm": 0.6021547556099237, |
|
"learning_rate": 1.9350232059978583e-05, |
|
"loss": 1.0433, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.6200762388818297, |
|
"grad_norm": 15.458265619359883, |
|
"learning_rate": 1.9340974212034385e-05, |
|
"loss": 0.9981, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.625158831003812, |
|
"grad_norm": 1.2140568290030576, |
|
"learning_rate": 1.9331656485806687e-05, |
|
"loss": 1.0703, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.6302414231257941, |
|
"grad_norm": 1.072369391194644, |
|
"learning_rate": 1.932227829848594e-05, |
|
"loss": 1.0085, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.6353240152477764, |
|
"grad_norm": 0.5560086533767725, |
|
"learning_rate": 1.9312839059674507e-05, |
|
"loss": 0.948, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.6404066073697586, |
|
"grad_norm": 0.6835673411297744, |
|
"learning_rate": 1.93033381712627e-05, |
|
"loss": 1.0326, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.6454891994917408, |
|
"grad_norm": 0.5632555836113124, |
|
"learning_rate": 1.9293775027302514e-05, |
|
"loss": 1.0066, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.650571791613723, |
|
"grad_norm": 0.756797864885407, |
|
"learning_rate": 1.9284149013878745e-05, |
|
"loss": 1.0946, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.6556543837357052, |
|
"grad_norm": 0.5284862562943021, |
|
"learning_rate": 1.927445950897765e-05, |
|
"loss": 1.006, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.6607369758576874, |
|
"grad_norm": 0.6743386059861107, |
|
"learning_rate": 1.926470588235294e-05, |
|
"loss": 1.0845, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.6658195679796697, |
|
"grad_norm": 0.600326744146643, |
|
"learning_rate": 1.9254887495389156e-05, |
|
"loss": 1.0703, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.6709021601016518, |
|
"grad_norm": 0.6781151858674082, |
|
"learning_rate": 1.9245003700962252e-05, |
|
"loss": 1.039, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.6759847522236341, |
|
"grad_norm": 0.6330220485375354, |
|
"learning_rate": 1.923505384329744e-05, |
|
"loss": 1.0711, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.6810673443456162, |
|
"grad_norm": 0.7355514488741505, |
|
"learning_rate": 1.9225037257824145e-05, |
|
"loss": 1.0303, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.6861499364675985, |
|
"grad_norm": 1.2520345572978264, |
|
"learning_rate": 1.921495327102804e-05, |
|
"loss": 1.1175, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.6912325285895807, |
|
"grad_norm": 0.6090687761757717, |
|
"learning_rate": 1.920480120030008e-05, |
|
"loss": 0.9646, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.6963151207115629, |
|
"grad_norm": 0.5292705285452848, |
|
"learning_rate": 1.9194580353782463e-05, |
|
"loss": 0.9466, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.7013977128335451, |
|
"grad_norm": 0.631666235944858, |
|
"learning_rate": 1.9184290030211485e-05, |
|
"loss": 1.0214, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.7064803049555273, |
|
"grad_norm": 0.5330412095712574, |
|
"learning_rate": 1.9173929518757108e-05, |
|
"loss": 1.0399, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.7115628970775095, |
|
"grad_norm": 0.5857060018855021, |
|
"learning_rate": 1.9163498098859318e-05, |
|
"loss": 0.8816, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.7166454891994918, |
|
"grad_norm": 0.5525598551714797, |
|
"learning_rate": 1.9152995040061047e-05, |
|
"loss": 1.059, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.7217280813214739, |
|
"grad_norm": 0.5680224766891685, |
|
"learning_rate": 1.9142419601837675e-05, |
|
"loss": 1.0108, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.7268106734434562, |
|
"grad_norm": 0.8442456969016195, |
|
"learning_rate": 1.9131771033422974e-05, |
|
"loss": 1.1537, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.7318932655654383, |
|
"grad_norm": 0.6777941101499753, |
|
"learning_rate": 1.912104857363146e-05, |
|
"loss": 0.9485, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.7369758576874206, |
|
"grad_norm": 0.6047611797963537, |
|
"learning_rate": 1.9110251450676987e-05, |
|
"loss": 1.1167, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.7420584498094028, |
|
"grad_norm": 0.7302224761524364, |
|
"learning_rate": 1.9099378881987576e-05, |
|
"loss": 1.0066, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.747141041931385, |
|
"grad_norm": 0.679593185673484, |
|
"learning_rate": 1.9088430074016365e-05, |
|
"loss": 1.0814, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.747141041931385, |
|
"eval_loss": NaN, |
|
"eval_runtime": 454.9184, |
|
"eval_samples_per_second": 3.062, |
|
"eval_steps_per_second": 0.385, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.7522236340533672, |
|
"grad_norm": 0.5396369338052232, |
|
"learning_rate": 1.9077404222048476e-05, |
|
"loss": 1.0294, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.7573062261753494, |
|
"grad_norm": 0.569056375341392, |
|
"learning_rate": 1.9066300510003925e-05, |
|
"loss": 1.0177, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.7623888182973316, |
|
"grad_norm": 0.6924084063596142, |
|
"learning_rate": 1.905511811023622e-05, |
|
"loss": 0.9484, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.7674714104193139, |
|
"grad_norm": 0.776918131135741, |
|
"learning_rate": 1.904385618332675e-05, |
|
"loss": 1.0091, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.772554002541296, |
|
"grad_norm": 0.5631403333949265, |
|
"learning_rate": 1.9032513877874704e-05, |
|
"loss": 1.0455, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.7776365946632783, |
|
"grad_norm": 0.5234633902025975, |
|
"learning_rate": 1.9021090330282533e-05, |
|
"loss": 0.8902, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.7827191867852605, |
|
"grad_norm": 0.8954844619126904, |
|
"learning_rate": 1.9009584664536742e-05, |
|
"loss": 1.1005, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.7878017789072427, |
|
"grad_norm": 0.9593051257588633, |
|
"learning_rate": 1.899799599198397e-05, |
|
"loss": 1.0136, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.7928843710292249, |
|
"grad_norm": 0.7462701977681836, |
|
"learning_rate": 1.8986323411102173e-05, |
|
"loss": 0.9644, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.7979669631512071, |
|
"grad_norm": 0.5185601200545743, |
|
"learning_rate": 1.8974566007266858e-05, |
|
"loss": 1.0351, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.8030495552731893, |
|
"grad_norm": 0.5198417937751084, |
|
"learning_rate": 1.896272285251216e-05, |
|
"loss": 1.0483, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.8081321473951716, |
|
"grad_norm": 0.6547184013524777, |
|
"learning_rate": 1.8950793005286704e-05, |
|
"loss": 1.064, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.8132147395171537, |
|
"grad_norm": 0.6353570340115701, |
|
"learning_rate": 1.8938775510204083e-05, |
|
"loss": 0.9942, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.818297331639136, |
|
"grad_norm": 0.6593647760506935, |
|
"learning_rate": 1.8926669397787795e-05, |
|
"loss": 0.9364, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.8233799237611181, |
|
"grad_norm": 0.7127754736535036, |
|
"learning_rate": 1.8914473684210527e-05, |
|
"loss": 1.0383, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.8284625158831004, |
|
"grad_norm": 0.5387503654944155, |
|
"learning_rate": 1.8902187371027652e-05, |
|
"loss": 1.013, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.8335451080050826, |
|
"grad_norm": 0.5486326058897425, |
|
"learning_rate": 1.8889809444904724e-05, |
|
"loss": 1.0064, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.8386277001270648, |
|
"grad_norm": 0.5469444493707462, |
|
"learning_rate": 1.887733887733888e-05, |
|
"loss": 0.9686, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.843710292249047, |
|
"grad_norm": 1.2362111579657413, |
|
"learning_rate": 1.8864774624373957e-05, |
|
"loss": 0.9969, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.8487928843710292, |
|
"grad_norm": 0.6714649440871081, |
|
"learning_rate": 1.8852115626309178e-05, |
|
"loss": 0.9914, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.8538754764930114, |
|
"grad_norm": 0.5790818230708058, |
|
"learning_rate": 1.883936080740118e-05, |
|
"loss": 0.9756, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.8589580686149937, |
|
"grad_norm": 0.7837618766260191, |
|
"learning_rate": 1.882650907555931e-05, |
|
"loss": 1.0127, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.8640406607369758, |
|
"grad_norm": 0.7086746606620099, |
|
"learning_rate": 1.88135593220339e-05, |
|
"loss": 1.1598, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.8691232528589581, |
|
"grad_norm": 0.6038564000981422, |
|
"learning_rate": 1.880051042109741e-05, |
|
"loss": 1.0328, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.8742058449809402, |
|
"grad_norm": 0.5480641081333995, |
|
"learning_rate": 1.878736122971819e-05, |
|
"loss": 1.0436, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.8792884371029225, |
|
"grad_norm": 0.5760210427645431, |
|
"learning_rate": 1.8774110587226747e-05, |
|
"loss": 0.9944, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.8843710292249047, |
|
"grad_norm": 0.5973246266208786, |
|
"learning_rate": 1.8760757314974183e-05, |
|
"loss": 0.9776, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.8894536213468869, |
|
"grad_norm": 0.5919506841693293, |
|
"learning_rate": 1.8747300215982725e-05, |
|
"loss": 0.9878, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.8945362134688691, |
|
"grad_norm": 1.0854485554789506, |
|
"learning_rate": 1.873373807458803e-05, |
|
"loss": 1.106, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.8996188055908514, |
|
"grad_norm": 0.5983944451554725, |
|
"learning_rate": 1.872006965607314e-05, |
|
"loss": 1.0879, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.9047013977128335, |
|
"grad_norm": 0.6634154303973875, |
|
"learning_rate": 1.870629370629371e-05, |
|
"loss": 0.9286, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.9097839898348158, |
|
"grad_norm": 0.5854692268990905, |
|
"learning_rate": 1.869240895129443e-05, |
|
"loss": 0.9738, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.9148665819567979, |
|
"grad_norm": 0.6129124494505737, |
|
"learning_rate": 1.8678414096916303e-05, |
|
"loss": 1.1037, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.9199491740787802, |
|
"grad_norm": 1.0109462911647256, |
|
"learning_rate": 1.8664307828394515e-05, |
|
"loss": 0.9844, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.9250317662007624, |
|
"grad_norm": 0.5765605814856808, |
|
"learning_rate": 1.8650088809946714e-05, |
|
"loss": 1.0584, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.9301143583227446, |
|
"grad_norm": 0.5399969782623486, |
|
"learning_rate": 1.8635755684351314e-05, |
|
"loss": 1.0444, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.9351969504447268, |
|
"grad_norm": 0.6015410734318279, |
|
"learning_rate": 1.862130707251567e-05, |
|
"loss": 1.0178, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.940279542566709, |
|
"grad_norm": 0.5028759554734284, |
|
"learning_rate": 1.8606741573033708e-05, |
|
"loss": 0.9699, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.9453621346886912, |
|
"grad_norm": 0.5458432185038177, |
|
"learning_rate": 1.8592057761732854e-05, |
|
"loss": 0.9319, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.9504447268106735, |
|
"grad_norm": 0.5646703109862106, |
|
"learning_rate": 1.8577254191209787e-05, |
|
"loss": 0.9879, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.9555273189326556, |
|
"grad_norm": 0.5213455747955001, |
|
"learning_rate": 1.856232939035487e-05, |
|
"loss": 0.93, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.9606099110546379, |
|
"grad_norm": 0.6160630051745865, |
|
"learning_rate": 1.854728186386478e-05, |
|
"loss": 0.9984, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.96569250317662, |
|
"grad_norm": 0.5717108146075167, |
|
"learning_rate": 1.853211009174312e-05, |
|
"loss": 0.9986, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.9707750952986023, |
|
"grad_norm": 1.1933909225187864, |
|
"learning_rate": 1.8516812528788576e-05, |
|
"loss": 0.9603, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.9758576874205845, |
|
"grad_norm": 0.5476464515272882, |
|
"learning_rate": 1.8501387604070308e-05, |
|
"loss": 1.0174, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.9809402795425667, |
|
"grad_norm": 0.6026655085443507, |
|
"learning_rate": 1.8485833720390153e-05, |
|
"loss": 0.9648, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.9860228716645489, |
|
"grad_norm": 0.5597982416459779, |
|
"learning_rate": 1.8470149253731344e-05, |
|
"loss": 1.064, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.9911054637865311, |
|
"grad_norm": 0.5256542216530053, |
|
"learning_rate": 1.845433255269321e-05, |
|
"loss": 0.9485, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.9961880559085133, |
|
"grad_norm": 0.5698011465117779, |
|
"learning_rate": 1.843838193791157e-05, |
|
"loss": 1.0465, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.9961880559085133, |
|
"eval_loss": NaN, |
|
"eval_runtime": 454.5687, |
|
"eval_samples_per_second": 3.064, |
|
"eval_steps_per_second": 0.385, |
|
"step": 196 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 392, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 2, |
|
"save_steps": 196, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 7.09553504082985e+17, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|