|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.6147570908816236, |
|
"eval_steps": 500, |
|
"global_step": 24416, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0007805320206966878, |
|
"grad_norm": 24.262310028076172, |
|
"learning_rate": 1.0157273918741808e-06, |
|
"loss": 9.1957, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.0015610640413933755, |
|
"grad_norm": 29.819591522216797, |
|
"learning_rate": 2.0314547837483616e-06, |
|
"loss": 7.9757, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.002341596062090063, |
|
"grad_norm": 14.961602210998535, |
|
"learning_rate": 3.0471821756225426e-06, |
|
"loss": 6.5436, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.003122128082786751, |
|
"grad_norm": 5.151533126831055, |
|
"learning_rate": 4.062909567496723e-06, |
|
"loss": 5.4812, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.003902660103483439, |
|
"grad_norm": 7.370818138122559, |
|
"learning_rate": 5.078636959370905e-06, |
|
"loss": 4.8332, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.004683192124180126, |
|
"grad_norm": 7.932437896728516, |
|
"learning_rate": 6.094364351245085e-06, |
|
"loss": 4.4411, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.005463724144876814, |
|
"grad_norm": 7.834314823150635, |
|
"learning_rate": 7.110091743119267e-06, |
|
"loss": 4.1978, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.006244256165573502, |
|
"grad_norm": 11.881230354309082, |
|
"learning_rate": 8.125819134993446e-06, |
|
"loss": 3.9897, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.00702478818627019, |
|
"grad_norm": 6.348697185516357, |
|
"learning_rate": 9.141546526867629e-06, |
|
"loss": 3.8375, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.007805320206966878, |
|
"grad_norm": 8.414978981018066, |
|
"learning_rate": 1.015727391874181e-05, |
|
"loss": 3.7681, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.008585852227663565, |
|
"grad_norm": 6.531826972961426, |
|
"learning_rate": 1.117300131061599e-05, |
|
"loss": 3.6599, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 0.009366384248360253, |
|
"grad_norm": 9.794255256652832, |
|
"learning_rate": 1.218872870249017e-05, |
|
"loss": 3.5324, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 0.01014691626905694, |
|
"grad_norm": 8.084766387939453, |
|
"learning_rate": 1.3204456094364351e-05, |
|
"loss": 3.4696, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 0.010927448289753628, |
|
"grad_norm": 7.659717559814453, |
|
"learning_rate": 1.4220183486238533e-05, |
|
"loss": 3.3832, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 0.011707980310450316, |
|
"grad_norm": 5.131711959838867, |
|
"learning_rate": 1.5235910878112714e-05, |
|
"loss": 3.3514, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.012488512331147004, |
|
"grad_norm": 6.3673601150512695, |
|
"learning_rate": 1.6251638269986893e-05, |
|
"loss": 3.2988, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 0.013269044351843692, |
|
"grad_norm": 4.5014142990112305, |
|
"learning_rate": 1.7267365661861077e-05, |
|
"loss": 3.239, |
|
"step": 527 |
|
}, |
|
{ |
|
"epoch": 0.01404957637254038, |
|
"grad_norm": 5.1420769691467285, |
|
"learning_rate": 1.8283093053735257e-05, |
|
"loss": 3.1892, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 0.014830108393237068, |
|
"grad_norm": 5.227200508117676, |
|
"learning_rate": 1.9298820445609438e-05, |
|
"loss": 3.1732, |
|
"step": 589 |
|
}, |
|
{ |
|
"epoch": 0.015610640413933756, |
|
"grad_norm": 4.4621734619140625, |
|
"learning_rate": 2.031454783748362e-05, |
|
"loss": 3.0898, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.016391172434630442, |
|
"grad_norm": 5.213263511657715, |
|
"learning_rate": 2.13302752293578e-05, |
|
"loss": 3.0832, |
|
"step": 651 |
|
}, |
|
{ |
|
"epoch": 0.01717170445532713, |
|
"grad_norm": 4.23059606552124, |
|
"learning_rate": 2.234600262123198e-05, |
|
"loss": 3.0341, |
|
"step": 682 |
|
}, |
|
{ |
|
"epoch": 0.017952236476023817, |
|
"grad_norm": 3.5276854038238525, |
|
"learning_rate": 2.336173001310616e-05, |
|
"loss": 3.0241, |
|
"step": 713 |
|
}, |
|
{ |
|
"epoch": 0.018732768496720505, |
|
"grad_norm": 3.199700355529785, |
|
"learning_rate": 2.437745740498034e-05, |
|
"loss": 2.9659, |
|
"step": 744 |
|
}, |
|
{ |
|
"epoch": 0.019513300517417193, |
|
"grad_norm": 9.667389869689941, |
|
"learning_rate": 2.5393184796854525e-05, |
|
"loss": 2.9292, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.02029383253811388, |
|
"grad_norm": 3.677541732788086, |
|
"learning_rate": 2.6408912188728702e-05, |
|
"loss": 2.919, |
|
"step": 806 |
|
}, |
|
{ |
|
"epoch": 0.02107436455881057, |
|
"grad_norm": 3.711683750152588, |
|
"learning_rate": 2.7424639580602886e-05, |
|
"loss": 2.9141, |
|
"step": 837 |
|
}, |
|
{ |
|
"epoch": 0.021854896579507257, |
|
"grad_norm": 3.1250174045562744, |
|
"learning_rate": 2.8440366972477066e-05, |
|
"loss": 2.8585, |
|
"step": 868 |
|
}, |
|
{ |
|
"epoch": 0.022635428600203945, |
|
"grad_norm": 3.4183220863342285, |
|
"learning_rate": 2.9456094364351244e-05, |
|
"loss": 2.8206, |
|
"step": 899 |
|
}, |
|
{ |
|
"epoch": 0.023415960620900633, |
|
"grad_norm": 2.8176679611206055, |
|
"learning_rate": 3.0471821756225428e-05, |
|
"loss": 2.787, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.02419649264159732, |
|
"grad_norm": 2.7993154525756836, |
|
"learning_rate": 3.148754914809961e-05, |
|
"loss": 2.7678, |
|
"step": 961 |
|
}, |
|
{ |
|
"epoch": 0.02497702466229401, |
|
"grad_norm": 2.8017027378082275, |
|
"learning_rate": 3.2503276539973785e-05, |
|
"loss": 2.7499, |
|
"step": 992 |
|
}, |
|
{ |
|
"epoch": 0.025757556682990696, |
|
"grad_norm": 2.958606004714966, |
|
"learning_rate": 3.351900393184797e-05, |
|
"loss": 2.7211, |
|
"step": 1023 |
|
}, |
|
{ |
|
"epoch": 0.026538088703687384, |
|
"grad_norm": 2.5942046642303467, |
|
"learning_rate": 3.453473132372215e-05, |
|
"loss": 2.6823, |
|
"step": 1054 |
|
}, |
|
{ |
|
"epoch": 0.027318620724384072, |
|
"grad_norm": 3.1978206634521484, |
|
"learning_rate": 3.555045871559633e-05, |
|
"loss": 2.6943, |
|
"step": 1085 |
|
}, |
|
{ |
|
"epoch": 0.02809915274508076, |
|
"grad_norm": 2.6260809898376465, |
|
"learning_rate": 3.6566186107470514e-05, |
|
"loss": 2.6574, |
|
"step": 1116 |
|
}, |
|
{ |
|
"epoch": 0.028879684765777448, |
|
"grad_norm": 2.6633715629577637, |
|
"learning_rate": 3.7581913499344695e-05, |
|
"loss": 2.6394, |
|
"step": 1147 |
|
}, |
|
{ |
|
"epoch": 0.029660216786474135, |
|
"grad_norm": 2.5701236724853516, |
|
"learning_rate": 3.8597640891218876e-05, |
|
"loss": 2.6336, |
|
"step": 1178 |
|
}, |
|
{ |
|
"epoch": 0.030440748807170823, |
|
"grad_norm": 2.4242734909057617, |
|
"learning_rate": 3.9613368283093056e-05, |
|
"loss": 2.6285, |
|
"step": 1209 |
|
}, |
|
{ |
|
"epoch": 0.03122128082786751, |
|
"grad_norm": 2.6410114765167236, |
|
"learning_rate": 4.062909567496724e-05, |
|
"loss": 2.5821, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.032001812848564196, |
|
"grad_norm": 2.419649839401245, |
|
"learning_rate": 4.164482306684142e-05, |
|
"loss": 2.5576, |
|
"step": 1271 |
|
}, |
|
{ |
|
"epoch": 0.032782344869260883, |
|
"grad_norm": 2.314750909805298, |
|
"learning_rate": 4.26605504587156e-05, |
|
"loss": 2.5438, |
|
"step": 1302 |
|
}, |
|
{ |
|
"epoch": 0.03356287688995757, |
|
"grad_norm": 2.288361072540283, |
|
"learning_rate": 4.367627785058978e-05, |
|
"loss": 2.5197, |
|
"step": 1333 |
|
}, |
|
{ |
|
"epoch": 0.03434340891065426, |
|
"grad_norm": 2.328840494155884, |
|
"learning_rate": 4.469200524246396e-05, |
|
"loss": 2.5193, |
|
"step": 1364 |
|
}, |
|
{ |
|
"epoch": 0.03512394093135095, |
|
"grad_norm": 2.2555599212646484, |
|
"learning_rate": 4.570773263433814e-05, |
|
"loss": 2.4907, |
|
"step": 1395 |
|
}, |
|
{ |
|
"epoch": 0.035904472952047635, |
|
"grad_norm": 2.1358766555786133, |
|
"learning_rate": 4.672346002621232e-05, |
|
"loss": 2.4737, |
|
"step": 1426 |
|
}, |
|
{ |
|
"epoch": 0.03668500497274432, |
|
"grad_norm": 2.1471140384674072, |
|
"learning_rate": 4.77391874180865e-05, |
|
"loss": 2.4733, |
|
"step": 1457 |
|
}, |
|
{ |
|
"epoch": 0.03746553699344101, |
|
"grad_norm": 2.1371941566467285, |
|
"learning_rate": 4.875491480996068e-05, |
|
"loss": 2.432, |
|
"step": 1488 |
|
}, |
|
{ |
|
"epoch": 0.0382460690141377, |
|
"grad_norm": 1.9789690971374512, |
|
"learning_rate": 4.977064220183487e-05, |
|
"loss": 2.4407, |
|
"step": 1519 |
|
}, |
|
{ |
|
"epoch": 0.039026601034834386, |
|
"grad_norm": 2.0055742263793945, |
|
"learning_rate": 4.9999915451558777e-05, |
|
"loss": 2.4178, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.039807133055531074, |
|
"grad_norm": 2.095475196838379, |
|
"learning_rate": 4.999955597496219e-05, |
|
"loss": 2.4198, |
|
"step": 1581 |
|
}, |
|
{ |
|
"epoch": 0.04058766507622776, |
|
"grad_norm": 2.046337842941284, |
|
"learning_rate": 4.9998914381774255e-05, |
|
"loss": 2.3936, |
|
"step": 1612 |
|
}, |
|
{ |
|
"epoch": 0.04136819709692445, |
|
"grad_norm": 1.8528848886489868, |
|
"learning_rate": 4.999799067923527e-05, |
|
"loss": 2.3893, |
|
"step": 1643 |
|
}, |
|
{ |
|
"epoch": 0.04214872911762114, |
|
"grad_norm": 1.883175253868103, |
|
"learning_rate": 4.999678487776908e-05, |
|
"loss": 2.3668, |
|
"step": 1674 |
|
}, |
|
{ |
|
"epoch": 0.042929261138317826, |
|
"grad_norm": 1.8084770441055298, |
|
"learning_rate": 4.9995296990983006e-05, |
|
"loss": 2.3702, |
|
"step": 1705 |
|
}, |
|
{ |
|
"epoch": 0.043709793159014514, |
|
"grad_norm": 1.8336963653564453, |
|
"learning_rate": 4.999352703566763e-05, |
|
"loss": 2.357, |
|
"step": 1736 |
|
}, |
|
{ |
|
"epoch": 0.0444903251797112, |
|
"grad_norm": 1.8369375467300415, |
|
"learning_rate": 4.999147503179668e-05, |
|
"loss": 2.3393, |
|
"step": 1767 |
|
}, |
|
{ |
|
"epoch": 0.04527085720040789, |
|
"grad_norm": 1.8411383628845215, |
|
"learning_rate": 4.998914100252672e-05, |
|
"loss": 2.319, |
|
"step": 1798 |
|
}, |
|
{ |
|
"epoch": 0.04605138922110458, |
|
"grad_norm": 1.785543441772461, |
|
"learning_rate": 4.998652497419696e-05, |
|
"loss": 2.3259, |
|
"step": 1829 |
|
}, |
|
{ |
|
"epoch": 0.046831921241801265, |
|
"grad_norm": 1.7194079160690308, |
|
"learning_rate": 4.9983626976328927e-05, |
|
"loss": 2.3046, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 0.04761245326249795, |
|
"grad_norm": 1.7909469604492188, |
|
"learning_rate": 4.998044704162613e-05, |
|
"loss": 2.307, |
|
"step": 1891 |
|
}, |
|
{ |
|
"epoch": 0.04839298528319464, |
|
"grad_norm": 1.725806474685669, |
|
"learning_rate": 4.9976985205973705e-05, |
|
"loss": 2.2992, |
|
"step": 1922 |
|
}, |
|
{ |
|
"epoch": 0.04917351730389133, |
|
"grad_norm": 1.67300546169281, |
|
"learning_rate": 4.997324150843799e-05, |
|
"loss": 2.2763, |
|
"step": 1953 |
|
}, |
|
{ |
|
"epoch": 0.04995404932458802, |
|
"grad_norm": 1.7416061162948608, |
|
"learning_rate": 4.99692159912661e-05, |
|
"loss": 2.2596, |
|
"step": 1984 |
|
}, |
|
{ |
|
"epoch": 0.050734581345284704, |
|
"grad_norm": 1.843024492263794, |
|
"learning_rate": 4.996490869988546e-05, |
|
"loss": 2.2644, |
|
"step": 2015 |
|
}, |
|
{ |
|
"epoch": 0.05151511336598139, |
|
"grad_norm": 1.6172534227371216, |
|
"learning_rate": 4.996031968290326e-05, |
|
"loss": 2.271, |
|
"step": 2046 |
|
}, |
|
{ |
|
"epoch": 0.05229564538667808, |
|
"grad_norm": 1.6510816812515259, |
|
"learning_rate": 4.995544899210594e-05, |
|
"loss": 2.2445, |
|
"step": 2077 |
|
}, |
|
{ |
|
"epoch": 0.05307617740737477, |
|
"grad_norm": 1.625775933265686, |
|
"learning_rate": 4.9950296682458583e-05, |
|
"loss": 2.2432, |
|
"step": 2108 |
|
}, |
|
{ |
|
"epoch": 0.053856709428071456, |
|
"grad_norm": 1.602299690246582, |
|
"learning_rate": 4.994486281210429e-05, |
|
"loss": 2.2266, |
|
"step": 2139 |
|
}, |
|
{ |
|
"epoch": 0.054637241448768144, |
|
"grad_norm": 1.7036757469177246, |
|
"learning_rate": 4.9939147442363566e-05, |
|
"loss": 2.2292, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 0.05541777346946483, |
|
"grad_norm": 1.5851162672042847, |
|
"learning_rate": 4.9933150637733574e-05, |
|
"loss": 2.2286, |
|
"step": 2201 |
|
}, |
|
{ |
|
"epoch": 0.05619830549016152, |
|
"grad_norm": 1.6328140497207642, |
|
"learning_rate": 4.992687246588743e-05, |
|
"loss": 2.2342, |
|
"step": 2232 |
|
}, |
|
{ |
|
"epoch": 0.05697883751085821, |
|
"grad_norm": 1.4901413917541504, |
|
"learning_rate": 4.992031299767347e-05, |
|
"loss": 2.2094, |
|
"step": 2263 |
|
}, |
|
{ |
|
"epoch": 0.057759369531554895, |
|
"grad_norm": 1.5897445678710938, |
|
"learning_rate": 4.9913472307114386e-05, |
|
"loss": 2.2126, |
|
"step": 2294 |
|
}, |
|
{ |
|
"epoch": 0.05853990155225158, |
|
"grad_norm": 1.5534899234771729, |
|
"learning_rate": 4.9906350471406446e-05, |
|
"loss": 2.2125, |
|
"step": 2325 |
|
}, |
|
{ |
|
"epoch": 0.05932043357294827, |
|
"grad_norm": 1.5713675022125244, |
|
"learning_rate": 4.989894757091861e-05, |
|
"loss": 2.2007, |
|
"step": 2356 |
|
}, |
|
{ |
|
"epoch": 0.06010096559364496, |
|
"grad_norm": 1.473955750465393, |
|
"learning_rate": 4.989126368919158e-05, |
|
"loss": 2.1819, |
|
"step": 2387 |
|
}, |
|
{ |
|
"epoch": 0.06088149761434165, |
|
"grad_norm": 1.5200265645980835, |
|
"learning_rate": 4.988329891293693e-05, |
|
"loss": 2.1944, |
|
"step": 2418 |
|
}, |
|
{ |
|
"epoch": 0.061662029635038335, |
|
"grad_norm": 1.5236589908599854, |
|
"learning_rate": 4.987505333203608e-05, |
|
"loss": 2.1816, |
|
"step": 2449 |
|
}, |
|
{ |
|
"epoch": 0.06244256165573502, |
|
"grad_norm": 1.5306485891342163, |
|
"learning_rate": 4.9866527039539276e-05, |
|
"loss": 2.187, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 0.06322309367643171, |
|
"grad_norm": 1.5017462968826294, |
|
"learning_rate": 4.9857720131664594e-05, |
|
"loss": 2.1706, |
|
"step": 2511 |
|
}, |
|
{ |
|
"epoch": 0.06400362569712839, |
|
"grad_norm": 4.610604763031006, |
|
"learning_rate": 4.9848632707796773e-05, |
|
"loss": 2.1528, |
|
"step": 2542 |
|
}, |
|
{ |
|
"epoch": 0.06478415771782509, |
|
"grad_norm": 1.4580930471420288, |
|
"learning_rate": 4.9839264870486155e-05, |
|
"loss": 2.1849, |
|
"step": 2573 |
|
}, |
|
{ |
|
"epoch": 0.06556468973852177, |
|
"grad_norm": 1.5506151914596558, |
|
"learning_rate": 4.9829616725447526e-05, |
|
"loss": 2.1504, |
|
"step": 2604 |
|
}, |
|
{ |
|
"epoch": 0.06634522175921846, |
|
"grad_norm": 1.4275034666061401, |
|
"learning_rate": 4.981968838155888e-05, |
|
"loss": 2.1492, |
|
"step": 2635 |
|
}, |
|
{ |
|
"epoch": 0.06712575377991514, |
|
"grad_norm": 1.4749577045440674, |
|
"learning_rate": 4.980947995086024e-05, |
|
"loss": 2.1579, |
|
"step": 2666 |
|
}, |
|
{ |
|
"epoch": 0.06790628580061184, |
|
"grad_norm": 1.4220107793807983, |
|
"learning_rate": 4.979899154855234e-05, |
|
"loss": 2.1522, |
|
"step": 2697 |
|
}, |
|
{ |
|
"epoch": 0.06868681782130852, |
|
"grad_norm": 1.4243831634521484, |
|
"learning_rate": 4.9788223292995386e-05, |
|
"loss": 2.1258, |
|
"step": 2728 |
|
}, |
|
{ |
|
"epoch": 0.06946734984200521, |
|
"grad_norm": 1.4925100803375244, |
|
"learning_rate": 4.977717530570768e-05, |
|
"loss": 2.1379, |
|
"step": 2759 |
|
}, |
|
{ |
|
"epoch": 0.0702478818627019, |
|
"grad_norm": 1.4839624166488647, |
|
"learning_rate": 4.976584771136425e-05, |
|
"loss": 2.1588, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 0.07102841388339859, |
|
"grad_norm": 1.4130090475082397, |
|
"learning_rate": 4.975424063779547e-05, |
|
"loss": 2.1461, |
|
"step": 2821 |
|
}, |
|
{ |
|
"epoch": 0.07180894590409527, |
|
"grad_norm": 1.3940472602844238, |
|
"learning_rate": 4.974235421598557e-05, |
|
"loss": 2.1416, |
|
"step": 2852 |
|
}, |
|
{ |
|
"epoch": 0.07258947792479196, |
|
"grad_norm": 1.4502291679382324, |
|
"learning_rate": 4.973018858007122e-05, |
|
"loss": 2.1315, |
|
"step": 2883 |
|
}, |
|
{ |
|
"epoch": 0.07337000994548865, |
|
"grad_norm": 1.4334006309509277, |
|
"learning_rate": 4.9717743867339963e-05, |
|
"loss": 2.122, |
|
"step": 2914 |
|
}, |
|
{ |
|
"epoch": 0.07415054196618534, |
|
"grad_norm": 1.4366185665130615, |
|
"learning_rate": 4.9705020218228695e-05, |
|
"loss": 2.1179, |
|
"step": 2945 |
|
}, |
|
{ |
|
"epoch": 0.07493107398688202, |
|
"grad_norm": 1.437264084815979, |
|
"learning_rate": 4.969201777632205e-05, |
|
"loss": 2.1171, |
|
"step": 2976 |
|
}, |
|
{ |
|
"epoch": 0.07571160600757872, |
|
"grad_norm": 1.3713815212249756, |
|
"learning_rate": 4.9678736688350846e-05, |
|
"loss": 2.1215, |
|
"step": 3007 |
|
}, |
|
{ |
|
"epoch": 0.0764921380282754, |
|
"grad_norm": 1.5930180549621582, |
|
"learning_rate": 4.966517710419033e-05, |
|
"loss": 2.0961, |
|
"step": 3038 |
|
}, |
|
{ |
|
"epoch": 0.07727267004897209, |
|
"grad_norm": 1.4208778142929077, |
|
"learning_rate": 4.965133917685858e-05, |
|
"loss": 2.1193, |
|
"step": 3069 |
|
}, |
|
{ |
|
"epoch": 0.07805320206966877, |
|
"grad_norm": 1.3563286066055298, |
|
"learning_rate": 4.9637223062514714e-05, |
|
"loss": 2.0957, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.07883373409036547, |
|
"grad_norm": 1.382861852645874, |
|
"learning_rate": 4.962282892045718e-05, |
|
"loss": 2.117, |
|
"step": 3131 |
|
}, |
|
{ |
|
"epoch": 0.07961426611106215, |
|
"grad_norm": 1.326231837272644, |
|
"learning_rate": 4.9608156913121904e-05, |
|
"loss": 2.0906, |
|
"step": 3162 |
|
}, |
|
{ |
|
"epoch": 0.08039479813175884, |
|
"grad_norm": 1.3465954065322876, |
|
"learning_rate": 4.959320720608049e-05, |
|
"loss": 2.09, |
|
"step": 3193 |
|
}, |
|
{ |
|
"epoch": 0.08117533015245552, |
|
"grad_norm": 1.3832777738571167, |
|
"learning_rate": 4.9577979968038354e-05, |
|
"loss": 2.089, |
|
"step": 3224 |
|
}, |
|
{ |
|
"epoch": 0.08195586217315222, |
|
"grad_norm": 1.4453243017196655, |
|
"learning_rate": 4.956247537083282e-05, |
|
"loss": 2.0867, |
|
"step": 3255 |
|
}, |
|
{ |
|
"epoch": 0.0827363941938489, |
|
"grad_norm": 1.310168743133545, |
|
"learning_rate": 4.9546693589431145e-05, |
|
"loss": 2.0915, |
|
"step": 3286 |
|
}, |
|
{ |
|
"epoch": 0.0835169262145456, |
|
"grad_norm": 1.354078769683838, |
|
"learning_rate": 4.9530634801928595e-05, |
|
"loss": 2.0888, |
|
"step": 3317 |
|
}, |
|
{ |
|
"epoch": 0.08429745823524228, |
|
"grad_norm": 1.3768595457077026, |
|
"learning_rate": 4.9514299189546395e-05, |
|
"loss": 2.0821, |
|
"step": 3348 |
|
}, |
|
{ |
|
"epoch": 0.08507799025593897, |
|
"grad_norm": 1.3402544260025024, |
|
"learning_rate": 4.949768693662973e-05, |
|
"loss": 2.0776, |
|
"step": 3379 |
|
}, |
|
{ |
|
"epoch": 0.08585852227663565, |
|
"grad_norm": 1.3178027868270874, |
|
"learning_rate": 4.948079823064559e-05, |
|
"loss": 2.0699, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 0.08663905429733235, |
|
"grad_norm": 1.7057585716247559, |
|
"learning_rate": 4.946363326218074e-05, |
|
"loss": 2.0615, |
|
"step": 3441 |
|
}, |
|
{ |
|
"epoch": 0.08741958631802903, |
|
"grad_norm": 1.3453648090362549, |
|
"learning_rate": 4.9446192224939525e-05, |
|
"loss": 2.0788, |
|
"step": 3472 |
|
}, |
|
{ |
|
"epoch": 0.08820011833872572, |
|
"grad_norm": 1.3319638967514038, |
|
"learning_rate": 4.942847531574167e-05, |
|
"loss": 2.0619, |
|
"step": 3503 |
|
}, |
|
{ |
|
"epoch": 0.0889806503594224, |
|
"grad_norm": 1.2845163345336914, |
|
"learning_rate": 4.941048273452008e-05, |
|
"loss": 2.0728, |
|
"step": 3534 |
|
}, |
|
{ |
|
"epoch": 0.0897611823801191, |
|
"grad_norm": 1.3615431785583496, |
|
"learning_rate": 4.9392214684318605e-05, |
|
"loss": 2.0538, |
|
"step": 3565 |
|
}, |
|
{ |
|
"epoch": 0.09054171440081578, |
|
"grad_norm": 1.2946122884750366, |
|
"learning_rate": 4.93736713712897e-05, |
|
"loss": 2.0689, |
|
"step": 3596 |
|
}, |
|
{ |
|
"epoch": 0.09132224642151247, |
|
"grad_norm": 1.294846534729004, |
|
"learning_rate": 4.9354853004692124e-05, |
|
"loss": 2.0706, |
|
"step": 3627 |
|
}, |
|
{ |
|
"epoch": 0.09210277844220915, |
|
"grad_norm": 1.298010230064392, |
|
"learning_rate": 4.93357597968886e-05, |
|
"loss": 2.0406, |
|
"step": 3658 |
|
}, |
|
{ |
|
"epoch": 0.09288331046290585, |
|
"grad_norm": 1.7569706439971924, |
|
"learning_rate": 4.931639196334338e-05, |
|
"loss": 2.0736, |
|
"step": 3689 |
|
}, |
|
{ |
|
"epoch": 0.09366384248360253, |
|
"grad_norm": 1.2595418691635132, |
|
"learning_rate": 4.9296749722619826e-05, |
|
"loss": 2.0654, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 0.09444437450429922, |
|
"grad_norm": 1.3139164447784424, |
|
"learning_rate": 4.9276833296377966e-05, |
|
"loss": 2.0467, |
|
"step": 3751 |
|
}, |
|
{ |
|
"epoch": 0.0952249065249959, |
|
"grad_norm": 1.290247917175293, |
|
"learning_rate": 4.925664290937196e-05, |
|
"loss": 2.0418, |
|
"step": 3782 |
|
}, |
|
{ |
|
"epoch": 0.0960054385456926, |
|
"grad_norm": 1.2755826711654663, |
|
"learning_rate": 4.9236178789447576e-05, |
|
"loss": 2.0244, |
|
"step": 3813 |
|
}, |
|
{ |
|
"epoch": 0.09678597056638928, |
|
"grad_norm": 1.2596077919006348, |
|
"learning_rate": 4.921544116753962e-05, |
|
"loss": 2.0609, |
|
"step": 3844 |
|
}, |
|
{ |
|
"epoch": 0.09756650258708598, |
|
"grad_norm": 1.249387264251709, |
|
"learning_rate": 4.919443027766935e-05, |
|
"loss": 2.0455, |
|
"step": 3875 |
|
}, |
|
{ |
|
"epoch": 0.09834703460778266, |
|
"grad_norm": 1.255379319190979, |
|
"learning_rate": 4.91731463569418e-05, |
|
"loss": 2.024, |
|
"step": 3906 |
|
}, |
|
{ |
|
"epoch": 0.09912756662847935, |
|
"grad_norm": 1.2620748281478882, |
|
"learning_rate": 4.915158964554312e-05, |
|
"loss": 2.0355, |
|
"step": 3937 |
|
}, |
|
{ |
|
"epoch": 0.09990809864917603, |
|
"grad_norm": 1.2736345529556274, |
|
"learning_rate": 4.912976038673786e-05, |
|
"loss": 2.0351, |
|
"step": 3968 |
|
}, |
|
{ |
|
"epoch": 0.10068863066987273, |
|
"grad_norm": 1.18524169921875, |
|
"learning_rate": 4.9107658826866254e-05, |
|
"loss": 2.029, |
|
"step": 3999 |
|
}, |
|
{ |
|
"epoch": 0.10146916269056941, |
|
"grad_norm": 10.429583549499512, |
|
"learning_rate": 4.908528521534139e-05, |
|
"loss": 2.0226, |
|
"step": 4030 |
|
}, |
|
{ |
|
"epoch": 0.1022496947112661, |
|
"grad_norm": 1.240295648574829, |
|
"learning_rate": 4.906263980464644e-05, |
|
"loss": 2.0401, |
|
"step": 4061 |
|
}, |
|
{ |
|
"epoch": 0.10303022673196278, |
|
"grad_norm": 1.2191627025604248, |
|
"learning_rate": 4.903972285033178e-05, |
|
"loss": 2.0079, |
|
"step": 4092 |
|
}, |
|
{ |
|
"epoch": 0.10381075875265948, |
|
"grad_norm": 1.2340337038040161, |
|
"learning_rate": 4.901653461101213e-05, |
|
"loss": 2.0305, |
|
"step": 4123 |
|
}, |
|
{ |
|
"epoch": 0.10459129077335616, |
|
"grad_norm": 1.284839153289795, |
|
"learning_rate": 4.8993075348363626e-05, |
|
"loss": 2.0212, |
|
"step": 4154 |
|
}, |
|
{ |
|
"epoch": 0.10537182279405286, |
|
"grad_norm": 1.2955284118652344, |
|
"learning_rate": 4.896934532712084e-05, |
|
"loss": 2.0328, |
|
"step": 4185 |
|
}, |
|
{ |
|
"epoch": 0.10615235481474954, |
|
"grad_norm": 1.165124535560608, |
|
"learning_rate": 4.8945344815073846e-05, |
|
"loss": 2.0248, |
|
"step": 4216 |
|
}, |
|
{ |
|
"epoch": 0.10693288683544623, |
|
"grad_norm": 1.1925835609436035, |
|
"learning_rate": 4.892107408306516e-05, |
|
"loss": 2.0086, |
|
"step": 4247 |
|
}, |
|
{ |
|
"epoch": 0.10771341885614291, |
|
"grad_norm": 1.1841633319854736, |
|
"learning_rate": 4.889653340498669e-05, |
|
"loss": 2.0166, |
|
"step": 4278 |
|
}, |
|
{ |
|
"epoch": 0.1084939508768396, |
|
"grad_norm": 1.2198415994644165, |
|
"learning_rate": 4.8871723057776664e-05, |
|
"loss": 2.0164, |
|
"step": 4309 |
|
}, |
|
{ |
|
"epoch": 0.10927448289753629, |
|
"grad_norm": 1.2284913063049316, |
|
"learning_rate": 4.8846643321416476e-05, |
|
"loss": 2.015, |
|
"step": 4340 |
|
}, |
|
{ |
|
"epoch": 0.11005501491823298, |
|
"grad_norm": 1.2090349197387695, |
|
"learning_rate": 4.882129447892753e-05, |
|
"loss": 2.0094, |
|
"step": 4371 |
|
}, |
|
{ |
|
"epoch": 0.11083554693892966, |
|
"grad_norm": 1.215232491493225, |
|
"learning_rate": 4.8795676816368076e-05, |
|
"loss": 2.0045, |
|
"step": 4402 |
|
}, |
|
{ |
|
"epoch": 0.11161607895962636, |
|
"grad_norm": 1.2804045677185059, |
|
"learning_rate": 4.876979062282995e-05, |
|
"loss": 1.9899, |
|
"step": 4433 |
|
}, |
|
{ |
|
"epoch": 0.11239661098032304, |
|
"grad_norm": 1.2405016422271729, |
|
"learning_rate": 4.8743636190435325e-05, |
|
"loss": 2.0093, |
|
"step": 4464 |
|
}, |
|
{ |
|
"epoch": 0.11317714300101973, |
|
"grad_norm": 1.1784151792526245, |
|
"learning_rate": 4.871721381433344e-05, |
|
"loss": 2.0036, |
|
"step": 4495 |
|
}, |
|
{ |
|
"epoch": 0.11395767502171641, |
|
"grad_norm": 1.185738205909729, |
|
"learning_rate": 4.869052379269719e-05, |
|
"loss": 2.0033, |
|
"step": 4526 |
|
}, |
|
{ |
|
"epoch": 0.11473820704241311, |
|
"grad_norm": 1.1984614133834839, |
|
"learning_rate": 4.866356642671985e-05, |
|
"loss": 2.025, |
|
"step": 4557 |
|
}, |
|
{ |
|
"epoch": 0.11551873906310979, |
|
"grad_norm": 1.1734029054641724, |
|
"learning_rate": 4.8636342020611634e-05, |
|
"loss": 2.0023, |
|
"step": 4588 |
|
}, |
|
{ |
|
"epoch": 0.11629927108380649, |
|
"grad_norm": 1.198611855506897, |
|
"learning_rate": 4.860885088159626e-05, |
|
"loss": 2.0031, |
|
"step": 4619 |
|
}, |
|
{ |
|
"epoch": 0.11707980310450317, |
|
"grad_norm": 1.1793649196624756, |
|
"learning_rate": 4.858109331990751e-05, |
|
"loss": 1.9806, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 0.11786033512519985, |
|
"grad_norm": 1.1945818662643433, |
|
"learning_rate": 4.855306964878567e-05, |
|
"loss": 1.9855, |
|
"step": 4681 |
|
}, |
|
{ |
|
"epoch": 0.11864086714589654, |
|
"grad_norm": 1.1875253915786743, |
|
"learning_rate": 4.8524780184474084e-05, |
|
"loss": 1.9883, |
|
"step": 4712 |
|
}, |
|
{ |
|
"epoch": 0.11942139916659322, |
|
"grad_norm": 1.1512328386306763, |
|
"learning_rate": 4.8496225246215496e-05, |
|
"loss": 1.9807, |
|
"step": 4743 |
|
}, |
|
{ |
|
"epoch": 0.12020193118728992, |
|
"grad_norm": 1.2059484720230103, |
|
"learning_rate": 4.8467405156248505e-05, |
|
"loss": 2.0103, |
|
"step": 4774 |
|
}, |
|
{ |
|
"epoch": 0.1209824632079866, |
|
"grad_norm": 1.1759779453277588, |
|
"learning_rate": 4.843832023980392e-05, |
|
"loss": 1.9754, |
|
"step": 4805 |
|
}, |
|
{ |
|
"epoch": 0.1217629952286833, |
|
"grad_norm": 1.2004414796829224, |
|
"learning_rate": 4.840897082510106e-05, |
|
"loss": 1.9893, |
|
"step": 4836 |
|
}, |
|
{ |
|
"epoch": 0.12254352724937997, |
|
"grad_norm": 1.1744251251220703, |
|
"learning_rate": 4.8379357243344084e-05, |
|
"loss": 1.9848, |
|
"step": 4867 |
|
}, |
|
{ |
|
"epoch": 0.12332405927007667, |
|
"grad_norm": 1.1731071472167969, |
|
"learning_rate": 4.8349479828718236e-05, |
|
"loss": 1.9945, |
|
"step": 4898 |
|
}, |
|
{ |
|
"epoch": 0.12410459129077335, |
|
"grad_norm": 1.2002074718475342, |
|
"learning_rate": 4.8319338918386075e-05, |
|
"loss": 1.9869, |
|
"step": 4929 |
|
}, |
|
{ |
|
"epoch": 0.12488512331147004, |
|
"grad_norm": 1.147085189819336, |
|
"learning_rate": 4.828893485248369e-05, |
|
"loss": 1.9876, |
|
"step": 4960 |
|
}, |
|
{ |
|
"epoch": 0.12566565533216673, |
|
"grad_norm": 1.1605637073516846, |
|
"learning_rate": 4.825826797411682e-05, |
|
"loss": 1.9858, |
|
"step": 4991 |
|
}, |
|
{ |
|
"epoch": 0.12644618735286342, |
|
"grad_norm": 1.1823869943618774, |
|
"learning_rate": 4.822733862935702e-05, |
|
"loss": 1.9617, |
|
"step": 5022 |
|
}, |
|
{ |
|
"epoch": 0.12722671937356012, |
|
"grad_norm": 1.1143817901611328, |
|
"learning_rate": 4.819614716723775e-05, |
|
"loss": 1.9756, |
|
"step": 5053 |
|
}, |
|
{ |
|
"epoch": 0.12800725139425678, |
|
"grad_norm": 1.1646877527236938, |
|
"learning_rate": 4.8164693939750425e-05, |
|
"loss": 2.0005, |
|
"step": 5084 |
|
}, |
|
{ |
|
"epoch": 0.12878778341495348, |
|
"grad_norm": 1.1390529870986938, |
|
"learning_rate": 4.813297930184042e-05, |
|
"loss": 1.9943, |
|
"step": 5115 |
|
}, |
|
{ |
|
"epoch": 0.12956831543565017, |
|
"grad_norm": 1.1759583950042725, |
|
"learning_rate": 4.810100361140314e-05, |
|
"loss": 1.9772, |
|
"step": 5146 |
|
}, |
|
{ |
|
"epoch": 0.13034884745634687, |
|
"grad_norm": 1.1465727090835571, |
|
"learning_rate": 4.8068767229279885e-05, |
|
"loss": 1.9563, |
|
"step": 5177 |
|
}, |
|
{ |
|
"epoch": 0.13112937947704353, |
|
"grad_norm": 1.2573471069335938, |
|
"learning_rate": 4.8036270519253854e-05, |
|
"loss": 1.9735, |
|
"step": 5208 |
|
}, |
|
{ |
|
"epoch": 0.13190991149774023, |
|
"grad_norm": 1.1301871538162231, |
|
"learning_rate": 4.8003513848046e-05, |
|
"loss": 1.987, |
|
"step": 5239 |
|
}, |
|
{ |
|
"epoch": 0.13269044351843692, |
|
"grad_norm": 1.1424096822738647, |
|
"learning_rate": 4.79704975853109e-05, |
|
"loss": 1.9661, |
|
"step": 5270 |
|
}, |
|
{ |
|
"epoch": 0.13347097553913362, |
|
"grad_norm": 1.131800889968872, |
|
"learning_rate": 4.793722210363262e-05, |
|
"loss": 1.9828, |
|
"step": 5301 |
|
}, |
|
{ |
|
"epoch": 0.13425150755983029, |
|
"grad_norm": 1.1227242946624756, |
|
"learning_rate": 4.7903687778520414e-05, |
|
"loss": 1.975, |
|
"step": 5332 |
|
}, |
|
{ |
|
"epoch": 0.13503203958052698, |
|
"grad_norm": 1.143832802772522, |
|
"learning_rate": 4.7869894988404593e-05, |
|
"loss": 1.9619, |
|
"step": 5363 |
|
}, |
|
{ |
|
"epoch": 0.13581257160122367, |
|
"grad_norm": 1.1281683444976807, |
|
"learning_rate": 4.783584411463221e-05, |
|
"loss": 1.9603, |
|
"step": 5394 |
|
}, |
|
{ |
|
"epoch": 0.13659310362192037, |
|
"grad_norm": 1.134901762008667, |
|
"learning_rate": 4.780153554146274e-05, |
|
"loss": 1.9888, |
|
"step": 5425 |
|
}, |
|
{ |
|
"epoch": 0.13737363564261704, |
|
"grad_norm": 1.1317532062530518, |
|
"learning_rate": 4.7766969656063766e-05, |
|
"loss": 1.9793, |
|
"step": 5456 |
|
}, |
|
{ |
|
"epoch": 0.13815416766331373, |
|
"grad_norm": 1.2516345977783203, |
|
"learning_rate": 4.773214684850662e-05, |
|
"loss": 1.9839, |
|
"step": 5487 |
|
}, |
|
{ |
|
"epoch": 0.13893469968401043, |
|
"grad_norm": 1.107391119003296, |
|
"learning_rate": 4.769706751176193e-05, |
|
"loss": 1.9606, |
|
"step": 5518 |
|
}, |
|
{ |
|
"epoch": 0.13971523170470712, |
|
"grad_norm": 1.1041492223739624, |
|
"learning_rate": 4.7661732041695264e-05, |
|
"loss": 1.9688, |
|
"step": 5549 |
|
}, |
|
{ |
|
"epoch": 0.1404957637254038, |
|
"grad_norm": 1.1099929809570312, |
|
"learning_rate": 4.762614083706258e-05, |
|
"loss": 1.9821, |
|
"step": 5580 |
|
}, |
|
{ |
|
"epoch": 0.14127629574610048, |
|
"grad_norm": 1.167151927947998, |
|
"learning_rate": 4.759029429950581e-05, |
|
"loss": 1.9669, |
|
"step": 5611 |
|
}, |
|
{ |
|
"epoch": 0.14205682776679718, |
|
"grad_norm": 1.0942577123641968, |
|
"learning_rate": 4.7554192833548235e-05, |
|
"loss": 1.9565, |
|
"step": 5642 |
|
}, |
|
{ |
|
"epoch": 0.14283735978749387, |
|
"grad_norm": 1.0859640836715698, |
|
"learning_rate": 4.751783684659e-05, |
|
"loss": 1.9634, |
|
"step": 5673 |
|
}, |
|
{ |
|
"epoch": 0.14361789180819054, |
|
"grad_norm": 2.420607089996338, |
|
"learning_rate": 4.748122674890348e-05, |
|
"loss": 1.9365, |
|
"step": 5704 |
|
}, |
|
{ |
|
"epoch": 0.14439842382888723, |
|
"grad_norm": 1.0996880531311035, |
|
"learning_rate": 4.7444362953628654e-05, |
|
"loss": 1.9718, |
|
"step": 5735 |
|
}, |
|
{ |
|
"epoch": 0.14517895584958393, |
|
"grad_norm": 1.1183338165283203, |
|
"learning_rate": 4.7407245876768424e-05, |
|
"loss": 1.9564, |
|
"step": 5766 |
|
}, |
|
{ |
|
"epoch": 0.14595948787028062, |
|
"grad_norm": 1.1093947887420654, |
|
"learning_rate": 4.736987593718397e-05, |
|
"loss": 1.9516, |
|
"step": 5797 |
|
}, |
|
{ |
|
"epoch": 0.1467400198909773, |
|
"grad_norm": 1.141797423362732, |
|
"learning_rate": 4.733225355658999e-05, |
|
"loss": 1.929, |
|
"step": 5828 |
|
}, |
|
{ |
|
"epoch": 0.14752055191167399, |
|
"grad_norm": 1.1214358806610107, |
|
"learning_rate": 4.7294379159549926e-05, |
|
"loss": 1.9297, |
|
"step": 5859 |
|
}, |
|
{ |
|
"epoch": 0.14830108393237068, |
|
"grad_norm": 1.1025431156158447, |
|
"learning_rate": 4.725625317347119e-05, |
|
"loss": 1.9226, |
|
"step": 5890 |
|
}, |
|
{ |
|
"epoch": 0.14908161595306738, |
|
"grad_norm": 1.1094226837158203, |
|
"learning_rate": 4.7217876028600374e-05, |
|
"loss": 1.946, |
|
"step": 5921 |
|
}, |
|
{ |
|
"epoch": 0.14986214797376404, |
|
"grad_norm": 1.0909287929534912, |
|
"learning_rate": 4.717924815801832e-05, |
|
"loss": 1.9511, |
|
"step": 5952 |
|
}, |
|
{ |
|
"epoch": 0.15064267999446074, |
|
"grad_norm": 1.0755189657211304, |
|
"learning_rate": 4.714036999763532e-05, |
|
"loss": 1.939, |
|
"step": 5983 |
|
}, |
|
{ |
|
"epoch": 0.15142321201515743, |
|
"grad_norm": 1.1426901817321777, |
|
"learning_rate": 4.7101241986186116e-05, |
|
"loss": 1.93, |
|
"step": 6014 |
|
}, |
|
{ |
|
"epoch": 0.15220374403585413, |
|
"grad_norm": 1.1208311319351196, |
|
"learning_rate": 4.7061864565225e-05, |
|
"loss": 1.9516, |
|
"step": 6045 |
|
}, |
|
{ |
|
"epoch": 0.1529842760565508, |
|
"grad_norm": 1.0931717157363892, |
|
"learning_rate": 4.702223817912081e-05, |
|
"loss": 1.9314, |
|
"step": 6076 |
|
}, |
|
{ |
|
"epoch": 0.1537648080772475, |
|
"grad_norm": 1.075864553451538, |
|
"learning_rate": 4.698236327505195e-05, |
|
"loss": 1.935, |
|
"step": 6107 |
|
}, |
|
{ |
|
"epoch": 0.15454534009794418, |
|
"grad_norm": 1.090559720993042, |
|
"learning_rate": 4.694224030300127e-05, |
|
"loss": 1.9431, |
|
"step": 6138 |
|
}, |
|
{ |
|
"epoch": 0.15532587211864088, |
|
"grad_norm": 1.0956670045852661, |
|
"learning_rate": 4.690186971575107e-05, |
|
"loss": 1.9678, |
|
"step": 6169 |
|
}, |
|
{ |
|
"epoch": 0.15610640413933755, |
|
"grad_norm": 1.125927448272705, |
|
"learning_rate": 4.6861251968877916e-05, |
|
"loss": 1.9292, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 0.15688693616003424, |
|
"grad_norm": 1.0844208002090454, |
|
"learning_rate": 4.68203875207476e-05, |
|
"loss": 1.9189, |
|
"step": 6231 |
|
}, |
|
{ |
|
"epoch": 0.15766746818073094, |
|
"grad_norm": 1.064327597618103, |
|
"learning_rate": 4.677927683250983e-05, |
|
"loss": 1.928, |
|
"step": 6262 |
|
}, |
|
{ |
|
"epoch": 0.15844800020142763, |
|
"grad_norm": 1.1281861066818237, |
|
"learning_rate": 4.6737920368093156e-05, |
|
"loss": 1.9437, |
|
"step": 6293 |
|
}, |
|
{ |
|
"epoch": 0.1592285322221243, |
|
"grad_norm": 1.0674384832382202, |
|
"learning_rate": 4.669631859419965e-05, |
|
"loss": 1.9347, |
|
"step": 6324 |
|
}, |
|
{ |
|
"epoch": 0.160009064242821, |
|
"grad_norm": 1.105566382408142, |
|
"learning_rate": 4.6654471980299676e-05, |
|
"loss": 1.9229, |
|
"step": 6355 |
|
}, |
|
{ |
|
"epoch": 0.1607895962635177, |
|
"grad_norm": 1.0582126379013062, |
|
"learning_rate": 4.661238099862658e-05, |
|
"loss": 1.9321, |
|
"step": 6386 |
|
}, |
|
{ |
|
"epoch": 0.16157012828421438, |
|
"grad_norm": 1.1147974729537964, |
|
"learning_rate": 4.657004612417138e-05, |
|
"loss": 1.9289, |
|
"step": 6417 |
|
}, |
|
{ |
|
"epoch": 0.16235066030491105, |
|
"grad_norm": 1.0804343223571777, |
|
"learning_rate": 4.6527467834677374e-05, |
|
"loss": 1.9307, |
|
"step": 6448 |
|
}, |
|
{ |
|
"epoch": 0.16313119232560774, |
|
"grad_norm": 1.0830508470535278, |
|
"learning_rate": 4.648464661063478e-05, |
|
"loss": 1.9408, |
|
"step": 6479 |
|
}, |
|
{ |
|
"epoch": 0.16391172434630444, |
|
"grad_norm": 1.0533562898635864, |
|
"learning_rate": 4.6441582935275264e-05, |
|
"loss": 1.9347, |
|
"step": 6510 |
|
}, |
|
{ |
|
"epoch": 0.16469225636700113, |
|
"grad_norm": 1.044925332069397, |
|
"learning_rate": 4.6398277294566586e-05, |
|
"loss": 1.9112, |
|
"step": 6541 |
|
}, |
|
{ |
|
"epoch": 0.1654727883876978, |
|
"grad_norm": 1.087296962738037, |
|
"learning_rate": 4.6354730177207e-05, |
|
"loss": 1.9381, |
|
"step": 6572 |
|
}, |
|
{ |
|
"epoch": 0.1662533204083945, |
|
"grad_norm": 1.0853668451309204, |
|
"learning_rate": 4.6310942074619787e-05, |
|
"loss": 1.9224, |
|
"step": 6603 |
|
}, |
|
{ |
|
"epoch": 0.1670338524290912, |
|
"grad_norm": 1.0661752223968506, |
|
"learning_rate": 4.626691348094777e-05, |
|
"loss": 1.9386, |
|
"step": 6634 |
|
}, |
|
{ |
|
"epoch": 0.16781438444978788, |
|
"grad_norm": 1.0393208265304565, |
|
"learning_rate": 4.622264489304762e-05, |
|
"loss": 1.9266, |
|
"step": 6665 |
|
}, |
|
{ |
|
"epoch": 0.16859491647048455, |
|
"grad_norm": 1.0324763059616089, |
|
"learning_rate": 4.617813681048434e-05, |
|
"loss": 1.919, |
|
"step": 6696 |
|
}, |
|
{ |
|
"epoch": 0.16937544849118125, |
|
"grad_norm": 1.0684878826141357, |
|
"learning_rate": 4.61333897355256e-05, |
|
"loss": 1.9321, |
|
"step": 6727 |
|
}, |
|
{ |
|
"epoch": 0.17015598051187794, |
|
"grad_norm": 1.0477505922317505, |
|
"learning_rate": 4.608840417313604e-05, |
|
"loss": 1.9449, |
|
"step": 6758 |
|
}, |
|
{ |
|
"epoch": 0.17093651253257464, |
|
"grad_norm": 1.0684869289398193, |
|
"learning_rate": 4.6043180630971646e-05, |
|
"loss": 1.9252, |
|
"step": 6789 |
|
}, |
|
{ |
|
"epoch": 0.1717170445532713, |
|
"grad_norm": 1.0869554281234741, |
|
"learning_rate": 4.599771961937391e-05, |
|
"loss": 1.9183, |
|
"step": 6820 |
|
}, |
|
{ |
|
"epoch": 0.172497576573968, |
|
"grad_norm": 1.0297125577926636, |
|
"learning_rate": 4.5952021651364204e-05, |
|
"loss": 1.9123, |
|
"step": 6851 |
|
}, |
|
{ |
|
"epoch": 0.1732781085946647, |
|
"grad_norm": 1.0670006275177002, |
|
"learning_rate": 4.590608724263786e-05, |
|
"loss": 1.9353, |
|
"step": 6882 |
|
}, |
|
{ |
|
"epoch": 0.1740586406153614, |
|
"grad_norm": 1.0744417905807495, |
|
"learning_rate": 4.585991691155845e-05, |
|
"loss": 1.8979, |
|
"step": 6913 |
|
}, |
|
{ |
|
"epoch": 0.17483917263605805, |
|
"grad_norm": 1.0910695791244507, |
|
"learning_rate": 4.581351117915188e-05, |
|
"loss": 1.9153, |
|
"step": 6944 |
|
}, |
|
{ |
|
"epoch": 0.17561970465675475, |
|
"grad_norm": 1.083776593208313, |
|
"learning_rate": 4.5766870569100534e-05, |
|
"loss": 1.9026, |
|
"step": 6975 |
|
}, |
|
{ |
|
"epoch": 0.17640023667745144, |
|
"grad_norm": 1.0857270956039429, |
|
"learning_rate": 4.571999560773736e-05, |
|
"loss": 1.9279, |
|
"step": 7006 |
|
}, |
|
{ |
|
"epoch": 0.1771807686981481, |
|
"grad_norm": 1.0350106954574585, |
|
"learning_rate": 4.5672886824039915e-05, |
|
"loss": 1.9236, |
|
"step": 7037 |
|
}, |
|
{ |
|
"epoch": 0.1779613007188448, |
|
"grad_norm": 1.0498634576797485, |
|
"learning_rate": 4.5625544749624435e-05, |
|
"loss": 1.8934, |
|
"step": 7068 |
|
}, |
|
{ |
|
"epoch": 0.1787418327395415, |
|
"grad_norm": 1.0836471319198608, |
|
"learning_rate": 4.5577969918739794e-05, |
|
"loss": 1.9131, |
|
"step": 7099 |
|
}, |
|
{ |
|
"epoch": 0.1795223647602382, |
|
"grad_norm": 1.0414760112762451, |
|
"learning_rate": 4.5530162868261486e-05, |
|
"loss": 1.8792, |
|
"step": 7130 |
|
}, |
|
{ |
|
"epoch": 0.18030289678093486, |
|
"grad_norm": 1.0747581720352173, |
|
"learning_rate": 4.548212413768558e-05, |
|
"loss": 1.8973, |
|
"step": 7161 |
|
}, |
|
{ |
|
"epoch": 0.18108342880163156, |
|
"grad_norm": 1.0464451313018799, |
|
"learning_rate": 4.543385426912261e-05, |
|
"loss": 1.907, |
|
"step": 7192 |
|
}, |
|
{ |
|
"epoch": 0.18186396082232825, |
|
"grad_norm": 1.0159403085708618, |
|
"learning_rate": 4.53853538072915e-05, |
|
"loss": 1.9111, |
|
"step": 7223 |
|
}, |
|
{ |
|
"epoch": 0.18264449284302495, |
|
"grad_norm": 1.093883991241455, |
|
"learning_rate": 4.533662329951336e-05, |
|
"loss": 1.8927, |
|
"step": 7254 |
|
}, |
|
{ |
|
"epoch": 0.18342502486372161, |
|
"grad_norm": 1.0358048677444458, |
|
"learning_rate": 4.528766329570536e-05, |
|
"loss": 1.9164, |
|
"step": 7285 |
|
}, |
|
{ |
|
"epoch": 0.1842055568844183, |
|
"grad_norm": 1.051798939704895, |
|
"learning_rate": 4.523847434837447e-05, |
|
"loss": 1.9084, |
|
"step": 7316 |
|
}, |
|
{ |
|
"epoch": 0.184986088905115, |
|
"grad_norm": 1.0729514360427856, |
|
"learning_rate": 4.518905701261128e-05, |
|
"loss": 1.9089, |
|
"step": 7347 |
|
}, |
|
{ |
|
"epoch": 0.1857666209258117, |
|
"grad_norm": 1.0399775505065918, |
|
"learning_rate": 4.5139411846083715e-05, |
|
"loss": 1.922, |
|
"step": 7378 |
|
}, |
|
{ |
|
"epoch": 0.18654715294650837, |
|
"grad_norm": 1.0500715970993042, |
|
"learning_rate": 4.508953940903073e-05, |
|
"loss": 1.9118, |
|
"step": 7409 |
|
}, |
|
{ |
|
"epoch": 0.18732768496720506, |
|
"grad_norm": 1.0770058631896973, |
|
"learning_rate": 4.5039440264255994e-05, |
|
"loss": 1.9059, |
|
"step": 7440 |
|
}, |
|
{ |
|
"epoch": 0.18810821698790176, |
|
"grad_norm": 1.0471183061599731, |
|
"learning_rate": 4.498911497712155e-05, |
|
"loss": 1.9119, |
|
"step": 7471 |
|
}, |
|
{ |
|
"epoch": 0.18888874900859845, |
|
"grad_norm": 1.038761019706726, |
|
"learning_rate": 4.493856411554142e-05, |
|
"loss": 1.8798, |
|
"step": 7502 |
|
}, |
|
{ |
|
"epoch": 0.18966928102929512, |
|
"grad_norm": 1.0365252494812012, |
|
"learning_rate": 4.4887788249975206e-05, |
|
"loss": 1.8983, |
|
"step": 7533 |
|
}, |
|
{ |
|
"epoch": 0.1904498130499918, |
|
"grad_norm": 1.1033800840377808, |
|
"learning_rate": 4.4836787953421656e-05, |
|
"loss": 1.8911, |
|
"step": 7564 |
|
}, |
|
{ |
|
"epoch": 0.1912303450706885, |
|
"grad_norm": 1.064213514328003, |
|
"learning_rate": 4.478556380141218e-05, |
|
"loss": 1.8995, |
|
"step": 7595 |
|
}, |
|
{ |
|
"epoch": 0.1920108770913852, |
|
"grad_norm": 0.997905969619751, |
|
"learning_rate": 4.4734116372004375e-05, |
|
"loss": 1.8705, |
|
"step": 7626 |
|
}, |
|
{ |
|
"epoch": 0.19279140911208187, |
|
"grad_norm": 1.0392402410507202, |
|
"learning_rate": 4.4682446245775477e-05, |
|
"loss": 1.9105, |
|
"step": 7657 |
|
}, |
|
{ |
|
"epoch": 0.19357194113277856, |
|
"grad_norm": 1.0340332984924316, |
|
"learning_rate": 4.463055400581586e-05, |
|
"loss": 1.9074, |
|
"step": 7688 |
|
}, |
|
{ |
|
"epoch": 0.19435247315347526, |
|
"grad_norm": 1.0196648836135864, |
|
"learning_rate": 4.4578440237722374e-05, |
|
"loss": 1.9059, |
|
"step": 7719 |
|
}, |
|
{ |
|
"epoch": 0.19513300517417195, |
|
"grad_norm": 0.99578458070755, |
|
"learning_rate": 4.452610552959183e-05, |
|
"loss": 1.8974, |
|
"step": 7750 |
|
}, |
|
{ |
|
"epoch": 0.19591353719486862, |
|
"grad_norm": 1.026090145111084, |
|
"learning_rate": 4.447355047201428e-05, |
|
"loss": 1.8886, |
|
"step": 7781 |
|
}, |
|
{ |
|
"epoch": 0.19669406921556531, |
|
"grad_norm": 0.9945096969604492, |
|
"learning_rate": 4.4420775658066414e-05, |
|
"loss": 1.8879, |
|
"step": 7812 |
|
}, |
|
{ |
|
"epoch": 0.197474601236262, |
|
"grad_norm": 1.0226850509643555, |
|
"learning_rate": 4.436778168330484e-05, |
|
"loss": 1.8936, |
|
"step": 7843 |
|
}, |
|
{ |
|
"epoch": 0.1982551332569587, |
|
"grad_norm": 1.0433967113494873, |
|
"learning_rate": 4.4314569145759353e-05, |
|
"loss": 1.882, |
|
"step": 7874 |
|
}, |
|
{ |
|
"epoch": 0.19903566527765537, |
|
"grad_norm": 1.0612221956253052, |
|
"learning_rate": 4.42611386459262e-05, |
|
"loss": 1.8879, |
|
"step": 7905 |
|
}, |
|
{ |
|
"epoch": 0.19981619729835207, |
|
"grad_norm": 1.0252665281295776, |
|
"learning_rate": 4.420749078676133e-05, |
|
"loss": 1.9005, |
|
"step": 7936 |
|
}, |
|
{ |
|
"epoch": 0.20059672931904876, |
|
"grad_norm": 1.015100359916687, |
|
"learning_rate": 4.4153626173673516e-05, |
|
"loss": 1.8944, |
|
"step": 7967 |
|
}, |
|
{ |
|
"epoch": 0.20137726133974546, |
|
"grad_norm": 1.0417771339416504, |
|
"learning_rate": 4.409954541451762e-05, |
|
"loss": 1.8837, |
|
"step": 7998 |
|
}, |
|
{ |
|
"epoch": 0.20215779336044212, |
|
"grad_norm": 1.026708960533142, |
|
"learning_rate": 4.404524911958764e-05, |
|
"loss": 1.92, |
|
"step": 8029 |
|
}, |
|
{ |
|
"epoch": 0.20293832538113882, |
|
"grad_norm": 1.0414159297943115, |
|
"learning_rate": 4.399073790160989e-05, |
|
"loss": 1.8709, |
|
"step": 8060 |
|
}, |
|
{ |
|
"epoch": 0.2037188574018355, |
|
"grad_norm": 1.014039397239685, |
|
"learning_rate": 4.393601237573607e-05, |
|
"loss": 1.8842, |
|
"step": 8091 |
|
}, |
|
{ |
|
"epoch": 0.2044993894225322, |
|
"grad_norm": 1.0103037357330322, |
|
"learning_rate": 4.388107315953628e-05, |
|
"loss": 1.8889, |
|
"step": 8122 |
|
}, |
|
{ |
|
"epoch": 0.20527992144322887, |
|
"grad_norm": 1.0166395902633667, |
|
"learning_rate": 4.382592087299212e-05, |
|
"loss": 1.8808, |
|
"step": 8153 |
|
}, |
|
{ |
|
"epoch": 0.20606045346392557, |
|
"grad_norm": 1.023850679397583, |
|
"learning_rate": 4.377055613848964e-05, |
|
"loss": 1.8918, |
|
"step": 8184 |
|
}, |
|
{ |
|
"epoch": 0.20684098548462226, |
|
"grad_norm": 1.0251572132110596, |
|
"learning_rate": 4.3714979580812355e-05, |
|
"loss": 1.882, |
|
"step": 8215 |
|
}, |
|
{ |
|
"epoch": 0.20762151750531896, |
|
"grad_norm": 1.0183281898498535, |
|
"learning_rate": 4.365919182713416e-05, |
|
"loss": 1.8844, |
|
"step": 8246 |
|
}, |
|
{ |
|
"epoch": 0.20840204952601563, |
|
"grad_norm": 1.0135486125946045, |
|
"learning_rate": 4.360319350701226e-05, |
|
"loss": 1.8814, |
|
"step": 8277 |
|
}, |
|
{ |
|
"epoch": 0.20918258154671232, |
|
"grad_norm": 1.0247812271118164, |
|
"learning_rate": 4.3546985252380115e-05, |
|
"loss": 1.8818, |
|
"step": 8308 |
|
}, |
|
{ |
|
"epoch": 0.20996311356740902, |
|
"grad_norm": 1.0123242139816284, |
|
"learning_rate": 4.349056769754021e-05, |
|
"loss": 1.8723, |
|
"step": 8339 |
|
}, |
|
{ |
|
"epoch": 0.2107436455881057, |
|
"grad_norm": 1.0139250755310059, |
|
"learning_rate": 4.3433941479156994e-05, |
|
"loss": 1.8858, |
|
"step": 8370 |
|
}, |
|
{ |
|
"epoch": 0.21152417760880238, |
|
"grad_norm": 4.008477687835693, |
|
"learning_rate": 4.3377107236249647e-05, |
|
"loss": 1.8735, |
|
"step": 8401 |
|
}, |
|
{ |
|
"epoch": 0.21230470962949907, |
|
"grad_norm": 1.0454871654510498, |
|
"learning_rate": 4.332006561018488e-05, |
|
"loss": 1.8824, |
|
"step": 8432 |
|
}, |
|
{ |
|
"epoch": 0.21308524165019577, |
|
"grad_norm": 1.0586799383163452, |
|
"learning_rate": 4.3262817244669683e-05, |
|
"loss": 1.8992, |
|
"step": 8463 |
|
}, |
|
{ |
|
"epoch": 0.21386577367089246, |
|
"grad_norm": 1.0073469877243042, |
|
"learning_rate": 4.3205362785744083e-05, |
|
"loss": 1.8766, |
|
"step": 8494 |
|
}, |
|
{ |
|
"epoch": 0.21464630569158913, |
|
"grad_norm": 0.9976999759674072, |
|
"learning_rate": 4.314770288177384e-05, |
|
"loss": 1.8864, |
|
"step": 8525 |
|
}, |
|
{ |
|
"epoch": 0.21542683771228582, |
|
"grad_norm": 0.9928112030029297, |
|
"learning_rate": 4.308983818344313e-05, |
|
"loss": 1.8874, |
|
"step": 8556 |
|
}, |
|
{ |
|
"epoch": 0.21620736973298252, |
|
"grad_norm": 1.038405418395996, |
|
"learning_rate": 4.3031769343747206e-05, |
|
"loss": 1.8725, |
|
"step": 8587 |
|
}, |
|
{ |
|
"epoch": 0.2169879017536792, |
|
"grad_norm": 1.0328835248947144, |
|
"learning_rate": 4.297349701798505e-05, |
|
"loss": 1.8884, |
|
"step": 8618 |
|
}, |
|
{ |
|
"epoch": 0.21776843377437588, |
|
"grad_norm": 1.0254933834075928, |
|
"learning_rate": 4.2915021863751916e-05, |
|
"loss": 1.8611, |
|
"step": 8649 |
|
}, |
|
{ |
|
"epoch": 0.21854896579507258, |
|
"grad_norm": 0.99531090259552, |
|
"learning_rate": 4.285634454093198e-05, |
|
"loss": 1.8889, |
|
"step": 8680 |
|
}, |
|
{ |
|
"epoch": 0.21932949781576927, |
|
"grad_norm": 1.0536468029022217, |
|
"learning_rate": 4.279746571169086e-05, |
|
"loss": 1.8875, |
|
"step": 8711 |
|
}, |
|
{ |
|
"epoch": 0.22011002983646596, |
|
"grad_norm": 0.9915075898170471, |
|
"learning_rate": 4.2738386040468136e-05, |
|
"loss": 1.8901, |
|
"step": 8742 |
|
}, |
|
{ |
|
"epoch": 0.22089056185716263, |
|
"grad_norm": 0.9883093237876892, |
|
"learning_rate": 4.2679106193969866e-05, |
|
"loss": 1.8704, |
|
"step": 8773 |
|
}, |
|
{ |
|
"epoch": 0.22167109387785933, |
|
"grad_norm": 1.0425434112548828, |
|
"learning_rate": 4.261962684116106e-05, |
|
"loss": 1.8643, |
|
"step": 8804 |
|
}, |
|
{ |
|
"epoch": 0.22245162589855602, |
|
"grad_norm": 1.0154091119766235, |
|
"learning_rate": 4.2559948653258145e-05, |
|
"loss": 1.8806, |
|
"step": 8835 |
|
}, |
|
{ |
|
"epoch": 0.22323215791925272, |
|
"grad_norm": 1.0432935953140259, |
|
"learning_rate": 4.250007230372134e-05, |
|
"loss": 1.8928, |
|
"step": 8866 |
|
}, |
|
{ |
|
"epoch": 0.22401268993994938, |
|
"grad_norm": 0.9709262251853943, |
|
"learning_rate": 4.2439998468247126e-05, |
|
"loss": 1.8591, |
|
"step": 8897 |
|
}, |
|
{ |
|
"epoch": 0.22479322196064608, |
|
"grad_norm": 0.9708088636398315, |
|
"learning_rate": 4.2379727824760566e-05, |
|
"loss": 1.8797, |
|
"step": 8928 |
|
}, |
|
{ |
|
"epoch": 0.22557375398134277, |
|
"grad_norm": 0.9911683201789856, |
|
"learning_rate": 4.231926105340768e-05, |
|
"loss": 1.8456, |
|
"step": 8959 |
|
}, |
|
{ |
|
"epoch": 0.22635428600203947, |
|
"grad_norm": 0.99161297082901, |
|
"learning_rate": 4.225859883654776e-05, |
|
"loss": 1.8712, |
|
"step": 8990 |
|
}, |
|
{ |
|
"epoch": 0.22713481802273613, |
|
"grad_norm": 1.0025807619094849, |
|
"learning_rate": 4.219774185874569e-05, |
|
"loss": 1.8814, |
|
"step": 9021 |
|
}, |
|
{ |
|
"epoch": 0.22791535004343283, |
|
"grad_norm": 0.9597026109695435, |
|
"learning_rate": 4.213669080676418e-05, |
|
"loss": 1.8799, |
|
"step": 9052 |
|
}, |
|
{ |
|
"epoch": 0.22869588206412952, |
|
"grad_norm": 1.0001273155212402, |
|
"learning_rate": 4.2075446369556056e-05, |
|
"loss": 1.8469, |
|
"step": 9083 |
|
}, |
|
{ |
|
"epoch": 0.22947641408482622, |
|
"grad_norm": 0.9670659303665161, |
|
"learning_rate": 4.201400923825648e-05, |
|
"loss": 1.8612, |
|
"step": 9114 |
|
}, |
|
{ |
|
"epoch": 0.23025694610552289, |
|
"grad_norm": 1.018050193786621, |
|
"learning_rate": 4.195238010617511e-05, |
|
"loss": 1.8656, |
|
"step": 9145 |
|
}, |
|
{ |
|
"epoch": 0.23103747812621958, |
|
"grad_norm": 0.9838584661483765, |
|
"learning_rate": 4.1890559668788344e-05, |
|
"loss": 1.8527, |
|
"step": 9176 |
|
}, |
|
{ |
|
"epoch": 0.23181801014691628, |
|
"grad_norm": 0.9968937635421753, |
|
"learning_rate": 4.1828548623731405e-05, |
|
"loss": 1.8753, |
|
"step": 9207 |
|
}, |
|
{ |
|
"epoch": 0.23259854216761297, |
|
"grad_norm": 0.9698464870452881, |
|
"learning_rate": 4.1766347670790506e-05, |
|
"loss": 1.8716, |
|
"step": 9238 |
|
}, |
|
{ |
|
"epoch": 0.23337907418830964, |
|
"grad_norm": 0.9592704772949219, |
|
"learning_rate": 4.170395751189495e-05, |
|
"loss": 1.8698, |
|
"step": 9269 |
|
}, |
|
{ |
|
"epoch": 0.23415960620900633, |
|
"grad_norm": 0.9904705286026001, |
|
"learning_rate": 4.164137885110921e-05, |
|
"loss": 1.8661, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 0.23494013822970303, |
|
"grad_norm": 1.0243868827819824, |
|
"learning_rate": 4.157861239462495e-05, |
|
"loss": 1.8658, |
|
"step": 9331 |
|
}, |
|
{ |
|
"epoch": 0.2357206702503997, |
|
"grad_norm": 0.9936602711677551, |
|
"learning_rate": 4.1515658850753114e-05, |
|
"loss": 1.8475, |
|
"step": 9362 |
|
}, |
|
{ |
|
"epoch": 0.2365012022710964, |
|
"grad_norm": 0.9626069068908691, |
|
"learning_rate": 4.145251892991588e-05, |
|
"loss": 1.8507, |
|
"step": 9393 |
|
}, |
|
{ |
|
"epoch": 0.23728173429179308, |
|
"grad_norm": 0.9688258767127991, |
|
"learning_rate": 4.138919334463868e-05, |
|
"loss": 1.8628, |
|
"step": 9424 |
|
}, |
|
{ |
|
"epoch": 0.23806226631248978, |
|
"grad_norm": 0.9996054172515869, |
|
"learning_rate": 4.1325682809542124e-05, |
|
"loss": 1.8734, |
|
"step": 9455 |
|
}, |
|
{ |
|
"epoch": 0.23884279833318645, |
|
"grad_norm": 1.0037505626678467, |
|
"learning_rate": 4.126198804133398e-05, |
|
"loss": 1.8778, |
|
"step": 9486 |
|
}, |
|
{ |
|
"epoch": 0.23962333035388314, |
|
"grad_norm": 0.9751474857330322, |
|
"learning_rate": 4.1198109758801055e-05, |
|
"loss": 1.8615, |
|
"step": 9517 |
|
}, |
|
{ |
|
"epoch": 0.24040386237457984, |
|
"grad_norm": 0.9760595560073853, |
|
"learning_rate": 4.113404868280107e-05, |
|
"loss": 1.8614, |
|
"step": 9548 |
|
}, |
|
{ |
|
"epoch": 0.24118439439527653, |
|
"grad_norm": 1.0009572505950928, |
|
"learning_rate": 4.106980553625457e-05, |
|
"loss": 1.8762, |
|
"step": 9579 |
|
}, |
|
{ |
|
"epoch": 0.2419649264159732, |
|
"grad_norm": 0.983182966709137, |
|
"learning_rate": 4.100538104413674e-05, |
|
"loss": 1.857, |
|
"step": 9610 |
|
}, |
|
{ |
|
"epoch": 0.2427454584366699, |
|
"grad_norm": 1.0748534202575684, |
|
"learning_rate": 4.09407759334692e-05, |
|
"loss": 1.8658, |
|
"step": 9641 |
|
}, |
|
{ |
|
"epoch": 0.2435259904573666, |
|
"grad_norm": 0.9747381806373596, |
|
"learning_rate": 4.087599093331186e-05, |
|
"loss": 1.8484, |
|
"step": 9672 |
|
}, |
|
{ |
|
"epoch": 0.24430652247806328, |
|
"grad_norm": 0.9784808158874512, |
|
"learning_rate": 4.081102677475462e-05, |
|
"loss": 1.8492, |
|
"step": 9703 |
|
}, |
|
{ |
|
"epoch": 0.24508705449875995, |
|
"grad_norm": 0.9820215106010437, |
|
"learning_rate": 4.0745884190909194e-05, |
|
"loss": 1.8516, |
|
"step": 9734 |
|
}, |
|
{ |
|
"epoch": 0.24586758651945664, |
|
"grad_norm": 0.9536774158477783, |
|
"learning_rate": 4.0680563916900796e-05, |
|
"loss": 1.8444, |
|
"step": 9765 |
|
}, |
|
{ |
|
"epoch": 0.24664811854015334, |
|
"grad_norm": 0.9706370234489441, |
|
"learning_rate": 4.0615066689859815e-05, |
|
"loss": 1.8477, |
|
"step": 9796 |
|
}, |
|
{ |
|
"epoch": 0.24742865056085003, |
|
"grad_norm": 0.9700275659561157, |
|
"learning_rate": 4.0549393248913584e-05, |
|
"loss": 1.8515, |
|
"step": 9827 |
|
}, |
|
{ |
|
"epoch": 0.2482091825815467, |
|
"grad_norm": 0.9522431492805481, |
|
"learning_rate": 4.048354433517794e-05, |
|
"loss": 1.8287, |
|
"step": 9858 |
|
}, |
|
{ |
|
"epoch": 0.2489897146022434, |
|
"grad_norm": 0.9691547155380249, |
|
"learning_rate": 4.0417520691748916e-05, |
|
"loss": 1.8594, |
|
"step": 9889 |
|
}, |
|
{ |
|
"epoch": 0.2497702466229401, |
|
"grad_norm": 0.9588534235954285, |
|
"learning_rate": 4.035132306369438e-05, |
|
"loss": 1.8459, |
|
"step": 9920 |
|
}, |
|
{ |
|
"epoch": 0.2505507786436368, |
|
"grad_norm": 0.9900926351547241, |
|
"learning_rate": 4.028495219804555e-05, |
|
"loss": 1.8608, |
|
"step": 9951 |
|
}, |
|
{ |
|
"epoch": 0.25133131066433345, |
|
"grad_norm": 0.9539284110069275, |
|
"learning_rate": 4.021840884378864e-05, |
|
"loss": 1.8522, |
|
"step": 9982 |
|
}, |
|
{ |
|
"epoch": 0.2521118426850302, |
|
"grad_norm": 0.9557245969772339, |
|
"learning_rate": 4.015169375185633e-05, |
|
"loss": 1.8743, |
|
"step": 10013 |
|
}, |
|
{ |
|
"epoch": 0.25289237470572684, |
|
"grad_norm": 0.9317564964294434, |
|
"learning_rate": 4.0084807675119396e-05, |
|
"loss": 1.8384, |
|
"step": 10044 |
|
}, |
|
{ |
|
"epoch": 0.2536729067264235, |
|
"grad_norm": 0.9944160580635071, |
|
"learning_rate": 4.0017751368378106e-05, |
|
"loss": 1.857, |
|
"step": 10075 |
|
}, |
|
{ |
|
"epoch": 0.25445343874712023, |
|
"grad_norm": 0.9859464168548584, |
|
"learning_rate": 3.995052558835377e-05, |
|
"loss": 1.8484, |
|
"step": 10106 |
|
}, |
|
{ |
|
"epoch": 0.2552339707678169, |
|
"grad_norm": 1.0093594789505005, |
|
"learning_rate": 3.988313109368017e-05, |
|
"loss": 1.8506, |
|
"step": 10137 |
|
}, |
|
{ |
|
"epoch": 0.25601450278851356, |
|
"grad_norm": 0.9451240301132202, |
|
"learning_rate": 3.981556864489504e-05, |
|
"loss": 1.8732, |
|
"step": 10168 |
|
}, |
|
{ |
|
"epoch": 0.2567950348092103, |
|
"grad_norm": 0.972819447517395, |
|
"learning_rate": 3.974783900443142e-05, |
|
"loss": 1.856, |
|
"step": 10199 |
|
}, |
|
{ |
|
"epoch": 0.25757556682990695, |
|
"grad_norm": 0.996379554271698, |
|
"learning_rate": 3.9679942936609095e-05, |
|
"loss": 1.8299, |
|
"step": 10230 |
|
}, |
|
{ |
|
"epoch": 0.2583560988506037, |
|
"grad_norm": 2.503967046737671, |
|
"learning_rate": 3.961188120762596e-05, |
|
"loss": 1.8484, |
|
"step": 10261 |
|
}, |
|
{ |
|
"epoch": 0.25913663087130034, |
|
"grad_norm": 0.9491912126541138, |
|
"learning_rate": 3.954365458554938e-05, |
|
"loss": 1.8528, |
|
"step": 10292 |
|
}, |
|
{ |
|
"epoch": 0.259917162891997, |
|
"grad_norm": 0.9642863273620605, |
|
"learning_rate": 3.947526384030751e-05, |
|
"loss": 1.863, |
|
"step": 10323 |
|
}, |
|
{ |
|
"epoch": 0.26069769491269373, |
|
"grad_norm": 0.9915285706520081, |
|
"learning_rate": 3.9406709743680624e-05, |
|
"loss": 1.8191, |
|
"step": 10354 |
|
}, |
|
{ |
|
"epoch": 0.2614782269333904, |
|
"grad_norm": 1.0176483392715454, |
|
"learning_rate": 3.9337993069292366e-05, |
|
"loss": 1.8396, |
|
"step": 10385 |
|
}, |
|
{ |
|
"epoch": 0.26225875895408707, |
|
"grad_norm": 0.9567158222198486, |
|
"learning_rate": 3.926911459260109e-05, |
|
"loss": 1.8617, |
|
"step": 10416 |
|
}, |
|
{ |
|
"epoch": 0.2630392909747838, |
|
"grad_norm": 0.9236714839935303, |
|
"learning_rate": 3.920007509089102e-05, |
|
"loss": 1.8362, |
|
"step": 10447 |
|
}, |
|
{ |
|
"epoch": 0.26381982299548046, |
|
"grad_norm": 0.9428004026412964, |
|
"learning_rate": 3.913087534326357e-05, |
|
"loss": 1.8447, |
|
"step": 10478 |
|
}, |
|
{ |
|
"epoch": 0.2646003550161772, |
|
"grad_norm": 0.941067099571228, |
|
"learning_rate": 3.9061516130628475e-05, |
|
"loss": 1.8329, |
|
"step": 10509 |
|
}, |
|
{ |
|
"epoch": 0.26538088703687385, |
|
"grad_norm": 0.9685031175613403, |
|
"learning_rate": 3.8991998235695025e-05, |
|
"loss": 1.8543, |
|
"step": 10540 |
|
}, |
|
{ |
|
"epoch": 0.2661614190575705, |
|
"grad_norm": 0.9629632234573364, |
|
"learning_rate": 3.8922322442963224e-05, |
|
"loss": 1.8118, |
|
"step": 10571 |
|
}, |
|
{ |
|
"epoch": 0.26694195107826724, |
|
"grad_norm": 0.9536260366439819, |
|
"learning_rate": 3.885248953871491e-05, |
|
"loss": 1.8379, |
|
"step": 10602 |
|
}, |
|
{ |
|
"epoch": 0.2677224830989639, |
|
"grad_norm": 0.9781749248504639, |
|
"learning_rate": 3.8782500311004915e-05, |
|
"loss": 1.8419, |
|
"step": 10633 |
|
}, |
|
{ |
|
"epoch": 0.26850301511966057, |
|
"grad_norm": 0.9313937425613403, |
|
"learning_rate": 3.871235554965218e-05, |
|
"loss": 1.8467, |
|
"step": 10664 |
|
}, |
|
{ |
|
"epoch": 0.2692835471403573, |
|
"grad_norm": 0.9775392413139343, |
|
"learning_rate": 3.864205604623078e-05, |
|
"loss": 1.8209, |
|
"step": 10695 |
|
}, |
|
{ |
|
"epoch": 0.27006407916105396, |
|
"grad_norm": 0.965977132320404, |
|
"learning_rate": 3.857160259406107e-05, |
|
"loss": 1.8558, |
|
"step": 10726 |
|
}, |
|
{ |
|
"epoch": 0.2708446111817507, |
|
"grad_norm": 0.9789881706237793, |
|
"learning_rate": 3.8500995988200674e-05, |
|
"loss": 1.8523, |
|
"step": 10757 |
|
}, |
|
{ |
|
"epoch": 0.27162514320244735, |
|
"grad_norm": 0.9083163738250732, |
|
"learning_rate": 3.843023702543556e-05, |
|
"loss": 1.8274, |
|
"step": 10788 |
|
}, |
|
{ |
|
"epoch": 0.272405675223144, |
|
"grad_norm": 0.9248176217079163, |
|
"learning_rate": 3.8359326504270984e-05, |
|
"loss": 1.8316, |
|
"step": 10819 |
|
}, |
|
{ |
|
"epoch": 0.27318620724384074, |
|
"grad_norm": 0.9702487587928772, |
|
"learning_rate": 3.828826522492255e-05, |
|
"loss": 1.8344, |
|
"step": 10850 |
|
}, |
|
{ |
|
"epoch": 0.2739667392645374, |
|
"grad_norm": 0.9324803948402405, |
|
"learning_rate": 3.821705398930713e-05, |
|
"loss": 1.8544, |
|
"step": 10881 |
|
}, |
|
{ |
|
"epoch": 0.2747472712852341, |
|
"grad_norm": 0.9495794773101807, |
|
"learning_rate": 3.814569360103385e-05, |
|
"loss": 1.8446, |
|
"step": 10912 |
|
}, |
|
{ |
|
"epoch": 0.2755278033059308, |
|
"grad_norm": 0.9734468460083008, |
|
"learning_rate": 3.807418486539499e-05, |
|
"loss": 1.8434, |
|
"step": 10943 |
|
}, |
|
{ |
|
"epoch": 0.27630833532662746, |
|
"grad_norm": 0.9376105666160583, |
|
"learning_rate": 3.80025285893569e-05, |
|
"loss": 1.8339, |
|
"step": 10974 |
|
}, |
|
{ |
|
"epoch": 0.2770888673473242, |
|
"grad_norm": 0.9637653827667236, |
|
"learning_rate": 3.793072558155093e-05, |
|
"loss": 1.8445, |
|
"step": 11005 |
|
}, |
|
{ |
|
"epoch": 0.27786939936802085, |
|
"grad_norm": 0.9537612199783325, |
|
"learning_rate": 3.785877665226426e-05, |
|
"loss": 1.8332, |
|
"step": 11036 |
|
}, |
|
{ |
|
"epoch": 0.2786499313887175, |
|
"grad_norm": 0.9774130582809448, |
|
"learning_rate": 3.778668261343079e-05, |
|
"loss": 1.8269, |
|
"step": 11067 |
|
}, |
|
{ |
|
"epoch": 0.27943046340941424, |
|
"grad_norm": 0.9648153781890869, |
|
"learning_rate": 3.771444427862192e-05, |
|
"loss": 1.8301, |
|
"step": 11098 |
|
}, |
|
{ |
|
"epoch": 0.2802109954301109, |
|
"grad_norm": 0.9317411184310913, |
|
"learning_rate": 3.7642062463037465e-05, |
|
"loss": 1.836, |
|
"step": 11129 |
|
}, |
|
{ |
|
"epoch": 0.2809915274508076, |
|
"grad_norm": 0.9821638464927673, |
|
"learning_rate": 3.7569537983496373e-05, |
|
"loss": 1.846, |
|
"step": 11160 |
|
}, |
|
{ |
|
"epoch": 0.2817720594715043, |
|
"grad_norm": 0.9444936513900757, |
|
"learning_rate": 3.749687165842753e-05, |
|
"loss": 1.8248, |
|
"step": 11191 |
|
}, |
|
{ |
|
"epoch": 0.28255259149220097, |
|
"grad_norm": 0.9421222805976868, |
|
"learning_rate": 3.7424064307860536e-05, |
|
"loss": 1.8521, |
|
"step": 11222 |
|
}, |
|
{ |
|
"epoch": 0.2833331235128977, |
|
"grad_norm": 0.9378920793533325, |
|
"learning_rate": 3.735111675341645e-05, |
|
"loss": 1.8361, |
|
"step": 11253 |
|
}, |
|
{ |
|
"epoch": 0.28411365553359436, |
|
"grad_norm": 0.9363348484039307, |
|
"learning_rate": 3.7278029818298524e-05, |
|
"loss": 1.8276, |
|
"step": 11284 |
|
}, |
|
{ |
|
"epoch": 0.284894187554291, |
|
"grad_norm": 0.9247562885284424, |
|
"learning_rate": 3.720480432728287e-05, |
|
"loss": 1.8354, |
|
"step": 11315 |
|
}, |
|
{ |
|
"epoch": 0.28567471957498775, |
|
"grad_norm": 0.9505523443222046, |
|
"learning_rate": 3.71314411067092e-05, |
|
"loss": 1.807, |
|
"step": 11346 |
|
}, |
|
{ |
|
"epoch": 0.2864552515956844, |
|
"grad_norm": 0.9292653799057007, |
|
"learning_rate": 3.70579409844715e-05, |
|
"loss": 1.8281, |
|
"step": 11377 |
|
}, |
|
{ |
|
"epoch": 0.2872357836163811, |
|
"grad_norm": 0.9625663161277771, |
|
"learning_rate": 3.698430479000865e-05, |
|
"loss": 1.8211, |
|
"step": 11408 |
|
}, |
|
{ |
|
"epoch": 0.2880163156370778, |
|
"grad_norm": 0.9280233979225159, |
|
"learning_rate": 3.691053335429509e-05, |
|
"loss": 1.8177, |
|
"step": 11439 |
|
}, |
|
{ |
|
"epoch": 0.28879684765777447, |
|
"grad_norm": 0.9617031216621399, |
|
"learning_rate": 3.683662750983147e-05, |
|
"loss": 1.831, |
|
"step": 11470 |
|
}, |
|
{ |
|
"epoch": 0.2895773796784712, |
|
"grad_norm": 0.9631845951080322, |
|
"learning_rate": 3.676258809063518e-05, |
|
"loss": 1.8154, |
|
"step": 11501 |
|
}, |
|
{ |
|
"epoch": 0.29035791169916786, |
|
"grad_norm": 0.9200331568717957, |
|
"learning_rate": 3.6688415932231004e-05, |
|
"loss": 1.8452, |
|
"step": 11532 |
|
}, |
|
{ |
|
"epoch": 0.2911384437198645, |
|
"grad_norm": 0.9431037902832031, |
|
"learning_rate": 3.661411187164166e-05, |
|
"loss": 1.8221, |
|
"step": 11563 |
|
}, |
|
{ |
|
"epoch": 0.29191897574056125, |
|
"grad_norm": 0.9307507276535034, |
|
"learning_rate": 3.65396767473784e-05, |
|
"loss": 1.8092, |
|
"step": 11594 |
|
}, |
|
{ |
|
"epoch": 0.2926995077612579, |
|
"grad_norm": 0.9896314144134521, |
|
"learning_rate": 3.6465111399431465e-05, |
|
"loss": 1.826, |
|
"step": 11625 |
|
}, |
|
{ |
|
"epoch": 0.2934800397819546, |
|
"grad_norm": 0.9361921548843384, |
|
"learning_rate": 3.6390416669260674e-05, |
|
"loss": 1.8207, |
|
"step": 11656 |
|
}, |
|
{ |
|
"epoch": 0.2942605718026513, |
|
"grad_norm": 0.9492716789245605, |
|
"learning_rate": 3.63155933997859e-05, |
|
"loss": 1.822, |
|
"step": 11687 |
|
}, |
|
{ |
|
"epoch": 0.29504110382334797, |
|
"grad_norm": 0.9642727375030518, |
|
"learning_rate": 3.624064243537758e-05, |
|
"loss": 1.8165, |
|
"step": 11718 |
|
}, |
|
{ |
|
"epoch": 0.29582163584404464, |
|
"grad_norm": 0.9700178503990173, |
|
"learning_rate": 3.616556462184716e-05, |
|
"loss": 1.8093, |
|
"step": 11749 |
|
}, |
|
{ |
|
"epoch": 0.29660216786474136, |
|
"grad_norm": 0.90894615650177, |
|
"learning_rate": 3.609036080643755e-05, |
|
"loss": 1.8312, |
|
"step": 11780 |
|
}, |
|
{ |
|
"epoch": 0.29738269988543803, |
|
"grad_norm": 0.9441593885421753, |
|
"learning_rate": 3.60150318378136e-05, |
|
"loss": 1.8324, |
|
"step": 11811 |
|
}, |
|
{ |
|
"epoch": 0.29816323190613475, |
|
"grad_norm": 0.9351882338523865, |
|
"learning_rate": 3.5939578566052465e-05, |
|
"loss": 1.8303, |
|
"step": 11842 |
|
}, |
|
{ |
|
"epoch": 0.2989437639268314, |
|
"grad_norm": 0.9438377618789673, |
|
"learning_rate": 3.586400184263408e-05, |
|
"loss": 1.8164, |
|
"step": 11873 |
|
}, |
|
{ |
|
"epoch": 0.2997242959475281, |
|
"grad_norm": 0.9775059819221497, |
|
"learning_rate": 3.578830252043148e-05, |
|
"loss": 1.8146, |
|
"step": 11904 |
|
}, |
|
{ |
|
"epoch": 0.3005048279682248, |
|
"grad_norm": 0.9342683553695679, |
|
"learning_rate": 3.571248145370125e-05, |
|
"loss": 1.8324, |
|
"step": 11935 |
|
}, |
|
{ |
|
"epoch": 0.3012853599889215, |
|
"grad_norm": 0.9539228081703186, |
|
"learning_rate": 3.5636539498073794e-05, |
|
"loss": 1.8112, |
|
"step": 11966 |
|
}, |
|
{ |
|
"epoch": 0.30206589200961814, |
|
"grad_norm": 1.0048060417175293, |
|
"learning_rate": 3.556047751054378e-05, |
|
"loss": 1.8261, |
|
"step": 11997 |
|
}, |
|
{ |
|
"epoch": 0.30284642403031486, |
|
"grad_norm": 0.9269846081733704, |
|
"learning_rate": 3.548429634946039e-05, |
|
"loss": 1.8256, |
|
"step": 12028 |
|
}, |
|
{ |
|
"epoch": 0.30362695605101153, |
|
"grad_norm": 0.937609076499939, |
|
"learning_rate": 3.540799687451768e-05, |
|
"loss": 1.8222, |
|
"step": 12059 |
|
}, |
|
{ |
|
"epoch": 0.30440748807170825, |
|
"grad_norm": 0.9538866281509399, |
|
"learning_rate": 3.533157994674485e-05, |
|
"loss": 1.8214, |
|
"step": 12090 |
|
}, |
|
{ |
|
"epoch": 0.3051880200924049, |
|
"grad_norm": 0.9644604325294495, |
|
"learning_rate": 3.5255046428496546e-05, |
|
"loss": 1.8117, |
|
"step": 12121 |
|
}, |
|
{ |
|
"epoch": 0.3059685521131016, |
|
"grad_norm": 1.0498855113983154, |
|
"learning_rate": 3.517839718344311e-05, |
|
"loss": 1.8249, |
|
"step": 12152 |
|
}, |
|
{ |
|
"epoch": 0.3067490841337983, |
|
"grad_norm": 0.952634334564209, |
|
"learning_rate": 3.510163307656086e-05, |
|
"loss": 1.8263, |
|
"step": 12183 |
|
}, |
|
{ |
|
"epoch": 0.307529616154495, |
|
"grad_norm": 0.9833852052688599, |
|
"learning_rate": 3.5024754974122324e-05, |
|
"loss": 1.8246, |
|
"step": 12214 |
|
}, |
|
{ |
|
"epoch": 0.30831014817519165, |
|
"grad_norm": 0.9421613812446594, |
|
"learning_rate": 3.494776374368643e-05, |
|
"loss": 1.8065, |
|
"step": 12245 |
|
}, |
|
{ |
|
"epoch": 0.30909068019588837, |
|
"grad_norm": 0.9119872450828552, |
|
"learning_rate": 3.4870660254088724e-05, |
|
"loss": 1.7958, |
|
"step": 12276 |
|
}, |
|
{ |
|
"epoch": 0.30987121221658503, |
|
"grad_norm": 0.9541532397270203, |
|
"learning_rate": 3.479344537543164e-05, |
|
"loss": 1.8156, |
|
"step": 12307 |
|
}, |
|
{ |
|
"epoch": 0.31065174423728176, |
|
"grad_norm": 0.9333822131156921, |
|
"learning_rate": 3.4716119979074565e-05, |
|
"loss": 1.8098, |
|
"step": 12338 |
|
}, |
|
{ |
|
"epoch": 0.3114322762579784, |
|
"grad_norm": 0.9501641392707825, |
|
"learning_rate": 3.463868493762412e-05, |
|
"loss": 1.815, |
|
"step": 12369 |
|
}, |
|
{ |
|
"epoch": 0.3122128082786751, |
|
"grad_norm": 0.9692359566688538, |
|
"learning_rate": 3.456114112492418e-05, |
|
"loss": 1.8192, |
|
"step": 12400 |
|
}, |
|
{ |
|
"epoch": 0.3129933402993718, |
|
"grad_norm": 0.9368917942047119, |
|
"learning_rate": 3.4483489416046164e-05, |
|
"loss": 1.8044, |
|
"step": 12431 |
|
}, |
|
{ |
|
"epoch": 0.3137738723200685, |
|
"grad_norm": 0.981076180934906, |
|
"learning_rate": 3.440573068727905e-05, |
|
"loss": 1.806, |
|
"step": 12462 |
|
}, |
|
{ |
|
"epoch": 0.31455440434076515, |
|
"grad_norm": 0.9535355567932129, |
|
"learning_rate": 3.4327865816119495e-05, |
|
"loss": 1.805, |
|
"step": 12493 |
|
}, |
|
{ |
|
"epoch": 0.31533493636146187, |
|
"grad_norm": 0.9825057983398438, |
|
"learning_rate": 3.4249895681262025e-05, |
|
"loss": 1.7878, |
|
"step": 12524 |
|
}, |
|
{ |
|
"epoch": 0.31611546838215854, |
|
"grad_norm": 0.9295104146003723, |
|
"learning_rate": 3.417182116258899e-05, |
|
"loss": 1.8091, |
|
"step": 12555 |
|
}, |
|
{ |
|
"epoch": 0.31689600040285526, |
|
"grad_norm": 0.9550284147262573, |
|
"learning_rate": 3.409364314116074e-05, |
|
"loss": 1.8053, |
|
"step": 12586 |
|
}, |
|
{ |
|
"epoch": 0.3176765324235519, |
|
"grad_norm": 0.9383823871612549, |
|
"learning_rate": 3.401536249920559e-05, |
|
"loss": 1.8034, |
|
"step": 12617 |
|
}, |
|
{ |
|
"epoch": 0.3184570644442486, |
|
"grad_norm": 0.9513770341873169, |
|
"learning_rate": 3.393698012010998e-05, |
|
"loss": 1.8097, |
|
"step": 12648 |
|
}, |
|
{ |
|
"epoch": 0.3192375964649453, |
|
"grad_norm": 0.9214389324188232, |
|
"learning_rate": 3.385849688840839e-05, |
|
"loss": 1.8089, |
|
"step": 12679 |
|
}, |
|
{ |
|
"epoch": 0.320018128485642, |
|
"grad_norm": 0.9370726346969604, |
|
"learning_rate": 3.3779913689773414e-05, |
|
"loss": 1.8022, |
|
"step": 12710 |
|
}, |
|
{ |
|
"epoch": 0.32079866050633865, |
|
"grad_norm": 0.9661527872085571, |
|
"learning_rate": 3.370123141100578e-05, |
|
"loss": 1.8151, |
|
"step": 12741 |
|
}, |
|
{ |
|
"epoch": 0.3215791925270354, |
|
"grad_norm": 0.9569831490516663, |
|
"learning_rate": 3.3622450940024305e-05, |
|
"loss": 1.8283, |
|
"step": 12772 |
|
}, |
|
{ |
|
"epoch": 0.32235972454773204, |
|
"grad_norm": 0.9162306785583496, |
|
"learning_rate": 3.35435731658559e-05, |
|
"loss": 1.8005, |
|
"step": 12803 |
|
}, |
|
{ |
|
"epoch": 0.32314025656842876, |
|
"grad_norm": 0.955159068107605, |
|
"learning_rate": 3.346459897862552e-05, |
|
"loss": 1.7929, |
|
"step": 12834 |
|
}, |
|
{ |
|
"epoch": 0.32392078858912543, |
|
"grad_norm": 0.9484529495239258, |
|
"learning_rate": 3.338552926954613e-05, |
|
"loss": 1.8355, |
|
"step": 12865 |
|
}, |
|
{ |
|
"epoch": 0.3247013206098221, |
|
"grad_norm": 0.9199361205101013, |
|
"learning_rate": 3.330636493090868e-05, |
|
"loss": 1.8096, |
|
"step": 12896 |
|
}, |
|
{ |
|
"epoch": 0.3254818526305188, |
|
"grad_norm": 0.928848147392273, |
|
"learning_rate": 3.322710685607193e-05, |
|
"loss": 1.7896, |
|
"step": 12927 |
|
}, |
|
{ |
|
"epoch": 0.3262623846512155, |
|
"grad_norm": 0.9141325354576111, |
|
"learning_rate": 3.314775593945251e-05, |
|
"loss": 1.8113, |
|
"step": 12958 |
|
}, |
|
{ |
|
"epoch": 0.32704291667191215, |
|
"grad_norm": 0.97685307264328, |
|
"learning_rate": 3.3068313076514714e-05, |
|
"loss": 1.8241, |
|
"step": 12989 |
|
}, |
|
{ |
|
"epoch": 0.3278234486926089, |
|
"grad_norm": 0.9171717166900635, |
|
"learning_rate": 3.298877916376047e-05, |
|
"loss": 1.7953, |
|
"step": 13020 |
|
}, |
|
{ |
|
"epoch": 0.32860398071330554, |
|
"grad_norm": 0.9607007503509521, |
|
"learning_rate": 3.290915509871915e-05, |
|
"loss": 1.819, |
|
"step": 13051 |
|
}, |
|
{ |
|
"epoch": 0.32938451273400227, |
|
"grad_norm": 0.9524408578872681, |
|
"learning_rate": 3.282944177993753e-05, |
|
"loss": 1.8133, |
|
"step": 13082 |
|
}, |
|
{ |
|
"epoch": 0.33016504475469893, |
|
"grad_norm": 0.9388900399208069, |
|
"learning_rate": 3.274964010696957e-05, |
|
"loss": 1.8103, |
|
"step": 13113 |
|
}, |
|
{ |
|
"epoch": 0.3309455767753956, |
|
"grad_norm": 0.8972403407096863, |
|
"learning_rate": 3.266975098036629e-05, |
|
"loss": 1.7997, |
|
"step": 13144 |
|
}, |
|
{ |
|
"epoch": 0.3317261087960923, |
|
"grad_norm": 0.9373508095741272, |
|
"learning_rate": 3.258977530166562e-05, |
|
"loss": 1.8038, |
|
"step": 13175 |
|
}, |
|
{ |
|
"epoch": 0.332506640816789, |
|
"grad_norm": 0.9278662204742432, |
|
"learning_rate": 3.250971397338227e-05, |
|
"loss": 1.803, |
|
"step": 13206 |
|
}, |
|
{ |
|
"epoch": 0.33328717283748566, |
|
"grad_norm": 0.9486290812492371, |
|
"learning_rate": 3.2429567898997404e-05, |
|
"loss": 1.8114, |
|
"step": 13237 |
|
}, |
|
{ |
|
"epoch": 0.3340677048581824, |
|
"grad_norm": 0.9786876440048218, |
|
"learning_rate": 3.234933798294859e-05, |
|
"loss": 1.7902, |
|
"step": 13268 |
|
}, |
|
{ |
|
"epoch": 0.33484823687887905, |
|
"grad_norm": 0.9001107215881348, |
|
"learning_rate": 3.2269025130619535e-05, |
|
"loss": 1.8108, |
|
"step": 13299 |
|
}, |
|
{ |
|
"epoch": 0.33562876889957577, |
|
"grad_norm": 0.9080937504768372, |
|
"learning_rate": 3.218863024832985e-05, |
|
"loss": 1.7951, |
|
"step": 13330 |
|
}, |
|
{ |
|
"epoch": 0.33640930092027244, |
|
"grad_norm": 0.93299800157547, |
|
"learning_rate": 3.2108154243324864e-05, |
|
"loss": 1.8129, |
|
"step": 13361 |
|
}, |
|
{ |
|
"epoch": 0.3371898329409691, |
|
"grad_norm": 0.9452754855155945, |
|
"learning_rate": 3.2027598023765345e-05, |
|
"loss": 1.8077, |
|
"step": 13392 |
|
}, |
|
{ |
|
"epoch": 0.3379703649616658, |
|
"grad_norm": 0.9515936970710754, |
|
"learning_rate": 3.194696249871729e-05, |
|
"loss": 1.7961, |
|
"step": 13423 |
|
}, |
|
{ |
|
"epoch": 0.3387508969823625, |
|
"grad_norm": 0.9395990371704102, |
|
"learning_rate": 3.186624857814164e-05, |
|
"loss": 1.7982, |
|
"step": 13454 |
|
}, |
|
{ |
|
"epoch": 0.33953142900305916, |
|
"grad_norm": 0.9647415280342102, |
|
"learning_rate": 3.178545717288401e-05, |
|
"loss": 1.7997, |
|
"step": 13485 |
|
}, |
|
{ |
|
"epoch": 0.3403119610237559, |
|
"grad_norm": 0.9437438249588013, |
|
"learning_rate": 3.170458919466444e-05, |
|
"loss": 1.8069, |
|
"step": 13516 |
|
}, |
|
{ |
|
"epoch": 0.34109249304445255, |
|
"grad_norm": 0.9350516200065613, |
|
"learning_rate": 3.1623645556067063e-05, |
|
"loss": 1.8024, |
|
"step": 13547 |
|
}, |
|
{ |
|
"epoch": 0.34187302506514927, |
|
"grad_norm": 0.9213714599609375, |
|
"learning_rate": 3.154262717052985e-05, |
|
"loss": 1.7963, |
|
"step": 13578 |
|
}, |
|
{ |
|
"epoch": 0.34265355708584594, |
|
"grad_norm": 0.9247968792915344, |
|
"learning_rate": 3.146153495233426e-05, |
|
"loss": 1.8172, |
|
"step": 13609 |
|
}, |
|
{ |
|
"epoch": 0.3434340891065426, |
|
"grad_norm": 0.8953402638435364, |
|
"learning_rate": 3.1380369816594944e-05, |
|
"loss": 1.8004, |
|
"step": 13640 |
|
}, |
|
{ |
|
"epoch": 0.34421462112723933, |
|
"grad_norm": 0.9214624762535095, |
|
"learning_rate": 3.129913267924946e-05, |
|
"loss": 1.7862, |
|
"step": 13671 |
|
}, |
|
{ |
|
"epoch": 0.344995153147936, |
|
"grad_norm": 0.946559727191925, |
|
"learning_rate": 3.121782445704782e-05, |
|
"loss": 1.7973, |
|
"step": 13702 |
|
}, |
|
{ |
|
"epoch": 0.34577568516863266, |
|
"grad_norm": 0.9344651103019714, |
|
"learning_rate": 3.11364460675423e-05, |
|
"loss": 1.7992, |
|
"step": 13733 |
|
}, |
|
{ |
|
"epoch": 0.3465562171893294, |
|
"grad_norm": 0.9140195846557617, |
|
"learning_rate": 3.1054998429076934e-05, |
|
"loss": 1.7846, |
|
"step": 13764 |
|
}, |
|
{ |
|
"epoch": 0.34733674921002605, |
|
"grad_norm": 0.9403822422027588, |
|
"learning_rate": 3.097348246077728e-05, |
|
"loss": 1.7814, |
|
"step": 13795 |
|
}, |
|
{ |
|
"epoch": 0.3481172812307228, |
|
"grad_norm": 0.954587996006012, |
|
"learning_rate": 3.0891899082539924e-05, |
|
"loss": 1.7864, |
|
"step": 13826 |
|
}, |
|
{ |
|
"epoch": 0.34889781325141944, |
|
"grad_norm": 0.9238544702529907, |
|
"learning_rate": 3.0810249215022233e-05, |
|
"loss": 1.7929, |
|
"step": 13857 |
|
}, |
|
{ |
|
"epoch": 0.3496783452721161, |
|
"grad_norm": 0.8994883894920349, |
|
"learning_rate": 3.0728533779631865e-05, |
|
"loss": 1.805, |
|
"step": 13888 |
|
}, |
|
{ |
|
"epoch": 0.35045887729281283, |
|
"grad_norm": 0.934203028678894, |
|
"learning_rate": 3.064675369851637e-05, |
|
"loss": 1.8046, |
|
"step": 13919 |
|
}, |
|
{ |
|
"epoch": 0.3512394093135095, |
|
"grad_norm": 0.9044675230979919, |
|
"learning_rate": 3.056490989455289e-05, |
|
"loss": 1.7866, |
|
"step": 13950 |
|
}, |
|
{ |
|
"epoch": 0.35201994133420617, |
|
"grad_norm": 0.915847897529602, |
|
"learning_rate": 3.0483003291337596e-05, |
|
"loss": 1.7861, |
|
"step": 13981 |
|
}, |
|
{ |
|
"epoch": 0.3528004733549029, |
|
"grad_norm": 0.9346082210540771, |
|
"learning_rate": 3.040103481317539e-05, |
|
"loss": 1.7891, |
|
"step": 14012 |
|
}, |
|
{ |
|
"epoch": 0.35358100537559956, |
|
"grad_norm": 0.9164626598358154, |
|
"learning_rate": 3.03190053850694e-05, |
|
"loss": 1.8021, |
|
"step": 14043 |
|
}, |
|
{ |
|
"epoch": 0.3543615373962962, |
|
"grad_norm": 0.9445552229881287, |
|
"learning_rate": 3.0236915932710573e-05, |
|
"loss": 1.7877, |
|
"step": 14074 |
|
}, |
|
{ |
|
"epoch": 0.35514206941699294, |
|
"grad_norm": 0.9604120254516602, |
|
"learning_rate": 3.0154767382467232e-05, |
|
"loss": 1.8087, |
|
"step": 14105 |
|
}, |
|
{ |
|
"epoch": 0.3559226014376896, |
|
"grad_norm": 0.9186177849769592, |
|
"learning_rate": 3.0072560661374582e-05, |
|
"loss": 1.7802, |
|
"step": 14136 |
|
}, |
|
{ |
|
"epoch": 0.35670313345838633, |
|
"grad_norm": 0.9228742122650146, |
|
"learning_rate": 2.999029669712431e-05, |
|
"loss": 1.8091, |
|
"step": 14167 |
|
}, |
|
{ |
|
"epoch": 0.357483665479083, |
|
"grad_norm": 0.9045746922492981, |
|
"learning_rate": 2.990797641805408e-05, |
|
"loss": 1.8013, |
|
"step": 14198 |
|
}, |
|
{ |
|
"epoch": 0.35826419749977967, |
|
"grad_norm": 0.9483916759490967, |
|
"learning_rate": 2.982560075313704e-05, |
|
"loss": 1.8044, |
|
"step": 14229 |
|
}, |
|
{ |
|
"epoch": 0.3590447295204764, |
|
"grad_norm": 0.9580939412117004, |
|
"learning_rate": 2.9743170631971368e-05, |
|
"loss": 1.8037, |
|
"step": 14260 |
|
}, |
|
{ |
|
"epoch": 0.35982526154117306, |
|
"grad_norm": 0.950960099697113, |
|
"learning_rate": 2.9660686984769792e-05, |
|
"loss": 1.8082, |
|
"step": 14291 |
|
}, |
|
{ |
|
"epoch": 0.3606057935618697, |
|
"grad_norm": 0.9356971383094788, |
|
"learning_rate": 2.9578150742349047e-05, |
|
"loss": 1.8049, |
|
"step": 14322 |
|
}, |
|
{ |
|
"epoch": 0.36138632558256645, |
|
"grad_norm": 0.9099254012107849, |
|
"learning_rate": 2.949556283611942e-05, |
|
"loss": 1.7991, |
|
"step": 14353 |
|
}, |
|
{ |
|
"epoch": 0.3621668576032631, |
|
"grad_norm": 0.9190629720687866, |
|
"learning_rate": 2.9412924198074206e-05, |
|
"loss": 1.8018, |
|
"step": 14384 |
|
}, |
|
{ |
|
"epoch": 0.36294738962395984, |
|
"grad_norm": 0.90566486120224, |
|
"learning_rate": 2.9330235760779208e-05, |
|
"loss": 1.789, |
|
"step": 14415 |
|
}, |
|
{ |
|
"epoch": 0.3637279216446565, |
|
"grad_norm": 0.970696747303009, |
|
"learning_rate": 2.9247498457362188e-05, |
|
"loss": 1.7851, |
|
"step": 14446 |
|
}, |
|
{ |
|
"epoch": 0.36450845366535317, |
|
"grad_norm": 0.9313214421272278, |
|
"learning_rate": 2.9164713221502373e-05, |
|
"loss": 1.7893, |
|
"step": 14477 |
|
}, |
|
{ |
|
"epoch": 0.3652889856860499, |
|
"grad_norm": 1.2887455224990845, |
|
"learning_rate": 2.9081880987419912e-05, |
|
"loss": 1.7836, |
|
"step": 14508 |
|
}, |
|
{ |
|
"epoch": 0.36606951770674656, |
|
"grad_norm": 0.922982394695282, |
|
"learning_rate": 2.8999002689865296e-05, |
|
"loss": 1.7676, |
|
"step": 14539 |
|
}, |
|
{ |
|
"epoch": 0.36685004972744323, |
|
"grad_norm": 0.9466812014579773, |
|
"learning_rate": 2.8916079264108852e-05, |
|
"loss": 1.7831, |
|
"step": 14570 |
|
}, |
|
{ |
|
"epoch": 0.36763058174813995, |
|
"grad_norm": 0.9107967615127563, |
|
"learning_rate": 2.883311164593017e-05, |
|
"loss": 1.7887, |
|
"step": 14601 |
|
}, |
|
{ |
|
"epoch": 0.3684111137688366, |
|
"grad_norm": 0.8859044313430786, |
|
"learning_rate": 2.875010077160754e-05, |
|
"loss": 1.7823, |
|
"step": 14632 |
|
}, |
|
{ |
|
"epoch": 0.36919164578953334, |
|
"grad_norm": 0.9182350039482117, |
|
"learning_rate": 2.866704757790741e-05, |
|
"loss": 1.7816, |
|
"step": 14663 |
|
}, |
|
{ |
|
"epoch": 0.36997217781023, |
|
"grad_norm": 1.1406787633895874, |
|
"learning_rate": 2.858395300207376e-05, |
|
"loss": 1.7962, |
|
"step": 14694 |
|
}, |
|
{ |
|
"epoch": 0.3707527098309267, |
|
"grad_norm": 0.9399011135101318, |
|
"learning_rate": 2.8500817981817607e-05, |
|
"loss": 1.7874, |
|
"step": 14725 |
|
}, |
|
{ |
|
"epoch": 0.3715332418516234, |
|
"grad_norm": 0.9223036766052246, |
|
"learning_rate": 2.8417643455306336e-05, |
|
"loss": 1.7661, |
|
"step": 14756 |
|
}, |
|
{ |
|
"epoch": 0.37231377387232006, |
|
"grad_norm": 0.9224306344985962, |
|
"learning_rate": 2.8334430361153185e-05, |
|
"loss": 1.7799, |
|
"step": 14787 |
|
}, |
|
{ |
|
"epoch": 0.37309430589301673, |
|
"grad_norm": 0.9353799819946289, |
|
"learning_rate": 2.8251179638406612e-05, |
|
"loss": 1.795, |
|
"step": 14818 |
|
}, |
|
{ |
|
"epoch": 0.37387483791371345, |
|
"grad_norm": 0.9208987355232239, |
|
"learning_rate": 2.8167892226539704e-05, |
|
"loss": 1.7967, |
|
"step": 14849 |
|
}, |
|
{ |
|
"epoch": 0.3746553699344101, |
|
"grad_norm": 0.9143754243850708, |
|
"learning_rate": 2.8084569065439588e-05, |
|
"loss": 1.7871, |
|
"step": 14880 |
|
}, |
|
{ |
|
"epoch": 0.37543590195510684, |
|
"grad_norm": 0.9162681698799133, |
|
"learning_rate": 2.8001211095396807e-05, |
|
"loss": 1.7791, |
|
"step": 14911 |
|
}, |
|
{ |
|
"epoch": 0.3762164339758035, |
|
"grad_norm": 0.9125809073448181, |
|
"learning_rate": 2.791781925709473e-05, |
|
"loss": 1.7847, |
|
"step": 14942 |
|
}, |
|
{ |
|
"epoch": 0.3769969659965002, |
|
"grad_norm": 0.9262179732322693, |
|
"learning_rate": 2.7834394491598908e-05, |
|
"loss": 1.789, |
|
"step": 14973 |
|
}, |
|
{ |
|
"epoch": 0.3777774980171969, |
|
"grad_norm": 0.9118053317070007, |
|
"learning_rate": 2.7750937740346485e-05, |
|
"loss": 1.7906, |
|
"step": 15004 |
|
}, |
|
{ |
|
"epoch": 0.37855803003789357, |
|
"grad_norm": 0.9322589635848999, |
|
"learning_rate": 2.7667449945135564e-05, |
|
"loss": 1.7811, |
|
"step": 15035 |
|
}, |
|
{ |
|
"epoch": 0.37933856205859023, |
|
"grad_norm": 0.9277715682983398, |
|
"learning_rate": 2.7583932048114557e-05, |
|
"loss": 1.7887, |
|
"step": 15066 |
|
}, |
|
{ |
|
"epoch": 0.38011909407928696, |
|
"grad_norm": 0.9804796576499939, |
|
"learning_rate": 2.7500384991771587e-05, |
|
"loss": 1.7945, |
|
"step": 15097 |
|
}, |
|
{ |
|
"epoch": 0.3808996260999836, |
|
"grad_norm": 0.9327893853187561, |
|
"learning_rate": 2.7416809718923825e-05, |
|
"loss": 1.785, |
|
"step": 15128 |
|
}, |
|
{ |
|
"epoch": 0.38168015812068035, |
|
"grad_norm": 0.9126656651496887, |
|
"learning_rate": 2.7333207172706864e-05, |
|
"loss": 1.7782, |
|
"step": 15159 |
|
}, |
|
{ |
|
"epoch": 0.382460690141377, |
|
"grad_norm": 0.9476912617683411, |
|
"learning_rate": 2.7249578296564088e-05, |
|
"loss": 1.7906, |
|
"step": 15190 |
|
}, |
|
{ |
|
"epoch": 0.3832412221620737, |
|
"grad_norm": 0.9388076663017273, |
|
"learning_rate": 2.7165924034235973e-05, |
|
"loss": 1.7993, |
|
"step": 15221 |
|
}, |
|
{ |
|
"epoch": 0.3840217541827704, |
|
"grad_norm": 0.9265156984329224, |
|
"learning_rate": 2.708224532974953e-05, |
|
"loss": 1.7624, |
|
"step": 15252 |
|
}, |
|
{ |
|
"epoch": 0.38480228620346707, |
|
"grad_norm": 0.9261599183082581, |
|
"learning_rate": 2.6998543127407538e-05, |
|
"loss": 1.7774, |
|
"step": 15283 |
|
}, |
|
{ |
|
"epoch": 0.38558281822416374, |
|
"grad_norm": 0.9569805860519409, |
|
"learning_rate": 2.6914818371777988e-05, |
|
"loss": 1.7695, |
|
"step": 15314 |
|
}, |
|
{ |
|
"epoch": 0.38636335024486046, |
|
"grad_norm": 0.8934130668640137, |
|
"learning_rate": 2.6831072007683373e-05, |
|
"loss": 1.7746, |
|
"step": 15345 |
|
}, |
|
{ |
|
"epoch": 0.3871438822655571, |
|
"grad_norm": 0.9160299301147461, |
|
"learning_rate": 2.6747304980190018e-05, |
|
"loss": 1.7698, |
|
"step": 15376 |
|
}, |
|
{ |
|
"epoch": 0.38792441428625385, |
|
"grad_norm": 0.9243568181991577, |
|
"learning_rate": 2.6663518234597453e-05, |
|
"loss": 1.7823, |
|
"step": 15407 |
|
}, |
|
{ |
|
"epoch": 0.3887049463069505, |
|
"grad_norm": 0.9243864417076111, |
|
"learning_rate": 2.6579712716427696e-05, |
|
"loss": 1.7819, |
|
"step": 15438 |
|
}, |
|
{ |
|
"epoch": 0.3894854783276472, |
|
"grad_norm": 0.9425201416015625, |
|
"learning_rate": 2.6495889371414652e-05, |
|
"loss": 1.7808, |
|
"step": 15469 |
|
}, |
|
{ |
|
"epoch": 0.3902660103483439, |
|
"grad_norm": 0.944649875164032, |
|
"learning_rate": 2.6412049145493367e-05, |
|
"loss": 1.781, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 0.3910465423690406, |
|
"grad_norm": 0.9222155809402466, |
|
"learning_rate": 2.632819298478939e-05, |
|
"loss": 1.7818, |
|
"step": 15531 |
|
}, |
|
{ |
|
"epoch": 0.39182707438973724, |
|
"grad_norm": 0.9028238654136658, |
|
"learning_rate": 2.6244321835608105e-05, |
|
"loss": 1.7622, |
|
"step": 15562 |
|
}, |
|
{ |
|
"epoch": 0.39260760641043396, |
|
"grad_norm": 0.9092904329299927, |
|
"learning_rate": 2.6160436644424024e-05, |
|
"loss": 1.7829, |
|
"step": 15593 |
|
}, |
|
{ |
|
"epoch": 0.39338813843113063, |
|
"grad_norm": 0.9831812381744385, |
|
"learning_rate": 2.6076538357870133e-05, |
|
"loss": 1.7927, |
|
"step": 15624 |
|
}, |
|
{ |
|
"epoch": 0.39416867045182735, |
|
"grad_norm": 0.9344479441642761, |
|
"learning_rate": 2.5992627922727196e-05, |
|
"loss": 1.7725, |
|
"step": 15655 |
|
}, |
|
{ |
|
"epoch": 0.394949202472524, |
|
"grad_norm": 0.9424457550048828, |
|
"learning_rate": 2.5908706285913066e-05, |
|
"loss": 1.77, |
|
"step": 15686 |
|
}, |
|
{ |
|
"epoch": 0.3957297344932207, |
|
"grad_norm": 0.9403957724571228, |
|
"learning_rate": 2.5824774394472008e-05, |
|
"loss": 1.7822, |
|
"step": 15717 |
|
}, |
|
{ |
|
"epoch": 0.3965102665139174, |
|
"grad_norm": 0.9497373104095459, |
|
"learning_rate": 2.5740833195563996e-05, |
|
"loss": 1.7874, |
|
"step": 15748 |
|
}, |
|
{ |
|
"epoch": 0.3972907985346141, |
|
"grad_norm": 0.9396623373031616, |
|
"learning_rate": 2.5656883636454067e-05, |
|
"loss": 1.7634, |
|
"step": 15779 |
|
}, |
|
{ |
|
"epoch": 0.39807133055531074, |
|
"grad_norm": 0.9426395297050476, |
|
"learning_rate": 2.557292666450159e-05, |
|
"loss": 1.7774, |
|
"step": 15810 |
|
}, |
|
{ |
|
"epoch": 0.39885186257600747, |
|
"grad_norm": 0.9340534806251526, |
|
"learning_rate": 2.5488963227149566e-05, |
|
"loss": 1.7619, |
|
"step": 15841 |
|
}, |
|
{ |
|
"epoch": 0.39963239459670413, |
|
"grad_norm": 0.9206343293190002, |
|
"learning_rate": 2.5404994271913983e-05, |
|
"loss": 1.7817, |
|
"step": 15872 |
|
}, |
|
{ |
|
"epoch": 0.40041292661740086, |
|
"grad_norm": 0.9418168067932129, |
|
"learning_rate": 2.5321020746373085e-05, |
|
"loss": 1.7788, |
|
"step": 15903 |
|
}, |
|
{ |
|
"epoch": 0.4011934586380975, |
|
"grad_norm": 0.9084994792938232, |
|
"learning_rate": 2.52370435981567e-05, |
|
"loss": 1.7784, |
|
"step": 15934 |
|
}, |
|
{ |
|
"epoch": 0.4019739906587942, |
|
"grad_norm": 0.9424017667770386, |
|
"learning_rate": 2.5153063774935533e-05, |
|
"loss": 1.7744, |
|
"step": 15965 |
|
}, |
|
{ |
|
"epoch": 0.4027545226794909, |
|
"grad_norm": 0.9338597655296326, |
|
"learning_rate": 2.506908222441045e-05, |
|
"loss": 1.7533, |
|
"step": 15996 |
|
}, |
|
{ |
|
"epoch": 0.4035350547001876, |
|
"grad_norm": 0.9237053394317627, |
|
"learning_rate": 2.498509989430187e-05, |
|
"loss": 1.7691, |
|
"step": 16027 |
|
}, |
|
{ |
|
"epoch": 0.40431558672088425, |
|
"grad_norm": 0.9271100759506226, |
|
"learning_rate": 2.4901117732338958e-05, |
|
"loss": 1.79, |
|
"step": 16058 |
|
}, |
|
{ |
|
"epoch": 0.40509611874158097, |
|
"grad_norm": 0.9305247068405151, |
|
"learning_rate": 2.481713668624899e-05, |
|
"loss": 1.7786, |
|
"step": 16089 |
|
}, |
|
{ |
|
"epoch": 0.40587665076227764, |
|
"grad_norm": 0.9237347841262817, |
|
"learning_rate": 2.4733157703746663e-05, |
|
"loss": 1.7951, |
|
"step": 16120 |
|
}, |
|
{ |
|
"epoch": 0.40665718278297436, |
|
"grad_norm": 0.8985100388526917, |
|
"learning_rate": 2.4649181732523392e-05, |
|
"loss": 1.7699, |
|
"step": 16151 |
|
}, |
|
{ |
|
"epoch": 0.407437714803671, |
|
"grad_norm": 0.9226688742637634, |
|
"learning_rate": 2.4565209720236582e-05, |
|
"loss": 1.769, |
|
"step": 16182 |
|
}, |
|
{ |
|
"epoch": 0.4082182468243677, |
|
"grad_norm": 0.9165879487991333, |
|
"learning_rate": 2.4481242614498975e-05, |
|
"loss": 1.7712, |
|
"step": 16213 |
|
}, |
|
{ |
|
"epoch": 0.4089987788450644, |
|
"grad_norm": 0.949297308921814, |
|
"learning_rate": 2.439728136286796e-05, |
|
"loss": 1.7709, |
|
"step": 16244 |
|
}, |
|
{ |
|
"epoch": 0.4097793108657611, |
|
"grad_norm": 0.930533230304718, |
|
"learning_rate": 2.4313326912834852e-05, |
|
"loss": 1.7512, |
|
"step": 16275 |
|
}, |
|
{ |
|
"epoch": 0.41055984288645775, |
|
"grad_norm": 0.926105797290802, |
|
"learning_rate": 2.4229380211814206e-05, |
|
"loss": 1.7769, |
|
"step": 16306 |
|
}, |
|
{ |
|
"epoch": 0.41134037490715447, |
|
"grad_norm": 0.9282335638999939, |
|
"learning_rate": 2.4145442207133124e-05, |
|
"loss": 1.7624, |
|
"step": 16337 |
|
}, |
|
{ |
|
"epoch": 0.41212090692785114, |
|
"grad_norm": 0.9377603530883789, |
|
"learning_rate": 2.406151384602059e-05, |
|
"loss": 1.7756, |
|
"step": 16368 |
|
}, |
|
{ |
|
"epoch": 0.4129014389485478, |
|
"grad_norm": 0.9203463792800903, |
|
"learning_rate": 2.3977596075596747e-05, |
|
"loss": 1.7839, |
|
"step": 16399 |
|
}, |
|
{ |
|
"epoch": 0.41368197096924453, |
|
"grad_norm": 0.9060366153717041, |
|
"learning_rate": 2.3893689842862223e-05, |
|
"loss": 1.7667, |
|
"step": 16430 |
|
}, |
|
{ |
|
"epoch": 0.4144625029899412, |
|
"grad_norm": 0.9065908193588257, |
|
"learning_rate": 2.3809796094687475e-05, |
|
"loss": 1.7691, |
|
"step": 16461 |
|
}, |
|
{ |
|
"epoch": 0.4152430350106379, |
|
"grad_norm": 0.9281785488128662, |
|
"learning_rate": 2.372591577780202e-05, |
|
"loss": 1.7573, |
|
"step": 16492 |
|
}, |
|
{ |
|
"epoch": 0.4160235670313346, |
|
"grad_norm": 0.880733847618103, |
|
"learning_rate": 2.3642049838783838e-05, |
|
"loss": 1.7697, |
|
"step": 16523 |
|
}, |
|
{ |
|
"epoch": 0.41680409905203125, |
|
"grad_norm": 0.9289157390594482, |
|
"learning_rate": 2.3558199224048666e-05, |
|
"loss": 1.7604, |
|
"step": 16554 |
|
}, |
|
{ |
|
"epoch": 0.417584631072728, |
|
"grad_norm": 0.9380747079849243, |
|
"learning_rate": 2.347436487983929e-05, |
|
"loss": 1.7686, |
|
"step": 16585 |
|
}, |
|
{ |
|
"epoch": 0.41836516309342464, |
|
"grad_norm": 0.9439517855644226, |
|
"learning_rate": 2.3390547752214888e-05, |
|
"loss": 1.7786, |
|
"step": 16616 |
|
}, |
|
{ |
|
"epoch": 0.4191456951141213, |
|
"grad_norm": 0.9420905113220215, |
|
"learning_rate": 2.330674878704035e-05, |
|
"loss": 1.762, |
|
"step": 16647 |
|
}, |
|
{ |
|
"epoch": 0.41992622713481803, |
|
"grad_norm": 0.9132199883460999, |
|
"learning_rate": 2.322296892997561e-05, |
|
"loss": 1.776, |
|
"step": 16678 |
|
}, |
|
{ |
|
"epoch": 0.4207067591555147, |
|
"grad_norm": 0.9286783933639526, |
|
"learning_rate": 2.313920912646497e-05, |
|
"loss": 1.7564, |
|
"step": 16709 |
|
}, |
|
{ |
|
"epoch": 0.4214872911762114, |
|
"grad_norm": 0.9031944870948792, |
|
"learning_rate": 2.305547032172643e-05, |
|
"loss": 1.75, |
|
"step": 16740 |
|
}, |
|
{ |
|
"epoch": 0.4222678231969081, |
|
"grad_norm": 0.9604383111000061, |
|
"learning_rate": 2.2971753460741014e-05, |
|
"loss": 1.7821, |
|
"step": 16771 |
|
}, |
|
{ |
|
"epoch": 0.42304835521760475, |
|
"grad_norm": 0.9145975112915039, |
|
"learning_rate": 2.288805948824212e-05, |
|
"loss": 1.7604, |
|
"step": 16802 |
|
}, |
|
{ |
|
"epoch": 0.4238288872383015, |
|
"grad_norm": 0.9214676022529602, |
|
"learning_rate": 2.2804389348704858e-05, |
|
"loss": 1.7553, |
|
"step": 16833 |
|
}, |
|
{ |
|
"epoch": 0.42460941925899814, |
|
"grad_norm": 0.9024298191070557, |
|
"learning_rate": 2.2720743986335374e-05, |
|
"loss": 1.7652, |
|
"step": 16864 |
|
}, |
|
{ |
|
"epoch": 0.4253899512796948, |
|
"grad_norm": 0.8926112055778503, |
|
"learning_rate": 2.2637124345060233e-05, |
|
"loss": 1.7583, |
|
"step": 16895 |
|
}, |
|
{ |
|
"epoch": 0.42617048330039153, |
|
"grad_norm": 0.9225366711616516, |
|
"learning_rate": 2.2553531368515695e-05, |
|
"loss": 1.7607, |
|
"step": 16926 |
|
}, |
|
{ |
|
"epoch": 0.4269510153210882, |
|
"grad_norm": 0.9340829253196716, |
|
"learning_rate": 2.2469966000037144e-05, |
|
"loss": 1.7659, |
|
"step": 16957 |
|
}, |
|
{ |
|
"epoch": 0.4277315473417849, |
|
"grad_norm": 0.9964073896408081, |
|
"learning_rate": 2.2386429182648417e-05, |
|
"loss": 1.7414, |
|
"step": 16988 |
|
}, |
|
{ |
|
"epoch": 0.4285120793624816, |
|
"grad_norm": 0.9147719144821167, |
|
"learning_rate": 2.230292185905114e-05, |
|
"loss": 1.764, |
|
"step": 17019 |
|
}, |
|
{ |
|
"epoch": 0.42929261138317826, |
|
"grad_norm": 0.9175845384597778, |
|
"learning_rate": 2.2219444971614116e-05, |
|
"loss": 1.7623, |
|
"step": 17050 |
|
}, |
|
{ |
|
"epoch": 0.430073143403875, |
|
"grad_norm": 0.9041738510131836, |
|
"learning_rate": 2.2135999462362655e-05, |
|
"loss": 1.7448, |
|
"step": 17081 |
|
}, |
|
{ |
|
"epoch": 0.43085367542457165, |
|
"grad_norm": 0.9334216713905334, |
|
"learning_rate": 2.2052586272968003e-05, |
|
"loss": 1.7548, |
|
"step": 17112 |
|
}, |
|
{ |
|
"epoch": 0.4316342074452683, |
|
"grad_norm": 0.9379672408103943, |
|
"learning_rate": 2.196920634473666e-05, |
|
"loss": 1.7703, |
|
"step": 17143 |
|
}, |
|
{ |
|
"epoch": 0.43241473946596504, |
|
"grad_norm": 0.9310106039047241, |
|
"learning_rate": 2.1885860618599787e-05, |
|
"loss": 1.7567, |
|
"step": 17174 |
|
}, |
|
{ |
|
"epoch": 0.4331952714866617, |
|
"grad_norm": 0.936827540397644, |
|
"learning_rate": 2.1802550035102577e-05, |
|
"loss": 1.7572, |
|
"step": 17205 |
|
}, |
|
{ |
|
"epoch": 0.4339758035073584, |
|
"grad_norm": 0.954757571220398, |
|
"learning_rate": 2.171927553439363e-05, |
|
"loss": 1.7773, |
|
"step": 17236 |
|
}, |
|
{ |
|
"epoch": 0.4347563355280551, |
|
"grad_norm": 0.9472478032112122, |
|
"learning_rate": 2.1636038056214376e-05, |
|
"loss": 1.7614, |
|
"step": 17267 |
|
}, |
|
{ |
|
"epoch": 0.43553686754875176, |
|
"grad_norm": 0.9342318773269653, |
|
"learning_rate": 2.155283853988844e-05, |
|
"loss": 1.7561, |
|
"step": 17298 |
|
}, |
|
{ |
|
"epoch": 0.4363173995694485, |
|
"grad_norm": 0.9363663792610168, |
|
"learning_rate": 2.146967792431106e-05, |
|
"loss": 1.7703, |
|
"step": 17329 |
|
}, |
|
{ |
|
"epoch": 0.43709793159014515, |
|
"grad_norm": 0.9496386647224426, |
|
"learning_rate": 2.138655714793849e-05, |
|
"loss": 1.7565, |
|
"step": 17360 |
|
}, |
|
{ |
|
"epoch": 0.4378784636108418, |
|
"grad_norm": 0.9076322913169861, |
|
"learning_rate": 2.1303477148777367e-05, |
|
"loss": 1.738, |
|
"step": 17391 |
|
}, |
|
{ |
|
"epoch": 0.43865899563153854, |
|
"grad_norm": 0.9229772686958313, |
|
"learning_rate": 2.122043886437421e-05, |
|
"loss": 1.7455, |
|
"step": 17422 |
|
}, |
|
{ |
|
"epoch": 0.4394395276522352, |
|
"grad_norm": 0.9290786385536194, |
|
"learning_rate": 2.1137443231804765e-05, |
|
"loss": 1.7635, |
|
"step": 17453 |
|
}, |
|
{ |
|
"epoch": 0.44022005967293193, |
|
"grad_norm": 0.9575942754745483, |
|
"learning_rate": 2.105449118766347e-05, |
|
"loss": 1.7763, |
|
"step": 17484 |
|
}, |
|
{ |
|
"epoch": 0.4410005916936286, |
|
"grad_norm": 0.9625800848007202, |
|
"learning_rate": 2.097158366805287e-05, |
|
"loss": 1.7514, |
|
"step": 17515 |
|
}, |
|
{ |
|
"epoch": 0.44178112371432526, |
|
"grad_norm": 0.9176467061042786, |
|
"learning_rate": 2.0888721608573047e-05, |
|
"loss": 1.7521, |
|
"step": 17546 |
|
}, |
|
{ |
|
"epoch": 0.442561655735022, |
|
"grad_norm": 0.92497318983078, |
|
"learning_rate": 2.0805905944311087e-05, |
|
"loss": 1.7582, |
|
"step": 17577 |
|
}, |
|
{ |
|
"epoch": 0.44334218775571865, |
|
"grad_norm": 0.9499204158782959, |
|
"learning_rate": 2.0723137609830497e-05, |
|
"loss": 1.762, |
|
"step": 17608 |
|
}, |
|
{ |
|
"epoch": 0.4441227197764153, |
|
"grad_norm": 0.9391850829124451, |
|
"learning_rate": 2.0640417539160686e-05, |
|
"loss": 1.7495, |
|
"step": 17639 |
|
}, |
|
{ |
|
"epoch": 0.44490325179711204, |
|
"grad_norm": 0.9657866358757019, |
|
"learning_rate": 2.0557746665786427e-05, |
|
"loss": 1.7642, |
|
"step": 17670 |
|
}, |
|
{ |
|
"epoch": 0.4456837838178087, |
|
"grad_norm": 0.955342173576355, |
|
"learning_rate": 2.0475125922637256e-05, |
|
"loss": 1.7645, |
|
"step": 17701 |
|
}, |
|
{ |
|
"epoch": 0.44646431583850543, |
|
"grad_norm": 0.9168228507041931, |
|
"learning_rate": 2.0392556242077047e-05, |
|
"loss": 1.7598, |
|
"step": 17732 |
|
}, |
|
{ |
|
"epoch": 0.4472448478592021, |
|
"grad_norm": 0.8854954242706299, |
|
"learning_rate": 2.031003855589343e-05, |
|
"loss": 1.7576, |
|
"step": 17763 |
|
}, |
|
{ |
|
"epoch": 0.44802537987989877, |
|
"grad_norm": 0.9424418210983276, |
|
"learning_rate": 2.022757379528727e-05, |
|
"loss": 1.766, |
|
"step": 17794 |
|
}, |
|
{ |
|
"epoch": 0.4488059119005955, |
|
"grad_norm": 0.9471886157989502, |
|
"learning_rate": 2.0145162890862184e-05, |
|
"loss": 1.7458, |
|
"step": 17825 |
|
}, |
|
{ |
|
"epoch": 0.44958644392129216, |
|
"grad_norm": 0.9373571872711182, |
|
"learning_rate": 2.0062806772614022e-05, |
|
"loss": 1.743, |
|
"step": 17856 |
|
}, |
|
{ |
|
"epoch": 0.4503669759419888, |
|
"grad_norm": 0.8964229226112366, |
|
"learning_rate": 1.9980506369920392e-05, |
|
"loss": 1.756, |
|
"step": 17887 |
|
}, |
|
{ |
|
"epoch": 0.45114750796268555, |
|
"grad_norm": 0.9414677619934082, |
|
"learning_rate": 1.989826261153015e-05, |
|
"loss": 1.7407, |
|
"step": 17918 |
|
}, |
|
{ |
|
"epoch": 0.4519280399833822, |
|
"grad_norm": 0.9373415112495422, |
|
"learning_rate": 1.9816076425552923e-05, |
|
"loss": 1.746, |
|
"step": 17949 |
|
}, |
|
{ |
|
"epoch": 0.45270857200407894, |
|
"grad_norm": 0.8916881084442139, |
|
"learning_rate": 1.9733948739448676e-05, |
|
"loss": 1.7564, |
|
"step": 17980 |
|
}, |
|
{ |
|
"epoch": 0.4534891040247756, |
|
"grad_norm": 0.9054657816886902, |
|
"learning_rate": 1.9651880480017155e-05, |
|
"loss": 1.7586, |
|
"step": 18011 |
|
}, |
|
{ |
|
"epoch": 0.45426963604547227, |
|
"grad_norm": 0.9446020126342773, |
|
"learning_rate": 1.9569872573387516e-05, |
|
"loss": 1.758, |
|
"step": 18042 |
|
}, |
|
{ |
|
"epoch": 0.455050168066169, |
|
"grad_norm": 0.9143036007881165, |
|
"learning_rate": 1.9487925945007854e-05, |
|
"loss": 1.7554, |
|
"step": 18073 |
|
}, |
|
{ |
|
"epoch": 0.45583070008686566, |
|
"grad_norm": 0.9555609822273254, |
|
"learning_rate": 1.9406041519634726e-05, |
|
"loss": 1.7533, |
|
"step": 18104 |
|
}, |
|
{ |
|
"epoch": 0.4566112321075623, |
|
"grad_norm": 0.929480791091919, |
|
"learning_rate": 1.932422022132275e-05, |
|
"loss": 1.7599, |
|
"step": 18135 |
|
}, |
|
{ |
|
"epoch": 0.45739176412825905, |
|
"grad_norm": 0.9011172652244568, |
|
"learning_rate": 1.924246297341414e-05, |
|
"loss": 1.7712, |
|
"step": 18166 |
|
}, |
|
{ |
|
"epoch": 0.4581722961489557, |
|
"grad_norm": 0.9395762085914612, |
|
"learning_rate": 1.9160770698528338e-05, |
|
"loss": 1.7715, |
|
"step": 18197 |
|
}, |
|
{ |
|
"epoch": 0.45895282816965244, |
|
"grad_norm": 0.9550304412841797, |
|
"learning_rate": 1.907914431855156e-05, |
|
"loss": 1.7315, |
|
"step": 18228 |
|
}, |
|
{ |
|
"epoch": 0.4597333601903491, |
|
"grad_norm": 0.9520135521888733, |
|
"learning_rate": 1.8997584754626412e-05, |
|
"loss": 1.7453, |
|
"step": 18259 |
|
}, |
|
{ |
|
"epoch": 0.46051389221104577, |
|
"grad_norm": 0.8951966762542725, |
|
"learning_rate": 1.8916092927141486e-05, |
|
"loss": 1.7496, |
|
"step": 18290 |
|
}, |
|
{ |
|
"epoch": 0.4612944242317425, |
|
"grad_norm": 0.936210036277771, |
|
"learning_rate": 1.883466975572098e-05, |
|
"loss": 1.7407, |
|
"step": 18321 |
|
}, |
|
{ |
|
"epoch": 0.46207495625243916, |
|
"grad_norm": 0.95763099193573, |
|
"learning_rate": 1.8753316159214312e-05, |
|
"loss": 1.7566, |
|
"step": 18352 |
|
}, |
|
{ |
|
"epoch": 0.46285548827313583, |
|
"grad_norm": 0.9345053434371948, |
|
"learning_rate": 1.8672033055685766e-05, |
|
"loss": 1.7769, |
|
"step": 18383 |
|
}, |
|
{ |
|
"epoch": 0.46363602029383255, |
|
"grad_norm": 0.9099903106689453, |
|
"learning_rate": 1.8590821362404116e-05, |
|
"loss": 1.7416, |
|
"step": 18414 |
|
}, |
|
{ |
|
"epoch": 0.4644165523145292, |
|
"grad_norm": 0.9202333092689514, |
|
"learning_rate": 1.8509681995832294e-05, |
|
"loss": 1.7372, |
|
"step": 18445 |
|
}, |
|
{ |
|
"epoch": 0.46519708433522594, |
|
"grad_norm": 0.9366357922554016, |
|
"learning_rate": 1.8428615871617004e-05, |
|
"loss": 1.7412, |
|
"step": 18476 |
|
}, |
|
{ |
|
"epoch": 0.4659776163559226, |
|
"grad_norm": 0.9029955267906189, |
|
"learning_rate": 1.8347623904578448e-05, |
|
"loss": 1.7465, |
|
"step": 18507 |
|
}, |
|
{ |
|
"epoch": 0.4667581483766193, |
|
"grad_norm": 0.9197161197662354, |
|
"learning_rate": 1.8266707008699975e-05, |
|
"loss": 1.7462, |
|
"step": 18538 |
|
}, |
|
{ |
|
"epoch": 0.467538680397316, |
|
"grad_norm": 0.9279952049255371, |
|
"learning_rate": 1.818586609711774e-05, |
|
"loss": 1.7496, |
|
"step": 18569 |
|
}, |
|
{ |
|
"epoch": 0.46831921241801266, |
|
"grad_norm": 0.9148679971694946, |
|
"learning_rate": 1.8105102082110462e-05, |
|
"loss": 1.7389, |
|
"step": 18600 |
|
}, |
|
{ |
|
"epoch": 0.46909974443870933, |
|
"grad_norm": 0.9217187166213989, |
|
"learning_rate": 1.8024415875089058e-05, |
|
"loss": 1.7377, |
|
"step": 18631 |
|
}, |
|
{ |
|
"epoch": 0.46988027645940605, |
|
"grad_norm": 0.9556980133056641, |
|
"learning_rate": 1.7943808386586407e-05, |
|
"loss": 1.7391, |
|
"step": 18662 |
|
}, |
|
{ |
|
"epoch": 0.4706608084801027, |
|
"grad_norm": 0.9437476992607117, |
|
"learning_rate": 1.7863280526247073e-05, |
|
"loss": 1.7401, |
|
"step": 18693 |
|
}, |
|
{ |
|
"epoch": 0.4714413405007994, |
|
"grad_norm": 0.923971951007843, |
|
"learning_rate": 1.7782833202817003e-05, |
|
"loss": 1.7492, |
|
"step": 18724 |
|
}, |
|
{ |
|
"epoch": 0.4722218725214961, |
|
"grad_norm": 0.9325092434883118, |
|
"learning_rate": 1.7702467324133327e-05, |
|
"loss": 1.7477, |
|
"step": 18755 |
|
}, |
|
{ |
|
"epoch": 0.4730024045421928, |
|
"grad_norm": 0.9427450299263, |
|
"learning_rate": 1.7622183797114042e-05, |
|
"loss": 1.7722, |
|
"step": 18786 |
|
}, |
|
{ |
|
"epoch": 0.4737829365628895, |
|
"grad_norm": 0.8803645372390747, |
|
"learning_rate": 1.7541983527747838e-05, |
|
"loss": 1.7428, |
|
"step": 18817 |
|
}, |
|
{ |
|
"epoch": 0.47456346858358617, |
|
"grad_norm": 0.9492975473403931, |
|
"learning_rate": 1.746186742108387e-05, |
|
"loss": 1.7558, |
|
"step": 18848 |
|
}, |
|
{ |
|
"epoch": 0.47534400060428283, |
|
"grad_norm": 0.9038426280021667, |
|
"learning_rate": 1.73818363812215e-05, |
|
"loss": 1.7576, |
|
"step": 18879 |
|
}, |
|
{ |
|
"epoch": 0.47612453262497956, |
|
"grad_norm": 0.9929136633872986, |
|
"learning_rate": 1.7301891311300153e-05, |
|
"loss": 1.7405, |
|
"step": 18910 |
|
}, |
|
{ |
|
"epoch": 0.4769050646456762, |
|
"grad_norm": 0.9464752078056335, |
|
"learning_rate": 1.7222033113489055e-05, |
|
"loss": 1.7343, |
|
"step": 18941 |
|
}, |
|
{ |
|
"epoch": 0.4776855966663729, |
|
"grad_norm": 0.9222598075866699, |
|
"learning_rate": 1.7142262688977127e-05, |
|
"loss": 1.7377, |
|
"step": 18972 |
|
}, |
|
{ |
|
"epoch": 0.4784661286870696, |
|
"grad_norm": 0.9139207005500793, |
|
"learning_rate": 1.7062580937962764e-05, |
|
"loss": 1.7448, |
|
"step": 19003 |
|
}, |
|
{ |
|
"epoch": 0.4792466607077663, |
|
"grad_norm": 0.9283340573310852, |
|
"learning_rate": 1.698298875964369e-05, |
|
"loss": 1.7577, |
|
"step": 19034 |
|
}, |
|
{ |
|
"epoch": 0.480027192728463, |
|
"grad_norm": 0.9210373163223267, |
|
"learning_rate": 1.690348705220684e-05, |
|
"loss": 1.7351, |
|
"step": 19065 |
|
}, |
|
{ |
|
"epoch": 0.48080772474915967, |
|
"grad_norm": 0.9250310063362122, |
|
"learning_rate": 1.6824076712818156e-05, |
|
"loss": 1.7554, |
|
"step": 19096 |
|
}, |
|
{ |
|
"epoch": 0.48158825676985634, |
|
"grad_norm": 0.9364836812019348, |
|
"learning_rate": 1.6744758637612533e-05, |
|
"loss": 1.7513, |
|
"step": 19127 |
|
}, |
|
{ |
|
"epoch": 0.48236878879055306, |
|
"grad_norm": 0.9413596391677856, |
|
"learning_rate": 1.6665533721683664e-05, |
|
"loss": 1.7358, |
|
"step": 19158 |
|
}, |
|
{ |
|
"epoch": 0.4831493208112497, |
|
"grad_norm": 0.9139796495437622, |
|
"learning_rate": 1.6586402859073974e-05, |
|
"loss": 1.74, |
|
"step": 19189 |
|
}, |
|
{ |
|
"epoch": 0.4839298528319464, |
|
"grad_norm": 0.9309468269348145, |
|
"learning_rate": 1.6507366942764463e-05, |
|
"loss": 1.7407, |
|
"step": 19220 |
|
}, |
|
{ |
|
"epoch": 0.4847103848526431, |
|
"grad_norm": 0.9395124316215515, |
|
"learning_rate": 1.6428426864664732e-05, |
|
"loss": 1.7541, |
|
"step": 19251 |
|
}, |
|
{ |
|
"epoch": 0.4854909168733398, |
|
"grad_norm": 0.9307234883308411, |
|
"learning_rate": 1.6349583515602816e-05, |
|
"loss": 1.7509, |
|
"step": 19282 |
|
}, |
|
{ |
|
"epoch": 0.4862714488940365, |
|
"grad_norm": 0.9114024043083191, |
|
"learning_rate": 1.6270837785315208e-05, |
|
"loss": 1.749, |
|
"step": 19313 |
|
}, |
|
{ |
|
"epoch": 0.4870519809147332, |
|
"grad_norm": 0.9282791018486023, |
|
"learning_rate": 1.619219056243676e-05, |
|
"loss": 1.7257, |
|
"step": 19344 |
|
}, |
|
{ |
|
"epoch": 0.48783251293542984, |
|
"grad_norm": 0.9253626465797424, |
|
"learning_rate": 1.6113642734490698e-05, |
|
"loss": 1.7298, |
|
"step": 19375 |
|
}, |
|
{ |
|
"epoch": 0.48861304495612656, |
|
"grad_norm": 0.9356294870376587, |
|
"learning_rate": 1.6035195187878577e-05, |
|
"loss": 1.7419, |
|
"step": 19406 |
|
}, |
|
{ |
|
"epoch": 0.48939357697682323, |
|
"grad_norm": 0.9312499761581421, |
|
"learning_rate": 1.5956848807870305e-05, |
|
"loss": 1.7409, |
|
"step": 19437 |
|
}, |
|
{ |
|
"epoch": 0.4901741089975199, |
|
"grad_norm": 0.91697758436203, |
|
"learning_rate": 1.587860447859413e-05, |
|
"loss": 1.7288, |
|
"step": 19468 |
|
}, |
|
{ |
|
"epoch": 0.4909546410182166, |
|
"grad_norm": 0.9301339983940125, |
|
"learning_rate": 1.5800463083026686e-05, |
|
"loss": 1.7479, |
|
"step": 19499 |
|
}, |
|
{ |
|
"epoch": 0.4917351730389133, |
|
"grad_norm": 0.9044349193572998, |
|
"learning_rate": 1.572242550298298e-05, |
|
"loss": 1.7264, |
|
"step": 19530 |
|
}, |
|
{ |
|
"epoch": 0.49251570505961, |
|
"grad_norm": 0.9050460457801819, |
|
"learning_rate": 1.56444926191065e-05, |
|
"loss": 1.7394, |
|
"step": 19561 |
|
}, |
|
{ |
|
"epoch": 0.4932962370803067, |
|
"grad_norm": 0.9435474872589111, |
|
"learning_rate": 1.5566665310859257e-05, |
|
"loss": 1.7438, |
|
"step": 19592 |
|
}, |
|
{ |
|
"epoch": 0.49407676910100334, |
|
"grad_norm": 0.9283745288848877, |
|
"learning_rate": 1.5488944456511846e-05, |
|
"loss": 1.7603, |
|
"step": 19623 |
|
}, |
|
{ |
|
"epoch": 0.49485730112170007, |
|
"grad_norm": 0.9714552164077759, |
|
"learning_rate": 1.5411330933133546e-05, |
|
"loss": 1.7316, |
|
"step": 19654 |
|
}, |
|
{ |
|
"epoch": 0.49563783314239673, |
|
"grad_norm": 0.92011958360672, |
|
"learning_rate": 1.533382561658241e-05, |
|
"loss": 1.7478, |
|
"step": 19685 |
|
}, |
|
{ |
|
"epoch": 0.4964183651630934, |
|
"grad_norm": 0.929989218711853, |
|
"learning_rate": 1.525642938149541e-05, |
|
"loss": 1.7497, |
|
"step": 19716 |
|
}, |
|
{ |
|
"epoch": 0.4971988971837901, |
|
"grad_norm": 0.9507896900177002, |
|
"learning_rate": 1.5179143101278536e-05, |
|
"loss": 1.7485, |
|
"step": 19747 |
|
}, |
|
{ |
|
"epoch": 0.4979794292044868, |
|
"grad_norm": 0.9416571855545044, |
|
"learning_rate": 1.5101967648096955e-05, |
|
"loss": 1.7289, |
|
"step": 19778 |
|
}, |
|
{ |
|
"epoch": 0.4987599612251835, |
|
"grad_norm": 0.9024575352668762, |
|
"learning_rate": 1.5024903892865172e-05, |
|
"loss": 1.7436, |
|
"step": 19809 |
|
}, |
|
{ |
|
"epoch": 0.4995404932458802, |
|
"grad_norm": 0.9383466243743896, |
|
"learning_rate": 1.4947952705237184e-05, |
|
"loss": 1.7362, |
|
"step": 19840 |
|
}, |
|
{ |
|
"epoch": 0.5003210252665768, |
|
"grad_norm": 0.9311878681182861, |
|
"learning_rate": 1.4871114953596682e-05, |
|
"loss": 1.7595, |
|
"step": 19871 |
|
}, |
|
{ |
|
"epoch": 0.5011015572872736, |
|
"grad_norm": 0.9517268538475037, |
|
"learning_rate": 1.4794391505047256e-05, |
|
"loss": 1.7367, |
|
"step": 19902 |
|
}, |
|
{ |
|
"epoch": 0.5018820893079703, |
|
"grad_norm": 0.911535918712616, |
|
"learning_rate": 1.4717783225402596e-05, |
|
"loss": 1.7268, |
|
"step": 19933 |
|
}, |
|
{ |
|
"epoch": 0.5026626213286669, |
|
"grad_norm": 0.9200896620750427, |
|
"learning_rate": 1.4641290979176735e-05, |
|
"loss": 1.7353, |
|
"step": 19964 |
|
}, |
|
{ |
|
"epoch": 0.5034431533493636, |
|
"grad_norm": 0.9185106754302979, |
|
"learning_rate": 1.4564915629574246e-05, |
|
"loss": 1.7441, |
|
"step": 19995 |
|
}, |
|
{ |
|
"epoch": 0.5042236853700603, |
|
"grad_norm": 0.9203515648841858, |
|
"learning_rate": 1.4488658038480601e-05, |
|
"loss": 1.7293, |
|
"step": 20026 |
|
}, |
|
{ |
|
"epoch": 0.505004217390757, |
|
"grad_norm": 0.9532755613327026, |
|
"learning_rate": 1.4412519066452323e-05, |
|
"loss": 1.7478, |
|
"step": 20057 |
|
}, |
|
{ |
|
"epoch": 0.5057847494114537, |
|
"grad_norm": 0.9353486895561218, |
|
"learning_rate": 1.4336499572707373e-05, |
|
"loss": 1.7468, |
|
"step": 20088 |
|
}, |
|
{ |
|
"epoch": 0.5065652814321504, |
|
"grad_norm": 0.9349709153175354, |
|
"learning_rate": 1.4260600415115433e-05, |
|
"loss": 1.7256, |
|
"step": 20119 |
|
}, |
|
{ |
|
"epoch": 0.507345813452847, |
|
"grad_norm": 0.9995957016944885, |
|
"learning_rate": 1.4184822450188137e-05, |
|
"loss": 1.7361, |
|
"step": 20150 |
|
}, |
|
{ |
|
"epoch": 0.5081263454735437, |
|
"grad_norm": 0.9351862072944641, |
|
"learning_rate": 1.410916653306954e-05, |
|
"loss": 1.7537, |
|
"step": 20181 |
|
}, |
|
{ |
|
"epoch": 0.5089068774942405, |
|
"grad_norm": 0.9246407747268677, |
|
"learning_rate": 1.403363351752639e-05, |
|
"loss": 1.7241, |
|
"step": 20212 |
|
}, |
|
{ |
|
"epoch": 0.5096874095149371, |
|
"grad_norm": 0.9154999256134033, |
|
"learning_rate": 1.3958224255938485e-05, |
|
"loss": 1.7207, |
|
"step": 20243 |
|
}, |
|
{ |
|
"epoch": 0.5104679415356338, |
|
"grad_norm": 0.9373302459716797, |
|
"learning_rate": 1.388293959928911e-05, |
|
"loss": 1.7538, |
|
"step": 20274 |
|
}, |
|
{ |
|
"epoch": 0.5112484735563305, |
|
"grad_norm": 0.924915611743927, |
|
"learning_rate": 1.3807780397155379e-05, |
|
"loss": 1.737, |
|
"step": 20305 |
|
}, |
|
{ |
|
"epoch": 0.5120290055770271, |
|
"grad_norm": 0.9185555577278137, |
|
"learning_rate": 1.3732747497698655e-05, |
|
"loss": 1.7105, |
|
"step": 20336 |
|
}, |
|
{ |
|
"epoch": 0.5128095375977239, |
|
"grad_norm": 0.9583287239074707, |
|
"learning_rate": 1.3657841747655038e-05, |
|
"loss": 1.7398, |
|
"step": 20367 |
|
}, |
|
{ |
|
"epoch": 0.5135900696184206, |
|
"grad_norm": 0.9166149497032166, |
|
"learning_rate": 1.3583063992325706e-05, |
|
"loss": 1.7428, |
|
"step": 20398 |
|
}, |
|
{ |
|
"epoch": 0.5143706016391173, |
|
"grad_norm": 0.9236160516738892, |
|
"learning_rate": 1.3508415075567496e-05, |
|
"loss": 1.7355, |
|
"step": 20429 |
|
}, |
|
{ |
|
"epoch": 0.5151511336598139, |
|
"grad_norm": 0.9086827635765076, |
|
"learning_rate": 1.343389583978327e-05, |
|
"loss": 1.737, |
|
"step": 20460 |
|
}, |
|
{ |
|
"epoch": 0.5159316656805106, |
|
"grad_norm": 0.9379333257675171, |
|
"learning_rate": 1.3359507125912468e-05, |
|
"loss": 1.7218, |
|
"step": 20491 |
|
}, |
|
{ |
|
"epoch": 0.5167121977012074, |
|
"grad_norm": 0.9533993005752563, |
|
"learning_rate": 1.3285249773421627e-05, |
|
"loss": 1.7339, |
|
"step": 20522 |
|
}, |
|
{ |
|
"epoch": 0.517492729721904, |
|
"grad_norm": 0.9649523496627808, |
|
"learning_rate": 1.3211124620294884e-05, |
|
"loss": 1.7459, |
|
"step": 20553 |
|
}, |
|
{ |
|
"epoch": 0.5182732617426007, |
|
"grad_norm": 0.9658921360969543, |
|
"learning_rate": 1.313713250302451e-05, |
|
"loss": 1.7391, |
|
"step": 20584 |
|
}, |
|
{ |
|
"epoch": 0.5190537937632974, |
|
"grad_norm": 0.960582435131073, |
|
"learning_rate": 1.3063274256601479e-05, |
|
"loss": 1.7303, |
|
"step": 20615 |
|
}, |
|
{ |
|
"epoch": 0.519834325783994, |
|
"grad_norm": 0.9230870008468628, |
|
"learning_rate": 1.2989550714506086e-05, |
|
"loss": 1.7223, |
|
"step": 20646 |
|
}, |
|
{ |
|
"epoch": 0.5206148578046907, |
|
"grad_norm": 0.9037120342254639, |
|
"learning_rate": 1.291596270869846e-05, |
|
"loss": 1.7191, |
|
"step": 20677 |
|
}, |
|
{ |
|
"epoch": 0.5213953898253875, |
|
"grad_norm": 0.9674888253211975, |
|
"learning_rate": 1.284251106960927e-05, |
|
"loss": 1.7177, |
|
"step": 20708 |
|
}, |
|
{ |
|
"epoch": 0.5221759218460841, |
|
"grad_norm": 0.908338725566864, |
|
"learning_rate": 1.2769196626130263e-05, |
|
"loss": 1.7243, |
|
"step": 20739 |
|
}, |
|
{ |
|
"epoch": 0.5229564538667808, |
|
"grad_norm": 0.9437979459762573, |
|
"learning_rate": 1.2696020205604969e-05, |
|
"loss": 1.7391, |
|
"step": 20770 |
|
}, |
|
{ |
|
"epoch": 0.5237369858874775, |
|
"grad_norm": 0.9256695508956909, |
|
"learning_rate": 1.2622982633819359e-05, |
|
"loss": 1.7403, |
|
"step": 20801 |
|
}, |
|
{ |
|
"epoch": 0.5245175179081741, |
|
"grad_norm": 0.9368360042572021, |
|
"learning_rate": 1.2550084734992484e-05, |
|
"loss": 1.7379, |
|
"step": 20832 |
|
}, |
|
{ |
|
"epoch": 0.5252980499288709, |
|
"grad_norm": 0.9218337535858154, |
|
"learning_rate": 1.247732733176724e-05, |
|
"loss": 1.7255, |
|
"step": 20863 |
|
}, |
|
{ |
|
"epoch": 0.5260785819495676, |
|
"grad_norm": 0.9407544136047363, |
|
"learning_rate": 1.2404711245201044e-05, |
|
"loss": 1.7231, |
|
"step": 20894 |
|
}, |
|
{ |
|
"epoch": 0.5268591139702643, |
|
"grad_norm": 0.9167141318321228, |
|
"learning_rate": 1.2332237294756535e-05, |
|
"loss": 1.7058, |
|
"step": 20925 |
|
}, |
|
{ |
|
"epoch": 0.5276396459909609, |
|
"grad_norm": 0.9397562742233276, |
|
"learning_rate": 1.225990629829241e-05, |
|
"loss": 1.7385, |
|
"step": 20956 |
|
}, |
|
{ |
|
"epoch": 0.5284201780116576, |
|
"grad_norm": 0.9333503842353821, |
|
"learning_rate": 1.2187719072054136e-05, |
|
"loss": 1.7451, |
|
"step": 20987 |
|
}, |
|
{ |
|
"epoch": 0.5292007100323544, |
|
"grad_norm": 0.9979040622711182, |
|
"learning_rate": 1.2115676430664735e-05, |
|
"loss": 1.7511, |
|
"step": 21018 |
|
}, |
|
{ |
|
"epoch": 0.529981242053051, |
|
"grad_norm": 0.9520873427391052, |
|
"learning_rate": 1.2043779187115647e-05, |
|
"loss": 1.7272, |
|
"step": 21049 |
|
}, |
|
{ |
|
"epoch": 0.5307617740737477, |
|
"grad_norm": 0.9319959282875061, |
|
"learning_rate": 1.1972028152757476e-05, |
|
"loss": 1.7381, |
|
"step": 21080 |
|
}, |
|
{ |
|
"epoch": 0.5315423060944444, |
|
"grad_norm": 0.997718870639801, |
|
"learning_rate": 1.1900424137290889e-05, |
|
"loss": 1.719, |
|
"step": 21111 |
|
}, |
|
{ |
|
"epoch": 0.532322838115141, |
|
"grad_norm": 0.95301353931427, |
|
"learning_rate": 1.1828967948757482e-05, |
|
"loss": 1.7272, |
|
"step": 21142 |
|
}, |
|
{ |
|
"epoch": 0.5331033701358378, |
|
"grad_norm": 0.9444928169250488, |
|
"learning_rate": 1.175766039353062e-05, |
|
"loss": 1.7327, |
|
"step": 21173 |
|
}, |
|
{ |
|
"epoch": 0.5338839021565345, |
|
"grad_norm": 0.9271036386489868, |
|
"learning_rate": 1.1686502276306382e-05, |
|
"loss": 1.7228, |
|
"step": 21204 |
|
}, |
|
{ |
|
"epoch": 0.5346644341772311, |
|
"grad_norm": 0.9555239081382751, |
|
"learning_rate": 1.1615494400094445e-05, |
|
"loss": 1.7344, |
|
"step": 21235 |
|
}, |
|
{ |
|
"epoch": 0.5354449661979278, |
|
"grad_norm": 0.9300810694694519, |
|
"learning_rate": 1.1544637566209029e-05, |
|
"loss": 1.7216, |
|
"step": 21266 |
|
}, |
|
{ |
|
"epoch": 0.5362254982186245, |
|
"grad_norm": 0.9187619686126709, |
|
"learning_rate": 1.1473932574259886e-05, |
|
"loss": 1.7382, |
|
"step": 21297 |
|
}, |
|
{ |
|
"epoch": 0.5370060302393211, |
|
"grad_norm": 0.9264461398124695, |
|
"learning_rate": 1.1403380222143247e-05, |
|
"loss": 1.7371, |
|
"step": 21328 |
|
}, |
|
{ |
|
"epoch": 0.5377865622600179, |
|
"grad_norm": 0.9823473691940308, |
|
"learning_rate": 1.1332981306032808e-05, |
|
"loss": 1.7292, |
|
"step": 21359 |
|
}, |
|
{ |
|
"epoch": 0.5385670942807146, |
|
"grad_norm": 0.9366657733917236, |
|
"learning_rate": 1.1262736620370762e-05, |
|
"loss": 1.7237, |
|
"step": 21390 |
|
}, |
|
{ |
|
"epoch": 0.5393476263014112, |
|
"grad_norm": 0.9524781107902527, |
|
"learning_rate": 1.1192646957858854e-05, |
|
"loss": 1.7155, |
|
"step": 21421 |
|
}, |
|
{ |
|
"epoch": 0.5401281583221079, |
|
"grad_norm": 0.947833776473999, |
|
"learning_rate": 1.1122713109449381e-05, |
|
"loss": 1.7106, |
|
"step": 21452 |
|
}, |
|
{ |
|
"epoch": 0.5409086903428046, |
|
"grad_norm": 0.9224444031715393, |
|
"learning_rate": 1.105293586433634e-05, |
|
"loss": 1.7265, |
|
"step": 21483 |
|
}, |
|
{ |
|
"epoch": 0.5416892223635014, |
|
"grad_norm": 0.9879010319709778, |
|
"learning_rate": 1.0983316009946446e-05, |
|
"loss": 1.7375, |
|
"step": 21514 |
|
}, |
|
{ |
|
"epoch": 0.542469754384198, |
|
"grad_norm": 0.9554262757301331, |
|
"learning_rate": 1.0913854331930282e-05, |
|
"loss": 1.7256, |
|
"step": 21545 |
|
}, |
|
{ |
|
"epoch": 0.5432502864048947, |
|
"grad_norm": 0.9443463087081909, |
|
"learning_rate": 1.0844551614153456e-05, |
|
"loss": 1.7001, |
|
"step": 21576 |
|
}, |
|
{ |
|
"epoch": 0.5440308184255914, |
|
"grad_norm": 0.9143596887588501, |
|
"learning_rate": 1.0775408638687725e-05, |
|
"loss": 1.7149, |
|
"step": 21607 |
|
}, |
|
{ |
|
"epoch": 0.544811350446288, |
|
"grad_norm": 0.9478058218955994, |
|
"learning_rate": 1.0706426185802165e-05, |
|
"loss": 1.7199, |
|
"step": 21638 |
|
}, |
|
{ |
|
"epoch": 0.5455918824669848, |
|
"grad_norm": 0.9589979648590088, |
|
"learning_rate": 1.0637605033954371e-05, |
|
"loss": 1.7145, |
|
"step": 21669 |
|
}, |
|
{ |
|
"epoch": 0.5463724144876815, |
|
"grad_norm": 0.9296464323997498, |
|
"learning_rate": 1.05689459597817e-05, |
|
"loss": 1.7279, |
|
"step": 21700 |
|
}, |
|
{ |
|
"epoch": 0.5471529465083781, |
|
"grad_norm": 0.9697522521018982, |
|
"learning_rate": 1.050044973809246e-05, |
|
"loss": 1.7341, |
|
"step": 21731 |
|
}, |
|
{ |
|
"epoch": 0.5479334785290748, |
|
"grad_norm": 0.9401953816413879, |
|
"learning_rate": 1.043211714185722e-05, |
|
"loss": 1.7276, |
|
"step": 21762 |
|
}, |
|
{ |
|
"epoch": 0.5487140105497715, |
|
"grad_norm": 0.92669677734375, |
|
"learning_rate": 1.036394894220003e-05, |
|
"loss": 1.7209, |
|
"step": 21793 |
|
}, |
|
{ |
|
"epoch": 0.5494945425704681, |
|
"grad_norm": 0.942746102809906, |
|
"learning_rate": 1.0295945908389751e-05, |
|
"loss": 1.731, |
|
"step": 21824 |
|
}, |
|
{ |
|
"epoch": 0.5502750745911649, |
|
"grad_norm": 0.9401955604553223, |
|
"learning_rate": 1.0228108807831393e-05, |
|
"loss": 1.7308, |
|
"step": 21855 |
|
}, |
|
{ |
|
"epoch": 0.5510556066118616, |
|
"grad_norm": 0.9710572957992554, |
|
"learning_rate": 1.01604384060574e-05, |
|
"loss": 1.7104, |
|
"step": 21886 |
|
}, |
|
{ |
|
"epoch": 0.5518361386325582, |
|
"grad_norm": 0.9225825071334839, |
|
"learning_rate": 1.009293546671907e-05, |
|
"loss": 1.7299, |
|
"step": 21917 |
|
}, |
|
{ |
|
"epoch": 0.5526166706532549, |
|
"grad_norm": 0.9352388381958008, |
|
"learning_rate": 1.002560075157791e-05, |
|
"loss": 1.7098, |
|
"step": 21948 |
|
}, |
|
{ |
|
"epoch": 0.5533972026739516, |
|
"grad_norm": 0.9479419589042664, |
|
"learning_rate": 9.958435020496995e-06, |
|
"loss": 1.7217, |
|
"step": 21979 |
|
}, |
|
{ |
|
"epoch": 0.5541777346946484, |
|
"grad_norm": 0.9639567732810974, |
|
"learning_rate": 9.89143903143249e-06, |
|
"loss": 1.715, |
|
"step": 22010 |
|
}, |
|
{ |
|
"epoch": 0.554958266715345, |
|
"grad_norm": 0.9379074573516846, |
|
"learning_rate": 9.824613540425038e-06, |
|
"loss": 1.7221, |
|
"step": 22041 |
|
}, |
|
{ |
|
"epoch": 0.5557387987360417, |
|
"grad_norm": 0.9399924874305725, |
|
"learning_rate": 9.757959301591197e-06, |
|
"loss": 1.7216, |
|
"step": 22072 |
|
}, |
|
{ |
|
"epoch": 0.5565193307567384, |
|
"grad_norm": 0.9303331971168518, |
|
"learning_rate": 9.691477067115017e-06, |
|
"loss": 1.7275, |
|
"step": 22103 |
|
}, |
|
{ |
|
"epoch": 0.557299862777435, |
|
"grad_norm": 0.92960524559021, |
|
"learning_rate": 9.625167587239467e-06, |
|
"loss": 1.7227, |
|
"step": 22134 |
|
}, |
|
{ |
|
"epoch": 0.5580803947981318, |
|
"grad_norm": 0.9338411688804626, |
|
"learning_rate": 9.559031610258007e-06, |
|
"loss": 1.7078, |
|
"step": 22165 |
|
}, |
|
{ |
|
"epoch": 0.5588609268188285, |
|
"grad_norm": 0.917374312877655, |
|
"learning_rate": 9.493069882506164e-06, |
|
"loss": 1.7395, |
|
"step": 22196 |
|
}, |
|
{ |
|
"epoch": 0.5596414588395251, |
|
"grad_norm": 0.9369717240333557, |
|
"learning_rate": 9.427283148353056e-06, |
|
"loss": 1.7157, |
|
"step": 22227 |
|
}, |
|
{ |
|
"epoch": 0.5604219908602218, |
|
"grad_norm": 0.9392804503440857, |
|
"learning_rate": 9.361672150193052e-06, |
|
"loss": 1.7107, |
|
"step": 22258 |
|
}, |
|
{ |
|
"epoch": 0.5612025228809185, |
|
"grad_norm": 0.926115095615387, |
|
"learning_rate": 9.29623762843734e-06, |
|
"loss": 1.7057, |
|
"step": 22289 |
|
}, |
|
{ |
|
"epoch": 0.5619830549016152, |
|
"grad_norm": 0.9339826703071594, |
|
"learning_rate": 9.230980321505594e-06, |
|
"loss": 1.726, |
|
"step": 22320 |
|
}, |
|
{ |
|
"epoch": 0.5627635869223119, |
|
"grad_norm": 0.916395902633667, |
|
"learning_rate": 9.165900965817668e-06, |
|
"loss": 1.7097, |
|
"step": 22351 |
|
}, |
|
{ |
|
"epoch": 0.5635441189430086, |
|
"grad_norm": 0.9421140551567078, |
|
"learning_rate": 9.101000295785245e-06, |
|
"loss": 1.7229, |
|
"step": 22382 |
|
}, |
|
{ |
|
"epoch": 0.5643246509637052, |
|
"grad_norm": 0.9253449440002441, |
|
"learning_rate": 9.036279043803565e-06, |
|
"loss": 1.7258, |
|
"step": 22413 |
|
}, |
|
{ |
|
"epoch": 0.5651051829844019, |
|
"grad_norm": 0.945318341255188, |
|
"learning_rate": 8.971737940243147e-06, |
|
"loss": 1.7322, |
|
"step": 22444 |
|
}, |
|
{ |
|
"epoch": 0.5658857150050987, |
|
"grad_norm": 0.9573378562927246, |
|
"learning_rate": 8.907377713441592e-06, |
|
"loss": 1.7183, |
|
"step": 22475 |
|
}, |
|
{ |
|
"epoch": 0.5666662470257954, |
|
"grad_norm": 0.9631168842315674, |
|
"learning_rate": 8.843199089695293e-06, |
|
"loss": 1.7139, |
|
"step": 22506 |
|
}, |
|
{ |
|
"epoch": 0.567446779046492, |
|
"grad_norm": 0.9413036704063416, |
|
"learning_rate": 8.779202793251311e-06, |
|
"loss": 1.7369, |
|
"step": 22537 |
|
}, |
|
{ |
|
"epoch": 0.5682273110671887, |
|
"grad_norm": 0.9556379914283752, |
|
"learning_rate": 8.715389546299149e-06, |
|
"loss": 1.7228, |
|
"step": 22568 |
|
}, |
|
{ |
|
"epoch": 0.5690078430878854, |
|
"grad_norm": 0.9588301181793213, |
|
"learning_rate": 8.651760068962617e-06, |
|
"loss": 1.7182, |
|
"step": 22599 |
|
}, |
|
{ |
|
"epoch": 0.569788375108582, |
|
"grad_norm": 0.9203631281852722, |
|
"learning_rate": 8.588315079291733e-06, |
|
"loss": 1.7032, |
|
"step": 22630 |
|
}, |
|
{ |
|
"epoch": 0.5705689071292788, |
|
"grad_norm": 0.9229585528373718, |
|
"learning_rate": 8.52505529325457e-06, |
|
"loss": 1.7246, |
|
"step": 22661 |
|
}, |
|
{ |
|
"epoch": 0.5713494391499755, |
|
"grad_norm": 0.9430622458457947, |
|
"learning_rate": 8.461981424729216e-06, |
|
"loss": 1.7339, |
|
"step": 22692 |
|
}, |
|
{ |
|
"epoch": 0.5721299711706721, |
|
"grad_norm": 0.9134554862976074, |
|
"learning_rate": 8.399094185495725e-06, |
|
"loss": 1.7086, |
|
"step": 22723 |
|
}, |
|
{ |
|
"epoch": 0.5729105031913688, |
|
"grad_norm": 0.9268332719802856, |
|
"learning_rate": 8.336394285228017e-06, |
|
"loss": 1.719, |
|
"step": 22754 |
|
}, |
|
{ |
|
"epoch": 0.5736910352120655, |
|
"grad_norm": 0.9411998987197876, |
|
"learning_rate": 8.273882431485952e-06, |
|
"loss": 1.704, |
|
"step": 22785 |
|
}, |
|
{ |
|
"epoch": 0.5744715672327622, |
|
"grad_norm": 1.0843279361724854, |
|
"learning_rate": 8.211559329707316e-06, |
|
"loss": 1.708, |
|
"step": 22816 |
|
}, |
|
{ |
|
"epoch": 0.5752520992534589, |
|
"grad_norm": 0.9296087026596069, |
|
"learning_rate": 8.149425683199823e-06, |
|
"loss": 1.7076, |
|
"step": 22847 |
|
}, |
|
{ |
|
"epoch": 0.5760326312741556, |
|
"grad_norm": 0.9319373965263367, |
|
"learning_rate": 8.08748219313325e-06, |
|
"loss": 1.7292, |
|
"step": 22878 |
|
}, |
|
{ |
|
"epoch": 0.5768131632948522, |
|
"grad_norm": 0.9682207107543945, |
|
"learning_rate": 8.025729558531453e-06, |
|
"loss": 1.7181, |
|
"step": 22909 |
|
}, |
|
{ |
|
"epoch": 0.5775936953155489, |
|
"grad_norm": 0.9323101043701172, |
|
"learning_rate": 7.964168476264508e-06, |
|
"loss": 1.7098, |
|
"step": 22940 |
|
}, |
|
{ |
|
"epoch": 0.5783742273362457, |
|
"grad_norm": 0.9210656881332397, |
|
"learning_rate": 7.902799641040884e-06, |
|
"loss": 1.7053, |
|
"step": 22971 |
|
}, |
|
{ |
|
"epoch": 0.5791547593569424, |
|
"grad_norm": 0.9448071122169495, |
|
"learning_rate": 7.841623745399523e-06, |
|
"loss": 1.7209, |
|
"step": 23002 |
|
}, |
|
{ |
|
"epoch": 0.579935291377639, |
|
"grad_norm": 0.9426878690719604, |
|
"learning_rate": 7.780641479702114e-06, |
|
"loss": 1.7286, |
|
"step": 23033 |
|
}, |
|
{ |
|
"epoch": 0.5807158233983357, |
|
"grad_norm": 0.9002141356468201, |
|
"learning_rate": 7.719853532125227e-06, |
|
"loss": 1.7074, |
|
"step": 23064 |
|
}, |
|
{ |
|
"epoch": 0.5814963554190324, |
|
"grad_norm": 0.9470099210739136, |
|
"learning_rate": 7.65926058865258e-06, |
|
"loss": 1.7092, |
|
"step": 23095 |
|
}, |
|
{ |
|
"epoch": 0.582276887439729, |
|
"grad_norm": 0.9272165894508362, |
|
"learning_rate": 7.598863333067313e-06, |
|
"loss": 1.7282, |
|
"step": 23126 |
|
}, |
|
{ |
|
"epoch": 0.5830574194604258, |
|
"grad_norm": 0.9540103673934937, |
|
"learning_rate": 7.538662446944253e-06, |
|
"loss": 1.7071, |
|
"step": 23157 |
|
}, |
|
{ |
|
"epoch": 0.5838379514811225, |
|
"grad_norm": 0.9641364812850952, |
|
"learning_rate": 7.478658609642211e-06, |
|
"loss": 1.7331, |
|
"step": 23188 |
|
}, |
|
{ |
|
"epoch": 0.5846184835018191, |
|
"grad_norm": 0.9338697195053101, |
|
"learning_rate": 7.418852498296327e-06, |
|
"loss": 1.708, |
|
"step": 23219 |
|
}, |
|
{ |
|
"epoch": 0.5853990155225158, |
|
"grad_norm": 0.9395743608474731, |
|
"learning_rate": 7.359244787810457e-06, |
|
"loss": 1.7168, |
|
"step": 23250 |
|
}, |
|
{ |
|
"epoch": 0.5861795475432126, |
|
"grad_norm": 0.9208729267120361, |
|
"learning_rate": 7.299836150849493e-06, |
|
"loss": 1.7199, |
|
"step": 23281 |
|
}, |
|
{ |
|
"epoch": 0.5869600795639092, |
|
"grad_norm": 0.9404511451721191, |
|
"learning_rate": 7.240627257831847e-06, |
|
"loss": 1.7126, |
|
"step": 23312 |
|
}, |
|
{ |
|
"epoch": 0.5877406115846059, |
|
"grad_norm": 0.977336049079895, |
|
"learning_rate": 7.1816187769218195e-06, |
|
"loss": 1.7337, |
|
"step": 23343 |
|
}, |
|
{ |
|
"epoch": 0.5885211436053026, |
|
"grad_norm": 0.9582931995391846, |
|
"learning_rate": 7.1228113740220895e-06, |
|
"loss": 1.7145, |
|
"step": 23374 |
|
}, |
|
{ |
|
"epoch": 0.5893016756259992, |
|
"grad_norm": 0.951636791229248, |
|
"learning_rate": 7.064205712766226e-06, |
|
"loss": 1.7109, |
|
"step": 23405 |
|
}, |
|
{ |
|
"epoch": 0.5900822076466959, |
|
"grad_norm": 0.954495370388031, |
|
"learning_rate": 7.005802454511129e-06, |
|
"loss": 1.7012, |
|
"step": 23436 |
|
}, |
|
{ |
|
"epoch": 0.5908627396673927, |
|
"grad_norm": 0.9662437438964844, |
|
"learning_rate": 6.947602258329639e-06, |
|
"loss": 1.6812, |
|
"step": 23467 |
|
}, |
|
{ |
|
"epoch": 0.5916432716880893, |
|
"grad_norm": 0.9717233777046204, |
|
"learning_rate": 6.889605781003078e-06, |
|
"loss": 1.7219, |
|
"step": 23498 |
|
}, |
|
{ |
|
"epoch": 0.592423803708786, |
|
"grad_norm": 0.9253862500190735, |
|
"learning_rate": 6.831813677013776e-06, |
|
"loss": 1.7164, |
|
"step": 23529 |
|
}, |
|
{ |
|
"epoch": 0.5932043357294827, |
|
"grad_norm": 0.9557362198829651, |
|
"learning_rate": 6.774226598537792e-06, |
|
"loss": 1.6982, |
|
"step": 23560 |
|
}, |
|
{ |
|
"epoch": 0.5939848677501794, |
|
"grad_norm": 0.9103747606277466, |
|
"learning_rate": 6.716845195437482e-06, |
|
"loss": 1.7287, |
|
"step": 23591 |
|
}, |
|
{ |
|
"epoch": 0.5947653997708761, |
|
"grad_norm": 0.9627851247787476, |
|
"learning_rate": 6.659670115254168e-06, |
|
"loss": 1.7108, |
|
"step": 23622 |
|
}, |
|
{ |
|
"epoch": 0.5955459317915728, |
|
"grad_norm": 0.9510253071784973, |
|
"learning_rate": 6.602702003200872e-06, |
|
"loss": 1.7212, |
|
"step": 23653 |
|
}, |
|
{ |
|
"epoch": 0.5963264638122695, |
|
"grad_norm": 0.9404213428497314, |
|
"learning_rate": 6.545941502154992e-06, |
|
"loss": 1.7105, |
|
"step": 23684 |
|
}, |
|
{ |
|
"epoch": 0.5971069958329661, |
|
"grad_norm": 0.938019871711731, |
|
"learning_rate": 6.489389252651057e-06, |
|
"loss": 1.7291, |
|
"step": 23715 |
|
}, |
|
{ |
|
"epoch": 0.5978875278536628, |
|
"grad_norm": 0.9484905004501343, |
|
"learning_rate": 6.4330458928735325e-06, |
|
"loss": 1.7272, |
|
"step": 23746 |
|
}, |
|
{ |
|
"epoch": 0.5986680598743596, |
|
"grad_norm": 0.9613280296325684, |
|
"learning_rate": 6.376912058649559e-06, |
|
"loss": 1.7094, |
|
"step": 23777 |
|
}, |
|
{ |
|
"epoch": 0.5994485918950562, |
|
"grad_norm": 0.9795516729354858, |
|
"learning_rate": 6.320988383441845e-06, |
|
"loss": 1.7129, |
|
"step": 23808 |
|
}, |
|
{ |
|
"epoch": 0.6002291239157529, |
|
"grad_norm": 0.965208888053894, |
|
"learning_rate": 6.265275498341452e-06, |
|
"loss": 1.7062, |
|
"step": 23839 |
|
}, |
|
{ |
|
"epoch": 0.6010096559364496, |
|
"grad_norm": 0.9511181116104126, |
|
"learning_rate": 6.209774032060714e-06, |
|
"loss": 1.7012, |
|
"step": 23870 |
|
}, |
|
{ |
|
"epoch": 0.6017901879571462, |
|
"grad_norm": 0.93194979429245, |
|
"learning_rate": 6.1544846109261365e-06, |
|
"loss": 1.724, |
|
"step": 23901 |
|
}, |
|
{ |
|
"epoch": 0.602570719977843, |
|
"grad_norm": 0.952998697757721, |
|
"learning_rate": 6.099407858871342e-06, |
|
"loss": 1.7067, |
|
"step": 23932 |
|
}, |
|
{ |
|
"epoch": 0.6033512519985397, |
|
"grad_norm": 0.9473235607147217, |
|
"learning_rate": 6.044544397429958e-06, |
|
"loss": 1.7171, |
|
"step": 23963 |
|
}, |
|
{ |
|
"epoch": 0.6041317840192363, |
|
"grad_norm": 0.9270508885383606, |
|
"learning_rate": 5.989894845728708e-06, |
|
"loss": 1.7113, |
|
"step": 23994 |
|
}, |
|
{ |
|
"epoch": 0.604912316039933, |
|
"grad_norm": 0.9286333322525024, |
|
"learning_rate": 5.9354598204803605e-06, |
|
"loss": 1.6881, |
|
"step": 24025 |
|
}, |
|
{ |
|
"epoch": 0.6056928480606297, |
|
"grad_norm": 0.9414156079292297, |
|
"learning_rate": 5.881239935976762e-06, |
|
"loss": 1.715, |
|
"step": 24056 |
|
}, |
|
{ |
|
"epoch": 0.6064733800813265, |
|
"grad_norm": 0.9520138502120972, |
|
"learning_rate": 5.827235804081954e-06, |
|
"loss": 1.7181, |
|
"step": 24087 |
|
}, |
|
{ |
|
"epoch": 0.6072539121020231, |
|
"grad_norm": 0.9608426690101624, |
|
"learning_rate": 5.773448034225221e-06, |
|
"loss": 1.7186, |
|
"step": 24118 |
|
}, |
|
{ |
|
"epoch": 0.6080344441227198, |
|
"grad_norm": 0.9268354773521423, |
|
"learning_rate": 5.719877233394228e-06, |
|
"loss": 1.7047, |
|
"step": 24149 |
|
}, |
|
{ |
|
"epoch": 0.6088149761434165, |
|
"grad_norm": 0.951617956161499, |
|
"learning_rate": 5.666524006128191e-06, |
|
"loss": 1.7162, |
|
"step": 24180 |
|
}, |
|
{ |
|
"epoch": 0.6095955081641131, |
|
"grad_norm": 0.9590893387794495, |
|
"learning_rate": 5.613388954511015e-06, |
|
"loss": 1.7105, |
|
"step": 24211 |
|
}, |
|
{ |
|
"epoch": 0.6103760401848098, |
|
"grad_norm": 0.9623116850852966, |
|
"learning_rate": 5.560472678164552e-06, |
|
"loss": 1.6998, |
|
"step": 24242 |
|
}, |
|
{ |
|
"epoch": 0.6111565722055066, |
|
"grad_norm": 0.9415599703788757, |
|
"learning_rate": 5.507775774241775e-06, |
|
"loss": 1.715, |
|
"step": 24273 |
|
}, |
|
{ |
|
"epoch": 0.6119371042262032, |
|
"grad_norm": 0.979045569896698, |
|
"learning_rate": 5.4552988374200945e-06, |
|
"loss": 1.7046, |
|
"step": 24304 |
|
}, |
|
{ |
|
"epoch": 0.6127176362468999, |
|
"grad_norm": 0.9478979706764221, |
|
"learning_rate": 5.403042459894597e-06, |
|
"loss": 1.7054, |
|
"step": 24335 |
|
}, |
|
{ |
|
"epoch": 0.6134981682675966, |
|
"grad_norm": 0.9414293766021729, |
|
"learning_rate": 5.3510072313714135e-06, |
|
"loss": 1.7, |
|
"step": 24366 |
|
}, |
|
{ |
|
"epoch": 0.6142787002882932, |
|
"grad_norm": 0.912340521812439, |
|
"learning_rate": 5.2991937390610205e-06, |
|
"loss": 1.7149, |
|
"step": 24397 |
|
} |
|
], |
|
"logging_steps": 31, |
|
"max_steps": 30517, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 3052, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.7053483356244673e+19, |
|
"train_batch_size": 16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|