{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.6916017272418264, "eval_steps": 500, "global_step": 27468, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0007805320206966878, "grad_norm": 24.262310028076172, "learning_rate": 1.0157273918741808e-06, "loss": 9.1957, "step": 31 }, { "epoch": 0.0015610640413933755, "grad_norm": 29.819591522216797, "learning_rate": 2.0314547837483616e-06, "loss": 7.9757, "step": 62 }, { "epoch": 0.002341596062090063, "grad_norm": 14.961602210998535, "learning_rate": 3.0471821756225426e-06, "loss": 6.5436, "step": 93 }, { "epoch": 0.003122128082786751, "grad_norm": 5.151533126831055, "learning_rate": 4.062909567496723e-06, "loss": 5.4812, "step": 124 }, { "epoch": 0.003902660103483439, "grad_norm": 7.370818138122559, "learning_rate": 5.078636959370905e-06, "loss": 4.8332, "step": 155 }, { "epoch": 0.004683192124180126, "grad_norm": 7.932437896728516, "learning_rate": 6.094364351245085e-06, "loss": 4.4411, "step": 186 }, { "epoch": 0.005463724144876814, "grad_norm": 7.834314823150635, "learning_rate": 7.110091743119267e-06, "loss": 4.1978, "step": 217 }, { "epoch": 0.006244256165573502, "grad_norm": 11.881230354309082, "learning_rate": 8.125819134993446e-06, "loss": 3.9897, "step": 248 }, { "epoch": 0.00702478818627019, "grad_norm": 6.348697185516357, "learning_rate": 9.141546526867629e-06, "loss": 3.8375, "step": 279 }, { "epoch": 0.007805320206966878, "grad_norm": 8.414978981018066, "learning_rate": 1.015727391874181e-05, "loss": 3.7681, "step": 310 }, { "epoch": 0.008585852227663565, "grad_norm": 6.531826972961426, "learning_rate": 1.117300131061599e-05, "loss": 3.6599, "step": 341 }, { "epoch": 0.009366384248360253, "grad_norm": 9.794255256652832, "learning_rate": 1.218872870249017e-05, "loss": 3.5324, "step": 372 }, { "epoch": 0.01014691626905694, "grad_norm": 8.084766387939453, "learning_rate": 1.3204456094364351e-05, "loss": 3.4696, "step": 403 }, { "epoch": 0.010927448289753628, "grad_norm": 7.659717559814453, "learning_rate": 1.4220183486238533e-05, "loss": 3.3832, "step": 434 }, { "epoch": 0.011707980310450316, "grad_norm": 5.131711959838867, "learning_rate": 1.5235910878112714e-05, "loss": 3.3514, "step": 465 }, { "epoch": 0.012488512331147004, "grad_norm": 6.3673601150512695, "learning_rate": 1.6251638269986893e-05, "loss": 3.2988, "step": 496 }, { "epoch": 0.013269044351843692, "grad_norm": 4.5014142990112305, "learning_rate": 1.7267365661861077e-05, "loss": 3.239, "step": 527 }, { "epoch": 0.01404957637254038, "grad_norm": 5.1420769691467285, "learning_rate": 1.8283093053735257e-05, "loss": 3.1892, "step": 558 }, { "epoch": 0.014830108393237068, "grad_norm": 5.227200508117676, "learning_rate": 1.9298820445609438e-05, "loss": 3.1732, "step": 589 }, { "epoch": 0.015610640413933756, "grad_norm": 4.4621734619140625, "learning_rate": 2.031454783748362e-05, "loss": 3.0898, "step": 620 }, { "epoch": 0.016391172434630442, "grad_norm": 5.213263511657715, "learning_rate": 2.13302752293578e-05, "loss": 3.0832, "step": 651 }, { "epoch": 0.01717170445532713, "grad_norm": 4.23059606552124, "learning_rate": 2.234600262123198e-05, "loss": 3.0341, "step": 682 }, { "epoch": 0.017952236476023817, "grad_norm": 3.5276854038238525, "learning_rate": 2.336173001310616e-05, "loss": 3.0241, "step": 713 }, { "epoch": 0.018732768496720505, "grad_norm": 3.199700355529785, "learning_rate": 2.437745740498034e-05, "loss": 2.9659, "step": 744 }, { "epoch": 0.019513300517417193, "grad_norm": 9.667389869689941, "learning_rate": 2.5393184796854525e-05, "loss": 2.9292, "step": 775 }, { "epoch": 0.02029383253811388, "grad_norm": 3.677541732788086, "learning_rate": 2.6408912188728702e-05, "loss": 2.919, "step": 806 }, { "epoch": 0.02107436455881057, "grad_norm": 3.711683750152588, "learning_rate": 2.7424639580602886e-05, "loss": 2.9141, "step": 837 }, { "epoch": 0.021854896579507257, "grad_norm": 3.1250174045562744, "learning_rate": 2.8440366972477066e-05, "loss": 2.8585, "step": 868 }, { "epoch": 0.022635428600203945, "grad_norm": 3.4183220863342285, "learning_rate": 2.9456094364351244e-05, "loss": 2.8206, "step": 899 }, { "epoch": 0.023415960620900633, "grad_norm": 2.8176679611206055, "learning_rate": 3.0471821756225428e-05, "loss": 2.787, "step": 930 }, { "epoch": 0.02419649264159732, "grad_norm": 2.7993154525756836, "learning_rate": 3.148754914809961e-05, "loss": 2.7678, "step": 961 }, { "epoch": 0.02497702466229401, "grad_norm": 2.8017027378082275, "learning_rate": 3.2503276539973785e-05, "loss": 2.7499, "step": 992 }, { "epoch": 0.025757556682990696, "grad_norm": 2.958606004714966, "learning_rate": 3.351900393184797e-05, "loss": 2.7211, "step": 1023 }, { "epoch": 0.026538088703687384, "grad_norm": 2.5942046642303467, "learning_rate": 3.453473132372215e-05, "loss": 2.6823, "step": 1054 }, { "epoch": 0.027318620724384072, "grad_norm": 3.1978206634521484, "learning_rate": 3.555045871559633e-05, "loss": 2.6943, "step": 1085 }, { "epoch": 0.02809915274508076, "grad_norm": 2.6260809898376465, "learning_rate": 3.6566186107470514e-05, "loss": 2.6574, "step": 1116 }, { "epoch": 0.028879684765777448, "grad_norm": 2.6633715629577637, "learning_rate": 3.7581913499344695e-05, "loss": 2.6394, "step": 1147 }, { "epoch": 0.029660216786474135, "grad_norm": 2.5701236724853516, "learning_rate": 3.8597640891218876e-05, "loss": 2.6336, "step": 1178 }, { "epoch": 0.030440748807170823, "grad_norm": 2.4242734909057617, "learning_rate": 3.9613368283093056e-05, "loss": 2.6285, "step": 1209 }, { "epoch": 0.03122128082786751, "grad_norm": 2.6410114765167236, "learning_rate": 4.062909567496724e-05, "loss": 2.5821, "step": 1240 }, { "epoch": 0.032001812848564196, "grad_norm": 2.419649839401245, "learning_rate": 4.164482306684142e-05, "loss": 2.5576, "step": 1271 }, { "epoch": 0.032782344869260883, "grad_norm": 2.314750909805298, "learning_rate": 4.26605504587156e-05, "loss": 2.5438, "step": 1302 }, { "epoch": 0.03356287688995757, "grad_norm": 2.288361072540283, "learning_rate": 4.367627785058978e-05, "loss": 2.5197, "step": 1333 }, { "epoch": 0.03434340891065426, "grad_norm": 2.328840494155884, "learning_rate": 4.469200524246396e-05, "loss": 2.5193, "step": 1364 }, { "epoch": 0.03512394093135095, "grad_norm": 2.2555599212646484, "learning_rate": 4.570773263433814e-05, "loss": 2.4907, "step": 1395 }, { "epoch": 0.035904472952047635, "grad_norm": 2.1358766555786133, "learning_rate": 4.672346002621232e-05, "loss": 2.4737, "step": 1426 }, { "epoch": 0.03668500497274432, "grad_norm": 2.1471140384674072, "learning_rate": 4.77391874180865e-05, "loss": 2.4733, "step": 1457 }, { "epoch": 0.03746553699344101, "grad_norm": 2.1371941566467285, "learning_rate": 4.875491480996068e-05, "loss": 2.432, "step": 1488 }, { "epoch": 0.0382460690141377, "grad_norm": 1.9789690971374512, "learning_rate": 4.977064220183487e-05, "loss": 2.4407, "step": 1519 }, { "epoch": 0.039026601034834386, "grad_norm": 2.0055742263793945, "learning_rate": 4.9999915451558777e-05, "loss": 2.4178, "step": 1550 }, { "epoch": 0.039807133055531074, "grad_norm": 2.095475196838379, "learning_rate": 4.999955597496219e-05, "loss": 2.4198, "step": 1581 }, { "epoch": 0.04058766507622776, "grad_norm": 2.046337842941284, "learning_rate": 4.9998914381774255e-05, "loss": 2.3936, "step": 1612 }, { "epoch": 0.04136819709692445, "grad_norm": 1.8528848886489868, "learning_rate": 4.999799067923527e-05, "loss": 2.3893, "step": 1643 }, { "epoch": 0.04214872911762114, "grad_norm": 1.883175253868103, "learning_rate": 4.999678487776908e-05, "loss": 2.3668, "step": 1674 }, { "epoch": 0.042929261138317826, "grad_norm": 1.8084770441055298, "learning_rate": 4.9995296990983006e-05, "loss": 2.3702, "step": 1705 }, { "epoch": 0.043709793159014514, "grad_norm": 1.8336963653564453, "learning_rate": 4.999352703566763e-05, "loss": 2.357, "step": 1736 }, { "epoch": 0.0444903251797112, "grad_norm": 1.8369375467300415, "learning_rate": 4.999147503179668e-05, "loss": 2.3393, "step": 1767 }, { "epoch": 0.04527085720040789, "grad_norm": 1.8411383628845215, "learning_rate": 4.998914100252672e-05, "loss": 2.319, "step": 1798 }, { "epoch": 0.04605138922110458, "grad_norm": 1.785543441772461, "learning_rate": 4.998652497419696e-05, "loss": 2.3259, "step": 1829 }, { "epoch": 0.046831921241801265, "grad_norm": 1.7194079160690308, "learning_rate": 4.9983626976328927e-05, "loss": 2.3046, "step": 1860 }, { "epoch": 0.04761245326249795, "grad_norm": 1.7909469604492188, "learning_rate": 4.998044704162613e-05, "loss": 2.307, "step": 1891 }, { "epoch": 0.04839298528319464, "grad_norm": 1.725806474685669, "learning_rate": 4.9976985205973705e-05, "loss": 2.2992, "step": 1922 }, { "epoch": 0.04917351730389133, "grad_norm": 1.67300546169281, "learning_rate": 4.997324150843799e-05, "loss": 2.2763, "step": 1953 }, { "epoch": 0.04995404932458802, "grad_norm": 1.7416061162948608, "learning_rate": 4.99692159912661e-05, "loss": 2.2596, "step": 1984 }, { "epoch": 0.050734581345284704, "grad_norm": 1.843024492263794, "learning_rate": 4.996490869988546e-05, "loss": 2.2644, "step": 2015 }, { "epoch": 0.05151511336598139, "grad_norm": 1.6172534227371216, "learning_rate": 4.996031968290326e-05, "loss": 2.271, "step": 2046 }, { "epoch": 0.05229564538667808, "grad_norm": 1.6510816812515259, "learning_rate": 4.995544899210594e-05, "loss": 2.2445, "step": 2077 }, { "epoch": 0.05307617740737477, "grad_norm": 1.625775933265686, "learning_rate": 4.9950296682458583e-05, "loss": 2.2432, "step": 2108 }, { "epoch": 0.053856709428071456, "grad_norm": 1.602299690246582, "learning_rate": 4.994486281210429e-05, "loss": 2.2266, "step": 2139 }, { "epoch": 0.054637241448768144, "grad_norm": 1.7036757469177246, "learning_rate": 4.9939147442363566e-05, "loss": 2.2292, "step": 2170 }, { "epoch": 0.05541777346946483, "grad_norm": 1.5851162672042847, "learning_rate": 4.9933150637733574e-05, "loss": 2.2286, "step": 2201 }, { "epoch": 0.05619830549016152, "grad_norm": 1.6328140497207642, "learning_rate": 4.992687246588743e-05, "loss": 2.2342, "step": 2232 }, { "epoch": 0.05697883751085821, "grad_norm": 1.4901413917541504, "learning_rate": 4.992031299767347e-05, "loss": 2.2094, "step": 2263 }, { "epoch": 0.057759369531554895, "grad_norm": 1.5897445678710938, "learning_rate": 4.9913472307114386e-05, "loss": 2.2126, "step": 2294 }, { "epoch": 0.05853990155225158, "grad_norm": 1.5534899234771729, "learning_rate": 4.9906350471406446e-05, "loss": 2.2125, "step": 2325 }, { "epoch": 0.05932043357294827, "grad_norm": 1.5713675022125244, "learning_rate": 4.989894757091861e-05, "loss": 2.2007, "step": 2356 }, { "epoch": 0.06010096559364496, "grad_norm": 1.473955750465393, "learning_rate": 4.989126368919158e-05, "loss": 2.1819, "step": 2387 }, { "epoch": 0.06088149761434165, "grad_norm": 1.5200265645980835, "learning_rate": 4.988329891293693e-05, "loss": 2.1944, "step": 2418 }, { "epoch": 0.061662029635038335, "grad_norm": 1.5236589908599854, "learning_rate": 4.987505333203608e-05, "loss": 2.1816, "step": 2449 }, { "epoch": 0.06244256165573502, "grad_norm": 1.5306485891342163, "learning_rate": 4.9866527039539276e-05, "loss": 2.187, "step": 2480 }, { "epoch": 0.06322309367643171, "grad_norm": 1.5017462968826294, "learning_rate": 4.9857720131664594e-05, "loss": 2.1706, "step": 2511 }, { "epoch": 0.06400362569712839, "grad_norm": 4.610604763031006, "learning_rate": 4.9848632707796773e-05, "loss": 2.1528, "step": 2542 }, { "epoch": 0.06478415771782509, "grad_norm": 1.4580930471420288, "learning_rate": 4.9839264870486155e-05, "loss": 2.1849, "step": 2573 }, { "epoch": 0.06556468973852177, "grad_norm": 1.5506151914596558, "learning_rate": 4.9829616725447526e-05, "loss": 2.1504, "step": 2604 }, { "epoch": 0.06634522175921846, "grad_norm": 1.4275034666061401, "learning_rate": 4.981968838155888e-05, "loss": 2.1492, "step": 2635 }, { "epoch": 0.06712575377991514, "grad_norm": 1.4749577045440674, "learning_rate": 4.980947995086024e-05, "loss": 2.1579, "step": 2666 }, { "epoch": 0.06790628580061184, "grad_norm": 1.4220107793807983, "learning_rate": 4.979899154855234e-05, "loss": 2.1522, "step": 2697 }, { "epoch": 0.06868681782130852, "grad_norm": 1.4243831634521484, "learning_rate": 4.9788223292995386e-05, "loss": 2.1258, "step": 2728 }, { "epoch": 0.06946734984200521, "grad_norm": 1.4925100803375244, "learning_rate": 4.977717530570768e-05, "loss": 2.1379, "step": 2759 }, { "epoch": 0.0702478818627019, "grad_norm": 1.4839624166488647, "learning_rate": 4.976584771136425e-05, "loss": 2.1588, "step": 2790 }, { "epoch": 0.07102841388339859, "grad_norm": 1.4130090475082397, "learning_rate": 4.975424063779547e-05, "loss": 2.1461, "step": 2821 }, { "epoch": 0.07180894590409527, "grad_norm": 1.3940472602844238, "learning_rate": 4.974235421598557e-05, "loss": 2.1416, "step": 2852 }, { "epoch": 0.07258947792479196, "grad_norm": 1.4502291679382324, "learning_rate": 4.973018858007122e-05, "loss": 2.1315, "step": 2883 }, { "epoch": 0.07337000994548865, "grad_norm": 1.4334006309509277, "learning_rate": 4.9717743867339963e-05, "loss": 2.122, "step": 2914 }, { "epoch": 0.07415054196618534, "grad_norm": 1.4366185665130615, "learning_rate": 4.9705020218228695e-05, "loss": 2.1179, "step": 2945 }, { "epoch": 0.07493107398688202, "grad_norm": 1.437264084815979, "learning_rate": 4.969201777632205e-05, "loss": 2.1171, "step": 2976 }, { "epoch": 0.07571160600757872, "grad_norm": 1.3713815212249756, "learning_rate": 4.9678736688350846e-05, "loss": 2.1215, "step": 3007 }, { "epoch": 0.0764921380282754, "grad_norm": 1.5930180549621582, "learning_rate": 4.966517710419033e-05, "loss": 2.0961, "step": 3038 }, { "epoch": 0.07727267004897209, "grad_norm": 1.4208778142929077, "learning_rate": 4.965133917685858e-05, "loss": 2.1193, "step": 3069 }, { "epoch": 0.07805320206966877, "grad_norm": 1.3563286066055298, "learning_rate": 4.9637223062514714e-05, "loss": 2.0957, "step": 3100 }, { "epoch": 0.07883373409036547, "grad_norm": 1.382861852645874, "learning_rate": 4.962282892045718e-05, "loss": 2.117, "step": 3131 }, { "epoch": 0.07961426611106215, "grad_norm": 1.326231837272644, "learning_rate": 4.9608156913121904e-05, "loss": 2.0906, "step": 3162 }, { "epoch": 0.08039479813175884, "grad_norm": 1.3465954065322876, "learning_rate": 4.959320720608049e-05, "loss": 2.09, "step": 3193 }, { "epoch": 0.08117533015245552, "grad_norm": 1.3832777738571167, "learning_rate": 4.9577979968038354e-05, "loss": 2.089, "step": 3224 }, { "epoch": 0.08195586217315222, "grad_norm": 1.4453243017196655, "learning_rate": 4.956247537083282e-05, "loss": 2.0867, "step": 3255 }, { "epoch": 0.0827363941938489, "grad_norm": 1.310168743133545, "learning_rate": 4.9546693589431145e-05, "loss": 2.0915, "step": 3286 }, { "epoch": 0.0835169262145456, "grad_norm": 1.354078769683838, "learning_rate": 4.9530634801928595e-05, "loss": 2.0888, "step": 3317 }, { "epoch": 0.08429745823524228, "grad_norm": 1.3768595457077026, "learning_rate": 4.9514299189546395e-05, "loss": 2.0821, "step": 3348 }, { "epoch": 0.08507799025593897, "grad_norm": 1.3402544260025024, "learning_rate": 4.949768693662973e-05, "loss": 2.0776, "step": 3379 }, { "epoch": 0.08585852227663565, "grad_norm": 1.3178027868270874, "learning_rate": 4.948079823064559e-05, "loss": 2.0699, "step": 3410 }, { "epoch": 0.08663905429733235, "grad_norm": 1.7057585716247559, "learning_rate": 4.946363326218074e-05, "loss": 2.0615, "step": 3441 }, { "epoch": 0.08741958631802903, "grad_norm": 1.3453648090362549, "learning_rate": 4.9446192224939525e-05, "loss": 2.0788, "step": 3472 }, { "epoch": 0.08820011833872572, "grad_norm": 1.3319638967514038, "learning_rate": 4.942847531574167e-05, "loss": 2.0619, "step": 3503 }, { "epoch": 0.0889806503594224, "grad_norm": 1.2845163345336914, "learning_rate": 4.941048273452008e-05, "loss": 2.0728, "step": 3534 }, { "epoch": 0.0897611823801191, "grad_norm": 1.3615431785583496, "learning_rate": 4.9392214684318605e-05, "loss": 2.0538, "step": 3565 }, { "epoch": 0.09054171440081578, "grad_norm": 1.2946122884750366, "learning_rate": 4.93736713712897e-05, "loss": 2.0689, "step": 3596 }, { "epoch": 0.09132224642151247, "grad_norm": 1.294846534729004, "learning_rate": 4.9354853004692124e-05, "loss": 2.0706, "step": 3627 }, { "epoch": 0.09210277844220915, "grad_norm": 1.298010230064392, "learning_rate": 4.93357597968886e-05, "loss": 2.0406, "step": 3658 }, { "epoch": 0.09288331046290585, "grad_norm": 1.7569706439971924, "learning_rate": 4.931639196334338e-05, "loss": 2.0736, "step": 3689 }, { "epoch": 0.09366384248360253, "grad_norm": 1.2595418691635132, "learning_rate": 4.9296749722619826e-05, "loss": 2.0654, "step": 3720 }, { "epoch": 0.09444437450429922, "grad_norm": 1.3139164447784424, "learning_rate": 4.9276833296377966e-05, "loss": 2.0467, "step": 3751 }, { "epoch": 0.0952249065249959, "grad_norm": 1.290247917175293, "learning_rate": 4.925664290937196e-05, "loss": 2.0418, "step": 3782 }, { "epoch": 0.0960054385456926, "grad_norm": 1.2755826711654663, "learning_rate": 4.9236178789447576e-05, "loss": 2.0244, "step": 3813 }, { "epoch": 0.09678597056638928, "grad_norm": 1.2596077919006348, "learning_rate": 4.921544116753962e-05, "loss": 2.0609, "step": 3844 }, { "epoch": 0.09756650258708598, "grad_norm": 1.249387264251709, "learning_rate": 4.919443027766935e-05, "loss": 2.0455, "step": 3875 }, { "epoch": 0.09834703460778266, "grad_norm": 1.255379319190979, "learning_rate": 4.91731463569418e-05, "loss": 2.024, "step": 3906 }, { "epoch": 0.09912756662847935, "grad_norm": 1.2620748281478882, "learning_rate": 4.915158964554312e-05, "loss": 2.0355, "step": 3937 }, { "epoch": 0.09990809864917603, "grad_norm": 1.2736345529556274, "learning_rate": 4.912976038673786e-05, "loss": 2.0351, "step": 3968 }, { "epoch": 0.10068863066987273, "grad_norm": 1.18524169921875, "learning_rate": 4.9107658826866254e-05, "loss": 2.029, "step": 3999 }, { "epoch": 0.10146916269056941, "grad_norm": 10.429583549499512, "learning_rate": 4.908528521534139e-05, "loss": 2.0226, "step": 4030 }, { "epoch": 0.1022496947112661, "grad_norm": 1.240295648574829, "learning_rate": 4.906263980464644e-05, "loss": 2.0401, "step": 4061 }, { "epoch": 0.10303022673196278, "grad_norm": 1.2191627025604248, "learning_rate": 4.903972285033178e-05, "loss": 2.0079, "step": 4092 }, { "epoch": 0.10381075875265948, "grad_norm": 1.2340337038040161, "learning_rate": 4.901653461101213e-05, "loss": 2.0305, "step": 4123 }, { "epoch": 0.10459129077335616, "grad_norm": 1.284839153289795, "learning_rate": 4.8993075348363626e-05, "loss": 2.0212, "step": 4154 }, { "epoch": 0.10537182279405286, "grad_norm": 1.2955284118652344, "learning_rate": 4.896934532712084e-05, "loss": 2.0328, "step": 4185 }, { "epoch": 0.10615235481474954, "grad_norm": 1.165124535560608, "learning_rate": 4.8945344815073846e-05, "loss": 2.0248, "step": 4216 }, { "epoch": 0.10693288683544623, "grad_norm": 1.1925835609436035, "learning_rate": 4.892107408306516e-05, "loss": 2.0086, "step": 4247 }, { "epoch": 0.10771341885614291, "grad_norm": 1.1841633319854736, "learning_rate": 4.889653340498669e-05, "loss": 2.0166, "step": 4278 }, { "epoch": 0.1084939508768396, "grad_norm": 1.2198415994644165, "learning_rate": 4.8871723057776664e-05, "loss": 2.0164, "step": 4309 }, { "epoch": 0.10927448289753629, "grad_norm": 1.2284913063049316, "learning_rate": 4.8846643321416476e-05, "loss": 2.015, "step": 4340 }, { "epoch": 0.11005501491823298, "grad_norm": 1.2090349197387695, "learning_rate": 4.882129447892753e-05, "loss": 2.0094, "step": 4371 }, { "epoch": 0.11083554693892966, "grad_norm": 1.215232491493225, "learning_rate": 4.8795676816368076e-05, "loss": 2.0045, "step": 4402 }, { "epoch": 0.11161607895962636, "grad_norm": 1.2804045677185059, "learning_rate": 4.876979062282995e-05, "loss": 1.9899, "step": 4433 }, { "epoch": 0.11239661098032304, "grad_norm": 1.2405016422271729, "learning_rate": 4.8743636190435325e-05, "loss": 2.0093, "step": 4464 }, { "epoch": 0.11317714300101973, "grad_norm": 1.1784151792526245, "learning_rate": 4.871721381433344e-05, "loss": 2.0036, "step": 4495 }, { "epoch": 0.11395767502171641, "grad_norm": 1.185738205909729, "learning_rate": 4.869052379269719e-05, "loss": 2.0033, "step": 4526 }, { "epoch": 0.11473820704241311, "grad_norm": 1.1984614133834839, "learning_rate": 4.866356642671985e-05, "loss": 2.025, "step": 4557 }, { "epoch": 0.11551873906310979, "grad_norm": 1.1734029054641724, "learning_rate": 4.8636342020611634e-05, "loss": 2.0023, "step": 4588 }, { "epoch": 0.11629927108380649, "grad_norm": 1.198611855506897, "learning_rate": 4.860885088159626e-05, "loss": 2.0031, "step": 4619 }, { "epoch": 0.11707980310450317, "grad_norm": 1.1793649196624756, "learning_rate": 4.858109331990751e-05, "loss": 1.9806, "step": 4650 }, { "epoch": 0.11786033512519985, "grad_norm": 1.1945818662643433, "learning_rate": 4.855306964878567e-05, "loss": 1.9855, "step": 4681 }, { "epoch": 0.11864086714589654, "grad_norm": 1.1875253915786743, "learning_rate": 4.8524780184474084e-05, "loss": 1.9883, "step": 4712 }, { "epoch": 0.11942139916659322, "grad_norm": 1.1512328386306763, "learning_rate": 4.8496225246215496e-05, "loss": 1.9807, "step": 4743 }, { "epoch": 0.12020193118728992, "grad_norm": 1.2059484720230103, "learning_rate": 4.8467405156248505e-05, "loss": 2.0103, "step": 4774 }, { "epoch": 0.1209824632079866, "grad_norm": 1.1759779453277588, "learning_rate": 4.843832023980392e-05, "loss": 1.9754, "step": 4805 }, { "epoch": 0.1217629952286833, "grad_norm": 1.2004414796829224, "learning_rate": 4.840897082510106e-05, "loss": 1.9893, "step": 4836 }, { "epoch": 0.12254352724937997, "grad_norm": 1.1744251251220703, "learning_rate": 4.8379357243344084e-05, "loss": 1.9848, "step": 4867 }, { "epoch": 0.12332405927007667, "grad_norm": 1.1731071472167969, "learning_rate": 4.8349479828718236e-05, "loss": 1.9945, "step": 4898 }, { "epoch": 0.12410459129077335, "grad_norm": 1.2002074718475342, "learning_rate": 4.8319338918386075e-05, "loss": 1.9869, "step": 4929 }, { "epoch": 0.12488512331147004, "grad_norm": 1.147085189819336, "learning_rate": 4.828893485248369e-05, "loss": 1.9876, "step": 4960 }, { "epoch": 0.12566565533216673, "grad_norm": 1.1605637073516846, "learning_rate": 4.825826797411682e-05, "loss": 1.9858, "step": 4991 }, { "epoch": 0.12644618735286342, "grad_norm": 1.1823869943618774, "learning_rate": 4.822733862935702e-05, "loss": 1.9617, "step": 5022 }, { "epoch": 0.12722671937356012, "grad_norm": 1.1143817901611328, "learning_rate": 4.819614716723775e-05, "loss": 1.9756, "step": 5053 }, { "epoch": 0.12800725139425678, "grad_norm": 1.1646877527236938, "learning_rate": 4.8164693939750425e-05, "loss": 2.0005, "step": 5084 }, { "epoch": 0.12878778341495348, "grad_norm": 1.1390529870986938, "learning_rate": 4.813297930184042e-05, "loss": 1.9943, "step": 5115 }, { "epoch": 0.12956831543565017, "grad_norm": 1.1759583950042725, "learning_rate": 4.810100361140314e-05, "loss": 1.9772, "step": 5146 }, { "epoch": 0.13034884745634687, "grad_norm": 1.1465727090835571, "learning_rate": 4.8068767229279885e-05, "loss": 1.9563, "step": 5177 }, { "epoch": 0.13112937947704353, "grad_norm": 1.2573471069335938, "learning_rate": 4.8036270519253854e-05, "loss": 1.9735, "step": 5208 }, { "epoch": 0.13190991149774023, "grad_norm": 1.1301871538162231, "learning_rate": 4.8003513848046e-05, "loss": 1.987, "step": 5239 }, { "epoch": 0.13269044351843692, "grad_norm": 1.1424096822738647, "learning_rate": 4.79704975853109e-05, "loss": 1.9661, "step": 5270 }, { "epoch": 0.13347097553913362, "grad_norm": 1.131800889968872, "learning_rate": 4.793722210363262e-05, "loss": 1.9828, "step": 5301 }, { "epoch": 0.13425150755983029, "grad_norm": 1.1227242946624756, "learning_rate": 4.7903687778520414e-05, "loss": 1.975, "step": 5332 }, { "epoch": 0.13503203958052698, "grad_norm": 1.143832802772522, "learning_rate": 4.7869894988404593e-05, "loss": 1.9619, "step": 5363 }, { "epoch": 0.13581257160122367, "grad_norm": 1.1281683444976807, "learning_rate": 4.783584411463221e-05, "loss": 1.9603, "step": 5394 }, { "epoch": 0.13659310362192037, "grad_norm": 1.134901762008667, "learning_rate": 4.780153554146274e-05, "loss": 1.9888, "step": 5425 }, { "epoch": 0.13737363564261704, "grad_norm": 1.1317532062530518, "learning_rate": 4.7766969656063766e-05, "loss": 1.9793, "step": 5456 }, { "epoch": 0.13815416766331373, "grad_norm": 1.2516345977783203, "learning_rate": 4.773214684850662e-05, "loss": 1.9839, "step": 5487 }, { "epoch": 0.13893469968401043, "grad_norm": 1.107391119003296, "learning_rate": 4.769706751176193e-05, "loss": 1.9606, "step": 5518 }, { "epoch": 0.13971523170470712, "grad_norm": 1.1041492223739624, "learning_rate": 4.7661732041695264e-05, "loss": 1.9688, "step": 5549 }, { "epoch": 0.1404957637254038, "grad_norm": 1.1099929809570312, "learning_rate": 4.762614083706258e-05, "loss": 1.9821, "step": 5580 }, { "epoch": 0.14127629574610048, "grad_norm": 1.167151927947998, "learning_rate": 4.759029429950581e-05, "loss": 1.9669, "step": 5611 }, { "epoch": 0.14205682776679718, "grad_norm": 1.0942577123641968, "learning_rate": 4.7554192833548235e-05, "loss": 1.9565, "step": 5642 }, { "epoch": 0.14283735978749387, "grad_norm": 1.0859640836715698, "learning_rate": 4.751783684659e-05, "loss": 1.9634, "step": 5673 }, { "epoch": 0.14361789180819054, "grad_norm": 2.420607089996338, "learning_rate": 4.748122674890348e-05, "loss": 1.9365, "step": 5704 }, { "epoch": 0.14439842382888723, "grad_norm": 1.0996880531311035, "learning_rate": 4.7444362953628654e-05, "loss": 1.9718, "step": 5735 }, { "epoch": 0.14517895584958393, "grad_norm": 1.1183338165283203, "learning_rate": 4.7407245876768424e-05, "loss": 1.9564, "step": 5766 }, { "epoch": 0.14595948787028062, "grad_norm": 1.1093947887420654, "learning_rate": 4.736987593718397e-05, "loss": 1.9516, "step": 5797 }, { "epoch": 0.1467400198909773, "grad_norm": 1.141797423362732, "learning_rate": 4.733225355658999e-05, "loss": 1.929, "step": 5828 }, { "epoch": 0.14752055191167399, "grad_norm": 1.1214358806610107, "learning_rate": 4.7294379159549926e-05, "loss": 1.9297, "step": 5859 }, { "epoch": 0.14830108393237068, "grad_norm": 1.1025431156158447, "learning_rate": 4.725625317347119e-05, "loss": 1.9226, "step": 5890 }, { "epoch": 0.14908161595306738, "grad_norm": 1.1094226837158203, "learning_rate": 4.7217876028600374e-05, "loss": 1.946, "step": 5921 }, { "epoch": 0.14986214797376404, "grad_norm": 1.0909287929534912, "learning_rate": 4.717924815801832e-05, "loss": 1.9511, "step": 5952 }, { "epoch": 0.15064267999446074, "grad_norm": 1.0755189657211304, "learning_rate": 4.714036999763532e-05, "loss": 1.939, "step": 5983 }, { "epoch": 0.15142321201515743, "grad_norm": 1.1426901817321777, "learning_rate": 4.7101241986186116e-05, "loss": 1.93, "step": 6014 }, { "epoch": 0.15220374403585413, "grad_norm": 1.1208311319351196, "learning_rate": 4.7061864565225e-05, "loss": 1.9516, "step": 6045 }, { "epoch": 0.1529842760565508, "grad_norm": 1.0931717157363892, "learning_rate": 4.702223817912081e-05, "loss": 1.9314, "step": 6076 }, { "epoch": 0.1537648080772475, "grad_norm": 1.075864553451538, "learning_rate": 4.698236327505195e-05, "loss": 1.935, "step": 6107 }, { "epoch": 0.15454534009794418, "grad_norm": 1.090559720993042, "learning_rate": 4.694224030300127e-05, "loss": 1.9431, "step": 6138 }, { "epoch": 0.15532587211864088, "grad_norm": 1.0956670045852661, "learning_rate": 4.690186971575107e-05, "loss": 1.9678, "step": 6169 }, { "epoch": 0.15610640413933755, "grad_norm": 1.125927448272705, "learning_rate": 4.6861251968877916e-05, "loss": 1.9292, "step": 6200 }, { "epoch": 0.15688693616003424, "grad_norm": 1.0844208002090454, "learning_rate": 4.68203875207476e-05, "loss": 1.9189, "step": 6231 }, { "epoch": 0.15766746818073094, "grad_norm": 1.064327597618103, "learning_rate": 4.677927683250983e-05, "loss": 1.928, "step": 6262 }, { "epoch": 0.15844800020142763, "grad_norm": 1.1281861066818237, "learning_rate": 4.6737920368093156e-05, "loss": 1.9437, "step": 6293 }, { "epoch": 0.1592285322221243, "grad_norm": 1.0674384832382202, "learning_rate": 4.669631859419965e-05, "loss": 1.9347, "step": 6324 }, { "epoch": 0.160009064242821, "grad_norm": 1.105566382408142, "learning_rate": 4.6654471980299676e-05, "loss": 1.9229, "step": 6355 }, { "epoch": 0.1607895962635177, "grad_norm": 1.0582126379013062, "learning_rate": 4.661238099862658e-05, "loss": 1.9321, "step": 6386 }, { "epoch": 0.16157012828421438, "grad_norm": 1.1147974729537964, "learning_rate": 4.657004612417138e-05, "loss": 1.9289, "step": 6417 }, { "epoch": 0.16235066030491105, "grad_norm": 1.0804343223571777, "learning_rate": 4.6527467834677374e-05, "loss": 1.9307, "step": 6448 }, { "epoch": 0.16313119232560774, "grad_norm": 1.0830508470535278, "learning_rate": 4.648464661063478e-05, "loss": 1.9408, "step": 6479 }, { "epoch": 0.16391172434630444, "grad_norm": 1.0533562898635864, "learning_rate": 4.6441582935275264e-05, "loss": 1.9347, "step": 6510 }, { "epoch": 0.16469225636700113, "grad_norm": 1.044925332069397, "learning_rate": 4.6398277294566586e-05, "loss": 1.9112, "step": 6541 }, { "epoch": 0.1654727883876978, "grad_norm": 1.087296962738037, "learning_rate": 4.6354730177207e-05, "loss": 1.9381, "step": 6572 }, { "epoch": 0.1662533204083945, "grad_norm": 1.0853668451309204, "learning_rate": 4.6310942074619787e-05, "loss": 1.9224, "step": 6603 }, { "epoch": 0.1670338524290912, "grad_norm": 1.0661752223968506, "learning_rate": 4.626691348094777e-05, "loss": 1.9386, "step": 6634 }, { "epoch": 0.16781438444978788, "grad_norm": 1.0393208265304565, "learning_rate": 4.622264489304762e-05, "loss": 1.9266, "step": 6665 }, { "epoch": 0.16859491647048455, "grad_norm": 1.0324763059616089, "learning_rate": 4.617813681048434e-05, "loss": 1.919, "step": 6696 }, { "epoch": 0.16937544849118125, "grad_norm": 1.0684878826141357, "learning_rate": 4.61333897355256e-05, "loss": 1.9321, "step": 6727 }, { "epoch": 0.17015598051187794, "grad_norm": 1.0477505922317505, "learning_rate": 4.608840417313604e-05, "loss": 1.9449, "step": 6758 }, { "epoch": 0.17093651253257464, "grad_norm": 1.0684869289398193, "learning_rate": 4.6043180630971646e-05, "loss": 1.9252, "step": 6789 }, { "epoch": 0.1717170445532713, "grad_norm": 1.0869554281234741, "learning_rate": 4.599771961937391e-05, "loss": 1.9183, "step": 6820 }, { "epoch": 0.172497576573968, "grad_norm": 1.0297125577926636, "learning_rate": 4.5952021651364204e-05, "loss": 1.9123, "step": 6851 }, { "epoch": 0.1732781085946647, "grad_norm": 1.0670006275177002, "learning_rate": 4.590608724263786e-05, "loss": 1.9353, "step": 6882 }, { "epoch": 0.1740586406153614, "grad_norm": 1.0744417905807495, "learning_rate": 4.585991691155845e-05, "loss": 1.8979, "step": 6913 }, { "epoch": 0.17483917263605805, "grad_norm": 1.0910695791244507, "learning_rate": 4.581351117915188e-05, "loss": 1.9153, "step": 6944 }, { "epoch": 0.17561970465675475, "grad_norm": 1.083776593208313, "learning_rate": 4.5766870569100534e-05, "loss": 1.9026, "step": 6975 }, { "epoch": 0.17640023667745144, "grad_norm": 1.0857270956039429, "learning_rate": 4.571999560773736e-05, "loss": 1.9279, "step": 7006 }, { "epoch": 0.1771807686981481, "grad_norm": 1.0350106954574585, "learning_rate": 4.5672886824039915e-05, "loss": 1.9236, "step": 7037 }, { "epoch": 0.1779613007188448, "grad_norm": 1.0498634576797485, "learning_rate": 4.5625544749624435e-05, "loss": 1.8934, "step": 7068 }, { "epoch": 0.1787418327395415, "grad_norm": 1.0836471319198608, "learning_rate": 4.5577969918739794e-05, "loss": 1.9131, "step": 7099 }, { "epoch": 0.1795223647602382, "grad_norm": 1.0414760112762451, "learning_rate": 4.5530162868261486e-05, "loss": 1.8792, "step": 7130 }, { "epoch": 0.18030289678093486, "grad_norm": 1.0747581720352173, "learning_rate": 4.548212413768558e-05, "loss": 1.8973, "step": 7161 }, { "epoch": 0.18108342880163156, "grad_norm": 1.0464451313018799, "learning_rate": 4.543385426912261e-05, "loss": 1.907, "step": 7192 }, { "epoch": 0.18186396082232825, "grad_norm": 1.0159403085708618, "learning_rate": 4.53853538072915e-05, "loss": 1.9111, "step": 7223 }, { "epoch": 0.18264449284302495, "grad_norm": 1.093883991241455, "learning_rate": 4.533662329951336e-05, "loss": 1.8927, "step": 7254 }, { "epoch": 0.18342502486372161, "grad_norm": 1.0358048677444458, "learning_rate": 4.528766329570536e-05, "loss": 1.9164, "step": 7285 }, { "epoch": 0.1842055568844183, "grad_norm": 1.051798939704895, "learning_rate": 4.523847434837447e-05, "loss": 1.9084, "step": 7316 }, { "epoch": 0.184986088905115, "grad_norm": 1.0729514360427856, "learning_rate": 4.518905701261128e-05, "loss": 1.9089, "step": 7347 }, { "epoch": 0.1857666209258117, "grad_norm": 1.0399775505065918, "learning_rate": 4.5139411846083715e-05, "loss": 1.922, "step": 7378 }, { "epoch": 0.18654715294650837, "grad_norm": 1.0500715970993042, "learning_rate": 4.508953940903073e-05, "loss": 1.9118, "step": 7409 }, { "epoch": 0.18732768496720506, "grad_norm": 1.0770058631896973, "learning_rate": 4.5039440264255994e-05, "loss": 1.9059, "step": 7440 }, { "epoch": 0.18810821698790176, "grad_norm": 1.0471183061599731, "learning_rate": 4.498911497712155e-05, "loss": 1.9119, "step": 7471 }, { "epoch": 0.18888874900859845, "grad_norm": 1.038761019706726, "learning_rate": 4.493856411554142e-05, "loss": 1.8798, "step": 7502 }, { "epoch": 0.18966928102929512, "grad_norm": 1.0365252494812012, "learning_rate": 4.4887788249975206e-05, "loss": 1.8983, "step": 7533 }, { "epoch": 0.1904498130499918, "grad_norm": 1.1033800840377808, "learning_rate": 4.4836787953421656e-05, "loss": 1.8911, "step": 7564 }, { "epoch": 0.1912303450706885, "grad_norm": 1.064213514328003, "learning_rate": 4.478556380141218e-05, "loss": 1.8995, "step": 7595 }, { "epoch": 0.1920108770913852, "grad_norm": 0.997905969619751, "learning_rate": 4.4734116372004375e-05, "loss": 1.8705, "step": 7626 }, { "epoch": 0.19279140911208187, "grad_norm": 1.0392402410507202, "learning_rate": 4.4682446245775477e-05, "loss": 1.9105, "step": 7657 }, { "epoch": 0.19357194113277856, "grad_norm": 1.0340332984924316, "learning_rate": 4.463055400581586e-05, "loss": 1.9074, "step": 7688 }, { "epoch": 0.19435247315347526, "grad_norm": 1.0196648836135864, "learning_rate": 4.4578440237722374e-05, "loss": 1.9059, "step": 7719 }, { "epoch": 0.19513300517417195, "grad_norm": 0.99578458070755, "learning_rate": 4.452610552959183e-05, "loss": 1.8974, "step": 7750 }, { "epoch": 0.19591353719486862, "grad_norm": 1.026090145111084, "learning_rate": 4.447355047201428e-05, "loss": 1.8886, "step": 7781 }, { "epoch": 0.19669406921556531, "grad_norm": 0.9945096969604492, "learning_rate": 4.4420775658066414e-05, "loss": 1.8879, "step": 7812 }, { "epoch": 0.197474601236262, "grad_norm": 1.0226850509643555, "learning_rate": 4.436778168330484e-05, "loss": 1.8936, "step": 7843 }, { "epoch": 0.1982551332569587, "grad_norm": 1.0433967113494873, "learning_rate": 4.4314569145759353e-05, "loss": 1.882, "step": 7874 }, { "epoch": 0.19903566527765537, "grad_norm": 1.0612221956253052, "learning_rate": 4.42611386459262e-05, "loss": 1.8879, "step": 7905 }, { "epoch": 0.19981619729835207, "grad_norm": 1.0252665281295776, "learning_rate": 4.420749078676133e-05, "loss": 1.9005, "step": 7936 }, { "epoch": 0.20059672931904876, "grad_norm": 1.015100359916687, "learning_rate": 4.4153626173673516e-05, "loss": 1.8944, "step": 7967 }, { "epoch": 0.20137726133974546, "grad_norm": 1.0417771339416504, "learning_rate": 4.409954541451762e-05, "loss": 1.8837, "step": 7998 }, { "epoch": 0.20215779336044212, "grad_norm": 1.026708960533142, "learning_rate": 4.404524911958764e-05, "loss": 1.92, "step": 8029 }, { "epoch": 0.20293832538113882, "grad_norm": 1.0414159297943115, "learning_rate": 4.399073790160989e-05, "loss": 1.8709, "step": 8060 }, { "epoch": 0.2037188574018355, "grad_norm": 1.014039397239685, "learning_rate": 4.393601237573607e-05, "loss": 1.8842, "step": 8091 }, { "epoch": 0.2044993894225322, "grad_norm": 1.0103037357330322, "learning_rate": 4.388107315953628e-05, "loss": 1.8889, "step": 8122 }, { "epoch": 0.20527992144322887, "grad_norm": 1.0166395902633667, "learning_rate": 4.382592087299212e-05, "loss": 1.8808, "step": 8153 }, { "epoch": 0.20606045346392557, "grad_norm": 1.023850679397583, "learning_rate": 4.377055613848964e-05, "loss": 1.8918, "step": 8184 }, { "epoch": 0.20684098548462226, "grad_norm": 1.0251572132110596, "learning_rate": 4.3714979580812355e-05, "loss": 1.882, "step": 8215 }, { "epoch": 0.20762151750531896, "grad_norm": 1.0183281898498535, "learning_rate": 4.365919182713416e-05, "loss": 1.8844, "step": 8246 }, { "epoch": 0.20840204952601563, "grad_norm": 1.0135486125946045, "learning_rate": 4.360319350701226e-05, "loss": 1.8814, "step": 8277 }, { "epoch": 0.20918258154671232, "grad_norm": 1.0247812271118164, "learning_rate": 4.3546985252380115e-05, "loss": 1.8818, "step": 8308 }, { "epoch": 0.20996311356740902, "grad_norm": 1.0123242139816284, "learning_rate": 4.349056769754021e-05, "loss": 1.8723, "step": 8339 }, { "epoch": 0.2107436455881057, "grad_norm": 1.0139250755310059, "learning_rate": 4.3433941479156994e-05, "loss": 1.8858, "step": 8370 }, { "epoch": 0.21152417760880238, "grad_norm": 4.008477687835693, "learning_rate": 4.3377107236249647e-05, "loss": 1.8735, "step": 8401 }, { "epoch": 0.21230470962949907, "grad_norm": 1.0454871654510498, "learning_rate": 4.332006561018488e-05, "loss": 1.8824, "step": 8432 }, { "epoch": 0.21308524165019577, "grad_norm": 1.0586799383163452, "learning_rate": 4.3262817244669683e-05, "loss": 1.8992, "step": 8463 }, { "epoch": 0.21386577367089246, "grad_norm": 1.0073469877243042, "learning_rate": 4.3205362785744083e-05, "loss": 1.8766, "step": 8494 }, { "epoch": 0.21464630569158913, "grad_norm": 0.9976999759674072, "learning_rate": 4.314770288177384e-05, "loss": 1.8864, "step": 8525 }, { "epoch": 0.21542683771228582, "grad_norm": 0.9928112030029297, "learning_rate": 4.308983818344313e-05, "loss": 1.8874, "step": 8556 }, { "epoch": 0.21620736973298252, "grad_norm": 1.038405418395996, "learning_rate": 4.3031769343747206e-05, "loss": 1.8725, "step": 8587 }, { "epoch": 0.2169879017536792, "grad_norm": 1.0328835248947144, "learning_rate": 4.297349701798505e-05, "loss": 1.8884, "step": 8618 }, { "epoch": 0.21776843377437588, "grad_norm": 1.0254933834075928, "learning_rate": 4.2915021863751916e-05, "loss": 1.8611, "step": 8649 }, { "epoch": 0.21854896579507258, "grad_norm": 0.99531090259552, "learning_rate": 4.285634454093198e-05, "loss": 1.8889, "step": 8680 }, { "epoch": 0.21932949781576927, "grad_norm": 1.0536468029022217, "learning_rate": 4.279746571169086e-05, "loss": 1.8875, "step": 8711 }, { "epoch": 0.22011002983646596, "grad_norm": 0.9915075898170471, "learning_rate": 4.2738386040468136e-05, "loss": 1.8901, "step": 8742 }, { "epoch": 0.22089056185716263, "grad_norm": 0.9883093237876892, "learning_rate": 4.2679106193969866e-05, "loss": 1.8704, "step": 8773 }, { "epoch": 0.22167109387785933, "grad_norm": 1.0425434112548828, "learning_rate": 4.261962684116106e-05, "loss": 1.8643, "step": 8804 }, { "epoch": 0.22245162589855602, "grad_norm": 1.0154091119766235, "learning_rate": 4.2559948653258145e-05, "loss": 1.8806, "step": 8835 }, { "epoch": 0.22323215791925272, "grad_norm": 1.0432935953140259, "learning_rate": 4.250007230372134e-05, "loss": 1.8928, "step": 8866 }, { "epoch": 0.22401268993994938, "grad_norm": 0.9709262251853943, "learning_rate": 4.2439998468247126e-05, "loss": 1.8591, "step": 8897 }, { "epoch": 0.22479322196064608, "grad_norm": 0.9708088636398315, "learning_rate": 4.2379727824760566e-05, "loss": 1.8797, "step": 8928 }, { "epoch": 0.22557375398134277, "grad_norm": 0.9911683201789856, "learning_rate": 4.231926105340768e-05, "loss": 1.8456, "step": 8959 }, { "epoch": 0.22635428600203947, "grad_norm": 0.99161297082901, "learning_rate": 4.225859883654776e-05, "loss": 1.8712, "step": 8990 }, { "epoch": 0.22713481802273613, "grad_norm": 1.0025807619094849, "learning_rate": 4.219774185874569e-05, "loss": 1.8814, "step": 9021 }, { "epoch": 0.22791535004343283, "grad_norm": 0.9597026109695435, "learning_rate": 4.213669080676418e-05, "loss": 1.8799, "step": 9052 }, { "epoch": 0.22869588206412952, "grad_norm": 1.0001273155212402, "learning_rate": 4.2075446369556056e-05, "loss": 1.8469, "step": 9083 }, { "epoch": 0.22947641408482622, "grad_norm": 0.9670659303665161, "learning_rate": 4.201400923825648e-05, "loss": 1.8612, "step": 9114 }, { "epoch": 0.23025694610552289, "grad_norm": 1.018050193786621, "learning_rate": 4.195238010617511e-05, "loss": 1.8656, "step": 9145 }, { "epoch": 0.23103747812621958, "grad_norm": 0.9838584661483765, "learning_rate": 4.1890559668788344e-05, "loss": 1.8527, "step": 9176 }, { "epoch": 0.23181801014691628, "grad_norm": 0.9968937635421753, "learning_rate": 4.1828548623731405e-05, "loss": 1.8753, "step": 9207 }, { "epoch": 0.23259854216761297, "grad_norm": 0.9698464870452881, "learning_rate": 4.1766347670790506e-05, "loss": 1.8716, "step": 9238 }, { "epoch": 0.23337907418830964, "grad_norm": 0.9592704772949219, "learning_rate": 4.170395751189495e-05, "loss": 1.8698, "step": 9269 }, { "epoch": 0.23415960620900633, "grad_norm": 0.9904705286026001, "learning_rate": 4.164137885110921e-05, "loss": 1.8661, "step": 9300 }, { "epoch": 0.23494013822970303, "grad_norm": 1.0243868827819824, "learning_rate": 4.157861239462495e-05, "loss": 1.8658, "step": 9331 }, { "epoch": 0.2357206702503997, "grad_norm": 0.9936602711677551, "learning_rate": 4.1515658850753114e-05, "loss": 1.8475, "step": 9362 }, { "epoch": 0.2365012022710964, "grad_norm": 0.9626069068908691, "learning_rate": 4.145251892991588e-05, "loss": 1.8507, "step": 9393 }, { "epoch": 0.23728173429179308, "grad_norm": 0.9688258767127991, "learning_rate": 4.138919334463868e-05, "loss": 1.8628, "step": 9424 }, { "epoch": 0.23806226631248978, "grad_norm": 0.9996054172515869, "learning_rate": 4.1325682809542124e-05, "loss": 1.8734, "step": 9455 }, { "epoch": 0.23884279833318645, "grad_norm": 1.0037505626678467, "learning_rate": 4.126198804133398e-05, "loss": 1.8778, "step": 9486 }, { "epoch": 0.23962333035388314, "grad_norm": 0.9751474857330322, "learning_rate": 4.1198109758801055e-05, "loss": 1.8615, "step": 9517 }, { "epoch": 0.24040386237457984, "grad_norm": 0.9760595560073853, "learning_rate": 4.113404868280107e-05, "loss": 1.8614, "step": 9548 }, { "epoch": 0.24118439439527653, "grad_norm": 1.0009572505950928, "learning_rate": 4.106980553625457e-05, "loss": 1.8762, "step": 9579 }, { "epoch": 0.2419649264159732, "grad_norm": 0.983182966709137, "learning_rate": 4.100538104413674e-05, "loss": 1.857, "step": 9610 }, { "epoch": 0.2427454584366699, "grad_norm": 1.0748534202575684, "learning_rate": 4.09407759334692e-05, "loss": 1.8658, "step": 9641 }, { "epoch": 0.2435259904573666, "grad_norm": 0.9747381806373596, "learning_rate": 4.087599093331186e-05, "loss": 1.8484, "step": 9672 }, { "epoch": 0.24430652247806328, "grad_norm": 0.9784808158874512, "learning_rate": 4.081102677475462e-05, "loss": 1.8492, "step": 9703 }, { "epoch": 0.24508705449875995, "grad_norm": 0.9820215106010437, "learning_rate": 4.0745884190909194e-05, "loss": 1.8516, "step": 9734 }, { "epoch": 0.24586758651945664, "grad_norm": 0.9536774158477783, "learning_rate": 4.0680563916900796e-05, "loss": 1.8444, "step": 9765 }, { "epoch": 0.24664811854015334, "grad_norm": 0.9706370234489441, "learning_rate": 4.0615066689859815e-05, "loss": 1.8477, "step": 9796 }, { "epoch": 0.24742865056085003, "grad_norm": 0.9700275659561157, "learning_rate": 4.0549393248913584e-05, "loss": 1.8515, "step": 9827 }, { "epoch": 0.2482091825815467, "grad_norm": 0.9522431492805481, "learning_rate": 4.048354433517794e-05, "loss": 1.8287, "step": 9858 }, { "epoch": 0.2489897146022434, "grad_norm": 0.9691547155380249, "learning_rate": 4.0417520691748916e-05, "loss": 1.8594, "step": 9889 }, { "epoch": 0.2497702466229401, "grad_norm": 0.9588534235954285, "learning_rate": 4.035132306369438e-05, "loss": 1.8459, "step": 9920 }, { "epoch": 0.2505507786436368, "grad_norm": 0.9900926351547241, "learning_rate": 4.028495219804555e-05, "loss": 1.8608, "step": 9951 }, { "epoch": 0.25133131066433345, "grad_norm": 0.9539284110069275, "learning_rate": 4.021840884378864e-05, "loss": 1.8522, "step": 9982 }, { "epoch": 0.2521118426850302, "grad_norm": 0.9557245969772339, "learning_rate": 4.015169375185633e-05, "loss": 1.8743, "step": 10013 }, { "epoch": 0.25289237470572684, "grad_norm": 0.9317564964294434, "learning_rate": 4.0084807675119396e-05, "loss": 1.8384, "step": 10044 }, { "epoch": 0.2536729067264235, "grad_norm": 0.9944160580635071, "learning_rate": 4.0017751368378106e-05, "loss": 1.857, "step": 10075 }, { "epoch": 0.25445343874712023, "grad_norm": 0.9859464168548584, "learning_rate": 3.995052558835377e-05, "loss": 1.8484, "step": 10106 }, { "epoch": 0.2552339707678169, "grad_norm": 1.0093594789505005, "learning_rate": 3.988313109368017e-05, "loss": 1.8506, "step": 10137 }, { "epoch": 0.25601450278851356, "grad_norm": 0.9451240301132202, "learning_rate": 3.981556864489504e-05, "loss": 1.8732, "step": 10168 }, { "epoch": 0.2567950348092103, "grad_norm": 0.972819447517395, "learning_rate": 3.974783900443142e-05, "loss": 1.856, "step": 10199 }, { "epoch": 0.25757556682990695, "grad_norm": 0.996379554271698, "learning_rate": 3.9679942936609095e-05, "loss": 1.8299, "step": 10230 }, { "epoch": 0.2583560988506037, "grad_norm": 2.503967046737671, "learning_rate": 3.961188120762596e-05, "loss": 1.8484, "step": 10261 }, { "epoch": 0.25913663087130034, "grad_norm": 0.9491912126541138, "learning_rate": 3.954365458554938e-05, "loss": 1.8528, "step": 10292 }, { "epoch": 0.259917162891997, "grad_norm": 0.9642863273620605, "learning_rate": 3.947526384030751e-05, "loss": 1.863, "step": 10323 }, { "epoch": 0.26069769491269373, "grad_norm": 0.9915285706520081, "learning_rate": 3.9406709743680624e-05, "loss": 1.8191, "step": 10354 }, { "epoch": 0.2614782269333904, "grad_norm": 1.0176483392715454, "learning_rate": 3.9337993069292366e-05, "loss": 1.8396, "step": 10385 }, { "epoch": 0.26225875895408707, "grad_norm": 0.9567158222198486, "learning_rate": 3.926911459260109e-05, "loss": 1.8617, "step": 10416 }, { "epoch": 0.2630392909747838, "grad_norm": 0.9236714839935303, "learning_rate": 3.920007509089102e-05, "loss": 1.8362, "step": 10447 }, { "epoch": 0.26381982299548046, "grad_norm": 0.9428004026412964, "learning_rate": 3.913087534326357e-05, "loss": 1.8447, "step": 10478 }, { "epoch": 0.2646003550161772, "grad_norm": 0.941067099571228, "learning_rate": 3.9061516130628475e-05, "loss": 1.8329, "step": 10509 }, { "epoch": 0.26538088703687385, "grad_norm": 0.9685031175613403, "learning_rate": 3.8991998235695025e-05, "loss": 1.8543, "step": 10540 }, { "epoch": 0.2661614190575705, "grad_norm": 0.9629632234573364, "learning_rate": 3.8922322442963224e-05, "loss": 1.8118, "step": 10571 }, { "epoch": 0.26694195107826724, "grad_norm": 0.9536260366439819, "learning_rate": 3.885248953871491e-05, "loss": 1.8379, "step": 10602 }, { "epoch": 0.2677224830989639, "grad_norm": 0.9781749248504639, "learning_rate": 3.8782500311004915e-05, "loss": 1.8419, "step": 10633 }, { "epoch": 0.26850301511966057, "grad_norm": 0.9313937425613403, "learning_rate": 3.871235554965218e-05, "loss": 1.8467, "step": 10664 }, { "epoch": 0.2692835471403573, "grad_norm": 0.9775392413139343, "learning_rate": 3.864205604623078e-05, "loss": 1.8209, "step": 10695 }, { "epoch": 0.27006407916105396, "grad_norm": 0.965977132320404, "learning_rate": 3.857160259406107e-05, "loss": 1.8558, "step": 10726 }, { "epoch": 0.2708446111817507, "grad_norm": 0.9789881706237793, "learning_rate": 3.8500995988200674e-05, "loss": 1.8523, "step": 10757 }, { "epoch": 0.27162514320244735, "grad_norm": 0.9083163738250732, "learning_rate": 3.843023702543556e-05, "loss": 1.8274, "step": 10788 }, { "epoch": 0.272405675223144, "grad_norm": 0.9248176217079163, "learning_rate": 3.8359326504270984e-05, "loss": 1.8316, "step": 10819 }, { "epoch": 0.27318620724384074, "grad_norm": 0.9702487587928772, "learning_rate": 3.828826522492255e-05, "loss": 1.8344, "step": 10850 }, { "epoch": 0.2739667392645374, "grad_norm": 0.9324803948402405, "learning_rate": 3.821705398930713e-05, "loss": 1.8544, "step": 10881 }, { "epoch": 0.2747472712852341, "grad_norm": 0.9495794773101807, "learning_rate": 3.814569360103385e-05, "loss": 1.8446, "step": 10912 }, { "epoch": 0.2755278033059308, "grad_norm": 0.9734468460083008, "learning_rate": 3.807418486539499e-05, "loss": 1.8434, "step": 10943 }, { "epoch": 0.27630833532662746, "grad_norm": 0.9376105666160583, "learning_rate": 3.80025285893569e-05, "loss": 1.8339, "step": 10974 }, { "epoch": 0.2770888673473242, "grad_norm": 0.9637653827667236, "learning_rate": 3.793072558155093e-05, "loss": 1.8445, "step": 11005 }, { "epoch": 0.27786939936802085, "grad_norm": 0.9537612199783325, "learning_rate": 3.785877665226426e-05, "loss": 1.8332, "step": 11036 }, { "epoch": 0.2786499313887175, "grad_norm": 0.9774130582809448, "learning_rate": 3.778668261343079e-05, "loss": 1.8269, "step": 11067 }, { "epoch": 0.27943046340941424, "grad_norm": 0.9648153781890869, "learning_rate": 3.771444427862192e-05, "loss": 1.8301, "step": 11098 }, { "epoch": 0.2802109954301109, "grad_norm": 0.9317411184310913, "learning_rate": 3.7642062463037465e-05, "loss": 1.836, "step": 11129 }, { "epoch": 0.2809915274508076, "grad_norm": 0.9821638464927673, "learning_rate": 3.7569537983496373e-05, "loss": 1.846, "step": 11160 }, { "epoch": 0.2817720594715043, "grad_norm": 0.9444936513900757, "learning_rate": 3.749687165842753e-05, "loss": 1.8248, "step": 11191 }, { "epoch": 0.28255259149220097, "grad_norm": 0.9421222805976868, "learning_rate": 3.7424064307860536e-05, "loss": 1.8521, "step": 11222 }, { "epoch": 0.2833331235128977, "grad_norm": 0.9378920793533325, "learning_rate": 3.735111675341645e-05, "loss": 1.8361, "step": 11253 }, { "epoch": 0.28411365553359436, "grad_norm": 0.9363348484039307, "learning_rate": 3.7278029818298524e-05, "loss": 1.8276, "step": 11284 }, { "epoch": 0.284894187554291, "grad_norm": 0.9247562885284424, "learning_rate": 3.720480432728287e-05, "loss": 1.8354, "step": 11315 }, { "epoch": 0.28567471957498775, "grad_norm": 0.9505523443222046, "learning_rate": 3.71314411067092e-05, "loss": 1.807, "step": 11346 }, { "epoch": 0.2864552515956844, "grad_norm": 0.9292653799057007, "learning_rate": 3.70579409844715e-05, "loss": 1.8281, "step": 11377 }, { "epoch": 0.2872357836163811, "grad_norm": 0.9625663161277771, "learning_rate": 3.698430479000865e-05, "loss": 1.8211, "step": 11408 }, { "epoch": 0.2880163156370778, "grad_norm": 0.9280233979225159, "learning_rate": 3.691053335429509e-05, "loss": 1.8177, "step": 11439 }, { "epoch": 0.28879684765777447, "grad_norm": 0.9617031216621399, "learning_rate": 3.683662750983147e-05, "loss": 1.831, "step": 11470 }, { "epoch": 0.2895773796784712, "grad_norm": 0.9631845951080322, "learning_rate": 3.676258809063518e-05, "loss": 1.8154, "step": 11501 }, { "epoch": 0.29035791169916786, "grad_norm": 0.9200331568717957, "learning_rate": 3.6688415932231004e-05, "loss": 1.8452, "step": 11532 }, { "epoch": 0.2911384437198645, "grad_norm": 0.9431037902832031, "learning_rate": 3.661411187164166e-05, "loss": 1.8221, "step": 11563 }, { "epoch": 0.29191897574056125, "grad_norm": 0.9307507276535034, "learning_rate": 3.65396767473784e-05, "loss": 1.8092, "step": 11594 }, { "epoch": 0.2926995077612579, "grad_norm": 0.9896314144134521, "learning_rate": 3.6465111399431465e-05, "loss": 1.826, "step": 11625 }, { "epoch": 0.2934800397819546, "grad_norm": 0.9361921548843384, "learning_rate": 3.6390416669260674e-05, "loss": 1.8207, "step": 11656 }, { "epoch": 0.2942605718026513, "grad_norm": 0.9492716789245605, "learning_rate": 3.63155933997859e-05, "loss": 1.822, "step": 11687 }, { "epoch": 0.29504110382334797, "grad_norm": 0.9642727375030518, "learning_rate": 3.624064243537758e-05, "loss": 1.8165, "step": 11718 }, { "epoch": 0.29582163584404464, "grad_norm": 0.9700178503990173, "learning_rate": 3.616556462184716e-05, "loss": 1.8093, "step": 11749 }, { "epoch": 0.29660216786474136, "grad_norm": 0.90894615650177, "learning_rate": 3.609036080643755e-05, "loss": 1.8312, "step": 11780 }, { "epoch": 0.29738269988543803, "grad_norm": 0.9441593885421753, "learning_rate": 3.60150318378136e-05, "loss": 1.8324, "step": 11811 }, { "epoch": 0.29816323190613475, "grad_norm": 0.9351882338523865, "learning_rate": 3.5939578566052465e-05, "loss": 1.8303, "step": 11842 }, { "epoch": 0.2989437639268314, "grad_norm": 0.9438377618789673, "learning_rate": 3.586400184263408e-05, "loss": 1.8164, "step": 11873 }, { "epoch": 0.2997242959475281, "grad_norm": 0.9775059819221497, "learning_rate": 3.578830252043148e-05, "loss": 1.8146, "step": 11904 }, { "epoch": 0.3005048279682248, "grad_norm": 0.9342683553695679, "learning_rate": 3.571248145370125e-05, "loss": 1.8324, "step": 11935 }, { "epoch": 0.3012853599889215, "grad_norm": 0.9539228081703186, "learning_rate": 3.5636539498073794e-05, "loss": 1.8112, "step": 11966 }, { "epoch": 0.30206589200961814, "grad_norm": 1.0048060417175293, "learning_rate": 3.556047751054378e-05, "loss": 1.8261, "step": 11997 }, { "epoch": 0.30284642403031486, "grad_norm": 0.9269846081733704, "learning_rate": 3.548429634946039e-05, "loss": 1.8256, "step": 12028 }, { "epoch": 0.30362695605101153, "grad_norm": 0.937609076499939, "learning_rate": 3.540799687451768e-05, "loss": 1.8222, "step": 12059 }, { "epoch": 0.30440748807170825, "grad_norm": 0.9538866281509399, "learning_rate": 3.533157994674485e-05, "loss": 1.8214, "step": 12090 }, { "epoch": 0.3051880200924049, "grad_norm": 0.9644604325294495, "learning_rate": 3.5255046428496546e-05, "loss": 1.8117, "step": 12121 }, { "epoch": 0.3059685521131016, "grad_norm": 1.0498855113983154, "learning_rate": 3.517839718344311e-05, "loss": 1.8249, "step": 12152 }, { "epoch": 0.3067490841337983, "grad_norm": 0.952634334564209, "learning_rate": 3.510163307656086e-05, "loss": 1.8263, "step": 12183 }, { "epoch": 0.307529616154495, "grad_norm": 0.9833852052688599, "learning_rate": 3.5024754974122324e-05, "loss": 1.8246, "step": 12214 }, { "epoch": 0.30831014817519165, "grad_norm": 0.9421613812446594, "learning_rate": 3.494776374368643e-05, "loss": 1.8065, "step": 12245 }, { "epoch": 0.30909068019588837, "grad_norm": 0.9119872450828552, "learning_rate": 3.4870660254088724e-05, "loss": 1.7958, "step": 12276 }, { "epoch": 0.30987121221658503, "grad_norm": 0.9541532397270203, "learning_rate": 3.479344537543164e-05, "loss": 1.8156, "step": 12307 }, { "epoch": 0.31065174423728176, "grad_norm": 0.9333822131156921, "learning_rate": 3.4716119979074565e-05, "loss": 1.8098, "step": 12338 }, { "epoch": 0.3114322762579784, "grad_norm": 0.9501641392707825, "learning_rate": 3.463868493762412e-05, "loss": 1.815, "step": 12369 }, { "epoch": 0.3122128082786751, "grad_norm": 0.9692359566688538, "learning_rate": 3.456114112492418e-05, "loss": 1.8192, "step": 12400 }, { "epoch": 0.3129933402993718, "grad_norm": 0.9368917942047119, "learning_rate": 3.4483489416046164e-05, "loss": 1.8044, "step": 12431 }, { "epoch": 0.3137738723200685, "grad_norm": 0.981076180934906, "learning_rate": 3.440573068727905e-05, "loss": 1.806, "step": 12462 }, { "epoch": 0.31455440434076515, "grad_norm": 0.9535355567932129, "learning_rate": 3.4327865816119495e-05, "loss": 1.805, "step": 12493 }, { "epoch": 0.31533493636146187, "grad_norm": 0.9825057983398438, "learning_rate": 3.4249895681262025e-05, "loss": 1.7878, "step": 12524 }, { "epoch": 0.31611546838215854, "grad_norm": 0.9295104146003723, "learning_rate": 3.417182116258899e-05, "loss": 1.8091, "step": 12555 }, { "epoch": 0.31689600040285526, "grad_norm": 0.9550284147262573, "learning_rate": 3.409364314116074e-05, "loss": 1.8053, "step": 12586 }, { "epoch": 0.3176765324235519, "grad_norm": 0.9383823871612549, "learning_rate": 3.401536249920559e-05, "loss": 1.8034, "step": 12617 }, { "epoch": 0.3184570644442486, "grad_norm": 0.9513770341873169, "learning_rate": 3.393698012010998e-05, "loss": 1.8097, "step": 12648 }, { "epoch": 0.3192375964649453, "grad_norm": 0.9214389324188232, "learning_rate": 3.385849688840839e-05, "loss": 1.8089, "step": 12679 }, { "epoch": 0.320018128485642, "grad_norm": 0.9370726346969604, "learning_rate": 3.3779913689773414e-05, "loss": 1.8022, "step": 12710 }, { "epoch": 0.32079866050633865, "grad_norm": 0.9661527872085571, "learning_rate": 3.370123141100578e-05, "loss": 1.8151, "step": 12741 }, { "epoch": 0.3215791925270354, "grad_norm": 0.9569831490516663, "learning_rate": 3.3622450940024305e-05, "loss": 1.8283, "step": 12772 }, { "epoch": 0.32235972454773204, "grad_norm": 0.9162306785583496, "learning_rate": 3.35435731658559e-05, "loss": 1.8005, "step": 12803 }, { "epoch": 0.32314025656842876, "grad_norm": 0.955159068107605, "learning_rate": 3.346459897862552e-05, "loss": 1.7929, "step": 12834 }, { "epoch": 0.32392078858912543, "grad_norm": 0.9484529495239258, "learning_rate": 3.338552926954613e-05, "loss": 1.8355, "step": 12865 }, { "epoch": 0.3247013206098221, "grad_norm": 0.9199361205101013, "learning_rate": 3.330636493090868e-05, "loss": 1.8096, "step": 12896 }, { "epoch": 0.3254818526305188, "grad_norm": 0.928848147392273, "learning_rate": 3.322710685607193e-05, "loss": 1.7896, "step": 12927 }, { "epoch": 0.3262623846512155, "grad_norm": 0.9141325354576111, "learning_rate": 3.314775593945251e-05, "loss": 1.8113, "step": 12958 }, { "epoch": 0.32704291667191215, "grad_norm": 0.97685307264328, "learning_rate": 3.3068313076514714e-05, "loss": 1.8241, "step": 12989 }, { "epoch": 0.3278234486926089, "grad_norm": 0.9171717166900635, "learning_rate": 3.298877916376047e-05, "loss": 1.7953, "step": 13020 }, { "epoch": 0.32860398071330554, "grad_norm": 0.9607007503509521, "learning_rate": 3.290915509871915e-05, "loss": 1.819, "step": 13051 }, { "epoch": 0.32938451273400227, "grad_norm": 0.9524408578872681, "learning_rate": 3.282944177993753e-05, "loss": 1.8133, "step": 13082 }, { "epoch": 0.33016504475469893, "grad_norm": 0.9388900399208069, "learning_rate": 3.274964010696957e-05, "loss": 1.8103, "step": 13113 }, { "epoch": 0.3309455767753956, "grad_norm": 0.8972403407096863, "learning_rate": 3.266975098036629e-05, "loss": 1.7997, "step": 13144 }, { "epoch": 0.3317261087960923, "grad_norm": 0.9373508095741272, "learning_rate": 3.258977530166562e-05, "loss": 1.8038, "step": 13175 }, { "epoch": 0.332506640816789, "grad_norm": 0.9278662204742432, "learning_rate": 3.250971397338227e-05, "loss": 1.803, "step": 13206 }, { "epoch": 0.33328717283748566, "grad_norm": 0.9486290812492371, "learning_rate": 3.2429567898997404e-05, "loss": 1.8114, "step": 13237 }, { "epoch": 0.3340677048581824, "grad_norm": 0.9786876440048218, "learning_rate": 3.234933798294859e-05, "loss": 1.7902, "step": 13268 }, { "epoch": 0.33484823687887905, "grad_norm": 0.9001107215881348, "learning_rate": 3.2269025130619535e-05, "loss": 1.8108, "step": 13299 }, { "epoch": 0.33562876889957577, "grad_norm": 0.9080937504768372, "learning_rate": 3.218863024832985e-05, "loss": 1.7951, "step": 13330 }, { "epoch": 0.33640930092027244, "grad_norm": 0.93299800157547, "learning_rate": 3.2108154243324864e-05, "loss": 1.8129, "step": 13361 }, { "epoch": 0.3371898329409691, "grad_norm": 0.9452754855155945, "learning_rate": 3.2027598023765345e-05, "loss": 1.8077, "step": 13392 }, { "epoch": 0.3379703649616658, "grad_norm": 0.9515936970710754, "learning_rate": 3.194696249871729e-05, "loss": 1.7961, "step": 13423 }, { "epoch": 0.3387508969823625, "grad_norm": 0.9395990371704102, "learning_rate": 3.186624857814164e-05, "loss": 1.7982, "step": 13454 }, { "epoch": 0.33953142900305916, "grad_norm": 0.9647415280342102, "learning_rate": 3.178545717288401e-05, "loss": 1.7997, "step": 13485 }, { "epoch": 0.3403119610237559, "grad_norm": 0.9437438249588013, "learning_rate": 3.170458919466444e-05, "loss": 1.8069, "step": 13516 }, { "epoch": 0.34109249304445255, "grad_norm": 0.9350516200065613, "learning_rate": 3.1623645556067063e-05, "loss": 1.8024, "step": 13547 }, { "epoch": 0.34187302506514927, "grad_norm": 0.9213714599609375, "learning_rate": 3.154262717052985e-05, "loss": 1.7963, "step": 13578 }, { "epoch": 0.34265355708584594, "grad_norm": 0.9247968792915344, "learning_rate": 3.146153495233426e-05, "loss": 1.8172, "step": 13609 }, { "epoch": 0.3434340891065426, "grad_norm": 0.8953402638435364, "learning_rate": 3.1380369816594944e-05, "loss": 1.8004, "step": 13640 }, { "epoch": 0.34421462112723933, "grad_norm": 0.9214624762535095, "learning_rate": 3.129913267924946e-05, "loss": 1.7862, "step": 13671 }, { "epoch": 0.344995153147936, "grad_norm": 0.946559727191925, "learning_rate": 3.121782445704782e-05, "loss": 1.7973, "step": 13702 }, { "epoch": 0.34577568516863266, "grad_norm": 0.9344651103019714, "learning_rate": 3.11364460675423e-05, "loss": 1.7992, "step": 13733 }, { "epoch": 0.3465562171893294, "grad_norm": 0.9140195846557617, "learning_rate": 3.1054998429076934e-05, "loss": 1.7846, "step": 13764 }, { "epoch": 0.34733674921002605, "grad_norm": 0.9403822422027588, "learning_rate": 3.097348246077728e-05, "loss": 1.7814, "step": 13795 }, { "epoch": 0.3481172812307228, "grad_norm": 0.954587996006012, "learning_rate": 3.0891899082539924e-05, "loss": 1.7864, "step": 13826 }, { "epoch": 0.34889781325141944, "grad_norm": 0.9238544702529907, "learning_rate": 3.0810249215022233e-05, "loss": 1.7929, "step": 13857 }, { "epoch": 0.3496783452721161, "grad_norm": 0.8994883894920349, "learning_rate": 3.0728533779631865e-05, "loss": 1.805, "step": 13888 }, { "epoch": 0.35045887729281283, "grad_norm": 0.934203028678894, "learning_rate": 3.064675369851637e-05, "loss": 1.8046, "step": 13919 }, { "epoch": 0.3512394093135095, "grad_norm": 0.9044675230979919, "learning_rate": 3.056490989455289e-05, "loss": 1.7866, "step": 13950 }, { "epoch": 0.35201994133420617, "grad_norm": 0.915847897529602, "learning_rate": 3.0483003291337596e-05, "loss": 1.7861, "step": 13981 }, { "epoch": 0.3528004733549029, "grad_norm": 0.9346082210540771, "learning_rate": 3.040103481317539e-05, "loss": 1.7891, "step": 14012 }, { "epoch": 0.35358100537559956, "grad_norm": 0.9164626598358154, "learning_rate": 3.03190053850694e-05, "loss": 1.8021, "step": 14043 }, { "epoch": 0.3543615373962962, "grad_norm": 0.9445552229881287, "learning_rate": 3.0236915932710573e-05, "loss": 1.7877, "step": 14074 }, { "epoch": 0.35514206941699294, "grad_norm": 0.9604120254516602, "learning_rate": 3.0154767382467232e-05, "loss": 1.8087, "step": 14105 }, { "epoch": 0.3559226014376896, "grad_norm": 0.9186177849769592, "learning_rate": 3.0072560661374582e-05, "loss": 1.7802, "step": 14136 }, { "epoch": 0.35670313345838633, "grad_norm": 0.9228742122650146, "learning_rate": 2.999029669712431e-05, "loss": 1.8091, "step": 14167 }, { "epoch": 0.357483665479083, "grad_norm": 0.9045746922492981, "learning_rate": 2.990797641805408e-05, "loss": 1.8013, "step": 14198 }, { "epoch": 0.35826419749977967, "grad_norm": 0.9483916759490967, "learning_rate": 2.982560075313704e-05, "loss": 1.8044, "step": 14229 }, { "epoch": 0.3590447295204764, "grad_norm": 0.9580939412117004, "learning_rate": 2.9743170631971368e-05, "loss": 1.8037, "step": 14260 }, { "epoch": 0.35982526154117306, "grad_norm": 0.950960099697113, "learning_rate": 2.9660686984769792e-05, "loss": 1.8082, "step": 14291 }, { "epoch": 0.3606057935618697, "grad_norm": 0.9356971383094788, "learning_rate": 2.9578150742349047e-05, "loss": 1.8049, "step": 14322 }, { "epoch": 0.36138632558256645, "grad_norm": 0.9099254012107849, "learning_rate": 2.949556283611942e-05, "loss": 1.7991, "step": 14353 }, { "epoch": 0.3621668576032631, "grad_norm": 0.9190629720687866, "learning_rate": 2.9412924198074206e-05, "loss": 1.8018, "step": 14384 }, { "epoch": 0.36294738962395984, "grad_norm": 0.90566486120224, "learning_rate": 2.9330235760779208e-05, "loss": 1.789, "step": 14415 }, { "epoch": 0.3637279216446565, "grad_norm": 0.970696747303009, "learning_rate": 2.9247498457362188e-05, "loss": 1.7851, "step": 14446 }, { "epoch": 0.36450845366535317, "grad_norm": 0.9313214421272278, "learning_rate": 2.9164713221502373e-05, "loss": 1.7893, "step": 14477 }, { "epoch": 0.3652889856860499, "grad_norm": 1.2887455224990845, "learning_rate": 2.9081880987419912e-05, "loss": 1.7836, "step": 14508 }, { "epoch": 0.36606951770674656, "grad_norm": 0.922982394695282, "learning_rate": 2.8999002689865296e-05, "loss": 1.7676, "step": 14539 }, { "epoch": 0.36685004972744323, "grad_norm": 0.9466812014579773, "learning_rate": 2.8916079264108852e-05, "loss": 1.7831, "step": 14570 }, { "epoch": 0.36763058174813995, "grad_norm": 0.9107967615127563, "learning_rate": 2.883311164593017e-05, "loss": 1.7887, "step": 14601 }, { "epoch": 0.3684111137688366, "grad_norm": 0.8859044313430786, "learning_rate": 2.875010077160754e-05, "loss": 1.7823, "step": 14632 }, { "epoch": 0.36919164578953334, "grad_norm": 0.9182350039482117, "learning_rate": 2.866704757790741e-05, "loss": 1.7816, "step": 14663 }, { "epoch": 0.36997217781023, "grad_norm": 1.1406787633895874, "learning_rate": 2.858395300207376e-05, "loss": 1.7962, "step": 14694 }, { "epoch": 0.3707527098309267, "grad_norm": 0.9399011135101318, "learning_rate": 2.8500817981817607e-05, "loss": 1.7874, "step": 14725 }, { "epoch": 0.3715332418516234, "grad_norm": 0.9223036766052246, "learning_rate": 2.8417643455306336e-05, "loss": 1.7661, "step": 14756 }, { "epoch": 0.37231377387232006, "grad_norm": 0.9224306344985962, "learning_rate": 2.8334430361153185e-05, "loss": 1.7799, "step": 14787 }, { "epoch": 0.37309430589301673, "grad_norm": 0.9353799819946289, "learning_rate": 2.8251179638406612e-05, "loss": 1.795, "step": 14818 }, { "epoch": 0.37387483791371345, "grad_norm": 0.9208987355232239, "learning_rate": 2.8167892226539704e-05, "loss": 1.7967, "step": 14849 }, { "epoch": 0.3746553699344101, "grad_norm": 0.9143754243850708, "learning_rate": 2.8084569065439588e-05, "loss": 1.7871, "step": 14880 }, { "epoch": 0.37543590195510684, "grad_norm": 0.9162681698799133, "learning_rate": 2.8001211095396807e-05, "loss": 1.7791, "step": 14911 }, { "epoch": 0.3762164339758035, "grad_norm": 0.9125809073448181, "learning_rate": 2.791781925709473e-05, "loss": 1.7847, "step": 14942 }, { "epoch": 0.3769969659965002, "grad_norm": 0.9262179732322693, "learning_rate": 2.7834394491598908e-05, "loss": 1.789, "step": 14973 }, { "epoch": 0.3777774980171969, "grad_norm": 0.9118053317070007, "learning_rate": 2.7750937740346485e-05, "loss": 1.7906, "step": 15004 }, { "epoch": 0.37855803003789357, "grad_norm": 0.9322589635848999, "learning_rate": 2.7667449945135564e-05, "loss": 1.7811, "step": 15035 }, { "epoch": 0.37933856205859023, "grad_norm": 0.9277715682983398, "learning_rate": 2.7583932048114557e-05, "loss": 1.7887, "step": 15066 }, { "epoch": 0.38011909407928696, "grad_norm": 0.9804796576499939, "learning_rate": 2.7500384991771587e-05, "loss": 1.7945, "step": 15097 }, { "epoch": 0.3808996260999836, "grad_norm": 0.9327893853187561, "learning_rate": 2.7416809718923825e-05, "loss": 1.785, "step": 15128 }, { "epoch": 0.38168015812068035, "grad_norm": 0.9126656651496887, "learning_rate": 2.7333207172706864e-05, "loss": 1.7782, "step": 15159 }, { "epoch": 0.382460690141377, "grad_norm": 0.9476912617683411, "learning_rate": 2.7249578296564088e-05, "loss": 1.7906, "step": 15190 }, { "epoch": 0.3832412221620737, "grad_norm": 0.9388076663017273, "learning_rate": 2.7165924034235973e-05, "loss": 1.7993, "step": 15221 }, { "epoch": 0.3840217541827704, "grad_norm": 0.9265156984329224, "learning_rate": 2.708224532974953e-05, "loss": 1.7624, "step": 15252 }, { "epoch": 0.38480228620346707, "grad_norm": 0.9261599183082581, "learning_rate": 2.6998543127407538e-05, "loss": 1.7774, "step": 15283 }, { "epoch": 0.38558281822416374, "grad_norm": 0.9569805860519409, "learning_rate": 2.6914818371777988e-05, "loss": 1.7695, "step": 15314 }, { "epoch": 0.38636335024486046, "grad_norm": 0.8934130668640137, "learning_rate": 2.6831072007683373e-05, "loss": 1.7746, "step": 15345 }, { "epoch": 0.3871438822655571, "grad_norm": 0.9160299301147461, "learning_rate": 2.6747304980190018e-05, "loss": 1.7698, "step": 15376 }, { "epoch": 0.38792441428625385, "grad_norm": 0.9243568181991577, "learning_rate": 2.6663518234597453e-05, "loss": 1.7823, "step": 15407 }, { "epoch": 0.3887049463069505, "grad_norm": 0.9243864417076111, "learning_rate": 2.6579712716427696e-05, "loss": 1.7819, "step": 15438 }, { "epoch": 0.3894854783276472, "grad_norm": 0.9425201416015625, "learning_rate": 2.6495889371414652e-05, "loss": 1.7808, "step": 15469 }, { "epoch": 0.3902660103483439, "grad_norm": 0.944649875164032, "learning_rate": 2.6412049145493367e-05, "loss": 1.781, "step": 15500 }, { "epoch": 0.3910465423690406, "grad_norm": 0.9222155809402466, "learning_rate": 2.632819298478939e-05, "loss": 1.7818, "step": 15531 }, { "epoch": 0.39182707438973724, "grad_norm": 0.9028238654136658, "learning_rate": 2.6244321835608105e-05, "loss": 1.7622, "step": 15562 }, { "epoch": 0.39260760641043396, "grad_norm": 0.9092904329299927, "learning_rate": 2.6160436644424024e-05, "loss": 1.7829, "step": 15593 }, { "epoch": 0.39338813843113063, "grad_norm": 0.9831812381744385, "learning_rate": 2.6076538357870133e-05, "loss": 1.7927, "step": 15624 }, { "epoch": 0.39416867045182735, "grad_norm": 0.9344479441642761, "learning_rate": 2.5992627922727196e-05, "loss": 1.7725, "step": 15655 }, { "epoch": 0.394949202472524, "grad_norm": 0.9424457550048828, "learning_rate": 2.5908706285913066e-05, "loss": 1.77, "step": 15686 }, { "epoch": 0.3957297344932207, "grad_norm": 0.9403957724571228, "learning_rate": 2.5824774394472008e-05, "loss": 1.7822, "step": 15717 }, { "epoch": 0.3965102665139174, "grad_norm": 0.9497373104095459, "learning_rate": 2.5740833195563996e-05, "loss": 1.7874, "step": 15748 }, { "epoch": 0.3972907985346141, "grad_norm": 0.9396623373031616, "learning_rate": 2.5656883636454067e-05, "loss": 1.7634, "step": 15779 }, { "epoch": 0.39807133055531074, "grad_norm": 0.9426395297050476, "learning_rate": 2.557292666450159e-05, "loss": 1.7774, "step": 15810 }, { "epoch": 0.39885186257600747, "grad_norm": 0.9340534806251526, "learning_rate": 2.5488963227149566e-05, "loss": 1.7619, "step": 15841 }, { "epoch": 0.39963239459670413, "grad_norm": 0.9206343293190002, "learning_rate": 2.5404994271913983e-05, "loss": 1.7817, "step": 15872 }, { "epoch": 0.40041292661740086, "grad_norm": 0.9418168067932129, "learning_rate": 2.5321020746373085e-05, "loss": 1.7788, "step": 15903 }, { "epoch": 0.4011934586380975, "grad_norm": 0.9084994792938232, "learning_rate": 2.52370435981567e-05, "loss": 1.7784, "step": 15934 }, { "epoch": 0.4019739906587942, "grad_norm": 0.9424017667770386, "learning_rate": 2.5153063774935533e-05, "loss": 1.7744, "step": 15965 }, { "epoch": 0.4027545226794909, "grad_norm": 0.9338597655296326, "learning_rate": 2.506908222441045e-05, "loss": 1.7533, "step": 15996 }, { "epoch": 0.4035350547001876, "grad_norm": 0.9237053394317627, "learning_rate": 2.498509989430187e-05, "loss": 1.7691, "step": 16027 }, { "epoch": 0.40431558672088425, "grad_norm": 0.9271100759506226, "learning_rate": 2.4901117732338958e-05, "loss": 1.79, "step": 16058 }, { "epoch": 0.40509611874158097, "grad_norm": 0.9305247068405151, "learning_rate": 2.481713668624899e-05, "loss": 1.7786, "step": 16089 }, { "epoch": 0.40587665076227764, "grad_norm": 0.9237347841262817, "learning_rate": 2.4733157703746663e-05, "loss": 1.7951, "step": 16120 }, { "epoch": 0.40665718278297436, "grad_norm": 0.8985100388526917, "learning_rate": 2.4649181732523392e-05, "loss": 1.7699, "step": 16151 }, { "epoch": 0.407437714803671, "grad_norm": 0.9226688742637634, "learning_rate": 2.4565209720236582e-05, "loss": 1.769, "step": 16182 }, { "epoch": 0.4082182468243677, "grad_norm": 0.9165879487991333, "learning_rate": 2.4481242614498975e-05, "loss": 1.7712, "step": 16213 }, { "epoch": 0.4089987788450644, "grad_norm": 0.949297308921814, "learning_rate": 2.439728136286796e-05, "loss": 1.7709, "step": 16244 }, { "epoch": 0.4097793108657611, "grad_norm": 0.930533230304718, "learning_rate": 2.4313326912834852e-05, "loss": 1.7512, "step": 16275 }, { "epoch": 0.41055984288645775, "grad_norm": 0.926105797290802, "learning_rate": 2.4229380211814206e-05, "loss": 1.7769, "step": 16306 }, { "epoch": 0.41134037490715447, "grad_norm": 0.9282335638999939, "learning_rate": 2.4145442207133124e-05, "loss": 1.7624, "step": 16337 }, { "epoch": 0.41212090692785114, "grad_norm": 0.9377603530883789, "learning_rate": 2.406151384602059e-05, "loss": 1.7756, "step": 16368 }, { "epoch": 0.4129014389485478, "grad_norm": 0.9203463792800903, "learning_rate": 2.3977596075596747e-05, "loss": 1.7839, "step": 16399 }, { "epoch": 0.41368197096924453, "grad_norm": 0.9060366153717041, "learning_rate": 2.3893689842862223e-05, "loss": 1.7667, "step": 16430 }, { "epoch": 0.4144625029899412, "grad_norm": 0.9065908193588257, "learning_rate": 2.3809796094687475e-05, "loss": 1.7691, "step": 16461 }, { "epoch": 0.4152430350106379, "grad_norm": 0.9281785488128662, "learning_rate": 2.372591577780202e-05, "loss": 1.7573, "step": 16492 }, { "epoch": 0.4160235670313346, "grad_norm": 0.880733847618103, "learning_rate": 2.3642049838783838e-05, "loss": 1.7697, "step": 16523 }, { "epoch": 0.41680409905203125, "grad_norm": 0.9289157390594482, "learning_rate": 2.3558199224048666e-05, "loss": 1.7604, "step": 16554 }, { "epoch": 0.417584631072728, "grad_norm": 0.9380747079849243, "learning_rate": 2.347436487983929e-05, "loss": 1.7686, "step": 16585 }, { "epoch": 0.41836516309342464, "grad_norm": 0.9439517855644226, "learning_rate": 2.3390547752214888e-05, "loss": 1.7786, "step": 16616 }, { "epoch": 0.4191456951141213, "grad_norm": 0.9420905113220215, "learning_rate": 2.330674878704035e-05, "loss": 1.762, "step": 16647 }, { "epoch": 0.41992622713481803, "grad_norm": 0.9132199883460999, "learning_rate": 2.322296892997561e-05, "loss": 1.776, "step": 16678 }, { "epoch": 0.4207067591555147, "grad_norm": 0.9286783933639526, "learning_rate": 2.313920912646497e-05, "loss": 1.7564, "step": 16709 }, { "epoch": 0.4214872911762114, "grad_norm": 0.9031944870948792, "learning_rate": 2.305547032172643e-05, "loss": 1.75, "step": 16740 }, { "epoch": 0.4222678231969081, "grad_norm": 0.9604383111000061, "learning_rate": 2.2971753460741014e-05, "loss": 1.7821, "step": 16771 }, { "epoch": 0.42304835521760475, "grad_norm": 0.9145975112915039, "learning_rate": 2.288805948824212e-05, "loss": 1.7604, "step": 16802 }, { "epoch": 0.4238288872383015, "grad_norm": 0.9214676022529602, "learning_rate": 2.2804389348704858e-05, "loss": 1.7553, "step": 16833 }, { "epoch": 0.42460941925899814, "grad_norm": 0.9024298191070557, "learning_rate": 2.2720743986335374e-05, "loss": 1.7652, "step": 16864 }, { "epoch": 0.4253899512796948, "grad_norm": 0.8926112055778503, "learning_rate": 2.2637124345060233e-05, "loss": 1.7583, "step": 16895 }, { "epoch": 0.42617048330039153, "grad_norm": 0.9225366711616516, "learning_rate": 2.2553531368515695e-05, "loss": 1.7607, "step": 16926 }, { "epoch": 0.4269510153210882, "grad_norm": 0.9340829253196716, "learning_rate": 2.2469966000037144e-05, "loss": 1.7659, "step": 16957 }, { "epoch": 0.4277315473417849, "grad_norm": 0.9964073896408081, "learning_rate": 2.2386429182648417e-05, "loss": 1.7414, "step": 16988 }, { "epoch": 0.4285120793624816, "grad_norm": 0.9147719144821167, "learning_rate": 2.230292185905114e-05, "loss": 1.764, "step": 17019 }, { "epoch": 0.42929261138317826, "grad_norm": 0.9175845384597778, "learning_rate": 2.2219444971614116e-05, "loss": 1.7623, "step": 17050 }, { "epoch": 0.430073143403875, "grad_norm": 0.9041738510131836, "learning_rate": 2.2135999462362655e-05, "loss": 1.7448, "step": 17081 }, { "epoch": 0.43085367542457165, "grad_norm": 0.9334216713905334, "learning_rate": 2.2052586272968003e-05, "loss": 1.7548, "step": 17112 }, { "epoch": 0.4316342074452683, "grad_norm": 0.9379672408103943, "learning_rate": 2.196920634473666e-05, "loss": 1.7703, "step": 17143 }, { "epoch": 0.43241473946596504, "grad_norm": 0.9310106039047241, "learning_rate": 2.1885860618599787e-05, "loss": 1.7567, "step": 17174 }, { "epoch": 0.4331952714866617, "grad_norm": 0.936827540397644, "learning_rate": 2.1802550035102577e-05, "loss": 1.7572, "step": 17205 }, { "epoch": 0.4339758035073584, "grad_norm": 0.954757571220398, "learning_rate": 2.171927553439363e-05, "loss": 1.7773, "step": 17236 }, { "epoch": 0.4347563355280551, "grad_norm": 0.9472478032112122, "learning_rate": 2.1636038056214376e-05, "loss": 1.7614, "step": 17267 }, { "epoch": 0.43553686754875176, "grad_norm": 0.9342318773269653, "learning_rate": 2.155283853988844e-05, "loss": 1.7561, "step": 17298 }, { "epoch": 0.4363173995694485, "grad_norm": 0.9363663792610168, "learning_rate": 2.146967792431106e-05, "loss": 1.7703, "step": 17329 }, { "epoch": 0.43709793159014515, "grad_norm": 0.9496386647224426, "learning_rate": 2.138655714793849e-05, "loss": 1.7565, "step": 17360 }, { "epoch": 0.4378784636108418, "grad_norm": 0.9076322913169861, "learning_rate": 2.1303477148777367e-05, "loss": 1.738, "step": 17391 }, { "epoch": 0.43865899563153854, "grad_norm": 0.9229772686958313, "learning_rate": 2.122043886437421e-05, "loss": 1.7455, "step": 17422 }, { "epoch": 0.4394395276522352, "grad_norm": 0.9290786385536194, "learning_rate": 2.1137443231804765e-05, "loss": 1.7635, "step": 17453 }, { "epoch": 0.44022005967293193, "grad_norm": 0.9575942754745483, "learning_rate": 2.105449118766347e-05, "loss": 1.7763, "step": 17484 }, { "epoch": 0.4410005916936286, "grad_norm": 0.9625800848007202, "learning_rate": 2.097158366805287e-05, "loss": 1.7514, "step": 17515 }, { "epoch": 0.44178112371432526, "grad_norm": 0.9176467061042786, "learning_rate": 2.0888721608573047e-05, "loss": 1.7521, "step": 17546 }, { "epoch": 0.442561655735022, "grad_norm": 0.92497318983078, "learning_rate": 2.0805905944311087e-05, "loss": 1.7582, "step": 17577 }, { "epoch": 0.44334218775571865, "grad_norm": 0.9499204158782959, "learning_rate": 2.0723137609830497e-05, "loss": 1.762, "step": 17608 }, { "epoch": 0.4441227197764153, "grad_norm": 0.9391850829124451, "learning_rate": 2.0640417539160686e-05, "loss": 1.7495, "step": 17639 }, { "epoch": 0.44490325179711204, "grad_norm": 0.9657866358757019, "learning_rate": 2.0557746665786427e-05, "loss": 1.7642, "step": 17670 }, { "epoch": 0.4456837838178087, "grad_norm": 0.955342173576355, "learning_rate": 2.0475125922637256e-05, "loss": 1.7645, "step": 17701 }, { "epoch": 0.44646431583850543, "grad_norm": 0.9168228507041931, "learning_rate": 2.0392556242077047e-05, "loss": 1.7598, "step": 17732 }, { "epoch": 0.4472448478592021, "grad_norm": 0.8854954242706299, "learning_rate": 2.031003855589343e-05, "loss": 1.7576, "step": 17763 }, { "epoch": 0.44802537987989877, "grad_norm": 0.9424418210983276, "learning_rate": 2.022757379528727e-05, "loss": 1.766, "step": 17794 }, { "epoch": 0.4488059119005955, "grad_norm": 0.9471886157989502, "learning_rate": 2.0145162890862184e-05, "loss": 1.7458, "step": 17825 }, { "epoch": 0.44958644392129216, "grad_norm": 0.9373571872711182, "learning_rate": 2.0062806772614022e-05, "loss": 1.743, "step": 17856 }, { "epoch": 0.4503669759419888, "grad_norm": 0.8964229226112366, "learning_rate": 1.9980506369920392e-05, "loss": 1.756, "step": 17887 }, { "epoch": 0.45114750796268555, "grad_norm": 0.9414677619934082, "learning_rate": 1.989826261153015e-05, "loss": 1.7407, "step": 17918 }, { "epoch": 0.4519280399833822, "grad_norm": 0.9373415112495422, "learning_rate": 1.9816076425552923e-05, "loss": 1.746, "step": 17949 }, { "epoch": 0.45270857200407894, "grad_norm": 0.8916881084442139, "learning_rate": 1.9733948739448676e-05, "loss": 1.7564, "step": 17980 }, { "epoch": 0.4534891040247756, "grad_norm": 0.9054657816886902, "learning_rate": 1.9651880480017155e-05, "loss": 1.7586, "step": 18011 }, { "epoch": 0.45426963604547227, "grad_norm": 0.9446020126342773, "learning_rate": 1.9569872573387516e-05, "loss": 1.758, "step": 18042 }, { "epoch": 0.455050168066169, "grad_norm": 0.9143036007881165, "learning_rate": 1.9487925945007854e-05, "loss": 1.7554, "step": 18073 }, { "epoch": 0.45583070008686566, "grad_norm": 0.9555609822273254, "learning_rate": 1.9406041519634726e-05, "loss": 1.7533, "step": 18104 }, { "epoch": 0.4566112321075623, "grad_norm": 0.929480791091919, "learning_rate": 1.932422022132275e-05, "loss": 1.7599, "step": 18135 }, { "epoch": 0.45739176412825905, "grad_norm": 0.9011172652244568, "learning_rate": 1.924246297341414e-05, "loss": 1.7712, "step": 18166 }, { "epoch": 0.4581722961489557, "grad_norm": 0.9395762085914612, "learning_rate": 1.9160770698528338e-05, "loss": 1.7715, "step": 18197 }, { "epoch": 0.45895282816965244, "grad_norm": 0.9550304412841797, "learning_rate": 1.907914431855156e-05, "loss": 1.7315, "step": 18228 }, { "epoch": 0.4597333601903491, "grad_norm": 0.9520135521888733, "learning_rate": 1.8997584754626412e-05, "loss": 1.7453, "step": 18259 }, { "epoch": 0.46051389221104577, "grad_norm": 0.8951966762542725, "learning_rate": 1.8916092927141486e-05, "loss": 1.7496, "step": 18290 }, { "epoch": 0.4612944242317425, "grad_norm": 0.936210036277771, "learning_rate": 1.883466975572098e-05, "loss": 1.7407, "step": 18321 }, { "epoch": 0.46207495625243916, "grad_norm": 0.95763099193573, "learning_rate": 1.8753316159214312e-05, "loss": 1.7566, "step": 18352 }, { "epoch": 0.46285548827313583, "grad_norm": 0.9345053434371948, "learning_rate": 1.8672033055685766e-05, "loss": 1.7769, "step": 18383 }, { "epoch": 0.46363602029383255, "grad_norm": 0.9099903106689453, "learning_rate": 1.8590821362404116e-05, "loss": 1.7416, "step": 18414 }, { "epoch": 0.4644165523145292, "grad_norm": 0.9202333092689514, "learning_rate": 1.8509681995832294e-05, "loss": 1.7372, "step": 18445 }, { "epoch": 0.46519708433522594, "grad_norm": 0.9366357922554016, "learning_rate": 1.8428615871617004e-05, "loss": 1.7412, "step": 18476 }, { "epoch": 0.4659776163559226, "grad_norm": 0.9029955267906189, "learning_rate": 1.8347623904578448e-05, "loss": 1.7465, "step": 18507 }, { "epoch": 0.4667581483766193, "grad_norm": 0.9197161197662354, "learning_rate": 1.8266707008699975e-05, "loss": 1.7462, "step": 18538 }, { "epoch": 0.467538680397316, "grad_norm": 0.9279952049255371, "learning_rate": 1.818586609711774e-05, "loss": 1.7496, "step": 18569 }, { "epoch": 0.46831921241801266, "grad_norm": 0.9148679971694946, "learning_rate": 1.8105102082110462e-05, "loss": 1.7389, "step": 18600 }, { "epoch": 0.46909974443870933, "grad_norm": 0.9217187166213989, "learning_rate": 1.8024415875089058e-05, "loss": 1.7377, "step": 18631 }, { "epoch": 0.46988027645940605, "grad_norm": 0.9556980133056641, "learning_rate": 1.7943808386586407e-05, "loss": 1.7391, "step": 18662 }, { "epoch": 0.4706608084801027, "grad_norm": 0.9437476992607117, "learning_rate": 1.7863280526247073e-05, "loss": 1.7401, "step": 18693 }, { "epoch": 0.4714413405007994, "grad_norm": 0.923971951007843, "learning_rate": 1.7782833202817003e-05, "loss": 1.7492, "step": 18724 }, { "epoch": 0.4722218725214961, "grad_norm": 0.9325092434883118, "learning_rate": 1.7702467324133327e-05, "loss": 1.7477, "step": 18755 }, { "epoch": 0.4730024045421928, "grad_norm": 0.9427450299263, "learning_rate": 1.7622183797114042e-05, "loss": 1.7722, "step": 18786 }, { "epoch": 0.4737829365628895, "grad_norm": 0.8803645372390747, "learning_rate": 1.7541983527747838e-05, "loss": 1.7428, "step": 18817 }, { "epoch": 0.47456346858358617, "grad_norm": 0.9492975473403931, "learning_rate": 1.746186742108387e-05, "loss": 1.7558, "step": 18848 }, { "epoch": 0.47534400060428283, "grad_norm": 0.9038426280021667, "learning_rate": 1.73818363812215e-05, "loss": 1.7576, "step": 18879 }, { "epoch": 0.47612453262497956, "grad_norm": 0.9929136633872986, "learning_rate": 1.7301891311300153e-05, "loss": 1.7405, "step": 18910 }, { "epoch": 0.4769050646456762, "grad_norm": 0.9464752078056335, "learning_rate": 1.7222033113489055e-05, "loss": 1.7343, "step": 18941 }, { "epoch": 0.4776855966663729, "grad_norm": 0.9222598075866699, "learning_rate": 1.7142262688977127e-05, "loss": 1.7377, "step": 18972 }, { "epoch": 0.4784661286870696, "grad_norm": 0.9139207005500793, "learning_rate": 1.7062580937962764e-05, "loss": 1.7448, "step": 19003 }, { "epoch": 0.4792466607077663, "grad_norm": 0.9283340573310852, "learning_rate": 1.698298875964369e-05, "loss": 1.7577, "step": 19034 }, { "epoch": 0.480027192728463, "grad_norm": 0.9210373163223267, "learning_rate": 1.690348705220684e-05, "loss": 1.7351, "step": 19065 }, { "epoch": 0.48080772474915967, "grad_norm": 0.9250310063362122, "learning_rate": 1.6824076712818156e-05, "loss": 1.7554, "step": 19096 }, { "epoch": 0.48158825676985634, "grad_norm": 0.9364836812019348, "learning_rate": 1.6744758637612533e-05, "loss": 1.7513, "step": 19127 }, { "epoch": 0.48236878879055306, "grad_norm": 0.9413596391677856, "learning_rate": 1.6665533721683664e-05, "loss": 1.7358, "step": 19158 }, { "epoch": 0.4831493208112497, "grad_norm": 0.9139796495437622, "learning_rate": 1.6586402859073974e-05, "loss": 1.74, "step": 19189 }, { "epoch": 0.4839298528319464, "grad_norm": 0.9309468269348145, "learning_rate": 1.6507366942764463e-05, "loss": 1.7407, "step": 19220 }, { "epoch": 0.4847103848526431, "grad_norm": 0.9395124316215515, "learning_rate": 1.6428426864664732e-05, "loss": 1.7541, "step": 19251 }, { "epoch": 0.4854909168733398, "grad_norm": 0.9307234883308411, "learning_rate": 1.6349583515602816e-05, "loss": 1.7509, "step": 19282 }, { "epoch": 0.4862714488940365, "grad_norm": 0.9114024043083191, "learning_rate": 1.6270837785315208e-05, "loss": 1.749, "step": 19313 }, { "epoch": 0.4870519809147332, "grad_norm": 0.9282791018486023, "learning_rate": 1.619219056243676e-05, "loss": 1.7257, "step": 19344 }, { "epoch": 0.48783251293542984, "grad_norm": 0.9253626465797424, "learning_rate": 1.6113642734490698e-05, "loss": 1.7298, "step": 19375 }, { "epoch": 0.48861304495612656, "grad_norm": 0.9356294870376587, "learning_rate": 1.6035195187878577e-05, "loss": 1.7419, "step": 19406 }, { "epoch": 0.48939357697682323, "grad_norm": 0.9312499761581421, "learning_rate": 1.5956848807870305e-05, "loss": 1.7409, "step": 19437 }, { "epoch": 0.4901741089975199, "grad_norm": 0.91697758436203, "learning_rate": 1.587860447859413e-05, "loss": 1.7288, "step": 19468 }, { "epoch": 0.4909546410182166, "grad_norm": 0.9301339983940125, "learning_rate": 1.5800463083026686e-05, "loss": 1.7479, "step": 19499 }, { "epoch": 0.4917351730389133, "grad_norm": 0.9044349193572998, "learning_rate": 1.572242550298298e-05, "loss": 1.7264, "step": 19530 }, { "epoch": 0.49251570505961, "grad_norm": 0.9050460457801819, "learning_rate": 1.56444926191065e-05, "loss": 1.7394, "step": 19561 }, { "epoch": 0.4932962370803067, "grad_norm": 0.9435474872589111, "learning_rate": 1.5566665310859257e-05, "loss": 1.7438, "step": 19592 }, { "epoch": 0.49407676910100334, "grad_norm": 0.9283745288848877, "learning_rate": 1.5488944456511846e-05, "loss": 1.7603, "step": 19623 }, { "epoch": 0.49485730112170007, "grad_norm": 0.9714552164077759, "learning_rate": 1.5411330933133546e-05, "loss": 1.7316, "step": 19654 }, { "epoch": 0.49563783314239673, "grad_norm": 0.92011958360672, "learning_rate": 1.533382561658241e-05, "loss": 1.7478, "step": 19685 }, { "epoch": 0.4964183651630934, "grad_norm": 0.929989218711853, "learning_rate": 1.525642938149541e-05, "loss": 1.7497, "step": 19716 }, { "epoch": 0.4971988971837901, "grad_norm": 0.9507896900177002, "learning_rate": 1.5179143101278536e-05, "loss": 1.7485, "step": 19747 }, { "epoch": 0.4979794292044868, "grad_norm": 0.9416571855545044, "learning_rate": 1.5101967648096955e-05, "loss": 1.7289, "step": 19778 }, { "epoch": 0.4987599612251835, "grad_norm": 0.9024575352668762, "learning_rate": 1.5024903892865172e-05, "loss": 1.7436, "step": 19809 }, { "epoch": 0.4995404932458802, "grad_norm": 0.9383466243743896, "learning_rate": 1.4947952705237184e-05, "loss": 1.7362, "step": 19840 }, { "epoch": 0.5003210252665768, "grad_norm": 0.9311878681182861, "learning_rate": 1.4871114953596682e-05, "loss": 1.7595, "step": 19871 }, { "epoch": 0.5011015572872736, "grad_norm": 0.9517268538475037, "learning_rate": 1.4794391505047256e-05, "loss": 1.7367, "step": 19902 }, { "epoch": 0.5018820893079703, "grad_norm": 0.911535918712616, "learning_rate": 1.4717783225402596e-05, "loss": 1.7268, "step": 19933 }, { "epoch": 0.5026626213286669, "grad_norm": 0.9200896620750427, "learning_rate": 1.4641290979176735e-05, "loss": 1.7353, "step": 19964 }, { "epoch": 0.5034431533493636, "grad_norm": 0.9185106754302979, "learning_rate": 1.4564915629574246e-05, "loss": 1.7441, "step": 19995 }, { "epoch": 0.5042236853700603, "grad_norm": 0.9203515648841858, "learning_rate": 1.4488658038480601e-05, "loss": 1.7293, "step": 20026 }, { "epoch": 0.505004217390757, "grad_norm": 0.9532755613327026, "learning_rate": 1.4412519066452323e-05, "loss": 1.7478, "step": 20057 }, { "epoch": 0.5057847494114537, "grad_norm": 0.9353486895561218, "learning_rate": 1.4336499572707373e-05, "loss": 1.7468, "step": 20088 }, { "epoch": 0.5065652814321504, "grad_norm": 0.9349709153175354, "learning_rate": 1.4260600415115433e-05, "loss": 1.7256, "step": 20119 }, { "epoch": 0.507345813452847, "grad_norm": 0.9995957016944885, "learning_rate": 1.4184822450188137e-05, "loss": 1.7361, "step": 20150 }, { "epoch": 0.5081263454735437, "grad_norm": 0.9351862072944641, "learning_rate": 1.410916653306954e-05, "loss": 1.7537, "step": 20181 }, { "epoch": 0.5089068774942405, "grad_norm": 0.9246407747268677, "learning_rate": 1.403363351752639e-05, "loss": 1.7241, "step": 20212 }, { "epoch": 0.5096874095149371, "grad_norm": 0.9154999256134033, "learning_rate": 1.3958224255938485e-05, "loss": 1.7207, "step": 20243 }, { "epoch": 0.5104679415356338, "grad_norm": 0.9373302459716797, "learning_rate": 1.388293959928911e-05, "loss": 1.7538, "step": 20274 }, { "epoch": 0.5112484735563305, "grad_norm": 0.924915611743927, "learning_rate": 1.3807780397155379e-05, "loss": 1.737, "step": 20305 }, { "epoch": 0.5120290055770271, "grad_norm": 0.9185555577278137, "learning_rate": 1.3732747497698655e-05, "loss": 1.7105, "step": 20336 }, { "epoch": 0.5128095375977239, "grad_norm": 0.9583287239074707, "learning_rate": 1.3657841747655038e-05, "loss": 1.7398, "step": 20367 }, { "epoch": 0.5135900696184206, "grad_norm": 0.9166149497032166, "learning_rate": 1.3583063992325706e-05, "loss": 1.7428, "step": 20398 }, { "epoch": 0.5143706016391173, "grad_norm": 0.9236160516738892, "learning_rate": 1.3508415075567496e-05, "loss": 1.7355, "step": 20429 }, { "epoch": 0.5151511336598139, "grad_norm": 0.9086827635765076, "learning_rate": 1.343389583978327e-05, "loss": 1.737, "step": 20460 }, { "epoch": 0.5159316656805106, "grad_norm": 0.9379333257675171, "learning_rate": 1.3359507125912468e-05, "loss": 1.7218, "step": 20491 }, { "epoch": 0.5167121977012074, "grad_norm": 0.9533993005752563, "learning_rate": 1.3285249773421627e-05, "loss": 1.7339, "step": 20522 }, { "epoch": 0.517492729721904, "grad_norm": 0.9649523496627808, "learning_rate": 1.3211124620294884e-05, "loss": 1.7459, "step": 20553 }, { "epoch": 0.5182732617426007, "grad_norm": 0.9658921360969543, "learning_rate": 1.313713250302451e-05, "loss": 1.7391, "step": 20584 }, { "epoch": 0.5190537937632974, "grad_norm": 0.960582435131073, "learning_rate": 1.3063274256601479e-05, "loss": 1.7303, "step": 20615 }, { "epoch": 0.519834325783994, "grad_norm": 0.9230870008468628, "learning_rate": 1.2989550714506086e-05, "loss": 1.7223, "step": 20646 }, { "epoch": 0.5206148578046907, "grad_norm": 0.9037120342254639, "learning_rate": 1.291596270869846e-05, "loss": 1.7191, "step": 20677 }, { "epoch": 0.5213953898253875, "grad_norm": 0.9674888253211975, "learning_rate": 1.284251106960927e-05, "loss": 1.7177, "step": 20708 }, { "epoch": 0.5221759218460841, "grad_norm": 0.908338725566864, "learning_rate": 1.2769196626130263e-05, "loss": 1.7243, "step": 20739 }, { "epoch": 0.5229564538667808, "grad_norm": 0.9437979459762573, "learning_rate": 1.2696020205604969e-05, "loss": 1.7391, "step": 20770 }, { "epoch": 0.5237369858874775, "grad_norm": 0.9256695508956909, "learning_rate": 1.2622982633819359e-05, "loss": 1.7403, "step": 20801 }, { "epoch": 0.5245175179081741, "grad_norm": 0.9368360042572021, "learning_rate": 1.2550084734992484e-05, "loss": 1.7379, "step": 20832 }, { "epoch": 0.5252980499288709, "grad_norm": 0.9218337535858154, "learning_rate": 1.247732733176724e-05, "loss": 1.7255, "step": 20863 }, { "epoch": 0.5260785819495676, "grad_norm": 0.9407544136047363, "learning_rate": 1.2404711245201044e-05, "loss": 1.7231, "step": 20894 }, { "epoch": 0.5268591139702643, "grad_norm": 0.9167141318321228, "learning_rate": 1.2332237294756535e-05, "loss": 1.7058, "step": 20925 }, { "epoch": 0.5276396459909609, "grad_norm": 0.9397562742233276, "learning_rate": 1.225990629829241e-05, "loss": 1.7385, "step": 20956 }, { "epoch": 0.5284201780116576, "grad_norm": 0.9333503842353821, "learning_rate": 1.2187719072054136e-05, "loss": 1.7451, "step": 20987 }, { "epoch": 0.5292007100323544, "grad_norm": 0.9979040622711182, "learning_rate": 1.2115676430664735e-05, "loss": 1.7511, "step": 21018 }, { "epoch": 0.529981242053051, "grad_norm": 0.9520873427391052, "learning_rate": 1.2043779187115647e-05, "loss": 1.7272, "step": 21049 }, { "epoch": 0.5307617740737477, "grad_norm": 0.9319959282875061, "learning_rate": 1.1972028152757476e-05, "loss": 1.7381, "step": 21080 }, { "epoch": 0.5315423060944444, "grad_norm": 0.997718870639801, "learning_rate": 1.1900424137290889e-05, "loss": 1.719, "step": 21111 }, { "epoch": 0.532322838115141, "grad_norm": 0.95301353931427, "learning_rate": 1.1828967948757482e-05, "loss": 1.7272, "step": 21142 }, { "epoch": 0.5331033701358378, "grad_norm": 0.9444928169250488, "learning_rate": 1.175766039353062e-05, "loss": 1.7327, "step": 21173 }, { "epoch": 0.5338839021565345, "grad_norm": 0.9271036386489868, "learning_rate": 1.1686502276306382e-05, "loss": 1.7228, "step": 21204 }, { "epoch": 0.5346644341772311, "grad_norm": 0.9555239081382751, "learning_rate": 1.1615494400094445e-05, "loss": 1.7344, "step": 21235 }, { "epoch": 0.5354449661979278, "grad_norm": 0.9300810694694519, "learning_rate": 1.1544637566209029e-05, "loss": 1.7216, "step": 21266 }, { "epoch": 0.5362254982186245, "grad_norm": 0.9187619686126709, "learning_rate": 1.1473932574259886e-05, "loss": 1.7382, "step": 21297 }, { "epoch": 0.5370060302393211, "grad_norm": 0.9264461398124695, "learning_rate": 1.1403380222143247e-05, "loss": 1.7371, "step": 21328 }, { "epoch": 0.5377865622600179, "grad_norm": 0.9823473691940308, "learning_rate": 1.1332981306032808e-05, "loss": 1.7292, "step": 21359 }, { "epoch": 0.5385670942807146, "grad_norm": 0.9366657733917236, "learning_rate": 1.1262736620370762e-05, "loss": 1.7237, "step": 21390 }, { "epoch": 0.5393476263014112, "grad_norm": 0.9524781107902527, "learning_rate": 1.1192646957858854e-05, "loss": 1.7155, "step": 21421 }, { "epoch": 0.5401281583221079, "grad_norm": 0.947833776473999, "learning_rate": 1.1122713109449381e-05, "loss": 1.7106, "step": 21452 }, { "epoch": 0.5409086903428046, "grad_norm": 0.9224444031715393, "learning_rate": 1.105293586433634e-05, "loss": 1.7265, "step": 21483 }, { "epoch": 0.5416892223635014, "grad_norm": 0.9879010319709778, "learning_rate": 1.0983316009946446e-05, "loss": 1.7375, "step": 21514 }, { "epoch": 0.542469754384198, "grad_norm": 0.9554262757301331, "learning_rate": 1.0913854331930282e-05, "loss": 1.7256, "step": 21545 }, { "epoch": 0.5432502864048947, "grad_norm": 0.9443463087081909, "learning_rate": 1.0844551614153456e-05, "loss": 1.7001, "step": 21576 }, { "epoch": 0.5440308184255914, "grad_norm": 0.9143596887588501, "learning_rate": 1.0775408638687725e-05, "loss": 1.7149, "step": 21607 }, { "epoch": 0.544811350446288, "grad_norm": 0.9478058218955994, "learning_rate": 1.0706426185802165e-05, "loss": 1.7199, "step": 21638 }, { "epoch": 0.5455918824669848, "grad_norm": 0.9589979648590088, "learning_rate": 1.0637605033954371e-05, "loss": 1.7145, "step": 21669 }, { "epoch": 0.5463724144876815, "grad_norm": 0.9296464323997498, "learning_rate": 1.05689459597817e-05, "loss": 1.7279, "step": 21700 }, { "epoch": 0.5471529465083781, "grad_norm": 0.9697522521018982, "learning_rate": 1.050044973809246e-05, "loss": 1.7341, "step": 21731 }, { "epoch": 0.5479334785290748, "grad_norm": 0.9401953816413879, "learning_rate": 1.043211714185722e-05, "loss": 1.7276, "step": 21762 }, { "epoch": 0.5487140105497715, "grad_norm": 0.92669677734375, "learning_rate": 1.036394894220003e-05, "loss": 1.7209, "step": 21793 }, { "epoch": 0.5494945425704681, "grad_norm": 0.942746102809906, "learning_rate": 1.0295945908389751e-05, "loss": 1.731, "step": 21824 }, { "epoch": 0.5502750745911649, "grad_norm": 0.9401955604553223, "learning_rate": 1.0228108807831393e-05, "loss": 1.7308, "step": 21855 }, { "epoch": 0.5510556066118616, "grad_norm": 0.9710572957992554, "learning_rate": 1.01604384060574e-05, "loss": 1.7104, "step": 21886 }, { "epoch": 0.5518361386325582, "grad_norm": 0.9225825071334839, "learning_rate": 1.009293546671907e-05, "loss": 1.7299, "step": 21917 }, { "epoch": 0.5526166706532549, "grad_norm": 0.9352388381958008, "learning_rate": 1.002560075157791e-05, "loss": 1.7098, "step": 21948 }, { "epoch": 0.5533972026739516, "grad_norm": 0.9479419589042664, "learning_rate": 9.958435020496995e-06, "loss": 1.7217, "step": 21979 }, { "epoch": 0.5541777346946484, "grad_norm": 0.9639567732810974, "learning_rate": 9.89143903143249e-06, "loss": 1.715, "step": 22010 }, { "epoch": 0.554958266715345, "grad_norm": 0.9379074573516846, "learning_rate": 9.824613540425038e-06, "loss": 1.7221, "step": 22041 }, { "epoch": 0.5557387987360417, "grad_norm": 0.9399924874305725, "learning_rate": 9.757959301591197e-06, "loss": 1.7216, "step": 22072 }, { "epoch": 0.5565193307567384, "grad_norm": 0.9303331971168518, "learning_rate": 9.691477067115017e-06, "loss": 1.7275, "step": 22103 }, { "epoch": 0.557299862777435, "grad_norm": 0.92960524559021, "learning_rate": 9.625167587239467e-06, "loss": 1.7227, "step": 22134 }, { "epoch": 0.5580803947981318, "grad_norm": 0.9338411688804626, "learning_rate": 9.559031610258007e-06, "loss": 1.7078, "step": 22165 }, { "epoch": 0.5588609268188285, "grad_norm": 0.917374312877655, "learning_rate": 9.493069882506164e-06, "loss": 1.7395, "step": 22196 }, { "epoch": 0.5596414588395251, "grad_norm": 0.9369717240333557, "learning_rate": 9.427283148353056e-06, "loss": 1.7157, "step": 22227 }, { "epoch": 0.5604219908602218, "grad_norm": 0.9392804503440857, "learning_rate": 9.361672150193052e-06, "loss": 1.7107, "step": 22258 }, { "epoch": 0.5612025228809185, "grad_norm": 0.926115095615387, "learning_rate": 9.29623762843734e-06, "loss": 1.7057, "step": 22289 }, { "epoch": 0.5619830549016152, "grad_norm": 0.9339826703071594, "learning_rate": 9.230980321505594e-06, "loss": 1.726, "step": 22320 }, { "epoch": 0.5627635869223119, "grad_norm": 0.916395902633667, "learning_rate": 9.165900965817668e-06, "loss": 1.7097, "step": 22351 }, { "epoch": 0.5635441189430086, "grad_norm": 0.9421140551567078, "learning_rate": 9.101000295785245e-06, "loss": 1.7229, "step": 22382 }, { "epoch": 0.5643246509637052, "grad_norm": 0.9253449440002441, "learning_rate": 9.036279043803565e-06, "loss": 1.7258, "step": 22413 }, { "epoch": 0.5651051829844019, "grad_norm": 0.945318341255188, "learning_rate": 8.971737940243147e-06, "loss": 1.7322, "step": 22444 }, { "epoch": 0.5658857150050987, "grad_norm": 0.9573378562927246, "learning_rate": 8.907377713441592e-06, "loss": 1.7183, "step": 22475 }, { "epoch": 0.5666662470257954, "grad_norm": 0.9631168842315674, "learning_rate": 8.843199089695293e-06, "loss": 1.7139, "step": 22506 }, { "epoch": 0.567446779046492, "grad_norm": 0.9413036704063416, "learning_rate": 8.779202793251311e-06, "loss": 1.7369, "step": 22537 }, { "epoch": 0.5682273110671887, "grad_norm": 0.9556379914283752, "learning_rate": 8.715389546299149e-06, "loss": 1.7228, "step": 22568 }, { "epoch": 0.5690078430878854, "grad_norm": 0.9588301181793213, "learning_rate": 8.651760068962617e-06, "loss": 1.7182, "step": 22599 }, { "epoch": 0.569788375108582, "grad_norm": 0.9203631281852722, "learning_rate": 8.588315079291733e-06, "loss": 1.7032, "step": 22630 }, { "epoch": 0.5705689071292788, "grad_norm": 0.9229585528373718, "learning_rate": 8.52505529325457e-06, "loss": 1.7246, "step": 22661 }, { "epoch": 0.5713494391499755, "grad_norm": 0.9430622458457947, "learning_rate": 8.461981424729216e-06, "loss": 1.7339, "step": 22692 }, { "epoch": 0.5721299711706721, "grad_norm": 0.9134554862976074, "learning_rate": 8.399094185495725e-06, "loss": 1.7086, "step": 22723 }, { "epoch": 0.5729105031913688, "grad_norm": 0.9268332719802856, "learning_rate": 8.336394285228017e-06, "loss": 1.719, "step": 22754 }, { "epoch": 0.5736910352120655, "grad_norm": 0.9411998987197876, "learning_rate": 8.273882431485952e-06, "loss": 1.704, "step": 22785 }, { "epoch": 0.5744715672327622, "grad_norm": 1.0843279361724854, "learning_rate": 8.211559329707316e-06, "loss": 1.708, "step": 22816 }, { "epoch": 0.5752520992534589, "grad_norm": 0.9296087026596069, "learning_rate": 8.149425683199823e-06, "loss": 1.7076, "step": 22847 }, { "epoch": 0.5760326312741556, "grad_norm": 0.9319373965263367, "learning_rate": 8.08748219313325e-06, "loss": 1.7292, "step": 22878 }, { "epoch": 0.5768131632948522, "grad_norm": 0.9682207107543945, "learning_rate": 8.025729558531453e-06, "loss": 1.7181, "step": 22909 }, { "epoch": 0.5775936953155489, "grad_norm": 0.9323101043701172, "learning_rate": 7.964168476264508e-06, "loss": 1.7098, "step": 22940 }, { "epoch": 0.5783742273362457, "grad_norm": 0.9210656881332397, "learning_rate": 7.902799641040884e-06, "loss": 1.7053, "step": 22971 }, { "epoch": 0.5791547593569424, "grad_norm": 0.9448071122169495, "learning_rate": 7.841623745399523e-06, "loss": 1.7209, "step": 23002 }, { "epoch": 0.579935291377639, "grad_norm": 0.9426878690719604, "learning_rate": 7.780641479702114e-06, "loss": 1.7286, "step": 23033 }, { "epoch": 0.5807158233983357, "grad_norm": 0.9002141356468201, "learning_rate": 7.719853532125227e-06, "loss": 1.7074, "step": 23064 }, { "epoch": 0.5814963554190324, "grad_norm": 0.9470099210739136, "learning_rate": 7.65926058865258e-06, "loss": 1.7092, "step": 23095 }, { "epoch": 0.582276887439729, "grad_norm": 0.9272165894508362, "learning_rate": 7.598863333067313e-06, "loss": 1.7282, "step": 23126 }, { "epoch": 0.5830574194604258, "grad_norm": 0.9540103673934937, "learning_rate": 7.538662446944253e-06, "loss": 1.7071, "step": 23157 }, { "epoch": 0.5838379514811225, "grad_norm": 0.9641364812850952, "learning_rate": 7.478658609642211e-06, "loss": 1.7331, "step": 23188 }, { "epoch": 0.5846184835018191, "grad_norm": 0.9338697195053101, "learning_rate": 7.418852498296327e-06, "loss": 1.708, "step": 23219 }, { "epoch": 0.5853990155225158, "grad_norm": 0.9395743608474731, "learning_rate": 7.359244787810457e-06, "loss": 1.7168, "step": 23250 }, { "epoch": 0.5861795475432126, "grad_norm": 0.9208729267120361, "learning_rate": 7.299836150849493e-06, "loss": 1.7199, "step": 23281 }, { "epoch": 0.5869600795639092, "grad_norm": 0.9404511451721191, "learning_rate": 7.240627257831847e-06, "loss": 1.7126, "step": 23312 }, { "epoch": 0.5877406115846059, "grad_norm": 0.977336049079895, "learning_rate": 7.1816187769218195e-06, "loss": 1.7337, "step": 23343 }, { "epoch": 0.5885211436053026, "grad_norm": 0.9582931995391846, "learning_rate": 7.1228113740220895e-06, "loss": 1.7145, "step": 23374 }, { "epoch": 0.5893016756259992, "grad_norm": 0.951636791229248, "learning_rate": 7.064205712766226e-06, "loss": 1.7109, "step": 23405 }, { "epoch": 0.5900822076466959, "grad_norm": 0.954495370388031, "learning_rate": 7.005802454511129e-06, "loss": 1.7012, "step": 23436 }, { "epoch": 0.5908627396673927, "grad_norm": 0.9662437438964844, "learning_rate": 6.947602258329639e-06, "loss": 1.6812, "step": 23467 }, { "epoch": 0.5916432716880893, "grad_norm": 0.9717233777046204, "learning_rate": 6.889605781003078e-06, "loss": 1.7219, "step": 23498 }, { "epoch": 0.592423803708786, "grad_norm": 0.9253862500190735, "learning_rate": 6.831813677013776e-06, "loss": 1.7164, "step": 23529 }, { "epoch": 0.5932043357294827, "grad_norm": 0.9557362198829651, "learning_rate": 6.774226598537792e-06, "loss": 1.6982, "step": 23560 }, { "epoch": 0.5939848677501794, "grad_norm": 0.9103747606277466, "learning_rate": 6.716845195437482e-06, "loss": 1.7287, "step": 23591 }, { "epoch": 0.5947653997708761, "grad_norm": 0.9627851247787476, "learning_rate": 6.659670115254168e-06, "loss": 1.7108, "step": 23622 }, { "epoch": 0.5955459317915728, "grad_norm": 0.9510253071784973, "learning_rate": 6.602702003200872e-06, "loss": 1.7212, "step": 23653 }, { "epoch": 0.5963264638122695, "grad_norm": 0.9404213428497314, "learning_rate": 6.545941502154992e-06, "loss": 1.7105, "step": 23684 }, { "epoch": 0.5971069958329661, "grad_norm": 0.938019871711731, "learning_rate": 6.489389252651057e-06, "loss": 1.7291, "step": 23715 }, { "epoch": 0.5978875278536628, "grad_norm": 0.9484905004501343, "learning_rate": 6.4330458928735325e-06, "loss": 1.7272, "step": 23746 }, { "epoch": 0.5986680598743596, "grad_norm": 0.9613280296325684, "learning_rate": 6.376912058649559e-06, "loss": 1.7094, "step": 23777 }, { "epoch": 0.5994485918950562, "grad_norm": 0.9795516729354858, "learning_rate": 6.320988383441845e-06, "loss": 1.7129, "step": 23808 }, { "epoch": 0.6002291239157529, "grad_norm": 0.965208888053894, "learning_rate": 6.265275498341452e-06, "loss": 1.7062, "step": 23839 }, { "epoch": 0.6010096559364496, "grad_norm": 0.9511181116104126, "learning_rate": 6.209774032060714e-06, "loss": 1.7012, "step": 23870 }, { "epoch": 0.6017901879571462, "grad_norm": 0.93194979429245, "learning_rate": 6.1544846109261365e-06, "loss": 1.724, "step": 23901 }, { "epoch": 0.602570719977843, "grad_norm": 0.952998697757721, "learning_rate": 6.099407858871342e-06, "loss": 1.7067, "step": 23932 }, { "epoch": 0.6033512519985397, "grad_norm": 0.9473235607147217, "learning_rate": 6.044544397429958e-06, "loss": 1.7171, "step": 23963 }, { "epoch": 0.6041317840192363, "grad_norm": 0.9270508885383606, "learning_rate": 5.989894845728708e-06, "loss": 1.7113, "step": 23994 }, { "epoch": 0.604912316039933, "grad_norm": 0.9286333322525024, "learning_rate": 5.9354598204803605e-06, "loss": 1.6881, "step": 24025 }, { "epoch": 0.6056928480606297, "grad_norm": 0.9414156079292297, "learning_rate": 5.881239935976762e-06, "loss": 1.715, "step": 24056 }, { "epoch": 0.6064733800813265, "grad_norm": 0.9520138502120972, "learning_rate": 5.827235804081954e-06, "loss": 1.7181, "step": 24087 }, { "epoch": 0.6072539121020231, "grad_norm": 0.9608426690101624, "learning_rate": 5.773448034225221e-06, "loss": 1.7186, "step": 24118 }, { "epoch": 0.6080344441227198, "grad_norm": 0.9268354773521423, "learning_rate": 5.719877233394228e-06, "loss": 1.7047, "step": 24149 }, { "epoch": 0.6088149761434165, "grad_norm": 0.951617956161499, "learning_rate": 5.666524006128191e-06, "loss": 1.7162, "step": 24180 }, { "epoch": 0.6095955081641131, "grad_norm": 0.9590893387794495, "learning_rate": 5.613388954511015e-06, "loss": 1.7105, "step": 24211 }, { "epoch": 0.6103760401848098, "grad_norm": 0.9623116850852966, "learning_rate": 5.560472678164552e-06, "loss": 1.6998, "step": 24242 }, { "epoch": 0.6111565722055066, "grad_norm": 0.9415599703788757, "learning_rate": 5.507775774241775e-06, "loss": 1.715, "step": 24273 }, { "epoch": 0.6119371042262032, "grad_norm": 0.979045569896698, "learning_rate": 5.4552988374200945e-06, "loss": 1.7046, "step": 24304 }, { "epoch": 0.6127176362468999, "grad_norm": 0.9478979706764221, "learning_rate": 5.403042459894597e-06, "loss": 1.7054, "step": 24335 }, { "epoch": 0.6134981682675966, "grad_norm": 0.9414293766021729, "learning_rate": 5.3510072313714135e-06, "loss": 1.7, "step": 24366 }, { "epoch": 0.6142787002882932, "grad_norm": 0.912340521812439, "learning_rate": 5.2991937390610205e-06, "loss": 1.7149, "step": 24397 }, { "epoch": 0.61505923230899, "grad_norm": 0.9561455845832825, "learning_rate": 5.247602567671625e-06, "loss": 1.7294, "step": 24428 }, { "epoch": 0.6158397643296867, "grad_norm": 0.9352984428405762, "learning_rate": 5.196234299402603e-06, "loss": 1.7266, "step": 24459 }, { "epoch": 0.6166202963503833, "grad_norm": 0.9225531220436096, "learning_rate": 5.145089513937865e-06, "loss": 1.6973, "step": 24490 }, { "epoch": 0.61740082837108, "grad_norm": 0.9766479134559631, "learning_rate": 5.094168788439369e-06, "loss": 1.7052, "step": 24521 }, { "epoch": 0.6181813603917767, "grad_norm": 0.9518314003944397, "learning_rate": 5.043472697540594e-06, "loss": 1.7026, "step": 24552 }, { "epoch": 0.6189618924124735, "grad_norm": 0.9657760858535767, "learning_rate": 4.993001813340012e-06, "loss": 1.7142, "step": 24583 }, { "epoch": 0.6197424244331701, "grad_norm": 0.9249383211135864, "learning_rate": 4.942756705394702e-06, "loss": 1.6983, "step": 24614 }, { "epoch": 0.6205229564538668, "grad_norm": 0.9539477825164795, "learning_rate": 4.892737940713884e-06, "loss": 1.7013, "step": 24645 }, { "epoch": 0.6213034884745635, "grad_norm": 0.959993302822113, "learning_rate": 4.842946083752511e-06, "loss": 1.7214, "step": 24676 }, { "epoch": 0.6220840204952601, "grad_norm": 0.9549261331558228, "learning_rate": 4.79338169640493e-06, "loss": 1.7196, "step": 24707 }, { "epoch": 0.6228645525159568, "grad_norm": 0.9377399682998657, "learning_rate": 4.74404533799851e-06, "loss": 1.7114, "step": 24738 }, { "epoch": 0.6236450845366536, "grad_norm": 0.9832403063774109, "learning_rate": 4.694937565287344e-06, "loss": 1.7185, "step": 24769 }, { "epoch": 0.6244256165573502, "grad_norm": 0.9397412538528442, "learning_rate": 4.646058932445985e-06, "loss": 1.7053, "step": 24800 }, { "epoch": 0.6252061485780469, "grad_norm": 0.9293022751808167, "learning_rate": 4.597409991063148e-06, "loss": 1.701, "step": 24831 }, { "epoch": 0.6259866805987436, "grad_norm": 0.9333553314208984, "learning_rate": 4.5489912901355375e-06, "loss": 1.7236, "step": 24862 }, { "epoch": 0.6267672126194402, "grad_norm": 0.939677894115448, "learning_rate": 4.500803376061608e-06, "loss": 1.7069, "step": 24893 }, { "epoch": 0.627547744640137, "grad_norm": 0.9724187850952148, "learning_rate": 4.45284679263541e-06, "loss": 1.7087, "step": 24924 }, { "epoch": 0.6283282766608337, "grad_norm": 0.9568901658058167, "learning_rate": 4.4051220810404775e-06, "loss": 1.7134, "step": 24955 }, { "epoch": 0.6291088086815303, "grad_norm": 0.9412270784378052, "learning_rate": 4.3576297798437025e-06, "loss": 1.7163, "step": 24986 }, { "epoch": 0.629889340702227, "grad_norm": 0.9624278545379639, "learning_rate": 4.3103704249892436e-06, "loss": 1.6989, "step": 25017 }, { "epoch": 0.6306698727229237, "grad_norm": 0.9283859729766846, "learning_rate": 4.263344549792487e-06, "loss": 1.7004, "step": 25048 }, { "epoch": 0.6314504047436205, "grad_norm": 0.9795125722885132, "learning_rate": 4.216552684934056e-06, "loss": 1.6937, "step": 25079 }, { "epoch": 0.6322309367643171, "grad_norm": 0.9757822751998901, "learning_rate": 4.169995358453777e-06, "loss": 1.6957, "step": 25110 }, { "epoch": 0.6330114687850138, "grad_norm": 0.9491864442825317, "learning_rate": 4.123673095744757e-06, "loss": 1.7041, "step": 25141 }, { "epoch": 0.6337920008057105, "grad_norm": 0.9521167278289795, "learning_rate": 4.077586419547435e-06, "loss": 1.7064, "step": 25172 }, { "epoch": 0.6345725328264071, "grad_norm": 0.9412490725517273, "learning_rate": 4.03173584994368e-06, "loss": 1.7025, "step": 25203 }, { "epoch": 0.6353530648471039, "grad_norm": 0.9738706350326538, "learning_rate": 3.986121904350948e-06, "loss": 1.6997, "step": 25234 }, { "epoch": 0.6361335968678006, "grad_norm": 0.9621003866195679, "learning_rate": 3.940745097516407e-06, "loss": 1.7047, "step": 25265 }, { "epoch": 0.6369141288884972, "grad_norm": 0.9555737972259521, "learning_rate": 3.89560594151116e-06, "loss": 1.7092, "step": 25296 }, { "epoch": 0.6376946609091939, "grad_norm": 0.9567461609840393, "learning_rate": 3.850704945724456e-06, "loss": 1.7252, "step": 25327 }, { "epoch": 0.6384751929298906, "grad_norm": 0.9793210029602051, "learning_rate": 3.8060426168579077e-06, "loss": 1.7053, "step": 25358 }, { "epoch": 0.6392557249505872, "grad_norm": 0.9398420453071594, "learning_rate": 3.7616194589198407e-06, "loss": 1.6869, "step": 25389 }, { "epoch": 0.640036256971284, "grad_norm": 0.9754509329795837, "learning_rate": 3.7174359732195574e-06, "loss": 1.7017, "step": 25420 }, { "epoch": 0.6408167889919807, "grad_norm": 0.933102011680603, "learning_rate": 3.673492658361677e-06, "loss": 1.7146, "step": 25451 }, { "epoch": 0.6415973210126773, "grad_norm": 0.9471291303634644, "learning_rate": 3.6297900102405467e-06, "loss": 1.7169, "step": 25482 }, { "epoch": 0.642377853033374, "grad_norm": 0.9755620956420898, "learning_rate": 3.586328522034607e-06, "loss": 1.721, "step": 25513 }, { "epoch": 0.6431583850540707, "grad_norm": 0.9221309423446655, "learning_rate": 3.543108684200838e-06, "loss": 1.7187, "step": 25544 }, { "epoch": 0.6439389170747675, "grad_norm": 0.9288764595985413, "learning_rate": 3.5001309844692464e-06, "loss": 1.7214, "step": 25575 }, { "epoch": 0.6447194490954641, "grad_norm": 0.9390939474105835, "learning_rate": 3.4573959078373215e-06, "loss": 1.7095, "step": 25606 }, { "epoch": 0.6454999811161608, "grad_norm": 0.9171627163887024, "learning_rate": 3.4149039365646063e-06, "loss": 1.678, "step": 25637 }, { "epoch": 0.6462805131368575, "grad_norm": 0.9164687395095825, "learning_rate": 3.3726555501672143e-06, "loss": 1.6947, "step": 25668 }, { "epoch": 0.6470610451575541, "grad_norm": 0.9990100264549255, "learning_rate": 3.33065122541244e-06, "loss": 1.6977, "step": 25699 }, { "epoch": 0.6478415771782509, "grad_norm": 0.9423044919967651, "learning_rate": 3.288891436313385e-06, "loss": 1.7054, "step": 25730 }, { "epoch": 0.6486221091989476, "grad_norm": 0.9439398646354675, "learning_rate": 3.2473766541235963e-06, "loss": 1.6939, "step": 25761 }, { "epoch": 0.6494026412196442, "grad_norm": 0.9685223698616028, "learning_rate": 3.2061073473317466e-06, "loss": 1.6861, "step": 25792 }, { "epoch": 0.6501831732403409, "grad_norm": 0.9343461990356445, "learning_rate": 3.1650839816563444e-06, "loss": 1.7053, "step": 25823 }, { "epoch": 0.6509637052610376, "grad_norm": 0.9476882815361023, "learning_rate": 3.1243070200405093e-06, "loss": 1.6878, "step": 25854 }, { "epoch": 0.6517442372817343, "grad_norm": 0.9528893828392029, "learning_rate": 3.0837769226467e-06, "loss": 1.6961, "step": 25885 }, { "epoch": 0.652524769302431, "grad_norm": 0.9664431214332581, "learning_rate": 3.0434941468515666e-06, "loss": 1.7044, "step": 25916 }, { "epoch": 0.6533053013231277, "grad_norm": 0.9796096682548523, "learning_rate": 3.003459147240753e-06, "loss": 1.7046, "step": 25947 }, { "epoch": 0.6540858333438243, "grad_norm": 0.9578924179077148, "learning_rate": 2.9636723756037875e-06, "loss": 1.7065, "step": 25978 }, { "epoch": 0.654866365364521, "grad_norm": 0.9339287281036377, "learning_rate": 2.9241342809289833e-06, "loss": 1.7154, "step": 26009 }, { "epoch": 0.6556468973852178, "grad_norm": 0.9628824591636658, "learning_rate": 2.8848453093983594e-06, "loss": 1.7059, "step": 26040 }, { "epoch": 0.6564274294059144, "grad_norm": 0.9443020820617676, "learning_rate": 2.8458059043826257e-06, "loss": 1.6936, "step": 26071 }, { "epoch": 0.6572079614266111, "grad_norm": 0.9134575128555298, "learning_rate": 2.807016506436172e-06, "loss": 1.6865, "step": 26102 }, { "epoch": 0.6579884934473078, "grad_norm": 0.9755576848983765, "learning_rate": 2.7684775532920566e-06, "loss": 1.6972, "step": 26133 }, { "epoch": 0.6587690254680045, "grad_norm": 0.9576392769813538, "learning_rate": 2.7301894798571425e-06, "loss": 1.7069, "step": 26164 }, { "epoch": 0.6595495574887011, "grad_norm": 0.9519221782684326, "learning_rate": 2.6921527182071386e-06, "loss": 1.6969, "step": 26195 }, { "epoch": 0.6603300895093979, "grad_norm": 0.9404755234718323, "learning_rate": 2.654367697581725e-06, "loss": 1.7054, "step": 26226 }, { "epoch": 0.6611106215300946, "grad_norm": 0.977472186088562, "learning_rate": 2.6168348443797175e-06, "loss": 1.7173, "step": 26257 }, { "epoch": 0.6618911535507912, "grad_norm": 0.9931960105895996, "learning_rate": 2.5795545821542757e-06, "loss": 1.6994, "step": 26288 }, { "epoch": 0.6626716855714879, "grad_norm": 0.9330551028251648, "learning_rate": 2.54252733160808e-06, "loss": 1.6796, "step": 26319 }, { "epoch": 0.6634522175921846, "grad_norm": 0.955140233039856, "learning_rate": 2.5057535105886294e-06, "loss": 1.7053, "step": 26350 }, { "epoch": 0.6642327496128813, "grad_norm": 0.962437629699707, "learning_rate": 2.4692335340834953e-06, "loss": 1.7038, "step": 26381 }, { "epoch": 0.665013281633578, "grad_norm": 0.9602552652359009, "learning_rate": 2.432967814215639e-06, "loss": 1.7006, "step": 26412 }, { "epoch": 0.6657938136542747, "grad_norm": 0.9575433731079102, "learning_rate": 2.396956760238794e-06, "loss": 1.7182, "step": 26443 }, { "epoch": 0.6665743456749713, "grad_norm": 0.9685966372489929, "learning_rate": 2.361200778532796e-06, "loss": 1.701, "step": 26474 }, { "epoch": 0.667354877695668, "grad_norm": 0.9528165459632874, "learning_rate": 2.325700272599049e-06, "loss": 1.7125, "step": 26505 }, { "epoch": 0.6681354097163648, "grad_norm": 0.9569064974784851, "learning_rate": 2.2904556430559415e-06, "loss": 1.7029, "step": 26536 }, { "epoch": 0.6689159417370614, "grad_norm": 0.9684230089187622, "learning_rate": 2.2554672876343106e-06, "loss": 1.7204, "step": 26567 }, { "epoch": 0.6696964737577581, "grad_norm": 0.9554132223129272, "learning_rate": 2.220735601173002e-06, "loss": 1.7026, "step": 26598 }, { "epoch": 0.6704770057784548, "grad_norm": 0.9310318231582642, "learning_rate": 2.186260975614382e-06, "loss": 1.6978, "step": 26629 }, { "epoch": 0.6712575377991515, "grad_norm": 0.9740652441978455, "learning_rate": 2.1520437999999034e-06, "loss": 1.6956, "step": 26660 }, { "epoch": 0.6720380698198481, "grad_norm": 0.9701205492019653, "learning_rate": 2.1180844604657526e-06, "loss": 1.7009, "step": 26691 }, { "epoch": 0.6728186018405449, "grad_norm": 0.9373700618743896, "learning_rate": 2.084383340238455e-06, "loss": 1.7104, "step": 26722 }, { "epoch": 0.6735991338612416, "grad_norm": 1.0050245523452759, "learning_rate": 2.0509408196305704e-06, "loss": 1.7123, "step": 26753 }, { "epoch": 0.6743796658819382, "grad_norm": 0.9987810850143433, "learning_rate": 2.017757276036403e-06, "loss": 1.7051, "step": 26784 }, { "epoch": 0.6751601979026349, "grad_norm": 0.951255738735199, "learning_rate": 1.984833083927726e-06, "loss": 1.6951, "step": 26815 }, { "epoch": 0.6759407299233317, "grad_norm": 0.9656464457511902, "learning_rate": 1.952168614849581e-06, "loss": 1.6894, "step": 26846 }, { "epoch": 0.6767212619440283, "grad_norm": 0.977102518081665, "learning_rate": 1.919764237416058e-06, "loss": 1.718, "step": 26877 }, { "epoch": 0.677501793964725, "grad_norm": 0.942525327205658, "learning_rate": 1.8876203173061463e-06, "loss": 1.7065, "step": 26908 }, { "epoch": 0.6782823259854217, "grad_norm": 0.963570773601532, "learning_rate": 1.8557372172596206e-06, "loss": 1.6983, "step": 26939 }, { "epoch": 0.6790628580061183, "grad_norm": 0.9753343462944031, "learning_rate": 1.8241152970729341e-06, "loss": 1.71, "step": 26970 }, { "epoch": 0.679843390026815, "grad_norm": 0.919635534286499, "learning_rate": 1.7927549135951572e-06, "loss": 1.7043, "step": 27001 }, { "epoch": 0.6806239220475118, "grad_norm": 0.9881942868232727, "learning_rate": 1.7616564207239477e-06, "loss": 1.7025, "step": 27032 }, { "epoch": 0.6814044540682084, "grad_norm": 0.9447351694107056, "learning_rate": 1.730820169401584e-06, "loss": 1.6991, "step": 27063 }, { "epoch": 0.6821849860889051, "grad_norm": 0.9544949531555176, "learning_rate": 1.7002465076109558e-06, "loss": 1.7078, "step": 27094 }, { "epoch": 0.6829655181096018, "grad_norm": 0.9443686604499817, "learning_rate": 1.6699357803716898e-06, "loss": 1.7168, "step": 27125 }, { "epoch": 0.6837460501302985, "grad_norm": 0.974307119846344, "learning_rate": 1.6398883297362305e-06, "loss": 1.7147, "step": 27156 }, { "epoch": 0.6845265821509952, "grad_norm": 0.9629453420639038, "learning_rate": 1.6101044947859606e-06, "loss": 1.7204, "step": 27187 }, { "epoch": 0.6853071141716919, "grad_norm": 0.95462965965271, "learning_rate": 1.5805846116274114e-06, "loss": 1.716, "step": 27218 }, { "epoch": 0.6860876461923886, "grad_norm": 0.945711612701416, "learning_rate": 1.5513290133884611e-06, "loss": 1.6973, "step": 27249 }, { "epoch": 0.6868681782130852, "grad_norm": 0.9599740505218506, "learning_rate": 1.5223380302145512e-06, "loss": 1.6957, "step": 27280 }, { "epoch": 0.6876487102337819, "grad_norm": 0.9428503513336182, "learning_rate": 1.4936119892649925e-06, "loss": 1.7194, "step": 27311 }, { "epoch": 0.6884292422544787, "grad_norm": 0.948523223400116, "learning_rate": 1.4651512147092482e-06, "loss": 1.6983, "step": 27342 }, { "epoch": 0.6892097742751753, "grad_norm": 0.9603691101074219, "learning_rate": 1.4369560277232908e-06, "loss": 1.7155, "step": 27373 }, { "epoch": 0.689990306295872, "grad_norm": 0.955329179763794, "learning_rate": 1.409026746485978e-06, "loss": 1.6903, "step": 27404 }, { "epoch": 0.6907708383165687, "grad_norm": 0.9583889842033386, "learning_rate": 1.3813636861754464e-06, "loss": 1.6942, "step": 27435 }, { "epoch": 0.6915513703372653, "grad_norm": 0.9602280259132385, "learning_rate": 1.3539671589655773e-06, "loss": 1.6889, "step": 27466 } ], "logging_steps": 31, "max_steps": 30517, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 3052, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.9185168775775257e+19, "train_batch_size": 16, "trial_name": null, "trial_params": null }