{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.0, "eval_steps": 500, "global_step": 2500, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.008, "grad_norm": 0.3058140892912027, "learning_rate": 1.6000000000000001e-06, "loss": 0.5786, "step": 10 }, { "epoch": 0.016, "grad_norm": 0.31255981400696337, "learning_rate": 3.2000000000000003e-06, "loss": 0.5032, "step": 20 }, { "epoch": 0.024, "grad_norm": 0.16719945295611136, "learning_rate": 4.800000000000001e-06, "loss": 0.3901, "step": 30 }, { "epoch": 0.032, "grad_norm": 0.11095534933859139, "learning_rate": 6.4000000000000006e-06, "loss": 0.2978, "step": 40 }, { "epoch": 0.04, "grad_norm": 0.1114769024399936, "learning_rate": 8.000000000000001e-06, "loss": 0.2631, "step": 50 }, { "epoch": 0.048, "grad_norm": 0.10411559733650633, "learning_rate": 9.600000000000001e-06, "loss": 0.242, "step": 60 }, { "epoch": 0.056, "grad_norm": 0.08350193155299954, "learning_rate": 1.1200000000000001e-05, "loss": 0.2214, "step": 70 }, { "epoch": 0.064, "grad_norm": 0.10529290900139675, "learning_rate": 1.2800000000000001e-05, "loss": 0.2134, "step": 80 }, { "epoch": 0.072, "grad_norm": 0.09284266871657494, "learning_rate": 1.4400000000000001e-05, "loss": 0.2122, "step": 90 }, { "epoch": 0.08, "grad_norm": 0.09938443320990606, "learning_rate": 1.6000000000000003e-05, "loss": 0.2035, "step": 100 }, { "epoch": 0.088, "grad_norm": 0.10023595945326495, "learning_rate": 1.76e-05, "loss": 0.1841, "step": 110 }, { "epoch": 0.096, "grad_norm": 0.0695781552706901, "learning_rate": 1.9200000000000003e-05, "loss": 0.1877, "step": 120 }, { "epoch": 0.104, "grad_norm": 0.07598876239857685, "learning_rate": 1.9999781283802247e-05, "loss": 0.1868, "step": 130 }, { "epoch": 0.112, "grad_norm": 0.08623180966304644, "learning_rate": 1.999803161162393e-05, "loss": 0.2111, "step": 140 }, { "epoch": 0.12, "grad_norm": 0.08260588578713098, "learning_rate": 1.999453257340926e-05, "loss": 0.1958, "step": 150 }, { "epoch": 0.128, "grad_norm": 0.07792179475378579, "learning_rate": 1.9989284781388617e-05, "loss": 0.1835, "step": 160 }, { "epoch": 0.136, "grad_norm": 0.07928527753977362, "learning_rate": 1.9982289153773648e-05, "loss": 0.1912, "step": 170 }, { "epoch": 0.144, "grad_norm": 0.08357416949011726, "learning_rate": 1.9973546914596622e-05, "loss": 0.1885, "step": 180 }, { "epoch": 0.152, "grad_norm": 0.08456641175983287, "learning_rate": 1.996305959349627e-05, "loss": 0.1841, "step": 190 }, { "epoch": 0.16, "grad_norm": 0.0901579451770012, "learning_rate": 1.9950829025450116e-05, "loss": 0.1806, "step": 200 }, { "epoch": 0.168, "grad_norm": 0.07982438004177704, "learning_rate": 1.993685735045343e-05, "loss": 0.1769, "step": 210 }, { "epoch": 0.176, "grad_norm": 0.10398225393987126, "learning_rate": 1.9921147013144782e-05, "loss": 0.1895, "step": 220 }, { "epoch": 0.184, "grad_norm": 0.1099351584849273, "learning_rate": 1.9903700762378303e-05, "loss": 0.18, "step": 230 }, { "epoch": 0.192, "grad_norm": 0.09703206171895983, "learning_rate": 1.9884521650742718e-05, "loss": 0.17, "step": 240 }, { "epoch": 0.2, "grad_norm": 0.09289058414749175, "learning_rate": 1.9863613034027224e-05, "loss": 0.1699, "step": 250 }, { "epoch": 0.208, "grad_norm": 0.08208598708089178, "learning_rate": 1.9840978570634338e-05, "loss": 0.1843, "step": 260 }, { "epoch": 0.216, "grad_norm": 0.08483871139912566, "learning_rate": 1.9816622220939762e-05, "loss": 0.1756, "step": 270 }, { "epoch": 0.224, "grad_norm": 0.08382234577690319, "learning_rate": 1.9790548246599447e-05, "loss": 0.1746, "step": 280 }, { "epoch": 0.232, "grad_norm": 0.11261586530879426, "learning_rate": 1.976276120980393e-05, "loss": 0.1739, "step": 290 }, { "epoch": 0.24, "grad_norm": 0.0777808773898071, "learning_rate": 1.973326597248006e-05, "loss": 0.1791, "step": 300 }, { "epoch": 0.248, "grad_norm": 0.0786352615035203, "learning_rate": 1.9702067695440333e-05, "loss": 0.1732, "step": 310 }, { "epoch": 0.256, "grad_norm": 0.07355004150569275, "learning_rate": 1.966917183747987e-05, "loss": 0.1737, "step": 320 }, { "epoch": 0.264, "grad_norm": 0.08213670035333703, "learning_rate": 1.9634584154421316e-05, "loss": 0.1777, "step": 330 }, { "epoch": 0.272, "grad_norm": 0.07217668674817151, "learning_rate": 1.95983106981077e-05, "loss": 0.1695, "step": 340 }, { "epoch": 0.28, "grad_norm": 0.07669208676566047, "learning_rate": 1.9560357815343577e-05, "loss": 0.1717, "step": 350 }, { "epoch": 0.288, "grad_norm": 0.07736091266235384, "learning_rate": 1.9520732146784493e-05, "loss": 0.1718, "step": 360 }, { "epoch": 0.296, "grad_norm": 0.08315079692157626, "learning_rate": 1.947944062577507e-05, "loss": 0.1776, "step": 370 }, { "epoch": 0.304, "grad_norm": 0.09135194147148734, "learning_rate": 1.9436490477135877e-05, "loss": 0.173, "step": 380 }, { "epoch": 0.312, "grad_norm": 0.0809078717299552, "learning_rate": 1.93918892158993e-05, "loss": 0.1773, "step": 390 }, { "epoch": 0.32, "grad_norm": 0.07427395281421514, "learning_rate": 1.934564464599461e-05, "loss": 0.17, "step": 400 }, { "epoch": 0.328, "grad_norm": 0.07398454431370932, "learning_rate": 1.9297764858882516e-05, "loss": 0.181, "step": 410 }, { "epoch": 0.336, "grad_norm": 0.0792882992648581, "learning_rate": 1.924825823213939e-05, "loss": 0.1627, "step": 420 }, { "epoch": 0.344, "grad_norm": 0.0861604357608397, "learning_rate": 1.9197133427991437e-05, "loss": 0.1731, "step": 430 }, { "epoch": 0.352, "grad_norm": 0.07755055594882805, "learning_rate": 1.9144399391799043e-05, "loss": 0.1703, "step": 440 }, { "epoch": 0.36, "grad_norm": 0.08634385297265851, "learning_rate": 1.909006535049163e-05, "loss": 0.1695, "step": 450 }, { "epoch": 0.368, "grad_norm": 0.07460615095270867, "learning_rate": 1.903414081095315e-05, "loss": 0.1808, "step": 460 }, { "epoch": 0.376, "grad_norm": 0.08366554725021544, "learning_rate": 1.897663555835872e-05, "loss": 0.1687, "step": 470 }, { "epoch": 0.384, "grad_norm": 0.11050801098711051, "learning_rate": 1.8917559654462474e-05, "loss": 0.1693, "step": 480 }, { "epoch": 0.392, "grad_norm": 0.07102922397109013, "learning_rate": 1.8856923435837024e-05, "loss": 0.1654, "step": 490 }, { "epoch": 0.4, "grad_norm": 0.07992253400672722, "learning_rate": 1.879473751206489e-05, "loss": 0.164, "step": 500 }, { "epoch": 0.408, "grad_norm": 0.08055086025032311, "learning_rate": 1.8731012763882132e-05, "loss": 0.1713, "step": 510 }, { "epoch": 0.416, "grad_norm": 0.0737582298851665, "learning_rate": 1.8665760341274505e-05, "loss": 0.1691, "step": 520 }, { "epoch": 0.424, "grad_norm": 0.07526226505268085, "learning_rate": 1.859899166152657e-05, "loss": 0.1612, "step": 530 }, { "epoch": 0.432, "grad_norm": 0.07994952744363151, "learning_rate": 1.8530718407223976e-05, "loss": 0.1741, "step": 540 }, { "epoch": 0.44, "grad_norm": 0.06771882882851786, "learning_rate": 1.8460952524209355e-05, "loss": 0.1771, "step": 550 }, { "epoch": 0.448, "grad_norm": 0.07925796602812565, "learning_rate": 1.8389706219492147e-05, "loss": 0.1658, "step": 560 }, { "epoch": 0.456, "grad_norm": 0.07737563218705418, "learning_rate": 1.831699195911272e-05, "loss": 0.1713, "step": 570 }, { "epoch": 0.464, "grad_norm": 0.07405191205739156, "learning_rate": 1.8242822465961177e-05, "loss": 0.1687, "step": 580 }, { "epoch": 0.472, "grad_norm": 0.08527959453273458, "learning_rate": 1.8167210717551224e-05, "loss": 0.1645, "step": 590 }, { "epoch": 0.48, "grad_norm": 0.08246899769527764, "learning_rate": 1.8090169943749477e-05, "loss": 0.1606, "step": 600 }, { "epoch": 0.488, "grad_norm": 0.07147580637020562, "learning_rate": 1.8011713624460608e-05, "loss": 0.1628, "step": 610 }, { "epoch": 0.496, "grad_norm": 0.07722727076416155, "learning_rate": 1.793185548726878e-05, "loss": 0.1759, "step": 620 }, { "epoch": 0.504, "grad_norm": 0.06844389954410786, "learning_rate": 1.785060950503568e-05, "loss": 0.1748, "step": 630 }, { "epoch": 0.512, "grad_norm": 0.07174329525930123, "learning_rate": 1.7767989893455696e-05, "loss": 0.1713, "step": 640 }, { "epoch": 0.52, "grad_norm": 0.08532370630535416, "learning_rate": 1.7684011108568593e-05, "loss": 0.1622, "step": 650 }, { "epoch": 0.528, "grad_norm": 0.06580211643615225, "learning_rate": 1.759868784423009e-05, "loss": 0.1593, "step": 660 }, { "epoch": 0.536, "grad_norm": 0.07333177241207892, "learning_rate": 1.7512035029540887e-05, "loss": 0.1611, "step": 670 }, { "epoch": 0.544, "grad_norm": 0.08247160155974288, "learning_rate": 1.74240678262345e-05, "loss": 0.157, "step": 680 }, { "epoch": 0.552, "grad_norm": 0.06614899961508676, "learning_rate": 1.73348016260244e-05, "loss": 0.1635, "step": 690 }, { "epoch": 0.56, "grad_norm": 0.08619484121470282, "learning_rate": 1.7244252047910893e-05, "loss": 0.1675, "step": 700 }, { "epoch": 0.568, "grad_norm": 0.08412526181881948, "learning_rate": 1.7152434935448257e-05, "loss": 0.1562, "step": 710 }, { "epoch": 0.576, "grad_norm": 0.07223298445136259, "learning_rate": 1.705936635397259e-05, "loss": 0.1687, "step": 720 }, { "epoch": 0.584, "grad_norm": 0.06857251106754064, "learning_rate": 1.6965062587790823e-05, "loss": 0.1663, "step": 730 }, { "epoch": 0.592, "grad_norm": 0.06696835945973133, "learning_rate": 1.6869540137331445e-05, "loss": 0.1674, "step": 740 }, { "epoch": 0.6, "grad_norm": 0.06686201633689924, "learning_rate": 1.6772815716257414e-05, "loss": 0.174, "step": 750 }, { "epoch": 0.608, "grad_norm": 0.08123286118785028, "learning_rate": 1.667490624854173e-05, "loss": 0.1667, "step": 760 }, { "epoch": 0.616, "grad_norm": 0.14148707567389088, "learning_rate": 1.6575828865506246e-05, "loss": 0.1682, "step": 770 }, { "epoch": 0.624, "grad_norm": 0.06485016340695926, "learning_rate": 1.647560090282419e-05, "loss": 0.1574, "step": 780 }, { "epoch": 0.632, "grad_norm": 0.0867889552062448, "learning_rate": 1.63742398974869e-05, "loss": 0.1629, "step": 790 }, { "epoch": 0.64, "grad_norm": 0.06811973631019821, "learning_rate": 1.6271763584735373e-05, "loss": 0.163, "step": 800 }, { "epoch": 0.648, "grad_norm": 0.08610829160689334, "learning_rate": 1.616818989495711e-05, "loss": 0.1584, "step": 810 }, { "epoch": 0.656, "grad_norm": 0.06829743878657456, "learning_rate": 1.6063536950548825e-05, "loss": 0.1611, "step": 820 }, { "epoch": 0.664, "grad_norm": 0.07996846689663462, "learning_rate": 1.595782306274553e-05, "loss": 0.1728, "step": 830 }, { "epoch": 0.672, "grad_norm": 0.06827784189424038, "learning_rate": 1.5851066728416617e-05, "loss": 0.1657, "step": 840 }, { "epoch": 0.68, "grad_norm": 0.09993992156719474, "learning_rate": 1.5743286626829437e-05, "loss": 0.1577, "step": 850 }, { "epoch": 0.688, "grad_norm": 0.0904528328527498, "learning_rate": 1.5634501616380967e-05, "loss": 0.1738, "step": 860 }, { "epoch": 0.696, "grad_norm": 0.07536916136620567, "learning_rate": 1.5524730731298136e-05, "loss": 0.1724, "step": 870 }, { "epoch": 0.704, "grad_norm": 0.07771163955093588, "learning_rate": 1.541399317830738e-05, "loss": 0.1587, "step": 880 }, { "epoch": 0.712, "grad_norm": 0.06870339623694065, "learning_rate": 1.530230833327405e-05, "loss": 0.1662, "step": 890 }, { "epoch": 0.72, "grad_norm": 0.08541842938568546, "learning_rate": 1.5189695737812153e-05, "loss": 0.1625, "step": 900 }, { "epoch": 0.728, "grad_norm": 0.07174614527960509, "learning_rate": 1.5076175095865171e-05, "loss": 0.1716, "step": 910 }, { "epoch": 0.736, "grad_norm": 0.07173726830388383, "learning_rate": 1.4961766270258422e-05, "loss": 0.1564, "step": 920 }, { "epoch": 0.744, "grad_norm": 0.07774148753609791, "learning_rate": 1.4846489279223653e-05, "loss": 0.1688, "step": 930 }, { "epoch": 0.752, "grad_norm": 0.07047837351321116, "learning_rate": 1.473036429289641e-05, "loss": 0.1671, "step": 940 }, { "epoch": 0.76, "grad_norm": 0.07209116510890862, "learning_rate": 1.461341162978688e-05, "loss": 0.1636, "step": 950 }, { "epoch": 0.768, "grad_norm": 0.0825643122647062, "learning_rate": 1.4495651753224706e-05, "loss": 0.1589, "step": 960 }, { "epoch": 0.776, "grad_norm": 0.09574339039828667, "learning_rate": 1.437710526777852e-05, "loss": 0.1663, "step": 970 }, { "epoch": 0.784, "grad_norm": 0.07179601153688872, "learning_rate": 1.4257792915650728e-05, "loss": 0.1583, "step": 980 }, { "epoch": 0.792, "grad_norm": 0.07652431647542295, "learning_rate": 1.4137735573048232e-05, "loss": 0.1514, "step": 990 }, { "epoch": 0.8, "grad_norm": 0.0686398495128086, "learning_rate": 1.4016954246529697e-05, "loss": 0.1504, "step": 1000 }, { "epoch": 0.808, "grad_norm": 0.06508028906617441, "learning_rate": 1.3895470069330003e-05, "loss": 0.1599, "step": 1010 }, { "epoch": 0.816, "grad_norm": 0.07608940173232373, "learning_rate": 1.377330429766256e-05, "loss": 0.1612, "step": 1020 }, { "epoch": 0.824, "grad_norm": 0.06899267345664216, "learning_rate": 1.3650478307000059e-05, "loss": 0.1575, "step": 1030 }, { "epoch": 0.832, "grad_norm": 0.06256387504627259, "learning_rate": 1.3527013588334415e-05, "loss": 0.1652, "step": 1040 }, { "epoch": 0.84, "grad_norm": 0.07396943401991835, "learning_rate": 1.3402931744416432e-05, "loss": 0.1538, "step": 1050 }, { "epoch": 0.848, "grad_norm": 0.06825533554005977, "learning_rate": 1.3278254485975977e-05, "loss": 0.16, "step": 1060 }, { "epoch": 0.856, "grad_norm": 0.07700315050313282, "learning_rate": 1.3153003627923217e-05, "loss": 0.1551, "step": 1070 }, { "epoch": 0.864, "grad_norm": 0.07759623789607299, "learning_rate": 1.3027201085531633e-05, "loss": 0.1635, "step": 1080 }, { "epoch": 0.872, "grad_norm": 0.07689613377715925, "learning_rate": 1.2900868870603502e-05, "loss": 0.1587, "step": 1090 }, { "epoch": 0.88, "grad_norm": 0.07385835492768326, "learning_rate": 1.2774029087618448e-05, "loss": 0.1709, "step": 1100 }, { "epoch": 0.888, "grad_norm": 0.06043191054467158, "learning_rate": 1.2646703929865817e-05, "loss": 0.1637, "step": 1110 }, { "epoch": 0.896, "grad_norm": 0.07533708016400092, "learning_rate": 1.2518915675561482e-05, "loss": 0.1682, "step": 1120 }, { "epoch": 0.904, "grad_norm": 0.06791483038650707, "learning_rate": 1.2390686683949799e-05, "loss": 0.1444, "step": 1130 }, { "epoch": 0.912, "grad_norm": 0.0815779013214324, "learning_rate": 1.2262039391391405e-05, "loss": 0.1569, "step": 1140 }, { "epoch": 0.92, "grad_norm": 0.07372313932075553, "learning_rate": 1.213299630743747e-05, "loss": 0.1523, "step": 1150 }, { "epoch": 0.928, "grad_norm": 0.08636256799085817, "learning_rate": 1.2003580010891214e-05, "loss": 0.1563, "step": 1160 }, { "epoch": 0.936, "grad_norm": 0.0717205138861145, "learning_rate": 1.187381314585725e-05, "loss": 0.1647, "step": 1170 }, { "epoch": 0.944, "grad_norm": 0.07005503393261057, "learning_rate": 1.1743718417779518e-05, "loss": 0.1663, "step": 1180 }, { "epoch": 0.952, "grad_norm": 0.07064996616100924, "learning_rate": 1.1613318589468512e-05, "loss": 0.1589, "step": 1190 }, { "epoch": 0.96, "grad_norm": 0.0662381053536627, "learning_rate": 1.148263647711842e-05, "loss": 0.161, "step": 1200 }, { "epoch": 0.968, "grad_norm": 0.07589095377012324, "learning_rate": 1.135169494631497e-05, "loss": 0.1676, "step": 1210 }, { "epoch": 0.976, "grad_norm": 0.07324628362518953, "learning_rate": 1.1220516908034602e-05, "loss": 0.1621, "step": 1220 }, { "epoch": 0.984, "grad_norm": 0.07087940558217348, "learning_rate": 1.1089125314635727e-05, "loss": 0.163, "step": 1230 }, { "epoch": 0.992, "grad_norm": 0.08079771757264623, "learning_rate": 1.0957543155842703e-05, "loss": 0.1594, "step": 1240 }, { "epoch": 1.0, "grad_norm": 0.06413004885280083, "learning_rate": 1.0825793454723325e-05, "loss": 0.1554, "step": 1250 }, { "epoch": 1.008, "grad_norm": 0.05476172368624532, "learning_rate": 1.0693899263660442e-05, "loss": 0.1102, "step": 1260 }, { "epoch": 1.016, "grad_norm": 0.05632412500843437, "learning_rate": 1.0561883660318456e-05, "loss": 0.1028, "step": 1270 }, { "epoch": 1.024, "grad_norm": 0.05646598003617287, "learning_rate": 1.0429769743605406e-05, "loss": 0.1108, "step": 1280 }, { "epoch": 1.032, "grad_norm": 0.06842048425083827, "learning_rate": 1.0297580629631324e-05, "loss": 0.1046, "step": 1290 }, { "epoch": 1.04, "grad_norm": 0.0597248090009715, "learning_rate": 1.0165339447663586e-05, "loss": 0.1035, "step": 1300 }, { "epoch": 1.048, "grad_norm": 0.07398432585579974, "learning_rate": 1.0033069336079952e-05, "loss": 0.0987, "step": 1310 }, { "epoch": 1.056, "grad_norm": 0.06505084369381099, "learning_rate": 9.900793438320037e-06, "loss": 0.0981, "step": 1320 }, { "epoch": 1.064, "grad_norm": 0.06284724491257883, "learning_rate": 9.768534898835864e-06, "loss": 0.1079, "step": 1330 }, { "epoch": 1.072, "grad_norm": 0.06844723724817173, "learning_rate": 9.636316859042258e-06, "loss": 0.1062, "step": 1340 }, { "epoch": 1.08, "grad_norm": 0.06837187272099314, "learning_rate": 9.504162453267776e-06, "loss": 0.1056, "step": 1350 }, { "epoch": 1.088, "grad_norm": 0.07008372679481019, "learning_rate": 9.372094804706867e-06, "loss": 0.1024, "step": 1360 }, { "epoch": 1.096, "grad_norm": 0.05573557646176466, "learning_rate": 9.24013702137397e-06, "loss": 0.1004, "step": 1370 }, { "epoch": 1.104, "grad_norm": 0.061630742090732805, "learning_rate": 9.108312192060298e-06, "loss": 0.0958, "step": 1380 }, { "epoch": 1.112, "grad_norm": 0.0626839943409606, "learning_rate": 8.97664338229395e-06, "loss": 0.1025, "step": 1390 }, { "epoch": 1.12, "grad_norm": 0.061119158952486805, "learning_rate": 8.84515363030414e-06, "loss": 0.0995, "step": 1400 }, { "epoch": 1.1280000000000001, "grad_norm": 0.05757909222944341, "learning_rate": 8.713865942990143e-06, "loss": 0.1027, "step": 1410 }, { "epoch": 1.1360000000000001, "grad_norm": 0.06947100250068435, "learning_rate": 8.582803291895758e-06, "loss": 0.1111, "step": 1420 }, { "epoch": 1.144, "grad_norm": 0.0562229609672269, "learning_rate": 8.451988609189987e-06, "loss": 0.108, "step": 1430 }, { "epoch": 1.152, "grad_norm": 0.06930898413364359, "learning_rate": 8.321444783654524e-06, "loss": 0.1078, "step": 1440 }, { "epoch": 1.16, "grad_norm": 0.06571389182101994, "learning_rate": 8.191194656678905e-06, "loss": 0.0988, "step": 1450 }, { "epoch": 1.168, "grad_norm": 0.06275542671231407, "learning_rate": 8.06126101826392e-06, "loss": 0.0961, "step": 1460 }, { "epoch": 1.176, "grad_norm": 0.06782279303820063, "learning_rate": 7.931666603034034e-06, "loss": 0.1024, "step": 1470 }, { "epoch": 1.184, "grad_norm": 0.06578749140882915, "learning_rate": 7.80243408625947e-06, "loss": 0.0961, "step": 1480 }, { "epoch": 1.192, "grad_norm": 0.061803419288976154, "learning_rate": 7.673586079888699e-06, "loss": 0.1062, "step": 1490 }, { "epoch": 1.2, "grad_norm": 0.0717440728997515, "learning_rate": 7.545145128592009e-06, "loss": 0.0989, "step": 1500 }, { "epoch": 1.208, "grad_norm": 0.06490247759799975, "learning_rate": 7.4171337058168365e-06, "loss": 0.1045, "step": 1510 }, { "epoch": 1.216, "grad_norm": 0.06226994133294724, "learning_rate": 7.28957420985556e-06, "loss": 0.0997, "step": 1520 }, { "epoch": 1.224, "grad_norm": 0.07083340324266124, "learning_rate": 7.16248895992645e-06, "loss": 0.1002, "step": 1530 }, { "epoch": 1.232, "grad_norm": 0.07767517880074136, "learning_rate": 7.035900192268464e-06, "loss": 0.1132, "step": 1540 }, { "epoch": 1.24, "grad_norm": 0.06462545486578795, "learning_rate": 6.909830056250527e-06, "loss": 0.1028, "step": 1550 }, { "epoch": 1.248, "grad_norm": 0.0754203548519236, "learning_rate": 6.784300610496049e-06, "loss": 0.1051, "step": 1560 }, { "epoch": 1.256, "grad_norm": 0.06592132565995444, "learning_rate": 6.659333819023291e-06, "loss": 0.1047, "step": 1570 }, { "epoch": 1.264, "grad_norm": 0.06961461995248241, "learning_rate": 6.534951547402322e-06, "loss": 0.1049, "step": 1580 }, { "epoch": 1.272, "grad_norm": 0.06511899962594735, "learning_rate": 6.411175558929152e-06, "loss": 0.106, "step": 1590 }, { "epoch": 1.28, "grad_norm": 0.05916934075813879, "learning_rate": 6.2880275108177915e-06, "loss": 0.1046, "step": 1600 }, { "epoch": 1.288, "grad_norm": 0.07134708360913396, "learning_rate": 6.165528950410884e-06, "loss": 0.1076, "step": 1610 }, { "epoch": 1.296, "grad_norm": 0.06849648011516535, "learning_rate": 6.04370131140952e-06, "loss": 0.1007, "step": 1620 }, { "epoch": 1.304, "grad_norm": 0.0672608182320331, "learning_rate": 5.922565910122967e-06, "loss": 0.0973, "step": 1630 }, { "epoch": 1.312, "grad_norm": 0.07780002144539479, "learning_rate": 5.802143941738945e-06, "loss": 0.1077, "step": 1640 }, { "epoch": 1.32, "grad_norm": 0.06521931109585341, "learning_rate": 5.6824564766150724e-06, "loss": 0.108, "step": 1650 }, { "epoch": 1.328, "grad_norm": 0.0837775649233135, "learning_rate": 5.563524456592163e-06, "loss": 0.1002, "step": 1660 }, { "epoch": 1.336, "grad_norm": 0.06954289449761725, "learning_rate": 5.445368691330008e-06, "loss": 0.1049, "step": 1670 }, { "epoch": 1.3439999999999999, "grad_norm": 0.06577591748805732, "learning_rate": 5.328009854666303e-06, "loss": 0.0984, "step": 1680 }, { "epoch": 1.3519999999999999, "grad_norm": 0.06001848210345022, "learning_rate": 5.211468480999304e-06, "loss": 0.1049, "step": 1690 }, { "epoch": 1.3599999999999999, "grad_norm": 0.05950122818785273, "learning_rate": 5.095764961694923e-06, "loss": 0.104, "step": 1700 }, { "epoch": 1.3679999999999999, "grad_norm": 0.056395399165830855, "learning_rate": 4.980919541518796e-06, "loss": 0.1005, "step": 1710 }, { "epoch": 1.376, "grad_norm": 0.06573253422421312, "learning_rate": 4.866952315094088e-06, "loss": 0.1009, "step": 1720 }, { "epoch": 1.384, "grad_norm": 0.07121035158602729, "learning_rate": 4.753883223385467e-06, "loss": 0.0983, "step": 1730 }, { "epoch": 1.392, "grad_norm": 0.06524819896068156, "learning_rate": 4.641732050210032e-06, "loss": 0.1051, "step": 1740 }, { "epoch": 1.4, "grad_norm": 0.06825918255074036, "learning_rate": 4.530518418775734e-06, "loss": 0.094, "step": 1750 }, { "epoch": 1.408, "grad_norm": 0.07443611934581895, "learning_rate": 4.420261788247841e-06, "loss": 0.1025, "step": 1760 }, { "epoch": 1.416, "grad_norm": 0.06453930101827404, "learning_rate": 4.3109814503441894e-06, "loss": 0.1023, "step": 1770 }, { "epoch": 1.424, "grad_norm": 0.06508573006276362, "learning_rate": 4.202696525959667e-06, "loss": 0.1042, "step": 1780 }, { "epoch": 1.432, "grad_norm": 0.059065263422893345, "learning_rate": 4.0954259618206295e-06, "loss": 0.109, "step": 1790 }, { "epoch": 1.44, "grad_norm": 0.06698134546589794, "learning_rate": 3.989188527169749e-06, "loss": 0.1031, "step": 1800 }, { "epoch": 1.448, "grad_norm": 0.07037002941100785, "learning_rate": 3.884002810481959e-06, "loss": 0.1106, "step": 1810 }, { "epoch": 1.456, "grad_norm": 0.05841985205150354, "learning_rate": 3.7798872162119948e-06, "loss": 0.0987, "step": 1820 }, { "epoch": 1.464, "grad_norm": 0.0593910919321036, "learning_rate": 3.676859961574162e-06, "loss": 0.0994, "step": 1830 }, { "epoch": 1.472, "grad_norm": 0.06578904279209549, "learning_rate": 3.5749390733548382e-06, "loss": 0.0998, "step": 1840 }, { "epoch": 1.48, "grad_norm": 0.06370094131872989, "learning_rate": 3.4741423847583134e-06, "loss": 0.1008, "step": 1850 }, { "epoch": 1.488, "grad_norm": 0.05739668138835314, "learning_rate": 3.3744875322865035e-06, "loss": 0.1016, "step": 1860 }, { "epoch": 1.496, "grad_norm": 0.062215363243071836, "learning_rate": 3.2759919526530536e-06, "loss": 0.1032, "step": 1870 }, { "epoch": 1.504, "grad_norm": 0.0715569328821687, "learning_rate": 3.178672879732435e-06, "loss": 0.0958, "step": 1880 }, { "epoch": 1.512, "grad_norm": 0.05841095850206598, "learning_rate": 3.0825473415445073e-06, "loss": 0.1002, "step": 1890 }, { "epoch": 1.52, "grad_norm": 0.0617717625489719, "learning_rate": 2.9876321572751143e-06, "loss": 0.1034, "step": 1900 }, { "epoch": 1.528, "grad_norm": 0.06062020772334433, "learning_rate": 2.8939439343332086e-06, "loss": 0.0973, "step": 1910 }, { "epoch": 1.536, "grad_norm": 0.059698562886373666, "learning_rate": 2.8014990654450325e-06, "loss": 0.1006, "step": 1920 }, { "epoch": 1.544, "grad_norm": 0.06376657349003535, "learning_rate": 2.7103137257858867e-06, "loss": 0.0935, "step": 1930 }, { "epoch": 1.552, "grad_norm": 0.07743391292616765, "learning_rate": 2.6204038701499056e-06, "loss": 0.1026, "step": 1940 }, { "epoch": 1.56, "grad_norm": 0.06924240453610515, "learning_rate": 2.5317852301584642e-06, "loss": 0.1053, "step": 1950 }, { "epoch": 1.568, "grad_norm": 0.05897566595897626, "learning_rate": 2.4444733115075823e-06, "loss": 0.1007, "step": 1960 }, { "epoch": 1.576, "grad_norm": 0.061747152948674104, "learning_rate": 2.3584833912548887e-06, "loss": 0.1023, "step": 1970 }, { "epoch": 1.584, "grad_norm": 0.06279470086154411, "learning_rate": 2.2738305151465646e-06, "loss": 0.0967, "step": 1980 }, { "epoch": 1.592, "grad_norm": 0.06372313419732084, "learning_rate": 2.190529494984782e-06, "loss": 0.0937, "step": 1990 }, { "epoch": 1.6, "grad_norm": 0.061625905799104846, "learning_rate": 2.1085949060360654e-06, "loss": 0.0911, "step": 2000 }, { "epoch": 1.608, "grad_norm": 0.0678362739569136, "learning_rate": 2.0280410844810426e-06, "loss": 0.1065, "step": 2010 }, { "epoch": 1.616, "grad_norm": 0.06643434332226122, "learning_rate": 1.9488821249060297e-06, "loss": 0.103, "step": 2020 }, { "epoch": 1.624, "grad_norm": 0.06165743807147769, "learning_rate": 1.8711318778368792e-06, "loss": 0.1019, "step": 2030 }, { "epoch": 1.6320000000000001, "grad_norm": 0.07358372056477895, "learning_rate": 1.7948039473155553e-06, "loss": 0.1041, "step": 2040 }, { "epoch": 1.6400000000000001, "grad_norm": 0.06806526736977143, "learning_rate": 1.7199116885197996e-06, "loss": 0.1032, "step": 2050 }, { "epoch": 1.6480000000000001, "grad_norm": 0.06919391837094059, "learning_rate": 1.646468205426377e-06, "loss": 0.098, "step": 2060 }, { "epoch": 1.6560000000000001, "grad_norm": 0.07710033571374911, "learning_rate": 1.5744863485182537e-06, "loss": 0.1039, "step": 2070 }, { "epoch": 1.6640000000000001, "grad_norm": 0.07419656560471355, "learning_rate": 1.5039787125361327e-06, "loss": 0.0992, "step": 2080 }, { "epoch": 1.6720000000000002, "grad_norm": 0.06441503783003091, "learning_rate": 1.4349576342747462e-06, "loss": 0.1009, "step": 2090 }, { "epoch": 1.6800000000000002, "grad_norm": 0.0587805726462111, "learning_rate": 1.367435190424261e-06, "loss": 0.1002, "step": 2100 }, { "epoch": 1.688, "grad_norm": 0.06818045135041786, "learning_rate": 1.3014231954572287e-06, "loss": 0.1032, "step": 2110 }, { "epoch": 1.696, "grad_norm": 0.06850284991866502, "learning_rate": 1.2369331995613664e-06, "loss": 0.1013, "step": 2120 }, { "epoch": 1.704, "grad_norm": 0.07505851534344943, "learning_rate": 1.1739764866186309e-06, "loss": 0.095, "step": 2130 }, { "epoch": 1.712, "grad_norm": 0.07008959733086773, "learning_rate": 1.112564072230863e-06, "loss": 0.102, "step": 2140 }, { "epoch": 1.72, "grad_norm": 0.05689031064270844, "learning_rate": 1.0527067017923654e-06, "loss": 0.1026, "step": 2150 }, { "epoch": 1.728, "grad_norm": 0.07596542582163048, "learning_rate": 9.944148486097793e-07, "loss": 0.1034, "step": 2160 }, { "epoch": 1.736, "grad_norm": 0.06093771036580793, "learning_rate": 9.376987120695547e-07, "loss": 0.0997, "step": 2170 }, { "epoch": 1.744, "grad_norm": 0.07428382141394303, "learning_rate": 8.825682158533555e-07, "loss": 0.1082, "step": 2180 }, { "epoch": 1.752, "grad_norm": 0.08399394161052269, "learning_rate": 8.290330062017015e-07, "loss": 0.1052, "step": 2190 }, { "epoch": 1.76, "grad_norm": 0.0668059758680316, "learning_rate": 7.771024502261526e-07, "loss": 0.099, "step": 2200 }, { "epoch": 1.768, "grad_norm": 0.0679292033158011, "learning_rate": 7.267856342703461e-07, "loss": 0.1066, "step": 2210 }, { "epoch": 1.776, "grad_norm": 0.058093075805745104, "learning_rate": 6.780913623201346e-07, "loss": 0.0989, "step": 2220 }, { "epoch": 1.784, "grad_norm": 0.06223534910150451, "learning_rate": 6.310281544631547e-07, "loss": 0.0975, "step": 2230 }, { "epoch": 1.792, "grad_norm": 0.0661429754113188, "learning_rate": 5.856042453980526e-07, "loss": 0.0987, "step": 2240 }, { "epoch": 1.8, "grad_norm": 0.061050405248751057, "learning_rate": 5.418275829936537e-07, "loss": 0.0998, "step": 2250 }, { "epoch": 1.808, "grad_norm": 0.056130252404826796, "learning_rate": 4.997058268983135e-07, "loss": 0.0932, "step": 2260 }, { "epoch": 1.8159999999999998, "grad_norm": 0.06066847959463214, "learning_rate": 4.592463471997022e-07, "loss": 0.0974, "step": 2270 }, { "epoch": 1.8239999999999998, "grad_norm": 0.0643399844655628, "learning_rate": 4.204562231352516e-07, "loss": 0.1057, "step": 2280 }, { "epoch": 1.8319999999999999, "grad_norm": 0.060150289880508086, "learning_rate": 3.83342241853496e-07, "loss": 0.0994, "step": 2290 }, { "epoch": 1.8399999999999999, "grad_norm": 0.06796944373259835, "learning_rate": 3.4791089722651437e-07, "loss": 0.1072, "step": 2300 }, { "epoch": 1.8479999999999999, "grad_norm": 0.06759971647919732, "learning_rate": 3.1416838871368925e-07, "loss": 0.1063, "step": 2310 }, { "epoch": 1.8559999999999999, "grad_norm": 0.067835192233575, "learning_rate": 2.8212062027698995e-07, "loss": 0.1006, "step": 2320 }, { "epoch": 1.8639999999999999, "grad_norm": 0.07157931341704744, "learning_rate": 2.5177319934793995e-07, "loss": 0.1042, "step": 2330 }, { "epoch": 1.8719999999999999, "grad_norm": 0.06616867154511055, "learning_rate": 2.2313143584648423e-07, "loss": 0.0978, "step": 2340 }, { "epoch": 1.88, "grad_norm": 0.10047393373201215, "learning_rate": 1.9620034125190645e-07, "loss": 0.0977, "step": 2350 }, { "epoch": 1.888, "grad_norm": 0.06073465869846798, "learning_rate": 1.7098462772596302e-07, "loss": 0.107, "step": 2360 }, { "epoch": 1.896, "grad_norm": 0.0619748402288604, "learning_rate": 1.474887072883935e-07, "loss": 0.098, "step": 2370 }, { "epoch": 1.904, "grad_norm": 0.06634025194054122, "learning_rate": 1.2571669104494254e-07, "loss": 0.0934, "step": 2380 }, { "epoch": 1.912, "grad_norm": 0.07694308017805654, "learning_rate": 1.0567238846803995e-07, "loss": 0.1058, "step": 2390 }, { "epoch": 1.92, "grad_norm": 0.060586005493788085, "learning_rate": 8.735930673024806e-08, "loss": 0.0929, "step": 2400 }, { "epoch": 1.928, "grad_norm": 0.07076099012369173, "learning_rate": 7.078065009060941e-08, "loss": 0.1053, "step": 2410 }, { "epoch": 1.936, "grad_norm": 0.07088076731988013, "learning_rate": 5.5939319333998546e-08, "loss": 0.0994, "step": 2420 }, { "epoch": 1.944, "grad_norm": 0.06859911115828961, "learning_rate": 4.2837911263562406e-08, "loss": 0.1046, "step": 2430 }, { "epoch": 1.952, "grad_norm": 0.09088957264396723, "learning_rate": 3.147871824635717e-08, "loss": 0.1021, "step": 2440 }, { "epoch": 1.96, "grad_norm": 0.06296850339427222, "learning_rate": 2.1863727812254653e-08, "loss": 0.0992, "step": 2450 }, { "epoch": 1.968, "grad_norm": 0.06311334916073906, "learning_rate": 1.3994622306173766e-08, "loss": 0.1047, "step": 2460 }, { "epoch": 1.976, "grad_norm": 0.06846458491142236, "learning_rate": 7.872778593728258e-09, "loss": 0.102, "step": 2470 }, { "epoch": 1.984, "grad_norm": 0.06333619567681512, "learning_rate": 3.499267820307184e-09, "loss": 0.0979, "step": 2480 }, { "epoch": 1.992, "grad_norm": 0.06011655948198505, "learning_rate": 8.748552236603758e-10, "loss": 0.1016, "step": 2490 }, { "epoch": 2.0, "grad_norm": 0.06939915779075108, "learning_rate": 0.0, "loss": 0.096, "step": 2500 }, { "epoch": 2.0, "step": 2500, "total_flos": 22262632857600.0, "train_loss": 0.140902588391304, "train_runtime": 9768.7089, "train_samples_per_second": 8.189, "train_steps_per_second": 0.256 } ], "logging_steps": 10, "max_steps": 2500, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 10000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 22262632857600.0, "train_batch_size": 2, "trial_name": null, "trial_params": null }