{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 2386, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0020955574182732607, "grad_norm": 0.5915184020996094, "learning_rate": 0.0001996647108130763, "loss": 1.2488, "step": 5 }, { "epoch": 0.004191114836546521, "grad_norm": 0.6416106224060059, "learning_rate": 0.00019924559932942164, "loss": 0.8074, "step": 10 }, { "epoch": 0.006286672254819782, "grad_norm": 0.47403833270072937, "learning_rate": 0.000198826487845767, "loss": 0.5298, "step": 15 }, { "epoch": 0.008382229673093043, "grad_norm": 0.49060583114624023, "learning_rate": 0.00019840737636211232, "loss": 0.4396, "step": 20 }, { "epoch": 0.010477787091366304, "grad_norm": 0.3994845449924469, "learning_rate": 0.00019798826487845767, "loss": 0.3997, "step": 25 }, { "epoch": 0.012573344509639563, "grad_norm": 5.743133068084717, "learning_rate": 0.00019756915339480302, "loss": 0.3571, "step": 30 }, { "epoch": 0.014668901927912825, "grad_norm": 0.4928306043148041, "learning_rate": 0.00019715004191114837, "loss": 0.3467, "step": 35 }, { "epoch": 0.016764459346186086, "grad_norm": 333.243896484375, "learning_rate": 0.00019673093042749373, "loss": 0.3173, "step": 40 }, { "epoch": 0.018860016764459347, "grad_norm": 0.5272846817970276, "learning_rate": 0.00019631181894383908, "loss": 0.2952, "step": 45 }, { "epoch": 0.020955574182732608, "grad_norm": 3.5901286602020264, "learning_rate": 0.00019589270746018443, "loss": 0.2935, "step": 50 }, { "epoch": 0.02305113160100587, "grad_norm": 0.5013518929481506, "learning_rate": 0.00019547359597652975, "loss": 0.3024, "step": 55 }, { "epoch": 0.025146689019279127, "grad_norm": 16.086902618408203, "learning_rate": 0.0001950544844928751, "loss": 0.2683, "step": 60 }, { "epoch": 0.027242246437552388, "grad_norm": 0.30715975165367126, "learning_rate": 0.00019463537300922046, "loss": 0.2653, "step": 65 }, { "epoch": 0.02933780385582565, "grad_norm": 0.33719402551651, "learning_rate": 0.0001942162615255658, "loss": 0.2575, "step": 70 }, { "epoch": 0.03143336127409891, "grad_norm": 0.3756738305091858, "learning_rate": 0.00019379715004191116, "loss": 0.2449, "step": 75 }, { "epoch": 0.03352891869237217, "grad_norm": 0.3635186553001404, "learning_rate": 0.0001933780385582565, "loss": 0.2538, "step": 80 }, { "epoch": 0.03562447611064543, "grad_norm": 0.3560231328010559, "learning_rate": 0.00019295892707460186, "loss": 0.2435, "step": 85 }, { "epoch": 0.037720033528918694, "grad_norm": 0.371216356754303, "learning_rate": 0.00019253981559094722, "loss": 0.2329, "step": 90 }, { "epoch": 0.039815590947191955, "grad_norm": 0.45535510778427124, "learning_rate": 0.00019212070410729254, "loss": 0.2338, "step": 95 }, { "epoch": 0.041911148365465216, "grad_norm": 15.063165664672852, "learning_rate": 0.0001917015926236379, "loss": 0.2353, "step": 100 }, { "epoch": 0.04400670578373848, "grad_norm": 0.308463454246521, "learning_rate": 0.00019128248113998324, "loss": 0.2323, "step": 105 }, { "epoch": 0.04610226320201174, "grad_norm": 0.29910165071487427, "learning_rate": 0.0001908633696563286, "loss": 0.2222, "step": 110 }, { "epoch": 0.04819782062028499, "grad_norm": 0.3254060745239258, "learning_rate": 0.00019044425817267395, "loss": 0.216, "step": 115 }, { "epoch": 0.050293378038558254, "grad_norm": 0.2623255252838135, "learning_rate": 0.0001900251466890193, "loss": 0.2136, "step": 120 }, { "epoch": 0.052388935456831515, "grad_norm": 0.2627584636211395, "learning_rate": 0.00018960603520536465, "loss": 0.2118, "step": 125 }, { "epoch": 0.054484492875104776, "grad_norm": 0.28595617413520813, "learning_rate": 0.00018918692372170998, "loss": 0.2088, "step": 130 }, { "epoch": 0.05658005029337804, "grad_norm": 0.550282895565033, "learning_rate": 0.00018876781223805533, "loss": 0.2198, "step": 135 }, { "epoch": 0.0586756077116513, "grad_norm": 0.296385258436203, "learning_rate": 0.00018834870075440068, "loss": 0.2219, "step": 140 }, { "epoch": 0.06077116512992456, "grad_norm": 0.3291824460029602, "learning_rate": 0.00018792958927074603, "loss": 0.2059, "step": 145 }, { "epoch": 0.06286672254819782, "grad_norm": 0.27933308482170105, "learning_rate": 0.00018751047778709136, "loss": 0.2001, "step": 150 }, { "epoch": 0.06496227996647108, "grad_norm": 0.2848331332206726, "learning_rate": 0.00018709136630343674, "loss": 0.1999, "step": 155 }, { "epoch": 0.06705783738474434, "grad_norm": 0.25534388422966003, "learning_rate": 0.0001866722548197821, "loss": 0.1977, "step": 160 }, { "epoch": 0.0691533948030176, "grad_norm": 2.9550936222076416, "learning_rate": 0.0001862531433361274, "loss": 0.2004, "step": 165 }, { "epoch": 0.07124895222129086, "grad_norm": 0.35537609457969666, "learning_rate": 0.00018583403185247276, "loss": 0.2131, "step": 170 }, { "epoch": 0.07334450963956413, "grad_norm": 47.72193908691406, "learning_rate": 0.00018541492036881812, "loss": 0.2016, "step": 175 }, { "epoch": 0.07544006705783739, "grad_norm": 0.3546775281429291, "learning_rate": 0.00018499580888516347, "loss": 0.1998, "step": 180 }, { "epoch": 0.07753562447611065, "grad_norm": 0.2701320946216583, "learning_rate": 0.0001845766974015088, "loss": 0.1948, "step": 185 }, { "epoch": 0.07963118189438391, "grad_norm": 0.2246120572090149, "learning_rate": 0.00018415758591785414, "loss": 0.1882, "step": 190 }, { "epoch": 0.08172673931265717, "grad_norm": 0.27891814708709717, "learning_rate": 0.00018373847443419952, "loss": 0.1879, "step": 195 }, { "epoch": 0.08382229673093043, "grad_norm": 0.4772103428840637, "learning_rate": 0.00018331936295054485, "loss": 0.2069, "step": 200 }, { "epoch": 0.08591785414920369, "grad_norm": 2.367286205291748, "learning_rate": 0.0001829002514668902, "loss": 0.2043, "step": 205 }, { "epoch": 0.08801341156747695, "grad_norm": 0.30191686749458313, "learning_rate": 0.00018248113998323555, "loss": 0.1978, "step": 210 }, { "epoch": 0.09010896898575022, "grad_norm": 0.22913698852062225, "learning_rate": 0.0001820620284995809, "loss": 0.1937, "step": 215 }, { "epoch": 0.09220452640402348, "grad_norm": 0.26689326763153076, "learning_rate": 0.00018164291701592623, "loss": 0.1892, "step": 220 }, { "epoch": 0.09430008382229674, "grad_norm": 0.26441484689712524, "learning_rate": 0.00018122380553227158, "loss": 0.1924, "step": 225 }, { "epoch": 0.09639564124056998, "grad_norm": 0.2602805197238922, "learning_rate": 0.00018080469404861696, "loss": 0.1856, "step": 230 }, { "epoch": 0.09849119865884325, "grad_norm": 0.21042250096797943, "learning_rate": 0.0001803855825649623, "loss": 0.1848, "step": 235 }, { "epoch": 0.10058675607711651, "grad_norm": 1.8752623796463013, "learning_rate": 0.00017996647108130763, "loss": 0.1822, "step": 240 }, { "epoch": 0.10268231349538977, "grad_norm": 0.3879601061344147, "learning_rate": 0.00017954735959765299, "loss": 0.1979, "step": 245 }, { "epoch": 0.10477787091366303, "grad_norm": 0.27347472310066223, "learning_rate": 0.00017912824811399834, "loss": 0.1903, "step": 250 }, { "epoch": 0.10687342833193629, "grad_norm": 0.24279960989952087, "learning_rate": 0.0001787091366303437, "loss": 0.1848, "step": 255 }, { "epoch": 0.10896898575020955, "grad_norm": 0.275285005569458, "learning_rate": 0.00017829002514668901, "loss": 0.1848, "step": 260 }, { "epoch": 0.11106454316848281, "grad_norm": 0.33457309007644653, "learning_rate": 0.00017787091366303437, "loss": 0.1851, "step": 265 }, { "epoch": 0.11316010058675607, "grad_norm": 0.5345426797866821, "learning_rate": 0.00017745180217937974, "loss": 0.189, "step": 270 }, { "epoch": 0.11525565800502934, "grad_norm": 0.3533399701118469, "learning_rate": 0.00017703269069572507, "loss": 0.1835, "step": 275 }, { "epoch": 0.1173512154233026, "grad_norm": 0.3257920742034912, "learning_rate": 0.00017661357921207042, "loss": 0.1892, "step": 280 }, { "epoch": 0.11944677284157586, "grad_norm": 0.21177901327610016, "learning_rate": 0.00017619446772841577, "loss": 0.1818, "step": 285 }, { "epoch": 0.12154233025984912, "grad_norm": 0.20798452198505402, "learning_rate": 0.00017577535624476112, "loss": 0.1792, "step": 290 }, { "epoch": 0.12363788767812238, "grad_norm": 0.21963848173618317, "learning_rate": 0.00017535624476110645, "loss": 0.1776, "step": 295 }, { "epoch": 0.12573344509639564, "grad_norm": 0.22277575731277466, "learning_rate": 0.0001749371332774518, "loss": 0.178, "step": 300 }, { "epoch": 0.12782900251466892, "grad_norm": 0.22803856432437897, "learning_rate": 0.00017451802179379715, "loss": 0.1732, "step": 305 }, { "epoch": 0.12992455993294216, "grad_norm": 0.20474207401275635, "learning_rate": 0.0001740989103101425, "loss": 0.173, "step": 310 }, { "epoch": 0.1320201173512154, "grad_norm": 0.23283089697360992, "learning_rate": 0.00017367979882648786, "loss": 0.1752, "step": 315 }, { "epoch": 0.13411567476948869, "grad_norm": 0.24834850430488586, "learning_rate": 0.0001732606873428332, "loss": 0.1727, "step": 320 }, { "epoch": 0.13621123218776193, "grad_norm": 0.2943226993083954, "learning_rate": 0.00017284157585917856, "loss": 0.1758, "step": 325 }, { "epoch": 0.1383067896060352, "grad_norm": 0.28577786684036255, "learning_rate": 0.00017242246437552388, "loss": 0.1835, "step": 330 }, { "epoch": 0.14040234702430845, "grad_norm": 0.22449374198913574, "learning_rate": 0.00017200335289186924, "loss": 0.1795, "step": 335 }, { "epoch": 0.14249790444258173, "grad_norm": 0.3676876425743103, "learning_rate": 0.0001715842414082146, "loss": 0.1771, "step": 340 }, { "epoch": 0.14459346186085498, "grad_norm": 0.4172014594078064, "learning_rate": 0.00017116512992455994, "loss": 0.197, "step": 345 }, { "epoch": 0.14668901927912825, "grad_norm": 0.26507022976875305, "learning_rate": 0.0001707460184409053, "loss": 0.181, "step": 350 }, { "epoch": 0.1487845766974015, "grad_norm": 0.2308947890996933, "learning_rate": 0.00017032690695725064, "loss": 0.1767, "step": 355 }, { "epoch": 0.15088013411567477, "grad_norm": 0.30771297216415405, "learning_rate": 0.000169907795473596, "loss": 0.1832, "step": 360 }, { "epoch": 0.15297569153394802, "grad_norm": 0.23101018369197845, "learning_rate": 0.00016948868398994132, "loss": 0.1811, "step": 365 }, { "epoch": 0.1550712489522213, "grad_norm": 0.20816421508789062, "learning_rate": 0.00016906957250628667, "loss": 0.1755, "step": 370 }, { "epoch": 0.15716680637049454, "grad_norm": 0.2107459455728531, "learning_rate": 0.00016865046102263202, "loss": 0.1726, "step": 375 }, { "epoch": 0.15926236378876782, "grad_norm": 2.224956750869751, "learning_rate": 0.00016823134953897737, "loss": 0.1718, "step": 380 }, { "epoch": 0.16135792120704107, "grad_norm": 0.22022977471351624, "learning_rate": 0.00016781223805532273, "loss": 0.1725, "step": 385 }, { "epoch": 0.16345347862531434, "grad_norm": 0.1845213919878006, "learning_rate": 0.00016739312657166808, "loss": 0.1677, "step": 390 }, { "epoch": 0.1655490360435876, "grad_norm": 0.21298770606517792, "learning_rate": 0.00016697401508801343, "loss": 0.1734, "step": 395 }, { "epoch": 0.16764459346186086, "grad_norm": 0.20876356959342957, "learning_rate": 0.00016655490360435878, "loss": 0.1761, "step": 400 }, { "epoch": 0.1697401508801341, "grad_norm": 0.21667592227458954, "learning_rate": 0.0001661357921207041, "loss": 0.1668, "step": 405 }, { "epoch": 0.17183570829840739, "grad_norm": 2.4301416873931885, "learning_rate": 0.00016571668063704946, "loss": 0.1697, "step": 410 }, { "epoch": 0.17393126571668063, "grad_norm": 0.43861570954322815, "learning_rate": 0.0001652975691533948, "loss": 0.1804, "step": 415 }, { "epoch": 0.1760268231349539, "grad_norm": 0.2521713078022003, "learning_rate": 0.00016487845766974016, "loss": 0.1762, "step": 420 }, { "epoch": 0.17812238055322716, "grad_norm": 0.20923659205436707, "learning_rate": 0.00016445934618608551, "loss": 0.1695, "step": 425 }, { "epoch": 0.18021793797150043, "grad_norm": 0.20591603219509125, "learning_rate": 0.00016404023470243087, "loss": 0.1729, "step": 430 }, { "epoch": 0.18231349538977368, "grad_norm": 2.5188469886779785, "learning_rate": 0.00016362112321877622, "loss": 0.1723, "step": 435 }, { "epoch": 0.18440905280804695, "grad_norm": 0.2523597478866577, "learning_rate": 0.00016320201173512154, "loss": 0.174, "step": 440 }, { "epoch": 0.1865046102263202, "grad_norm": 0.23274292051792145, "learning_rate": 0.0001627829002514669, "loss": 0.1707, "step": 445 }, { "epoch": 0.18860016764459347, "grad_norm": 0.26748543977737427, "learning_rate": 0.00016236378876781225, "loss": 0.1686, "step": 450 }, { "epoch": 0.19069572506286672, "grad_norm": 0.2876422107219696, "learning_rate": 0.0001619446772841576, "loss": 0.1723, "step": 455 }, { "epoch": 0.19279128248113997, "grad_norm": 0.5347093939781189, "learning_rate": 0.00016152556580050292, "loss": 0.1873, "step": 460 }, { "epoch": 0.19488683989941324, "grad_norm": 0.2463475912809372, "learning_rate": 0.0001611064543168483, "loss": 0.1817, "step": 465 }, { "epoch": 0.1969823973176865, "grad_norm": 0.21807396411895752, "learning_rate": 0.00016068734283319365, "loss": 0.1751, "step": 470 }, { "epoch": 0.19907795473595977, "grad_norm": 0.1950008124113083, "learning_rate": 0.00016026823134953898, "loss": 0.1723, "step": 475 }, { "epoch": 0.20117351215423301, "grad_norm": 0.19139733910560608, "learning_rate": 0.00015984911986588433, "loss": 0.168, "step": 480 }, { "epoch": 0.2032690695725063, "grad_norm": 0.24496974050998688, "learning_rate": 0.00015943000838222968, "loss": 0.1713, "step": 485 }, { "epoch": 0.20536462699077954, "grad_norm": 0.2883533239364624, "learning_rate": 0.00015901089689857503, "loss": 0.1758, "step": 490 }, { "epoch": 0.2074601844090528, "grad_norm": 0.2685905694961548, "learning_rate": 0.00015859178541492036, "loss": 0.1789, "step": 495 }, { "epoch": 0.20955574182732606, "grad_norm": 0.22861076891422272, "learning_rate": 0.0001581726739312657, "loss": 0.1751, "step": 500 }, { "epoch": 0.21165129924559933, "grad_norm": 0.4216708242893219, "learning_rate": 0.0001577535624476111, "loss": 0.1765, "step": 505 }, { "epoch": 0.21374685666387258, "grad_norm": 0.33128198981285095, "learning_rate": 0.0001573344509639564, "loss": 0.1908, "step": 510 }, { "epoch": 0.21584241408214586, "grad_norm": 0.21954227983951569, "learning_rate": 0.00015691533948030176, "loss": 0.1707, "step": 515 }, { "epoch": 0.2179379715004191, "grad_norm": 0.22218674421310425, "learning_rate": 0.00015649622799664712, "loss": 0.1695, "step": 520 }, { "epoch": 0.22003352891869238, "grad_norm": 0.23518136143684387, "learning_rate": 0.00015607711651299247, "loss": 0.1677, "step": 525 }, { "epoch": 0.22212908633696563, "grad_norm": 0.4071219861507416, "learning_rate": 0.0001556580050293378, "loss": 0.1753, "step": 530 }, { "epoch": 0.2242246437552389, "grad_norm": 0.25883597135543823, "learning_rate": 0.00015523889354568314, "loss": 0.1741, "step": 535 }, { "epoch": 0.22632020117351215, "grad_norm": 0.19170518219470978, "learning_rate": 0.00015481978206202852, "loss": 0.17, "step": 540 }, { "epoch": 0.22841575859178542, "grad_norm": 0.18559418618679047, "learning_rate": 0.00015440067057837387, "loss": 0.1701, "step": 545 }, { "epoch": 0.23051131601005867, "grad_norm": 0.2035888284444809, "learning_rate": 0.0001539815590947192, "loss": 0.1623, "step": 550 }, { "epoch": 0.23260687342833195, "grad_norm": 7.4185709953308105, "learning_rate": 0.00015356244761106455, "loss": 0.1654, "step": 555 }, { "epoch": 0.2347024308466052, "grad_norm": 0.21483619511127472, "learning_rate": 0.0001531433361274099, "loss": 0.1664, "step": 560 }, { "epoch": 0.23679798826487847, "grad_norm": 0.2908990979194641, "learning_rate": 0.00015272422464375525, "loss": 0.1733, "step": 565 }, { "epoch": 0.23889354568315171, "grad_norm": 0.34150460362434387, "learning_rate": 0.00015230511316010058, "loss": 0.1721, "step": 570 }, { "epoch": 0.240989103101425, "grad_norm": 0.26667365431785583, "learning_rate": 0.00015188600167644593, "loss": 0.1778, "step": 575 }, { "epoch": 0.24308466051969824, "grad_norm": 0.2024029940366745, "learning_rate": 0.0001514668901927913, "loss": 0.1681, "step": 580 }, { "epoch": 0.2451802179379715, "grad_norm": 0.1918814480304718, "learning_rate": 0.00015104777870913663, "loss": 0.1672, "step": 585 }, { "epoch": 0.24727577535624476, "grad_norm": 0.22173915803432465, "learning_rate": 0.00015062866722548199, "loss": 0.1629, "step": 590 }, { "epoch": 0.24937133277451803, "grad_norm": 0.20861753821372986, "learning_rate": 0.00015020955574182734, "loss": 0.1695, "step": 595 }, { "epoch": 0.2514668901927913, "grad_norm": 0.2476891726255417, "learning_rate": 0.0001497904442581727, "loss": 0.1687, "step": 600 }, { "epoch": 0.25356244761106456, "grad_norm": 0.23498353362083435, "learning_rate": 0.00014937133277451801, "loss": 0.1721, "step": 605 }, { "epoch": 0.25565800502933783, "grad_norm": 0.24014067649841309, "learning_rate": 0.00014895222129086337, "loss": 0.1643, "step": 610 }, { "epoch": 0.25775356244761105, "grad_norm": 1.1056888103485107, "learning_rate": 0.00014853310980720872, "loss": 0.1652, "step": 615 }, { "epoch": 0.2598491198658843, "grad_norm": 0.3467954397201538, "learning_rate": 0.00014811399832355407, "loss": 0.1741, "step": 620 }, { "epoch": 0.2619446772841576, "grad_norm": 0.29533931612968445, "learning_rate": 0.00014769488683989942, "loss": 0.1768, "step": 625 }, { "epoch": 0.2640402347024308, "grad_norm": 0.2000960111618042, "learning_rate": 0.00014727577535624477, "loss": 0.1662, "step": 630 }, { "epoch": 0.2661357921207041, "grad_norm": 0.18396534025669098, "learning_rate": 0.00014685666387259013, "loss": 0.166, "step": 635 }, { "epoch": 0.26823134953897737, "grad_norm": 0.18797029554843903, "learning_rate": 0.00014643755238893545, "loss": 0.1624, "step": 640 }, { "epoch": 0.27032690695725065, "grad_norm": 0.1914837509393692, "learning_rate": 0.0001460184409052808, "loss": 0.1625, "step": 645 }, { "epoch": 0.27242246437552387, "grad_norm": 0.18006779253482819, "learning_rate": 0.00014559932942162615, "loss": 0.1602, "step": 650 }, { "epoch": 0.27451802179379714, "grad_norm": 0.19413350522518158, "learning_rate": 0.0001451802179379715, "loss": 0.1632, "step": 655 }, { "epoch": 0.2766135792120704, "grad_norm": 0.19674436748027802, "learning_rate": 0.00014476110645431686, "loss": 0.1593, "step": 660 }, { "epoch": 0.2787091366303437, "grad_norm": 0.16716861724853516, "learning_rate": 0.0001443419949706622, "loss": 0.161, "step": 665 }, { "epoch": 0.2808046940486169, "grad_norm": 0.1591351330280304, "learning_rate": 0.00014392288348700756, "loss": 0.163, "step": 670 }, { "epoch": 0.2829002514668902, "grad_norm": 0.16973218321800232, "learning_rate": 0.0001435037720033529, "loss": 0.1665, "step": 675 }, { "epoch": 0.28499580888516346, "grad_norm": 0.16811800003051758, "learning_rate": 0.00014308466051969824, "loss": 0.1586, "step": 680 }, { "epoch": 0.28709136630343673, "grad_norm": 0.21386906504631042, "learning_rate": 0.0001426655490360436, "loss": 0.1589, "step": 685 }, { "epoch": 0.28918692372170995, "grad_norm": 0.17704640328884125, "learning_rate": 0.00014224643755238894, "loss": 0.1597, "step": 690 }, { "epoch": 0.29128248113998323, "grad_norm": 0.1727607399225235, "learning_rate": 0.0001418273260687343, "loss": 0.1565, "step": 695 }, { "epoch": 0.2933780385582565, "grad_norm": 0.18027518689632416, "learning_rate": 0.00014140821458507964, "loss": 0.1597, "step": 700 }, { "epoch": 0.2954735959765298, "grad_norm": 0.18142397701740265, "learning_rate": 0.000140989103101425, "loss": 0.1618, "step": 705 }, { "epoch": 0.297569153394803, "grad_norm": 0.18902671337127686, "learning_rate": 0.00014056999161777035, "loss": 0.1606, "step": 710 }, { "epoch": 0.2996647108130763, "grad_norm": 0.2578323185443878, "learning_rate": 0.00014015088013411567, "loss": 0.1658, "step": 715 }, { "epoch": 0.30176026823134955, "grad_norm": 0.2432331144809723, "learning_rate": 0.00013973176865046102, "loss": 0.1694, "step": 720 }, { "epoch": 0.3038558256496228, "grad_norm": 0.1846640259027481, "learning_rate": 0.00013931265716680638, "loss": 0.1655, "step": 725 }, { "epoch": 0.30595138306789604, "grad_norm": 0.1582545042037964, "learning_rate": 0.00013889354568315173, "loss": 0.1639, "step": 730 }, { "epoch": 0.3080469404861693, "grad_norm": 0.16749900579452515, "learning_rate": 0.00013847443419949708, "loss": 0.1596, "step": 735 }, { "epoch": 0.3101424979044426, "grad_norm": 0.17118974030017853, "learning_rate": 0.00013805532271584243, "loss": 0.1569, "step": 740 }, { "epoch": 0.31223805532271587, "grad_norm": 0.16258081793785095, "learning_rate": 0.00013763621123218778, "loss": 0.1599, "step": 745 }, { "epoch": 0.3143336127409891, "grad_norm": 0.17213059961795807, "learning_rate": 0.0001372170997485331, "loss": 0.1588, "step": 750 }, { "epoch": 0.31642917015926236, "grad_norm": 0.18207783997058868, "learning_rate": 0.00013679798826487846, "loss": 0.1571, "step": 755 }, { "epoch": 0.31852472757753564, "grad_norm": 0.19723886251449585, "learning_rate": 0.0001363788767812238, "loss": 0.1552, "step": 760 }, { "epoch": 0.32062028499580886, "grad_norm": 0.17248809337615967, "learning_rate": 0.00013595976529756916, "loss": 0.1564, "step": 765 }, { "epoch": 0.32271584241408213, "grad_norm": 0.18213661015033722, "learning_rate": 0.0001355406538139145, "loss": 0.1597, "step": 770 }, { "epoch": 0.3248113998323554, "grad_norm": 0.1856526881456375, "learning_rate": 0.00013512154233025987, "loss": 0.1609, "step": 775 }, { "epoch": 0.3269069572506287, "grad_norm": 0.16957145929336548, "learning_rate": 0.00013470243084660522, "loss": 0.1595, "step": 780 }, { "epoch": 0.3290025146689019, "grad_norm": 0.17088639736175537, "learning_rate": 0.00013428331936295054, "loss": 0.161, "step": 785 }, { "epoch": 0.3310980720871752, "grad_norm": 1.320410132408142, "learning_rate": 0.0001338642078792959, "loss": 0.165, "step": 790 }, { "epoch": 0.33319362950544845, "grad_norm": 0.19440437853336334, "learning_rate": 0.00013344509639564125, "loss": 0.1602, "step": 795 }, { "epoch": 0.3352891869237217, "grad_norm": 0.19590894877910614, "learning_rate": 0.0001330259849119866, "loss": 0.1616, "step": 800 }, { "epoch": 0.33738474434199495, "grad_norm": 0.19055521488189697, "learning_rate": 0.00013260687342833192, "loss": 0.1638, "step": 805 }, { "epoch": 0.3394803017602682, "grad_norm": 0.1577232927083969, "learning_rate": 0.00013218776194467727, "loss": 0.1582, "step": 810 }, { "epoch": 0.3415758591785415, "grad_norm": 0.16225744783878326, "learning_rate": 0.00013176865046102265, "loss": 0.1555, "step": 815 }, { "epoch": 0.34367141659681477, "grad_norm": 0.1565002202987671, "learning_rate": 0.000131349538977368, "loss": 0.1567, "step": 820 }, { "epoch": 0.345766974015088, "grad_norm": 0.18790805339813232, "learning_rate": 0.00013093042749371333, "loss": 0.1581, "step": 825 }, { "epoch": 0.34786253143336127, "grad_norm": 0.18632063269615173, "learning_rate": 0.00013051131601005868, "loss": 0.1621, "step": 830 }, { "epoch": 0.34995808885163454, "grad_norm": 0.17684835195541382, "learning_rate": 0.00013009220452640403, "loss": 0.1591, "step": 835 }, { "epoch": 0.3520536462699078, "grad_norm": 0.16213147342205048, "learning_rate": 0.00012967309304274938, "loss": 0.1593, "step": 840 }, { "epoch": 0.35414920368818104, "grad_norm": 0.1656450480222702, "learning_rate": 0.0001292539815590947, "loss": 0.1617, "step": 845 }, { "epoch": 0.3562447611064543, "grad_norm": 0.16887331008911133, "learning_rate": 0.0001288348700754401, "loss": 0.1601, "step": 850 }, { "epoch": 0.3583403185247276, "grad_norm": 0.17660242319107056, "learning_rate": 0.00012841575859178544, "loss": 0.1575, "step": 855 }, { "epoch": 0.36043587594300086, "grad_norm": 0.17625996470451355, "learning_rate": 0.00012799664710813076, "loss": 0.1597, "step": 860 }, { "epoch": 0.3625314333612741, "grad_norm": 0.15301348268985748, "learning_rate": 0.00012757753562447612, "loss": 0.1597, "step": 865 }, { "epoch": 0.36462699077954736, "grad_norm": 0.15253449976444244, "learning_rate": 0.00012715842414082147, "loss": 0.157, "step": 870 }, { "epoch": 0.36672254819782063, "grad_norm": 0.16985070705413818, "learning_rate": 0.00012673931265716682, "loss": 0.1558, "step": 875 }, { "epoch": 0.3688181056160939, "grad_norm": 0.1545540988445282, "learning_rate": 0.00012632020117351214, "loss": 0.1596, "step": 880 }, { "epoch": 0.3709136630343671, "grad_norm": 0.1766255795955658, "learning_rate": 0.0001259010896898575, "loss": 0.1537, "step": 885 }, { "epoch": 0.3730092204526404, "grad_norm": 0.15094798803329468, "learning_rate": 0.00012548197820620288, "loss": 0.1562, "step": 890 }, { "epoch": 0.3751047778709137, "grad_norm": 0.16060136258602142, "learning_rate": 0.0001250628667225482, "loss": 0.158, "step": 895 }, { "epoch": 0.37720033528918695, "grad_norm": 0.17947901785373688, "learning_rate": 0.00012464375523889355, "loss": 0.1583, "step": 900 }, { "epoch": 0.37929589270746017, "grad_norm": 0.15727129578590393, "learning_rate": 0.0001242246437552389, "loss": 0.1529, "step": 905 }, { "epoch": 0.38139145012573344, "grad_norm": 0.14456555247306824, "learning_rate": 0.00012380553227158426, "loss": 0.1573, "step": 910 }, { "epoch": 0.3834870075440067, "grad_norm": 0.15629969537258148, "learning_rate": 0.00012338642078792958, "loss": 0.1607, "step": 915 }, { "epoch": 0.38558256496227994, "grad_norm": 0.1958230435848236, "learning_rate": 0.00012296730930427493, "loss": 0.1572, "step": 920 }, { "epoch": 0.3876781223805532, "grad_norm": 0.18159349262714386, "learning_rate": 0.00012254819782062028, "loss": 0.1565, "step": 925 }, { "epoch": 0.3897736797988265, "grad_norm": 0.19391848146915436, "learning_rate": 0.00012212908633696564, "loss": 0.1552, "step": 930 }, { "epoch": 0.39186923721709976, "grad_norm": 0.18596895039081573, "learning_rate": 0.00012170997485331099, "loss": 0.1623, "step": 935 }, { "epoch": 0.393964794635373, "grad_norm": 0.175604447722435, "learning_rate": 0.00012129086336965634, "loss": 0.1593, "step": 940 }, { "epoch": 0.39606035205364626, "grad_norm": 0.18805819749832153, "learning_rate": 0.00012087175188600168, "loss": 0.1603, "step": 945 }, { "epoch": 0.39815590947191953, "grad_norm": 0.15223102271556854, "learning_rate": 0.00012045264040234703, "loss": 0.1568, "step": 950 }, { "epoch": 0.4002514668901928, "grad_norm": 0.13771827518939972, "learning_rate": 0.00012003352891869237, "loss": 0.1553, "step": 955 }, { "epoch": 0.40234702430846603, "grad_norm": 0.1633366495370865, "learning_rate": 0.00011961441743503772, "loss": 0.156, "step": 960 }, { "epoch": 0.4044425817267393, "grad_norm": 0.17143379151821136, "learning_rate": 0.00011919530595138306, "loss": 0.1542, "step": 965 }, { "epoch": 0.4065381391450126, "grad_norm": 0.16767437756061554, "learning_rate": 0.00011877619446772844, "loss": 0.1572, "step": 970 }, { "epoch": 0.40863369656328585, "grad_norm": 0.16299773752689362, "learning_rate": 0.00011835708298407377, "loss": 0.1583, "step": 975 }, { "epoch": 0.4107292539815591, "grad_norm": 0.15164397656917572, "learning_rate": 0.00011793797150041913, "loss": 0.1526, "step": 980 }, { "epoch": 0.41282481139983235, "grad_norm": 0.1567896157503128, "learning_rate": 0.00011751886001676446, "loss": 0.154, "step": 985 }, { "epoch": 0.4149203688181056, "grad_norm": 0.15943744778633118, "learning_rate": 0.00011709974853310982, "loss": 0.1533, "step": 990 }, { "epoch": 0.4170159262363789, "grad_norm": 0.16572755575180054, "learning_rate": 0.00011668063704945515, "loss": 0.1559, "step": 995 }, { "epoch": 0.4191114836546521, "grad_norm": 0.15907670557498932, "learning_rate": 0.0001162615255658005, "loss": 0.1558, "step": 1000 }, { "epoch": 0.4212070410729254, "grad_norm": 0.1808643937110901, "learning_rate": 0.00011584241408214587, "loss": 0.1555, "step": 1005 }, { "epoch": 0.42330259849119867, "grad_norm": 0.23819687962532043, "learning_rate": 0.00011542330259849121, "loss": 0.1579, "step": 1010 }, { "epoch": 0.42539815590947194, "grad_norm": 0.24894015491008759, "learning_rate": 0.00011500419111483656, "loss": 0.1637, "step": 1015 }, { "epoch": 0.42749371332774516, "grad_norm": 0.18128713965415955, "learning_rate": 0.0001145850796311819, "loss": 0.1605, "step": 1020 }, { "epoch": 0.42958927074601844, "grad_norm": 0.1779192090034485, "learning_rate": 0.00011416596814752725, "loss": 0.1586, "step": 1025 }, { "epoch": 0.4316848281642917, "grad_norm": 0.1617233157157898, "learning_rate": 0.00011374685666387259, "loss": 0.1548, "step": 1030 }, { "epoch": 0.433780385582565, "grad_norm": 0.146457701921463, "learning_rate": 0.00011332774518021794, "loss": 0.1581, "step": 1035 }, { "epoch": 0.4358759430008382, "grad_norm": 18.378135681152344, "learning_rate": 0.00011290863369656328, "loss": 0.1546, "step": 1040 }, { "epoch": 0.4379715004191115, "grad_norm": 0.16601739823818207, "learning_rate": 0.00011248952221290864, "loss": 0.1573, "step": 1045 }, { "epoch": 0.44006705783738476, "grad_norm": 0.18232333660125732, "learning_rate": 0.000112070410729254, "loss": 0.1546, "step": 1050 }, { "epoch": 0.44216261525565803, "grad_norm": 0.15297749638557434, "learning_rate": 0.00011165129924559933, "loss": 0.1561, "step": 1055 }, { "epoch": 0.44425817267393125, "grad_norm": 7.5409440994262695, "learning_rate": 0.00011123218776194469, "loss": 0.1614, "step": 1060 }, { "epoch": 0.4463537300922045, "grad_norm": 0.1807660162448883, "learning_rate": 0.00011081307627829002, "loss": 0.1624, "step": 1065 }, { "epoch": 0.4484492875104778, "grad_norm": 0.18783360719680786, "learning_rate": 0.00011039396479463538, "loss": 0.1609, "step": 1070 }, { "epoch": 0.450544844928751, "grad_norm": 0.23045672476291656, "learning_rate": 0.00010997485331098071, "loss": 0.1618, "step": 1075 }, { "epoch": 0.4526404023470243, "grad_norm": 0.2051040381193161, "learning_rate": 0.00010955574182732607, "loss": 0.164, "step": 1080 }, { "epoch": 0.45473595976529757, "grad_norm": 0.17596812546253204, "learning_rate": 0.00010913663034367143, "loss": 0.1578, "step": 1085 }, { "epoch": 0.45683151718357085, "grad_norm": 0.1564697027206421, "learning_rate": 0.00010871751886001677, "loss": 0.1549, "step": 1090 }, { "epoch": 0.45892707460184406, "grad_norm": 0.2508351802825928, "learning_rate": 0.00010829840737636212, "loss": 0.1541, "step": 1095 }, { "epoch": 0.46102263202011734, "grad_norm": 0.17002500593662262, "learning_rate": 0.00010787929589270746, "loss": 0.1606, "step": 1100 }, { "epoch": 0.4631181894383906, "grad_norm": 0.1768285632133484, "learning_rate": 0.00010746018440905281, "loss": 0.1533, "step": 1105 }, { "epoch": 0.4652137468566639, "grad_norm": 0.21098843216896057, "learning_rate": 0.00010704107292539815, "loss": 0.1598, "step": 1110 }, { "epoch": 0.4673093042749371, "grad_norm": 0.1743684709072113, "learning_rate": 0.0001066219614417435, "loss": 0.1577, "step": 1115 }, { "epoch": 0.4694048616932104, "grad_norm": 0.18270978331565857, "learning_rate": 0.00010620284995808884, "loss": 0.1533, "step": 1120 }, { "epoch": 0.47150041911148366, "grad_norm": 0.19265097379684448, "learning_rate": 0.00010578373847443422, "loss": 0.1545, "step": 1125 }, { "epoch": 0.47359597652975693, "grad_norm": 0.19420358538627625, "learning_rate": 0.00010536462699077956, "loss": 0.1587, "step": 1130 }, { "epoch": 0.47569153394803015, "grad_norm": 0.1721310317516327, "learning_rate": 0.00010494551550712491, "loss": 0.1577, "step": 1135 }, { "epoch": 0.47778709136630343, "grad_norm": 0.1586717814207077, "learning_rate": 0.00010452640402347025, "loss": 0.1551, "step": 1140 }, { "epoch": 0.4798826487845767, "grad_norm": 1.5393892526626587, "learning_rate": 0.0001041072925398156, "loss": 0.1537, "step": 1145 }, { "epoch": 0.48197820620285, "grad_norm": 0.17265351116657257, "learning_rate": 0.00010368818105616094, "loss": 0.1514, "step": 1150 }, { "epoch": 0.4840737636211232, "grad_norm": 0.15501521527767181, "learning_rate": 0.00010326906957250629, "loss": 0.1559, "step": 1155 }, { "epoch": 0.4861693210393965, "grad_norm": 0.1425572782754898, "learning_rate": 0.00010284995808885165, "loss": 0.1532, "step": 1160 }, { "epoch": 0.48826487845766975, "grad_norm": 0.1547863483428955, "learning_rate": 0.00010243084660519699, "loss": 0.1557, "step": 1165 }, { "epoch": 0.490360435875943, "grad_norm": 0.1579703986644745, "learning_rate": 0.00010201173512154234, "loss": 0.1568, "step": 1170 }, { "epoch": 0.49245599329421624, "grad_norm": 4.8748297691345215, "learning_rate": 0.00010159262363788768, "loss": 0.1542, "step": 1175 }, { "epoch": 0.4945515507124895, "grad_norm": 0.16955487430095673, "learning_rate": 0.00010117351215423303, "loss": 0.1573, "step": 1180 }, { "epoch": 0.4966471081307628, "grad_norm": 0.17641142010688782, "learning_rate": 0.00010075440067057837, "loss": 0.1563, "step": 1185 }, { "epoch": 0.49874266554903607, "grad_norm": 0.19048817455768585, "learning_rate": 0.00010033528918692372, "loss": 0.1565, "step": 1190 }, { "epoch": 0.5008382229673093, "grad_norm": 0.1500770002603531, "learning_rate": 9.991617770326908e-05, "loss": 0.1561, "step": 1195 }, { "epoch": 0.5029337803855826, "grad_norm": 0.1577143371105194, "learning_rate": 9.949706621961443e-05, "loss": 0.1555, "step": 1200 }, { "epoch": 0.5050293378038558, "grad_norm": 0.14767815172672272, "learning_rate": 9.907795473595977e-05, "loss": 0.1546, "step": 1205 }, { "epoch": 0.5071248952221291, "grad_norm": 0.1728920042514801, "learning_rate": 9.865884325230512e-05, "loss": 0.1582, "step": 1210 }, { "epoch": 0.5092204526404024, "grad_norm": 0.13907761871814728, "learning_rate": 9.823973176865047e-05, "loss": 0.154, "step": 1215 }, { "epoch": 0.5113160100586757, "grad_norm": 0.14644944667816162, "learning_rate": 9.782062028499581e-05, "loss": 0.151, "step": 1220 }, { "epoch": 0.5134115674769488, "grad_norm": 0.1749754697084427, "learning_rate": 9.740150880134116e-05, "loss": 0.1558, "step": 1225 }, { "epoch": 0.5155071248952221, "grad_norm": 1.7467832565307617, "learning_rate": 9.698239731768651e-05, "loss": 0.154, "step": 1230 }, { "epoch": 0.5176026823134954, "grad_norm": 0.3709283471107483, "learning_rate": 9.656328583403186e-05, "loss": 0.1648, "step": 1235 }, { "epoch": 0.5196982397317687, "grad_norm": 0.28035739064216614, "learning_rate": 9.61441743503772e-05, "loss": 0.1697, "step": 1240 }, { "epoch": 0.5217937971500419, "grad_norm": 0.1671634167432785, "learning_rate": 9.572506286672255e-05, "loss": 0.1611, "step": 1245 }, { "epoch": 0.5238893545683152, "grad_norm": 0.15249969065189362, "learning_rate": 9.53059513830679e-05, "loss": 0.1557, "step": 1250 }, { "epoch": 0.5259849119865885, "grad_norm": 0.14879916608333588, "learning_rate": 9.488683989941326e-05, "loss": 0.1516, "step": 1255 }, { "epoch": 0.5280804694048616, "grad_norm": 2.7410941123962402, "learning_rate": 9.44677284157586e-05, "loss": 0.1497, "step": 1260 }, { "epoch": 0.5301760268231349, "grad_norm": 0.188863143324852, "learning_rate": 9.404861693210395e-05, "loss": 0.1567, "step": 1265 }, { "epoch": 0.5322715842414082, "grad_norm": 0.22812993824481964, "learning_rate": 9.36295054484493e-05, "loss": 0.1586, "step": 1270 }, { "epoch": 0.5343671416596815, "grad_norm": 0.23461438715457916, "learning_rate": 9.321039396479464e-05, "loss": 0.1593, "step": 1275 }, { "epoch": 0.5364626990779547, "grad_norm": 0.23611848056316376, "learning_rate": 9.279128248113999e-05, "loss": 0.1575, "step": 1280 }, { "epoch": 0.538558256496228, "grad_norm": 0.20142588019371033, "learning_rate": 9.237217099748533e-05, "loss": 0.1611, "step": 1285 }, { "epoch": 0.5406538139145013, "grad_norm": 0.18650312721729279, "learning_rate": 9.195305951383069e-05, "loss": 0.1585, "step": 1290 }, { "epoch": 0.5427493713327746, "grad_norm": 0.1717916578054428, "learning_rate": 9.153394803017603e-05, "loss": 0.1566, "step": 1295 }, { "epoch": 0.5448449287510477, "grad_norm": 0.154686838388443, "learning_rate": 9.111483654652138e-05, "loss": 0.154, "step": 1300 }, { "epoch": 0.546940486169321, "grad_norm": 0.17387616634368896, "learning_rate": 9.069572506286673e-05, "loss": 0.1574, "step": 1305 }, { "epoch": 0.5490360435875943, "grad_norm": 0.1707635372877121, "learning_rate": 9.027661357921207e-05, "loss": 0.1595, "step": 1310 }, { "epoch": 0.5511316010058676, "grad_norm": 0.1505287140607834, "learning_rate": 8.985750209555742e-05, "loss": 0.1572, "step": 1315 }, { "epoch": 0.5532271584241408, "grad_norm": 0.17044112086296082, "learning_rate": 8.943839061190276e-05, "loss": 0.1526, "step": 1320 }, { "epoch": 0.5553227158424141, "grad_norm": 0.20292381942272186, "learning_rate": 8.901927912824813e-05, "loss": 0.157, "step": 1325 }, { "epoch": 0.5574182732606874, "grad_norm": 0.22127684950828552, "learning_rate": 8.860016764459346e-05, "loss": 0.1569, "step": 1330 }, { "epoch": 0.5595138306789607, "grad_norm": 0.169328510761261, "learning_rate": 8.818105616093882e-05, "loss": 0.1574, "step": 1335 }, { "epoch": 0.5616093880972338, "grad_norm": 0.1888064444065094, "learning_rate": 8.776194467728415e-05, "loss": 0.152, "step": 1340 }, { "epoch": 0.5637049455155071, "grad_norm": 2.7211835384368896, "learning_rate": 8.734283319362952e-05, "loss": 0.1557, "step": 1345 }, { "epoch": 0.5658005029337804, "grad_norm": 0.18435344099998474, "learning_rate": 8.692372170997486e-05, "loss": 0.1527, "step": 1350 }, { "epoch": 0.5678960603520536, "grad_norm": 0.2031932920217514, "learning_rate": 8.650461022632021e-05, "loss": 0.1573, "step": 1355 }, { "epoch": 0.5699916177703269, "grad_norm": 0.1941346377134323, "learning_rate": 8.608549874266555e-05, "loss": 0.1538, "step": 1360 }, { "epoch": 0.5720871751886002, "grad_norm": 0.18645240366458893, "learning_rate": 8.56663872590109e-05, "loss": 0.1542, "step": 1365 }, { "epoch": 0.5741827326068735, "grad_norm": 0.1776381880044937, "learning_rate": 8.524727577535625e-05, "loss": 0.1567, "step": 1370 }, { "epoch": 0.5762782900251467, "grad_norm": 0.17730367183685303, "learning_rate": 8.482816429170159e-05, "loss": 0.1552, "step": 1375 }, { "epoch": 0.5783738474434199, "grad_norm": 0.1614091843366623, "learning_rate": 8.440905280804694e-05, "loss": 0.1545, "step": 1380 }, { "epoch": 0.5804694048616932, "grad_norm": 0.2010612040758133, "learning_rate": 8.398994132439229e-05, "loss": 0.1525, "step": 1385 }, { "epoch": 0.5825649622799665, "grad_norm": 0.29623404145240784, "learning_rate": 8.357082984073764e-05, "loss": 0.1632, "step": 1390 }, { "epoch": 0.5846605196982397, "grad_norm": 0.26536107063293457, "learning_rate": 8.315171835708298e-05, "loss": 0.161, "step": 1395 }, { "epoch": 0.586756077116513, "grad_norm": 0.21128953993320465, "learning_rate": 8.273260687342833e-05, "loss": 0.1585, "step": 1400 }, { "epoch": 0.5888516345347863, "grad_norm": 3.6902313232421875, "learning_rate": 8.231349538977369e-05, "loss": 0.153, "step": 1405 }, { "epoch": 0.5909471919530596, "grad_norm": 0.18794021010398865, "learning_rate": 8.189438390611904e-05, "loss": 0.1585, "step": 1410 }, { "epoch": 0.5930427493713327, "grad_norm": 0.17947795987129211, "learning_rate": 8.147527242246438e-05, "loss": 0.1584, "step": 1415 }, { "epoch": 0.595138306789606, "grad_norm": 0.15674275159835815, "learning_rate": 8.105616093880973e-05, "loss": 0.1547, "step": 1420 }, { "epoch": 0.5972338642078793, "grad_norm": 0.15368077158927917, "learning_rate": 8.063704945515508e-05, "loss": 0.154, "step": 1425 }, { "epoch": 0.5993294216261525, "grad_norm": 0.1675073206424713, "learning_rate": 8.021793797150042e-05, "loss": 0.1519, "step": 1430 }, { "epoch": 0.6014249790444258, "grad_norm": 0.15424901247024536, "learning_rate": 7.979882648784577e-05, "loss": 0.1532, "step": 1435 }, { "epoch": 0.6035205364626991, "grad_norm": 1.752557396888733, "learning_rate": 7.937971500419112e-05, "loss": 0.1521, "step": 1440 }, { "epoch": 0.6056160938809724, "grad_norm": 0.179281547665596, "learning_rate": 7.896060352053647e-05, "loss": 0.1568, "step": 1445 }, { "epoch": 0.6077116512992456, "grad_norm": 0.18618442118167877, "learning_rate": 7.854149203688181e-05, "loss": 0.1555, "step": 1450 }, { "epoch": 0.6098072087175188, "grad_norm": 0.1767512410879135, "learning_rate": 7.812238055322716e-05, "loss": 0.156, "step": 1455 }, { "epoch": 0.6119027661357921, "grad_norm": 0.1729685217142105, "learning_rate": 7.770326906957252e-05, "loss": 0.1521, "step": 1460 }, { "epoch": 0.6139983235540654, "grad_norm": 0.16246432065963745, "learning_rate": 7.728415758591785e-05, "loss": 0.1541, "step": 1465 }, { "epoch": 0.6160938809723386, "grad_norm": 0.15939489006996155, "learning_rate": 7.68650461022632e-05, "loss": 0.1551, "step": 1470 }, { "epoch": 0.6181894383906119, "grad_norm": 0.1684809774160385, "learning_rate": 7.644593461860854e-05, "loss": 0.1542, "step": 1475 }, { "epoch": 0.6202849958088852, "grad_norm": 0.18968342244625092, "learning_rate": 7.602682313495391e-05, "loss": 0.1537, "step": 1480 }, { "epoch": 0.6223805532271585, "grad_norm": 0.20144881308078766, "learning_rate": 7.560771165129925e-05, "loss": 0.158, "step": 1485 }, { "epoch": 0.6244761106454317, "grad_norm": 0.1785353720188141, "learning_rate": 7.51886001676446e-05, "loss": 0.1537, "step": 1490 }, { "epoch": 0.6265716680637049, "grad_norm": 3.9302639961242676, "learning_rate": 7.476948868398994e-05, "loss": 0.1539, "step": 1495 }, { "epoch": 0.6286672254819782, "grad_norm": 0.17349500954151154, "learning_rate": 7.43503772003353e-05, "loss": 0.1549, "step": 1500 }, { "epoch": 0.6307627829002515, "grad_norm": 0.181074857711792, "learning_rate": 7.393126571668064e-05, "loss": 0.1523, "step": 1505 }, { "epoch": 0.6328583403185247, "grad_norm": 0.1837494969367981, "learning_rate": 7.351215423302599e-05, "loss": 0.1543, "step": 1510 }, { "epoch": 0.634953897736798, "grad_norm": 0.17972221970558167, "learning_rate": 7.309304274937133e-05, "loss": 0.1539, "step": 1515 }, { "epoch": 0.6370494551550713, "grad_norm": 0.18084241449832916, "learning_rate": 7.267393126571668e-05, "loss": 0.1504, "step": 1520 }, { "epoch": 0.6391450125733446, "grad_norm": 0.16487205028533936, "learning_rate": 7.225481978206203e-05, "loss": 0.1547, "step": 1525 }, { "epoch": 0.6412405699916177, "grad_norm": 3.8815670013427734, "learning_rate": 7.183570829840737e-05, "loss": 0.1481, "step": 1530 }, { "epoch": 0.643336127409891, "grad_norm": 0.18625426292419434, "learning_rate": 7.141659681475272e-05, "loss": 0.1551, "step": 1535 }, { "epoch": 0.6454316848281643, "grad_norm": 0.2006104439496994, "learning_rate": 7.099748533109808e-05, "loss": 0.1567, "step": 1540 }, { "epoch": 0.6475272422464375, "grad_norm": 0.20722784101963043, "learning_rate": 7.057837384744343e-05, "loss": 0.1599, "step": 1545 }, { "epoch": 0.6496227996647108, "grad_norm": 0.18843336403369904, "learning_rate": 7.015926236378877e-05, "loss": 0.1568, "step": 1550 }, { "epoch": 0.6517183570829841, "grad_norm": 0.16453127562999725, "learning_rate": 6.974015088013412e-05, "loss": 0.1564, "step": 1555 }, { "epoch": 0.6538139145012574, "grad_norm": 0.1718008667230606, "learning_rate": 6.932103939647947e-05, "loss": 0.1507, "step": 1560 }, { "epoch": 0.6559094719195306, "grad_norm": 1.9749072790145874, "learning_rate": 6.890192791282482e-05, "loss": 0.1546, "step": 1565 }, { "epoch": 0.6580050293378038, "grad_norm": 0.846435546875, "learning_rate": 6.848281642917016e-05, "loss": 0.1553, "step": 1570 }, { "epoch": 0.6601005867560771, "grad_norm": 0.21918132901191711, "learning_rate": 6.806370494551551e-05, "loss": 0.1544, "step": 1575 }, { "epoch": 0.6621961441743504, "grad_norm": 0.20493735373020172, "learning_rate": 6.764459346186086e-05, "loss": 0.1588, "step": 1580 }, { "epoch": 0.6642917015926236, "grad_norm": 0.16465319693088531, "learning_rate": 6.72254819782062e-05, "loss": 0.1556, "step": 1585 }, { "epoch": 0.6663872590108969, "grad_norm": 0.1511167287826538, "learning_rate": 6.680637049455155e-05, "loss": 0.1551, "step": 1590 }, { "epoch": 0.6684828164291702, "grad_norm": 0.15197286009788513, "learning_rate": 6.63872590108969e-05, "loss": 0.1513, "step": 1595 }, { "epoch": 0.6705783738474435, "grad_norm": 0.14843755960464478, "learning_rate": 6.596814752724226e-05, "loss": 0.1476, "step": 1600 }, { "epoch": 0.6726739312657167, "grad_norm": 0.14049085974693298, "learning_rate": 6.55490360435876e-05, "loss": 0.1507, "step": 1605 }, { "epoch": 0.6747694886839899, "grad_norm": 0.1492355465888977, "learning_rate": 6.512992455993295e-05, "loss": 0.1499, "step": 1610 }, { "epoch": 0.6768650461022632, "grad_norm": 0.15042835474014282, "learning_rate": 6.47108130762783e-05, "loss": 0.15, "step": 1615 }, { "epoch": 0.6789606035205364, "grad_norm": 0.14849700033664703, "learning_rate": 6.429170159262365e-05, "loss": 0.152, "step": 1620 }, { "epoch": 0.6810561609388097, "grad_norm": 0.15401747822761536, "learning_rate": 6.387259010896899e-05, "loss": 0.1507, "step": 1625 }, { "epoch": 0.683151718357083, "grad_norm": 0.13168840110301971, "learning_rate": 6.345347862531434e-05, "loss": 0.153, "step": 1630 }, { "epoch": 0.6852472757753563, "grad_norm": 0.1513977199792862, "learning_rate": 6.303436714165969e-05, "loss": 0.1516, "step": 1635 }, { "epoch": 0.6873428331936295, "grad_norm": 0.14483259618282318, "learning_rate": 6.261525565800503e-05, "loss": 0.15, "step": 1640 }, { "epoch": 0.6894383906119028, "grad_norm": 0.15542860329151154, "learning_rate": 6.219614417435038e-05, "loss": 0.1493, "step": 1645 }, { "epoch": 0.691533948030176, "grad_norm": 0.14083847403526306, "learning_rate": 6.177703269069572e-05, "loss": 0.1511, "step": 1650 }, { "epoch": 0.6936295054484493, "grad_norm": 0.14129005372524261, "learning_rate": 6.135792120704108e-05, "loss": 0.1525, "step": 1655 }, { "epoch": 0.6957250628667225, "grad_norm": 0.14137649536132812, "learning_rate": 6.093880972338642e-05, "loss": 0.1486, "step": 1660 }, { "epoch": 0.6978206202849958, "grad_norm": 0.163183331489563, "learning_rate": 6.051969823973177e-05, "loss": 0.1526, "step": 1665 }, { "epoch": 0.6999161777032691, "grad_norm": 0.15466581284999847, "learning_rate": 6.010058675607711e-05, "loss": 0.1523, "step": 1670 }, { "epoch": 0.7020117351215424, "grad_norm": 0.1582300364971161, "learning_rate": 5.968147527242247e-05, "loss": 0.1501, "step": 1675 }, { "epoch": 0.7041072925398156, "grad_norm": 0.15225544571876526, "learning_rate": 5.9262363788767817e-05, "loss": 0.1529, "step": 1680 }, { "epoch": 0.7062028499580888, "grad_norm": 0.1497444212436676, "learning_rate": 5.884325230511316e-05, "loss": 0.1504, "step": 1685 }, { "epoch": 0.7082984073763621, "grad_norm": 0.15386608242988586, "learning_rate": 5.8424140821458507e-05, "loss": 0.1507, "step": 1690 }, { "epoch": 0.7103939647946353, "grad_norm": 0.16286878287792206, "learning_rate": 5.8005029337803865e-05, "loss": 0.1483, "step": 1695 }, { "epoch": 0.7124895222129086, "grad_norm": 0.16049712896347046, "learning_rate": 5.758591785414921e-05, "loss": 0.1511, "step": 1700 }, { "epoch": 0.7145850796311819, "grad_norm": 0.16950005292892456, "learning_rate": 5.7166806370494555e-05, "loss": 0.1497, "step": 1705 }, { "epoch": 0.7166806370494552, "grad_norm": 0.14905424416065216, "learning_rate": 5.67476948868399e-05, "loss": 0.1518, "step": 1710 }, { "epoch": 0.7187761944677284, "grad_norm": 0.16250193119049072, "learning_rate": 5.632858340318525e-05, "loss": 0.1518, "step": 1715 }, { "epoch": 0.7208717518860017, "grad_norm": 0.15719325840473175, "learning_rate": 5.59094719195306e-05, "loss": 0.1509, "step": 1720 }, { "epoch": 0.7229673093042749, "grad_norm": 0.13987241685390472, "learning_rate": 5.549036043587594e-05, "loss": 0.1493, "step": 1725 }, { "epoch": 0.7250628667225482, "grad_norm": 0.15766645967960358, "learning_rate": 5.507124895222129e-05, "loss": 0.151, "step": 1730 }, { "epoch": 0.7271584241408214, "grad_norm": 0.1426958292722702, "learning_rate": 5.4652137468566645e-05, "loss": 0.1477, "step": 1735 }, { "epoch": 0.7292539815590947, "grad_norm": 0.15173014998435974, "learning_rate": 5.423302598491199e-05, "loss": 0.1498, "step": 1740 }, { "epoch": 0.731349538977368, "grad_norm": 0.13781285285949707, "learning_rate": 5.3813914501257335e-05, "loss": 0.1461, "step": 1745 }, { "epoch": 0.7334450963956413, "grad_norm": 0.15598557889461517, "learning_rate": 5.339480301760269e-05, "loss": 0.1468, "step": 1750 }, { "epoch": 0.7355406538139145, "grad_norm": 0.1487160325050354, "learning_rate": 5.297569153394803e-05, "loss": 0.1515, "step": 1755 }, { "epoch": 0.7376362112321878, "grad_norm": 0.1766008883714676, "learning_rate": 5.255658005029338e-05, "loss": 0.1496, "step": 1760 }, { "epoch": 0.739731768650461, "grad_norm": 0.1634136140346527, "learning_rate": 5.213746856663872e-05, "loss": 0.152, "step": 1765 }, { "epoch": 0.7418273260687342, "grad_norm": 0.16231022775173187, "learning_rate": 5.171835708298408e-05, "loss": 0.1511, "step": 1770 }, { "epoch": 0.7439228834870075, "grad_norm": 0.1736147552728653, "learning_rate": 5.1299245599329425e-05, "loss": 0.1519, "step": 1775 }, { "epoch": 0.7460184409052808, "grad_norm": 0.15696606040000916, "learning_rate": 5.088013411567477e-05, "loss": 0.1522, "step": 1780 }, { "epoch": 0.7481139983235541, "grad_norm": 0.16596518456935883, "learning_rate": 5.0461022632020115e-05, "loss": 0.1494, "step": 1785 }, { "epoch": 0.7502095557418273, "grad_norm": 0.16456949710845947, "learning_rate": 5.0041911148365474e-05, "loss": 0.1485, "step": 1790 }, { "epoch": 0.7523051131601006, "grad_norm": 0.16951188445091248, "learning_rate": 4.962279966471082e-05, "loss": 0.1536, "step": 1795 }, { "epoch": 0.7544006705783739, "grad_norm": 0.17507892847061157, "learning_rate": 4.9203688181056164e-05, "loss": 0.1524, "step": 1800 }, { "epoch": 0.7564962279966471, "grad_norm": 0.1613740175962448, "learning_rate": 4.8784576697401516e-05, "loss": 0.151, "step": 1805 }, { "epoch": 0.7585917854149203, "grad_norm": 0.1735554188489914, "learning_rate": 4.836546521374686e-05, "loss": 0.149, "step": 1810 }, { "epoch": 0.7606873428331936, "grad_norm": 0.15241263806819916, "learning_rate": 4.7946353730092206e-05, "loss": 0.1475, "step": 1815 }, { "epoch": 0.7627829002514669, "grad_norm": 0.15598662197589874, "learning_rate": 4.752724224643755e-05, "loss": 0.1504, "step": 1820 }, { "epoch": 0.7648784576697402, "grad_norm": 0.1638031303882599, "learning_rate": 4.71081307627829e-05, "loss": 0.1483, "step": 1825 }, { "epoch": 0.7669740150880134, "grad_norm": 0.15647321939468384, "learning_rate": 4.668901927912825e-05, "loss": 0.1487, "step": 1830 }, { "epoch": 0.7690695725062867, "grad_norm": 0.15045635402202606, "learning_rate": 4.62699077954736e-05, "loss": 0.1489, "step": 1835 }, { "epoch": 0.7711651299245599, "grad_norm": 0.1481800377368927, "learning_rate": 4.5850796311818944e-05, "loss": 0.1498, "step": 1840 }, { "epoch": 0.7732606873428332, "grad_norm": 0.16117189824581146, "learning_rate": 4.5431684828164296e-05, "loss": 0.147, "step": 1845 }, { "epoch": 0.7753562447611064, "grad_norm": 0.15690650045871735, "learning_rate": 4.501257334450964e-05, "loss": 0.152, "step": 1850 }, { "epoch": 0.7774518021793797, "grad_norm": 0.15324068069458008, "learning_rate": 4.459346186085499e-05, "loss": 0.1509, "step": 1855 }, { "epoch": 0.779547359597653, "grad_norm": 0.1670048087835312, "learning_rate": 4.417435037720034e-05, "loss": 0.1481, "step": 1860 }, { "epoch": 0.7816429170159263, "grad_norm": 0.16738423705101013, "learning_rate": 4.375523889354568e-05, "loss": 0.1525, "step": 1865 }, { "epoch": 0.7837384744341995, "grad_norm": 0.15926587581634521, "learning_rate": 4.333612740989103e-05, "loss": 0.1499, "step": 1870 }, { "epoch": 0.7858340318524728, "grad_norm": 0.15555323660373688, "learning_rate": 4.291701592623638e-05, "loss": 0.1487, "step": 1875 }, { "epoch": 0.787929589270746, "grad_norm": 0.14897003769874573, "learning_rate": 4.2497904442581724e-05, "loss": 0.153, "step": 1880 }, { "epoch": 0.7900251466890192, "grad_norm": 0.16712865233421326, "learning_rate": 4.2078792958927076e-05, "loss": 0.1522, "step": 1885 }, { "epoch": 0.7921207041072925, "grad_norm": 0.16521087288856506, "learning_rate": 4.165968147527242e-05, "loss": 0.1554, "step": 1890 }, { "epoch": 0.7942162615255658, "grad_norm": 0.15836192667484283, "learning_rate": 4.124056999161777e-05, "loss": 0.1515, "step": 1895 }, { "epoch": 0.7963118189438391, "grad_norm": 0.1507478654384613, "learning_rate": 4.0821458507963125e-05, "loss": 0.1525, "step": 1900 }, { "epoch": 0.7984073763621123, "grad_norm": 0.6994775533676147, "learning_rate": 4.040234702430847e-05, "loss": 0.1522, "step": 1905 }, { "epoch": 0.8005029337803856, "grad_norm": 0.1698751598596573, "learning_rate": 3.998323554065382e-05, "loss": 0.152, "step": 1910 }, { "epoch": 0.8025984911986589, "grad_norm": 0.16350072622299194, "learning_rate": 3.9564124056999166e-05, "loss": 0.1499, "step": 1915 }, { "epoch": 0.8046940486169321, "grad_norm": 0.15547800064086914, "learning_rate": 3.914501257334451e-05, "loss": 0.1496, "step": 1920 }, { "epoch": 0.8067896060352053, "grad_norm": 0.1584520936012268, "learning_rate": 3.8725901089689856e-05, "loss": 0.1521, "step": 1925 }, { "epoch": 0.8088851634534786, "grad_norm": 0.15392720699310303, "learning_rate": 3.830678960603521e-05, "loss": 0.1513, "step": 1930 }, { "epoch": 0.8109807208717519, "grad_norm": 0.15251478552818298, "learning_rate": 3.788767812238055e-05, "loss": 0.1503, "step": 1935 }, { "epoch": 0.8130762782900252, "grad_norm": 0.1555824875831604, "learning_rate": 3.7468566638725905e-05, "loss": 0.148, "step": 1940 }, { "epoch": 0.8151718357082984, "grad_norm": 1.0603052377700806, "learning_rate": 3.704945515507125e-05, "loss": 0.1511, "step": 1945 }, { "epoch": 0.8172673931265717, "grad_norm": 0.21994204819202423, "learning_rate": 3.66303436714166e-05, "loss": 0.1536, "step": 1950 }, { "epoch": 0.819362950544845, "grad_norm": 0.2238265573978424, "learning_rate": 3.6211232187761947e-05, "loss": 0.1549, "step": 1955 }, { "epoch": 0.8214585079631181, "grad_norm": 0.212602898478508, "learning_rate": 3.57921207041073e-05, "loss": 0.1567, "step": 1960 }, { "epoch": 0.8235540653813914, "grad_norm": 0.18257835507392883, "learning_rate": 3.537300922045264e-05, "loss": 0.1513, "step": 1965 }, { "epoch": 0.8256496227996647, "grad_norm": 0.17607073485851288, "learning_rate": 3.495389773679799e-05, "loss": 0.1539, "step": 1970 }, { "epoch": 0.827745180217938, "grad_norm": 0.14804469048976898, "learning_rate": 3.453478625314333e-05, "loss": 0.1502, "step": 1975 }, { "epoch": 0.8298407376362112, "grad_norm": 0.15939727425575256, "learning_rate": 3.4115674769488685e-05, "loss": 0.1482, "step": 1980 }, { "epoch": 0.8319362950544845, "grad_norm": 0.16081516444683075, "learning_rate": 3.369656328583403e-05, "loss": 0.152, "step": 1985 }, { "epoch": 0.8340318524727578, "grad_norm": 0.7829539179801941, "learning_rate": 3.327745180217938e-05, "loss": 0.1497, "step": 1990 }, { "epoch": 0.836127409891031, "grad_norm": 0.15778230130672455, "learning_rate": 3.285834031852473e-05, "loss": 0.1485, "step": 1995 }, { "epoch": 0.8382229673093042, "grad_norm": 0.16156698763370514, "learning_rate": 3.243922883487008e-05, "loss": 0.1508, "step": 2000 }, { "epoch": 0.8403185247275775, "grad_norm": 0.15631239116191864, "learning_rate": 3.2020117351215424e-05, "loss": 0.1484, "step": 2005 }, { "epoch": 0.8424140821458508, "grad_norm": 0.1522732824087143, "learning_rate": 3.1601005867560775e-05, "loss": 0.1508, "step": 2010 }, { "epoch": 0.8445096395641241, "grad_norm": 0.1581004112958908, "learning_rate": 3.118189438390612e-05, "loss": 0.149, "step": 2015 }, { "epoch": 0.8466051969823973, "grad_norm": 0.1513088047504425, "learning_rate": 3.076278290025147e-05, "loss": 0.15, "step": 2020 }, { "epoch": 0.8487007544006706, "grad_norm": 0.1780407577753067, "learning_rate": 3.0343671416596814e-05, "loss": 0.1494, "step": 2025 }, { "epoch": 0.8507963118189439, "grad_norm": 0.16309364140033722, "learning_rate": 2.9924559932942165e-05, "loss": 0.1474, "step": 2030 }, { "epoch": 0.852891869237217, "grad_norm": 0.15739715099334717, "learning_rate": 2.950544844928751e-05, "loss": 0.1496, "step": 2035 }, { "epoch": 0.8549874266554903, "grad_norm": 0.1566157042980194, "learning_rate": 2.908633696563286e-05, "loss": 0.1482, "step": 2040 }, { "epoch": 0.8570829840737636, "grad_norm": 0.14910633862018585, "learning_rate": 2.8667225481978204e-05, "loss": 0.1493, "step": 2045 }, { "epoch": 0.8591785414920369, "grad_norm": 0.15060973167419434, "learning_rate": 2.8248113998323556e-05, "loss": 0.1462, "step": 2050 }, { "epoch": 0.8612740989103101, "grad_norm": 0.15438151359558105, "learning_rate": 2.7829002514668907e-05, "loss": 0.1509, "step": 2055 }, { "epoch": 0.8633696563285834, "grad_norm": 0.15441566705703735, "learning_rate": 2.7409891031014252e-05, "loss": 0.1481, "step": 2060 }, { "epoch": 0.8654652137468567, "grad_norm": 0.15039560198783875, "learning_rate": 2.69907795473596e-05, "loss": 0.1466, "step": 2065 }, { "epoch": 0.86756077116513, "grad_norm": 0.14182990789413452, "learning_rate": 2.6571668063704946e-05, "loss": 0.1475, "step": 2070 }, { "epoch": 0.8696563285834031, "grad_norm": 0.1383397877216339, "learning_rate": 2.6152556580050297e-05, "loss": 0.1498, "step": 2075 }, { "epoch": 0.8717518860016764, "grad_norm": 0.1353161633014679, "learning_rate": 2.5733445096395642e-05, "loss": 0.1485, "step": 2080 }, { "epoch": 0.8738474434199497, "grad_norm": 0.14230230450630188, "learning_rate": 2.5314333612740994e-05, "loss": 0.1488, "step": 2085 }, { "epoch": 0.875943000838223, "grad_norm": 0.14749757945537567, "learning_rate": 2.4895222129086336e-05, "loss": 0.1464, "step": 2090 }, { "epoch": 0.8780385582564962, "grad_norm": 0.1343354731798172, "learning_rate": 2.4476110645431684e-05, "loss": 0.1476, "step": 2095 }, { "epoch": 0.8801341156747695, "grad_norm": 0.15284012258052826, "learning_rate": 2.4056999161777032e-05, "loss": 0.1499, "step": 2100 }, { "epoch": 0.8822296730930428, "grad_norm": 0.14641784131526947, "learning_rate": 2.363788767812238e-05, "loss": 0.1473, "step": 2105 }, { "epoch": 0.8843252305113161, "grad_norm": 0.13705027103424072, "learning_rate": 2.321877619446773e-05, "loss": 0.1464, "step": 2110 }, { "epoch": 0.8864207879295892, "grad_norm": 0.15073353052139282, "learning_rate": 2.2799664710813078e-05, "loss": 0.1464, "step": 2115 }, { "epoch": 0.8885163453478625, "grad_norm": 0.14341919124126434, "learning_rate": 2.2380553227158423e-05, "loss": 0.1479, "step": 2120 }, { "epoch": 0.8906119027661358, "grad_norm": 0.14984482526779175, "learning_rate": 2.196144174350377e-05, "loss": 0.1468, "step": 2125 }, { "epoch": 0.892707460184409, "grad_norm": 0.1421850025653839, "learning_rate": 2.1542330259849123e-05, "loss": 0.1431, "step": 2130 }, { "epoch": 0.8948030176026823, "grad_norm": 0.15779058635234833, "learning_rate": 2.112321877619447e-05, "loss": 0.1484, "step": 2135 }, { "epoch": 0.8968985750209556, "grad_norm": 0.14414595067501068, "learning_rate": 2.0704107292539816e-05, "loss": 0.1477, "step": 2140 }, { "epoch": 0.8989941324392289, "grad_norm": 0.15078584849834442, "learning_rate": 2.0284995808885164e-05, "loss": 0.1471, "step": 2145 }, { "epoch": 0.901089689857502, "grad_norm": 0.6652347445487976, "learning_rate": 1.9865884325230513e-05, "loss": 0.146, "step": 2150 }, { "epoch": 0.9031852472757753, "grad_norm": 0.1553775519132614, "learning_rate": 1.944677284157586e-05, "loss": 0.1455, "step": 2155 }, { "epoch": 0.9052808046940486, "grad_norm": 0.14678525924682617, "learning_rate": 1.902766135792121e-05, "loss": 0.1469, "step": 2160 }, { "epoch": 0.9073763621123219, "grad_norm": 0.1497979462146759, "learning_rate": 1.8608549874266558e-05, "loss": 0.1488, "step": 2165 }, { "epoch": 0.9094719195305951, "grad_norm": 0.14247293770313263, "learning_rate": 1.8189438390611903e-05, "loss": 0.147, "step": 2170 }, { "epoch": 0.9115674769488684, "grad_norm": 0.1436048001050949, "learning_rate": 1.777032690695725e-05, "loss": 0.148, "step": 2175 }, { "epoch": 0.9136630343671417, "grad_norm": 0.1506524235010147, "learning_rate": 1.73512154233026e-05, "loss": 0.147, "step": 2180 }, { "epoch": 0.915758591785415, "grad_norm": 0.14503729343414307, "learning_rate": 1.6932103939647948e-05, "loss": 0.1475, "step": 2185 }, { "epoch": 0.9178541492036881, "grad_norm": 0.14085812866687775, "learning_rate": 1.6512992455993296e-05, "loss": 0.1468, "step": 2190 }, { "epoch": 0.9199497066219614, "grad_norm": 0.13786457479000092, "learning_rate": 1.609388097233864e-05, "loss": 0.1449, "step": 2195 }, { "epoch": 0.9220452640402347, "grad_norm": 0.14363813400268555, "learning_rate": 1.567476948868399e-05, "loss": 0.144, "step": 2200 }, { "epoch": 0.924140821458508, "grad_norm": 0.14791929721832275, "learning_rate": 1.5255658005029338e-05, "loss": 0.147, "step": 2205 }, { "epoch": 0.9262363788767812, "grad_norm": 0.14655451476573944, "learning_rate": 1.4836546521374687e-05, "loss": 0.1496, "step": 2210 }, { "epoch": 0.9283319362950545, "grad_norm": 0.14993825554847717, "learning_rate": 1.4417435037720033e-05, "loss": 0.1465, "step": 2215 }, { "epoch": 0.9304274937133278, "grad_norm": 0.1357424557209015, "learning_rate": 1.3998323554065382e-05, "loss": 0.1478, "step": 2220 }, { "epoch": 0.932523051131601, "grad_norm": 0.15557856857776642, "learning_rate": 1.357921207041073e-05, "loss": 0.1429, "step": 2225 }, { "epoch": 0.9346186085498742, "grad_norm": 0.16618798673152924, "learning_rate": 1.3160100586756077e-05, "loss": 0.1475, "step": 2230 }, { "epoch": 0.9367141659681475, "grad_norm": 0.14325110614299774, "learning_rate": 1.2740989103101425e-05, "loss": 0.1453, "step": 2235 }, { "epoch": 0.9388097233864208, "grad_norm": 0.15656636655330658, "learning_rate": 1.2321877619446773e-05, "loss": 0.1463, "step": 2240 }, { "epoch": 0.940905280804694, "grad_norm": 0.15591678023338318, "learning_rate": 1.1902766135792122e-05, "loss": 0.1475, "step": 2245 }, { "epoch": 0.9430008382229673, "grad_norm": 0.14923661947250366, "learning_rate": 1.148365465213747e-05, "loss": 0.1474, "step": 2250 }, { "epoch": 0.9450963956412406, "grad_norm": 1.0343477725982666, "learning_rate": 1.1064543168482817e-05, "loss": 0.1434, "step": 2255 }, { "epoch": 0.9471919530595139, "grad_norm": 0.16045677661895752, "learning_rate": 1.0645431684828165e-05, "loss": 0.1448, "step": 2260 }, { "epoch": 0.9492875104777871, "grad_norm": 0.14702020585536957, "learning_rate": 1.0226320201173514e-05, "loss": 0.1468, "step": 2265 }, { "epoch": 0.9513830678960603, "grad_norm": 0.14469429850578308, "learning_rate": 9.80720871751886e-06, "loss": 0.1434, "step": 2270 }, { "epoch": 0.9534786253143336, "grad_norm": 0.1469959169626236, "learning_rate": 9.388097233864209e-06, "loss": 0.1487, "step": 2275 }, { "epoch": 0.9555741827326069, "grad_norm": 0.13562710583209991, "learning_rate": 8.968985750209557e-06, "loss": 0.1477, "step": 2280 }, { "epoch": 0.9576697401508801, "grad_norm": 0.15792514383792877, "learning_rate": 8.549874266554904e-06, "loss": 0.149, "step": 2285 }, { "epoch": 0.9597652975691534, "grad_norm": 0.13964731991291046, "learning_rate": 8.130762782900252e-06, "loss": 0.144, "step": 2290 }, { "epoch": 0.9618608549874267, "grad_norm": 0.1598869264125824, "learning_rate": 7.711651299245599e-06, "loss": 0.1486, "step": 2295 }, { "epoch": 0.9639564124057, "grad_norm": 11.353910446166992, "learning_rate": 7.292539815590947e-06, "loss": 0.1441, "step": 2300 }, { "epoch": 0.9660519698239731, "grad_norm": 0.9498974084854126, "learning_rate": 6.873428331936295e-06, "loss": 0.149, "step": 2305 }, { "epoch": 0.9681475272422464, "grad_norm": 0.15554401278495789, "learning_rate": 6.454316848281643e-06, "loss": 0.1449, "step": 2310 }, { "epoch": 0.9702430846605197, "grad_norm": 0.16430126130580902, "learning_rate": 6.035205364626991e-06, "loss": 0.1451, "step": 2315 }, { "epoch": 0.972338642078793, "grad_norm": 0.15480098128318787, "learning_rate": 5.616093880972339e-06, "loss": 0.146, "step": 2320 }, { "epoch": 0.9744341994970662, "grad_norm": 0.15997706353664398, "learning_rate": 5.1969823973176864e-06, "loss": 0.1483, "step": 2325 }, { "epoch": 0.9765297569153395, "grad_norm": 0.15260910987854004, "learning_rate": 4.777870913663034e-06, "loss": 0.1472, "step": 2330 }, { "epoch": 0.9786253143336128, "grad_norm": 0.15502935647964478, "learning_rate": 4.358759430008382e-06, "loss": 0.1456, "step": 2335 }, { "epoch": 0.980720871751886, "grad_norm": 0.14989130198955536, "learning_rate": 3.939647946353731e-06, "loss": 0.1464, "step": 2340 }, { "epoch": 0.9828164291701592, "grad_norm": 0.1601138710975647, "learning_rate": 3.5205364626990782e-06, "loss": 0.1479, "step": 2345 }, { "epoch": 0.9849119865884325, "grad_norm": 0.15280233323574066, "learning_rate": 3.101424979044426e-06, "loss": 0.1459, "step": 2350 }, { "epoch": 0.9870075440067058, "grad_norm": 0.15722650289535522, "learning_rate": 2.6823134953897737e-06, "loss": 0.1484, "step": 2355 }, { "epoch": 0.989103101424979, "grad_norm": 0.15498095750808716, "learning_rate": 2.2632020117351217e-06, "loss": 0.1486, "step": 2360 }, { "epoch": 0.9911986588432523, "grad_norm": 0.16407634317874908, "learning_rate": 1.8440905280804694e-06, "loss": 0.1508, "step": 2365 }, { "epoch": 0.9932942162615256, "grad_norm": 0.1465112566947937, "learning_rate": 1.4249790444258174e-06, "loss": 0.1482, "step": 2370 }, { "epoch": 0.9953897736797989, "grad_norm": 0.1511894315481186, "learning_rate": 1.005867560771165e-06, "loss": 0.1458, "step": 2375 }, { "epoch": 0.9974853310980721, "grad_norm": 0.15166474878787994, "learning_rate": 5.86756077116513e-07, "loss": 0.147, "step": 2380 }, { "epoch": 0.9995808885163453, "grad_norm": 1.638846516609192, "learning_rate": 1.6764459346186086e-07, "loss": 0.1469, "step": 2385 }, { "epoch": 1.0, "eval_loss": 0.14722344279289246, "eval_runtime": 1611.1392, "eval_samples_per_second": 5.265, "eval_steps_per_second": 0.659, "step": 2386 } ], "logging_steps": 5, "max_steps": 2386, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 6.667359824212132e+17, "train_batch_size": 32, "trial_name": null, "trial_params": null }