LoL_Build-Llama3B / trainer_state.json
HatimF's picture
End of training
6dfeafb verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.0,
"eval_steps": 500,
"global_step": 2386,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0020955574182732607,
"grad_norm": 0.5915184020996094,
"learning_rate": 0.0001996647108130763,
"loss": 1.2488,
"step": 5
},
{
"epoch": 0.004191114836546521,
"grad_norm": 0.6416106224060059,
"learning_rate": 0.00019924559932942164,
"loss": 0.8074,
"step": 10
},
{
"epoch": 0.006286672254819782,
"grad_norm": 0.47403833270072937,
"learning_rate": 0.000198826487845767,
"loss": 0.5298,
"step": 15
},
{
"epoch": 0.008382229673093043,
"grad_norm": 0.49060583114624023,
"learning_rate": 0.00019840737636211232,
"loss": 0.4396,
"step": 20
},
{
"epoch": 0.010477787091366304,
"grad_norm": 0.3994845449924469,
"learning_rate": 0.00019798826487845767,
"loss": 0.3997,
"step": 25
},
{
"epoch": 0.012573344509639563,
"grad_norm": 5.743133068084717,
"learning_rate": 0.00019756915339480302,
"loss": 0.3571,
"step": 30
},
{
"epoch": 0.014668901927912825,
"grad_norm": 0.4928306043148041,
"learning_rate": 0.00019715004191114837,
"loss": 0.3467,
"step": 35
},
{
"epoch": 0.016764459346186086,
"grad_norm": 333.243896484375,
"learning_rate": 0.00019673093042749373,
"loss": 0.3173,
"step": 40
},
{
"epoch": 0.018860016764459347,
"grad_norm": 0.5272846817970276,
"learning_rate": 0.00019631181894383908,
"loss": 0.2952,
"step": 45
},
{
"epoch": 0.020955574182732608,
"grad_norm": 3.5901286602020264,
"learning_rate": 0.00019589270746018443,
"loss": 0.2935,
"step": 50
},
{
"epoch": 0.02305113160100587,
"grad_norm": 0.5013518929481506,
"learning_rate": 0.00019547359597652975,
"loss": 0.3024,
"step": 55
},
{
"epoch": 0.025146689019279127,
"grad_norm": 16.086902618408203,
"learning_rate": 0.0001950544844928751,
"loss": 0.2683,
"step": 60
},
{
"epoch": 0.027242246437552388,
"grad_norm": 0.30715975165367126,
"learning_rate": 0.00019463537300922046,
"loss": 0.2653,
"step": 65
},
{
"epoch": 0.02933780385582565,
"grad_norm": 0.33719402551651,
"learning_rate": 0.0001942162615255658,
"loss": 0.2575,
"step": 70
},
{
"epoch": 0.03143336127409891,
"grad_norm": 0.3756738305091858,
"learning_rate": 0.00019379715004191116,
"loss": 0.2449,
"step": 75
},
{
"epoch": 0.03352891869237217,
"grad_norm": 0.3635186553001404,
"learning_rate": 0.0001933780385582565,
"loss": 0.2538,
"step": 80
},
{
"epoch": 0.03562447611064543,
"grad_norm": 0.3560231328010559,
"learning_rate": 0.00019295892707460186,
"loss": 0.2435,
"step": 85
},
{
"epoch": 0.037720033528918694,
"grad_norm": 0.371216356754303,
"learning_rate": 0.00019253981559094722,
"loss": 0.2329,
"step": 90
},
{
"epoch": 0.039815590947191955,
"grad_norm": 0.45535510778427124,
"learning_rate": 0.00019212070410729254,
"loss": 0.2338,
"step": 95
},
{
"epoch": 0.041911148365465216,
"grad_norm": 15.063165664672852,
"learning_rate": 0.0001917015926236379,
"loss": 0.2353,
"step": 100
},
{
"epoch": 0.04400670578373848,
"grad_norm": 0.308463454246521,
"learning_rate": 0.00019128248113998324,
"loss": 0.2323,
"step": 105
},
{
"epoch": 0.04610226320201174,
"grad_norm": 0.29910165071487427,
"learning_rate": 0.0001908633696563286,
"loss": 0.2222,
"step": 110
},
{
"epoch": 0.04819782062028499,
"grad_norm": 0.3254060745239258,
"learning_rate": 0.00019044425817267395,
"loss": 0.216,
"step": 115
},
{
"epoch": 0.050293378038558254,
"grad_norm": 0.2623255252838135,
"learning_rate": 0.0001900251466890193,
"loss": 0.2136,
"step": 120
},
{
"epoch": 0.052388935456831515,
"grad_norm": 0.2627584636211395,
"learning_rate": 0.00018960603520536465,
"loss": 0.2118,
"step": 125
},
{
"epoch": 0.054484492875104776,
"grad_norm": 0.28595617413520813,
"learning_rate": 0.00018918692372170998,
"loss": 0.2088,
"step": 130
},
{
"epoch": 0.05658005029337804,
"grad_norm": 0.550282895565033,
"learning_rate": 0.00018876781223805533,
"loss": 0.2198,
"step": 135
},
{
"epoch": 0.0586756077116513,
"grad_norm": 0.296385258436203,
"learning_rate": 0.00018834870075440068,
"loss": 0.2219,
"step": 140
},
{
"epoch": 0.06077116512992456,
"grad_norm": 0.3291824460029602,
"learning_rate": 0.00018792958927074603,
"loss": 0.2059,
"step": 145
},
{
"epoch": 0.06286672254819782,
"grad_norm": 0.27933308482170105,
"learning_rate": 0.00018751047778709136,
"loss": 0.2001,
"step": 150
},
{
"epoch": 0.06496227996647108,
"grad_norm": 0.2848331332206726,
"learning_rate": 0.00018709136630343674,
"loss": 0.1999,
"step": 155
},
{
"epoch": 0.06705783738474434,
"grad_norm": 0.25534388422966003,
"learning_rate": 0.0001866722548197821,
"loss": 0.1977,
"step": 160
},
{
"epoch": 0.0691533948030176,
"grad_norm": 2.9550936222076416,
"learning_rate": 0.0001862531433361274,
"loss": 0.2004,
"step": 165
},
{
"epoch": 0.07124895222129086,
"grad_norm": 0.35537609457969666,
"learning_rate": 0.00018583403185247276,
"loss": 0.2131,
"step": 170
},
{
"epoch": 0.07334450963956413,
"grad_norm": 47.72193908691406,
"learning_rate": 0.00018541492036881812,
"loss": 0.2016,
"step": 175
},
{
"epoch": 0.07544006705783739,
"grad_norm": 0.3546775281429291,
"learning_rate": 0.00018499580888516347,
"loss": 0.1998,
"step": 180
},
{
"epoch": 0.07753562447611065,
"grad_norm": 0.2701320946216583,
"learning_rate": 0.0001845766974015088,
"loss": 0.1948,
"step": 185
},
{
"epoch": 0.07963118189438391,
"grad_norm": 0.2246120572090149,
"learning_rate": 0.00018415758591785414,
"loss": 0.1882,
"step": 190
},
{
"epoch": 0.08172673931265717,
"grad_norm": 0.27891814708709717,
"learning_rate": 0.00018373847443419952,
"loss": 0.1879,
"step": 195
},
{
"epoch": 0.08382229673093043,
"grad_norm": 0.4772103428840637,
"learning_rate": 0.00018331936295054485,
"loss": 0.2069,
"step": 200
},
{
"epoch": 0.08591785414920369,
"grad_norm": 2.367286205291748,
"learning_rate": 0.0001829002514668902,
"loss": 0.2043,
"step": 205
},
{
"epoch": 0.08801341156747695,
"grad_norm": 0.30191686749458313,
"learning_rate": 0.00018248113998323555,
"loss": 0.1978,
"step": 210
},
{
"epoch": 0.09010896898575022,
"grad_norm": 0.22913698852062225,
"learning_rate": 0.0001820620284995809,
"loss": 0.1937,
"step": 215
},
{
"epoch": 0.09220452640402348,
"grad_norm": 0.26689326763153076,
"learning_rate": 0.00018164291701592623,
"loss": 0.1892,
"step": 220
},
{
"epoch": 0.09430008382229674,
"grad_norm": 0.26441484689712524,
"learning_rate": 0.00018122380553227158,
"loss": 0.1924,
"step": 225
},
{
"epoch": 0.09639564124056998,
"grad_norm": 0.2602805197238922,
"learning_rate": 0.00018080469404861696,
"loss": 0.1856,
"step": 230
},
{
"epoch": 0.09849119865884325,
"grad_norm": 0.21042250096797943,
"learning_rate": 0.0001803855825649623,
"loss": 0.1848,
"step": 235
},
{
"epoch": 0.10058675607711651,
"grad_norm": 1.8752623796463013,
"learning_rate": 0.00017996647108130763,
"loss": 0.1822,
"step": 240
},
{
"epoch": 0.10268231349538977,
"grad_norm": 0.3879601061344147,
"learning_rate": 0.00017954735959765299,
"loss": 0.1979,
"step": 245
},
{
"epoch": 0.10477787091366303,
"grad_norm": 0.27347472310066223,
"learning_rate": 0.00017912824811399834,
"loss": 0.1903,
"step": 250
},
{
"epoch": 0.10687342833193629,
"grad_norm": 0.24279960989952087,
"learning_rate": 0.0001787091366303437,
"loss": 0.1848,
"step": 255
},
{
"epoch": 0.10896898575020955,
"grad_norm": 0.275285005569458,
"learning_rate": 0.00017829002514668901,
"loss": 0.1848,
"step": 260
},
{
"epoch": 0.11106454316848281,
"grad_norm": 0.33457309007644653,
"learning_rate": 0.00017787091366303437,
"loss": 0.1851,
"step": 265
},
{
"epoch": 0.11316010058675607,
"grad_norm": 0.5345426797866821,
"learning_rate": 0.00017745180217937974,
"loss": 0.189,
"step": 270
},
{
"epoch": 0.11525565800502934,
"grad_norm": 0.3533399701118469,
"learning_rate": 0.00017703269069572507,
"loss": 0.1835,
"step": 275
},
{
"epoch": 0.1173512154233026,
"grad_norm": 0.3257920742034912,
"learning_rate": 0.00017661357921207042,
"loss": 0.1892,
"step": 280
},
{
"epoch": 0.11944677284157586,
"grad_norm": 0.21177901327610016,
"learning_rate": 0.00017619446772841577,
"loss": 0.1818,
"step": 285
},
{
"epoch": 0.12154233025984912,
"grad_norm": 0.20798452198505402,
"learning_rate": 0.00017577535624476112,
"loss": 0.1792,
"step": 290
},
{
"epoch": 0.12363788767812238,
"grad_norm": 0.21963848173618317,
"learning_rate": 0.00017535624476110645,
"loss": 0.1776,
"step": 295
},
{
"epoch": 0.12573344509639564,
"grad_norm": 0.22277575731277466,
"learning_rate": 0.0001749371332774518,
"loss": 0.178,
"step": 300
},
{
"epoch": 0.12782900251466892,
"grad_norm": 0.22803856432437897,
"learning_rate": 0.00017451802179379715,
"loss": 0.1732,
"step": 305
},
{
"epoch": 0.12992455993294216,
"grad_norm": 0.20474207401275635,
"learning_rate": 0.0001740989103101425,
"loss": 0.173,
"step": 310
},
{
"epoch": 0.1320201173512154,
"grad_norm": 0.23283089697360992,
"learning_rate": 0.00017367979882648786,
"loss": 0.1752,
"step": 315
},
{
"epoch": 0.13411567476948869,
"grad_norm": 0.24834850430488586,
"learning_rate": 0.0001732606873428332,
"loss": 0.1727,
"step": 320
},
{
"epoch": 0.13621123218776193,
"grad_norm": 0.2943226993083954,
"learning_rate": 0.00017284157585917856,
"loss": 0.1758,
"step": 325
},
{
"epoch": 0.1383067896060352,
"grad_norm": 0.28577786684036255,
"learning_rate": 0.00017242246437552388,
"loss": 0.1835,
"step": 330
},
{
"epoch": 0.14040234702430845,
"grad_norm": 0.22449374198913574,
"learning_rate": 0.00017200335289186924,
"loss": 0.1795,
"step": 335
},
{
"epoch": 0.14249790444258173,
"grad_norm": 0.3676876425743103,
"learning_rate": 0.0001715842414082146,
"loss": 0.1771,
"step": 340
},
{
"epoch": 0.14459346186085498,
"grad_norm": 0.4172014594078064,
"learning_rate": 0.00017116512992455994,
"loss": 0.197,
"step": 345
},
{
"epoch": 0.14668901927912825,
"grad_norm": 0.26507022976875305,
"learning_rate": 0.0001707460184409053,
"loss": 0.181,
"step": 350
},
{
"epoch": 0.1487845766974015,
"grad_norm": 0.2308947890996933,
"learning_rate": 0.00017032690695725064,
"loss": 0.1767,
"step": 355
},
{
"epoch": 0.15088013411567477,
"grad_norm": 0.30771297216415405,
"learning_rate": 0.000169907795473596,
"loss": 0.1832,
"step": 360
},
{
"epoch": 0.15297569153394802,
"grad_norm": 0.23101018369197845,
"learning_rate": 0.00016948868398994132,
"loss": 0.1811,
"step": 365
},
{
"epoch": 0.1550712489522213,
"grad_norm": 0.20816421508789062,
"learning_rate": 0.00016906957250628667,
"loss": 0.1755,
"step": 370
},
{
"epoch": 0.15716680637049454,
"grad_norm": 0.2107459455728531,
"learning_rate": 0.00016865046102263202,
"loss": 0.1726,
"step": 375
},
{
"epoch": 0.15926236378876782,
"grad_norm": 2.224956750869751,
"learning_rate": 0.00016823134953897737,
"loss": 0.1718,
"step": 380
},
{
"epoch": 0.16135792120704107,
"grad_norm": 0.22022977471351624,
"learning_rate": 0.00016781223805532273,
"loss": 0.1725,
"step": 385
},
{
"epoch": 0.16345347862531434,
"grad_norm": 0.1845213919878006,
"learning_rate": 0.00016739312657166808,
"loss": 0.1677,
"step": 390
},
{
"epoch": 0.1655490360435876,
"grad_norm": 0.21298770606517792,
"learning_rate": 0.00016697401508801343,
"loss": 0.1734,
"step": 395
},
{
"epoch": 0.16764459346186086,
"grad_norm": 0.20876356959342957,
"learning_rate": 0.00016655490360435878,
"loss": 0.1761,
"step": 400
},
{
"epoch": 0.1697401508801341,
"grad_norm": 0.21667592227458954,
"learning_rate": 0.0001661357921207041,
"loss": 0.1668,
"step": 405
},
{
"epoch": 0.17183570829840739,
"grad_norm": 2.4301416873931885,
"learning_rate": 0.00016571668063704946,
"loss": 0.1697,
"step": 410
},
{
"epoch": 0.17393126571668063,
"grad_norm": 0.43861570954322815,
"learning_rate": 0.0001652975691533948,
"loss": 0.1804,
"step": 415
},
{
"epoch": 0.1760268231349539,
"grad_norm": 0.2521713078022003,
"learning_rate": 0.00016487845766974016,
"loss": 0.1762,
"step": 420
},
{
"epoch": 0.17812238055322716,
"grad_norm": 0.20923659205436707,
"learning_rate": 0.00016445934618608551,
"loss": 0.1695,
"step": 425
},
{
"epoch": 0.18021793797150043,
"grad_norm": 0.20591603219509125,
"learning_rate": 0.00016404023470243087,
"loss": 0.1729,
"step": 430
},
{
"epoch": 0.18231349538977368,
"grad_norm": 2.5188469886779785,
"learning_rate": 0.00016362112321877622,
"loss": 0.1723,
"step": 435
},
{
"epoch": 0.18440905280804695,
"grad_norm": 0.2523597478866577,
"learning_rate": 0.00016320201173512154,
"loss": 0.174,
"step": 440
},
{
"epoch": 0.1865046102263202,
"grad_norm": 0.23274292051792145,
"learning_rate": 0.0001627829002514669,
"loss": 0.1707,
"step": 445
},
{
"epoch": 0.18860016764459347,
"grad_norm": 0.26748543977737427,
"learning_rate": 0.00016236378876781225,
"loss": 0.1686,
"step": 450
},
{
"epoch": 0.19069572506286672,
"grad_norm": 0.2876422107219696,
"learning_rate": 0.0001619446772841576,
"loss": 0.1723,
"step": 455
},
{
"epoch": 0.19279128248113997,
"grad_norm": 0.5347093939781189,
"learning_rate": 0.00016152556580050292,
"loss": 0.1873,
"step": 460
},
{
"epoch": 0.19488683989941324,
"grad_norm": 0.2463475912809372,
"learning_rate": 0.0001611064543168483,
"loss": 0.1817,
"step": 465
},
{
"epoch": 0.1969823973176865,
"grad_norm": 0.21807396411895752,
"learning_rate": 0.00016068734283319365,
"loss": 0.1751,
"step": 470
},
{
"epoch": 0.19907795473595977,
"grad_norm": 0.1950008124113083,
"learning_rate": 0.00016026823134953898,
"loss": 0.1723,
"step": 475
},
{
"epoch": 0.20117351215423301,
"grad_norm": 0.19139733910560608,
"learning_rate": 0.00015984911986588433,
"loss": 0.168,
"step": 480
},
{
"epoch": 0.2032690695725063,
"grad_norm": 0.24496974050998688,
"learning_rate": 0.00015943000838222968,
"loss": 0.1713,
"step": 485
},
{
"epoch": 0.20536462699077954,
"grad_norm": 0.2883533239364624,
"learning_rate": 0.00015901089689857503,
"loss": 0.1758,
"step": 490
},
{
"epoch": 0.2074601844090528,
"grad_norm": 0.2685905694961548,
"learning_rate": 0.00015859178541492036,
"loss": 0.1789,
"step": 495
},
{
"epoch": 0.20955574182732606,
"grad_norm": 0.22861076891422272,
"learning_rate": 0.0001581726739312657,
"loss": 0.1751,
"step": 500
},
{
"epoch": 0.21165129924559933,
"grad_norm": 0.4216708242893219,
"learning_rate": 0.0001577535624476111,
"loss": 0.1765,
"step": 505
},
{
"epoch": 0.21374685666387258,
"grad_norm": 0.33128198981285095,
"learning_rate": 0.0001573344509639564,
"loss": 0.1908,
"step": 510
},
{
"epoch": 0.21584241408214586,
"grad_norm": 0.21954227983951569,
"learning_rate": 0.00015691533948030176,
"loss": 0.1707,
"step": 515
},
{
"epoch": 0.2179379715004191,
"grad_norm": 0.22218674421310425,
"learning_rate": 0.00015649622799664712,
"loss": 0.1695,
"step": 520
},
{
"epoch": 0.22003352891869238,
"grad_norm": 0.23518136143684387,
"learning_rate": 0.00015607711651299247,
"loss": 0.1677,
"step": 525
},
{
"epoch": 0.22212908633696563,
"grad_norm": 0.4071219861507416,
"learning_rate": 0.0001556580050293378,
"loss": 0.1753,
"step": 530
},
{
"epoch": 0.2242246437552389,
"grad_norm": 0.25883597135543823,
"learning_rate": 0.00015523889354568314,
"loss": 0.1741,
"step": 535
},
{
"epoch": 0.22632020117351215,
"grad_norm": 0.19170518219470978,
"learning_rate": 0.00015481978206202852,
"loss": 0.17,
"step": 540
},
{
"epoch": 0.22841575859178542,
"grad_norm": 0.18559418618679047,
"learning_rate": 0.00015440067057837387,
"loss": 0.1701,
"step": 545
},
{
"epoch": 0.23051131601005867,
"grad_norm": 0.2035888284444809,
"learning_rate": 0.0001539815590947192,
"loss": 0.1623,
"step": 550
},
{
"epoch": 0.23260687342833195,
"grad_norm": 7.4185709953308105,
"learning_rate": 0.00015356244761106455,
"loss": 0.1654,
"step": 555
},
{
"epoch": 0.2347024308466052,
"grad_norm": 0.21483619511127472,
"learning_rate": 0.0001531433361274099,
"loss": 0.1664,
"step": 560
},
{
"epoch": 0.23679798826487847,
"grad_norm": 0.2908990979194641,
"learning_rate": 0.00015272422464375525,
"loss": 0.1733,
"step": 565
},
{
"epoch": 0.23889354568315171,
"grad_norm": 0.34150460362434387,
"learning_rate": 0.00015230511316010058,
"loss": 0.1721,
"step": 570
},
{
"epoch": 0.240989103101425,
"grad_norm": 0.26667365431785583,
"learning_rate": 0.00015188600167644593,
"loss": 0.1778,
"step": 575
},
{
"epoch": 0.24308466051969824,
"grad_norm": 0.2024029940366745,
"learning_rate": 0.0001514668901927913,
"loss": 0.1681,
"step": 580
},
{
"epoch": 0.2451802179379715,
"grad_norm": 0.1918814480304718,
"learning_rate": 0.00015104777870913663,
"loss": 0.1672,
"step": 585
},
{
"epoch": 0.24727577535624476,
"grad_norm": 0.22173915803432465,
"learning_rate": 0.00015062866722548199,
"loss": 0.1629,
"step": 590
},
{
"epoch": 0.24937133277451803,
"grad_norm": 0.20861753821372986,
"learning_rate": 0.00015020955574182734,
"loss": 0.1695,
"step": 595
},
{
"epoch": 0.2514668901927913,
"grad_norm": 0.2476891726255417,
"learning_rate": 0.0001497904442581727,
"loss": 0.1687,
"step": 600
},
{
"epoch": 0.25356244761106456,
"grad_norm": 0.23498353362083435,
"learning_rate": 0.00014937133277451801,
"loss": 0.1721,
"step": 605
},
{
"epoch": 0.25565800502933783,
"grad_norm": 0.24014067649841309,
"learning_rate": 0.00014895222129086337,
"loss": 0.1643,
"step": 610
},
{
"epoch": 0.25775356244761105,
"grad_norm": 1.1056888103485107,
"learning_rate": 0.00014853310980720872,
"loss": 0.1652,
"step": 615
},
{
"epoch": 0.2598491198658843,
"grad_norm": 0.3467954397201538,
"learning_rate": 0.00014811399832355407,
"loss": 0.1741,
"step": 620
},
{
"epoch": 0.2619446772841576,
"grad_norm": 0.29533931612968445,
"learning_rate": 0.00014769488683989942,
"loss": 0.1768,
"step": 625
},
{
"epoch": 0.2640402347024308,
"grad_norm": 0.2000960111618042,
"learning_rate": 0.00014727577535624477,
"loss": 0.1662,
"step": 630
},
{
"epoch": 0.2661357921207041,
"grad_norm": 0.18396534025669098,
"learning_rate": 0.00014685666387259013,
"loss": 0.166,
"step": 635
},
{
"epoch": 0.26823134953897737,
"grad_norm": 0.18797029554843903,
"learning_rate": 0.00014643755238893545,
"loss": 0.1624,
"step": 640
},
{
"epoch": 0.27032690695725065,
"grad_norm": 0.1914837509393692,
"learning_rate": 0.0001460184409052808,
"loss": 0.1625,
"step": 645
},
{
"epoch": 0.27242246437552387,
"grad_norm": 0.18006779253482819,
"learning_rate": 0.00014559932942162615,
"loss": 0.1602,
"step": 650
},
{
"epoch": 0.27451802179379714,
"grad_norm": 0.19413350522518158,
"learning_rate": 0.0001451802179379715,
"loss": 0.1632,
"step": 655
},
{
"epoch": 0.2766135792120704,
"grad_norm": 0.19674436748027802,
"learning_rate": 0.00014476110645431686,
"loss": 0.1593,
"step": 660
},
{
"epoch": 0.2787091366303437,
"grad_norm": 0.16716861724853516,
"learning_rate": 0.0001443419949706622,
"loss": 0.161,
"step": 665
},
{
"epoch": 0.2808046940486169,
"grad_norm": 0.1591351330280304,
"learning_rate": 0.00014392288348700756,
"loss": 0.163,
"step": 670
},
{
"epoch": 0.2829002514668902,
"grad_norm": 0.16973218321800232,
"learning_rate": 0.0001435037720033529,
"loss": 0.1665,
"step": 675
},
{
"epoch": 0.28499580888516346,
"grad_norm": 0.16811800003051758,
"learning_rate": 0.00014308466051969824,
"loss": 0.1586,
"step": 680
},
{
"epoch": 0.28709136630343673,
"grad_norm": 0.21386906504631042,
"learning_rate": 0.0001426655490360436,
"loss": 0.1589,
"step": 685
},
{
"epoch": 0.28918692372170995,
"grad_norm": 0.17704640328884125,
"learning_rate": 0.00014224643755238894,
"loss": 0.1597,
"step": 690
},
{
"epoch": 0.29128248113998323,
"grad_norm": 0.1727607399225235,
"learning_rate": 0.0001418273260687343,
"loss": 0.1565,
"step": 695
},
{
"epoch": 0.2933780385582565,
"grad_norm": 0.18027518689632416,
"learning_rate": 0.00014140821458507964,
"loss": 0.1597,
"step": 700
},
{
"epoch": 0.2954735959765298,
"grad_norm": 0.18142397701740265,
"learning_rate": 0.000140989103101425,
"loss": 0.1618,
"step": 705
},
{
"epoch": 0.297569153394803,
"grad_norm": 0.18902671337127686,
"learning_rate": 0.00014056999161777035,
"loss": 0.1606,
"step": 710
},
{
"epoch": 0.2996647108130763,
"grad_norm": 0.2578323185443878,
"learning_rate": 0.00014015088013411567,
"loss": 0.1658,
"step": 715
},
{
"epoch": 0.30176026823134955,
"grad_norm": 0.2432331144809723,
"learning_rate": 0.00013973176865046102,
"loss": 0.1694,
"step": 720
},
{
"epoch": 0.3038558256496228,
"grad_norm": 0.1846640259027481,
"learning_rate": 0.00013931265716680638,
"loss": 0.1655,
"step": 725
},
{
"epoch": 0.30595138306789604,
"grad_norm": 0.1582545042037964,
"learning_rate": 0.00013889354568315173,
"loss": 0.1639,
"step": 730
},
{
"epoch": 0.3080469404861693,
"grad_norm": 0.16749900579452515,
"learning_rate": 0.00013847443419949708,
"loss": 0.1596,
"step": 735
},
{
"epoch": 0.3101424979044426,
"grad_norm": 0.17118974030017853,
"learning_rate": 0.00013805532271584243,
"loss": 0.1569,
"step": 740
},
{
"epoch": 0.31223805532271587,
"grad_norm": 0.16258081793785095,
"learning_rate": 0.00013763621123218778,
"loss": 0.1599,
"step": 745
},
{
"epoch": 0.3143336127409891,
"grad_norm": 0.17213059961795807,
"learning_rate": 0.0001372170997485331,
"loss": 0.1588,
"step": 750
},
{
"epoch": 0.31642917015926236,
"grad_norm": 0.18207783997058868,
"learning_rate": 0.00013679798826487846,
"loss": 0.1571,
"step": 755
},
{
"epoch": 0.31852472757753564,
"grad_norm": 0.19723886251449585,
"learning_rate": 0.0001363788767812238,
"loss": 0.1552,
"step": 760
},
{
"epoch": 0.32062028499580886,
"grad_norm": 0.17248809337615967,
"learning_rate": 0.00013595976529756916,
"loss": 0.1564,
"step": 765
},
{
"epoch": 0.32271584241408213,
"grad_norm": 0.18213661015033722,
"learning_rate": 0.0001355406538139145,
"loss": 0.1597,
"step": 770
},
{
"epoch": 0.3248113998323554,
"grad_norm": 0.1856526881456375,
"learning_rate": 0.00013512154233025987,
"loss": 0.1609,
"step": 775
},
{
"epoch": 0.3269069572506287,
"grad_norm": 0.16957145929336548,
"learning_rate": 0.00013470243084660522,
"loss": 0.1595,
"step": 780
},
{
"epoch": 0.3290025146689019,
"grad_norm": 0.17088639736175537,
"learning_rate": 0.00013428331936295054,
"loss": 0.161,
"step": 785
},
{
"epoch": 0.3310980720871752,
"grad_norm": 1.320410132408142,
"learning_rate": 0.0001338642078792959,
"loss": 0.165,
"step": 790
},
{
"epoch": 0.33319362950544845,
"grad_norm": 0.19440437853336334,
"learning_rate": 0.00013344509639564125,
"loss": 0.1602,
"step": 795
},
{
"epoch": 0.3352891869237217,
"grad_norm": 0.19590894877910614,
"learning_rate": 0.0001330259849119866,
"loss": 0.1616,
"step": 800
},
{
"epoch": 0.33738474434199495,
"grad_norm": 0.19055521488189697,
"learning_rate": 0.00013260687342833192,
"loss": 0.1638,
"step": 805
},
{
"epoch": 0.3394803017602682,
"grad_norm": 0.1577232927083969,
"learning_rate": 0.00013218776194467727,
"loss": 0.1582,
"step": 810
},
{
"epoch": 0.3415758591785415,
"grad_norm": 0.16225744783878326,
"learning_rate": 0.00013176865046102265,
"loss": 0.1555,
"step": 815
},
{
"epoch": 0.34367141659681477,
"grad_norm": 0.1565002202987671,
"learning_rate": 0.000131349538977368,
"loss": 0.1567,
"step": 820
},
{
"epoch": 0.345766974015088,
"grad_norm": 0.18790805339813232,
"learning_rate": 0.00013093042749371333,
"loss": 0.1581,
"step": 825
},
{
"epoch": 0.34786253143336127,
"grad_norm": 0.18632063269615173,
"learning_rate": 0.00013051131601005868,
"loss": 0.1621,
"step": 830
},
{
"epoch": 0.34995808885163454,
"grad_norm": 0.17684835195541382,
"learning_rate": 0.00013009220452640403,
"loss": 0.1591,
"step": 835
},
{
"epoch": 0.3520536462699078,
"grad_norm": 0.16213147342205048,
"learning_rate": 0.00012967309304274938,
"loss": 0.1593,
"step": 840
},
{
"epoch": 0.35414920368818104,
"grad_norm": 0.1656450480222702,
"learning_rate": 0.0001292539815590947,
"loss": 0.1617,
"step": 845
},
{
"epoch": 0.3562447611064543,
"grad_norm": 0.16887331008911133,
"learning_rate": 0.0001288348700754401,
"loss": 0.1601,
"step": 850
},
{
"epoch": 0.3583403185247276,
"grad_norm": 0.17660242319107056,
"learning_rate": 0.00012841575859178544,
"loss": 0.1575,
"step": 855
},
{
"epoch": 0.36043587594300086,
"grad_norm": 0.17625996470451355,
"learning_rate": 0.00012799664710813076,
"loss": 0.1597,
"step": 860
},
{
"epoch": 0.3625314333612741,
"grad_norm": 0.15301348268985748,
"learning_rate": 0.00012757753562447612,
"loss": 0.1597,
"step": 865
},
{
"epoch": 0.36462699077954736,
"grad_norm": 0.15253449976444244,
"learning_rate": 0.00012715842414082147,
"loss": 0.157,
"step": 870
},
{
"epoch": 0.36672254819782063,
"grad_norm": 0.16985070705413818,
"learning_rate": 0.00012673931265716682,
"loss": 0.1558,
"step": 875
},
{
"epoch": 0.3688181056160939,
"grad_norm": 0.1545540988445282,
"learning_rate": 0.00012632020117351214,
"loss": 0.1596,
"step": 880
},
{
"epoch": 0.3709136630343671,
"grad_norm": 0.1766255795955658,
"learning_rate": 0.0001259010896898575,
"loss": 0.1537,
"step": 885
},
{
"epoch": 0.3730092204526404,
"grad_norm": 0.15094798803329468,
"learning_rate": 0.00012548197820620288,
"loss": 0.1562,
"step": 890
},
{
"epoch": 0.3751047778709137,
"grad_norm": 0.16060136258602142,
"learning_rate": 0.0001250628667225482,
"loss": 0.158,
"step": 895
},
{
"epoch": 0.37720033528918695,
"grad_norm": 0.17947901785373688,
"learning_rate": 0.00012464375523889355,
"loss": 0.1583,
"step": 900
},
{
"epoch": 0.37929589270746017,
"grad_norm": 0.15727129578590393,
"learning_rate": 0.0001242246437552389,
"loss": 0.1529,
"step": 905
},
{
"epoch": 0.38139145012573344,
"grad_norm": 0.14456555247306824,
"learning_rate": 0.00012380553227158426,
"loss": 0.1573,
"step": 910
},
{
"epoch": 0.3834870075440067,
"grad_norm": 0.15629969537258148,
"learning_rate": 0.00012338642078792958,
"loss": 0.1607,
"step": 915
},
{
"epoch": 0.38558256496227994,
"grad_norm": 0.1958230435848236,
"learning_rate": 0.00012296730930427493,
"loss": 0.1572,
"step": 920
},
{
"epoch": 0.3876781223805532,
"grad_norm": 0.18159349262714386,
"learning_rate": 0.00012254819782062028,
"loss": 0.1565,
"step": 925
},
{
"epoch": 0.3897736797988265,
"grad_norm": 0.19391848146915436,
"learning_rate": 0.00012212908633696564,
"loss": 0.1552,
"step": 930
},
{
"epoch": 0.39186923721709976,
"grad_norm": 0.18596895039081573,
"learning_rate": 0.00012170997485331099,
"loss": 0.1623,
"step": 935
},
{
"epoch": 0.393964794635373,
"grad_norm": 0.175604447722435,
"learning_rate": 0.00012129086336965634,
"loss": 0.1593,
"step": 940
},
{
"epoch": 0.39606035205364626,
"grad_norm": 0.18805819749832153,
"learning_rate": 0.00012087175188600168,
"loss": 0.1603,
"step": 945
},
{
"epoch": 0.39815590947191953,
"grad_norm": 0.15223102271556854,
"learning_rate": 0.00012045264040234703,
"loss": 0.1568,
"step": 950
},
{
"epoch": 0.4002514668901928,
"grad_norm": 0.13771827518939972,
"learning_rate": 0.00012003352891869237,
"loss": 0.1553,
"step": 955
},
{
"epoch": 0.40234702430846603,
"grad_norm": 0.1633366495370865,
"learning_rate": 0.00011961441743503772,
"loss": 0.156,
"step": 960
},
{
"epoch": 0.4044425817267393,
"grad_norm": 0.17143379151821136,
"learning_rate": 0.00011919530595138306,
"loss": 0.1542,
"step": 965
},
{
"epoch": 0.4065381391450126,
"grad_norm": 0.16767437756061554,
"learning_rate": 0.00011877619446772844,
"loss": 0.1572,
"step": 970
},
{
"epoch": 0.40863369656328585,
"grad_norm": 0.16299773752689362,
"learning_rate": 0.00011835708298407377,
"loss": 0.1583,
"step": 975
},
{
"epoch": 0.4107292539815591,
"grad_norm": 0.15164397656917572,
"learning_rate": 0.00011793797150041913,
"loss": 0.1526,
"step": 980
},
{
"epoch": 0.41282481139983235,
"grad_norm": 0.1567896157503128,
"learning_rate": 0.00011751886001676446,
"loss": 0.154,
"step": 985
},
{
"epoch": 0.4149203688181056,
"grad_norm": 0.15943744778633118,
"learning_rate": 0.00011709974853310982,
"loss": 0.1533,
"step": 990
},
{
"epoch": 0.4170159262363789,
"grad_norm": 0.16572755575180054,
"learning_rate": 0.00011668063704945515,
"loss": 0.1559,
"step": 995
},
{
"epoch": 0.4191114836546521,
"grad_norm": 0.15907670557498932,
"learning_rate": 0.0001162615255658005,
"loss": 0.1558,
"step": 1000
},
{
"epoch": 0.4212070410729254,
"grad_norm": 0.1808643937110901,
"learning_rate": 0.00011584241408214587,
"loss": 0.1555,
"step": 1005
},
{
"epoch": 0.42330259849119867,
"grad_norm": 0.23819687962532043,
"learning_rate": 0.00011542330259849121,
"loss": 0.1579,
"step": 1010
},
{
"epoch": 0.42539815590947194,
"grad_norm": 0.24894015491008759,
"learning_rate": 0.00011500419111483656,
"loss": 0.1637,
"step": 1015
},
{
"epoch": 0.42749371332774516,
"grad_norm": 0.18128713965415955,
"learning_rate": 0.0001145850796311819,
"loss": 0.1605,
"step": 1020
},
{
"epoch": 0.42958927074601844,
"grad_norm": 0.1779192090034485,
"learning_rate": 0.00011416596814752725,
"loss": 0.1586,
"step": 1025
},
{
"epoch": 0.4316848281642917,
"grad_norm": 0.1617233157157898,
"learning_rate": 0.00011374685666387259,
"loss": 0.1548,
"step": 1030
},
{
"epoch": 0.433780385582565,
"grad_norm": 0.146457701921463,
"learning_rate": 0.00011332774518021794,
"loss": 0.1581,
"step": 1035
},
{
"epoch": 0.4358759430008382,
"grad_norm": 18.378135681152344,
"learning_rate": 0.00011290863369656328,
"loss": 0.1546,
"step": 1040
},
{
"epoch": 0.4379715004191115,
"grad_norm": 0.16601739823818207,
"learning_rate": 0.00011248952221290864,
"loss": 0.1573,
"step": 1045
},
{
"epoch": 0.44006705783738476,
"grad_norm": 0.18232333660125732,
"learning_rate": 0.000112070410729254,
"loss": 0.1546,
"step": 1050
},
{
"epoch": 0.44216261525565803,
"grad_norm": 0.15297749638557434,
"learning_rate": 0.00011165129924559933,
"loss": 0.1561,
"step": 1055
},
{
"epoch": 0.44425817267393125,
"grad_norm": 7.5409440994262695,
"learning_rate": 0.00011123218776194469,
"loss": 0.1614,
"step": 1060
},
{
"epoch": 0.4463537300922045,
"grad_norm": 0.1807660162448883,
"learning_rate": 0.00011081307627829002,
"loss": 0.1624,
"step": 1065
},
{
"epoch": 0.4484492875104778,
"grad_norm": 0.18783360719680786,
"learning_rate": 0.00011039396479463538,
"loss": 0.1609,
"step": 1070
},
{
"epoch": 0.450544844928751,
"grad_norm": 0.23045672476291656,
"learning_rate": 0.00010997485331098071,
"loss": 0.1618,
"step": 1075
},
{
"epoch": 0.4526404023470243,
"grad_norm": 0.2051040381193161,
"learning_rate": 0.00010955574182732607,
"loss": 0.164,
"step": 1080
},
{
"epoch": 0.45473595976529757,
"grad_norm": 0.17596812546253204,
"learning_rate": 0.00010913663034367143,
"loss": 0.1578,
"step": 1085
},
{
"epoch": 0.45683151718357085,
"grad_norm": 0.1564697027206421,
"learning_rate": 0.00010871751886001677,
"loss": 0.1549,
"step": 1090
},
{
"epoch": 0.45892707460184406,
"grad_norm": 0.2508351802825928,
"learning_rate": 0.00010829840737636212,
"loss": 0.1541,
"step": 1095
},
{
"epoch": 0.46102263202011734,
"grad_norm": 0.17002500593662262,
"learning_rate": 0.00010787929589270746,
"loss": 0.1606,
"step": 1100
},
{
"epoch": 0.4631181894383906,
"grad_norm": 0.1768285632133484,
"learning_rate": 0.00010746018440905281,
"loss": 0.1533,
"step": 1105
},
{
"epoch": 0.4652137468566639,
"grad_norm": 0.21098843216896057,
"learning_rate": 0.00010704107292539815,
"loss": 0.1598,
"step": 1110
},
{
"epoch": 0.4673093042749371,
"grad_norm": 0.1743684709072113,
"learning_rate": 0.0001066219614417435,
"loss": 0.1577,
"step": 1115
},
{
"epoch": 0.4694048616932104,
"grad_norm": 0.18270978331565857,
"learning_rate": 0.00010620284995808884,
"loss": 0.1533,
"step": 1120
},
{
"epoch": 0.47150041911148366,
"grad_norm": 0.19265097379684448,
"learning_rate": 0.00010578373847443422,
"loss": 0.1545,
"step": 1125
},
{
"epoch": 0.47359597652975693,
"grad_norm": 0.19420358538627625,
"learning_rate": 0.00010536462699077956,
"loss": 0.1587,
"step": 1130
},
{
"epoch": 0.47569153394803015,
"grad_norm": 0.1721310317516327,
"learning_rate": 0.00010494551550712491,
"loss": 0.1577,
"step": 1135
},
{
"epoch": 0.47778709136630343,
"grad_norm": 0.1586717814207077,
"learning_rate": 0.00010452640402347025,
"loss": 0.1551,
"step": 1140
},
{
"epoch": 0.4798826487845767,
"grad_norm": 1.5393892526626587,
"learning_rate": 0.0001041072925398156,
"loss": 0.1537,
"step": 1145
},
{
"epoch": 0.48197820620285,
"grad_norm": 0.17265351116657257,
"learning_rate": 0.00010368818105616094,
"loss": 0.1514,
"step": 1150
},
{
"epoch": 0.4840737636211232,
"grad_norm": 0.15501521527767181,
"learning_rate": 0.00010326906957250629,
"loss": 0.1559,
"step": 1155
},
{
"epoch": 0.4861693210393965,
"grad_norm": 0.1425572782754898,
"learning_rate": 0.00010284995808885165,
"loss": 0.1532,
"step": 1160
},
{
"epoch": 0.48826487845766975,
"grad_norm": 0.1547863483428955,
"learning_rate": 0.00010243084660519699,
"loss": 0.1557,
"step": 1165
},
{
"epoch": 0.490360435875943,
"grad_norm": 0.1579703986644745,
"learning_rate": 0.00010201173512154234,
"loss": 0.1568,
"step": 1170
},
{
"epoch": 0.49245599329421624,
"grad_norm": 4.8748297691345215,
"learning_rate": 0.00010159262363788768,
"loss": 0.1542,
"step": 1175
},
{
"epoch": 0.4945515507124895,
"grad_norm": 0.16955487430095673,
"learning_rate": 0.00010117351215423303,
"loss": 0.1573,
"step": 1180
},
{
"epoch": 0.4966471081307628,
"grad_norm": 0.17641142010688782,
"learning_rate": 0.00010075440067057837,
"loss": 0.1563,
"step": 1185
},
{
"epoch": 0.49874266554903607,
"grad_norm": 0.19048817455768585,
"learning_rate": 0.00010033528918692372,
"loss": 0.1565,
"step": 1190
},
{
"epoch": 0.5008382229673093,
"grad_norm": 0.1500770002603531,
"learning_rate": 9.991617770326908e-05,
"loss": 0.1561,
"step": 1195
},
{
"epoch": 0.5029337803855826,
"grad_norm": 0.1577143371105194,
"learning_rate": 9.949706621961443e-05,
"loss": 0.1555,
"step": 1200
},
{
"epoch": 0.5050293378038558,
"grad_norm": 0.14767815172672272,
"learning_rate": 9.907795473595977e-05,
"loss": 0.1546,
"step": 1205
},
{
"epoch": 0.5071248952221291,
"grad_norm": 0.1728920042514801,
"learning_rate": 9.865884325230512e-05,
"loss": 0.1582,
"step": 1210
},
{
"epoch": 0.5092204526404024,
"grad_norm": 0.13907761871814728,
"learning_rate": 9.823973176865047e-05,
"loss": 0.154,
"step": 1215
},
{
"epoch": 0.5113160100586757,
"grad_norm": 0.14644944667816162,
"learning_rate": 9.782062028499581e-05,
"loss": 0.151,
"step": 1220
},
{
"epoch": 0.5134115674769488,
"grad_norm": 0.1749754697084427,
"learning_rate": 9.740150880134116e-05,
"loss": 0.1558,
"step": 1225
},
{
"epoch": 0.5155071248952221,
"grad_norm": 1.7467832565307617,
"learning_rate": 9.698239731768651e-05,
"loss": 0.154,
"step": 1230
},
{
"epoch": 0.5176026823134954,
"grad_norm": 0.3709283471107483,
"learning_rate": 9.656328583403186e-05,
"loss": 0.1648,
"step": 1235
},
{
"epoch": 0.5196982397317687,
"grad_norm": 0.28035739064216614,
"learning_rate": 9.61441743503772e-05,
"loss": 0.1697,
"step": 1240
},
{
"epoch": 0.5217937971500419,
"grad_norm": 0.1671634167432785,
"learning_rate": 9.572506286672255e-05,
"loss": 0.1611,
"step": 1245
},
{
"epoch": 0.5238893545683152,
"grad_norm": 0.15249969065189362,
"learning_rate": 9.53059513830679e-05,
"loss": 0.1557,
"step": 1250
},
{
"epoch": 0.5259849119865885,
"grad_norm": 0.14879916608333588,
"learning_rate": 9.488683989941326e-05,
"loss": 0.1516,
"step": 1255
},
{
"epoch": 0.5280804694048616,
"grad_norm": 2.7410941123962402,
"learning_rate": 9.44677284157586e-05,
"loss": 0.1497,
"step": 1260
},
{
"epoch": 0.5301760268231349,
"grad_norm": 0.188863143324852,
"learning_rate": 9.404861693210395e-05,
"loss": 0.1567,
"step": 1265
},
{
"epoch": 0.5322715842414082,
"grad_norm": 0.22812993824481964,
"learning_rate": 9.36295054484493e-05,
"loss": 0.1586,
"step": 1270
},
{
"epoch": 0.5343671416596815,
"grad_norm": 0.23461438715457916,
"learning_rate": 9.321039396479464e-05,
"loss": 0.1593,
"step": 1275
},
{
"epoch": 0.5364626990779547,
"grad_norm": 0.23611848056316376,
"learning_rate": 9.279128248113999e-05,
"loss": 0.1575,
"step": 1280
},
{
"epoch": 0.538558256496228,
"grad_norm": 0.20142588019371033,
"learning_rate": 9.237217099748533e-05,
"loss": 0.1611,
"step": 1285
},
{
"epoch": 0.5406538139145013,
"grad_norm": 0.18650312721729279,
"learning_rate": 9.195305951383069e-05,
"loss": 0.1585,
"step": 1290
},
{
"epoch": 0.5427493713327746,
"grad_norm": 0.1717916578054428,
"learning_rate": 9.153394803017603e-05,
"loss": 0.1566,
"step": 1295
},
{
"epoch": 0.5448449287510477,
"grad_norm": 0.154686838388443,
"learning_rate": 9.111483654652138e-05,
"loss": 0.154,
"step": 1300
},
{
"epoch": 0.546940486169321,
"grad_norm": 0.17387616634368896,
"learning_rate": 9.069572506286673e-05,
"loss": 0.1574,
"step": 1305
},
{
"epoch": 0.5490360435875943,
"grad_norm": 0.1707635372877121,
"learning_rate": 9.027661357921207e-05,
"loss": 0.1595,
"step": 1310
},
{
"epoch": 0.5511316010058676,
"grad_norm": 0.1505287140607834,
"learning_rate": 8.985750209555742e-05,
"loss": 0.1572,
"step": 1315
},
{
"epoch": 0.5532271584241408,
"grad_norm": 0.17044112086296082,
"learning_rate": 8.943839061190276e-05,
"loss": 0.1526,
"step": 1320
},
{
"epoch": 0.5553227158424141,
"grad_norm": 0.20292381942272186,
"learning_rate": 8.901927912824813e-05,
"loss": 0.157,
"step": 1325
},
{
"epoch": 0.5574182732606874,
"grad_norm": 0.22127684950828552,
"learning_rate": 8.860016764459346e-05,
"loss": 0.1569,
"step": 1330
},
{
"epoch": 0.5595138306789607,
"grad_norm": 0.169328510761261,
"learning_rate": 8.818105616093882e-05,
"loss": 0.1574,
"step": 1335
},
{
"epoch": 0.5616093880972338,
"grad_norm": 0.1888064444065094,
"learning_rate": 8.776194467728415e-05,
"loss": 0.152,
"step": 1340
},
{
"epoch": 0.5637049455155071,
"grad_norm": 2.7211835384368896,
"learning_rate": 8.734283319362952e-05,
"loss": 0.1557,
"step": 1345
},
{
"epoch": 0.5658005029337804,
"grad_norm": 0.18435344099998474,
"learning_rate": 8.692372170997486e-05,
"loss": 0.1527,
"step": 1350
},
{
"epoch": 0.5678960603520536,
"grad_norm": 0.2031932920217514,
"learning_rate": 8.650461022632021e-05,
"loss": 0.1573,
"step": 1355
},
{
"epoch": 0.5699916177703269,
"grad_norm": 0.1941346377134323,
"learning_rate": 8.608549874266555e-05,
"loss": 0.1538,
"step": 1360
},
{
"epoch": 0.5720871751886002,
"grad_norm": 0.18645240366458893,
"learning_rate": 8.56663872590109e-05,
"loss": 0.1542,
"step": 1365
},
{
"epoch": 0.5741827326068735,
"grad_norm": 0.1776381880044937,
"learning_rate": 8.524727577535625e-05,
"loss": 0.1567,
"step": 1370
},
{
"epoch": 0.5762782900251467,
"grad_norm": 0.17730367183685303,
"learning_rate": 8.482816429170159e-05,
"loss": 0.1552,
"step": 1375
},
{
"epoch": 0.5783738474434199,
"grad_norm": 0.1614091843366623,
"learning_rate": 8.440905280804694e-05,
"loss": 0.1545,
"step": 1380
},
{
"epoch": 0.5804694048616932,
"grad_norm": 0.2010612040758133,
"learning_rate": 8.398994132439229e-05,
"loss": 0.1525,
"step": 1385
},
{
"epoch": 0.5825649622799665,
"grad_norm": 0.29623404145240784,
"learning_rate": 8.357082984073764e-05,
"loss": 0.1632,
"step": 1390
},
{
"epoch": 0.5846605196982397,
"grad_norm": 0.26536107063293457,
"learning_rate": 8.315171835708298e-05,
"loss": 0.161,
"step": 1395
},
{
"epoch": 0.586756077116513,
"grad_norm": 0.21128953993320465,
"learning_rate": 8.273260687342833e-05,
"loss": 0.1585,
"step": 1400
},
{
"epoch": 0.5888516345347863,
"grad_norm": 3.6902313232421875,
"learning_rate": 8.231349538977369e-05,
"loss": 0.153,
"step": 1405
},
{
"epoch": 0.5909471919530596,
"grad_norm": 0.18794021010398865,
"learning_rate": 8.189438390611904e-05,
"loss": 0.1585,
"step": 1410
},
{
"epoch": 0.5930427493713327,
"grad_norm": 0.17947795987129211,
"learning_rate": 8.147527242246438e-05,
"loss": 0.1584,
"step": 1415
},
{
"epoch": 0.595138306789606,
"grad_norm": 0.15674275159835815,
"learning_rate": 8.105616093880973e-05,
"loss": 0.1547,
"step": 1420
},
{
"epoch": 0.5972338642078793,
"grad_norm": 0.15368077158927917,
"learning_rate": 8.063704945515508e-05,
"loss": 0.154,
"step": 1425
},
{
"epoch": 0.5993294216261525,
"grad_norm": 0.1675073206424713,
"learning_rate": 8.021793797150042e-05,
"loss": 0.1519,
"step": 1430
},
{
"epoch": 0.6014249790444258,
"grad_norm": 0.15424901247024536,
"learning_rate": 7.979882648784577e-05,
"loss": 0.1532,
"step": 1435
},
{
"epoch": 0.6035205364626991,
"grad_norm": 1.752557396888733,
"learning_rate": 7.937971500419112e-05,
"loss": 0.1521,
"step": 1440
},
{
"epoch": 0.6056160938809724,
"grad_norm": 0.179281547665596,
"learning_rate": 7.896060352053647e-05,
"loss": 0.1568,
"step": 1445
},
{
"epoch": 0.6077116512992456,
"grad_norm": 0.18618442118167877,
"learning_rate": 7.854149203688181e-05,
"loss": 0.1555,
"step": 1450
},
{
"epoch": 0.6098072087175188,
"grad_norm": 0.1767512410879135,
"learning_rate": 7.812238055322716e-05,
"loss": 0.156,
"step": 1455
},
{
"epoch": 0.6119027661357921,
"grad_norm": 0.1729685217142105,
"learning_rate": 7.770326906957252e-05,
"loss": 0.1521,
"step": 1460
},
{
"epoch": 0.6139983235540654,
"grad_norm": 0.16246432065963745,
"learning_rate": 7.728415758591785e-05,
"loss": 0.1541,
"step": 1465
},
{
"epoch": 0.6160938809723386,
"grad_norm": 0.15939489006996155,
"learning_rate": 7.68650461022632e-05,
"loss": 0.1551,
"step": 1470
},
{
"epoch": 0.6181894383906119,
"grad_norm": 0.1684809774160385,
"learning_rate": 7.644593461860854e-05,
"loss": 0.1542,
"step": 1475
},
{
"epoch": 0.6202849958088852,
"grad_norm": 0.18968342244625092,
"learning_rate": 7.602682313495391e-05,
"loss": 0.1537,
"step": 1480
},
{
"epoch": 0.6223805532271585,
"grad_norm": 0.20144881308078766,
"learning_rate": 7.560771165129925e-05,
"loss": 0.158,
"step": 1485
},
{
"epoch": 0.6244761106454317,
"grad_norm": 0.1785353720188141,
"learning_rate": 7.51886001676446e-05,
"loss": 0.1537,
"step": 1490
},
{
"epoch": 0.6265716680637049,
"grad_norm": 3.9302639961242676,
"learning_rate": 7.476948868398994e-05,
"loss": 0.1539,
"step": 1495
},
{
"epoch": 0.6286672254819782,
"grad_norm": 0.17349500954151154,
"learning_rate": 7.43503772003353e-05,
"loss": 0.1549,
"step": 1500
},
{
"epoch": 0.6307627829002515,
"grad_norm": 0.181074857711792,
"learning_rate": 7.393126571668064e-05,
"loss": 0.1523,
"step": 1505
},
{
"epoch": 0.6328583403185247,
"grad_norm": 0.1837494969367981,
"learning_rate": 7.351215423302599e-05,
"loss": 0.1543,
"step": 1510
},
{
"epoch": 0.634953897736798,
"grad_norm": 0.17972221970558167,
"learning_rate": 7.309304274937133e-05,
"loss": 0.1539,
"step": 1515
},
{
"epoch": 0.6370494551550713,
"grad_norm": 0.18084241449832916,
"learning_rate": 7.267393126571668e-05,
"loss": 0.1504,
"step": 1520
},
{
"epoch": 0.6391450125733446,
"grad_norm": 0.16487205028533936,
"learning_rate": 7.225481978206203e-05,
"loss": 0.1547,
"step": 1525
},
{
"epoch": 0.6412405699916177,
"grad_norm": 3.8815670013427734,
"learning_rate": 7.183570829840737e-05,
"loss": 0.1481,
"step": 1530
},
{
"epoch": 0.643336127409891,
"grad_norm": 0.18625426292419434,
"learning_rate": 7.141659681475272e-05,
"loss": 0.1551,
"step": 1535
},
{
"epoch": 0.6454316848281643,
"grad_norm": 0.2006104439496994,
"learning_rate": 7.099748533109808e-05,
"loss": 0.1567,
"step": 1540
},
{
"epoch": 0.6475272422464375,
"grad_norm": 0.20722784101963043,
"learning_rate": 7.057837384744343e-05,
"loss": 0.1599,
"step": 1545
},
{
"epoch": 0.6496227996647108,
"grad_norm": 0.18843336403369904,
"learning_rate": 7.015926236378877e-05,
"loss": 0.1568,
"step": 1550
},
{
"epoch": 0.6517183570829841,
"grad_norm": 0.16453127562999725,
"learning_rate": 6.974015088013412e-05,
"loss": 0.1564,
"step": 1555
},
{
"epoch": 0.6538139145012574,
"grad_norm": 0.1718008667230606,
"learning_rate": 6.932103939647947e-05,
"loss": 0.1507,
"step": 1560
},
{
"epoch": 0.6559094719195306,
"grad_norm": 1.9749072790145874,
"learning_rate": 6.890192791282482e-05,
"loss": 0.1546,
"step": 1565
},
{
"epoch": 0.6580050293378038,
"grad_norm": 0.846435546875,
"learning_rate": 6.848281642917016e-05,
"loss": 0.1553,
"step": 1570
},
{
"epoch": 0.6601005867560771,
"grad_norm": 0.21918132901191711,
"learning_rate": 6.806370494551551e-05,
"loss": 0.1544,
"step": 1575
},
{
"epoch": 0.6621961441743504,
"grad_norm": 0.20493735373020172,
"learning_rate": 6.764459346186086e-05,
"loss": 0.1588,
"step": 1580
},
{
"epoch": 0.6642917015926236,
"grad_norm": 0.16465319693088531,
"learning_rate": 6.72254819782062e-05,
"loss": 0.1556,
"step": 1585
},
{
"epoch": 0.6663872590108969,
"grad_norm": 0.1511167287826538,
"learning_rate": 6.680637049455155e-05,
"loss": 0.1551,
"step": 1590
},
{
"epoch": 0.6684828164291702,
"grad_norm": 0.15197286009788513,
"learning_rate": 6.63872590108969e-05,
"loss": 0.1513,
"step": 1595
},
{
"epoch": 0.6705783738474435,
"grad_norm": 0.14843755960464478,
"learning_rate": 6.596814752724226e-05,
"loss": 0.1476,
"step": 1600
},
{
"epoch": 0.6726739312657167,
"grad_norm": 0.14049085974693298,
"learning_rate": 6.55490360435876e-05,
"loss": 0.1507,
"step": 1605
},
{
"epoch": 0.6747694886839899,
"grad_norm": 0.1492355465888977,
"learning_rate": 6.512992455993295e-05,
"loss": 0.1499,
"step": 1610
},
{
"epoch": 0.6768650461022632,
"grad_norm": 0.15042835474014282,
"learning_rate": 6.47108130762783e-05,
"loss": 0.15,
"step": 1615
},
{
"epoch": 0.6789606035205364,
"grad_norm": 0.14849700033664703,
"learning_rate": 6.429170159262365e-05,
"loss": 0.152,
"step": 1620
},
{
"epoch": 0.6810561609388097,
"grad_norm": 0.15401747822761536,
"learning_rate": 6.387259010896899e-05,
"loss": 0.1507,
"step": 1625
},
{
"epoch": 0.683151718357083,
"grad_norm": 0.13168840110301971,
"learning_rate": 6.345347862531434e-05,
"loss": 0.153,
"step": 1630
},
{
"epoch": 0.6852472757753563,
"grad_norm": 0.1513977199792862,
"learning_rate": 6.303436714165969e-05,
"loss": 0.1516,
"step": 1635
},
{
"epoch": 0.6873428331936295,
"grad_norm": 0.14483259618282318,
"learning_rate": 6.261525565800503e-05,
"loss": 0.15,
"step": 1640
},
{
"epoch": 0.6894383906119028,
"grad_norm": 0.15542860329151154,
"learning_rate": 6.219614417435038e-05,
"loss": 0.1493,
"step": 1645
},
{
"epoch": 0.691533948030176,
"grad_norm": 0.14083847403526306,
"learning_rate": 6.177703269069572e-05,
"loss": 0.1511,
"step": 1650
},
{
"epoch": 0.6936295054484493,
"grad_norm": 0.14129005372524261,
"learning_rate": 6.135792120704108e-05,
"loss": 0.1525,
"step": 1655
},
{
"epoch": 0.6957250628667225,
"grad_norm": 0.14137649536132812,
"learning_rate": 6.093880972338642e-05,
"loss": 0.1486,
"step": 1660
},
{
"epoch": 0.6978206202849958,
"grad_norm": 0.163183331489563,
"learning_rate": 6.051969823973177e-05,
"loss": 0.1526,
"step": 1665
},
{
"epoch": 0.6999161777032691,
"grad_norm": 0.15466581284999847,
"learning_rate": 6.010058675607711e-05,
"loss": 0.1523,
"step": 1670
},
{
"epoch": 0.7020117351215424,
"grad_norm": 0.1582300364971161,
"learning_rate": 5.968147527242247e-05,
"loss": 0.1501,
"step": 1675
},
{
"epoch": 0.7041072925398156,
"grad_norm": 0.15225544571876526,
"learning_rate": 5.9262363788767817e-05,
"loss": 0.1529,
"step": 1680
},
{
"epoch": 0.7062028499580888,
"grad_norm": 0.1497444212436676,
"learning_rate": 5.884325230511316e-05,
"loss": 0.1504,
"step": 1685
},
{
"epoch": 0.7082984073763621,
"grad_norm": 0.15386608242988586,
"learning_rate": 5.8424140821458507e-05,
"loss": 0.1507,
"step": 1690
},
{
"epoch": 0.7103939647946353,
"grad_norm": 0.16286878287792206,
"learning_rate": 5.8005029337803865e-05,
"loss": 0.1483,
"step": 1695
},
{
"epoch": 0.7124895222129086,
"grad_norm": 0.16049712896347046,
"learning_rate": 5.758591785414921e-05,
"loss": 0.1511,
"step": 1700
},
{
"epoch": 0.7145850796311819,
"grad_norm": 0.16950005292892456,
"learning_rate": 5.7166806370494555e-05,
"loss": 0.1497,
"step": 1705
},
{
"epoch": 0.7166806370494552,
"grad_norm": 0.14905424416065216,
"learning_rate": 5.67476948868399e-05,
"loss": 0.1518,
"step": 1710
},
{
"epoch": 0.7187761944677284,
"grad_norm": 0.16250193119049072,
"learning_rate": 5.632858340318525e-05,
"loss": 0.1518,
"step": 1715
},
{
"epoch": 0.7208717518860017,
"grad_norm": 0.15719325840473175,
"learning_rate": 5.59094719195306e-05,
"loss": 0.1509,
"step": 1720
},
{
"epoch": 0.7229673093042749,
"grad_norm": 0.13987241685390472,
"learning_rate": 5.549036043587594e-05,
"loss": 0.1493,
"step": 1725
},
{
"epoch": 0.7250628667225482,
"grad_norm": 0.15766645967960358,
"learning_rate": 5.507124895222129e-05,
"loss": 0.151,
"step": 1730
},
{
"epoch": 0.7271584241408214,
"grad_norm": 0.1426958292722702,
"learning_rate": 5.4652137468566645e-05,
"loss": 0.1477,
"step": 1735
},
{
"epoch": 0.7292539815590947,
"grad_norm": 0.15173014998435974,
"learning_rate": 5.423302598491199e-05,
"loss": 0.1498,
"step": 1740
},
{
"epoch": 0.731349538977368,
"grad_norm": 0.13781285285949707,
"learning_rate": 5.3813914501257335e-05,
"loss": 0.1461,
"step": 1745
},
{
"epoch": 0.7334450963956413,
"grad_norm": 0.15598557889461517,
"learning_rate": 5.339480301760269e-05,
"loss": 0.1468,
"step": 1750
},
{
"epoch": 0.7355406538139145,
"grad_norm": 0.1487160325050354,
"learning_rate": 5.297569153394803e-05,
"loss": 0.1515,
"step": 1755
},
{
"epoch": 0.7376362112321878,
"grad_norm": 0.1766008883714676,
"learning_rate": 5.255658005029338e-05,
"loss": 0.1496,
"step": 1760
},
{
"epoch": 0.739731768650461,
"grad_norm": 0.1634136140346527,
"learning_rate": 5.213746856663872e-05,
"loss": 0.152,
"step": 1765
},
{
"epoch": 0.7418273260687342,
"grad_norm": 0.16231022775173187,
"learning_rate": 5.171835708298408e-05,
"loss": 0.1511,
"step": 1770
},
{
"epoch": 0.7439228834870075,
"grad_norm": 0.1736147552728653,
"learning_rate": 5.1299245599329425e-05,
"loss": 0.1519,
"step": 1775
},
{
"epoch": 0.7460184409052808,
"grad_norm": 0.15696606040000916,
"learning_rate": 5.088013411567477e-05,
"loss": 0.1522,
"step": 1780
},
{
"epoch": 0.7481139983235541,
"grad_norm": 0.16596518456935883,
"learning_rate": 5.0461022632020115e-05,
"loss": 0.1494,
"step": 1785
},
{
"epoch": 0.7502095557418273,
"grad_norm": 0.16456949710845947,
"learning_rate": 5.0041911148365474e-05,
"loss": 0.1485,
"step": 1790
},
{
"epoch": 0.7523051131601006,
"grad_norm": 0.16951188445091248,
"learning_rate": 4.962279966471082e-05,
"loss": 0.1536,
"step": 1795
},
{
"epoch": 0.7544006705783739,
"grad_norm": 0.17507892847061157,
"learning_rate": 4.9203688181056164e-05,
"loss": 0.1524,
"step": 1800
},
{
"epoch": 0.7564962279966471,
"grad_norm": 0.1613740175962448,
"learning_rate": 4.8784576697401516e-05,
"loss": 0.151,
"step": 1805
},
{
"epoch": 0.7585917854149203,
"grad_norm": 0.1735554188489914,
"learning_rate": 4.836546521374686e-05,
"loss": 0.149,
"step": 1810
},
{
"epoch": 0.7606873428331936,
"grad_norm": 0.15241263806819916,
"learning_rate": 4.7946353730092206e-05,
"loss": 0.1475,
"step": 1815
},
{
"epoch": 0.7627829002514669,
"grad_norm": 0.15598662197589874,
"learning_rate": 4.752724224643755e-05,
"loss": 0.1504,
"step": 1820
},
{
"epoch": 0.7648784576697402,
"grad_norm": 0.1638031303882599,
"learning_rate": 4.71081307627829e-05,
"loss": 0.1483,
"step": 1825
},
{
"epoch": 0.7669740150880134,
"grad_norm": 0.15647321939468384,
"learning_rate": 4.668901927912825e-05,
"loss": 0.1487,
"step": 1830
},
{
"epoch": 0.7690695725062867,
"grad_norm": 0.15045635402202606,
"learning_rate": 4.62699077954736e-05,
"loss": 0.1489,
"step": 1835
},
{
"epoch": 0.7711651299245599,
"grad_norm": 0.1481800377368927,
"learning_rate": 4.5850796311818944e-05,
"loss": 0.1498,
"step": 1840
},
{
"epoch": 0.7732606873428332,
"grad_norm": 0.16117189824581146,
"learning_rate": 4.5431684828164296e-05,
"loss": 0.147,
"step": 1845
},
{
"epoch": 0.7753562447611064,
"grad_norm": 0.15690650045871735,
"learning_rate": 4.501257334450964e-05,
"loss": 0.152,
"step": 1850
},
{
"epoch": 0.7774518021793797,
"grad_norm": 0.15324068069458008,
"learning_rate": 4.459346186085499e-05,
"loss": 0.1509,
"step": 1855
},
{
"epoch": 0.779547359597653,
"grad_norm": 0.1670048087835312,
"learning_rate": 4.417435037720034e-05,
"loss": 0.1481,
"step": 1860
},
{
"epoch": 0.7816429170159263,
"grad_norm": 0.16738423705101013,
"learning_rate": 4.375523889354568e-05,
"loss": 0.1525,
"step": 1865
},
{
"epoch": 0.7837384744341995,
"grad_norm": 0.15926587581634521,
"learning_rate": 4.333612740989103e-05,
"loss": 0.1499,
"step": 1870
},
{
"epoch": 0.7858340318524728,
"grad_norm": 0.15555323660373688,
"learning_rate": 4.291701592623638e-05,
"loss": 0.1487,
"step": 1875
},
{
"epoch": 0.787929589270746,
"grad_norm": 0.14897003769874573,
"learning_rate": 4.2497904442581724e-05,
"loss": 0.153,
"step": 1880
},
{
"epoch": 0.7900251466890192,
"grad_norm": 0.16712865233421326,
"learning_rate": 4.2078792958927076e-05,
"loss": 0.1522,
"step": 1885
},
{
"epoch": 0.7921207041072925,
"grad_norm": 0.16521087288856506,
"learning_rate": 4.165968147527242e-05,
"loss": 0.1554,
"step": 1890
},
{
"epoch": 0.7942162615255658,
"grad_norm": 0.15836192667484283,
"learning_rate": 4.124056999161777e-05,
"loss": 0.1515,
"step": 1895
},
{
"epoch": 0.7963118189438391,
"grad_norm": 0.1507478654384613,
"learning_rate": 4.0821458507963125e-05,
"loss": 0.1525,
"step": 1900
},
{
"epoch": 0.7984073763621123,
"grad_norm": 0.6994775533676147,
"learning_rate": 4.040234702430847e-05,
"loss": 0.1522,
"step": 1905
},
{
"epoch": 0.8005029337803856,
"grad_norm": 0.1698751598596573,
"learning_rate": 3.998323554065382e-05,
"loss": 0.152,
"step": 1910
},
{
"epoch": 0.8025984911986589,
"grad_norm": 0.16350072622299194,
"learning_rate": 3.9564124056999166e-05,
"loss": 0.1499,
"step": 1915
},
{
"epoch": 0.8046940486169321,
"grad_norm": 0.15547800064086914,
"learning_rate": 3.914501257334451e-05,
"loss": 0.1496,
"step": 1920
},
{
"epoch": 0.8067896060352053,
"grad_norm": 0.1584520936012268,
"learning_rate": 3.8725901089689856e-05,
"loss": 0.1521,
"step": 1925
},
{
"epoch": 0.8088851634534786,
"grad_norm": 0.15392720699310303,
"learning_rate": 3.830678960603521e-05,
"loss": 0.1513,
"step": 1930
},
{
"epoch": 0.8109807208717519,
"grad_norm": 0.15251478552818298,
"learning_rate": 3.788767812238055e-05,
"loss": 0.1503,
"step": 1935
},
{
"epoch": 0.8130762782900252,
"grad_norm": 0.1555824875831604,
"learning_rate": 3.7468566638725905e-05,
"loss": 0.148,
"step": 1940
},
{
"epoch": 0.8151718357082984,
"grad_norm": 1.0603052377700806,
"learning_rate": 3.704945515507125e-05,
"loss": 0.1511,
"step": 1945
},
{
"epoch": 0.8172673931265717,
"grad_norm": 0.21994204819202423,
"learning_rate": 3.66303436714166e-05,
"loss": 0.1536,
"step": 1950
},
{
"epoch": 0.819362950544845,
"grad_norm": 0.2238265573978424,
"learning_rate": 3.6211232187761947e-05,
"loss": 0.1549,
"step": 1955
},
{
"epoch": 0.8214585079631181,
"grad_norm": 0.212602898478508,
"learning_rate": 3.57921207041073e-05,
"loss": 0.1567,
"step": 1960
},
{
"epoch": 0.8235540653813914,
"grad_norm": 0.18257835507392883,
"learning_rate": 3.537300922045264e-05,
"loss": 0.1513,
"step": 1965
},
{
"epoch": 0.8256496227996647,
"grad_norm": 0.17607073485851288,
"learning_rate": 3.495389773679799e-05,
"loss": 0.1539,
"step": 1970
},
{
"epoch": 0.827745180217938,
"grad_norm": 0.14804469048976898,
"learning_rate": 3.453478625314333e-05,
"loss": 0.1502,
"step": 1975
},
{
"epoch": 0.8298407376362112,
"grad_norm": 0.15939727425575256,
"learning_rate": 3.4115674769488685e-05,
"loss": 0.1482,
"step": 1980
},
{
"epoch": 0.8319362950544845,
"grad_norm": 0.16081516444683075,
"learning_rate": 3.369656328583403e-05,
"loss": 0.152,
"step": 1985
},
{
"epoch": 0.8340318524727578,
"grad_norm": 0.7829539179801941,
"learning_rate": 3.327745180217938e-05,
"loss": 0.1497,
"step": 1990
},
{
"epoch": 0.836127409891031,
"grad_norm": 0.15778230130672455,
"learning_rate": 3.285834031852473e-05,
"loss": 0.1485,
"step": 1995
},
{
"epoch": 0.8382229673093042,
"grad_norm": 0.16156698763370514,
"learning_rate": 3.243922883487008e-05,
"loss": 0.1508,
"step": 2000
},
{
"epoch": 0.8403185247275775,
"grad_norm": 0.15631239116191864,
"learning_rate": 3.2020117351215424e-05,
"loss": 0.1484,
"step": 2005
},
{
"epoch": 0.8424140821458508,
"grad_norm": 0.1522732824087143,
"learning_rate": 3.1601005867560775e-05,
"loss": 0.1508,
"step": 2010
},
{
"epoch": 0.8445096395641241,
"grad_norm": 0.1581004112958908,
"learning_rate": 3.118189438390612e-05,
"loss": 0.149,
"step": 2015
},
{
"epoch": 0.8466051969823973,
"grad_norm": 0.1513088047504425,
"learning_rate": 3.076278290025147e-05,
"loss": 0.15,
"step": 2020
},
{
"epoch": 0.8487007544006706,
"grad_norm": 0.1780407577753067,
"learning_rate": 3.0343671416596814e-05,
"loss": 0.1494,
"step": 2025
},
{
"epoch": 0.8507963118189439,
"grad_norm": 0.16309364140033722,
"learning_rate": 2.9924559932942165e-05,
"loss": 0.1474,
"step": 2030
},
{
"epoch": 0.852891869237217,
"grad_norm": 0.15739715099334717,
"learning_rate": 2.950544844928751e-05,
"loss": 0.1496,
"step": 2035
},
{
"epoch": 0.8549874266554903,
"grad_norm": 0.1566157042980194,
"learning_rate": 2.908633696563286e-05,
"loss": 0.1482,
"step": 2040
},
{
"epoch": 0.8570829840737636,
"grad_norm": 0.14910633862018585,
"learning_rate": 2.8667225481978204e-05,
"loss": 0.1493,
"step": 2045
},
{
"epoch": 0.8591785414920369,
"grad_norm": 0.15060973167419434,
"learning_rate": 2.8248113998323556e-05,
"loss": 0.1462,
"step": 2050
},
{
"epoch": 0.8612740989103101,
"grad_norm": 0.15438151359558105,
"learning_rate": 2.7829002514668907e-05,
"loss": 0.1509,
"step": 2055
},
{
"epoch": 0.8633696563285834,
"grad_norm": 0.15441566705703735,
"learning_rate": 2.7409891031014252e-05,
"loss": 0.1481,
"step": 2060
},
{
"epoch": 0.8654652137468567,
"grad_norm": 0.15039560198783875,
"learning_rate": 2.69907795473596e-05,
"loss": 0.1466,
"step": 2065
},
{
"epoch": 0.86756077116513,
"grad_norm": 0.14182990789413452,
"learning_rate": 2.6571668063704946e-05,
"loss": 0.1475,
"step": 2070
},
{
"epoch": 0.8696563285834031,
"grad_norm": 0.1383397877216339,
"learning_rate": 2.6152556580050297e-05,
"loss": 0.1498,
"step": 2075
},
{
"epoch": 0.8717518860016764,
"grad_norm": 0.1353161633014679,
"learning_rate": 2.5733445096395642e-05,
"loss": 0.1485,
"step": 2080
},
{
"epoch": 0.8738474434199497,
"grad_norm": 0.14230230450630188,
"learning_rate": 2.5314333612740994e-05,
"loss": 0.1488,
"step": 2085
},
{
"epoch": 0.875943000838223,
"grad_norm": 0.14749757945537567,
"learning_rate": 2.4895222129086336e-05,
"loss": 0.1464,
"step": 2090
},
{
"epoch": 0.8780385582564962,
"grad_norm": 0.1343354731798172,
"learning_rate": 2.4476110645431684e-05,
"loss": 0.1476,
"step": 2095
},
{
"epoch": 0.8801341156747695,
"grad_norm": 0.15284012258052826,
"learning_rate": 2.4056999161777032e-05,
"loss": 0.1499,
"step": 2100
},
{
"epoch": 0.8822296730930428,
"grad_norm": 0.14641784131526947,
"learning_rate": 2.363788767812238e-05,
"loss": 0.1473,
"step": 2105
},
{
"epoch": 0.8843252305113161,
"grad_norm": 0.13705027103424072,
"learning_rate": 2.321877619446773e-05,
"loss": 0.1464,
"step": 2110
},
{
"epoch": 0.8864207879295892,
"grad_norm": 0.15073353052139282,
"learning_rate": 2.2799664710813078e-05,
"loss": 0.1464,
"step": 2115
},
{
"epoch": 0.8885163453478625,
"grad_norm": 0.14341919124126434,
"learning_rate": 2.2380553227158423e-05,
"loss": 0.1479,
"step": 2120
},
{
"epoch": 0.8906119027661358,
"grad_norm": 0.14984482526779175,
"learning_rate": 2.196144174350377e-05,
"loss": 0.1468,
"step": 2125
},
{
"epoch": 0.892707460184409,
"grad_norm": 0.1421850025653839,
"learning_rate": 2.1542330259849123e-05,
"loss": 0.1431,
"step": 2130
},
{
"epoch": 0.8948030176026823,
"grad_norm": 0.15779058635234833,
"learning_rate": 2.112321877619447e-05,
"loss": 0.1484,
"step": 2135
},
{
"epoch": 0.8968985750209556,
"grad_norm": 0.14414595067501068,
"learning_rate": 2.0704107292539816e-05,
"loss": 0.1477,
"step": 2140
},
{
"epoch": 0.8989941324392289,
"grad_norm": 0.15078584849834442,
"learning_rate": 2.0284995808885164e-05,
"loss": 0.1471,
"step": 2145
},
{
"epoch": 0.901089689857502,
"grad_norm": 0.6652347445487976,
"learning_rate": 1.9865884325230513e-05,
"loss": 0.146,
"step": 2150
},
{
"epoch": 0.9031852472757753,
"grad_norm": 0.1553775519132614,
"learning_rate": 1.944677284157586e-05,
"loss": 0.1455,
"step": 2155
},
{
"epoch": 0.9052808046940486,
"grad_norm": 0.14678525924682617,
"learning_rate": 1.902766135792121e-05,
"loss": 0.1469,
"step": 2160
},
{
"epoch": 0.9073763621123219,
"grad_norm": 0.1497979462146759,
"learning_rate": 1.8608549874266558e-05,
"loss": 0.1488,
"step": 2165
},
{
"epoch": 0.9094719195305951,
"grad_norm": 0.14247293770313263,
"learning_rate": 1.8189438390611903e-05,
"loss": 0.147,
"step": 2170
},
{
"epoch": 0.9115674769488684,
"grad_norm": 0.1436048001050949,
"learning_rate": 1.777032690695725e-05,
"loss": 0.148,
"step": 2175
},
{
"epoch": 0.9136630343671417,
"grad_norm": 0.1506524235010147,
"learning_rate": 1.73512154233026e-05,
"loss": 0.147,
"step": 2180
},
{
"epoch": 0.915758591785415,
"grad_norm": 0.14503729343414307,
"learning_rate": 1.6932103939647948e-05,
"loss": 0.1475,
"step": 2185
},
{
"epoch": 0.9178541492036881,
"grad_norm": 0.14085812866687775,
"learning_rate": 1.6512992455993296e-05,
"loss": 0.1468,
"step": 2190
},
{
"epoch": 0.9199497066219614,
"grad_norm": 0.13786457479000092,
"learning_rate": 1.609388097233864e-05,
"loss": 0.1449,
"step": 2195
},
{
"epoch": 0.9220452640402347,
"grad_norm": 0.14363813400268555,
"learning_rate": 1.567476948868399e-05,
"loss": 0.144,
"step": 2200
},
{
"epoch": 0.924140821458508,
"grad_norm": 0.14791929721832275,
"learning_rate": 1.5255658005029338e-05,
"loss": 0.147,
"step": 2205
},
{
"epoch": 0.9262363788767812,
"grad_norm": 0.14655451476573944,
"learning_rate": 1.4836546521374687e-05,
"loss": 0.1496,
"step": 2210
},
{
"epoch": 0.9283319362950545,
"grad_norm": 0.14993825554847717,
"learning_rate": 1.4417435037720033e-05,
"loss": 0.1465,
"step": 2215
},
{
"epoch": 0.9304274937133278,
"grad_norm": 0.1357424557209015,
"learning_rate": 1.3998323554065382e-05,
"loss": 0.1478,
"step": 2220
},
{
"epoch": 0.932523051131601,
"grad_norm": 0.15557856857776642,
"learning_rate": 1.357921207041073e-05,
"loss": 0.1429,
"step": 2225
},
{
"epoch": 0.9346186085498742,
"grad_norm": 0.16618798673152924,
"learning_rate": 1.3160100586756077e-05,
"loss": 0.1475,
"step": 2230
},
{
"epoch": 0.9367141659681475,
"grad_norm": 0.14325110614299774,
"learning_rate": 1.2740989103101425e-05,
"loss": 0.1453,
"step": 2235
},
{
"epoch": 0.9388097233864208,
"grad_norm": 0.15656636655330658,
"learning_rate": 1.2321877619446773e-05,
"loss": 0.1463,
"step": 2240
},
{
"epoch": 0.940905280804694,
"grad_norm": 0.15591678023338318,
"learning_rate": 1.1902766135792122e-05,
"loss": 0.1475,
"step": 2245
},
{
"epoch": 0.9430008382229673,
"grad_norm": 0.14923661947250366,
"learning_rate": 1.148365465213747e-05,
"loss": 0.1474,
"step": 2250
},
{
"epoch": 0.9450963956412406,
"grad_norm": 1.0343477725982666,
"learning_rate": 1.1064543168482817e-05,
"loss": 0.1434,
"step": 2255
},
{
"epoch": 0.9471919530595139,
"grad_norm": 0.16045677661895752,
"learning_rate": 1.0645431684828165e-05,
"loss": 0.1448,
"step": 2260
},
{
"epoch": 0.9492875104777871,
"grad_norm": 0.14702020585536957,
"learning_rate": 1.0226320201173514e-05,
"loss": 0.1468,
"step": 2265
},
{
"epoch": 0.9513830678960603,
"grad_norm": 0.14469429850578308,
"learning_rate": 9.80720871751886e-06,
"loss": 0.1434,
"step": 2270
},
{
"epoch": 0.9534786253143336,
"grad_norm": 0.1469959169626236,
"learning_rate": 9.388097233864209e-06,
"loss": 0.1487,
"step": 2275
},
{
"epoch": 0.9555741827326069,
"grad_norm": 0.13562710583209991,
"learning_rate": 8.968985750209557e-06,
"loss": 0.1477,
"step": 2280
},
{
"epoch": 0.9576697401508801,
"grad_norm": 0.15792514383792877,
"learning_rate": 8.549874266554904e-06,
"loss": 0.149,
"step": 2285
},
{
"epoch": 0.9597652975691534,
"grad_norm": 0.13964731991291046,
"learning_rate": 8.130762782900252e-06,
"loss": 0.144,
"step": 2290
},
{
"epoch": 0.9618608549874267,
"grad_norm": 0.1598869264125824,
"learning_rate": 7.711651299245599e-06,
"loss": 0.1486,
"step": 2295
},
{
"epoch": 0.9639564124057,
"grad_norm": 11.353910446166992,
"learning_rate": 7.292539815590947e-06,
"loss": 0.1441,
"step": 2300
},
{
"epoch": 0.9660519698239731,
"grad_norm": 0.9498974084854126,
"learning_rate": 6.873428331936295e-06,
"loss": 0.149,
"step": 2305
},
{
"epoch": 0.9681475272422464,
"grad_norm": 0.15554401278495789,
"learning_rate": 6.454316848281643e-06,
"loss": 0.1449,
"step": 2310
},
{
"epoch": 0.9702430846605197,
"grad_norm": 0.16430126130580902,
"learning_rate": 6.035205364626991e-06,
"loss": 0.1451,
"step": 2315
},
{
"epoch": 0.972338642078793,
"grad_norm": 0.15480098128318787,
"learning_rate": 5.616093880972339e-06,
"loss": 0.146,
"step": 2320
},
{
"epoch": 0.9744341994970662,
"grad_norm": 0.15997706353664398,
"learning_rate": 5.1969823973176864e-06,
"loss": 0.1483,
"step": 2325
},
{
"epoch": 0.9765297569153395,
"grad_norm": 0.15260910987854004,
"learning_rate": 4.777870913663034e-06,
"loss": 0.1472,
"step": 2330
},
{
"epoch": 0.9786253143336128,
"grad_norm": 0.15502935647964478,
"learning_rate": 4.358759430008382e-06,
"loss": 0.1456,
"step": 2335
},
{
"epoch": 0.980720871751886,
"grad_norm": 0.14989130198955536,
"learning_rate": 3.939647946353731e-06,
"loss": 0.1464,
"step": 2340
},
{
"epoch": 0.9828164291701592,
"grad_norm": 0.1601138710975647,
"learning_rate": 3.5205364626990782e-06,
"loss": 0.1479,
"step": 2345
},
{
"epoch": 0.9849119865884325,
"grad_norm": 0.15280233323574066,
"learning_rate": 3.101424979044426e-06,
"loss": 0.1459,
"step": 2350
},
{
"epoch": 0.9870075440067058,
"grad_norm": 0.15722650289535522,
"learning_rate": 2.6823134953897737e-06,
"loss": 0.1484,
"step": 2355
},
{
"epoch": 0.989103101424979,
"grad_norm": 0.15498095750808716,
"learning_rate": 2.2632020117351217e-06,
"loss": 0.1486,
"step": 2360
},
{
"epoch": 0.9911986588432523,
"grad_norm": 0.16407634317874908,
"learning_rate": 1.8440905280804694e-06,
"loss": 0.1508,
"step": 2365
},
{
"epoch": 0.9932942162615256,
"grad_norm": 0.1465112566947937,
"learning_rate": 1.4249790444258174e-06,
"loss": 0.1482,
"step": 2370
},
{
"epoch": 0.9953897736797989,
"grad_norm": 0.1511894315481186,
"learning_rate": 1.005867560771165e-06,
"loss": 0.1458,
"step": 2375
},
{
"epoch": 0.9974853310980721,
"grad_norm": 0.15166474878787994,
"learning_rate": 5.86756077116513e-07,
"loss": 0.147,
"step": 2380
},
{
"epoch": 0.9995808885163453,
"grad_norm": 1.638846516609192,
"learning_rate": 1.6764459346186086e-07,
"loss": 0.1469,
"step": 2385
},
{
"epoch": 1.0,
"eval_loss": 0.14722344279289246,
"eval_runtime": 1611.1392,
"eval_samples_per_second": 5.265,
"eval_steps_per_second": 0.659,
"step": 2386
}
],
"logging_steps": 5,
"max_steps": 2386,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 6.667359824212132e+17,
"train_batch_size": 32,
"trial_name": null,
"trial_params": null
}