search-llm / trainer_state.json
Espresso-AI's picture
Upload folder using huggingface_hub
8228657 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 4.0,
"eval_steps": 500,
"global_step": 24996,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.008001280204832774,
"grad_norm": 0.9324016571044922,
"learning_rate": 2.0000000000000003e-06,
"loss": 2.0562,
"step": 50
},
{
"epoch": 0.016002560409665547,
"grad_norm": 0.4582217037677765,
"learning_rate": 4.000000000000001e-06,
"loss": 1.9037,
"step": 100
},
{
"epoch": 0.02400384061449832,
"grad_norm": 0.6721668243408203,
"learning_rate": 6e-06,
"loss": 1.6957,
"step": 150
},
{
"epoch": 0.032005120819331094,
"grad_norm": 0.6238133907318115,
"learning_rate": 8.000000000000001e-06,
"loss": 1.5897,
"step": 200
},
{
"epoch": 0.04000640102416387,
"grad_norm": 0.6062737703323364,
"learning_rate": 1e-05,
"loss": 1.5664,
"step": 250
},
{
"epoch": 0.04800768122899664,
"grad_norm": 0.6185649633407593,
"learning_rate": 1.2e-05,
"loss": 1.5315,
"step": 300
},
{
"epoch": 0.05600896143382941,
"grad_norm": 0.6351732611656189,
"learning_rate": 1.4000000000000001e-05,
"loss": 1.4996,
"step": 350
},
{
"epoch": 0.06401024163866219,
"grad_norm": 0.6717058420181274,
"learning_rate": 1.6000000000000003e-05,
"loss": 1.4784,
"step": 400
},
{
"epoch": 0.07201152184349496,
"grad_norm": 0.6665578484535217,
"learning_rate": 1.8e-05,
"loss": 1.4795,
"step": 450
},
{
"epoch": 0.08001280204832774,
"grad_norm": 0.5694233775138855,
"learning_rate": 2e-05,
"loss": 1.4583,
"step": 500
},
{
"epoch": 0.0880140822531605,
"grad_norm": 0.6044241786003113,
"learning_rate": 2.2000000000000003e-05,
"loss": 1.4392,
"step": 550
},
{
"epoch": 0.09601536245799328,
"grad_norm": 0.6153579354286194,
"learning_rate": 2.4e-05,
"loss": 1.436,
"step": 600
},
{
"epoch": 0.10401664266282605,
"grad_norm": 0.5287560224533081,
"learning_rate": 2.6000000000000002e-05,
"loss": 1.4244,
"step": 650
},
{
"epoch": 0.11201792286765883,
"grad_norm": 0.5597755908966064,
"learning_rate": 2.8000000000000003e-05,
"loss": 1.4369,
"step": 700
},
{
"epoch": 0.1200192030724916,
"grad_norm": 0.4981745183467865,
"learning_rate": 3e-05,
"loss": 1.3981,
"step": 750
},
{
"epoch": 0.12802048327732438,
"grad_norm": 0.48743560910224915,
"learning_rate": 3.2000000000000005e-05,
"loss": 1.4009,
"step": 800
},
{
"epoch": 0.13602176348215714,
"grad_norm": 0.5637604594230652,
"learning_rate": 3.4000000000000007e-05,
"loss": 1.4028,
"step": 850
},
{
"epoch": 0.14402304368698993,
"grad_norm": 0.470486581325531,
"learning_rate": 3.6e-05,
"loss": 1.3938,
"step": 900
},
{
"epoch": 0.1520243238918227,
"grad_norm": 0.5075104236602783,
"learning_rate": 3.8e-05,
"loss": 1.3859,
"step": 950
},
{
"epoch": 0.16002560409665548,
"grad_norm": 0.4708271324634552,
"learning_rate": 4e-05,
"loss": 1.3865,
"step": 1000
},
{
"epoch": 0.16802688430148824,
"grad_norm": 0.3994545638561249,
"learning_rate": 4.2e-05,
"loss": 1.4072,
"step": 1050
},
{
"epoch": 0.176028164506321,
"grad_norm": 0.4936317801475525,
"learning_rate": 4.4000000000000006e-05,
"loss": 1.3828,
"step": 1100
},
{
"epoch": 0.1840294447111538,
"grad_norm": 0.3775925934314728,
"learning_rate": 4.600000000000001e-05,
"loss": 1.4046,
"step": 1150
},
{
"epoch": 0.19203072491598655,
"grad_norm": 0.45916563272476196,
"learning_rate": 4.8e-05,
"loss": 1.3725,
"step": 1200
},
{
"epoch": 0.20003200512081934,
"grad_norm": 0.42247775197029114,
"learning_rate": 5e-05,
"loss": 1.371,
"step": 1250
},
{
"epoch": 0.2080332853256521,
"grad_norm": 0.4264281392097473,
"learning_rate": 5.2000000000000004e-05,
"loss": 1.3729,
"step": 1300
},
{
"epoch": 0.2160345655304849,
"grad_norm": 0.3741685152053833,
"learning_rate": 5.4000000000000005e-05,
"loss": 1.3621,
"step": 1350
},
{
"epoch": 0.22403584573531765,
"grad_norm": 0.392640084028244,
"learning_rate": 5.6000000000000006e-05,
"loss": 1.3557,
"step": 1400
},
{
"epoch": 0.2320371259401504,
"grad_norm": 0.3837718367576599,
"learning_rate": 5.8e-05,
"loss": 1.3599,
"step": 1450
},
{
"epoch": 0.2400384061449832,
"grad_norm": 0.36931145191192627,
"learning_rate": 6e-05,
"loss": 1.3527,
"step": 1500
},
{
"epoch": 0.24803968634981596,
"grad_norm": 0.4106380343437195,
"learning_rate": 6.2e-05,
"loss": 1.3531,
"step": 1550
},
{
"epoch": 0.25604096655464875,
"grad_norm": 0.41261059045791626,
"learning_rate": 6.400000000000001e-05,
"loss": 1.3722,
"step": 1600
},
{
"epoch": 0.2640422467594815,
"grad_norm": 0.3928689658641815,
"learning_rate": 6.6e-05,
"loss": 1.3423,
"step": 1650
},
{
"epoch": 0.2720435269643143,
"grad_norm": 0.4072926640510559,
"learning_rate": 6.800000000000001e-05,
"loss": 1.3458,
"step": 1700
},
{
"epoch": 0.28004480716914704,
"grad_norm": 0.3700864315032959,
"learning_rate": 7e-05,
"loss": 1.3566,
"step": 1750
},
{
"epoch": 0.28804608737397985,
"grad_norm": 0.36805135011672974,
"learning_rate": 7.2e-05,
"loss": 1.3124,
"step": 1800
},
{
"epoch": 0.2960473675788126,
"grad_norm": 0.39832988381385803,
"learning_rate": 7.4e-05,
"loss": 1.3408,
"step": 1850
},
{
"epoch": 0.3040486477836454,
"grad_norm": 0.4116155505180359,
"learning_rate": 7.6e-05,
"loss": 1.3259,
"step": 1900
},
{
"epoch": 0.31204992798847814,
"grad_norm": 0.3844102919101715,
"learning_rate": 7.800000000000001e-05,
"loss": 1.3223,
"step": 1950
},
{
"epoch": 0.32005120819331095,
"grad_norm": 0.36435940861701965,
"learning_rate": 8e-05,
"loss": 1.3493,
"step": 2000
},
{
"epoch": 0.3280524883981437,
"grad_norm": 0.40303346514701843,
"learning_rate": 8.2e-05,
"loss": 1.313,
"step": 2050
},
{
"epoch": 0.3360537686029765,
"grad_norm": 0.3959302604198456,
"learning_rate": 8.4e-05,
"loss": 1.3116,
"step": 2100
},
{
"epoch": 0.34405504880780924,
"grad_norm": 0.42148756980895996,
"learning_rate": 8.6e-05,
"loss": 1.3172,
"step": 2150
},
{
"epoch": 0.352056329012642,
"grad_norm": 0.35398295521736145,
"learning_rate": 8.800000000000001e-05,
"loss": 1.3123,
"step": 2200
},
{
"epoch": 0.3600576092174748,
"grad_norm": 0.32611459493637085,
"learning_rate": 9e-05,
"loss": 1.3291,
"step": 2250
},
{
"epoch": 0.3680588894223076,
"grad_norm": 0.3089679777622223,
"learning_rate": 9.200000000000001e-05,
"loss": 1.3162,
"step": 2300
},
{
"epoch": 0.37606016962714034,
"grad_norm": 0.38052898645401,
"learning_rate": 9.4e-05,
"loss": 1.2888,
"step": 2350
},
{
"epoch": 0.3840614498319731,
"grad_norm": 0.40810805559158325,
"learning_rate": 9.6e-05,
"loss": 1.292,
"step": 2400
},
{
"epoch": 0.39206273003680586,
"grad_norm": 0.3573181927204132,
"learning_rate": 9.8e-05,
"loss": 1.3094,
"step": 2450
},
{
"epoch": 0.4000640102416387,
"grad_norm": 0.3486098349094391,
"learning_rate": 0.0001,
"loss": 1.2908,
"step": 2500
},
{
"epoch": 0.40806529044647144,
"grad_norm": 0.4237920045852661,
"learning_rate": 9.999878110192404e-05,
"loss": 1.2868,
"step": 2550
},
{
"epoch": 0.4160665706513042,
"grad_norm": 0.3493655025959015,
"learning_rate": 9.999512446712467e-05,
"loss": 1.3221,
"step": 2600
},
{
"epoch": 0.42406785085613696,
"grad_norm": 0.3716041147708893,
"learning_rate": 9.998903027388448e-05,
"loss": 1.2728,
"step": 2650
},
{
"epoch": 0.4320691310609698,
"grad_norm": 0.43094709515571594,
"learning_rate": 9.998049881933149e-05,
"loss": 1.2864,
"step": 2700
},
{
"epoch": 0.44007041126580254,
"grad_norm": 0.379719078540802,
"learning_rate": 9.996953051942463e-05,
"loss": 1.2956,
"step": 2750
},
{
"epoch": 0.4480716914706353,
"grad_norm": 0.45424267649650574,
"learning_rate": 9.995612590893352e-05,
"loss": 1.2927,
"step": 2800
},
{
"epoch": 0.45607297167546806,
"grad_norm": 0.34709858894348145,
"learning_rate": 9.99402856414123e-05,
"loss": 1.2801,
"step": 2850
},
{
"epoch": 0.4640742518803008,
"grad_norm": 0.3956676423549652,
"learning_rate": 9.992201048916783e-05,
"loss": 1.2869,
"step": 2900
},
{
"epoch": 0.47207553208513364,
"grad_norm": 0.3462512493133545,
"learning_rate": 9.990130134322201e-05,
"loss": 1.268,
"step": 2950
},
{
"epoch": 0.4800768122899664,
"grad_norm": 0.3835497796535492,
"learning_rate": 9.987815921326842e-05,
"loss": 1.254,
"step": 3000
},
{
"epoch": 0.48807809249479917,
"grad_norm": 0.37989285588264465,
"learning_rate": 9.98525852276229e-05,
"loss": 1.2869,
"step": 3050
},
{
"epoch": 0.4960793726996319,
"grad_norm": 0.3443322777748108,
"learning_rate": 9.982458063316877e-05,
"loss": 1.2933,
"step": 3100
},
{
"epoch": 0.5040806529044647,
"grad_norm": 0.32426556944847107,
"learning_rate": 9.979414679529588e-05,
"loss": 1.2756,
"step": 3150
},
{
"epoch": 0.5120819331092975,
"grad_norm": 0.4362022578716278,
"learning_rate": 9.976128519783405e-05,
"loss": 1.251,
"step": 3200
},
{
"epoch": 0.5200832133141302,
"grad_norm": 0.3969637155532837,
"learning_rate": 9.972599744298086e-05,
"loss": 1.2593,
"step": 3250
},
{
"epoch": 0.528084493518963,
"grad_norm": 0.41199642419815063,
"learning_rate": 9.96882852512233e-05,
"loss": 1.2484,
"step": 3300
},
{
"epoch": 0.5360857737237958,
"grad_norm": 0.3282572329044342,
"learning_rate": 9.964815046125412e-05,
"loss": 1.2559,
"step": 3350
},
{
"epoch": 0.5440870539286286,
"grad_norm": 0.4450869858264923,
"learning_rate": 9.960559502988208e-05,
"loss": 1.2488,
"step": 3400
},
{
"epoch": 0.5520883341334614,
"grad_norm": 0.356630802154541,
"learning_rate": 9.956062103193648e-05,
"loss": 1.2688,
"step": 3450
},
{
"epoch": 0.5600896143382941,
"grad_norm": 0.35085442662239075,
"learning_rate": 9.95132306601661e-05,
"loss": 1.2398,
"step": 3500
},
{
"epoch": 0.5680908945431269,
"grad_norm": 0.4403521716594696,
"learning_rate": 9.946342622513229e-05,
"loss": 1.2442,
"step": 3550
},
{
"epoch": 0.5760921747479597,
"grad_norm": 0.4456493854522705,
"learning_rate": 9.941121015509621e-05,
"loss": 1.2413,
"step": 3600
},
{
"epoch": 0.5840934549527924,
"grad_norm": 0.4137299954891205,
"learning_rate": 9.935658499590059e-05,
"loss": 1.2546,
"step": 3650
},
{
"epoch": 0.5920947351576252,
"grad_norm": 0.3572884798049927,
"learning_rate": 9.929955341084547e-05,
"loss": 1.2533,
"step": 3700
},
{
"epoch": 0.600096015362458,
"grad_norm": 0.42966127395629883,
"learning_rate": 9.924011818055843e-05,
"loss": 1.2389,
"step": 3750
},
{
"epoch": 0.6080972955672908,
"grad_norm": 0.4403945803642273,
"learning_rate": 9.917828220285896e-05,
"loss": 1.2198,
"step": 3800
},
{
"epoch": 0.6160985757721236,
"grad_norm": 0.4667624235153198,
"learning_rate": 9.911404849261726e-05,
"loss": 1.2417,
"step": 3850
},
{
"epoch": 0.6240998559769563,
"grad_norm": 0.6005256772041321,
"learning_rate": 9.904742018160715e-05,
"loss": 1.2006,
"step": 3900
},
{
"epoch": 0.6321011361817891,
"grad_norm": 0.3864199221134186,
"learning_rate": 9.897840051835343e-05,
"loss": 1.2164,
"step": 3950
},
{
"epoch": 0.6401024163866219,
"grad_norm": 0.5411766171455383,
"learning_rate": 9.89069928679735e-05,
"loss": 1.2187,
"step": 4000
},
{
"epoch": 0.6481036965914546,
"grad_norm": 0.38500306010246277,
"learning_rate": 9.883320071201325e-05,
"loss": 1.2363,
"step": 4050
},
{
"epoch": 0.6561049767962874,
"grad_norm": 0.4081592559814453,
"learning_rate": 9.875702764827737e-05,
"loss": 1.2247,
"step": 4100
},
{
"epoch": 0.6641062570011201,
"grad_norm": 0.43678033351898193,
"learning_rate": 9.86784773906539e-05,
"loss": 1.2323,
"step": 4150
},
{
"epoch": 0.672107537205953,
"grad_norm": 0.4319170117378235,
"learning_rate": 9.859755376893314e-05,
"loss": 1.1906,
"step": 4200
},
{
"epoch": 0.6801088174107858,
"grad_norm": 0.4729995131492615,
"learning_rate": 9.851426072862098e-05,
"loss": 1.2098,
"step": 4250
},
{
"epoch": 0.6881100976156185,
"grad_norm": 0.4233373701572418,
"learning_rate": 9.842860233074645e-05,
"loss": 1.204,
"step": 4300
},
{
"epoch": 0.6961113778204513,
"grad_norm": 0.48369497060775757,
"learning_rate": 9.834058275166383e-05,
"loss": 1.2046,
"step": 4350
},
{
"epoch": 0.704112658025284,
"grad_norm": 0.36634331941604614,
"learning_rate": 9.825020628284896e-05,
"loss": 1.2117,
"step": 4400
},
{
"epoch": 0.7121139382301168,
"grad_norm": 0.41213107109069824,
"learning_rate": 9.815747733068995e-05,
"loss": 1.1977,
"step": 4450
},
{
"epoch": 0.7201152184349496,
"grad_norm": 0.37744736671447754,
"learning_rate": 9.80624004162725e-05,
"loss": 1.2372,
"step": 4500
},
{
"epoch": 0.7281164986397823,
"grad_norm": 0.49204540252685547,
"learning_rate": 9.79649801751593e-05,
"loss": 1.2191,
"step": 4550
},
{
"epoch": 0.7361177788446152,
"grad_norm": 0.42844048142433167,
"learning_rate": 9.786522135716415e-05,
"loss": 1.2114,
"step": 4600
},
{
"epoch": 0.7441190590494479,
"grad_norm": 0.5176280736923218,
"learning_rate": 9.776312882612028e-05,
"loss": 1.2041,
"step": 4650
},
{
"epoch": 0.7521203392542807,
"grad_norm": 0.44180744886398315,
"learning_rate": 9.765870755964331e-05,
"loss": 1.2096,
"step": 4700
},
{
"epoch": 0.7601216194591135,
"grad_norm": 0.4751160442829132,
"learning_rate": 9.755196264888846e-05,
"loss": 1.1928,
"step": 4750
},
{
"epoch": 0.7681228996639462,
"grad_norm": 0.4147936999797821,
"learning_rate": 9.744289929830236e-05,
"loss": 1.2183,
"step": 4800
},
{
"epoch": 0.776124179868779,
"grad_norm": 0.45143744349479675,
"learning_rate": 9.733152282536936e-05,
"loss": 1.1798,
"step": 4850
},
{
"epoch": 0.7841254600736117,
"grad_norm": 0.46376484632492065,
"learning_rate": 9.721783866035218e-05,
"loss": 1.1871,
"step": 4900
},
{
"epoch": 0.7921267402784445,
"grad_norm": 0.4516090154647827,
"learning_rate": 9.710185234602727e-05,
"loss": 1.1901,
"step": 4950
},
{
"epoch": 0.8001280204832774,
"grad_norm": 0.4624463617801666,
"learning_rate": 9.698356953741438e-05,
"loss": 1.1848,
"step": 5000
},
{
"epoch": 0.8081293006881101,
"grad_norm": 0.41177043318748474,
"learning_rate": 9.686299600150108e-05,
"loss": 1.1934,
"step": 5050
},
{
"epoch": 0.8161305808929429,
"grad_norm": 0.46030393242836,
"learning_rate": 9.674013761696136e-05,
"loss": 1.1695,
"step": 5100
},
{
"epoch": 0.8241318610977757,
"grad_norm": 0.48872360587120056,
"learning_rate": 9.661500037386921e-05,
"loss": 1.192,
"step": 5150
},
{
"epoch": 0.8321331413026084,
"grad_norm": 0.36913803219795227,
"learning_rate": 9.648759037340638e-05,
"loss": 1.1624,
"step": 5200
},
{
"epoch": 0.8401344215074412,
"grad_norm": 0.5374457836151123,
"learning_rate": 9.635791382756508e-05,
"loss": 1.1847,
"step": 5250
},
{
"epoch": 0.8481357017122739,
"grad_norm": 0.43082547187805176,
"learning_rate": 9.622597705884496e-05,
"loss": 1.1798,
"step": 5300
},
{
"epoch": 0.8561369819171067,
"grad_norm": 0.4757367968559265,
"learning_rate": 9.609178649994499e-05,
"loss": 1.185,
"step": 5350
},
{
"epoch": 0.8641382621219396,
"grad_norm": 0.41058582067489624,
"learning_rate": 9.595534869344974e-05,
"loss": 1.18,
"step": 5400
},
{
"epoch": 0.8721395423267723,
"grad_norm": 0.4382982552051544,
"learning_rate": 9.581667029151039e-05,
"loss": 1.1977,
"step": 5450
},
{
"epoch": 0.8801408225316051,
"grad_norm": 0.44662296772003174,
"learning_rate": 9.567575805552042e-05,
"loss": 1.1841,
"step": 5500
},
{
"epoch": 0.8881421027364378,
"grad_norm": 0.4811069071292877,
"learning_rate": 9.553261885578597e-05,
"loss": 1.1765,
"step": 5550
},
{
"epoch": 0.8961433829412706,
"grad_norm": 0.4336834251880646,
"learning_rate": 9.538725967119087e-05,
"loss": 1.1879,
"step": 5600
},
{
"epoch": 0.9041446631461034,
"grad_norm": 0.44644805788993835,
"learning_rate": 9.523968758885631e-05,
"loss": 1.2058,
"step": 5650
},
{
"epoch": 0.9121459433509361,
"grad_norm": 0.4499703347682953,
"learning_rate": 9.508990980379537e-05,
"loss": 1.1397,
"step": 5700
},
{
"epoch": 0.920147223555769,
"grad_norm": 0.4322624206542969,
"learning_rate": 9.493793361856226e-05,
"loss": 1.1396,
"step": 5750
},
{
"epoch": 0.9281485037606017,
"grad_norm": 0.43696922063827515,
"learning_rate": 9.478376644289613e-05,
"loss": 1.1698,
"step": 5800
},
{
"epoch": 0.9361497839654345,
"grad_norm": 0.4423019587993622,
"learning_rate": 9.462741579335992e-05,
"loss": 1.1365,
"step": 5850
},
{
"epoch": 0.9441510641702673,
"grad_norm": 0.4975908696651459,
"learning_rate": 9.44688892929739e-05,
"loss": 1.1775,
"step": 5900
},
{
"epoch": 0.9521523443751,
"grad_norm": 0.49656620621681213,
"learning_rate": 9.430819467084392e-05,
"loss": 1.141,
"step": 5950
},
{
"epoch": 0.9601536245799328,
"grad_norm": 0.4862751066684723,
"learning_rate": 9.414533976178458e-05,
"loss": 1.144,
"step": 6000
},
{
"epoch": 0.9681549047847655,
"grad_norm": 0.5158004760742188,
"learning_rate": 9.398033250593733e-05,
"loss": 1.1603,
"step": 6050
},
{
"epoch": 0.9761561849895983,
"grad_norm": 0.4406384229660034,
"learning_rate": 9.381318094838321e-05,
"loss": 1.1351,
"step": 6100
},
{
"epoch": 0.9841574651944311,
"grad_norm": 0.4819541573524475,
"learning_rate": 9.36438932387507e-05,
"loss": 1.172,
"step": 6150
},
{
"epoch": 0.9921587453992639,
"grad_norm": 0.4627222418785095,
"learning_rate": 9.347247763081835e-05,
"loss": 1.1587,
"step": 6200
},
{
"epoch": 1.0001600256040966,
"grad_norm": 0.4336191713809967,
"learning_rate": 9.329894248211233e-05,
"loss": 1.1324,
"step": 6250
},
{
"epoch": 1.0081613058089294,
"grad_norm": 0.5482844114303589,
"learning_rate": 9.312329625349902e-05,
"loss": 1.0647,
"step": 6300
},
{
"epoch": 1.0161625860137622,
"grad_norm": 0.48142924904823303,
"learning_rate": 9.294554750877242e-05,
"loss": 1.0681,
"step": 6350
},
{
"epoch": 1.024163866218595,
"grad_norm": 0.45703792572021484,
"learning_rate": 9.276570491423664e-05,
"loss": 1.0956,
"step": 6400
},
{
"epoch": 1.0321651464234278,
"grad_norm": 0.5305693745613098,
"learning_rate": 9.258377723828338e-05,
"loss": 1.0476,
"step": 6450
},
{
"epoch": 1.0401664266282604,
"grad_norm": 0.5061540603637695,
"learning_rate": 9.23997733509644e-05,
"loss": 1.0613,
"step": 6500
},
{
"epoch": 1.0481677068330932,
"grad_norm": 0.4534345865249634,
"learning_rate": 9.221370222355908e-05,
"loss": 1.0604,
"step": 6550
},
{
"epoch": 1.056168987037926,
"grad_norm": 0.5327995419502258,
"learning_rate": 9.202557292813697e-05,
"loss": 1.0396,
"step": 6600
},
{
"epoch": 1.0641702672427589,
"grad_norm": 0.5349058508872986,
"learning_rate": 9.183539463711555e-05,
"loss": 1.0909,
"step": 6650
},
{
"epoch": 1.0721715474475917,
"grad_norm": 0.5274565815925598,
"learning_rate": 9.16431766228129e-05,
"loss": 1.0539,
"step": 6700
},
{
"epoch": 1.0801728276524245,
"grad_norm": 0.5548765659332275,
"learning_rate": 9.144892825699577e-05,
"loss": 1.0438,
"step": 6750
},
{
"epoch": 1.088174107857257,
"grad_norm": 0.5511574745178223,
"learning_rate": 9.125265901042251e-05,
"loss": 1.091,
"step": 6800
},
{
"epoch": 1.09617538806209,
"grad_norm": 0.5760679841041565,
"learning_rate": 9.105437845238141e-05,
"loss": 1.0522,
"step": 6850
},
{
"epoch": 1.1041766682669227,
"grad_norm": 0.6349498629570007,
"learning_rate": 9.085409625022411e-05,
"loss": 1.0592,
"step": 6900
},
{
"epoch": 1.1121779484717556,
"grad_norm": 0.6243582963943481,
"learning_rate": 9.065182216889423e-05,
"loss": 1.0683,
"step": 6950
},
{
"epoch": 1.1201792286765881,
"grad_norm": 0.5389096140861511,
"learning_rate": 9.04475660704513e-05,
"loss": 1.0549,
"step": 7000
},
{
"epoch": 1.128180508881421,
"grad_norm": 0.4826034605503082,
"learning_rate": 9.024133791358995e-05,
"loss": 1.078,
"step": 7050
},
{
"epoch": 1.1361817890862538,
"grad_norm": 0.517662525177002,
"learning_rate": 9.003314775315433e-05,
"loss": 1.0857,
"step": 7100
},
{
"epoch": 1.1441830692910866,
"grad_norm": 0.5285195112228394,
"learning_rate": 8.982300573964788e-05,
"loss": 1.0449,
"step": 7150
},
{
"epoch": 1.1521843494959194,
"grad_norm": 0.47838062047958374,
"learning_rate": 8.961092211873841e-05,
"loss": 1.035,
"step": 7200
},
{
"epoch": 1.1601856297007522,
"grad_norm": 0.5411238670349121,
"learning_rate": 8.939690723075865e-05,
"loss": 1.058,
"step": 7250
},
{
"epoch": 1.1681869099055848,
"grad_norm": 0.49349942803382874,
"learning_rate": 8.9180971510202e-05,
"loss": 1.0567,
"step": 7300
},
{
"epoch": 1.1761881901104176,
"grad_norm": 0.5397252440452576,
"learning_rate": 8.896312548521382e-05,
"loss": 1.0842,
"step": 7350
},
{
"epoch": 1.1841894703152505,
"grad_norm": 0.5732976794242859,
"learning_rate": 8.874337977707817e-05,
"loss": 1.0581,
"step": 7400
},
{
"epoch": 1.1921907505200833,
"grad_norm": 0.4691354036331177,
"learning_rate": 8.852174509969984e-05,
"loss": 1.0205,
"step": 7450
},
{
"epoch": 1.2001920307249159,
"grad_norm": 0.6526204943656921,
"learning_rate": 8.829823225908214e-05,
"loss": 1.0387,
"step": 7500
},
{
"epoch": 1.2081933109297487,
"grad_norm": 0.46403875946998596,
"learning_rate": 8.80728521527999e-05,
"loss": 1.0554,
"step": 7550
},
{
"epoch": 1.2161945911345815,
"grad_norm": 0.462835431098938,
"learning_rate": 8.784561576946824e-05,
"loss": 1.0503,
"step": 7600
},
{
"epoch": 1.2241958713394143,
"grad_norm": 0.5204878449440002,
"learning_rate": 8.76165341882068e-05,
"loss": 1.0815,
"step": 7650
},
{
"epoch": 1.2321971515442471,
"grad_norm": 0.5061752796173096,
"learning_rate": 8.738561857809949e-05,
"loss": 1.0534,
"step": 7700
},
{
"epoch": 1.24019843174908,
"grad_norm": 0.5491166710853577,
"learning_rate": 8.715288019765006e-05,
"loss": 1.0625,
"step": 7750
},
{
"epoch": 1.2481997119539125,
"grad_norm": 0.4824821650981903,
"learning_rate": 8.691833039423305e-05,
"loss": 1.0473,
"step": 7800
},
{
"epoch": 1.2562009921587454,
"grad_norm": 0.5545431971549988,
"learning_rate": 8.668198060354065e-05,
"loss": 1.0432,
"step": 7850
},
{
"epoch": 1.2642022723635782,
"grad_norm": 0.5635075569152832,
"learning_rate": 8.644384234902503e-05,
"loss": 1.0473,
"step": 7900
},
{
"epoch": 1.272203552568411,
"grad_norm": 0.49760839343070984,
"learning_rate": 8.620392724133661e-05,
"loss": 1.0616,
"step": 7950
},
{
"epoch": 1.2802048327732436,
"grad_norm": 0.6228436231613159,
"learning_rate": 8.596224697775795e-05,
"loss": 1.0722,
"step": 8000
},
{
"epoch": 1.2882061129780764,
"grad_norm": 0.5831848978996277,
"learning_rate": 8.571881334163334e-05,
"loss": 1.0581,
"step": 8050
},
{
"epoch": 1.2962073931829092,
"grad_norm": 0.4925413131713867,
"learning_rate": 8.547363820179442e-05,
"loss": 1.0429,
"step": 8100
},
{
"epoch": 1.304208673387742,
"grad_norm": 0.5401794910430908,
"learning_rate": 8.522673351198143e-05,
"loss": 1.0366,
"step": 8150
},
{
"epoch": 1.3122099535925749,
"grad_norm": 0.5081512928009033,
"learning_rate": 8.497811131026046e-05,
"loss": 1.0464,
"step": 8200
},
{
"epoch": 1.3202112337974077,
"grad_norm": 0.5704318881034851,
"learning_rate": 8.472778371843641e-05,
"loss": 1.0509,
"step": 8250
},
{
"epoch": 1.3282125140022405,
"grad_norm": 0.505493700504303,
"learning_rate": 8.447576294146209e-05,
"loss": 1.0713,
"step": 8300
},
{
"epoch": 1.336213794207073,
"grad_norm": 0.5492855906486511,
"learning_rate": 8.422206126684311e-05,
"loss": 1.0385,
"step": 8350
},
{
"epoch": 1.344215074411906,
"grad_norm": 0.4849604666233063,
"learning_rate": 8.39666910640388e-05,
"loss": 1.0422,
"step": 8400
},
{
"epoch": 1.3522163546167387,
"grad_norm": 0.5335291624069214,
"learning_rate": 8.370966478385907e-05,
"loss": 1.0235,
"step": 8450
},
{
"epoch": 1.3602176348215713,
"grad_norm": 0.5804494619369507,
"learning_rate": 8.345099495785753e-05,
"loss": 1.0069,
"step": 8500
},
{
"epoch": 1.3682189150264041,
"grad_norm": 0.5363398194313049,
"learning_rate": 8.319069419772023e-05,
"loss": 1.0402,
"step": 8550
},
{
"epoch": 1.376220195231237,
"grad_norm": 0.6018884181976318,
"learning_rate": 8.292877519465105e-05,
"loss": 1.031,
"step": 8600
},
{
"epoch": 1.3842214754360698,
"grad_norm": 0.5088986158370972,
"learning_rate": 8.266525071875273e-05,
"loss": 1.0349,
"step": 8650
},
{
"epoch": 1.3922227556409026,
"grad_norm": 0.575141191482544,
"learning_rate": 8.240013361840436e-05,
"loss": 1.019,
"step": 8700
},
{
"epoch": 1.4002240358457354,
"grad_norm": 0.5023386478424072,
"learning_rate": 8.21334368196348e-05,
"loss": 1.0428,
"step": 8750
},
{
"epoch": 1.4082253160505682,
"grad_norm": 0.5338187217712402,
"learning_rate": 8.186517332549276e-05,
"loss": 1.0188,
"step": 8800
},
{
"epoch": 1.4162265962554008,
"grad_norm": 0.5415542721748352,
"learning_rate": 8.159535621541243e-05,
"loss": 1.0172,
"step": 8850
},
{
"epoch": 1.4242278764602336,
"grad_norm": 0.4934610426425934,
"learning_rate": 8.13239986445761e-05,
"loss": 1.0249,
"step": 8900
},
{
"epoch": 1.4322291566650664,
"grad_norm": 0.5592052936553955,
"learning_rate": 8.10511138432726e-05,
"loss": 1.0295,
"step": 8950
},
{
"epoch": 1.4402304368698993,
"grad_norm": 0.5904792547225952,
"learning_rate": 8.07767151162523e-05,
"loss": 1.0427,
"step": 9000
},
{
"epoch": 1.4482317170747319,
"grad_norm": 0.5952058434486389,
"learning_rate": 8.050081584207843e-05,
"loss": 1.0504,
"step": 9050
},
{
"epoch": 1.4562329972795647,
"grad_norm": 0.49933668971061707,
"learning_rate": 8.022342947247474e-05,
"loss": 1.0391,
"step": 9100
},
{
"epoch": 1.4642342774843975,
"grad_norm": 0.5853399634361267,
"learning_rate": 7.994456953166973e-05,
"loss": 1.0013,
"step": 9150
},
{
"epoch": 1.4722355576892303,
"grad_norm": 0.6290839910507202,
"learning_rate": 7.966424961573722e-05,
"loss": 1.0159,
"step": 9200
},
{
"epoch": 1.4802368378940631,
"grad_norm": 0.6153242588043213,
"learning_rate": 7.938248339193346e-05,
"loss": 1.0312,
"step": 9250
},
{
"epoch": 1.488238118098896,
"grad_norm": 0.5723234415054321,
"learning_rate": 7.909928459803077e-05,
"loss": 1.0198,
"step": 9300
},
{
"epoch": 1.4962393983037285,
"grad_norm": 0.5766580700874329,
"learning_rate": 7.881466704164774e-05,
"loss": 0.9915,
"step": 9350
},
{
"epoch": 1.5042406785085614,
"grad_norm": 0.6211101412773132,
"learning_rate": 7.852864459957605e-05,
"loss": 1.0324,
"step": 9400
},
{
"epoch": 1.5122419587133942,
"grad_norm": 0.5500127673149109,
"learning_rate": 7.824123121710387e-05,
"loss": 1.0316,
"step": 9450
},
{
"epoch": 1.5202432389182268,
"grad_norm": 0.6087145209312439,
"learning_rate": 7.795244090733597e-05,
"loss": 1.0334,
"step": 9500
},
{
"epoch": 1.5282445191230596,
"grad_norm": 0.5456790924072266,
"learning_rate": 7.766228775051045e-05,
"loss": 1.0243,
"step": 9550
},
{
"epoch": 1.5362457993278924,
"grad_norm": 0.613003671169281,
"learning_rate": 7.737078589331229e-05,
"loss": 0.9954,
"step": 9600
},
{
"epoch": 1.5442470795327252,
"grad_norm": 0.5732637047767639,
"learning_rate": 7.707794954818363e-05,
"loss": 1.0172,
"step": 9650
},
{
"epoch": 1.552248359737558,
"grad_norm": 0.5933035612106323,
"learning_rate": 7.678379299263076e-05,
"loss": 1.0049,
"step": 9700
},
{
"epoch": 1.5602496399423909,
"grad_norm": 0.5576400756835938,
"learning_rate": 7.648833056852807e-05,
"loss": 0.9918,
"step": 9750
},
{
"epoch": 1.5682509201472237,
"grad_norm": 0.4915982782840729,
"learning_rate": 7.619157668141877e-05,
"loss": 0.9949,
"step": 9800
},
{
"epoch": 1.5762522003520565,
"grad_norm": 0.5492275357246399,
"learning_rate": 7.58935457998125e-05,
"loss": 1.0233,
"step": 9850
},
{
"epoch": 1.584253480556889,
"grad_norm": 0.5720072984695435,
"learning_rate": 7.559425245448006e-05,
"loss": 1.053,
"step": 9900
},
{
"epoch": 1.592254760761722,
"grad_norm": 0.5385164022445679,
"learning_rate": 7.529371123774471e-05,
"loss": 1.0329,
"step": 9950
},
{
"epoch": 1.6002560409665545,
"grad_norm": 0.5508076548576355,
"learning_rate": 7.49919368027709e-05,
"loss": 1.0156,
"step": 10000
},
{
"epoch": 1.6082573211713873,
"grad_norm": 0.5111932754516602,
"learning_rate": 7.468894386284976e-05,
"loss": 1.0061,
"step": 10050
},
{
"epoch": 1.6162586013762201,
"grad_norm": 0.5652381777763367,
"learning_rate": 7.438474719068173e-05,
"loss": 0.999,
"step": 10100
},
{
"epoch": 1.624259881581053,
"grad_norm": 0.5519747138023376,
"learning_rate": 7.407936161765637e-05,
"loss": 1.0129,
"step": 10150
},
{
"epoch": 1.6322611617858858,
"grad_norm": 0.5380906462669373,
"learning_rate": 7.377280203312917e-05,
"loss": 0.9991,
"step": 10200
},
{
"epoch": 1.6402624419907186,
"grad_norm": 0.635221540927887,
"learning_rate": 7.346508338369563e-05,
"loss": 0.9889,
"step": 10250
},
{
"epoch": 1.6482637221955514,
"grad_norm": 0.4963364005088806,
"learning_rate": 7.315622067246254e-05,
"loss": 0.9808,
"step": 10300
},
{
"epoch": 1.6562650024003842,
"grad_norm": 0.5576931238174438,
"learning_rate": 7.28462289583165e-05,
"loss": 1.0122,
"step": 10350
},
{
"epoch": 1.6642662826052168,
"grad_norm": 0.5357556343078613,
"learning_rate": 7.253512335518964e-05,
"loss": 1.0193,
"step": 10400
},
{
"epoch": 1.6722675628100496,
"grad_norm": 0.5757668018341064,
"learning_rate": 7.22229190313228e-05,
"loss": 1.0032,
"step": 10450
},
{
"epoch": 1.6802688430148822,
"grad_norm": 0.5368041396141052,
"learning_rate": 7.1909631208526e-05,
"loss": 1.0159,
"step": 10500
},
{
"epoch": 1.688270123219715,
"grad_norm": 0.5617921948432922,
"learning_rate": 7.15952751614362e-05,
"loss": 1.0178,
"step": 10550
},
{
"epoch": 1.6962714034245479,
"grad_norm": 0.6092725396156311,
"learning_rate": 7.127986621677263e-05,
"loss": 1.0112,
"step": 10600
},
{
"epoch": 1.7042726836293807,
"grad_norm": 0.6592617034912109,
"learning_rate": 7.096341975258953e-05,
"loss": 0.9944,
"step": 10650
},
{
"epoch": 1.7122739638342135,
"grad_norm": 0.5574333667755127,
"learning_rate": 7.064595119752636e-05,
"loss": 0.9896,
"step": 10700
},
{
"epoch": 1.7202752440390463,
"grad_norm": 0.5796050429344177,
"learning_rate": 7.032747603005552e-05,
"loss": 1.0099,
"step": 10750
},
{
"epoch": 1.7282765242438791,
"grad_norm": 0.5359869599342346,
"learning_rate": 7.00080097777278e-05,
"loss": 0.976,
"step": 10800
},
{
"epoch": 1.736277804448712,
"grad_norm": 0.533762514591217,
"learning_rate": 6.968756801641523e-05,
"loss": 1.011,
"step": 10850
},
{
"epoch": 1.7442790846535445,
"grad_norm": 0.6239184141159058,
"learning_rate": 6.936616636955164e-05,
"loss": 0.9971,
"step": 10900
},
{
"epoch": 1.7522803648583773,
"grad_norm": 0.6219179034233093,
"learning_rate": 6.904382050737096e-05,
"loss": 1.0195,
"step": 10950
},
{
"epoch": 1.7602816450632102,
"grad_norm": 0.6290085315704346,
"learning_rate": 6.872054614614328e-05,
"loss": 1.0393,
"step": 11000
},
{
"epoch": 1.7682829252680428,
"grad_norm": 0.6031250953674316,
"learning_rate": 6.839635904740846e-05,
"loss": 1.013,
"step": 11050
},
{
"epoch": 1.7762842054728756,
"grad_norm": 0.5307977199554443,
"learning_rate": 6.807127501720773e-05,
"loss": 0.9924,
"step": 11100
},
{
"epoch": 1.7842854856777084,
"grad_norm": 0.593413233757019,
"learning_rate": 6.774530990531308e-05,
"loss": 0.9848,
"step": 11150
},
{
"epoch": 1.7922867658825412,
"grad_norm": 0.5930994749069214,
"learning_rate": 6.741847960445437e-05,
"loss": 0.9891,
"step": 11200
},
{
"epoch": 1.800288046087374,
"grad_norm": 0.5843376517295837,
"learning_rate": 6.709080004954465e-05,
"loss": 1.0075,
"step": 11250
},
{
"epoch": 1.8082893262922068,
"grad_norm": 0.6507195830345154,
"learning_rate": 6.676228721690301e-05,
"loss": 0.9872,
"step": 11300
},
{
"epoch": 1.8162906064970397,
"grad_norm": 0.502108097076416,
"learning_rate": 6.64329571234759e-05,
"loss": 0.9898,
"step": 11350
},
{
"epoch": 1.8242918867018723,
"grad_norm": 0.6289730072021484,
"learning_rate": 6.610282582605597e-05,
"loss": 0.9886,
"step": 11400
},
{
"epoch": 1.832293166906705,
"grad_norm": 0.5781500935554504,
"learning_rate": 6.577190942049937e-05,
"loss": 1.0211,
"step": 11450
},
{
"epoch": 1.840294447111538,
"grad_norm": 0.6279985308647156,
"learning_rate": 6.544022404094092e-05,
"loss": 1.0123,
"step": 11500
},
{
"epoch": 1.8482957273163705,
"grad_norm": 0.6612405180931091,
"learning_rate": 6.51077858590074e-05,
"loss": 0.9978,
"step": 11550
},
{
"epoch": 1.8562970075212033,
"grad_norm": 0.6027297973632812,
"learning_rate": 6.477461108302927e-05,
"loss": 0.9786,
"step": 11600
},
{
"epoch": 1.8642982877260361,
"grad_norm": 0.6454290151596069,
"learning_rate": 6.444071595725026e-05,
"loss": 0.9692,
"step": 11650
},
{
"epoch": 1.872299567930869,
"grad_norm": 0.5385527014732361,
"learning_rate": 6.410611676103542e-05,
"loss": 0.9789,
"step": 11700
},
{
"epoch": 1.8803008481357018,
"grad_norm": 0.5324643850326538,
"learning_rate": 6.37708298080774e-05,
"loss": 0.9892,
"step": 11750
},
{
"epoch": 1.8883021283405346,
"grad_norm": 0.5693587064743042,
"learning_rate": 6.343487144560108e-05,
"loss": 0.9881,
"step": 11800
},
{
"epoch": 1.8963034085453674,
"grad_norm": 0.6578769683837891,
"learning_rate": 6.309825805356654e-05,
"loss": 0.9871,
"step": 11850
},
{
"epoch": 1.9043046887502,
"grad_norm": 0.5471733212471008,
"learning_rate": 6.276100604387039e-05,
"loss": 1.0013,
"step": 11900
},
{
"epoch": 1.9123059689550328,
"grad_norm": 0.6353397369384766,
"learning_rate": 6.24231318595457e-05,
"loss": 0.9653,
"step": 11950
},
{
"epoch": 1.9203072491598656,
"grad_norm": 0.5183307528495789,
"learning_rate": 6.208465197396013e-05,
"loss": 0.9833,
"step": 12000
},
{
"epoch": 1.9283085293646982,
"grad_norm": 0.539452850818634,
"learning_rate": 6.1745582890013e-05,
"loss": 0.9686,
"step": 12050
},
{
"epoch": 1.936309809569531,
"grad_norm": 0.681061863899231,
"learning_rate": 6.140594113933042e-05,
"loss": 0.9814,
"step": 12100
},
{
"epoch": 1.9443110897743638,
"grad_norm": 0.6232041716575623,
"learning_rate": 6.106574328145949e-05,
"loss": 0.9858,
"step": 12150
},
{
"epoch": 1.9523123699791967,
"grad_norm": 0.5435238480567932,
"learning_rate": 6.072500590306076e-05,
"loss": 0.9921,
"step": 12200
},
{
"epoch": 1.9603136501840295,
"grad_norm": 0.6599912047386169,
"learning_rate": 6.038374561709964e-05,
"loss": 0.9627,
"step": 12250
},
{
"epoch": 1.9683149303888623,
"grad_norm": 0.6491255760192871,
"learning_rate": 6.0041979062036356e-05,
"loss": 1.0074,
"step": 12300
},
{
"epoch": 1.976316210593695,
"grad_norm": 0.545872688293457,
"learning_rate": 5.969972290101478e-05,
"loss": 0.9769,
"step": 12350
},
{
"epoch": 1.984317490798528,
"grad_norm": 0.5526396632194519,
"learning_rate": 5.935699382104994e-05,
"loss": 0.9853,
"step": 12400
},
{
"epoch": 1.9923187710033605,
"grad_norm": 0.5161740183830261,
"learning_rate": 5.901380853221448e-05,
"loss": 1.0122,
"step": 12450
},
{
"epoch": 2.000320051208193,
"grad_norm": 0.501007080078125,
"learning_rate": 5.867018376682396e-05,
"loss": 1.0062,
"step": 12500
},
{
"epoch": 2.008321331413026,
"grad_norm": 0.6925072073936462,
"learning_rate": 5.832613627862097e-05,
"loss": 0.8524,
"step": 12550
},
{
"epoch": 2.0163226116178588,
"grad_norm": 0.6264834403991699,
"learning_rate": 5.798168284195837e-05,
"loss": 0.8474,
"step": 12600
},
{
"epoch": 2.0243238918226916,
"grad_norm": 0.543258786201477,
"learning_rate": 5.7636840250981405e-05,
"loss": 0.819,
"step": 12650
},
{
"epoch": 2.0323251720275244,
"grad_norm": 0.5597353577613831,
"learning_rate": 5.729162531880892e-05,
"loss": 0.8635,
"step": 12700
},
{
"epoch": 2.040326452232357,
"grad_norm": 0.8554184436798096,
"learning_rate": 5.694605487671357e-05,
"loss": 0.8283,
"step": 12750
},
{
"epoch": 2.04832773243719,
"grad_norm": 0.5572032332420349,
"learning_rate": 5.6600145773301206e-05,
"loss": 0.8744,
"step": 12800
},
{
"epoch": 2.056329012642023,
"grad_norm": 0.6553796529769897,
"learning_rate": 5.625391487368949e-05,
"loss": 0.8572,
"step": 12850
},
{
"epoch": 2.0643302928468557,
"grad_norm": 0.6666269302368164,
"learning_rate": 5.590737905868551e-05,
"loss": 0.8625,
"step": 12900
},
{
"epoch": 2.0723315730516885,
"grad_norm": 0.6420098543167114,
"learning_rate": 5.5560555223962784e-05,
"loss": 0.8571,
"step": 12950
},
{
"epoch": 2.080332853256521,
"grad_norm": 0.6100978255271912,
"learning_rate": 5.5213460279237496e-05,
"loss": 0.8579,
"step": 13000
},
{
"epoch": 2.0883341334613537,
"grad_norm": 0.5606010556221008,
"learning_rate": 5.4866111147444085e-05,
"loss": 0.8715,
"step": 13050
},
{
"epoch": 2.0963354136661865,
"grad_norm": 0.6860626339912415,
"learning_rate": 5.4518524763910076e-05,
"loss": 0.8513,
"step": 13100
},
{
"epoch": 2.1043366938710193,
"grad_norm": 0.7012873291969299,
"learning_rate": 5.4170718075530404e-05,
"loss": 0.8454,
"step": 13150
},
{
"epoch": 2.112337974075852,
"grad_norm": 0.6773386597633362,
"learning_rate": 5.382270803994126e-05,
"loss": 0.881,
"step": 13200
},
{
"epoch": 2.120339254280685,
"grad_norm": 0.7174808382987976,
"learning_rate": 5.347451162469309e-05,
"loss": 0.858,
"step": 13250
},
{
"epoch": 2.1283405344855177,
"grad_norm": 0.6617708206176758,
"learning_rate": 5.3126145806423575e-05,
"loss": 0.8531,
"step": 13300
},
{
"epoch": 2.1363418146903506,
"grad_norm": 0.7019938826560974,
"learning_rate": 5.277762757002971e-05,
"loss": 0.8665,
"step": 13350
},
{
"epoch": 2.1443430948951834,
"grad_norm": 0.6727155447006226,
"learning_rate": 5.2428973907839804e-05,
"loss": 0.8528,
"step": 13400
},
{
"epoch": 2.152344375100016,
"grad_norm": 0.6648008227348328,
"learning_rate": 5.208020181878499e-05,
"loss": 0.8399,
"step": 13450
},
{
"epoch": 2.160345655304849,
"grad_norm": 0.6274258494377136,
"learning_rate": 5.173132830757039e-05,
"loss": 0.8662,
"step": 13500
},
{
"epoch": 2.1683469355096814,
"grad_norm": 0.6117835640907288,
"learning_rate": 5.138237038384608e-05,
"loss": 0.8503,
"step": 13550
},
{
"epoch": 2.176348215714514,
"grad_norm": 0.6691561341285706,
"learning_rate": 5.103334506137772e-05,
"loss": 0.8645,
"step": 13600
},
{
"epoch": 2.184349495919347,
"grad_norm": 0.5877872109413147,
"learning_rate": 5.0684269357217084e-05,
"loss": 0.8555,
"step": 13650
},
{
"epoch": 2.19235077612418,
"grad_norm": 0.717145562171936,
"learning_rate": 5.033516029087231e-05,
"loss": 0.8799,
"step": 13700
},
{
"epoch": 2.2003520563290127,
"grad_norm": 0.6587306261062622,
"learning_rate": 4.99860348834782e-05,
"loss": 0.88,
"step": 13750
},
{
"epoch": 2.2083533365338455,
"grad_norm": 0.6128435134887695,
"learning_rate": 4.963691015696624e-05,
"loss": 0.853,
"step": 13800
},
{
"epoch": 2.2163546167386783,
"grad_norm": 0.5783178210258484,
"learning_rate": 4.928780313323473e-05,
"loss": 0.8863,
"step": 13850
},
{
"epoch": 2.224355896943511,
"grad_norm": 0.6364635825157166,
"learning_rate": 4.893873083331882e-05,
"loss": 0.8378,
"step": 13900
},
{
"epoch": 2.232357177148344,
"grad_norm": 0.7266676425933838,
"learning_rate": 4.8589710276560744e-05,
"loss": 0.8647,
"step": 13950
},
{
"epoch": 2.2403584573531763,
"grad_norm": 0.6661261916160583,
"learning_rate": 4.8240758479779884e-05,
"loss": 0.835,
"step": 14000
},
{
"epoch": 2.248359737558009,
"grad_norm": 0.7814577221870422,
"learning_rate": 4.789189245644318e-05,
"loss": 0.8379,
"step": 14050
},
{
"epoch": 2.256361017762842,
"grad_norm": 0.6463228464126587,
"learning_rate": 4.754312921583564e-05,
"loss": 0.8628,
"step": 14100
},
{
"epoch": 2.2643622979676747,
"grad_norm": 0.7221981883049011,
"learning_rate": 4.719448576223096e-05,
"loss": 0.8565,
"step": 14150
},
{
"epoch": 2.2723635781725076,
"grad_norm": 0.6834704279899597,
"learning_rate": 4.684597909406253e-05,
"loss": 0.8776,
"step": 14200
},
{
"epoch": 2.2803648583773404,
"grad_norm": 0.6285929083824158,
"learning_rate": 4.649762620309466e-05,
"loss": 0.8356,
"step": 14250
},
{
"epoch": 2.288366138582173,
"grad_norm": 0.6223785281181335,
"learning_rate": 4.61494440735941e-05,
"loss": 0.834,
"step": 14300
},
{
"epoch": 2.296367418787006,
"grad_norm": 0.7635972499847412,
"learning_rate": 4.580144968150192e-05,
"loss": 0.8603,
"step": 14350
},
{
"epoch": 2.304368698991839,
"grad_norm": 0.6658472418785095,
"learning_rate": 4.5453659993605944e-05,
"loss": 0.8393,
"step": 14400
},
{
"epoch": 2.3123699791966716,
"grad_norm": 0.6321608424186707,
"learning_rate": 4.510609196671345e-05,
"loss": 0.8416,
"step": 14450
},
{
"epoch": 2.3203712594015045,
"grad_norm": 0.6526052355766296,
"learning_rate": 4.4758762546824364e-05,
"loss": 0.8696,
"step": 14500
},
{
"epoch": 2.328372539606337,
"grad_norm": 0.6651396751403809,
"learning_rate": 4.441168866830516e-05,
"loss": 0.8619,
"step": 14550
},
{
"epoch": 2.3363738198111696,
"grad_norm": 0.6737195253372192,
"learning_rate": 4.406488725306317e-05,
"loss": 0.8447,
"step": 14600
},
{
"epoch": 2.3443751000160025,
"grad_norm": 0.5813534259796143,
"learning_rate": 4.3718375209721505e-05,
"loss": 0.8468,
"step": 14650
},
{
"epoch": 2.3523763802208353,
"grad_norm": 0.7838544249534607,
"learning_rate": 4.337216943279464e-05,
"loss": 0.8344,
"step": 14700
},
{
"epoch": 2.360377660425668,
"grad_norm": 0.6356272101402283,
"learning_rate": 4.3026286801864854e-05,
"loss": 0.8699,
"step": 14750
},
{
"epoch": 2.368378940630501,
"grad_norm": 0.6201728582382202,
"learning_rate": 4.268074418075903e-05,
"loss": 0.8515,
"step": 14800
},
{
"epoch": 2.3763802208353337,
"grad_norm": 0.6424387693405151,
"learning_rate": 4.233555841672663e-05,
"loss": 0.8429,
"step": 14850
},
{
"epoch": 2.3843815010401666,
"grad_norm": 0.6518192291259766,
"learning_rate": 4.1990746339618216e-05,
"loss": 0.8666,
"step": 14900
},
{
"epoch": 2.3923827812449994,
"grad_norm": 0.7236935496330261,
"learning_rate": 4.164632476106484e-05,
"loss": 0.8478,
"step": 14950
},
{
"epoch": 2.4003840614498317,
"grad_norm": 0.692077100276947,
"learning_rate": 4.13023104736585e-05,
"loss": 0.8756,
"step": 15000
},
{
"epoch": 2.4083853416546646,
"grad_norm": 0.6063551306724548,
"learning_rate": 4.095872025013333e-05,
"loss": 0.8221,
"step": 15050
},
{
"epoch": 2.4163866218594974,
"grad_norm": 0.5967947840690613,
"learning_rate": 4.0615570842547815e-05,
"loss": 0.8731,
"step": 15100
},
{
"epoch": 2.42438790206433,
"grad_norm": 0.7256248593330383,
"learning_rate": 4.0272878981468045e-05,
"loss": 0.8614,
"step": 15150
},
{
"epoch": 2.432389182269163,
"grad_norm": 0.6148253083229065,
"learning_rate": 3.9930661375152045e-05,
"loss": 0.8337,
"step": 15200
},
{
"epoch": 2.440390462473996,
"grad_norm": 0.6977733373641968,
"learning_rate": 3.958893470873509e-05,
"loss": 0.8235,
"step": 15250
},
{
"epoch": 2.4483917426788286,
"grad_norm": 0.6713569164276123,
"learning_rate": 3.924771564341621e-05,
"loss": 0.8301,
"step": 15300
},
{
"epoch": 2.4563930228836615,
"grad_norm": 0.637150228023529,
"learning_rate": 3.890702081564593e-05,
"loss": 0.8355,
"step": 15350
},
{
"epoch": 2.4643943030884943,
"grad_norm": 0.6515012383460999,
"learning_rate": 3.856686683631502e-05,
"loss": 0.8547,
"step": 15400
},
{
"epoch": 2.472395583293327,
"grad_norm": 0.6041115522384644,
"learning_rate": 3.822727028994471e-05,
"loss": 0.8421,
"step": 15450
},
{
"epoch": 2.48039686349816,
"grad_norm": 0.6472122073173523,
"learning_rate": 3.788824773387811e-05,
"loss": 0.838,
"step": 15500
},
{
"epoch": 2.4883981437029927,
"grad_norm": 0.7198679447174072,
"learning_rate": 3.754981569747287e-05,
"loss": 0.8213,
"step": 15550
},
{
"epoch": 2.496399423907825,
"grad_norm": 0.6885024309158325,
"learning_rate": 3.721199068129527e-05,
"loss": 0.8354,
"step": 15600
},
{
"epoch": 2.504400704112658,
"grad_norm": 0.6485710740089417,
"learning_rate": 3.6874789156315835e-05,
"loss": 0.8633,
"step": 15650
},
{
"epoch": 2.5124019843174907,
"grad_norm": 0.6844043135643005,
"learning_rate": 3.6538227563106164e-05,
"loss": 0.8105,
"step": 15700
},
{
"epoch": 2.5204032645223235,
"grad_norm": 0.8702924847602844,
"learning_rate": 3.6202322311037374e-05,
"loss": 0.8121,
"step": 15750
},
{
"epoch": 2.5284045447271564,
"grad_norm": 0.7209028005599976,
"learning_rate": 3.586708977748012e-05,
"loss": 0.8758,
"step": 15800
},
{
"epoch": 2.536405824931989,
"grad_norm": 0.6202672123908997,
"learning_rate": 3.553254630700597e-05,
"loss": 0.8474,
"step": 15850
},
{
"epoch": 2.544407105136822,
"grad_norm": 0.6669834852218628,
"learning_rate": 3.5198708210590626e-05,
"loss": 0.8482,
"step": 15900
},
{
"epoch": 2.552408385341655,
"grad_norm": 0.6584005951881409,
"learning_rate": 3.486559176481865e-05,
"loss": 0.8416,
"step": 15950
},
{
"epoch": 2.560409665546487,
"grad_norm": 0.5252745747566223,
"learning_rate": 3.4533213211089825e-05,
"loss": 0.8401,
"step": 16000
},
{
"epoch": 2.56841094575132,
"grad_norm": 0.6504681706428528,
"learning_rate": 3.4201588754827314e-05,
"loss": 0.8486,
"step": 16050
},
{
"epoch": 2.576412225956153,
"grad_norm": 0.7693225741386414,
"learning_rate": 3.387073456468761e-05,
"loss": 0.845,
"step": 16100
},
{
"epoch": 2.5844135061609856,
"grad_norm": 0.6965412497520447,
"learning_rate": 3.354066677177214e-05,
"loss": 0.8568,
"step": 16150
},
{
"epoch": 2.5924147863658185,
"grad_norm": 0.6364871859550476,
"learning_rate": 3.321140146884081e-05,
"loss": 0.8693,
"step": 16200
},
{
"epoch": 2.6004160665706513,
"grad_norm": 0.7464070916175842,
"learning_rate": 3.288295470952737e-05,
"loss": 0.8566,
"step": 16250
},
{
"epoch": 2.608417346775484,
"grad_norm": 0.6804706454277039,
"learning_rate": 3.255534250755678e-05,
"loss": 0.8708,
"step": 16300
},
{
"epoch": 2.616418626980317,
"grad_norm": 0.6775168776512146,
"learning_rate": 3.222858083596431e-05,
"loss": 0.8558,
"step": 16350
},
{
"epoch": 2.6244199071851497,
"grad_norm": 0.7037110924720764,
"learning_rate": 3.190268562631688e-05,
"loss": 0.8313,
"step": 16400
},
{
"epoch": 2.6324211873899825,
"grad_norm": 0.5500233173370361,
"learning_rate": 3.157767276793626e-05,
"loss": 0.7998,
"step": 16450
},
{
"epoch": 2.6404224675948154,
"grad_norm": 0.7002319097518921,
"learning_rate": 3.125355810712435e-05,
"loss": 0.8247,
"step": 16500
},
{
"epoch": 2.648423747799648,
"grad_norm": 0.7857663631439209,
"learning_rate": 3.093035744639061e-05,
"loss": 0.8318,
"step": 16550
},
{
"epoch": 2.656425028004481,
"grad_norm": 0.7866588830947876,
"learning_rate": 3.06080865436816e-05,
"loss": 0.814,
"step": 16600
},
{
"epoch": 2.6644263082093134,
"grad_norm": 0.6732119917869568,
"learning_rate": 3.0286761111612626e-05,
"loss": 0.8527,
"step": 16650
},
{
"epoch": 2.672427588414146,
"grad_norm": 0.6847143769264221,
"learning_rate": 2.9966396816701725e-05,
"loss": 0.8012,
"step": 16700
},
{
"epoch": 2.680428868618979,
"grad_norm": 0.6862397193908691,
"learning_rate": 2.964700927860581e-05,
"loss": 0.817,
"step": 16750
},
{
"epoch": 2.688430148823812,
"grad_norm": 0.6373656988143921,
"learning_rate": 2.9328614069359128e-05,
"loss": 0.8625,
"step": 16800
},
{
"epoch": 2.6964314290286446,
"grad_norm": 0.7099259495735168,
"learning_rate": 2.9011226712613937e-05,
"loss": 0.8445,
"step": 16850
},
{
"epoch": 2.7044327092334774,
"grad_norm": 0.6120832562446594,
"learning_rate": 2.8694862682883866e-05,
"loss": 0.7982,
"step": 16900
},
{
"epoch": 2.7124339894383103,
"grad_norm": 0.8442961573600769,
"learning_rate": 2.8379537404789124e-05,
"loss": 0.8307,
"step": 16950
},
{
"epoch": 2.7204352696431426,
"grad_norm": 0.6093111634254456,
"learning_rate": 2.8065266252304712e-05,
"loss": 0.7811,
"step": 17000
},
{
"epoch": 2.7284365498479755,
"grad_norm": 0.6981213092803955,
"learning_rate": 2.775206454801079e-05,
"loss": 0.8533,
"step": 17050
},
{
"epoch": 2.7364378300528083,
"grad_norm": 0.696796178817749,
"learning_rate": 2.7439947562345546e-05,
"loss": 0.8182,
"step": 17100
},
{
"epoch": 2.744439110257641,
"grad_norm": 0.6803203225135803,
"learning_rate": 2.7128930512860658e-05,
"loss": 0.8391,
"step": 17150
},
{
"epoch": 2.752440390462474,
"grad_norm": 0.6710325479507446,
"learning_rate": 2.6819028563479505e-05,
"loss": 0.8229,
"step": 17200
},
{
"epoch": 2.7604416706673067,
"grad_norm": 0.7354364395141602,
"learning_rate": 2.6510256823757667e-05,
"loss": 0.8121,
"step": 17250
},
{
"epoch": 2.7684429508721395,
"grad_norm": 0.7637900710105896,
"learning_rate": 2.6202630348146324e-05,
"loss": 0.84,
"step": 17300
},
{
"epoch": 2.7764442310769724,
"grad_norm": 0.587841808795929,
"learning_rate": 2.589616413525824e-05,
"loss": 0.8117,
"step": 17350
},
{
"epoch": 2.784445511281805,
"grad_norm": 0.667797327041626,
"learning_rate": 2.559087312713651e-05,
"loss": 0.8131,
"step": 17400
},
{
"epoch": 2.792446791486638,
"grad_norm": 0.5983754992485046,
"learning_rate": 2.5286772208526027e-05,
"loss": 0.8309,
"step": 17450
},
{
"epoch": 2.800448071691471,
"grad_norm": 0.7884683012962341,
"learning_rate": 2.4983876206147776e-05,
"loss": 0.8208,
"step": 17500
},
{
"epoch": 2.8084493518963036,
"grad_norm": 0.6957002878189087,
"learning_rate": 2.4682199887975938e-05,
"loss": 0.8699,
"step": 17550
},
{
"epoch": 2.8164506321011364,
"grad_norm": 0.5889870524406433,
"learning_rate": 2.438175796251786e-05,
"loss": 0.8282,
"step": 17600
},
{
"epoch": 2.824451912305969,
"grad_norm": 0.7357444167137146,
"learning_rate": 2.4082565078096935e-05,
"loss": 0.8025,
"step": 17650
},
{
"epoch": 2.8324531925108016,
"grad_norm": 0.8791075348854065,
"learning_rate": 2.3784635822138424e-05,
"loss": 0.8338,
"step": 17700
},
{
"epoch": 2.8404544727156344,
"grad_norm": 0.6486666202545166,
"learning_rate": 2.348798472045819e-05,
"loss": 0.8221,
"step": 17750
},
{
"epoch": 2.8484557529204673,
"grad_norm": 0.6927144527435303,
"learning_rate": 2.3192626236554516e-05,
"loss": 0.8249,
"step": 17800
},
{
"epoch": 2.8564570331253,
"grad_norm": 0.7874731421470642,
"learning_rate": 2.2898574770902914e-05,
"loss": 0.8267,
"step": 17850
},
{
"epoch": 2.864458313330133,
"grad_norm": 0.6948744654655457,
"learning_rate": 2.260584466025401e-05,
"loss": 0.8139,
"step": 17900
},
{
"epoch": 2.8724595935349657,
"grad_norm": 0.6847252249717712,
"learning_rate": 2.231445017693454e-05,
"loss": 0.8432,
"step": 17950
},
{
"epoch": 2.8804608737397985,
"grad_norm": 0.6781948804855347,
"learning_rate": 2.2024405528151547e-05,
"loss": 0.8298,
"step": 18000
},
{
"epoch": 2.888462153944631,
"grad_norm": 0.6336878538131714,
"learning_rate": 2.1735724855299567e-05,
"loss": 0.81,
"step": 18050
},
{
"epoch": 2.8964634341494637,
"grad_norm": 0.6111786365509033,
"learning_rate": 2.1448422233271254e-05,
"loss": 0.7977,
"step": 18100
},
{
"epoch": 2.9044647143542965,
"grad_norm": 0.7334992289543152,
"learning_rate": 2.116251166977118e-05,
"loss": 0.8221,
"step": 18150
},
{
"epoch": 2.9124659945591294,
"grad_norm": 0.8408239483833313,
"learning_rate": 2.0878007104632775e-05,
"loss": 0.8597,
"step": 18200
},
{
"epoch": 2.920467274763962,
"grad_norm": 0.6892300248146057,
"learning_rate": 2.059492240913866e-05,
"loss": 0.8088,
"step": 18250
},
{
"epoch": 2.928468554968795,
"grad_norm": 0.5887216925621033,
"learning_rate": 2.0313271385344522e-05,
"loss": 0.8207,
"step": 18300
},
{
"epoch": 2.936469835173628,
"grad_norm": 0.6477887630462646,
"learning_rate": 2.0033067765406004e-05,
"loss": 0.839,
"step": 18350
},
{
"epoch": 2.9444711153784606,
"grad_norm": 0.7348781228065491,
"learning_rate": 1.9754325210909174e-05,
"loss": 0.8412,
"step": 18400
},
{
"epoch": 2.9524723955832934,
"grad_norm": 0.6994746923446655,
"learning_rate": 1.947705731220462e-05,
"loss": 0.8069,
"step": 18450
},
{
"epoch": 2.9604736757881263,
"grad_norm": 0.6778689026832581,
"learning_rate": 1.920127758774466e-05,
"loss": 0.8119,
"step": 18500
},
{
"epoch": 2.968474955992959,
"grad_norm": 0.7225350737571716,
"learning_rate": 1.8926999483424313e-05,
"loss": 0.8504,
"step": 18550
},
{
"epoch": 2.976476236197792,
"grad_norm": 0.8543927073478699,
"learning_rate": 1.8654236371925727e-05,
"loss": 0.8124,
"step": 18600
},
{
"epoch": 2.9844775164026243,
"grad_norm": 0.6733004450798035,
"learning_rate": 1.8383001552066164e-05,
"loss": 0.8299,
"step": 18650
},
{
"epoch": 2.992478796607457,
"grad_norm": 0.7595524787902832,
"learning_rate": 1.8113308248149635e-05,
"loss": 0.8177,
"step": 18700
},
{
"epoch": 3.00048007681229,
"grad_norm": 0.6430355310440063,
"learning_rate": 1.784516960932211e-05,
"loss": 0.8311,
"step": 18750
},
{
"epoch": 3.0084813570171227,
"grad_norm": 0.6314593553543091,
"learning_rate": 1.7578598708930428e-05,
"loss": 0.7241,
"step": 18800
},
{
"epoch": 3.0164826372219555,
"grad_norm": 0.6709548234939575,
"learning_rate": 1.7313608543884868e-05,
"loss": 0.7459,
"step": 18850
},
{
"epoch": 3.0244839174267883,
"grad_norm": 0.6981055736541748,
"learning_rate": 1.7050212034025575e-05,
"loss": 0.7399,
"step": 18900
},
{
"epoch": 3.032485197631621,
"grad_norm": 0.7344752550125122,
"learning_rate": 1.6788422021492472e-05,
"loss": 0.7197,
"step": 18950
},
{
"epoch": 3.040486477836454,
"grad_norm": 0.7181177139282227,
"learning_rate": 1.6528251270099256e-05,
"loss": 0.7161,
"step": 19000
},
{
"epoch": 3.048487758041287,
"grad_norm": 0.5588538646697998,
"learning_rate": 1.6269712464711105e-05,
"loss": 0.7488,
"step": 19050
},
{
"epoch": 3.056489038246119,
"grad_norm": 0.7480596303939819,
"learning_rate": 1.6012818210626075e-05,
"loss": 0.712,
"step": 19100
},
{
"epoch": 3.064490318450952,
"grad_norm": 0.5872476696968079,
"learning_rate": 1.5757581032960638e-05,
"loss": 0.7154,
"step": 19150
},
{
"epoch": 3.072491598655785,
"grad_norm": 0.6517592072486877,
"learning_rate": 1.550401337603902e-05,
"loss": 0.7134,
"step": 19200
},
{
"epoch": 3.0804928788606176,
"grad_norm": 0.6740624308586121,
"learning_rate": 1.5252127602786397e-05,
"loss": 0.7446,
"step": 19250
},
{
"epoch": 3.0884941590654504,
"grad_norm": 0.6558243632316589,
"learning_rate": 1.5001935994126104e-05,
"loss": 0.7354,
"step": 19300
},
{
"epoch": 3.0964954392702833,
"grad_norm": 0.6562448740005493,
"learning_rate": 1.4753450748380953e-05,
"loss": 0.7035,
"step": 19350
},
{
"epoch": 3.104496719475116,
"grad_norm": 0.7197704315185547,
"learning_rate": 1.4506683980678503e-05,
"loss": 0.7313,
"step": 19400
},
{
"epoch": 3.112497999679949,
"grad_norm": 0.7322149872779846,
"learning_rate": 1.4261647722360278e-05,
"loss": 0.7357,
"step": 19450
},
{
"epoch": 3.1204992798847817,
"grad_norm": 0.7477754354476929,
"learning_rate": 1.4018353920395193e-05,
"loss": 0.7334,
"step": 19500
},
{
"epoch": 3.1285005600896145,
"grad_norm": 0.5789663791656494,
"learning_rate": 1.3776814436797176e-05,
"loss": 0.7435,
"step": 19550
},
{
"epoch": 3.1365018402944473,
"grad_norm": 0.6711790561676025,
"learning_rate": 1.3537041048046695e-05,
"loss": 0.713,
"step": 19600
},
{
"epoch": 3.1445031204992797,
"grad_norm": 0.6365526914596558,
"learning_rate": 1.32990454445166e-05,
"loss": 0.7679,
"step": 19650
},
{
"epoch": 3.1525044007041125,
"grad_norm": 0.6711440682411194,
"learning_rate": 1.3062839229902263e-05,
"loss": 0.7199,
"step": 19700
},
{
"epoch": 3.1605056809089453,
"grad_norm": 0.7575156688690186,
"learning_rate": 1.2828433920655687e-05,
"loss": 0.7052,
"step": 19750
},
{
"epoch": 3.168506961113778,
"grad_norm": 0.6248161792755127,
"learning_rate": 1.2595840945424093e-05,
"loss": 0.7129,
"step": 19800
},
{
"epoch": 3.176508241318611,
"grad_norm": 0.7342621684074402,
"learning_rate": 1.2365071644492682e-05,
"loss": 0.7187,
"step": 19850
},
{
"epoch": 3.184509521523444,
"grad_norm": 0.6555837988853455,
"learning_rate": 1.2136137269231723e-05,
"loss": 0.7378,
"step": 19900
},
{
"epoch": 3.1925108017282766,
"grad_norm": 0.7092923521995544,
"learning_rate": 1.1909048981547998e-05,
"loss": 0.7259,
"step": 19950
},
{
"epoch": 3.2005120819331094,
"grad_norm": 0.6612015962600708,
"learning_rate": 1.1683817853340595e-05,
"loss": 0.7168,
"step": 20000
},
{
"epoch": 3.2085133621379422,
"grad_norm": 0.7098072171211243,
"learning_rate": 1.1460454865961056e-05,
"loss": 0.7226,
"step": 20050
},
{
"epoch": 3.2165146423427746,
"grad_norm": 0.7153401970863342,
"learning_rate": 1.1238970909677993e-05,
"loss": 0.7304,
"step": 20100
},
{
"epoch": 3.2245159225476074,
"grad_norm": 0.6184985637664795,
"learning_rate": 1.1019376783146174e-05,
"loss": 0.7113,
"step": 20150
},
{
"epoch": 3.2325172027524403,
"grad_norm": 0.658746063709259,
"learning_rate": 1.080168319287989e-05,
"loss": 0.7545,
"step": 20200
},
{
"epoch": 3.240518482957273,
"grad_norm": 0.6574704647064209,
"learning_rate": 1.0585900752731077e-05,
"loss": 0.7103,
"step": 20250
},
{
"epoch": 3.248519763162106,
"grad_norm": 0.7436890006065369,
"learning_rate": 1.0372039983371818e-05,
"loss": 0.7335,
"step": 20300
},
{
"epoch": 3.2565210433669387,
"grad_norm": 0.7010487914085388,
"learning_rate": 1.0160111311781284e-05,
"loss": 0.7463,
"step": 20350
},
{
"epoch": 3.2645223235717715,
"grad_norm": 0.7422741651535034,
"learning_rate": 9.950125070737476e-06,
"loss": 0.7253,
"step": 20400
},
{
"epoch": 3.2725236037766043,
"grad_norm": 0.6463279724121094,
"learning_rate": 9.742091498313426e-06,
"loss": 0.7334,
"step": 20450
},
{
"epoch": 3.280524883981437,
"grad_norm": 0.7407013177871704,
"learning_rate": 9.536020737377993e-06,
"loss": 0.7104,
"step": 20500
},
{
"epoch": 3.28852616418627,
"grad_norm": 0.6367645263671875,
"learning_rate": 9.331922835101282e-06,
"loss": 0.7105,
"step": 20550
},
{
"epoch": 3.296527444391103,
"grad_norm": 0.7848844528198242,
"learning_rate": 9.129807742464957e-06,
"loss": 0.7155,
"step": 20600
},
{
"epoch": 3.304528724595935,
"grad_norm": 0.6976688504219055,
"learning_rate": 8.929685313776903e-06,
"loss": 0.7363,
"step": 20650
},
{
"epoch": 3.312530004800768,
"grad_norm": 0.7478333711624146,
"learning_rate": 8.731565306190852e-06,
"loss": 0.7612,
"step": 20700
},
{
"epoch": 3.320531285005601,
"grad_norm": 0.6891793012619019,
"learning_rate": 8.535457379230649e-06,
"loss": 0.7221,
"step": 20750
},
{
"epoch": 3.3285325652104336,
"grad_norm": 0.6285429000854492,
"learning_rate": 8.341371094319289e-06,
"loss": 0.7554,
"step": 20800
},
{
"epoch": 3.3365338454152664,
"grad_norm": 0.7216114401817322,
"learning_rate": 8.149315914312733e-06,
"loss": 0.7297,
"step": 20850
},
{
"epoch": 3.3445351256200992,
"grad_norm": 0.6681531071662903,
"learning_rate": 7.959301203038566e-06,
"loss": 0.7283,
"step": 20900
},
{
"epoch": 3.352536405824932,
"grad_norm": 0.6501981019973755,
"learning_rate": 7.771336224839425e-06,
"loss": 0.723,
"step": 20950
},
{
"epoch": 3.360537686029765,
"grad_norm": 0.7048876285552979,
"learning_rate": 7.585430144121319e-06,
"loss": 0.7239,
"step": 21000
},
{
"epoch": 3.3685389662345977,
"grad_norm": 0.6659078001976013,
"learning_rate": 7.401592024906812e-06,
"loss": 0.7417,
"step": 21050
},
{
"epoch": 3.37654024643943,
"grad_norm": 0.632502019405365,
"learning_rate": 7.219830830393093e-06,
"loss": 0.7115,
"step": 21100
},
{
"epoch": 3.384541526644263,
"grad_norm": 0.6977764964103699,
"learning_rate": 7.040155422514977e-06,
"loss": 0.6854,
"step": 21150
},
{
"epoch": 3.3925428068490957,
"grad_norm": 0.6409167051315308,
"learning_rate": 6.862574561512825e-06,
"loss": 0.7016,
"step": 21200
},
{
"epoch": 3.4005440870539285,
"grad_norm": 0.6645964980125427,
"learning_rate": 6.6870969055054246e-06,
"loss": 0.7224,
"step": 21250
},
{
"epoch": 3.4085453672587613,
"grad_norm": 0.6525962352752686,
"learning_rate": 6.51373101006787e-06,
"loss": 0.7248,
"step": 21300
},
{
"epoch": 3.416546647463594,
"grad_norm": 0.7221346497535706,
"learning_rate": 6.3424853278144015e-06,
"loss": 0.7159,
"step": 21350
},
{
"epoch": 3.424547927668427,
"grad_norm": 0.7742597460746765,
"learning_rate": 6.173368207986358e-06,
"loss": 0.7158,
"step": 21400
},
{
"epoch": 3.43254920787326,
"grad_norm": 0.7201706171035767,
"learning_rate": 6.00638789604499e-06,
"loss": 0.7351,
"step": 21450
},
{
"epoch": 3.4405504880780926,
"grad_norm": 0.6694920063018799,
"learning_rate": 5.841552533269534e-06,
"loss": 0.7313,
"step": 21500
},
{
"epoch": 3.4485517682829254,
"grad_norm": 0.773014485836029,
"learning_rate": 5.6788701563602665e-06,
"loss": 0.7526,
"step": 21550
},
{
"epoch": 3.4565530484877582,
"grad_norm": 0.602942168712616,
"learning_rate": 5.518348697046644e-06,
"loss": 0.7304,
"step": 21600
},
{
"epoch": 3.464554328692591,
"grad_norm": 0.7288883924484253,
"learning_rate": 5.359995981700544e-06,
"loss": 0.7086,
"step": 21650
},
{
"epoch": 3.4725556088974234,
"grad_norm": 0.8319910168647766,
"learning_rate": 5.203819730954806e-06,
"loss": 0.7356,
"step": 21700
},
{
"epoch": 3.4805568891022562,
"grad_norm": 0.7911546230316162,
"learning_rate": 5.049827559326686e-06,
"loss": 0.7192,
"step": 21750
},
{
"epoch": 3.488558169307089,
"grad_norm": 0.704937756061554,
"learning_rate": 4.898026974846631e-06,
"loss": 0.7209,
"step": 21800
},
{
"epoch": 3.496559449511922,
"grad_norm": 0.7878553867340088,
"learning_rate": 4.748425378692278e-06,
"loss": 0.7449,
"step": 21850
},
{
"epoch": 3.5045607297167547,
"grad_norm": 0.7807219624519348,
"learning_rate": 4.601030064827527e-06,
"loss": 0.722,
"step": 21900
},
{
"epoch": 3.5125620099215875,
"grad_norm": 0.7362565994262695,
"learning_rate": 4.455848219646957e-06,
"loss": 0.7069,
"step": 21950
},
{
"epoch": 3.5205632901264203,
"grad_norm": 0.6918688416481018,
"learning_rate": 4.3128869216254366e-06,
"loss": 0.7198,
"step": 22000
},
{
"epoch": 3.528564570331253,
"grad_norm": 0.6748279929161072,
"learning_rate": 4.172153140973012e-06,
"loss": 0.7429,
"step": 22050
},
{
"epoch": 3.5365658505360855,
"grad_norm": 0.7261555194854736,
"learning_rate": 4.033653739295062e-06,
"loss": 0.7453,
"step": 22100
},
{
"epoch": 3.5445671307409183,
"grad_norm": 0.7590264081954956,
"learning_rate": 3.897395469257759e-06,
"loss": 0.7435,
"step": 22150
},
{
"epoch": 3.552568410945751,
"grad_norm": 0.6835177540779114,
"learning_rate": 3.7633849742588213e-06,
"loss": 0.7305,
"step": 22200
},
{
"epoch": 3.560569691150584,
"grad_norm": 0.7218915224075317,
"learning_rate": 3.6316287881036306e-06,
"loss": 0.7357,
"step": 22250
},
{
"epoch": 3.568570971355417,
"grad_norm": 0.7100145220756531,
"learning_rate": 3.502133334686669e-06,
"loss": 0.7244,
"step": 22300
},
{
"epoch": 3.5765722515602496,
"grad_norm": 0.6982618570327759,
"learning_rate": 3.374904927678285e-06,
"loss": 0.7172,
"step": 22350
},
{
"epoch": 3.5845735317650824,
"grad_norm": 0.6808996200561523,
"learning_rate": 3.2499497702169035e-06,
"loss": 0.7112,
"step": 22400
},
{
"epoch": 3.5925748119699152,
"grad_norm": 0.6580341458320618,
"learning_rate": 3.1272739546065746e-06,
"loss": 0.7375,
"step": 22450
},
{
"epoch": 3.600576092174748,
"grad_norm": 0.7925476431846619,
"learning_rate": 3.0068834620199103e-06,
"loss": 0.7351,
"step": 22500
},
{
"epoch": 3.608577372379581,
"grad_norm": 0.6711053848266602,
"learning_rate": 2.888784162206504e-06,
"loss": 0.7281,
"step": 22550
},
{
"epoch": 3.6165786525844137,
"grad_norm": 0.6397190690040588,
"learning_rate": 2.7729818132067264e-06,
"loss": 0.7277,
"step": 22600
},
{
"epoch": 3.6245799327892465,
"grad_norm": 0.6673882007598877,
"learning_rate": 2.6594820610710057e-06,
"loss": 0.7045,
"step": 22650
},
{
"epoch": 3.6325812129940793,
"grad_norm": 0.6923602223396301,
"learning_rate": 2.5482904395845107e-06,
"loss": 0.7435,
"step": 22700
},
{
"epoch": 3.6405824931989117,
"grad_norm": 0.7044751048088074,
"learning_rate": 2.439412369997374e-06,
"loss": 0.7135,
"step": 22750
},
{
"epoch": 3.6485837734037445,
"grad_norm": 0.7120568156242371,
"learning_rate": 2.33285316076039e-06,
"loss": 0.7115,
"step": 22800
},
{
"epoch": 3.6565850536085773,
"grad_norm": 0.6994568705558777,
"learning_rate": 2.22861800726617e-06,
"loss": 0.7212,
"step": 22850
},
{
"epoch": 3.66458633381341,
"grad_norm": 0.7410824298858643,
"learning_rate": 2.1267119915958088e-06,
"loss": 0.6973,
"step": 22900
},
{
"epoch": 3.672587614018243,
"grad_norm": 0.7019298076629639,
"learning_rate": 2.0271400822711894e-06,
"loss": 0.724,
"step": 22950
},
{
"epoch": 3.680588894223076,
"grad_norm": 0.7187603116035461,
"learning_rate": 1.929907134012654e-06,
"loss": 0.699,
"step": 23000
},
{
"epoch": 3.6885901744279086,
"grad_norm": 0.7453562617301941,
"learning_rate": 1.8350178875023272e-06,
"loss": 0.7243,
"step": 23050
},
{
"epoch": 3.696591454632741,
"grad_norm": 0.8272750973701477,
"learning_rate": 1.7424769691530284e-06,
"loss": 0.7178,
"step": 23100
},
{
"epoch": 3.704592734837574,
"grad_norm": 0.6875506043434143,
"learning_rate": 1.6522888908826473e-06,
"loss": 0.7328,
"step": 23150
},
{
"epoch": 3.7125940150424066,
"grad_norm": 0.6446670889854431,
"learning_rate": 1.5644580498941851e-06,
"loss": 0.7121,
"step": 23200
},
{
"epoch": 3.7205952952472394,
"grad_norm": 0.6092432141304016,
"learning_rate": 1.478988728461367e-06,
"loss": 0.7045,
"step": 23250
},
{
"epoch": 3.7285965754520722,
"grad_norm": 0.6800952553749084,
"learning_rate": 1.3958850937198453e-06,
"loss": 0.7275,
"step": 23300
},
{
"epoch": 3.736597855656905,
"grad_norm": 0.684956431388855,
"learning_rate": 1.315151197464043e-06,
"loss": 0.7272,
"step": 23350
},
{
"epoch": 3.744599135861738,
"grad_norm": 0.7142546772956848,
"learning_rate": 1.236790975949592e-06,
"loss": 0.7213,
"step": 23400
},
{
"epoch": 3.7526004160665707,
"grad_norm": 0.7309266328811646,
"learning_rate": 1.1608082497014228e-06,
"loss": 0.7212,
"step": 23450
},
{
"epoch": 3.7606016962714035,
"grad_norm": 0.7835919260978699,
"learning_rate": 1.087206723327483e-06,
"loss": 0.7411,
"step": 23500
},
{
"epoch": 3.7686029764762363,
"grad_norm": 0.6577751040458679,
"learning_rate": 1.0159899853381394e-06,
"loss": 0.7104,
"step": 23550
},
{
"epoch": 3.776604256681069,
"grad_norm": 0.7286819219589233,
"learning_rate": 9.471615079711838e-07,
"loss": 0.7284,
"step": 23600
},
{
"epoch": 3.784605536885902,
"grad_norm": 0.7102928757667542,
"learning_rate": 8.807246470225517e-07,
"loss": 0.7415,
"step": 23650
},
{
"epoch": 3.7926068170907348,
"grad_norm": 0.6017107963562012,
"learning_rate": 8.166826416827422e-07,
"loss": 0.7063,
"step": 23700
},
{
"epoch": 3.800608097295567,
"grad_norm": 0.7728732824325562,
"learning_rate": 7.550386143788224e-07,
"loss": 0.6959,
"step": 23750
},
{
"epoch": 3.8086093775004,
"grad_norm": 0.7075885534286499,
"learning_rate": 6.95795570622243e-07,
"loss": 0.6977,
"step": 23800
},
{
"epoch": 3.816610657705233,
"grad_norm": 0.7628594636917114,
"learning_rate": 6.389563988622948e-07,
"loss": 0.7189,
"step": 23850
},
{
"epoch": 3.8246119379100656,
"grad_norm": 0.6857520341873169,
"learning_rate": 5.845238703452604e-07,
"loss": 0.7281,
"step": 23900
},
{
"epoch": 3.8326132181148984,
"grad_norm": 0.676717221736908,
"learning_rate": 5.325006389793053e-07,
"loss": 0.7275,
"step": 23950
},
{
"epoch": 3.8406144983197312,
"grad_norm": 0.6261276602745056,
"learning_rate": 4.828892412050978e-07,
"loss": 0.7162,
"step": 24000
},
{
"epoch": 3.848615778524564,
"grad_norm": 0.761638343334198,
"learning_rate": 4.356920958721311e-07,
"loss": 0.7405,
"step": 24050
},
{
"epoch": 3.856617058729397,
"grad_norm": 0.697512686252594,
"learning_rate": 3.909115041207889e-07,
"loss": 0.7156,
"step": 24100
},
{
"epoch": 3.8646183389342292,
"grad_norm": 0.6379271149635315,
"learning_rate": 3.4854964927015235e-07,
"loss": 0.7443,
"step": 24150
},
{
"epoch": 3.872619619139062,
"grad_norm": 0.7762777209281921,
"learning_rate": 3.086085967115626e-07,
"loss": 0.7186,
"step": 24200
},
{
"epoch": 3.880620899343895,
"grad_norm": 0.6799700260162354,
"learning_rate": 2.7109029380790186e-07,
"loss": 0.6928,
"step": 24250
},
{
"epoch": 3.8886221795487277,
"grad_norm": 0.8152979016304016,
"learning_rate": 2.3599656979866325e-07,
"loss": 0.7145,
"step": 24300
},
{
"epoch": 3.8966234597535605,
"grad_norm": 0.6055078506469727,
"learning_rate": 2.0332913571074476e-07,
"loss": 0.7105,
"step": 24350
},
{
"epoch": 3.9046247399583933,
"grad_norm": 0.7122170329093933,
"learning_rate": 1.7308958427505462e-07,
"loss": 0.7322,
"step": 24400
},
{
"epoch": 3.912626020163226,
"grad_norm": 0.7174406051635742,
"learning_rate": 1.4527938984883471e-07,
"loss": 0.7294,
"step": 24450
},
{
"epoch": 3.920627300368059,
"grad_norm": 0.7419302463531494,
"learning_rate": 1.1989990834378462e-07,
"loss": 0.7272,
"step": 24500
},
{
"epoch": 3.9286285805728918,
"grad_norm": 0.7208146452903748,
"learning_rate": 9.695237715994788e-08,
"loss": 0.7066,
"step": 24550
},
{
"epoch": 3.9366298607777246,
"grad_norm": 0.8183807730674744,
"learning_rate": 7.6437915125388e-08,
"loss": 0.721,
"step": 24600
},
{
"epoch": 3.9446311409825574,
"grad_norm": 0.7065560817718506,
"learning_rate": 5.835752244164883e-08,
"loss": 0.737,
"step": 24650
},
{
"epoch": 3.95263242118739,
"grad_norm": 0.7247999310493469,
"learning_rate": 4.2712080634949024e-08,
"loss": 0.7346,
"step": 24700
},
{
"epoch": 3.9606337013922226,
"grad_norm": 0.6564909219741821,
"learning_rate": 2.9502352513255394e-08,
"loss": 0.7305,
"step": 24750
},
{
"epoch": 3.9686349815970554,
"grad_norm": 0.7844890356063843,
"learning_rate": 1.8728982129051497e-08,
"loss": 0.7076,
"step": 24800
},
{
"epoch": 3.9766362618018882,
"grad_norm": 0.7181993722915649,
"learning_rate": 1.0392494747957227e-08,
"loss": 0.7439,
"step": 24850
},
{
"epoch": 3.984637542006721,
"grad_norm": 0.6785070300102234,
"learning_rate": 4.493296823104842e-09,
"loss": 0.7222,
"step": 24900
},
{
"epoch": 3.992638822211554,
"grad_norm": 0.6972596645355225,
"learning_rate": 1.0316759753381534e-09,
"loss": 0.7258,
"step": 24950
}
],
"logging_steps": 50,
"max_steps": 24996,
"num_input_tokens_seen": 0,
"num_train_epochs": 4,
"save_steps": 2000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 5.928976828545164e+19,
"train_batch_size": 16,
"trial_name": null,
"trial_params": null
}