zephyr-7b-gemma-sft / trainer_state.json
li-muyang's picture
Model save
33df837 verified
raw
history blame contribute delete
38.7 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 3.0,
"eval_steps": 100,
"global_step": 1026,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0029239766081871343,
"grad_norm": 6.1624870962638925,
"learning_rate": 1.941747572815534e-07,
"loss": 0.913,
"step": 1
},
{
"epoch": 0.014619883040935672,
"grad_norm": 11.812860926129314,
"learning_rate": 9.70873786407767e-07,
"loss": 0.8859,
"step": 5
},
{
"epoch": 0.029239766081871343,
"grad_norm": 2.6140064947786015,
"learning_rate": 1.941747572815534e-06,
"loss": 0.8409,
"step": 10
},
{
"epoch": 0.043859649122807015,
"grad_norm": 2.2391899324318154,
"learning_rate": 2.912621359223301e-06,
"loss": 0.8325,
"step": 15
},
{
"epoch": 0.05847953216374269,
"grad_norm": 2.11742528178852,
"learning_rate": 3.883495145631068e-06,
"loss": 0.8169,
"step": 20
},
{
"epoch": 0.07309941520467836,
"grad_norm": 2.088530291916816,
"learning_rate": 4.854368932038836e-06,
"loss": 0.8168,
"step": 25
},
{
"epoch": 0.08771929824561403,
"grad_norm": 2.028716635962563,
"learning_rate": 5.825242718446602e-06,
"loss": 0.8075,
"step": 30
},
{
"epoch": 0.1023391812865497,
"grad_norm": 2.3256620167562234,
"learning_rate": 6.79611650485437e-06,
"loss": 0.7953,
"step": 35
},
{
"epoch": 0.11695906432748537,
"grad_norm": 2.349007546868107,
"learning_rate": 7.766990291262136e-06,
"loss": 0.7841,
"step": 40
},
{
"epoch": 0.13157894736842105,
"grad_norm": 2.2423399474385666,
"learning_rate": 8.737864077669904e-06,
"loss": 0.7888,
"step": 45
},
{
"epoch": 0.14619883040935672,
"grad_norm": 1.942782247095399,
"learning_rate": 9.708737864077671e-06,
"loss": 0.8042,
"step": 50
},
{
"epoch": 0.1608187134502924,
"grad_norm": 2.1945226195890033,
"learning_rate": 1.0679611650485437e-05,
"loss": 0.8018,
"step": 55
},
{
"epoch": 0.17543859649122806,
"grad_norm": 1.8960792380527152,
"learning_rate": 1.1650485436893204e-05,
"loss": 0.7832,
"step": 60
},
{
"epoch": 0.19005847953216373,
"grad_norm": 2.050931671194075,
"learning_rate": 1.2621359223300974e-05,
"loss": 0.8048,
"step": 65
},
{
"epoch": 0.2046783625730994,
"grad_norm": 2.0914042301099043,
"learning_rate": 1.359223300970874e-05,
"loss": 0.7988,
"step": 70
},
{
"epoch": 0.21929824561403508,
"grad_norm": 2.1780402538204804,
"learning_rate": 1.4563106796116507e-05,
"loss": 0.8046,
"step": 75
},
{
"epoch": 0.23391812865497075,
"grad_norm": 2.2507384711871077,
"learning_rate": 1.5533980582524273e-05,
"loss": 0.8101,
"step": 80
},
{
"epoch": 0.24853801169590642,
"grad_norm": 2.0740107954310987,
"learning_rate": 1.650485436893204e-05,
"loss": 0.7992,
"step": 85
},
{
"epoch": 0.2631578947368421,
"grad_norm": 2.1703649782676426,
"learning_rate": 1.7475728155339808e-05,
"loss": 0.8168,
"step": 90
},
{
"epoch": 0.2777777777777778,
"grad_norm": 2.4978786575354817,
"learning_rate": 1.8446601941747575e-05,
"loss": 0.806,
"step": 95
},
{
"epoch": 0.29239766081871343,
"grad_norm": 2.0698254347384983,
"learning_rate": 1.9417475728155343e-05,
"loss": 0.8286,
"step": 100
},
{
"epoch": 0.29239766081871343,
"eval_loss": 0.8580158948898315,
"eval_runtime": 94.9126,
"eval_samples_per_second": 22.99,
"eval_steps_per_second": 0.369,
"step": 100
},
{
"epoch": 0.30701754385964913,
"grad_norm": 2.0735037748593723,
"learning_rate": 1.999976830073192e-05,
"loss": 0.8281,
"step": 105
},
{
"epoch": 0.3216374269005848,
"grad_norm": 2.095073626207817,
"learning_rate": 1.99971618072711e-05,
"loss": 0.8174,
"step": 110
},
{
"epoch": 0.3362573099415205,
"grad_norm": 2.1603082170396455,
"learning_rate": 1.9991659953668803e-05,
"loss": 0.8217,
"step": 115
},
{
"epoch": 0.3508771929824561,
"grad_norm": 2.0326092354167637,
"learning_rate": 1.998326433336204e-05,
"loss": 0.8381,
"step": 120
},
{
"epoch": 0.3654970760233918,
"grad_norm": 2.226630427942925,
"learning_rate": 1.997197737787537e-05,
"loss": 0.8261,
"step": 125
},
{
"epoch": 0.38011695906432746,
"grad_norm": 2.119406491731454,
"learning_rate": 1.9957802356116665e-05,
"loss": 0.8208,
"step": 130
},
{
"epoch": 0.39473684210526316,
"grad_norm": 2.8253278464362084,
"learning_rate": 1.994074337343041e-05,
"loss": 0.8129,
"step": 135
},
{
"epoch": 0.4093567251461988,
"grad_norm": 2.4882232695538464,
"learning_rate": 1.9920805370408695e-05,
"loss": 0.8135,
"step": 140
},
{
"epoch": 0.4239766081871345,
"grad_norm": 4.780582760365185,
"learning_rate": 1.9897994121460322e-05,
"loss": 0.8364,
"step": 145
},
{
"epoch": 0.43859649122807015,
"grad_norm": 11.5182408239238,
"learning_rate": 1.9872316233138463e-05,
"loss": 0.8277,
"step": 150
},
{
"epoch": 0.45321637426900585,
"grad_norm": 12.627096663789404,
"learning_rate": 1.9843779142227258e-05,
"loss": 0.8196,
"step": 155
},
{
"epoch": 0.4678362573099415,
"grad_norm": 2.740111270404608,
"learning_rate": 1.981239111358799e-05,
"loss": 0.8042,
"step": 160
},
{
"epoch": 0.4824561403508772,
"grad_norm": 2.174778620165311,
"learning_rate": 1.9778161237765438e-05,
"loss": 0.8064,
"step": 165
},
{
"epoch": 0.49707602339181284,
"grad_norm": 1.9246861870071321,
"learning_rate": 1.9741099428355075e-05,
"loss": 0.8024,
"step": 170
},
{
"epoch": 0.5116959064327485,
"grad_norm": 9.798315852556506,
"learning_rate": 1.9701216419131934e-05,
"loss": 0.8352,
"step": 175
},
{
"epoch": 0.5263157894736842,
"grad_norm": 2.9139360459829082,
"learning_rate": 1.9658523760941882e-05,
"loss": 0.8033,
"step": 180
},
{
"epoch": 0.5409356725146199,
"grad_norm": 14.543721579870974,
"learning_rate": 1.9613033818356322e-05,
"loss": 0.8631,
"step": 185
},
{
"epoch": 0.5555555555555556,
"grad_norm": 11.532560091003726,
"learning_rate": 1.956475976609114e-05,
"loss": 0.8304,
"step": 190
},
{
"epoch": 0.5701754385964912,
"grad_norm": 10.191690209836944,
"learning_rate": 1.951371558519111e-05,
"loss": 0.8021,
"step": 195
},
{
"epoch": 0.5847953216374269,
"grad_norm": 2.490323350347935,
"learning_rate": 1.9459916058980707e-05,
"loss": 0.798,
"step": 200
},
{
"epoch": 0.5847953216374269,
"eval_loss": 0.8706345558166504,
"eval_runtime": 94.8283,
"eval_samples_per_second": 23.01,
"eval_steps_per_second": 0.369,
"step": 200
},
{
"epoch": 0.5994152046783626,
"grad_norm": 2.025711297854962,
"learning_rate": 1.9403376768782592e-05,
"loss": 0.7884,
"step": 205
},
{
"epoch": 0.6140350877192983,
"grad_norm": 1.9488358870140705,
"learning_rate": 1.9344114089404983e-05,
"loss": 0.7804,
"step": 210
},
{
"epoch": 0.6286549707602339,
"grad_norm": 1.8564140890417253,
"learning_rate": 1.9282145184399197e-05,
"loss": 0.808,
"step": 215
},
{
"epoch": 0.6432748538011696,
"grad_norm": 2.0881170345282727,
"learning_rate": 1.9217488001088784e-05,
"loss": 0.799,
"step": 220
},
{
"epoch": 0.6578947368421053,
"grad_norm": 1.9678597098298427,
"learning_rate": 1.9150161265371663e-05,
"loss": 0.795,
"step": 225
},
{
"epoch": 0.672514619883041,
"grad_norm": 2.865460519803586,
"learning_rate": 1.908018447629674e-05,
"loss": 0.7915,
"step": 230
},
{
"epoch": 0.6871345029239766,
"grad_norm": 2.398699191637217,
"learning_rate": 1.9007577900416648e-05,
"loss": 0.7814,
"step": 235
},
{
"epoch": 0.7017543859649122,
"grad_norm": 2.1260051093101353,
"learning_rate": 1.8932362565918167e-05,
"loss": 0.7983,
"step": 240
},
{
"epoch": 0.716374269005848,
"grad_norm": 5.80073712189035,
"learning_rate": 1.8854560256532098e-05,
"loss": 0.7691,
"step": 245
},
{
"epoch": 0.7309941520467836,
"grad_norm": 2.5341535901655896,
"learning_rate": 1.877419350522429e-05,
"loss": 0.8155,
"step": 250
},
{
"epoch": 0.7456140350877193,
"grad_norm": 1.95626689403789,
"learning_rate": 1.869128558766965e-05,
"loss": 0.7878,
"step": 255
},
{
"epoch": 0.7602339181286549,
"grad_norm": 15.120401535093198,
"learning_rate": 1.8605860515511158e-05,
"loss": 0.7883,
"step": 260
},
{
"epoch": 0.7748538011695907,
"grad_norm": 2.2607309079656135,
"learning_rate": 1.8517943029405577e-05,
"loss": 0.7999,
"step": 265
},
{
"epoch": 0.7894736842105263,
"grad_norm": 2.295750513860874,
"learning_rate": 1.8427558591858164e-05,
"loss": 0.7697,
"step": 270
},
{
"epoch": 0.804093567251462,
"grad_norm": 4.1135088544849445,
"learning_rate": 1.833473337984823e-05,
"loss": 0.8065,
"step": 275
},
{
"epoch": 0.8187134502923976,
"grad_norm": 3.2241188905244895,
"learning_rate": 1.823949427724785e-05,
"loss": 0.799,
"step": 280
},
{
"epoch": 0.8333333333333334,
"grad_norm": 1.6843445283218286,
"learning_rate": 1.8141868867035745e-05,
"loss": 0.7909,
"step": 285
},
{
"epoch": 0.847953216374269,
"grad_norm": 1.7626650921963332,
"learning_rate": 1.8041885423308808e-05,
"loss": 0.781,
"step": 290
},
{
"epoch": 0.8625730994152047,
"grad_norm": 10.38812175281534,
"learning_rate": 1.7939572903093383e-05,
"loss": 0.7923,
"step": 295
},
{
"epoch": 0.8771929824561403,
"grad_norm": 3.1451964354292596,
"learning_rate": 1.7834960937958775e-05,
"loss": 0.7677,
"step": 300
},
{
"epoch": 0.8771929824561403,
"eval_loss": 0.8726791739463806,
"eval_runtime": 93.9416,
"eval_samples_per_second": 23.227,
"eval_steps_per_second": 0.373,
"step": 300
},
{
"epoch": 0.8918128654970761,
"grad_norm": 1.9970765983454766,
"learning_rate": 1.7728079825435426e-05,
"loss": 0.7966,
"step": 305
},
{
"epoch": 0.9064327485380117,
"grad_norm": 1.8313570465257998,
"learning_rate": 1.761896052024019e-05,
"loss": 0.7784,
"step": 310
},
{
"epoch": 0.9210526315789473,
"grad_norm": 2.028945141590537,
"learning_rate": 1.7507634625311283e-05,
"loss": 0.7569,
"step": 315
},
{
"epoch": 0.935672514619883,
"grad_norm": 3.1385176933971706,
"learning_rate": 1.7394134382655496e-05,
"loss": 0.7805,
"step": 320
},
{
"epoch": 0.9502923976608187,
"grad_norm": 1.7723196199368927,
"learning_rate": 1.7278492664010326e-05,
"loss": 0.7681,
"step": 325
},
{
"epoch": 0.9649122807017544,
"grad_norm": 1.7577488585231846,
"learning_rate": 1.7160742961323714e-05,
"loss": 0.7672,
"step": 330
},
{
"epoch": 0.97953216374269,
"grad_norm": 2.3636063665348757,
"learning_rate": 1.7040919377054182e-05,
"loss": 0.7795,
"step": 335
},
{
"epoch": 0.9941520467836257,
"grad_norm": 1.7709587208644213,
"learning_rate": 1.6919056614294133e-05,
"loss": 0.7624,
"step": 340
},
{
"epoch": 1.0087719298245614,
"grad_norm": 3.8073195101773765,
"learning_rate": 1.6795189966719182e-05,
"loss": 0.6426,
"step": 345
},
{
"epoch": 1.023391812865497,
"grad_norm": 2.787413972058394,
"learning_rate": 1.666935530836651e-05,
"loss": 0.5366,
"step": 350
},
{
"epoch": 1.0380116959064327,
"grad_norm": 2.362170165993678,
"learning_rate": 1.654158908324504e-05,
"loss": 0.5377,
"step": 355
},
{
"epoch": 1.0526315789473684,
"grad_norm": 2.311139734884461,
"learning_rate": 1.6411928294780626e-05,
"loss": 0.5252,
"step": 360
},
{
"epoch": 1.067251461988304,
"grad_norm": 1.886267380269756,
"learning_rate": 1.6280410495099165e-05,
"loss": 0.5028,
"step": 365
},
{
"epoch": 1.0818713450292399,
"grad_norm": 1.8701236760118256,
"learning_rate": 1.6147073774150834e-05,
"loss": 0.5121,
"step": 370
},
{
"epoch": 1.0964912280701755,
"grad_norm": 1.7743194110147134,
"learning_rate": 1.601195674867853e-05,
"loss": 0.5189,
"step": 375
},
{
"epoch": 1.1111111111111112,
"grad_norm": 36.55479744540479,
"learning_rate": 1.5875098551033765e-05,
"loss": 0.5071,
"step": 380
},
{
"epoch": 1.1257309941520468,
"grad_norm": 2.1840870340337823,
"learning_rate": 1.5736538817843228e-05,
"loss": 0.5489,
"step": 385
},
{
"epoch": 1.1403508771929824,
"grad_norm": 2.096047543627006,
"learning_rate": 1.559631767852929e-05,
"loss": 0.5115,
"step": 390
},
{
"epoch": 1.154970760233918,
"grad_norm": 8.700584160621043,
"learning_rate": 1.54544757436878e-05,
"loss": 0.5061,
"step": 395
},
{
"epoch": 1.1695906432748537,
"grad_norm": 1.7774756036197978,
"learning_rate": 1.5311054093326508e-05,
"loss": 0.5584,
"step": 400
},
{
"epoch": 1.1695906432748537,
"eval_loss": 0.9159147143363953,
"eval_runtime": 94.2366,
"eval_samples_per_second": 23.154,
"eval_steps_per_second": 0.371,
"step": 400
},
{
"epoch": 1.1842105263157894,
"grad_norm": 2.550427212943654,
"learning_rate": 1.5166094264967553e-05,
"loss": 0.5166,
"step": 405
},
{
"epoch": 1.198830409356725,
"grad_norm": 1.9907997528305226,
"learning_rate": 1.5019638241617429e-05,
"loss": 0.5131,
"step": 410
},
{
"epoch": 1.213450292397661,
"grad_norm": 1.7648223360572315,
"learning_rate": 1.4871728439607967e-05,
"loss": 0.5182,
"step": 415
},
{
"epoch": 1.2280701754385965,
"grad_norm": 2.23899567188947,
"learning_rate": 1.472240769631176e-05,
"loss": 0.5062,
"step": 420
},
{
"epoch": 1.2426900584795322,
"grad_norm": 1.763198949927548,
"learning_rate": 1.4571719257735702e-05,
"loss": 0.5203,
"step": 425
},
{
"epoch": 1.2573099415204678,
"grad_norm": 4.030641844892038,
"learning_rate": 1.4419706765996153e-05,
"loss": 0.5176,
"step": 430
},
{
"epoch": 1.2719298245614035,
"grad_norm": 1.8591284929042133,
"learning_rate": 1.4266414246679379e-05,
"loss": 0.5099,
"step": 435
},
{
"epoch": 1.286549707602339,
"grad_norm": 1.8559628967021533,
"learning_rate": 1.4111886096090953e-05,
"loss": 0.5129,
"step": 440
},
{
"epoch": 1.301169590643275,
"grad_norm": 2.862306979981545,
"learning_rate": 1.3956167068397756e-05,
"loss": 0.5075,
"step": 445
},
{
"epoch": 1.3157894736842106,
"grad_norm": 1.8928522544167012,
"learning_rate": 1.3799302262666388e-05,
"loss": 0.5177,
"step": 450
},
{
"epoch": 1.3304093567251463,
"grad_norm": 1.8170476120999561,
"learning_rate": 1.364133710980162e-05,
"loss": 0.516,
"step": 455
},
{
"epoch": 1.345029239766082,
"grad_norm": 1.735070100858528,
"learning_rate": 1.3482317359388806e-05,
"loss": 0.5086,
"step": 460
},
{
"epoch": 1.3596491228070176,
"grad_norm": 2.0140408558653737,
"learning_rate": 1.3322289066443947e-05,
"loss": 0.5169,
"step": 465
},
{
"epoch": 1.3742690058479532,
"grad_norm": 1.7311234684729015,
"learning_rate": 1.316129857807534e-05,
"loss": 0.5077,
"step": 470
},
{
"epoch": 1.3888888888888888,
"grad_norm": 1.7834637989309219,
"learning_rate": 1.2999392520060594e-05,
"loss": 0.5181,
"step": 475
},
{
"epoch": 1.4035087719298245,
"grad_norm": 1.725680916190487,
"learning_rate": 1.2836617783342968e-05,
"loss": 0.5102,
"step": 480
},
{
"epoch": 1.4181286549707601,
"grad_norm": 1.8726265727670828,
"learning_rate": 1.2673021510450893e-05,
"loss": 0.5169,
"step": 485
},
{
"epoch": 1.4327485380116958,
"grad_norm": 2.056653138512768,
"learning_rate": 1.250865108184464e-05,
"loss": 0.5085,
"step": 490
},
{
"epoch": 1.4473684210526316,
"grad_norm": 1.6627221599390933,
"learning_rate": 1.2343554102194073e-05,
"loss": 0.4981,
"step": 495
},
{
"epoch": 1.4619883040935673,
"grad_norm": 1.7476089665186805,
"learning_rate": 1.2177778386591475e-05,
"loss": 0.5186,
"step": 500
},
{
"epoch": 1.4619883040935673,
"eval_loss": 0.8920474648475647,
"eval_runtime": 93.9752,
"eval_samples_per_second": 23.219,
"eval_steps_per_second": 0.372,
"step": 500
},
{
"epoch": 1.476608187134503,
"grad_norm": 1.6946408040860308,
"learning_rate": 1.2011371946703416e-05,
"loss": 0.5197,
"step": 505
},
{
"epoch": 1.4912280701754386,
"grad_norm": 1.7569189927952173,
"learning_rate": 1.1844382976865714e-05,
"loss": 0.515,
"step": 510
},
{
"epoch": 1.5058479532163744,
"grad_norm": 1.9990864745103767,
"learning_rate": 1.1676859840125468e-05,
"loss": 0.5188,
"step": 515
},
{
"epoch": 1.52046783625731,
"grad_norm": 1.659671819522036,
"learning_rate": 1.1508851054234236e-05,
"loss": 0.4952,
"step": 520
},
{
"epoch": 1.5350877192982457,
"grad_norm": 1.8125623588878788,
"learning_rate": 1.1340405277596427e-05,
"loss": 0.5029,
"step": 525
},
{
"epoch": 1.5497076023391814,
"grad_norm": 1.6209359235184773,
"learning_rate": 1.1171571295176915e-05,
"loss": 0.5016,
"step": 530
},
{
"epoch": 1.564327485380117,
"grad_norm": 1.7381191339168205,
"learning_rate": 1.1002398004372048e-05,
"loss": 0.4938,
"step": 535
},
{
"epoch": 1.5789473684210527,
"grad_norm": 1.7440328988243268,
"learning_rate": 1.0832934400848063e-05,
"loss": 0.508,
"step": 540
},
{
"epoch": 1.5935672514619883,
"grad_norm": 1.6873580611700445,
"learning_rate": 1.066322956435104e-05,
"loss": 0.4994,
"step": 545
},
{
"epoch": 1.608187134502924,
"grad_norm": 1.6955737592784719,
"learning_rate": 1.0493332644492534e-05,
"loss": 0.494,
"step": 550
},
{
"epoch": 1.6228070175438596,
"grad_norm": 1.8639586363579819,
"learning_rate": 1.0323292846514927e-05,
"loss": 0.5086,
"step": 555
},
{
"epoch": 1.6374269005847952,
"grad_norm": 1.6602403945581443,
"learning_rate": 1.015315941704071e-05,
"loss": 0.4958,
"step": 560
},
{
"epoch": 1.6520467836257309,
"grad_norm": 1.6463228018665035,
"learning_rate": 9.982981629809776e-06,
"loss": 0.4947,
"step": 565
},
{
"epoch": 1.6666666666666665,
"grad_norm": 1.8439520823673021,
"learning_rate": 9.812808771408804e-06,
"loss": 0.4943,
"step": 570
},
{
"epoch": 1.6812865497076024,
"grad_norm": 1.6567753225002235,
"learning_rate": 9.64269012699702e-06,
"loss": 0.4815,
"step": 575
},
{
"epoch": 1.695906432748538,
"grad_norm": 1.6625638574779524,
"learning_rate": 9.472674966032276e-06,
"loss": 0.5006,
"step": 580
},
{
"epoch": 1.7105263157894737,
"grad_norm": 1.898719426059897,
"learning_rate": 9.302812528001741e-06,
"loss": 0.5086,
"step": 585
},
{
"epoch": 1.7251461988304093,
"grad_norm": 1.8795535435362318,
"learning_rate": 9.133152008161235e-06,
"loss": 0.5094,
"step": 590
},
{
"epoch": 1.7397660818713452,
"grad_norm": 1.7067878022572074,
"learning_rate": 8.963742543287397e-06,
"loss": 0.5001,
"step": 595
},
{
"epoch": 1.7543859649122808,
"grad_norm": 1.6684322823190847,
"learning_rate": 8.79463319744677e-06,
"loss": 0.5067,
"step": 600
},
{
"epoch": 1.7543859649122808,
"eval_loss": 0.8780078291893005,
"eval_runtime": 94.1156,
"eval_samples_per_second": 23.184,
"eval_steps_per_second": 0.372,
"step": 600
},
{
"epoch": 1.7690058479532165,
"grad_norm": 1.575226401438412,
"learning_rate": 8.625872947785968e-06,
"loss": 0.5163,
"step": 605
},
{
"epoch": 1.7836257309941521,
"grad_norm": 1.665839831059325,
"learning_rate": 8.457510670346976e-06,
"loss": 0.5185,
"step": 610
},
{
"epoch": 1.7982456140350878,
"grad_norm": 1.668476308962127,
"learning_rate": 8.2895951259118e-06,
"loss": 0.4891,
"step": 615
},
{
"epoch": 1.8128654970760234,
"grad_norm": 1.6099746797995087,
"learning_rate": 8.122174945880409e-06,
"loss": 0.4941,
"step": 620
},
{
"epoch": 1.827485380116959,
"grad_norm": 1.5972670521003787,
"learning_rate": 7.955298618186227e-06,
"loss": 0.4983,
"step": 625
},
{
"epoch": 1.8421052631578947,
"grad_norm": 1.6567002288888941,
"learning_rate": 7.78901447325314e-06,
"loss": 0.4923,
"step": 630
},
{
"epoch": 1.8567251461988303,
"grad_norm": 1.633625117649634,
"learning_rate": 7.623370669998115e-06,
"loss": 0.4952,
"step": 635
},
{
"epoch": 1.871345029239766,
"grad_norm": 1.6702231621007573,
"learning_rate": 7.458415181883506e-06,
"loss": 0.4781,
"step": 640
},
{
"epoch": 1.8859649122807016,
"grad_norm": 1.5695315130710974,
"learning_rate": 7.294195783023072e-06,
"loss": 0.4772,
"step": 645
},
{
"epoch": 1.9005847953216373,
"grad_norm": 1.5932464772328656,
"learning_rate": 7.130760034345708e-06,
"loss": 0.4766,
"step": 650
},
{
"epoch": 1.9152046783625731,
"grad_norm": 1.5938375889822713,
"learning_rate": 6.968155269820951e-06,
"loss": 0.4954,
"step": 655
},
{
"epoch": 1.9298245614035088,
"grad_norm": 1.6068162124707588,
"learning_rate": 6.806428582750191e-06,
"loss": 0.5031,
"step": 660
},
{
"epoch": 1.9444444444444444,
"grad_norm": 1.6369099221197159,
"learning_rate": 6.645626812127588e-06,
"loss": 0.4871,
"step": 665
},
{
"epoch": 1.95906432748538,
"grad_norm": 1.8488127966750827,
"learning_rate": 6.485796529074662e-06,
"loss": 0.4793,
"step": 670
},
{
"epoch": 1.973684210526316,
"grad_norm": 1.5795572120040184,
"learning_rate": 6.326984023352435e-06,
"loss": 0.4728,
"step": 675
},
{
"epoch": 1.9883040935672516,
"grad_norm": 1.8151786682783004,
"learning_rate": 6.169235289955073e-06,
"loss": 0.4659,
"step": 680
},
{
"epoch": 2.002923976608187,
"grad_norm": 3.8300611957194977,
"learning_rate": 6.012596015788903e-06,
"loss": 0.4261,
"step": 685
},
{
"epoch": 2.017543859649123,
"grad_norm": 2.621763538813387,
"learning_rate": 5.8571115664406655e-06,
"loss": 0.2426,
"step": 690
},
{
"epoch": 2.0321637426900585,
"grad_norm": 1.8343232864531938,
"learning_rate": 5.702826973038776e-06,
"loss": 0.24,
"step": 695
},
{
"epoch": 2.046783625730994,
"grad_norm": 1.9497619965124309,
"learning_rate": 5.549786919211532e-06,
"loss": 0.2223,
"step": 700
},
{
"epoch": 2.046783625730994,
"eval_loss": 0.951133131980896,
"eval_runtime": 93.9878,
"eval_samples_per_second": 23.216,
"eval_steps_per_second": 0.372,
"step": 700
},
{
"epoch": 2.06140350877193,
"grad_norm": 1.7439743078429233,
"learning_rate": 5.3980357281459e-06,
"loss": 0.229,
"step": 705
},
{
"epoch": 2.0760233918128654,
"grad_norm": 1.5354726993504848,
"learning_rate": 5.247617349750717e-06,
"loss": 0.2083,
"step": 710
},
{
"epoch": 2.090643274853801,
"grad_norm": 7.788063462658685,
"learning_rate": 5.0985753479279824e-06,
"loss": 0.2391,
"step": 715
},
{
"epoch": 2.1052631578947367,
"grad_norm": 1.5764797068116867,
"learning_rate": 4.950952887955992e-06,
"loss": 0.2146,
"step": 720
},
{
"epoch": 2.1198830409356724,
"grad_norm": 1.5338239957853983,
"learning_rate": 4.80479272398786e-06,
"loss": 0.2145,
"step": 725
},
{
"epoch": 2.134502923976608,
"grad_norm": 1.5400470674701994,
"learning_rate": 4.660137186669131e-06,
"loss": 0.2131,
"step": 730
},
{
"epoch": 2.1491228070175437,
"grad_norm": 1.5891715585976705,
"learning_rate": 4.5170281708780865e-06,
"loss": 0.2273,
"step": 735
},
{
"epoch": 2.1637426900584797,
"grad_norm": 1.6323196857228797,
"learning_rate": 4.375507123592194e-06,
"loss": 0.2221,
"step": 740
},
{
"epoch": 2.1783625730994154,
"grad_norm": 3.4202645186771896,
"learning_rate": 4.235615031884326e-06,
"loss": 0.2139,
"step": 745
},
{
"epoch": 2.192982456140351,
"grad_norm": 1.6425009664630226,
"learning_rate": 4.097392411052149e-06,
"loss": 0.2165,
"step": 750
},
{
"epoch": 2.2076023391812867,
"grad_norm": 1.5960953283192305,
"learning_rate": 3.9608792928841596e-06,
"loss": 0.201,
"step": 755
},
{
"epoch": 2.2222222222222223,
"grad_norm": 1.5527267754127971,
"learning_rate": 3.826115214065739e-06,
"loss": 0.2162,
"step": 760
},
{
"epoch": 2.236842105263158,
"grad_norm": 1.5345330165293922,
"learning_rate": 3.693139204728623e-06,
"loss": 0.2033,
"step": 765
},
{
"epoch": 2.2514619883040936,
"grad_norm": 1.5602344960819443,
"learning_rate": 3.561989777147059e-06,
"loss": 0.2175,
"step": 770
},
{
"epoch": 2.2660818713450293,
"grad_norm": 1.543288744735399,
"learning_rate": 3.4327049145839496e-06,
"loss": 0.2044,
"step": 775
},
{
"epoch": 2.280701754385965,
"grad_norm": 1.5668680143830462,
"learning_rate": 3.3053220602902057e-06,
"loss": 0.2067,
"step": 780
},
{
"epoch": 2.2953216374269005,
"grad_norm": 1.4653721970803488,
"learning_rate": 3.1798781066605076e-06,
"loss": 0.2072,
"step": 785
},
{
"epoch": 2.309941520467836,
"grad_norm": 1.5895314411422212,
"learning_rate": 3.056409384548575e-06,
"loss": 0.2077,
"step": 790
},
{
"epoch": 2.324561403508772,
"grad_norm": 1.5324432023481178,
"learning_rate": 2.934951652745123e-06,
"loss": 0.2067,
"step": 795
},
{
"epoch": 2.3391812865497075,
"grad_norm": 1.6877695775778203,
"learning_rate": 2.8155400876214365e-06,
"loss": 0.2114,
"step": 800
},
{
"epoch": 2.3391812865497075,
"eval_loss": 0.9491869807243347,
"eval_runtime": 94.0592,
"eval_samples_per_second": 23.198,
"eval_steps_per_second": 0.372,
"step": 800
},
{
"epoch": 2.353801169590643,
"grad_norm": 1.797333972810049,
"learning_rate": 2.698209272941659e-06,
"loss": 0.2066,
"step": 805
},
{
"epoch": 2.3684210526315788,
"grad_norm": 1.4817024828655516,
"learning_rate": 2.5829931898467143e-06,
"loss": 0.2003,
"step": 810
},
{
"epoch": 2.383040935672515,
"grad_norm": 1.5944694035375513,
"learning_rate": 2.469925207012741e-06,
"loss": 0.2111,
"step": 815
},
{
"epoch": 2.39766081871345,
"grad_norm": 1.5437718528241764,
"learning_rate": 2.3590380709869175e-06,
"loss": 0.2036,
"step": 820
},
{
"epoch": 2.412280701754386,
"grad_norm": 1.5545155179428847,
"learning_rate": 2.2503638967034668e-06,
"loss": 0.2119,
"step": 825
},
{
"epoch": 2.426900584795322,
"grad_norm": 1.5090412576695187,
"learning_rate": 2.1439341581825855e-06,
"loss": 0.2079,
"step": 830
},
{
"epoch": 2.4415204678362574,
"grad_norm": 1.5651792272296057,
"learning_rate": 2.039779679414996e-06,
"loss": 0.2049,
"step": 835
},
{
"epoch": 2.456140350877193,
"grad_norm": 1.629599771732464,
"learning_rate": 1.9379306254347487e-06,
"loss": 0.2063,
"step": 840
},
{
"epoch": 2.4707602339181287,
"grad_norm": 1.517424552525727,
"learning_rate": 1.838416493582893e-06,
"loss": 0.2054,
"step": 845
},
{
"epoch": 2.4853801169590644,
"grad_norm": 1.5665838306665794,
"learning_rate": 1.7412661049645097e-06,
"loss": 0.2072,
"step": 850
},
{
"epoch": 2.5,
"grad_norm": 1.5709890486248266,
"learning_rate": 1.6465075961015697e-06,
"loss": 0.1969,
"step": 855
},
{
"epoch": 2.5146198830409356,
"grad_norm": 1.433926747309705,
"learning_rate": 1.554168410784117e-06,
"loss": 0.2025,
"step": 860
},
{
"epoch": 2.5292397660818713,
"grad_norm": 1.56024876676377,
"learning_rate": 1.4642752921220272e-06,
"loss": 0.2019,
"step": 865
},
{
"epoch": 2.543859649122807,
"grad_norm": 1.4917168419123434,
"learning_rate": 1.3768542747997215e-06,
"loss": 0.2092,
"step": 870
},
{
"epoch": 2.5584795321637426,
"grad_norm": 1.5176036989413515,
"learning_rate": 1.2919306775360495e-06,
"loss": 0.2053,
"step": 875
},
{
"epoch": 2.573099415204678,
"grad_norm": 1.5727527454387282,
"learning_rate": 1.209529095751527e-06,
"loss": 0.2014,
"step": 880
},
{
"epoch": 2.587719298245614,
"grad_norm": 1.489918411889695,
"learning_rate": 1.1296733944450445e-06,
"loss": 0.2012,
"step": 885
},
{
"epoch": 2.60233918128655,
"grad_norm": 1.5532738891332698,
"learning_rate": 1.0523867012821444e-06,
"loss": 0.2031,
"step": 890
},
{
"epoch": 2.616959064327485,
"grad_norm": 1.4483198492727338,
"learning_rate": 9.776913998968196e-07,
"loss": 0.2012,
"step": 895
},
{
"epoch": 2.6315789473684212,
"grad_norm": 1.5331083693362297,
"learning_rate": 9.056091234088038e-07,
"loss": 0.1983,
"step": 900
},
{
"epoch": 2.6315789473684212,
"eval_loss": 0.9508408904075623,
"eval_runtime": 93.8904,
"eval_samples_per_second": 23.24,
"eval_steps_per_second": 0.373,
"step": 900
},
{
"epoch": 2.646198830409357,
"grad_norm": 1.4959941578534954,
"learning_rate": 8.361607481582312e-07,
"loss": 0.2031,
"step": 905
},
{
"epoch": 2.6608187134502925,
"grad_norm": 1.533846604598029,
"learning_rate": 7.693663876594648e-07,
"loss": 0.2086,
"step": 910
},
{
"epoch": 2.675438596491228,
"grad_norm": 1.4882252878887594,
"learning_rate": 7.052453867758525e-07,
"loss": 0.2044,
"step": 915
},
{
"epoch": 2.690058479532164,
"grad_norm": 1.463561383109587,
"learning_rate": 6.438163161171096e-07,
"loss": 0.199,
"step": 920
},
{
"epoch": 2.7046783625730995,
"grad_norm": 1.4904086029449222,
"learning_rate": 5.850969666609363e-07,
"loss": 0.1949,
"step": 925
},
{
"epoch": 2.719298245614035,
"grad_norm": 1.4484608706071034,
"learning_rate": 5.291043446004074e-07,
"loss": 0.1941,
"step": 930
},
{
"epoch": 2.7339181286549707,
"grad_norm": 1.4693647586602272,
"learning_rate": 4.7585466641868696e-07,
"loss": 0.1925,
"step": 935
},
{
"epoch": 2.7485380116959064,
"grad_norm": 1.4804496226260073,
"learning_rate": 4.25363354192434e-07,
"loss": 0.2027,
"step": 940
},
{
"epoch": 2.763157894736842,
"grad_norm": 1.5177590267070082,
"learning_rate": 3.776450311252866e-07,
"loss": 0.2068,
"step": 945
},
{
"epoch": 2.7777777777777777,
"grad_norm": 1.4962603716375011,
"learning_rate": 3.3271351731271717e-07,
"loss": 0.1993,
"step": 950
},
{
"epoch": 2.7923976608187133,
"grad_norm": 1.53541672293609,
"learning_rate": 2.905818257394799e-07,
"loss": 0.2126,
"step": 955
},
{
"epoch": 2.807017543859649,
"grad_norm": 1.4669550373356948,
"learning_rate": 2.512621585108155e-07,
"loss": 0.2021,
"step": 960
},
{
"epoch": 2.821637426900585,
"grad_norm": 1.4475885818879743,
"learning_rate": 2.1476590331849566e-07,
"loss": 0.195,
"step": 965
},
{
"epoch": 2.8362573099415203,
"grad_norm": 1.487795902053431,
"learning_rate": 1.811036301427449e-07,
"loss": 0.2073,
"step": 970
},
{
"epoch": 2.8508771929824563,
"grad_norm": 1.4972431353793751,
"learning_rate": 1.502850881909801e-07,
"loss": 0.2017,
"step": 975
},
{
"epoch": 2.8654970760233915,
"grad_norm": 1.506977096918749,
"learning_rate": 1.2231920307425927e-07,
"loss": 0.2081,
"step": 980
},
{
"epoch": 2.8801169590643276,
"grad_norm": 1.5752046829905806,
"learning_rate": 9.721407422226492e-08,
"loss": 0.2089,
"step": 985
},
{
"epoch": 2.8947368421052633,
"grad_norm": 1.4564645802512894,
"learning_rate": 7.497697253756265e-08,
"loss": 0.1933,
"step": 990
},
{
"epoch": 2.909356725146199,
"grad_norm": 1.493498674859337,
"learning_rate": 5.5614338289812216e-08,
"loss": 0.1974,
"step": 995
},
{
"epoch": 2.9239766081871346,
"grad_norm": 1.4762567685805037,
"learning_rate": 3.913177925055189e-08,
"loss": 0.2014,
"step": 1000
},
{
"epoch": 2.9239766081871346,
"eval_loss": 0.946516215801239,
"eval_runtime": 94.0428,
"eval_samples_per_second": 23.202,
"eval_steps_per_second": 0.372,
"step": 1000
},
{
"epoch": 2.93859649122807,
"grad_norm": 1.47647651184245,
"learning_rate": 2.5534069069081957e-08,
"loss": 0.2044,
"step": 1005
},
{
"epoch": 2.953216374269006,
"grad_norm": 1.4575732503367402,
"learning_rate": 1.482514588993067e-08,
"loss": 0.2001,
"step": 1010
},
{
"epoch": 2.9678362573099415,
"grad_norm": 1.5453006830413416,
"learning_rate": 7.0081112122966086e-09,
"loss": 0.2042,
"step": 1015
},
{
"epoch": 2.982456140350877,
"grad_norm": 1.7014593329310401,
"learning_rate": 2.0852289917971947e-09,
"loss": 0.1995,
"step": 1020
},
{
"epoch": 2.997076023391813,
"grad_norm": 1.5593115324810627,
"learning_rate": 5.792498478651709e-11,
"loss": 0.2055,
"step": 1025
},
{
"epoch": 3.0,
"step": 1026,
"total_flos": 429437673799680.0,
"train_loss": 0.5057662985420134,
"train_runtime": 26805.8841,
"train_samples_per_second": 4.896,
"train_steps_per_second": 0.038
}
],
"logging_steps": 5,
"max_steps": 1026,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 100,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 429437673799680.0,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}