satyanshu404's picture
Model save
e319cfe verified
raw
history blame contribute delete
No virus
75 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.0,
"eval_steps": 500,
"global_step": 9013,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0022190169754798626,
"grad_norm": 7.625,
"learning_rate": 3.3277870216306157e-07,
"loss": 1.2617,
"step": 20
},
{
"epoch": 0.004438033950959725,
"grad_norm": 5.5625,
"learning_rate": 6.655574043261231e-07,
"loss": 1.2121,
"step": 40
},
{
"epoch": 0.006657050926439588,
"grad_norm": 4.78125,
"learning_rate": 9.983361064891848e-07,
"loss": 1.2737,
"step": 60
},
{
"epoch": 0.00887606790191945,
"grad_norm": 5.46875,
"learning_rate": 1.3311148086522463e-06,
"loss": 1.1634,
"step": 80
},
{
"epoch": 0.011095084877399313,
"grad_norm": 6.46875,
"learning_rate": 1.6638935108153078e-06,
"loss": 1.1582,
"step": 100
},
{
"epoch": 0.013314101852879175,
"grad_norm": 6.09375,
"learning_rate": 1.9966722129783695e-06,
"loss": 1.1071,
"step": 120
},
{
"epoch": 0.015533118828359036,
"grad_norm": 5.5625,
"learning_rate": 2.329450915141431e-06,
"loss": 1.0944,
"step": 140
},
{
"epoch": 0.0177521358038389,
"grad_norm": 3.3125,
"learning_rate": 2.6622296173044925e-06,
"loss": 0.9622,
"step": 160
},
{
"epoch": 0.01997115277931876,
"grad_norm": 5.65625,
"learning_rate": 2.995008319467554e-06,
"loss": 0.7219,
"step": 180
},
{
"epoch": 0.022190169754798626,
"grad_norm": 3.234375,
"learning_rate": 3.3277870216306156e-06,
"loss": 0.6194,
"step": 200
},
{
"epoch": 0.024409186730278486,
"grad_norm": 3.359375,
"learning_rate": 3.6605657237936775e-06,
"loss": 0.5223,
"step": 220
},
{
"epoch": 0.02662820370575835,
"grad_norm": 2.234375,
"learning_rate": 3.993344425956739e-06,
"loss": 0.5673,
"step": 240
},
{
"epoch": 0.02884722068123821,
"grad_norm": 2.109375,
"learning_rate": 4.326123128119801e-06,
"loss": 0.5215,
"step": 260
},
{
"epoch": 0.031066237656718072,
"grad_norm": 2.765625,
"learning_rate": 4.658901830282862e-06,
"loss": 0.5166,
"step": 280
},
{
"epoch": 0.03328525463219793,
"grad_norm": 2.0625,
"learning_rate": 4.991680532445923e-06,
"loss": 0.5023,
"step": 300
},
{
"epoch": 0.0355042716076778,
"grad_norm": 2.0625,
"learning_rate": 5.324459234608985e-06,
"loss": 0.4611,
"step": 320
},
{
"epoch": 0.03772328858315766,
"grad_norm": 2.09375,
"learning_rate": 5.657237936772047e-06,
"loss": 0.5203,
"step": 340
},
{
"epoch": 0.03994230555863752,
"grad_norm": 2.34375,
"learning_rate": 5.990016638935108e-06,
"loss": 0.5498,
"step": 360
},
{
"epoch": 0.04216132253411738,
"grad_norm": 2.03125,
"learning_rate": 6.32279534109817e-06,
"loss": 0.5597,
"step": 380
},
{
"epoch": 0.04438033950959725,
"grad_norm": 2.453125,
"learning_rate": 6.655574043261231e-06,
"loss": 0.5133,
"step": 400
},
{
"epoch": 0.04659935648507711,
"grad_norm": 2.15625,
"learning_rate": 6.988352745424292e-06,
"loss": 0.4843,
"step": 420
},
{
"epoch": 0.04881837346055697,
"grad_norm": 2.265625,
"learning_rate": 7.321131447587355e-06,
"loss": 0.5358,
"step": 440
},
{
"epoch": 0.051037390436036834,
"grad_norm": 2.6875,
"learning_rate": 7.653910149750416e-06,
"loss": 0.3936,
"step": 460
},
{
"epoch": 0.0532564074115167,
"grad_norm": 2.125,
"learning_rate": 7.986688851913478e-06,
"loss": 0.5245,
"step": 480
},
{
"epoch": 0.05547542438699656,
"grad_norm": 2.796875,
"learning_rate": 8.319467554076538e-06,
"loss": 0.5094,
"step": 500
},
{
"epoch": 0.05769444136247642,
"grad_norm": 1.84375,
"learning_rate": 8.652246256239602e-06,
"loss": 0.4476,
"step": 520
},
{
"epoch": 0.059913458337956284,
"grad_norm": 2.09375,
"learning_rate": 8.985024958402662e-06,
"loss": 0.4836,
"step": 540
},
{
"epoch": 0.062132475313436145,
"grad_norm": 1.6328125,
"learning_rate": 9.317803660565724e-06,
"loss": 0.4805,
"step": 560
},
{
"epoch": 0.06435149228891601,
"grad_norm": 1.8828125,
"learning_rate": 9.650582362728786e-06,
"loss": 0.425,
"step": 580
},
{
"epoch": 0.06657050926439587,
"grad_norm": 2.6875,
"learning_rate": 9.983361064891846e-06,
"loss": 0.5223,
"step": 600
},
{
"epoch": 0.06878952623987573,
"grad_norm": 2.140625,
"learning_rate": 1.031613976705491e-05,
"loss": 0.5,
"step": 620
},
{
"epoch": 0.0710085432153556,
"grad_norm": 2.203125,
"learning_rate": 1.064891846921797e-05,
"loss": 0.5271,
"step": 640
},
{
"epoch": 0.07322756019083546,
"grad_norm": 1.9765625,
"learning_rate": 1.0981697171381032e-05,
"loss": 0.4981,
"step": 660
},
{
"epoch": 0.07544657716631532,
"grad_norm": 1.2421875,
"learning_rate": 1.1314475873544094e-05,
"loss": 0.5133,
"step": 680
},
{
"epoch": 0.07766559414179519,
"grad_norm": 2.515625,
"learning_rate": 1.1647254575707154e-05,
"loss": 0.4693,
"step": 700
},
{
"epoch": 0.07988461111727505,
"grad_norm": 2.03125,
"learning_rate": 1.1980033277870216e-05,
"loss": 0.4844,
"step": 720
},
{
"epoch": 0.08210362809275491,
"grad_norm": 1.578125,
"learning_rate": 1.2312811980033278e-05,
"loss": 0.4943,
"step": 740
},
{
"epoch": 0.08432264506823477,
"grad_norm": 2.75,
"learning_rate": 1.264559068219634e-05,
"loss": 0.5,
"step": 760
},
{
"epoch": 0.08654166204371463,
"grad_norm": 1.46875,
"learning_rate": 1.2978369384359402e-05,
"loss": 0.4318,
"step": 780
},
{
"epoch": 0.0887606790191945,
"grad_norm": 2.484375,
"learning_rate": 1.3311148086522462e-05,
"loss": 0.4745,
"step": 800
},
{
"epoch": 0.09097969599467436,
"grad_norm": 2.28125,
"learning_rate": 1.3643926788685524e-05,
"loss": 0.5459,
"step": 820
},
{
"epoch": 0.09319871297015422,
"grad_norm": 2.078125,
"learning_rate": 1.3976705490848584e-05,
"loss": 0.5021,
"step": 840
},
{
"epoch": 0.09541772994563408,
"grad_norm": 0.97265625,
"learning_rate": 1.4309484193011648e-05,
"loss": 0.4271,
"step": 860
},
{
"epoch": 0.09763674692111395,
"grad_norm": 1.828125,
"learning_rate": 1.464226289517471e-05,
"loss": 0.4924,
"step": 880
},
{
"epoch": 0.09985576389659381,
"grad_norm": 2.25,
"learning_rate": 1.497504159733777e-05,
"loss": 0.514,
"step": 900
},
{
"epoch": 0.10207478087207367,
"grad_norm": 1.578125,
"learning_rate": 1.5307820299500832e-05,
"loss": 0.4569,
"step": 920
},
{
"epoch": 0.10429379784755353,
"grad_norm": 1.4140625,
"learning_rate": 1.5640599001663892e-05,
"loss": 0.4751,
"step": 940
},
{
"epoch": 0.1065128148230334,
"grad_norm": 2.1875,
"learning_rate": 1.5973377703826956e-05,
"loss": 0.4719,
"step": 960
},
{
"epoch": 0.10873183179851326,
"grad_norm": 1.921875,
"learning_rate": 1.6306156405990016e-05,
"loss": 0.4455,
"step": 980
},
{
"epoch": 0.11095084877399312,
"grad_norm": 1.7421875,
"learning_rate": 1.6638935108153077e-05,
"loss": 0.5068,
"step": 1000
},
{
"epoch": 0.11316986574947298,
"grad_norm": 1.96875,
"learning_rate": 1.697171381031614e-05,
"loss": 0.5076,
"step": 1020
},
{
"epoch": 0.11538888272495285,
"grad_norm": 2.0625,
"learning_rate": 1.7304492512479204e-05,
"loss": 0.449,
"step": 1040
},
{
"epoch": 0.11760789970043271,
"grad_norm": 2.859375,
"learning_rate": 1.7637271214642264e-05,
"loss": 0.4298,
"step": 1060
},
{
"epoch": 0.11982691667591257,
"grad_norm": 2.0625,
"learning_rate": 1.7970049916805324e-05,
"loss": 0.5298,
"step": 1080
},
{
"epoch": 0.12204593365139244,
"grad_norm": 1.90625,
"learning_rate": 1.8302828618968388e-05,
"loss": 0.5398,
"step": 1100
},
{
"epoch": 0.12426495062687229,
"grad_norm": 1.40625,
"learning_rate": 1.8635607321131448e-05,
"loss": 0.5533,
"step": 1120
},
{
"epoch": 0.12648396760235217,
"grad_norm": 2.21875,
"learning_rate": 1.896838602329451e-05,
"loss": 0.422,
"step": 1140
},
{
"epoch": 0.12870298457783202,
"grad_norm": 1.96875,
"learning_rate": 1.9301164725457572e-05,
"loss": 0.4807,
"step": 1160
},
{
"epoch": 0.13092200155331188,
"grad_norm": 1.484375,
"learning_rate": 1.9633943427620632e-05,
"loss": 0.5209,
"step": 1180
},
{
"epoch": 0.13314101852879173,
"grad_norm": 1.9765625,
"learning_rate": 1.9966722129783693e-05,
"loss": 0.4522,
"step": 1200
},
{
"epoch": 0.13536003550427161,
"grad_norm": 1.3828125,
"learning_rate": 2.0299500831946756e-05,
"loss": 0.4684,
"step": 1220
},
{
"epoch": 0.13757905247975147,
"grad_norm": 1.59375,
"learning_rate": 2.063227953410982e-05,
"loss": 0.3982,
"step": 1240
},
{
"epoch": 0.13979806945523132,
"grad_norm": 1.7734375,
"learning_rate": 2.096505823627288e-05,
"loss": 0.5518,
"step": 1260
},
{
"epoch": 0.1420170864307112,
"grad_norm": 1.6875,
"learning_rate": 2.129783693843594e-05,
"loss": 0.5107,
"step": 1280
},
{
"epoch": 0.14423610340619106,
"grad_norm": 1.453125,
"learning_rate": 2.1630615640599004e-05,
"loss": 0.4422,
"step": 1300
},
{
"epoch": 0.1464551203816709,
"grad_norm": 2.015625,
"learning_rate": 2.1963394342762064e-05,
"loss": 0.5578,
"step": 1320
},
{
"epoch": 0.1486741373571508,
"grad_norm": 1.4921875,
"learning_rate": 2.2296173044925124e-05,
"loss": 0.4516,
"step": 1340
},
{
"epoch": 0.15089315433263065,
"grad_norm": 2.15625,
"learning_rate": 2.2628951747088188e-05,
"loss": 0.3979,
"step": 1360
},
{
"epoch": 0.1531121713081105,
"grad_norm": 1.4140625,
"learning_rate": 2.296173044925125e-05,
"loss": 0.4895,
"step": 1380
},
{
"epoch": 0.15533118828359038,
"grad_norm": 1.453125,
"learning_rate": 2.329450915141431e-05,
"loss": 0.4505,
"step": 1400
},
{
"epoch": 0.15755020525907024,
"grad_norm": 2.265625,
"learning_rate": 2.3627287853577372e-05,
"loss": 0.5008,
"step": 1420
},
{
"epoch": 0.1597692222345501,
"grad_norm": 1.5390625,
"learning_rate": 2.3960066555740432e-05,
"loss": 0.523,
"step": 1440
},
{
"epoch": 0.16198823921002994,
"grad_norm": 1.5546875,
"learning_rate": 2.4292845257903493e-05,
"loss": 0.4813,
"step": 1460
},
{
"epoch": 0.16420725618550983,
"grad_norm": 1.46875,
"learning_rate": 2.4625623960066556e-05,
"loss": 0.4995,
"step": 1480
},
{
"epoch": 0.16642627316098968,
"grad_norm": 1.8359375,
"learning_rate": 2.495840266222962e-05,
"loss": 0.474,
"step": 1500
},
{
"epoch": 0.16864529013646953,
"grad_norm": 1.6328125,
"learning_rate": 2.529118136439268e-05,
"loss": 0.4201,
"step": 1520
},
{
"epoch": 0.17086430711194942,
"grad_norm": 2.15625,
"learning_rate": 2.562396006655574e-05,
"loss": 0.5332,
"step": 1540
},
{
"epoch": 0.17308332408742927,
"grad_norm": 2.53125,
"learning_rate": 2.5956738768718804e-05,
"loss": 0.5105,
"step": 1560
},
{
"epoch": 0.17530234106290912,
"grad_norm": 1.734375,
"learning_rate": 2.6289517470881864e-05,
"loss": 0.4679,
"step": 1580
},
{
"epoch": 0.177521358038389,
"grad_norm": 1.4453125,
"learning_rate": 2.6622296173044925e-05,
"loss": 0.4276,
"step": 1600
},
{
"epoch": 0.17974037501386886,
"grad_norm": 1.9609375,
"learning_rate": 2.6955074875207988e-05,
"loss": 0.4462,
"step": 1620
},
{
"epoch": 0.1819593919893487,
"grad_norm": 1.734375,
"learning_rate": 2.728785357737105e-05,
"loss": 0.4534,
"step": 1640
},
{
"epoch": 0.1841784089648286,
"grad_norm": 1.5703125,
"learning_rate": 2.762063227953411e-05,
"loss": 0.4699,
"step": 1660
},
{
"epoch": 0.18639742594030845,
"grad_norm": 1.828125,
"learning_rate": 2.795341098169717e-05,
"loss": 0.483,
"step": 1680
},
{
"epoch": 0.1886164429157883,
"grad_norm": 2.765625,
"learning_rate": 2.8286189683860236e-05,
"loss": 0.4776,
"step": 1700
},
{
"epoch": 0.19083545989126816,
"grad_norm": 2.0,
"learning_rate": 2.8618968386023296e-05,
"loss": 0.4361,
"step": 1720
},
{
"epoch": 0.19305447686674804,
"grad_norm": 1.640625,
"learning_rate": 2.8951747088186356e-05,
"loss": 0.4566,
"step": 1740
},
{
"epoch": 0.1952734938422279,
"grad_norm": 1.640625,
"learning_rate": 2.928452579034942e-05,
"loss": 0.4609,
"step": 1760
},
{
"epoch": 0.19749251081770774,
"grad_norm": 1.9140625,
"learning_rate": 2.961730449251248e-05,
"loss": 0.5019,
"step": 1780
},
{
"epoch": 0.19971152779318763,
"grad_norm": 2.125,
"learning_rate": 2.995008319467554e-05,
"loss": 0.4844,
"step": 1800
},
{
"epoch": 0.20193054476866748,
"grad_norm": 2.015625,
"learning_rate": 2.999958848436878e-05,
"loss": 0.4974,
"step": 1820
},
{
"epoch": 0.20414956174414733,
"grad_norm": 1.828125,
"learning_rate": 2.9998050673796383e-05,
"loss": 0.4591,
"step": 1840
},
{
"epoch": 0.20636857871962722,
"grad_norm": 1.8671875,
"learning_rate": 2.999537386964595e-05,
"loss": 0.4319,
"step": 1860
},
{
"epoch": 0.20858759569510707,
"grad_norm": 1.953125,
"learning_rate": 2.9991558275201416e-05,
"loss": 0.4425,
"step": 1880
},
{
"epoch": 0.21080661267058692,
"grad_norm": 1.703125,
"learning_rate": 2.998660418022961e-05,
"loss": 0.4729,
"step": 1900
},
{
"epoch": 0.2130256296460668,
"grad_norm": 1.6484375,
"learning_rate": 2.9980511960958247e-05,
"loss": 0.4049,
"step": 1920
},
{
"epoch": 0.21524464662154666,
"grad_norm": 1.671875,
"learning_rate": 2.9973282080047365e-05,
"loss": 0.4646,
"step": 1940
},
{
"epoch": 0.2174636635970265,
"grad_norm": 1.6875,
"learning_rate": 2.996491508655417e-05,
"loss": 0.4725,
"step": 1960
},
{
"epoch": 0.21968268057250637,
"grad_norm": 1.953125,
"learning_rate": 2.995541161589137e-05,
"loss": 0.4448,
"step": 1980
},
{
"epoch": 0.22190169754798625,
"grad_norm": 1.3046875,
"learning_rate": 2.994477238977889e-05,
"loss": 0.4673,
"step": 2000
},
{
"epoch": 0.2241207145234661,
"grad_norm": 1.71875,
"learning_rate": 2.9932998216189077e-05,
"loss": 0.423,
"step": 2020
},
{
"epoch": 0.22633973149894596,
"grad_norm": 1.703125,
"learning_rate": 2.992008998928534e-05,
"loss": 0.4803,
"step": 2040
},
{
"epoch": 0.22855874847442584,
"grad_norm": 1.4296875,
"learning_rate": 2.990604868935424e-05,
"loss": 0.4604,
"step": 2060
},
{
"epoch": 0.2307777654499057,
"grad_norm": 1.890625,
"learning_rate": 2.989087538273105e-05,
"loss": 0.5155,
"step": 2080
},
{
"epoch": 0.23299678242538555,
"grad_norm": 1.7421875,
"learning_rate": 2.9874571221718776e-05,
"loss": 0.5122,
"step": 2100
},
{
"epoch": 0.23521579940086543,
"grad_norm": 1.765625,
"learning_rate": 2.985713744450063e-05,
"loss": 0.4475,
"step": 2120
},
{
"epoch": 0.23743481637634528,
"grad_norm": 2.15625,
"learning_rate": 2.9838575375046034e-05,
"loss": 0.4281,
"step": 2140
},
{
"epoch": 0.23965383335182514,
"grad_norm": 1.9296875,
"learning_rate": 2.9818886423010024e-05,
"loss": 0.4383,
"step": 2160
},
{
"epoch": 0.24187285032730502,
"grad_norm": 1.7421875,
"learning_rate": 2.979807208362625e-05,
"loss": 0.449,
"step": 2180
},
{
"epoch": 0.24409186730278487,
"grad_norm": 1.4453125,
"learning_rate": 2.9776133937593395e-05,
"loss": 0.5002,
"step": 2200
},
{
"epoch": 0.24631088427826472,
"grad_norm": 1.6484375,
"learning_rate": 2.9753073650955128e-05,
"loss": 0.4667,
"step": 2220
},
{
"epoch": 0.24852990125374458,
"grad_norm": 2.203125,
"learning_rate": 2.9728892974973592e-05,
"loss": 0.4431,
"step": 2240
},
{
"epoch": 0.25074891822922446,
"grad_norm": 1.78125,
"learning_rate": 2.970359374599641e-05,
"loss": 0.4684,
"step": 2260
},
{
"epoch": 0.25296793520470434,
"grad_norm": 2.15625,
"learning_rate": 2.967717788531722e-05,
"loss": 0.4599,
"step": 2280
},
{
"epoch": 0.25518695218018417,
"grad_norm": 1.4453125,
"learning_rate": 2.9649647399029764e-05,
"loss": 0.5047,
"step": 2300
},
{
"epoch": 0.25740596915566405,
"grad_norm": 2.265625,
"learning_rate": 2.9621004377875558e-05,
"loss": 0.4824,
"step": 2320
},
{
"epoch": 0.2596249861311439,
"grad_norm": 1.3515625,
"learning_rate": 2.959125099708509e-05,
"loss": 0.4928,
"step": 2340
},
{
"epoch": 0.26184400310662376,
"grad_norm": 1.5703125,
"learning_rate": 2.9560389516212638e-05,
"loss": 0.5386,
"step": 2360
},
{
"epoch": 0.26406302008210364,
"grad_norm": 0.96875,
"learning_rate": 2.9528422278964687e-05,
"loss": 0.4092,
"step": 2380
},
{
"epoch": 0.26628203705758346,
"grad_norm": 1.5703125,
"learning_rate": 2.949535171302192e-05,
"loss": 0.4929,
"step": 2400
},
{
"epoch": 0.26850105403306335,
"grad_norm": 1.5859375,
"learning_rate": 2.9461180329854867e-05,
"loss": 0.4893,
"step": 2420
},
{
"epoch": 0.27072007100854323,
"grad_norm": 1.890625,
"learning_rate": 2.9425910724533165e-05,
"loss": 0.4554,
"step": 2440
},
{
"epoch": 0.27293908798402305,
"grad_norm": 1.65625,
"learning_rate": 2.9389545575528496e-05,
"loss": 0.4702,
"step": 2460
},
{
"epoch": 0.27515810495950294,
"grad_norm": 1.375,
"learning_rate": 2.9352087644511162e-05,
"loss": 0.5056,
"step": 2480
},
{
"epoch": 0.2773771219349828,
"grad_norm": 2.859375,
"learning_rate": 2.9313539776140362e-05,
"loss": 0.428,
"step": 2500
},
{
"epoch": 0.27959613891046264,
"grad_norm": 1.7890625,
"learning_rate": 2.9273904897848174e-05,
"loss": 0.4827,
"step": 2520
},
{
"epoch": 0.2818151558859425,
"grad_norm": 1.359375,
"learning_rate": 2.9233186019617214e-05,
"loss": 0.5082,
"step": 2540
},
{
"epoch": 0.2840341728614224,
"grad_norm": 1.421875,
"learning_rate": 2.9191386233752062e-05,
"loss": 0.4803,
"step": 2560
},
{
"epoch": 0.28625318983690223,
"grad_norm": 1.859375,
"learning_rate": 2.9148508714644424e-05,
"loss": 0.5205,
"step": 2580
},
{
"epoch": 0.2884722068123821,
"grad_norm": 1.546875,
"learning_rate": 2.9104556718532054e-05,
"loss": 0.4965,
"step": 2600
},
{
"epoch": 0.290691223787862,
"grad_norm": 1.5078125,
"learning_rate": 2.9059533583251487e-05,
"loss": 0.4376,
"step": 2620
},
{
"epoch": 0.2929102407633418,
"grad_norm": 1.7578125,
"learning_rate": 2.9013442727984517e-05,
"loss": 0.4618,
"step": 2640
},
{
"epoch": 0.2951292577388217,
"grad_norm": 1.640625,
"learning_rate": 2.8966287652998562e-05,
"loss": 0.4943,
"step": 2660
},
{
"epoch": 0.2973482747143016,
"grad_norm": 1.703125,
"learning_rate": 2.891807193938085e-05,
"loss": 0.4582,
"step": 2680
},
{
"epoch": 0.2995672916897814,
"grad_norm": 1.6875,
"learning_rate": 2.8868799248766436e-05,
"loss": 0.5133,
"step": 2700
},
{
"epoch": 0.3017863086652613,
"grad_norm": 1.71875,
"learning_rate": 2.8818473323060143e-05,
"loss": 0.4628,
"step": 2720
},
{
"epoch": 0.3040053256407412,
"grad_norm": 1.609375,
"learning_rate": 2.87670979841524e-05,
"loss": 0.4408,
"step": 2740
},
{
"epoch": 0.306224342616221,
"grad_norm": 1.734375,
"learning_rate": 2.8714677133628963e-05,
"loss": 0.5004,
"step": 2760
},
{
"epoch": 0.3084433595917009,
"grad_norm": 1.6640625,
"learning_rate": 2.866121475247467e-05,
"loss": 0.4031,
"step": 2780
},
{
"epoch": 0.31066237656718076,
"grad_norm": 1.8046875,
"learning_rate": 2.8606714900771055e-05,
"loss": 0.5192,
"step": 2800
},
{
"epoch": 0.3128813935426606,
"grad_norm": 2.03125,
"learning_rate": 2.8551181717388066e-05,
"loss": 0.5397,
"step": 2820
},
{
"epoch": 0.3151004105181405,
"grad_norm": 1.9609375,
"learning_rate": 2.849461941966972e-05,
"loss": 0.4446,
"step": 2840
},
{
"epoch": 0.3173194274936203,
"grad_norm": 1.9296875,
"learning_rate": 2.8437032303113823e-05,
"loss": 0.4464,
"step": 2860
},
{
"epoch": 0.3195384444691002,
"grad_norm": 1.734375,
"learning_rate": 2.8378424741045773e-05,
"loss": 0.526,
"step": 2880
},
{
"epoch": 0.32175746144458006,
"grad_norm": 1.7265625,
"learning_rate": 2.831880118428644e-05,
"loss": 0.4087,
"step": 2900
},
{
"epoch": 0.3239764784200599,
"grad_norm": 1.125,
"learning_rate": 2.8258166160814135e-05,
"loss": 0.4833,
"step": 2920
},
{
"epoch": 0.32619549539553977,
"grad_norm": 1.34375,
"learning_rate": 2.8196524275420758e-05,
"loss": 0.4403,
"step": 2940
},
{
"epoch": 0.32841451237101965,
"grad_norm": 1.7421875,
"learning_rate": 2.813388020936211e-05,
"loss": 0.4728,
"step": 2960
},
{
"epoch": 0.3306335293464995,
"grad_norm": 1.0390625,
"learning_rate": 2.8070238720002364e-05,
"loss": 0.4389,
"step": 2980
},
{
"epoch": 0.33285254632197936,
"grad_norm": 0.90625,
"learning_rate": 2.800560464045278e-05,
"loss": 0.482,
"step": 3000
},
{
"epoch": 0.33507156329745924,
"grad_norm": 1.6796875,
"learning_rate": 2.7939982879204686e-05,
"loss": 0.4157,
"step": 3020
},
{
"epoch": 0.33729058027293907,
"grad_norm": 1.9140625,
"learning_rate": 2.787337841975668e-05,
"loss": 0.4593,
"step": 3040
},
{
"epoch": 0.33950959724841895,
"grad_norm": 1.2421875,
"learning_rate": 2.78057963202362e-05,
"loss": 0.4071,
"step": 3060
},
{
"epoch": 0.34172861422389883,
"grad_norm": 1.671875,
"learning_rate": 2.773724171301538e-05,
"loss": 0.4307,
"step": 3080
},
{
"epoch": 0.34394763119937866,
"grad_norm": 2.75,
"learning_rate": 2.7667719804321285e-05,
"loss": 0.5196,
"step": 3100
},
{
"epoch": 0.34616664817485854,
"grad_norm": 1.71875,
"learning_rate": 2.7597235873840544e-05,
"loss": 0.4529,
"step": 3120
},
{
"epoch": 0.3483856651503384,
"grad_norm": 1.453125,
"learning_rate": 2.7525795274318386e-05,
"loss": 0.5671,
"step": 3140
},
{
"epoch": 0.35060468212581825,
"grad_norm": 1.6484375,
"learning_rate": 2.745340343115213e-05,
"loss": 0.506,
"step": 3160
},
{
"epoch": 0.3528236991012981,
"grad_norm": 1.828125,
"learning_rate": 2.7380065841979196e-05,
"loss": 0.4834,
"step": 3180
},
{
"epoch": 0.355042716076778,
"grad_norm": 1.9296875,
"learning_rate": 2.7305788076259565e-05,
"loss": 0.4635,
"step": 3200
},
{
"epoch": 0.35726173305225783,
"grad_norm": 1.7109375,
"learning_rate": 2.7230575774852843e-05,
"loss": 0.4873,
"step": 3220
},
{
"epoch": 0.3594807500277377,
"grad_norm": 1.578125,
"learning_rate": 2.715443464958986e-05,
"loss": 0.4133,
"step": 3240
},
{
"epoch": 0.3616997670032176,
"grad_norm": 2.03125,
"learning_rate": 2.707737048283891e-05,
"loss": 0.4954,
"step": 3260
},
{
"epoch": 0.3639187839786974,
"grad_norm": 1.3671875,
"learning_rate": 2.699938912706663e-05,
"loss": 0.4585,
"step": 3280
},
{
"epoch": 0.3661378009541773,
"grad_norm": 1.921875,
"learning_rate": 2.6920496504393507e-05,
"loss": 0.5479,
"step": 3300
},
{
"epoch": 0.3683568179296572,
"grad_norm": 2.515625,
"learning_rate": 2.6840698606144197e-05,
"loss": 0.4712,
"step": 3320
},
{
"epoch": 0.370575834905137,
"grad_norm": 2.28125,
"learning_rate": 2.6760001492392474e-05,
"loss": 0.5046,
"step": 3340
},
{
"epoch": 0.3727948518806169,
"grad_norm": 1.53125,
"learning_rate": 2.6678411291501038e-05,
"loss": 0.522,
"step": 3360
},
{
"epoch": 0.3750138688560967,
"grad_norm": 1.6796875,
"learning_rate": 2.6595934199656108e-05,
"loss": 0.4852,
"step": 3380
},
{
"epoch": 0.3772328858315766,
"grad_norm": 1.9140625,
"learning_rate": 2.6512576480396862e-05,
"loss": 0.4745,
"step": 3400
},
{
"epoch": 0.3794519028070565,
"grad_norm": 1.71875,
"learning_rate": 2.6428344464139756e-05,
"loss": 0.4903,
"step": 3420
},
{
"epoch": 0.3816709197825363,
"grad_norm": 1.625,
"learning_rate": 2.6343244547697798e-05,
"loss": 0.3906,
"step": 3440
},
{
"epoch": 0.3838899367580162,
"grad_norm": 1.859375,
"learning_rate": 2.6257283193794742e-05,
"loss": 0.471,
"step": 3460
},
{
"epoch": 0.3861089537334961,
"grad_norm": 1.90625,
"learning_rate": 2.617046693057429e-05,
"loss": 0.4878,
"step": 3480
},
{
"epoch": 0.3883279707089759,
"grad_norm": 1.765625,
"learning_rate": 2.6082802351104317e-05,
"loss": 0.4718,
"step": 3500
},
{
"epoch": 0.3905469876844558,
"grad_norm": 1.8828125,
"learning_rate": 2.5994296112876222e-05,
"loss": 0.4649,
"step": 3520
},
{
"epoch": 0.39276600465993566,
"grad_norm": 1.2109375,
"learning_rate": 2.5904954937299267e-05,
"loss": 0.3973,
"step": 3540
},
{
"epoch": 0.3949850216354155,
"grad_norm": 1.4609375,
"learning_rate": 2.5814785609190197e-05,
"loss": 0.4785,
"step": 3560
},
{
"epoch": 0.39720403861089537,
"grad_norm": 2.171875,
"learning_rate": 2.5723794976257947e-05,
"loss": 0.4512,
"step": 3580
},
{
"epoch": 0.39942305558637525,
"grad_norm": 1.7265625,
"learning_rate": 2.5631989948583623e-05,
"loss": 0.4931,
"step": 3600
},
{
"epoch": 0.4016420725618551,
"grad_norm": 1.1796875,
"learning_rate": 2.553937749809572e-05,
"loss": 0.4183,
"step": 3620
},
{
"epoch": 0.40386108953733496,
"grad_norm": 1.7890625,
"learning_rate": 2.544596465804068e-05,
"loss": 0.4321,
"step": 3640
},
{
"epoch": 0.40608010651281484,
"grad_norm": 2.109375,
"learning_rate": 2.5351758522448724e-05,
"loss": 0.489,
"step": 3660
},
{
"epoch": 0.40829912348829467,
"grad_norm": 2.03125,
"learning_rate": 2.5256766245595166e-05,
"loss": 0.4398,
"step": 3680
},
{
"epoch": 0.41051814046377455,
"grad_norm": 1.6484375,
"learning_rate": 2.516099504145703e-05,
"loss": 0.4831,
"step": 3700
},
{
"epoch": 0.41273715743925443,
"grad_norm": 1.90625,
"learning_rate": 2.5064452183165283e-05,
"loss": 0.4365,
"step": 3720
},
{
"epoch": 0.41495617441473426,
"grad_norm": 1.359375,
"learning_rate": 2.496714500245241e-05,
"loss": 0.4309,
"step": 3740
},
{
"epoch": 0.41717519139021414,
"grad_norm": 1.765625,
"learning_rate": 2.4869080889095693e-05,
"loss": 0.5378,
"step": 3760
},
{
"epoch": 0.419394208365694,
"grad_norm": 1.75,
"learning_rate": 2.477026729035595e-05,
"loss": 0.4905,
"step": 3780
},
{
"epoch": 0.42161322534117385,
"grad_norm": 1.859375,
"learning_rate": 2.4670711710412026e-05,
"loss": 0.4187,
"step": 3800
},
{
"epoch": 0.42383224231665373,
"grad_norm": 1.9140625,
"learning_rate": 2.457042170979086e-05,
"loss": 0.4817,
"step": 3820
},
{
"epoch": 0.4260512592921336,
"grad_norm": 1.6953125,
"learning_rate": 2.4469404904793338e-05,
"loss": 0.5108,
"step": 3840
},
{
"epoch": 0.42827027626761344,
"grad_norm": 1.46875,
"learning_rate": 2.4367668966915885e-05,
"loss": 0.5112,
"step": 3860
},
{
"epoch": 0.4304892932430933,
"grad_norm": 3.046875,
"learning_rate": 2.4265221622267876e-05,
"loss": 0.5353,
"step": 3880
},
{
"epoch": 0.43270831021857314,
"grad_norm": 1.6640625,
"learning_rate": 2.4162070650984893e-05,
"loss": 0.4684,
"step": 3900
},
{
"epoch": 0.434927327194053,
"grad_norm": 1.5078125,
"learning_rate": 2.4058223886637872e-05,
"loss": 0.4374,
"step": 3920
},
{
"epoch": 0.4371463441695329,
"grad_norm": 1.6953125,
"learning_rate": 2.3953689215638194e-05,
"loss": 0.3753,
"step": 3940
},
{
"epoch": 0.43936536114501273,
"grad_norm": 1.4921875,
"learning_rate": 2.3848474576638807e-05,
"loss": 0.4427,
"step": 3960
},
{
"epoch": 0.4415843781204926,
"grad_norm": 1.25,
"learning_rate": 2.3742587959931285e-05,
"loss": 0.5074,
"step": 3980
},
{
"epoch": 0.4438033950959725,
"grad_norm": 1.7265625,
"learning_rate": 2.3636037406839076e-05,
"loss": 0.4841,
"step": 4000
},
{
"epoch": 0.4460224120714523,
"grad_norm": 1.8203125,
"learning_rate": 2.3528831009106786e-05,
"loss": 0.4643,
"step": 4020
},
{
"epoch": 0.4482414290469322,
"grad_norm": 1.25,
"learning_rate": 2.3420976908285687e-05,
"loss": 0.4976,
"step": 4040
},
{
"epoch": 0.4504604460224121,
"grad_norm": 1.828125,
"learning_rate": 2.3312483295115424e-05,
"loss": 0.5453,
"step": 4060
},
{
"epoch": 0.4526794629978919,
"grad_norm": 2.921875,
"learning_rate": 2.320335840890198e-05,
"loss": 0.46,
"step": 4080
},
{
"epoch": 0.4548984799733718,
"grad_norm": 2.03125,
"learning_rate": 2.3093610536891965e-05,
"loss": 0.5129,
"step": 4100
},
{
"epoch": 0.4571174969488517,
"grad_norm": 1.4375,
"learning_rate": 2.2983248013643253e-05,
"loss": 0.4429,
"step": 4120
},
{
"epoch": 0.4593365139243315,
"grad_norm": 2.03125,
"learning_rate": 2.2872279220392054e-05,
"loss": 0.4817,
"step": 4140
},
{
"epoch": 0.4615555308998114,
"grad_norm": 1.8984375,
"learning_rate": 2.2760712584416386e-05,
"loss": 0.51,
"step": 4160
},
{
"epoch": 0.46377454787529127,
"grad_norm": 1.2734375,
"learning_rate": 2.2648556578396107e-05,
"loss": 0.5001,
"step": 4180
},
{
"epoch": 0.4659935648507711,
"grad_norm": 2.34375,
"learning_rate": 2.2535819719769487e-05,
"loss": 0.4739,
"step": 4200
},
{
"epoch": 0.468212581826251,
"grad_norm": 1.484375,
"learning_rate": 2.242251057008633e-05,
"loss": 0.474,
"step": 4220
},
{
"epoch": 0.47043159880173085,
"grad_norm": 1.7421875,
"learning_rate": 2.2308637734357826e-05,
"loss": 0.4622,
"step": 4240
},
{
"epoch": 0.4726506157772107,
"grad_norm": 1.4296875,
"learning_rate": 2.219420986040305e-05,
"loss": 0.5075,
"step": 4260
},
{
"epoch": 0.47486963275269056,
"grad_norm": 1.7734375,
"learning_rate": 2.2079235638192203e-05,
"loss": 0.4817,
"step": 4280
},
{
"epoch": 0.47708864972817044,
"grad_norm": 1.3671875,
"learning_rate": 2.1963723799186706e-05,
"loss": 0.445,
"step": 4300
},
{
"epoch": 0.47930766670365027,
"grad_norm": 1.6796875,
"learning_rate": 2.184768311567608e-05,
"loss": 0.475,
"step": 4320
},
{
"epoch": 0.48152668367913015,
"grad_norm": 1.640625,
"learning_rate": 2.1731122400111764e-05,
"loss": 0.4418,
"step": 4340
},
{
"epoch": 0.48374570065461003,
"grad_norm": 1.765625,
"learning_rate": 2.161405050443789e-05,
"loss": 0.4722,
"step": 4360
},
{
"epoch": 0.48596471763008986,
"grad_norm": 1.75,
"learning_rate": 2.1496476319419002e-05,
"loss": 0.4814,
"step": 4380
},
{
"epoch": 0.48818373460556974,
"grad_norm": 1.59375,
"learning_rate": 2.137840877396491e-05,
"loss": 0.5435,
"step": 4400
},
{
"epoch": 0.49040275158104957,
"grad_norm": 1.734375,
"learning_rate": 2.125985683445258e-05,
"loss": 0.4521,
"step": 4420
},
{
"epoch": 0.49262176855652945,
"grad_norm": 1.6640625,
"learning_rate": 2.114082950404519e-05,
"loss": 0.452,
"step": 4440
},
{
"epoch": 0.49484078553200933,
"grad_norm": 1.2734375,
"learning_rate": 2.1021335822008447e-05,
"loss": 0.5176,
"step": 4460
},
{
"epoch": 0.49705980250748916,
"grad_norm": 1.6328125,
"learning_rate": 2.0901384863024078e-05,
"loss": 0.4307,
"step": 4480
},
{
"epoch": 0.49927881948296904,
"grad_norm": 1.6640625,
"learning_rate": 2.0780985736500696e-05,
"loss": 0.4856,
"step": 4500
},
{
"epoch": 0.5014978364584489,
"grad_norm": 1.9609375,
"learning_rate": 2.0660147585881994e-05,
"loss": 0.4339,
"step": 4520
},
{
"epoch": 0.5037168534339288,
"grad_norm": 2.046875,
"learning_rate": 2.0538879587952382e-05,
"loss": 0.4902,
"step": 4540
},
{
"epoch": 0.5059358704094087,
"grad_norm": 1.8671875,
"learning_rate": 2.0417190952140064e-05,
"loss": 0.4343,
"step": 4560
},
{
"epoch": 0.5081548873848885,
"grad_norm": 1.7265625,
"learning_rate": 2.029509091981765e-05,
"loss": 0.4416,
"step": 4580
},
{
"epoch": 0.5103739043603683,
"grad_norm": 1.9375,
"learning_rate": 2.0172588763600335e-05,
"loss": 0.4802,
"step": 4600
},
{
"epoch": 0.5125929213358482,
"grad_norm": 1.703125,
"learning_rate": 2.0049693786641734e-05,
"loss": 0.4794,
"step": 4620
},
{
"epoch": 0.5148119383113281,
"grad_norm": 2.8125,
"learning_rate": 1.9926415321927347e-05,
"loss": 0.5092,
"step": 4640
},
{
"epoch": 0.517030955286808,
"grad_norm": 1.3984375,
"learning_rate": 1.980276273156581e-05,
"loss": 0.4595,
"step": 4660
},
{
"epoch": 0.5192499722622878,
"grad_norm": 1.484375,
"learning_rate": 1.9678745406077886e-05,
"loss": 0.452,
"step": 4680
},
{
"epoch": 0.5214689892377676,
"grad_norm": 1.703125,
"learning_rate": 1.9554372763683337e-05,
"loss": 0.4528,
"step": 4700
},
{
"epoch": 0.5236880062132475,
"grad_norm": 1.5,
"learning_rate": 1.9429654249585684e-05,
"loss": 0.4743,
"step": 4720
},
{
"epoch": 0.5259070231887274,
"grad_norm": 1.5703125,
"learning_rate": 1.9304599335254894e-05,
"loss": 0.4476,
"step": 4740
},
{
"epoch": 0.5281260401642073,
"grad_norm": 1.546875,
"learning_rate": 1.9179217517708117e-05,
"loss": 0.447,
"step": 4760
},
{
"epoch": 0.5303450571396872,
"grad_norm": 1.625,
"learning_rate": 1.9053518318788428e-05,
"loss": 0.4606,
"step": 4780
},
{
"epoch": 0.5325640741151669,
"grad_norm": 1.6875,
"learning_rate": 1.8927511284441722e-05,
"loss": 0.4561,
"step": 4800
},
{
"epoch": 0.5347830910906468,
"grad_norm": 1.8359375,
"learning_rate": 1.880120598399178e-05,
"loss": 0.4522,
"step": 4820
},
{
"epoch": 0.5370021080661267,
"grad_norm": 1.9921875,
"learning_rate": 1.8674612009413536e-05,
"loss": 0.4943,
"step": 4840
},
{
"epoch": 0.5392211250416066,
"grad_norm": 1.703125,
"learning_rate": 1.8547738974604623e-05,
"loss": 0.5822,
"step": 4860
},
{
"epoch": 0.5414401420170865,
"grad_norm": 1.7734375,
"learning_rate": 1.842059651465531e-05,
"loss": 0.4273,
"step": 4880
},
{
"epoch": 0.5436591589925663,
"grad_norm": 1.6953125,
"learning_rate": 1.829319428511673e-05,
"loss": 0.4704,
"step": 4900
},
{
"epoch": 0.5458781759680461,
"grad_norm": 1.671875,
"learning_rate": 1.816554196126767e-05,
"loss": 0.5452,
"step": 4920
},
{
"epoch": 0.548097192943526,
"grad_norm": 1.8203125,
"learning_rate": 1.803764923737974e-05,
"loss": 0.436,
"step": 4940
},
{
"epoch": 0.5503162099190059,
"grad_norm": 1.7578125,
"learning_rate": 1.7909525825981214e-05,
"loss": 0.4518,
"step": 4960
},
{
"epoch": 0.5525352268944858,
"grad_norm": 2.03125,
"learning_rate": 1.778118145711942e-05,
"loss": 0.4138,
"step": 4980
},
{
"epoch": 0.5547542438699656,
"grad_norm": 2.46875,
"learning_rate": 1.7652625877621793e-05,
"loss": 0.4332,
"step": 5000
},
{
"epoch": 0.5569732608454455,
"grad_norm": 2.09375,
"learning_rate": 1.7523868850355704e-05,
"loss": 0.4842,
"step": 5020
},
{
"epoch": 0.5591922778209253,
"grad_norm": 1.6328125,
"learning_rate": 1.7394920153487022e-05,
"loss": 0.4935,
"step": 5040
},
{
"epoch": 0.5614112947964052,
"grad_norm": 1.6953125,
"learning_rate": 1.7265789579737528e-05,
"loss": 0.5129,
"step": 5060
},
{
"epoch": 0.563630311771885,
"grad_norm": 1.65625,
"learning_rate": 1.7136486935641256e-05,
"loss": 0.4281,
"step": 5080
},
{
"epoch": 0.5658493287473649,
"grad_norm": 2.015625,
"learning_rate": 1.7007022040799726e-05,
"loss": 0.4634,
"step": 5100
},
{
"epoch": 0.5680683457228448,
"grad_norm": 2.15625,
"learning_rate": 1.687740472713623e-05,
"loss": 0.5225,
"step": 5120
},
{
"epoch": 0.5702873626983247,
"grad_norm": 1.9921875,
"learning_rate": 1.674764483814918e-05,
"loss": 0.4838,
"step": 5140
},
{
"epoch": 0.5725063796738045,
"grad_norm": 1.7421875,
"learning_rate": 1.661775222816453e-05,
"loss": 0.4291,
"step": 5160
},
{
"epoch": 0.5747253966492843,
"grad_norm": 1.671875,
"learning_rate": 1.648773676158747e-05,
"loss": 0.3925,
"step": 5180
},
{
"epoch": 0.5769444136247642,
"grad_norm": 1.734375,
"learning_rate": 1.6357608312153223e-05,
"loss": 0.4385,
"step": 5200
},
{
"epoch": 0.5791634306002441,
"grad_norm": 1.71875,
"learning_rate": 1.6227376762177272e-05,
"loss": 0.4144,
"step": 5220
},
{
"epoch": 0.581382447575724,
"grad_norm": 2.71875,
"learning_rate": 1.6097052001804825e-05,
"loss": 0.5116,
"step": 5240
},
{
"epoch": 0.5836014645512038,
"grad_norm": 1.953125,
"learning_rate": 1.5966643928259753e-05,
"loss": 0.4869,
"step": 5260
},
{
"epoch": 0.5858204815266836,
"grad_norm": 1.9140625,
"learning_rate": 1.5836162445092963e-05,
"loss": 0.4644,
"step": 5280
},
{
"epoch": 0.5880394985021635,
"grad_norm": 1.8203125,
"learning_rate": 1.5705617461430282e-05,
"loss": 0.3844,
"step": 5300
},
{
"epoch": 0.5902585154776434,
"grad_norm": 1.75,
"learning_rate": 1.5575018891219944e-05,
"loss": 0.4233,
"step": 5320
},
{
"epoch": 0.5924775324531233,
"grad_norm": 1.140625,
"learning_rate": 1.5444376652479706e-05,
"loss": 0.5273,
"step": 5340
},
{
"epoch": 0.5946965494286032,
"grad_norm": 1.921875,
"learning_rate": 1.531370066654362e-05,
"loss": 0.4783,
"step": 5360
},
{
"epoch": 0.5969155664040829,
"grad_norm": 1.796875,
"learning_rate": 1.5183000857308604e-05,
"loss": 0.4747,
"step": 5380
},
{
"epoch": 0.5991345833795628,
"grad_norm": 1.6484375,
"learning_rate": 1.5052287150480774e-05,
"loss": 0.4335,
"step": 5400
},
{
"epoch": 0.6013536003550427,
"grad_norm": 1.75,
"learning_rate": 1.4921569472821673e-05,
"loss": 0.4489,
"step": 5420
},
{
"epoch": 0.6035726173305226,
"grad_norm": 2.21875,
"learning_rate": 1.4790857751394398e-05,
"loss": 0.4495,
"step": 5440
},
{
"epoch": 0.6057916343060025,
"grad_norm": 1.5703125,
"learning_rate": 1.4660161912809718e-05,
"loss": 0.4516,
"step": 5460
},
{
"epoch": 0.6080106512814824,
"grad_norm": 1.4765625,
"learning_rate": 1.4529491882472209e-05,
"loss": 0.4418,
"step": 5480
},
{
"epoch": 0.6102296682569621,
"grad_norm": 1.6015625,
"learning_rate": 1.4398857583826501e-05,
"loss": 0.4701,
"step": 5500
},
{
"epoch": 0.612448685232442,
"grad_norm": 2.390625,
"learning_rate": 1.4268268937603659e-05,
"loss": 0.4957,
"step": 5520
},
{
"epoch": 0.6146677022079219,
"grad_norm": 1.7421875,
"learning_rate": 1.413773586106777e-05,
"loss": 0.5176,
"step": 5540
},
{
"epoch": 0.6168867191834018,
"grad_norm": 1.8125,
"learning_rate": 1.400726826726282e-05,
"loss": 0.4252,
"step": 5560
},
{
"epoch": 0.6191057361588816,
"grad_norm": 1.4453125,
"learning_rate": 1.3876876064259836e-05,
"loss": 0.473,
"step": 5580
},
{
"epoch": 0.6213247531343615,
"grad_norm": 1.6796875,
"learning_rate": 1.3746569154404477e-05,
"loss": 0.4589,
"step": 5600
},
{
"epoch": 0.6235437701098413,
"grad_norm": 2.734375,
"learning_rate": 1.3616357433564993e-05,
"loss": 0.4083,
"step": 5620
},
{
"epoch": 0.6257627870853212,
"grad_norm": 1.75,
"learning_rate": 1.348625079038071e-05,
"loss": 0.4638,
"step": 5640
},
{
"epoch": 0.6279818040608011,
"grad_norm": 2.0625,
"learning_rate": 1.335625910551108e-05,
"loss": 0.4703,
"step": 5660
},
{
"epoch": 0.630200821036281,
"grad_norm": 1.7109375,
"learning_rate": 1.3226392250885288e-05,
"loss": 0.4683,
"step": 5680
},
{
"epoch": 0.6324198380117608,
"grad_norm": 1.2265625,
"learning_rate": 1.3096660088952581e-05,
"loss": 0.4641,
"step": 5700
},
{
"epoch": 0.6346388549872406,
"grad_norm": 1.6953125,
"learning_rate": 1.2967072471933255e-05,
"loss": 0.4324,
"step": 5720
},
{
"epoch": 0.6368578719627205,
"grad_norm": 1.6953125,
"learning_rate": 1.283763924107046e-05,
"loss": 0.5183,
"step": 5740
},
{
"epoch": 0.6390768889382004,
"grad_norm": 1.4921875,
"learning_rate": 1.2708370225882848e-05,
"loss": 0.4178,
"step": 5760
},
{
"epoch": 0.6412959059136802,
"grad_norm": 2.046875,
"learning_rate": 1.2579275243418074e-05,
"loss": 0.4503,
"step": 5780
},
{
"epoch": 0.6435149228891601,
"grad_norm": 1.21875,
"learning_rate": 1.245036409750725e-05,
"loss": 0.4796,
"step": 5800
},
{
"epoch": 0.64573393986464,
"grad_norm": 1.6015625,
"learning_rate": 1.2321646578020452e-05,
"loss": 0.4437,
"step": 5820
},
{
"epoch": 0.6479529568401198,
"grad_norm": 2.09375,
"learning_rate": 1.219313246012321e-05,
"loss": 0.4504,
"step": 5840
},
{
"epoch": 0.6501719738155997,
"grad_norm": 1.2265625,
"learning_rate": 1.2064831503534185e-05,
"loss": 0.5122,
"step": 5860
},
{
"epoch": 0.6523909907910795,
"grad_norm": 1.7890625,
"learning_rate": 1.1936753451783973e-05,
"loss": 0.4294,
"step": 5880
},
{
"epoch": 0.6546100077665594,
"grad_norm": 2.296875,
"learning_rate": 1.1808908031475151e-05,
"loss": 0.4895,
"step": 5900
},
{
"epoch": 0.6568290247420393,
"grad_norm": 1.140625,
"learning_rate": 1.1681304951543635e-05,
"loss": 0.4824,
"step": 5920
},
{
"epoch": 0.6590480417175192,
"grad_norm": 1.875,
"learning_rate": 1.1553953902521321e-05,
"loss": 0.525,
"step": 5940
},
{
"epoch": 0.661267058692999,
"grad_norm": 1.8359375,
"learning_rate": 1.1426864555800195e-05,
"loss": 0.4289,
"step": 5960
},
{
"epoch": 0.6634860756684788,
"grad_norm": 1.9375,
"learning_rate": 1.1300046562897837e-05,
"loss": 0.4695,
"step": 5980
},
{
"epoch": 0.6657050926439587,
"grad_norm": 1.7109375,
"learning_rate": 1.1173509554724461e-05,
"loss": 0.465,
"step": 6000
},
{
"epoch": 0.6679241096194386,
"grad_norm": 1.5625,
"learning_rate": 1.104726314085153e-05,
"loss": 0.4641,
"step": 6020
},
{
"epoch": 0.6701431265949185,
"grad_norm": 1.84375,
"learning_rate": 1.0921316908781965e-05,
"loss": 0.4559,
"step": 6040
},
{
"epoch": 0.6723621435703984,
"grad_norm": 1.4375,
"learning_rate": 1.079568042322205e-05,
"loss": 0.4938,
"step": 6060
},
{
"epoch": 0.6745811605458781,
"grad_norm": 1.2734375,
"learning_rate": 1.0670363225355054e-05,
"loss": 0.4247,
"step": 6080
},
{
"epoch": 0.676800177521358,
"grad_norm": 1.5390625,
"learning_rate": 1.0545374832116658e-05,
"loss": 0.4739,
"step": 6100
},
{
"epoch": 0.6790191944968379,
"grad_norm": 1.640625,
"learning_rate": 1.042072473547221e-05,
"loss": 0.4923,
"step": 6120
},
{
"epoch": 0.6812382114723178,
"grad_norm": 1.5,
"learning_rate": 1.0296422401695867e-05,
"loss": 0.5248,
"step": 6140
},
{
"epoch": 0.6834572284477977,
"grad_norm": 1.5390625,
"learning_rate": 1.017247727065172e-05,
"loss": 0.4336,
"step": 6160
},
{
"epoch": 0.6856762454232775,
"grad_norm": 1.7265625,
"learning_rate": 1.0048898755076885e-05,
"loss": 0.4915,
"step": 6180
},
{
"epoch": 0.6878952623987573,
"grad_norm": 1.515625,
"learning_rate": 9.925696239866679e-06,
"loss": 0.4908,
"step": 6200
},
{
"epoch": 0.6901142793742372,
"grad_norm": 1.4375,
"learning_rate": 9.802879081361927e-06,
"loss": 0.4512,
"step": 6220
},
{
"epoch": 0.6923332963497171,
"grad_norm": 2.21875,
"learning_rate": 9.680456606638376e-06,
"loss": 0.4356,
"step": 6240
},
{
"epoch": 0.694552313325197,
"grad_norm": 1.796875,
"learning_rate": 9.558438112798397e-06,
"loss": 0.4321,
"step": 6260
},
{
"epoch": 0.6967713303006768,
"grad_norm": 1.421875,
"learning_rate": 9.436832866264942e-06,
"loss": 0.4288,
"step": 6280
},
{
"epoch": 0.6989903472761566,
"grad_norm": 2.140625,
"learning_rate": 9.3156501020778e-06,
"loss": 0.4119,
"step": 6300
},
{
"epoch": 0.7012093642516365,
"grad_norm": 1.7421875,
"learning_rate": 9.194899023192295e-06,
"loss": 0.4729,
"step": 6320
},
{
"epoch": 0.7034283812271164,
"grad_norm": 2.03125,
"learning_rate": 9.074588799780359e-06,
"loss": 0.4438,
"step": 6340
},
{
"epoch": 0.7056473982025963,
"grad_norm": 1.703125,
"learning_rate": 8.95472856853414e-06,
"loss": 0.4509,
"step": 6360
},
{
"epoch": 0.7078664151780761,
"grad_norm": 1.453125,
"learning_rate": 8.835327431972136e-06,
"loss": 0.4812,
"step": 6380
},
{
"epoch": 0.710085432153556,
"grad_norm": 1.4375,
"learning_rate": 8.716394457747915e-06,
"loss": 0.4796,
"step": 6400
},
{
"epoch": 0.7123044491290358,
"grad_norm": 0.9375,
"learning_rate": 8.597938677961505e-06,
"loss": 0.4138,
"step": 6420
},
{
"epoch": 0.7145234661045157,
"grad_norm": 1.71875,
"learning_rate": 8.479969088473462e-06,
"loss": 0.4161,
"step": 6440
},
{
"epoch": 0.7167424830799956,
"grad_norm": 2.203125,
"learning_rate": 8.362494648221697e-06,
"loss": 0.4685,
"step": 6460
},
{
"epoch": 0.7189615000554754,
"grad_norm": 1.6875,
"learning_rate": 8.245524278541116e-06,
"loss": 0.4476,
"step": 6480
},
{
"epoch": 0.7211805170309553,
"grad_norm": 1.671875,
"learning_rate": 8.129066862486115e-06,
"loss": 0.5104,
"step": 6500
},
{
"epoch": 0.7233995340064352,
"grad_norm": 1.5625,
"learning_rate": 8.013131244155964e-06,
"loss": 0.4467,
"step": 6520
},
{
"epoch": 0.725618550981915,
"grad_norm": 2.234375,
"learning_rate": 7.89772622802316e-06,
"loss": 0.4317,
"step": 6540
},
{
"epoch": 0.7278375679573948,
"grad_norm": 2.109375,
"learning_rate": 7.782860578264806e-06,
"loss": 0.4398,
"step": 6560
},
{
"epoch": 0.7300565849328747,
"grad_norm": 1.6875,
"learning_rate": 7.668543018097014e-06,
"loss": 0.5054,
"step": 6580
},
{
"epoch": 0.7322756019083546,
"grad_norm": 1.4453125,
"learning_rate": 7.5547822291124715e-06,
"loss": 0.4968,
"step": 6600
},
{
"epoch": 0.7344946188838345,
"grad_norm": 1.5,
"learning_rate": 7.441586850621102e-06,
"loss": 0.4202,
"step": 6620
},
{
"epoch": 0.7367136358593144,
"grad_norm": 1.765625,
"learning_rate": 7.328965478993994e-06,
"loss": 0.4447,
"step": 6640
},
{
"epoch": 0.7389326528347941,
"grad_norm": 1.90625,
"learning_rate": 7.2169266670105555e-06,
"loss": 0.4869,
"step": 6660
},
{
"epoch": 0.741151669810274,
"grad_norm": 1.3203125,
"learning_rate": 7.105478923209001e-06,
"loss": 0.4744,
"step": 6680
},
{
"epoch": 0.7433706867857539,
"grad_norm": 0.76171875,
"learning_rate": 6.994630711240201e-06,
"loss": 0.4054,
"step": 6700
},
{
"epoch": 0.7455897037612338,
"grad_norm": 1.9765625,
"learning_rate": 6.884390449224898e-06,
"loss": 0.4307,
"step": 6720
},
{
"epoch": 0.7478087207367137,
"grad_norm": 1.921875,
"learning_rate": 6.774766509114435e-06,
"loss": 0.4728,
"step": 6740
},
{
"epoch": 0.7500277377121934,
"grad_norm": 1.53125,
"learning_rate": 6.66576721605496e-06,
"loss": 0.4254,
"step": 6760
},
{
"epoch": 0.7522467546876733,
"grad_norm": 1.78125,
"learning_rate": 6.557400847755183e-06,
"loss": 0.4508,
"step": 6780
},
{
"epoch": 0.7544657716631532,
"grad_norm": 1.828125,
"learning_rate": 6.449675633857772e-06,
"loss": 0.4814,
"step": 6800
},
{
"epoch": 0.7566847886386331,
"grad_norm": 1.578125,
"learning_rate": 6.3425997553143315e-06,
"loss": 0.459,
"step": 6820
},
{
"epoch": 0.758903805614113,
"grad_norm": 1.7421875,
"learning_rate": 6.236181343764144e-06,
"loss": 0.464,
"step": 6840
},
{
"epoch": 0.7611228225895929,
"grad_norm": 2.21875,
"learning_rate": 6.130428480916626e-06,
"loss": 0.4897,
"step": 6860
},
{
"epoch": 0.7633418395650726,
"grad_norm": 1.5625,
"learning_rate": 6.025349197937577e-06,
"loss": 0.4614,
"step": 6880
},
{
"epoch": 0.7655608565405525,
"grad_norm": 1.71875,
"learning_rate": 5.920951474839266e-06,
"loss": 0.444,
"step": 6900
},
{
"epoch": 0.7677798735160324,
"grad_norm": 1.1640625,
"learning_rate": 5.817243239874434e-06,
"loss": 0.4649,
"step": 6920
},
{
"epoch": 0.7699988904915123,
"grad_norm": 1.8046875,
"learning_rate": 5.714232368934163e-06,
"loss": 0.4758,
"step": 6940
},
{
"epoch": 0.7722179074669921,
"grad_norm": 2.15625,
"learning_rate": 5.611926684949779e-06,
"loss": 0.4519,
"step": 6960
},
{
"epoch": 0.774436924442472,
"grad_norm": 1.5234375,
"learning_rate": 5.510333957298756e-06,
"loss": 0.4729,
"step": 6980
},
{
"epoch": 0.7766559414179518,
"grad_norm": 1.28125,
"learning_rate": 5.409461901214679e-06,
"loss": 0.4557,
"step": 7000
},
{
"epoch": 0.7788749583934317,
"grad_norm": 1.2265625,
"learning_rate": 5.3093181772013545e-06,
"loss": 0.4433,
"step": 7020
},
{
"epoch": 0.7810939753689116,
"grad_norm": 1.4375,
"learning_rate": 5.209910390451007e-06,
"loss": 0.4767,
"step": 7040
},
{
"epoch": 0.7833129923443914,
"grad_norm": 2.140625,
"learning_rate": 5.111246090266763e-06,
"loss": 0.4422,
"step": 7060
},
{
"epoch": 0.7855320093198713,
"grad_norm": 1.7265625,
"learning_rate": 5.0133327694893035e-06,
"loss": 0.4276,
"step": 7080
},
{
"epoch": 0.7877510262953512,
"grad_norm": 1.5859375,
"learning_rate": 4.916177863927856e-06,
"loss": 0.482,
"step": 7100
},
{
"epoch": 0.789970043270831,
"grad_norm": 1.71875,
"learning_rate": 4.819788751795485e-06,
"loss": 0.4817,
"step": 7120
},
{
"epoch": 0.7921890602463109,
"grad_norm": 1.6484375,
"learning_rate": 4.7241727531487925e-06,
"loss": 0.4702,
"step": 7140
},
{
"epoch": 0.7944080772217907,
"grad_norm": 1.9296875,
"learning_rate": 4.629337129331983e-06,
"loss": 0.4494,
"step": 7160
},
{
"epoch": 0.7966270941972706,
"grad_norm": 1.9453125,
"learning_rate": 4.535289082425438e-06,
"loss": 0.4411,
"step": 7180
},
{
"epoch": 0.7988461111727505,
"grad_norm": 2.015625,
"learning_rate": 4.442035754698759e-06,
"loss": 0.4671,
"step": 7200
},
{
"epoch": 0.8010651281482304,
"grad_norm": 2.359375,
"learning_rate": 4.349584228068369e-06,
"loss": 0.4077,
"step": 7220
},
{
"epoch": 0.8032841451237102,
"grad_norm": 1.3984375,
"learning_rate": 4.257941523559703e-06,
"loss": 0.5378,
"step": 7240
},
{
"epoch": 0.80550316209919,
"grad_norm": 1.84375,
"learning_rate": 4.167114600773983e-06,
"loss": 0.4719,
"step": 7260
},
{
"epoch": 0.8077221790746699,
"grad_norm": 1.140625,
"learning_rate": 4.0771103573597125e-06,
"loss": 0.4596,
"step": 7280
},
{
"epoch": 0.8099411960501498,
"grad_norm": 1.5703125,
"learning_rate": 3.987935628488841e-06,
"loss": 0.4812,
"step": 7300
},
{
"epoch": 0.8121602130256297,
"grad_norm": 1.65625,
"learning_rate": 3.899597186337676e-06,
"loss": 0.4636,
"step": 7320
},
{
"epoch": 0.8143792300011095,
"grad_norm": 1.609375,
"learning_rate": 3.812101739572605e-06,
"loss": 0.4605,
"step": 7340
},
{
"epoch": 0.8165982469765893,
"grad_norm": 2.09375,
"learning_rate": 3.725455932840593e-06,
"loss": 0.4368,
"step": 7360
},
{
"epoch": 0.8188172639520692,
"grad_norm": 1.8125,
"learning_rate": 3.6396663462645917e-06,
"loss": 0.4686,
"step": 7380
},
{
"epoch": 0.8210362809275491,
"grad_norm": 1.875,
"learning_rate": 3.554739494943813e-06,
"loss": 0.4701,
"step": 7400
},
{
"epoch": 0.823255297903029,
"grad_norm": 1.8984375,
"learning_rate": 3.470681828458962e-06,
"loss": 0.4595,
"step": 7420
},
{
"epoch": 0.8254743148785089,
"grad_norm": 1.328125,
"learning_rate": 3.3874997303824416e-06,
"loss": 0.4265,
"step": 7440
},
{
"epoch": 0.8276933318539886,
"grad_norm": 1.5,
"learning_rate": 3.305199517793557e-06,
"loss": 0.4927,
"step": 7460
},
{
"epoch": 0.8299123488294685,
"grad_norm": 1.7109375,
"learning_rate": 3.2237874407987776e-06,
"loss": 0.4211,
"step": 7480
},
{
"epoch": 0.8321313658049484,
"grad_norm": 2.28125,
"learning_rate": 3.1432696820570993e-06,
"loss": 0.4771,
"step": 7500
},
{
"epoch": 0.8343503827804283,
"grad_norm": 1.7109375,
"learning_rate": 3.0636523563104985e-06,
"loss": 0.4934,
"step": 7520
},
{
"epoch": 0.8365693997559082,
"grad_norm": 1.6328125,
"learning_rate": 2.9849415099195886e-06,
"loss": 0.5184,
"step": 7540
},
{
"epoch": 0.838788416731388,
"grad_norm": 1.234375,
"learning_rate": 2.9071431204044123e-06,
"loss": 0.4622,
"step": 7560
},
{
"epoch": 0.8410074337068678,
"grad_norm": 1.6171875,
"learning_rate": 2.8302630959905084e-06,
"loss": 0.4491,
"step": 7580
},
{
"epoch": 0.8432264506823477,
"grad_norm": 2.46875,
"learning_rate": 2.7543072751602246e-06,
"loss": 0.4093,
"step": 7600
},
{
"epoch": 0.8454454676578276,
"grad_norm": 1.671875,
"learning_rate": 2.6792814262093214e-06,
"loss": 0.4312,
"step": 7620
},
{
"epoch": 0.8476644846333075,
"grad_norm": 1.71875,
"learning_rate": 2.605191246808912e-06,
"loss": 0.455,
"step": 7640
},
{
"epoch": 0.8498835016087873,
"grad_norm": 2.109375,
"learning_rate": 2.5320423635727824e-06,
"loss": 0.4839,
"step": 7660
},
{
"epoch": 0.8521025185842672,
"grad_norm": 2.15625,
"learning_rate": 2.45984033163006e-06,
"loss": 0.5647,
"step": 7680
},
{
"epoch": 0.854321535559747,
"grad_norm": 1.6484375,
"learning_rate": 2.388590634203366e-06,
"loss": 0.4108,
"step": 7700
},
{
"epoch": 0.8565405525352269,
"grad_norm": 1.90625,
"learning_rate": 2.3182986821923934e-06,
"loss": 0.3933,
"step": 7720
},
{
"epoch": 0.8587595695107068,
"grad_norm": 2.015625,
"learning_rate": 2.2489698137629904e-06,
"loss": 0.413,
"step": 7740
},
{
"epoch": 0.8609785864861866,
"grad_norm": 2.140625,
"learning_rate": 2.1806092939417732e-06,
"loss": 0.516,
"step": 7760
},
{
"epoch": 0.8631976034616665,
"grad_norm": 1.40625,
"learning_rate": 2.1132223142162714e-06,
"loss": 0.4403,
"step": 7780
},
{
"epoch": 0.8654166204371463,
"grad_norm": 1.859375,
"learning_rate": 2.046813992140679e-06,
"loss": 0.4355,
"step": 7800
},
{
"epoch": 0.8676356374126262,
"grad_norm": 1.4375,
"learning_rate": 1.981389370947218e-06,
"loss": 0.4744,
"step": 7820
},
{
"epoch": 0.869854654388106,
"grad_norm": 2.359375,
"learning_rate": 1.9169534191631243e-06,
"loss": 0.4106,
"step": 7840
},
{
"epoch": 0.8720736713635859,
"grad_norm": 2.046875,
"learning_rate": 1.853511030233354e-06,
"loss": 0.4519,
"step": 7860
},
{
"epoch": 0.8742926883390658,
"grad_norm": 1.7421875,
"learning_rate": 1.79106702214893e-06,
"loss": 0.4568,
"step": 7880
},
{
"epoch": 0.8765117053145457,
"grad_norm": 1.5234375,
"learning_rate": 1.7296261370810695e-06,
"loss": 0.4734,
"step": 7900
},
{
"epoch": 0.8787307222900255,
"grad_norm": 1.828125,
"learning_rate": 1.669193041021041e-06,
"loss": 0.547,
"step": 7920
},
{
"epoch": 0.8809497392655053,
"grad_norm": 1.4453125,
"learning_rate": 1.6097723234258188e-06,
"loss": 0.5001,
"step": 7940
},
{
"epoch": 0.8831687562409852,
"grad_norm": 1.6875,
"learning_rate": 1.5513684968695574e-06,
"loss": 0.447,
"step": 7960
},
{
"epoch": 0.8853877732164651,
"grad_norm": 1.578125,
"learning_rate": 1.4939859967008768e-06,
"loss": 0.4161,
"step": 7980
},
{
"epoch": 0.887606790191945,
"grad_norm": 1.78125,
"learning_rate": 1.437629180706037e-06,
"loss": 0.4606,
"step": 8000
},
{
"epoch": 0.8898258071674249,
"grad_norm": 1.6015625,
"learning_rate": 1.382302328778e-06,
"loss": 0.4126,
"step": 8020
},
{
"epoch": 0.8920448241429046,
"grad_norm": 1.6015625,
"learning_rate": 1.328009642591394e-06,
"loss": 0.4313,
"step": 8040
},
{
"epoch": 0.8942638411183845,
"grad_norm": 1.4296875,
"learning_rate": 1.2747552452834388e-06,
"loss": 0.4436,
"step": 8060
},
{
"epoch": 0.8964828580938644,
"grad_norm": 1.6015625,
"learning_rate": 1.2225431811408133e-06,
"loss": 0.4591,
"step": 8080
},
{
"epoch": 0.8987018750693443,
"grad_norm": 1.484375,
"learning_rate": 1.1713774152925195e-06,
"loss": 0.4103,
"step": 8100
},
{
"epoch": 0.9009208920448242,
"grad_norm": 1.890625,
"learning_rate": 1.1212618334087693e-06,
"loss": 0.4196,
"step": 8120
},
{
"epoch": 0.903139909020304,
"grad_norm": 2.125,
"learning_rate": 1.0722002414058868e-06,
"loss": 0.4701,
"step": 8140
},
{
"epoch": 0.9053589259957838,
"grad_norm": 1.8671875,
"learning_rate": 1.0241963651572867e-06,
"loss": 0.4452,
"step": 8160
},
{
"epoch": 0.9075779429712637,
"grad_norm": 2.21875,
"learning_rate": 9.772538502105093e-07,
"loss": 0.4122,
"step": 8180
},
{
"epoch": 0.9097969599467436,
"grad_norm": 1.625,
"learning_rate": 9.313762615103761e-07,
"loss": 0.4265,
"step": 8200
},
{
"epoch": 0.9120159769222235,
"grad_norm": 1.9296875,
"learning_rate": 8.865670831282513e-07,
"loss": 0.4678,
"step": 8220
},
{
"epoch": 0.9142349938977034,
"grad_norm": 1.671875,
"learning_rate": 8.42829717997457e-07,
"loss": 0.4137,
"step": 8240
},
{
"epoch": 0.9164540108731832,
"grad_norm": 1.6484375,
"learning_rate": 8.001674876548471e-07,
"loss": 0.4939,
"step": 8260
},
{
"epoch": 0.918673027848663,
"grad_norm": 1.640625,
"learning_rate": 7.585836319885525e-07,
"loss": 0.4888,
"step": 8280
},
{
"epoch": 0.9208920448241429,
"grad_norm": 1.578125,
"learning_rate": 7.180813089919403e-07,
"loss": 0.4352,
"step": 8300
},
{
"epoch": 0.9231110617996228,
"grad_norm": 1.6640625,
"learning_rate": 6.78663594523788e-07,
"loss": 0.5086,
"step": 8320
},
{
"epoch": 0.9253300787751026,
"grad_norm": 1.96875,
"learning_rate": 6.403334820746876e-07,
"loss": 0.4695,
"step": 8340
},
{
"epoch": 0.9275490957505825,
"grad_norm": 1.5234375,
"learning_rate": 6.030938825397225e-07,
"loss": 0.3946,
"step": 8360
},
{
"epoch": 0.9297681127260623,
"grad_norm": 1.609375,
"learning_rate": 5.669476239973975e-07,
"loss": 0.4797,
"step": 8380
},
{
"epoch": 0.9319871297015422,
"grad_norm": 1.796875,
"learning_rate": 5.318974514948672e-07,
"loss": 0.4274,
"step": 8400
},
{
"epoch": 0.9342061466770221,
"grad_norm": 1.453125,
"learning_rate": 4.979460268394726e-07,
"loss": 0.4369,
"step": 8420
},
{
"epoch": 0.936425163652502,
"grad_norm": 1.6640625,
"learning_rate": 4.6509592839659666e-07,
"loss": 0.4452,
"step": 8440
},
{
"epoch": 0.9386441806279818,
"grad_norm": 1.7734375,
"learning_rate": 4.333496508938506e-07,
"loss": 0.4949,
"step": 8460
},
{
"epoch": 0.9408631976034617,
"grad_norm": 2.078125,
"learning_rate": 4.02709605231627e-07,
"loss": 0.4815,
"step": 8480
},
{
"epoch": 0.9430822145789415,
"grad_norm": 2.109375,
"learning_rate": 3.731781182999983e-07,
"loss": 0.5533,
"step": 8500
},
{
"epoch": 0.9453012315544214,
"grad_norm": 1.5234375,
"learning_rate": 3.447574328020109e-07,
"loss": 0.4741,
"step": 8520
},
{
"epoch": 0.9475202485299012,
"grad_norm": 1.5546875,
"learning_rate": 3.1744970708337205e-07,
"loss": 0.5182,
"step": 8540
},
{
"epoch": 0.9497392655053811,
"grad_norm": 1.4296875,
"learning_rate": 2.912570149685323e-07,
"loss": 0.4612,
"step": 8560
},
{
"epoch": 0.951958282480861,
"grad_norm": 1.3984375,
"learning_rate": 2.661813456032014e-07,
"loss": 0.4253,
"step": 8580
},
{
"epoch": 0.9541772994563409,
"grad_norm": 2.125,
"learning_rate": 2.4222460330327933e-07,
"loss": 0.5029,
"step": 8600
},
{
"epoch": 0.9563963164318207,
"grad_norm": 1.9921875,
"learning_rate": 2.1938860741023858e-07,
"loss": 0.4444,
"step": 8620
},
{
"epoch": 0.9586153334073005,
"grad_norm": 1.390625,
"learning_rate": 1.9767509215296297e-07,
"loss": 0.4607,
"step": 8640
},
{
"epoch": 0.9608343503827804,
"grad_norm": 1.7890625,
"learning_rate": 1.7708570651604306e-07,
"loss": 0.5045,
"step": 8660
},
{
"epoch": 0.9630533673582603,
"grad_norm": 1.484375,
"learning_rate": 1.5762201411454626e-07,
"loss": 0.4525,
"step": 8680
},
{
"epoch": 0.9652723843337402,
"grad_norm": 1.28125,
"learning_rate": 1.3928549307527183e-07,
"loss": 0.4484,
"step": 8700
},
{
"epoch": 0.9674914013092201,
"grad_norm": 1.7265625,
"learning_rate": 1.2207753592450078e-07,
"loss": 0.4194,
"step": 8720
},
{
"epoch": 0.9697104182846998,
"grad_norm": 1.7734375,
"learning_rate": 1.059994494822386e-07,
"loss": 0.4822,
"step": 8740
},
{
"epoch": 0.9719294352601797,
"grad_norm": 1.7421875,
"learning_rate": 9.105245476297653e-08,
"loss": 0.4584,
"step": 8760
},
{
"epoch": 0.9741484522356596,
"grad_norm": 1.5546875,
"learning_rate": 7.723768688296217e-08,
"loss": 0.4508,
"step": 8780
},
{
"epoch": 0.9763674692111395,
"grad_norm": 1.4921875,
"learning_rate": 6.455619497399534e-08,
"loss": 0.4953,
"step": 8800
},
{
"epoch": 0.9785864861866194,
"grad_norm": 1.9375,
"learning_rate": 5.300894210375329e-08,
"loss": 0.4791,
"step": 8820
},
{
"epoch": 0.9808055031620991,
"grad_norm": 1.4765625,
"learning_rate": 4.259680520265596e-08,
"loss": 0.4083,
"step": 8840
},
{
"epoch": 0.983024520137579,
"grad_norm": 1.6328125,
"learning_rate": 3.3320574997267595e-08,
"loss": 0.449,
"step": 8860
},
{
"epoch": 0.9852435371130589,
"grad_norm": 1.84375,
"learning_rate": 2.5180955950243056e-08,
"loss": 0.5022,
"step": 8880
},
{
"epoch": 0.9874625540885388,
"grad_norm": 1.203125,
"learning_rate": 1.8178566206837334e-08,
"loss": 0.54,
"step": 8900
},
{
"epoch": 0.9896815710640187,
"grad_norm": 1.8203125,
"learning_rate": 1.231393754795307e-08,
"loss": 0.4788,
"step": 8920
},
{
"epoch": 0.9919005880394985,
"grad_norm": 1.5546875,
"learning_rate": 7.587515349762874e-09,
"loss": 0.5039,
"step": 8940
},
{
"epoch": 0.9941196050149783,
"grad_norm": 1.8984375,
"learning_rate": 3.9996585498797145e-09,
"loss": 0.5317,
"step": 8960
},
{
"epoch": 0.9963386219904582,
"grad_norm": 2.0,
"learning_rate": 1.550639620103711e-09,
"loss": 0.4578,
"step": 8980
},
{
"epoch": 0.9985576389659381,
"grad_norm": 1.5234375,
"learning_rate": 2.406445457253659e-10,
"loss": 0.474,
"step": 9000
},
{
"epoch": 1.0,
"step": 9013,
"total_flos": 5.485972481640161e+17,
"train_loss": 0.4824531834622566,
"train_runtime": 26326.043,
"train_samples_per_second": 0.685,
"train_steps_per_second": 0.342
}
],
"logging_steps": 20,
"max_steps": 9013,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 100,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 5.485972481640161e+17,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}