llama_8b_lima_43 / trainer_state.json
OpenLeecher's picture
End of training
b4be6ac verified
raw
history blame
40.5 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.9996249062265566,
"eval_steps": 80,
"global_step": 1066,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0046886721680420105,
"grad_norm": 131.48858057089643,
"learning_rate": 5.5e-07,
"loss": 1.2775,
"step": 5
},
{
"epoch": 0.009377344336084021,
"grad_norm": 33.55585411699123,
"learning_rate": 1.1e-06,
"loss": 1.3472,
"step": 10
},
{
"epoch": 0.014066016504126031,
"grad_norm": 10.392565895406074,
"learning_rate": 1.6499999999999999e-06,
"loss": 1.1186,
"step": 15
},
{
"epoch": 0.018754688672168042,
"grad_norm": 13.30940048870882,
"learning_rate": 2.2e-06,
"loss": 1.1381,
"step": 20
},
{
"epoch": 0.023443360840210054,
"grad_norm": 25.46432441054564,
"learning_rate": 2.75e-06,
"loss": 1.1455,
"step": 25
},
{
"epoch": 0.028132033008252063,
"grad_norm": 26.00538399730873,
"learning_rate": 3.2999999999999997e-06,
"loss": 1.049,
"step": 30
},
{
"epoch": 0.032820705176294075,
"grad_norm": 9.35384218444162,
"learning_rate": 3.8499999999999996e-06,
"loss": 0.8965,
"step": 35
},
{
"epoch": 0.037509377344336084,
"grad_norm": 2.9429262000270113,
"learning_rate": 4.4e-06,
"loss": 1.1271,
"step": 40
},
{
"epoch": 0.04219804951237809,
"grad_norm": 3.2422986577839716,
"learning_rate": 4.95e-06,
"loss": 1.0061,
"step": 45
},
{
"epoch": 0.04688672168042011,
"grad_norm": 6.503070007858331,
"learning_rate": 5.5e-06,
"loss": 0.9871,
"step": 50
},
{
"epoch": 0.05157539384846212,
"grad_norm": 5.434874712126211,
"learning_rate": 5.414406436166232e-06,
"loss": 1.0584,
"step": 55
},
{
"epoch": 0.056264066016504126,
"grad_norm": 2.394040425022367,
"learning_rate": 5.32986463435603e-06,
"loss": 1.0672,
"step": 60
},
{
"epoch": 0.060952738184546135,
"grad_norm": 2.7812207561170776,
"learning_rate": 5.246366801851234e-06,
"loss": 0.9798,
"step": 65
},
{
"epoch": 0.06564141035258815,
"grad_norm": 3.4321299409489927,
"learning_rate": 5.163905165275343e-06,
"loss": 1.0684,
"step": 70
},
{
"epoch": 0.07033008252063015,
"grad_norm": 2.4331799242183005,
"learning_rate": 5.082471970641763e-06,
"loss": 1.1214,
"step": 75
},
{
"epoch": 0.07501875468867217,
"grad_norm": 2.66350051316822,
"learning_rate": 5.002059483402411e-06,
"loss": 1.0422,
"step": 80
},
{
"epoch": 0.07501875468867217,
"eval_loss": 1.0044387578964233,
"eval_runtime": 21.7707,
"eval_samples_per_second": 9.187,
"eval_steps_per_second": 2.297,
"step": 80
},
{
"epoch": 0.07970742685671418,
"grad_norm": 2.457085516495112,
"learning_rate": 4.922659988496696e-06,
"loss": 1.004,
"step": 85
},
{
"epoch": 0.08439609902475619,
"grad_norm": 2.673304521252877,
"learning_rate": 4.844265790400869e-06,
"loss": 1.1774,
"step": 90
},
{
"epoch": 0.0890847711927982,
"grad_norm": 2.9697389911450105,
"learning_rate": 4.766869213177739e-06,
"loss": 1.029,
"step": 95
},
{
"epoch": 0.09377344336084022,
"grad_norm": 2.7485846803270175,
"learning_rate": 4.690462600526791e-06,
"loss": 1.0735,
"step": 100
},
{
"epoch": 0.09846211552888222,
"grad_norm": 2.3597539988620047,
"learning_rate": 4.615038315834675e-06,
"loss": 1.0426,
"step": 105
},
{
"epoch": 0.10315078769692423,
"grad_norm": 10.287138782877946,
"learning_rate": 4.5405887422260886e-06,
"loss": 1.1257,
"step": 110
},
{
"epoch": 0.10783945986496624,
"grad_norm": 6.22099162659968,
"learning_rate": 4.467106282615065e-06,
"loss": 1.0203,
"step": 115
},
{
"epoch": 0.11252813203300825,
"grad_norm": 6.918830274001748,
"learning_rate": 4.394583359756651e-06,
"loss": 1.1145,
"step": 120
},
{
"epoch": 0.11721680420105027,
"grad_norm": 5.8742101902111,
"learning_rate": 4.323012416298999e-06,
"loss": 1.0943,
"step": 125
},
{
"epoch": 0.12190547636909227,
"grad_norm": 2.6642490667255525,
"learning_rate": 4.252385914835873e-06,
"loss": 1.087,
"step": 130
},
{
"epoch": 0.12659414853713427,
"grad_norm": 2.6965562685918023,
"learning_rate": 4.182696337959566e-06,
"loss": 1.0873,
"step": 135
},
{
"epoch": 0.1312828207051763,
"grad_norm": 2.951044122767734,
"learning_rate": 4.113936188314245e-06,
"loss": 0.8704,
"step": 140
},
{
"epoch": 0.1359714928732183,
"grad_norm": 2.101235354697959,
"learning_rate": 4.046097988649726e-06,
"loss": 1.1235,
"step": 145
},
{
"epoch": 0.1406601650412603,
"grad_norm": 3.2330279427425617,
"learning_rate": 3.979174281875685e-06,
"loss": 0.997,
"step": 150
},
{
"epoch": 0.14534883720930233,
"grad_norm": 3.3889833141888728,
"learning_rate": 3.9131576311163e-06,
"loss": 1.0179,
"step": 155
},
{
"epoch": 0.15003750937734434,
"grad_norm": 4.973298096388189,
"learning_rate": 3.848040619765356e-06,
"loss": 0.9127,
"step": 160
},
{
"epoch": 0.15003750937734434,
"eval_loss": 0.9606707692146301,
"eval_runtime": 20.2418,
"eval_samples_per_second": 9.881,
"eval_steps_per_second": 2.47,
"step": 160
},
{
"epoch": 0.15472618154538634,
"grad_norm": 8.29768606535662,
"learning_rate": 3.7838158515417857e-06,
"loss": 1.0887,
"step": 165
},
{
"epoch": 0.15941485371342837,
"grad_norm": 3.7810993101840147,
"learning_rate": 3.7204759505456866e-06,
"loss": 0.8966,
"step": 170
},
{
"epoch": 0.16410352588147037,
"grad_norm": 2.644220736384783,
"learning_rate": 3.65801356131479e-06,
"loss": 1.0543,
"step": 175
},
{
"epoch": 0.16879219804951237,
"grad_norm": 3.614698732279852,
"learning_rate": 3.596421348881407e-06,
"loss": 0.9387,
"step": 180
},
{
"epoch": 0.1734808702175544,
"grad_norm": 2.8015834886932787,
"learning_rate": 3.535691998829856e-06,
"loss": 1.0746,
"step": 185
},
{
"epoch": 0.1781695423855964,
"grad_norm": 2.564293653383082,
"learning_rate": 3.4758182173543725e-06,
"loss": 0.8235,
"step": 190
},
{
"epoch": 0.1828582145536384,
"grad_norm": 3.118218626122133,
"learning_rate": 3.4167927313175065e-06,
"loss": 0.9392,
"step": 195
},
{
"epoch": 0.18754688672168043,
"grad_norm": 3.1320015313206078,
"learning_rate": 3.358608288309036e-06,
"loss": 0.9951,
"step": 200
},
{
"epoch": 0.19223555888972244,
"grad_norm": 2.8693840620672026,
"learning_rate": 3.3012576567053635e-06,
"loss": 1.0835,
"step": 205
},
{
"epoch": 0.19692423105776444,
"grad_norm": 3.961730546242666,
"learning_rate": 3.2447336257294427e-06,
"loss": 0.8606,
"step": 210
},
{
"epoch": 0.20161290322580644,
"grad_norm": 3.380982676943891,
"learning_rate": 3.189029005511225e-06,
"loss": 1.0039,
"step": 215
},
{
"epoch": 0.20630157539384847,
"grad_norm": 3.072769902074738,
"learning_rate": 3.134136627148626e-06,
"loss": 0.9235,
"step": 220
},
{
"epoch": 0.21099024756189047,
"grad_norm": 2.245002916223262,
"learning_rate": 3.080049342769041e-06,
"loss": 1.06,
"step": 225
},
{
"epoch": 0.21567891972993247,
"grad_norm": 2.541443132900714,
"learning_rate": 3.026760025591393e-06,
"loss": 0.9432,
"step": 230
},
{
"epoch": 0.2203675918979745,
"grad_norm": 3.123240987262189,
"learning_rate": 2.97426156998874e-06,
"loss": 1.0341,
"step": 235
},
{
"epoch": 0.2250562640660165,
"grad_norm": 2.404173594460732,
"learning_rate": 2.9225468915514425e-06,
"loss": 0.8605,
"step": 240
},
{
"epoch": 0.2250562640660165,
"eval_loss": 0.944010317325592,
"eval_runtime": 20.3476,
"eval_samples_per_second": 9.829,
"eval_steps_per_second": 2.457,
"step": 240
},
{
"epoch": 0.2297449362340585,
"grad_norm": 2.839083125403713,
"learning_rate": 2.8716089271509e-06,
"loss": 0.986,
"step": 245
},
{
"epoch": 0.23443360840210054,
"grad_norm": 3.109434158155652,
"learning_rate": 2.8214406350038632e-06,
"loss": 0.8936,
"step": 250
},
{
"epoch": 0.23912228057014254,
"grad_norm": 2.5603812627943325,
"learning_rate": 2.772034994737337e-06,
"loss": 1.0787,
"step": 255
},
{
"epoch": 0.24381095273818454,
"grad_norm": 2.843918945845048,
"learning_rate": 2.7233850074540736e-06,
"loss": 0.8681,
"step": 260
},
{
"epoch": 0.24849962490622657,
"grad_norm": 3.295967379586164,
"learning_rate": 2.6754836957986757e-06,
"loss": 1.063,
"step": 265
},
{
"epoch": 0.25318829707426854,
"grad_norm": 2.6450503587985867,
"learning_rate": 2.6283241040243133e-06,
"loss": 0.9345,
"step": 270
},
{
"epoch": 0.25787696924231057,
"grad_norm": 2.6669681551281936,
"learning_rate": 2.5818992980600576e-06,
"loss": 0.9366,
"step": 275
},
{
"epoch": 0.2625656414103526,
"grad_norm": 2.365910075512645,
"learning_rate": 2.5362023655788563e-06,
"loss": 0.9222,
"step": 280
},
{
"epoch": 0.2672543135783946,
"grad_norm": 2.4773250316490336,
"learning_rate": 2.491226416066151e-06,
"loss": 0.9816,
"step": 285
},
{
"epoch": 0.2719429857464366,
"grad_norm": 2.051844651359742,
"learning_rate": 2.4469645808891426e-06,
"loss": 1.0592,
"step": 290
},
{
"epoch": 0.27663165791447863,
"grad_norm": 2.7110963822173875,
"learning_rate": 2.40341001336673e-06,
"loss": 0.9458,
"step": 295
},
{
"epoch": 0.2813203300825206,
"grad_norm": 2.1769151172554047,
"learning_rate": 2.3605558888401135e-06,
"loss": 0.8555,
"step": 300
},
{
"epoch": 0.28600900225056264,
"grad_norm": 2.626727731539155,
"learning_rate": 2.318395404744094e-06,
"loss": 1.1516,
"step": 305
},
{
"epoch": 0.29069767441860467,
"grad_norm": 2.306595833971501,
"learning_rate": 2.276921780679061e-06,
"loss": 0.8737,
"step": 310
},
{
"epoch": 0.29538634658664664,
"grad_norm": 2.569147248252582,
"learning_rate": 2.2361282584836925e-06,
"loss": 1.0032,
"step": 315
},
{
"epoch": 0.30007501875468867,
"grad_norm": 1.9670956149750123,
"learning_rate": 2.1960081023083778e-06,
"loss": 0.9068,
"step": 320
},
{
"epoch": 0.30007501875468867,
"eval_loss": 0.9342896938323975,
"eval_runtime": 20.3147,
"eval_samples_per_second": 9.845,
"eval_steps_per_second": 2.461,
"step": 320
},
{
"epoch": 0.3047636909227307,
"grad_norm": 2.0859945243647395,
"learning_rate": 2.156554598689365e-06,
"loss": 1.0191,
"step": 325
},
{
"epoch": 0.3094523630907727,
"grad_norm": 1.9655466100226573,
"learning_rate": 2.117761056623659e-06,
"loss": 0.9328,
"step": 330
},
{
"epoch": 0.3141410352588147,
"grad_norm": 2.581123721405676,
"learning_rate": 2.0796208076446752e-06,
"loss": 1.0101,
"step": 335
},
{
"epoch": 0.31882970742685673,
"grad_norm": 2.443562264938451,
"learning_rate": 2.0421272058986607e-06,
"loss": 0.9154,
"step": 340
},
{
"epoch": 0.3235183795948987,
"grad_norm": 2.3271716313484987,
"learning_rate": 2.0052736282219008e-06,
"loss": 1.044,
"step": 345
},
{
"epoch": 0.32820705176294074,
"grad_norm": 2.4236414480748834,
"learning_rate": 1.9690534742187182e-06,
"loss": 1.0251,
"step": 350
},
{
"epoch": 0.33289572393098277,
"grad_norm": 2.5311887907377253,
"learning_rate": 1.9334601663402865e-06,
"loss": 0.9126,
"step": 355
},
{
"epoch": 0.33758439609902474,
"grad_norm": 2.19719594303758,
"learning_rate": 1.898487149964267e-06,
"loss": 0.9877,
"step": 360
},
{
"epoch": 0.34227306826706677,
"grad_norm": 3.0147002131335805,
"learning_rate": 1.8641278934752799e-06,
"loss": 0.9057,
"step": 365
},
{
"epoch": 0.3469617404351088,
"grad_norm": 2.134323364449907,
"learning_rate": 1.8303758883462328e-06,
"loss": 1.0681,
"step": 370
},
{
"epoch": 0.3516504126031508,
"grad_norm": 2.3509776870701424,
"learning_rate": 1.7972246492205194e-06,
"loss": 0.9568,
"step": 375
},
{
"epoch": 0.3563390847711928,
"grad_norm": 2.1090286048701814,
"learning_rate": 1.7646677139950976e-06,
"loss": 1.032,
"step": 380
},
{
"epoch": 0.36102775693923483,
"grad_norm": 2.36149114890146,
"learning_rate": 1.7326986439044696e-06,
"loss": 0.9867,
"step": 385
},
{
"epoch": 0.3657164291072768,
"grad_norm": 2.255024349073518,
"learning_rate": 1.701311023605583e-06,
"loss": 0.8747,
"step": 390
},
{
"epoch": 0.37040510127531884,
"grad_norm": 2.6380764656009923,
"learning_rate": 1.6704984612636572e-06,
"loss": 0.9224,
"step": 395
},
{
"epoch": 0.37509377344336087,
"grad_norm": 2.4870934478859126,
"learning_rate": 1.6402545886389659e-06,
"loss": 0.9147,
"step": 400
},
{
"epoch": 0.37509377344336087,
"eval_loss": 0.9293374419212341,
"eval_runtime": 20.325,
"eval_samples_per_second": 9.84,
"eval_steps_per_second": 2.46,
"step": 400
},
{
"epoch": 0.37978244561140284,
"grad_norm": 2.5774447639043236,
"learning_rate": 1.610573061174586e-06,
"loss": 0.9285,
"step": 405
},
{
"epoch": 0.38447111777944487,
"grad_norm": 2.2678788716470817,
"learning_rate": 1.5814475580851346e-06,
"loss": 0.9994,
"step": 410
},
{
"epoch": 0.38915978994748684,
"grad_norm": 2.8722747503884114,
"learning_rate": 1.5528717824465089e-06,
"loss": 0.8864,
"step": 415
},
{
"epoch": 0.3938484621155289,
"grad_norm": 2.436535560556849,
"learning_rate": 1.5248394612866496e-06,
"loss": 1.1302,
"step": 420
},
{
"epoch": 0.3985371342835709,
"grad_norm": 1.9720574518399083,
"learning_rate": 1.4973443456773522e-06,
"loss": 0.9394,
"step": 425
},
{
"epoch": 0.4032258064516129,
"grad_norm": 2.6156706265956258,
"learning_rate": 1.4703802108271373e-06,
"loss": 0.7922,
"step": 430
},
{
"epoch": 0.4079144786196549,
"grad_norm": 2.295310925977984,
"learning_rate": 1.4439408561752077e-06,
"loss": 0.9312,
"step": 435
},
{
"epoch": 0.41260315078769694,
"grad_norm": 2.147812597078933,
"learning_rate": 1.4180201054865116e-06,
"loss": 0.9859,
"step": 440
},
{
"epoch": 0.4172918229557389,
"grad_norm": 2.029009456836816,
"learning_rate": 1.392611806947934e-06,
"loss": 1.0307,
"step": 445
},
{
"epoch": 0.42198049512378094,
"grad_norm": 2.590837380389044,
"learning_rate": 1.3677098332656357e-06,
"loss": 0.92,
"step": 450
},
{
"epoch": 0.42666916729182297,
"grad_norm": 2.3763009955883505,
"learning_rate": 1.3433080817635696e-06,
"loss": 1.0955,
"step": 455
},
{
"epoch": 0.43135783945986494,
"grad_norm": 2.80941534379534,
"learning_rate": 1.3194004744831898e-06,
"loss": 0.903,
"step": 460
},
{
"epoch": 0.436046511627907,
"grad_norm": 2.14656443429063,
"learning_rate": 1.2959809582843855e-06,
"loss": 0.8284,
"step": 465
},
{
"epoch": 0.440735183795949,
"grad_norm": 2.7676121711339423,
"learning_rate": 1.273043504947661e-06,
"loss": 1.0462,
"step": 470
},
{
"epoch": 0.445423855963991,
"grad_norm": 2.44008408182776,
"learning_rate": 1.2505821112775862e-06,
"loss": 0.979,
"step": 475
},
{
"epoch": 0.450112528132033,
"grad_norm": 2.584889518443889,
"learning_rate": 1.2285907992075474e-06,
"loss": 1.0192,
"step": 480
},
{
"epoch": 0.450112528132033,
"eval_loss": 0.9250276684761047,
"eval_runtime": 20.5471,
"eval_samples_per_second": 9.734,
"eval_steps_per_second": 2.433,
"step": 480
},
{
"epoch": 0.45480120030007504,
"grad_norm": 2.496373694809835,
"learning_rate": 1.207063615905829e-06,
"loss": 0.9037,
"step": 485
},
{
"epoch": 0.459489872468117,
"grad_norm": 2.371873747415523,
"learning_rate": 1.1859946338830404e-06,
"loss": 1.0312,
"step": 490
},
{
"epoch": 0.46417854463615904,
"grad_norm": 2.2570285962672894,
"learning_rate": 1.1653779511009372e-06,
"loss": 0.9113,
"step": 495
},
{
"epoch": 0.46886721680420107,
"grad_norm": 2.6279198777243606,
"learning_rate": 1.145207691082648e-06,
"loss": 0.8337,
"step": 500
},
{
"epoch": 0.47355588897224304,
"grad_norm": 2.7761777412837767,
"learning_rate": 1.1254780030243539e-06,
"loss": 0.9602,
"step": 505
},
{
"epoch": 0.4782445611402851,
"grad_norm": 2.5445962149324663,
"learning_rate": 1.1061830619084358e-06,
"loss": 0.9804,
"step": 510
},
{
"epoch": 0.4829332333083271,
"grad_norm": 2.3633364319931514,
"learning_rate": 1.087317068618139e-06,
"loss": 0.977,
"step": 515
},
{
"epoch": 0.4876219054763691,
"grad_norm": 1.9986528249443352,
"learning_rate": 1.0688742500537784e-06,
"loss": 0.9425,
"step": 520
},
{
"epoch": 0.4923105776444111,
"grad_norm": 2.4501412442964803,
"learning_rate": 1.0508488592505175e-06,
"loss": 0.9995,
"step": 525
},
{
"epoch": 0.49699924981245314,
"grad_norm": 2.4394994412302786,
"learning_rate": 1.0332351754977698e-06,
"loss": 0.9329,
"step": 530
},
{
"epoch": 0.5016879219804952,
"grad_norm": 2.2074019259444544,
"learning_rate": 1.016027504460246e-06,
"loss": 0.9203,
"step": 535
},
{
"epoch": 0.5063765941485371,
"grad_norm": 2.610216953517823,
"learning_rate": 9.992201783006927e-07,
"loss": 1.0291,
"step": 540
},
{
"epoch": 0.5110652663165791,
"grad_norm": 2.718383436345948,
"learning_rate": 9.828075558043617e-07,
"loss": 0.9292,
"step": 545
},
{
"epoch": 0.5157539384846211,
"grad_norm": 3.6229517535213143,
"learning_rate": 9.667840225052484e-07,
"loss": 1.0165,
"step": 550
},
{
"epoch": 0.5204426106526632,
"grad_norm": 2.566929441984276,
"learning_rate": 9.511439908141446e-07,
"loss": 1.012,
"step": 555
},
{
"epoch": 0.5251312828207052,
"grad_norm": 2.2133821736785015,
"learning_rate": 9.358819001485473e-07,
"loss": 0.8303,
"step": 560
},
{
"epoch": 0.5251312828207052,
"eval_loss": 0.9191934466362,
"eval_runtime": 20.5275,
"eval_samples_per_second": 9.743,
"eval_steps_per_second": 2.436,
"step": 560
},
{
"epoch": 0.5298199549887472,
"grad_norm": 2.530396363584249,
"learning_rate": 9.209922170644708e-07,
"loss": 1.022,
"step": 565
},
{
"epoch": 0.5345086271567892,
"grad_norm": 3.0761740344436017,
"learning_rate": 9.06469435390206e-07,
"loss": 0.9437,
"step": 570
},
{
"epoch": 0.5391972993248312,
"grad_norm": 2.679955680751534,
"learning_rate": 8.923080763620794e-07,
"loss": 0.9776,
"step": 575
},
{
"epoch": 0.5438859714928732,
"grad_norm": 2.477813948279551,
"learning_rate": 8.785026887622588e-07,
"loss": 0.8858,
"step": 580
},
{
"epoch": 0.5485746436609152,
"grad_norm": 3.021641574122816,
"learning_rate": 8.650478490586582e-07,
"loss": 0.9392,
"step": 585
},
{
"epoch": 0.5532633158289573,
"grad_norm": 3.0622821568316985,
"learning_rate": 8.519381615469985e-07,
"loss": 1.0067,
"step": 590
},
{
"epoch": 0.5579519879969993,
"grad_norm": 2.0240771327720606,
"learning_rate": 8.391682584950767e-07,
"loss": 0.8645,
"step": 595
},
{
"epoch": 0.5626406601650412,
"grad_norm": 2.4592599613454453,
"learning_rate": 8.267328002892997e-07,
"loss": 0.8116,
"step": 600
},
{
"epoch": 0.5673293323330832,
"grad_norm": 2.3279060484215304,
"learning_rate": 8.146264755835511e-07,
"loss": 1.0685,
"step": 605
},
{
"epoch": 0.5720180045011253,
"grad_norm": 2.1915343114528394,
"learning_rate": 8.028440014504431e-07,
"loss": 1.0312,
"step": 610
},
{
"epoch": 0.5767066766691673,
"grad_norm": 2.353311242466667,
"learning_rate": 7.913801235350256e-07,
"loss": 0.9753,
"step": 615
},
{
"epoch": 0.5813953488372093,
"grad_norm": 1.9822254771453949,
"learning_rate": 7.80229616211014e-07,
"loss": 1.0626,
"step": 620
},
{
"epoch": 0.5860840210052514,
"grad_norm": 2.229570765456582,
"learning_rate": 7.693872827396111e-07,
"loss": 0.8915,
"step": 625
},
{
"epoch": 0.5907726931732933,
"grad_norm": 2.1333212031989017,
"learning_rate": 7.58847955430991e-07,
"loss": 0.9256,
"step": 630
},
{
"epoch": 0.5954613653413353,
"grad_norm": 2.3285962864619107,
"learning_rate": 7.486064958085216e-07,
"loss": 0.8844,
"step": 635
},
{
"epoch": 0.6001500375093773,
"grad_norm": 2.4239895162113654,
"learning_rate": 7.386577947758049e-07,
"loss": 1.0284,
"step": 640
},
{
"epoch": 0.6001500375093773,
"eval_loss": 0.9292100667953491,
"eval_runtime": 20.2551,
"eval_samples_per_second": 9.874,
"eval_steps_per_second": 2.469,
"step": 640
},
{
"epoch": 0.6048387096774194,
"grad_norm": 2.592229243089013,
"learning_rate": 7.289967727866171e-07,
"loss": 0.8607,
"step": 645
},
{
"epoch": 0.6095273818454614,
"grad_norm": 2.557096114560489,
"learning_rate": 7.196183800178289e-07,
"loss": 1.0461,
"step": 650
},
{
"epoch": 0.6142160540135033,
"grad_norm": 2.7258364788939278,
"learning_rate": 7.105175965454019e-07,
"loss": 0.8923,
"step": 655
},
{
"epoch": 0.6189047261815454,
"grad_norm": 2.500897088452433,
"learning_rate": 7.016894325235454e-07,
"loss": 0.908,
"step": 660
},
{
"epoch": 0.6235933983495874,
"grad_norm": 3.4364862124732873,
"learning_rate": 6.931289283671353e-07,
"loss": 0.8488,
"step": 665
},
{
"epoch": 0.6282820705176294,
"grad_norm": 2.6544071650230405,
"learning_rate": 6.84831154937491e-07,
"loss": 0.815,
"step": 670
},
{
"epoch": 0.6329707426856714,
"grad_norm": 2.9017321933643356,
"learning_rate": 6.767912137316187e-07,
"loss": 1.0641,
"step": 675
},
{
"epoch": 0.6376594148537135,
"grad_norm": 3.2461446427637854,
"learning_rate": 6.690042370750264e-07,
"loss": 0.9388,
"step": 680
},
{
"epoch": 0.6423480870217554,
"grad_norm": 2.5936370960610398,
"learning_rate": 6.614653883182271e-07,
"loss": 0.83,
"step": 685
},
{
"epoch": 0.6470367591897974,
"grad_norm": 3.027451608426823,
"learning_rate": 6.541698620370481e-07,
"loss": 0.9852,
"step": 690
},
{
"epoch": 0.6517254313578394,
"grad_norm": 2.8965181885799933,
"learning_rate": 6.471128842368711e-07,
"loss": 0.8914,
"step": 695
},
{
"epoch": 0.6564141035258815,
"grad_norm": 2.9030689776107,
"learning_rate": 6.402897125609332e-07,
"loss": 0.9833,
"step": 700
},
{
"epoch": 0.6611027756939235,
"grad_norm": 2.6265372108049303,
"learning_rate": 6.336956365028259e-07,
"loss": 1.0902,
"step": 705
},
{
"epoch": 0.6657914478619655,
"grad_norm": 2.186738270502514,
"learning_rate": 6.273259776233337e-07,
"loss": 0.8316,
"step": 710
},
{
"epoch": 0.6704801200300075,
"grad_norm": 10.89006237530139,
"learning_rate": 6.211760897717641e-07,
"loss": 1.0283,
"step": 715
},
{
"epoch": 0.6751687921980495,
"grad_norm": 2.7796220586562344,
"learning_rate": 6.152413593119235e-07,
"loss": 0.9183,
"step": 720
},
{
"epoch": 0.6751687921980495,
"eval_loss": 0.9389083385467529,
"eval_runtime": 20.3673,
"eval_samples_per_second": 9.82,
"eval_steps_per_second": 2.455,
"step": 720
},
{
"epoch": 0.6798574643660915,
"grad_norm": 2.906995236208513,
"learning_rate": 6.095172053529076e-07,
"loss": 1.034,
"step": 725
},
{
"epoch": 0.6845461365341335,
"grad_norm": 2.3367410989611943,
"learning_rate": 6.039990799848741e-07,
"loss": 0.898,
"step": 730
},
{
"epoch": 0.6892348087021756,
"grad_norm": 2.3266907504669976,
"learning_rate": 5.986824685199863e-07,
"loss": 0.7855,
"step": 735
},
{
"epoch": 0.6939234808702176,
"grad_norm": 2.1566038379464842,
"learning_rate": 5.935628897387149e-07,
"loss": 1.0417,
"step": 740
},
{
"epoch": 0.6986121530382595,
"grad_norm": 2.3519944231674685,
"learning_rate": 5.886358961416999e-07,
"loss": 0.9567,
"step": 745
},
{
"epoch": 0.7033008252063015,
"grad_norm": 2.5749570284624457,
"learning_rate": 5.838970742073876e-07,
"loss": 0.7935,
"step": 750
},
{
"epoch": 0.7079894973743436,
"grad_norm": 2.0032057277402844,
"learning_rate": 5.793420446556638e-07,
"loss": 0.9967,
"step": 755
},
{
"epoch": 0.7126781695423856,
"grad_norm": 2.8692030696523703,
"learning_rate": 5.74966462717722e-07,
"loss": 1.0786,
"step": 760
},
{
"epoch": 0.7173668417104276,
"grad_norm": 2.1378421674086994,
"learning_rate": 5.707660184124143e-07,
"loss": 1.115,
"step": 765
},
{
"epoch": 0.7220555138784697,
"grad_norm": 2.344543312884903,
"learning_rate": 5.667364368293497e-07,
"loss": 0.9502,
"step": 770
},
{
"epoch": 0.7267441860465116,
"grad_norm": 1.9323329729740275,
"learning_rate": 5.6287347841902e-07,
"loss": 0.9964,
"step": 775
},
{
"epoch": 0.7314328582145536,
"grad_norm": 2.948582260278745,
"learning_rate": 5.591729392902467e-07,
"loss": 1.0077,
"step": 780
},
{
"epoch": 0.7361215303825956,
"grad_norm": 2.1985962635350442,
"learning_rate": 5.556306515152638e-07,
"loss": 0.7058,
"step": 785
},
{
"epoch": 0.7408102025506377,
"grad_norm": 4.099442910663203,
"learning_rate": 5.522424834427688e-07,
"loss": 0.856,
"step": 790
},
{
"epoch": 0.7454988747186797,
"grad_norm": 2.3260610179486787,
"learning_rate": 5.490043400192936e-07,
"loss": 0.9852,
"step": 795
},
{
"epoch": 0.7501875468867217,
"grad_norm": 2.2249724666107693,
"learning_rate": 5.459121631192727e-07,
"loss": 0.9897,
"step": 800
},
{
"epoch": 0.7501875468867217,
"eval_loss": 0.9337027072906494,
"eval_runtime": 20.543,
"eval_samples_per_second": 9.736,
"eval_steps_per_second": 2.434,
"step": 800
},
{
"epoch": 0.7548762190547637,
"grad_norm": 2.3238862226735244,
"learning_rate": 5.429619318842062e-07,
"loss": 0.9693,
"step": 805
},
{
"epoch": 0.7595648912228057,
"grad_norm": 2.5983979780721005,
"learning_rate": 5.401496630713439e-07,
"loss": 0.9363,
"step": 810
},
{
"epoch": 0.7642535633908477,
"grad_norm": 3.203392479652258,
"learning_rate": 5.374714114123462e-07,
"loss": 0.9007,
"step": 815
},
{
"epoch": 0.7689422355588897,
"grad_norm": 3.0116714434819523,
"learning_rate": 5.34923269982403e-07,
"loss": 0.8538,
"step": 820
},
{
"epoch": 0.7736309077269318,
"grad_norm": 2.735036803752382,
"learning_rate": 5.325013705803326e-07,
"loss": 0.9429,
"step": 825
},
{
"epoch": 0.7783195798949737,
"grad_norm": 2.7192900196409693,
"learning_rate": 5.302018841202155e-07,
"loss": 0.862,
"step": 830
},
{
"epoch": 0.7830082520630157,
"grad_norm": 3.310126483594716,
"learning_rate": 5.28021021035156e-07,
"loss": 0.8743,
"step": 835
},
{
"epoch": 0.7876969242310577,
"grad_norm": 3.5788231445835623,
"learning_rate": 5.25955031693814e-07,
"loss": 0.8728,
"step": 840
},
{
"epoch": 0.7923855963990998,
"grad_norm": 2.3031948200618784,
"learning_rate": 5.240002068303935e-07,
"loss": 0.9804,
"step": 845
},
{
"epoch": 0.7970742685671418,
"grad_norm": 2.27441145876717,
"learning_rate": 5.22152877988829e-07,
"loss": 0.9592,
"step": 850
},
{
"epoch": 0.8017629407351838,
"grad_norm": 2.399760421263353,
"learning_rate": 5.204094179819663e-07,
"loss": 1.1113,
"step": 855
},
{
"epoch": 0.8064516129032258,
"grad_norm": 2.786650153581553,
"learning_rate": 5.187662413666055e-07,
"loss": 0.896,
"step": 860
},
{
"epoch": 0.8111402850712678,
"grad_norm": 2.3375248528715886,
"learning_rate": 5.137741469209312e-07,
"loss": 0.9484,
"step": 865
},
{
"epoch": 0.8158289572393098,
"grad_norm": 2.107239092918484,
"learning_rate": 5.034812576940423e-07,
"loss": 0.9107,
"step": 870
},
{
"epoch": 0.8205176294073518,
"grad_norm": 3.2120652639502136,
"learning_rate": 4.931883684671535e-07,
"loss": 1.0274,
"step": 875
},
{
"epoch": 0.8252063015753939,
"grad_norm": 2.1470972788703353,
"learning_rate": 4.828954792402647e-07,
"loss": 1.0209,
"step": 880
},
{
"epoch": 0.8252063015753939,
"eval_loss": 0.93896484375,
"eval_runtime": 20.3111,
"eval_samples_per_second": 9.847,
"eval_steps_per_second": 2.462,
"step": 880
},
{
"epoch": 0.8298949737434359,
"grad_norm": 2.0424483656754724,
"learning_rate": 4.7260259001337577e-07,
"loss": 0.9472,
"step": 885
},
{
"epoch": 0.8345836459114778,
"grad_norm": 2.5878530403298434,
"learning_rate": 4.6230970078648696e-07,
"loss": 0.9458,
"step": 890
},
{
"epoch": 0.8392723180795199,
"grad_norm": 2.497909901428293,
"learning_rate": 4.5201681155959816e-07,
"loss": 0.7928,
"step": 895
},
{
"epoch": 0.8439609902475619,
"grad_norm": 2.1704945555054462,
"learning_rate": 4.417239223327093e-07,
"loss": 0.8992,
"step": 900
},
{
"epoch": 0.8486496624156039,
"grad_norm": 1.9807380776486385,
"learning_rate": 4.314310331058205e-07,
"loss": 1.0681,
"step": 905
},
{
"epoch": 0.8533383345836459,
"grad_norm": 2.2057348642236625,
"learning_rate": 4.2113814387893164e-07,
"loss": 1.057,
"step": 910
},
{
"epoch": 0.858027006751688,
"grad_norm": 2.434197081864654,
"learning_rate": 4.108452546520428e-07,
"loss": 0.9609,
"step": 915
},
{
"epoch": 0.8627156789197299,
"grad_norm": 1.641371038860168,
"learning_rate": 4.00552365425154e-07,
"loss": 0.9775,
"step": 920
},
{
"epoch": 0.8674043510877719,
"grad_norm": 2.207714901698735,
"learning_rate": 3.9025947619826517e-07,
"loss": 0.7712,
"step": 925
},
{
"epoch": 0.872093023255814,
"grad_norm": 2.011772170698376,
"learning_rate": 3.7996658697137626e-07,
"loss": 0.9408,
"step": 930
},
{
"epoch": 0.876781695423856,
"grad_norm": 2.5252414435017747,
"learning_rate": 3.6967369774448745e-07,
"loss": 0.8363,
"step": 935
},
{
"epoch": 0.881470367591898,
"grad_norm": 2.3365392350792287,
"learning_rate": 3.5938080851759865e-07,
"loss": 0.9977,
"step": 940
},
{
"epoch": 0.88615903975994,
"grad_norm": 2.3882731527862346,
"learning_rate": 3.490879192907099e-07,
"loss": 0.9326,
"step": 945
},
{
"epoch": 0.890847711927982,
"grad_norm": 2.180049227770247,
"learning_rate": 3.38795030063821e-07,
"loss": 0.9909,
"step": 950
},
{
"epoch": 0.895536384096024,
"grad_norm": 2.4546116544168965,
"learning_rate": 3.285021408369321e-07,
"loss": 0.8207,
"step": 955
},
{
"epoch": 0.900225056264066,
"grad_norm": 3.062713753398875,
"learning_rate": 3.1820925161004326e-07,
"loss": 0.9118,
"step": 960
},
{
"epoch": 0.900225056264066,
"eval_loss": 0.9374144077301025,
"eval_runtime": 20.3928,
"eval_samples_per_second": 9.807,
"eval_steps_per_second": 2.452,
"step": 960
},
{
"epoch": 0.904913728432108,
"grad_norm": 2.9305276635268633,
"learning_rate": 3.0791636238315446e-07,
"loss": 0.8345,
"step": 965
},
{
"epoch": 0.9096024006001501,
"grad_norm": 2.9508689565092046,
"learning_rate": 2.9762347315626565e-07,
"loss": 0.7999,
"step": 970
},
{
"epoch": 0.9142910727681921,
"grad_norm": 3.9095096230897237,
"learning_rate": 2.873305839293768e-07,
"loss": 0.8812,
"step": 975
},
{
"epoch": 0.918979744936234,
"grad_norm": 2.954610613967905,
"learning_rate": 2.7703769470248794e-07,
"loss": 0.9124,
"step": 980
},
{
"epoch": 0.923668417104276,
"grad_norm": 2.3909984231204446,
"learning_rate": 2.6674480547559913e-07,
"loss": 0.8418,
"step": 985
},
{
"epoch": 0.9283570892723181,
"grad_norm": 2.407545423074375,
"learning_rate": 2.5645191624871027e-07,
"loss": 0.9194,
"step": 990
},
{
"epoch": 0.9330457614403601,
"grad_norm": 2.868767477778071,
"learning_rate": 2.4615902702182147e-07,
"loss": 0.9541,
"step": 995
},
{
"epoch": 0.9377344336084021,
"grad_norm": 2.1293635803564066,
"learning_rate": 2.3586613779493258e-07,
"loss": 0.8087,
"step": 1000
},
{
"epoch": 0.9424231057764441,
"grad_norm": 2.0557494272602264,
"learning_rate": 2.255732485680438e-07,
"loss": 0.9081,
"step": 1005
},
{
"epoch": 0.9471117779444861,
"grad_norm": 2.1864930223853056,
"learning_rate": 2.1528035934115495e-07,
"loss": 0.859,
"step": 1010
},
{
"epoch": 0.9518004501125281,
"grad_norm": 2.05122621453041,
"learning_rate": 2.0498747011426614e-07,
"loss": 0.8725,
"step": 1015
},
{
"epoch": 0.9564891222805701,
"grad_norm": 3.2726273841892923,
"learning_rate": 1.9469458088737728e-07,
"loss": 0.7904,
"step": 1020
},
{
"epoch": 0.9611777944486122,
"grad_norm": 2.259537550017741,
"learning_rate": 1.8440169166048842e-07,
"loss": 1.0669,
"step": 1025
},
{
"epoch": 0.9658664666166542,
"grad_norm": 2.406310953068651,
"learning_rate": 1.7410880243359964e-07,
"loss": 0.8434,
"step": 1030
},
{
"epoch": 0.9705551387846961,
"grad_norm": 1.9259292180053644,
"learning_rate": 1.6381591320671076e-07,
"loss": 1.0234,
"step": 1035
},
{
"epoch": 0.9752438109527382,
"grad_norm": 2.0241322819114504,
"learning_rate": 1.5352302397982195e-07,
"loss": 0.9077,
"step": 1040
},
{
"epoch": 0.9752438109527382,
"eval_loss": 0.9355975389480591,
"eval_runtime": 20.3082,
"eval_samples_per_second": 9.848,
"eval_steps_per_second": 2.462,
"step": 1040
},
{
"epoch": 0.9799324831207802,
"grad_norm": 1.9205593868577922,
"learning_rate": 1.432301347529331e-07,
"loss": 0.7646,
"step": 1045
},
{
"epoch": 0.9846211552888222,
"grad_norm": 3.2012232274694834,
"learning_rate": 1.329372455260443e-07,
"loss": 0.8593,
"step": 1050
},
{
"epoch": 0.9893098274568642,
"grad_norm": 2.3484948634638205,
"learning_rate": 1.2264435629915543e-07,
"loss": 1.0972,
"step": 1055
},
{
"epoch": 0.9939984996249063,
"grad_norm": 2.7160625332606685,
"learning_rate": 1.1235146707226661e-07,
"loss": 0.7634,
"step": 1060
},
{
"epoch": 0.9986871717929482,
"grad_norm": 2.3964216948732173,
"learning_rate": 1.0205857784537777e-07,
"loss": 0.9783,
"step": 1065
},
{
"epoch": 0.9996249062265566,
"step": 1066,
"total_flos": 1.5465201990107136e+17,
"train_loss": 0.9640034305221815,
"train_runtime": 13404.8173,
"train_samples_per_second": 2.386,
"train_steps_per_second": 0.08
}
],
"logging_steps": 5,
"max_steps": 1066,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 1066,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.5465201990107136e+17,
"train_batch_size": 3,
"trial_name": null,
"trial_params": null
}