cilooor's picture
Training in progress, step 200, checkpoint
12eb785 verified
{
"best_metric": 1.0697510242462158,
"best_model_checkpoint": "miner_id_24/checkpoint-200",
"epoch": 0.056589092452429796,
"eval_steps": 50,
"global_step": 200,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.00028294546226214896,
"grad_norm": 0.18001686036586761,
"learning_rate": 6e-06,
"loss": 1.2994,
"step": 1
},
{
"epoch": 0.00028294546226214896,
"eval_loss": 1.2716970443725586,
"eval_runtime": 846.7814,
"eval_samples_per_second": 7.03,
"eval_steps_per_second": 1.758,
"step": 1
},
{
"epoch": 0.0005658909245242979,
"grad_norm": 0.1731778383255005,
"learning_rate": 1.2e-05,
"loss": 1.3714,
"step": 2
},
{
"epoch": 0.0008488363867864469,
"grad_norm": 0.16550840437412262,
"learning_rate": 1.8e-05,
"loss": 1.2408,
"step": 3
},
{
"epoch": 0.0011317818490485959,
"grad_norm": 0.16617779433727264,
"learning_rate": 2.4e-05,
"loss": 1.398,
"step": 4
},
{
"epoch": 0.001414727311310745,
"grad_norm": 0.15455280244350433,
"learning_rate": 3e-05,
"loss": 1.2861,
"step": 5
},
{
"epoch": 0.0016976727735728938,
"grad_norm": 0.15682223439216614,
"learning_rate": 3.6e-05,
"loss": 1.293,
"step": 6
},
{
"epoch": 0.001980618235835043,
"grad_norm": 0.14510343968868256,
"learning_rate": 4.2e-05,
"loss": 1.2524,
"step": 7
},
{
"epoch": 0.0022635636980971917,
"grad_norm": 0.14818604290485382,
"learning_rate": 4.8e-05,
"loss": 1.2562,
"step": 8
},
{
"epoch": 0.0025465091603593406,
"grad_norm": 0.14594592154026031,
"learning_rate": 5.4000000000000005e-05,
"loss": 1.3486,
"step": 9
},
{
"epoch": 0.00282945462262149,
"grad_norm": 0.14386072754859924,
"learning_rate": 6e-05,
"loss": 1.2519,
"step": 10
},
{
"epoch": 0.0031124000848836387,
"grad_norm": 0.1916978359222412,
"learning_rate": 5.999589914977407e-05,
"loss": 1.1521,
"step": 11
},
{
"epoch": 0.0033953455471457876,
"grad_norm": 0.16554145514965057,
"learning_rate": 5.998359772022778e-05,
"loss": 1.248,
"step": 12
},
{
"epoch": 0.0036782910094079364,
"grad_norm": 0.1682235598564148,
"learning_rate": 5.996309907444915e-05,
"loss": 1.2722,
"step": 13
},
{
"epoch": 0.003961236471670086,
"grad_norm": 0.15518620610237122,
"learning_rate": 5.9934408816563236e-05,
"loss": 1.1769,
"step": 14
},
{
"epoch": 0.004244181933932234,
"grad_norm": 0.15147370100021362,
"learning_rate": 5.98975347902001e-05,
"loss": 1.3129,
"step": 15
},
{
"epoch": 0.0045271273961943834,
"grad_norm": 0.15708550810813904,
"learning_rate": 5.9852487076350345e-05,
"loss": 1.2437,
"step": 16
},
{
"epoch": 0.004810072858456533,
"grad_norm": 0.16699694097042084,
"learning_rate": 5.979927799060915e-05,
"loss": 1.2656,
"step": 17
},
{
"epoch": 0.005093018320718681,
"grad_norm": 0.15669851005077362,
"learning_rate": 5.9737922079809257e-05,
"loss": 1.2197,
"step": 18
},
{
"epoch": 0.0053759637829808304,
"grad_norm": 0.15826715528964996,
"learning_rate": 5.9668436118044054e-05,
"loss": 1.2848,
"step": 19
},
{
"epoch": 0.00565890924524298,
"grad_norm": 0.15940728783607483,
"learning_rate": 5.959083910208167e-05,
"loss": 1.2357,
"step": 20
},
{
"epoch": 0.005941854707505128,
"grad_norm": 0.14362244307994843,
"learning_rate": 5.9505152246171474e-05,
"loss": 1.1756,
"step": 21
},
{
"epoch": 0.0062248001697672774,
"grad_norm": 0.1448201835155487,
"learning_rate": 5.941139897624428e-05,
"loss": 1.1705,
"step": 22
},
{
"epoch": 0.006507745632029427,
"grad_norm": 0.14286337792873383,
"learning_rate": 5.9309604923507984e-05,
"loss": 1.256,
"step": 23
},
{
"epoch": 0.006790691094291575,
"grad_norm": 0.14091116189956665,
"learning_rate": 5.9199797917440176e-05,
"loss": 1.2419,
"step": 24
},
{
"epoch": 0.0070736365565537244,
"grad_norm": 0.1461900919675827,
"learning_rate": 5.908200797817991e-05,
"loss": 1.3056,
"step": 25
},
{
"epoch": 0.007356582018815873,
"grad_norm": 0.1513061672449112,
"learning_rate": 5.895626730832046e-05,
"loss": 1.2286,
"step": 26
},
{
"epoch": 0.007639527481078022,
"grad_norm": 0.14222566783428192,
"learning_rate": 5.882261028410545e-05,
"loss": 1.3284,
"step": 27
},
{
"epoch": 0.007922472943340171,
"grad_norm": 0.14666450023651123,
"learning_rate": 5.8681073446030734e-05,
"loss": 1.1703,
"step": 28
},
{
"epoch": 0.00820541840560232,
"grad_norm": 0.1654808074235916,
"learning_rate": 5.853169548885461e-05,
"loss": 1.1861,
"step": 29
},
{
"epoch": 0.008488363867864468,
"grad_norm": 0.1394944190979004,
"learning_rate": 5.8374517251019035e-05,
"loss": 1.2514,
"step": 30
},
{
"epoch": 0.008771309330126618,
"grad_norm": 0.16221028566360474,
"learning_rate": 5.820958170348484e-05,
"loss": 1.2141,
"step": 31
},
{
"epoch": 0.009054254792388767,
"grad_norm": 0.17233219742774963,
"learning_rate": 5.8036933937983825e-05,
"loss": 1.1461,
"step": 32
},
{
"epoch": 0.009337200254650915,
"grad_norm": 0.16100017726421356,
"learning_rate": 5.7856621154691217e-05,
"loss": 1.1342,
"step": 33
},
{
"epoch": 0.009620145716913065,
"grad_norm": 0.16232934594154358,
"learning_rate": 5.766869264932154e-05,
"loss": 1.0513,
"step": 34
},
{
"epoch": 0.009903091179175214,
"grad_norm": 0.169218510389328,
"learning_rate": 5.747319979965172e-05,
"loss": 1.176,
"step": 35
},
{
"epoch": 0.010186036641437362,
"grad_norm": 0.18459808826446533,
"learning_rate": 5.727019605147488e-05,
"loss": 1.1991,
"step": 36
},
{
"epoch": 0.010468982103699512,
"grad_norm": 0.21113476157188416,
"learning_rate": 5.7059736903988775e-05,
"loss": 1.1908,
"step": 37
},
{
"epoch": 0.010751927565961661,
"grad_norm": 0.255100816488266,
"learning_rate": 5.684187989462291e-05,
"loss": 1.0502,
"step": 38
},
{
"epoch": 0.01103487302822381,
"grad_norm": 0.2532814145088196,
"learning_rate": 5.661668458330836e-05,
"loss": 1.1723,
"step": 39
},
{
"epoch": 0.01131781849048596,
"grad_norm": 0.24460916221141815,
"learning_rate": 5.638421253619467e-05,
"loss": 1.092,
"step": 40
},
{
"epoch": 0.011600763952748108,
"grad_norm": 0.33700716495513916,
"learning_rate": 5.614452730881832e-05,
"loss": 1.0077,
"step": 41
},
{
"epoch": 0.011883709415010256,
"grad_norm": 0.3086978793144226,
"learning_rate": 5.589769442872722e-05,
"loss": 0.984,
"step": 42
},
{
"epoch": 0.012166654877272406,
"grad_norm": 0.34127745032310486,
"learning_rate": 5.5643781377566175e-05,
"loss": 0.9426,
"step": 43
},
{
"epoch": 0.012449600339534555,
"grad_norm": 0.3650180697441101,
"learning_rate": 5.538285757262806e-05,
"loss": 0.9358,
"step": 44
},
{
"epoch": 0.012732545801796703,
"grad_norm": 0.3661767542362213,
"learning_rate": 5.5114994347875856e-05,
"loss": 0.9236,
"step": 45
},
{
"epoch": 0.013015491264058853,
"grad_norm": 0.35510310530662537,
"learning_rate": 5.48402649344406e-05,
"loss": 0.9134,
"step": 46
},
{
"epoch": 0.013298436726321002,
"grad_norm": 0.3974863588809967,
"learning_rate": 5.455874444060078e-05,
"loss": 1.1377,
"step": 47
},
{
"epoch": 0.01358138218858315,
"grad_norm": 0.3589693009853363,
"learning_rate": 5.427050983124843e-05,
"loss": 1.2237,
"step": 48
},
{
"epoch": 0.013864327650845299,
"grad_norm": 0.36883828043937683,
"learning_rate": 5.397563990684774e-05,
"loss": 1.1407,
"step": 49
},
{
"epoch": 0.014147273113107449,
"grad_norm": 0.3884880840778351,
"learning_rate": 5.367421528189181e-05,
"loss": 1.1999,
"step": 50
},
{
"epoch": 0.014147273113107449,
"eval_loss": 1.1299543380737305,
"eval_runtime": 853.0261,
"eval_samples_per_second": 6.979,
"eval_steps_per_second": 1.746,
"step": 50
},
{
"epoch": 0.014430218575369597,
"grad_norm": 0.2056826651096344,
"learning_rate": 5.336631836286338e-05,
"loss": 1.2011,
"step": 51
},
{
"epoch": 0.014713164037631746,
"grad_norm": 0.19640228152275085,
"learning_rate": 5.3052033325705774e-05,
"loss": 1.1475,
"step": 52
},
{
"epoch": 0.014996109499893896,
"grad_norm": 0.19956982135772705,
"learning_rate": 5.2731446092810044e-05,
"loss": 1.1226,
"step": 53
},
{
"epoch": 0.015279054962156044,
"grad_norm": 0.15548309683799744,
"learning_rate": 5.240464430952462e-05,
"loss": 1.1501,
"step": 54
},
{
"epoch": 0.015562000424418193,
"grad_norm": 0.1360340118408203,
"learning_rate": 5.207171732019395e-05,
"loss": 1.1462,
"step": 55
},
{
"epoch": 0.015844945886680343,
"grad_norm": 0.14059096574783325,
"learning_rate": 5.1732756143732675e-05,
"loss": 1.1009,
"step": 56
},
{
"epoch": 0.01612789134894249,
"grad_norm": 0.13691364228725433,
"learning_rate": 5.1387853448741916e-05,
"loss": 1.1818,
"step": 57
},
{
"epoch": 0.01641083681120464,
"grad_norm": 0.13487009704113007,
"learning_rate": 5.103710352817465e-05,
"loss": 1.1531,
"step": 58
},
{
"epoch": 0.016693782273466788,
"grad_norm": 0.15809845924377441,
"learning_rate": 5.068060227355698e-05,
"loss": 1.1472,
"step": 59
},
{
"epoch": 0.016976727735728937,
"grad_norm": 0.15477801859378815,
"learning_rate": 5.0318447148772234e-05,
"loss": 1.1809,
"step": 60
},
{
"epoch": 0.01725967319799109,
"grad_norm": 0.1484566628932953,
"learning_rate": 4.995073716341545e-05,
"loss": 1.2081,
"step": 61
},
{
"epoch": 0.017542618660253237,
"grad_norm": 0.1421346217393875,
"learning_rate": 4.957757284572506e-05,
"loss": 1.224,
"step": 62
},
{
"epoch": 0.017825564122515385,
"grad_norm": 0.1333300918340683,
"learning_rate": 4.91990562150995e-05,
"loss": 1.1276,
"step": 63
},
{
"epoch": 0.018108509584777534,
"grad_norm": 0.12150599807500839,
"learning_rate": 4.881529075420611e-05,
"loss": 1.2021,
"step": 64
},
{
"epoch": 0.018391455047039682,
"grad_norm": 0.12204714864492416,
"learning_rate": 4.8426381380690036e-05,
"loss": 1.114,
"step": 65
},
{
"epoch": 0.01867440050930183,
"grad_norm": 0.12727297842502594,
"learning_rate": 4.8032434418490753e-05,
"loss": 1.1536,
"step": 66
},
{
"epoch": 0.018957345971563982,
"grad_norm": 0.11255544424057007,
"learning_rate": 4.7633557568774194e-05,
"loss": 1.1206,
"step": 67
},
{
"epoch": 0.01924029143382613,
"grad_norm": 0.1191386803984642,
"learning_rate": 4.722985988048831e-05,
"loss": 1.1546,
"step": 68
},
{
"epoch": 0.01952323689608828,
"grad_norm": 0.13010723888874054,
"learning_rate": 4.6821451720550184e-05,
"loss": 1.1694,
"step": 69
},
{
"epoch": 0.019806182358350428,
"grad_norm": 0.12365347146987915,
"learning_rate": 4.640844474367282e-05,
"loss": 1.136,
"step": 70
},
{
"epoch": 0.020089127820612576,
"grad_norm": 0.14611800014972687,
"learning_rate": 4.5990951861839815e-05,
"loss": 1.1683,
"step": 71
},
{
"epoch": 0.020372073282874725,
"grad_norm": 0.1299329400062561,
"learning_rate": 4.5569087213436455e-05,
"loss": 1.2127,
"step": 72
},
{
"epoch": 0.020655018745136876,
"grad_norm": 0.1349126100540161,
"learning_rate": 4.514296613204532e-05,
"loss": 1.1545,
"step": 73
},
{
"epoch": 0.020937964207399025,
"grad_norm": 0.12816999852657318,
"learning_rate": 4.471270511491525e-05,
"loss": 1.198,
"step": 74
},
{
"epoch": 0.021220909669661173,
"grad_norm": 0.14442527294158936,
"learning_rate": 4.427842179111221e-05,
"loss": 1.1331,
"step": 75
},
{
"epoch": 0.021503855131923322,
"grad_norm": 0.1318608820438385,
"learning_rate": 4.3840234889360634e-05,
"loss": 1.1442,
"step": 76
},
{
"epoch": 0.02178680059418547,
"grad_norm": 0.13773271441459656,
"learning_rate": 4.33982642055842e-05,
"loss": 1.1503,
"step": 77
},
{
"epoch": 0.02206974605644762,
"grad_norm": 0.12418694794178009,
"learning_rate": 4.2952630570154785e-05,
"loss": 1.0393,
"step": 78
},
{
"epoch": 0.022352691518709767,
"grad_norm": 0.14450491964817047,
"learning_rate": 4.250345581485871e-05,
"loss": 1.1194,
"step": 79
},
{
"epoch": 0.02263563698097192,
"grad_norm": 0.15314346551895142,
"learning_rate": 4.205086273958909e-05,
"loss": 1.0763,
"step": 80
},
{
"epoch": 0.022918582443234067,
"grad_norm": 0.1480676680803299,
"learning_rate": 4.1594975078773565e-05,
"loss": 1.2132,
"step": 81
},
{
"epoch": 0.023201527905496216,
"grad_norm": 0.14896009862422943,
"learning_rate": 4.113591746754662e-05,
"loss": 1.1389,
"step": 82
},
{
"epoch": 0.023484473367758364,
"grad_norm": 0.17366276681423187,
"learning_rate": 4.06738154076755e-05,
"loss": 1.2691,
"step": 83
},
{
"epoch": 0.023767418830020513,
"grad_norm": 0.16257454454898834,
"learning_rate": 4.020879523324929e-05,
"loss": 1.0568,
"step": 84
},
{
"epoch": 0.02405036429228266,
"grad_norm": 0.16570693254470825,
"learning_rate": 3.974098407614051e-05,
"loss": 1.1422,
"step": 85
},
{
"epoch": 0.024333309754544813,
"grad_norm": 0.1792537122964859,
"learning_rate": 3.927050983124842e-05,
"loss": 1.1421,
"step": 86
},
{
"epoch": 0.02461625521680696,
"grad_norm": 0.18916811048984528,
"learning_rate": 3.8797501121533946e-05,
"loss": 1.1065,
"step": 87
},
{
"epoch": 0.02489920067906911,
"grad_norm": 0.2118011862039566,
"learning_rate": 3.832208726285534e-05,
"loss": 1.0761,
"step": 88
},
{
"epoch": 0.025182146141331258,
"grad_norm": 0.21735292673110962,
"learning_rate": 3.784439822861459e-05,
"loss": 0.8525,
"step": 89
},
{
"epoch": 0.025465091603593407,
"grad_norm": 0.23854799568653107,
"learning_rate": 3.7364564614223976e-05,
"loss": 0.9597,
"step": 90
},
{
"epoch": 0.025748037065855555,
"grad_norm": 0.23937970399856567,
"learning_rate": 3.688271760140255e-05,
"loss": 0.8657,
"step": 91
},
{
"epoch": 0.026030982528117707,
"grad_norm": 0.26307588815689087,
"learning_rate": 3.6398988922312406e-05,
"loss": 0.953,
"step": 92
},
{
"epoch": 0.026313927990379855,
"grad_norm": 0.27083978056907654,
"learning_rate": 3.591351082354441e-05,
"loss": 0.7971,
"step": 93
},
{
"epoch": 0.026596873452642004,
"grad_norm": 0.2805502116680145,
"learning_rate": 3.54264160299633e-05,
"loss": 0.9744,
"step": 94
},
{
"epoch": 0.026879818914904152,
"grad_norm": 0.35504117608070374,
"learning_rate": 3.493783770842202e-05,
"loss": 0.9262,
"step": 95
},
{
"epoch": 0.0271627643771663,
"grad_norm": 0.33214688301086426,
"learning_rate": 3.444790943135526e-05,
"loss": 1.0821,
"step": 96
},
{
"epoch": 0.02744570983942845,
"grad_norm": 0.3750684857368469,
"learning_rate": 3.3956765140262074e-05,
"loss": 0.8112,
"step": 97
},
{
"epoch": 0.027728655301690597,
"grad_norm": 0.3210698962211609,
"learning_rate": 3.346453910908759e-05,
"loss": 1.004,
"step": 98
},
{
"epoch": 0.02801160076395275,
"grad_norm": 0.3541051745414734,
"learning_rate": 3.297136590751389e-05,
"loss": 1.16,
"step": 99
},
{
"epoch": 0.028294546226214898,
"grad_norm": 0.36050865054130554,
"learning_rate": 3.247738036416998e-05,
"loss": 1.1555,
"step": 100
},
{
"epoch": 0.028294546226214898,
"eval_loss": 1.0897471904754639,
"eval_runtime": 853.3729,
"eval_samples_per_second": 6.976,
"eval_steps_per_second": 1.745,
"step": 100
},
{
"epoch": 0.028577491688477046,
"grad_norm": 0.1543198525905609,
"learning_rate": 3.1982717529770985e-05,
"loss": 1.2425,
"step": 101
},
{
"epoch": 0.028860437150739195,
"grad_norm": 0.19346995651721954,
"learning_rate": 3.148751264019667e-05,
"loss": 1.349,
"step": 102
},
{
"epoch": 0.029143382613001343,
"grad_norm": 0.1380472481250763,
"learning_rate": 3.099190107951924e-05,
"loss": 1.1654,
"step": 103
},
{
"epoch": 0.02942632807526349,
"grad_norm": 0.1306401491165161,
"learning_rate": 3.049601834299076e-05,
"loss": 1.1968,
"step": 104
},
{
"epoch": 0.029709273537525643,
"grad_norm": 0.14598701894283295,
"learning_rate": 3e-05,
"loss": 1.1663,
"step": 105
},
{
"epoch": 0.029992218999787792,
"grad_norm": 0.14115272462368011,
"learning_rate": 2.9503981657009246e-05,
"loss": 1.212,
"step": 106
},
{
"epoch": 0.03027516446204994,
"grad_norm": 0.13328498601913452,
"learning_rate": 2.9008098920480752e-05,
"loss": 1.1183,
"step": 107
},
{
"epoch": 0.03055810992431209,
"grad_norm": 0.12777693569660187,
"learning_rate": 2.851248735980333e-05,
"loss": 1.113,
"step": 108
},
{
"epoch": 0.030841055386574237,
"grad_norm": 0.1404535323381424,
"learning_rate": 2.801728247022902e-05,
"loss": 1.128,
"step": 109
},
{
"epoch": 0.031124000848836385,
"grad_norm": 0.11834380775690079,
"learning_rate": 2.7522619635830034e-05,
"loss": 1.166,
"step": 110
},
{
"epoch": 0.031406946311098534,
"grad_norm": 0.14574697613716125,
"learning_rate": 2.702863409248612e-05,
"loss": 1.1504,
"step": 111
},
{
"epoch": 0.031689891773360686,
"grad_norm": 0.13284307718276978,
"learning_rate": 2.6535460890912416e-05,
"loss": 1.1715,
"step": 112
},
{
"epoch": 0.03197283723562283,
"grad_norm": 0.128324955701828,
"learning_rate": 2.604323485973793e-05,
"loss": 1.0557,
"step": 113
},
{
"epoch": 0.03225578269788498,
"grad_norm": 0.13001534342765808,
"learning_rate": 2.555209056864474e-05,
"loss": 1.2398,
"step": 114
},
{
"epoch": 0.032538728160147135,
"grad_norm": 0.13345256447792053,
"learning_rate": 2.5062162291577978e-05,
"loss": 1.2067,
"step": 115
},
{
"epoch": 0.03282167362240928,
"grad_norm": 0.12472221255302429,
"learning_rate": 2.4573583970036712e-05,
"loss": 1.1886,
"step": 116
},
{
"epoch": 0.03310461908467143,
"grad_norm": 0.14320768415927887,
"learning_rate": 2.4086489176455595e-05,
"loss": 1.2183,
"step": 117
},
{
"epoch": 0.033387564546933576,
"grad_norm": 0.1415589302778244,
"learning_rate": 2.36010110776876e-05,
"loss": 1.2193,
"step": 118
},
{
"epoch": 0.03367051000919573,
"grad_norm": 0.14643894135951996,
"learning_rate": 2.3117282398597456e-05,
"loss": 1.2352,
"step": 119
},
{
"epoch": 0.03395345547145787,
"grad_norm": 0.1440449059009552,
"learning_rate": 2.263543538577603e-05,
"loss": 1.1977,
"step": 120
},
{
"epoch": 0.034236400933720025,
"grad_norm": 0.14856146275997162,
"learning_rate": 2.215560177138541e-05,
"loss": 1.1351,
"step": 121
},
{
"epoch": 0.03451934639598218,
"grad_norm": 0.13947530090808868,
"learning_rate": 2.167791273714467e-05,
"loss": 1.169,
"step": 122
},
{
"epoch": 0.03480229185824432,
"grad_norm": 0.13417986035346985,
"learning_rate": 2.1202498878466062e-05,
"loss": 1.1483,
"step": 123
},
{
"epoch": 0.035085237320506474,
"grad_norm": 0.13690292835235596,
"learning_rate": 2.072949016875158e-05,
"loss": 1.0759,
"step": 124
},
{
"epoch": 0.03536818278276862,
"grad_norm": 0.1381928026676178,
"learning_rate": 2.0259015923859498e-05,
"loss": 1.2043,
"step": 125
},
{
"epoch": 0.03565112824503077,
"grad_norm": 0.14060895144939423,
"learning_rate": 1.979120476675071e-05,
"loss": 1.1919,
"step": 126
},
{
"epoch": 0.03593407370729292,
"grad_norm": 0.12901578843593597,
"learning_rate": 1.9326184592324503e-05,
"loss": 1.0793,
"step": 127
},
{
"epoch": 0.03621701916955507,
"grad_norm": 0.1504661738872528,
"learning_rate": 1.8864082532453373e-05,
"loss": 1.2077,
"step": 128
},
{
"epoch": 0.03649996463181722,
"grad_norm": 0.16385625302791595,
"learning_rate": 1.840502492122644e-05,
"loss": 1.1345,
"step": 129
},
{
"epoch": 0.036782910094079364,
"grad_norm": 0.16052238643169403,
"learning_rate": 1.7949137260410924e-05,
"loss": 1.1935,
"step": 130
},
{
"epoch": 0.037065855556341516,
"grad_norm": 0.16583813726902008,
"learning_rate": 1.7496544185141295e-05,
"loss": 1.0112,
"step": 131
},
{
"epoch": 0.03734880101860366,
"grad_norm": 0.17477497458457947,
"learning_rate": 1.7047369429845216e-05,
"loss": 1.0563,
"step": 132
},
{
"epoch": 0.03763174648086581,
"grad_norm": 0.17088453471660614,
"learning_rate": 1.6601735794415806e-05,
"loss": 1.0716,
"step": 133
},
{
"epoch": 0.037914691943127965,
"grad_norm": 0.18345321714878082,
"learning_rate": 1.615976511063937e-05,
"loss": 0.9948,
"step": 134
},
{
"epoch": 0.03819763740539011,
"grad_norm": 0.20441484451293945,
"learning_rate": 1.5721578208887793e-05,
"loss": 0.9666,
"step": 135
},
{
"epoch": 0.03848058286765226,
"grad_norm": 0.1876918077468872,
"learning_rate": 1.5287294885084766e-05,
"loss": 0.9862,
"step": 136
},
{
"epoch": 0.03876352832991441,
"grad_norm": 0.22819191217422485,
"learning_rate": 1.4857033867954697e-05,
"loss": 1.1309,
"step": 137
},
{
"epoch": 0.03904647379217656,
"grad_norm": 0.2145112156867981,
"learning_rate": 1.4430912786563554e-05,
"loss": 0.9526,
"step": 138
},
{
"epoch": 0.039329419254438704,
"grad_norm": 0.21892738342285156,
"learning_rate": 1.4009048138160195e-05,
"loss": 1.0422,
"step": 139
},
{
"epoch": 0.039612364716700856,
"grad_norm": 0.23647260665893555,
"learning_rate": 1.3591555256327199e-05,
"loss": 0.9467,
"step": 140
},
{
"epoch": 0.03989531017896301,
"grad_norm": 0.25753793120384216,
"learning_rate": 1.3178548279449822e-05,
"loss": 0.8659,
"step": 141
},
{
"epoch": 0.04017825564122515,
"grad_norm": 0.24101592600345612,
"learning_rate": 1.2770140119511693e-05,
"loss": 0.8691,
"step": 142
},
{
"epoch": 0.040461201103487304,
"grad_norm": 0.2642737329006195,
"learning_rate": 1.2366442431225809e-05,
"loss": 1.0043,
"step": 143
},
{
"epoch": 0.04074414656574945,
"grad_norm": 0.27606478333473206,
"learning_rate": 1.1967565581509248e-05,
"loss": 0.8247,
"step": 144
},
{
"epoch": 0.0410270920280116,
"grad_norm": 0.27429643273353577,
"learning_rate": 1.1573618619309965e-05,
"loss": 0.8922,
"step": 145
},
{
"epoch": 0.04131003749027375,
"grad_norm": 0.29994338750839233,
"learning_rate": 1.1184709245793889e-05,
"loss": 1.0195,
"step": 146
},
{
"epoch": 0.0415929829525359,
"grad_norm": 0.3414655029773712,
"learning_rate": 1.0800943784900502e-05,
"loss": 0.8346,
"step": 147
},
{
"epoch": 0.04187592841479805,
"grad_norm": 0.30294448137283325,
"learning_rate": 1.042242715427494e-05,
"loss": 1.1227,
"step": 148
},
{
"epoch": 0.042158873877060195,
"grad_norm": 0.2909342646598816,
"learning_rate": 1.004926283658455e-05,
"loss": 0.8504,
"step": 149
},
{
"epoch": 0.04244181933932235,
"grad_norm": 0.35827234387397766,
"learning_rate": 9.681552851227774e-06,
"loss": 1.1022,
"step": 150
},
{
"epoch": 0.04244181933932235,
"eval_loss": 1.0741040706634521,
"eval_runtime": 853.591,
"eval_samples_per_second": 6.974,
"eval_steps_per_second": 1.744,
"step": 150
},
{
"epoch": 0.04272476480158449,
"grad_norm": 0.15892794728279114,
"learning_rate": 9.319397726443026e-06,
"loss": 1.18,
"step": 151
},
{
"epoch": 0.043007710263846644,
"grad_norm": 0.15233199298381805,
"learning_rate": 8.962896471825342e-06,
"loss": 1.1396,
"step": 152
},
{
"epoch": 0.043290655726108795,
"grad_norm": 0.15636806190013885,
"learning_rate": 8.61214655125809e-06,
"loss": 1.3293,
"step": 153
},
{
"epoch": 0.04357360118837094,
"grad_norm": 0.17083175480365753,
"learning_rate": 8.267243856267331e-06,
"loss": 1.2175,
"step": 154
},
{
"epoch": 0.04385654665063309,
"grad_norm": 0.14700956642627716,
"learning_rate": 7.928282679806052e-06,
"loss": 1.2229,
"step": 155
},
{
"epoch": 0.04413949211289524,
"grad_norm": 0.14753887057304382,
"learning_rate": 7.595355690475393e-06,
"loss": 1.1916,
"step": 156
},
{
"epoch": 0.04442243757515739,
"grad_norm": 0.16037601232528687,
"learning_rate": 7.268553907189964e-06,
"loss": 1.1722,
"step": 157
},
{
"epoch": 0.044705383037419534,
"grad_norm": 0.125972181558609,
"learning_rate": 6.947966674294236e-06,
"loss": 1.162,
"step": 158
},
{
"epoch": 0.044988328499681686,
"grad_norm": 0.15123845636844635,
"learning_rate": 6.6336816371366305e-06,
"loss": 1.206,
"step": 159
},
{
"epoch": 0.04527127396194384,
"grad_norm": 0.14084303379058838,
"learning_rate": 6.325784718108196e-06,
"loss": 1.0952,
"step": 160
},
{
"epoch": 0.04555421942420598,
"grad_norm": 0.1478935182094574,
"learning_rate": 6.0243600931522595e-06,
"loss": 1.0997,
"step": 161
},
{
"epoch": 0.045837164886468135,
"grad_norm": 0.1472909152507782,
"learning_rate": 5.72949016875158e-06,
"loss": 1.1209,
"step": 162
},
{
"epoch": 0.04612011034873028,
"grad_norm": 0.15565665066242218,
"learning_rate": 5.44125555939923e-06,
"loss": 1.2026,
"step": 163
},
{
"epoch": 0.04640305581099243,
"grad_norm": 0.15660125017166138,
"learning_rate": 5.159735065559399e-06,
"loss": 1.1838,
"step": 164
},
{
"epoch": 0.04668600127325458,
"grad_norm": 0.1355641484260559,
"learning_rate": 4.885005652124144e-06,
"loss": 1.2546,
"step": 165
},
{
"epoch": 0.04696894673551673,
"grad_norm": 0.13826073706150055,
"learning_rate": 4.617142427371934e-06,
"loss": 1.1198,
"step": 166
},
{
"epoch": 0.04725189219777888,
"grad_norm": 0.13007190823554993,
"learning_rate": 4.3562186224338265e-06,
"loss": 1.0406,
"step": 167
},
{
"epoch": 0.047534837660041025,
"grad_norm": 0.13735632598400116,
"learning_rate": 4.102305571272783e-06,
"loss": 1.2001,
"step": 168
},
{
"epoch": 0.04781778312230318,
"grad_norm": 0.13487175107002258,
"learning_rate": 3.855472691181678e-06,
"loss": 1.1674,
"step": 169
},
{
"epoch": 0.04810072858456532,
"grad_norm": 0.1506294310092926,
"learning_rate": 3.615787463805331e-06,
"loss": 1.1782,
"step": 170
},
{
"epoch": 0.048383674046827474,
"grad_norm": 0.13945788145065308,
"learning_rate": 3.383315416691646e-06,
"loss": 1.1162,
"step": 171
},
{
"epoch": 0.048666619509089626,
"grad_norm": 0.14078959822654724,
"learning_rate": 3.158120105377096e-06,
"loss": 1.1837,
"step": 172
},
{
"epoch": 0.04894956497135177,
"grad_norm": 0.15105940401554108,
"learning_rate": 2.940263096011233e-06,
"loss": 1.2026,
"step": 173
},
{
"epoch": 0.04923251043361392,
"grad_norm": 0.14473502337932587,
"learning_rate": 2.729803948525125e-06,
"loss": 1.0766,
"step": 174
},
{
"epoch": 0.04951545589587607,
"grad_norm": 0.14269044995307922,
"learning_rate": 2.526800200348275e-06,
"loss": 1.1419,
"step": 175
},
{
"epoch": 0.04979840135813822,
"grad_norm": 0.1475040763616562,
"learning_rate": 2.3313073506784575e-06,
"loss": 1.2384,
"step": 176
},
{
"epoch": 0.050081346820400365,
"grad_norm": 0.13880924880504608,
"learning_rate": 2.143378845308791e-06,
"loss": 1.1159,
"step": 177
},
{
"epoch": 0.050364292282662516,
"grad_norm": 0.14378739893436432,
"learning_rate": 1.9630660620161777e-06,
"loss": 1.1177,
"step": 178
},
{
"epoch": 0.05064723774492467,
"grad_norm": 0.17214110493659973,
"learning_rate": 1.790418296515165e-06,
"loss": 1.1171,
"step": 179
},
{
"epoch": 0.05093018320718681,
"grad_norm": 0.16885404288768768,
"learning_rate": 1.625482748980961e-06,
"loss": 1.2154,
"step": 180
},
{
"epoch": 0.051213128669448965,
"grad_norm": 0.15598832070827484,
"learning_rate": 1.4683045111453942e-06,
"loss": 1.1079,
"step": 181
},
{
"epoch": 0.05149607413171111,
"grad_norm": 0.16221092641353607,
"learning_rate": 1.3189265539692707e-06,
"loss": 1.1572,
"step": 182
},
{
"epoch": 0.05177901959397326,
"grad_norm": 0.15152089297771454,
"learning_rate": 1.1773897158945557e-06,
"loss": 1.08,
"step": 183
},
{
"epoch": 0.052061965056235414,
"grad_norm": 0.1765381395816803,
"learning_rate": 1.0437326916795432e-06,
"loss": 1.0874,
"step": 184
},
{
"epoch": 0.05234491051849756,
"grad_norm": 0.18851357698440552,
"learning_rate": 9.179920218200888e-07,
"loss": 1.031,
"step": 185
},
{
"epoch": 0.05262785598075971,
"grad_norm": 0.190265491604805,
"learning_rate": 8.002020825598277e-07,
"loss": 1.0912,
"step": 186
},
{
"epoch": 0.052910801443021856,
"grad_norm": 0.1942068487405777,
"learning_rate": 6.90395076492022e-07,
"loss": 1.0471,
"step": 187
},
{
"epoch": 0.05319374690528401,
"grad_norm": 0.23885029554367065,
"learning_rate": 5.886010237557194e-07,
"loss": 0.9656,
"step": 188
},
{
"epoch": 0.05347669236754615,
"grad_norm": 0.22708922624588013,
"learning_rate": 4.94847753828529e-07,
"loss": 1.0172,
"step": 189
},
{
"epoch": 0.053759637829808304,
"grad_norm": 0.23104092478752136,
"learning_rate": 4.091608979183303e-07,
"loss": 0.9197,
"step": 190
},
{
"epoch": 0.054042583292070456,
"grad_norm": 0.28225743770599365,
"learning_rate": 3.315638819559452e-07,
"loss": 0.8387,
"step": 191
},
{
"epoch": 0.0543255287543326,
"grad_norm": 0.29937031865119934,
"learning_rate": 2.6207792019074414e-07,
"loss": 0.8365,
"step": 192
},
{
"epoch": 0.05460847421659475,
"grad_norm": 0.27068015933036804,
"learning_rate": 2.0072200939085573e-07,
"loss": 0.8807,
"step": 193
},
{
"epoch": 0.0548914196788569,
"grad_norm": 0.3136865198612213,
"learning_rate": 1.475129236496575e-07,
"loss": 0.8064,
"step": 194
},
{
"epoch": 0.05517436514111905,
"grad_norm": 0.30548372864723206,
"learning_rate": 1.0246520979990459e-07,
"loss": 1.0338,
"step": 195
},
{
"epoch": 0.055457310603381195,
"grad_norm": 0.29919055104255676,
"learning_rate": 6.559118343676396e-08,
"loss": 0.8476,
"step": 196
},
{
"epoch": 0.05574025606564335,
"grad_norm": 0.333148330450058,
"learning_rate": 3.690092555085789e-08,
"loss": 0.7565,
"step": 197
},
{
"epoch": 0.0560232015279055,
"grad_norm": 0.31107303500175476,
"learning_rate": 1.640227977221853e-08,
"loss": 0.8806,
"step": 198
},
{
"epoch": 0.056306146990167644,
"grad_norm": 0.3069625198841095,
"learning_rate": 4.1008502259298755e-09,
"loss": 1.1436,
"step": 199
},
{
"epoch": 0.056589092452429796,
"grad_norm": 0.35064685344696045,
"learning_rate": 0.0,
"loss": 1.0373,
"step": 200
},
{
"epoch": 0.056589092452429796,
"eval_loss": 1.0697510242462158,
"eval_runtime": 853.5988,
"eval_samples_per_second": 6.974,
"eval_steps_per_second": 1.744,
"step": 200
}
],
"logging_steps": 1,
"max_steps": 200,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 50,
"stateful_callbacks": {
"EarlyStoppingCallback": {
"args": {
"early_stopping_patience": 4,
"early_stopping_threshold": 0.0
},
"attributes": {
"early_stopping_patience_counter": 0
}
},
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 5.275415744945848e+17,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}