{ "best_metric": 10.48843765258789, "best_model_checkpoint": "miner_id_24/checkpoint-150", "epoch": 0.0989282769991756, "eval_steps": 25, "global_step": 150, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0006595218466611707, "grad_norm": 5.548171520233154, "learning_rate": 2.9999999999999997e-05, "loss": 44.5605, "step": 1 }, { "epoch": 0.0006595218466611707, "eval_loss": 11.123749732971191, "eval_runtime": 0.5337, "eval_samples_per_second": 93.678, "eval_steps_per_second": 13.115, "step": 1 }, { "epoch": 0.0013190436933223414, "grad_norm": 5.983486175537109, "learning_rate": 5.9999999999999995e-05, "loss": 44.5645, "step": 2 }, { "epoch": 0.001978565539983512, "grad_norm": 5.650338172912598, "learning_rate": 8.999999999999999e-05, "loss": 44.5039, "step": 3 }, { "epoch": 0.0026380873866446827, "grad_norm": 5.475784778594971, "learning_rate": 0.00011999999999999999, "loss": 44.5117, "step": 4 }, { "epoch": 0.0032976092333058533, "grad_norm": 5.486029148101807, "learning_rate": 0.00015, "loss": 44.4688, "step": 5 }, { "epoch": 0.003957131079967024, "grad_norm": 5.552286148071289, "learning_rate": 0.00017999999999999998, "loss": 44.4121, "step": 6 }, { "epoch": 0.004616652926628195, "grad_norm": 5.922216415405273, "learning_rate": 0.00020999999999999998, "loss": 44.3496, "step": 7 }, { "epoch": 0.0052761747732893655, "grad_norm": 5.640804767608643, "learning_rate": 0.00023999999999999998, "loss": 44.2832, "step": 8 }, { "epoch": 0.005935696619950536, "grad_norm": 5.632114887237549, "learning_rate": 0.00027, "loss": 44.2227, "step": 9 }, { "epoch": 0.006595218466611707, "grad_norm": 5.427684307098389, "learning_rate": 0.0003, "loss": 44.1895, "step": 10 }, { "epoch": 0.007254740313272877, "grad_norm": 5.3626389503479, "learning_rate": 0.0002999794957488703, "loss": 44.1113, "step": 11 }, { "epoch": 0.007914262159934049, "grad_norm": 5.239843368530273, "learning_rate": 0.0002999179886011389, "loss": 43.9922, "step": 12 }, { "epoch": 0.008573784006595218, "grad_norm": 5.086913585662842, "learning_rate": 0.0002998154953722457, "loss": 43.8848, "step": 13 }, { "epoch": 0.00923330585325639, "grad_norm": 5.158645153045654, "learning_rate": 0.00029967204408281613, "loss": 43.7676, "step": 14 }, { "epoch": 0.00989282769991756, "grad_norm": 4.871565341949463, "learning_rate": 0.00029948767395100045, "loss": 43.7051, "step": 15 }, { "epoch": 0.010552349546578731, "grad_norm": 4.838162422180176, "learning_rate": 0.0002992624353817517, "loss": 43.6191, "step": 16 }, { "epoch": 0.0112118713932399, "grad_norm": 4.602661609649658, "learning_rate": 0.0002989963899530457, "loss": 43.5488, "step": 17 }, { "epoch": 0.011871393239901072, "grad_norm": 4.472682476043701, "learning_rate": 0.00029868961039904624, "loss": 43.4824, "step": 18 }, { "epoch": 0.012530915086562242, "grad_norm": 4.483953952789307, "learning_rate": 0.00029834218059022024, "loss": 43.4609, "step": 19 }, { "epoch": 0.013190436933223413, "grad_norm": 4.183454513549805, "learning_rate": 0.00029795419551040833, "loss": 43.3125, "step": 20 }, { "epoch": 0.013849958779884583, "grad_norm": 4.2280449867248535, "learning_rate": 0.00029752576123085736, "loss": 43.1465, "step": 21 }, { "epoch": 0.014509480626545754, "grad_norm": 3.8879261016845703, "learning_rate": 0.0002970569948812214, "loss": 43.2305, "step": 22 }, { "epoch": 0.015169002473206924, "grad_norm": 3.6963751316070557, "learning_rate": 0.0002965480246175399, "loss": 43.1582, "step": 23 }, { "epoch": 0.015828524319868097, "grad_norm": 3.540698289871216, "learning_rate": 0.0002959989895872009, "loss": 43.0605, "step": 24 }, { "epoch": 0.016488046166529265, "grad_norm": 3.258104085922241, "learning_rate": 0.0002954100398908995, "loss": 42.9844, "step": 25 }, { "epoch": 0.016488046166529265, "eval_loss": 10.72249984741211, "eval_runtime": 0.1131, "eval_samples_per_second": 441.99, "eval_steps_per_second": 61.879, "step": 25 }, { "epoch": 0.017147568013190437, "grad_norm": 3.194683074951172, "learning_rate": 0.0002947813365416023, "loss": 42.8789, "step": 26 }, { "epoch": 0.017807089859851608, "grad_norm": 3.04011607170105, "learning_rate": 0.0002941130514205272, "loss": 42.8496, "step": 27 }, { "epoch": 0.01846661170651278, "grad_norm": 2.8812403678894043, "learning_rate": 0.0002934053672301536, "loss": 42.7383, "step": 28 }, { "epoch": 0.019126133553173948, "grad_norm": 2.8334062099456787, "learning_rate": 0.00029265847744427303, "loss": 42.791, "step": 29 }, { "epoch": 0.01978565539983512, "grad_norm": 2.608891248703003, "learning_rate": 0.00029187258625509513, "loss": 42.7285, "step": 30 }, { "epoch": 0.02044517724649629, "grad_norm": 2.658057451248169, "learning_rate": 0.00029104790851742417, "loss": 42.7129, "step": 31 }, { "epoch": 0.021104699093157462, "grad_norm": 2.4135985374450684, "learning_rate": 0.0002901846696899191, "loss": 42.6523, "step": 32 }, { "epoch": 0.02176422093981863, "grad_norm": 2.3150999546051025, "learning_rate": 0.00028928310577345606, "loss": 42.6055, "step": 33 }, { "epoch": 0.0224237427864798, "grad_norm": 2.3153157234191895, "learning_rate": 0.0002883434632466077, "loss": 42.5723, "step": 34 }, { "epoch": 0.023083264633140973, "grad_norm": 2.148899555206299, "learning_rate": 0.00028736599899825856, "loss": 42.5176, "step": 35 }, { "epoch": 0.023742786479802144, "grad_norm": 2.181697130203247, "learning_rate": 0.00028635098025737434, "loss": 42.5156, "step": 36 }, { "epoch": 0.024402308326463316, "grad_norm": 2.0900909900665283, "learning_rate": 0.00028529868451994384, "loss": 42.4102, "step": 37 }, { "epoch": 0.025061830173124484, "grad_norm": 1.945241928100586, "learning_rate": 0.0002842093994731145, "loss": 42.4102, "step": 38 }, { "epoch": 0.025721352019785655, "grad_norm": 1.914509892463684, "learning_rate": 0.00028308342291654174, "loss": 42.3848, "step": 39 }, { "epoch": 0.026380873866446827, "grad_norm": 2.0310447216033936, "learning_rate": 0.00028192106268097334, "loss": 42.3613, "step": 40 }, { "epoch": 0.027040395713107998, "grad_norm": 2.335347890853882, "learning_rate": 0.00028072263654409154, "loss": 42.3184, "step": 41 }, { "epoch": 0.027699917559769166, "grad_norm": 1.846063256263733, "learning_rate": 0.0002794884721436361, "loss": 42.291, "step": 42 }, { "epoch": 0.028359439406430337, "grad_norm": 1.7599420547485352, "learning_rate": 0.00027821890688783083, "loss": 42.2578, "step": 43 }, { "epoch": 0.02901896125309151, "grad_norm": 1.7382268905639648, "learning_rate": 0.0002769142878631403, "loss": 42.2266, "step": 44 }, { "epoch": 0.02967848309975268, "grad_norm": 1.5867624282836914, "learning_rate": 0.00027557497173937923, "loss": 42.1758, "step": 45 }, { "epoch": 0.03033800494641385, "grad_norm": 1.6499996185302734, "learning_rate": 0.000274201324672203, "loss": 42.1758, "step": 46 }, { "epoch": 0.03099752679307502, "grad_norm": 1.8098398447036743, "learning_rate": 0.00027279372220300385, "loss": 42.1719, "step": 47 }, { "epoch": 0.031657048639736195, "grad_norm": 1.7325196266174316, "learning_rate": 0.0002713525491562421, "loss": 42.1387, "step": 48 }, { "epoch": 0.03231657048639736, "grad_norm": 1.4527415037155151, "learning_rate": 0.00026987819953423867, "loss": 42.1289, "step": 49 }, { "epoch": 0.03297609233305853, "grad_norm": 1.7800573110580444, "learning_rate": 0.00026837107640945905, "loss": 42.0781, "step": 50 }, { "epoch": 0.03297609233305853, "eval_loss": 10.520312309265137, "eval_runtime": 0.1132, "eval_samples_per_second": 441.817, "eval_steps_per_second": 61.854, "step": 50 }, { "epoch": 0.033635614179719706, "grad_norm": 1.5558362007141113, "learning_rate": 0.0002668315918143169, "loss": 42.1289, "step": 51 }, { "epoch": 0.034295136026380874, "grad_norm": 1.4583945274353027, "learning_rate": 0.00026526016662852886, "loss": 42.0273, "step": 52 }, { "epoch": 0.03495465787304204, "grad_norm": 1.2193175554275513, "learning_rate": 0.00026365723046405023, "loss": 42.0605, "step": 53 }, { "epoch": 0.035614179719703216, "grad_norm": 1.212590217590332, "learning_rate": 0.0002620232215476231, "loss": 42.0234, "step": 54 }, { "epoch": 0.036273701566364384, "grad_norm": 0.9861271381378174, "learning_rate": 0.0002603585866009697, "loss": 42.0625, "step": 55 }, { "epoch": 0.03693322341302556, "grad_norm": 0.9094260931015015, "learning_rate": 0.00025866378071866334, "loss": 42.0195, "step": 56 }, { "epoch": 0.03759274525968673, "grad_norm": 0.9702939391136169, "learning_rate": 0.00025693926724370956, "loss": 42.0547, "step": 57 }, { "epoch": 0.038252267106347895, "grad_norm": 0.8315839767456055, "learning_rate": 0.00025518551764087326, "loss": 42.0605, "step": 58 }, { "epoch": 0.03891178895300907, "grad_norm": 0.9290674328804016, "learning_rate": 0.00025340301136778483, "loss": 42.0391, "step": 59 }, { "epoch": 0.03957131079967024, "grad_norm": 1.1367566585540771, "learning_rate": 0.00025159223574386114, "loss": 41.9785, "step": 60 }, { "epoch": 0.04023083264633141, "grad_norm": 0.842949390411377, "learning_rate": 0.0002497536858170772, "loss": 42.0508, "step": 61 }, { "epoch": 0.04089035449299258, "grad_norm": 0.7557629346847534, "learning_rate": 0.00024788786422862526, "loss": 42.0332, "step": 62 }, { "epoch": 0.04154987633965375, "grad_norm": 0.6981768608093262, "learning_rate": 0.00024599528107549745, "loss": 42.0215, "step": 63 }, { "epoch": 0.042209398186314924, "grad_norm": 0.6374576687812805, "learning_rate": 0.00024407645377103054, "loss": 42.0039, "step": 64 }, { "epoch": 0.04286892003297609, "grad_norm": 0.7283995151519775, "learning_rate": 0.00024213190690345018, "loss": 41.998, "step": 65 }, { "epoch": 0.04352844187963726, "grad_norm": 0.7016935348510742, "learning_rate": 0.00024016217209245374, "loss": 41.9688, "step": 66 }, { "epoch": 0.044187963726298435, "grad_norm": 0.7185855507850647, "learning_rate": 0.00023816778784387094, "loss": 41.9727, "step": 67 }, { "epoch": 0.0448474855729596, "grad_norm": 0.5180225372314453, "learning_rate": 0.0002361492994024415, "loss": 41.9844, "step": 68 }, { "epoch": 0.04550700741962078, "grad_norm": 0.7438845038414001, "learning_rate": 0.0002341072586027509, "loss": 41.959, "step": 69 }, { "epoch": 0.046166529266281946, "grad_norm": 0.6051533818244934, "learning_rate": 0.00023204222371836405, "loss": 41.9551, "step": 70 }, { "epoch": 0.046826051112943114, "grad_norm": 0.5988097190856934, "learning_rate": 0.00022995475930919905, "loss": 42.0098, "step": 71 }, { "epoch": 0.04748557295960429, "grad_norm": 0.6344655156135559, "learning_rate": 0.00022784543606718227, "loss": 42.0039, "step": 72 }, { "epoch": 0.04814509480626546, "grad_norm": 0.6249358057975769, "learning_rate": 0.00022571483066022657, "loss": 41.998, "step": 73 }, { "epoch": 0.04880461665292663, "grad_norm": 0.6966153383255005, "learning_rate": 0.0002235635255745762, "loss": 41.9434, "step": 74 }, { "epoch": 0.0494641384995878, "grad_norm": 0.6041496396064758, "learning_rate": 0.00022139210895556104, "loss": 41.9766, "step": 75 }, { "epoch": 0.0494641384995878, "eval_loss": 10.489999771118164, "eval_runtime": 0.1147, "eval_samples_per_second": 436.085, "eval_steps_per_second": 61.052, "step": 75 }, { "epoch": 0.05012366034624897, "grad_norm": 0.6852412223815918, "learning_rate": 0.00021920117444680317, "loss": 42.0117, "step": 76 }, { "epoch": 0.05078318219291014, "grad_norm": 0.43333616852760315, "learning_rate": 0.00021699132102792097, "loss": 41.9805, "step": 77 }, { "epoch": 0.05144270403957131, "grad_norm": 0.6226640343666077, "learning_rate": 0.0002147631528507739, "loss": 41.9941, "step": 78 }, { "epoch": 0.05210222588623248, "grad_norm": 0.57357257604599, "learning_rate": 0.00021251727907429355, "loss": 42.0098, "step": 79 }, { "epoch": 0.05276174773289365, "grad_norm": 0.4521426260471344, "learning_rate": 0.0002102543136979454, "loss": 42.0098, "step": 80 }, { "epoch": 0.05342126957955482, "grad_norm": 0.4179379642009735, "learning_rate": 0.0002079748753938678, "loss": 41.9688, "step": 81 }, { "epoch": 0.054080791426215996, "grad_norm": 0.421734094619751, "learning_rate": 0.0002056795873377331, "loss": 41.9785, "step": 82 }, { "epoch": 0.054740313272877164, "grad_norm": 0.3774226903915405, "learning_rate": 0.00020336907703837748, "loss": 41.9785, "step": 83 }, { "epoch": 0.05539983511953833, "grad_norm": 0.3708513379096985, "learning_rate": 0.00020104397616624645, "loss": 41.9727, "step": 84 }, { "epoch": 0.05605935696619951, "grad_norm": 0.42900216579437256, "learning_rate": 0.00019870492038070252, "loss": 41.957, "step": 85 }, { "epoch": 0.056718878812860675, "grad_norm": 0.46859726309776306, "learning_rate": 0.0001963525491562421, "loss": 41.9707, "step": 86 }, { "epoch": 0.05737840065952185, "grad_norm": 0.34871360659599304, "learning_rate": 0.0001939875056076697, "loss": 41.9727, "step": 87 }, { "epoch": 0.05803792250618302, "grad_norm": 0.3781905472278595, "learning_rate": 0.00019161043631427666, "loss": 41.9902, "step": 88 }, { "epoch": 0.058697444352844186, "grad_norm": 0.6564770936965942, "learning_rate": 0.00018922199114307294, "loss": 41.9258, "step": 89 }, { "epoch": 0.05935696619950536, "grad_norm": 0.7199106812477112, "learning_rate": 0.00018682282307111987, "loss": 41.9258, "step": 90 }, { "epoch": 0.06001648804616653, "grad_norm": 0.3589068651199341, "learning_rate": 0.00018441358800701273, "loss": 41.9805, "step": 91 }, { "epoch": 0.0606760098928277, "grad_norm": 0.4372813403606415, "learning_rate": 0.00018199494461156203, "loss": 41.998, "step": 92 }, { "epoch": 0.06133553173948887, "grad_norm": 0.4405193626880646, "learning_rate": 0.000179567554117722, "loss": 41.9453, "step": 93 }, { "epoch": 0.06199505358615004, "grad_norm": 0.7273338437080383, "learning_rate": 0.00017713208014981648, "loss": 41.9199, "step": 94 }, { "epoch": 0.06265457543281121, "grad_norm": 0.4150626063346863, "learning_rate": 0.00017468918854211007, "loss": 41.9609, "step": 95 }, { "epoch": 0.06331409727947239, "grad_norm": 0.4945806860923767, "learning_rate": 0.00017223954715677627, "loss": 41.9844, "step": 96 }, { "epoch": 0.06397361912613356, "grad_norm": 0.36564674973487854, "learning_rate": 0.00016978382570131034, "loss": 41.9512, "step": 97 }, { "epoch": 0.06463314097279473, "grad_norm": 0.632477879524231, "learning_rate": 0.00016732269554543794, "loss": 41.918, "step": 98 }, { "epoch": 0.0652926628194559, "grad_norm": 0.24244551360607147, "learning_rate": 0.00016485682953756942, "loss": 41.9531, "step": 99 }, { "epoch": 0.06595218466611706, "grad_norm": 0.39558830857276917, "learning_rate": 0.00016238690182084986, "loss": 41.9551, "step": 100 }, { "epoch": 0.06595218466611706, "eval_loss": 10.496562957763672, "eval_runtime": 0.1143, "eval_samples_per_second": 437.308, "eval_steps_per_second": 61.223, "step": 100 }, { "epoch": 0.06661170651277823, "grad_norm": 0.6666596531867981, "learning_rate": 0.0001599135876488549, "loss": 42.0117, "step": 101 }, { "epoch": 0.06727122835943941, "grad_norm": 0.546597421169281, "learning_rate": 0.00015743756320098332, "loss": 41.9258, "step": 102 }, { "epoch": 0.06793075020610058, "grad_norm": 0.40258997678756714, "learning_rate": 0.0001549595053975962, "loss": 41.9414, "step": 103 }, { "epoch": 0.06859027205276175, "grad_norm": 0.4524000883102417, "learning_rate": 0.00015248009171495378, "loss": 41.9805, "step": 104 }, { "epoch": 0.06924979389942292, "grad_norm": 0.39214250445365906, "learning_rate": 0.00015, "loss": 41.9355, "step": 105 }, { "epoch": 0.06990931574608408, "grad_norm": 0.5733072757720947, "learning_rate": 0.00014751990828504622, "loss": 42.0059, "step": 106 }, { "epoch": 0.07056883759274526, "grad_norm": 0.5044758319854736, "learning_rate": 0.00014504049460240375, "loss": 42.002, "step": 107 }, { "epoch": 0.07122835943940643, "grad_norm": 0.6467098593711853, "learning_rate": 0.00014256243679901663, "loss": 42.0098, "step": 108 }, { "epoch": 0.0718878812860676, "grad_norm": 0.43010029196739197, "learning_rate": 0.00014008641235114508, "loss": 41.9902, "step": 109 }, { "epoch": 0.07254740313272877, "grad_norm": 0.6928155422210693, "learning_rate": 0.00013761309817915014, "loss": 42.0234, "step": 110 }, { "epoch": 0.07320692497938994, "grad_norm": 0.44245922565460205, "learning_rate": 0.00013514317046243058, "loss": 41.9961, "step": 111 }, { "epoch": 0.07386644682605112, "grad_norm": 0.3800172805786133, "learning_rate": 0.00013267730445456208, "loss": 41.9805, "step": 112 }, { "epoch": 0.07452596867271229, "grad_norm": 0.7035552263259888, "learning_rate": 0.00013021617429868963, "loss": 42.0156, "step": 113 }, { "epoch": 0.07518549051937345, "grad_norm": 0.42883816361427307, "learning_rate": 0.00012776045284322368, "loss": 41.9395, "step": 114 }, { "epoch": 0.07584501236603462, "grad_norm": 0.32898375391960144, "learning_rate": 0.00012531081145788987, "loss": 41.9883, "step": 115 }, { "epoch": 0.07650453421269579, "grad_norm": 0.3533617854118347, "learning_rate": 0.00012286791985018355, "loss": 41.9688, "step": 116 }, { "epoch": 0.07716405605935697, "grad_norm": 0.48832958936691284, "learning_rate": 0.00012043244588227796, "loss": 41.9922, "step": 117 }, { "epoch": 0.07782357790601814, "grad_norm": 0.42974764108657837, "learning_rate": 0.00011800505538843798, "loss": 41.9609, "step": 118 }, { "epoch": 0.07848309975267931, "grad_norm": 0.32136955857276917, "learning_rate": 0.00011558641199298727, "loss": 41.9648, "step": 119 }, { "epoch": 0.07914262159934048, "grad_norm": 0.3226495385169983, "learning_rate": 0.00011317717692888012, "loss": 41.9551, "step": 120 }, { "epoch": 0.07980214344600164, "grad_norm": 0.31127500534057617, "learning_rate": 0.00011077800885692702, "loss": 41.9531, "step": 121 }, { "epoch": 0.08046166529266283, "grad_norm": 0.4595889151096344, "learning_rate": 0.00010838956368572334, "loss": 41.9473, "step": 122 }, { "epoch": 0.081121187139324, "grad_norm": 0.382497102022171, "learning_rate": 0.0001060124943923303, "loss": 41.9473, "step": 123 }, { "epoch": 0.08178070898598516, "grad_norm": 0.27015018463134766, "learning_rate": 0.0001036474508437579, "loss": 41.9668, "step": 124 }, { "epoch": 0.08244023083264633, "grad_norm": 0.28897708654403687, "learning_rate": 0.00010129507961929748, "loss": 41.9609, "step": 125 }, { "epoch": 0.08244023083264633, "eval_loss": 10.48843765258789, "eval_runtime": 0.114, "eval_samples_per_second": 438.508, "eval_steps_per_second": 61.391, "step": 125 }, { "epoch": 0.0830997526793075, "grad_norm": 0.39089083671569824, "learning_rate": 9.895602383375353e-05, "loss": 41.9727, "step": 126 }, { "epoch": 0.08375927452596867, "grad_norm": 0.7614229321479797, "learning_rate": 9.663092296162251e-05, "loss": 41.9121, "step": 127 }, { "epoch": 0.08441879637262985, "grad_norm": 0.3535849153995514, "learning_rate": 9.432041266226686e-05, "loss": 41.9707, "step": 128 }, { "epoch": 0.08507831821929102, "grad_norm": 0.5981976985931396, "learning_rate": 9.202512460613219e-05, "loss": 41.9141, "step": 129 }, { "epoch": 0.08573784006595218, "grad_norm": 0.7550039291381836, "learning_rate": 8.97456863020546e-05, "loss": 42.0059, "step": 130 }, { "epoch": 0.08639736191261335, "grad_norm": 0.294890820980072, "learning_rate": 8.748272092570646e-05, "loss": 41.959, "step": 131 }, { "epoch": 0.08705688375927452, "grad_norm": 0.4648720920085907, "learning_rate": 8.523684714922608e-05, "loss": 41.9805, "step": 132 }, { "epoch": 0.0877164056059357, "grad_norm": 0.6221141219139099, "learning_rate": 8.300867897207903e-05, "loss": 42.002, "step": 133 }, { "epoch": 0.08837592745259687, "grad_norm": 0.7409745454788208, "learning_rate": 8.079882555319684e-05, "loss": 41.9043, "step": 134 }, { "epoch": 0.08903544929925804, "grad_norm": 0.36831387877464294, "learning_rate": 7.860789104443896e-05, "loss": 41.9746, "step": 135 }, { "epoch": 0.0896949711459192, "grad_norm": 0.3386104702949524, "learning_rate": 7.643647442542382e-05, "loss": 41.9551, "step": 136 }, { "epoch": 0.09035449299258037, "grad_norm": 0.342767596244812, "learning_rate": 7.428516933977347e-05, "loss": 41.9473, "step": 137 }, { "epoch": 0.09101401483924156, "grad_norm": 0.5757667422294617, "learning_rate": 7.215456393281776e-05, "loss": 41.998, "step": 138 }, { "epoch": 0.09167353668590272, "grad_norm": 0.38113224506378174, "learning_rate": 7.004524069080096e-05, "loss": 41.9746, "step": 139 }, { "epoch": 0.09233305853256389, "grad_norm": 0.4469098746776581, "learning_rate": 6.795777628163599e-05, "loss": 41.9219, "step": 140 }, { "epoch": 0.09299258037922506, "grad_norm": 0.3035033643245697, "learning_rate": 6.58927413972491e-05, "loss": 41.9707, "step": 141 }, { "epoch": 0.09365210222588623, "grad_norm": 0.44389528036117554, "learning_rate": 6.385070059755846e-05, "loss": 41.9824, "step": 142 }, { "epoch": 0.09431162407254741, "grad_norm": 0.5598168969154358, "learning_rate": 6.183221215612904e-05, "loss": 41.9473, "step": 143 }, { "epoch": 0.09497114591920858, "grad_norm": 0.33077386021614075, "learning_rate": 5.983782790754623e-05, "loss": 41.9492, "step": 144 }, { "epoch": 0.09563066776586975, "grad_norm": 0.3129209578037262, "learning_rate": 5.786809309654982e-05, "loss": 41.9727, "step": 145 }, { "epoch": 0.09629018961253091, "grad_norm": 0.24582117795944214, "learning_rate": 5.592354622896944e-05, "loss": 41.9609, "step": 146 }, { "epoch": 0.09694971145919208, "grad_norm": 0.49745967984199524, "learning_rate": 5.40047189245025e-05, "loss": 41.9277, "step": 147 }, { "epoch": 0.09760923330585326, "grad_norm": 0.313666969537735, "learning_rate": 5.211213577137469e-05, "loss": 41.957, "step": 148 }, { "epoch": 0.09826875515251443, "grad_norm": 0.4432874917984009, "learning_rate": 5.024631418292274e-05, "loss": 41.9395, "step": 149 }, { "epoch": 0.0989282769991756, "grad_norm": 0.26742538809776306, "learning_rate": 4.840776425613886e-05, "loss": 41.9688, "step": 150 }, { "epoch": 0.0989282769991756, "eval_loss": 10.48843765258789, "eval_runtime": 0.1139, "eval_samples_per_second": 439.033, "eval_steps_per_second": 61.465, "step": 150 } ], "logging_steps": 1, "max_steps": 200, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 50, "stateful_callbacks": { "EarlyStoppingCallback": { "args": { "early_stopping_patience": 1, "early_stopping_threshold": 0.0 }, "attributes": { "early_stopping_patience_counter": 0 } }, "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1193213952000.0, "train_batch_size": 8, "trial_name": null, "trial_params": null }