{ "best_metric": 0.530606210231781, "best_model_checkpoint": "miner_id_24/checkpoint-300", "epoch": 0.068575347162695, "eval_steps": 50, "global_step": 300, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0002285844905423167, "grad_norm": 12.226268768310547, "learning_rate": 3.3333333333333333e-06, "loss": 3.9387, "step": 1 }, { "epoch": 0.0002285844905423167, "eval_loss": 4.918857097625732, "eval_runtime": 1326.574, "eval_samples_per_second": 5.554, "eval_steps_per_second": 2.777, "step": 1 }, { "epoch": 0.0004571689810846334, "grad_norm": 16.508821487426758, "learning_rate": 6.666666666666667e-06, "loss": 4.288, "step": 2 }, { "epoch": 0.0006857534716269502, "grad_norm": 19.46694564819336, "learning_rate": 1e-05, "loss": 4.2996, "step": 3 }, { "epoch": 0.0009143379621692668, "grad_norm": 16.385190963745117, "learning_rate": 1.3333333333333333e-05, "loss": 3.5999, "step": 4 }, { "epoch": 0.0011429224527115835, "grad_norm": 19.315223693847656, "learning_rate": 1.6666666666666667e-05, "loss": 3.9415, "step": 5 }, { "epoch": 0.0013715069432539003, "grad_norm": 20.83527946472168, "learning_rate": 2e-05, "loss": 3.7644, "step": 6 }, { "epoch": 0.001600091433796217, "grad_norm": 50.4894905090332, "learning_rate": 2.3333333333333336e-05, "loss": 5.543, "step": 7 }, { "epoch": 0.0018286759243385335, "grad_norm": 109.4534912109375, "learning_rate": 2.6666666666666667e-05, "loss": 9.1326, "step": 8 }, { "epoch": 0.00205726041488085, "grad_norm": 209.9546356201172, "learning_rate": 3e-05, "loss": 11.0199, "step": 9 }, { "epoch": 0.002285844905423167, "grad_norm": 126.74362182617188, "learning_rate": 3.3333333333333335e-05, "loss": 6.5991, "step": 10 }, { "epoch": 0.002514429395965484, "grad_norm": 62.07261657714844, "learning_rate": 3.6666666666666666e-05, "loss": 4.5192, "step": 11 }, { "epoch": 0.0027430138865078006, "grad_norm": 57.047672271728516, "learning_rate": 4e-05, "loss": 2.792, "step": 12 }, { "epoch": 0.002971598377050117, "grad_norm": 40.239585876464844, "learning_rate": 4.3333333333333334e-05, "loss": 1.6243, "step": 13 }, { "epoch": 0.003200182867592434, "grad_norm": 147.44810485839844, "learning_rate": 4.666666666666667e-05, "loss": 7.7768, "step": 14 }, { "epoch": 0.0034287673581347507, "grad_norm": 129.61868286132812, "learning_rate": 5e-05, "loss": 3.3601, "step": 15 }, { "epoch": 0.003657351848677067, "grad_norm": 67.8729476928711, "learning_rate": 5.333333333333333e-05, "loss": 2.3788, "step": 16 }, { "epoch": 0.003885936339219384, "grad_norm": 53.81801986694336, "learning_rate": 5.666666666666667e-05, "loss": 1.5561, "step": 17 }, { "epoch": 0.0041145208297617, "grad_norm": 108.63700866699219, "learning_rate": 6e-05, "loss": 3.4077, "step": 18 }, { "epoch": 0.004343105320304018, "grad_norm": 92.20858764648438, "learning_rate": 6.333333333333333e-05, "loss": 5.118, "step": 19 }, { "epoch": 0.004571689810846334, "grad_norm": 107.2628402709961, "learning_rate": 6.666666666666667e-05, "loss": 4.0284, "step": 20 }, { "epoch": 0.00480027430138865, "grad_norm": 84.7047348022461, "learning_rate": 7e-05, "loss": 2.3014, "step": 21 }, { "epoch": 0.005028858791930968, "grad_norm": 80.38473510742188, "learning_rate": 7.333333333333333e-05, "loss": 2.9712, "step": 22 }, { "epoch": 0.005257443282473284, "grad_norm": 58.196502685546875, "learning_rate": 7.666666666666667e-05, "loss": 3.1879, "step": 23 }, { "epoch": 0.005486027773015601, "grad_norm": 217.900146484375, "learning_rate": 8e-05, "loss": 10.9929, "step": 24 }, { "epoch": 0.005714612263557918, "grad_norm": 137.6026153564453, "learning_rate": 8.333333333333334e-05, "loss": 5.2346, "step": 25 }, { "epoch": 0.005943196754100234, "grad_norm": 67.36244201660156, "learning_rate": 8.666666666666667e-05, "loss": 1.8588, "step": 26 }, { "epoch": 0.006171781244642551, "grad_norm": 53.583919525146484, "learning_rate": 9e-05, "loss": 1.208, "step": 27 }, { "epoch": 0.006400365735184868, "grad_norm": 247.46914672851562, "learning_rate": 9.333333333333334e-05, "loss": 23.0121, "step": 28 }, { "epoch": 0.006628950225727184, "grad_norm": 309.0799255371094, "learning_rate": 9.666666666666667e-05, "loss": 11.6811, "step": 29 }, { "epoch": 0.006857534716269501, "grad_norm": 88.57025146484375, "learning_rate": 0.0001, "loss": 6.6625, "step": 30 }, { "epoch": 0.007086119206811818, "grad_norm": 37.52093505859375, "learning_rate": 9.999661540018812e-05, "loss": 5.0423, "step": 31 }, { "epoch": 0.007314703697354134, "grad_norm": 59.48794937133789, "learning_rate": 9.998646205897309e-05, "loss": 4.7137, "step": 32 }, { "epoch": 0.007543288187896451, "grad_norm": 31.38121223449707, "learning_rate": 9.99695413509548e-05, "loss": 2.6695, "step": 33 }, { "epoch": 0.007771872678438768, "grad_norm": 45.99576950073242, "learning_rate": 9.994585556692624e-05, "loss": 3.0682, "step": 34 }, { "epoch": 0.008000457168981085, "grad_norm": 26.696224212646484, "learning_rate": 9.991540791356342e-05, "loss": 3.2612, "step": 35 }, { "epoch": 0.0082290416595234, "grad_norm": 46.0620002746582, "learning_rate": 9.987820251299122e-05, "loss": 3.0265, "step": 36 }, { "epoch": 0.008457626150065718, "grad_norm": 28.95342254638672, "learning_rate": 9.983424440222531e-05, "loss": 3.1935, "step": 37 }, { "epoch": 0.008686210640608035, "grad_norm": 20.682689666748047, "learning_rate": 9.978353953249022e-05, "loss": 2.1278, "step": 38 }, { "epoch": 0.00891479513115035, "grad_norm": 95.37089538574219, "learning_rate": 9.972609476841367e-05, "loss": 4.2482, "step": 39 }, { "epoch": 0.009143379621692668, "grad_norm": 36.080413818359375, "learning_rate": 9.966191788709716e-05, "loss": 3.0929, "step": 40 }, { "epoch": 0.009371964112234985, "grad_norm": 42.082088470458984, "learning_rate": 9.959101757706308e-05, "loss": 2.736, "step": 41 }, { "epoch": 0.0096005486027773, "grad_norm": 22.620975494384766, "learning_rate": 9.951340343707852e-05, "loss": 3.1542, "step": 42 }, { "epoch": 0.009829133093319618, "grad_norm": 31.406814575195312, "learning_rate": 9.942908597485558e-05, "loss": 3.6269, "step": 43 }, { "epoch": 0.010057717583861935, "grad_norm": 25.586286544799805, "learning_rate": 9.933807660562898e-05, "loss": 6.2693, "step": 44 }, { "epoch": 0.010286302074404252, "grad_norm": 29.92267608642578, "learning_rate": 9.924038765061042e-05, "loss": 7.0683, "step": 45 }, { "epoch": 0.010514886564946568, "grad_norm": 20.818403244018555, "learning_rate": 9.913603233532067e-05, "loss": 6.0885, "step": 46 }, { "epoch": 0.010743471055488885, "grad_norm": 20.533842086791992, "learning_rate": 9.902502478779896e-05, "loss": 5.4988, "step": 47 }, { "epoch": 0.010972055546031203, "grad_norm": 16.80043601989746, "learning_rate": 9.890738003669029e-05, "loss": 4.4107, "step": 48 }, { "epoch": 0.011200640036573518, "grad_norm": 13.855122566223145, "learning_rate": 9.878311400921072e-05, "loss": 4.2749, "step": 49 }, { "epoch": 0.011429224527115835, "grad_norm": 21.536596298217773, "learning_rate": 9.865224352899119e-05, "loss": 5.1643, "step": 50 }, { "epoch": 0.011429224527115835, "eval_loss": 2.806919813156128, "eval_runtime": 1332.9908, "eval_samples_per_second": 5.527, "eval_steps_per_second": 2.764, "step": 50 }, { "epoch": 0.011657809017658153, "grad_norm": 15.51828670501709, "learning_rate": 9.851478631379982e-05, "loss": 4.3807, "step": 51 }, { "epoch": 0.011886393508200468, "grad_norm": 11.271062850952148, "learning_rate": 9.837076097314319e-05, "loss": 4.1682, "step": 52 }, { "epoch": 0.012114977998742785, "grad_norm": 8.433566093444824, "learning_rate": 9.822018700574695e-05, "loss": 3.3631, "step": 53 }, { "epoch": 0.012343562489285103, "grad_norm": 8.091526985168457, "learning_rate": 9.806308479691595e-05, "loss": 2.9425, "step": 54 }, { "epoch": 0.012572146979827418, "grad_norm": 8.342974662780762, "learning_rate": 9.789947561577445e-05, "loss": 2.9333, "step": 55 }, { "epoch": 0.012800731470369735, "grad_norm": 17.649492263793945, "learning_rate": 9.77293816123866e-05, "loss": 3.3319, "step": 56 }, { "epoch": 0.013029315960912053, "grad_norm": 28.43054962158203, "learning_rate": 9.755282581475769e-05, "loss": 3.0355, "step": 57 }, { "epoch": 0.013257900451454368, "grad_norm": 28.488826751708984, "learning_rate": 9.736983212571646e-05, "loss": 2.6905, "step": 58 }, { "epoch": 0.013486484941996685, "grad_norm": 68.5689926147461, "learning_rate": 9.718042531967918e-05, "loss": 2.8557, "step": 59 }, { "epoch": 0.013715069432539003, "grad_norm": 7.864617347717285, "learning_rate": 9.698463103929542e-05, "loss": 0.1518, "step": 60 }, { "epoch": 0.013943653923081318, "grad_norm": 4.028836727142334, "learning_rate": 9.678247579197657e-05, "loss": 0.0393, "step": 61 }, { "epoch": 0.014172238413623636, "grad_norm": 78.4482192993164, "learning_rate": 9.657398694630712e-05, "loss": 0.2168, "step": 62 }, { "epoch": 0.014400822904165953, "grad_norm": 212.4618682861328, "learning_rate": 9.635919272833938e-05, "loss": 3.4277, "step": 63 }, { "epoch": 0.014629407394708268, "grad_norm": 142.9414825439453, "learning_rate": 9.613812221777212e-05, "loss": 11.2092, "step": 64 }, { "epoch": 0.014857991885250586, "grad_norm": 106.03768157958984, "learning_rate": 9.591080534401371e-05, "loss": 3.2856, "step": 65 }, { "epoch": 0.015086576375792903, "grad_norm": 130.8965606689453, "learning_rate": 9.567727288213005e-05, "loss": 3.508, "step": 66 }, { "epoch": 0.015315160866335218, "grad_norm": 86.60301208496094, "learning_rate": 9.543755644867822e-05, "loss": 2.7746, "step": 67 }, { "epoch": 0.015543745356877536, "grad_norm": 62.65277862548828, "learning_rate": 9.519168849742604e-05, "loss": 1.4813, "step": 68 }, { "epoch": 0.01577232984741985, "grad_norm": 44.15681076049805, "learning_rate": 9.493970231495835e-05, "loss": 3.2382, "step": 69 }, { "epoch": 0.01600091433796217, "grad_norm": 54.2675666809082, "learning_rate": 9.468163201617062e-05, "loss": 3.0064, "step": 70 }, { "epoch": 0.016229498828504486, "grad_norm": 20.503276824951172, "learning_rate": 9.441751253965021e-05, "loss": 1.6639, "step": 71 }, { "epoch": 0.0164580833190468, "grad_norm": 42.72779846191406, "learning_rate": 9.414737964294636e-05, "loss": 1.9632, "step": 72 }, { "epoch": 0.01668666780958912, "grad_norm": 94.91614532470703, "learning_rate": 9.38712698977291e-05, "loss": 7.9977, "step": 73 }, { "epoch": 0.016915252300131436, "grad_norm": 120.97164154052734, "learning_rate": 9.358922068483812e-05, "loss": 6.0946, "step": 74 }, { "epoch": 0.01714383679067375, "grad_norm": 65.04187774658203, "learning_rate": 9.330127018922194e-05, "loss": 2.1872, "step": 75 }, { "epoch": 0.01737242128121607, "grad_norm": 102.3792495727539, "learning_rate": 9.300745739476829e-05, "loss": 1.955, "step": 76 }, { "epoch": 0.017601005771758386, "grad_norm": 78.6114501953125, "learning_rate": 9.270782207902629e-05, "loss": 1.5882, "step": 77 }, { "epoch": 0.0178295902623007, "grad_norm": 112.37207794189453, "learning_rate": 9.24024048078213e-05, "loss": 9.1085, "step": 78 }, { "epoch": 0.01805817475284302, "grad_norm": 123.03736114501953, "learning_rate": 9.209124692976287e-05, "loss": 12.134, "step": 79 }, { "epoch": 0.018286759243385336, "grad_norm": 64.75898742675781, "learning_rate": 9.177439057064683e-05, "loss": 8.9277, "step": 80 }, { "epoch": 0.01851534373392765, "grad_norm": 35.20903015136719, "learning_rate": 9.145187862775209e-05, "loss": 5.1302, "step": 81 }, { "epoch": 0.01874392822446997, "grad_norm": 62.34042739868164, "learning_rate": 9.112375476403312e-05, "loss": 5.2516, "step": 82 }, { "epoch": 0.018972512715012286, "grad_norm": 45.33003616333008, "learning_rate": 9.079006340220862e-05, "loss": 4.3035, "step": 83 }, { "epoch": 0.0192010972055546, "grad_norm": 71.29220581054688, "learning_rate": 9.045084971874738e-05, "loss": 4.2392, "step": 84 }, { "epoch": 0.01942968169609692, "grad_norm": 19.70421600341797, "learning_rate": 9.01061596377522e-05, "loss": 3.9666, "step": 85 }, { "epoch": 0.019658266186639236, "grad_norm": 14.824080467224121, "learning_rate": 8.97560398247424e-05, "loss": 2.4165, "step": 86 }, { "epoch": 0.019886850677181555, "grad_norm": 14.691969871520996, "learning_rate": 8.940053768033609e-05, "loss": 2.522, "step": 87 }, { "epoch": 0.02011543516772387, "grad_norm": 20.41330337524414, "learning_rate": 8.903970133383297e-05, "loss": 2.6812, "step": 88 }, { "epoch": 0.020344019658266186, "grad_norm": 16.38035774230957, "learning_rate": 8.86735796366982e-05, "loss": 1.7804, "step": 89 }, { "epoch": 0.020572604148808505, "grad_norm": 14.719985008239746, "learning_rate": 8.83022221559489e-05, "loss": 2.0308, "step": 90 }, { "epoch": 0.02080118863935082, "grad_norm": 21.014904022216797, "learning_rate": 8.792567916744346e-05, "loss": 2.2666, "step": 91 }, { "epoch": 0.021029773129893136, "grad_norm": 27.58352279663086, "learning_rate": 8.754400164907497e-05, "loss": 3.785, "step": 92 }, { "epoch": 0.021258357620435455, "grad_norm": 24.515430450439453, "learning_rate": 8.715724127386972e-05, "loss": 5.0201, "step": 93 }, { "epoch": 0.02148694211097777, "grad_norm": 28.423784255981445, "learning_rate": 8.676545040299145e-05, "loss": 6.3986, "step": 94 }, { "epoch": 0.021715526601520086, "grad_norm": 19.08306312561035, "learning_rate": 8.636868207865244e-05, "loss": 6.0112, "step": 95 }, { "epoch": 0.021944111092062405, "grad_norm": 18.091100692749023, "learning_rate": 8.596699001693255e-05, "loss": 4.6697, "step": 96 }, { "epoch": 0.02217269558260472, "grad_norm": 16.675926208496094, "learning_rate": 8.556042860050687e-05, "loss": 5.068, "step": 97 }, { "epoch": 0.022401280073147036, "grad_norm": 13.843223571777344, "learning_rate": 8.51490528712831e-05, "loss": 4.7174, "step": 98 }, { "epoch": 0.022629864563689355, "grad_norm": 11.593822479248047, "learning_rate": 8.473291852294987e-05, "loss": 3.1727, "step": 99 }, { "epoch": 0.02285844905423167, "grad_norm": 14.757932662963867, "learning_rate": 8.43120818934367e-05, "loss": 4.6322, "step": 100 }, { "epoch": 0.02285844905423167, "eval_loss": 2.1855103969573975, "eval_runtime": 1333.7195, "eval_samples_per_second": 5.524, "eval_steps_per_second": 2.762, "step": 100 }, { "epoch": 0.023087033544773986, "grad_norm": 14.086247444152832, "learning_rate": 8.388659995728663e-05, "loss": 4.5384, "step": 101 }, { "epoch": 0.023315618035316305, "grad_norm": 10.055537223815918, "learning_rate": 8.345653031794292e-05, "loss": 4.1292, "step": 102 }, { "epoch": 0.02354420252585862, "grad_norm": 7.018960952758789, "learning_rate": 8.302193119995039e-05, "loss": 3.673, "step": 103 }, { "epoch": 0.023772787016400936, "grad_norm": 7.652724266052246, "learning_rate": 8.258286144107276e-05, "loss": 3.1104, "step": 104 }, { "epoch": 0.024001371506943255, "grad_norm": 9.536213874816895, "learning_rate": 8.213938048432697e-05, "loss": 2.8206, "step": 105 }, { "epoch": 0.02422995599748557, "grad_norm": 9.198083877563477, "learning_rate": 8.169154836993551e-05, "loss": 2.8642, "step": 106 }, { "epoch": 0.024458540488027886, "grad_norm": 14.823854446411133, "learning_rate": 8.1239425727198e-05, "loss": 2.7631, "step": 107 }, { "epoch": 0.024687124978570205, "grad_norm": 44.76399612426758, "learning_rate": 8.07830737662829e-05, "loss": 2.9486, "step": 108 }, { "epoch": 0.02491570946911252, "grad_norm": 29.669414520263672, "learning_rate": 8.032255426994069e-05, "loss": 2.7715, "step": 109 }, { "epoch": 0.025144293959654836, "grad_norm": 60.78176498413086, "learning_rate": 7.985792958513931e-05, "loss": 2.6962, "step": 110 }, { "epoch": 0.025372878450197155, "grad_norm": 17.208572387695312, "learning_rate": 7.938926261462366e-05, "loss": 0.353, "step": 111 }, { "epoch": 0.02560146294073947, "grad_norm": 3.3832650184631348, "learning_rate": 7.891661680839932e-05, "loss": 0.0653, "step": 112 }, { "epoch": 0.025830047431281786, "grad_norm": 7.614874839782715, "learning_rate": 7.844005615514259e-05, "loss": 0.0529, "step": 113 }, { "epoch": 0.026058631921824105, "grad_norm": 56.453670501708984, "learning_rate": 7.795964517353735e-05, "loss": 3.9549, "step": 114 }, { "epoch": 0.02628721641236642, "grad_norm": 91.68329620361328, "learning_rate": 7.74754489035403e-05, "loss": 4.0682, "step": 115 }, { "epoch": 0.026515800902908736, "grad_norm": 57.930580139160156, "learning_rate": 7.698753289757565e-05, "loss": 1.4162, "step": 116 }, { "epoch": 0.026744385393451055, "grad_norm": 112.63847351074219, "learning_rate": 7.649596321166024e-05, "loss": 1.2008, "step": 117 }, { "epoch": 0.02697296988399337, "grad_norm": 149.06593322753906, "learning_rate": 7.600080639646077e-05, "loss": 1.9499, "step": 118 }, { "epoch": 0.027201554374535687, "grad_norm": 399.54156494140625, "learning_rate": 7.550212948828377e-05, "loss": 2.0733, "step": 119 }, { "epoch": 0.027430138865078006, "grad_norm": 148.1618194580078, "learning_rate": 7.500000000000001e-05, "loss": 4.0831, "step": 120 }, { "epoch": 0.02765872335562032, "grad_norm": 76.7452621459961, "learning_rate": 7.449448591190435e-05, "loss": 3.1251, "step": 121 }, { "epoch": 0.027887307846162637, "grad_norm": 71.67847442626953, "learning_rate": 7.398565566251232e-05, "loss": 2.4827, "step": 122 }, { "epoch": 0.028115892336704956, "grad_norm": 23.501968383789062, "learning_rate": 7.347357813929454e-05, "loss": 1.5631, "step": 123 }, { "epoch": 0.02834447682724727, "grad_norm": 136.9478302001953, "learning_rate": 7.295832266935059e-05, "loss": 7.535, "step": 124 }, { "epoch": 0.028573061317789587, "grad_norm": 91.296142578125, "learning_rate": 7.243995901002312e-05, "loss": 3.9653, "step": 125 }, { "epoch": 0.028801645808331906, "grad_norm": 48.54848861694336, "learning_rate": 7.191855733945387e-05, "loss": 2.1426, "step": 126 }, { "epoch": 0.02903023029887422, "grad_norm": 202.33116149902344, "learning_rate": 7.139418824708272e-05, "loss": 2.8557, "step": 127 }, { "epoch": 0.029258814789416537, "grad_norm": 1095.19873046875, "learning_rate": 7.08669227240909e-05, "loss": 4.8724, "step": 128 }, { "epoch": 0.029487399279958856, "grad_norm": 1942.80859375, "learning_rate": 7.033683215379002e-05, "loss": 20.7307, "step": 129 }, { "epoch": 0.02971598377050117, "grad_norm": 439.0259094238281, "learning_rate": 6.980398830195785e-05, "loss": 15.7319, "step": 130 }, { "epoch": 0.029944568261043487, "grad_norm": 145.61013793945312, "learning_rate": 6.926846330712242e-05, "loss": 9.0764, "step": 131 }, { "epoch": 0.030173152751585806, "grad_norm": 65.11675262451172, "learning_rate": 6.873032967079561e-05, "loss": 7.8135, "step": 132 }, { "epoch": 0.03040173724212812, "grad_norm": 371.9282531738281, "learning_rate": 6.818966024765758e-05, "loss": 7.8735, "step": 133 }, { "epoch": 0.030630321732670437, "grad_norm": 113.37095642089844, "learning_rate": 6.764652823569344e-05, "loss": 5.4435, "step": 134 }, { "epoch": 0.030858906223212756, "grad_norm": 152.44851684570312, "learning_rate": 6.710100716628344e-05, "loss": 4.9697, "step": 135 }, { "epoch": 0.03108749071375507, "grad_norm": 29.180082321166992, "learning_rate": 6.65531708942479e-05, "loss": 3.2838, "step": 136 }, { "epoch": 0.03131607520429739, "grad_norm": 26.18000030517578, "learning_rate": 6.600309358784857e-05, "loss": 3.4387, "step": 137 }, { "epoch": 0.0315446596948397, "grad_norm": 39.24956512451172, "learning_rate": 6.545084971874738e-05, "loss": 3.3899, "step": 138 }, { "epoch": 0.03177324418538202, "grad_norm": 27.828189849853516, "learning_rate": 6.48965140519241e-05, "loss": 2.435, "step": 139 }, { "epoch": 0.03200182867592434, "grad_norm": 28.83315086364746, "learning_rate": 6.434016163555452e-05, "loss": 3.0311, "step": 140 }, { "epoch": 0.03223041316646665, "grad_norm": 23.669389724731445, "learning_rate": 6.378186779084995e-05, "loss": 2.4059, "step": 141 }, { "epoch": 0.03245899765700897, "grad_norm": 13.909722328186035, "learning_rate": 6.322170810186012e-05, "loss": 1.7042, "step": 142 }, { "epoch": 0.03268758214755129, "grad_norm": 14.179429054260254, "learning_rate": 6.26597584052401e-05, "loss": 3.3267, "step": 143 }, { "epoch": 0.0329161666380936, "grad_norm": 17.85869026184082, "learning_rate": 6.209609477998338e-05, "loss": 5.1052, "step": 144 }, { "epoch": 0.03314475112863592, "grad_norm": 23.01960563659668, "learning_rate": 6.153079353712201e-05, "loss": 6.2258, "step": 145 }, { "epoch": 0.03337333561917824, "grad_norm": 17.689586639404297, "learning_rate": 6.096393120939516e-05, "loss": 5.5296, "step": 146 }, { "epoch": 0.03360192010972055, "grad_norm": 15.267470359802246, "learning_rate": 6.0395584540887963e-05, "loss": 5.1839, "step": 147 }, { "epoch": 0.03383050460026287, "grad_norm": 13.088833808898926, "learning_rate": 5.982583047664151e-05, "loss": 4.1851, "step": 148 }, { "epoch": 0.03405908909080519, "grad_norm": 13.637064933776855, "learning_rate": 5.925474615223573e-05, "loss": 4.2342, "step": 149 }, { "epoch": 0.0342876735813475, "grad_norm": 11.020079612731934, "learning_rate": 5.868240888334653e-05, "loss": 4.5138, "step": 150 }, { "epoch": 0.0342876735813475, "eval_loss": 1.670865774154663, "eval_runtime": 1333.0768, "eval_samples_per_second": 5.527, "eval_steps_per_second": 2.764, "step": 150 }, { "epoch": 0.03451625807188982, "grad_norm": 15.293791770935059, "learning_rate": 5.810889615527838e-05, "loss": 4.6994, "step": 151 }, { "epoch": 0.03474484256243214, "grad_norm": 9.839681625366211, "learning_rate": 5.753428561247416e-05, "loss": 3.8388, "step": 152 }, { "epoch": 0.03497342705297445, "grad_norm": 7.527024745941162, "learning_rate": 5.695865504800327e-05, "loss": 3.2164, "step": 153 }, { "epoch": 0.03520201154351677, "grad_norm": 8.136163711547852, "learning_rate": 5.6382082393029746e-05, "loss": 3.0072, "step": 154 }, { "epoch": 0.03543059603405909, "grad_norm": 7.6138691902160645, "learning_rate": 5.5804645706261514e-05, "loss": 2.8556, "step": 155 }, { "epoch": 0.0356591805246014, "grad_norm": 23.783597946166992, "learning_rate": 5.522642316338268e-05, "loss": 2.5473, "step": 156 }, { "epoch": 0.03588776501514372, "grad_norm": 20.85479164123535, "learning_rate": 5.464749304646962e-05, "loss": 2.6145, "step": 157 }, { "epoch": 0.03611634950568604, "grad_norm": 24.58341407775879, "learning_rate": 5.4067933733392915e-05, "loss": 2.7502, "step": 158 }, { "epoch": 0.03634493399622835, "grad_norm": 30.31068992614746, "learning_rate": 5.348782368720626e-05, "loss": 1.2587, "step": 159 }, { "epoch": 0.03657351848677067, "grad_norm": 9.1329984664917, "learning_rate": 5.290724144552379e-05, "loss": 0.1588, "step": 160 }, { "epoch": 0.03680210297731299, "grad_norm": 30.12510108947754, "learning_rate": 5.232626560988735e-05, "loss": 0.4367, "step": 161 }, { "epoch": 0.0370306874678553, "grad_norm": 21.02296257019043, "learning_rate": 5.174497483512506e-05, "loss": 0.4107, "step": 162 }, { "epoch": 0.03725927195839762, "grad_norm": 39.78958511352539, "learning_rate": 5.116344781870281e-05, "loss": 2.3939, "step": 163 }, { "epoch": 0.03748785644893994, "grad_norm": 89.31826782226562, "learning_rate": 5.0581763290069865e-05, "loss": 6.3699, "step": 164 }, { "epoch": 0.03771644093948225, "grad_norm": 50.574974060058594, "learning_rate": 5e-05, "loss": 3.5648, "step": 165 }, { "epoch": 0.03794502543002457, "grad_norm": 30.325014114379883, "learning_rate": 4.941823670993016e-05, "loss": 2.0516, "step": 166 }, { "epoch": 0.03817360992056689, "grad_norm": 19.043434143066406, "learning_rate": 4.883655218129719e-05, "loss": 1.2083, "step": 167 }, { "epoch": 0.0384021944111092, "grad_norm": 29.482452392578125, "learning_rate": 4.825502516487497e-05, "loss": 2.1119, "step": 168 }, { "epoch": 0.03863077890165152, "grad_norm": 37.51374816894531, "learning_rate": 4.767373439011267e-05, "loss": 2.4501, "step": 169 }, { "epoch": 0.03885936339219384, "grad_norm": 23.2747745513916, "learning_rate": 4.709275855447621e-05, "loss": 2.0485, "step": 170 }, { "epoch": 0.03908794788273615, "grad_norm": 15.753173828125, "learning_rate": 4.6512176312793736e-05, "loss": 1.2318, "step": 171 }, { "epoch": 0.03931653237327847, "grad_norm": 26.802000045776367, "learning_rate": 4.593206626660709e-05, "loss": 1.7534, "step": 172 }, { "epoch": 0.03954511686382079, "grad_norm": 94.17913818359375, "learning_rate": 4.535250695353039e-05, "loss": 6.9058, "step": 173 }, { "epoch": 0.03977370135436311, "grad_norm": 129.48439025878906, "learning_rate": 4.477357683661734e-05, "loss": 7.7345, "step": 174 }, { "epoch": 0.04000228584490542, "grad_norm": 83.24310302734375, "learning_rate": 4.4195354293738484e-05, "loss": 5.7297, "step": 175 }, { "epoch": 0.04023087033544774, "grad_norm": 116.52914428710938, "learning_rate": 4.361791760697027e-05, "loss": 4.7677, "step": 176 }, { "epoch": 0.04045945482599006, "grad_norm": 39.86652374267578, "learning_rate": 4.3041344951996746e-05, "loss": 1.596, "step": 177 }, { "epoch": 0.04068803931653237, "grad_norm": 14.975143432617188, "learning_rate": 4.246571438752585e-05, "loss": 1.2558, "step": 178 }, { "epoch": 0.04091662380707469, "grad_norm": 86.05208587646484, "learning_rate": 4.1891103844721636e-05, "loss": 8.7519, "step": 179 }, { "epoch": 0.04114520829761701, "grad_norm": 40.95692825317383, "learning_rate": 4.131759111665349e-05, "loss": 6.2332, "step": 180 }, { "epoch": 0.04137379278815932, "grad_norm": 82.19147491455078, "learning_rate": 4.074525384776428e-05, "loss": 7.8078, "step": 181 }, { "epoch": 0.04160237727870164, "grad_norm": 46.78995132446289, "learning_rate": 4.017416952335849e-05, "loss": 5.3358, "step": 182 }, { "epoch": 0.04183096176924396, "grad_norm": 23.36475372314453, "learning_rate": 3.960441545911204e-05, "loss": 4.3148, "step": 183 }, { "epoch": 0.04205954625978627, "grad_norm": 25.982982635498047, "learning_rate": 3.903606879060483e-05, "loss": 2.5694, "step": 184 }, { "epoch": 0.04228813075032859, "grad_norm": 16.722654342651367, "learning_rate": 3.846920646287799e-05, "loss": 2.6081, "step": 185 }, { "epoch": 0.04251671524087091, "grad_norm": 18.733665466308594, "learning_rate": 3.790390522001662e-05, "loss": 2.1414, "step": 186 }, { "epoch": 0.04274529973141322, "grad_norm": 22.727380752563477, "learning_rate": 3.7340241594759916e-05, "loss": 3.6351, "step": 187 }, { "epoch": 0.04297388422195554, "grad_norm": 16.99645233154297, "learning_rate": 3.67782918981399e-05, "loss": 2.4938, "step": 188 }, { "epoch": 0.04320246871249786, "grad_norm": 27.48732566833496, "learning_rate": 3.6218132209150045e-05, "loss": 2.0775, "step": 189 }, { "epoch": 0.04343105320304017, "grad_norm": 15.861635208129883, "learning_rate": 3.5659838364445505e-05, "loss": 2.5506, "step": 190 }, { "epoch": 0.04365963769358249, "grad_norm": 25.20842933654785, "learning_rate": 3.51034859480759e-05, "loss": 2.6434, "step": 191 }, { "epoch": 0.04388822218412481, "grad_norm": 28.7867374420166, "learning_rate": 3.4549150281252636e-05, "loss": 4.0947, "step": 192 }, { "epoch": 0.04411680667466712, "grad_norm": 16.924789428710938, "learning_rate": 3.399690641215142e-05, "loss": 4.607, "step": 193 }, { "epoch": 0.04434539116520944, "grad_norm": 19.492359161376953, "learning_rate": 3.34468291057521e-05, "loss": 5.7136, "step": 194 }, { "epoch": 0.04457397565575176, "grad_norm": 20.93442726135254, "learning_rate": 3.289899283371657e-05, "loss": 5.2217, "step": 195 }, { "epoch": 0.04480256014629407, "grad_norm": 15.588977813720703, "learning_rate": 3.235347176430656e-05, "loss": 5.381, "step": 196 }, { "epoch": 0.04503114463683639, "grad_norm": 13.808337211608887, "learning_rate": 3.1810339752342446e-05, "loss": 5.4432, "step": 197 }, { "epoch": 0.04525972912737871, "grad_norm": 14.99952507019043, "learning_rate": 3.12696703292044e-05, "loss": 4.5804, "step": 198 }, { "epoch": 0.04548831361792102, "grad_norm": 13.91661548614502, "learning_rate": 3.073153669287759e-05, "loss": 4.1034, "step": 199 }, { "epoch": 0.04571689810846334, "grad_norm": 13.113777160644531, "learning_rate": 3.019601169804216e-05, "loss": 4.2222, "step": 200 }, { "epoch": 0.04571689810846334, "eval_loss": 1.62479567527771, "eval_runtime": 1331.6429, "eval_samples_per_second": 5.533, "eval_steps_per_second": 2.767, "step": 200 }, { "epoch": 0.04594548259900566, "grad_norm": 10.6249418258667, "learning_rate": 2.9663167846209998e-05, "loss": 3.5265, "step": 201 }, { "epoch": 0.04617406708954797, "grad_norm": 8.47287368774414, "learning_rate": 2.9133077275909108e-05, "loss": 3.4879, "step": 202 }, { "epoch": 0.04640265158009029, "grad_norm": 7.781945705413818, "learning_rate": 2.86058117529173e-05, "loss": 3.0601, "step": 203 }, { "epoch": 0.04663123607063261, "grad_norm": 7.176991939544678, "learning_rate": 2.8081442660546125e-05, "loss": 2.6794, "step": 204 }, { "epoch": 0.04685982056117492, "grad_norm": 6.956686496734619, "learning_rate": 2.7560040989976892e-05, "loss": 2.3741, "step": 205 }, { "epoch": 0.04708840505171724, "grad_norm": 7.541036605834961, "learning_rate": 2.7041677330649407e-05, "loss": 2.4602, "step": 206 }, { "epoch": 0.04731698954225956, "grad_norm": 26.17896270751953, "learning_rate": 2.6526421860705473e-05, "loss": 1.7172, "step": 207 }, { "epoch": 0.04754557403280187, "grad_norm": 44.718421936035156, "learning_rate": 2.6014344337487707e-05, "loss": 1.8271, "step": 208 }, { "epoch": 0.04777415852334419, "grad_norm": 23.585582733154297, "learning_rate": 2.5505514088095655e-05, "loss": 1.1668, "step": 209 }, { "epoch": 0.04800274301388651, "grad_norm": 18.206127166748047, "learning_rate": 2.500000000000001e-05, "loss": 0.7469, "step": 210 }, { "epoch": 0.04823132750442882, "grad_norm": 4.393265724182129, "learning_rate": 2.4497870511716235e-05, "loss": 0.1077, "step": 211 }, { "epoch": 0.04845991199497114, "grad_norm": 4.106040954589844, "learning_rate": 2.399919360353923e-05, "loss": 0.2907, "step": 212 }, { "epoch": 0.04868849648551346, "grad_norm": 13.828383445739746, "learning_rate": 2.350403678833976e-05, "loss": 0.8737, "step": 213 }, { "epoch": 0.04891708097605577, "grad_norm": 59.581298828125, "learning_rate": 2.3012467102424373e-05, "loss": 4.8994, "step": 214 }, { "epoch": 0.04914566546659809, "grad_norm": 52.219642639160156, "learning_rate": 2.25245510964597e-05, "loss": 4.3086, "step": 215 }, { "epoch": 0.04937424995714041, "grad_norm": 57.01164627075195, "learning_rate": 2.2040354826462668e-05, "loss": 3.0226, "step": 216 }, { "epoch": 0.04960283444768272, "grad_norm": 35.1441764831543, "learning_rate": 2.1559943844857422e-05, "loss": 1.8326, "step": 217 }, { "epoch": 0.04983141893822504, "grad_norm": 23.194089889526367, "learning_rate": 2.1083383191600674e-05, "loss": 0.9515, "step": 218 }, { "epoch": 0.05006000342876736, "grad_norm": 20.370725631713867, "learning_rate": 2.061073738537635e-05, "loss": 2.2338, "step": 219 }, { "epoch": 0.05028858791930967, "grad_norm": 31.7033634185791, "learning_rate": 2.0142070414860704e-05, "loss": 2.838, "step": 220 }, { "epoch": 0.05051717240985199, "grad_norm": 30.212299346923828, "learning_rate": 1.9677445730059346e-05, "loss": 3.5437, "step": 221 }, { "epoch": 0.05074575690039431, "grad_norm": 24.48371124267578, "learning_rate": 1.9216926233717085e-05, "loss": 2.1516, "step": 222 }, { "epoch": 0.05097434139093662, "grad_norm": 26.28635597229004, "learning_rate": 1.8760574272802e-05, "loss": 1.8107, "step": 223 }, { "epoch": 0.05120292588147894, "grad_norm": 64.8018569946289, "learning_rate": 1.8308451630064484e-05, "loss": 3.8766, "step": 224 }, { "epoch": 0.05143151037202126, "grad_norm": 79.22872161865234, "learning_rate": 1.7860619515673033e-05, "loss": 5.5396, "step": 225 }, { "epoch": 0.05166009486256357, "grad_norm": 81.44451904296875, "learning_rate": 1.7417138558927244e-05, "loss": 5.3564, "step": 226 }, { "epoch": 0.05188867935310589, "grad_norm": 74.99747467041016, "learning_rate": 1.6978068800049624e-05, "loss": 4.2999, "step": 227 }, { "epoch": 0.05211726384364821, "grad_norm": 30.039758682250977, "learning_rate": 1.6543469682057106e-05, "loss": 4.5715, "step": 228 }, { "epoch": 0.05234584833419052, "grad_norm": 25.890417098999023, "learning_rate": 1.611340004271339e-05, "loss": 4.2147, "step": 229 }, { "epoch": 0.05257443282473284, "grad_norm": 24.455820083618164, "learning_rate": 1.5687918106563326e-05, "loss": 3.6963, "step": 230 }, { "epoch": 0.05280301731527516, "grad_norm": 13.776949882507324, "learning_rate": 1.526708147705013e-05, "loss": 2.5523, "step": 231 }, { "epoch": 0.05303160180581747, "grad_norm": 14.489748001098633, "learning_rate": 1.4850947128716913e-05, "loss": 2.8481, "step": 232 }, { "epoch": 0.05326018629635979, "grad_norm": 14.851402282714844, "learning_rate": 1.4439571399493146e-05, "loss": 2.9429, "step": 233 }, { "epoch": 0.05348877078690211, "grad_norm": 14.959819793701172, "learning_rate": 1.4033009983067452e-05, "loss": 2.7784, "step": 234 }, { "epoch": 0.05371735527744442, "grad_norm": 10.266538619995117, "learning_rate": 1.3631317921347563e-05, "loss": 2.1112, "step": 235 }, { "epoch": 0.05394593976798674, "grad_norm": 13.498571395874023, "learning_rate": 1.3234549597008571e-05, "loss": 2.61, "step": 236 }, { "epoch": 0.05417452425852906, "grad_norm": 13.794647216796875, "learning_rate": 1.2842758726130283e-05, "loss": 2.655, "step": 237 }, { "epoch": 0.05440310874907137, "grad_norm": 12.700047492980957, "learning_rate": 1.245599835092504e-05, "loss": 2.1982, "step": 238 }, { "epoch": 0.05463169323961369, "grad_norm": 11.33856201171875, "learning_rate": 1.2074320832556556e-05, "loss": 1.8852, "step": 239 }, { "epoch": 0.05486027773015601, "grad_norm": 12.628767013549805, "learning_rate": 1.1697777844051105e-05, "loss": 2.6147, "step": 240 }, { "epoch": 0.05508886222069832, "grad_norm": 12.457136154174805, "learning_rate": 1.132642036330181e-05, "loss": 2.2797, "step": 241 }, { "epoch": 0.05531744671124064, "grad_norm": 16.78705406188965, "learning_rate": 1.096029866616704e-05, "loss": 2.1772, "step": 242 }, { "epoch": 0.05554603120178296, "grad_norm": 20.40379524230957, "learning_rate": 1.0599462319663905e-05, "loss": 4.4865, "step": 243 }, { "epoch": 0.05577461569232527, "grad_norm": 47.940460205078125, "learning_rate": 1.0243960175257606e-05, "loss": 4.1248, "step": 244 }, { "epoch": 0.05600320018286759, "grad_norm": 20.962236404418945, "learning_rate": 9.893840362247809e-06, "loss": 5.0609, "step": 245 }, { "epoch": 0.05623178467340991, "grad_norm": 17.944730758666992, "learning_rate": 9.549150281252633e-06, "loss": 5.8087, "step": 246 }, { "epoch": 0.05646036916395222, "grad_norm": 14.370197296142578, "learning_rate": 9.209936597791407e-06, "loss": 4.9449, "step": 247 }, { "epoch": 0.05668895365449454, "grad_norm": 13.268957138061523, "learning_rate": 8.876245235966885e-06, "loss": 4.4975, "step": 248 }, { "epoch": 0.05691753814503686, "grad_norm": 13.039083480834961, "learning_rate": 8.548121372247918e-06, "loss": 3.907, "step": 249 }, { "epoch": 0.05714612263557917, "grad_norm": 14.2072172164917, "learning_rate": 8.225609429353187e-06, "loss": 3.7406, "step": 250 }, { "epoch": 0.05714612263557917, "eval_loss": 0.5557529926300049, "eval_runtime": 1331.7964, "eval_samples_per_second": 5.532, "eval_steps_per_second": 2.766, "step": 250 }, { "epoch": 0.05737470712612149, "grad_norm": 31.539043426513672, "learning_rate": 7.908753070237123e-06, "loss": 4.1344, "step": 251 }, { "epoch": 0.05760329161666381, "grad_norm": 10.60650634765625, "learning_rate": 7.597595192178702e-06, "loss": 3.037, "step": 252 }, { "epoch": 0.05783187610720612, "grad_norm": 5.5060715675354, "learning_rate": 7.292177920973725e-06, "loss": 2.9713, "step": 253 }, { "epoch": 0.05806046059774844, "grad_norm": 5.154425144195557, "learning_rate": 6.992542605231739e-06, "loss": 2.2515, "step": 254 }, { "epoch": 0.05828904508829076, "grad_norm": 5.6335129737854, "learning_rate": 6.698729810778065e-06, "loss": 2.3608, "step": 255 }, { "epoch": 0.05851762957883307, "grad_norm": 6.264554023742676, "learning_rate": 6.410779315161886e-06, "loss": 2.2597, "step": 256 }, { "epoch": 0.05874621406937539, "grad_norm": 9.82129955291748, "learning_rate": 6.128730102270897e-06, "loss": 1.9233, "step": 257 }, { "epoch": 0.05897479855991771, "grad_norm": 12.5000581741333, "learning_rate": 5.852620357053651e-06, "loss": 2.1487, "step": 258 }, { "epoch": 0.05920338305046002, "grad_norm": 17.345787048339844, "learning_rate": 5.582487460349805e-06, "loss": 2.1851, "step": 259 }, { "epoch": 0.05943196754100234, "grad_norm": 15.88868236541748, "learning_rate": 5.318367983829392e-06, "loss": 0.7703, "step": 260 }, { "epoch": 0.05966055203154466, "grad_norm": 18.099214553833008, "learning_rate": 5.060297685041659e-06, "loss": 0.5869, "step": 261 }, { "epoch": 0.05988913652208697, "grad_norm": 11.634132385253906, "learning_rate": 4.8083115025739756e-06, "loss": 0.375, "step": 262 }, { "epoch": 0.06011772101262929, "grad_norm": 12.680456161499023, "learning_rate": 4.562443551321788e-06, "loss": 0.6281, "step": 263 }, { "epoch": 0.06034630550317161, "grad_norm": 10.923900604248047, "learning_rate": 4.322727117869951e-06, "loss": 0.5998, "step": 264 }, { "epoch": 0.06057488999371392, "grad_norm": 33.76637268066406, "learning_rate": 4.089194655986306e-06, "loss": 2.3875, "step": 265 }, { "epoch": 0.06080347448425624, "grad_norm": 29.129016876220703, "learning_rate": 3.861877782227885e-06, "loss": 2.1657, "step": 266 }, { "epoch": 0.06103205897479856, "grad_norm": 25.44561767578125, "learning_rate": 3.6408072716606346e-06, "loss": 2.1342, "step": 267 }, { "epoch": 0.061260643465340874, "grad_norm": 29.16153335571289, "learning_rate": 3.426013053692878e-06, "loss": 1.9011, "step": 268 }, { "epoch": 0.06148922795588319, "grad_norm": 27.153125762939453, "learning_rate": 3.2175242080234313e-06, "loss": 1.852, "step": 269 }, { "epoch": 0.06171781244642551, "grad_norm": 23.434755325317383, "learning_rate": 3.0153689607045845e-06, "loss": 2.194, "step": 270 }, { "epoch": 0.061946396936967824, "grad_norm": 31.765398025512695, "learning_rate": 2.8195746803208244e-06, "loss": 2.348, "step": 271 }, { "epoch": 0.06217498142751014, "grad_norm": 20.402849197387695, "learning_rate": 2.63016787428354e-06, "loss": 2.2186, "step": 272 }, { "epoch": 0.06240356591805246, "grad_norm": 43.326622009277344, "learning_rate": 2.4471741852423237e-06, "loss": 1.9633, "step": 273 }, { "epoch": 0.06263215040859478, "grad_norm": 55.71512222290039, "learning_rate": 2.2706183876134045e-06, "loss": 2.5231, "step": 274 }, { "epoch": 0.0628607348991371, "grad_norm": 34.0222282409668, "learning_rate": 2.100524384225555e-06, "loss": 2.5434, "step": 275 }, { "epoch": 0.0630893193896794, "grad_norm": 30.28105354309082, "learning_rate": 1.9369152030840556e-06, "loss": 2.2744, "step": 276 }, { "epoch": 0.06331790388022172, "grad_norm": 70.08159637451172, "learning_rate": 1.7798129942530551e-06, "loss": 2.8638, "step": 277 }, { "epoch": 0.06354648837076404, "grad_norm": 30.30040740966797, "learning_rate": 1.6292390268568104e-06, "loss": 2.3145, "step": 278 }, { "epoch": 0.06377507286130636, "grad_norm": 12.503889083862305, "learning_rate": 1.4852136862001764e-06, "loss": 1.9287, "step": 279 }, { "epoch": 0.06400365735184868, "grad_norm": 10.321149826049805, "learning_rate": 1.3477564710088098e-06, "loss": 1.7086, "step": 280 }, { "epoch": 0.064232241842391, "grad_norm": 13.305763244628906, "learning_rate": 1.2168859907892904e-06, "loss": 2.1641, "step": 281 }, { "epoch": 0.0644608263329333, "grad_norm": 11.012710571289062, "learning_rate": 1.0926199633097157e-06, "loss": 2.1994, "step": 282 }, { "epoch": 0.06468941082347562, "grad_norm": 11.165496826171875, "learning_rate": 9.749752122010346e-07, "loss": 1.8865, "step": 283 }, { "epoch": 0.06491799531401794, "grad_norm": 8.645009994506836, "learning_rate": 8.639676646793382e-07, "loss": 1.4106, "step": 284 }, { "epoch": 0.06514657980456026, "grad_norm": 12.307551383972168, "learning_rate": 7.596123493895991e-07, "loss": 1.9273, "step": 285 }, { "epoch": 0.06537516429510258, "grad_norm": 11.830700874328613, "learning_rate": 6.61923394371039e-07, "loss": 2.3673, "step": 286 }, { "epoch": 0.0656037487856449, "grad_norm": 10.462738037109375, "learning_rate": 5.7091402514442e-07, "loss": 1.6951, "step": 287 }, { "epoch": 0.0658323332761872, "grad_norm": 11.972997665405273, "learning_rate": 4.865965629214819e-07, "loss": 2.0444, "step": 288 }, { "epoch": 0.06606091776672952, "grad_norm": 18.008352279663086, "learning_rate": 4.089824229369155e-07, "loss": 2.4141, "step": 289 }, { "epoch": 0.06628950225727184, "grad_norm": 11.071226119995117, "learning_rate": 3.380821129028489e-07, "loss": 1.8623, "step": 290 }, { "epoch": 0.06651808674781416, "grad_norm": 15.886430740356445, "learning_rate": 2.7390523158633554e-07, "loss": 2.2248, "step": 291 }, { "epoch": 0.06674667123835648, "grad_norm": 14.11369800567627, "learning_rate": 2.1646046750978254e-07, "loss": 3.1426, "step": 292 }, { "epoch": 0.0669752557288988, "grad_norm": 17.461822509765625, "learning_rate": 1.657555977746972e-07, "loss": 3.738, "step": 293 }, { "epoch": 0.0672038402194411, "grad_norm": 9.49361801147461, "learning_rate": 1.2179748700879012e-07, "loss": 3.9952, "step": 294 }, { "epoch": 0.06743242470998342, "grad_norm": 11.396438598632812, "learning_rate": 8.459208643659122e-08, "loss": 4.4705, "step": 295 }, { "epoch": 0.06766100920052574, "grad_norm": 10.36643123626709, "learning_rate": 5.4144433073771707e-08, "loss": 4.337, "step": 296 }, { "epoch": 0.06788959369106806, "grad_norm": 11.684181213378906, "learning_rate": 3.04586490452119e-08, "loss": 4.1834, "step": 297 }, { "epoch": 0.06811817818161038, "grad_norm": 8.283185958862305, "learning_rate": 1.3537941026914303e-08, "loss": 2.6821, "step": 298 }, { "epoch": 0.0683467626721527, "grad_norm": 10.205315589904785, "learning_rate": 3.384599811889766e-09, "loss": 3.7379, "step": 299 }, { "epoch": 0.068575347162695, "grad_norm": 11.798584938049316, "learning_rate": 0.0, "loss": 3.3263, "step": 300 }, { "epoch": 0.068575347162695, "eval_loss": 0.530606210231781, "eval_runtime": 1331.9933, "eval_samples_per_second": 5.532, "eval_steps_per_second": 2.766, "step": 300 } ], "logging_steps": 1, "max_steps": 300, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 50, "stateful_callbacks": { "EarlyStoppingCallback": { "args": { "early_stopping_patience": 5, "early_stopping_threshold": 0.0 }, "attributes": { "early_stopping_patience_counter": 0 } }, "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 6.686338196994785e+17, "train_batch_size": 8, "trial_name": null, "trial_params": null }