{ "best_metric": 0.28952550888061523, "best_model_checkpoint": "miner_id_24/checkpoint-100", "epoch": 2.9953488372093022, "eval_steps": 50, "global_step": 161, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.018604651162790697, "grad_norm": 1.01740562915802, "learning_rate": 7e-06, "loss": 1.7104, "step": 1 }, { "epoch": 0.018604651162790697, "eval_loss": 0.6080694794654846, "eval_runtime": 10.5925, "eval_samples_per_second": 8.591, "eval_steps_per_second": 2.171, "step": 1 }, { "epoch": 0.037209302325581395, "grad_norm": 1.2670235633850098, "learning_rate": 1.4e-05, "loss": 1.8752, "step": 2 }, { "epoch": 0.05581395348837209, "grad_norm": 1.3142772912979126, "learning_rate": 2.1e-05, "loss": 1.7304, "step": 3 }, { "epoch": 0.07441860465116279, "grad_norm": 1.2250651121139526, "learning_rate": 2.8e-05, "loss": 1.9433, "step": 4 }, { "epoch": 0.09302325581395349, "grad_norm": 1.5665745735168457, "learning_rate": 3.5e-05, "loss": 2.1169, "step": 5 }, { "epoch": 0.11162790697674418, "grad_norm": 1.5763330459594727, "learning_rate": 4.2e-05, "loss": 2.4289, "step": 6 }, { "epoch": 0.13023255813953488, "grad_norm": 1.7122726440429688, "learning_rate": 4.899999999999999e-05, "loss": 2.4843, "step": 7 }, { "epoch": 0.14883720930232558, "grad_norm": 1.7638782262802124, "learning_rate": 5.6e-05, "loss": 1.914, "step": 8 }, { "epoch": 0.16744186046511628, "grad_norm": 2.0044069290161133, "learning_rate": 6.3e-05, "loss": 2.2168, "step": 9 }, { "epoch": 0.18604651162790697, "grad_norm": 2.2012875080108643, "learning_rate": 7e-05, "loss": 2.3124, "step": 10 }, { "epoch": 0.20465116279069767, "grad_norm": 2.3812570571899414, "learning_rate": 6.99924252512213e-05, "loss": 2.1676, "step": 11 }, { "epoch": 0.22325581395348837, "grad_norm": 1.740301251411438, "learning_rate": 6.996970428356058e-05, "loss": 1.9273, "step": 12 }, { "epoch": 0.24186046511627907, "grad_norm": 4.242422580718994, "learning_rate": 6.993184693162481e-05, "loss": 3.0044, "step": 13 }, { "epoch": 0.26046511627906976, "grad_norm": 1.112087368965149, "learning_rate": 6.987886958169575e-05, "loss": 1.5393, "step": 14 }, { "epoch": 0.27906976744186046, "grad_norm": 0.9684474468231201, "learning_rate": 6.981079516463718e-05, "loss": 1.1536, "step": 15 }, { "epoch": 0.29767441860465116, "grad_norm": 0.8984664082527161, "learning_rate": 6.972765314596951e-05, "loss": 1.1399, "step": 16 }, { "epoch": 0.31627906976744186, "grad_norm": 0.9321298599243164, "learning_rate": 6.962947951311591e-05, "loss": 1.4814, "step": 17 }, { "epoch": 0.33488372093023255, "grad_norm": 0.8456665277481079, "learning_rate": 6.951631675982521e-05, "loss": 1.3894, "step": 18 }, { "epoch": 0.35348837209302325, "grad_norm": 0.7932624220848083, "learning_rate": 6.938821386777899e-05, "loss": 1.4062, "step": 19 }, { "epoch": 0.37209302325581395, "grad_norm": 1.1389163732528687, "learning_rate": 6.924522628539012e-05, "loss": 1.4328, "step": 20 }, { "epoch": 0.39069767441860465, "grad_norm": 1.0435296297073364, "learning_rate": 6.908741590380231e-05, "loss": 1.496, "step": 21 }, { "epoch": 0.40930232558139534, "grad_norm": 1.1448780298233032, "learning_rate": 6.891485103010098e-05, "loss": 1.2627, "step": 22 }, { "epoch": 0.42790697674418604, "grad_norm": 1.52295982837677, "learning_rate": 6.87276063577471e-05, "loss": 1.4529, "step": 23 }, { "epoch": 0.44651162790697674, "grad_norm": 1.8864622116088867, "learning_rate": 6.852576293424654e-05, "loss": 2.0397, "step": 24 }, { "epoch": 0.46511627906976744, "grad_norm": 1.491450548171997, "learning_rate": 6.830940812606936e-05, "loss": 1.5753, "step": 25 }, { "epoch": 0.48372093023255813, "grad_norm": 2.425402879714966, "learning_rate": 6.807863558083377e-05, "loss": 1.2114, "step": 26 }, { "epoch": 0.5023255813953489, "grad_norm": 0.5629173517227173, "learning_rate": 6.78335451867715e-05, "loss": 1.6225, "step": 27 }, { "epoch": 0.5209302325581395, "grad_norm": 0.7677317261695862, "learning_rate": 6.757424302949186e-05, "loss": 1.6867, "step": 28 }, { "epoch": 0.5395348837209303, "grad_norm": 0.9299004673957825, "learning_rate": 6.73008413460634e-05, "loss": 1.288, "step": 29 }, { "epoch": 0.5581395348837209, "grad_norm": 0.8549377918243408, "learning_rate": 6.701345847643282e-05, "loss": 1.1299, "step": 30 }, { "epoch": 0.5767441860465117, "grad_norm": 1.0633025169372559, "learning_rate": 6.671221881220246e-05, "loss": 1.5311, "step": 31 }, { "epoch": 0.5953488372093023, "grad_norm": 1.01902174949646, "learning_rate": 6.639725274278824e-05, "loss": 1.3447, "step": 32 }, { "epoch": 0.6139534883720931, "grad_norm": 1.1506015062332153, "learning_rate": 6.606869659898156e-05, "loss": 1.3272, "step": 33 }, { "epoch": 0.6325581395348837, "grad_norm": 1.0536060333251953, "learning_rate": 6.572669259393953e-05, "loss": 1.317, "step": 34 }, { "epoch": 0.6511627906976745, "grad_norm": 1.214861512184143, "learning_rate": 6.537138876162896e-05, "loss": 1.2711, "step": 35 }, { "epoch": 0.6697674418604651, "grad_norm": 1.3296098709106445, "learning_rate": 6.5002938892751e-05, "loss": 1.471, "step": 36 }, { "epoch": 0.6883720930232559, "grad_norm": 1.5142545700073242, "learning_rate": 6.462150246817388e-05, "loss": 1.5411, "step": 37 }, { "epoch": 0.7069767441860465, "grad_norm": 1.7185128927230835, "learning_rate": 6.422724458990284e-05, "loss": 1.1718, "step": 38 }, { "epoch": 0.7255813953488373, "grad_norm": 2.9415876865386963, "learning_rate": 6.38203359096168e-05, "loss": 1.1685, "step": 39 }, { "epoch": 0.7441860465116279, "grad_norm": 0.6370990872383118, "learning_rate": 6.340095255480317e-05, "loss": 1.2429, "step": 40 }, { "epoch": 0.7627906976744186, "grad_norm": 0.7854994535446167, "learning_rate": 6.296927605252219e-05, "loss": 1.1277, "step": 41 }, { "epoch": 0.7813953488372093, "grad_norm": 0.6397520899772644, "learning_rate": 6.252549325083437e-05, "loss": 0.9854, "step": 42 }, { "epoch": 0.8, "grad_norm": 0.8113309741020203, "learning_rate": 6.206979623792457e-05, "loss": 1.5336, "step": 43 }, { "epoch": 0.8186046511627907, "grad_norm": 0.9286227226257324, "learning_rate": 6.160238225895803e-05, "loss": 1.32, "step": 44 }, { "epoch": 0.8372093023255814, "grad_norm": 0.9448747634887695, "learning_rate": 6.112345363070428e-05, "loss": 1.2975, "step": 45 }, { "epoch": 0.8558139534883721, "grad_norm": 0.92212975025177, "learning_rate": 6.063321765396568e-05, "loss": 1.1939, "step": 46 }, { "epoch": 0.8744186046511628, "grad_norm": 1.1105095148086548, "learning_rate": 6.013188652384889e-05, "loss": 1.0333, "step": 47 }, { "epoch": 0.8930232558139535, "grad_norm": 1.1572589874267578, "learning_rate": 5.9619677237917655e-05, "loss": 1.2029, "step": 48 }, { "epoch": 0.9116279069767442, "grad_norm": 1.4601818323135376, "learning_rate": 5.9096811502266993e-05, "loss": 1.3201, "step": 49 }, { "epoch": 0.9302325581395349, "grad_norm": 1.729429006576538, "learning_rate": 5.8563515635559344e-05, "loss": 1.3054, "step": 50 }, { "epoch": 0.9302325581395349, "eval_loss": 0.3099779188632965, "eval_runtime": 10.8877, "eval_samples_per_second": 8.358, "eval_steps_per_second": 2.112, "step": 50 }, { "epoch": 0.9488372093023256, "grad_norm": 1.666062593460083, "learning_rate": 5.8020020471064145e-05, "loss": 1.1947, "step": 51 }, { "epoch": 0.9674418604651163, "grad_norm": 2.129988431930542, "learning_rate": 5.746656125674329e-05, "loss": 1.003, "step": 52 }, { "epoch": 0.986046511627907, "grad_norm": 0.9616814255714417, "learning_rate": 5.690337755342581e-05, "loss": 1.3023, "step": 53 }, { "epoch": 1.0046511627906978, "grad_norm": 1.7374155521392822, "learning_rate": 5.633071313111564e-05, "loss": 1.3017, "step": 54 }, { "epoch": 1.0232558139534884, "grad_norm": 0.5342385768890381, "learning_rate": 5.574881586347755e-05, "loss": 0.8393, "step": 55 }, { "epoch": 1.041860465116279, "grad_norm": 0.7542828917503357, "learning_rate": 5.5157937620546804e-05, "loss": 0.9745, "step": 56 }, { "epoch": 1.0604651162790697, "grad_norm": 0.7869417071342468, "learning_rate": 5.4558334159709074e-05, "loss": 1.1747, "step": 57 }, { "epoch": 1.0790697674418606, "grad_norm": 0.773287296295166, "learning_rate": 5.3950265014997655e-05, "loss": 0.8018, "step": 58 }, { "epoch": 1.0976744186046512, "grad_norm": 1.2457112073898315, "learning_rate": 5.3333993384756045e-05, "loss": 0.9032, "step": 59 }, { "epoch": 1.1162790697674418, "grad_norm": 1.0725675821304321, "learning_rate": 5.270978601771444e-05, "loss": 0.9193, "step": 60 }, { "epoch": 1.1348837209302325, "grad_norm": 0.763644814491272, "learning_rate": 5.207791309752948e-05, "loss": 0.7066, "step": 61 }, { "epoch": 1.1534883720930234, "grad_norm": 0.8121412396430969, "learning_rate": 5.1438648125837204e-05, "loss": 0.7076, "step": 62 }, { "epoch": 1.172093023255814, "grad_norm": 1.0602554082870483, "learning_rate": 5.07922678038698e-05, "loss": 0.7972, "step": 63 }, { "epoch": 1.1906976744186046, "grad_norm": 1.2864611148834229, "learning_rate": 5.013905191268754e-05, "loss": 0.9987, "step": 64 }, { "epoch": 1.2093023255813953, "grad_norm": 1.507265567779541, "learning_rate": 4.9479283192077487e-05, "loss": 0.9133, "step": 65 }, { "epoch": 1.2279069767441861, "grad_norm": 1.0964022874832153, "learning_rate": 4.881324721817167e-05, "loss": 0.6193, "step": 66 }, { "epoch": 1.2465116279069768, "grad_norm": 1.7399965524673462, "learning_rate": 4.8141232279837554e-05, "loss": 0.8116, "step": 67 }, { "epoch": 1.2651162790697674, "grad_norm": 0.8102202415466309, "learning_rate": 4.746352925389414e-05, "loss": 1.1357, "step": 68 }, { "epoch": 1.283720930232558, "grad_norm": 0.9682488441467285, "learning_rate": 4.678043147920822e-05, "loss": 1.0843, "step": 69 }, { "epoch": 1.302325581395349, "grad_norm": 0.8420053720474243, "learning_rate": 4.60922346297246e-05, "loss": 1.052, "step": 70 }, { "epoch": 1.3209302325581396, "grad_norm": 0.9388038516044617, "learning_rate": 4.539923658648585e-05, "loss": 0.8371, "step": 71 }, { "epoch": 1.3395348837209302, "grad_norm": 0.8845089077949524, "learning_rate": 4.4701737308696665e-05, "loss": 0.7301, "step": 72 }, { "epoch": 1.3581395348837209, "grad_norm": 1.6632778644561768, "learning_rate": 4.4000038703888604e-05, "loss": 1.0172, "step": 73 }, { "epoch": 1.3767441860465115, "grad_norm": 1.184276819229126, "learning_rate": 4.3294444497241656e-05, "loss": 0.8308, "step": 74 }, { "epoch": 1.3953488372093024, "grad_norm": 1.2161909341812134, "learning_rate": 4.2585260100118973e-05, "loss": 0.6478, "step": 75 }, { "epoch": 1.413953488372093, "grad_norm": 1.1866018772125244, "learning_rate": 4.187279247787175e-05, "loss": 0.6407, "step": 76 }, { "epoch": 1.4325581395348836, "grad_norm": 1.264217495918274, "learning_rate": 4.115735001697149e-05, "loss": 0.8542, "step": 77 }, { "epoch": 1.4511627906976745, "grad_norm": 1.6465506553649902, "learning_rate": 4.0439242391527176e-05, "loss": 0.7179, "step": 78 }, { "epoch": 1.4697674418604652, "grad_norm": 1.624822735786438, "learning_rate": 3.971878042924501e-05, "loss": 0.6409, "step": 79 }, { "epoch": 1.4883720930232558, "grad_norm": 1.7168909311294556, "learning_rate": 3.899627597688895e-05, "loss": 0.7917, "step": 80 }, { "epoch": 1.5069767441860464, "grad_norm": 0.8661062121391296, "learning_rate": 3.8272041765299995e-05, "loss": 0.9741, "step": 81 }, { "epoch": 1.525581395348837, "grad_norm": 0.9207056164741516, "learning_rate": 3.754639127403304e-05, "loss": 1.2139, "step": 82 }, { "epoch": 1.544186046511628, "grad_norm": 1.0560969114303589, "learning_rate": 3.6819638595669306e-05, "loss": 0.8169, "step": 83 }, { "epoch": 1.5627906976744186, "grad_norm": 1.0724343061447144, "learning_rate": 3.609209829986387e-05, "loss": 0.9448, "step": 84 }, { "epoch": 1.5813953488372094, "grad_norm": 1.2073761224746704, "learning_rate": 3.536408529718625e-05, "loss": 0.9814, "step": 85 }, { "epoch": 1.6, "grad_norm": 1.2773185968399048, "learning_rate": 3.463591470281375e-05, "loss": 0.7854, "step": 86 }, { "epoch": 1.6186046511627907, "grad_norm": 1.2945756912231445, "learning_rate": 3.390790170013612e-05, "loss": 0.8327, "step": 87 }, { "epoch": 1.6372093023255814, "grad_norm": 1.2657415866851807, "learning_rate": 3.318036140433069e-05, "loss": 0.7456, "step": 88 }, { "epoch": 1.655813953488372, "grad_norm": 1.308043360710144, "learning_rate": 3.245360872596697e-05, "loss": 0.5729, "step": 89 }, { "epoch": 1.6744186046511627, "grad_norm": 1.3613492250442505, "learning_rate": 3.172795823469999e-05, "loss": 0.6906, "step": 90 }, { "epoch": 1.6930232558139535, "grad_norm": 1.2736989259719849, "learning_rate": 3.100372402311106e-05, "loss": 0.3997, "step": 91 }, { "epoch": 1.7116279069767442, "grad_norm": 1.814155101776123, "learning_rate": 3.0281219570754987e-05, "loss": 0.5502, "step": 92 }, { "epoch": 1.730232558139535, "grad_norm": 1.9878065586090088, "learning_rate": 2.956075760847283e-05, "loss": 1.1875, "step": 93 }, { "epoch": 1.7488372093023257, "grad_norm": 0.9894658923149109, "learning_rate": 2.8842649983028517e-05, "loss": 1.081, "step": 94 }, { "epoch": 1.7674418604651163, "grad_norm": 0.9422222375869751, "learning_rate": 2.8127207522128252e-05, "loss": 0.6992, "step": 95 }, { "epoch": 1.786046511627907, "grad_norm": 1.0437465906143188, "learning_rate": 2.7414739899881024e-05, "loss": 0.7245, "step": 96 }, { "epoch": 1.8046511627906976, "grad_norm": 1.2193009853363037, "learning_rate": 2.670555550275834e-05, "loss": 0.7591, "step": 97 }, { "epoch": 1.8232558139534882, "grad_norm": 1.1166754961013794, "learning_rate": 2.5999961296111394e-05, "loss": 0.9399, "step": 98 }, { "epoch": 1.841860465116279, "grad_norm": 1.4289640188217163, "learning_rate": 2.5298262691303332e-05, "loss": 0.9413, "step": 99 }, { "epoch": 1.8604651162790697, "grad_norm": 1.2337830066680908, "learning_rate": 2.4600763413514138e-05, "loss": 0.7486, "step": 100 }, { "epoch": 1.8604651162790697, "eval_loss": 0.28952550888061523, "eval_runtime": 10.8771, "eval_samples_per_second": 8.366, "eval_steps_per_second": 2.115, "step": 100 }, { "epoch": 1.8790697674418606, "grad_norm": 1.4255279302597046, "learning_rate": 2.3907765370275404e-05, "loss": 0.765, "step": 101 }, { "epoch": 1.8976744186046512, "grad_norm": 1.7940795421600342, "learning_rate": 2.3219568520791783e-05, "loss": 0.6342, "step": 102 }, { "epoch": 1.916279069767442, "grad_norm": 1.6805592775344849, "learning_rate": 2.2536470746105853e-05, "loss": 0.7094, "step": 103 }, { "epoch": 1.9348837209302325, "grad_norm": 2.0769453048706055, "learning_rate": 2.1858767720162456e-05, "loss": 0.8721, "step": 104 }, { "epoch": 1.9534883720930232, "grad_norm": 1.9850454330444336, "learning_rate": 2.1186752781828328e-05, "loss": 0.5382, "step": 105 }, { "epoch": 1.9720930232558138, "grad_norm": 1.6767321825027466, "learning_rate": 2.0520716807922514e-05, "loss": 0.5035, "step": 106 }, { "epoch": 1.9906976744186047, "grad_norm": 1.6948860883712769, "learning_rate": 1.986094808731247e-05, "loss": 0.9199, "step": 107 }, { "epoch": 2.0093023255813955, "grad_norm": 1.2903045415878296, "learning_rate": 1.92077321961302e-05, "loss": 0.6539, "step": 108 }, { "epoch": 2.027906976744186, "grad_norm": 0.6793351173400879, "learning_rate": 1.8561351874162783e-05, "loss": 0.6779, "step": 109 }, { "epoch": 2.046511627906977, "grad_norm": 1.0761351585388184, "learning_rate": 1.7922086902470502e-05, "loss": 0.6817, "step": 110 }, { "epoch": 2.0651162790697675, "grad_norm": 0.825526773929596, "learning_rate": 1.7290213982285564e-05, "loss": 0.5877, "step": 111 }, { "epoch": 2.083720930232558, "grad_norm": 0.9566294550895691, "learning_rate": 1.666600661524396e-05, "loss": 0.5537, "step": 112 }, { "epoch": 2.1023255813953488, "grad_norm": 0.9278555512428284, "learning_rate": 1.604973498500235e-05, "loss": 0.6015, "step": 113 }, { "epoch": 2.1209302325581394, "grad_norm": 0.9701972007751465, "learning_rate": 1.5441665840290927e-05, "loss": 0.547, "step": 114 }, { "epoch": 2.13953488372093, "grad_norm": 1.187429666519165, "learning_rate": 1.4842062379453195e-05, "loss": 0.5522, "step": 115 }, { "epoch": 2.158139534883721, "grad_norm": 0.8903448581695557, "learning_rate": 1.4251184136522453e-05, "loss": 0.403, "step": 116 }, { "epoch": 2.1767441860465118, "grad_norm": 0.9742254614830017, "learning_rate": 1.3669286868884356e-05, "loss": 0.279, "step": 117 }, { "epoch": 2.1953488372093024, "grad_norm": 0.9774483442306519, "learning_rate": 1.3096622446574184e-05, "loss": 0.3753, "step": 118 }, { "epoch": 2.213953488372093, "grad_norm": 1.1328277587890625, "learning_rate": 1.2533438743256703e-05, "loss": 0.2929, "step": 119 }, { "epoch": 2.2325581395348837, "grad_norm": 1.241883397102356, "learning_rate": 1.1979979528935865e-05, "loss": 0.3254, "step": 120 }, { "epoch": 2.2511627906976743, "grad_norm": 0.8701224327087402, "learning_rate": 1.1436484364440646e-05, "loss": 0.5831, "step": 121 }, { "epoch": 2.269767441860465, "grad_norm": 0.9288952350616455, "learning_rate": 1.0903188497732999e-05, "loss": 0.8483, "step": 122 }, { "epoch": 2.2883720930232556, "grad_norm": 0.7865915298461914, "learning_rate": 1.038032276208234e-05, "loss": 0.5125, "step": 123 }, { "epoch": 2.3069767441860467, "grad_norm": 1.1094341278076172, "learning_rate": 9.86811347615111e-06, "loss": 0.453, "step": 124 }, { "epoch": 2.3255813953488373, "grad_norm": 1.1128486394882202, "learning_rate": 9.36678234603432e-06, "loss": 0.4629, "step": 125 }, { "epoch": 2.344186046511628, "grad_norm": 1.2837432622909546, "learning_rate": 8.876546369295717e-06, "loss": 0.6856, "step": 126 }, { "epoch": 2.3627906976744186, "grad_norm": 1.148620367050171, "learning_rate": 8.397617741041962e-06, "loss": 0.5077, "step": 127 }, { "epoch": 2.3813953488372093, "grad_norm": 1.2100828886032104, "learning_rate": 7.930203762075426e-06, "loss": 0.4271, "step": 128 }, { "epoch": 2.4, "grad_norm": 1.403386116027832, "learning_rate": 7.47450674916562e-06, "loss": 0.3401, "step": 129 }, { "epoch": 2.4186046511627906, "grad_norm": 1.1033246517181396, "learning_rate": 7.030723947477803e-06, "loss": 0.2728, "step": 130 }, { "epoch": 2.4372093023255816, "grad_norm": 1.4235916137695312, "learning_rate": 6.599047445196836e-06, "loss": 0.3717, "step": 131 }, { "epoch": 2.4558139534883723, "grad_norm": 1.2089860439300537, "learning_rate": 6.1796640903832005e-06, "loss": 0.3064, "step": 132 }, { "epoch": 2.474418604651163, "grad_norm": 1.1210060119628906, "learning_rate": 5.772755410097166e-06, "loss": 0.2141, "step": 133 }, { "epoch": 2.4930232558139536, "grad_norm": 1.133145809173584, "learning_rate": 5.378497531826109e-06, "loss": 0.5761, "step": 134 }, { "epoch": 2.511627906976744, "grad_norm": 1.5591737031936646, "learning_rate": 4.997061107248992e-06, "loss": 0.8019, "step": 135 }, { "epoch": 2.530232558139535, "grad_norm": 1.521730899810791, "learning_rate": 4.6286112383710245e-06, "loss": 0.58, "step": 136 }, { "epoch": 2.5488372093023255, "grad_norm": 1.5007507801055908, "learning_rate": 4.273307406060465e-06, "loss": 0.6187, "step": 137 }, { "epoch": 2.567441860465116, "grad_norm": 1.2669973373413086, "learning_rate": 3.931303401018427e-06, "loss": 0.5699, "step": 138 }, { "epoch": 2.5860465116279068, "grad_norm": 1.4318617582321167, "learning_rate": 3.602747257211758e-06, "loss": 0.5587, "step": 139 }, { "epoch": 2.604651162790698, "grad_norm": 1.6378320455551147, "learning_rate": 3.2877811877975422e-06, "loss": 0.4079, "step": 140 }, { "epoch": 2.6232558139534885, "grad_norm": 1.452219009399414, "learning_rate": 2.9865415235671784e-06, "loss": 0.4239, "step": 141 }, { "epoch": 2.641860465116279, "grad_norm": 1.3756489753723145, "learning_rate": 2.6991586539366018e-06, "loss": 0.4746, "step": 142 }, { "epoch": 2.66046511627907, "grad_norm": 1.2296236753463745, "learning_rate": 2.425756970508135e-06, "loss": 0.3361, "step": 143 }, { "epoch": 2.6790697674418604, "grad_norm": 1.558907389640808, "learning_rate": 2.166454813228505e-06, "loss": 0.4134, "step": 144 }, { "epoch": 2.697674418604651, "grad_norm": 1.4058880805969238, "learning_rate": 1.9213644191662354e-06, "loss": 0.2047, "step": 145 }, { "epoch": 2.7162790697674417, "grad_norm": 1.1016334295272827, "learning_rate": 1.6905918739306423e-06, "loss": 0.149, "step": 146 }, { "epoch": 2.734883720930233, "grad_norm": 1.425627589225769, "learning_rate": 1.4742370657534552e-06, "loss": 0.9796, "step": 147 }, { "epoch": 2.753488372093023, "grad_norm": 1.2110799551010132, "learning_rate": 1.2723936422528996e-06, "loss": 0.6505, "step": 148 }, { "epoch": 2.772093023255814, "grad_norm": 1.069684386253357, "learning_rate": 1.0851489698990146e-06, "loss": 0.6269, "step": 149 }, { "epoch": 2.7906976744186047, "grad_norm": 1.4680191278457642, "learning_rate": 9.125840961976982e-07, "loss": 0.8263, "step": 150 }, { "epoch": 2.7906976744186047, "eval_loss": 0.3216976821422577, "eval_runtime": 10.869, "eval_samples_per_second": 8.372, "eval_steps_per_second": 2.116, "step": 150 }, { "epoch": 2.8093023255813954, "grad_norm": 1.4344302415847778, "learning_rate": 7.547737146098809e-07, "loss": 0.768, "step": 151 }, { "epoch": 2.827906976744186, "grad_norm": 1.6189731359481812, "learning_rate": 6.117861322210028e-07, "loss": 0.5337, "step": 152 }, { "epoch": 2.8465116279069766, "grad_norm": 1.4234650135040283, "learning_rate": 4.836832401747892e-07, "loss": 0.5149, "step": 153 }, { "epoch": 2.8651162790697673, "grad_norm": 1.1778002977371216, "learning_rate": 3.705204868840894e-07, "loss": 0.4272, "step": 154 }, { "epoch": 2.883720930232558, "grad_norm": 1.35899019241333, "learning_rate": 2.7234685403047243e-07, "loss": 0.3179, "step": 155 }, { "epoch": 2.902325581395349, "grad_norm": 1.4125027656555176, "learning_rate": 1.892048353628217e-07, "loss": 0.365, "step": 156 }, { "epoch": 2.9209302325581397, "grad_norm": 1.287215232849121, "learning_rate": 1.211304183042494e-07, "loss": 0.3343, "step": 157 }, { "epoch": 2.9395348837209303, "grad_norm": 1.398272156715393, "learning_rate": 6.81530683751752e-08, "loss": 0.3152, "step": 158 }, { "epoch": 2.958139534883721, "grad_norm": 1.168246865272522, "learning_rate": 3.029571643941353e-08, "loss": 0.1939, "step": 159 }, { "epoch": 2.9767441860465116, "grad_norm": 1.5252883434295654, "learning_rate": 7.574748778697015e-09, "loss": 0.4645, "step": 160 }, { "epoch": 2.9953488372093022, "grad_norm": 1.418809413909912, "learning_rate": 0.0, "loss": 0.3772, "step": 161 } ], "logging_steps": 1, "max_steps": 161, "num_input_tokens_seen": 0, "num_train_epochs": 4, "save_steps": 50, "stateful_callbacks": { "EarlyStoppingCallback": { "args": { "early_stopping_patience": 4, "early_stopping_threshold": 0.0 }, "attributes": { "early_stopping_patience_counter": 1 } }, "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 4.532890163478528e+17, "train_batch_size": 8, "trial_name": null, "trial_params": null }