{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.18814675446848542, "eval_steps": 500, "global_step": 300, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0006271558482282847, "grad_norm": 1.1329874992370605, "learning_rate": 0.0001, "loss": 2.5314, "step": 1 }, { "epoch": 0.0012543116964565694, "grad_norm": 1.0171780586242676, "learning_rate": 9.966555183946489e-05, "loss": 2.291, "step": 2 }, { "epoch": 0.0018814675446848542, "grad_norm": 0.9779753088951111, "learning_rate": 9.933110367892977e-05, "loss": 2.5455, "step": 3 }, { "epoch": 0.002508623392913139, "grad_norm": 1.0696159601211548, "learning_rate": 9.899665551839465e-05, "loss": 2.385, "step": 4 }, { "epoch": 0.0031357792411414237, "grad_norm": 1.1393558979034424, "learning_rate": 9.866220735785953e-05, "loss": 2.22, "step": 5 }, { "epoch": 0.0037629350893697085, "grad_norm": 1.214498519897461, "learning_rate": 9.832775919732441e-05, "loss": 1.8638, "step": 6 }, { "epoch": 0.004390090937597993, "grad_norm": 1.7331534624099731, "learning_rate": 9.799331103678931e-05, "loss": 1.7735, "step": 7 }, { "epoch": 0.005017246785826278, "grad_norm": 1.0437425374984741, "learning_rate": 9.765886287625419e-05, "loss": 1.5449, "step": 8 }, { "epoch": 0.005644402634054563, "grad_norm": 0.9133126139640808, "learning_rate": 9.732441471571907e-05, "loss": 1.5503, "step": 9 }, { "epoch": 0.006271558482282847, "grad_norm": 0.8929418325424194, "learning_rate": 9.698996655518396e-05, "loss": 1.5382, "step": 10 }, { "epoch": 0.006898714330511132, "grad_norm": 0.7974631190299988, "learning_rate": 9.665551839464884e-05, "loss": 1.3685, "step": 11 }, { "epoch": 0.007525870178739417, "grad_norm": 0.6889926791191101, "learning_rate": 9.632107023411372e-05, "loss": 1.4076, "step": 12 }, { "epoch": 0.008153026026967701, "grad_norm": 0.463778555393219, "learning_rate": 9.59866220735786e-05, "loss": 1.3502, "step": 13 }, { "epoch": 0.008780181875195987, "grad_norm": 0.4665314853191376, "learning_rate": 9.565217391304348e-05, "loss": 1.3215, "step": 14 }, { "epoch": 0.00940733772342427, "grad_norm": 0.49210086464881897, "learning_rate": 9.531772575250837e-05, "loss": 1.2521, "step": 15 }, { "epoch": 0.010034493571652555, "grad_norm": 0.5524080395698547, "learning_rate": 9.498327759197325e-05, "loss": 1.3825, "step": 16 }, { "epoch": 0.01066164941988084, "grad_norm": 0.5872707366943359, "learning_rate": 9.464882943143813e-05, "loss": 1.2788, "step": 17 }, { "epoch": 0.011288805268109126, "grad_norm": 0.6367695331573486, "learning_rate": 9.431438127090302e-05, "loss": 1.2561, "step": 18 }, { "epoch": 0.01191596111633741, "grad_norm": 0.5402896404266357, "learning_rate": 9.39799331103679e-05, "loss": 1.3212, "step": 19 }, { "epoch": 0.012543116964565695, "grad_norm": 0.5474902987480164, "learning_rate": 9.364548494983279e-05, "loss": 1.2272, "step": 20 }, { "epoch": 0.01317027281279398, "grad_norm": 0.5391371846199036, "learning_rate": 9.331103678929767e-05, "loss": 1.2131, "step": 21 }, { "epoch": 0.013797428661022263, "grad_norm": 0.6083835959434509, "learning_rate": 9.297658862876255e-05, "loss": 1.2583, "step": 22 }, { "epoch": 0.014424584509250549, "grad_norm": 0.625644326210022, "learning_rate": 9.264214046822743e-05, "loss": 1.2724, "step": 23 }, { "epoch": 0.015051740357478834, "grad_norm": 0.537401020526886, "learning_rate": 9.230769230769232e-05, "loss": 1.2605, "step": 24 }, { "epoch": 0.01567889620570712, "grad_norm": 0.6273273825645447, "learning_rate": 9.19732441471572e-05, "loss": 1.3261, "step": 25 }, { "epoch": 0.016306052053935403, "grad_norm": 0.637491762638092, "learning_rate": 9.163879598662207e-05, "loss": 1.1808, "step": 26 }, { "epoch": 0.016933207902163686, "grad_norm": 0.6607061624526978, "learning_rate": 9.130434782608696e-05, "loss": 1.2826, "step": 27 }, { "epoch": 0.017560363750391973, "grad_norm": 0.7268609404563904, "learning_rate": 9.096989966555184e-05, "loss": 1.2573, "step": 28 }, { "epoch": 0.018187519598620257, "grad_norm": 0.7763082385063171, "learning_rate": 9.063545150501673e-05, "loss": 1.266, "step": 29 }, { "epoch": 0.01881467544684854, "grad_norm": 0.72670978307724, "learning_rate": 9.030100334448161e-05, "loss": 1.1736, "step": 30 }, { "epoch": 0.019441831295076827, "grad_norm": 0.7848783731460571, "learning_rate": 8.996655518394649e-05, "loss": 1.1741, "step": 31 }, { "epoch": 0.02006898714330511, "grad_norm": 0.8250067234039307, "learning_rate": 8.963210702341137e-05, "loss": 1.0689, "step": 32 }, { "epoch": 0.020696142991533398, "grad_norm": 0.8322198390960693, "learning_rate": 8.929765886287625e-05, "loss": 1.1708, "step": 33 }, { "epoch": 0.02132329883976168, "grad_norm": 0.8828574419021606, "learning_rate": 8.896321070234114e-05, "loss": 1.1479, "step": 34 }, { "epoch": 0.021950454687989965, "grad_norm": 0.8523293137550354, "learning_rate": 8.862876254180602e-05, "loss": 1.1957, "step": 35 }, { "epoch": 0.022577610536218252, "grad_norm": 0.8329194784164429, "learning_rate": 8.82943143812709e-05, "loss": 1.2659, "step": 36 }, { "epoch": 0.023204766384446535, "grad_norm": 0.5968635082244873, "learning_rate": 8.795986622073578e-05, "loss": 1.0579, "step": 37 }, { "epoch": 0.02383192223267482, "grad_norm": 0.4709320068359375, "learning_rate": 8.762541806020068e-05, "loss": 1.1495, "step": 38 }, { "epoch": 0.024459078080903106, "grad_norm": 0.4913414418697357, "learning_rate": 8.729096989966556e-05, "loss": 1.2192, "step": 39 }, { "epoch": 0.02508623392913139, "grad_norm": 0.4877021014690399, "learning_rate": 8.695652173913044e-05, "loss": 1.3065, "step": 40 }, { "epoch": 0.025713389777359673, "grad_norm": 0.3754381835460663, "learning_rate": 8.662207357859532e-05, "loss": 1.1311, "step": 41 }, { "epoch": 0.02634054562558796, "grad_norm": 0.4108287990093231, "learning_rate": 8.62876254180602e-05, "loss": 1.2202, "step": 42 }, { "epoch": 0.026967701473816243, "grad_norm": 0.40012553334236145, "learning_rate": 8.595317725752509e-05, "loss": 1.1415, "step": 43 }, { "epoch": 0.027594857322044527, "grad_norm": 0.39889755845069885, "learning_rate": 8.561872909698997e-05, "loss": 1.1504, "step": 44 }, { "epoch": 0.028222013170272814, "grad_norm": 0.46676573157310486, "learning_rate": 8.528428093645485e-05, "loss": 1.1806, "step": 45 }, { "epoch": 0.028849169018501097, "grad_norm": 0.4341486692428589, "learning_rate": 8.494983277591973e-05, "loss": 1.1706, "step": 46 }, { "epoch": 0.02947632486672938, "grad_norm": 0.5124387145042419, "learning_rate": 8.461538461538461e-05, "loss": 1.1873, "step": 47 }, { "epoch": 0.030103480714957668, "grad_norm": 0.40826505422592163, "learning_rate": 8.42809364548495e-05, "loss": 1.0895, "step": 48 }, { "epoch": 0.03073063656318595, "grad_norm": 0.4583950936794281, "learning_rate": 8.394648829431439e-05, "loss": 1.2264, "step": 49 }, { "epoch": 0.03135779241141424, "grad_norm": 0.4081198275089264, "learning_rate": 8.361204013377927e-05, "loss": 1.1446, "step": 50 }, { "epoch": 0.03198494825964252, "grad_norm": 0.41521233320236206, "learning_rate": 8.327759197324416e-05, "loss": 1.1491, "step": 51 }, { "epoch": 0.032612104107870805, "grad_norm": 0.40511155128479004, "learning_rate": 8.294314381270904e-05, "loss": 1.1759, "step": 52 }, { "epoch": 0.03323925995609909, "grad_norm": 0.5162859559059143, "learning_rate": 8.260869565217392e-05, "loss": 1.1583, "step": 53 }, { "epoch": 0.03386641580432737, "grad_norm": 0.465262770652771, "learning_rate": 8.22742474916388e-05, "loss": 1.2034, "step": 54 }, { "epoch": 0.03449357165255566, "grad_norm": 0.42902040481567383, "learning_rate": 8.193979933110368e-05, "loss": 1.172, "step": 55 }, { "epoch": 0.035120727500783946, "grad_norm": 0.42888331413269043, "learning_rate": 8.160535117056857e-05, "loss": 1.082, "step": 56 }, { "epoch": 0.03574788334901223, "grad_norm": 0.49385911226272583, "learning_rate": 8.127090301003345e-05, "loss": 1.1224, "step": 57 }, { "epoch": 0.036375039197240513, "grad_norm": 0.47527334094047546, "learning_rate": 8.093645484949833e-05, "loss": 1.1926, "step": 58 }, { "epoch": 0.0370021950454688, "grad_norm": 0.5191435813903809, "learning_rate": 8.060200668896321e-05, "loss": 1.2253, "step": 59 }, { "epoch": 0.03762935089369708, "grad_norm": 0.48137006163597107, "learning_rate": 8.026755852842809e-05, "loss": 1.17, "step": 60 }, { "epoch": 0.03825650674192537, "grad_norm": 0.44135233759880066, "learning_rate": 7.993311036789299e-05, "loss": 1.1388, "step": 61 }, { "epoch": 0.038883662590153655, "grad_norm": 0.46292147040367126, "learning_rate": 7.959866220735787e-05, "loss": 1.1531, "step": 62 }, { "epoch": 0.03951081843838194, "grad_norm": 0.4255847930908203, "learning_rate": 7.926421404682275e-05, "loss": 1.1398, "step": 63 }, { "epoch": 0.04013797428661022, "grad_norm": 0.4323138892650604, "learning_rate": 7.892976588628763e-05, "loss": 1.1635, "step": 64 }, { "epoch": 0.040765130134838505, "grad_norm": 0.4316229522228241, "learning_rate": 7.859531772575252e-05, "loss": 1.1673, "step": 65 }, { "epoch": 0.041392285983066796, "grad_norm": 0.42433494329452515, "learning_rate": 7.82608695652174e-05, "loss": 1.085, "step": 66 }, { "epoch": 0.04201944183129508, "grad_norm": 0.42199933528900146, "learning_rate": 7.792642140468228e-05, "loss": 1.1098, "step": 67 }, { "epoch": 0.04264659767952336, "grad_norm": 0.4090598523616791, "learning_rate": 7.759197324414716e-05, "loss": 1.0722, "step": 68 }, { "epoch": 0.043273753527751646, "grad_norm": 0.46552446484565735, "learning_rate": 7.725752508361204e-05, "loss": 1.1798, "step": 69 }, { "epoch": 0.04390090937597993, "grad_norm": 0.4066402018070221, "learning_rate": 7.692307692307693e-05, "loss": 1.0683, "step": 70 }, { "epoch": 0.04452806522420821, "grad_norm": 0.46165624260902405, "learning_rate": 7.658862876254181e-05, "loss": 1.0879, "step": 71 }, { "epoch": 0.045155221072436504, "grad_norm": 0.4765890836715698, "learning_rate": 7.62541806020067e-05, "loss": 1.17, "step": 72 }, { "epoch": 0.04578237692066479, "grad_norm": 0.46825796365737915, "learning_rate": 7.591973244147159e-05, "loss": 1.1548, "step": 73 }, { "epoch": 0.04640953276889307, "grad_norm": 0.42665427923202515, "learning_rate": 7.558528428093647e-05, "loss": 1.2517, "step": 74 }, { "epoch": 0.047036688617121354, "grad_norm": 0.4694182574748993, "learning_rate": 7.525083612040135e-05, "loss": 1.1806, "step": 75 }, { "epoch": 0.04766384446534964, "grad_norm": 0.45235225558280945, "learning_rate": 7.491638795986622e-05, "loss": 1.1515, "step": 76 }, { "epoch": 0.04829100031357792, "grad_norm": 0.4057731628417969, "learning_rate": 7.45819397993311e-05, "loss": 1.0757, "step": 77 }, { "epoch": 0.04891815616180621, "grad_norm": 0.44630780816078186, "learning_rate": 7.424749163879598e-05, "loss": 1.1323, "step": 78 }, { "epoch": 0.049545312010034495, "grad_norm": 0.41524800658226013, "learning_rate": 7.391304347826086e-05, "loss": 1.167, "step": 79 }, { "epoch": 0.05017246785826278, "grad_norm": 0.46300187706947327, "learning_rate": 7.357859531772575e-05, "loss": 1.1519, "step": 80 }, { "epoch": 0.05079962370649106, "grad_norm": 0.4147876799106598, "learning_rate": 7.324414715719064e-05, "loss": 1.1566, "step": 81 }, { "epoch": 0.051426779554719346, "grad_norm": 0.44954177737236023, "learning_rate": 7.290969899665552e-05, "loss": 1.1737, "step": 82 }, { "epoch": 0.05205393540294763, "grad_norm": 0.45056983828544617, "learning_rate": 7.25752508361204e-05, "loss": 1.1597, "step": 83 }, { "epoch": 0.05268109125117592, "grad_norm": 0.5045411586761475, "learning_rate": 7.224080267558529e-05, "loss": 1.1294, "step": 84 }, { "epoch": 0.0533082470994042, "grad_norm": 0.5283576846122742, "learning_rate": 7.190635451505017e-05, "loss": 1.2078, "step": 85 }, { "epoch": 0.05393540294763249, "grad_norm": 0.434832900762558, "learning_rate": 7.157190635451505e-05, "loss": 1.1519, "step": 86 }, { "epoch": 0.05456255879586077, "grad_norm": 0.4940570294857025, "learning_rate": 7.123745819397993e-05, "loss": 1.2029, "step": 87 }, { "epoch": 0.055189714644089054, "grad_norm": 0.47406020760536194, "learning_rate": 7.090301003344481e-05, "loss": 1.095, "step": 88 }, { "epoch": 0.055816870492317344, "grad_norm": 0.46511971950531006, "learning_rate": 7.05685618729097e-05, "loss": 1.1762, "step": 89 }, { "epoch": 0.05644402634054563, "grad_norm": 0.47649258375167847, "learning_rate": 7.023411371237458e-05, "loss": 1.1546, "step": 90 }, { "epoch": 0.05707118218877391, "grad_norm": 0.4064464867115021, "learning_rate": 6.989966555183946e-05, "loss": 1.0969, "step": 91 }, { "epoch": 0.057698338037002195, "grad_norm": 0.4659188389778137, "learning_rate": 6.956521739130436e-05, "loss": 1.1868, "step": 92 }, { "epoch": 0.05832549388523048, "grad_norm": 0.4586854577064514, "learning_rate": 6.923076923076924e-05, "loss": 1.1075, "step": 93 }, { "epoch": 0.05895264973345876, "grad_norm": 0.4805111289024353, "learning_rate": 6.889632107023412e-05, "loss": 1.1154, "step": 94 }, { "epoch": 0.05957980558168705, "grad_norm": 0.38674503564834595, "learning_rate": 6.8561872909699e-05, "loss": 1.0882, "step": 95 }, { "epoch": 0.060206961429915336, "grad_norm": 0.4246669411659241, "learning_rate": 6.822742474916388e-05, "loss": 1.0373, "step": 96 }, { "epoch": 0.06083411727814362, "grad_norm": 0.4012146592140198, "learning_rate": 6.789297658862876e-05, "loss": 1.1813, "step": 97 }, { "epoch": 0.0614612731263719, "grad_norm": 0.4579727053642273, "learning_rate": 6.755852842809365e-05, "loss": 1.1539, "step": 98 }, { "epoch": 0.062088428974600186, "grad_norm": 0.44659194350242615, "learning_rate": 6.722408026755853e-05, "loss": 1.104, "step": 99 }, { "epoch": 0.06271558482282848, "grad_norm": 0.43122145533561707, "learning_rate": 6.688963210702341e-05, "loss": 1.1289, "step": 100 }, { "epoch": 0.06334274067105676, "grad_norm": 0.40788164734840393, "learning_rate": 6.655518394648829e-05, "loss": 1.1314, "step": 101 }, { "epoch": 0.06396989651928504, "grad_norm": 0.45048585534095764, "learning_rate": 6.622073578595317e-05, "loss": 1.0983, "step": 102 }, { "epoch": 0.06459705236751333, "grad_norm": 0.4325787425041199, "learning_rate": 6.588628762541807e-05, "loss": 1.0949, "step": 103 }, { "epoch": 0.06522420821574161, "grad_norm": 0.5192950367927551, "learning_rate": 6.555183946488295e-05, "loss": 1.2203, "step": 104 }, { "epoch": 0.0658513640639699, "grad_norm": 0.46830520033836365, "learning_rate": 6.521739130434783e-05, "loss": 1.1678, "step": 105 }, { "epoch": 0.06647851991219818, "grad_norm": 0.4136311709880829, "learning_rate": 6.488294314381272e-05, "loss": 1.0903, "step": 106 }, { "epoch": 0.06710567576042646, "grad_norm": 0.5483879446983337, "learning_rate": 6.45484949832776e-05, "loss": 1.0997, "step": 107 }, { "epoch": 0.06773283160865474, "grad_norm": 0.4444606602191925, "learning_rate": 6.421404682274248e-05, "loss": 1.0104, "step": 108 }, { "epoch": 0.06835998745688304, "grad_norm": 0.41186246275901794, "learning_rate": 6.387959866220736e-05, "loss": 1.0611, "step": 109 }, { "epoch": 0.06898714330511133, "grad_norm": 0.47986000776290894, "learning_rate": 6.354515050167224e-05, "loss": 1.1197, "step": 110 }, { "epoch": 0.06961429915333961, "grad_norm": 0.4698418080806732, "learning_rate": 6.321070234113713e-05, "loss": 1.2814, "step": 111 }, { "epoch": 0.07024145500156789, "grad_norm": 0.4690100848674774, "learning_rate": 6.287625418060201e-05, "loss": 1.1208, "step": 112 }, { "epoch": 0.07086861084979618, "grad_norm": 0.5585668683052063, "learning_rate": 6.254180602006689e-05, "loss": 1.1289, "step": 113 }, { "epoch": 0.07149576669802446, "grad_norm": 0.45809587836265564, "learning_rate": 6.220735785953178e-05, "loss": 1.1284, "step": 114 }, { "epoch": 0.07212292254625274, "grad_norm": 0.48366600275039673, "learning_rate": 6.187290969899667e-05, "loss": 1.145, "step": 115 }, { "epoch": 0.07275007839448103, "grad_norm": 0.4505147337913513, "learning_rate": 6.153846153846155e-05, "loss": 1.1071, "step": 116 }, { "epoch": 0.07337723424270931, "grad_norm": 0.5526301264762878, "learning_rate": 6.120401337792643e-05, "loss": 1.2021, "step": 117 }, { "epoch": 0.0740043900909376, "grad_norm": 0.44302597641944885, "learning_rate": 6.086956521739131e-05, "loss": 1.0973, "step": 118 }, { "epoch": 0.07463154593916588, "grad_norm": 0.38735851645469666, "learning_rate": 6.0535117056856194e-05, "loss": 0.9937, "step": 119 }, { "epoch": 0.07525870178739416, "grad_norm": 0.45709213614463806, "learning_rate": 6.0200668896321076e-05, "loss": 1.0586, "step": 120 }, { "epoch": 0.07588585763562246, "grad_norm": 0.4670064151287079, "learning_rate": 5.986622073578596e-05, "loss": 1.1423, "step": 121 }, { "epoch": 0.07651301348385074, "grad_norm": 0.4870319664478302, "learning_rate": 5.953177257525085e-05, "loss": 1.149, "step": 122 }, { "epoch": 0.07714016933207903, "grad_norm": 0.43948817253112793, "learning_rate": 5.919732441471573e-05, "loss": 1.1102, "step": 123 }, { "epoch": 0.07776732518030731, "grad_norm": 0.5276496410369873, "learning_rate": 5.886287625418061e-05, "loss": 1.179, "step": 124 }, { "epoch": 0.07839448102853559, "grad_norm": 0.456938773393631, "learning_rate": 5.852842809364549e-05, "loss": 1.0614, "step": 125 }, { "epoch": 0.07902163687676388, "grad_norm": 0.48240378499031067, "learning_rate": 5.819397993311037e-05, "loss": 1.1562, "step": 126 }, { "epoch": 0.07964879272499216, "grad_norm": 0.48209676146507263, "learning_rate": 5.785953177257525e-05, "loss": 1.0598, "step": 127 }, { "epoch": 0.08027594857322044, "grad_norm": 0.5053116679191589, "learning_rate": 5.752508361204013e-05, "loss": 1.2358, "step": 128 }, { "epoch": 0.08090310442144873, "grad_norm": 0.48982295393943787, "learning_rate": 5.7190635451505014e-05, "loss": 1.1236, "step": 129 }, { "epoch": 0.08153026026967701, "grad_norm": 0.473768025636673, "learning_rate": 5.6856187290969896e-05, "loss": 1.1223, "step": 130 }, { "epoch": 0.0821574161179053, "grad_norm": 0.47272294759750366, "learning_rate": 5.652173913043478e-05, "loss": 1.1766, "step": 131 }, { "epoch": 0.08278457196613359, "grad_norm": 0.42775315046310425, "learning_rate": 5.6187290969899666e-05, "loss": 1.0777, "step": 132 }, { "epoch": 0.08341172781436187, "grad_norm": 0.43032482266426086, "learning_rate": 5.585284280936455e-05, "loss": 1.0509, "step": 133 }, { "epoch": 0.08403888366259016, "grad_norm": 0.4302811324596405, "learning_rate": 5.551839464882943e-05, "loss": 1.1031, "step": 134 }, { "epoch": 0.08466603951081844, "grad_norm": 0.48152413964271545, "learning_rate": 5.518394648829431e-05, "loss": 1.0831, "step": 135 }, { "epoch": 0.08529319535904673, "grad_norm": 0.44327837228775024, "learning_rate": 5.4849498327759194e-05, "loss": 1.0824, "step": 136 }, { "epoch": 0.08592035120727501, "grad_norm": 0.4980611801147461, "learning_rate": 5.451505016722408e-05, "loss": 1.0317, "step": 137 }, { "epoch": 0.08654750705550329, "grad_norm": 0.5285961031913757, "learning_rate": 5.4180602006688965e-05, "loss": 1.0671, "step": 138 }, { "epoch": 0.08717466290373158, "grad_norm": 0.47221964597702026, "learning_rate": 5.384615384615385e-05, "loss": 1.1313, "step": 139 }, { "epoch": 0.08780181875195986, "grad_norm": 0.4523843228816986, "learning_rate": 5.351170568561873e-05, "loss": 1.1363, "step": 140 }, { "epoch": 0.08842897460018814, "grad_norm": 0.44657716155052185, "learning_rate": 5.317725752508361e-05, "loss": 1.0926, "step": 141 }, { "epoch": 0.08905613044841643, "grad_norm": 0.4858958423137665, "learning_rate": 5.284280936454849e-05, "loss": 1.142, "step": 142 }, { "epoch": 0.08968328629664471, "grad_norm": 0.5070914626121521, "learning_rate": 5.250836120401338e-05, "loss": 1.1487, "step": 143 }, { "epoch": 0.09031044214487301, "grad_norm": 0.5086584687232971, "learning_rate": 5.217391304347826e-05, "loss": 1.1316, "step": 144 }, { "epoch": 0.09093759799310129, "grad_norm": 0.5099964737892151, "learning_rate": 5.1839464882943145e-05, "loss": 1.1523, "step": 145 }, { "epoch": 0.09156475384132957, "grad_norm": 0.42420920729637146, "learning_rate": 5.150501672240803e-05, "loss": 1.0709, "step": 146 }, { "epoch": 0.09219190968955786, "grad_norm": 0.4656766355037689, "learning_rate": 5.117056856187291e-05, "loss": 1.0626, "step": 147 }, { "epoch": 0.09281906553778614, "grad_norm": 0.44983309507369995, "learning_rate": 5.08361204013378e-05, "loss": 1.1076, "step": 148 }, { "epoch": 0.09344622138601442, "grad_norm": 0.49652299284935, "learning_rate": 5.050167224080268e-05, "loss": 1.158, "step": 149 }, { "epoch": 0.09407337723424271, "grad_norm": 0.5079907774925232, "learning_rate": 5.016722408026756e-05, "loss": 1.1208, "step": 150 }, { "epoch": 0.09470053308247099, "grad_norm": 0.5696375966072083, "learning_rate": 4.983277591973244e-05, "loss": 1.1515, "step": 151 }, { "epoch": 0.09532768893069928, "grad_norm": 0.44712352752685547, "learning_rate": 4.9498327759197325e-05, "loss": 1.045, "step": 152 }, { "epoch": 0.09595484477892756, "grad_norm": 0.41328972578048706, "learning_rate": 4.916387959866221e-05, "loss": 1.0785, "step": 153 }, { "epoch": 0.09658200062715584, "grad_norm": 0.4514102637767792, "learning_rate": 4.8829431438127096e-05, "loss": 1.0989, "step": 154 }, { "epoch": 0.09720915647538414, "grad_norm": 0.5154950618743896, "learning_rate": 4.849498327759198e-05, "loss": 1.1652, "step": 155 }, { "epoch": 0.09783631232361242, "grad_norm": 0.4378693997859955, "learning_rate": 4.816053511705686e-05, "loss": 1.1153, "step": 156 }, { "epoch": 0.0984634681718407, "grad_norm": 0.5264151096343994, "learning_rate": 4.782608695652174e-05, "loss": 1.1305, "step": 157 }, { "epoch": 0.09909062402006899, "grad_norm": 0.5044124126434326, "learning_rate": 4.7491638795986624e-05, "loss": 1.1603, "step": 158 }, { "epoch": 0.09971777986829727, "grad_norm": 0.5369151830673218, "learning_rate": 4.715719063545151e-05, "loss": 1.0011, "step": 159 }, { "epoch": 0.10034493571652556, "grad_norm": 0.4663112163543701, "learning_rate": 4.6822742474916394e-05, "loss": 1.1319, "step": 160 }, { "epoch": 0.10097209156475384, "grad_norm": 0.497859388589859, "learning_rate": 4.6488294314381276e-05, "loss": 1.117, "step": 161 }, { "epoch": 0.10159924741298212, "grad_norm": 0.5183404684066772, "learning_rate": 4.615384615384616e-05, "loss": 1.1978, "step": 162 }, { "epoch": 0.10222640326121041, "grad_norm": 0.45794522762298584, "learning_rate": 4.581939799331103e-05, "loss": 1.1036, "step": 163 }, { "epoch": 0.10285355910943869, "grad_norm": 0.5128252506256104, "learning_rate": 4.548494983277592e-05, "loss": 1.1711, "step": 164 }, { "epoch": 0.10348071495766697, "grad_norm": 0.481505423784256, "learning_rate": 4.5150501672240804e-05, "loss": 1.1143, "step": 165 }, { "epoch": 0.10410787080589526, "grad_norm": 0.48832452297210693, "learning_rate": 4.4816053511705686e-05, "loss": 1.1723, "step": 166 }, { "epoch": 0.10473502665412356, "grad_norm": 0.4828014373779297, "learning_rate": 4.448160535117057e-05, "loss": 1.1016, "step": 167 }, { "epoch": 0.10536218250235184, "grad_norm": 0.5178865790367126, "learning_rate": 4.414715719063545e-05, "loss": 1.1412, "step": 168 }, { "epoch": 0.10598933835058012, "grad_norm": 0.4863995313644409, "learning_rate": 4.381270903010034e-05, "loss": 1.1386, "step": 169 }, { "epoch": 0.1066164941988084, "grad_norm": 0.48305442929267883, "learning_rate": 4.347826086956522e-05, "loss": 1.1376, "step": 170 }, { "epoch": 0.10724365004703669, "grad_norm": 0.46624070405960083, "learning_rate": 4.31438127090301e-05, "loss": 1.1082, "step": 171 }, { "epoch": 0.10787080589526497, "grad_norm": 0.49723443388938904, "learning_rate": 4.2809364548494984e-05, "loss": 1.087, "step": 172 }, { "epoch": 0.10849796174349326, "grad_norm": 0.47998717427253723, "learning_rate": 4.2474916387959866e-05, "loss": 1.1228, "step": 173 }, { "epoch": 0.10912511759172154, "grad_norm": 0.4895256459712982, "learning_rate": 4.214046822742475e-05, "loss": 1.1387, "step": 174 }, { "epoch": 0.10975227343994982, "grad_norm": 0.4830639660358429, "learning_rate": 4.180602006688964e-05, "loss": 1.1239, "step": 175 }, { "epoch": 0.11037942928817811, "grad_norm": 0.43242347240448, "learning_rate": 4.147157190635452e-05, "loss": 1.0437, "step": 176 }, { "epoch": 0.11100658513640639, "grad_norm": 0.4875499904155731, "learning_rate": 4.11371237458194e-05, "loss": 1.1805, "step": 177 }, { "epoch": 0.11163374098463469, "grad_norm": 0.5745497941970825, "learning_rate": 4.080267558528428e-05, "loss": 1.1453, "step": 178 }, { "epoch": 0.11226089683286297, "grad_norm": 0.5793723464012146, "learning_rate": 4.0468227424749165e-05, "loss": 1.1364, "step": 179 }, { "epoch": 0.11288805268109126, "grad_norm": 0.48804497718811035, "learning_rate": 4.0133779264214046e-05, "loss": 1.0968, "step": 180 }, { "epoch": 0.11351520852931954, "grad_norm": 0.46328264474868774, "learning_rate": 3.9799331103678935e-05, "loss": 1.0965, "step": 181 }, { "epoch": 0.11414236437754782, "grad_norm": 0.4533807039260864, "learning_rate": 3.946488294314382e-05, "loss": 1.0968, "step": 182 }, { "epoch": 0.1147695202257761, "grad_norm": 0.4990158975124359, "learning_rate": 3.91304347826087e-05, "loss": 1.133, "step": 183 }, { "epoch": 0.11539667607400439, "grad_norm": 0.5183166265487671, "learning_rate": 3.879598662207358e-05, "loss": 1.1555, "step": 184 }, { "epoch": 0.11602383192223267, "grad_norm": 0.5148091912269592, "learning_rate": 3.846153846153846e-05, "loss": 1.0681, "step": 185 }, { "epoch": 0.11665098777046096, "grad_norm": 0.4919118285179138, "learning_rate": 3.812709030100335e-05, "loss": 1.184, "step": 186 }, { "epoch": 0.11727814361868924, "grad_norm": 0.5206012725830078, "learning_rate": 3.7792642140468233e-05, "loss": 1.1401, "step": 187 }, { "epoch": 0.11790529946691752, "grad_norm": 0.4834771156311035, "learning_rate": 3.745819397993311e-05, "loss": 1.088, "step": 188 }, { "epoch": 0.11853245531514581, "grad_norm": 0.48845043778419495, "learning_rate": 3.712374581939799e-05, "loss": 1.1958, "step": 189 }, { "epoch": 0.1191596111633741, "grad_norm": 0.49845704436302185, "learning_rate": 3.678929765886287e-05, "loss": 1.1188, "step": 190 }, { "epoch": 0.11978676701160239, "grad_norm": 0.4738442003726959, "learning_rate": 3.645484949832776e-05, "loss": 1.153, "step": 191 }, { "epoch": 0.12041392285983067, "grad_norm": 0.4959152042865753, "learning_rate": 3.612040133779264e-05, "loss": 1.125, "step": 192 }, { "epoch": 0.12104107870805896, "grad_norm": 0.5036417245864868, "learning_rate": 3.5785953177257525e-05, "loss": 1.1273, "step": 193 }, { "epoch": 0.12166823455628724, "grad_norm": 0.5091941952705383, "learning_rate": 3.545150501672241e-05, "loss": 1.0774, "step": 194 }, { "epoch": 0.12229539040451552, "grad_norm": 0.4901737570762634, "learning_rate": 3.511705685618729e-05, "loss": 1.1179, "step": 195 }, { "epoch": 0.1229225462527438, "grad_norm": 0.45219141244888306, "learning_rate": 3.478260869565218e-05, "loss": 1.1085, "step": 196 }, { "epoch": 0.12354970210097209, "grad_norm": 0.4779893159866333, "learning_rate": 3.444816053511706e-05, "loss": 1.093, "step": 197 }, { "epoch": 0.12417685794920037, "grad_norm": 0.4967818260192871, "learning_rate": 3.411371237458194e-05, "loss": 1.1042, "step": 198 }, { "epoch": 0.12480401379742866, "grad_norm": 0.5056282877922058, "learning_rate": 3.3779264214046823e-05, "loss": 1.1222, "step": 199 }, { "epoch": 0.12543116964565695, "grad_norm": 0.4826442003250122, "learning_rate": 3.3444816053511705e-05, "loss": 1.0909, "step": 200 }, { "epoch": 0.12605832549388524, "grad_norm": 0.4724636673927307, "learning_rate": 3.311036789297659e-05, "loss": 1.0632, "step": 201 }, { "epoch": 0.12668548134211352, "grad_norm": 0.48456090688705444, "learning_rate": 3.2775919732441476e-05, "loss": 1.1662, "step": 202 }, { "epoch": 0.1273126371903418, "grad_norm": 0.43576645851135254, "learning_rate": 3.244147157190636e-05, "loss": 1.0106, "step": 203 }, { "epoch": 0.1279397930385701, "grad_norm": 0.5297894477844238, "learning_rate": 3.210702341137124e-05, "loss": 1.1668, "step": 204 }, { "epoch": 0.12856694888679837, "grad_norm": 0.5179861783981323, "learning_rate": 3.177257525083612e-05, "loss": 1.0566, "step": 205 }, { "epoch": 0.12919410473502665, "grad_norm": 0.46285516023635864, "learning_rate": 3.1438127090301004e-05, "loss": 1.1157, "step": 206 }, { "epoch": 0.12982126058325494, "grad_norm": 0.511254072189331, "learning_rate": 3.110367892976589e-05, "loss": 1.1425, "step": 207 }, { "epoch": 0.13044841643148322, "grad_norm": 0.5081919431686401, "learning_rate": 3.0769230769230774e-05, "loss": 1.0436, "step": 208 }, { "epoch": 0.1310755722797115, "grad_norm": 0.4743541479110718, "learning_rate": 3.0434782608695656e-05, "loss": 1.0811, "step": 209 }, { "epoch": 0.1317027281279398, "grad_norm": 0.5982806086540222, "learning_rate": 3.0100334448160538e-05, "loss": 1.1674, "step": 210 }, { "epoch": 0.13232988397616807, "grad_norm": 0.47165796160697937, "learning_rate": 2.9765886287625424e-05, "loss": 1.1274, "step": 211 }, { "epoch": 0.13295703982439636, "grad_norm": 0.5538997054100037, "learning_rate": 2.9431438127090305e-05, "loss": 1.1679, "step": 212 }, { "epoch": 0.13358419567262464, "grad_norm": 0.5390966534614563, "learning_rate": 2.9096989966555184e-05, "loss": 1.1613, "step": 213 }, { "epoch": 0.13421135152085292, "grad_norm": 0.6614457368850708, "learning_rate": 2.8762541806020066e-05, "loss": 1.2044, "step": 214 }, { "epoch": 0.1348385073690812, "grad_norm": 0.4833295941352844, "learning_rate": 2.8428093645484948e-05, "loss": 1.1787, "step": 215 }, { "epoch": 0.1354656632173095, "grad_norm": 0.4769502282142639, "learning_rate": 2.8093645484949833e-05, "loss": 1.1303, "step": 216 }, { "epoch": 0.13609281906553777, "grad_norm": 0.5252038240432739, "learning_rate": 2.7759197324414715e-05, "loss": 1.1373, "step": 217 }, { "epoch": 0.13671997491376608, "grad_norm": 0.5152972936630249, "learning_rate": 2.7424749163879597e-05, "loss": 1.1386, "step": 218 }, { "epoch": 0.13734713076199437, "grad_norm": 0.45123931765556335, "learning_rate": 2.7090301003344482e-05, "loss": 1.0077, "step": 219 }, { "epoch": 0.13797428661022265, "grad_norm": 0.45800599455833435, "learning_rate": 2.6755852842809364e-05, "loss": 1.1186, "step": 220 }, { "epoch": 0.13860144245845094, "grad_norm": 0.4793985188007355, "learning_rate": 2.6421404682274246e-05, "loss": 1.1138, "step": 221 }, { "epoch": 0.13922859830667922, "grad_norm": 0.5087202787399292, "learning_rate": 2.608695652173913e-05, "loss": 1.0913, "step": 222 }, { "epoch": 0.1398557541549075, "grad_norm": 0.519314169883728, "learning_rate": 2.5752508361204013e-05, "loss": 1.1132, "step": 223 }, { "epoch": 0.14048291000313579, "grad_norm": 0.4722985625267029, "learning_rate": 2.54180602006689e-05, "loss": 1.1185, "step": 224 }, { "epoch": 0.14111006585136407, "grad_norm": 0.5009178519248962, "learning_rate": 2.508361204013378e-05, "loss": 1.1063, "step": 225 }, { "epoch": 0.14173722169959235, "grad_norm": 0.5109993815422058, "learning_rate": 2.4749163879598663e-05, "loss": 1.1522, "step": 226 }, { "epoch": 0.14236437754782064, "grad_norm": 0.46906983852386475, "learning_rate": 2.4414715719063548e-05, "loss": 1.0815, "step": 227 }, { "epoch": 0.14299153339604892, "grad_norm": 0.48216167092323303, "learning_rate": 2.408026755852843e-05, "loss": 1.1103, "step": 228 }, { "epoch": 0.1436186892442772, "grad_norm": 0.4617651402950287, "learning_rate": 2.3745819397993312e-05, "loss": 1.0704, "step": 229 }, { "epoch": 0.1442458450925055, "grad_norm": 0.5504018664360046, "learning_rate": 2.3411371237458197e-05, "loss": 1.1774, "step": 230 }, { "epoch": 0.14487300094073377, "grad_norm": 0.48636704683303833, "learning_rate": 2.307692307692308e-05, "loss": 1.1166, "step": 231 }, { "epoch": 0.14550015678896205, "grad_norm": 0.46996790170669556, "learning_rate": 2.274247491638796e-05, "loss": 1.1043, "step": 232 }, { "epoch": 0.14612731263719034, "grad_norm": 0.4818289577960968, "learning_rate": 2.2408026755852843e-05, "loss": 1.0713, "step": 233 }, { "epoch": 0.14675446848541862, "grad_norm": 0.48282650113105774, "learning_rate": 2.2073578595317725e-05, "loss": 1.0679, "step": 234 }, { "epoch": 0.1473816243336469, "grad_norm": 0.4748537242412567, "learning_rate": 2.173913043478261e-05, "loss": 1.0467, "step": 235 }, { "epoch": 0.1480087801818752, "grad_norm": 0.4510209560394287, "learning_rate": 2.1404682274247492e-05, "loss": 1.0525, "step": 236 }, { "epoch": 0.14863593603010347, "grad_norm": 0.5060710310935974, "learning_rate": 2.1070234113712374e-05, "loss": 1.1351, "step": 237 }, { "epoch": 0.14926309187833176, "grad_norm": 0.5170108675956726, "learning_rate": 2.073578595317726e-05, "loss": 1.1409, "step": 238 }, { "epoch": 0.14989024772656004, "grad_norm": 0.5064777731895447, "learning_rate": 2.040133779264214e-05, "loss": 1.0937, "step": 239 }, { "epoch": 0.15051740357478832, "grad_norm": 0.5006868243217468, "learning_rate": 2.0066889632107023e-05, "loss": 1.0439, "step": 240 }, { "epoch": 0.15114455942301663, "grad_norm": 0.44411641359329224, "learning_rate": 1.973244147157191e-05, "loss": 0.9983, "step": 241 }, { "epoch": 0.15177171527124492, "grad_norm": 0.4972614645957947, "learning_rate": 1.939799331103679e-05, "loss": 1.1376, "step": 242 }, { "epoch": 0.1523988711194732, "grad_norm": 0.5033297538757324, "learning_rate": 1.9063545150501676e-05, "loss": 1.1652, "step": 243 }, { "epoch": 0.15302602696770148, "grad_norm": 0.4621533751487732, "learning_rate": 1.8729096989966554e-05, "loss": 1.0376, "step": 244 }, { "epoch": 0.15365318281592977, "grad_norm": 0.4779740273952484, "learning_rate": 1.8394648829431436e-05, "loss": 1.0468, "step": 245 }, { "epoch": 0.15428033866415805, "grad_norm": 0.4619259536266327, "learning_rate": 1.806020066889632e-05, "loss": 1.0634, "step": 246 }, { "epoch": 0.15490749451238633, "grad_norm": 0.5598598718643188, "learning_rate": 1.7725752508361204e-05, "loss": 1.2436, "step": 247 }, { "epoch": 0.15553465036061462, "grad_norm": 0.4703880250453949, "learning_rate": 1.739130434782609e-05, "loss": 1.1322, "step": 248 }, { "epoch": 0.1561618062088429, "grad_norm": 0.4648728668689728, "learning_rate": 1.705685618729097e-05, "loss": 1.0827, "step": 249 }, { "epoch": 0.15678896205707119, "grad_norm": 0.5145319700241089, "learning_rate": 1.6722408026755853e-05, "loss": 1.1761, "step": 250 }, { "epoch": 0.15741611790529947, "grad_norm": 0.5132664442062378, "learning_rate": 1.6387959866220738e-05, "loss": 1.0624, "step": 251 }, { "epoch": 0.15804327375352775, "grad_norm": 0.4508794844150543, "learning_rate": 1.605351170568562e-05, "loss": 1.0546, "step": 252 }, { "epoch": 0.15867042960175604, "grad_norm": 0.4499465823173523, "learning_rate": 1.5719063545150502e-05, "loss": 1.0157, "step": 253 }, { "epoch": 0.15929758544998432, "grad_norm": 0.458778977394104, "learning_rate": 1.5384615384615387e-05, "loss": 1.0634, "step": 254 }, { "epoch": 0.1599247412982126, "grad_norm": 0.5598222017288208, "learning_rate": 1.5050167224080269e-05, "loss": 1.1146, "step": 255 }, { "epoch": 0.1605518971464409, "grad_norm": 0.5084378123283386, "learning_rate": 1.4715719063545153e-05, "loss": 1.075, "step": 256 }, { "epoch": 0.16117905299466917, "grad_norm": 0.4847070276737213, "learning_rate": 1.4381270903010033e-05, "loss": 1.0363, "step": 257 }, { "epoch": 0.16180620884289745, "grad_norm": 0.4656808376312256, "learning_rate": 1.4046822742474917e-05, "loss": 1.0491, "step": 258 }, { "epoch": 0.16243336469112574, "grad_norm": 0.5068191289901733, "learning_rate": 1.3712374581939799e-05, "loss": 1.0642, "step": 259 }, { "epoch": 0.16306052053935402, "grad_norm": 0.48283588886260986, "learning_rate": 1.3377926421404682e-05, "loss": 1.1085, "step": 260 }, { "epoch": 0.1636876763875823, "grad_norm": 0.5512163043022156, "learning_rate": 1.3043478260869566e-05, "loss": 1.1055, "step": 261 }, { "epoch": 0.1643148322358106, "grad_norm": 0.5653620362281799, "learning_rate": 1.270903010033445e-05, "loss": 1.0846, "step": 262 }, { "epoch": 0.16494198808403887, "grad_norm": 0.48929548263549805, "learning_rate": 1.2374581939799331e-05, "loss": 1.0273, "step": 263 }, { "epoch": 0.16556914393226718, "grad_norm": 0.5095831155776978, "learning_rate": 1.2040133779264215e-05, "loss": 1.1798, "step": 264 }, { "epoch": 0.16619629978049547, "grad_norm": 0.5254250168800354, "learning_rate": 1.1705685618729099e-05, "loss": 1.0542, "step": 265 }, { "epoch": 0.16682345562872375, "grad_norm": 0.505216658115387, "learning_rate": 1.137123745819398e-05, "loss": 1.1344, "step": 266 }, { "epoch": 0.16745061147695203, "grad_norm": 0.5194329023361206, "learning_rate": 1.1036789297658862e-05, "loss": 1.0841, "step": 267 }, { "epoch": 0.16807776732518032, "grad_norm": 0.5093269944190979, "learning_rate": 1.0702341137123746e-05, "loss": 1.1222, "step": 268 }, { "epoch": 0.1687049231734086, "grad_norm": 0.5413921475410461, "learning_rate": 1.036789297658863e-05, "loss": 1.063, "step": 269 }, { "epoch": 0.16933207902163688, "grad_norm": 0.49521705508232117, "learning_rate": 1.0033444816053512e-05, "loss": 1.0666, "step": 270 }, { "epoch": 0.16995923486986517, "grad_norm": 0.479844868183136, "learning_rate": 9.698996655518395e-06, "loss": 1.0231, "step": 271 }, { "epoch": 0.17058639071809345, "grad_norm": 0.5199151635169983, "learning_rate": 9.364548494983277e-06, "loss": 1.0276, "step": 272 }, { "epoch": 0.17121354656632173, "grad_norm": 0.48625025153160095, "learning_rate": 9.03010033444816e-06, "loss": 1.0848, "step": 273 }, { "epoch": 0.17184070241455002, "grad_norm": 0.4806617498397827, "learning_rate": 8.695652173913044e-06, "loss": 1.1002, "step": 274 }, { "epoch": 0.1724678582627783, "grad_norm": 0.4978601336479187, "learning_rate": 8.361204013377926e-06, "loss": 1.1587, "step": 275 }, { "epoch": 0.17309501411100658, "grad_norm": 0.4522283375263214, "learning_rate": 8.02675585284281e-06, "loss": 1.079, "step": 276 }, { "epoch": 0.17372216995923487, "grad_norm": 0.5176008939743042, "learning_rate": 7.692307692307694e-06, "loss": 1.137, "step": 277 }, { "epoch": 0.17434932580746315, "grad_norm": 0.5522940158843994, "learning_rate": 7.357859531772576e-06, "loss": 1.0736, "step": 278 }, { "epoch": 0.17497648165569143, "grad_norm": 0.5581282377243042, "learning_rate": 7.023411371237458e-06, "loss": 1.1654, "step": 279 }, { "epoch": 0.17560363750391972, "grad_norm": 0.48599082231521606, "learning_rate": 6.688963210702341e-06, "loss": 1.103, "step": 280 }, { "epoch": 0.176230793352148, "grad_norm": 0.43440407514572144, "learning_rate": 6.354515050167225e-06, "loss": 0.9973, "step": 281 }, { "epoch": 0.17685794920037629, "grad_norm": 0.5466330051422119, "learning_rate": 6.0200668896321075e-06, "loss": 1.0821, "step": 282 }, { "epoch": 0.17748510504860457, "grad_norm": 0.5091434717178345, "learning_rate": 5.68561872909699e-06, "loss": 1.0652, "step": 283 }, { "epoch": 0.17811226089683285, "grad_norm": 0.5003758072853088, "learning_rate": 5.351170568561873e-06, "loss": 1.0895, "step": 284 }, { "epoch": 0.17873941674506114, "grad_norm": 0.47549617290496826, "learning_rate": 5.016722408026756e-06, "loss": 1.1288, "step": 285 }, { "epoch": 0.17936657259328942, "grad_norm": 0.6036182641983032, "learning_rate": 4.682274247491639e-06, "loss": 1.1673, "step": 286 }, { "epoch": 0.17999372844151773, "grad_norm": 0.4893932044506073, "learning_rate": 4.347826086956522e-06, "loss": 1.1532, "step": 287 }, { "epoch": 0.18062088428974601, "grad_norm": 0.5435315370559692, "learning_rate": 4.013377926421405e-06, "loss": 1.0154, "step": 288 }, { "epoch": 0.1812480401379743, "grad_norm": 0.4814181625843048, "learning_rate": 3.678929765886288e-06, "loss": 1.0401, "step": 289 }, { "epoch": 0.18187519598620258, "grad_norm": 0.4699884057044983, "learning_rate": 3.3444816053511705e-06, "loss": 1.0851, "step": 290 }, { "epoch": 0.18250235183443086, "grad_norm": 0.5204551815986633, "learning_rate": 3.0100334448160537e-06, "loss": 1.0937, "step": 291 }, { "epoch": 0.18312950768265915, "grad_norm": 0.506284773349762, "learning_rate": 2.6755852842809365e-06, "loss": 1.1464, "step": 292 }, { "epoch": 0.18375666353088743, "grad_norm": 0.4346812963485718, "learning_rate": 2.3411371237458193e-06, "loss": 1.054, "step": 293 }, { "epoch": 0.18438381937911572, "grad_norm": 0.5776488780975342, "learning_rate": 2.0066889632107025e-06, "loss": 1.1026, "step": 294 }, { "epoch": 0.185010975227344, "grad_norm": 0.49304401874542236, "learning_rate": 1.6722408026755853e-06, "loss": 1.0577, "step": 295 }, { "epoch": 0.18563813107557228, "grad_norm": 0.5319218039512634, "learning_rate": 1.3377926421404683e-06, "loss": 1.0623, "step": 296 }, { "epoch": 0.18626528692380057, "grad_norm": 0.45602694153785706, "learning_rate": 1.0033444816053512e-06, "loss": 1.0911, "step": 297 }, { "epoch": 0.18689244277202885, "grad_norm": 0.5336869955062866, "learning_rate": 6.688963210702341e-07, "loss": 1.111, "step": 298 }, { "epoch": 0.18751959862025713, "grad_norm": 0.4785906970500946, "learning_rate": 3.3444816053511706e-07, "loss": 1.0679, "step": 299 }, { "epoch": 0.18814675446848542, "grad_norm": 0.5171682834625244, "learning_rate": 0.0, "loss": 1.1548, "step": 300 } ], "logging_steps": 1, "max_steps": 300, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 3.181837914803405e+16, "train_batch_size": 4, "trial_name": null, "trial_params": null }