{ "best_metric": null, "best_model_checkpoint": null, "epoch": 4.0, "eval_steps": 2000, "global_step": 21900, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0091324200913242, "grad_norm": 141.86129760742188, "learning_rate": 9.982648401826485e-06, "loss": 6.156, "step": 50 }, { "epoch": 0.0182648401826484, "grad_norm": 83.82975769042969, "learning_rate": 9.959817351598175e-06, "loss": 4.1566, "step": 100 }, { "epoch": 0.0273972602739726, "grad_norm": 171.21966552734375, "learning_rate": 9.936986301369864e-06, "loss": 3.1539, "step": 150 }, { "epoch": 0.0365296803652968, "grad_norm": 86.14761352539062, "learning_rate": 9.914155251141553e-06, "loss": 2.7091, "step": 200 }, { "epoch": 0.045662100456621, "grad_norm": 155.69285583496094, "learning_rate": 9.891324200913242e-06, "loss": 2.3896, "step": 250 }, { "epoch": 0.0547945205479452, "grad_norm": 3496.547607421875, "learning_rate": 9.868493150684932e-06, "loss": 2.186, "step": 300 }, { "epoch": 0.0639269406392694, "grad_norm": 53.159969329833984, "learning_rate": 9.845662100456623e-06, "loss": 2.3572, "step": 350 }, { "epoch": 0.0730593607305936, "grad_norm": 3373.076416015625, "learning_rate": 9.82283105022831e-06, "loss": 2.2758, "step": 400 }, { "epoch": 0.0821917808219178, "grad_norm": 105.64806365966797, "learning_rate": 9.800000000000001e-06, "loss": 2.0534, "step": 450 }, { "epoch": 0.091324200913242, "grad_norm": 77.49466705322266, "learning_rate": 9.77716894977169e-06, "loss": 2.0835, "step": 500 }, { "epoch": 0.1004566210045662, "grad_norm": 111.51605224609375, "learning_rate": 9.75433789954338e-06, "loss": 1.9298, "step": 550 }, { "epoch": 0.1095890410958904, "grad_norm": 43.85201644897461, "learning_rate": 9.731506849315069e-06, "loss": 1.9758, "step": 600 }, { "epoch": 0.1187214611872146, "grad_norm": 55.28047180175781, "learning_rate": 9.708675799086758e-06, "loss": 1.9304, "step": 650 }, { "epoch": 0.1278538812785388, "grad_norm": 50.79043197631836, "learning_rate": 9.685844748858449e-06, "loss": 1.9333, "step": 700 }, { "epoch": 0.136986301369863, "grad_norm": 26.703508377075195, "learning_rate": 9.663013698630138e-06, "loss": 1.817, "step": 750 }, { "epoch": 0.1461187214611872, "grad_norm": 36.239036560058594, "learning_rate": 9.640182648401828e-06, "loss": 1.6908, "step": 800 }, { "epoch": 0.1552511415525114, "grad_norm": 207.3531494140625, "learning_rate": 9.617351598173517e-06, "loss": 1.727, "step": 850 }, { "epoch": 0.1643835616438356, "grad_norm": 49.595481872558594, "learning_rate": 9.594520547945206e-06, "loss": 1.6568, "step": 900 }, { "epoch": 0.1735159817351598, "grad_norm": 49.645843505859375, "learning_rate": 9.571689497716895e-06, "loss": 1.6459, "step": 950 }, { "epoch": 0.182648401826484, "grad_norm": 28.5211181640625, "learning_rate": 9.548858447488585e-06, "loss": 1.5953, "step": 1000 }, { "epoch": 0.1917808219178082, "grad_norm": 43.102718353271484, "learning_rate": 9.526027397260274e-06, "loss": 1.507, "step": 1050 }, { "epoch": 0.2009132420091324, "grad_norm": 26.59748077392578, "learning_rate": 9.503196347031965e-06, "loss": 1.4658, "step": 1100 }, { "epoch": 0.2100456621004566, "grad_norm": 33.241085052490234, "learning_rate": 9.480365296803654e-06, "loss": 1.5042, "step": 1150 }, { "epoch": 0.2191780821917808, "grad_norm": 31.159038543701172, "learning_rate": 9.457534246575343e-06, "loss": 1.4255, "step": 1200 }, { "epoch": 0.228310502283105, "grad_norm": 45.05030059814453, "learning_rate": 9.434703196347033e-06, "loss": 1.4393, "step": 1250 }, { "epoch": 0.2374429223744292, "grad_norm": 25.268295288085938, "learning_rate": 9.411872146118722e-06, "loss": 1.4258, "step": 1300 }, { "epoch": 0.2465753424657534, "grad_norm": 24.656484603881836, "learning_rate": 9.389041095890413e-06, "loss": 1.3565, "step": 1350 }, { "epoch": 0.2557077625570776, "grad_norm": 23.01006507873535, "learning_rate": 9.3662100456621e-06, "loss": 1.3265, "step": 1400 }, { "epoch": 0.2648401826484018, "grad_norm": 30.002086639404297, "learning_rate": 9.343378995433791e-06, "loss": 1.3825, "step": 1450 }, { "epoch": 0.273972602739726, "grad_norm": 27.9675350189209, "learning_rate": 9.32054794520548e-06, "loss": 1.305, "step": 1500 }, { "epoch": 0.2831050228310502, "grad_norm": 40.30059814453125, "learning_rate": 9.29771689497717e-06, "loss": 1.3627, "step": 1550 }, { "epoch": 0.2922374429223744, "grad_norm": 29.09018898010254, "learning_rate": 9.274885844748859e-06, "loss": 1.3124, "step": 1600 }, { "epoch": 0.3013698630136986, "grad_norm": 49.747859954833984, "learning_rate": 9.252054794520548e-06, "loss": 1.3085, "step": 1650 }, { "epoch": 0.3105022831050228, "grad_norm": 34.45816421508789, "learning_rate": 9.229223744292238e-06, "loss": 1.2565, "step": 1700 }, { "epoch": 0.319634703196347, "grad_norm": 25.871952056884766, "learning_rate": 9.206392694063929e-06, "loss": 1.3172, "step": 1750 }, { "epoch": 0.3287671232876712, "grad_norm": 40.712486267089844, "learning_rate": 9.183561643835616e-06, "loss": 1.3525, "step": 1800 }, { "epoch": 0.3378995433789954, "grad_norm": 41.31215286254883, "learning_rate": 9.160730593607307e-06, "loss": 1.3379, "step": 1850 }, { "epoch": 0.3470319634703196, "grad_norm": 34.724308013916016, "learning_rate": 9.137899543378996e-06, "loss": 1.3613, "step": 1900 }, { "epoch": 0.3561643835616438, "grad_norm": 32.3390007019043, "learning_rate": 9.115068493150686e-06, "loss": 1.2672, "step": 1950 }, { "epoch": 0.365296803652968, "grad_norm": 41.678157806396484, "learning_rate": 9.092237442922375e-06, "loss": 1.2615, "step": 2000 }, { "epoch": 0.365296803652968, "eval_exact_match": 71.52317880794702, "eval_f1": 80.78902012347558, "eval_runtime": 396.1868, "eval_samples_per_second": 26.679, "eval_steps_per_second": 1.668, "step": 2000 }, { "epoch": 0.3744292237442922, "grad_norm": 25.548471450805664, "learning_rate": 9.069406392694064e-06, "loss": 1.2814, "step": 2050 }, { "epoch": 0.3835616438356164, "grad_norm": 50.78987503051758, "learning_rate": 9.046575342465755e-06, "loss": 1.258, "step": 2100 }, { "epoch": 0.3926940639269406, "grad_norm": 27.325096130371094, "learning_rate": 9.023744292237444e-06, "loss": 1.2259, "step": 2150 }, { "epoch": 0.4018264840182648, "grad_norm": 45.96411895751953, "learning_rate": 9.000913242009132e-06, "loss": 1.1546, "step": 2200 }, { "epoch": 0.410958904109589, "grad_norm": 32.445091247558594, "learning_rate": 8.978082191780823e-06, "loss": 1.2579, "step": 2250 }, { "epoch": 0.4200913242009132, "grad_norm": 23.029979705810547, "learning_rate": 8.955251141552512e-06, "loss": 1.169, "step": 2300 }, { "epoch": 0.4292237442922374, "grad_norm": 30.386211395263672, "learning_rate": 8.932420091324201e-06, "loss": 1.1398, "step": 2350 }, { "epoch": 0.4383561643835616, "grad_norm": 17.771953582763672, "learning_rate": 8.90958904109589e-06, "loss": 1.1934, "step": 2400 }, { "epoch": 0.4474885844748858, "grad_norm": 12.856189727783203, "learning_rate": 8.88675799086758e-06, "loss": 1.1591, "step": 2450 }, { "epoch": 0.45662100456621, "grad_norm": 29.3809871673584, "learning_rate": 8.863926940639271e-06, "loss": 1.199, "step": 2500 }, { "epoch": 0.4657534246575342, "grad_norm": 14.796148300170898, "learning_rate": 8.84109589041096e-06, "loss": 1.1566, "step": 2550 }, { "epoch": 0.4748858447488584, "grad_norm": 32.102691650390625, "learning_rate": 8.81826484018265e-06, "loss": 1.2326, "step": 2600 }, { "epoch": 0.4840182648401826, "grad_norm": 27.475114822387695, "learning_rate": 8.795433789954339e-06, "loss": 1.1362, "step": 2650 }, { "epoch": 0.4931506849315068, "grad_norm": 20.31199073791504, "learning_rate": 8.772602739726028e-06, "loss": 1.1284, "step": 2700 }, { "epoch": 0.502283105022831, "grad_norm": 34.34742736816406, "learning_rate": 8.749771689497719e-06, "loss": 1.1765, "step": 2750 }, { "epoch": 0.5114155251141552, "grad_norm": 38.820980072021484, "learning_rate": 8.726940639269406e-06, "loss": 1.2815, "step": 2800 }, { "epoch": 0.5205479452054794, "grad_norm": 39.47126388549805, "learning_rate": 8.704109589041096e-06, "loss": 1.2214, "step": 2850 }, { "epoch": 0.5296803652968036, "grad_norm": 23.623933792114258, "learning_rate": 8.681278538812787e-06, "loss": 1.2233, "step": 2900 }, { "epoch": 0.5388127853881278, "grad_norm": 28.29753875732422, "learning_rate": 8.658447488584476e-06, "loss": 1.2184, "step": 2950 }, { "epoch": 0.547945205479452, "grad_norm": 11.690893173217773, "learning_rate": 8.635616438356165e-06, "loss": 1.2078, "step": 3000 }, { "epoch": 0.5570776255707762, "grad_norm": 22.935726165771484, "learning_rate": 8.612785388127854e-06, "loss": 1.1662, "step": 3050 }, { "epoch": 0.5662100456621004, "grad_norm": 41.53331756591797, "learning_rate": 8.589954337899544e-06, "loss": 1.1703, "step": 3100 }, { "epoch": 0.5753424657534246, "grad_norm": 39.44691467285156, "learning_rate": 8.567123287671235e-06, "loss": 1.0978, "step": 3150 }, { "epoch": 0.5844748858447488, "grad_norm": 24.385053634643555, "learning_rate": 8.544292237442922e-06, "loss": 1.164, "step": 3200 }, { "epoch": 0.593607305936073, "grad_norm": 25.12101173400879, "learning_rate": 8.521461187214613e-06, "loss": 1.0858, "step": 3250 }, { "epoch": 0.6027397260273972, "grad_norm": 26.9351806640625, "learning_rate": 8.498630136986302e-06, "loss": 1.119, "step": 3300 }, { "epoch": 0.6118721461187214, "grad_norm": 17.235322952270508, "learning_rate": 8.475799086757992e-06, "loss": 1.1505, "step": 3350 }, { "epoch": 0.6210045662100456, "grad_norm": 27.588298797607422, "learning_rate": 8.452968036529681e-06, "loss": 1.0935, "step": 3400 }, { "epoch": 0.6301369863013698, "grad_norm": 27.88182258605957, "learning_rate": 8.43013698630137e-06, "loss": 1.138, "step": 3450 }, { "epoch": 0.639269406392694, "grad_norm": 38.748836517333984, "learning_rate": 8.40730593607306e-06, "loss": 1.0731, "step": 3500 }, { "epoch": 0.6484018264840182, "grad_norm": 38.5189094543457, "learning_rate": 8.38447488584475e-06, "loss": 1.0474, "step": 3550 }, { "epoch": 0.6575342465753424, "grad_norm": 33.10224151611328, "learning_rate": 8.361643835616438e-06, "loss": 0.8802, "step": 3600 }, { "epoch": 0.6666666666666666, "grad_norm": 45.91608810424805, "learning_rate": 8.338812785388129e-06, "loss": 1.0351, "step": 3650 }, { "epoch": 0.6757990867579908, "grad_norm": 59.66465759277344, "learning_rate": 8.315981735159818e-06, "loss": 1.0275, "step": 3700 }, { "epoch": 0.684931506849315, "grad_norm": 20.09067726135254, "learning_rate": 8.293150684931507e-06, "loss": 1.0122, "step": 3750 }, { "epoch": 0.6940639269406392, "grad_norm": 17.678356170654297, "learning_rate": 8.270319634703197e-06, "loss": 1.0117, "step": 3800 }, { "epoch": 0.7031963470319634, "grad_norm": 15.248215675354004, "learning_rate": 8.247488584474886e-06, "loss": 1.0043, "step": 3850 }, { "epoch": 0.7123287671232876, "grad_norm": 15.972944259643555, "learning_rate": 8.224657534246577e-06, "loss": 1.075, "step": 3900 }, { "epoch": 0.7214611872146118, "grad_norm": 35.40599822998047, "learning_rate": 8.201826484018266e-06, "loss": 1.0646, "step": 3950 }, { "epoch": 0.730593607305936, "grad_norm": 50.43526840209961, "learning_rate": 8.178995433789955e-06, "loss": 1.04, "step": 4000 }, { "epoch": 0.730593607305936, "eval_exact_match": 78.41059602649007, "eval_f1": 86.58121239815173, "eval_runtime": 396.4641, "eval_samples_per_second": 26.661, "eval_steps_per_second": 1.667, "step": 4000 }, { "epoch": 0.7397260273972602, "grad_norm": 38.34847640991211, "learning_rate": 8.156164383561645e-06, "loss": 1.0143, "step": 4050 }, { "epoch": 0.7488584474885844, "grad_norm": 24.275487899780273, "learning_rate": 8.133333333333334e-06, "loss": 1.0638, "step": 4100 }, { "epoch": 0.7579908675799086, "grad_norm": 57.73759078979492, "learning_rate": 8.110502283105023e-06, "loss": 1.0726, "step": 4150 }, { "epoch": 0.7671232876712328, "grad_norm": 26.460296630859375, "learning_rate": 8.087671232876712e-06, "loss": 0.9874, "step": 4200 }, { "epoch": 0.776255707762557, "grad_norm": 34.39996337890625, "learning_rate": 8.064840182648402e-06, "loss": 1.0369, "step": 4250 }, { "epoch": 0.7853881278538812, "grad_norm": 31.487375259399414, "learning_rate": 8.042009132420093e-06, "loss": 1.033, "step": 4300 }, { "epoch": 0.7945205479452054, "grad_norm": 26.804845809936523, "learning_rate": 8.019178082191782e-06, "loss": 1.0823, "step": 4350 }, { "epoch": 0.8036529680365296, "grad_norm": 17.534334182739258, "learning_rate": 7.996347031963471e-06, "loss": 1.0338, "step": 4400 }, { "epoch": 0.8127853881278538, "grad_norm": 16.264385223388672, "learning_rate": 7.97351598173516e-06, "loss": 0.9379, "step": 4450 }, { "epoch": 0.821917808219178, "grad_norm": 15.467246055603027, "learning_rate": 7.95068493150685e-06, "loss": 0.9996, "step": 4500 }, { "epoch": 0.8310502283105022, "grad_norm": 38.00102615356445, "learning_rate": 7.92785388127854e-06, "loss": 0.9505, "step": 4550 }, { "epoch": 0.8401826484018264, "grad_norm": 25.31998634338379, "learning_rate": 7.905022831050228e-06, "loss": 1.0487, "step": 4600 }, { "epoch": 0.8493150684931506, "grad_norm": 14.750944137573242, "learning_rate": 7.88219178082192e-06, "loss": 1.0197, "step": 4650 }, { "epoch": 0.8584474885844748, "grad_norm": 34.27394485473633, "learning_rate": 7.859360730593608e-06, "loss": 0.9946, "step": 4700 }, { "epoch": 0.867579908675799, "grad_norm": 48.27435302734375, "learning_rate": 7.836529680365298e-06, "loss": 0.9748, "step": 4750 }, { "epoch": 0.8767123287671232, "grad_norm": 26.16417121887207, "learning_rate": 7.813698630136987e-06, "loss": 0.9385, "step": 4800 }, { "epoch": 0.8858447488584474, "grad_norm": 28.09895896911621, "learning_rate": 7.790867579908676e-06, "loss": 0.9434, "step": 4850 }, { "epoch": 0.8949771689497716, "grad_norm": 31.583642959594727, "learning_rate": 7.768036529680366e-06, "loss": 0.9019, "step": 4900 }, { "epoch": 0.9041095890410958, "grad_norm": 11.045271873474121, "learning_rate": 7.745205479452056e-06, "loss": 0.9086, "step": 4950 }, { "epoch": 0.91324200913242, "grad_norm": 33.99599838256836, "learning_rate": 7.722374429223744e-06, "loss": 0.9362, "step": 5000 }, { "epoch": 0.9223744292237442, "grad_norm": 20.473243713378906, "learning_rate": 7.699543378995435e-06, "loss": 0.9518, "step": 5050 }, { "epoch": 0.9315068493150684, "grad_norm": 22.298221588134766, "learning_rate": 7.676712328767124e-06, "loss": 0.8972, "step": 5100 }, { "epoch": 0.9406392694063926, "grad_norm": 23.762950897216797, "learning_rate": 7.653881278538813e-06, "loss": 0.9735, "step": 5150 }, { "epoch": 0.9497716894977168, "grad_norm": 31.01708221435547, "learning_rate": 7.631050228310503e-06, "loss": 0.9931, "step": 5200 }, { "epoch": 0.958904109589041, "grad_norm": 27.428834915161133, "learning_rate": 7.608219178082192e-06, "loss": 0.9892, "step": 5250 }, { "epoch": 0.9680365296803652, "grad_norm": 23.573469161987305, "learning_rate": 7.585388127853882e-06, "loss": 0.9576, "step": 5300 }, { "epoch": 0.9771689497716894, "grad_norm": 30.324687957763672, "learning_rate": 7.562557077625572e-06, "loss": 0.9869, "step": 5350 }, { "epoch": 0.9863013698630136, "grad_norm": 35.9589958190918, "learning_rate": 7.539726027397261e-06, "loss": 0.9965, "step": 5400 }, { "epoch": 0.9954337899543378, "grad_norm": 13.001126289367676, "learning_rate": 7.516894977168951e-06, "loss": 0.9539, "step": 5450 }, { "epoch": 1.004566210045662, "grad_norm": 10.314200401306152, "learning_rate": 7.49406392694064e-06, "loss": 0.8357, "step": 5500 }, { "epoch": 1.0136986301369864, "grad_norm": 22.927391052246094, "learning_rate": 7.47123287671233e-06, "loss": 0.7973, "step": 5550 }, { "epoch": 1.0228310502283104, "grad_norm": 18.42430305480957, "learning_rate": 7.4484018264840185e-06, "loss": 0.7766, "step": 5600 }, { "epoch": 1.0319634703196348, "grad_norm": 35.02809524536133, "learning_rate": 7.425570776255709e-06, "loss": 0.8057, "step": 5650 }, { "epoch": 1.0410958904109588, "grad_norm": 9.175677299499512, "learning_rate": 7.402739726027398e-06, "loss": 0.7897, "step": 5700 }, { "epoch": 1.0502283105022832, "grad_norm": 72.41110229492188, "learning_rate": 7.379908675799088e-06, "loss": 0.766, "step": 5750 }, { "epoch": 1.0593607305936072, "grad_norm": 25.297029495239258, "learning_rate": 7.357077625570776e-06, "loss": 0.6862, "step": 5800 }, { "epoch": 1.0684931506849316, "grad_norm": 17.325456619262695, "learning_rate": 7.3342465753424665e-06, "loss": 0.7479, "step": 5850 }, { "epoch": 1.0776255707762556, "grad_norm": 21.178491592407227, "learning_rate": 7.311415525114156e-06, "loss": 0.715, "step": 5900 }, { "epoch": 1.08675799086758, "grad_norm": 30.497148513793945, "learning_rate": 7.288584474885846e-06, "loss": 0.7737, "step": 5950 }, { "epoch": 1.095890410958904, "grad_norm": 30.200098037719727, "learning_rate": 7.265753424657534e-06, "loss": 0.7385, "step": 6000 }, { "epoch": 1.095890410958904, "eval_exact_match": 80.99337748344371, "eval_f1": 88.52122863303049, "eval_runtime": 395.8625, "eval_samples_per_second": 26.701, "eval_steps_per_second": 1.67, "step": 6000 }, { "epoch": 1.1050228310502284, "grad_norm": 13.1331205368042, "learning_rate": 7.242922374429224e-06, "loss": 0.7753, "step": 6050 }, { "epoch": 1.1141552511415524, "grad_norm": 17.219268798828125, "learning_rate": 7.220091324200914e-06, "loss": 0.8186, "step": 6100 }, { "epoch": 1.1232876712328768, "grad_norm": 21.086429595947266, "learning_rate": 7.197260273972604e-06, "loss": 0.823, "step": 6150 }, { "epoch": 1.1324200913242009, "grad_norm": 36.774166107177734, "learning_rate": 7.174429223744292e-06, "loss": 0.8048, "step": 6200 }, { "epoch": 1.1415525114155252, "grad_norm": 14.021933555603027, "learning_rate": 7.151598173515982e-06, "loss": 0.771, "step": 6250 }, { "epoch": 1.1506849315068493, "grad_norm": 16.90314292907715, "learning_rate": 7.128767123287672e-06, "loss": 0.6957, "step": 6300 }, { "epoch": 1.1598173515981736, "grad_norm": 29.586990356445312, "learning_rate": 7.105936073059362e-06, "loss": 0.7806, "step": 6350 }, { "epoch": 1.1689497716894977, "grad_norm": 13.954606056213379, "learning_rate": 7.083105022831051e-06, "loss": 0.7764, "step": 6400 }, { "epoch": 1.178082191780822, "grad_norm": 27.69614601135254, "learning_rate": 7.06027397260274e-06, "loss": 0.7411, "step": 6450 }, { "epoch": 1.187214611872146, "grad_norm": 14.492104530334473, "learning_rate": 7.03744292237443e-06, "loss": 0.821, "step": 6500 }, { "epoch": 1.1963470319634704, "grad_norm": 27.99992561340332, "learning_rate": 7.0146118721461195e-06, "loss": 0.7502, "step": 6550 }, { "epoch": 1.2054794520547945, "grad_norm": 26.75726890563965, "learning_rate": 6.991780821917809e-06, "loss": 0.8265, "step": 6600 }, { "epoch": 1.2146118721461188, "grad_norm": 20.587421417236328, "learning_rate": 6.968949771689498e-06, "loss": 0.8594, "step": 6650 }, { "epoch": 1.2237442922374429, "grad_norm": 15.334819793701172, "learning_rate": 6.946118721461188e-06, "loss": 0.8162, "step": 6700 }, { "epoch": 1.2328767123287672, "grad_norm": 18.709548950195312, "learning_rate": 6.9232876712328774e-06, "loss": 0.7414, "step": 6750 }, { "epoch": 1.2420091324200913, "grad_norm": 23.877904891967773, "learning_rate": 6.900456621004567e-06, "loss": 0.7779, "step": 6800 }, { "epoch": 1.2511415525114156, "grad_norm": 15.881481170654297, "learning_rate": 6.877625570776256e-06, "loss": 0.7659, "step": 6850 }, { "epoch": 1.2602739726027397, "grad_norm": 16.046165466308594, "learning_rate": 6.854794520547946e-06, "loss": 0.7246, "step": 6900 }, { "epoch": 1.269406392694064, "grad_norm": 19.089651107788086, "learning_rate": 6.831963470319636e-06, "loss": 0.7905, "step": 6950 }, { "epoch": 1.278538812785388, "grad_norm": 23.612180709838867, "learning_rate": 6.809132420091325e-06, "loss": 0.671, "step": 7000 }, { "epoch": 1.2876712328767124, "grad_norm": 33.00151443481445, "learning_rate": 6.786301369863014e-06, "loss": 0.8364, "step": 7050 }, { "epoch": 1.2968036529680365, "grad_norm": 21.602069854736328, "learning_rate": 6.763470319634704e-06, "loss": 0.7281, "step": 7100 }, { "epoch": 1.3059360730593608, "grad_norm": 9.662558555603027, "learning_rate": 6.740639269406394e-06, "loss": 0.8133, "step": 7150 }, { "epoch": 1.3150684931506849, "grad_norm": 101.09040069580078, "learning_rate": 6.7178082191780825e-06, "loss": 0.7267, "step": 7200 }, { "epoch": 1.3242009132420092, "grad_norm": 29.580852508544922, "learning_rate": 6.6949771689497726e-06, "loss": 0.7759, "step": 7250 }, { "epoch": 1.3333333333333333, "grad_norm": 16.52943992614746, "learning_rate": 6.672146118721462e-06, "loss": 0.74, "step": 7300 }, { "epoch": 1.3424657534246576, "grad_norm": 26.262174606323242, "learning_rate": 6.649315068493152e-06, "loss": 0.8074, "step": 7350 }, { "epoch": 1.3515981735159817, "grad_norm": 28.12957191467285, "learning_rate": 6.62648401826484e-06, "loss": 0.7235, "step": 7400 }, { "epoch": 1.360730593607306, "grad_norm": 29.391780853271484, "learning_rate": 6.6036529680365305e-06, "loss": 0.7643, "step": 7450 }, { "epoch": 1.36986301369863, "grad_norm": 15.40013313293457, "learning_rate": 6.58082191780822e-06, "loss": 0.7499, "step": 7500 }, { "epoch": 1.3789954337899544, "grad_norm": 10.625802993774414, "learning_rate": 6.557990867579909e-06, "loss": 0.8129, "step": 7550 }, { "epoch": 1.3881278538812785, "grad_norm": 21.512968063354492, "learning_rate": 6.535159817351598e-06, "loss": 0.7053, "step": 7600 }, { "epoch": 1.3972602739726028, "grad_norm": 43.220603942871094, "learning_rate": 6.512328767123288e-06, "loss": 0.7325, "step": 7650 }, { "epoch": 1.4063926940639269, "grad_norm": 20.33064842224121, "learning_rate": 6.489497716894978e-06, "loss": 0.6695, "step": 7700 }, { "epoch": 1.4155251141552512, "grad_norm": 12.011942863464355, "learning_rate": 6.466666666666667e-06, "loss": 0.7438, "step": 7750 }, { "epoch": 1.4246575342465753, "grad_norm": 17.656845092773438, "learning_rate": 6.443835616438356e-06, "loss": 0.7761, "step": 7800 }, { "epoch": 1.4337899543378996, "grad_norm": 19.555429458618164, "learning_rate": 6.421004566210046e-06, "loss": 0.7815, "step": 7850 }, { "epoch": 1.4429223744292237, "grad_norm": 15.19387149810791, "learning_rate": 6.398173515981736e-06, "loss": 0.7897, "step": 7900 }, { "epoch": 1.452054794520548, "grad_norm": 48.02432632446289, "learning_rate": 6.375342465753425e-06, "loss": 0.7725, "step": 7950 }, { "epoch": 1.461187214611872, "grad_norm": 18.34738540649414, "learning_rate": 6.352511415525114e-06, "loss": 0.8278, "step": 8000 }, { "epoch": 1.461187214611872, "eval_exact_match": 81.57994323557237, "eval_f1": 88.68935584305143, "eval_runtime": 395.554, "eval_samples_per_second": 26.722, "eval_steps_per_second": 1.671, "step": 8000 }, { "epoch": 1.4703196347031964, "grad_norm": 13.668383598327637, "learning_rate": 6.329680365296804e-06, "loss": 0.7063, "step": 8050 }, { "epoch": 1.4794520547945205, "grad_norm": 21.045486450195312, "learning_rate": 6.306849315068494e-06, "loss": 0.7451, "step": 8100 }, { "epoch": 1.4885844748858448, "grad_norm": 24.719881057739258, "learning_rate": 6.284018264840183e-06, "loss": 0.7129, "step": 8150 }, { "epoch": 1.4977168949771689, "grad_norm": 8.752976417541504, "learning_rate": 6.261187214611873e-06, "loss": 0.7768, "step": 8200 }, { "epoch": 1.5068493150684932, "grad_norm": 12.758523941040039, "learning_rate": 6.238356164383562e-06, "loss": 0.7322, "step": 8250 }, { "epoch": 1.5159817351598175, "grad_norm": 12.468632698059082, "learning_rate": 6.215525114155252e-06, "loss": 0.7916, "step": 8300 }, { "epoch": 1.5251141552511416, "grad_norm": 20.91657257080078, "learning_rate": 6.1926940639269405e-06, "loss": 0.7525, "step": 8350 }, { "epoch": 1.5342465753424657, "grad_norm": 38.200714111328125, "learning_rate": 6.169863013698631e-06, "loss": 0.7487, "step": 8400 }, { "epoch": 1.54337899543379, "grad_norm": 28.027910232543945, "learning_rate": 6.14703196347032e-06, "loss": 0.783, "step": 8450 }, { "epoch": 1.5525114155251143, "grad_norm": 9.391847610473633, "learning_rate": 6.12420091324201e-06, "loss": 0.7869, "step": 8500 }, { "epoch": 1.5616438356164384, "grad_norm": 122.94676971435547, "learning_rate": 6.101369863013698e-06, "loss": 0.8702, "step": 8550 }, { "epoch": 1.5707762557077625, "grad_norm": 32.576805114746094, "learning_rate": 6.0785388127853885e-06, "loss": 0.7748, "step": 8600 }, { "epoch": 1.5799086757990868, "grad_norm": 17.054048538208008, "learning_rate": 6.055707762557078e-06, "loss": 0.7431, "step": 8650 }, { "epoch": 1.589041095890411, "grad_norm": 17.969711303710938, "learning_rate": 6.032876712328768e-06, "loss": 0.8278, "step": 8700 }, { "epoch": 1.5981735159817352, "grad_norm": 33.43639373779297, "learning_rate": 6.010045662100456e-06, "loss": 0.7789, "step": 8750 }, { "epoch": 1.6073059360730593, "grad_norm": 28.036745071411133, "learning_rate": 5.987214611872146e-06, "loss": 0.7694, "step": 8800 }, { "epoch": 1.6164383561643836, "grad_norm": 17.829118728637695, "learning_rate": 5.9643835616438365e-06, "loss": 0.7598, "step": 8850 }, { "epoch": 1.625570776255708, "grad_norm": 19.825571060180664, "learning_rate": 5.941552511415526e-06, "loss": 0.6907, "step": 8900 }, { "epoch": 1.634703196347032, "grad_norm": 29.223722457885742, "learning_rate": 5.918721461187214e-06, "loss": 0.7, "step": 8950 }, { "epoch": 1.643835616438356, "grad_norm": 9.084179878234863, "learning_rate": 5.895890410958904e-06, "loss": 0.7237, "step": 9000 }, { "epoch": 1.6529680365296804, "grad_norm": 18.593826293945312, "learning_rate": 5.873059360730594e-06, "loss": 0.7623, "step": 9050 }, { "epoch": 1.6621004566210047, "grad_norm": 21.121200561523438, "learning_rate": 5.850228310502284e-06, "loss": 0.7438, "step": 9100 }, { "epoch": 1.6712328767123288, "grad_norm": 22.33557891845703, "learning_rate": 5.827397260273973e-06, "loss": 0.7657, "step": 9150 }, { "epoch": 1.6803652968036529, "grad_norm": 19.565675735473633, "learning_rate": 5.804566210045662e-06, "loss": 0.7604, "step": 9200 }, { "epoch": 1.6894977168949772, "grad_norm": 24.26027488708496, "learning_rate": 5.781735159817352e-06, "loss": 0.8224, "step": 9250 }, { "epoch": 1.6986301369863015, "grad_norm": 13.819406509399414, "learning_rate": 5.7589041095890415e-06, "loss": 0.7744, "step": 9300 }, { "epoch": 1.7077625570776256, "grad_norm": 20.518844604492188, "learning_rate": 5.736073059360731e-06, "loss": 0.7758, "step": 9350 }, { "epoch": 1.7168949771689497, "grad_norm": 13.40528678894043, "learning_rate": 5.71324200913242e-06, "loss": 0.7066, "step": 9400 }, { "epoch": 1.726027397260274, "grad_norm": 8.37540054321289, "learning_rate": 5.69041095890411e-06, "loss": 0.6852, "step": 9450 }, { "epoch": 1.7351598173515983, "grad_norm": 21.453828811645508, "learning_rate": 5.6675799086758e-06, "loss": 0.675, "step": 9500 }, { "epoch": 1.7442922374429224, "grad_norm": 12.05217456817627, "learning_rate": 5.644748858447489e-06, "loss": 0.6603, "step": 9550 }, { "epoch": 1.7534246575342465, "grad_norm": 26.813020706176758, "learning_rate": 5.621917808219178e-06, "loss": 0.8166, "step": 9600 }, { "epoch": 1.7625570776255708, "grad_norm": 35.790435791015625, "learning_rate": 5.599086757990868e-06, "loss": 0.671, "step": 9650 }, { "epoch": 1.771689497716895, "grad_norm": 21.900400161743164, "learning_rate": 5.576255707762558e-06, "loss": 0.7431, "step": 9700 }, { "epoch": 1.7808219178082192, "grad_norm": 31.134845733642578, "learning_rate": 5.553424657534247e-06, "loss": 0.7594, "step": 9750 }, { "epoch": 1.7899543378995433, "grad_norm": 14.03296947479248, "learning_rate": 5.530593607305937e-06, "loss": 0.7507, "step": 9800 }, { "epoch": 1.7990867579908676, "grad_norm": 94.07928466796875, "learning_rate": 5.507762557077626e-06, "loss": 0.8151, "step": 9850 }, { "epoch": 1.808219178082192, "grad_norm": 11.925488471984863, "learning_rate": 5.484931506849316e-06, "loss": 0.7601, "step": 9900 }, { "epoch": 1.817351598173516, "grad_norm": 13.049914360046387, "learning_rate": 5.4621004566210045e-06, "loss": 0.7119, "step": 9950 }, { "epoch": 1.82648401826484, "grad_norm": 16.289236068725586, "learning_rate": 5.4392694063926946e-06, "loss": 0.7411, "step": 10000 }, { "epoch": 1.82648401826484, "eval_exact_match": 81.82592242194892, "eval_f1": 89.05234914493408, "eval_runtime": 395.3505, "eval_samples_per_second": 26.736, "eval_steps_per_second": 1.672, "step": 10000 }, { "epoch": 1.8356164383561644, "grad_norm": 25.6737117767334, "learning_rate": 5.416438356164384e-06, "loss": 0.7937, "step": 10050 }, { "epoch": 1.8447488584474887, "grad_norm": 17.59447479248047, "learning_rate": 5.393607305936074e-06, "loss": 0.7665, "step": 10100 }, { "epoch": 1.8538812785388128, "grad_norm": 11.786163330078125, "learning_rate": 5.370776255707762e-06, "loss": 0.811, "step": 10150 }, { "epoch": 1.8630136986301369, "grad_norm": 19.147621154785156, "learning_rate": 5.3479452054794525e-06, "loss": 0.7869, "step": 10200 }, { "epoch": 1.8721461187214612, "grad_norm": 25.873458862304688, "learning_rate": 5.325114155251142e-06, "loss": 0.8007, "step": 10250 }, { "epoch": 1.8812785388127855, "grad_norm": 21.098203659057617, "learning_rate": 5.302283105022832e-06, "loss": 0.7293, "step": 10300 }, { "epoch": 1.8904109589041096, "grad_norm": 12.302563667297363, "learning_rate": 5.27945205479452e-06, "loss": 0.712, "step": 10350 }, { "epoch": 1.8995433789954337, "grad_norm": 15.331055641174316, "learning_rate": 5.25662100456621e-06, "loss": 0.6805, "step": 10400 }, { "epoch": 1.908675799086758, "grad_norm": 43.43006134033203, "learning_rate": 5.2337899543379005e-06, "loss": 0.7232, "step": 10450 }, { "epoch": 1.9178082191780823, "grad_norm": 73.6401596069336, "learning_rate": 5.21095890410959e-06, "loss": 0.7803, "step": 10500 }, { "epoch": 1.9269406392694064, "grad_norm": 7.5964436531066895, "learning_rate": 5.188127853881278e-06, "loss": 0.7418, "step": 10550 }, { "epoch": 1.9360730593607305, "grad_norm": 18.22124671936035, "learning_rate": 5.165296803652968e-06, "loss": 0.8099, "step": 10600 }, { "epoch": 1.9452054794520548, "grad_norm": 27.608963012695312, "learning_rate": 5.142465753424658e-06, "loss": 0.7369, "step": 10650 }, { "epoch": 1.954337899543379, "grad_norm": 45.187503814697266, "learning_rate": 5.119634703196348e-06, "loss": 0.8291, "step": 10700 }, { "epoch": 1.9634703196347032, "grad_norm": 23.95409393310547, "learning_rate": 5.096803652968037e-06, "loss": 0.7422, "step": 10750 }, { "epoch": 1.9726027397260273, "grad_norm": 14.64238452911377, "learning_rate": 5.073972602739726e-06, "loss": 0.7852, "step": 10800 }, { "epoch": 1.9817351598173516, "grad_norm": 16.381633758544922, "learning_rate": 5.051141552511416e-06, "loss": 0.7173, "step": 10850 }, { "epoch": 1.990867579908676, "grad_norm": 16.627071380615234, "learning_rate": 5.0283105022831055e-06, "loss": 0.8126, "step": 10900 }, { "epoch": 2.0, "grad_norm": 23.54231071472168, "learning_rate": 5.005479452054795e-06, "loss": 0.7654, "step": 10950 }, { "epoch": 2.009132420091324, "grad_norm": 16.47466278076172, "learning_rate": 4.982648401826484e-06, "loss": 0.5292, "step": 11000 }, { "epoch": 2.018264840182648, "grad_norm": 20.614046096801758, "learning_rate": 4.959817351598174e-06, "loss": 0.5507, "step": 11050 }, { "epoch": 2.0273972602739727, "grad_norm": 22.253305435180664, "learning_rate": 4.93744292237443e-06, "loss": 0.506, "step": 11100 }, { "epoch": 2.036529680365297, "grad_norm": 38.29316711425781, "learning_rate": 4.9146118721461185e-06, "loss": 0.5594, "step": 11150 }, { "epoch": 2.045662100456621, "grad_norm": 21.850147247314453, "learning_rate": 4.891780821917809e-06, "loss": 0.5626, "step": 11200 }, { "epoch": 2.0547945205479454, "grad_norm": 13.217841148376465, "learning_rate": 4.868949771689498e-06, "loss": 0.5738, "step": 11250 }, { "epoch": 2.0639269406392695, "grad_norm": 18.664854049682617, "learning_rate": 4.846118721461188e-06, "loss": 0.6301, "step": 11300 }, { "epoch": 2.0730593607305936, "grad_norm": 11.781886100769043, "learning_rate": 4.823287671232877e-06, "loss": 0.534, "step": 11350 }, { "epoch": 2.0821917808219177, "grad_norm": 19.580331802368164, "learning_rate": 4.8004566210045665e-06, "loss": 0.5434, "step": 11400 }, { "epoch": 2.091324200913242, "grad_norm": 15.862279891967773, "learning_rate": 4.777625570776256e-06, "loss": 0.5809, "step": 11450 }, { "epoch": 2.1004566210045663, "grad_norm": 35.96525955200195, "learning_rate": 4.754794520547946e-06, "loss": 0.5169, "step": 11500 }, { "epoch": 2.1095890410958904, "grad_norm": 14.418628692626953, "learning_rate": 4.731963470319635e-06, "loss": 0.5428, "step": 11550 }, { "epoch": 2.1187214611872145, "grad_norm": 19.833417892456055, "learning_rate": 4.709132420091324e-06, "loss": 0.554, "step": 11600 }, { "epoch": 2.127853881278539, "grad_norm": 7.7394118309021, "learning_rate": 4.686301369863014e-06, "loss": 0.5387, "step": 11650 }, { "epoch": 2.136986301369863, "grad_norm": 13.092580795288086, "learning_rate": 4.663470319634704e-06, "loss": 0.5263, "step": 11700 }, { "epoch": 2.146118721461187, "grad_norm": 11.311351776123047, "learning_rate": 4.640639269406393e-06, "loss": 0.5652, "step": 11750 }, { "epoch": 2.1552511415525113, "grad_norm": 8.475380897521973, "learning_rate": 4.617808219178082e-06, "loss": 0.5572, "step": 11800 }, { "epoch": 2.1643835616438354, "grad_norm": 9.318852424621582, "learning_rate": 4.5949771689497715e-06, "loss": 0.5704, "step": 11850 }, { "epoch": 2.17351598173516, "grad_norm": 17.163570404052734, "learning_rate": 4.572146118721462e-06, "loss": 0.5533, "step": 11900 }, { "epoch": 2.182648401826484, "grad_norm": 11.318022727966309, "learning_rate": 4.549315068493151e-06, "loss": 0.4844, "step": 11950 }, { "epoch": 2.191780821917808, "grad_norm": 25.644908905029297, "learning_rate": 4.526484018264841e-06, "loss": 0.5628, "step": 12000 }, { "epoch": 2.191780821917808, "eval_exact_match": 82.33680227057711, "eval_f1": 89.23466367243802, "eval_runtime": 395.5495, "eval_samples_per_second": 26.722, "eval_steps_per_second": 1.671, "step": 12000 }, { "epoch": 2.2009132420091326, "grad_norm": 22.05693817138672, "learning_rate": 4.50365296803653e-06, "loss": 0.4721, "step": 12050 }, { "epoch": 2.2100456621004567, "grad_norm": 48.57001495361328, "learning_rate": 4.4808219178082195e-06, "loss": 0.5225, "step": 12100 }, { "epoch": 2.219178082191781, "grad_norm": 18.33832359313965, "learning_rate": 4.457990867579909e-06, "loss": 0.5864, "step": 12150 }, { "epoch": 2.228310502283105, "grad_norm": 17.940277099609375, "learning_rate": 4.435159817351599e-06, "loss": 0.5558, "step": 12200 }, { "epoch": 2.237442922374429, "grad_norm": 15.108060836791992, "learning_rate": 4.412328767123288e-06, "loss": 0.5825, "step": 12250 }, { "epoch": 2.2465753424657535, "grad_norm": 87.06202697753906, "learning_rate": 4.389497716894977e-06, "loss": 0.5755, "step": 12300 }, { "epoch": 2.2557077625570776, "grad_norm": 25.141014099121094, "learning_rate": 4.366666666666667e-06, "loss": 0.662, "step": 12350 }, { "epoch": 2.2648401826484017, "grad_norm": 17.8304386138916, "learning_rate": 4.343835616438357e-06, "loss": 0.5106, "step": 12400 }, { "epoch": 2.2739726027397262, "grad_norm": 30.3131160736084, "learning_rate": 4.321004566210046e-06, "loss": 0.5553, "step": 12450 }, { "epoch": 2.2831050228310503, "grad_norm": 16.719921112060547, "learning_rate": 4.298173515981735e-06, "loss": 0.5601, "step": 12500 }, { "epoch": 2.2922374429223744, "grad_norm": 17.55047035217285, "learning_rate": 4.2753424657534245e-06, "loss": 0.5909, "step": 12550 }, { "epoch": 2.3013698630136985, "grad_norm": 25.325210571289062, "learning_rate": 4.252511415525115e-06, "loss": 0.5292, "step": 12600 }, { "epoch": 2.3105022831050226, "grad_norm": 7.231600284576416, "learning_rate": 4.229680365296804e-06, "loss": 0.5618, "step": 12650 }, { "epoch": 2.319634703196347, "grad_norm": 45.177276611328125, "learning_rate": 4.206849315068494e-06, "loss": 0.6545, "step": 12700 }, { "epoch": 2.328767123287671, "grad_norm": 14.02021312713623, "learning_rate": 4.184018264840182e-06, "loss": 0.5223, "step": 12750 }, { "epoch": 2.3378995433789953, "grad_norm": 15.296303749084473, "learning_rate": 4.1611872146118725e-06, "loss": 0.5214, "step": 12800 }, { "epoch": 2.34703196347032, "grad_norm": 12.6821870803833, "learning_rate": 4.138356164383562e-06, "loss": 0.5621, "step": 12850 }, { "epoch": 2.356164383561644, "grad_norm": 11.630056381225586, "learning_rate": 4.115525114155252e-06, "loss": 0.5318, "step": 12900 }, { "epoch": 2.365296803652968, "grad_norm": 18.717533111572266, "learning_rate": 4.092694063926941e-06, "loss": 0.6217, "step": 12950 }, { "epoch": 2.374429223744292, "grad_norm": 18.65459442138672, "learning_rate": 4.06986301369863e-06, "loss": 0.5254, "step": 13000 }, { "epoch": 2.383561643835616, "grad_norm": 51.29315948486328, "learning_rate": 4.04703196347032e-06, "loss": 0.5516, "step": 13050 }, { "epoch": 2.3926940639269407, "grad_norm": 28.213186264038086, "learning_rate": 4.02420091324201e-06, "loss": 0.5661, "step": 13100 }, { "epoch": 2.401826484018265, "grad_norm": 25.394929885864258, "learning_rate": 4.001369863013699e-06, "loss": 0.5811, "step": 13150 }, { "epoch": 2.410958904109589, "grad_norm": 11.663569450378418, "learning_rate": 3.978538812785388e-06, "loss": 0.5785, "step": 13200 }, { "epoch": 2.4200913242009134, "grad_norm": 23.070152282714844, "learning_rate": 3.9557077625570776e-06, "loss": 0.5997, "step": 13250 }, { "epoch": 2.4292237442922375, "grad_norm": 9.592535018920898, "learning_rate": 3.932876712328768e-06, "loss": 0.5743, "step": 13300 }, { "epoch": 2.4383561643835616, "grad_norm": 23.151607513427734, "learning_rate": 3.910045662100457e-06, "loss": 0.5721, "step": 13350 }, { "epoch": 2.4474885844748857, "grad_norm": 22.872838973999023, "learning_rate": 3.887214611872146e-06, "loss": 0.5103, "step": 13400 }, { "epoch": 2.45662100456621, "grad_norm": 49.325408935546875, "learning_rate": 3.8643835616438354e-06, "loss": 0.552, "step": 13450 }, { "epoch": 2.4657534246575343, "grad_norm": 13.299301147460938, "learning_rate": 3.8415525114155256e-06, "loss": 0.6012, "step": 13500 }, { "epoch": 2.4748858447488584, "grad_norm": 10.96218204498291, "learning_rate": 3.818721461187215e-06, "loss": 0.5902, "step": 13550 }, { "epoch": 2.4840182648401825, "grad_norm": 25.126304626464844, "learning_rate": 3.7958904109589045e-06, "loss": 0.5583, "step": 13600 }, { "epoch": 2.493150684931507, "grad_norm": 88.76774597167969, "learning_rate": 3.7730593607305938e-06, "loss": 0.513, "step": 13650 }, { "epoch": 2.502283105022831, "grad_norm": 19.230432510375977, "learning_rate": 3.7502283105022834e-06, "loss": 0.5879, "step": 13700 }, { "epoch": 2.5114155251141552, "grad_norm": 9.016436576843262, "learning_rate": 3.7273972602739727e-06, "loss": 0.5277, "step": 13750 }, { "epoch": 2.5205479452054793, "grad_norm": 9.230749130249023, "learning_rate": 3.7045662100456624e-06, "loss": 0.6139, "step": 13800 }, { "epoch": 2.5296803652968034, "grad_norm": 10.221979141235352, "learning_rate": 3.6817351598173516e-06, "loss": 0.5702, "step": 13850 }, { "epoch": 2.538812785388128, "grad_norm": 26.807771682739258, "learning_rate": 3.6589041095890413e-06, "loss": 0.594, "step": 13900 }, { "epoch": 2.547945205479452, "grad_norm": 15.471819877624512, "learning_rate": 3.6360730593607306e-06, "loss": 0.665, "step": 13950 }, { "epoch": 2.557077625570776, "grad_norm": 29.882976531982422, "learning_rate": 3.6132420091324207e-06, "loss": 0.5556, "step": 14000 }, { "epoch": 2.557077625570776, "eval_exact_match": 82.63008514664143, "eval_f1": 89.62653441826156, "eval_runtime": 395.4302, "eval_samples_per_second": 26.73, "eval_steps_per_second": 1.672, "step": 14000 }, { "epoch": 2.5662100456621006, "grad_norm": 20.645360946655273, "learning_rate": 3.5904109589041095e-06, "loss": 0.5454, "step": 14050 }, { "epoch": 2.5753424657534247, "grad_norm": 22.412696838378906, "learning_rate": 3.5675799086757996e-06, "loss": 0.5218, "step": 14100 }, { "epoch": 2.584474885844749, "grad_norm": 22.36018180847168, "learning_rate": 3.544748858447489e-06, "loss": 0.5433, "step": 14150 }, { "epoch": 2.593607305936073, "grad_norm": 30.506481170654297, "learning_rate": 3.5219178082191786e-06, "loss": 0.5584, "step": 14200 }, { "epoch": 2.602739726027397, "grad_norm": 18.273223876953125, "learning_rate": 3.499086757990868e-06, "loss": 0.4913, "step": 14250 }, { "epoch": 2.6118721461187215, "grad_norm": 29.807708740234375, "learning_rate": 3.4762557077625575e-06, "loss": 0.5721, "step": 14300 }, { "epoch": 2.6210045662100456, "grad_norm": 59.90718078613281, "learning_rate": 3.453881278538813e-06, "loss": 0.4875, "step": 14350 }, { "epoch": 2.6301369863013697, "grad_norm": 8.444619178771973, "learning_rate": 3.4310502283105023e-06, "loss": 0.5851, "step": 14400 }, { "epoch": 2.6392694063926943, "grad_norm": 119.5136947631836, "learning_rate": 3.408219178082192e-06, "loss": 0.5803, "step": 14450 }, { "epoch": 2.6484018264840183, "grad_norm": 14.274751663208008, "learning_rate": 3.3853881278538813e-06, "loss": 0.5416, "step": 14500 }, { "epoch": 2.6575342465753424, "grad_norm": 23.54309844970703, "learning_rate": 3.3625570776255714e-06, "loss": 0.5252, "step": 14550 }, { "epoch": 2.6666666666666665, "grad_norm": 17.312332153320312, "learning_rate": 3.33972602739726e-06, "loss": 0.5931, "step": 14600 }, { "epoch": 2.6757990867579906, "grad_norm": 23.978649139404297, "learning_rate": 3.3168949771689503e-06, "loss": 0.5912, "step": 14650 }, { "epoch": 2.684931506849315, "grad_norm": 27.63211441040039, "learning_rate": 3.2940639269406396e-06, "loss": 0.6038, "step": 14700 }, { "epoch": 2.6940639269406392, "grad_norm": 9.839302062988281, "learning_rate": 3.2712328767123292e-06, "loss": 0.565, "step": 14750 }, { "epoch": 2.7031963470319633, "grad_norm": 12.635977745056152, "learning_rate": 3.2484018264840185e-06, "loss": 0.5678, "step": 14800 }, { "epoch": 2.712328767123288, "grad_norm": 21.576257705688477, "learning_rate": 3.225570776255708e-06, "loss": 0.4886, "step": 14850 }, { "epoch": 2.721461187214612, "grad_norm": 9.049654960632324, "learning_rate": 3.2027397260273974e-06, "loss": 0.5295, "step": 14900 }, { "epoch": 2.730593607305936, "grad_norm": 54.493343353271484, "learning_rate": 3.179908675799087e-06, "loss": 0.5178, "step": 14950 }, { "epoch": 2.73972602739726, "grad_norm": 21.172515869140625, "learning_rate": 3.1575342465753427e-06, "loss": 0.6128, "step": 15000 }, { "epoch": 2.748858447488584, "grad_norm": 23.262418746948242, "learning_rate": 3.134703196347032e-06, "loss": 0.5992, "step": 15050 }, { "epoch": 2.7579908675799087, "grad_norm": 6.960672855377197, "learning_rate": 3.1118721461187216e-06, "loss": 0.5634, "step": 15100 }, { "epoch": 2.767123287671233, "grad_norm": 53.032554626464844, "learning_rate": 3.089041095890411e-06, "loss": 0.5313, "step": 15150 }, { "epoch": 2.776255707762557, "grad_norm": 17.943723678588867, "learning_rate": 3.066210045662101e-06, "loss": 0.5871, "step": 15200 }, { "epoch": 2.7853881278538815, "grad_norm": 26.733182907104492, "learning_rate": 3.04337899543379e-06, "loss": 0.5628, "step": 15250 }, { "epoch": 2.7945205479452055, "grad_norm": 13.597286224365234, "learning_rate": 3.02054794520548e-06, "loss": 0.563, "step": 15300 }, { "epoch": 2.8036529680365296, "grad_norm": 23.258501052856445, "learning_rate": 2.997716894977169e-06, "loss": 0.5781, "step": 15350 }, { "epoch": 2.8127853881278537, "grad_norm": 32.529056549072266, "learning_rate": 2.974885844748859e-06, "loss": 0.5328, "step": 15400 }, { "epoch": 2.821917808219178, "grad_norm": 44.435787200927734, "learning_rate": 2.952054794520548e-06, "loss": 0.5254, "step": 15450 }, { "epoch": 2.8310502283105023, "grad_norm": 10.251275062561035, "learning_rate": 2.929223744292238e-06, "loss": 0.5194, "step": 15500 }, { "epoch": 2.8401826484018264, "grad_norm": 24.379085540771484, "learning_rate": 2.906392694063927e-06, "loss": 0.5232, "step": 15550 }, { "epoch": 2.8493150684931505, "grad_norm": 32.740516662597656, "learning_rate": 2.8835616438356167e-06, "loss": 0.5731, "step": 15600 }, { "epoch": 2.858447488584475, "grad_norm": 24.7740478515625, "learning_rate": 2.860730593607306e-06, "loss": 0.5251, "step": 15650 }, { "epoch": 2.867579908675799, "grad_norm": 29.77228355407715, "learning_rate": 2.8378995433789957e-06, "loss": 0.5457, "step": 15700 }, { "epoch": 2.8767123287671232, "grad_norm": 24.371246337890625, "learning_rate": 2.815068493150685e-06, "loss": 0.574, "step": 15750 }, { "epoch": 2.8858447488584473, "grad_norm": 10.619139671325684, "learning_rate": 2.7922374429223746e-06, "loss": 0.5337, "step": 15800 }, { "epoch": 2.8949771689497714, "grad_norm": 31.601932525634766, "learning_rate": 2.769406392694064e-06, "loss": 0.5693, "step": 15850 }, { "epoch": 2.904109589041096, "grad_norm": 28.158267974853516, "learning_rate": 2.7465753424657536e-06, "loss": 0.561, "step": 15900 }, { "epoch": 2.91324200913242, "grad_norm": 23.712093353271484, "learning_rate": 2.723744292237443e-06, "loss": 0.5312, "step": 15950 }, { "epoch": 2.922374429223744, "grad_norm": 16.564476013183594, "learning_rate": 2.700913242009133e-06, "loss": 0.4938, "step": 16000 }, { "epoch": 2.922374429223744, "eval_exact_match": 82.58278145695364, "eval_f1": 89.75516423119765, "eval_runtime": 395.7187, "eval_samples_per_second": 26.711, "eval_steps_per_second": 1.67, "step": 16000 }, { "epoch": 2.9315068493150687, "grad_norm": 10.199018478393555, "learning_rate": 2.6780821917808218e-06, "loss": 0.5761, "step": 16050 }, { "epoch": 2.9406392694063928, "grad_norm": 12.591775894165039, "learning_rate": 2.655251141552512e-06, "loss": 0.5681, "step": 16100 }, { "epoch": 2.949771689497717, "grad_norm": 20.624439239501953, "learning_rate": 2.632420091324201e-06, "loss": 0.4934, "step": 16150 }, { "epoch": 2.958904109589041, "grad_norm": 38.65128707885742, "learning_rate": 2.609589041095891e-06, "loss": 0.5956, "step": 16200 }, { "epoch": 2.968036529680365, "grad_norm": 24.270273208618164, "learning_rate": 2.58675799086758e-06, "loss": 0.5219, "step": 16250 }, { "epoch": 2.9771689497716896, "grad_norm": 24.47715950012207, "learning_rate": 2.5639269406392698e-06, "loss": 0.5067, "step": 16300 }, { "epoch": 2.9863013698630136, "grad_norm": 7.162631988525391, "learning_rate": 2.541095890410959e-06, "loss": 0.5377, "step": 16350 }, { "epoch": 2.9954337899543377, "grad_norm": 9.583894729614258, "learning_rate": 2.5182648401826487e-06, "loss": 0.5518, "step": 16400 }, { "epoch": 3.0045662100456623, "grad_norm": 25.51879119873047, "learning_rate": 2.4958904109589042e-06, "loss": 0.4965, "step": 16450 }, { "epoch": 3.0136986301369864, "grad_norm": 142.86093139648438, "learning_rate": 2.473059360730594e-06, "loss": 0.4489, "step": 16500 }, { "epoch": 3.0228310502283104, "grad_norm": 13.863718032836914, "learning_rate": 2.450228310502283e-06, "loss": 0.4735, "step": 16550 }, { "epoch": 3.0319634703196345, "grad_norm": 12.609000205993652, "learning_rate": 2.427397260273973e-06, "loss": 0.4123, "step": 16600 }, { "epoch": 3.041095890410959, "grad_norm": 12.495716094970703, "learning_rate": 2.4045662100456625e-06, "loss": 0.4467, "step": 16650 }, { "epoch": 3.050228310502283, "grad_norm": 23.454723358154297, "learning_rate": 2.381735159817352e-06, "loss": 0.4544, "step": 16700 }, { "epoch": 3.0593607305936072, "grad_norm": 21.551776885986328, "learning_rate": 2.3589041095890415e-06, "loss": 0.4199, "step": 16750 }, { "epoch": 3.0684931506849313, "grad_norm": 29.826936721801758, "learning_rate": 2.3360730593607308e-06, "loss": 0.4447, "step": 16800 }, { "epoch": 3.077625570776256, "grad_norm": 24.2338809967041, "learning_rate": 2.3132420091324204e-06, "loss": 0.4281, "step": 16850 }, { "epoch": 3.08675799086758, "grad_norm": 12.561247825622559, "learning_rate": 2.2904109589041097e-06, "loss": 0.4572, "step": 16900 }, { "epoch": 3.095890410958904, "grad_norm": 26.721580505371094, "learning_rate": 2.2675799086757994e-06, "loss": 0.407, "step": 16950 }, { "epoch": 3.105022831050228, "grad_norm": 53.11214828491211, "learning_rate": 2.244748858447489e-06, "loss": 0.4183, "step": 17000 }, { "epoch": 3.1141552511415527, "grad_norm": 13.374695777893066, "learning_rate": 2.2219178082191783e-06, "loss": 0.4651, "step": 17050 }, { "epoch": 3.1232876712328768, "grad_norm": 17.132118225097656, "learning_rate": 2.199086757990868e-06, "loss": 0.3833, "step": 17100 }, { "epoch": 3.132420091324201, "grad_norm": 12.751568794250488, "learning_rate": 2.1762557077625573e-06, "loss": 0.4563, "step": 17150 }, { "epoch": 3.141552511415525, "grad_norm": 11.56992244720459, "learning_rate": 2.153424657534247e-06, "loss": 0.4315, "step": 17200 }, { "epoch": 3.1506849315068495, "grad_norm": 18.173213958740234, "learning_rate": 2.130593607305936e-06, "loss": 0.4013, "step": 17250 }, { "epoch": 3.1598173515981736, "grad_norm": 15.345487594604492, "learning_rate": 2.107762557077626e-06, "loss": 0.4176, "step": 17300 }, { "epoch": 3.1689497716894977, "grad_norm": 11.01826286315918, "learning_rate": 2.084931506849315e-06, "loss": 0.4426, "step": 17350 }, { "epoch": 3.1780821917808217, "grad_norm": 27.82256507873535, "learning_rate": 2.062100456621005e-06, "loss": 0.4511, "step": 17400 }, { "epoch": 3.1872146118721463, "grad_norm": 23.4956111907959, "learning_rate": 2.0392694063926945e-06, "loss": 0.4195, "step": 17450 }, { "epoch": 3.1963470319634704, "grad_norm": 15.364335060119629, "learning_rate": 2.0164383561643838e-06, "loss": 0.423, "step": 17500 }, { "epoch": 3.2054794520547945, "grad_norm": 6.171648025512695, "learning_rate": 1.9936073059360735e-06, "loss": 0.4323, "step": 17550 }, { "epoch": 3.2146118721461185, "grad_norm": 10.051441192626953, "learning_rate": 1.9707762557077627e-06, "loss": 0.4307, "step": 17600 }, { "epoch": 3.223744292237443, "grad_norm": 15.94997501373291, "learning_rate": 1.9479452054794524e-06, "loss": 0.4034, "step": 17650 }, { "epoch": 3.232876712328767, "grad_norm": 43.07652282714844, "learning_rate": 1.9251141552511417e-06, "loss": 0.4106, "step": 17700 }, { "epoch": 3.2420091324200913, "grad_norm": 17.467111587524414, "learning_rate": 1.9022831050228313e-06, "loss": 0.4648, "step": 17750 }, { "epoch": 3.2511415525114153, "grad_norm": 10.493595123291016, "learning_rate": 1.8794520547945208e-06, "loss": 0.4619, "step": 17800 }, { "epoch": 3.26027397260274, "grad_norm": 10.744498252868652, "learning_rate": 1.8566210045662103e-06, "loss": 0.3861, "step": 17850 }, { "epoch": 3.269406392694064, "grad_norm": 12.96605110168457, "learning_rate": 1.8337899543378998e-06, "loss": 0.4077, "step": 17900 }, { "epoch": 3.278538812785388, "grad_norm": 6.190354347229004, "learning_rate": 1.8109589041095892e-06, "loss": 0.3937, "step": 17950 }, { "epoch": 3.287671232876712, "grad_norm": 15.968260765075684, "learning_rate": 1.7881278538812787e-06, "loss": 0.4852, "step": 18000 }, { "epoch": 3.287671232876712, "eval_exact_match": 82.42194891201514, "eval_f1": 89.61372740496692, "eval_runtime": 413.8098, "eval_samples_per_second": 25.543, "eval_steps_per_second": 1.597, "step": 18000 }, { "epoch": 3.2968036529680367, "grad_norm": 39.12847900390625, "learning_rate": 1.7652968036529684e-06, "loss": 0.3748, "step": 18050 }, { "epoch": 3.3059360730593608, "grad_norm": 19.359882354736328, "learning_rate": 1.7424657534246579e-06, "loss": 0.4165, "step": 18100 }, { "epoch": 3.315068493150685, "grad_norm": 6.587915897369385, "learning_rate": 1.7196347031963473e-06, "loss": 0.4489, "step": 18150 }, { "epoch": 3.324200913242009, "grad_norm": 10.52927017211914, "learning_rate": 1.6968036529680368e-06, "loss": 0.4074, "step": 18200 }, { "epoch": 3.3333333333333335, "grad_norm": 25.514862060546875, "learning_rate": 1.6739726027397263e-06, "loss": 0.4552, "step": 18250 }, { "epoch": 3.3424657534246576, "grad_norm": 12.100693702697754, "learning_rate": 1.6511415525114157e-06, "loss": 0.4463, "step": 18300 }, { "epoch": 3.3515981735159817, "grad_norm": 31.30626106262207, "learning_rate": 1.6283105022831052e-06, "loss": 0.393, "step": 18350 }, { "epoch": 3.3607305936073057, "grad_norm": 19.230247497558594, "learning_rate": 1.6054794520547947e-06, "loss": 0.4318, "step": 18400 }, { "epoch": 3.3698630136986303, "grad_norm": 10.644558906555176, "learning_rate": 1.5826484018264844e-06, "loss": 0.453, "step": 18450 }, { "epoch": 3.3789954337899544, "grad_norm": 9.888466835021973, "learning_rate": 1.5598173515981738e-06, "loss": 0.4205, "step": 18500 }, { "epoch": 3.3881278538812785, "grad_norm": 14.341341972351074, "learning_rate": 1.5369863013698633e-06, "loss": 0.4418, "step": 18550 }, { "epoch": 3.3972602739726026, "grad_norm": 26.137784957885742, "learning_rate": 1.5141552511415528e-06, "loss": 0.421, "step": 18600 }, { "epoch": 3.406392694063927, "grad_norm": 8.944993019104004, "learning_rate": 1.4913242009132423e-06, "loss": 0.4003, "step": 18650 }, { "epoch": 3.415525114155251, "grad_norm": 49.0703125, "learning_rate": 1.4684931506849317e-06, "loss": 0.4567, "step": 18700 }, { "epoch": 3.4246575342465753, "grad_norm": 23.05057716369629, "learning_rate": 1.4456621004566212e-06, "loss": 0.4215, "step": 18750 }, { "epoch": 3.4337899543378994, "grad_norm": 21.53809356689453, "learning_rate": 1.4228310502283107e-06, "loss": 0.3586, "step": 18800 }, { "epoch": 3.442922374429224, "grad_norm": 13.685781478881836, "learning_rate": 1.4000000000000001e-06, "loss": 0.3776, "step": 18850 }, { "epoch": 3.452054794520548, "grad_norm": 20.39994239807129, "learning_rate": 1.3771689497716898e-06, "loss": 0.4484, "step": 18900 }, { "epoch": 3.461187214611872, "grad_norm": 35.2899169921875, "learning_rate": 1.3543378995433793e-06, "loss": 0.4426, "step": 18950 }, { "epoch": 3.470319634703196, "grad_norm": 54.691349029541016, "learning_rate": 1.3315068493150688e-06, "loss": 0.4736, "step": 19000 }, { "epoch": 3.4794520547945207, "grad_norm": 13.618513107299805, "learning_rate": 1.3086757990867582e-06, "loss": 0.377, "step": 19050 }, { "epoch": 3.4885844748858448, "grad_norm": 23.14171600341797, "learning_rate": 1.2858447488584475e-06, "loss": 0.4071, "step": 19100 }, { "epoch": 3.497716894977169, "grad_norm": 25.513710021972656, "learning_rate": 1.263013698630137e-06, "loss": 0.385, "step": 19150 }, { "epoch": 3.506849315068493, "grad_norm": 15.305421829223633, "learning_rate": 1.2401826484018267e-06, "loss": 0.4366, "step": 19200 }, { "epoch": 3.5159817351598175, "grad_norm": 15.072031021118164, "learning_rate": 1.2173515981735161e-06, "loss": 0.4637, "step": 19250 }, { "epoch": 3.5251141552511416, "grad_norm": 16.01192283630371, "learning_rate": 1.1945205479452056e-06, "loss": 0.4313, "step": 19300 }, { "epoch": 3.5342465753424657, "grad_norm": 18.003978729248047, "learning_rate": 1.1721461187214613e-06, "loss": 0.4291, "step": 19350 }, { "epoch": 3.54337899543379, "grad_norm": 33.20173263549805, "learning_rate": 1.1493150684931508e-06, "loss": 0.4792, "step": 19400 }, { "epoch": 3.5525114155251143, "grad_norm": 11.555415153503418, "learning_rate": 1.1264840182648403e-06, "loss": 0.4561, "step": 19450 }, { "epoch": 3.5616438356164384, "grad_norm": 12.765213966369629, "learning_rate": 1.1036529680365298e-06, "loss": 0.4258, "step": 19500 }, { "epoch": 3.5707762557077625, "grad_norm": 14.976128578186035, "learning_rate": 1.0808219178082192e-06, "loss": 0.4376, "step": 19550 }, { "epoch": 3.5799086757990866, "grad_norm": 21.439231872558594, "learning_rate": 1.0579908675799087e-06, "loss": 0.451, "step": 19600 }, { "epoch": 3.589041095890411, "grad_norm": 13.795149803161621, "learning_rate": 1.0351598173515982e-06, "loss": 0.389, "step": 19650 }, { "epoch": 3.598173515981735, "grad_norm": 10.679841995239258, "learning_rate": 1.0123287671232876e-06, "loss": 0.4535, "step": 19700 }, { "epoch": 3.6073059360730593, "grad_norm": 18.03670310974121, "learning_rate": 9.894977168949773e-07, "loss": 0.4252, "step": 19750 }, { "epoch": 3.616438356164384, "grad_norm": 68.05775451660156, "learning_rate": 9.666666666666668e-07, "loss": 0.4625, "step": 19800 }, { "epoch": 3.625570776255708, "grad_norm": 32.91834259033203, "learning_rate": 9.438356164383563e-07, "loss": 0.4295, "step": 19850 }, { "epoch": 3.634703196347032, "grad_norm": 34.28315734863281, "learning_rate": 9.210045662100457e-07, "loss": 0.3863, "step": 19900 }, { "epoch": 3.643835616438356, "grad_norm": 13.401795387268066, "learning_rate": 8.981735159817352e-07, "loss": 0.4354, "step": 19950 }, { "epoch": 3.65296803652968, "grad_norm": 9.942870140075684, "learning_rate": 8.753424657534247e-07, "loss": 0.3813, "step": 20000 }, { "epoch": 3.65296803652968, "eval_exact_match": 82.50709555345317, "eval_f1": 89.67676155395135, "eval_runtime": 810.5169, "eval_samples_per_second": 13.041, "eval_steps_per_second": 0.816, "step": 20000 }, { "epoch": 3.6621004566210047, "grad_norm": 33.140907287597656, "learning_rate": 8.525114155251143e-07, "loss": 0.4385, "step": 20050 }, { "epoch": 3.671232876712329, "grad_norm": 86.72850799560547, "learning_rate": 8.296803652968037e-07, "loss": 0.4785, "step": 20100 }, { "epoch": 3.680365296803653, "grad_norm": 11.17636775970459, "learning_rate": 8.068493150684932e-07, "loss": 0.4134, "step": 20150 }, { "epoch": 3.6894977168949774, "grad_norm": 14.498661994934082, "learning_rate": 7.840182648401827e-07, "loss": 0.3903, "step": 20200 }, { "epoch": 3.6986301369863015, "grad_norm": 18.49641227722168, "learning_rate": 7.611872146118722e-07, "loss": 0.4131, "step": 20250 }, { "epoch": 3.7077625570776256, "grad_norm": 23.517004013061523, "learning_rate": 7.383561643835617e-07, "loss": 0.4074, "step": 20300 }, { "epoch": 3.7168949771689497, "grad_norm": 14.428693771362305, "learning_rate": 7.155251141552512e-07, "loss": 0.3898, "step": 20350 }, { "epoch": 3.7260273972602738, "grad_norm": 13.074502944946289, "learning_rate": 6.926940639269407e-07, "loss": 0.4251, "step": 20400 }, { "epoch": 3.7351598173515983, "grad_norm": 10.06019401550293, "learning_rate": 6.698630136986301e-07, "loss": 0.409, "step": 20450 }, { "epoch": 3.7442922374429224, "grad_norm": 17.26006507873535, "learning_rate": 6.470319634703197e-07, "loss": 0.374, "step": 20500 }, { "epoch": 3.7534246575342465, "grad_norm": 19.47539710998535, "learning_rate": 6.242009132420092e-07, "loss": 0.3683, "step": 20550 }, { "epoch": 3.762557077625571, "grad_norm": 20.99232292175293, "learning_rate": 6.013698630136987e-07, "loss": 0.4675, "step": 20600 }, { "epoch": 3.771689497716895, "grad_norm": 48.530582427978516, "learning_rate": 5.785388127853881e-07, "loss": 0.4275, "step": 20650 }, { "epoch": 3.780821917808219, "grad_norm": 15.112560272216797, "learning_rate": 5.557077625570777e-07, "loss": 0.38, "step": 20700 }, { "epoch": 3.7899543378995433, "grad_norm": 13.57143497467041, "learning_rate": 5.328767123287672e-07, "loss": 0.3885, "step": 20750 }, { "epoch": 3.7990867579908674, "grad_norm": 24.962566375732422, "learning_rate": 5.100456621004567e-07, "loss": 0.4148, "step": 20800 }, { "epoch": 3.808219178082192, "grad_norm": 35.816993713378906, "learning_rate": 4.872146118721461e-07, "loss": 0.4211, "step": 20850 }, { "epoch": 3.817351598173516, "grad_norm": 29.423664093017578, "learning_rate": 4.6438356164383565e-07, "loss": 0.5014, "step": 20900 }, { "epoch": 3.82648401826484, "grad_norm": 39.580039978027344, "learning_rate": 4.415525114155252e-07, "loss": 0.4521, "step": 20950 }, { "epoch": 3.8356164383561646, "grad_norm": 6.263591766357422, "learning_rate": 4.1872146118721465e-07, "loss": 0.452, "step": 21000 }, { "epoch": 3.8447488584474887, "grad_norm": 68.41936492919922, "learning_rate": 3.958904109589042e-07, "loss": 0.4276, "step": 21050 }, { "epoch": 3.853881278538813, "grad_norm": 22.434158325195312, "learning_rate": 3.7305936073059365e-07, "loss": 0.4951, "step": 21100 }, { "epoch": 3.863013698630137, "grad_norm": 16.98661231994629, "learning_rate": 3.5022831050228317e-07, "loss": 0.4212, "step": 21150 }, { "epoch": 3.872146118721461, "grad_norm": 12.409255981445312, "learning_rate": 3.2739726027397264e-07, "loss": 0.4156, "step": 21200 }, { "epoch": 3.8812785388127855, "grad_norm": 26.019704818725586, "learning_rate": 3.0456621004566217e-07, "loss": 0.4105, "step": 21250 }, { "epoch": 3.8904109589041096, "grad_norm": 28.032352447509766, "learning_rate": 2.817351598173516e-07, "loss": 0.4564, "step": 21300 }, { "epoch": 3.8995433789954337, "grad_norm": 18.951528549194336, "learning_rate": 2.589041095890411e-07, "loss": 0.4687, "step": 21350 }, { "epoch": 3.908675799086758, "grad_norm": 36.14161682128906, "learning_rate": 2.360730593607306e-07, "loss": 0.4194, "step": 21400 }, { "epoch": 3.9178082191780823, "grad_norm": 22.935070037841797, "learning_rate": 2.132420091324201e-07, "loss": 0.4582, "step": 21450 }, { "epoch": 3.9269406392694064, "grad_norm": 15.434907913208008, "learning_rate": 1.904109589041096e-07, "loss": 0.3789, "step": 21500 }, { "epoch": 3.9360730593607305, "grad_norm": 5.678657054901123, "learning_rate": 1.675799086757991e-07, "loss": 0.4196, "step": 21550 }, { "epoch": 3.9452054794520546, "grad_norm": 14.900694847106934, "learning_rate": 1.447488584474886e-07, "loss": 0.4129, "step": 21600 }, { "epoch": 3.954337899543379, "grad_norm": 13.7803316116333, "learning_rate": 1.2191780821917807e-07, "loss": 0.4436, "step": 21650 }, { "epoch": 3.963470319634703, "grad_norm": 15.580832481384277, "learning_rate": 9.908675799086758e-08, "loss": 0.3984, "step": 21700 }, { "epoch": 3.9726027397260273, "grad_norm": 12.353429794311523, "learning_rate": 7.625570776255708e-08, "loss": 0.4279, "step": 21750 }, { "epoch": 3.981735159817352, "grad_norm": 10.456271171569824, "learning_rate": 5.3424657534246586e-08, "loss": 0.4438, "step": 21800 }, { "epoch": 3.990867579908676, "grad_norm": 17.142772674560547, "learning_rate": 3.059360730593608e-08, "loss": 0.4569, "step": 21850 }, { "epoch": 4.0, "grad_norm": 8.582050323486328, "learning_rate": 7.762557077625571e-09, "loss": 0.4129, "step": 21900 }, { "epoch": 4.0, "step": 21900, "total_flos": 2.6750162778225377e+18, "train_loss": 0.7768389247214957, "train_runtime": 96436.1031, "train_samples_per_second": 3.633, "train_steps_per_second": 0.227 } ], "logging_steps": 50, "max_steps": 21900, "num_input_tokens_seen": 0, "num_train_epochs": 4, "save_steps": 5000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.6750162778225377e+18, "train_batch_size": 4, "trial_name": null, "trial_params": null }