{ "best_metric": null, "best_model_checkpoint": null, "epoch": 5.988118811881188, "eval_steps": 500, "global_step": 1134, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.005280528052805281, "grad_norm": 7.677456378936768, "learning_rate": 4.385964912280702e-07, "loss": 0.9639, "step": 1 }, { "epoch": 0.010561056105610561, "grad_norm": 6.208090305328369, "learning_rate": 8.771929824561404e-07, "loss": 0.8832, "step": 2 }, { "epoch": 0.015841584158415842, "grad_norm": 7.966090202331543, "learning_rate": 1.3157894736842106e-06, "loss": 0.9749, "step": 3 }, { "epoch": 0.021122112211221122, "grad_norm": 7.089982986450195, "learning_rate": 1.7543859649122807e-06, "loss": 0.9166, "step": 4 }, { "epoch": 0.026402640264026403, "grad_norm": 9.171483039855957, "learning_rate": 2.1929824561403507e-06, "loss": 1.02, "step": 5 }, { "epoch": 0.031683168316831684, "grad_norm": 7.198622703552246, "learning_rate": 2.631578947368421e-06, "loss": 0.9205, "step": 6 }, { "epoch": 0.036963696369636964, "grad_norm": 6.673768997192383, "learning_rate": 3.070175438596491e-06, "loss": 0.9179, "step": 7 }, { "epoch": 0.042244224422442245, "grad_norm": 5.332284450531006, "learning_rate": 3.5087719298245615e-06, "loss": 0.7845, "step": 8 }, { "epoch": 0.047524752475247525, "grad_norm": 5.0864458084106445, "learning_rate": 3.9473684210526315e-06, "loss": 0.7676, "step": 9 }, { "epoch": 0.052805280528052806, "grad_norm": 4.446084022521973, "learning_rate": 4.3859649122807014e-06, "loss": 0.6778, "step": 10 }, { "epoch": 0.058085808580858087, "grad_norm": 4.680896759033203, "learning_rate": 4.824561403508772e-06, "loss": 0.6388, "step": 11 }, { "epoch": 0.06336633663366337, "grad_norm": 3.745345115661621, "learning_rate": 5.263157894736842e-06, "loss": 0.7111, "step": 12 }, { "epoch": 0.06864686468646865, "grad_norm": 3.1842403411865234, "learning_rate": 5.701754385964912e-06, "loss": 0.6327, "step": 13 }, { "epoch": 0.07392739273927393, "grad_norm": 2.8893682956695557, "learning_rate": 6.140350877192982e-06, "loss": 0.6334, "step": 14 }, { "epoch": 0.07920792079207921, "grad_norm": 2.5611019134521484, "learning_rate": 6.578947368421053e-06, "loss": 0.6284, "step": 15 }, { "epoch": 0.08448844884488449, "grad_norm": 2.8038105964660645, "learning_rate": 7.017543859649123e-06, "loss": 0.6224, "step": 16 }, { "epoch": 0.08976897689768977, "grad_norm": 2.51741099357605, "learning_rate": 7.456140350877193e-06, "loss": 0.6035, "step": 17 }, { "epoch": 0.09504950495049505, "grad_norm": 2.309906482696533, "learning_rate": 7.894736842105263e-06, "loss": 0.4823, "step": 18 }, { "epoch": 0.10033003300330033, "grad_norm": 2.719104528427124, "learning_rate": 8.333333333333334e-06, "loss": 0.5194, "step": 19 }, { "epoch": 0.10561056105610561, "grad_norm": 2.3248131275177, "learning_rate": 8.771929824561403e-06, "loss": 0.5342, "step": 20 }, { "epoch": 0.11089108910891089, "grad_norm": 2.2312052249908447, "learning_rate": 9.210526315789474e-06, "loss": 0.5275, "step": 21 }, { "epoch": 0.11617161716171617, "grad_norm": 2.045426368713379, "learning_rate": 9.649122807017545e-06, "loss": 0.5499, "step": 22 }, { "epoch": 0.12145214521452145, "grad_norm": 2.17621111869812, "learning_rate": 1.0087719298245614e-05, "loss": 0.5891, "step": 23 }, { "epoch": 0.12673267326732673, "grad_norm": 2.363523244857788, "learning_rate": 1.0526315789473684e-05, "loss": 0.667, "step": 24 }, { "epoch": 0.132013201320132, "grad_norm": 2.3387694358825684, "learning_rate": 1.0964912280701754e-05, "loss": 0.5842, "step": 25 }, { "epoch": 0.1372937293729373, "grad_norm": 2.1026556491851807, "learning_rate": 1.1403508771929824e-05, "loss": 0.6199, "step": 26 }, { "epoch": 0.14257425742574256, "grad_norm": 2.3579792976379395, "learning_rate": 1.1842105263157895e-05, "loss": 0.5985, "step": 27 }, { "epoch": 0.14785478547854786, "grad_norm": 2.136988639831543, "learning_rate": 1.2280701754385964e-05, "loss": 0.5759, "step": 28 }, { "epoch": 0.15313531353135312, "grad_norm": 1.9653126001358032, "learning_rate": 1.2719298245614037e-05, "loss": 0.5132, "step": 29 }, { "epoch": 0.15841584158415842, "grad_norm": 2.0811052322387695, "learning_rate": 1.3157894736842106e-05, "loss": 0.514, "step": 30 }, { "epoch": 0.16369636963696368, "grad_norm": 2.0110175609588623, "learning_rate": 1.3596491228070177e-05, "loss": 0.588, "step": 31 }, { "epoch": 0.16897689768976898, "grad_norm": 2.020909070968628, "learning_rate": 1.4035087719298246e-05, "loss": 0.5896, "step": 32 }, { "epoch": 0.17425742574257425, "grad_norm": 2.001084804534912, "learning_rate": 1.4473684210526317e-05, "loss": 0.5028, "step": 33 }, { "epoch": 0.17953795379537954, "grad_norm": 1.9398471117019653, "learning_rate": 1.4912280701754386e-05, "loss": 0.5199, "step": 34 }, { "epoch": 0.1848184818481848, "grad_norm": 2.3714287281036377, "learning_rate": 1.5350877192982457e-05, "loss": 0.6112, "step": 35 }, { "epoch": 0.1900990099009901, "grad_norm": 2.054084062576294, "learning_rate": 1.5789473684210526e-05, "loss": 0.4697, "step": 36 }, { "epoch": 0.19537953795379537, "grad_norm": 2.1482019424438477, "learning_rate": 1.62280701754386e-05, "loss": 0.442, "step": 37 }, { "epoch": 0.20066006600660066, "grad_norm": 2.005889892578125, "learning_rate": 1.6666666666666667e-05, "loss": 0.4276, "step": 38 }, { "epoch": 0.20594059405940593, "grad_norm": 2.1889655590057373, "learning_rate": 1.7105263157894737e-05, "loss": 0.5114, "step": 39 }, { "epoch": 0.21122112211221122, "grad_norm": 1.9033912420272827, "learning_rate": 1.7543859649122806e-05, "loss": 0.5266, "step": 40 }, { "epoch": 0.2165016501650165, "grad_norm": 2.01960825920105, "learning_rate": 1.7982456140350878e-05, "loss": 0.485, "step": 41 }, { "epoch": 0.22178217821782178, "grad_norm": 2.0285496711730957, "learning_rate": 1.8421052631578947e-05, "loss": 0.4915, "step": 42 }, { "epoch": 0.22706270627062705, "grad_norm": 1.9628126621246338, "learning_rate": 1.885964912280702e-05, "loss": 0.4059, "step": 43 }, { "epoch": 0.23234323432343235, "grad_norm": 2.2826972007751465, "learning_rate": 1.929824561403509e-05, "loss": 0.5476, "step": 44 }, { "epoch": 0.2376237623762376, "grad_norm": 2.3612606525421143, "learning_rate": 1.9736842105263158e-05, "loss": 0.5466, "step": 45 }, { "epoch": 0.2429042904290429, "grad_norm": 1.8648459911346436, "learning_rate": 2.0175438596491227e-05, "loss": 0.505, "step": 46 }, { "epoch": 0.24818481848184817, "grad_norm": 1.9400116205215454, "learning_rate": 2.06140350877193e-05, "loss": 0.4488, "step": 47 }, { "epoch": 0.25346534653465347, "grad_norm": 1.8791626691818237, "learning_rate": 2.105263157894737e-05, "loss": 0.4311, "step": 48 }, { "epoch": 0.25874587458745874, "grad_norm": 1.8579607009887695, "learning_rate": 2.149122807017544e-05, "loss": 0.4188, "step": 49 }, { "epoch": 0.264026402640264, "grad_norm": 1.8994585275650024, "learning_rate": 2.1929824561403507e-05, "loss": 0.4455, "step": 50 }, { "epoch": 0.2693069306930693, "grad_norm": 1.9908592700958252, "learning_rate": 2.236842105263158e-05, "loss": 0.4684, "step": 51 }, { "epoch": 0.2745874587458746, "grad_norm": 2.282810926437378, "learning_rate": 2.280701754385965e-05, "loss": 0.6328, "step": 52 }, { "epoch": 0.27986798679867986, "grad_norm": 2.017083168029785, "learning_rate": 2.324561403508772e-05, "loss": 0.4561, "step": 53 }, { "epoch": 0.2851485148514851, "grad_norm": 2.0290071964263916, "learning_rate": 2.368421052631579e-05, "loss": 0.4749, "step": 54 }, { "epoch": 0.29042904290429045, "grad_norm": 2.1724143028259277, "learning_rate": 2.412280701754386e-05, "loss": 0.6216, "step": 55 }, { "epoch": 0.2957095709570957, "grad_norm": 1.8445512056350708, "learning_rate": 2.456140350877193e-05, "loss": 0.4473, "step": 56 }, { "epoch": 0.300990099009901, "grad_norm": 1.9536579847335815, "learning_rate": 2.5e-05, "loss": 0.4746, "step": 57 }, { "epoch": 0.30627062706270625, "grad_norm": 2.0544443130493164, "learning_rate": 2.5438596491228074e-05, "loss": 0.4069, "step": 58 }, { "epoch": 0.31155115511551157, "grad_norm": 2.315953016281128, "learning_rate": 2.5877192982456143e-05, "loss": 0.4296, "step": 59 }, { "epoch": 0.31683168316831684, "grad_norm": 2.234273672103882, "learning_rate": 2.6315789473684212e-05, "loss": 0.4941, "step": 60 }, { "epoch": 0.3221122112211221, "grad_norm": 2.1946239471435547, "learning_rate": 2.675438596491228e-05, "loss": 0.4416, "step": 61 }, { "epoch": 0.32739273927392737, "grad_norm": 2.0075912475585938, "learning_rate": 2.7192982456140354e-05, "loss": 0.4202, "step": 62 }, { "epoch": 0.3326732673267327, "grad_norm": 4.797417640686035, "learning_rate": 2.7631578947368426e-05, "loss": 0.452, "step": 63 }, { "epoch": 0.33795379537953796, "grad_norm": 3.3362960815429688, "learning_rate": 2.8070175438596492e-05, "loss": 0.5405, "step": 64 }, { "epoch": 0.3432343234323432, "grad_norm": 1.9122159481048584, "learning_rate": 2.850877192982456e-05, "loss": 0.3242, "step": 65 }, { "epoch": 0.3485148514851485, "grad_norm": 1.8941394090652466, "learning_rate": 2.8947368421052634e-05, "loss": 0.4896, "step": 66 }, { "epoch": 0.3537953795379538, "grad_norm": 1.8021477460861206, "learning_rate": 2.9385964912280706e-05, "loss": 0.3874, "step": 67 }, { "epoch": 0.3590759075907591, "grad_norm": 2.036555767059326, "learning_rate": 2.9824561403508772e-05, "loss": 0.4886, "step": 68 }, { "epoch": 0.36435643564356435, "grad_norm": 2.1943323612213135, "learning_rate": 3.0263157894736844e-05, "loss": 0.421, "step": 69 }, { "epoch": 0.3696369636963696, "grad_norm": 2.077173948287964, "learning_rate": 3.0701754385964913e-05, "loss": 0.5161, "step": 70 }, { "epoch": 0.37491749174917494, "grad_norm": 2.0704095363616943, "learning_rate": 3.1140350877192986e-05, "loss": 0.4241, "step": 71 }, { "epoch": 0.3801980198019802, "grad_norm": 2.082000970840454, "learning_rate": 3.157894736842105e-05, "loss": 0.4373, "step": 72 }, { "epoch": 0.38547854785478547, "grad_norm": 1.8969218730926514, "learning_rate": 3.2017543859649124e-05, "loss": 0.4705, "step": 73 }, { "epoch": 0.39075907590759074, "grad_norm": 2.206298351287842, "learning_rate": 3.24561403508772e-05, "loss": 0.4938, "step": 74 }, { "epoch": 0.39603960396039606, "grad_norm": 2.0572750568389893, "learning_rate": 3.289473684210527e-05, "loss": 0.5007, "step": 75 }, { "epoch": 0.4013201320132013, "grad_norm": 1.94302237033844, "learning_rate": 3.3333333333333335e-05, "loss": 0.3689, "step": 76 }, { "epoch": 0.4066006600660066, "grad_norm": 2.0126149654388428, "learning_rate": 3.377192982456141e-05, "loss": 0.4881, "step": 77 }, { "epoch": 0.41188118811881186, "grad_norm": 1.887984037399292, "learning_rate": 3.421052631578947e-05, "loss": 0.4761, "step": 78 }, { "epoch": 0.4171617161716172, "grad_norm": 1.9020264148712158, "learning_rate": 3.4649122807017546e-05, "loss": 0.4343, "step": 79 }, { "epoch": 0.42244224422442245, "grad_norm": 1.942435622215271, "learning_rate": 3.508771929824561e-05, "loss": 0.4563, "step": 80 }, { "epoch": 0.4277227722772277, "grad_norm": 2.269737482070923, "learning_rate": 3.5526315789473684e-05, "loss": 0.4508, "step": 81 }, { "epoch": 0.433003300330033, "grad_norm": 2.0216665267944336, "learning_rate": 3.5964912280701756e-05, "loss": 0.4971, "step": 82 }, { "epoch": 0.4382838283828383, "grad_norm": 2.1765635013580322, "learning_rate": 3.640350877192983e-05, "loss": 0.452, "step": 83 }, { "epoch": 0.44356435643564357, "grad_norm": 2.25856876373291, "learning_rate": 3.6842105263157895e-05, "loss": 0.4824, "step": 84 }, { "epoch": 0.44884488448844884, "grad_norm": 2.2144601345062256, "learning_rate": 3.728070175438597e-05, "loss": 0.4479, "step": 85 }, { "epoch": 0.4541254125412541, "grad_norm": 1.97480309009552, "learning_rate": 3.771929824561404e-05, "loss": 0.3071, "step": 86 }, { "epoch": 0.4594059405940594, "grad_norm": 1.9352009296417236, "learning_rate": 3.815789473684211e-05, "loss": 0.4513, "step": 87 }, { "epoch": 0.4646864686468647, "grad_norm": 2.055535316467285, "learning_rate": 3.859649122807018e-05, "loss": 0.4508, "step": 88 }, { "epoch": 0.46996699669966996, "grad_norm": 1.93705415725708, "learning_rate": 3.9035087719298244e-05, "loss": 0.4083, "step": 89 }, { "epoch": 0.4752475247524752, "grad_norm": 1.9412288665771484, "learning_rate": 3.9473684210526316e-05, "loss": 0.3715, "step": 90 }, { "epoch": 0.48052805280528055, "grad_norm": 2.098421335220337, "learning_rate": 3.991228070175439e-05, "loss": 0.414, "step": 91 }, { "epoch": 0.4858085808580858, "grad_norm": 2.2177186012268066, "learning_rate": 4.0350877192982455e-05, "loss": 0.4736, "step": 92 }, { "epoch": 0.4910891089108911, "grad_norm": 2.056107759475708, "learning_rate": 4.078947368421053e-05, "loss": 0.4177, "step": 93 }, { "epoch": 0.49636963696369635, "grad_norm": 1.9581352472305298, "learning_rate": 4.12280701754386e-05, "loss": 0.3688, "step": 94 }, { "epoch": 0.5016501650165016, "grad_norm": 2.2061662673950195, "learning_rate": 4.166666666666667e-05, "loss": 0.4705, "step": 95 }, { "epoch": 0.5069306930693069, "grad_norm": 1.7467211484909058, "learning_rate": 4.210526315789474e-05, "loss": 0.3303, "step": 96 }, { "epoch": 0.5122112211221123, "grad_norm": 1.9702417850494385, "learning_rate": 4.254385964912281e-05, "loss": 0.5031, "step": 97 }, { "epoch": 0.5174917491749175, "grad_norm": 2.079378604888916, "learning_rate": 4.298245614035088e-05, "loss": 0.456, "step": 98 }, { "epoch": 0.5227722772277228, "grad_norm": 1.8181231021881104, "learning_rate": 4.342105263157895e-05, "loss": 0.4231, "step": 99 }, { "epoch": 0.528052805280528, "grad_norm": 2.1575117111206055, "learning_rate": 4.3859649122807014e-05, "loss": 0.457, "step": 100 }, { "epoch": 0.5333333333333333, "grad_norm": 2.3540713787078857, "learning_rate": 4.429824561403509e-05, "loss": 0.397, "step": 101 }, { "epoch": 0.5386138613861386, "grad_norm": 2.3277106285095215, "learning_rate": 4.473684210526316e-05, "loss": 0.474, "step": 102 }, { "epoch": 0.5438943894389439, "grad_norm": 2.0837771892547607, "learning_rate": 4.517543859649123e-05, "loss": 0.4911, "step": 103 }, { "epoch": 0.5491749174917492, "grad_norm": 2.315387487411499, "learning_rate": 4.56140350877193e-05, "loss": 0.5138, "step": 104 }, { "epoch": 0.5544554455445545, "grad_norm": 1.9372241497039795, "learning_rate": 4.605263157894737e-05, "loss": 0.402, "step": 105 }, { "epoch": 0.5597359735973597, "grad_norm": 2.0722286701202393, "learning_rate": 4.649122807017544e-05, "loss": 0.3484, "step": 106 }, { "epoch": 0.565016501650165, "grad_norm": 1.8825434446334839, "learning_rate": 4.6929824561403515e-05, "loss": 0.4057, "step": 107 }, { "epoch": 0.5702970297029702, "grad_norm": 2.0918331146240234, "learning_rate": 4.736842105263158e-05, "loss": 0.4073, "step": 108 }, { "epoch": 0.5755775577557756, "grad_norm": 2.246974468231201, "learning_rate": 4.780701754385965e-05, "loss": 0.4104, "step": 109 }, { "epoch": 0.5808580858085809, "grad_norm": 1.8505111932754517, "learning_rate": 4.824561403508772e-05, "loss": 0.351, "step": 110 }, { "epoch": 0.5861386138613861, "grad_norm": 2.2233192920684814, "learning_rate": 4.868421052631579e-05, "loss": 0.4916, "step": 111 }, { "epoch": 0.5914191419141914, "grad_norm": 2.3233530521392822, "learning_rate": 4.912280701754386e-05, "loss": 0.4886, "step": 112 }, { "epoch": 0.5966996699669967, "grad_norm": 2.298288345336914, "learning_rate": 4.956140350877193e-05, "loss": 0.497, "step": 113 }, { "epoch": 0.601980198019802, "grad_norm": 1.9848483800888062, "learning_rate": 5e-05, "loss": 0.3982, "step": 114 }, { "epoch": 0.6072607260726073, "grad_norm": 1.7758945226669312, "learning_rate": 4.995098039215686e-05, "loss": 0.4284, "step": 115 }, { "epoch": 0.6125412541254125, "grad_norm": 2.1073226928710938, "learning_rate": 4.990196078431373e-05, "loss": 0.4741, "step": 116 }, { "epoch": 0.6178217821782178, "grad_norm": 2.1385958194732666, "learning_rate": 4.985294117647059e-05, "loss": 0.3853, "step": 117 }, { "epoch": 0.6231023102310231, "grad_norm": 2.053973436355591, "learning_rate": 4.980392156862745e-05, "loss": 0.4385, "step": 118 }, { "epoch": 0.6283828382838283, "grad_norm": 1.7011091709136963, "learning_rate": 4.975490196078432e-05, "loss": 0.3604, "step": 119 }, { "epoch": 0.6336633663366337, "grad_norm": 2.1312694549560547, "learning_rate": 4.970588235294118e-05, "loss": 0.4636, "step": 120 }, { "epoch": 0.638943894389439, "grad_norm": 1.9020744562149048, "learning_rate": 4.9656862745098046e-05, "loss": 0.3795, "step": 121 }, { "epoch": 0.6442244224422442, "grad_norm": 1.9798043966293335, "learning_rate": 4.960784313725491e-05, "loss": 0.382, "step": 122 }, { "epoch": 0.6495049504950495, "grad_norm": 1.8981302976608276, "learning_rate": 4.955882352941177e-05, "loss": 0.4038, "step": 123 }, { "epoch": 0.6547854785478547, "grad_norm": 1.9499566555023193, "learning_rate": 4.9509803921568634e-05, "loss": 0.434, "step": 124 }, { "epoch": 0.6600660066006601, "grad_norm": 1.912457823753357, "learning_rate": 4.9460784313725495e-05, "loss": 0.3187, "step": 125 }, { "epoch": 0.6653465346534654, "grad_norm": 2.0483133792877197, "learning_rate": 4.9411764705882355e-05, "loss": 0.3916, "step": 126 }, { "epoch": 0.6706270627062706, "grad_norm": 1.924310564994812, "learning_rate": 4.936274509803922e-05, "loss": 0.378, "step": 127 }, { "epoch": 0.6759075907590759, "grad_norm": 2.1889538764953613, "learning_rate": 4.931372549019608e-05, "loss": 0.4093, "step": 128 }, { "epoch": 0.6811881188118812, "grad_norm": 1.8973898887634277, "learning_rate": 4.9264705882352944e-05, "loss": 0.3571, "step": 129 }, { "epoch": 0.6864686468646864, "grad_norm": 2.0125250816345215, "learning_rate": 4.9215686274509804e-05, "loss": 0.359, "step": 130 }, { "epoch": 0.6917491749174918, "grad_norm": 1.9622538089752197, "learning_rate": 4.9166666666666665e-05, "loss": 0.3238, "step": 131 }, { "epoch": 0.697029702970297, "grad_norm": 1.7294894456863403, "learning_rate": 4.911764705882353e-05, "loss": 0.3555, "step": 132 }, { "epoch": 0.7023102310231023, "grad_norm": 2.0299930572509766, "learning_rate": 4.906862745098039e-05, "loss": 0.4667, "step": 133 }, { "epoch": 0.7075907590759076, "grad_norm": 1.813370704650879, "learning_rate": 4.901960784313725e-05, "loss": 0.3004, "step": 134 }, { "epoch": 0.7128712871287128, "grad_norm": 2.090129852294922, "learning_rate": 4.897058823529412e-05, "loss": 0.3845, "step": 135 }, { "epoch": 0.7181518151815182, "grad_norm": 2.3778114318847656, "learning_rate": 4.892156862745098e-05, "loss": 0.4731, "step": 136 }, { "epoch": 0.7234323432343235, "grad_norm": 2.1827681064605713, "learning_rate": 4.887254901960784e-05, "loss": 0.4078, "step": 137 }, { "epoch": 0.7287128712871287, "grad_norm": 2.18556547164917, "learning_rate": 4.882352941176471e-05, "loss": 0.4042, "step": 138 }, { "epoch": 0.733993399339934, "grad_norm": 1.9759682416915894, "learning_rate": 4.877450980392157e-05, "loss": 0.3694, "step": 139 }, { "epoch": 0.7392739273927392, "grad_norm": 1.8205828666687012, "learning_rate": 4.872549019607843e-05, "loss": 0.244, "step": 140 }, { "epoch": 0.7445544554455445, "grad_norm": 1.8210268020629883, "learning_rate": 4.86764705882353e-05, "loss": 0.4084, "step": 141 }, { "epoch": 0.7498349834983499, "grad_norm": 2.2197041511535645, "learning_rate": 4.862745098039216e-05, "loss": 0.5236, "step": 142 }, { "epoch": 0.7551155115511551, "grad_norm": 2.137676239013672, "learning_rate": 4.8578431372549024e-05, "loss": 0.3302, "step": 143 }, { "epoch": 0.7603960396039604, "grad_norm": 2.126865863800049, "learning_rate": 4.8529411764705885e-05, "loss": 0.4265, "step": 144 }, { "epoch": 0.7656765676567657, "grad_norm": 1.828809380531311, "learning_rate": 4.8480392156862745e-05, "loss": 0.4363, "step": 145 }, { "epoch": 0.7709570957095709, "grad_norm": 1.7918983697891235, "learning_rate": 4.843137254901961e-05, "loss": 0.3568, "step": 146 }, { "epoch": 0.7762376237623763, "grad_norm": 1.7884886264801025, "learning_rate": 4.838235294117647e-05, "loss": 0.3338, "step": 147 }, { "epoch": 0.7815181518151815, "grad_norm": 1.8494501113891602, "learning_rate": 4.8333333333333334e-05, "loss": 0.3311, "step": 148 }, { "epoch": 0.7867986798679868, "grad_norm": 2.0265438556671143, "learning_rate": 4.82843137254902e-05, "loss": 0.4381, "step": 149 }, { "epoch": 0.7920792079207921, "grad_norm": 2.0035383701324463, "learning_rate": 4.823529411764706e-05, "loss": 0.4291, "step": 150 }, { "epoch": 0.7973597359735973, "grad_norm": 1.8268975019454956, "learning_rate": 4.818627450980392e-05, "loss": 0.464, "step": 151 }, { "epoch": 0.8026402640264027, "grad_norm": 1.9028264284133911, "learning_rate": 4.813725490196079e-05, "loss": 0.3807, "step": 152 }, { "epoch": 0.807920792079208, "grad_norm": 1.8994662761688232, "learning_rate": 4.808823529411765e-05, "loss": 0.3904, "step": 153 }, { "epoch": 0.8132013201320132, "grad_norm": 1.832381248474121, "learning_rate": 4.803921568627452e-05, "loss": 0.3856, "step": 154 }, { "epoch": 0.8184818481848185, "grad_norm": 1.877752661705017, "learning_rate": 4.799019607843138e-05, "loss": 0.3185, "step": 155 }, { "epoch": 0.8237623762376237, "grad_norm": 2.0872726440429688, "learning_rate": 4.794117647058824e-05, "loss": 0.3217, "step": 156 }, { "epoch": 0.829042904290429, "grad_norm": 1.8779263496398926, "learning_rate": 4.7892156862745105e-05, "loss": 0.3654, "step": 157 }, { "epoch": 0.8343234323432344, "grad_norm": 1.996422529220581, "learning_rate": 4.7843137254901966e-05, "loss": 0.3524, "step": 158 }, { "epoch": 0.8396039603960396, "grad_norm": 1.5910488367080688, "learning_rate": 4.7794117647058826e-05, "loss": 0.2859, "step": 159 }, { "epoch": 0.8448844884488449, "grad_norm": 1.9424618482589722, "learning_rate": 4.774509803921569e-05, "loss": 0.3422, "step": 160 }, { "epoch": 0.8501650165016502, "grad_norm": 1.9187934398651123, "learning_rate": 4.7696078431372554e-05, "loss": 0.3807, "step": 161 }, { "epoch": 0.8554455445544554, "grad_norm": 1.7809456586837769, "learning_rate": 4.7647058823529414e-05, "loss": 0.2991, "step": 162 }, { "epoch": 0.8607260726072608, "grad_norm": 1.9575221538543701, "learning_rate": 4.7598039215686275e-05, "loss": 0.4688, "step": 163 }, { "epoch": 0.866006600660066, "grad_norm": 1.8438433408737183, "learning_rate": 4.7549019607843135e-05, "loss": 0.3691, "step": 164 }, { "epoch": 0.8712871287128713, "grad_norm": 1.9522879123687744, "learning_rate": 4.75e-05, "loss": 0.3987, "step": 165 }, { "epoch": 0.8765676567656766, "grad_norm": 2.243354320526123, "learning_rate": 4.745098039215686e-05, "loss": 0.4836, "step": 166 }, { "epoch": 0.8818481848184818, "grad_norm": 1.8762164115905762, "learning_rate": 4.7401960784313724e-05, "loss": 0.4691, "step": 167 }, { "epoch": 0.8871287128712871, "grad_norm": 1.8055609464645386, "learning_rate": 4.735294117647059e-05, "loss": 0.2506, "step": 168 }, { "epoch": 0.8924092409240925, "grad_norm": 1.8521029949188232, "learning_rate": 4.730392156862745e-05, "loss": 0.3291, "step": 169 }, { "epoch": 0.8976897689768977, "grad_norm": 1.939030647277832, "learning_rate": 4.725490196078431e-05, "loss": 0.4104, "step": 170 }, { "epoch": 0.902970297029703, "grad_norm": 1.8853607177734375, "learning_rate": 4.720588235294118e-05, "loss": 0.3797, "step": 171 }, { "epoch": 0.9082508250825082, "grad_norm": 2.0953316688537598, "learning_rate": 4.715686274509804e-05, "loss": 0.324, "step": 172 }, { "epoch": 0.9135313531353135, "grad_norm": 1.9342799186706543, "learning_rate": 4.71078431372549e-05, "loss": 0.4306, "step": 173 }, { "epoch": 0.9188118811881189, "grad_norm": 1.8248006105422974, "learning_rate": 4.705882352941177e-05, "loss": 0.3388, "step": 174 }, { "epoch": 0.9240924092409241, "grad_norm": 1.9689913988113403, "learning_rate": 4.700980392156863e-05, "loss": 0.3591, "step": 175 }, { "epoch": 0.9293729372937294, "grad_norm": 2.017063856124878, "learning_rate": 4.6960784313725495e-05, "loss": 0.4723, "step": 176 }, { "epoch": 0.9346534653465347, "grad_norm": 1.9692254066467285, "learning_rate": 4.6911764705882356e-05, "loss": 0.3893, "step": 177 }, { "epoch": 0.9399339933993399, "grad_norm": 1.9935567378997803, "learning_rate": 4.6862745098039216e-05, "loss": 0.3938, "step": 178 }, { "epoch": 0.9452145214521452, "grad_norm": 1.7153037786483765, "learning_rate": 4.681372549019608e-05, "loss": 0.2368, "step": 179 }, { "epoch": 0.9504950495049505, "grad_norm": 1.4944133758544922, "learning_rate": 4.6764705882352944e-05, "loss": 0.2181, "step": 180 }, { "epoch": 0.9557755775577558, "grad_norm": 1.9143524169921875, "learning_rate": 4.6715686274509804e-05, "loss": 0.413, "step": 181 }, { "epoch": 0.9610561056105611, "grad_norm": 2.162576675415039, "learning_rate": 4.666666666666667e-05, "loss": 0.3745, "step": 182 }, { "epoch": 0.9663366336633663, "grad_norm": 1.8236726522445679, "learning_rate": 4.661764705882353e-05, "loss": 0.3056, "step": 183 }, { "epoch": 0.9716171617161716, "grad_norm": 1.9680614471435547, "learning_rate": 4.656862745098039e-05, "loss": 0.3875, "step": 184 }, { "epoch": 0.976897689768977, "grad_norm": 1.575900912284851, "learning_rate": 4.651960784313726e-05, "loss": 0.1831, "step": 185 }, { "epoch": 0.9821782178217822, "grad_norm": 2.6015613079071045, "learning_rate": 4.647058823529412e-05, "loss": 0.4227, "step": 186 }, { "epoch": 0.9874587458745875, "grad_norm": 2.066946268081665, "learning_rate": 4.642156862745098e-05, "loss": 0.4256, "step": 187 }, { "epoch": 0.9927392739273927, "grad_norm": 2.1683449745178223, "learning_rate": 4.637254901960785e-05, "loss": 0.3943, "step": 188 }, { "epoch": 0.998019801980198, "grad_norm": 1.876991629600525, "learning_rate": 4.632352941176471e-05, "loss": 0.4049, "step": 189 }, { "epoch": 1.0033003300330032, "grad_norm": 2.2598772048950195, "learning_rate": 4.6274509803921576e-05, "loss": 0.3749, "step": 190 }, { "epoch": 1.0085808580858087, "grad_norm": 1.8292592763900757, "learning_rate": 4.6225490196078436e-05, "loss": 0.2654, "step": 191 }, { "epoch": 1.0138613861386139, "grad_norm": 2.6986138820648193, "learning_rate": 4.61764705882353e-05, "loss": 0.3662, "step": 192 }, { "epoch": 1.019141914191419, "grad_norm": 1.9084346294403076, "learning_rate": 4.6127450980392164e-05, "loss": 0.2929, "step": 193 }, { "epoch": 1.0244224422442245, "grad_norm": 1.6963775157928467, "learning_rate": 4.607843137254902e-05, "loss": 0.2481, "step": 194 }, { "epoch": 1.0297029702970297, "grad_norm": 2.524332046508789, "learning_rate": 4.6029411764705885e-05, "loss": 0.2213, "step": 195 }, { "epoch": 1.034983498349835, "grad_norm": 1.945142388343811, "learning_rate": 4.5980392156862746e-05, "loss": 0.2892, "step": 196 }, { "epoch": 1.0402640264026402, "grad_norm": 2.3402678966522217, "learning_rate": 4.5931372549019606e-05, "loss": 0.2232, "step": 197 }, { "epoch": 1.0455445544554456, "grad_norm": 1.7755571603775024, "learning_rate": 4.588235294117647e-05, "loss": 0.2416, "step": 198 }, { "epoch": 1.0508250825082508, "grad_norm": 2.110517740249634, "learning_rate": 4.5833333333333334e-05, "loss": 0.2557, "step": 199 }, { "epoch": 1.056105610561056, "grad_norm": 1.7219949960708618, "learning_rate": 4.5784313725490194e-05, "loss": 0.2274, "step": 200 }, { "epoch": 1.0613861386138614, "grad_norm": 1.9697656631469727, "learning_rate": 4.573529411764706e-05, "loss": 0.2588, "step": 201 }, { "epoch": 1.0666666666666667, "grad_norm": 1.5107744932174683, "learning_rate": 4.568627450980392e-05, "loss": 0.2175, "step": 202 }, { "epoch": 1.0719471947194719, "grad_norm": 1.8557658195495605, "learning_rate": 4.563725490196078e-05, "loss": 0.2901, "step": 203 }, { "epoch": 1.0772277227722773, "grad_norm": 1.764145851135254, "learning_rate": 4.558823529411765e-05, "loss": 0.176, "step": 204 }, { "epoch": 1.0825082508250825, "grad_norm": 1.8334012031555176, "learning_rate": 4.553921568627451e-05, "loss": 0.2572, "step": 205 }, { "epoch": 1.0877887788778877, "grad_norm": 1.8206666707992554, "learning_rate": 4.549019607843137e-05, "loss": 0.248, "step": 206 }, { "epoch": 1.0930693069306932, "grad_norm": 2.0776381492614746, "learning_rate": 4.544117647058824e-05, "loss": 0.3113, "step": 207 }, { "epoch": 1.0983498349834984, "grad_norm": 1.8429386615753174, "learning_rate": 4.53921568627451e-05, "loss": 0.2849, "step": 208 }, { "epoch": 1.1036303630363036, "grad_norm": 1.7320504188537598, "learning_rate": 4.5343137254901966e-05, "loss": 0.2145, "step": 209 }, { "epoch": 1.108910891089109, "grad_norm": 1.7207646369934082, "learning_rate": 4.5294117647058826e-05, "loss": 0.2013, "step": 210 }, { "epoch": 1.1141914191419142, "grad_norm": 2.134873151779175, "learning_rate": 4.524509803921569e-05, "loss": 0.3293, "step": 211 }, { "epoch": 1.1194719471947194, "grad_norm": 1.7931280136108398, "learning_rate": 4.5196078431372554e-05, "loss": 0.2292, "step": 212 }, { "epoch": 1.1247524752475249, "grad_norm": 2.1878650188446045, "learning_rate": 4.5147058823529415e-05, "loss": 0.3173, "step": 213 }, { "epoch": 1.13003300330033, "grad_norm": 1.8994349241256714, "learning_rate": 4.5098039215686275e-05, "loss": 0.2742, "step": 214 }, { "epoch": 1.1353135313531353, "grad_norm": 1.9204659461975098, "learning_rate": 4.504901960784314e-05, "loss": 0.2713, "step": 215 }, { "epoch": 1.1405940594059407, "grad_norm": 1.9061977863311768, "learning_rate": 4.5e-05, "loss": 0.2433, "step": 216 }, { "epoch": 1.145874587458746, "grad_norm": 2.254232168197632, "learning_rate": 4.495098039215686e-05, "loss": 0.3523, "step": 217 }, { "epoch": 1.1511551155115511, "grad_norm": 1.9071446657180786, "learning_rate": 4.490196078431373e-05, "loss": 0.2196, "step": 218 }, { "epoch": 1.1564356435643564, "grad_norm": 1.8089710474014282, "learning_rate": 4.485294117647059e-05, "loss": 0.2907, "step": 219 }, { "epoch": 1.1617161716171618, "grad_norm": 1.9056932926177979, "learning_rate": 4.480392156862745e-05, "loss": 0.2564, "step": 220 }, { "epoch": 1.166996699669967, "grad_norm": 1.9336401224136353, "learning_rate": 4.475490196078432e-05, "loss": 0.205, "step": 221 }, { "epoch": 1.1722772277227722, "grad_norm": 1.9126192331314087, "learning_rate": 4.470588235294118e-05, "loss": 0.2647, "step": 222 }, { "epoch": 1.1775577557755776, "grad_norm": 1.8508714437484741, "learning_rate": 4.4656862745098047e-05, "loss": 0.262, "step": 223 }, { "epoch": 1.1828382838283829, "grad_norm": 1.478278398513794, "learning_rate": 4.460784313725491e-05, "loss": 0.1711, "step": 224 }, { "epoch": 1.188118811881188, "grad_norm": 1.7818694114685059, "learning_rate": 4.455882352941177e-05, "loss": 0.2043, "step": 225 }, { "epoch": 1.1933993399339933, "grad_norm": 1.916344404220581, "learning_rate": 4.450980392156863e-05, "loss": 0.3064, "step": 226 }, { "epoch": 1.1986798679867987, "grad_norm": 1.8637932538986206, "learning_rate": 4.446078431372549e-05, "loss": 0.3247, "step": 227 }, { "epoch": 1.203960396039604, "grad_norm": 1.7257781028747559, "learning_rate": 4.4411764705882356e-05, "loss": 0.1981, "step": 228 }, { "epoch": 1.2092409240924091, "grad_norm": 1.9121214151382446, "learning_rate": 4.4362745098039216e-05, "loss": 0.2963, "step": 229 }, { "epoch": 1.2145214521452146, "grad_norm": 1.5968807935714722, "learning_rate": 4.431372549019608e-05, "loss": 0.1802, "step": 230 }, { "epoch": 1.2198019801980198, "grad_norm": 1.7256313562393188, "learning_rate": 4.4264705882352944e-05, "loss": 0.2442, "step": 231 }, { "epoch": 1.225082508250825, "grad_norm": 1.7865731716156006, "learning_rate": 4.4215686274509805e-05, "loss": 0.2683, "step": 232 }, { "epoch": 1.2303630363036304, "grad_norm": 1.7361854314804077, "learning_rate": 4.4166666666666665e-05, "loss": 0.2285, "step": 233 }, { "epoch": 1.2356435643564356, "grad_norm": 1.8758944272994995, "learning_rate": 4.411764705882353e-05, "loss": 0.3081, "step": 234 }, { "epoch": 1.2409240924092408, "grad_norm": 2.000033140182495, "learning_rate": 4.406862745098039e-05, "loss": 0.2405, "step": 235 }, { "epoch": 1.2462046204620463, "grad_norm": 1.8750522136688232, "learning_rate": 4.401960784313725e-05, "loss": 0.2778, "step": 236 }, { "epoch": 1.2514851485148515, "grad_norm": 1.7535063028335571, "learning_rate": 4.397058823529412e-05, "loss": 0.2045, "step": 237 }, { "epoch": 1.2567656765676567, "grad_norm": 1.9849064350128174, "learning_rate": 4.392156862745098e-05, "loss": 0.3018, "step": 238 }, { "epoch": 1.2620462046204621, "grad_norm": 1.8400393724441528, "learning_rate": 4.387254901960784e-05, "loss": 0.2555, "step": 239 }, { "epoch": 1.2673267326732673, "grad_norm": 1.8575385808944702, "learning_rate": 4.382352941176471e-05, "loss": 0.2783, "step": 240 }, { "epoch": 1.2726072607260726, "grad_norm": 1.8452024459838867, "learning_rate": 4.377450980392157e-05, "loss": 0.3091, "step": 241 }, { "epoch": 1.277887788778878, "grad_norm": 1.9682793617248535, "learning_rate": 4.3725490196078437e-05, "loss": 0.2169, "step": 242 }, { "epoch": 1.2831683168316832, "grad_norm": 1.7926579713821411, "learning_rate": 4.36764705882353e-05, "loss": 0.27, "step": 243 }, { "epoch": 1.2884488448844884, "grad_norm": 1.690425157546997, "learning_rate": 4.362745098039216e-05, "loss": 0.2336, "step": 244 }, { "epoch": 1.2937293729372938, "grad_norm": 1.775240421295166, "learning_rate": 4.3578431372549025e-05, "loss": 0.1981, "step": 245 }, { "epoch": 1.299009900990099, "grad_norm": 1.7951467037200928, "learning_rate": 4.3529411764705885e-05, "loss": 0.2195, "step": 246 }, { "epoch": 1.3042904290429043, "grad_norm": 1.5444797277450562, "learning_rate": 4.3480392156862746e-05, "loss": 0.2264, "step": 247 }, { "epoch": 1.3095709570957097, "grad_norm": 1.731550931930542, "learning_rate": 4.343137254901961e-05, "loss": 0.1676, "step": 248 }, { "epoch": 1.314851485148515, "grad_norm": 1.747083306312561, "learning_rate": 4.3382352941176474e-05, "loss": 0.28, "step": 249 }, { "epoch": 1.3201320132013201, "grad_norm": 1.7961376905441284, "learning_rate": 4.3333333333333334e-05, "loss": 0.2003, "step": 250 }, { "epoch": 1.3254125412541253, "grad_norm": 1.8366891145706177, "learning_rate": 4.32843137254902e-05, "loss": 0.2093, "step": 251 }, { "epoch": 1.3306930693069308, "grad_norm": 1.4352390766143799, "learning_rate": 4.323529411764706e-05, "loss": 0.1597, "step": 252 }, { "epoch": 1.335973597359736, "grad_norm": 1.9146888256072998, "learning_rate": 4.318627450980392e-05, "loss": 0.2425, "step": 253 }, { "epoch": 1.3412541254125412, "grad_norm": 1.4383189678192139, "learning_rate": 4.313725490196079e-05, "loss": 0.1943, "step": 254 }, { "epoch": 1.3465346534653464, "grad_norm": 1.5246001482009888, "learning_rate": 4.308823529411765e-05, "loss": 0.1543, "step": 255 }, { "epoch": 1.3518151815181518, "grad_norm": 1.4863159656524658, "learning_rate": 4.303921568627452e-05, "loss": 0.2441, "step": 256 }, { "epoch": 1.357095709570957, "grad_norm": 1.6645705699920654, "learning_rate": 4.299019607843138e-05, "loss": 0.2329, "step": 257 }, { "epoch": 1.3623762376237623, "grad_norm": 1.946554183959961, "learning_rate": 4.294117647058823e-05, "loss": 0.1952, "step": 258 }, { "epoch": 1.3676567656765677, "grad_norm": 2.0372443199157715, "learning_rate": 4.28921568627451e-05, "loss": 0.332, "step": 259 }, { "epoch": 1.372937293729373, "grad_norm": 1.846138834953308, "learning_rate": 4.284313725490196e-05, "loss": 0.1408, "step": 260 }, { "epoch": 1.378217821782178, "grad_norm": 1.5724695920944214, "learning_rate": 4.2794117647058827e-05, "loss": 0.1926, "step": 261 }, { "epoch": 1.3834983498349835, "grad_norm": 2.1506614685058594, "learning_rate": 4.274509803921569e-05, "loss": 0.217, "step": 262 }, { "epoch": 1.3887788778877888, "grad_norm": 2.0763325691223145, "learning_rate": 4.269607843137255e-05, "loss": 0.2871, "step": 263 }, { "epoch": 1.394059405940594, "grad_norm": 1.9296153783798218, "learning_rate": 4.2647058823529415e-05, "loss": 0.2655, "step": 264 }, { "epoch": 1.3993399339933994, "grad_norm": 1.7979801893234253, "learning_rate": 4.2598039215686275e-05, "loss": 0.2715, "step": 265 }, { "epoch": 1.4046204620462046, "grad_norm": 1.4527943134307861, "learning_rate": 4.2549019607843136e-05, "loss": 0.203, "step": 266 }, { "epoch": 1.4099009900990098, "grad_norm": 1.8454203605651855, "learning_rate": 4.25e-05, "loss": 0.198, "step": 267 }, { "epoch": 1.4151815181518153, "grad_norm": 1.6438169479370117, "learning_rate": 4.2450980392156864e-05, "loss": 0.2056, "step": 268 }, { "epoch": 1.4204620462046205, "grad_norm": 1.5819754600524902, "learning_rate": 4.2401960784313724e-05, "loss": 0.2154, "step": 269 }, { "epoch": 1.4257425742574257, "grad_norm": 1.800973653793335, "learning_rate": 4.235294117647059e-05, "loss": 0.2536, "step": 270 }, { "epoch": 1.431023102310231, "grad_norm": 1.6425402164459229, "learning_rate": 4.230392156862745e-05, "loss": 0.2111, "step": 271 }, { "epoch": 1.4363036303630363, "grad_norm": 1.908632755279541, "learning_rate": 4.225490196078431e-05, "loss": 0.2822, "step": 272 }, { "epoch": 1.4415841584158415, "grad_norm": 2.028026580810547, "learning_rate": 4.220588235294118e-05, "loss": 0.3118, "step": 273 }, { "epoch": 1.446864686468647, "grad_norm": 1.9891923666000366, "learning_rate": 4.215686274509804e-05, "loss": 0.2469, "step": 274 }, { "epoch": 1.4521452145214522, "grad_norm": 1.771210789680481, "learning_rate": 4.210784313725491e-05, "loss": 0.2057, "step": 275 }, { "epoch": 1.4574257425742574, "grad_norm": 1.8335461616516113, "learning_rate": 4.205882352941177e-05, "loss": 0.203, "step": 276 }, { "epoch": 1.4627062706270628, "grad_norm": 1.4403390884399414, "learning_rate": 4.200980392156863e-05, "loss": 0.1652, "step": 277 }, { "epoch": 1.467986798679868, "grad_norm": 2.0193352699279785, "learning_rate": 4.1960784313725496e-05, "loss": 0.2914, "step": 278 }, { "epoch": 1.4732673267326732, "grad_norm": 1.5873808860778809, "learning_rate": 4.1911764705882356e-05, "loss": 0.1871, "step": 279 }, { "epoch": 1.4785478547854787, "grad_norm": 2.0619425773620605, "learning_rate": 4.1862745098039217e-05, "loss": 0.2449, "step": 280 }, { "epoch": 1.4838283828382839, "grad_norm": 1.777978539466858, "learning_rate": 4.1813725490196084e-05, "loss": 0.2585, "step": 281 }, { "epoch": 1.489108910891089, "grad_norm": 1.7166889905929565, "learning_rate": 4.1764705882352944e-05, "loss": 0.2648, "step": 282 }, { "epoch": 1.4943894389438943, "grad_norm": 1.7009400129318237, "learning_rate": 4.1715686274509805e-05, "loss": 0.1847, "step": 283 }, { "epoch": 1.4996699669966997, "grad_norm": 2.067512035369873, "learning_rate": 4.166666666666667e-05, "loss": 0.2803, "step": 284 }, { "epoch": 1.504950495049505, "grad_norm": 1.6885766983032227, "learning_rate": 4.161764705882353e-05, "loss": 0.2037, "step": 285 }, { "epoch": 1.5102310231023104, "grad_norm": 1.9202988147735596, "learning_rate": 4.156862745098039e-05, "loss": 0.2791, "step": 286 }, { "epoch": 1.5155115511551154, "grad_norm": 1.6683584451675415, "learning_rate": 4.151960784313726e-05, "loss": 0.2228, "step": 287 }, { "epoch": 1.5207920792079208, "grad_norm": 1.783361792564392, "learning_rate": 4.147058823529412e-05, "loss": 0.2524, "step": 288 }, { "epoch": 1.526072607260726, "grad_norm": 1.9562329053878784, "learning_rate": 4.142156862745099e-05, "loss": 0.2773, "step": 289 }, { "epoch": 1.5313531353135312, "grad_norm": 1.4846049547195435, "learning_rate": 4.137254901960784e-05, "loss": 0.2528, "step": 290 }, { "epoch": 1.5366336633663367, "grad_norm": 1.6068270206451416, "learning_rate": 4.13235294117647e-05, "loss": 0.1513, "step": 291 }, { "epoch": 1.5419141914191419, "grad_norm": 1.8563951253890991, "learning_rate": 4.127450980392157e-05, "loss": 0.2518, "step": 292 }, { "epoch": 1.547194719471947, "grad_norm": 1.9122540950775146, "learning_rate": 4.122549019607843e-05, "loss": 0.2235, "step": 293 }, { "epoch": 1.5524752475247525, "grad_norm": 1.619687557220459, "learning_rate": 4.11764705882353e-05, "loss": 0.236, "step": 294 }, { "epoch": 1.5577557755775577, "grad_norm": 1.9006292819976807, "learning_rate": 4.112745098039216e-05, "loss": 0.2688, "step": 295 }, { "epoch": 1.563036303630363, "grad_norm": 1.5912319421768188, "learning_rate": 4.107843137254902e-05, "loss": 0.2036, "step": 296 }, { "epoch": 1.5683168316831684, "grad_norm": 1.9365366697311401, "learning_rate": 4.1029411764705886e-05, "loss": 0.2708, "step": 297 }, { "epoch": 1.5735973597359736, "grad_norm": 1.535831332206726, "learning_rate": 4.0980392156862746e-05, "loss": 0.1875, "step": 298 }, { "epoch": 1.5788778877887788, "grad_norm": 2.117027997970581, "learning_rate": 4.0931372549019607e-05, "loss": 0.3118, "step": 299 }, { "epoch": 1.5841584158415842, "grad_norm": 1.8837215900421143, "learning_rate": 4.0882352941176474e-05, "loss": 0.2084, "step": 300 }, { "epoch": 1.5894389438943894, "grad_norm": 1.502886176109314, "learning_rate": 4.0833333333333334e-05, "loss": 0.1348, "step": 301 }, { "epoch": 1.5947194719471947, "grad_norm": 1.6588914394378662, "learning_rate": 4.0784313725490195e-05, "loss": 0.1775, "step": 302 }, { "epoch": 1.6, "grad_norm": 1.6253366470336914, "learning_rate": 4.073529411764706e-05, "loss": 0.2719, "step": 303 }, { "epoch": 1.6052805280528053, "grad_norm": 1.9440994262695312, "learning_rate": 4.068627450980392e-05, "loss": 0.2278, "step": 304 }, { "epoch": 1.6105610561056105, "grad_norm": 1.7787673473358154, "learning_rate": 4.063725490196078e-05, "loss": 0.2513, "step": 305 }, { "epoch": 1.615841584158416, "grad_norm": 1.3408194780349731, "learning_rate": 4.058823529411765e-05, "loss": 0.1134, "step": 306 }, { "epoch": 1.6211221122112212, "grad_norm": 1.739343523979187, "learning_rate": 4.053921568627451e-05, "loss": 0.194, "step": 307 }, { "epoch": 1.6264026402640264, "grad_norm": 1.7391927242279053, "learning_rate": 4.049019607843138e-05, "loss": 0.2273, "step": 308 }, { "epoch": 1.6316831683168318, "grad_norm": 2.134809732437134, "learning_rate": 4.044117647058824e-05, "loss": 0.349, "step": 309 }, { "epoch": 1.636963696369637, "grad_norm": 1.8628054857254028, "learning_rate": 4.03921568627451e-05, "loss": 0.2687, "step": 310 }, { "epoch": 1.6422442244224422, "grad_norm": 1.8895047903060913, "learning_rate": 4.0343137254901966e-05, "loss": 0.2684, "step": 311 }, { "epoch": 1.6475247524752477, "grad_norm": 1.784513235092163, "learning_rate": 4.029411764705883e-05, "loss": 0.2728, "step": 312 }, { "epoch": 1.6528052805280526, "grad_norm": 1.505031704902649, "learning_rate": 4.024509803921569e-05, "loss": 0.1806, "step": 313 }, { "epoch": 1.658085808580858, "grad_norm": 1.9005438089370728, "learning_rate": 4.0196078431372555e-05, "loss": 0.3178, "step": 314 }, { "epoch": 1.6633663366336635, "grad_norm": 1.769161581993103, "learning_rate": 4.0147058823529415e-05, "loss": 0.2392, "step": 315 }, { "epoch": 1.6686468646864685, "grad_norm": 1.8065416812896729, "learning_rate": 4.0098039215686276e-05, "loss": 0.2972, "step": 316 }, { "epoch": 1.673927392739274, "grad_norm": 1.6391319036483765, "learning_rate": 4.004901960784314e-05, "loss": 0.2132, "step": 317 }, { "epoch": 1.6792079207920794, "grad_norm": 1.3760650157928467, "learning_rate": 4e-05, "loss": 0.1106, "step": 318 }, { "epoch": 1.6844884488448844, "grad_norm": 1.5832698345184326, "learning_rate": 3.9950980392156864e-05, "loss": 0.1965, "step": 319 }, { "epoch": 1.6897689768976898, "grad_norm": 1.830043911933899, "learning_rate": 3.990196078431373e-05, "loss": 0.2658, "step": 320 }, { "epoch": 1.695049504950495, "grad_norm": 1.8273866176605225, "learning_rate": 3.985294117647059e-05, "loss": 0.2284, "step": 321 }, { "epoch": 1.7003300330033002, "grad_norm": 1.6484299898147583, "learning_rate": 3.980392156862745e-05, "loss": 0.1503, "step": 322 }, { "epoch": 1.7056105610561056, "grad_norm": 1.7297075986862183, "learning_rate": 3.975490196078431e-05, "loss": 0.1818, "step": 323 }, { "epoch": 1.7108910891089109, "grad_norm": 1.9557067155838013, "learning_rate": 3.970588235294117e-05, "loss": 0.3179, "step": 324 }, { "epoch": 1.716171617161716, "grad_norm": 1.8135654926300049, "learning_rate": 3.965686274509804e-05, "loss": 0.1871, "step": 325 }, { "epoch": 1.7214521452145215, "grad_norm": 1.9970617294311523, "learning_rate": 3.96078431372549e-05, "loss": 0.2435, "step": 326 }, { "epoch": 1.7267326732673267, "grad_norm": 1.584839105606079, "learning_rate": 3.955882352941177e-05, "loss": 0.1813, "step": 327 }, { "epoch": 1.732013201320132, "grad_norm": 1.5239087343215942, "learning_rate": 3.950980392156863e-05, "loss": 0.1193, "step": 328 }, { "epoch": 1.7372937293729374, "grad_norm": 1.514541506767273, "learning_rate": 3.946078431372549e-05, "loss": 0.1658, "step": 329 }, { "epoch": 1.7425742574257426, "grad_norm": 1.8841454982757568, "learning_rate": 3.9411764705882356e-05, "loss": 0.3376, "step": 330 }, { "epoch": 1.7478547854785478, "grad_norm": 1.5418506860733032, "learning_rate": 3.936274509803922e-05, "loss": 0.2179, "step": 331 }, { "epoch": 1.7531353135313532, "grad_norm": 1.637239933013916, "learning_rate": 3.931372549019608e-05, "loss": 0.1976, "step": 332 }, { "epoch": 1.7584158415841584, "grad_norm": 1.8015220165252686, "learning_rate": 3.9264705882352945e-05, "loss": 0.2629, "step": 333 }, { "epoch": 1.7636963696369636, "grad_norm": 1.7710020542144775, "learning_rate": 3.9215686274509805e-05, "loss": 0.2208, "step": 334 }, { "epoch": 1.768976897689769, "grad_norm": 1.9169963598251343, "learning_rate": 3.9166666666666665e-05, "loss": 0.2811, "step": 335 }, { "epoch": 1.7742574257425743, "grad_norm": 1.738980770111084, "learning_rate": 3.911764705882353e-05, "loss": 0.2009, "step": 336 }, { "epoch": 1.7795379537953795, "grad_norm": 1.888925313949585, "learning_rate": 3.906862745098039e-05, "loss": 0.2615, "step": 337 }, { "epoch": 1.784818481848185, "grad_norm": 1.8248200416564941, "learning_rate": 3.9019607843137254e-05, "loss": 0.2284, "step": 338 }, { "epoch": 1.7900990099009901, "grad_norm": 1.6162784099578857, "learning_rate": 3.897058823529412e-05, "loss": 0.2206, "step": 339 }, { "epoch": 1.7953795379537953, "grad_norm": 1.441273808479309, "learning_rate": 3.892156862745098e-05, "loss": 0.1806, "step": 340 }, { "epoch": 1.8006600660066008, "grad_norm": 1.815674066543579, "learning_rate": 3.887254901960785e-05, "loss": 0.2169, "step": 341 }, { "epoch": 1.805940594059406, "grad_norm": 1.6441361904144287, "learning_rate": 3.882352941176471e-05, "loss": 0.2177, "step": 342 }, { "epoch": 1.8112211221122112, "grad_norm": 1.923663854598999, "learning_rate": 3.877450980392157e-05, "loss": 0.2942, "step": 343 }, { "epoch": 1.8165016501650166, "grad_norm": 1.978797197341919, "learning_rate": 3.872549019607844e-05, "loss": 0.2693, "step": 344 }, { "epoch": 1.8217821782178216, "grad_norm": 1.8134146928787231, "learning_rate": 3.86764705882353e-05, "loss": 0.2713, "step": 345 }, { "epoch": 1.827062706270627, "grad_norm": 1.7703922986984253, "learning_rate": 3.862745098039216e-05, "loss": 0.2413, "step": 346 }, { "epoch": 1.8323432343234325, "grad_norm": 1.7030301094055176, "learning_rate": 3.8578431372549025e-05, "loss": 0.1866, "step": 347 }, { "epoch": 1.8376237623762375, "grad_norm": 1.521941065788269, "learning_rate": 3.8529411764705886e-05, "loss": 0.174, "step": 348 }, { "epoch": 1.842904290429043, "grad_norm": 1.8277267217636108, "learning_rate": 3.8480392156862746e-05, "loss": 0.2077, "step": 349 }, { "epoch": 1.8481848184818483, "grad_norm": 2.028367042541504, "learning_rate": 3.8431372549019614e-05, "loss": 0.2073, "step": 350 }, { "epoch": 1.8534653465346533, "grad_norm": 1.373708724975586, "learning_rate": 3.8382352941176474e-05, "loss": 0.1383, "step": 351 }, { "epoch": 1.8587458745874588, "grad_norm": 1.6797735691070557, "learning_rate": 3.8333333333333334e-05, "loss": 0.1971, "step": 352 }, { "epoch": 1.864026402640264, "grad_norm": 1.4328402280807495, "learning_rate": 3.82843137254902e-05, "loss": 0.1601, "step": 353 }, { "epoch": 1.8693069306930692, "grad_norm": 1.8246557712554932, "learning_rate": 3.8235294117647055e-05, "loss": 0.2408, "step": 354 }, { "epoch": 1.8745874587458746, "grad_norm": 1.6448115110397339, "learning_rate": 3.818627450980392e-05, "loss": 0.2677, "step": 355 }, { "epoch": 1.8798679867986798, "grad_norm": 1.6840052604675293, "learning_rate": 3.813725490196078e-05, "loss": 0.1955, "step": 356 }, { "epoch": 1.885148514851485, "grad_norm": 1.989355206489563, "learning_rate": 3.8088235294117644e-05, "loss": 0.2761, "step": 357 }, { "epoch": 1.8904290429042905, "grad_norm": 1.5983843803405762, "learning_rate": 3.803921568627451e-05, "loss": 0.2057, "step": 358 }, { "epoch": 1.8957095709570957, "grad_norm": 1.5310300588607788, "learning_rate": 3.799019607843137e-05, "loss": 0.1787, "step": 359 }, { "epoch": 1.900990099009901, "grad_norm": 1.835742712020874, "learning_rate": 3.794117647058824e-05, "loss": 0.2972, "step": 360 }, { "epoch": 1.9062706270627063, "grad_norm": 1.890306830406189, "learning_rate": 3.78921568627451e-05, "loss": 0.2972, "step": 361 }, { "epoch": 1.9115511551155115, "grad_norm": 2.0529651641845703, "learning_rate": 3.784313725490196e-05, "loss": 0.3286, "step": 362 }, { "epoch": 1.9168316831683168, "grad_norm": 1.742149829864502, "learning_rate": 3.779411764705883e-05, "loss": 0.2106, "step": 363 }, { "epoch": 1.9221122112211222, "grad_norm": 1.8192955255508423, "learning_rate": 3.774509803921569e-05, "loss": 0.2289, "step": 364 }, { "epoch": 1.9273927392739274, "grad_norm": 1.7159464359283447, "learning_rate": 3.769607843137255e-05, "loss": 0.2324, "step": 365 }, { "epoch": 1.9326732673267326, "grad_norm": 1.7470611333847046, "learning_rate": 3.7647058823529415e-05, "loss": 0.1381, "step": 366 }, { "epoch": 1.937953795379538, "grad_norm": 1.7340490818023682, "learning_rate": 3.7598039215686276e-05, "loss": 0.2136, "step": 367 }, { "epoch": 1.9432343234323433, "grad_norm": 1.6391620635986328, "learning_rate": 3.7549019607843136e-05, "loss": 0.2178, "step": 368 }, { "epoch": 1.9485148514851485, "grad_norm": 1.5655213594436646, "learning_rate": 3.7500000000000003e-05, "loss": 0.2091, "step": 369 }, { "epoch": 1.953795379537954, "grad_norm": 1.7240495681762695, "learning_rate": 3.7450980392156864e-05, "loss": 0.1779, "step": 370 }, { "epoch": 1.9590759075907591, "grad_norm": 1.5939252376556396, "learning_rate": 3.7401960784313724e-05, "loss": 0.1863, "step": 371 }, { "epoch": 1.9643564356435643, "grad_norm": 1.5324146747589111, "learning_rate": 3.735294117647059e-05, "loss": 0.1767, "step": 372 }, { "epoch": 1.9696369636963698, "grad_norm": 1.689475417137146, "learning_rate": 3.730392156862745e-05, "loss": 0.2135, "step": 373 }, { "epoch": 1.974917491749175, "grad_norm": 1.9409862756729126, "learning_rate": 3.725490196078432e-05, "loss": 0.2633, "step": 374 }, { "epoch": 1.9801980198019802, "grad_norm": 1.743085265159607, "learning_rate": 3.720588235294118e-05, "loss": 0.2098, "step": 375 }, { "epoch": 1.9854785478547856, "grad_norm": 1.6699271202087402, "learning_rate": 3.715686274509804e-05, "loss": 0.2288, "step": 376 }, { "epoch": 1.9907590759075906, "grad_norm": 1.662705421447754, "learning_rate": 3.710784313725491e-05, "loss": 0.2385, "step": 377 }, { "epoch": 1.996039603960396, "grad_norm": 1.4365413188934326, "learning_rate": 3.705882352941177e-05, "loss": 0.1793, "step": 378 }, { "epoch": 2.0013201320132015, "grad_norm": 1.5893417596817017, "learning_rate": 3.700980392156863e-05, "loss": 0.2293, "step": 379 }, { "epoch": 2.0066006600660065, "grad_norm": 1.2844809293746948, "learning_rate": 3.6960784313725496e-05, "loss": 0.1413, "step": 380 }, { "epoch": 2.011881188118812, "grad_norm": 1.3603469133377075, "learning_rate": 3.6911764705882356e-05, "loss": 0.1292, "step": 381 }, { "epoch": 2.0171617161716173, "grad_norm": 1.5470837354660034, "learning_rate": 3.686274509803922e-05, "loss": 0.1572, "step": 382 }, { "epoch": 2.0224422442244223, "grad_norm": 1.8591777086257935, "learning_rate": 3.6813725490196084e-05, "loss": 0.1545, "step": 383 }, { "epoch": 2.0277227722772277, "grad_norm": 1.2977832555770874, "learning_rate": 3.6764705882352945e-05, "loss": 0.0742, "step": 384 }, { "epoch": 2.033003300330033, "grad_norm": 1.9429755210876465, "learning_rate": 3.6715686274509805e-05, "loss": 0.1088, "step": 385 }, { "epoch": 2.038283828382838, "grad_norm": 2.0210816860198975, "learning_rate": 3.6666666666666666e-05, "loss": 0.1492, "step": 386 }, { "epoch": 2.0435643564356436, "grad_norm": 1.8192780017852783, "learning_rate": 3.6617647058823526e-05, "loss": 0.1585, "step": 387 }, { "epoch": 2.048844884488449, "grad_norm": 1.5634256601333618, "learning_rate": 3.6568627450980393e-05, "loss": 0.1152, "step": 388 }, { "epoch": 2.054125412541254, "grad_norm": 1.7004332542419434, "learning_rate": 3.6519607843137254e-05, "loss": 0.1629, "step": 389 }, { "epoch": 2.0594059405940595, "grad_norm": 1.7340906858444214, "learning_rate": 3.6470588235294114e-05, "loss": 0.1523, "step": 390 }, { "epoch": 2.064686468646865, "grad_norm": 1.4663294553756714, "learning_rate": 3.642156862745098e-05, "loss": 0.1138, "step": 391 }, { "epoch": 2.06996699669967, "grad_norm": 1.6033658981323242, "learning_rate": 3.637254901960784e-05, "loss": 0.1485, "step": 392 }, { "epoch": 2.0752475247524753, "grad_norm": 1.3706963062286377, "learning_rate": 3.632352941176471e-05, "loss": 0.0883, "step": 393 }, { "epoch": 2.0805280528052803, "grad_norm": 1.0890157222747803, "learning_rate": 3.627450980392157e-05, "loss": 0.0526, "step": 394 }, { "epoch": 2.0858085808580857, "grad_norm": 1.5241326093673706, "learning_rate": 3.622549019607843e-05, "loss": 0.1417, "step": 395 }, { "epoch": 2.091089108910891, "grad_norm": 1.389540433883667, "learning_rate": 3.61764705882353e-05, "loss": 0.0972, "step": 396 }, { "epoch": 2.096369636963696, "grad_norm": 1.8035510778427124, "learning_rate": 3.612745098039216e-05, "loss": 0.1539, "step": 397 }, { "epoch": 2.1016501650165016, "grad_norm": 1.5949468612670898, "learning_rate": 3.607843137254902e-05, "loss": 0.1086, "step": 398 }, { "epoch": 2.106930693069307, "grad_norm": 1.5921865701675415, "learning_rate": 3.6029411764705886e-05, "loss": 0.1227, "step": 399 }, { "epoch": 2.112211221122112, "grad_norm": 1.6403027772903442, "learning_rate": 3.5980392156862746e-05, "loss": 0.1315, "step": 400 }, { "epoch": 2.1174917491749174, "grad_norm": 1.7026506662368774, "learning_rate": 3.593137254901961e-05, "loss": 0.1178, "step": 401 }, { "epoch": 2.122772277227723, "grad_norm": 1.5574462413787842, "learning_rate": 3.5882352941176474e-05, "loss": 0.1316, "step": 402 }, { "epoch": 2.128052805280528, "grad_norm": 1.6316189765930176, "learning_rate": 3.5833333333333335e-05, "loss": 0.1351, "step": 403 }, { "epoch": 2.1333333333333333, "grad_norm": 1.2187312841415405, "learning_rate": 3.5784313725490195e-05, "loss": 0.0708, "step": 404 }, { "epoch": 2.1386138613861387, "grad_norm": 1.619545817375183, "learning_rate": 3.573529411764706e-05, "loss": 0.1369, "step": 405 }, { "epoch": 2.1438943894389437, "grad_norm": 1.4654717445373535, "learning_rate": 3.568627450980392e-05, "loss": 0.1499, "step": 406 }, { "epoch": 2.149174917491749, "grad_norm": 1.393074870109558, "learning_rate": 3.563725490196079e-05, "loss": 0.1277, "step": 407 }, { "epoch": 2.1544554455445546, "grad_norm": 1.7896983623504639, "learning_rate": 3.558823529411765e-05, "loss": 0.166, "step": 408 }, { "epoch": 2.1597359735973596, "grad_norm": 1.5691279172897339, "learning_rate": 3.553921568627451e-05, "loss": 0.0936, "step": 409 }, { "epoch": 2.165016501650165, "grad_norm": 1.6163969039916992, "learning_rate": 3.549019607843138e-05, "loss": 0.1234, "step": 410 }, { "epoch": 2.1702970297029704, "grad_norm": 1.7184455394744873, "learning_rate": 3.544117647058824e-05, "loss": 0.1287, "step": 411 }, { "epoch": 2.1755775577557754, "grad_norm": 1.8023498058319092, "learning_rate": 3.53921568627451e-05, "loss": 0.1433, "step": 412 }, { "epoch": 2.180858085808581, "grad_norm": 1.6063473224639893, "learning_rate": 3.534313725490197e-05, "loss": 0.1158, "step": 413 }, { "epoch": 2.1861386138613863, "grad_norm": 1.6521129608154297, "learning_rate": 3.529411764705883e-05, "loss": 0.115, "step": 414 }, { "epoch": 2.1914191419141913, "grad_norm": 1.5446815490722656, "learning_rate": 3.524509803921569e-05, "loss": 0.115, "step": 415 }, { "epoch": 2.1966996699669967, "grad_norm": 1.4110487699508667, "learning_rate": 3.5196078431372555e-05, "loss": 0.132, "step": 416 }, { "epoch": 2.201980198019802, "grad_norm": 1.732271671295166, "learning_rate": 3.514705882352941e-05, "loss": 0.116, "step": 417 }, { "epoch": 2.207260726072607, "grad_norm": 1.8033490180969238, "learning_rate": 3.5098039215686276e-05, "loss": 0.1675, "step": 418 }, { "epoch": 2.2125412541254126, "grad_norm": 1.6612602472305298, "learning_rate": 3.5049019607843136e-05, "loss": 0.1549, "step": 419 }, { "epoch": 2.217821782178218, "grad_norm": 1.4840703010559082, "learning_rate": 3.5e-05, "loss": 0.116, "step": 420 }, { "epoch": 2.223102310231023, "grad_norm": 1.5208748579025269, "learning_rate": 3.4950980392156864e-05, "loss": 0.1286, "step": 421 }, { "epoch": 2.2283828382838284, "grad_norm": 1.6537951231002808, "learning_rate": 3.4901960784313725e-05, "loss": 0.0995, "step": 422 }, { "epoch": 2.233663366336634, "grad_norm": 1.8060580492019653, "learning_rate": 3.4852941176470585e-05, "loss": 0.1264, "step": 423 }, { "epoch": 2.238943894389439, "grad_norm": 1.8712666034698486, "learning_rate": 3.480392156862745e-05, "loss": 0.1722, "step": 424 }, { "epoch": 2.2442244224422443, "grad_norm": 1.5156561136245728, "learning_rate": 3.475490196078431e-05, "loss": 0.117, "step": 425 }, { "epoch": 2.2495049504950497, "grad_norm": 1.6250231266021729, "learning_rate": 3.470588235294118e-05, "loss": 0.1299, "step": 426 }, { "epoch": 2.2547854785478547, "grad_norm": 2.060224771499634, "learning_rate": 3.465686274509804e-05, "loss": 0.1552, "step": 427 }, { "epoch": 2.26006600660066, "grad_norm": 1.6682344675064087, "learning_rate": 3.46078431372549e-05, "loss": 0.137, "step": 428 }, { "epoch": 2.2653465346534656, "grad_norm": 1.5238713026046753, "learning_rate": 3.455882352941177e-05, "loss": 0.0956, "step": 429 }, { "epoch": 2.2706270627062706, "grad_norm": 1.7260777950286865, "learning_rate": 3.450980392156863e-05, "loss": 0.127, "step": 430 }, { "epoch": 2.275907590759076, "grad_norm": 1.5617260932922363, "learning_rate": 3.446078431372549e-05, "loss": 0.1495, "step": 431 }, { "epoch": 2.2811881188118814, "grad_norm": 1.6972553730010986, "learning_rate": 3.441176470588236e-05, "loss": 0.1521, "step": 432 }, { "epoch": 2.2864686468646864, "grad_norm": 1.4397108554840088, "learning_rate": 3.436274509803922e-05, "loss": 0.1218, "step": 433 }, { "epoch": 2.291749174917492, "grad_norm": 1.099647045135498, "learning_rate": 3.431372549019608e-05, "loss": 0.0629, "step": 434 }, { "epoch": 2.297029702970297, "grad_norm": 1.1678043603897095, "learning_rate": 3.4264705882352945e-05, "loss": 0.0905, "step": 435 }, { "epoch": 2.3023102310231023, "grad_norm": 1.3565727472305298, "learning_rate": 3.4215686274509805e-05, "loss": 0.1028, "step": 436 }, { "epoch": 2.3075907590759077, "grad_norm": 1.846362590789795, "learning_rate": 3.4166666666666666e-05, "loss": 0.1715, "step": 437 }, { "epoch": 2.3128712871287127, "grad_norm": 1.637338399887085, "learning_rate": 3.411764705882353e-05, "loss": 0.131, "step": 438 }, { "epoch": 2.318151815181518, "grad_norm": 1.5920330286026, "learning_rate": 3.4068627450980394e-05, "loss": 0.1279, "step": 439 }, { "epoch": 2.3234323432343236, "grad_norm": 1.6633886098861694, "learning_rate": 3.401960784313726e-05, "loss": 0.1284, "step": 440 }, { "epoch": 2.3287128712871286, "grad_norm": 1.677240252494812, "learning_rate": 3.397058823529412e-05, "loss": 0.132, "step": 441 }, { "epoch": 2.333993399339934, "grad_norm": 1.6058099269866943, "learning_rate": 3.392156862745098e-05, "loss": 0.116, "step": 442 }, { "epoch": 2.3392739273927394, "grad_norm": 1.586955189704895, "learning_rate": 3.387254901960785e-05, "loss": 0.1201, "step": 443 }, { "epoch": 2.3445544554455444, "grad_norm": 2.02844500541687, "learning_rate": 3.382352941176471e-05, "loss": 0.1732, "step": 444 }, { "epoch": 2.34983498349835, "grad_norm": 1.858375072479248, "learning_rate": 3.377450980392157e-05, "loss": 0.1591, "step": 445 }, { "epoch": 2.3551155115511553, "grad_norm": 1.9212247133255005, "learning_rate": 3.372549019607844e-05, "loss": 0.2203, "step": 446 }, { "epoch": 2.3603960396039603, "grad_norm": 1.327735424041748, "learning_rate": 3.36764705882353e-05, "loss": 0.0983, "step": 447 }, { "epoch": 2.3656765676567657, "grad_norm": 1.6712234020233154, "learning_rate": 3.362745098039216e-05, "loss": 0.1282, "step": 448 }, { "epoch": 2.370957095709571, "grad_norm": 1.812563180923462, "learning_rate": 3.357843137254902e-05, "loss": 0.1417, "step": 449 }, { "epoch": 2.376237623762376, "grad_norm": 1.4844976663589478, "learning_rate": 3.352941176470588e-05, "loss": 0.0892, "step": 450 }, { "epoch": 2.3815181518151816, "grad_norm": 1.7959266901016235, "learning_rate": 3.348039215686275e-05, "loss": 0.1753, "step": 451 }, { "epoch": 2.3867986798679866, "grad_norm": 1.4532088041305542, "learning_rate": 3.343137254901961e-05, "loss": 0.1154, "step": 452 }, { "epoch": 2.392079207920792, "grad_norm": 1.618485927581787, "learning_rate": 3.338235294117647e-05, "loss": 0.1268, "step": 453 }, { "epoch": 2.3973597359735974, "grad_norm": 1.537070870399475, "learning_rate": 3.3333333333333335e-05, "loss": 0.1304, "step": 454 }, { "epoch": 2.4026402640264024, "grad_norm": 1.4429882764816284, "learning_rate": 3.3284313725490195e-05, "loss": 0.1521, "step": 455 }, { "epoch": 2.407920792079208, "grad_norm": 1.5398533344268799, "learning_rate": 3.3235294117647056e-05, "loss": 0.1237, "step": 456 }, { "epoch": 2.4132013201320133, "grad_norm": 1.7345346212387085, "learning_rate": 3.318627450980392e-05, "loss": 0.1765, "step": 457 }, { "epoch": 2.4184818481848183, "grad_norm": 1.6664462089538574, "learning_rate": 3.3137254901960784e-05, "loss": 0.1363, "step": 458 }, { "epoch": 2.4237623762376237, "grad_norm": 1.2778750658035278, "learning_rate": 3.308823529411765e-05, "loss": 0.0798, "step": 459 }, { "epoch": 2.429042904290429, "grad_norm": 1.5715030431747437, "learning_rate": 3.303921568627451e-05, "loss": 0.1376, "step": 460 }, { "epoch": 2.434323432343234, "grad_norm": 1.3661701679229736, "learning_rate": 3.299019607843137e-05, "loss": 0.1179, "step": 461 }, { "epoch": 2.4396039603960396, "grad_norm": 1.6188709735870361, "learning_rate": 3.294117647058824e-05, "loss": 0.102, "step": 462 }, { "epoch": 2.444884488448845, "grad_norm": 1.5044273138046265, "learning_rate": 3.28921568627451e-05, "loss": 0.1048, "step": 463 }, { "epoch": 2.45016501650165, "grad_norm": 1.8329862356185913, "learning_rate": 3.284313725490196e-05, "loss": 0.1501, "step": 464 }, { "epoch": 2.4554455445544554, "grad_norm": 1.4117523431777954, "learning_rate": 3.279411764705883e-05, "loss": 0.0899, "step": 465 }, { "epoch": 2.460726072607261, "grad_norm": 1.4140467643737793, "learning_rate": 3.274509803921569e-05, "loss": 0.1143, "step": 466 }, { "epoch": 2.466006600660066, "grad_norm": 1.5512601137161255, "learning_rate": 3.269607843137255e-05, "loss": 0.1294, "step": 467 }, { "epoch": 2.4712871287128713, "grad_norm": 1.366589069366455, "learning_rate": 3.2647058823529416e-05, "loss": 0.1044, "step": 468 }, { "epoch": 2.4765676567656767, "grad_norm": 1.4923369884490967, "learning_rate": 3.2598039215686276e-05, "loss": 0.1209, "step": 469 }, { "epoch": 2.4818481848184817, "grad_norm": 1.4540033340454102, "learning_rate": 3.254901960784314e-05, "loss": 0.1094, "step": 470 }, { "epoch": 2.487128712871287, "grad_norm": 1.7414342164993286, "learning_rate": 3.2500000000000004e-05, "loss": 0.2082, "step": 471 }, { "epoch": 2.4924092409240926, "grad_norm": 1.3162891864776611, "learning_rate": 3.2450980392156864e-05, "loss": 0.0972, "step": 472 }, { "epoch": 2.4976897689768975, "grad_norm": 1.5640320777893066, "learning_rate": 3.240196078431373e-05, "loss": 0.1221, "step": 473 }, { "epoch": 2.502970297029703, "grad_norm": 1.3759536743164062, "learning_rate": 3.235294117647059e-05, "loss": 0.1026, "step": 474 }, { "epoch": 2.5082508250825084, "grad_norm": 1.3398675918579102, "learning_rate": 3.230392156862745e-05, "loss": 0.0904, "step": 475 }, { "epoch": 2.5135313531353134, "grad_norm": 1.1896995306015015, "learning_rate": 3.225490196078432e-05, "loss": 0.0847, "step": 476 }, { "epoch": 2.518811881188119, "grad_norm": 1.3623279333114624, "learning_rate": 3.220588235294118e-05, "loss": 0.1159, "step": 477 }, { "epoch": 2.5240924092409243, "grad_norm": 1.3050968647003174, "learning_rate": 3.215686274509804e-05, "loss": 0.0967, "step": 478 }, { "epoch": 2.5293729372937293, "grad_norm": 1.8732277154922485, "learning_rate": 3.210784313725491e-05, "loss": 0.1273, "step": 479 }, { "epoch": 2.5346534653465347, "grad_norm": 1.5116699934005737, "learning_rate": 3.205882352941177e-05, "loss": 0.1245, "step": 480 }, { "epoch": 2.53993399339934, "grad_norm": 1.6075270175933838, "learning_rate": 3.200980392156863e-05, "loss": 0.1215, "step": 481 }, { "epoch": 2.545214521452145, "grad_norm": 1.6938683986663818, "learning_rate": 3.196078431372549e-05, "loss": 0.1226, "step": 482 }, { "epoch": 2.5504950495049505, "grad_norm": 1.361075520515442, "learning_rate": 3.191176470588235e-05, "loss": 0.108, "step": 483 }, { "epoch": 2.555775577557756, "grad_norm": 1.388642430305481, "learning_rate": 3.186274509803922e-05, "loss": 0.1223, "step": 484 }, { "epoch": 2.561056105610561, "grad_norm": 1.637742519378662, "learning_rate": 3.181372549019608e-05, "loss": 0.1395, "step": 485 }, { "epoch": 2.5663366336633664, "grad_norm": 1.377528429031372, "learning_rate": 3.176470588235294e-05, "loss": 0.1064, "step": 486 }, { "epoch": 2.571617161716172, "grad_norm": 1.3829468488693237, "learning_rate": 3.1715686274509806e-05, "loss": 0.1101, "step": 487 }, { "epoch": 2.576897689768977, "grad_norm": 1.142531394958496, "learning_rate": 3.1666666666666666e-05, "loss": 0.0885, "step": 488 }, { "epoch": 2.5821782178217823, "grad_norm": 1.226916790008545, "learning_rate": 3.161764705882353e-05, "loss": 0.0713, "step": 489 }, { "epoch": 2.5874587458745877, "grad_norm": 1.3948824405670166, "learning_rate": 3.1568627450980394e-05, "loss": 0.0945, "step": 490 }, { "epoch": 2.5927392739273927, "grad_norm": 1.679543375968933, "learning_rate": 3.1519607843137254e-05, "loss": 0.1122, "step": 491 }, { "epoch": 2.598019801980198, "grad_norm": 1.4996442794799805, "learning_rate": 3.147058823529412e-05, "loss": 0.1502, "step": 492 }, { "epoch": 2.6033003300330035, "grad_norm": 1.6479462385177612, "learning_rate": 3.142156862745098e-05, "loss": 0.1335, "step": 493 }, { "epoch": 2.6085808580858085, "grad_norm": 1.9503371715545654, "learning_rate": 3.137254901960784e-05, "loss": 0.1378, "step": 494 }, { "epoch": 2.613861386138614, "grad_norm": 1.6987338066101074, "learning_rate": 3.132352941176471e-05, "loss": 0.174, "step": 495 }, { "epoch": 2.6191419141914194, "grad_norm": 1.4869428873062134, "learning_rate": 3.127450980392157e-05, "loss": 0.1444, "step": 496 }, { "epoch": 2.6244224422442244, "grad_norm": 1.4220130443572998, "learning_rate": 3.122549019607843e-05, "loss": 0.1146, "step": 497 }, { "epoch": 2.62970297029703, "grad_norm": 1.8738462924957275, "learning_rate": 3.11764705882353e-05, "loss": 0.1999, "step": 498 }, { "epoch": 2.6349834983498353, "grad_norm": 1.6982347965240479, "learning_rate": 3.112745098039216e-05, "loss": 0.1241, "step": 499 }, { "epoch": 2.6402640264026402, "grad_norm": 1.5183193683624268, "learning_rate": 3.107843137254902e-05, "loss": 0.1198, "step": 500 }, { "epoch": 2.6455445544554457, "grad_norm": 1.4872636795043945, "learning_rate": 3.1029411764705886e-05, "loss": 0.0927, "step": 501 }, { "epoch": 2.6508250825082507, "grad_norm": 1.6051157712936401, "learning_rate": 3.098039215686275e-05, "loss": 0.1524, "step": 502 }, { "epoch": 2.656105610561056, "grad_norm": 1.5139557123184204, "learning_rate": 3.093137254901961e-05, "loss": 0.1237, "step": 503 }, { "epoch": 2.6613861386138615, "grad_norm": 1.5860090255737305, "learning_rate": 3.0882352941176475e-05, "loss": 0.1409, "step": 504 }, { "epoch": 2.6666666666666665, "grad_norm": 1.7378284931182861, "learning_rate": 3.0833333333333335e-05, "loss": 0.1551, "step": 505 }, { "epoch": 2.671947194719472, "grad_norm": 1.5908591747283936, "learning_rate": 3.07843137254902e-05, "loss": 0.1169, "step": 506 }, { "epoch": 2.6772277227722774, "grad_norm": 1.4210619926452637, "learning_rate": 3.073529411764706e-05, "loss": 0.1217, "step": 507 }, { "epoch": 2.6825082508250824, "grad_norm": 1.4068129062652588, "learning_rate": 3.0686274509803923e-05, "loss": 0.1484, "step": 508 }, { "epoch": 2.687788778877888, "grad_norm": 1.4808945655822754, "learning_rate": 3.063725490196079e-05, "loss": 0.1402, "step": 509 }, { "epoch": 2.693069306930693, "grad_norm": 1.342729926109314, "learning_rate": 3.058823529411765e-05, "loss": 0.0826, "step": 510 }, { "epoch": 2.6983498349834982, "grad_norm": 1.457270860671997, "learning_rate": 3.053921568627451e-05, "loss": 0.0809, "step": 511 }, { "epoch": 2.7036303630363037, "grad_norm": 1.3564640283584595, "learning_rate": 3.0490196078431376e-05, "loss": 0.1333, "step": 512 }, { "epoch": 2.7089108910891087, "grad_norm": 1.5419988632202148, "learning_rate": 3.0441176470588233e-05, "loss": 0.1145, "step": 513 }, { "epoch": 2.714191419141914, "grad_norm": 1.4028230905532837, "learning_rate": 3.0392156862745097e-05, "loss": 0.1156, "step": 514 }, { "epoch": 2.7194719471947195, "grad_norm": 1.5217773914337158, "learning_rate": 3.034313725490196e-05, "loss": 0.0962, "step": 515 }, { "epoch": 2.7247524752475245, "grad_norm": 1.2986435890197754, "learning_rate": 3.0294117647058824e-05, "loss": 0.1039, "step": 516 }, { "epoch": 2.73003300330033, "grad_norm": 1.3331955671310425, "learning_rate": 3.0245098039215685e-05, "loss": 0.1077, "step": 517 }, { "epoch": 2.7353135313531354, "grad_norm": 1.7780873775482178, "learning_rate": 3.019607843137255e-05, "loss": 0.2033, "step": 518 }, { "epoch": 2.7405940594059404, "grad_norm": 1.1954641342163086, "learning_rate": 3.0147058823529413e-05, "loss": 0.1134, "step": 519 }, { "epoch": 2.745874587458746, "grad_norm": 1.465927004814148, "learning_rate": 3.0098039215686273e-05, "loss": 0.1329, "step": 520 }, { "epoch": 2.7511551155115512, "grad_norm": 1.359671711921692, "learning_rate": 3.0049019607843137e-05, "loss": 0.0876, "step": 521 }, { "epoch": 2.756435643564356, "grad_norm": 1.570184350013733, "learning_rate": 3e-05, "loss": 0.1125, "step": 522 }, { "epoch": 2.7617161716171617, "grad_norm": 1.286363959312439, "learning_rate": 2.9950980392156865e-05, "loss": 0.0978, "step": 523 }, { "epoch": 2.766996699669967, "grad_norm": 1.6705877780914307, "learning_rate": 2.9901960784313725e-05, "loss": 0.1546, "step": 524 }, { "epoch": 2.772277227722772, "grad_norm": 1.475122332572937, "learning_rate": 2.985294117647059e-05, "loss": 0.1083, "step": 525 }, { "epoch": 2.7775577557755775, "grad_norm": 1.4407463073730469, "learning_rate": 2.9803921568627453e-05, "loss": 0.1123, "step": 526 }, { "epoch": 2.782838283828383, "grad_norm": 1.4154635667800903, "learning_rate": 2.9754901960784313e-05, "loss": 0.1232, "step": 527 }, { "epoch": 2.788118811881188, "grad_norm": 1.6671222448349, "learning_rate": 2.9705882352941177e-05, "loss": 0.1767, "step": 528 }, { "epoch": 2.7933993399339934, "grad_norm": 1.4719637632369995, "learning_rate": 2.965686274509804e-05, "loss": 0.1325, "step": 529 }, { "epoch": 2.798679867986799, "grad_norm": 0.979076087474823, "learning_rate": 2.9607843137254905e-05, "loss": 0.0692, "step": 530 }, { "epoch": 2.803960396039604, "grad_norm": 1.507076621055603, "learning_rate": 2.9558823529411766e-05, "loss": 0.1275, "step": 531 }, { "epoch": 2.809240924092409, "grad_norm": 1.6182982921600342, "learning_rate": 2.950980392156863e-05, "loss": 0.1528, "step": 532 }, { "epoch": 2.8145214521452147, "grad_norm": 1.5272071361541748, "learning_rate": 2.9460784313725493e-05, "loss": 0.1344, "step": 533 }, { "epoch": 2.8198019801980196, "grad_norm": 1.3018370866775513, "learning_rate": 2.9411764705882354e-05, "loss": 0.101, "step": 534 }, { "epoch": 2.825082508250825, "grad_norm": 1.2394366264343262, "learning_rate": 2.9362745098039218e-05, "loss": 0.0893, "step": 535 }, { "epoch": 2.8303630363036305, "grad_norm": 1.5351279973983765, "learning_rate": 2.931372549019608e-05, "loss": 0.1226, "step": 536 }, { "epoch": 2.8356435643564355, "grad_norm": 1.7062324285507202, "learning_rate": 2.9264705882352945e-05, "loss": 0.1363, "step": 537 }, { "epoch": 2.840924092409241, "grad_norm": 1.666043996810913, "learning_rate": 2.9215686274509806e-05, "loss": 0.137, "step": 538 }, { "epoch": 2.8462046204620464, "grad_norm": 1.3965767621994019, "learning_rate": 2.916666666666667e-05, "loss": 0.1018, "step": 539 }, { "epoch": 2.8514851485148514, "grad_norm": 1.4355634450912476, "learning_rate": 2.9117647058823534e-05, "loss": 0.1009, "step": 540 }, { "epoch": 2.856765676567657, "grad_norm": 1.369983434677124, "learning_rate": 2.9068627450980394e-05, "loss": 0.0964, "step": 541 }, { "epoch": 2.862046204620462, "grad_norm": 1.5841052532196045, "learning_rate": 2.9019607843137258e-05, "loss": 0.148, "step": 542 }, { "epoch": 2.867326732673267, "grad_norm": 1.360392451286316, "learning_rate": 2.8970588235294122e-05, "loss": 0.0744, "step": 543 }, { "epoch": 2.8726072607260726, "grad_norm": 1.379198431968689, "learning_rate": 2.8921568627450986e-05, "loss": 0.1047, "step": 544 }, { "epoch": 2.877887788778878, "grad_norm": 1.3699851036071777, "learning_rate": 2.8872549019607843e-05, "loss": 0.0813, "step": 545 }, { "epoch": 2.883168316831683, "grad_norm": 1.162501335144043, "learning_rate": 2.8823529411764703e-05, "loss": 0.0848, "step": 546 }, { "epoch": 2.8884488448844885, "grad_norm": 1.6580568552017212, "learning_rate": 2.8774509803921567e-05, "loss": 0.1252, "step": 547 }, { "epoch": 2.893729372937294, "grad_norm": 1.8794211149215698, "learning_rate": 2.872549019607843e-05, "loss": 0.1581, "step": 548 }, { "epoch": 2.899009900990099, "grad_norm": 1.7686625719070435, "learning_rate": 2.8676470588235295e-05, "loss": 0.172, "step": 549 }, { "epoch": 2.9042904290429044, "grad_norm": 1.371221899986267, "learning_rate": 2.8627450980392155e-05, "loss": 0.0977, "step": 550 }, { "epoch": 2.90957095709571, "grad_norm": 1.4571088552474976, "learning_rate": 2.857843137254902e-05, "loss": 0.1251, "step": 551 }, { "epoch": 2.9148514851485148, "grad_norm": 1.1055774688720703, "learning_rate": 2.8529411764705883e-05, "loss": 0.0761, "step": 552 }, { "epoch": 2.92013201320132, "grad_norm": 1.4305754899978638, "learning_rate": 2.8480392156862744e-05, "loss": 0.143, "step": 553 }, { "epoch": 2.9254125412541256, "grad_norm": 1.3293156623840332, "learning_rate": 2.8431372549019608e-05, "loss": 0.1218, "step": 554 }, { "epoch": 2.9306930693069306, "grad_norm": 1.7073545455932617, "learning_rate": 2.838235294117647e-05, "loss": 0.1495, "step": 555 }, { "epoch": 2.935973597359736, "grad_norm": 1.0066826343536377, "learning_rate": 2.8333333333333335e-05, "loss": 0.0698, "step": 556 }, { "epoch": 2.9412541254125415, "grad_norm": 1.5967711210250854, "learning_rate": 2.8284313725490196e-05, "loss": 0.1436, "step": 557 }, { "epoch": 2.9465346534653465, "grad_norm": 1.7498648166656494, "learning_rate": 2.823529411764706e-05, "loss": 0.1793, "step": 558 }, { "epoch": 2.951815181518152, "grad_norm": 1.4582027196884155, "learning_rate": 2.8186274509803924e-05, "loss": 0.1152, "step": 559 }, { "epoch": 2.9570957095709574, "grad_norm": 1.1277716159820557, "learning_rate": 2.8137254901960784e-05, "loss": 0.0807, "step": 560 }, { "epoch": 2.9623762376237623, "grad_norm": 1.4396110773086548, "learning_rate": 2.8088235294117648e-05, "loss": 0.1141, "step": 561 }, { "epoch": 2.9676567656765678, "grad_norm": 1.3722310066223145, "learning_rate": 2.8039215686274512e-05, "loss": 0.0952, "step": 562 }, { "epoch": 2.972937293729373, "grad_norm": 1.6217355728149414, "learning_rate": 2.7990196078431376e-05, "loss": 0.1329, "step": 563 }, { "epoch": 2.978217821782178, "grad_norm": 1.3327069282531738, "learning_rate": 2.7941176470588236e-05, "loss": 0.109, "step": 564 }, { "epoch": 2.9834983498349836, "grad_norm": 1.6068836450576782, "learning_rate": 2.78921568627451e-05, "loss": 0.1136, "step": 565 }, { "epoch": 2.9887788778877886, "grad_norm": 1.6134989261627197, "learning_rate": 2.7843137254901964e-05, "loss": 0.1452, "step": 566 }, { "epoch": 2.994059405940594, "grad_norm": 1.301640272140503, "learning_rate": 2.7794117647058824e-05, "loss": 0.1692, "step": 567 }, { "epoch": 2.9993399339933995, "grad_norm": 1.5887155532836914, "learning_rate": 2.774509803921569e-05, "loss": 0.1529, "step": 568 }, { "epoch": 3.0046204620462045, "grad_norm": 1.1964720487594604, "learning_rate": 2.7696078431372552e-05, "loss": 0.05, "step": 569 }, { "epoch": 3.00990099009901, "grad_norm": 1.26799476146698, "learning_rate": 2.7647058823529416e-05, "loss": 0.0582, "step": 570 }, { "epoch": 3.0151815181518153, "grad_norm": 1.0527023077011108, "learning_rate": 2.7598039215686277e-05, "loss": 0.0661, "step": 571 }, { "epoch": 3.0204620462046203, "grad_norm": 1.0215531587600708, "learning_rate": 2.754901960784314e-05, "loss": 0.0487, "step": 572 }, { "epoch": 3.0257425742574258, "grad_norm": 0.9712955951690674, "learning_rate": 2.7500000000000004e-05, "loss": 0.0533, "step": 573 }, { "epoch": 3.031023102310231, "grad_norm": 0.7851030826568604, "learning_rate": 2.7450980392156865e-05, "loss": 0.0253, "step": 574 }, { "epoch": 3.036303630363036, "grad_norm": 1.6078637838363647, "learning_rate": 2.740196078431373e-05, "loss": 0.0841, "step": 575 }, { "epoch": 3.0415841584158416, "grad_norm": 1.459649920463562, "learning_rate": 2.7352941176470593e-05, "loss": 0.0654, "step": 576 }, { "epoch": 3.046864686468647, "grad_norm": 1.6587159633636475, "learning_rate": 2.730392156862745e-05, "loss": 0.0524, "step": 577 }, { "epoch": 3.052145214521452, "grad_norm": 1.8531984090805054, "learning_rate": 2.7254901960784314e-05, "loss": 0.0682, "step": 578 }, { "epoch": 3.0574257425742575, "grad_norm": 2.0687427520751953, "learning_rate": 2.7205882352941174e-05, "loss": 0.0809, "step": 579 }, { "epoch": 3.062706270627063, "grad_norm": 1.7597553730010986, "learning_rate": 2.7156862745098038e-05, "loss": 0.0906, "step": 580 }, { "epoch": 3.067986798679868, "grad_norm": 1.593573808670044, "learning_rate": 2.7107843137254902e-05, "loss": 0.0798, "step": 581 }, { "epoch": 3.0732673267326733, "grad_norm": 1.6800479888916016, "learning_rate": 2.7058823529411766e-05, "loss": 0.0769, "step": 582 }, { "epoch": 3.0785478547854783, "grad_norm": 1.2938110828399658, "learning_rate": 2.7009803921568626e-05, "loss": 0.0588, "step": 583 }, { "epoch": 3.0838283828382838, "grad_norm": 0.9052571654319763, "learning_rate": 2.696078431372549e-05, "loss": 0.0343, "step": 584 }, { "epoch": 3.089108910891089, "grad_norm": 1.4364440441131592, "learning_rate": 2.6911764705882354e-05, "loss": 0.0659, "step": 585 }, { "epoch": 3.094389438943894, "grad_norm": 1.2938659191131592, "learning_rate": 2.6862745098039214e-05, "loss": 0.0653, "step": 586 }, { "epoch": 3.0996699669966996, "grad_norm": 1.2038064002990723, "learning_rate": 2.681372549019608e-05, "loss": 0.0731, "step": 587 }, { "epoch": 3.104950495049505, "grad_norm": 1.1313140392303467, "learning_rate": 2.6764705882352942e-05, "loss": 0.0488, "step": 588 }, { "epoch": 3.11023102310231, "grad_norm": 1.426941990852356, "learning_rate": 2.6715686274509806e-05, "loss": 0.0724, "step": 589 }, { "epoch": 3.1155115511551155, "grad_norm": 1.415529489517212, "learning_rate": 2.6666666666666667e-05, "loss": 0.0902, "step": 590 }, { "epoch": 3.120792079207921, "grad_norm": 1.4976341724395752, "learning_rate": 2.661764705882353e-05, "loss": 0.0837, "step": 591 }, { "epoch": 3.126072607260726, "grad_norm": 1.3381704092025757, "learning_rate": 2.6568627450980394e-05, "loss": 0.0735, "step": 592 }, { "epoch": 3.1313531353135313, "grad_norm": 1.0349838733673096, "learning_rate": 2.6519607843137255e-05, "loss": 0.0482, "step": 593 }, { "epoch": 3.1366336633663368, "grad_norm": 1.4465690851211548, "learning_rate": 2.647058823529412e-05, "loss": 0.0823, "step": 594 }, { "epoch": 3.1419141914191417, "grad_norm": 1.113172173500061, "learning_rate": 2.6421568627450983e-05, "loss": 0.0608, "step": 595 }, { "epoch": 3.147194719471947, "grad_norm": 1.3824833631515503, "learning_rate": 2.6372549019607846e-05, "loss": 0.0566, "step": 596 }, { "epoch": 3.1524752475247526, "grad_norm": 1.2780243158340454, "learning_rate": 2.6323529411764707e-05, "loss": 0.0566, "step": 597 }, { "epoch": 3.1577557755775576, "grad_norm": 1.7640819549560547, "learning_rate": 2.627450980392157e-05, "loss": 0.0627, "step": 598 }, { "epoch": 3.163036303630363, "grad_norm": 1.8267886638641357, "learning_rate": 2.6225490196078435e-05, "loss": 0.1135, "step": 599 }, { "epoch": 3.1683168316831685, "grad_norm": 1.4951374530792236, "learning_rate": 2.6176470588235295e-05, "loss": 0.0824, "step": 600 }, { "epoch": 3.1735973597359735, "grad_norm": 1.1276224851608276, "learning_rate": 2.612745098039216e-05, "loss": 0.0481, "step": 601 }, { "epoch": 3.178877887788779, "grad_norm": 1.3539289236068726, "learning_rate": 2.6078431372549023e-05, "loss": 0.0774, "step": 602 }, { "epoch": 3.1841584158415843, "grad_norm": 1.1666077375411987, "learning_rate": 2.6029411764705887e-05, "loss": 0.0484, "step": 603 }, { "epoch": 3.1894389438943893, "grad_norm": 1.4730373620986938, "learning_rate": 2.5980392156862747e-05, "loss": 0.0576, "step": 604 }, { "epoch": 3.1947194719471947, "grad_norm": 1.3001021146774292, "learning_rate": 2.593137254901961e-05, "loss": 0.0657, "step": 605 }, { "epoch": 3.2, "grad_norm": 1.2780375480651855, "learning_rate": 2.5882352941176475e-05, "loss": 0.0538, "step": 606 }, { "epoch": 3.205280528052805, "grad_norm": 1.1565543413162231, "learning_rate": 2.5833333333333336e-05, "loss": 0.0505, "step": 607 }, { "epoch": 3.2105610561056106, "grad_norm": 1.4991666078567505, "learning_rate": 2.57843137254902e-05, "loss": 0.058, "step": 608 }, { "epoch": 3.215841584158416, "grad_norm": 1.1679130792617798, "learning_rate": 2.5735294117647057e-05, "loss": 0.0525, "step": 609 }, { "epoch": 3.221122112211221, "grad_norm": 1.251451849937439, "learning_rate": 2.568627450980392e-05, "loss": 0.0525, "step": 610 }, { "epoch": 3.2264026402640265, "grad_norm": 1.4389278888702393, "learning_rate": 2.5637254901960784e-05, "loss": 0.0776, "step": 611 }, { "epoch": 3.231683168316832, "grad_norm": 1.2829740047454834, "learning_rate": 2.5588235294117645e-05, "loss": 0.0529, "step": 612 }, { "epoch": 3.236963696369637, "grad_norm": 1.0701279640197754, "learning_rate": 2.553921568627451e-05, "loss": 0.0409, "step": 613 }, { "epoch": 3.2422442244224423, "grad_norm": 1.363618016242981, "learning_rate": 2.5490196078431373e-05, "loss": 0.0797, "step": 614 }, { "epoch": 3.2475247524752477, "grad_norm": 1.2213047742843628, "learning_rate": 2.5441176470588236e-05, "loss": 0.0621, "step": 615 }, { "epoch": 3.2528052805280527, "grad_norm": 1.4697519540786743, "learning_rate": 2.5392156862745097e-05, "loss": 0.0779, "step": 616 }, { "epoch": 3.258085808580858, "grad_norm": 1.4408804178237915, "learning_rate": 2.534313725490196e-05, "loss": 0.0642, "step": 617 }, { "epoch": 3.2633663366336636, "grad_norm": 1.207322120666504, "learning_rate": 2.5294117647058825e-05, "loss": 0.0588, "step": 618 }, { "epoch": 3.2686468646864686, "grad_norm": 1.399373173713684, "learning_rate": 2.5245098039215685e-05, "loss": 0.0789, "step": 619 }, { "epoch": 3.273927392739274, "grad_norm": 1.1810945272445679, "learning_rate": 2.519607843137255e-05, "loss": 0.0574, "step": 620 }, { "epoch": 3.2792079207920795, "grad_norm": 1.0468858480453491, "learning_rate": 2.5147058823529413e-05, "loss": 0.0466, "step": 621 }, { "epoch": 3.2844884488448844, "grad_norm": 1.3326268196105957, "learning_rate": 2.5098039215686277e-05, "loss": 0.0674, "step": 622 }, { "epoch": 3.28976897689769, "grad_norm": 1.3134809732437134, "learning_rate": 2.5049019607843137e-05, "loss": 0.0588, "step": 623 }, { "epoch": 3.295049504950495, "grad_norm": 1.1663881540298462, "learning_rate": 2.5e-05, "loss": 0.0465, "step": 624 }, { "epoch": 3.3003300330033003, "grad_norm": 0.9541448950767517, "learning_rate": 2.4950980392156865e-05, "loss": 0.0434, "step": 625 }, { "epoch": 3.3056105610561057, "grad_norm": 1.1596111059188843, "learning_rate": 2.4901960784313726e-05, "loss": 0.0573, "step": 626 }, { "epoch": 3.3108910891089107, "grad_norm": 1.3220415115356445, "learning_rate": 2.485294117647059e-05, "loss": 0.058, "step": 627 }, { "epoch": 3.316171617161716, "grad_norm": 1.2258095741271973, "learning_rate": 2.4803921568627453e-05, "loss": 0.0471, "step": 628 }, { "epoch": 3.3214521452145216, "grad_norm": 1.2251251935958862, "learning_rate": 2.4754901960784317e-05, "loss": 0.0647, "step": 629 }, { "epoch": 3.3267326732673266, "grad_norm": 1.299035906791687, "learning_rate": 2.4705882352941178e-05, "loss": 0.056, "step": 630 }, { "epoch": 3.332013201320132, "grad_norm": 1.5529186725616455, "learning_rate": 2.465686274509804e-05, "loss": 0.084, "step": 631 }, { "epoch": 3.3372937293729374, "grad_norm": 1.0892446041107178, "learning_rate": 2.4607843137254902e-05, "loss": 0.0453, "step": 632 }, { "epoch": 3.3425742574257424, "grad_norm": 1.1661828756332397, "learning_rate": 2.4558823529411766e-05, "loss": 0.0485, "step": 633 }, { "epoch": 3.347854785478548, "grad_norm": 1.0938224792480469, "learning_rate": 2.4509803921568626e-05, "loss": 0.0469, "step": 634 }, { "epoch": 3.3531353135313533, "grad_norm": 1.2503447532653809, "learning_rate": 2.446078431372549e-05, "loss": 0.0777, "step": 635 }, { "epoch": 3.3584158415841583, "grad_norm": 1.069814682006836, "learning_rate": 2.4411764705882354e-05, "loss": 0.049, "step": 636 }, { "epoch": 3.3636963696369637, "grad_norm": 1.321007251739502, "learning_rate": 2.4362745098039215e-05, "loss": 0.0596, "step": 637 }, { "epoch": 3.368976897689769, "grad_norm": 1.5562186241149902, "learning_rate": 2.431372549019608e-05, "loss": 0.0883, "step": 638 }, { "epoch": 3.374257425742574, "grad_norm": 1.3883391618728638, "learning_rate": 2.4264705882352942e-05, "loss": 0.0777, "step": 639 }, { "epoch": 3.3795379537953796, "grad_norm": 1.6028392314910889, "learning_rate": 2.4215686274509806e-05, "loss": 0.0769, "step": 640 }, { "epoch": 3.384818481848185, "grad_norm": 1.777230978012085, "learning_rate": 2.4166666666666667e-05, "loss": 0.0943, "step": 641 }, { "epoch": 3.39009900990099, "grad_norm": 1.0344568490982056, "learning_rate": 2.411764705882353e-05, "loss": 0.0452, "step": 642 }, { "epoch": 3.3953795379537954, "grad_norm": 1.2247084379196167, "learning_rate": 2.4068627450980395e-05, "loss": 0.0516, "step": 643 }, { "epoch": 3.4006600660066004, "grad_norm": 1.2037266492843628, "learning_rate": 2.401960784313726e-05, "loss": 0.0486, "step": 644 }, { "epoch": 3.405940594059406, "grad_norm": 1.3176463842391968, "learning_rate": 2.397058823529412e-05, "loss": 0.0522, "step": 645 }, { "epoch": 3.4112211221122113, "grad_norm": 1.6888436079025269, "learning_rate": 2.3921568627450983e-05, "loss": 0.1001, "step": 646 }, { "epoch": 3.4165016501650163, "grad_norm": 1.296751856803894, "learning_rate": 2.3872549019607847e-05, "loss": 0.0656, "step": 647 }, { "epoch": 3.4217821782178217, "grad_norm": 1.126638650894165, "learning_rate": 2.3823529411764707e-05, "loss": 0.0477, "step": 648 }, { "epoch": 3.427062706270627, "grad_norm": 1.394085168838501, "learning_rate": 2.3774509803921568e-05, "loss": 0.0761, "step": 649 }, { "epoch": 3.432343234323432, "grad_norm": 1.4828298091888428, "learning_rate": 2.372549019607843e-05, "loss": 0.0683, "step": 650 }, { "epoch": 3.4376237623762376, "grad_norm": 1.4716849327087402, "learning_rate": 2.3676470588235295e-05, "loss": 0.0855, "step": 651 }, { "epoch": 3.442904290429043, "grad_norm": 1.3064403533935547, "learning_rate": 2.3627450980392156e-05, "loss": 0.0644, "step": 652 }, { "epoch": 3.448184818481848, "grad_norm": 1.4186874628067017, "learning_rate": 2.357843137254902e-05, "loss": 0.0556, "step": 653 }, { "epoch": 3.4534653465346534, "grad_norm": 1.3081494569778442, "learning_rate": 2.3529411764705884e-05, "loss": 0.0663, "step": 654 }, { "epoch": 3.458745874587459, "grad_norm": 1.174717664718628, "learning_rate": 2.3480392156862748e-05, "loss": 0.0704, "step": 655 }, { "epoch": 3.464026402640264, "grad_norm": 1.310571551322937, "learning_rate": 2.3431372549019608e-05, "loss": 0.0555, "step": 656 }, { "epoch": 3.4693069306930693, "grad_norm": 1.3356480598449707, "learning_rate": 2.3382352941176472e-05, "loss": 0.0565, "step": 657 }, { "epoch": 3.4745874587458747, "grad_norm": 1.5645304918289185, "learning_rate": 2.3333333333333336e-05, "loss": 0.0885, "step": 658 }, { "epoch": 3.4798679867986797, "grad_norm": 1.6264077425003052, "learning_rate": 2.3284313725490196e-05, "loss": 0.0936, "step": 659 }, { "epoch": 3.485148514851485, "grad_norm": 1.1637386083602905, "learning_rate": 2.323529411764706e-05, "loss": 0.0499, "step": 660 }, { "epoch": 3.4904290429042906, "grad_norm": 1.3853108882904053, "learning_rate": 2.3186274509803924e-05, "loss": 0.0781, "step": 661 }, { "epoch": 3.4957095709570956, "grad_norm": 1.1978479623794556, "learning_rate": 2.3137254901960788e-05, "loss": 0.0524, "step": 662 }, { "epoch": 3.500990099009901, "grad_norm": 1.1041313409805298, "learning_rate": 2.308823529411765e-05, "loss": 0.054, "step": 663 }, { "epoch": 3.5062706270627064, "grad_norm": 0.9631710052490234, "learning_rate": 2.303921568627451e-05, "loss": 0.0393, "step": 664 }, { "epoch": 3.5115511551155114, "grad_norm": 1.166167974472046, "learning_rate": 2.2990196078431373e-05, "loss": 0.0401, "step": 665 }, { "epoch": 3.516831683168317, "grad_norm": 1.7196093797683716, "learning_rate": 2.2941176470588237e-05, "loss": 0.0686, "step": 666 }, { "epoch": 3.5221122112211223, "grad_norm": 1.4491196870803833, "learning_rate": 2.2892156862745097e-05, "loss": 0.0721, "step": 667 }, { "epoch": 3.5273927392739273, "grad_norm": 1.724544882774353, "learning_rate": 2.284313725490196e-05, "loss": 0.0789, "step": 668 }, { "epoch": 3.5326732673267327, "grad_norm": 1.2871899604797363, "learning_rate": 2.2794117647058825e-05, "loss": 0.0502, "step": 669 }, { "epoch": 3.537953795379538, "grad_norm": 1.7234700918197632, "learning_rate": 2.2745098039215685e-05, "loss": 0.0896, "step": 670 }, { "epoch": 3.543234323432343, "grad_norm": 1.6124573945999146, "learning_rate": 2.269607843137255e-05, "loss": 0.0899, "step": 671 }, { "epoch": 3.5485148514851486, "grad_norm": 1.1992857456207275, "learning_rate": 2.2647058823529413e-05, "loss": 0.0514, "step": 672 }, { "epoch": 3.553795379537954, "grad_norm": 1.2639023065567017, "learning_rate": 2.2598039215686277e-05, "loss": 0.0584, "step": 673 }, { "epoch": 3.559075907590759, "grad_norm": 1.1662899255752563, "learning_rate": 2.2549019607843138e-05, "loss": 0.0551, "step": 674 }, { "epoch": 3.5643564356435644, "grad_norm": 1.27886962890625, "learning_rate": 2.25e-05, "loss": 0.0706, "step": 675 }, { "epoch": 3.56963696369637, "grad_norm": 1.487029790878296, "learning_rate": 2.2450980392156865e-05, "loss": 0.0973, "step": 676 }, { "epoch": 3.574917491749175, "grad_norm": 1.0966662168502808, "learning_rate": 2.2401960784313726e-05, "loss": 0.0483, "step": 677 }, { "epoch": 3.5801980198019803, "grad_norm": 1.592883825302124, "learning_rate": 2.235294117647059e-05, "loss": 0.097, "step": 678 }, { "epoch": 3.5854785478547857, "grad_norm": 1.1297260522842407, "learning_rate": 2.2303921568627454e-05, "loss": 0.052, "step": 679 }, { "epoch": 3.5907590759075907, "grad_norm": 1.3549542427062988, "learning_rate": 2.2254901960784314e-05, "loss": 0.0636, "step": 680 }, { "epoch": 3.596039603960396, "grad_norm": 1.4858061075210571, "learning_rate": 2.2205882352941178e-05, "loss": 0.0823, "step": 681 }, { "epoch": 3.6013201320132016, "grad_norm": 1.2688394784927368, "learning_rate": 2.215686274509804e-05, "loss": 0.0621, "step": 682 }, { "epoch": 3.6066006600660065, "grad_norm": 1.2265433073043823, "learning_rate": 2.2107843137254902e-05, "loss": 0.0521, "step": 683 }, { "epoch": 3.611881188118812, "grad_norm": 1.5389267206192017, "learning_rate": 2.2058823529411766e-05, "loss": 0.0644, "step": 684 }, { "epoch": 3.6171617161716174, "grad_norm": 1.0617897510528564, "learning_rate": 2.2009803921568627e-05, "loss": 0.0487, "step": 685 }, { "epoch": 3.6224422442244224, "grad_norm": 1.5505889654159546, "learning_rate": 2.196078431372549e-05, "loss": 0.0727, "step": 686 }, { "epoch": 3.627722772277228, "grad_norm": 1.6774746179580688, "learning_rate": 2.1911764705882354e-05, "loss": 0.0763, "step": 687 }, { "epoch": 3.6330033003300333, "grad_norm": 1.253771185874939, "learning_rate": 2.1862745098039218e-05, "loss": 0.0629, "step": 688 }, { "epoch": 3.6382838283828383, "grad_norm": 1.324569821357727, "learning_rate": 2.181372549019608e-05, "loss": 0.0673, "step": 689 }, { "epoch": 3.6435643564356437, "grad_norm": 1.7756513357162476, "learning_rate": 2.1764705882352943e-05, "loss": 0.1118, "step": 690 }, { "epoch": 3.6488448844884487, "grad_norm": 1.2172956466674805, "learning_rate": 2.1715686274509807e-05, "loss": 0.0553, "step": 691 }, { "epoch": 3.654125412541254, "grad_norm": 1.201130747795105, "learning_rate": 2.1666666666666667e-05, "loss": 0.0654, "step": 692 }, { "epoch": 3.6594059405940595, "grad_norm": 1.3230106830596924, "learning_rate": 2.161764705882353e-05, "loss": 0.0647, "step": 693 }, { "epoch": 3.6646864686468645, "grad_norm": 1.3373692035675049, "learning_rate": 2.1568627450980395e-05, "loss": 0.0691, "step": 694 }, { "epoch": 3.66996699669967, "grad_norm": 1.423130750656128, "learning_rate": 2.151960784313726e-05, "loss": 0.0683, "step": 695 }, { "epoch": 3.6752475247524754, "grad_norm": 1.2763397693634033, "learning_rate": 2.1470588235294116e-05, "loss": 0.0589, "step": 696 }, { "epoch": 3.6805280528052804, "grad_norm": 1.749027967453003, "learning_rate": 2.142156862745098e-05, "loss": 0.0918, "step": 697 }, { "epoch": 3.685808580858086, "grad_norm": 1.3163336515426636, "learning_rate": 2.1372549019607844e-05, "loss": 0.0585, "step": 698 }, { "epoch": 3.691089108910891, "grad_norm": 1.6769047975540161, "learning_rate": 2.1323529411764707e-05, "loss": 0.0967, "step": 699 }, { "epoch": 3.6963696369636962, "grad_norm": 1.2727911472320557, "learning_rate": 2.1274509803921568e-05, "loss": 0.0713, "step": 700 }, { "epoch": 3.7016501650165017, "grad_norm": 1.4433225393295288, "learning_rate": 2.1225490196078432e-05, "loss": 0.0783, "step": 701 }, { "epoch": 3.7069306930693067, "grad_norm": 1.3980076313018799, "learning_rate": 2.1176470588235296e-05, "loss": 0.0677, "step": 702 }, { "epoch": 3.712211221122112, "grad_norm": 1.3446606397628784, "learning_rate": 2.1127450980392156e-05, "loss": 0.0746, "step": 703 }, { "epoch": 3.7174917491749175, "grad_norm": 1.2781853675842285, "learning_rate": 2.107843137254902e-05, "loss": 0.0545, "step": 704 }, { "epoch": 3.7227722772277225, "grad_norm": 1.4710532426834106, "learning_rate": 2.1029411764705884e-05, "loss": 0.0612, "step": 705 }, { "epoch": 3.728052805280528, "grad_norm": 1.2814794778823853, "learning_rate": 2.0980392156862748e-05, "loss": 0.0718, "step": 706 }, { "epoch": 3.7333333333333334, "grad_norm": 1.4661047458648682, "learning_rate": 2.0931372549019608e-05, "loss": 0.0791, "step": 707 }, { "epoch": 3.7386138613861384, "grad_norm": 1.37482750415802, "learning_rate": 2.0882352941176472e-05, "loss": 0.0737, "step": 708 }, { "epoch": 3.743894389438944, "grad_norm": 1.3684672117233276, "learning_rate": 2.0833333333333336e-05, "loss": 0.0743, "step": 709 }, { "epoch": 3.7491749174917492, "grad_norm": 0.859553873538971, "learning_rate": 2.0784313725490197e-05, "loss": 0.0356, "step": 710 }, { "epoch": 3.7544554455445542, "grad_norm": 1.2247698307037354, "learning_rate": 2.073529411764706e-05, "loss": 0.0671, "step": 711 }, { "epoch": 3.7597359735973597, "grad_norm": 1.2793521881103516, "learning_rate": 2.068627450980392e-05, "loss": 0.0692, "step": 712 }, { "epoch": 3.765016501650165, "grad_norm": 1.26449453830719, "learning_rate": 2.0637254901960785e-05, "loss": 0.0622, "step": 713 }, { "epoch": 3.77029702970297, "grad_norm": 1.1848664283752441, "learning_rate": 2.058823529411765e-05, "loss": 0.0535, "step": 714 }, { "epoch": 3.7755775577557755, "grad_norm": 1.034454107284546, "learning_rate": 2.053921568627451e-05, "loss": 0.0403, "step": 715 }, { "epoch": 3.780858085808581, "grad_norm": 1.005811333656311, "learning_rate": 2.0490196078431373e-05, "loss": 0.0489, "step": 716 }, { "epoch": 3.786138613861386, "grad_norm": 1.294252872467041, "learning_rate": 2.0441176470588237e-05, "loss": 0.0599, "step": 717 }, { "epoch": 3.7914191419141914, "grad_norm": 1.2683149576187134, "learning_rate": 2.0392156862745097e-05, "loss": 0.0527, "step": 718 }, { "epoch": 3.796699669966997, "grad_norm": 1.4859176874160767, "learning_rate": 2.034313725490196e-05, "loss": 0.0748, "step": 719 }, { "epoch": 3.801980198019802, "grad_norm": 1.5484555959701538, "learning_rate": 2.0294117647058825e-05, "loss": 0.0694, "step": 720 }, { "epoch": 3.8072607260726072, "grad_norm": 1.0861327648162842, "learning_rate": 2.024509803921569e-05, "loss": 0.0489, "step": 721 }, { "epoch": 3.8125412541254127, "grad_norm": 2.1083528995513916, "learning_rate": 2.019607843137255e-05, "loss": 0.0996, "step": 722 }, { "epoch": 3.8178217821782177, "grad_norm": 1.1636090278625488, "learning_rate": 2.0147058823529413e-05, "loss": 0.0439, "step": 723 }, { "epoch": 3.823102310231023, "grad_norm": 1.4956183433532715, "learning_rate": 2.0098039215686277e-05, "loss": 0.0771, "step": 724 }, { "epoch": 3.8283828382838285, "grad_norm": 1.4790761470794678, "learning_rate": 2.0049019607843138e-05, "loss": 0.0561, "step": 725 }, { "epoch": 3.8336633663366335, "grad_norm": 1.4751077890396118, "learning_rate": 2e-05, "loss": 0.0788, "step": 726 }, { "epoch": 3.838943894389439, "grad_norm": 1.2702194452285767, "learning_rate": 1.9950980392156866e-05, "loss": 0.07, "step": 727 }, { "epoch": 3.8442244224422444, "grad_norm": 0.9263429045677185, "learning_rate": 1.9901960784313726e-05, "loss": 0.0359, "step": 728 }, { "epoch": 3.8495049504950494, "grad_norm": 1.4288864135742188, "learning_rate": 1.9852941176470586e-05, "loss": 0.0869, "step": 729 }, { "epoch": 3.854785478547855, "grad_norm": 1.3054994344711304, "learning_rate": 1.980392156862745e-05, "loss": 0.0755, "step": 730 }, { "epoch": 3.8600660066006602, "grad_norm": 1.422020435333252, "learning_rate": 1.9754901960784314e-05, "loss": 0.072, "step": 731 }, { "epoch": 3.8653465346534652, "grad_norm": 1.2863701581954956, "learning_rate": 1.9705882352941178e-05, "loss": 0.0676, "step": 732 }, { "epoch": 3.8706270627062707, "grad_norm": 1.4083125591278076, "learning_rate": 1.965686274509804e-05, "loss": 0.0805, "step": 733 }, { "epoch": 3.875907590759076, "grad_norm": 1.275930404663086, "learning_rate": 1.9607843137254903e-05, "loss": 0.0737, "step": 734 }, { "epoch": 3.881188118811881, "grad_norm": 1.2777963876724243, "learning_rate": 1.9558823529411766e-05, "loss": 0.0674, "step": 735 }, { "epoch": 3.8864686468646865, "grad_norm": 1.145695686340332, "learning_rate": 1.9509803921568627e-05, "loss": 0.0593, "step": 736 }, { "epoch": 3.891749174917492, "grad_norm": 1.13607919216156, "learning_rate": 1.946078431372549e-05, "loss": 0.0481, "step": 737 }, { "epoch": 3.897029702970297, "grad_norm": 1.5105829238891602, "learning_rate": 1.9411764705882355e-05, "loss": 0.1056, "step": 738 }, { "epoch": 3.9023102310231024, "grad_norm": 1.080869197845459, "learning_rate": 1.936274509803922e-05, "loss": 0.0469, "step": 739 }, { "epoch": 3.907590759075908, "grad_norm": 1.6012327671051025, "learning_rate": 1.931372549019608e-05, "loss": 0.0883, "step": 740 }, { "epoch": 3.912871287128713, "grad_norm": 1.3881632089614868, "learning_rate": 1.9264705882352943e-05, "loss": 0.0736, "step": 741 }, { "epoch": 3.9181518151815182, "grad_norm": 1.2698092460632324, "learning_rate": 1.9215686274509807e-05, "loss": 0.061, "step": 742 }, { "epoch": 3.9234323432343237, "grad_norm": 1.7859208583831787, "learning_rate": 1.9166666666666667e-05, "loss": 0.0981, "step": 743 }, { "epoch": 3.9287128712871286, "grad_norm": 1.1902238130569458, "learning_rate": 1.9117647058823528e-05, "loss": 0.0489, "step": 744 }, { "epoch": 3.933993399339934, "grad_norm": 1.3295652866363525, "learning_rate": 1.906862745098039e-05, "loss": 0.0634, "step": 745 }, { "epoch": 3.9392739273927395, "grad_norm": 1.2844423055648804, "learning_rate": 1.9019607843137255e-05, "loss": 0.0759, "step": 746 }, { "epoch": 3.9445544554455445, "grad_norm": 1.062388300895691, "learning_rate": 1.897058823529412e-05, "loss": 0.0449, "step": 747 }, { "epoch": 3.94983498349835, "grad_norm": 1.51730477809906, "learning_rate": 1.892156862745098e-05, "loss": 0.1383, "step": 748 }, { "epoch": 3.9551155115511554, "grad_norm": 1.5683430433273315, "learning_rate": 1.8872549019607844e-05, "loss": 0.0915, "step": 749 }, { "epoch": 3.9603960396039604, "grad_norm": 1.20012366771698, "learning_rate": 1.8823529411764708e-05, "loss": 0.057, "step": 750 }, { "epoch": 3.965676567656766, "grad_norm": 1.351365089416504, "learning_rate": 1.8774509803921568e-05, "loss": 0.0663, "step": 751 }, { "epoch": 3.9709570957095712, "grad_norm": 1.5210295915603638, "learning_rate": 1.8725490196078432e-05, "loss": 0.0832, "step": 752 }, { "epoch": 3.976237623762376, "grad_norm": 1.5740433931350708, "learning_rate": 1.8676470588235296e-05, "loss": 0.0951, "step": 753 }, { "epoch": 3.9815181518151816, "grad_norm": 1.2916804552078247, "learning_rate": 1.862745098039216e-05, "loss": 0.0638, "step": 754 }, { "epoch": 3.9867986798679866, "grad_norm": 1.4912750720977783, "learning_rate": 1.857843137254902e-05, "loss": 0.0718, "step": 755 }, { "epoch": 3.992079207920792, "grad_norm": 1.2898180484771729, "learning_rate": 1.8529411764705884e-05, "loss": 0.0672, "step": 756 }, { "epoch": 3.9973597359735975, "grad_norm": 0.9569465517997742, "learning_rate": 1.8480392156862748e-05, "loss": 0.0425, "step": 757 }, { "epoch": 4.002640264026403, "grad_norm": 1.1869324445724487, "learning_rate": 1.843137254901961e-05, "loss": 0.0527, "step": 758 }, { "epoch": 4.007920792079208, "grad_norm": 0.7702716588973999, "learning_rate": 1.8382352941176472e-05, "loss": 0.0358, "step": 759 }, { "epoch": 4.013201320132013, "grad_norm": 0.6784669160842896, "learning_rate": 1.8333333333333333e-05, "loss": 0.0268, "step": 760 }, { "epoch": 4.018481848184819, "grad_norm": 0.8795797228813171, "learning_rate": 1.8284313725490197e-05, "loss": 0.0363, "step": 761 }, { "epoch": 4.023762376237624, "grad_norm": 0.9152675271034241, "learning_rate": 1.8235294117647057e-05, "loss": 0.0303, "step": 762 }, { "epoch": 4.029042904290429, "grad_norm": 0.685607373714447, "learning_rate": 1.818627450980392e-05, "loss": 0.0249, "step": 763 }, { "epoch": 4.034323432343235, "grad_norm": 0.8284955620765686, "learning_rate": 1.8137254901960785e-05, "loss": 0.0257, "step": 764 }, { "epoch": 4.03960396039604, "grad_norm": 0.8242619037628174, "learning_rate": 1.808823529411765e-05, "loss": 0.027, "step": 765 }, { "epoch": 4.044884488448845, "grad_norm": 1.0114595890045166, "learning_rate": 1.803921568627451e-05, "loss": 0.0222, "step": 766 }, { "epoch": 4.0501650165016505, "grad_norm": 0.6319472789764404, "learning_rate": 1.7990196078431373e-05, "loss": 0.0182, "step": 767 }, { "epoch": 4.0554455445544555, "grad_norm": 0.9475066661834717, "learning_rate": 1.7941176470588237e-05, "loss": 0.0269, "step": 768 }, { "epoch": 4.0607260726072605, "grad_norm": 1.0184468030929565, "learning_rate": 1.7892156862745098e-05, "loss": 0.0258, "step": 769 }, { "epoch": 4.066006600660066, "grad_norm": 1.226219654083252, "learning_rate": 1.784313725490196e-05, "loss": 0.036, "step": 770 }, { "epoch": 4.071287128712871, "grad_norm": 0.839829683303833, "learning_rate": 1.7794117647058825e-05, "loss": 0.027, "step": 771 }, { "epoch": 4.076567656765676, "grad_norm": 1.2068742513656616, "learning_rate": 1.774509803921569e-05, "loss": 0.0345, "step": 772 }, { "epoch": 4.081848184818482, "grad_norm": 1.1739568710327148, "learning_rate": 1.769607843137255e-05, "loss": 0.0334, "step": 773 }, { "epoch": 4.087128712871287, "grad_norm": 1.167466640472412, "learning_rate": 1.7647058823529414e-05, "loss": 0.0308, "step": 774 }, { "epoch": 4.092409240924092, "grad_norm": 1.0854191780090332, "learning_rate": 1.7598039215686277e-05, "loss": 0.0288, "step": 775 }, { "epoch": 4.097689768976898, "grad_norm": 0.8785208463668823, "learning_rate": 1.7549019607843138e-05, "loss": 0.0298, "step": 776 }, { "epoch": 4.102970297029703, "grad_norm": 1.252079725265503, "learning_rate": 1.75e-05, "loss": 0.044, "step": 777 }, { "epoch": 4.108250825082508, "grad_norm": 1.1048779487609863, "learning_rate": 1.7450980392156862e-05, "loss": 0.0285, "step": 778 }, { "epoch": 4.113531353135314, "grad_norm": 1.3396317958831787, "learning_rate": 1.7401960784313726e-05, "loss": 0.0457, "step": 779 }, { "epoch": 4.118811881188119, "grad_norm": 1.165590763092041, "learning_rate": 1.735294117647059e-05, "loss": 0.0319, "step": 780 }, { "epoch": 4.124092409240924, "grad_norm": 0.9108593463897705, "learning_rate": 1.730392156862745e-05, "loss": 0.0286, "step": 781 }, { "epoch": 4.12937293729373, "grad_norm": 0.9101956486701965, "learning_rate": 1.7254901960784314e-05, "loss": 0.0291, "step": 782 }, { "epoch": 4.134653465346535, "grad_norm": 1.055116891860962, "learning_rate": 1.720588235294118e-05, "loss": 0.0326, "step": 783 }, { "epoch": 4.13993399339934, "grad_norm": 0.9137353897094727, "learning_rate": 1.715686274509804e-05, "loss": 0.0312, "step": 784 }, { "epoch": 4.145214521452146, "grad_norm": 0.9962388277053833, "learning_rate": 1.7107843137254903e-05, "loss": 0.0317, "step": 785 }, { "epoch": 4.150495049504951, "grad_norm": 0.891434907913208, "learning_rate": 1.7058823529411767e-05, "loss": 0.0357, "step": 786 }, { "epoch": 4.155775577557756, "grad_norm": 0.9022417068481445, "learning_rate": 1.700980392156863e-05, "loss": 0.0346, "step": 787 }, { "epoch": 4.161056105610561, "grad_norm": 0.886080265045166, "learning_rate": 1.696078431372549e-05, "loss": 0.0308, "step": 788 }, { "epoch": 4.1663366336633665, "grad_norm": 1.155617356300354, "learning_rate": 1.6911764705882355e-05, "loss": 0.0904, "step": 789 }, { "epoch": 4.1716171617161715, "grad_norm": 1.101927399635315, "learning_rate": 1.686274509803922e-05, "loss": 0.0289, "step": 790 }, { "epoch": 4.1768976897689765, "grad_norm": 1.2381272315979004, "learning_rate": 1.681372549019608e-05, "loss": 0.0473, "step": 791 }, { "epoch": 4.182178217821782, "grad_norm": 1.2841603755950928, "learning_rate": 1.676470588235294e-05, "loss": 0.0392, "step": 792 }, { "epoch": 4.187458745874587, "grad_norm": 0.9901617765426636, "learning_rate": 1.6715686274509804e-05, "loss": 0.0274, "step": 793 }, { "epoch": 4.192739273927392, "grad_norm": 1.011318325996399, "learning_rate": 1.6666666666666667e-05, "loss": 0.0329, "step": 794 }, { "epoch": 4.198019801980198, "grad_norm": 0.8173012733459473, "learning_rate": 1.6617647058823528e-05, "loss": 0.0272, "step": 795 }, { "epoch": 4.203300330033003, "grad_norm": 1.3152134418487549, "learning_rate": 1.6568627450980392e-05, "loss": 0.0309, "step": 796 }, { "epoch": 4.208580858085808, "grad_norm": 0.9203322529792786, "learning_rate": 1.6519607843137256e-05, "loss": 0.0287, "step": 797 }, { "epoch": 4.213861386138614, "grad_norm": 0.8306743502616882, "learning_rate": 1.647058823529412e-05, "loss": 0.0291, "step": 798 }, { "epoch": 4.219141914191419, "grad_norm": 1.1384245157241821, "learning_rate": 1.642156862745098e-05, "loss": 0.0295, "step": 799 }, { "epoch": 4.224422442244224, "grad_norm": 1.0017738342285156, "learning_rate": 1.6372549019607844e-05, "loss": 0.0347, "step": 800 }, { "epoch": 4.22970297029703, "grad_norm": 0.7871435284614563, "learning_rate": 1.6323529411764708e-05, "loss": 0.0277, "step": 801 }, { "epoch": 4.234983498349835, "grad_norm": 0.9366597533226013, "learning_rate": 1.627450980392157e-05, "loss": 0.0323, "step": 802 }, { "epoch": 4.24026402640264, "grad_norm": 0.9586142301559448, "learning_rate": 1.6225490196078432e-05, "loss": 0.0212, "step": 803 }, { "epoch": 4.245544554455446, "grad_norm": 1.21640145778656, "learning_rate": 1.6176470588235296e-05, "loss": 0.0357, "step": 804 }, { "epoch": 4.250825082508251, "grad_norm": 0.9384037852287292, "learning_rate": 1.612745098039216e-05, "loss": 0.0315, "step": 805 }, { "epoch": 4.256105610561056, "grad_norm": 1.0717084407806396, "learning_rate": 1.607843137254902e-05, "loss": 0.0337, "step": 806 }, { "epoch": 4.261386138613862, "grad_norm": 0.9708730578422546, "learning_rate": 1.6029411764705884e-05, "loss": 0.0319, "step": 807 }, { "epoch": 4.266666666666667, "grad_norm": 0.7535268068313599, "learning_rate": 1.5980392156862745e-05, "loss": 0.0294, "step": 808 }, { "epoch": 4.271947194719472, "grad_norm": 1.0513904094696045, "learning_rate": 1.593137254901961e-05, "loss": 0.0262, "step": 809 }, { "epoch": 4.2772277227722775, "grad_norm": 1.4030704498291016, "learning_rate": 1.588235294117647e-05, "loss": 0.0436, "step": 810 }, { "epoch": 4.2825082508250825, "grad_norm": 0.5738725662231445, "learning_rate": 1.5833333333333333e-05, "loss": 0.0201, "step": 811 }, { "epoch": 4.2877887788778875, "grad_norm": 1.305862307548523, "learning_rate": 1.5784313725490197e-05, "loss": 0.043, "step": 812 }, { "epoch": 4.293069306930693, "grad_norm": 0.9697719812393188, "learning_rate": 1.573529411764706e-05, "loss": 0.0306, "step": 813 }, { "epoch": 4.298349834983498, "grad_norm": 0.8352718353271484, "learning_rate": 1.568627450980392e-05, "loss": 0.0269, "step": 814 }, { "epoch": 4.303630363036303, "grad_norm": 1.000013828277588, "learning_rate": 1.5637254901960785e-05, "loss": 0.0272, "step": 815 }, { "epoch": 4.308910891089109, "grad_norm": 0.6394527554512024, "learning_rate": 1.558823529411765e-05, "loss": 0.0224, "step": 816 }, { "epoch": 4.314191419141914, "grad_norm": 0.9506754875183105, "learning_rate": 1.553921568627451e-05, "loss": 0.0293, "step": 817 }, { "epoch": 4.319471947194719, "grad_norm": 0.9406654834747314, "learning_rate": 1.5490196078431373e-05, "loss": 0.0261, "step": 818 }, { "epoch": 4.324752475247525, "grad_norm": 1.1264249086380005, "learning_rate": 1.5441176470588237e-05, "loss": 0.0347, "step": 819 }, { "epoch": 4.33003300330033, "grad_norm": 0.9461225271224976, "learning_rate": 1.53921568627451e-05, "loss": 0.0261, "step": 820 }, { "epoch": 4.335313531353135, "grad_norm": 1.1620982885360718, "learning_rate": 1.5343137254901962e-05, "loss": 0.0309, "step": 821 }, { "epoch": 4.340594059405941, "grad_norm": 0.8539828658103943, "learning_rate": 1.5294117647058826e-05, "loss": 0.028, "step": 822 }, { "epoch": 4.345874587458746, "grad_norm": 1.1822220087051392, "learning_rate": 1.5245098039215688e-05, "loss": 0.0352, "step": 823 }, { "epoch": 4.351155115511551, "grad_norm": 1.000964641571045, "learning_rate": 1.5196078431372548e-05, "loss": 0.0325, "step": 824 }, { "epoch": 4.356435643564357, "grad_norm": 1.211593508720398, "learning_rate": 1.5147058823529412e-05, "loss": 0.0394, "step": 825 }, { "epoch": 4.361716171617162, "grad_norm": 1.0759446620941162, "learning_rate": 1.5098039215686274e-05, "loss": 0.0299, "step": 826 }, { "epoch": 4.366996699669967, "grad_norm": 0.7947413921356201, "learning_rate": 1.5049019607843137e-05, "loss": 0.0269, "step": 827 }, { "epoch": 4.372277227722773, "grad_norm": 1.283584713935852, "learning_rate": 1.5e-05, "loss": 0.0396, "step": 828 }, { "epoch": 4.377557755775578, "grad_norm": 1.0424344539642334, "learning_rate": 1.4950980392156863e-05, "loss": 0.0303, "step": 829 }, { "epoch": 4.382838283828383, "grad_norm": 1.1539040803909302, "learning_rate": 1.4901960784313726e-05, "loss": 0.0369, "step": 830 }, { "epoch": 4.3881188118811885, "grad_norm": 0.9597378373146057, "learning_rate": 1.4852941176470589e-05, "loss": 0.0341, "step": 831 }, { "epoch": 4.3933993399339935, "grad_norm": 1.133381724357605, "learning_rate": 1.4803921568627453e-05, "loss": 0.032, "step": 832 }, { "epoch": 4.398679867986798, "grad_norm": 0.8584814071655273, "learning_rate": 1.4754901960784315e-05, "loss": 0.0274, "step": 833 }, { "epoch": 4.403960396039604, "grad_norm": 0.7152499556541443, "learning_rate": 1.4705882352941177e-05, "loss": 0.0228, "step": 834 }, { "epoch": 4.409240924092409, "grad_norm": 1.0020297765731812, "learning_rate": 1.465686274509804e-05, "loss": 0.0405, "step": 835 }, { "epoch": 4.414521452145214, "grad_norm": 1.236913800239563, "learning_rate": 1.4607843137254903e-05, "loss": 0.0399, "step": 836 }, { "epoch": 4.41980198019802, "grad_norm": 1.0684503316879272, "learning_rate": 1.4558823529411767e-05, "loss": 0.0375, "step": 837 }, { "epoch": 4.425082508250825, "grad_norm": 1.2516707181930542, "learning_rate": 1.4509803921568629e-05, "loss": 0.0379, "step": 838 }, { "epoch": 4.43036303630363, "grad_norm": 1.063377857208252, "learning_rate": 1.4460784313725493e-05, "loss": 0.0389, "step": 839 }, { "epoch": 4.435643564356436, "grad_norm": 0.740829586982727, "learning_rate": 1.4411764705882352e-05, "loss": 0.0258, "step": 840 }, { "epoch": 4.440924092409241, "grad_norm": 0.9484118223190308, "learning_rate": 1.4362745098039216e-05, "loss": 0.0339, "step": 841 }, { "epoch": 4.446204620462046, "grad_norm": 1.024625539779663, "learning_rate": 1.4313725490196078e-05, "loss": 0.0329, "step": 842 }, { "epoch": 4.451485148514852, "grad_norm": 0.8846271634101868, "learning_rate": 1.4264705882352942e-05, "loss": 0.0293, "step": 843 }, { "epoch": 4.456765676567657, "grad_norm": 0.6733059883117676, "learning_rate": 1.4215686274509804e-05, "loss": 0.0224, "step": 844 }, { "epoch": 4.462046204620462, "grad_norm": 1.042506456375122, "learning_rate": 1.4166666666666668e-05, "loss": 0.0273, "step": 845 }, { "epoch": 4.467326732673268, "grad_norm": 0.9040712118148804, "learning_rate": 1.411764705882353e-05, "loss": 0.0293, "step": 846 }, { "epoch": 4.472607260726073, "grad_norm": 0.8816120028495789, "learning_rate": 1.4068627450980392e-05, "loss": 0.0264, "step": 847 }, { "epoch": 4.477887788778878, "grad_norm": 0.9503970146179199, "learning_rate": 1.4019607843137256e-05, "loss": 0.0386, "step": 848 }, { "epoch": 4.483168316831684, "grad_norm": 0.6163874268531799, "learning_rate": 1.3970588235294118e-05, "loss": 0.0179, "step": 849 }, { "epoch": 4.488448844884489, "grad_norm": 1.139955759048462, "learning_rate": 1.3921568627450982e-05, "loss": 0.0364, "step": 850 }, { "epoch": 4.493729372937294, "grad_norm": 0.8715453147888184, "learning_rate": 1.3872549019607844e-05, "loss": 0.0254, "step": 851 }, { "epoch": 4.4990099009900995, "grad_norm": 0.9393417835235596, "learning_rate": 1.3823529411764708e-05, "loss": 0.033, "step": 852 }, { "epoch": 4.504290429042904, "grad_norm": 1.1215251684188843, "learning_rate": 1.377450980392157e-05, "loss": 0.0357, "step": 853 }, { "epoch": 4.509570957095709, "grad_norm": 0.9020299911499023, "learning_rate": 1.3725490196078432e-05, "loss": 0.0257, "step": 854 }, { "epoch": 4.514851485148515, "grad_norm": 1.0256519317626953, "learning_rate": 1.3676470588235296e-05, "loss": 0.0271, "step": 855 }, { "epoch": 4.52013201320132, "grad_norm": 0.7362510561943054, "learning_rate": 1.3627450980392157e-05, "loss": 0.0214, "step": 856 }, { "epoch": 4.525412541254125, "grad_norm": 1.1379399299621582, "learning_rate": 1.3578431372549019e-05, "loss": 0.0337, "step": 857 }, { "epoch": 4.530693069306931, "grad_norm": 1.2522259950637817, "learning_rate": 1.3529411764705883e-05, "loss": 0.0401, "step": 858 }, { "epoch": 4.535973597359736, "grad_norm": 1.050032138824463, "learning_rate": 1.3480392156862745e-05, "loss": 0.0346, "step": 859 }, { "epoch": 4.541254125412541, "grad_norm": 0.9979643821716309, "learning_rate": 1.3431372549019607e-05, "loss": 0.0283, "step": 860 }, { "epoch": 4.546534653465347, "grad_norm": 0.9979017972946167, "learning_rate": 1.3382352941176471e-05, "loss": 0.0302, "step": 861 }, { "epoch": 4.551815181518152, "grad_norm": 0.9042947292327881, "learning_rate": 1.3333333333333333e-05, "loss": 0.0299, "step": 862 }, { "epoch": 4.557095709570957, "grad_norm": 1.0963839292526245, "learning_rate": 1.3284313725490197e-05, "loss": 0.0331, "step": 863 }, { "epoch": 4.562376237623763, "grad_norm": 0.9561034440994263, "learning_rate": 1.323529411764706e-05, "loss": 0.025, "step": 864 }, { "epoch": 4.567656765676568, "grad_norm": 1.147443175315857, "learning_rate": 1.3186274509803923e-05, "loss": 0.0363, "step": 865 }, { "epoch": 4.572937293729373, "grad_norm": 1.173801064491272, "learning_rate": 1.3137254901960785e-05, "loss": 0.0384, "step": 866 }, { "epoch": 4.578217821782178, "grad_norm": 1.3326085805892944, "learning_rate": 1.3088235294117648e-05, "loss": 0.0428, "step": 867 }, { "epoch": 4.583498349834984, "grad_norm": 0.8449905514717102, "learning_rate": 1.3039215686274511e-05, "loss": 0.02, "step": 868 }, { "epoch": 4.588778877887789, "grad_norm": 0.9651418924331665, "learning_rate": 1.2990196078431374e-05, "loss": 0.03, "step": 869 }, { "epoch": 4.594059405940594, "grad_norm": 0.7175216674804688, "learning_rate": 1.2941176470588238e-05, "loss": 0.0243, "step": 870 }, { "epoch": 4.5993399339934, "grad_norm": 0.6259102821350098, "learning_rate": 1.28921568627451e-05, "loss": 0.0212, "step": 871 }, { "epoch": 4.604620462046205, "grad_norm": 0.7783799171447754, "learning_rate": 1.284313725490196e-05, "loss": 0.0261, "step": 872 }, { "epoch": 4.6099009900990096, "grad_norm": 1.1052485704421997, "learning_rate": 1.2794117647058822e-05, "loss": 0.0413, "step": 873 }, { "epoch": 4.615181518151815, "grad_norm": 1.0588634014129639, "learning_rate": 1.2745098039215686e-05, "loss": 0.0324, "step": 874 }, { "epoch": 4.62046204620462, "grad_norm": 0.8361983299255371, "learning_rate": 1.2696078431372548e-05, "loss": 0.0263, "step": 875 }, { "epoch": 4.625742574257425, "grad_norm": 0.9796653985977173, "learning_rate": 1.2647058823529412e-05, "loss": 0.0357, "step": 876 }, { "epoch": 4.631023102310231, "grad_norm": 1.612053394317627, "learning_rate": 1.2598039215686275e-05, "loss": 0.0404, "step": 877 }, { "epoch": 4.636303630363036, "grad_norm": 0.853406548500061, "learning_rate": 1.2549019607843138e-05, "loss": 0.0266, "step": 878 }, { "epoch": 4.641584158415841, "grad_norm": 1.0649423599243164, "learning_rate": 1.25e-05, "loss": 0.0373, "step": 879 }, { "epoch": 4.646864686468647, "grad_norm": 1.1256874799728394, "learning_rate": 1.2450980392156863e-05, "loss": 0.0251, "step": 880 }, { "epoch": 4.652145214521452, "grad_norm": 0.8546039462089539, "learning_rate": 1.2401960784313727e-05, "loss": 0.0272, "step": 881 }, { "epoch": 4.657425742574257, "grad_norm": 0.7499862313270569, "learning_rate": 1.2352941176470589e-05, "loss": 0.0227, "step": 882 }, { "epoch": 4.662706270627063, "grad_norm": 1.2291607856750488, "learning_rate": 1.2303921568627451e-05, "loss": 0.0416, "step": 883 }, { "epoch": 4.667986798679868, "grad_norm": 0.919152021408081, "learning_rate": 1.2254901960784313e-05, "loss": 0.0325, "step": 884 }, { "epoch": 4.673267326732673, "grad_norm": 0.8878404498100281, "learning_rate": 1.2205882352941177e-05, "loss": 0.0296, "step": 885 }, { "epoch": 4.678547854785479, "grad_norm": 1.1350431442260742, "learning_rate": 1.215686274509804e-05, "loss": 0.0372, "step": 886 }, { "epoch": 4.683828382838284, "grad_norm": 0.9311625957489014, "learning_rate": 1.2107843137254903e-05, "loss": 0.0324, "step": 887 }, { "epoch": 4.689108910891089, "grad_norm": 1.2680948972702026, "learning_rate": 1.2058823529411765e-05, "loss": 0.0427, "step": 888 }, { "epoch": 4.694389438943895, "grad_norm": 1.0892651081085205, "learning_rate": 1.200980392156863e-05, "loss": 0.0289, "step": 889 }, { "epoch": 4.6996699669967, "grad_norm": 0.9546079635620117, "learning_rate": 1.1960784313725491e-05, "loss": 0.0355, "step": 890 }, { "epoch": 4.704950495049505, "grad_norm": 1.0489920377731323, "learning_rate": 1.1911764705882354e-05, "loss": 0.0287, "step": 891 }, { "epoch": 4.710231023102311, "grad_norm": 0.8627532720565796, "learning_rate": 1.1862745098039216e-05, "loss": 0.03, "step": 892 }, { "epoch": 4.7155115511551156, "grad_norm": 1.1037335395812988, "learning_rate": 1.1813725490196078e-05, "loss": 0.0369, "step": 893 }, { "epoch": 4.7207920792079205, "grad_norm": 0.8661030530929565, "learning_rate": 1.1764705882352942e-05, "loss": 0.0267, "step": 894 }, { "epoch": 4.726072607260726, "grad_norm": 0.7932984828948975, "learning_rate": 1.1715686274509804e-05, "loss": 0.0254, "step": 895 }, { "epoch": 4.731353135313531, "grad_norm": 1.1760293245315552, "learning_rate": 1.1666666666666668e-05, "loss": 0.0408, "step": 896 }, { "epoch": 4.736633663366336, "grad_norm": 1.098027229309082, "learning_rate": 1.161764705882353e-05, "loss": 0.0304, "step": 897 }, { "epoch": 4.741914191419142, "grad_norm": 1.1287803649902344, "learning_rate": 1.1568627450980394e-05, "loss": 0.0305, "step": 898 }, { "epoch": 4.747194719471947, "grad_norm": 1.1614177227020264, "learning_rate": 1.1519607843137254e-05, "loss": 0.0436, "step": 899 }, { "epoch": 4.752475247524752, "grad_norm": 0.8995096683502197, "learning_rate": 1.1470588235294118e-05, "loss": 0.027, "step": 900 }, { "epoch": 4.757755775577558, "grad_norm": 1.0893324613571167, "learning_rate": 1.142156862745098e-05, "loss": 0.0304, "step": 901 }, { "epoch": 4.763036303630363, "grad_norm": 0.8086807131767273, "learning_rate": 1.1372549019607843e-05, "loss": 0.0304, "step": 902 }, { "epoch": 4.768316831683168, "grad_norm": 0.6953551769256592, "learning_rate": 1.1323529411764707e-05, "loss": 0.0248, "step": 903 }, { "epoch": 4.773597359735973, "grad_norm": 0.707534670829773, "learning_rate": 1.1274509803921569e-05, "loss": 0.0219, "step": 904 }, { "epoch": 4.778877887788779, "grad_norm": 0.9743644595146179, "learning_rate": 1.1225490196078433e-05, "loss": 0.0327, "step": 905 }, { "epoch": 4.784158415841584, "grad_norm": 0.9171704053878784, "learning_rate": 1.1176470588235295e-05, "loss": 0.0277, "step": 906 }, { "epoch": 4.789438943894389, "grad_norm": 1.019524097442627, "learning_rate": 1.1127450980392157e-05, "loss": 0.0308, "step": 907 }, { "epoch": 4.794719471947195, "grad_norm": 0.8753800392150879, "learning_rate": 1.107843137254902e-05, "loss": 0.0282, "step": 908 }, { "epoch": 4.8, "grad_norm": 1.0594887733459473, "learning_rate": 1.1029411764705883e-05, "loss": 0.0283, "step": 909 }, { "epoch": 4.805280528052805, "grad_norm": 0.9229967594146729, "learning_rate": 1.0980392156862745e-05, "loss": 0.0298, "step": 910 }, { "epoch": 4.810561056105611, "grad_norm": 0.8133540153503418, "learning_rate": 1.0931372549019609e-05, "loss": 0.0276, "step": 911 }, { "epoch": 4.815841584158416, "grad_norm": 0.8808728456497192, "learning_rate": 1.0882352941176471e-05, "loss": 0.0304, "step": 912 }, { "epoch": 4.821122112211221, "grad_norm": 1.0457746982574463, "learning_rate": 1.0833333333333334e-05, "loss": 0.0313, "step": 913 }, { "epoch": 4.8264026402640265, "grad_norm": 1.004420518875122, "learning_rate": 1.0784313725490197e-05, "loss": 0.0341, "step": 914 }, { "epoch": 4.8316831683168315, "grad_norm": 0.8167664408683777, "learning_rate": 1.0735294117647058e-05, "loss": 0.0254, "step": 915 }, { "epoch": 4.8369636963696365, "grad_norm": 0.591806948184967, "learning_rate": 1.0686274509803922e-05, "loss": 0.0191, "step": 916 }, { "epoch": 4.842244224422442, "grad_norm": 1.1454893350601196, "learning_rate": 1.0637254901960784e-05, "loss": 0.0394, "step": 917 }, { "epoch": 4.847524752475247, "grad_norm": 0.9010074734687805, "learning_rate": 1.0588235294117648e-05, "loss": 0.0287, "step": 918 }, { "epoch": 4.852805280528052, "grad_norm": 0.9918347597122192, "learning_rate": 1.053921568627451e-05, "loss": 0.0288, "step": 919 }, { "epoch": 4.858085808580858, "grad_norm": 0.9720319509506226, "learning_rate": 1.0490196078431374e-05, "loss": 0.0356, "step": 920 }, { "epoch": 4.863366336633663, "grad_norm": 0.8841050267219543, "learning_rate": 1.0441176470588236e-05, "loss": 0.0301, "step": 921 }, { "epoch": 4.868646864686468, "grad_norm": 1.0269768238067627, "learning_rate": 1.0392156862745098e-05, "loss": 0.0315, "step": 922 }, { "epoch": 4.873927392739274, "grad_norm": 0.7187484502792358, "learning_rate": 1.034313725490196e-05, "loss": 0.0198, "step": 923 }, { "epoch": 4.879207920792079, "grad_norm": 0.8568077087402344, "learning_rate": 1.0294117647058824e-05, "loss": 0.0298, "step": 924 }, { "epoch": 4.884488448844884, "grad_norm": 1.0901175737380981, "learning_rate": 1.0245098039215687e-05, "loss": 0.0321, "step": 925 }, { "epoch": 4.88976897689769, "grad_norm": 0.7445207834243774, "learning_rate": 1.0196078431372549e-05, "loss": 0.0328, "step": 926 }, { "epoch": 4.895049504950495, "grad_norm": 1.0356990098953247, "learning_rate": 1.0147058823529413e-05, "loss": 0.0301, "step": 927 }, { "epoch": 4.9003300330033, "grad_norm": 1.0626378059387207, "learning_rate": 1.0098039215686275e-05, "loss": 0.0317, "step": 928 }, { "epoch": 4.905610561056106, "grad_norm": 0.9375091195106506, "learning_rate": 1.0049019607843139e-05, "loss": 0.0306, "step": 929 }, { "epoch": 4.910891089108911, "grad_norm": 0.7572767734527588, "learning_rate": 1e-05, "loss": 0.0233, "step": 930 }, { "epoch": 4.916171617161716, "grad_norm": 0.7957236170768738, "learning_rate": 9.950980392156863e-06, "loss": 0.0286, "step": 931 }, { "epoch": 4.921452145214522, "grad_norm": 0.723996639251709, "learning_rate": 9.901960784313725e-06, "loss": 0.0221, "step": 932 }, { "epoch": 4.926732673267327, "grad_norm": 0.7637007236480713, "learning_rate": 9.852941176470589e-06, "loss": 0.0222, "step": 933 }, { "epoch": 4.932013201320132, "grad_norm": 1.145386815071106, "learning_rate": 9.803921568627451e-06, "loss": 0.0401, "step": 934 }, { "epoch": 4.9372937293729375, "grad_norm": 1.1786422729492188, "learning_rate": 9.754901960784313e-06, "loss": 0.04, "step": 935 }, { "epoch": 4.9425742574257425, "grad_norm": 0.881608784198761, "learning_rate": 9.705882352941177e-06, "loss": 0.0292, "step": 936 }, { "epoch": 4.9478547854785475, "grad_norm": 1.056412935256958, "learning_rate": 9.65686274509804e-06, "loss": 0.0325, "step": 937 }, { "epoch": 4.953135313531353, "grad_norm": 1.0825217962265015, "learning_rate": 9.607843137254903e-06, "loss": 0.039, "step": 938 }, { "epoch": 4.958415841584158, "grad_norm": 0.6380865573883057, "learning_rate": 9.558823529411764e-06, "loss": 0.0231, "step": 939 }, { "epoch": 4.963696369636963, "grad_norm": 0.8444051742553711, "learning_rate": 9.509803921568628e-06, "loss": 0.0284, "step": 940 }, { "epoch": 4.968976897689769, "grad_norm": 0.8642618656158447, "learning_rate": 9.46078431372549e-06, "loss": 0.0265, "step": 941 }, { "epoch": 4.974257425742574, "grad_norm": 1.0572503805160522, "learning_rate": 9.411764705882354e-06, "loss": 0.0341, "step": 942 }, { "epoch": 4.979537953795379, "grad_norm": 0.8778902888298035, "learning_rate": 9.362745098039216e-06, "loss": 0.0304, "step": 943 }, { "epoch": 4.984818481848185, "grad_norm": 0.8647822737693787, "learning_rate": 9.31372549019608e-06, "loss": 0.0289, "step": 944 }, { "epoch": 4.99009900990099, "grad_norm": 1.0456162691116333, "learning_rate": 9.264705882352942e-06, "loss": 0.0322, "step": 945 }, { "epoch": 4.995379537953795, "grad_norm": 1.451906681060791, "learning_rate": 9.215686274509804e-06, "loss": 0.0414, "step": 946 }, { "epoch": 5.000660066006601, "grad_norm": 1.2711623907089233, "learning_rate": 9.166666666666666e-06, "loss": 0.0378, "step": 947 }, { "epoch": 5.005940594059406, "grad_norm": 0.5490008592605591, "learning_rate": 9.117647058823529e-06, "loss": 0.0236, "step": 948 }, { "epoch": 5.011221122112211, "grad_norm": 0.4675862789154053, "learning_rate": 9.068627450980392e-06, "loss": 0.0193, "step": 949 }, { "epoch": 5.016501650165017, "grad_norm": 0.5535048246383667, "learning_rate": 9.019607843137255e-06, "loss": 0.0198, "step": 950 }, { "epoch": 5.021782178217822, "grad_norm": 0.759076714515686, "learning_rate": 8.970588235294119e-06, "loss": 0.024, "step": 951 }, { "epoch": 5.027062706270627, "grad_norm": 0.551156222820282, "learning_rate": 8.92156862745098e-06, "loss": 0.0182, "step": 952 }, { "epoch": 5.032343234323433, "grad_norm": 0.561464786529541, "learning_rate": 8.872549019607845e-06, "loss": 0.0211, "step": 953 }, { "epoch": 5.037623762376238, "grad_norm": 0.3056110441684723, "learning_rate": 8.823529411764707e-06, "loss": 0.0146, "step": 954 }, { "epoch": 5.042904290429043, "grad_norm": 0.6087129712104797, "learning_rate": 8.774509803921569e-06, "loss": 0.0186, "step": 955 }, { "epoch": 5.0481848184818485, "grad_norm": 0.37718823552131653, "learning_rate": 8.725490196078431e-06, "loss": 0.0157, "step": 956 }, { "epoch": 5.0534653465346535, "grad_norm": 0.5187584161758423, "learning_rate": 8.676470588235295e-06, "loss": 0.0206, "step": 957 }, { "epoch": 5.0587458745874585, "grad_norm": 0.4911420941352844, "learning_rate": 8.627450980392157e-06, "loss": 0.0198, "step": 958 }, { "epoch": 5.064026402640264, "grad_norm": 0.5944388508796692, "learning_rate": 8.57843137254902e-06, "loss": 0.0199, "step": 959 }, { "epoch": 5.069306930693069, "grad_norm": 0.4171479344367981, "learning_rate": 8.529411764705883e-06, "loss": 0.018, "step": 960 }, { "epoch": 5.074587458745874, "grad_norm": 0.4937235414981842, "learning_rate": 8.480392156862745e-06, "loss": 0.0182, "step": 961 }, { "epoch": 5.07986798679868, "grad_norm": 0.44039833545684814, "learning_rate": 8.43137254901961e-06, "loss": 0.0166, "step": 962 }, { "epoch": 5.085148514851485, "grad_norm": 0.5266954302787781, "learning_rate": 8.38235294117647e-06, "loss": 0.0199, "step": 963 }, { "epoch": 5.09042904290429, "grad_norm": 0.5795379877090454, "learning_rate": 8.333333333333334e-06, "loss": 0.0168, "step": 964 }, { "epoch": 5.095709570957096, "grad_norm": 0.7336843013763428, "learning_rate": 8.284313725490196e-06, "loss": 0.0265, "step": 965 }, { "epoch": 5.100990099009901, "grad_norm": 0.4279429018497467, "learning_rate": 8.23529411764706e-06, "loss": 0.0177, "step": 966 }, { "epoch": 5.106270627062706, "grad_norm": 0.6605322957038879, "learning_rate": 8.186274509803922e-06, "loss": 0.0211, "step": 967 }, { "epoch": 5.111551155115512, "grad_norm": 0.6620057225227356, "learning_rate": 8.137254901960784e-06, "loss": 0.0186, "step": 968 }, { "epoch": 5.116831683168317, "grad_norm": 0.36124753952026367, "learning_rate": 8.088235294117648e-06, "loss": 0.015, "step": 969 }, { "epoch": 5.122112211221122, "grad_norm": 0.64441978931427, "learning_rate": 8.03921568627451e-06, "loss": 0.0176, "step": 970 }, { "epoch": 5.127392739273928, "grad_norm": 0.45318105816841125, "learning_rate": 7.990196078431372e-06, "loss": 0.0155, "step": 971 }, { "epoch": 5.132673267326733, "grad_norm": 0.528228223323822, "learning_rate": 7.941176470588235e-06, "loss": 0.0169, "step": 972 }, { "epoch": 5.137953795379538, "grad_norm": 0.5010206699371338, "learning_rate": 7.892156862745098e-06, "loss": 0.018, "step": 973 }, { "epoch": 5.143234323432344, "grad_norm": 0.4574146568775177, "learning_rate": 7.84313725490196e-06, "loss": 0.0172, "step": 974 }, { "epoch": 5.148514851485149, "grad_norm": 0.5487357974052429, "learning_rate": 7.794117647058825e-06, "loss": 0.0173, "step": 975 }, { "epoch": 5.153795379537954, "grad_norm": 0.7028687596321106, "learning_rate": 7.745098039215687e-06, "loss": 0.0219, "step": 976 }, { "epoch": 5.1590759075907595, "grad_norm": 0.5689717531204224, "learning_rate": 7.69607843137255e-06, "loss": 0.0167, "step": 977 }, { "epoch": 5.1643564356435645, "grad_norm": 0.4024108052253723, "learning_rate": 7.647058823529413e-06, "loss": 0.0139, "step": 978 }, { "epoch": 5.1696369636963695, "grad_norm": 0.81581050157547, "learning_rate": 7.598039215686274e-06, "loss": 0.0156, "step": 979 }, { "epoch": 5.174917491749175, "grad_norm": 0.4765579104423523, "learning_rate": 7.549019607843137e-06, "loss": 0.0155, "step": 980 }, { "epoch": 5.18019801980198, "grad_norm": 0.43915316462516785, "learning_rate": 7.5e-06, "loss": 0.0188, "step": 981 }, { "epoch": 5.185478547854785, "grad_norm": 0.5227847099304199, "learning_rate": 7.450980392156863e-06, "loss": 0.0206, "step": 982 }, { "epoch": 5.19075907590759, "grad_norm": 0.6932447552680969, "learning_rate": 7.401960784313726e-06, "loss": 0.0215, "step": 983 }, { "epoch": 5.196039603960396, "grad_norm": 0.6111219525337219, "learning_rate": 7.3529411764705884e-06, "loss": 0.0212, "step": 984 }, { "epoch": 5.201320132013201, "grad_norm": 0.4277690649032593, "learning_rate": 7.3039215686274515e-06, "loss": 0.0159, "step": 985 }, { "epoch": 5.206600660066006, "grad_norm": 0.47906622290611267, "learning_rate": 7.2549019607843145e-06, "loss": 0.0148, "step": 986 }, { "epoch": 5.211881188118812, "grad_norm": 0.565126359462738, "learning_rate": 7.205882352941176e-06, "loss": 0.0195, "step": 987 }, { "epoch": 5.217161716171617, "grad_norm": 0.46082907915115356, "learning_rate": 7.156862745098039e-06, "loss": 0.0183, "step": 988 }, { "epoch": 5.222442244224422, "grad_norm": 0.6407860517501831, "learning_rate": 7.107843137254902e-06, "loss": 0.0186, "step": 989 }, { "epoch": 5.227722772277228, "grad_norm": 0.4975033104419708, "learning_rate": 7.058823529411765e-06, "loss": 0.018, "step": 990 }, { "epoch": 5.233003300330033, "grad_norm": 0.8421338200569153, "learning_rate": 7.009803921568628e-06, "loss": 0.0236, "step": 991 }, { "epoch": 5.238283828382838, "grad_norm": 0.5029832124710083, "learning_rate": 6.960784313725491e-06, "loss": 0.017, "step": 992 }, { "epoch": 5.243564356435644, "grad_norm": 0.6490949988365173, "learning_rate": 6.911764705882354e-06, "loss": 0.0225, "step": 993 }, { "epoch": 5.248844884488449, "grad_norm": 0.4616677165031433, "learning_rate": 6.862745098039216e-06, "loss": 0.0141, "step": 994 }, { "epoch": 5.254125412541254, "grad_norm": 0.5079948306083679, "learning_rate": 6.813725490196078e-06, "loss": 0.0201, "step": 995 }, { "epoch": 5.25940594059406, "grad_norm": 0.6577730178833008, "learning_rate": 6.7647058823529414e-06, "loss": 0.0166, "step": 996 }, { "epoch": 5.264686468646865, "grad_norm": 0.4667012691497803, "learning_rate": 6.715686274509804e-06, "loss": 0.0142, "step": 997 }, { "epoch": 5.26996699669967, "grad_norm": 0.4957946836948395, "learning_rate": 6.666666666666667e-06, "loss": 0.0183, "step": 998 }, { "epoch": 5.2752475247524755, "grad_norm": 0.5439987182617188, "learning_rate": 6.61764705882353e-06, "loss": 0.0203, "step": 999 }, { "epoch": 5.2805280528052805, "grad_norm": 0.6871292591094971, "learning_rate": 6.568627450980393e-06, "loss": 0.0187, "step": 1000 }, { "epoch": 5.2858085808580855, "grad_norm": 1.2314088344573975, "learning_rate": 6.519607843137256e-06, "loss": 0.0218, "step": 1001 }, { "epoch": 5.291089108910891, "grad_norm": 0.9941577911376953, "learning_rate": 6.470588235294119e-06, "loss": 0.0231, "step": 1002 }, { "epoch": 5.296369636963696, "grad_norm": 0.5083587765693665, "learning_rate": 6.42156862745098e-06, "loss": 0.0175, "step": 1003 }, { "epoch": 5.301650165016501, "grad_norm": 0.548037052154541, "learning_rate": 6.372549019607843e-06, "loss": 0.019, "step": 1004 }, { "epoch": 5.306930693069307, "grad_norm": 0.5824829339981079, "learning_rate": 6.323529411764706e-06, "loss": 0.0229, "step": 1005 }, { "epoch": 5.312211221122112, "grad_norm": 0.40190955996513367, "learning_rate": 6.274509803921569e-06, "loss": 0.0177, "step": 1006 }, { "epoch": 5.317491749174917, "grad_norm": 0.5402151346206665, "learning_rate": 6.225490196078431e-06, "loss": 0.0204, "step": 1007 }, { "epoch": 5.322772277227723, "grad_norm": 0.5250375866889954, "learning_rate": 6.1764705882352944e-06, "loss": 0.0179, "step": 1008 }, { "epoch": 5.328052805280528, "grad_norm": 0.823615550994873, "learning_rate": 6.127450980392157e-06, "loss": 0.018, "step": 1009 }, { "epoch": 5.333333333333333, "grad_norm": 0.5749222636222839, "learning_rate": 6.07843137254902e-06, "loss": 0.0179, "step": 1010 }, { "epoch": 5.338613861386139, "grad_norm": 0.44801655411720276, "learning_rate": 6.029411764705883e-06, "loss": 0.0196, "step": 1011 }, { "epoch": 5.343894389438944, "grad_norm": 0.41333383321762085, "learning_rate": 5.980392156862746e-06, "loss": 0.0174, "step": 1012 }, { "epoch": 5.349174917491749, "grad_norm": 0.6496670842170715, "learning_rate": 5.931372549019608e-06, "loss": 0.0185, "step": 1013 }, { "epoch": 5.354455445544555, "grad_norm": 0.5372949838638306, "learning_rate": 5.882352941176471e-06, "loss": 0.0196, "step": 1014 }, { "epoch": 5.35973597359736, "grad_norm": 0.6921015381813049, "learning_rate": 5.833333333333334e-06, "loss": 0.0208, "step": 1015 }, { "epoch": 5.365016501650165, "grad_norm": 0.40831905603408813, "learning_rate": 5.784313725490197e-06, "loss": 0.0156, "step": 1016 }, { "epoch": 5.370297029702971, "grad_norm": 0.746163010597229, "learning_rate": 5.735294117647059e-06, "loss": 0.0247, "step": 1017 }, { "epoch": 5.375577557755776, "grad_norm": 0.3688511252403259, "learning_rate": 5.686274509803921e-06, "loss": 0.0149, "step": 1018 }, { "epoch": 5.380858085808581, "grad_norm": 0.6566423773765564, "learning_rate": 5.637254901960784e-06, "loss": 0.0207, "step": 1019 }, { "epoch": 5.3861386138613865, "grad_norm": 0.5603022575378418, "learning_rate": 5.588235294117647e-06, "loss": 0.015, "step": 1020 }, { "epoch": 5.3914191419141915, "grad_norm": 0.48195961117744446, "learning_rate": 5.53921568627451e-06, "loss": 0.0196, "step": 1021 }, { "epoch": 5.3966996699669965, "grad_norm": 0.8574143052101135, "learning_rate": 5.490196078431373e-06, "loss": 0.0213, "step": 1022 }, { "epoch": 5.401980198019802, "grad_norm": 0.6029947996139526, "learning_rate": 5.441176470588236e-06, "loss": 0.0157, "step": 1023 }, { "epoch": 5.407260726072607, "grad_norm": 0.6817245483398438, "learning_rate": 5.392156862745099e-06, "loss": 0.021, "step": 1024 }, { "epoch": 5.412541254125412, "grad_norm": 0.5209661722183228, "learning_rate": 5.343137254901961e-06, "loss": 0.0174, "step": 1025 }, { "epoch": 5.417821782178218, "grad_norm": 0.637236475944519, "learning_rate": 5.294117647058824e-06, "loss": 0.0227, "step": 1026 }, { "epoch": 5.423102310231023, "grad_norm": 0.4161434471607208, "learning_rate": 5.245098039215687e-06, "loss": 0.0186, "step": 1027 }, { "epoch": 5.428382838283828, "grad_norm": 0.4232980012893677, "learning_rate": 5.196078431372549e-06, "loss": 0.0157, "step": 1028 }, { "epoch": 5.433663366336634, "grad_norm": 0.8480959534645081, "learning_rate": 5.147058823529412e-06, "loss": 0.0236, "step": 1029 }, { "epoch": 5.438943894389439, "grad_norm": 0.7071532011032104, "learning_rate": 5.098039215686274e-06, "loss": 0.0232, "step": 1030 }, { "epoch": 5.444224422442244, "grad_norm": 0.46706661581993103, "learning_rate": 5.049019607843137e-06, "loss": 0.0154, "step": 1031 }, { "epoch": 5.44950495049505, "grad_norm": 0.38799214363098145, "learning_rate": 5e-06, "loss": 0.0153, "step": 1032 }, { "epoch": 5.454785478547855, "grad_norm": 0.6109775900840759, "learning_rate": 4.950980392156863e-06, "loss": 0.0185, "step": 1033 }, { "epoch": 5.46006600660066, "grad_norm": 0.4341484308242798, "learning_rate": 4.901960784313726e-06, "loss": 0.0167, "step": 1034 }, { "epoch": 5.465346534653466, "grad_norm": 0.7449372410774231, "learning_rate": 4.852941176470589e-06, "loss": 0.0198, "step": 1035 }, { "epoch": 5.470627062706271, "grad_norm": 0.36592140793800354, "learning_rate": 4.803921568627452e-06, "loss": 0.0157, "step": 1036 }, { "epoch": 5.475907590759076, "grad_norm": 0.5300789475440979, "learning_rate": 4.754901960784314e-06, "loss": 0.018, "step": 1037 }, { "epoch": 5.481188118811881, "grad_norm": 0.5365732908248901, "learning_rate": 4.705882352941177e-06, "loss": 0.0192, "step": 1038 }, { "epoch": 5.486468646864687, "grad_norm": 0.5415321588516235, "learning_rate": 4.65686274509804e-06, "loss": 0.0176, "step": 1039 }, { "epoch": 5.491749174917492, "grad_norm": 0.4487341344356537, "learning_rate": 4.607843137254902e-06, "loss": 0.0149, "step": 1040 }, { "epoch": 5.497029702970297, "grad_norm": 0.5181805491447449, "learning_rate": 4.558823529411764e-06, "loss": 0.0167, "step": 1041 }, { "epoch": 5.5023102310231025, "grad_norm": 0.7573541402816772, "learning_rate": 4.509803921568627e-06, "loss": 0.0192, "step": 1042 }, { "epoch": 5.5075907590759074, "grad_norm": 2.3085265159606934, "learning_rate": 4.46078431372549e-06, "loss": 0.0912, "step": 1043 }, { "epoch": 5.512871287128712, "grad_norm": 0.5817523002624512, "learning_rate": 4.411764705882353e-06, "loss": 0.015, "step": 1044 }, { "epoch": 5.518151815181518, "grad_norm": 0.6805261969566345, "learning_rate": 4.362745098039216e-06, "loss": 0.0215, "step": 1045 }, { "epoch": 5.523432343234323, "grad_norm": 0.597396969795227, "learning_rate": 4.313725490196079e-06, "loss": 0.0206, "step": 1046 }, { "epoch": 5.528712871287128, "grad_norm": 0.934479296207428, "learning_rate": 4.264705882352942e-06, "loss": 0.0207, "step": 1047 }, { "epoch": 5.533993399339934, "grad_norm": 0.4807792603969574, "learning_rate": 4.215686274509805e-06, "loss": 0.0158, "step": 1048 }, { "epoch": 5.539273927392739, "grad_norm": 0.6328734159469604, "learning_rate": 4.166666666666667e-06, "loss": 0.0202, "step": 1049 }, { "epoch": 5.544554455445544, "grad_norm": 0.4848245084285736, "learning_rate": 4.11764705882353e-06, "loss": 0.0199, "step": 1050 }, { "epoch": 5.54983498349835, "grad_norm": 0.4785003066062927, "learning_rate": 4.068627450980392e-06, "loss": 0.0188, "step": 1051 }, { "epoch": 5.555115511551155, "grad_norm": 0.4844956696033478, "learning_rate": 4.019607843137255e-06, "loss": 0.0175, "step": 1052 }, { "epoch": 5.56039603960396, "grad_norm": 0.40522444248199463, "learning_rate": 3.970588235294117e-06, "loss": 0.0159, "step": 1053 }, { "epoch": 5.565676567656766, "grad_norm": 0.40739139914512634, "learning_rate": 3.92156862745098e-06, "loss": 0.0165, "step": 1054 }, { "epoch": 5.570957095709571, "grad_norm": 0.42678526043891907, "learning_rate": 3.872549019607843e-06, "loss": 0.0152, "step": 1055 }, { "epoch": 5.576237623762376, "grad_norm": 0.5168190598487854, "learning_rate": 3.823529411764706e-06, "loss": 0.0164, "step": 1056 }, { "epoch": 5.581518151815182, "grad_norm": 0.4231308400630951, "learning_rate": 3.7745098039215686e-06, "loss": 0.016, "step": 1057 }, { "epoch": 5.586798679867987, "grad_norm": 0.604710578918457, "learning_rate": 3.7254901960784316e-06, "loss": 0.0225, "step": 1058 }, { "epoch": 5.592079207920792, "grad_norm": 0.6866090893745422, "learning_rate": 3.6764705882352942e-06, "loss": 0.0195, "step": 1059 }, { "epoch": 5.597359735973598, "grad_norm": 0.34970760345458984, "learning_rate": 3.6274509803921573e-06, "loss": 0.014, "step": 1060 }, { "epoch": 5.602640264026403, "grad_norm": 0.4601968228816986, "learning_rate": 3.5784313725490194e-06, "loss": 0.018, "step": 1061 }, { "epoch": 5.607920792079208, "grad_norm": 0.3815479874610901, "learning_rate": 3.5294117647058825e-06, "loss": 0.0152, "step": 1062 }, { "epoch": 5.6132013201320134, "grad_norm": 0.36140522360801697, "learning_rate": 3.4803921568627455e-06, "loss": 0.0149, "step": 1063 }, { "epoch": 5.618481848184818, "grad_norm": 0.485866904258728, "learning_rate": 3.431372549019608e-06, "loss": 0.0171, "step": 1064 }, { "epoch": 5.623762376237623, "grad_norm": 0.5746606588363647, "learning_rate": 3.3823529411764707e-06, "loss": 0.0168, "step": 1065 }, { "epoch": 5.629042904290429, "grad_norm": 0.5221585035324097, "learning_rate": 3.3333333333333333e-06, "loss": 0.0199, "step": 1066 }, { "epoch": 5.634323432343234, "grad_norm": 0.538971483707428, "learning_rate": 3.2843137254901964e-06, "loss": 0.0225, "step": 1067 }, { "epoch": 5.639603960396039, "grad_norm": 0.4141756594181061, "learning_rate": 3.2352941176470594e-06, "loss": 0.0171, "step": 1068 }, { "epoch": 5.644884488448845, "grad_norm": 0.3699265420436859, "learning_rate": 3.1862745098039216e-06, "loss": 0.0127, "step": 1069 }, { "epoch": 5.65016501650165, "grad_norm": 0.8097654581069946, "learning_rate": 3.1372549019607846e-06, "loss": 0.0168, "step": 1070 }, { "epoch": 5.655445544554455, "grad_norm": 0.8609626889228821, "learning_rate": 3.0882352941176472e-06, "loss": 0.0202, "step": 1071 }, { "epoch": 5.660726072607261, "grad_norm": 0.7239216566085815, "learning_rate": 3.03921568627451e-06, "loss": 0.0173, "step": 1072 }, { "epoch": 5.666006600660066, "grad_norm": 0.5894525647163391, "learning_rate": 2.990196078431373e-06, "loss": 0.0248, "step": 1073 }, { "epoch": 5.671287128712871, "grad_norm": 0.39227187633514404, "learning_rate": 2.9411764705882355e-06, "loss": 0.0173, "step": 1074 }, { "epoch": 5.676567656765677, "grad_norm": 0.5740650296211243, "learning_rate": 2.8921568627450985e-06, "loss": 0.0196, "step": 1075 }, { "epoch": 5.681848184818482, "grad_norm": 0.5081339478492737, "learning_rate": 2.8431372549019607e-06, "loss": 0.0178, "step": 1076 }, { "epoch": 5.687128712871287, "grad_norm": 0.45653530955314636, "learning_rate": 2.7941176470588237e-06, "loss": 0.017, "step": 1077 }, { "epoch": 5.692409240924093, "grad_norm": 0.5196064710617065, "learning_rate": 2.7450980392156863e-06, "loss": 0.0158, "step": 1078 }, { "epoch": 5.697689768976898, "grad_norm": 0.5605891942977905, "learning_rate": 2.6960784313725493e-06, "loss": 0.0172, "step": 1079 }, { "epoch": 5.702970297029703, "grad_norm": 0.6851004958152771, "learning_rate": 2.647058823529412e-06, "loss": 0.0172, "step": 1080 }, { "epoch": 5.708250825082509, "grad_norm": 0.4508073031902313, "learning_rate": 2.5980392156862746e-06, "loss": 0.0171, "step": 1081 }, { "epoch": 5.713531353135314, "grad_norm": 0.4393492639064789, "learning_rate": 2.549019607843137e-06, "loss": 0.0167, "step": 1082 }, { "epoch": 5.718811881188119, "grad_norm": 0.5042679309844971, "learning_rate": 2.5e-06, "loss": 0.0197, "step": 1083 }, { "epoch": 5.724092409240924, "grad_norm": 0.40561428666114807, "learning_rate": 2.450980392156863e-06, "loss": 0.0175, "step": 1084 }, { "epoch": 5.729372937293729, "grad_norm": 0.390462189912796, "learning_rate": 2.401960784313726e-06, "loss": 0.0166, "step": 1085 }, { "epoch": 5.734653465346534, "grad_norm": 0.44142773747444153, "learning_rate": 2.3529411764705885e-06, "loss": 0.0161, "step": 1086 }, { "epoch": 5.73993399339934, "grad_norm": 0.6280815601348877, "learning_rate": 2.303921568627451e-06, "loss": 0.0191, "step": 1087 }, { "epoch": 5.745214521452145, "grad_norm": 0.6288333535194397, "learning_rate": 2.2549019607843137e-06, "loss": 0.0169, "step": 1088 }, { "epoch": 5.75049504950495, "grad_norm": 0.39587247371673584, "learning_rate": 2.2058823529411767e-06, "loss": 0.0154, "step": 1089 }, { "epoch": 5.755775577557756, "grad_norm": 0.8085327744483948, "learning_rate": 2.1568627450980393e-06, "loss": 0.0214, "step": 1090 }, { "epoch": 5.761056105610561, "grad_norm": 0.31861352920532227, "learning_rate": 2.1078431372549023e-06, "loss": 0.0113, "step": 1091 }, { "epoch": 5.766336633663366, "grad_norm": 0.6090717315673828, "learning_rate": 2.058823529411765e-06, "loss": 0.0176, "step": 1092 }, { "epoch": 5.771617161716172, "grad_norm": 0.37864407896995544, "learning_rate": 2.0098039215686276e-06, "loss": 0.013, "step": 1093 }, { "epoch": 5.776897689768977, "grad_norm": 0.39943984150886536, "learning_rate": 1.96078431372549e-06, "loss": 0.0183, "step": 1094 }, { "epoch": 5.782178217821782, "grad_norm": 0.3896962106227875, "learning_rate": 1.911764705882353e-06, "loss": 0.0147, "step": 1095 }, { "epoch": 5.787458745874588, "grad_norm": 0.8300476670265198, "learning_rate": 1.8627450980392158e-06, "loss": 0.0219, "step": 1096 }, { "epoch": 5.792739273927393, "grad_norm": 0.7217696905136108, "learning_rate": 1.8137254901960786e-06, "loss": 0.0193, "step": 1097 }, { "epoch": 5.798019801980198, "grad_norm": 0.7046459317207336, "learning_rate": 1.7647058823529412e-06, "loss": 0.019, "step": 1098 }, { "epoch": 5.803300330033004, "grad_norm": 0.5788043141365051, "learning_rate": 1.715686274509804e-06, "loss": 0.0197, "step": 1099 }, { "epoch": 5.808580858085809, "grad_norm": 0.3641752004623413, "learning_rate": 1.6666666666666667e-06, "loss": 0.0139, "step": 1100 }, { "epoch": 5.813861386138614, "grad_norm": 0.5467818379402161, "learning_rate": 1.6176470588235297e-06, "loss": 0.0185, "step": 1101 }, { "epoch": 5.81914191419142, "grad_norm": 0.7358901500701904, "learning_rate": 1.5686274509803923e-06, "loss": 0.0218, "step": 1102 }, { "epoch": 5.824422442244225, "grad_norm": 0.39868056774139404, "learning_rate": 1.519607843137255e-06, "loss": 0.0168, "step": 1103 }, { "epoch": 5.8297029702970296, "grad_norm": 0.5790492296218872, "learning_rate": 1.4705882352941177e-06, "loss": 0.0154, "step": 1104 }, { "epoch": 5.834983498349835, "grad_norm": 0.6561235189437866, "learning_rate": 1.4215686274509803e-06, "loss": 0.0184, "step": 1105 }, { "epoch": 5.84026402640264, "grad_norm": 0.4628670811653137, "learning_rate": 1.3725490196078432e-06, "loss": 0.0178, "step": 1106 }, { "epoch": 5.845544554455445, "grad_norm": 0.5217213034629822, "learning_rate": 1.323529411764706e-06, "loss": 0.0179, "step": 1107 }, { "epoch": 5.850825082508251, "grad_norm": 0.6204583644866943, "learning_rate": 1.2745098039215686e-06, "loss": 0.0197, "step": 1108 }, { "epoch": 5.856105610561056, "grad_norm": 0.48575830459594727, "learning_rate": 1.2254901960784314e-06, "loss": 0.0191, "step": 1109 }, { "epoch": 5.861386138613861, "grad_norm": 0.8481599688529968, "learning_rate": 1.1764705882352942e-06, "loss": 0.023, "step": 1110 }, { "epoch": 5.866666666666667, "grad_norm": 0.5180693864822388, "learning_rate": 1.1274509803921568e-06, "loss": 0.016, "step": 1111 }, { "epoch": 5.871947194719472, "grad_norm": 0.47926902770996094, "learning_rate": 1.0784313725490197e-06, "loss": 0.018, "step": 1112 }, { "epoch": 5.877227722772277, "grad_norm": 0.4460168182849884, "learning_rate": 1.0294117647058825e-06, "loss": 0.0155, "step": 1113 }, { "epoch": 5.882508250825083, "grad_norm": 0.6543490886688232, "learning_rate": 9.80392156862745e-07, "loss": 0.0193, "step": 1114 }, { "epoch": 5.887788778877888, "grad_norm": 0.5319867134094238, "learning_rate": 9.313725490196079e-07, "loss": 0.0164, "step": 1115 }, { "epoch": 5.893069306930693, "grad_norm": 0.5237565636634827, "learning_rate": 8.823529411764706e-07, "loss": 0.0143, "step": 1116 }, { "epoch": 5.898349834983498, "grad_norm": 0.43923959136009216, "learning_rate": 8.333333333333333e-07, "loss": 0.0172, "step": 1117 }, { "epoch": 5.903630363036304, "grad_norm": 0.5296475291252136, "learning_rate": 7.843137254901962e-07, "loss": 0.0182, "step": 1118 }, { "epoch": 5.908910891089109, "grad_norm": 0.38380494713783264, "learning_rate": 7.352941176470589e-07, "loss": 0.0159, "step": 1119 }, { "epoch": 5.914191419141914, "grad_norm": 0.6031242609024048, "learning_rate": 6.862745098039216e-07, "loss": 0.0152, "step": 1120 }, { "epoch": 5.91947194719472, "grad_norm": 0.5882181525230408, "learning_rate": 6.372549019607843e-07, "loss": 0.0171, "step": 1121 }, { "epoch": 5.924752475247525, "grad_norm": 0.41396641731262207, "learning_rate": 5.882352941176471e-07, "loss": 0.0162, "step": 1122 }, { "epoch": 5.93003300330033, "grad_norm": 0.595112144947052, "learning_rate": 5.392156862745098e-07, "loss": 0.0174, "step": 1123 }, { "epoch": 5.9353135313531356, "grad_norm": 0.6171261072158813, "learning_rate": 4.901960784313725e-07, "loss": 0.022, "step": 1124 }, { "epoch": 5.9405940594059405, "grad_norm": 0.40438172221183777, "learning_rate": 4.411764705882353e-07, "loss": 0.0144, "step": 1125 }, { "epoch": 5.9458745874587455, "grad_norm": 0.6478520035743713, "learning_rate": 3.921568627450981e-07, "loss": 0.0169, "step": 1126 }, { "epoch": 5.951155115511551, "grad_norm": 0.3631436824798584, "learning_rate": 3.431372549019608e-07, "loss": 0.0149, "step": 1127 }, { "epoch": 5.956435643564356, "grad_norm": 0.7141023278236389, "learning_rate": 2.9411764705882356e-07, "loss": 0.0169, "step": 1128 }, { "epoch": 5.961716171617161, "grad_norm": 0.6038995981216431, "learning_rate": 2.4509803921568627e-07, "loss": 0.0197, "step": 1129 }, { "epoch": 5.966996699669967, "grad_norm": 0.5263422727584839, "learning_rate": 1.9607843137254904e-07, "loss": 0.0201, "step": 1130 }, { "epoch": 5.972277227722772, "grad_norm": 0.604172945022583, "learning_rate": 1.4705882352941178e-07, "loss": 0.0181, "step": 1131 }, { "epoch": 5.977557755775577, "grad_norm": 0.6950435638427734, "learning_rate": 9.803921568627452e-08, "loss": 0.0211, "step": 1132 }, { "epoch": 5.982838283828383, "grad_norm": 0.5159468650817871, "learning_rate": 4.901960784313726e-08, "loss": 0.0153, "step": 1133 }, { "epoch": 5.988118811881188, "grad_norm": 0.7329381108283997, "learning_rate": 0.0, "loss": 0.0224, "step": 1134 } ], "logging_steps": 1, "max_steps": 1134, "num_input_tokens_seen": 0, "num_train_epochs": 6, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 6.176433819923251e+17, "train_batch_size": 2, "trial_name": null, "trial_params": null }