{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 0.999129677980853, "eval_steps": 72, "global_step": 574, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0, "eval_loss": 1.136976718902588, "eval_runtime": 64.1168, "eval_samples_per_second": 73.117, "eval_steps_per_second": 9.14, "step": 0 }, { "epoch": 0.0017406440382941688, "grad_norm": 1.859375, "learning_rate": 0.0, "loss": 1.1426, "step": 1 }, { "epoch": 0.0034812880765883376, "grad_norm": 1.9765625, "learning_rate": 1.7543859649122808e-07, "loss": 1.1274, "step": 2 }, { "epoch": 0.005221932114882507, "grad_norm": 1.828125, "learning_rate": 3.5087719298245616e-07, "loss": 1.1205, "step": 3 }, { "epoch": 0.006962576153176675, "grad_norm": 1.8515625, "learning_rate": 5.263157894736843e-07, "loss": 1.1383, "step": 4 }, { "epoch": 0.008703220191470844, "grad_norm": 1.796875, "learning_rate": 7.017543859649123e-07, "loss": 1.14, "step": 5 }, { "epoch": 0.010443864229765013, "grad_norm": 1.734375, "learning_rate": 8.771929824561404e-07, "loss": 1.1031, "step": 6 }, { "epoch": 0.012184508268059183, "grad_norm": 2.015625, "learning_rate": 1.0526315789473685e-06, "loss": 1.1794, "step": 7 }, { "epoch": 0.01392515230635335, "grad_norm": 1.765625, "learning_rate": 1.2280701754385965e-06, "loss": 1.1766, "step": 8 }, { "epoch": 0.015665796344647518, "grad_norm": 1.8046875, "learning_rate": 1.4035087719298246e-06, "loss": 1.1251, "step": 9 }, { "epoch": 0.017406440382941687, "grad_norm": 1.6796875, "learning_rate": 1.5789473684210526e-06, "loss": 1.1622, "step": 10 }, { "epoch": 0.019147084421235857, "grad_norm": 1.8671875, "learning_rate": 1.7543859649122807e-06, "loss": 1.0846, "step": 11 }, { "epoch": 0.020887728459530026, "grad_norm": 1.765625, "learning_rate": 1.929824561403509e-06, "loss": 1.1644, "step": 12 }, { "epoch": 0.022628372497824196, "grad_norm": 1.7265625, "learning_rate": 2.105263157894737e-06, "loss": 1.1361, "step": 13 }, { "epoch": 0.024369016536118365, "grad_norm": 1.8671875, "learning_rate": 2.280701754385965e-06, "loss": 1.1783, "step": 14 }, { "epoch": 0.02610966057441253, "grad_norm": 1.828125, "learning_rate": 2.456140350877193e-06, "loss": 1.127, "step": 15 }, { "epoch": 0.0278503046127067, "grad_norm": 1.640625, "learning_rate": 2.631578947368421e-06, "loss": 1.1587, "step": 16 }, { "epoch": 0.02959094865100087, "grad_norm": 1.671875, "learning_rate": 2.8070175438596493e-06, "loss": 1.1088, "step": 17 }, { "epoch": 0.031331592689295036, "grad_norm": 1.59375, "learning_rate": 2.9824561403508774e-06, "loss": 1.174, "step": 18 }, { "epoch": 0.03307223672758921, "grad_norm": 1.71875, "learning_rate": 3.157894736842105e-06, "loss": 1.1453, "step": 19 }, { "epoch": 0.034812880765883375, "grad_norm": 1.8515625, "learning_rate": 3.3333333333333333e-06, "loss": 1.1922, "step": 20 }, { "epoch": 0.03655352480417755, "grad_norm": 1.5390625, "learning_rate": 3.5087719298245615e-06, "loss": 1.1541, "step": 21 }, { "epoch": 0.038294168842471714, "grad_norm": 1.5390625, "learning_rate": 3.6842105263157896e-06, "loss": 1.0909, "step": 22 }, { "epoch": 0.04003481288076589, "grad_norm": 1.515625, "learning_rate": 3.859649122807018e-06, "loss": 1.1498, "step": 23 }, { "epoch": 0.04177545691906005, "grad_norm": 1.5703125, "learning_rate": 4.035087719298246e-06, "loss": 1.1096, "step": 24 }, { "epoch": 0.04351610095735422, "grad_norm": 1.453125, "learning_rate": 4.210526315789474e-06, "loss": 1.1152, "step": 25 }, { "epoch": 0.04525674499564839, "grad_norm": 1.390625, "learning_rate": 4.385964912280702e-06, "loss": 1.1111, "step": 26 }, { "epoch": 0.04699738903394256, "grad_norm": 1.296875, "learning_rate": 4.56140350877193e-06, "loss": 1.161, "step": 27 }, { "epoch": 0.04873803307223673, "grad_norm": 1.390625, "learning_rate": 4.736842105263158e-06, "loss": 1.1557, "step": 28 }, { "epoch": 0.050478677110530897, "grad_norm": 1.3125, "learning_rate": 4.912280701754386e-06, "loss": 1.1065, "step": 29 }, { "epoch": 0.05221932114882506, "grad_norm": 1.28125, "learning_rate": 5.087719298245615e-06, "loss": 1.0493, "step": 30 }, { "epoch": 0.053959965187119235, "grad_norm": 1.1953125, "learning_rate": 5.263157894736842e-06, "loss": 1.0652, "step": 31 }, { "epoch": 0.0557006092254134, "grad_norm": 1.1484375, "learning_rate": 5.438596491228071e-06, "loss": 1.0389, "step": 32 }, { "epoch": 0.057441253263707574, "grad_norm": 1.1015625, "learning_rate": 5.6140350877192985e-06, "loss": 1.0349, "step": 33 }, { "epoch": 0.05918189730200174, "grad_norm": 1.0390625, "learning_rate": 5.789473684210527e-06, "loss": 1.1109, "step": 34 }, { "epoch": 0.060922541340295906, "grad_norm": 1.03125, "learning_rate": 5.964912280701755e-06, "loss": 1.0541, "step": 35 }, { "epoch": 0.06266318537859007, "grad_norm": 0.9453125, "learning_rate": 6.140350877192983e-06, "loss": 1.1176, "step": 36 }, { "epoch": 0.06440382941688425, "grad_norm": 0.9140625, "learning_rate": 6.31578947368421e-06, "loss": 1.0774, "step": 37 }, { "epoch": 0.06614447345517842, "grad_norm": 0.9453125, "learning_rate": 6.491228070175439e-06, "loss": 1.031, "step": 38 }, { "epoch": 0.06788511749347259, "grad_norm": 0.85546875, "learning_rate": 6.666666666666667e-06, "loss": 1.0977, "step": 39 }, { "epoch": 0.06962576153176675, "grad_norm": 0.8203125, "learning_rate": 6.842105263157896e-06, "loss": 1.0871, "step": 40 }, { "epoch": 0.07136640557006092, "grad_norm": 0.83984375, "learning_rate": 7.017543859649123e-06, "loss": 1.0856, "step": 41 }, { "epoch": 0.0731070496083551, "grad_norm": 0.82421875, "learning_rate": 7.192982456140352e-06, "loss": 1.0189, "step": 42 }, { "epoch": 0.07484769364664925, "grad_norm": 0.76953125, "learning_rate": 7.368421052631579e-06, "loss": 1.0633, "step": 43 }, { "epoch": 0.07658833768494343, "grad_norm": 0.7578125, "learning_rate": 7.5438596491228074e-06, "loss": 1.0582, "step": 44 }, { "epoch": 0.0783289817232376, "grad_norm": 0.78125, "learning_rate": 7.719298245614036e-06, "loss": 1.0261, "step": 45 }, { "epoch": 0.08006962576153177, "grad_norm": 0.66796875, "learning_rate": 7.894736842105265e-06, "loss": 0.9962, "step": 46 }, { "epoch": 0.08181026979982593, "grad_norm": 0.6953125, "learning_rate": 8.070175438596492e-06, "loss": 1.0041, "step": 47 }, { "epoch": 0.0835509138381201, "grad_norm": 0.6640625, "learning_rate": 8.24561403508772e-06, "loss": 1.0953, "step": 48 }, { "epoch": 0.08529155787641428, "grad_norm": 0.671875, "learning_rate": 8.421052631578948e-06, "loss": 1.0397, "step": 49 }, { "epoch": 0.08703220191470844, "grad_norm": 0.66015625, "learning_rate": 8.596491228070176e-06, "loss": 1.0366, "step": 50 }, { "epoch": 0.08877284595300261, "grad_norm": 0.65625, "learning_rate": 8.771929824561405e-06, "loss": 1.005, "step": 51 }, { "epoch": 0.09051348999129678, "grad_norm": 0.62109375, "learning_rate": 8.947368421052632e-06, "loss": 1.1106, "step": 52 }, { "epoch": 0.09225413402959094, "grad_norm": 0.59375, "learning_rate": 9.12280701754386e-06, "loss": 1.002, "step": 53 }, { "epoch": 0.09399477806788512, "grad_norm": 0.5859375, "learning_rate": 9.298245614035088e-06, "loss": 1.0599, "step": 54 }, { "epoch": 0.09573542210617929, "grad_norm": 0.59765625, "learning_rate": 9.473684210526315e-06, "loss": 1.0736, "step": 55 }, { "epoch": 0.09747606614447346, "grad_norm": 0.56640625, "learning_rate": 9.649122807017545e-06, "loss": 1.0179, "step": 56 }, { "epoch": 0.09921671018276762, "grad_norm": 0.58203125, "learning_rate": 9.824561403508772e-06, "loss": 1.0308, "step": 57 }, { "epoch": 0.10095735422106179, "grad_norm": 0.5625, "learning_rate": 1e-05, "loss": 1.0135, "step": 58 }, { "epoch": 0.10269799825935597, "grad_norm": 0.5859375, "learning_rate": 9.999979270446263e-06, "loss": 0.9598, "step": 59 }, { "epoch": 0.10443864229765012, "grad_norm": 0.5390625, "learning_rate": 9.999917081956933e-06, "loss": 1.0041, "step": 60 }, { "epoch": 0.1061792863359443, "grad_norm": 0.53125, "learning_rate": 9.999813435047668e-06, "loss": 1.0726, "step": 61 }, { "epoch": 0.10791993037423847, "grad_norm": 0.55078125, "learning_rate": 9.99966833057789e-06, "loss": 0.9915, "step": 62 }, { "epoch": 0.10966057441253264, "grad_norm": 0.55078125, "learning_rate": 9.999481769750779e-06, "loss": 1.0266, "step": 63 }, { "epoch": 0.1114012184508268, "grad_norm": 0.5234375, "learning_rate": 9.999253754113263e-06, "loss": 1.0524, "step": 64 }, { "epoch": 0.11314186248912098, "grad_norm": 0.53125, "learning_rate": 9.998984285556008e-06, "loss": 1.0254, "step": 65 }, { "epoch": 0.11488250652741515, "grad_norm": 0.53515625, "learning_rate": 9.998673366313399e-06, "loss": 1.0132, "step": 66 }, { "epoch": 0.11662315056570931, "grad_norm": 0.5234375, "learning_rate": 9.998320998963523e-06, "loss": 0.9604, "step": 67 }, { "epoch": 0.11836379460400348, "grad_norm": 0.54296875, "learning_rate": 9.997927186428145e-06, "loss": 1.0147, "step": 68 }, { "epoch": 0.12010443864229765, "grad_norm": 0.52734375, "learning_rate": 9.997491931972694e-06, "loss": 0.998, "step": 69 }, { "epoch": 0.12184508268059181, "grad_norm": 0.51953125, "learning_rate": 9.997015239206216e-06, "loss": 1.034, "step": 70 }, { "epoch": 0.12358572671888599, "grad_norm": 0.53125, "learning_rate": 9.996497112081365e-06, "loss": 1.0103, "step": 71 }, { "epoch": 0.12532637075718014, "grad_norm": 0.5078125, "learning_rate": 9.99593755489436e-06, "loss": 1.0053, "step": 72 }, { "epoch": 0.12532637075718014, "eval_loss": 0.9893413186073303, "eval_runtime": 60.3364, "eval_samples_per_second": 77.698, "eval_steps_per_second": 9.712, "step": 72 }, { "epoch": 0.12706701479547433, "grad_norm": 0.50390625, "learning_rate": 9.995336572284945e-06, "loss": 0.9841, "step": 73 }, { "epoch": 0.1288076588337685, "grad_norm": 0.53125, "learning_rate": 9.994694169236366e-06, "loss": 0.9298, "step": 74 }, { "epoch": 0.13054830287206268, "grad_norm": 0.51171875, "learning_rate": 9.99401035107531e-06, "loss": 1.0449, "step": 75 }, { "epoch": 0.13228894691035684, "grad_norm": 0.515625, "learning_rate": 9.993285123471878e-06, "loss": 1.0139, "step": 76 }, { "epoch": 0.134029590948651, "grad_norm": 0.5078125, "learning_rate": 9.992518492439526e-06, "loss": 1.0129, "step": 77 }, { "epoch": 0.13577023498694518, "grad_norm": 0.54296875, "learning_rate": 9.991710464335022e-06, "loss": 0.9696, "step": 78 }, { "epoch": 0.13751087902523934, "grad_norm": 0.53125, "learning_rate": 9.990861045858392e-06, "loss": 1.0114, "step": 79 }, { "epoch": 0.1392515230635335, "grad_norm": 0.515625, "learning_rate": 9.989970244052861e-06, "loss": 0.957, "step": 80 }, { "epoch": 0.1409921671018277, "grad_norm": 0.5078125, "learning_rate": 9.9890380663048e-06, "loss": 0.9603, "step": 81 }, { "epoch": 0.14273281114012185, "grad_norm": 0.52734375, "learning_rate": 9.98806452034366e-06, "loss": 1.0523, "step": 82 }, { "epoch": 0.144473455178416, "grad_norm": 0.5078125, "learning_rate": 9.987049614241907e-06, "loss": 0.9804, "step": 83 }, { "epoch": 0.1462140992167102, "grad_norm": 0.54296875, "learning_rate": 9.985993356414965e-06, "loss": 1.0307, "step": 84 }, { "epoch": 0.14795474325500435, "grad_norm": 0.515625, "learning_rate": 9.984895755621136e-06, "loss": 0.9802, "step": 85 }, { "epoch": 0.1496953872932985, "grad_norm": 0.494140625, "learning_rate": 9.983756820961528e-06, "loss": 0.9547, "step": 86 }, { "epoch": 0.1514360313315927, "grad_norm": 0.5, "learning_rate": 9.982576561879984e-06, "loss": 1.014, "step": 87 }, { "epoch": 0.15317667536988686, "grad_norm": 0.494140625, "learning_rate": 9.981354988163002e-06, "loss": 0.9708, "step": 88 }, { "epoch": 0.15491731940818101, "grad_norm": 0.51171875, "learning_rate": 9.980092109939651e-06, "loss": 0.9717, "step": 89 }, { "epoch": 0.1566579634464752, "grad_norm": 0.484375, "learning_rate": 9.978787937681496e-06, "loss": 0.9862, "step": 90 }, { "epoch": 0.15839860748476936, "grad_norm": 0.498046875, "learning_rate": 9.977442482202498e-06, "loss": 0.9813, "step": 91 }, { "epoch": 0.16013925152306355, "grad_norm": 0.494140625, "learning_rate": 9.976055754658935e-06, "loss": 0.9551, "step": 92 }, { "epoch": 0.1618798955613577, "grad_norm": 0.51171875, "learning_rate": 9.974627766549301e-06, "loss": 0.9777, "step": 93 }, { "epoch": 0.16362053959965187, "grad_norm": 0.50390625, "learning_rate": 9.973158529714224e-06, "loss": 0.9648, "step": 94 }, { "epoch": 0.16536118363794605, "grad_norm": 0.48828125, "learning_rate": 9.971648056336349e-06, "loss": 0.964, "step": 95 }, { "epoch": 0.1671018276762402, "grad_norm": 0.5078125, "learning_rate": 9.97009635894025e-06, "loss": 0.9635, "step": 96 }, { "epoch": 0.16884247171453437, "grad_norm": 0.515625, "learning_rate": 9.968503450392332e-06, "loss": 1.0115, "step": 97 }, { "epoch": 0.17058311575282856, "grad_norm": 0.51171875, "learning_rate": 9.966869343900702e-06, "loss": 0.9841, "step": 98 }, { "epoch": 0.17232375979112272, "grad_norm": 0.484375, "learning_rate": 9.965194053015083e-06, "loss": 1.0241, "step": 99 }, { "epoch": 0.17406440382941687, "grad_norm": 0.50390625, "learning_rate": 9.963477591626687e-06, "loss": 0.968, "step": 100 }, { "epoch": 0.17580504786771106, "grad_norm": 0.50390625, "learning_rate": 9.961719973968102e-06, "loss": 0.946, "step": 101 }, { "epoch": 0.17754569190600522, "grad_norm": 0.5234375, "learning_rate": 9.959921214613187e-06, "loss": 0.9284, "step": 102 }, { "epoch": 0.17928633594429938, "grad_norm": 0.494140625, "learning_rate": 9.958081328476926e-06, "loss": 0.958, "step": 103 }, { "epoch": 0.18102697998259357, "grad_norm": 0.486328125, "learning_rate": 9.956200330815329e-06, "loss": 0.9361, "step": 104 }, { "epoch": 0.18276762402088773, "grad_norm": 0.5078125, "learning_rate": 9.954278237225296e-06, "loss": 1.0116, "step": 105 }, { "epoch": 0.18450826805918188, "grad_norm": 0.5078125, "learning_rate": 9.952315063644479e-06, "loss": 1.0041, "step": 106 }, { "epoch": 0.18624891209747607, "grad_norm": 0.5, "learning_rate": 9.950310826351168e-06, "loss": 0.9725, "step": 107 }, { "epoch": 0.18798955613577023, "grad_norm": 0.498046875, "learning_rate": 9.948265541964136e-06, "loss": 0.9763, "step": 108 }, { "epoch": 0.18973020017406442, "grad_norm": 0.5, "learning_rate": 9.946179227442521e-06, "loss": 0.9814, "step": 109 }, { "epoch": 0.19147084421235858, "grad_norm": 0.50390625, "learning_rate": 9.944051900085668e-06, "loss": 1.0092, "step": 110 }, { "epoch": 0.19321148825065274, "grad_norm": 0.515625, "learning_rate": 9.941883577532993e-06, "loss": 0.9836, "step": 111 }, { "epoch": 0.19495213228894692, "grad_norm": 0.49609375, "learning_rate": 9.939674277763845e-06, "loss": 0.9598, "step": 112 }, { "epoch": 0.19669277632724108, "grad_norm": 0.48828125, "learning_rate": 9.937424019097337e-06, "loss": 0.988, "step": 113 }, { "epoch": 0.19843342036553524, "grad_norm": 0.498046875, "learning_rate": 9.935132820192218e-06, "loss": 0.9512, "step": 114 }, { "epoch": 0.20017406440382943, "grad_norm": 0.498046875, "learning_rate": 9.932800700046697e-06, "loss": 0.9916, "step": 115 }, { "epoch": 0.20191470844212359, "grad_norm": 0.51171875, "learning_rate": 9.9304276779983e-06, "loss": 0.9397, "step": 116 }, { "epoch": 0.20365535248041775, "grad_norm": 0.48828125, "learning_rate": 9.9280137737237e-06, "loss": 0.9791, "step": 117 }, { "epoch": 0.20539599651871193, "grad_norm": 0.515625, "learning_rate": 9.925559007238564e-06, "loss": 0.9231, "step": 118 }, { "epoch": 0.2071366405570061, "grad_norm": 0.5, "learning_rate": 9.923063398897372e-06, "loss": 0.9854, "step": 119 }, { "epoch": 0.20887728459530025, "grad_norm": 0.50390625, "learning_rate": 9.920526969393267e-06, "loss": 1.0411, "step": 120 }, { "epoch": 0.21061792863359444, "grad_norm": 0.5078125, "learning_rate": 9.917949739757869e-06, "loss": 1.0254, "step": 121 }, { "epoch": 0.2123585726718886, "grad_norm": 0.5, "learning_rate": 9.915331731361104e-06, "loss": 0.9208, "step": 122 }, { "epoch": 0.21409921671018275, "grad_norm": 0.5078125, "learning_rate": 9.912672965911034e-06, "loss": 0.9195, "step": 123 }, { "epoch": 0.21583986074847694, "grad_norm": 0.48046875, "learning_rate": 9.909973465453666e-06, "loss": 0.9938, "step": 124 }, { "epoch": 0.2175805047867711, "grad_norm": 0.5078125, "learning_rate": 9.907233252372775e-06, "loss": 0.904, "step": 125 }, { "epoch": 0.2193211488250653, "grad_norm": 0.48828125, "learning_rate": 9.904452349389717e-06, "loss": 0.9882, "step": 126 }, { "epoch": 0.22106179286335945, "grad_norm": 0.50390625, "learning_rate": 9.901630779563247e-06, "loss": 0.9688, "step": 127 }, { "epoch": 0.2228024369016536, "grad_norm": 0.51953125, "learning_rate": 9.898768566289316e-06, "loss": 1.0522, "step": 128 }, { "epoch": 0.2245430809399478, "grad_norm": 0.48828125, "learning_rate": 9.895865733300887e-06, "loss": 0.9551, "step": 129 }, { "epoch": 0.22628372497824195, "grad_norm": 0.498046875, "learning_rate": 9.89292230466773e-06, "loss": 0.9623, "step": 130 }, { "epoch": 0.2280243690165361, "grad_norm": 0.515625, "learning_rate": 9.889938304796236e-06, "loss": 0.9702, "step": 131 }, { "epoch": 0.2297650130548303, "grad_norm": 0.5078125, "learning_rate": 9.886913758429194e-06, "loss": 0.9679, "step": 132 }, { "epoch": 0.23150565709312446, "grad_norm": 0.466796875, "learning_rate": 9.883848690645601e-06, "loss": 0.9719, "step": 133 }, { "epoch": 0.23324630113141862, "grad_norm": 0.4921875, "learning_rate": 9.880743126860458e-06, "loss": 0.9717, "step": 134 }, { "epoch": 0.2349869451697128, "grad_norm": 0.4921875, "learning_rate": 9.87759709282454e-06, "loss": 0.9365, "step": 135 }, { "epoch": 0.23672758920800696, "grad_norm": 0.478515625, "learning_rate": 9.874410614624202e-06, "loss": 0.9254, "step": 136 }, { "epoch": 0.23846823324630112, "grad_norm": 0.498046875, "learning_rate": 9.871183718681153e-06, "loss": 1.0045, "step": 137 }, { "epoch": 0.2402088772845953, "grad_norm": 0.51171875, "learning_rate": 9.867916431752237e-06, "loss": 0.9693, "step": 138 }, { "epoch": 0.24194952132288947, "grad_norm": 0.474609375, "learning_rate": 9.864608780929218e-06, "loss": 0.9981, "step": 139 }, { "epoch": 0.24369016536118362, "grad_norm": 0.4921875, "learning_rate": 9.861260793638539e-06, "loss": 0.9569, "step": 140 }, { "epoch": 0.2454308093994778, "grad_norm": 0.490234375, "learning_rate": 9.857872497641117e-06, "loss": 1.0422, "step": 141 }, { "epoch": 0.24717145343777197, "grad_norm": 0.49609375, "learning_rate": 9.854443921032098e-06, "loss": 1.0408, "step": 142 }, { "epoch": 0.24891209747606616, "grad_norm": 0.484375, "learning_rate": 9.850975092240625e-06, "loss": 0.8893, "step": 143 }, { "epoch": 0.2506527415143603, "grad_norm": 0.5078125, "learning_rate": 9.84746604002961e-06, "loss": 0.9679, "step": 144 }, { "epoch": 0.2506527415143603, "eval_loss": 0.9576423168182373, "eval_runtime": 59.1215, "eval_samples_per_second": 79.294, "eval_steps_per_second": 9.912, "step": 144 }, { "epoch": 0.2523933855526545, "grad_norm": 0.5, "learning_rate": 9.843916793495487e-06, "loss": 0.9071, "step": 145 }, { "epoch": 0.25413402959094866, "grad_norm": 0.498046875, "learning_rate": 9.840327382067972e-06, "loss": 0.9496, "step": 146 }, { "epoch": 0.2558746736292428, "grad_norm": 0.5078125, "learning_rate": 9.836697835509827e-06, "loss": 0.9864, "step": 147 }, { "epoch": 0.257615317667537, "grad_norm": 0.5, "learning_rate": 9.833028183916601e-06, "loss": 1.0082, "step": 148 }, { "epoch": 0.25935596170583114, "grad_norm": 0.494140625, "learning_rate": 9.829318457716395e-06, "loss": 0.9591, "step": 149 }, { "epoch": 0.26109660574412535, "grad_norm": 0.515625, "learning_rate": 9.82556868766959e-06, "loss": 1.0204, "step": 150 }, { "epoch": 0.2628372497824195, "grad_norm": 0.4921875, "learning_rate": 9.821778904868616e-06, "loss": 0.9805, "step": 151 }, { "epoch": 0.26457789382071367, "grad_norm": 0.5078125, "learning_rate": 9.817949140737672e-06, "loss": 0.9761, "step": 152 }, { "epoch": 0.26631853785900783, "grad_norm": 0.50390625, "learning_rate": 9.81407942703248e-06, "loss": 0.9789, "step": 153 }, { "epoch": 0.268059181897302, "grad_norm": 0.49609375, "learning_rate": 9.810169795840012e-06, "loss": 0.952, "step": 154 }, { "epoch": 0.26979982593559615, "grad_norm": 0.484375, "learning_rate": 9.806220279578236e-06, "loss": 0.9431, "step": 155 }, { "epoch": 0.27154046997389036, "grad_norm": 0.5078125, "learning_rate": 9.802230910995833e-06, "loss": 1.0015, "step": 156 }, { "epoch": 0.2732811140121845, "grad_norm": 0.49609375, "learning_rate": 9.798201723171938e-06, "loss": 0.9513, "step": 157 }, { "epoch": 0.2750217580504787, "grad_norm": 0.48828125, "learning_rate": 9.794132749515854e-06, "loss": 0.9454, "step": 158 }, { "epoch": 0.27676240208877284, "grad_norm": 0.494140625, "learning_rate": 9.790024023766789e-06, "loss": 0.9581, "step": 159 }, { "epoch": 0.278503046127067, "grad_norm": 0.482421875, "learning_rate": 9.785875579993558e-06, "loss": 0.9874, "step": 160 }, { "epoch": 0.28024369016536116, "grad_norm": 0.49609375, "learning_rate": 9.781687452594318e-06, "loss": 0.9417, "step": 161 }, { "epoch": 0.2819843342036554, "grad_norm": 0.478515625, "learning_rate": 9.777459676296276e-06, "loss": 0.9589, "step": 162 }, { "epoch": 0.28372497824194953, "grad_norm": 0.515625, "learning_rate": 9.773192286155395e-06, "loss": 0.9851, "step": 163 }, { "epoch": 0.2854656222802437, "grad_norm": 0.49609375, "learning_rate": 9.768885317556116e-06, "loss": 0.98, "step": 164 }, { "epoch": 0.28720626631853785, "grad_norm": 0.5078125, "learning_rate": 9.764538806211052e-06, "loss": 0.9651, "step": 165 }, { "epoch": 0.288946910356832, "grad_norm": 0.50390625, "learning_rate": 9.760152788160697e-06, "loss": 0.9407, "step": 166 }, { "epoch": 0.2906875543951262, "grad_norm": 0.46875, "learning_rate": 9.755727299773135e-06, "loss": 0.9553, "step": 167 }, { "epoch": 0.2924281984334204, "grad_norm": 0.470703125, "learning_rate": 9.75126237774372e-06, "loss": 0.9823, "step": 168 }, { "epoch": 0.29416884247171454, "grad_norm": 0.5, "learning_rate": 9.746758059094791e-06, "loss": 0.9832, "step": 169 }, { "epoch": 0.2959094865100087, "grad_norm": 0.50390625, "learning_rate": 9.742214381175355e-06, "loss": 0.8976, "step": 170 }, { "epoch": 0.29765013054830286, "grad_norm": 0.48828125, "learning_rate": 9.737631381660777e-06, "loss": 0.9331, "step": 171 }, { "epoch": 0.299390774586597, "grad_norm": 0.50390625, "learning_rate": 9.733009098552473e-06, "loss": 0.9666, "step": 172 }, { "epoch": 0.30113141862489123, "grad_norm": 0.490234375, "learning_rate": 9.728347570177587e-06, "loss": 0.9781, "step": 173 }, { "epoch": 0.3028720626631854, "grad_norm": 0.484375, "learning_rate": 9.723646835188681e-06, "loss": 0.969, "step": 174 }, { "epoch": 0.30461270670147955, "grad_norm": 0.51171875, "learning_rate": 9.71890693256341e-06, "loss": 0.9481, "step": 175 }, { "epoch": 0.3063533507397737, "grad_norm": 0.49609375, "learning_rate": 9.7141279016042e-06, "loss": 1.0301, "step": 176 }, { "epoch": 0.30809399477806787, "grad_norm": 0.48828125, "learning_rate": 9.709309781937925e-06, "loss": 1.023, "step": 177 }, { "epoch": 0.30983463881636203, "grad_norm": 0.490234375, "learning_rate": 9.704452613515571e-06, "loss": 0.9598, "step": 178 }, { "epoch": 0.31157528285465624, "grad_norm": 0.482421875, "learning_rate": 9.699556436611912e-06, "loss": 0.9699, "step": 179 }, { "epoch": 0.3133159268929504, "grad_norm": 0.498046875, "learning_rate": 9.694621291825174e-06, "loss": 0.903, "step": 180 }, { "epoch": 0.31505657093124456, "grad_norm": 0.515625, "learning_rate": 9.689647220076696e-06, "loss": 0.9581, "step": 181 }, { "epoch": 0.3167972149695387, "grad_norm": 0.490234375, "learning_rate": 9.684634262610593e-06, "loss": 0.939, "step": 182 }, { "epoch": 0.3185378590078329, "grad_norm": 0.498046875, "learning_rate": 9.679582460993413e-06, "loss": 0.9363, "step": 183 }, { "epoch": 0.3202785030461271, "grad_norm": 0.482421875, "learning_rate": 9.674491857113792e-06, "loss": 1.0215, "step": 184 }, { "epoch": 0.32201914708442125, "grad_norm": 0.48828125, "learning_rate": 9.669362493182112e-06, "loss": 0.9464, "step": 185 }, { "epoch": 0.3237597911227154, "grad_norm": 0.482421875, "learning_rate": 9.66419441173014e-06, "loss": 0.8955, "step": 186 }, { "epoch": 0.32550043516100957, "grad_norm": 0.494140625, "learning_rate": 9.658987655610687e-06, "loss": 0.9503, "step": 187 }, { "epoch": 0.32724107919930373, "grad_norm": 0.474609375, "learning_rate": 9.653742267997245e-06, "loss": 0.9808, "step": 188 }, { "epoch": 0.3289817232375979, "grad_norm": 0.48046875, "learning_rate": 9.648458292383631e-06, "loss": 1.0155, "step": 189 }, { "epoch": 0.3307223672758921, "grad_norm": 0.5078125, "learning_rate": 9.643135772583627e-06, "loss": 0.929, "step": 190 }, { "epoch": 0.33246301131418626, "grad_norm": 0.5, "learning_rate": 9.63777475273062e-06, "loss": 0.9214, "step": 191 }, { "epoch": 0.3342036553524804, "grad_norm": 0.47265625, "learning_rate": 9.632375277277226e-06, "loss": 0.9762, "step": 192 }, { "epoch": 0.3359442993907746, "grad_norm": 0.474609375, "learning_rate": 9.626937390994932e-06, "loss": 0.9734, "step": 193 }, { "epoch": 0.33768494342906874, "grad_norm": 0.498046875, "learning_rate": 9.621461138973725e-06, "loss": 0.9599, "step": 194 }, { "epoch": 0.3394255874673629, "grad_norm": 0.50390625, "learning_rate": 9.615946566621704e-06, "loss": 0.9448, "step": 195 }, { "epoch": 0.3411662315056571, "grad_norm": 0.494140625, "learning_rate": 9.61039371966472e-06, "loss": 0.9721, "step": 196 }, { "epoch": 0.3429068755439513, "grad_norm": 0.48828125, "learning_rate": 9.60480264414599e-06, "loss": 0.97, "step": 197 }, { "epoch": 0.34464751958224543, "grad_norm": 0.49609375, "learning_rate": 9.599173386425711e-06, "loss": 0.9964, "step": 198 }, { "epoch": 0.3463881636205396, "grad_norm": 0.486328125, "learning_rate": 9.593505993180687e-06, "loss": 0.9715, "step": 199 }, { "epoch": 0.34812880765883375, "grad_norm": 0.4921875, "learning_rate": 9.587800511403931e-06, "loss": 0.9655, "step": 200 }, { "epoch": 0.34986945169712796, "grad_norm": 0.5078125, "learning_rate": 9.582056988404276e-06, "loss": 0.9606, "step": 201 }, { "epoch": 0.3516100957354221, "grad_norm": 0.50390625, "learning_rate": 9.576275471805993e-06, "loss": 0.9166, "step": 202 }, { "epoch": 0.3533507397737163, "grad_norm": 0.494140625, "learning_rate": 9.570456009548383e-06, "loss": 0.9063, "step": 203 }, { "epoch": 0.35509138381201044, "grad_norm": 0.484375, "learning_rate": 9.564598649885391e-06, "loss": 0.9295, "step": 204 }, { "epoch": 0.3568320278503046, "grad_norm": 0.48046875, "learning_rate": 9.558703441385195e-06, "loss": 0.9933, "step": 205 }, { "epoch": 0.35857267188859876, "grad_norm": 0.48046875, "learning_rate": 9.552770432929812e-06, "loss": 0.9572, "step": 206 }, { "epoch": 0.360313315926893, "grad_norm": 0.482421875, "learning_rate": 9.54679967371469e-06, "loss": 0.9484, "step": 207 }, { "epoch": 0.36205395996518713, "grad_norm": 0.47265625, "learning_rate": 9.540791213248299e-06, "loss": 0.9266, "step": 208 }, { "epoch": 0.3637946040034813, "grad_norm": 0.5, "learning_rate": 9.534745101351719e-06, "loss": 0.9526, "step": 209 }, { "epoch": 0.36553524804177545, "grad_norm": 0.5, "learning_rate": 9.528661388158234e-06, "loss": 1.0046, "step": 210 }, { "epoch": 0.3672758920800696, "grad_norm": 0.51171875, "learning_rate": 9.522540124112902e-06, "loss": 0.9757, "step": 211 }, { "epoch": 0.36901653611836377, "grad_norm": 0.48046875, "learning_rate": 9.516381359972157e-06, "loss": 1.0195, "step": 212 }, { "epoch": 0.370757180156658, "grad_norm": 0.490234375, "learning_rate": 9.51018514680337e-06, "loss": 0.971, "step": 213 }, { "epoch": 0.37249782419495214, "grad_norm": 0.5, "learning_rate": 9.503951535984434e-06, "loss": 0.9336, "step": 214 }, { "epoch": 0.3742384682332463, "grad_norm": 0.494140625, "learning_rate": 9.49768057920334e-06, "loss": 0.9887, "step": 215 }, { "epoch": 0.37597911227154046, "grad_norm": 0.486328125, "learning_rate": 9.491372328457737e-06, "loss": 0.966, "step": 216 }, { "epoch": 0.37597911227154046, "eval_loss": 0.9439952373504639, "eval_runtime": 59.4647, "eval_samples_per_second": 78.837, "eval_steps_per_second": 9.855, "step": 216 }, { "epoch": 0.3777197563098346, "grad_norm": 0.494140625, "learning_rate": 9.485026836054519e-06, "loss": 0.9771, "step": 217 }, { "epoch": 0.37946040034812883, "grad_norm": 0.498046875, "learning_rate": 9.478644154609372e-06, "loss": 0.9988, "step": 218 }, { "epoch": 0.381201044386423, "grad_norm": 0.470703125, "learning_rate": 9.472224337046357e-06, "loss": 0.9875, "step": 219 }, { "epoch": 0.38294168842471715, "grad_norm": 0.482421875, "learning_rate": 9.46576743659745e-06, "loss": 0.9664, "step": 220 }, { "epoch": 0.3846823324630113, "grad_norm": 0.478515625, "learning_rate": 9.45927350680212e-06, "loss": 0.9693, "step": 221 }, { "epoch": 0.38642297650130547, "grad_norm": 0.48828125, "learning_rate": 9.452742601506873e-06, "loss": 0.9137, "step": 222 }, { "epoch": 0.38816362053959963, "grad_norm": 0.5078125, "learning_rate": 9.446174774864808e-06, "loss": 0.9047, "step": 223 }, { "epoch": 0.38990426457789384, "grad_norm": 0.486328125, "learning_rate": 9.439570081335173e-06, "loss": 0.9926, "step": 224 }, { "epoch": 0.391644908616188, "grad_norm": 0.478515625, "learning_rate": 9.432928575682908e-06, "loss": 0.9693, "step": 225 }, { "epoch": 0.39338555265448216, "grad_norm": 0.50390625, "learning_rate": 9.426250312978191e-06, "loss": 0.9106, "step": 226 }, { "epoch": 0.3951261966927763, "grad_norm": 0.5, "learning_rate": 9.419535348595985e-06, "loss": 0.9639, "step": 227 }, { "epoch": 0.3968668407310705, "grad_norm": 0.49609375, "learning_rate": 9.412783738215576e-06, "loss": 0.9338, "step": 228 }, { "epoch": 0.39860748476936464, "grad_norm": 0.48828125, "learning_rate": 9.405995537820111e-06, "loss": 1.0216, "step": 229 }, { "epoch": 0.40034812880765885, "grad_norm": 0.498046875, "learning_rate": 9.399170803696139e-06, "loss": 0.942, "step": 230 }, { "epoch": 0.402088772845953, "grad_norm": 0.486328125, "learning_rate": 9.392309592433134e-06, "loss": 0.9184, "step": 231 }, { "epoch": 0.40382941688424717, "grad_norm": 0.5234375, "learning_rate": 9.385411960923036e-06, "loss": 0.9178, "step": 232 }, { "epoch": 0.40557006092254133, "grad_norm": 0.4921875, "learning_rate": 9.378477966359773e-06, "loss": 0.9303, "step": 233 }, { "epoch": 0.4073107049608355, "grad_norm": 0.5, "learning_rate": 9.371507666238793e-06, "loss": 0.9563, "step": 234 }, { "epoch": 0.4090513489991297, "grad_norm": 0.47265625, "learning_rate": 9.364501118356579e-06, "loss": 0.92, "step": 235 }, { "epoch": 0.41079199303742386, "grad_norm": 0.4765625, "learning_rate": 9.357458380810175e-06, "loss": 0.9532, "step": 236 }, { "epoch": 0.412532637075718, "grad_norm": 0.482421875, "learning_rate": 9.350379511996706e-06, "loss": 0.9604, "step": 237 }, { "epoch": 0.4142732811140122, "grad_norm": 0.48828125, "learning_rate": 9.343264570612883e-06, "loss": 0.9415, "step": 238 }, { "epoch": 0.41601392515230634, "grad_norm": 0.48828125, "learning_rate": 9.336113615654535e-06, "loss": 0.9752, "step": 239 }, { "epoch": 0.4177545691906005, "grad_norm": 0.484375, "learning_rate": 9.328926706416102e-06, "loss": 0.9517, "step": 240 }, { "epoch": 0.4194952132288947, "grad_norm": 0.484375, "learning_rate": 9.321703902490152e-06, "loss": 0.9245, "step": 241 }, { "epoch": 0.4212358572671889, "grad_norm": 0.50390625, "learning_rate": 9.314445263766888e-06, "loss": 0.9341, "step": 242 }, { "epoch": 0.42297650130548303, "grad_norm": 0.47265625, "learning_rate": 9.307150850433643e-06, "loss": 0.9399, "step": 243 }, { "epoch": 0.4247171453437772, "grad_norm": 0.48828125, "learning_rate": 9.299820722974396e-06, "loss": 0.9865, "step": 244 }, { "epoch": 0.42645778938207135, "grad_norm": 0.5078125, "learning_rate": 9.29245494216925e-06, "loss": 0.9538, "step": 245 }, { "epoch": 0.4281984334203655, "grad_norm": 0.5, "learning_rate": 9.285053569093948e-06, "loss": 1.0095, "step": 246 }, { "epoch": 0.4299390774586597, "grad_norm": 0.494140625, "learning_rate": 9.277616665119352e-06, "loss": 0.9691, "step": 247 }, { "epoch": 0.4316797214969539, "grad_norm": 0.4921875, "learning_rate": 9.27014429191094e-06, "loss": 0.9854, "step": 248 }, { "epoch": 0.43342036553524804, "grad_norm": 0.498046875, "learning_rate": 9.262636511428304e-06, "loss": 0.9179, "step": 249 }, { "epoch": 0.4351610095735422, "grad_norm": 0.490234375, "learning_rate": 9.255093385924616e-06, "loss": 0.9388, "step": 250 }, { "epoch": 0.43690165361183636, "grad_norm": 0.51953125, "learning_rate": 9.247514977946124e-06, "loss": 0.9788, "step": 251 }, { "epoch": 0.4386422976501306, "grad_norm": 0.515625, "learning_rate": 9.239901350331635e-06, "loss": 0.9301, "step": 252 }, { "epoch": 0.44038294168842473, "grad_norm": 0.474609375, "learning_rate": 9.232252566211993e-06, "loss": 0.9656, "step": 253 }, { "epoch": 0.4421235857267189, "grad_norm": 0.4921875, "learning_rate": 9.224568689009548e-06, "loss": 1.0119, "step": 254 }, { "epoch": 0.44386422976501305, "grad_norm": 0.49609375, "learning_rate": 9.216849782437637e-06, "loss": 0.874, "step": 255 }, { "epoch": 0.4456048738033072, "grad_norm": 0.5, "learning_rate": 9.20909591050006e-06, "loss": 0.9191, "step": 256 }, { "epoch": 0.44734551784160137, "grad_norm": 0.50390625, "learning_rate": 9.201307137490536e-06, "loss": 0.9017, "step": 257 }, { "epoch": 0.4490861618798956, "grad_norm": 0.4921875, "learning_rate": 9.19348352799218e-06, "loss": 0.9363, "step": 258 }, { "epoch": 0.45082680591818974, "grad_norm": 0.494140625, "learning_rate": 9.185625146876966e-06, "loss": 0.9921, "step": 259 }, { "epoch": 0.4525674499564839, "grad_norm": 0.482421875, "learning_rate": 9.177732059305187e-06, "loss": 0.9358, "step": 260 }, { "epoch": 0.45430809399477806, "grad_norm": 0.494140625, "learning_rate": 9.169804330724916e-06, "loss": 0.9257, "step": 261 }, { "epoch": 0.4560487380330722, "grad_norm": 0.49609375, "learning_rate": 9.161842026871465e-06, "loss": 0.9201, "step": 262 }, { "epoch": 0.4577893820713664, "grad_norm": 0.498046875, "learning_rate": 9.153845213766837e-06, "loss": 0.9212, "step": 263 }, { "epoch": 0.4595300261096606, "grad_norm": 0.5, "learning_rate": 9.145813957719174e-06, "loss": 0.9735, "step": 264 }, { "epoch": 0.46127067014795475, "grad_norm": 0.5078125, "learning_rate": 9.137748325322223e-06, "loss": 0.9585, "step": 265 }, { "epoch": 0.4630113141862489, "grad_norm": 0.50390625, "learning_rate": 9.129648383454764e-06, "loss": 0.9781, "step": 266 }, { "epoch": 0.46475195822454307, "grad_norm": 0.486328125, "learning_rate": 9.121514199280072e-06, "loss": 0.9759, "step": 267 }, { "epoch": 0.46649260226283723, "grad_norm": 0.5078125, "learning_rate": 9.113345840245348e-06, "loss": 0.9688, "step": 268 }, { "epoch": 0.46823324630113144, "grad_norm": 0.49609375, "learning_rate": 9.105143374081167e-06, "loss": 0.9092, "step": 269 }, { "epoch": 0.4699738903394256, "grad_norm": 0.48046875, "learning_rate": 9.096906868800917e-06, "loss": 0.9357, "step": 270 }, { "epoch": 0.47171453437771976, "grad_norm": 0.4921875, "learning_rate": 9.088636392700227e-06, "loss": 1.0134, "step": 271 }, { "epoch": 0.4734551784160139, "grad_norm": 0.49609375, "learning_rate": 9.08033201435641e-06, "loss": 0.9494, "step": 272 }, { "epoch": 0.4751958224543081, "grad_norm": 0.48046875, "learning_rate": 9.071993802627887e-06, "loss": 0.9446, "step": 273 }, { "epoch": 0.47693646649260224, "grad_norm": 0.494140625, "learning_rate": 9.063621826653624e-06, "loss": 0.8926, "step": 274 }, { "epoch": 0.47867711053089645, "grad_norm": 0.484375, "learning_rate": 9.055216155852548e-06, "loss": 0.9216, "step": 275 }, { "epoch": 0.4804177545691906, "grad_norm": 0.470703125, "learning_rate": 9.046776859922983e-06, "loss": 0.9442, "step": 276 }, { "epoch": 0.4821583986074848, "grad_norm": 0.484375, "learning_rate": 9.038304008842064e-06, "loss": 0.9456, "step": 277 }, { "epoch": 0.48389904264577893, "grad_norm": 0.482421875, "learning_rate": 9.02979767286516e-06, "loss": 0.909, "step": 278 }, { "epoch": 0.4856396866840731, "grad_norm": 0.50390625, "learning_rate": 9.021257922525289e-06, "loss": 0.9597, "step": 279 }, { "epoch": 0.48738033072236725, "grad_norm": 0.5, "learning_rate": 9.012684828632538e-06, "loss": 0.9646, "step": 280 }, { "epoch": 0.48912097476066146, "grad_norm": 0.50390625, "learning_rate": 9.004078462273471e-06, "loss": 0.9679, "step": 281 }, { "epoch": 0.4908616187989556, "grad_norm": 0.5078125, "learning_rate": 8.995438894810541e-06, "loss": 0.952, "step": 282 }, { "epoch": 0.4926022628372498, "grad_norm": 0.484375, "learning_rate": 8.9867661978815e-06, "loss": 0.9792, "step": 283 }, { "epoch": 0.49434290687554394, "grad_norm": 0.5, "learning_rate": 8.978060443398802e-06, "loss": 0.939, "step": 284 }, { "epoch": 0.4960835509138381, "grad_norm": 0.486328125, "learning_rate": 8.96932170354901e-06, "loss": 0.9974, "step": 285 }, { "epoch": 0.4978241949521323, "grad_norm": 0.48828125, "learning_rate": 8.960550050792194e-06, "loss": 0.9947, "step": 286 }, { "epoch": 0.4995648389904265, "grad_norm": 0.50390625, "learning_rate": 8.951745557861333e-06, "loss": 0.9139, "step": 287 }, { "epoch": 0.5013054830287206, "grad_norm": 0.49609375, "learning_rate": 8.942908297761712e-06, "loss": 0.9397, "step": 288 }, { "epoch": 0.5013054830287206, "eval_loss": 0.9357889294624329, "eval_runtime": 60.8471, "eval_samples_per_second": 77.046, "eval_steps_per_second": 9.631, "step": 288 }, { "epoch": 0.5030461270670148, "grad_norm": 0.482421875, "learning_rate": 8.934038343770312e-06, "loss": 0.9656, "step": 289 }, { "epoch": 0.504786771105309, "grad_norm": 0.5078125, "learning_rate": 8.925135769435211e-06, "loss": 0.9896, "step": 290 }, { "epoch": 0.5065274151436031, "grad_norm": 0.5, "learning_rate": 8.916200648574964e-06, "loss": 0.905, "step": 291 }, { "epoch": 0.5082680591818973, "grad_norm": 0.490234375, "learning_rate": 8.907233055277999e-06, "loss": 0.9309, "step": 292 }, { "epoch": 0.5100087032201914, "grad_norm": 0.48046875, "learning_rate": 8.898233063902e-06, "loss": 0.9796, "step": 293 }, { "epoch": 0.5117493472584856, "grad_norm": 0.5234375, "learning_rate": 8.889200749073285e-06, "loss": 0.9335, "step": 294 }, { "epoch": 0.5134899912967799, "grad_norm": 0.498046875, "learning_rate": 8.880136185686202e-06, "loss": 0.9292, "step": 295 }, { "epoch": 0.515230635335074, "grad_norm": 0.478515625, "learning_rate": 8.871039448902488e-06, "loss": 0.9116, "step": 296 }, { "epoch": 0.5169712793733682, "grad_norm": 0.47265625, "learning_rate": 8.861910614150662e-06, "loss": 0.9315, "step": 297 }, { "epoch": 0.5187119234116623, "grad_norm": 0.478515625, "learning_rate": 8.852749757125392e-06, "loss": 0.9283, "step": 298 }, { "epoch": 0.5204525674499565, "grad_norm": 0.5078125, "learning_rate": 8.843556953786872e-06, "loss": 0.952, "step": 299 }, { "epoch": 0.5221932114882507, "grad_norm": 0.48828125, "learning_rate": 8.834332280360181e-06, "loss": 0.9999, "step": 300 }, { "epoch": 0.5239338555265448, "grad_norm": 0.498046875, "learning_rate": 8.82507581333467e-06, "loss": 0.9453, "step": 301 }, { "epoch": 0.525674499564839, "grad_norm": 0.482421875, "learning_rate": 8.815787629463306e-06, "loss": 0.8678, "step": 302 }, { "epoch": 0.5274151436031331, "grad_norm": 0.5, "learning_rate": 8.806467805762056e-06, "loss": 0.9878, "step": 303 }, { "epoch": 0.5291557876414273, "grad_norm": 0.49609375, "learning_rate": 8.797116419509232e-06, "loss": 0.8964, "step": 304 }, { "epoch": 0.5308964316797214, "grad_norm": 0.474609375, "learning_rate": 8.78773354824486e-06, "loss": 0.9584, "step": 305 }, { "epoch": 0.5326370757180157, "grad_norm": 0.482421875, "learning_rate": 8.778319269770033e-06, "loss": 0.9715, "step": 306 }, { "epoch": 0.5343777197563099, "grad_norm": 0.484375, "learning_rate": 8.768873662146271e-06, "loss": 0.9034, "step": 307 }, { "epoch": 0.536118363794604, "grad_norm": 0.494140625, "learning_rate": 8.759396803694863e-06, "loss": 0.9189, "step": 308 }, { "epoch": 0.5378590078328982, "grad_norm": 0.484375, "learning_rate": 8.749888772996226e-06, "loss": 1.0066, "step": 309 }, { "epoch": 0.5395996518711923, "grad_norm": 0.47265625, "learning_rate": 8.74034964888926e-06, "loss": 1.0147, "step": 310 }, { "epoch": 0.5413402959094865, "grad_norm": 0.484375, "learning_rate": 8.730779510470672e-06, "loss": 0.9504, "step": 311 }, { "epoch": 0.5430809399477807, "grad_norm": 0.482421875, "learning_rate": 8.721178437094346e-06, "loss": 0.9239, "step": 312 }, { "epoch": 0.5448215839860748, "grad_norm": 0.5078125, "learning_rate": 8.711546508370666e-06, "loss": 0.9145, "step": 313 }, { "epoch": 0.546562228024369, "grad_norm": 0.498046875, "learning_rate": 8.701883804165867e-06, "loss": 0.897, "step": 314 }, { "epoch": 0.5483028720626631, "grad_norm": 0.490234375, "learning_rate": 8.692190404601368e-06, "loss": 0.8796, "step": 315 }, { "epoch": 0.5500435161009574, "grad_norm": 0.53515625, "learning_rate": 8.682466390053106e-06, "loss": 1.03, "step": 316 }, { "epoch": 0.5517841601392516, "grad_norm": 0.482421875, "learning_rate": 8.672711841150877e-06, "loss": 0.9676, "step": 317 }, { "epoch": 0.5535248041775457, "grad_norm": 0.490234375, "learning_rate": 8.662926838777657e-06, "loss": 1.0098, "step": 318 }, { "epoch": 0.5552654482158399, "grad_norm": 0.5078125, "learning_rate": 8.653111464068937e-06, "loss": 0.9344, "step": 319 }, { "epoch": 0.557006092254134, "grad_norm": 0.51171875, "learning_rate": 8.643265798412057e-06, "loss": 0.9555, "step": 320 }, { "epoch": 0.5587467362924282, "grad_norm": 0.5, "learning_rate": 8.633389923445515e-06, "loss": 0.8845, "step": 321 }, { "epoch": 0.5604873803307223, "grad_norm": 0.48046875, "learning_rate": 8.623483921058304e-06, "loss": 0.9403, "step": 322 }, { "epoch": 0.5622280243690165, "grad_norm": 0.482421875, "learning_rate": 8.613547873389228e-06, "loss": 0.9654, "step": 323 }, { "epoch": 0.5639686684073107, "grad_norm": 0.490234375, "learning_rate": 8.603581862826222e-06, "loss": 1.0108, "step": 324 }, { "epoch": 0.5657093124456049, "grad_norm": 0.4921875, "learning_rate": 8.593585972005665e-06, "loss": 0.9708, "step": 325 }, { "epoch": 0.5674499564838991, "grad_norm": 0.484375, "learning_rate": 8.5835602838117e-06, "loss": 0.9333, "step": 326 }, { "epoch": 0.5691906005221932, "grad_norm": 0.478515625, "learning_rate": 8.573504881375543e-06, "loss": 0.9067, "step": 327 }, { "epoch": 0.5709312445604874, "grad_norm": 0.486328125, "learning_rate": 8.563419848074798e-06, "loss": 0.9388, "step": 328 }, { "epoch": 0.5726718885987816, "grad_norm": 0.4921875, "learning_rate": 8.55330526753276e-06, "loss": 0.9246, "step": 329 }, { "epoch": 0.5744125326370757, "grad_norm": 0.498046875, "learning_rate": 8.543161223617724e-06, "loss": 0.9222, "step": 330 }, { "epoch": 0.5761531766753699, "grad_norm": 0.490234375, "learning_rate": 8.532987800442292e-06, "loss": 1.0001, "step": 331 }, { "epoch": 0.577893820713664, "grad_norm": 0.498046875, "learning_rate": 8.522785082362675e-06, "loss": 1.0003, "step": 332 }, { "epoch": 0.5796344647519582, "grad_norm": 0.498046875, "learning_rate": 8.512553153977988e-06, "loss": 0.9103, "step": 333 }, { "epoch": 0.5813751087902524, "grad_norm": 0.498046875, "learning_rate": 8.502292100129553e-06, "loss": 0.9895, "step": 334 }, { "epoch": 0.5831157528285466, "grad_norm": 0.4921875, "learning_rate": 8.492002005900201e-06, "loss": 0.9981, "step": 335 }, { "epoch": 0.5848563968668408, "grad_norm": 0.4765625, "learning_rate": 8.481682956613555e-06, "loss": 0.8758, "step": 336 }, { "epoch": 0.5865970409051349, "grad_norm": 0.49609375, "learning_rate": 8.471335037833328e-06, "loss": 0.9845, "step": 337 }, { "epoch": 0.5883376849434291, "grad_norm": 0.474609375, "learning_rate": 8.460958335362617e-06, "loss": 0.9974, "step": 338 }, { "epoch": 0.5900783289817232, "grad_norm": 0.4921875, "learning_rate": 8.450552935243186e-06, "loss": 0.9725, "step": 339 }, { "epoch": 0.5918189730200174, "grad_norm": 0.48046875, "learning_rate": 8.440118923754757e-06, "loss": 0.9362, "step": 340 }, { "epoch": 0.5935596170583116, "grad_norm": 0.47265625, "learning_rate": 8.429656387414289e-06, "loss": 0.9659, "step": 341 }, { "epoch": 0.5953002610966057, "grad_norm": 0.5078125, "learning_rate": 8.419165412975265e-06, "loss": 0.9346, "step": 342 }, { "epoch": 0.5970409051348999, "grad_norm": 0.48828125, "learning_rate": 8.408646087426975e-06, "loss": 0.8759, "step": 343 }, { "epoch": 0.598781549173194, "grad_norm": 0.5, "learning_rate": 8.398098497993785e-06, "loss": 0.9359, "step": 344 }, { "epoch": 0.6005221932114883, "grad_norm": 0.5, "learning_rate": 8.387522732134428e-06, "loss": 0.9171, "step": 345 }, { "epoch": 0.6022628372497825, "grad_norm": 0.49609375, "learning_rate": 8.376918877541263e-06, "loss": 0.919, "step": 346 }, { "epoch": 0.6040034812880766, "grad_norm": 0.5234375, "learning_rate": 8.36628702213956e-06, "loss": 0.9211, "step": 347 }, { "epoch": 0.6057441253263708, "grad_norm": 0.494140625, "learning_rate": 8.355627254086771e-06, "loss": 0.9191, "step": 348 }, { "epoch": 0.6074847693646649, "grad_norm": 0.49609375, "learning_rate": 8.344939661771784e-06, "loss": 0.926, "step": 349 }, { "epoch": 0.6092254134029591, "grad_norm": 0.5, "learning_rate": 8.334224333814209e-06, "loss": 0.9088, "step": 350 }, { "epoch": 0.6109660574412533, "grad_norm": 0.50390625, "learning_rate": 8.323481359063631e-06, "loss": 0.9455, "step": 351 }, { "epoch": 0.6127067014795474, "grad_norm": 0.50390625, "learning_rate": 8.312710826598884e-06, "loss": 0.8873, "step": 352 }, { "epoch": 0.6144473455178416, "grad_norm": 0.49609375, "learning_rate": 8.301912825727294e-06, "loss": 0.9562, "step": 353 }, { "epoch": 0.6161879895561357, "grad_norm": 0.48828125, "learning_rate": 8.29108744598396e-06, "loss": 0.9324, "step": 354 }, { "epoch": 0.61792863359443, "grad_norm": 0.51171875, "learning_rate": 8.280234777131e-06, "loss": 0.9037, "step": 355 }, { "epoch": 0.6196692776327241, "grad_norm": 0.5078125, "learning_rate": 8.269354909156803e-06, "loss": 0.9127, "step": 356 }, { "epoch": 0.6214099216710183, "grad_norm": 0.494140625, "learning_rate": 8.258447932275296e-06, "loss": 0.9003, "step": 357 }, { "epoch": 0.6231505657093125, "grad_norm": 0.494140625, "learning_rate": 8.247513936925182e-06, "loss": 0.9871, "step": 358 }, { "epoch": 0.6248912097476066, "grad_norm": 0.50390625, "learning_rate": 8.236553013769198e-06, "loss": 0.8732, "step": 359 }, { "epoch": 0.6266318537859008, "grad_norm": 0.486328125, "learning_rate": 8.225565253693365e-06, "loss": 0.9563, "step": 360 }, { "epoch": 0.6266318537859008, "eval_loss": 0.9300395846366882, "eval_runtime": 59.3628, "eval_samples_per_second": 78.972, "eval_steps_per_second": 9.871, "step": 360 }, { "epoch": 0.6283724978241949, "grad_norm": 0.5078125, "learning_rate": 8.214550747806227e-06, "loss": 0.9173, "step": 361 }, { "epoch": 0.6301131418624891, "grad_norm": 0.494140625, "learning_rate": 8.2035095874381e-06, "loss": 0.9297, "step": 362 }, { "epoch": 0.6318537859007833, "grad_norm": 0.50390625, "learning_rate": 8.192441864140314e-06, "loss": 0.9582, "step": 363 }, { "epoch": 0.6335944299390774, "grad_norm": 0.50390625, "learning_rate": 8.181347669684456e-06, "loss": 0.9176, "step": 364 }, { "epoch": 0.6353350739773717, "grad_norm": 0.494140625, "learning_rate": 8.170227096061607e-06, "loss": 0.9181, "step": 365 }, { "epoch": 0.6370757180156658, "grad_norm": 0.486328125, "learning_rate": 8.15908023548158e-06, "loss": 0.9708, "step": 366 }, { "epoch": 0.63881636205396, "grad_norm": 0.48046875, "learning_rate": 8.147907180372147e-06, "loss": 0.9683, "step": 367 }, { "epoch": 0.6405570060922542, "grad_norm": 0.4765625, "learning_rate": 8.136708023378292e-06, "loss": 0.9728, "step": 368 }, { "epoch": 0.6422976501305483, "grad_norm": 0.498046875, "learning_rate": 8.125482857361426e-06, "loss": 0.9696, "step": 369 }, { "epoch": 0.6440382941688425, "grad_norm": 0.49609375, "learning_rate": 8.114231775398618e-06, "loss": 0.9189, "step": 370 }, { "epoch": 0.6457789382071366, "grad_norm": 0.486328125, "learning_rate": 8.102954870781831e-06, "loss": 0.9361, "step": 371 }, { "epoch": 0.6475195822454308, "grad_norm": 0.490234375, "learning_rate": 8.091652237017152e-06, "loss": 0.9612, "step": 372 }, { "epoch": 0.6492602262837249, "grad_norm": 0.498046875, "learning_rate": 8.080323967823993e-06, "loss": 0.9442, "step": 373 }, { "epoch": 0.6510008703220191, "grad_norm": 0.48828125, "learning_rate": 8.068970157134349e-06, "loss": 0.9346, "step": 374 }, { "epoch": 0.6527415143603134, "grad_norm": 0.486328125, "learning_rate": 8.057590899091985e-06, "loss": 0.8999, "step": 375 }, { "epoch": 0.6544821583986075, "grad_norm": 0.498046875, "learning_rate": 8.046186288051681e-06, "loss": 0.9674, "step": 376 }, { "epoch": 0.6562228024369017, "grad_norm": 0.48046875, "learning_rate": 8.034756418578434e-06, "loss": 0.9364, "step": 377 }, { "epoch": 0.6579634464751958, "grad_norm": 0.484375, "learning_rate": 8.023301385446682e-06, "loss": 0.9341, "step": 378 }, { "epoch": 0.65970409051349, "grad_norm": 0.4921875, "learning_rate": 8.011821283639515e-06, "loss": 0.8986, "step": 379 }, { "epoch": 0.6614447345517842, "grad_norm": 0.515625, "learning_rate": 8.000316208347891e-06, "loss": 0.9266, "step": 380 }, { "epoch": 0.6631853785900783, "grad_norm": 0.5078125, "learning_rate": 7.988786254969837e-06, "loss": 0.8911, "step": 381 }, { "epoch": 0.6649260226283725, "grad_norm": 0.5078125, "learning_rate": 7.977231519109665e-06, "loss": 0.9221, "step": 382 }, { "epoch": 0.6666666666666666, "grad_norm": 0.466796875, "learning_rate": 7.965652096577188e-06, "loss": 0.9635, "step": 383 }, { "epoch": 0.6684073107049608, "grad_norm": 0.48828125, "learning_rate": 7.954048083386909e-06, "loss": 0.9023, "step": 384 }, { "epoch": 0.6701479547432551, "grad_norm": 0.48828125, "learning_rate": 7.942419575757235e-06, "loss": 0.9277, "step": 385 }, { "epoch": 0.6718885987815492, "grad_norm": 0.486328125, "learning_rate": 7.930766670109675e-06, "loss": 0.9826, "step": 386 }, { "epoch": 0.6736292428198434, "grad_norm": 0.490234375, "learning_rate": 7.919089463068038e-06, "loss": 0.9223, "step": 387 }, { "epoch": 0.6753698868581375, "grad_norm": 0.51953125, "learning_rate": 7.907388051457647e-06, "loss": 1.0468, "step": 388 }, { "epoch": 0.6771105308964317, "grad_norm": 0.4921875, "learning_rate": 7.895662532304516e-06, "loss": 0.9233, "step": 389 }, { "epoch": 0.6788511749347258, "grad_norm": 0.478515625, "learning_rate": 7.883913002834555e-06, "loss": 0.9756, "step": 390 }, { "epoch": 0.68059181897302, "grad_norm": 0.494140625, "learning_rate": 7.872139560472767e-06, "loss": 0.9, "step": 391 }, { "epoch": 0.6823324630113142, "grad_norm": 0.51953125, "learning_rate": 7.860342302842434e-06, "loss": 0.894, "step": 392 }, { "epoch": 0.6840731070496083, "grad_norm": 0.5, "learning_rate": 7.848521327764309e-06, "loss": 0.9369, "step": 393 }, { "epoch": 0.6858137510879025, "grad_norm": 0.5, "learning_rate": 7.836676733255809e-06, "loss": 0.9416, "step": 394 }, { "epoch": 0.6875543951261966, "grad_norm": 0.494140625, "learning_rate": 7.824808617530197e-06, "loss": 0.8992, "step": 395 }, { "epoch": 0.6892950391644909, "grad_norm": 0.486328125, "learning_rate": 7.812917078995769e-06, "loss": 0.9457, "step": 396 }, { "epoch": 0.6910356832027851, "grad_norm": 0.4765625, "learning_rate": 7.801002216255042e-06, "loss": 0.9484, "step": 397 }, { "epoch": 0.6927763272410792, "grad_norm": 0.50390625, "learning_rate": 7.78906412810393e-06, "loss": 0.9297, "step": 398 }, { "epoch": 0.6945169712793734, "grad_norm": 0.498046875, "learning_rate": 7.777102913530927e-06, "loss": 0.951, "step": 399 }, { "epoch": 0.6962576153176675, "grad_norm": 0.4921875, "learning_rate": 7.76511867171629e-06, "loss": 0.9091, "step": 400 }, { "epoch": 0.6979982593559617, "grad_norm": 0.49609375, "learning_rate": 7.753111502031214e-06, "loss": 0.9041, "step": 401 }, { "epoch": 0.6997389033942559, "grad_norm": 0.50390625, "learning_rate": 7.741081504037009e-06, "loss": 0.9463, "step": 402 }, { "epoch": 0.70147954743255, "grad_norm": 0.486328125, "learning_rate": 7.729028777484266e-06, "loss": 0.9494, "step": 403 }, { "epoch": 0.7032201914708442, "grad_norm": 0.482421875, "learning_rate": 7.716953422312044e-06, "loss": 0.9266, "step": 404 }, { "epoch": 0.7049608355091384, "grad_norm": 0.5078125, "learning_rate": 7.704855538647033e-06, "loss": 0.9731, "step": 405 }, { "epoch": 0.7067014795474326, "grad_norm": 0.490234375, "learning_rate": 7.692735226802729e-06, "loss": 0.961, "step": 406 }, { "epoch": 0.7084421235857267, "grad_norm": 0.482421875, "learning_rate": 7.680592587278585e-06, "loss": 0.9506, "step": 407 }, { "epoch": 0.7101827676240209, "grad_norm": 0.5, "learning_rate": 7.668427720759207e-06, "loss": 0.9084, "step": 408 }, { "epoch": 0.7119234116623151, "grad_norm": 0.5, "learning_rate": 7.656240728113493e-06, "loss": 0.9147, "step": 409 }, { "epoch": 0.7136640557006092, "grad_norm": 0.4921875, "learning_rate": 7.644031710393815e-06, "loss": 0.9498, "step": 410 }, { "epoch": 0.7154046997389034, "grad_norm": 0.490234375, "learning_rate": 7.631800768835167e-06, "loss": 0.9145, "step": 411 }, { "epoch": 0.7171453437771975, "grad_norm": 0.498046875, "learning_rate": 7.619548004854332e-06, "loss": 0.8884, "step": 412 }, { "epoch": 0.7188859878154917, "grad_norm": 0.48828125, "learning_rate": 7.607273520049041e-06, "loss": 0.9777, "step": 413 }, { "epoch": 0.720626631853786, "grad_norm": 0.478515625, "learning_rate": 7.594977416197134e-06, "loss": 0.9802, "step": 414 }, { "epoch": 0.72236727589208, "grad_norm": 0.49609375, "learning_rate": 7.582659795255707e-06, "loss": 0.9484, "step": 415 }, { "epoch": 0.7241079199303743, "grad_norm": 0.5078125, "learning_rate": 7.570320759360273e-06, "loss": 0.9231, "step": 416 }, { "epoch": 0.7258485639686684, "grad_norm": 0.498046875, "learning_rate": 7.557960410823917e-06, "loss": 0.9416, "step": 417 }, { "epoch": 0.7275892080069626, "grad_norm": 0.4921875, "learning_rate": 7.545578852136443e-06, "loss": 1.0058, "step": 418 }, { "epoch": 0.7293298520452568, "grad_norm": 0.5, "learning_rate": 7.533176185963523e-06, "loss": 0.9426, "step": 419 }, { "epoch": 0.7310704960835509, "grad_norm": 0.498046875, "learning_rate": 7.520752515145855e-06, "loss": 0.949, "step": 420 }, { "epoch": 0.7328111401218451, "grad_norm": 0.484375, "learning_rate": 7.508307942698296e-06, "loss": 0.9727, "step": 421 }, { "epoch": 0.7345517841601392, "grad_norm": 0.515625, "learning_rate": 7.495842571809021e-06, "loss": 0.8763, "step": 422 }, { "epoch": 0.7362924281984334, "grad_norm": 0.5, "learning_rate": 7.4833565058386595e-06, "loss": 0.8914, "step": 423 }, { "epoch": 0.7380330722367275, "grad_norm": 0.49609375, "learning_rate": 7.470849848319443e-06, "loss": 0.9157, "step": 424 }, { "epoch": 0.7397737162750218, "grad_norm": 0.490234375, "learning_rate": 7.458322702954342e-06, "loss": 0.8703, "step": 425 }, { "epoch": 0.741514360313316, "grad_norm": 0.48828125, "learning_rate": 7.44577517361621e-06, "loss": 0.9106, "step": 426 }, { "epoch": 0.7432550043516101, "grad_norm": 0.5078125, "learning_rate": 7.4332073643469196e-06, "loss": 0.8928, "step": 427 }, { "epoch": 0.7449956483899043, "grad_norm": 0.5234375, "learning_rate": 7.420619379356504e-06, "loss": 0.9402, "step": 428 }, { "epoch": 0.7467362924281984, "grad_norm": 0.490234375, "learning_rate": 7.408011323022286e-06, "loss": 0.8997, "step": 429 }, { "epoch": 0.7484769364664926, "grad_norm": 0.48828125, "learning_rate": 7.395383299888019e-06, "loss": 0.9831, "step": 430 }, { "epoch": 0.7502175805047868, "grad_norm": 0.494140625, "learning_rate": 7.382735414663017e-06, "loss": 1.04, "step": 431 }, { "epoch": 0.7519582245430809, "grad_norm": 0.49609375, "learning_rate": 7.370067772221285e-06, "loss": 0.9034, "step": 432 }, { "epoch": 0.7519582245430809, "eval_loss": 0.9259106516838074, "eval_runtime": 59.5681, "eval_samples_per_second": 78.7, "eval_steps_per_second": 9.837, "step": 432 }, { "epoch": 0.7536988685813751, "grad_norm": 0.4921875, "learning_rate": 7.357380477600654e-06, "loss": 0.9201, "step": 433 }, { "epoch": 0.7554395126196692, "grad_norm": 0.5078125, "learning_rate": 7.3446736360019065e-06, "loss": 0.9291, "step": 434 }, { "epoch": 0.7571801566579635, "grad_norm": 0.47265625, "learning_rate": 7.331947352787905e-06, "loss": 0.951, "step": 435 }, { "epoch": 0.7589208006962577, "grad_norm": 0.49609375, "learning_rate": 7.319201733482715e-06, "loss": 0.9208, "step": 436 }, { "epoch": 0.7606614447345518, "grad_norm": 0.490234375, "learning_rate": 7.3064368837707425e-06, "loss": 0.9116, "step": 437 }, { "epoch": 0.762402088772846, "grad_norm": 0.50390625, "learning_rate": 7.2936529094958365e-06, "loss": 0.9165, "step": 438 }, { "epoch": 0.7641427328111401, "grad_norm": 0.494140625, "learning_rate": 7.280849916660434e-06, "loss": 0.9513, "step": 439 }, { "epoch": 0.7658833768494343, "grad_norm": 0.48828125, "learning_rate": 7.268028011424664e-06, "loss": 0.9725, "step": 440 }, { "epoch": 0.7676240208877284, "grad_norm": 0.494140625, "learning_rate": 7.255187300105477e-06, "loss": 0.9328, "step": 441 }, { "epoch": 0.7693646649260226, "grad_norm": 0.50390625, "learning_rate": 7.24232788917576e-06, "loss": 0.8776, "step": 442 }, { "epoch": 0.7711053089643168, "grad_norm": 0.5078125, "learning_rate": 7.229449885263451e-06, "loss": 0.9432, "step": 443 }, { "epoch": 0.7728459530026109, "grad_norm": 0.478515625, "learning_rate": 7.21655339515066e-06, "loss": 0.8951, "step": 444 }, { "epoch": 0.7745865970409052, "grad_norm": 0.4921875, "learning_rate": 7.203638525772783e-06, "loss": 0.961, "step": 445 }, { "epoch": 0.7763272410791993, "grad_norm": 0.48828125, "learning_rate": 7.1907053842176075e-06, "loss": 0.9498, "step": 446 }, { "epoch": 0.7780678851174935, "grad_norm": 0.498046875, "learning_rate": 7.17775407772444e-06, "loss": 0.9638, "step": 447 }, { "epoch": 0.7798085291557877, "grad_norm": 0.490234375, "learning_rate": 7.164784713683197e-06, "loss": 0.9516, "step": 448 }, { "epoch": 0.7815491731940818, "grad_norm": 0.466796875, "learning_rate": 7.1517973996335335e-06, "loss": 0.9108, "step": 449 }, { "epoch": 0.783289817232376, "grad_norm": 0.50390625, "learning_rate": 7.138792243263936e-06, "loss": 0.9089, "step": 450 }, { "epoch": 0.7850304612706701, "grad_norm": 0.50390625, "learning_rate": 7.125769352410845e-06, "loss": 0.9667, "step": 451 }, { "epoch": 0.7867711053089643, "grad_norm": 0.484375, "learning_rate": 7.112728835057742e-06, "loss": 0.9458, "step": 452 }, { "epoch": 0.7885117493472585, "grad_norm": 0.478515625, "learning_rate": 7.099670799334269e-06, "loss": 0.899, "step": 453 }, { "epoch": 0.7902523933855526, "grad_norm": 0.498046875, "learning_rate": 7.08659535351533e-06, "loss": 0.9031, "step": 454 }, { "epoch": 0.7919930374238469, "grad_norm": 0.49609375, "learning_rate": 7.073502606020187e-06, "loss": 0.9771, "step": 455 }, { "epoch": 0.793733681462141, "grad_norm": 0.52734375, "learning_rate": 7.060392665411564e-06, "loss": 0.9113, "step": 456 }, { "epoch": 0.7954743255004352, "grad_norm": 0.498046875, "learning_rate": 7.0472656403947505e-06, "loss": 1.0172, "step": 457 }, { "epoch": 0.7972149695387293, "grad_norm": 0.4765625, "learning_rate": 7.034121639816691e-06, "loss": 0.9282, "step": 458 }, { "epoch": 0.7989556135770235, "grad_norm": 0.50390625, "learning_rate": 7.020960772665096e-06, "loss": 0.896, "step": 459 }, { "epoch": 0.8006962576153177, "grad_norm": 0.490234375, "learning_rate": 7.007783148067524e-06, "loss": 0.881, "step": 460 }, { "epoch": 0.8024369016536118, "grad_norm": 0.5078125, "learning_rate": 6.994588875290488e-06, "loss": 0.9155, "step": 461 }, { "epoch": 0.804177545691906, "grad_norm": 0.5078125, "learning_rate": 6.9813780637385385e-06, "loss": 0.9012, "step": 462 }, { "epoch": 0.8059181897302001, "grad_norm": 0.50390625, "learning_rate": 6.968150822953372e-06, "loss": 0.9085, "step": 463 }, { "epoch": 0.8076588337684943, "grad_norm": 0.498046875, "learning_rate": 6.954907262612906e-06, "loss": 0.9818, "step": 464 }, { "epoch": 0.8093994778067886, "grad_norm": 0.484375, "learning_rate": 6.941647492530378e-06, "loss": 0.9717, "step": 465 }, { "epoch": 0.8111401218450827, "grad_norm": 0.5, "learning_rate": 6.928371622653434e-06, "loss": 0.9369, "step": 466 }, { "epoch": 0.8128807658833769, "grad_norm": 0.486328125, "learning_rate": 6.91507976306322e-06, "loss": 0.8943, "step": 467 }, { "epoch": 0.814621409921671, "grad_norm": 0.498046875, "learning_rate": 6.901772023973459e-06, "loss": 0.9113, "step": 468 }, { "epoch": 0.8163620539599652, "grad_norm": 0.49609375, "learning_rate": 6.888448515729552e-06, "loss": 0.8995, "step": 469 }, { "epoch": 0.8181026979982594, "grad_norm": 0.486328125, "learning_rate": 6.8751093488076485e-06, "loss": 0.9159, "step": 470 }, { "epoch": 0.8198433420365535, "grad_norm": 0.498046875, "learning_rate": 6.86175463381374e-06, "loss": 0.9388, "step": 471 }, { "epoch": 0.8215839860748477, "grad_norm": 0.49609375, "learning_rate": 6.8483844814827405e-06, "loss": 0.9524, "step": 472 }, { "epoch": 0.8233246301131418, "grad_norm": 0.5078125, "learning_rate": 6.8349990026775656e-06, "loss": 0.9458, "step": 473 }, { "epoch": 0.825065274151436, "grad_norm": 0.5, "learning_rate": 6.821598308388217e-06, "loss": 0.9258, "step": 474 }, { "epoch": 0.8268059181897301, "grad_norm": 0.482421875, "learning_rate": 6.8081825097308584e-06, "loss": 0.9159, "step": 475 }, { "epoch": 0.8285465622280244, "grad_norm": 0.5, "learning_rate": 6.794751717946897e-06, "loss": 0.9029, "step": 476 }, { "epoch": 0.8302872062663186, "grad_norm": 0.484375, "learning_rate": 6.781306044402064e-06, "loss": 0.8852, "step": 477 }, { "epoch": 0.8320278503046127, "grad_norm": 0.482421875, "learning_rate": 6.767845600585479e-06, "loss": 0.945, "step": 478 }, { "epoch": 0.8337684943429069, "grad_norm": 0.474609375, "learning_rate": 6.754370498108747e-06, "loss": 0.9776, "step": 479 }, { "epoch": 0.835509138381201, "grad_norm": 0.490234375, "learning_rate": 6.740880848705005e-06, "loss": 0.9229, "step": 480 }, { "epoch": 0.8372497824194952, "grad_norm": 0.50390625, "learning_rate": 6.72737676422802e-06, "loss": 0.9686, "step": 481 }, { "epoch": 0.8389904264577894, "grad_norm": 0.484375, "learning_rate": 6.713858356651253e-06, "loss": 0.9459, "step": 482 }, { "epoch": 0.8407310704960835, "grad_norm": 0.486328125, "learning_rate": 6.700325738066923e-06, "loss": 0.9472, "step": 483 }, { "epoch": 0.8424717145343777, "grad_norm": 0.5, "learning_rate": 6.686779020685089e-06, "loss": 0.9417, "step": 484 }, { "epoch": 0.8442123585726719, "grad_norm": 0.490234375, "learning_rate": 6.6732183168327146e-06, "loss": 0.9666, "step": 485 }, { "epoch": 0.8459530026109661, "grad_norm": 0.490234375, "learning_rate": 6.659643738952732e-06, "loss": 0.9368, "step": 486 }, { "epoch": 0.8476936466492603, "grad_norm": 0.484375, "learning_rate": 6.646055399603122e-06, "loss": 0.9466, "step": 487 }, { "epoch": 0.8494342906875544, "grad_norm": 0.5, "learning_rate": 6.6324534114559656e-06, "loss": 0.9159, "step": 488 }, { "epoch": 0.8511749347258486, "grad_norm": 0.5078125, "learning_rate": 6.618837887296523e-06, "loss": 0.9067, "step": 489 }, { "epoch": 0.8529155787641427, "grad_norm": 0.48046875, "learning_rate": 6.605208940022289e-06, "loss": 0.9608, "step": 490 }, { "epoch": 0.8546562228024369, "grad_norm": 0.498046875, "learning_rate": 6.591566682642061e-06, "loss": 0.9663, "step": 491 }, { "epoch": 0.856396866840731, "grad_norm": 0.5234375, "learning_rate": 6.5779112282750035e-06, "loss": 0.8931, "step": 492 }, { "epoch": 0.8581375108790252, "grad_norm": 0.498046875, "learning_rate": 6.564242690149705e-06, "loss": 0.9383, "step": 493 }, { "epoch": 0.8598781549173194, "grad_norm": 0.484375, "learning_rate": 6.550561181603244e-06, "loss": 0.9343, "step": 494 }, { "epoch": 0.8616187989556136, "grad_norm": 0.486328125, "learning_rate": 6.536866816080247e-06, "loss": 0.9999, "step": 495 }, { "epoch": 0.8633594429939078, "grad_norm": 0.4765625, "learning_rate": 6.523159707131951e-06, "loss": 0.9955, "step": 496 }, { "epoch": 0.8651000870322019, "grad_norm": 0.50390625, "learning_rate": 6.509439968415252e-06, "loss": 0.8897, "step": 497 }, { "epoch": 0.8668407310704961, "grad_norm": 0.498046875, "learning_rate": 6.4957077136917776e-06, "loss": 0.9065, "step": 498 }, { "epoch": 0.8685813751087903, "grad_norm": 0.4921875, "learning_rate": 6.481963056826932e-06, "loss": 0.9224, "step": 499 }, { "epoch": 0.8703220191470844, "grad_norm": 0.50390625, "learning_rate": 6.468206111788957e-06, "loss": 0.9638, "step": 500 }, { "epoch": 0.8720626631853786, "grad_norm": 0.5, "learning_rate": 6.454436992647984e-06, "loss": 0.9232, "step": 501 }, { "epoch": 0.8738033072236727, "grad_norm": 0.51171875, "learning_rate": 6.440655813575093e-06, "loss": 0.9412, "step": 502 }, { "epoch": 0.8755439512619669, "grad_norm": 0.5078125, "learning_rate": 6.426862688841359e-06, "loss": 0.9051, "step": 503 }, { "epoch": 0.8772845953002611, "grad_norm": 0.5078125, "learning_rate": 6.413057732816911e-06, "loss": 0.9214, "step": 504 }, { "epoch": 0.8772845953002611, "eval_loss": 0.9230473637580872, "eval_runtime": 59.2532, "eval_samples_per_second": 79.118, "eval_steps_per_second": 9.89, "step": 504 }, { "epoch": 0.8790252393385553, "grad_norm": 0.49609375, "learning_rate": 6.3992410599699786e-06, "loss": 0.908, "step": 505 }, { "epoch": 0.8807658833768495, "grad_norm": 0.48828125, "learning_rate": 6.385412784865948e-06, "loss": 0.9613, "step": 506 }, { "epoch": 0.8825065274151436, "grad_norm": 0.498046875, "learning_rate": 6.371573022166409e-06, "loss": 0.9061, "step": 507 }, { "epoch": 0.8842471714534378, "grad_norm": 0.5, "learning_rate": 6.357721886628201e-06, "loss": 0.9117, "step": 508 }, { "epoch": 0.8859878154917319, "grad_norm": 0.48046875, "learning_rate": 6.34385949310247e-06, "loss": 0.9219, "step": 509 }, { "epoch": 0.8877284595300261, "grad_norm": 0.5078125, "learning_rate": 6.329985956533708e-06, "loss": 0.976, "step": 510 }, { "epoch": 0.8894691035683203, "grad_norm": 0.515625, "learning_rate": 6.3161013919588e-06, "loss": 0.9131, "step": 511 }, { "epoch": 0.8912097476066144, "grad_norm": 0.5, "learning_rate": 6.302205914506083e-06, "loss": 0.908, "step": 512 }, { "epoch": 0.8929503916449086, "grad_norm": 0.4921875, "learning_rate": 6.2882996393943706e-06, "loss": 0.9786, "step": 513 }, { "epoch": 0.8946910356832027, "grad_norm": 0.48828125, "learning_rate": 6.274382681932019e-06, "loss": 0.8728, "step": 514 }, { "epoch": 0.896431679721497, "grad_norm": 0.482421875, "learning_rate": 6.2604551575159476e-06, "loss": 0.9491, "step": 515 }, { "epoch": 0.8981723237597912, "grad_norm": 0.49609375, "learning_rate": 6.24651718163071e-06, "loss": 0.995, "step": 516 }, { "epoch": 0.8999129677980853, "grad_norm": 0.482421875, "learning_rate": 6.2325688698475106e-06, "loss": 0.9574, "step": 517 }, { "epoch": 0.9016536118363795, "grad_norm": 0.486328125, "learning_rate": 6.218610337823262e-06, "loss": 0.9004, "step": 518 }, { "epoch": 0.9033942558746736, "grad_norm": 0.48046875, "learning_rate": 6.2046417012996195e-06, "loss": 0.9165, "step": 519 }, { "epoch": 0.9051348999129678, "grad_norm": 0.498046875, "learning_rate": 6.1906630761020245e-06, "loss": 0.9534, "step": 520 }, { "epoch": 0.906875543951262, "grad_norm": 0.5, "learning_rate": 6.17667457813874e-06, "loss": 0.8883, "step": 521 }, { "epoch": 0.9086161879895561, "grad_norm": 0.51171875, "learning_rate": 6.162676323399898e-06, "loss": 0.919, "step": 522 }, { "epoch": 0.9103568320278503, "grad_norm": 0.4921875, "learning_rate": 6.148668427956523e-06, "loss": 0.9187, "step": 523 }, { "epoch": 0.9120974760661444, "grad_norm": 0.494140625, "learning_rate": 6.134651007959586e-06, "loss": 0.8984, "step": 524 }, { "epoch": 0.9138381201044387, "grad_norm": 0.4921875, "learning_rate": 6.120624179639032e-06, "loss": 0.9585, "step": 525 }, { "epoch": 0.9155787641427328, "grad_norm": 0.49609375, "learning_rate": 6.106588059302818e-06, "loss": 0.9848, "step": 526 }, { "epoch": 0.917319408181027, "grad_norm": 0.4921875, "learning_rate": 6.092542763335947e-06, "loss": 0.9542, "step": 527 }, { "epoch": 0.9190600522193212, "grad_norm": 0.5, "learning_rate": 6.0784884081995065e-06, "loss": 0.9476, "step": 528 }, { "epoch": 0.9208006962576153, "grad_norm": 0.498046875, "learning_rate": 6.0644251104296995e-06, "loss": 0.9425, "step": 529 }, { "epoch": 0.9225413402959095, "grad_norm": 0.5, "learning_rate": 6.0503529866368824e-06, "loss": 0.9532, "step": 530 }, { "epoch": 0.9242819843342036, "grad_norm": 0.49609375, "learning_rate": 6.036272153504592e-06, "loss": 0.9243, "step": 531 }, { "epoch": 0.9260226283724978, "grad_norm": 0.490234375, "learning_rate": 6.022182727788586e-06, "loss": 0.9452, "step": 532 }, { "epoch": 0.927763272410792, "grad_norm": 0.5234375, "learning_rate": 6.008084826315863e-06, "loss": 0.9243, "step": 533 }, { "epoch": 0.9295039164490861, "grad_norm": 0.498046875, "learning_rate": 5.993978565983709e-06, "loss": 0.8917, "step": 534 }, { "epoch": 0.9312445604873804, "grad_norm": 0.4921875, "learning_rate": 5.979864063758717e-06, "loss": 0.9562, "step": 535 }, { "epoch": 0.9329852045256745, "grad_norm": 0.478515625, "learning_rate": 5.965741436675816e-06, "loss": 0.9904, "step": 536 }, { "epoch": 0.9347258485639687, "grad_norm": 0.5, "learning_rate": 5.9516108018373145e-06, "loss": 0.8953, "step": 537 }, { "epoch": 0.9364664926022629, "grad_norm": 0.494140625, "learning_rate": 5.937472276411909e-06, "loss": 0.8674, "step": 538 }, { "epoch": 0.938207136640557, "grad_norm": 0.498046875, "learning_rate": 5.923325977633732e-06, "loss": 1.0281, "step": 539 }, { "epoch": 0.9399477806788512, "grad_norm": 0.494140625, "learning_rate": 5.909172022801364e-06, "loss": 0.9278, "step": 540 }, { "epoch": 0.9416884247171453, "grad_norm": 0.486328125, "learning_rate": 5.8950105292768754e-06, "loss": 0.8704, "step": 541 }, { "epoch": 0.9434290687554395, "grad_norm": 0.515625, "learning_rate": 5.880841614484841e-06, "loss": 0.9561, "step": 542 }, { "epoch": 0.9451697127937336, "grad_norm": 0.47265625, "learning_rate": 5.866665395911375e-06, "loss": 0.9407, "step": 543 }, { "epoch": 0.9469103568320278, "grad_norm": 0.484375, "learning_rate": 5.852481991103149e-06, "loss": 0.8833, "step": 544 }, { "epoch": 0.9486510008703221, "grad_norm": 0.5, "learning_rate": 5.838291517666427e-06, "loss": 0.993, "step": 545 }, { "epoch": 0.9503916449086162, "grad_norm": 0.494140625, "learning_rate": 5.824094093266077e-06, "loss": 0.9231, "step": 546 }, { "epoch": 0.9521322889469104, "grad_norm": 0.490234375, "learning_rate": 5.809889835624611e-06, "loss": 0.9512, "step": 547 }, { "epoch": 0.9538729329852045, "grad_norm": 0.494140625, "learning_rate": 5.795678862521197e-06, "loss": 0.9175, "step": 548 }, { "epoch": 0.9556135770234987, "grad_norm": 0.486328125, "learning_rate": 5.781461291790687e-06, "loss": 0.8922, "step": 549 }, { "epoch": 0.9573542210617929, "grad_norm": 0.494140625, "learning_rate": 5.767237241322641e-06, "loss": 0.9701, "step": 550 }, { "epoch": 0.959094865100087, "grad_norm": 0.49609375, "learning_rate": 5.753006829060343e-06, "loss": 0.9464, "step": 551 }, { "epoch": 0.9608355091383812, "grad_norm": 0.5, "learning_rate": 5.738770172999835e-06, "loss": 0.9335, "step": 552 }, { "epoch": 0.9625761531766753, "grad_norm": 0.458984375, "learning_rate": 5.724527391188927e-06, "loss": 0.9689, "step": 553 }, { "epoch": 0.9643167972149695, "grad_norm": 0.50390625, "learning_rate": 5.710278601726222e-06, "loss": 0.8756, "step": 554 }, { "epoch": 0.9660574412532638, "grad_norm": 0.484375, "learning_rate": 5.696023922760141e-06, "loss": 0.9459, "step": 555 }, { "epoch": 0.9677980852915579, "grad_norm": 0.50390625, "learning_rate": 5.681763472487933e-06, "loss": 0.9305, "step": 556 }, { "epoch": 0.9695387293298521, "grad_norm": 0.50390625, "learning_rate": 5.667497369154712e-06, "loss": 0.8707, "step": 557 }, { "epoch": 0.9712793733681462, "grad_norm": 0.47265625, "learning_rate": 5.6532257310524565e-06, "loss": 0.8977, "step": 558 }, { "epoch": 0.9730200174064404, "grad_norm": 0.494140625, "learning_rate": 5.638948676519043e-06, "loss": 0.8735, "step": 559 }, { "epoch": 0.9747606614447345, "grad_norm": 0.50390625, "learning_rate": 5.624666323937257e-06, "loss": 0.9788, "step": 560 }, { "epoch": 0.9765013054830287, "grad_norm": 0.49609375, "learning_rate": 5.610378791733821e-06, "loss": 0.9348, "step": 561 }, { "epoch": 0.9782419495213229, "grad_norm": 0.48046875, "learning_rate": 5.596086198378399e-06, "loss": 0.9258, "step": 562 }, { "epoch": 0.979982593559617, "grad_norm": 0.490234375, "learning_rate": 5.5817886623826245e-06, "loss": 0.9184, "step": 563 }, { "epoch": 0.9817232375979112, "grad_norm": 0.490234375, "learning_rate": 5.567486302299112e-06, "loss": 0.9439, "step": 564 }, { "epoch": 0.9834638816362054, "grad_norm": 0.474609375, "learning_rate": 5.553179236720482e-06, "loss": 0.9887, "step": 565 }, { "epoch": 0.9852045256744996, "grad_norm": 0.48828125, "learning_rate": 5.5388675842783644e-06, "loss": 0.9439, "step": 566 }, { "epoch": 0.9869451697127938, "grad_norm": 0.490234375, "learning_rate": 5.524551463642429e-06, "loss": 0.971, "step": 567 }, { "epoch": 0.9886858137510879, "grad_norm": 0.486328125, "learning_rate": 5.510230993519391e-06, "loss": 0.9583, "step": 568 }, { "epoch": 0.9904264577893821, "grad_norm": 0.50390625, "learning_rate": 5.495906292652035e-06, "loss": 0.9788, "step": 569 }, { "epoch": 0.9921671018276762, "grad_norm": 0.4921875, "learning_rate": 5.48157747981822e-06, "loss": 0.9634, "step": 570 }, { "epoch": 0.9939077458659704, "grad_norm": 0.5078125, "learning_rate": 5.467244673829908e-06, "loss": 0.8834, "step": 571 }, { "epoch": 0.9956483899042646, "grad_norm": 0.49609375, "learning_rate": 5.452907993532164e-06, "loss": 0.9113, "step": 572 }, { "epoch": 0.9973890339425587, "grad_norm": 0.50390625, "learning_rate": 5.438567557802186e-06, "loss": 0.8828, "step": 573 }, { "epoch": 0.999129677980853, "grad_norm": 0.48828125, "learning_rate": 5.424223485548303e-06, "loss": 0.9363, "step": 574 } ], "logging_steps": 1, "max_steps": 1148, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 574, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 3.613017336006574e+18, "train_batch_size": 8, "trial_name": null, "trial_params": null }