|
{ |
|
"best_global_step": null, |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.999129677980853, |
|
"eval_steps": 72, |
|
"global_step": 574, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0, |
|
"eval_loss": 1.136976718902588, |
|
"eval_runtime": 64.1168, |
|
"eval_samples_per_second": 73.117, |
|
"eval_steps_per_second": 9.14, |
|
"step": 0 |
|
}, |
|
{ |
|
"epoch": 0.0017406440382941688, |
|
"grad_norm": 1.859375, |
|
"learning_rate": 0.0, |
|
"loss": 1.1426, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0034812880765883376, |
|
"grad_norm": 1.9765625, |
|
"learning_rate": 1.7543859649122808e-07, |
|
"loss": 1.1274, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.005221932114882507, |
|
"grad_norm": 1.828125, |
|
"learning_rate": 3.5087719298245616e-07, |
|
"loss": 1.1205, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.006962576153176675, |
|
"grad_norm": 1.8515625, |
|
"learning_rate": 5.263157894736843e-07, |
|
"loss": 1.1383, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.008703220191470844, |
|
"grad_norm": 1.796875, |
|
"learning_rate": 7.017543859649123e-07, |
|
"loss": 1.14, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.010443864229765013, |
|
"grad_norm": 1.734375, |
|
"learning_rate": 8.771929824561404e-07, |
|
"loss": 1.1031, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.012184508268059183, |
|
"grad_norm": 2.015625, |
|
"learning_rate": 1.0526315789473685e-06, |
|
"loss": 1.1794, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.01392515230635335, |
|
"grad_norm": 1.765625, |
|
"learning_rate": 1.2280701754385965e-06, |
|
"loss": 1.1766, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.015665796344647518, |
|
"grad_norm": 1.8046875, |
|
"learning_rate": 1.4035087719298246e-06, |
|
"loss": 1.1251, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.017406440382941687, |
|
"grad_norm": 1.6796875, |
|
"learning_rate": 1.5789473684210526e-06, |
|
"loss": 1.1622, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.019147084421235857, |
|
"grad_norm": 1.8671875, |
|
"learning_rate": 1.7543859649122807e-06, |
|
"loss": 1.0846, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.020887728459530026, |
|
"grad_norm": 1.765625, |
|
"learning_rate": 1.929824561403509e-06, |
|
"loss": 1.1644, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.022628372497824196, |
|
"grad_norm": 1.7265625, |
|
"learning_rate": 2.105263157894737e-06, |
|
"loss": 1.1361, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.024369016536118365, |
|
"grad_norm": 1.8671875, |
|
"learning_rate": 2.280701754385965e-06, |
|
"loss": 1.1783, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.02610966057441253, |
|
"grad_norm": 1.828125, |
|
"learning_rate": 2.456140350877193e-06, |
|
"loss": 1.127, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.0278503046127067, |
|
"grad_norm": 1.640625, |
|
"learning_rate": 2.631578947368421e-06, |
|
"loss": 1.1587, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.02959094865100087, |
|
"grad_norm": 1.671875, |
|
"learning_rate": 2.8070175438596493e-06, |
|
"loss": 1.1088, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.031331592689295036, |
|
"grad_norm": 1.59375, |
|
"learning_rate": 2.9824561403508774e-06, |
|
"loss": 1.174, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.03307223672758921, |
|
"grad_norm": 1.71875, |
|
"learning_rate": 3.157894736842105e-06, |
|
"loss": 1.1453, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.034812880765883375, |
|
"grad_norm": 1.8515625, |
|
"learning_rate": 3.3333333333333333e-06, |
|
"loss": 1.1922, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.03655352480417755, |
|
"grad_norm": 1.5390625, |
|
"learning_rate": 3.5087719298245615e-06, |
|
"loss": 1.1541, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.038294168842471714, |
|
"grad_norm": 1.5390625, |
|
"learning_rate": 3.6842105263157896e-06, |
|
"loss": 1.0909, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.04003481288076589, |
|
"grad_norm": 1.515625, |
|
"learning_rate": 3.859649122807018e-06, |
|
"loss": 1.1498, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.04177545691906005, |
|
"grad_norm": 1.5703125, |
|
"learning_rate": 4.035087719298246e-06, |
|
"loss": 1.1096, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.04351610095735422, |
|
"grad_norm": 1.453125, |
|
"learning_rate": 4.210526315789474e-06, |
|
"loss": 1.1152, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.04525674499564839, |
|
"grad_norm": 1.390625, |
|
"learning_rate": 4.385964912280702e-06, |
|
"loss": 1.1111, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.04699738903394256, |
|
"grad_norm": 1.296875, |
|
"learning_rate": 4.56140350877193e-06, |
|
"loss": 1.161, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.04873803307223673, |
|
"grad_norm": 1.390625, |
|
"learning_rate": 4.736842105263158e-06, |
|
"loss": 1.1557, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.050478677110530897, |
|
"grad_norm": 1.3125, |
|
"learning_rate": 4.912280701754386e-06, |
|
"loss": 1.1065, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.05221932114882506, |
|
"grad_norm": 1.28125, |
|
"learning_rate": 5.087719298245615e-06, |
|
"loss": 1.0493, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.053959965187119235, |
|
"grad_norm": 1.1953125, |
|
"learning_rate": 5.263157894736842e-06, |
|
"loss": 1.0652, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.0557006092254134, |
|
"grad_norm": 1.1484375, |
|
"learning_rate": 5.438596491228071e-06, |
|
"loss": 1.0389, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.057441253263707574, |
|
"grad_norm": 1.1015625, |
|
"learning_rate": 5.6140350877192985e-06, |
|
"loss": 1.0349, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.05918189730200174, |
|
"grad_norm": 1.0390625, |
|
"learning_rate": 5.789473684210527e-06, |
|
"loss": 1.1109, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.060922541340295906, |
|
"grad_norm": 1.03125, |
|
"learning_rate": 5.964912280701755e-06, |
|
"loss": 1.0541, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.06266318537859007, |
|
"grad_norm": 0.9453125, |
|
"learning_rate": 6.140350877192983e-06, |
|
"loss": 1.1176, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.06440382941688425, |
|
"grad_norm": 0.9140625, |
|
"learning_rate": 6.31578947368421e-06, |
|
"loss": 1.0774, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.06614447345517842, |
|
"grad_norm": 0.9453125, |
|
"learning_rate": 6.491228070175439e-06, |
|
"loss": 1.031, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.06788511749347259, |
|
"grad_norm": 0.85546875, |
|
"learning_rate": 6.666666666666667e-06, |
|
"loss": 1.0977, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.06962576153176675, |
|
"grad_norm": 0.8203125, |
|
"learning_rate": 6.842105263157896e-06, |
|
"loss": 1.0871, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.07136640557006092, |
|
"grad_norm": 0.83984375, |
|
"learning_rate": 7.017543859649123e-06, |
|
"loss": 1.0856, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.0731070496083551, |
|
"grad_norm": 0.82421875, |
|
"learning_rate": 7.192982456140352e-06, |
|
"loss": 1.0189, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.07484769364664925, |
|
"grad_norm": 0.76953125, |
|
"learning_rate": 7.368421052631579e-06, |
|
"loss": 1.0633, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.07658833768494343, |
|
"grad_norm": 0.7578125, |
|
"learning_rate": 7.5438596491228074e-06, |
|
"loss": 1.0582, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.0783289817232376, |
|
"grad_norm": 0.78125, |
|
"learning_rate": 7.719298245614036e-06, |
|
"loss": 1.0261, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.08006962576153177, |
|
"grad_norm": 0.66796875, |
|
"learning_rate": 7.894736842105265e-06, |
|
"loss": 0.9962, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.08181026979982593, |
|
"grad_norm": 0.6953125, |
|
"learning_rate": 8.070175438596492e-06, |
|
"loss": 1.0041, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.0835509138381201, |
|
"grad_norm": 0.6640625, |
|
"learning_rate": 8.24561403508772e-06, |
|
"loss": 1.0953, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.08529155787641428, |
|
"grad_norm": 0.671875, |
|
"learning_rate": 8.421052631578948e-06, |
|
"loss": 1.0397, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.08703220191470844, |
|
"grad_norm": 0.66015625, |
|
"learning_rate": 8.596491228070176e-06, |
|
"loss": 1.0366, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.08877284595300261, |
|
"grad_norm": 0.65625, |
|
"learning_rate": 8.771929824561405e-06, |
|
"loss": 1.005, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.09051348999129678, |
|
"grad_norm": 0.62109375, |
|
"learning_rate": 8.947368421052632e-06, |
|
"loss": 1.1106, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.09225413402959094, |
|
"grad_norm": 0.59375, |
|
"learning_rate": 9.12280701754386e-06, |
|
"loss": 1.002, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.09399477806788512, |
|
"grad_norm": 0.5859375, |
|
"learning_rate": 9.298245614035088e-06, |
|
"loss": 1.0599, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.09573542210617929, |
|
"grad_norm": 0.59765625, |
|
"learning_rate": 9.473684210526315e-06, |
|
"loss": 1.0736, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.09747606614447346, |
|
"grad_norm": 0.56640625, |
|
"learning_rate": 9.649122807017545e-06, |
|
"loss": 1.0179, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.09921671018276762, |
|
"grad_norm": 0.58203125, |
|
"learning_rate": 9.824561403508772e-06, |
|
"loss": 1.0308, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.10095735422106179, |
|
"grad_norm": 0.5625, |
|
"learning_rate": 1e-05, |
|
"loss": 1.0135, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.10269799825935597, |
|
"grad_norm": 0.5859375, |
|
"learning_rate": 9.999979270446263e-06, |
|
"loss": 0.9598, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.10443864229765012, |
|
"grad_norm": 0.5390625, |
|
"learning_rate": 9.999917081956933e-06, |
|
"loss": 1.0041, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.1061792863359443, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 9.999813435047668e-06, |
|
"loss": 1.0726, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.10791993037423847, |
|
"grad_norm": 0.55078125, |
|
"learning_rate": 9.99966833057789e-06, |
|
"loss": 0.9915, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.10966057441253264, |
|
"grad_norm": 0.55078125, |
|
"learning_rate": 9.999481769750779e-06, |
|
"loss": 1.0266, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.1114012184508268, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 9.999253754113263e-06, |
|
"loss": 1.0524, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.11314186248912098, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 9.998984285556008e-06, |
|
"loss": 1.0254, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.11488250652741515, |
|
"grad_norm": 0.53515625, |
|
"learning_rate": 9.998673366313399e-06, |
|
"loss": 1.0132, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.11662315056570931, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 9.998320998963523e-06, |
|
"loss": 0.9604, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.11836379460400348, |
|
"grad_norm": 0.54296875, |
|
"learning_rate": 9.997927186428145e-06, |
|
"loss": 1.0147, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.12010443864229765, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 9.997491931972694e-06, |
|
"loss": 0.998, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.12184508268059181, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 9.997015239206216e-06, |
|
"loss": 1.034, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.12358572671888599, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 9.996497112081365e-06, |
|
"loss": 1.0103, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.12532637075718014, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 9.99593755489436e-06, |
|
"loss": 1.0053, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.12532637075718014, |
|
"eval_loss": 0.9893413186073303, |
|
"eval_runtime": 60.3364, |
|
"eval_samples_per_second": 77.698, |
|
"eval_steps_per_second": 9.712, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.12706701479547433, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 9.995336572284945e-06, |
|
"loss": 0.9841, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.1288076588337685, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 9.994694169236366e-06, |
|
"loss": 0.9298, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.13054830287206268, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 9.99401035107531e-06, |
|
"loss": 1.0449, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.13228894691035684, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 9.993285123471878e-06, |
|
"loss": 1.0139, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.134029590948651, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 9.992518492439526e-06, |
|
"loss": 1.0129, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.13577023498694518, |
|
"grad_norm": 0.54296875, |
|
"learning_rate": 9.991710464335022e-06, |
|
"loss": 0.9696, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.13751087902523934, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 9.990861045858392e-06, |
|
"loss": 1.0114, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.1392515230635335, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 9.989970244052861e-06, |
|
"loss": 0.957, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.1409921671018277, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 9.9890380663048e-06, |
|
"loss": 0.9603, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.14273281114012185, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 9.98806452034366e-06, |
|
"loss": 1.0523, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.144473455178416, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 9.987049614241907e-06, |
|
"loss": 0.9804, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.1462140992167102, |
|
"grad_norm": 0.54296875, |
|
"learning_rate": 9.985993356414965e-06, |
|
"loss": 1.0307, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.14795474325500435, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 9.984895755621136e-06, |
|
"loss": 0.9802, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.1496953872932985, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 9.983756820961528e-06, |
|
"loss": 0.9547, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.1514360313315927, |
|
"grad_norm": 0.5, |
|
"learning_rate": 9.982576561879984e-06, |
|
"loss": 1.014, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.15317667536988686, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 9.981354988163002e-06, |
|
"loss": 0.9708, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.15491731940818101, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 9.980092109939651e-06, |
|
"loss": 0.9717, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.1566579634464752, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 9.978787937681496e-06, |
|
"loss": 0.9862, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.15839860748476936, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 9.977442482202498e-06, |
|
"loss": 0.9813, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.16013925152306355, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 9.976055754658935e-06, |
|
"loss": 0.9551, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.1618798955613577, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 9.974627766549301e-06, |
|
"loss": 0.9777, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.16362053959965187, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 9.973158529714224e-06, |
|
"loss": 0.9648, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.16536118363794605, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 9.971648056336349e-06, |
|
"loss": 0.964, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.1671018276762402, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 9.97009635894025e-06, |
|
"loss": 0.9635, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.16884247171453437, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 9.968503450392332e-06, |
|
"loss": 1.0115, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.17058311575282856, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 9.966869343900702e-06, |
|
"loss": 0.9841, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.17232375979112272, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 9.965194053015083e-06, |
|
"loss": 1.0241, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.17406440382941687, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 9.963477591626687e-06, |
|
"loss": 0.968, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.17580504786771106, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 9.961719973968102e-06, |
|
"loss": 0.946, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.17754569190600522, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 9.959921214613187e-06, |
|
"loss": 0.9284, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.17928633594429938, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 9.958081328476926e-06, |
|
"loss": 0.958, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.18102697998259357, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 9.956200330815329e-06, |
|
"loss": 0.9361, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.18276762402088773, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 9.954278237225296e-06, |
|
"loss": 1.0116, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.18450826805918188, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 9.952315063644479e-06, |
|
"loss": 1.0041, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.18624891209747607, |
|
"grad_norm": 0.5, |
|
"learning_rate": 9.950310826351168e-06, |
|
"loss": 0.9725, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.18798955613577023, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 9.948265541964136e-06, |
|
"loss": 0.9763, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.18973020017406442, |
|
"grad_norm": 0.5, |
|
"learning_rate": 9.946179227442521e-06, |
|
"loss": 0.9814, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.19147084421235858, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 9.944051900085668e-06, |
|
"loss": 1.0092, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.19321148825065274, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 9.941883577532993e-06, |
|
"loss": 0.9836, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.19495213228894692, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 9.939674277763845e-06, |
|
"loss": 0.9598, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.19669277632724108, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 9.937424019097337e-06, |
|
"loss": 0.988, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.19843342036553524, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 9.935132820192218e-06, |
|
"loss": 0.9512, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.20017406440382943, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 9.932800700046697e-06, |
|
"loss": 0.9916, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.20191470844212359, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 9.9304276779983e-06, |
|
"loss": 0.9397, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.20365535248041775, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 9.9280137737237e-06, |
|
"loss": 0.9791, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.20539599651871193, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 9.925559007238564e-06, |
|
"loss": 0.9231, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.2071366405570061, |
|
"grad_norm": 0.5, |
|
"learning_rate": 9.923063398897372e-06, |
|
"loss": 0.9854, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.20887728459530025, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 9.920526969393267e-06, |
|
"loss": 1.0411, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.21061792863359444, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 9.917949739757869e-06, |
|
"loss": 1.0254, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.2123585726718886, |
|
"grad_norm": 0.5, |
|
"learning_rate": 9.915331731361104e-06, |
|
"loss": 0.9208, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.21409921671018275, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 9.912672965911034e-06, |
|
"loss": 0.9195, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.21583986074847694, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 9.909973465453666e-06, |
|
"loss": 0.9938, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.2175805047867711, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 9.907233252372775e-06, |
|
"loss": 0.904, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.2193211488250653, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 9.904452349389717e-06, |
|
"loss": 0.9882, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.22106179286335945, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 9.901630779563247e-06, |
|
"loss": 0.9688, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.2228024369016536, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 9.898768566289316e-06, |
|
"loss": 1.0522, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.2245430809399478, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 9.895865733300887e-06, |
|
"loss": 0.9551, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.22628372497824195, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 9.89292230466773e-06, |
|
"loss": 0.9623, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.2280243690165361, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 9.889938304796236e-06, |
|
"loss": 0.9702, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.2297650130548303, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 9.886913758429194e-06, |
|
"loss": 0.9679, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.23150565709312446, |
|
"grad_norm": 0.466796875, |
|
"learning_rate": 9.883848690645601e-06, |
|
"loss": 0.9719, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.23324630113141862, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 9.880743126860458e-06, |
|
"loss": 0.9717, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.2349869451697128, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 9.87759709282454e-06, |
|
"loss": 0.9365, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.23672758920800696, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 9.874410614624202e-06, |
|
"loss": 0.9254, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.23846823324630112, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 9.871183718681153e-06, |
|
"loss": 1.0045, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.2402088772845953, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 9.867916431752237e-06, |
|
"loss": 0.9693, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.24194952132288947, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 9.864608780929218e-06, |
|
"loss": 0.9981, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.24369016536118362, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 9.861260793638539e-06, |
|
"loss": 0.9569, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.2454308093994778, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 9.857872497641117e-06, |
|
"loss": 1.0422, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.24717145343777197, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 9.854443921032098e-06, |
|
"loss": 1.0408, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.24891209747606616, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 9.850975092240625e-06, |
|
"loss": 0.8893, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.2506527415143603, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 9.84746604002961e-06, |
|
"loss": 0.9679, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.2506527415143603, |
|
"eval_loss": 0.9576423168182373, |
|
"eval_runtime": 59.1215, |
|
"eval_samples_per_second": 79.294, |
|
"eval_steps_per_second": 9.912, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.2523933855526545, |
|
"grad_norm": 0.5, |
|
"learning_rate": 9.843916793495487e-06, |
|
"loss": 0.9071, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.25413402959094866, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 9.840327382067972e-06, |
|
"loss": 0.9496, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.2558746736292428, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 9.836697835509827e-06, |
|
"loss": 0.9864, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.257615317667537, |
|
"grad_norm": 0.5, |
|
"learning_rate": 9.833028183916601e-06, |
|
"loss": 1.0082, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.25935596170583114, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 9.829318457716395e-06, |
|
"loss": 0.9591, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.26109660574412535, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 9.82556868766959e-06, |
|
"loss": 1.0204, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.2628372497824195, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 9.821778904868616e-06, |
|
"loss": 0.9805, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.26457789382071367, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 9.817949140737672e-06, |
|
"loss": 0.9761, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.26631853785900783, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 9.81407942703248e-06, |
|
"loss": 0.9789, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.268059181897302, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 9.810169795840012e-06, |
|
"loss": 0.952, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.26979982593559615, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 9.806220279578236e-06, |
|
"loss": 0.9431, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.27154046997389036, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 9.802230910995833e-06, |
|
"loss": 1.0015, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.2732811140121845, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 9.798201723171938e-06, |
|
"loss": 0.9513, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.2750217580504787, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 9.794132749515854e-06, |
|
"loss": 0.9454, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.27676240208877284, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 9.790024023766789e-06, |
|
"loss": 0.9581, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.278503046127067, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 9.785875579993558e-06, |
|
"loss": 0.9874, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.28024369016536116, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 9.781687452594318e-06, |
|
"loss": 0.9417, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.2819843342036554, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 9.777459676296276e-06, |
|
"loss": 0.9589, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.28372497824194953, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 9.773192286155395e-06, |
|
"loss": 0.9851, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.2854656222802437, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 9.768885317556116e-06, |
|
"loss": 0.98, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.28720626631853785, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 9.764538806211052e-06, |
|
"loss": 0.9651, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.288946910356832, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 9.760152788160697e-06, |
|
"loss": 0.9407, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.2906875543951262, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 9.755727299773135e-06, |
|
"loss": 0.9553, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.2924281984334204, |
|
"grad_norm": 0.470703125, |
|
"learning_rate": 9.75126237774372e-06, |
|
"loss": 0.9823, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.29416884247171454, |
|
"grad_norm": 0.5, |
|
"learning_rate": 9.746758059094791e-06, |
|
"loss": 0.9832, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.2959094865100087, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 9.742214381175355e-06, |
|
"loss": 0.8976, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.29765013054830286, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 9.737631381660777e-06, |
|
"loss": 0.9331, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.299390774586597, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 9.733009098552473e-06, |
|
"loss": 0.9666, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.30113141862489123, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 9.728347570177587e-06, |
|
"loss": 0.9781, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.3028720626631854, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 9.723646835188681e-06, |
|
"loss": 0.969, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.30461270670147955, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 9.71890693256341e-06, |
|
"loss": 0.9481, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.3063533507397737, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 9.7141279016042e-06, |
|
"loss": 1.0301, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.30809399477806787, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 9.709309781937925e-06, |
|
"loss": 1.023, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.30983463881636203, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 9.704452613515571e-06, |
|
"loss": 0.9598, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.31157528285465624, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 9.699556436611912e-06, |
|
"loss": 0.9699, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.3133159268929504, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 9.694621291825174e-06, |
|
"loss": 0.903, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.31505657093124456, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 9.689647220076696e-06, |
|
"loss": 0.9581, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.3167972149695387, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 9.684634262610593e-06, |
|
"loss": 0.939, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.3185378590078329, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 9.679582460993413e-06, |
|
"loss": 0.9363, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.3202785030461271, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 9.674491857113792e-06, |
|
"loss": 1.0215, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.32201914708442125, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 9.669362493182112e-06, |
|
"loss": 0.9464, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.3237597911227154, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 9.66419441173014e-06, |
|
"loss": 0.8955, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.32550043516100957, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 9.658987655610687e-06, |
|
"loss": 0.9503, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.32724107919930373, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 9.653742267997245e-06, |
|
"loss": 0.9808, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.3289817232375979, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 9.648458292383631e-06, |
|
"loss": 1.0155, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.3307223672758921, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 9.643135772583627e-06, |
|
"loss": 0.929, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.33246301131418626, |
|
"grad_norm": 0.5, |
|
"learning_rate": 9.63777475273062e-06, |
|
"loss": 0.9214, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.3342036553524804, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 9.632375277277226e-06, |
|
"loss": 0.9762, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.3359442993907746, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 9.626937390994932e-06, |
|
"loss": 0.9734, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.33768494342906874, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 9.621461138973725e-06, |
|
"loss": 0.9599, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.3394255874673629, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 9.615946566621704e-06, |
|
"loss": 0.9448, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.3411662315056571, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 9.61039371966472e-06, |
|
"loss": 0.9721, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.3429068755439513, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 9.60480264414599e-06, |
|
"loss": 0.97, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.34464751958224543, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 9.599173386425711e-06, |
|
"loss": 0.9964, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.3463881636205396, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 9.593505993180687e-06, |
|
"loss": 0.9715, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.34812880765883375, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 9.587800511403931e-06, |
|
"loss": 0.9655, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.34986945169712796, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 9.582056988404276e-06, |
|
"loss": 0.9606, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.3516100957354221, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 9.576275471805993e-06, |
|
"loss": 0.9166, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.3533507397737163, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 9.570456009548383e-06, |
|
"loss": 0.9063, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.35509138381201044, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 9.564598649885391e-06, |
|
"loss": 0.9295, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.3568320278503046, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 9.558703441385195e-06, |
|
"loss": 0.9933, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.35857267188859876, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 9.552770432929812e-06, |
|
"loss": 0.9572, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.360313315926893, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 9.54679967371469e-06, |
|
"loss": 0.9484, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.36205395996518713, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 9.540791213248299e-06, |
|
"loss": 0.9266, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.3637946040034813, |
|
"grad_norm": 0.5, |
|
"learning_rate": 9.534745101351719e-06, |
|
"loss": 0.9526, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.36553524804177545, |
|
"grad_norm": 0.5, |
|
"learning_rate": 9.528661388158234e-06, |
|
"loss": 1.0046, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.3672758920800696, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 9.522540124112902e-06, |
|
"loss": 0.9757, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.36901653611836377, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 9.516381359972157e-06, |
|
"loss": 1.0195, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.370757180156658, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 9.51018514680337e-06, |
|
"loss": 0.971, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.37249782419495214, |
|
"grad_norm": 0.5, |
|
"learning_rate": 9.503951535984434e-06, |
|
"loss": 0.9336, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.3742384682332463, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 9.49768057920334e-06, |
|
"loss": 0.9887, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.37597911227154046, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 9.491372328457737e-06, |
|
"loss": 0.966, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.37597911227154046, |
|
"eval_loss": 0.9439952373504639, |
|
"eval_runtime": 59.4647, |
|
"eval_samples_per_second": 78.837, |
|
"eval_steps_per_second": 9.855, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.3777197563098346, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 9.485026836054519e-06, |
|
"loss": 0.9771, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.37946040034812883, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 9.478644154609372e-06, |
|
"loss": 0.9988, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.381201044386423, |
|
"grad_norm": 0.470703125, |
|
"learning_rate": 9.472224337046357e-06, |
|
"loss": 0.9875, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.38294168842471715, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 9.46576743659745e-06, |
|
"loss": 0.9664, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.3846823324630113, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 9.45927350680212e-06, |
|
"loss": 0.9693, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.38642297650130547, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 9.452742601506873e-06, |
|
"loss": 0.9137, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.38816362053959963, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 9.446174774864808e-06, |
|
"loss": 0.9047, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.38990426457789384, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 9.439570081335173e-06, |
|
"loss": 0.9926, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.391644908616188, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 9.432928575682908e-06, |
|
"loss": 0.9693, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.39338555265448216, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 9.426250312978191e-06, |
|
"loss": 0.9106, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.3951261966927763, |
|
"grad_norm": 0.5, |
|
"learning_rate": 9.419535348595985e-06, |
|
"loss": 0.9639, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.3968668407310705, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 9.412783738215576e-06, |
|
"loss": 0.9338, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.39860748476936464, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 9.405995537820111e-06, |
|
"loss": 1.0216, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.40034812880765885, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 9.399170803696139e-06, |
|
"loss": 0.942, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.402088772845953, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 9.392309592433134e-06, |
|
"loss": 0.9184, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.40382941688424717, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 9.385411960923036e-06, |
|
"loss": 0.9178, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.40557006092254133, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 9.378477966359773e-06, |
|
"loss": 0.9303, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.4073107049608355, |
|
"grad_norm": 0.5, |
|
"learning_rate": 9.371507666238793e-06, |
|
"loss": 0.9563, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.4090513489991297, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 9.364501118356579e-06, |
|
"loss": 0.92, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.41079199303742386, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 9.357458380810175e-06, |
|
"loss": 0.9532, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.412532637075718, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 9.350379511996706e-06, |
|
"loss": 0.9604, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.4142732811140122, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 9.343264570612883e-06, |
|
"loss": 0.9415, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.41601392515230634, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 9.336113615654535e-06, |
|
"loss": 0.9752, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.4177545691906005, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 9.328926706416102e-06, |
|
"loss": 0.9517, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.4194952132288947, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 9.321703902490152e-06, |
|
"loss": 0.9245, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.4212358572671889, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 9.314445263766888e-06, |
|
"loss": 0.9341, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.42297650130548303, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 9.307150850433643e-06, |
|
"loss": 0.9399, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.4247171453437772, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 9.299820722974396e-06, |
|
"loss": 0.9865, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.42645778938207135, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 9.29245494216925e-06, |
|
"loss": 0.9538, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.4281984334203655, |
|
"grad_norm": 0.5, |
|
"learning_rate": 9.285053569093948e-06, |
|
"loss": 1.0095, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.4299390774586597, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 9.277616665119352e-06, |
|
"loss": 0.9691, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.4316797214969539, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 9.27014429191094e-06, |
|
"loss": 0.9854, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.43342036553524804, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 9.262636511428304e-06, |
|
"loss": 0.9179, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.4351610095735422, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 9.255093385924616e-06, |
|
"loss": 0.9388, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.43690165361183636, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 9.247514977946124e-06, |
|
"loss": 0.9788, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 0.4386422976501306, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 9.239901350331635e-06, |
|
"loss": 0.9301, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.44038294168842473, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 9.232252566211993e-06, |
|
"loss": 0.9656, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.4421235857267189, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 9.224568689009548e-06, |
|
"loss": 1.0119, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.44386422976501305, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 9.216849782437637e-06, |
|
"loss": 0.874, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.4456048738033072, |
|
"grad_norm": 0.5, |
|
"learning_rate": 9.20909591050006e-06, |
|
"loss": 0.9191, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.44734551784160137, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 9.201307137490536e-06, |
|
"loss": 0.9017, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 0.4490861618798956, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 9.19348352799218e-06, |
|
"loss": 0.9363, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.45082680591818974, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 9.185625146876966e-06, |
|
"loss": 0.9921, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 0.4525674499564839, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 9.177732059305187e-06, |
|
"loss": 0.9358, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.45430809399477806, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 9.169804330724916e-06, |
|
"loss": 0.9257, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.4560487380330722, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 9.161842026871465e-06, |
|
"loss": 0.9201, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.4577893820713664, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 9.153845213766837e-06, |
|
"loss": 0.9212, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 0.4595300261096606, |
|
"grad_norm": 0.5, |
|
"learning_rate": 9.145813957719174e-06, |
|
"loss": 0.9735, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.46127067014795475, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 9.137748325322223e-06, |
|
"loss": 0.9585, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.4630113141862489, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 9.129648383454764e-06, |
|
"loss": 0.9781, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.46475195822454307, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 9.121514199280072e-06, |
|
"loss": 0.9759, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.46649260226283723, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 9.113345840245348e-06, |
|
"loss": 0.9688, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.46823324630113144, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 9.105143374081167e-06, |
|
"loss": 0.9092, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 0.4699738903394256, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 9.096906868800917e-06, |
|
"loss": 0.9357, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.47171453437771976, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 9.088636392700227e-06, |
|
"loss": 1.0134, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 0.4734551784160139, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 9.08033201435641e-06, |
|
"loss": 0.9494, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.4751958224543081, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 9.071993802627887e-06, |
|
"loss": 0.9446, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.47693646649260224, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 9.063621826653624e-06, |
|
"loss": 0.8926, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 0.47867711053089645, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 9.055216155852548e-06, |
|
"loss": 0.9216, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.4804177545691906, |
|
"grad_norm": 0.470703125, |
|
"learning_rate": 9.046776859922983e-06, |
|
"loss": 0.9442, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.4821583986074848, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 9.038304008842064e-06, |
|
"loss": 0.9456, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 0.48389904264577893, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 9.02979767286516e-06, |
|
"loss": 0.909, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 0.4856396866840731, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 9.021257922525289e-06, |
|
"loss": 0.9597, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.48738033072236725, |
|
"grad_norm": 0.5, |
|
"learning_rate": 9.012684828632538e-06, |
|
"loss": 0.9646, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.48912097476066146, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 9.004078462273471e-06, |
|
"loss": 0.9679, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 0.4908616187989556, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 8.995438894810541e-06, |
|
"loss": 0.952, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.4926022628372498, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 8.9867661978815e-06, |
|
"loss": 0.9792, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 0.49434290687554394, |
|
"grad_norm": 0.5, |
|
"learning_rate": 8.978060443398802e-06, |
|
"loss": 0.939, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 0.4960835509138381, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 8.96932170354901e-06, |
|
"loss": 0.9974, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.4978241949521323, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 8.960550050792194e-06, |
|
"loss": 0.9947, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 0.4995648389904265, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 8.951745557861333e-06, |
|
"loss": 0.9139, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 0.5013054830287206, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 8.942908297761712e-06, |
|
"loss": 0.9397, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.5013054830287206, |
|
"eval_loss": 0.9357889294624329, |
|
"eval_runtime": 60.8471, |
|
"eval_samples_per_second": 77.046, |
|
"eval_steps_per_second": 9.631, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.5030461270670148, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 8.934038343770312e-06, |
|
"loss": 0.9656, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 0.504786771105309, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 8.925135769435211e-06, |
|
"loss": 0.9896, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.5065274151436031, |
|
"grad_norm": 0.5, |
|
"learning_rate": 8.916200648574964e-06, |
|
"loss": 0.905, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 0.5082680591818973, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 8.907233055277999e-06, |
|
"loss": 0.9309, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 0.5100087032201914, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 8.898233063902e-06, |
|
"loss": 0.9796, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 0.5117493472584856, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 8.889200749073285e-06, |
|
"loss": 0.9335, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.5134899912967799, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 8.880136185686202e-06, |
|
"loss": 0.9292, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.515230635335074, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 8.871039448902488e-06, |
|
"loss": 0.9116, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 0.5169712793733682, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 8.861910614150662e-06, |
|
"loss": 0.9315, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 0.5187119234116623, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 8.852749757125392e-06, |
|
"loss": 0.9283, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 0.5204525674499565, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 8.843556953786872e-06, |
|
"loss": 0.952, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 0.5221932114882507, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 8.834332280360181e-06, |
|
"loss": 0.9999, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.5239338555265448, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 8.82507581333467e-06, |
|
"loss": 0.9453, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 0.525674499564839, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 8.815787629463306e-06, |
|
"loss": 0.8678, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 0.5274151436031331, |
|
"grad_norm": 0.5, |
|
"learning_rate": 8.806467805762056e-06, |
|
"loss": 0.9878, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 0.5291557876414273, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 8.797116419509232e-06, |
|
"loss": 0.8964, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 0.5308964316797214, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 8.78773354824486e-06, |
|
"loss": 0.9584, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.5326370757180157, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 8.778319269770033e-06, |
|
"loss": 0.9715, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 0.5343777197563099, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 8.768873662146271e-06, |
|
"loss": 0.9034, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 0.536118363794604, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 8.759396803694863e-06, |
|
"loss": 0.9189, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 0.5378590078328982, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 8.749888772996226e-06, |
|
"loss": 1.0066, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 0.5395996518711923, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 8.74034964888926e-06, |
|
"loss": 1.0147, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.5413402959094865, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 8.730779510470672e-06, |
|
"loss": 0.9504, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 0.5430809399477807, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 8.721178437094346e-06, |
|
"loss": 0.9239, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 0.5448215839860748, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 8.711546508370666e-06, |
|
"loss": 0.9145, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 0.546562228024369, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 8.701883804165867e-06, |
|
"loss": 0.897, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 0.5483028720626631, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 8.692190404601368e-06, |
|
"loss": 0.8796, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.5500435161009574, |
|
"grad_norm": 0.53515625, |
|
"learning_rate": 8.682466390053106e-06, |
|
"loss": 1.03, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 0.5517841601392516, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 8.672711841150877e-06, |
|
"loss": 0.9676, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 0.5535248041775457, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 8.662926838777657e-06, |
|
"loss": 1.0098, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 0.5552654482158399, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 8.653111464068937e-06, |
|
"loss": 0.9344, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 0.557006092254134, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 8.643265798412057e-06, |
|
"loss": 0.9555, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.5587467362924282, |
|
"grad_norm": 0.5, |
|
"learning_rate": 8.633389923445515e-06, |
|
"loss": 0.8845, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 0.5604873803307223, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 8.623483921058304e-06, |
|
"loss": 0.9403, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 0.5622280243690165, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 8.613547873389228e-06, |
|
"loss": 0.9654, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 0.5639686684073107, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 8.603581862826222e-06, |
|
"loss": 1.0108, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 0.5657093124456049, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 8.593585972005665e-06, |
|
"loss": 0.9708, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.5674499564838991, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 8.5835602838117e-06, |
|
"loss": 0.9333, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 0.5691906005221932, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 8.573504881375543e-06, |
|
"loss": 0.9067, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 0.5709312445604874, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 8.563419848074798e-06, |
|
"loss": 0.9388, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 0.5726718885987816, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 8.55330526753276e-06, |
|
"loss": 0.9246, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 0.5744125326370757, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 8.543161223617724e-06, |
|
"loss": 0.9222, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.5761531766753699, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 8.532987800442292e-06, |
|
"loss": 1.0001, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 0.577893820713664, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 8.522785082362675e-06, |
|
"loss": 1.0003, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 0.5796344647519582, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 8.512553153977988e-06, |
|
"loss": 0.9103, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 0.5813751087902524, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 8.502292100129553e-06, |
|
"loss": 0.9895, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 0.5831157528285466, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 8.492002005900201e-06, |
|
"loss": 0.9981, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.5848563968668408, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 8.481682956613555e-06, |
|
"loss": 0.8758, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 0.5865970409051349, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 8.471335037833328e-06, |
|
"loss": 0.9845, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 0.5883376849434291, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 8.460958335362617e-06, |
|
"loss": 0.9974, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 0.5900783289817232, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 8.450552935243186e-06, |
|
"loss": 0.9725, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 0.5918189730200174, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 8.440118923754757e-06, |
|
"loss": 0.9362, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.5935596170583116, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 8.429656387414289e-06, |
|
"loss": 0.9659, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 0.5953002610966057, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 8.419165412975265e-06, |
|
"loss": 0.9346, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 0.5970409051348999, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 8.408646087426975e-06, |
|
"loss": 0.8759, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 0.598781549173194, |
|
"grad_norm": 0.5, |
|
"learning_rate": 8.398098497993785e-06, |
|
"loss": 0.9359, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 0.6005221932114883, |
|
"grad_norm": 0.5, |
|
"learning_rate": 8.387522732134428e-06, |
|
"loss": 0.9171, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.6022628372497825, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 8.376918877541263e-06, |
|
"loss": 0.919, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 0.6040034812880766, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 8.36628702213956e-06, |
|
"loss": 0.9211, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 0.6057441253263708, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 8.355627254086771e-06, |
|
"loss": 0.9191, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 0.6074847693646649, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 8.344939661771784e-06, |
|
"loss": 0.926, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 0.6092254134029591, |
|
"grad_norm": 0.5, |
|
"learning_rate": 8.334224333814209e-06, |
|
"loss": 0.9088, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.6109660574412533, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 8.323481359063631e-06, |
|
"loss": 0.9455, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 0.6127067014795474, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 8.312710826598884e-06, |
|
"loss": 0.8873, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 0.6144473455178416, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 8.301912825727294e-06, |
|
"loss": 0.9562, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 0.6161879895561357, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 8.29108744598396e-06, |
|
"loss": 0.9324, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 0.61792863359443, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 8.280234777131e-06, |
|
"loss": 0.9037, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.6196692776327241, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 8.269354909156803e-06, |
|
"loss": 0.9127, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 0.6214099216710183, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 8.258447932275296e-06, |
|
"loss": 0.9003, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 0.6231505657093125, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 8.247513936925182e-06, |
|
"loss": 0.9871, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 0.6248912097476066, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 8.236553013769198e-06, |
|
"loss": 0.8732, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 0.6266318537859008, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 8.225565253693365e-06, |
|
"loss": 0.9563, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.6266318537859008, |
|
"eval_loss": 0.9300395846366882, |
|
"eval_runtime": 59.3628, |
|
"eval_samples_per_second": 78.972, |
|
"eval_steps_per_second": 9.871, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.6283724978241949, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 8.214550747806227e-06, |
|
"loss": 0.9173, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 0.6301131418624891, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 8.2035095874381e-06, |
|
"loss": 0.9297, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 0.6318537859007833, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 8.192441864140314e-06, |
|
"loss": 0.9582, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 0.6335944299390774, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 8.181347669684456e-06, |
|
"loss": 0.9176, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 0.6353350739773717, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 8.170227096061607e-06, |
|
"loss": 0.9181, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.6370757180156658, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 8.15908023548158e-06, |
|
"loss": 0.9708, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 0.63881636205396, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 8.147907180372147e-06, |
|
"loss": 0.9683, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 0.6405570060922542, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 8.136708023378292e-06, |
|
"loss": 0.9728, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 0.6422976501305483, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 8.125482857361426e-06, |
|
"loss": 0.9696, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 0.6440382941688425, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 8.114231775398618e-06, |
|
"loss": 0.9189, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.6457789382071366, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 8.102954870781831e-06, |
|
"loss": 0.9361, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 0.6475195822454308, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 8.091652237017152e-06, |
|
"loss": 0.9612, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 0.6492602262837249, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 8.080323967823993e-06, |
|
"loss": 0.9442, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 0.6510008703220191, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 8.068970157134349e-06, |
|
"loss": 0.9346, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 0.6527415143603134, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 8.057590899091985e-06, |
|
"loss": 0.8999, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.6544821583986075, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 8.046186288051681e-06, |
|
"loss": 0.9674, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 0.6562228024369017, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 8.034756418578434e-06, |
|
"loss": 0.9364, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 0.6579634464751958, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 8.023301385446682e-06, |
|
"loss": 0.9341, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 0.65970409051349, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 8.011821283639515e-06, |
|
"loss": 0.8986, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 0.6614447345517842, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 8.000316208347891e-06, |
|
"loss": 0.9266, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.6631853785900783, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 7.988786254969837e-06, |
|
"loss": 0.8911, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 0.6649260226283725, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 7.977231519109665e-06, |
|
"loss": 0.9221, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 0.6666666666666666, |
|
"grad_norm": 0.466796875, |
|
"learning_rate": 7.965652096577188e-06, |
|
"loss": 0.9635, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 0.6684073107049608, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 7.954048083386909e-06, |
|
"loss": 0.9023, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 0.6701479547432551, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 7.942419575757235e-06, |
|
"loss": 0.9277, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.6718885987815492, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 7.930766670109675e-06, |
|
"loss": 0.9826, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 0.6736292428198434, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 7.919089463068038e-06, |
|
"loss": 0.9223, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 0.6753698868581375, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 7.907388051457647e-06, |
|
"loss": 1.0468, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 0.6771105308964317, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 7.895662532304516e-06, |
|
"loss": 0.9233, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 0.6788511749347258, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 7.883913002834555e-06, |
|
"loss": 0.9756, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.68059181897302, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 7.872139560472767e-06, |
|
"loss": 0.9, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 0.6823324630113142, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 7.860342302842434e-06, |
|
"loss": 0.894, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 0.6840731070496083, |
|
"grad_norm": 0.5, |
|
"learning_rate": 7.848521327764309e-06, |
|
"loss": 0.9369, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 0.6858137510879025, |
|
"grad_norm": 0.5, |
|
"learning_rate": 7.836676733255809e-06, |
|
"loss": 0.9416, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 0.6875543951261966, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 7.824808617530197e-06, |
|
"loss": 0.8992, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.6892950391644909, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 7.812917078995769e-06, |
|
"loss": 0.9457, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 0.6910356832027851, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 7.801002216255042e-06, |
|
"loss": 0.9484, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 0.6927763272410792, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 7.78906412810393e-06, |
|
"loss": 0.9297, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 0.6945169712793734, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 7.777102913530927e-06, |
|
"loss": 0.951, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 0.6962576153176675, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 7.76511867171629e-06, |
|
"loss": 0.9091, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.6979982593559617, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 7.753111502031214e-06, |
|
"loss": 0.9041, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 0.6997389033942559, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 7.741081504037009e-06, |
|
"loss": 0.9463, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 0.70147954743255, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 7.729028777484266e-06, |
|
"loss": 0.9494, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 0.7032201914708442, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 7.716953422312044e-06, |
|
"loss": 0.9266, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 0.7049608355091384, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 7.704855538647033e-06, |
|
"loss": 0.9731, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.7067014795474326, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 7.692735226802729e-06, |
|
"loss": 0.961, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 0.7084421235857267, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 7.680592587278585e-06, |
|
"loss": 0.9506, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 0.7101827676240209, |
|
"grad_norm": 0.5, |
|
"learning_rate": 7.668427720759207e-06, |
|
"loss": 0.9084, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 0.7119234116623151, |
|
"grad_norm": 0.5, |
|
"learning_rate": 7.656240728113493e-06, |
|
"loss": 0.9147, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 0.7136640557006092, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 7.644031710393815e-06, |
|
"loss": 0.9498, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.7154046997389034, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 7.631800768835167e-06, |
|
"loss": 0.9145, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 0.7171453437771975, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 7.619548004854332e-06, |
|
"loss": 0.8884, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 0.7188859878154917, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 7.607273520049041e-06, |
|
"loss": 0.9777, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 0.720626631853786, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 7.594977416197134e-06, |
|
"loss": 0.9802, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 0.72236727589208, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 7.582659795255707e-06, |
|
"loss": 0.9484, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.7241079199303743, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 7.570320759360273e-06, |
|
"loss": 0.9231, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 0.7258485639686684, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 7.557960410823917e-06, |
|
"loss": 0.9416, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 0.7275892080069626, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 7.545578852136443e-06, |
|
"loss": 1.0058, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 0.7293298520452568, |
|
"grad_norm": 0.5, |
|
"learning_rate": 7.533176185963523e-06, |
|
"loss": 0.9426, |
|
"step": 419 |
|
}, |
|
{ |
|
"epoch": 0.7310704960835509, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 7.520752515145855e-06, |
|
"loss": 0.949, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.7328111401218451, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 7.508307942698296e-06, |
|
"loss": 0.9727, |
|
"step": 421 |
|
}, |
|
{ |
|
"epoch": 0.7345517841601392, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 7.495842571809021e-06, |
|
"loss": 0.8763, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 0.7362924281984334, |
|
"grad_norm": 0.5, |
|
"learning_rate": 7.4833565058386595e-06, |
|
"loss": 0.8914, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 0.7380330722367275, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 7.470849848319443e-06, |
|
"loss": 0.9157, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 0.7397737162750218, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 7.458322702954342e-06, |
|
"loss": 0.8703, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.741514360313316, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 7.44577517361621e-06, |
|
"loss": 0.9106, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 0.7432550043516101, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 7.4332073643469196e-06, |
|
"loss": 0.8928, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 0.7449956483899043, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 7.420619379356504e-06, |
|
"loss": 0.9402, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 0.7467362924281984, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 7.408011323022286e-06, |
|
"loss": 0.8997, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 0.7484769364664926, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 7.395383299888019e-06, |
|
"loss": 0.9831, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.7502175805047868, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 7.382735414663017e-06, |
|
"loss": 1.04, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 0.7519582245430809, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 7.370067772221285e-06, |
|
"loss": 0.9034, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 0.7519582245430809, |
|
"eval_loss": 0.9259106516838074, |
|
"eval_runtime": 59.5681, |
|
"eval_samples_per_second": 78.7, |
|
"eval_steps_per_second": 9.837, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 0.7536988685813751, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 7.357380477600654e-06, |
|
"loss": 0.9201, |
|
"step": 433 |
|
}, |
|
{ |
|
"epoch": 0.7554395126196692, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 7.3446736360019065e-06, |
|
"loss": 0.9291, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 0.7571801566579635, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 7.331947352787905e-06, |
|
"loss": 0.951, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.7589208006962577, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 7.319201733482715e-06, |
|
"loss": 0.9208, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 0.7606614447345518, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 7.3064368837707425e-06, |
|
"loss": 0.9116, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 0.762402088772846, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 7.2936529094958365e-06, |
|
"loss": 0.9165, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 0.7641427328111401, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 7.280849916660434e-06, |
|
"loss": 0.9513, |
|
"step": 439 |
|
}, |
|
{ |
|
"epoch": 0.7658833768494343, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 7.268028011424664e-06, |
|
"loss": 0.9725, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.7676240208877284, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 7.255187300105477e-06, |
|
"loss": 0.9328, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 0.7693646649260226, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 7.24232788917576e-06, |
|
"loss": 0.8776, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 0.7711053089643168, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 7.229449885263451e-06, |
|
"loss": 0.9432, |
|
"step": 443 |
|
}, |
|
{ |
|
"epoch": 0.7728459530026109, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 7.21655339515066e-06, |
|
"loss": 0.8951, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 0.7745865970409052, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 7.203638525772783e-06, |
|
"loss": 0.961, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.7763272410791993, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 7.1907053842176075e-06, |
|
"loss": 0.9498, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 0.7780678851174935, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 7.17775407772444e-06, |
|
"loss": 0.9638, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 0.7798085291557877, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 7.164784713683197e-06, |
|
"loss": 0.9516, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 0.7815491731940818, |
|
"grad_norm": 0.466796875, |
|
"learning_rate": 7.1517973996335335e-06, |
|
"loss": 0.9108, |
|
"step": 449 |
|
}, |
|
{ |
|
"epoch": 0.783289817232376, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 7.138792243263936e-06, |
|
"loss": 0.9089, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.7850304612706701, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 7.125769352410845e-06, |
|
"loss": 0.9667, |
|
"step": 451 |
|
}, |
|
{ |
|
"epoch": 0.7867711053089643, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 7.112728835057742e-06, |
|
"loss": 0.9458, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 0.7885117493472585, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 7.099670799334269e-06, |
|
"loss": 0.899, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 0.7902523933855526, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 7.08659535351533e-06, |
|
"loss": 0.9031, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 0.7919930374238469, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 7.073502606020187e-06, |
|
"loss": 0.9771, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.793733681462141, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 7.060392665411564e-06, |
|
"loss": 0.9113, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 0.7954743255004352, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 7.0472656403947505e-06, |
|
"loss": 1.0172, |
|
"step": 457 |
|
}, |
|
{ |
|
"epoch": 0.7972149695387293, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 7.034121639816691e-06, |
|
"loss": 0.9282, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 0.7989556135770235, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 7.020960772665096e-06, |
|
"loss": 0.896, |
|
"step": 459 |
|
}, |
|
{ |
|
"epoch": 0.8006962576153177, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 7.007783148067524e-06, |
|
"loss": 0.881, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.8024369016536118, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 6.994588875290488e-06, |
|
"loss": 0.9155, |
|
"step": 461 |
|
}, |
|
{ |
|
"epoch": 0.804177545691906, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 6.9813780637385385e-06, |
|
"loss": 0.9012, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 0.8059181897302001, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 6.968150822953372e-06, |
|
"loss": 0.9085, |
|
"step": 463 |
|
}, |
|
{ |
|
"epoch": 0.8076588337684943, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 6.954907262612906e-06, |
|
"loss": 0.9818, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 0.8093994778067886, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 6.941647492530378e-06, |
|
"loss": 0.9717, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.8111401218450827, |
|
"grad_norm": 0.5, |
|
"learning_rate": 6.928371622653434e-06, |
|
"loss": 0.9369, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 0.8128807658833769, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 6.91507976306322e-06, |
|
"loss": 0.8943, |
|
"step": 467 |
|
}, |
|
{ |
|
"epoch": 0.814621409921671, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 6.901772023973459e-06, |
|
"loss": 0.9113, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 0.8163620539599652, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 6.888448515729552e-06, |
|
"loss": 0.8995, |
|
"step": 469 |
|
}, |
|
{ |
|
"epoch": 0.8181026979982594, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 6.8751093488076485e-06, |
|
"loss": 0.9159, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.8198433420365535, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 6.86175463381374e-06, |
|
"loss": 0.9388, |
|
"step": 471 |
|
}, |
|
{ |
|
"epoch": 0.8215839860748477, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 6.8483844814827405e-06, |
|
"loss": 0.9524, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 0.8233246301131418, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 6.8349990026775656e-06, |
|
"loss": 0.9458, |
|
"step": 473 |
|
}, |
|
{ |
|
"epoch": 0.825065274151436, |
|
"grad_norm": 0.5, |
|
"learning_rate": 6.821598308388217e-06, |
|
"loss": 0.9258, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 0.8268059181897301, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 6.8081825097308584e-06, |
|
"loss": 0.9159, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.8285465622280244, |
|
"grad_norm": 0.5, |
|
"learning_rate": 6.794751717946897e-06, |
|
"loss": 0.9029, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 0.8302872062663186, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 6.781306044402064e-06, |
|
"loss": 0.8852, |
|
"step": 477 |
|
}, |
|
{ |
|
"epoch": 0.8320278503046127, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 6.767845600585479e-06, |
|
"loss": 0.945, |
|
"step": 478 |
|
}, |
|
{ |
|
"epoch": 0.8337684943429069, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 6.754370498108747e-06, |
|
"loss": 0.9776, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 0.835509138381201, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 6.740880848705005e-06, |
|
"loss": 0.9229, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.8372497824194952, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 6.72737676422802e-06, |
|
"loss": 0.9686, |
|
"step": 481 |
|
}, |
|
{ |
|
"epoch": 0.8389904264577894, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 6.713858356651253e-06, |
|
"loss": 0.9459, |
|
"step": 482 |
|
}, |
|
{ |
|
"epoch": 0.8407310704960835, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 6.700325738066923e-06, |
|
"loss": 0.9472, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 0.8424717145343777, |
|
"grad_norm": 0.5, |
|
"learning_rate": 6.686779020685089e-06, |
|
"loss": 0.9417, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 0.8442123585726719, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 6.6732183168327146e-06, |
|
"loss": 0.9666, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 0.8459530026109661, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 6.659643738952732e-06, |
|
"loss": 0.9368, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 0.8476936466492603, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 6.646055399603122e-06, |
|
"loss": 0.9466, |
|
"step": 487 |
|
}, |
|
{ |
|
"epoch": 0.8494342906875544, |
|
"grad_norm": 0.5, |
|
"learning_rate": 6.6324534114559656e-06, |
|
"loss": 0.9159, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 0.8511749347258486, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 6.618837887296523e-06, |
|
"loss": 0.9067, |
|
"step": 489 |
|
}, |
|
{ |
|
"epoch": 0.8529155787641427, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 6.605208940022289e-06, |
|
"loss": 0.9608, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.8546562228024369, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 6.591566682642061e-06, |
|
"loss": 0.9663, |
|
"step": 491 |
|
}, |
|
{ |
|
"epoch": 0.856396866840731, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 6.5779112282750035e-06, |
|
"loss": 0.8931, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 0.8581375108790252, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 6.564242690149705e-06, |
|
"loss": 0.9383, |
|
"step": 493 |
|
}, |
|
{ |
|
"epoch": 0.8598781549173194, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 6.550561181603244e-06, |
|
"loss": 0.9343, |
|
"step": 494 |
|
}, |
|
{ |
|
"epoch": 0.8616187989556136, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 6.536866816080247e-06, |
|
"loss": 0.9999, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 0.8633594429939078, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 6.523159707131951e-06, |
|
"loss": 0.9955, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 0.8651000870322019, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 6.509439968415252e-06, |
|
"loss": 0.8897, |
|
"step": 497 |
|
}, |
|
{ |
|
"epoch": 0.8668407310704961, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 6.4957077136917776e-06, |
|
"loss": 0.9065, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 0.8685813751087903, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 6.481963056826932e-06, |
|
"loss": 0.9224, |
|
"step": 499 |
|
}, |
|
{ |
|
"epoch": 0.8703220191470844, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 6.468206111788957e-06, |
|
"loss": 0.9638, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.8720626631853786, |
|
"grad_norm": 0.5, |
|
"learning_rate": 6.454436992647984e-06, |
|
"loss": 0.9232, |
|
"step": 501 |
|
}, |
|
{ |
|
"epoch": 0.8738033072236727, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 6.440655813575093e-06, |
|
"loss": 0.9412, |
|
"step": 502 |
|
}, |
|
{ |
|
"epoch": 0.8755439512619669, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 6.426862688841359e-06, |
|
"loss": 0.9051, |
|
"step": 503 |
|
}, |
|
{ |
|
"epoch": 0.8772845953002611, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 6.413057732816911e-06, |
|
"loss": 0.9214, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 0.8772845953002611, |
|
"eval_loss": 0.9230473637580872, |
|
"eval_runtime": 59.2532, |
|
"eval_samples_per_second": 79.118, |
|
"eval_steps_per_second": 9.89, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 0.8790252393385553, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 6.3992410599699786e-06, |
|
"loss": 0.908, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 0.8807658833768495, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 6.385412784865948e-06, |
|
"loss": 0.9613, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 0.8825065274151436, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 6.371573022166409e-06, |
|
"loss": 0.9061, |
|
"step": 507 |
|
}, |
|
{ |
|
"epoch": 0.8842471714534378, |
|
"grad_norm": 0.5, |
|
"learning_rate": 6.357721886628201e-06, |
|
"loss": 0.9117, |
|
"step": 508 |
|
}, |
|
{ |
|
"epoch": 0.8859878154917319, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 6.34385949310247e-06, |
|
"loss": 0.9219, |
|
"step": 509 |
|
}, |
|
{ |
|
"epoch": 0.8877284595300261, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 6.329985956533708e-06, |
|
"loss": 0.976, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.8894691035683203, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 6.3161013919588e-06, |
|
"loss": 0.9131, |
|
"step": 511 |
|
}, |
|
{ |
|
"epoch": 0.8912097476066144, |
|
"grad_norm": 0.5, |
|
"learning_rate": 6.302205914506083e-06, |
|
"loss": 0.908, |
|
"step": 512 |
|
}, |
|
{ |
|
"epoch": 0.8929503916449086, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 6.2882996393943706e-06, |
|
"loss": 0.9786, |
|
"step": 513 |
|
}, |
|
{ |
|
"epoch": 0.8946910356832027, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 6.274382681932019e-06, |
|
"loss": 0.8728, |
|
"step": 514 |
|
}, |
|
{ |
|
"epoch": 0.896431679721497, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 6.2604551575159476e-06, |
|
"loss": 0.9491, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 0.8981723237597912, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 6.24651718163071e-06, |
|
"loss": 0.995, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 0.8999129677980853, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 6.2325688698475106e-06, |
|
"loss": 0.9574, |
|
"step": 517 |
|
}, |
|
{ |
|
"epoch": 0.9016536118363795, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 6.218610337823262e-06, |
|
"loss": 0.9004, |
|
"step": 518 |
|
}, |
|
{ |
|
"epoch": 0.9033942558746736, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 6.2046417012996195e-06, |
|
"loss": 0.9165, |
|
"step": 519 |
|
}, |
|
{ |
|
"epoch": 0.9051348999129678, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 6.1906630761020245e-06, |
|
"loss": 0.9534, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.906875543951262, |
|
"grad_norm": 0.5, |
|
"learning_rate": 6.17667457813874e-06, |
|
"loss": 0.8883, |
|
"step": 521 |
|
}, |
|
{ |
|
"epoch": 0.9086161879895561, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 6.162676323399898e-06, |
|
"loss": 0.919, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 0.9103568320278503, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 6.148668427956523e-06, |
|
"loss": 0.9187, |
|
"step": 523 |
|
}, |
|
{ |
|
"epoch": 0.9120974760661444, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 6.134651007959586e-06, |
|
"loss": 0.8984, |
|
"step": 524 |
|
}, |
|
{ |
|
"epoch": 0.9138381201044387, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 6.120624179639032e-06, |
|
"loss": 0.9585, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.9155787641427328, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 6.106588059302818e-06, |
|
"loss": 0.9848, |
|
"step": 526 |
|
}, |
|
{ |
|
"epoch": 0.917319408181027, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 6.092542763335947e-06, |
|
"loss": 0.9542, |
|
"step": 527 |
|
}, |
|
{ |
|
"epoch": 0.9190600522193212, |
|
"grad_norm": 0.5, |
|
"learning_rate": 6.0784884081995065e-06, |
|
"loss": 0.9476, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 0.9208006962576153, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 6.0644251104296995e-06, |
|
"loss": 0.9425, |
|
"step": 529 |
|
}, |
|
{ |
|
"epoch": 0.9225413402959095, |
|
"grad_norm": 0.5, |
|
"learning_rate": 6.0503529866368824e-06, |
|
"loss": 0.9532, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.9242819843342036, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 6.036272153504592e-06, |
|
"loss": 0.9243, |
|
"step": 531 |
|
}, |
|
{ |
|
"epoch": 0.9260226283724978, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 6.022182727788586e-06, |
|
"loss": 0.9452, |
|
"step": 532 |
|
}, |
|
{ |
|
"epoch": 0.927763272410792, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 6.008084826315863e-06, |
|
"loss": 0.9243, |
|
"step": 533 |
|
}, |
|
{ |
|
"epoch": 0.9295039164490861, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 5.993978565983709e-06, |
|
"loss": 0.8917, |
|
"step": 534 |
|
}, |
|
{ |
|
"epoch": 0.9312445604873804, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 5.979864063758717e-06, |
|
"loss": 0.9562, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 0.9329852045256745, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 5.965741436675816e-06, |
|
"loss": 0.9904, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 0.9347258485639687, |
|
"grad_norm": 0.5, |
|
"learning_rate": 5.9516108018373145e-06, |
|
"loss": 0.8953, |
|
"step": 537 |
|
}, |
|
{ |
|
"epoch": 0.9364664926022629, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 5.937472276411909e-06, |
|
"loss": 0.8674, |
|
"step": 538 |
|
}, |
|
{ |
|
"epoch": 0.938207136640557, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 5.923325977633732e-06, |
|
"loss": 1.0281, |
|
"step": 539 |
|
}, |
|
{ |
|
"epoch": 0.9399477806788512, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 5.909172022801364e-06, |
|
"loss": 0.9278, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.9416884247171453, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 5.8950105292768754e-06, |
|
"loss": 0.8704, |
|
"step": 541 |
|
}, |
|
{ |
|
"epoch": 0.9434290687554395, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 5.880841614484841e-06, |
|
"loss": 0.9561, |
|
"step": 542 |
|
}, |
|
{ |
|
"epoch": 0.9451697127937336, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 5.866665395911375e-06, |
|
"loss": 0.9407, |
|
"step": 543 |
|
}, |
|
{ |
|
"epoch": 0.9469103568320278, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 5.852481991103149e-06, |
|
"loss": 0.8833, |
|
"step": 544 |
|
}, |
|
{ |
|
"epoch": 0.9486510008703221, |
|
"grad_norm": 0.5, |
|
"learning_rate": 5.838291517666427e-06, |
|
"loss": 0.993, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 0.9503916449086162, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 5.824094093266077e-06, |
|
"loss": 0.9231, |
|
"step": 546 |
|
}, |
|
{ |
|
"epoch": 0.9521322889469104, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 5.809889835624611e-06, |
|
"loss": 0.9512, |
|
"step": 547 |
|
}, |
|
{ |
|
"epoch": 0.9538729329852045, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 5.795678862521197e-06, |
|
"loss": 0.9175, |
|
"step": 548 |
|
}, |
|
{ |
|
"epoch": 0.9556135770234987, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 5.781461291790687e-06, |
|
"loss": 0.8922, |
|
"step": 549 |
|
}, |
|
{ |
|
"epoch": 0.9573542210617929, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 5.767237241322641e-06, |
|
"loss": 0.9701, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.959094865100087, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 5.753006829060343e-06, |
|
"loss": 0.9464, |
|
"step": 551 |
|
}, |
|
{ |
|
"epoch": 0.9608355091383812, |
|
"grad_norm": 0.5, |
|
"learning_rate": 5.738770172999835e-06, |
|
"loss": 0.9335, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 0.9625761531766753, |
|
"grad_norm": 0.458984375, |
|
"learning_rate": 5.724527391188927e-06, |
|
"loss": 0.9689, |
|
"step": 553 |
|
}, |
|
{ |
|
"epoch": 0.9643167972149695, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 5.710278601726222e-06, |
|
"loss": 0.8756, |
|
"step": 554 |
|
}, |
|
{ |
|
"epoch": 0.9660574412532638, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 5.696023922760141e-06, |
|
"loss": 0.9459, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 0.9677980852915579, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 5.681763472487933e-06, |
|
"loss": 0.9305, |
|
"step": 556 |
|
}, |
|
{ |
|
"epoch": 0.9695387293298521, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 5.667497369154712e-06, |
|
"loss": 0.8707, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.9712793733681462, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 5.6532257310524565e-06, |
|
"loss": 0.8977, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 0.9730200174064404, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 5.638948676519043e-06, |
|
"loss": 0.8735, |
|
"step": 559 |
|
}, |
|
{ |
|
"epoch": 0.9747606614447345, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 5.624666323937257e-06, |
|
"loss": 0.9788, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.9765013054830287, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 5.610378791733821e-06, |
|
"loss": 0.9348, |
|
"step": 561 |
|
}, |
|
{ |
|
"epoch": 0.9782419495213229, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 5.596086198378399e-06, |
|
"loss": 0.9258, |
|
"step": 562 |
|
}, |
|
{ |
|
"epoch": 0.979982593559617, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 5.5817886623826245e-06, |
|
"loss": 0.9184, |
|
"step": 563 |
|
}, |
|
{ |
|
"epoch": 0.9817232375979112, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 5.567486302299112e-06, |
|
"loss": 0.9439, |
|
"step": 564 |
|
}, |
|
{ |
|
"epoch": 0.9834638816362054, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 5.553179236720482e-06, |
|
"loss": 0.9887, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 0.9852045256744996, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 5.5388675842783644e-06, |
|
"loss": 0.9439, |
|
"step": 566 |
|
}, |
|
{ |
|
"epoch": 0.9869451697127938, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 5.524551463642429e-06, |
|
"loss": 0.971, |
|
"step": 567 |
|
}, |
|
{ |
|
"epoch": 0.9886858137510879, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 5.510230993519391e-06, |
|
"loss": 0.9583, |
|
"step": 568 |
|
}, |
|
{ |
|
"epoch": 0.9904264577893821, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 5.495906292652035e-06, |
|
"loss": 0.9788, |
|
"step": 569 |
|
}, |
|
{ |
|
"epoch": 0.9921671018276762, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 5.48157747981822e-06, |
|
"loss": 0.9634, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.9939077458659704, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 5.467244673829908e-06, |
|
"loss": 0.8834, |
|
"step": 571 |
|
}, |
|
{ |
|
"epoch": 0.9956483899042646, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 5.452907993532164e-06, |
|
"loss": 0.9113, |
|
"step": 572 |
|
}, |
|
{ |
|
"epoch": 0.9973890339425587, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 5.438567557802186e-06, |
|
"loss": 0.8828, |
|
"step": 573 |
|
}, |
|
{ |
|
"epoch": 0.999129677980853, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 5.424223485548303e-06, |
|
"loss": 0.9363, |
|
"step": 574 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 1148, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 2, |
|
"save_steps": 574, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 3.613017336006574e+18, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|