|
{ |
|
"best_global_step": null, |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.0, |
|
"eval_steps": 106, |
|
"global_step": 848, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0023584905660377358, |
|
"grad_norm": 5.1080776865120034, |
|
"learning_rate": 0.0, |
|
"loss": 0.9592, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0023584905660377358, |
|
"eval_loss": 0.9177566170692444, |
|
"eval_runtime": 83.3901, |
|
"eval_samples_per_second": 8.454, |
|
"eval_steps_per_second": 0.54, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0047169811320754715, |
|
"grad_norm": 5.216656877082355, |
|
"learning_rate": 4e-08, |
|
"loss": 0.9643, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.007075471698113208, |
|
"grad_norm": 5.86435572061478, |
|
"learning_rate": 8e-08, |
|
"loss": 0.9448, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.009433962264150943, |
|
"grad_norm": 4.885606218308886, |
|
"learning_rate": 1.2e-07, |
|
"loss": 0.9499, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.01179245283018868, |
|
"grad_norm": 4.863782725318177, |
|
"learning_rate": 1.6e-07, |
|
"loss": 0.9547, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.014150943396226415, |
|
"grad_norm": 4.84448661904324, |
|
"learning_rate": 2e-07, |
|
"loss": 0.9378, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.01650943396226415, |
|
"grad_norm": 4.594764533947918, |
|
"learning_rate": 2.4e-07, |
|
"loss": 0.967, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.018867924528301886, |
|
"grad_norm": 5.311507883731841, |
|
"learning_rate": 2.8e-07, |
|
"loss": 0.9534, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.02122641509433962, |
|
"grad_norm": 4.780331285112587, |
|
"learning_rate": 3.2e-07, |
|
"loss": 0.979, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.02358490566037736, |
|
"grad_norm": 5.172814213529754, |
|
"learning_rate": 3.6e-07, |
|
"loss": 0.9284, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.025943396226415096, |
|
"grad_norm": 4.693805464386623, |
|
"learning_rate": 4e-07, |
|
"loss": 0.9452, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.02830188679245283, |
|
"grad_norm": 5.378045080991313, |
|
"learning_rate": 4.3999999999999997e-07, |
|
"loss": 0.9575, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.030660377358490566, |
|
"grad_norm": 4.718804488820069, |
|
"learning_rate": 4.8e-07, |
|
"loss": 0.9601, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.0330188679245283, |
|
"grad_norm": 4.8688696728838785, |
|
"learning_rate": 5.2e-07, |
|
"loss": 0.9578, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.03537735849056604, |
|
"grad_norm": 4.41639309685216, |
|
"learning_rate": 5.6e-07, |
|
"loss": 0.9433, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.03773584905660377, |
|
"grad_norm": 4.330783791521376, |
|
"learning_rate": 6e-07, |
|
"loss": 0.9498, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.04009433962264151, |
|
"grad_norm": 4.580748421983631, |
|
"learning_rate": 6.4e-07, |
|
"loss": 0.9607, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.04245283018867924, |
|
"grad_norm": 3.895307917106911, |
|
"learning_rate": 6.800000000000001e-07, |
|
"loss": 0.9561, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.04481132075471698, |
|
"grad_norm": 3.9063376740559894, |
|
"learning_rate": 7.2e-07, |
|
"loss": 0.9423, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.04716981132075472, |
|
"grad_norm": 3.9785213934452335, |
|
"learning_rate": 7.599999999999999e-07, |
|
"loss": 0.9422, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.049528301886792456, |
|
"grad_norm": 3.5176355010251044, |
|
"learning_rate": 8e-07, |
|
"loss": 0.9289, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.05188679245283019, |
|
"grad_norm": 5.022294330418908, |
|
"learning_rate": 8.399999999999999e-07, |
|
"loss": 0.9446, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.054245283018867926, |
|
"grad_norm": 2.6661772621278637, |
|
"learning_rate": 8.799999999999999e-07, |
|
"loss": 0.9396, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.05660377358490566, |
|
"grad_norm": 2.8934776201699757, |
|
"learning_rate": 9.2e-07, |
|
"loss": 0.9424, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.0589622641509434, |
|
"grad_norm": 3.310951896187245, |
|
"learning_rate": 9.6e-07, |
|
"loss": 0.9317, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.06132075471698113, |
|
"grad_norm": 3.180088649692653, |
|
"learning_rate": 1e-06, |
|
"loss": 0.8733, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.06367924528301887, |
|
"grad_norm": 3.0819569242393032, |
|
"learning_rate": 1.04e-06, |
|
"loss": 0.9422, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.0660377358490566, |
|
"grad_norm": 4.5985187060206405, |
|
"learning_rate": 1.08e-06, |
|
"loss": 0.9522, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.06839622641509434, |
|
"grad_norm": 2.614037871400199, |
|
"learning_rate": 1.12e-06, |
|
"loss": 0.9168, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.07075471698113207, |
|
"grad_norm": 3.4637889884655007, |
|
"learning_rate": 1.16e-06, |
|
"loss": 0.8819, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.07311320754716981, |
|
"grad_norm": 2.7607684537358623, |
|
"learning_rate": 1.2e-06, |
|
"loss": 0.9443, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.07547169811320754, |
|
"grad_norm": 2.45835136955183, |
|
"learning_rate": 1.24e-06, |
|
"loss": 0.8971, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.07783018867924528, |
|
"grad_norm": 2.8319894880867724, |
|
"learning_rate": 1.28e-06, |
|
"loss": 0.9311, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.08018867924528301, |
|
"grad_norm": 1.880755586715356, |
|
"learning_rate": 1.32e-06, |
|
"loss": 0.895, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.08254716981132075, |
|
"grad_norm": 3.1062915440189713, |
|
"learning_rate": 1.3600000000000001e-06, |
|
"loss": 0.9074, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.08490566037735849, |
|
"grad_norm": 2.745579181683214, |
|
"learning_rate": 1.4e-06, |
|
"loss": 0.9066, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.08726415094339622, |
|
"grad_norm": 2.51965653055992, |
|
"learning_rate": 1.44e-06, |
|
"loss": 0.8737, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.08962264150943396, |
|
"grad_norm": 2.324394856248302, |
|
"learning_rate": 1.48e-06, |
|
"loss": 0.9239, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.09198113207547169, |
|
"grad_norm": 2.7275453116030834, |
|
"learning_rate": 1.5199999999999998e-06, |
|
"loss": 0.9029, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.09433962264150944, |
|
"grad_norm": 2.8394394816071498, |
|
"learning_rate": 1.5599999999999999e-06, |
|
"loss": 0.885, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.09669811320754718, |
|
"grad_norm": 2.3797648877001323, |
|
"learning_rate": 1.6e-06, |
|
"loss": 0.9084, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.09905660377358491, |
|
"grad_norm": 2.298049507841523, |
|
"learning_rate": 1.6399999999999998e-06, |
|
"loss": 0.8786, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.10141509433962265, |
|
"grad_norm": 2.0882957626327814, |
|
"learning_rate": 1.6799999999999998e-06, |
|
"loss": 0.885, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.10377358490566038, |
|
"grad_norm": 1.9805393143682433, |
|
"learning_rate": 1.7199999999999998e-06, |
|
"loss": 0.9202, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.10613207547169812, |
|
"grad_norm": 2.1136265646711605, |
|
"learning_rate": 1.7599999999999999e-06, |
|
"loss": 0.8515, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.10849056603773585, |
|
"grad_norm": 1.7161578870903316, |
|
"learning_rate": 1.8e-06, |
|
"loss": 0.9035, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.11084905660377359, |
|
"grad_norm": 2.2352887975077795, |
|
"learning_rate": 1.84e-06, |
|
"loss": 0.8615, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.11320754716981132, |
|
"grad_norm": 1.997343145683379, |
|
"learning_rate": 1.8799999999999998e-06, |
|
"loss": 0.8958, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.11556603773584906, |
|
"grad_norm": 2.1520613046888024, |
|
"learning_rate": 1.92e-06, |
|
"loss": 0.8706, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.1179245283018868, |
|
"grad_norm": 1.610078026570572, |
|
"learning_rate": 1.96e-06, |
|
"loss": 0.9127, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.12028301886792453, |
|
"grad_norm": 1.6267119325099246, |
|
"learning_rate": 2e-06, |
|
"loss": 0.8604, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.12264150943396226, |
|
"grad_norm": 1.7305415242899538, |
|
"learning_rate": 1.99999225068337e-06, |
|
"loss": 0.8704, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.125, |
|
"grad_norm": 4.837245274184259, |
|
"learning_rate": 1.9999690028535855e-06, |
|
"loss": 0.9003, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.12735849056603774, |
|
"grad_norm": 2.006265922883068, |
|
"learning_rate": 1.9999302568709546e-06, |
|
"loss": 0.8759, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.12971698113207547, |
|
"grad_norm": 1.8749359565723283, |
|
"learning_rate": 1.999876013335988e-06, |
|
"loss": 0.8686, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.1320754716981132, |
|
"grad_norm": 1.5041642645696787, |
|
"learning_rate": 1.9998062730893862e-06, |
|
"loss": 0.8727, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.13443396226415094, |
|
"grad_norm": 1.4630749706626824, |
|
"learning_rate": 1.9997210372120272e-06, |
|
"loss": 0.8718, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.13679245283018868, |
|
"grad_norm": 1.4357346298179297, |
|
"learning_rate": 1.9996203070249514e-06, |
|
"loss": 0.9, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.1391509433962264, |
|
"grad_norm": 1.589349788137753, |
|
"learning_rate": 1.9995040840893383e-06, |
|
"loss": 0.8832, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.14150943396226415, |
|
"grad_norm": 1.3078655659788383, |
|
"learning_rate": 1.9993723702064853e-06, |
|
"loss": 0.8689, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.14386792452830188, |
|
"grad_norm": 1.2162726319357984, |
|
"learning_rate": 1.9992251674177763e-06, |
|
"loss": 0.8565, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.14622641509433962, |
|
"grad_norm": 1.4682400986060873, |
|
"learning_rate": 1.999062478004655e-06, |
|
"loss": 0.8768, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.14858490566037735, |
|
"grad_norm": 1.475835127346073, |
|
"learning_rate": 1.9988843044885837e-06, |
|
"loss": 0.8485, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.1509433962264151, |
|
"grad_norm": 1.313661603517591, |
|
"learning_rate": 1.998690649631009e-06, |
|
"loss": 0.8791, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.15330188679245282, |
|
"grad_norm": 1.2318794472546806, |
|
"learning_rate": 1.998481516433316e-06, |
|
"loss": 0.8418, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.15566037735849056, |
|
"grad_norm": 1.224129101101457, |
|
"learning_rate": 1.9982569081367843e-06, |
|
"loss": 0.8472, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.1580188679245283, |
|
"grad_norm": 1.4354409800877264, |
|
"learning_rate": 1.9980168282225344e-06, |
|
"loss": 0.8682, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.16037735849056603, |
|
"grad_norm": 1.3882277723705556, |
|
"learning_rate": 1.9977612804114775e-06, |
|
"loss": 0.8929, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.16273584905660377, |
|
"grad_norm": 1.1353588967534465, |
|
"learning_rate": 1.9974902686642557e-06, |
|
"loss": 0.8123, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.1650943396226415, |
|
"grad_norm": 1.1680380889234738, |
|
"learning_rate": 1.9972037971811797e-06, |
|
"loss": 0.8573, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.16745283018867924, |
|
"grad_norm": 1.1797181726680495, |
|
"learning_rate": 1.9969018704021675e-06, |
|
"loss": 0.8518, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.16981132075471697, |
|
"grad_norm": 1.3076313839386762, |
|
"learning_rate": 1.9965844930066696e-06, |
|
"loss": 0.8425, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.1721698113207547, |
|
"grad_norm": 1.2192501506435032, |
|
"learning_rate": 1.9962516699136036e-06, |
|
"loss": 0.8633, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.17452830188679244, |
|
"grad_norm": 1.1885172994924504, |
|
"learning_rate": 1.9959034062812714e-06, |
|
"loss": 0.8417, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.17688679245283018, |
|
"grad_norm": 1.0228274366877697, |
|
"learning_rate": 1.9955397075072833e-06, |
|
"loss": 0.8645, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.1792452830188679, |
|
"grad_norm": 1.0362433430443114, |
|
"learning_rate": 1.9951605792284742e-06, |
|
"loss": 0.8442, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.18160377358490565, |
|
"grad_norm": 1.3774246732359812, |
|
"learning_rate": 1.9947660273208134e-06, |
|
"loss": 0.8355, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.18396226415094338, |
|
"grad_norm": 1.3163064309206076, |
|
"learning_rate": 1.9943560578993165e-06, |
|
"loss": 0.8246, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.18632075471698112, |
|
"grad_norm": 1.0822503373501942, |
|
"learning_rate": 1.9939306773179494e-06, |
|
"loss": 0.8315, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.18867924528301888, |
|
"grad_norm": 1.0361871014520703, |
|
"learning_rate": 1.9934898921695292e-06, |
|
"loss": 0.8346, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.19103773584905662, |
|
"grad_norm": 1.0371978245512512, |
|
"learning_rate": 1.993033709285624e-06, |
|
"loss": 0.8706, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.19339622641509435, |
|
"grad_norm": 1.0574050613888015, |
|
"learning_rate": 1.992562135736444e-06, |
|
"loss": 0.8308, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.1957547169811321, |
|
"grad_norm": 1.241296877885062, |
|
"learning_rate": 1.992075178830736e-06, |
|
"loss": 0.872, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.19811320754716982, |
|
"grad_norm": 1.152024986745144, |
|
"learning_rate": 1.9915728461156654e-06, |
|
"loss": 0.8365, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.20047169811320756, |
|
"grad_norm": 1.0972286027771576, |
|
"learning_rate": 1.991055145376703e-06, |
|
"loss": 0.8511, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.2028301886792453, |
|
"grad_norm": 1.3363467615851807, |
|
"learning_rate": 1.990522084637503e-06, |
|
"loss": 0.8604, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.20518867924528303, |
|
"grad_norm": 0.9249543605316475, |
|
"learning_rate": 1.9899736721597786e-06, |
|
"loss": 0.8078, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.20754716981132076, |
|
"grad_norm": 1.028197337537026, |
|
"learning_rate": 1.9894099164431722e-06, |
|
"loss": 0.8572, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.2099056603773585, |
|
"grad_norm": 0.9345635713942412, |
|
"learning_rate": 1.9888308262251284e-06, |
|
"loss": 0.814, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.21226415094339623, |
|
"grad_norm": 1.0476770833253457, |
|
"learning_rate": 1.9882364104807535e-06, |
|
"loss": 0.8358, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.21462264150943397, |
|
"grad_norm": 1.0517946592201646, |
|
"learning_rate": 1.9876266784226794e-06, |
|
"loss": 0.8263, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.2169811320754717, |
|
"grad_norm": 1.1715532913432805, |
|
"learning_rate": 1.987001639500919e-06, |
|
"loss": 0.8268, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.21933962264150944, |
|
"grad_norm": 0.928474382958498, |
|
"learning_rate": 1.9863613034027223e-06, |
|
"loss": 0.8278, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.22169811320754718, |
|
"grad_norm": 0.9243352298229107, |
|
"learning_rate": 1.985705680052423e-06, |
|
"loss": 0.817, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.2240566037735849, |
|
"grad_norm": 0.9825897380291061, |
|
"learning_rate": 1.985034779611287e-06, |
|
"loss": 0.8185, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.22641509433962265, |
|
"grad_norm": 1.0129581531905947, |
|
"learning_rate": 1.9843486124773543e-06, |
|
"loss": 0.8261, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.22877358490566038, |
|
"grad_norm": 0.9634164693430555, |
|
"learning_rate": 1.9836471892852777e-06, |
|
"loss": 0.8448, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.23113207547169812, |
|
"grad_norm": 0.9208269074792377, |
|
"learning_rate": 1.982930520906158e-06, |
|
"loss": 0.8435, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.23349056603773585, |
|
"grad_norm": 1.2744902927155426, |
|
"learning_rate": 1.9821986184473754e-06, |
|
"loss": 0.811, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.2358490566037736, |
|
"grad_norm": 0.9132891996164993, |
|
"learning_rate": 1.9814514932524176e-06, |
|
"loss": 0.8629, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.23820754716981132, |
|
"grad_norm": 1.3924445022644105, |
|
"learning_rate": 1.9806891569007048e-06, |
|
"loss": 0.8157, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.24056603773584906, |
|
"grad_norm": 0.9910775720488755, |
|
"learning_rate": 1.9799116212074075e-06, |
|
"loss": 0.8133, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.2429245283018868, |
|
"grad_norm": 0.8521457201237292, |
|
"learning_rate": 1.979118898223267e-06, |
|
"loss": 0.818, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.24528301886792453, |
|
"grad_norm": 0.961507205526783, |
|
"learning_rate": 1.978311000234406e-06, |
|
"loss": 0.8312, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.24764150943396226, |
|
"grad_norm": 0.9234203353918131, |
|
"learning_rate": 1.9774879397621383e-06, |
|
"loss": 0.8307, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.939970736839533, |
|
"learning_rate": 1.9766497295627777e-06, |
|
"loss": 0.8121, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"eval_loss": 0.7522591948509216, |
|
"eval_runtime": 82.8066, |
|
"eval_samples_per_second": 8.514, |
|
"eval_steps_per_second": 0.543, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.25235849056603776, |
|
"grad_norm": 1.009636311779919, |
|
"learning_rate": 1.9757963826274354e-06, |
|
"loss": 0.8321, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.25471698113207547, |
|
"grad_norm": 0.8979616538385209, |
|
"learning_rate": 1.9749279121818236e-06, |
|
"loss": 0.8442, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.25707547169811323, |
|
"grad_norm": 0.953771354540708, |
|
"learning_rate": 1.9740443316860463e-06, |
|
"loss": 0.8484, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.25943396226415094, |
|
"grad_norm": 1.0957061062800166, |
|
"learning_rate": 1.9731456548343944e-06, |
|
"loss": 0.8204, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.2617924528301887, |
|
"grad_norm": 0.8834656211178301, |
|
"learning_rate": 1.9722318955551303e-06, |
|
"loss": 0.7817, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.2641509433962264, |
|
"grad_norm": 1.0070698705880765, |
|
"learning_rate": 1.9713030680102743e-06, |
|
"loss": 0.8309, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.2665094339622642, |
|
"grad_norm": 1.0103166521916906, |
|
"learning_rate": 1.970359186595384e-06, |
|
"loss": 0.8454, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.2688679245283019, |
|
"grad_norm": 0.7965830780073067, |
|
"learning_rate": 1.9694002659393305e-06, |
|
"loss": 0.7659, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.27122641509433965, |
|
"grad_norm": 0.9676729525657803, |
|
"learning_rate": 1.968426320904074e-06, |
|
"loss": 0.8076, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.27358490566037735, |
|
"grad_norm": 0.8547052227174852, |
|
"learning_rate": 1.967437366584431e-06, |
|
"loss": 0.8305, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.2759433962264151, |
|
"grad_norm": 0.9129048151834617, |
|
"learning_rate": 1.9664334183078425e-06, |
|
"loss": 0.8443, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.2783018867924528, |
|
"grad_norm": 0.8521299847115695, |
|
"learning_rate": 1.965414491634134e-06, |
|
"loss": 0.8244, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.2806603773584906, |
|
"grad_norm": 0.8875033746133125, |
|
"learning_rate": 1.964380602355277e-06, |
|
"loss": 0.7855, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.2830188679245283, |
|
"grad_norm": 0.8322961791869938, |
|
"learning_rate": 1.9633317664951417e-06, |
|
"loss": 0.8246, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.28537735849056606, |
|
"grad_norm": 0.8996547969201056, |
|
"learning_rate": 1.9622680003092503e-06, |
|
"loss": 0.7911, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.28773584905660377, |
|
"grad_norm": 0.8421734340822833, |
|
"learning_rate": 1.9611893202845253e-06, |
|
"loss": 0.8075, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.29009433962264153, |
|
"grad_norm": 0.9715000278619699, |
|
"learning_rate": 1.9600957431390324e-06, |
|
"loss": 0.8258, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.29245283018867924, |
|
"grad_norm": 0.9486047429972293, |
|
"learning_rate": 1.9589872858217233e-06, |
|
"loss": 0.8249, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.294811320754717, |
|
"grad_norm": 0.8708681162275084, |
|
"learning_rate": 1.9578639655121707e-06, |
|
"loss": 0.8061, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.2971698113207547, |
|
"grad_norm": 0.815600627546263, |
|
"learning_rate": 1.9567257996203046e-06, |
|
"loss": 0.8117, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.29952830188679247, |
|
"grad_norm": 0.899644949674199, |
|
"learning_rate": 1.955572805786141e-06, |
|
"loss": 0.8324, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.3018867924528302, |
|
"grad_norm": 0.7797228121513826, |
|
"learning_rate": 1.9544050018795075e-06, |
|
"loss": 0.8085, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.30424528301886794, |
|
"grad_norm": 1.4425571525030387, |
|
"learning_rate": 1.953222405999769e-06, |
|
"loss": 0.8043, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.30660377358490565, |
|
"grad_norm": 1.3224705998967385, |
|
"learning_rate": 1.9520250364755458e-06, |
|
"loss": 0.8104, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.3089622641509434, |
|
"grad_norm": 0.8148761244743041, |
|
"learning_rate": 1.9508129118644293e-06, |
|
"loss": 0.828, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.3113207547169811, |
|
"grad_norm": 1.0994942919129091, |
|
"learning_rate": 1.949586050952693e-06, |
|
"loss": 0.8007, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.3136792452830189, |
|
"grad_norm": 0.9838294161632027, |
|
"learning_rate": 1.9483444727550054e-06, |
|
"loss": 0.8304, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.3160377358490566, |
|
"grad_norm": 1.0206044534349468, |
|
"learning_rate": 1.9470881965141307e-06, |
|
"loss": 0.8044, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.31839622641509435, |
|
"grad_norm": 0.8186471659261382, |
|
"learning_rate": 1.9458172417006346e-06, |
|
"loss": 0.7921, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.32075471698113206, |
|
"grad_norm": 1.040687215576322, |
|
"learning_rate": 1.944531628012579e-06, |
|
"loss": 0.8309, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.3231132075471698, |
|
"grad_norm": 0.8067576806018053, |
|
"learning_rate": 1.9432313753752194e-06, |
|
"loss": 0.8006, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.32547169811320753, |
|
"grad_norm": 1.1895556658600497, |
|
"learning_rate": 1.941916503940694e-06, |
|
"loss": 0.8055, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.3278301886792453, |
|
"grad_norm": 0.877104803725933, |
|
"learning_rate": 1.9405870340877135e-06, |
|
"loss": 0.815, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.330188679245283, |
|
"grad_norm": 1.0755873765351904, |
|
"learning_rate": 1.9392429864212433e-06, |
|
"loss": 0.8357, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.33254716981132076, |
|
"grad_norm": 0.8389051921712227, |
|
"learning_rate": 1.9378843817721854e-06, |
|
"loss": 0.8054, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.33490566037735847, |
|
"grad_norm": 0.8492954330518299, |
|
"learning_rate": 1.9365112411970546e-06, |
|
"loss": 0.8125, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.33726415094339623, |
|
"grad_norm": 0.8117108742521268, |
|
"learning_rate": 1.9351235859776537e-06, |
|
"loss": 0.7996, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.33962264150943394, |
|
"grad_norm": 0.8734747925195555, |
|
"learning_rate": 1.9337214376207417e-06, |
|
"loss": 0.7987, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.3419811320754717, |
|
"grad_norm": 0.870843122026526, |
|
"learning_rate": 1.932304817857702e-06, |
|
"loss": 0.8071, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.3443396226415094, |
|
"grad_norm": 0.7823387766729867, |
|
"learning_rate": 1.930873748644204e-06, |
|
"loss": 0.781, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.3466981132075472, |
|
"grad_norm": 1.0441516146481582, |
|
"learning_rate": 1.9294282521598657e-06, |
|
"loss": 0.8211, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.3490566037735849, |
|
"grad_norm": 0.8689106447919821, |
|
"learning_rate": 1.927968350807906e-06, |
|
"loss": 0.7827, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.35141509433962265, |
|
"grad_norm": 0.8369054883974562, |
|
"learning_rate": 1.9264940672148015e-06, |
|
"loss": 0.7987, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.35377358490566035, |
|
"grad_norm": 1.628472390625535, |
|
"learning_rate": 1.9250054242299326e-06, |
|
"loss": 0.8245, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.3561320754716981, |
|
"grad_norm": 0.8210347316488275, |
|
"learning_rate": 1.9235024449252305e-06, |
|
"loss": 0.8026, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.3584905660377358, |
|
"grad_norm": 0.8976272281891767, |
|
"learning_rate": 1.9219851525948203e-06, |
|
"loss": 0.8343, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.3608490566037736, |
|
"grad_norm": 0.8688278315807118, |
|
"learning_rate": 1.92045357075466e-06, |
|
"loss": 0.8378, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.3632075471698113, |
|
"grad_norm": 0.8259924861279039, |
|
"learning_rate": 1.9189077231421746e-06, |
|
"loss": 0.771, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.36556603773584906, |
|
"grad_norm": 0.850617328037459, |
|
"learning_rate": 1.917347633715889e-06, |
|
"loss": 0.8116, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.36792452830188677, |
|
"grad_norm": 0.830159427983018, |
|
"learning_rate": 1.915773326655057e-06, |
|
"loss": 0.8142, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.37028301886792453, |
|
"grad_norm": 0.9047033822843099, |
|
"learning_rate": 1.9141848263592873e-06, |
|
"loss": 0.7867, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.37264150943396224, |
|
"grad_norm": 0.8043279871245154, |
|
"learning_rate": 1.9125821574481623e-06, |
|
"loss": 0.7821, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.375, |
|
"grad_norm": 0.9855797219219302, |
|
"learning_rate": 1.9109653447608605e-06, |
|
"loss": 0.838, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.37735849056603776, |
|
"grad_norm": 0.8671907333946726, |
|
"learning_rate": 1.909334413355768e-06, |
|
"loss": 0.8199, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.37971698113207547, |
|
"grad_norm": 0.8038519157610383, |
|
"learning_rate": 1.9076893885100934e-06, |
|
"loss": 0.7914, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.38207547169811323, |
|
"grad_norm": 0.8534138715990678, |
|
"learning_rate": 1.906030295719473e-06, |
|
"loss": 0.803, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.38443396226415094, |
|
"grad_norm": 1.272101513607124, |
|
"learning_rate": 1.9043571606975775e-06, |
|
"loss": 0.8243, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.3867924528301887, |
|
"grad_norm": 0.8724688206578753, |
|
"learning_rate": 1.9026700093757129e-06, |
|
"loss": 0.8, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.3891509433962264, |
|
"grad_norm": 0.8272714642025372, |
|
"learning_rate": 1.9009688679024189e-06, |
|
"loss": 0.7941, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.3915094339622642, |
|
"grad_norm": 1.609360832798017, |
|
"learning_rate": 1.8992537626430636e-06, |
|
"loss": 0.8184, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.3938679245283019, |
|
"grad_norm": 1.068913128886596, |
|
"learning_rate": 1.897524720179434e-06, |
|
"loss": 0.7551, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.39622641509433965, |
|
"grad_norm": 0.876062488501413, |
|
"learning_rate": 1.8957817673093256e-06, |
|
"loss": 0.8202, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.39858490566037735, |
|
"grad_norm": 0.7804659272450435, |
|
"learning_rate": 1.894024931046125e-06, |
|
"loss": 0.8096, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.4009433962264151, |
|
"grad_norm": 1.1020958270083518, |
|
"learning_rate": 1.8922542386183939e-06, |
|
"loss": 0.8023, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.4033018867924528, |
|
"grad_norm": 0.8307091625451186, |
|
"learning_rate": 1.8904697174694446e-06, |
|
"loss": 0.797, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.4056603773584906, |
|
"grad_norm": 1.4660516537985, |
|
"learning_rate": 1.8886713952569156e-06, |
|
"loss": 0.8001, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.4080188679245283, |
|
"grad_norm": 0.9042037911584216, |
|
"learning_rate": 1.8868592998523436e-06, |
|
"loss": 0.7798, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.41037735849056606, |
|
"grad_norm": 0.8127436933722597, |
|
"learning_rate": 1.885033459340731e-06, |
|
"loss": 0.791, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.41273584905660377, |
|
"grad_norm": 0.8906980453056944, |
|
"learning_rate": 1.8831939020201096e-06, |
|
"loss": 0.8117, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.41509433962264153, |
|
"grad_norm": 0.7811131450862968, |
|
"learning_rate": 1.8813406564011044e-06, |
|
"loss": 0.742, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.41745283018867924, |
|
"grad_norm": 2.0285717851138974, |
|
"learning_rate": 1.8794737512064888e-06, |
|
"loss": 0.803, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.419811320754717, |
|
"grad_norm": 0.9681303046265977, |
|
"learning_rate": 1.8775932153707426e-06, |
|
"loss": 0.7857, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.4221698113207547, |
|
"grad_norm": 0.8312813398366425, |
|
"learning_rate": 1.8756990780396006e-06, |
|
"loss": 0.8091, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.42452830188679247, |
|
"grad_norm": 0.8325963532837448, |
|
"learning_rate": 1.8737913685696027e-06, |
|
"loss": 0.7936, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.4268867924528302, |
|
"grad_norm": 1.0242582555350113, |
|
"learning_rate": 1.8718701165276383e-06, |
|
"loss": 0.7679, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.42924528301886794, |
|
"grad_norm": 0.8568509432407337, |
|
"learning_rate": 1.869935351690488e-06, |
|
"loss": 0.7813, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.43160377358490565, |
|
"grad_norm": 1.1130973261677584, |
|
"learning_rate": 1.867987104044363e-06, |
|
"loss": 0.8173, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.4339622641509434, |
|
"grad_norm": 0.8076547906696275, |
|
"learning_rate": 1.8660254037844386e-06, |
|
"loss": 0.7823, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.4363207547169811, |
|
"grad_norm": 0.8462497637993398, |
|
"learning_rate": 1.864050281314388e-06, |
|
"loss": 0.7964, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.4386792452830189, |
|
"grad_norm": 0.8157104951412149, |
|
"learning_rate": 1.8620617672459096e-06, |
|
"loss": 0.7851, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.4410377358490566, |
|
"grad_norm": 0.8016398797843809, |
|
"learning_rate": 1.8600598923982537e-06, |
|
"loss": 0.7939, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.44339622641509435, |
|
"grad_norm": 0.9872324739847712, |
|
"learning_rate": 1.858044687797745e-06, |
|
"loss": 0.8105, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.44575471698113206, |
|
"grad_norm": 0.8070225439686687, |
|
"learning_rate": 1.8560161846773e-06, |
|
"loss": 0.7972, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.4481132075471698, |
|
"grad_norm": 0.8159111285487299, |
|
"learning_rate": 1.8539744144759447e-06, |
|
"loss": 0.7956, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.45047169811320753, |
|
"grad_norm": 0.8271007147512706, |
|
"learning_rate": 1.851919408838327e-06, |
|
"loss": 0.801, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.4528301886792453, |
|
"grad_norm": 0.8240075084440164, |
|
"learning_rate": 1.8498511996142253e-06, |
|
"loss": 0.8131, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.455188679245283, |
|
"grad_norm": 0.8100827961239685, |
|
"learning_rate": 1.8477698188580557e-06, |
|
"loss": 0.7953, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.45754716981132076, |
|
"grad_norm": 0.8090963621477647, |
|
"learning_rate": 1.8456752988283757e-06, |
|
"loss": 0.7713, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.45990566037735847, |
|
"grad_norm": 0.8646115465094427, |
|
"learning_rate": 1.8435676719873827e-06, |
|
"loss": 0.8218, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.46226415094339623, |
|
"grad_norm": 0.8283897647434588, |
|
"learning_rate": 1.8414469710004124e-06, |
|
"loss": 0.7937, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.46462264150943394, |
|
"grad_norm": 0.8130051258008277, |
|
"learning_rate": 1.839313228735431e-06, |
|
"loss": 0.7822, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.4669811320754717, |
|
"grad_norm": 0.8754673959382613, |
|
"learning_rate": 1.8371664782625285e-06, |
|
"loss": 0.7903, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.4693396226415094, |
|
"grad_norm": 1.287939397478769, |
|
"learning_rate": 1.8350067528534024e-06, |
|
"loss": 0.7718, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.4716981132075472, |
|
"grad_norm": 0.8236848990785427, |
|
"learning_rate": 1.8328340859808446e-06, |
|
"loss": 0.8012, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.4740566037735849, |
|
"grad_norm": 0.8938042537260998, |
|
"learning_rate": 1.8306485113182229e-06, |
|
"loss": 0.7721, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.47641509433962265, |
|
"grad_norm": 0.8496295751844541, |
|
"learning_rate": 1.8284500627389567e-06, |
|
"loss": 0.8043, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.47877358490566035, |
|
"grad_norm": 0.8325387527055562, |
|
"learning_rate": 1.8262387743159948e-06, |
|
"loss": 0.7936, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.4811320754716981, |
|
"grad_norm": 0.8207579274147835, |
|
"learning_rate": 1.824014680321285e-06, |
|
"loss": 0.8153, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.4834905660377358, |
|
"grad_norm": 0.9100175298073298, |
|
"learning_rate": 1.821777815225245e-06, |
|
"loss": 0.777, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.4858490566037736, |
|
"grad_norm": 0.9563101605978098, |
|
"learning_rate": 1.8195282136962264e-06, |
|
"loss": 0.7991, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.4882075471698113, |
|
"grad_norm": 0.8404681193002833, |
|
"learning_rate": 1.817265910599978e-06, |
|
"loss": 0.7968, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.49056603773584906, |
|
"grad_norm": 0.898412624742003, |
|
"learning_rate": 1.814990940999106e-06, |
|
"loss": 0.7585, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.49292452830188677, |
|
"grad_norm": 0.8495327539856999, |
|
"learning_rate": 1.81270334015253e-06, |
|
"loss": 0.8133, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.49528301886792453, |
|
"grad_norm": 1.1225882373174563, |
|
"learning_rate": 1.8104031435149362e-06, |
|
"loss": 0.8121, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.49764150943396224, |
|
"grad_norm": 0.9329252376646318, |
|
"learning_rate": 1.8080903867362293e-06, |
|
"loss": 0.815, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 1.0591364512578068, |
|
"learning_rate": 1.8057651056609782e-06, |
|
"loss": 0.7649, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"eval_loss": 0.714023768901825, |
|
"eval_runtime": 83.0267, |
|
"eval_samples_per_second": 8.491, |
|
"eval_steps_per_second": 0.542, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.5023584905660378, |
|
"grad_norm": 0.8614390398527076, |
|
"learning_rate": 1.803427336327861e-06, |
|
"loss": 0.8039, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.5047169811320755, |
|
"grad_norm": 0.903956531352285, |
|
"learning_rate": 1.8010771149691082e-06, |
|
"loss": 0.784, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.5070754716981132, |
|
"grad_norm": 0.8574588324038701, |
|
"learning_rate": 1.7987144780099376e-06, |
|
"loss": 0.7956, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.5094339622641509, |
|
"grad_norm": 0.8976197887421247, |
|
"learning_rate": 1.7963394620679942e-06, |
|
"loss": 0.8261, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.5117924528301887, |
|
"grad_norm": 0.8410804453164143, |
|
"learning_rate": 1.7939521039527781e-06, |
|
"loss": 0.7816, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.5141509433962265, |
|
"grad_norm": 0.903488950014599, |
|
"learning_rate": 1.7915524406650775e-06, |
|
"loss": 0.7701, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.5165094339622641, |
|
"grad_norm": 0.8374711197328489, |
|
"learning_rate": 1.7891405093963937e-06, |
|
"loss": 0.7761, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.5188679245283019, |
|
"grad_norm": 0.938187408138231, |
|
"learning_rate": 1.7867163475283646e-06, |
|
"loss": 0.8247, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.5212264150943396, |
|
"grad_norm": 1.974177068847498, |
|
"learning_rate": 1.7842799926321863e-06, |
|
"loss": 0.7984, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.5235849056603774, |
|
"grad_norm": 0.8445314748349818, |
|
"learning_rate": 1.7818314824680298e-06, |
|
"loss": 0.7783, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.5259433962264151, |
|
"grad_norm": 0.8043844477824647, |
|
"learning_rate": 1.779370854984456e-06, |
|
"loss": 0.7748, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.5283018867924528, |
|
"grad_norm": 0.8199345888971309, |
|
"learning_rate": 1.7768981483178279e-06, |
|
"loss": 0.7961, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.5306603773584906, |
|
"grad_norm": 0.8198791332861125, |
|
"learning_rate": 1.7744134007917194e-06, |
|
"loss": 0.7809, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.5330188679245284, |
|
"grad_norm": 0.8038089067621491, |
|
"learning_rate": 1.7719166509163208e-06, |
|
"loss": 0.752, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.535377358490566, |
|
"grad_norm": 0.8115733374512414, |
|
"learning_rate": 1.7694079373878433e-06, |
|
"loss": 0.7977, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.5377358490566038, |
|
"grad_norm": 0.8325145007355951, |
|
"learning_rate": 1.7668872990879173e-06, |
|
"loss": 0.775, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.5400943396226415, |
|
"grad_norm": 0.8029142353641874, |
|
"learning_rate": 1.7643547750829918e-06, |
|
"loss": 0.7897, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.5424528301886793, |
|
"grad_norm": 0.8068710232788708, |
|
"learning_rate": 1.7618104046237274e-06, |
|
"loss": 0.7869, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.5448113207547169, |
|
"grad_norm": 0.8179053896349096, |
|
"learning_rate": 1.7592542271443887e-06, |
|
"loss": 0.7737, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.5471698113207547, |
|
"grad_norm": 0.9236859332969564, |
|
"learning_rate": 1.7566862822622328e-06, |
|
"loss": 0.7779, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.5495283018867925, |
|
"grad_norm": 0.8105651817226325, |
|
"learning_rate": 1.754106609776896e-06, |
|
"loss": 0.7841, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.5518867924528302, |
|
"grad_norm": 0.8955884533729873, |
|
"learning_rate": 1.7515152496697763e-06, |
|
"loss": 0.8077, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.5542452830188679, |
|
"grad_norm": 1.015222493683735, |
|
"learning_rate": 1.748912242103413e-06, |
|
"loss": 0.8013, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.5566037735849056, |
|
"grad_norm": 0.9309116956021466, |
|
"learning_rate": 1.746297627420866e-06, |
|
"loss": 0.7672, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.5589622641509434, |
|
"grad_norm": 0.9599117437990355, |
|
"learning_rate": 1.7436714461450897e-06, |
|
"loss": 0.7957, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.5613207547169812, |
|
"grad_norm": 0.8340801449927656, |
|
"learning_rate": 1.7410337389783041e-06, |
|
"loss": 0.7994, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.5636792452830188, |
|
"grad_norm": 0.788362523807564, |
|
"learning_rate": 1.7383845468013654e-06, |
|
"loss": 0.7707, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.5660377358490566, |
|
"grad_norm": 0.8572171206767668, |
|
"learning_rate": 1.7357239106731317e-06, |
|
"loss": 0.7931, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.5683962264150944, |
|
"grad_norm": 4.180247592376519, |
|
"learning_rate": 1.733051871829826e-06, |
|
"loss": 0.798, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.5707547169811321, |
|
"grad_norm": 0.8738642652249169, |
|
"learning_rate": 1.7303684716843995e-06, |
|
"loss": 0.7768, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.5731132075471698, |
|
"grad_norm": 0.8413513234438594, |
|
"learning_rate": 1.727673751825886e-06, |
|
"loss": 0.8139, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.5754716981132075, |
|
"grad_norm": 0.8032830413975406, |
|
"learning_rate": 1.7249677540187609e-06, |
|
"loss": 0.7877, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.5778301886792453, |
|
"grad_norm": 0.7992641395875911, |
|
"learning_rate": 1.7222505202022913e-06, |
|
"loss": 0.7712, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.5801886792452831, |
|
"grad_norm": 0.8865998909749021, |
|
"learning_rate": 1.7195220924898882e-06, |
|
"loss": 0.7584, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.5825471698113207, |
|
"grad_norm": 0.8256395052056026, |
|
"learning_rate": 1.7167825131684511e-06, |
|
"loss": 0.7905, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.5849056603773585, |
|
"grad_norm": 0.8134932247816357, |
|
"learning_rate": 1.7140318246977163e-06, |
|
"loss": 0.7932, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.5872641509433962, |
|
"grad_norm": 0.8289362753913454, |
|
"learning_rate": 1.7112700697095953e-06, |
|
"loss": 0.7727, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.589622641509434, |
|
"grad_norm": 0.7919440339387878, |
|
"learning_rate": 1.7084972910075154e-06, |
|
"loss": 0.819, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.5919811320754716, |
|
"grad_norm": 0.7892796730141778, |
|
"learning_rate": 1.7057135315657567e-06, |
|
"loss": 0.7994, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 0.5943396226415094, |
|
"grad_norm": 0.818593844058535, |
|
"learning_rate": 1.7029188345287865e-06, |
|
"loss": 0.7833, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.5966981132075472, |
|
"grad_norm": 0.8550235074282425, |
|
"learning_rate": 1.7001132432105894e-06, |
|
"loss": 0.7627, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.5990566037735849, |
|
"grad_norm": 0.7825131867584777, |
|
"learning_rate": 1.6972968010939952e-06, |
|
"loss": 0.7863, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.6014150943396226, |
|
"grad_norm": 0.7872637454110324, |
|
"learning_rate": 1.6944695518300084e-06, |
|
"loss": 0.8202, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.6037735849056604, |
|
"grad_norm": 0.8264865629926671, |
|
"learning_rate": 1.6916315392371283e-06, |
|
"loss": 0.7794, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.6061320754716981, |
|
"grad_norm": 0.8112171041686048, |
|
"learning_rate": 1.688782807300671e-06, |
|
"loss": 0.7975, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 0.6084905660377359, |
|
"grad_norm": 0.8017599214256371, |
|
"learning_rate": 1.685923400172088e-06, |
|
"loss": 0.7821, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.6108490566037735, |
|
"grad_norm": 0.7855915924335556, |
|
"learning_rate": 1.683053362168282e-06, |
|
"loss": 0.7834, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 0.6132075471698113, |
|
"grad_norm": 0.8227864852025443, |
|
"learning_rate": 1.6801727377709191e-06, |
|
"loss": 0.7715, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.6155660377358491, |
|
"grad_norm": 0.8347676603852344, |
|
"learning_rate": 1.6772815716257411e-06, |
|
"loss": 0.7783, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.6179245283018868, |
|
"grad_norm": 0.847840576553053, |
|
"learning_rate": 1.6743799085418709e-06, |
|
"loss": 0.7796, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.6202830188679245, |
|
"grad_norm": 0.8187554443504805, |
|
"learning_rate": 1.671467793491121e-06, |
|
"loss": 0.8223, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 0.6226415094339622, |
|
"grad_norm": 0.8560599665847224, |
|
"learning_rate": 1.6685452716072942e-06, |
|
"loss": 0.7769, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.625, |
|
"grad_norm": 0.8756516850796994, |
|
"learning_rate": 1.6656123881854858e-06, |
|
"loss": 0.782, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.6273584905660378, |
|
"grad_norm": 0.861305723586375, |
|
"learning_rate": 1.6626691886813802e-06, |
|
"loss": 0.7907, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.6297169811320755, |
|
"grad_norm": 0.8565858633218812, |
|
"learning_rate": 1.6597157187105474e-06, |
|
"loss": 0.7586, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.6320754716981132, |
|
"grad_norm": 0.7796005038585514, |
|
"learning_rate": 1.6567520240477343e-06, |
|
"loss": 0.7996, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.6344339622641509, |
|
"grad_norm": 0.9853760454692906, |
|
"learning_rate": 1.6537781506261586e-06, |
|
"loss": 0.803, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 0.6367924528301887, |
|
"grad_norm": 0.8214336899114236, |
|
"learning_rate": 1.6507941445367934e-06, |
|
"loss": 0.7484, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.6391509433962265, |
|
"grad_norm": 0.8871384572382998, |
|
"learning_rate": 1.647800052027655e-06, |
|
"loss": 0.7876, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 0.6415094339622641, |
|
"grad_norm": 0.8564572450153076, |
|
"learning_rate": 1.6447959195030849e-06, |
|
"loss": 0.8054, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.6438679245283019, |
|
"grad_norm": 0.8408989291477506, |
|
"learning_rate": 1.6417817935230316e-06, |
|
"loss": 0.7668, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.6462264150943396, |
|
"grad_norm": 0.7945270947755096, |
|
"learning_rate": 1.6387577208023279e-06, |
|
"loss": 0.7893, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 0.6485849056603774, |
|
"grad_norm": 0.8844404937902336, |
|
"learning_rate": 1.6357237482099683e-06, |
|
"loss": 0.794, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.6509433962264151, |
|
"grad_norm": 0.8518070202862351, |
|
"learning_rate": 1.6326799227683803e-06, |
|
"loss": 0.7967, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.6533018867924528, |
|
"grad_norm": 0.819470621959587, |
|
"learning_rate": 1.6296262916526995e-06, |
|
"loss": 0.7633, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 0.6556603773584906, |
|
"grad_norm": 0.7989369195204531, |
|
"learning_rate": 1.626562902190034e-06, |
|
"loss": 0.807, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 0.6580188679245284, |
|
"grad_norm": 0.9412075176216419, |
|
"learning_rate": 1.6234898018587336e-06, |
|
"loss": 0.7752, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.660377358490566, |
|
"grad_norm": 0.7895838126966979, |
|
"learning_rate": 1.6204070382876531e-06, |
|
"loss": 0.7786, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.6627358490566038, |
|
"grad_norm": 0.7981016976096619, |
|
"learning_rate": 1.6173146592554152e-06, |
|
"loss": 0.7958, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 0.6650943396226415, |
|
"grad_norm": 0.8907323576889667, |
|
"learning_rate": 1.6142127126896679e-06, |
|
"loss": 0.807, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.6674528301886793, |
|
"grad_norm": 0.8053567442309949, |
|
"learning_rate": 1.6111012466663434e-06, |
|
"loss": 0.788, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 0.6698113207547169, |
|
"grad_norm": 0.8378739800286575, |
|
"learning_rate": 1.6079803094089125e-06, |
|
"loss": 0.736, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 0.6721698113207547, |
|
"grad_norm": 0.8858100985238717, |
|
"learning_rate": 1.6048499492876375e-06, |
|
"loss": 0.8106, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.6745283018867925, |
|
"grad_norm": 0.8276972316955101, |
|
"learning_rate": 1.6017102148188215e-06, |
|
"loss": 0.7835, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 0.6768867924528302, |
|
"grad_norm": 0.8853131510592382, |
|
"learning_rate": 1.598561154664058e-06, |
|
"loss": 0.7965, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 0.6792452830188679, |
|
"grad_norm": 0.8786321199638693, |
|
"learning_rate": 1.5954028176294746e-06, |
|
"loss": 0.8268, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.6816037735849056, |
|
"grad_norm": 0.8669272571254311, |
|
"learning_rate": 1.5922352526649801e-06, |
|
"loss": 0.7823, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 0.6839622641509434, |
|
"grad_norm": 0.9644127048111335, |
|
"learning_rate": 1.589058508863501e-06, |
|
"loss": 0.78, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.6863207547169812, |
|
"grad_norm": 0.8073310279373501, |
|
"learning_rate": 1.5858726354602248e-06, |
|
"loss": 0.7765, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 0.6886792452830188, |
|
"grad_norm": 0.8775721979922628, |
|
"learning_rate": 1.5826776818318348e-06, |
|
"loss": 0.7597, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 0.6910377358490566, |
|
"grad_norm": 1.015547726125361, |
|
"learning_rate": 1.5794736974957464e-06, |
|
"loss": 0.786, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 0.6933962264150944, |
|
"grad_norm": 0.7708601481727361, |
|
"learning_rate": 1.5762607321093366e-06, |
|
"loss": 0.7817, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.6957547169811321, |
|
"grad_norm": 0.8426421786890617, |
|
"learning_rate": 1.5730388354691785e-06, |
|
"loss": 0.7847, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.6981132075471698, |
|
"grad_norm": 0.7855014020345946, |
|
"learning_rate": 1.569808057510266e-06, |
|
"loss": 0.7748, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 0.7004716981132075, |
|
"grad_norm": 0.8389890263875862, |
|
"learning_rate": 1.5665684483052424e-06, |
|
"loss": 0.7856, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 0.7028301886792453, |
|
"grad_norm": 0.8482309028462507, |
|
"learning_rate": 1.563320058063622e-06, |
|
"loss": 0.7469, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 0.7051886792452831, |
|
"grad_norm": 1.0888639918368348, |
|
"learning_rate": 1.5600629371310144e-06, |
|
"loss": 0.7905, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 0.7075471698113207, |
|
"grad_norm": 0.8011947485403833, |
|
"learning_rate": 1.556797135988342e-06, |
|
"loss": 0.7902, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.7099056603773585, |
|
"grad_norm": 0.7584432625187374, |
|
"learning_rate": 1.5535227052510588e-06, |
|
"loss": 0.7935, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 0.7122641509433962, |
|
"grad_norm": 1.1353449141437797, |
|
"learning_rate": 1.5502396956683667e-06, |
|
"loss": 0.7992, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 0.714622641509434, |
|
"grad_norm": 0.8144382945981636, |
|
"learning_rate": 1.5469481581224271e-06, |
|
"loss": 0.761, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 0.7169811320754716, |
|
"grad_norm": 0.9448262210280223, |
|
"learning_rate": 1.5436481436275724e-06, |
|
"loss": 0.7839, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 0.7193396226415094, |
|
"grad_norm": 0.8145609006654517, |
|
"learning_rate": 1.5403397033295179e-06, |
|
"loss": 0.7975, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.7216981132075472, |
|
"grad_norm": 0.9825256520905712, |
|
"learning_rate": 1.537022888504566e-06, |
|
"loss": 0.7918, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 0.7240566037735849, |
|
"grad_norm": 0.9647183394075436, |
|
"learning_rate": 1.5336977505588134e-06, |
|
"loss": 0.7973, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 0.7264150943396226, |
|
"grad_norm": 0.802175509091533, |
|
"learning_rate": 1.530364341027354e-06, |
|
"loss": 0.7928, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 0.7287735849056604, |
|
"grad_norm": 0.8122620856594104, |
|
"learning_rate": 1.5270227115734789e-06, |
|
"loss": 0.7767, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 0.7311320754716981, |
|
"grad_norm": 0.8656607812881103, |
|
"learning_rate": 1.5236729139878778e-06, |
|
"loss": 0.7746, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.7334905660377359, |
|
"grad_norm": 0.8721224552709389, |
|
"learning_rate": 1.5203150001878353e-06, |
|
"loss": 0.7929, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 0.7358490566037735, |
|
"grad_norm": 0.7856050732773205, |
|
"learning_rate": 1.5169490222164254e-06, |
|
"loss": 0.7733, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 0.7382075471698113, |
|
"grad_norm": 0.8639502364522528, |
|
"learning_rate": 1.5135750322417066e-06, |
|
"loss": 0.7958, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 0.7405660377358491, |
|
"grad_norm": 0.9986722246195658, |
|
"learning_rate": 1.5101930825559124e-06, |
|
"loss": 0.7944, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 0.7429245283018868, |
|
"grad_norm": 0.8861549532629037, |
|
"learning_rate": 1.50680322557464e-06, |
|
"loss": 0.7887, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.7452830188679245, |
|
"grad_norm": 0.8526619540985027, |
|
"learning_rate": 1.5034055138360398e-06, |
|
"loss": 0.7388, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 0.7476415094339622, |
|
"grad_norm": 0.8760084110715984, |
|
"learning_rate": 1.5e-06, |
|
"loss": 0.7719, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 0.8038573705056793, |
|
"learning_rate": 1.4965867368473306e-06, |
|
"loss": 0.7514, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"eval_loss": 0.6982797384262085, |
|
"eval_runtime": 83.2645, |
|
"eval_samples_per_second": 8.467, |
|
"eval_steps_per_second": 0.54, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 0.7523584905660378, |
|
"grad_norm": 0.8193168846563599, |
|
"learning_rate": 1.4931657772789457e-06, |
|
"loss": 0.7834, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 0.7547169811320755, |
|
"grad_norm": 0.9092282687501353, |
|
"learning_rate": 1.4897371743150423e-06, |
|
"loss": 0.753, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.7570754716981132, |
|
"grad_norm": 0.944599516853919, |
|
"learning_rate": 1.4863009810942813e-06, |
|
"loss": 0.7855, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 0.7594339622641509, |
|
"grad_norm": 0.9097780581691608, |
|
"learning_rate": 1.4828572508729606e-06, |
|
"loss": 0.7837, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 0.7617924528301887, |
|
"grad_norm": 0.8635246603088246, |
|
"learning_rate": 1.479406037024192e-06, |
|
"loss": 0.7862, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 0.7641509433962265, |
|
"grad_norm": 0.8041740773336183, |
|
"learning_rate": 1.4759473930370736e-06, |
|
"loss": 0.7705, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 0.7665094339622641, |
|
"grad_norm": 0.8563734416095982, |
|
"learning_rate": 1.4724813725158596e-06, |
|
"loss": 0.7944, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.7688679245283019, |
|
"grad_norm": 0.7885016820225573, |
|
"learning_rate": 1.4690080291791312e-06, |
|
"loss": 0.78, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 0.7712264150943396, |
|
"grad_norm": 0.9769999602785767, |
|
"learning_rate": 1.4655274168589633e-06, |
|
"loss": 0.7506, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 0.7735849056603774, |
|
"grad_norm": 0.8241115466345136, |
|
"learning_rate": 1.462039589500089e-06, |
|
"loss": 0.7826, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 0.7759433962264151, |
|
"grad_norm": 0.8261080001645498, |
|
"learning_rate": 1.4585446011590658e-06, |
|
"loss": 0.7592, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 0.7783018867924528, |
|
"grad_norm": 1.1493611817217861, |
|
"learning_rate": 1.4550425060034365e-06, |
|
"loss": 0.7674, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.7806603773584906, |
|
"grad_norm": 0.8343201494559767, |
|
"learning_rate": 1.4515333583108893e-06, |
|
"loss": 0.8072, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 0.7830188679245284, |
|
"grad_norm": 1.0664747829909942, |
|
"learning_rate": 1.4480172124684172e-06, |
|
"loss": 0.7802, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 0.785377358490566, |
|
"grad_norm": 0.7792110149695123, |
|
"learning_rate": 1.4444941229714758e-06, |
|
"loss": 0.7704, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 0.7877358490566038, |
|
"grad_norm": 0.8718279852088779, |
|
"learning_rate": 1.440964144423136e-06, |
|
"loss": 0.7963, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 0.7900943396226415, |
|
"grad_norm": 1.02769301319547, |
|
"learning_rate": 1.4374273315332415e-06, |
|
"loss": 0.7863, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.7924528301886793, |
|
"grad_norm": 0.7843647199877115, |
|
"learning_rate": 1.433883739117558e-06, |
|
"loss": 0.7705, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 0.7948113207547169, |
|
"grad_norm": 0.815818120801617, |
|
"learning_rate": 1.430333422096925e-06, |
|
"loss": 0.7651, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 0.7971698113207547, |
|
"grad_norm": 0.7835815188776669, |
|
"learning_rate": 1.4267764354964037e-06, |
|
"loss": 0.789, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 0.7995283018867925, |
|
"grad_norm": 0.8382956450308263, |
|
"learning_rate": 1.423212834444425e-06, |
|
"loss": 0.817, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 0.8018867924528302, |
|
"grad_norm": 0.9606287069411898, |
|
"learning_rate": 1.4196426741719345e-06, |
|
"loss": 0.7801, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.8042452830188679, |
|
"grad_norm": 0.7833416575809102, |
|
"learning_rate": 1.4160660100115373e-06, |
|
"loss": 0.8139, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 0.8066037735849056, |
|
"grad_norm": 0.8112490810199738, |
|
"learning_rate": 1.4124828973966392e-06, |
|
"loss": 0.7951, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 0.8089622641509434, |
|
"grad_norm": 0.8533925894384254, |
|
"learning_rate": 1.4088933918605887e-06, |
|
"loss": 0.7555, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 0.8113207547169812, |
|
"grad_norm": 0.8917123790523681, |
|
"learning_rate": 1.405297549035816e-06, |
|
"loss": 0.7941, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 0.8136792452830188, |
|
"grad_norm": 0.8690146736748873, |
|
"learning_rate": 1.4016954246529694e-06, |
|
"loss": 0.7966, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.8160377358490566, |
|
"grad_norm": 0.8372555162768984, |
|
"learning_rate": 1.3980870745400544e-06, |
|
"loss": 0.7972, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 0.8183962264150944, |
|
"grad_norm": 0.8356712411477044, |
|
"learning_rate": 1.3944725546215662e-06, |
|
"loss": 0.753, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 0.8207547169811321, |
|
"grad_norm": 0.8366058600346219, |
|
"learning_rate": 1.3908519209176225e-06, |
|
"loss": 0.7813, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 0.8231132075471698, |
|
"grad_norm": 0.8373524566281616, |
|
"learning_rate": 1.3872252295430986e-06, |
|
"loss": 0.7956, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 0.8254716981132075, |
|
"grad_norm": 0.7833296789989939, |
|
"learning_rate": 1.3835925367067529e-06, |
|
"loss": 0.7449, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.8278301886792453, |
|
"grad_norm": 0.7981265071211454, |
|
"learning_rate": 1.37995389871036e-06, |
|
"loss": 0.7756, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 0.8301886792452831, |
|
"grad_norm": 0.8840540972881347, |
|
"learning_rate": 1.3763093719478357e-06, |
|
"loss": 0.7984, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 0.8325471698113207, |
|
"grad_norm": 0.985313340190558, |
|
"learning_rate": 1.372659012904363e-06, |
|
"loss": 0.7821, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 0.8349056603773585, |
|
"grad_norm": 0.8138177618630202, |
|
"learning_rate": 1.369002878155519e-06, |
|
"loss": 0.7618, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 0.8372641509433962, |
|
"grad_norm": 0.7948448582342686, |
|
"learning_rate": 1.3653410243663951e-06, |
|
"loss": 0.7559, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.839622641509434, |
|
"grad_norm": 0.8151246949343306, |
|
"learning_rate": 1.3616735082907196e-06, |
|
"loss": 0.7722, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 0.8419811320754716, |
|
"grad_norm": 0.8162887157403834, |
|
"learning_rate": 1.35800038676998e-06, |
|
"loss": 0.7711, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 0.8443396226415094, |
|
"grad_norm": 1.1322690706169278, |
|
"learning_rate": 1.3543217167325388e-06, |
|
"loss": 0.762, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 0.8466981132075472, |
|
"grad_norm": 0.8102205098745584, |
|
"learning_rate": 1.3506375551927544e-06, |
|
"loss": 0.7532, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 0.8490566037735849, |
|
"grad_norm": 0.9369994865443904, |
|
"learning_rate": 1.3469479592500951e-06, |
|
"loss": 0.737, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.8514150943396226, |
|
"grad_norm": 0.9319609634032745, |
|
"learning_rate": 1.3432529860882556e-06, |
|
"loss": 0.8074, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 0.8537735849056604, |
|
"grad_norm": 0.844911038169109, |
|
"learning_rate": 1.3395526929742691e-06, |
|
"loss": 0.7576, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 0.8561320754716981, |
|
"grad_norm": 0.8213852435481517, |
|
"learning_rate": 1.3358471372576227e-06, |
|
"loss": 0.8091, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 0.8584905660377359, |
|
"grad_norm": 0.8679917359639049, |
|
"learning_rate": 1.3321363763693643e-06, |
|
"loss": 0.7649, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 0.8608490566037735, |
|
"grad_norm": 0.863728546344944, |
|
"learning_rate": 1.3284204678212167e-06, |
|
"loss": 0.7364, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.8632075471698113, |
|
"grad_norm": 0.9151925878031607, |
|
"learning_rate": 1.3246994692046835e-06, |
|
"loss": 0.7617, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 0.8655660377358491, |
|
"grad_norm": 0.7676651686086633, |
|
"learning_rate": 1.3209734381901578e-06, |
|
"loss": 0.7703, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 0.8679245283018868, |
|
"grad_norm": 0.79154626898034, |
|
"learning_rate": 1.3172424325260272e-06, |
|
"loss": 0.7816, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 0.8702830188679245, |
|
"grad_norm": 0.8831690649426513, |
|
"learning_rate": 1.3135065100377814e-06, |
|
"loss": 0.7646, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 0.8726415094339622, |
|
"grad_norm": 0.8098189405568664, |
|
"learning_rate": 1.3097657286271116e-06, |
|
"loss": 0.7673, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.875, |
|
"grad_norm": 0.8071739591609438, |
|
"learning_rate": 1.3060201462710176e-06, |
|
"loss": 0.7438, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 0.8773584905660378, |
|
"grad_norm": 0.8864737282217798, |
|
"learning_rate": 1.3022698210209066e-06, |
|
"loss": 0.7393, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 0.8797169811320755, |
|
"grad_norm": 0.8193555244410853, |
|
"learning_rate": 1.2985148110016947e-06, |
|
"loss": 0.7658, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 0.8820754716981132, |
|
"grad_norm": 0.7819668232955883, |
|
"learning_rate": 1.2947551744109043e-06, |
|
"loss": 0.7876, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 0.8844339622641509, |
|
"grad_norm": 0.9661764152661253, |
|
"learning_rate": 1.2909909695177645e-06, |
|
"loss": 0.7556, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.8867924528301887, |
|
"grad_norm": 0.7916178501632968, |
|
"learning_rate": 1.2872222546623065e-06, |
|
"loss": 0.7776, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 0.8891509433962265, |
|
"grad_norm": 0.908391562439544, |
|
"learning_rate": 1.2834490882544598e-06, |
|
"loss": 0.7604, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 0.8915094339622641, |
|
"grad_norm": 0.7673617521747667, |
|
"learning_rate": 1.2796715287731461e-06, |
|
"loss": 0.7786, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 0.8938679245283019, |
|
"grad_norm": 1.1737766347467777, |
|
"learning_rate": 1.2758896347653752e-06, |
|
"loss": 0.7692, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 0.8962264150943396, |
|
"grad_norm": 0.7823714073715833, |
|
"learning_rate": 1.272103464845335e-06, |
|
"loss": 0.7389, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.8985849056603774, |
|
"grad_norm": 1.0030453246218272, |
|
"learning_rate": 1.2683130776934848e-06, |
|
"loss": 0.7556, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 0.9009433962264151, |
|
"grad_norm": 0.9688917314182516, |
|
"learning_rate": 1.2645185320556444e-06, |
|
"loss": 0.7415, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 0.9033018867924528, |
|
"grad_norm": 0.8539895113482738, |
|
"learning_rate": 1.2607198867420858e-06, |
|
"loss": 0.7772, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 0.9056603773584906, |
|
"grad_norm": 0.8011530846319137, |
|
"learning_rate": 1.256917200626619e-06, |
|
"loss": 0.7426, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 0.9080188679245284, |
|
"grad_norm": 0.9859270894449156, |
|
"learning_rate": 1.253110532645682e-06, |
|
"loss": 0.7942, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.910377358490566, |
|
"grad_norm": 0.7970067775774904, |
|
"learning_rate": 1.2492999417974253e-06, |
|
"loss": 0.7914, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 0.9127358490566038, |
|
"grad_norm": 0.9049836680757578, |
|
"learning_rate": 1.245485487140799e-06, |
|
"loss": 0.7787, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 0.9150943396226415, |
|
"grad_norm": 0.866322963165458, |
|
"learning_rate": 1.2416672277946373e-06, |
|
"loss": 0.781, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 0.9174528301886793, |
|
"grad_norm": 0.9269306886834423, |
|
"learning_rate": 1.2378452229367413e-06, |
|
"loss": 0.808, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 0.9198113207547169, |
|
"grad_norm": 0.8286820002413788, |
|
"learning_rate": 1.2340195318029622e-06, |
|
"loss": 0.7623, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.9221698113207547, |
|
"grad_norm": 0.8143258703482895, |
|
"learning_rate": 1.2301902136862848e-06, |
|
"loss": 0.7965, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 0.9245283018867925, |
|
"grad_norm": 0.8041175896213196, |
|
"learning_rate": 1.2263573279359053e-06, |
|
"loss": 0.7613, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 0.9268867924528302, |
|
"grad_norm": 0.9087369855701409, |
|
"learning_rate": 1.2225209339563143e-06, |
|
"loss": 0.7669, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 0.9292452830188679, |
|
"grad_norm": 0.8126237661274147, |
|
"learning_rate": 1.2186810912063758e-06, |
|
"loss": 0.7531, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 0.9316037735849056, |
|
"grad_norm": 0.8354891878817764, |
|
"learning_rate": 1.214837859198404e-06, |
|
"loss": 0.768, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.9339622641509434, |
|
"grad_norm": 0.7793219061787805, |
|
"learning_rate": 1.2109912974972422e-06, |
|
"loss": 0.7831, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 0.9363207547169812, |
|
"grad_norm": 0.7813175142444524, |
|
"learning_rate": 1.2071414657193396e-06, |
|
"loss": 0.7671, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 0.9386792452830188, |
|
"grad_norm": 0.8960835581240535, |
|
"learning_rate": 1.2032884235318268e-06, |
|
"loss": 0.775, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 0.9410377358490566, |
|
"grad_norm": 0.8773817072709057, |
|
"learning_rate": 1.1994322306515925e-06, |
|
"loss": 0.7915, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 0.9433962264150944, |
|
"grad_norm": 0.8191284697051531, |
|
"learning_rate": 1.1955729468443546e-06, |
|
"loss": 0.7548, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.9457547169811321, |
|
"grad_norm": 0.8704070228218153, |
|
"learning_rate": 1.1917106319237384e-06, |
|
"loss": 0.773, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 0.9481132075471698, |
|
"grad_norm": 0.8119790656590767, |
|
"learning_rate": 1.1878453457503464e-06, |
|
"loss": 0.7933, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 0.9504716981132075, |
|
"grad_norm": 0.8290570173312977, |
|
"learning_rate": 1.1839771482308308e-06, |
|
"loss": 0.7515, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 0.9528301886792453, |
|
"grad_norm": 0.7808059259378133, |
|
"learning_rate": 1.1801060993169666e-06, |
|
"loss": 0.771, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 0.9551886792452831, |
|
"grad_norm": 0.8554505156061943, |
|
"learning_rate": 1.1762322590047219e-06, |
|
"loss": 0.767, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.9575471698113207, |
|
"grad_norm": 0.8132652328205165, |
|
"learning_rate": 1.172355687333326e-06, |
|
"loss": 0.7467, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 0.9599056603773585, |
|
"grad_norm": 0.7767004491954446, |
|
"learning_rate": 1.168476444384342e-06, |
|
"loss": 0.7667, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 0.9622641509433962, |
|
"grad_norm": 0.8619052365176207, |
|
"learning_rate": 1.164594590280734e-06, |
|
"loss": 0.7993, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 0.964622641509434, |
|
"grad_norm": 0.8115144668833016, |
|
"learning_rate": 1.1607101851859345e-06, |
|
"loss": 0.7787, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 0.9669811320754716, |
|
"grad_norm": 0.8000429071733963, |
|
"learning_rate": 1.156823289302914e-06, |
|
"loss": 0.7767, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.9693396226415094, |
|
"grad_norm": 0.8228751842965996, |
|
"learning_rate": 1.152933962873246e-06, |
|
"loss": 0.7829, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 0.9716981132075472, |
|
"grad_norm": 0.8685312714228164, |
|
"learning_rate": 1.1490422661761743e-06, |
|
"loss": 0.7775, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 0.9740566037735849, |
|
"grad_norm": 0.9383805515060808, |
|
"learning_rate": 1.1451482595276796e-06, |
|
"loss": 0.7716, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 0.9764150943396226, |
|
"grad_norm": 0.978907741116596, |
|
"learning_rate": 1.1412520032795419e-06, |
|
"loss": 0.79, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 0.9787735849056604, |
|
"grad_norm": 0.936909408413621, |
|
"learning_rate": 1.1373535578184082e-06, |
|
"loss": 0.7463, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.9811320754716981, |
|
"grad_norm": 0.7779360289097986, |
|
"learning_rate": 1.1334529835648552e-06, |
|
"loss": 0.7544, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 0.9834905660377359, |
|
"grad_norm": 0.8442404415567627, |
|
"learning_rate": 1.1295503409724525e-06, |
|
"loss": 0.7968, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 0.9858490566037735, |
|
"grad_norm": 0.8856697438669298, |
|
"learning_rate": 1.1256456905268263e-06, |
|
"loss": 0.7622, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 0.9882075471698113, |
|
"grad_norm": 0.8362007813195715, |
|
"learning_rate": 1.1217390927447225e-06, |
|
"loss": 0.7475, |
|
"step": 419 |
|
}, |
|
{ |
|
"epoch": 0.9905660377358491, |
|
"grad_norm": 0.7824485352158383, |
|
"learning_rate": 1.1178306081730664e-06, |
|
"loss": 0.7705, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.9929245283018868, |
|
"grad_norm": 0.7992205810856007, |
|
"learning_rate": 1.113920297388028e-06, |
|
"loss": 0.7545, |
|
"step": 421 |
|
}, |
|
{ |
|
"epoch": 0.9952830188679245, |
|
"grad_norm": 0.785520038953663, |
|
"learning_rate": 1.1100082209940793e-06, |
|
"loss": 0.7647, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 0.9976415094339622, |
|
"grad_norm": 0.8327122327201606, |
|
"learning_rate": 1.106094439623058e-06, |
|
"loss": 0.7632, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.7736149741657676, |
|
"learning_rate": 1.1021790139332264e-06, |
|
"loss": 0.7861, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 0.6892534494400024, |
|
"eval_runtime": 81.9429, |
|
"eval_samples_per_second": 8.604, |
|
"eval_steps_per_second": 0.549, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 1.0023584905660377, |
|
"grad_norm": 0.8035685998252425, |
|
"learning_rate": 1.0982620046083305e-06, |
|
"loss": 0.7506, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 1.0047169811320755, |
|
"grad_norm": 1.001273686570143, |
|
"learning_rate": 1.0943434723566623e-06, |
|
"loss": 0.783, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 1.0070754716981132, |
|
"grad_norm": 0.7620350953234064, |
|
"learning_rate": 1.0904234779101158e-06, |
|
"loss": 0.7436, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 1.009433962264151, |
|
"grad_norm": 0.8051824532909724, |
|
"learning_rate": 1.0865020820232466e-06, |
|
"loss": 0.7408, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 1.0117924528301887, |
|
"grad_norm": 1.1441709749307019, |
|
"learning_rate": 1.0825793454723324e-06, |
|
"loss": 0.7619, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 1.0141509433962264, |
|
"grad_norm": 0.934287576351282, |
|
"learning_rate": 1.0786553290544278e-06, |
|
"loss": 0.7291, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 1.0165094339622642, |
|
"grad_norm": 0.867146390883491, |
|
"learning_rate": 1.0747300935864243e-06, |
|
"loss": 0.7697, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 1.0188679245283019, |
|
"grad_norm": 0.7934743429540405, |
|
"learning_rate": 1.070803699904107e-06, |
|
"loss": 0.7774, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 1.0212264150943395, |
|
"grad_norm": 1.204890281675565, |
|
"learning_rate": 1.0668762088612114e-06, |
|
"loss": 0.7659, |
|
"step": 433 |
|
}, |
|
{ |
|
"epoch": 1.0235849056603774, |
|
"grad_norm": 0.7748660244184836, |
|
"learning_rate": 1.0629476813284807e-06, |
|
"loss": 0.7812, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 1.025943396226415, |
|
"grad_norm": 0.8559552505505363, |
|
"learning_rate": 1.0590181781927227e-06, |
|
"loss": 0.7641, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 1.028301886792453, |
|
"grad_norm": 0.8278708288686452, |
|
"learning_rate": 1.0550877603558654e-06, |
|
"loss": 0.7339, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 1.0306603773584906, |
|
"grad_norm": 0.8199114258570241, |
|
"learning_rate": 1.0511564887340135e-06, |
|
"loss": 0.7884, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 1.0330188679245282, |
|
"grad_norm": 0.8220578039981541, |
|
"learning_rate": 1.0472244242565034e-06, |
|
"loss": 0.7636, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 1.0353773584905661, |
|
"grad_norm": 0.9899659514811034, |
|
"learning_rate": 1.043291627864961e-06, |
|
"loss": 0.7507, |
|
"step": 439 |
|
}, |
|
{ |
|
"epoch": 1.0377358490566038, |
|
"grad_norm": 0.9164260231900845, |
|
"learning_rate": 1.0393581605123552e-06, |
|
"loss": 0.7604, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 1.0400943396226414, |
|
"grad_norm": 0.780514004370331, |
|
"learning_rate": 1.035424083162054e-06, |
|
"loss": 0.7661, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 1.0424528301886793, |
|
"grad_norm": 0.9654942156178127, |
|
"learning_rate": 1.031489456786879e-06, |
|
"loss": 0.732, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 1.044811320754717, |
|
"grad_norm": 0.9028629851443349, |
|
"learning_rate": 1.0275543423681621e-06, |
|
"loss": 0.7704, |
|
"step": 443 |
|
}, |
|
{ |
|
"epoch": 1.0471698113207548, |
|
"grad_norm": 0.8997990850767794, |
|
"learning_rate": 1.0236188008947978e-06, |
|
"loss": 0.7832, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 1.0495283018867925, |
|
"grad_norm": 0.8380399486601815, |
|
"learning_rate": 1.0196828933623008e-06, |
|
"loss": 0.7513, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 1.0518867924528301, |
|
"grad_norm": 0.818805158462816, |
|
"learning_rate": 1.0157466807718577e-06, |
|
"loss": 0.7595, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 1.054245283018868, |
|
"grad_norm": 0.8325448461017404, |
|
"learning_rate": 1.0118102241293847e-06, |
|
"loss": 0.7634, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 1.0566037735849056, |
|
"grad_norm": 0.7936938025099423, |
|
"learning_rate": 1.0078735844445788e-06, |
|
"loss": 0.768, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 1.0589622641509433, |
|
"grad_norm": 0.8378311364099181, |
|
"learning_rate": 1.0039368227299753e-06, |
|
"loss": 0.7599, |
|
"step": 449 |
|
}, |
|
{ |
|
"epoch": 1.0613207547169812, |
|
"grad_norm": 0.8299911067496856, |
|
"learning_rate": 1e-06, |
|
"loss": 0.7488, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 1.0636792452830188, |
|
"grad_norm": 0.8677844259185759, |
|
"learning_rate": 9.960631772700248e-07, |
|
"loss": 0.7698, |
|
"step": 451 |
|
}, |
|
{ |
|
"epoch": 1.0660377358490567, |
|
"grad_norm": 0.8463491375440975, |
|
"learning_rate": 9.92126415555421e-07, |
|
"loss": 0.7493, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 1.0683962264150944, |
|
"grad_norm": 0.8004441596711467, |
|
"learning_rate": 9.881897758706154e-07, |
|
"loss": 0.7366, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 1.070754716981132, |
|
"grad_norm": 1.2321632539336993, |
|
"learning_rate": 9.842533192281422e-07, |
|
"loss": 0.7481, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 1.0731132075471699, |
|
"grad_norm": 0.8194799035445371, |
|
"learning_rate": 9.803171066376995e-07, |
|
"loss": 0.7834, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 1.0754716981132075, |
|
"grad_norm": 0.7953471718099565, |
|
"learning_rate": 9.763811991052019e-07, |
|
"loss": 0.7508, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 1.0778301886792452, |
|
"grad_norm": 0.8500817934393053, |
|
"learning_rate": 9.72445657631838e-07, |
|
"loss": 0.7639, |
|
"step": 457 |
|
}, |
|
{ |
|
"epoch": 1.080188679245283, |
|
"grad_norm": 0.8053711636082244, |
|
"learning_rate": 9.68510543213121e-07, |
|
"loss": 0.7546, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 1.0825471698113207, |
|
"grad_norm": 1.1272698654628237, |
|
"learning_rate": 9.645759168379461e-07, |
|
"loss": 0.735, |
|
"step": 459 |
|
}, |
|
{ |
|
"epoch": 1.0849056603773586, |
|
"grad_norm": 0.7772007598787661, |
|
"learning_rate": 9.606418394876447e-07, |
|
"loss": 0.747, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 1.0872641509433962, |
|
"grad_norm": 0.8414360815423256, |
|
"learning_rate": 9.567083721350388e-07, |
|
"loss": 0.7411, |
|
"step": 461 |
|
}, |
|
{ |
|
"epoch": 1.0896226415094339, |
|
"grad_norm": 0.7791267607388218, |
|
"learning_rate": 9.527755757434966e-07, |
|
"loss": 0.7436, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 1.0919811320754718, |
|
"grad_norm": 0.80528466203603, |
|
"learning_rate": 9.48843511265987e-07, |
|
"loss": 0.7656, |
|
"step": 463 |
|
}, |
|
{ |
|
"epoch": 1.0943396226415094, |
|
"grad_norm": 0.7758351013188943, |
|
"learning_rate": 9.449122396441343e-07, |
|
"loss": 0.737, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 1.0966981132075473, |
|
"grad_norm": 1.2494162832647924, |
|
"learning_rate": 9.409818218072772e-07, |
|
"loss": 0.7613, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 1.099056603773585, |
|
"grad_norm": 0.8049303948390758, |
|
"learning_rate": 9.370523186715194e-07, |
|
"loss": 0.7628, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 1.1014150943396226, |
|
"grad_norm": 0.7896423211268341, |
|
"learning_rate": 9.331237911387888e-07, |
|
"loss": 0.7301, |
|
"step": 467 |
|
}, |
|
{ |
|
"epoch": 1.1037735849056605, |
|
"grad_norm": 0.7947232451183359, |
|
"learning_rate": 9.291963000958931e-07, |
|
"loss": 0.7803, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 1.1061320754716981, |
|
"grad_norm": 0.8128944094386984, |
|
"learning_rate": 9.252699064135758e-07, |
|
"loss": 0.7693, |
|
"step": 469 |
|
}, |
|
{ |
|
"epoch": 1.1084905660377358, |
|
"grad_norm": 0.9181658602259837, |
|
"learning_rate": 9.213446709455722e-07, |
|
"loss": 0.7794, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 1.1108490566037736, |
|
"grad_norm": 0.753282034988688, |
|
"learning_rate": 9.174206545276677e-07, |
|
"loss": 0.7383, |
|
"step": 471 |
|
}, |
|
{ |
|
"epoch": 1.1132075471698113, |
|
"grad_norm": 0.8890774956158611, |
|
"learning_rate": 9.134979179767532e-07, |
|
"loss": 0.7528, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 1.115566037735849, |
|
"grad_norm": 0.9503365963029163, |
|
"learning_rate": 9.095765220898843e-07, |
|
"loss": 0.7757, |
|
"step": 473 |
|
}, |
|
{ |
|
"epoch": 1.1179245283018868, |
|
"grad_norm": 0.9284151146362815, |
|
"learning_rate": 9.056565276433377e-07, |
|
"loss": 0.7442, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 1.1202830188679245, |
|
"grad_norm": 0.9704301526052668, |
|
"learning_rate": 9.017379953916695e-07, |
|
"loss": 0.7417, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 1.1226415094339623, |
|
"grad_norm": 0.8572079632472874, |
|
"learning_rate": 8.978209860667738e-07, |
|
"loss": 0.7724, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 1.125, |
|
"grad_norm": 0.7912726011032718, |
|
"learning_rate": 8.93905560376942e-07, |
|
"loss": 0.7741, |
|
"step": 477 |
|
}, |
|
{ |
|
"epoch": 1.1273584905660377, |
|
"grad_norm": 0.7762605349979439, |
|
"learning_rate": 8.899917790059207e-07, |
|
"loss": 0.7211, |
|
"step": 478 |
|
}, |
|
{ |
|
"epoch": 1.1297169811320755, |
|
"grad_norm": 0.8625085844504403, |
|
"learning_rate": 8.860797026119721e-07, |
|
"loss": 0.7327, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 1.1320754716981132, |
|
"grad_norm": 0.8012017130865744, |
|
"learning_rate": 8.821693918269333e-07, |
|
"loss": 0.7293, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 1.134433962264151, |
|
"grad_norm": 0.7832218701492647, |
|
"learning_rate": 8.782609072552776e-07, |
|
"loss": 0.7584, |
|
"step": 481 |
|
}, |
|
{ |
|
"epoch": 1.1367924528301887, |
|
"grad_norm": 0.8223365605395953, |
|
"learning_rate": 8.743543094731737e-07, |
|
"loss": 0.757, |
|
"step": 482 |
|
}, |
|
{ |
|
"epoch": 1.1391509433962264, |
|
"grad_norm": 0.769269936037437, |
|
"learning_rate": 8.704496590275477e-07, |
|
"loss": 0.7265, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 1.1415094339622642, |
|
"grad_norm": 0.8470727703331676, |
|
"learning_rate": 8.665470164351447e-07, |
|
"loss": 0.7446, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 1.1438679245283019, |
|
"grad_norm": 0.8609116334367811, |
|
"learning_rate": 8.626464421815918e-07, |
|
"loss": 0.7357, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 1.1462264150943395, |
|
"grad_norm": 0.7644474626055862, |
|
"learning_rate": 8.587479967204582e-07, |
|
"loss": 0.7502, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 1.1485849056603774, |
|
"grad_norm": 0.7974217428241436, |
|
"learning_rate": 8.548517404723206e-07, |
|
"loss": 0.7255, |
|
"step": 487 |
|
}, |
|
{ |
|
"epoch": 1.150943396226415, |
|
"grad_norm": 0.8124702999828396, |
|
"learning_rate": 8.509577338238254e-07, |
|
"loss": 0.766, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 1.1533018867924527, |
|
"grad_norm": 0.8292245412487236, |
|
"learning_rate": 8.47066037126754e-07, |
|
"loss": 0.7302, |
|
"step": 489 |
|
}, |
|
{ |
|
"epoch": 1.1556603773584906, |
|
"grad_norm": 0.819810501024351, |
|
"learning_rate": 8.43176710697086e-07, |
|
"loss": 0.7163, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 1.1580188679245282, |
|
"grad_norm": 0.8385443524556274, |
|
"learning_rate": 8.392898148140656e-07, |
|
"loss": 0.7465, |
|
"step": 491 |
|
}, |
|
{ |
|
"epoch": 1.1603773584905661, |
|
"grad_norm": 0.778726751410171, |
|
"learning_rate": 8.354054097192659e-07, |
|
"loss": 0.7758, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 1.1627358490566038, |
|
"grad_norm": 0.788415584362694, |
|
"learning_rate": 8.315235556156579e-07, |
|
"loss": 0.7498, |
|
"step": 493 |
|
}, |
|
{ |
|
"epoch": 1.1650943396226414, |
|
"grad_norm": 0.7703907755533744, |
|
"learning_rate": 8.276443126666742e-07, |
|
"loss": 0.7276, |
|
"step": 494 |
|
}, |
|
{ |
|
"epoch": 1.1674528301886793, |
|
"grad_norm": 0.8704987882807484, |
|
"learning_rate": 8.237677409952784e-07, |
|
"loss": 0.766, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 1.169811320754717, |
|
"grad_norm": 0.7948667823022109, |
|
"learning_rate": 8.198939006830333e-07, |
|
"loss": 0.7656, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 1.1721698113207548, |
|
"grad_norm": 0.8090640349585008, |
|
"learning_rate": 8.160228517691692e-07, |
|
"loss": 0.7257, |
|
"step": 497 |
|
}, |
|
{ |
|
"epoch": 1.1745283018867925, |
|
"grad_norm": 1.159992161506865, |
|
"learning_rate": 8.121546542496538e-07, |
|
"loss": 0.7593, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 1.1768867924528301, |
|
"grad_norm": 0.8230774285047441, |
|
"learning_rate": 8.082893680762618e-07, |
|
"loss": 0.7511, |
|
"step": 499 |
|
}, |
|
{ |
|
"epoch": 1.179245283018868, |
|
"grad_norm": 0.8682718782355334, |
|
"learning_rate": 8.044270531556452e-07, |
|
"loss": 0.7647, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.1816037735849056, |
|
"grad_norm": 0.8274347964467509, |
|
"learning_rate": 8.005677693484076e-07, |
|
"loss": 0.7536, |
|
"step": 501 |
|
}, |
|
{ |
|
"epoch": 1.1839622641509433, |
|
"grad_norm": 0.8334334904558071, |
|
"learning_rate": 7.967115764681731e-07, |
|
"loss": 0.782, |
|
"step": 502 |
|
}, |
|
{ |
|
"epoch": 1.1863207547169812, |
|
"grad_norm": 0.8985376970213992, |
|
"learning_rate": 7.928585342806607e-07, |
|
"loss": 0.7461, |
|
"step": 503 |
|
}, |
|
{ |
|
"epoch": 1.1886792452830188, |
|
"grad_norm": 0.7846783242941104, |
|
"learning_rate": 7.890087025027579e-07, |
|
"loss": 0.7461, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 1.1910377358490567, |
|
"grad_norm": 0.800897718044774, |
|
"learning_rate": 7.85162140801596e-07, |
|
"loss": 0.7525, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 1.1933962264150944, |
|
"grad_norm": 0.7742939894445084, |
|
"learning_rate": 7.813189087936242e-07, |
|
"loss": 0.7082, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 1.195754716981132, |
|
"grad_norm": 0.8945053331857878, |
|
"learning_rate": 7.774790660436857e-07, |
|
"loss": 0.7328, |
|
"step": 507 |
|
}, |
|
{ |
|
"epoch": 1.1981132075471699, |
|
"grad_norm": 0.9098394537893619, |
|
"learning_rate": 7.736426720640948e-07, |
|
"loss": 0.7802, |
|
"step": 508 |
|
}, |
|
{ |
|
"epoch": 1.2004716981132075, |
|
"grad_norm": 0.7969225975663269, |
|
"learning_rate": 7.698097863137152e-07, |
|
"loss": 0.7423, |
|
"step": 509 |
|
}, |
|
{ |
|
"epoch": 1.2028301886792452, |
|
"grad_norm": 0.8867765640390939, |
|
"learning_rate": 7.659804681970377e-07, |
|
"loss": 0.7443, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 1.205188679245283, |
|
"grad_norm": 0.8936277250989298, |
|
"learning_rate": 7.621547770632589e-07, |
|
"loss": 0.7446, |
|
"step": 511 |
|
}, |
|
{ |
|
"epoch": 1.2075471698113207, |
|
"grad_norm": 0.7650448661091079, |
|
"learning_rate": 7.583327722053626e-07, |
|
"loss": 0.7437, |
|
"step": 512 |
|
}, |
|
{ |
|
"epoch": 1.2099056603773586, |
|
"grad_norm": 0.7718047147502091, |
|
"learning_rate": 7.545145128592008e-07, |
|
"loss": 0.7507, |
|
"step": 513 |
|
}, |
|
{ |
|
"epoch": 1.2122641509433962, |
|
"grad_norm": 0.9572262834139598, |
|
"learning_rate": 7.507000582025748e-07, |
|
"loss": 0.7329, |
|
"step": 514 |
|
}, |
|
{ |
|
"epoch": 1.2146226415094339, |
|
"grad_norm": 0.8375942206336273, |
|
"learning_rate": 7.468894673543181e-07, |
|
"loss": 0.7393, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 1.2169811320754718, |
|
"grad_norm": 0.8300278395620482, |
|
"learning_rate": 7.430827993733808e-07, |
|
"loss": 0.729, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 1.2193396226415094, |
|
"grad_norm": 0.924932117248118, |
|
"learning_rate": 7.39280113257914e-07, |
|
"loss": 0.7418, |
|
"step": 517 |
|
}, |
|
{ |
|
"epoch": 1.2216981132075473, |
|
"grad_norm": 0.8751602815649364, |
|
"learning_rate": 7.354814679443556e-07, |
|
"loss": 0.7695, |
|
"step": 518 |
|
}, |
|
{ |
|
"epoch": 1.224056603773585, |
|
"grad_norm": 0.7891993502461758, |
|
"learning_rate": 7.316869223065155e-07, |
|
"loss": 0.7755, |
|
"step": 519 |
|
}, |
|
{ |
|
"epoch": 1.2264150943396226, |
|
"grad_norm": 0.8088776711707174, |
|
"learning_rate": 7.278965351546648e-07, |
|
"loss": 0.7572, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 1.2287735849056605, |
|
"grad_norm": 0.7734655717390242, |
|
"learning_rate": 7.241103652346248e-07, |
|
"loss": 0.731, |
|
"step": 521 |
|
}, |
|
{ |
|
"epoch": 1.2311320754716981, |
|
"grad_norm": 0.8075042544518887, |
|
"learning_rate": 7.20328471226854e-07, |
|
"loss": 0.7681, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 1.2334905660377358, |
|
"grad_norm": 0.8682085195199817, |
|
"learning_rate": 7.165509117455406e-07, |
|
"loss": 0.7869, |
|
"step": 523 |
|
}, |
|
{ |
|
"epoch": 1.2358490566037736, |
|
"grad_norm": 0.7941356815577187, |
|
"learning_rate": 7.127777453376935e-07, |
|
"loss": 0.7839, |
|
"step": 524 |
|
}, |
|
{ |
|
"epoch": 1.2382075471698113, |
|
"grad_norm": 0.797011639565925, |
|
"learning_rate": 7.090090304822355e-07, |
|
"loss": 0.7704, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 1.240566037735849, |
|
"grad_norm": 0.778511764914176, |
|
"learning_rate": 7.052448255890957e-07, |
|
"loss": 0.7684, |
|
"step": 526 |
|
}, |
|
{ |
|
"epoch": 1.2429245283018868, |
|
"grad_norm": 0.9154469456152399, |
|
"learning_rate": 7.014851889983057e-07, |
|
"loss": 0.747, |
|
"step": 527 |
|
}, |
|
{ |
|
"epoch": 1.2452830188679245, |
|
"grad_norm": 0.7896606577750244, |
|
"learning_rate": 6.97730178979093e-07, |
|
"loss": 0.7243, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 1.2476415094339623, |
|
"grad_norm": 0.8076356582882437, |
|
"learning_rate": 6.939798537289825e-07, |
|
"loss": 0.7879, |
|
"step": 529 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 0.8038075591560918, |
|
"learning_rate": 6.902342713728886e-07, |
|
"loss": 0.7304, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"eval_loss": 0.6839449405670166, |
|
"eval_runtime": 82.7741, |
|
"eval_samples_per_second": 8.517, |
|
"eval_steps_per_second": 0.544, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 1.2523584905660377, |
|
"grad_norm": 0.7543374055980275, |
|
"learning_rate": 6.864934899622191e-07, |
|
"loss": 0.7671, |
|
"step": 531 |
|
}, |
|
{ |
|
"epoch": 1.2547169811320755, |
|
"grad_norm": 0.7940675058566147, |
|
"learning_rate": 6.827575674739725e-07, |
|
"loss": 0.7509, |
|
"step": 532 |
|
}, |
|
{ |
|
"epoch": 1.2570754716981132, |
|
"grad_norm": 0.7898168085125467, |
|
"learning_rate": 6.790265618098423e-07, |
|
"loss": 0.7819, |
|
"step": 533 |
|
}, |
|
{ |
|
"epoch": 1.259433962264151, |
|
"grad_norm": 0.8290907307209445, |
|
"learning_rate": 6.753005307953165e-07, |
|
"loss": 0.7433, |
|
"step": 534 |
|
}, |
|
{ |
|
"epoch": 1.2617924528301887, |
|
"grad_norm": 0.8051377335366591, |
|
"learning_rate": 6.715795321787836e-07, |
|
"loss": 0.7583, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 1.2641509433962264, |
|
"grad_norm": 0.7964463188550859, |
|
"learning_rate": 6.678636236306357e-07, |
|
"loss": 0.7733, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 1.2665094339622642, |
|
"grad_norm": 1.0183118136986125, |
|
"learning_rate": 6.641528627423774e-07, |
|
"loss": 0.7648, |
|
"step": 537 |
|
}, |
|
{ |
|
"epoch": 1.2688679245283019, |
|
"grad_norm": 0.7893378901295127, |
|
"learning_rate": 6.604473070257308e-07, |
|
"loss": 0.7543, |
|
"step": 538 |
|
}, |
|
{ |
|
"epoch": 1.2712264150943398, |
|
"grad_norm": 0.962839563526039, |
|
"learning_rate": 6.567470139117447e-07, |
|
"loss": 0.7815, |
|
"step": 539 |
|
}, |
|
{ |
|
"epoch": 1.2735849056603774, |
|
"grad_norm": 0.779909096149425, |
|
"learning_rate": 6.530520407499049e-07, |
|
"loss": 0.7321, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 1.275943396226415, |
|
"grad_norm": 1.0157939390615558, |
|
"learning_rate": 6.493624448072457e-07, |
|
"loss": 0.7622, |
|
"step": 541 |
|
}, |
|
{ |
|
"epoch": 1.2783018867924527, |
|
"grad_norm": 0.9839155144434717, |
|
"learning_rate": 6.456782832674613e-07, |
|
"loss": 0.7498, |
|
"step": 542 |
|
}, |
|
{ |
|
"epoch": 1.2806603773584906, |
|
"grad_norm": 0.8287088174891809, |
|
"learning_rate": 6.419996132300203e-07, |
|
"loss": 0.7479, |
|
"step": 543 |
|
}, |
|
{ |
|
"epoch": 1.2830188679245282, |
|
"grad_norm": 0.8401170082742024, |
|
"learning_rate": 6.383264917092801e-07, |
|
"loss": 0.7398, |
|
"step": 544 |
|
}, |
|
{ |
|
"epoch": 1.2853773584905661, |
|
"grad_norm": 0.8649864952648277, |
|
"learning_rate": 6.34658975633605e-07, |
|
"loss": 0.7629, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 1.2877358490566038, |
|
"grad_norm": 0.7978032312823685, |
|
"learning_rate": 6.30997121844481e-07, |
|
"loss": 0.7455, |
|
"step": 546 |
|
}, |
|
{ |
|
"epoch": 1.2900943396226414, |
|
"grad_norm": 0.78443421199377, |
|
"learning_rate": 6.273409870956369e-07, |
|
"loss": 0.7512, |
|
"step": 547 |
|
}, |
|
{ |
|
"epoch": 1.2924528301886793, |
|
"grad_norm": 0.9742229809355082, |
|
"learning_rate": 6.236906280521646e-07, |
|
"loss": 0.7629, |
|
"step": 548 |
|
}, |
|
{ |
|
"epoch": 1.294811320754717, |
|
"grad_norm": 0.8605023916238441, |
|
"learning_rate": 6.200461012896401e-07, |
|
"loss": 0.7389, |
|
"step": 549 |
|
}, |
|
{ |
|
"epoch": 1.2971698113207548, |
|
"grad_norm": 0.7976345004555616, |
|
"learning_rate": 6.164074632932472e-07, |
|
"loss": 0.7675, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 1.2995283018867925, |
|
"grad_norm": 0.8766199870243025, |
|
"learning_rate": 6.127747704569015e-07, |
|
"loss": 0.7392, |
|
"step": 551 |
|
}, |
|
{ |
|
"epoch": 1.3018867924528301, |
|
"grad_norm": 0.8388074881151663, |
|
"learning_rate": 6.091480790823771e-07, |
|
"loss": 0.7594, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 1.304245283018868, |
|
"grad_norm": 0.7816816297148448, |
|
"learning_rate": 6.055274453784338e-07, |
|
"loss": 0.7754, |
|
"step": 553 |
|
}, |
|
{ |
|
"epoch": 1.3066037735849056, |
|
"grad_norm": 0.8051939468008078, |
|
"learning_rate": 6.019129254599456e-07, |
|
"loss": 0.7456, |
|
"step": 554 |
|
}, |
|
{ |
|
"epoch": 1.3089622641509435, |
|
"grad_norm": 0.7954437915413475, |
|
"learning_rate": 5.983045753470307e-07, |
|
"loss": 0.7804, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 1.3113207547169812, |
|
"grad_norm": 0.7995393304495568, |
|
"learning_rate": 5.947024509641841e-07, |
|
"loss": 0.7346, |
|
"step": 556 |
|
}, |
|
{ |
|
"epoch": 1.3136792452830188, |
|
"grad_norm": 0.7666328856774342, |
|
"learning_rate": 5.911066081394112e-07, |
|
"loss": 0.7641, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 1.3160377358490565, |
|
"grad_norm": 0.9135245763682434, |
|
"learning_rate": 5.875171026033608e-07, |
|
"loss": 0.7493, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 1.3183962264150944, |
|
"grad_norm": 0.7843209916172604, |
|
"learning_rate": 5.839339899884628e-07, |
|
"loss": 0.754, |
|
"step": 559 |
|
}, |
|
{ |
|
"epoch": 1.320754716981132, |
|
"grad_norm": 0.7798421629061796, |
|
"learning_rate": 5.803573258280654e-07, |
|
"loss": 0.7453, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 1.3231132075471699, |
|
"grad_norm": 0.8026516589712169, |
|
"learning_rate": 5.76787165555575e-07, |
|
"loss": 0.7791, |
|
"step": 561 |
|
}, |
|
{ |
|
"epoch": 1.3254716981132075, |
|
"grad_norm": 0.8161986502002226, |
|
"learning_rate": 5.732235645035963e-07, |
|
"loss": 0.7435, |
|
"step": 562 |
|
}, |
|
{ |
|
"epoch": 1.3278301886792452, |
|
"grad_norm": 0.7982394811556922, |
|
"learning_rate": 5.696665779030749e-07, |
|
"loss": 0.7312, |
|
"step": 563 |
|
}, |
|
{ |
|
"epoch": 1.330188679245283, |
|
"grad_norm": 0.8424060129449961, |
|
"learning_rate": 5.661162608824419e-07, |
|
"loss": 0.7209, |
|
"step": 564 |
|
}, |
|
{ |
|
"epoch": 1.3325471698113207, |
|
"grad_norm": 0.8393522976074216, |
|
"learning_rate": 5.625726684667585e-07, |
|
"loss": 0.7497, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 1.3349056603773586, |
|
"grad_norm": 0.8458006058371306, |
|
"learning_rate": 5.590358555768642e-07, |
|
"loss": 0.7415, |
|
"step": 566 |
|
}, |
|
{ |
|
"epoch": 1.3372641509433962, |
|
"grad_norm": 0.7731963506173513, |
|
"learning_rate": 5.555058770285246e-07, |
|
"loss": 0.7289, |
|
"step": 567 |
|
}, |
|
{ |
|
"epoch": 1.3396226415094339, |
|
"grad_norm": 0.919340196876583, |
|
"learning_rate": 5.519827875315823e-07, |
|
"loss": 0.7372, |
|
"step": 568 |
|
}, |
|
{ |
|
"epoch": 1.3419811320754718, |
|
"grad_norm": 0.8047418954501285, |
|
"learning_rate": 5.484666416891108e-07, |
|
"loss": 0.7854, |
|
"step": 569 |
|
}, |
|
{ |
|
"epoch": 1.3443396226415094, |
|
"grad_norm": 0.8420862799710107, |
|
"learning_rate": 5.449574939965636e-07, |
|
"loss": 0.7487, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 1.3466981132075473, |
|
"grad_norm": 0.862354237686273, |
|
"learning_rate": 5.414553988409343e-07, |
|
"loss": 0.7615, |
|
"step": 571 |
|
}, |
|
{ |
|
"epoch": 1.349056603773585, |
|
"grad_norm": 0.764673226355618, |
|
"learning_rate": 5.379604104999109e-07, |
|
"loss": 0.7402, |
|
"step": 572 |
|
}, |
|
{ |
|
"epoch": 1.3514150943396226, |
|
"grad_norm": 0.9229852280599696, |
|
"learning_rate": 5.344725831410368e-07, |
|
"loss": 0.746, |
|
"step": 573 |
|
}, |
|
{ |
|
"epoch": 1.3537735849056602, |
|
"grad_norm": 0.7749939321603743, |
|
"learning_rate": 5.309919708208686e-07, |
|
"loss": 0.7393, |
|
"step": 574 |
|
}, |
|
{ |
|
"epoch": 1.3561320754716981, |
|
"grad_norm": 1.0855469699332105, |
|
"learning_rate": 5.275186274841404e-07, |
|
"loss": 0.7292, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 1.3584905660377358, |
|
"grad_norm": 1.3079826302270743, |
|
"learning_rate": 5.240526069629264e-07, |
|
"loss": 0.7551, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 1.3608490566037736, |
|
"grad_norm": 0.8097304218867772, |
|
"learning_rate": 5.205939629758078e-07, |
|
"loss": 0.7554, |
|
"step": 577 |
|
}, |
|
{ |
|
"epoch": 1.3632075471698113, |
|
"grad_norm": 0.7629186235386943, |
|
"learning_rate": 5.171427491270393e-07, |
|
"loss": 0.7448, |
|
"step": 578 |
|
}, |
|
{ |
|
"epoch": 1.365566037735849, |
|
"grad_norm": 0.8231834244100196, |
|
"learning_rate": 5.136990189057187e-07, |
|
"loss": 0.7675, |
|
"step": 579 |
|
}, |
|
{ |
|
"epoch": 1.3679245283018868, |
|
"grad_norm": 0.7758543745731015, |
|
"learning_rate": 5.102628256849575e-07, |
|
"loss": 0.7643, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 1.3702830188679245, |
|
"grad_norm": 0.7915005392947032, |
|
"learning_rate": 5.068342227210545e-07, |
|
"loss": 0.732, |
|
"step": 581 |
|
}, |
|
{ |
|
"epoch": 1.3726415094339623, |
|
"grad_norm": 0.821233361833331, |
|
"learning_rate": 5.034132631526695e-07, |
|
"loss": 0.7145, |
|
"step": 582 |
|
}, |
|
{ |
|
"epoch": 1.375, |
|
"grad_norm": 0.8288831867947671, |
|
"learning_rate": 5.000000000000002e-07, |
|
"loss": 0.7717, |
|
"step": 583 |
|
}, |
|
{ |
|
"epoch": 1.3773584905660377, |
|
"grad_norm": 0.9380885957065556, |
|
"learning_rate": 4.9659448616396e-07, |
|
"loss": 0.7655, |
|
"step": 584 |
|
}, |
|
{ |
|
"epoch": 1.3797169811320755, |
|
"grad_norm": 0.769393956851887, |
|
"learning_rate": 4.9319677442536e-07, |
|
"loss": 0.7481, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 1.3820754716981132, |
|
"grad_norm": 0.7624074760874653, |
|
"learning_rate": 4.89806917444088e-07, |
|
"loss": 0.7408, |
|
"step": 586 |
|
}, |
|
{ |
|
"epoch": 1.384433962264151, |
|
"grad_norm": 0.8836661905432738, |
|
"learning_rate": 4.864249677582935e-07, |
|
"loss": 0.7423, |
|
"step": 587 |
|
}, |
|
{ |
|
"epoch": 1.3867924528301887, |
|
"grad_norm": 1.7675171950723623, |
|
"learning_rate": 4.830509777835744e-07, |
|
"loss": 0.7414, |
|
"step": 588 |
|
}, |
|
{ |
|
"epoch": 1.3891509433962264, |
|
"grad_norm": 0.8297208161937744, |
|
"learning_rate": 4.796849998121647e-07, |
|
"loss": 0.7571, |
|
"step": 589 |
|
}, |
|
{ |
|
"epoch": 1.3915094339622642, |
|
"grad_norm": 0.8105952844684988, |
|
"learning_rate": 4.7632708601212215e-07, |
|
"loss": 0.7473, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 1.3938679245283019, |
|
"grad_norm": 0.943401465741068, |
|
"learning_rate": 4.7297728842652116e-07, |
|
"loss": 0.7885, |
|
"step": 591 |
|
}, |
|
{ |
|
"epoch": 1.3962264150943398, |
|
"grad_norm": 0.8407405967900344, |
|
"learning_rate": 4.6963565897264623e-07, |
|
"loss": 0.7768, |
|
"step": 592 |
|
}, |
|
{ |
|
"epoch": 1.3985849056603774, |
|
"grad_norm": 0.7742620223687944, |
|
"learning_rate": 4.663022494411866e-07, |
|
"loss": 0.7679, |
|
"step": 593 |
|
}, |
|
{ |
|
"epoch": 1.400943396226415, |
|
"grad_norm": 0.7984390061772534, |
|
"learning_rate": 4.6297711149543405e-07, |
|
"loss": 0.7503, |
|
"step": 594 |
|
}, |
|
{ |
|
"epoch": 1.4033018867924527, |
|
"grad_norm": 0.7687242314529935, |
|
"learning_rate": 4.596602966704822e-07, |
|
"loss": 0.7593, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 1.4056603773584906, |
|
"grad_norm": 0.7752552260078952, |
|
"learning_rate": 4.5635185637242735e-07, |
|
"loss": 0.7236, |
|
"step": 596 |
|
}, |
|
{ |
|
"epoch": 1.4080188679245282, |
|
"grad_norm": 0.7909608031012165, |
|
"learning_rate": 4.530518418775733e-07, |
|
"loss": 0.7487, |
|
"step": 597 |
|
}, |
|
{ |
|
"epoch": 1.4103773584905661, |
|
"grad_norm": 1.4293658648035819, |
|
"learning_rate": 4.4976030433163337e-07, |
|
"loss": 0.7619, |
|
"step": 598 |
|
}, |
|
{ |
|
"epoch": 1.4127358490566038, |
|
"grad_norm": 0.9601629598020603, |
|
"learning_rate": 4.4647729474894123e-07, |
|
"loss": 0.7301, |
|
"step": 599 |
|
}, |
|
{ |
|
"epoch": 1.4150943396226414, |
|
"grad_norm": 0.8524971604621474, |
|
"learning_rate": 4.432028640116581e-07, |
|
"loss": 0.7743, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 1.4174528301886793, |
|
"grad_norm": 0.7955465630229851, |
|
"learning_rate": 4.399370628689857e-07, |
|
"loss": 0.7549, |
|
"step": 601 |
|
}, |
|
{ |
|
"epoch": 1.419811320754717, |
|
"grad_norm": 0.8461044127574078, |
|
"learning_rate": 4.366799419363779e-07, |
|
"loss": 0.7923, |
|
"step": 602 |
|
}, |
|
{ |
|
"epoch": 1.4221698113207548, |
|
"grad_norm": 0.7959852549975004, |
|
"learning_rate": 4.3343155169475797e-07, |
|
"loss": 0.7313, |
|
"step": 603 |
|
}, |
|
{ |
|
"epoch": 1.4245283018867925, |
|
"grad_norm": 0.8115147827054038, |
|
"learning_rate": 4.3019194248973377e-07, |
|
"loss": 0.7353, |
|
"step": 604 |
|
}, |
|
{ |
|
"epoch": 1.4268867924528301, |
|
"grad_norm": 0.7884615566733161, |
|
"learning_rate": 4.269611645308214e-07, |
|
"loss": 0.7396, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 1.429245283018868, |
|
"grad_norm": 0.8709874327087314, |
|
"learning_rate": 4.237392678906633e-07, |
|
"loss": 0.7557, |
|
"step": 606 |
|
}, |
|
{ |
|
"epoch": 1.4316037735849056, |
|
"grad_norm": 0.9486373793118293, |
|
"learning_rate": 4.205263025042538e-07, |
|
"loss": 0.7397, |
|
"step": 607 |
|
}, |
|
{ |
|
"epoch": 1.4339622641509435, |
|
"grad_norm": 0.8162741407892988, |
|
"learning_rate": 4.173223181681651e-07, |
|
"loss": 0.7327, |
|
"step": 608 |
|
}, |
|
{ |
|
"epoch": 1.4363207547169812, |
|
"grad_norm": 0.8564525506081095, |
|
"learning_rate": 4.141273645397754e-07, |
|
"loss": 0.7745, |
|
"step": 609 |
|
}, |
|
{ |
|
"epoch": 1.4386792452830188, |
|
"grad_norm": 0.8291423725092533, |
|
"learning_rate": 4.109414911364992e-07, |
|
"loss": 0.761, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 1.4410377358490565, |
|
"grad_norm": 0.9110035252316134, |
|
"learning_rate": 4.0776474733502007e-07, |
|
"loss": 0.7227, |
|
"step": 611 |
|
}, |
|
{ |
|
"epoch": 1.4433962264150944, |
|
"grad_norm": 7.248992388941307, |
|
"learning_rate": 4.045971823705249e-07, |
|
"loss": 0.7515, |
|
"step": 612 |
|
}, |
|
{ |
|
"epoch": 1.445754716981132, |
|
"grad_norm": 0.737128592265643, |
|
"learning_rate": 4.0143884533594197e-07, |
|
"loss": 0.7308, |
|
"step": 613 |
|
}, |
|
{ |
|
"epoch": 1.4481132075471699, |
|
"grad_norm": 0.7878766896489899, |
|
"learning_rate": 3.982897851811786e-07, |
|
"loss": 0.7323, |
|
"step": 614 |
|
}, |
|
{ |
|
"epoch": 1.4504716981132075, |
|
"grad_norm": 0.7859241609615767, |
|
"learning_rate": 3.951500507123627e-07, |
|
"loss": 0.7814, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 1.4528301886792452, |
|
"grad_norm": 0.8312739174249127, |
|
"learning_rate": 3.920196905910873e-07, |
|
"loss": 0.7446, |
|
"step": 616 |
|
}, |
|
{ |
|
"epoch": 1.455188679245283, |
|
"grad_norm": 0.8160254075021761, |
|
"learning_rate": 3.888987533336566e-07, |
|
"loss": 0.745, |
|
"step": 617 |
|
}, |
|
{ |
|
"epoch": 1.4575471698113207, |
|
"grad_norm": 0.9271092228733877, |
|
"learning_rate": 3.8578728731033214e-07, |
|
"loss": 0.7875, |
|
"step": 618 |
|
}, |
|
{ |
|
"epoch": 1.4599056603773586, |
|
"grad_norm": 1.3059448042170938, |
|
"learning_rate": 3.826853407445848e-07, |
|
"loss": 0.7758, |
|
"step": 619 |
|
}, |
|
{ |
|
"epoch": 1.4622641509433962, |
|
"grad_norm": 0.7940556961073607, |
|
"learning_rate": 3.795929617123468e-07, |
|
"loss": 0.7289, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 1.4646226415094339, |
|
"grad_norm": 0.7908440739021586, |
|
"learning_rate": 3.765101981412665e-07, |
|
"loss": 0.7247, |
|
"step": 621 |
|
}, |
|
{ |
|
"epoch": 1.4669811320754718, |
|
"grad_norm": 0.807553140785391, |
|
"learning_rate": 3.7343709780996614e-07, |
|
"loss": 0.7365, |
|
"step": 622 |
|
}, |
|
{ |
|
"epoch": 1.4693396226415094, |
|
"grad_norm": 2.5050420828496787, |
|
"learning_rate": 3.703737083473005e-07, |
|
"loss": 0.7729, |
|
"step": 623 |
|
}, |
|
{ |
|
"epoch": 1.4716981132075473, |
|
"grad_norm": 0.8163112108481482, |
|
"learning_rate": 3.673200772316193e-07, |
|
"loss": 0.7269, |
|
"step": 624 |
|
}, |
|
{ |
|
"epoch": 1.474056603773585, |
|
"grad_norm": 0.8451238760005333, |
|
"learning_rate": 3.6427625179003217e-07, |
|
"loss": 0.7435, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 1.4764150943396226, |
|
"grad_norm": 1.0290141575096419, |
|
"learning_rate": 3.6124227919767227e-07, |
|
"loss": 0.7488, |
|
"step": 626 |
|
}, |
|
{ |
|
"epoch": 1.4787735849056602, |
|
"grad_norm": 0.7955100632691958, |
|
"learning_rate": 3.5821820647696864e-07, |
|
"loss": 0.7556, |
|
"step": 627 |
|
}, |
|
{ |
|
"epoch": 1.4811320754716981, |
|
"grad_norm": 0.8238206028160229, |
|
"learning_rate": 3.552040804969149e-07, |
|
"loss": 0.7541, |
|
"step": 628 |
|
}, |
|
{ |
|
"epoch": 1.4834905660377358, |
|
"grad_norm": 0.8519020092733177, |
|
"learning_rate": 3.5219994797234476e-07, |
|
"loss": 0.7783, |
|
"step": 629 |
|
}, |
|
{ |
|
"epoch": 1.4858490566037736, |
|
"grad_norm": 0.7845182991866497, |
|
"learning_rate": 3.4920585546320625e-07, |
|
"loss": 0.7457, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 1.4882075471698113, |
|
"grad_norm": 0.7643446889279868, |
|
"learning_rate": 3.4622184937384156e-07, |
|
"loss": 0.7637, |
|
"step": 631 |
|
}, |
|
{ |
|
"epoch": 1.490566037735849, |
|
"grad_norm": 0.8986522134788724, |
|
"learning_rate": 3.4324797595226564e-07, |
|
"loss": 0.7431, |
|
"step": 632 |
|
}, |
|
{ |
|
"epoch": 1.4929245283018868, |
|
"grad_norm": 0.788015002889114, |
|
"learning_rate": 3.4028428128945286e-07, |
|
"loss": 0.746, |
|
"step": 633 |
|
}, |
|
{ |
|
"epoch": 1.4952830188679245, |
|
"grad_norm": 0.7754842750109397, |
|
"learning_rate": 3.3733081131861975e-07, |
|
"loss": 0.7688, |
|
"step": 634 |
|
}, |
|
{ |
|
"epoch": 1.4976415094339623, |
|
"grad_norm": 0.8009965193531273, |
|
"learning_rate": 3.343876118145141e-07, |
|
"loss": 0.7501, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"grad_norm": 1.006640510210976, |
|
"learning_rate": 3.314547283927057e-07, |
|
"loss": 0.7239, |
|
"step": 636 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"eval_loss": 0.6808757185935974, |
|
"eval_runtime": 83.2321, |
|
"eval_samples_per_second": 8.47, |
|
"eval_steps_per_second": 0.541, |
|
"step": 636 |
|
}, |
|
{ |
|
"epoch": 1.5023584905660377, |
|
"grad_norm": 0.8598327085719165, |
|
"learning_rate": 3.2853220650887913e-07, |
|
"loss": 0.7149, |
|
"step": 637 |
|
}, |
|
{ |
|
"epoch": 1.5047169811320755, |
|
"grad_norm": 0.8482092774985639, |
|
"learning_rate": 3.256200914581292e-07, |
|
"loss": 0.7556, |
|
"step": 638 |
|
}, |
|
{ |
|
"epoch": 1.5070754716981132, |
|
"grad_norm": 0.7837630460537407, |
|
"learning_rate": 3.227184283742591e-07, |
|
"loss": 0.7432, |
|
"step": 639 |
|
}, |
|
{ |
|
"epoch": 1.509433962264151, |
|
"grad_norm": 0.8392002898875265, |
|
"learning_rate": 3.198272622290804e-07, |
|
"loss": 0.7395, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 1.5117924528301887, |
|
"grad_norm": 0.826051018451035, |
|
"learning_rate": 3.169466378317177e-07, |
|
"loss": 0.7849, |
|
"step": 641 |
|
}, |
|
{ |
|
"epoch": 1.5141509433962264, |
|
"grad_norm": 1.0059361369860251, |
|
"learning_rate": 3.1407659982791204e-07, |
|
"loss": 0.7541, |
|
"step": 642 |
|
}, |
|
{ |
|
"epoch": 1.516509433962264, |
|
"grad_norm": 0.8092490329577119, |
|
"learning_rate": 3.112171926993291e-07, |
|
"loss": 0.7348, |
|
"step": 643 |
|
}, |
|
{ |
|
"epoch": 1.5188679245283019, |
|
"grad_norm": 0.7280559308125196, |
|
"learning_rate": 3.0836846076287146e-07, |
|
"loss": 0.6873, |
|
"step": 644 |
|
}, |
|
{ |
|
"epoch": 1.5212264150943398, |
|
"grad_norm": 0.8517206567349668, |
|
"learning_rate": 3.055304481699913e-07, |
|
"loss": 0.7839, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 1.5235849056603774, |
|
"grad_norm": 0.8340097914857253, |
|
"learning_rate": 3.027031989060046e-07, |
|
"loss": 0.7561, |
|
"step": 646 |
|
}, |
|
{ |
|
"epoch": 1.525943396226415, |
|
"grad_norm": 0.8002849424485019, |
|
"learning_rate": 2.998867567894108e-07, |
|
"loss": 0.7747, |
|
"step": 647 |
|
}, |
|
{ |
|
"epoch": 1.5283018867924527, |
|
"grad_norm": 0.8623306309909063, |
|
"learning_rate": 2.970811654712133e-07, |
|
"loss": 0.7337, |
|
"step": 648 |
|
}, |
|
{ |
|
"epoch": 1.5306603773584906, |
|
"grad_norm": 0.8335419959901207, |
|
"learning_rate": 2.942864684342432e-07, |
|
"loss": 0.741, |
|
"step": 649 |
|
}, |
|
{ |
|
"epoch": 1.5330188679245285, |
|
"grad_norm": 0.8731062418731657, |
|
"learning_rate": 2.91502708992485e-07, |
|
"loss": 0.7424, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 1.5353773584905661, |
|
"grad_norm": 0.819811746630834, |
|
"learning_rate": 2.8872993029040506e-07, |
|
"loss": 0.7355, |
|
"step": 651 |
|
}, |
|
{ |
|
"epoch": 1.5377358490566038, |
|
"grad_norm": 0.8074813557073929, |
|
"learning_rate": 2.859681753022838e-07, |
|
"loss": 0.7594, |
|
"step": 652 |
|
}, |
|
{ |
|
"epoch": 1.5400943396226414, |
|
"grad_norm": 0.820619090005774, |
|
"learning_rate": 2.8321748683154887e-07, |
|
"loss": 0.7443, |
|
"step": 653 |
|
}, |
|
{ |
|
"epoch": 1.5424528301886793, |
|
"grad_norm": 0.8554565499942881, |
|
"learning_rate": 2.8047790751011216e-07, |
|
"loss": 0.6991, |
|
"step": 654 |
|
}, |
|
{ |
|
"epoch": 1.544811320754717, |
|
"grad_norm": 0.7784690870870672, |
|
"learning_rate": 2.777494797977088e-07, |
|
"loss": 0.7328, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 1.5471698113207548, |
|
"grad_norm": 1.7923607795927756, |
|
"learning_rate": 2.7503224598123895e-07, |
|
"loss": 0.7266, |
|
"step": 656 |
|
}, |
|
{ |
|
"epoch": 1.5495283018867925, |
|
"grad_norm": 0.7677443512564135, |
|
"learning_rate": 2.7232624817411376e-07, |
|
"loss": 0.737, |
|
"step": 657 |
|
}, |
|
{ |
|
"epoch": 1.5518867924528301, |
|
"grad_norm": 0.8055758349794541, |
|
"learning_rate": 2.6963152831560066e-07, |
|
"loss": 0.7444, |
|
"step": 658 |
|
}, |
|
{ |
|
"epoch": 1.5542452830188678, |
|
"grad_norm": 0.9008016637756693, |
|
"learning_rate": 2.6694812817017387e-07, |
|
"loss": 0.7128, |
|
"step": 659 |
|
}, |
|
{ |
|
"epoch": 1.5566037735849056, |
|
"grad_norm": 0.8106690468627914, |
|
"learning_rate": 2.642760893268684e-07, |
|
"loss": 0.7457, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 1.5589622641509435, |
|
"grad_norm": 1.027228042593183, |
|
"learning_rate": 2.616154531986345e-07, |
|
"loss": 0.7251, |
|
"step": 661 |
|
}, |
|
{ |
|
"epoch": 1.5613207547169812, |
|
"grad_norm": 0.7957920778904417, |
|
"learning_rate": 2.5896626102169594e-07, |
|
"loss": 0.7481, |
|
"step": 662 |
|
}, |
|
{ |
|
"epoch": 1.5636792452830188, |
|
"grad_norm": 0.7514802697133819, |
|
"learning_rate": 2.5632855385491037e-07, |
|
"loss": 0.757, |
|
"step": 663 |
|
}, |
|
{ |
|
"epoch": 1.5660377358490565, |
|
"grad_norm": 0.8814866050056973, |
|
"learning_rate": 2.53702372579134e-07, |
|
"loss": 0.7361, |
|
"step": 664 |
|
}, |
|
{ |
|
"epoch": 1.5683962264150944, |
|
"grad_norm": 0.8346312832230348, |
|
"learning_rate": 2.51087757896587e-07, |
|
"loss": 0.7378, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 1.5707547169811322, |
|
"grad_norm": 0.8164772219257521, |
|
"learning_rate": 2.4848475033022377e-07, |
|
"loss": 0.7631, |
|
"step": 666 |
|
}, |
|
{ |
|
"epoch": 1.5731132075471699, |
|
"grad_norm": 0.7933267558316177, |
|
"learning_rate": 2.458933902231038e-07, |
|
"loss": 0.7288, |
|
"step": 667 |
|
}, |
|
{ |
|
"epoch": 1.5754716981132075, |
|
"grad_norm": 0.8420080383383726, |
|
"learning_rate": 2.4331371773776687e-07, |
|
"loss": 0.77, |
|
"step": 668 |
|
}, |
|
{ |
|
"epoch": 1.5778301886792452, |
|
"grad_norm": 5.289720197423799, |
|
"learning_rate": 2.407457728556115e-07, |
|
"loss": 0.7184, |
|
"step": 669 |
|
}, |
|
{ |
|
"epoch": 1.580188679245283, |
|
"grad_norm": 1.0842649090634284, |
|
"learning_rate": 2.3818959537627282e-07, |
|
"loss": 0.7383, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 1.5825471698113207, |
|
"grad_norm": 0.8519744691939684, |
|
"learning_rate": 2.3564522491700833e-07, |
|
"loss": 0.769, |
|
"step": 671 |
|
}, |
|
{ |
|
"epoch": 1.5849056603773586, |
|
"grad_norm": 0.8168668326966967, |
|
"learning_rate": 2.3311270091208256e-07, |
|
"loss": 0.7444, |
|
"step": 672 |
|
}, |
|
{ |
|
"epoch": 1.5872641509433962, |
|
"grad_norm": 0.7514913975722562, |
|
"learning_rate": 2.3059206261215668e-07, |
|
"loss": 0.691, |
|
"step": 673 |
|
}, |
|
{ |
|
"epoch": 1.5896226415094339, |
|
"grad_norm": 0.7519874181403705, |
|
"learning_rate": 2.2808334908367909e-07, |
|
"loss": 0.7425, |
|
"step": 674 |
|
}, |
|
{ |
|
"epoch": 1.5919811320754715, |
|
"grad_norm": 0.8353526410754343, |
|
"learning_rate": 2.2558659920828095e-07, |
|
"loss": 0.7616, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 1.5943396226415094, |
|
"grad_norm": 0.7738579301027763, |
|
"learning_rate": 2.2310185168217212e-07, |
|
"loss": 0.7756, |
|
"step": 676 |
|
}, |
|
{ |
|
"epoch": 1.5966981132075473, |
|
"grad_norm": 0.813046680487735, |
|
"learning_rate": 2.206291450155441e-07, |
|
"loss": 0.7675, |
|
"step": 677 |
|
}, |
|
{ |
|
"epoch": 1.599056603773585, |
|
"grad_norm": 0.7815916852088219, |
|
"learning_rate": 2.181685175319702e-07, |
|
"loss": 0.7459, |
|
"step": 678 |
|
}, |
|
{ |
|
"epoch": 1.6014150943396226, |
|
"grad_norm": 0.7482293961455181, |
|
"learning_rate": 2.157200073678137e-07, |
|
"loss": 0.7237, |
|
"step": 679 |
|
}, |
|
{ |
|
"epoch": 1.6037735849056602, |
|
"grad_norm": 0.9475759536537903, |
|
"learning_rate": 2.132836524716355e-07, |
|
"loss": 0.718, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 1.6061320754716981, |
|
"grad_norm": 0.8731724054191227, |
|
"learning_rate": 2.1085949060360653e-07, |
|
"loss": 0.7344, |
|
"step": 681 |
|
}, |
|
{ |
|
"epoch": 1.608490566037736, |
|
"grad_norm": 0.8124879077769539, |
|
"learning_rate": 2.0844755933492263e-07, |
|
"loss": 0.7515, |
|
"step": 682 |
|
}, |
|
{ |
|
"epoch": 1.6108490566037736, |
|
"grad_norm": 0.7772794085957034, |
|
"learning_rate": 2.0604789604722205e-07, |
|
"loss": 0.7471, |
|
"step": 683 |
|
}, |
|
{ |
|
"epoch": 1.6132075471698113, |
|
"grad_norm": 0.8144653370194169, |
|
"learning_rate": 2.0366053793200565e-07, |
|
"loss": 0.7724, |
|
"step": 684 |
|
}, |
|
{ |
|
"epoch": 1.615566037735849, |
|
"grad_norm": 1.1024190577460233, |
|
"learning_rate": 2.0128552199006198e-07, |
|
"loss": 0.7389, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 1.6179245283018868, |
|
"grad_norm": 0.7813510804481604, |
|
"learning_rate": 1.9892288503089205e-07, |
|
"loss": 0.7688, |
|
"step": 686 |
|
}, |
|
{ |
|
"epoch": 1.6202830188679245, |
|
"grad_norm": 0.7649385235808642, |
|
"learning_rate": 1.9657266367213898e-07, |
|
"loss": 0.7279, |
|
"step": 687 |
|
}, |
|
{ |
|
"epoch": 1.6226415094339623, |
|
"grad_norm": 0.8664030315613933, |
|
"learning_rate": 1.9423489433902184e-07, |
|
"loss": 0.7604, |
|
"step": 688 |
|
}, |
|
{ |
|
"epoch": 1.625, |
|
"grad_norm": 0.8134285929715391, |
|
"learning_rate": 1.9190961326377053e-07, |
|
"loss": 0.7628, |
|
"step": 689 |
|
}, |
|
{ |
|
"epoch": 1.6273584905660377, |
|
"grad_norm": 0.8126410627826297, |
|
"learning_rate": 1.8959685648506362e-07, |
|
"loss": 0.7479, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 1.6297169811320755, |
|
"grad_norm": 0.8243887159653667, |
|
"learning_rate": 1.8729665984747e-07, |
|
"loss": 0.743, |
|
"step": 691 |
|
}, |
|
{ |
|
"epoch": 1.6320754716981132, |
|
"grad_norm": 0.8107223249523531, |
|
"learning_rate": 1.8500905900089403e-07, |
|
"loss": 0.7601, |
|
"step": 692 |
|
}, |
|
{ |
|
"epoch": 1.634433962264151, |
|
"grad_norm": 0.8038068007057135, |
|
"learning_rate": 1.82734089400022e-07, |
|
"loss": 0.7643, |
|
"step": 693 |
|
}, |
|
{ |
|
"epoch": 1.6367924528301887, |
|
"grad_norm": 0.9712605139438796, |
|
"learning_rate": 1.804717863037737e-07, |
|
"loss": 0.7445, |
|
"step": 694 |
|
}, |
|
{ |
|
"epoch": 1.6391509433962264, |
|
"grad_norm": 0.7879085057169681, |
|
"learning_rate": 1.7822218477475494e-07, |
|
"loss": 0.7485, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 1.641509433962264, |
|
"grad_norm": 0.8550428267297568, |
|
"learning_rate": 1.7598531967871465e-07, |
|
"loss": 0.7783, |
|
"step": 696 |
|
}, |
|
{ |
|
"epoch": 1.6438679245283019, |
|
"grad_norm": 0.7795732830836477, |
|
"learning_rate": 1.737612256840053e-07, |
|
"loss": 0.7484, |
|
"step": 697 |
|
}, |
|
{ |
|
"epoch": 1.6462264150943398, |
|
"grad_norm": 0.8133414594705825, |
|
"learning_rate": 1.7154993726104328e-07, |
|
"loss": 0.7751, |
|
"step": 698 |
|
}, |
|
{ |
|
"epoch": 1.6485849056603774, |
|
"grad_norm": 0.7994545997274072, |
|
"learning_rate": 1.6935148868177718e-07, |
|
"loss": 0.7501, |
|
"step": 699 |
|
}, |
|
{ |
|
"epoch": 1.650943396226415, |
|
"grad_norm": 0.7496147135106944, |
|
"learning_rate": 1.6716591401915502e-07, |
|
"loss": 0.759, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 1.6533018867924527, |
|
"grad_norm": 0.7274661448787512, |
|
"learning_rate": 1.6499324714659758e-07, |
|
"loss": 0.735, |
|
"step": 701 |
|
}, |
|
{ |
|
"epoch": 1.6556603773584906, |
|
"grad_norm": 0.8892069878489233, |
|
"learning_rate": 1.6283352173747146e-07, |
|
"loss": 0.7694, |
|
"step": 702 |
|
}, |
|
{ |
|
"epoch": 1.6580188679245285, |
|
"grad_norm": 0.7851702666302485, |
|
"learning_rate": 1.6068677126456897e-07, |
|
"loss": 0.7373, |
|
"step": 703 |
|
}, |
|
{ |
|
"epoch": 1.6603773584905661, |
|
"grad_norm": 0.7771975549062173, |
|
"learning_rate": 1.585530289995878e-07, |
|
"loss": 0.7101, |
|
"step": 704 |
|
}, |
|
{ |
|
"epoch": 1.6627358490566038, |
|
"grad_norm": 0.8852742002247123, |
|
"learning_rate": 1.564323280126173e-07, |
|
"loss": 0.7525, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 1.6650943396226414, |
|
"grad_norm": 1.0210362257937011, |
|
"learning_rate": 1.5432470117162433e-07, |
|
"loss": 0.7752, |
|
"step": 706 |
|
}, |
|
{ |
|
"epoch": 1.6674528301886793, |
|
"grad_norm": 0.8077571397037019, |
|
"learning_rate": 1.522301811419442e-07, |
|
"loss": 0.7395, |
|
"step": 707 |
|
}, |
|
{ |
|
"epoch": 1.669811320754717, |
|
"grad_norm": 0.8331764729151993, |
|
"learning_rate": 1.5014880038577482e-07, |
|
"loss": 0.7416, |
|
"step": 708 |
|
}, |
|
{ |
|
"epoch": 1.6721698113207548, |
|
"grad_norm": 2.1774846943729935, |
|
"learning_rate": 1.4808059116167303e-07, |
|
"loss": 0.7492, |
|
"step": 709 |
|
}, |
|
{ |
|
"epoch": 1.6745283018867925, |
|
"grad_norm": 0.8939523687224223, |
|
"learning_rate": 1.460255855240552e-07, |
|
"loss": 0.7358, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 1.6768867924528301, |
|
"grad_norm": 0.7563678829167328, |
|
"learning_rate": 1.4398381532269998e-07, |
|
"loss": 0.7593, |
|
"step": 711 |
|
}, |
|
{ |
|
"epoch": 1.6792452830188678, |
|
"grad_norm": 0.7668313174457733, |
|
"learning_rate": 1.4195531220225487e-07, |
|
"loss": 0.7275, |
|
"step": 712 |
|
}, |
|
{ |
|
"epoch": 1.6816037735849056, |
|
"grad_norm": 0.7791231762469476, |
|
"learning_rate": 1.39940107601746e-07, |
|
"loss": 0.7692, |
|
"step": 713 |
|
}, |
|
{ |
|
"epoch": 1.6839622641509435, |
|
"grad_norm": 1.014530101254706, |
|
"learning_rate": 1.3793823275409066e-07, |
|
"loss": 0.7475, |
|
"step": 714 |
|
}, |
|
{ |
|
"epoch": 1.6863207547169812, |
|
"grad_norm": 0.754925190142359, |
|
"learning_rate": 1.3594971868561232e-07, |
|
"loss": 0.7248, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 1.6886792452830188, |
|
"grad_norm": 0.7608021380299098, |
|
"learning_rate": 1.3397459621556128e-07, |
|
"loss": 0.7558, |
|
"step": 716 |
|
}, |
|
{ |
|
"epoch": 1.6910377358490565, |
|
"grad_norm": 1.320307190976547, |
|
"learning_rate": 1.320128959556369e-07, |
|
"loss": 0.7393, |
|
"step": 717 |
|
}, |
|
{ |
|
"epoch": 1.6933962264150944, |
|
"grad_norm": 0.8339214047495644, |
|
"learning_rate": 1.300646483095118e-07, |
|
"loss": 0.7597, |
|
"step": 718 |
|
}, |
|
{ |
|
"epoch": 1.6957547169811322, |
|
"grad_norm": 0.8826119280765871, |
|
"learning_rate": 1.2812988347236166e-07, |
|
"loss": 0.7416, |
|
"step": 719 |
|
}, |
|
{ |
|
"epoch": 1.6981132075471699, |
|
"grad_norm": 0.8928251184591149, |
|
"learning_rate": 1.262086314303973e-07, |
|
"loss": 0.7318, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 1.7004716981132075, |
|
"grad_norm": 0.7951715947182522, |
|
"learning_rate": 1.243009219603993e-07, |
|
"loss": 0.7482, |
|
"step": 721 |
|
}, |
|
{ |
|
"epoch": 1.7028301886792452, |
|
"grad_norm": 0.8741704408992066, |
|
"learning_rate": 1.2240678462925723e-07, |
|
"loss": 0.7287, |
|
"step": 722 |
|
}, |
|
{ |
|
"epoch": 1.705188679245283, |
|
"grad_norm": 0.7786335970762307, |
|
"learning_rate": 1.2052624879351103e-07, |
|
"loss": 0.719, |
|
"step": 723 |
|
}, |
|
{ |
|
"epoch": 1.7075471698113207, |
|
"grad_norm": 0.9985291709831805, |
|
"learning_rate": 1.1865934359889573e-07, |
|
"loss": 0.7453, |
|
"step": 724 |
|
}, |
|
{ |
|
"epoch": 1.7099056603773586, |
|
"grad_norm": 0.7859405679363147, |
|
"learning_rate": 1.1680609797989038e-07, |
|
"loss": 0.7452, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 1.7122641509433962, |
|
"grad_norm": 0.8028725796240722, |
|
"learning_rate": 1.1496654065926925e-07, |
|
"loss": 0.7173, |
|
"step": 726 |
|
}, |
|
{ |
|
"epoch": 1.7146226415094339, |
|
"grad_norm": 0.8165982522602734, |
|
"learning_rate": 1.1314070014765642e-07, |
|
"loss": 0.7218, |
|
"step": 727 |
|
}, |
|
{ |
|
"epoch": 1.7169811320754715, |
|
"grad_norm": 0.7860295163288017, |
|
"learning_rate": 1.1132860474308436e-07, |
|
"loss": 0.7614, |
|
"step": 728 |
|
}, |
|
{ |
|
"epoch": 1.7193396226415094, |
|
"grad_norm": 0.7889724037710595, |
|
"learning_rate": 1.0953028253055541e-07, |
|
"loss": 0.7195, |
|
"step": 729 |
|
}, |
|
{ |
|
"epoch": 1.7216981132075473, |
|
"grad_norm": 0.8067588097960763, |
|
"learning_rate": 1.0774576138160596e-07, |
|
"loss": 0.7482, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 1.724056603773585, |
|
"grad_norm": 0.786061208841395, |
|
"learning_rate": 1.0597506895387499e-07, |
|
"loss": 0.7472, |
|
"step": 731 |
|
}, |
|
{ |
|
"epoch": 1.7264150943396226, |
|
"grad_norm": 0.7917438727246212, |
|
"learning_rate": 1.0421823269067442e-07, |
|
"loss": 0.7751, |
|
"step": 732 |
|
}, |
|
{ |
|
"epoch": 1.7287735849056602, |
|
"grad_norm": 0.7907162871021037, |
|
"learning_rate": 1.024752798205658e-07, |
|
"loss": 0.7202, |
|
"step": 733 |
|
}, |
|
{ |
|
"epoch": 1.7311320754716981, |
|
"grad_norm": 0.9995843557820588, |
|
"learning_rate": 1.0074623735693633e-07, |
|
"loss": 0.7471, |
|
"step": 734 |
|
}, |
|
{ |
|
"epoch": 1.733490566037736, |
|
"grad_norm": 1.3558698569505305, |
|
"learning_rate": 9.903113209758096e-08, |
|
"loss": 0.7564, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 1.7358490566037736, |
|
"grad_norm": 0.7870396429020904, |
|
"learning_rate": 9.732999062428704e-08, |
|
"loss": 0.7641, |
|
"step": 736 |
|
}, |
|
{ |
|
"epoch": 1.7382075471698113, |
|
"grad_norm": 0.8525068172515468, |
|
"learning_rate": 9.564283930242257e-08, |
|
"loss": 0.7404, |
|
"step": 737 |
|
}, |
|
{ |
|
"epoch": 1.740566037735849, |
|
"grad_norm": 0.8014286873652903, |
|
"learning_rate": 9.396970428052697e-08, |
|
"loss": 0.7194, |
|
"step": 738 |
|
}, |
|
{ |
|
"epoch": 1.7429245283018868, |
|
"grad_norm": 0.7676759180901973, |
|
"learning_rate": 9.231061148990648e-08, |
|
"loss": 0.7221, |
|
"step": 739 |
|
}, |
|
{ |
|
"epoch": 1.7452830188679245, |
|
"grad_norm": 0.9131948150344564, |
|
"learning_rate": 9.066558664423163e-08, |
|
"loss": 0.7572, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 1.7476415094339623, |
|
"grad_norm": 0.7734044180311592, |
|
"learning_rate": 8.903465523913955e-08, |
|
"loss": 0.7757, |
|
"step": 741 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"grad_norm": 0.755029746442279, |
|
"learning_rate": 8.741784255183759e-08, |
|
"loss": 0.7411, |
|
"step": 742 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"eval_loss": 0.6796497702598572, |
|
"eval_runtime": 82.4746, |
|
"eval_samples_per_second": 8.548, |
|
"eval_steps_per_second": 0.546, |
|
"step": 742 |
|
}, |
|
{ |
|
"epoch": 1.7523584905660377, |
|
"grad_norm": 0.8497097319379918, |
|
"learning_rate": 8.581517364071267e-08, |
|
"loss": 0.7203, |
|
"step": 743 |
|
}, |
|
{ |
|
"epoch": 1.7547169811320755, |
|
"grad_norm": 0.8020779183255858, |
|
"learning_rate": 8.422667334494249e-08, |
|
"loss": 0.7558, |
|
"step": 744 |
|
}, |
|
{ |
|
"epoch": 1.7570754716981132, |
|
"grad_norm": 0.7616565389684601, |
|
"learning_rate": 8.265236628411087e-08, |
|
"loss": 0.7422, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 1.759433962264151, |
|
"grad_norm": 0.9323853208771872, |
|
"learning_rate": 8.109227685782538e-08, |
|
"loss": 0.7819, |
|
"step": 746 |
|
}, |
|
{ |
|
"epoch": 1.7617924528301887, |
|
"grad_norm": 1.350644299549973, |
|
"learning_rate": 7.954642924533994e-08, |
|
"loss": 0.7492, |
|
"step": 747 |
|
}, |
|
{ |
|
"epoch": 1.7641509433962264, |
|
"grad_norm": 0.8092647636710026, |
|
"learning_rate": 7.801484740517939e-08, |
|
"loss": 0.7638, |
|
"step": 748 |
|
}, |
|
{ |
|
"epoch": 1.766509433962264, |
|
"grad_norm": 0.8308445659034376, |
|
"learning_rate": 7.649755507476952e-08, |
|
"loss": 0.7444, |
|
"step": 749 |
|
}, |
|
{ |
|
"epoch": 1.7688679245283019, |
|
"grad_norm": 2.5649621576701587, |
|
"learning_rate": 7.499457577006751e-08, |
|
"loss": 0.7713, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 1.7712264150943398, |
|
"grad_norm": 0.786323334872225, |
|
"learning_rate": 7.350593278519823e-08, |
|
"loss": 0.7426, |
|
"step": 751 |
|
}, |
|
{ |
|
"epoch": 1.7735849056603774, |
|
"grad_norm": 0.7795649745069329, |
|
"learning_rate": 7.203164919209359e-08, |
|
"loss": 0.7674, |
|
"step": 752 |
|
}, |
|
{ |
|
"epoch": 1.775943396226415, |
|
"grad_norm": 0.8003158325581134, |
|
"learning_rate": 7.057174784013431e-08, |
|
"loss": 0.7531, |
|
"step": 753 |
|
}, |
|
{ |
|
"epoch": 1.7783018867924527, |
|
"grad_norm": 0.7834816769548146, |
|
"learning_rate": 6.912625135579586e-08, |
|
"loss": 0.7212, |
|
"step": 754 |
|
}, |
|
{ |
|
"epoch": 1.7806603773584906, |
|
"grad_norm": 0.8289452420732738, |
|
"learning_rate": 6.76951821422982e-08, |
|
"loss": 0.7715, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 1.7830188679245285, |
|
"grad_norm": 0.8410465172590653, |
|
"learning_rate": 6.627856237925811e-08, |
|
"loss": 0.7276, |
|
"step": 756 |
|
}, |
|
{ |
|
"epoch": 1.7853773584905661, |
|
"grad_norm": 0.8394202851068477, |
|
"learning_rate": 6.487641402234612e-08, |
|
"loss": 0.744, |
|
"step": 757 |
|
}, |
|
{ |
|
"epoch": 1.7877358490566038, |
|
"grad_norm": 0.8383760649776106, |
|
"learning_rate": 6.348875880294535e-08, |
|
"loss": 0.7384, |
|
"step": 758 |
|
}, |
|
{ |
|
"epoch": 1.7900943396226414, |
|
"grad_norm": 1.453114388460075, |
|
"learning_rate": 6.211561822781474e-08, |
|
"loss": 0.7765, |
|
"step": 759 |
|
}, |
|
{ |
|
"epoch": 1.7924528301886793, |
|
"grad_norm": 0.8499896987869552, |
|
"learning_rate": 6.075701357875662e-08, |
|
"loss": 0.7284, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 1.794811320754717, |
|
"grad_norm": 0.8652146496280481, |
|
"learning_rate": 5.9412965912286396e-08, |
|
"loss": 0.7561, |
|
"step": 761 |
|
}, |
|
{ |
|
"epoch": 1.7971698113207548, |
|
"grad_norm": 0.8299032649447099, |
|
"learning_rate": 5.808349605930585e-08, |
|
"loss": 0.7565, |
|
"step": 762 |
|
}, |
|
{ |
|
"epoch": 1.7995283018867925, |
|
"grad_norm": 0.7812451643671983, |
|
"learning_rate": 5.6768624624780604e-08, |
|
"loss": 0.725, |
|
"step": 763 |
|
}, |
|
{ |
|
"epoch": 1.8018867924528301, |
|
"grad_norm": 0.8477713992666978, |
|
"learning_rate": 5.5468371987420936e-08, |
|
"loss": 0.7466, |
|
"step": 764 |
|
}, |
|
{ |
|
"epoch": 1.8042452830188678, |
|
"grad_norm": 0.8337239533995923, |
|
"learning_rate": 5.4182758299365364e-08, |
|
"loss": 0.735, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 1.8066037735849056, |
|
"grad_norm": 0.7691176408664865, |
|
"learning_rate": 5.29118034858691e-08, |
|
"loss": 0.772, |
|
"step": 766 |
|
}, |
|
{ |
|
"epoch": 1.8089622641509435, |
|
"grad_norm": 0.8049526588971233, |
|
"learning_rate": 5.165552724499478e-08, |
|
"loss": 0.7315, |
|
"step": 767 |
|
}, |
|
{ |
|
"epoch": 1.8113207547169812, |
|
"grad_norm": 0.8319464046786564, |
|
"learning_rate": 5.0413949047306894e-08, |
|
"loss": 0.7726, |
|
"step": 768 |
|
}, |
|
{ |
|
"epoch": 1.8136792452830188, |
|
"grad_norm": 0.8474092435937203, |
|
"learning_rate": 4.918708813557093e-08, |
|
"loss": 0.7487, |
|
"step": 769 |
|
}, |
|
{ |
|
"epoch": 1.8160377358490565, |
|
"grad_norm": 0.7943675010786753, |
|
"learning_rate": 4.797496352445396e-08, |
|
"loss": 0.7691, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 1.8183962264150944, |
|
"grad_norm": 0.8217628363291084, |
|
"learning_rate": 4.677759400023085e-08, |
|
"loss": 0.7437, |
|
"step": 771 |
|
}, |
|
{ |
|
"epoch": 1.8207547169811322, |
|
"grad_norm": 0.832358842867652, |
|
"learning_rate": 4.55949981204925e-08, |
|
"loss": 0.7568, |
|
"step": 772 |
|
}, |
|
{ |
|
"epoch": 1.8231132075471699, |
|
"grad_norm": 0.8613869764376921, |
|
"learning_rate": 4.442719421385921e-08, |
|
"loss": 0.7274, |
|
"step": 773 |
|
}, |
|
{ |
|
"epoch": 1.8254716981132075, |
|
"grad_norm": 0.7931363978514411, |
|
"learning_rate": 4.3274200379695315e-08, |
|
"loss": 0.7209, |
|
"step": 774 |
|
}, |
|
{ |
|
"epoch": 1.8278301886792452, |
|
"grad_norm": 0.7551383801460104, |
|
"learning_rate": 4.213603448782932e-08, |
|
"loss": 0.7533, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 1.830188679245283, |
|
"grad_norm": 0.8476707134264969, |
|
"learning_rate": 4.101271417827668e-08, |
|
"loss": 0.7585, |
|
"step": 776 |
|
}, |
|
{ |
|
"epoch": 1.8325471698113207, |
|
"grad_norm": 0.8877158557106878, |
|
"learning_rate": 3.9904256860967433e-08, |
|
"loss": 0.7222, |
|
"step": 777 |
|
}, |
|
{ |
|
"epoch": 1.8349056603773586, |
|
"grad_norm": 0.8633047498966683, |
|
"learning_rate": 3.881067971547469e-08, |
|
"loss": 0.7502, |
|
"step": 778 |
|
}, |
|
{ |
|
"epoch": 1.8372641509433962, |
|
"grad_norm": 1.0520376373477054, |
|
"learning_rate": 3.7731999690749585e-08, |
|
"loss": 0.7443, |
|
"step": 779 |
|
}, |
|
{ |
|
"epoch": 1.8396226415094339, |
|
"grad_norm": 0.7492875546723151, |
|
"learning_rate": 3.666823350485848e-08, |
|
"loss": 0.7781, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 1.8419811320754715, |
|
"grad_norm": 0.7847944050831147, |
|
"learning_rate": 3.561939764472299e-08, |
|
"loss": 0.737, |
|
"step": 781 |
|
}, |
|
{ |
|
"epoch": 1.8443396226415094, |
|
"grad_norm": 1.368419762468144, |
|
"learning_rate": 3.458550836586582e-08, |
|
"loss": 0.7523, |
|
"step": 782 |
|
}, |
|
{ |
|
"epoch": 1.8466981132075473, |
|
"grad_norm": 0.8704033001785062, |
|
"learning_rate": 3.356658169215743e-08, |
|
"loss": 0.742, |
|
"step": 783 |
|
}, |
|
{ |
|
"epoch": 1.849056603773585, |
|
"grad_norm": 0.764412444543419, |
|
"learning_rate": 3.2562633415568754e-08, |
|
"loss": 0.7327, |
|
"step": 784 |
|
}, |
|
{ |
|
"epoch": 1.8514150943396226, |
|
"grad_norm": 0.8695554372256408, |
|
"learning_rate": 3.157367909592601e-08, |
|
"loss": 0.7338, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 1.8537735849056602, |
|
"grad_norm": 0.7794614378925147, |
|
"learning_rate": 3.0599734060669626e-08, |
|
"loss": 0.7446, |
|
"step": 786 |
|
}, |
|
{ |
|
"epoch": 1.8561320754716981, |
|
"grad_norm": 0.7712181969029962, |
|
"learning_rate": 2.9640813404616327e-08, |
|
"loss": 0.7377, |
|
"step": 787 |
|
}, |
|
{ |
|
"epoch": 1.858490566037736, |
|
"grad_norm": 0.8826462854163208, |
|
"learning_rate": 2.869693198972556e-08, |
|
"loss": 0.7555, |
|
"step": 788 |
|
}, |
|
{ |
|
"epoch": 1.8608490566037736, |
|
"grad_norm": 1.029113085742105, |
|
"learning_rate": 2.7768104444869434e-08, |
|
"loss": 0.7795, |
|
"step": 789 |
|
}, |
|
{ |
|
"epoch": 1.8632075471698113, |
|
"grad_norm": 0.8318363964351398, |
|
"learning_rate": 2.6854345165605474e-08, |
|
"loss": 0.7351, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 1.865566037735849, |
|
"grad_norm": 0.7865924156566725, |
|
"learning_rate": 2.595566831395346e-08, |
|
"loss": 0.7222, |
|
"step": 791 |
|
}, |
|
{ |
|
"epoch": 1.8679245283018868, |
|
"grad_norm": 0.7955098420784998, |
|
"learning_rate": 2.507208781817638e-08, |
|
"loss": 0.7515, |
|
"step": 792 |
|
}, |
|
{ |
|
"epoch": 1.8702830188679245, |
|
"grad_norm": 0.9146875621357804, |
|
"learning_rate": 2.4203617372564378e-08, |
|
"loss": 0.7173, |
|
"step": 793 |
|
}, |
|
{ |
|
"epoch": 1.8726415094339623, |
|
"grad_norm": 0.7544835268896075, |
|
"learning_rate": 2.3350270437222374e-08, |
|
"loss": 0.7307, |
|
"step": 794 |
|
}, |
|
{ |
|
"epoch": 1.875, |
|
"grad_norm": 1.806767663849715, |
|
"learning_rate": 2.2512060237861452e-08, |
|
"loss": 0.7301, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 1.8773584905660377, |
|
"grad_norm": 0.804925906379938, |
|
"learning_rate": 2.1688999765594018e-08, |
|
"loss": 0.7552, |
|
"step": 796 |
|
}, |
|
{ |
|
"epoch": 1.8797169811320755, |
|
"grad_norm": 0.7773879448134918, |
|
"learning_rate": 2.0881101776732967e-08, |
|
"loss": 0.7619, |
|
"step": 797 |
|
}, |
|
{ |
|
"epoch": 1.8820754716981132, |
|
"grad_norm": 0.871709115043609, |
|
"learning_rate": 2.0088378792592286e-08, |
|
"loss": 0.7686, |
|
"step": 798 |
|
}, |
|
{ |
|
"epoch": 1.884433962264151, |
|
"grad_norm": 0.7446217889352331, |
|
"learning_rate": 1.9310843099295204e-08, |
|
"loss": 0.7128, |
|
"step": 799 |
|
}, |
|
{ |
|
"epoch": 1.8867924528301887, |
|
"grad_norm": 0.7661065857274754, |
|
"learning_rate": 1.8548506747582128e-08, |
|
"loss": 0.7284, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 1.8891509433962264, |
|
"grad_norm": 0.7916406522727306, |
|
"learning_rate": 1.780138155262456e-08, |
|
"loss": 0.7321, |
|
"step": 801 |
|
}, |
|
{ |
|
"epoch": 1.891509433962264, |
|
"grad_norm": 0.9641121942586747, |
|
"learning_rate": 1.7069479093842042e-08, |
|
"loss": 0.7411, |
|
"step": 802 |
|
}, |
|
{ |
|
"epoch": 1.8938679245283019, |
|
"grad_norm": 0.7930873801533905, |
|
"learning_rate": 1.6352810714722387e-08, |
|
"loss": 0.7345, |
|
"step": 803 |
|
}, |
|
{ |
|
"epoch": 1.8962264150943398, |
|
"grad_norm": 0.9298677041430008, |
|
"learning_rate": 1.565138752264572e-08, |
|
"loss": 0.7616, |
|
"step": 804 |
|
}, |
|
{ |
|
"epoch": 1.8985849056603774, |
|
"grad_norm": 0.885574237101463, |
|
"learning_rate": 1.496522038871295e-08, |
|
"loss": 0.7572, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 1.900943396226415, |
|
"grad_norm": 0.8153461825360933, |
|
"learning_rate": 1.4294319947577017e-08, |
|
"loss": 0.75, |
|
"step": 806 |
|
}, |
|
{ |
|
"epoch": 1.9033018867924527, |
|
"grad_norm": 0.7660700579693493, |
|
"learning_rate": 1.3638696597277677e-08, |
|
"loss": 0.7421, |
|
"step": 807 |
|
}, |
|
{ |
|
"epoch": 1.9056603773584906, |
|
"grad_norm": 0.805964323807633, |
|
"learning_rate": 1.2998360499080763e-08, |
|
"loss": 0.7958, |
|
"step": 808 |
|
}, |
|
{ |
|
"epoch": 1.9080188679245285, |
|
"grad_norm": 0.9429626916004815, |
|
"learning_rate": 1.2373321577320628e-08, |
|
"loss": 0.734, |
|
"step": 809 |
|
}, |
|
{ |
|
"epoch": 1.9103773584905661, |
|
"grad_norm": 0.7904102497155711, |
|
"learning_rate": 1.1763589519246387e-08, |
|
"loss": 0.7478, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 1.9127358490566038, |
|
"grad_norm": 0.7928898152839706, |
|
"learning_rate": 1.1169173774871477e-08, |
|
"loss": 0.7373, |
|
"step": 811 |
|
}, |
|
{ |
|
"epoch": 1.9150943396226414, |
|
"grad_norm": 0.7814723884700895, |
|
"learning_rate": 1.0590083556827556e-08, |
|
"loss": 0.7615, |
|
"step": 812 |
|
}, |
|
{ |
|
"epoch": 1.9174528301886793, |
|
"grad_norm": 0.7802344182322299, |
|
"learning_rate": 1.0026327840221727e-08, |
|
"loss": 0.7503, |
|
"step": 813 |
|
}, |
|
{ |
|
"epoch": 1.919811320754717, |
|
"grad_norm": 1.0812085791485797, |
|
"learning_rate": 9.477915362496758e-09, |
|
"loss": 0.7614, |
|
"step": 814 |
|
}, |
|
{ |
|
"epoch": 1.9221698113207548, |
|
"grad_norm": 0.820203697559199, |
|
"learning_rate": 8.94485462329675e-09, |
|
"loss": 0.7374, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 1.9245283018867925, |
|
"grad_norm": 0.8722587060636673, |
|
"learning_rate": 8.42715388433446e-09, |
|
"loss": 0.7328, |
|
"step": 816 |
|
}, |
|
{ |
|
"epoch": 1.9268867924528301, |
|
"grad_norm": 0.7840984315269908, |
|
"learning_rate": 7.924821169263963e-09, |
|
"loss": 0.7542, |
|
"step": 817 |
|
}, |
|
{ |
|
"epoch": 1.9292452830188678, |
|
"grad_norm": 0.8440330231818731, |
|
"learning_rate": 7.437864263555638e-09, |
|
"loss": 0.7317, |
|
"step": 818 |
|
}, |
|
{ |
|
"epoch": 1.9316037735849056, |
|
"grad_norm": 0.895812932209063, |
|
"learning_rate": 6.966290714375933e-09, |
|
"loss": 0.7409, |
|
"step": 819 |
|
}, |
|
{ |
|
"epoch": 1.9339622641509435, |
|
"grad_norm": 0.899638041623908, |
|
"learning_rate": 6.510107830470568e-09, |
|
"loss": 0.7446, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 1.9363207547169812, |
|
"grad_norm": 0.8152793437016046, |
|
"learning_rate": 6.069322682050515e-09, |
|
"loss": 0.7634, |
|
"step": 821 |
|
}, |
|
{ |
|
"epoch": 1.9386792452830188, |
|
"grad_norm": 0.7735814965641481, |
|
"learning_rate": 5.643942100683308e-09, |
|
"loss": 0.7809, |
|
"step": 822 |
|
}, |
|
{ |
|
"epoch": 1.9410377358490565, |
|
"grad_norm": 0.8351447564551893, |
|
"learning_rate": 5.23397267918646e-09, |
|
"loss": 0.7588, |
|
"step": 823 |
|
}, |
|
{ |
|
"epoch": 1.9433962264150944, |
|
"grad_norm": 0.7614597204440534, |
|
"learning_rate": 4.83942077152577e-09, |
|
"loss": 0.7231, |
|
"step": 824 |
|
}, |
|
{ |
|
"epoch": 1.9457547169811322, |
|
"grad_norm": 0.8548976791386363, |
|
"learning_rate": 4.460292492716511e-09, |
|
"loss": 0.754, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 1.9481132075471699, |
|
"grad_norm": 0.8381107415157564, |
|
"learning_rate": 4.0965937187287246e-09, |
|
"loss": 0.7351, |
|
"step": 826 |
|
}, |
|
{ |
|
"epoch": 1.9504716981132075, |
|
"grad_norm": 5.772455437843555, |
|
"learning_rate": 3.748330086396523e-09, |
|
"loss": 0.7657, |
|
"step": 827 |
|
}, |
|
{ |
|
"epoch": 1.9528301886792452, |
|
"grad_norm": 0.797010233582818, |
|
"learning_rate": 3.415506993330153e-09, |
|
"loss": 0.7248, |
|
"step": 828 |
|
}, |
|
{ |
|
"epoch": 1.955188679245283, |
|
"grad_norm": 0.777781094479457, |
|
"learning_rate": 3.0981295978326216e-09, |
|
"loss": 0.7604, |
|
"step": 829 |
|
}, |
|
{ |
|
"epoch": 1.9575471698113207, |
|
"grad_norm": 0.8449725908943951, |
|
"learning_rate": 2.7962028188198706e-09, |
|
"loss": 0.7736, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 1.9599056603773586, |
|
"grad_norm": 0.7801120467589358, |
|
"learning_rate": 2.5097313357442806e-09, |
|
"loss": 0.7178, |
|
"step": 831 |
|
}, |
|
{ |
|
"epoch": 1.9622641509433962, |
|
"grad_norm": 0.7762312098767981, |
|
"learning_rate": 2.2387195885221756e-09, |
|
"loss": 0.7575, |
|
"step": 832 |
|
}, |
|
{ |
|
"epoch": 1.9646226415094339, |
|
"grad_norm": 0.824362916729911, |
|
"learning_rate": 1.983171777465431e-09, |
|
"loss": 0.76, |
|
"step": 833 |
|
}, |
|
{ |
|
"epoch": 1.9669811320754715, |
|
"grad_norm": 0.9204175438402101, |
|
"learning_rate": 1.743091863215751e-09, |
|
"loss": 0.7312, |
|
"step": 834 |
|
}, |
|
{ |
|
"epoch": 1.9693396226415094, |
|
"grad_norm": 0.8426151332993567, |
|
"learning_rate": 1.5184835666838258e-09, |
|
"loss": 0.7489, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 1.9716981132075473, |
|
"grad_norm": 0.7997774511020088, |
|
"learning_rate": 1.3093503689910467e-09, |
|
"loss": 0.7243, |
|
"step": 836 |
|
}, |
|
{ |
|
"epoch": 1.974056603773585, |
|
"grad_norm": 1.0915014977180026, |
|
"learning_rate": 1.1156955114162147e-09, |
|
"loss": 0.7411, |
|
"step": 837 |
|
}, |
|
{ |
|
"epoch": 1.9764150943396226, |
|
"grad_norm": 0.7394147838620203, |
|
"learning_rate": 9.375219953450253e-10, |
|
"loss": 0.746, |
|
"step": 838 |
|
}, |
|
{ |
|
"epoch": 1.9787735849056602, |
|
"grad_norm": 0.7690476142622722, |
|
"learning_rate": 7.748325822234392e-10, |
|
"loss": 0.729, |
|
"step": 839 |
|
}, |
|
{ |
|
"epoch": 1.9811320754716981, |
|
"grad_norm": 0.9039884352435014, |
|
"learning_rate": 6.276297935149388e-10, |
|
"loss": 0.7569, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 1.983490566037736, |
|
"grad_norm": 1.2814279255523484, |
|
"learning_rate": 4.959159106615596e-10, |
|
"loss": 0.7351, |
|
"step": 841 |
|
}, |
|
{ |
|
"epoch": 1.9858490566037736, |
|
"grad_norm": 0.8004748840113347, |
|
"learning_rate": 3.7969297504858443e-10, |
|
"loss": 0.724, |
|
"step": 842 |
|
}, |
|
{ |
|
"epoch": 1.9882075471698113, |
|
"grad_norm": 0.9234344448728469, |
|
"learning_rate": 2.789627879725698e-10, |
|
"loss": 0.749, |
|
"step": 843 |
|
}, |
|
{ |
|
"epoch": 1.990566037735849, |
|
"grad_norm": 1.0793509856974828, |
|
"learning_rate": 1.9372691061381175e-10, |
|
"loss": 0.7296, |
|
"step": 844 |
|
}, |
|
{ |
|
"epoch": 1.9929245283018868, |
|
"grad_norm": 0.8234018950018985, |
|
"learning_rate": 1.2398666401181035e-10, |
|
"loss": 0.738, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 1.9952830188679245, |
|
"grad_norm": 0.7591071692242605, |
|
"learning_rate": 6.974312904517443e-11, |
|
"loss": 0.7713, |
|
"step": 846 |
|
}, |
|
{ |
|
"epoch": 1.9976415094339623, |
|
"grad_norm": 0.8367565544655438, |
|
"learning_rate": 3.099714641452422e-11, |
|
"loss": 0.7812, |
|
"step": 847 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 0.8006380089754, |
|
"learning_rate": 7.749316629612756e-12, |
|
"loss": 0.769, |
|
"step": 848 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 0.6794618964195251, |
|
"eval_runtime": 82.3418, |
|
"eval_samples_per_second": 8.562, |
|
"eval_steps_per_second": 0.547, |
|
"step": 848 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 848, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 2, |
|
"save_steps": 212, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 8.964690701765837e+17, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|