{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 2.0, "eval_steps": 106, "global_step": 848, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0023584905660377358, "grad_norm": 5.1080776865120034, "learning_rate": 0.0, "loss": 0.9592, "step": 1 }, { "epoch": 0.0023584905660377358, "eval_loss": 0.9177566170692444, "eval_runtime": 83.3901, "eval_samples_per_second": 8.454, "eval_steps_per_second": 0.54, "step": 1 }, { "epoch": 0.0047169811320754715, "grad_norm": 5.216656877082355, "learning_rate": 4e-08, "loss": 0.9643, "step": 2 }, { "epoch": 0.007075471698113208, "grad_norm": 5.86435572061478, "learning_rate": 8e-08, "loss": 0.9448, "step": 3 }, { "epoch": 0.009433962264150943, "grad_norm": 4.885606218308886, "learning_rate": 1.2e-07, "loss": 0.9499, "step": 4 }, { "epoch": 0.01179245283018868, "grad_norm": 4.863782725318177, "learning_rate": 1.6e-07, "loss": 0.9547, "step": 5 }, { "epoch": 0.014150943396226415, "grad_norm": 4.84448661904324, "learning_rate": 2e-07, "loss": 0.9378, "step": 6 }, { "epoch": 0.01650943396226415, "grad_norm": 4.594764533947918, "learning_rate": 2.4e-07, "loss": 0.967, "step": 7 }, { "epoch": 0.018867924528301886, "grad_norm": 5.311507883731841, "learning_rate": 2.8e-07, "loss": 0.9534, "step": 8 }, { "epoch": 0.02122641509433962, "grad_norm": 4.780331285112587, "learning_rate": 3.2e-07, "loss": 0.979, "step": 9 }, { "epoch": 0.02358490566037736, "grad_norm": 5.172814213529754, "learning_rate": 3.6e-07, "loss": 0.9284, "step": 10 }, { "epoch": 0.025943396226415096, "grad_norm": 4.693805464386623, "learning_rate": 4e-07, "loss": 0.9452, "step": 11 }, { "epoch": 0.02830188679245283, "grad_norm": 5.378045080991313, "learning_rate": 4.3999999999999997e-07, "loss": 0.9575, "step": 12 }, { "epoch": 0.030660377358490566, "grad_norm": 4.718804488820069, "learning_rate": 4.8e-07, "loss": 0.9601, "step": 13 }, { "epoch": 0.0330188679245283, "grad_norm": 4.8688696728838785, "learning_rate": 5.2e-07, "loss": 0.9578, "step": 14 }, { "epoch": 0.03537735849056604, "grad_norm": 4.41639309685216, "learning_rate": 5.6e-07, "loss": 0.9433, "step": 15 }, { "epoch": 0.03773584905660377, "grad_norm": 4.330783791521376, "learning_rate": 6e-07, "loss": 0.9498, "step": 16 }, { "epoch": 0.04009433962264151, "grad_norm": 4.580748421983631, "learning_rate": 6.4e-07, "loss": 0.9607, "step": 17 }, { "epoch": 0.04245283018867924, "grad_norm": 3.895307917106911, "learning_rate": 6.800000000000001e-07, "loss": 0.9561, "step": 18 }, { "epoch": 0.04481132075471698, "grad_norm": 3.9063376740559894, "learning_rate": 7.2e-07, "loss": 0.9423, "step": 19 }, { "epoch": 0.04716981132075472, "grad_norm": 3.9785213934452335, "learning_rate": 7.599999999999999e-07, "loss": 0.9422, "step": 20 }, { "epoch": 0.049528301886792456, "grad_norm": 3.5176355010251044, "learning_rate": 8e-07, "loss": 0.9289, "step": 21 }, { "epoch": 0.05188679245283019, "grad_norm": 5.022294330418908, "learning_rate": 8.399999999999999e-07, "loss": 0.9446, "step": 22 }, { "epoch": 0.054245283018867926, "grad_norm": 2.6661772621278637, "learning_rate": 8.799999999999999e-07, "loss": 0.9396, "step": 23 }, { "epoch": 0.05660377358490566, "grad_norm": 2.8934776201699757, "learning_rate": 9.2e-07, "loss": 0.9424, "step": 24 }, { "epoch": 0.0589622641509434, "grad_norm": 3.310951896187245, "learning_rate": 9.6e-07, "loss": 0.9317, "step": 25 }, { "epoch": 0.06132075471698113, "grad_norm": 3.180088649692653, "learning_rate": 1e-06, "loss": 0.8733, "step": 26 }, { "epoch": 0.06367924528301887, "grad_norm": 3.0819569242393032, "learning_rate": 1.04e-06, "loss": 0.9422, "step": 27 }, { "epoch": 0.0660377358490566, "grad_norm": 4.5985187060206405, "learning_rate": 1.08e-06, "loss": 0.9522, "step": 28 }, { "epoch": 0.06839622641509434, "grad_norm": 2.614037871400199, "learning_rate": 1.12e-06, "loss": 0.9168, "step": 29 }, { "epoch": 0.07075471698113207, "grad_norm": 3.4637889884655007, "learning_rate": 1.16e-06, "loss": 0.8819, "step": 30 }, { "epoch": 0.07311320754716981, "grad_norm": 2.7607684537358623, "learning_rate": 1.2e-06, "loss": 0.9443, "step": 31 }, { "epoch": 0.07547169811320754, "grad_norm": 2.45835136955183, "learning_rate": 1.24e-06, "loss": 0.8971, "step": 32 }, { "epoch": 0.07783018867924528, "grad_norm": 2.8319894880867724, "learning_rate": 1.28e-06, "loss": 0.9311, "step": 33 }, { "epoch": 0.08018867924528301, "grad_norm": 1.880755586715356, "learning_rate": 1.32e-06, "loss": 0.895, "step": 34 }, { "epoch": 0.08254716981132075, "grad_norm": 3.1062915440189713, "learning_rate": 1.3600000000000001e-06, "loss": 0.9074, "step": 35 }, { "epoch": 0.08490566037735849, "grad_norm": 2.745579181683214, "learning_rate": 1.4e-06, "loss": 0.9066, "step": 36 }, { "epoch": 0.08726415094339622, "grad_norm": 2.51965653055992, "learning_rate": 1.44e-06, "loss": 0.8737, "step": 37 }, { "epoch": 0.08962264150943396, "grad_norm": 2.324394856248302, "learning_rate": 1.48e-06, "loss": 0.9239, "step": 38 }, { "epoch": 0.09198113207547169, "grad_norm": 2.7275453116030834, "learning_rate": 1.5199999999999998e-06, "loss": 0.9029, "step": 39 }, { "epoch": 0.09433962264150944, "grad_norm": 2.8394394816071498, "learning_rate": 1.5599999999999999e-06, "loss": 0.885, "step": 40 }, { "epoch": 0.09669811320754718, "grad_norm": 2.3797648877001323, "learning_rate": 1.6e-06, "loss": 0.9084, "step": 41 }, { "epoch": 0.09905660377358491, "grad_norm": 2.298049507841523, "learning_rate": 1.6399999999999998e-06, "loss": 0.8786, "step": 42 }, { "epoch": 0.10141509433962265, "grad_norm": 2.0882957626327814, "learning_rate": 1.6799999999999998e-06, "loss": 0.885, "step": 43 }, { "epoch": 0.10377358490566038, "grad_norm": 1.9805393143682433, "learning_rate": 1.7199999999999998e-06, "loss": 0.9202, "step": 44 }, { "epoch": 0.10613207547169812, "grad_norm": 2.1136265646711605, "learning_rate": 1.7599999999999999e-06, "loss": 0.8515, "step": 45 }, { "epoch": 0.10849056603773585, "grad_norm": 1.7161578870903316, "learning_rate": 1.8e-06, "loss": 0.9035, "step": 46 }, { "epoch": 0.11084905660377359, "grad_norm": 2.2352887975077795, "learning_rate": 1.84e-06, "loss": 0.8615, "step": 47 }, { "epoch": 0.11320754716981132, "grad_norm": 1.997343145683379, "learning_rate": 1.8799999999999998e-06, "loss": 0.8958, "step": 48 }, { "epoch": 0.11556603773584906, "grad_norm": 2.1520613046888024, "learning_rate": 1.92e-06, "loss": 0.8706, "step": 49 }, { "epoch": 0.1179245283018868, "grad_norm": 1.610078026570572, "learning_rate": 1.96e-06, "loss": 0.9127, "step": 50 }, { "epoch": 0.12028301886792453, "grad_norm": 1.6267119325099246, "learning_rate": 2e-06, "loss": 0.8604, "step": 51 }, { "epoch": 0.12264150943396226, "grad_norm": 1.7305415242899538, "learning_rate": 1.99999225068337e-06, "loss": 0.8704, "step": 52 }, { "epoch": 0.125, "grad_norm": 4.837245274184259, "learning_rate": 1.9999690028535855e-06, "loss": 0.9003, "step": 53 }, { "epoch": 0.12735849056603774, "grad_norm": 2.006265922883068, "learning_rate": 1.9999302568709546e-06, "loss": 0.8759, "step": 54 }, { "epoch": 0.12971698113207547, "grad_norm": 1.8749359565723283, "learning_rate": 1.999876013335988e-06, "loss": 0.8686, "step": 55 }, { "epoch": 0.1320754716981132, "grad_norm": 1.5041642645696787, "learning_rate": 1.9998062730893862e-06, "loss": 0.8727, "step": 56 }, { "epoch": 0.13443396226415094, "grad_norm": 1.4630749706626824, "learning_rate": 1.9997210372120272e-06, "loss": 0.8718, "step": 57 }, { "epoch": 0.13679245283018868, "grad_norm": 1.4357346298179297, "learning_rate": 1.9996203070249514e-06, "loss": 0.9, "step": 58 }, { "epoch": 0.1391509433962264, "grad_norm": 1.589349788137753, "learning_rate": 1.9995040840893383e-06, "loss": 0.8832, "step": 59 }, { "epoch": 0.14150943396226415, "grad_norm": 1.3078655659788383, "learning_rate": 1.9993723702064853e-06, "loss": 0.8689, "step": 60 }, { "epoch": 0.14386792452830188, "grad_norm": 1.2162726319357984, "learning_rate": 1.9992251674177763e-06, "loss": 0.8565, "step": 61 }, { "epoch": 0.14622641509433962, "grad_norm": 1.4682400986060873, "learning_rate": 1.999062478004655e-06, "loss": 0.8768, "step": 62 }, { "epoch": 0.14858490566037735, "grad_norm": 1.475835127346073, "learning_rate": 1.9988843044885837e-06, "loss": 0.8485, "step": 63 }, { "epoch": 0.1509433962264151, "grad_norm": 1.313661603517591, "learning_rate": 1.998690649631009e-06, "loss": 0.8791, "step": 64 }, { "epoch": 0.15330188679245282, "grad_norm": 1.2318794472546806, "learning_rate": 1.998481516433316e-06, "loss": 0.8418, "step": 65 }, { "epoch": 0.15566037735849056, "grad_norm": 1.224129101101457, "learning_rate": 1.9982569081367843e-06, "loss": 0.8472, "step": 66 }, { "epoch": 0.1580188679245283, "grad_norm": 1.4354409800877264, "learning_rate": 1.9980168282225344e-06, "loss": 0.8682, "step": 67 }, { "epoch": 0.16037735849056603, "grad_norm": 1.3882277723705556, "learning_rate": 1.9977612804114775e-06, "loss": 0.8929, "step": 68 }, { "epoch": 0.16273584905660377, "grad_norm": 1.1353588967534465, "learning_rate": 1.9974902686642557e-06, "loss": 0.8123, "step": 69 }, { "epoch": 0.1650943396226415, "grad_norm": 1.1680380889234738, "learning_rate": 1.9972037971811797e-06, "loss": 0.8573, "step": 70 }, { "epoch": 0.16745283018867924, "grad_norm": 1.1797181726680495, "learning_rate": 1.9969018704021675e-06, "loss": 0.8518, "step": 71 }, { "epoch": 0.16981132075471697, "grad_norm": 1.3076313839386762, "learning_rate": 1.9965844930066696e-06, "loss": 0.8425, "step": 72 }, { "epoch": 0.1721698113207547, "grad_norm": 1.2192501506435032, "learning_rate": 1.9962516699136036e-06, "loss": 0.8633, "step": 73 }, { "epoch": 0.17452830188679244, "grad_norm": 1.1885172994924504, "learning_rate": 1.9959034062812714e-06, "loss": 0.8417, "step": 74 }, { "epoch": 0.17688679245283018, "grad_norm": 1.0228274366877697, "learning_rate": 1.9955397075072833e-06, "loss": 0.8645, "step": 75 }, { "epoch": 0.1792452830188679, "grad_norm": 1.0362433430443114, "learning_rate": 1.9951605792284742e-06, "loss": 0.8442, "step": 76 }, { "epoch": 0.18160377358490565, "grad_norm": 1.3774246732359812, "learning_rate": 1.9947660273208134e-06, "loss": 0.8355, "step": 77 }, { "epoch": 0.18396226415094338, "grad_norm": 1.3163064309206076, "learning_rate": 1.9943560578993165e-06, "loss": 0.8246, "step": 78 }, { "epoch": 0.18632075471698112, "grad_norm": 1.0822503373501942, "learning_rate": 1.9939306773179494e-06, "loss": 0.8315, "step": 79 }, { "epoch": 0.18867924528301888, "grad_norm": 1.0361871014520703, "learning_rate": 1.9934898921695292e-06, "loss": 0.8346, "step": 80 }, { "epoch": 0.19103773584905662, "grad_norm": 1.0371978245512512, "learning_rate": 1.993033709285624e-06, "loss": 0.8706, "step": 81 }, { "epoch": 0.19339622641509435, "grad_norm": 1.0574050613888015, "learning_rate": 1.992562135736444e-06, "loss": 0.8308, "step": 82 }, { "epoch": 0.1957547169811321, "grad_norm": 1.241296877885062, "learning_rate": 1.992075178830736e-06, "loss": 0.872, "step": 83 }, { "epoch": 0.19811320754716982, "grad_norm": 1.152024986745144, "learning_rate": 1.9915728461156654e-06, "loss": 0.8365, "step": 84 }, { "epoch": 0.20047169811320756, "grad_norm": 1.0972286027771576, "learning_rate": 1.991055145376703e-06, "loss": 0.8511, "step": 85 }, { "epoch": 0.2028301886792453, "grad_norm": 1.3363467615851807, "learning_rate": 1.990522084637503e-06, "loss": 0.8604, "step": 86 }, { "epoch": 0.20518867924528303, "grad_norm": 0.9249543605316475, "learning_rate": 1.9899736721597786e-06, "loss": 0.8078, "step": 87 }, { "epoch": 0.20754716981132076, "grad_norm": 1.028197337537026, "learning_rate": 1.9894099164431722e-06, "loss": 0.8572, "step": 88 }, { "epoch": 0.2099056603773585, "grad_norm": 0.9345635713942412, "learning_rate": 1.9888308262251284e-06, "loss": 0.814, "step": 89 }, { "epoch": 0.21226415094339623, "grad_norm": 1.0476770833253457, "learning_rate": 1.9882364104807535e-06, "loss": 0.8358, "step": 90 }, { "epoch": 0.21462264150943397, "grad_norm": 1.0517946592201646, "learning_rate": 1.9876266784226794e-06, "loss": 0.8263, "step": 91 }, { "epoch": 0.2169811320754717, "grad_norm": 1.1715532913432805, "learning_rate": 1.987001639500919e-06, "loss": 0.8268, "step": 92 }, { "epoch": 0.21933962264150944, "grad_norm": 0.928474382958498, "learning_rate": 1.9863613034027223e-06, "loss": 0.8278, "step": 93 }, { "epoch": 0.22169811320754718, "grad_norm": 0.9243352298229107, "learning_rate": 1.985705680052423e-06, "loss": 0.817, "step": 94 }, { "epoch": 0.2240566037735849, "grad_norm": 0.9825897380291061, "learning_rate": 1.985034779611287e-06, "loss": 0.8185, "step": 95 }, { "epoch": 0.22641509433962265, "grad_norm": 1.0129581531905947, "learning_rate": 1.9843486124773543e-06, "loss": 0.8261, "step": 96 }, { "epoch": 0.22877358490566038, "grad_norm": 0.9634164693430555, "learning_rate": 1.9836471892852777e-06, "loss": 0.8448, "step": 97 }, { "epoch": 0.23113207547169812, "grad_norm": 0.9208269074792377, "learning_rate": 1.982930520906158e-06, "loss": 0.8435, "step": 98 }, { "epoch": 0.23349056603773585, "grad_norm": 1.2744902927155426, "learning_rate": 1.9821986184473754e-06, "loss": 0.811, "step": 99 }, { "epoch": 0.2358490566037736, "grad_norm": 0.9132891996164993, "learning_rate": 1.9814514932524176e-06, "loss": 0.8629, "step": 100 }, { "epoch": 0.23820754716981132, "grad_norm": 1.3924445022644105, "learning_rate": 1.9806891569007048e-06, "loss": 0.8157, "step": 101 }, { "epoch": 0.24056603773584906, "grad_norm": 0.9910775720488755, "learning_rate": 1.9799116212074075e-06, "loss": 0.8133, "step": 102 }, { "epoch": 0.2429245283018868, "grad_norm": 0.8521457201237292, "learning_rate": 1.979118898223267e-06, "loss": 0.818, "step": 103 }, { "epoch": 0.24528301886792453, "grad_norm": 0.961507205526783, "learning_rate": 1.978311000234406e-06, "loss": 0.8312, "step": 104 }, { "epoch": 0.24764150943396226, "grad_norm": 0.9234203353918131, "learning_rate": 1.9774879397621383e-06, "loss": 0.8307, "step": 105 }, { "epoch": 0.25, "grad_norm": 0.939970736839533, "learning_rate": 1.9766497295627777e-06, "loss": 0.8121, "step": 106 }, { "epoch": 0.25, "eval_loss": 0.7522591948509216, "eval_runtime": 82.8066, "eval_samples_per_second": 8.514, "eval_steps_per_second": 0.543, "step": 106 }, { "epoch": 0.25235849056603776, "grad_norm": 1.009636311779919, "learning_rate": 1.9757963826274354e-06, "loss": 0.8321, "step": 107 }, { "epoch": 0.25471698113207547, "grad_norm": 0.8979616538385209, "learning_rate": 1.9749279121818236e-06, "loss": 0.8442, "step": 108 }, { "epoch": 0.25707547169811323, "grad_norm": 0.953771354540708, "learning_rate": 1.9740443316860463e-06, "loss": 0.8484, "step": 109 }, { "epoch": 0.25943396226415094, "grad_norm": 1.0957061062800166, "learning_rate": 1.9731456548343944e-06, "loss": 0.8204, "step": 110 }, { "epoch": 0.2617924528301887, "grad_norm": 0.8834656211178301, "learning_rate": 1.9722318955551303e-06, "loss": 0.7817, "step": 111 }, { "epoch": 0.2641509433962264, "grad_norm": 1.0070698705880765, "learning_rate": 1.9713030680102743e-06, "loss": 0.8309, "step": 112 }, { "epoch": 0.2665094339622642, "grad_norm": 1.0103166521916906, "learning_rate": 1.970359186595384e-06, "loss": 0.8454, "step": 113 }, { "epoch": 0.2688679245283019, "grad_norm": 0.7965830780073067, "learning_rate": 1.9694002659393305e-06, "loss": 0.7659, "step": 114 }, { "epoch": 0.27122641509433965, "grad_norm": 0.9676729525657803, "learning_rate": 1.968426320904074e-06, "loss": 0.8076, "step": 115 }, { "epoch": 0.27358490566037735, "grad_norm": 0.8547052227174852, "learning_rate": 1.967437366584431e-06, "loss": 0.8305, "step": 116 }, { "epoch": 0.2759433962264151, "grad_norm": 0.9129048151834617, "learning_rate": 1.9664334183078425e-06, "loss": 0.8443, "step": 117 }, { "epoch": 0.2783018867924528, "grad_norm": 0.8521299847115695, "learning_rate": 1.965414491634134e-06, "loss": 0.8244, "step": 118 }, { "epoch": 0.2806603773584906, "grad_norm": 0.8875033746133125, "learning_rate": 1.964380602355277e-06, "loss": 0.7855, "step": 119 }, { "epoch": 0.2830188679245283, "grad_norm": 0.8322961791869938, "learning_rate": 1.9633317664951417e-06, "loss": 0.8246, "step": 120 }, { "epoch": 0.28537735849056606, "grad_norm": 0.8996547969201056, "learning_rate": 1.9622680003092503e-06, "loss": 0.7911, "step": 121 }, { "epoch": 0.28773584905660377, "grad_norm": 0.8421734340822833, "learning_rate": 1.9611893202845253e-06, "loss": 0.8075, "step": 122 }, { "epoch": 0.29009433962264153, "grad_norm": 0.9715000278619699, "learning_rate": 1.9600957431390324e-06, "loss": 0.8258, "step": 123 }, { "epoch": 0.29245283018867924, "grad_norm": 0.9486047429972293, "learning_rate": 1.9589872858217233e-06, "loss": 0.8249, "step": 124 }, { "epoch": 0.294811320754717, "grad_norm": 0.8708681162275084, "learning_rate": 1.9578639655121707e-06, "loss": 0.8061, "step": 125 }, { "epoch": 0.2971698113207547, "grad_norm": 0.815600627546263, "learning_rate": 1.9567257996203046e-06, "loss": 0.8117, "step": 126 }, { "epoch": 0.29952830188679247, "grad_norm": 0.899644949674199, "learning_rate": 1.955572805786141e-06, "loss": 0.8324, "step": 127 }, { "epoch": 0.3018867924528302, "grad_norm": 0.7797228121513826, "learning_rate": 1.9544050018795075e-06, "loss": 0.8085, "step": 128 }, { "epoch": 0.30424528301886794, "grad_norm": 1.4425571525030387, "learning_rate": 1.953222405999769e-06, "loss": 0.8043, "step": 129 }, { "epoch": 0.30660377358490565, "grad_norm": 1.3224705998967385, "learning_rate": 1.9520250364755458e-06, "loss": 0.8104, "step": 130 }, { "epoch": 0.3089622641509434, "grad_norm": 0.8148761244743041, "learning_rate": 1.9508129118644293e-06, "loss": 0.828, "step": 131 }, { "epoch": 0.3113207547169811, "grad_norm": 1.0994942919129091, "learning_rate": 1.949586050952693e-06, "loss": 0.8007, "step": 132 }, { "epoch": 0.3136792452830189, "grad_norm": 0.9838294161632027, "learning_rate": 1.9483444727550054e-06, "loss": 0.8304, "step": 133 }, { "epoch": 0.3160377358490566, "grad_norm": 1.0206044534349468, "learning_rate": 1.9470881965141307e-06, "loss": 0.8044, "step": 134 }, { "epoch": 0.31839622641509435, "grad_norm": 0.8186471659261382, "learning_rate": 1.9458172417006346e-06, "loss": 0.7921, "step": 135 }, { "epoch": 0.32075471698113206, "grad_norm": 1.040687215576322, "learning_rate": 1.944531628012579e-06, "loss": 0.8309, "step": 136 }, { "epoch": 0.3231132075471698, "grad_norm": 0.8067576806018053, "learning_rate": 1.9432313753752194e-06, "loss": 0.8006, "step": 137 }, { "epoch": 0.32547169811320753, "grad_norm": 1.1895556658600497, "learning_rate": 1.941916503940694e-06, "loss": 0.8055, "step": 138 }, { "epoch": 0.3278301886792453, "grad_norm": 0.877104803725933, "learning_rate": 1.9405870340877135e-06, "loss": 0.815, "step": 139 }, { "epoch": 0.330188679245283, "grad_norm": 1.0755873765351904, "learning_rate": 1.9392429864212433e-06, "loss": 0.8357, "step": 140 }, { "epoch": 0.33254716981132076, "grad_norm": 0.8389051921712227, "learning_rate": 1.9378843817721854e-06, "loss": 0.8054, "step": 141 }, { "epoch": 0.33490566037735847, "grad_norm": 0.8492954330518299, "learning_rate": 1.9365112411970546e-06, "loss": 0.8125, "step": 142 }, { "epoch": 0.33726415094339623, "grad_norm": 0.8117108742521268, "learning_rate": 1.9351235859776537e-06, "loss": 0.7996, "step": 143 }, { "epoch": 0.33962264150943394, "grad_norm": 0.8734747925195555, "learning_rate": 1.9337214376207417e-06, "loss": 0.7987, "step": 144 }, { "epoch": 0.3419811320754717, "grad_norm": 0.870843122026526, "learning_rate": 1.932304817857702e-06, "loss": 0.8071, "step": 145 }, { "epoch": 0.3443396226415094, "grad_norm": 0.7823387766729867, "learning_rate": 1.930873748644204e-06, "loss": 0.781, "step": 146 }, { "epoch": 0.3466981132075472, "grad_norm": 1.0441516146481582, "learning_rate": 1.9294282521598657e-06, "loss": 0.8211, "step": 147 }, { "epoch": 0.3490566037735849, "grad_norm": 0.8689106447919821, "learning_rate": 1.927968350807906e-06, "loss": 0.7827, "step": 148 }, { "epoch": 0.35141509433962265, "grad_norm": 0.8369054883974562, "learning_rate": 1.9264940672148015e-06, "loss": 0.7987, "step": 149 }, { "epoch": 0.35377358490566035, "grad_norm": 1.628472390625535, "learning_rate": 1.9250054242299326e-06, "loss": 0.8245, "step": 150 }, { "epoch": 0.3561320754716981, "grad_norm": 0.8210347316488275, "learning_rate": 1.9235024449252305e-06, "loss": 0.8026, "step": 151 }, { "epoch": 0.3584905660377358, "grad_norm": 0.8976272281891767, "learning_rate": 1.9219851525948203e-06, "loss": 0.8343, "step": 152 }, { "epoch": 0.3608490566037736, "grad_norm": 0.8688278315807118, "learning_rate": 1.92045357075466e-06, "loss": 0.8378, "step": 153 }, { "epoch": 0.3632075471698113, "grad_norm": 0.8259924861279039, "learning_rate": 1.9189077231421746e-06, "loss": 0.771, "step": 154 }, { "epoch": 0.36556603773584906, "grad_norm": 0.850617328037459, "learning_rate": 1.917347633715889e-06, "loss": 0.8116, "step": 155 }, { "epoch": 0.36792452830188677, "grad_norm": 0.830159427983018, "learning_rate": 1.915773326655057e-06, "loss": 0.8142, "step": 156 }, { "epoch": 0.37028301886792453, "grad_norm": 0.9047033822843099, "learning_rate": 1.9141848263592873e-06, "loss": 0.7867, "step": 157 }, { "epoch": 0.37264150943396224, "grad_norm": 0.8043279871245154, "learning_rate": 1.9125821574481623e-06, "loss": 0.7821, "step": 158 }, { "epoch": 0.375, "grad_norm": 0.9855797219219302, "learning_rate": 1.9109653447608605e-06, "loss": 0.838, "step": 159 }, { "epoch": 0.37735849056603776, "grad_norm": 0.8671907333946726, "learning_rate": 1.909334413355768e-06, "loss": 0.8199, "step": 160 }, { "epoch": 0.37971698113207547, "grad_norm": 0.8038519157610383, "learning_rate": 1.9076893885100934e-06, "loss": 0.7914, "step": 161 }, { "epoch": 0.38207547169811323, "grad_norm": 0.8534138715990678, "learning_rate": 1.906030295719473e-06, "loss": 0.803, "step": 162 }, { "epoch": 0.38443396226415094, "grad_norm": 1.272101513607124, "learning_rate": 1.9043571606975775e-06, "loss": 0.8243, "step": 163 }, { "epoch": 0.3867924528301887, "grad_norm": 0.8724688206578753, "learning_rate": 1.9026700093757129e-06, "loss": 0.8, "step": 164 }, { "epoch": 0.3891509433962264, "grad_norm": 0.8272714642025372, "learning_rate": 1.9009688679024189e-06, "loss": 0.7941, "step": 165 }, { "epoch": 0.3915094339622642, "grad_norm": 1.609360832798017, "learning_rate": 1.8992537626430636e-06, "loss": 0.8184, "step": 166 }, { "epoch": 0.3938679245283019, "grad_norm": 1.068913128886596, "learning_rate": 1.897524720179434e-06, "loss": 0.7551, "step": 167 }, { "epoch": 0.39622641509433965, "grad_norm": 0.876062488501413, "learning_rate": 1.8957817673093256e-06, "loss": 0.8202, "step": 168 }, { "epoch": 0.39858490566037735, "grad_norm": 0.7804659272450435, "learning_rate": 1.894024931046125e-06, "loss": 0.8096, "step": 169 }, { "epoch": 0.4009433962264151, "grad_norm": 1.1020958270083518, "learning_rate": 1.8922542386183939e-06, "loss": 0.8023, "step": 170 }, { "epoch": 0.4033018867924528, "grad_norm": 0.8307091625451186, "learning_rate": 1.8904697174694446e-06, "loss": 0.797, "step": 171 }, { "epoch": 0.4056603773584906, "grad_norm": 1.4660516537985, "learning_rate": 1.8886713952569156e-06, "loss": 0.8001, "step": 172 }, { "epoch": 0.4080188679245283, "grad_norm": 0.9042037911584216, "learning_rate": 1.8868592998523436e-06, "loss": 0.7798, "step": 173 }, { "epoch": 0.41037735849056606, "grad_norm": 0.8127436933722597, "learning_rate": 1.885033459340731e-06, "loss": 0.791, "step": 174 }, { "epoch": 0.41273584905660377, "grad_norm": 0.8906980453056944, "learning_rate": 1.8831939020201096e-06, "loss": 0.8117, "step": 175 }, { "epoch": 0.41509433962264153, "grad_norm": 0.7811131450862968, "learning_rate": 1.8813406564011044e-06, "loss": 0.742, "step": 176 }, { "epoch": 0.41745283018867924, "grad_norm": 2.0285717851138974, "learning_rate": 1.8794737512064888e-06, "loss": 0.803, "step": 177 }, { "epoch": 0.419811320754717, "grad_norm": 0.9681303046265977, "learning_rate": 1.8775932153707426e-06, "loss": 0.7857, "step": 178 }, { "epoch": 0.4221698113207547, "grad_norm": 0.8312813398366425, "learning_rate": 1.8756990780396006e-06, "loss": 0.8091, "step": 179 }, { "epoch": 0.42452830188679247, "grad_norm": 0.8325963532837448, "learning_rate": 1.8737913685696027e-06, "loss": 0.7936, "step": 180 }, { "epoch": 0.4268867924528302, "grad_norm": 1.0242582555350113, "learning_rate": 1.8718701165276383e-06, "loss": 0.7679, "step": 181 }, { "epoch": 0.42924528301886794, "grad_norm": 0.8568509432407337, "learning_rate": 1.869935351690488e-06, "loss": 0.7813, "step": 182 }, { "epoch": 0.43160377358490565, "grad_norm": 1.1130973261677584, "learning_rate": 1.867987104044363e-06, "loss": 0.8173, "step": 183 }, { "epoch": 0.4339622641509434, "grad_norm": 0.8076547906696275, "learning_rate": 1.8660254037844386e-06, "loss": 0.7823, "step": 184 }, { "epoch": 0.4363207547169811, "grad_norm": 0.8462497637993398, "learning_rate": 1.864050281314388e-06, "loss": 0.7964, "step": 185 }, { "epoch": 0.4386792452830189, "grad_norm": 0.8157104951412149, "learning_rate": 1.8620617672459096e-06, "loss": 0.7851, "step": 186 }, { "epoch": 0.4410377358490566, "grad_norm": 0.8016398797843809, "learning_rate": 1.8600598923982537e-06, "loss": 0.7939, "step": 187 }, { "epoch": 0.44339622641509435, "grad_norm": 0.9872324739847712, "learning_rate": 1.858044687797745e-06, "loss": 0.8105, "step": 188 }, { "epoch": 0.44575471698113206, "grad_norm": 0.8070225439686687, "learning_rate": 1.8560161846773e-06, "loss": 0.7972, "step": 189 }, { "epoch": 0.4481132075471698, "grad_norm": 0.8159111285487299, "learning_rate": 1.8539744144759447e-06, "loss": 0.7956, "step": 190 }, { "epoch": 0.45047169811320753, "grad_norm": 0.8271007147512706, "learning_rate": 1.851919408838327e-06, "loss": 0.801, "step": 191 }, { "epoch": 0.4528301886792453, "grad_norm": 0.8240075084440164, "learning_rate": 1.8498511996142253e-06, "loss": 0.8131, "step": 192 }, { "epoch": 0.455188679245283, "grad_norm": 0.8100827961239685, "learning_rate": 1.8477698188580557e-06, "loss": 0.7953, "step": 193 }, { "epoch": 0.45754716981132076, "grad_norm": 0.8090963621477647, "learning_rate": 1.8456752988283757e-06, "loss": 0.7713, "step": 194 }, { "epoch": 0.45990566037735847, "grad_norm": 0.8646115465094427, "learning_rate": 1.8435676719873827e-06, "loss": 0.8218, "step": 195 }, { "epoch": 0.46226415094339623, "grad_norm": 0.8283897647434588, "learning_rate": 1.8414469710004124e-06, "loss": 0.7937, "step": 196 }, { "epoch": 0.46462264150943394, "grad_norm": 0.8130051258008277, "learning_rate": 1.839313228735431e-06, "loss": 0.7822, "step": 197 }, { "epoch": 0.4669811320754717, "grad_norm": 0.8754673959382613, "learning_rate": 1.8371664782625285e-06, "loss": 0.7903, "step": 198 }, { "epoch": 0.4693396226415094, "grad_norm": 1.287939397478769, "learning_rate": 1.8350067528534024e-06, "loss": 0.7718, "step": 199 }, { "epoch": 0.4716981132075472, "grad_norm": 0.8236848990785427, "learning_rate": 1.8328340859808446e-06, "loss": 0.8012, "step": 200 }, { "epoch": 0.4740566037735849, "grad_norm": 0.8938042537260998, "learning_rate": 1.8306485113182229e-06, "loss": 0.7721, "step": 201 }, { "epoch": 0.47641509433962265, "grad_norm": 0.8496295751844541, "learning_rate": 1.8284500627389567e-06, "loss": 0.8043, "step": 202 }, { "epoch": 0.47877358490566035, "grad_norm": 0.8325387527055562, "learning_rate": 1.8262387743159948e-06, "loss": 0.7936, "step": 203 }, { "epoch": 0.4811320754716981, "grad_norm": 0.8207579274147835, "learning_rate": 1.824014680321285e-06, "loss": 0.8153, "step": 204 }, { "epoch": 0.4834905660377358, "grad_norm": 0.9100175298073298, "learning_rate": 1.821777815225245e-06, "loss": 0.777, "step": 205 }, { "epoch": 0.4858490566037736, "grad_norm": 0.9563101605978098, "learning_rate": 1.8195282136962264e-06, "loss": 0.7991, "step": 206 }, { "epoch": 0.4882075471698113, "grad_norm": 0.8404681193002833, "learning_rate": 1.817265910599978e-06, "loss": 0.7968, "step": 207 }, { "epoch": 0.49056603773584906, "grad_norm": 0.898412624742003, "learning_rate": 1.814990940999106e-06, "loss": 0.7585, "step": 208 }, { "epoch": 0.49292452830188677, "grad_norm": 0.8495327539856999, "learning_rate": 1.81270334015253e-06, "loss": 0.8133, "step": 209 }, { "epoch": 0.49528301886792453, "grad_norm": 1.1225882373174563, "learning_rate": 1.8104031435149362e-06, "loss": 0.8121, "step": 210 }, { "epoch": 0.49764150943396224, "grad_norm": 0.9329252376646318, "learning_rate": 1.8080903867362293e-06, "loss": 0.815, "step": 211 }, { "epoch": 0.5, "grad_norm": 1.0591364512578068, "learning_rate": 1.8057651056609782e-06, "loss": 0.7649, "step": 212 }, { "epoch": 0.5, "eval_loss": 0.714023768901825, "eval_runtime": 83.0267, "eval_samples_per_second": 8.491, "eval_steps_per_second": 0.542, "step": 212 }, { "epoch": 0.5023584905660378, "grad_norm": 0.8614390398527076, "learning_rate": 1.803427336327861e-06, "loss": 0.8039, "step": 213 }, { "epoch": 0.5047169811320755, "grad_norm": 0.903956531352285, "learning_rate": 1.8010771149691082e-06, "loss": 0.784, "step": 214 }, { "epoch": 0.5070754716981132, "grad_norm": 0.8574588324038701, "learning_rate": 1.7987144780099376e-06, "loss": 0.7956, "step": 215 }, { "epoch": 0.5094339622641509, "grad_norm": 0.8976197887421247, "learning_rate": 1.7963394620679942e-06, "loss": 0.8261, "step": 216 }, { "epoch": 0.5117924528301887, "grad_norm": 0.8410804453164143, "learning_rate": 1.7939521039527781e-06, "loss": 0.7816, "step": 217 }, { "epoch": 0.5141509433962265, "grad_norm": 0.903488950014599, "learning_rate": 1.7915524406650775e-06, "loss": 0.7701, "step": 218 }, { "epoch": 0.5165094339622641, "grad_norm": 0.8374711197328489, "learning_rate": 1.7891405093963937e-06, "loss": 0.7761, "step": 219 }, { "epoch": 0.5188679245283019, "grad_norm": 0.938187408138231, "learning_rate": 1.7867163475283646e-06, "loss": 0.8247, "step": 220 }, { "epoch": 0.5212264150943396, "grad_norm": 1.974177068847498, "learning_rate": 1.7842799926321863e-06, "loss": 0.7984, "step": 221 }, { "epoch": 0.5235849056603774, "grad_norm": 0.8445314748349818, "learning_rate": 1.7818314824680298e-06, "loss": 0.7783, "step": 222 }, { "epoch": 0.5259433962264151, "grad_norm": 0.8043844477824647, "learning_rate": 1.779370854984456e-06, "loss": 0.7748, "step": 223 }, { "epoch": 0.5283018867924528, "grad_norm": 0.8199345888971309, "learning_rate": 1.7768981483178279e-06, "loss": 0.7961, "step": 224 }, { "epoch": 0.5306603773584906, "grad_norm": 0.8198791332861125, "learning_rate": 1.7744134007917194e-06, "loss": 0.7809, "step": 225 }, { "epoch": 0.5330188679245284, "grad_norm": 0.8038089067621491, "learning_rate": 1.7719166509163208e-06, "loss": 0.752, "step": 226 }, { "epoch": 0.535377358490566, "grad_norm": 0.8115733374512414, "learning_rate": 1.7694079373878433e-06, "loss": 0.7977, "step": 227 }, { "epoch": 0.5377358490566038, "grad_norm": 0.8325145007355951, "learning_rate": 1.7668872990879173e-06, "loss": 0.775, "step": 228 }, { "epoch": 0.5400943396226415, "grad_norm": 0.8029142353641874, "learning_rate": 1.7643547750829918e-06, "loss": 0.7897, "step": 229 }, { "epoch": 0.5424528301886793, "grad_norm": 0.8068710232788708, "learning_rate": 1.7618104046237274e-06, "loss": 0.7869, "step": 230 }, { "epoch": 0.5448113207547169, "grad_norm": 0.8179053896349096, "learning_rate": 1.7592542271443887e-06, "loss": 0.7737, "step": 231 }, { "epoch": 0.5471698113207547, "grad_norm": 0.9236859332969564, "learning_rate": 1.7566862822622328e-06, "loss": 0.7779, "step": 232 }, { "epoch": 0.5495283018867925, "grad_norm": 0.8105651817226325, "learning_rate": 1.754106609776896e-06, "loss": 0.7841, "step": 233 }, { "epoch": 0.5518867924528302, "grad_norm": 0.8955884533729873, "learning_rate": 1.7515152496697763e-06, "loss": 0.8077, "step": 234 }, { "epoch": 0.5542452830188679, "grad_norm": 1.015222493683735, "learning_rate": 1.748912242103413e-06, "loss": 0.8013, "step": 235 }, { "epoch": 0.5566037735849056, "grad_norm": 0.9309116956021466, "learning_rate": 1.746297627420866e-06, "loss": 0.7672, "step": 236 }, { "epoch": 0.5589622641509434, "grad_norm": 0.9599117437990355, "learning_rate": 1.7436714461450897e-06, "loss": 0.7957, "step": 237 }, { "epoch": 0.5613207547169812, "grad_norm": 0.8340801449927656, "learning_rate": 1.7410337389783041e-06, "loss": 0.7994, "step": 238 }, { "epoch": 0.5636792452830188, "grad_norm": 0.788362523807564, "learning_rate": 1.7383845468013654e-06, "loss": 0.7707, "step": 239 }, { "epoch": 0.5660377358490566, "grad_norm": 0.8572171206767668, "learning_rate": 1.7357239106731317e-06, "loss": 0.7931, "step": 240 }, { "epoch": 0.5683962264150944, "grad_norm": 4.180247592376519, "learning_rate": 1.733051871829826e-06, "loss": 0.798, "step": 241 }, { "epoch": 0.5707547169811321, "grad_norm": 0.8738642652249169, "learning_rate": 1.7303684716843995e-06, "loss": 0.7768, "step": 242 }, { "epoch": 0.5731132075471698, "grad_norm": 0.8413513234438594, "learning_rate": 1.727673751825886e-06, "loss": 0.8139, "step": 243 }, { "epoch": 0.5754716981132075, "grad_norm": 0.8032830413975406, "learning_rate": 1.7249677540187609e-06, "loss": 0.7877, "step": 244 }, { "epoch": 0.5778301886792453, "grad_norm": 0.7992641395875911, "learning_rate": 1.7222505202022913e-06, "loss": 0.7712, "step": 245 }, { "epoch": 0.5801886792452831, "grad_norm": 0.8865998909749021, "learning_rate": 1.7195220924898882e-06, "loss": 0.7584, "step": 246 }, { "epoch": 0.5825471698113207, "grad_norm": 0.8256395052056026, "learning_rate": 1.7167825131684511e-06, "loss": 0.7905, "step": 247 }, { "epoch": 0.5849056603773585, "grad_norm": 0.8134932247816357, "learning_rate": 1.7140318246977163e-06, "loss": 0.7932, "step": 248 }, { "epoch": 0.5872641509433962, "grad_norm": 0.8289362753913454, "learning_rate": 1.7112700697095953e-06, "loss": 0.7727, "step": 249 }, { "epoch": 0.589622641509434, "grad_norm": 0.7919440339387878, "learning_rate": 1.7084972910075154e-06, "loss": 0.819, "step": 250 }, { "epoch": 0.5919811320754716, "grad_norm": 0.7892796730141778, "learning_rate": 1.7057135315657567e-06, "loss": 0.7994, "step": 251 }, { "epoch": 0.5943396226415094, "grad_norm": 0.818593844058535, "learning_rate": 1.7029188345287865e-06, "loss": 0.7833, "step": 252 }, { "epoch": 0.5966981132075472, "grad_norm": 0.8550235074282425, "learning_rate": 1.7001132432105894e-06, "loss": 0.7627, "step": 253 }, { "epoch": 0.5990566037735849, "grad_norm": 0.7825131867584777, "learning_rate": 1.6972968010939952e-06, "loss": 0.7863, "step": 254 }, { "epoch": 0.6014150943396226, "grad_norm": 0.7872637454110324, "learning_rate": 1.6944695518300084e-06, "loss": 0.8202, "step": 255 }, { "epoch": 0.6037735849056604, "grad_norm": 0.8264865629926671, "learning_rate": 1.6916315392371283e-06, "loss": 0.7794, "step": 256 }, { "epoch": 0.6061320754716981, "grad_norm": 0.8112171041686048, "learning_rate": 1.688782807300671e-06, "loss": 0.7975, "step": 257 }, { "epoch": 0.6084905660377359, "grad_norm": 0.8017599214256371, "learning_rate": 1.685923400172088e-06, "loss": 0.7821, "step": 258 }, { "epoch": 0.6108490566037735, "grad_norm": 0.7855915924335556, "learning_rate": 1.683053362168282e-06, "loss": 0.7834, "step": 259 }, { "epoch": 0.6132075471698113, "grad_norm": 0.8227864852025443, "learning_rate": 1.6801727377709191e-06, "loss": 0.7715, "step": 260 }, { "epoch": 0.6155660377358491, "grad_norm": 0.8347676603852344, "learning_rate": 1.6772815716257411e-06, "loss": 0.7783, "step": 261 }, { "epoch": 0.6179245283018868, "grad_norm": 0.847840576553053, "learning_rate": 1.6743799085418709e-06, "loss": 0.7796, "step": 262 }, { "epoch": 0.6202830188679245, "grad_norm": 0.8187554443504805, "learning_rate": 1.671467793491121e-06, "loss": 0.8223, "step": 263 }, { "epoch": 0.6226415094339622, "grad_norm": 0.8560599665847224, "learning_rate": 1.6685452716072942e-06, "loss": 0.7769, "step": 264 }, { "epoch": 0.625, "grad_norm": 0.8756516850796994, "learning_rate": 1.6656123881854858e-06, "loss": 0.782, "step": 265 }, { "epoch": 0.6273584905660378, "grad_norm": 0.861305723586375, "learning_rate": 1.6626691886813802e-06, "loss": 0.7907, "step": 266 }, { "epoch": 0.6297169811320755, "grad_norm": 0.8565858633218812, "learning_rate": 1.6597157187105474e-06, "loss": 0.7586, "step": 267 }, { "epoch": 0.6320754716981132, "grad_norm": 0.7796005038585514, "learning_rate": 1.6567520240477343e-06, "loss": 0.7996, "step": 268 }, { "epoch": 0.6344339622641509, "grad_norm": 0.9853760454692906, "learning_rate": 1.6537781506261586e-06, "loss": 0.803, "step": 269 }, { "epoch": 0.6367924528301887, "grad_norm": 0.8214336899114236, "learning_rate": 1.6507941445367934e-06, "loss": 0.7484, "step": 270 }, { "epoch": 0.6391509433962265, "grad_norm": 0.8871384572382998, "learning_rate": 1.647800052027655e-06, "loss": 0.7876, "step": 271 }, { "epoch": 0.6415094339622641, "grad_norm": 0.8564572450153076, "learning_rate": 1.6447959195030849e-06, "loss": 0.8054, "step": 272 }, { "epoch": 0.6438679245283019, "grad_norm": 0.8408989291477506, "learning_rate": 1.6417817935230316e-06, "loss": 0.7668, "step": 273 }, { "epoch": 0.6462264150943396, "grad_norm": 0.7945270947755096, "learning_rate": 1.6387577208023279e-06, "loss": 0.7893, "step": 274 }, { "epoch": 0.6485849056603774, "grad_norm": 0.8844404937902336, "learning_rate": 1.6357237482099683e-06, "loss": 0.794, "step": 275 }, { "epoch": 0.6509433962264151, "grad_norm": 0.8518070202862351, "learning_rate": 1.6326799227683803e-06, "loss": 0.7967, "step": 276 }, { "epoch": 0.6533018867924528, "grad_norm": 0.819470621959587, "learning_rate": 1.6296262916526995e-06, "loss": 0.7633, "step": 277 }, { "epoch": 0.6556603773584906, "grad_norm": 0.7989369195204531, "learning_rate": 1.626562902190034e-06, "loss": 0.807, "step": 278 }, { "epoch": 0.6580188679245284, "grad_norm": 0.9412075176216419, "learning_rate": 1.6234898018587336e-06, "loss": 0.7752, "step": 279 }, { "epoch": 0.660377358490566, "grad_norm": 0.7895838126966979, "learning_rate": 1.6204070382876531e-06, "loss": 0.7786, "step": 280 }, { "epoch": 0.6627358490566038, "grad_norm": 0.7981016976096619, "learning_rate": 1.6173146592554152e-06, "loss": 0.7958, "step": 281 }, { "epoch": 0.6650943396226415, "grad_norm": 0.8907323576889667, "learning_rate": 1.6142127126896679e-06, "loss": 0.807, "step": 282 }, { "epoch": 0.6674528301886793, "grad_norm": 0.8053567442309949, "learning_rate": 1.6111012466663434e-06, "loss": 0.788, "step": 283 }, { "epoch": 0.6698113207547169, "grad_norm": 0.8378739800286575, "learning_rate": 1.6079803094089125e-06, "loss": 0.736, "step": 284 }, { "epoch": 0.6721698113207547, "grad_norm": 0.8858100985238717, "learning_rate": 1.6048499492876375e-06, "loss": 0.8106, "step": 285 }, { "epoch": 0.6745283018867925, "grad_norm": 0.8276972316955101, "learning_rate": 1.6017102148188215e-06, "loss": 0.7835, "step": 286 }, { "epoch": 0.6768867924528302, "grad_norm": 0.8853131510592382, "learning_rate": 1.598561154664058e-06, "loss": 0.7965, "step": 287 }, { "epoch": 0.6792452830188679, "grad_norm": 0.8786321199638693, "learning_rate": 1.5954028176294746e-06, "loss": 0.8268, "step": 288 }, { "epoch": 0.6816037735849056, "grad_norm": 0.8669272571254311, "learning_rate": 1.5922352526649801e-06, "loss": 0.7823, "step": 289 }, { "epoch": 0.6839622641509434, "grad_norm": 0.9644127048111335, "learning_rate": 1.589058508863501e-06, "loss": 0.78, "step": 290 }, { "epoch": 0.6863207547169812, "grad_norm": 0.8073310279373501, "learning_rate": 1.5858726354602248e-06, "loss": 0.7765, "step": 291 }, { "epoch": 0.6886792452830188, "grad_norm": 0.8775721979922628, "learning_rate": 1.5826776818318348e-06, "loss": 0.7597, "step": 292 }, { "epoch": 0.6910377358490566, "grad_norm": 1.015547726125361, "learning_rate": 1.5794736974957464e-06, "loss": 0.786, "step": 293 }, { "epoch": 0.6933962264150944, "grad_norm": 0.7708601481727361, "learning_rate": 1.5762607321093366e-06, "loss": 0.7817, "step": 294 }, { "epoch": 0.6957547169811321, "grad_norm": 0.8426421786890617, "learning_rate": 1.5730388354691785e-06, "loss": 0.7847, "step": 295 }, { "epoch": 0.6981132075471698, "grad_norm": 0.7855014020345946, "learning_rate": 1.569808057510266e-06, "loss": 0.7748, "step": 296 }, { "epoch": 0.7004716981132075, "grad_norm": 0.8389890263875862, "learning_rate": 1.5665684483052424e-06, "loss": 0.7856, "step": 297 }, { "epoch": 0.7028301886792453, "grad_norm": 0.8482309028462507, "learning_rate": 1.563320058063622e-06, "loss": 0.7469, "step": 298 }, { "epoch": 0.7051886792452831, "grad_norm": 1.0888639918368348, "learning_rate": 1.5600629371310144e-06, "loss": 0.7905, "step": 299 }, { "epoch": 0.7075471698113207, "grad_norm": 0.8011947485403833, "learning_rate": 1.556797135988342e-06, "loss": 0.7902, "step": 300 }, { "epoch": 0.7099056603773585, "grad_norm": 0.7584432625187374, "learning_rate": 1.5535227052510588e-06, "loss": 0.7935, "step": 301 }, { "epoch": 0.7122641509433962, "grad_norm": 1.1353449141437797, "learning_rate": 1.5502396956683667e-06, "loss": 0.7992, "step": 302 }, { "epoch": 0.714622641509434, "grad_norm": 0.8144382945981636, "learning_rate": 1.5469481581224271e-06, "loss": 0.761, "step": 303 }, { "epoch": 0.7169811320754716, "grad_norm": 0.9448262210280223, "learning_rate": 1.5436481436275724e-06, "loss": 0.7839, "step": 304 }, { "epoch": 0.7193396226415094, "grad_norm": 0.8145609006654517, "learning_rate": 1.5403397033295179e-06, "loss": 0.7975, "step": 305 }, { "epoch": 0.7216981132075472, "grad_norm": 0.9825256520905712, "learning_rate": 1.537022888504566e-06, "loss": 0.7918, "step": 306 }, { "epoch": 0.7240566037735849, "grad_norm": 0.9647183394075436, "learning_rate": 1.5336977505588134e-06, "loss": 0.7973, "step": 307 }, { "epoch": 0.7264150943396226, "grad_norm": 0.802175509091533, "learning_rate": 1.530364341027354e-06, "loss": 0.7928, "step": 308 }, { "epoch": 0.7287735849056604, "grad_norm": 0.8122620856594104, "learning_rate": 1.5270227115734789e-06, "loss": 0.7767, "step": 309 }, { "epoch": 0.7311320754716981, "grad_norm": 0.8656607812881103, "learning_rate": 1.5236729139878778e-06, "loss": 0.7746, "step": 310 }, { "epoch": 0.7334905660377359, "grad_norm": 0.8721224552709389, "learning_rate": 1.5203150001878353e-06, "loss": 0.7929, "step": 311 }, { "epoch": 0.7358490566037735, "grad_norm": 0.7856050732773205, "learning_rate": 1.5169490222164254e-06, "loss": 0.7733, "step": 312 }, { "epoch": 0.7382075471698113, "grad_norm": 0.8639502364522528, "learning_rate": 1.5135750322417066e-06, "loss": 0.7958, "step": 313 }, { "epoch": 0.7405660377358491, "grad_norm": 0.9986722246195658, "learning_rate": 1.5101930825559124e-06, "loss": 0.7944, "step": 314 }, { "epoch": 0.7429245283018868, "grad_norm": 0.8861549532629037, "learning_rate": 1.50680322557464e-06, "loss": 0.7887, "step": 315 }, { "epoch": 0.7452830188679245, "grad_norm": 0.8526619540985027, "learning_rate": 1.5034055138360398e-06, "loss": 0.7388, "step": 316 }, { "epoch": 0.7476415094339622, "grad_norm": 0.8760084110715984, "learning_rate": 1.5e-06, "loss": 0.7719, "step": 317 }, { "epoch": 0.75, "grad_norm": 0.8038573705056793, "learning_rate": 1.4965867368473306e-06, "loss": 0.7514, "step": 318 }, { "epoch": 0.75, "eval_loss": 0.6982797384262085, "eval_runtime": 83.2645, "eval_samples_per_second": 8.467, "eval_steps_per_second": 0.54, "step": 318 }, { "epoch": 0.7523584905660378, "grad_norm": 0.8193168846563599, "learning_rate": 1.4931657772789457e-06, "loss": 0.7834, "step": 319 }, { "epoch": 0.7547169811320755, "grad_norm": 0.9092282687501353, "learning_rate": 1.4897371743150423e-06, "loss": 0.753, "step": 320 }, { "epoch": 0.7570754716981132, "grad_norm": 0.944599516853919, "learning_rate": 1.4863009810942813e-06, "loss": 0.7855, "step": 321 }, { "epoch": 0.7594339622641509, "grad_norm": 0.9097780581691608, "learning_rate": 1.4828572508729606e-06, "loss": 0.7837, "step": 322 }, { "epoch": 0.7617924528301887, "grad_norm": 0.8635246603088246, "learning_rate": 1.479406037024192e-06, "loss": 0.7862, "step": 323 }, { "epoch": 0.7641509433962265, "grad_norm": 0.8041740773336183, "learning_rate": 1.4759473930370736e-06, "loss": 0.7705, "step": 324 }, { "epoch": 0.7665094339622641, "grad_norm": 0.8563734416095982, "learning_rate": 1.4724813725158596e-06, "loss": 0.7944, "step": 325 }, { "epoch": 0.7688679245283019, "grad_norm": 0.7885016820225573, "learning_rate": 1.4690080291791312e-06, "loss": 0.78, "step": 326 }, { "epoch": 0.7712264150943396, "grad_norm": 0.9769999602785767, "learning_rate": 1.4655274168589633e-06, "loss": 0.7506, "step": 327 }, { "epoch": 0.7735849056603774, "grad_norm": 0.8241115466345136, "learning_rate": 1.462039589500089e-06, "loss": 0.7826, "step": 328 }, { "epoch": 0.7759433962264151, "grad_norm": 0.8261080001645498, "learning_rate": 1.4585446011590658e-06, "loss": 0.7592, "step": 329 }, { "epoch": 0.7783018867924528, "grad_norm": 1.1493611817217861, "learning_rate": 1.4550425060034365e-06, "loss": 0.7674, "step": 330 }, { "epoch": 0.7806603773584906, "grad_norm": 0.8343201494559767, "learning_rate": 1.4515333583108893e-06, "loss": 0.8072, "step": 331 }, { "epoch": 0.7830188679245284, "grad_norm": 1.0664747829909942, "learning_rate": 1.4480172124684172e-06, "loss": 0.7802, "step": 332 }, { "epoch": 0.785377358490566, "grad_norm": 0.7792110149695123, "learning_rate": 1.4444941229714758e-06, "loss": 0.7704, "step": 333 }, { "epoch": 0.7877358490566038, "grad_norm": 0.8718279852088779, "learning_rate": 1.440964144423136e-06, "loss": 0.7963, "step": 334 }, { "epoch": 0.7900943396226415, "grad_norm": 1.02769301319547, "learning_rate": 1.4374273315332415e-06, "loss": 0.7863, "step": 335 }, { "epoch": 0.7924528301886793, "grad_norm": 0.7843647199877115, "learning_rate": 1.433883739117558e-06, "loss": 0.7705, "step": 336 }, { "epoch": 0.7948113207547169, "grad_norm": 0.815818120801617, "learning_rate": 1.430333422096925e-06, "loss": 0.7651, "step": 337 }, { "epoch": 0.7971698113207547, "grad_norm": 0.7835815188776669, "learning_rate": 1.4267764354964037e-06, "loss": 0.789, "step": 338 }, { "epoch": 0.7995283018867925, "grad_norm": 0.8382956450308263, "learning_rate": 1.423212834444425e-06, "loss": 0.817, "step": 339 }, { "epoch": 0.8018867924528302, "grad_norm": 0.9606287069411898, "learning_rate": 1.4196426741719345e-06, "loss": 0.7801, "step": 340 }, { "epoch": 0.8042452830188679, "grad_norm": 0.7833416575809102, "learning_rate": 1.4160660100115373e-06, "loss": 0.8139, "step": 341 }, { "epoch": 0.8066037735849056, "grad_norm": 0.8112490810199738, "learning_rate": 1.4124828973966392e-06, "loss": 0.7951, "step": 342 }, { "epoch": 0.8089622641509434, "grad_norm": 0.8533925894384254, "learning_rate": 1.4088933918605887e-06, "loss": 0.7555, "step": 343 }, { "epoch": 0.8113207547169812, "grad_norm": 0.8917123790523681, "learning_rate": 1.405297549035816e-06, "loss": 0.7941, "step": 344 }, { "epoch": 0.8136792452830188, "grad_norm": 0.8690146736748873, "learning_rate": 1.4016954246529694e-06, "loss": 0.7966, "step": 345 }, { "epoch": 0.8160377358490566, "grad_norm": 0.8372555162768984, "learning_rate": 1.3980870745400544e-06, "loss": 0.7972, "step": 346 }, { "epoch": 0.8183962264150944, "grad_norm": 0.8356712411477044, "learning_rate": 1.3944725546215662e-06, "loss": 0.753, "step": 347 }, { "epoch": 0.8207547169811321, "grad_norm": 0.8366058600346219, "learning_rate": 1.3908519209176225e-06, "loss": 0.7813, "step": 348 }, { "epoch": 0.8231132075471698, "grad_norm": 0.8373524566281616, "learning_rate": 1.3872252295430986e-06, "loss": 0.7956, "step": 349 }, { "epoch": 0.8254716981132075, "grad_norm": 0.7833296789989939, "learning_rate": 1.3835925367067529e-06, "loss": 0.7449, "step": 350 }, { "epoch": 0.8278301886792453, "grad_norm": 0.7981265071211454, "learning_rate": 1.37995389871036e-06, "loss": 0.7756, "step": 351 }, { "epoch": 0.8301886792452831, "grad_norm": 0.8840540972881347, "learning_rate": 1.3763093719478357e-06, "loss": 0.7984, "step": 352 }, { "epoch": 0.8325471698113207, "grad_norm": 0.985313340190558, "learning_rate": 1.372659012904363e-06, "loss": 0.7821, "step": 353 }, { "epoch": 0.8349056603773585, "grad_norm": 0.8138177618630202, "learning_rate": 1.369002878155519e-06, "loss": 0.7618, "step": 354 }, { "epoch": 0.8372641509433962, "grad_norm": 0.7948448582342686, "learning_rate": 1.3653410243663951e-06, "loss": 0.7559, "step": 355 }, { "epoch": 0.839622641509434, "grad_norm": 0.8151246949343306, "learning_rate": 1.3616735082907196e-06, "loss": 0.7722, "step": 356 }, { "epoch": 0.8419811320754716, "grad_norm": 0.8162887157403834, "learning_rate": 1.35800038676998e-06, "loss": 0.7711, "step": 357 }, { "epoch": 0.8443396226415094, "grad_norm": 1.1322690706169278, "learning_rate": 1.3543217167325388e-06, "loss": 0.762, "step": 358 }, { "epoch": 0.8466981132075472, "grad_norm": 0.8102205098745584, "learning_rate": 1.3506375551927544e-06, "loss": 0.7532, "step": 359 }, { "epoch": 0.8490566037735849, "grad_norm": 0.9369994865443904, "learning_rate": 1.3469479592500951e-06, "loss": 0.737, "step": 360 }, { "epoch": 0.8514150943396226, "grad_norm": 0.9319609634032745, "learning_rate": 1.3432529860882556e-06, "loss": 0.8074, "step": 361 }, { "epoch": 0.8537735849056604, "grad_norm": 0.844911038169109, "learning_rate": 1.3395526929742691e-06, "loss": 0.7576, "step": 362 }, { "epoch": 0.8561320754716981, "grad_norm": 0.8213852435481517, "learning_rate": 1.3358471372576227e-06, "loss": 0.8091, "step": 363 }, { "epoch": 0.8584905660377359, "grad_norm": 0.8679917359639049, "learning_rate": 1.3321363763693643e-06, "loss": 0.7649, "step": 364 }, { "epoch": 0.8608490566037735, "grad_norm": 0.863728546344944, "learning_rate": 1.3284204678212167e-06, "loss": 0.7364, "step": 365 }, { "epoch": 0.8632075471698113, "grad_norm": 0.9151925878031607, "learning_rate": 1.3246994692046835e-06, "loss": 0.7617, "step": 366 }, { "epoch": 0.8655660377358491, "grad_norm": 0.7676651686086633, "learning_rate": 1.3209734381901578e-06, "loss": 0.7703, "step": 367 }, { "epoch": 0.8679245283018868, "grad_norm": 0.79154626898034, "learning_rate": 1.3172424325260272e-06, "loss": 0.7816, "step": 368 }, { "epoch": 0.8702830188679245, "grad_norm": 0.8831690649426513, "learning_rate": 1.3135065100377814e-06, "loss": 0.7646, "step": 369 }, { "epoch": 0.8726415094339622, "grad_norm": 0.8098189405568664, "learning_rate": 1.3097657286271116e-06, "loss": 0.7673, "step": 370 }, { "epoch": 0.875, "grad_norm": 0.8071739591609438, "learning_rate": 1.3060201462710176e-06, "loss": 0.7438, "step": 371 }, { "epoch": 0.8773584905660378, "grad_norm": 0.8864737282217798, "learning_rate": 1.3022698210209066e-06, "loss": 0.7393, "step": 372 }, { "epoch": 0.8797169811320755, "grad_norm": 0.8193555244410853, "learning_rate": 1.2985148110016947e-06, "loss": 0.7658, "step": 373 }, { "epoch": 0.8820754716981132, "grad_norm": 0.7819668232955883, "learning_rate": 1.2947551744109043e-06, "loss": 0.7876, "step": 374 }, { "epoch": 0.8844339622641509, "grad_norm": 0.9661764152661253, "learning_rate": 1.2909909695177645e-06, "loss": 0.7556, "step": 375 }, { "epoch": 0.8867924528301887, "grad_norm": 0.7916178501632968, "learning_rate": 1.2872222546623065e-06, "loss": 0.7776, "step": 376 }, { "epoch": 0.8891509433962265, "grad_norm": 0.908391562439544, "learning_rate": 1.2834490882544598e-06, "loss": 0.7604, "step": 377 }, { "epoch": 0.8915094339622641, "grad_norm": 0.7673617521747667, "learning_rate": 1.2796715287731461e-06, "loss": 0.7786, "step": 378 }, { "epoch": 0.8938679245283019, "grad_norm": 1.1737766347467777, "learning_rate": 1.2758896347653752e-06, "loss": 0.7692, "step": 379 }, { "epoch": 0.8962264150943396, "grad_norm": 0.7823714073715833, "learning_rate": 1.272103464845335e-06, "loss": 0.7389, "step": 380 }, { "epoch": 0.8985849056603774, "grad_norm": 1.0030453246218272, "learning_rate": 1.2683130776934848e-06, "loss": 0.7556, "step": 381 }, { "epoch": 0.9009433962264151, "grad_norm": 0.9688917314182516, "learning_rate": 1.2645185320556444e-06, "loss": 0.7415, "step": 382 }, { "epoch": 0.9033018867924528, "grad_norm": 0.8539895113482738, "learning_rate": 1.2607198867420858e-06, "loss": 0.7772, "step": 383 }, { "epoch": 0.9056603773584906, "grad_norm": 0.8011530846319137, "learning_rate": 1.256917200626619e-06, "loss": 0.7426, "step": 384 }, { "epoch": 0.9080188679245284, "grad_norm": 0.9859270894449156, "learning_rate": 1.253110532645682e-06, "loss": 0.7942, "step": 385 }, { "epoch": 0.910377358490566, "grad_norm": 0.7970067775774904, "learning_rate": 1.2492999417974253e-06, "loss": 0.7914, "step": 386 }, { "epoch": 0.9127358490566038, "grad_norm": 0.9049836680757578, "learning_rate": 1.245485487140799e-06, "loss": 0.7787, "step": 387 }, { "epoch": 0.9150943396226415, "grad_norm": 0.866322963165458, "learning_rate": 1.2416672277946373e-06, "loss": 0.781, "step": 388 }, { "epoch": 0.9174528301886793, "grad_norm": 0.9269306886834423, "learning_rate": 1.2378452229367413e-06, "loss": 0.808, "step": 389 }, { "epoch": 0.9198113207547169, "grad_norm": 0.8286820002413788, "learning_rate": 1.2340195318029622e-06, "loss": 0.7623, "step": 390 }, { "epoch": 0.9221698113207547, "grad_norm": 0.8143258703482895, "learning_rate": 1.2301902136862848e-06, "loss": 0.7965, "step": 391 }, { "epoch": 0.9245283018867925, "grad_norm": 0.8041175896213196, "learning_rate": 1.2263573279359053e-06, "loss": 0.7613, "step": 392 }, { "epoch": 0.9268867924528302, "grad_norm": 0.9087369855701409, "learning_rate": 1.2225209339563143e-06, "loss": 0.7669, "step": 393 }, { "epoch": 0.9292452830188679, "grad_norm": 0.8126237661274147, "learning_rate": 1.2186810912063758e-06, "loss": 0.7531, "step": 394 }, { "epoch": 0.9316037735849056, "grad_norm": 0.8354891878817764, "learning_rate": 1.214837859198404e-06, "loss": 0.768, "step": 395 }, { "epoch": 0.9339622641509434, "grad_norm": 0.7793219061787805, "learning_rate": 1.2109912974972422e-06, "loss": 0.7831, "step": 396 }, { "epoch": 0.9363207547169812, "grad_norm": 0.7813175142444524, "learning_rate": 1.2071414657193396e-06, "loss": 0.7671, "step": 397 }, { "epoch": 0.9386792452830188, "grad_norm": 0.8960835581240535, "learning_rate": 1.2032884235318268e-06, "loss": 0.775, "step": 398 }, { "epoch": 0.9410377358490566, "grad_norm": 0.8773817072709057, "learning_rate": 1.1994322306515925e-06, "loss": 0.7915, "step": 399 }, { "epoch": 0.9433962264150944, "grad_norm": 0.8191284697051531, "learning_rate": 1.1955729468443546e-06, "loss": 0.7548, "step": 400 }, { "epoch": 0.9457547169811321, "grad_norm": 0.8704070228218153, "learning_rate": 1.1917106319237384e-06, "loss": 0.773, "step": 401 }, { "epoch": 0.9481132075471698, "grad_norm": 0.8119790656590767, "learning_rate": 1.1878453457503464e-06, "loss": 0.7933, "step": 402 }, { "epoch": 0.9504716981132075, "grad_norm": 0.8290570173312977, "learning_rate": 1.1839771482308308e-06, "loss": 0.7515, "step": 403 }, { "epoch": 0.9528301886792453, "grad_norm": 0.7808059259378133, "learning_rate": 1.1801060993169666e-06, "loss": 0.771, "step": 404 }, { "epoch": 0.9551886792452831, "grad_norm": 0.8554505156061943, "learning_rate": 1.1762322590047219e-06, "loss": 0.767, "step": 405 }, { "epoch": 0.9575471698113207, "grad_norm": 0.8132652328205165, "learning_rate": 1.172355687333326e-06, "loss": 0.7467, "step": 406 }, { "epoch": 0.9599056603773585, "grad_norm": 0.7767004491954446, "learning_rate": 1.168476444384342e-06, "loss": 0.7667, "step": 407 }, { "epoch": 0.9622641509433962, "grad_norm": 0.8619052365176207, "learning_rate": 1.164594590280734e-06, "loss": 0.7993, "step": 408 }, { "epoch": 0.964622641509434, "grad_norm": 0.8115144668833016, "learning_rate": 1.1607101851859345e-06, "loss": 0.7787, "step": 409 }, { "epoch": 0.9669811320754716, "grad_norm": 0.8000429071733963, "learning_rate": 1.156823289302914e-06, "loss": 0.7767, "step": 410 }, { "epoch": 0.9693396226415094, "grad_norm": 0.8228751842965996, "learning_rate": 1.152933962873246e-06, "loss": 0.7829, "step": 411 }, { "epoch": 0.9716981132075472, "grad_norm": 0.8685312714228164, "learning_rate": 1.1490422661761743e-06, "loss": 0.7775, "step": 412 }, { "epoch": 0.9740566037735849, "grad_norm": 0.9383805515060808, "learning_rate": 1.1451482595276796e-06, "loss": 0.7716, "step": 413 }, { "epoch": 0.9764150943396226, "grad_norm": 0.978907741116596, "learning_rate": 1.1412520032795419e-06, "loss": 0.79, "step": 414 }, { "epoch": 0.9787735849056604, "grad_norm": 0.936909408413621, "learning_rate": 1.1373535578184082e-06, "loss": 0.7463, "step": 415 }, { "epoch": 0.9811320754716981, "grad_norm": 0.7779360289097986, "learning_rate": 1.1334529835648552e-06, "loss": 0.7544, "step": 416 }, { "epoch": 0.9834905660377359, "grad_norm": 0.8442404415567627, "learning_rate": 1.1295503409724525e-06, "loss": 0.7968, "step": 417 }, { "epoch": 0.9858490566037735, "grad_norm": 0.8856697438669298, "learning_rate": 1.1256456905268263e-06, "loss": 0.7622, "step": 418 }, { "epoch": 0.9882075471698113, "grad_norm": 0.8362007813195715, "learning_rate": 1.1217390927447225e-06, "loss": 0.7475, "step": 419 }, { "epoch": 0.9905660377358491, "grad_norm": 0.7824485352158383, "learning_rate": 1.1178306081730664e-06, "loss": 0.7705, "step": 420 }, { "epoch": 0.9929245283018868, "grad_norm": 0.7992205810856007, "learning_rate": 1.113920297388028e-06, "loss": 0.7545, "step": 421 }, { "epoch": 0.9952830188679245, "grad_norm": 0.785520038953663, "learning_rate": 1.1100082209940793e-06, "loss": 0.7647, "step": 422 }, { "epoch": 0.9976415094339622, "grad_norm": 0.8327122327201606, "learning_rate": 1.106094439623058e-06, "loss": 0.7632, "step": 423 }, { "epoch": 1.0, "grad_norm": 0.7736149741657676, "learning_rate": 1.1021790139332264e-06, "loss": 0.7861, "step": 424 }, { "epoch": 1.0, "eval_loss": 0.6892534494400024, "eval_runtime": 81.9429, "eval_samples_per_second": 8.604, "eval_steps_per_second": 0.549, "step": 424 }, { "epoch": 1.0023584905660377, "grad_norm": 0.8035685998252425, "learning_rate": 1.0982620046083305e-06, "loss": 0.7506, "step": 425 }, { "epoch": 1.0047169811320755, "grad_norm": 1.001273686570143, "learning_rate": 1.0943434723566623e-06, "loss": 0.783, "step": 426 }, { "epoch": 1.0070754716981132, "grad_norm": 0.7620350953234064, "learning_rate": 1.0904234779101158e-06, "loss": 0.7436, "step": 427 }, { "epoch": 1.009433962264151, "grad_norm": 0.8051824532909724, "learning_rate": 1.0865020820232466e-06, "loss": 0.7408, "step": 428 }, { "epoch": 1.0117924528301887, "grad_norm": 1.1441709749307019, "learning_rate": 1.0825793454723324e-06, "loss": 0.7619, "step": 429 }, { "epoch": 1.0141509433962264, "grad_norm": 0.934287576351282, "learning_rate": 1.0786553290544278e-06, "loss": 0.7291, "step": 430 }, { "epoch": 1.0165094339622642, "grad_norm": 0.867146390883491, "learning_rate": 1.0747300935864243e-06, "loss": 0.7697, "step": 431 }, { "epoch": 1.0188679245283019, "grad_norm": 0.7934743429540405, "learning_rate": 1.070803699904107e-06, "loss": 0.7774, "step": 432 }, { "epoch": 1.0212264150943395, "grad_norm": 1.204890281675565, "learning_rate": 1.0668762088612114e-06, "loss": 0.7659, "step": 433 }, { "epoch": 1.0235849056603774, "grad_norm": 0.7748660244184836, "learning_rate": 1.0629476813284807e-06, "loss": 0.7812, "step": 434 }, { "epoch": 1.025943396226415, "grad_norm": 0.8559552505505363, "learning_rate": 1.0590181781927227e-06, "loss": 0.7641, "step": 435 }, { "epoch": 1.028301886792453, "grad_norm": 0.8278708288686452, "learning_rate": 1.0550877603558654e-06, "loss": 0.7339, "step": 436 }, { "epoch": 1.0306603773584906, "grad_norm": 0.8199114258570241, "learning_rate": 1.0511564887340135e-06, "loss": 0.7884, "step": 437 }, { "epoch": 1.0330188679245282, "grad_norm": 0.8220578039981541, "learning_rate": 1.0472244242565034e-06, "loss": 0.7636, "step": 438 }, { "epoch": 1.0353773584905661, "grad_norm": 0.9899659514811034, "learning_rate": 1.043291627864961e-06, "loss": 0.7507, "step": 439 }, { "epoch": 1.0377358490566038, "grad_norm": 0.9164260231900845, "learning_rate": 1.0393581605123552e-06, "loss": 0.7604, "step": 440 }, { "epoch": 1.0400943396226414, "grad_norm": 0.780514004370331, "learning_rate": 1.035424083162054e-06, "loss": 0.7661, "step": 441 }, { "epoch": 1.0424528301886793, "grad_norm": 0.9654942156178127, "learning_rate": 1.031489456786879e-06, "loss": 0.732, "step": 442 }, { "epoch": 1.044811320754717, "grad_norm": 0.9028629851443349, "learning_rate": 1.0275543423681621e-06, "loss": 0.7704, "step": 443 }, { "epoch": 1.0471698113207548, "grad_norm": 0.8997990850767794, "learning_rate": 1.0236188008947978e-06, "loss": 0.7832, "step": 444 }, { "epoch": 1.0495283018867925, "grad_norm": 0.8380399486601815, "learning_rate": 1.0196828933623008e-06, "loss": 0.7513, "step": 445 }, { "epoch": 1.0518867924528301, "grad_norm": 0.818805158462816, "learning_rate": 1.0157466807718577e-06, "loss": 0.7595, "step": 446 }, { "epoch": 1.054245283018868, "grad_norm": 0.8325448461017404, "learning_rate": 1.0118102241293847e-06, "loss": 0.7634, "step": 447 }, { "epoch": 1.0566037735849056, "grad_norm": 0.7936938025099423, "learning_rate": 1.0078735844445788e-06, "loss": 0.768, "step": 448 }, { "epoch": 1.0589622641509433, "grad_norm": 0.8378311364099181, "learning_rate": 1.0039368227299753e-06, "loss": 0.7599, "step": 449 }, { "epoch": 1.0613207547169812, "grad_norm": 0.8299911067496856, "learning_rate": 1e-06, "loss": 0.7488, "step": 450 }, { "epoch": 1.0636792452830188, "grad_norm": 0.8677844259185759, "learning_rate": 9.960631772700248e-07, "loss": 0.7698, "step": 451 }, { "epoch": 1.0660377358490567, "grad_norm": 0.8463491375440975, "learning_rate": 9.92126415555421e-07, "loss": 0.7493, "step": 452 }, { "epoch": 1.0683962264150944, "grad_norm": 0.8004441596711467, "learning_rate": 9.881897758706154e-07, "loss": 0.7366, "step": 453 }, { "epoch": 1.070754716981132, "grad_norm": 1.2321632539336993, "learning_rate": 9.842533192281422e-07, "loss": 0.7481, "step": 454 }, { "epoch": 1.0731132075471699, "grad_norm": 0.8194799035445371, "learning_rate": 9.803171066376995e-07, "loss": 0.7834, "step": 455 }, { "epoch": 1.0754716981132075, "grad_norm": 0.7953471718099565, "learning_rate": 9.763811991052019e-07, "loss": 0.7508, "step": 456 }, { "epoch": 1.0778301886792452, "grad_norm": 0.8500817934393053, "learning_rate": 9.72445657631838e-07, "loss": 0.7639, "step": 457 }, { "epoch": 1.080188679245283, "grad_norm": 0.8053711636082244, "learning_rate": 9.68510543213121e-07, "loss": 0.7546, "step": 458 }, { "epoch": 1.0825471698113207, "grad_norm": 1.1272698654628237, "learning_rate": 9.645759168379461e-07, "loss": 0.735, "step": 459 }, { "epoch": 1.0849056603773586, "grad_norm": 0.7772007598787661, "learning_rate": 9.606418394876447e-07, "loss": 0.747, "step": 460 }, { "epoch": 1.0872641509433962, "grad_norm": 0.8414360815423256, "learning_rate": 9.567083721350388e-07, "loss": 0.7411, "step": 461 }, { "epoch": 1.0896226415094339, "grad_norm": 0.7791267607388218, "learning_rate": 9.527755757434966e-07, "loss": 0.7436, "step": 462 }, { "epoch": 1.0919811320754718, "grad_norm": 0.80528466203603, "learning_rate": 9.48843511265987e-07, "loss": 0.7656, "step": 463 }, { "epoch": 1.0943396226415094, "grad_norm": 0.7758351013188943, "learning_rate": 9.449122396441343e-07, "loss": 0.737, "step": 464 }, { "epoch": 1.0966981132075473, "grad_norm": 1.2494162832647924, "learning_rate": 9.409818218072772e-07, "loss": 0.7613, "step": 465 }, { "epoch": 1.099056603773585, "grad_norm": 0.8049303948390758, "learning_rate": 9.370523186715194e-07, "loss": 0.7628, "step": 466 }, { "epoch": 1.1014150943396226, "grad_norm": 0.7896423211268341, "learning_rate": 9.331237911387888e-07, "loss": 0.7301, "step": 467 }, { "epoch": 1.1037735849056605, "grad_norm": 0.7947232451183359, "learning_rate": 9.291963000958931e-07, "loss": 0.7803, "step": 468 }, { "epoch": 1.1061320754716981, "grad_norm": 0.8128944094386984, "learning_rate": 9.252699064135758e-07, "loss": 0.7693, "step": 469 }, { "epoch": 1.1084905660377358, "grad_norm": 0.9181658602259837, "learning_rate": 9.213446709455722e-07, "loss": 0.7794, "step": 470 }, { "epoch": 1.1108490566037736, "grad_norm": 0.753282034988688, "learning_rate": 9.174206545276677e-07, "loss": 0.7383, "step": 471 }, { "epoch": 1.1132075471698113, "grad_norm": 0.8890774956158611, "learning_rate": 9.134979179767532e-07, "loss": 0.7528, "step": 472 }, { "epoch": 1.115566037735849, "grad_norm": 0.9503365963029163, "learning_rate": 9.095765220898843e-07, "loss": 0.7757, "step": 473 }, { "epoch": 1.1179245283018868, "grad_norm": 0.9284151146362815, "learning_rate": 9.056565276433377e-07, "loss": 0.7442, "step": 474 }, { "epoch": 1.1202830188679245, "grad_norm": 0.9704301526052668, "learning_rate": 9.017379953916695e-07, "loss": 0.7417, "step": 475 }, { "epoch": 1.1226415094339623, "grad_norm": 0.8572079632472874, "learning_rate": 8.978209860667738e-07, "loss": 0.7724, "step": 476 }, { "epoch": 1.125, "grad_norm": 0.7912726011032718, "learning_rate": 8.93905560376942e-07, "loss": 0.7741, "step": 477 }, { "epoch": 1.1273584905660377, "grad_norm": 0.7762605349979439, "learning_rate": 8.899917790059207e-07, "loss": 0.7211, "step": 478 }, { "epoch": 1.1297169811320755, "grad_norm": 0.8625085844504403, "learning_rate": 8.860797026119721e-07, "loss": 0.7327, "step": 479 }, { "epoch": 1.1320754716981132, "grad_norm": 0.8012017130865744, "learning_rate": 8.821693918269333e-07, "loss": 0.7293, "step": 480 }, { "epoch": 1.134433962264151, "grad_norm": 0.7832218701492647, "learning_rate": 8.782609072552776e-07, "loss": 0.7584, "step": 481 }, { "epoch": 1.1367924528301887, "grad_norm": 0.8223365605395953, "learning_rate": 8.743543094731737e-07, "loss": 0.757, "step": 482 }, { "epoch": 1.1391509433962264, "grad_norm": 0.769269936037437, "learning_rate": 8.704496590275477e-07, "loss": 0.7265, "step": 483 }, { "epoch": 1.1415094339622642, "grad_norm": 0.8470727703331676, "learning_rate": 8.665470164351447e-07, "loss": 0.7446, "step": 484 }, { "epoch": 1.1438679245283019, "grad_norm": 0.8609116334367811, "learning_rate": 8.626464421815918e-07, "loss": 0.7357, "step": 485 }, { "epoch": 1.1462264150943395, "grad_norm": 0.7644474626055862, "learning_rate": 8.587479967204582e-07, "loss": 0.7502, "step": 486 }, { "epoch": 1.1485849056603774, "grad_norm": 0.7974217428241436, "learning_rate": 8.548517404723206e-07, "loss": 0.7255, "step": 487 }, { "epoch": 1.150943396226415, "grad_norm": 0.8124702999828396, "learning_rate": 8.509577338238254e-07, "loss": 0.766, "step": 488 }, { "epoch": 1.1533018867924527, "grad_norm": 0.8292245412487236, "learning_rate": 8.47066037126754e-07, "loss": 0.7302, "step": 489 }, { "epoch": 1.1556603773584906, "grad_norm": 0.819810501024351, "learning_rate": 8.43176710697086e-07, "loss": 0.7163, "step": 490 }, { "epoch": 1.1580188679245282, "grad_norm": 0.8385443524556274, "learning_rate": 8.392898148140656e-07, "loss": 0.7465, "step": 491 }, { "epoch": 1.1603773584905661, "grad_norm": 0.778726751410171, "learning_rate": 8.354054097192659e-07, "loss": 0.7758, "step": 492 }, { "epoch": 1.1627358490566038, "grad_norm": 0.788415584362694, "learning_rate": 8.315235556156579e-07, "loss": 0.7498, "step": 493 }, { "epoch": 1.1650943396226414, "grad_norm": 0.7703907755533744, "learning_rate": 8.276443126666742e-07, "loss": 0.7276, "step": 494 }, { "epoch": 1.1674528301886793, "grad_norm": 0.8704987882807484, "learning_rate": 8.237677409952784e-07, "loss": 0.766, "step": 495 }, { "epoch": 1.169811320754717, "grad_norm": 0.7948667823022109, "learning_rate": 8.198939006830333e-07, "loss": 0.7656, "step": 496 }, { "epoch": 1.1721698113207548, "grad_norm": 0.8090640349585008, "learning_rate": 8.160228517691692e-07, "loss": 0.7257, "step": 497 }, { "epoch": 1.1745283018867925, "grad_norm": 1.159992161506865, "learning_rate": 8.121546542496538e-07, "loss": 0.7593, "step": 498 }, { "epoch": 1.1768867924528301, "grad_norm": 0.8230774285047441, "learning_rate": 8.082893680762618e-07, "loss": 0.7511, "step": 499 }, { "epoch": 1.179245283018868, "grad_norm": 0.8682718782355334, "learning_rate": 8.044270531556452e-07, "loss": 0.7647, "step": 500 }, { "epoch": 1.1816037735849056, "grad_norm": 0.8274347964467509, "learning_rate": 8.005677693484076e-07, "loss": 0.7536, "step": 501 }, { "epoch": 1.1839622641509433, "grad_norm": 0.8334334904558071, "learning_rate": 7.967115764681731e-07, "loss": 0.782, "step": 502 }, { "epoch": 1.1863207547169812, "grad_norm": 0.8985376970213992, "learning_rate": 7.928585342806607e-07, "loss": 0.7461, "step": 503 }, { "epoch": 1.1886792452830188, "grad_norm": 0.7846783242941104, "learning_rate": 7.890087025027579e-07, "loss": 0.7461, "step": 504 }, { "epoch": 1.1910377358490567, "grad_norm": 0.800897718044774, "learning_rate": 7.85162140801596e-07, "loss": 0.7525, "step": 505 }, { "epoch": 1.1933962264150944, "grad_norm": 0.7742939894445084, "learning_rate": 7.813189087936242e-07, "loss": 0.7082, "step": 506 }, { "epoch": 1.195754716981132, "grad_norm": 0.8945053331857878, "learning_rate": 7.774790660436857e-07, "loss": 0.7328, "step": 507 }, { "epoch": 1.1981132075471699, "grad_norm": 0.9098394537893619, "learning_rate": 7.736426720640948e-07, "loss": 0.7802, "step": 508 }, { "epoch": 1.2004716981132075, "grad_norm": 0.7969225975663269, "learning_rate": 7.698097863137152e-07, "loss": 0.7423, "step": 509 }, { "epoch": 1.2028301886792452, "grad_norm": 0.8867765640390939, "learning_rate": 7.659804681970377e-07, "loss": 0.7443, "step": 510 }, { "epoch": 1.205188679245283, "grad_norm": 0.8936277250989298, "learning_rate": 7.621547770632589e-07, "loss": 0.7446, "step": 511 }, { "epoch": 1.2075471698113207, "grad_norm": 0.7650448661091079, "learning_rate": 7.583327722053626e-07, "loss": 0.7437, "step": 512 }, { "epoch": 1.2099056603773586, "grad_norm": 0.7718047147502091, "learning_rate": 7.545145128592008e-07, "loss": 0.7507, "step": 513 }, { "epoch": 1.2122641509433962, "grad_norm": 0.9572262834139598, "learning_rate": 7.507000582025748e-07, "loss": 0.7329, "step": 514 }, { "epoch": 1.2146226415094339, "grad_norm": 0.8375942206336273, "learning_rate": 7.468894673543181e-07, "loss": 0.7393, "step": 515 }, { "epoch": 1.2169811320754718, "grad_norm": 0.8300278395620482, "learning_rate": 7.430827993733808e-07, "loss": 0.729, "step": 516 }, { "epoch": 1.2193396226415094, "grad_norm": 0.924932117248118, "learning_rate": 7.39280113257914e-07, "loss": 0.7418, "step": 517 }, { "epoch": 1.2216981132075473, "grad_norm": 0.8751602815649364, "learning_rate": 7.354814679443556e-07, "loss": 0.7695, "step": 518 }, { "epoch": 1.224056603773585, "grad_norm": 0.7891993502461758, "learning_rate": 7.316869223065155e-07, "loss": 0.7755, "step": 519 }, { "epoch": 1.2264150943396226, "grad_norm": 0.8088776711707174, "learning_rate": 7.278965351546648e-07, "loss": 0.7572, "step": 520 }, { "epoch": 1.2287735849056605, "grad_norm": 0.7734655717390242, "learning_rate": 7.241103652346248e-07, "loss": 0.731, "step": 521 }, { "epoch": 1.2311320754716981, "grad_norm": 0.8075042544518887, "learning_rate": 7.20328471226854e-07, "loss": 0.7681, "step": 522 }, { "epoch": 1.2334905660377358, "grad_norm": 0.8682085195199817, "learning_rate": 7.165509117455406e-07, "loss": 0.7869, "step": 523 }, { "epoch": 1.2358490566037736, "grad_norm": 0.7941356815577187, "learning_rate": 7.127777453376935e-07, "loss": 0.7839, "step": 524 }, { "epoch": 1.2382075471698113, "grad_norm": 0.797011639565925, "learning_rate": 7.090090304822355e-07, "loss": 0.7704, "step": 525 }, { "epoch": 1.240566037735849, "grad_norm": 0.778511764914176, "learning_rate": 7.052448255890957e-07, "loss": 0.7684, "step": 526 }, { "epoch": 1.2429245283018868, "grad_norm": 0.9154469456152399, "learning_rate": 7.014851889983057e-07, "loss": 0.747, "step": 527 }, { "epoch": 1.2452830188679245, "grad_norm": 0.7896606577750244, "learning_rate": 6.97730178979093e-07, "loss": 0.7243, "step": 528 }, { "epoch": 1.2476415094339623, "grad_norm": 0.8076356582882437, "learning_rate": 6.939798537289825e-07, "loss": 0.7879, "step": 529 }, { "epoch": 1.25, "grad_norm": 0.8038075591560918, "learning_rate": 6.902342713728886e-07, "loss": 0.7304, "step": 530 }, { "epoch": 1.25, "eval_loss": 0.6839449405670166, "eval_runtime": 82.7741, "eval_samples_per_second": 8.517, "eval_steps_per_second": 0.544, "step": 530 }, { "epoch": 1.2523584905660377, "grad_norm": 0.7543374055980275, "learning_rate": 6.864934899622191e-07, "loss": 0.7671, "step": 531 }, { "epoch": 1.2547169811320755, "grad_norm": 0.7940675058566147, "learning_rate": 6.827575674739725e-07, "loss": 0.7509, "step": 532 }, { "epoch": 1.2570754716981132, "grad_norm": 0.7898168085125467, "learning_rate": 6.790265618098423e-07, "loss": 0.7819, "step": 533 }, { "epoch": 1.259433962264151, "grad_norm": 0.8290907307209445, "learning_rate": 6.753005307953165e-07, "loss": 0.7433, "step": 534 }, { "epoch": 1.2617924528301887, "grad_norm": 0.8051377335366591, "learning_rate": 6.715795321787836e-07, "loss": 0.7583, "step": 535 }, { "epoch": 1.2641509433962264, "grad_norm": 0.7964463188550859, "learning_rate": 6.678636236306357e-07, "loss": 0.7733, "step": 536 }, { "epoch": 1.2665094339622642, "grad_norm": 1.0183118136986125, "learning_rate": 6.641528627423774e-07, "loss": 0.7648, "step": 537 }, { "epoch": 1.2688679245283019, "grad_norm": 0.7893378901295127, "learning_rate": 6.604473070257308e-07, "loss": 0.7543, "step": 538 }, { "epoch": 1.2712264150943398, "grad_norm": 0.962839563526039, "learning_rate": 6.567470139117447e-07, "loss": 0.7815, "step": 539 }, { "epoch": 1.2735849056603774, "grad_norm": 0.779909096149425, "learning_rate": 6.530520407499049e-07, "loss": 0.7321, "step": 540 }, { "epoch": 1.275943396226415, "grad_norm": 1.0157939390615558, "learning_rate": 6.493624448072457e-07, "loss": 0.7622, "step": 541 }, { "epoch": 1.2783018867924527, "grad_norm": 0.9839155144434717, "learning_rate": 6.456782832674613e-07, "loss": 0.7498, "step": 542 }, { "epoch": 1.2806603773584906, "grad_norm": 0.8287088174891809, "learning_rate": 6.419996132300203e-07, "loss": 0.7479, "step": 543 }, { "epoch": 1.2830188679245282, "grad_norm": 0.8401170082742024, "learning_rate": 6.383264917092801e-07, "loss": 0.7398, "step": 544 }, { "epoch": 1.2853773584905661, "grad_norm": 0.8649864952648277, "learning_rate": 6.34658975633605e-07, "loss": 0.7629, "step": 545 }, { "epoch": 1.2877358490566038, "grad_norm": 0.7978032312823685, "learning_rate": 6.30997121844481e-07, "loss": 0.7455, "step": 546 }, { "epoch": 1.2900943396226414, "grad_norm": 0.78443421199377, "learning_rate": 6.273409870956369e-07, "loss": 0.7512, "step": 547 }, { "epoch": 1.2924528301886793, "grad_norm": 0.9742229809355082, "learning_rate": 6.236906280521646e-07, "loss": 0.7629, "step": 548 }, { "epoch": 1.294811320754717, "grad_norm": 0.8605023916238441, "learning_rate": 6.200461012896401e-07, "loss": 0.7389, "step": 549 }, { "epoch": 1.2971698113207548, "grad_norm": 0.7976345004555616, "learning_rate": 6.164074632932472e-07, "loss": 0.7675, "step": 550 }, { "epoch": 1.2995283018867925, "grad_norm": 0.8766199870243025, "learning_rate": 6.127747704569015e-07, "loss": 0.7392, "step": 551 }, { "epoch": 1.3018867924528301, "grad_norm": 0.8388074881151663, "learning_rate": 6.091480790823771e-07, "loss": 0.7594, "step": 552 }, { "epoch": 1.304245283018868, "grad_norm": 0.7816816297148448, "learning_rate": 6.055274453784338e-07, "loss": 0.7754, "step": 553 }, { "epoch": 1.3066037735849056, "grad_norm": 0.8051939468008078, "learning_rate": 6.019129254599456e-07, "loss": 0.7456, "step": 554 }, { "epoch": 1.3089622641509435, "grad_norm": 0.7954437915413475, "learning_rate": 5.983045753470307e-07, "loss": 0.7804, "step": 555 }, { "epoch": 1.3113207547169812, "grad_norm": 0.7995393304495568, "learning_rate": 5.947024509641841e-07, "loss": 0.7346, "step": 556 }, { "epoch": 1.3136792452830188, "grad_norm": 0.7666328856774342, "learning_rate": 5.911066081394112e-07, "loss": 0.7641, "step": 557 }, { "epoch": 1.3160377358490565, "grad_norm": 0.9135245763682434, "learning_rate": 5.875171026033608e-07, "loss": 0.7493, "step": 558 }, { "epoch": 1.3183962264150944, "grad_norm": 0.7843209916172604, "learning_rate": 5.839339899884628e-07, "loss": 0.754, "step": 559 }, { "epoch": 1.320754716981132, "grad_norm": 0.7798421629061796, "learning_rate": 5.803573258280654e-07, "loss": 0.7453, "step": 560 }, { "epoch": 1.3231132075471699, "grad_norm": 0.8026516589712169, "learning_rate": 5.76787165555575e-07, "loss": 0.7791, "step": 561 }, { "epoch": 1.3254716981132075, "grad_norm": 0.8161986502002226, "learning_rate": 5.732235645035963e-07, "loss": 0.7435, "step": 562 }, { "epoch": 1.3278301886792452, "grad_norm": 0.7982394811556922, "learning_rate": 5.696665779030749e-07, "loss": 0.7312, "step": 563 }, { "epoch": 1.330188679245283, "grad_norm": 0.8424060129449961, "learning_rate": 5.661162608824419e-07, "loss": 0.7209, "step": 564 }, { "epoch": 1.3325471698113207, "grad_norm": 0.8393522976074216, "learning_rate": 5.625726684667585e-07, "loss": 0.7497, "step": 565 }, { "epoch": 1.3349056603773586, "grad_norm": 0.8458006058371306, "learning_rate": 5.590358555768642e-07, "loss": 0.7415, "step": 566 }, { "epoch": 1.3372641509433962, "grad_norm": 0.7731963506173513, "learning_rate": 5.555058770285246e-07, "loss": 0.7289, "step": 567 }, { "epoch": 1.3396226415094339, "grad_norm": 0.919340196876583, "learning_rate": 5.519827875315823e-07, "loss": 0.7372, "step": 568 }, { "epoch": 1.3419811320754718, "grad_norm": 0.8047418954501285, "learning_rate": 5.484666416891108e-07, "loss": 0.7854, "step": 569 }, { "epoch": 1.3443396226415094, "grad_norm": 0.8420862799710107, "learning_rate": 5.449574939965636e-07, "loss": 0.7487, "step": 570 }, { "epoch": 1.3466981132075473, "grad_norm": 0.862354237686273, "learning_rate": 5.414553988409343e-07, "loss": 0.7615, "step": 571 }, { "epoch": 1.349056603773585, "grad_norm": 0.764673226355618, "learning_rate": 5.379604104999109e-07, "loss": 0.7402, "step": 572 }, { "epoch": 1.3514150943396226, "grad_norm": 0.9229852280599696, "learning_rate": 5.344725831410368e-07, "loss": 0.746, "step": 573 }, { "epoch": 1.3537735849056602, "grad_norm": 0.7749939321603743, "learning_rate": 5.309919708208686e-07, "loss": 0.7393, "step": 574 }, { "epoch": 1.3561320754716981, "grad_norm": 1.0855469699332105, "learning_rate": 5.275186274841404e-07, "loss": 0.7292, "step": 575 }, { "epoch": 1.3584905660377358, "grad_norm": 1.3079826302270743, "learning_rate": 5.240526069629264e-07, "loss": 0.7551, "step": 576 }, { "epoch": 1.3608490566037736, "grad_norm": 0.8097304218867772, "learning_rate": 5.205939629758078e-07, "loss": 0.7554, "step": 577 }, { "epoch": 1.3632075471698113, "grad_norm": 0.7629186235386943, "learning_rate": 5.171427491270393e-07, "loss": 0.7448, "step": 578 }, { "epoch": 1.365566037735849, "grad_norm": 0.8231834244100196, "learning_rate": 5.136990189057187e-07, "loss": 0.7675, "step": 579 }, { "epoch": 1.3679245283018868, "grad_norm": 0.7758543745731015, "learning_rate": 5.102628256849575e-07, "loss": 0.7643, "step": 580 }, { "epoch": 1.3702830188679245, "grad_norm": 0.7915005392947032, "learning_rate": 5.068342227210545e-07, "loss": 0.732, "step": 581 }, { "epoch": 1.3726415094339623, "grad_norm": 0.821233361833331, "learning_rate": 5.034132631526695e-07, "loss": 0.7145, "step": 582 }, { "epoch": 1.375, "grad_norm": 0.8288831867947671, "learning_rate": 5.000000000000002e-07, "loss": 0.7717, "step": 583 }, { "epoch": 1.3773584905660377, "grad_norm": 0.9380885957065556, "learning_rate": 4.9659448616396e-07, "loss": 0.7655, "step": 584 }, { "epoch": 1.3797169811320755, "grad_norm": 0.769393956851887, "learning_rate": 4.9319677442536e-07, "loss": 0.7481, "step": 585 }, { "epoch": 1.3820754716981132, "grad_norm": 0.7624074760874653, "learning_rate": 4.89806917444088e-07, "loss": 0.7408, "step": 586 }, { "epoch": 1.384433962264151, "grad_norm": 0.8836661905432738, "learning_rate": 4.864249677582935e-07, "loss": 0.7423, "step": 587 }, { "epoch": 1.3867924528301887, "grad_norm": 1.7675171950723623, "learning_rate": 4.830509777835744e-07, "loss": 0.7414, "step": 588 }, { "epoch": 1.3891509433962264, "grad_norm": 0.8297208161937744, "learning_rate": 4.796849998121647e-07, "loss": 0.7571, "step": 589 }, { "epoch": 1.3915094339622642, "grad_norm": 0.8105952844684988, "learning_rate": 4.7632708601212215e-07, "loss": 0.7473, "step": 590 }, { "epoch": 1.3938679245283019, "grad_norm": 0.943401465741068, "learning_rate": 4.7297728842652116e-07, "loss": 0.7885, "step": 591 }, { "epoch": 1.3962264150943398, "grad_norm": 0.8407405967900344, "learning_rate": 4.6963565897264623e-07, "loss": 0.7768, "step": 592 }, { "epoch": 1.3985849056603774, "grad_norm": 0.7742620223687944, "learning_rate": 4.663022494411866e-07, "loss": 0.7679, "step": 593 }, { "epoch": 1.400943396226415, "grad_norm": 0.7984390061772534, "learning_rate": 4.6297711149543405e-07, "loss": 0.7503, "step": 594 }, { "epoch": 1.4033018867924527, "grad_norm": 0.7687242314529935, "learning_rate": 4.596602966704822e-07, "loss": 0.7593, "step": 595 }, { "epoch": 1.4056603773584906, "grad_norm": 0.7752552260078952, "learning_rate": 4.5635185637242735e-07, "loss": 0.7236, "step": 596 }, { "epoch": 1.4080188679245282, "grad_norm": 0.7909608031012165, "learning_rate": 4.530518418775733e-07, "loss": 0.7487, "step": 597 }, { "epoch": 1.4103773584905661, "grad_norm": 1.4293658648035819, "learning_rate": 4.4976030433163337e-07, "loss": 0.7619, "step": 598 }, { "epoch": 1.4127358490566038, "grad_norm": 0.9601629598020603, "learning_rate": 4.4647729474894123e-07, "loss": 0.7301, "step": 599 }, { "epoch": 1.4150943396226414, "grad_norm": 0.8524971604621474, "learning_rate": 4.432028640116581e-07, "loss": 0.7743, "step": 600 }, { "epoch": 1.4174528301886793, "grad_norm": 0.7955465630229851, "learning_rate": 4.399370628689857e-07, "loss": 0.7549, "step": 601 }, { "epoch": 1.419811320754717, "grad_norm": 0.8461044127574078, "learning_rate": 4.366799419363779e-07, "loss": 0.7923, "step": 602 }, { "epoch": 1.4221698113207548, "grad_norm": 0.7959852549975004, "learning_rate": 4.3343155169475797e-07, "loss": 0.7313, "step": 603 }, { "epoch": 1.4245283018867925, "grad_norm": 0.8115147827054038, "learning_rate": 4.3019194248973377e-07, "loss": 0.7353, "step": 604 }, { "epoch": 1.4268867924528301, "grad_norm": 0.7884615566733161, "learning_rate": 4.269611645308214e-07, "loss": 0.7396, "step": 605 }, { "epoch": 1.429245283018868, "grad_norm": 0.8709874327087314, "learning_rate": 4.237392678906633e-07, "loss": 0.7557, "step": 606 }, { "epoch": 1.4316037735849056, "grad_norm": 0.9486373793118293, "learning_rate": 4.205263025042538e-07, "loss": 0.7397, "step": 607 }, { "epoch": 1.4339622641509435, "grad_norm": 0.8162741407892988, "learning_rate": 4.173223181681651e-07, "loss": 0.7327, "step": 608 }, { "epoch": 1.4363207547169812, "grad_norm": 0.8564525506081095, "learning_rate": 4.141273645397754e-07, "loss": 0.7745, "step": 609 }, { "epoch": 1.4386792452830188, "grad_norm": 0.8291423725092533, "learning_rate": 4.109414911364992e-07, "loss": 0.761, "step": 610 }, { "epoch": 1.4410377358490565, "grad_norm": 0.9110035252316134, "learning_rate": 4.0776474733502007e-07, "loss": 0.7227, "step": 611 }, { "epoch": 1.4433962264150944, "grad_norm": 7.248992388941307, "learning_rate": 4.045971823705249e-07, "loss": 0.7515, "step": 612 }, { "epoch": 1.445754716981132, "grad_norm": 0.737128592265643, "learning_rate": 4.0143884533594197e-07, "loss": 0.7308, "step": 613 }, { "epoch": 1.4481132075471699, "grad_norm": 0.7878766896489899, "learning_rate": 3.982897851811786e-07, "loss": 0.7323, "step": 614 }, { "epoch": 1.4504716981132075, "grad_norm": 0.7859241609615767, "learning_rate": 3.951500507123627e-07, "loss": 0.7814, "step": 615 }, { "epoch": 1.4528301886792452, "grad_norm": 0.8312739174249127, "learning_rate": 3.920196905910873e-07, "loss": 0.7446, "step": 616 }, { "epoch": 1.455188679245283, "grad_norm": 0.8160254075021761, "learning_rate": 3.888987533336566e-07, "loss": 0.745, "step": 617 }, { "epoch": 1.4575471698113207, "grad_norm": 0.9271092228733877, "learning_rate": 3.8578728731033214e-07, "loss": 0.7875, "step": 618 }, { "epoch": 1.4599056603773586, "grad_norm": 1.3059448042170938, "learning_rate": 3.826853407445848e-07, "loss": 0.7758, "step": 619 }, { "epoch": 1.4622641509433962, "grad_norm": 0.7940556961073607, "learning_rate": 3.795929617123468e-07, "loss": 0.7289, "step": 620 }, { "epoch": 1.4646226415094339, "grad_norm": 0.7908440739021586, "learning_rate": 3.765101981412665e-07, "loss": 0.7247, "step": 621 }, { "epoch": 1.4669811320754718, "grad_norm": 0.807553140785391, "learning_rate": 3.7343709780996614e-07, "loss": 0.7365, "step": 622 }, { "epoch": 1.4693396226415094, "grad_norm": 2.5050420828496787, "learning_rate": 3.703737083473005e-07, "loss": 0.7729, "step": 623 }, { "epoch": 1.4716981132075473, "grad_norm": 0.8163112108481482, "learning_rate": 3.673200772316193e-07, "loss": 0.7269, "step": 624 }, { "epoch": 1.474056603773585, "grad_norm": 0.8451238760005333, "learning_rate": 3.6427625179003217e-07, "loss": 0.7435, "step": 625 }, { "epoch": 1.4764150943396226, "grad_norm": 1.0290141575096419, "learning_rate": 3.6124227919767227e-07, "loss": 0.7488, "step": 626 }, { "epoch": 1.4787735849056602, "grad_norm": 0.7955100632691958, "learning_rate": 3.5821820647696864e-07, "loss": 0.7556, "step": 627 }, { "epoch": 1.4811320754716981, "grad_norm": 0.8238206028160229, "learning_rate": 3.552040804969149e-07, "loss": 0.7541, "step": 628 }, { "epoch": 1.4834905660377358, "grad_norm": 0.8519020092733177, "learning_rate": 3.5219994797234476e-07, "loss": 0.7783, "step": 629 }, { "epoch": 1.4858490566037736, "grad_norm": 0.7845182991866497, "learning_rate": 3.4920585546320625e-07, "loss": 0.7457, "step": 630 }, { "epoch": 1.4882075471698113, "grad_norm": 0.7643446889279868, "learning_rate": 3.4622184937384156e-07, "loss": 0.7637, "step": 631 }, { "epoch": 1.490566037735849, "grad_norm": 0.8986522134788724, "learning_rate": 3.4324797595226564e-07, "loss": 0.7431, "step": 632 }, { "epoch": 1.4929245283018868, "grad_norm": 0.788015002889114, "learning_rate": 3.4028428128945286e-07, "loss": 0.746, "step": 633 }, { "epoch": 1.4952830188679245, "grad_norm": 0.7754842750109397, "learning_rate": 3.3733081131861975e-07, "loss": 0.7688, "step": 634 }, { "epoch": 1.4976415094339623, "grad_norm": 0.8009965193531273, "learning_rate": 3.343876118145141e-07, "loss": 0.7501, "step": 635 }, { "epoch": 1.5, "grad_norm": 1.006640510210976, "learning_rate": 3.314547283927057e-07, "loss": 0.7239, "step": 636 }, { "epoch": 1.5, "eval_loss": 0.6808757185935974, "eval_runtime": 83.2321, "eval_samples_per_second": 8.47, "eval_steps_per_second": 0.541, "step": 636 }, { "epoch": 1.5023584905660377, "grad_norm": 0.8598327085719165, "learning_rate": 3.2853220650887913e-07, "loss": 0.7149, "step": 637 }, { "epoch": 1.5047169811320755, "grad_norm": 0.8482092774985639, "learning_rate": 3.256200914581292e-07, "loss": 0.7556, "step": 638 }, { "epoch": 1.5070754716981132, "grad_norm": 0.7837630460537407, "learning_rate": 3.227184283742591e-07, "loss": 0.7432, "step": 639 }, { "epoch": 1.509433962264151, "grad_norm": 0.8392002898875265, "learning_rate": 3.198272622290804e-07, "loss": 0.7395, "step": 640 }, { "epoch": 1.5117924528301887, "grad_norm": 0.826051018451035, "learning_rate": 3.169466378317177e-07, "loss": 0.7849, "step": 641 }, { "epoch": 1.5141509433962264, "grad_norm": 1.0059361369860251, "learning_rate": 3.1407659982791204e-07, "loss": 0.7541, "step": 642 }, { "epoch": 1.516509433962264, "grad_norm": 0.8092490329577119, "learning_rate": 3.112171926993291e-07, "loss": 0.7348, "step": 643 }, { "epoch": 1.5188679245283019, "grad_norm": 0.7280559308125196, "learning_rate": 3.0836846076287146e-07, "loss": 0.6873, "step": 644 }, { "epoch": 1.5212264150943398, "grad_norm": 0.8517206567349668, "learning_rate": 3.055304481699913e-07, "loss": 0.7839, "step": 645 }, { "epoch": 1.5235849056603774, "grad_norm": 0.8340097914857253, "learning_rate": 3.027031989060046e-07, "loss": 0.7561, "step": 646 }, { "epoch": 1.525943396226415, "grad_norm": 0.8002849424485019, "learning_rate": 2.998867567894108e-07, "loss": 0.7747, "step": 647 }, { "epoch": 1.5283018867924527, "grad_norm": 0.8623306309909063, "learning_rate": 2.970811654712133e-07, "loss": 0.7337, "step": 648 }, { "epoch": 1.5306603773584906, "grad_norm": 0.8335419959901207, "learning_rate": 2.942864684342432e-07, "loss": 0.741, "step": 649 }, { "epoch": 1.5330188679245285, "grad_norm": 0.8731062418731657, "learning_rate": 2.91502708992485e-07, "loss": 0.7424, "step": 650 }, { "epoch": 1.5353773584905661, "grad_norm": 0.819811746630834, "learning_rate": 2.8872993029040506e-07, "loss": 0.7355, "step": 651 }, { "epoch": 1.5377358490566038, "grad_norm": 0.8074813557073929, "learning_rate": 2.859681753022838e-07, "loss": 0.7594, "step": 652 }, { "epoch": 1.5400943396226414, "grad_norm": 0.820619090005774, "learning_rate": 2.8321748683154887e-07, "loss": 0.7443, "step": 653 }, { "epoch": 1.5424528301886793, "grad_norm": 0.8554565499942881, "learning_rate": 2.8047790751011216e-07, "loss": 0.6991, "step": 654 }, { "epoch": 1.544811320754717, "grad_norm": 0.7784690870870672, "learning_rate": 2.777494797977088e-07, "loss": 0.7328, "step": 655 }, { "epoch": 1.5471698113207548, "grad_norm": 1.7923607795927756, "learning_rate": 2.7503224598123895e-07, "loss": 0.7266, "step": 656 }, { "epoch": 1.5495283018867925, "grad_norm": 0.7677443512564135, "learning_rate": 2.7232624817411376e-07, "loss": 0.737, "step": 657 }, { "epoch": 1.5518867924528301, "grad_norm": 0.8055758349794541, "learning_rate": 2.6963152831560066e-07, "loss": 0.7444, "step": 658 }, { "epoch": 1.5542452830188678, "grad_norm": 0.9008016637756693, "learning_rate": 2.6694812817017387e-07, "loss": 0.7128, "step": 659 }, { "epoch": 1.5566037735849056, "grad_norm": 0.8106690468627914, "learning_rate": 2.642760893268684e-07, "loss": 0.7457, "step": 660 }, { "epoch": 1.5589622641509435, "grad_norm": 1.027228042593183, "learning_rate": 2.616154531986345e-07, "loss": 0.7251, "step": 661 }, { "epoch": 1.5613207547169812, "grad_norm": 0.7957920778904417, "learning_rate": 2.5896626102169594e-07, "loss": 0.7481, "step": 662 }, { "epoch": 1.5636792452830188, "grad_norm": 0.7514802697133819, "learning_rate": 2.5632855385491037e-07, "loss": 0.757, "step": 663 }, { "epoch": 1.5660377358490565, "grad_norm": 0.8814866050056973, "learning_rate": 2.53702372579134e-07, "loss": 0.7361, "step": 664 }, { "epoch": 1.5683962264150944, "grad_norm": 0.8346312832230348, "learning_rate": 2.51087757896587e-07, "loss": 0.7378, "step": 665 }, { "epoch": 1.5707547169811322, "grad_norm": 0.8164772219257521, "learning_rate": 2.4848475033022377e-07, "loss": 0.7631, "step": 666 }, { "epoch": 1.5731132075471699, "grad_norm": 0.7933267558316177, "learning_rate": 2.458933902231038e-07, "loss": 0.7288, "step": 667 }, { "epoch": 1.5754716981132075, "grad_norm": 0.8420080383383726, "learning_rate": 2.4331371773776687e-07, "loss": 0.77, "step": 668 }, { "epoch": 1.5778301886792452, "grad_norm": 5.289720197423799, "learning_rate": 2.407457728556115e-07, "loss": 0.7184, "step": 669 }, { "epoch": 1.580188679245283, "grad_norm": 1.0842649090634284, "learning_rate": 2.3818959537627282e-07, "loss": 0.7383, "step": 670 }, { "epoch": 1.5825471698113207, "grad_norm": 0.8519744691939684, "learning_rate": 2.3564522491700833e-07, "loss": 0.769, "step": 671 }, { "epoch": 1.5849056603773586, "grad_norm": 0.8168668326966967, "learning_rate": 2.3311270091208256e-07, "loss": 0.7444, "step": 672 }, { "epoch": 1.5872641509433962, "grad_norm": 0.7514913975722562, "learning_rate": 2.3059206261215668e-07, "loss": 0.691, "step": 673 }, { "epoch": 1.5896226415094339, "grad_norm": 0.7519874181403705, "learning_rate": 2.2808334908367909e-07, "loss": 0.7425, "step": 674 }, { "epoch": 1.5919811320754715, "grad_norm": 0.8353526410754343, "learning_rate": 2.2558659920828095e-07, "loss": 0.7616, "step": 675 }, { "epoch": 1.5943396226415094, "grad_norm": 0.7738579301027763, "learning_rate": 2.2310185168217212e-07, "loss": 0.7756, "step": 676 }, { "epoch": 1.5966981132075473, "grad_norm": 0.813046680487735, "learning_rate": 2.206291450155441e-07, "loss": 0.7675, "step": 677 }, { "epoch": 1.599056603773585, "grad_norm": 0.7815916852088219, "learning_rate": 2.181685175319702e-07, "loss": 0.7459, "step": 678 }, { "epoch": 1.6014150943396226, "grad_norm": 0.7482293961455181, "learning_rate": 2.157200073678137e-07, "loss": 0.7237, "step": 679 }, { "epoch": 1.6037735849056602, "grad_norm": 0.9475759536537903, "learning_rate": 2.132836524716355e-07, "loss": 0.718, "step": 680 }, { "epoch": 1.6061320754716981, "grad_norm": 0.8731724054191227, "learning_rate": 2.1085949060360653e-07, "loss": 0.7344, "step": 681 }, { "epoch": 1.608490566037736, "grad_norm": 0.8124879077769539, "learning_rate": 2.0844755933492263e-07, "loss": 0.7515, "step": 682 }, { "epoch": 1.6108490566037736, "grad_norm": 0.7772794085957034, "learning_rate": 2.0604789604722205e-07, "loss": 0.7471, "step": 683 }, { "epoch": 1.6132075471698113, "grad_norm": 0.8144653370194169, "learning_rate": 2.0366053793200565e-07, "loss": 0.7724, "step": 684 }, { "epoch": 1.615566037735849, "grad_norm": 1.1024190577460233, "learning_rate": 2.0128552199006198e-07, "loss": 0.7389, "step": 685 }, { "epoch": 1.6179245283018868, "grad_norm": 0.7813510804481604, "learning_rate": 1.9892288503089205e-07, "loss": 0.7688, "step": 686 }, { "epoch": 1.6202830188679245, "grad_norm": 0.7649385235808642, "learning_rate": 1.9657266367213898e-07, "loss": 0.7279, "step": 687 }, { "epoch": 1.6226415094339623, "grad_norm": 0.8664030315613933, "learning_rate": 1.9423489433902184e-07, "loss": 0.7604, "step": 688 }, { "epoch": 1.625, "grad_norm": 0.8134285929715391, "learning_rate": 1.9190961326377053e-07, "loss": 0.7628, "step": 689 }, { "epoch": 1.6273584905660377, "grad_norm": 0.8126410627826297, "learning_rate": 1.8959685648506362e-07, "loss": 0.7479, "step": 690 }, { "epoch": 1.6297169811320755, "grad_norm": 0.8243887159653667, "learning_rate": 1.8729665984747e-07, "loss": 0.743, "step": 691 }, { "epoch": 1.6320754716981132, "grad_norm": 0.8107223249523531, "learning_rate": 1.8500905900089403e-07, "loss": 0.7601, "step": 692 }, { "epoch": 1.634433962264151, "grad_norm": 0.8038068007057135, "learning_rate": 1.82734089400022e-07, "loss": 0.7643, "step": 693 }, { "epoch": 1.6367924528301887, "grad_norm": 0.9712605139438796, "learning_rate": 1.804717863037737e-07, "loss": 0.7445, "step": 694 }, { "epoch": 1.6391509433962264, "grad_norm": 0.7879085057169681, "learning_rate": 1.7822218477475494e-07, "loss": 0.7485, "step": 695 }, { "epoch": 1.641509433962264, "grad_norm": 0.8550428267297568, "learning_rate": 1.7598531967871465e-07, "loss": 0.7783, "step": 696 }, { "epoch": 1.6438679245283019, "grad_norm": 0.7795732830836477, "learning_rate": 1.737612256840053e-07, "loss": 0.7484, "step": 697 }, { "epoch": 1.6462264150943398, "grad_norm": 0.8133414594705825, "learning_rate": 1.7154993726104328e-07, "loss": 0.7751, "step": 698 }, { "epoch": 1.6485849056603774, "grad_norm": 0.7994545997274072, "learning_rate": 1.6935148868177718e-07, "loss": 0.7501, "step": 699 }, { "epoch": 1.650943396226415, "grad_norm": 0.7496147135106944, "learning_rate": 1.6716591401915502e-07, "loss": 0.759, "step": 700 }, { "epoch": 1.6533018867924527, "grad_norm": 0.7274661448787512, "learning_rate": 1.6499324714659758e-07, "loss": 0.735, "step": 701 }, { "epoch": 1.6556603773584906, "grad_norm": 0.8892069878489233, "learning_rate": 1.6283352173747146e-07, "loss": 0.7694, "step": 702 }, { "epoch": 1.6580188679245285, "grad_norm": 0.7851702666302485, "learning_rate": 1.6068677126456897e-07, "loss": 0.7373, "step": 703 }, { "epoch": 1.6603773584905661, "grad_norm": 0.7771975549062173, "learning_rate": 1.585530289995878e-07, "loss": 0.7101, "step": 704 }, { "epoch": 1.6627358490566038, "grad_norm": 0.8852742002247123, "learning_rate": 1.564323280126173e-07, "loss": 0.7525, "step": 705 }, { "epoch": 1.6650943396226414, "grad_norm": 1.0210362257937011, "learning_rate": 1.5432470117162433e-07, "loss": 0.7752, "step": 706 }, { "epoch": 1.6674528301886793, "grad_norm": 0.8077571397037019, "learning_rate": 1.522301811419442e-07, "loss": 0.7395, "step": 707 }, { "epoch": 1.669811320754717, "grad_norm": 0.8331764729151993, "learning_rate": 1.5014880038577482e-07, "loss": 0.7416, "step": 708 }, { "epoch": 1.6721698113207548, "grad_norm": 2.1774846943729935, "learning_rate": 1.4808059116167303e-07, "loss": 0.7492, "step": 709 }, { "epoch": 1.6745283018867925, "grad_norm": 0.8939523687224223, "learning_rate": 1.460255855240552e-07, "loss": 0.7358, "step": 710 }, { "epoch": 1.6768867924528301, "grad_norm": 0.7563678829167328, "learning_rate": 1.4398381532269998e-07, "loss": 0.7593, "step": 711 }, { "epoch": 1.6792452830188678, "grad_norm": 0.7668313174457733, "learning_rate": 1.4195531220225487e-07, "loss": 0.7275, "step": 712 }, { "epoch": 1.6816037735849056, "grad_norm": 0.7791231762469476, "learning_rate": 1.39940107601746e-07, "loss": 0.7692, "step": 713 }, { "epoch": 1.6839622641509435, "grad_norm": 1.014530101254706, "learning_rate": 1.3793823275409066e-07, "loss": 0.7475, "step": 714 }, { "epoch": 1.6863207547169812, "grad_norm": 0.754925190142359, "learning_rate": 1.3594971868561232e-07, "loss": 0.7248, "step": 715 }, { "epoch": 1.6886792452830188, "grad_norm": 0.7608021380299098, "learning_rate": 1.3397459621556128e-07, "loss": 0.7558, "step": 716 }, { "epoch": 1.6910377358490565, "grad_norm": 1.320307190976547, "learning_rate": 1.320128959556369e-07, "loss": 0.7393, "step": 717 }, { "epoch": 1.6933962264150944, "grad_norm": 0.8339214047495644, "learning_rate": 1.300646483095118e-07, "loss": 0.7597, "step": 718 }, { "epoch": 1.6957547169811322, "grad_norm": 0.8826119280765871, "learning_rate": 1.2812988347236166e-07, "loss": 0.7416, "step": 719 }, { "epoch": 1.6981132075471699, "grad_norm": 0.8928251184591149, "learning_rate": 1.262086314303973e-07, "loss": 0.7318, "step": 720 }, { "epoch": 1.7004716981132075, "grad_norm": 0.7951715947182522, "learning_rate": 1.243009219603993e-07, "loss": 0.7482, "step": 721 }, { "epoch": 1.7028301886792452, "grad_norm": 0.8741704408992066, "learning_rate": 1.2240678462925723e-07, "loss": 0.7287, "step": 722 }, { "epoch": 1.705188679245283, "grad_norm": 0.7786335970762307, "learning_rate": 1.2052624879351103e-07, "loss": 0.719, "step": 723 }, { "epoch": 1.7075471698113207, "grad_norm": 0.9985291709831805, "learning_rate": 1.1865934359889573e-07, "loss": 0.7453, "step": 724 }, { "epoch": 1.7099056603773586, "grad_norm": 0.7859405679363147, "learning_rate": 1.1680609797989038e-07, "loss": 0.7452, "step": 725 }, { "epoch": 1.7122641509433962, "grad_norm": 0.8028725796240722, "learning_rate": 1.1496654065926925e-07, "loss": 0.7173, "step": 726 }, { "epoch": 1.7146226415094339, "grad_norm": 0.8165982522602734, "learning_rate": 1.1314070014765642e-07, "loss": 0.7218, "step": 727 }, { "epoch": 1.7169811320754715, "grad_norm": 0.7860295163288017, "learning_rate": 1.1132860474308436e-07, "loss": 0.7614, "step": 728 }, { "epoch": 1.7193396226415094, "grad_norm": 0.7889724037710595, "learning_rate": 1.0953028253055541e-07, "loss": 0.7195, "step": 729 }, { "epoch": 1.7216981132075473, "grad_norm": 0.8067588097960763, "learning_rate": 1.0774576138160596e-07, "loss": 0.7482, "step": 730 }, { "epoch": 1.724056603773585, "grad_norm": 0.786061208841395, "learning_rate": 1.0597506895387499e-07, "loss": 0.7472, "step": 731 }, { "epoch": 1.7264150943396226, "grad_norm": 0.7917438727246212, "learning_rate": 1.0421823269067442e-07, "loss": 0.7751, "step": 732 }, { "epoch": 1.7287735849056602, "grad_norm": 0.7907162871021037, "learning_rate": 1.024752798205658e-07, "loss": 0.7202, "step": 733 }, { "epoch": 1.7311320754716981, "grad_norm": 0.9995843557820588, "learning_rate": 1.0074623735693633e-07, "loss": 0.7471, "step": 734 }, { "epoch": 1.733490566037736, "grad_norm": 1.3558698569505305, "learning_rate": 9.903113209758096e-08, "loss": 0.7564, "step": 735 }, { "epoch": 1.7358490566037736, "grad_norm": 0.7870396429020904, "learning_rate": 9.732999062428704e-08, "loss": 0.7641, "step": 736 }, { "epoch": 1.7382075471698113, "grad_norm": 0.8525068172515468, "learning_rate": 9.564283930242257e-08, "loss": 0.7404, "step": 737 }, { "epoch": 1.740566037735849, "grad_norm": 0.8014286873652903, "learning_rate": 9.396970428052697e-08, "loss": 0.7194, "step": 738 }, { "epoch": 1.7429245283018868, "grad_norm": 0.7676759180901973, "learning_rate": 9.231061148990648e-08, "loss": 0.7221, "step": 739 }, { "epoch": 1.7452830188679245, "grad_norm": 0.9131948150344564, "learning_rate": 9.066558664423163e-08, "loss": 0.7572, "step": 740 }, { "epoch": 1.7476415094339623, "grad_norm": 0.7734044180311592, "learning_rate": 8.903465523913955e-08, "loss": 0.7757, "step": 741 }, { "epoch": 1.75, "grad_norm": 0.755029746442279, "learning_rate": 8.741784255183759e-08, "loss": 0.7411, "step": 742 }, { "epoch": 1.75, "eval_loss": 0.6796497702598572, "eval_runtime": 82.4746, "eval_samples_per_second": 8.548, "eval_steps_per_second": 0.546, "step": 742 }, { "epoch": 1.7523584905660377, "grad_norm": 0.8497097319379918, "learning_rate": 8.581517364071267e-08, "loss": 0.7203, "step": 743 }, { "epoch": 1.7547169811320755, "grad_norm": 0.8020779183255858, "learning_rate": 8.422667334494249e-08, "loss": 0.7558, "step": 744 }, { "epoch": 1.7570754716981132, "grad_norm": 0.7616565389684601, "learning_rate": 8.265236628411087e-08, "loss": 0.7422, "step": 745 }, { "epoch": 1.759433962264151, "grad_norm": 0.9323853208771872, "learning_rate": 8.109227685782538e-08, "loss": 0.7819, "step": 746 }, { "epoch": 1.7617924528301887, "grad_norm": 1.350644299549973, "learning_rate": 7.954642924533994e-08, "loss": 0.7492, "step": 747 }, { "epoch": 1.7641509433962264, "grad_norm": 0.8092647636710026, "learning_rate": 7.801484740517939e-08, "loss": 0.7638, "step": 748 }, { "epoch": 1.766509433962264, "grad_norm": 0.8308445659034376, "learning_rate": 7.649755507476952e-08, "loss": 0.7444, "step": 749 }, { "epoch": 1.7688679245283019, "grad_norm": 2.5649621576701587, "learning_rate": 7.499457577006751e-08, "loss": 0.7713, "step": 750 }, { "epoch": 1.7712264150943398, "grad_norm": 0.786323334872225, "learning_rate": 7.350593278519823e-08, "loss": 0.7426, "step": 751 }, { "epoch": 1.7735849056603774, "grad_norm": 0.7795649745069329, "learning_rate": 7.203164919209359e-08, "loss": 0.7674, "step": 752 }, { "epoch": 1.775943396226415, "grad_norm": 0.8003158325581134, "learning_rate": 7.057174784013431e-08, "loss": 0.7531, "step": 753 }, { "epoch": 1.7783018867924527, "grad_norm": 0.7834816769548146, "learning_rate": 6.912625135579586e-08, "loss": 0.7212, "step": 754 }, { "epoch": 1.7806603773584906, "grad_norm": 0.8289452420732738, "learning_rate": 6.76951821422982e-08, "loss": 0.7715, "step": 755 }, { "epoch": 1.7830188679245285, "grad_norm": 0.8410465172590653, "learning_rate": 6.627856237925811e-08, "loss": 0.7276, "step": 756 }, { "epoch": 1.7853773584905661, "grad_norm": 0.8394202851068477, "learning_rate": 6.487641402234612e-08, "loss": 0.744, "step": 757 }, { "epoch": 1.7877358490566038, "grad_norm": 0.8383760649776106, "learning_rate": 6.348875880294535e-08, "loss": 0.7384, "step": 758 }, { "epoch": 1.7900943396226414, "grad_norm": 1.453114388460075, "learning_rate": 6.211561822781474e-08, "loss": 0.7765, "step": 759 }, { "epoch": 1.7924528301886793, "grad_norm": 0.8499896987869552, "learning_rate": 6.075701357875662e-08, "loss": 0.7284, "step": 760 }, { "epoch": 1.794811320754717, "grad_norm": 0.8652146496280481, "learning_rate": 5.9412965912286396e-08, "loss": 0.7561, "step": 761 }, { "epoch": 1.7971698113207548, "grad_norm": 0.8299032649447099, "learning_rate": 5.808349605930585e-08, "loss": 0.7565, "step": 762 }, { "epoch": 1.7995283018867925, "grad_norm": 0.7812451643671983, "learning_rate": 5.6768624624780604e-08, "loss": 0.725, "step": 763 }, { "epoch": 1.8018867924528301, "grad_norm": 0.8477713992666978, "learning_rate": 5.5468371987420936e-08, "loss": 0.7466, "step": 764 }, { "epoch": 1.8042452830188678, "grad_norm": 0.8337239533995923, "learning_rate": 5.4182758299365364e-08, "loss": 0.735, "step": 765 }, { "epoch": 1.8066037735849056, "grad_norm": 0.7691176408664865, "learning_rate": 5.29118034858691e-08, "loss": 0.772, "step": 766 }, { "epoch": 1.8089622641509435, "grad_norm": 0.8049526588971233, "learning_rate": 5.165552724499478e-08, "loss": 0.7315, "step": 767 }, { "epoch": 1.8113207547169812, "grad_norm": 0.8319464046786564, "learning_rate": 5.0413949047306894e-08, "loss": 0.7726, "step": 768 }, { "epoch": 1.8136792452830188, "grad_norm": 0.8474092435937203, "learning_rate": 4.918708813557093e-08, "loss": 0.7487, "step": 769 }, { "epoch": 1.8160377358490565, "grad_norm": 0.7943675010786753, "learning_rate": 4.797496352445396e-08, "loss": 0.7691, "step": 770 }, { "epoch": 1.8183962264150944, "grad_norm": 0.8217628363291084, "learning_rate": 4.677759400023085e-08, "loss": 0.7437, "step": 771 }, { "epoch": 1.8207547169811322, "grad_norm": 0.832358842867652, "learning_rate": 4.55949981204925e-08, "loss": 0.7568, "step": 772 }, { "epoch": 1.8231132075471699, "grad_norm": 0.8613869764376921, "learning_rate": 4.442719421385921e-08, "loss": 0.7274, "step": 773 }, { "epoch": 1.8254716981132075, "grad_norm": 0.7931363978514411, "learning_rate": 4.3274200379695315e-08, "loss": 0.7209, "step": 774 }, { "epoch": 1.8278301886792452, "grad_norm": 0.7551383801460104, "learning_rate": 4.213603448782932e-08, "loss": 0.7533, "step": 775 }, { "epoch": 1.830188679245283, "grad_norm": 0.8476707134264969, "learning_rate": 4.101271417827668e-08, "loss": 0.7585, "step": 776 }, { "epoch": 1.8325471698113207, "grad_norm": 0.8877158557106878, "learning_rate": 3.9904256860967433e-08, "loss": 0.7222, "step": 777 }, { "epoch": 1.8349056603773586, "grad_norm": 0.8633047498966683, "learning_rate": 3.881067971547469e-08, "loss": 0.7502, "step": 778 }, { "epoch": 1.8372641509433962, "grad_norm": 1.0520376373477054, "learning_rate": 3.7731999690749585e-08, "loss": 0.7443, "step": 779 }, { "epoch": 1.8396226415094339, "grad_norm": 0.7492875546723151, "learning_rate": 3.666823350485848e-08, "loss": 0.7781, "step": 780 }, { "epoch": 1.8419811320754715, "grad_norm": 0.7847944050831147, "learning_rate": 3.561939764472299e-08, "loss": 0.737, "step": 781 }, { "epoch": 1.8443396226415094, "grad_norm": 1.368419762468144, "learning_rate": 3.458550836586582e-08, "loss": 0.7523, "step": 782 }, { "epoch": 1.8466981132075473, "grad_norm": 0.8704033001785062, "learning_rate": 3.356658169215743e-08, "loss": 0.742, "step": 783 }, { "epoch": 1.849056603773585, "grad_norm": 0.764412444543419, "learning_rate": 3.2562633415568754e-08, "loss": 0.7327, "step": 784 }, { "epoch": 1.8514150943396226, "grad_norm": 0.8695554372256408, "learning_rate": 3.157367909592601e-08, "loss": 0.7338, "step": 785 }, { "epoch": 1.8537735849056602, "grad_norm": 0.7794614378925147, "learning_rate": 3.0599734060669626e-08, "loss": 0.7446, "step": 786 }, { "epoch": 1.8561320754716981, "grad_norm": 0.7712181969029962, "learning_rate": 2.9640813404616327e-08, "loss": 0.7377, "step": 787 }, { "epoch": 1.858490566037736, "grad_norm": 0.8826462854163208, "learning_rate": 2.869693198972556e-08, "loss": 0.7555, "step": 788 }, { "epoch": 1.8608490566037736, "grad_norm": 1.029113085742105, "learning_rate": 2.7768104444869434e-08, "loss": 0.7795, "step": 789 }, { "epoch": 1.8632075471698113, "grad_norm": 0.8318363964351398, "learning_rate": 2.6854345165605474e-08, "loss": 0.7351, "step": 790 }, { "epoch": 1.865566037735849, "grad_norm": 0.7865924156566725, "learning_rate": 2.595566831395346e-08, "loss": 0.7222, "step": 791 }, { "epoch": 1.8679245283018868, "grad_norm": 0.7955098420784998, "learning_rate": 2.507208781817638e-08, "loss": 0.7515, "step": 792 }, { "epoch": 1.8702830188679245, "grad_norm": 0.9146875621357804, "learning_rate": 2.4203617372564378e-08, "loss": 0.7173, "step": 793 }, { "epoch": 1.8726415094339623, "grad_norm": 0.7544835268896075, "learning_rate": 2.3350270437222374e-08, "loss": 0.7307, "step": 794 }, { "epoch": 1.875, "grad_norm": 1.806767663849715, "learning_rate": 2.2512060237861452e-08, "loss": 0.7301, "step": 795 }, { "epoch": 1.8773584905660377, "grad_norm": 0.804925906379938, "learning_rate": 2.1688999765594018e-08, "loss": 0.7552, "step": 796 }, { "epoch": 1.8797169811320755, "grad_norm": 0.7773879448134918, "learning_rate": 2.0881101776732967e-08, "loss": 0.7619, "step": 797 }, { "epoch": 1.8820754716981132, "grad_norm": 0.871709115043609, "learning_rate": 2.0088378792592286e-08, "loss": 0.7686, "step": 798 }, { "epoch": 1.884433962264151, "grad_norm": 0.7446217889352331, "learning_rate": 1.9310843099295204e-08, "loss": 0.7128, "step": 799 }, { "epoch": 1.8867924528301887, "grad_norm": 0.7661065857274754, "learning_rate": 1.8548506747582128e-08, "loss": 0.7284, "step": 800 }, { "epoch": 1.8891509433962264, "grad_norm": 0.7916406522727306, "learning_rate": 1.780138155262456e-08, "loss": 0.7321, "step": 801 }, { "epoch": 1.891509433962264, "grad_norm": 0.9641121942586747, "learning_rate": 1.7069479093842042e-08, "loss": 0.7411, "step": 802 }, { "epoch": 1.8938679245283019, "grad_norm": 0.7930873801533905, "learning_rate": 1.6352810714722387e-08, "loss": 0.7345, "step": 803 }, { "epoch": 1.8962264150943398, "grad_norm": 0.9298677041430008, "learning_rate": 1.565138752264572e-08, "loss": 0.7616, "step": 804 }, { "epoch": 1.8985849056603774, "grad_norm": 0.885574237101463, "learning_rate": 1.496522038871295e-08, "loss": 0.7572, "step": 805 }, { "epoch": 1.900943396226415, "grad_norm": 0.8153461825360933, "learning_rate": 1.4294319947577017e-08, "loss": 0.75, "step": 806 }, { "epoch": 1.9033018867924527, "grad_norm": 0.7660700579693493, "learning_rate": 1.3638696597277677e-08, "loss": 0.7421, "step": 807 }, { "epoch": 1.9056603773584906, "grad_norm": 0.805964323807633, "learning_rate": 1.2998360499080763e-08, "loss": 0.7958, "step": 808 }, { "epoch": 1.9080188679245285, "grad_norm": 0.9429626916004815, "learning_rate": 1.2373321577320628e-08, "loss": 0.734, "step": 809 }, { "epoch": 1.9103773584905661, "grad_norm": 0.7904102497155711, "learning_rate": 1.1763589519246387e-08, "loss": 0.7478, "step": 810 }, { "epoch": 1.9127358490566038, "grad_norm": 0.7928898152839706, "learning_rate": 1.1169173774871477e-08, "loss": 0.7373, "step": 811 }, { "epoch": 1.9150943396226414, "grad_norm": 0.7814723884700895, "learning_rate": 1.0590083556827556e-08, "loss": 0.7615, "step": 812 }, { "epoch": 1.9174528301886793, "grad_norm": 0.7802344182322299, "learning_rate": 1.0026327840221727e-08, "loss": 0.7503, "step": 813 }, { "epoch": 1.919811320754717, "grad_norm": 1.0812085791485797, "learning_rate": 9.477915362496758e-09, "loss": 0.7614, "step": 814 }, { "epoch": 1.9221698113207548, "grad_norm": 0.820203697559199, "learning_rate": 8.94485462329675e-09, "loss": 0.7374, "step": 815 }, { "epoch": 1.9245283018867925, "grad_norm": 0.8722587060636673, "learning_rate": 8.42715388433446e-09, "loss": 0.7328, "step": 816 }, { "epoch": 1.9268867924528301, "grad_norm": 0.7840984315269908, "learning_rate": 7.924821169263963e-09, "loss": 0.7542, "step": 817 }, { "epoch": 1.9292452830188678, "grad_norm": 0.8440330231818731, "learning_rate": 7.437864263555638e-09, "loss": 0.7317, "step": 818 }, { "epoch": 1.9316037735849056, "grad_norm": 0.895812932209063, "learning_rate": 6.966290714375933e-09, "loss": 0.7409, "step": 819 }, { "epoch": 1.9339622641509435, "grad_norm": 0.899638041623908, "learning_rate": 6.510107830470568e-09, "loss": 0.7446, "step": 820 }, { "epoch": 1.9363207547169812, "grad_norm": 0.8152793437016046, "learning_rate": 6.069322682050515e-09, "loss": 0.7634, "step": 821 }, { "epoch": 1.9386792452830188, "grad_norm": 0.7735814965641481, "learning_rate": 5.643942100683308e-09, "loss": 0.7809, "step": 822 }, { "epoch": 1.9410377358490565, "grad_norm": 0.8351447564551893, "learning_rate": 5.23397267918646e-09, "loss": 0.7588, "step": 823 }, { "epoch": 1.9433962264150944, "grad_norm": 0.7614597204440534, "learning_rate": 4.83942077152577e-09, "loss": 0.7231, "step": 824 }, { "epoch": 1.9457547169811322, "grad_norm": 0.8548976791386363, "learning_rate": 4.460292492716511e-09, "loss": 0.754, "step": 825 }, { "epoch": 1.9481132075471699, "grad_norm": 0.8381107415157564, "learning_rate": 4.0965937187287246e-09, "loss": 0.7351, "step": 826 }, { "epoch": 1.9504716981132075, "grad_norm": 5.772455437843555, "learning_rate": 3.748330086396523e-09, "loss": 0.7657, "step": 827 }, { "epoch": 1.9528301886792452, "grad_norm": 0.797010233582818, "learning_rate": 3.415506993330153e-09, "loss": 0.7248, "step": 828 }, { "epoch": 1.955188679245283, "grad_norm": 0.777781094479457, "learning_rate": 3.0981295978326216e-09, "loss": 0.7604, "step": 829 }, { "epoch": 1.9575471698113207, "grad_norm": 0.8449725908943951, "learning_rate": 2.7962028188198706e-09, "loss": 0.7736, "step": 830 }, { "epoch": 1.9599056603773586, "grad_norm": 0.7801120467589358, "learning_rate": 2.5097313357442806e-09, "loss": 0.7178, "step": 831 }, { "epoch": 1.9622641509433962, "grad_norm": 0.7762312098767981, "learning_rate": 2.2387195885221756e-09, "loss": 0.7575, "step": 832 }, { "epoch": 1.9646226415094339, "grad_norm": 0.824362916729911, "learning_rate": 1.983171777465431e-09, "loss": 0.76, "step": 833 }, { "epoch": 1.9669811320754715, "grad_norm": 0.9204175438402101, "learning_rate": 1.743091863215751e-09, "loss": 0.7312, "step": 834 }, { "epoch": 1.9693396226415094, "grad_norm": 0.8426151332993567, "learning_rate": 1.5184835666838258e-09, "loss": 0.7489, "step": 835 }, { "epoch": 1.9716981132075473, "grad_norm": 0.7997774511020088, "learning_rate": 1.3093503689910467e-09, "loss": 0.7243, "step": 836 }, { "epoch": 1.974056603773585, "grad_norm": 1.0915014977180026, "learning_rate": 1.1156955114162147e-09, "loss": 0.7411, "step": 837 }, { "epoch": 1.9764150943396226, "grad_norm": 0.7394147838620203, "learning_rate": 9.375219953450253e-10, "loss": 0.746, "step": 838 }, { "epoch": 1.9787735849056602, "grad_norm": 0.7690476142622722, "learning_rate": 7.748325822234392e-10, "loss": 0.729, "step": 839 }, { "epoch": 1.9811320754716981, "grad_norm": 0.9039884352435014, "learning_rate": 6.276297935149388e-10, "loss": 0.7569, "step": 840 }, { "epoch": 1.983490566037736, "grad_norm": 1.2814279255523484, "learning_rate": 4.959159106615596e-10, "loss": 0.7351, "step": 841 }, { "epoch": 1.9858490566037736, "grad_norm": 0.8004748840113347, "learning_rate": 3.7969297504858443e-10, "loss": 0.724, "step": 842 }, { "epoch": 1.9882075471698113, "grad_norm": 0.9234344448728469, "learning_rate": 2.789627879725698e-10, "loss": 0.749, "step": 843 }, { "epoch": 1.990566037735849, "grad_norm": 1.0793509856974828, "learning_rate": 1.9372691061381175e-10, "loss": 0.7296, "step": 844 }, { "epoch": 1.9929245283018868, "grad_norm": 0.8234018950018985, "learning_rate": 1.2398666401181035e-10, "loss": 0.738, "step": 845 }, { "epoch": 1.9952830188679245, "grad_norm": 0.7591071692242605, "learning_rate": 6.974312904517443e-11, "loss": 0.7713, "step": 846 }, { "epoch": 1.9976415094339623, "grad_norm": 0.8367565544655438, "learning_rate": 3.099714641452422e-11, "loss": 0.7812, "step": 847 }, { "epoch": 2.0, "grad_norm": 0.8006380089754, "learning_rate": 7.749316629612756e-12, "loss": 0.769, "step": 848 }, { "epoch": 2.0, "eval_loss": 0.6794618964195251, "eval_runtime": 82.3418, "eval_samples_per_second": 8.562, "eval_steps_per_second": 0.547, "step": 848 } ], "logging_steps": 1, "max_steps": 848, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 212, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 8.964690701765837e+17, "train_batch_size": 2, "trial_name": null, "trial_params": null }