{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.0, "eval_steps": 500, "global_step": 1284, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.001557632398753894, "grad_norm": 1.7232097279069922, "learning_rate": 9.999985033870806e-06, "loss": 0.6355, "step": 1 }, { "epoch": 0.003115264797507788, "grad_norm": 0.7663661005911214, "learning_rate": 9.999940135572817e-06, "loss": 0.2544, "step": 2 }, { "epoch": 0.004672897196261682, "grad_norm": 0.8839734168626316, "learning_rate": 9.999865305374812e-06, "loss": 0.3767, "step": 3 }, { "epoch": 0.006230529595015576, "grad_norm": 0.8042784289646617, "learning_rate": 9.999760543724761e-06, "loss": 0.184, "step": 4 }, { "epoch": 0.00778816199376947, "grad_norm": 0.6765931399234855, "learning_rate": 9.999625851249816e-06, "loss": 0.2941, "step": 5 }, { "epoch": 0.009345794392523364, "grad_norm": 0.7193614710776666, "learning_rate": 9.999461228756304e-06, "loss": 0.2705, "step": 6 }, { "epoch": 0.010903426791277258, "grad_norm": 0.7344615984114217, "learning_rate": 9.99926667722973e-06, "loss": 0.3259, "step": 7 }, { "epoch": 0.012461059190031152, "grad_norm": 0.73017739071341, "learning_rate": 9.999042197834769e-06, "loss": 0.2852, "step": 8 }, { "epoch": 0.014018691588785047, "grad_norm": 0.592507619306345, "learning_rate": 9.998787791915254e-06, "loss": 0.2649, "step": 9 }, { "epoch": 0.01557632398753894, "grad_norm": 0.9399500241178238, "learning_rate": 9.998503460994176e-06, "loss": 0.2521, "step": 10 }, { "epoch": 0.017133956386292833, "grad_norm": 0.7765731965837874, "learning_rate": 9.998189206773665e-06, "loss": 0.3556, "step": 11 }, { "epoch": 0.018691588785046728, "grad_norm": 0.6693444621163488, "learning_rate": 9.997845031134992e-06, "loss": 0.2239, "step": 12 }, { "epoch": 0.020249221183800622, "grad_norm": 1.4936804808243616, "learning_rate": 9.997470936138547e-06, "loss": 0.8968, "step": 13 }, { "epoch": 0.021806853582554516, "grad_norm": 0.7687864170878725, "learning_rate": 9.997066924023832e-06, "loss": 0.294, "step": 14 }, { "epoch": 0.02336448598130841, "grad_norm": 1.0982871522503537, "learning_rate": 9.996632997209444e-06, "loss": 0.3007, "step": 15 }, { "epoch": 0.024922118380062305, "grad_norm": 0.9371896720117513, "learning_rate": 9.996169158293066e-06, "loss": 0.2438, "step": 16 }, { "epoch": 0.0264797507788162, "grad_norm": 0.7798940457815309, "learning_rate": 9.995675410051448e-06, "loss": 0.2212, "step": 17 }, { "epoch": 0.028037383177570093, "grad_norm": 0.7355604168049059, "learning_rate": 9.99515175544039e-06, "loss": 0.2793, "step": 18 }, { "epoch": 0.029595015576323987, "grad_norm": 0.8219475835131775, "learning_rate": 9.994598197594723e-06, "loss": 0.2868, "step": 19 }, { "epoch": 0.03115264797507788, "grad_norm": 0.8165917677127098, "learning_rate": 9.994014739828298e-06, "loss": 0.2703, "step": 20 }, { "epoch": 0.03271028037383177, "grad_norm": 0.733767269375867, "learning_rate": 9.993401385633951e-06, "loss": 0.2059, "step": 21 }, { "epoch": 0.03426791277258567, "grad_norm": 1.2694949029078932, "learning_rate": 9.992758138683502e-06, "loss": 0.9151, "step": 22 }, { "epoch": 0.03582554517133956, "grad_norm": 0.7777986747788719, "learning_rate": 9.992085002827719e-06, "loss": 0.2599, "step": 23 }, { "epoch": 0.037383177570093455, "grad_norm": 0.610655735988334, "learning_rate": 9.991381982096293e-06, "loss": 0.2347, "step": 24 }, { "epoch": 0.03894080996884735, "grad_norm": 0.8562287986067203, "learning_rate": 9.990649080697825e-06, "loss": 0.2359, "step": 25 }, { "epoch": 0.040498442367601244, "grad_norm": 0.6673172477309072, "learning_rate": 9.989886303019793e-06, "loss": 0.2271, "step": 26 }, { "epoch": 0.04205607476635514, "grad_norm": 0.8629676432514921, "learning_rate": 9.989093653628532e-06, "loss": 0.2304, "step": 27 }, { "epoch": 0.04361370716510903, "grad_norm": 0.6898071747588274, "learning_rate": 9.988271137269197e-06, "loss": 0.3192, "step": 28 }, { "epoch": 0.045171339563862926, "grad_norm": 0.5010462374255644, "learning_rate": 9.987418758865743e-06, "loss": 0.2406, "step": 29 }, { "epoch": 0.04672897196261682, "grad_norm": 0.8259320779398918, "learning_rate": 9.98653652352089e-06, "loss": 0.3266, "step": 30 }, { "epoch": 0.048286604361370715, "grad_norm": 0.6076733459531145, "learning_rate": 9.9856244365161e-06, "loss": 0.2997, "step": 31 }, { "epoch": 0.04984423676012461, "grad_norm": 0.7298955355144913, "learning_rate": 9.984682503311536e-06, "loss": 0.2967, "step": 32 }, { "epoch": 0.0514018691588785, "grad_norm": 0.7346554751645953, "learning_rate": 9.983710729546038e-06, "loss": 0.3105, "step": 33 }, { "epoch": 0.0529595015576324, "grad_norm": 0.668130630437843, "learning_rate": 9.98270912103708e-06, "loss": 0.3535, "step": 34 }, { "epoch": 0.05451713395638629, "grad_norm": 0.6379634068346577, "learning_rate": 9.981677683780743e-06, "loss": 0.3709, "step": 35 }, { "epoch": 0.056074766355140186, "grad_norm": 0.798860986544702, "learning_rate": 9.98061642395168e-06, "loss": 0.2243, "step": 36 }, { "epoch": 0.05763239875389408, "grad_norm": 0.6593735789582216, "learning_rate": 9.979525347903067e-06, "loss": 0.2541, "step": 37 }, { "epoch": 0.059190031152647975, "grad_norm": 0.7224987244711324, "learning_rate": 9.978404462166582e-06, "loss": 0.2736, "step": 38 }, { "epoch": 0.06074766355140187, "grad_norm": 0.6440291059790796, "learning_rate": 9.977253773452349e-06, "loss": 0.2764, "step": 39 }, { "epoch": 0.06230529595015576, "grad_norm": 0.6222848807943162, "learning_rate": 9.976073288648913e-06, "loss": 0.2698, "step": 40 }, { "epoch": 0.06386292834890965, "grad_norm": 0.7394026438854199, "learning_rate": 9.97486301482319e-06, "loss": 0.2522, "step": 41 }, { "epoch": 0.06542056074766354, "grad_norm": 0.7520093536738988, "learning_rate": 9.973622959220426e-06, "loss": 0.2137, "step": 42 }, { "epoch": 0.06697819314641744, "grad_norm": 0.744961609793494, "learning_rate": 9.97235312926415e-06, "loss": 0.2328, "step": 43 }, { "epoch": 0.06853582554517133, "grad_norm": 0.7656967225499874, "learning_rate": 9.971053532556143e-06, "loss": 0.2669, "step": 44 }, { "epoch": 0.07009345794392523, "grad_norm": 0.6382236962718162, "learning_rate": 9.969724176876373e-06, "loss": 0.2192, "step": 45 }, { "epoch": 0.07165109034267912, "grad_norm": 0.6472374848660734, "learning_rate": 9.968365070182967e-06, "loss": 0.3144, "step": 46 }, { "epoch": 0.07320872274143302, "grad_norm": 0.6391395623819942, "learning_rate": 9.966976220612151e-06, "loss": 0.2048, "step": 47 }, { "epoch": 0.07476635514018691, "grad_norm": 0.7757553970772143, "learning_rate": 9.965557636478203e-06, "loss": 0.2579, "step": 48 }, { "epoch": 0.0763239875389408, "grad_norm": 0.8133958395260938, "learning_rate": 9.964109326273411e-06, "loss": 0.3226, "step": 49 }, { "epoch": 0.0778816199376947, "grad_norm": 0.4637065889612604, "learning_rate": 9.962631298668015e-06, "loss": 0.185, "step": 50 }, { "epoch": 0.0794392523364486, "grad_norm": 0.6273381176066575, "learning_rate": 9.961123562510153e-06, "loss": 0.3509, "step": 51 }, { "epoch": 0.08099688473520249, "grad_norm": 0.6419234115207709, "learning_rate": 9.959586126825818e-06, "loss": 0.2806, "step": 52 }, { "epoch": 0.08255451713395638, "grad_norm": 0.7160133507575562, "learning_rate": 9.95801900081879e-06, "loss": 0.1838, "step": 53 }, { "epoch": 0.08411214953271028, "grad_norm": 0.7848165329694581, "learning_rate": 9.956422193870597e-06, "loss": 0.255, "step": 54 }, { "epoch": 0.08566978193146417, "grad_norm": 0.6040970248107416, "learning_rate": 9.954795715540447e-06, "loss": 0.2142, "step": 55 }, { "epoch": 0.08722741433021806, "grad_norm": 0.6115559611935754, "learning_rate": 9.953139575565169e-06, "loss": 0.2668, "step": 56 }, { "epoch": 0.08878504672897196, "grad_norm": 0.6298402847496053, "learning_rate": 9.95145378385917e-06, "loss": 0.2881, "step": 57 }, { "epoch": 0.09034267912772585, "grad_norm": 0.5535053280930566, "learning_rate": 9.949738350514358e-06, "loss": 0.2202, "step": 58 }, { "epoch": 0.09190031152647975, "grad_norm": 0.6352050541969153, "learning_rate": 9.947993285800093e-06, "loss": 0.3398, "step": 59 }, { "epoch": 0.09345794392523364, "grad_norm": 0.826557642193822, "learning_rate": 9.94621860016312e-06, "loss": 0.3364, "step": 60 }, { "epoch": 0.09501557632398754, "grad_norm": 0.7298038729692221, "learning_rate": 9.944414304227508e-06, "loss": 0.2634, "step": 61 }, { "epoch": 0.09657320872274143, "grad_norm": 1.1507218720730263, "learning_rate": 9.94258040879459e-06, "loss": 0.8284, "step": 62 }, { "epoch": 0.09813084112149532, "grad_norm": 0.6885391578463873, "learning_rate": 9.94071692484289e-06, "loss": 0.2336, "step": 63 }, { "epoch": 0.09968847352024922, "grad_norm": 0.7334194656077329, "learning_rate": 9.938823863528065e-06, "loss": 0.2652, "step": 64 }, { "epoch": 0.10124610591900311, "grad_norm": 1.0514064266169973, "learning_rate": 9.936901236182836e-06, "loss": 0.835, "step": 65 }, { "epoch": 0.102803738317757, "grad_norm": 0.6334750887159478, "learning_rate": 9.934949054316917e-06, "loss": 0.2812, "step": 66 }, { "epoch": 0.1043613707165109, "grad_norm": 0.4857684609448803, "learning_rate": 9.932967329616953e-06, "loss": 0.202, "step": 67 }, { "epoch": 0.1059190031152648, "grad_norm": 0.690411540980916, "learning_rate": 9.930956073946442e-06, "loss": 0.2828, "step": 68 }, { "epoch": 0.10747663551401869, "grad_norm": 0.6862713871911896, "learning_rate": 9.928915299345669e-06, "loss": 0.3246, "step": 69 }, { "epoch": 0.10903426791277258, "grad_norm": 0.6617081291471549, "learning_rate": 9.926845018031632e-06, "loss": 0.2621, "step": 70 }, { "epoch": 0.11059190031152648, "grad_norm": 0.5745643134324515, "learning_rate": 9.924745242397968e-06, "loss": 0.2224, "step": 71 }, { "epoch": 0.11214953271028037, "grad_norm": 0.7300598195282793, "learning_rate": 9.922615985014887e-06, "loss": 0.2297, "step": 72 }, { "epoch": 0.11370716510903427, "grad_norm": 0.6648229017872759, "learning_rate": 9.920457258629081e-06, "loss": 0.2461, "step": 73 }, { "epoch": 0.11526479750778816, "grad_norm": 0.6868885744716628, "learning_rate": 9.918269076163664e-06, "loss": 0.2728, "step": 74 }, { "epoch": 0.11682242990654206, "grad_norm": 0.6982895714335172, "learning_rate": 9.916051450718085e-06, "loss": 0.257, "step": 75 }, { "epoch": 0.11838006230529595, "grad_norm": 0.5677598559231926, "learning_rate": 9.91380439556805e-06, "loss": 0.2614, "step": 76 }, { "epoch": 0.11993769470404984, "grad_norm": 0.5071293538430349, "learning_rate": 9.911527924165445e-06, "loss": 0.2793, "step": 77 }, { "epoch": 0.12149532710280374, "grad_norm": 0.7281461877976642, "learning_rate": 9.909222050138259e-06, "loss": 0.2861, "step": 78 }, { "epoch": 0.12305295950155763, "grad_norm": 0.6491525810286177, "learning_rate": 9.906886787290492e-06, "loss": 0.2799, "step": 79 }, { "epoch": 0.12461059190031153, "grad_norm": 0.9626139969649712, "learning_rate": 9.904522149602087e-06, "loss": 0.2251, "step": 80 }, { "epoch": 0.1261682242990654, "grad_norm": 0.705551656345552, "learning_rate": 9.902128151228827e-06, "loss": 0.2899, "step": 81 }, { "epoch": 0.1277258566978193, "grad_norm": 0.5796094318767273, "learning_rate": 9.899704806502272e-06, "loss": 0.3023, "step": 82 }, { "epoch": 0.1292834890965732, "grad_norm": 0.5778885783153778, "learning_rate": 9.89725212992966e-06, "loss": 0.2596, "step": 83 }, { "epoch": 0.1308411214953271, "grad_norm": 0.6407326481544323, "learning_rate": 9.894770136193814e-06, "loss": 0.3182, "step": 84 }, { "epoch": 0.13239875389408098, "grad_norm": 0.8761396102635524, "learning_rate": 9.892258840153075e-06, "loss": 0.3261, "step": 85 }, { "epoch": 0.13395638629283488, "grad_norm": 0.8248387465515045, "learning_rate": 9.889718256841195e-06, "loss": 0.3105, "step": 86 }, { "epoch": 0.13551401869158877, "grad_norm": 2.168823178408131, "learning_rate": 9.88714840146725e-06, "loss": 0.7869, "step": 87 }, { "epoch": 0.13707165109034267, "grad_norm": 0.6163273279750748, "learning_rate": 9.884549289415556e-06, "loss": 0.2249, "step": 88 }, { "epoch": 0.13862928348909656, "grad_norm": 1.5331833228458356, "learning_rate": 9.881920936245577e-06, "loss": 0.7778, "step": 89 }, { "epoch": 0.14018691588785046, "grad_norm": 1.5428867178521768, "learning_rate": 9.879263357691815e-06, "loss": 0.8951, "step": 90 }, { "epoch": 0.14174454828660435, "grad_norm": 0.6177160477898084, "learning_rate": 9.876576569663739e-06, "loss": 0.224, "step": 91 }, { "epoch": 0.14330218068535824, "grad_norm": 0.7981644114495157, "learning_rate": 9.873860588245675e-06, "loss": 0.2892, "step": 92 }, { "epoch": 0.14485981308411214, "grad_norm": 0.600801453450908, "learning_rate": 9.871115429696716e-06, "loss": 0.2782, "step": 93 }, { "epoch": 0.14641744548286603, "grad_norm": 0.6547888169420532, "learning_rate": 9.868341110450618e-06, "loss": 0.3347, "step": 94 }, { "epoch": 0.14797507788161993, "grad_norm": 0.5823853177715245, "learning_rate": 9.865537647115713e-06, "loss": 0.2778, "step": 95 }, { "epoch": 0.14953271028037382, "grad_norm": 0.9116578300075924, "learning_rate": 9.862705056474795e-06, "loss": 0.3503, "step": 96 }, { "epoch": 0.15109034267912771, "grad_norm": 0.5426782446240797, "learning_rate": 9.859843355485033e-06, "loss": 0.2386, "step": 97 }, { "epoch": 0.1526479750778816, "grad_norm": 0.5543482391953797, "learning_rate": 9.856952561277862e-06, "loss": 0.2344, "step": 98 }, { "epoch": 0.1542056074766355, "grad_norm": 0.7660668413572433, "learning_rate": 9.854032691158881e-06, "loss": 0.2692, "step": 99 }, { "epoch": 0.1557632398753894, "grad_norm": 0.7745787606896338, "learning_rate": 9.851083762607752e-06, "loss": 0.2374, "step": 100 }, { "epoch": 0.1573208722741433, "grad_norm": 0.8283533162770704, "learning_rate": 9.848105793278092e-06, "loss": 0.2699, "step": 101 }, { "epoch": 0.1588785046728972, "grad_norm": 0.5323847631668214, "learning_rate": 9.84509880099737e-06, "loss": 0.2177, "step": 102 }, { "epoch": 0.16043613707165108, "grad_norm": 0.7022062850923283, "learning_rate": 9.842062803766804e-06, "loss": 0.2258, "step": 103 }, { "epoch": 0.16199376947040497, "grad_norm": 1.9739756098890826, "learning_rate": 9.838997819761241e-06, "loss": 0.9263, "step": 104 }, { "epoch": 0.16355140186915887, "grad_norm": 0.7651110334163559, "learning_rate": 9.835903867329061e-06, "loss": 0.2654, "step": 105 }, { "epoch": 0.16510903426791276, "grad_norm": 0.6634416278579489, "learning_rate": 9.832780964992059e-06, "loss": 0.2959, "step": 106 }, { "epoch": 0.16666666666666666, "grad_norm": 0.62947915393871, "learning_rate": 9.829629131445342e-06, "loss": 0.2608, "step": 107 }, { "epoch": 0.16822429906542055, "grad_norm": 1.55385845539561, "learning_rate": 9.826448385557208e-06, "loss": 0.842, "step": 108 }, { "epoch": 0.16978193146417445, "grad_norm": 0.6067362647236724, "learning_rate": 9.823238746369038e-06, "loss": 0.3388, "step": 109 }, { "epoch": 0.17133956386292834, "grad_norm": 0.7095716481506402, "learning_rate": 9.820000233095179e-06, "loss": 0.2953, "step": 110 }, { "epoch": 0.17289719626168223, "grad_norm": 1.4374378842104667, "learning_rate": 9.81673286512284e-06, "loss": 0.7775, "step": 111 }, { "epoch": 0.17445482866043613, "grad_norm": 0.6761770106689733, "learning_rate": 9.813436662011958e-06, "loss": 0.2929, "step": 112 }, { "epoch": 0.17601246105919002, "grad_norm": 0.7954842040051252, "learning_rate": 9.810111643495095e-06, "loss": 0.2734, "step": 113 }, { "epoch": 0.17757009345794392, "grad_norm": 1.271851983971207, "learning_rate": 9.806757829477313e-06, "loss": 0.7987, "step": 114 }, { "epoch": 0.1791277258566978, "grad_norm": 0.526209190777812, "learning_rate": 9.803375240036057e-06, "loss": 0.3054, "step": 115 }, { "epoch": 0.1806853582554517, "grad_norm": 0.6440505513615228, "learning_rate": 9.799963895421036e-06, "loss": 0.2416, "step": 116 }, { "epoch": 0.1822429906542056, "grad_norm": 0.9367999131179932, "learning_rate": 9.7965238160541e-06, "loss": 0.3409, "step": 117 }, { "epoch": 0.1838006230529595, "grad_norm": 0.8415567444030029, "learning_rate": 9.793055022529116e-06, "loss": 0.2847, "step": 118 }, { "epoch": 0.1853582554517134, "grad_norm": 0.5786248320276434, "learning_rate": 9.789557535611853e-06, "loss": 0.2644, "step": 119 }, { "epoch": 0.18691588785046728, "grad_norm": 0.687463433264138, "learning_rate": 9.786031376239842e-06, "loss": 0.2569, "step": 120 }, { "epoch": 0.18847352024922118, "grad_norm": 0.45347828103414345, "learning_rate": 9.78247656552227e-06, "loss": 0.2415, "step": 121 }, { "epoch": 0.19003115264797507, "grad_norm": 0.5650726061282849, "learning_rate": 9.778893124739836e-06, "loss": 0.2547, "step": 122 }, { "epoch": 0.19158878504672897, "grad_norm": 0.5103395837306397, "learning_rate": 9.775281075344639e-06, "loss": 0.3065, "step": 123 }, { "epoch": 0.19314641744548286, "grad_norm": 0.5183636563691485, "learning_rate": 9.771640438960037e-06, "loss": 0.2922, "step": 124 }, { "epoch": 0.19470404984423675, "grad_norm": 0.5026042818767394, "learning_rate": 9.76797123738052e-06, "loss": 0.2588, "step": 125 }, { "epoch": 0.19626168224299065, "grad_norm": 0.6794540572358105, "learning_rate": 9.76427349257159e-06, "loss": 0.3041, "step": 126 }, { "epoch": 0.19781931464174454, "grad_norm": 0.6283325050944261, "learning_rate": 9.76054722666962e-06, "loss": 0.2991, "step": 127 }, { "epoch": 0.19937694704049844, "grad_norm": 0.7184815540245492, "learning_rate": 9.756792461981713e-06, "loss": 0.2599, "step": 128 }, { "epoch": 0.20093457943925233, "grad_norm": 0.701839843872524, "learning_rate": 9.753009220985593e-06, "loss": 0.2711, "step": 129 }, { "epoch": 0.20249221183800623, "grad_norm": 0.6505643462960924, "learning_rate": 9.749197526329446e-06, "loss": 0.2947, "step": 130 }, { "epoch": 0.20404984423676012, "grad_norm": 0.6104336914925119, "learning_rate": 9.745357400831799e-06, "loss": 0.2331, "step": 131 }, { "epoch": 0.205607476635514, "grad_norm": 0.908932761888463, "learning_rate": 9.741488867481377e-06, "loss": 0.2377, "step": 132 }, { "epoch": 0.2071651090342679, "grad_norm": 0.5980236644974856, "learning_rate": 9.737591949436969e-06, "loss": 0.1601, "step": 133 }, { "epoch": 0.2087227414330218, "grad_norm": 0.5937884904455855, "learning_rate": 9.733666670027288e-06, "loss": 0.2542, "step": 134 }, { "epoch": 0.2102803738317757, "grad_norm": 0.7666646448496817, "learning_rate": 9.729713052750827e-06, "loss": 0.2465, "step": 135 }, { "epoch": 0.2118380062305296, "grad_norm": 0.5386370084617528, "learning_rate": 9.725731121275725e-06, "loss": 0.3172, "step": 136 }, { "epoch": 0.21339563862928349, "grad_norm": 0.5517332858518383, "learning_rate": 9.721720899439621e-06, "loss": 0.2864, "step": 137 }, { "epoch": 0.21495327102803738, "grad_norm": 0.60004714901346, "learning_rate": 9.71768241124952e-06, "loss": 0.3101, "step": 138 }, { "epoch": 0.21651090342679127, "grad_norm": 0.7525697267510544, "learning_rate": 9.71361568088163e-06, "loss": 0.2235, "step": 139 }, { "epoch": 0.21806853582554517, "grad_norm": 0.5929490381364206, "learning_rate": 9.709520732681238e-06, "loss": 0.231, "step": 140 }, { "epoch": 0.21962616822429906, "grad_norm": 0.7027523851155623, "learning_rate": 9.705397591162556e-06, "loss": 0.2379, "step": 141 }, { "epoch": 0.22118380062305296, "grad_norm": 0.6504000514836087, "learning_rate": 9.701246281008568e-06, "loss": 0.2462, "step": 142 }, { "epoch": 0.22274143302180685, "grad_norm": 0.5770728592720746, "learning_rate": 9.697066827070894e-06, "loss": 0.2859, "step": 143 }, { "epoch": 0.22429906542056074, "grad_norm": 0.5798579381876525, "learning_rate": 9.692859254369631e-06, "loss": 0.2663, "step": 144 }, { "epoch": 0.22585669781931464, "grad_norm": 0.5033709259076921, "learning_rate": 9.68862358809321e-06, "loss": 0.1913, "step": 145 }, { "epoch": 0.22741433021806853, "grad_norm": 0.5493196423085126, "learning_rate": 9.684359853598245e-06, "loss": 0.2197, "step": 146 }, { "epoch": 0.22897196261682243, "grad_norm": 0.5652921319820055, "learning_rate": 9.680068076409373e-06, "loss": 0.2601, "step": 147 }, { "epoch": 0.23052959501557632, "grad_norm": 0.6783593909741826, "learning_rate": 9.675748282219114e-06, "loss": 0.2922, "step": 148 }, { "epoch": 0.23208722741433022, "grad_norm": 0.7088765017570264, "learning_rate": 9.671400496887704e-06, "loss": 0.2997, "step": 149 }, { "epoch": 0.2336448598130841, "grad_norm": 0.6064540912244738, "learning_rate": 9.667024746442953e-06, "loss": 0.3283, "step": 150 }, { "epoch": 0.235202492211838, "grad_norm": 0.6900794118386824, "learning_rate": 9.662621057080077e-06, "loss": 0.2447, "step": 151 }, { "epoch": 0.2367601246105919, "grad_norm": 0.8898873494327856, "learning_rate": 9.65818945516155e-06, "loss": 0.2857, "step": 152 }, { "epoch": 0.2383177570093458, "grad_norm": 0.7596388046727889, "learning_rate": 9.653729967216944e-06, "loss": 0.2765, "step": 153 }, { "epoch": 0.2398753894080997, "grad_norm": 0.5897245550654492, "learning_rate": 9.64924261994277e-06, "loss": 0.2629, "step": 154 }, { "epoch": 0.24143302180685358, "grad_norm": 0.6261944121589039, "learning_rate": 9.644727440202308e-06, "loss": 0.3554, "step": 155 }, { "epoch": 0.24299065420560748, "grad_norm": 0.7471727372694826, "learning_rate": 9.640184455025472e-06, "loss": 0.3595, "step": 156 }, { "epoch": 0.24454828660436137, "grad_norm": 0.6539521944383881, "learning_rate": 9.635613691608619e-06, "loss": 0.2351, "step": 157 }, { "epoch": 0.24610591900311526, "grad_norm": 0.7372613573398835, "learning_rate": 9.631015177314402e-06, "loss": 0.3061, "step": 158 }, { "epoch": 0.24766355140186916, "grad_norm": 0.802899194753761, "learning_rate": 9.62638893967161e-06, "loss": 0.3606, "step": 159 }, { "epoch": 0.24922118380062305, "grad_norm": 0.754935748177354, "learning_rate": 9.621735006374984e-06, "loss": 0.223, "step": 160 }, { "epoch": 0.2507788161993769, "grad_norm": 0.4173684214144698, "learning_rate": 9.617053405285077e-06, "loss": 0.2319, "step": 161 }, { "epoch": 0.2523364485981308, "grad_norm": 0.6435417325636267, "learning_rate": 9.612344164428063e-06, "loss": 0.3996, "step": 162 }, { "epoch": 0.2538940809968847, "grad_norm": 0.8830545123878781, "learning_rate": 9.607607311995587e-06, "loss": 0.3724, "step": 163 }, { "epoch": 0.2554517133956386, "grad_norm": 2.2533700279001554, "learning_rate": 9.602842876344589e-06, "loss": 0.8261, "step": 164 }, { "epoch": 0.2570093457943925, "grad_norm": 0.6641726752237889, "learning_rate": 9.59805088599713e-06, "loss": 0.3339, "step": 165 }, { "epoch": 0.2585669781931464, "grad_norm": 0.6276989828906703, "learning_rate": 9.59323136964023e-06, "loss": 0.219, "step": 166 }, { "epoch": 0.2601246105919003, "grad_norm": 0.8210416456426778, "learning_rate": 9.588384356125691e-06, "loss": 0.282, "step": 167 }, { "epoch": 0.2616822429906542, "grad_norm": 0.751594963703605, "learning_rate": 9.583509874469924e-06, "loss": 0.3373, "step": 168 }, { "epoch": 0.2632398753894081, "grad_norm": 0.7261273311893066, "learning_rate": 9.578607953853777e-06, "loss": 0.28, "step": 169 }, { "epoch": 0.26479750778816197, "grad_norm": 0.6922505407762392, "learning_rate": 9.573678623622364e-06, "loss": 0.3412, "step": 170 }, { "epoch": 0.26635514018691586, "grad_norm": 0.6419868376151807, "learning_rate": 9.568721913284879e-06, "loss": 0.2588, "step": 171 }, { "epoch": 0.26791277258566976, "grad_norm": 0.49324052267306684, "learning_rate": 9.563737852514432e-06, "loss": 0.2332, "step": 172 }, { "epoch": 0.26947040498442365, "grad_norm": 0.5770500149070207, "learning_rate": 9.558726471147862e-06, "loss": 0.2937, "step": 173 }, { "epoch": 0.27102803738317754, "grad_norm": 0.7329648250931949, "learning_rate": 9.553687799185556e-06, "loss": 0.3309, "step": 174 }, { "epoch": 0.27258566978193144, "grad_norm": 0.4864393923841514, "learning_rate": 9.548621866791286e-06, "loss": 0.2043, "step": 175 }, { "epoch": 0.27414330218068533, "grad_norm": 1.136074014302222, "learning_rate": 9.54352870429201e-06, "loss": 0.7893, "step": 176 }, { "epoch": 0.2757009345794392, "grad_norm": 0.6443586968700755, "learning_rate": 9.538408342177699e-06, "loss": 0.3025, "step": 177 }, { "epoch": 0.2772585669781931, "grad_norm": 0.6403852564269055, "learning_rate": 9.533260811101152e-06, "loss": 0.2629, "step": 178 }, { "epoch": 0.278816199376947, "grad_norm": 0.6869290500301357, "learning_rate": 9.528086141877817e-06, "loss": 0.301, "step": 179 }, { "epoch": 0.2803738317757009, "grad_norm": 0.9338061428659953, "learning_rate": 9.522884365485599e-06, "loss": 0.7925, "step": 180 }, { "epoch": 0.2819314641744548, "grad_norm": 0.7304512334819939, "learning_rate": 9.517655513064682e-06, "loss": 0.44, "step": 181 }, { "epoch": 0.2834890965732087, "grad_norm": 0.6207928517482323, "learning_rate": 9.512399615917342e-06, "loss": 0.228, "step": 182 }, { "epoch": 0.2850467289719626, "grad_norm": 0.8847617104487071, "learning_rate": 9.507116705507748e-06, "loss": 0.4137, "step": 183 }, { "epoch": 0.2866043613707165, "grad_norm": 0.6631070285038642, "learning_rate": 9.50180681346179e-06, "loss": 0.3438, "step": 184 }, { "epoch": 0.2881619937694704, "grad_norm": 1.0647904328891042, "learning_rate": 9.49646997156688e-06, "loss": 0.2452, "step": 185 }, { "epoch": 0.2897196261682243, "grad_norm": 0.4959301062995678, "learning_rate": 9.491106211771765e-06, "loss": 0.2915, "step": 186 }, { "epoch": 0.29127725856697817, "grad_norm": 0.705003606824088, "learning_rate": 9.485715566186333e-06, "loss": 0.2945, "step": 187 }, { "epoch": 0.29283489096573206, "grad_norm": 1.045178016891709, "learning_rate": 9.480298067081422e-06, "loss": 0.5356, "step": 188 }, { "epoch": 0.29439252336448596, "grad_norm": 0.6198122538394608, "learning_rate": 9.474853746888631e-06, "loss": 0.2744, "step": 189 }, { "epoch": 0.29595015576323985, "grad_norm": 0.631725313682397, "learning_rate": 9.469382638200119e-06, "loss": 0.3064, "step": 190 }, { "epoch": 0.29750778816199375, "grad_norm": 0.49405112326099454, "learning_rate": 9.463884773768413e-06, "loss": 0.2016, "step": 191 }, { "epoch": 0.29906542056074764, "grad_norm": 0.9255242318461628, "learning_rate": 9.458360186506212e-06, "loss": 0.318, "step": 192 }, { "epoch": 0.30062305295950154, "grad_norm": 0.5172541239349757, "learning_rate": 9.452808909486195e-06, "loss": 0.2182, "step": 193 }, { "epoch": 0.30218068535825543, "grad_norm": 0.6548225129226617, "learning_rate": 9.447230975940808e-06, "loss": 0.3501, "step": 194 }, { "epoch": 0.3037383177570093, "grad_norm": 0.6373263585258429, "learning_rate": 9.441626419262084e-06, "loss": 0.2331, "step": 195 }, { "epoch": 0.3052959501557632, "grad_norm": 0.8451103465765266, "learning_rate": 9.43599527300143e-06, "loss": 0.2379, "step": 196 }, { "epoch": 0.3068535825545171, "grad_norm": 0.6257655537331657, "learning_rate": 9.430337570869432e-06, "loss": 0.262, "step": 197 }, { "epoch": 0.308411214953271, "grad_norm": 0.7885047351755374, "learning_rate": 9.424653346735649e-06, "loss": 0.2783, "step": 198 }, { "epoch": 0.3099688473520249, "grad_norm": 1.2716968661440529, "learning_rate": 9.418942634628413e-06, "loss": 0.3266, "step": 199 }, { "epoch": 0.3115264797507788, "grad_norm": 0.709734211796566, "learning_rate": 9.413205468734628e-06, "loss": 0.298, "step": 200 }, { "epoch": 0.3130841121495327, "grad_norm": 0.6966111704003821, "learning_rate": 9.40744188339956e-06, "loss": 0.3872, "step": 201 }, { "epoch": 0.3146417445482866, "grad_norm": 0.6328947624746845, "learning_rate": 9.401651913126634e-06, "loss": 0.2855, "step": 202 }, { "epoch": 0.3161993769470405, "grad_norm": 0.6782499058815041, "learning_rate": 9.395835592577228e-06, "loss": 0.2555, "step": 203 }, { "epoch": 0.3177570093457944, "grad_norm": 0.6111146053952768, "learning_rate": 9.389992956570463e-06, "loss": 0.2119, "step": 204 }, { "epoch": 0.31931464174454827, "grad_norm": 2.1014154014060646, "learning_rate": 9.384124040082996e-06, "loss": 1.3393, "step": 205 }, { "epoch": 0.32087227414330216, "grad_norm": 0.6937660819701476, "learning_rate": 9.378228878248814e-06, "loss": 0.3003, "step": 206 }, { "epoch": 0.32242990654205606, "grad_norm": 0.7798606308310687, "learning_rate": 9.372307506359019e-06, "loss": 0.2394, "step": 207 }, { "epoch": 0.32398753894080995, "grad_norm": 0.8064220607757047, "learning_rate": 9.366359959861615e-06, "loss": 0.2457, "step": 208 }, { "epoch": 0.32554517133956384, "grad_norm": 0.8663909019847741, "learning_rate": 9.360386274361305e-06, "loss": 0.2363, "step": 209 }, { "epoch": 0.32710280373831774, "grad_norm": 0.7765065772183143, "learning_rate": 9.354386485619264e-06, "loss": 0.2519, "step": 210 }, { "epoch": 0.32866043613707163, "grad_norm": 0.6924828379419826, "learning_rate": 9.348360629552941e-06, "loss": 0.2237, "step": 211 }, { "epoch": 0.3302180685358255, "grad_norm": 0.6564177523605338, "learning_rate": 9.342308742235831e-06, "loss": 0.2369, "step": 212 }, { "epoch": 0.3317757009345794, "grad_norm": 0.6166208484170821, "learning_rate": 9.336230859897266e-06, "loss": 0.2531, "step": 213 }, { "epoch": 0.3333333333333333, "grad_norm": 0.6325656097098429, "learning_rate": 9.330127018922195e-06, "loss": 0.228, "step": 214 }, { "epoch": 0.3348909657320872, "grad_norm": 0.9005060609426595, "learning_rate": 9.323997255850965e-06, "loss": 0.2068, "step": 215 }, { "epoch": 0.3364485981308411, "grad_norm": 0.6030970162930338, "learning_rate": 9.317841607379106e-06, "loss": 0.2269, "step": 216 }, { "epoch": 0.338006230529595, "grad_norm": 0.5757510215326483, "learning_rate": 9.311660110357116e-06, "loss": 0.2277, "step": 217 }, { "epoch": 0.3395638629283489, "grad_norm": 0.48332154427815277, "learning_rate": 9.30545280179022e-06, "loss": 0.2088, "step": 218 }, { "epoch": 0.3411214953271028, "grad_norm": 1.6587632436476218, "learning_rate": 9.299219718838174e-06, "loss": 0.8341, "step": 219 }, { "epoch": 0.3426791277258567, "grad_norm": 0.9279443473544912, "learning_rate": 9.292960898815032e-06, "loss": 0.2415, "step": 220 }, { "epoch": 0.3442367601246106, "grad_norm": 1.6924178170865158, "learning_rate": 9.286676379188913e-06, "loss": 0.9856, "step": 221 }, { "epoch": 0.34579439252336447, "grad_norm": 1.8214244233005599, "learning_rate": 9.280366197581792e-06, "loss": 0.826, "step": 222 }, { "epoch": 0.34735202492211836, "grad_norm": 0.783390501917877, "learning_rate": 9.274030391769264e-06, "loss": 0.8049, "step": 223 }, { "epoch": 0.34890965732087226, "grad_norm": 0.5859949211179467, "learning_rate": 9.267668999680328e-06, "loss": 0.2939, "step": 224 }, { "epoch": 0.35046728971962615, "grad_norm": 0.7610274820950346, "learning_rate": 9.261282059397145e-06, "loss": 0.2581, "step": 225 }, { "epoch": 0.35202492211838005, "grad_norm": 0.6066786542121007, "learning_rate": 9.25486960915483e-06, "loss": 0.2758, "step": 226 }, { "epoch": 0.35358255451713394, "grad_norm": 1.1073512044256302, "learning_rate": 9.248431687341203e-06, "loss": 0.237, "step": 227 }, { "epoch": 0.35514018691588783, "grad_norm": 1.0141799395602682, "learning_rate": 9.241968332496576e-06, "loss": 0.2568, "step": 228 }, { "epoch": 0.35669781931464173, "grad_norm": 0.5568072088382595, "learning_rate": 9.235479583313504e-06, "loss": 0.2334, "step": 229 }, { "epoch": 0.3582554517133956, "grad_norm": 0.7930065901392893, "learning_rate": 9.228965478636575e-06, "loss": 0.2793, "step": 230 }, { "epoch": 0.3598130841121495, "grad_norm": 0.7048596889355911, "learning_rate": 9.222426057462162e-06, "loss": 0.3062, "step": 231 }, { "epoch": 0.3613707165109034, "grad_norm": 0.8914753088496622, "learning_rate": 9.215861358938191e-06, "loss": 0.3289, "step": 232 }, { "epoch": 0.3629283489096573, "grad_norm": 0.8233950771824091, "learning_rate": 9.209271422363915e-06, "loss": 0.3021, "step": 233 }, { "epoch": 0.3644859813084112, "grad_norm": 0.8572614177235147, "learning_rate": 9.20265628718967e-06, "loss": 0.3435, "step": 234 }, { "epoch": 0.3660436137071651, "grad_norm": 0.6533204687228193, "learning_rate": 9.196015993016642e-06, "loss": 0.2537, "step": 235 }, { "epoch": 0.367601246105919, "grad_norm": 0.676657852907951, "learning_rate": 9.189350579596635e-06, "loss": 0.273, "step": 236 }, { "epoch": 0.3691588785046729, "grad_norm": 0.8090061594796749, "learning_rate": 9.182660086831819e-06, "loss": 0.3786, "step": 237 }, { "epoch": 0.3707165109034268, "grad_norm": 0.6195251333212162, "learning_rate": 9.175944554774507e-06, "loss": 0.2799, "step": 238 }, { "epoch": 0.37227414330218067, "grad_norm": 0.5501170288605118, "learning_rate": 9.169204023626911e-06, "loss": 0.2577, "step": 239 }, { "epoch": 0.37383177570093457, "grad_norm": 0.6087726744520703, "learning_rate": 9.162438533740891e-06, "loss": 0.2578, "step": 240 }, { "epoch": 0.37538940809968846, "grad_norm": 0.6423864822269919, "learning_rate": 9.15564812561773e-06, "loss": 0.2865, "step": 241 }, { "epoch": 0.37694704049844235, "grad_norm": 0.6443057475504381, "learning_rate": 9.148832839907871e-06, "loss": 0.2898, "step": 242 }, { "epoch": 0.37850467289719625, "grad_norm": 0.5688149258379418, "learning_rate": 9.141992717410697e-06, "loss": 0.239, "step": 243 }, { "epoch": 0.38006230529595014, "grad_norm": 0.6629923621745606, "learning_rate": 9.135127799074271e-06, "loss": 0.2614, "step": 244 }, { "epoch": 0.38161993769470404, "grad_norm": 0.6300463035343699, "learning_rate": 9.128238125995093e-06, "loss": 0.24, "step": 245 }, { "epoch": 0.38317757009345793, "grad_norm": 0.643027164394109, "learning_rate": 9.121323739417858e-06, "loss": 0.2365, "step": 246 }, { "epoch": 0.3847352024922118, "grad_norm": 2.8455942968073282, "learning_rate": 9.114384680735211e-06, "loss": 0.7814, "step": 247 }, { "epoch": 0.3862928348909657, "grad_norm": 0.630104563255545, "learning_rate": 9.107420991487488e-06, "loss": 0.2234, "step": 248 }, { "epoch": 0.3878504672897196, "grad_norm": 0.905027542232206, "learning_rate": 9.100432713362477e-06, "loss": 0.3142, "step": 249 }, { "epoch": 0.3894080996884735, "grad_norm": 0.6960898261331736, "learning_rate": 9.09341988819517e-06, "loss": 0.3195, "step": 250 }, { "epoch": 0.3909657320872274, "grad_norm": 0.7350311907562319, "learning_rate": 9.086382557967507e-06, "loss": 0.2479, "step": 251 }, { "epoch": 0.3925233644859813, "grad_norm": 0.7212085527913289, "learning_rate": 9.07932076480812e-06, "loss": 0.2334, "step": 252 }, { "epoch": 0.3940809968847352, "grad_norm": 0.6340762358232386, "learning_rate": 9.072234550992099e-06, "loss": 0.282, "step": 253 }, { "epoch": 0.3956386292834891, "grad_norm": 0.5688729860657064, "learning_rate": 9.065123958940716e-06, "loss": 0.2928, "step": 254 }, { "epoch": 0.397196261682243, "grad_norm": 0.8922598360101682, "learning_rate": 9.057989031221187e-06, "loss": 0.281, "step": 255 }, { "epoch": 0.3987538940809969, "grad_norm": 0.7114134585553089, "learning_rate": 9.050829810546416e-06, "loss": 0.2635, "step": 256 }, { "epoch": 0.40031152647975077, "grad_norm": 0.6966492509810489, "learning_rate": 9.043646339774726e-06, "loss": 0.3394, "step": 257 }, { "epoch": 0.40186915887850466, "grad_norm": 0.5958512506208158, "learning_rate": 9.036438661909624e-06, "loss": 0.2718, "step": 258 }, { "epoch": 0.40342679127725856, "grad_norm": 0.9116531361266647, "learning_rate": 9.029206820099518e-06, "loss": 0.2785, "step": 259 }, { "epoch": 0.40498442367601245, "grad_norm": 0.5887944110529719, "learning_rate": 9.021950857637486e-06, "loss": 0.2668, "step": 260 }, { "epoch": 0.40654205607476634, "grad_norm": 0.6108632031180715, "learning_rate": 9.014670817960993e-06, "loss": 0.1969, "step": 261 }, { "epoch": 0.40809968847352024, "grad_norm": 0.6907915569936867, "learning_rate": 9.007366744651646e-06, "loss": 0.2268, "step": 262 }, { "epoch": 0.40965732087227413, "grad_norm": 0.5793225985517833, "learning_rate": 9.000038681434925e-06, "loss": 0.1931, "step": 263 }, { "epoch": 0.411214953271028, "grad_norm": 1.0784967522591073, "learning_rate": 8.99268667217993e-06, "loss": 0.2329, "step": 264 }, { "epoch": 0.4127725856697819, "grad_norm": 0.602627598917929, "learning_rate": 8.985310760899108e-06, "loss": 0.2574, "step": 265 }, { "epoch": 0.4143302180685358, "grad_norm": 0.6724448335681737, "learning_rate": 8.977910991747993e-06, "loss": 0.2781, "step": 266 }, { "epoch": 0.4158878504672897, "grad_norm": 0.713294514234468, "learning_rate": 8.970487409024949e-06, "loss": 0.2284, "step": 267 }, { "epoch": 0.4174454828660436, "grad_norm": 0.8427860833648895, "learning_rate": 8.96304005717089e-06, "loss": 0.2953, "step": 268 }, { "epoch": 0.4190031152647975, "grad_norm": 0.5840913639256132, "learning_rate": 8.955568980769033e-06, "loss": 0.2974, "step": 269 }, { "epoch": 0.4205607476635514, "grad_norm": 0.6131987426518342, "learning_rate": 8.948074224544615e-06, "loss": 0.2773, "step": 270 }, { "epoch": 0.4221183800623053, "grad_norm": 0.5947515631497642, "learning_rate": 8.94055583336463e-06, "loss": 0.275, "step": 271 }, { "epoch": 0.4236760124610592, "grad_norm": 0.5946955459161034, "learning_rate": 8.933013852237564e-06, "loss": 0.3307, "step": 272 }, { "epoch": 0.4252336448598131, "grad_norm": 0.7451920245650221, "learning_rate": 8.925448326313125e-06, "loss": 0.286, "step": 273 }, { "epoch": 0.42679127725856697, "grad_norm": 0.924856917547545, "learning_rate": 8.917859300881965e-06, "loss": 0.2879, "step": 274 }, { "epoch": 0.42834890965732086, "grad_norm": 1.0327909293428048, "learning_rate": 8.91024682137542e-06, "loss": 0.2408, "step": 275 }, { "epoch": 0.42990654205607476, "grad_norm": 0.880531058290941, "learning_rate": 8.90261093336523e-06, "loss": 0.3045, "step": 276 }, { "epoch": 0.43146417445482865, "grad_norm": 0.580559025082547, "learning_rate": 8.89495168256327e-06, "loss": 0.2805, "step": 277 }, { "epoch": 0.43302180685358255, "grad_norm": 0.5475610266945151, "learning_rate": 8.887269114821275e-06, "loss": 0.2338, "step": 278 }, { "epoch": 0.43457943925233644, "grad_norm": 0.6043881248930308, "learning_rate": 8.879563276130567e-06, "loss": 0.2516, "step": 279 }, { "epoch": 0.43613707165109034, "grad_norm": 0.6621273162261064, "learning_rate": 8.871834212621773e-06, "loss": 0.2485, "step": 280 }, { "epoch": 0.43769470404984423, "grad_norm": 0.6108881149273678, "learning_rate": 8.86408197056456e-06, "loss": 0.2082, "step": 281 }, { "epoch": 0.4392523364485981, "grad_norm": 0.6626887696196829, "learning_rate": 8.856306596367351e-06, "loss": 0.2741, "step": 282 }, { "epoch": 0.440809968847352, "grad_norm": 0.6517640250027363, "learning_rate": 8.84850813657705e-06, "loss": 0.3733, "step": 283 }, { "epoch": 0.4423676012461059, "grad_norm": 0.5979891922361272, "learning_rate": 8.840686637878756e-06, "loss": 0.2615, "step": 284 }, { "epoch": 0.4439252336448598, "grad_norm": 0.5447033448285953, "learning_rate": 8.832842147095495e-06, "loss": 0.2204, "step": 285 }, { "epoch": 0.4454828660436137, "grad_norm": 0.6176926760203407, "learning_rate": 8.82497471118793e-06, "loss": 0.259, "step": 286 }, { "epoch": 0.4470404984423676, "grad_norm": 0.6803871655362536, "learning_rate": 8.817084377254089e-06, "loss": 0.3406, "step": 287 }, { "epoch": 0.4485981308411215, "grad_norm": 1.0924030353117906, "learning_rate": 8.809171192529074e-06, "loss": 0.7712, "step": 288 }, { "epoch": 0.4501557632398754, "grad_norm": 0.7261898907998104, "learning_rate": 8.80123520438478e-06, "loss": 0.2439, "step": 289 }, { "epoch": 0.4517133956386293, "grad_norm": 0.5278888030187282, "learning_rate": 8.79327646032962e-06, "loss": 0.2838, "step": 290 }, { "epoch": 0.4532710280373832, "grad_norm": 0.7336946619047033, "learning_rate": 8.785295008008227e-06, "loss": 0.2863, "step": 291 }, { "epoch": 0.45482866043613707, "grad_norm": 0.5663680560662904, "learning_rate": 8.777290895201186e-06, "loss": 0.2487, "step": 292 }, { "epoch": 0.45638629283489096, "grad_norm": 0.6674793012574268, "learning_rate": 8.769264169824725e-06, "loss": 0.2115, "step": 293 }, { "epoch": 0.45794392523364486, "grad_norm": 0.629476332991769, "learning_rate": 8.761214879930452e-06, "loss": 0.2698, "step": 294 }, { "epoch": 0.45950155763239875, "grad_norm": 0.5733049643694853, "learning_rate": 8.753143073705048e-06, "loss": 0.2755, "step": 295 }, { "epoch": 0.46105919003115264, "grad_norm": 0.7496857759281685, "learning_rate": 8.745048799469996e-06, "loss": 0.2066, "step": 296 }, { "epoch": 0.46261682242990654, "grad_norm": 0.5833814287468523, "learning_rate": 8.736932105681272e-06, "loss": 0.2233, "step": 297 }, { "epoch": 0.46417445482866043, "grad_norm": 0.7333726802931715, "learning_rate": 8.728793040929075e-06, "loss": 0.2529, "step": 298 }, { "epoch": 0.4657320872274143, "grad_norm": 0.6752352954357181, "learning_rate": 8.720631653937522e-06, "loss": 0.2487, "step": 299 }, { "epoch": 0.4672897196261682, "grad_norm": 0.6281610479983413, "learning_rate": 8.712447993564362e-06, "loss": 0.2513, "step": 300 }, { "epoch": 0.4688473520249221, "grad_norm": 1.719855082212449, "learning_rate": 8.70424210880068e-06, "loss": 1.2841, "step": 301 }, { "epoch": 0.470404984423676, "grad_norm": 0.8697991461292468, "learning_rate": 8.696014048770611e-06, "loss": 0.1752, "step": 302 }, { "epoch": 0.4719626168224299, "grad_norm": 0.7388848487578813, "learning_rate": 8.68776386273104e-06, "loss": 0.3559, "step": 303 }, { "epoch": 0.4735202492211838, "grad_norm": 0.6740217209253545, "learning_rate": 8.679491600071304e-06, "loss": 0.2506, "step": 304 }, { "epoch": 0.4750778816199377, "grad_norm": 0.8324393939489836, "learning_rate": 8.671197310312905e-06, "loss": 0.2951, "step": 305 }, { "epoch": 0.4766355140186916, "grad_norm": 0.5859142473050233, "learning_rate": 8.662881043109208e-06, "loss": 0.2737, "step": 306 }, { "epoch": 0.4781931464174455, "grad_norm": 0.5144143552921643, "learning_rate": 8.654542848245146e-06, "loss": 0.2555, "step": 307 }, { "epoch": 0.4797507788161994, "grad_norm": 0.9870060592323339, "learning_rate": 8.646182775636917e-06, "loss": 0.3268, "step": 308 }, { "epoch": 0.48130841121495327, "grad_norm": 1.22444825315241, "learning_rate": 8.637800875331693e-06, "loss": 0.8927, "step": 309 }, { "epoch": 0.48286604361370716, "grad_norm": 0.6219488136095235, "learning_rate": 8.629397197507315e-06, "loss": 0.1925, "step": 310 }, { "epoch": 0.48442367601246106, "grad_norm": 0.6663733783513853, "learning_rate": 8.620971792471994e-06, "loss": 0.2525, "step": 311 }, { "epoch": 0.48598130841121495, "grad_norm": 0.7779309034351384, "learning_rate": 8.612524710664012e-06, "loss": 0.2857, "step": 312 }, { "epoch": 0.48753894080996885, "grad_norm": 0.5815264215390993, "learning_rate": 8.604056002651414e-06, "loss": 0.2132, "step": 313 }, { "epoch": 0.48909657320872274, "grad_norm": 0.639985193258881, "learning_rate": 8.595565719131711e-06, "loss": 0.2927, "step": 314 }, { "epoch": 0.49065420560747663, "grad_norm": 0.5843630349664094, "learning_rate": 8.587053910931576e-06, "loss": 0.2073, "step": 315 }, { "epoch": 0.49221183800623053, "grad_norm": 0.5718073343739106, "learning_rate": 8.578520629006537e-06, "loss": 0.2701, "step": 316 }, { "epoch": 0.4937694704049844, "grad_norm": 0.8093573764947887, "learning_rate": 8.569965924440675e-06, "loss": 0.3701, "step": 317 }, { "epoch": 0.4953271028037383, "grad_norm": 0.45875585137104996, "learning_rate": 8.561389848446314e-06, "loss": 0.2535, "step": 318 }, { "epoch": 0.4968847352024922, "grad_norm": 0.6800533406672229, "learning_rate": 8.55279245236372e-06, "loss": 0.2542, "step": 319 }, { "epoch": 0.4984423676012461, "grad_norm": 0.5791491500061517, "learning_rate": 8.544173787660788e-06, "loss": 0.2233, "step": 320 }, { "epoch": 0.5, "grad_norm": 0.6853895685598861, "learning_rate": 8.535533905932739e-06, "loss": 0.1762, "step": 321 }, { "epoch": 0.5015576323987538, "grad_norm": 0.5400182911842227, "learning_rate": 8.526872858901806e-06, "loss": 0.2565, "step": 322 }, { "epoch": 0.5031152647975078, "grad_norm": 0.7099762938056243, "learning_rate": 8.518190698416929e-06, "loss": 0.2901, "step": 323 }, { "epoch": 0.5046728971962616, "grad_norm": 0.7351812911105055, "learning_rate": 8.509487476453442e-06, "loss": 0.2829, "step": 324 }, { "epoch": 0.5062305295950156, "grad_norm": 0.6208369210464704, "learning_rate": 8.500763245112763e-06, "loss": 0.2916, "step": 325 }, { "epoch": 0.5077881619937694, "grad_norm": 0.7390752110429559, "learning_rate": 8.492018056622083e-06, "loss": 0.3912, "step": 326 }, { "epoch": 0.5093457943925234, "grad_norm": 0.5930641088378493, "learning_rate": 8.483251963334047e-06, "loss": 0.2193, "step": 327 }, { "epoch": 0.5109034267912772, "grad_norm": 0.6344017285315496, "learning_rate": 8.474465017726452e-06, "loss": 0.2191, "step": 328 }, { "epoch": 0.5124610591900312, "grad_norm": 0.7250776947740573, "learning_rate": 8.465657272401921e-06, "loss": 0.2435, "step": 329 }, { "epoch": 0.514018691588785, "grad_norm": 0.6175903219629659, "learning_rate": 8.456828780087598e-06, "loss": 0.2005, "step": 330 }, { "epoch": 0.5155763239875389, "grad_norm": 0.6186407482326058, "learning_rate": 8.447979593634823e-06, "loss": 0.2969, "step": 331 }, { "epoch": 0.5171339563862928, "grad_norm": 0.6221498367824214, "learning_rate": 8.439109766018825e-06, "loss": 0.2532, "step": 332 }, { "epoch": 0.5186915887850467, "grad_norm": 0.5991537960888588, "learning_rate": 8.430219350338398e-06, "loss": 0.2029, "step": 333 }, { "epoch": 0.5202492211838006, "grad_norm": 0.5074549508649242, "learning_rate": 8.421308399815586e-06, "loss": 0.2452, "step": 334 }, { "epoch": 0.5218068535825545, "grad_norm": 1.0447367430008974, "learning_rate": 8.412376967795362e-06, "loss": 0.7821, "step": 335 }, { "epoch": 0.5233644859813084, "grad_norm": 0.6715190449661054, "learning_rate": 8.403425107745315e-06, "loss": 0.1873, "step": 336 }, { "epoch": 0.5249221183800623, "grad_norm": 0.8046073365053287, "learning_rate": 8.394452873255321e-06, "loss": 0.2459, "step": 337 }, { "epoch": 0.5264797507788161, "grad_norm": 0.6185327399453182, "learning_rate": 8.385460318037228e-06, "loss": 0.2347, "step": 338 }, { "epoch": 0.5280373831775701, "grad_norm": 0.6523073370428902, "learning_rate": 8.376447495924533e-06, "loss": 0.2494, "step": 339 }, { "epoch": 0.5295950155763239, "grad_norm": 0.8230020627407991, "learning_rate": 8.367414460872064e-06, "loss": 0.7898, "step": 340 }, { "epoch": 0.5311526479750779, "grad_norm": 0.6739918439315772, "learning_rate": 8.358361266955641e-06, "loss": 0.3159, "step": 341 }, { "epoch": 0.5327102803738317, "grad_norm": 0.7519773686193855, "learning_rate": 8.34928796837178e-06, "loss": 0.2593, "step": 342 }, { "epoch": 0.5342679127725857, "grad_norm": 0.5445971280446563, "learning_rate": 8.34019461943734e-06, "loss": 0.2923, "step": 343 }, { "epoch": 0.5358255451713395, "grad_norm": 0.7071155258569332, "learning_rate": 8.331081274589217e-06, "loss": 0.2598, "step": 344 }, { "epoch": 0.5373831775700935, "grad_norm": 0.719502132159577, "learning_rate": 8.321947988384006e-06, "loss": 0.276, "step": 345 }, { "epoch": 0.5389408099688473, "grad_norm": 0.5170866766320925, "learning_rate": 8.312794815497688e-06, "loss": 0.2236, "step": 346 }, { "epoch": 0.5404984423676013, "grad_norm": 0.6654228108717598, "learning_rate": 8.303621810725287e-06, "loss": 0.2666, "step": 347 }, { "epoch": 0.5420560747663551, "grad_norm": 0.6117403127181333, "learning_rate": 8.294429028980555e-06, "loss": 0.2767, "step": 348 }, { "epoch": 0.543613707165109, "grad_norm": 0.8584300691210696, "learning_rate": 8.285216525295636e-06, "loss": 0.2521, "step": 349 }, { "epoch": 0.5451713395638629, "grad_norm": 0.7222215782352946, "learning_rate": 8.275984354820736e-06, "loss": 0.2499, "step": 350 }, { "epoch": 0.5467289719626168, "grad_norm": 0.643947565152816, "learning_rate": 8.266732572823799e-06, "loss": 0.2092, "step": 351 }, { "epoch": 0.5482866043613707, "grad_norm": 0.6231517894213087, "learning_rate": 8.25746123469017e-06, "loss": 0.2598, "step": 352 }, { "epoch": 0.5498442367601246, "grad_norm": 0.5983797307315736, "learning_rate": 8.248170395922266e-06, "loss": 0.3241, "step": 353 }, { "epoch": 0.5514018691588785, "grad_norm": 0.6611331737862588, "learning_rate": 8.238860112139246e-06, "loss": 0.2292, "step": 354 }, { "epoch": 0.5529595015576324, "grad_norm": 0.667013942050513, "learning_rate": 8.229530439076674e-06, "loss": 0.2911, "step": 355 }, { "epoch": 0.5545171339563862, "grad_norm": 0.5711025724816864, "learning_rate": 8.220181432586187e-06, "loss": 0.3828, "step": 356 }, { "epoch": 0.5560747663551402, "grad_norm": 0.7326983314478233, "learning_rate": 8.210813148635158e-06, "loss": 0.4075, "step": 357 }, { "epoch": 0.557632398753894, "grad_norm": 0.5806269356738779, "learning_rate": 8.201425643306367e-06, "loss": 0.2283, "step": 358 }, { "epoch": 0.559190031152648, "grad_norm": 0.8188821554767335, "learning_rate": 8.192018972797665e-06, "loss": 0.2549, "step": 359 }, { "epoch": 0.5607476635514018, "grad_norm": 0.6809831648038962, "learning_rate": 8.182593193421625e-06, "loss": 0.2543, "step": 360 }, { "epoch": 0.5623052959501558, "grad_norm": 0.6203982930897892, "learning_rate": 8.173148361605224e-06, "loss": 0.3205, "step": 361 }, { "epoch": 0.5638629283489096, "grad_norm": 0.4686220806935477, "learning_rate": 8.163684533889489e-06, "loss": 0.2356, "step": 362 }, { "epoch": 0.5654205607476636, "grad_norm": 0.7261908034014758, "learning_rate": 8.154201766929167e-06, "loss": 0.2763, "step": 363 }, { "epoch": 0.5669781931464174, "grad_norm": 0.6301811228833201, "learning_rate": 8.144700117492386e-06, "loss": 0.2369, "step": 364 }, { "epoch": 0.5685358255451713, "grad_norm": 0.7734460446532929, "learning_rate": 8.135179642460308e-06, "loss": 0.2722, "step": 365 }, { "epoch": 0.5700934579439252, "grad_norm": 0.6997558366317673, "learning_rate": 8.125640398826803e-06, "loss": 0.2543, "step": 366 }, { "epoch": 0.5716510903426791, "grad_norm": 0.8988136993440748, "learning_rate": 8.116082443698085e-06, "loss": 0.2835, "step": 367 }, { "epoch": 0.573208722741433, "grad_norm": 0.8141909161490042, "learning_rate": 8.106505834292396e-06, "loss": 0.2538, "step": 368 }, { "epoch": 0.5747663551401869, "grad_norm": 1.3911734406138567, "learning_rate": 8.09691062793964e-06, "loss": 0.7141, "step": 369 }, { "epoch": 0.5763239875389408, "grad_norm": 0.6359597508417911, "learning_rate": 8.087296882081062e-06, "loss": 0.2794, "step": 370 }, { "epoch": 0.5778816199376947, "grad_norm": 0.5841003451023182, "learning_rate": 8.077664654268883e-06, "loss": 0.2973, "step": 371 }, { "epoch": 0.5794392523364486, "grad_norm": 0.6441301027816297, "learning_rate": 8.06801400216597e-06, "loss": 0.2477, "step": 372 }, { "epoch": 0.5809968847352025, "grad_norm": 0.6922555735413835, "learning_rate": 8.058344983545486e-06, "loss": 0.2524, "step": 373 }, { "epoch": 0.5825545171339563, "grad_norm": 0.6083842427437397, "learning_rate": 8.048657656290545e-06, "loss": 0.2837, "step": 374 }, { "epoch": 0.5841121495327103, "grad_norm": 0.5803164555571954, "learning_rate": 8.03895207839386e-06, "loss": 0.3084, "step": 375 }, { "epoch": 0.5856697819314641, "grad_norm": 0.8102916914102428, "learning_rate": 8.029228307957408e-06, "loss": 0.2462, "step": 376 }, { "epoch": 0.5872274143302181, "grad_norm": 0.7061482090328073, "learning_rate": 8.019486403192069e-06, "loss": 0.2487, "step": 377 }, { "epoch": 0.5887850467289719, "grad_norm": 0.5113763283575798, "learning_rate": 8.009726422417286e-06, "loss": 0.2931, "step": 378 }, { "epoch": 0.5903426791277259, "grad_norm": 0.6897001371312232, "learning_rate": 7.99994842406071e-06, "loss": 0.2186, "step": 379 }, { "epoch": 0.5919003115264797, "grad_norm": 0.9072221405351328, "learning_rate": 7.99015246665786e-06, "loss": 0.8348, "step": 380 }, { "epoch": 0.5934579439252337, "grad_norm": 0.6531934878590409, "learning_rate": 7.980338608851756e-06, "loss": 0.2228, "step": 381 }, { "epoch": 0.5950155763239875, "grad_norm": 0.7327462885118111, "learning_rate": 7.970506909392588e-06, "loss": 0.2433, "step": 382 }, { "epoch": 0.5965732087227414, "grad_norm": 1.1643513095965004, "learning_rate": 7.960657427137347e-06, "loss": 0.3147, "step": 383 }, { "epoch": 0.5981308411214953, "grad_norm": 0.6452843748972524, "learning_rate": 7.950790221049485e-06, "loss": 0.3303, "step": 384 }, { "epoch": 0.5996884735202492, "grad_norm": 0.7274077390061954, "learning_rate": 7.940905350198553e-06, "loss": 0.3347, "step": 385 }, { "epoch": 0.6012461059190031, "grad_norm": 0.6178112722471735, "learning_rate": 7.931002873759852e-06, "loss": 0.2276, "step": 386 }, { "epoch": 0.602803738317757, "grad_norm": 0.7461711392201733, "learning_rate": 7.921082851014079e-06, "loss": 0.2634, "step": 387 }, { "epoch": 0.6043613707165109, "grad_norm": 0.8795743223985983, "learning_rate": 7.911145341346972e-06, "loss": 0.2896, "step": 388 }, { "epoch": 0.6059190031152648, "grad_norm": 0.8319649658997385, "learning_rate": 7.901190404248952e-06, "loss": 0.7066, "step": 389 }, { "epoch": 0.6074766355140186, "grad_norm": 0.5267541152010679, "learning_rate": 7.89121809931477e-06, "loss": 0.2416, "step": 390 }, { "epoch": 0.6090342679127726, "grad_norm": 0.46625213633043144, "learning_rate": 7.881228486243144e-06, "loss": 0.2482, "step": 391 }, { "epoch": 0.6105919003115264, "grad_norm": 0.6924717742437231, "learning_rate": 7.871221624836414e-06, "loss": 0.3523, "step": 392 }, { "epoch": 0.6121495327102804, "grad_norm": 0.5838406566254681, "learning_rate": 7.861197575000168e-06, "loss": 0.228, "step": 393 }, { "epoch": 0.6137071651090342, "grad_norm": 0.8869116056189287, "learning_rate": 7.8511563967429e-06, "loss": 0.2496, "step": 394 }, { "epoch": 0.6152647975077882, "grad_norm": 0.45593356473166835, "learning_rate": 7.841098150175636e-06, "loss": 0.2643, "step": 395 }, { "epoch": 0.616822429906542, "grad_norm": 0.8435131653971922, "learning_rate": 7.831022895511586e-06, "loss": 0.2496, "step": 396 }, { "epoch": 0.618380062305296, "grad_norm": 0.4987004271913618, "learning_rate": 7.820930693065771e-06, "loss": 0.2063, "step": 397 }, { "epoch": 0.6199376947040498, "grad_norm": 0.7942215586211049, "learning_rate": 7.810821603254677e-06, "loss": 0.2181, "step": 398 }, { "epoch": 0.6214953271028038, "grad_norm": 0.6794367411830937, "learning_rate": 7.800695686595879e-06, "loss": 0.3009, "step": 399 }, { "epoch": 0.6230529595015576, "grad_norm": 0.7273130133089402, "learning_rate": 7.790553003707691e-06, "loss": 0.2469, "step": 400 }, { "epoch": 0.6246105919003115, "grad_norm": 0.6902330948327355, "learning_rate": 7.780393615308787e-06, "loss": 0.2508, "step": 401 }, { "epoch": 0.6261682242990654, "grad_norm": 0.7084362576430354, "learning_rate": 7.770217582217863e-06, "loss": 0.2551, "step": 402 }, { "epoch": 0.6277258566978193, "grad_norm": 0.8029205384416798, "learning_rate": 7.760024965353246e-06, "loss": 0.2333, "step": 403 }, { "epoch": 0.6292834890965732, "grad_norm": 0.6112029983652504, "learning_rate": 7.749815825732543e-06, "loss": 0.298, "step": 404 }, { "epoch": 0.6308411214953271, "grad_norm": 0.7494581341489577, "learning_rate": 7.739590224472275e-06, "loss": 0.3462, "step": 405 }, { "epoch": 0.632398753894081, "grad_norm": 0.7347669711126691, "learning_rate": 7.729348222787514e-06, "loss": 0.3149, "step": 406 }, { "epoch": 0.6339563862928349, "grad_norm": 0.6796064286407987, "learning_rate": 7.719089881991503e-06, "loss": 0.2873, "step": 407 }, { "epoch": 0.6355140186915887, "grad_norm": 0.7425509324765857, "learning_rate": 7.708815263495307e-06, "loss": 0.3278, "step": 408 }, { "epoch": 0.6370716510903427, "grad_norm": 0.609414275478013, "learning_rate": 7.698524428807431e-06, "loss": 0.2708, "step": 409 }, { "epoch": 0.6386292834890965, "grad_norm": 0.7757117977400942, "learning_rate": 7.68821743953346e-06, "loss": 0.2555, "step": 410 }, { "epoch": 0.6401869158878505, "grad_norm": 0.6642687790623766, "learning_rate": 7.677894357375689e-06, "loss": 0.3625, "step": 411 }, { "epoch": 0.6417445482866043, "grad_norm": 0.5791966784356082, "learning_rate": 7.667555244132749e-06, "loss": 0.2661, "step": 412 }, { "epoch": 0.6433021806853583, "grad_norm": 0.5594732951892226, "learning_rate": 7.65720016169924e-06, "loss": 0.2995, "step": 413 }, { "epoch": 0.6448598130841121, "grad_norm": 0.6021900759219545, "learning_rate": 7.646829172065367e-06, "loss": 0.3099, "step": 414 }, { "epoch": 0.6464174454828661, "grad_norm": 0.5562483872284556, "learning_rate": 7.636442337316555e-06, "loss": 0.2376, "step": 415 }, { "epoch": 0.6479750778816199, "grad_norm": 0.5829741964791303, "learning_rate": 7.6260397196330895e-06, "loss": 0.2774, "step": 416 }, { "epoch": 0.6495327102803738, "grad_norm": 0.7958468559537486, "learning_rate": 7.615621381289737e-06, "loss": 0.2316, "step": 417 }, { "epoch": 0.6510903426791277, "grad_norm": 0.6088648202059304, "learning_rate": 7.6051873846553795e-06, "loss": 0.33, "step": 418 }, { "epoch": 0.6526479750778816, "grad_norm": 0.523142097250351, "learning_rate": 7.594737792192629e-06, "loss": 0.2589, "step": 419 }, { "epoch": 0.6542056074766355, "grad_norm": 0.6233483869502079, "learning_rate": 7.584272666457471e-06, "loss": 0.2409, "step": 420 }, { "epoch": 0.6557632398753894, "grad_norm": 0.8831745439168878, "learning_rate": 7.573792070098873e-06, "loss": 0.3156, "step": 421 }, { "epoch": 0.6573208722741433, "grad_norm": 0.646939395915981, "learning_rate": 7.5632960658584184e-06, "loss": 0.1882, "step": 422 }, { "epoch": 0.6588785046728972, "grad_norm": 0.7493677482681486, "learning_rate": 7.5527847165699295e-06, "loss": 0.2533, "step": 423 }, { "epoch": 0.660436137071651, "grad_norm": 0.6895089900125264, "learning_rate": 7.542258085159091e-06, "loss": 0.2239, "step": 424 }, { "epoch": 0.661993769470405, "grad_norm": 0.8218899831192643, "learning_rate": 7.531716234643071e-06, "loss": 0.3025, "step": 425 }, { "epoch": 0.6635514018691588, "grad_norm": 0.8511691627825192, "learning_rate": 7.5211592281301525e-06, "loss": 0.2081, "step": 426 }, { "epoch": 0.6651090342679128, "grad_norm": 0.6202374314769092, "learning_rate": 7.510587128819341e-06, "loss": 0.2159, "step": 427 }, { "epoch": 0.6666666666666666, "grad_norm": 0.8637856332283039, "learning_rate": 7.500000000000001e-06, "loss": 0.6963, "step": 428 }, { "epoch": 0.6682242990654206, "grad_norm": 0.7070722564579784, "learning_rate": 7.489397905051465e-06, "loss": 0.3265, "step": 429 }, { "epoch": 0.6697819314641744, "grad_norm": 1.0912368137134154, "learning_rate": 7.478780907442665e-06, "loss": 0.3064, "step": 430 }, { "epoch": 0.6713395638629284, "grad_norm": 0.4993170737898787, "learning_rate": 7.468149070731742e-06, "loss": 0.3532, "step": 431 }, { "epoch": 0.6728971962616822, "grad_norm": 0.6200661171236782, "learning_rate": 7.457502458565673e-06, "loss": 0.2325, "step": 432 }, { "epoch": 0.6744548286604362, "grad_norm": 0.6152606478186087, "learning_rate": 7.446841134679888e-06, "loss": 0.2538, "step": 433 }, { "epoch": 0.67601246105919, "grad_norm": 0.5515844149625706, "learning_rate": 7.436165162897886e-06, "loss": 0.2619, "step": 434 }, { "epoch": 0.677570093457944, "grad_norm": 0.7008609334925875, "learning_rate": 7.425474607130858e-06, "loss": 0.3168, "step": 435 }, { "epoch": 0.6791277258566978, "grad_norm": 0.6379535213002501, "learning_rate": 7.414769531377298e-06, "loss": 0.268, "step": 436 }, { "epoch": 0.6806853582554517, "grad_norm": 0.5317208819493666, "learning_rate": 7.4040499997226245e-06, "loss": 0.2193, "step": 437 }, { "epoch": 0.6822429906542056, "grad_norm": 0.7290549365391932, "learning_rate": 7.393316076338798e-06, "loss": 0.3694, "step": 438 }, { "epoch": 0.6838006230529595, "grad_norm": 0.8546888899097251, "learning_rate": 7.382567825483929e-06, "loss": 0.2822, "step": 439 }, { "epoch": 0.6853582554517134, "grad_norm": 0.7390434139959143, "learning_rate": 7.371805311501905e-06, "loss": 0.24, "step": 440 }, { "epoch": 0.6869158878504673, "grad_norm": 0.7021761813392882, "learning_rate": 7.361028598821993e-06, "loss": 0.3065, "step": 441 }, { "epoch": 0.6884735202492211, "grad_norm": 0.5340954968447894, "learning_rate": 7.350237751958466e-06, "loss": 0.221, "step": 442 }, { "epoch": 0.6900311526479751, "grad_norm": 0.5761626294301733, "learning_rate": 7.339432835510203e-06, "loss": 0.2345, "step": 443 }, { "epoch": 0.6915887850467289, "grad_norm": 0.662018391928594, "learning_rate": 7.328613914160319e-06, "loss": 0.3171, "step": 444 }, { "epoch": 0.6931464174454829, "grad_norm": 0.8130781056088618, "learning_rate": 7.3177810526757594e-06, "loss": 0.2909, "step": 445 }, { "epoch": 0.6947040498442367, "grad_norm": 0.7219882547975953, "learning_rate": 7.3069343159069296e-06, "loss": 0.2481, "step": 446 }, { "epoch": 0.6962616822429907, "grad_norm": 0.6369674341462834, "learning_rate": 7.296073768787293e-06, "loss": 0.3649, "step": 447 }, { "epoch": 0.6978193146417445, "grad_norm": 0.7223244796104977, "learning_rate": 7.285199476332991e-06, "loss": 0.3488, "step": 448 }, { "epoch": 0.6993769470404985, "grad_norm": 0.9091117254585579, "learning_rate": 7.27431150364245e-06, "loss": 0.3168, "step": 449 }, { "epoch": 0.7009345794392523, "grad_norm": 0.7868966044967969, "learning_rate": 7.263409915895992e-06, "loss": 0.259, "step": 450 }, { "epoch": 0.7024922118380063, "grad_norm": 0.6563385278402535, "learning_rate": 7.252494778355444e-06, "loss": 0.25, "step": 451 }, { "epoch": 0.7040498442367601, "grad_norm": 0.641132207138942, "learning_rate": 7.2415661563637506e-06, "loss": 0.3307, "step": 452 }, { "epoch": 0.705607476635514, "grad_norm": 0.7073578438725788, "learning_rate": 7.23062411534458e-06, "loss": 0.2261, "step": 453 }, { "epoch": 0.7071651090342679, "grad_norm": 1.040324988179143, "learning_rate": 7.2196687208019315e-06, "loss": 0.2057, "step": 454 }, { "epoch": 0.7087227414330218, "grad_norm": 0.6121914696936145, "learning_rate": 7.208700038319744e-06, "loss": 0.3199, "step": 455 }, { "epoch": 0.7102803738317757, "grad_norm": 0.48700575482675645, "learning_rate": 7.1977181335615085e-06, "loss": 0.2259, "step": 456 }, { "epoch": 0.7118380062305296, "grad_norm": 0.545525370035186, "learning_rate": 7.186723072269863e-06, "loss": 0.268, "step": 457 }, { "epoch": 0.7133956386292835, "grad_norm": 0.846012722177333, "learning_rate": 7.175714920266214e-06, "loss": 0.7256, "step": 458 }, { "epoch": 0.7149532710280374, "grad_norm": 0.5989442738821008, "learning_rate": 7.164693743450329e-06, "loss": 0.3005, "step": 459 }, { "epoch": 0.7165109034267912, "grad_norm": 0.6556232944526054, "learning_rate": 7.153659607799952e-06, "loss": 0.2745, "step": 460 }, { "epoch": 0.7180685358255452, "grad_norm": 0.6984028449665124, "learning_rate": 7.142612579370402e-06, "loss": 0.2272, "step": 461 }, { "epoch": 0.719626168224299, "grad_norm": 0.5324547293774875, "learning_rate": 7.131552724294181e-06, "loss": 0.2518, "step": 462 }, { "epoch": 0.721183800623053, "grad_norm": 0.5943008852162496, "learning_rate": 7.1204801087805765e-06, "loss": 0.2663, "step": 463 }, { "epoch": 0.7227414330218068, "grad_norm": 0.5954388212811877, "learning_rate": 7.109394799115268e-06, "loss": 0.25, "step": 464 }, { "epoch": 0.7242990654205608, "grad_norm": 0.5237590199785461, "learning_rate": 7.098296861659925e-06, "loss": 0.2451, "step": 465 }, { "epoch": 0.7258566978193146, "grad_norm": 0.6108790141608955, "learning_rate": 7.0871863628518136e-06, "loss": 0.2782, "step": 466 }, { "epoch": 0.7274143302180686, "grad_norm": 0.5784066614984076, "learning_rate": 7.0760633692033975e-06, "loss": 0.2588, "step": 467 }, { "epoch": 0.7289719626168224, "grad_norm": 0.5736026805586273, "learning_rate": 7.064927947301942e-06, "loss": 0.3319, "step": 468 }, { "epoch": 0.7305295950155763, "grad_norm": 0.660786534496975, "learning_rate": 7.0537801638091116e-06, "loss": 0.3207, "step": 469 }, { "epoch": 0.7320872274143302, "grad_norm": 0.6164260678789174, "learning_rate": 7.042620085460574e-06, "loss": 0.2759, "step": 470 }, { "epoch": 0.7336448598130841, "grad_norm": 1.0298833742062845, "learning_rate": 7.0314477790656e-06, "loss": 0.2769, "step": 471 }, { "epoch": 0.735202492211838, "grad_norm": 0.7848249743313419, "learning_rate": 7.020263311506659e-06, "loss": 0.3963, "step": 472 }, { "epoch": 0.7367601246105919, "grad_norm": 0.5488287365596327, "learning_rate": 7.009066749739026e-06, "loss": 0.2244, "step": 473 }, { "epoch": 0.7383177570093458, "grad_norm": 0.5479634675942974, "learning_rate": 6.99785816079038e-06, "loss": 0.277, "step": 474 }, { "epoch": 0.7398753894080997, "grad_norm": 0.632641842156797, "learning_rate": 6.986637611760394e-06, "loss": 0.2948, "step": 475 }, { "epoch": 0.7414330218068536, "grad_norm": 0.8957280703439034, "learning_rate": 6.975405169820344e-06, "loss": 0.353, "step": 476 }, { "epoch": 0.7429906542056075, "grad_norm": 0.575855558736389, "learning_rate": 6.9641609022127e-06, "loss": 0.2667, "step": 477 }, { "epoch": 0.7445482866043613, "grad_norm": 0.6675031465700932, "learning_rate": 6.952904876250729e-06, "loss": 0.239, "step": 478 }, { "epoch": 0.7461059190031153, "grad_norm": 0.5488652674770181, "learning_rate": 6.941637159318083e-06, "loss": 0.2605, "step": 479 }, { "epoch": 0.7476635514018691, "grad_norm": 0.765674305969199, "learning_rate": 6.9303578188684085e-06, "loss": 0.2668, "step": 480 }, { "epoch": 0.7492211838006231, "grad_norm": 0.6239515815918181, "learning_rate": 6.919066922424931e-06, "loss": 0.2883, "step": 481 }, { "epoch": 0.7507788161993769, "grad_norm": 0.8869720139234101, "learning_rate": 6.907764537580053e-06, "loss": 0.2726, "step": 482 }, { "epoch": 0.7523364485981309, "grad_norm": 0.7182796918869947, "learning_rate": 6.896450731994959e-06, "loss": 0.2575, "step": 483 }, { "epoch": 0.7538940809968847, "grad_norm": 0.689714104473123, "learning_rate": 6.8851255733992006e-06, "loss": 0.2548, "step": 484 }, { "epoch": 0.7554517133956387, "grad_norm": 0.8752253075858156, "learning_rate": 6.873789129590287e-06, "loss": 0.2598, "step": 485 }, { "epoch": 0.7570093457943925, "grad_norm": 0.6547980788626615, "learning_rate": 6.862441468433298e-06, "loss": 0.274, "step": 486 }, { "epoch": 0.7585669781931464, "grad_norm": 0.6955009265885427, "learning_rate": 6.851082657860453e-06, "loss": 0.286, "step": 487 }, { "epoch": 0.7601246105919003, "grad_norm": 0.6057981135550708, "learning_rate": 6.839712765870725e-06, "loss": 0.3072, "step": 488 }, { "epoch": 0.7616822429906542, "grad_norm": 0.5562050274960125, "learning_rate": 6.828331860529422e-06, "loss": 0.2765, "step": 489 }, { "epoch": 0.7632398753894081, "grad_norm": 0.9242326126038012, "learning_rate": 6.816940009967787e-06, "loss": 0.8322, "step": 490 }, { "epoch": 0.764797507788162, "grad_norm": 0.7207873437414208, "learning_rate": 6.805537282382581e-06, "loss": 0.2175, "step": 491 }, { "epoch": 0.7663551401869159, "grad_norm": 0.5928431496932391, "learning_rate": 6.79412374603568e-06, "loss": 0.24, "step": 492 }, { "epoch": 0.7679127725856698, "grad_norm": 0.8086943486132299, "learning_rate": 6.782699469253671e-06, "loss": 0.3252, "step": 493 }, { "epoch": 0.7694704049844237, "grad_norm": 0.6500702055304157, "learning_rate": 6.771264520427432e-06, "loss": 0.2831, "step": 494 }, { "epoch": 0.7710280373831776, "grad_norm": 0.5699110226071109, "learning_rate": 6.759818968011731e-06, "loss": 0.2604, "step": 495 }, { "epoch": 0.7725856697819314, "grad_norm": 0.5580360203832775, "learning_rate": 6.748362880524819e-06, "loss": 0.2684, "step": 496 }, { "epoch": 0.7741433021806854, "grad_norm": 0.7886971673525824, "learning_rate": 6.736896326548006e-06, "loss": 0.2123, "step": 497 }, { "epoch": 0.7757009345794392, "grad_norm": 0.6054794222872896, "learning_rate": 6.7254193747252645e-06, "loss": 0.3127, "step": 498 }, { "epoch": 0.7772585669781932, "grad_norm": 0.7462264966697667, "learning_rate": 6.713932093762811e-06, "loss": 0.3051, "step": 499 }, { "epoch": 0.778816199376947, "grad_norm": 0.6730107352048917, "learning_rate": 6.702434552428702e-06, "loss": 0.3007, "step": 500 }, { "epoch": 0.778816199376947, "eval_loss": 0.366039514541626, "eval_runtime": 2.8278, "eval_samples_per_second": 9.194, "eval_steps_per_second": 2.475, "step": 500 }, { "epoch": 0.780373831775701, "grad_norm": 0.6861562261314369, "learning_rate": 6.690926819552408e-06, "loss": 0.287, "step": 501 }, { "epoch": 0.7819314641744548, "grad_norm": 0.9308842751027873, "learning_rate": 6.679408964024419e-06, "loss": 0.8811, "step": 502 }, { "epoch": 0.7834890965732088, "grad_norm": 0.6729340583401545, "learning_rate": 6.667881054795818e-06, "loss": 0.2304, "step": 503 }, { "epoch": 0.7850467289719626, "grad_norm": 0.7848693380567189, "learning_rate": 6.65634316087788e-06, "loss": 0.2965, "step": 504 }, { "epoch": 0.7866043613707165, "grad_norm": 0.7035209610164758, "learning_rate": 6.6447953513416474e-06, "loss": 0.2589, "step": 505 }, { "epoch": 0.7881619937694704, "grad_norm": 0.5912497530045528, "learning_rate": 6.633237695317523e-06, "loss": 0.2566, "step": 506 }, { "epoch": 0.7897196261682243, "grad_norm": 0.7247030582803601, "learning_rate": 6.621670261994857e-06, "loss": 0.2726, "step": 507 }, { "epoch": 0.7912772585669782, "grad_norm": 0.6656199289111854, "learning_rate": 6.610093120621532e-06, "loss": 0.2999, "step": 508 }, { "epoch": 0.7928348909657321, "grad_norm": 0.8158049430019846, "learning_rate": 6.598506340503541e-06, "loss": 0.2453, "step": 509 }, { "epoch": 0.794392523364486, "grad_norm": 0.6984808487227331, "learning_rate": 6.586909991004587e-06, "loss": 0.3149, "step": 510 }, { "epoch": 0.7959501557632399, "grad_norm": 0.6126995511808185, "learning_rate": 6.575304141545653e-06, "loss": 0.2666, "step": 511 }, { "epoch": 0.7975077881619937, "grad_norm": 0.5505067863178127, "learning_rate": 6.5636888616046e-06, "loss": 0.2998, "step": 512 }, { "epoch": 0.7990654205607477, "grad_norm": 0.7685023837309144, "learning_rate": 6.552064220715737e-06, "loss": 0.1876, "step": 513 }, { "epoch": 0.8006230529595015, "grad_norm": 0.6186176349687995, "learning_rate": 6.5404302884694145e-06, "loss": 0.2823, "step": 514 }, { "epoch": 0.8021806853582555, "grad_norm": 0.6245560858038731, "learning_rate": 6.528787134511608e-06, "loss": 0.2063, "step": 515 }, { "epoch": 0.8037383177570093, "grad_norm": 0.6027333943405707, "learning_rate": 6.5171348285434965e-06, "loss": 0.3079, "step": 516 }, { "epoch": 0.8052959501557633, "grad_norm": 0.7334879584524152, "learning_rate": 6.505473440321044e-06, "loss": 0.2906, "step": 517 }, { "epoch": 0.8068535825545171, "grad_norm": 0.5611150628723894, "learning_rate": 6.493803039654589e-06, "loss": 0.2437, "step": 518 }, { "epoch": 0.8084112149532711, "grad_norm": 0.5709098533110826, "learning_rate": 6.48212369640842e-06, "loss": 0.2695, "step": 519 }, { "epoch": 0.8099688473520249, "grad_norm": 0.7469745364074795, "learning_rate": 6.4704354805003626e-06, "loss": 0.2828, "step": 520 }, { "epoch": 0.8115264797507789, "grad_norm": 0.7580041865120294, "learning_rate": 6.458738461901354e-06, "loss": 0.2456, "step": 521 }, { "epoch": 0.8130841121495327, "grad_norm": 0.7356278600835281, "learning_rate": 6.447032710635035e-06, "loss": 0.2325, "step": 522 }, { "epoch": 0.8146417445482866, "grad_norm": 0.5694798902062997, "learning_rate": 6.435318296777316e-06, "loss": 0.2763, "step": 523 }, { "epoch": 0.8161993769470405, "grad_norm": 0.6227419240061058, "learning_rate": 6.423595290455971e-06, "loss": 0.2871, "step": 524 }, { "epoch": 0.8177570093457944, "grad_norm": 0.7005734890264759, "learning_rate": 6.41186376185021e-06, "loss": 0.3003, "step": 525 }, { "epoch": 0.8193146417445483, "grad_norm": 0.8909957470129115, "learning_rate": 6.400123781190265e-06, "loss": 0.3328, "step": 526 }, { "epoch": 0.8208722741433022, "grad_norm": 0.5534613081412874, "learning_rate": 6.388375418756959e-06, "loss": 0.2816, "step": 527 }, { "epoch": 0.822429906542056, "grad_norm": 0.7546735553270361, "learning_rate": 6.3766187448813e-06, "loss": 0.249, "step": 528 }, { "epoch": 0.82398753894081, "grad_norm": 1.08724358600162, "learning_rate": 6.3648538299440444e-06, "loss": 0.2978, "step": 529 }, { "epoch": 0.8255451713395638, "grad_norm": 0.5510668935985298, "learning_rate": 6.35308074437529e-06, "loss": 0.2586, "step": 530 }, { "epoch": 0.8271028037383178, "grad_norm": 0.9295747471096518, "learning_rate": 6.341299558654042e-06, "loss": 0.4423, "step": 531 }, { "epoch": 0.8286604361370716, "grad_norm": 0.6078601432381958, "learning_rate": 6.329510343307801e-06, "loss": 0.3089, "step": 532 }, { "epoch": 0.8302180685358256, "grad_norm": 0.7009464939112006, "learning_rate": 6.3177131689121325e-06, "loss": 0.276, "step": 533 }, { "epoch": 0.8317757009345794, "grad_norm": 0.6055023924718476, "learning_rate": 6.305908106090255e-06, "loss": 0.289, "step": 534 }, { "epoch": 0.8333333333333334, "grad_norm": 1.0146910918408667, "learning_rate": 6.294095225512604e-06, "loss": 0.2117, "step": 535 }, { "epoch": 0.8348909657320872, "grad_norm": 0.5684609592585866, "learning_rate": 6.282274597896421e-06, "loss": 0.268, "step": 536 }, { "epoch": 0.8364485981308412, "grad_norm": 0.6324240721524141, "learning_rate": 6.2704462940053165e-06, "loss": 0.2348, "step": 537 }, { "epoch": 0.838006230529595, "grad_norm": 0.582281483203043, "learning_rate": 6.2586103846488654e-06, "loss": 0.2975, "step": 538 }, { "epoch": 0.839563862928349, "grad_norm": 0.9776397911217686, "learning_rate": 6.246766940682165e-06, "loss": 0.7799, "step": 539 }, { "epoch": 0.8411214953271028, "grad_norm": 0.5174311636719064, "learning_rate": 6.234916033005421e-06, "loss": 0.1973, "step": 540 }, { "epoch": 0.8426791277258567, "grad_norm": 0.6331348424871293, "learning_rate": 6.22305773256352e-06, "loss": 0.2749, "step": 541 }, { "epoch": 0.8442367601246106, "grad_norm": 0.5229024799327089, "learning_rate": 6.211192110345603e-06, "loss": 0.2811, "step": 542 }, { "epoch": 0.8457943925233645, "grad_norm": 0.5575336291274628, "learning_rate": 6.199319237384645e-06, "loss": 0.2534, "step": 543 }, { "epoch": 0.8473520249221184, "grad_norm": 0.7289277957152529, "learning_rate": 6.187439184757025e-06, "loss": 0.7677, "step": 544 }, { "epoch": 0.8489096573208723, "grad_norm": 0.540862000313681, "learning_rate": 6.1755520235821055e-06, "loss": 0.3294, "step": 545 }, { "epoch": 0.8504672897196262, "grad_norm": 0.7035928510596402, "learning_rate": 6.163657825021802e-06, "loss": 0.3147, "step": 546 }, { "epoch": 0.8520249221183801, "grad_norm": 0.5438019081147566, "learning_rate": 6.1517566602801596e-06, "loss": 0.2003, "step": 547 }, { "epoch": 0.8535825545171339, "grad_norm": 0.8025432437697821, "learning_rate": 6.139848600602926e-06, "loss": 0.2756, "step": 548 }, { "epoch": 0.8551401869158879, "grad_norm": 0.654724311621637, "learning_rate": 6.127933717277123e-06, "loss": 0.2934, "step": 549 }, { "epoch": 0.8566978193146417, "grad_norm": 1.328494349119985, "learning_rate": 6.116012081630629e-06, "loss": 0.2731, "step": 550 }, { "epoch": 0.8582554517133957, "grad_norm": 0.5435085655801555, "learning_rate": 6.104083765031734e-06, "loss": 0.1934, "step": 551 }, { "epoch": 0.8598130841121495, "grad_norm": 0.7875228878328122, "learning_rate": 6.0921488388887315e-06, "loss": 0.7651, "step": 552 }, { "epoch": 0.8613707165109035, "grad_norm": 0.9138141754922854, "learning_rate": 6.080207374649482e-06, "loss": 0.6927, "step": 553 }, { "epoch": 0.8629283489096573, "grad_norm": 0.658537276564411, "learning_rate": 6.068259443800981e-06, "loss": 0.3088, "step": 554 }, { "epoch": 0.8644859813084113, "grad_norm": 0.8424646536545584, "learning_rate": 6.0563051178689395e-06, "loss": 0.6504, "step": 555 }, { "epoch": 0.8660436137071651, "grad_norm": 0.5487578856405264, "learning_rate": 6.0443444684173524e-06, "loss": 0.2504, "step": 556 }, { "epoch": 0.867601246105919, "grad_norm": 0.577742497246078, "learning_rate": 6.032377567048071e-06, "loss": 0.2724, "step": 557 }, { "epoch": 0.8691588785046729, "grad_norm": 0.5720861179090082, "learning_rate": 6.0204044854003705e-06, "loss": 0.2494, "step": 558 }, { "epoch": 0.8707165109034268, "grad_norm": 0.614579567677496, "learning_rate": 6.008425295150526e-06, "loss": 0.2431, "step": 559 }, { "epoch": 0.8722741433021807, "grad_norm": 0.7053311054530548, "learning_rate": 5.996440068011383e-06, "loss": 0.3007, "step": 560 }, { "epoch": 0.8738317757009346, "grad_norm": 0.6676390464189279, "learning_rate": 5.9844488757319205e-06, "loss": 0.2309, "step": 561 }, { "epoch": 0.8753894080996885, "grad_norm": 0.7749136512043995, "learning_rate": 5.972451790096837e-06, "loss": 0.3327, "step": 562 }, { "epoch": 0.8769470404984424, "grad_norm": 0.6384638289071073, "learning_rate": 5.960448882926101e-06, "loss": 0.3447, "step": 563 }, { "epoch": 0.8785046728971962, "grad_norm": 0.5992691951009588, "learning_rate": 5.948440226074539e-06, "loss": 0.2181, "step": 564 }, { "epoch": 0.8800623052959502, "grad_norm": 0.7553595588283479, "learning_rate": 5.936425891431394e-06, "loss": 0.2307, "step": 565 }, { "epoch": 0.881619937694704, "grad_norm": 0.7852167594459105, "learning_rate": 5.924405950919902e-06, "loss": 0.3119, "step": 566 }, { "epoch": 0.883177570093458, "grad_norm": 0.5241767306732831, "learning_rate": 5.91238047649685e-06, "loss": 0.2293, "step": 567 }, { "epoch": 0.8847352024922118, "grad_norm": 0.7471757161407457, "learning_rate": 5.900349540152167e-06, "loss": 0.3251, "step": 568 }, { "epoch": 0.8862928348909658, "grad_norm": 0.7843562558921093, "learning_rate": 5.888313213908468e-06, "loss": 0.2868, "step": 569 }, { "epoch": 0.8878504672897196, "grad_norm": 0.7503757117692468, "learning_rate": 5.876271569820638e-06, "loss": 0.2555, "step": 570 }, { "epoch": 0.8894080996884736, "grad_norm": 0.7275473381141497, "learning_rate": 5.864224679975399e-06, "loss": 0.2945, "step": 571 }, { "epoch": 0.8909657320872274, "grad_norm": 0.7108035276261796, "learning_rate": 5.852172616490875e-06, "loss": 0.2826, "step": 572 }, { "epoch": 0.8925233644859814, "grad_norm": 0.7292991034746364, "learning_rate": 5.84011545151616e-06, "loss": 0.2052, "step": 573 }, { "epoch": 0.8940809968847352, "grad_norm": 0.7701552350107903, "learning_rate": 5.828053257230893e-06, "loss": 0.272, "step": 574 }, { "epoch": 0.8956386292834891, "grad_norm": 0.5385532832886137, "learning_rate": 5.815986105844813e-06, "loss": 0.2859, "step": 575 }, { "epoch": 0.897196261682243, "grad_norm": 0.4784130573714912, "learning_rate": 5.803914069597342e-06, "loss": 0.2385, "step": 576 }, { "epoch": 0.8987538940809969, "grad_norm": 0.6661872649899948, "learning_rate": 5.791837220757139e-06, "loss": 0.2601, "step": 577 }, { "epoch": 0.9003115264797508, "grad_norm": 0.6423784935357807, "learning_rate": 5.779755631621679e-06, "loss": 0.2861, "step": 578 }, { "epoch": 0.9018691588785047, "grad_norm": 0.5564393840451491, "learning_rate": 5.767669374516807e-06, "loss": 0.2247, "step": 579 }, { "epoch": 0.9034267912772586, "grad_norm": 1.0831258133460262, "learning_rate": 5.755578521796321e-06, "loss": 0.7525, "step": 580 }, { "epoch": 0.9049844236760125, "grad_norm": 0.7513349003919912, "learning_rate": 5.743483145841525e-06, "loss": 0.2417, "step": 581 }, { "epoch": 0.9065420560747663, "grad_norm": 0.6795049942946425, "learning_rate": 5.731383319060805e-06, "loss": 0.3177, "step": 582 }, { "epoch": 0.9080996884735203, "grad_norm": 0.7390882203257174, "learning_rate": 5.719279113889184e-06, "loss": 0.3581, "step": 583 }, { "epoch": 0.9096573208722741, "grad_norm": 0.582238701325137, "learning_rate": 5.707170602787908e-06, "loss": 0.2755, "step": 584 }, { "epoch": 0.9112149532710281, "grad_norm": 0.5600730304265147, "learning_rate": 5.695057858243989e-06, "loss": 0.2745, "step": 585 }, { "epoch": 0.9127725856697819, "grad_norm": 0.638261445819756, "learning_rate": 5.682940952769788e-06, "loss": 0.177, "step": 586 }, { "epoch": 0.9143302180685359, "grad_norm": 0.6191376433403027, "learning_rate": 5.670819958902576e-06, "loss": 0.2447, "step": 587 }, { "epoch": 0.9158878504672897, "grad_norm": 0.8780779667116613, "learning_rate": 5.658694949204094e-06, "loss": 0.2438, "step": 588 }, { "epoch": 0.9174454828660437, "grad_norm": 0.7198384217051569, "learning_rate": 5.646565996260129e-06, "loss": 0.2408, "step": 589 }, { "epoch": 0.9190031152647975, "grad_norm": 1.0167707545321327, "learning_rate": 5.634433172680072e-06, "loss": 0.7316, "step": 590 }, { "epoch": 0.9205607476635514, "grad_norm": 0.6546640817603999, "learning_rate": 5.622296551096481e-06, "loss": 0.3253, "step": 591 }, { "epoch": 0.9221183800623053, "grad_norm": 0.8382914937621136, "learning_rate": 5.61015620416466e-06, "loss": 0.2981, "step": 592 }, { "epoch": 0.9236760124610592, "grad_norm": 0.6120676244562511, "learning_rate": 5.598012204562204e-06, "loss": 0.2647, "step": 593 }, { "epoch": 0.9252336448598131, "grad_norm": 0.700304920368459, "learning_rate": 5.5858646249885855e-06, "loss": 0.2249, "step": 594 }, { "epoch": 0.926791277258567, "grad_norm": 0.6840607666164885, "learning_rate": 5.573713538164698e-06, "loss": 0.2795, "step": 595 }, { "epoch": 0.9283489096573209, "grad_norm": 0.8513439406827232, "learning_rate": 5.561559016832438e-06, "loss": 0.2931, "step": 596 }, { "epoch": 0.9299065420560748, "grad_norm": 0.7770854828276103, "learning_rate": 5.549401133754259e-06, "loss": 0.2819, "step": 597 }, { "epoch": 0.9314641744548287, "grad_norm": 0.5874428152071298, "learning_rate": 5.5372399617127415e-06, "loss": 0.4483, "step": 598 }, { "epoch": 0.9330218068535826, "grad_norm": 0.6488291930752155, "learning_rate": 5.525075573510154e-06, "loss": 0.3873, "step": 599 }, { "epoch": 0.9345794392523364, "grad_norm": 0.5286049009581415, "learning_rate": 5.512908041968018e-06, "loss": 0.3113, "step": 600 }, { "epoch": 0.9361370716510904, "grad_norm": 0.5929114341675884, "learning_rate": 5.500737439926674e-06, "loss": 0.334, "step": 601 }, { "epoch": 0.9376947040498442, "grad_norm": 0.847988678139219, "learning_rate": 5.488563840244843e-06, "loss": 0.3026, "step": 602 }, { "epoch": 0.9392523364485982, "grad_norm": 0.5635536494534855, "learning_rate": 5.476387315799189e-06, "loss": 0.2146, "step": 603 }, { "epoch": 0.940809968847352, "grad_norm": 0.5570339586064699, "learning_rate": 5.464207939483891e-06, "loss": 0.2407, "step": 604 }, { "epoch": 0.942367601246106, "grad_norm": 0.5996622051462993, "learning_rate": 5.452025784210193e-06, "loss": 0.2301, "step": 605 }, { "epoch": 0.9439252336448598, "grad_norm": 0.6134433053870486, "learning_rate": 5.439840922905982e-06, "loss": 0.2881, "step": 606 }, { "epoch": 0.9454828660436138, "grad_norm": 0.6534399078221188, "learning_rate": 5.42765342851534e-06, "loss": 0.2991, "step": 607 }, { "epoch": 0.9470404984423676, "grad_norm": 0.690575160461211, "learning_rate": 5.415463373998112e-06, "loss": 0.3353, "step": 608 }, { "epoch": 0.9485981308411215, "grad_norm": 0.8921403858615625, "learning_rate": 5.403270832329473e-06, "loss": 0.2008, "step": 609 }, { "epoch": 0.9501557632398754, "grad_norm": 0.7025072704858318, "learning_rate": 5.391075876499483e-06, "loss": 0.2621, "step": 610 }, { "epoch": 0.9517133956386293, "grad_norm": 0.6757376481036776, "learning_rate": 5.3788785795126554e-06, "loss": 0.2469, "step": 611 }, { "epoch": 0.9532710280373832, "grad_norm": 0.6875817981934039, "learning_rate": 5.36667901438752e-06, "loss": 0.236, "step": 612 }, { "epoch": 0.9548286604361371, "grad_norm": 0.7529907066389188, "learning_rate": 5.354477254156184e-06, "loss": 0.2755, "step": 613 }, { "epoch": 0.956386292834891, "grad_norm": 0.5896914001022201, "learning_rate": 5.342273371863895e-06, "loss": 0.2634, "step": 614 }, { "epoch": 0.9579439252336449, "grad_norm": 0.7027739875415374, "learning_rate": 5.330067440568605e-06, "loss": 0.2829, "step": 615 }, { "epoch": 0.9595015576323987, "grad_norm": 0.7667507882778263, "learning_rate": 5.317859533340532e-06, "loss": 0.3506, "step": 616 }, { "epoch": 0.9610591900311527, "grad_norm": 0.8456849495117152, "learning_rate": 5.30564972326172e-06, "loss": 0.3054, "step": 617 }, { "epoch": 0.9626168224299065, "grad_norm": 0.51248054747711, "learning_rate": 5.293438083425611e-06, "loss": 0.2301, "step": 618 }, { "epoch": 0.9641744548286605, "grad_norm": 0.7467447701014356, "learning_rate": 5.281224686936594e-06, "loss": 0.3769, "step": 619 }, { "epoch": 0.9657320872274143, "grad_norm": 0.6414852824849853, "learning_rate": 5.26900960690958e-06, "loss": 0.2779, "step": 620 }, { "epoch": 0.9672897196261683, "grad_norm": 0.6295561918474766, "learning_rate": 5.256792916469552e-06, "loss": 0.2586, "step": 621 }, { "epoch": 0.9688473520249221, "grad_norm": 0.6049915483456579, "learning_rate": 5.244574688751138e-06, "loss": 0.2195, "step": 622 }, { "epoch": 0.9704049844236761, "grad_norm": 1.0443726591172307, "learning_rate": 5.23235499689817e-06, "loss": 0.7974, "step": 623 }, { "epoch": 0.9719626168224299, "grad_norm": 0.4947295180592083, "learning_rate": 5.220133914063239e-06, "loss": 0.2173, "step": 624 }, { "epoch": 0.9735202492211839, "grad_norm": 0.774384784856486, "learning_rate": 5.20791151340727e-06, "loss": 0.2196, "step": 625 }, { "epoch": 0.9750778816199377, "grad_norm": 0.6848166382044595, "learning_rate": 5.195687868099073e-06, "loss": 0.2784, "step": 626 }, { "epoch": 0.9766355140186916, "grad_norm": 1.5114886266488685, "learning_rate": 5.1834630513149086e-06, "loss": 0.8531, "step": 627 }, { "epoch": 0.9781931464174455, "grad_norm": 1.0107267590162416, "learning_rate": 5.171237136238054e-06, "loss": 0.2692, "step": 628 }, { "epoch": 0.9797507788161994, "grad_norm": 0.7292686375997546, "learning_rate": 5.159010196058356e-06, "loss": 0.2305, "step": 629 }, { "epoch": 0.9813084112149533, "grad_norm": 0.565137268345777, "learning_rate": 5.1467823039718046e-06, "loss": 0.3076, "step": 630 }, { "epoch": 0.9828660436137072, "grad_norm": 1.0043606696953649, "learning_rate": 5.134553533180082e-06, "loss": 0.7515, "step": 631 }, { "epoch": 0.9844236760124611, "grad_norm": 0.6233222851344706, "learning_rate": 5.122323956890136e-06, "loss": 0.3019, "step": 632 }, { "epoch": 0.985981308411215, "grad_norm": 0.730841019679576, "learning_rate": 5.110093648313732e-06, "loss": 0.3483, "step": 633 }, { "epoch": 0.9875389408099688, "grad_norm": 0.6632089778706831, "learning_rate": 5.097862680667024e-06, "loss": 0.3608, "step": 634 }, { "epoch": 0.9890965732087228, "grad_norm": 0.6691349603071851, "learning_rate": 5.085631127170106e-06, "loss": 0.2524, "step": 635 }, { "epoch": 0.9906542056074766, "grad_norm": 0.73951871814352, "learning_rate": 5.073399061046584e-06, "loss": 0.3134, "step": 636 }, { "epoch": 0.9922118380062306, "grad_norm": 0.6079224831882191, "learning_rate": 5.061166555523129e-06, "loss": 0.308, "step": 637 }, { "epoch": 0.9937694704049844, "grad_norm": 0.6166501321049228, "learning_rate": 5.048933683829046e-06, "loss": 0.3073, "step": 638 }, { "epoch": 0.9953271028037384, "grad_norm": 1.003680827370098, "learning_rate": 5.0367005191958275e-06, "loss": 0.3229, "step": 639 }, { "epoch": 0.9968847352024922, "grad_norm": 0.7517097037555416, "learning_rate": 5.024467134856725e-06, "loss": 0.2744, "step": 640 }, { "epoch": 0.9984423676012462, "grad_norm": 0.5399606823979543, "learning_rate": 5.012233604046303e-06, "loss": 0.247, "step": 641 }, { "epoch": 1.0, "grad_norm": 0.602288219366613, "learning_rate": 5e-06, "loss": 0.2401, "step": 642 }, { "epoch": 1.0015576323987538, "grad_norm": 0.5865025697495287, "learning_rate": 4.987766395953699e-06, "loss": 0.1907, "step": 643 }, { "epoch": 1.0031152647975077, "grad_norm": 0.6271078194072086, "learning_rate": 4.975532865143277e-06, "loss": 0.1837, "step": 644 }, { "epoch": 1.0046728971962617, "grad_norm": 0.5433097442594456, "learning_rate": 4.963299480804173e-06, "loss": 0.2122, "step": 645 }, { "epoch": 1.0062305295950156, "grad_norm": 0.6939024837652152, "learning_rate": 4.951066316170956e-06, "loss": 0.2464, "step": 646 }, { "epoch": 1.0077881619937694, "grad_norm": 0.6120591075726564, "learning_rate": 4.938833444476873e-06, "loss": 0.2943, "step": 647 }, { "epoch": 1.0093457943925233, "grad_norm": 0.5279986678775014, "learning_rate": 4.926600938953418e-06, "loss": 0.2139, "step": 648 }, { "epoch": 1.0109034267912773, "grad_norm": 0.5465124083851354, "learning_rate": 4.9143688728298946e-06, "loss": 0.2946, "step": 649 }, { "epoch": 1.0124610591900312, "grad_norm": 0.598599797464655, "learning_rate": 4.9021373193329775e-06, "loss": 0.2467, "step": 650 }, { "epoch": 1.014018691588785, "grad_norm": 0.5911973059402644, "learning_rate": 4.889906351686269e-06, "loss": 0.2439, "step": 651 }, { "epoch": 1.0155763239875388, "grad_norm": 0.5007183880780492, "learning_rate": 4.8776760431098665e-06, "loss": 0.2448, "step": 652 }, { "epoch": 1.017133956386293, "grad_norm": 0.6198547585957133, "learning_rate": 4.865446466819918e-06, "loss": 0.2107, "step": 653 }, { "epoch": 1.0186915887850467, "grad_norm": 0.6141060329229784, "learning_rate": 4.853217696028197e-06, "loss": 0.2124, "step": 654 }, { "epoch": 1.0202492211838006, "grad_norm": 0.5500970083546799, "learning_rate": 4.840989803941645e-06, "loss": 0.2413, "step": 655 }, { "epoch": 1.0218068535825544, "grad_norm": 0.5785491153029504, "learning_rate": 4.828762863761948e-06, "loss": 0.322, "step": 656 }, { "epoch": 1.0233644859813085, "grad_norm": 0.5158799094788992, "learning_rate": 4.816536948685091e-06, "loss": 0.2458, "step": 657 }, { "epoch": 1.0249221183800623, "grad_norm": 0.7110301489590548, "learning_rate": 4.804312131900929e-06, "loss": 0.1491, "step": 658 }, { "epoch": 1.0264797507788161, "grad_norm": 0.49142641785128816, "learning_rate": 4.792088486592731e-06, "loss": 0.2403, "step": 659 }, { "epoch": 1.02803738317757, "grad_norm": 0.6365446048004576, "learning_rate": 4.779866085936762e-06, "loss": 0.2723, "step": 660 }, { "epoch": 1.029595015576324, "grad_norm": 0.5622051291501104, "learning_rate": 4.767645003101831e-06, "loss": 0.1508, "step": 661 }, { "epoch": 1.0311526479750779, "grad_norm": 0.7978535815313595, "learning_rate": 4.755425311248863e-06, "loss": 0.7126, "step": 662 }, { "epoch": 1.0327102803738317, "grad_norm": 0.49688606974260013, "learning_rate": 4.7432070835304494e-06, "loss": 0.1821, "step": 663 }, { "epoch": 1.0342679127725856, "grad_norm": 0.9843736648935775, "learning_rate": 4.730990393090422e-06, "loss": 0.6915, "step": 664 }, { "epoch": 1.0358255451713396, "grad_norm": 0.5197585726631249, "learning_rate": 4.718775313063406e-06, "loss": 0.1985, "step": 665 }, { "epoch": 1.0373831775700935, "grad_norm": 0.7127374001801975, "learning_rate": 4.70656191657439e-06, "loss": 0.2209, "step": 666 }, { "epoch": 1.0389408099688473, "grad_norm": 0.5805515887865611, "learning_rate": 4.6943502767382815e-06, "loss": 0.1781, "step": 667 }, { "epoch": 1.0404984423676011, "grad_norm": 0.5812301249853751, "learning_rate": 4.6821404666594715e-06, "loss": 0.2362, "step": 668 }, { "epoch": 1.0420560747663552, "grad_norm": 0.5137436961715455, "learning_rate": 4.669932559431396e-06, "loss": 0.2381, "step": 669 }, { "epoch": 1.043613707165109, "grad_norm": 0.49254347681935357, "learning_rate": 4.657726628136105e-06, "loss": 0.2335, "step": 670 }, { "epoch": 1.0451713395638629, "grad_norm": 0.807702133830089, "learning_rate": 4.645522745843817e-06, "loss": 0.1636, "step": 671 }, { "epoch": 1.0467289719626167, "grad_norm": 0.6598785992415666, "learning_rate": 4.6333209856124814e-06, "loss": 0.2039, "step": 672 }, { "epoch": 1.0482866043613708, "grad_norm": 0.5735476926058175, "learning_rate": 4.621121420487345e-06, "loss": 0.2344, "step": 673 }, { "epoch": 1.0498442367601246, "grad_norm": 0.5406292279451508, "learning_rate": 4.608924123500519e-06, "loss": 0.1802, "step": 674 }, { "epoch": 1.0514018691588785, "grad_norm": 0.5983081571916266, "learning_rate": 4.596729167670529e-06, "loss": 0.1676, "step": 675 }, { "epoch": 1.0529595015576323, "grad_norm": 0.5810265835484077, "learning_rate": 4.58453662600189e-06, "loss": 0.2822, "step": 676 }, { "epoch": 1.0545171339563864, "grad_norm": 0.6855576677633736, "learning_rate": 4.572346571484661e-06, "loss": 0.1977, "step": 677 }, { "epoch": 1.0560747663551402, "grad_norm": 0.5335540200996262, "learning_rate": 4.5601590770940195e-06, "loss": 0.2152, "step": 678 }, { "epoch": 1.057632398753894, "grad_norm": 0.6453393789551976, "learning_rate": 4.547974215789808e-06, "loss": 0.1886, "step": 679 }, { "epoch": 1.0591900311526479, "grad_norm": 0.5915758123605618, "learning_rate": 4.535792060516112e-06, "loss": 0.2209, "step": 680 }, { "epoch": 1.060747663551402, "grad_norm": 0.6302498485071517, "learning_rate": 4.523612684200811e-06, "loss": 0.239, "step": 681 }, { "epoch": 1.0623052959501558, "grad_norm": 0.7064325822119844, "learning_rate": 4.511436159755159e-06, "loss": 0.1988, "step": 682 }, { "epoch": 1.0638629283489096, "grad_norm": 0.5229568055974967, "learning_rate": 4.499262560073328e-06, "loss": 0.2219, "step": 683 }, { "epoch": 1.0654205607476634, "grad_norm": 0.848853210404809, "learning_rate": 4.487091958031984e-06, "loss": 0.1813, "step": 684 }, { "epoch": 1.0669781931464175, "grad_norm": 0.62716757784117, "learning_rate": 4.474924426489847e-06, "loss": 0.2433, "step": 685 }, { "epoch": 1.0685358255451713, "grad_norm": 0.5961449452514619, "learning_rate": 4.46276003828726e-06, "loss": 0.2576, "step": 686 }, { "epoch": 1.0700934579439252, "grad_norm": 0.5336164214991632, "learning_rate": 4.450598866245743e-06, "loss": 0.2094, "step": 687 }, { "epoch": 1.071651090342679, "grad_norm": 0.5350296348658355, "learning_rate": 4.438440983167564e-06, "loss": 0.2883, "step": 688 }, { "epoch": 1.073208722741433, "grad_norm": 0.53810152034016, "learning_rate": 4.426286461835303e-06, "loss": 0.1968, "step": 689 }, { "epoch": 1.074766355140187, "grad_norm": 0.4674678946148812, "learning_rate": 4.414135375011416e-06, "loss": 0.1468, "step": 690 }, { "epoch": 1.0763239875389408, "grad_norm": 0.5672398628672646, "learning_rate": 4.401987795437797e-06, "loss": 0.1796, "step": 691 }, { "epoch": 1.0778816199376946, "grad_norm": 0.6409062495534954, "learning_rate": 4.3898437958353435e-06, "loss": 0.1536, "step": 692 }, { "epoch": 1.0794392523364487, "grad_norm": 0.7979975112099554, "learning_rate": 4.377703448903519e-06, "loss": 0.7814, "step": 693 }, { "epoch": 1.0809968847352025, "grad_norm": 0.6128214407671699, "learning_rate": 4.3655668273199305e-06, "loss": 0.249, "step": 694 }, { "epoch": 1.0825545171339563, "grad_norm": 0.6671186706045426, "learning_rate": 4.353434003739872e-06, "loss": 0.162, "step": 695 }, { "epoch": 1.0841121495327102, "grad_norm": 0.5657453591665842, "learning_rate": 4.341305050795907e-06, "loss": 0.2355, "step": 696 }, { "epoch": 1.0856697819314642, "grad_norm": 0.5876907165277706, "learning_rate": 4.329180041097425e-06, "loss": 0.2898, "step": 697 }, { "epoch": 1.087227414330218, "grad_norm": 0.5418335332942893, "learning_rate": 4.3170590472302125e-06, "loss": 0.2049, "step": 698 }, { "epoch": 1.088785046728972, "grad_norm": 0.6194906159202879, "learning_rate": 4.304942141756012e-06, "loss": 0.2716, "step": 699 }, { "epoch": 1.0903426791277258, "grad_norm": 0.6321743236263448, "learning_rate": 4.292829397212094e-06, "loss": 0.1858, "step": 700 }, { "epoch": 1.0919003115264798, "grad_norm": 0.5884478735032194, "learning_rate": 4.280720886110815e-06, "loss": 0.2181, "step": 701 }, { "epoch": 1.0934579439252337, "grad_norm": 0.512888102374138, "learning_rate": 4.268616680939197e-06, "loss": 0.2495, "step": 702 }, { "epoch": 1.0950155763239875, "grad_norm": 0.6806148269647997, "learning_rate": 4.256516854158476e-06, "loss": 0.1811, "step": 703 }, { "epoch": 1.0965732087227413, "grad_norm": 0.590183243846604, "learning_rate": 4.244421478203681e-06, "loss": 0.1759, "step": 704 }, { "epoch": 1.0981308411214954, "grad_norm": 0.7141876070652246, "learning_rate": 4.232330625483194e-06, "loss": 0.2155, "step": 705 }, { "epoch": 1.0996884735202492, "grad_norm": 0.5606859986060387, "learning_rate": 4.220244368378324e-06, "loss": 0.1748, "step": 706 }, { "epoch": 1.101246105919003, "grad_norm": 0.5374363163472021, "learning_rate": 4.208162779242862e-06, "loss": 0.15, "step": 707 }, { "epoch": 1.102803738317757, "grad_norm": 0.6337046437481759, "learning_rate": 4.19608593040266e-06, "loss": 0.2094, "step": 708 }, { "epoch": 1.104361370716511, "grad_norm": 0.6441679702023525, "learning_rate": 4.184013894155187e-06, "loss": 0.1347, "step": 709 }, { "epoch": 1.1059190031152648, "grad_norm": 0.5105399247363097, "learning_rate": 4.171946742769109e-06, "loss": 0.2492, "step": 710 }, { "epoch": 1.1074766355140186, "grad_norm": 0.835104618126293, "learning_rate": 4.1598845484838405e-06, "loss": 0.6552, "step": 711 }, { "epoch": 1.1090342679127725, "grad_norm": 0.5091044357624278, "learning_rate": 4.147827383509127e-06, "loss": 0.2459, "step": 712 }, { "epoch": 1.1105919003115265, "grad_norm": 0.4173053058204635, "learning_rate": 4.135775320024601e-06, "loss": 0.1834, "step": 713 }, { "epoch": 1.1121495327102804, "grad_norm": 0.516073956820414, "learning_rate": 4.123728430179363e-06, "loss": 0.2096, "step": 714 }, { "epoch": 1.1137071651090342, "grad_norm": 0.5229385439265322, "learning_rate": 4.111686786091534e-06, "loss": 0.1619, "step": 715 }, { "epoch": 1.115264797507788, "grad_norm": 0.697243292089673, "learning_rate": 4.099650459847835e-06, "loss": 0.2181, "step": 716 }, { "epoch": 1.1168224299065421, "grad_norm": 0.7935991971149668, "learning_rate": 4.087619523503149e-06, "loss": 0.7976, "step": 717 }, { "epoch": 1.118380062305296, "grad_norm": 0.7468568843345416, "learning_rate": 4.0755940490801e-06, "loss": 0.1844, "step": 718 }, { "epoch": 1.1199376947040498, "grad_norm": 0.655225509433663, "learning_rate": 4.0635741085686065e-06, "loss": 0.2016, "step": 719 }, { "epoch": 1.1214953271028036, "grad_norm": 0.6590392753571896, "learning_rate": 4.051559773925462e-06, "loss": 0.2175, "step": 720 }, { "epoch": 1.1230529595015577, "grad_norm": 0.5673140068671504, "learning_rate": 4.039551117073899e-06, "loss": 0.1377, "step": 721 }, { "epoch": 1.1246105919003115, "grad_norm": 0.5308315587462934, "learning_rate": 4.027548209903165e-06, "loss": 0.1628, "step": 722 }, { "epoch": 1.1261682242990654, "grad_norm": 0.5915750734678845, "learning_rate": 4.01555112426808e-06, "loss": 0.1005, "step": 723 }, { "epoch": 1.1277258566978192, "grad_norm": 0.6605084387776982, "learning_rate": 4.00355993198862e-06, "loss": 0.1835, "step": 724 }, { "epoch": 1.1292834890965733, "grad_norm": 0.5150684098328038, "learning_rate": 3.991574704849474e-06, "loss": 0.2452, "step": 725 }, { "epoch": 1.1308411214953271, "grad_norm": 0.5762133251543371, "learning_rate": 3.97959551459963e-06, "loss": 0.7117, "step": 726 }, { "epoch": 1.132398753894081, "grad_norm": 0.5992236027329398, "learning_rate": 3.967622432951931e-06, "loss": 0.2487, "step": 727 }, { "epoch": 1.1339563862928348, "grad_norm": 0.6328864728376592, "learning_rate": 3.955655531582649e-06, "loss": 0.237, "step": 728 }, { "epoch": 1.1355140186915889, "grad_norm": 0.5851714493036804, "learning_rate": 3.943694882131061e-06, "loss": 0.2293, "step": 729 }, { "epoch": 1.1370716510903427, "grad_norm": 0.6659466040289207, "learning_rate": 3.931740556199021e-06, "loss": 0.7144, "step": 730 }, { "epoch": 1.1386292834890965, "grad_norm": 0.8181017630361279, "learning_rate": 3.91979262535052e-06, "loss": 0.4389, "step": 731 }, { "epoch": 1.1401869158878504, "grad_norm": 0.6310519685249386, "learning_rate": 3.907851161111269e-06, "loss": 0.1969, "step": 732 }, { "epoch": 1.1417445482866044, "grad_norm": 0.5414429317292473, "learning_rate": 3.895916234968267e-06, "loss": 0.1969, "step": 733 }, { "epoch": 1.1433021806853583, "grad_norm": 0.5694107952685924, "learning_rate": 3.883987918369373e-06, "loss": 0.2409, "step": 734 }, { "epoch": 1.144859813084112, "grad_norm": 0.5303093457895504, "learning_rate": 3.8720662827228774e-06, "loss": 0.1599, "step": 735 }, { "epoch": 1.146417445482866, "grad_norm": 0.5798145908998529, "learning_rate": 3.860151399397077e-06, "loss": 0.2864, "step": 736 }, { "epoch": 1.14797507788162, "grad_norm": 0.7216318366333795, "learning_rate": 3.848243339719841e-06, "loss": 0.1274, "step": 737 }, { "epoch": 1.1495327102803738, "grad_norm": 0.5363385976748041, "learning_rate": 3.836342174978199e-06, "loss": 0.1679, "step": 738 }, { "epoch": 1.1510903426791277, "grad_norm": 0.5904714591263907, "learning_rate": 3.824447976417897e-06, "loss": 0.1803, "step": 739 }, { "epoch": 1.1526479750778815, "grad_norm": 0.7957820087907058, "learning_rate": 3.8125608152429777e-06, "loss": 0.2525, "step": 740 }, { "epoch": 1.1542056074766356, "grad_norm": 0.6852233460884398, "learning_rate": 3.8006807626153565e-06, "loss": 0.6679, "step": 741 }, { "epoch": 1.1557632398753894, "grad_norm": 0.7058769394932962, "learning_rate": 3.7888078896543984e-06, "loss": 0.1596, "step": 742 }, { "epoch": 1.1573208722741433, "grad_norm": 0.5829336602902025, "learning_rate": 3.776942267436482e-06, "loss": 0.1256, "step": 743 }, { "epoch": 1.158878504672897, "grad_norm": 0.4936789586579176, "learning_rate": 3.7650839669945804e-06, "loss": 0.2213, "step": 744 }, { "epoch": 1.1604361370716512, "grad_norm": 0.6225558777529572, "learning_rate": 3.7532330593178356e-06, "loss": 0.1667, "step": 745 }, { "epoch": 1.161993769470405, "grad_norm": 0.5614342757375034, "learning_rate": 3.741389615351136e-06, "loss": 0.1611, "step": 746 }, { "epoch": 1.1635514018691588, "grad_norm": 0.5754099281842366, "learning_rate": 3.729553705994685e-06, "loss": 0.2156, "step": 747 }, { "epoch": 1.1651090342679127, "grad_norm": 0.5573995209639809, "learning_rate": 3.7177254021035824e-06, "loss": 0.1777, "step": 748 }, { "epoch": 1.1666666666666667, "grad_norm": 0.6243355246035951, "learning_rate": 3.705904774487396e-06, "loss": 0.2092, "step": 749 }, { "epoch": 1.1682242990654206, "grad_norm": 0.7072395196568583, "learning_rate": 3.694091893909746e-06, "loss": 0.1916, "step": 750 }, { "epoch": 1.1697819314641744, "grad_norm": 0.5876756814093886, "learning_rate": 3.6822868310878683e-06, "loss": 0.1514, "step": 751 }, { "epoch": 1.1713395638629283, "grad_norm": 0.51963445472172, "learning_rate": 3.670489656692202e-06, "loss": 0.2446, "step": 752 }, { "epoch": 1.1728971962616823, "grad_norm": 0.6338437719274572, "learning_rate": 3.658700441345959e-06, "loss": 0.2041, "step": 753 }, { "epoch": 1.1744548286604362, "grad_norm": 0.7029067603645407, "learning_rate": 3.646919255624711e-06, "loss": 0.1852, "step": 754 }, { "epoch": 1.17601246105919, "grad_norm": 2.8451831521816353, "learning_rate": 3.6351461700559564e-06, "loss": 0.6003, "step": 755 }, { "epoch": 1.1775700934579438, "grad_norm": 0.5800825083133982, "learning_rate": 3.623381255118702e-06, "loss": 0.1983, "step": 756 }, { "epoch": 1.179127725856698, "grad_norm": 0.7108249246289187, "learning_rate": 3.6116245812430404e-06, "loss": 0.1904, "step": 757 }, { "epoch": 1.1806853582554517, "grad_norm": 0.5537550002098082, "learning_rate": 3.5998762188097364e-06, "loss": 0.2697, "step": 758 }, { "epoch": 1.1822429906542056, "grad_norm": 0.5354252171867416, "learning_rate": 3.588136238149791e-06, "loss": 0.2256, "step": 759 }, { "epoch": 1.1838006230529594, "grad_norm": 0.5746205607851972, "learning_rate": 3.5764047095440313e-06, "loss": 0.2216, "step": 760 }, { "epoch": 1.1853582554517135, "grad_norm": 0.6474421569647941, "learning_rate": 3.5646817032226855e-06, "loss": 0.18, "step": 761 }, { "epoch": 1.1869158878504673, "grad_norm": 0.6499250531970322, "learning_rate": 3.552967289364967e-06, "loss": 0.1564, "step": 762 }, { "epoch": 1.1884735202492211, "grad_norm": 0.6915421275577466, "learning_rate": 3.541261538098647e-06, "loss": 0.1937, "step": 763 }, { "epoch": 1.190031152647975, "grad_norm": 0.7484332653622916, "learning_rate": 3.529564519499641e-06, "loss": 0.8181, "step": 764 }, { "epoch": 1.191588785046729, "grad_norm": 0.5487998303234818, "learning_rate": 3.517876303591581e-06, "loss": 0.2187, "step": 765 }, { "epoch": 1.1931464174454829, "grad_norm": 0.5468867601412329, "learning_rate": 3.506196960345413e-06, "loss": 0.2741, "step": 766 }, { "epoch": 1.1947040498442367, "grad_norm": 0.5333599222192844, "learning_rate": 3.494526559678958e-06, "loss": 0.196, "step": 767 }, { "epoch": 1.1962616822429906, "grad_norm": 0.6163139809793089, "learning_rate": 3.4828651714565056e-06, "loss": 0.1689, "step": 768 }, { "epoch": 1.1978193146417446, "grad_norm": 0.6118788241013663, "learning_rate": 3.4712128654883915e-06, "loss": 0.6684, "step": 769 }, { "epoch": 1.1993769470404985, "grad_norm": 0.7588288371409719, "learning_rate": 3.459569711530586e-06, "loss": 0.1844, "step": 770 }, { "epoch": 1.2009345794392523, "grad_norm": 0.512022978994329, "learning_rate": 3.447935779284265e-06, "loss": 0.2842, "step": 771 }, { "epoch": 1.2024922118380061, "grad_norm": 0.5591196557510058, "learning_rate": 3.436311138395402e-06, "loss": 0.1664, "step": 772 }, { "epoch": 1.2040498442367602, "grad_norm": 0.8115349813699498, "learning_rate": 3.424695858454347e-06, "loss": 0.307, "step": 773 }, { "epoch": 1.205607476635514, "grad_norm": 0.5218970414061956, "learning_rate": 3.4130900089954142e-06, "loss": 0.2348, "step": 774 }, { "epoch": 1.2071651090342679, "grad_norm": 0.6595592104442388, "learning_rate": 3.4014936594964608e-06, "loss": 0.2381, "step": 775 }, { "epoch": 1.2087227414330217, "grad_norm": 0.5724303041429597, "learning_rate": 3.3899068793784717e-06, "loss": 0.2087, "step": 776 }, { "epoch": 1.2102803738317758, "grad_norm": 0.5855841777284334, "learning_rate": 3.378329738005144e-06, "loss": 0.2215, "step": 777 }, { "epoch": 1.2118380062305296, "grad_norm": 0.5794484038490832, "learning_rate": 3.3667623046824783e-06, "loss": 0.1275, "step": 778 }, { "epoch": 1.2133956386292835, "grad_norm": 0.8252092993882152, "learning_rate": 3.3552046486583547e-06, "loss": 0.2076, "step": 779 }, { "epoch": 1.2149532710280373, "grad_norm": 0.5054517090774381, "learning_rate": 3.3436568391221215e-06, "loss": 0.2269, "step": 780 }, { "epoch": 1.2165109034267914, "grad_norm": 0.6361245982706095, "learning_rate": 3.3321189452041814e-06, "loss": 0.1765, "step": 781 }, { "epoch": 1.2180685358255452, "grad_norm": 0.6359846065805195, "learning_rate": 3.3205910359755823e-06, "loss": 0.213, "step": 782 }, { "epoch": 1.219626168224299, "grad_norm": 0.575827824012367, "learning_rate": 3.309073180447593e-06, "loss": 0.2128, "step": 783 }, { "epoch": 1.2211838006230529, "grad_norm": 0.594472470089166, "learning_rate": 3.2975654475713005e-06, "loss": 0.1867, "step": 784 }, { "epoch": 1.222741433021807, "grad_norm": 0.6841555235014364, "learning_rate": 3.286067906237188e-06, "loss": 0.1615, "step": 785 }, { "epoch": 1.2242990654205608, "grad_norm": 0.5190015178190842, "learning_rate": 3.274580625274737e-06, "loss": 0.2139, "step": 786 }, { "epoch": 1.2258566978193146, "grad_norm": 0.675548730024154, "learning_rate": 3.263103673451996e-06, "loss": 0.2535, "step": 787 }, { "epoch": 1.2274143302180685, "grad_norm": 0.7466550657712839, "learning_rate": 3.2516371194751838e-06, "loss": 0.2706, "step": 788 }, { "epoch": 1.2289719626168225, "grad_norm": 0.6666523188287753, "learning_rate": 3.24018103198827e-06, "loss": 0.1727, "step": 789 }, { "epoch": 1.2305295950155763, "grad_norm": 0.6337742576262074, "learning_rate": 3.22873547957257e-06, "loss": 0.1979, "step": 790 }, { "epoch": 1.2320872274143302, "grad_norm": 0.5592619596660579, "learning_rate": 3.217300530746331e-06, "loss": 0.2195, "step": 791 }, { "epoch": 1.233644859813084, "grad_norm": 0.6098769962820866, "learning_rate": 3.2058762539643214e-06, "loss": 0.231, "step": 792 }, { "epoch": 1.235202492211838, "grad_norm": 0.5644158815873026, "learning_rate": 3.1944627176174204e-06, "loss": 0.2373, "step": 793 }, { "epoch": 1.236760124610592, "grad_norm": 0.9387436364990357, "learning_rate": 3.1830599900322135e-06, "loss": 0.8066, "step": 794 }, { "epoch": 1.2383177570093458, "grad_norm": 0.5202721107932754, "learning_rate": 3.1716681394705783e-06, "loss": 0.2198, "step": 795 }, { "epoch": 1.2398753894080996, "grad_norm": 0.5496594883977267, "learning_rate": 3.1602872341292772e-06, "loss": 0.1961, "step": 796 }, { "epoch": 1.2414330218068537, "grad_norm": 0.6098209049046375, "learning_rate": 3.148917342139548e-06, "loss": 0.2209, "step": 797 }, { "epoch": 1.2429906542056075, "grad_norm": 0.6346192193491992, "learning_rate": 3.1375585315667047e-06, "loss": 0.1802, "step": 798 }, { "epoch": 1.2445482866043613, "grad_norm": 0.765060316147878, "learning_rate": 3.1262108704097137e-06, "loss": 0.7635, "step": 799 }, { "epoch": 1.2461059190031152, "grad_norm": 0.5914294283180993, "learning_rate": 3.1148744266008024e-06, "loss": 0.1632, "step": 800 }, { "epoch": 1.2476635514018692, "grad_norm": 0.6570214224554458, "learning_rate": 3.1035492680050416e-06, "loss": 0.2406, "step": 801 }, { "epoch": 1.249221183800623, "grad_norm": 0.5452895252692876, "learning_rate": 3.0922354624199487e-06, "loss": 0.2458, "step": 802 }, { "epoch": 1.250778816199377, "grad_norm": 0.8021393101552716, "learning_rate": 3.080933077575071e-06, "loss": 0.2422, "step": 803 }, { "epoch": 1.2523364485981308, "grad_norm": 0.5980931498074656, "learning_rate": 3.0696421811315923e-06, "loss": 0.3071, "step": 804 }, { "epoch": 1.2538940809968846, "grad_norm": 0.7151706000107999, "learning_rate": 3.0583628406819167e-06, "loss": 0.1693, "step": 805 }, { "epoch": 1.2554517133956387, "grad_norm": 0.7557667483815892, "learning_rate": 3.0470951237492724e-06, "loss": 0.6472, "step": 806 }, { "epoch": 1.2570093457943925, "grad_norm": 0.5383018776710425, "learning_rate": 3.035839097787301e-06, "loss": 0.2179, "step": 807 }, { "epoch": 1.2585669781931463, "grad_norm": 0.5591749963571674, "learning_rate": 3.024594830179658e-06, "loss": 0.2537, "step": 808 }, { "epoch": 1.2601246105919004, "grad_norm": 0.5434411022961728, "learning_rate": 3.0133623882396067e-06, "loss": 0.1933, "step": 809 }, { "epoch": 1.2616822429906542, "grad_norm": 0.6201199414157127, "learning_rate": 3.0021418392096215e-06, "loss": 0.2193, "step": 810 }, { "epoch": 1.263239875389408, "grad_norm": 0.669872737304252, "learning_rate": 2.9909332502609744e-06, "loss": 0.1446, "step": 811 }, { "epoch": 1.264797507788162, "grad_norm": 0.5828212724381946, "learning_rate": 2.979736688493343e-06, "loss": 0.253, "step": 812 }, { "epoch": 1.2663551401869158, "grad_norm": 0.5644246073635343, "learning_rate": 2.968552220934402e-06, "loss": 0.2241, "step": 813 }, { "epoch": 1.2679127725856698, "grad_norm": 0.5607183966563147, "learning_rate": 2.9573799145394266e-06, "loss": 0.2745, "step": 814 }, { "epoch": 1.2694704049844237, "grad_norm": 0.474278573645314, "learning_rate": 2.9462198361908893e-06, "loss": 0.1337, "step": 815 }, { "epoch": 1.2710280373831775, "grad_norm": 1.3740104676242695, "learning_rate": 2.9350720526980592e-06, "loss": 0.7275, "step": 816 }, { "epoch": 1.2725856697819315, "grad_norm": 0.711919465609596, "learning_rate": 2.9239366307966033e-06, "loss": 0.7703, "step": 817 }, { "epoch": 1.2741433021806854, "grad_norm": 0.6630095165519386, "learning_rate": 2.912813637148187e-06, "loss": 0.1362, "step": 818 }, { "epoch": 1.2757009345794392, "grad_norm": 0.602978479954925, "learning_rate": 2.9017031383400772e-06, "loss": 0.1538, "step": 819 }, { "epoch": 1.277258566978193, "grad_norm": 0.7257094178132808, "learning_rate": 2.890605200884733e-06, "loss": 0.1533, "step": 820 }, { "epoch": 1.278816199376947, "grad_norm": 0.5041273431174441, "learning_rate": 2.879519891219424e-06, "loss": 0.195, "step": 821 }, { "epoch": 1.280373831775701, "grad_norm": 0.5315316010733497, "learning_rate": 2.868447275705821e-06, "loss": 0.1634, "step": 822 }, { "epoch": 1.2819314641744548, "grad_norm": 0.6131393193633405, "learning_rate": 2.8573874206296005e-06, "loss": 0.2044, "step": 823 }, { "epoch": 1.2834890965732086, "grad_norm": 0.5970183436040595, "learning_rate": 2.846340392200051e-06, "loss": 0.2539, "step": 824 }, { "epoch": 1.2850467289719627, "grad_norm": 0.5065607828613443, "learning_rate": 2.8353062565496715e-06, "loss": 0.166, "step": 825 }, { "epoch": 1.2866043613707165, "grad_norm": 0.6837585450327539, "learning_rate": 2.824285079733788e-06, "loss": 0.2406, "step": 826 }, { "epoch": 1.2881619937694704, "grad_norm": 0.5599124367538377, "learning_rate": 2.8132769277301374e-06, "loss": 0.157, "step": 827 }, { "epoch": 1.2897196261682242, "grad_norm": 0.49511182258884123, "learning_rate": 2.8022818664384945e-06, "loss": 0.1824, "step": 828 }, { "epoch": 1.291277258566978, "grad_norm": 0.5896227768783965, "learning_rate": 2.791299961680255e-06, "loss": 0.258, "step": 829 }, { "epoch": 1.2928348909657321, "grad_norm": 0.5730839746888057, "learning_rate": 2.7803312791980697e-06, "loss": 0.2379, "step": 830 }, { "epoch": 1.294392523364486, "grad_norm": 0.6736027436106719, "learning_rate": 2.769375884655421e-06, "loss": 0.2515, "step": 831 }, { "epoch": 1.2959501557632398, "grad_norm": 0.6242264709025079, "learning_rate": 2.758433843636252e-06, "loss": 0.1923, "step": 832 }, { "epoch": 1.2975077881619939, "grad_norm": 0.6286501649370759, "learning_rate": 2.7475052216445588e-06, "loss": 0.1842, "step": 833 }, { "epoch": 1.2990654205607477, "grad_norm": 0.5914028874575505, "learning_rate": 2.7365900841040104e-06, "loss": 0.1473, "step": 834 }, { "epoch": 1.3006230529595015, "grad_norm": 0.6185390608342917, "learning_rate": 2.7256884963575536e-06, "loss": 0.2347, "step": 835 }, { "epoch": 1.3021806853582554, "grad_norm": 0.5490349592608532, "learning_rate": 2.714800523667011e-06, "loss": 0.1875, "step": 836 }, { "epoch": 1.3037383177570092, "grad_norm": 0.619509184599072, "learning_rate": 2.703926231212708e-06, "loss": 0.2472, "step": 837 }, { "epoch": 1.3052959501557633, "grad_norm": 0.7158852617243849, "learning_rate": 2.6930656840930713e-06, "loss": 0.2468, "step": 838 }, { "epoch": 1.3068535825545171, "grad_norm": 0.723996693315863, "learning_rate": 2.6822189473242422e-06, "loss": 0.2276, "step": 839 }, { "epoch": 1.308411214953271, "grad_norm": 0.5632679246702721, "learning_rate": 2.671386085839682e-06, "loss": 0.2215, "step": 840 }, { "epoch": 1.309968847352025, "grad_norm": 0.5985281299233187, "learning_rate": 2.6605671644897967e-06, "loss": 0.3572, "step": 841 }, { "epoch": 1.3115264797507789, "grad_norm": 0.686365038362172, "learning_rate": 2.6497622480415346e-06, "loss": 0.3266, "step": 842 }, { "epoch": 1.3130841121495327, "grad_norm": 0.6582167035193466, "learning_rate": 2.6389714011780078e-06, "loss": 0.2669, "step": 843 }, { "epoch": 1.3146417445482865, "grad_norm": 0.5788593198598405, "learning_rate": 2.628194688498096e-06, "loss": 0.2572, "step": 844 }, { "epoch": 1.3161993769470404, "grad_norm": 0.5457859405003289, "learning_rate": 2.6174321745160707e-06, "loss": 0.2016, "step": 845 }, { "epoch": 1.3177570093457944, "grad_norm": 0.5093352344419979, "learning_rate": 2.606683923661204e-06, "loss": 0.2152, "step": 846 }, { "epoch": 1.3193146417445483, "grad_norm": 0.6332315738079737, "learning_rate": 2.5959500002773763e-06, "loss": 0.1449, "step": 847 }, { "epoch": 1.320872274143302, "grad_norm": 0.4685817262067038, "learning_rate": 2.585230468622705e-06, "loss": 0.2152, "step": 848 }, { "epoch": 1.3224299065420562, "grad_norm": 0.49981674536826676, "learning_rate": 2.574525392869143e-06, "loss": 0.1783, "step": 849 }, { "epoch": 1.32398753894081, "grad_norm": 1.0724140185603979, "learning_rate": 2.563834837102115e-06, "loss": 0.6521, "step": 850 }, { "epoch": 1.3255451713395638, "grad_norm": 0.5582924411201017, "learning_rate": 2.5531588653201134e-06, "loss": 0.1696, "step": 851 }, { "epoch": 1.3271028037383177, "grad_norm": 0.5064036048508646, "learning_rate": 2.542497541434329e-06, "loss": 0.1616, "step": 852 }, { "epoch": 1.3286604361370715, "grad_norm": 0.7424483742643145, "learning_rate": 2.531850929268258e-06, "loss": 0.3, "step": 853 }, { "epoch": 1.3302180685358256, "grad_norm": 0.6230465177616844, "learning_rate": 2.5212190925573366e-06, "loss": 0.638, "step": 854 }, { "epoch": 1.3317757009345794, "grad_norm": 0.5918692120070366, "learning_rate": 2.510602094948535e-06, "loss": 0.1742, "step": 855 }, { "epoch": 1.3333333333333333, "grad_norm": 0.6010294004442239, "learning_rate": 2.5000000000000015e-06, "loss": 0.2177, "step": 856 }, { "epoch": 1.3348909657320873, "grad_norm": 0.8792131655034126, "learning_rate": 2.4894128711806603e-06, "loss": 0.1769, "step": 857 }, { "epoch": 1.3364485981308412, "grad_norm": 0.5439058673773991, "learning_rate": 2.4788407718698487e-06, "loss": 0.1423, "step": 858 }, { "epoch": 1.338006230529595, "grad_norm": 0.6698324814866782, "learning_rate": 2.468283765356931e-06, "loss": 0.2016, "step": 859 }, { "epoch": 1.3395638629283488, "grad_norm": 0.7209231410413558, "learning_rate": 2.4577419148409123e-06, "loss": 0.2526, "step": 860 }, { "epoch": 1.3411214953271027, "grad_norm": 0.5554513132189719, "learning_rate": 2.447215283430072e-06, "loss": 0.2054, "step": 861 }, { "epoch": 1.3426791277258567, "grad_norm": 0.6204181777681318, "learning_rate": 2.436703934141583e-06, "loss": 0.2055, "step": 862 }, { "epoch": 1.3442367601246106, "grad_norm": 0.6174847596208328, "learning_rate": 2.4262079299011287e-06, "loss": 0.2214, "step": 863 }, { "epoch": 1.3457943925233644, "grad_norm": 0.596347448095447, "learning_rate": 2.4157273335425296e-06, "loss": 0.2396, "step": 864 }, { "epoch": 1.3473520249221185, "grad_norm": 0.6945784756475979, "learning_rate": 2.4052622078073714e-06, "loss": 0.6505, "step": 865 }, { "epoch": 1.3489096573208723, "grad_norm": 0.6484352021895371, "learning_rate": 2.394812615344622e-06, "loss": 0.2738, "step": 866 }, { "epoch": 1.3504672897196262, "grad_norm": 0.5351329498045785, "learning_rate": 2.384378618710265e-06, "loss": 0.1667, "step": 867 }, { "epoch": 1.35202492211838, "grad_norm": 0.5884847134807437, "learning_rate": 2.3739602803669126e-06, "loss": 0.2267, "step": 868 }, { "epoch": 1.3535825545171338, "grad_norm": 0.7327148616338638, "learning_rate": 2.363557662683446e-06, "loss": 0.7327, "step": 869 }, { "epoch": 1.355140186915888, "grad_norm": 0.712258388986809, "learning_rate": 2.3531708279346347e-06, "loss": 0.2281, "step": 870 }, { "epoch": 1.3566978193146417, "grad_norm": 0.598978404959667, "learning_rate": 2.3427998383007605e-06, "loss": 0.2433, "step": 871 }, { "epoch": 1.3582554517133956, "grad_norm": 0.6295698566133894, "learning_rate": 2.3324447558672543e-06, "loss": 0.2569, "step": 872 }, { "epoch": 1.3598130841121496, "grad_norm": 0.5986998801181967, "learning_rate": 2.3221056426243112e-06, "loss": 0.1962, "step": 873 }, { "epoch": 1.3613707165109035, "grad_norm": 0.5674067782697227, "learning_rate": 2.3117825604665405e-06, "loss": 0.2158, "step": 874 }, { "epoch": 1.3629283489096573, "grad_norm": 0.6884839665928056, "learning_rate": 2.3014755711925695e-06, "loss": 0.2329, "step": 875 }, { "epoch": 1.3644859813084111, "grad_norm": 0.6885485571502845, "learning_rate": 2.291184736504695e-06, "loss": 0.1955, "step": 876 }, { "epoch": 1.366043613707165, "grad_norm": 0.5473669303308986, "learning_rate": 2.2809101180084963e-06, "loss": 0.2171, "step": 877 }, { "epoch": 1.367601246105919, "grad_norm": 0.5860469422390239, "learning_rate": 2.2706517772124875e-06, "loss": 0.3142, "step": 878 }, { "epoch": 1.3691588785046729, "grad_norm": 0.4589130720603424, "learning_rate": 2.2604097755277244e-06, "loss": 0.1892, "step": 879 }, { "epoch": 1.3707165109034267, "grad_norm": 0.56817784386682, "learning_rate": 2.250184174267459e-06, "loss": 0.2104, "step": 880 }, { "epoch": 1.3722741433021808, "grad_norm": 2.9170440438070244, "learning_rate": 2.239975034646756e-06, "loss": 0.4075, "step": 881 }, { "epoch": 1.3738317757009346, "grad_norm": 0.5936408877749505, "learning_rate": 2.2297824177821374e-06, "loss": 0.1725, "step": 882 }, { "epoch": 1.3753894080996885, "grad_norm": 0.5570669743525561, "learning_rate": 2.2196063846912142e-06, "loss": 0.1089, "step": 883 }, { "epoch": 1.3769470404984423, "grad_norm": 0.7043574775467037, "learning_rate": 2.209446996292312e-06, "loss": 0.2412, "step": 884 }, { "epoch": 1.3785046728971961, "grad_norm": 0.6356878135461869, "learning_rate": 2.1993043134041213e-06, "loss": 0.2314, "step": 885 }, { "epoch": 1.3800623052959502, "grad_norm": 0.6088173529635026, "learning_rate": 2.1891783967453235e-06, "loss": 0.1986, "step": 886 }, { "epoch": 1.381619937694704, "grad_norm": 0.6505134414780895, "learning_rate": 2.1790693069342293e-06, "loss": 0.2621, "step": 887 }, { "epoch": 1.3831775700934579, "grad_norm": 0.5229082442758375, "learning_rate": 2.168977104488415e-06, "loss": 0.1857, "step": 888 }, { "epoch": 1.384735202492212, "grad_norm": 0.6219752894683416, "learning_rate": 2.158901849824363e-06, "loss": 0.2393, "step": 889 }, { "epoch": 1.3862928348909658, "grad_norm": 0.6712717762459158, "learning_rate": 2.1488436032571e-06, "loss": 0.1975, "step": 890 }, { "epoch": 1.3878504672897196, "grad_norm": 0.5882444001428154, "learning_rate": 2.138802424999833e-06, "loss": 0.1938, "step": 891 }, { "epoch": 1.3894080996884735, "grad_norm": 0.7887785551105917, "learning_rate": 2.128778375163588e-06, "loss": 0.2438, "step": 892 }, { "epoch": 1.3909657320872273, "grad_norm": 0.6726478463164812, "learning_rate": 2.1187715137568564e-06, "loss": 0.1609, "step": 893 }, { "epoch": 1.3925233644859814, "grad_norm": 1.1772556659919378, "learning_rate": 2.1087819006852327e-06, "loss": 0.1874, "step": 894 }, { "epoch": 1.3940809968847352, "grad_norm": 0.6141735195326121, "learning_rate": 2.098809595751049e-06, "loss": 0.1771, "step": 895 }, { "epoch": 1.395638629283489, "grad_norm": 0.574616048911299, "learning_rate": 2.08885465865303e-06, "loss": 0.1973, "step": 896 }, { "epoch": 1.397196261682243, "grad_norm": 0.7291041551854346, "learning_rate": 2.078917148985921e-06, "loss": 0.7534, "step": 897 }, { "epoch": 1.398753894080997, "grad_norm": 0.5907252824302475, "learning_rate": 2.06899712624015e-06, "loss": 0.2802, "step": 898 }, { "epoch": 1.4003115264797508, "grad_norm": 0.5328199626427012, "learning_rate": 2.0590946498014493e-06, "loss": 0.2303, "step": 899 }, { "epoch": 1.4018691588785046, "grad_norm": 0.6150130691322249, "learning_rate": 2.049209778950518e-06, "loss": 0.2424, "step": 900 }, { "epoch": 1.4034267912772584, "grad_norm": 0.6127344165171512, "learning_rate": 2.0393425728626526e-06, "loss": 0.1674, "step": 901 }, { "epoch": 1.4049844236760125, "grad_norm": 0.6455010665412071, "learning_rate": 2.029493090607413e-06, "loss": 0.201, "step": 902 }, { "epoch": 1.4065420560747663, "grad_norm": 0.6680633935331085, "learning_rate": 2.0196613911482445e-06, "loss": 0.1652, "step": 903 }, { "epoch": 1.4080996884735202, "grad_norm": 0.5230495421575583, "learning_rate": 2.0098475333421426e-06, "loss": 0.1869, "step": 904 }, { "epoch": 1.4096573208722742, "grad_norm": 0.6484127593933029, "learning_rate": 2.0000515759392904e-06, "loss": 0.2167, "step": 905 }, { "epoch": 1.411214953271028, "grad_norm": 0.6594541288417016, "learning_rate": 1.9902735775827146e-06, "loss": 0.3135, "step": 906 }, { "epoch": 1.412772585669782, "grad_norm": 0.6546159891601314, "learning_rate": 1.980513596807932e-06, "loss": 0.1546, "step": 907 }, { "epoch": 1.4143302180685358, "grad_norm": 0.6853904172880749, "learning_rate": 1.9707716920425923e-06, "loss": 0.2063, "step": 908 }, { "epoch": 1.4158878504672896, "grad_norm": 0.6131385223592029, "learning_rate": 1.9610479216061393e-06, "loss": 0.2792, "step": 909 }, { "epoch": 1.4174454828660437, "grad_norm": 0.47832488985489935, "learning_rate": 1.951342343709456e-06, "loss": 0.2819, "step": 910 }, { "epoch": 1.4190031152647975, "grad_norm": 0.7260584028544175, "learning_rate": 1.941655016454515e-06, "loss": 0.1664, "step": 911 }, { "epoch": 1.4205607476635513, "grad_norm": 0.531050945261339, "learning_rate": 1.9319859978340312e-06, "loss": 0.218, "step": 912 }, { "epoch": 1.4221183800623054, "grad_norm": 0.7077461606190044, "learning_rate": 1.9223353457311178e-06, "loss": 0.2129, "step": 913 }, { "epoch": 1.4236760124610592, "grad_norm": 0.6914074130890262, "learning_rate": 1.912703117918938e-06, "loss": 0.1846, "step": 914 }, { "epoch": 1.425233644859813, "grad_norm": 0.7199368018705896, "learning_rate": 1.9030893720603604e-06, "loss": 0.2253, "step": 915 }, { "epoch": 1.426791277258567, "grad_norm": 0.5583301531155793, "learning_rate": 1.8934941657076057e-06, "loss": 0.2059, "step": 916 }, { "epoch": 1.4283489096573208, "grad_norm": 0.5646047767722963, "learning_rate": 1.8839175563019145e-06, "loss": 0.2514, "step": 917 }, { "epoch": 1.4299065420560748, "grad_norm": 0.6817330899431177, "learning_rate": 1.8743596011731984e-06, "loss": 0.2018, "step": 918 }, { "epoch": 1.4314641744548287, "grad_norm": 0.6478720709719429, "learning_rate": 1.8648203575396912e-06, "loss": 0.1172, "step": 919 }, { "epoch": 1.4330218068535825, "grad_norm": 0.6481753359929265, "learning_rate": 1.8552998825076163e-06, "loss": 0.2396, "step": 920 }, { "epoch": 1.4345794392523366, "grad_norm": 0.6132125740119907, "learning_rate": 1.845798233070833e-06, "loss": 0.2065, "step": 921 }, { "epoch": 1.4361370716510904, "grad_norm": 0.5448210616281429, "learning_rate": 1.8363154661105131e-06, "loss": 0.2022, "step": 922 }, { "epoch": 1.4376947040498442, "grad_norm": 0.5931140592473454, "learning_rate": 1.826851638394777e-06, "loss": 0.2241, "step": 923 }, { "epoch": 1.439252336448598, "grad_norm": 0.5732273344982183, "learning_rate": 1.8174068065783768e-06, "loss": 0.238, "step": 924 }, { "epoch": 1.440809968847352, "grad_norm": 0.6031290862741796, "learning_rate": 1.8079810272023373e-06, "loss": 0.2065, "step": 925 }, { "epoch": 1.442367601246106, "grad_norm": 0.7027323074692515, "learning_rate": 1.7985743566936338e-06, "loss": 0.1353, "step": 926 }, { "epoch": 1.4439252336448598, "grad_norm": 0.679661751499936, "learning_rate": 1.7891868513648436e-06, "loss": 0.1636, "step": 927 }, { "epoch": 1.4454828660436136, "grad_norm": 0.6123098696997167, "learning_rate": 1.7798185674138163e-06, "loss": 0.7362, "step": 928 }, { "epoch": 1.4470404984423677, "grad_norm": 0.6533459270958951, "learning_rate": 1.7704695609233275e-06, "loss": 0.2479, "step": 929 }, { "epoch": 1.4485981308411215, "grad_norm": 0.6018511814849575, "learning_rate": 1.7611398878607544e-06, "loss": 0.2135, "step": 930 }, { "epoch": 1.4501557632398754, "grad_norm": 0.6980529317039801, "learning_rate": 1.7518296040777355e-06, "loss": 0.1639, "step": 931 }, { "epoch": 1.4517133956386292, "grad_norm": 0.6062016861868827, "learning_rate": 1.742538765309832e-06, "loss": 0.2538, "step": 932 }, { "epoch": 1.453271028037383, "grad_norm": 0.6539624285807091, "learning_rate": 1.7332674271762018e-06, "loss": 0.1811, "step": 933 }, { "epoch": 1.4548286604361371, "grad_norm": 0.7122687298587204, "learning_rate": 1.724015645179264e-06, "loss": 0.1891, "step": 934 }, { "epoch": 1.456386292834891, "grad_norm": 0.7014287750312252, "learning_rate": 1.7147834747043651e-06, "loss": 0.1866, "step": 935 }, { "epoch": 1.4579439252336448, "grad_norm": 0.6319275863012598, "learning_rate": 1.7055709710194452e-06, "loss": 0.1921, "step": 936 }, { "epoch": 1.4595015576323989, "grad_norm": 0.6620029301577026, "learning_rate": 1.6963781892747128e-06, "loss": 0.1889, "step": 937 }, { "epoch": 1.4610591900311527, "grad_norm": 0.6613697258053355, "learning_rate": 1.6872051845023124e-06, "loss": 0.2277, "step": 938 }, { "epoch": 1.4626168224299065, "grad_norm": 0.6709827699093153, "learning_rate": 1.6780520116159954e-06, "loss": 0.2345, "step": 939 }, { "epoch": 1.4641744548286604, "grad_norm": 0.6308994122097046, "learning_rate": 1.6689187254107852e-06, "loss": 0.2384, "step": 940 }, { "epoch": 1.4657320872274142, "grad_norm": 0.49997735693518447, "learning_rate": 1.6598053805626607e-06, "loss": 0.2885, "step": 941 }, { "epoch": 1.4672897196261683, "grad_norm": 0.5769855416317906, "learning_rate": 1.6507120316282222e-06, "loss": 0.2262, "step": 942 }, { "epoch": 1.4688473520249221, "grad_norm": 0.7766043611879518, "learning_rate": 1.6416387330443594e-06, "loss": 0.2012, "step": 943 }, { "epoch": 1.470404984423676, "grad_norm": 0.6306011227357949, "learning_rate": 1.6325855391279399e-06, "loss": 0.1672, "step": 944 }, { "epoch": 1.47196261682243, "grad_norm": 0.6252880509484167, "learning_rate": 1.623552504075467e-06, "loss": 0.21, "step": 945 }, { "epoch": 1.4735202492211839, "grad_norm": 0.5733165288371812, "learning_rate": 1.6145396819627734e-06, "loss": 0.203, "step": 946 }, { "epoch": 1.4750778816199377, "grad_norm": 0.7571650504143702, "learning_rate": 1.6055471267446804e-06, "loss": 0.2244, "step": 947 }, { "epoch": 1.4766355140186915, "grad_norm": 0.7478203882298783, "learning_rate": 1.5965748922546876e-06, "loss": 0.1799, "step": 948 }, { "epoch": 1.4781931464174454, "grad_norm": 0.6868312970173284, "learning_rate": 1.5876230322046393e-06, "loss": 0.6746, "step": 949 }, { "epoch": 1.4797507788161994, "grad_norm": 0.6037632401087822, "learning_rate": 1.578691600184416e-06, "loss": 0.2057, "step": 950 }, { "epoch": 1.4813084112149533, "grad_norm": 0.6046826777481786, "learning_rate": 1.569780649661603e-06, "loss": 0.2926, "step": 951 }, { "epoch": 1.482866043613707, "grad_norm": 0.6848611609536812, "learning_rate": 1.5608902339811765e-06, "loss": 0.2328, "step": 952 }, { "epoch": 1.4844236760124612, "grad_norm": 0.5462582273781476, "learning_rate": 1.5520204063651784e-06, "loss": 0.2703, "step": 953 }, { "epoch": 1.485981308411215, "grad_norm": 0.5459862014182448, "learning_rate": 1.5431712199124033e-06, "loss": 0.2268, "step": 954 }, { "epoch": 1.4875389408099688, "grad_norm": 0.5909228089011523, "learning_rate": 1.5343427275980804e-06, "loss": 0.1918, "step": 955 }, { "epoch": 1.4890965732087227, "grad_norm": 0.6922392218824316, "learning_rate": 1.5255349822735494e-06, "loss": 0.2065, "step": 956 }, { "epoch": 1.4906542056074765, "grad_norm": 0.6347770175659823, "learning_rate": 1.5167480366659538e-06, "loss": 0.1683, "step": 957 }, { "epoch": 1.4922118380062306, "grad_norm": 0.5516070303093964, "learning_rate": 1.507981943377918e-06, "loss": 0.2173, "step": 958 }, { "epoch": 1.4937694704049844, "grad_norm": 0.6071049706243482, "learning_rate": 1.4992367548872383e-06, "loss": 0.2458, "step": 959 }, { "epoch": 1.4953271028037383, "grad_norm": 0.6720642183082266, "learning_rate": 1.490512523546559e-06, "loss": 0.1915, "step": 960 }, { "epoch": 1.4968847352024923, "grad_norm": 0.5147991401453758, "learning_rate": 1.481809301583072e-06, "loss": 0.2315, "step": 961 }, { "epoch": 1.4984423676012462, "grad_norm": 0.5193622138782028, "learning_rate": 1.4731271410981961e-06, "loss": 0.2994, "step": 962 }, { "epoch": 1.5, "grad_norm": 0.6483302843412619, "learning_rate": 1.4644660940672628e-06, "loss": 0.1486, "step": 963 }, { "epoch": 1.5015576323987538, "grad_norm": 0.6211856895290482, "learning_rate": 1.4558262123392125e-06, "loss": 0.1979, "step": 964 }, { "epoch": 1.5031152647975077, "grad_norm": 0.6267158569891346, "learning_rate": 1.4472075476362819e-06, "loss": 0.226, "step": 965 }, { "epoch": 1.5046728971962615, "grad_norm": 0.6878081812868871, "learning_rate": 1.4386101515536865e-06, "loss": 0.215, "step": 966 }, { "epoch": 1.5062305295950156, "grad_norm": 0.6013152247422209, "learning_rate": 1.4300340755593256e-06, "loss": 0.1853, "step": 967 }, { "epoch": 1.5077881619937694, "grad_norm": 0.5961333640534994, "learning_rate": 1.4214793709934644e-06, "loss": 0.2134, "step": 968 }, { "epoch": 1.5093457943925235, "grad_norm": 0.6714185068729337, "learning_rate": 1.4129460890684255e-06, "loss": 0.2285, "step": 969 }, { "epoch": 1.5109034267912773, "grad_norm": 0.5287976544223729, "learning_rate": 1.4044342808682904e-06, "loss": 0.226, "step": 970 }, { "epoch": 1.5124610591900312, "grad_norm": 0.7244263964282488, "learning_rate": 1.3959439973485872e-06, "loss": 0.1919, "step": 971 }, { "epoch": 1.514018691588785, "grad_norm": 0.7411473501185608, "learning_rate": 1.38747528933599e-06, "loss": 0.1573, "step": 972 }, { "epoch": 1.5155763239875388, "grad_norm": 0.7535346088046625, "learning_rate": 1.3790282075280064e-06, "loss": 0.1939, "step": 973 }, { "epoch": 1.5171339563862927, "grad_norm": 0.7535256340431424, "learning_rate": 1.3706028024926855e-06, "loss": 0.1736, "step": 974 }, { "epoch": 1.5186915887850467, "grad_norm": 0.5709920085521653, "learning_rate": 1.3621991246683093e-06, "loss": 0.222, "step": 975 }, { "epoch": 1.5202492211838006, "grad_norm": 0.7297081050093409, "learning_rate": 1.3538172243630848e-06, "loss": 0.1788, "step": 976 }, { "epoch": 1.5218068535825546, "grad_norm": 0.6180657113624616, "learning_rate": 1.3454571517548554e-06, "loss": 0.177, "step": 977 }, { "epoch": 1.5233644859813085, "grad_norm": 0.591364040153287, "learning_rate": 1.3371189568907915e-06, "loss": 0.2128, "step": 978 }, { "epoch": 1.5249221183800623, "grad_norm": 0.5022261849806481, "learning_rate": 1.3288026896870964e-06, "loss": 0.2267, "step": 979 }, { "epoch": 1.5264797507788161, "grad_norm": 0.6203532747393247, "learning_rate": 1.3205083999286973e-06, "loss": 0.1955, "step": 980 }, { "epoch": 1.52803738317757, "grad_norm": 0.6729761565050358, "learning_rate": 1.3122361372689624e-06, "loss": 0.1657, "step": 981 }, { "epoch": 1.5295950155763238, "grad_norm": 0.6536059366129146, "learning_rate": 1.3039859512293885e-06, "loss": 0.2198, "step": 982 }, { "epoch": 1.5311526479750779, "grad_norm": 0.695539877357984, "learning_rate": 1.2957578911993212e-06, "loss": 0.2023, "step": 983 }, { "epoch": 1.5327102803738317, "grad_norm": 0.4424609062997814, "learning_rate": 1.28755200643564e-06, "loss": 0.2238, "step": 984 }, { "epoch": 1.5342679127725858, "grad_norm": 0.5843664425742057, "learning_rate": 1.27936834606248e-06, "loss": 0.2197, "step": 985 }, { "epoch": 1.5358255451713396, "grad_norm": 0.6148367721669108, "learning_rate": 1.2712069590709265e-06, "loss": 0.2004, "step": 986 }, { "epoch": 1.5373831775700935, "grad_norm": 0.562472309572901, "learning_rate": 1.2630678943187292e-06, "loss": 0.1835, "step": 987 }, { "epoch": 1.5389408099688473, "grad_norm": 0.7987485306921659, "learning_rate": 1.2549512005300068e-06, "loss": 0.1995, "step": 988 }, { "epoch": 1.5404984423676011, "grad_norm": 0.6656097374020112, "learning_rate": 1.246856926294953e-06, "loss": 0.1874, "step": 989 }, { "epoch": 1.542056074766355, "grad_norm": 0.7982996720144556, "learning_rate": 1.23878512006955e-06, "loss": 0.351, "step": 990 }, { "epoch": 1.543613707165109, "grad_norm": 0.7237322332450993, "learning_rate": 1.2307358301752753e-06, "loss": 0.6736, "step": 991 }, { "epoch": 1.5451713395638629, "grad_norm": 0.6606058936315011, "learning_rate": 1.222709104798816e-06, "loss": 0.2446, "step": 992 }, { "epoch": 1.546728971962617, "grad_norm": 0.7491652827965601, "learning_rate": 1.214704991991773e-06, "loss": 0.1912, "step": 993 }, { "epoch": 1.5482866043613708, "grad_norm": 0.5316071822548012, "learning_rate": 1.2067235396703819e-06, "loss": 0.2006, "step": 994 }, { "epoch": 1.5498442367601246, "grad_norm": 0.4870862919772457, "learning_rate": 1.198764795615221e-06, "loss": 0.1721, "step": 995 }, { "epoch": 1.5514018691588785, "grad_norm": 0.45786723638909316, "learning_rate": 1.190828807470929e-06, "loss": 0.2847, "step": 996 }, { "epoch": 1.5529595015576323, "grad_norm": 0.5220478168315765, "learning_rate": 1.182915622745912e-06, "loss": 0.1942, "step": 997 }, { "epoch": 1.5545171339563861, "grad_norm": 1.0355443091689227, "learning_rate": 1.17502528881207e-06, "loss": 0.6648, "step": 998 }, { "epoch": 1.5560747663551402, "grad_norm": 0.6645917404798355, "learning_rate": 1.1671578529045075e-06, "loss": 0.1686, "step": 999 }, { "epoch": 1.557632398753894, "grad_norm": 0.5772980090930082, "learning_rate": 1.1593133621212454e-06, "loss": 0.1601, "step": 1000 }, { "epoch": 1.557632398753894, "eval_loss": 0.36974087357521057, "eval_runtime": 2.7835, "eval_samples_per_second": 9.341, "eval_steps_per_second": 2.515, "step": 1000 }, { "epoch": 1.559190031152648, "grad_norm": 0.6634169549696792, "learning_rate": 1.151491863422951e-06, "loss": 0.178, "step": 1001 }, { "epoch": 1.560747663551402, "grad_norm": 0.6293088334490837, "learning_rate": 1.1436934036326485e-06, "loss": 0.2054, "step": 1002 }, { "epoch": 1.5623052959501558, "grad_norm": 0.6633410484590091, "learning_rate": 1.135918029435441e-06, "loss": 0.2654, "step": 1003 }, { "epoch": 1.5638629283489096, "grad_norm": 0.6622964143754947, "learning_rate": 1.1281657873782287e-06, "loss": 0.2793, "step": 1004 }, { "epoch": 1.5654205607476634, "grad_norm": 0.6119616717415093, "learning_rate": 1.1204367238694357e-06, "loss": 0.2711, "step": 1005 }, { "epoch": 1.5669781931464173, "grad_norm": 0.47949709495438925, "learning_rate": 1.1127308851787244e-06, "loss": 0.2021, "step": 1006 }, { "epoch": 1.5685358255451713, "grad_norm": 0.6934647914219084, "learning_rate": 1.1050483174367304e-06, "loss": 0.203, "step": 1007 }, { "epoch": 1.5700934579439252, "grad_norm": 0.5481726220781126, "learning_rate": 1.0973890666347703e-06, "loss": 0.2368, "step": 1008 }, { "epoch": 1.5716510903426792, "grad_norm": 0.6509857110936647, "learning_rate": 1.0897531786245819e-06, "loss": 0.2125, "step": 1009 }, { "epoch": 1.573208722741433, "grad_norm": 0.5617926695948265, "learning_rate": 1.0821406991180367e-06, "loss": 0.2398, "step": 1010 }, { "epoch": 1.574766355140187, "grad_norm": 0.6111199895380683, "learning_rate": 1.0745516736868766e-06, "loss": 0.2485, "step": 1011 }, { "epoch": 1.5763239875389408, "grad_norm": 0.7124396857915882, "learning_rate": 1.0669861477624376e-06, "loss": 0.2517, "step": 1012 }, { "epoch": 1.5778816199376946, "grad_norm": 0.5990035454998449, "learning_rate": 1.0594441666353722e-06, "loss": 0.151, "step": 1013 }, { "epoch": 1.5794392523364484, "grad_norm": 0.6520260923864126, "learning_rate": 1.0519257754553868e-06, "loss": 0.1888, "step": 1014 }, { "epoch": 1.5809968847352025, "grad_norm": 0.7169934380455352, "learning_rate": 1.0444310192309675e-06, "loss": 0.2318, "step": 1015 }, { "epoch": 1.5825545171339563, "grad_norm": 0.5993914903355098, "learning_rate": 1.0369599428291116e-06, "loss": 0.2124, "step": 1016 }, { "epoch": 1.5841121495327104, "grad_norm": 0.49513044936794753, "learning_rate": 1.0295125909750537e-06, "loss": 0.2402, "step": 1017 }, { "epoch": 1.5856697819314642, "grad_norm": 0.7226758243950167, "learning_rate": 1.0220890082520074e-06, "loss": 0.2398, "step": 1018 }, { "epoch": 1.587227414330218, "grad_norm": 0.6288423009834784, "learning_rate": 1.0146892391008929e-06, "loss": 0.1948, "step": 1019 }, { "epoch": 1.588785046728972, "grad_norm": 0.596604250578145, "learning_rate": 1.0073133278200702e-06, "loss": 0.232, "step": 1020 }, { "epoch": 1.5903426791277258, "grad_norm": 0.6135147035860976, "learning_rate": 9.999613185650748e-07, "loss": 0.2614, "step": 1021 }, { "epoch": 1.5919003115264796, "grad_norm": 0.5404711008426897, "learning_rate": 9.926332553483547e-07, "loss": 0.1885, "step": 1022 }, { "epoch": 1.5934579439252337, "grad_norm": 0.7031398847787177, "learning_rate": 9.853291820390087e-07, "loss": 0.2325, "step": 1023 }, { "epoch": 1.5950155763239875, "grad_norm": 0.6348294111355727, "learning_rate": 9.780491423625154e-07, "loss": 0.2385, "step": 1024 }, { "epoch": 1.5965732087227416, "grad_norm": 0.5521768442130486, "learning_rate": 9.70793179900484e-07, "loss": 0.1988, "step": 1025 }, { "epoch": 1.5981308411214954, "grad_norm": 0.613980141544231, "learning_rate": 9.63561338090378e-07, "loss": 0.1027, "step": 1026 }, { "epoch": 1.5996884735202492, "grad_norm": 0.7720687456198224, "learning_rate": 9.563536602252749e-07, "loss": 0.2905, "step": 1027 }, { "epoch": 1.601246105919003, "grad_norm": 0.6319851908391722, "learning_rate": 9.49170189453586e-07, "loss": 0.2617, "step": 1028 }, { "epoch": 1.602803738317757, "grad_norm": 0.6736383488893641, "learning_rate": 9.420109687788148e-07, "loss": 0.2641, "step": 1029 }, { "epoch": 1.6043613707165107, "grad_norm": 0.6656334468054872, "learning_rate": 9.348760410592855e-07, "loss": 0.2171, "step": 1030 }, { "epoch": 1.6059190031152648, "grad_norm": 0.6957076685070199, "learning_rate": 9.277654490079035e-07, "loss": 0.236, "step": 1031 }, { "epoch": 1.6074766355140186, "grad_norm": 0.6288117317451123, "learning_rate": 9.206792351918809e-07, "loss": 0.1562, "step": 1032 }, { "epoch": 1.6090342679127727, "grad_norm": 0.6616920441316214, "learning_rate": 9.136174420324962e-07, "loss": 0.218, "step": 1033 }, { "epoch": 1.6105919003115265, "grad_norm": 0.6342551825259773, "learning_rate": 9.065801118048312e-07, "loss": 0.2252, "step": 1034 }, { "epoch": 1.6121495327102804, "grad_norm": 0.5589154301426593, "learning_rate": 8.995672866375237e-07, "loss": 0.1551, "step": 1035 }, { "epoch": 1.6137071651090342, "grad_norm": 0.4635682927354973, "learning_rate": 8.925790085125152e-07, "loss": 0.1833, "step": 1036 }, { "epoch": 1.615264797507788, "grad_norm": 0.5543844541586774, "learning_rate": 8.856153192647909e-07, "loss": 0.2067, "step": 1037 }, { "epoch": 1.616822429906542, "grad_norm": 0.6110967305529611, "learning_rate": 8.786762605821419e-07, "loss": 0.2331, "step": 1038 }, { "epoch": 1.618380062305296, "grad_norm": 0.6930362377044883, "learning_rate": 8.717618740049083e-07, "loss": 0.2451, "step": 1039 }, { "epoch": 1.6199376947040498, "grad_norm": 0.8137005439911986, "learning_rate": 8.648722009257315e-07, "loss": 0.1626, "step": 1040 }, { "epoch": 1.6214953271028039, "grad_norm": 0.7004708747608873, "learning_rate": 8.580072825893049e-07, "loss": 0.1831, "step": 1041 }, { "epoch": 1.6230529595015577, "grad_norm": 0.607273962654228, "learning_rate": 8.511671600921306e-07, "loss": 0.161, "step": 1042 }, { "epoch": 1.6246105919003115, "grad_norm": 0.5132546282344246, "learning_rate": 8.443518743822726e-07, "loss": 0.2764, "step": 1043 }, { "epoch": 1.6261682242990654, "grad_norm": 0.6893277741210059, "learning_rate": 8.375614662591097e-07, "loss": 0.1793, "step": 1044 }, { "epoch": 1.6277258566978192, "grad_norm": 0.46872986327742755, "learning_rate": 8.307959763730899e-07, "loss": 0.1761, "step": 1045 }, { "epoch": 1.629283489096573, "grad_norm": 0.5290839127332908, "learning_rate": 8.240554452254929e-07, "loss": 0.2329, "step": 1046 }, { "epoch": 1.6308411214953271, "grad_norm": 0.6109039006881162, "learning_rate": 8.173399131681831e-07, "loss": 0.2121, "step": 1047 }, { "epoch": 1.632398753894081, "grad_norm": 0.6091367601010639, "learning_rate": 8.106494204033677e-07, "loss": 0.1873, "step": 1048 }, { "epoch": 1.633956386292835, "grad_norm": 0.5867126273845332, "learning_rate": 8.039840069833594e-07, "loss": 0.7795, "step": 1049 }, { "epoch": 1.6355140186915889, "grad_norm": 0.5685700476106482, "learning_rate": 7.973437128103306e-07, "loss": 0.2459, "step": 1050 }, { "epoch": 1.6370716510903427, "grad_norm": 0.5135498312661578, "learning_rate": 7.907285776360863e-07, "loss": 0.255, "step": 1051 }, { "epoch": 1.6386292834890965, "grad_norm": 0.6667857649972504, "learning_rate": 7.841386410618096e-07, "loss": 0.2476, "step": 1052 }, { "epoch": 1.6401869158878504, "grad_norm": 0.5960262952580946, "learning_rate": 7.775739425378398e-07, "loss": 0.2462, "step": 1053 }, { "epoch": 1.6417445482866042, "grad_norm": 0.8148918825910239, "learning_rate": 7.710345213634247e-07, "loss": 0.2342, "step": 1054 }, { "epoch": 1.6433021806853583, "grad_norm": 0.6117412041332031, "learning_rate": 7.645204166864967e-07, "loss": 0.1889, "step": 1055 }, { "epoch": 1.644859813084112, "grad_norm": 0.6194608199292435, "learning_rate": 7.580316675034255e-07, "loss": 0.1807, "step": 1056 }, { "epoch": 1.6464174454828662, "grad_norm": 0.9323712553721532, "learning_rate": 7.515683126587975e-07, "loss": 0.1879, "step": 1057 }, { "epoch": 1.64797507788162, "grad_norm": 0.5493635025344051, "learning_rate": 7.451303908451707e-07, "loss": 0.2106, "step": 1058 }, { "epoch": 1.6495327102803738, "grad_norm": 0.7158277912716375, "learning_rate": 7.387179406028549e-07, "loss": 0.2354, "step": 1059 }, { "epoch": 1.6510903426791277, "grad_norm": 0.5571552725370883, "learning_rate": 7.323310003196749e-07, "loss": 0.2145, "step": 1060 }, { "epoch": 1.6526479750778815, "grad_norm": 0.7538580273599934, "learning_rate": 7.259696082307372e-07, "loss": 0.2482, "step": 1061 }, { "epoch": 1.6542056074766354, "grad_norm": 0.5327326528906995, "learning_rate": 7.196338024182098e-07, "loss": 0.2354, "step": 1062 }, { "epoch": 1.6557632398753894, "grad_norm": 0.557287992992955, "learning_rate": 7.133236208110878e-07, "loss": 0.1693, "step": 1063 }, { "epoch": 1.6573208722741433, "grad_norm": 0.5104215870170742, "learning_rate": 7.070391011849698e-07, "loss": 0.1717, "step": 1064 }, { "epoch": 1.6588785046728973, "grad_norm": 0.5917806139945608, "learning_rate": 7.007802811618258e-07, "loss": 0.2085, "step": 1065 }, { "epoch": 1.6604361370716512, "grad_norm": 0.6306946462467933, "learning_rate": 6.945471982097818e-07, "loss": 0.2145, "step": 1066 }, { "epoch": 1.661993769470405, "grad_norm": 0.7196660102146798, "learning_rate": 6.883398896428867e-07, "loss": 0.2321, "step": 1067 }, { "epoch": 1.6635514018691588, "grad_norm": 0.532187998071398, "learning_rate": 6.821583926208947e-07, "loss": 0.2366, "step": 1068 }, { "epoch": 1.6651090342679127, "grad_norm": 0.5827926596867945, "learning_rate": 6.760027441490369e-07, "loss": 0.285, "step": 1069 }, { "epoch": 1.6666666666666665, "grad_norm": 0.7057955872693081, "learning_rate": 6.698729810778065e-07, "loss": 0.6118, "step": 1070 }, { "epoch": 1.6682242990654206, "grad_norm": 0.7121732240730015, "learning_rate": 6.637691401027352e-07, "loss": 0.8544, "step": 1071 }, { "epoch": 1.6697819314641744, "grad_norm": 0.49184886655898746, "learning_rate": 6.576912577641697e-07, "loss": 0.2325, "step": 1072 }, { "epoch": 1.6713395638629285, "grad_norm": 0.6760569677497837, "learning_rate": 6.516393704470608e-07, "loss": 0.8096, "step": 1073 }, { "epoch": 1.6728971962616823, "grad_norm": 0.7478627429100672, "learning_rate": 6.456135143807369e-07, "loss": 0.1594, "step": 1074 }, { "epoch": 1.6744548286604362, "grad_norm": 0.5379337402997096, "learning_rate": 6.396137256386975e-07, "loss": 0.1994, "step": 1075 }, { "epoch": 1.67601246105919, "grad_norm": 0.5032927663415819, "learning_rate": 6.336400401383857e-07, "loss": 0.2132, "step": 1076 }, { "epoch": 1.6775700934579438, "grad_norm": 0.47582414487022096, "learning_rate": 6.276924936409829e-07, "loss": 0.2478, "step": 1077 }, { "epoch": 1.6791277258566977, "grad_norm": 0.6434383172973982, "learning_rate": 6.217711217511857e-07, "loss": 0.2143, "step": 1078 }, { "epoch": 1.6806853582554517, "grad_norm": 0.6604513508530203, "learning_rate": 6.158759599170045e-07, "loss": 0.19, "step": 1079 }, { "epoch": 1.6822429906542056, "grad_norm": 0.5696994691557421, "learning_rate": 6.100070434295379e-07, "loss": 0.2234, "step": 1080 }, { "epoch": 1.6838006230529596, "grad_norm": 0.668728026470969, "learning_rate": 6.041644074227738e-07, "loss": 0.1627, "step": 1081 }, { "epoch": 1.6853582554517135, "grad_norm": 0.5959783592005928, "learning_rate": 5.983480868733666e-07, "loss": 0.2239, "step": 1082 }, { "epoch": 1.6869158878504673, "grad_norm": 0.7134998370022319, "learning_rate": 5.925581166004407e-07, "loss": 0.1782, "step": 1083 }, { "epoch": 1.6884735202492211, "grad_norm": 0.5805335392567044, "learning_rate": 5.867945312653733e-07, "loss": 0.1725, "step": 1084 }, { "epoch": 1.690031152647975, "grad_norm": 0.938179646578548, "learning_rate": 5.810573653715884e-07, "loss": 0.6935, "step": 1085 }, { "epoch": 1.6915887850467288, "grad_norm": 0.6255817452393061, "learning_rate": 5.753466532643531e-07, "loss": 0.2057, "step": 1086 }, { "epoch": 1.6931464174454829, "grad_norm": 0.6164273521801443, "learning_rate": 5.696624291305692e-07, "loss": 0.2354, "step": 1087 }, { "epoch": 1.6947040498442367, "grad_norm": 0.6024358987476546, "learning_rate": 5.64004726998571e-07, "loss": 0.1628, "step": 1088 }, { "epoch": 1.6962616822429908, "grad_norm": 0.583734062156969, "learning_rate": 5.583735807379165e-07, "loss": 0.258, "step": 1089 }, { "epoch": 1.6978193146417446, "grad_norm": 0.5381732861463082, "learning_rate": 5.527690240591927e-07, "loss": 0.259, "step": 1090 }, { "epoch": 1.6993769470404985, "grad_norm": 0.6432738953861825, "learning_rate": 5.471910905138062e-07, "loss": 0.2233, "step": 1091 }, { "epoch": 1.7009345794392523, "grad_norm": 0.7405741920054726, "learning_rate": 5.416398134937878e-07, "loss": 0.2732, "step": 1092 }, { "epoch": 1.7024922118380061, "grad_norm": 0.5681851173547261, "learning_rate": 5.361152262315883e-07, "loss": 0.1976, "step": 1093 }, { "epoch": 1.70404984423676, "grad_norm": 0.5473658208153401, "learning_rate": 5.306173617998823e-07, "loss": 0.2139, "step": 1094 }, { "epoch": 1.705607476635514, "grad_norm": 0.5255770004268006, "learning_rate": 5.251462531113705e-07, "loss": 0.1707, "step": 1095 }, { "epoch": 1.7071651090342679, "grad_norm": 0.5374196236910275, "learning_rate": 5.197019329185787e-07, "loss": 0.2281, "step": 1096 }, { "epoch": 1.708722741433022, "grad_norm": 0.5260272455503919, "learning_rate": 5.142844338136693e-07, "loss": 0.2471, "step": 1097 }, { "epoch": 1.7102803738317758, "grad_norm": 0.5646648215018458, "learning_rate": 5.088937882282358e-07, "loss": 0.1951, "step": 1098 }, { "epoch": 1.7118380062305296, "grad_norm": 0.5458034326374251, "learning_rate": 5.035300284331213e-07, "loss": 0.2432, "step": 1099 }, { "epoch": 1.7133956386292835, "grad_norm": 0.6037542349280097, "learning_rate": 4.981931865382111e-07, "loss": 0.2618, "step": 1100 }, { "epoch": 1.7149532710280373, "grad_norm": 0.5948181143891571, "learning_rate": 4.92883294492254e-07, "loss": 0.1635, "step": 1101 }, { "epoch": 1.7165109034267911, "grad_norm": 0.5776080511589572, "learning_rate": 4.876003840826588e-07, "loss": 0.2451, "step": 1102 }, { "epoch": 1.7180685358255452, "grad_norm": 0.5026789603534472, "learning_rate": 4.823444869353177e-07, "loss": 0.1562, "step": 1103 }, { "epoch": 1.719626168224299, "grad_norm": 0.5367686440727194, "learning_rate": 4.771156345144018e-07, "loss": 0.2041, "step": 1104 }, { "epoch": 1.721183800623053, "grad_norm": 0.658875304880407, "learning_rate": 4.7191385812218547e-07, "loss": 0.7385, "step": 1105 }, { "epoch": 1.722741433021807, "grad_norm": 0.6189314088368518, "learning_rate": 4.6673918889884985e-07, "loss": 0.269, "step": 1106 }, { "epoch": 1.7242990654205608, "grad_norm": 0.7000827321303451, "learning_rate": 4.615916578223029e-07, "loss": 0.219, "step": 1107 }, { "epoch": 1.7258566978193146, "grad_norm": 0.6218605150183268, "learning_rate": 4.5647129570799196e-07, "loss": 0.2765, "step": 1108 }, { "epoch": 1.7274143302180685, "grad_norm": 0.7349959501432963, "learning_rate": 4.513781332087153e-07, "loss": 0.2457, "step": 1109 }, { "epoch": 1.7289719626168223, "grad_norm": 0.7008089188891375, "learning_rate": 4.463122008144449e-07, "loss": 0.2083, "step": 1110 }, { "epoch": 1.7305295950155763, "grad_norm": 0.5607384923036217, "learning_rate": 4.4127352885214026e-07, "loss": 0.2166, "step": 1111 }, { "epoch": 1.7320872274143302, "grad_norm": 0.645885067353582, "learning_rate": 4.362621474855688e-07, "loss": 0.2312, "step": 1112 }, { "epoch": 1.7336448598130842, "grad_norm": 1.0208563939368382, "learning_rate": 4.3127808671512114e-07, "loss": 0.8117, "step": 1113 }, { "epoch": 1.735202492211838, "grad_norm": 0.6252228702704483, "learning_rate": 4.263213763776369e-07, "loss": 0.2111, "step": 1114 }, { "epoch": 1.736760124610592, "grad_norm": 0.5295697965398305, "learning_rate": 4.213920461462234e-07, "loss": 0.1693, "step": 1115 }, { "epoch": 1.7383177570093458, "grad_norm": 0.5321708104559851, "learning_rate": 4.1649012553007795e-07, "loss": 0.1598, "step": 1116 }, { "epoch": 1.7398753894080996, "grad_norm": 0.6297092340917954, "learning_rate": 4.116156438743102e-07, "loss": 0.6563, "step": 1117 }, { "epoch": 1.7414330218068534, "grad_norm": 0.6502646751207796, "learning_rate": 4.067686303597701e-07, "loss": 0.828, "step": 1118 }, { "epoch": 1.7429906542056075, "grad_norm": 0.6299291575701418, "learning_rate": 4.0194911400287084e-07, "loss": 0.1618, "step": 1119 }, { "epoch": 1.7445482866043613, "grad_norm": 0.6763449045114679, "learning_rate": 3.971571236554117e-07, "loss": 0.1755, "step": 1120 }, { "epoch": 1.7461059190031154, "grad_norm": 0.6097387148482865, "learning_rate": 3.923926880044132e-07, "loss": 0.2144, "step": 1121 }, { "epoch": 1.7476635514018692, "grad_norm": 0.6555185956249271, "learning_rate": 3.876558355719373e-07, "loss": 0.7322, "step": 1122 }, { "epoch": 1.749221183800623, "grad_norm": 0.529186428154948, "learning_rate": 3.8294659471492466e-07, "loss": 0.2299, "step": 1123 }, { "epoch": 1.750778816199377, "grad_norm": 0.5210926587424783, "learning_rate": 3.7826499362501657e-07, "loss": 0.7365, "step": 1124 }, { "epoch": 1.7523364485981308, "grad_norm": 0.5705242040491778, "learning_rate": 3.7361106032839266e-07, "loss": 0.1769, "step": 1125 }, { "epoch": 1.7538940809968846, "grad_norm": 0.6221568867702569, "learning_rate": 3.6898482268559745e-07, "loss": 0.2641, "step": 1126 }, { "epoch": 1.7554517133956387, "grad_norm": 0.5761942177039651, "learning_rate": 3.643863083913829e-07, "loss": 0.1756, "step": 1127 }, { "epoch": 1.7570093457943925, "grad_norm": 0.5598040118881634, "learning_rate": 3.5981554497452886e-07, "loss": 0.1968, "step": 1128 }, { "epoch": 1.7585669781931466, "grad_norm": 0.6115168676953736, "learning_rate": 3.552725597976925e-07, "loss": 0.1656, "step": 1129 }, { "epoch": 1.7601246105919004, "grad_norm": 0.5618155080059914, "learning_rate": 3.507573800572328e-07, "loss": 0.2284, "step": 1130 }, { "epoch": 1.7616822429906542, "grad_norm": 0.5746401068264577, "learning_rate": 3.462700327830559e-07, "loss": 0.2942, "step": 1131 }, { "epoch": 1.763239875389408, "grad_norm": 0.5812252313415237, "learning_rate": 3.418105448384507e-07, "loss": 0.2478, "step": 1132 }, { "epoch": 1.764797507788162, "grad_norm": 0.5205912713872596, "learning_rate": 3.3737894291992426e-07, "loss": 0.1793, "step": 1133 }, { "epoch": 1.7663551401869158, "grad_norm": 0.59827648712384, "learning_rate": 3.32975253557048e-07, "loss": 0.1643, "step": 1134 }, { "epoch": 1.7679127725856698, "grad_norm": 0.5476307417528222, "learning_rate": 3.2859950311229625e-07, "loss": 0.2551, "step": 1135 }, { "epoch": 1.7694704049844237, "grad_norm": 0.5912924193469346, "learning_rate": 3.242517177808874e-07, "loss": 0.2958, "step": 1136 }, { "epoch": 1.7710280373831777, "grad_norm": 0.6149100468721329, "learning_rate": 3.199319235906273e-07, "loss": 0.2391, "step": 1137 }, { "epoch": 1.7725856697819315, "grad_norm": 0.5135930360411095, "learning_rate": 3.1564014640175613e-07, "loss": 0.1651, "step": 1138 }, { "epoch": 1.7741433021806854, "grad_norm": 0.545428012439618, "learning_rate": 3.1137641190678967e-07, "loss": 0.1866, "step": 1139 }, { "epoch": 1.7757009345794392, "grad_norm": 0.7534351814752701, "learning_rate": 3.0714074563037043e-07, "loss": 0.1752, "step": 1140 }, { "epoch": 1.777258566978193, "grad_norm": 0.6368630911991421, "learning_rate": 3.029331729291074e-07, "loss": 0.2049, "step": 1141 }, { "epoch": 1.778816199376947, "grad_norm": 0.594879170975228, "learning_rate": 2.9875371899143246e-07, "loss": 0.6515, "step": 1142 }, { "epoch": 1.780373831775701, "grad_norm": 0.5107491983811178, "learning_rate": 2.9460240883744496e-07, "loss": 0.1768, "step": 1143 }, { "epoch": 1.7819314641744548, "grad_norm": 0.5417424283009235, "learning_rate": 2.9047926731876177e-07, "loss": 0.2087, "step": 1144 }, { "epoch": 1.7834890965732089, "grad_norm": 0.6798250278910997, "learning_rate": 2.8638431911837126e-07, "loss": 0.28, "step": 1145 }, { "epoch": 1.7850467289719627, "grad_norm": 1.0039697477055525, "learning_rate": 2.8231758875048087e-07, "loss": 0.2178, "step": 1146 }, { "epoch": 1.7866043613707165, "grad_norm": 0.6262879884675681, "learning_rate": 2.7827910056037857e-07, "loss": 0.1775, "step": 1147 }, { "epoch": 1.7881619937694704, "grad_norm": 0.6215897703245564, "learning_rate": 2.742688787242764e-07, "loss": 0.1726, "step": 1148 }, { "epoch": 1.7897196261682242, "grad_norm": 0.7119013996583601, "learning_rate": 2.702869472491748e-07, "loss": 0.6972, "step": 1149 }, { "epoch": 1.791277258566978, "grad_norm": 0.5300668663016045, "learning_rate": 2.6633332997271277e-07, "loss": 0.1797, "step": 1150 }, { "epoch": 1.7928348909657321, "grad_norm": 0.6010057607669876, "learning_rate": 2.624080505630311e-07, "loss": 0.2206, "step": 1151 }, { "epoch": 1.794392523364486, "grad_norm": 0.5566227596606641, "learning_rate": 2.585111325186235e-07, "loss": 0.2991, "step": 1152 }, { "epoch": 1.79595015576324, "grad_norm": 0.7373120599532028, "learning_rate": 2.5464259916820276e-07, "loss": 0.2155, "step": 1153 }, { "epoch": 1.7975077881619939, "grad_norm": 0.5444732937470245, "learning_rate": 2.508024736705561e-07, "loss": 0.2394, "step": 1154 }, { "epoch": 1.7990654205607477, "grad_norm": 0.6187155054450917, "learning_rate": 2.469907790144088e-07, "loss": 0.1835, "step": 1155 }, { "epoch": 1.8006230529595015, "grad_norm": 0.45905420748321907, "learning_rate": 2.4320753801828853e-07, "loss": 0.1604, "step": 1156 }, { "epoch": 1.8021806853582554, "grad_norm": 0.43879992123704614, "learning_rate": 2.394527733303831e-07, "loss": 0.1723, "step": 1157 }, { "epoch": 1.8037383177570092, "grad_norm": 0.5693210977954104, "learning_rate": 2.3572650742840985e-07, "loss": 0.1148, "step": 1158 }, { "epoch": 1.8052959501557633, "grad_norm": 0.5824148680871707, "learning_rate": 2.3202876261948004e-07, "loss": 0.2304, "step": 1159 }, { "epoch": 1.8068535825545171, "grad_norm": 0.692163156120945, "learning_rate": 2.2835956103996525e-07, "loss": 0.1877, "step": 1160 }, { "epoch": 1.8084112149532712, "grad_norm": 0.627351603757966, "learning_rate": 2.247189246553616e-07, "loss": 0.2133, "step": 1161 }, { "epoch": 1.809968847352025, "grad_norm": 0.6447153050268107, "learning_rate": 2.2110687526016418e-07, "loss": 0.1753, "step": 1162 }, { "epoch": 1.8115264797507789, "grad_norm": 0.6498742455536043, "learning_rate": 2.1752343447773105e-07, "loss": 0.2801, "step": 1163 }, { "epoch": 1.8130841121495327, "grad_norm": 0.5556829005135521, "learning_rate": 2.1396862376015904e-07, "loss": 0.2613, "step": 1164 }, { "epoch": 1.8146417445482865, "grad_norm": 0.5643429928213758, "learning_rate": 2.1044246438814918e-07, "loss": 0.2529, "step": 1165 }, { "epoch": 1.8161993769470404, "grad_norm": 0.747941354496335, "learning_rate": 2.0694497747088428e-07, "loss": 0.2321, "step": 1166 }, { "epoch": 1.8177570093457944, "grad_norm": 0.6102806034003578, "learning_rate": 2.034761839459015e-07, "loss": 0.218, "step": 1167 }, { "epoch": 1.8193146417445483, "grad_norm": 0.7249346044063573, "learning_rate": 2.0003610457896506e-07, "loss": 0.2317, "step": 1168 }, { "epoch": 1.8208722741433023, "grad_norm": 0.5858879975789372, "learning_rate": 1.966247599639448e-07, "loss": 0.1886, "step": 1169 }, { "epoch": 1.8224299065420562, "grad_norm": 0.458859760014875, "learning_rate": 1.9324217052268835e-07, "loss": 0.2037, "step": 1170 }, { "epoch": 1.82398753894081, "grad_norm": 0.6009222197974706, "learning_rate": 1.8988835650490645e-07, "loss": 0.1845, "step": 1171 }, { "epoch": 1.8255451713395638, "grad_norm": 0.7980290452291322, "learning_rate": 1.865633379880427e-07, "loss": 0.2341, "step": 1172 }, { "epoch": 1.8271028037383177, "grad_norm": 0.5432836689095761, "learning_rate": 1.832671348771614e-07, "loss": 0.2282, "step": 1173 }, { "epoch": 1.8286604361370715, "grad_norm": 0.628743562125201, "learning_rate": 1.799997669048209e-07, "loss": 0.261, "step": 1174 }, { "epoch": 1.8302180685358256, "grad_norm": 0.5832048925707848, "learning_rate": 1.7676125363096431e-07, "loss": 0.2157, "step": 1175 }, { "epoch": 1.8317757009345794, "grad_norm": 0.6281683541540782, "learning_rate": 1.7355161444279346e-07, "loss": 0.2799, "step": 1176 }, { "epoch": 1.8333333333333335, "grad_norm": 0.6990945601104308, "learning_rate": 1.7037086855465902e-07, "loss": 0.2248, "step": 1177 }, { "epoch": 1.8348909657320873, "grad_norm": 0.48096314961131165, "learning_rate": 1.6721903500794112e-07, "loss": 0.2014, "step": 1178 }, { "epoch": 1.8364485981308412, "grad_norm": 0.6720925885316696, "learning_rate": 1.6409613267094006e-07, "loss": 0.1839, "step": 1179 }, { "epoch": 1.838006230529595, "grad_norm": 0.5023236174498488, "learning_rate": 1.6100218023876013e-07, "loss": 0.1729, "step": 1180 }, { "epoch": 1.8395638629283488, "grad_norm": 0.6472364425599902, "learning_rate": 1.5793719623319713e-07, "loss": 0.2147, "step": 1181 }, { "epoch": 1.8411214953271027, "grad_norm": 0.5750750734641096, "learning_rate": 1.5490119900263e-07, "loss": 0.1734, "step": 1182 }, { "epoch": 1.8426791277258567, "grad_norm": 0.640400103725089, "learning_rate": 1.5189420672190924e-07, "loss": 0.2149, "step": 1183 }, { "epoch": 1.8442367601246106, "grad_norm": 0.6159833044318871, "learning_rate": 1.489162373922498e-07, "loss": 0.2426, "step": 1184 }, { "epoch": 1.8457943925233646, "grad_norm": 0.5891969287939168, "learning_rate": 1.4596730884112008e-07, "loss": 0.2164, "step": 1185 }, { "epoch": 1.8473520249221185, "grad_norm": 0.6060235472721057, "learning_rate": 1.4304743872213868e-07, "loss": 0.2567, "step": 1186 }, { "epoch": 1.8489096573208723, "grad_norm": 0.6985533236531171, "learning_rate": 1.4015664451496713e-07, "loss": 0.1688, "step": 1187 }, { "epoch": 1.8504672897196262, "grad_norm": 0.5859326752612632, "learning_rate": 1.372949435252058e-07, "loss": 0.2311, "step": 1188 }, { "epoch": 1.85202492211838, "grad_norm": 0.7294577853076925, "learning_rate": 1.3446235288428867e-07, "loss": 0.2023, "step": 1189 }, { "epoch": 1.8535825545171338, "grad_norm": 0.6888020410981194, "learning_rate": 1.31658889549382e-07, "loss": 0.2295, "step": 1190 }, { "epoch": 1.855140186915888, "grad_norm": 0.6480785185138133, "learning_rate": 1.288845703032854e-07, "loss": 0.2359, "step": 1191 }, { "epoch": 1.8566978193146417, "grad_norm": 0.7015546454899397, "learning_rate": 1.2613941175432577e-07, "loss": 0.2293, "step": 1192 }, { "epoch": 1.8582554517133958, "grad_norm": 0.5363209042476926, "learning_rate": 1.2342343033626248e-07, "loss": 0.18, "step": 1193 }, { "epoch": 1.8598130841121496, "grad_norm": 0.6459401688040521, "learning_rate": 1.2073664230818571e-07, "loss": 0.1598, "step": 1194 }, { "epoch": 1.8613707165109035, "grad_norm": 0.638563320329426, "learning_rate": 1.180790637544249e-07, "loss": 0.2476, "step": 1195 }, { "epoch": 1.8629283489096573, "grad_norm": 0.6002853300206833, "learning_rate": 1.1545071058444324e-07, "loss": 0.2199, "step": 1196 }, { "epoch": 1.8644859813084111, "grad_norm": 0.5764195151127395, "learning_rate": 1.1285159853275107e-07, "loss": 0.1227, "step": 1197 }, { "epoch": 1.866043613707165, "grad_norm": 0.6123103816053441, "learning_rate": 1.1028174315880657e-07, "loss": 0.815, "step": 1198 }, { "epoch": 1.867601246105919, "grad_norm": 0.8123905890394445, "learning_rate": 1.0774115984692523e-07, "loss": 0.7364, "step": 1199 }, { "epoch": 1.8691588785046729, "grad_norm": 0.5718174425382163, "learning_rate": 1.0522986380618606e-07, "loss": 0.2227, "step": 1200 }, { "epoch": 1.870716510903427, "grad_norm": 0.5280345967170433, "learning_rate": 1.0274787007034215e-07, "loss": 0.232, "step": 1201 }, { "epoch": 1.8722741433021808, "grad_norm": 0.6572165953398976, "learning_rate": 1.0029519349772754e-07, "loss": 0.2738, "step": 1202 }, { "epoch": 1.8738317757009346, "grad_norm": 0.5244811712323645, "learning_rate": 9.787184877117328e-08, "loss": 0.2594, "step": 1203 }, { "epoch": 1.8753894080996885, "grad_norm": 0.6676225172095591, "learning_rate": 9.547785039791535e-08, "loss": 0.1547, "step": 1204 }, { "epoch": 1.8769470404984423, "grad_norm": 0.5476881051496193, "learning_rate": 9.311321270950801e-08, "loss": 0.2128, "step": 1205 }, { "epoch": 1.8785046728971961, "grad_norm": 0.5345787062303978, "learning_rate": 9.077794986174226e-08, "loss": 0.229, "step": 1206 }, { "epoch": 1.8800623052959502, "grad_norm": 0.6008670505158255, "learning_rate": 8.84720758345553e-08, "loss": 0.2073, "step": 1207 }, { "epoch": 1.881619937694704, "grad_norm": 0.6417543491732723, "learning_rate": 8.619560443195174e-08, "loss": 0.2276, "step": 1208 }, { "epoch": 1.883177570093458, "grad_norm": 0.5552051709339759, "learning_rate": 8.394854928191587e-08, "loss": 0.2693, "step": 1209 }, { "epoch": 1.884735202492212, "grad_norm": 0.6694794375263459, "learning_rate": 8.173092383633563e-08, "loss": 0.1767, "step": 1210 }, { "epoch": 1.8862928348909658, "grad_norm": 0.5551294915003306, "learning_rate": 7.954274137091877e-08, "loss": 0.147, "step": 1211 }, { "epoch": 1.8878504672897196, "grad_norm": 0.534343522912461, "learning_rate": 7.738401498511406e-08, "loss": 0.1891, "step": 1212 }, { "epoch": 1.8894080996884735, "grad_norm": 0.6889315198186039, "learning_rate": 7.525475760203239e-08, "loss": 0.2832, "step": 1213 }, { "epoch": 1.8909657320872273, "grad_norm": 0.6872767665873238, "learning_rate": 7.315498196836968e-08, "loss": 0.1472, "step": 1214 }, { "epoch": 1.8925233644859814, "grad_norm": 0.6612078267914279, "learning_rate": 7.108470065433193e-08, "loss": 0.2704, "step": 1215 }, { "epoch": 1.8940809968847352, "grad_norm": 0.560144578359172, "learning_rate": 6.904392605355803e-08, "loss": 0.2943, "step": 1216 }, { "epoch": 1.8956386292834893, "grad_norm": 0.590357610353342, "learning_rate": 6.703267038304706e-08, "loss": 0.229, "step": 1217 }, { "epoch": 1.897196261682243, "grad_norm": 0.496234632076301, "learning_rate": 6.505094568308223e-08, "loss": 0.2807, "step": 1218 }, { "epoch": 1.898753894080997, "grad_norm": 0.6744039849106092, "learning_rate": 6.309876381716429e-08, "loss": 0.1744, "step": 1219 }, { "epoch": 1.9003115264797508, "grad_norm": 0.6282540133425786, "learning_rate": 6.117613647193543e-08, "loss": 0.1896, "step": 1220 }, { "epoch": 1.9018691588785046, "grad_norm": 0.545913837904785, "learning_rate": 5.928307515711107e-08, "loss": 0.1798, "step": 1221 }, { "epoch": 1.9034267912772584, "grad_norm": 0.5750165821891594, "learning_rate": 5.741959120541096e-08, "loss": 0.2328, "step": 1222 }, { "epoch": 1.9049844236760125, "grad_norm": 0.6138340557718204, "learning_rate": 5.558569577249207e-08, "loss": 0.1179, "step": 1223 }, { "epoch": 1.9065420560747663, "grad_norm": 0.5250776885173793, "learning_rate": 5.378139983688135e-08, "loss": 0.178, "step": 1224 }, { "epoch": 1.9080996884735204, "grad_norm": 0.5353000767702344, "learning_rate": 5.2006714199908106e-08, "loss": 0.1881, "step": 1225 }, { "epoch": 1.9096573208722742, "grad_norm": 0.6835275834635214, "learning_rate": 5.026164948564283e-08, "loss": 0.2075, "step": 1226 }, { "epoch": 1.911214953271028, "grad_norm": 0.48910598318494225, "learning_rate": 4.854621614083122e-08, "loss": 0.1864, "step": 1227 }, { "epoch": 1.912772585669782, "grad_norm": 0.5349299444820935, "learning_rate": 4.686042443483196e-08, "loss": 0.297, "step": 1228 }, { "epoch": 1.9143302180685358, "grad_norm": 0.42493561251864087, "learning_rate": 4.520428445955571e-08, "loss": 0.1656, "step": 1229 }, { "epoch": 1.9158878504672896, "grad_norm": 0.5786034487132414, "learning_rate": 4.357780612940343e-08, "loss": 0.2243, "step": 1230 }, { "epoch": 1.9174454828660437, "grad_norm": 0.6628360349913834, "learning_rate": 4.1980999181210344e-08, "loss": 0.237, "step": 1231 }, { "epoch": 1.9190031152647975, "grad_norm": 0.5726232559997532, "learning_rate": 4.041387317418377e-08, "loss": 0.19, "step": 1232 }, { "epoch": 1.9205607476635516, "grad_norm": 0.5957569462038291, "learning_rate": 3.88764374898476e-08, "loss": 0.1922, "step": 1233 }, { "epoch": 1.9221183800623054, "grad_norm": 0.5981083547680797, "learning_rate": 3.736870133198622e-08, "loss": 0.1861, "step": 1234 }, { "epoch": 1.9236760124610592, "grad_norm": 0.5039034286835927, "learning_rate": 3.589067372658961e-08, "loss": 0.1587, "step": 1235 }, { "epoch": 1.925233644859813, "grad_norm": 0.5946390323111295, "learning_rate": 3.444236352179831e-08, "loss": 0.1658, "step": 1236 }, { "epoch": 1.926791277258567, "grad_norm": 0.595459693543413, "learning_rate": 3.302377938785128e-08, "loss": 0.1509, "step": 1237 }, { "epoch": 1.9283489096573208, "grad_norm": 0.6496361440227659, "learning_rate": 3.163492981703431e-08, "loss": 0.693, "step": 1238 }, { "epoch": 1.9299065420560748, "grad_norm": 0.6473442377574977, "learning_rate": 3.027582312362776e-08, "loss": 0.2082, "step": 1239 }, { "epoch": 1.9314641744548287, "grad_norm": 0.6541508019021366, "learning_rate": 2.8946467443858873e-08, "loss": 0.1911, "step": 1240 }, { "epoch": 1.9330218068535827, "grad_norm": 0.5343926005691645, "learning_rate": 2.764687073585015e-08, "loss": 0.1724, "step": 1241 }, { "epoch": 1.9345794392523366, "grad_norm": 0.6477434345058528, "learning_rate": 2.6377040779574924e-08, "loss": 0.2109, "step": 1242 }, { "epoch": 1.9361370716510904, "grad_norm": 0.4788618278691479, "learning_rate": 2.5136985176809625e-08, "loss": 0.2257, "step": 1243 }, { "epoch": 1.9376947040498442, "grad_norm": 0.7161454071348646, "learning_rate": 2.3926711351086594e-08, "loss": 0.2201, "step": 1244 }, { "epoch": 1.939252336448598, "grad_norm": 0.6162622985277744, "learning_rate": 2.2746226547651352e-08, "loss": 0.2353, "step": 1245 }, { "epoch": 1.940809968847352, "grad_norm": 0.6572303875322422, "learning_rate": 2.1595537833419276e-08, "loss": 0.1238, "step": 1246 }, { "epoch": 1.942367601246106, "grad_norm": 0.4905959486856584, "learning_rate": 2.047465209693289e-08, "loss": 0.1913, "step": 1247 }, { "epoch": 1.9439252336448598, "grad_norm": 0.6262698157072888, "learning_rate": 1.9383576048320752e-08, "loss": 0.2135, "step": 1248 }, { "epoch": 1.9454828660436139, "grad_norm": 0.5977256427085367, "learning_rate": 1.8322316219256396e-08, "loss": 0.3103, "step": 1249 }, { "epoch": 1.9470404984423677, "grad_norm": 0.5597696638121235, "learning_rate": 1.7290878962920587e-08, "loss": 0.2003, "step": 1250 }, { "epoch": 1.9485981308411215, "grad_norm": 0.7414334515835818, "learning_rate": 1.6289270453963e-08, "loss": 0.133, "step": 1251 }, { "epoch": 1.9501557632398754, "grad_norm": 0.6577156096873663, "learning_rate": 1.531749668846394e-08, "loss": 0.1938, "step": 1252 }, { "epoch": 1.9517133956386292, "grad_norm": 0.44233229164345883, "learning_rate": 1.4375563483901011e-08, "loss": 0.2196, "step": 1253 }, { "epoch": 1.953271028037383, "grad_norm": 0.5494650237391789, "learning_rate": 1.3463476479110837e-08, "loss": 0.239, "step": 1254 }, { "epoch": 1.9548286604361371, "grad_norm": 0.4597714878638595, "learning_rate": 1.2581241134258515e-08, "loss": 0.1569, "step": 1255 }, { "epoch": 1.956386292834891, "grad_norm": 0.5344018379047227, "learning_rate": 1.1728862730803759e-08, "loss": 0.2506, "step": 1256 }, { "epoch": 1.957943925233645, "grad_norm": 0.6170756968114266, "learning_rate": 1.0906346371468147e-08, "loss": 0.1756, "step": 1257 }, { "epoch": 1.9595015576323989, "grad_norm": 0.7298638061579327, "learning_rate": 1.011369698020681e-08, "loss": 1.2315, "step": 1258 }, { "epoch": 1.9610591900311527, "grad_norm": 0.7726214050156128, "learning_rate": 9.350919302176242e-09, "loss": 0.2132, "step": 1259 }, { "epoch": 1.9626168224299065, "grad_norm": 0.7662101654548159, "learning_rate": 8.618017903708198e-09, "loss": 0.1979, "step": 1260 }, { "epoch": 1.9641744548286604, "grad_norm": 0.7759433714405526, "learning_rate": 7.914997172282502e-09, "loss": 0.7825, "step": 1261 }, { "epoch": 1.9657320872274142, "grad_norm": 0.5561454562523979, "learning_rate": 7.241861316497623e-09, "loss": 0.1787, "step": 1262 }, { "epoch": 1.9672897196261683, "grad_norm": 0.6088898793952118, "learning_rate": 6.598614366049027e-09, "loss": 0.1525, "step": 1263 }, { "epoch": 1.9688473520249221, "grad_norm": 0.6388680112674057, "learning_rate": 5.985260171704199e-09, "loss": 0.2538, "step": 1264 }, { "epoch": 1.9704049844236762, "grad_norm": 1.2537539847150645, "learning_rate": 5.401802405277656e-09, "loss": 0.208, "step": 1265 }, { "epoch": 1.97196261682243, "grad_norm": 0.5173311007016309, "learning_rate": 4.848244559610971e-09, "loss": 0.1927, "step": 1266 }, { "epoch": 1.9735202492211839, "grad_norm": 0.7196153604341586, "learning_rate": 4.324589948552227e-09, "loss": 0.1292, "step": 1267 }, { "epoch": 1.9750778816199377, "grad_norm": 0.8257531526111838, "learning_rate": 3.830841706934374e-09, "loss": 0.2268, "step": 1268 }, { "epoch": 1.9766355140186915, "grad_norm": 0.6450834847941774, "learning_rate": 3.367002790556906e-09, "loss": 0.8063, "step": 1269 }, { "epoch": 1.9781931464174454, "grad_norm": 0.5873649988075496, "learning_rate": 2.9330759761692086e-09, "loss": 0.1333, "step": 1270 }, { "epoch": 1.9797507788161994, "grad_norm": 0.5338211522410232, "learning_rate": 2.5290638614533516e-09, "loss": 0.1651, "step": 1271 }, { "epoch": 1.9813084112149533, "grad_norm": 0.5778154838677994, "learning_rate": 2.154968865007989e-09, "loss": 0.1375, "step": 1272 }, { "epoch": 1.9828660436137073, "grad_norm": 0.6159025357834766, "learning_rate": 1.8107932263350392e-09, "loss": 0.2551, "step": 1273 }, { "epoch": 1.9844236760124612, "grad_norm": 0.5276126425838753, "learning_rate": 1.4965390058258035e-09, "loss": 0.2857, "step": 1274 }, { "epoch": 1.985981308411215, "grad_norm": 0.578702087814479, "learning_rate": 1.2122080847470907e-09, "loss": 0.1644, "step": 1275 }, { "epoch": 1.9875389408099688, "grad_norm": 0.6156843300297827, "learning_rate": 9.578021652323356e-10, "loss": 0.203, "step": 1276 }, { "epoch": 1.9890965732087227, "grad_norm": 0.5234412510622443, "learning_rate": 7.3332277027105e-10, "loss": 0.2086, "step": 1277 }, { "epoch": 1.9906542056074765, "grad_norm": 0.6889557509874759, "learning_rate": 5.38771243697167e-10, "loss": 0.2412, "step": 1278 }, { "epoch": 1.9922118380062306, "grad_norm": 0.5571525262541802, "learning_rate": 3.7414875018515483e-10, "loss": 0.314, "step": 1279 }, { "epoch": 1.9937694704049844, "grad_norm": 0.48476554367462193, "learning_rate": 2.3945627523891403e-10, "loss": 0.1878, "step": 1280 }, { "epoch": 1.9953271028037385, "grad_norm": 0.5759369634737167, "learning_rate": 1.3469462518844733e-10, "loss": 0.181, "step": 1281 }, { "epoch": 1.9968847352024923, "grad_norm": 0.6640918787171276, "learning_rate": 5.986442718486363e-11, "loss": 0.1886, "step": 1282 }, { "epoch": 1.9984423676012462, "grad_norm": 0.6150458129744069, "learning_rate": 1.496612919482665e-11, "loss": 0.2397, "step": 1283 }, { "epoch": 2.0, "grad_norm": 0.6727485887005468, "learning_rate": 0.0, "loss": 0.2332, "step": 1284 }, { "epoch": 2.0, "step": 1284, "total_flos": 17255618273280.0, "train_loss": 0.27858535300162723, "train_runtime": 1476.77, "train_samples_per_second": 3.477, "train_steps_per_second": 0.869 } ], "logging_steps": 1, "max_steps": 1284, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 70000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 17255618273280.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }