{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 2884, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00034674063800277393, "grad_norm": 3.4529595375061035, "learning_rate": 0.0, "loss": 0.2571, "step": 1 }, { "epoch": 0.0006934812760055479, "grad_norm": 1.7922414541244507, "learning_rate": 1e-06, "loss": 0.2304, "step": 2 }, { "epoch": 0.0010402219140083217, "grad_norm": 2.7978198528289795, "learning_rate": 1e-06, "loss": 0.2421, "step": 3 }, { "epoch": 0.0013869625520110957, "grad_norm": 1.7440896034240723, "learning_rate": 1e-06, "loss": 0.2437, "step": 4 }, { "epoch": 0.0017337031900138697, "grad_norm": 2.0253705978393555, "learning_rate": 1e-06, "loss": 0.2456, "step": 5 }, { "epoch": 0.0020804438280166435, "grad_norm": 7.390465259552002, "learning_rate": 1e-06, "loss": 0.285, "step": 6 }, { "epoch": 0.0024271844660194173, "grad_norm": 2.6305766105651855, "learning_rate": 1e-06, "loss": 0.2477, "step": 7 }, { "epoch": 0.0027739251040221915, "grad_norm": 3.094043254852295, "learning_rate": 1e-06, "loss": 0.2229, "step": 8 }, { "epoch": 0.0031206657420249652, "grad_norm": 2.3640098571777344, "learning_rate": 1e-06, "loss": 0.2233, "step": 9 }, { "epoch": 0.0034674063800277394, "grad_norm": 2.151405096054077, "learning_rate": 1e-06, "loss": 0.2211, "step": 10 }, { "epoch": 0.0038141470180305132, "grad_norm": 5.151231288909912, "learning_rate": 1e-06, "loss": 0.2354, "step": 11 }, { "epoch": 0.004160887656033287, "grad_norm": 2.285297155380249, "learning_rate": 1e-06, "loss": 0.2483, "step": 12 }, { "epoch": 0.004507628294036061, "grad_norm": 2.348036050796509, "learning_rate": 1e-06, "loss": 0.2419, "step": 13 }, { "epoch": 0.0048543689320388345, "grad_norm": 2.321516275405884, "learning_rate": 1e-06, "loss": 0.2659, "step": 14 }, { "epoch": 0.005201109570041609, "grad_norm": 2.291966438293457, "learning_rate": 1e-06, "loss": 0.2315, "step": 15 }, { "epoch": 0.005547850208044383, "grad_norm": 2.938875198364258, "learning_rate": 1e-06, "loss": 0.2479, "step": 16 }, { "epoch": 0.005894590846047157, "grad_norm": 1.7737529277801514, "learning_rate": 1e-06, "loss": 0.2242, "step": 17 }, { "epoch": 0.0062413314840499305, "grad_norm": 2.331916332244873, "learning_rate": 1e-06, "loss": 0.2386, "step": 18 }, { "epoch": 0.006588072122052704, "grad_norm": 2.8218319416046143, "learning_rate": 1e-06, "loss": 0.2655, "step": 19 }, { "epoch": 0.006934812760055479, "grad_norm": 1.9200022220611572, "learning_rate": 1e-06, "loss": 0.2384, "step": 20 }, { "epoch": 0.007281553398058253, "grad_norm": 2.6691722869873047, "learning_rate": 1e-06, "loss": 0.2594, "step": 21 }, { "epoch": 0.0076282940360610264, "grad_norm": 3.2480461597442627, "learning_rate": 1e-06, "loss": 0.1936, "step": 22 }, { "epoch": 0.0079750346740638, "grad_norm": 4.234226226806641, "learning_rate": 1e-06, "loss": 0.2672, "step": 23 }, { "epoch": 0.008321775312066574, "grad_norm": 4.796227931976318, "learning_rate": 1e-06, "loss": 0.2711, "step": 24 }, { "epoch": 0.008668515950069348, "grad_norm": 4.371148109436035, "learning_rate": 1e-06, "loss": 0.2256, "step": 25 }, { "epoch": 0.009015256588072122, "grad_norm": 2.540862798690796, "learning_rate": 1e-06, "loss": 0.2418, "step": 26 }, { "epoch": 0.009361997226074895, "grad_norm": 1.8659073114395142, "learning_rate": 1e-06, "loss": 0.2439, "step": 27 }, { "epoch": 0.009708737864077669, "grad_norm": 2.3605165481567383, "learning_rate": 1e-06, "loss": 0.2177, "step": 28 }, { "epoch": 0.010055478502080445, "grad_norm": 2.2104761600494385, "learning_rate": 1e-06, "loss": 0.2504, "step": 29 }, { "epoch": 0.010402219140083218, "grad_norm": 2.1602864265441895, "learning_rate": 1e-06, "loss": 0.2372, "step": 30 }, { "epoch": 0.010748959778085992, "grad_norm": 2.1729562282562256, "learning_rate": 1e-06, "loss": 0.2562, "step": 31 }, { "epoch": 0.011095700416088766, "grad_norm": 2.284585475921631, "learning_rate": 1e-06, "loss": 0.2355, "step": 32 }, { "epoch": 0.01144244105409154, "grad_norm": 2.268206834793091, "learning_rate": 1e-06, "loss": 0.2454, "step": 33 }, { "epoch": 0.011789181692094313, "grad_norm": 2.5299553871154785, "learning_rate": 1e-06, "loss": 0.2471, "step": 34 }, { "epoch": 0.012135922330097087, "grad_norm": 3.3298118114471436, "learning_rate": 1e-06, "loss": 0.2728, "step": 35 }, { "epoch": 0.012482662968099861, "grad_norm": 4.213827133178711, "learning_rate": 1e-06, "loss": 0.2271, "step": 36 }, { "epoch": 0.012829403606102635, "grad_norm": 4.147632122039795, "learning_rate": 1e-06, "loss": 0.2474, "step": 37 }, { "epoch": 0.013176144244105409, "grad_norm": 3.017467737197876, "learning_rate": 1e-06, "loss": 0.2443, "step": 38 }, { "epoch": 0.013522884882108182, "grad_norm": 3.92279314994812, "learning_rate": 1e-06, "loss": 0.2309, "step": 39 }, { "epoch": 0.013869625520110958, "grad_norm": 2.6705775260925293, "learning_rate": 1e-06, "loss": 0.2122, "step": 40 }, { "epoch": 0.014216366158113732, "grad_norm": 2.2579407691955566, "learning_rate": 1e-06, "loss": 0.2189, "step": 41 }, { "epoch": 0.014563106796116505, "grad_norm": 3.3366434574127197, "learning_rate": 1e-06, "loss": 0.2268, "step": 42 }, { "epoch": 0.014909847434119279, "grad_norm": 1.8895633220672607, "learning_rate": 1e-06, "loss": 0.2219, "step": 43 }, { "epoch": 0.015256588072122053, "grad_norm": 4.174144268035889, "learning_rate": 1e-06, "loss": 0.209, "step": 44 }, { "epoch": 0.015603328710124827, "grad_norm": 2.0545594692230225, "learning_rate": 1e-06, "loss": 0.2464, "step": 45 }, { "epoch": 0.0159500693481276, "grad_norm": 6.63257360458374, "learning_rate": 1e-06, "loss": 0.2905, "step": 46 }, { "epoch": 0.016296809986130374, "grad_norm": 3.2842488288879395, "learning_rate": 1e-06, "loss": 0.2674, "step": 47 }, { "epoch": 0.016643550624133148, "grad_norm": 2.0572311878204346, "learning_rate": 1e-06, "loss": 0.237, "step": 48 }, { "epoch": 0.01699029126213592, "grad_norm": 1.9547162055969238, "learning_rate": 1e-06, "loss": 0.2268, "step": 49 }, { "epoch": 0.017337031900138695, "grad_norm": 2.775322914123535, "learning_rate": 1e-06, "loss": 0.2315, "step": 50 }, { "epoch": 0.01768377253814147, "grad_norm": 1.8670415878295898, "learning_rate": 1e-06, "loss": 0.21, "step": 51 }, { "epoch": 0.018030513176144243, "grad_norm": 2.3691036701202393, "learning_rate": 1e-06, "loss": 0.2465, "step": 52 }, { "epoch": 0.018377253814147017, "grad_norm": 4.806734561920166, "learning_rate": 1e-06, "loss": 0.2437, "step": 53 }, { "epoch": 0.01872399445214979, "grad_norm": 4.114093780517578, "learning_rate": 1e-06, "loss": 0.2402, "step": 54 }, { "epoch": 0.019070735090152564, "grad_norm": 2.5198612213134766, "learning_rate": 1e-06, "loss": 0.2409, "step": 55 }, { "epoch": 0.019417475728155338, "grad_norm": 2.066634178161621, "learning_rate": 1e-06, "loss": 0.2435, "step": 56 }, { "epoch": 0.019764216366158115, "grad_norm": 2.1365854740142822, "learning_rate": 1e-06, "loss": 0.2321, "step": 57 }, { "epoch": 0.02011095700416089, "grad_norm": 3.2070484161376953, "learning_rate": 1e-06, "loss": 0.2347, "step": 58 }, { "epoch": 0.020457697642163663, "grad_norm": 2.717198133468628, "learning_rate": 1e-06, "loss": 0.1966, "step": 59 }, { "epoch": 0.020804438280166437, "grad_norm": 3.56074857711792, "learning_rate": 1e-06, "loss": 0.2505, "step": 60 }, { "epoch": 0.02115117891816921, "grad_norm": 3.655073642730713, "learning_rate": 1e-06, "loss": 0.2467, "step": 61 }, { "epoch": 0.021497919556171984, "grad_norm": 3.183065414428711, "learning_rate": 1e-06, "loss": 0.2201, "step": 62 }, { "epoch": 0.021844660194174758, "grad_norm": 3.160886764526367, "learning_rate": 1e-06, "loss": 0.2187, "step": 63 }, { "epoch": 0.022191400832177532, "grad_norm": 1.6657497882843018, "learning_rate": 1e-06, "loss": 0.2096, "step": 64 }, { "epoch": 0.022538141470180306, "grad_norm": 1.8679834604263306, "learning_rate": 1e-06, "loss": 0.2478, "step": 65 }, { "epoch": 0.02288488210818308, "grad_norm": 1.8762993812561035, "learning_rate": 1e-06, "loss": 0.218, "step": 66 }, { "epoch": 0.023231622746185853, "grad_norm": 2.1899123191833496, "learning_rate": 1e-06, "loss": 0.2422, "step": 67 }, { "epoch": 0.023578363384188627, "grad_norm": 2.1151392459869385, "learning_rate": 1e-06, "loss": 0.267, "step": 68 }, { "epoch": 0.0239251040221914, "grad_norm": 3.3461737632751465, "learning_rate": 1e-06, "loss": 0.2615, "step": 69 }, { "epoch": 0.024271844660194174, "grad_norm": 1.9115928411483765, "learning_rate": 1e-06, "loss": 0.2385, "step": 70 }, { "epoch": 0.024618585298196948, "grad_norm": 1.9966187477111816, "learning_rate": 1e-06, "loss": 0.2341, "step": 71 }, { "epoch": 0.024965325936199722, "grad_norm": 4.439002990722656, "learning_rate": 1e-06, "loss": 0.2683, "step": 72 }, { "epoch": 0.025312066574202496, "grad_norm": 3.926988124847412, "learning_rate": 1e-06, "loss": 0.2489, "step": 73 }, { "epoch": 0.02565880721220527, "grad_norm": 2.2309978008270264, "learning_rate": 1e-06, "loss": 0.2471, "step": 74 }, { "epoch": 0.026005547850208043, "grad_norm": 3.369230031967163, "learning_rate": 1e-06, "loss": 0.2282, "step": 75 }, { "epoch": 0.026352288488210817, "grad_norm": 4.4944047927856445, "learning_rate": 1e-06, "loss": 0.227, "step": 76 }, { "epoch": 0.02669902912621359, "grad_norm": 5.281230449676514, "learning_rate": 1e-06, "loss": 0.2358, "step": 77 }, { "epoch": 0.027045769764216365, "grad_norm": 2.726041555404663, "learning_rate": 1e-06, "loss": 0.2569, "step": 78 }, { "epoch": 0.027392510402219142, "grad_norm": 2.2781665325164795, "learning_rate": 1e-06, "loss": 0.2436, "step": 79 }, { "epoch": 0.027739251040221916, "grad_norm": 5.479579925537109, "learning_rate": 1e-06, "loss": 0.2588, "step": 80 }, { "epoch": 0.02808599167822469, "grad_norm": 5.336531162261963, "learning_rate": 1e-06, "loss": 0.2315, "step": 81 }, { "epoch": 0.028432732316227463, "grad_norm": 1.9904639720916748, "learning_rate": 1e-06, "loss": 0.249, "step": 82 }, { "epoch": 0.028779472954230237, "grad_norm": 2.493849277496338, "learning_rate": 1e-06, "loss": 0.2586, "step": 83 }, { "epoch": 0.02912621359223301, "grad_norm": 2.0761678218841553, "learning_rate": 1e-06, "loss": 0.2256, "step": 84 }, { "epoch": 0.029472954230235784, "grad_norm": 3.9057183265686035, "learning_rate": 1e-06, "loss": 0.2605, "step": 85 }, { "epoch": 0.029819694868238558, "grad_norm": 1.8108359575271606, "learning_rate": 1e-06, "loss": 0.209, "step": 86 }, { "epoch": 0.030166435506241332, "grad_norm": 3.9006757736206055, "learning_rate": 1e-06, "loss": 0.2302, "step": 87 }, { "epoch": 0.030513176144244106, "grad_norm": 3.529353141784668, "learning_rate": 1e-06, "loss": 0.2497, "step": 88 }, { "epoch": 0.03085991678224688, "grad_norm": 1.9021555185317993, "learning_rate": 1e-06, "loss": 0.2343, "step": 89 }, { "epoch": 0.031206657420249653, "grad_norm": 2.0818121433258057, "learning_rate": 1e-06, "loss": 0.2296, "step": 90 }, { "epoch": 0.03155339805825243, "grad_norm": 2.9069199562072754, "learning_rate": 1e-06, "loss": 0.2336, "step": 91 }, { "epoch": 0.0319001386962552, "grad_norm": 2.5499114990234375, "learning_rate": 1e-06, "loss": 0.2326, "step": 92 }, { "epoch": 0.032246879334257975, "grad_norm": 4.565133571624756, "learning_rate": 1e-06, "loss": 0.2325, "step": 93 }, { "epoch": 0.03259361997226075, "grad_norm": 1.965366005897522, "learning_rate": 1e-06, "loss": 0.2436, "step": 94 }, { "epoch": 0.03294036061026352, "grad_norm": 3.4335970878601074, "learning_rate": 1e-06, "loss": 0.2517, "step": 95 }, { "epoch": 0.033287101248266296, "grad_norm": 2.7447807788848877, "learning_rate": 1e-06, "loss": 0.219, "step": 96 }, { "epoch": 0.03363384188626907, "grad_norm": 1.9556831121444702, "learning_rate": 1e-06, "loss": 0.2071, "step": 97 }, { "epoch": 0.03398058252427184, "grad_norm": 2.175020694732666, "learning_rate": 1e-06, "loss": 0.2481, "step": 98 }, { "epoch": 0.03432732316227462, "grad_norm": 3.7034220695495605, "learning_rate": 1e-06, "loss": 0.2525, "step": 99 }, { "epoch": 0.03467406380027739, "grad_norm": 4.264119625091553, "learning_rate": 1e-06, "loss": 0.21, "step": 100 }, { "epoch": 0.035020804438280165, "grad_norm": 4.019951820373535, "learning_rate": 1e-06, "loss": 0.2495, "step": 101 }, { "epoch": 0.03536754507628294, "grad_norm": 2.4060921669006348, "learning_rate": 1e-06, "loss": 0.2295, "step": 102 }, { "epoch": 0.03571428571428571, "grad_norm": 2.3502883911132812, "learning_rate": 1e-06, "loss": 0.232, "step": 103 }, { "epoch": 0.036061026352288486, "grad_norm": 4.202142715454102, "learning_rate": 1e-06, "loss": 0.2159, "step": 104 }, { "epoch": 0.03640776699029126, "grad_norm": 6.274311542510986, "learning_rate": 1e-06, "loss": 0.2825, "step": 105 }, { "epoch": 0.036754507628294034, "grad_norm": 2.756082057952881, "learning_rate": 1e-06, "loss": 0.251, "step": 106 }, { "epoch": 0.03710124826629681, "grad_norm": 1.8495687246322632, "learning_rate": 1e-06, "loss": 0.2371, "step": 107 }, { "epoch": 0.03744798890429958, "grad_norm": 2.5674338340759277, "learning_rate": 1e-06, "loss": 0.2157, "step": 108 }, { "epoch": 0.037794729542302355, "grad_norm": 3.304002285003662, "learning_rate": 1e-06, "loss": 0.2552, "step": 109 }, { "epoch": 0.03814147018030513, "grad_norm": 4.199469566345215, "learning_rate": 1e-06, "loss": 0.2165, "step": 110 }, { "epoch": 0.0384882108183079, "grad_norm": 2.1667439937591553, "learning_rate": 1e-06, "loss": 0.2639, "step": 111 }, { "epoch": 0.038834951456310676, "grad_norm": 2.8193678855895996, "learning_rate": 1e-06, "loss": 0.2174, "step": 112 }, { "epoch": 0.03918169209431346, "grad_norm": 1.8696953058242798, "learning_rate": 1e-06, "loss": 0.2057, "step": 113 }, { "epoch": 0.03952843273231623, "grad_norm": 2.3659722805023193, "learning_rate": 1e-06, "loss": 0.2731, "step": 114 }, { "epoch": 0.039875173370319004, "grad_norm": 1.782591700553894, "learning_rate": 1e-06, "loss": 0.2287, "step": 115 }, { "epoch": 0.04022191400832178, "grad_norm": 1.8214987516403198, "learning_rate": 1e-06, "loss": 0.2225, "step": 116 }, { "epoch": 0.04056865464632455, "grad_norm": 1.8799686431884766, "learning_rate": 1e-06, "loss": 0.2216, "step": 117 }, { "epoch": 0.040915395284327326, "grad_norm": 2.0166141986846924, "learning_rate": 1e-06, "loss": 0.2429, "step": 118 }, { "epoch": 0.0412621359223301, "grad_norm": 5.748719215393066, "learning_rate": 1e-06, "loss": 0.2297, "step": 119 }, { "epoch": 0.04160887656033287, "grad_norm": 2.980369806289673, "learning_rate": 1e-06, "loss": 0.2321, "step": 120 }, { "epoch": 0.04195561719833565, "grad_norm": 2.343538999557495, "learning_rate": 1e-06, "loss": 0.2129, "step": 121 }, { "epoch": 0.04230235783633842, "grad_norm": 1.8937276601791382, "learning_rate": 1e-06, "loss": 0.2297, "step": 122 }, { "epoch": 0.042649098474341195, "grad_norm": 7.0609354972839355, "learning_rate": 1e-06, "loss": 0.2407, "step": 123 }, { "epoch": 0.04299583911234397, "grad_norm": 1.7186024188995361, "learning_rate": 1e-06, "loss": 0.2042, "step": 124 }, { "epoch": 0.04334257975034674, "grad_norm": 2.7151777744293213, "learning_rate": 1e-06, "loss": 0.2323, "step": 125 }, { "epoch": 0.043689320388349516, "grad_norm": 2.631126642227173, "learning_rate": 1e-06, "loss": 0.258, "step": 126 }, { "epoch": 0.04403606102635229, "grad_norm": 2.087282657623291, "learning_rate": 1e-06, "loss": 0.2013, "step": 127 }, { "epoch": 0.044382801664355064, "grad_norm": 3.90163254737854, "learning_rate": 1e-06, "loss": 0.21, "step": 128 }, { "epoch": 0.04472954230235784, "grad_norm": 5.360821723937988, "learning_rate": 1e-06, "loss": 0.2395, "step": 129 }, { "epoch": 0.04507628294036061, "grad_norm": 2.5944528579711914, "learning_rate": 1e-06, "loss": 0.2365, "step": 130 }, { "epoch": 0.045423023578363385, "grad_norm": 2.6796722412109375, "learning_rate": 1e-06, "loss": 0.2361, "step": 131 }, { "epoch": 0.04576976421636616, "grad_norm": 2.1970760822296143, "learning_rate": 1e-06, "loss": 0.2208, "step": 132 }, { "epoch": 0.04611650485436893, "grad_norm": 2.5466551780700684, "learning_rate": 1e-06, "loss": 0.2339, "step": 133 }, { "epoch": 0.046463245492371706, "grad_norm": 4.48819637298584, "learning_rate": 1e-06, "loss": 0.2848, "step": 134 }, { "epoch": 0.04680998613037448, "grad_norm": 1.9005647897720337, "learning_rate": 1e-06, "loss": 0.21, "step": 135 }, { "epoch": 0.047156726768377254, "grad_norm": 9.14037036895752, "learning_rate": 1e-06, "loss": 0.2805, "step": 136 }, { "epoch": 0.04750346740638003, "grad_norm": 1.840706467628479, "learning_rate": 1e-06, "loss": 0.2, "step": 137 }, { "epoch": 0.0478502080443828, "grad_norm": 5.4056315422058105, "learning_rate": 1e-06, "loss": 0.2754, "step": 138 }, { "epoch": 0.048196948682385575, "grad_norm": 1.9696849584579468, "learning_rate": 1e-06, "loss": 0.1991, "step": 139 }, { "epoch": 0.04854368932038835, "grad_norm": 1.8743115663528442, "learning_rate": 1e-06, "loss": 0.2051, "step": 140 }, { "epoch": 0.04889042995839112, "grad_norm": 2.6441774368286133, "learning_rate": 1e-06, "loss": 0.2093, "step": 141 }, { "epoch": 0.049237170596393896, "grad_norm": 1.965954303741455, "learning_rate": 1e-06, "loss": 0.2259, "step": 142 }, { "epoch": 0.04958391123439667, "grad_norm": 3.0601775646209717, "learning_rate": 1e-06, "loss": 0.2273, "step": 143 }, { "epoch": 0.049930651872399444, "grad_norm": 2.259979009628296, "learning_rate": 1e-06, "loss": 0.2355, "step": 144 }, { "epoch": 0.05027739251040222, "grad_norm": 3.249948024749756, "learning_rate": 1e-06, "loss": 0.2634, "step": 145 }, { "epoch": 0.05062413314840499, "grad_norm": 4.351866245269775, "learning_rate": 1e-06, "loss": 0.2353, "step": 146 }, { "epoch": 0.050970873786407765, "grad_norm": 1.9071911573410034, "learning_rate": 1e-06, "loss": 0.2508, "step": 147 }, { "epoch": 0.05131761442441054, "grad_norm": 4.363613128662109, "learning_rate": 1e-06, "loss": 0.2179, "step": 148 }, { "epoch": 0.05166435506241331, "grad_norm": 3.450490713119507, "learning_rate": 1e-06, "loss": 0.2519, "step": 149 }, { "epoch": 0.052011095700416086, "grad_norm": 4.706002712249756, "learning_rate": 1e-06, "loss": 0.2135, "step": 150 }, { "epoch": 0.05235783633841886, "grad_norm": 2.632718324661255, "learning_rate": 1e-06, "loss": 0.2218, "step": 151 }, { "epoch": 0.052704576976421634, "grad_norm": 3.9972386360168457, "learning_rate": 1e-06, "loss": 0.2227, "step": 152 }, { "epoch": 0.05305131761442441, "grad_norm": 1.96170175075531, "learning_rate": 1e-06, "loss": 0.2189, "step": 153 }, { "epoch": 0.05339805825242718, "grad_norm": 4.996349811553955, "learning_rate": 1e-06, "loss": 0.2225, "step": 154 }, { "epoch": 0.053744798890429955, "grad_norm": 5.351596832275391, "learning_rate": 1e-06, "loss": 0.2043, "step": 155 }, { "epoch": 0.05409153952843273, "grad_norm": 2.1169140338897705, "learning_rate": 1e-06, "loss": 0.2483, "step": 156 }, { "epoch": 0.0544382801664355, "grad_norm": 2.441293478012085, "learning_rate": 1e-06, "loss": 0.216, "step": 157 }, { "epoch": 0.054785020804438284, "grad_norm": 2.3679656982421875, "learning_rate": 1e-06, "loss": 0.2292, "step": 158 }, { "epoch": 0.05513176144244106, "grad_norm": 5.575881004333496, "learning_rate": 1e-06, "loss": 0.2512, "step": 159 }, { "epoch": 0.05547850208044383, "grad_norm": 2.48142671585083, "learning_rate": 1e-06, "loss": 0.2249, "step": 160 }, { "epoch": 0.055825242718446605, "grad_norm": 3.078421115875244, "learning_rate": 1e-06, "loss": 0.2205, "step": 161 }, { "epoch": 0.05617198335644938, "grad_norm": 2.7186150550842285, "learning_rate": 1e-06, "loss": 0.2208, "step": 162 }, { "epoch": 0.05651872399445215, "grad_norm": 2.42346453666687, "learning_rate": 1e-06, "loss": 0.2404, "step": 163 }, { "epoch": 0.056865464632454926, "grad_norm": 2.095087766647339, "learning_rate": 1e-06, "loss": 0.2354, "step": 164 }, { "epoch": 0.0572122052704577, "grad_norm": 2.2342190742492676, "learning_rate": 1e-06, "loss": 0.2002, "step": 165 }, { "epoch": 0.057558945908460474, "grad_norm": 1.9590411186218262, "learning_rate": 1e-06, "loss": 0.2073, "step": 166 }, { "epoch": 0.05790568654646325, "grad_norm": 4.556865692138672, "learning_rate": 1e-06, "loss": 0.2742, "step": 167 }, { "epoch": 0.05825242718446602, "grad_norm": 2.7081010341644287, "learning_rate": 1e-06, "loss": 0.2455, "step": 168 }, { "epoch": 0.058599167822468795, "grad_norm": 2.4702224731445312, "learning_rate": 1e-06, "loss": 0.2209, "step": 169 }, { "epoch": 0.05894590846047157, "grad_norm": 2.41284441947937, "learning_rate": 1e-06, "loss": 0.259, "step": 170 }, { "epoch": 0.05929264909847434, "grad_norm": 3.244028091430664, "learning_rate": 1e-06, "loss": 0.2358, "step": 171 }, { "epoch": 0.059639389736477116, "grad_norm": 2.7794923782348633, "learning_rate": 1e-06, "loss": 0.1973, "step": 172 }, { "epoch": 0.05998613037447989, "grad_norm": 3.4682586193084717, "learning_rate": 1e-06, "loss": 0.2523, "step": 173 }, { "epoch": 0.060332871012482664, "grad_norm": 2.3559231758117676, "learning_rate": 1e-06, "loss": 0.2529, "step": 174 }, { "epoch": 0.06067961165048544, "grad_norm": 3.0808186531066895, "learning_rate": 1e-06, "loss": 0.1952, "step": 175 }, { "epoch": 0.06102635228848821, "grad_norm": 2.975972890853882, "learning_rate": 1e-06, "loss": 0.2226, "step": 176 }, { "epoch": 0.061373092926490985, "grad_norm": 2.133094072341919, "learning_rate": 1e-06, "loss": 0.2202, "step": 177 }, { "epoch": 0.06171983356449376, "grad_norm": 1.9771744012832642, "learning_rate": 1e-06, "loss": 0.2025, "step": 178 }, { "epoch": 0.06206657420249653, "grad_norm": 1.7458380460739136, "learning_rate": 1e-06, "loss": 0.1922, "step": 179 }, { "epoch": 0.06241331484049931, "grad_norm": 2.3998160362243652, "learning_rate": 1e-06, "loss": 0.1807, "step": 180 }, { "epoch": 0.06276005547850208, "grad_norm": 2.3493542671203613, "learning_rate": 1e-06, "loss": 0.2146, "step": 181 }, { "epoch": 0.06310679611650485, "grad_norm": 6.937662124633789, "learning_rate": 1e-06, "loss": 0.2579, "step": 182 }, { "epoch": 0.06345353675450763, "grad_norm": 1.6203336715698242, "learning_rate": 1e-06, "loss": 0.198, "step": 183 }, { "epoch": 0.0638002773925104, "grad_norm": 1.5239827632904053, "learning_rate": 1e-06, "loss": 0.189, "step": 184 }, { "epoch": 0.06414701803051318, "grad_norm": 3.796680450439453, "learning_rate": 1e-06, "loss": 0.2651, "step": 185 }, { "epoch": 0.06449375866851595, "grad_norm": 2.5366694927215576, "learning_rate": 1e-06, "loss": 0.2121, "step": 186 }, { "epoch": 0.06484049930651872, "grad_norm": 1.8307462930679321, "learning_rate": 1e-06, "loss": 0.2303, "step": 187 }, { "epoch": 0.0651872399445215, "grad_norm": 4.409234523773193, "learning_rate": 1e-06, "loss": 0.2159, "step": 188 }, { "epoch": 0.06553398058252427, "grad_norm": 4.105565547943115, "learning_rate": 1e-06, "loss": 0.2293, "step": 189 }, { "epoch": 0.06588072122052704, "grad_norm": 4.292323589324951, "learning_rate": 1e-06, "loss": 0.2063, "step": 190 }, { "epoch": 0.06622746185852982, "grad_norm": 2.141298294067383, "learning_rate": 1e-06, "loss": 0.1966, "step": 191 }, { "epoch": 0.06657420249653259, "grad_norm": 2.1823067665100098, "learning_rate": 1e-06, "loss": 0.2019, "step": 192 }, { "epoch": 0.06692094313453537, "grad_norm": 4.099167823791504, "learning_rate": 1e-06, "loss": 0.237, "step": 193 }, { "epoch": 0.06726768377253814, "grad_norm": 2.6066272258758545, "learning_rate": 1e-06, "loss": 0.2275, "step": 194 }, { "epoch": 0.06761442441054091, "grad_norm": 2.4590089321136475, "learning_rate": 1e-06, "loss": 0.2467, "step": 195 }, { "epoch": 0.06796116504854369, "grad_norm": 3.860490083694458, "learning_rate": 1e-06, "loss": 0.2552, "step": 196 }, { "epoch": 0.06830790568654646, "grad_norm": 2.6223294734954834, "learning_rate": 1e-06, "loss": 0.2669, "step": 197 }, { "epoch": 0.06865464632454923, "grad_norm": 3.189718723297119, "learning_rate": 1e-06, "loss": 0.2096, "step": 198 }, { "epoch": 0.06900138696255201, "grad_norm": 5.834605693817139, "learning_rate": 1e-06, "loss": 0.2272, "step": 199 }, { "epoch": 0.06934812760055478, "grad_norm": 4.342336177825928, "learning_rate": 1e-06, "loss": 0.205, "step": 200 }, { "epoch": 0.06969486823855756, "grad_norm": 5.050157070159912, "learning_rate": 1e-06, "loss": 0.2252, "step": 201 }, { "epoch": 0.07004160887656033, "grad_norm": 1.8504279851913452, "learning_rate": 1e-06, "loss": 0.1912, "step": 202 }, { "epoch": 0.0703883495145631, "grad_norm": 2.362457275390625, "learning_rate": 1e-06, "loss": 0.2248, "step": 203 }, { "epoch": 0.07073509015256588, "grad_norm": 4.938218116760254, "learning_rate": 1e-06, "loss": 0.2119, "step": 204 }, { "epoch": 0.07108183079056865, "grad_norm": 4.182404041290283, "learning_rate": 1e-06, "loss": 0.2224, "step": 205 }, { "epoch": 0.07142857142857142, "grad_norm": 1.9187567234039307, "learning_rate": 1e-06, "loss": 0.2193, "step": 206 }, { "epoch": 0.0717753120665742, "grad_norm": 4.429518699645996, "learning_rate": 1e-06, "loss": 0.2174, "step": 207 }, { "epoch": 0.07212205270457697, "grad_norm": 5.2877044677734375, "learning_rate": 1e-06, "loss": 0.2216, "step": 208 }, { "epoch": 0.07246879334257975, "grad_norm": 4.316680431365967, "learning_rate": 1e-06, "loss": 0.254, "step": 209 }, { "epoch": 0.07281553398058252, "grad_norm": 2.9150466918945312, "learning_rate": 1e-06, "loss": 0.2083, "step": 210 }, { "epoch": 0.0731622746185853, "grad_norm": 1.7052415609359741, "learning_rate": 1e-06, "loss": 0.2126, "step": 211 }, { "epoch": 0.07350901525658807, "grad_norm": 3.053804636001587, "learning_rate": 1e-06, "loss": 0.2239, "step": 212 }, { "epoch": 0.07385575589459084, "grad_norm": 2.949108123779297, "learning_rate": 1e-06, "loss": 0.2101, "step": 213 }, { "epoch": 0.07420249653259361, "grad_norm": 1.8862674236297607, "learning_rate": 1e-06, "loss": 0.2066, "step": 214 }, { "epoch": 0.07454923717059639, "grad_norm": 2.371164321899414, "learning_rate": 1e-06, "loss": 0.2532, "step": 215 }, { "epoch": 0.07489597780859916, "grad_norm": 2.709270477294922, "learning_rate": 1e-06, "loss": 0.1968, "step": 216 }, { "epoch": 0.07524271844660194, "grad_norm": 5.096902847290039, "learning_rate": 1e-06, "loss": 0.2111, "step": 217 }, { "epoch": 0.07558945908460471, "grad_norm": 2.7359046936035156, "learning_rate": 1e-06, "loss": 0.1927, "step": 218 }, { "epoch": 0.07593619972260748, "grad_norm": 1.800238847732544, "learning_rate": 1e-06, "loss": 0.1898, "step": 219 }, { "epoch": 0.07628294036061026, "grad_norm": 4.802062511444092, "learning_rate": 1e-06, "loss": 0.221, "step": 220 }, { "epoch": 0.07662968099861303, "grad_norm": 2.447291374206543, "learning_rate": 1e-06, "loss": 0.2287, "step": 221 }, { "epoch": 0.0769764216366158, "grad_norm": 7.738630771636963, "learning_rate": 1e-06, "loss": 0.2055, "step": 222 }, { "epoch": 0.07732316227461858, "grad_norm": 2.3890295028686523, "learning_rate": 1e-06, "loss": 0.2246, "step": 223 }, { "epoch": 0.07766990291262135, "grad_norm": 4.384284496307373, "learning_rate": 1e-06, "loss": 0.2348, "step": 224 }, { "epoch": 0.07801664355062413, "grad_norm": 2.8400914669036865, "learning_rate": 1e-06, "loss": 0.2089, "step": 225 }, { "epoch": 0.07836338418862691, "grad_norm": 2.7480292320251465, "learning_rate": 1e-06, "loss": 0.2736, "step": 226 }, { "epoch": 0.07871012482662969, "grad_norm": 5.927616596221924, "learning_rate": 1e-06, "loss": 0.2613, "step": 227 }, { "epoch": 0.07905686546463246, "grad_norm": 5.0669846534729, "learning_rate": 1e-06, "loss": 0.24, "step": 228 }, { "epoch": 0.07940360610263524, "grad_norm": 5.679924011230469, "learning_rate": 1e-06, "loss": 0.2188, "step": 229 }, { "epoch": 0.07975034674063801, "grad_norm": 3.200345277786255, "learning_rate": 1e-06, "loss": 0.2018, "step": 230 }, { "epoch": 0.08009708737864078, "grad_norm": 3.164886474609375, "learning_rate": 1e-06, "loss": 0.2325, "step": 231 }, { "epoch": 0.08044382801664356, "grad_norm": 3.412248373031616, "learning_rate": 1e-06, "loss": 0.2029, "step": 232 }, { "epoch": 0.08079056865464633, "grad_norm": 4.144454479217529, "learning_rate": 1e-06, "loss": 0.1915, "step": 233 }, { "epoch": 0.0811373092926491, "grad_norm": 2.3590505123138428, "learning_rate": 1e-06, "loss": 0.2341, "step": 234 }, { "epoch": 0.08148404993065188, "grad_norm": 1.8131296634674072, "learning_rate": 1e-06, "loss": 0.2044, "step": 235 }, { "epoch": 0.08183079056865465, "grad_norm": 3.05785870552063, "learning_rate": 1e-06, "loss": 0.2044, "step": 236 }, { "epoch": 0.08217753120665743, "grad_norm": 3.1392011642456055, "learning_rate": 1e-06, "loss": 0.203, "step": 237 }, { "epoch": 0.0825242718446602, "grad_norm": 2.5274972915649414, "learning_rate": 1e-06, "loss": 0.2357, "step": 238 }, { "epoch": 0.08287101248266297, "grad_norm": 1.6625158786773682, "learning_rate": 1e-06, "loss": 0.2181, "step": 239 }, { "epoch": 0.08321775312066575, "grad_norm": 1.920361876487732, "learning_rate": 1e-06, "loss": 0.2186, "step": 240 }, { "epoch": 0.08356449375866852, "grad_norm": 3.1764955520629883, "learning_rate": 1e-06, "loss": 0.1844, "step": 241 }, { "epoch": 0.0839112343966713, "grad_norm": 3.478422164916992, "learning_rate": 1e-06, "loss": 0.2248, "step": 242 }, { "epoch": 0.08425797503467407, "grad_norm": 2.3174679279327393, "learning_rate": 1e-06, "loss": 0.2345, "step": 243 }, { "epoch": 0.08460471567267684, "grad_norm": 3.9944794178009033, "learning_rate": 1e-06, "loss": 0.1961, "step": 244 }, { "epoch": 0.08495145631067962, "grad_norm": 4.681294918060303, "learning_rate": 1e-06, "loss": 0.2107, "step": 245 }, { "epoch": 0.08529819694868239, "grad_norm": 1.9410604238510132, "learning_rate": 1e-06, "loss": 0.2098, "step": 246 }, { "epoch": 0.08564493758668516, "grad_norm": 6.453781604766846, "learning_rate": 1e-06, "loss": 0.2285, "step": 247 }, { "epoch": 0.08599167822468794, "grad_norm": 1.9601918458938599, "learning_rate": 1e-06, "loss": 0.2293, "step": 248 }, { "epoch": 0.08633841886269071, "grad_norm": 3.6088569164276123, "learning_rate": 1e-06, "loss": 0.2397, "step": 249 }, { "epoch": 0.08668515950069348, "grad_norm": 6.8316826820373535, "learning_rate": 1e-06, "loss": 0.2361, "step": 250 }, { "epoch": 0.08703190013869626, "grad_norm": 4.1882147789001465, "learning_rate": 1e-06, "loss": 0.2352, "step": 251 }, { "epoch": 0.08737864077669903, "grad_norm": 2.6391122341156006, "learning_rate": 1e-06, "loss": 0.2207, "step": 252 }, { "epoch": 0.0877253814147018, "grad_norm": 1.6957664489746094, "learning_rate": 1e-06, "loss": 0.1932, "step": 253 }, { "epoch": 0.08807212205270458, "grad_norm": 2.1076745986938477, "learning_rate": 1e-06, "loss": 0.2082, "step": 254 }, { "epoch": 0.08841886269070735, "grad_norm": 2.3015823364257812, "learning_rate": 1e-06, "loss": 0.2228, "step": 255 }, { "epoch": 0.08876560332871013, "grad_norm": 4.031891822814941, "learning_rate": 1e-06, "loss": 0.2055, "step": 256 }, { "epoch": 0.0891123439667129, "grad_norm": 2.306821346282959, "learning_rate": 1e-06, "loss": 0.2286, "step": 257 }, { "epoch": 0.08945908460471567, "grad_norm": 2.1814613342285156, "learning_rate": 1e-06, "loss": 0.2046, "step": 258 }, { "epoch": 0.08980582524271845, "grad_norm": 5.232451438903809, "learning_rate": 1e-06, "loss": 0.216, "step": 259 }, { "epoch": 0.09015256588072122, "grad_norm": 7.794690132141113, "learning_rate": 1e-06, "loss": 0.1838, "step": 260 }, { "epoch": 0.090499306518724, "grad_norm": 3.107283115386963, "learning_rate": 1e-06, "loss": 0.2261, "step": 261 }, { "epoch": 0.09084604715672677, "grad_norm": 1.7422479391098022, "learning_rate": 1e-06, "loss": 0.2222, "step": 262 }, { "epoch": 0.09119278779472954, "grad_norm": 2.579590320587158, "learning_rate": 1e-06, "loss": 0.2015, "step": 263 }, { "epoch": 0.09153952843273232, "grad_norm": 3.788679838180542, "learning_rate": 1e-06, "loss": 0.2373, "step": 264 }, { "epoch": 0.09188626907073509, "grad_norm": 2.0979104042053223, "learning_rate": 1e-06, "loss": 0.2183, "step": 265 }, { "epoch": 0.09223300970873786, "grad_norm": 2.019399642944336, "learning_rate": 1e-06, "loss": 0.2195, "step": 266 }, { "epoch": 0.09257975034674064, "grad_norm": 2.5858969688415527, "learning_rate": 1e-06, "loss": 0.234, "step": 267 }, { "epoch": 0.09292649098474341, "grad_norm": 2.279453754425049, "learning_rate": 1e-06, "loss": 0.2312, "step": 268 }, { "epoch": 0.09327323162274619, "grad_norm": 3.3690133094787598, "learning_rate": 1e-06, "loss": 0.2499, "step": 269 }, { "epoch": 0.09361997226074896, "grad_norm": 1.591291904449463, "learning_rate": 1e-06, "loss": 0.1986, "step": 270 }, { "epoch": 0.09396671289875173, "grad_norm": 2.8157074451446533, "learning_rate": 1e-06, "loss": 0.2185, "step": 271 }, { "epoch": 0.09431345353675451, "grad_norm": 2.9287891387939453, "learning_rate": 1e-06, "loss": 0.2151, "step": 272 }, { "epoch": 0.09466019417475728, "grad_norm": 3.26141357421875, "learning_rate": 1e-06, "loss": 0.2295, "step": 273 }, { "epoch": 0.09500693481276005, "grad_norm": 1.793872594833374, "learning_rate": 1e-06, "loss": 0.2053, "step": 274 }, { "epoch": 0.09535367545076283, "grad_norm": 6.906111717224121, "learning_rate": 1e-06, "loss": 0.1894, "step": 275 }, { "epoch": 0.0957004160887656, "grad_norm": 5.939563751220703, "learning_rate": 1e-06, "loss": 0.2454, "step": 276 }, { "epoch": 0.09604715672676838, "grad_norm": 4.869255065917969, "learning_rate": 1e-06, "loss": 0.2029, "step": 277 }, { "epoch": 0.09639389736477115, "grad_norm": 2.0359835624694824, "learning_rate": 1e-06, "loss": 0.2193, "step": 278 }, { "epoch": 0.09674063800277392, "grad_norm": 1.732591152191162, "learning_rate": 1e-06, "loss": 0.2223, "step": 279 }, { "epoch": 0.0970873786407767, "grad_norm": 2.112185478210449, "learning_rate": 1e-06, "loss": 0.2261, "step": 280 }, { "epoch": 0.09743411927877947, "grad_norm": 1.9228016138076782, "learning_rate": 1e-06, "loss": 0.2168, "step": 281 }, { "epoch": 0.09778085991678225, "grad_norm": 2.5188448429107666, "learning_rate": 1e-06, "loss": 0.2526, "step": 282 }, { "epoch": 0.09812760055478502, "grad_norm": 6.440067291259766, "learning_rate": 1e-06, "loss": 0.1962, "step": 283 }, { "epoch": 0.09847434119278779, "grad_norm": 2.1563827991485596, "learning_rate": 1e-06, "loss": 0.2155, "step": 284 }, { "epoch": 0.09882108183079057, "grad_norm": 2.890427589416504, "learning_rate": 1e-06, "loss": 0.204, "step": 285 }, { "epoch": 0.09916782246879334, "grad_norm": 3.064075469970703, "learning_rate": 1e-06, "loss": 0.2124, "step": 286 }, { "epoch": 0.09951456310679611, "grad_norm": 2.013902187347412, "learning_rate": 1e-06, "loss": 0.2011, "step": 287 }, { "epoch": 0.09986130374479889, "grad_norm": 5.711750507354736, "learning_rate": 1e-06, "loss": 0.198, "step": 288 }, { "epoch": 0.10020804438280166, "grad_norm": 4.605215072631836, "learning_rate": 1e-06, "loss": 0.2074, "step": 289 }, { "epoch": 0.10055478502080444, "grad_norm": 3.5413124561309814, "learning_rate": 1e-06, "loss": 0.2516, "step": 290 }, { "epoch": 0.10090152565880721, "grad_norm": 2.2203333377838135, "learning_rate": 1e-06, "loss": 0.2353, "step": 291 }, { "epoch": 0.10124826629680998, "grad_norm": 3.4780688285827637, "learning_rate": 1e-06, "loss": 0.2278, "step": 292 }, { "epoch": 0.10159500693481276, "grad_norm": 2.2636001110076904, "learning_rate": 1e-06, "loss": 0.1992, "step": 293 }, { "epoch": 0.10194174757281553, "grad_norm": 4.070486545562744, "learning_rate": 1e-06, "loss": 0.2037, "step": 294 }, { "epoch": 0.1022884882108183, "grad_norm": 2.2457258701324463, "learning_rate": 1e-06, "loss": 0.2409, "step": 295 }, { "epoch": 0.10263522884882108, "grad_norm": 2.1032586097717285, "learning_rate": 1e-06, "loss": 0.2105, "step": 296 }, { "epoch": 0.10298196948682385, "grad_norm": 2.8881735801696777, "learning_rate": 1e-06, "loss": 0.2115, "step": 297 }, { "epoch": 0.10332871012482663, "grad_norm": 2.166180372238159, "learning_rate": 1e-06, "loss": 0.2231, "step": 298 }, { "epoch": 0.1036754507628294, "grad_norm": 2.0594778060913086, "learning_rate": 1e-06, "loss": 0.2409, "step": 299 }, { "epoch": 0.10402219140083217, "grad_norm": 1.8544009923934937, "learning_rate": 1e-06, "loss": 0.1873, "step": 300 }, { "epoch": 0.10436893203883495, "grad_norm": 1.979411005973816, "learning_rate": 1e-06, "loss": 0.2177, "step": 301 }, { "epoch": 0.10471567267683772, "grad_norm": 2.5190136432647705, "learning_rate": 1e-06, "loss": 0.2206, "step": 302 }, { "epoch": 0.1050624133148405, "grad_norm": 2.110924482345581, "learning_rate": 1e-06, "loss": 0.2087, "step": 303 }, { "epoch": 0.10540915395284327, "grad_norm": 5.314784526824951, "learning_rate": 1e-06, "loss": 0.1707, "step": 304 }, { "epoch": 0.10575589459084604, "grad_norm": 3.8186211585998535, "learning_rate": 1e-06, "loss": 0.2255, "step": 305 }, { "epoch": 0.10610263522884882, "grad_norm": 2.197397232055664, "learning_rate": 1e-06, "loss": 0.2026, "step": 306 }, { "epoch": 0.10644937586685159, "grad_norm": 3.364658832550049, "learning_rate": 1e-06, "loss": 0.237, "step": 307 }, { "epoch": 0.10679611650485436, "grad_norm": 3.6538567543029785, "learning_rate": 1e-06, "loss": 0.2135, "step": 308 }, { "epoch": 0.10714285714285714, "grad_norm": 2.080467462539673, "learning_rate": 1e-06, "loss": 0.2011, "step": 309 }, { "epoch": 0.10748959778085991, "grad_norm": 2.394052267074585, "learning_rate": 1e-06, "loss": 0.2081, "step": 310 }, { "epoch": 0.10783633841886268, "grad_norm": 3.270670175552368, "learning_rate": 1e-06, "loss": 0.2139, "step": 311 }, { "epoch": 0.10818307905686546, "grad_norm": 4.5176682472229, "learning_rate": 1e-06, "loss": 0.2214, "step": 312 }, { "epoch": 0.10852981969486823, "grad_norm": 2.9834976196289062, "learning_rate": 1e-06, "loss": 0.2235, "step": 313 }, { "epoch": 0.108876560332871, "grad_norm": 3.658803701400757, "learning_rate": 1e-06, "loss": 0.1903, "step": 314 }, { "epoch": 0.10922330097087378, "grad_norm": 2.972036600112915, "learning_rate": 1e-06, "loss": 0.2044, "step": 315 }, { "epoch": 0.10957004160887657, "grad_norm": 2.014233350753784, "learning_rate": 1e-06, "loss": 0.2177, "step": 316 }, { "epoch": 0.10991678224687934, "grad_norm": 3.3329989910125732, "learning_rate": 1e-06, "loss": 0.212, "step": 317 }, { "epoch": 0.11026352288488211, "grad_norm": 4.286015510559082, "learning_rate": 1e-06, "loss": 0.2164, "step": 318 }, { "epoch": 0.11061026352288489, "grad_norm": 2.517414093017578, "learning_rate": 1e-06, "loss": 0.2443, "step": 319 }, { "epoch": 0.11095700416088766, "grad_norm": 2.3605127334594727, "learning_rate": 1e-06, "loss": 0.2136, "step": 320 }, { "epoch": 0.11130374479889044, "grad_norm": 2.0342857837677, "learning_rate": 1e-06, "loss": 0.2117, "step": 321 }, { "epoch": 0.11165048543689321, "grad_norm": 3.313946485519409, "learning_rate": 1e-06, "loss": 0.2159, "step": 322 }, { "epoch": 0.11199722607489598, "grad_norm": 2.8215742111206055, "learning_rate": 1e-06, "loss": 0.1666, "step": 323 }, { "epoch": 0.11234396671289876, "grad_norm": 3.7066569328308105, "learning_rate": 1e-06, "loss": 0.2115, "step": 324 }, { "epoch": 0.11269070735090153, "grad_norm": 2.824295997619629, "learning_rate": 1e-06, "loss": 0.2068, "step": 325 }, { "epoch": 0.1130374479889043, "grad_norm": 3.302109956741333, "learning_rate": 1e-06, "loss": 0.2007, "step": 326 }, { "epoch": 0.11338418862690708, "grad_norm": 1.7297828197479248, "learning_rate": 1e-06, "loss": 0.1796, "step": 327 }, { "epoch": 0.11373092926490985, "grad_norm": 2.3774609565734863, "learning_rate": 1e-06, "loss": 0.197, "step": 328 }, { "epoch": 0.11407766990291263, "grad_norm": 3.876222610473633, "learning_rate": 1e-06, "loss": 0.2106, "step": 329 }, { "epoch": 0.1144244105409154, "grad_norm": 2.891040086746216, "learning_rate": 1e-06, "loss": 0.1809, "step": 330 }, { "epoch": 0.11477115117891817, "grad_norm": 3.9112064838409424, "learning_rate": 1e-06, "loss": 0.2031, "step": 331 }, { "epoch": 0.11511789181692095, "grad_norm": 1.772921085357666, "learning_rate": 1e-06, "loss": 0.2011, "step": 332 }, { "epoch": 0.11546463245492372, "grad_norm": 2.6547513008117676, "learning_rate": 1e-06, "loss": 0.2123, "step": 333 }, { "epoch": 0.1158113730929265, "grad_norm": 5.290209770202637, "learning_rate": 1e-06, "loss": 0.2275, "step": 334 }, { "epoch": 0.11615811373092927, "grad_norm": 5.404000759124756, "learning_rate": 1e-06, "loss": 0.2422, "step": 335 }, { "epoch": 0.11650485436893204, "grad_norm": 2.187218427658081, "learning_rate": 1e-06, "loss": 0.1854, "step": 336 }, { "epoch": 0.11685159500693482, "grad_norm": 3.028299570083618, "learning_rate": 1e-06, "loss": 0.2331, "step": 337 }, { "epoch": 0.11719833564493759, "grad_norm": 2.9846150875091553, "learning_rate": 1e-06, "loss": 0.2288, "step": 338 }, { "epoch": 0.11754507628294036, "grad_norm": 3.630133867263794, "learning_rate": 1e-06, "loss": 0.1559, "step": 339 }, { "epoch": 0.11789181692094314, "grad_norm": 6.968530654907227, "learning_rate": 1e-06, "loss": 0.2637, "step": 340 }, { "epoch": 0.11823855755894591, "grad_norm": 2.9122650623321533, "learning_rate": 1e-06, "loss": 0.1821, "step": 341 }, { "epoch": 0.11858529819694869, "grad_norm": 2.7790746688842773, "learning_rate": 1e-06, "loss": 0.1917, "step": 342 }, { "epoch": 0.11893203883495146, "grad_norm": 4.776673793792725, "learning_rate": 1e-06, "loss": 0.1928, "step": 343 }, { "epoch": 0.11927877947295423, "grad_norm": 2.7039010524749756, "learning_rate": 1e-06, "loss": 0.2258, "step": 344 }, { "epoch": 0.119625520110957, "grad_norm": 2.733076333999634, "learning_rate": 1e-06, "loss": 0.1988, "step": 345 }, { "epoch": 0.11997226074895978, "grad_norm": 3.1806342601776123, "learning_rate": 1e-06, "loss": 0.2147, "step": 346 }, { "epoch": 0.12031900138696255, "grad_norm": 2.9872305393218994, "learning_rate": 1e-06, "loss": 0.1832, "step": 347 }, { "epoch": 0.12066574202496533, "grad_norm": 6.822136402130127, "learning_rate": 1e-06, "loss": 0.2202, "step": 348 }, { "epoch": 0.1210124826629681, "grad_norm": 2.3304481506347656, "learning_rate": 1e-06, "loss": 0.2352, "step": 349 }, { "epoch": 0.12135922330097088, "grad_norm": 2.4243297576904297, "learning_rate": 1e-06, "loss": 0.2004, "step": 350 }, { "epoch": 0.12170596393897365, "grad_norm": 3.050774097442627, "learning_rate": 1e-06, "loss": 0.2103, "step": 351 }, { "epoch": 0.12205270457697642, "grad_norm": 2.177351951599121, "learning_rate": 1e-06, "loss": 0.2416, "step": 352 }, { "epoch": 0.1223994452149792, "grad_norm": 4.790650844573975, "learning_rate": 1e-06, "loss": 0.2174, "step": 353 }, { "epoch": 0.12274618585298197, "grad_norm": 1.6365033388137817, "learning_rate": 1e-06, "loss": 0.1874, "step": 354 }, { "epoch": 0.12309292649098474, "grad_norm": 3.9262518882751465, "learning_rate": 1e-06, "loss": 0.2175, "step": 355 }, { "epoch": 0.12343966712898752, "grad_norm": 1.7626748085021973, "learning_rate": 1e-06, "loss": 0.2129, "step": 356 }, { "epoch": 0.12378640776699029, "grad_norm": 1.7612718343734741, "learning_rate": 1e-06, "loss": 0.1816, "step": 357 }, { "epoch": 0.12413314840499307, "grad_norm": 1.9998083114624023, "learning_rate": 1e-06, "loss": 0.1955, "step": 358 }, { "epoch": 0.12447988904299584, "grad_norm": 1.9432895183563232, "learning_rate": 1e-06, "loss": 0.1793, "step": 359 }, { "epoch": 0.12482662968099861, "grad_norm": 2.2494137287139893, "learning_rate": 1e-06, "loss": 0.2182, "step": 360 }, { "epoch": 0.1251733703190014, "grad_norm": 2.6575872898101807, "learning_rate": 1e-06, "loss": 0.2126, "step": 361 }, { "epoch": 0.12552011095700416, "grad_norm": 2.1404902935028076, "learning_rate": 1e-06, "loss": 0.1727, "step": 362 }, { "epoch": 0.12586685159500693, "grad_norm": 3.4852030277252197, "learning_rate": 1e-06, "loss": 0.2132, "step": 363 }, { "epoch": 0.1262135922330097, "grad_norm": 1.9237537384033203, "learning_rate": 1e-06, "loss": 0.1882, "step": 364 }, { "epoch": 0.12656033287101248, "grad_norm": 2.6047005653381348, "learning_rate": 1e-06, "loss": 0.2276, "step": 365 }, { "epoch": 0.12690707350901526, "grad_norm": 4.0389838218688965, "learning_rate": 1e-06, "loss": 0.2294, "step": 366 }, { "epoch": 0.12725381414701803, "grad_norm": 2.9058427810668945, "learning_rate": 1e-06, "loss": 0.1704, "step": 367 }, { "epoch": 0.1276005547850208, "grad_norm": 2.214973211288452, "learning_rate": 1e-06, "loss": 0.1933, "step": 368 }, { "epoch": 0.12794729542302358, "grad_norm": 2.8029608726501465, "learning_rate": 1e-06, "loss": 0.238, "step": 369 }, { "epoch": 0.12829403606102635, "grad_norm": 3.4244496822357178, "learning_rate": 1e-06, "loss": 0.2404, "step": 370 }, { "epoch": 0.12864077669902912, "grad_norm": 3.9695568084716797, "learning_rate": 1e-06, "loss": 0.2299, "step": 371 }, { "epoch": 0.1289875173370319, "grad_norm": 2.100534439086914, "learning_rate": 1e-06, "loss": 0.1878, "step": 372 }, { "epoch": 0.12933425797503467, "grad_norm": 1.942069172859192, "learning_rate": 1e-06, "loss": 0.227, "step": 373 }, { "epoch": 0.12968099861303745, "grad_norm": 2.7720251083374023, "learning_rate": 1e-06, "loss": 0.2054, "step": 374 }, { "epoch": 0.13002773925104022, "grad_norm": 1.916839599609375, "learning_rate": 1e-06, "loss": 0.2135, "step": 375 }, { "epoch": 0.130374479889043, "grad_norm": 2.759535312652588, "learning_rate": 1e-06, "loss": 0.2327, "step": 376 }, { "epoch": 0.13072122052704577, "grad_norm": 2.059061288833618, "learning_rate": 1e-06, "loss": 0.199, "step": 377 }, { "epoch": 0.13106796116504854, "grad_norm": 1.892815113067627, "learning_rate": 1e-06, "loss": 0.1972, "step": 378 }, { "epoch": 0.13141470180305131, "grad_norm": 2.945600748062134, "learning_rate": 1e-06, "loss": 0.2046, "step": 379 }, { "epoch": 0.1317614424410541, "grad_norm": 2.75258207321167, "learning_rate": 1e-06, "loss": 0.2171, "step": 380 }, { "epoch": 0.13210818307905686, "grad_norm": 3.2966949939727783, "learning_rate": 1e-06, "loss": 0.2423, "step": 381 }, { "epoch": 0.13245492371705964, "grad_norm": 2.6719970703125, "learning_rate": 1e-06, "loss": 0.1777, "step": 382 }, { "epoch": 0.1328016643550624, "grad_norm": 3.278883695602417, "learning_rate": 1e-06, "loss": 0.16, "step": 383 }, { "epoch": 0.13314840499306518, "grad_norm": 3.322058916091919, "learning_rate": 1e-06, "loss": 0.1984, "step": 384 }, { "epoch": 0.13349514563106796, "grad_norm": 2.3730714321136475, "learning_rate": 1e-06, "loss": 0.1836, "step": 385 }, { "epoch": 0.13384188626907073, "grad_norm": 1.6652333736419678, "learning_rate": 1e-06, "loss": 0.1896, "step": 386 }, { "epoch": 0.1341886269070735, "grad_norm": 3.528460741043091, "learning_rate": 1e-06, "loss": 0.2363, "step": 387 }, { "epoch": 0.13453536754507628, "grad_norm": 2.009749174118042, "learning_rate": 1e-06, "loss": 0.1814, "step": 388 }, { "epoch": 0.13488210818307905, "grad_norm": 2.015735149383545, "learning_rate": 1e-06, "loss": 0.2067, "step": 389 }, { "epoch": 0.13522884882108183, "grad_norm": 1.9921092987060547, "learning_rate": 1e-06, "loss": 0.1973, "step": 390 }, { "epoch": 0.1355755894590846, "grad_norm": 2.4047930240631104, "learning_rate": 1e-06, "loss": 0.1924, "step": 391 }, { "epoch": 0.13592233009708737, "grad_norm": 2.010021924972534, "learning_rate": 1e-06, "loss": 0.1744, "step": 392 }, { "epoch": 0.13626907073509015, "grad_norm": 2.5969536304473877, "learning_rate": 1e-06, "loss": 0.2092, "step": 393 }, { "epoch": 0.13661581137309292, "grad_norm": 3.3786892890930176, "learning_rate": 1e-06, "loss": 0.2044, "step": 394 }, { "epoch": 0.1369625520110957, "grad_norm": 1.805240511894226, "learning_rate": 1e-06, "loss": 0.2015, "step": 395 }, { "epoch": 0.13730929264909847, "grad_norm": 1.8784292936325073, "learning_rate": 1e-06, "loss": 0.2059, "step": 396 }, { "epoch": 0.13765603328710124, "grad_norm": 2.3622231483459473, "learning_rate": 1e-06, "loss": 0.1786, "step": 397 }, { "epoch": 0.13800277392510402, "grad_norm": 1.7961145639419556, "learning_rate": 1e-06, "loss": 0.2168, "step": 398 }, { "epoch": 0.1383495145631068, "grad_norm": 1.9970123767852783, "learning_rate": 1e-06, "loss": 0.1922, "step": 399 }, { "epoch": 0.13869625520110956, "grad_norm": 5.615191459655762, "learning_rate": 1e-06, "loss": 0.2319, "step": 400 }, { "epoch": 0.13904299583911234, "grad_norm": 2.6220815181732178, "learning_rate": 1e-06, "loss": 0.1899, "step": 401 }, { "epoch": 0.1393897364771151, "grad_norm": 3.2528562545776367, "learning_rate": 1e-06, "loss": 0.1744, "step": 402 }, { "epoch": 0.13973647711511789, "grad_norm": 2.6345772743225098, "learning_rate": 1e-06, "loss": 0.2226, "step": 403 }, { "epoch": 0.14008321775312066, "grad_norm": 5.005814075469971, "learning_rate": 1e-06, "loss": 0.2174, "step": 404 }, { "epoch": 0.14042995839112343, "grad_norm": 2.4105288982391357, "learning_rate": 1e-06, "loss": 0.2154, "step": 405 }, { "epoch": 0.1407766990291262, "grad_norm": 3.282689332962036, "learning_rate": 1e-06, "loss": 0.246, "step": 406 }, { "epoch": 0.14112343966712898, "grad_norm": 1.84548020362854, "learning_rate": 1e-06, "loss": 0.204, "step": 407 }, { "epoch": 0.14147018030513175, "grad_norm": 3.254951000213623, "learning_rate": 1e-06, "loss": 0.2355, "step": 408 }, { "epoch": 0.14181692094313453, "grad_norm": 2.499875068664551, "learning_rate": 1e-06, "loss": 0.2076, "step": 409 }, { "epoch": 0.1421636615811373, "grad_norm": 1.6304558515548706, "learning_rate": 1e-06, "loss": 0.1925, "step": 410 }, { "epoch": 0.14251040221914008, "grad_norm": 2.8476548194885254, "learning_rate": 1e-06, "loss": 0.2329, "step": 411 }, { "epoch": 0.14285714285714285, "grad_norm": 2.403111696243286, "learning_rate": 1e-06, "loss": 0.2016, "step": 412 }, { "epoch": 0.14320388349514562, "grad_norm": 2.050206422805786, "learning_rate": 1e-06, "loss": 0.227, "step": 413 }, { "epoch": 0.1435506241331484, "grad_norm": 3.8210604190826416, "learning_rate": 1e-06, "loss": 0.1825, "step": 414 }, { "epoch": 0.14389736477115117, "grad_norm": 2.2794923782348633, "learning_rate": 1e-06, "loss": 0.1822, "step": 415 }, { "epoch": 0.14424410540915394, "grad_norm": 2.152036666870117, "learning_rate": 1e-06, "loss": 0.2087, "step": 416 }, { "epoch": 0.14459084604715672, "grad_norm": 2.4530344009399414, "learning_rate": 1e-06, "loss": 0.2381, "step": 417 }, { "epoch": 0.1449375866851595, "grad_norm": 2.0212299823760986, "learning_rate": 1e-06, "loss": 0.2166, "step": 418 }, { "epoch": 0.14528432732316227, "grad_norm": 5.47108268737793, "learning_rate": 1e-06, "loss": 0.1952, "step": 419 }, { "epoch": 0.14563106796116504, "grad_norm": 2.0776853561401367, "learning_rate": 1e-06, "loss": 0.2232, "step": 420 }, { "epoch": 0.1459778085991678, "grad_norm": 5.302248477935791, "learning_rate": 1e-06, "loss": 0.1972, "step": 421 }, { "epoch": 0.1463245492371706, "grad_norm": 6.040057182312012, "learning_rate": 1e-06, "loss": 0.236, "step": 422 }, { "epoch": 0.14667128987517336, "grad_norm": 2.0244548320770264, "learning_rate": 1e-06, "loss": 0.189, "step": 423 }, { "epoch": 0.14701803051317613, "grad_norm": 1.9456743001937866, "learning_rate": 1e-06, "loss": 0.2195, "step": 424 }, { "epoch": 0.1473647711511789, "grad_norm": 2.1498379707336426, "learning_rate": 1e-06, "loss": 0.1891, "step": 425 }, { "epoch": 0.14771151178918168, "grad_norm": 2.6949310302734375, "learning_rate": 1e-06, "loss": 0.1933, "step": 426 }, { "epoch": 0.14805825242718446, "grad_norm": 2.2090682983398438, "learning_rate": 1e-06, "loss": 0.2001, "step": 427 }, { "epoch": 0.14840499306518723, "grad_norm": 2.047943592071533, "learning_rate": 1e-06, "loss": 0.2008, "step": 428 }, { "epoch": 0.14875173370319, "grad_norm": 3.49019718170166, "learning_rate": 1e-06, "loss": 0.25, "step": 429 }, { "epoch": 0.14909847434119278, "grad_norm": 2.07293963432312, "learning_rate": 1e-06, "loss": 0.176, "step": 430 }, { "epoch": 0.14944521497919555, "grad_norm": 1.9637081623077393, "learning_rate": 1e-06, "loss": 0.1918, "step": 431 }, { "epoch": 0.14979195561719832, "grad_norm": 4.7957658767700195, "learning_rate": 1e-06, "loss": 0.2399, "step": 432 }, { "epoch": 0.1501386962552011, "grad_norm": 2.5030736923217773, "learning_rate": 1e-06, "loss": 0.1782, "step": 433 }, { "epoch": 0.15048543689320387, "grad_norm": 1.744457483291626, "learning_rate": 1e-06, "loss": 0.1907, "step": 434 }, { "epoch": 0.15083217753120665, "grad_norm": 2.6001105308532715, "learning_rate": 1e-06, "loss": 0.1946, "step": 435 }, { "epoch": 0.15117891816920942, "grad_norm": 4.785390377044678, "learning_rate": 1e-06, "loss": 0.2152, "step": 436 }, { "epoch": 0.1515256588072122, "grad_norm": 1.5918031930923462, "learning_rate": 1e-06, "loss": 0.1817, "step": 437 }, { "epoch": 0.15187239944521497, "grad_norm": 2.5173332691192627, "learning_rate": 1e-06, "loss": 0.2512, "step": 438 }, { "epoch": 0.15221914008321774, "grad_norm": 2.2436468601226807, "learning_rate": 1e-06, "loss": 0.2083, "step": 439 }, { "epoch": 0.15256588072122051, "grad_norm": 5.61021089553833, "learning_rate": 1e-06, "loss": 0.2143, "step": 440 }, { "epoch": 0.1529126213592233, "grad_norm": 3.092174768447876, "learning_rate": 1e-06, "loss": 0.2081, "step": 441 }, { "epoch": 0.15325936199722606, "grad_norm": 1.9928513765335083, "learning_rate": 1e-06, "loss": 0.1816, "step": 442 }, { "epoch": 0.15360610263522884, "grad_norm": 2.1547915935516357, "learning_rate": 1e-06, "loss": 0.1919, "step": 443 }, { "epoch": 0.1539528432732316, "grad_norm": 4.531169414520264, "learning_rate": 1e-06, "loss": 0.2085, "step": 444 }, { "epoch": 0.15429958391123438, "grad_norm": 1.7542948722839355, "learning_rate": 1e-06, "loss": 0.1803, "step": 445 }, { "epoch": 0.15464632454923716, "grad_norm": 2.1614997386932373, "learning_rate": 1e-06, "loss": 0.1946, "step": 446 }, { "epoch": 0.15499306518723993, "grad_norm": 5.018011093139648, "learning_rate": 1e-06, "loss": 0.1974, "step": 447 }, { "epoch": 0.1553398058252427, "grad_norm": 2.907313346862793, "learning_rate": 1e-06, "loss": 0.1916, "step": 448 }, { "epoch": 0.15568654646324548, "grad_norm": 2.7045724391937256, "learning_rate": 1e-06, "loss": 0.1847, "step": 449 }, { "epoch": 0.15603328710124825, "grad_norm": 1.8427867889404297, "learning_rate": 1e-06, "loss": 0.1918, "step": 450 }, { "epoch": 0.15638002773925105, "grad_norm": 2.3798131942749023, "learning_rate": 1e-06, "loss": 0.2125, "step": 451 }, { "epoch": 0.15672676837725383, "grad_norm": 3.0421640872955322, "learning_rate": 1e-06, "loss": 0.1729, "step": 452 }, { "epoch": 0.1570735090152566, "grad_norm": 2.487687110900879, "learning_rate": 1e-06, "loss": 0.1791, "step": 453 }, { "epoch": 0.15742024965325938, "grad_norm": 3.9708778858184814, "learning_rate": 1e-06, "loss": 0.1902, "step": 454 }, { "epoch": 0.15776699029126215, "grad_norm": 2.342562437057495, "learning_rate": 1e-06, "loss": 0.1757, "step": 455 }, { "epoch": 0.15811373092926492, "grad_norm": 3.1214346885681152, "learning_rate": 1e-06, "loss": 0.1758, "step": 456 }, { "epoch": 0.1584604715672677, "grad_norm": 3.499249219894409, "learning_rate": 1e-06, "loss": 0.1947, "step": 457 }, { "epoch": 0.15880721220527047, "grad_norm": 1.9343475103378296, "learning_rate": 1e-06, "loss": 0.1835, "step": 458 }, { "epoch": 0.15915395284327324, "grad_norm": 3.174394130706787, "learning_rate": 1e-06, "loss": 0.2083, "step": 459 }, { "epoch": 0.15950069348127602, "grad_norm": 1.8774772882461548, "learning_rate": 1e-06, "loss": 0.1937, "step": 460 }, { "epoch": 0.1598474341192788, "grad_norm": 1.7514275312423706, "learning_rate": 1e-06, "loss": 0.1865, "step": 461 }, { "epoch": 0.16019417475728157, "grad_norm": 2.4523355960845947, "learning_rate": 1e-06, "loss": 0.1719, "step": 462 }, { "epoch": 0.16054091539528434, "grad_norm": 5.636368274688721, "learning_rate": 1e-06, "loss": 0.1751, "step": 463 }, { "epoch": 0.1608876560332871, "grad_norm": 2.248699426651001, "learning_rate": 1e-06, "loss": 0.2034, "step": 464 }, { "epoch": 0.1612343966712899, "grad_norm": 1.9587477445602417, "learning_rate": 1e-06, "loss": 0.1991, "step": 465 }, { "epoch": 0.16158113730929266, "grad_norm": 1.5505951642990112, "learning_rate": 1e-06, "loss": 0.1662, "step": 466 }, { "epoch": 0.16192787794729543, "grad_norm": 1.778409719467163, "learning_rate": 1e-06, "loss": 0.206, "step": 467 }, { "epoch": 0.1622746185852982, "grad_norm": 3.1874828338623047, "learning_rate": 1e-06, "loss": 0.1773, "step": 468 }, { "epoch": 0.16262135922330098, "grad_norm": 3.800231695175171, "learning_rate": 1e-06, "loss": 0.209, "step": 469 }, { "epoch": 0.16296809986130376, "grad_norm": 2.2624082565307617, "learning_rate": 1e-06, "loss": 0.1968, "step": 470 }, { "epoch": 0.16331484049930653, "grad_norm": 2.0975418090820312, "learning_rate": 1e-06, "loss": 0.235, "step": 471 }, { "epoch": 0.1636615811373093, "grad_norm": 5.558737277984619, "learning_rate": 1e-06, "loss": 0.1581, "step": 472 }, { "epoch": 0.16400832177531208, "grad_norm": 3.688441514968872, "learning_rate": 1e-06, "loss": 0.2038, "step": 473 }, { "epoch": 0.16435506241331485, "grad_norm": 1.7553553581237793, "learning_rate": 1e-06, "loss": 0.178, "step": 474 }, { "epoch": 0.16470180305131762, "grad_norm": 4.0227952003479, "learning_rate": 1e-06, "loss": 0.1732, "step": 475 }, { "epoch": 0.1650485436893204, "grad_norm": 1.9997745752334595, "learning_rate": 1e-06, "loss": 0.2068, "step": 476 }, { "epoch": 0.16539528432732317, "grad_norm": 3.5058529376983643, "learning_rate": 1e-06, "loss": 0.2066, "step": 477 }, { "epoch": 0.16574202496532595, "grad_norm": 3.745401382446289, "learning_rate": 1e-06, "loss": 0.2022, "step": 478 }, { "epoch": 0.16608876560332872, "grad_norm": 1.898235559463501, "learning_rate": 1e-06, "loss": 0.1787, "step": 479 }, { "epoch": 0.1664355062413315, "grad_norm": 2.2373104095458984, "learning_rate": 1e-06, "loss": 0.202, "step": 480 }, { "epoch": 0.16678224687933427, "grad_norm": 3.4441022872924805, "learning_rate": 1e-06, "loss": 0.2223, "step": 481 }, { "epoch": 0.16712898751733704, "grad_norm": 2.0082736015319824, "learning_rate": 1e-06, "loss": 0.1436, "step": 482 }, { "epoch": 0.16747572815533981, "grad_norm": 7.100375175476074, "learning_rate": 1e-06, "loss": 0.2622, "step": 483 }, { "epoch": 0.1678224687933426, "grad_norm": 3.85292649269104, "learning_rate": 1e-06, "loss": 0.2141, "step": 484 }, { "epoch": 0.16816920943134536, "grad_norm": 2.212933301925659, "learning_rate": 1e-06, "loss": 0.2052, "step": 485 }, { "epoch": 0.16851595006934814, "grad_norm": 2.0661942958831787, "learning_rate": 1e-06, "loss": 0.1676, "step": 486 }, { "epoch": 0.1688626907073509, "grad_norm": 4.262602806091309, "learning_rate": 1e-06, "loss": 0.2382, "step": 487 }, { "epoch": 0.16920943134535368, "grad_norm": 2.734023094177246, "learning_rate": 1e-06, "loss": 0.1833, "step": 488 }, { "epoch": 0.16955617198335646, "grad_norm": 1.9461660385131836, "learning_rate": 1e-06, "loss": 0.1833, "step": 489 }, { "epoch": 0.16990291262135923, "grad_norm": 2.6612391471862793, "learning_rate": 1e-06, "loss": 0.1996, "step": 490 }, { "epoch": 0.170249653259362, "grad_norm": 2.2470898628234863, "learning_rate": 1e-06, "loss": 0.1882, "step": 491 }, { "epoch": 0.17059639389736478, "grad_norm": 4.1081743240356445, "learning_rate": 1e-06, "loss": 0.1783, "step": 492 }, { "epoch": 0.17094313453536755, "grad_norm": 2.1462888717651367, "learning_rate": 1e-06, "loss": 0.2158, "step": 493 }, { "epoch": 0.17128987517337033, "grad_norm": 1.8906127214431763, "learning_rate": 1e-06, "loss": 0.196, "step": 494 }, { "epoch": 0.1716366158113731, "grad_norm": 3.4147090911865234, "learning_rate": 1e-06, "loss": 0.2561, "step": 495 }, { "epoch": 0.17198335644937587, "grad_norm": 3.48602032661438, "learning_rate": 1e-06, "loss": 0.17, "step": 496 }, { "epoch": 0.17233009708737865, "grad_norm": 1.873856544494629, "learning_rate": 1e-06, "loss": 0.1979, "step": 497 }, { "epoch": 0.17267683772538142, "grad_norm": 3.3756508827209473, "learning_rate": 1e-06, "loss": 0.1932, "step": 498 }, { "epoch": 0.1730235783633842, "grad_norm": 1.7137960195541382, "learning_rate": 1e-06, "loss": 0.1957, "step": 499 }, { "epoch": 0.17337031900138697, "grad_norm": 2.221604347229004, "learning_rate": 1e-06, "loss": 0.2075, "step": 500 }, { "epoch": 0.17371705963938974, "grad_norm": 1.7929978370666504, "learning_rate": 1e-06, "loss": 0.179, "step": 501 }, { "epoch": 0.17406380027739252, "grad_norm": 2.439054250717163, "learning_rate": 1e-06, "loss": 0.1962, "step": 502 }, { "epoch": 0.1744105409153953, "grad_norm": 2.1946470737457275, "learning_rate": 1e-06, "loss": 0.1734, "step": 503 }, { "epoch": 0.17475728155339806, "grad_norm": 2.1295015811920166, "learning_rate": 1e-06, "loss": 0.1943, "step": 504 }, { "epoch": 0.17510402219140084, "grad_norm": 3.222607135772705, "learning_rate": 1e-06, "loss": 0.192, "step": 505 }, { "epoch": 0.1754507628294036, "grad_norm": 3.0692672729492188, "learning_rate": 1e-06, "loss": 0.1785, "step": 506 }, { "epoch": 0.17579750346740639, "grad_norm": 3.299917459487915, "learning_rate": 1e-06, "loss": 0.1768, "step": 507 }, { "epoch": 0.17614424410540916, "grad_norm": 2.0762133598327637, "learning_rate": 1e-06, "loss": 0.1948, "step": 508 }, { "epoch": 0.17649098474341193, "grad_norm": 2.0623905658721924, "learning_rate": 1e-06, "loss": 0.2167, "step": 509 }, { "epoch": 0.1768377253814147, "grad_norm": 2.0946035385131836, "learning_rate": 1e-06, "loss": 0.2028, "step": 510 }, { "epoch": 0.17718446601941748, "grad_norm": 1.861733078956604, "learning_rate": 1e-06, "loss": 0.2162, "step": 511 }, { "epoch": 0.17753120665742025, "grad_norm": 2.4840567111968994, "learning_rate": 1e-06, "loss": 0.1863, "step": 512 }, { "epoch": 0.17787794729542303, "grad_norm": 1.8076410293579102, "learning_rate": 1e-06, "loss": 0.2008, "step": 513 }, { "epoch": 0.1782246879334258, "grad_norm": 2.526137351989746, "learning_rate": 1e-06, "loss": 0.1904, "step": 514 }, { "epoch": 0.17857142857142858, "grad_norm": 1.856165885925293, "learning_rate": 1e-06, "loss": 0.2123, "step": 515 }, { "epoch": 0.17891816920943135, "grad_norm": 3.9912686347961426, "learning_rate": 1e-06, "loss": 0.1335, "step": 516 }, { "epoch": 0.17926490984743412, "grad_norm": 2.328289270401001, "learning_rate": 1e-06, "loss": 0.2084, "step": 517 }, { "epoch": 0.1796116504854369, "grad_norm": 1.7976534366607666, "learning_rate": 1e-06, "loss": 0.1879, "step": 518 }, { "epoch": 0.17995839112343967, "grad_norm": 2.629750967025757, "learning_rate": 1e-06, "loss": 0.2154, "step": 519 }, { "epoch": 0.18030513176144244, "grad_norm": 2.200639009475708, "learning_rate": 1e-06, "loss": 0.1987, "step": 520 }, { "epoch": 0.18065187239944522, "grad_norm": 2.636253833770752, "learning_rate": 1e-06, "loss": 0.2092, "step": 521 }, { "epoch": 0.180998613037448, "grad_norm": 2.329237222671509, "learning_rate": 1e-06, "loss": 0.1865, "step": 522 }, { "epoch": 0.18134535367545077, "grad_norm": 3.2682480812072754, "learning_rate": 1e-06, "loss": 0.2204, "step": 523 }, { "epoch": 0.18169209431345354, "grad_norm": 2.117788791656494, "learning_rate": 1e-06, "loss": 0.1871, "step": 524 }, { "epoch": 0.1820388349514563, "grad_norm": 1.9039613008499146, "learning_rate": 1e-06, "loss": 0.1856, "step": 525 }, { "epoch": 0.1823855755894591, "grad_norm": 2.3658318519592285, "learning_rate": 1e-06, "loss": 0.1799, "step": 526 }, { "epoch": 0.18273231622746186, "grad_norm": 3.818437337875366, "learning_rate": 1e-06, "loss": 0.1944, "step": 527 }, { "epoch": 0.18307905686546463, "grad_norm": 2.8354179859161377, "learning_rate": 1e-06, "loss": 0.1592, "step": 528 }, { "epoch": 0.1834257975034674, "grad_norm": 1.872366189956665, "learning_rate": 1e-06, "loss": 0.1782, "step": 529 }, { "epoch": 0.18377253814147018, "grad_norm": 2.1778335571289062, "learning_rate": 1e-06, "loss": 0.1787, "step": 530 }, { "epoch": 0.18411927877947296, "grad_norm": 2.6627323627471924, "learning_rate": 1e-06, "loss": 0.1775, "step": 531 }, { "epoch": 0.18446601941747573, "grad_norm": 2.1031336784362793, "learning_rate": 1e-06, "loss": 0.2051, "step": 532 }, { "epoch": 0.1848127600554785, "grad_norm": 4.2585272789001465, "learning_rate": 1e-06, "loss": 0.2088, "step": 533 }, { "epoch": 0.18515950069348128, "grad_norm": 3.1983723640441895, "learning_rate": 1e-06, "loss": 0.177, "step": 534 }, { "epoch": 0.18550624133148405, "grad_norm": 2.121945381164551, "learning_rate": 1e-06, "loss": 0.212, "step": 535 }, { "epoch": 0.18585298196948682, "grad_norm": 4.327321529388428, "learning_rate": 1e-06, "loss": 0.2301, "step": 536 }, { "epoch": 0.1861997226074896, "grad_norm": 3.546032190322876, "learning_rate": 1e-06, "loss": 0.2165, "step": 537 }, { "epoch": 0.18654646324549237, "grad_norm": 2.3898680210113525, "learning_rate": 1e-06, "loss": 0.1892, "step": 538 }, { "epoch": 0.18689320388349515, "grad_norm": 2.9088211059570312, "learning_rate": 1e-06, "loss": 0.2018, "step": 539 }, { "epoch": 0.18723994452149792, "grad_norm": 2.1512014865875244, "learning_rate": 1e-06, "loss": 0.2037, "step": 540 }, { "epoch": 0.1875866851595007, "grad_norm": 2.608860731124878, "learning_rate": 1e-06, "loss": 0.1905, "step": 541 }, { "epoch": 0.18793342579750347, "grad_norm": 2.0679538249969482, "learning_rate": 1e-06, "loss": 0.1969, "step": 542 }, { "epoch": 0.18828016643550624, "grad_norm": 2.928675889968872, "learning_rate": 1e-06, "loss": 0.2332, "step": 543 }, { "epoch": 0.18862690707350901, "grad_norm": 6.560424327850342, "learning_rate": 1e-06, "loss": 0.2278, "step": 544 }, { "epoch": 0.1889736477115118, "grad_norm": 1.9088329076766968, "learning_rate": 1e-06, "loss": 0.2027, "step": 545 }, { "epoch": 0.18932038834951456, "grad_norm": 2.4593074321746826, "learning_rate": 1e-06, "loss": 0.1856, "step": 546 }, { "epoch": 0.18966712898751734, "grad_norm": 4.2560858726501465, "learning_rate": 1e-06, "loss": 0.1764, "step": 547 }, { "epoch": 0.1900138696255201, "grad_norm": 1.7965539693832397, "learning_rate": 1e-06, "loss": 0.1932, "step": 548 }, { "epoch": 0.19036061026352288, "grad_norm": 1.9651305675506592, "learning_rate": 1e-06, "loss": 0.2102, "step": 549 }, { "epoch": 0.19070735090152566, "grad_norm": 1.5734628438949585, "learning_rate": 1e-06, "loss": 0.1769, "step": 550 }, { "epoch": 0.19105409153952843, "grad_norm": 3.6712100505828857, "learning_rate": 1e-06, "loss": 0.2034, "step": 551 }, { "epoch": 0.1914008321775312, "grad_norm": 4.593487739562988, "learning_rate": 1e-06, "loss": 0.1643, "step": 552 }, { "epoch": 0.19174757281553398, "grad_norm": 2.358043909072876, "learning_rate": 1e-06, "loss": 0.1652, "step": 553 }, { "epoch": 0.19209431345353675, "grad_norm": 2.64416241645813, "learning_rate": 1e-06, "loss": 0.239, "step": 554 }, { "epoch": 0.19244105409153953, "grad_norm": 4.924618721008301, "learning_rate": 1e-06, "loss": 0.1842, "step": 555 }, { "epoch": 0.1927877947295423, "grad_norm": 2.018336534500122, "learning_rate": 1e-06, "loss": 0.2024, "step": 556 }, { "epoch": 0.19313453536754507, "grad_norm": 2.0920143127441406, "learning_rate": 1e-06, "loss": 0.2031, "step": 557 }, { "epoch": 0.19348127600554785, "grad_norm": 1.8479957580566406, "learning_rate": 1e-06, "loss": 0.1916, "step": 558 }, { "epoch": 0.19382801664355062, "grad_norm": 5.079397201538086, "learning_rate": 1e-06, "loss": 0.2203, "step": 559 }, { "epoch": 0.1941747572815534, "grad_norm": 4.152866363525391, "learning_rate": 1e-06, "loss": 0.1936, "step": 560 }, { "epoch": 0.19452149791955617, "grad_norm": 1.6684824228286743, "learning_rate": 1e-06, "loss": 0.163, "step": 561 }, { "epoch": 0.19486823855755894, "grad_norm": 2.2925472259521484, "learning_rate": 1e-06, "loss": 0.1826, "step": 562 }, { "epoch": 0.19521497919556172, "grad_norm": 1.73776113986969, "learning_rate": 1e-06, "loss": 0.175, "step": 563 }, { "epoch": 0.1955617198335645, "grad_norm": 6.04637336730957, "learning_rate": 1e-06, "loss": 0.2135, "step": 564 }, { "epoch": 0.19590846047156726, "grad_norm": 2.765383005142212, "learning_rate": 1e-06, "loss": 0.1885, "step": 565 }, { "epoch": 0.19625520110957004, "grad_norm": 2.205034017562866, "learning_rate": 1e-06, "loss": 0.1965, "step": 566 }, { "epoch": 0.1966019417475728, "grad_norm": 2.0300064086914062, "learning_rate": 1e-06, "loss": 0.206, "step": 567 }, { "epoch": 0.19694868238557559, "grad_norm": 1.9674218893051147, "learning_rate": 1e-06, "loss": 0.162, "step": 568 }, { "epoch": 0.19729542302357836, "grad_norm": 2.875797986984253, "learning_rate": 1e-06, "loss": 0.2251, "step": 569 }, { "epoch": 0.19764216366158113, "grad_norm": 1.855223536491394, "learning_rate": 1e-06, "loss": 0.2211, "step": 570 }, { "epoch": 0.1979889042995839, "grad_norm": 2.8994317054748535, "learning_rate": 1e-06, "loss": 0.206, "step": 571 }, { "epoch": 0.19833564493758668, "grad_norm": 2.2230451107025146, "learning_rate": 1e-06, "loss": 0.1947, "step": 572 }, { "epoch": 0.19868238557558945, "grad_norm": 5.828412055969238, "learning_rate": 1e-06, "loss": 0.21, "step": 573 }, { "epoch": 0.19902912621359223, "grad_norm": 2.926950454711914, "learning_rate": 1e-06, "loss": 0.2018, "step": 574 }, { "epoch": 0.199375866851595, "grad_norm": 3.7960805892944336, "learning_rate": 1e-06, "loss": 0.188, "step": 575 }, { "epoch": 0.19972260748959778, "grad_norm": 1.8869881629943848, "learning_rate": 1e-06, "loss": 0.2025, "step": 576 }, { "epoch": 0.20006934812760055, "grad_norm": 3.9236044883728027, "learning_rate": 1e-06, "loss": 0.1571, "step": 577 }, { "epoch": 0.20041608876560332, "grad_norm": 1.937950849533081, "learning_rate": 1e-06, "loss": 0.1944, "step": 578 }, { "epoch": 0.2007628294036061, "grad_norm": 2.139031410217285, "learning_rate": 1e-06, "loss": 0.1991, "step": 579 }, { "epoch": 0.20110957004160887, "grad_norm": 4.336677551269531, "learning_rate": 1e-06, "loss": 0.2159, "step": 580 }, { "epoch": 0.20145631067961164, "grad_norm": 3.819993495941162, "learning_rate": 1e-06, "loss": 0.1833, "step": 581 }, { "epoch": 0.20180305131761442, "grad_norm": 1.7961671352386475, "learning_rate": 1e-06, "loss": 0.1807, "step": 582 }, { "epoch": 0.2021497919556172, "grad_norm": 2.0944950580596924, "learning_rate": 1e-06, "loss": 0.2336, "step": 583 }, { "epoch": 0.20249653259361997, "grad_norm": 2.7267916202545166, "learning_rate": 1e-06, "loss": 0.2282, "step": 584 }, { "epoch": 0.20284327323162274, "grad_norm": 1.8287086486816406, "learning_rate": 1e-06, "loss": 0.1684, "step": 585 }, { "epoch": 0.2031900138696255, "grad_norm": 3.0344362258911133, "learning_rate": 1e-06, "loss": 0.1965, "step": 586 }, { "epoch": 0.2035367545076283, "grad_norm": 4.7607340812683105, "learning_rate": 1e-06, "loss": 0.193, "step": 587 }, { "epoch": 0.20388349514563106, "grad_norm": 1.6755648851394653, "learning_rate": 1e-06, "loss": 0.1764, "step": 588 }, { "epoch": 0.20423023578363383, "grad_norm": 2.2113454341888428, "learning_rate": 1e-06, "loss": 0.1614, "step": 589 }, { "epoch": 0.2045769764216366, "grad_norm": 2.667480945587158, "learning_rate": 1e-06, "loss": 0.173, "step": 590 }, { "epoch": 0.20492371705963938, "grad_norm": 2.3278677463531494, "learning_rate": 1e-06, "loss": 0.1876, "step": 591 }, { "epoch": 0.20527045769764216, "grad_norm": 3.0954275131225586, "learning_rate": 1e-06, "loss": 0.219, "step": 592 }, { "epoch": 0.20561719833564493, "grad_norm": 1.8866382837295532, "learning_rate": 1e-06, "loss": 0.1709, "step": 593 }, { "epoch": 0.2059639389736477, "grad_norm": 2.9739086627960205, "learning_rate": 1e-06, "loss": 0.1719, "step": 594 }, { "epoch": 0.20631067961165048, "grad_norm": 1.8454285860061646, "learning_rate": 1e-06, "loss": 0.1959, "step": 595 }, { "epoch": 0.20665742024965325, "grad_norm": 1.798344373703003, "learning_rate": 1e-06, "loss": 0.1655, "step": 596 }, { "epoch": 0.20700416088765602, "grad_norm": 3.577150821685791, "learning_rate": 1e-06, "loss": 0.2016, "step": 597 }, { "epoch": 0.2073509015256588, "grad_norm": 1.9027544260025024, "learning_rate": 1e-06, "loss": 0.1939, "step": 598 }, { "epoch": 0.20769764216366157, "grad_norm": 4.212296962738037, "learning_rate": 1e-06, "loss": 0.1663, "step": 599 }, { "epoch": 0.20804438280166435, "grad_norm": 2.9982285499572754, "learning_rate": 1e-06, "loss": 0.2316, "step": 600 }, { "epoch": 0.20839112343966712, "grad_norm": 2.4148337841033936, "learning_rate": 1e-06, "loss": 0.194, "step": 601 }, { "epoch": 0.2087378640776699, "grad_norm": 5.278850078582764, "learning_rate": 1e-06, "loss": 0.1835, "step": 602 }, { "epoch": 0.20908460471567267, "grad_norm": 2.83439564704895, "learning_rate": 1e-06, "loss": 0.1839, "step": 603 }, { "epoch": 0.20943134535367544, "grad_norm": 2.2179036140441895, "learning_rate": 1e-06, "loss": 0.2021, "step": 604 }, { "epoch": 0.20977808599167821, "grad_norm": 5.919695854187012, "learning_rate": 1e-06, "loss": 0.2135, "step": 605 }, { "epoch": 0.210124826629681, "grad_norm": 1.5985592603683472, "learning_rate": 1e-06, "loss": 0.1672, "step": 606 }, { "epoch": 0.21047156726768376, "grad_norm": 3.6065750122070312, "learning_rate": 1e-06, "loss": 0.2133, "step": 607 }, { "epoch": 0.21081830790568654, "grad_norm": 3.56968355178833, "learning_rate": 1e-06, "loss": 0.1867, "step": 608 }, { "epoch": 0.2111650485436893, "grad_norm": 2.662672758102417, "learning_rate": 1e-06, "loss": 0.1773, "step": 609 }, { "epoch": 0.21151178918169208, "grad_norm": 1.9275373220443726, "learning_rate": 1e-06, "loss": 0.1926, "step": 610 }, { "epoch": 0.21185852981969486, "grad_norm": 4.133257865905762, "learning_rate": 1e-06, "loss": 0.1931, "step": 611 }, { "epoch": 0.21220527045769763, "grad_norm": 1.791725516319275, "learning_rate": 1e-06, "loss": 0.1874, "step": 612 }, { "epoch": 0.2125520110957004, "grad_norm": 2.296025514602661, "learning_rate": 1e-06, "loss": 0.1968, "step": 613 }, { "epoch": 0.21289875173370318, "grad_norm": 1.7383009195327759, "learning_rate": 1e-06, "loss": 0.1597, "step": 614 }, { "epoch": 0.21324549237170595, "grad_norm": 2.899918556213379, "learning_rate": 1e-06, "loss": 0.1843, "step": 615 }, { "epoch": 0.21359223300970873, "grad_norm": 7.615829944610596, "learning_rate": 1e-06, "loss": 0.2117, "step": 616 }, { "epoch": 0.2139389736477115, "grad_norm": 1.9387335777282715, "learning_rate": 1e-06, "loss": 0.1798, "step": 617 }, { "epoch": 0.21428571428571427, "grad_norm": 4.627980709075928, "learning_rate": 1e-06, "loss": 0.1829, "step": 618 }, { "epoch": 0.21463245492371705, "grad_norm": 7.2196526527404785, "learning_rate": 1e-06, "loss": 0.2323, "step": 619 }, { "epoch": 0.21497919556171982, "grad_norm": 1.4321414232254028, "learning_rate": 1e-06, "loss": 0.157, "step": 620 }, { "epoch": 0.2153259361997226, "grad_norm": 1.6646205186843872, "learning_rate": 1e-06, "loss": 0.1922, "step": 621 }, { "epoch": 0.21567267683772537, "grad_norm": 1.867525577545166, "learning_rate": 1e-06, "loss": 0.1995, "step": 622 }, { "epoch": 0.21601941747572814, "grad_norm": 2.072211503982544, "learning_rate": 1e-06, "loss": 0.1848, "step": 623 }, { "epoch": 0.21636615811373092, "grad_norm": 3.1472299098968506, "learning_rate": 1e-06, "loss": 0.17, "step": 624 }, { "epoch": 0.2167128987517337, "grad_norm": 5.170080184936523, "learning_rate": 1e-06, "loss": 0.1975, "step": 625 }, { "epoch": 0.21705963938973646, "grad_norm": 1.9573615789413452, "learning_rate": 1e-06, "loss": 0.2022, "step": 626 }, { "epoch": 0.21740638002773924, "grad_norm": 4.052754878997803, "learning_rate": 1e-06, "loss": 0.1595, "step": 627 }, { "epoch": 0.217753120665742, "grad_norm": 2.469994306564331, "learning_rate": 1e-06, "loss": 0.2114, "step": 628 }, { "epoch": 0.21809986130374479, "grad_norm": 1.9860141277313232, "learning_rate": 1e-06, "loss": 0.1842, "step": 629 }, { "epoch": 0.21844660194174756, "grad_norm": 3.6322503089904785, "learning_rate": 1e-06, "loss": 0.183, "step": 630 }, { "epoch": 0.21879334257975036, "grad_norm": 3.5005853176116943, "learning_rate": 1e-06, "loss": 0.1921, "step": 631 }, { "epoch": 0.21914008321775313, "grad_norm": 1.7438961267471313, "learning_rate": 1e-06, "loss": 0.1883, "step": 632 }, { "epoch": 0.2194868238557559, "grad_norm": 2.330274820327759, "learning_rate": 1e-06, "loss": 0.1777, "step": 633 }, { "epoch": 0.21983356449375868, "grad_norm": 2.0113935470581055, "learning_rate": 1e-06, "loss": 0.198, "step": 634 }, { "epoch": 0.22018030513176146, "grad_norm": 2.120816946029663, "learning_rate": 1e-06, "loss": 0.195, "step": 635 }, { "epoch": 0.22052704576976423, "grad_norm": 3.88839054107666, "learning_rate": 1e-06, "loss": 0.2156, "step": 636 }, { "epoch": 0.220873786407767, "grad_norm": 2.330038070678711, "learning_rate": 1e-06, "loss": 0.1525, "step": 637 }, { "epoch": 0.22122052704576978, "grad_norm": 2.0380382537841797, "learning_rate": 1e-06, "loss": 0.1892, "step": 638 }, { "epoch": 0.22156726768377255, "grad_norm": 2.988846778869629, "learning_rate": 1e-06, "loss": 0.1784, "step": 639 }, { "epoch": 0.22191400832177532, "grad_norm": 3.193902015686035, "learning_rate": 1e-06, "loss": 0.2036, "step": 640 }, { "epoch": 0.2222607489597781, "grad_norm": 3.138087511062622, "learning_rate": 1e-06, "loss": 0.1995, "step": 641 }, { "epoch": 0.22260748959778087, "grad_norm": 3.0959274768829346, "learning_rate": 1e-06, "loss": 0.212, "step": 642 }, { "epoch": 0.22295423023578365, "grad_norm": 3.1513280868530273, "learning_rate": 1e-06, "loss": 0.1928, "step": 643 }, { "epoch": 0.22330097087378642, "grad_norm": 3.7054238319396973, "learning_rate": 1e-06, "loss": 0.1547, "step": 644 }, { "epoch": 0.2236477115117892, "grad_norm": 1.7885769605636597, "learning_rate": 1e-06, "loss": 0.1764, "step": 645 }, { "epoch": 0.22399445214979197, "grad_norm": 4.171994686126709, "learning_rate": 1e-06, "loss": 0.1813, "step": 646 }, { "epoch": 0.22434119278779474, "grad_norm": 2.313520669937134, "learning_rate": 1e-06, "loss": 0.2098, "step": 647 }, { "epoch": 0.22468793342579751, "grad_norm": 3.154613971710205, "learning_rate": 1e-06, "loss": 0.1776, "step": 648 }, { "epoch": 0.2250346740638003, "grad_norm": 5.905630111694336, "learning_rate": 1e-06, "loss": 0.1632, "step": 649 }, { "epoch": 0.22538141470180306, "grad_norm": 3.555774211883545, "learning_rate": 1e-06, "loss": 0.1747, "step": 650 }, { "epoch": 0.22572815533980584, "grad_norm": 2.0930142402648926, "learning_rate": 1e-06, "loss": 0.1962, "step": 651 }, { "epoch": 0.2260748959778086, "grad_norm": 3.322031021118164, "learning_rate": 1e-06, "loss": 0.1887, "step": 652 }, { "epoch": 0.22642163661581138, "grad_norm": 3.3001041412353516, "learning_rate": 1e-06, "loss": 0.1579, "step": 653 }, { "epoch": 0.22676837725381416, "grad_norm": 2.0050160884857178, "learning_rate": 1e-06, "loss": 0.1696, "step": 654 }, { "epoch": 0.22711511789181693, "grad_norm": 2.4040985107421875, "learning_rate": 1e-06, "loss": 0.1937, "step": 655 }, { "epoch": 0.2274618585298197, "grad_norm": 2.627812147140503, "learning_rate": 1e-06, "loss": 0.1667, "step": 656 }, { "epoch": 0.22780859916782248, "grad_norm": 4.34411096572876, "learning_rate": 1e-06, "loss": 0.1815, "step": 657 }, { "epoch": 0.22815533980582525, "grad_norm": 4.466109275817871, "learning_rate": 1e-06, "loss": 0.1798, "step": 658 }, { "epoch": 0.22850208044382803, "grad_norm": 1.537264347076416, "learning_rate": 1e-06, "loss": 0.1758, "step": 659 }, { "epoch": 0.2288488210818308, "grad_norm": 3.8718326091766357, "learning_rate": 1e-06, "loss": 0.1796, "step": 660 }, { "epoch": 0.22919556171983357, "grad_norm": 3.2782208919525146, "learning_rate": 1e-06, "loss": 0.158, "step": 661 }, { "epoch": 0.22954230235783635, "grad_norm": 1.7151904106140137, "learning_rate": 1e-06, "loss": 0.1769, "step": 662 }, { "epoch": 0.22988904299583912, "grad_norm": 1.8331800699234009, "learning_rate": 1e-06, "loss": 0.1671, "step": 663 }, { "epoch": 0.2302357836338419, "grad_norm": 2.3678629398345947, "learning_rate": 1e-06, "loss": 0.164, "step": 664 }, { "epoch": 0.23058252427184467, "grad_norm": 1.5758439302444458, "learning_rate": 1e-06, "loss": 0.1689, "step": 665 }, { "epoch": 0.23092926490984744, "grad_norm": 2.456174850463867, "learning_rate": 1e-06, "loss": 0.1681, "step": 666 }, { "epoch": 0.23127600554785022, "grad_norm": 1.6386404037475586, "learning_rate": 1e-06, "loss": 0.1722, "step": 667 }, { "epoch": 0.231622746185853, "grad_norm": 1.6972070932388306, "learning_rate": 1e-06, "loss": 0.1699, "step": 668 }, { "epoch": 0.23196948682385576, "grad_norm": 2.991753101348877, "learning_rate": 1e-06, "loss": 0.194, "step": 669 }, { "epoch": 0.23231622746185854, "grad_norm": 1.5393069982528687, "learning_rate": 1e-06, "loss": 0.1717, "step": 670 }, { "epoch": 0.2326629680998613, "grad_norm": 2.0264406204223633, "learning_rate": 1e-06, "loss": 0.1961, "step": 671 }, { "epoch": 0.23300970873786409, "grad_norm": 4.824713706970215, "learning_rate": 1e-06, "loss": 0.1516, "step": 672 }, { "epoch": 0.23335644937586686, "grad_norm": 1.823652982711792, "learning_rate": 1e-06, "loss": 0.1913, "step": 673 }, { "epoch": 0.23370319001386963, "grad_norm": 2.0635838508605957, "learning_rate": 1e-06, "loss": 0.1702, "step": 674 }, { "epoch": 0.2340499306518724, "grad_norm": 1.7453879117965698, "learning_rate": 1e-06, "loss": 0.171, "step": 675 }, { "epoch": 0.23439667128987518, "grad_norm": 3.4135522842407227, "learning_rate": 1e-06, "loss": 0.2, "step": 676 }, { "epoch": 0.23474341192787795, "grad_norm": 4.439614295959473, "learning_rate": 1e-06, "loss": 0.1779, "step": 677 }, { "epoch": 0.23509015256588073, "grad_norm": 1.9684357643127441, "learning_rate": 1e-06, "loss": 0.1828, "step": 678 }, { "epoch": 0.2354368932038835, "grad_norm": 1.9556607007980347, "learning_rate": 1e-06, "loss": 0.1967, "step": 679 }, { "epoch": 0.23578363384188628, "grad_norm": 3.652463436126709, "learning_rate": 1e-06, "loss": 0.2005, "step": 680 }, { "epoch": 0.23613037447988905, "grad_norm": 1.9398771524429321, "learning_rate": 1e-06, "loss": 0.158, "step": 681 }, { "epoch": 0.23647711511789182, "grad_norm": 3.950073480606079, "learning_rate": 1e-06, "loss": 0.192, "step": 682 }, { "epoch": 0.2368238557558946, "grad_norm": 3.032841205596924, "learning_rate": 1e-06, "loss": 0.2069, "step": 683 }, { "epoch": 0.23717059639389737, "grad_norm": 1.8892643451690674, "learning_rate": 1e-06, "loss": 0.1994, "step": 684 }, { "epoch": 0.23751733703190014, "grad_norm": 2.4548637866973877, "learning_rate": 1e-06, "loss": 0.1818, "step": 685 }, { "epoch": 0.23786407766990292, "grad_norm": 1.8674260377883911, "learning_rate": 1e-06, "loss": 0.1712, "step": 686 }, { "epoch": 0.2382108183079057, "grad_norm": 5.607412338256836, "learning_rate": 1e-06, "loss": 0.205, "step": 687 }, { "epoch": 0.23855755894590847, "grad_norm": 1.955690860748291, "learning_rate": 1e-06, "loss": 0.192, "step": 688 }, { "epoch": 0.23890429958391124, "grad_norm": 3.3477094173431396, "learning_rate": 1e-06, "loss": 0.1538, "step": 689 }, { "epoch": 0.239251040221914, "grad_norm": 4.284628868103027, "learning_rate": 1e-06, "loss": 0.1592, "step": 690 }, { "epoch": 0.2395977808599168, "grad_norm": 2.2527334690093994, "learning_rate": 1e-06, "loss": 0.1675, "step": 691 }, { "epoch": 0.23994452149791956, "grad_norm": 2.17332124710083, "learning_rate": 1e-06, "loss": 0.1988, "step": 692 }, { "epoch": 0.24029126213592233, "grad_norm": 3.59987211227417, "learning_rate": 1e-06, "loss": 0.2038, "step": 693 }, { "epoch": 0.2406380027739251, "grad_norm": 4.105143070220947, "learning_rate": 1e-06, "loss": 0.2028, "step": 694 }, { "epoch": 0.24098474341192788, "grad_norm": 1.9248536825180054, "learning_rate": 1e-06, "loss": 0.2031, "step": 695 }, { "epoch": 0.24133148404993066, "grad_norm": 2.1153578758239746, "learning_rate": 1e-06, "loss": 0.2164, "step": 696 }, { "epoch": 0.24167822468793343, "grad_norm": 3.2150540351867676, "learning_rate": 1e-06, "loss": 0.1985, "step": 697 }, { "epoch": 0.2420249653259362, "grad_norm": 2.063730478286743, "learning_rate": 1e-06, "loss": 0.1914, "step": 698 }, { "epoch": 0.24237170596393898, "grad_norm": 1.997815728187561, "learning_rate": 1e-06, "loss": 0.1956, "step": 699 }, { "epoch": 0.24271844660194175, "grad_norm": 2.1964597702026367, "learning_rate": 1e-06, "loss": 0.1665, "step": 700 }, { "epoch": 0.24306518723994452, "grad_norm": 1.6123956441879272, "learning_rate": 1e-06, "loss": 0.1647, "step": 701 }, { "epoch": 0.2434119278779473, "grad_norm": 4.534806251525879, "learning_rate": 1e-06, "loss": 0.1932, "step": 702 }, { "epoch": 0.24375866851595007, "grad_norm": 3.9443535804748535, "learning_rate": 1e-06, "loss": 0.1928, "step": 703 }, { "epoch": 0.24410540915395285, "grad_norm": 6.974303245544434, "learning_rate": 1e-06, "loss": 0.1648, "step": 704 }, { "epoch": 0.24445214979195562, "grad_norm": 2.646906614303589, "learning_rate": 1e-06, "loss": 0.2277, "step": 705 }, { "epoch": 0.2447988904299584, "grad_norm": 2.9131650924682617, "learning_rate": 1e-06, "loss": 0.1663, "step": 706 }, { "epoch": 0.24514563106796117, "grad_norm": 1.7801005840301514, "learning_rate": 1e-06, "loss": 0.1656, "step": 707 }, { "epoch": 0.24549237170596394, "grad_norm": 2.639249563217163, "learning_rate": 1e-06, "loss": 0.1751, "step": 708 }, { "epoch": 0.24583911234396671, "grad_norm": 1.4407970905303955, "learning_rate": 1e-06, "loss": 0.1702, "step": 709 }, { "epoch": 0.2461858529819695, "grad_norm": 1.4568276405334473, "learning_rate": 1e-06, "loss": 0.1659, "step": 710 }, { "epoch": 0.24653259361997226, "grad_norm": 4.625824928283691, "learning_rate": 1e-06, "loss": 0.2295, "step": 711 }, { "epoch": 0.24687933425797504, "grad_norm": 4.630171775817871, "learning_rate": 1e-06, "loss": 0.1668, "step": 712 }, { "epoch": 0.2472260748959778, "grad_norm": 1.7043591737747192, "learning_rate": 1e-06, "loss": 0.179, "step": 713 }, { "epoch": 0.24757281553398058, "grad_norm": 2.682840585708618, "learning_rate": 1e-06, "loss": 0.1708, "step": 714 }, { "epoch": 0.24791955617198336, "grad_norm": 1.8791319131851196, "learning_rate": 1e-06, "loss": 0.195, "step": 715 }, { "epoch": 0.24826629680998613, "grad_norm": 1.7238434553146362, "learning_rate": 1e-06, "loss": 0.1711, "step": 716 }, { "epoch": 0.2486130374479889, "grad_norm": 3.407539129257202, "learning_rate": 1e-06, "loss": 0.1666, "step": 717 }, { "epoch": 0.24895977808599168, "grad_norm": 5.695379257202148, "learning_rate": 1e-06, "loss": 0.1979, "step": 718 }, { "epoch": 0.24930651872399445, "grad_norm": 4.801501274108887, "learning_rate": 1e-06, "loss": 0.1888, "step": 719 }, { "epoch": 0.24965325936199723, "grad_norm": 4.1193389892578125, "learning_rate": 1e-06, "loss": 0.202, "step": 720 }, { "epoch": 0.25, "grad_norm": 2.602652072906494, "learning_rate": 1e-06, "loss": 0.1926, "step": 721 }, { "epoch": 0.2503467406380028, "grad_norm": 2.5733113288879395, "learning_rate": 1e-06, "loss": 0.1964, "step": 722 }, { "epoch": 0.25069348127600555, "grad_norm": 1.5140520334243774, "learning_rate": 1e-06, "loss": 0.1719, "step": 723 }, { "epoch": 0.2510402219140083, "grad_norm": 1.8638113737106323, "learning_rate": 1e-06, "loss": 0.1745, "step": 724 }, { "epoch": 0.2513869625520111, "grad_norm": 1.9770817756652832, "learning_rate": 1e-06, "loss": 0.2094, "step": 725 }, { "epoch": 0.25173370319001387, "grad_norm": 3.411306619644165, "learning_rate": 1e-06, "loss": 0.1674, "step": 726 }, { "epoch": 0.25208044382801664, "grad_norm": 5.473649024963379, "learning_rate": 1e-06, "loss": 0.1843, "step": 727 }, { "epoch": 0.2524271844660194, "grad_norm": 3.618953227996826, "learning_rate": 1e-06, "loss": 0.1691, "step": 728 }, { "epoch": 0.2527739251040222, "grad_norm": 2.7553508281707764, "learning_rate": 1e-06, "loss": 0.1745, "step": 729 }, { "epoch": 0.25312066574202496, "grad_norm": 3.168771266937256, "learning_rate": 1e-06, "loss": 0.1952, "step": 730 }, { "epoch": 0.25346740638002774, "grad_norm": 2.5667014122009277, "learning_rate": 1e-06, "loss": 0.1685, "step": 731 }, { "epoch": 0.2538141470180305, "grad_norm": 2.024369716644287, "learning_rate": 1e-06, "loss": 0.1782, "step": 732 }, { "epoch": 0.2541608876560333, "grad_norm": 3.3991565704345703, "learning_rate": 1e-06, "loss": 0.1628, "step": 733 }, { "epoch": 0.25450762829403606, "grad_norm": 3.4920365810394287, "learning_rate": 1e-06, "loss": 0.1869, "step": 734 }, { "epoch": 0.25485436893203883, "grad_norm": 2.4127650260925293, "learning_rate": 1e-06, "loss": 0.2054, "step": 735 }, { "epoch": 0.2552011095700416, "grad_norm": 2.5283799171447754, "learning_rate": 1e-06, "loss": 0.2128, "step": 736 }, { "epoch": 0.2555478502080444, "grad_norm": 3.1239235401153564, "learning_rate": 1e-06, "loss": 0.189, "step": 737 }, { "epoch": 0.25589459084604715, "grad_norm": 4.53823184967041, "learning_rate": 1e-06, "loss": 0.2216, "step": 738 }, { "epoch": 0.25624133148404993, "grad_norm": 1.9192531108856201, "learning_rate": 1e-06, "loss": 0.1937, "step": 739 }, { "epoch": 0.2565880721220527, "grad_norm": 3.241487979888916, "learning_rate": 1e-06, "loss": 0.1744, "step": 740 }, { "epoch": 0.2569348127600555, "grad_norm": 4.408164024353027, "learning_rate": 1e-06, "loss": 0.1792, "step": 741 }, { "epoch": 0.25728155339805825, "grad_norm": 2.0512523651123047, "learning_rate": 1e-06, "loss": 0.158, "step": 742 }, { "epoch": 0.257628294036061, "grad_norm": 1.6755858659744263, "learning_rate": 1e-06, "loss": 0.1819, "step": 743 }, { "epoch": 0.2579750346740638, "grad_norm": 1.6671578884124756, "learning_rate": 1e-06, "loss": 0.1634, "step": 744 }, { "epoch": 0.25832177531206657, "grad_norm": 2.0156643390655518, "learning_rate": 1e-06, "loss": 0.169, "step": 745 }, { "epoch": 0.25866851595006934, "grad_norm": 3.1058249473571777, "learning_rate": 1e-06, "loss": 0.1633, "step": 746 }, { "epoch": 0.2590152565880721, "grad_norm": 2.4028232097625732, "learning_rate": 1e-06, "loss": 0.1907, "step": 747 }, { "epoch": 0.2593619972260749, "grad_norm": 5.4095354080200195, "learning_rate": 1e-06, "loss": 0.1564, "step": 748 }, { "epoch": 0.25970873786407767, "grad_norm": 2.967435359954834, "learning_rate": 1e-06, "loss": 0.1811, "step": 749 }, { "epoch": 0.26005547850208044, "grad_norm": 2.7029571533203125, "learning_rate": 1e-06, "loss": 0.2192, "step": 750 }, { "epoch": 0.2604022191400832, "grad_norm": 2.9570839405059814, "learning_rate": 1e-06, "loss": 0.1605, "step": 751 }, { "epoch": 0.260748959778086, "grad_norm": 2.031963348388672, "learning_rate": 1e-06, "loss": 0.2144, "step": 752 }, { "epoch": 0.26109570041608876, "grad_norm": 1.6963039636611938, "learning_rate": 1e-06, "loss": 0.1506, "step": 753 }, { "epoch": 0.26144244105409153, "grad_norm": 1.8354779481887817, "learning_rate": 1e-06, "loss": 0.1736, "step": 754 }, { "epoch": 0.2617891816920943, "grad_norm": 2.998814582824707, "learning_rate": 1e-06, "loss": 0.1888, "step": 755 }, { "epoch": 0.2621359223300971, "grad_norm": 1.5606882572174072, "learning_rate": 1e-06, "loss": 0.1434, "step": 756 }, { "epoch": 0.26248266296809986, "grad_norm": 6.14403772354126, "learning_rate": 1e-06, "loss": 0.1722, "step": 757 }, { "epoch": 0.26282940360610263, "grad_norm": 4.8302321434021, "learning_rate": 1e-06, "loss": 0.1674, "step": 758 }, { "epoch": 0.2631761442441054, "grad_norm": 6.868436813354492, "learning_rate": 1e-06, "loss": 0.2065, "step": 759 }, { "epoch": 0.2635228848821082, "grad_norm": 5.897651672363281, "learning_rate": 1e-06, "loss": 0.1934, "step": 760 }, { "epoch": 0.26386962552011095, "grad_norm": 3.1365790367126465, "learning_rate": 1e-06, "loss": 0.2086, "step": 761 }, { "epoch": 0.2642163661581137, "grad_norm": 4.416586875915527, "learning_rate": 1e-06, "loss": 0.1834, "step": 762 }, { "epoch": 0.2645631067961165, "grad_norm": 3.5971829891204834, "learning_rate": 1e-06, "loss": 0.1734, "step": 763 }, { "epoch": 0.26490984743411927, "grad_norm": 2.017256259918213, "learning_rate": 1e-06, "loss": 0.2148, "step": 764 }, { "epoch": 0.26525658807212205, "grad_norm": 1.7896546125411987, "learning_rate": 1e-06, "loss": 0.1934, "step": 765 }, { "epoch": 0.2656033287101248, "grad_norm": 2.142129898071289, "learning_rate": 1e-06, "loss": 0.1691, "step": 766 }, { "epoch": 0.2659500693481276, "grad_norm": 2.0466034412384033, "learning_rate": 1e-06, "loss": 0.1889, "step": 767 }, { "epoch": 0.26629680998613037, "grad_norm": 5.9090189933776855, "learning_rate": 1e-06, "loss": 0.181, "step": 768 }, { "epoch": 0.26664355062413314, "grad_norm": 3.580826997756958, "learning_rate": 1e-06, "loss": 0.1827, "step": 769 }, { "epoch": 0.2669902912621359, "grad_norm": 4.003768444061279, "learning_rate": 1e-06, "loss": 0.1917, "step": 770 }, { "epoch": 0.2673370319001387, "grad_norm": 1.884700059890747, "learning_rate": 1e-06, "loss": 0.182, "step": 771 }, { "epoch": 0.26768377253814146, "grad_norm": 2.163311243057251, "learning_rate": 1e-06, "loss": 0.2078, "step": 772 }, { "epoch": 0.26803051317614424, "grad_norm": 4.079468250274658, "learning_rate": 1e-06, "loss": 0.1485, "step": 773 }, { "epoch": 0.268377253814147, "grad_norm": 3.486384391784668, "learning_rate": 1e-06, "loss": 0.1655, "step": 774 }, { "epoch": 0.2687239944521498, "grad_norm": 3.0869040489196777, "learning_rate": 1e-06, "loss": 0.1543, "step": 775 }, { "epoch": 0.26907073509015256, "grad_norm": 1.9200345277786255, "learning_rate": 1e-06, "loss": 0.1797, "step": 776 }, { "epoch": 0.26941747572815533, "grad_norm": 2.677326202392578, "learning_rate": 1e-06, "loss": 0.1624, "step": 777 }, { "epoch": 0.2697642163661581, "grad_norm": 4.965575695037842, "learning_rate": 1e-06, "loss": 0.1843, "step": 778 }, { "epoch": 0.2701109570041609, "grad_norm": 3.100609064102173, "learning_rate": 1e-06, "loss": 0.2043, "step": 779 }, { "epoch": 0.27045769764216365, "grad_norm": 3.7525222301483154, "learning_rate": 1e-06, "loss": 0.1853, "step": 780 }, { "epoch": 0.2708044382801664, "grad_norm": 3.449674606323242, "learning_rate": 1e-06, "loss": 0.1619, "step": 781 }, { "epoch": 0.2711511789181692, "grad_norm": 2.731919050216675, "learning_rate": 1e-06, "loss": 0.2145, "step": 782 }, { "epoch": 0.271497919556172, "grad_norm": 5.897451877593994, "learning_rate": 1e-06, "loss": 0.196, "step": 783 }, { "epoch": 0.27184466019417475, "grad_norm": 3.070488452911377, "learning_rate": 1e-06, "loss": 0.1985, "step": 784 }, { "epoch": 0.2721914008321775, "grad_norm": 2.275068521499634, "learning_rate": 1e-06, "loss": 0.1865, "step": 785 }, { "epoch": 0.2725381414701803, "grad_norm": 2.4451684951782227, "learning_rate": 1e-06, "loss": 0.1983, "step": 786 }, { "epoch": 0.27288488210818307, "grad_norm": 2.198415517807007, "learning_rate": 1e-06, "loss": 0.1975, "step": 787 }, { "epoch": 0.27323162274618584, "grad_norm": 4.776117324829102, "learning_rate": 1e-06, "loss": 0.1747, "step": 788 }, { "epoch": 0.2735783633841886, "grad_norm": 4.163407325744629, "learning_rate": 1e-06, "loss": 0.1743, "step": 789 }, { "epoch": 0.2739251040221914, "grad_norm": 1.9131619930267334, "learning_rate": 1e-06, "loss": 0.1942, "step": 790 }, { "epoch": 0.27427184466019416, "grad_norm": 2.953847885131836, "learning_rate": 1e-06, "loss": 0.1653, "step": 791 }, { "epoch": 0.27461858529819694, "grad_norm": 1.98551344871521, "learning_rate": 1e-06, "loss": 0.1992, "step": 792 }, { "epoch": 0.2749653259361997, "grad_norm": 5.063632965087891, "learning_rate": 1e-06, "loss": 0.1385, "step": 793 }, { "epoch": 0.2753120665742025, "grad_norm": 3.6445019245147705, "learning_rate": 1e-06, "loss": 0.1653, "step": 794 }, { "epoch": 0.27565880721220526, "grad_norm": 1.9771358966827393, "learning_rate": 1e-06, "loss": 0.2067, "step": 795 }, { "epoch": 0.27600554785020803, "grad_norm": 1.9333562850952148, "learning_rate": 1e-06, "loss": 0.1988, "step": 796 }, { "epoch": 0.2763522884882108, "grad_norm": 1.9832173585891724, "learning_rate": 1e-06, "loss": 0.1646, "step": 797 }, { "epoch": 0.2766990291262136, "grad_norm": 2.332988977432251, "learning_rate": 1e-06, "loss": 0.164, "step": 798 }, { "epoch": 0.27704576976421635, "grad_norm": 2.837818145751953, "learning_rate": 1e-06, "loss": 0.1404, "step": 799 }, { "epoch": 0.27739251040221913, "grad_norm": 3.3492045402526855, "learning_rate": 1e-06, "loss": 0.1635, "step": 800 }, { "epoch": 0.2777392510402219, "grad_norm": 1.856004238128662, "learning_rate": 1e-06, "loss": 0.152, "step": 801 }, { "epoch": 0.2780859916782247, "grad_norm": 5.804831504821777, "learning_rate": 1e-06, "loss": 0.1833, "step": 802 }, { "epoch": 0.27843273231622745, "grad_norm": 1.7844274044036865, "learning_rate": 1e-06, "loss": 0.1768, "step": 803 }, { "epoch": 0.2787794729542302, "grad_norm": 2.5231499671936035, "learning_rate": 1e-06, "loss": 0.193, "step": 804 }, { "epoch": 0.279126213592233, "grad_norm": 3.219759464263916, "learning_rate": 1e-06, "loss": 0.1958, "step": 805 }, { "epoch": 0.27947295423023577, "grad_norm": 5.376789093017578, "learning_rate": 1e-06, "loss": 0.197, "step": 806 }, { "epoch": 0.27981969486823854, "grad_norm": 2.796621084213257, "learning_rate": 1e-06, "loss": 0.2056, "step": 807 }, { "epoch": 0.2801664355062413, "grad_norm": 1.9890904426574707, "learning_rate": 1e-06, "loss": 0.1505, "step": 808 }, { "epoch": 0.2805131761442441, "grad_norm": 3.3186163902282715, "learning_rate": 1e-06, "loss": 0.1487, "step": 809 }, { "epoch": 0.28085991678224687, "grad_norm": 3.1521759033203125, "learning_rate": 1e-06, "loss": 0.1981, "step": 810 }, { "epoch": 0.28120665742024964, "grad_norm": 1.7996686697006226, "learning_rate": 1e-06, "loss": 0.2118, "step": 811 }, { "epoch": 0.2815533980582524, "grad_norm": 6.712550640106201, "learning_rate": 1e-06, "loss": 0.1817, "step": 812 }, { "epoch": 0.2819001386962552, "grad_norm": 4.741528511047363, "learning_rate": 1e-06, "loss": 0.1332, "step": 813 }, { "epoch": 0.28224687933425796, "grad_norm": 1.8851327896118164, "learning_rate": 1e-06, "loss": 0.2028, "step": 814 }, { "epoch": 0.28259361997226073, "grad_norm": 2.5891757011413574, "learning_rate": 1e-06, "loss": 0.2061, "step": 815 }, { "epoch": 0.2829403606102635, "grad_norm": 2.130458116531372, "learning_rate": 1e-06, "loss": 0.1825, "step": 816 }, { "epoch": 0.2832871012482663, "grad_norm": 2.5108346939086914, "learning_rate": 1e-06, "loss": 0.1774, "step": 817 }, { "epoch": 0.28363384188626906, "grad_norm": 1.7813079357147217, "learning_rate": 1e-06, "loss": 0.1809, "step": 818 }, { "epoch": 0.28398058252427183, "grad_norm": 1.8457926511764526, "learning_rate": 1e-06, "loss": 0.1468, "step": 819 }, { "epoch": 0.2843273231622746, "grad_norm": 2.4858312606811523, "learning_rate": 1e-06, "loss": 0.1721, "step": 820 }, { "epoch": 0.2846740638002774, "grad_norm": 2.8823843002319336, "learning_rate": 1e-06, "loss": 0.1908, "step": 821 }, { "epoch": 0.28502080443828015, "grad_norm": 2.092439889907837, "learning_rate": 1e-06, "loss": 0.2033, "step": 822 }, { "epoch": 0.2853675450762829, "grad_norm": 2.2995755672454834, "learning_rate": 1e-06, "loss": 0.1974, "step": 823 }, { "epoch": 0.2857142857142857, "grad_norm": 1.6263433694839478, "learning_rate": 1e-06, "loss": 0.1792, "step": 824 }, { "epoch": 0.28606102635228847, "grad_norm": 3.589700222015381, "learning_rate": 1e-06, "loss": 0.1548, "step": 825 }, { "epoch": 0.28640776699029125, "grad_norm": 2.6296792030334473, "learning_rate": 1e-06, "loss": 0.177, "step": 826 }, { "epoch": 0.286754507628294, "grad_norm": 3.712181568145752, "learning_rate": 1e-06, "loss": 0.1846, "step": 827 }, { "epoch": 0.2871012482662968, "grad_norm": 8.185032844543457, "learning_rate": 1e-06, "loss": 0.2371, "step": 828 }, { "epoch": 0.28744798890429957, "grad_norm": 3.00469708442688, "learning_rate": 1e-06, "loss": 0.1415, "step": 829 }, { "epoch": 0.28779472954230234, "grad_norm": 3.0410897731781006, "learning_rate": 1e-06, "loss": 0.1644, "step": 830 }, { "epoch": 0.2881414701803051, "grad_norm": 2.678393840789795, "learning_rate": 1e-06, "loss": 0.2029, "step": 831 }, { "epoch": 0.2884882108183079, "grad_norm": 2.8907225131988525, "learning_rate": 1e-06, "loss": 0.1449, "step": 832 }, { "epoch": 0.28883495145631066, "grad_norm": 1.4863390922546387, "learning_rate": 1e-06, "loss": 0.1468, "step": 833 }, { "epoch": 0.28918169209431344, "grad_norm": 1.769559621810913, "learning_rate": 1e-06, "loss": 0.1932, "step": 834 }, { "epoch": 0.2895284327323162, "grad_norm": 1.6298843622207642, "learning_rate": 1e-06, "loss": 0.1409, "step": 835 }, { "epoch": 0.289875173370319, "grad_norm": 4.213280200958252, "learning_rate": 1e-06, "loss": 0.1718, "step": 836 }, { "epoch": 0.29022191400832176, "grad_norm": 5.342798233032227, "learning_rate": 1e-06, "loss": 0.2073, "step": 837 }, { "epoch": 0.29056865464632453, "grad_norm": 2.607419967651367, "learning_rate": 1e-06, "loss": 0.175, "step": 838 }, { "epoch": 0.2909153952843273, "grad_norm": 3.2880568504333496, "learning_rate": 1e-06, "loss": 0.2236, "step": 839 }, { "epoch": 0.2912621359223301, "grad_norm": 2.370192527770996, "learning_rate": 1e-06, "loss": 0.1806, "step": 840 }, { "epoch": 0.29160887656033285, "grad_norm": 1.8948553800582886, "learning_rate": 1e-06, "loss": 0.1832, "step": 841 }, { "epoch": 0.2919556171983356, "grad_norm": 2.092898368835449, "learning_rate": 1e-06, "loss": 0.1868, "step": 842 }, { "epoch": 0.2923023578363384, "grad_norm": 1.5037286281585693, "learning_rate": 1e-06, "loss": 0.1581, "step": 843 }, { "epoch": 0.2926490984743412, "grad_norm": 3.512408494949341, "learning_rate": 1e-06, "loss": 0.1883, "step": 844 }, { "epoch": 0.29299583911234395, "grad_norm": 7.901118278503418, "learning_rate": 1e-06, "loss": 0.1555, "step": 845 }, { "epoch": 0.2933425797503467, "grad_norm": 5.720544815063477, "learning_rate": 1e-06, "loss": 0.1626, "step": 846 }, { "epoch": 0.2936893203883495, "grad_norm": 2.572035312652588, "learning_rate": 1e-06, "loss": 0.2006, "step": 847 }, { "epoch": 0.29403606102635227, "grad_norm": 1.5427334308624268, "learning_rate": 1e-06, "loss": 0.1666, "step": 848 }, { "epoch": 0.29438280166435504, "grad_norm": 2.345066547393799, "learning_rate": 1e-06, "loss": 0.1867, "step": 849 }, { "epoch": 0.2947295423023578, "grad_norm": 2.536377429962158, "learning_rate": 1e-06, "loss": 0.1918, "step": 850 }, { "epoch": 0.2950762829403606, "grad_norm": 1.5710004568099976, "learning_rate": 1e-06, "loss": 0.1747, "step": 851 }, { "epoch": 0.29542302357836336, "grad_norm": 1.5936557054519653, "learning_rate": 1e-06, "loss": 0.1569, "step": 852 }, { "epoch": 0.29576976421636614, "grad_norm": 4.063443660736084, "learning_rate": 1e-06, "loss": 0.1922, "step": 853 }, { "epoch": 0.2961165048543689, "grad_norm": 1.6471939086914062, "learning_rate": 1e-06, "loss": 0.1688, "step": 854 }, { "epoch": 0.2964632454923717, "grad_norm": 3.7626683712005615, "learning_rate": 1e-06, "loss": 0.1441, "step": 855 }, { "epoch": 0.29680998613037446, "grad_norm": 3.8700716495513916, "learning_rate": 1e-06, "loss": 0.1727, "step": 856 }, { "epoch": 0.29715672676837723, "grad_norm": 3.989415407180786, "learning_rate": 1e-06, "loss": 0.1821, "step": 857 }, { "epoch": 0.29750346740638, "grad_norm": 5.3771538734436035, "learning_rate": 1e-06, "loss": 0.2133, "step": 858 }, { "epoch": 0.2978502080443828, "grad_norm": 1.7356910705566406, "learning_rate": 1e-06, "loss": 0.1702, "step": 859 }, { "epoch": 0.29819694868238555, "grad_norm": 1.795180320739746, "learning_rate": 1e-06, "loss": 0.1836, "step": 860 }, { "epoch": 0.29854368932038833, "grad_norm": 4.196115970611572, "learning_rate": 1e-06, "loss": 0.173, "step": 861 }, { "epoch": 0.2988904299583911, "grad_norm": 2.782108783721924, "learning_rate": 1e-06, "loss": 0.1905, "step": 862 }, { "epoch": 0.2992371705963939, "grad_norm": 3.027663230895996, "learning_rate": 1e-06, "loss": 0.1974, "step": 863 }, { "epoch": 0.29958391123439665, "grad_norm": 1.9452866315841675, "learning_rate": 1e-06, "loss": 0.1655, "step": 864 }, { "epoch": 0.2999306518723994, "grad_norm": 2.2103707790374756, "learning_rate": 1e-06, "loss": 0.2024, "step": 865 }, { "epoch": 0.3002773925104022, "grad_norm": 2.538545846939087, "learning_rate": 1e-06, "loss": 0.1733, "step": 866 }, { "epoch": 0.30062413314840497, "grad_norm": 2.4375507831573486, "learning_rate": 1e-06, "loss": 0.1841, "step": 867 }, { "epoch": 0.30097087378640774, "grad_norm": 2.466561794281006, "learning_rate": 1e-06, "loss": 0.1835, "step": 868 }, { "epoch": 0.3013176144244105, "grad_norm": 1.5926258563995361, "learning_rate": 1e-06, "loss": 0.1591, "step": 869 }, { "epoch": 0.3016643550624133, "grad_norm": 1.525660514831543, "learning_rate": 1e-06, "loss": 0.1355, "step": 870 }, { "epoch": 0.30201109570041607, "grad_norm": 1.8583039045333862, "learning_rate": 1e-06, "loss": 0.164, "step": 871 }, { "epoch": 0.30235783633841884, "grad_norm": 7.7814531326293945, "learning_rate": 1e-06, "loss": 0.2342, "step": 872 }, { "epoch": 0.3027045769764216, "grad_norm": 1.736899495124817, "learning_rate": 1e-06, "loss": 0.1928, "step": 873 }, { "epoch": 0.3030513176144244, "grad_norm": 2.617612361907959, "learning_rate": 1e-06, "loss": 0.2155, "step": 874 }, { "epoch": 0.30339805825242716, "grad_norm": 4.537561416625977, "learning_rate": 1e-06, "loss": 0.1747, "step": 875 }, { "epoch": 0.30374479889042993, "grad_norm": 3.117304563522339, "learning_rate": 1e-06, "loss": 0.1786, "step": 876 }, { "epoch": 0.3040915395284327, "grad_norm": 1.9936450719833374, "learning_rate": 1e-06, "loss": 0.1963, "step": 877 }, { "epoch": 0.3044382801664355, "grad_norm": 2.502594470977783, "learning_rate": 1e-06, "loss": 0.1719, "step": 878 }, { "epoch": 0.30478502080443826, "grad_norm": 2.9981226921081543, "learning_rate": 1e-06, "loss": 0.1975, "step": 879 }, { "epoch": 0.30513176144244103, "grad_norm": 1.8054769039154053, "learning_rate": 1e-06, "loss": 0.1663, "step": 880 }, { "epoch": 0.3054785020804438, "grad_norm": 1.8026496171951294, "learning_rate": 1e-06, "loss": 0.183, "step": 881 }, { "epoch": 0.3058252427184466, "grad_norm": 2.0367887020111084, "learning_rate": 1e-06, "loss": 0.1754, "step": 882 }, { "epoch": 0.30617198335644935, "grad_norm": 1.685020089149475, "learning_rate": 1e-06, "loss": 0.1749, "step": 883 }, { "epoch": 0.3065187239944521, "grad_norm": 2.1757829189300537, "learning_rate": 1e-06, "loss": 0.2091, "step": 884 }, { "epoch": 0.3068654646324549, "grad_norm": 5.054877281188965, "learning_rate": 1e-06, "loss": 0.1787, "step": 885 }, { "epoch": 0.30721220527045767, "grad_norm": 1.8964219093322754, "learning_rate": 1e-06, "loss": 0.1853, "step": 886 }, { "epoch": 0.30755894590846045, "grad_norm": 6.154984951019287, "learning_rate": 1e-06, "loss": 0.1642, "step": 887 }, { "epoch": 0.3079056865464632, "grad_norm": 3.810920238494873, "learning_rate": 1e-06, "loss": 0.1482, "step": 888 }, { "epoch": 0.308252427184466, "grad_norm": 1.7902008295059204, "learning_rate": 1e-06, "loss": 0.2041, "step": 889 }, { "epoch": 0.30859916782246877, "grad_norm": 2.2116265296936035, "learning_rate": 1e-06, "loss": 0.1945, "step": 890 }, { "epoch": 0.30894590846047154, "grad_norm": 4.08473014831543, "learning_rate": 1e-06, "loss": 0.1814, "step": 891 }, { "epoch": 0.3092926490984743, "grad_norm": 4.12503719329834, "learning_rate": 1e-06, "loss": 0.1697, "step": 892 }, { "epoch": 0.3096393897364771, "grad_norm": 2.198702096939087, "learning_rate": 1e-06, "loss": 0.1646, "step": 893 }, { "epoch": 0.30998613037447986, "grad_norm": 2.104871988296509, "learning_rate": 1e-06, "loss": 0.1837, "step": 894 }, { "epoch": 0.31033287101248264, "grad_norm": 1.7803840637207031, "learning_rate": 1e-06, "loss": 0.1687, "step": 895 }, { "epoch": 0.3106796116504854, "grad_norm": 2.5232043266296387, "learning_rate": 1e-06, "loss": 0.1852, "step": 896 }, { "epoch": 0.3110263522884882, "grad_norm": 1.5908442735671997, "learning_rate": 1e-06, "loss": 0.161, "step": 897 }, { "epoch": 0.31137309292649096, "grad_norm": 3.704270839691162, "learning_rate": 1e-06, "loss": 0.1547, "step": 898 }, { "epoch": 0.31171983356449373, "grad_norm": 1.6646925210952759, "learning_rate": 1e-06, "loss": 0.1621, "step": 899 }, { "epoch": 0.3120665742024965, "grad_norm": 1.6272737979888916, "learning_rate": 1e-06, "loss": 0.1738, "step": 900 }, { "epoch": 0.3124133148404993, "grad_norm": 2.2414157390594482, "learning_rate": 1e-06, "loss": 0.1857, "step": 901 }, { "epoch": 0.3127600554785021, "grad_norm": 4.613629341125488, "learning_rate": 1e-06, "loss": 0.2038, "step": 902 }, { "epoch": 0.3131067961165049, "grad_norm": 2.6360721588134766, "learning_rate": 1e-06, "loss": 0.1689, "step": 903 }, { "epoch": 0.31345353675450766, "grad_norm": 2.4351446628570557, "learning_rate": 1e-06, "loss": 0.1699, "step": 904 }, { "epoch": 0.31380027739251043, "grad_norm": 1.645674228668213, "learning_rate": 1e-06, "loss": 0.1747, "step": 905 }, { "epoch": 0.3141470180305132, "grad_norm": 2.1112935543060303, "learning_rate": 1e-06, "loss": 0.1725, "step": 906 }, { "epoch": 0.314493758668516, "grad_norm": 1.8450239896774292, "learning_rate": 1e-06, "loss": 0.1783, "step": 907 }, { "epoch": 0.31484049930651875, "grad_norm": 3.959017276763916, "learning_rate": 1e-06, "loss": 0.165, "step": 908 }, { "epoch": 0.3151872399445215, "grad_norm": 2.1411192417144775, "learning_rate": 1e-06, "loss": 0.1566, "step": 909 }, { "epoch": 0.3155339805825243, "grad_norm": 2.4546244144439697, "learning_rate": 1e-06, "loss": 0.193, "step": 910 }, { "epoch": 0.31588072122052707, "grad_norm": 4.3430280685424805, "learning_rate": 1e-06, "loss": 0.1721, "step": 911 }, { "epoch": 0.31622746185852985, "grad_norm": 3.988929271697998, "learning_rate": 1e-06, "loss": 0.159, "step": 912 }, { "epoch": 0.3165742024965326, "grad_norm": 3.2247719764709473, "learning_rate": 1e-06, "loss": 0.1624, "step": 913 }, { "epoch": 0.3169209431345354, "grad_norm": 1.647119402885437, "learning_rate": 1e-06, "loss": 0.1659, "step": 914 }, { "epoch": 0.31726768377253817, "grad_norm": 3.0910604000091553, "learning_rate": 1e-06, "loss": 0.18, "step": 915 }, { "epoch": 0.31761442441054094, "grad_norm": 1.7037394046783447, "learning_rate": 1e-06, "loss": 0.2035, "step": 916 }, { "epoch": 0.3179611650485437, "grad_norm": 3.9424848556518555, "learning_rate": 1e-06, "loss": 0.1755, "step": 917 }, { "epoch": 0.3183079056865465, "grad_norm": 1.8996752500534058, "learning_rate": 1e-06, "loss": 0.1674, "step": 918 }, { "epoch": 0.31865464632454926, "grad_norm": 1.9568631649017334, "learning_rate": 1e-06, "loss": 0.1741, "step": 919 }, { "epoch": 0.31900138696255204, "grad_norm": 2.990325927734375, "learning_rate": 1e-06, "loss": 0.1907, "step": 920 }, { "epoch": 0.3193481276005548, "grad_norm": 1.8824876546859741, "learning_rate": 1e-06, "loss": 0.1751, "step": 921 }, { "epoch": 0.3196948682385576, "grad_norm": 1.7028950452804565, "learning_rate": 1e-06, "loss": 0.1362, "step": 922 }, { "epoch": 0.32004160887656036, "grad_norm": 2.010843515396118, "learning_rate": 1e-06, "loss": 0.1758, "step": 923 }, { "epoch": 0.32038834951456313, "grad_norm": 1.7268744707107544, "learning_rate": 1e-06, "loss": 0.1853, "step": 924 }, { "epoch": 0.3207350901525659, "grad_norm": 2.266728162765503, "learning_rate": 1e-06, "loss": 0.1644, "step": 925 }, { "epoch": 0.3210818307905687, "grad_norm": 5.072295188903809, "learning_rate": 1e-06, "loss": 0.1793, "step": 926 }, { "epoch": 0.32142857142857145, "grad_norm": 2.210855484008789, "learning_rate": 1e-06, "loss": 0.1425, "step": 927 }, { "epoch": 0.3217753120665742, "grad_norm": 2.0803961753845215, "learning_rate": 1e-06, "loss": 0.1651, "step": 928 }, { "epoch": 0.322122052704577, "grad_norm": 2.160520076751709, "learning_rate": 1e-06, "loss": 0.1596, "step": 929 }, { "epoch": 0.3224687933425798, "grad_norm": 1.9977649450302124, "learning_rate": 1e-06, "loss": 0.1876, "step": 930 }, { "epoch": 0.32281553398058255, "grad_norm": 1.5417920351028442, "learning_rate": 1e-06, "loss": 0.1503, "step": 931 }, { "epoch": 0.3231622746185853, "grad_norm": 2.099778175354004, "learning_rate": 1e-06, "loss": 0.1836, "step": 932 }, { "epoch": 0.3235090152565881, "grad_norm": 3.9819977283477783, "learning_rate": 1e-06, "loss": 0.1587, "step": 933 }, { "epoch": 0.32385575589459087, "grad_norm": 1.6170867681503296, "learning_rate": 1e-06, "loss": 0.147, "step": 934 }, { "epoch": 0.32420249653259364, "grad_norm": 3.3660576343536377, "learning_rate": 1e-06, "loss": 0.1445, "step": 935 }, { "epoch": 0.3245492371705964, "grad_norm": 2.2826545238494873, "learning_rate": 1e-06, "loss": 0.1579, "step": 936 }, { "epoch": 0.3248959778085992, "grad_norm": 5.107556343078613, "learning_rate": 1e-06, "loss": 0.1964, "step": 937 }, { "epoch": 0.32524271844660196, "grad_norm": 3.9381065368652344, "learning_rate": 1e-06, "loss": 0.2003, "step": 938 }, { "epoch": 0.32558945908460474, "grad_norm": 2.5233023166656494, "learning_rate": 1e-06, "loss": 0.1747, "step": 939 }, { "epoch": 0.3259361997226075, "grad_norm": 1.8627601861953735, "learning_rate": 1e-06, "loss": 0.177, "step": 940 }, { "epoch": 0.3262829403606103, "grad_norm": 1.7760263681411743, "learning_rate": 1e-06, "loss": 0.1548, "step": 941 }, { "epoch": 0.32662968099861306, "grad_norm": 2.0792856216430664, "learning_rate": 1e-06, "loss": 0.1762, "step": 942 }, { "epoch": 0.32697642163661583, "grad_norm": 4.968944549560547, "learning_rate": 1e-06, "loss": 0.163, "step": 943 }, { "epoch": 0.3273231622746186, "grad_norm": 2.9295568466186523, "learning_rate": 1e-06, "loss": 0.1928, "step": 944 }, { "epoch": 0.3276699029126214, "grad_norm": 3.2143867015838623, "learning_rate": 1e-06, "loss": 0.1401, "step": 945 }, { "epoch": 0.32801664355062415, "grad_norm": 4.367483139038086, "learning_rate": 1e-06, "loss": 0.2002, "step": 946 }, { "epoch": 0.32836338418862693, "grad_norm": 2.355031967163086, "learning_rate": 1e-06, "loss": 0.1719, "step": 947 }, { "epoch": 0.3287101248266297, "grad_norm": 2.9974913597106934, "learning_rate": 1e-06, "loss": 0.1559, "step": 948 }, { "epoch": 0.3290568654646325, "grad_norm": 2.6766953468322754, "learning_rate": 1e-06, "loss": 0.1834, "step": 949 }, { "epoch": 0.32940360610263525, "grad_norm": 3.276123523712158, "learning_rate": 1e-06, "loss": 0.1751, "step": 950 }, { "epoch": 0.329750346740638, "grad_norm": 1.6151905059814453, "learning_rate": 1e-06, "loss": 0.1809, "step": 951 }, { "epoch": 0.3300970873786408, "grad_norm": 1.8168013095855713, "learning_rate": 1e-06, "loss": 0.1436, "step": 952 }, { "epoch": 0.33044382801664357, "grad_norm": 4.220736503601074, "learning_rate": 1e-06, "loss": 0.1696, "step": 953 }, { "epoch": 0.33079056865464634, "grad_norm": 2.646458625793457, "learning_rate": 1e-06, "loss": 0.14, "step": 954 }, { "epoch": 0.3311373092926491, "grad_norm": 5.339474201202393, "learning_rate": 1e-06, "loss": 0.2126, "step": 955 }, { "epoch": 0.3314840499306519, "grad_norm": 3.148782730102539, "learning_rate": 1e-06, "loss": 0.1692, "step": 956 }, { "epoch": 0.33183079056865467, "grad_norm": 3.1843745708465576, "learning_rate": 1e-06, "loss": 0.1804, "step": 957 }, { "epoch": 0.33217753120665744, "grad_norm": 2.0136914253234863, "learning_rate": 1e-06, "loss": 0.1817, "step": 958 }, { "epoch": 0.3325242718446602, "grad_norm": 2.250756025314331, "learning_rate": 1e-06, "loss": 0.1885, "step": 959 }, { "epoch": 0.332871012482663, "grad_norm": 3.832793951034546, "learning_rate": 1e-06, "loss": 0.1617, "step": 960 }, { "epoch": 0.33321775312066576, "grad_norm": 3.097174644470215, "learning_rate": 1e-06, "loss": 0.163, "step": 961 }, { "epoch": 0.33356449375866853, "grad_norm": 3.634685516357422, "learning_rate": 1e-06, "loss": 0.1465, "step": 962 }, { "epoch": 0.3339112343966713, "grad_norm": 1.968932032585144, "learning_rate": 1e-06, "loss": 0.1819, "step": 963 }, { "epoch": 0.3342579750346741, "grad_norm": 3.506141424179077, "learning_rate": 1e-06, "loss": 0.1623, "step": 964 }, { "epoch": 0.33460471567267686, "grad_norm": 2.201019048690796, "learning_rate": 1e-06, "loss": 0.1987, "step": 965 }, { "epoch": 0.33495145631067963, "grad_norm": 1.871758222579956, "learning_rate": 1e-06, "loss": 0.1651, "step": 966 }, { "epoch": 0.3352981969486824, "grad_norm": 2.852182626724243, "learning_rate": 1e-06, "loss": 0.1751, "step": 967 }, { "epoch": 0.3356449375866852, "grad_norm": 2.830751657485962, "learning_rate": 1e-06, "loss": 0.1934, "step": 968 }, { "epoch": 0.33599167822468795, "grad_norm": 2.9230191707611084, "learning_rate": 1e-06, "loss": 0.1689, "step": 969 }, { "epoch": 0.3363384188626907, "grad_norm": 3.4389359951019287, "learning_rate": 1e-06, "loss": 0.1606, "step": 970 }, { "epoch": 0.3366851595006935, "grad_norm": 3.4921634197235107, "learning_rate": 1e-06, "loss": 0.1891, "step": 971 }, { "epoch": 0.33703190013869627, "grad_norm": 2.9009199142456055, "learning_rate": 1e-06, "loss": 0.1508, "step": 972 }, { "epoch": 0.33737864077669905, "grad_norm": 1.8419466018676758, "learning_rate": 1e-06, "loss": 0.173, "step": 973 }, { "epoch": 0.3377253814147018, "grad_norm": 4.729164123535156, "learning_rate": 1e-06, "loss": 0.1863, "step": 974 }, { "epoch": 0.3380721220527046, "grad_norm": 4.153334617614746, "learning_rate": 1e-06, "loss": 0.189, "step": 975 }, { "epoch": 0.33841886269070737, "grad_norm": 2.406278371810913, "learning_rate": 1e-06, "loss": 0.1726, "step": 976 }, { "epoch": 0.33876560332871014, "grad_norm": 1.854493498802185, "learning_rate": 1e-06, "loss": 0.1859, "step": 977 }, { "epoch": 0.3391123439667129, "grad_norm": 1.7517307996749878, "learning_rate": 1e-06, "loss": 0.1849, "step": 978 }, { "epoch": 0.3394590846047157, "grad_norm": 3.5556588172912598, "learning_rate": 1e-06, "loss": 0.1691, "step": 979 }, { "epoch": 0.33980582524271846, "grad_norm": 3.863245725631714, "learning_rate": 1e-06, "loss": 0.1758, "step": 980 }, { "epoch": 0.34015256588072124, "grad_norm": 4.660958766937256, "learning_rate": 1e-06, "loss": 0.1492, "step": 981 }, { "epoch": 0.340499306518724, "grad_norm": 2.8175435066223145, "learning_rate": 1e-06, "loss": 0.1559, "step": 982 }, { "epoch": 0.3408460471567268, "grad_norm": 3.012770891189575, "learning_rate": 1e-06, "loss": 0.1876, "step": 983 }, { "epoch": 0.34119278779472956, "grad_norm": 3.963909387588501, "learning_rate": 1e-06, "loss": 0.1512, "step": 984 }, { "epoch": 0.34153952843273233, "grad_norm": 1.6655025482177734, "learning_rate": 1e-06, "loss": 0.1429, "step": 985 }, { "epoch": 0.3418862690707351, "grad_norm": 2.783829689025879, "learning_rate": 1e-06, "loss": 0.2169, "step": 986 }, { "epoch": 0.3422330097087379, "grad_norm": 1.5330981016159058, "learning_rate": 1e-06, "loss": 0.1479, "step": 987 }, { "epoch": 0.34257975034674065, "grad_norm": 1.6944447755813599, "learning_rate": 1e-06, "loss": 0.1762, "step": 988 }, { "epoch": 0.3429264909847434, "grad_norm": 2.4283335208892822, "learning_rate": 1e-06, "loss": 0.1577, "step": 989 }, { "epoch": 0.3432732316227462, "grad_norm": 2.7645480632781982, "learning_rate": 1e-06, "loss": 0.1697, "step": 990 }, { "epoch": 0.343619972260749, "grad_norm": 4.196599960327148, "learning_rate": 1e-06, "loss": 0.1777, "step": 991 }, { "epoch": 0.34396671289875175, "grad_norm": 1.866174340248108, "learning_rate": 1e-06, "loss": 0.1461, "step": 992 }, { "epoch": 0.3443134535367545, "grad_norm": 2.0059845447540283, "learning_rate": 1e-06, "loss": 0.182, "step": 993 }, { "epoch": 0.3446601941747573, "grad_norm": 1.9645408391952515, "learning_rate": 1e-06, "loss": 0.1908, "step": 994 }, { "epoch": 0.34500693481276007, "grad_norm": 2.1238350868225098, "learning_rate": 1e-06, "loss": 0.2052, "step": 995 }, { "epoch": 0.34535367545076284, "grad_norm": 1.8180639743804932, "learning_rate": 1e-06, "loss": 0.1684, "step": 996 }, { "epoch": 0.3457004160887656, "grad_norm": 2.0264620780944824, "learning_rate": 1e-06, "loss": 0.1827, "step": 997 }, { "epoch": 0.3460471567267684, "grad_norm": 2.302675724029541, "learning_rate": 1e-06, "loss": 0.1999, "step": 998 }, { "epoch": 0.34639389736477116, "grad_norm": 2.18147873878479, "learning_rate": 1e-06, "loss": 0.1694, "step": 999 }, { "epoch": 0.34674063800277394, "grad_norm": 7.13607931137085, "learning_rate": 1e-06, "loss": 0.1692, "step": 1000 }, { "epoch": 0.3470873786407767, "grad_norm": 3.1698849201202393, "learning_rate": 1e-06, "loss": 0.1861, "step": 1001 }, { "epoch": 0.3474341192787795, "grad_norm": 2.142284393310547, "learning_rate": 1e-06, "loss": 0.157, "step": 1002 }, { "epoch": 0.34778085991678226, "grad_norm": 2.648524284362793, "learning_rate": 1e-06, "loss": 0.1809, "step": 1003 }, { "epoch": 0.34812760055478503, "grad_norm": 1.6139206886291504, "learning_rate": 1e-06, "loss": 0.1659, "step": 1004 }, { "epoch": 0.3484743411927878, "grad_norm": 1.9868088960647583, "learning_rate": 1e-06, "loss": 0.142, "step": 1005 }, { "epoch": 0.3488210818307906, "grad_norm": 1.7184799909591675, "learning_rate": 1e-06, "loss": 0.1547, "step": 1006 }, { "epoch": 0.34916782246879335, "grad_norm": 1.670831561088562, "learning_rate": 1e-06, "loss": 0.1624, "step": 1007 }, { "epoch": 0.34951456310679613, "grad_norm": 1.932572364807129, "learning_rate": 1e-06, "loss": 0.1428, "step": 1008 }, { "epoch": 0.3498613037447989, "grad_norm": 3.854353189468384, "learning_rate": 1e-06, "loss": 0.1436, "step": 1009 }, { "epoch": 0.3502080443828017, "grad_norm": 1.7538644075393677, "learning_rate": 1e-06, "loss": 0.1491, "step": 1010 }, { "epoch": 0.35055478502080445, "grad_norm": 1.8496301174163818, "learning_rate": 1e-06, "loss": 0.1756, "step": 1011 }, { "epoch": 0.3509015256588072, "grad_norm": 4.00302267074585, "learning_rate": 1e-06, "loss": 0.1809, "step": 1012 }, { "epoch": 0.35124826629681, "grad_norm": 1.931908369064331, "learning_rate": 1e-06, "loss": 0.1675, "step": 1013 }, { "epoch": 0.35159500693481277, "grad_norm": 1.933568000793457, "learning_rate": 1e-06, "loss": 0.1556, "step": 1014 }, { "epoch": 0.35194174757281554, "grad_norm": 1.5976754426956177, "learning_rate": 1e-06, "loss": 0.1566, "step": 1015 }, { "epoch": 0.3522884882108183, "grad_norm": 1.9067904949188232, "learning_rate": 1e-06, "loss": 0.157, "step": 1016 }, { "epoch": 0.3526352288488211, "grad_norm": 2.5264508724212646, "learning_rate": 1e-06, "loss": 0.1512, "step": 1017 }, { "epoch": 0.35298196948682387, "grad_norm": 3.613884925842285, "learning_rate": 1e-06, "loss": 0.1775, "step": 1018 }, { "epoch": 0.35332871012482664, "grad_norm": 2.958832263946533, "learning_rate": 1e-06, "loss": 0.1838, "step": 1019 }, { "epoch": 0.3536754507628294, "grad_norm": 1.8982412815093994, "learning_rate": 1e-06, "loss": 0.1416, "step": 1020 }, { "epoch": 0.3540221914008322, "grad_norm": 2.6603124141693115, "learning_rate": 1e-06, "loss": 0.1839, "step": 1021 }, { "epoch": 0.35436893203883496, "grad_norm": 3.2685227394104004, "learning_rate": 1e-06, "loss": 0.1893, "step": 1022 }, { "epoch": 0.35471567267683773, "grad_norm": 2.6243014335632324, "learning_rate": 1e-06, "loss": 0.1805, "step": 1023 }, { "epoch": 0.3550624133148405, "grad_norm": 3.69974946975708, "learning_rate": 1e-06, "loss": 0.199, "step": 1024 }, { "epoch": 0.3554091539528433, "grad_norm": 1.906920313835144, "learning_rate": 1e-06, "loss": 0.1526, "step": 1025 }, { "epoch": 0.35575589459084606, "grad_norm": 4.538094520568848, "learning_rate": 1e-06, "loss": 0.2041, "step": 1026 }, { "epoch": 0.35610263522884883, "grad_norm": 1.9397320747375488, "learning_rate": 1e-06, "loss": 0.1619, "step": 1027 }, { "epoch": 0.3564493758668516, "grad_norm": 3.0926973819732666, "learning_rate": 1e-06, "loss": 0.193, "step": 1028 }, { "epoch": 0.3567961165048544, "grad_norm": 6.100024223327637, "learning_rate": 1e-06, "loss": 0.1759, "step": 1029 }, { "epoch": 0.35714285714285715, "grad_norm": 1.7386844158172607, "learning_rate": 1e-06, "loss": 0.1573, "step": 1030 }, { "epoch": 0.3574895977808599, "grad_norm": 2.905871629714966, "learning_rate": 1e-06, "loss": 0.1522, "step": 1031 }, { "epoch": 0.3578363384188627, "grad_norm": 2.167205333709717, "learning_rate": 1e-06, "loss": 0.1628, "step": 1032 }, { "epoch": 0.35818307905686547, "grad_norm": 2.418584108352661, "learning_rate": 1e-06, "loss": 0.1755, "step": 1033 }, { "epoch": 0.35852981969486825, "grad_norm": 2.9872934818267822, "learning_rate": 1e-06, "loss": 0.174, "step": 1034 }, { "epoch": 0.358876560332871, "grad_norm": 3.5263776779174805, "learning_rate": 1e-06, "loss": 0.1802, "step": 1035 }, { "epoch": 0.3592233009708738, "grad_norm": 3.891584634780884, "learning_rate": 1e-06, "loss": 0.1996, "step": 1036 }, { "epoch": 0.35957004160887657, "grad_norm": 2.081064224243164, "learning_rate": 1e-06, "loss": 0.1578, "step": 1037 }, { "epoch": 0.35991678224687934, "grad_norm": 2.100369930267334, "learning_rate": 1e-06, "loss": 0.1572, "step": 1038 }, { "epoch": 0.3602635228848821, "grad_norm": 2.148822069168091, "learning_rate": 1e-06, "loss": 0.1788, "step": 1039 }, { "epoch": 0.3606102635228849, "grad_norm": 2.7405498027801514, "learning_rate": 1e-06, "loss": 0.1863, "step": 1040 }, { "epoch": 0.36095700416088766, "grad_norm": 4.158656597137451, "learning_rate": 1e-06, "loss": 0.1699, "step": 1041 }, { "epoch": 0.36130374479889044, "grad_norm": 2.858583688735962, "learning_rate": 1e-06, "loss": 0.1855, "step": 1042 }, { "epoch": 0.3616504854368932, "grad_norm": 5.225622177124023, "learning_rate": 1e-06, "loss": 0.1613, "step": 1043 }, { "epoch": 0.361997226074896, "grad_norm": 2.721017599105835, "learning_rate": 1e-06, "loss": 0.1532, "step": 1044 }, { "epoch": 0.36234396671289876, "grad_norm": 3.189161539077759, "learning_rate": 1e-06, "loss": 0.1966, "step": 1045 }, { "epoch": 0.36269070735090153, "grad_norm": 2.8146345615386963, "learning_rate": 1e-06, "loss": 0.1786, "step": 1046 }, { "epoch": 0.3630374479889043, "grad_norm": 3.777596950531006, "learning_rate": 1e-06, "loss": 0.1579, "step": 1047 }, { "epoch": 0.3633841886269071, "grad_norm": 2.2806339263916016, "learning_rate": 1e-06, "loss": 0.1447, "step": 1048 }, { "epoch": 0.36373092926490985, "grad_norm": 2.7215936183929443, "learning_rate": 1e-06, "loss": 0.187, "step": 1049 }, { "epoch": 0.3640776699029126, "grad_norm": 3.6170296669006348, "learning_rate": 1e-06, "loss": 0.192, "step": 1050 }, { "epoch": 0.3644244105409154, "grad_norm": 2.5755667686462402, "learning_rate": 1e-06, "loss": 0.1592, "step": 1051 }, { "epoch": 0.3647711511789182, "grad_norm": 2.1443443298339844, "learning_rate": 1e-06, "loss": 0.1704, "step": 1052 }, { "epoch": 0.36511789181692095, "grad_norm": 2.3216934204101562, "learning_rate": 1e-06, "loss": 0.1787, "step": 1053 }, { "epoch": 0.3654646324549237, "grad_norm": 2.0402567386627197, "learning_rate": 1e-06, "loss": 0.1474, "step": 1054 }, { "epoch": 0.3658113730929265, "grad_norm": 2.2533066272735596, "learning_rate": 1e-06, "loss": 0.2121, "step": 1055 }, { "epoch": 0.36615811373092927, "grad_norm": 2.638406753540039, "learning_rate": 1e-06, "loss": 0.165, "step": 1056 }, { "epoch": 0.36650485436893204, "grad_norm": 1.7857331037521362, "learning_rate": 1e-06, "loss": 0.1653, "step": 1057 }, { "epoch": 0.3668515950069348, "grad_norm": 4.700150966644287, "learning_rate": 1e-06, "loss": 0.1723, "step": 1058 }, { "epoch": 0.3671983356449376, "grad_norm": 1.8558968305587769, "learning_rate": 1e-06, "loss": 0.1674, "step": 1059 }, { "epoch": 0.36754507628294036, "grad_norm": 2.047917604446411, "learning_rate": 1e-06, "loss": 0.1757, "step": 1060 }, { "epoch": 0.36789181692094314, "grad_norm": 2.157421588897705, "learning_rate": 1e-06, "loss": 0.1655, "step": 1061 }, { "epoch": 0.3682385575589459, "grad_norm": 4.857705593109131, "learning_rate": 1e-06, "loss": 0.2087, "step": 1062 }, { "epoch": 0.3685852981969487, "grad_norm": 2.143380641937256, "learning_rate": 1e-06, "loss": 0.1595, "step": 1063 }, { "epoch": 0.36893203883495146, "grad_norm": 3.1953229904174805, "learning_rate": 1e-06, "loss": 0.1592, "step": 1064 }, { "epoch": 0.36927877947295423, "grad_norm": 1.5903429985046387, "learning_rate": 1e-06, "loss": 0.1705, "step": 1065 }, { "epoch": 0.369625520110957, "grad_norm": 2.1189424991607666, "learning_rate": 1e-06, "loss": 0.1611, "step": 1066 }, { "epoch": 0.3699722607489598, "grad_norm": 2.089324712753296, "learning_rate": 1e-06, "loss": 0.1625, "step": 1067 }, { "epoch": 0.37031900138696255, "grad_norm": 2.1971182823181152, "learning_rate": 1e-06, "loss": 0.2072, "step": 1068 }, { "epoch": 0.37066574202496533, "grad_norm": 3.347195625305176, "learning_rate": 1e-06, "loss": 0.1455, "step": 1069 }, { "epoch": 0.3710124826629681, "grad_norm": 3.526571035385132, "learning_rate": 1e-06, "loss": 0.154, "step": 1070 }, { "epoch": 0.3713592233009709, "grad_norm": 3.1888411045074463, "learning_rate": 1e-06, "loss": 0.1559, "step": 1071 }, { "epoch": 0.37170596393897365, "grad_norm": 2.727715253829956, "learning_rate": 1e-06, "loss": 0.1538, "step": 1072 }, { "epoch": 0.3720527045769764, "grad_norm": 2.437049150466919, "learning_rate": 1e-06, "loss": 0.1637, "step": 1073 }, { "epoch": 0.3723994452149792, "grad_norm": 4.141301155090332, "learning_rate": 1e-06, "loss": 0.192, "step": 1074 }, { "epoch": 0.37274618585298197, "grad_norm": 3.7499258518218994, "learning_rate": 1e-06, "loss": 0.1791, "step": 1075 }, { "epoch": 0.37309292649098474, "grad_norm": 1.5035871267318726, "learning_rate": 1e-06, "loss": 0.1553, "step": 1076 }, { "epoch": 0.3734396671289875, "grad_norm": 1.849098563194275, "learning_rate": 1e-06, "loss": 0.1851, "step": 1077 }, { "epoch": 0.3737864077669903, "grad_norm": 3.0276803970336914, "learning_rate": 1e-06, "loss": 0.1682, "step": 1078 }, { "epoch": 0.37413314840499307, "grad_norm": 3.6269848346710205, "learning_rate": 1e-06, "loss": 0.1584, "step": 1079 }, { "epoch": 0.37447988904299584, "grad_norm": 1.632621169090271, "learning_rate": 1e-06, "loss": 0.1598, "step": 1080 }, { "epoch": 0.3748266296809986, "grad_norm": 4.8789472579956055, "learning_rate": 1e-06, "loss": 0.1871, "step": 1081 }, { "epoch": 0.3751733703190014, "grad_norm": 2.575338363647461, "learning_rate": 1e-06, "loss": 0.1781, "step": 1082 }, { "epoch": 0.37552011095700416, "grad_norm": 5.345282077789307, "learning_rate": 1e-06, "loss": 0.1923, "step": 1083 }, { "epoch": 0.37586685159500693, "grad_norm": 1.7190324068069458, "learning_rate": 1e-06, "loss": 0.155, "step": 1084 }, { "epoch": 0.3762135922330097, "grad_norm": 2.7497055530548096, "learning_rate": 1e-06, "loss": 0.1667, "step": 1085 }, { "epoch": 0.3765603328710125, "grad_norm": 2.812009334564209, "learning_rate": 1e-06, "loss": 0.2117, "step": 1086 }, { "epoch": 0.37690707350901526, "grad_norm": 1.6642462015151978, "learning_rate": 1e-06, "loss": 0.1749, "step": 1087 }, { "epoch": 0.37725381414701803, "grad_norm": 1.9300440549850464, "learning_rate": 1e-06, "loss": 0.1683, "step": 1088 }, { "epoch": 0.3776005547850208, "grad_norm": 3.1472578048706055, "learning_rate": 1e-06, "loss": 0.2015, "step": 1089 }, { "epoch": 0.3779472954230236, "grad_norm": 2.2062807083129883, "learning_rate": 1e-06, "loss": 0.1713, "step": 1090 }, { "epoch": 0.37829403606102635, "grad_norm": 2.400383472442627, "learning_rate": 1e-06, "loss": 0.2022, "step": 1091 }, { "epoch": 0.3786407766990291, "grad_norm": 1.869706153869629, "learning_rate": 1e-06, "loss": 0.1863, "step": 1092 }, { "epoch": 0.3789875173370319, "grad_norm": 4.637445449829102, "learning_rate": 1e-06, "loss": 0.1642, "step": 1093 }, { "epoch": 0.37933425797503467, "grad_norm": 3.4108481407165527, "learning_rate": 1e-06, "loss": 0.1535, "step": 1094 }, { "epoch": 0.37968099861303745, "grad_norm": 2.6168668270111084, "learning_rate": 1e-06, "loss": 0.1479, "step": 1095 }, { "epoch": 0.3800277392510402, "grad_norm": 2.7562978267669678, "learning_rate": 1e-06, "loss": 0.2129, "step": 1096 }, { "epoch": 0.380374479889043, "grad_norm": 2.557950019836426, "learning_rate": 1e-06, "loss": 0.1644, "step": 1097 }, { "epoch": 0.38072122052704577, "grad_norm": 4.20210075378418, "learning_rate": 1e-06, "loss": 0.177, "step": 1098 }, { "epoch": 0.38106796116504854, "grad_norm": 2.3209750652313232, "learning_rate": 1e-06, "loss": 0.1768, "step": 1099 }, { "epoch": 0.3814147018030513, "grad_norm": 2.318815231323242, "learning_rate": 1e-06, "loss": 0.2034, "step": 1100 }, { "epoch": 0.3817614424410541, "grad_norm": 4.568731307983398, "learning_rate": 1e-06, "loss": 0.171, "step": 1101 }, { "epoch": 0.38210818307905686, "grad_norm": 1.882871150970459, "learning_rate": 1e-06, "loss": 0.1263, "step": 1102 }, { "epoch": 0.38245492371705964, "grad_norm": 2.3090946674346924, "learning_rate": 1e-06, "loss": 0.2043, "step": 1103 }, { "epoch": 0.3828016643550624, "grad_norm": 2.3740408420562744, "learning_rate": 1e-06, "loss": 0.1483, "step": 1104 }, { "epoch": 0.3831484049930652, "grad_norm": 3.6114842891693115, "learning_rate": 1e-06, "loss": 0.1848, "step": 1105 }, { "epoch": 0.38349514563106796, "grad_norm": 1.9591243267059326, "learning_rate": 1e-06, "loss": 0.1751, "step": 1106 }, { "epoch": 0.38384188626907073, "grad_norm": 1.7068411111831665, "learning_rate": 1e-06, "loss": 0.1449, "step": 1107 }, { "epoch": 0.3841886269070735, "grad_norm": 1.9355899095535278, "learning_rate": 1e-06, "loss": 0.1902, "step": 1108 }, { "epoch": 0.3845353675450763, "grad_norm": 6.148166179656982, "learning_rate": 1e-06, "loss": 0.1864, "step": 1109 }, { "epoch": 0.38488210818307905, "grad_norm": 3.4833147525787354, "learning_rate": 1e-06, "loss": 0.1521, "step": 1110 }, { "epoch": 0.3852288488210818, "grad_norm": 5.457380294799805, "learning_rate": 1e-06, "loss": 0.17, "step": 1111 }, { "epoch": 0.3855755894590846, "grad_norm": 2.54488468170166, "learning_rate": 1e-06, "loss": 0.187, "step": 1112 }, { "epoch": 0.3859223300970874, "grad_norm": 1.795543909072876, "learning_rate": 1e-06, "loss": 0.2005, "step": 1113 }, { "epoch": 0.38626907073509015, "grad_norm": 2.521865129470825, "learning_rate": 1e-06, "loss": 0.2035, "step": 1114 }, { "epoch": 0.3866158113730929, "grad_norm": 2.5379583835601807, "learning_rate": 1e-06, "loss": 0.184, "step": 1115 }, { "epoch": 0.3869625520110957, "grad_norm": 1.7145198583602905, "learning_rate": 1e-06, "loss": 0.1709, "step": 1116 }, { "epoch": 0.38730929264909847, "grad_norm": 3.1036429405212402, "learning_rate": 1e-06, "loss": 0.2025, "step": 1117 }, { "epoch": 0.38765603328710124, "grad_norm": 2.6453635692596436, "learning_rate": 1e-06, "loss": 0.1301, "step": 1118 }, { "epoch": 0.388002773925104, "grad_norm": 2.3852477073669434, "learning_rate": 1e-06, "loss": 0.1355, "step": 1119 }, { "epoch": 0.3883495145631068, "grad_norm": 1.9018865823745728, "learning_rate": 1e-06, "loss": 0.1618, "step": 1120 }, { "epoch": 0.38869625520110956, "grad_norm": 2.8111538887023926, "learning_rate": 1e-06, "loss": 0.1624, "step": 1121 }, { "epoch": 0.38904299583911234, "grad_norm": 2.226997137069702, "learning_rate": 1e-06, "loss": 0.1671, "step": 1122 }, { "epoch": 0.3893897364771151, "grad_norm": 2.0597381591796875, "learning_rate": 1e-06, "loss": 0.1604, "step": 1123 }, { "epoch": 0.3897364771151179, "grad_norm": 5.358310699462891, "learning_rate": 1e-06, "loss": 0.2211, "step": 1124 }, { "epoch": 0.39008321775312066, "grad_norm": 2.202601909637451, "learning_rate": 1e-06, "loss": 0.1544, "step": 1125 }, { "epoch": 0.39042995839112343, "grad_norm": 1.6089842319488525, "learning_rate": 1e-06, "loss": 0.1582, "step": 1126 }, { "epoch": 0.3907766990291262, "grad_norm": 2.50658917427063, "learning_rate": 1e-06, "loss": 0.1792, "step": 1127 }, { "epoch": 0.391123439667129, "grad_norm": 1.6953543424606323, "learning_rate": 1e-06, "loss": 0.1624, "step": 1128 }, { "epoch": 0.39147018030513175, "grad_norm": 1.8754124641418457, "learning_rate": 1e-06, "loss": 0.1586, "step": 1129 }, { "epoch": 0.39181692094313453, "grad_norm": 4.671680450439453, "learning_rate": 1e-06, "loss": 0.1969, "step": 1130 }, { "epoch": 0.3921636615811373, "grad_norm": 2.1152446269989014, "learning_rate": 1e-06, "loss": 0.1634, "step": 1131 }, { "epoch": 0.3925104022191401, "grad_norm": 4.22470235824585, "learning_rate": 1e-06, "loss": 0.1453, "step": 1132 }, { "epoch": 0.39285714285714285, "grad_norm": 1.6116793155670166, "learning_rate": 1e-06, "loss": 0.1672, "step": 1133 }, { "epoch": 0.3932038834951456, "grad_norm": 1.9633543491363525, "learning_rate": 1e-06, "loss": 0.1529, "step": 1134 }, { "epoch": 0.3935506241331484, "grad_norm": 2.0340752601623535, "learning_rate": 1e-06, "loss": 0.1708, "step": 1135 }, { "epoch": 0.39389736477115117, "grad_norm": 2.2635862827301025, "learning_rate": 1e-06, "loss": 0.1787, "step": 1136 }, { "epoch": 0.39424410540915394, "grad_norm": 2.268153667449951, "learning_rate": 1e-06, "loss": 0.1992, "step": 1137 }, { "epoch": 0.3945908460471567, "grad_norm": 1.7230418920516968, "learning_rate": 1e-06, "loss": 0.1372, "step": 1138 }, { "epoch": 0.3949375866851595, "grad_norm": 1.6801517009735107, "learning_rate": 1e-06, "loss": 0.1428, "step": 1139 }, { "epoch": 0.39528432732316227, "grad_norm": 2.6721580028533936, "learning_rate": 1e-06, "loss": 0.1597, "step": 1140 }, { "epoch": 0.39563106796116504, "grad_norm": 2.748892307281494, "learning_rate": 1e-06, "loss": 0.1554, "step": 1141 }, { "epoch": 0.3959778085991678, "grad_norm": 2.1693079471588135, "learning_rate": 1e-06, "loss": 0.1856, "step": 1142 }, { "epoch": 0.3963245492371706, "grad_norm": 3.472423553466797, "learning_rate": 1e-06, "loss": 0.1748, "step": 1143 }, { "epoch": 0.39667128987517336, "grad_norm": 3.2652645111083984, "learning_rate": 1e-06, "loss": 0.1756, "step": 1144 }, { "epoch": 0.39701803051317613, "grad_norm": 1.8527421951293945, "learning_rate": 1e-06, "loss": 0.1743, "step": 1145 }, { "epoch": 0.3973647711511789, "grad_norm": 1.7605152130126953, "learning_rate": 1e-06, "loss": 0.1524, "step": 1146 }, { "epoch": 0.3977115117891817, "grad_norm": 1.6430308818817139, "learning_rate": 1e-06, "loss": 0.1571, "step": 1147 }, { "epoch": 0.39805825242718446, "grad_norm": 1.800592064857483, "learning_rate": 1e-06, "loss": 0.1962, "step": 1148 }, { "epoch": 0.39840499306518723, "grad_norm": 1.925662875175476, "learning_rate": 1e-06, "loss": 0.1923, "step": 1149 }, { "epoch": 0.39875173370319, "grad_norm": 2.714669942855835, "learning_rate": 1e-06, "loss": 0.1562, "step": 1150 }, { "epoch": 0.3990984743411928, "grad_norm": 1.951225996017456, "learning_rate": 1e-06, "loss": 0.16, "step": 1151 }, { "epoch": 0.39944521497919555, "grad_norm": 3.028454542160034, "learning_rate": 1e-06, "loss": 0.1745, "step": 1152 }, { "epoch": 0.3997919556171983, "grad_norm": 5.070191860198975, "learning_rate": 1e-06, "loss": 0.1875, "step": 1153 }, { "epoch": 0.4001386962552011, "grad_norm": 1.9278351068496704, "learning_rate": 1e-06, "loss": 0.1608, "step": 1154 }, { "epoch": 0.40048543689320387, "grad_norm": 2.457918643951416, "learning_rate": 1e-06, "loss": 0.2081, "step": 1155 }, { "epoch": 0.40083217753120665, "grad_norm": 2.267364501953125, "learning_rate": 1e-06, "loss": 0.147, "step": 1156 }, { "epoch": 0.4011789181692094, "grad_norm": 2.4611542224884033, "learning_rate": 1e-06, "loss": 0.1763, "step": 1157 }, { "epoch": 0.4015256588072122, "grad_norm": 1.9886502027511597, "learning_rate": 1e-06, "loss": 0.183, "step": 1158 }, { "epoch": 0.40187239944521497, "grad_norm": 2.998033285140991, "learning_rate": 1e-06, "loss": 0.1725, "step": 1159 }, { "epoch": 0.40221914008321774, "grad_norm": 5.4129319190979, "learning_rate": 1e-06, "loss": 0.2027, "step": 1160 }, { "epoch": 0.4025658807212205, "grad_norm": 3.6759653091430664, "learning_rate": 1e-06, "loss": 0.1682, "step": 1161 }, { "epoch": 0.4029126213592233, "grad_norm": 2.2087016105651855, "learning_rate": 1e-06, "loss": 0.1633, "step": 1162 }, { "epoch": 0.40325936199722606, "grad_norm": 3.113842725753784, "learning_rate": 1e-06, "loss": 0.1823, "step": 1163 }, { "epoch": 0.40360610263522884, "grad_norm": 1.9368499517440796, "learning_rate": 1e-06, "loss": 0.1677, "step": 1164 }, { "epoch": 0.4039528432732316, "grad_norm": 3.0826315879821777, "learning_rate": 1e-06, "loss": 0.1492, "step": 1165 }, { "epoch": 0.4042995839112344, "grad_norm": 4.97300386428833, "learning_rate": 1e-06, "loss": 0.1753, "step": 1166 }, { "epoch": 0.40464632454923716, "grad_norm": 4.992250442504883, "learning_rate": 1e-06, "loss": 0.1532, "step": 1167 }, { "epoch": 0.40499306518723993, "grad_norm": 7.608388423919678, "learning_rate": 1e-06, "loss": 0.1693, "step": 1168 }, { "epoch": 0.4053398058252427, "grad_norm": 2.3100359439849854, "learning_rate": 1e-06, "loss": 0.1581, "step": 1169 }, { "epoch": 0.4056865464632455, "grad_norm": 2.648092746734619, "learning_rate": 1e-06, "loss": 0.1489, "step": 1170 }, { "epoch": 0.40603328710124825, "grad_norm": 1.7646437883377075, "learning_rate": 1e-06, "loss": 0.1578, "step": 1171 }, { "epoch": 0.406380027739251, "grad_norm": 2.959559679031372, "learning_rate": 1e-06, "loss": 0.1574, "step": 1172 }, { "epoch": 0.4067267683772538, "grad_norm": 2.033961057662964, "learning_rate": 1e-06, "loss": 0.1493, "step": 1173 }, { "epoch": 0.4070735090152566, "grad_norm": 1.6733283996582031, "learning_rate": 1e-06, "loss": 0.1473, "step": 1174 }, { "epoch": 0.40742024965325935, "grad_norm": 2.8953778743743896, "learning_rate": 1e-06, "loss": 0.1538, "step": 1175 }, { "epoch": 0.4077669902912621, "grad_norm": 2.2271323204040527, "learning_rate": 1e-06, "loss": 0.1749, "step": 1176 }, { "epoch": 0.4081137309292649, "grad_norm": 1.6847202777862549, "learning_rate": 1e-06, "loss": 0.1325, "step": 1177 }, { "epoch": 0.40846047156726767, "grad_norm": 5.325228214263916, "learning_rate": 1e-06, "loss": 0.2158, "step": 1178 }, { "epoch": 0.40880721220527044, "grad_norm": 2.9201323986053467, "learning_rate": 1e-06, "loss": 0.2057, "step": 1179 }, { "epoch": 0.4091539528432732, "grad_norm": 3.783559560775757, "learning_rate": 1e-06, "loss": 0.1741, "step": 1180 }, { "epoch": 0.409500693481276, "grad_norm": 2.83846378326416, "learning_rate": 1e-06, "loss": 0.1893, "step": 1181 }, { "epoch": 0.40984743411927876, "grad_norm": 3.131998300552368, "learning_rate": 1e-06, "loss": 0.1753, "step": 1182 }, { "epoch": 0.41019417475728154, "grad_norm": 3.4768800735473633, "learning_rate": 1e-06, "loss": 0.1638, "step": 1183 }, { "epoch": 0.4105409153952843, "grad_norm": 3.8055405616760254, "learning_rate": 1e-06, "loss": 0.1453, "step": 1184 }, { "epoch": 0.4108876560332871, "grad_norm": 3.457226514816284, "learning_rate": 1e-06, "loss": 0.1789, "step": 1185 }, { "epoch": 0.41123439667128986, "grad_norm": 2.19962477684021, "learning_rate": 1e-06, "loss": 0.1832, "step": 1186 }, { "epoch": 0.41158113730929263, "grad_norm": 4.11622428894043, "learning_rate": 1e-06, "loss": 0.1351, "step": 1187 }, { "epoch": 0.4119278779472954, "grad_norm": 2.369816541671753, "learning_rate": 1e-06, "loss": 0.1943, "step": 1188 }, { "epoch": 0.4122746185852982, "grad_norm": 6.653234958648682, "learning_rate": 1e-06, "loss": 0.1635, "step": 1189 }, { "epoch": 0.41262135922330095, "grad_norm": 7.1791887283325195, "learning_rate": 1e-06, "loss": 0.1677, "step": 1190 }, { "epoch": 0.41296809986130373, "grad_norm": 3.2327146530151367, "learning_rate": 1e-06, "loss": 0.1546, "step": 1191 }, { "epoch": 0.4133148404993065, "grad_norm": 3.696610927581787, "learning_rate": 1e-06, "loss": 0.1586, "step": 1192 }, { "epoch": 0.4136615811373093, "grad_norm": 2.113956928253174, "learning_rate": 1e-06, "loss": 0.1764, "step": 1193 }, { "epoch": 0.41400832177531205, "grad_norm": 3.988048553466797, "learning_rate": 1e-06, "loss": 0.1883, "step": 1194 }, { "epoch": 0.4143550624133148, "grad_norm": 2.866166353225708, "learning_rate": 1e-06, "loss": 0.1633, "step": 1195 }, { "epoch": 0.4147018030513176, "grad_norm": 2.072659492492676, "learning_rate": 1e-06, "loss": 0.1714, "step": 1196 }, { "epoch": 0.41504854368932037, "grad_norm": 1.681881308555603, "learning_rate": 1e-06, "loss": 0.1443, "step": 1197 }, { "epoch": 0.41539528432732314, "grad_norm": 1.793905258178711, "learning_rate": 1e-06, "loss": 0.1258, "step": 1198 }, { "epoch": 0.4157420249653259, "grad_norm": 5.42755126953125, "learning_rate": 1e-06, "loss": 0.2034, "step": 1199 }, { "epoch": 0.4160887656033287, "grad_norm": 2.7761476039886475, "learning_rate": 1e-06, "loss": 0.1226, "step": 1200 }, { "epoch": 0.41643550624133147, "grad_norm": 3.929356575012207, "learning_rate": 1e-06, "loss": 0.1742, "step": 1201 }, { "epoch": 0.41678224687933424, "grad_norm": 1.8514819145202637, "learning_rate": 1e-06, "loss": 0.1527, "step": 1202 }, { "epoch": 0.417128987517337, "grad_norm": 1.546459436416626, "learning_rate": 1e-06, "loss": 0.1307, "step": 1203 }, { "epoch": 0.4174757281553398, "grad_norm": 2.12691330909729, "learning_rate": 1e-06, "loss": 0.2007, "step": 1204 }, { "epoch": 0.41782246879334256, "grad_norm": 5.694792747497559, "learning_rate": 1e-06, "loss": 0.1964, "step": 1205 }, { "epoch": 0.41816920943134533, "grad_norm": 5.02055549621582, "learning_rate": 1e-06, "loss": 0.1751, "step": 1206 }, { "epoch": 0.4185159500693481, "grad_norm": 2.231020927429199, "learning_rate": 1e-06, "loss": 0.1956, "step": 1207 }, { "epoch": 0.4188626907073509, "grad_norm": 1.8750131130218506, "learning_rate": 1e-06, "loss": 0.1278, "step": 1208 }, { "epoch": 0.41920943134535366, "grad_norm": 2.0554685592651367, "learning_rate": 1e-06, "loss": 0.1586, "step": 1209 }, { "epoch": 0.41955617198335643, "grad_norm": 1.9559741020202637, "learning_rate": 1e-06, "loss": 0.1896, "step": 1210 }, { "epoch": 0.4199029126213592, "grad_norm": 4.545790672302246, "learning_rate": 1e-06, "loss": 0.1508, "step": 1211 }, { "epoch": 0.420249653259362, "grad_norm": 4.00286340713501, "learning_rate": 1e-06, "loss": 0.142, "step": 1212 }, { "epoch": 0.42059639389736475, "grad_norm": 5.119708061218262, "learning_rate": 1e-06, "loss": 0.1608, "step": 1213 }, { "epoch": 0.4209431345353675, "grad_norm": 1.9498828649520874, "learning_rate": 1e-06, "loss": 0.1715, "step": 1214 }, { "epoch": 0.4212898751733703, "grad_norm": 3.187913417816162, "learning_rate": 1e-06, "loss": 0.1756, "step": 1215 }, { "epoch": 0.42163661581137307, "grad_norm": 2.6275475025177, "learning_rate": 1e-06, "loss": 0.1494, "step": 1216 }, { "epoch": 0.42198335644937585, "grad_norm": 4.876014232635498, "learning_rate": 1e-06, "loss": 0.2037, "step": 1217 }, { "epoch": 0.4223300970873786, "grad_norm": 1.6075856685638428, "learning_rate": 1e-06, "loss": 0.1337, "step": 1218 }, { "epoch": 0.4226768377253814, "grad_norm": 2.4928295612335205, "learning_rate": 1e-06, "loss": 0.1601, "step": 1219 }, { "epoch": 0.42302357836338417, "grad_norm": 2.0485427379608154, "learning_rate": 1e-06, "loss": 0.1533, "step": 1220 }, { "epoch": 0.42337031900138694, "grad_norm": 2.0547242164611816, "learning_rate": 1e-06, "loss": 0.1723, "step": 1221 }, { "epoch": 0.4237170596393897, "grad_norm": 1.4388387203216553, "learning_rate": 1e-06, "loss": 0.1404, "step": 1222 }, { "epoch": 0.4240638002773925, "grad_norm": 3.0887644290924072, "learning_rate": 1e-06, "loss": 0.185, "step": 1223 }, { "epoch": 0.42441054091539526, "grad_norm": 2.907358169555664, "learning_rate": 1e-06, "loss": 0.1759, "step": 1224 }, { "epoch": 0.42475728155339804, "grad_norm": 1.9200416803359985, "learning_rate": 1e-06, "loss": 0.1861, "step": 1225 }, { "epoch": 0.4251040221914008, "grad_norm": 3.701133966445923, "learning_rate": 1e-06, "loss": 0.1473, "step": 1226 }, { "epoch": 0.4254507628294036, "grad_norm": 1.728384256362915, "learning_rate": 1e-06, "loss": 0.1704, "step": 1227 }, { "epoch": 0.42579750346740636, "grad_norm": 2.6459665298461914, "learning_rate": 1e-06, "loss": 0.1521, "step": 1228 }, { "epoch": 0.42614424410540913, "grad_norm": 3.3006246089935303, "learning_rate": 1e-06, "loss": 0.1552, "step": 1229 }, { "epoch": 0.4264909847434119, "grad_norm": 1.900184154510498, "learning_rate": 1e-06, "loss": 0.1576, "step": 1230 }, { "epoch": 0.4268377253814147, "grad_norm": 1.7001935243606567, "learning_rate": 1e-06, "loss": 0.1413, "step": 1231 }, { "epoch": 0.42718446601941745, "grad_norm": 2.1855411529541016, "learning_rate": 1e-06, "loss": 0.1584, "step": 1232 }, { "epoch": 0.4275312066574202, "grad_norm": 1.9435559511184692, "learning_rate": 1e-06, "loss": 0.1849, "step": 1233 }, { "epoch": 0.427877947295423, "grad_norm": 1.7659322023391724, "learning_rate": 1e-06, "loss": 0.145, "step": 1234 }, { "epoch": 0.4282246879334258, "grad_norm": 2.6840810775756836, "learning_rate": 1e-06, "loss": 0.1511, "step": 1235 }, { "epoch": 0.42857142857142855, "grad_norm": 1.843663215637207, "learning_rate": 1e-06, "loss": 0.1653, "step": 1236 }, { "epoch": 0.4289181692094313, "grad_norm": 1.7395163774490356, "learning_rate": 1e-06, "loss": 0.167, "step": 1237 }, { "epoch": 0.4292649098474341, "grad_norm": 5.5851664543151855, "learning_rate": 1e-06, "loss": 0.1967, "step": 1238 }, { "epoch": 0.42961165048543687, "grad_norm": 3.6831417083740234, "learning_rate": 1e-06, "loss": 0.167, "step": 1239 }, { "epoch": 0.42995839112343964, "grad_norm": 2.4813308715820312, "learning_rate": 1e-06, "loss": 0.1743, "step": 1240 }, { "epoch": 0.4303051317614424, "grad_norm": 1.4829434156417847, "learning_rate": 1e-06, "loss": 0.1298, "step": 1241 }, { "epoch": 0.4306518723994452, "grad_norm": 6.014243125915527, "learning_rate": 1e-06, "loss": 0.1911, "step": 1242 }, { "epoch": 0.43099861303744796, "grad_norm": 3.8425610065460205, "learning_rate": 1e-06, "loss": 0.1447, "step": 1243 }, { "epoch": 0.43134535367545074, "grad_norm": 2.8617472648620605, "learning_rate": 1e-06, "loss": 0.1752, "step": 1244 }, { "epoch": 0.4316920943134535, "grad_norm": 1.48086678981781, "learning_rate": 1e-06, "loss": 0.1398, "step": 1245 }, { "epoch": 0.4320388349514563, "grad_norm": 1.9533336162567139, "learning_rate": 1e-06, "loss": 0.1919, "step": 1246 }, { "epoch": 0.43238557558945906, "grad_norm": 2.6993250846862793, "learning_rate": 1e-06, "loss": 0.1761, "step": 1247 }, { "epoch": 0.43273231622746183, "grad_norm": 2.825752019882202, "learning_rate": 1e-06, "loss": 0.1489, "step": 1248 }, { "epoch": 0.4330790568654646, "grad_norm": 2.8645107746124268, "learning_rate": 1e-06, "loss": 0.1537, "step": 1249 }, { "epoch": 0.4334257975034674, "grad_norm": 1.9935320615768433, "learning_rate": 1e-06, "loss": 0.1725, "step": 1250 }, { "epoch": 0.43377253814147015, "grad_norm": 3.51171612739563, "learning_rate": 1e-06, "loss": 0.1773, "step": 1251 }, { "epoch": 0.43411927877947293, "grad_norm": 1.9769134521484375, "learning_rate": 1e-06, "loss": 0.1814, "step": 1252 }, { "epoch": 0.4344660194174757, "grad_norm": 2.077732801437378, "learning_rate": 1e-06, "loss": 0.1867, "step": 1253 }, { "epoch": 0.4348127600554785, "grad_norm": 1.734952688217163, "learning_rate": 1e-06, "loss": 0.1546, "step": 1254 }, { "epoch": 0.43515950069348125, "grad_norm": 1.789803385734558, "learning_rate": 1e-06, "loss": 0.1589, "step": 1255 }, { "epoch": 0.435506241331484, "grad_norm": 2.2564432621002197, "learning_rate": 1e-06, "loss": 0.2033, "step": 1256 }, { "epoch": 0.4358529819694868, "grad_norm": 2.3573691844940186, "learning_rate": 1e-06, "loss": 0.2015, "step": 1257 }, { "epoch": 0.43619972260748957, "grad_norm": 1.9252046346664429, "learning_rate": 1e-06, "loss": 0.1751, "step": 1258 }, { "epoch": 0.43654646324549234, "grad_norm": 1.6210649013519287, "learning_rate": 1e-06, "loss": 0.1694, "step": 1259 }, { "epoch": 0.4368932038834951, "grad_norm": 2.1824262142181396, "learning_rate": 1e-06, "loss": 0.1634, "step": 1260 }, { "epoch": 0.4372399445214979, "grad_norm": 1.7464162111282349, "learning_rate": 1e-06, "loss": 0.1703, "step": 1261 }, { "epoch": 0.4375866851595007, "grad_norm": 2.9391233921051025, "learning_rate": 1e-06, "loss": 0.212, "step": 1262 }, { "epoch": 0.4379334257975035, "grad_norm": 5.7660722732543945, "learning_rate": 1e-06, "loss": 0.1808, "step": 1263 }, { "epoch": 0.43828016643550627, "grad_norm": 1.7276206016540527, "learning_rate": 1e-06, "loss": 0.1597, "step": 1264 }, { "epoch": 0.43862690707350904, "grad_norm": 1.5537587404251099, "learning_rate": 1e-06, "loss": 0.1627, "step": 1265 }, { "epoch": 0.4389736477115118, "grad_norm": 2.254763126373291, "learning_rate": 1e-06, "loss": 0.163, "step": 1266 }, { "epoch": 0.4393203883495146, "grad_norm": 1.903600811958313, "learning_rate": 1e-06, "loss": 0.1533, "step": 1267 }, { "epoch": 0.43966712898751736, "grad_norm": 2.922935962677002, "learning_rate": 1e-06, "loss": 0.1979, "step": 1268 }, { "epoch": 0.44001386962552014, "grad_norm": 4.926792621612549, "learning_rate": 1e-06, "loss": 0.1434, "step": 1269 }, { "epoch": 0.4403606102635229, "grad_norm": 2.8258774280548096, "learning_rate": 1e-06, "loss": 0.1578, "step": 1270 }, { "epoch": 0.4407073509015257, "grad_norm": 3.5319342613220215, "learning_rate": 1e-06, "loss": 0.1085, "step": 1271 }, { "epoch": 0.44105409153952846, "grad_norm": 1.9517085552215576, "learning_rate": 1e-06, "loss": 0.1877, "step": 1272 }, { "epoch": 0.44140083217753123, "grad_norm": 1.7757554054260254, "learning_rate": 1e-06, "loss": 0.1774, "step": 1273 }, { "epoch": 0.441747572815534, "grad_norm": 2.927002191543579, "learning_rate": 1e-06, "loss": 0.1516, "step": 1274 }, { "epoch": 0.4420943134535368, "grad_norm": 2.922168493270874, "learning_rate": 1e-06, "loss": 0.1405, "step": 1275 }, { "epoch": 0.44244105409153955, "grad_norm": 1.9187227487564087, "learning_rate": 1e-06, "loss": 0.1424, "step": 1276 }, { "epoch": 0.44278779472954233, "grad_norm": 3.046382427215576, "learning_rate": 1e-06, "loss": 0.1587, "step": 1277 }, { "epoch": 0.4431345353675451, "grad_norm": 2.2099106311798096, "learning_rate": 1e-06, "loss": 0.1518, "step": 1278 }, { "epoch": 0.4434812760055479, "grad_norm": 3.655850887298584, "learning_rate": 1e-06, "loss": 0.1733, "step": 1279 }, { "epoch": 0.44382801664355065, "grad_norm": 3.842125415802002, "learning_rate": 1e-06, "loss": 0.1927, "step": 1280 }, { "epoch": 0.4441747572815534, "grad_norm": 1.9096598625183105, "learning_rate": 1e-06, "loss": 0.128, "step": 1281 }, { "epoch": 0.4445214979195562, "grad_norm": 3.1731748580932617, "learning_rate": 1e-06, "loss": 0.1563, "step": 1282 }, { "epoch": 0.44486823855755897, "grad_norm": 2.3779473304748535, "learning_rate": 1e-06, "loss": 0.171, "step": 1283 }, { "epoch": 0.44521497919556174, "grad_norm": 4.150517463684082, "learning_rate": 1e-06, "loss": 0.1452, "step": 1284 }, { "epoch": 0.4455617198335645, "grad_norm": 3.774120807647705, "learning_rate": 1e-06, "loss": 0.1926, "step": 1285 }, { "epoch": 0.4459084604715673, "grad_norm": 3.833528995513916, "learning_rate": 1e-06, "loss": 0.1771, "step": 1286 }, { "epoch": 0.44625520110957007, "grad_norm": 3.3106303215026855, "learning_rate": 1e-06, "loss": 0.1918, "step": 1287 }, { "epoch": 0.44660194174757284, "grad_norm": 3.2806928157806396, "learning_rate": 1e-06, "loss": 0.147, "step": 1288 }, { "epoch": 0.4469486823855756, "grad_norm": 2.7862508296966553, "learning_rate": 1e-06, "loss": 0.1768, "step": 1289 }, { "epoch": 0.4472954230235784, "grad_norm": 2.7099242210388184, "learning_rate": 1e-06, "loss": 0.2095, "step": 1290 }, { "epoch": 0.44764216366158116, "grad_norm": 3.24961256980896, "learning_rate": 1e-06, "loss": 0.171, "step": 1291 }, { "epoch": 0.44798890429958393, "grad_norm": 3.220797061920166, "learning_rate": 1e-06, "loss": 0.1641, "step": 1292 }, { "epoch": 0.4483356449375867, "grad_norm": 4.3876237869262695, "learning_rate": 1e-06, "loss": 0.1526, "step": 1293 }, { "epoch": 0.4486823855755895, "grad_norm": 2.9639668464660645, "learning_rate": 1e-06, "loss": 0.1649, "step": 1294 }, { "epoch": 0.44902912621359226, "grad_norm": 2.0333526134490967, "learning_rate": 1e-06, "loss": 0.1648, "step": 1295 }, { "epoch": 0.44937586685159503, "grad_norm": 3.781501531600952, "learning_rate": 1e-06, "loss": 0.1571, "step": 1296 }, { "epoch": 0.4497226074895978, "grad_norm": 1.5544551610946655, "learning_rate": 1e-06, "loss": 0.1496, "step": 1297 }, { "epoch": 0.4500693481276006, "grad_norm": 5.4716596603393555, "learning_rate": 1e-06, "loss": 0.199, "step": 1298 }, { "epoch": 0.45041608876560335, "grad_norm": 1.998415470123291, "learning_rate": 1e-06, "loss": 0.1236, "step": 1299 }, { "epoch": 0.4507628294036061, "grad_norm": 5.753319263458252, "learning_rate": 1e-06, "loss": 0.2119, "step": 1300 }, { "epoch": 0.4511095700416089, "grad_norm": 3.596370220184326, "learning_rate": 1e-06, "loss": 0.1901, "step": 1301 }, { "epoch": 0.45145631067961167, "grad_norm": 3.2227611541748047, "learning_rate": 1e-06, "loss": 0.1738, "step": 1302 }, { "epoch": 0.45180305131761445, "grad_norm": 3.6115000247955322, "learning_rate": 1e-06, "loss": 0.2224, "step": 1303 }, { "epoch": 0.4521497919556172, "grad_norm": 3.304997682571411, "learning_rate": 1e-06, "loss": 0.1662, "step": 1304 }, { "epoch": 0.45249653259362, "grad_norm": 2.065157175064087, "learning_rate": 1e-06, "loss": 0.1958, "step": 1305 }, { "epoch": 0.45284327323162277, "grad_norm": 3.3705663681030273, "learning_rate": 1e-06, "loss": 0.133, "step": 1306 }, { "epoch": 0.45319001386962554, "grad_norm": 3.053819417953491, "learning_rate": 1e-06, "loss": 0.1409, "step": 1307 }, { "epoch": 0.4535367545076283, "grad_norm": 2.586345911026001, "learning_rate": 1e-06, "loss": 0.1698, "step": 1308 }, { "epoch": 0.4538834951456311, "grad_norm": 1.9915062189102173, "learning_rate": 1e-06, "loss": 0.1884, "step": 1309 }, { "epoch": 0.45423023578363386, "grad_norm": 2.1440911293029785, "learning_rate": 1e-06, "loss": 0.1587, "step": 1310 }, { "epoch": 0.45457697642163664, "grad_norm": 2.2233524322509766, "learning_rate": 1e-06, "loss": 0.1383, "step": 1311 }, { "epoch": 0.4549237170596394, "grad_norm": 1.7290481328964233, "learning_rate": 1e-06, "loss": 0.1783, "step": 1312 }, { "epoch": 0.4552704576976422, "grad_norm": 1.7241450548171997, "learning_rate": 1e-06, "loss": 0.1505, "step": 1313 }, { "epoch": 0.45561719833564496, "grad_norm": 2.3768575191497803, "learning_rate": 1e-06, "loss": 0.1672, "step": 1314 }, { "epoch": 0.45596393897364773, "grad_norm": 2.1226019859313965, "learning_rate": 1e-06, "loss": 0.1497, "step": 1315 }, { "epoch": 0.4563106796116505, "grad_norm": 1.9038270711898804, "learning_rate": 1e-06, "loss": 0.1664, "step": 1316 }, { "epoch": 0.4566574202496533, "grad_norm": 1.9552680253982544, "learning_rate": 1e-06, "loss": 0.1473, "step": 1317 }, { "epoch": 0.45700416088765605, "grad_norm": 1.8354042768478394, "learning_rate": 1e-06, "loss": 0.1628, "step": 1318 }, { "epoch": 0.4573509015256588, "grad_norm": 2.84702467918396, "learning_rate": 1e-06, "loss": 0.2017, "step": 1319 }, { "epoch": 0.4576976421636616, "grad_norm": 2.1125497817993164, "learning_rate": 1e-06, "loss": 0.1753, "step": 1320 }, { "epoch": 0.4580443828016644, "grad_norm": 1.6211894750595093, "learning_rate": 1e-06, "loss": 0.1473, "step": 1321 }, { "epoch": 0.45839112343966715, "grad_norm": 3.268822431564331, "learning_rate": 1e-06, "loss": 0.1463, "step": 1322 }, { "epoch": 0.4587378640776699, "grad_norm": 1.8956842422485352, "learning_rate": 1e-06, "loss": 0.1445, "step": 1323 }, { "epoch": 0.4590846047156727, "grad_norm": 3.552382230758667, "learning_rate": 1e-06, "loss": 0.1357, "step": 1324 }, { "epoch": 0.45943134535367547, "grad_norm": 1.8421423435211182, "learning_rate": 1e-06, "loss": 0.1479, "step": 1325 }, { "epoch": 0.45977808599167824, "grad_norm": 4.14149808883667, "learning_rate": 1e-06, "loss": 0.1504, "step": 1326 }, { "epoch": 0.460124826629681, "grad_norm": 1.7335238456726074, "learning_rate": 1e-06, "loss": 0.1548, "step": 1327 }, { "epoch": 0.4604715672676838, "grad_norm": 4.671210289001465, "learning_rate": 1e-06, "loss": 0.2263, "step": 1328 }, { "epoch": 0.46081830790568656, "grad_norm": 2.813931941986084, "learning_rate": 1e-06, "loss": 0.1668, "step": 1329 }, { "epoch": 0.46116504854368934, "grad_norm": 2.9509613513946533, "learning_rate": 1e-06, "loss": 0.1484, "step": 1330 }, { "epoch": 0.4615117891816921, "grad_norm": 2.949826240539551, "learning_rate": 1e-06, "loss": 0.1827, "step": 1331 }, { "epoch": 0.4618585298196949, "grad_norm": 1.8698536157608032, "learning_rate": 1e-06, "loss": 0.168, "step": 1332 }, { "epoch": 0.46220527045769766, "grad_norm": 3.174593210220337, "learning_rate": 1e-06, "loss": 0.2109, "step": 1333 }, { "epoch": 0.46255201109570043, "grad_norm": 1.977563738822937, "learning_rate": 1e-06, "loss": 0.1426, "step": 1334 }, { "epoch": 0.4628987517337032, "grad_norm": 1.8068509101867676, "learning_rate": 1e-06, "loss": 0.1172, "step": 1335 }, { "epoch": 0.463245492371706, "grad_norm": 2.076296329498291, "learning_rate": 1e-06, "loss": 0.1608, "step": 1336 }, { "epoch": 0.46359223300970875, "grad_norm": 2.3091671466827393, "learning_rate": 1e-06, "loss": 0.1665, "step": 1337 }, { "epoch": 0.46393897364771153, "grad_norm": 1.894162654876709, "learning_rate": 1e-06, "loss": 0.1522, "step": 1338 }, { "epoch": 0.4642857142857143, "grad_norm": 1.876305341720581, "learning_rate": 1e-06, "loss": 0.1444, "step": 1339 }, { "epoch": 0.4646324549237171, "grad_norm": 2.8168094158172607, "learning_rate": 1e-06, "loss": 0.1686, "step": 1340 }, { "epoch": 0.46497919556171985, "grad_norm": 2.988266706466675, "learning_rate": 1e-06, "loss": 0.1545, "step": 1341 }, { "epoch": 0.4653259361997226, "grad_norm": 1.7541871070861816, "learning_rate": 1e-06, "loss": 0.1423, "step": 1342 }, { "epoch": 0.4656726768377254, "grad_norm": 4.07562255859375, "learning_rate": 1e-06, "loss": 0.1447, "step": 1343 }, { "epoch": 0.46601941747572817, "grad_norm": 4.45770263671875, "learning_rate": 1e-06, "loss": 0.1639, "step": 1344 }, { "epoch": 0.46636615811373094, "grad_norm": 1.9045706987380981, "learning_rate": 1e-06, "loss": 0.1765, "step": 1345 }, { "epoch": 0.4667128987517337, "grad_norm": 1.927842378616333, "learning_rate": 1e-06, "loss": 0.1485, "step": 1346 }, { "epoch": 0.4670596393897365, "grad_norm": 7.443863868713379, "learning_rate": 1e-06, "loss": 0.1642, "step": 1347 }, { "epoch": 0.46740638002773927, "grad_norm": 2.3793625831604004, "learning_rate": 1e-06, "loss": 0.1906, "step": 1348 }, { "epoch": 0.46775312066574204, "grad_norm": 2.352898120880127, "learning_rate": 1e-06, "loss": 0.1796, "step": 1349 }, { "epoch": 0.4680998613037448, "grad_norm": 3.9866578578948975, "learning_rate": 1e-06, "loss": 0.1445, "step": 1350 }, { "epoch": 0.4684466019417476, "grad_norm": 1.6399085521697998, "learning_rate": 1e-06, "loss": 0.1484, "step": 1351 }, { "epoch": 0.46879334257975036, "grad_norm": 1.9889776706695557, "learning_rate": 1e-06, "loss": 0.174, "step": 1352 }, { "epoch": 0.46914008321775313, "grad_norm": 2.0575404167175293, "learning_rate": 1e-06, "loss": 0.1832, "step": 1353 }, { "epoch": 0.4694868238557559, "grad_norm": 2.2382092475891113, "learning_rate": 1e-06, "loss": 0.1551, "step": 1354 }, { "epoch": 0.4698335644937587, "grad_norm": 2.102724552154541, "learning_rate": 1e-06, "loss": 0.1566, "step": 1355 }, { "epoch": 0.47018030513176146, "grad_norm": 2.692265033721924, "learning_rate": 1e-06, "loss": 0.1921, "step": 1356 }, { "epoch": 0.47052704576976423, "grad_norm": 1.9280555248260498, "learning_rate": 1e-06, "loss": 0.1678, "step": 1357 }, { "epoch": 0.470873786407767, "grad_norm": 1.9627320766448975, "learning_rate": 1e-06, "loss": 0.1591, "step": 1358 }, { "epoch": 0.4712205270457698, "grad_norm": 1.7975282669067383, "learning_rate": 1e-06, "loss": 0.1312, "step": 1359 }, { "epoch": 0.47156726768377255, "grad_norm": 3.5527241230010986, "learning_rate": 1e-06, "loss": 0.1574, "step": 1360 }, { "epoch": 0.4719140083217753, "grad_norm": 3.669248104095459, "learning_rate": 1e-06, "loss": 0.2004, "step": 1361 }, { "epoch": 0.4722607489597781, "grad_norm": 1.9283638000488281, "learning_rate": 1e-06, "loss": 0.1776, "step": 1362 }, { "epoch": 0.47260748959778087, "grad_norm": 1.8212238550186157, "learning_rate": 1e-06, "loss": 0.1401, "step": 1363 }, { "epoch": 0.47295423023578365, "grad_norm": 3.3076553344726562, "learning_rate": 1e-06, "loss": 0.1734, "step": 1364 }, { "epoch": 0.4733009708737864, "grad_norm": 2.9192631244659424, "learning_rate": 1e-06, "loss": 0.1458, "step": 1365 }, { "epoch": 0.4736477115117892, "grad_norm": 5.290735244750977, "learning_rate": 1e-06, "loss": 0.18, "step": 1366 }, { "epoch": 0.47399445214979197, "grad_norm": 3.645962953567505, "learning_rate": 1e-06, "loss": 0.1608, "step": 1367 }, { "epoch": 0.47434119278779474, "grad_norm": 4.024303913116455, "learning_rate": 1e-06, "loss": 0.1925, "step": 1368 }, { "epoch": 0.4746879334257975, "grad_norm": 2.006380081176758, "learning_rate": 1e-06, "loss": 0.1215, "step": 1369 }, { "epoch": 0.4750346740638003, "grad_norm": 3.3509087562561035, "learning_rate": 1e-06, "loss": 0.1481, "step": 1370 }, { "epoch": 0.47538141470180306, "grad_norm": 1.8464128971099854, "learning_rate": 1e-06, "loss": 0.1433, "step": 1371 }, { "epoch": 0.47572815533980584, "grad_norm": 2.2263832092285156, "learning_rate": 1e-06, "loss": 0.1968, "step": 1372 }, { "epoch": 0.4760748959778086, "grad_norm": 2.3892805576324463, "learning_rate": 1e-06, "loss": 0.1736, "step": 1373 }, { "epoch": 0.4764216366158114, "grad_norm": 2.452542781829834, "learning_rate": 1e-06, "loss": 0.1808, "step": 1374 }, { "epoch": 0.47676837725381416, "grad_norm": 2.9188780784606934, "learning_rate": 1e-06, "loss": 0.1254, "step": 1375 }, { "epoch": 0.47711511789181693, "grad_norm": 1.651590347290039, "learning_rate": 1e-06, "loss": 0.1679, "step": 1376 }, { "epoch": 0.4774618585298197, "grad_norm": 1.9459658861160278, "learning_rate": 1e-06, "loss": 0.1622, "step": 1377 }, { "epoch": 0.4778085991678225, "grad_norm": 2.0248286724090576, "learning_rate": 1e-06, "loss": 0.1365, "step": 1378 }, { "epoch": 0.47815533980582525, "grad_norm": 4.144098281860352, "learning_rate": 1e-06, "loss": 0.1347, "step": 1379 }, { "epoch": 0.478502080443828, "grad_norm": 1.46219801902771, "learning_rate": 1e-06, "loss": 0.1468, "step": 1380 }, { "epoch": 0.4788488210818308, "grad_norm": 2.1017038822174072, "learning_rate": 1e-06, "loss": 0.1694, "step": 1381 }, { "epoch": 0.4791955617198336, "grad_norm": 2.1770832538604736, "learning_rate": 1e-06, "loss": 0.1829, "step": 1382 }, { "epoch": 0.47954230235783635, "grad_norm": 3.3750691413879395, "learning_rate": 1e-06, "loss": 0.1604, "step": 1383 }, { "epoch": 0.4798890429958391, "grad_norm": 2.4006078243255615, "learning_rate": 1e-06, "loss": 0.1437, "step": 1384 }, { "epoch": 0.4802357836338419, "grad_norm": 2.0251758098602295, "learning_rate": 1e-06, "loss": 0.163, "step": 1385 }, { "epoch": 0.48058252427184467, "grad_norm": 1.4575427770614624, "learning_rate": 1e-06, "loss": 0.144, "step": 1386 }, { "epoch": 0.48092926490984744, "grad_norm": 2.5154523849487305, "learning_rate": 1e-06, "loss": 0.1608, "step": 1387 }, { "epoch": 0.4812760055478502, "grad_norm": 2.151294231414795, "learning_rate": 1e-06, "loss": 0.1779, "step": 1388 }, { "epoch": 0.481622746185853, "grad_norm": 2.023602247238159, "learning_rate": 1e-06, "loss": 0.1693, "step": 1389 }, { "epoch": 0.48196948682385576, "grad_norm": 2.615605115890503, "learning_rate": 1e-06, "loss": 0.1406, "step": 1390 }, { "epoch": 0.48231622746185854, "grad_norm": 1.9372549057006836, "learning_rate": 1e-06, "loss": 0.1482, "step": 1391 }, { "epoch": 0.4826629680998613, "grad_norm": 2.0867562294006348, "learning_rate": 1e-06, "loss": 0.1745, "step": 1392 }, { "epoch": 0.4830097087378641, "grad_norm": 2.231527090072632, "learning_rate": 1e-06, "loss": 0.1585, "step": 1393 }, { "epoch": 0.48335644937586686, "grad_norm": 1.7819068431854248, "learning_rate": 1e-06, "loss": 0.1544, "step": 1394 }, { "epoch": 0.48370319001386963, "grad_norm": 2.0249552726745605, "learning_rate": 1e-06, "loss": 0.1983, "step": 1395 }, { "epoch": 0.4840499306518724, "grad_norm": 2.4923202991485596, "learning_rate": 1e-06, "loss": 0.1355, "step": 1396 }, { "epoch": 0.4843966712898752, "grad_norm": 1.6963531970977783, "learning_rate": 1e-06, "loss": 0.1464, "step": 1397 }, { "epoch": 0.48474341192787795, "grad_norm": 2.0951385498046875, "learning_rate": 1e-06, "loss": 0.1533, "step": 1398 }, { "epoch": 0.48509015256588073, "grad_norm": 3.5178959369659424, "learning_rate": 1e-06, "loss": 0.1273, "step": 1399 }, { "epoch": 0.4854368932038835, "grad_norm": 2.0446174144744873, "learning_rate": 1e-06, "loss": 0.1815, "step": 1400 }, { "epoch": 0.4857836338418863, "grad_norm": 4.695765972137451, "learning_rate": 1e-06, "loss": 0.1558, "step": 1401 }, { "epoch": 0.48613037447988905, "grad_norm": 2.182650089263916, "learning_rate": 1e-06, "loss": 0.1541, "step": 1402 }, { "epoch": 0.4864771151178918, "grad_norm": 2.1337432861328125, "learning_rate": 1e-06, "loss": 0.157, "step": 1403 }, { "epoch": 0.4868238557558946, "grad_norm": 2.5917625427246094, "learning_rate": 1e-06, "loss": 0.1789, "step": 1404 }, { "epoch": 0.48717059639389737, "grad_norm": 1.9805132150650024, "learning_rate": 1e-06, "loss": 0.1449, "step": 1405 }, { "epoch": 0.48751733703190014, "grad_norm": 2.243947982788086, "learning_rate": 1e-06, "loss": 0.1448, "step": 1406 }, { "epoch": 0.4878640776699029, "grad_norm": 1.755910873413086, "learning_rate": 1e-06, "loss": 0.1519, "step": 1407 }, { "epoch": 0.4882108183079057, "grad_norm": 1.5781508684158325, "learning_rate": 1e-06, "loss": 0.1566, "step": 1408 }, { "epoch": 0.48855755894590847, "grad_norm": 1.7904160022735596, "learning_rate": 1e-06, "loss": 0.1246, "step": 1409 }, { "epoch": 0.48890429958391124, "grad_norm": 3.9564223289489746, "learning_rate": 1e-06, "loss": 0.1654, "step": 1410 }, { "epoch": 0.489251040221914, "grad_norm": 1.6851087808609009, "learning_rate": 1e-06, "loss": 0.1368, "step": 1411 }, { "epoch": 0.4895977808599168, "grad_norm": 3.7889657020568848, "learning_rate": 1e-06, "loss": 0.1632, "step": 1412 }, { "epoch": 0.48994452149791956, "grad_norm": 4.374680995941162, "learning_rate": 1e-06, "loss": 0.1564, "step": 1413 }, { "epoch": 0.49029126213592233, "grad_norm": 3.0737457275390625, "learning_rate": 1e-06, "loss": 0.1501, "step": 1414 }, { "epoch": 0.4906380027739251, "grad_norm": 2.6151061058044434, "learning_rate": 1e-06, "loss": 0.1595, "step": 1415 }, { "epoch": 0.4909847434119279, "grad_norm": 2.651824951171875, "learning_rate": 1e-06, "loss": 0.1469, "step": 1416 }, { "epoch": 0.49133148404993066, "grad_norm": 2.8488190174102783, "learning_rate": 1e-06, "loss": 0.1519, "step": 1417 }, { "epoch": 0.49167822468793343, "grad_norm": 2.576077699661255, "learning_rate": 1e-06, "loss": 0.153, "step": 1418 }, { "epoch": 0.4920249653259362, "grad_norm": 1.594548225402832, "learning_rate": 1e-06, "loss": 0.1542, "step": 1419 }, { "epoch": 0.492371705963939, "grad_norm": 2.900624990463257, "learning_rate": 1e-06, "loss": 0.1542, "step": 1420 }, { "epoch": 0.49271844660194175, "grad_norm": 3.2703654766082764, "learning_rate": 1e-06, "loss": 0.1285, "step": 1421 }, { "epoch": 0.4930651872399445, "grad_norm": 1.6843549013137817, "learning_rate": 1e-06, "loss": 0.1513, "step": 1422 }, { "epoch": 0.4934119278779473, "grad_norm": 1.9763399362564087, "learning_rate": 1e-06, "loss": 0.1504, "step": 1423 }, { "epoch": 0.49375866851595007, "grad_norm": 1.569488763809204, "learning_rate": 1e-06, "loss": 0.1452, "step": 1424 }, { "epoch": 0.49410540915395285, "grad_norm": 2.2819178104400635, "learning_rate": 1e-06, "loss": 0.1346, "step": 1425 }, { "epoch": 0.4944521497919556, "grad_norm": 1.7116472721099854, "learning_rate": 1e-06, "loss": 0.1288, "step": 1426 }, { "epoch": 0.4947988904299584, "grad_norm": 2.4184670448303223, "learning_rate": 1e-06, "loss": 0.1365, "step": 1427 }, { "epoch": 0.49514563106796117, "grad_norm": 2.8264670372009277, "learning_rate": 1e-06, "loss": 0.1521, "step": 1428 }, { "epoch": 0.49549237170596394, "grad_norm": 1.987285852432251, "learning_rate": 1e-06, "loss": 0.1617, "step": 1429 }, { "epoch": 0.4958391123439667, "grad_norm": 2.5394463539123535, "learning_rate": 1e-06, "loss": 0.1418, "step": 1430 }, { "epoch": 0.4961858529819695, "grad_norm": 2.5165560245513916, "learning_rate": 1e-06, "loss": 0.1605, "step": 1431 }, { "epoch": 0.49653259361997226, "grad_norm": 2.378039836883545, "learning_rate": 1e-06, "loss": 0.1866, "step": 1432 }, { "epoch": 0.49687933425797504, "grad_norm": 2.0523364543914795, "learning_rate": 1e-06, "loss": 0.1298, "step": 1433 }, { "epoch": 0.4972260748959778, "grad_norm": 3.141749382019043, "learning_rate": 1e-06, "loss": 0.1296, "step": 1434 }, { "epoch": 0.4975728155339806, "grad_norm": 2.1089577674865723, "learning_rate": 1e-06, "loss": 0.135, "step": 1435 }, { "epoch": 0.49791955617198336, "grad_norm": 1.6610194444656372, "learning_rate": 1e-06, "loss": 0.1194, "step": 1436 }, { "epoch": 0.49826629680998613, "grad_norm": 3.3861851692199707, "learning_rate": 1e-06, "loss": 0.1649, "step": 1437 }, { "epoch": 0.4986130374479889, "grad_norm": 2.972479820251465, "learning_rate": 1e-06, "loss": 0.1565, "step": 1438 }, { "epoch": 0.4989597780859917, "grad_norm": 4.443920612335205, "learning_rate": 1e-06, "loss": 0.1518, "step": 1439 }, { "epoch": 0.49930651872399445, "grad_norm": 3.384650707244873, "learning_rate": 1e-06, "loss": 0.1399, "step": 1440 }, { "epoch": 0.4996532593619972, "grad_norm": 2.9029901027679443, "learning_rate": 1e-06, "loss": 0.1523, "step": 1441 }, { "epoch": 0.5, "grad_norm": 1.911129117012024, "learning_rate": 1e-06, "loss": 0.1748, "step": 1442 }, { "epoch": 0.5003467406380028, "grad_norm": 3.056776285171509, "learning_rate": 1e-06, "loss": 0.1394, "step": 1443 }, { "epoch": 0.5006934812760055, "grad_norm": 2.821357011795044, "learning_rate": 1e-06, "loss": 0.1946, "step": 1444 }, { "epoch": 0.5010402219140083, "grad_norm": 3.346324920654297, "learning_rate": 1e-06, "loss": 0.1793, "step": 1445 }, { "epoch": 0.5013869625520111, "grad_norm": 3.7908129692077637, "learning_rate": 1e-06, "loss": 0.1635, "step": 1446 }, { "epoch": 0.5017337031900139, "grad_norm": 1.877949595451355, "learning_rate": 1e-06, "loss": 0.1678, "step": 1447 }, { "epoch": 0.5020804438280166, "grad_norm": 1.9401321411132812, "learning_rate": 1e-06, "loss": 0.1575, "step": 1448 }, { "epoch": 0.5024271844660194, "grad_norm": 1.9344282150268555, "learning_rate": 1e-06, "loss": 0.1646, "step": 1449 }, { "epoch": 0.5027739251040222, "grad_norm": 2.282867193222046, "learning_rate": 1e-06, "loss": 0.1888, "step": 1450 }, { "epoch": 0.503120665742025, "grad_norm": 3.797800064086914, "learning_rate": 1e-06, "loss": 0.1256, "step": 1451 }, { "epoch": 0.5034674063800277, "grad_norm": 3.1414899826049805, "learning_rate": 1e-06, "loss": 0.1426, "step": 1452 }, { "epoch": 0.5038141470180305, "grad_norm": 1.9722864627838135, "learning_rate": 1e-06, "loss": 0.1408, "step": 1453 }, { "epoch": 0.5041608876560333, "grad_norm": 1.5774776935577393, "learning_rate": 1e-06, "loss": 0.1137, "step": 1454 }, { "epoch": 0.5045076282940361, "grad_norm": 3.0094971656799316, "learning_rate": 1e-06, "loss": 0.1554, "step": 1455 }, { "epoch": 0.5048543689320388, "grad_norm": 3.39162015914917, "learning_rate": 1e-06, "loss": 0.1611, "step": 1456 }, { "epoch": 0.5052011095700416, "grad_norm": 2.489313840866089, "learning_rate": 1e-06, "loss": 0.1535, "step": 1457 }, { "epoch": 0.5055478502080444, "grad_norm": 4.133990287780762, "learning_rate": 1e-06, "loss": 0.1775, "step": 1458 }, { "epoch": 0.5058945908460472, "grad_norm": 2.7449488639831543, "learning_rate": 1e-06, "loss": 0.1495, "step": 1459 }, { "epoch": 0.5062413314840499, "grad_norm": 4.382802486419678, "learning_rate": 1e-06, "loss": 0.133, "step": 1460 }, { "epoch": 0.5065880721220527, "grad_norm": 2.110130786895752, "learning_rate": 1e-06, "loss": 0.1564, "step": 1461 }, { "epoch": 0.5069348127600555, "grad_norm": 2.527834415435791, "learning_rate": 1e-06, "loss": 0.1757, "step": 1462 }, { "epoch": 0.5072815533980582, "grad_norm": 2.9167022705078125, "learning_rate": 1e-06, "loss": 0.1686, "step": 1463 }, { "epoch": 0.507628294036061, "grad_norm": 2.2591302394866943, "learning_rate": 1e-06, "loss": 0.1355, "step": 1464 }, { "epoch": 0.5079750346740638, "grad_norm": 4.148419380187988, "learning_rate": 1e-06, "loss": 0.1391, "step": 1465 }, { "epoch": 0.5083217753120666, "grad_norm": 4.044282913208008, "learning_rate": 1e-06, "loss": 0.1642, "step": 1466 }, { "epoch": 0.5086685159500693, "grad_norm": 1.8170924186706543, "learning_rate": 1e-06, "loss": 0.1218, "step": 1467 }, { "epoch": 0.5090152565880721, "grad_norm": 2.241746664047241, "learning_rate": 1e-06, "loss": 0.1643, "step": 1468 }, { "epoch": 0.5093619972260749, "grad_norm": 3.467628240585327, "learning_rate": 1e-06, "loss": 0.1355, "step": 1469 }, { "epoch": 0.5097087378640777, "grad_norm": 3.2897260189056396, "learning_rate": 1e-06, "loss": 0.145, "step": 1470 }, { "epoch": 0.5100554785020804, "grad_norm": 1.7362409830093384, "learning_rate": 1e-06, "loss": 0.1376, "step": 1471 }, { "epoch": 0.5104022191400832, "grad_norm": 2.2274043560028076, "learning_rate": 1e-06, "loss": 0.1581, "step": 1472 }, { "epoch": 0.510748959778086, "grad_norm": 1.8008100986480713, "learning_rate": 1e-06, "loss": 0.1359, "step": 1473 }, { "epoch": 0.5110957004160888, "grad_norm": 1.5842705965042114, "learning_rate": 1e-06, "loss": 0.1264, "step": 1474 }, { "epoch": 0.5114424410540915, "grad_norm": 3.425924301147461, "learning_rate": 1e-06, "loss": 0.1281, "step": 1475 }, { "epoch": 0.5117891816920943, "grad_norm": 2.418201208114624, "learning_rate": 1e-06, "loss": 0.1695, "step": 1476 }, { "epoch": 0.5121359223300971, "grad_norm": 2.0753753185272217, "learning_rate": 1e-06, "loss": 0.1415, "step": 1477 }, { "epoch": 0.5124826629680999, "grad_norm": 2.378450632095337, "learning_rate": 1e-06, "loss": 0.2064, "step": 1478 }, { "epoch": 0.5128294036061026, "grad_norm": 2.1089658737182617, "learning_rate": 1e-06, "loss": 0.1471, "step": 1479 }, { "epoch": 0.5131761442441054, "grad_norm": 2.442574977874756, "learning_rate": 1e-06, "loss": 0.175, "step": 1480 }, { "epoch": 0.5135228848821082, "grad_norm": 2.503943681716919, "learning_rate": 1e-06, "loss": 0.1389, "step": 1481 }, { "epoch": 0.513869625520111, "grad_norm": 1.4603307247161865, "learning_rate": 1e-06, "loss": 0.1318, "step": 1482 }, { "epoch": 0.5142163661581137, "grad_norm": 2.270416498184204, "learning_rate": 1e-06, "loss": 0.1572, "step": 1483 }, { "epoch": 0.5145631067961165, "grad_norm": 2.438401699066162, "learning_rate": 1e-06, "loss": 0.1337, "step": 1484 }, { "epoch": 0.5149098474341193, "grad_norm": 2.581489086151123, "learning_rate": 1e-06, "loss": 0.1554, "step": 1485 }, { "epoch": 0.515256588072122, "grad_norm": 2.019855260848999, "learning_rate": 1e-06, "loss": 0.1386, "step": 1486 }, { "epoch": 0.5156033287101248, "grad_norm": 4.10310173034668, "learning_rate": 1e-06, "loss": 0.1458, "step": 1487 }, { "epoch": 0.5159500693481276, "grad_norm": 1.7705767154693604, "learning_rate": 1e-06, "loss": 0.1464, "step": 1488 }, { "epoch": 0.5162968099861304, "grad_norm": 2.2230114936828613, "learning_rate": 1e-06, "loss": 0.1376, "step": 1489 }, { "epoch": 0.5166435506241331, "grad_norm": 6.302447319030762, "learning_rate": 1e-06, "loss": 0.1822, "step": 1490 }, { "epoch": 0.5169902912621359, "grad_norm": 3.1601343154907227, "learning_rate": 1e-06, "loss": 0.1539, "step": 1491 }, { "epoch": 0.5173370319001387, "grad_norm": 2.8884189128875732, "learning_rate": 1e-06, "loss": 0.1727, "step": 1492 }, { "epoch": 0.5176837725381415, "grad_norm": 2.00549578666687, "learning_rate": 1e-06, "loss": 0.135, "step": 1493 }, { "epoch": 0.5180305131761442, "grad_norm": 3.5855584144592285, "learning_rate": 1e-06, "loss": 0.2021, "step": 1494 }, { "epoch": 0.518377253814147, "grad_norm": 4.1129069328308105, "learning_rate": 1e-06, "loss": 0.1407, "step": 1495 }, { "epoch": 0.5187239944521498, "grad_norm": 2.877685546875, "learning_rate": 1e-06, "loss": 0.195, "step": 1496 }, { "epoch": 0.5190707350901526, "grad_norm": 2.007178544998169, "learning_rate": 1e-06, "loss": 0.1874, "step": 1497 }, { "epoch": 0.5194174757281553, "grad_norm": 1.9459450244903564, "learning_rate": 1e-06, "loss": 0.1465, "step": 1498 }, { "epoch": 0.5197642163661581, "grad_norm": 3.181014060974121, "learning_rate": 1e-06, "loss": 0.1346, "step": 1499 }, { "epoch": 0.5201109570041609, "grad_norm": 2.48248553276062, "learning_rate": 1e-06, "loss": 0.1531, "step": 1500 }, { "epoch": 0.5204576976421637, "grad_norm": 2.8141188621520996, "learning_rate": 1e-06, "loss": 0.1957, "step": 1501 }, { "epoch": 0.5208044382801664, "grad_norm": 1.9651800394058228, "learning_rate": 1e-06, "loss": 0.1665, "step": 1502 }, { "epoch": 0.5211511789181692, "grad_norm": 1.5798146724700928, "learning_rate": 1e-06, "loss": 0.1452, "step": 1503 }, { "epoch": 0.521497919556172, "grad_norm": 1.6036591529846191, "learning_rate": 1e-06, "loss": 0.1268, "step": 1504 }, { "epoch": 0.5218446601941747, "grad_norm": 2.003415107727051, "learning_rate": 1e-06, "loss": 0.1337, "step": 1505 }, { "epoch": 0.5221914008321775, "grad_norm": 3.0233986377716064, "learning_rate": 1e-06, "loss": 0.2012, "step": 1506 }, { "epoch": 0.5225381414701803, "grad_norm": 1.6863069534301758, "learning_rate": 1e-06, "loss": 0.1401, "step": 1507 }, { "epoch": 0.5228848821081831, "grad_norm": 4.177762985229492, "learning_rate": 1e-06, "loss": 0.1805, "step": 1508 }, { "epoch": 0.5232316227461858, "grad_norm": 1.8863307237625122, "learning_rate": 1e-06, "loss": 0.1557, "step": 1509 }, { "epoch": 0.5235783633841886, "grad_norm": 4.922123432159424, "learning_rate": 1e-06, "loss": 0.211, "step": 1510 }, { "epoch": 0.5239251040221914, "grad_norm": 2.77559232711792, "learning_rate": 1e-06, "loss": 0.1621, "step": 1511 }, { "epoch": 0.5242718446601942, "grad_norm": 1.7231181859970093, "learning_rate": 1e-06, "loss": 0.1443, "step": 1512 }, { "epoch": 0.5246185852981969, "grad_norm": 2.0931448936462402, "learning_rate": 1e-06, "loss": 0.1776, "step": 1513 }, { "epoch": 0.5249653259361997, "grad_norm": 2.012320041656494, "learning_rate": 1e-06, "loss": 0.1765, "step": 1514 }, { "epoch": 0.5253120665742025, "grad_norm": 1.9464123249053955, "learning_rate": 1e-06, "loss": 0.1426, "step": 1515 }, { "epoch": 0.5256588072122053, "grad_norm": 4.585052013397217, "learning_rate": 1e-06, "loss": 0.1579, "step": 1516 }, { "epoch": 0.526005547850208, "grad_norm": 2.2083330154418945, "learning_rate": 1e-06, "loss": 0.1476, "step": 1517 }, { "epoch": 0.5263522884882108, "grad_norm": 2.2266805171966553, "learning_rate": 1e-06, "loss": 0.1477, "step": 1518 }, { "epoch": 0.5266990291262136, "grad_norm": 2.1281206607818604, "learning_rate": 1e-06, "loss": 0.137, "step": 1519 }, { "epoch": 0.5270457697642164, "grad_norm": 2.8681931495666504, "learning_rate": 1e-06, "loss": 0.1794, "step": 1520 }, { "epoch": 0.5273925104022191, "grad_norm": 2.800225257873535, "learning_rate": 1e-06, "loss": 0.1278, "step": 1521 }, { "epoch": 0.5277392510402219, "grad_norm": 4.79030704498291, "learning_rate": 1e-06, "loss": 0.2031, "step": 1522 }, { "epoch": 0.5280859916782247, "grad_norm": 1.678977370262146, "learning_rate": 1e-06, "loss": 0.1378, "step": 1523 }, { "epoch": 0.5284327323162274, "grad_norm": 2.589163064956665, "learning_rate": 1e-06, "loss": 0.1535, "step": 1524 }, { "epoch": 0.5287794729542302, "grad_norm": 2.825577974319458, "learning_rate": 1e-06, "loss": 0.1411, "step": 1525 }, { "epoch": 0.529126213592233, "grad_norm": 2.0116469860076904, "learning_rate": 1e-06, "loss": 0.1568, "step": 1526 }, { "epoch": 0.5294729542302358, "grad_norm": 1.699302077293396, "learning_rate": 1e-06, "loss": 0.1558, "step": 1527 }, { "epoch": 0.5298196948682385, "grad_norm": 2.1007838249206543, "learning_rate": 1e-06, "loss": 0.1772, "step": 1528 }, { "epoch": 0.5301664355062413, "grad_norm": 2.103917360305786, "learning_rate": 1e-06, "loss": 0.1601, "step": 1529 }, { "epoch": 0.5305131761442441, "grad_norm": 3.3394370079040527, "learning_rate": 1e-06, "loss": 0.1823, "step": 1530 }, { "epoch": 0.5308599167822469, "grad_norm": 2.7562904357910156, "learning_rate": 1e-06, "loss": 0.1509, "step": 1531 }, { "epoch": 0.5312066574202496, "grad_norm": 2.3599746227264404, "learning_rate": 1e-06, "loss": 0.1563, "step": 1532 }, { "epoch": 0.5315533980582524, "grad_norm": 2.014448404312134, "learning_rate": 1e-06, "loss": 0.1273, "step": 1533 }, { "epoch": 0.5319001386962552, "grad_norm": 1.5994056463241577, "learning_rate": 1e-06, "loss": 0.1494, "step": 1534 }, { "epoch": 0.532246879334258, "grad_norm": 1.998543620109558, "learning_rate": 1e-06, "loss": 0.1479, "step": 1535 }, { "epoch": 0.5325936199722607, "grad_norm": 1.7799046039581299, "learning_rate": 1e-06, "loss": 0.1466, "step": 1536 }, { "epoch": 0.5329403606102635, "grad_norm": 2.455554723739624, "learning_rate": 1e-06, "loss": 0.1201, "step": 1537 }, { "epoch": 0.5332871012482663, "grad_norm": 1.9349944591522217, "learning_rate": 1e-06, "loss": 0.1483, "step": 1538 }, { "epoch": 0.5336338418862691, "grad_norm": 2.089660406112671, "learning_rate": 1e-06, "loss": 0.1631, "step": 1539 }, { "epoch": 0.5339805825242718, "grad_norm": 2.356295108795166, "learning_rate": 1e-06, "loss": 0.1547, "step": 1540 }, { "epoch": 0.5343273231622746, "grad_norm": 2.6225013732910156, "learning_rate": 1e-06, "loss": 0.1634, "step": 1541 }, { "epoch": 0.5346740638002774, "grad_norm": 4.054690361022949, "learning_rate": 1e-06, "loss": 0.1597, "step": 1542 }, { "epoch": 0.5350208044382802, "grad_norm": 2.222895860671997, "learning_rate": 1e-06, "loss": 0.1719, "step": 1543 }, { "epoch": 0.5353675450762829, "grad_norm": 1.8733757734298706, "learning_rate": 1e-06, "loss": 0.1387, "step": 1544 }, { "epoch": 0.5357142857142857, "grad_norm": 3.893399715423584, "learning_rate": 1e-06, "loss": 0.1212, "step": 1545 }, { "epoch": 0.5360610263522885, "grad_norm": 2.2785661220550537, "learning_rate": 1e-06, "loss": 0.1573, "step": 1546 }, { "epoch": 0.5364077669902912, "grad_norm": 2.716165065765381, "learning_rate": 1e-06, "loss": 0.1809, "step": 1547 }, { "epoch": 0.536754507628294, "grad_norm": 2.239154100418091, "learning_rate": 1e-06, "loss": 0.1484, "step": 1548 }, { "epoch": 0.5371012482662968, "grad_norm": 2.2777199745178223, "learning_rate": 1e-06, "loss": 0.1776, "step": 1549 }, { "epoch": 0.5374479889042996, "grad_norm": 3.0242958068847656, "learning_rate": 1e-06, "loss": 0.1308, "step": 1550 }, { "epoch": 0.5377947295423023, "grad_norm": 2.226205587387085, "learning_rate": 1e-06, "loss": 0.1737, "step": 1551 }, { "epoch": 0.5381414701803051, "grad_norm": 2.850829601287842, "learning_rate": 1e-06, "loss": 0.1871, "step": 1552 }, { "epoch": 0.5384882108183079, "grad_norm": 3.371608257293701, "learning_rate": 1e-06, "loss": 0.1564, "step": 1553 }, { "epoch": 0.5388349514563107, "grad_norm": 2.1789321899414062, "learning_rate": 1e-06, "loss": 0.1589, "step": 1554 }, { "epoch": 0.5391816920943134, "grad_norm": 1.6266393661499023, "learning_rate": 1e-06, "loss": 0.1507, "step": 1555 }, { "epoch": 0.5395284327323162, "grad_norm": 1.8946685791015625, "learning_rate": 1e-06, "loss": 0.1556, "step": 1556 }, { "epoch": 0.539875173370319, "grad_norm": 3.5388271808624268, "learning_rate": 1e-06, "loss": 0.1408, "step": 1557 }, { "epoch": 0.5402219140083218, "grad_norm": 3.0828049182891846, "learning_rate": 1e-06, "loss": 0.1673, "step": 1558 }, { "epoch": 0.5405686546463245, "grad_norm": 2.063697576522827, "learning_rate": 1e-06, "loss": 0.1394, "step": 1559 }, { "epoch": 0.5409153952843273, "grad_norm": 2.2390551567077637, "learning_rate": 1e-06, "loss": 0.1453, "step": 1560 }, { "epoch": 0.5412621359223301, "grad_norm": 4.824660778045654, "learning_rate": 1e-06, "loss": 0.1882, "step": 1561 }, { "epoch": 0.5416088765603329, "grad_norm": 1.9755252599716187, "learning_rate": 1e-06, "loss": 0.1711, "step": 1562 }, { "epoch": 0.5419556171983356, "grad_norm": 1.8800153732299805, "learning_rate": 1e-06, "loss": 0.1458, "step": 1563 }, { "epoch": 0.5423023578363384, "grad_norm": 3.3188271522521973, "learning_rate": 1e-06, "loss": 0.1564, "step": 1564 }, { "epoch": 0.5426490984743412, "grad_norm": 1.953864574432373, "learning_rate": 1e-06, "loss": 0.1517, "step": 1565 }, { "epoch": 0.542995839112344, "grad_norm": 2.689492702484131, "learning_rate": 1e-06, "loss": 0.1671, "step": 1566 }, { "epoch": 0.5433425797503467, "grad_norm": 2.0836610794067383, "learning_rate": 1e-06, "loss": 0.1683, "step": 1567 }, { "epoch": 0.5436893203883495, "grad_norm": 2.045745849609375, "learning_rate": 1e-06, "loss": 0.1545, "step": 1568 }, { "epoch": 0.5440360610263523, "grad_norm": 4.5482401847839355, "learning_rate": 1e-06, "loss": 0.1518, "step": 1569 }, { "epoch": 0.544382801664355, "grad_norm": 2.2771973609924316, "learning_rate": 1e-06, "loss": 0.1972, "step": 1570 }, { "epoch": 0.5447295423023578, "grad_norm": 2.4855430126190186, "learning_rate": 1e-06, "loss": 0.1717, "step": 1571 }, { "epoch": 0.5450762829403606, "grad_norm": 1.9103244543075562, "learning_rate": 1e-06, "loss": 0.1662, "step": 1572 }, { "epoch": 0.5454230235783634, "grad_norm": 2.2929623126983643, "learning_rate": 1e-06, "loss": 0.1974, "step": 1573 }, { "epoch": 0.5457697642163661, "grad_norm": 2.0059826374053955, "learning_rate": 1e-06, "loss": 0.1622, "step": 1574 }, { "epoch": 0.5461165048543689, "grad_norm": 2.1234233379364014, "learning_rate": 1e-06, "loss": 0.178, "step": 1575 }, { "epoch": 0.5464632454923717, "grad_norm": 4.070279598236084, "learning_rate": 1e-06, "loss": 0.215, "step": 1576 }, { "epoch": 0.5468099861303745, "grad_norm": 2.5643842220306396, "learning_rate": 1e-06, "loss": 0.159, "step": 1577 }, { "epoch": 0.5471567267683772, "grad_norm": 1.6535149812698364, "learning_rate": 1e-06, "loss": 0.146, "step": 1578 }, { "epoch": 0.54750346740638, "grad_norm": 2.068358898162842, "learning_rate": 1e-06, "loss": 0.1691, "step": 1579 }, { "epoch": 0.5478502080443828, "grad_norm": 2.275562286376953, "learning_rate": 1e-06, "loss": 0.1564, "step": 1580 }, { "epoch": 0.5481969486823856, "grad_norm": 2.2625136375427246, "learning_rate": 1e-06, "loss": 0.1399, "step": 1581 }, { "epoch": 0.5485436893203883, "grad_norm": 2.553527593612671, "learning_rate": 1e-06, "loss": 0.1339, "step": 1582 }, { "epoch": 0.5488904299583911, "grad_norm": 2.405510425567627, "learning_rate": 1e-06, "loss": 0.1518, "step": 1583 }, { "epoch": 0.5492371705963939, "grad_norm": 2.205770969390869, "learning_rate": 1e-06, "loss": 0.1656, "step": 1584 }, { "epoch": 0.5495839112343966, "grad_norm": 2.961263418197632, "learning_rate": 1e-06, "loss": 0.1549, "step": 1585 }, { "epoch": 0.5499306518723994, "grad_norm": 2.239316701889038, "learning_rate": 1e-06, "loss": 0.1637, "step": 1586 }, { "epoch": 0.5502773925104022, "grad_norm": 2.7622323036193848, "learning_rate": 1e-06, "loss": 0.1726, "step": 1587 }, { "epoch": 0.550624133148405, "grad_norm": 2.3954594135284424, "learning_rate": 1e-06, "loss": 0.152, "step": 1588 }, { "epoch": 0.5509708737864077, "grad_norm": 2.528770923614502, "learning_rate": 1e-06, "loss": 0.1328, "step": 1589 }, { "epoch": 0.5513176144244105, "grad_norm": 4.014219760894775, "learning_rate": 1e-06, "loss": 0.1167, "step": 1590 }, { "epoch": 0.5516643550624133, "grad_norm": 1.444198727607727, "learning_rate": 1e-06, "loss": 0.1387, "step": 1591 }, { "epoch": 0.5520110957004161, "grad_norm": 1.6705647706985474, "learning_rate": 1e-06, "loss": 0.1104, "step": 1592 }, { "epoch": 0.5523578363384188, "grad_norm": 2.243098735809326, "learning_rate": 1e-06, "loss": 0.1426, "step": 1593 }, { "epoch": 0.5527045769764216, "grad_norm": 2.3548195362091064, "learning_rate": 1e-06, "loss": 0.1432, "step": 1594 }, { "epoch": 0.5530513176144244, "grad_norm": 1.7457523345947266, "learning_rate": 1e-06, "loss": 0.131, "step": 1595 }, { "epoch": 0.5533980582524272, "grad_norm": 2.6490869522094727, "learning_rate": 1e-06, "loss": 0.1571, "step": 1596 }, { "epoch": 0.5537447988904299, "grad_norm": 3.868406057357788, "learning_rate": 1e-06, "loss": 0.1713, "step": 1597 }, { "epoch": 0.5540915395284327, "grad_norm": 3.2075355052948, "learning_rate": 1e-06, "loss": 0.1467, "step": 1598 }, { "epoch": 0.5544382801664355, "grad_norm": 2.0262351036071777, "learning_rate": 1e-06, "loss": 0.1598, "step": 1599 }, { "epoch": 0.5547850208044383, "grad_norm": 2.029439926147461, "learning_rate": 1e-06, "loss": 0.1578, "step": 1600 }, { "epoch": 0.555131761442441, "grad_norm": 1.949021339416504, "learning_rate": 1e-06, "loss": 0.1524, "step": 1601 }, { "epoch": 0.5554785020804438, "grad_norm": 4.777602195739746, "learning_rate": 1e-06, "loss": 0.1553, "step": 1602 }, { "epoch": 0.5558252427184466, "grad_norm": 2.1241836547851562, "learning_rate": 1e-06, "loss": 0.1868, "step": 1603 }, { "epoch": 0.5561719833564494, "grad_norm": 2.2839200496673584, "learning_rate": 1e-06, "loss": 0.151, "step": 1604 }, { "epoch": 0.5565187239944521, "grad_norm": 2.7008562088012695, "learning_rate": 1e-06, "loss": 0.132, "step": 1605 }, { "epoch": 0.5568654646324549, "grad_norm": 2.1138343811035156, "learning_rate": 1e-06, "loss": 0.1349, "step": 1606 }, { "epoch": 0.5572122052704577, "grad_norm": 2.0788066387176514, "learning_rate": 1e-06, "loss": 0.1803, "step": 1607 }, { "epoch": 0.5575589459084604, "grad_norm": 2.323345899581909, "learning_rate": 1e-06, "loss": 0.1749, "step": 1608 }, { "epoch": 0.5579056865464632, "grad_norm": 3.4080355167388916, "learning_rate": 1e-06, "loss": 0.1453, "step": 1609 }, { "epoch": 0.558252427184466, "grad_norm": 2.0123281478881836, "learning_rate": 1e-06, "loss": 0.1345, "step": 1610 }, { "epoch": 0.5585991678224688, "grad_norm": 2.1220619678497314, "learning_rate": 1e-06, "loss": 0.1823, "step": 1611 }, { "epoch": 0.5589459084604715, "grad_norm": 2.4971022605895996, "learning_rate": 1e-06, "loss": 0.1591, "step": 1612 }, { "epoch": 0.5592926490984743, "grad_norm": 1.527266025543213, "learning_rate": 1e-06, "loss": 0.1459, "step": 1613 }, { "epoch": 0.5596393897364771, "grad_norm": 2.1191792488098145, "learning_rate": 1e-06, "loss": 0.1411, "step": 1614 }, { "epoch": 0.5599861303744799, "grad_norm": 1.978819489479065, "learning_rate": 1e-06, "loss": 0.1459, "step": 1615 }, { "epoch": 0.5603328710124826, "grad_norm": 2.3482275009155273, "learning_rate": 1e-06, "loss": 0.1784, "step": 1616 }, { "epoch": 0.5606796116504854, "grad_norm": 1.5257140398025513, "learning_rate": 1e-06, "loss": 0.142, "step": 1617 }, { "epoch": 0.5610263522884882, "grad_norm": 3.646221399307251, "learning_rate": 1e-06, "loss": 0.1756, "step": 1618 }, { "epoch": 0.561373092926491, "grad_norm": 1.7928833961486816, "learning_rate": 1e-06, "loss": 0.1423, "step": 1619 }, { "epoch": 0.5617198335644937, "grad_norm": 2.0367002487182617, "learning_rate": 1e-06, "loss": 0.1382, "step": 1620 }, { "epoch": 0.5620665742024965, "grad_norm": 1.9482024908065796, "learning_rate": 1e-06, "loss": 0.1573, "step": 1621 }, { "epoch": 0.5624133148404993, "grad_norm": 2.336115598678589, "learning_rate": 1e-06, "loss": 0.147, "step": 1622 }, { "epoch": 0.562760055478502, "grad_norm": 4.602109909057617, "learning_rate": 1e-06, "loss": 0.1847, "step": 1623 }, { "epoch": 0.5631067961165048, "grad_norm": 2.374413251876831, "learning_rate": 1e-06, "loss": 0.1668, "step": 1624 }, { "epoch": 0.5634535367545076, "grad_norm": 3.371469497680664, "learning_rate": 1e-06, "loss": 0.1654, "step": 1625 }, { "epoch": 0.5638002773925104, "grad_norm": 1.608858585357666, "learning_rate": 1e-06, "loss": 0.1259, "step": 1626 }, { "epoch": 0.5641470180305131, "grad_norm": 1.8223319053649902, "learning_rate": 1e-06, "loss": 0.1599, "step": 1627 }, { "epoch": 0.5644937586685159, "grad_norm": 3.2663540840148926, "learning_rate": 1e-06, "loss": 0.1525, "step": 1628 }, { "epoch": 0.5648404993065187, "grad_norm": 2.5820138454437256, "learning_rate": 1e-06, "loss": 0.1748, "step": 1629 }, { "epoch": 0.5651872399445215, "grad_norm": 2.0734665393829346, "learning_rate": 1e-06, "loss": 0.1456, "step": 1630 }, { "epoch": 0.5655339805825242, "grad_norm": 2.0002732276916504, "learning_rate": 1e-06, "loss": 0.1547, "step": 1631 }, { "epoch": 0.565880721220527, "grad_norm": 2.328211545944214, "learning_rate": 1e-06, "loss": 0.1603, "step": 1632 }, { "epoch": 0.5662274618585298, "grad_norm": 3.090592861175537, "learning_rate": 1e-06, "loss": 0.1467, "step": 1633 }, { "epoch": 0.5665742024965326, "grad_norm": 1.7330372333526611, "learning_rate": 1e-06, "loss": 0.1289, "step": 1634 }, { "epoch": 0.5669209431345353, "grad_norm": 1.6095702648162842, "learning_rate": 1e-06, "loss": 0.1386, "step": 1635 }, { "epoch": 0.5672676837725381, "grad_norm": 2.5205414295196533, "learning_rate": 1e-06, "loss": 0.1817, "step": 1636 }, { "epoch": 0.5676144244105409, "grad_norm": 2.5302205085754395, "learning_rate": 1e-06, "loss": 0.1515, "step": 1637 }, { "epoch": 0.5679611650485437, "grad_norm": 2.0151143074035645, "learning_rate": 1e-06, "loss": 0.1376, "step": 1638 }, { "epoch": 0.5683079056865464, "grad_norm": 4.463467121124268, "learning_rate": 1e-06, "loss": 0.1865, "step": 1639 }, { "epoch": 0.5686546463245492, "grad_norm": 2.7108094692230225, "learning_rate": 1e-06, "loss": 0.1527, "step": 1640 }, { "epoch": 0.569001386962552, "grad_norm": 3.0464956760406494, "learning_rate": 1e-06, "loss": 0.1671, "step": 1641 }, { "epoch": 0.5693481276005548, "grad_norm": 2.9522790908813477, "learning_rate": 1e-06, "loss": 0.1863, "step": 1642 }, { "epoch": 0.5696948682385575, "grad_norm": 2.414750099182129, "learning_rate": 1e-06, "loss": 0.1537, "step": 1643 }, { "epoch": 0.5700416088765603, "grad_norm": 1.9789800643920898, "learning_rate": 1e-06, "loss": 0.1405, "step": 1644 }, { "epoch": 0.5703883495145631, "grad_norm": 2.12092661857605, "learning_rate": 1e-06, "loss": 0.1452, "step": 1645 }, { "epoch": 0.5707350901525658, "grad_norm": 2.106600046157837, "learning_rate": 1e-06, "loss": 0.1769, "step": 1646 }, { "epoch": 0.5710818307905686, "grad_norm": 1.6850626468658447, "learning_rate": 1e-06, "loss": 0.1415, "step": 1647 }, { "epoch": 0.5714285714285714, "grad_norm": 2.36781907081604, "learning_rate": 1e-06, "loss": 0.1574, "step": 1648 }, { "epoch": 0.5717753120665742, "grad_norm": 3.573758840560913, "learning_rate": 1e-06, "loss": 0.1754, "step": 1649 }, { "epoch": 0.5721220527045769, "grad_norm": 2.212064743041992, "learning_rate": 1e-06, "loss": 0.1398, "step": 1650 }, { "epoch": 0.5724687933425797, "grad_norm": 3.7138683795928955, "learning_rate": 1e-06, "loss": 0.1486, "step": 1651 }, { "epoch": 0.5728155339805825, "grad_norm": 2.328507900238037, "learning_rate": 1e-06, "loss": 0.181, "step": 1652 }, { "epoch": 0.5731622746185853, "grad_norm": 2.0306859016418457, "learning_rate": 1e-06, "loss": 0.1678, "step": 1653 }, { "epoch": 0.573509015256588, "grad_norm": 2.6543784141540527, "learning_rate": 1e-06, "loss": 0.1798, "step": 1654 }, { "epoch": 0.5738557558945908, "grad_norm": 2.0634257793426514, "learning_rate": 1e-06, "loss": 0.1597, "step": 1655 }, { "epoch": 0.5742024965325936, "grad_norm": 3.299668788909912, "learning_rate": 1e-06, "loss": 0.1498, "step": 1656 }, { "epoch": 0.5745492371705964, "grad_norm": 1.8128036260604858, "learning_rate": 1e-06, "loss": 0.1252, "step": 1657 }, { "epoch": 0.5748959778085991, "grad_norm": 2.365511894226074, "learning_rate": 1e-06, "loss": 0.1581, "step": 1658 }, { "epoch": 0.5752427184466019, "grad_norm": 2.4852473735809326, "learning_rate": 1e-06, "loss": 0.1702, "step": 1659 }, { "epoch": 0.5755894590846047, "grad_norm": 2.0349810123443604, "learning_rate": 1e-06, "loss": 0.1515, "step": 1660 }, { "epoch": 0.5759361997226075, "grad_norm": 1.8448779582977295, "learning_rate": 1e-06, "loss": 0.1553, "step": 1661 }, { "epoch": 0.5762829403606102, "grad_norm": 3.345630645751953, "learning_rate": 1e-06, "loss": 0.1508, "step": 1662 }, { "epoch": 0.576629680998613, "grad_norm": 2.9367878437042236, "learning_rate": 1e-06, "loss": 0.1637, "step": 1663 }, { "epoch": 0.5769764216366158, "grad_norm": 1.9184025526046753, "learning_rate": 1e-06, "loss": 0.1452, "step": 1664 }, { "epoch": 0.5773231622746186, "grad_norm": 3.0208446979522705, "learning_rate": 1e-06, "loss": 0.1319, "step": 1665 }, { "epoch": 0.5776699029126213, "grad_norm": 1.9824481010437012, "learning_rate": 1e-06, "loss": 0.1285, "step": 1666 }, { "epoch": 0.5780166435506241, "grad_norm": 1.768882393836975, "learning_rate": 1e-06, "loss": 0.1589, "step": 1667 }, { "epoch": 0.5783633841886269, "grad_norm": 5.792402267456055, "learning_rate": 1e-06, "loss": 0.1643, "step": 1668 }, { "epoch": 0.5787101248266296, "grad_norm": 2.2808361053466797, "learning_rate": 1e-06, "loss": 0.1925, "step": 1669 }, { "epoch": 0.5790568654646324, "grad_norm": 1.661110520362854, "learning_rate": 1e-06, "loss": 0.1383, "step": 1670 }, { "epoch": 0.5794036061026352, "grad_norm": 2.764539957046509, "learning_rate": 1e-06, "loss": 0.1346, "step": 1671 }, { "epoch": 0.579750346740638, "grad_norm": 2.591721296310425, "learning_rate": 1e-06, "loss": 0.1683, "step": 1672 }, { "epoch": 0.5800970873786407, "grad_norm": 1.6234064102172852, "learning_rate": 1e-06, "loss": 0.1638, "step": 1673 }, { "epoch": 0.5804438280166435, "grad_norm": 3.9577250480651855, "learning_rate": 1e-06, "loss": 0.1731, "step": 1674 }, { "epoch": 0.5807905686546463, "grad_norm": 3.799262762069702, "learning_rate": 1e-06, "loss": 0.1556, "step": 1675 }, { "epoch": 0.5811373092926491, "grad_norm": 2.9457106590270996, "learning_rate": 1e-06, "loss": 0.1611, "step": 1676 }, { "epoch": 0.5814840499306518, "grad_norm": 2.1177031993865967, "learning_rate": 1e-06, "loss": 0.1723, "step": 1677 }, { "epoch": 0.5818307905686546, "grad_norm": 1.5257512331008911, "learning_rate": 1e-06, "loss": 0.132, "step": 1678 }, { "epoch": 0.5821775312066574, "grad_norm": 2.6540706157684326, "learning_rate": 1e-06, "loss": 0.1648, "step": 1679 }, { "epoch": 0.5825242718446602, "grad_norm": 2.4282069206237793, "learning_rate": 1e-06, "loss": 0.1462, "step": 1680 }, { "epoch": 0.5828710124826629, "grad_norm": 1.7106441259384155, "learning_rate": 1e-06, "loss": 0.1528, "step": 1681 }, { "epoch": 0.5832177531206657, "grad_norm": 2.3406167030334473, "learning_rate": 1e-06, "loss": 0.1909, "step": 1682 }, { "epoch": 0.5835644937586685, "grad_norm": 2.504379987716675, "learning_rate": 1e-06, "loss": 0.1444, "step": 1683 }, { "epoch": 0.5839112343966713, "grad_norm": 4.548439979553223, "learning_rate": 1e-06, "loss": 0.1736, "step": 1684 }, { "epoch": 0.584257975034674, "grad_norm": 2.725050449371338, "learning_rate": 1e-06, "loss": 0.1406, "step": 1685 }, { "epoch": 0.5846047156726768, "grad_norm": 2.026104211807251, "learning_rate": 1e-06, "loss": 0.1531, "step": 1686 }, { "epoch": 0.5849514563106796, "grad_norm": 2.1058292388916016, "learning_rate": 1e-06, "loss": 0.1434, "step": 1687 }, { "epoch": 0.5852981969486823, "grad_norm": 2.1308064460754395, "learning_rate": 1e-06, "loss": 0.1529, "step": 1688 }, { "epoch": 0.5856449375866851, "grad_norm": 4.247373104095459, "learning_rate": 1e-06, "loss": 0.1962, "step": 1689 }, { "epoch": 0.5859916782246879, "grad_norm": 4.272634029388428, "learning_rate": 1e-06, "loss": 0.1319, "step": 1690 }, { "epoch": 0.5863384188626907, "grad_norm": 1.9161146879196167, "learning_rate": 1e-06, "loss": 0.1794, "step": 1691 }, { "epoch": 0.5866851595006934, "grad_norm": 3.2819652557373047, "learning_rate": 1e-06, "loss": 0.1388, "step": 1692 }, { "epoch": 0.5870319001386962, "grad_norm": 2.6425933837890625, "learning_rate": 1e-06, "loss": 0.1508, "step": 1693 }, { "epoch": 0.587378640776699, "grad_norm": 2.2003285884857178, "learning_rate": 1e-06, "loss": 0.1315, "step": 1694 }, { "epoch": 0.5877253814147018, "grad_norm": 2.436519145965576, "learning_rate": 1e-06, "loss": 0.1888, "step": 1695 }, { "epoch": 0.5880721220527045, "grad_norm": 2.4235095977783203, "learning_rate": 1e-06, "loss": 0.1574, "step": 1696 }, { "epoch": 0.5884188626907073, "grad_norm": 2.7736899852752686, "learning_rate": 1e-06, "loss": 0.1324, "step": 1697 }, { "epoch": 0.5887656033287101, "grad_norm": 6.031193256378174, "learning_rate": 1e-06, "loss": 0.1635, "step": 1698 }, { "epoch": 0.5891123439667129, "grad_norm": 1.7690778970718384, "learning_rate": 1e-06, "loss": 0.1351, "step": 1699 }, { "epoch": 0.5894590846047156, "grad_norm": 1.9819053411483765, "learning_rate": 1e-06, "loss": 0.1649, "step": 1700 }, { "epoch": 0.5898058252427184, "grad_norm": 2.2417149543762207, "learning_rate": 1e-06, "loss": 0.1551, "step": 1701 }, { "epoch": 0.5901525658807212, "grad_norm": 1.7193907499313354, "learning_rate": 1e-06, "loss": 0.1373, "step": 1702 }, { "epoch": 0.590499306518724, "grad_norm": 2.16274356842041, "learning_rate": 1e-06, "loss": 0.1629, "step": 1703 }, { "epoch": 0.5908460471567267, "grad_norm": 2.157911539077759, "learning_rate": 1e-06, "loss": 0.1289, "step": 1704 }, { "epoch": 0.5911927877947295, "grad_norm": 3.0244007110595703, "learning_rate": 1e-06, "loss": 0.157, "step": 1705 }, { "epoch": 0.5915395284327323, "grad_norm": 1.9083997011184692, "learning_rate": 1e-06, "loss": 0.1736, "step": 1706 }, { "epoch": 0.591886269070735, "grad_norm": 2.4870333671569824, "learning_rate": 1e-06, "loss": 0.1541, "step": 1707 }, { "epoch": 0.5922330097087378, "grad_norm": 2.714510202407837, "learning_rate": 1e-06, "loss": 0.1561, "step": 1708 }, { "epoch": 0.5925797503467406, "grad_norm": 1.7150081396102905, "learning_rate": 1e-06, "loss": 0.1437, "step": 1709 }, { "epoch": 0.5929264909847434, "grad_norm": 2.5376031398773193, "learning_rate": 1e-06, "loss": 0.1588, "step": 1710 }, { "epoch": 0.5932732316227461, "grad_norm": 2.173816680908203, "learning_rate": 1e-06, "loss": 0.1668, "step": 1711 }, { "epoch": 0.5936199722607489, "grad_norm": 3.2527916431427, "learning_rate": 1e-06, "loss": 0.1579, "step": 1712 }, { "epoch": 0.5939667128987517, "grad_norm": 2.0413970947265625, "learning_rate": 1e-06, "loss": 0.1535, "step": 1713 }, { "epoch": 0.5943134535367545, "grad_norm": 2.2500665187835693, "learning_rate": 1e-06, "loss": 0.1744, "step": 1714 }, { "epoch": 0.5946601941747572, "grad_norm": 2.931854248046875, "learning_rate": 1e-06, "loss": 0.1729, "step": 1715 }, { "epoch": 0.59500693481276, "grad_norm": 1.800407886505127, "learning_rate": 1e-06, "loss": 0.148, "step": 1716 }, { "epoch": 0.5953536754507628, "grad_norm": 3.381312370300293, "learning_rate": 1e-06, "loss": 0.1358, "step": 1717 }, { "epoch": 0.5957004160887656, "grad_norm": 3.778513193130493, "learning_rate": 1e-06, "loss": 0.156, "step": 1718 }, { "epoch": 0.5960471567267683, "grad_norm": 1.6460663080215454, "learning_rate": 1e-06, "loss": 0.1479, "step": 1719 }, { "epoch": 0.5963938973647711, "grad_norm": 2.169105052947998, "learning_rate": 1e-06, "loss": 0.1712, "step": 1720 }, { "epoch": 0.5967406380027739, "grad_norm": 1.761500358581543, "learning_rate": 1e-06, "loss": 0.1442, "step": 1721 }, { "epoch": 0.5970873786407767, "grad_norm": 1.7620950937271118, "learning_rate": 1e-06, "loss": 0.1423, "step": 1722 }, { "epoch": 0.5974341192787794, "grad_norm": 2.3129122257232666, "learning_rate": 1e-06, "loss": 0.1765, "step": 1723 }, { "epoch": 0.5977808599167822, "grad_norm": 2.858874559402466, "learning_rate": 1e-06, "loss": 0.1467, "step": 1724 }, { "epoch": 0.598127600554785, "grad_norm": 3.7320244312286377, "learning_rate": 1e-06, "loss": 0.1796, "step": 1725 }, { "epoch": 0.5984743411927878, "grad_norm": 3.407935380935669, "learning_rate": 1e-06, "loss": 0.1386, "step": 1726 }, { "epoch": 0.5988210818307905, "grad_norm": 2.2188820838928223, "learning_rate": 1e-06, "loss": 0.1723, "step": 1727 }, { "epoch": 0.5991678224687933, "grad_norm": 1.947733759880066, "learning_rate": 1e-06, "loss": 0.1716, "step": 1728 }, { "epoch": 0.5995145631067961, "grad_norm": 2.274256467819214, "learning_rate": 1e-06, "loss": 0.1519, "step": 1729 }, { "epoch": 0.5998613037447988, "grad_norm": 1.881277084350586, "learning_rate": 1e-06, "loss": 0.127, "step": 1730 }, { "epoch": 0.6002080443828016, "grad_norm": 2.522675037384033, "learning_rate": 1e-06, "loss": 0.1806, "step": 1731 }, { "epoch": 0.6005547850208044, "grad_norm": 3.52107834815979, "learning_rate": 1e-06, "loss": 0.1581, "step": 1732 }, { "epoch": 0.6009015256588072, "grad_norm": 2.1973533630371094, "learning_rate": 1e-06, "loss": 0.126, "step": 1733 }, { "epoch": 0.6012482662968099, "grad_norm": 2.215819835662842, "learning_rate": 1e-06, "loss": 0.1454, "step": 1734 }, { "epoch": 0.6015950069348127, "grad_norm": 1.8062198162078857, "learning_rate": 1e-06, "loss": 0.1583, "step": 1735 }, { "epoch": 0.6019417475728155, "grad_norm": 2.090193510055542, "learning_rate": 1e-06, "loss": 0.1434, "step": 1736 }, { "epoch": 0.6022884882108183, "grad_norm": 1.7331854104995728, "learning_rate": 1e-06, "loss": 0.1323, "step": 1737 }, { "epoch": 0.602635228848821, "grad_norm": 2.4180166721343994, "learning_rate": 1e-06, "loss": 0.1384, "step": 1738 }, { "epoch": 0.6029819694868238, "grad_norm": 3.263031482696533, "learning_rate": 1e-06, "loss": 0.1324, "step": 1739 }, { "epoch": 0.6033287101248266, "grad_norm": 2.341269016265869, "learning_rate": 1e-06, "loss": 0.1579, "step": 1740 }, { "epoch": 0.6036754507628294, "grad_norm": 3.2346274852752686, "learning_rate": 1e-06, "loss": 0.1382, "step": 1741 }, { "epoch": 0.6040221914008321, "grad_norm": 1.9541387557983398, "learning_rate": 1e-06, "loss": 0.1673, "step": 1742 }, { "epoch": 0.6043689320388349, "grad_norm": 3.1931421756744385, "learning_rate": 1e-06, "loss": 0.1437, "step": 1743 }, { "epoch": 0.6047156726768377, "grad_norm": 1.9747329950332642, "learning_rate": 1e-06, "loss": 0.1346, "step": 1744 }, { "epoch": 0.6050624133148405, "grad_norm": 3.360968589782715, "learning_rate": 1e-06, "loss": 0.1584, "step": 1745 }, { "epoch": 0.6054091539528432, "grad_norm": 2.160388469696045, "learning_rate": 1e-06, "loss": 0.132, "step": 1746 }, { "epoch": 0.605755894590846, "grad_norm": 4.136070728302002, "learning_rate": 1e-06, "loss": 0.1506, "step": 1747 }, { "epoch": 0.6061026352288488, "grad_norm": 2.1393015384674072, "learning_rate": 1e-06, "loss": 0.1568, "step": 1748 }, { "epoch": 0.6064493758668515, "grad_norm": 2.9097952842712402, "learning_rate": 1e-06, "loss": 0.1343, "step": 1749 }, { "epoch": 0.6067961165048543, "grad_norm": 1.9220006465911865, "learning_rate": 1e-06, "loss": 0.1446, "step": 1750 }, { "epoch": 0.6071428571428571, "grad_norm": 2.2562577724456787, "learning_rate": 1e-06, "loss": 0.1751, "step": 1751 }, { "epoch": 0.6074895977808599, "grad_norm": 2.5833182334899902, "learning_rate": 1e-06, "loss": 0.1606, "step": 1752 }, { "epoch": 0.6078363384188626, "grad_norm": 6.893828392028809, "learning_rate": 1e-06, "loss": 0.1516, "step": 1753 }, { "epoch": 0.6081830790568654, "grad_norm": 6.116687774658203, "learning_rate": 1e-06, "loss": 0.1175, "step": 1754 }, { "epoch": 0.6085298196948682, "grad_norm": 1.879164218902588, "learning_rate": 1e-06, "loss": 0.1381, "step": 1755 }, { "epoch": 0.608876560332871, "grad_norm": 3.298027753829956, "learning_rate": 1e-06, "loss": 0.1726, "step": 1756 }, { "epoch": 0.6092233009708737, "grad_norm": 2.1097750663757324, "learning_rate": 1e-06, "loss": 0.165, "step": 1757 }, { "epoch": 0.6095700416088765, "grad_norm": 4.932060241699219, "learning_rate": 1e-06, "loss": 0.1399, "step": 1758 }, { "epoch": 0.6099167822468793, "grad_norm": 1.9408305883407593, "learning_rate": 1e-06, "loss": 0.1661, "step": 1759 }, { "epoch": 0.6102635228848821, "grad_norm": 1.8838698863983154, "learning_rate": 1e-06, "loss": 0.1617, "step": 1760 }, { "epoch": 0.6106102635228848, "grad_norm": 2.995274782180786, "learning_rate": 1e-06, "loss": 0.1361, "step": 1761 }, { "epoch": 0.6109570041608876, "grad_norm": 2.9905359745025635, "learning_rate": 1e-06, "loss": 0.1485, "step": 1762 }, { "epoch": 0.6113037447988904, "grad_norm": 2.9674386978149414, "learning_rate": 1e-06, "loss": 0.1503, "step": 1763 }, { "epoch": 0.6116504854368932, "grad_norm": 2.390683889389038, "learning_rate": 1e-06, "loss": 0.1585, "step": 1764 }, { "epoch": 0.6119972260748959, "grad_norm": 2.8863577842712402, "learning_rate": 1e-06, "loss": 0.1579, "step": 1765 }, { "epoch": 0.6123439667128987, "grad_norm": 3.183380365371704, "learning_rate": 1e-06, "loss": 0.1292, "step": 1766 }, { "epoch": 0.6126907073509015, "grad_norm": 2.080923318862915, "learning_rate": 1e-06, "loss": 0.1166, "step": 1767 }, { "epoch": 0.6130374479889042, "grad_norm": 2.8261148929595947, "learning_rate": 1e-06, "loss": 0.159, "step": 1768 }, { "epoch": 0.613384188626907, "grad_norm": 2.0629355907440186, "learning_rate": 1e-06, "loss": 0.1433, "step": 1769 }, { "epoch": 0.6137309292649098, "grad_norm": 1.8249051570892334, "learning_rate": 1e-06, "loss": 0.1448, "step": 1770 }, { "epoch": 0.6140776699029126, "grad_norm": 2.489915132522583, "learning_rate": 1e-06, "loss": 0.1975, "step": 1771 }, { "epoch": 0.6144244105409153, "grad_norm": 2.171525716781616, "learning_rate": 1e-06, "loss": 0.143, "step": 1772 }, { "epoch": 0.6147711511789181, "grad_norm": 3.3242056369781494, "learning_rate": 1e-06, "loss": 0.1337, "step": 1773 }, { "epoch": 0.6151178918169209, "grad_norm": 3.56597900390625, "learning_rate": 1e-06, "loss": 0.1278, "step": 1774 }, { "epoch": 0.6154646324549237, "grad_norm": 1.8567633628845215, "learning_rate": 1e-06, "loss": 0.1546, "step": 1775 }, { "epoch": 0.6158113730929264, "grad_norm": 2.3509087562561035, "learning_rate": 1e-06, "loss": 0.1677, "step": 1776 }, { "epoch": 0.6161581137309292, "grad_norm": 2.3500568866729736, "learning_rate": 1e-06, "loss": 0.1573, "step": 1777 }, { "epoch": 0.616504854368932, "grad_norm": 1.687183141708374, "learning_rate": 1e-06, "loss": 0.1384, "step": 1778 }, { "epoch": 0.6168515950069348, "grad_norm": 2.0229380130767822, "learning_rate": 1e-06, "loss": 0.1231, "step": 1779 }, { "epoch": 0.6171983356449375, "grad_norm": 2.1885972023010254, "learning_rate": 1e-06, "loss": 0.1656, "step": 1780 }, { "epoch": 0.6175450762829403, "grad_norm": 4.203535556793213, "learning_rate": 1e-06, "loss": 0.1725, "step": 1781 }, { "epoch": 0.6178918169209431, "grad_norm": 2.7054553031921387, "learning_rate": 1e-06, "loss": 0.1316, "step": 1782 }, { "epoch": 0.6182385575589459, "grad_norm": 2.386046886444092, "learning_rate": 1e-06, "loss": 0.1086, "step": 1783 }, { "epoch": 0.6185852981969486, "grad_norm": 2.22220516204834, "learning_rate": 1e-06, "loss": 0.199, "step": 1784 }, { "epoch": 0.6189320388349514, "grad_norm": 2.0740716457366943, "learning_rate": 1e-06, "loss": 0.1167, "step": 1785 }, { "epoch": 0.6192787794729542, "grad_norm": 3.928887128829956, "learning_rate": 1e-06, "loss": 0.1659, "step": 1786 }, { "epoch": 0.619625520110957, "grad_norm": 3.2319023609161377, "learning_rate": 1e-06, "loss": 0.1538, "step": 1787 }, { "epoch": 0.6199722607489597, "grad_norm": 2.2219529151916504, "learning_rate": 1e-06, "loss": 0.1689, "step": 1788 }, { "epoch": 0.6203190013869625, "grad_norm": 1.9555144309997559, "learning_rate": 1e-06, "loss": 0.1572, "step": 1789 }, { "epoch": 0.6206657420249653, "grad_norm": 1.8324187994003296, "learning_rate": 1e-06, "loss": 0.1549, "step": 1790 }, { "epoch": 0.621012482662968, "grad_norm": 4.268178462982178, "learning_rate": 1e-06, "loss": 0.1676, "step": 1791 }, { "epoch": 0.6213592233009708, "grad_norm": 2.28480863571167, "learning_rate": 1e-06, "loss": 0.1555, "step": 1792 }, { "epoch": 0.6217059639389736, "grad_norm": 2.970492362976074, "learning_rate": 1e-06, "loss": 0.1369, "step": 1793 }, { "epoch": 0.6220527045769764, "grad_norm": 2.2355594635009766, "learning_rate": 1e-06, "loss": 0.2016, "step": 1794 }, { "epoch": 0.6223994452149791, "grad_norm": 2.1921377182006836, "learning_rate": 1e-06, "loss": 0.1664, "step": 1795 }, { "epoch": 0.6227461858529819, "grad_norm": 2.3036246299743652, "learning_rate": 1e-06, "loss": 0.1617, "step": 1796 }, { "epoch": 0.6230929264909847, "grad_norm": 2.2555198669433594, "learning_rate": 1e-06, "loss": 0.1693, "step": 1797 }, { "epoch": 0.6234396671289875, "grad_norm": 2.416224718093872, "learning_rate": 1e-06, "loss": 0.1745, "step": 1798 }, { "epoch": 0.6237864077669902, "grad_norm": 1.9002783298492432, "learning_rate": 1e-06, "loss": 0.1618, "step": 1799 }, { "epoch": 0.624133148404993, "grad_norm": 1.8858561515808105, "learning_rate": 1e-06, "loss": 0.1885, "step": 1800 }, { "epoch": 0.6244798890429958, "grad_norm": 3.2046573162078857, "learning_rate": 1e-06, "loss": 0.1588, "step": 1801 }, { "epoch": 0.6248266296809986, "grad_norm": 3.2242980003356934, "learning_rate": 1e-06, "loss": 0.1758, "step": 1802 }, { "epoch": 0.6251733703190014, "grad_norm": 4.321735858917236, "learning_rate": 1e-06, "loss": 0.1745, "step": 1803 }, { "epoch": 0.6255201109570042, "grad_norm": 5.258564472198486, "learning_rate": 1e-06, "loss": 0.157, "step": 1804 }, { "epoch": 0.625866851595007, "grad_norm": 2.0180282592773438, "learning_rate": 1e-06, "loss": 0.1327, "step": 1805 }, { "epoch": 0.6262135922330098, "grad_norm": 2.150374412536621, "learning_rate": 1e-06, "loss": 0.129, "step": 1806 }, { "epoch": 0.6265603328710125, "grad_norm": 2.426352024078369, "learning_rate": 1e-06, "loss": 0.155, "step": 1807 }, { "epoch": 0.6269070735090153, "grad_norm": 3.8777828216552734, "learning_rate": 1e-06, "loss": 0.134, "step": 1808 }, { "epoch": 0.6272538141470181, "grad_norm": 2.8608689308166504, "learning_rate": 1e-06, "loss": 0.1723, "step": 1809 }, { "epoch": 0.6276005547850209, "grad_norm": 2.592209577560425, "learning_rate": 1e-06, "loss": 0.1473, "step": 1810 }, { "epoch": 0.6279472954230236, "grad_norm": 1.9494845867156982, "learning_rate": 1e-06, "loss": 0.1771, "step": 1811 }, { "epoch": 0.6282940360610264, "grad_norm": 1.7008180618286133, "learning_rate": 1e-06, "loss": 0.1429, "step": 1812 }, { "epoch": 0.6286407766990292, "grad_norm": 1.7186037302017212, "learning_rate": 1e-06, "loss": 0.161, "step": 1813 }, { "epoch": 0.628987517337032, "grad_norm": 2.3491291999816895, "learning_rate": 1e-06, "loss": 0.1767, "step": 1814 }, { "epoch": 0.6293342579750347, "grad_norm": 2.4897589683532715, "learning_rate": 1e-06, "loss": 0.1528, "step": 1815 }, { "epoch": 0.6296809986130375, "grad_norm": 1.9620110988616943, "learning_rate": 1e-06, "loss": 0.1481, "step": 1816 }, { "epoch": 0.6300277392510403, "grad_norm": 2.08915114402771, "learning_rate": 1e-06, "loss": 0.1465, "step": 1817 }, { "epoch": 0.630374479889043, "grad_norm": 3.368116617202759, "learning_rate": 1e-06, "loss": 0.1516, "step": 1818 }, { "epoch": 0.6307212205270458, "grad_norm": 2.13322377204895, "learning_rate": 1e-06, "loss": 0.153, "step": 1819 }, { "epoch": 0.6310679611650486, "grad_norm": 2.479769229888916, "learning_rate": 1e-06, "loss": 0.1646, "step": 1820 }, { "epoch": 0.6314147018030514, "grad_norm": 3.7209994792938232, "learning_rate": 1e-06, "loss": 0.1795, "step": 1821 }, { "epoch": 0.6317614424410541, "grad_norm": 4.989049434661865, "learning_rate": 1e-06, "loss": 0.1582, "step": 1822 }, { "epoch": 0.6321081830790569, "grad_norm": 2.99127459526062, "learning_rate": 1e-06, "loss": 0.1149, "step": 1823 }, { "epoch": 0.6324549237170597, "grad_norm": 3.3212974071502686, "learning_rate": 1e-06, "loss": 0.1359, "step": 1824 }, { "epoch": 0.6328016643550625, "grad_norm": 2.3676962852478027, "learning_rate": 1e-06, "loss": 0.1448, "step": 1825 }, { "epoch": 0.6331484049930652, "grad_norm": 2.0515060424804688, "learning_rate": 1e-06, "loss": 0.1557, "step": 1826 }, { "epoch": 0.633495145631068, "grad_norm": 3.021026372909546, "learning_rate": 1e-06, "loss": 0.143, "step": 1827 }, { "epoch": 0.6338418862690708, "grad_norm": 2.328517198562622, "learning_rate": 1e-06, "loss": 0.1265, "step": 1828 }, { "epoch": 0.6341886269070736, "grad_norm": 1.7838841676712036, "learning_rate": 1e-06, "loss": 0.1188, "step": 1829 }, { "epoch": 0.6345353675450763, "grad_norm": 2.873117685317993, "learning_rate": 1e-06, "loss": 0.1251, "step": 1830 }, { "epoch": 0.6348821081830791, "grad_norm": 2.515712022781372, "learning_rate": 1e-06, "loss": 0.1258, "step": 1831 }, { "epoch": 0.6352288488210819, "grad_norm": 4.897085189819336, "learning_rate": 1e-06, "loss": 0.1528, "step": 1832 }, { "epoch": 0.6355755894590847, "grad_norm": 2.860135793685913, "learning_rate": 1e-06, "loss": 0.1385, "step": 1833 }, { "epoch": 0.6359223300970874, "grad_norm": 2.492570400238037, "learning_rate": 1e-06, "loss": 0.1761, "step": 1834 }, { "epoch": 0.6362690707350902, "grad_norm": 1.7522510290145874, "learning_rate": 1e-06, "loss": 0.1309, "step": 1835 }, { "epoch": 0.636615811373093, "grad_norm": 2.5122199058532715, "learning_rate": 1e-06, "loss": 0.1502, "step": 1836 }, { "epoch": 0.6369625520110958, "grad_norm": 1.8004939556121826, "learning_rate": 1e-06, "loss": 0.1482, "step": 1837 }, { "epoch": 0.6373092926490985, "grad_norm": 2.451542854309082, "learning_rate": 1e-06, "loss": 0.1656, "step": 1838 }, { "epoch": 0.6376560332871013, "grad_norm": 3.232908248901367, "learning_rate": 1e-06, "loss": 0.1625, "step": 1839 }, { "epoch": 0.6380027739251041, "grad_norm": 2.172776460647583, "learning_rate": 1e-06, "loss": 0.1713, "step": 1840 }, { "epoch": 0.6383495145631068, "grad_norm": 2.3084630966186523, "learning_rate": 1e-06, "loss": 0.1474, "step": 1841 }, { "epoch": 0.6386962552011096, "grad_norm": 1.8688488006591797, "learning_rate": 1e-06, "loss": 0.1386, "step": 1842 }, { "epoch": 0.6390429958391124, "grad_norm": 2.7231826782226562, "learning_rate": 1e-06, "loss": 0.1523, "step": 1843 }, { "epoch": 0.6393897364771152, "grad_norm": 2.4542200565338135, "learning_rate": 1e-06, "loss": 0.1505, "step": 1844 }, { "epoch": 0.6397364771151179, "grad_norm": 2.8579893112182617, "learning_rate": 1e-06, "loss": 0.1565, "step": 1845 }, { "epoch": 0.6400832177531207, "grad_norm": 3.261667251586914, "learning_rate": 1e-06, "loss": 0.1216, "step": 1846 }, { "epoch": 0.6404299583911235, "grad_norm": 2.358067035675049, "learning_rate": 1e-06, "loss": 0.17, "step": 1847 }, { "epoch": 0.6407766990291263, "grad_norm": 3.6049678325653076, "learning_rate": 1e-06, "loss": 0.1488, "step": 1848 }, { "epoch": 0.641123439667129, "grad_norm": 2.246796131134033, "learning_rate": 1e-06, "loss": 0.1601, "step": 1849 }, { "epoch": 0.6414701803051318, "grad_norm": 1.8105849027633667, "learning_rate": 1e-06, "loss": 0.1093, "step": 1850 }, { "epoch": 0.6418169209431346, "grad_norm": 1.771470546722412, "learning_rate": 1e-06, "loss": 0.1511, "step": 1851 }, { "epoch": 0.6421636615811374, "grad_norm": 2.7330245971679688, "learning_rate": 1e-06, "loss": 0.176, "step": 1852 }, { "epoch": 0.6425104022191401, "grad_norm": 4.204245567321777, "learning_rate": 1e-06, "loss": 0.1684, "step": 1853 }, { "epoch": 0.6428571428571429, "grad_norm": 2.937845468521118, "learning_rate": 1e-06, "loss": 0.1684, "step": 1854 }, { "epoch": 0.6432038834951457, "grad_norm": 3.5551681518554688, "learning_rate": 1e-06, "loss": 0.1421, "step": 1855 }, { "epoch": 0.6435506241331485, "grad_norm": 3.067307472229004, "learning_rate": 1e-06, "loss": 0.1413, "step": 1856 }, { "epoch": 0.6438973647711512, "grad_norm": 2.2362422943115234, "learning_rate": 1e-06, "loss": 0.1781, "step": 1857 }, { "epoch": 0.644244105409154, "grad_norm": 2.397939443588257, "learning_rate": 1e-06, "loss": 0.1694, "step": 1858 }, { "epoch": 0.6445908460471568, "grad_norm": 2.447730779647827, "learning_rate": 1e-06, "loss": 0.1727, "step": 1859 }, { "epoch": 0.6449375866851595, "grad_norm": 2.1447441577911377, "learning_rate": 1e-06, "loss": 0.1688, "step": 1860 }, { "epoch": 0.6452843273231623, "grad_norm": 2.5228350162506104, "learning_rate": 1e-06, "loss": 0.1489, "step": 1861 }, { "epoch": 0.6456310679611651, "grad_norm": 3.115823268890381, "learning_rate": 1e-06, "loss": 0.1276, "step": 1862 }, { "epoch": 0.6459778085991679, "grad_norm": 1.9809269905090332, "learning_rate": 1e-06, "loss": 0.1688, "step": 1863 }, { "epoch": 0.6463245492371706, "grad_norm": 1.765738606452942, "learning_rate": 1e-06, "loss": 0.1619, "step": 1864 }, { "epoch": 0.6466712898751734, "grad_norm": 2.3409016132354736, "learning_rate": 1e-06, "loss": 0.1492, "step": 1865 }, { "epoch": 0.6470180305131762, "grad_norm": 2.63494873046875, "learning_rate": 1e-06, "loss": 0.155, "step": 1866 }, { "epoch": 0.647364771151179, "grad_norm": 2.3334195613861084, "learning_rate": 1e-06, "loss": 0.1476, "step": 1867 }, { "epoch": 0.6477115117891817, "grad_norm": 2.024077892303467, "learning_rate": 1e-06, "loss": 0.1419, "step": 1868 }, { "epoch": 0.6480582524271845, "grad_norm": 2.0084269046783447, "learning_rate": 1e-06, "loss": 0.152, "step": 1869 }, { "epoch": 0.6484049930651873, "grad_norm": 2.7949881553649902, "learning_rate": 1e-06, "loss": 0.2076, "step": 1870 }, { "epoch": 0.6487517337031901, "grad_norm": 4.183343887329102, "learning_rate": 1e-06, "loss": 0.1928, "step": 1871 }, { "epoch": 0.6490984743411928, "grad_norm": 3.4971086978912354, "learning_rate": 1e-06, "loss": 0.1486, "step": 1872 }, { "epoch": 0.6494452149791956, "grad_norm": 2.022676944732666, "learning_rate": 1e-06, "loss": 0.1332, "step": 1873 }, { "epoch": 0.6497919556171984, "grad_norm": 3.9578323364257812, "learning_rate": 1e-06, "loss": 0.1666, "step": 1874 }, { "epoch": 0.6501386962552012, "grad_norm": 2.3028218746185303, "learning_rate": 1e-06, "loss": 0.1662, "step": 1875 }, { "epoch": 0.6504854368932039, "grad_norm": 2.505436897277832, "learning_rate": 1e-06, "loss": 0.1291, "step": 1876 }, { "epoch": 0.6508321775312067, "grad_norm": 3.1190056800842285, "learning_rate": 1e-06, "loss": 0.1731, "step": 1877 }, { "epoch": 0.6511789181692095, "grad_norm": 2.1968798637390137, "learning_rate": 1e-06, "loss": 0.1481, "step": 1878 }, { "epoch": 0.6515256588072122, "grad_norm": 2.4867115020751953, "learning_rate": 1e-06, "loss": 0.1194, "step": 1879 }, { "epoch": 0.651872399445215, "grad_norm": 2.8347103595733643, "learning_rate": 1e-06, "loss": 0.1717, "step": 1880 }, { "epoch": 0.6522191400832178, "grad_norm": 3.8689327239990234, "learning_rate": 1e-06, "loss": 0.1226, "step": 1881 }, { "epoch": 0.6525658807212206, "grad_norm": 2.115302801132202, "learning_rate": 1e-06, "loss": 0.1491, "step": 1882 }, { "epoch": 0.6529126213592233, "grad_norm": 3.8725671768188477, "learning_rate": 1e-06, "loss": 0.1623, "step": 1883 }, { "epoch": 0.6532593619972261, "grad_norm": 2.957167863845825, "learning_rate": 1e-06, "loss": 0.1231, "step": 1884 }, { "epoch": 0.6536061026352289, "grad_norm": 2.131582021713257, "learning_rate": 1e-06, "loss": 0.1722, "step": 1885 }, { "epoch": 0.6539528432732317, "grad_norm": 2.6788721084594727, "learning_rate": 1e-06, "loss": 0.143, "step": 1886 }, { "epoch": 0.6542995839112344, "grad_norm": 2.4989116191864014, "learning_rate": 1e-06, "loss": 0.1345, "step": 1887 }, { "epoch": 0.6546463245492372, "grad_norm": 2.2083873748779297, "learning_rate": 1e-06, "loss": 0.1272, "step": 1888 }, { "epoch": 0.65499306518724, "grad_norm": 3.461536407470703, "learning_rate": 1e-06, "loss": 0.1506, "step": 1889 }, { "epoch": 0.6553398058252428, "grad_norm": 1.8483829498291016, "learning_rate": 1e-06, "loss": 0.1281, "step": 1890 }, { "epoch": 0.6556865464632455, "grad_norm": 5.3467302322387695, "learning_rate": 1e-06, "loss": 0.163, "step": 1891 }, { "epoch": 0.6560332871012483, "grad_norm": 3.6457359790802, "learning_rate": 1e-06, "loss": 0.1626, "step": 1892 }, { "epoch": 0.6563800277392511, "grad_norm": 3.4049503803253174, "learning_rate": 1e-06, "loss": 0.1452, "step": 1893 }, { "epoch": 0.6567267683772539, "grad_norm": 5.646260738372803, "learning_rate": 1e-06, "loss": 0.1739, "step": 1894 }, { "epoch": 0.6570735090152566, "grad_norm": 3.8253393173217773, "learning_rate": 1e-06, "loss": 0.1399, "step": 1895 }, { "epoch": 0.6574202496532594, "grad_norm": 3.993398904800415, "learning_rate": 1e-06, "loss": 0.1651, "step": 1896 }, { "epoch": 0.6577669902912622, "grad_norm": 4.906624794006348, "learning_rate": 1e-06, "loss": 0.109, "step": 1897 }, { "epoch": 0.658113730929265, "grad_norm": 2.4523298740386963, "learning_rate": 1e-06, "loss": 0.1374, "step": 1898 }, { "epoch": 0.6584604715672677, "grad_norm": 4.38106632232666, "learning_rate": 1e-06, "loss": 0.1601, "step": 1899 }, { "epoch": 0.6588072122052705, "grad_norm": 2.361201047897339, "learning_rate": 1e-06, "loss": 0.1647, "step": 1900 }, { "epoch": 0.6591539528432733, "grad_norm": 1.7961761951446533, "learning_rate": 1e-06, "loss": 0.1503, "step": 1901 }, { "epoch": 0.659500693481276, "grad_norm": 4.200663089752197, "learning_rate": 1e-06, "loss": 0.1362, "step": 1902 }, { "epoch": 0.6598474341192788, "grad_norm": 2.141098976135254, "learning_rate": 1e-06, "loss": 0.1287, "step": 1903 }, { "epoch": 0.6601941747572816, "grad_norm": 3.659383773803711, "learning_rate": 1e-06, "loss": 0.1679, "step": 1904 }, { "epoch": 0.6605409153952844, "grad_norm": 2.159740447998047, "learning_rate": 1e-06, "loss": 0.151, "step": 1905 }, { "epoch": 0.6608876560332871, "grad_norm": 2.612746238708496, "learning_rate": 1e-06, "loss": 0.1331, "step": 1906 }, { "epoch": 0.6612343966712899, "grad_norm": 2.067161798477173, "learning_rate": 1e-06, "loss": 0.1375, "step": 1907 }, { "epoch": 0.6615811373092927, "grad_norm": 2.2887136936187744, "learning_rate": 1e-06, "loss": 0.1445, "step": 1908 }, { "epoch": 0.6619278779472955, "grad_norm": 3.688062906265259, "learning_rate": 1e-06, "loss": 0.1943, "step": 1909 }, { "epoch": 0.6622746185852982, "grad_norm": 2.4100046157836914, "learning_rate": 1e-06, "loss": 0.1798, "step": 1910 }, { "epoch": 0.662621359223301, "grad_norm": 1.7269797325134277, "learning_rate": 1e-06, "loss": 0.1404, "step": 1911 }, { "epoch": 0.6629680998613038, "grad_norm": 1.8225007057189941, "learning_rate": 1e-06, "loss": 0.1252, "step": 1912 }, { "epoch": 0.6633148404993066, "grad_norm": 2.511077404022217, "learning_rate": 1e-06, "loss": 0.1458, "step": 1913 }, { "epoch": 0.6636615811373093, "grad_norm": 4.242014408111572, "learning_rate": 1e-06, "loss": 0.1833, "step": 1914 }, { "epoch": 0.6640083217753121, "grad_norm": 2.4541306495666504, "learning_rate": 1e-06, "loss": 0.1639, "step": 1915 }, { "epoch": 0.6643550624133149, "grad_norm": 4.093621730804443, "learning_rate": 1e-06, "loss": 0.1533, "step": 1916 }, { "epoch": 0.6647018030513177, "grad_norm": 1.6138478517532349, "learning_rate": 1e-06, "loss": 0.1365, "step": 1917 }, { "epoch": 0.6650485436893204, "grad_norm": 2.1072335243225098, "learning_rate": 1e-06, "loss": 0.1502, "step": 1918 }, { "epoch": 0.6653952843273232, "grad_norm": 4.632135391235352, "learning_rate": 1e-06, "loss": 0.1454, "step": 1919 }, { "epoch": 0.665742024965326, "grad_norm": 2.38916015625, "learning_rate": 1e-06, "loss": 0.1426, "step": 1920 }, { "epoch": 0.6660887656033287, "grad_norm": 1.7111142873764038, "learning_rate": 1e-06, "loss": 0.1491, "step": 1921 }, { "epoch": 0.6664355062413315, "grad_norm": 1.6361645460128784, "learning_rate": 1e-06, "loss": 0.1181, "step": 1922 }, { "epoch": 0.6667822468793343, "grad_norm": 2.845128297805786, "learning_rate": 1e-06, "loss": 0.1508, "step": 1923 }, { "epoch": 0.6671289875173371, "grad_norm": 3.6008825302124023, "learning_rate": 1e-06, "loss": 0.1773, "step": 1924 }, { "epoch": 0.6674757281553398, "grad_norm": 2.7898175716400146, "learning_rate": 1e-06, "loss": 0.1485, "step": 1925 }, { "epoch": 0.6678224687933426, "grad_norm": 2.1968812942504883, "learning_rate": 1e-06, "loss": 0.1538, "step": 1926 }, { "epoch": 0.6681692094313454, "grad_norm": 3.2181034088134766, "learning_rate": 1e-06, "loss": 0.1637, "step": 1927 }, { "epoch": 0.6685159500693482, "grad_norm": 3.3751060962677, "learning_rate": 1e-06, "loss": 0.1967, "step": 1928 }, { "epoch": 0.6688626907073509, "grad_norm": 4.066105842590332, "learning_rate": 1e-06, "loss": 0.1598, "step": 1929 }, { "epoch": 0.6692094313453537, "grad_norm": 3.0509390830993652, "learning_rate": 1e-06, "loss": 0.2023, "step": 1930 }, { "epoch": 0.6695561719833565, "grad_norm": 2.0706655979156494, "learning_rate": 1e-06, "loss": 0.1549, "step": 1931 }, { "epoch": 0.6699029126213593, "grad_norm": 2.6645240783691406, "learning_rate": 1e-06, "loss": 0.1451, "step": 1932 }, { "epoch": 0.670249653259362, "grad_norm": 2.8021857738494873, "learning_rate": 1e-06, "loss": 0.1238, "step": 1933 }, { "epoch": 0.6705963938973648, "grad_norm": 2.668379068374634, "learning_rate": 1e-06, "loss": 0.1606, "step": 1934 }, { "epoch": 0.6709431345353676, "grad_norm": 2.5608723163604736, "learning_rate": 1e-06, "loss": 0.1425, "step": 1935 }, { "epoch": 0.6712898751733704, "grad_norm": 2.992034435272217, "learning_rate": 1e-06, "loss": 0.1378, "step": 1936 }, { "epoch": 0.6716366158113731, "grad_norm": 1.993120789527893, "learning_rate": 1e-06, "loss": 0.1504, "step": 1937 }, { "epoch": 0.6719833564493759, "grad_norm": 1.990553379058838, "learning_rate": 1e-06, "loss": 0.1648, "step": 1938 }, { "epoch": 0.6723300970873787, "grad_norm": 3.997328758239746, "learning_rate": 1e-06, "loss": 0.1319, "step": 1939 }, { "epoch": 0.6726768377253814, "grad_norm": 3.1100356578826904, "learning_rate": 1e-06, "loss": 0.1824, "step": 1940 }, { "epoch": 0.6730235783633842, "grad_norm": 1.7388405799865723, "learning_rate": 1e-06, "loss": 0.1388, "step": 1941 }, { "epoch": 0.673370319001387, "grad_norm": 2.0104892253875732, "learning_rate": 1e-06, "loss": 0.1537, "step": 1942 }, { "epoch": 0.6737170596393898, "grad_norm": 4.981960296630859, "learning_rate": 1e-06, "loss": 0.1823, "step": 1943 }, { "epoch": 0.6740638002773925, "grad_norm": 1.5539988279342651, "learning_rate": 1e-06, "loss": 0.1507, "step": 1944 }, { "epoch": 0.6744105409153953, "grad_norm": 3.975370168685913, "learning_rate": 1e-06, "loss": 0.1869, "step": 1945 }, { "epoch": 0.6747572815533981, "grad_norm": 2.3512701988220215, "learning_rate": 1e-06, "loss": 0.1363, "step": 1946 }, { "epoch": 0.6751040221914009, "grad_norm": 1.7181259393692017, "learning_rate": 1e-06, "loss": 0.1472, "step": 1947 }, { "epoch": 0.6754507628294036, "grad_norm": 1.7866528034210205, "learning_rate": 1e-06, "loss": 0.1358, "step": 1948 }, { "epoch": 0.6757975034674064, "grad_norm": 2.635709524154663, "learning_rate": 1e-06, "loss": 0.1659, "step": 1949 }, { "epoch": 0.6761442441054092, "grad_norm": 1.9648034572601318, "learning_rate": 1e-06, "loss": 0.1443, "step": 1950 }, { "epoch": 0.676490984743412, "grad_norm": 2.1203794479370117, "learning_rate": 1e-06, "loss": 0.1283, "step": 1951 }, { "epoch": 0.6768377253814147, "grad_norm": 2.54052996635437, "learning_rate": 1e-06, "loss": 0.1447, "step": 1952 }, { "epoch": 0.6771844660194175, "grad_norm": 2.9203107357025146, "learning_rate": 1e-06, "loss": 0.1362, "step": 1953 }, { "epoch": 0.6775312066574203, "grad_norm": 2.1829326152801514, "learning_rate": 1e-06, "loss": 0.1267, "step": 1954 }, { "epoch": 0.6778779472954231, "grad_norm": 2.608271837234497, "learning_rate": 1e-06, "loss": 0.1673, "step": 1955 }, { "epoch": 0.6782246879334258, "grad_norm": 2.003397226333618, "learning_rate": 1e-06, "loss": 0.1687, "step": 1956 }, { "epoch": 0.6785714285714286, "grad_norm": 2.5364773273468018, "learning_rate": 1e-06, "loss": 0.1331, "step": 1957 }, { "epoch": 0.6789181692094314, "grad_norm": 2.5657753944396973, "learning_rate": 1e-06, "loss": 0.1485, "step": 1958 }, { "epoch": 0.6792649098474342, "grad_norm": 2.2307422161102295, "learning_rate": 1e-06, "loss": 0.16, "step": 1959 }, { "epoch": 0.6796116504854369, "grad_norm": 2.0911974906921387, "learning_rate": 1e-06, "loss": 0.1338, "step": 1960 }, { "epoch": 0.6799583911234397, "grad_norm": 3.5447793006896973, "learning_rate": 1e-06, "loss": 0.1949, "step": 1961 }, { "epoch": 0.6803051317614425, "grad_norm": 2.7882168292999268, "learning_rate": 1e-06, "loss": 0.1912, "step": 1962 }, { "epoch": 0.6806518723994452, "grad_norm": 1.985961675643921, "learning_rate": 1e-06, "loss": 0.1228, "step": 1963 }, { "epoch": 0.680998613037448, "grad_norm": 4.209602355957031, "learning_rate": 1e-06, "loss": 0.1476, "step": 1964 }, { "epoch": 0.6813453536754508, "grad_norm": 2.3590171337127686, "learning_rate": 1e-06, "loss": 0.1246, "step": 1965 }, { "epoch": 0.6816920943134536, "grad_norm": 2.680187225341797, "learning_rate": 1e-06, "loss": 0.1812, "step": 1966 }, { "epoch": 0.6820388349514563, "grad_norm": 2.99859881401062, "learning_rate": 1e-06, "loss": 0.1483, "step": 1967 }, { "epoch": 0.6823855755894591, "grad_norm": 2.140812635421753, "learning_rate": 1e-06, "loss": 0.1315, "step": 1968 }, { "epoch": 0.6827323162274619, "grad_norm": 2.1156227588653564, "learning_rate": 1e-06, "loss": 0.1576, "step": 1969 }, { "epoch": 0.6830790568654647, "grad_norm": 2.218289613723755, "learning_rate": 1e-06, "loss": 0.1378, "step": 1970 }, { "epoch": 0.6834257975034674, "grad_norm": 2.0559825897216797, "learning_rate": 1e-06, "loss": 0.158, "step": 1971 }, { "epoch": 0.6837725381414702, "grad_norm": 2.555392026901245, "learning_rate": 1e-06, "loss": 0.1377, "step": 1972 }, { "epoch": 0.684119278779473, "grad_norm": 2.068892240524292, "learning_rate": 1e-06, "loss": 0.134, "step": 1973 }, { "epoch": 0.6844660194174758, "grad_norm": 1.8548657894134521, "learning_rate": 1e-06, "loss": 0.1438, "step": 1974 }, { "epoch": 0.6848127600554785, "grad_norm": 5.5719499588012695, "learning_rate": 1e-06, "loss": 0.1765, "step": 1975 }, { "epoch": 0.6851595006934813, "grad_norm": 3.041820526123047, "learning_rate": 1e-06, "loss": 0.129, "step": 1976 }, { "epoch": 0.6855062413314841, "grad_norm": 7.963346004486084, "learning_rate": 1e-06, "loss": 0.2188, "step": 1977 }, { "epoch": 0.6858529819694869, "grad_norm": 2.143159866333008, "learning_rate": 1e-06, "loss": 0.1669, "step": 1978 }, { "epoch": 0.6861997226074896, "grad_norm": 2.7842423915863037, "learning_rate": 1e-06, "loss": 0.1568, "step": 1979 }, { "epoch": 0.6865464632454924, "grad_norm": 2.4829189777374268, "learning_rate": 1e-06, "loss": 0.1609, "step": 1980 }, { "epoch": 0.6868932038834952, "grad_norm": 2.9691684246063232, "learning_rate": 1e-06, "loss": 0.1566, "step": 1981 }, { "epoch": 0.687239944521498, "grad_norm": 1.885637640953064, "learning_rate": 1e-06, "loss": 0.1251, "step": 1982 }, { "epoch": 0.6875866851595007, "grad_norm": 3.081054210662842, "learning_rate": 1e-06, "loss": 0.1216, "step": 1983 }, { "epoch": 0.6879334257975035, "grad_norm": 2.235093832015991, "learning_rate": 1e-06, "loss": 0.1424, "step": 1984 }, { "epoch": 0.6882801664355063, "grad_norm": 2.051729679107666, "learning_rate": 1e-06, "loss": 0.1411, "step": 1985 }, { "epoch": 0.688626907073509, "grad_norm": 3.0905375480651855, "learning_rate": 1e-06, "loss": 0.1706, "step": 1986 }, { "epoch": 0.6889736477115118, "grad_norm": 2.348816156387329, "learning_rate": 1e-06, "loss": 0.1609, "step": 1987 }, { "epoch": 0.6893203883495146, "grad_norm": 3.2591960430145264, "learning_rate": 1e-06, "loss": 0.189, "step": 1988 }, { "epoch": 0.6896671289875174, "grad_norm": 5.137045860290527, "learning_rate": 1e-06, "loss": 0.1329, "step": 1989 }, { "epoch": 0.6900138696255201, "grad_norm": 3.611876964569092, "learning_rate": 1e-06, "loss": 0.13, "step": 1990 }, { "epoch": 0.6903606102635229, "grad_norm": 3.3672244548797607, "learning_rate": 1e-06, "loss": 0.1583, "step": 1991 }, { "epoch": 0.6907073509015257, "grad_norm": 2.499140501022339, "learning_rate": 1e-06, "loss": 0.12, "step": 1992 }, { "epoch": 0.6910540915395285, "grad_norm": 2.993394136428833, "learning_rate": 1e-06, "loss": 0.1329, "step": 1993 }, { "epoch": 0.6914008321775312, "grad_norm": 3.3767669200897217, "learning_rate": 1e-06, "loss": 0.1607, "step": 1994 }, { "epoch": 0.691747572815534, "grad_norm": 4.418931484222412, "learning_rate": 1e-06, "loss": 0.17, "step": 1995 }, { "epoch": 0.6920943134535368, "grad_norm": 2.544867753982544, "learning_rate": 1e-06, "loss": 0.1642, "step": 1996 }, { "epoch": 0.6924410540915396, "grad_norm": 4.367650508880615, "learning_rate": 1e-06, "loss": 0.161, "step": 1997 }, { "epoch": 0.6927877947295423, "grad_norm": 2.103868007659912, "learning_rate": 1e-06, "loss": 0.1351, "step": 1998 }, { "epoch": 0.6931345353675451, "grad_norm": 2.2191128730773926, "learning_rate": 1e-06, "loss": 0.1622, "step": 1999 }, { "epoch": 0.6934812760055479, "grad_norm": 2.239701747894287, "learning_rate": 1e-06, "loss": 0.1547, "step": 2000 }, { "epoch": 0.6938280166435506, "grad_norm": 2.166487216949463, "learning_rate": 1e-06, "loss": 0.1431, "step": 2001 }, { "epoch": 0.6941747572815534, "grad_norm": 2.3545658588409424, "learning_rate": 1e-06, "loss": 0.1271, "step": 2002 }, { "epoch": 0.6945214979195562, "grad_norm": 2.040910005569458, "learning_rate": 1e-06, "loss": 0.1502, "step": 2003 }, { "epoch": 0.694868238557559, "grad_norm": 2.179579973220825, "learning_rate": 1e-06, "loss": 0.1385, "step": 2004 }, { "epoch": 0.6952149791955617, "grad_norm": 2.0149500370025635, "learning_rate": 1e-06, "loss": 0.1396, "step": 2005 }, { "epoch": 0.6955617198335645, "grad_norm": 3.165092706680298, "learning_rate": 1e-06, "loss": 0.1217, "step": 2006 }, { "epoch": 0.6959084604715673, "grad_norm": 2.45447039604187, "learning_rate": 1e-06, "loss": 0.1455, "step": 2007 }, { "epoch": 0.6962552011095701, "grad_norm": 4.395773410797119, "learning_rate": 1e-06, "loss": 0.1568, "step": 2008 }, { "epoch": 0.6966019417475728, "grad_norm": 4.389420986175537, "learning_rate": 1e-06, "loss": 0.1294, "step": 2009 }, { "epoch": 0.6969486823855756, "grad_norm": 1.8730595111846924, "learning_rate": 1e-06, "loss": 0.1371, "step": 2010 }, { "epoch": 0.6972954230235784, "grad_norm": 2.394237518310547, "learning_rate": 1e-06, "loss": 0.1588, "step": 2011 }, { "epoch": 0.6976421636615812, "grad_norm": 2.2952566146850586, "learning_rate": 1e-06, "loss": 0.1362, "step": 2012 }, { "epoch": 0.6979889042995839, "grad_norm": 3.177366018295288, "learning_rate": 1e-06, "loss": 0.1169, "step": 2013 }, { "epoch": 0.6983356449375867, "grad_norm": 2.6569299697875977, "learning_rate": 1e-06, "loss": 0.1503, "step": 2014 }, { "epoch": 0.6986823855755895, "grad_norm": 2.000425338745117, "learning_rate": 1e-06, "loss": 0.1399, "step": 2015 }, { "epoch": 0.6990291262135923, "grad_norm": 3.407623529434204, "learning_rate": 1e-06, "loss": 0.1704, "step": 2016 }, { "epoch": 0.699375866851595, "grad_norm": 2.73795747756958, "learning_rate": 1e-06, "loss": 0.1476, "step": 2017 }, { "epoch": 0.6997226074895978, "grad_norm": 3.8470706939697266, "learning_rate": 1e-06, "loss": 0.1694, "step": 2018 }, { "epoch": 0.7000693481276006, "grad_norm": 2.0070230960845947, "learning_rate": 1e-06, "loss": 0.1629, "step": 2019 }, { "epoch": 0.7004160887656034, "grad_norm": 2.150434970855713, "learning_rate": 1e-06, "loss": 0.1228, "step": 2020 }, { "epoch": 0.7007628294036061, "grad_norm": 1.901470422744751, "learning_rate": 1e-06, "loss": 0.114, "step": 2021 }, { "epoch": 0.7011095700416089, "grad_norm": 2.0084474086761475, "learning_rate": 1e-06, "loss": 0.1431, "step": 2022 }, { "epoch": 0.7014563106796117, "grad_norm": 2.262711763381958, "learning_rate": 1e-06, "loss": 0.1298, "step": 2023 }, { "epoch": 0.7018030513176144, "grad_norm": 2.019993782043457, "learning_rate": 1e-06, "loss": 0.1568, "step": 2024 }, { "epoch": 0.7021497919556172, "grad_norm": 3.4103362560272217, "learning_rate": 1e-06, "loss": 0.1283, "step": 2025 }, { "epoch": 0.70249653259362, "grad_norm": 2.4157567024230957, "learning_rate": 1e-06, "loss": 0.1739, "step": 2026 }, { "epoch": 0.7028432732316228, "grad_norm": 2.405488967895508, "learning_rate": 1e-06, "loss": 0.1439, "step": 2027 }, { "epoch": 0.7031900138696255, "grad_norm": 1.8161323070526123, "learning_rate": 1e-06, "loss": 0.1458, "step": 2028 }, { "epoch": 0.7035367545076283, "grad_norm": 4.244541168212891, "learning_rate": 1e-06, "loss": 0.1481, "step": 2029 }, { "epoch": 0.7038834951456311, "grad_norm": 2.1027145385742188, "learning_rate": 1e-06, "loss": 0.1376, "step": 2030 }, { "epoch": 0.7042302357836339, "grad_norm": 1.911086916923523, "learning_rate": 1e-06, "loss": 0.1393, "step": 2031 }, { "epoch": 0.7045769764216366, "grad_norm": 3.033745527267456, "learning_rate": 1e-06, "loss": 0.1313, "step": 2032 }, { "epoch": 0.7049237170596394, "grad_norm": 2.4053955078125, "learning_rate": 1e-06, "loss": 0.1483, "step": 2033 }, { "epoch": 0.7052704576976422, "grad_norm": 1.796697735786438, "learning_rate": 1e-06, "loss": 0.1207, "step": 2034 }, { "epoch": 0.705617198335645, "grad_norm": 2.3117001056671143, "learning_rate": 1e-06, "loss": 0.1391, "step": 2035 }, { "epoch": 0.7059639389736477, "grad_norm": 3.2871193885803223, "learning_rate": 1e-06, "loss": 0.169, "step": 2036 }, { "epoch": 0.7063106796116505, "grad_norm": 2.3646295070648193, "learning_rate": 1e-06, "loss": 0.1827, "step": 2037 }, { "epoch": 0.7066574202496533, "grad_norm": 1.7376772165298462, "learning_rate": 1e-06, "loss": 0.1375, "step": 2038 }, { "epoch": 0.707004160887656, "grad_norm": 2.200528383255005, "learning_rate": 1e-06, "loss": 0.1417, "step": 2039 }, { "epoch": 0.7073509015256588, "grad_norm": 2.742612838745117, "learning_rate": 1e-06, "loss": 0.156, "step": 2040 }, { "epoch": 0.7076976421636616, "grad_norm": 2.012239694595337, "learning_rate": 1e-06, "loss": 0.1768, "step": 2041 }, { "epoch": 0.7080443828016644, "grad_norm": 1.6217942237854004, "learning_rate": 1e-06, "loss": 0.1093, "step": 2042 }, { "epoch": 0.7083911234396671, "grad_norm": 3.5539634227752686, "learning_rate": 1e-06, "loss": 0.1688, "step": 2043 }, { "epoch": 0.7087378640776699, "grad_norm": 2.4913277626037598, "learning_rate": 1e-06, "loss": 0.1956, "step": 2044 }, { "epoch": 0.7090846047156727, "grad_norm": 1.9440995454788208, "learning_rate": 1e-06, "loss": 0.1648, "step": 2045 }, { "epoch": 0.7094313453536755, "grad_norm": 2.7142701148986816, "learning_rate": 1e-06, "loss": 0.1566, "step": 2046 }, { "epoch": 0.7097780859916782, "grad_norm": 2.7249319553375244, "learning_rate": 1e-06, "loss": 0.1664, "step": 2047 }, { "epoch": 0.710124826629681, "grad_norm": 2.3271358013153076, "learning_rate": 1e-06, "loss": 0.1705, "step": 2048 }, { "epoch": 0.7104715672676838, "grad_norm": 1.7891216278076172, "learning_rate": 1e-06, "loss": 0.1293, "step": 2049 }, { "epoch": 0.7108183079056866, "grad_norm": 2.274380922317505, "learning_rate": 1e-06, "loss": 0.1559, "step": 2050 }, { "epoch": 0.7111650485436893, "grad_norm": 2.742744207382202, "learning_rate": 1e-06, "loss": 0.1763, "step": 2051 }, { "epoch": 0.7115117891816921, "grad_norm": 1.602330207824707, "learning_rate": 1e-06, "loss": 0.1516, "step": 2052 }, { "epoch": 0.7118585298196949, "grad_norm": 1.677738904953003, "learning_rate": 1e-06, "loss": 0.1205, "step": 2053 }, { "epoch": 0.7122052704576977, "grad_norm": 2.515608549118042, "learning_rate": 1e-06, "loss": 0.1495, "step": 2054 }, { "epoch": 0.7125520110957004, "grad_norm": 2.719072103500366, "learning_rate": 1e-06, "loss": 0.142, "step": 2055 }, { "epoch": 0.7128987517337032, "grad_norm": 2.3802108764648438, "learning_rate": 1e-06, "loss": 0.1374, "step": 2056 }, { "epoch": 0.713245492371706, "grad_norm": 3.3662099838256836, "learning_rate": 1e-06, "loss": 0.1593, "step": 2057 }, { "epoch": 0.7135922330097088, "grad_norm": 2.0012762546539307, "learning_rate": 1e-06, "loss": 0.1675, "step": 2058 }, { "epoch": 0.7139389736477115, "grad_norm": 1.8083027601242065, "learning_rate": 1e-06, "loss": 0.1215, "step": 2059 }, { "epoch": 0.7142857142857143, "grad_norm": 3.830106735229492, "learning_rate": 1e-06, "loss": 0.1446, "step": 2060 }, { "epoch": 0.7146324549237171, "grad_norm": 2.1997570991516113, "learning_rate": 1e-06, "loss": 0.1455, "step": 2061 }, { "epoch": 0.7149791955617198, "grad_norm": 1.8245985507965088, "learning_rate": 1e-06, "loss": 0.1576, "step": 2062 }, { "epoch": 0.7153259361997226, "grad_norm": 2.8333470821380615, "learning_rate": 1e-06, "loss": 0.1414, "step": 2063 }, { "epoch": 0.7156726768377254, "grad_norm": 1.8098995685577393, "learning_rate": 1e-06, "loss": 0.1155, "step": 2064 }, { "epoch": 0.7160194174757282, "grad_norm": 2.2318859100341797, "learning_rate": 1e-06, "loss": 0.145, "step": 2065 }, { "epoch": 0.7163661581137309, "grad_norm": 3.0203652381896973, "learning_rate": 1e-06, "loss": 0.1889, "step": 2066 }, { "epoch": 0.7167128987517337, "grad_norm": 3.068648099899292, "learning_rate": 1e-06, "loss": 0.1269, "step": 2067 }, { "epoch": 0.7170596393897365, "grad_norm": 2.7916111946105957, "learning_rate": 1e-06, "loss": 0.156, "step": 2068 }, { "epoch": 0.7174063800277393, "grad_norm": 1.9404290914535522, "learning_rate": 1e-06, "loss": 0.1426, "step": 2069 }, { "epoch": 0.717753120665742, "grad_norm": 2.287555694580078, "learning_rate": 1e-06, "loss": 0.121, "step": 2070 }, { "epoch": 0.7180998613037448, "grad_norm": 2.1457648277282715, "learning_rate": 1e-06, "loss": 0.1232, "step": 2071 }, { "epoch": 0.7184466019417476, "grad_norm": 4.4433441162109375, "learning_rate": 1e-06, "loss": 0.1315, "step": 2072 }, { "epoch": 0.7187933425797504, "grad_norm": 3.2125916481018066, "learning_rate": 1e-06, "loss": 0.1623, "step": 2073 }, { "epoch": 0.7191400832177531, "grad_norm": 2.4481723308563232, "learning_rate": 1e-06, "loss": 0.1458, "step": 2074 }, { "epoch": 0.7194868238557559, "grad_norm": 2.894010543823242, "learning_rate": 1e-06, "loss": 0.1623, "step": 2075 }, { "epoch": 0.7198335644937587, "grad_norm": 2.1684398651123047, "learning_rate": 1e-06, "loss": 0.1323, "step": 2076 }, { "epoch": 0.7201803051317615, "grad_norm": 2.481157064437866, "learning_rate": 1e-06, "loss": 0.1576, "step": 2077 }, { "epoch": 0.7205270457697642, "grad_norm": 2.519806385040283, "learning_rate": 1e-06, "loss": 0.1507, "step": 2078 }, { "epoch": 0.720873786407767, "grad_norm": 2.813413381576538, "learning_rate": 1e-06, "loss": 0.143, "step": 2079 }, { "epoch": 0.7212205270457698, "grad_norm": 2.9213216304779053, "learning_rate": 1e-06, "loss": 0.157, "step": 2080 }, { "epoch": 0.7215672676837726, "grad_norm": 2.8391432762145996, "learning_rate": 1e-06, "loss": 0.1652, "step": 2081 }, { "epoch": 0.7219140083217753, "grad_norm": 2.9582345485687256, "learning_rate": 1e-06, "loss": 0.1511, "step": 2082 }, { "epoch": 0.7222607489597781, "grad_norm": 2.236300230026245, "learning_rate": 1e-06, "loss": 0.1272, "step": 2083 }, { "epoch": 0.7226074895977809, "grad_norm": 2.000270366668701, "learning_rate": 1e-06, "loss": 0.1579, "step": 2084 }, { "epoch": 0.7229542302357836, "grad_norm": 2.317187547683716, "learning_rate": 1e-06, "loss": 0.16, "step": 2085 }, { "epoch": 0.7233009708737864, "grad_norm": 2.866093397140503, "learning_rate": 1e-06, "loss": 0.1539, "step": 2086 }, { "epoch": 0.7236477115117892, "grad_norm": 3.5143120288848877, "learning_rate": 1e-06, "loss": 0.1434, "step": 2087 }, { "epoch": 0.723994452149792, "grad_norm": 3.5509307384490967, "learning_rate": 1e-06, "loss": 0.1357, "step": 2088 }, { "epoch": 0.7243411927877947, "grad_norm": 1.9065073728561401, "learning_rate": 1e-06, "loss": 0.1227, "step": 2089 }, { "epoch": 0.7246879334257975, "grad_norm": 2.9716525077819824, "learning_rate": 1e-06, "loss": 0.1538, "step": 2090 }, { "epoch": 0.7250346740638003, "grad_norm": 2.858944892883301, "learning_rate": 1e-06, "loss": 0.1333, "step": 2091 }, { "epoch": 0.7253814147018031, "grad_norm": 4.433228969573975, "learning_rate": 1e-06, "loss": 0.1487, "step": 2092 }, { "epoch": 0.7257281553398058, "grad_norm": 2.223188877105713, "learning_rate": 1e-06, "loss": 0.152, "step": 2093 }, { "epoch": 0.7260748959778086, "grad_norm": 2.040834665298462, "learning_rate": 1e-06, "loss": 0.1542, "step": 2094 }, { "epoch": 0.7264216366158114, "grad_norm": 2.6200249195098877, "learning_rate": 1e-06, "loss": 0.1551, "step": 2095 }, { "epoch": 0.7267683772538142, "grad_norm": 2.9047865867614746, "learning_rate": 1e-06, "loss": 0.1368, "step": 2096 }, { "epoch": 0.7271151178918169, "grad_norm": 2.617482900619507, "learning_rate": 1e-06, "loss": 0.1762, "step": 2097 }, { "epoch": 0.7274618585298197, "grad_norm": 2.3887550830841064, "learning_rate": 1e-06, "loss": 0.1652, "step": 2098 }, { "epoch": 0.7278085991678225, "grad_norm": 2.3276662826538086, "learning_rate": 1e-06, "loss": 0.1696, "step": 2099 }, { "epoch": 0.7281553398058253, "grad_norm": 2.4688990116119385, "learning_rate": 1e-06, "loss": 0.1865, "step": 2100 }, { "epoch": 0.728502080443828, "grad_norm": 3.6115245819091797, "learning_rate": 1e-06, "loss": 0.1667, "step": 2101 }, { "epoch": 0.7288488210818308, "grad_norm": 4.399360179901123, "learning_rate": 1e-06, "loss": 0.1383, "step": 2102 }, { "epoch": 0.7291955617198336, "grad_norm": 5.1829729080200195, "learning_rate": 1e-06, "loss": 0.1576, "step": 2103 }, { "epoch": 0.7295423023578363, "grad_norm": 4.441248416900635, "learning_rate": 1e-06, "loss": 0.139, "step": 2104 }, { "epoch": 0.7298890429958391, "grad_norm": 3.7266438007354736, "learning_rate": 1e-06, "loss": 0.1497, "step": 2105 }, { "epoch": 0.7302357836338419, "grad_norm": 2.000267505645752, "learning_rate": 1e-06, "loss": 0.1643, "step": 2106 }, { "epoch": 0.7305825242718447, "grad_norm": 2.909482955932617, "learning_rate": 1e-06, "loss": 0.1409, "step": 2107 }, { "epoch": 0.7309292649098474, "grad_norm": 3.0315585136413574, "learning_rate": 1e-06, "loss": 0.1822, "step": 2108 }, { "epoch": 0.7312760055478502, "grad_norm": 2.9591143131256104, "learning_rate": 1e-06, "loss": 0.1946, "step": 2109 }, { "epoch": 0.731622746185853, "grad_norm": 3.1562037467956543, "learning_rate": 1e-06, "loss": 0.1889, "step": 2110 }, { "epoch": 0.7319694868238558, "grad_norm": 3.8917124271392822, "learning_rate": 1e-06, "loss": 0.1746, "step": 2111 }, { "epoch": 0.7323162274618585, "grad_norm": 2.4440743923187256, "learning_rate": 1e-06, "loss": 0.1837, "step": 2112 }, { "epoch": 0.7326629680998613, "grad_norm": 2.5348663330078125, "learning_rate": 1e-06, "loss": 0.1625, "step": 2113 }, { "epoch": 0.7330097087378641, "grad_norm": 1.7352176904678345, "learning_rate": 1e-06, "loss": 0.1518, "step": 2114 }, { "epoch": 0.7333564493758669, "grad_norm": 2.3291268348693848, "learning_rate": 1e-06, "loss": 0.1341, "step": 2115 }, { "epoch": 0.7337031900138696, "grad_norm": 3.349982738494873, "learning_rate": 1e-06, "loss": 0.166, "step": 2116 }, { "epoch": 0.7340499306518724, "grad_norm": 2.1478614807128906, "learning_rate": 1e-06, "loss": 0.161, "step": 2117 }, { "epoch": 0.7343966712898752, "grad_norm": 2.126044750213623, "learning_rate": 1e-06, "loss": 0.1441, "step": 2118 }, { "epoch": 0.734743411927878, "grad_norm": 2.5404157638549805, "learning_rate": 1e-06, "loss": 0.164, "step": 2119 }, { "epoch": 0.7350901525658807, "grad_norm": 2.651552677154541, "learning_rate": 1e-06, "loss": 0.1599, "step": 2120 }, { "epoch": 0.7354368932038835, "grad_norm": 4.563192367553711, "learning_rate": 1e-06, "loss": 0.124, "step": 2121 }, { "epoch": 0.7357836338418863, "grad_norm": 1.917719841003418, "learning_rate": 1e-06, "loss": 0.165, "step": 2122 }, { "epoch": 0.736130374479889, "grad_norm": 3.211930274963379, "learning_rate": 1e-06, "loss": 0.1625, "step": 2123 }, { "epoch": 0.7364771151178918, "grad_norm": 2.4502320289611816, "learning_rate": 1e-06, "loss": 0.1318, "step": 2124 }, { "epoch": 0.7368238557558946, "grad_norm": 2.0151150226593018, "learning_rate": 1e-06, "loss": 0.1832, "step": 2125 }, { "epoch": 0.7371705963938974, "grad_norm": 3.2877564430236816, "learning_rate": 1e-06, "loss": 0.149, "step": 2126 }, { "epoch": 0.7375173370319001, "grad_norm": 2.4384498596191406, "learning_rate": 1e-06, "loss": 0.1717, "step": 2127 }, { "epoch": 0.7378640776699029, "grad_norm": 1.685181736946106, "learning_rate": 1e-06, "loss": 0.1361, "step": 2128 }, { "epoch": 0.7382108183079057, "grad_norm": 1.982335090637207, "learning_rate": 1e-06, "loss": 0.1598, "step": 2129 }, { "epoch": 0.7385575589459085, "grad_norm": 1.7395703792572021, "learning_rate": 1e-06, "loss": 0.1523, "step": 2130 }, { "epoch": 0.7389042995839112, "grad_norm": 4.207993507385254, "learning_rate": 1e-06, "loss": 0.1749, "step": 2131 }, { "epoch": 0.739251040221914, "grad_norm": 2.950206995010376, "learning_rate": 1e-06, "loss": 0.1748, "step": 2132 }, { "epoch": 0.7395977808599168, "grad_norm": 2.3146812915802, "learning_rate": 1e-06, "loss": 0.1379, "step": 2133 }, { "epoch": 0.7399445214979196, "grad_norm": 3.0385754108428955, "learning_rate": 1e-06, "loss": 0.149, "step": 2134 }, { "epoch": 0.7402912621359223, "grad_norm": 2.4939308166503906, "learning_rate": 1e-06, "loss": 0.1621, "step": 2135 }, { "epoch": 0.7406380027739251, "grad_norm": 3.453312397003174, "learning_rate": 1e-06, "loss": 0.1817, "step": 2136 }, { "epoch": 0.7409847434119279, "grad_norm": 1.8451257944107056, "learning_rate": 1e-06, "loss": 0.1169, "step": 2137 }, { "epoch": 0.7413314840499307, "grad_norm": 1.7449294328689575, "learning_rate": 1e-06, "loss": 0.1307, "step": 2138 }, { "epoch": 0.7416782246879334, "grad_norm": 2.023353338241577, "learning_rate": 1e-06, "loss": 0.1229, "step": 2139 }, { "epoch": 0.7420249653259362, "grad_norm": 2.990633726119995, "learning_rate": 1e-06, "loss": 0.1469, "step": 2140 }, { "epoch": 0.742371705963939, "grad_norm": 2.25154709815979, "learning_rate": 1e-06, "loss": 0.1447, "step": 2141 }, { "epoch": 0.7427184466019418, "grad_norm": 2.082965135574341, "learning_rate": 1e-06, "loss": 0.1315, "step": 2142 }, { "epoch": 0.7430651872399445, "grad_norm": 2.342172145843506, "learning_rate": 1e-06, "loss": 0.1613, "step": 2143 }, { "epoch": 0.7434119278779473, "grad_norm": 1.9159072637557983, "learning_rate": 1e-06, "loss": 0.128, "step": 2144 }, { "epoch": 0.7437586685159501, "grad_norm": 2.7071194648742676, "learning_rate": 1e-06, "loss": 0.1023, "step": 2145 }, { "epoch": 0.7441054091539528, "grad_norm": 1.955583095550537, "learning_rate": 1e-06, "loss": 0.1694, "step": 2146 }, { "epoch": 0.7444521497919556, "grad_norm": 1.6116520166397095, "learning_rate": 1e-06, "loss": 0.1452, "step": 2147 }, { "epoch": 0.7447988904299584, "grad_norm": 2.063347101211548, "learning_rate": 1e-06, "loss": 0.1365, "step": 2148 }, { "epoch": 0.7451456310679612, "grad_norm": 2.245347023010254, "learning_rate": 1e-06, "loss": 0.1437, "step": 2149 }, { "epoch": 0.7454923717059639, "grad_norm": 2.005086660385132, "learning_rate": 1e-06, "loss": 0.1031, "step": 2150 }, { "epoch": 0.7458391123439667, "grad_norm": 3.6264290809631348, "learning_rate": 1e-06, "loss": 0.1509, "step": 2151 }, { "epoch": 0.7461858529819695, "grad_norm": 3.764726161956787, "learning_rate": 1e-06, "loss": 0.1373, "step": 2152 }, { "epoch": 0.7465325936199723, "grad_norm": 5.636833667755127, "learning_rate": 1e-06, "loss": 0.1851, "step": 2153 }, { "epoch": 0.746879334257975, "grad_norm": 3.568516969680786, "learning_rate": 1e-06, "loss": 0.1337, "step": 2154 }, { "epoch": 0.7472260748959778, "grad_norm": 1.9982422590255737, "learning_rate": 1e-06, "loss": 0.1399, "step": 2155 }, { "epoch": 0.7475728155339806, "grad_norm": 2.262208938598633, "learning_rate": 1e-06, "loss": 0.1318, "step": 2156 }, { "epoch": 0.7479195561719834, "grad_norm": 2.338360071182251, "learning_rate": 1e-06, "loss": 0.1532, "step": 2157 }, { "epoch": 0.7482662968099861, "grad_norm": 2.570775032043457, "learning_rate": 1e-06, "loss": 0.157, "step": 2158 }, { "epoch": 0.7486130374479889, "grad_norm": 2.5094552040100098, "learning_rate": 1e-06, "loss": 0.1252, "step": 2159 }, { "epoch": 0.7489597780859917, "grad_norm": 3.4843361377716064, "learning_rate": 1e-06, "loss": 0.1238, "step": 2160 }, { "epoch": 0.7493065187239945, "grad_norm": 2.467364549636841, "learning_rate": 1e-06, "loss": 0.1446, "step": 2161 }, { "epoch": 0.7496532593619972, "grad_norm": 2.1688084602355957, "learning_rate": 1e-06, "loss": 0.146, "step": 2162 }, { "epoch": 0.75, "grad_norm": 2.409573793411255, "learning_rate": 1e-06, "loss": 0.1702, "step": 2163 }, { "epoch": 0.7503467406380028, "grad_norm": 3.1284704208374023, "learning_rate": 1e-06, "loss": 0.1131, "step": 2164 }, { "epoch": 0.7506934812760055, "grad_norm": 2.3599703311920166, "learning_rate": 1e-06, "loss": 0.1569, "step": 2165 }, { "epoch": 0.7510402219140083, "grad_norm": 2.9270598888397217, "learning_rate": 1e-06, "loss": 0.1723, "step": 2166 }, { "epoch": 0.7513869625520111, "grad_norm": 3.9771506786346436, "learning_rate": 1e-06, "loss": 0.1508, "step": 2167 }, { "epoch": 0.7517337031900139, "grad_norm": 2.2273499965667725, "learning_rate": 1e-06, "loss": 0.1313, "step": 2168 }, { "epoch": 0.7520804438280166, "grad_norm": 2.1485869884490967, "learning_rate": 1e-06, "loss": 0.1242, "step": 2169 }, { "epoch": 0.7524271844660194, "grad_norm": 3.504756212234497, "learning_rate": 1e-06, "loss": 0.1515, "step": 2170 }, { "epoch": 0.7527739251040222, "grad_norm": 3.7918713092803955, "learning_rate": 1e-06, "loss": 0.1426, "step": 2171 }, { "epoch": 0.753120665742025, "grad_norm": 3.0239686965942383, "learning_rate": 1e-06, "loss": 0.1645, "step": 2172 }, { "epoch": 0.7534674063800277, "grad_norm": 2.892434597015381, "learning_rate": 1e-06, "loss": 0.133, "step": 2173 }, { "epoch": 0.7538141470180305, "grad_norm": 3.0266504287719727, "learning_rate": 1e-06, "loss": 0.1423, "step": 2174 }, { "epoch": 0.7541608876560333, "grad_norm": 2.232647180557251, "learning_rate": 1e-06, "loss": 0.1101, "step": 2175 }, { "epoch": 0.7545076282940361, "grad_norm": 2.5927672386169434, "learning_rate": 1e-06, "loss": 0.1474, "step": 2176 }, { "epoch": 0.7548543689320388, "grad_norm": 4.152268409729004, "learning_rate": 1e-06, "loss": 0.1451, "step": 2177 }, { "epoch": 0.7552011095700416, "grad_norm": 2.1622960567474365, "learning_rate": 1e-06, "loss": 0.1412, "step": 2178 }, { "epoch": 0.7555478502080444, "grad_norm": 3.911947250366211, "learning_rate": 1e-06, "loss": 0.1432, "step": 2179 }, { "epoch": 0.7558945908460472, "grad_norm": 3.2365176677703857, "learning_rate": 1e-06, "loss": 0.1284, "step": 2180 }, { "epoch": 0.7562413314840499, "grad_norm": 2.945903778076172, "learning_rate": 1e-06, "loss": 0.1461, "step": 2181 }, { "epoch": 0.7565880721220527, "grad_norm": 3.0714845657348633, "learning_rate": 1e-06, "loss": 0.1857, "step": 2182 }, { "epoch": 0.7569348127600555, "grad_norm": 2.2900753021240234, "learning_rate": 1e-06, "loss": 0.1419, "step": 2183 }, { "epoch": 0.7572815533980582, "grad_norm": 2.031371593475342, "learning_rate": 1e-06, "loss": 0.1216, "step": 2184 }, { "epoch": 0.757628294036061, "grad_norm": 2.7789697647094727, "learning_rate": 1e-06, "loss": 0.1275, "step": 2185 }, { "epoch": 0.7579750346740638, "grad_norm": 2.324058771133423, "learning_rate": 1e-06, "loss": 0.1307, "step": 2186 }, { "epoch": 0.7583217753120666, "grad_norm": 3.2857003211975098, "learning_rate": 1e-06, "loss": 0.1331, "step": 2187 }, { "epoch": 0.7586685159500693, "grad_norm": 3.0428361892700195, "learning_rate": 1e-06, "loss": 0.1491, "step": 2188 }, { "epoch": 0.7590152565880721, "grad_norm": 2.756199598312378, "learning_rate": 1e-06, "loss": 0.115, "step": 2189 }, { "epoch": 0.7593619972260749, "grad_norm": 2.3134708404541016, "learning_rate": 1e-06, "loss": 0.1233, "step": 2190 }, { "epoch": 0.7597087378640777, "grad_norm": 2.598783016204834, "learning_rate": 1e-06, "loss": 0.1333, "step": 2191 }, { "epoch": 0.7600554785020804, "grad_norm": 3.0439116954803467, "learning_rate": 1e-06, "loss": 0.1508, "step": 2192 }, { "epoch": 0.7604022191400832, "grad_norm": 2.078136682510376, "learning_rate": 1e-06, "loss": 0.1192, "step": 2193 }, { "epoch": 0.760748959778086, "grad_norm": 2.642509937286377, "learning_rate": 1e-06, "loss": 0.1577, "step": 2194 }, { "epoch": 0.7610957004160888, "grad_norm": 2.7496683597564697, "learning_rate": 1e-06, "loss": 0.1658, "step": 2195 }, { "epoch": 0.7614424410540915, "grad_norm": 2.2078957557678223, "learning_rate": 1e-06, "loss": 0.1016, "step": 2196 }, { "epoch": 0.7617891816920943, "grad_norm": 1.7725173234939575, "learning_rate": 1e-06, "loss": 0.1178, "step": 2197 }, { "epoch": 0.7621359223300971, "grad_norm": 2.821902275085449, "learning_rate": 1e-06, "loss": 0.1306, "step": 2198 }, { "epoch": 0.7624826629680999, "grad_norm": 1.9677958488464355, "learning_rate": 1e-06, "loss": 0.1499, "step": 2199 }, { "epoch": 0.7628294036061026, "grad_norm": 1.9492521286010742, "learning_rate": 1e-06, "loss": 0.1408, "step": 2200 }, { "epoch": 0.7631761442441054, "grad_norm": 2.5732200145721436, "learning_rate": 1e-06, "loss": 0.1239, "step": 2201 }, { "epoch": 0.7635228848821082, "grad_norm": 5.270609378814697, "learning_rate": 1e-06, "loss": 0.1715, "step": 2202 }, { "epoch": 0.763869625520111, "grad_norm": 2.55741286277771, "learning_rate": 1e-06, "loss": 0.1183, "step": 2203 }, { "epoch": 0.7642163661581137, "grad_norm": 2.0879695415496826, "learning_rate": 1e-06, "loss": 0.1571, "step": 2204 }, { "epoch": 0.7645631067961165, "grad_norm": 3.306610584259033, "learning_rate": 1e-06, "loss": 0.1819, "step": 2205 }, { "epoch": 0.7649098474341193, "grad_norm": 5.571447849273682, "learning_rate": 1e-06, "loss": 0.1288, "step": 2206 }, { "epoch": 0.765256588072122, "grad_norm": 2.29826283454895, "learning_rate": 1e-06, "loss": 0.1707, "step": 2207 }, { "epoch": 0.7656033287101248, "grad_norm": 2.8895623683929443, "learning_rate": 1e-06, "loss": 0.1426, "step": 2208 }, { "epoch": 0.7659500693481276, "grad_norm": 2.1935150623321533, "learning_rate": 1e-06, "loss": 0.148, "step": 2209 }, { "epoch": 0.7662968099861304, "grad_norm": 2.136641502380371, "learning_rate": 1e-06, "loss": 0.1317, "step": 2210 }, { "epoch": 0.7666435506241331, "grad_norm": 2.338508367538452, "learning_rate": 1e-06, "loss": 0.143, "step": 2211 }, { "epoch": 0.7669902912621359, "grad_norm": 1.9527918100357056, "learning_rate": 1e-06, "loss": 0.129, "step": 2212 }, { "epoch": 0.7673370319001387, "grad_norm": 3.5570404529571533, "learning_rate": 1e-06, "loss": 0.1631, "step": 2213 }, { "epoch": 0.7676837725381415, "grad_norm": 1.9036303758621216, "learning_rate": 1e-06, "loss": 0.1453, "step": 2214 }, { "epoch": 0.7680305131761442, "grad_norm": 3.279881238937378, "learning_rate": 1e-06, "loss": 0.1561, "step": 2215 }, { "epoch": 0.768377253814147, "grad_norm": 3.6818034648895264, "learning_rate": 1e-06, "loss": 0.1404, "step": 2216 }, { "epoch": 0.7687239944521498, "grad_norm": 2.868618965148926, "learning_rate": 1e-06, "loss": 0.1492, "step": 2217 }, { "epoch": 0.7690707350901526, "grad_norm": 2.422971725463867, "learning_rate": 1e-06, "loss": 0.1465, "step": 2218 }, { "epoch": 0.7694174757281553, "grad_norm": 3.9464166164398193, "learning_rate": 1e-06, "loss": 0.1541, "step": 2219 }, { "epoch": 0.7697642163661581, "grad_norm": 2.302083969116211, "learning_rate": 1e-06, "loss": 0.1585, "step": 2220 }, { "epoch": 0.7701109570041609, "grad_norm": 2.771343946456909, "learning_rate": 1e-06, "loss": 0.127, "step": 2221 }, { "epoch": 0.7704576976421637, "grad_norm": 2.0313382148742676, "learning_rate": 1e-06, "loss": 0.1635, "step": 2222 }, { "epoch": 0.7708044382801664, "grad_norm": 3.974736213684082, "learning_rate": 1e-06, "loss": 0.1493, "step": 2223 }, { "epoch": 0.7711511789181692, "grad_norm": 2.2979185581207275, "learning_rate": 1e-06, "loss": 0.1205, "step": 2224 }, { "epoch": 0.771497919556172, "grad_norm": 2.6536171436309814, "learning_rate": 1e-06, "loss": 0.1431, "step": 2225 }, { "epoch": 0.7718446601941747, "grad_norm": 2.917592763900757, "learning_rate": 1e-06, "loss": 0.1473, "step": 2226 }, { "epoch": 0.7721914008321775, "grad_norm": 2.4181203842163086, "learning_rate": 1e-06, "loss": 0.1502, "step": 2227 }, { "epoch": 0.7725381414701803, "grad_norm": 2.419362783432007, "learning_rate": 1e-06, "loss": 0.1183, "step": 2228 }, { "epoch": 0.7728848821081831, "grad_norm": 2.5661659240722656, "learning_rate": 1e-06, "loss": 0.1402, "step": 2229 }, { "epoch": 0.7732316227461858, "grad_norm": 1.9720306396484375, "learning_rate": 1e-06, "loss": 0.1253, "step": 2230 }, { "epoch": 0.7735783633841886, "grad_norm": 3.7642719745635986, "learning_rate": 1e-06, "loss": 0.1552, "step": 2231 }, { "epoch": 0.7739251040221914, "grad_norm": 8.331310272216797, "learning_rate": 1e-06, "loss": 0.2239, "step": 2232 }, { "epoch": 0.7742718446601942, "grad_norm": 2.8353569507598877, "learning_rate": 1e-06, "loss": 0.1348, "step": 2233 }, { "epoch": 0.7746185852981969, "grad_norm": 3.4421985149383545, "learning_rate": 1e-06, "loss": 0.1813, "step": 2234 }, { "epoch": 0.7749653259361997, "grad_norm": 3.0186569690704346, "learning_rate": 1e-06, "loss": 0.1741, "step": 2235 }, { "epoch": 0.7753120665742025, "grad_norm": 1.918044924736023, "learning_rate": 1e-06, "loss": 0.1231, "step": 2236 }, { "epoch": 0.7756588072122053, "grad_norm": 2.8227319717407227, "learning_rate": 1e-06, "loss": 0.141, "step": 2237 }, { "epoch": 0.776005547850208, "grad_norm": 1.9563891887664795, "learning_rate": 1e-06, "loss": 0.1403, "step": 2238 }, { "epoch": 0.7763522884882108, "grad_norm": 3.1032557487487793, "learning_rate": 1e-06, "loss": 0.1455, "step": 2239 }, { "epoch": 0.7766990291262136, "grad_norm": 2.261861801147461, "learning_rate": 1e-06, "loss": 0.1397, "step": 2240 }, { "epoch": 0.7770457697642164, "grad_norm": 3.733213186264038, "learning_rate": 1e-06, "loss": 0.167, "step": 2241 }, { "epoch": 0.7773925104022191, "grad_norm": 2.189244270324707, "learning_rate": 1e-06, "loss": 0.1243, "step": 2242 }, { "epoch": 0.7777392510402219, "grad_norm": 3.7104716300964355, "learning_rate": 1e-06, "loss": 0.1295, "step": 2243 }, { "epoch": 0.7780859916782247, "grad_norm": 1.9497910737991333, "learning_rate": 1e-06, "loss": 0.1408, "step": 2244 }, { "epoch": 0.7784327323162274, "grad_norm": 3.3845322132110596, "learning_rate": 1e-06, "loss": 0.1478, "step": 2245 }, { "epoch": 0.7787794729542302, "grad_norm": 1.5930759906768799, "learning_rate": 1e-06, "loss": 0.1126, "step": 2246 }, { "epoch": 0.779126213592233, "grad_norm": 2.993636131286621, "learning_rate": 1e-06, "loss": 0.1232, "step": 2247 }, { "epoch": 0.7794729542302358, "grad_norm": 2.2873573303222656, "learning_rate": 1e-06, "loss": 0.1456, "step": 2248 }, { "epoch": 0.7798196948682385, "grad_norm": 3.0485949516296387, "learning_rate": 1e-06, "loss": 0.1277, "step": 2249 }, { "epoch": 0.7801664355062413, "grad_norm": 3.5445034503936768, "learning_rate": 1e-06, "loss": 0.145, "step": 2250 }, { "epoch": 0.7805131761442441, "grad_norm": 2.8255326747894287, "learning_rate": 1e-06, "loss": 0.1321, "step": 2251 }, { "epoch": 0.7808599167822469, "grad_norm": 2.1605257987976074, "learning_rate": 1e-06, "loss": 0.1152, "step": 2252 }, { "epoch": 0.7812066574202496, "grad_norm": 2.541396379470825, "learning_rate": 1e-06, "loss": 0.1488, "step": 2253 }, { "epoch": 0.7815533980582524, "grad_norm": 2.67930006980896, "learning_rate": 1e-06, "loss": 0.1334, "step": 2254 }, { "epoch": 0.7819001386962552, "grad_norm": 2.717165470123291, "learning_rate": 1e-06, "loss": 0.1487, "step": 2255 }, { "epoch": 0.782246879334258, "grad_norm": 2.5122649669647217, "learning_rate": 1e-06, "loss": 0.128, "step": 2256 }, { "epoch": 0.7825936199722607, "grad_norm": 3.2565906047821045, "learning_rate": 1e-06, "loss": 0.1541, "step": 2257 }, { "epoch": 0.7829403606102635, "grad_norm": 2.227794647216797, "learning_rate": 1e-06, "loss": 0.1562, "step": 2258 }, { "epoch": 0.7832871012482663, "grad_norm": 1.8468554019927979, "learning_rate": 1e-06, "loss": 0.1258, "step": 2259 }, { "epoch": 0.7836338418862691, "grad_norm": 2.6450536251068115, "learning_rate": 1e-06, "loss": 0.1218, "step": 2260 }, { "epoch": 0.7839805825242718, "grad_norm": 3.5764429569244385, "learning_rate": 1e-06, "loss": 0.1399, "step": 2261 }, { "epoch": 0.7843273231622746, "grad_norm": 3.0315048694610596, "learning_rate": 1e-06, "loss": 0.1651, "step": 2262 }, { "epoch": 0.7846740638002774, "grad_norm": 3.2928550243377686, "learning_rate": 1e-06, "loss": 0.1461, "step": 2263 }, { "epoch": 0.7850208044382802, "grad_norm": 2.096841812133789, "learning_rate": 1e-06, "loss": 0.1297, "step": 2264 }, { "epoch": 0.7853675450762829, "grad_norm": 2.753180980682373, "learning_rate": 1e-06, "loss": 0.1545, "step": 2265 }, { "epoch": 0.7857142857142857, "grad_norm": 3.1257476806640625, "learning_rate": 1e-06, "loss": 0.1371, "step": 2266 }, { "epoch": 0.7860610263522885, "grad_norm": 3.747185707092285, "learning_rate": 1e-06, "loss": 0.148, "step": 2267 }, { "epoch": 0.7864077669902912, "grad_norm": 2.7665867805480957, "learning_rate": 1e-06, "loss": 0.1316, "step": 2268 }, { "epoch": 0.786754507628294, "grad_norm": 4.6021318435668945, "learning_rate": 1e-06, "loss": 0.1613, "step": 2269 }, { "epoch": 0.7871012482662968, "grad_norm": 3.030388116836548, "learning_rate": 1e-06, "loss": 0.1333, "step": 2270 }, { "epoch": 0.7874479889042996, "grad_norm": 2.6754000186920166, "learning_rate": 1e-06, "loss": 0.1391, "step": 2271 }, { "epoch": 0.7877947295423023, "grad_norm": 3.4339709281921387, "learning_rate": 1e-06, "loss": 0.1393, "step": 2272 }, { "epoch": 0.7881414701803051, "grad_norm": 2.5777359008789062, "learning_rate": 1e-06, "loss": 0.1269, "step": 2273 }, { "epoch": 0.7884882108183079, "grad_norm": 3.1862616539001465, "learning_rate": 1e-06, "loss": 0.175, "step": 2274 }, { "epoch": 0.7888349514563107, "grad_norm": 2.623753786087036, "learning_rate": 1e-06, "loss": 0.1374, "step": 2275 }, { "epoch": 0.7891816920943134, "grad_norm": 2.4776697158813477, "learning_rate": 1e-06, "loss": 0.1435, "step": 2276 }, { "epoch": 0.7895284327323162, "grad_norm": 2.2111032009124756, "learning_rate": 1e-06, "loss": 0.1139, "step": 2277 }, { "epoch": 0.789875173370319, "grad_norm": 2.097119092941284, "learning_rate": 1e-06, "loss": 0.1246, "step": 2278 }, { "epoch": 0.7902219140083218, "grad_norm": 2.417785882949829, "learning_rate": 1e-06, "loss": 0.1433, "step": 2279 }, { "epoch": 0.7905686546463245, "grad_norm": 3.1172475814819336, "learning_rate": 1e-06, "loss": 0.1155, "step": 2280 }, { "epoch": 0.7909153952843273, "grad_norm": 2.320932626724243, "learning_rate": 1e-06, "loss": 0.1368, "step": 2281 }, { "epoch": 0.7912621359223301, "grad_norm": 1.8932015895843506, "learning_rate": 1e-06, "loss": 0.1325, "step": 2282 }, { "epoch": 0.7916088765603329, "grad_norm": 2.369478225708008, "learning_rate": 1e-06, "loss": 0.1814, "step": 2283 }, { "epoch": 0.7919556171983356, "grad_norm": 1.7178212404251099, "learning_rate": 1e-06, "loss": 0.1213, "step": 2284 }, { "epoch": 0.7923023578363384, "grad_norm": 2.493131399154663, "learning_rate": 1e-06, "loss": 0.1378, "step": 2285 }, { "epoch": 0.7926490984743412, "grad_norm": 2.1028831005096436, "learning_rate": 1e-06, "loss": 0.137, "step": 2286 }, { "epoch": 0.792995839112344, "grad_norm": 3.674328088760376, "learning_rate": 1e-06, "loss": 0.156, "step": 2287 }, { "epoch": 0.7933425797503467, "grad_norm": 2.515397310256958, "learning_rate": 1e-06, "loss": 0.1278, "step": 2288 }, { "epoch": 0.7936893203883495, "grad_norm": 2.559921979904175, "learning_rate": 1e-06, "loss": 0.1363, "step": 2289 }, { "epoch": 0.7940360610263523, "grad_norm": 3.420501232147217, "learning_rate": 1e-06, "loss": 0.1326, "step": 2290 }, { "epoch": 0.794382801664355, "grad_norm": 1.9830430746078491, "learning_rate": 1e-06, "loss": 0.1242, "step": 2291 }, { "epoch": 0.7947295423023578, "grad_norm": 1.8026798963546753, "learning_rate": 1e-06, "loss": 0.116, "step": 2292 }, { "epoch": 0.7950762829403606, "grad_norm": 2.667938232421875, "learning_rate": 1e-06, "loss": 0.14, "step": 2293 }, { "epoch": 0.7954230235783634, "grad_norm": 3.140871524810791, "learning_rate": 1e-06, "loss": 0.137, "step": 2294 }, { "epoch": 0.7957697642163661, "grad_norm": 3.2242789268493652, "learning_rate": 1e-06, "loss": 0.1559, "step": 2295 }, { "epoch": 0.7961165048543689, "grad_norm": 2.4139838218688965, "learning_rate": 1e-06, "loss": 0.1286, "step": 2296 }, { "epoch": 0.7964632454923717, "grad_norm": 1.957815408706665, "learning_rate": 1e-06, "loss": 0.1292, "step": 2297 }, { "epoch": 0.7968099861303745, "grad_norm": 4.165436267852783, "learning_rate": 1e-06, "loss": 0.1342, "step": 2298 }, { "epoch": 0.7971567267683772, "grad_norm": 2.260488748550415, "learning_rate": 1e-06, "loss": 0.126, "step": 2299 }, { "epoch": 0.79750346740638, "grad_norm": 2.7062175273895264, "learning_rate": 1e-06, "loss": 0.1285, "step": 2300 }, { "epoch": 0.7978502080443828, "grad_norm": 3.5681710243225098, "learning_rate": 1e-06, "loss": 0.1115, "step": 2301 }, { "epoch": 0.7981969486823856, "grad_norm": 4.171554088592529, "learning_rate": 1e-06, "loss": 0.12, "step": 2302 }, { "epoch": 0.7985436893203883, "grad_norm": 1.9496839046478271, "learning_rate": 1e-06, "loss": 0.1323, "step": 2303 }, { "epoch": 0.7988904299583911, "grad_norm": 2.0774519443511963, "learning_rate": 1e-06, "loss": 0.1353, "step": 2304 }, { "epoch": 0.7992371705963939, "grad_norm": 2.0676655769348145, "learning_rate": 1e-06, "loss": 0.1294, "step": 2305 }, { "epoch": 0.7995839112343966, "grad_norm": 2.6793768405914307, "learning_rate": 1e-06, "loss": 0.1349, "step": 2306 }, { "epoch": 0.7999306518723994, "grad_norm": 1.8901208639144897, "learning_rate": 1e-06, "loss": 0.1318, "step": 2307 }, { "epoch": 0.8002773925104022, "grad_norm": 2.2167181968688965, "learning_rate": 1e-06, "loss": 0.1393, "step": 2308 }, { "epoch": 0.800624133148405, "grad_norm": 3.0722689628601074, "learning_rate": 1e-06, "loss": 0.1565, "step": 2309 }, { "epoch": 0.8009708737864077, "grad_norm": 2.4346120357513428, "learning_rate": 1e-06, "loss": 0.1498, "step": 2310 }, { "epoch": 0.8013176144244105, "grad_norm": 1.9527393579483032, "learning_rate": 1e-06, "loss": 0.1459, "step": 2311 }, { "epoch": 0.8016643550624133, "grad_norm": 2.2069756984710693, "learning_rate": 1e-06, "loss": 0.166, "step": 2312 }, { "epoch": 0.8020110957004161, "grad_norm": 3.2333059310913086, "learning_rate": 1e-06, "loss": 0.1786, "step": 2313 }, { "epoch": 0.8023578363384188, "grad_norm": 3.749083995819092, "learning_rate": 1e-06, "loss": 0.1594, "step": 2314 }, { "epoch": 0.8027045769764216, "grad_norm": 2.32126522064209, "learning_rate": 1e-06, "loss": 0.1161, "step": 2315 }, { "epoch": 0.8030513176144244, "grad_norm": 3.9740304946899414, "learning_rate": 1e-06, "loss": 0.1492, "step": 2316 }, { "epoch": 0.8033980582524272, "grad_norm": 2.4023282527923584, "learning_rate": 1e-06, "loss": 0.1319, "step": 2317 }, { "epoch": 0.8037447988904299, "grad_norm": 2.1061739921569824, "learning_rate": 1e-06, "loss": 0.1239, "step": 2318 }, { "epoch": 0.8040915395284327, "grad_norm": 2.2139999866485596, "learning_rate": 1e-06, "loss": 0.1479, "step": 2319 }, { "epoch": 0.8044382801664355, "grad_norm": 2.9679105281829834, "learning_rate": 1e-06, "loss": 0.1466, "step": 2320 }, { "epoch": 0.8047850208044383, "grad_norm": 2.461606740951538, "learning_rate": 1e-06, "loss": 0.1505, "step": 2321 }, { "epoch": 0.805131761442441, "grad_norm": 2.9136569499969482, "learning_rate": 1e-06, "loss": 0.1626, "step": 2322 }, { "epoch": 0.8054785020804438, "grad_norm": 1.9064357280731201, "learning_rate": 1e-06, "loss": 0.138, "step": 2323 }, { "epoch": 0.8058252427184466, "grad_norm": 2.3706328868865967, "learning_rate": 1e-06, "loss": 0.1311, "step": 2324 }, { "epoch": 0.8061719833564494, "grad_norm": 3.0558860301971436, "learning_rate": 1e-06, "loss": 0.1171, "step": 2325 }, { "epoch": 0.8065187239944521, "grad_norm": 4.301748275756836, "learning_rate": 1e-06, "loss": 0.1781, "step": 2326 }, { "epoch": 0.8068654646324549, "grad_norm": 3.325167655944824, "learning_rate": 1e-06, "loss": 0.1443, "step": 2327 }, { "epoch": 0.8072122052704577, "grad_norm": 2.0337326526641846, "learning_rate": 1e-06, "loss": 0.1317, "step": 2328 }, { "epoch": 0.8075589459084604, "grad_norm": 2.0139780044555664, "learning_rate": 1e-06, "loss": 0.163, "step": 2329 }, { "epoch": 0.8079056865464632, "grad_norm": 2.4566049575805664, "learning_rate": 1e-06, "loss": 0.1758, "step": 2330 }, { "epoch": 0.808252427184466, "grad_norm": 2.1362626552581787, "learning_rate": 1e-06, "loss": 0.1435, "step": 2331 }, { "epoch": 0.8085991678224688, "grad_norm": 1.9229339361190796, "learning_rate": 1e-06, "loss": 0.1444, "step": 2332 }, { "epoch": 0.8089459084604715, "grad_norm": 3.5861167907714844, "learning_rate": 1e-06, "loss": 0.1486, "step": 2333 }, { "epoch": 0.8092926490984743, "grad_norm": 1.9148346185684204, "learning_rate": 1e-06, "loss": 0.1355, "step": 2334 }, { "epoch": 0.8096393897364771, "grad_norm": 2.1705756187438965, "learning_rate": 1e-06, "loss": 0.1402, "step": 2335 }, { "epoch": 0.8099861303744799, "grad_norm": 1.75710928440094, "learning_rate": 1e-06, "loss": 0.1188, "step": 2336 }, { "epoch": 0.8103328710124826, "grad_norm": 1.7081013917922974, "learning_rate": 1e-06, "loss": 0.1314, "step": 2337 }, { "epoch": 0.8106796116504854, "grad_norm": 3.3994858264923096, "learning_rate": 1e-06, "loss": 0.1017, "step": 2338 }, { "epoch": 0.8110263522884882, "grad_norm": 3.6731457710266113, "learning_rate": 1e-06, "loss": 0.1161, "step": 2339 }, { "epoch": 0.811373092926491, "grad_norm": 2.8746140003204346, "learning_rate": 1e-06, "loss": 0.1453, "step": 2340 }, { "epoch": 0.8117198335644937, "grad_norm": 2.074737310409546, "learning_rate": 1e-06, "loss": 0.1477, "step": 2341 }, { "epoch": 0.8120665742024965, "grad_norm": 3.902578592300415, "learning_rate": 1e-06, "loss": 0.1372, "step": 2342 }, { "epoch": 0.8124133148404993, "grad_norm": 2.1763041019439697, "learning_rate": 1e-06, "loss": 0.1246, "step": 2343 }, { "epoch": 0.812760055478502, "grad_norm": 3.487217903137207, "learning_rate": 1e-06, "loss": 0.1074, "step": 2344 }, { "epoch": 0.8131067961165048, "grad_norm": 1.8151427507400513, "learning_rate": 1e-06, "loss": 0.0811, "step": 2345 }, { "epoch": 0.8134535367545076, "grad_norm": 2.7768118381500244, "learning_rate": 1e-06, "loss": 0.17, "step": 2346 }, { "epoch": 0.8138002773925104, "grad_norm": 2.1280786991119385, "learning_rate": 1e-06, "loss": 0.152, "step": 2347 }, { "epoch": 0.8141470180305131, "grad_norm": 3.410444974899292, "learning_rate": 1e-06, "loss": 0.1596, "step": 2348 }, { "epoch": 0.8144937586685159, "grad_norm": 1.9879850149154663, "learning_rate": 1e-06, "loss": 0.1277, "step": 2349 }, { "epoch": 0.8148404993065187, "grad_norm": 2.854005813598633, "learning_rate": 1e-06, "loss": 0.1358, "step": 2350 }, { "epoch": 0.8151872399445215, "grad_norm": 2.091987133026123, "learning_rate": 1e-06, "loss": 0.1321, "step": 2351 }, { "epoch": 0.8155339805825242, "grad_norm": 2.376237392425537, "learning_rate": 1e-06, "loss": 0.1499, "step": 2352 }, { "epoch": 0.815880721220527, "grad_norm": 2.1092143058776855, "learning_rate": 1e-06, "loss": 0.1543, "step": 2353 }, { "epoch": 0.8162274618585298, "grad_norm": 2.8994827270507812, "learning_rate": 1e-06, "loss": 0.1463, "step": 2354 }, { "epoch": 0.8165742024965326, "grad_norm": 2.607811689376831, "learning_rate": 1e-06, "loss": 0.1171, "step": 2355 }, { "epoch": 0.8169209431345353, "grad_norm": 3.0912883281707764, "learning_rate": 1e-06, "loss": 0.1444, "step": 2356 }, { "epoch": 0.8172676837725381, "grad_norm": 2.976580858230591, "learning_rate": 1e-06, "loss": 0.1761, "step": 2357 }, { "epoch": 0.8176144244105409, "grad_norm": 2.745866060256958, "learning_rate": 1e-06, "loss": 0.1282, "step": 2358 }, { "epoch": 0.8179611650485437, "grad_norm": 2.7969820499420166, "learning_rate": 1e-06, "loss": 0.117, "step": 2359 }, { "epoch": 0.8183079056865464, "grad_norm": 2.0547497272491455, "learning_rate": 1e-06, "loss": 0.1322, "step": 2360 }, { "epoch": 0.8186546463245492, "grad_norm": 2.9290833473205566, "learning_rate": 1e-06, "loss": 0.18, "step": 2361 }, { "epoch": 0.819001386962552, "grad_norm": 2.636199474334717, "learning_rate": 1e-06, "loss": 0.1636, "step": 2362 }, { "epoch": 0.8193481276005548, "grad_norm": 2.2920427322387695, "learning_rate": 1e-06, "loss": 0.1362, "step": 2363 }, { "epoch": 0.8196948682385575, "grad_norm": 2.04429030418396, "learning_rate": 1e-06, "loss": 0.1486, "step": 2364 }, { "epoch": 0.8200416088765603, "grad_norm": 2.0373737812042236, "learning_rate": 1e-06, "loss": 0.1257, "step": 2365 }, { "epoch": 0.8203883495145631, "grad_norm": 2.8232688903808594, "learning_rate": 1e-06, "loss": 0.1549, "step": 2366 }, { "epoch": 0.8207350901525658, "grad_norm": 2.8030526638031006, "learning_rate": 1e-06, "loss": 0.1429, "step": 2367 }, { "epoch": 0.8210818307905686, "grad_norm": 3.216235399246216, "learning_rate": 1e-06, "loss": 0.1781, "step": 2368 }, { "epoch": 0.8214285714285714, "grad_norm": 2.1101338863372803, "learning_rate": 1e-06, "loss": 0.1353, "step": 2369 }, { "epoch": 0.8217753120665742, "grad_norm": 3.3085856437683105, "learning_rate": 1e-06, "loss": 0.1526, "step": 2370 }, { "epoch": 0.8221220527045769, "grad_norm": 2.0516085624694824, "learning_rate": 1e-06, "loss": 0.148, "step": 2371 }, { "epoch": 0.8224687933425797, "grad_norm": 2.2691268920898438, "learning_rate": 1e-06, "loss": 0.1306, "step": 2372 }, { "epoch": 0.8228155339805825, "grad_norm": 2.2179465293884277, "learning_rate": 1e-06, "loss": 0.1412, "step": 2373 }, { "epoch": 0.8231622746185853, "grad_norm": 2.9574997425079346, "learning_rate": 1e-06, "loss": 0.1928, "step": 2374 }, { "epoch": 0.823509015256588, "grad_norm": 2.3397562503814697, "learning_rate": 1e-06, "loss": 0.1478, "step": 2375 }, { "epoch": 0.8238557558945908, "grad_norm": 2.1710567474365234, "learning_rate": 1e-06, "loss": 0.1619, "step": 2376 }, { "epoch": 0.8242024965325936, "grad_norm": 2.402174472808838, "learning_rate": 1e-06, "loss": 0.1495, "step": 2377 }, { "epoch": 0.8245492371705964, "grad_norm": 2.147606372833252, "learning_rate": 1e-06, "loss": 0.1261, "step": 2378 }, { "epoch": 0.8248959778085991, "grad_norm": 2.674955129623413, "learning_rate": 1e-06, "loss": 0.1174, "step": 2379 }, { "epoch": 0.8252427184466019, "grad_norm": 3.098621129989624, "learning_rate": 1e-06, "loss": 0.1686, "step": 2380 }, { "epoch": 0.8255894590846047, "grad_norm": 2.411407232284546, "learning_rate": 1e-06, "loss": 0.1175, "step": 2381 }, { "epoch": 0.8259361997226075, "grad_norm": 3.399182081222534, "learning_rate": 1e-06, "loss": 0.1345, "step": 2382 }, { "epoch": 0.8262829403606102, "grad_norm": 3.7523751258850098, "learning_rate": 1e-06, "loss": 0.139, "step": 2383 }, { "epoch": 0.826629680998613, "grad_norm": 2.860729217529297, "learning_rate": 1e-06, "loss": 0.1782, "step": 2384 }, { "epoch": 0.8269764216366158, "grad_norm": 4.691393852233887, "learning_rate": 1e-06, "loss": 0.1536, "step": 2385 }, { "epoch": 0.8273231622746186, "grad_norm": 2.227851152420044, "learning_rate": 1e-06, "loss": 0.1156, "step": 2386 }, { "epoch": 0.8276699029126213, "grad_norm": 2.429328680038452, "learning_rate": 1e-06, "loss": 0.1305, "step": 2387 }, { "epoch": 0.8280166435506241, "grad_norm": 1.7986654043197632, "learning_rate": 1e-06, "loss": 0.1271, "step": 2388 }, { "epoch": 0.8283633841886269, "grad_norm": 2.666243314743042, "learning_rate": 1e-06, "loss": 0.1173, "step": 2389 }, { "epoch": 0.8287101248266296, "grad_norm": 3.4500293731689453, "learning_rate": 1e-06, "loss": 0.1979, "step": 2390 }, { "epoch": 0.8290568654646324, "grad_norm": 2.057880401611328, "learning_rate": 1e-06, "loss": 0.1349, "step": 2391 }, { "epoch": 0.8294036061026352, "grad_norm": 2.378908634185791, "learning_rate": 1e-06, "loss": 0.129, "step": 2392 }, { "epoch": 0.829750346740638, "grad_norm": 1.690531849861145, "learning_rate": 1e-06, "loss": 0.1249, "step": 2393 }, { "epoch": 0.8300970873786407, "grad_norm": 3.056180953979492, "learning_rate": 1e-06, "loss": 0.1269, "step": 2394 }, { "epoch": 0.8304438280166435, "grad_norm": 4.274582862854004, "learning_rate": 1e-06, "loss": 0.1123, "step": 2395 }, { "epoch": 0.8307905686546463, "grad_norm": 2.3732125759124756, "learning_rate": 1e-06, "loss": 0.1626, "step": 2396 }, { "epoch": 0.8311373092926491, "grad_norm": 2.0551693439483643, "learning_rate": 1e-06, "loss": 0.1381, "step": 2397 }, { "epoch": 0.8314840499306518, "grad_norm": 3.1651957035064697, "learning_rate": 1e-06, "loss": 0.143, "step": 2398 }, { "epoch": 0.8318307905686546, "grad_norm": 1.9696643352508545, "learning_rate": 1e-06, "loss": 0.1238, "step": 2399 }, { "epoch": 0.8321775312066574, "grad_norm": 2.9673922061920166, "learning_rate": 1e-06, "loss": 0.1699, "step": 2400 }, { "epoch": 0.8325242718446602, "grad_norm": 2.1757307052612305, "learning_rate": 1e-06, "loss": 0.1287, "step": 2401 }, { "epoch": 0.8328710124826629, "grad_norm": 2.689530611038208, "learning_rate": 1e-06, "loss": 0.1407, "step": 2402 }, { "epoch": 0.8332177531206657, "grad_norm": 3.808122396469116, "learning_rate": 1e-06, "loss": 0.1323, "step": 2403 }, { "epoch": 0.8335644937586685, "grad_norm": 2.106861114501953, "learning_rate": 1e-06, "loss": 0.1375, "step": 2404 }, { "epoch": 0.8339112343966713, "grad_norm": 3.0327420234680176, "learning_rate": 1e-06, "loss": 0.1323, "step": 2405 }, { "epoch": 0.834257975034674, "grad_norm": 2.0406875610351562, "learning_rate": 1e-06, "loss": 0.1349, "step": 2406 }, { "epoch": 0.8346047156726768, "grad_norm": 5.959853172302246, "learning_rate": 1e-06, "loss": 0.1603, "step": 2407 }, { "epoch": 0.8349514563106796, "grad_norm": 2.409048080444336, "learning_rate": 1e-06, "loss": 0.1401, "step": 2408 }, { "epoch": 0.8352981969486823, "grad_norm": 2.135956287384033, "learning_rate": 1e-06, "loss": 0.1406, "step": 2409 }, { "epoch": 0.8356449375866851, "grad_norm": 2.256753921508789, "learning_rate": 1e-06, "loss": 0.1348, "step": 2410 }, { "epoch": 0.8359916782246879, "grad_norm": 2.43808913230896, "learning_rate": 1e-06, "loss": 0.1495, "step": 2411 }, { "epoch": 0.8363384188626907, "grad_norm": 2.6997463703155518, "learning_rate": 1e-06, "loss": 0.1227, "step": 2412 }, { "epoch": 0.8366851595006934, "grad_norm": 3.3730976581573486, "learning_rate": 1e-06, "loss": 0.1212, "step": 2413 }, { "epoch": 0.8370319001386962, "grad_norm": 3.1687965393066406, "learning_rate": 1e-06, "loss": 0.1268, "step": 2414 }, { "epoch": 0.837378640776699, "grad_norm": 2.4090065956115723, "learning_rate": 1e-06, "loss": 0.1449, "step": 2415 }, { "epoch": 0.8377253814147018, "grad_norm": 2.7174880504608154, "learning_rate": 1e-06, "loss": 0.1413, "step": 2416 }, { "epoch": 0.8380721220527045, "grad_norm": 2.5149476528167725, "learning_rate": 1e-06, "loss": 0.1568, "step": 2417 }, { "epoch": 0.8384188626907073, "grad_norm": 4.081730842590332, "learning_rate": 1e-06, "loss": 0.1159, "step": 2418 }, { "epoch": 0.8387656033287101, "grad_norm": 2.8811569213867188, "learning_rate": 1e-06, "loss": 0.1729, "step": 2419 }, { "epoch": 0.8391123439667129, "grad_norm": 2.758857011795044, "learning_rate": 1e-06, "loss": 0.1391, "step": 2420 }, { "epoch": 0.8394590846047156, "grad_norm": 2.5798659324645996, "learning_rate": 1e-06, "loss": 0.1211, "step": 2421 }, { "epoch": 0.8398058252427184, "grad_norm": 5.093081951141357, "learning_rate": 1e-06, "loss": 0.1562, "step": 2422 }, { "epoch": 0.8401525658807212, "grad_norm": 2.7829928398132324, "learning_rate": 1e-06, "loss": 0.116, "step": 2423 }, { "epoch": 0.840499306518724, "grad_norm": 5.886650562286377, "learning_rate": 1e-06, "loss": 0.1584, "step": 2424 }, { "epoch": 0.8408460471567267, "grad_norm": 2.695272207260132, "learning_rate": 1e-06, "loss": 0.1519, "step": 2425 }, { "epoch": 0.8411927877947295, "grad_norm": 1.8741930723190308, "learning_rate": 1e-06, "loss": 0.1183, "step": 2426 }, { "epoch": 0.8415395284327323, "grad_norm": 2.3263349533081055, "learning_rate": 1e-06, "loss": 0.1567, "step": 2427 }, { "epoch": 0.841886269070735, "grad_norm": 2.279158353805542, "learning_rate": 1e-06, "loss": 0.1386, "step": 2428 }, { "epoch": 0.8422330097087378, "grad_norm": 2.275637626647949, "learning_rate": 1e-06, "loss": 0.1325, "step": 2429 }, { "epoch": 0.8425797503467406, "grad_norm": 2.705381393432617, "learning_rate": 1e-06, "loss": 0.1842, "step": 2430 }, { "epoch": 0.8429264909847434, "grad_norm": 2.2983760833740234, "learning_rate": 1e-06, "loss": 0.1288, "step": 2431 }, { "epoch": 0.8432732316227461, "grad_norm": 2.5210258960723877, "learning_rate": 1e-06, "loss": 0.1506, "step": 2432 }, { "epoch": 0.8436199722607489, "grad_norm": 2.4116249084472656, "learning_rate": 1e-06, "loss": 0.1377, "step": 2433 }, { "epoch": 0.8439667128987517, "grad_norm": 3.0181736946105957, "learning_rate": 1e-06, "loss": 0.1206, "step": 2434 }, { "epoch": 0.8443134535367545, "grad_norm": 2.6214163303375244, "learning_rate": 1e-06, "loss": 0.1807, "step": 2435 }, { "epoch": 0.8446601941747572, "grad_norm": 2.952951669692993, "learning_rate": 1e-06, "loss": 0.1392, "step": 2436 }, { "epoch": 0.84500693481276, "grad_norm": 2.5941317081451416, "learning_rate": 1e-06, "loss": 0.1359, "step": 2437 }, { "epoch": 0.8453536754507628, "grad_norm": 3.742173433303833, "learning_rate": 1e-06, "loss": 0.1257, "step": 2438 }, { "epoch": 0.8457004160887656, "grad_norm": 2.2225148677825928, "learning_rate": 1e-06, "loss": 0.1341, "step": 2439 }, { "epoch": 0.8460471567267683, "grad_norm": 2.1798903942108154, "learning_rate": 1e-06, "loss": 0.1436, "step": 2440 }, { "epoch": 0.8463938973647711, "grad_norm": 2.250204563140869, "learning_rate": 1e-06, "loss": 0.1274, "step": 2441 }, { "epoch": 0.8467406380027739, "grad_norm": 2.7729578018188477, "learning_rate": 1e-06, "loss": 0.1432, "step": 2442 }, { "epoch": 0.8470873786407767, "grad_norm": 2.45955491065979, "learning_rate": 1e-06, "loss": 0.1196, "step": 2443 }, { "epoch": 0.8474341192787794, "grad_norm": 2.97157621383667, "learning_rate": 1e-06, "loss": 0.1218, "step": 2444 }, { "epoch": 0.8477808599167822, "grad_norm": 2.3688995838165283, "learning_rate": 1e-06, "loss": 0.1171, "step": 2445 }, { "epoch": 0.848127600554785, "grad_norm": 2.0294275283813477, "learning_rate": 1e-06, "loss": 0.1428, "step": 2446 }, { "epoch": 0.8484743411927878, "grad_norm": 2.387620687484741, "learning_rate": 1e-06, "loss": 0.1597, "step": 2447 }, { "epoch": 0.8488210818307905, "grad_norm": 2.6711483001708984, "learning_rate": 1e-06, "loss": 0.1353, "step": 2448 }, { "epoch": 0.8491678224687933, "grad_norm": 1.8898664712905884, "learning_rate": 1e-06, "loss": 0.1174, "step": 2449 }, { "epoch": 0.8495145631067961, "grad_norm": 3.391063928604126, "learning_rate": 1e-06, "loss": 0.1485, "step": 2450 }, { "epoch": 0.8498613037447988, "grad_norm": 2.1252994537353516, "learning_rate": 1e-06, "loss": 0.1212, "step": 2451 }, { "epoch": 0.8502080443828016, "grad_norm": 2.0090253353118896, "learning_rate": 1e-06, "loss": 0.1171, "step": 2452 }, { "epoch": 0.8505547850208044, "grad_norm": 2.170214891433716, "learning_rate": 1e-06, "loss": 0.1487, "step": 2453 }, { "epoch": 0.8509015256588072, "grad_norm": 2.922431468963623, "learning_rate": 1e-06, "loss": 0.1148, "step": 2454 }, { "epoch": 0.8512482662968099, "grad_norm": 2.2969465255737305, "learning_rate": 1e-06, "loss": 0.1417, "step": 2455 }, { "epoch": 0.8515950069348127, "grad_norm": 2.0030033588409424, "learning_rate": 1e-06, "loss": 0.1146, "step": 2456 }, { "epoch": 0.8519417475728155, "grad_norm": 2.1828982830047607, "learning_rate": 1e-06, "loss": 0.1204, "step": 2457 }, { "epoch": 0.8522884882108183, "grad_norm": 1.8292583227157593, "learning_rate": 1e-06, "loss": 0.1263, "step": 2458 }, { "epoch": 0.852635228848821, "grad_norm": 2.6252031326293945, "learning_rate": 1e-06, "loss": 0.1443, "step": 2459 }, { "epoch": 0.8529819694868238, "grad_norm": 2.30623459815979, "learning_rate": 1e-06, "loss": 0.1436, "step": 2460 }, { "epoch": 0.8533287101248266, "grad_norm": 2.8300130367279053, "learning_rate": 1e-06, "loss": 0.1644, "step": 2461 }, { "epoch": 0.8536754507628294, "grad_norm": 2.417668104171753, "learning_rate": 1e-06, "loss": 0.1432, "step": 2462 }, { "epoch": 0.8540221914008321, "grad_norm": 2.5126116275787354, "learning_rate": 1e-06, "loss": 0.1228, "step": 2463 }, { "epoch": 0.8543689320388349, "grad_norm": 2.0431835651397705, "learning_rate": 1e-06, "loss": 0.135, "step": 2464 }, { "epoch": 0.8547156726768377, "grad_norm": 3.008413076400757, "learning_rate": 1e-06, "loss": 0.1218, "step": 2465 }, { "epoch": 0.8550624133148405, "grad_norm": 2.4672775268554688, "learning_rate": 1e-06, "loss": 0.157, "step": 2466 }, { "epoch": 0.8554091539528432, "grad_norm": 3.271653652191162, "learning_rate": 1e-06, "loss": 0.1474, "step": 2467 }, { "epoch": 0.855755894590846, "grad_norm": 2.1060640811920166, "learning_rate": 1e-06, "loss": 0.112, "step": 2468 }, { "epoch": 0.8561026352288488, "grad_norm": 1.9942084550857544, "learning_rate": 1e-06, "loss": 0.1288, "step": 2469 }, { "epoch": 0.8564493758668515, "grad_norm": 2.6707653999328613, "learning_rate": 1e-06, "loss": 0.1613, "step": 2470 }, { "epoch": 0.8567961165048543, "grad_norm": 3.239352226257324, "learning_rate": 1e-06, "loss": 0.1444, "step": 2471 }, { "epoch": 0.8571428571428571, "grad_norm": 2.7429757118225098, "learning_rate": 1e-06, "loss": 0.1516, "step": 2472 }, { "epoch": 0.8574895977808599, "grad_norm": 2.1264328956604004, "learning_rate": 1e-06, "loss": 0.1657, "step": 2473 }, { "epoch": 0.8578363384188626, "grad_norm": 2.4109182357788086, "learning_rate": 1e-06, "loss": 0.1331, "step": 2474 }, { "epoch": 0.8581830790568654, "grad_norm": 2.9945809841156006, "learning_rate": 1e-06, "loss": 0.1349, "step": 2475 }, { "epoch": 0.8585298196948682, "grad_norm": 2.7945027351379395, "learning_rate": 1e-06, "loss": 0.151, "step": 2476 }, { "epoch": 0.858876560332871, "grad_norm": 1.5665547847747803, "learning_rate": 1e-06, "loss": 0.1231, "step": 2477 }, { "epoch": 0.8592233009708737, "grad_norm": 2.601025342941284, "learning_rate": 1e-06, "loss": 0.1207, "step": 2478 }, { "epoch": 0.8595700416088765, "grad_norm": 3.027022123336792, "learning_rate": 1e-06, "loss": 0.1592, "step": 2479 }, { "epoch": 0.8599167822468793, "grad_norm": 2.1945347785949707, "learning_rate": 1e-06, "loss": 0.142, "step": 2480 }, { "epoch": 0.8602635228848821, "grad_norm": 2.1606740951538086, "learning_rate": 1e-06, "loss": 0.1387, "step": 2481 }, { "epoch": 0.8606102635228848, "grad_norm": 3.8076674938201904, "learning_rate": 1e-06, "loss": 0.1128, "step": 2482 }, { "epoch": 0.8609570041608876, "grad_norm": 3.2100863456726074, "learning_rate": 1e-06, "loss": 0.1345, "step": 2483 }, { "epoch": 0.8613037447988904, "grad_norm": 2.945937156677246, "learning_rate": 1e-06, "loss": 0.1253, "step": 2484 }, { "epoch": 0.8616504854368932, "grad_norm": 2.1155993938446045, "learning_rate": 1e-06, "loss": 0.1366, "step": 2485 }, { "epoch": 0.8619972260748959, "grad_norm": 2.137275457382202, "learning_rate": 1e-06, "loss": 0.1434, "step": 2486 }, { "epoch": 0.8623439667128987, "grad_norm": 3.105685234069824, "learning_rate": 1e-06, "loss": 0.1242, "step": 2487 }, { "epoch": 0.8626907073509015, "grad_norm": 2.3001792430877686, "learning_rate": 1e-06, "loss": 0.1322, "step": 2488 }, { "epoch": 0.8630374479889042, "grad_norm": 4.752837657928467, "learning_rate": 1e-06, "loss": 0.1471, "step": 2489 }, { "epoch": 0.863384188626907, "grad_norm": 4.404865264892578, "learning_rate": 1e-06, "loss": 0.1753, "step": 2490 }, { "epoch": 0.8637309292649098, "grad_norm": 2.0198276042938232, "learning_rate": 1e-06, "loss": 0.1351, "step": 2491 }, { "epoch": 0.8640776699029126, "grad_norm": 4.447784900665283, "learning_rate": 1e-06, "loss": 0.1736, "step": 2492 }, { "epoch": 0.8644244105409153, "grad_norm": 2.295395612716675, "learning_rate": 1e-06, "loss": 0.1327, "step": 2493 }, { "epoch": 0.8647711511789181, "grad_norm": 1.9984354972839355, "learning_rate": 1e-06, "loss": 0.1164, "step": 2494 }, { "epoch": 0.8651178918169209, "grad_norm": 2.847456693649292, "learning_rate": 1e-06, "loss": 0.1441, "step": 2495 }, { "epoch": 0.8654646324549237, "grad_norm": 3.244466543197632, "learning_rate": 1e-06, "loss": 0.1434, "step": 2496 }, { "epoch": 0.8658113730929264, "grad_norm": 2.110191583633423, "learning_rate": 1e-06, "loss": 0.1269, "step": 2497 }, { "epoch": 0.8661581137309292, "grad_norm": 2.554386615753174, "learning_rate": 1e-06, "loss": 0.1522, "step": 2498 }, { "epoch": 0.866504854368932, "grad_norm": 3.446542501449585, "learning_rate": 1e-06, "loss": 0.1483, "step": 2499 }, { "epoch": 0.8668515950069348, "grad_norm": 2.2236502170562744, "learning_rate": 1e-06, "loss": 0.1384, "step": 2500 }, { "epoch": 0.8671983356449375, "grad_norm": 2.7683444023132324, "learning_rate": 1e-06, "loss": 0.1283, "step": 2501 }, { "epoch": 0.8675450762829403, "grad_norm": 3.002000093460083, "learning_rate": 1e-06, "loss": 0.1742, "step": 2502 }, { "epoch": 0.8678918169209431, "grad_norm": 5.195772647857666, "learning_rate": 1e-06, "loss": 0.1753, "step": 2503 }, { "epoch": 0.8682385575589459, "grad_norm": 3.325066566467285, "learning_rate": 1e-06, "loss": 0.1238, "step": 2504 }, { "epoch": 0.8685852981969486, "grad_norm": 2.8764262199401855, "learning_rate": 1e-06, "loss": 0.1269, "step": 2505 }, { "epoch": 0.8689320388349514, "grad_norm": 4.183605670928955, "learning_rate": 1e-06, "loss": 0.173, "step": 2506 }, { "epoch": 0.8692787794729542, "grad_norm": 2.656033515930176, "learning_rate": 1e-06, "loss": 0.1767, "step": 2507 }, { "epoch": 0.869625520110957, "grad_norm": 2.7101523876190186, "learning_rate": 1e-06, "loss": 0.1274, "step": 2508 }, { "epoch": 0.8699722607489597, "grad_norm": 2.912912368774414, "learning_rate": 1e-06, "loss": 0.1603, "step": 2509 }, { "epoch": 0.8703190013869625, "grad_norm": 2.7677292823791504, "learning_rate": 1e-06, "loss": 0.1403, "step": 2510 }, { "epoch": 0.8706657420249653, "grad_norm": 7.316020965576172, "learning_rate": 1e-06, "loss": 0.1854, "step": 2511 }, { "epoch": 0.871012482662968, "grad_norm": 2.9252498149871826, "learning_rate": 1e-06, "loss": 0.1385, "step": 2512 }, { "epoch": 0.8713592233009708, "grad_norm": 2.6173603534698486, "learning_rate": 1e-06, "loss": 0.1405, "step": 2513 }, { "epoch": 0.8717059639389736, "grad_norm": 3.6852455139160156, "learning_rate": 1e-06, "loss": 0.1621, "step": 2514 }, { "epoch": 0.8720527045769764, "grad_norm": 3.2382307052612305, "learning_rate": 1e-06, "loss": 0.0994, "step": 2515 }, { "epoch": 0.8723994452149791, "grad_norm": 2.688394784927368, "learning_rate": 1e-06, "loss": 0.1423, "step": 2516 }, { "epoch": 0.8727461858529819, "grad_norm": 3.035778522491455, "learning_rate": 1e-06, "loss": 0.1563, "step": 2517 }, { "epoch": 0.8730929264909847, "grad_norm": 2.3126778602600098, "learning_rate": 1e-06, "loss": 0.1394, "step": 2518 }, { "epoch": 0.8734396671289875, "grad_norm": 2.3858559131622314, "learning_rate": 1e-06, "loss": 0.1215, "step": 2519 }, { "epoch": 0.8737864077669902, "grad_norm": 3.0119643211364746, "learning_rate": 1e-06, "loss": 0.1157, "step": 2520 }, { "epoch": 0.874133148404993, "grad_norm": 2.6011157035827637, "learning_rate": 1e-06, "loss": 0.1646, "step": 2521 }, { "epoch": 0.8744798890429958, "grad_norm": 2.6988086700439453, "learning_rate": 1e-06, "loss": 0.1424, "step": 2522 }, { "epoch": 0.8748266296809986, "grad_norm": 2.830876588821411, "learning_rate": 1e-06, "loss": 0.1378, "step": 2523 }, { "epoch": 0.8751733703190014, "grad_norm": 1.7022074460983276, "learning_rate": 1e-06, "loss": 0.1197, "step": 2524 }, { "epoch": 0.8755201109570042, "grad_norm": 2.1738150119781494, "learning_rate": 1e-06, "loss": 0.123, "step": 2525 }, { "epoch": 0.875866851595007, "grad_norm": 3.662080764770508, "learning_rate": 1e-06, "loss": 0.1436, "step": 2526 }, { "epoch": 0.8762135922330098, "grad_norm": 1.7804447412490845, "learning_rate": 1e-06, "loss": 0.0925, "step": 2527 }, { "epoch": 0.8765603328710125, "grad_norm": 2.030134916305542, "learning_rate": 1e-06, "loss": 0.1294, "step": 2528 }, { "epoch": 0.8769070735090153, "grad_norm": 1.968540072441101, "learning_rate": 1e-06, "loss": 0.1247, "step": 2529 }, { "epoch": 0.8772538141470181, "grad_norm": 3.905348539352417, "learning_rate": 1e-06, "loss": 0.1248, "step": 2530 }, { "epoch": 0.8776005547850209, "grad_norm": 2.988689661026001, "learning_rate": 1e-06, "loss": 0.1408, "step": 2531 }, { "epoch": 0.8779472954230236, "grad_norm": 2.6646010875701904, "learning_rate": 1e-06, "loss": 0.137, "step": 2532 }, { "epoch": 0.8782940360610264, "grad_norm": 3.523129940032959, "learning_rate": 1e-06, "loss": 0.1305, "step": 2533 }, { "epoch": 0.8786407766990292, "grad_norm": 7.069967269897461, "learning_rate": 1e-06, "loss": 0.151, "step": 2534 }, { "epoch": 0.878987517337032, "grad_norm": 2.2202308177948, "learning_rate": 1e-06, "loss": 0.1281, "step": 2535 }, { "epoch": 0.8793342579750347, "grad_norm": 4.117976188659668, "learning_rate": 1e-06, "loss": 0.1423, "step": 2536 }, { "epoch": 0.8796809986130375, "grad_norm": 2.210732936859131, "learning_rate": 1e-06, "loss": 0.1138, "step": 2537 }, { "epoch": 0.8800277392510403, "grad_norm": 2.42452335357666, "learning_rate": 1e-06, "loss": 0.1292, "step": 2538 }, { "epoch": 0.880374479889043, "grad_norm": 4.303539276123047, "learning_rate": 1e-06, "loss": 0.1631, "step": 2539 }, { "epoch": 0.8807212205270458, "grad_norm": 3.3438968658447266, "learning_rate": 1e-06, "loss": 0.1535, "step": 2540 }, { "epoch": 0.8810679611650486, "grad_norm": 3.1318681240081787, "learning_rate": 1e-06, "loss": 0.1806, "step": 2541 }, { "epoch": 0.8814147018030514, "grad_norm": 2.541802167892456, "learning_rate": 1e-06, "loss": 0.132, "step": 2542 }, { "epoch": 0.8817614424410541, "grad_norm": 3.502878427505493, "learning_rate": 1e-06, "loss": 0.127, "step": 2543 }, { "epoch": 0.8821081830790569, "grad_norm": 5.988008499145508, "learning_rate": 1e-06, "loss": 0.1334, "step": 2544 }, { "epoch": 0.8824549237170597, "grad_norm": 3.4589040279388428, "learning_rate": 1e-06, "loss": 0.1198, "step": 2545 }, { "epoch": 0.8828016643550625, "grad_norm": 4.258084774017334, "learning_rate": 1e-06, "loss": 0.1556, "step": 2546 }, { "epoch": 0.8831484049930652, "grad_norm": 2.9290645122528076, "learning_rate": 1e-06, "loss": 0.1383, "step": 2547 }, { "epoch": 0.883495145631068, "grad_norm": 5.24904727935791, "learning_rate": 1e-06, "loss": 0.1404, "step": 2548 }, { "epoch": 0.8838418862690708, "grad_norm": 2.883603572845459, "learning_rate": 1e-06, "loss": 0.1282, "step": 2549 }, { "epoch": 0.8841886269070736, "grad_norm": 2.8196918964385986, "learning_rate": 1e-06, "loss": 0.1478, "step": 2550 }, { "epoch": 0.8845353675450763, "grad_norm": 7.683722972869873, "learning_rate": 1e-06, "loss": 0.2069, "step": 2551 }, { "epoch": 0.8848821081830791, "grad_norm": 3.080634832382202, "learning_rate": 1e-06, "loss": 0.1451, "step": 2552 }, { "epoch": 0.8852288488210819, "grad_norm": 2.3532392978668213, "learning_rate": 1e-06, "loss": 0.1343, "step": 2553 }, { "epoch": 0.8855755894590847, "grad_norm": 2.798870325088501, "learning_rate": 1e-06, "loss": 0.1315, "step": 2554 }, { "epoch": 0.8859223300970874, "grad_norm": 2.4134509563446045, "learning_rate": 1e-06, "loss": 0.1507, "step": 2555 }, { "epoch": 0.8862690707350902, "grad_norm": 3.6513829231262207, "learning_rate": 1e-06, "loss": 0.1658, "step": 2556 }, { "epoch": 0.886615811373093, "grad_norm": 4.6206817626953125, "learning_rate": 1e-06, "loss": 0.1489, "step": 2557 }, { "epoch": 0.8869625520110958, "grad_norm": 4.091174125671387, "learning_rate": 1e-06, "loss": 0.152, "step": 2558 }, { "epoch": 0.8873092926490985, "grad_norm": 1.8325977325439453, "learning_rate": 1e-06, "loss": 0.1384, "step": 2559 }, { "epoch": 0.8876560332871013, "grad_norm": 2.203538656234741, "learning_rate": 1e-06, "loss": 0.1476, "step": 2560 }, { "epoch": 0.8880027739251041, "grad_norm": 3.418199300765991, "learning_rate": 1e-06, "loss": 0.11, "step": 2561 }, { "epoch": 0.8883495145631068, "grad_norm": 3.521099805831909, "learning_rate": 1e-06, "loss": 0.1088, "step": 2562 }, { "epoch": 0.8886962552011096, "grad_norm": 5.303309917449951, "learning_rate": 1e-06, "loss": 0.1504, "step": 2563 }, { "epoch": 0.8890429958391124, "grad_norm": 2.706254243850708, "learning_rate": 1e-06, "loss": 0.1522, "step": 2564 }, { "epoch": 0.8893897364771152, "grad_norm": 3.3216984272003174, "learning_rate": 1e-06, "loss": 0.1324, "step": 2565 }, { "epoch": 0.8897364771151179, "grad_norm": 3.598803997039795, "learning_rate": 1e-06, "loss": 0.1523, "step": 2566 }, { "epoch": 0.8900832177531207, "grad_norm": 2.6814308166503906, "learning_rate": 1e-06, "loss": 0.1539, "step": 2567 }, { "epoch": 0.8904299583911235, "grad_norm": 2.5854735374450684, "learning_rate": 1e-06, "loss": 0.1025, "step": 2568 }, { "epoch": 0.8907766990291263, "grad_norm": 2.8812289237976074, "learning_rate": 1e-06, "loss": 0.1313, "step": 2569 }, { "epoch": 0.891123439667129, "grad_norm": 3.4628384113311768, "learning_rate": 1e-06, "loss": 0.1494, "step": 2570 }, { "epoch": 0.8914701803051318, "grad_norm": 2.1708106994628906, "learning_rate": 1e-06, "loss": 0.1214, "step": 2571 }, { "epoch": 0.8918169209431346, "grad_norm": 2.2976863384246826, "learning_rate": 1e-06, "loss": 0.1184, "step": 2572 }, { "epoch": 0.8921636615811374, "grad_norm": 2.7640933990478516, "learning_rate": 1e-06, "loss": 0.1234, "step": 2573 }, { "epoch": 0.8925104022191401, "grad_norm": 3.4392740726470947, "learning_rate": 1e-06, "loss": 0.1495, "step": 2574 }, { "epoch": 0.8928571428571429, "grad_norm": 3.151865005493164, "learning_rate": 1e-06, "loss": 0.1204, "step": 2575 }, { "epoch": 0.8932038834951457, "grad_norm": 2.665188789367676, "learning_rate": 1e-06, "loss": 0.1312, "step": 2576 }, { "epoch": 0.8935506241331485, "grad_norm": 2.2830886840820312, "learning_rate": 1e-06, "loss": 0.1069, "step": 2577 }, { "epoch": 0.8938973647711512, "grad_norm": 3.063302755355835, "learning_rate": 1e-06, "loss": 0.1506, "step": 2578 }, { "epoch": 0.894244105409154, "grad_norm": 2.6693129539489746, "learning_rate": 1e-06, "loss": 0.1242, "step": 2579 }, { "epoch": 0.8945908460471568, "grad_norm": 3.1677896976470947, "learning_rate": 1e-06, "loss": 0.127, "step": 2580 }, { "epoch": 0.8949375866851595, "grad_norm": 3.562784194946289, "learning_rate": 1e-06, "loss": 0.1502, "step": 2581 }, { "epoch": 0.8952843273231623, "grad_norm": 2.2113330364227295, "learning_rate": 1e-06, "loss": 0.1146, "step": 2582 }, { "epoch": 0.8956310679611651, "grad_norm": 3.4925997257232666, "learning_rate": 1e-06, "loss": 0.133, "step": 2583 }, { "epoch": 0.8959778085991679, "grad_norm": 2.26218843460083, "learning_rate": 1e-06, "loss": 0.14, "step": 2584 }, { "epoch": 0.8963245492371706, "grad_norm": 2.437492847442627, "learning_rate": 1e-06, "loss": 0.1292, "step": 2585 }, { "epoch": 0.8966712898751734, "grad_norm": 3.4927961826324463, "learning_rate": 1e-06, "loss": 0.1195, "step": 2586 }, { "epoch": 0.8970180305131762, "grad_norm": 4.102070331573486, "learning_rate": 1e-06, "loss": 0.1356, "step": 2587 }, { "epoch": 0.897364771151179, "grad_norm": 5.60887336730957, "learning_rate": 1e-06, "loss": 0.1479, "step": 2588 }, { "epoch": 0.8977115117891817, "grad_norm": 2.31508731842041, "learning_rate": 1e-06, "loss": 0.1579, "step": 2589 }, { "epoch": 0.8980582524271845, "grad_norm": 2.7380573749542236, "learning_rate": 1e-06, "loss": 0.1463, "step": 2590 }, { "epoch": 0.8984049930651873, "grad_norm": 1.9896982908248901, "learning_rate": 1e-06, "loss": 0.1162, "step": 2591 }, { "epoch": 0.8987517337031901, "grad_norm": 2.7056241035461426, "learning_rate": 1e-06, "loss": 0.1591, "step": 2592 }, { "epoch": 0.8990984743411928, "grad_norm": 1.9622806310653687, "learning_rate": 1e-06, "loss": 0.1138, "step": 2593 }, { "epoch": 0.8994452149791956, "grad_norm": 3.957826614379883, "learning_rate": 1e-06, "loss": 0.1561, "step": 2594 }, { "epoch": 0.8997919556171984, "grad_norm": 2.8110544681549072, "learning_rate": 1e-06, "loss": 0.1379, "step": 2595 }, { "epoch": 0.9001386962552012, "grad_norm": 3.7810873985290527, "learning_rate": 1e-06, "loss": 0.1533, "step": 2596 }, { "epoch": 0.9004854368932039, "grad_norm": 2.166959762573242, "learning_rate": 1e-06, "loss": 0.1411, "step": 2597 }, { "epoch": 0.9008321775312067, "grad_norm": 2.021667718887329, "learning_rate": 1e-06, "loss": 0.1501, "step": 2598 }, { "epoch": 0.9011789181692095, "grad_norm": 3.390395402908325, "learning_rate": 1e-06, "loss": 0.1468, "step": 2599 }, { "epoch": 0.9015256588072122, "grad_norm": 2.1694798469543457, "learning_rate": 1e-06, "loss": 0.1283, "step": 2600 }, { "epoch": 0.901872399445215, "grad_norm": 3.0354669094085693, "learning_rate": 1e-06, "loss": 0.1327, "step": 2601 }, { "epoch": 0.9022191400832178, "grad_norm": 3.5150482654571533, "learning_rate": 1e-06, "loss": 0.1492, "step": 2602 }, { "epoch": 0.9025658807212206, "grad_norm": 2.1176624298095703, "learning_rate": 1e-06, "loss": 0.1426, "step": 2603 }, { "epoch": 0.9029126213592233, "grad_norm": 2.6938793659210205, "learning_rate": 1e-06, "loss": 0.1662, "step": 2604 }, { "epoch": 0.9032593619972261, "grad_norm": 3.2767770290374756, "learning_rate": 1e-06, "loss": 0.1132, "step": 2605 }, { "epoch": 0.9036061026352289, "grad_norm": 3.4244790077209473, "learning_rate": 1e-06, "loss": 0.152, "step": 2606 }, { "epoch": 0.9039528432732317, "grad_norm": 2.3270678520202637, "learning_rate": 1e-06, "loss": 0.164, "step": 2607 }, { "epoch": 0.9042995839112344, "grad_norm": 2.2908077239990234, "learning_rate": 1e-06, "loss": 0.1561, "step": 2608 }, { "epoch": 0.9046463245492372, "grad_norm": 1.976806879043579, "learning_rate": 1e-06, "loss": 0.1219, "step": 2609 }, { "epoch": 0.90499306518724, "grad_norm": 2.578418731689453, "learning_rate": 1e-06, "loss": 0.1454, "step": 2610 }, { "epoch": 0.9053398058252428, "grad_norm": 2.3116960525512695, "learning_rate": 1e-06, "loss": 0.1499, "step": 2611 }, { "epoch": 0.9056865464632455, "grad_norm": 3.2435667514801025, "learning_rate": 1e-06, "loss": 0.1213, "step": 2612 }, { "epoch": 0.9060332871012483, "grad_norm": 1.842995524406433, "learning_rate": 1e-06, "loss": 0.1074, "step": 2613 }, { "epoch": 0.9063800277392511, "grad_norm": 2.7041714191436768, "learning_rate": 1e-06, "loss": 0.1504, "step": 2614 }, { "epoch": 0.9067267683772539, "grad_norm": 3.233328104019165, "learning_rate": 1e-06, "loss": 0.1284, "step": 2615 }, { "epoch": 0.9070735090152566, "grad_norm": 2.0158610343933105, "learning_rate": 1e-06, "loss": 0.1097, "step": 2616 }, { "epoch": 0.9074202496532594, "grad_norm": 2.651226043701172, "learning_rate": 1e-06, "loss": 0.1276, "step": 2617 }, { "epoch": 0.9077669902912622, "grad_norm": 3.511500835418701, "learning_rate": 1e-06, "loss": 0.1523, "step": 2618 }, { "epoch": 0.908113730929265, "grad_norm": 2.868360996246338, "learning_rate": 1e-06, "loss": 0.1482, "step": 2619 }, { "epoch": 0.9084604715672677, "grad_norm": 5.283854961395264, "learning_rate": 1e-06, "loss": 0.1529, "step": 2620 }, { "epoch": 0.9088072122052705, "grad_norm": 2.312080144882202, "learning_rate": 1e-06, "loss": 0.1197, "step": 2621 }, { "epoch": 0.9091539528432733, "grad_norm": 2.4903831481933594, "learning_rate": 1e-06, "loss": 0.1488, "step": 2622 }, { "epoch": 0.909500693481276, "grad_norm": 3.8020405769348145, "learning_rate": 1e-06, "loss": 0.1167, "step": 2623 }, { "epoch": 0.9098474341192788, "grad_norm": 4.8802666664123535, "learning_rate": 1e-06, "loss": 0.1598, "step": 2624 }, { "epoch": 0.9101941747572816, "grad_norm": 2.2983484268188477, "learning_rate": 1e-06, "loss": 0.1255, "step": 2625 }, { "epoch": 0.9105409153952844, "grad_norm": 2.7895936965942383, "learning_rate": 1e-06, "loss": 0.1464, "step": 2626 }, { "epoch": 0.9108876560332871, "grad_norm": 3.0631260871887207, "learning_rate": 1e-06, "loss": 0.1165, "step": 2627 }, { "epoch": 0.9112343966712899, "grad_norm": 5.066497802734375, "learning_rate": 1e-06, "loss": 0.1336, "step": 2628 }, { "epoch": 0.9115811373092927, "grad_norm": 2.471651792526245, "learning_rate": 1e-06, "loss": 0.1418, "step": 2629 }, { "epoch": 0.9119278779472955, "grad_norm": 2.3465819358825684, "learning_rate": 1e-06, "loss": 0.1309, "step": 2630 }, { "epoch": 0.9122746185852982, "grad_norm": 2.3085386753082275, "learning_rate": 1e-06, "loss": 0.1099, "step": 2631 }, { "epoch": 0.912621359223301, "grad_norm": 6.42535400390625, "learning_rate": 1e-06, "loss": 0.1657, "step": 2632 }, { "epoch": 0.9129680998613038, "grad_norm": 2.8673324584960938, "learning_rate": 1e-06, "loss": 0.1509, "step": 2633 }, { "epoch": 0.9133148404993066, "grad_norm": 2.016261339187622, "learning_rate": 1e-06, "loss": 0.1464, "step": 2634 }, { "epoch": 0.9136615811373093, "grad_norm": 2.854090690612793, "learning_rate": 1e-06, "loss": 0.1341, "step": 2635 }, { "epoch": 0.9140083217753121, "grad_norm": 2.522303342819214, "learning_rate": 1e-06, "loss": 0.1272, "step": 2636 }, { "epoch": 0.9143550624133149, "grad_norm": 2.4291818141937256, "learning_rate": 1e-06, "loss": 0.1476, "step": 2637 }, { "epoch": 0.9147018030513177, "grad_norm": 1.8062353134155273, "learning_rate": 1e-06, "loss": 0.1041, "step": 2638 }, { "epoch": 0.9150485436893204, "grad_norm": 2.9814038276672363, "learning_rate": 1e-06, "loss": 0.1324, "step": 2639 }, { "epoch": 0.9153952843273232, "grad_norm": 2.3664748668670654, "learning_rate": 1e-06, "loss": 0.1533, "step": 2640 }, { "epoch": 0.915742024965326, "grad_norm": 2.347658634185791, "learning_rate": 1e-06, "loss": 0.1217, "step": 2641 }, { "epoch": 0.9160887656033287, "grad_norm": 3.068976640701294, "learning_rate": 1e-06, "loss": 0.1383, "step": 2642 }, { "epoch": 0.9164355062413315, "grad_norm": 2.92535400390625, "learning_rate": 1e-06, "loss": 0.1459, "step": 2643 }, { "epoch": 0.9167822468793343, "grad_norm": 2.56888747215271, "learning_rate": 1e-06, "loss": 0.1308, "step": 2644 }, { "epoch": 0.9171289875173371, "grad_norm": 2.389007806777954, "learning_rate": 1e-06, "loss": 0.1628, "step": 2645 }, { "epoch": 0.9174757281553398, "grad_norm": 2.4458770751953125, "learning_rate": 1e-06, "loss": 0.1135, "step": 2646 }, { "epoch": 0.9178224687933426, "grad_norm": 2.0837292671203613, "learning_rate": 1e-06, "loss": 0.1154, "step": 2647 }, { "epoch": 0.9181692094313454, "grad_norm": 5.257343769073486, "learning_rate": 1e-06, "loss": 0.1318, "step": 2648 }, { "epoch": 0.9185159500693482, "grad_norm": 4.321478366851807, "learning_rate": 1e-06, "loss": 0.1418, "step": 2649 }, { "epoch": 0.9188626907073509, "grad_norm": 2.9317991733551025, "learning_rate": 1e-06, "loss": 0.1491, "step": 2650 }, { "epoch": 0.9192094313453537, "grad_norm": 3.1797657012939453, "learning_rate": 1e-06, "loss": 0.1566, "step": 2651 }, { "epoch": 0.9195561719833565, "grad_norm": 2.146343231201172, "learning_rate": 1e-06, "loss": 0.1431, "step": 2652 }, { "epoch": 0.9199029126213593, "grad_norm": 2.9985647201538086, "learning_rate": 1e-06, "loss": 0.1404, "step": 2653 }, { "epoch": 0.920249653259362, "grad_norm": 1.452409267425537, "learning_rate": 1e-06, "loss": 0.0951, "step": 2654 }, { "epoch": 0.9205963938973648, "grad_norm": 3.1620914936065674, "learning_rate": 1e-06, "loss": 0.1582, "step": 2655 }, { "epoch": 0.9209431345353676, "grad_norm": 4.072134971618652, "learning_rate": 1e-06, "loss": 0.1415, "step": 2656 }, { "epoch": 0.9212898751733704, "grad_norm": 2.393535852432251, "learning_rate": 1e-06, "loss": 0.1214, "step": 2657 }, { "epoch": 0.9216366158113731, "grad_norm": 1.983525037765503, "learning_rate": 1e-06, "loss": 0.1196, "step": 2658 }, { "epoch": 0.9219833564493759, "grad_norm": 2.7609565258026123, "learning_rate": 1e-06, "loss": 0.1005, "step": 2659 }, { "epoch": 0.9223300970873787, "grad_norm": 2.900486707687378, "learning_rate": 1e-06, "loss": 0.1112, "step": 2660 }, { "epoch": 0.9226768377253814, "grad_norm": 2.1307523250579834, "learning_rate": 1e-06, "loss": 0.114, "step": 2661 }, { "epoch": 0.9230235783633842, "grad_norm": 3.795403242111206, "learning_rate": 1e-06, "loss": 0.1341, "step": 2662 }, { "epoch": 0.923370319001387, "grad_norm": 2.027860403060913, "learning_rate": 1e-06, "loss": 0.1359, "step": 2663 }, { "epoch": 0.9237170596393898, "grad_norm": 1.9218776226043701, "learning_rate": 1e-06, "loss": 0.1206, "step": 2664 }, { "epoch": 0.9240638002773925, "grad_norm": 2.235786199569702, "learning_rate": 1e-06, "loss": 0.1399, "step": 2665 }, { "epoch": 0.9244105409153953, "grad_norm": 2.2658088207244873, "learning_rate": 1e-06, "loss": 0.1469, "step": 2666 }, { "epoch": 0.9247572815533981, "grad_norm": 2.2411727905273438, "learning_rate": 1e-06, "loss": 0.1334, "step": 2667 }, { "epoch": 0.9251040221914009, "grad_norm": 2.2818808555603027, "learning_rate": 1e-06, "loss": 0.1323, "step": 2668 }, { "epoch": 0.9254507628294036, "grad_norm": 2.430490255355835, "learning_rate": 1e-06, "loss": 0.129, "step": 2669 }, { "epoch": 0.9257975034674064, "grad_norm": 2.8982584476470947, "learning_rate": 1e-06, "loss": 0.1231, "step": 2670 }, { "epoch": 0.9261442441054092, "grad_norm": 3.195540428161621, "learning_rate": 1e-06, "loss": 0.1309, "step": 2671 }, { "epoch": 0.926490984743412, "grad_norm": 2.102155923843384, "learning_rate": 1e-06, "loss": 0.1392, "step": 2672 }, { "epoch": 0.9268377253814147, "grad_norm": 2.8941521644592285, "learning_rate": 1e-06, "loss": 0.1402, "step": 2673 }, { "epoch": 0.9271844660194175, "grad_norm": 1.982313632965088, "learning_rate": 1e-06, "loss": 0.1215, "step": 2674 }, { "epoch": 0.9275312066574203, "grad_norm": 3.4784913063049316, "learning_rate": 1e-06, "loss": 0.1597, "step": 2675 }, { "epoch": 0.9278779472954231, "grad_norm": 2.5161221027374268, "learning_rate": 1e-06, "loss": 0.1721, "step": 2676 }, { "epoch": 0.9282246879334258, "grad_norm": 2.5937342643737793, "learning_rate": 1e-06, "loss": 0.1466, "step": 2677 }, { "epoch": 0.9285714285714286, "grad_norm": 2.7261319160461426, "learning_rate": 1e-06, "loss": 0.1446, "step": 2678 }, { "epoch": 0.9289181692094314, "grad_norm": 2.406301498413086, "learning_rate": 1e-06, "loss": 0.143, "step": 2679 }, { "epoch": 0.9292649098474342, "grad_norm": 3.284876823425293, "learning_rate": 1e-06, "loss": 0.1859, "step": 2680 }, { "epoch": 0.9296116504854369, "grad_norm": 2.8308169841766357, "learning_rate": 1e-06, "loss": 0.1483, "step": 2681 }, { "epoch": 0.9299583911234397, "grad_norm": 2.309736490249634, "learning_rate": 1e-06, "loss": 0.1341, "step": 2682 }, { "epoch": 0.9303051317614425, "grad_norm": 2.3416709899902344, "learning_rate": 1e-06, "loss": 0.09, "step": 2683 }, { "epoch": 0.9306518723994452, "grad_norm": 2.0100162029266357, "learning_rate": 1e-06, "loss": 0.1185, "step": 2684 }, { "epoch": 0.930998613037448, "grad_norm": 2.5902180671691895, "learning_rate": 1e-06, "loss": 0.1393, "step": 2685 }, { "epoch": 0.9313453536754508, "grad_norm": 4.1081767082214355, "learning_rate": 1e-06, "loss": 0.1468, "step": 2686 }, { "epoch": 0.9316920943134536, "grad_norm": 2.6473231315612793, "learning_rate": 1e-06, "loss": 0.1258, "step": 2687 }, { "epoch": 0.9320388349514563, "grad_norm": 2.6274592876434326, "learning_rate": 1e-06, "loss": 0.1216, "step": 2688 }, { "epoch": 0.9323855755894591, "grad_norm": 2.321573495864868, "learning_rate": 1e-06, "loss": 0.1156, "step": 2689 }, { "epoch": 0.9327323162274619, "grad_norm": 2.83542799949646, "learning_rate": 1e-06, "loss": 0.113, "step": 2690 }, { "epoch": 0.9330790568654647, "grad_norm": 3.1709511280059814, "learning_rate": 1e-06, "loss": 0.1644, "step": 2691 }, { "epoch": 0.9334257975034674, "grad_norm": 1.8596209287643433, "learning_rate": 1e-06, "loss": 0.126, "step": 2692 }, { "epoch": 0.9337725381414702, "grad_norm": 2.7414655685424805, "learning_rate": 1e-06, "loss": 0.1305, "step": 2693 }, { "epoch": 0.934119278779473, "grad_norm": 3.0172135829925537, "learning_rate": 1e-06, "loss": 0.1463, "step": 2694 }, { "epoch": 0.9344660194174758, "grad_norm": 2.8152413368225098, "learning_rate": 1e-06, "loss": 0.1396, "step": 2695 }, { "epoch": 0.9348127600554785, "grad_norm": 2.0890872478485107, "learning_rate": 1e-06, "loss": 0.1034, "step": 2696 }, { "epoch": 0.9351595006934813, "grad_norm": 2.5807907581329346, "learning_rate": 1e-06, "loss": 0.1488, "step": 2697 }, { "epoch": 0.9355062413314841, "grad_norm": 2.900175094604492, "learning_rate": 1e-06, "loss": 0.1427, "step": 2698 }, { "epoch": 0.9358529819694869, "grad_norm": 2.0167062282562256, "learning_rate": 1e-06, "loss": 0.1071, "step": 2699 }, { "epoch": 0.9361997226074896, "grad_norm": 2.26310658454895, "learning_rate": 1e-06, "loss": 0.1393, "step": 2700 }, { "epoch": 0.9365464632454924, "grad_norm": 2.191679000854492, "learning_rate": 1e-06, "loss": 0.1299, "step": 2701 }, { "epoch": 0.9368932038834952, "grad_norm": 2.8167102336883545, "learning_rate": 1e-06, "loss": 0.1317, "step": 2702 }, { "epoch": 0.937239944521498, "grad_norm": 2.700887441635132, "learning_rate": 1e-06, "loss": 0.1726, "step": 2703 }, { "epoch": 0.9375866851595007, "grad_norm": 2.511220932006836, "learning_rate": 1e-06, "loss": 0.1579, "step": 2704 }, { "epoch": 0.9379334257975035, "grad_norm": 3.294839859008789, "learning_rate": 1e-06, "loss": 0.142, "step": 2705 }, { "epoch": 0.9382801664355063, "grad_norm": 2.9182393550872803, "learning_rate": 1e-06, "loss": 0.1242, "step": 2706 }, { "epoch": 0.938626907073509, "grad_norm": 1.9845600128173828, "learning_rate": 1e-06, "loss": 0.1103, "step": 2707 }, { "epoch": 0.9389736477115118, "grad_norm": 3.3602540493011475, "learning_rate": 1e-06, "loss": 0.1374, "step": 2708 }, { "epoch": 0.9393203883495146, "grad_norm": 2.03686261177063, "learning_rate": 1e-06, "loss": 0.1203, "step": 2709 }, { "epoch": 0.9396671289875174, "grad_norm": 2.6273975372314453, "learning_rate": 1e-06, "loss": 0.1296, "step": 2710 }, { "epoch": 0.9400138696255201, "grad_norm": 2.5899906158447266, "learning_rate": 1e-06, "loss": 0.1411, "step": 2711 }, { "epoch": 0.9403606102635229, "grad_norm": 2.5806844234466553, "learning_rate": 1e-06, "loss": 0.1371, "step": 2712 }, { "epoch": 0.9407073509015257, "grad_norm": 2.3415279388427734, "learning_rate": 1e-06, "loss": 0.1231, "step": 2713 }, { "epoch": 0.9410540915395285, "grad_norm": 3.5045316219329834, "learning_rate": 1e-06, "loss": 0.1382, "step": 2714 }, { "epoch": 0.9414008321775312, "grad_norm": 2.0843143463134766, "learning_rate": 1e-06, "loss": 0.1078, "step": 2715 }, { "epoch": 0.941747572815534, "grad_norm": 2.7572216987609863, "learning_rate": 1e-06, "loss": 0.1407, "step": 2716 }, { "epoch": 0.9420943134535368, "grad_norm": 2.4735355377197266, "learning_rate": 1e-06, "loss": 0.1305, "step": 2717 }, { "epoch": 0.9424410540915396, "grad_norm": 2.1515815258026123, "learning_rate": 1e-06, "loss": 0.1231, "step": 2718 }, { "epoch": 0.9427877947295423, "grad_norm": 3.494328737258911, "learning_rate": 1e-06, "loss": 0.1017, "step": 2719 }, { "epoch": 0.9431345353675451, "grad_norm": 2.391047239303589, "learning_rate": 1e-06, "loss": 0.125, "step": 2720 }, { "epoch": 0.9434812760055479, "grad_norm": 1.7509092092514038, "learning_rate": 1e-06, "loss": 0.1099, "step": 2721 }, { "epoch": 0.9438280166435506, "grad_norm": 2.7417235374450684, "learning_rate": 1e-06, "loss": 0.1167, "step": 2722 }, { "epoch": 0.9441747572815534, "grad_norm": 2.740318536758423, "learning_rate": 1e-06, "loss": 0.1291, "step": 2723 }, { "epoch": 0.9445214979195562, "grad_norm": 3.362915515899658, "learning_rate": 1e-06, "loss": 0.1361, "step": 2724 }, { "epoch": 0.944868238557559, "grad_norm": 2.27724552154541, "learning_rate": 1e-06, "loss": 0.1218, "step": 2725 }, { "epoch": 0.9452149791955617, "grad_norm": 2.4609215259552, "learning_rate": 1e-06, "loss": 0.1397, "step": 2726 }, { "epoch": 0.9455617198335645, "grad_norm": 3.4139716625213623, "learning_rate": 1e-06, "loss": 0.1238, "step": 2727 }, { "epoch": 0.9459084604715673, "grad_norm": 2.5708959102630615, "learning_rate": 1e-06, "loss": 0.1173, "step": 2728 }, { "epoch": 0.9462552011095701, "grad_norm": 2.1245861053466797, "learning_rate": 1e-06, "loss": 0.1325, "step": 2729 }, { "epoch": 0.9466019417475728, "grad_norm": 3.1224300861358643, "learning_rate": 1e-06, "loss": 0.1674, "step": 2730 }, { "epoch": 0.9469486823855756, "grad_norm": 2.9298243522644043, "learning_rate": 1e-06, "loss": 0.1628, "step": 2731 }, { "epoch": 0.9472954230235784, "grad_norm": 2.5283334255218506, "learning_rate": 1e-06, "loss": 0.1238, "step": 2732 }, { "epoch": 0.9476421636615812, "grad_norm": 1.95565927028656, "learning_rate": 1e-06, "loss": 0.1154, "step": 2733 }, { "epoch": 0.9479889042995839, "grad_norm": 2.800057888031006, "learning_rate": 1e-06, "loss": 0.1088, "step": 2734 }, { "epoch": 0.9483356449375867, "grad_norm": 2.9999020099639893, "learning_rate": 1e-06, "loss": 0.1263, "step": 2735 }, { "epoch": 0.9486823855755895, "grad_norm": 2.755236864089966, "learning_rate": 1e-06, "loss": 0.1282, "step": 2736 }, { "epoch": 0.9490291262135923, "grad_norm": 2.161933422088623, "learning_rate": 1e-06, "loss": 0.1441, "step": 2737 }, { "epoch": 0.949375866851595, "grad_norm": 2.8767547607421875, "learning_rate": 1e-06, "loss": 0.1047, "step": 2738 }, { "epoch": 0.9497226074895978, "grad_norm": 2.172853946685791, "learning_rate": 1e-06, "loss": 0.1201, "step": 2739 }, { "epoch": 0.9500693481276006, "grad_norm": 2.792728900909424, "learning_rate": 1e-06, "loss": 0.1136, "step": 2740 }, { "epoch": 0.9504160887656034, "grad_norm": 2.2738940715789795, "learning_rate": 1e-06, "loss": 0.0994, "step": 2741 }, { "epoch": 0.9507628294036061, "grad_norm": 3.087293863296509, "learning_rate": 1e-06, "loss": 0.1417, "step": 2742 }, { "epoch": 0.9511095700416089, "grad_norm": 2.7275638580322266, "learning_rate": 1e-06, "loss": 0.1498, "step": 2743 }, { "epoch": 0.9514563106796117, "grad_norm": 3.0619494915008545, "learning_rate": 1e-06, "loss": 0.1445, "step": 2744 }, { "epoch": 0.9518030513176144, "grad_norm": 2.5986380577087402, "learning_rate": 1e-06, "loss": 0.1363, "step": 2745 }, { "epoch": 0.9521497919556172, "grad_norm": 3.417404890060425, "learning_rate": 1e-06, "loss": 0.1448, "step": 2746 }, { "epoch": 0.95249653259362, "grad_norm": 4.566203594207764, "learning_rate": 1e-06, "loss": 0.1653, "step": 2747 }, { "epoch": 0.9528432732316228, "grad_norm": 3.027130126953125, "learning_rate": 1e-06, "loss": 0.1543, "step": 2748 }, { "epoch": 0.9531900138696255, "grad_norm": 3.81862211227417, "learning_rate": 1e-06, "loss": 0.1651, "step": 2749 }, { "epoch": 0.9535367545076283, "grad_norm": 2.1442902088165283, "learning_rate": 1e-06, "loss": 0.1094, "step": 2750 }, { "epoch": 0.9538834951456311, "grad_norm": 2.5918867588043213, "learning_rate": 1e-06, "loss": 0.1348, "step": 2751 }, { "epoch": 0.9542302357836339, "grad_norm": 2.379676103591919, "learning_rate": 1e-06, "loss": 0.1376, "step": 2752 }, { "epoch": 0.9545769764216366, "grad_norm": 2.339801549911499, "learning_rate": 1e-06, "loss": 0.1239, "step": 2753 }, { "epoch": 0.9549237170596394, "grad_norm": 3.523301839828491, "learning_rate": 1e-06, "loss": 0.1241, "step": 2754 }, { "epoch": 0.9552704576976422, "grad_norm": 5.052760124206543, "learning_rate": 1e-06, "loss": 0.132, "step": 2755 }, { "epoch": 0.955617198335645, "grad_norm": 2.946012020111084, "learning_rate": 1e-06, "loss": 0.1742, "step": 2756 }, { "epoch": 0.9559639389736477, "grad_norm": 3.70878529548645, "learning_rate": 1e-06, "loss": 0.1526, "step": 2757 }, { "epoch": 0.9563106796116505, "grad_norm": 2.7530357837677, "learning_rate": 1e-06, "loss": 0.1757, "step": 2758 }, { "epoch": 0.9566574202496533, "grad_norm": 5.642550945281982, "learning_rate": 1e-06, "loss": 0.1614, "step": 2759 }, { "epoch": 0.957004160887656, "grad_norm": 3.5328824520111084, "learning_rate": 1e-06, "loss": 0.1212, "step": 2760 }, { "epoch": 0.9573509015256588, "grad_norm": 2.6719205379486084, "learning_rate": 1e-06, "loss": 0.136, "step": 2761 }, { "epoch": 0.9576976421636616, "grad_norm": 3.643240213394165, "learning_rate": 1e-06, "loss": 0.1279, "step": 2762 }, { "epoch": 0.9580443828016644, "grad_norm": 3.318526029586792, "learning_rate": 1e-06, "loss": 0.1327, "step": 2763 }, { "epoch": 0.9583911234396671, "grad_norm": 2.845430850982666, "learning_rate": 1e-06, "loss": 0.1291, "step": 2764 }, { "epoch": 0.9587378640776699, "grad_norm": 2.6309032440185547, "learning_rate": 1e-06, "loss": 0.1295, "step": 2765 }, { "epoch": 0.9590846047156727, "grad_norm": 2.3172414302825928, "learning_rate": 1e-06, "loss": 0.1494, "step": 2766 }, { "epoch": 0.9594313453536755, "grad_norm": 2.4579062461853027, "learning_rate": 1e-06, "loss": 0.1252, "step": 2767 }, { "epoch": 0.9597780859916782, "grad_norm": 2.277562379837036, "learning_rate": 1e-06, "loss": 0.127, "step": 2768 }, { "epoch": 0.960124826629681, "grad_norm": 2.561744451522827, "learning_rate": 1e-06, "loss": 0.1183, "step": 2769 }, { "epoch": 0.9604715672676838, "grad_norm": 2.368710517883301, "learning_rate": 1e-06, "loss": 0.1209, "step": 2770 }, { "epoch": 0.9608183079056866, "grad_norm": 2.4350576400756836, "learning_rate": 1e-06, "loss": 0.1615, "step": 2771 }, { "epoch": 0.9611650485436893, "grad_norm": 2.933718204498291, "learning_rate": 1e-06, "loss": 0.1183, "step": 2772 }, { "epoch": 0.9615117891816921, "grad_norm": 2.6752936840057373, "learning_rate": 1e-06, "loss": 0.1351, "step": 2773 }, { "epoch": 0.9618585298196949, "grad_norm": 2.8626444339752197, "learning_rate": 1e-06, "loss": 0.1445, "step": 2774 }, { "epoch": 0.9622052704576977, "grad_norm": 2.7919156551361084, "learning_rate": 1e-06, "loss": 0.1015, "step": 2775 }, { "epoch": 0.9625520110957004, "grad_norm": 2.687023878097534, "learning_rate": 1e-06, "loss": 0.1398, "step": 2776 }, { "epoch": 0.9628987517337032, "grad_norm": 3.4772815704345703, "learning_rate": 1e-06, "loss": 0.1276, "step": 2777 }, { "epoch": 0.963245492371706, "grad_norm": 3.7592716217041016, "learning_rate": 1e-06, "loss": 0.1196, "step": 2778 }, { "epoch": 0.9635922330097088, "grad_norm": 2.2832236289978027, "learning_rate": 1e-06, "loss": 0.1149, "step": 2779 }, { "epoch": 0.9639389736477115, "grad_norm": 2.2533531188964844, "learning_rate": 1e-06, "loss": 0.1361, "step": 2780 }, { "epoch": 0.9642857142857143, "grad_norm": 2.458859920501709, "learning_rate": 1e-06, "loss": 0.1248, "step": 2781 }, { "epoch": 0.9646324549237171, "grad_norm": 2.3881120681762695, "learning_rate": 1e-06, "loss": 0.1365, "step": 2782 }, { "epoch": 0.9649791955617198, "grad_norm": 2.8706626892089844, "learning_rate": 1e-06, "loss": 0.1324, "step": 2783 }, { "epoch": 0.9653259361997226, "grad_norm": 2.1797640323638916, "learning_rate": 1e-06, "loss": 0.1178, "step": 2784 }, { "epoch": 0.9656726768377254, "grad_norm": 3.3081295490264893, "learning_rate": 1e-06, "loss": 0.1058, "step": 2785 }, { "epoch": 0.9660194174757282, "grad_norm": 4.193579196929932, "learning_rate": 1e-06, "loss": 0.1522, "step": 2786 }, { "epoch": 0.9663661581137309, "grad_norm": 2.6109237670898438, "learning_rate": 1e-06, "loss": 0.1091, "step": 2787 }, { "epoch": 0.9667128987517337, "grad_norm": 2.3282434940338135, "learning_rate": 1e-06, "loss": 0.1333, "step": 2788 }, { "epoch": 0.9670596393897365, "grad_norm": 2.578936815261841, "learning_rate": 1e-06, "loss": 0.1161, "step": 2789 }, { "epoch": 0.9674063800277393, "grad_norm": 3.2605860233306885, "learning_rate": 1e-06, "loss": 0.1535, "step": 2790 }, { "epoch": 0.967753120665742, "grad_norm": 2.45969295501709, "learning_rate": 1e-06, "loss": 0.1199, "step": 2791 }, { "epoch": 0.9680998613037448, "grad_norm": 3.5829577445983887, "learning_rate": 1e-06, "loss": 0.1001, "step": 2792 }, { "epoch": 0.9684466019417476, "grad_norm": 2.413280963897705, "learning_rate": 1e-06, "loss": 0.1261, "step": 2793 }, { "epoch": 0.9687933425797504, "grad_norm": 2.541856050491333, "learning_rate": 1e-06, "loss": 0.106, "step": 2794 }, { "epoch": 0.9691400832177531, "grad_norm": 1.7989451885223389, "learning_rate": 1e-06, "loss": 0.1139, "step": 2795 }, { "epoch": 0.9694868238557559, "grad_norm": 5.392777442932129, "learning_rate": 1e-06, "loss": 0.1379, "step": 2796 }, { "epoch": 0.9698335644937587, "grad_norm": 2.561255931854248, "learning_rate": 1e-06, "loss": 0.1301, "step": 2797 }, { "epoch": 0.9701803051317615, "grad_norm": 2.3186416625976562, "learning_rate": 1e-06, "loss": 0.1049, "step": 2798 }, { "epoch": 0.9705270457697642, "grad_norm": 2.255427122116089, "learning_rate": 1e-06, "loss": 0.1084, "step": 2799 }, { "epoch": 0.970873786407767, "grad_norm": 3.2925469875335693, "learning_rate": 1e-06, "loss": 0.1509, "step": 2800 }, { "epoch": 0.9712205270457698, "grad_norm": 2.2866995334625244, "learning_rate": 1e-06, "loss": 0.1221, "step": 2801 }, { "epoch": 0.9715672676837726, "grad_norm": 2.409914016723633, "learning_rate": 1e-06, "loss": 0.1279, "step": 2802 }, { "epoch": 0.9719140083217753, "grad_norm": 2.746307849884033, "learning_rate": 1e-06, "loss": 0.1119, "step": 2803 }, { "epoch": 0.9722607489597781, "grad_norm": 2.264927387237549, "learning_rate": 1e-06, "loss": 0.1333, "step": 2804 }, { "epoch": 0.9726074895977809, "grad_norm": 1.9397131204605103, "learning_rate": 1e-06, "loss": 0.0929, "step": 2805 }, { "epoch": 0.9729542302357836, "grad_norm": 4.742498397827148, "learning_rate": 1e-06, "loss": 0.1577, "step": 2806 }, { "epoch": 0.9733009708737864, "grad_norm": 3.283451795578003, "learning_rate": 1e-06, "loss": 0.114, "step": 2807 }, { "epoch": 0.9736477115117892, "grad_norm": 2.6818947792053223, "learning_rate": 1e-06, "loss": 0.1446, "step": 2808 }, { "epoch": 0.973994452149792, "grad_norm": 4.0017476081848145, "learning_rate": 1e-06, "loss": 0.1241, "step": 2809 }, { "epoch": 0.9743411927877947, "grad_norm": 2.343931198120117, "learning_rate": 1e-06, "loss": 0.1157, "step": 2810 }, { "epoch": 0.9746879334257975, "grad_norm": 2.369596004486084, "learning_rate": 1e-06, "loss": 0.1451, "step": 2811 }, { "epoch": 0.9750346740638003, "grad_norm": 2.386434316635132, "learning_rate": 1e-06, "loss": 0.1205, "step": 2812 }, { "epoch": 0.9753814147018031, "grad_norm": 1.8188420534133911, "learning_rate": 1e-06, "loss": 0.0877, "step": 2813 }, { "epoch": 0.9757281553398058, "grad_norm": 2.927330493927002, "learning_rate": 1e-06, "loss": 0.1493, "step": 2814 }, { "epoch": 0.9760748959778086, "grad_norm": 2.0784151554107666, "learning_rate": 1e-06, "loss": 0.1269, "step": 2815 }, { "epoch": 0.9764216366158114, "grad_norm": 2.062995433807373, "learning_rate": 1e-06, "loss": 0.1287, "step": 2816 }, { "epoch": 0.9767683772538142, "grad_norm": 2.0508055686950684, "learning_rate": 1e-06, "loss": 0.0901, "step": 2817 }, { "epoch": 0.9771151178918169, "grad_norm": 3.7478532791137695, "learning_rate": 1e-06, "loss": 0.1304, "step": 2818 }, { "epoch": 0.9774618585298197, "grad_norm": 2.1847145557403564, "learning_rate": 1e-06, "loss": 0.1275, "step": 2819 }, { "epoch": 0.9778085991678225, "grad_norm": 2.580650568008423, "learning_rate": 1e-06, "loss": 0.157, "step": 2820 }, { "epoch": 0.9781553398058253, "grad_norm": 2.9201722145080566, "learning_rate": 1e-06, "loss": 0.1486, "step": 2821 }, { "epoch": 0.978502080443828, "grad_norm": 5.508189678192139, "learning_rate": 1e-06, "loss": 0.1458, "step": 2822 }, { "epoch": 0.9788488210818308, "grad_norm": 3.1913814544677734, "learning_rate": 1e-06, "loss": 0.1162, "step": 2823 }, { "epoch": 0.9791955617198336, "grad_norm": 4.063955783843994, "learning_rate": 1e-06, "loss": 0.159, "step": 2824 }, { "epoch": 0.9795423023578363, "grad_norm": 2.553622245788574, "learning_rate": 1e-06, "loss": 0.1343, "step": 2825 }, { "epoch": 0.9798890429958391, "grad_norm": 2.749812364578247, "learning_rate": 1e-06, "loss": 0.1393, "step": 2826 }, { "epoch": 0.9802357836338419, "grad_norm": 3.2445449829101562, "learning_rate": 1e-06, "loss": 0.1138, "step": 2827 }, { "epoch": 0.9805825242718447, "grad_norm": 2.1383461952209473, "learning_rate": 1e-06, "loss": 0.1053, "step": 2828 }, { "epoch": 0.9809292649098474, "grad_norm": 3.216644763946533, "learning_rate": 1e-06, "loss": 0.0934, "step": 2829 }, { "epoch": 0.9812760055478502, "grad_norm": 2.851259231567383, "learning_rate": 1e-06, "loss": 0.1455, "step": 2830 }, { "epoch": 0.981622746185853, "grad_norm": 3.037447452545166, "learning_rate": 1e-06, "loss": 0.1392, "step": 2831 }, { "epoch": 0.9819694868238558, "grad_norm": 3.316592216491699, "learning_rate": 1e-06, "loss": 0.1772, "step": 2832 }, { "epoch": 0.9823162274618585, "grad_norm": 2.238222599029541, "learning_rate": 1e-06, "loss": 0.1162, "step": 2833 }, { "epoch": 0.9826629680998613, "grad_norm": 2.7036538124084473, "learning_rate": 1e-06, "loss": 0.1116, "step": 2834 }, { "epoch": 0.9830097087378641, "grad_norm": 2.6380455493927, "learning_rate": 1e-06, "loss": 0.0934, "step": 2835 }, { "epoch": 0.9833564493758669, "grad_norm": 2.5659232139587402, "learning_rate": 1e-06, "loss": 0.1418, "step": 2836 }, { "epoch": 0.9837031900138696, "grad_norm": 3.315857172012329, "learning_rate": 1e-06, "loss": 0.16, "step": 2837 }, { "epoch": 0.9840499306518724, "grad_norm": 2.2236742973327637, "learning_rate": 1e-06, "loss": 0.1433, "step": 2838 }, { "epoch": 0.9843966712898752, "grad_norm": 2.794739246368408, "learning_rate": 1e-06, "loss": 0.1276, "step": 2839 }, { "epoch": 0.984743411927878, "grad_norm": 1.8826152086257935, "learning_rate": 1e-06, "loss": 0.1023, "step": 2840 }, { "epoch": 0.9850901525658807, "grad_norm": 2.971917152404785, "learning_rate": 1e-06, "loss": 0.1238, "step": 2841 }, { "epoch": 0.9854368932038835, "grad_norm": 3.801140069961548, "learning_rate": 1e-06, "loss": 0.1215, "step": 2842 }, { "epoch": 0.9857836338418863, "grad_norm": 5.8529276847839355, "learning_rate": 1e-06, "loss": 0.1557, "step": 2843 }, { "epoch": 0.986130374479889, "grad_norm": 3.4524965286254883, "learning_rate": 1e-06, "loss": 0.1246, "step": 2844 }, { "epoch": 0.9864771151178918, "grad_norm": 2.8966329097747803, "learning_rate": 1e-06, "loss": 0.1083, "step": 2845 }, { "epoch": 0.9868238557558946, "grad_norm": 3.0084590911865234, "learning_rate": 1e-06, "loss": 0.1523, "step": 2846 }, { "epoch": 0.9871705963938974, "grad_norm": 2.85929536819458, "learning_rate": 1e-06, "loss": 0.1506, "step": 2847 }, { "epoch": 0.9875173370319001, "grad_norm": 3.1562740802764893, "learning_rate": 1e-06, "loss": 0.1013, "step": 2848 }, { "epoch": 0.9878640776699029, "grad_norm": 2.7616796493530273, "learning_rate": 1e-06, "loss": 0.139, "step": 2849 }, { "epoch": 0.9882108183079057, "grad_norm": 2.781879186630249, "learning_rate": 1e-06, "loss": 0.164, "step": 2850 }, { "epoch": 0.9885575589459085, "grad_norm": 2.50597882270813, "learning_rate": 1e-06, "loss": 0.1022, "step": 2851 }, { "epoch": 0.9889042995839112, "grad_norm": 4.0737833976745605, "learning_rate": 1e-06, "loss": 0.1652, "step": 2852 }, { "epoch": 0.989251040221914, "grad_norm": 3.9901700019836426, "learning_rate": 1e-06, "loss": 0.1411, "step": 2853 }, { "epoch": 0.9895977808599168, "grad_norm": 2.159637928009033, "learning_rate": 1e-06, "loss": 0.1058, "step": 2854 }, { "epoch": 0.9899445214979196, "grad_norm": 3.102248430252075, "learning_rate": 1e-06, "loss": 0.136, "step": 2855 }, { "epoch": 0.9902912621359223, "grad_norm": 2.2245311737060547, "learning_rate": 1e-06, "loss": 0.1498, "step": 2856 }, { "epoch": 0.9906380027739251, "grad_norm": 3.013869524002075, "learning_rate": 1e-06, "loss": 0.1276, "step": 2857 }, { "epoch": 0.9909847434119279, "grad_norm": 2.9369828701019287, "learning_rate": 1e-06, "loss": 0.1353, "step": 2858 }, { "epoch": 0.9913314840499307, "grad_norm": 2.4606969356536865, "learning_rate": 1e-06, "loss": 0.1251, "step": 2859 }, { "epoch": 0.9916782246879334, "grad_norm": 2.1041226387023926, "learning_rate": 1e-06, "loss": 0.1168, "step": 2860 }, { "epoch": 0.9920249653259362, "grad_norm": 2.6059367656707764, "learning_rate": 1e-06, "loss": 0.1419, "step": 2861 }, { "epoch": 0.992371705963939, "grad_norm": 2.68571138381958, "learning_rate": 1e-06, "loss": 0.147, "step": 2862 }, { "epoch": 0.9927184466019418, "grad_norm": 2.2065820693969727, "learning_rate": 1e-06, "loss": 0.1141, "step": 2863 }, { "epoch": 0.9930651872399445, "grad_norm": 2.7700083255767822, "learning_rate": 1e-06, "loss": 0.1321, "step": 2864 }, { "epoch": 0.9934119278779473, "grad_norm": 1.815712571144104, "learning_rate": 1e-06, "loss": 0.104, "step": 2865 }, { "epoch": 0.9937586685159501, "grad_norm": 2.593801259994507, "learning_rate": 1e-06, "loss": 0.1425, "step": 2866 }, { "epoch": 0.9941054091539528, "grad_norm": 2.9171457290649414, "learning_rate": 1e-06, "loss": 0.1342, "step": 2867 }, { "epoch": 0.9944521497919556, "grad_norm": 4.014968395233154, "learning_rate": 1e-06, "loss": 0.1385, "step": 2868 }, { "epoch": 0.9947988904299584, "grad_norm": 2.8752143383026123, "learning_rate": 1e-06, "loss": 0.1386, "step": 2869 }, { "epoch": 0.9951456310679612, "grad_norm": 3.8755412101745605, "learning_rate": 1e-06, "loss": 0.1815, "step": 2870 }, { "epoch": 0.9954923717059639, "grad_norm": 2.6354281902313232, "learning_rate": 1e-06, "loss": 0.1115, "step": 2871 }, { "epoch": 0.9958391123439667, "grad_norm": 2.4390392303466797, "learning_rate": 1e-06, "loss": 0.1543, "step": 2872 }, { "epoch": 0.9961858529819695, "grad_norm": 2.396690607070923, "learning_rate": 1e-06, "loss": 0.1096, "step": 2873 }, { "epoch": 0.9965325936199723, "grad_norm": 2.379737615585327, "learning_rate": 1e-06, "loss": 0.1426, "step": 2874 }, { "epoch": 0.996879334257975, "grad_norm": 3.5744974613189697, "learning_rate": 1e-06, "loss": 0.1485, "step": 2875 }, { "epoch": 0.9972260748959778, "grad_norm": 2.513230323791504, "learning_rate": 1e-06, "loss": 0.1598, "step": 2876 }, { "epoch": 0.9975728155339806, "grad_norm": 2.260457992553711, "learning_rate": 1e-06, "loss": 0.1226, "step": 2877 }, { "epoch": 0.9979195561719834, "grad_norm": 2.7690110206604004, "learning_rate": 1e-06, "loss": 0.1366, "step": 2878 }, { "epoch": 0.9982662968099861, "grad_norm": 2.5981392860412598, "learning_rate": 1e-06, "loss": 0.1333, "step": 2879 }, { "epoch": 0.9986130374479889, "grad_norm": 3.166869878768921, "learning_rate": 1e-06, "loss": 0.1061, "step": 2880 }, { "epoch": 0.9989597780859917, "grad_norm": 1.9656025171279907, "learning_rate": 1e-06, "loss": 0.12, "step": 2881 }, { "epoch": 0.9993065187239945, "grad_norm": 2.704719066619873, "learning_rate": 1e-06, "loss": 0.124, "step": 2882 }, { "epoch": 0.9996532593619972, "grad_norm": 2.3944320678710938, "learning_rate": 1e-06, "loss": 0.1216, "step": 2883 }, { "epoch": 1.0, "grad_norm": 2.362536907196045, "learning_rate": 1e-06, "loss": 0.1326, "step": 2884 }, { "epoch": 1.0, "step": 2884, "total_flos": 9.774378103106175e+18, "train_loss": 0.16740735881443908, "train_runtime": 15269.0756, "train_samples_per_second": 24.172, "train_steps_per_second": 0.189 } ], "logging_steps": 1.0, "max_steps": 2884, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 9.774378103106175e+18, "train_batch_size": 16, "trial_name": null, "trial_params": null }