{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.9989550679205852, "eval_steps": 500, "global_step": 239, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0041797283176593526, "grad_norm": 3.793109431652244, "learning_rate": 2.5e-06, "loss": 1.3773, "step": 1 }, { "epoch": 0.008359456635318705, "grad_norm": 3.860176928138404, "learning_rate": 5e-06, "loss": 1.3697, "step": 2 }, { "epoch": 0.012539184952978056, "grad_norm": 3.472574916996383, "learning_rate": 7.500000000000001e-06, "loss": 1.3685, "step": 3 }, { "epoch": 0.01671891327063741, "grad_norm": 2.426017601116389, "learning_rate": 1e-05, "loss": 1.2106, "step": 4 }, { "epoch": 0.02089864158829676, "grad_norm": 2.369903447078516, "learning_rate": 1.25e-05, "loss": 1.2334, "step": 5 }, { "epoch": 0.025078369905956112, "grad_norm": 6.10573157991006, "learning_rate": 1.5000000000000002e-05, "loss": 1.1659, "step": 6 }, { "epoch": 0.029258098223615466, "grad_norm": 2.828089970728912, "learning_rate": 1.7500000000000002e-05, "loss": 1.1133, "step": 7 }, { "epoch": 0.03343782654127482, "grad_norm": 3.7326281687198226, "learning_rate": 2e-05, "loss": 1.031, "step": 8 }, { "epoch": 0.03761755485893417, "grad_norm": 2.8226722768014127, "learning_rate": 1.9999075218579184e-05, "loss": 1.0627, "step": 9 }, { "epoch": 0.04179728317659352, "grad_norm": 2.3941660915552196, "learning_rate": 1.9996301045360874e-05, "loss": 0.8942, "step": 10 }, { "epoch": 0.04597701149425287, "grad_norm": 2.333154808520554, "learning_rate": 1.9991677993445832e-05, "loss": 0.8475, "step": 11 }, { "epoch": 0.050156739811912224, "grad_norm": 2.1073095025076003, "learning_rate": 1.9985206917896563e-05, "loss": 0.7987, "step": 12 }, { "epoch": 0.054336468129571575, "grad_norm": 2.0147413946396506, "learning_rate": 1.9976889015579167e-05, "loss": 0.7122, "step": 13 }, { "epoch": 0.05851619644723093, "grad_norm": 1.7762710304793987, "learning_rate": 1.9966725824941933e-05, "loss": 0.6907, "step": 14 }, { "epoch": 0.06269592476489028, "grad_norm": 1.854790476944044, "learning_rate": 1.9954719225730847e-05, "loss": 0.583, "step": 15 }, { "epoch": 0.06687565308254964, "grad_norm": 1.6431498121799215, "learning_rate": 1.994087143864188e-05, "loss": 0.5144, "step": 16 }, { "epoch": 0.07105538140020899, "grad_norm": 1.5849345385422475, "learning_rate": 1.992518502491028e-05, "loss": 0.4662, "step": 17 }, { "epoch": 0.07523510971786834, "grad_norm": 2.0976691689662488, "learning_rate": 1.9907662885836836e-05, "loss": 0.5641, "step": 18 }, { "epoch": 0.0794148380355277, "grad_norm": 1.8591436531727428, "learning_rate": 1.9888308262251286e-05, "loss": 0.5513, "step": 19 }, { "epoch": 0.08359456635318704, "grad_norm": 1.6077371820716417, "learning_rate": 1.986712473391289e-05, "loss": 0.514, "step": 20 }, { "epoch": 0.0877742946708464, "grad_norm": 1.6369595572017508, "learning_rate": 1.9844116218848335e-05, "loss": 0.4729, "step": 21 }, { "epoch": 0.09195402298850575, "grad_norm": 1.4143725432673173, "learning_rate": 1.9819286972627066e-05, "loss": 0.4404, "step": 22 }, { "epoch": 0.0961337513061651, "grad_norm": 1.9394850939365578, "learning_rate": 1.9792641587574212e-05, "loss": 0.4529, "step": 23 }, { "epoch": 0.10031347962382445, "grad_norm": 1.4086541686064722, "learning_rate": 1.9764184991921178e-05, "loss": 0.4101, "step": 24 }, { "epoch": 0.1044932079414838, "grad_norm": 2.563989330153671, "learning_rate": 1.973392244889415e-05, "loss": 0.3327, "step": 25 }, { "epoch": 0.10867293625914315, "grad_norm": 1.4054095165182117, "learning_rate": 1.9701859555740647e-05, "loss": 0.3673, "step": 26 }, { "epoch": 0.11285266457680251, "grad_norm": 1.3347083701406377, "learning_rate": 1.966800224269424e-05, "loss": 0.4013, "step": 27 }, { "epoch": 0.11703239289446186, "grad_norm": 1.3599751397794146, "learning_rate": 1.9632356771877735e-05, "loss": 0.3765, "step": 28 }, { "epoch": 0.12121212121212122, "grad_norm": 1.2994011828520489, "learning_rate": 1.9594929736144978e-05, "loss": 0.3092, "step": 29 }, { "epoch": 0.12539184952978055, "grad_norm": 1.4307464855692031, "learning_rate": 1.955572805786141e-05, "loss": 0.352, "step": 30 }, { "epoch": 0.12957157784743992, "grad_norm": 1.1156432595068917, "learning_rate": 1.9514758987623784e-05, "loss": 0.2338, "step": 31 }, { "epoch": 0.13375130616509928, "grad_norm": 1.2383396612081838, "learning_rate": 1.9472030102919102e-05, "loss": 0.3058, "step": 32 }, { "epoch": 0.13793103448275862, "grad_norm": 1.3484241033815327, "learning_rate": 1.94275493067231e-05, "loss": 0.3032, "step": 33 }, { "epoch": 0.14211076280041798, "grad_norm": 1.1814763693141177, "learning_rate": 1.938132482603856e-05, "loss": 0.2606, "step": 34 }, { "epoch": 0.14629049111807732, "grad_norm": 1.1152475267807018, "learning_rate": 1.9333365210373668e-05, "loss": 0.2655, "step": 35 }, { "epoch": 0.15047021943573669, "grad_norm": 1.0274673205489633, "learning_rate": 1.9283679330160726e-05, "loss": 0.2393, "step": 36 }, { "epoch": 0.15464994775339602, "grad_norm": 1.1446059182649253, "learning_rate": 1.9232276375115517e-05, "loss": 0.3053, "step": 37 }, { "epoch": 0.1588296760710554, "grad_norm": 1.2083159051790577, "learning_rate": 1.9179165852537596e-05, "loss": 0.2887, "step": 38 }, { "epoch": 0.16300940438871472, "grad_norm": 1.5658760108502938, "learning_rate": 1.9124357585551872e-05, "loss": 0.3527, "step": 39 }, { "epoch": 0.1671891327063741, "grad_norm": 1.3279092492462565, "learning_rate": 1.9067861711291744e-05, "loss": 0.3117, "step": 40 }, { "epoch": 0.17136886102403343, "grad_norm": 1.177122725805587, "learning_rate": 1.900968867902419e-05, "loss": 0.2835, "step": 41 }, { "epoch": 0.1755485893416928, "grad_norm": 1.0091524997429624, "learning_rate": 1.89498492482171e-05, "loss": 0.2474, "step": 42 }, { "epoch": 0.17972831765935215, "grad_norm": 0.993604546498234, "learning_rate": 1.8888354486549238e-05, "loss": 0.2289, "step": 43 }, { "epoch": 0.1839080459770115, "grad_norm": 1.3521657755483074, "learning_rate": 1.8825215767863215e-05, "loss": 0.3526, "step": 44 }, { "epoch": 0.18808777429467086, "grad_norm": 1.1636752310323113, "learning_rate": 1.876044477006183e-05, "loss": 0.3034, "step": 45 }, { "epoch": 0.1922675026123302, "grad_norm": 0.9201202294799887, "learning_rate": 1.8694053472948154e-05, "loss": 0.2322, "step": 46 }, { "epoch": 0.19644723092998956, "grad_norm": 1.0591767636272813, "learning_rate": 1.8626054156009807e-05, "loss": 0.1968, "step": 47 }, { "epoch": 0.2006269592476489, "grad_norm": 1.1802357695207026, "learning_rate": 1.8556459396147777e-05, "loss": 0.248, "step": 48 }, { "epoch": 0.20480668756530826, "grad_norm": 1.05844609547146, "learning_rate": 1.8485282065350237e-05, "loss": 0.2361, "step": 49 }, { "epoch": 0.2089864158829676, "grad_norm": 0.8493872312825937, "learning_rate": 1.8412535328311813e-05, "loss": 0.163, "step": 50 }, { "epoch": 0.21316614420062696, "grad_norm": 1.1938936546063497, "learning_rate": 1.8338232639998672e-05, "loss": 0.2011, "step": 51 }, { "epoch": 0.2173458725182863, "grad_norm": 0.9821596876965665, "learning_rate": 1.826238774315995e-05, "loss": 0.1797, "step": 52 }, { "epoch": 0.22152560083594566, "grad_norm": 1.0890806155634258, "learning_rate": 1.8185014665785936e-05, "loss": 0.2356, "step": 53 }, { "epoch": 0.22570532915360503, "grad_norm": 3.1623969781776875, "learning_rate": 1.810612771851352e-05, "loss": 0.2341, "step": 54 }, { "epoch": 0.22988505747126436, "grad_norm": 1.0495429246657253, "learning_rate": 1.8025741491979326e-05, "loss": 0.2286, "step": 55 }, { "epoch": 0.23406478578892373, "grad_norm": 1.0998745172943698, "learning_rate": 1.7943870854121126e-05, "loss": 0.228, "step": 56 }, { "epoch": 0.23824451410658307, "grad_norm": 1.0611257780697756, "learning_rate": 1.7860530947427878e-05, "loss": 0.1926, "step": 57 }, { "epoch": 0.24242424242424243, "grad_norm": 1.3963883396571588, "learning_rate": 1.777573718613904e-05, "loss": 0.2455, "step": 58 }, { "epoch": 0.24660397074190177, "grad_norm": 0.8817393289325651, "learning_rate": 1.768950525339362e-05, "loss": 0.186, "step": 59 }, { "epoch": 0.2507836990595611, "grad_norm": 1.004709836073224, "learning_rate": 1.7601851098329484e-05, "loss": 0.2003, "step": 60 }, { "epoch": 0.2549634273772205, "grad_norm": 1.009032284413222, "learning_rate": 1.7512790933133435e-05, "loss": 0.2389, "step": 61 }, { "epoch": 0.25914315569487983, "grad_norm": 1.0842648970922986, "learning_rate": 1.74223412300427e-05, "loss": 0.2324, "step": 62 }, { "epoch": 0.26332288401253917, "grad_norm": 1.1362114248016657, "learning_rate": 1.7330518718298263e-05, "loss": 0.2636, "step": 63 }, { "epoch": 0.26750261233019856, "grad_norm": 1.2490219244647938, "learning_rate": 1.72373403810507e-05, "loss": 0.2392, "step": 64 }, { "epoch": 0.2716823406478579, "grad_norm": 0.9229560381350037, "learning_rate": 1.7142823452219036e-05, "loss": 0.2207, "step": 65 }, { "epoch": 0.27586206896551724, "grad_norm": 0.9194163633451438, "learning_rate": 1.7046985413303215e-05, "loss": 0.2208, "step": 66 }, { "epoch": 0.2800417972831766, "grad_norm": 0.9431299303955618, "learning_rate": 1.6949843990150798e-05, "loss": 0.1877, "step": 67 }, { "epoch": 0.28422152560083597, "grad_norm": 1.0695148700733295, "learning_rate": 1.6851417149678442e-05, "loss": 0.2431, "step": 68 }, { "epoch": 0.2884012539184953, "grad_norm": 0.9981918594046101, "learning_rate": 1.6751723096548834e-05, "loss": 0.2134, "step": 69 }, { "epoch": 0.29258098223615464, "grad_norm": 0.7779428741827196, "learning_rate": 1.6650780269803587e-05, "loss": 0.1772, "step": 70 }, { "epoch": 0.296760710553814, "grad_norm": 0.7017283200072758, "learning_rate": 1.6548607339452853e-05, "loss": 0.154, "step": 71 }, { "epoch": 0.30094043887147337, "grad_norm": 1.2657237042635219, "learning_rate": 1.644522320302217e-05, "loss": 0.1989, "step": 72 }, { "epoch": 0.3051201671891327, "grad_norm": 0.9141216726104325, "learning_rate": 1.634064698205725e-05, "loss": 0.1632, "step": 73 }, { "epoch": 0.30929989550679204, "grad_norm": 0.853915517926657, "learning_rate": 1.6234898018587336e-05, "loss": 0.2127, "step": 74 }, { "epoch": 0.31347962382445144, "grad_norm": 1.0323735847419797, "learning_rate": 1.612799587154777e-05, "loss": 0.2167, "step": 75 }, { "epoch": 0.3176593521421108, "grad_norm": 0.930984497217397, "learning_rate": 1.6019960313162436e-05, "loss": 0.1956, "step": 76 }, { "epoch": 0.3218390804597701, "grad_norm": 1.006790474010485, "learning_rate": 1.5910811325286768e-05, "loss": 0.2224, "step": 77 }, { "epoch": 0.32601880877742945, "grad_norm": 0.6839140410559078, "learning_rate": 1.5800569095711983e-05, "loss": 0.1534, "step": 78 }, { "epoch": 0.33019853709508884, "grad_norm": 0.9218639125486777, "learning_rate": 1.5689254014431225e-05, "loss": 0.2173, "step": 79 }, { "epoch": 0.3343782654127482, "grad_norm": 0.9628453893674432, "learning_rate": 1.5576886669868297e-05, "loss": 0.2144, "step": 80 }, { "epoch": 0.3385579937304075, "grad_norm": 1.0976706938526015, "learning_rate": 1.5463487845069708e-05, "loss": 0.2869, "step": 81 }, { "epoch": 0.34273772204806685, "grad_norm": 0.843877293720617, "learning_rate": 1.5349078513860728e-05, "loss": 0.1644, "step": 82 }, { "epoch": 0.34691745036572624, "grad_norm": 1.0395155619197012, "learning_rate": 1.5233679836966122e-05, "loss": 0.2048, "step": 83 }, { "epoch": 0.3510971786833856, "grad_norm": 0.62654878377622, "learning_rate": 1.5117313158096371e-05, "loss": 0.1011, "step": 84 }, { "epoch": 0.3552769070010449, "grad_norm": 0.8765627543230693, "learning_rate": 1.5000000000000002e-05, "loss": 0.2282, "step": 85 }, { "epoch": 0.3594566353187043, "grad_norm": 0.8493012101164317, "learning_rate": 1.4881762060482814e-05, "loss": 0.2303, "step": 86 }, { "epoch": 0.36363636363636365, "grad_norm": 0.9061911256912499, "learning_rate": 1.476262120839475e-05, "loss": 0.1772, "step": 87 }, { "epoch": 0.367816091954023, "grad_norm": 0.7714054432768603, "learning_rate": 1.4642599479585106e-05, "loss": 0.1827, "step": 88 }, { "epoch": 0.3719958202716823, "grad_norm": 1.0302753021934954, "learning_rate": 1.4521719072826858e-05, "loss": 0.2556, "step": 89 }, { "epoch": 0.3761755485893417, "grad_norm": 0.8170194188293493, "learning_rate": 1.4400002345710871e-05, "loss": 0.2131, "step": 90 }, { "epoch": 0.38035527690700105, "grad_norm": 0.8353229514513874, "learning_rate": 1.427747181051071e-05, "loss": 0.1774, "step": 91 }, { "epoch": 0.3845350052246604, "grad_norm": 0.8996783507158662, "learning_rate": 1.4154150130018867e-05, "loss": 0.2074, "step": 92 }, { "epoch": 0.3887147335423197, "grad_norm": 1.0120263532766836, "learning_rate": 1.4030060113355118e-05, "loss": 0.2291, "step": 93 }, { "epoch": 0.3928944618599791, "grad_norm": 0.8009583936768994, "learning_rate": 1.3905224711747844e-05, "loss": 0.1793, "step": 94 }, { "epoch": 0.39707419017763845, "grad_norm": 0.8764394706115101, "learning_rate": 1.3779667014289067e-05, "loss": 0.1639, "step": 95 }, { "epoch": 0.4012539184952978, "grad_norm": 1.2665682983369955, "learning_rate": 1.3653410243663953e-05, "loss": 0.2469, "step": 96 }, { "epoch": 0.4054336468129572, "grad_norm": 2.02790558792064, "learning_rate": 1.3526477751855645e-05, "loss": 0.2458, "step": 97 }, { "epoch": 0.4096133751306165, "grad_norm": 0.8765418818205535, "learning_rate": 1.3398893015826166e-05, "loss": 0.2141, "step": 98 }, { "epoch": 0.41379310344827586, "grad_norm": 0.8402084285563245, "learning_rate": 1.3270679633174219e-05, "loss": 0.17, "step": 99 }, { "epoch": 0.4179728317659352, "grad_norm": 0.8927111902324383, "learning_rate": 1.3141861317770628e-05, "loss": 0.2419, "step": 100 }, { "epoch": 0.4221525600835946, "grad_norm": 0.9307548734337794, "learning_rate": 1.3012461895372343e-05, "loss": 0.2211, "step": 101 }, { "epoch": 0.4263322884012539, "grad_norm": 0.8182965551610069, "learning_rate": 1.2882505299215711e-05, "loss": 0.1546, "step": 102 }, { "epoch": 0.43051201671891326, "grad_norm": 0.5513459236136672, "learning_rate": 1.2752015565589852e-05, "loss": 0.1028, "step": 103 }, { "epoch": 0.4346917450365726, "grad_norm": 0.7895344671438952, "learning_rate": 1.2621016829391022e-05, "loss": 0.147, "step": 104 }, { "epoch": 0.438871473354232, "grad_norm": 0.7347726170918327, "learning_rate": 1.2489533319658703e-05, "loss": 0.1499, "step": 105 }, { "epoch": 0.4430512016718913, "grad_norm": 0.9930172706392331, "learning_rate": 1.2357589355094275e-05, "loss": 0.1999, "step": 106 }, { "epoch": 0.44723092998955066, "grad_norm": 0.700736172068482, "learning_rate": 1.2225209339563144e-05, "loss": 0.1727, "step": 107 }, { "epoch": 0.45141065830721006, "grad_norm": 0.9578584031611264, "learning_rate": 1.2092417757581085e-05, "loss": 0.2408, "step": 108 }, { "epoch": 0.4555903866248694, "grad_norm": 0.8203132426685221, "learning_rate": 1.1959239169785668e-05, "loss": 0.1951, "step": 109 }, { "epoch": 0.45977011494252873, "grad_norm": 0.9270712768052833, "learning_rate": 1.182569820839362e-05, "loss": 0.2222, "step": 110 }, { "epoch": 0.46394984326018807, "grad_norm": 0.9063435343774252, "learning_rate": 1.1691819572644941e-05, "loss": 0.1827, "step": 111 }, { "epoch": 0.46812957157784746, "grad_norm": 0.9742349956564444, "learning_rate": 1.155762802423463e-05, "loss": 0.2166, "step": 112 }, { "epoch": 0.4723092998955068, "grad_norm": 0.740191368731416, "learning_rate": 1.1423148382732854e-05, "loss": 0.162, "step": 113 }, { "epoch": 0.47648902821316613, "grad_norm": 2.152737189786801, "learning_rate": 1.128840552099439e-05, "loss": 0.2347, "step": 114 }, { "epoch": 0.48066875653082547, "grad_norm": 0.6837438960691173, "learning_rate": 1.1153424360558268e-05, "loss": 0.1344, "step": 115 }, { "epoch": 0.48484848484848486, "grad_norm": 0.8928037881326962, "learning_rate": 1.1018229867038358e-05, "loss": 0.1723, "step": 116 }, { "epoch": 0.4890282131661442, "grad_norm": 1.004544035476834, "learning_rate": 1.0882847045505809e-05, "loss": 0.1962, "step": 117 }, { "epoch": 0.49320794148380354, "grad_norm": 0.841348260273773, "learning_rate": 1.0747300935864245e-05, "loss": 0.1803, "step": 118 }, { "epoch": 0.49738766980146293, "grad_norm": 0.8496561849066544, "learning_rate": 1.0611616608218429e-05, "loss": 0.1905, "step": 119 }, { "epoch": 0.5015673981191222, "grad_norm": 1.1461318580041029, "learning_rate": 1.0475819158237426e-05, "loss": 0.2435, "step": 120 }, { "epoch": 0.5057471264367817, "grad_norm": 0.9171413538232467, "learning_rate": 1.0339933702512978e-05, "loss": 0.253, "step": 121 }, { "epoch": 0.509926854754441, "grad_norm": 0.7285731725801685, "learning_rate": 1.0203985373914056e-05, "loss": 0.1548, "step": 122 }, { "epoch": 0.5141065830721003, "grad_norm": 0.8652478932288088, "learning_rate": 1.0067999316938348e-05, "loss": 0.204, "step": 123 }, { "epoch": 0.5182863113897597, "grad_norm": 1.38915125451696, "learning_rate": 9.932000683061654e-06, "loss": 0.2442, "step": 124 }, { "epoch": 0.522466039707419, "grad_norm": 0.9268619467285734, "learning_rate": 9.79601462608595e-06, "loss": 0.2297, "step": 125 }, { "epoch": 0.5266457680250783, "grad_norm": 0.7880807499102447, "learning_rate": 9.660066297487024e-06, "loss": 0.181, "step": 126 }, { "epoch": 0.5308254963427377, "grad_norm": 0.754337068341988, "learning_rate": 9.524180841762577e-06, "loss": 0.1641, "step": 127 }, { "epoch": 0.5350052246603971, "grad_norm": 1.1050970913178217, "learning_rate": 9.388383391781576e-06, "loss": 0.2412, "step": 128 }, { "epoch": 0.5391849529780565, "grad_norm": 0.911580865548536, "learning_rate": 9.252699064135759e-06, "loss": 0.2092, "step": 129 }, { "epoch": 0.5433646812957158, "grad_norm": 0.8679286773035769, "learning_rate": 9.117152954494195e-06, "loss": 0.1902, "step": 130 }, { "epoch": 0.5475444096133751, "grad_norm": 0.76185165053938, "learning_rate": 8.981770132961649e-06, "loss": 0.2015, "step": 131 }, { "epoch": 0.5517241379310345, "grad_norm": 0.6153250846939452, "learning_rate": 8.846575639441732e-06, "loss": 0.1301, "step": 132 }, { "epoch": 0.5559038662486938, "grad_norm": 0.8129974861108137, "learning_rate": 8.711594479005614e-06, "loss": 0.1688, "step": 133 }, { "epoch": 0.5600835945663531, "grad_norm": 0.8932210165728961, "learning_rate": 8.576851617267151e-06, "loss": 0.142, "step": 134 }, { "epoch": 0.5642633228840125, "grad_norm": 1.2480488477623177, "learning_rate": 8.442371975765368e-06, "loss": 0.2117, "step": 135 }, { "epoch": 0.5684430512016719, "grad_norm": 1.2436168746130591, "learning_rate": 8.308180427355062e-06, "loss": 0.2376, "step": 136 }, { "epoch": 0.5726227795193313, "grad_norm": 0.8002711459729407, "learning_rate": 8.174301791606384e-06, "loss": 0.1827, "step": 137 }, { "epoch": 0.5768025078369906, "grad_norm": 0.8475588556028547, "learning_rate": 8.040760830214334e-06, "loss": 0.1658, "step": 138 }, { "epoch": 0.58098223615465, "grad_norm": 0.8639184920307346, "learning_rate": 7.907582242418916e-06, "loss": 0.17, "step": 139 }, { "epoch": 0.5851619644723093, "grad_norm": 0.8598440958255673, "learning_rate": 7.774790660436857e-06, "loss": 0.2112, "step": 140 }, { "epoch": 0.5893416927899686, "grad_norm": 0.8792127016276399, "learning_rate": 7.642410644905726e-06, "loss": 0.2157, "step": 141 }, { "epoch": 0.593521421107628, "grad_norm": 0.5310794775676358, "learning_rate": 7.5104666803413015e-06, "loss": 0.0995, "step": 142 }, { "epoch": 0.5977011494252874, "grad_norm": 0.6031042659903776, "learning_rate": 7.378983170608982e-06, "loss": 0.1226, "step": 143 }, { "epoch": 0.6018808777429467, "grad_norm": 0.9322591026745743, "learning_rate": 7.24798443441015e-06, "loss": 0.2269, "step": 144 }, { "epoch": 0.6060606060606061, "grad_norm": 0.7650887090451826, "learning_rate": 7.117494700784292e-06, "loss": 0.1972, "step": 145 }, { "epoch": 0.6102403343782654, "grad_norm": 1.08702051590229, "learning_rate": 6.9875381046276605e-06, "loss": 0.2312, "step": 146 }, { "epoch": 0.6144200626959248, "grad_norm": 1.1938077188160876, "learning_rate": 6.8581386822293765e-06, "loss": 0.1944, "step": 147 }, { "epoch": 0.6185997910135841, "grad_norm": 0.8286648541039914, "learning_rate": 6.729320366825785e-06, "loss": 0.2012, "step": 148 }, { "epoch": 0.6227795193312434, "grad_norm": 1.0448385340552901, "learning_rate": 6.601106984173835e-06, "loss": 0.176, "step": 149 }, { "epoch": 0.6269592476489029, "grad_norm": 0.6431098205020193, "learning_rate": 6.473522248144359e-06, "loss": 0.1611, "step": 150 }, { "epoch": 0.6311389759665622, "grad_norm": 0.6902192399005279, "learning_rate": 6.34658975633605e-06, "loss": 0.1571, "step": 151 }, { "epoch": 0.6353187042842215, "grad_norm": 1.4126319469067181, "learning_rate": 6.220332985710936e-06, "loss": 0.2253, "step": 152 }, { "epoch": 0.6394984326018809, "grad_norm": 0.9250333234508288, "learning_rate": 6.094775288252157e-06, "loss": 0.1659, "step": 153 }, { "epoch": 0.6436781609195402, "grad_norm": 1.0996139187104108, "learning_rate": 5.9699398866448846e-06, "loss": 0.1927, "step": 154 }, { "epoch": 0.6478578892371996, "grad_norm": 0.6600170395773419, "learning_rate": 5.845849869981137e-06, "loss": 0.1504, "step": 155 }, { "epoch": 0.6520376175548589, "grad_norm": 0.8091133996097659, "learning_rate": 5.722528189489294e-06, "loss": 0.1728, "step": 156 }, { "epoch": 0.6562173458725182, "grad_norm": 0.6787441434180406, "learning_rate": 5.599997654289129e-06, "loss": 0.1306, "step": 157 }, { "epoch": 0.6603970741901777, "grad_norm": 0.8602639720659184, "learning_rate": 5.478280927173145e-06, "loss": 0.168, "step": 158 }, { "epoch": 0.664576802507837, "grad_norm": 0.6826930291022159, "learning_rate": 5.357400520414898e-06, "loss": 0.1205, "step": 159 }, { "epoch": 0.6687565308254964, "grad_norm": 0.6701217983786535, "learning_rate": 5.237378791605249e-06, "loss": 0.1452, "step": 160 }, { "epoch": 0.6729362591431557, "grad_norm": 0.7578669897505922, "learning_rate": 5.11823793951719e-06, "loss": 0.1626, "step": 161 }, { "epoch": 0.677115987460815, "grad_norm": 0.8367106457596803, "learning_rate": 5.000000000000003e-06, "loss": 0.1981, "step": 162 }, { "epoch": 0.6812957157784744, "grad_norm": 0.8324385169673411, "learning_rate": 4.882686841903627e-06, "loss": 0.1945, "step": 163 }, { "epoch": 0.6854754440961337, "grad_norm": 0.8336564812001488, "learning_rate": 4.766320163033882e-06, "loss": 0.1497, "step": 164 }, { "epoch": 0.6896551724137931, "grad_norm": 0.9106572698316981, "learning_rate": 4.6509214861392785e-06, "loss": 0.2122, "step": 165 }, { "epoch": 0.6938349007314525, "grad_norm": 0.9899598996363259, "learning_rate": 4.5365121549302916e-06, "loss": 0.291, "step": 166 }, { "epoch": 0.6980146290491118, "grad_norm": 0.7526211809216892, "learning_rate": 4.423113330131708e-06, "loss": 0.1398, "step": 167 }, { "epoch": 0.7021943573667712, "grad_norm": 0.580108623854109, "learning_rate": 4.310745985568779e-06, "loss": 0.1228, "step": 168 }, { "epoch": 0.7063740856844305, "grad_norm": 0.5902701942035865, "learning_rate": 4.19943090428802e-06, "loss": 0.122, "step": 169 }, { "epoch": 0.7105538140020898, "grad_norm": 0.8908115982207944, "learning_rate": 4.0891886747132356e-06, "loss": 0.1679, "step": 170 }, { "epoch": 0.7147335423197492, "grad_norm": 0.7610727588905307, "learning_rate": 3.9800396868375675e-06, "loss": 0.1331, "step": 171 }, { "epoch": 0.7189132706374086, "grad_norm": 0.9084312218170052, "learning_rate": 3.872004128452231e-06, "loss": 0.2037, "step": 172 }, { "epoch": 0.723092998955068, "grad_norm": 0.7951214412363438, "learning_rate": 3.7651019814126656e-06, "loss": 0.152, "step": 173 }, { "epoch": 0.7272727272727273, "grad_norm": 1.0007664865334354, "learning_rate": 3.659353017942754e-06, "loss": 0.199, "step": 174 }, { "epoch": 0.7314524555903866, "grad_norm": 0.8000807402307464, "learning_rate": 3.5547767969778355e-06, "loss": 0.1874, "step": 175 }, { "epoch": 0.735632183908046, "grad_norm": 0.7056309894011483, "learning_rate": 3.4513926605471504e-06, "loss": 0.1408, "step": 176 }, { "epoch": 0.7398119122257053, "grad_norm": 1.0154810770219924, "learning_rate": 3.3492197301964145e-06, "loss": 0.1516, "step": 177 }, { "epoch": 0.7439916405433646, "grad_norm": 0.6774580884476967, "learning_rate": 3.248276903451171e-06, "loss": 0.1389, "step": 178 }, { "epoch": 0.7481713688610241, "grad_norm": 1.383727265413641, "learning_rate": 3.1485828503215588e-06, "loss": 0.1652, "step": 179 }, { "epoch": 0.7523510971786834, "grad_norm": 0.9084226985349156, "learning_rate": 3.0501560098492056e-06, "loss": 0.22, "step": 180 }, { "epoch": 0.7565308254963428, "grad_norm": 0.7904229164202767, "learning_rate": 2.9530145866967897e-06, "loss": 0.1688, "step": 181 }, { "epoch": 0.7607105538140021, "grad_norm": 0.9537175407913363, "learning_rate": 2.8571765477809645e-06, "loss": 0.2203, "step": 182 }, { "epoch": 0.7648902821316614, "grad_norm": 0.9045213538470349, "learning_rate": 2.7626596189492983e-06, "loss": 0.2235, "step": 183 }, { "epoch": 0.7690700104493208, "grad_norm": 0.8437761162278906, "learning_rate": 2.669481281701739e-06, "loss": 0.2405, "step": 184 }, { "epoch": 0.7732497387669801, "grad_norm": 0.9170366332680835, "learning_rate": 2.5776587699573007e-06, "loss": 0.2356, "step": 185 }, { "epoch": 0.7774294670846394, "grad_norm": 0.859807092495294, "learning_rate": 2.487209066866565e-06, "loss": 0.1581, "step": 186 }, { "epoch": 0.7816091954022989, "grad_norm": 0.7541718313410231, "learning_rate": 2.398148901670521e-06, "loss": 0.1735, "step": 187 }, { "epoch": 0.7857889237199582, "grad_norm": 0.5298691809105464, "learning_rate": 2.3104947466063785e-06, "loss": 0.1247, "step": 188 }, { "epoch": 0.7899686520376176, "grad_norm": 0.5509131740535222, "learning_rate": 2.224262813860962e-06, "loss": 0.1135, "step": 189 }, { "epoch": 0.7941483803552769, "grad_norm": 1.211014988349712, "learning_rate": 2.1394690525721275e-06, "loss": 0.2396, "step": 190 }, { "epoch": 0.7983281086729362, "grad_norm": 0.8304883104248448, "learning_rate": 2.0561291458788736e-06, "loss": 0.1495, "step": 191 }, { "epoch": 0.8025078369905956, "grad_norm": 0.6575918108532034, "learning_rate": 1.9742585080206754e-06, "loss": 0.1705, "step": 192 }, { "epoch": 0.8066875653082549, "grad_norm": 0.6483828691185768, "learning_rate": 1.8938722814864863e-06, "loss": 0.1179, "step": 193 }, { "epoch": 0.8108672936259144, "grad_norm": 0.7874326139345168, "learning_rate": 1.8149853342140644e-06, "loss": 0.1911, "step": 194 }, { "epoch": 0.8150470219435737, "grad_norm": 0.9396325288901851, "learning_rate": 1.7376122568400533e-06, "loss": 0.2313, "step": 195 }, { "epoch": 0.819226750261233, "grad_norm": 0.8980422324440791, "learning_rate": 1.6617673600013295e-06, "loss": 0.2637, "step": 196 }, { "epoch": 0.8234064785788924, "grad_norm": 0.8393671732445989, "learning_rate": 1.587464671688187e-06, "loss": 0.2008, "step": 197 }, { "epoch": 0.8275862068965517, "grad_norm": 0.9628013968768121, "learning_rate": 1.5147179346497665e-06, "loss": 0.1803, "step": 198 }, { "epoch": 0.831765935214211, "grad_norm": 0.8055867006150553, "learning_rate": 1.443540603852227e-06, "loss": 0.2058, "step": 199 }, { "epoch": 0.8359456635318704, "grad_norm": 0.7402053267386335, "learning_rate": 1.373945843990192e-06, "loss": 0.1678, "step": 200 }, { "epoch": 0.8401253918495298, "grad_norm": 0.6345259422046625, "learning_rate": 1.3059465270518469e-06, "loss": 0.1499, "step": 201 }, { "epoch": 0.8443051201671892, "grad_norm": 0.9278886576312714, "learning_rate": 1.2395552299381742e-06, "loss": 0.1376, "step": 202 }, { "epoch": 0.8484848484848485, "grad_norm": 0.7305237779713204, "learning_rate": 1.1747842321367886e-06, "loss": 0.1689, "step": 203 }, { "epoch": 0.8526645768025078, "grad_norm": 0.785339323897186, "learning_rate": 1.1116455134507665e-06, "loss": 0.1515, "step": 204 }, { "epoch": 0.8568443051201672, "grad_norm": 0.8232817926726907, "learning_rate": 1.0501507517829012e-06, "loss": 0.2148, "step": 205 }, { "epoch": 0.8610240334378265, "grad_norm": 0.920951874634867, "learning_rate": 9.903113209758098e-07, "loss": 0.1981, "step": 206 }, { "epoch": 0.8652037617554859, "grad_norm": 0.778697980092621, "learning_rate": 9.321382887082564e-07, "loss": 0.1697, "step": 207 }, { "epoch": 0.8693834900731452, "grad_norm": 0.552864168373498, "learning_rate": 8.756424144481313e-07, "loss": 0.1197, "step": 208 }, { "epoch": 0.8735632183908046, "grad_norm": 0.7721394559949915, "learning_rate": 8.208341474624071e-07, "loss": 0.1636, "step": 209 }, { "epoch": 0.877742946708464, "grad_norm": 0.5887750742792935, "learning_rate": 7.677236248844855e-07, "loss": 0.1238, "step": 210 }, { "epoch": 0.8819226750261233, "grad_norm": 1.1706298717117896, "learning_rate": 7.163206698392744e-07, "loss": 0.2685, "step": 211 }, { "epoch": 0.8861024033437827, "grad_norm": 0.7156481787290525, "learning_rate": 6.666347896263326e-07, "loss": 0.1272, "step": 212 }, { "epoch": 0.890282131661442, "grad_norm": 0.9148297160556185, "learning_rate": 6.186751739614405e-07, "loss": 0.1878, "step": 213 }, { "epoch": 0.8944618599791013, "grad_norm": 0.8443233125008259, "learning_rate": 5.724506932769014e-07, "loss": 0.1872, "step": 214 }, { "epoch": 0.8986415882967607, "grad_norm": 0.7906265849992707, "learning_rate": 5.279698970809011e-07, "loss": 0.1672, "step": 215 }, { "epoch": 0.9028213166144201, "grad_norm": 0.8000045890229907, "learning_rate": 4.852410123762164e-07, "loss": 0.1357, "step": 216 }, { "epoch": 0.9070010449320794, "grad_norm": 0.7692285277924558, "learning_rate": 4.4427194213859216e-07, "loss": 0.1733, "step": 217 }, { "epoch": 0.9111807732497388, "grad_norm": 0.6828595942343869, "learning_rate": 4.0507026385502747e-07, "loss": 0.1377, "step": 218 }, { "epoch": 0.9153605015673981, "grad_norm": 0.7616205791512801, "learning_rate": 3.6764322812226416e-07, "loss": 0.1743, "step": 219 }, { "epoch": 0.9195402298850575, "grad_norm": 0.6920015184607431, "learning_rate": 3.319977573057642e-07, "loss": 0.178, "step": 220 }, { "epoch": 0.9237199582027168, "grad_norm": 0.8599083603091006, "learning_rate": 2.9814044425935605e-07, "loss": 0.211, "step": 221 }, { "epoch": 0.9278996865203761, "grad_norm": 0.8372366140139196, "learning_rate": 2.6607755110584886e-07, "loss": 0.1693, "step": 222 }, { "epoch": 0.9320794148380356, "grad_norm": 0.740216908527879, "learning_rate": 2.3581500807882462e-07, "loss": 0.1677, "step": 223 }, { "epoch": 0.9362591431556949, "grad_norm": 0.7486951708296801, "learning_rate": 2.0735841242578992e-07, "loss": 0.1367, "step": 224 }, { "epoch": 0.9404388714733543, "grad_norm": 0.6059460975199559, "learning_rate": 1.8071302737293294e-07, "loss": 0.1287, "step": 225 }, { "epoch": 0.9446185997910136, "grad_norm": 0.6725131427516452, "learning_rate": 1.558837811516667e-07, "loss": 0.1355, "step": 226 }, { "epoch": 0.9487983281086729, "grad_norm": 0.89103460281586, "learning_rate": 1.3287526608711132e-07, "loss": 0.2224, "step": 227 }, { "epoch": 0.9529780564263323, "grad_norm": 0.7294258012671165, "learning_rate": 1.1169173774871478e-07, "loss": 0.1247, "step": 228 }, { "epoch": 0.9571577847439916, "grad_norm": 0.5563178146719318, "learning_rate": 9.233711416316571e-08, "loss": 0.1229, "step": 229 }, { "epoch": 0.9613375130616509, "grad_norm": 1.2052590543331665, "learning_rate": 7.481497508972313e-08, "loss": 0.1689, "step": 230 }, { "epoch": 0.9655172413793104, "grad_norm": 0.8195585207091408, "learning_rate": 5.912856135812051e-08, "loss": 0.1942, "step": 231 }, { "epoch": 0.9696969696969697, "grad_norm": 2.0024730819109773, "learning_rate": 4.528077426915412e-08, "loss": 0.1719, "step": 232 }, { "epoch": 0.9738766980146291, "grad_norm": 0.6177697556146066, "learning_rate": 3.327417505806785e-08, "loss": 0.1286, "step": 233 }, { "epoch": 0.9780564263322884, "grad_norm": 2.319773776301439, "learning_rate": 2.311098442083659e-08, "loss": 0.2415, "step": 234 }, { "epoch": 0.9822361546499477, "grad_norm": 1.1389478132400794, "learning_rate": 1.4793082103435885e-08, "loss": 0.1913, "step": 235 }, { "epoch": 0.9864158829676071, "grad_norm": 0.8823489571757573, "learning_rate": 8.322006554171147e-09, "loss": 0.2037, "step": 236 }, { "epoch": 0.9905956112852664, "grad_norm": 0.6469295501938342, "learning_rate": 3.698954639129726e-09, "loss": 0.1321, "step": 237 }, { "epoch": 0.9947753396029259, "grad_norm": 0.8129747939307076, "learning_rate": 9.24781420816867e-10, "loss": 0.194, "step": 238 }, { "epoch": 0.9989550679205852, "grad_norm": 0.9261130483396426, "learning_rate": 0.0, "loss": 0.1793, "step": 239 }, { "epoch": 0.9989550679205852, "step": 239, "total_flos": 449784335171584.0, "train_loss": 0.26206765498451606, "train_runtime": 2412.5124, "train_samples_per_second": 12.688, "train_steps_per_second": 0.099 } ], "logging_steps": 1.0, "max_steps": 239, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 50000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": false, "should_training_stop": false }, "attributes": {} } }, "total_flos": 449784335171584.0, "train_batch_size": 2, "trial_name": null, "trial_params": null }