{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 0.07545835052761894, "eval_steps": 160, "global_step": 480, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00015720489693253945, "grad_norm": 1.3751904964447021, "learning_rate": 0.0, "loss": 3.5741, "step": 1 }, { "epoch": 0.00015720489693253945, "eval_loss": 3.4173049926757812, "eval_runtime": 2315.7248, "eval_samples_per_second": 3.998, "eval_steps_per_second": 1.999, "step": 1 }, { "epoch": 0.0003144097938650789, "grad_norm": 1.231239676475525, "learning_rate": 5e-06, "loss": 3.3021, "step": 2 }, { "epoch": 0.00047161469079761836, "grad_norm": 1.3657807111740112, "learning_rate": 1e-05, "loss": 3.6333, "step": 3 }, { "epoch": 0.0006288195877301578, "grad_norm": 1.3117496967315674, "learning_rate": 1.5e-05, "loss": 3.3731, "step": 4 }, { "epoch": 0.0007860244846626972, "grad_norm": 1.4118576049804688, "learning_rate": 2e-05, "loss": 3.612, "step": 5 }, { "epoch": 0.0009432293815952367, "grad_norm": 1.3155895471572876, "learning_rate": 2.5e-05, "loss": 3.3296, "step": 6 }, { "epoch": 0.001100434278527776, "grad_norm": 1.2847192287445068, "learning_rate": 3e-05, "loss": 3.2168, "step": 7 }, { "epoch": 0.0012576391754603156, "grad_norm": 1.1421078443527222, "learning_rate": 3.5e-05, "loss": 3.085, "step": 8 }, { "epoch": 0.0014148440723928551, "grad_norm": 0.9923035502433777, "learning_rate": 4e-05, "loss": 3.0472, "step": 9 }, { "epoch": 0.0015720489693253944, "grad_norm": 0.795043408870697, "learning_rate": 4.5e-05, "loss": 2.6666, "step": 10 }, { "epoch": 0.001729253866257934, "grad_norm": 0.5987974405288696, "learning_rate": 5e-05, "loss": 2.473, "step": 11 }, { "epoch": 0.0018864587631904734, "grad_norm": 0.4488905668258667, "learning_rate": 4.9999999236547564e-05, "loss": 2.3731, "step": 12 }, { "epoch": 0.002043663660123013, "grad_norm": 0.3517301380634308, "learning_rate": 4.999999694619029e-05, "loss": 2.2158, "step": 13 }, { "epoch": 0.002200868557055552, "grad_norm": 0.3045121431350708, "learning_rate": 4.999999312892831e-05, "loss": 2.3351, "step": 14 }, { "epoch": 0.002358073453988092, "grad_norm": 0.24488244950771332, "learning_rate": 4.9999987784761884e-05, "loss": 2.2693, "step": 15 }, { "epoch": 0.0025152783509206312, "grad_norm": 0.22892728447914124, "learning_rate": 4.999998091369132e-05, "loss": 2.1006, "step": 16 }, { "epoch": 0.0026724832478531705, "grad_norm": 0.23219206929206848, "learning_rate": 4.999997251571704e-05, "loss": 2.215, "step": 17 }, { "epoch": 0.0028296881447857102, "grad_norm": 0.24427154660224915, "learning_rate": 4.999996259083956e-05, "loss": 2.1708, "step": 18 }, { "epoch": 0.0029868930417182495, "grad_norm": 0.2640205919742584, "learning_rate": 4.999995113905947e-05, "loss": 2.1709, "step": 19 }, { "epoch": 0.003144097938650789, "grad_norm": 0.26644033193588257, "learning_rate": 4.999993816037749e-05, "loss": 2.1733, "step": 20 }, { "epoch": 0.0033013028355833285, "grad_norm": 0.2621535062789917, "learning_rate": 4.9999923654794414e-05, "loss": 2.0059, "step": 21 }, { "epoch": 0.003458507732515868, "grad_norm": 0.2586187422275543, "learning_rate": 4.999990762231111e-05, "loss": 2.0336, "step": 22 }, { "epoch": 0.003615712629448407, "grad_norm": 0.26732271909713745, "learning_rate": 4.9999890062928566e-05, "loss": 2.0566, "step": 23 }, { "epoch": 0.003772917526380947, "grad_norm": 0.2357867807149887, "learning_rate": 4.999987097664787e-05, "loss": 1.9529, "step": 24 }, { "epoch": 0.003930122423313486, "grad_norm": 0.2297009825706482, "learning_rate": 4.999985036347016e-05, "loss": 2.0369, "step": 25 }, { "epoch": 0.004087327320246026, "grad_norm": 0.20529747009277344, "learning_rate": 4.9999828223396705e-05, "loss": 1.9781, "step": 26 }, { "epoch": 0.004244532217178565, "grad_norm": 0.18342873454093933, "learning_rate": 4.999980455642887e-05, "loss": 1.9986, "step": 27 }, { "epoch": 0.004401737114111104, "grad_norm": 0.16487397253513336, "learning_rate": 4.999977936256809e-05, "loss": 1.9063, "step": 28 }, { "epoch": 0.004558942011043644, "grad_norm": 0.1762266606092453, "learning_rate": 4.99997526418159e-05, "loss": 1.9517, "step": 29 }, { "epoch": 0.004716146907976184, "grad_norm": 0.16371938586235046, "learning_rate": 4.999972439417394e-05, "loss": 1.7734, "step": 30 }, { "epoch": 0.004873351804908723, "grad_norm": 0.17309769988059998, "learning_rate": 4.999969461964392e-05, "loss": 1.8732, "step": 31 }, { "epoch": 0.0050305567018412625, "grad_norm": 0.15772338211536407, "learning_rate": 4.9999663318227683e-05, "loss": 1.7537, "step": 32 }, { "epoch": 0.005187761598773802, "grad_norm": 0.17521986365318298, "learning_rate": 4.9999630489927126e-05, "loss": 2.0077, "step": 33 }, { "epoch": 0.005344966495706341, "grad_norm": 0.15462292730808258, "learning_rate": 4.999959613474425e-05, "loss": 1.8576, "step": 34 }, { "epoch": 0.005502171392638881, "grad_norm": 0.15280336141586304, "learning_rate": 4.999956025268117e-05, "loss": 1.862, "step": 35 }, { "epoch": 0.0056593762895714205, "grad_norm": 0.14518432319164276, "learning_rate": 4.999952284374006e-05, "loss": 1.8893, "step": 36 }, { "epoch": 0.005816581186503959, "grad_norm": 0.16087624430656433, "learning_rate": 4.999948390792321e-05, "loss": 1.8658, "step": 37 }, { "epoch": 0.005973786083436499, "grad_norm": 0.17504698038101196, "learning_rate": 4.999944344523301e-05, "loss": 1.7647, "step": 38 }, { "epoch": 0.006130990980369039, "grad_norm": 0.17786233127117157, "learning_rate": 4.999940145567191e-05, "loss": 1.8133, "step": 39 }, { "epoch": 0.006288195877301578, "grad_norm": 0.1628972887992859, "learning_rate": 4.999935793924249e-05, "loss": 1.7731, "step": 40 }, { "epoch": 0.006445400774234117, "grad_norm": 0.13461466133594513, "learning_rate": 4.9999312895947406e-05, "loss": 1.7558, "step": 41 }, { "epoch": 0.006602605671166657, "grad_norm": 0.12960125505924225, "learning_rate": 4.99992663257894e-05, "loss": 1.7639, "step": 42 }, { "epoch": 0.006759810568099196, "grad_norm": 0.10991287231445312, "learning_rate": 4.9999218228771324e-05, "loss": 1.7538, "step": 43 }, { "epoch": 0.006917015465031736, "grad_norm": 0.11583230644464493, "learning_rate": 4.999916860489612e-05, "loss": 1.715, "step": 44 }, { "epoch": 0.007074220361964275, "grad_norm": 0.10344280302524567, "learning_rate": 4.999911745416681e-05, "loss": 1.6907, "step": 45 }, { "epoch": 0.007231425258896814, "grad_norm": 0.10546118766069412, "learning_rate": 4.999906477658651e-05, "loss": 1.7294, "step": 46 }, { "epoch": 0.007388630155829354, "grad_norm": 0.11775675415992737, "learning_rate": 4.9999010572158465e-05, "loss": 1.7146, "step": 47 }, { "epoch": 0.007545835052761894, "grad_norm": 0.11109112203121185, "learning_rate": 4.999895484088596e-05, "loss": 1.6939, "step": 48 }, { "epoch": 0.007703039949694433, "grad_norm": 0.1116517186164856, "learning_rate": 4.999889758277242e-05, "loss": 1.7271, "step": 49 }, { "epoch": 0.007860244846626972, "grad_norm": 0.11245547980070114, "learning_rate": 4.999883879782132e-05, "loss": 1.7333, "step": 50 }, { "epoch": 0.008017449743559512, "grad_norm": 0.1150551363825798, "learning_rate": 4.999877848603626e-05, "loss": 1.7036, "step": 51 }, { "epoch": 0.008174654640492052, "grad_norm": 0.10856381803750992, "learning_rate": 4.999871664742093e-05, "loss": 1.7493, "step": 52 }, { "epoch": 0.008331859537424591, "grad_norm": 0.10760089010000229, "learning_rate": 4.9998653281979095e-05, "loss": 1.6292, "step": 53 }, { "epoch": 0.00848906443435713, "grad_norm": 0.0932115837931633, "learning_rate": 4.9998588389714634e-05, "loss": 1.6608, "step": 54 }, { "epoch": 0.00864626933128967, "grad_norm": 0.09837482124567032, "learning_rate": 4.9998521970631504e-05, "loss": 1.7834, "step": 55 }, { "epoch": 0.008803474228222209, "grad_norm": 0.08872833847999573, "learning_rate": 4.9998454024733775e-05, "loss": 1.6484, "step": 56 }, { "epoch": 0.008960679125154749, "grad_norm": 0.08829163759946823, "learning_rate": 4.9998384552025577e-05, "loss": 1.5913, "step": 57 }, { "epoch": 0.009117884022087288, "grad_norm": 0.09087682515382767, "learning_rate": 4.999831355251117e-05, "loss": 1.6809, "step": 58 }, { "epoch": 0.009275088919019828, "grad_norm": 0.08675853163003922, "learning_rate": 4.9998241026194884e-05, "loss": 1.6519, "step": 59 }, { "epoch": 0.009432293815952368, "grad_norm": 0.08463481813669205, "learning_rate": 4.999816697308114e-05, "loss": 1.6234, "step": 60 }, { "epoch": 0.009589498712884906, "grad_norm": 0.08403950184583664, "learning_rate": 4.999809139317448e-05, "loss": 1.6533, "step": 61 }, { "epoch": 0.009746703609817445, "grad_norm": 0.08155622333288193, "learning_rate": 4.99980142864795e-05, "loss": 1.6726, "step": 62 }, { "epoch": 0.009903908506749985, "grad_norm": 0.08056480437517166, "learning_rate": 4.999793565300093e-05, "loss": 1.5881, "step": 63 }, { "epoch": 0.010061113403682525, "grad_norm": 0.07879023998975754, "learning_rate": 4.999785549274355e-05, "loss": 1.5568, "step": 64 }, { "epoch": 0.010218318300615065, "grad_norm": 0.07828455418348312, "learning_rate": 4.9997773805712265e-05, "loss": 1.6464, "step": 65 }, { "epoch": 0.010375523197547604, "grad_norm": 0.08054805546998978, "learning_rate": 4.9997690591912075e-05, "loss": 1.6213, "step": 66 }, { "epoch": 0.010532728094480142, "grad_norm": 0.07610727101564407, "learning_rate": 4.999760585134805e-05, "loss": 1.5729, "step": 67 }, { "epoch": 0.010689932991412682, "grad_norm": 0.07693428546190262, "learning_rate": 4.999751958402537e-05, "loss": 1.5444, "step": 68 }, { "epoch": 0.010847137888345222, "grad_norm": 0.0810319185256958, "learning_rate": 4.99974317899493e-05, "loss": 1.7045, "step": 69 }, { "epoch": 0.011004342785277762, "grad_norm": 0.07729896157979965, "learning_rate": 4.9997342469125205e-05, "loss": 1.6268, "step": 70 }, { "epoch": 0.011161547682210301, "grad_norm": 0.07730107754468918, "learning_rate": 4.999725162155855e-05, "loss": 1.658, "step": 71 }, { "epoch": 0.011318752579142841, "grad_norm": 0.08072328567504883, "learning_rate": 4.9997159247254864e-05, "loss": 1.5045, "step": 72 }, { "epoch": 0.011475957476075379, "grad_norm": 0.08120577782392502, "learning_rate": 4.9997065346219805e-05, "loss": 1.568, "step": 73 }, { "epoch": 0.011633162373007919, "grad_norm": 0.08131498098373413, "learning_rate": 4.99969699184591e-05, "loss": 1.6035, "step": 74 }, { "epoch": 0.011790367269940458, "grad_norm": 0.08395873010158539, "learning_rate": 4.9996872963978584e-05, "loss": 1.5844, "step": 75 }, { "epoch": 0.011947572166872998, "grad_norm": 0.08502068370580673, "learning_rate": 4.999677448278417e-05, "loss": 1.6661, "step": 76 }, { "epoch": 0.012104777063805538, "grad_norm": 0.08467952907085419, "learning_rate": 4.999667447488188e-05, "loss": 1.5537, "step": 77 }, { "epoch": 0.012261981960738078, "grad_norm": 0.19682182371616364, "learning_rate": 4.999657294027782e-05, "loss": 1.5051, "step": 78 }, { "epoch": 0.012419186857670617, "grad_norm": 0.08586428314447403, "learning_rate": 4.999646987897818e-05, "loss": 1.565, "step": 79 }, { "epoch": 0.012576391754603155, "grad_norm": 0.08156823366880417, "learning_rate": 4.999636529098928e-05, "loss": 1.6627, "step": 80 }, { "epoch": 0.012733596651535695, "grad_norm": 0.08715341240167618, "learning_rate": 4.9996259176317486e-05, "loss": 1.5862, "step": 81 }, { "epoch": 0.012890801548468235, "grad_norm": 0.09664586186408997, "learning_rate": 4.999615153496928e-05, "loss": 1.5741, "step": 82 }, { "epoch": 0.013048006445400774, "grad_norm": 0.08438891172409058, "learning_rate": 4.999604236695125e-05, "loss": 1.5933, "step": 83 }, { "epoch": 0.013205211342333314, "grad_norm": 0.08333732932806015, "learning_rate": 4.999593167227006e-05, "loss": 1.5904, "step": 84 }, { "epoch": 0.013362416239265854, "grad_norm": 0.07945791631937027, "learning_rate": 4.9995819450932455e-05, "loss": 1.5763, "step": 85 }, { "epoch": 0.013519621136198392, "grad_norm": 0.07682961225509644, "learning_rate": 4.9995705702945304e-05, "loss": 1.5197, "step": 86 }, { "epoch": 0.013676826033130932, "grad_norm": 0.07547677308320999, "learning_rate": 4.999559042831555e-05, "loss": 1.6825, "step": 87 }, { "epoch": 0.013834030930063471, "grad_norm": 0.07293456047773361, "learning_rate": 4.999547362705025e-05, "loss": 1.5466, "step": 88 }, { "epoch": 0.013991235826996011, "grad_norm": 0.07730914652347565, "learning_rate": 4.999535529915651e-05, "loss": 1.5775, "step": 89 }, { "epoch": 0.01414844072392855, "grad_norm": 0.07689664512872696, "learning_rate": 4.9995235444641565e-05, "loss": 1.5881, "step": 90 }, { "epoch": 0.01430564562086109, "grad_norm": 0.07754997909069061, "learning_rate": 4.999511406351275e-05, "loss": 1.5037, "step": 91 }, { "epoch": 0.014462850517793628, "grad_norm": 0.07229866087436676, "learning_rate": 4.999499115577746e-05, "loss": 1.5077, "step": 92 }, { "epoch": 0.014620055414726168, "grad_norm": 0.07491567730903625, "learning_rate": 4.9994866721443215e-05, "loss": 1.5461, "step": 93 }, { "epoch": 0.014777260311658708, "grad_norm": 0.07258685678243637, "learning_rate": 4.9994740760517605e-05, "loss": 1.5516, "step": 94 }, { "epoch": 0.014934465208591248, "grad_norm": 0.07643327116966248, "learning_rate": 4.9994613273008334e-05, "loss": 1.6223, "step": 95 }, { "epoch": 0.015091670105523787, "grad_norm": 0.0740588903427124, "learning_rate": 4.999448425892318e-05, "loss": 1.5322, "step": 96 }, { "epoch": 0.015248875002456327, "grad_norm": 0.44172239303588867, "learning_rate": 4.999435371827003e-05, "loss": 1.5498, "step": 97 }, { "epoch": 0.015406079899388867, "grad_norm": 0.0756363570690155, "learning_rate": 4.999422165105684e-05, "loss": 1.559, "step": 98 }, { "epoch": 0.015563284796321405, "grad_norm": 0.07251248508691788, "learning_rate": 4.99940880572917e-05, "loss": 1.5903, "step": 99 }, { "epoch": 0.015720489693253945, "grad_norm": 0.06931837648153305, "learning_rate": 4.999395293698275e-05, "loss": 1.4849, "step": 100 }, { "epoch": 0.015877694590186484, "grad_norm": 0.07403590530157089, "learning_rate": 4.9993816290138254e-05, "loss": 1.5191, "step": 101 }, { "epoch": 0.016034899487119024, "grad_norm": 0.07027724385261536, "learning_rate": 4.999367811676655e-05, "loss": 1.5655, "step": 102 }, { "epoch": 0.016192104384051564, "grad_norm": 0.07320379465818405, "learning_rate": 4.9993538416876093e-05, "loss": 1.4869, "step": 103 }, { "epoch": 0.016349309280984103, "grad_norm": 0.0726180374622345, "learning_rate": 4.9993397190475396e-05, "loss": 1.4629, "step": 104 }, { "epoch": 0.016506514177916643, "grad_norm": 0.07542011886835098, "learning_rate": 4.999325443757309e-05, "loss": 1.5976, "step": 105 }, { "epoch": 0.016663719074849183, "grad_norm": 0.07440067082643509, "learning_rate": 4.9993110158177895e-05, "loss": 1.5469, "step": 106 }, { "epoch": 0.016820923971781723, "grad_norm": 0.07547372579574585, "learning_rate": 4.999296435229863e-05, "loss": 1.5328, "step": 107 }, { "epoch": 0.01697812886871426, "grad_norm": 0.07532137632369995, "learning_rate": 4.999281701994419e-05, "loss": 1.6742, "step": 108 }, { "epoch": 0.0171353337656468, "grad_norm": 0.07249438762664795, "learning_rate": 4.999266816112358e-05, "loss": 1.4799, "step": 109 }, { "epoch": 0.01729253866257934, "grad_norm": 0.07399806380271912, "learning_rate": 4.999251777584589e-05, "loss": 1.5438, "step": 110 }, { "epoch": 0.017449743559511878, "grad_norm": 0.08135057240724564, "learning_rate": 4.99923658641203e-05, "loss": 1.5608, "step": 111 }, { "epoch": 0.017606948456444418, "grad_norm": 0.07508935779333115, "learning_rate": 4.99922124259561e-05, "loss": 1.5894, "step": 112 }, { "epoch": 0.017764153353376957, "grad_norm": 0.07432372123003006, "learning_rate": 4.999205746136265e-05, "loss": 1.4818, "step": 113 }, { "epoch": 0.017921358250309497, "grad_norm": 0.07694194465875626, "learning_rate": 4.999190097034942e-05, "loss": 1.5629, "step": 114 }, { "epoch": 0.018078563147242037, "grad_norm": 0.07384433597326279, "learning_rate": 4.999174295292597e-05, "loss": 1.4829, "step": 115 }, { "epoch": 0.018235768044174577, "grad_norm": 0.07152919471263885, "learning_rate": 4.999158340910195e-05, "loss": 1.4748, "step": 116 }, { "epoch": 0.018392972941107116, "grad_norm": 0.07719701528549194, "learning_rate": 4.999142233888709e-05, "loss": 1.5524, "step": 117 }, { "epoch": 0.018550177838039656, "grad_norm": 0.07540587335824966, "learning_rate": 4.999125974229125e-05, "loss": 1.4661, "step": 118 }, { "epoch": 0.018707382734972196, "grad_norm": 0.0787581130862236, "learning_rate": 4.9991095619324344e-05, "loss": 1.6455, "step": 119 }, { "epoch": 0.018864587631904736, "grad_norm": 0.07454577833414078, "learning_rate": 4.999092996999641e-05, "loss": 1.5083, "step": 120 }, { "epoch": 0.019021792528837272, "grad_norm": 0.0751076266169548, "learning_rate": 4.9990762794317545e-05, "loss": 1.4874, "step": 121 }, { "epoch": 0.01917899742576981, "grad_norm": 0.07733119279146194, "learning_rate": 4.999059409229798e-05, "loss": 1.6308, "step": 122 }, { "epoch": 0.01933620232270235, "grad_norm": 0.07897089421749115, "learning_rate": 4.999042386394802e-05, "loss": 1.5906, "step": 123 }, { "epoch": 0.01949340721963489, "grad_norm": 0.07758141309022903, "learning_rate": 4.999025210927804e-05, "loss": 1.5604, "step": 124 }, { "epoch": 0.01965061211656743, "grad_norm": 0.07845707982778549, "learning_rate": 4.9990078828298544e-05, "loss": 1.5901, "step": 125 }, { "epoch": 0.01980781701349997, "grad_norm": 0.0772818773984909, "learning_rate": 4.998990402102012e-05, "loss": 1.4516, "step": 126 }, { "epoch": 0.01996502191043251, "grad_norm": 0.07795504480600357, "learning_rate": 4.998972768745344e-05, "loss": 1.4642, "step": 127 }, { "epoch": 0.02012222680736505, "grad_norm": 0.0784008800983429, "learning_rate": 4.998954982760926e-05, "loss": 1.5936, "step": 128 }, { "epoch": 0.02027943170429759, "grad_norm": 0.07791212201118469, "learning_rate": 4.9989370441498465e-05, "loss": 1.4705, "step": 129 }, { "epoch": 0.02043663660123013, "grad_norm": 0.07785367220640182, "learning_rate": 4.9989189529132004e-05, "loss": 1.5085, "step": 130 }, { "epoch": 0.02059384149816267, "grad_norm": 0.07916689664125443, "learning_rate": 4.9989007090520925e-05, "loss": 1.5365, "step": 131 }, { "epoch": 0.02075104639509521, "grad_norm": 0.0775083601474762, "learning_rate": 4.9988823125676367e-05, "loss": 1.5286, "step": 132 }, { "epoch": 0.020908251292027745, "grad_norm": 0.08110442757606506, "learning_rate": 4.998863763460956e-05, "loss": 1.5779, "step": 133 }, { "epoch": 0.021065456188960285, "grad_norm": 0.0814640000462532, "learning_rate": 4.998845061733185e-05, "loss": 1.4778, "step": 134 }, { "epoch": 0.021222661085892824, "grad_norm": 0.08069492131471634, "learning_rate": 4.998826207385465e-05, "loss": 1.5317, "step": 135 }, { "epoch": 0.021379865982825364, "grad_norm": 0.07377774268388748, "learning_rate": 4.998807200418948e-05, "loss": 1.5258, "step": 136 }, { "epoch": 0.021537070879757904, "grad_norm": 0.0787922590970993, "learning_rate": 4.9987880408347945e-05, "loss": 1.5185, "step": 137 }, { "epoch": 0.021694275776690444, "grad_norm": 0.07662995159626007, "learning_rate": 4.9987687286341745e-05, "loss": 1.4637, "step": 138 }, { "epoch": 0.021851480673622983, "grad_norm": 0.08528955280780792, "learning_rate": 4.9987492638182676e-05, "loss": 1.4776, "step": 139 }, { "epoch": 0.022008685570555523, "grad_norm": 0.08089053630828857, "learning_rate": 4.9987296463882626e-05, "loss": 1.5885, "step": 140 }, { "epoch": 0.022165890467488063, "grad_norm": 0.08029694855213165, "learning_rate": 4.998709876345358e-05, "loss": 1.4557, "step": 141 }, { "epoch": 0.022323095364420602, "grad_norm": 0.07918502390384674, "learning_rate": 4.9986899536907614e-05, "loss": 1.4285, "step": 142 }, { "epoch": 0.022480300261353142, "grad_norm": 0.0813126415014267, "learning_rate": 4.998669878425689e-05, "loss": 1.5958, "step": 143 }, { "epoch": 0.022637505158285682, "grad_norm": 0.07935188710689545, "learning_rate": 4.998649650551368e-05, "loss": 1.5249, "step": 144 }, { "epoch": 0.02279471005521822, "grad_norm": 0.08163304626941681, "learning_rate": 4.9986292700690324e-05, "loss": 1.483, "step": 145 }, { "epoch": 0.022951914952150758, "grad_norm": 0.08277447521686554, "learning_rate": 4.998608736979928e-05, "loss": 1.6212, "step": 146 }, { "epoch": 0.023109119849083298, "grad_norm": 0.08285827934741974, "learning_rate": 4.9985880512853076e-05, "loss": 1.4495, "step": 147 }, { "epoch": 0.023266324746015837, "grad_norm": 0.082750603556633, "learning_rate": 4.998567212986437e-05, "loss": 1.4335, "step": 148 }, { "epoch": 0.023423529642948377, "grad_norm": 0.07986058294773102, "learning_rate": 4.998546222084587e-05, "loss": 1.4704, "step": 149 }, { "epoch": 0.023580734539880917, "grad_norm": 0.08105576783418655, "learning_rate": 4.9985250785810396e-05, "loss": 1.5183, "step": 150 }, { "epoch": 0.023737939436813457, "grad_norm": 0.08202917128801346, "learning_rate": 4.9985037824770866e-05, "loss": 1.5423, "step": 151 }, { "epoch": 0.023895144333745996, "grad_norm": 0.08937894552946091, "learning_rate": 4.998482333774029e-05, "loss": 1.5731, "step": 152 }, { "epoch": 0.024052349230678536, "grad_norm": 0.08333728462457657, "learning_rate": 4.9984607324731766e-05, "loss": 1.5133, "step": 153 }, { "epoch": 0.024209554127611076, "grad_norm": 0.08529175072908401, "learning_rate": 4.998438978575849e-05, "loss": 1.516, "step": 154 }, { "epoch": 0.024366759024543615, "grad_norm": 0.08508963882923126, "learning_rate": 4.998417072083374e-05, "loss": 1.5646, "step": 155 }, { "epoch": 0.024523963921476155, "grad_norm": 0.08971578627824783, "learning_rate": 4.99839501299709e-05, "loss": 1.4714, "step": 156 }, { "epoch": 0.024681168818408695, "grad_norm": 0.08380109816789627, "learning_rate": 4.998372801318345e-05, "loss": 1.4476, "step": 157 }, { "epoch": 0.024838373715341235, "grad_norm": 0.08533143252134323, "learning_rate": 4.9983504370484945e-05, "loss": 1.4866, "step": 158 }, { "epoch": 0.02499557861227377, "grad_norm": 0.08318709582090378, "learning_rate": 4.998327920188905e-05, "loss": 1.5274, "step": 159 }, { "epoch": 0.02515278350920631, "grad_norm": 0.08486370742321014, "learning_rate": 4.9983052507409525e-05, "loss": 1.4713, "step": 160 }, { "epoch": 0.02515278350920631, "eval_loss": 1.5136528015136719, "eval_runtime": 2318.8971, "eval_samples_per_second": 3.992, "eval_steps_per_second": 1.996, "step": 160 }, { "epoch": 0.02530998840613885, "grad_norm": 0.08242359757423401, "learning_rate": 4.9982824287060195e-05, "loss": 1.5069, "step": 161 }, { "epoch": 0.02546719330307139, "grad_norm": 0.08547423779964447, "learning_rate": 4.9982594540855014e-05, "loss": 1.4973, "step": 162 }, { "epoch": 0.02562439820000393, "grad_norm": 0.08345580101013184, "learning_rate": 4.9982363268808016e-05, "loss": 1.5078, "step": 163 }, { "epoch": 0.02578160309693647, "grad_norm": 0.0830339640378952, "learning_rate": 4.9982130470933316e-05, "loss": 1.4098, "step": 164 }, { "epoch": 0.02593880799386901, "grad_norm": 0.08568515628576279, "learning_rate": 4.998189614724514e-05, "loss": 1.4628, "step": 165 }, { "epoch": 0.02609601289080155, "grad_norm": 0.08261829614639282, "learning_rate": 4.998166029775779e-05, "loss": 1.4492, "step": 166 }, { "epoch": 0.02625321778773409, "grad_norm": 0.08944887667894363, "learning_rate": 4.998142292248569e-05, "loss": 1.5633, "step": 167 }, { "epoch": 0.02641042268466663, "grad_norm": 0.08632911741733551, "learning_rate": 4.998118402144332e-05, "loss": 1.5106, "step": 168 }, { "epoch": 0.026567627581599168, "grad_norm": 0.08733859658241272, "learning_rate": 4.998094359464528e-05, "loss": 1.5607, "step": 169 }, { "epoch": 0.026724832478531708, "grad_norm": 0.08667927235364914, "learning_rate": 4.9980701642106245e-05, "loss": 1.4544, "step": 170 }, { "epoch": 0.026882037375464244, "grad_norm": 0.08655022084712982, "learning_rate": 4.9980458163841006e-05, "loss": 1.5264, "step": 171 }, { "epoch": 0.027039242272396784, "grad_norm": 0.08899988234043121, "learning_rate": 4.9980213159864426e-05, "loss": 1.4778, "step": 172 }, { "epoch": 0.027196447169329323, "grad_norm": 0.09411856532096863, "learning_rate": 4.997996663019147e-05, "loss": 1.5269, "step": 173 }, { "epoch": 0.027353652066261863, "grad_norm": 0.087191641330719, "learning_rate": 4.997971857483719e-05, "loss": 1.5166, "step": 174 }, { "epoch": 0.027510856963194403, "grad_norm": 0.08959636092185974, "learning_rate": 4.997946899381675e-05, "loss": 1.5503, "step": 175 }, { "epoch": 0.027668061860126943, "grad_norm": 0.0951187014579773, "learning_rate": 4.997921788714537e-05, "loss": 1.4879, "step": 176 }, { "epoch": 0.027825266757059482, "grad_norm": 0.09324768930673599, "learning_rate": 4.997896525483841e-05, "loss": 1.5714, "step": 177 }, { "epoch": 0.027982471653992022, "grad_norm": 0.08633986115455627, "learning_rate": 4.997871109691129e-05, "loss": 1.4198, "step": 178 }, { "epoch": 0.028139676550924562, "grad_norm": 0.08947525173425674, "learning_rate": 4.9978455413379535e-05, "loss": 1.4702, "step": 179 }, { "epoch": 0.0282968814478571, "grad_norm": 0.09275490790605545, "learning_rate": 4.9978198204258766e-05, "loss": 1.5252, "step": 180 }, { "epoch": 0.02845408634478964, "grad_norm": 0.08761609345674515, "learning_rate": 4.9977939469564676e-05, "loss": 1.505, "step": 181 }, { "epoch": 0.02861129124172218, "grad_norm": 0.08683087676763535, "learning_rate": 4.997767920931308e-05, "loss": 1.5059, "step": 182 }, { "epoch": 0.02876849613865472, "grad_norm": 0.08931361883878708, "learning_rate": 4.997741742351988e-05, "loss": 1.5003, "step": 183 }, { "epoch": 0.028925701035587257, "grad_norm": 0.08820109069347382, "learning_rate": 4.997715411220105e-05, "loss": 1.5132, "step": 184 }, { "epoch": 0.029082905932519797, "grad_norm": 0.09284964948892593, "learning_rate": 4.997688927537268e-05, "loss": 1.4561, "step": 185 }, { "epoch": 0.029240110829452336, "grad_norm": 0.09472864121198654, "learning_rate": 4.997662291305094e-05, "loss": 1.4729, "step": 186 }, { "epoch": 0.029397315726384876, "grad_norm": 0.08725330233573914, "learning_rate": 4.997635502525211e-05, "loss": 1.3994, "step": 187 }, { "epoch": 0.029554520623317416, "grad_norm": 0.09085626900196075, "learning_rate": 4.9976085611992536e-05, "loss": 1.4695, "step": 188 }, { "epoch": 0.029711725520249956, "grad_norm": 0.09322400391101837, "learning_rate": 4.9975814673288684e-05, "loss": 1.4753, "step": 189 }, { "epoch": 0.029868930417182495, "grad_norm": 0.08927160501480103, "learning_rate": 4.99755422091571e-05, "loss": 1.4465, "step": 190 }, { "epoch": 0.030026135314115035, "grad_norm": 0.09317070990800858, "learning_rate": 4.997526821961442e-05, "loss": 1.5124, "step": 191 }, { "epoch": 0.030183340211047575, "grad_norm": 0.08911167085170746, "learning_rate": 4.9974992704677385e-05, "loss": 1.4515, "step": 192 }, { "epoch": 0.030340545107980114, "grad_norm": 0.09432853013277054, "learning_rate": 4.997471566436282e-05, "loss": 1.4623, "step": 193 }, { "epoch": 0.030497750004912654, "grad_norm": 0.09417332708835602, "learning_rate": 4.997443709868764e-05, "loss": 1.5103, "step": 194 }, { "epoch": 0.030654954901845194, "grad_norm": 0.09564542025327682, "learning_rate": 4.997415700766887e-05, "loss": 1.4929, "step": 195 }, { "epoch": 0.030812159798777734, "grad_norm": 0.09101004898548126, "learning_rate": 4.997387539132361e-05, "loss": 1.4225, "step": 196 }, { "epoch": 0.03096936469571027, "grad_norm": 0.09196274727582932, "learning_rate": 4.997359224966906e-05, "loss": 1.4701, "step": 197 }, { "epoch": 0.03112656959264281, "grad_norm": 0.09573279321193695, "learning_rate": 4.997330758272251e-05, "loss": 1.4425, "step": 198 }, { "epoch": 0.03128377448957535, "grad_norm": 0.09180758893489838, "learning_rate": 4.9973021390501354e-05, "loss": 1.4426, "step": 199 }, { "epoch": 0.03144097938650789, "grad_norm": 0.09583238512277603, "learning_rate": 4.997273367302306e-05, "loss": 1.5158, "step": 200 }, { "epoch": 0.03159818428344043, "grad_norm": 0.09394747018814087, "learning_rate": 4.997244443030521e-05, "loss": 1.4306, "step": 201 }, { "epoch": 0.03175538918037297, "grad_norm": 0.09470199793577194, "learning_rate": 4.9972153662365474e-05, "loss": 1.5286, "step": 202 }, { "epoch": 0.031912594077305505, "grad_norm": 0.09274959564208984, "learning_rate": 4.997186136922161e-05, "loss": 1.4803, "step": 203 }, { "epoch": 0.03206979897423805, "grad_norm": 0.09344369918107986, "learning_rate": 4.997156755089145e-05, "loss": 1.5449, "step": 204 }, { "epoch": 0.032227003871170584, "grad_norm": 0.09794919937849045, "learning_rate": 4.997127220739296e-05, "loss": 1.4383, "step": 205 }, { "epoch": 0.03238420876810313, "grad_norm": 0.09698093682527542, "learning_rate": 4.997097533874418e-05, "loss": 1.4462, "step": 206 }, { "epoch": 0.032541413665035664, "grad_norm": 0.09690559655427933, "learning_rate": 4.997067694496323e-05, "loss": 1.4735, "step": 207 }, { "epoch": 0.03269861856196821, "grad_norm": 0.09657544642686844, "learning_rate": 4.9970377026068336e-05, "loss": 1.5672, "step": 208 }, { "epoch": 0.03285582345890074, "grad_norm": 0.09483659267425537, "learning_rate": 4.9970075582077825e-05, "loss": 1.4931, "step": 209 }, { "epoch": 0.033013028355833286, "grad_norm": 0.09744243323802948, "learning_rate": 4.9969772613010104e-05, "loss": 1.4638, "step": 210 }, { "epoch": 0.03317023325276582, "grad_norm": 0.09521006047725677, "learning_rate": 4.9969468118883665e-05, "loss": 1.4127, "step": 211 }, { "epoch": 0.033327438149698366, "grad_norm": 0.09646004438400269, "learning_rate": 4.996916209971713e-05, "loss": 1.5139, "step": 212 }, { "epoch": 0.0334846430466309, "grad_norm": 0.09292810410261154, "learning_rate": 4.996885455552916e-05, "loss": 1.4399, "step": 213 }, { "epoch": 0.033641847943563445, "grad_norm": 0.09986516088247299, "learning_rate": 4.996854548633857e-05, "loss": 1.4637, "step": 214 }, { "epoch": 0.03379905284049598, "grad_norm": 0.09723702073097229, "learning_rate": 4.996823489216421e-05, "loss": 1.5673, "step": 215 }, { "epoch": 0.03395625773742852, "grad_norm": 0.09608977288007736, "learning_rate": 4.996792277302507e-05, "loss": 1.4428, "step": 216 }, { "epoch": 0.03411346263436106, "grad_norm": 0.09329380095005035, "learning_rate": 4.99676091289402e-05, "loss": 1.3892, "step": 217 }, { "epoch": 0.0342706675312936, "grad_norm": 0.0959913358092308, "learning_rate": 4.996729395992875e-05, "loss": 1.5219, "step": 218 }, { "epoch": 0.03442787242822614, "grad_norm": 0.09832671284675598, "learning_rate": 4.996697726600999e-05, "loss": 1.5259, "step": 219 }, { "epoch": 0.03458507732515868, "grad_norm": 0.10061636567115784, "learning_rate": 4.996665904720325e-05, "loss": 1.5216, "step": 220 }, { "epoch": 0.03474228222209122, "grad_norm": 0.09742400050163269, "learning_rate": 4.9966339303527965e-05, "loss": 1.3819, "step": 221 }, { "epoch": 0.034899487119023756, "grad_norm": 0.09629969298839569, "learning_rate": 4.996601803500367e-05, "loss": 1.5341, "step": 222 }, { "epoch": 0.0350566920159563, "grad_norm": 0.09776200354099274, "learning_rate": 4.996569524164998e-05, "loss": 1.5054, "step": 223 }, { "epoch": 0.035213896912888835, "grad_norm": 0.1008530780673027, "learning_rate": 4.996537092348661e-05, "loss": 1.5333, "step": 224 }, { "epoch": 0.03537110180982138, "grad_norm": 0.09749735891819, "learning_rate": 4.996504508053338e-05, "loss": 1.3899, "step": 225 }, { "epoch": 0.035528306706753915, "grad_norm": 0.10522401332855225, "learning_rate": 4.9964717712810175e-05, "loss": 1.5413, "step": 226 }, { "epoch": 0.03568551160368646, "grad_norm": 0.09566272795200348, "learning_rate": 4.9964388820336996e-05, "loss": 1.435, "step": 227 }, { "epoch": 0.035842716500618994, "grad_norm": 0.10133984684944153, "learning_rate": 4.996405840313393e-05, "loss": 1.445, "step": 228 }, { "epoch": 0.03599992139755153, "grad_norm": 0.09702739119529724, "learning_rate": 4.996372646122116e-05, "loss": 1.4287, "step": 229 }, { "epoch": 0.036157126294484074, "grad_norm": 0.1012992411851883, "learning_rate": 4.996339299461896e-05, "loss": 1.382, "step": 230 }, { "epoch": 0.03631433119141661, "grad_norm": 0.09877166152000427, "learning_rate": 4.99630580033477e-05, "loss": 1.5729, "step": 231 }, { "epoch": 0.03647153608834915, "grad_norm": 0.1033129170536995, "learning_rate": 4.996272148742783e-05, "loss": 1.4754, "step": 232 }, { "epoch": 0.03662874098528169, "grad_norm": 0.09901215881109238, "learning_rate": 4.9962383446879914e-05, "loss": 1.5153, "step": 233 }, { "epoch": 0.03678594588221423, "grad_norm": 0.10241983830928802, "learning_rate": 4.996204388172458e-05, "loss": 1.5131, "step": 234 }, { "epoch": 0.03694315077914677, "grad_norm": 0.09574593603610992, "learning_rate": 4.9961702791982594e-05, "loss": 1.5285, "step": 235 }, { "epoch": 0.03710035567607931, "grad_norm": 0.10309838503599167, "learning_rate": 4.996136017767477e-05, "loss": 1.5751, "step": 236 }, { "epoch": 0.03725756057301185, "grad_norm": 0.09928470849990845, "learning_rate": 4.996101603882204e-05, "loss": 1.5108, "step": 237 }, { "epoch": 0.03741476546994439, "grad_norm": 0.10514767467975616, "learning_rate": 4.996067037544542e-05, "loss": 1.4206, "step": 238 }, { "epoch": 0.03757197036687693, "grad_norm": 0.10411518812179565, "learning_rate": 4.996032318756601e-05, "loss": 1.5628, "step": 239 }, { "epoch": 0.03772917526380947, "grad_norm": 0.0989808738231659, "learning_rate": 4.9959974475205045e-05, "loss": 1.4444, "step": 240 }, { "epoch": 0.03788638016074201, "grad_norm": 0.10069911926984787, "learning_rate": 4.9959624238383804e-05, "loss": 1.4805, "step": 241 }, { "epoch": 0.038043585057674544, "grad_norm": 0.10637518763542175, "learning_rate": 4.995927247712367e-05, "loss": 1.5289, "step": 242 }, { "epoch": 0.03820078995460709, "grad_norm": 0.10085684061050415, "learning_rate": 4.995891919144614e-05, "loss": 1.5288, "step": 243 }, { "epoch": 0.03835799485153962, "grad_norm": 0.09989017248153687, "learning_rate": 4.995856438137279e-05, "loss": 1.5444, "step": 244 }, { "epoch": 0.038515199748472166, "grad_norm": 0.10382463037967682, "learning_rate": 4.9958208046925294e-05, "loss": 1.4621, "step": 245 }, { "epoch": 0.0386724046454047, "grad_norm": 0.10208063572645187, "learning_rate": 4.99578501881254e-05, "loss": 1.5003, "step": 246 }, { "epoch": 0.038829609542337246, "grad_norm": 0.1028011366724968, "learning_rate": 4.9957490804994977e-05, "loss": 1.516, "step": 247 }, { "epoch": 0.03898681443926978, "grad_norm": 0.10475701838731766, "learning_rate": 4.995712989755598e-05, "loss": 1.5333, "step": 248 }, { "epoch": 0.039144019336202325, "grad_norm": 0.1038154736161232, "learning_rate": 4.995676746583044e-05, "loss": 1.4779, "step": 249 }, { "epoch": 0.03930122423313486, "grad_norm": 0.10413440316915512, "learning_rate": 4.99564035098405e-05, "loss": 1.5241, "step": 250 }, { "epoch": 0.039458429130067404, "grad_norm": 0.09869382530450821, "learning_rate": 4.995603802960838e-05, "loss": 1.442, "step": 251 }, { "epoch": 0.03961563402699994, "grad_norm": 0.10138234496116638, "learning_rate": 4.995567102515641e-05, "loss": 1.5393, "step": 252 }, { "epoch": 0.039772838923932484, "grad_norm": 0.10225867480039597, "learning_rate": 4.995530249650701e-05, "loss": 1.4516, "step": 253 }, { "epoch": 0.03993004382086502, "grad_norm": 0.09942895174026489, "learning_rate": 4.995493244368268e-05, "loss": 1.4543, "step": 254 }, { "epoch": 0.040087248717797556, "grad_norm": 0.11218860745429993, "learning_rate": 4.995456086670602e-05, "loss": 1.4985, "step": 255 }, { "epoch": 0.0402444536147301, "grad_norm": 0.10839337855577469, "learning_rate": 4.9954187765599736e-05, "loss": 1.4805, "step": 256 }, { "epoch": 0.040401658511662636, "grad_norm": 0.10317599028348923, "learning_rate": 4.9953813140386595e-05, "loss": 1.4412, "step": 257 }, { "epoch": 0.04055886340859518, "grad_norm": 0.10285656154155731, "learning_rate": 4.99534369910895e-05, "loss": 1.476, "step": 258 }, { "epoch": 0.040716068305527715, "grad_norm": 0.10330680012702942, "learning_rate": 4.995305931773141e-05, "loss": 1.5157, "step": 259 }, { "epoch": 0.04087327320246026, "grad_norm": 0.1086694598197937, "learning_rate": 4.99526801203354e-05, "loss": 1.4999, "step": 260 }, { "epoch": 0.041030478099392795, "grad_norm": 0.10800144821405411, "learning_rate": 4.995229939892464e-05, "loss": 1.4764, "step": 261 }, { "epoch": 0.04118768299632534, "grad_norm": 0.10645303875207901, "learning_rate": 4.9951917153522355e-05, "loss": 1.4404, "step": 262 }, { "epoch": 0.041344887893257874, "grad_norm": 0.10440964996814728, "learning_rate": 4.9951533384151906e-05, "loss": 1.3678, "step": 263 }, { "epoch": 0.04150209279019042, "grad_norm": 0.10993078351020813, "learning_rate": 4.995114809083673e-05, "loss": 1.5064, "step": 264 }, { "epoch": 0.041659297687122954, "grad_norm": 0.10710245370864868, "learning_rate": 4.9950761273600366e-05, "loss": 1.4134, "step": 265 }, { "epoch": 0.04181650258405549, "grad_norm": 0.11030582338571548, "learning_rate": 4.995037293246644e-05, "loss": 1.5299, "step": 266 }, { "epoch": 0.04197370748098803, "grad_norm": 0.1058267131447792, "learning_rate": 4.994998306745866e-05, "loss": 1.3654, "step": 267 }, { "epoch": 0.04213091237792057, "grad_norm": 0.10541702806949615, "learning_rate": 4.994959167860084e-05, "loss": 1.4297, "step": 268 }, { "epoch": 0.04228811727485311, "grad_norm": 0.11085420846939087, "learning_rate": 4.994919876591689e-05, "loss": 1.4876, "step": 269 }, { "epoch": 0.04244532217178565, "grad_norm": 0.11054470390081406, "learning_rate": 4.994880432943081e-05, "loss": 1.574, "step": 270 }, { "epoch": 0.04260252706871819, "grad_norm": 0.11234510689973831, "learning_rate": 4.994840836916668e-05, "loss": 1.5079, "step": 271 }, { "epoch": 0.04275973196565073, "grad_norm": 0.11040106415748596, "learning_rate": 4.994801088514869e-05, "loss": 1.5091, "step": 272 }, { "epoch": 0.04291693686258327, "grad_norm": 0.10639887303113937, "learning_rate": 4.994761187740111e-05, "loss": 1.4495, "step": 273 }, { "epoch": 0.04307414175951581, "grad_norm": 0.11268071085214615, "learning_rate": 4.994721134594833e-05, "loss": 1.5057, "step": 274 }, { "epoch": 0.04323134665644835, "grad_norm": 0.10079260170459747, "learning_rate": 4.994680929081479e-05, "loss": 1.4145, "step": 275 }, { "epoch": 0.04338855155338089, "grad_norm": 0.11474710702896118, "learning_rate": 4.994640571202506e-05, "loss": 1.5061, "step": 276 }, { "epoch": 0.04354575645031343, "grad_norm": 0.10946876555681229, "learning_rate": 4.994600060960377e-05, "loss": 1.5306, "step": 277 }, { "epoch": 0.04370296134724597, "grad_norm": 0.11192137002944946, "learning_rate": 4.994559398357569e-05, "loss": 1.5347, "step": 278 }, { "epoch": 0.0438601662441785, "grad_norm": 0.10744784027338028, "learning_rate": 4.994518583396564e-05, "loss": 1.4686, "step": 279 }, { "epoch": 0.044017371141111046, "grad_norm": 0.11113352328538895, "learning_rate": 4.9944776160798544e-05, "loss": 1.4101, "step": 280 }, { "epoch": 0.04417457603804358, "grad_norm": 0.11456230282783508, "learning_rate": 4.994436496409943e-05, "loss": 1.4036, "step": 281 }, { "epoch": 0.044331780934976125, "grad_norm": 0.11608672887086868, "learning_rate": 4.994395224389342e-05, "loss": 1.4949, "step": 282 }, { "epoch": 0.04448898583190866, "grad_norm": 0.1232326403260231, "learning_rate": 4.9943538000205705e-05, "loss": 1.5501, "step": 283 }, { "epoch": 0.044646190728841205, "grad_norm": 0.11791515350341797, "learning_rate": 4.994312223306159e-05, "loss": 1.4542, "step": 284 }, { "epoch": 0.04480339562577374, "grad_norm": 0.11657550930976868, "learning_rate": 4.9942704942486476e-05, "loss": 1.4724, "step": 285 }, { "epoch": 0.044960600522706284, "grad_norm": 0.11560262739658356, "learning_rate": 4.994228612850584e-05, "loss": 1.4036, "step": 286 }, { "epoch": 0.04511780541963882, "grad_norm": 0.10999175906181335, "learning_rate": 4.994186579114527e-05, "loss": 1.4489, "step": 287 }, { "epoch": 0.045275010316571364, "grad_norm": 0.11586826294660568, "learning_rate": 4.9941443930430436e-05, "loss": 1.5486, "step": 288 }, { "epoch": 0.0454322152135039, "grad_norm": 0.11349951475858688, "learning_rate": 4.994102054638711e-05, "loss": 1.5698, "step": 289 }, { "epoch": 0.04558942011043644, "grad_norm": 0.11978698521852493, "learning_rate": 4.9940595639041134e-05, "loss": 1.3933, "step": 290 }, { "epoch": 0.04574662500736898, "grad_norm": 0.11438622325658798, "learning_rate": 4.994016920841846e-05, "loss": 1.5005, "step": 291 }, { "epoch": 0.045903829904301516, "grad_norm": 0.11395915597677231, "learning_rate": 4.9939741254545155e-05, "loss": 1.4521, "step": 292 }, { "epoch": 0.04606103480123406, "grad_norm": 0.11659599095582962, "learning_rate": 4.993931177744734e-05, "loss": 1.5166, "step": 293 }, { "epoch": 0.046218239698166595, "grad_norm": 0.11053171753883362, "learning_rate": 4.9938880777151254e-05, "loss": 1.4459, "step": 294 }, { "epoch": 0.04637544459509914, "grad_norm": 0.11428084224462509, "learning_rate": 4.993844825368321e-05, "loss": 1.4448, "step": 295 }, { "epoch": 0.046532649492031675, "grad_norm": 0.10734150558710098, "learning_rate": 4.993801420706964e-05, "loss": 1.3388, "step": 296 }, { "epoch": 0.04668985438896422, "grad_norm": 0.11137369275093079, "learning_rate": 4.993757863733703e-05, "loss": 1.4155, "step": 297 }, { "epoch": 0.046847059285896754, "grad_norm": 0.1221408098936081, "learning_rate": 4.993714154451202e-05, "loss": 1.4884, "step": 298 }, { "epoch": 0.0470042641828293, "grad_norm": 0.11707969009876251, "learning_rate": 4.993670292862127e-05, "loss": 1.4605, "step": 299 }, { "epoch": 0.047161469079761834, "grad_norm": 0.11751751601696014, "learning_rate": 4.993626278969158e-05, "loss": 1.5538, "step": 300 }, { "epoch": 0.04731867397669438, "grad_norm": 0.11617731302976608, "learning_rate": 4.993582112774984e-05, "loss": 1.438, "step": 301 }, { "epoch": 0.04747587887362691, "grad_norm": 0.15164637565612793, "learning_rate": 4.993537794282302e-05, "loss": 1.4607, "step": 302 }, { "epoch": 0.047633083770559456, "grad_norm": 0.12434446811676025, "learning_rate": 4.9934933234938193e-05, "loss": 1.4167, "step": 303 }, { "epoch": 0.04779028866749199, "grad_norm": 0.12518739700317383, "learning_rate": 4.993448700412251e-05, "loss": 1.4003, "step": 304 }, { "epoch": 0.04794749356442453, "grad_norm": 0.11146944761276245, "learning_rate": 4.993403925040323e-05, "loss": 1.3913, "step": 305 }, { "epoch": 0.04810469846135707, "grad_norm": 0.11682326346635818, "learning_rate": 4.993358997380771e-05, "loss": 1.3415, "step": 306 }, { "epoch": 0.04826190335828961, "grad_norm": 0.1197504773736, "learning_rate": 4.993313917436336e-05, "loss": 1.515, "step": 307 }, { "epoch": 0.04841910825522215, "grad_norm": 0.14647473394870758, "learning_rate": 4.993268685209775e-05, "loss": 1.4529, "step": 308 }, { "epoch": 0.04857631315215469, "grad_norm": 0.12431525439023972, "learning_rate": 4.9932233007038484e-05, "loss": 1.5426, "step": 309 }, { "epoch": 0.04873351804908723, "grad_norm": 0.11715538799762726, "learning_rate": 4.9931777639213284e-05, "loss": 1.4615, "step": 310 }, { "epoch": 0.04889072294601977, "grad_norm": 0.12391876429319382, "learning_rate": 4.993132074864997e-05, "loss": 1.4138, "step": 311 }, { "epoch": 0.04904792784295231, "grad_norm": 0.11894181370735168, "learning_rate": 4.9930862335376444e-05, "loss": 1.4383, "step": 312 }, { "epoch": 0.049205132739884846, "grad_norm": 0.1225295439362526, "learning_rate": 4.9930402399420695e-05, "loss": 1.3847, "step": 313 }, { "epoch": 0.04936233763681739, "grad_norm": 0.11435995995998383, "learning_rate": 4.9929940940810825e-05, "loss": 1.4254, "step": 314 }, { "epoch": 0.049519542533749926, "grad_norm": 0.11988761276006699, "learning_rate": 4.9929477959575024e-05, "loss": 1.4787, "step": 315 }, { "epoch": 0.04967674743068247, "grad_norm": 0.11983373016119003, "learning_rate": 4.992901345574155e-05, "loss": 1.4341, "step": 316 }, { "epoch": 0.049833952327615005, "grad_norm": 0.13395054638385773, "learning_rate": 4.992854742933878e-05, "loss": 1.4315, "step": 317 }, { "epoch": 0.04999115722454754, "grad_norm": 0.12578143179416656, "learning_rate": 4.9928079880395186e-05, "loss": 1.4143, "step": 318 }, { "epoch": 0.050148362121480085, "grad_norm": 0.1401878446340561, "learning_rate": 4.992761080893932e-05, "loss": 1.4665, "step": 319 }, { "epoch": 0.05030556701841262, "grad_norm": 0.13048145174980164, "learning_rate": 4.9927140214999826e-05, "loss": 1.4266, "step": 320 }, { "epoch": 0.05030556701841262, "eval_loss": 1.450086236000061, "eval_runtime": 2316.1877, "eval_samples_per_second": 3.997, "eval_steps_per_second": 1.999, "step": 320 }, { "epoch": 0.050462771915345164, "grad_norm": 0.13121232390403748, "learning_rate": 4.992666809860545e-05, "loss": 1.4946, "step": 321 }, { "epoch": 0.0506199768122777, "grad_norm": 0.13547195494174957, "learning_rate": 4.9926194459785015e-05, "loss": 1.5532, "step": 322 }, { "epoch": 0.050777181709210244, "grad_norm": 0.11797169595956802, "learning_rate": 4.992571929856747e-05, "loss": 1.4118, "step": 323 }, { "epoch": 0.05093438660614278, "grad_norm": 0.12734922766685486, "learning_rate": 4.992524261498183e-05, "loss": 1.4427, "step": 324 }, { "epoch": 0.05109159150307532, "grad_norm": 0.12444902211427689, "learning_rate": 4.99247644090572e-05, "loss": 1.4369, "step": 325 }, { "epoch": 0.05124879640000786, "grad_norm": 0.12244518846273422, "learning_rate": 4.99242846808228e-05, "loss": 1.4587, "step": 326 }, { "epoch": 0.0514060012969404, "grad_norm": 0.12424397468566895, "learning_rate": 4.9923803430307916e-05, "loss": 1.3949, "step": 327 }, { "epoch": 0.05156320619387294, "grad_norm": 0.1352718621492386, "learning_rate": 4.9923320657541944e-05, "loss": 1.504, "step": 328 }, { "epoch": 0.05172041109080548, "grad_norm": 0.12855666875839233, "learning_rate": 4.992283636255438e-05, "loss": 1.4271, "step": 329 }, { "epoch": 0.05187761598773802, "grad_norm": 0.129829540848732, "learning_rate": 4.99223505453748e-05, "loss": 1.455, "step": 330 }, { "epoch": 0.052034820884670555, "grad_norm": 0.12780050933361053, "learning_rate": 4.992186320603286e-05, "loss": 1.4045, "step": 331 }, { "epoch": 0.0521920257816031, "grad_norm": 0.13515712320804596, "learning_rate": 4.992137434455834e-05, "loss": 1.4335, "step": 332 }, { "epoch": 0.052349230678535634, "grad_norm": 0.15026766061782837, "learning_rate": 4.99208839609811e-05, "loss": 1.5386, "step": 333 }, { "epoch": 0.05250643557546818, "grad_norm": 0.13422101736068726, "learning_rate": 4.992039205533108e-05, "loss": 1.454, "step": 334 }, { "epoch": 0.05266364047240071, "grad_norm": 0.13735777139663696, "learning_rate": 4.991989862763833e-05, "loss": 1.4415, "step": 335 }, { "epoch": 0.05282084536933326, "grad_norm": 0.12985137104988098, "learning_rate": 4.9919403677932994e-05, "loss": 1.385, "step": 336 }, { "epoch": 0.05297805026626579, "grad_norm": 0.1301167607307434, "learning_rate": 4.9918907206245285e-05, "loss": 1.4364, "step": 337 }, { "epoch": 0.053135255163198336, "grad_norm": 0.1407599002122879, "learning_rate": 4.991840921260553e-05, "loss": 1.4454, "step": 338 }, { "epoch": 0.05329246006013087, "grad_norm": 0.12763133645057678, "learning_rate": 4.9917909697044164e-05, "loss": 1.4008, "step": 339 }, { "epoch": 0.053449664957063416, "grad_norm": 0.1443052589893341, "learning_rate": 4.991740865959167e-05, "loss": 1.5184, "step": 340 }, { "epoch": 0.05360686985399595, "grad_norm": 0.13496418297290802, "learning_rate": 4.991690610027866e-05, "loss": 1.3888, "step": 341 }, { "epoch": 0.05376407475092849, "grad_norm": 0.12681293487548828, "learning_rate": 4.991640201913583e-05, "loss": 1.42, "step": 342 }, { "epoch": 0.05392127964786103, "grad_norm": 0.13178062438964844, "learning_rate": 4.9915896416193965e-05, "loss": 1.4178, "step": 343 }, { "epoch": 0.05407848454479357, "grad_norm": 0.14452503621578217, "learning_rate": 4.991538929148394e-05, "loss": 1.4248, "step": 344 }, { "epoch": 0.05423568944172611, "grad_norm": 0.1352955400943756, "learning_rate": 4.991488064503674e-05, "loss": 1.4304, "step": 345 }, { "epoch": 0.05439289433865865, "grad_norm": 0.14846469461917877, "learning_rate": 4.991437047688343e-05, "loss": 1.4784, "step": 346 }, { "epoch": 0.05455009923559119, "grad_norm": 0.12475849688053131, "learning_rate": 4.9913858787055156e-05, "loss": 1.4131, "step": 347 }, { "epoch": 0.054707304132523726, "grad_norm": 0.13835409283638, "learning_rate": 4.991334557558318e-05, "loss": 1.4913, "step": 348 }, { "epoch": 0.05486450902945627, "grad_norm": 0.13921529054641724, "learning_rate": 4.991283084249885e-05, "loss": 1.3713, "step": 349 }, { "epoch": 0.055021713926388806, "grad_norm": 0.13188250362873077, "learning_rate": 4.9912314587833586e-05, "loss": 1.3608, "step": 350 }, { "epoch": 0.05517891882332135, "grad_norm": 0.12457428872585297, "learning_rate": 4.991179681161895e-05, "loss": 1.4427, "step": 351 }, { "epoch": 0.055336123720253885, "grad_norm": 0.12452542781829834, "learning_rate": 4.9911277513886535e-05, "loss": 1.4179, "step": 352 }, { "epoch": 0.05549332861718643, "grad_norm": 0.14799195528030396, "learning_rate": 4.9910756694668074e-05, "loss": 1.4532, "step": 353 }, { "epoch": 0.055650533514118965, "grad_norm": 0.13485541939735413, "learning_rate": 4.991023435399538e-05, "loss": 1.4114, "step": 354 }, { "epoch": 0.0558077384110515, "grad_norm": 0.1422443389892578, "learning_rate": 4.990971049190034e-05, "loss": 1.377, "step": 355 }, { "epoch": 0.055964943307984044, "grad_norm": 0.12994804978370667, "learning_rate": 4.990918510841496e-05, "loss": 1.4474, "step": 356 }, { "epoch": 0.05612214820491658, "grad_norm": 0.1429785192012787, "learning_rate": 4.990865820357133e-05, "loss": 1.4435, "step": 357 }, { "epoch": 0.056279353101849124, "grad_norm": 0.12979790568351746, "learning_rate": 4.9908129777401625e-05, "loss": 1.4039, "step": 358 }, { "epoch": 0.05643655799878166, "grad_norm": 0.1332644671201706, "learning_rate": 4.990759982993812e-05, "loss": 1.4377, "step": 359 }, { "epoch": 0.0565937628957142, "grad_norm": 0.13796579837799072, "learning_rate": 4.99070683612132e-05, "loss": 1.3951, "step": 360 }, { "epoch": 0.05675096779264674, "grad_norm": 0.14315246045589447, "learning_rate": 4.9906535371259294e-05, "loss": 1.4042, "step": 361 }, { "epoch": 0.05690817268957928, "grad_norm": 0.1463768631219864, "learning_rate": 4.9906000860108974e-05, "loss": 1.461, "step": 362 }, { "epoch": 0.05706537758651182, "grad_norm": 0.14041170477867126, "learning_rate": 4.9905464827794884e-05, "loss": 1.4147, "step": 363 }, { "epoch": 0.05722258248344436, "grad_norm": 0.19242697954177856, "learning_rate": 4.990492727434976e-05, "loss": 1.3435, "step": 364 }, { "epoch": 0.0573797873803769, "grad_norm": 0.1556611955165863, "learning_rate": 4.990438819980644e-05, "loss": 1.4075, "step": 365 }, { "epoch": 0.05753699227730944, "grad_norm": 0.13157570362091064, "learning_rate": 4.990384760419784e-05, "loss": 1.3334, "step": 366 }, { "epoch": 0.05769419717424198, "grad_norm": 0.17953743040561676, "learning_rate": 4.990330548755698e-05, "loss": 1.4609, "step": 367 }, { "epoch": 0.057851402071174514, "grad_norm": 0.14179491996765137, "learning_rate": 4.990276184991697e-05, "loss": 1.4344, "step": 368 }, { "epoch": 0.05800860696810706, "grad_norm": 0.16522593796253204, "learning_rate": 4.9902216691311024e-05, "loss": 1.3794, "step": 369 }, { "epoch": 0.05816581186503959, "grad_norm": 0.12736016511917114, "learning_rate": 4.9901670011772425e-05, "loss": 1.4167, "step": 370 }, { "epoch": 0.058323016761972137, "grad_norm": 0.15869787335395813, "learning_rate": 4.990112181133456e-05, "loss": 1.4293, "step": 371 }, { "epoch": 0.05848022165890467, "grad_norm": 0.14410504698753357, "learning_rate": 4.990057209003093e-05, "loss": 1.4357, "step": 372 }, { "epoch": 0.058637426555837216, "grad_norm": 0.1567080020904541, "learning_rate": 4.9900020847895086e-05, "loss": 1.4146, "step": 373 }, { "epoch": 0.05879463145276975, "grad_norm": 0.1430107057094574, "learning_rate": 4.989946808496071e-05, "loss": 1.3415, "step": 374 }, { "epoch": 0.058951836349702295, "grad_norm": 0.146332785487175, "learning_rate": 4.989891380126156e-05, "loss": 1.496, "step": 375 }, { "epoch": 0.05910904124663483, "grad_norm": 0.13674487173557281, "learning_rate": 4.989835799683149e-05, "loss": 1.3611, "step": 376 }, { "epoch": 0.059266246143567375, "grad_norm": 0.1321984827518463, "learning_rate": 4.989780067170444e-05, "loss": 1.4695, "step": 377 }, { "epoch": 0.05942345104049991, "grad_norm": 0.1535942554473877, "learning_rate": 4.9897241825914464e-05, "loss": 1.3564, "step": 378 }, { "epoch": 0.059580655937432454, "grad_norm": 0.1538037806749344, "learning_rate": 4.989668145949568e-05, "loss": 1.3502, "step": 379 }, { "epoch": 0.05973786083436499, "grad_norm": 0.15744829177856445, "learning_rate": 4.989611957248232e-05, "loss": 1.4318, "step": 380 }, { "epoch": 0.05989506573129753, "grad_norm": 0.17178332805633545, "learning_rate": 4.98955561649087e-05, "loss": 1.4643, "step": 381 }, { "epoch": 0.06005227062823007, "grad_norm": 0.15913072228431702, "learning_rate": 4.989499123680923e-05, "loss": 1.487, "step": 382 }, { "epoch": 0.060209475525162606, "grad_norm": 0.15134060382843018, "learning_rate": 4.9894424788218415e-05, "loss": 1.4705, "step": 383 }, { "epoch": 0.06036668042209515, "grad_norm": 0.13704389333724976, "learning_rate": 4.989385681917085e-05, "loss": 1.4756, "step": 384 }, { "epoch": 0.060523885319027686, "grad_norm": 0.14025503396987915, "learning_rate": 4.989328732970122e-05, "loss": 1.443, "step": 385 }, { "epoch": 0.06068109021596023, "grad_norm": 0.1822325438261032, "learning_rate": 4.9892716319844325e-05, "loss": 1.3996, "step": 386 }, { "epoch": 0.060838295112892765, "grad_norm": 0.15639656782150269, "learning_rate": 4.989214378963502e-05, "loss": 1.3656, "step": 387 }, { "epoch": 0.06099550000982531, "grad_norm": 0.15097728371620178, "learning_rate": 4.989156973910828e-05, "loss": 1.4055, "step": 388 }, { "epoch": 0.061152704906757845, "grad_norm": 0.18977142870426178, "learning_rate": 4.989099416829917e-05, "loss": 1.4472, "step": 389 }, { "epoch": 0.06130990980369039, "grad_norm": 0.1596304178237915, "learning_rate": 4.989041707724284e-05, "loss": 1.4373, "step": 390 }, { "epoch": 0.061467114700622924, "grad_norm": 0.171820729970932, "learning_rate": 4.988983846597454e-05, "loss": 1.468, "step": 391 }, { "epoch": 0.06162431959755547, "grad_norm": 0.14266176521778107, "learning_rate": 4.98892583345296e-05, "loss": 1.4037, "step": 392 }, { "epoch": 0.061781524494488003, "grad_norm": 0.13375528156757355, "learning_rate": 4.988867668294346e-05, "loss": 1.437, "step": 393 }, { "epoch": 0.06193872939142054, "grad_norm": 0.13332228362560272, "learning_rate": 4.988809351125165e-05, "loss": 1.3892, "step": 394 }, { "epoch": 0.06209593428835308, "grad_norm": 0.17180980741977692, "learning_rate": 4.988750881948977e-05, "loss": 1.3494, "step": 395 }, { "epoch": 0.06225313918528562, "grad_norm": 0.1419111043214798, "learning_rate": 4.988692260769355e-05, "loss": 1.3748, "step": 396 }, { "epoch": 0.06241034408221816, "grad_norm": 0.17256620526313782, "learning_rate": 4.9886334875898776e-05, "loss": 1.3549, "step": 397 }, { "epoch": 0.0625675489791507, "grad_norm": 0.2243422418832779, "learning_rate": 4.988574562414137e-05, "loss": 1.4465, "step": 398 }, { "epoch": 0.06272475387608324, "grad_norm": 0.15700723230838776, "learning_rate": 4.9885154852457294e-05, "loss": 1.4477, "step": 399 }, { "epoch": 0.06288195877301578, "grad_norm": 0.14497259259223938, "learning_rate": 4.988456256088264e-05, "loss": 1.3861, "step": 400 }, { "epoch": 0.06303916366994831, "grad_norm": 0.14747034013271332, "learning_rate": 4.988396874945359e-05, "loss": 1.4206, "step": 401 }, { "epoch": 0.06319636856688086, "grad_norm": 0.17671054601669312, "learning_rate": 4.98833734182064e-05, "loss": 1.2475, "step": 402 }, { "epoch": 0.0633535734638134, "grad_norm": 0.16974316537380219, "learning_rate": 4.9882776567177446e-05, "loss": 1.4955, "step": 403 }, { "epoch": 0.06351077836074594, "grad_norm": 0.15419775247573853, "learning_rate": 4.988217819640317e-05, "loss": 1.4209, "step": 404 }, { "epoch": 0.06366798325767847, "grad_norm": 0.13987664878368378, "learning_rate": 4.988157830592012e-05, "loss": 1.456, "step": 405 }, { "epoch": 0.06382518815461101, "grad_norm": 0.24560455977916718, "learning_rate": 4.988097689576493e-05, "loss": 1.3567, "step": 406 }, { "epoch": 0.06398239305154356, "grad_norm": 0.13870076835155487, "learning_rate": 4.9880373965974334e-05, "loss": 1.3752, "step": 407 }, { "epoch": 0.0641395979484761, "grad_norm": 0.16167718172073364, "learning_rate": 4.987976951658517e-05, "loss": 1.4766, "step": 408 }, { "epoch": 0.06429680284540863, "grad_norm": 0.1700398474931717, "learning_rate": 4.9879163547634346e-05, "loss": 1.427, "step": 409 }, { "epoch": 0.06445400774234117, "grad_norm": 0.15502458810806274, "learning_rate": 4.987855605915887e-05, "loss": 1.3965, "step": 410 }, { "epoch": 0.06461121263927372, "grad_norm": 0.14834032952785492, "learning_rate": 4.987794705119584e-05, "loss": 1.4399, "step": 411 }, { "epoch": 0.06476841753620625, "grad_norm": 0.22443649172782898, "learning_rate": 4.987733652378246e-05, "loss": 1.3736, "step": 412 }, { "epoch": 0.06492562243313879, "grad_norm": 0.14396560192108154, "learning_rate": 4.9876724476956015e-05, "loss": 1.4648, "step": 413 }, { "epoch": 0.06508282733007133, "grad_norm": 0.15352006256580353, "learning_rate": 4.987611091075389e-05, "loss": 1.4988, "step": 414 }, { "epoch": 0.06524003222700388, "grad_norm": 0.13210074603557587, "learning_rate": 4.987549582521356e-05, "loss": 1.3705, "step": 415 }, { "epoch": 0.06539723712393641, "grad_norm": 0.16056782007217407, "learning_rate": 4.98748792203726e-05, "loss": 1.3388, "step": 416 }, { "epoch": 0.06555444202086895, "grad_norm": 0.18992343544960022, "learning_rate": 4.9874261096268647e-05, "loss": 1.3842, "step": 417 }, { "epoch": 0.06571164691780149, "grad_norm": 0.1789916455745697, "learning_rate": 4.9873641452939466e-05, "loss": 1.3622, "step": 418 }, { "epoch": 0.06586885181473402, "grad_norm": 0.21043789386749268, "learning_rate": 4.9873020290422915e-05, "loss": 1.3477, "step": 419 }, { "epoch": 0.06602605671166657, "grad_norm": 0.15355254709720612, "learning_rate": 4.987239760875691e-05, "loss": 1.3643, "step": 420 }, { "epoch": 0.06618326160859911, "grad_norm": 0.1433190107345581, "learning_rate": 4.9871773407979496e-05, "loss": 1.3753, "step": 421 }, { "epoch": 0.06634046650553165, "grad_norm": 0.17479249835014343, "learning_rate": 4.987114768812879e-05, "loss": 1.3809, "step": 422 }, { "epoch": 0.06649767140246418, "grad_norm": 0.186944842338562, "learning_rate": 4.987052044924302e-05, "loss": 1.3616, "step": 423 }, { "epoch": 0.06665487629939673, "grad_norm": 0.15202952921390533, "learning_rate": 4.986989169136048e-05, "loss": 1.4479, "step": 424 }, { "epoch": 0.06681208119632927, "grad_norm": 0.16295532882213593, "learning_rate": 4.9869261414519575e-05, "loss": 1.3713, "step": 425 }, { "epoch": 0.0669692860932618, "grad_norm": 0.19577625393867493, "learning_rate": 4.986862961875881e-05, "loss": 1.4199, "step": 426 }, { "epoch": 0.06712649099019434, "grad_norm": 0.22768542170524597, "learning_rate": 4.986799630411677e-05, "loss": 1.3529, "step": 427 }, { "epoch": 0.06728369588712689, "grad_norm": 0.25184011459350586, "learning_rate": 4.986736147063212e-05, "loss": 1.3944, "step": 428 }, { "epoch": 0.06744090078405943, "grad_norm": 0.15565118193626404, "learning_rate": 4.986672511834366e-05, "loss": 1.4505, "step": 429 }, { "epoch": 0.06759810568099196, "grad_norm": 0.16559922695159912, "learning_rate": 4.986608724729024e-05, "loss": 1.3742, "step": 430 }, { "epoch": 0.0677553105779245, "grad_norm": 0.14826242625713348, "learning_rate": 4.986544785751081e-05, "loss": 1.4008, "step": 431 }, { "epoch": 0.06791251547485704, "grad_norm": 0.16543184220790863, "learning_rate": 4.986480694904444e-05, "loss": 1.3433, "step": 432 }, { "epoch": 0.06806972037178959, "grad_norm": 0.15332931280136108, "learning_rate": 4.986416452193027e-05, "loss": 1.4459, "step": 433 }, { "epoch": 0.06822692526872212, "grad_norm": 0.18880733847618103, "learning_rate": 4.986352057620752e-05, "loss": 1.3902, "step": 434 }, { "epoch": 0.06838413016565466, "grad_norm": 0.1513829231262207, "learning_rate": 4.986287511191554e-05, "loss": 1.3485, "step": 435 }, { "epoch": 0.0685413350625872, "grad_norm": 0.15241704881191254, "learning_rate": 4.9862228129093745e-05, "loss": 1.3051, "step": 436 }, { "epoch": 0.06869853995951974, "grad_norm": 0.1956702321767807, "learning_rate": 4.986157962778165e-05, "loss": 1.4647, "step": 437 }, { "epoch": 0.06885574485645228, "grad_norm": 0.2027936428785324, "learning_rate": 4.9860929608018866e-05, "loss": 1.3602, "step": 438 }, { "epoch": 0.06901294975338482, "grad_norm": 0.1623186320066452, "learning_rate": 4.986027806984509e-05, "loss": 1.4154, "step": 439 }, { "epoch": 0.06917015465031735, "grad_norm": 0.16111283004283905, "learning_rate": 4.985962501330011e-05, "loss": 1.4311, "step": 440 }, { "epoch": 0.0693273595472499, "grad_norm": 0.16754299402236938, "learning_rate": 4.985897043842382e-05, "loss": 1.349, "step": 441 }, { "epoch": 0.06948456444418244, "grad_norm": 0.1766330897808075, "learning_rate": 4.985831434525621e-05, "loss": 1.3714, "step": 442 }, { "epoch": 0.06964176934111498, "grad_norm": 0.1742810308933258, "learning_rate": 4.985765673383733e-05, "loss": 1.4161, "step": 443 }, { "epoch": 0.06979897423804751, "grad_norm": 0.17025281488895416, "learning_rate": 4.985699760420736e-05, "loss": 1.3925, "step": 444 }, { "epoch": 0.06995617913498005, "grad_norm": 0.19201375544071198, "learning_rate": 4.985633695640655e-05, "loss": 1.4158, "step": 445 }, { "epoch": 0.0701133840319126, "grad_norm": 0.1636267751455307, "learning_rate": 4.985567479047524e-05, "loss": 1.4071, "step": 446 }, { "epoch": 0.07027058892884513, "grad_norm": 0.19676333665847778, "learning_rate": 4.9855011106453894e-05, "loss": 1.3449, "step": 447 }, { "epoch": 0.07042779382577767, "grad_norm": 0.17712907493114471, "learning_rate": 4.985434590438303e-05, "loss": 1.3421, "step": 448 }, { "epoch": 0.07058499872271021, "grad_norm": 0.18515101075172424, "learning_rate": 4.985367918430329e-05, "loss": 1.4051, "step": 449 }, { "epoch": 0.07074220361964276, "grad_norm": 0.17168915271759033, "learning_rate": 4.985301094625538e-05, "loss": 1.3093, "step": 450 }, { "epoch": 0.0708994085165753, "grad_norm": 0.1891397386789322, "learning_rate": 4.9852341190280127e-05, "loss": 1.3075, "step": 451 }, { "epoch": 0.07105661341350783, "grad_norm": 0.17731457948684692, "learning_rate": 4.985166991641843e-05, "loss": 1.3986, "step": 452 }, { "epoch": 0.07121381831044037, "grad_norm": 0.18817296624183655, "learning_rate": 4.985099712471129e-05, "loss": 1.3531, "step": 453 }, { "epoch": 0.07137102320737292, "grad_norm": 0.1782791018486023, "learning_rate": 4.9850322815199795e-05, "loss": 1.4064, "step": 454 }, { "epoch": 0.07152822810430545, "grad_norm": 0.18053874373435974, "learning_rate": 4.984964698792514e-05, "loss": 1.4607, "step": 455 }, { "epoch": 0.07168543300123799, "grad_norm": 0.286338746547699, "learning_rate": 4.984896964292858e-05, "loss": 1.3036, "step": 456 }, { "epoch": 0.07184263789817052, "grad_norm": 0.2560707926750183, "learning_rate": 4.98482907802515e-05, "loss": 1.3428, "step": 457 }, { "epoch": 0.07199984279510306, "grad_norm": 0.19296897947788239, "learning_rate": 4.984761039993537e-05, "loss": 1.3502, "step": 458 }, { "epoch": 0.07215704769203561, "grad_norm": 0.19685949385166168, "learning_rate": 4.9846928502021725e-05, "loss": 1.4015, "step": 459 }, { "epoch": 0.07231425258896815, "grad_norm": 0.1548481583595276, "learning_rate": 4.984624508655223e-05, "loss": 1.3698, "step": 460 }, { "epoch": 0.07247145748590068, "grad_norm": 0.16076034307479858, "learning_rate": 4.984556015356862e-05, "loss": 1.3627, "step": 461 }, { "epoch": 0.07262866238283322, "grad_norm": 0.18571603298187256, "learning_rate": 4.9844873703112726e-05, "loss": 1.3506, "step": 462 }, { "epoch": 0.07278586727976577, "grad_norm": 0.1540035605430603, "learning_rate": 4.984418573522648e-05, "loss": 1.4483, "step": 463 }, { "epoch": 0.0729430721766983, "grad_norm": 0.1730145364999771, "learning_rate": 4.984349624995188e-05, "loss": 1.3678, "step": 464 }, { "epoch": 0.07310027707363084, "grad_norm": 0.26254212856292725, "learning_rate": 4.984280524733107e-05, "loss": 1.401, "step": 465 }, { "epoch": 0.07325748197056338, "grad_norm": 0.2079063057899475, "learning_rate": 4.984211272740623e-05, "loss": 1.3655, "step": 466 }, { "epoch": 0.07341468686749593, "grad_norm": 0.21711499989032745, "learning_rate": 4.9841418690219653e-05, "loss": 1.4011, "step": 467 }, { "epoch": 0.07357189176442847, "grad_norm": 0.18226252496242523, "learning_rate": 4.984072313581375e-05, "loss": 1.4213, "step": 468 }, { "epoch": 0.073729096661361, "grad_norm": 0.1463780552148819, "learning_rate": 4.9840026064230984e-05, "loss": 1.4519, "step": 469 }, { "epoch": 0.07388630155829354, "grad_norm": 0.18232892453670502, "learning_rate": 4.983932747551394e-05, "loss": 1.3657, "step": 470 }, { "epoch": 0.07404350645522607, "grad_norm": 0.19644559919834137, "learning_rate": 4.9838627369705285e-05, "loss": 1.3988, "step": 471 }, { "epoch": 0.07420071135215862, "grad_norm": 0.16292576491832733, "learning_rate": 4.983792574684776e-05, "loss": 1.4369, "step": 472 }, { "epoch": 0.07435791624909116, "grad_norm": 0.2244543433189392, "learning_rate": 4.983722260698425e-05, "loss": 1.4269, "step": 473 }, { "epoch": 0.0745151211460237, "grad_norm": 0.2582489848136902, "learning_rate": 4.9836517950157666e-05, "loss": 1.3986, "step": 474 }, { "epoch": 0.07467232604295623, "grad_norm": 0.15564194321632385, "learning_rate": 4.983581177641108e-05, "loss": 1.3871, "step": 475 }, { "epoch": 0.07482953093988878, "grad_norm": 0.2301008552312851, "learning_rate": 4.9835104085787596e-05, "loss": 1.3572, "step": 476 }, { "epoch": 0.07498673583682132, "grad_norm": 0.21603424847126007, "learning_rate": 4.9834394878330444e-05, "loss": 1.3803, "step": 477 }, { "epoch": 0.07514394073375386, "grad_norm": 0.16744717955589294, "learning_rate": 4.9833684154082937e-05, "loss": 1.4233, "step": 478 }, { "epoch": 0.07530114563068639, "grad_norm": 0.23016415536403656, "learning_rate": 4.98329719130885e-05, "loss": 1.3962, "step": 479 }, { "epoch": 0.07545835052761894, "grad_norm": 0.19687114655971527, "learning_rate": 4.983225815539061e-05, "loss": 1.3667, "step": 480 }, { "epoch": 0.07545835052761894, "eval_loss": 1.3748993873596191, "eval_runtime": 2315.5952, "eval_samples_per_second": 3.998, "eval_steps_per_second": 1.999, "step": 480 } ], "logging_steps": 1, "max_steps": 12722, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 160, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 2.4587428485188813e+18, "train_batch_size": 1, "trial_name": null, "trial_params": null }