{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.9999685610387433, "eval_steps": 500, "global_step": 23855, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 4.19186150089601e-05, "grad_norm": 11.043977737426758, "learning_rate": 1.3966480446927376e-08, "loss": 1.4028, "step": 1 }, { "epoch": 8.38372300179202e-05, "grad_norm": 12.333194732666016, "learning_rate": 2.793296089385475e-08, "loss": 1.2168, "step": 2 }, { "epoch": 0.0001257558450268803, "grad_norm": 11.61270523071289, "learning_rate": 4.189944134078213e-08, "loss": 1.3109, "step": 3 }, { "epoch": 0.0001676744600358404, "grad_norm": 9.490097045898438, "learning_rate": 5.58659217877095e-08, "loss": 1.3597, "step": 4 }, { "epoch": 0.00020959307504480053, "grad_norm": 10.877546310424805, "learning_rate": 6.983240223463688e-08, "loss": 1.4159, "step": 5 }, { "epoch": 0.0002515116900537606, "grad_norm": 11.821202278137207, "learning_rate": 8.379888268156426e-08, "loss": 1.4009, "step": 6 }, { "epoch": 0.0002934303050627207, "grad_norm": 8.896172523498535, "learning_rate": 9.776536312849162e-08, "loss": 1.2255, "step": 7 }, { "epoch": 0.0003353489200716808, "grad_norm": 10.6831636428833, "learning_rate": 1.11731843575419e-07, "loss": 1.3841, "step": 8 }, { "epoch": 0.00037726753508064096, "grad_norm": 9.90141487121582, "learning_rate": 1.2569832402234637e-07, "loss": 1.345, "step": 9 }, { "epoch": 0.00041918615008960106, "grad_norm": 11.198246002197266, "learning_rate": 1.3966480446927375e-07, "loss": 1.3203, "step": 10 }, { "epoch": 0.00046110476509856116, "grad_norm": 9.780250549316406, "learning_rate": 1.5363128491620113e-07, "loss": 1.3553, "step": 11 }, { "epoch": 0.0005030233801075212, "grad_norm": 9.907234191894531, "learning_rate": 1.6759776536312851e-07, "loss": 1.3481, "step": 12 }, { "epoch": 0.0005449419951164814, "grad_norm": 10.445611953735352, "learning_rate": 1.815642458100559e-07, "loss": 1.2618, "step": 13 }, { "epoch": 0.0005868606101254414, "grad_norm": 11.794252395629883, "learning_rate": 1.9553072625698325e-07, "loss": 1.3253, "step": 14 }, { "epoch": 0.0006287792251344016, "grad_norm": 10.397599220275879, "learning_rate": 2.0949720670391063e-07, "loss": 1.3597, "step": 15 }, { "epoch": 0.0006706978401433616, "grad_norm": 9.914612770080566, "learning_rate": 2.23463687150838e-07, "loss": 1.4242, "step": 16 }, { "epoch": 0.0007126164551523218, "grad_norm": 10.728616714477539, "learning_rate": 2.3743016759776536e-07, "loss": 1.2816, "step": 17 }, { "epoch": 0.0007545350701612819, "grad_norm": 9.650260925292969, "learning_rate": 2.5139664804469275e-07, "loss": 1.2085, "step": 18 }, { "epoch": 0.000796453685170242, "grad_norm": 8.366731643676758, "learning_rate": 2.6536312849162015e-07, "loss": 1.2565, "step": 19 }, { "epoch": 0.0008383723001792021, "grad_norm": 9.327199935913086, "learning_rate": 2.793296089385475e-07, "loss": 1.2295, "step": 20 }, { "epoch": 0.0008802909151881622, "grad_norm": 10.049495697021484, "learning_rate": 2.932960893854749e-07, "loss": 1.3847, "step": 21 }, { "epoch": 0.0009222095301971223, "grad_norm": 10.264235496520996, "learning_rate": 3.0726256983240227e-07, "loss": 1.4026, "step": 22 }, { "epoch": 0.0009641281452060824, "grad_norm": 9.986045837402344, "learning_rate": 3.212290502793296e-07, "loss": 1.3589, "step": 23 }, { "epoch": 0.0010060467602150424, "grad_norm": 9.068108558654785, "learning_rate": 3.3519553072625703e-07, "loss": 1.306, "step": 24 }, { "epoch": 0.0010479653752240026, "grad_norm": 8.677799224853516, "learning_rate": 3.491620111731844e-07, "loss": 1.319, "step": 25 }, { "epoch": 0.0010898839902329627, "grad_norm": 8.349884033203125, "learning_rate": 3.631284916201118e-07, "loss": 1.2788, "step": 26 }, { "epoch": 0.0011318026052419229, "grad_norm": 8.106966972351074, "learning_rate": 3.7709497206703914e-07, "loss": 1.2824, "step": 27 }, { "epoch": 0.0011737212202508828, "grad_norm": 8.445157051086426, "learning_rate": 3.910614525139665e-07, "loss": 1.1866, "step": 28 }, { "epoch": 0.001215639835259843, "grad_norm": 7.624929904937744, "learning_rate": 4.0502793296089385e-07, "loss": 1.2078, "step": 29 }, { "epoch": 0.0012575584502688031, "grad_norm": 7.446774005889893, "learning_rate": 4.1899441340782126e-07, "loss": 1.2591, "step": 30 }, { "epoch": 0.0012994770652777633, "grad_norm": 8.298200607299805, "learning_rate": 4.3296089385474867e-07, "loss": 1.1396, "step": 31 }, { "epoch": 0.0013413956802867232, "grad_norm": 7.57865571975708, "learning_rate": 4.46927374301676e-07, "loss": 1.2271, "step": 32 }, { "epoch": 0.0013833142952956834, "grad_norm": 8.09827995300293, "learning_rate": 4.608938547486034e-07, "loss": 1.2735, "step": 33 }, { "epoch": 0.0014252329103046435, "grad_norm": 6.101046562194824, "learning_rate": 4.7486033519553073e-07, "loss": 1.2001, "step": 34 }, { "epoch": 0.0014671515253136037, "grad_norm": 7.004069805145264, "learning_rate": 4.888268156424581e-07, "loss": 1.2802, "step": 35 }, { "epoch": 0.0015090701403225638, "grad_norm": 6.164199352264404, "learning_rate": 5.027932960893855e-07, "loss": 1.1305, "step": 36 }, { "epoch": 0.0015509887553315238, "grad_norm": 5.447147846221924, "learning_rate": 5.16759776536313e-07, "loss": 1.1148, "step": 37 }, { "epoch": 0.001592907370340484, "grad_norm": 5.174515247344971, "learning_rate": 5.307262569832403e-07, "loss": 1.1261, "step": 38 }, { "epoch": 0.001634825985349444, "grad_norm": 5.3310441970825195, "learning_rate": 5.446927374301677e-07, "loss": 1.0658, "step": 39 }, { "epoch": 0.0016767446003584042, "grad_norm": 4.908824443817139, "learning_rate": 5.58659217877095e-07, "loss": 0.9762, "step": 40 }, { "epoch": 0.0017186632153673642, "grad_norm": 5.3785858154296875, "learning_rate": 5.726256983240224e-07, "loss": 1.0261, "step": 41 }, { "epoch": 0.0017605818303763243, "grad_norm": 4.65134334564209, "learning_rate": 5.865921787709498e-07, "loss": 1.0826, "step": 42 }, { "epoch": 0.0018025004453852845, "grad_norm": 5.081592559814453, "learning_rate": 6.005586592178772e-07, "loss": 1.052, "step": 43 }, { "epoch": 0.0018444190603942446, "grad_norm": 4.959555625915527, "learning_rate": 6.145251396648045e-07, "loss": 1.0315, "step": 44 }, { "epoch": 0.0018863376754032046, "grad_norm": 4.905389785766602, "learning_rate": 6.284916201117319e-07, "loss": 1.0928, "step": 45 }, { "epoch": 0.0019282562904121647, "grad_norm": 4.698017597198486, "learning_rate": 6.424581005586592e-07, "loss": 1.0737, "step": 46 }, { "epoch": 0.0019701749054211247, "grad_norm": 4.49393892288208, "learning_rate": 6.564245810055866e-07, "loss": 1.0312, "step": 47 }, { "epoch": 0.002012093520430085, "grad_norm": 4.619397163391113, "learning_rate": 6.703910614525141e-07, "loss": 1.0277, "step": 48 }, { "epoch": 0.002054012135439045, "grad_norm": 4.435727596282959, "learning_rate": 6.843575418994414e-07, "loss": 1.0043, "step": 49 }, { "epoch": 0.002095930750448005, "grad_norm": 4.4498610496521, "learning_rate": 6.983240223463688e-07, "loss": 0.9735, "step": 50 }, { "epoch": 0.0021378493654569653, "grad_norm": 4.359391212463379, "learning_rate": 7.122905027932961e-07, "loss": 1.0174, "step": 51 }, { "epoch": 0.0021797679804659255, "grad_norm": 4.059515953063965, "learning_rate": 7.262569832402236e-07, "loss": 1.0338, "step": 52 }, { "epoch": 0.0022216865954748856, "grad_norm": 4.3099212646484375, "learning_rate": 7.402234636871509e-07, "loss": 0.9917, "step": 53 }, { "epoch": 0.0022636052104838458, "grad_norm": 3.765662431716919, "learning_rate": 7.541899441340783e-07, "loss": 1.0286, "step": 54 }, { "epoch": 0.002305523825492806, "grad_norm": 3.8867156505584717, "learning_rate": 7.681564245810056e-07, "loss": 0.9758, "step": 55 }, { "epoch": 0.0023474424405017656, "grad_norm": 4.202647686004639, "learning_rate": 7.82122905027933e-07, "loss": 1.0605, "step": 56 }, { "epoch": 0.002389361055510726, "grad_norm": 3.5507137775421143, "learning_rate": 7.960893854748604e-07, "loss": 0.9087, "step": 57 }, { "epoch": 0.002431279670519686, "grad_norm": 3.678880453109741, "learning_rate": 8.100558659217877e-07, "loss": 0.9527, "step": 58 }, { "epoch": 0.002473198285528646, "grad_norm": 4.039539337158203, "learning_rate": 8.240223463687152e-07, "loss": 0.9402, "step": 59 }, { "epoch": 0.0025151169005376063, "grad_norm": 3.592109441757202, "learning_rate": 8.379888268156425e-07, "loss": 0.8972, "step": 60 }, { "epoch": 0.0025570355155465664, "grad_norm": 4.118957042694092, "learning_rate": 8.519553072625699e-07, "loss": 0.9465, "step": 61 }, { "epoch": 0.0025989541305555266, "grad_norm": 3.8275771141052246, "learning_rate": 8.659217877094973e-07, "loss": 0.9771, "step": 62 }, { "epoch": 0.0026408727455644867, "grad_norm": 3.8876028060913086, "learning_rate": 8.798882681564247e-07, "loss": 0.9235, "step": 63 }, { "epoch": 0.0026827913605734464, "grad_norm": 3.2331321239471436, "learning_rate": 8.93854748603352e-07, "loss": 0.9001, "step": 64 }, { "epoch": 0.0027247099755824066, "grad_norm": 3.5235989093780518, "learning_rate": 9.078212290502794e-07, "loss": 0.9, "step": 65 }, { "epoch": 0.0027666285905913668, "grad_norm": 3.322288990020752, "learning_rate": 9.217877094972068e-07, "loss": 0.9597, "step": 66 }, { "epoch": 0.002808547205600327, "grad_norm": 3.75331449508667, "learning_rate": 9.357541899441341e-07, "loss": 0.9727, "step": 67 }, { "epoch": 0.002850465820609287, "grad_norm": 3.956204652786255, "learning_rate": 9.497206703910615e-07, "loss": 0.9365, "step": 68 }, { "epoch": 0.0028923844356182472, "grad_norm": 3.469529151916504, "learning_rate": 9.63687150837989e-07, "loss": 1.0074, "step": 69 }, { "epoch": 0.0029343030506272074, "grad_norm": 3.170593023300171, "learning_rate": 9.776536312849163e-07, "loss": 0.8921, "step": 70 }, { "epoch": 0.0029762216656361675, "grad_norm": 3.210873603820801, "learning_rate": 9.916201117318436e-07, "loss": 0.9514, "step": 71 }, { "epoch": 0.0030181402806451277, "grad_norm": 6.245716094970703, "learning_rate": 1.005586592178771e-06, "loss": 0.9468, "step": 72 }, { "epoch": 0.0030600588956540874, "grad_norm": 3.095719814300537, "learning_rate": 1.0195530726256985e-06, "loss": 0.9108, "step": 73 }, { "epoch": 0.0031019775106630476, "grad_norm": 3.0561957359313965, "learning_rate": 1.033519553072626e-06, "loss": 0.9891, "step": 74 }, { "epoch": 0.0031438961256720077, "grad_norm": 3.075324058532715, "learning_rate": 1.0474860335195533e-06, "loss": 0.8748, "step": 75 }, { "epoch": 0.003185814740680968, "grad_norm": 3.866781234741211, "learning_rate": 1.0614525139664806e-06, "loss": 0.8934, "step": 76 }, { "epoch": 0.003227733355689928, "grad_norm": 3.119115114212036, "learning_rate": 1.075418994413408e-06, "loss": 0.9103, "step": 77 }, { "epoch": 0.003269651970698888, "grad_norm": 2.709261178970337, "learning_rate": 1.0893854748603353e-06, "loss": 0.9553, "step": 78 }, { "epoch": 0.0033115705857078483, "grad_norm": 2.87870454788208, "learning_rate": 1.1033519553072627e-06, "loss": 0.8527, "step": 79 }, { "epoch": 0.0033534892007168085, "grad_norm": 3.117891788482666, "learning_rate": 1.11731843575419e-06, "loss": 0.8931, "step": 80 }, { "epoch": 0.003395407815725768, "grad_norm": 3.472135066986084, "learning_rate": 1.1312849162011174e-06, "loss": 0.9911, "step": 81 }, { "epoch": 0.0034373264307347284, "grad_norm": 3.512698173522949, "learning_rate": 1.1452513966480447e-06, "loss": 0.8922, "step": 82 }, { "epoch": 0.0034792450457436885, "grad_norm": 3.275850534439087, "learning_rate": 1.159217877094972e-06, "loss": 0.9024, "step": 83 }, { "epoch": 0.0035211636607526487, "grad_norm": 2.6427974700927734, "learning_rate": 1.1731843575418997e-06, "loss": 0.9037, "step": 84 }, { "epoch": 0.003563082275761609, "grad_norm": 3.282602310180664, "learning_rate": 1.187150837988827e-06, "loss": 0.8657, "step": 85 }, { "epoch": 0.003605000890770569, "grad_norm": 2.972087860107422, "learning_rate": 1.2011173184357544e-06, "loss": 0.8961, "step": 86 }, { "epoch": 0.003646919505779529, "grad_norm": 2.87984299659729, "learning_rate": 1.2150837988826817e-06, "loss": 0.7979, "step": 87 }, { "epoch": 0.0036888381207884893, "grad_norm": 2.9522838592529297, "learning_rate": 1.229050279329609e-06, "loss": 0.8691, "step": 88 }, { "epoch": 0.0037307567357974494, "grad_norm": 3.820802688598633, "learning_rate": 1.2430167597765364e-06, "loss": 0.937, "step": 89 }, { "epoch": 0.003772675350806409, "grad_norm": 3.1461021900177, "learning_rate": 1.2569832402234638e-06, "loss": 0.8526, "step": 90 }, { "epoch": 0.0038145939658153693, "grad_norm": 3.0969464778900146, "learning_rate": 1.2709497206703911e-06, "loss": 0.8688, "step": 91 }, { "epoch": 0.0038565125808243295, "grad_norm": 2.7656970024108887, "learning_rate": 1.2849162011173185e-06, "loss": 0.8305, "step": 92 }, { "epoch": 0.0038984311958332896, "grad_norm": 3.389080762863159, "learning_rate": 1.2988826815642458e-06, "loss": 0.8599, "step": 93 }, { "epoch": 0.003940349810842249, "grad_norm": 2.853841781616211, "learning_rate": 1.3128491620111732e-06, "loss": 0.8256, "step": 94 }, { "epoch": 0.00398226842585121, "grad_norm": 2.6804730892181396, "learning_rate": 1.3268156424581008e-06, "loss": 0.7704, "step": 95 }, { "epoch": 0.00402418704086017, "grad_norm": 3.852142810821533, "learning_rate": 1.3407821229050281e-06, "loss": 0.9018, "step": 96 }, { "epoch": 0.00406610565586913, "grad_norm": 2.8969810009002686, "learning_rate": 1.3547486033519555e-06, "loss": 0.8545, "step": 97 }, { "epoch": 0.00410802427087809, "grad_norm": 2.6420414447784424, "learning_rate": 1.3687150837988828e-06, "loss": 0.8558, "step": 98 }, { "epoch": 0.0041499428858870506, "grad_norm": 3.532609224319458, "learning_rate": 1.3826815642458102e-06, "loss": 0.8855, "step": 99 }, { "epoch": 0.00419186150089601, "grad_norm": 3.096846103668213, "learning_rate": 1.3966480446927375e-06, "loss": 0.8856, "step": 100 }, { "epoch": 0.004233780115904971, "grad_norm": 2.629955291748047, "learning_rate": 1.4106145251396649e-06, "loss": 0.7997, "step": 101 }, { "epoch": 0.004275698730913931, "grad_norm": 2.7514114379882812, "learning_rate": 1.4245810055865922e-06, "loss": 0.7959, "step": 102 }, { "epoch": 0.00431761734592289, "grad_norm": 3.4642157554626465, "learning_rate": 1.4385474860335196e-06, "loss": 0.8461, "step": 103 }, { "epoch": 0.004359535960931851, "grad_norm": 3.1932907104492188, "learning_rate": 1.4525139664804472e-06, "loss": 0.8287, "step": 104 }, { "epoch": 0.004401454575940811, "grad_norm": 3.1007745265960693, "learning_rate": 1.4664804469273743e-06, "loss": 0.8895, "step": 105 }, { "epoch": 0.004443373190949771, "grad_norm": 2.880373001098633, "learning_rate": 1.4804469273743019e-06, "loss": 0.8739, "step": 106 }, { "epoch": 0.004485291805958731, "grad_norm": 3.0392465591430664, "learning_rate": 1.4944134078212292e-06, "loss": 0.8524, "step": 107 }, { "epoch": 0.0045272104209676915, "grad_norm": 3.643795967102051, "learning_rate": 1.5083798882681566e-06, "loss": 0.8332, "step": 108 }, { "epoch": 0.004569129035976651, "grad_norm": 3.018862724304199, "learning_rate": 1.522346368715084e-06, "loss": 0.7999, "step": 109 }, { "epoch": 0.004611047650985612, "grad_norm": 2.6595449447631836, "learning_rate": 1.5363128491620113e-06, "loss": 0.9011, "step": 110 }, { "epoch": 0.0046529662659945716, "grad_norm": 2.9376864433288574, "learning_rate": 1.5502793296089386e-06, "loss": 0.8167, "step": 111 }, { "epoch": 0.004694884881003531, "grad_norm": 2.927375316619873, "learning_rate": 1.564245810055866e-06, "loss": 0.922, "step": 112 }, { "epoch": 0.004736803496012492, "grad_norm": 3.0302398204803467, "learning_rate": 1.5782122905027933e-06, "loss": 0.7909, "step": 113 }, { "epoch": 0.004778722111021452, "grad_norm": 2.6513712406158447, "learning_rate": 1.5921787709497207e-06, "loss": 0.8088, "step": 114 }, { "epoch": 0.004820640726030412, "grad_norm": 3.24501895904541, "learning_rate": 1.6061452513966483e-06, "loss": 0.8809, "step": 115 }, { "epoch": 0.004862559341039372, "grad_norm": 4.411985874176025, "learning_rate": 1.6201117318435754e-06, "loss": 0.8372, "step": 116 }, { "epoch": 0.0049044779560483325, "grad_norm": 2.618208885192871, "learning_rate": 1.634078212290503e-06, "loss": 0.7806, "step": 117 }, { "epoch": 0.004946396571057292, "grad_norm": 3.4706871509552, "learning_rate": 1.6480446927374303e-06, "loss": 0.8385, "step": 118 }, { "epoch": 0.004988315186066253, "grad_norm": 2.9874985218048096, "learning_rate": 1.6620111731843577e-06, "loss": 0.7559, "step": 119 }, { "epoch": 0.0050302338010752125, "grad_norm": 2.927576780319214, "learning_rate": 1.675977653631285e-06, "loss": 0.8403, "step": 120 }, { "epoch": 0.005072152416084172, "grad_norm": 4.5653886795043945, "learning_rate": 1.6899441340782124e-06, "loss": 0.8248, "step": 121 }, { "epoch": 0.005114071031093133, "grad_norm": 3.5408616065979004, "learning_rate": 1.7039106145251397e-06, "loss": 0.8462, "step": 122 }, { "epoch": 0.0051559896461020925, "grad_norm": 3.509172201156616, "learning_rate": 1.717877094972067e-06, "loss": 0.8156, "step": 123 }, { "epoch": 0.005197908261111053, "grad_norm": 3.104031562805176, "learning_rate": 1.7318435754189947e-06, "loss": 0.8286, "step": 124 }, { "epoch": 0.005239826876120013, "grad_norm": 2.805504560470581, "learning_rate": 1.7458100558659218e-06, "loss": 0.8363, "step": 125 }, { "epoch": 0.0052817454911289734, "grad_norm": 2.7023589611053467, "learning_rate": 1.7597765363128494e-06, "loss": 0.8299, "step": 126 }, { "epoch": 0.005323664106137933, "grad_norm": 2.272128105163574, "learning_rate": 1.7737430167597765e-06, "loss": 0.7574, "step": 127 }, { "epoch": 0.005365582721146893, "grad_norm": 3.2294626235961914, "learning_rate": 1.787709497206704e-06, "loss": 0.8317, "step": 128 }, { "epoch": 0.0054075013361558535, "grad_norm": 3.1404478549957275, "learning_rate": 1.8016759776536314e-06, "loss": 0.8228, "step": 129 }, { "epoch": 0.005449419951164813, "grad_norm": 2.7537453174591064, "learning_rate": 1.8156424581005588e-06, "loss": 0.7876, "step": 130 }, { "epoch": 0.005491338566173774, "grad_norm": 3.156856060028076, "learning_rate": 1.8296089385474861e-06, "loss": 0.8094, "step": 131 }, { "epoch": 0.0055332571811827335, "grad_norm": 2.822392702102661, "learning_rate": 1.8435754189944135e-06, "loss": 0.7407, "step": 132 }, { "epoch": 0.005575175796191694, "grad_norm": 2.899127960205078, "learning_rate": 1.857541899441341e-06, "loss": 0.7459, "step": 133 }, { "epoch": 0.005617094411200654, "grad_norm": 3.5183870792388916, "learning_rate": 1.8715083798882682e-06, "loss": 0.7979, "step": 134 }, { "epoch": 0.005659013026209614, "grad_norm": 2.3882830142974854, "learning_rate": 1.8854748603351958e-06, "loss": 0.7713, "step": 135 }, { "epoch": 0.005700931641218574, "grad_norm": 2.919219493865967, "learning_rate": 1.899441340782123e-06, "loss": 0.7247, "step": 136 }, { "epoch": 0.005742850256227534, "grad_norm": 2.847912549972534, "learning_rate": 1.9134078212290505e-06, "loss": 0.8031, "step": 137 }, { "epoch": 0.0057847688712364944, "grad_norm": 3.1799099445343018, "learning_rate": 1.927374301675978e-06, "loss": 0.8376, "step": 138 }, { "epoch": 0.005826687486245454, "grad_norm": 4.1890788078308105, "learning_rate": 1.941340782122905e-06, "loss": 0.7457, "step": 139 }, { "epoch": 0.005868606101254415, "grad_norm": 3.087230920791626, "learning_rate": 1.9553072625698325e-06, "loss": 0.7848, "step": 140 }, { "epoch": 0.0059105247162633745, "grad_norm": 4.844302654266357, "learning_rate": 1.96927374301676e-06, "loss": 0.8305, "step": 141 }, { "epoch": 0.005952443331272335, "grad_norm": 2.881718158721924, "learning_rate": 1.9832402234636873e-06, "loss": 0.7746, "step": 142 }, { "epoch": 0.005994361946281295, "grad_norm": 2.837083101272583, "learning_rate": 1.9972067039106146e-06, "loss": 0.8009, "step": 143 }, { "epoch": 0.006036280561290255, "grad_norm": 3.235560417175293, "learning_rate": 2.011173184357542e-06, "loss": 0.7824, "step": 144 }, { "epoch": 0.006078199176299215, "grad_norm": 3.165954828262329, "learning_rate": 2.0251396648044693e-06, "loss": 0.8579, "step": 145 }, { "epoch": 0.006120117791308175, "grad_norm": 6.9068732261657715, "learning_rate": 2.039106145251397e-06, "loss": 0.8571, "step": 146 }, { "epoch": 0.006162036406317135, "grad_norm": 2.7789406776428223, "learning_rate": 2.053072625698324e-06, "loss": 0.8025, "step": 147 }, { "epoch": 0.006203955021326095, "grad_norm": 3.560725212097168, "learning_rate": 2.067039106145252e-06, "loss": 0.7463, "step": 148 }, { "epoch": 0.006245873636335056, "grad_norm": 2.554861545562744, "learning_rate": 2.0810055865921787e-06, "loss": 0.7824, "step": 149 }, { "epoch": 0.006287792251344015, "grad_norm": 2.869584560394287, "learning_rate": 2.0949720670391065e-06, "loss": 0.7648, "step": 150 }, { "epoch": 0.006329710866352976, "grad_norm": 2.8584115505218506, "learning_rate": 2.108938547486034e-06, "loss": 0.8709, "step": 151 }, { "epoch": 0.006371629481361936, "grad_norm": 2.587827682495117, "learning_rate": 2.1229050279329612e-06, "loss": 0.705, "step": 152 }, { "epoch": 0.006413548096370896, "grad_norm": 3.909536123275757, "learning_rate": 2.1368715083798886e-06, "loss": 0.7473, "step": 153 }, { "epoch": 0.006455466711379856, "grad_norm": 2.9165918827056885, "learning_rate": 2.150837988826816e-06, "loss": 0.7328, "step": 154 }, { "epoch": 0.006497385326388816, "grad_norm": 7.03745174407959, "learning_rate": 2.1648044692737433e-06, "loss": 0.8068, "step": 155 }, { "epoch": 0.006539303941397776, "grad_norm": 3.1578030586242676, "learning_rate": 2.1787709497206706e-06, "loss": 0.7399, "step": 156 }, { "epoch": 0.006581222556406736, "grad_norm": 3.001016616821289, "learning_rate": 2.192737430167598e-06, "loss": 0.7332, "step": 157 }, { "epoch": 0.006623141171415697, "grad_norm": 3.170642852783203, "learning_rate": 2.2067039106145253e-06, "loss": 0.8439, "step": 158 }, { "epoch": 0.006665059786424656, "grad_norm": 3.07902193069458, "learning_rate": 2.2206703910614527e-06, "loss": 0.7343, "step": 159 }, { "epoch": 0.006706978401433617, "grad_norm": 2.8774187564849854, "learning_rate": 2.23463687150838e-06, "loss": 0.8452, "step": 160 }, { "epoch": 0.006748897016442577, "grad_norm": 2.823042392730713, "learning_rate": 2.2486033519553074e-06, "loss": 0.7802, "step": 161 }, { "epoch": 0.006790815631451536, "grad_norm": 2.8480947017669678, "learning_rate": 2.2625698324022348e-06, "loss": 0.7446, "step": 162 }, { "epoch": 0.006832734246460497, "grad_norm": 3.1317436695098877, "learning_rate": 2.276536312849162e-06, "loss": 0.7851, "step": 163 }, { "epoch": 0.006874652861469457, "grad_norm": 3.1690399646759033, "learning_rate": 2.2905027932960895e-06, "loss": 0.825, "step": 164 }, { "epoch": 0.006916571476478417, "grad_norm": 3.4310126304626465, "learning_rate": 2.304469273743017e-06, "loss": 0.7387, "step": 165 }, { "epoch": 0.006958490091487377, "grad_norm": 2.942338466644287, "learning_rate": 2.318435754189944e-06, "loss": 0.7533, "step": 166 }, { "epoch": 0.007000408706496338, "grad_norm": 2.445305109024048, "learning_rate": 2.3324022346368715e-06, "loss": 0.7195, "step": 167 }, { "epoch": 0.007042327321505297, "grad_norm": 2.878413438796997, "learning_rate": 2.3463687150837993e-06, "loss": 0.8234, "step": 168 }, { "epoch": 0.007084245936514258, "grad_norm": 2.6861331462860107, "learning_rate": 2.3603351955307262e-06, "loss": 0.772, "step": 169 }, { "epoch": 0.007126164551523218, "grad_norm": 2.8491790294647217, "learning_rate": 2.374301675977654e-06, "loss": 0.7383, "step": 170 }, { "epoch": 0.007168083166532177, "grad_norm": 2.6254913806915283, "learning_rate": 2.3882681564245814e-06, "loss": 0.7231, "step": 171 }, { "epoch": 0.007210001781541138, "grad_norm": 2.673827886581421, "learning_rate": 2.4022346368715087e-06, "loss": 0.7819, "step": 172 }, { "epoch": 0.007251920396550098, "grad_norm": 2.696357011795044, "learning_rate": 2.416201117318436e-06, "loss": 0.8283, "step": 173 }, { "epoch": 0.007293839011559058, "grad_norm": 3.0996456146240234, "learning_rate": 2.4301675977653634e-06, "loss": 0.7338, "step": 174 }, { "epoch": 0.007335757626568018, "grad_norm": 2.5453052520751953, "learning_rate": 2.4441340782122908e-06, "loss": 0.7632, "step": 175 }, { "epoch": 0.007377676241576979, "grad_norm": 2.9189231395721436, "learning_rate": 2.458100558659218e-06, "loss": 0.8123, "step": 176 }, { "epoch": 0.007419594856585938, "grad_norm": 2.913210868835449, "learning_rate": 2.4720670391061455e-06, "loss": 0.7645, "step": 177 }, { "epoch": 0.007461513471594899, "grad_norm": 2.9248435497283936, "learning_rate": 2.486033519553073e-06, "loss": 0.687, "step": 178 }, { "epoch": 0.007503432086603859, "grad_norm": 3.0863420963287354, "learning_rate": 2.5e-06, "loss": 0.8398, "step": 179 }, { "epoch": 0.007545350701612818, "grad_norm": 3.0230460166931152, "learning_rate": 2.5139664804469276e-06, "loss": 0.7921, "step": 180 }, { "epoch": 0.007587269316621779, "grad_norm": 2.7897536754608154, "learning_rate": 2.5279329608938553e-06, "loss": 0.8153, "step": 181 }, { "epoch": 0.007629187931630739, "grad_norm": 2.815363645553589, "learning_rate": 2.5418994413407823e-06, "loss": 0.8191, "step": 182 }, { "epoch": 0.007671106546639699, "grad_norm": 4.857927322387695, "learning_rate": 2.5558659217877096e-06, "loss": 0.7371, "step": 183 }, { "epoch": 0.007713025161648659, "grad_norm": 3.084327220916748, "learning_rate": 2.569832402234637e-06, "loss": 0.7725, "step": 184 }, { "epoch": 0.0077549437766576195, "grad_norm": 2.924710750579834, "learning_rate": 2.5837988826815648e-06, "loss": 0.7868, "step": 185 }, { "epoch": 0.007796862391666579, "grad_norm": 2.5783791542053223, "learning_rate": 2.5977653631284917e-06, "loss": 0.7231, "step": 186 }, { "epoch": 0.00783878100667554, "grad_norm": 4.226300239562988, "learning_rate": 2.611731843575419e-06, "loss": 0.699, "step": 187 }, { "epoch": 0.007880699621684499, "grad_norm": 2.8547775745391846, "learning_rate": 2.6256983240223464e-06, "loss": 0.7552, "step": 188 }, { "epoch": 0.00792261823669346, "grad_norm": 2.822194814682007, "learning_rate": 2.639664804469274e-06, "loss": 0.7327, "step": 189 }, { "epoch": 0.00796453685170242, "grad_norm": 2.5158987045288086, "learning_rate": 2.6536312849162015e-06, "loss": 0.7101, "step": 190 }, { "epoch": 0.00800645546671138, "grad_norm": 2.4685606956481934, "learning_rate": 2.6675977653631285e-06, "loss": 0.707, "step": 191 }, { "epoch": 0.00804837408172034, "grad_norm": 3.3183419704437256, "learning_rate": 2.6815642458100562e-06, "loss": 0.7405, "step": 192 }, { "epoch": 0.0080902926967293, "grad_norm": 2.983234405517578, "learning_rate": 2.6955307262569836e-06, "loss": 0.7343, "step": 193 }, { "epoch": 0.00813221131173826, "grad_norm": 3.9679718017578125, "learning_rate": 2.709497206703911e-06, "loss": 0.7034, "step": 194 }, { "epoch": 0.008174129926747221, "grad_norm": 3.1161153316497803, "learning_rate": 2.723463687150838e-06, "loss": 0.785, "step": 195 }, { "epoch": 0.00821604854175618, "grad_norm": 2.7596611976623535, "learning_rate": 2.7374301675977656e-06, "loss": 0.7332, "step": 196 }, { "epoch": 0.00825796715676514, "grad_norm": 2.901865243911743, "learning_rate": 2.751396648044693e-06, "loss": 0.7596, "step": 197 }, { "epoch": 0.008299885771774101, "grad_norm": 2.9487693309783936, "learning_rate": 2.7653631284916204e-06, "loss": 0.7835, "step": 198 }, { "epoch": 0.00834180438678306, "grad_norm": 2.833528757095337, "learning_rate": 2.779329608938548e-06, "loss": 0.7088, "step": 199 }, { "epoch": 0.00838372300179202, "grad_norm": 2.8041226863861084, "learning_rate": 2.793296089385475e-06, "loss": 0.7629, "step": 200 }, { "epoch": 0.008425641616800981, "grad_norm": 2.7175652980804443, "learning_rate": 2.8072625698324024e-06, "loss": 0.8026, "step": 201 }, { "epoch": 0.008467560231809942, "grad_norm": 2.6939637660980225, "learning_rate": 2.8212290502793298e-06, "loss": 0.7779, "step": 202 }, { "epoch": 0.0085094788468189, "grad_norm": 4.562593460083008, "learning_rate": 2.8351955307262576e-06, "loss": 0.7133, "step": 203 }, { "epoch": 0.008551397461827861, "grad_norm": 3.401461362838745, "learning_rate": 2.8491620111731845e-06, "loss": 0.8009, "step": 204 }, { "epoch": 0.008593316076836822, "grad_norm": 3.1963589191436768, "learning_rate": 2.863128491620112e-06, "loss": 0.7093, "step": 205 }, { "epoch": 0.00863523469184578, "grad_norm": 3.010451316833496, "learning_rate": 2.877094972067039e-06, "loss": 0.7599, "step": 206 }, { "epoch": 0.008677153306854741, "grad_norm": 4.295202255249023, "learning_rate": 2.891061452513967e-06, "loss": 0.7521, "step": 207 }, { "epoch": 0.008719071921863702, "grad_norm": 3.1220977306365967, "learning_rate": 2.9050279329608943e-06, "loss": 0.7666, "step": 208 }, { "epoch": 0.008760990536872662, "grad_norm": 2.8612353801727295, "learning_rate": 2.9189944134078213e-06, "loss": 0.6839, "step": 209 }, { "epoch": 0.008802909151881621, "grad_norm": 2.45243239402771, "learning_rate": 2.9329608938547486e-06, "loss": 0.7356, "step": 210 }, { "epoch": 0.008844827766890582, "grad_norm": 2.2479686737060547, "learning_rate": 2.9469273743016764e-06, "loss": 0.6044, "step": 211 }, { "epoch": 0.008886746381899542, "grad_norm": 2.902117967605591, "learning_rate": 2.9608938547486037e-06, "loss": 0.7558, "step": 212 }, { "epoch": 0.008928664996908501, "grad_norm": 2.455935001373291, "learning_rate": 2.9748603351955307e-06, "loss": 0.7156, "step": 213 }, { "epoch": 0.008970583611917462, "grad_norm": 2.6825640201568604, "learning_rate": 2.9888268156424584e-06, "loss": 0.7143, "step": 214 }, { "epoch": 0.009012502226926422, "grad_norm": 4.025918960571289, "learning_rate": 3.002793296089386e-06, "loss": 0.7103, "step": 215 }, { "epoch": 0.009054420841935383, "grad_norm": 2.8707778453826904, "learning_rate": 3.016759776536313e-06, "loss": 0.7338, "step": 216 }, { "epoch": 0.009096339456944342, "grad_norm": 4.088874340057373, "learning_rate": 3.030726256983241e-06, "loss": 0.7119, "step": 217 }, { "epoch": 0.009138258071953302, "grad_norm": 2.8198301792144775, "learning_rate": 3.044692737430168e-06, "loss": 0.6392, "step": 218 }, { "epoch": 0.009180176686962263, "grad_norm": 3.557756185531616, "learning_rate": 3.0586592178770952e-06, "loss": 0.6871, "step": 219 }, { "epoch": 0.009222095301971224, "grad_norm": 3.385762929916382, "learning_rate": 3.0726256983240226e-06, "loss": 0.7717, "step": 220 }, { "epoch": 0.009264013916980183, "grad_norm": 3.633345127105713, "learning_rate": 3.0865921787709503e-06, "loss": 0.7511, "step": 221 }, { "epoch": 0.009305932531989143, "grad_norm": 3.783918857574463, "learning_rate": 3.1005586592178773e-06, "loss": 0.7912, "step": 222 }, { "epoch": 0.009347851146998104, "grad_norm": 3.423413038253784, "learning_rate": 3.1145251396648046e-06, "loss": 0.7675, "step": 223 }, { "epoch": 0.009389769762007063, "grad_norm": 2.5598065853118896, "learning_rate": 3.128491620111732e-06, "loss": 0.7074, "step": 224 }, { "epoch": 0.009431688377016023, "grad_norm": 3.2848143577575684, "learning_rate": 3.1424581005586598e-06, "loss": 0.7383, "step": 225 }, { "epoch": 0.009473606992024984, "grad_norm": 2.9649925231933594, "learning_rate": 3.1564245810055867e-06, "loss": 0.7373, "step": 226 }, { "epoch": 0.009515525607033944, "grad_norm": 6.60773229598999, "learning_rate": 3.170391061452514e-06, "loss": 0.825, "step": 227 }, { "epoch": 0.009557444222042903, "grad_norm": 3.2311365604400635, "learning_rate": 3.1843575418994414e-06, "loss": 0.7301, "step": 228 }, { "epoch": 0.009599362837051864, "grad_norm": 2.8749194145202637, "learning_rate": 3.198324022346369e-06, "loss": 0.7313, "step": 229 }, { "epoch": 0.009641281452060824, "grad_norm": 2.8163068294525146, "learning_rate": 3.2122905027932965e-06, "loss": 0.6866, "step": 230 }, { "epoch": 0.009683200067069783, "grad_norm": 2.788806915283203, "learning_rate": 3.2262569832402235e-06, "loss": 0.7481, "step": 231 }, { "epoch": 0.009725118682078744, "grad_norm": 2.8023436069488525, "learning_rate": 3.240223463687151e-06, "loss": 0.7197, "step": 232 }, { "epoch": 0.009767037297087704, "grad_norm": 2.73837947845459, "learning_rate": 3.2541899441340786e-06, "loss": 0.7316, "step": 233 }, { "epoch": 0.009808955912096665, "grad_norm": 2.8294026851654053, "learning_rate": 3.268156424581006e-06, "loss": 0.7077, "step": 234 }, { "epoch": 0.009850874527105624, "grad_norm": 3.933138370513916, "learning_rate": 3.282122905027933e-06, "loss": 0.7969, "step": 235 }, { "epoch": 0.009892793142114584, "grad_norm": 2.6609456539154053, "learning_rate": 3.2960893854748607e-06, "loss": 0.7207, "step": 236 }, { "epoch": 0.009934711757123545, "grad_norm": 2.6862380504608154, "learning_rate": 3.310055865921788e-06, "loss": 0.7207, "step": 237 }, { "epoch": 0.009976630372132506, "grad_norm": 2.951667308807373, "learning_rate": 3.3240223463687154e-06, "loss": 0.7946, "step": 238 }, { "epoch": 0.010018548987141464, "grad_norm": 2.411879539489746, "learning_rate": 3.337988826815643e-06, "loss": 0.6793, "step": 239 }, { "epoch": 0.010060467602150425, "grad_norm": 2.7471470832824707, "learning_rate": 3.35195530726257e-06, "loss": 0.7034, "step": 240 }, { "epoch": 0.010102386217159386, "grad_norm": 2.7636725902557373, "learning_rate": 3.3659217877094974e-06, "loss": 0.737, "step": 241 }, { "epoch": 0.010144304832168344, "grad_norm": 25.283689498901367, "learning_rate": 3.3798882681564248e-06, "loss": 0.6365, "step": 242 }, { "epoch": 0.010186223447177305, "grad_norm": 3.0434410572052, "learning_rate": 3.3938547486033526e-06, "loss": 0.6741, "step": 243 }, { "epoch": 0.010228142062186266, "grad_norm": 2.6149771213531494, "learning_rate": 3.4078212290502795e-06, "loss": 0.6714, "step": 244 }, { "epoch": 0.010270060677195226, "grad_norm": 3.5512335300445557, "learning_rate": 3.421787709497207e-06, "loss": 0.7086, "step": 245 }, { "epoch": 0.010311979292204185, "grad_norm": 2.3900697231292725, "learning_rate": 3.435754189944134e-06, "loss": 0.6438, "step": 246 }, { "epoch": 0.010353897907213146, "grad_norm": 2.5329132080078125, "learning_rate": 3.449720670391062e-06, "loss": 0.733, "step": 247 }, { "epoch": 0.010395816522222106, "grad_norm": 3.146763324737549, "learning_rate": 3.4636871508379893e-06, "loss": 0.7543, "step": 248 }, { "epoch": 0.010437735137231065, "grad_norm": 2.5793352127075195, "learning_rate": 3.4776536312849163e-06, "loss": 0.6908, "step": 249 }, { "epoch": 0.010479653752240026, "grad_norm": 2.56219482421875, "learning_rate": 3.4916201117318436e-06, "loss": 0.6928, "step": 250 }, { "epoch": 0.010521572367248986, "grad_norm": 3.2761824131011963, "learning_rate": 3.5055865921787714e-06, "loss": 0.7713, "step": 251 }, { "epoch": 0.010563490982257947, "grad_norm": 3.1227619647979736, "learning_rate": 3.5195530726256988e-06, "loss": 0.6947, "step": 252 }, { "epoch": 0.010605409597266906, "grad_norm": 3.0068044662475586, "learning_rate": 3.5335195530726257e-06, "loss": 0.7393, "step": 253 }, { "epoch": 0.010647328212275866, "grad_norm": 2.8906219005584717, "learning_rate": 3.547486033519553e-06, "loss": 0.7109, "step": 254 }, { "epoch": 0.010689246827284827, "grad_norm": 2.7756505012512207, "learning_rate": 3.561452513966481e-06, "loss": 0.7541, "step": 255 }, { "epoch": 0.010731165442293786, "grad_norm": 4.215346813201904, "learning_rate": 3.575418994413408e-06, "loss": 0.6869, "step": 256 }, { "epoch": 0.010773084057302746, "grad_norm": 2.955638885498047, "learning_rate": 3.5893854748603355e-06, "loss": 0.7109, "step": 257 }, { "epoch": 0.010815002672311707, "grad_norm": 2.5789742469787598, "learning_rate": 3.603351955307263e-06, "loss": 0.6988, "step": 258 }, { "epoch": 0.010856921287320668, "grad_norm": 2.4248883724212646, "learning_rate": 3.6173184357541902e-06, "loss": 0.7131, "step": 259 }, { "epoch": 0.010898839902329626, "grad_norm": 2.708299160003662, "learning_rate": 3.6312849162011176e-06, "loss": 0.673, "step": 260 }, { "epoch": 0.010940758517338587, "grad_norm": 3.324216842651367, "learning_rate": 3.6452513966480454e-06, "loss": 0.7355, "step": 261 }, { "epoch": 0.010982677132347548, "grad_norm": 2.857961416244507, "learning_rate": 3.6592178770949723e-06, "loss": 0.7032, "step": 262 }, { "epoch": 0.011024595747356508, "grad_norm": 3.5762405395507812, "learning_rate": 3.6731843575418996e-06, "loss": 0.6898, "step": 263 }, { "epoch": 0.011066514362365467, "grad_norm": 2.324134588241577, "learning_rate": 3.687150837988827e-06, "loss": 0.6616, "step": 264 }, { "epoch": 0.011108432977374428, "grad_norm": 2.6841907501220703, "learning_rate": 3.7011173184357548e-06, "loss": 0.7448, "step": 265 }, { "epoch": 0.011150351592383388, "grad_norm": 5.289700031280518, "learning_rate": 3.715083798882682e-06, "loss": 0.7044, "step": 266 }, { "epoch": 0.011192270207392347, "grad_norm": 3.1530003547668457, "learning_rate": 3.729050279329609e-06, "loss": 0.7001, "step": 267 }, { "epoch": 0.011234188822401308, "grad_norm": 2.4380669593811035, "learning_rate": 3.7430167597765364e-06, "loss": 0.6908, "step": 268 }, { "epoch": 0.011276107437410268, "grad_norm": 3.653078556060791, "learning_rate": 3.756983240223464e-06, "loss": 0.6793, "step": 269 }, { "epoch": 0.011318026052419229, "grad_norm": 2.4566493034362793, "learning_rate": 3.7709497206703915e-06, "loss": 0.6856, "step": 270 }, { "epoch": 0.011359944667428188, "grad_norm": 2.4401514530181885, "learning_rate": 3.7849162011173185e-06, "loss": 0.6776, "step": 271 }, { "epoch": 0.011401863282437148, "grad_norm": 2.6851251125335693, "learning_rate": 3.798882681564246e-06, "loss": 0.7032, "step": 272 }, { "epoch": 0.011443781897446109, "grad_norm": 4.420756816864014, "learning_rate": 3.8128491620111736e-06, "loss": 0.6871, "step": 273 }, { "epoch": 0.011485700512455068, "grad_norm": 2.81062912940979, "learning_rate": 3.826815642458101e-06, "loss": 0.766, "step": 274 }, { "epoch": 0.011527619127464028, "grad_norm": 2.6485209465026855, "learning_rate": 3.840782122905028e-06, "loss": 0.6633, "step": 275 }, { "epoch": 0.011569537742472989, "grad_norm": 2.8563404083251953, "learning_rate": 3.854748603351956e-06, "loss": 0.7145, "step": 276 }, { "epoch": 0.01161145635748195, "grad_norm": 2.3327572345733643, "learning_rate": 3.868715083798883e-06, "loss": 0.7146, "step": 277 }, { "epoch": 0.011653374972490908, "grad_norm": 2.755711555480957, "learning_rate": 3.88268156424581e-06, "loss": 0.685, "step": 278 }, { "epoch": 0.011695293587499869, "grad_norm": 3.0609724521636963, "learning_rate": 3.896648044692738e-06, "loss": 0.7488, "step": 279 }, { "epoch": 0.01173721220250883, "grad_norm": 2.857595920562744, "learning_rate": 3.910614525139665e-06, "loss": 0.7036, "step": 280 }, { "epoch": 0.01177913081751779, "grad_norm": 2.373518705368042, "learning_rate": 3.9245810055865924e-06, "loss": 0.6217, "step": 281 }, { "epoch": 0.011821049432526749, "grad_norm": 2.7846057415008545, "learning_rate": 3.93854748603352e-06, "loss": 0.6632, "step": 282 }, { "epoch": 0.01186296804753571, "grad_norm": 3.3408989906311035, "learning_rate": 3.952513966480447e-06, "loss": 0.8263, "step": 283 }, { "epoch": 0.01190488666254467, "grad_norm": 3.7452585697174072, "learning_rate": 3.9664804469273745e-06, "loss": 0.7376, "step": 284 }, { "epoch": 0.011946805277553629, "grad_norm": 2.6985201835632324, "learning_rate": 3.980446927374302e-06, "loss": 0.7167, "step": 285 }, { "epoch": 0.01198872389256259, "grad_norm": 3.0542120933532715, "learning_rate": 3.994413407821229e-06, "loss": 0.7255, "step": 286 }, { "epoch": 0.01203064250757155, "grad_norm": 2.5447919368743896, "learning_rate": 4.0083798882681566e-06, "loss": 0.6644, "step": 287 }, { "epoch": 0.01207256112258051, "grad_norm": 3.3139374256134033, "learning_rate": 4.022346368715084e-06, "loss": 0.6748, "step": 288 }, { "epoch": 0.01211447973758947, "grad_norm": 2.9251606464385986, "learning_rate": 4.036312849162011e-06, "loss": 0.6863, "step": 289 }, { "epoch": 0.01215639835259843, "grad_norm": 2.424278736114502, "learning_rate": 4.050279329608939e-06, "loss": 0.7021, "step": 290 }, { "epoch": 0.01219831696760739, "grad_norm": 2.3473854064941406, "learning_rate": 4.064245810055866e-06, "loss": 0.7494, "step": 291 }, { "epoch": 0.01224023558261635, "grad_norm": 2.930497884750366, "learning_rate": 4.078212290502794e-06, "loss": 0.7206, "step": 292 }, { "epoch": 0.01228215419762531, "grad_norm": 2.687784433364868, "learning_rate": 4.0921787709497215e-06, "loss": 0.695, "step": 293 }, { "epoch": 0.01232407281263427, "grad_norm": 3.0904457569122314, "learning_rate": 4.106145251396648e-06, "loss": 0.6381, "step": 294 }, { "epoch": 0.012365991427643231, "grad_norm": 2.638981342315674, "learning_rate": 4.120111731843575e-06, "loss": 0.7004, "step": 295 }, { "epoch": 0.01240791004265219, "grad_norm": 4.9086713790893555, "learning_rate": 4.134078212290504e-06, "loss": 0.7008, "step": 296 }, { "epoch": 0.01244982865766115, "grad_norm": 2.3048059940338135, "learning_rate": 4.148044692737431e-06, "loss": 0.6678, "step": 297 }, { "epoch": 0.012491747272670111, "grad_norm": 2.5617387294769287, "learning_rate": 4.1620111731843575e-06, "loss": 0.6629, "step": 298 }, { "epoch": 0.01253366588767907, "grad_norm": 2.465421438217163, "learning_rate": 4.175977653631285e-06, "loss": 0.677, "step": 299 }, { "epoch": 0.01257558450268803, "grad_norm": 3.2878613471984863, "learning_rate": 4.189944134078213e-06, "loss": 0.6535, "step": 300 }, { "epoch": 0.012617503117696991, "grad_norm": 2.942570209503174, "learning_rate": 4.20391061452514e-06, "loss": 0.7607, "step": 301 }, { "epoch": 0.012659421732705952, "grad_norm": 2.4307117462158203, "learning_rate": 4.217877094972068e-06, "loss": 0.6877, "step": 302 }, { "epoch": 0.012701340347714911, "grad_norm": 2.6186230182647705, "learning_rate": 4.231843575418994e-06, "loss": 0.7264, "step": 303 }, { "epoch": 0.012743258962723871, "grad_norm": 2.6421711444854736, "learning_rate": 4.2458100558659224e-06, "loss": 0.7114, "step": 304 }, { "epoch": 0.012785177577732832, "grad_norm": 2.91774845123291, "learning_rate": 4.25977653631285e-06, "loss": 0.659, "step": 305 }, { "epoch": 0.012827096192741793, "grad_norm": 2.6977829933166504, "learning_rate": 4.273743016759777e-06, "loss": 0.6651, "step": 306 }, { "epoch": 0.012869014807750752, "grad_norm": 3.339097738265991, "learning_rate": 4.287709497206704e-06, "loss": 0.7236, "step": 307 }, { "epoch": 0.012910933422759712, "grad_norm": 2.486833333969116, "learning_rate": 4.301675977653632e-06, "loss": 0.6675, "step": 308 }, { "epoch": 0.012952852037768673, "grad_norm": 2.527785301208496, "learning_rate": 4.315642458100559e-06, "loss": 0.7616, "step": 309 }, { "epoch": 0.012994770652777632, "grad_norm": 4.481642723083496, "learning_rate": 4.3296089385474866e-06, "loss": 0.7252, "step": 310 }, { "epoch": 0.013036689267786592, "grad_norm": 2.8283214569091797, "learning_rate": 4.343575418994414e-06, "loss": 0.6831, "step": 311 }, { "epoch": 0.013078607882795553, "grad_norm": 2.742154121398926, "learning_rate": 4.357541899441341e-06, "loss": 0.7016, "step": 312 }, { "epoch": 0.013120526497804513, "grad_norm": 3.5507469177246094, "learning_rate": 4.371508379888269e-06, "loss": 0.7538, "step": 313 }, { "epoch": 0.013162445112813472, "grad_norm": 2.7475197315216064, "learning_rate": 4.385474860335196e-06, "loss": 0.6906, "step": 314 }, { "epoch": 0.013204363727822433, "grad_norm": 3.097378730773926, "learning_rate": 4.399441340782123e-06, "loss": 0.6593, "step": 315 }, { "epoch": 0.013246282342831393, "grad_norm": 2.8879802227020264, "learning_rate": 4.413407821229051e-06, "loss": 0.7261, "step": 316 }, { "epoch": 0.013288200957840352, "grad_norm": 3.024611234664917, "learning_rate": 4.427374301675978e-06, "loss": 0.735, "step": 317 }, { "epoch": 0.013330119572849313, "grad_norm": 3.1278154850006104, "learning_rate": 4.441340782122905e-06, "loss": 0.7312, "step": 318 }, { "epoch": 0.013372038187858273, "grad_norm": 5.5577192306518555, "learning_rate": 4.455307262569833e-06, "loss": 0.7353, "step": 319 }, { "epoch": 0.013413956802867234, "grad_norm": 2.8218960762023926, "learning_rate": 4.46927374301676e-06, "loss": 0.7405, "step": 320 }, { "epoch": 0.013455875417876193, "grad_norm": 2.28967022895813, "learning_rate": 4.4832402234636875e-06, "loss": 0.6548, "step": 321 }, { "epoch": 0.013497794032885153, "grad_norm": 2.6633613109588623, "learning_rate": 4.497206703910615e-06, "loss": 0.6388, "step": 322 }, { "epoch": 0.013539712647894114, "grad_norm": 2.5754964351654053, "learning_rate": 4.511173184357542e-06, "loss": 0.6904, "step": 323 }, { "epoch": 0.013581631262903073, "grad_norm": 2.6649813652038574, "learning_rate": 4.5251396648044695e-06, "loss": 0.6901, "step": 324 }, { "epoch": 0.013623549877912033, "grad_norm": 2.43988037109375, "learning_rate": 4.539106145251397e-06, "loss": 0.703, "step": 325 }, { "epoch": 0.013665468492920994, "grad_norm": 2.679094076156616, "learning_rate": 4.553072625698324e-06, "loss": 0.6435, "step": 326 }, { "epoch": 0.013707387107929955, "grad_norm": 2.8632240295410156, "learning_rate": 4.567039106145252e-06, "loss": 0.6892, "step": 327 }, { "epoch": 0.013749305722938913, "grad_norm": 3.1725411415100098, "learning_rate": 4.581005586592179e-06, "loss": 0.7357, "step": 328 }, { "epoch": 0.013791224337947874, "grad_norm": 2.260277032852173, "learning_rate": 4.594972067039106e-06, "loss": 0.674, "step": 329 }, { "epoch": 0.013833142952956835, "grad_norm": 3.465122938156128, "learning_rate": 4.608938547486034e-06, "loss": 0.7397, "step": 330 }, { "epoch": 0.013875061567965795, "grad_norm": 2.504032850265503, "learning_rate": 4.622905027932961e-06, "loss": 0.6651, "step": 331 }, { "epoch": 0.013916980182974754, "grad_norm": 2.6629602909088135, "learning_rate": 4.636871508379888e-06, "loss": 0.7048, "step": 332 }, { "epoch": 0.013958898797983715, "grad_norm": 2.5031206607818604, "learning_rate": 4.6508379888268166e-06, "loss": 0.6887, "step": 333 }, { "epoch": 0.014000817412992675, "grad_norm": 2.9903151988983154, "learning_rate": 4.664804469273743e-06, "loss": 0.7257, "step": 334 }, { "epoch": 0.014042736028001634, "grad_norm": 2.9478399753570557, "learning_rate": 4.67877094972067e-06, "loss": 0.7024, "step": 335 }, { "epoch": 0.014084654643010595, "grad_norm": 2.189058780670166, "learning_rate": 4.692737430167599e-06, "loss": 0.6491, "step": 336 }, { "epoch": 0.014126573258019555, "grad_norm": 3.0139029026031494, "learning_rate": 4.706703910614526e-06, "loss": 0.7686, "step": 337 }, { "epoch": 0.014168491873028516, "grad_norm": 3.8560128211975098, "learning_rate": 4.7206703910614525e-06, "loss": 0.6989, "step": 338 }, { "epoch": 0.014210410488037475, "grad_norm": 2.7407500743865967, "learning_rate": 4.73463687150838e-06, "loss": 0.7013, "step": 339 }, { "epoch": 0.014252329103046435, "grad_norm": 3.0531363487243652, "learning_rate": 4.748603351955308e-06, "loss": 0.6864, "step": 340 }, { "epoch": 0.014294247718055396, "grad_norm": 2.6591856479644775, "learning_rate": 4.762569832402235e-06, "loss": 0.7378, "step": 341 }, { "epoch": 0.014336166333064355, "grad_norm": 2.8588860034942627, "learning_rate": 4.776536312849163e-06, "loss": 0.7473, "step": 342 }, { "epoch": 0.014378084948073315, "grad_norm": 2.364124059677124, "learning_rate": 4.790502793296089e-06, "loss": 0.6533, "step": 343 }, { "epoch": 0.014420003563082276, "grad_norm": 4.2260637283325195, "learning_rate": 4.8044692737430175e-06, "loss": 0.6987, "step": 344 }, { "epoch": 0.014461922178091237, "grad_norm": 4.1112284660339355, "learning_rate": 4.818435754189945e-06, "loss": 0.6969, "step": 345 }, { "epoch": 0.014503840793100195, "grad_norm": 2.8002099990844727, "learning_rate": 4.832402234636872e-06, "loss": 0.7032, "step": 346 }, { "epoch": 0.014545759408109156, "grad_norm": 2.612104892730713, "learning_rate": 4.846368715083799e-06, "loss": 0.6785, "step": 347 }, { "epoch": 0.014587678023118117, "grad_norm": 2.6508865356445312, "learning_rate": 4.860335195530727e-06, "loss": 0.6879, "step": 348 }, { "epoch": 0.014629596638127077, "grad_norm": 2.5518290996551514, "learning_rate": 4.874301675977654e-06, "loss": 0.6703, "step": 349 }, { "epoch": 0.014671515253136036, "grad_norm": 3.3849527835845947, "learning_rate": 4.8882681564245816e-06, "loss": 0.7399, "step": 350 }, { "epoch": 0.014713433868144997, "grad_norm": 2.8031623363494873, "learning_rate": 4.902234636871509e-06, "loss": 0.6023, "step": 351 }, { "epoch": 0.014755352483153957, "grad_norm": 2.250138759613037, "learning_rate": 4.916201117318436e-06, "loss": 0.6386, "step": 352 }, { "epoch": 0.014797271098162916, "grad_norm": 4.354822158813477, "learning_rate": 4.930167597765364e-06, "loss": 0.659, "step": 353 }, { "epoch": 0.014839189713171877, "grad_norm": 2.7632617950439453, "learning_rate": 4.944134078212291e-06, "loss": 0.7507, "step": 354 }, { "epoch": 0.014881108328180837, "grad_norm": 2.602531671524048, "learning_rate": 4.958100558659218e-06, "loss": 0.6929, "step": 355 }, { "epoch": 0.014923026943189798, "grad_norm": 2.548358917236328, "learning_rate": 4.972067039106146e-06, "loss": 0.6822, "step": 356 }, { "epoch": 0.014964945558198757, "grad_norm": 2.584974527359009, "learning_rate": 4.986033519553073e-06, "loss": 0.6245, "step": 357 }, { "epoch": 0.015006864173207717, "grad_norm": 2.6537554264068604, "learning_rate": 5e-06, "loss": 0.6632, "step": 358 }, { "epoch": 0.015048782788216678, "grad_norm": 2.6608128547668457, "learning_rate": 5.013966480446928e-06, "loss": 0.7304, "step": 359 }, { "epoch": 0.015090701403225637, "grad_norm": 2.427600383758545, "learning_rate": 5.027932960893855e-06, "loss": 0.6971, "step": 360 }, { "epoch": 0.015132620018234597, "grad_norm": 2.5650017261505127, "learning_rate": 5.041899441340783e-06, "loss": 0.6566, "step": 361 }, { "epoch": 0.015174538633243558, "grad_norm": 2.561750888824463, "learning_rate": 5.055865921787711e-06, "loss": 0.6486, "step": 362 }, { "epoch": 0.015216457248252518, "grad_norm": 2.8175137042999268, "learning_rate": 5.069832402234637e-06, "loss": 0.6905, "step": 363 }, { "epoch": 0.015258375863261477, "grad_norm": 2.332737922668457, "learning_rate": 5.0837988826815645e-06, "loss": 0.69, "step": 364 }, { "epoch": 0.015300294478270438, "grad_norm": 2.814790725708008, "learning_rate": 5.097765363128492e-06, "loss": 0.6666, "step": 365 }, { "epoch": 0.015342213093279398, "grad_norm": 2.7276952266693115, "learning_rate": 5.111731843575419e-06, "loss": 0.7144, "step": 366 }, { "epoch": 0.015384131708288357, "grad_norm": 2.4987082481384277, "learning_rate": 5.125698324022347e-06, "loss": 0.6947, "step": 367 }, { "epoch": 0.015426050323297318, "grad_norm": 2.7696533203125, "learning_rate": 5.139664804469274e-06, "loss": 0.6333, "step": 368 }, { "epoch": 0.015467968938306279, "grad_norm": 2.889282464981079, "learning_rate": 5.153631284916202e-06, "loss": 0.7047, "step": 369 }, { "epoch": 0.015509887553315239, "grad_norm": 2.7080438137054443, "learning_rate": 5.1675977653631295e-06, "loss": 0.6957, "step": 370 }, { "epoch": 0.015551806168324198, "grad_norm": 2.8144917488098145, "learning_rate": 5.181564245810057e-06, "loss": 0.6509, "step": 371 }, { "epoch": 0.015593724783333159, "grad_norm": 2.4290153980255127, "learning_rate": 5.195530726256983e-06, "loss": 0.7266, "step": 372 }, { "epoch": 0.015635643398342117, "grad_norm": 2.3816773891448975, "learning_rate": 5.209497206703911e-06, "loss": 0.6152, "step": 373 }, { "epoch": 0.01567756201335108, "grad_norm": 3.150020122528076, "learning_rate": 5.223463687150838e-06, "loss": 0.7825, "step": 374 }, { "epoch": 0.01571948062836004, "grad_norm": 2.575658082962036, "learning_rate": 5.2374301675977654e-06, "loss": 0.6739, "step": 375 }, { "epoch": 0.015761399243368997, "grad_norm": 2.753528118133545, "learning_rate": 5.251396648044693e-06, "loss": 0.6385, "step": 376 }, { "epoch": 0.01580331785837796, "grad_norm": 2.6365485191345215, "learning_rate": 5.265363128491621e-06, "loss": 0.6632, "step": 377 }, { "epoch": 0.01584523647338692, "grad_norm": 2.6769771575927734, "learning_rate": 5.279329608938548e-06, "loss": 0.6615, "step": 378 }, { "epoch": 0.01588715508839588, "grad_norm": 2.855189800262451, "learning_rate": 5.293296089385476e-06, "loss": 0.6463, "step": 379 }, { "epoch": 0.01592907370340484, "grad_norm": 2.5988221168518066, "learning_rate": 5.307262569832403e-06, "loss": 0.7192, "step": 380 }, { "epoch": 0.0159709923184138, "grad_norm": 2.5702664852142334, "learning_rate": 5.3212290502793296e-06, "loss": 0.7416, "step": 381 }, { "epoch": 0.01601291093342276, "grad_norm": 2.2259557247161865, "learning_rate": 5.335195530726257e-06, "loss": 0.6202, "step": 382 }, { "epoch": 0.01605482954843172, "grad_norm": 2.611996650695801, "learning_rate": 5.349162011173184e-06, "loss": 0.6727, "step": 383 }, { "epoch": 0.01609674816344068, "grad_norm": 3.738086700439453, "learning_rate": 5.3631284916201125e-06, "loss": 0.6891, "step": 384 }, { "epoch": 0.01613866677844964, "grad_norm": 2.6065804958343506, "learning_rate": 5.37709497206704e-06, "loss": 0.6889, "step": 385 }, { "epoch": 0.0161805853934586, "grad_norm": 2.6075551509857178, "learning_rate": 5.391061452513967e-06, "loss": 0.7169, "step": 386 }, { "epoch": 0.01622250400846756, "grad_norm": 2.5495686531066895, "learning_rate": 5.4050279329608945e-06, "loss": 0.7083, "step": 387 }, { "epoch": 0.01626442262347652, "grad_norm": 2.6504456996917725, "learning_rate": 5.418994413407822e-06, "loss": 0.5785, "step": 388 }, { "epoch": 0.01630634123848548, "grad_norm": 2.9117431640625, "learning_rate": 5.432960893854749e-06, "loss": 0.6725, "step": 389 }, { "epoch": 0.016348259853494442, "grad_norm": 4.445065975189209, "learning_rate": 5.446927374301676e-06, "loss": 0.6897, "step": 390 }, { "epoch": 0.0163901784685034, "grad_norm": 2.7848117351531982, "learning_rate": 5.460893854748603e-06, "loss": 0.6449, "step": 391 }, { "epoch": 0.01643209708351236, "grad_norm": 2.5062150955200195, "learning_rate": 5.474860335195531e-06, "loss": 0.7132, "step": 392 }, { "epoch": 0.016474015698521322, "grad_norm": 2.5208873748779297, "learning_rate": 5.488826815642459e-06, "loss": 0.6922, "step": 393 }, { "epoch": 0.01651593431353028, "grad_norm": 2.3664815425872803, "learning_rate": 5.502793296089386e-06, "loss": 0.7202, "step": 394 }, { "epoch": 0.01655785292853924, "grad_norm": 2.6732680797576904, "learning_rate": 5.516759776536313e-06, "loss": 0.6719, "step": 395 }, { "epoch": 0.016599771543548202, "grad_norm": 2.5037856101989746, "learning_rate": 5.530726256983241e-06, "loss": 0.685, "step": 396 }, { "epoch": 0.01664169015855716, "grad_norm": 2.707822322845459, "learning_rate": 5.544692737430168e-06, "loss": 0.7035, "step": 397 }, { "epoch": 0.01668360877356612, "grad_norm": 2.6246120929718018, "learning_rate": 5.558659217877096e-06, "loss": 0.6835, "step": 398 }, { "epoch": 0.016725527388575082, "grad_norm": 2.231182098388672, "learning_rate": 5.572625698324022e-06, "loss": 0.6547, "step": 399 }, { "epoch": 0.01676744600358404, "grad_norm": 2.8230369091033936, "learning_rate": 5.58659217877095e-06, "loss": 0.7142, "step": 400 }, { "epoch": 0.016809364618593, "grad_norm": 2.784086227416992, "learning_rate": 5.6005586592178775e-06, "loss": 0.703, "step": 401 }, { "epoch": 0.016851283233601962, "grad_norm": 2.3837430477142334, "learning_rate": 5.614525139664805e-06, "loss": 0.7352, "step": 402 }, { "epoch": 0.01689320184861092, "grad_norm": 2.4660985469818115, "learning_rate": 5.628491620111732e-06, "loss": 0.5892, "step": 403 }, { "epoch": 0.016935120463619883, "grad_norm": 2.9437410831451416, "learning_rate": 5.6424581005586595e-06, "loss": 0.7003, "step": 404 }, { "epoch": 0.016977039078628842, "grad_norm": 2.8202409744262695, "learning_rate": 5.656424581005588e-06, "loss": 0.6934, "step": 405 }, { "epoch": 0.0170189576936378, "grad_norm": 2.4496681690216064, "learning_rate": 5.670391061452515e-06, "loss": 0.6745, "step": 406 }, { "epoch": 0.017060876308646764, "grad_norm": 2.893758535385132, "learning_rate": 5.6843575418994425e-06, "loss": 0.6693, "step": 407 }, { "epoch": 0.017102794923655722, "grad_norm": 2.626572370529175, "learning_rate": 5.698324022346369e-06, "loss": 0.7112, "step": 408 }, { "epoch": 0.01714471353866468, "grad_norm": 2.985039472579956, "learning_rate": 5.712290502793296e-06, "loss": 0.6873, "step": 409 }, { "epoch": 0.017186632153673644, "grad_norm": 2.744976043701172, "learning_rate": 5.726256983240224e-06, "loss": 0.6747, "step": 410 }, { "epoch": 0.017228550768682602, "grad_norm": 3.4895448684692383, "learning_rate": 5.740223463687151e-06, "loss": 0.7154, "step": 411 }, { "epoch": 0.01727046938369156, "grad_norm": 2.5664021968841553, "learning_rate": 5.754189944134078e-06, "loss": 0.6383, "step": 412 }, { "epoch": 0.017312387998700524, "grad_norm": 2.9696762561798096, "learning_rate": 5.768156424581007e-06, "loss": 0.6187, "step": 413 }, { "epoch": 0.017354306613709482, "grad_norm": 2.6104531288146973, "learning_rate": 5.782122905027934e-06, "loss": 0.6526, "step": 414 }, { "epoch": 0.017396225228718445, "grad_norm": 2.3787293434143066, "learning_rate": 5.796089385474861e-06, "loss": 0.6574, "step": 415 }, { "epoch": 0.017438143843727404, "grad_norm": 4.052406311035156, "learning_rate": 5.810055865921789e-06, "loss": 0.6312, "step": 416 }, { "epoch": 0.017480062458736362, "grad_norm": 2.731464147567749, "learning_rate": 5.824022346368715e-06, "loss": 0.7272, "step": 417 }, { "epoch": 0.017521981073745325, "grad_norm": 2.4729630947113037, "learning_rate": 5.8379888268156425e-06, "loss": 0.6308, "step": 418 }, { "epoch": 0.017563899688754284, "grad_norm": 2.6406426429748535, "learning_rate": 5.85195530726257e-06, "loss": 0.6913, "step": 419 }, { "epoch": 0.017605818303763242, "grad_norm": 2.585611581802368, "learning_rate": 5.865921787709497e-06, "loss": 0.7102, "step": 420 }, { "epoch": 0.017647736918772205, "grad_norm": 2.3380510807037354, "learning_rate": 5.879888268156425e-06, "loss": 0.6388, "step": 421 }, { "epoch": 0.017689655533781164, "grad_norm": 2.6468875408172607, "learning_rate": 5.893854748603353e-06, "loss": 0.7767, "step": 422 }, { "epoch": 0.017731574148790123, "grad_norm": 2.614097833633423, "learning_rate": 5.90782122905028e-06, "loss": 0.649, "step": 423 }, { "epoch": 0.017773492763799085, "grad_norm": 2.3821287155151367, "learning_rate": 5.9217877094972075e-06, "loss": 0.6816, "step": 424 }, { "epoch": 0.017815411378808044, "grad_norm": 2.4621341228485107, "learning_rate": 5.935754189944135e-06, "loss": 0.6799, "step": 425 }, { "epoch": 0.017857329993817003, "grad_norm": 2.655632257461548, "learning_rate": 5.949720670391061e-06, "loss": 0.6392, "step": 426 }, { "epoch": 0.017899248608825965, "grad_norm": 2.5188260078430176, "learning_rate": 5.963687150837989e-06, "loss": 0.6855, "step": 427 }, { "epoch": 0.017941167223834924, "grad_norm": 2.735036611557007, "learning_rate": 5.977653631284917e-06, "loss": 0.6512, "step": 428 }, { "epoch": 0.017983085838843886, "grad_norm": 2.4352264404296875, "learning_rate": 5.991620111731844e-06, "loss": 0.6902, "step": 429 }, { "epoch": 0.018025004453852845, "grad_norm": 2.429781436920166, "learning_rate": 6.005586592178772e-06, "loss": 0.6601, "step": 430 }, { "epoch": 0.018066923068861804, "grad_norm": 2.2725296020507812, "learning_rate": 6.019553072625699e-06, "loss": 0.6641, "step": 431 }, { "epoch": 0.018108841683870766, "grad_norm": 2.323840618133545, "learning_rate": 6.033519553072626e-06, "loss": 0.6197, "step": 432 }, { "epoch": 0.018150760298879725, "grad_norm": 4.280886173248291, "learning_rate": 6.047486033519554e-06, "loss": 0.6783, "step": 433 }, { "epoch": 0.018192678913888684, "grad_norm": 4.413292407989502, "learning_rate": 6.061452513966482e-06, "loss": 0.673, "step": 434 }, { "epoch": 0.018234597528897646, "grad_norm": 2.440016031265259, "learning_rate": 6.0754189944134075e-06, "loss": 0.7042, "step": 435 }, { "epoch": 0.018276516143906605, "grad_norm": 2.786083221435547, "learning_rate": 6.089385474860336e-06, "loss": 0.6772, "step": 436 }, { "epoch": 0.018318434758915564, "grad_norm": 2.5190529823303223, "learning_rate": 6.103351955307263e-06, "loss": 0.6531, "step": 437 }, { "epoch": 0.018360353373924526, "grad_norm": 2.94730806350708, "learning_rate": 6.1173184357541904e-06, "loss": 0.7501, "step": 438 }, { "epoch": 0.018402271988933485, "grad_norm": 2.471013069152832, "learning_rate": 6.131284916201118e-06, "loss": 0.7138, "step": 439 }, { "epoch": 0.018444190603942447, "grad_norm": 2.405035972595215, "learning_rate": 6.145251396648045e-06, "loss": 0.6584, "step": 440 }, { "epoch": 0.018486109218951406, "grad_norm": 2.173051118850708, "learning_rate": 6.1592178770949725e-06, "loss": 0.6972, "step": 441 }, { "epoch": 0.018528027833960365, "grad_norm": 2.629324197769165, "learning_rate": 6.173184357541901e-06, "loss": 0.7494, "step": 442 }, { "epoch": 0.018569946448969327, "grad_norm": 2.6673340797424316, "learning_rate": 6.187150837988828e-06, "loss": 0.5957, "step": 443 }, { "epoch": 0.018611865063978286, "grad_norm": 2.618642568588257, "learning_rate": 6.2011173184357546e-06, "loss": 0.6597, "step": 444 }, { "epoch": 0.018653783678987245, "grad_norm": 3.0562009811401367, "learning_rate": 6.215083798882682e-06, "loss": 0.6597, "step": 445 }, { "epoch": 0.018695702293996207, "grad_norm": 3.3780555725097656, "learning_rate": 6.229050279329609e-06, "loss": 0.5884, "step": 446 }, { "epoch": 0.018737620909005166, "grad_norm": 2.284836530685425, "learning_rate": 6.243016759776537e-06, "loss": 0.6891, "step": 447 }, { "epoch": 0.018779539524014125, "grad_norm": 2.9965620040893555, "learning_rate": 6.256983240223464e-06, "loss": 0.7057, "step": 448 }, { "epoch": 0.018821458139023087, "grad_norm": 2.5688085556030273, "learning_rate": 6.270949720670392e-06, "loss": 0.6439, "step": 449 }, { "epoch": 0.018863376754032046, "grad_norm": 2.4804508686065674, "learning_rate": 6.2849162011173195e-06, "loss": 0.6679, "step": 450 }, { "epoch": 0.01890529536904101, "grad_norm": 2.287537097930908, "learning_rate": 6.298882681564247e-06, "loss": 0.6361, "step": 451 }, { "epoch": 0.018947213984049967, "grad_norm": 2.4971697330474854, "learning_rate": 6.312849162011173e-06, "loss": 0.6315, "step": 452 }, { "epoch": 0.018989132599058926, "grad_norm": 2.558323860168457, "learning_rate": 6.326815642458101e-06, "loss": 0.6538, "step": 453 }, { "epoch": 0.01903105121406789, "grad_norm": 3.106698989868164, "learning_rate": 6.340782122905028e-06, "loss": 0.6597, "step": 454 }, { "epoch": 0.019072969829076847, "grad_norm": 2.3365285396575928, "learning_rate": 6.3547486033519555e-06, "loss": 0.6473, "step": 455 }, { "epoch": 0.019114888444085806, "grad_norm": 2.492600202560425, "learning_rate": 6.368715083798883e-06, "loss": 0.7194, "step": 456 }, { "epoch": 0.01915680705909477, "grad_norm": 2.1446070671081543, "learning_rate": 6.382681564245811e-06, "loss": 0.6955, "step": 457 }, { "epoch": 0.019198725674103728, "grad_norm": 2.553858518600464, "learning_rate": 6.396648044692738e-06, "loss": 0.6142, "step": 458 }, { "epoch": 0.019240644289112686, "grad_norm": 2.1677637100219727, "learning_rate": 6.410614525139666e-06, "loss": 0.6993, "step": 459 }, { "epoch": 0.01928256290412165, "grad_norm": 2.6346821784973145, "learning_rate": 6.424581005586593e-06, "loss": 0.6955, "step": 460 }, { "epoch": 0.019324481519130608, "grad_norm": 2.6257708072662354, "learning_rate": 6.43854748603352e-06, "loss": 0.6415, "step": 461 }, { "epoch": 0.019366400134139566, "grad_norm": 2.5083322525024414, "learning_rate": 6.452513966480447e-06, "loss": 0.7208, "step": 462 }, { "epoch": 0.01940831874914853, "grad_norm": 2.6328489780426025, "learning_rate": 6.466480446927374e-06, "loss": 0.7343, "step": 463 }, { "epoch": 0.019450237364157488, "grad_norm": 2.7963204383850098, "learning_rate": 6.480446927374302e-06, "loss": 0.6178, "step": 464 }, { "epoch": 0.01949215597916645, "grad_norm": 3.3501534461975098, "learning_rate": 6.49441340782123e-06, "loss": 0.6847, "step": 465 }, { "epoch": 0.01953407459417541, "grad_norm": 2.721029281616211, "learning_rate": 6.508379888268157e-06, "loss": 0.6233, "step": 466 }, { "epoch": 0.019575993209184368, "grad_norm": 2.452599048614502, "learning_rate": 6.5223463687150846e-06, "loss": 0.6567, "step": 467 }, { "epoch": 0.01961791182419333, "grad_norm": 2.35260272026062, "learning_rate": 6.536312849162012e-06, "loss": 0.6073, "step": 468 }, { "epoch": 0.01965983043920229, "grad_norm": 2.5046584606170654, "learning_rate": 6.550279329608939e-06, "loss": 0.5776, "step": 469 }, { "epoch": 0.019701749054211248, "grad_norm": 2.6965157985687256, "learning_rate": 6.564245810055866e-06, "loss": 0.7187, "step": 470 }, { "epoch": 0.01974366766922021, "grad_norm": 2.376108407974243, "learning_rate": 6.578212290502793e-06, "loss": 0.6956, "step": 471 }, { "epoch": 0.01978558628422917, "grad_norm": 2.7525837421417236, "learning_rate": 6.592178770949721e-06, "loss": 0.6843, "step": 472 }, { "epoch": 0.019827504899238128, "grad_norm": 3.7484428882598877, "learning_rate": 6.606145251396649e-06, "loss": 0.6033, "step": 473 }, { "epoch": 0.01986942351424709, "grad_norm": 3.784064292907715, "learning_rate": 6.620111731843576e-06, "loss": 0.6459, "step": 474 }, { "epoch": 0.01991134212925605, "grad_norm": 2.867446184158325, "learning_rate": 6.634078212290503e-06, "loss": 0.6614, "step": 475 }, { "epoch": 0.01995326074426501, "grad_norm": 2.92148756980896, "learning_rate": 6.648044692737431e-06, "loss": 0.665, "step": 476 }, { "epoch": 0.01999517935927397, "grad_norm": 2.326934337615967, "learning_rate": 6.662011173184358e-06, "loss": 0.6708, "step": 477 }, { "epoch": 0.02003709797428293, "grad_norm": 2.836622953414917, "learning_rate": 6.675977653631286e-06, "loss": 0.7029, "step": 478 }, { "epoch": 0.02007901658929189, "grad_norm": 2.948094129562378, "learning_rate": 6.689944134078212e-06, "loss": 0.7326, "step": 479 }, { "epoch": 0.02012093520430085, "grad_norm": 2.755100965499878, "learning_rate": 6.70391061452514e-06, "loss": 0.7085, "step": 480 }, { "epoch": 0.02016285381930981, "grad_norm": 2.4562394618988037, "learning_rate": 6.7178770949720675e-06, "loss": 0.6153, "step": 481 }, { "epoch": 0.02020477243431877, "grad_norm": 2.301795721054077, "learning_rate": 6.731843575418995e-06, "loss": 0.6307, "step": 482 }, { "epoch": 0.02024669104932773, "grad_norm": 2.928086519241333, "learning_rate": 6.745810055865922e-06, "loss": 0.706, "step": 483 }, { "epoch": 0.02028860966433669, "grad_norm": 2.6611623764038086, "learning_rate": 6.7597765363128496e-06, "loss": 0.6155, "step": 484 }, { "epoch": 0.02033052827934565, "grad_norm": 2.544773578643799, "learning_rate": 6.773743016759777e-06, "loss": 0.6512, "step": 485 }, { "epoch": 0.02037244689435461, "grad_norm": 2.637178659439087, "learning_rate": 6.787709497206705e-06, "loss": 0.6212, "step": 486 }, { "epoch": 0.02041436550936357, "grad_norm": 2.5372962951660156, "learning_rate": 6.8016759776536325e-06, "loss": 0.6527, "step": 487 }, { "epoch": 0.02045628412437253, "grad_norm": 2.680229425430298, "learning_rate": 6.815642458100559e-06, "loss": 0.6472, "step": 488 }, { "epoch": 0.02049820273938149, "grad_norm": 2.714492082595825, "learning_rate": 6.829608938547486e-06, "loss": 0.704, "step": 489 }, { "epoch": 0.020540121354390452, "grad_norm": 2.3450028896331787, "learning_rate": 6.843575418994414e-06, "loss": 0.6346, "step": 490 }, { "epoch": 0.02058203996939941, "grad_norm": 2.889641284942627, "learning_rate": 6.857541899441341e-06, "loss": 0.6161, "step": 491 }, { "epoch": 0.02062395858440837, "grad_norm": 3.3372135162353516, "learning_rate": 6.871508379888268e-06, "loss": 0.6799, "step": 492 }, { "epoch": 0.020665877199417333, "grad_norm": 2.5409061908721924, "learning_rate": 6.885474860335197e-06, "loss": 0.7192, "step": 493 }, { "epoch": 0.02070779581442629, "grad_norm": 2.417322874069214, "learning_rate": 6.899441340782124e-06, "loss": 0.6582, "step": 494 }, { "epoch": 0.02074971442943525, "grad_norm": 2.5529394149780273, "learning_rate": 6.913407821229051e-06, "loss": 0.7645, "step": 495 }, { "epoch": 0.020791633044444213, "grad_norm": 2.045924186706543, "learning_rate": 6.927374301675979e-06, "loss": 0.6693, "step": 496 }, { "epoch": 0.02083355165945317, "grad_norm": 2.3331449031829834, "learning_rate": 6.941340782122905e-06, "loss": 0.626, "step": 497 }, { "epoch": 0.02087547027446213, "grad_norm": 6.668644905090332, "learning_rate": 6.9553072625698325e-06, "loss": 0.6667, "step": 498 }, { "epoch": 0.020917388889471093, "grad_norm": 2.604750394821167, "learning_rate": 6.96927374301676e-06, "loss": 0.6897, "step": 499 }, { "epoch": 0.02095930750448005, "grad_norm": 2.945009469985962, "learning_rate": 6.983240223463687e-06, "loss": 0.6091, "step": 500 }, { "epoch": 0.021001226119489014, "grad_norm": 3.357870101928711, "learning_rate": 6.9972067039106154e-06, "loss": 0.67, "step": 501 }, { "epoch": 0.021043144734497973, "grad_norm": 2.6712093353271484, "learning_rate": 7.011173184357543e-06, "loss": 0.7224, "step": 502 }, { "epoch": 0.02108506334950693, "grad_norm": 2.281604528427124, "learning_rate": 7.02513966480447e-06, "loss": 0.6903, "step": 503 }, { "epoch": 0.021126981964515894, "grad_norm": 2.699956178665161, "learning_rate": 7.0391061452513975e-06, "loss": 0.6741, "step": 504 }, { "epoch": 0.021168900579524853, "grad_norm": 2.2988040447235107, "learning_rate": 7.053072625698325e-06, "loss": 0.6664, "step": 505 }, { "epoch": 0.02121081919453381, "grad_norm": 2.154871702194214, "learning_rate": 7.067039106145251e-06, "loss": 0.6583, "step": 506 }, { "epoch": 0.021252737809542774, "grad_norm": 2.3224594593048096, "learning_rate": 7.081005586592179e-06, "loss": 0.6096, "step": 507 }, { "epoch": 0.021294656424551733, "grad_norm": 2.204451560974121, "learning_rate": 7.094972067039106e-06, "loss": 0.5885, "step": 508 }, { "epoch": 0.02133657503956069, "grad_norm": 2.5341761112213135, "learning_rate": 7.108938547486034e-06, "loss": 0.6668, "step": 509 }, { "epoch": 0.021378493654569654, "grad_norm": 2.5178170204162598, "learning_rate": 7.122905027932962e-06, "loss": 0.5709, "step": 510 }, { "epoch": 0.021420412269578613, "grad_norm": 2.894789218902588, "learning_rate": 7.136871508379889e-06, "loss": 0.631, "step": 511 }, { "epoch": 0.02146233088458757, "grad_norm": 3.123939275741577, "learning_rate": 7.150837988826816e-06, "loss": 0.6602, "step": 512 }, { "epoch": 0.021504249499596534, "grad_norm": 2.2762420177459717, "learning_rate": 7.164804469273744e-06, "loss": 0.6456, "step": 513 }, { "epoch": 0.021546168114605493, "grad_norm": 2.521265745162964, "learning_rate": 7.178770949720671e-06, "loss": 0.6601, "step": 514 }, { "epoch": 0.021588086729614455, "grad_norm": 2.475508213043213, "learning_rate": 7.1927374301675976e-06, "loss": 0.6187, "step": 515 }, { "epoch": 0.021630005344623414, "grad_norm": 2.666090250015259, "learning_rate": 7.206703910614526e-06, "loss": 0.6687, "step": 516 }, { "epoch": 0.021671923959632373, "grad_norm": 2.1866204738616943, "learning_rate": 7.220670391061453e-06, "loss": 0.7349, "step": 517 }, { "epoch": 0.021713842574641335, "grad_norm": 2.3896963596343994, "learning_rate": 7.2346368715083805e-06, "loss": 0.7272, "step": 518 }, { "epoch": 0.021755761189650294, "grad_norm": 2.4141321182250977, "learning_rate": 7.248603351955308e-06, "loss": 0.641, "step": 519 }, { "epoch": 0.021797679804659253, "grad_norm": 2.722669839859009, "learning_rate": 7.262569832402235e-06, "loss": 0.6538, "step": 520 }, { "epoch": 0.021839598419668215, "grad_norm": 2.400012493133545, "learning_rate": 7.2765363128491625e-06, "loss": 0.7032, "step": 521 }, { "epoch": 0.021881517034677174, "grad_norm": 2.5579464435577393, "learning_rate": 7.290502793296091e-06, "loss": 0.6709, "step": 522 }, { "epoch": 0.021923435649686133, "grad_norm": 4.950404644012451, "learning_rate": 7.304469273743018e-06, "loss": 0.7094, "step": 523 }, { "epoch": 0.021965354264695095, "grad_norm": 3.55713152885437, "learning_rate": 7.318435754189945e-06, "loss": 0.6979, "step": 524 }, { "epoch": 0.022007272879704054, "grad_norm": 2.3400328159332275, "learning_rate": 7.332402234636872e-06, "loss": 0.6973, "step": 525 }, { "epoch": 0.022049191494713016, "grad_norm": 2.1587042808532715, "learning_rate": 7.346368715083799e-06, "loss": 0.6596, "step": 526 }, { "epoch": 0.022091110109721975, "grad_norm": 2.2695236206054688, "learning_rate": 7.360335195530727e-06, "loss": 0.6707, "step": 527 }, { "epoch": 0.022133028724730934, "grad_norm": 2.1349868774414062, "learning_rate": 7.374301675977654e-06, "loss": 0.609, "step": 528 }, { "epoch": 0.022174947339739896, "grad_norm": 2.8137712478637695, "learning_rate": 7.388268156424581e-06, "loss": 0.6983, "step": 529 }, { "epoch": 0.022216865954748855, "grad_norm": 2.269854784011841, "learning_rate": 7.4022346368715096e-06, "loss": 0.6497, "step": 530 }, { "epoch": 0.022258784569757814, "grad_norm": 3.2189855575561523, "learning_rate": 7.416201117318437e-06, "loss": 0.6615, "step": 531 }, { "epoch": 0.022300703184766776, "grad_norm": 6.72660493850708, "learning_rate": 7.430167597765364e-06, "loss": 0.6664, "step": 532 }, { "epoch": 0.022342621799775735, "grad_norm": 2.0794622898101807, "learning_rate": 7.444134078212291e-06, "loss": 0.5903, "step": 533 }, { "epoch": 0.022384540414784694, "grad_norm": 2.545872449874878, "learning_rate": 7.458100558659218e-06, "loss": 0.6856, "step": 534 }, { "epoch": 0.022426459029793656, "grad_norm": 2.622898817062378, "learning_rate": 7.4720670391061455e-06, "loss": 0.6882, "step": 535 }, { "epoch": 0.022468377644802615, "grad_norm": 2.356590509414673, "learning_rate": 7.486033519553073e-06, "loss": 0.6257, "step": 536 }, { "epoch": 0.022510296259811574, "grad_norm": 2.586193323135376, "learning_rate": 7.500000000000001e-06, "loss": 0.7233, "step": 537 }, { "epoch": 0.022552214874820536, "grad_norm": 2.4106850624084473, "learning_rate": 7.513966480446928e-06, "loss": 0.6595, "step": 538 }, { "epoch": 0.022594133489829495, "grad_norm": 3.877833127975464, "learning_rate": 7.527932960893856e-06, "loss": 0.6817, "step": 539 }, { "epoch": 0.022636052104838458, "grad_norm": 2.527883768081665, "learning_rate": 7.541899441340783e-06, "loss": 0.7446, "step": 540 }, { "epoch": 0.022677970719847416, "grad_norm": 2.9128737449645996, "learning_rate": 7.5558659217877105e-06, "loss": 0.6693, "step": 541 }, { "epoch": 0.022719889334856375, "grad_norm": 3.008596897125244, "learning_rate": 7.569832402234637e-06, "loss": 0.658, "step": 542 }, { "epoch": 0.022761807949865338, "grad_norm": 2.568277597427368, "learning_rate": 7.583798882681564e-06, "loss": 0.6942, "step": 543 }, { "epoch": 0.022803726564874297, "grad_norm": 2.3093831539154053, "learning_rate": 7.597765363128492e-06, "loss": 0.6086, "step": 544 }, { "epoch": 0.022845645179883255, "grad_norm": 2.4647295475006104, "learning_rate": 7.61173184357542e-06, "loss": 0.6684, "step": 545 }, { "epoch": 0.022887563794892218, "grad_norm": 2.123410224914551, "learning_rate": 7.625698324022347e-06, "loss": 0.6324, "step": 546 }, { "epoch": 0.022929482409901177, "grad_norm": 2.2342522144317627, "learning_rate": 7.639664804469275e-06, "loss": 0.6904, "step": 547 }, { "epoch": 0.022971401024910135, "grad_norm": 2.9604039192199707, "learning_rate": 7.653631284916202e-06, "loss": 0.7172, "step": 548 }, { "epoch": 0.023013319639919098, "grad_norm": 2.5419046878814697, "learning_rate": 7.66759776536313e-06, "loss": 0.6141, "step": 549 }, { "epoch": 0.023055238254928057, "grad_norm": 2.4476802349090576, "learning_rate": 7.681564245810057e-06, "loss": 0.7047, "step": 550 }, { "epoch": 0.02309715686993702, "grad_norm": 2.3174829483032227, "learning_rate": 7.695530726256984e-06, "loss": 0.6696, "step": 551 }, { "epoch": 0.023139075484945978, "grad_norm": 4.127678871154785, "learning_rate": 7.709497206703911e-06, "loss": 0.6379, "step": 552 }, { "epoch": 0.023180994099954937, "grad_norm": 2.3269200325012207, "learning_rate": 7.723463687150839e-06, "loss": 0.6794, "step": 553 }, { "epoch": 0.0232229127149639, "grad_norm": 3.0065884590148926, "learning_rate": 7.737430167597766e-06, "loss": 0.573, "step": 554 }, { "epoch": 0.023264831329972858, "grad_norm": 2.4457650184631348, "learning_rate": 7.751396648044693e-06, "loss": 0.6396, "step": 555 }, { "epoch": 0.023306749944981817, "grad_norm": 2.391122817993164, "learning_rate": 7.76536312849162e-06, "loss": 0.5956, "step": 556 }, { "epoch": 0.02334866855999078, "grad_norm": 2.497328281402588, "learning_rate": 7.779329608938548e-06, "loss": 0.6707, "step": 557 }, { "epoch": 0.023390587174999738, "grad_norm": 2.543797492980957, "learning_rate": 7.793296089385475e-06, "loss": 0.6299, "step": 558 }, { "epoch": 0.023432505790008697, "grad_norm": 3.0376203060150146, "learning_rate": 7.807262569832403e-06, "loss": 0.6567, "step": 559 }, { "epoch": 0.02347442440501766, "grad_norm": 2.291801929473877, "learning_rate": 7.82122905027933e-06, "loss": 0.6814, "step": 560 }, { "epoch": 0.023516343020026618, "grad_norm": 2.1596291065216064, "learning_rate": 7.835195530726258e-06, "loss": 0.6773, "step": 561 }, { "epoch": 0.02355826163503558, "grad_norm": 2.7445342540740967, "learning_rate": 7.849162011173185e-06, "loss": 0.7398, "step": 562 }, { "epoch": 0.02360018025004454, "grad_norm": 3.7520413398742676, "learning_rate": 7.863128491620112e-06, "loss": 0.6464, "step": 563 }, { "epoch": 0.023642098865053498, "grad_norm": 2.368675708770752, "learning_rate": 7.87709497206704e-06, "loss": 0.6403, "step": 564 }, { "epoch": 0.02368401748006246, "grad_norm": 2.5533103942871094, "learning_rate": 7.891061452513967e-06, "loss": 0.6965, "step": 565 }, { "epoch": 0.02372593609507142, "grad_norm": 2.343559503555298, "learning_rate": 7.905027932960894e-06, "loss": 0.6517, "step": 566 }, { "epoch": 0.023767854710080378, "grad_norm": 2.2151381969451904, "learning_rate": 7.918994413407822e-06, "loss": 0.6231, "step": 567 }, { "epoch": 0.02380977332508934, "grad_norm": 2.6884868144989014, "learning_rate": 7.932960893854749e-06, "loss": 0.7155, "step": 568 }, { "epoch": 0.0238516919400983, "grad_norm": 2.3405961990356445, "learning_rate": 7.946927374301676e-06, "loss": 0.6762, "step": 569 }, { "epoch": 0.023893610555107258, "grad_norm": 2.6449568271636963, "learning_rate": 7.960893854748604e-06, "loss": 0.6463, "step": 570 }, { "epoch": 0.02393552917011622, "grad_norm": 2.328225612640381, "learning_rate": 7.974860335195531e-06, "loss": 0.6053, "step": 571 }, { "epoch": 0.02397744778512518, "grad_norm": 2.4250075817108154, "learning_rate": 7.988826815642458e-06, "loss": 0.612, "step": 572 }, { "epoch": 0.024019366400134138, "grad_norm": 2.3136439323425293, "learning_rate": 8.002793296089386e-06, "loss": 0.6167, "step": 573 }, { "epoch": 0.0240612850151431, "grad_norm": 2.605642795562744, "learning_rate": 8.016759776536313e-06, "loss": 0.6024, "step": 574 }, { "epoch": 0.02410320363015206, "grad_norm": 2.554685354232788, "learning_rate": 8.03072625698324e-06, "loss": 0.644, "step": 575 }, { "epoch": 0.02414512224516102, "grad_norm": 2.562774896621704, "learning_rate": 8.044692737430168e-06, "loss": 0.689, "step": 576 }, { "epoch": 0.02418704086016998, "grad_norm": 2.1206629276275635, "learning_rate": 8.058659217877097e-06, "loss": 0.6602, "step": 577 }, { "epoch": 0.02422895947517894, "grad_norm": 2.676403284072876, "learning_rate": 8.072625698324023e-06, "loss": 0.6887, "step": 578 }, { "epoch": 0.0242708780901879, "grad_norm": 2.280531406402588, "learning_rate": 8.08659217877095e-06, "loss": 0.6149, "step": 579 }, { "epoch": 0.02431279670519686, "grad_norm": 2.5814526081085205, "learning_rate": 8.100558659217877e-06, "loss": 0.64, "step": 580 }, { "epoch": 0.02435471532020582, "grad_norm": 2.306455135345459, "learning_rate": 8.114525139664805e-06, "loss": 0.6384, "step": 581 }, { "epoch": 0.02439663393521478, "grad_norm": 2.36468243598938, "learning_rate": 8.128491620111732e-06, "loss": 0.6947, "step": 582 }, { "epoch": 0.02443855255022374, "grad_norm": 2.75958251953125, "learning_rate": 8.14245810055866e-06, "loss": 0.6529, "step": 583 }, { "epoch": 0.0244804711652327, "grad_norm": 2.912837505340576, "learning_rate": 8.156424581005588e-06, "loss": 0.6813, "step": 584 }, { "epoch": 0.02452238978024166, "grad_norm": 2.515711545944214, "learning_rate": 8.170391061452516e-06, "loss": 0.6542, "step": 585 }, { "epoch": 0.02456430839525062, "grad_norm": 2.0811500549316406, "learning_rate": 8.184357541899443e-06, "loss": 0.6151, "step": 586 }, { "epoch": 0.024606227010259583, "grad_norm": 2.4596381187438965, "learning_rate": 8.198324022346369e-06, "loss": 0.67, "step": 587 }, { "epoch": 0.02464814562526854, "grad_norm": 2.574840545654297, "learning_rate": 8.212290502793296e-06, "loss": 0.5454, "step": 588 }, { "epoch": 0.0246900642402775, "grad_norm": 2.358583688735962, "learning_rate": 8.226256983240223e-06, "loss": 0.6415, "step": 589 }, { "epoch": 0.024731982855286463, "grad_norm": 2.2676570415496826, "learning_rate": 8.24022346368715e-06, "loss": 0.6612, "step": 590 }, { "epoch": 0.02477390147029542, "grad_norm": 2.2409465312957764, "learning_rate": 8.254189944134078e-06, "loss": 0.6206, "step": 591 }, { "epoch": 0.02481582008530438, "grad_norm": 2.664630174636841, "learning_rate": 8.268156424581007e-06, "loss": 0.6847, "step": 592 }, { "epoch": 0.024857738700313343, "grad_norm": 2.45497727394104, "learning_rate": 8.282122905027935e-06, "loss": 0.6355, "step": 593 }, { "epoch": 0.0248996573153223, "grad_norm": 3.5538523197174072, "learning_rate": 8.296089385474862e-06, "loss": 0.6702, "step": 594 }, { "epoch": 0.02494157593033126, "grad_norm": 2.1242623329162598, "learning_rate": 8.31005586592179e-06, "loss": 0.6042, "step": 595 }, { "epoch": 0.024983494545340223, "grad_norm": 2.2941930294036865, "learning_rate": 8.324022346368715e-06, "loss": 0.6756, "step": 596 }, { "epoch": 0.02502541316034918, "grad_norm": 2.2682604789733887, "learning_rate": 8.337988826815642e-06, "loss": 0.6336, "step": 597 }, { "epoch": 0.02506733177535814, "grad_norm": 2.370041608810425, "learning_rate": 8.35195530726257e-06, "loss": 0.6872, "step": 598 }, { "epoch": 0.025109250390367103, "grad_norm": 2.2761971950531006, "learning_rate": 8.365921787709497e-06, "loss": 0.7175, "step": 599 }, { "epoch": 0.02515116900537606, "grad_norm": 2.47746205329895, "learning_rate": 8.379888268156426e-06, "loss": 0.6269, "step": 600 }, { "epoch": 0.025193087620385024, "grad_norm": 2.57621693611145, "learning_rate": 8.393854748603353e-06, "loss": 0.6445, "step": 601 }, { "epoch": 0.025235006235393983, "grad_norm": 2.147331714630127, "learning_rate": 8.40782122905028e-06, "loss": 0.6666, "step": 602 }, { "epoch": 0.025276924850402942, "grad_norm": 2.630632162094116, "learning_rate": 8.421787709497208e-06, "loss": 0.6524, "step": 603 }, { "epoch": 0.025318843465411904, "grad_norm": 2.0624170303344727, "learning_rate": 8.435754189944135e-06, "loss": 0.6571, "step": 604 }, { "epoch": 0.025360762080420863, "grad_norm": 1.937938928604126, "learning_rate": 8.449720670391061e-06, "loss": 0.6109, "step": 605 }, { "epoch": 0.025402680695429822, "grad_norm": 1.8631874322891235, "learning_rate": 8.463687150837988e-06, "loss": 0.6525, "step": 606 }, { "epoch": 0.025444599310438784, "grad_norm": 3.037720203399658, "learning_rate": 8.477653631284918e-06, "loss": 0.592, "step": 607 }, { "epoch": 0.025486517925447743, "grad_norm": 2.673577308654785, "learning_rate": 8.491620111731845e-06, "loss": 0.6573, "step": 608 }, { "epoch": 0.025528436540456702, "grad_norm": 2.2288999557495117, "learning_rate": 8.505586592178772e-06, "loss": 0.6084, "step": 609 }, { "epoch": 0.025570355155465664, "grad_norm": 2.2973146438598633, "learning_rate": 8.5195530726257e-06, "loss": 0.6805, "step": 610 }, { "epoch": 0.025612273770474623, "grad_norm": 2.1445624828338623, "learning_rate": 8.533519553072627e-06, "loss": 0.6309, "step": 611 }, { "epoch": 0.025654192385483585, "grad_norm": 2.0550131797790527, "learning_rate": 8.547486033519554e-06, "loss": 0.6244, "step": 612 }, { "epoch": 0.025696111000492544, "grad_norm": 2.1979005336761475, "learning_rate": 8.561452513966482e-06, "loss": 0.6303, "step": 613 }, { "epoch": 0.025738029615501503, "grad_norm": 2.220811128616333, "learning_rate": 8.575418994413407e-06, "loss": 0.6805, "step": 614 }, { "epoch": 0.025779948230510465, "grad_norm": 2.4565558433532715, "learning_rate": 8.589385474860336e-06, "loss": 0.6617, "step": 615 }, { "epoch": 0.025821866845519424, "grad_norm": 2.1726880073547363, "learning_rate": 8.603351955307264e-06, "loss": 0.7068, "step": 616 }, { "epoch": 0.025863785460528383, "grad_norm": 2.45832896232605, "learning_rate": 8.617318435754191e-06, "loss": 0.674, "step": 617 }, { "epoch": 0.025905704075537345, "grad_norm": 2.4205663204193115, "learning_rate": 8.631284916201118e-06, "loss": 0.6853, "step": 618 }, { "epoch": 0.025947622690546304, "grad_norm": 3.4726197719573975, "learning_rate": 8.645251396648046e-06, "loss": 0.5867, "step": 619 }, { "epoch": 0.025989541305555263, "grad_norm": 1.9893414974212646, "learning_rate": 8.659217877094973e-06, "loss": 0.5639, "step": 620 }, { "epoch": 0.026031459920564225, "grad_norm": 2.7024102210998535, "learning_rate": 8.6731843575419e-06, "loss": 0.6298, "step": 621 }, { "epoch": 0.026073378535573184, "grad_norm": 2.793827533721924, "learning_rate": 8.687150837988828e-06, "loss": 0.6851, "step": 622 }, { "epoch": 0.026115297150582143, "grad_norm": 2.191304922103882, "learning_rate": 8.701117318435755e-06, "loss": 0.7041, "step": 623 }, { "epoch": 0.026157215765591105, "grad_norm": 2.2656030654907227, "learning_rate": 8.715083798882683e-06, "loss": 0.669, "step": 624 }, { "epoch": 0.026199134380600064, "grad_norm": 2.1486339569091797, "learning_rate": 8.72905027932961e-06, "loss": 0.5976, "step": 625 }, { "epoch": 0.026241052995609027, "grad_norm": 2.6539909839630127, "learning_rate": 8.743016759776537e-06, "loss": 0.7127, "step": 626 }, { "epoch": 0.026282971610617985, "grad_norm": 2.5260841846466064, "learning_rate": 8.756983240223465e-06, "loss": 0.6017, "step": 627 }, { "epoch": 0.026324890225626944, "grad_norm": 2.2829489707946777, "learning_rate": 8.770949720670392e-06, "loss": 0.6496, "step": 628 }, { "epoch": 0.026366808840635907, "grad_norm": 2.329037666320801, "learning_rate": 8.78491620111732e-06, "loss": 0.6404, "step": 629 }, { "epoch": 0.026408727455644865, "grad_norm": 3.115675210952759, "learning_rate": 8.798882681564247e-06, "loss": 0.616, "step": 630 }, { "epoch": 0.026450646070653824, "grad_norm": 2.5623931884765625, "learning_rate": 8.812849162011174e-06, "loss": 0.6565, "step": 631 }, { "epoch": 0.026492564685662787, "grad_norm": 2.3510684967041016, "learning_rate": 8.826815642458101e-06, "loss": 0.6732, "step": 632 }, { "epoch": 0.026534483300671746, "grad_norm": 2.254058361053467, "learning_rate": 8.840782122905029e-06, "loss": 0.6938, "step": 633 }, { "epoch": 0.026576401915680704, "grad_norm": 2.357813596725464, "learning_rate": 8.854748603351956e-06, "loss": 0.6564, "step": 634 }, { "epoch": 0.026618320530689667, "grad_norm": 2.092827320098877, "learning_rate": 8.868715083798883e-06, "loss": 0.7403, "step": 635 }, { "epoch": 0.026660239145698626, "grad_norm": 2.344557523727417, "learning_rate": 8.88268156424581e-06, "loss": 0.606, "step": 636 }, { "epoch": 0.026702157760707588, "grad_norm": 3.6882519721984863, "learning_rate": 8.896648044692738e-06, "loss": 0.6749, "step": 637 }, { "epoch": 0.026744076375716547, "grad_norm": 2.3419227600097656, "learning_rate": 8.910614525139666e-06, "loss": 0.6789, "step": 638 }, { "epoch": 0.026785994990725506, "grad_norm": 2.6330130100250244, "learning_rate": 8.924581005586593e-06, "loss": 0.6801, "step": 639 }, { "epoch": 0.026827913605734468, "grad_norm": 2.5487372875213623, "learning_rate": 8.93854748603352e-06, "loss": 0.6436, "step": 640 }, { "epoch": 0.026869832220743427, "grad_norm": 2.32186222076416, "learning_rate": 8.952513966480448e-06, "loss": 0.6451, "step": 641 }, { "epoch": 0.026911750835752386, "grad_norm": 3.68650221824646, "learning_rate": 8.966480446927375e-06, "loss": 0.6592, "step": 642 }, { "epoch": 0.026953669450761348, "grad_norm": 2.2560391426086426, "learning_rate": 8.980446927374302e-06, "loss": 0.6665, "step": 643 }, { "epoch": 0.026995588065770307, "grad_norm": 2.2607083320617676, "learning_rate": 8.99441340782123e-06, "loss": 0.6173, "step": 644 }, { "epoch": 0.027037506680779266, "grad_norm": 2.390232801437378, "learning_rate": 9.008379888268157e-06, "loss": 0.6045, "step": 645 }, { "epoch": 0.027079425295788228, "grad_norm": 2.2526586055755615, "learning_rate": 9.022346368715084e-06, "loss": 0.6904, "step": 646 }, { "epoch": 0.027121343910797187, "grad_norm": 2.138493537902832, "learning_rate": 9.036312849162012e-06, "loss": 0.6474, "step": 647 }, { "epoch": 0.027163262525806146, "grad_norm": 2.8702101707458496, "learning_rate": 9.050279329608939e-06, "loss": 0.6497, "step": 648 }, { "epoch": 0.027205181140815108, "grad_norm": 2.1779701709747314, "learning_rate": 9.064245810055866e-06, "loss": 0.5991, "step": 649 }, { "epoch": 0.027247099755824067, "grad_norm": 2.408578872680664, "learning_rate": 9.078212290502794e-06, "loss": 0.7225, "step": 650 }, { "epoch": 0.02728901837083303, "grad_norm": 2.7729334831237793, "learning_rate": 9.092178770949721e-06, "loss": 0.6402, "step": 651 }, { "epoch": 0.027330936985841988, "grad_norm": 3.0806005001068115, "learning_rate": 9.106145251396648e-06, "loss": 0.6968, "step": 652 }, { "epoch": 0.027372855600850947, "grad_norm": 2.2014987468719482, "learning_rate": 9.120111731843576e-06, "loss": 0.6594, "step": 653 }, { "epoch": 0.02741477421585991, "grad_norm": 2.19582462310791, "learning_rate": 9.134078212290503e-06, "loss": 0.6767, "step": 654 }, { "epoch": 0.027456692830868868, "grad_norm": 2.152064323425293, "learning_rate": 9.14804469273743e-06, "loss": 0.6556, "step": 655 }, { "epoch": 0.027498611445877827, "grad_norm": 2.2734298706054688, "learning_rate": 9.162011173184358e-06, "loss": 0.6152, "step": 656 }, { "epoch": 0.02754053006088679, "grad_norm": 2.8368523120880127, "learning_rate": 9.175977653631287e-06, "loss": 0.6186, "step": 657 }, { "epoch": 0.027582448675895748, "grad_norm": 2.4587786197662354, "learning_rate": 9.189944134078213e-06, "loss": 0.6119, "step": 658 }, { "epoch": 0.027624367290904707, "grad_norm": 2.152632236480713, "learning_rate": 9.20391061452514e-06, "loss": 0.687, "step": 659 }, { "epoch": 0.02766628590591367, "grad_norm": 2.2641217708587646, "learning_rate": 9.217877094972067e-06, "loss": 0.6769, "step": 660 }, { "epoch": 0.027708204520922628, "grad_norm": 2.2633421421051025, "learning_rate": 9.231843575418995e-06, "loss": 0.68, "step": 661 }, { "epoch": 0.02775012313593159, "grad_norm": 2.180310010910034, "learning_rate": 9.245810055865922e-06, "loss": 0.6789, "step": 662 }, { "epoch": 0.02779204175094055, "grad_norm": 2.6634669303894043, "learning_rate": 9.25977653631285e-06, "loss": 0.652, "step": 663 }, { "epoch": 0.027833960365949508, "grad_norm": 2.5330326557159424, "learning_rate": 9.273743016759777e-06, "loss": 0.7277, "step": 664 }, { "epoch": 0.02787587898095847, "grad_norm": 3.3534345626831055, "learning_rate": 9.287709497206706e-06, "loss": 0.6224, "step": 665 }, { "epoch": 0.02791779759596743, "grad_norm": 2.9641172885894775, "learning_rate": 9.301675977653633e-06, "loss": 0.7273, "step": 666 }, { "epoch": 0.027959716210976388, "grad_norm": 2.370046377182007, "learning_rate": 9.315642458100559e-06, "loss": 0.6523, "step": 667 }, { "epoch": 0.02800163482598535, "grad_norm": 2.3549857139587402, "learning_rate": 9.329608938547486e-06, "loss": 0.7147, "step": 668 }, { "epoch": 0.02804355344099431, "grad_norm": 2.2561306953430176, "learning_rate": 9.343575418994413e-06, "loss": 0.6757, "step": 669 }, { "epoch": 0.028085472056003268, "grad_norm": 2.076265573501587, "learning_rate": 9.35754189944134e-06, "loss": 0.6699, "step": 670 }, { "epoch": 0.02812739067101223, "grad_norm": 2.3372890949249268, "learning_rate": 9.371508379888268e-06, "loss": 0.6751, "step": 671 }, { "epoch": 0.02816930928602119, "grad_norm": 1.9389113187789917, "learning_rate": 9.385474860335197e-06, "loss": 0.6177, "step": 672 }, { "epoch": 0.028211227901030148, "grad_norm": 2.3378565311431885, "learning_rate": 9.399441340782125e-06, "loss": 0.6608, "step": 673 }, { "epoch": 0.02825314651603911, "grad_norm": 2.1655216217041016, "learning_rate": 9.413407821229052e-06, "loss": 0.6227, "step": 674 }, { "epoch": 0.02829506513104807, "grad_norm": 2.2022480964660645, "learning_rate": 9.42737430167598e-06, "loss": 0.615, "step": 675 }, { "epoch": 0.028336983746057032, "grad_norm": 2.083022117614746, "learning_rate": 9.441340782122905e-06, "loss": 0.6365, "step": 676 }, { "epoch": 0.02837890236106599, "grad_norm": 2.1779723167419434, "learning_rate": 9.455307262569832e-06, "loss": 0.6255, "step": 677 }, { "epoch": 0.02842082097607495, "grad_norm": 2.324958324432373, "learning_rate": 9.46927374301676e-06, "loss": 0.6691, "step": 678 }, { "epoch": 0.028462739591083912, "grad_norm": 2.029717445373535, "learning_rate": 9.483240223463687e-06, "loss": 0.6372, "step": 679 }, { "epoch": 0.02850465820609287, "grad_norm": 2.416611433029175, "learning_rate": 9.497206703910616e-06, "loss": 0.6585, "step": 680 }, { "epoch": 0.02854657682110183, "grad_norm": 2.362761974334717, "learning_rate": 9.511173184357543e-06, "loss": 0.7062, "step": 681 }, { "epoch": 0.028588495436110792, "grad_norm": 2.506455659866333, "learning_rate": 9.52513966480447e-06, "loss": 0.6718, "step": 682 }, { "epoch": 0.02863041405111975, "grad_norm": 2.1242499351501465, "learning_rate": 9.539106145251398e-06, "loss": 0.6247, "step": 683 }, { "epoch": 0.02867233266612871, "grad_norm": 2.009378433227539, "learning_rate": 9.553072625698325e-06, "loss": 0.6655, "step": 684 }, { "epoch": 0.028714251281137672, "grad_norm": 1.9742814302444458, "learning_rate": 9.567039106145251e-06, "loss": 0.6166, "step": 685 }, { "epoch": 0.02875616989614663, "grad_norm": 2.1838138103485107, "learning_rate": 9.581005586592178e-06, "loss": 0.601, "step": 686 }, { "epoch": 0.028798088511155593, "grad_norm": 2.097086191177368, "learning_rate": 9.594972067039106e-06, "loss": 0.5984, "step": 687 }, { "epoch": 0.028840007126164552, "grad_norm": 2.4223294258117676, "learning_rate": 9.608938547486035e-06, "loss": 0.6202, "step": 688 }, { "epoch": 0.02888192574117351, "grad_norm": 1.9200360774993896, "learning_rate": 9.622905027932962e-06, "loss": 0.6067, "step": 689 }, { "epoch": 0.028923844356182473, "grad_norm": 2.105224132537842, "learning_rate": 9.63687150837989e-06, "loss": 0.6067, "step": 690 }, { "epoch": 0.028965762971191432, "grad_norm": 2.300150156021118, "learning_rate": 9.650837988826817e-06, "loss": 0.6649, "step": 691 }, { "epoch": 0.02900768158620039, "grad_norm": 2.342790126800537, "learning_rate": 9.664804469273744e-06, "loss": 0.6468, "step": 692 }, { "epoch": 0.029049600201209353, "grad_norm": 1.9945627450942993, "learning_rate": 9.678770949720672e-06, "loss": 0.6218, "step": 693 }, { "epoch": 0.029091518816218312, "grad_norm": 2.0632734298706055, "learning_rate": 9.692737430167597e-06, "loss": 0.6705, "step": 694 }, { "epoch": 0.02913343743122727, "grad_norm": 1.9253826141357422, "learning_rate": 9.706703910614526e-06, "loss": 0.5956, "step": 695 }, { "epoch": 0.029175356046236233, "grad_norm": 2.0828006267547607, "learning_rate": 9.720670391061454e-06, "loss": 0.6457, "step": 696 }, { "epoch": 0.029217274661245192, "grad_norm": 2.241994619369507, "learning_rate": 9.734636871508381e-06, "loss": 0.6581, "step": 697 }, { "epoch": 0.029259193276254154, "grad_norm": 2.4707107543945312, "learning_rate": 9.748603351955308e-06, "loss": 0.61, "step": 698 }, { "epoch": 0.029301111891263113, "grad_norm": 2.1014795303344727, "learning_rate": 9.762569832402236e-06, "loss": 0.6515, "step": 699 }, { "epoch": 0.029343030506272072, "grad_norm": 1.9869345426559448, "learning_rate": 9.776536312849163e-06, "loss": 0.6057, "step": 700 }, { "epoch": 0.029384949121281034, "grad_norm": 2.161334753036499, "learning_rate": 9.79050279329609e-06, "loss": 0.6438, "step": 701 }, { "epoch": 0.029426867736289993, "grad_norm": 2.1327807903289795, "learning_rate": 9.804469273743018e-06, "loss": 0.6234, "step": 702 }, { "epoch": 0.029468786351298952, "grad_norm": 2.159215211868286, "learning_rate": 9.818435754189945e-06, "loss": 0.6963, "step": 703 }, { "epoch": 0.029510704966307914, "grad_norm": 2.1804020404815674, "learning_rate": 9.832402234636873e-06, "loss": 0.6842, "step": 704 }, { "epoch": 0.029552623581316873, "grad_norm": 2.9227845668792725, "learning_rate": 9.8463687150838e-06, "loss": 0.6372, "step": 705 }, { "epoch": 0.029594542196325832, "grad_norm": 2.6973304748535156, "learning_rate": 9.860335195530727e-06, "loss": 0.6247, "step": 706 }, { "epoch": 0.029636460811334794, "grad_norm": 2.110806703567505, "learning_rate": 9.874301675977655e-06, "loss": 0.6202, "step": 707 }, { "epoch": 0.029678379426343753, "grad_norm": 1.92245352268219, "learning_rate": 9.888268156424582e-06, "loss": 0.6525, "step": 708 }, { "epoch": 0.029720298041352712, "grad_norm": 1.8969814777374268, "learning_rate": 9.90223463687151e-06, "loss": 0.6299, "step": 709 }, { "epoch": 0.029762216656361674, "grad_norm": 2.4143826961517334, "learning_rate": 9.916201117318437e-06, "loss": 0.6009, "step": 710 }, { "epoch": 0.029804135271370633, "grad_norm": 1.9404470920562744, "learning_rate": 9.930167597765364e-06, "loss": 0.5633, "step": 711 }, { "epoch": 0.029846053886379596, "grad_norm": 3.9270551204681396, "learning_rate": 9.944134078212291e-06, "loss": 0.6068, "step": 712 }, { "epoch": 0.029887972501388554, "grad_norm": 2.1888234615325928, "learning_rate": 9.958100558659219e-06, "loss": 0.6474, "step": 713 }, { "epoch": 0.029929891116397513, "grad_norm": 2.618354558944702, "learning_rate": 9.972067039106146e-06, "loss": 0.6259, "step": 714 }, { "epoch": 0.029971809731406476, "grad_norm": 3.4958300590515137, "learning_rate": 9.986033519553073e-06, "loss": 0.5993, "step": 715 }, { "epoch": 0.030013728346415434, "grad_norm": 2.0776946544647217, "learning_rate": 1e-05, "loss": 0.6924, "step": 716 }, { "epoch": 0.030055646961424393, "grad_norm": 2.1698451042175293, "learning_rate": 9.999999953915957e-06, "loss": 0.6595, "step": 717 }, { "epoch": 0.030097565576433356, "grad_norm": 2.1589391231536865, "learning_rate": 9.999999815663825e-06, "loss": 0.6791, "step": 718 }, { "epoch": 0.030139484191442314, "grad_norm": 2.1829891204833984, "learning_rate": 9.999999585243611e-06, "loss": 0.6682, "step": 719 }, { "epoch": 0.030181402806451273, "grad_norm": 1.921593427658081, "learning_rate": 9.999999262655315e-06, "loss": 0.5958, "step": 720 }, { "epoch": 0.030223321421460236, "grad_norm": 2.5506343841552734, "learning_rate": 9.999998847898946e-06, "loss": 0.6246, "step": 721 }, { "epoch": 0.030265240036469195, "grad_norm": 2.0642521381378174, "learning_rate": 9.999998340974511e-06, "loss": 0.6371, "step": 722 }, { "epoch": 0.030307158651478157, "grad_norm": 2.260650396347046, "learning_rate": 9.999997741882019e-06, "loss": 0.6382, "step": 723 }, { "epoch": 0.030349077266487116, "grad_norm": 2.367037534713745, "learning_rate": 9.99999705062148e-06, "loss": 0.6369, "step": 724 }, { "epoch": 0.030390995881496075, "grad_norm": 2.329249858856201, "learning_rate": 9.999996267192908e-06, "loss": 0.6795, "step": 725 }, { "epoch": 0.030432914496505037, "grad_norm": 2.8582522869110107, "learning_rate": 9.999995391596315e-06, "loss": 0.6301, "step": 726 }, { "epoch": 0.030474833111513996, "grad_norm": 2.533318519592285, "learning_rate": 9.999994423831723e-06, "loss": 0.6371, "step": 727 }, { "epoch": 0.030516751726522955, "grad_norm": 2.104336738586426, "learning_rate": 9.999993363899144e-06, "loss": 0.6462, "step": 728 }, { "epoch": 0.030558670341531917, "grad_norm": 2.224116325378418, "learning_rate": 9.999992211798599e-06, "loss": 0.6955, "step": 729 }, { "epoch": 0.030600588956540876, "grad_norm": 2.233973264694214, "learning_rate": 9.999990967530111e-06, "loss": 0.6465, "step": 730 }, { "epoch": 0.030642507571549835, "grad_norm": 2.1715362071990967, "learning_rate": 9.999989631093702e-06, "loss": 0.6083, "step": 731 }, { "epoch": 0.030684426186558797, "grad_norm": 2.40329647064209, "learning_rate": 9.999988202489395e-06, "loss": 0.686, "step": 732 }, { "epoch": 0.030726344801567756, "grad_norm": 2.2688918113708496, "learning_rate": 9.99998668171722e-06, "loss": 0.6534, "step": 733 }, { "epoch": 0.030768263416576715, "grad_norm": 1.9682135581970215, "learning_rate": 9.9999850687772e-06, "loss": 0.6693, "step": 734 }, { "epoch": 0.030810182031585677, "grad_norm": 2.8932173252105713, "learning_rate": 9.99998336366937e-06, "loss": 0.6705, "step": 735 }, { "epoch": 0.030852100646594636, "grad_norm": 2.14335036277771, "learning_rate": 9.999981566393758e-06, "loss": 0.7055, "step": 736 }, { "epoch": 0.030894019261603598, "grad_norm": 2.570866584777832, "learning_rate": 9.999979676950398e-06, "loss": 0.716, "step": 737 }, { "epoch": 0.030935937876612557, "grad_norm": 2.7873876094818115, "learning_rate": 9.999977695339324e-06, "loss": 0.5881, "step": 738 }, { "epoch": 0.030977856491621516, "grad_norm": 2.2102718353271484, "learning_rate": 9.999975621560575e-06, "loss": 0.653, "step": 739 }, { "epoch": 0.031019775106630478, "grad_norm": 2.680661916732788, "learning_rate": 9.999973455614188e-06, "loss": 0.6524, "step": 740 }, { "epoch": 0.031061693721639437, "grad_norm": 2.1248762607574463, "learning_rate": 9.9999711975002e-06, "loss": 0.6585, "step": 741 }, { "epoch": 0.031103612336648396, "grad_norm": 2.1659164428710938, "learning_rate": 9.999968847218657e-06, "loss": 0.6369, "step": 742 }, { "epoch": 0.031145530951657358, "grad_norm": 2.4370532035827637, "learning_rate": 9.999966404769601e-06, "loss": 0.7115, "step": 743 }, { "epoch": 0.031187449566666317, "grad_norm": 2.2942252159118652, "learning_rate": 9.999963870153077e-06, "loss": 0.6771, "step": 744 }, { "epoch": 0.031229368181675276, "grad_norm": 2.4939041137695312, "learning_rate": 9.99996124336913e-06, "loss": 0.6407, "step": 745 }, { "epoch": 0.031271286796684235, "grad_norm": 2.019752264022827, "learning_rate": 9.99995852441781e-06, "loss": 0.6606, "step": 746 }, { "epoch": 0.0313132054116932, "grad_norm": 2.3534154891967773, "learning_rate": 9.999955713299167e-06, "loss": 0.6095, "step": 747 }, { "epoch": 0.03135512402670216, "grad_norm": 1.9930251836776733, "learning_rate": 9.999952810013252e-06, "loss": 0.6895, "step": 748 }, { "epoch": 0.031397042641711115, "grad_norm": 1.9856761693954468, "learning_rate": 9.99994981456012e-06, "loss": 0.6459, "step": 749 }, { "epoch": 0.03143896125672008, "grad_norm": 2.278073310852051, "learning_rate": 9.999946726939825e-06, "loss": 0.711, "step": 750 }, { "epoch": 0.03148087987172904, "grad_norm": 2.158524751663208, "learning_rate": 9.999943547152425e-06, "loss": 0.6496, "step": 751 }, { "epoch": 0.031522798486737995, "grad_norm": 2.1586833000183105, "learning_rate": 9.999940275197977e-06, "loss": 0.6229, "step": 752 }, { "epoch": 0.03156471710174696, "grad_norm": 2.1443099975585938, "learning_rate": 9.999936911076542e-06, "loss": 0.6544, "step": 753 }, { "epoch": 0.03160663571675592, "grad_norm": 2.245804786682129, "learning_rate": 9.999933454788184e-06, "loss": 0.7083, "step": 754 }, { "epoch": 0.03164855433176488, "grad_norm": 1.8603520393371582, "learning_rate": 9.999929906332963e-06, "loss": 0.5913, "step": 755 }, { "epoch": 0.03169047294677384, "grad_norm": 2.2342495918273926, "learning_rate": 9.999926265710946e-06, "loss": 0.6503, "step": 756 }, { "epoch": 0.0317323915617828, "grad_norm": 1.9558509588241577, "learning_rate": 9.999922532922203e-06, "loss": 0.6647, "step": 757 }, { "epoch": 0.03177431017679176, "grad_norm": 2.0300614833831787, "learning_rate": 9.999918707966798e-06, "loss": 0.7172, "step": 758 }, { "epoch": 0.03181622879180072, "grad_norm": 2.274648427963257, "learning_rate": 9.999914790844806e-06, "loss": 0.6035, "step": 759 }, { "epoch": 0.03185814740680968, "grad_norm": 2.0076394081115723, "learning_rate": 9.999910781556294e-06, "loss": 0.6573, "step": 760 }, { "epoch": 0.03190006602181864, "grad_norm": 2.182783842086792, "learning_rate": 9.999906680101341e-06, "loss": 0.7257, "step": 761 }, { "epoch": 0.0319419846368276, "grad_norm": 2.8411145210266113, "learning_rate": 9.999902486480022e-06, "loss": 0.6781, "step": 762 }, { "epoch": 0.03198390325183656, "grad_norm": 2.881937026977539, "learning_rate": 9.99989820069241e-06, "loss": 0.6505, "step": 763 }, { "epoch": 0.03202582186684552, "grad_norm": 2.1319310665130615, "learning_rate": 9.999893822738589e-06, "loss": 0.6565, "step": 764 }, { "epoch": 0.03206774048185448, "grad_norm": 2.0614030361175537, "learning_rate": 9.999889352618634e-06, "loss": 0.6718, "step": 765 }, { "epoch": 0.03210965909686344, "grad_norm": 2.5795786380767822, "learning_rate": 9.999884790332635e-06, "loss": 0.693, "step": 766 }, { "epoch": 0.0321515777118724, "grad_norm": 2.389796495437622, "learning_rate": 9.999880135880669e-06, "loss": 0.6182, "step": 767 }, { "epoch": 0.03219349632688136, "grad_norm": 2.1625802516937256, "learning_rate": 9.999875389262828e-06, "loss": 0.6513, "step": 768 }, { "epoch": 0.03223541494189032, "grad_norm": 2.315321922302246, "learning_rate": 9.999870550479194e-06, "loss": 0.6033, "step": 769 }, { "epoch": 0.03227733355689928, "grad_norm": 1.8646318912506104, "learning_rate": 9.999865619529859e-06, "loss": 0.6427, "step": 770 }, { "epoch": 0.03231925217190824, "grad_norm": 2.0160062313079834, "learning_rate": 9.999860596414912e-06, "loss": 0.568, "step": 771 }, { "epoch": 0.0323611707869172, "grad_norm": 2.043740749359131, "learning_rate": 9.999855481134449e-06, "loss": 0.6753, "step": 772 }, { "epoch": 0.03240308940192616, "grad_norm": 2.0608816146850586, "learning_rate": 9.999850273688562e-06, "loss": 0.5753, "step": 773 }, { "epoch": 0.03244500801693512, "grad_norm": 2.6602981090545654, "learning_rate": 9.999844974077346e-06, "loss": 0.6272, "step": 774 }, { "epoch": 0.03248692663194408, "grad_norm": 1.9657292366027832, "learning_rate": 9.999839582300901e-06, "loss": 0.6678, "step": 775 }, { "epoch": 0.03252884524695304, "grad_norm": 2.5170111656188965, "learning_rate": 9.999834098359326e-06, "loss": 0.6857, "step": 776 }, { "epoch": 0.032570763861962, "grad_norm": 2.0246846675872803, "learning_rate": 9.999828522252721e-06, "loss": 0.6329, "step": 777 }, { "epoch": 0.03261268247697096, "grad_norm": 1.9355851411819458, "learning_rate": 9.99982285398119e-06, "loss": 0.6545, "step": 778 }, { "epoch": 0.03265460109197992, "grad_norm": 1.939918041229248, "learning_rate": 9.999817093544835e-06, "loss": 0.5995, "step": 779 }, { "epoch": 0.032696519706988884, "grad_norm": 2.1070618629455566, "learning_rate": 9.999811240943764e-06, "loss": 0.635, "step": 780 }, { "epoch": 0.03273843832199784, "grad_norm": 2.419384479522705, "learning_rate": 9.999805296178086e-06, "loss": 0.6787, "step": 781 }, { "epoch": 0.0327803569370068, "grad_norm": 1.8461296558380127, "learning_rate": 9.999799259247909e-06, "loss": 0.6289, "step": 782 }, { "epoch": 0.032822275552015764, "grad_norm": 1.9640430212020874, "learning_rate": 9.999793130153344e-06, "loss": 0.6483, "step": 783 }, { "epoch": 0.03286419416702472, "grad_norm": 1.8324068784713745, "learning_rate": 9.999786908894505e-06, "loss": 0.6452, "step": 784 }, { "epoch": 0.03290611278203368, "grad_norm": 2.241764545440674, "learning_rate": 9.999780595471506e-06, "loss": 0.7177, "step": 785 }, { "epoch": 0.032948031397042644, "grad_norm": 1.950531244277954, "learning_rate": 9.999774189884464e-06, "loss": 0.6189, "step": 786 }, { "epoch": 0.0329899500120516, "grad_norm": 3.139158248901367, "learning_rate": 9.999767692133498e-06, "loss": 0.6758, "step": 787 }, { "epoch": 0.03303186862706056, "grad_norm": 1.947472333908081, "learning_rate": 9.999761102218725e-06, "loss": 0.6006, "step": 788 }, { "epoch": 0.033073787242069524, "grad_norm": 2.054529905319214, "learning_rate": 9.99975442014027e-06, "loss": 0.6515, "step": 789 }, { "epoch": 0.03311570585707848, "grad_norm": 2.0914242267608643, "learning_rate": 9.999747645898252e-06, "loss": 0.6796, "step": 790 }, { "epoch": 0.03315762447208744, "grad_norm": 2.0671627521514893, "learning_rate": 9.9997407794928e-06, "loss": 0.6543, "step": 791 }, { "epoch": 0.033199543087096405, "grad_norm": 2.087556838989258, "learning_rate": 9.999733820924036e-06, "loss": 0.6299, "step": 792 }, { "epoch": 0.03324146170210536, "grad_norm": 2.0677614212036133, "learning_rate": 9.999726770192093e-06, "loss": 0.6268, "step": 793 }, { "epoch": 0.03328338031711432, "grad_norm": 3.043769121170044, "learning_rate": 9.999719627297099e-06, "loss": 0.6334, "step": 794 }, { "epoch": 0.033325298932123285, "grad_norm": 1.7962676286697388, "learning_rate": 9.999712392239185e-06, "loss": 0.5867, "step": 795 }, { "epoch": 0.03336721754713224, "grad_norm": 1.9781280755996704, "learning_rate": 9.999705065018485e-06, "loss": 0.6047, "step": 796 }, { "epoch": 0.0334091361621412, "grad_norm": 1.9063185453414917, "learning_rate": 9.999697645635133e-06, "loss": 0.6103, "step": 797 }, { "epoch": 0.033451054777150165, "grad_norm": 1.878420114517212, "learning_rate": 9.999690134089269e-06, "loss": 0.649, "step": 798 }, { "epoch": 0.03349297339215912, "grad_norm": 1.9801173210144043, "learning_rate": 9.999682530381028e-06, "loss": 0.6668, "step": 799 }, { "epoch": 0.03353489200716808, "grad_norm": 1.9846618175506592, "learning_rate": 9.999674834510552e-06, "loss": 0.6865, "step": 800 }, { "epoch": 0.033576810622177045, "grad_norm": 2.082828998565674, "learning_rate": 9.999667046477982e-06, "loss": 0.6293, "step": 801 }, { "epoch": 0.033618729237186, "grad_norm": 2.1333506107330322, "learning_rate": 9.99965916628346e-06, "loss": 0.6334, "step": 802 }, { "epoch": 0.03366064785219496, "grad_norm": 2.034817934036255, "learning_rate": 9.999651193927134e-06, "loss": 0.7208, "step": 803 }, { "epoch": 0.033702566467203925, "grad_norm": 1.9576637744903564, "learning_rate": 9.99964312940915e-06, "loss": 0.6974, "step": 804 }, { "epoch": 0.03374448508221289, "grad_norm": 2.2368080615997314, "learning_rate": 9.999634972729658e-06, "loss": 0.6815, "step": 805 }, { "epoch": 0.03378640369722184, "grad_norm": 1.856459140777588, "learning_rate": 9.999626723888806e-06, "loss": 0.6717, "step": 806 }, { "epoch": 0.033828322312230805, "grad_norm": 2.1301348209381104, "learning_rate": 9.999618382886749e-06, "loss": 0.7119, "step": 807 }, { "epoch": 0.03387024092723977, "grad_norm": 2.161363363265991, "learning_rate": 9.999609949723637e-06, "loss": 0.6212, "step": 808 }, { "epoch": 0.03391215954224872, "grad_norm": 1.874588966369629, "learning_rate": 9.999601424399627e-06, "loss": 0.6573, "step": 809 }, { "epoch": 0.033954078157257685, "grad_norm": 2.2790310382843018, "learning_rate": 9.999592806914878e-06, "loss": 0.6385, "step": 810 }, { "epoch": 0.03399599677226665, "grad_norm": 1.8422726392745972, "learning_rate": 9.999584097269546e-06, "loss": 0.6441, "step": 811 }, { "epoch": 0.0340379153872756, "grad_norm": 1.8333834409713745, "learning_rate": 9.999575295463794e-06, "loss": 0.6292, "step": 812 }, { "epoch": 0.034079834002284565, "grad_norm": 5.092312335968018, "learning_rate": 9.999566401497783e-06, "loss": 0.6105, "step": 813 }, { "epoch": 0.03412175261729353, "grad_norm": 2.234649181365967, "learning_rate": 9.999557415371677e-06, "loss": 0.5895, "step": 814 }, { "epoch": 0.03416367123230248, "grad_norm": 2.331089973449707, "learning_rate": 9.999548337085641e-06, "loss": 0.6354, "step": 815 }, { "epoch": 0.034205589847311445, "grad_norm": 2.1508147716522217, "learning_rate": 9.999539166639842e-06, "loss": 0.5943, "step": 816 }, { "epoch": 0.03424750846232041, "grad_norm": 2.214442014694214, "learning_rate": 9.999529904034454e-06, "loss": 0.6025, "step": 817 }, { "epoch": 0.03428942707732936, "grad_norm": 2.141712188720703, "learning_rate": 9.999520549269639e-06, "loss": 0.6346, "step": 818 }, { "epoch": 0.034331345692338325, "grad_norm": 2.361402750015259, "learning_rate": 9.999511102345578e-06, "loss": 0.6725, "step": 819 }, { "epoch": 0.03437326430734729, "grad_norm": 2.0400681495666504, "learning_rate": 9.99950156326244e-06, "loss": 0.6365, "step": 820 }, { "epoch": 0.03441518292235624, "grad_norm": 1.8471722602844238, "learning_rate": 9.999491932020401e-06, "loss": 0.6445, "step": 821 }, { "epoch": 0.034457101537365205, "grad_norm": 2.7184407711029053, "learning_rate": 9.999482208619641e-06, "loss": 0.6765, "step": 822 }, { "epoch": 0.03449902015237417, "grad_norm": 1.9119573831558228, "learning_rate": 9.99947239306034e-06, "loss": 0.6069, "step": 823 }, { "epoch": 0.03454093876738312, "grad_norm": 1.9898536205291748, "learning_rate": 9.999462485342675e-06, "loss": 0.6257, "step": 824 }, { "epoch": 0.034582857382392085, "grad_norm": 2.4578464031219482, "learning_rate": 9.999452485466832e-06, "loss": 0.618, "step": 825 }, { "epoch": 0.03462477599740105, "grad_norm": 2.04687762260437, "learning_rate": 9.999442393432992e-06, "loss": 0.6188, "step": 826 }, { "epoch": 0.03466669461241, "grad_norm": 2.0053558349609375, "learning_rate": 9.999432209241345e-06, "loss": 0.5964, "step": 827 }, { "epoch": 0.034708613227418965, "grad_norm": 2.1637418270111084, "learning_rate": 9.999421932892075e-06, "loss": 0.6305, "step": 828 }, { "epoch": 0.03475053184242793, "grad_norm": 2.1743037700653076, "learning_rate": 9.999411564385372e-06, "loss": 0.6066, "step": 829 }, { "epoch": 0.03479245045743689, "grad_norm": 2.4274322986602783, "learning_rate": 9.999401103721431e-06, "loss": 0.6979, "step": 830 }, { "epoch": 0.034834369072445845, "grad_norm": 2.159470558166504, "learning_rate": 9.999390550900441e-06, "loss": 0.6586, "step": 831 }, { "epoch": 0.03487628768745481, "grad_norm": 2.2080342769622803, "learning_rate": 9.9993799059226e-06, "loss": 0.7225, "step": 832 }, { "epoch": 0.03491820630246377, "grad_norm": 1.9519104957580566, "learning_rate": 9.999369168788098e-06, "loss": 0.6444, "step": 833 }, { "epoch": 0.034960124917472725, "grad_norm": 2.201483726501465, "learning_rate": 9.999358339497139e-06, "loss": 0.6527, "step": 834 }, { "epoch": 0.03500204353248169, "grad_norm": 1.9946222305297852, "learning_rate": 9.99934741804992e-06, "loss": 0.609, "step": 835 }, { "epoch": 0.03504396214749065, "grad_norm": 2.2267982959747314, "learning_rate": 9.999336404446643e-06, "loss": 0.6034, "step": 836 }, { "epoch": 0.035085880762499605, "grad_norm": 2.1866965293884277, "learning_rate": 9.99932529868751e-06, "loss": 0.5896, "step": 837 }, { "epoch": 0.03512779937750857, "grad_norm": 2.2730767726898193, "learning_rate": 9.999314100772729e-06, "loss": 0.6211, "step": 838 }, { "epoch": 0.03516971799251753, "grad_norm": 3.748840570449829, "learning_rate": 9.999302810702501e-06, "loss": 0.6296, "step": 839 }, { "epoch": 0.035211636607526485, "grad_norm": 3.1725950241088867, "learning_rate": 9.999291428477037e-06, "loss": 0.634, "step": 840 }, { "epoch": 0.03525355522253545, "grad_norm": 2.264741897583008, "learning_rate": 9.999279954096548e-06, "loss": 0.604, "step": 841 }, { "epoch": 0.03529547383754441, "grad_norm": 2.0657284259796143, "learning_rate": 9.999268387561245e-06, "loss": 0.5995, "step": 842 }, { "epoch": 0.035337392452553365, "grad_norm": 2.039092540740967, "learning_rate": 9.999256728871338e-06, "loss": 0.6138, "step": 843 }, { "epoch": 0.03537931106756233, "grad_norm": 2.5151424407958984, "learning_rate": 9.999244978027046e-06, "loss": 0.6155, "step": 844 }, { "epoch": 0.03542122968257129, "grad_norm": 2.0595638751983643, "learning_rate": 9.999233135028586e-06, "loss": 0.5565, "step": 845 }, { "epoch": 0.035463148297580245, "grad_norm": 2.079683780670166, "learning_rate": 9.999221199876172e-06, "loss": 0.6852, "step": 846 }, { "epoch": 0.03550506691258921, "grad_norm": 1.891739010810852, "learning_rate": 9.999209172570026e-06, "loss": 0.6252, "step": 847 }, { "epoch": 0.03554698552759817, "grad_norm": 7.136612892150879, "learning_rate": 9.999197053110371e-06, "loss": 0.6913, "step": 848 }, { "epoch": 0.035588904142607125, "grad_norm": 2.1642067432403564, "learning_rate": 9.99918484149743e-06, "loss": 0.6422, "step": 849 }, { "epoch": 0.03563082275761609, "grad_norm": 3.205826759338379, "learning_rate": 9.999172537731427e-06, "loss": 0.6071, "step": 850 }, { "epoch": 0.03567274137262505, "grad_norm": 1.94369375705719, "learning_rate": 9.99916014181259e-06, "loss": 0.5608, "step": 851 }, { "epoch": 0.035714659987634005, "grad_norm": 2.27901291847229, "learning_rate": 9.999147653741146e-06, "loss": 0.6151, "step": 852 }, { "epoch": 0.03575657860264297, "grad_norm": 4.263737201690674, "learning_rate": 9.999135073517326e-06, "loss": 0.6094, "step": 853 }, { "epoch": 0.03579849721765193, "grad_norm": 2.0487706661224365, "learning_rate": 9.999122401141364e-06, "loss": 0.6351, "step": 854 }, { "epoch": 0.03584041583266089, "grad_norm": 2.4582996368408203, "learning_rate": 9.999109636613489e-06, "loss": 0.6141, "step": 855 }, { "epoch": 0.03588233444766985, "grad_norm": 2.8045034408569336, "learning_rate": 9.999096779933941e-06, "loss": 0.6427, "step": 856 }, { "epoch": 0.03592425306267881, "grad_norm": 2.344547986984253, "learning_rate": 9.999083831102953e-06, "loss": 0.6696, "step": 857 }, { "epoch": 0.03596617167768777, "grad_norm": 2.2919669151306152, "learning_rate": 9.999070790120768e-06, "loss": 0.6446, "step": 858 }, { "epoch": 0.03600809029269673, "grad_norm": 2.579813003540039, "learning_rate": 9.999057656987622e-06, "loss": 0.6729, "step": 859 }, { "epoch": 0.03605000890770569, "grad_norm": 3.104531764984131, "learning_rate": 9.99904443170376e-06, "loss": 0.5987, "step": 860 }, { "epoch": 0.03609192752271465, "grad_norm": 1.8666026592254639, "learning_rate": 9.999031114269424e-06, "loss": 0.5862, "step": 861 }, { "epoch": 0.03613384613772361, "grad_norm": 2.042297840118408, "learning_rate": 9.999017704684861e-06, "loss": 0.6355, "step": 862 }, { "epoch": 0.03617576475273257, "grad_norm": 2.058321475982666, "learning_rate": 9.999004202950317e-06, "loss": 0.6024, "step": 863 }, { "epoch": 0.03621768336774153, "grad_norm": 1.8740770816802979, "learning_rate": 9.998990609066042e-06, "loss": 0.6122, "step": 864 }, { "epoch": 0.03625960198275049, "grad_norm": 2.3495712280273438, "learning_rate": 9.998976923032287e-06, "loss": 0.6359, "step": 865 }, { "epoch": 0.03630152059775945, "grad_norm": 1.8214011192321777, "learning_rate": 9.998963144849303e-06, "loss": 0.6057, "step": 866 }, { "epoch": 0.03634343921276841, "grad_norm": 2.3904340267181396, "learning_rate": 9.998949274517341e-06, "loss": 0.653, "step": 867 }, { "epoch": 0.03638535782777737, "grad_norm": 2.204500436782837, "learning_rate": 9.998935312036664e-06, "loss": 0.6549, "step": 868 }, { "epoch": 0.03642727644278633, "grad_norm": 2.072605848312378, "learning_rate": 9.998921257407523e-06, "loss": 0.6267, "step": 869 }, { "epoch": 0.03646919505779529, "grad_norm": 2.400864839553833, "learning_rate": 9.99890711063018e-06, "loss": 0.6694, "step": 870 }, { "epoch": 0.03651111367280425, "grad_norm": 2.1897146701812744, "learning_rate": 9.998892871704896e-06, "loss": 0.6107, "step": 871 }, { "epoch": 0.03655303228781321, "grad_norm": 2.2820122241973877, "learning_rate": 9.998878540631933e-06, "loss": 0.6294, "step": 872 }, { "epoch": 0.03659495090282217, "grad_norm": 2.174614429473877, "learning_rate": 9.998864117411553e-06, "loss": 0.6277, "step": 873 }, { "epoch": 0.03663686951783113, "grad_norm": 2.152735471725464, "learning_rate": 9.998849602044025e-06, "loss": 0.6794, "step": 874 }, { "epoch": 0.03667878813284009, "grad_norm": 2.051175355911255, "learning_rate": 9.998834994529616e-06, "loss": 0.5986, "step": 875 }, { "epoch": 0.03672070674784905, "grad_norm": 2.0627896785736084, "learning_rate": 9.998820294868592e-06, "loss": 0.6503, "step": 876 }, { "epoch": 0.036762625362858015, "grad_norm": 2.885378360748291, "learning_rate": 9.998805503061228e-06, "loss": 0.6386, "step": 877 }, { "epoch": 0.03680454397786697, "grad_norm": 1.9922055006027222, "learning_rate": 9.998790619107797e-06, "loss": 0.5589, "step": 878 }, { "epoch": 0.03684646259287593, "grad_norm": 2.215498924255371, "learning_rate": 9.99877564300857e-06, "loss": 0.7185, "step": 879 }, { "epoch": 0.036888381207884895, "grad_norm": 2.0374271869659424, "learning_rate": 9.998760574763824e-06, "loss": 0.633, "step": 880 }, { "epoch": 0.03693029982289385, "grad_norm": 2.0164170265197754, "learning_rate": 9.998745414373837e-06, "loss": 0.5943, "step": 881 }, { "epoch": 0.03697221843790281, "grad_norm": 2.413146734237671, "learning_rate": 9.99873016183889e-06, "loss": 0.7601, "step": 882 }, { "epoch": 0.037014137052911775, "grad_norm": 2.4408116340637207, "learning_rate": 9.998714817159264e-06, "loss": 0.6866, "step": 883 }, { "epoch": 0.03705605566792073, "grad_norm": 2.1920294761657715, "learning_rate": 9.998699380335237e-06, "loss": 0.6116, "step": 884 }, { "epoch": 0.03709797428292969, "grad_norm": 2.503025531768799, "learning_rate": 9.998683851367101e-06, "loss": 0.6492, "step": 885 }, { "epoch": 0.037139892897938655, "grad_norm": 2.2482995986938477, "learning_rate": 9.998668230255138e-06, "loss": 0.6968, "step": 886 }, { "epoch": 0.03718181151294761, "grad_norm": 2.7362990379333496, "learning_rate": 9.998652516999636e-06, "loss": 0.6193, "step": 887 }, { "epoch": 0.03722373012795657, "grad_norm": 1.9171456098556519, "learning_rate": 9.998636711600886e-06, "loss": 0.6425, "step": 888 }, { "epoch": 0.037265648742965535, "grad_norm": 2.097720146179199, "learning_rate": 9.998620814059177e-06, "loss": 0.6315, "step": 889 }, { "epoch": 0.03730756735797449, "grad_norm": 2.6054506301879883, "learning_rate": 9.998604824374804e-06, "loss": 0.6239, "step": 890 }, { "epoch": 0.03734948597298345, "grad_norm": 2.1212351322174072, "learning_rate": 9.998588742548064e-06, "loss": 0.6628, "step": 891 }, { "epoch": 0.037391404587992415, "grad_norm": 2.3125998973846436, "learning_rate": 9.99857256857925e-06, "loss": 0.6837, "step": 892 }, { "epoch": 0.03743332320300137, "grad_norm": 2.4319841861724854, "learning_rate": 9.998556302468658e-06, "loss": 0.6986, "step": 893 }, { "epoch": 0.03747524181801033, "grad_norm": 2.1367547512054443, "learning_rate": 9.998539944216594e-06, "loss": 0.6711, "step": 894 }, { "epoch": 0.037517160433019295, "grad_norm": 1.828328251838684, "learning_rate": 9.998523493823356e-06, "loss": 0.63, "step": 895 }, { "epoch": 0.03755907904802825, "grad_norm": 2.0193355083465576, "learning_rate": 9.998506951289246e-06, "loss": 0.5716, "step": 896 }, { "epoch": 0.03760099766303721, "grad_norm": 2.6404244899749756, "learning_rate": 9.998490316614572e-06, "loss": 0.5993, "step": 897 }, { "epoch": 0.037642916278046175, "grad_norm": 1.8520944118499756, "learning_rate": 9.998473589799638e-06, "loss": 0.5827, "step": 898 }, { "epoch": 0.03768483489305513, "grad_norm": 2.200181007385254, "learning_rate": 9.998456770844755e-06, "loss": 0.6664, "step": 899 }, { "epoch": 0.03772675350806409, "grad_norm": 2.537867307662964, "learning_rate": 9.998439859750231e-06, "loss": 0.6534, "step": 900 }, { "epoch": 0.037768672123073055, "grad_norm": 2.1963348388671875, "learning_rate": 9.998422856516378e-06, "loss": 0.7234, "step": 901 }, { "epoch": 0.03781059073808202, "grad_norm": 2.070733070373535, "learning_rate": 9.998405761143511e-06, "loss": 0.6896, "step": 902 }, { "epoch": 0.03785250935309097, "grad_norm": 2.1760902404785156, "learning_rate": 9.998388573631943e-06, "loss": 0.6237, "step": 903 }, { "epoch": 0.037894427968099935, "grad_norm": 1.9165219068527222, "learning_rate": 9.998371293981991e-06, "loss": 0.6381, "step": 904 }, { "epoch": 0.0379363465831089, "grad_norm": 2.0758018493652344, "learning_rate": 9.998353922193976e-06, "loss": 0.6278, "step": 905 }, { "epoch": 0.03797826519811785, "grad_norm": 2.194253444671631, "learning_rate": 9.998336458268213e-06, "loss": 0.591, "step": 906 }, { "epoch": 0.038020183813126815, "grad_norm": 2.325324296951294, "learning_rate": 9.998318902205031e-06, "loss": 0.6229, "step": 907 }, { "epoch": 0.03806210242813578, "grad_norm": 2.2922940254211426, "learning_rate": 9.99830125400475e-06, "loss": 0.7176, "step": 908 }, { "epoch": 0.03810402104314473, "grad_norm": 2.108165740966797, "learning_rate": 9.998283513667693e-06, "loss": 0.5962, "step": 909 }, { "epoch": 0.038145939658153695, "grad_norm": 1.9316000938415527, "learning_rate": 9.99826568119419e-06, "loss": 0.6489, "step": 910 }, { "epoch": 0.03818785827316266, "grad_norm": 5.856729984283447, "learning_rate": 9.99824775658457e-06, "loss": 0.6765, "step": 911 }, { "epoch": 0.03822977688817161, "grad_norm": 2.3831493854522705, "learning_rate": 9.998229739839162e-06, "loss": 0.5908, "step": 912 }, { "epoch": 0.038271695503180575, "grad_norm": 4.7122602462768555, "learning_rate": 9.998211630958299e-06, "loss": 0.6427, "step": 913 }, { "epoch": 0.03831361411818954, "grad_norm": 2.300511598587036, "learning_rate": 9.998193429942314e-06, "loss": 0.6369, "step": 914 }, { "epoch": 0.03835553273319849, "grad_norm": 1.9987046718597412, "learning_rate": 9.998175136791542e-06, "loss": 0.6851, "step": 915 }, { "epoch": 0.038397451348207455, "grad_norm": 1.8227089643478394, "learning_rate": 9.998156751506323e-06, "loss": 0.5636, "step": 916 }, { "epoch": 0.03843936996321642, "grad_norm": 2.324416160583496, "learning_rate": 9.998138274086993e-06, "loss": 0.6368, "step": 917 }, { "epoch": 0.03848128857822537, "grad_norm": 2.520789623260498, "learning_rate": 9.998119704533893e-06, "loss": 0.6435, "step": 918 }, { "epoch": 0.038523207193234335, "grad_norm": 2.594089984893799, "learning_rate": 9.998101042847368e-06, "loss": 0.6698, "step": 919 }, { "epoch": 0.0385651258082433, "grad_norm": 2.699319839477539, "learning_rate": 9.998082289027758e-06, "loss": 0.6201, "step": 920 }, { "epoch": 0.03860704442325225, "grad_norm": 1.9221361875534058, "learning_rate": 9.99806344307541e-06, "loss": 0.6228, "step": 921 }, { "epoch": 0.038648963038261215, "grad_norm": 4.1987223625183105, "learning_rate": 9.998044504990674e-06, "loss": 0.5916, "step": 922 }, { "epoch": 0.03869088165327018, "grad_norm": 2.0511937141418457, "learning_rate": 9.998025474773896e-06, "loss": 0.5728, "step": 923 }, { "epoch": 0.03873280026827913, "grad_norm": 2.385354995727539, "learning_rate": 9.998006352425428e-06, "loss": 0.6559, "step": 924 }, { "epoch": 0.038774718883288095, "grad_norm": 2.0336406230926514, "learning_rate": 9.997987137945623e-06, "loss": 0.6065, "step": 925 }, { "epoch": 0.03881663749829706, "grad_norm": 2.1945173740386963, "learning_rate": 9.997967831334834e-06, "loss": 0.6457, "step": 926 }, { "epoch": 0.03885855611330602, "grad_norm": 1.9292148351669312, "learning_rate": 9.997948432593419e-06, "loss": 0.6409, "step": 927 }, { "epoch": 0.038900474728314975, "grad_norm": 2.071479320526123, "learning_rate": 9.997928941721733e-06, "loss": 0.6532, "step": 928 }, { "epoch": 0.03894239334332394, "grad_norm": 2.0521202087402344, "learning_rate": 9.997909358720136e-06, "loss": 0.6671, "step": 929 }, { "epoch": 0.0389843119583329, "grad_norm": 2.0067126750946045, "learning_rate": 9.99788968358899e-06, "loss": 0.6426, "step": 930 }, { "epoch": 0.039026230573341855, "grad_norm": 2.2265005111694336, "learning_rate": 9.997869916328658e-06, "loss": 0.5826, "step": 931 }, { "epoch": 0.03906814918835082, "grad_norm": 2.11007022857666, "learning_rate": 9.9978500569395e-06, "loss": 0.5928, "step": 932 }, { "epoch": 0.03911006780335978, "grad_norm": 2.6718568801879883, "learning_rate": 9.997830105421888e-06, "loss": 0.6708, "step": 933 }, { "epoch": 0.039151986418368735, "grad_norm": 2.082888126373291, "learning_rate": 9.997810061776188e-06, "loss": 0.6466, "step": 934 }, { "epoch": 0.0391939050333777, "grad_norm": 2.0685365200042725, "learning_rate": 9.997789926002768e-06, "loss": 0.6043, "step": 935 }, { "epoch": 0.03923582364838666, "grad_norm": 2.2471463680267334, "learning_rate": 9.997769698102e-06, "loss": 0.6256, "step": 936 }, { "epoch": 0.039277742263395615, "grad_norm": 2.053041696548462, "learning_rate": 9.997749378074257e-06, "loss": 0.6462, "step": 937 }, { "epoch": 0.03931966087840458, "grad_norm": 2.614366054534912, "learning_rate": 9.997728965919912e-06, "loss": 0.6781, "step": 938 }, { "epoch": 0.03936157949341354, "grad_norm": 2.191256046295166, "learning_rate": 9.997708461639344e-06, "loss": 0.6805, "step": 939 }, { "epoch": 0.039403498108422495, "grad_norm": 2.4432551860809326, "learning_rate": 9.99768786523293e-06, "loss": 0.6528, "step": 940 }, { "epoch": 0.03944541672343146, "grad_norm": 2.2578818798065186, "learning_rate": 9.997667176701046e-06, "loss": 0.6496, "step": 941 }, { "epoch": 0.03948733533844042, "grad_norm": 2.4796969890594482, "learning_rate": 9.997646396044081e-06, "loss": 0.5965, "step": 942 }, { "epoch": 0.039529253953449375, "grad_norm": 3.9915294647216797, "learning_rate": 9.997625523262412e-06, "loss": 0.6373, "step": 943 }, { "epoch": 0.03957117256845834, "grad_norm": 1.9042834043502808, "learning_rate": 9.997604558356426e-06, "loss": 0.6198, "step": 944 }, { "epoch": 0.0396130911834673, "grad_norm": 6.074185371398926, "learning_rate": 9.997583501326508e-06, "loss": 0.6204, "step": 945 }, { "epoch": 0.039655009798476255, "grad_norm": 1.980825424194336, "learning_rate": 9.997562352173047e-06, "loss": 0.6598, "step": 946 }, { "epoch": 0.03969692841348522, "grad_norm": 2.0216293334960938, "learning_rate": 9.997541110896434e-06, "loss": 0.6515, "step": 947 }, { "epoch": 0.03973884702849418, "grad_norm": 1.7800298929214478, "learning_rate": 9.997519777497059e-06, "loss": 0.6442, "step": 948 }, { "epoch": 0.039780765643503135, "grad_norm": 2.4337451457977295, "learning_rate": 9.997498351975317e-06, "loss": 0.618, "step": 949 }, { "epoch": 0.0398226842585121, "grad_norm": 2.2343862056732178, "learning_rate": 9.997476834331598e-06, "loss": 0.6259, "step": 950 }, { "epoch": 0.03986460287352106, "grad_norm": 2.449711322784424, "learning_rate": 9.997455224566306e-06, "loss": 0.6106, "step": 951 }, { "epoch": 0.03990652148853002, "grad_norm": 1.8188543319702148, "learning_rate": 9.997433522679834e-06, "loss": 0.6306, "step": 952 }, { "epoch": 0.03994844010353898, "grad_norm": 1.8320624828338623, "learning_rate": 9.997411728672585e-06, "loss": 0.6038, "step": 953 }, { "epoch": 0.03999035871854794, "grad_norm": 2.2548768520355225, "learning_rate": 9.997389842544959e-06, "loss": 0.5763, "step": 954 }, { "epoch": 0.0400322773335569, "grad_norm": 2.9717891216278076, "learning_rate": 9.99736786429736e-06, "loss": 0.67, "step": 955 }, { "epoch": 0.04007419594856586, "grad_norm": 2.7783923149108887, "learning_rate": 9.997345793930193e-06, "loss": 0.6373, "step": 956 }, { "epoch": 0.04011611456357482, "grad_norm": 2.0498077869415283, "learning_rate": 9.997323631443865e-06, "loss": 0.6892, "step": 957 }, { "epoch": 0.04015803317858378, "grad_norm": 2.5229403972625732, "learning_rate": 9.997301376838784e-06, "loss": 0.6315, "step": 958 }, { "epoch": 0.04019995179359274, "grad_norm": 4.66330099105835, "learning_rate": 9.997279030115361e-06, "loss": 0.6377, "step": 959 }, { "epoch": 0.0402418704086017, "grad_norm": 1.921441674232483, "learning_rate": 9.997256591274006e-06, "loss": 0.5911, "step": 960 }, { "epoch": 0.04028378902361066, "grad_norm": 2.0007731914520264, "learning_rate": 9.997234060315137e-06, "loss": 0.6494, "step": 961 }, { "epoch": 0.04032570763861962, "grad_norm": 2.3910374641418457, "learning_rate": 9.997211437239164e-06, "loss": 0.6391, "step": 962 }, { "epoch": 0.04036762625362858, "grad_norm": 2.269239902496338, "learning_rate": 9.997188722046509e-06, "loss": 0.6309, "step": 963 }, { "epoch": 0.04040954486863754, "grad_norm": 2.431304693222046, "learning_rate": 9.997165914737587e-06, "loss": 0.6399, "step": 964 }, { "epoch": 0.0404514634836465, "grad_norm": 2.062647581100464, "learning_rate": 9.99714301531282e-06, "loss": 0.5574, "step": 965 }, { "epoch": 0.04049338209865546, "grad_norm": 2.0744240283966064, "learning_rate": 9.99712002377263e-06, "loss": 0.6077, "step": 966 }, { "epoch": 0.04053530071366442, "grad_norm": 4.451521873474121, "learning_rate": 9.99709694011744e-06, "loss": 0.6642, "step": 967 }, { "epoch": 0.04057721932867338, "grad_norm": 2.0042037963867188, "learning_rate": 9.997073764347677e-06, "loss": 0.6319, "step": 968 }, { "epoch": 0.04061913794368234, "grad_norm": 2.0429298877716064, "learning_rate": 9.997050496463767e-06, "loss": 0.6372, "step": 969 }, { "epoch": 0.0406610565586913, "grad_norm": 2.396198034286499, "learning_rate": 9.99702713646614e-06, "loss": 0.6294, "step": 970 }, { "epoch": 0.04070297517370026, "grad_norm": 2.1630561351776123, "learning_rate": 9.997003684355224e-06, "loss": 0.5553, "step": 971 }, { "epoch": 0.04074489378870922, "grad_norm": 1.9896035194396973, "learning_rate": 9.996980140131457e-06, "loss": 0.5903, "step": 972 }, { "epoch": 0.04078681240371818, "grad_norm": 2.3065271377563477, "learning_rate": 9.996956503795265e-06, "loss": 0.6016, "step": 973 }, { "epoch": 0.04082873101872714, "grad_norm": 2.1208674907684326, "learning_rate": 9.99693277534709e-06, "loss": 0.631, "step": 974 }, { "epoch": 0.0408706496337361, "grad_norm": 1.904353380203247, "learning_rate": 9.996908954787367e-06, "loss": 0.6293, "step": 975 }, { "epoch": 0.04091256824874506, "grad_norm": 2.2521729469299316, "learning_rate": 9.996885042116534e-06, "loss": 0.66, "step": 976 }, { "epoch": 0.040954486863754025, "grad_norm": 2.0567831993103027, "learning_rate": 9.996861037335035e-06, "loss": 0.7101, "step": 977 }, { "epoch": 0.04099640547876298, "grad_norm": 2.6107919216156006, "learning_rate": 9.996836940443311e-06, "loss": 0.6427, "step": 978 }, { "epoch": 0.04103832409377194, "grad_norm": 2.2128021717071533, "learning_rate": 9.996812751441806e-06, "loss": 0.6137, "step": 979 }, { "epoch": 0.041080242708780905, "grad_norm": 2.941446304321289, "learning_rate": 9.996788470330964e-06, "loss": 0.6489, "step": 980 }, { "epoch": 0.04112216132378986, "grad_norm": 2.0812060832977295, "learning_rate": 9.996764097111235e-06, "loss": 0.6512, "step": 981 }, { "epoch": 0.04116407993879882, "grad_norm": 2.48330020904541, "learning_rate": 9.99673963178307e-06, "loss": 0.6145, "step": 982 }, { "epoch": 0.041205998553807785, "grad_norm": 1.981804609298706, "learning_rate": 9.996715074346916e-06, "loss": 0.6187, "step": 983 }, { "epoch": 0.04124791716881674, "grad_norm": 2.110888957977295, "learning_rate": 9.996690424803226e-06, "loss": 0.6751, "step": 984 }, { "epoch": 0.0412898357838257, "grad_norm": 2.209141731262207, "learning_rate": 9.996665683152458e-06, "loss": 0.6049, "step": 985 }, { "epoch": 0.041331754398834665, "grad_norm": 2.005582332611084, "learning_rate": 9.996640849395066e-06, "loss": 0.6105, "step": 986 }, { "epoch": 0.04137367301384362, "grad_norm": 2.084472179412842, "learning_rate": 9.996615923531506e-06, "loss": 0.7045, "step": 987 }, { "epoch": 0.04141559162885258, "grad_norm": 2.0114452838897705, "learning_rate": 9.996590905562237e-06, "loss": 0.6849, "step": 988 }, { "epoch": 0.041457510243861545, "grad_norm": 1.982556939125061, "learning_rate": 9.996565795487727e-06, "loss": 0.6051, "step": 989 }, { "epoch": 0.0414994288588705, "grad_norm": 2.0266144275665283, "learning_rate": 9.996540593308432e-06, "loss": 0.6106, "step": 990 }, { "epoch": 0.04154134747387946, "grad_norm": 2.17915940284729, "learning_rate": 9.996515299024819e-06, "loss": 0.6251, "step": 991 }, { "epoch": 0.041583266088888425, "grad_norm": 1.9621455669403076, "learning_rate": 9.996489912637353e-06, "loss": 0.6608, "step": 992 }, { "epoch": 0.04162518470389738, "grad_norm": 2.7644379138946533, "learning_rate": 9.996464434146501e-06, "loss": 0.6456, "step": 993 }, { "epoch": 0.04166710331890634, "grad_norm": 2.5632481575012207, "learning_rate": 9.996438863552738e-06, "loss": 0.5511, "step": 994 }, { "epoch": 0.041709021933915305, "grad_norm": 2.063150644302368, "learning_rate": 9.996413200856532e-06, "loss": 0.6421, "step": 995 }, { "epoch": 0.04175094054892426, "grad_norm": 2.2503323554992676, "learning_rate": 9.996387446058354e-06, "loss": 0.5667, "step": 996 }, { "epoch": 0.04179285916393322, "grad_norm": 2.0378832817077637, "learning_rate": 9.996361599158682e-06, "loss": 0.6401, "step": 997 }, { "epoch": 0.041834777778942185, "grad_norm": 1.9427117109298706, "learning_rate": 9.996335660157989e-06, "loss": 0.6278, "step": 998 }, { "epoch": 0.04187669639395114, "grad_norm": 1.8632545471191406, "learning_rate": 9.99630962905676e-06, "loss": 0.6301, "step": 999 }, { "epoch": 0.0419186150089601, "grad_norm": 3.069451093673706, "learning_rate": 9.996283505855467e-06, "loss": 0.6016, "step": 1000 }, { "epoch": 0.041960533623969065, "grad_norm": 2.8802921772003174, "learning_rate": 9.996257290554593e-06, "loss": 0.6534, "step": 1001 }, { "epoch": 0.04200245223897803, "grad_norm": 1.844536542892456, "learning_rate": 9.996230983154626e-06, "loss": 0.6292, "step": 1002 }, { "epoch": 0.04204437085398698, "grad_norm": 2.0763580799102783, "learning_rate": 9.996204583656047e-06, "loss": 0.6387, "step": 1003 }, { "epoch": 0.042086289468995945, "grad_norm": 2.1079463958740234, "learning_rate": 9.996178092059343e-06, "loss": 0.6788, "step": 1004 }, { "epoch": 0.04212820808400491, "grad_norm": 2.2011196613311768, "learning_rate": 9.996151508365003e-06, "loss": 0.6632, "step": 1005 }, { "epoch": 0.04217012669901386, "grad_norm": 2.18424916267395, "learning_rate": 9.996124832573517e-06, "loss": 0.678, "step": 1006 }, { "epoch": 0.042212045314022825, "grad_norm": 2.021512746810913, "learning_rate": 9.996098064685377e-06, "loss": 0.6061, "step": 1007 }, { "epoch": 0.04225396392903179, "grad_norm": 2.06307315826416, "learning_rate": 9.996071204701077e-06, "loss": 0.6264, "step": 1008 }, { "epoch": 0.04229588254404074, "grad_norm": 2.1571643352508545, "learning_rate": 9.99604425262111e-06, "loss": 0.5168, "step": 1009 }, { "epoch": 0.042337801159049705, "grad_norm": 2.09653902053833, "learning_rate": 9.996017208445972e-06, "loss": 0.6804, "step": 1010 }, { "epoch": 0.04237971977405867, "grad_norm": 2.2900936603546143, "learning_rate": 9.995990072176167e-06, "loss": 0.6203, "step": 1011 }, { "epoch": 0.04242163838906762, "grad_norm": 2.1626384258270264, "learning_rate": 9.99596284381219e-06, "loss": 0.6028, "step": 1012 }, { "epoch": 0.042463557004076585, "grad_norm": 2.124253034591675, "learning_rate": 9.995935523354546e-06, "loss": 0.6744, "step": 1013 }, { "epoch": 0.04250547561908555, "grad_norm": 1.9403799772262573, "learning_rate": 9.995908110803736e-06, "loss": 0.5924, "step": 1014 }, { "epoch": 0.0425473942340945, "grad_norm": 2.127476215362549, "learning_rate": 9.995880606160266e-06, "loss": 0.6791, "step": 1015 }, { "epoch": 0.042589312849103465, "grad_norm": 2.028974771499634, "learning_rate": 9.995853009424646e-06, "loss": 0.6026, "step": 1016 }, { "epoch": 0.04263123146411243, "grad_norm": 2.0882980823516846, "learning_rate": 9.995825320597381e-06, "loss": 0.6301, "step": 1017 }, { "epoch": 0.04267315007912138, "grad_norm": 2.6975317001342773, "learning_rate": 9.995797539678982e-06, "loss": 0.6341, "step": 1018 }, { "epoch": 0.042715068694130345, "grad_norm": 2.1981918811798096, "learning_rate": 9.995769666669963e-06, "loss": 0.6217, "step": 1019 }, { "epoch": 0.04275698730913931, "grad_norm": 2.3533995151519775, "learning_rate": 9.995741701570836e-06, "loss": 0.6119, "step": 1020 }, { "epoch": 0.04279890592414826, "grad_norm": 2.3561511039733887, "learning_rate": 9.995713644382118e-06, "loss": 0.6791, "step": 1021 }, { "epoch": 0.042840824539157225, "grad_norm": 3.067324161529541, "learning_rate": 9.995685495104324e-06, "loss": 0.6621, "step": 1022 }, { "epoch": 0.04288274315416619, "grad_norm": 2.0364584922790527, "learning_rate": 9.995657253737975e-06, "loss": 0.6283, "step": 1023 }, { "epoch": 0.04292466176917514, "grad_norm": 2.288529634475708, "learning_rate": 9.99562892028359e-06, "loss": 0.6211, "step": 1024 }, { "epoch": 0.042966580384184105, "grad_norm": 1.9871643781661987, "learning_rate": 9.995600494741693e-06, "loss": 0.6121, "step": 1025 }, { "epoch": 0.04300849899919307, "grad_norm": 1.995348334312439, "learning_rate": 9.995571977112809e-06, "loss": 0.5868, "step": 1026 }, { "epoch": 0.04305041761420203, "grad_norm": 2.2786927223205566, "learning_rate": 9.995543367397458e-06, "loss": 0.6211, "step": 1027 }, { "epoch": 0.043092336229210985, "grad_norm": 1.7956774234771729, "learning_rate": 9.995514665596174e-06, "loss": 0.5763, "step": 1028 }, { "epoch": 0.04313425484421995, "grad_norm": 2.1730258464813232, "learning_rate": 9.99548587170948e-06, "loss": 0.6138, "step": 1029 }, { "epoch": 0.04317617345922891, "grad_norm": 2.154412269592285, "learning_rate": 9.995456985737912e-06, "loss": 0.6377, "step": 1030 }, { "epoch": 0.043218092074237865, "grad_norm": 1.994537353515625, "learning_rate": 9.995428007682002e-06, "loss": 0.5923, "step": 1031 }, { "epoch": 0.04326001068924683, "grad_norm": 3.485304117202759, "learning_rate": 9.99539893754228e-06, "loss": 0.6427, "step": 1032 }, { "epoch": 0.04330192930425579, "grad_norm": 2.0804967880249023, "learning_rate": 9.995369775319285e-06, "loss": 0.6332, "step": 1033 }, { "epoch": 0.043343847919264746, "grad_norm": 2.177534341812134, "learning_rate": 9.995340521013557e-06, "loss": 0.6733, "step": 1034 }, { "epoch": 0.04338576653427371, "grad_norm": 1.9438097476959229, "learning_rate": 9.99531117462563e-06, "loss": 0.6669, "step": 1035 }, { "epoch": 0.04342768514928267, "grad_norm": 2.351550579071045, "learning_rate": 9.995281736156047e-06, "loss": 0.6261, "step": 1036 }, { "epoch": 0.043469603764291626, "grad_norm": 2.133162021636963, "learning_rate": 9.995252205605352e-06, "loss": 0.5859, "step": 1037 }, { "epoch": 0.04351152237930059, "grad_norm": 2.0393927097320557, "learning_rate": 9.995222582974087e-06, "loss": 0.5746, "step": 1038 }, { "epoch": 0.04355344099430955, "grad_norm": 2.347597599029541, "learning_rate": 9.9951928682628e-06, "loss": 0.6132, "step": 1039 }, { "epoch": 0.043595359609318506, "grad_norm": 1.946039080619812, "learning_rate": 9.995163061472038e-06, "loss": 0.582, "step": 1040 }, { "epoch": 0.04363727822432747, "grad_norm": 1.8698797225952148, "learning_rate": 9.99513316260235e-06, "loss": 0.5678, "step": 1041 }, { "epoch": 0.04367919683933643, "grad_norm": 2.1470561027526855, "learning_rate": 9.99510317165429e-06, "loss": 0.6572, "step": 1042 }, { "epoch": 0.043721115454345386, "grad_norm": 1.869652509689331, "learning_rate": 9.995073088628406e-06, "loss": 0.5678, "step": 1043 }, { "epoch": 0.04376303406935435, "grad_norm": 2.126990795135498, "learning_rate": 9.995042913525257e-06, "loss": 0.636, "step": 1044 }, { "epoch": 0.04380495268436331, "grad_norm": 2.3576486110687256, "learning_rate": 9.995012646345397e-06, "loss": 0.6778, "step": 1045 }, { "epoch": 0.043846871299372266, "grad_norm": 1.9637084007263184, "learning_rate": 9.994982287089384e-06, "loss": 0.689, "step": 1046 }, { "epoch": 0.04388878991438123, "grad_norm": 2.85406231880188, "learning_rate": 9.994951835757778e-06, "loss": 0.68, "step": 1047 }, { "epoch": 0.04393070852939019, "grad_norm": 2.0926449298858643, "learning_rate": 9.994921292351139e-06, "loss": 0.675, "step": 1048 }, { "epoch": 0.043972627144399146, "grad_norm": 1.9877572059631348, "learning_rate": 9.994890656870031e-06, "loss": 0.6117, "step": 1049 }, { "epoch": 0.04401454575940811, "grad_norm": 1.8922791481018066, "learning_rate": 9.994859929315021e-06, "loss": 0.5768, "step": 1050 }, { "epoch": 0.04405646437441707, "grad_norm": 2.4085676670074463, "learning_rate": 9.994829109686673e-06, "loss": 0.6262, "step": 1051 }, { "epoch": 0.04409838298942603, "grad_norm": 1.9577785730361938, "learning_rate": 9.994798197985555e-06, "loss": 0.6639, "step": 1052 }, { "epoch": 0.04414030160443499, "grad_norm": 1.777238368988037, "learning_rate": 9.994767194212237e-06, "loss": 0.6041, "step": 1053 }, { "epoch": 0.04418222021944395, "grad_norm": 2.2575814723968506, "learning_rate": 9.994736098367291e-06, "loss": 0.585, "step": 1054 }, { "epoch": 0.04422413883445291, "grad_norm": 2.174894332885742, "learning_rate": 9.994704910451289e-06, "loss": 0.6474, "step": 1055 }, { "epoch": 0.04426605744946187, "grad_norm": 1.9328409433364868, "learning_rate": 9.994673630464809e-06, "loss": 0.7042, "step": 1056 }, { "epoch": 0.04430797606447083, "grad_norm": 1.8970657587051392, "learning_rate": 9.994642258408425e-06, "loss": 0.6076, "step": 1057 }, { "epoch": 0.04434989467947979, "grad_norm": 2.1585581302642822, "learning_rate": 9.994610794282715e-06, "loss": 0.6344, "step": 1058 }, { "epoch": 0.04439181329448875, "grad_norm": 1.962539553642273, "learning_rate": 9.99457923808826e-06, "loss": 0.631, "step": 1059 }, { "epoch": 0.04443373190949771, "grad_norm": 1.7902785539627075, "learning_rate": 9.994547589825642e-06, "loss": 0.6082, "step": 1060 }, { "epoch": 0.04447565052450667, "grad_norm": 2.2888734340667725, "learning_rate": 9.994515849495445e-06, "loss": 0.6295, "step": 1061 }, { "epoch": 0.04451756913951563, "grad_norm": 1.8478760719299316, "learning_rate": 9.994484017098251e-06, "loss": 0.5548, "step": 1062 }, { "epoch": 0.04455948775452459, "grad_norm": 2.479475736618042, "learning_rate": 9.99445209263465e-06, "loss": 0.5597, "step": 1063 }, { "epoch": 0.04460140636953355, "grad_norm": 2.7640035152435303, "learning_rate": 9.99442007610523e-06, "loss": 0.644, "step": 1064 }, { "epoch": 0.04464332498454251, "grad_norm": 2.2510435581207275, "learning_rate": 9.99438796751058e-06, "loss": 0.592, "step": 1065 }, { "epoch": 0.04468524359955147, "grad_norm": 2.368692398071289, "learning_rate": 9.99435576685129e-06, "loss": 0.573, "step": 1066 }, { "epoch": 0.04472716221456043, "grad_norm": 1.9845982789993286, "learning_rate": 9.994323474127957e-06, "loss": 0.6268, "step": 1067 }, { "epoch": 0.04476908082956939, "grad_norm": 2.552448272705078, "learning_rate": 9.994291089341177e-06, "loss": 0.6324, "step": 1068 }, { "epoch": 0.04481099944457835, "grad_norm": 3.058661937713623, "learning_rate": 9.994258612491544e-06, "loss": 0.6461, "step": 1069 }, { "epoch": 0.04485291805958731, "grad_norm": 1.9615867137908936, "learning_rate": 9.994226043579658e-06, "loss": 0.5603, "step": 1070 }, { "epoch": 0.04489483667459627, "grad_norm": 3.7914857864379883, "learning_rate": 9.994193382606118e-06, "loss": 0.6099, "step": 1071 }, { "epoch": 0.04493675528960523, "grad_norm": 1.812886118888855, "learning_rate": 9.99416062957153e-06, "loss": 0.5827, "step": 1072 }, { "epoch": 0.04497867390461419, "grad_norm": 1.9697136878967285, "learning_rate": 9.994127784476493e-06, "loss": 0.6612, "step": 1073 }, { "epoch": 0.04502059251962315, "grad_norm": 2.405611515045166, "learning_rate": 9.994094847321616e-06, "loss": 0.6351, "step": 1074 }, { "epoch": 0.04506251113463211, "grad_norm": 1.904634952545166, "learning_rate": 9.994061818107504e-06, "loss": 0.6598, "step": 1075 }, { "epoch": 0.04510442974964107, "grad_norm": 2.1117782592773438, "learning_rate": 9.994028696834764e-06, "loss": 0.6445, "step": 1076 }, { "epoch": 0.045146348364650035, "grad_norm": 7.411186695098877, "learning_rate": 9.99399548350401e-06, "loss": 0.6495, "step": 1077 }, { "epoch": 0.04518826697965899, "grad_norm": 1.905332088470459, "learning_rate": 9.993962178115856e-06, "loss": 0.5965, "step": 1078 }, { "epoch": 0.04523018559466795, "grad_norm": 2.819453239440918, "learning_rate": 9.993928780670912e-06, "loss": 0.6053, "step": 1079 }, { "epoch": 0.045272104209676915, "grad_norm": 2.178849697113037, "learning_rate": 9.993895291169792e-06, "loss": 0.6664, "step": 1080 }, { "epoch": 0.04531402282468587, "grad_norm": 2.0287623405456543, "learning_rate": 9.99386170961312e-06, "loss": 0.6974, "step": 1081 }, { "epoch": 0.04535594143969483, "grad_norm": 1.8999123573303223, "learning_rate": 9.99382803600151e-06, "loss": 0.5858, "step": 1082 }, { "epoch": 0.045397860054703795, "grad_norm": 2.24186635017395, "learning_rate": 9.993794270335582e-06, "loss": 0.6711, "step": 1083 }, { "epoch": 0.04543977866971275, "grad_norm": 3.0034432411193848, "learning_rate": 9.993760412615962e-06, "loss": 0.6183, "step": 1084 }, { "epoch": 0.04548169728472171, "grad_norm": 2.0216448307037354, "learning_rate": 9.993726462843273e-06, "loss": 0.7291, "step": 1085 }, { "epoch": 0.045523615899730675, "grad_norm": 2.080998182296753, "learning_rate": 9.99369242101814e-06, "loss": 0.6138, "step": 1086 }, { "epoch": 0.04556553451473963, "grad_norm": 2.398409366607666, "learning_rate": 9.99365828714119e-06, "loss": 0.6816, "step": 1087 }, { "epoch": 0.04560745312974859, "grad_norm": 2.37554669380188, "learning_rate": 9.993624061213054e-06, "loss": 0.6751, "step": 1088 }, { "epoch": 0.045649371744757555, "grad_norm": 1.9067797660827637, "learning_rate": 9.99358974323436e-06, "loss": 0.6467, "step": 1089 }, { "epoch": 0.04569129035976651, "grad_norm": 1.9838926792144775, "learning_rate": 9.993555333205745e-06, "loss": 0.6125, "step": 1090 }, { "epoch": 0.04573320897477547, "grad_norm": 2.5507521629333496, "learning_rate": 9.993520831127838e-06, "loss": 0.6272, "step": 1091 }, { "epoch": 0.045775127589784435, "grad_norm": 1.9268264770507812, "learning_rate": 9.99348623700128e-06, "loss": 0.6366, "step": 1092 }, { "epoch": 0.04581704620479339, "grad_norm": 1.9477548599243164, "learning_rate": 9.993451550826707e-06, "loss": 0.6526, "step": 1093 }, { "epoch": 0.04585896481980235, "grad_norm": 1.7760612964630127, "learning_rate": 9.993416772604755e-06, "loss": 0.6931, "step": 1094 }, { "epoch": 0.045900883434811315, "grad_norm": 2.166259765625, "learning_rate": 9.993381902336072e-06, "loss": 0.6508, "step": 1095 }, { "epoch": 0.04594280204982027, "grad_norm": 2.222029685974121, "learning_rate": 9.993346940021294e-06, "loss": 0.6319, "step": 1096 }, { "epoch": 0.04598472066482923, "grad_norm": 2.081387758255005, "learning_rate": 9.99331188566107e-06, "loss": 0.6465, "step": 1097 }, { "epoch": 0.046026639279838195, "grad_norm": 1.890896201133728, "learning_rate": 9.993276739256042e-06, "loss": 0.6023, "step": 1098 }, { "epoch": 0.04606855789484715, "grad_norm": 2.581195831298828, "learning_rate": 9.993241500806861e-06, "loss": 0.6156, "step": 1099 }, { "epoch": 0.04611047650985611, "grad_norm": 1.9135334491729736, "learning_rate": 9.993206170314178e-06, "loss": 0.6356, "step": 1100 }, { "epoch": 0.046152395124865075, "grad_norm": 2.375643491744995, "learning_rate": 9.99317074777864e-06, "loss": 0.6231, "step": 1101 }, { "epoch": 0.04619431373987404, "grad_norm": 2.099245548248291, "learning_rate": 9.993135233200903e-06, "loss": 0.6298, "step": 1102 }, { "epoch": 0.04623623235488299, "grad_norm": 1.7767843008041382, "learning_rate": 9.993099626581621e-06, "loss": 0.6004, "step": 1103 }, { "epoch": 0.046278150969891955, "grad_norm": 2.076544761657715, "learning_rate": 9.99306392792145e-06, "loss": 0.6094, "step": 1104 }, { "epoch": 0.04632006958490092, "grad_norm": 2.256321907043457, "learning_rate": 9.993028137221047e-06, "loss": 0.5972, "step": 1105 }, { "epoch": 0.04636198819990987, "grad_norm": 2.3097305297851562, "learning_rate": 9.992992254481073e-06, "loss": 0.6594, "step": 1106 }, { "epoch": 0.046403906814918836, "grad_norm": 2.317251205444336, "learning_rate": 9.99295627970219e-06, "loss": 0.6178, "step": 1107 }, { "epoch": 0.0464458254299278, "grad_norm": 2.1791155338287354, "learning_rate": 9.992920212885062e-06, "loss": 0.5902, "step": 1108 }, { "epoch": 0.04648774404493675, "grad_norm": 2.019660711288452, "learning_rate": 9.992884054030351e-06, "loss": 0.6361, "step": 1109 }, { "epoch": 0.046529662659945716, "grad_norm": 4.299177646636963, "learning_rate": 9.992847803138725e-06, "loss": 0.5925, "step": 1110 }, { "epoch": 0.04657158127495468, "grad_norm": 2.0206100940704346, "learning_rate": 9.992811460210851e-06, "loss": 0.66, "step": 1111 }, { "epoch": 0.04661349988996363, "grad_norm": 2.428382635116577, "learning_rate": 9.992775025247403e-06, "loss": 0.701, "step": 1112 }, { "epoch": 0.046655418504972596, "grad_norm": 1.9598227739334106, "learning_rate": 9.992738498249048e-06, "loss": 0.6232, "step": 1113 }, { "epoch": 0.04669733711998156, "grad_norm": 2.0380563735961914, "learning_rate": 9.99270187921646e-06, "loss": 0.682, "step": 1114 }, { "epoch": 0.04673925573499051, "grad_norm": 1.9752901792526245, "learning_rate": 9.992665168150317e-06, "loss": 0.6176, "step": 1115 }, { "epoch": 0.046781174349999476, "grad_norm": 1.9655952453613281, "learning_rate": 9.992628365051293e-06, "loss": 0.6305, "step": 1116 }, { "epoch": 0.04682309296500844, "grad_norm": 1.9678301811218262, "learning_rate": 9.992591469920066e-06, "loss": 0.686, "step": 1117 }, { "epoch": 0.04686501158001739, "grad_norm": 1.8623244762420654, "learning_rate": 9.99255448275732e-06, "loss": 0.6694, "step": 1118 }, { "epoch": 0.046906930195026356, "grad_norm": 1.9267112016677856, "learning_rate": 9.992517403563731e-06, "loss": 0.6035, "step": 1119 }, { "epoch": 0.04694884881003532, "grad_norm": 1.8859097957611084, "learning_rate": 9.992480232339987e-06, "loss": 0.6205, "step": 1120 }, { "epoch": 0.04699076742504427, "grad_norm": 2.091428279876709, "learning_rate": 9.992442969086772e-06, "loss": 0.6004, "step": 1121 }, { "epoch": 0.047032686040053236, "grad_norm": 2.4798660278320312, "learning_rate": 9.992405613804773e-06, "loss": 0.6825, "step": 1122 }, { "epoch": 0.0470746046550622, "grad_norm": 2.1905014514923096, "learning_rate": 9.992368166494678e-06, "loss": 0.6006, "step": 1123 }, { "epoch": 0.04711652327007116, "grad_norm": 1.7403011322021484, "learning_rate": 9.992330627157177e-06, "loss": 0.6093, "step": 1124 }, { "epoch": 0.047158441885080116, "grad_norm": 2.3027503490448, "learning_rate": 9.992292995792964e-06, "loss": 0.6629, "step": 1125 }, { "epoch": 0.04720036050008908, "grad_norm": 1.8752154111862183, "learning_rate": 9.99225527240273e-06, "loss": 0.5833, "step": 1126 }, { "epoch": 0.04724227911509804, "grad_norm": 2.0070784091949463, "learning_rate": 9.992217456987172e-06, "loss": 0.603, "step": 1127 }, { "epoch": 0.047284197730106996, "grad_norm": 2.1037521362304688, "learning_rate": 9.992179549546987e-06, "loss": 0.6394, "step": 1128 }, { "epoch": 0.04732611634511596, "grad_norm": 2.331955909729004, "learning_rate": 9.992141550082872e-06, "loss": 0.66, "step": 1129 }, { "epoch": 0.04736803496012492, "grad_norm": 2.249666690826416, "learning_rate": 9.992103458595532e-06, "loss": 0.5943, "step": 1130 }, { "epoch": 0.047409953575133876, "grad_norm": 2.205582857131958, "learning_rate": 9.992065275085662e-06, "loss": 0.6247, "step": 1131 }, { "epoch": 0.04745187219014284, "grad_norm": 2.347663164138794, "learning_rate": 9.992026999553974e-06, "loss": 0.6075, "step": 1132 }, { "epoch": 0.0474937908051518, "grad_norm": 1.9325251579284668, "learning_rate": 9.991988632001168e-06, "loss": 0.5973, "step": 1133 }, { "epoch": 0.047535709420160756, "grad_norm": 2.1144466400146484, "learning_rate": 9.991950172427951e-06, "loss": 0.5684, "step": 1134 }, { "epoch": 0.04757762803516972, "grad_norm": 2.1840922832489014, "learning_rate": 9.991911620835035e-06, "loss": 0.6225, "step": 1135 }, { "epoch": 0.04761954665017868, "grad_norm": 2.2210495471954346, "learning_rate": 9.991872977223131e-06, "loss": 0.5908, "step": 1136 }, { "epoch": 0.047661465265187636, "grad_norm": 2.359999418258667, "learning_rate": 9.991834241592947e-06, "loss": 0.5559, "step": 1137 }, { "epoch": 0.0477033838801966, "grad_norm": 2.3766582012176514, "learning_rate": 9.991795413945202e-06, "loss": 0.5898, "step": 1138 }, { "epoch": 0.04774530249520556, "grad_norm": 1.998710036277771, "learning_rate": 9.991756494280609e-06, "loss": 0.6176, "step": 1139 }, { "epoch": 0.047787221110214516, "grad_norm": 2.089348554611206, "learning_rate": 9.991717482599886e-06, "loss": 0.6787, "step": 1140 }, { "epoch": 0.04782913972522348, "grad_norm": 2.0354106426239014, "learning_rate": 9.99167837890375e-06, "loss": 0.6074, "step": 1141 }, { "epoch": 0.04787105834023244, "grad_norm": 2.54428768157959, "learning_rate": 9.991639183192928e-06, "loss": 0.603, "step": 1142 }, { "epoch": 0.047912976955241396, "grad_norm": 2.1968533992767334, "learning_rate": 9.991599895468134e-06, "loss": 0.6735, "step": 1143 }, { "epoch": 0.04795489557025036, "grad_norm": 2.0636916160583496, "learning_rate": 9.9915605157301e-06, "loss": 0.6159, "step": 1144 }, { "epoch": 0.04799681418525932, "grad_norm": 2.1939308643341064, "learning_rate": 9.991521043979546e-06, "loss": 0.5927, "step": 1145 }, { "epoch": 0.048038732800268276, "grad_norm": 2.290015459060669, "learning_rate": 9.991481480217204e-06, "loss": 0.5916, "step": 1146 }, { "epoch": 0.04808065141527724, "grad_norm": 2.076064348220825, "learning_rate": 9.9914418244438e-06, "loss": 0.5665, "step": 1147 }, { "epoch": 0.0481225700302862, "grad_norm": 2.1515698432922363, "learning_rate": 9.991402076660066e-06, "loss": 0.6264, "step": 1148 }, { "epoch": 0.04816448864529516, "grad_norm": 1.943031907081604, "learning_rate": 9.991362236866738e-06, "loss": 0.5904, "step": 1149 }, { "epoch": 0.04820640726030412, "grad_norm": 1.9283581972122192, "learning_rate": 9.991322305064544e-06, "loss": 0.674, "step": 1150 }, { "epoch": 0.04824832587531308, "grad_norm": 2.0417697429656982, "learning_rate": 9.991282281254227e-06, "loss": 0.6303, "step": 1151 }, { "epoch": 0.04829024449032204, "grad_norm": 2.523505449295044, "learning_rate": 9.991242165436519e-06, "loss": 0.7049, "step": 1152 }, { "epoch": 0.048332163105331, "grad_norm": 1.995453953742981, "learning_rate": 9.991201957612163e-06, "loss": 0.6078, "step": 1153 }, { "epoch": 0.04837408172033996, "grad_norm": 1.926548719406128, "learning_rate": 9.991161657781899e-06, "loss": 0.5481, "step": 1154 }, { "epoch": 0.04841600033534892, "grad_norm": 2.3380024433135986, "learning_rate": 9.991121265946469e-06, "loss": 0.5858, "step": 1155 }, { "epoch": 0.04845791895035788, "grad_norm": 2.2467124462127686, "learning_rate": 9.991080782106618e-06, "loss": 0.5967, "step": 1156 }, { "epoch": 0.04849983756536684, "grad_norm": 2.179980516433716, "learning_rate": 9.991040206263094e-06, "loss": 0.6373, "step": 1157 }, { "epoch": 0.0485417561803758, "grad_norm": 2.273350954055786, "learning_rate": 9.990999538416644e-06, "loss": 0.6545, "step": 1158 }, { "epoch": 0.04858367479538476, "grad_norm": 2.4673960208892822, "learning_rate": 9.990958778568016e-06, "loss": 0.6219, "step": 1159 }, { "epoch": 0.04862559341039372, "grad_norm": 2.486846923828125, "learning_rate": 9.990917926717962e-06, "loss": 0.6563, "step": 1160 }, { "epoch": 0.04866751202540268, "grad_norm": 1.8913716077804565, "learning_rate": 9.990876982867237e-06, "loss": 0.5623, "step": 1161 }, { "epoch": 0.04870943064041164, "grad_norm": 2.228980541229248, "learning_rate": 9.990835947016596e-06, "loss": 0.6817, "step": 1162 }, { "epoch": 0.0487513492554206, "grad_norm": 3.0874855518341064, "learning_rate": 9.990794819166793e-06, "loss": 0.5911, "step": 1163 }, { "epoch": 0.04879326787042956, "grad_norm": 2.216672420501709, "learning_rate": 9.990753599318586e-06, "loss": 0.6475, "step": 1164 }, { "epoch": 0.04883518648543852, "grad_norm": 1.991873860359192, "learning_rate": 9.990712287472737e-06, "loss": 0.6397, "step": 1165 }, { "epoch": 0.04887710510044748, "grad_norm": 1.9528381824493408, "learning_rate": 9.990670883630006e-06, "loss": 0.615, "step": 1166 }, { "epoch": 0.04891902371545644, "grad_norm": 2.015815019607544, "learning_rate": 9.990629387791158e-06, "loss": 0.6984, "step": 1167 }, { "epoch": 0.0489609423304654, "grad_norm": 1.8780571222305298, "learning_rate": 9.990587799956955e-06, "loss": 0.5905, "step": 1168 }, { "epoch": 0.04900286094547436, "grad_norm": 2.1173012256622314, "learning_rate": 9.990546120128164e-06, "loss": 0.6473, "step": 1169 }, { "epoch": 0.04904477956048332, "grad_norm": 2.0327606201171875, "learning_rate": 9.990504348305557e-06, "loss": 0.6211, "step": 1170 }, { "epoch": 0.04908669817549228, "grad_norm": 2.2667288780212402, "learning_rate": 9.9904624844899e-06, "loss": 0.6064, "step": 1171 }, { "epoch": 0.04912861679050124, "grad_norm": 1.9518396854400635, "learning_rate": 9.990420528681967e-06, "loss": 0.6536, "step": 1172 }, { "epoch": 0.0491705354055102, "grad_norm": 2.524078607559204, "learning_rate": 9.99037848088253e-06, "loss": 0.6299, "step": 1173 }, { "epoch": 0.049212454020519165, "grad_norm": 1.7728079557418823, "learning_rate": 9.990336341092365e-06, "loss": 0.6067, "step": 1174 }, { "epoch": 0.04925437263552812, "grad_norm": 3.555034875869751, "learning_rate": 9.990294109312248e-06, "loss": 0.6067, "step": 1175 }, { "epoch": 0.04929629125053708, "grad_norm": 1.9354556798934937, "learning_rate": 9.990251785542959e-06, "loss": 0.6156, "step": 1176 }, { "epoch": 0.049338209865546045, "grad_norm": 2.2311172485351562, "learning_rate": 9.990209369785277e-06, "loss": 0.5655, "step": 1177 }, { "epoch": 0.049380128480555, "grad_norm": 2.203660726547241, "learning_rate": 9.990166862039982e-06, "loss": 0.5984, "step": 1178 }, { "epoch": 0.04942204709556396, "grad_norm": 2.20307993888855, "learning_rate": 9.990124262307862e-06, "loss": 0.5835, "step": 1179 }, { "epoch": 0.049463965710572926, "grad_norm": 2.044196367263794, "learning_rate": 9.990081570589699e-06, "loss": 0.6469, "step": 1180 }, { "epoch": 0.04950588432558188, "grad_norm": 2.2629971504211426, "learning_rate": 9.990038786886281e-06, "loss": 0.6209, "step": 1181 }, { "epoch": 0.04954780294059084, "grad_norm": 4.178192138671875, "learning_rate": 9.989995911198397e-06, "loss": 0.6647, "step": 1182 }, { "epoch": 0.049589721555599806, "grad_norm": 2.6004273891448975, "learning_rate": 9.989952943526835e-06, "loss": 0.6237, "step": 1183 }, { "epoch": 0.04963164017060876, "grad_norm": 1.7885719537734985, "learning_rate": 9.98990988387239e-06, "loss": 0.5506, "step": 1184 }, { "epoch": 0.04967355878561772, "grad_norm": 1.9313968420028687, "learning_rate": 9.989866732235855e-06, "loss": 0.5973, "step": 1185 }, { "epoch": 0.049715477400626686, "grad_norm": 1.9803202152252197, "learning_rate": 9.989823488618025e-06, "loss": 0.6526, "step": 1186 }, { "epoch": 0.04975739601563564, "grad_norm": 2.1742217540740967, "learning_rate": 9.989780153019698e-06, "loss": 0.5985, "step": 1187 }, { "epoch": 0.0497993146306446, "grad_norm": 2.18009614944458, "learning_rate": 9.98973672544167e-06, "loss": 0.6238, "step": 1188 }, { "epoch": 0.049841233245653566, "grad_norm": 2.0900304317474365, "learning_rate": 9.989693205884745e-06, "loss": 0.6484, "step": 1189 }, { "epoch": 0.04988315186066252, "grad_norm": 1.9218544960021973, "learning_rate": 9.989649594349725e-06, "loss": 0.6832, "step": 1190 }, { "epoch": 0.04992507047567148, "grad_norm": 2.0665152072906494, "learning_rate": 9.98960589083741e-06, "loss": 0.6237, "step": 1191 }, { "epoch": 0.049966989090680446, "grad_norm": 1.8591840267181396, "learning_rate": 9.98956209534861e-06, "loss": 0.6137, "step": 1192 }, { "epoch": 0.0500089077056894, "grad_norm": 1.9862662553787231, "learning_rate": 9.98951820788413e-06, "loss": 0.6424, "step": 1193 }, { "epoch": 0.05005082632069836, "grad_norm": 2.0562708377838135, "learning_rate": 9.98947422844478e-06, "loss": 0.6097, "step": 1194 }, { "epoch": 0.050092744935707326, "grad_norm": 2.0050606727600098, "learning_rate": 9.98943015703137e-06, "loss": 0.6009, "step": 1195 }, { "epoch": 0.05013466355071628, "grad_norm": 2.192657470703125, "learning_rate": 9.989385993644711e-06, "loss": 0.6367, "step": 1196 }, { "epoch": 0.05017658216572524, "grad_norm": 1.9123454093933105, "learning_rate": 9.989341738285621e-06, "loss": 0.6643, "step": 1197 }, { "epoch": 0.050218500780734206, "grad_norm": 1.829519510269165, "learning_rate": 9.989297390954914e-06, "loss": 0.6361, "step": 1198 }, { "epoch": 0.05026041939574317, "grad_norm": 2.223609209060669, "learning_rate": 9.989252951653407e-06, "loss": 0.6112, "step": 1199 }, { "epoch": 0.05030233801075212, "grad_norm": 2.00823712348938, "learning_rate": 9.989208420381917e-06, "loss": 0.6007, "step": 1200 }, { "epoch": 0.050344256625761086, "grad_norm": 1.9087591171264648, "learning_rate": 9.98916379714127e-06, "loss": 0.6509, "step": 1201 }, { "epoch": 0.05038617524077005, "grad_norm": 2.153404712677002, "learning_rate": 9.989119081932283e-06, "loss": 0.5996, "step": 1202 }, { "epoch": 0.050428093855779, "grad_norm": 2.0067319869995117, "learning_rate": 9.989074274755785e-06, "loss": 0.6836, "step": 1203 }, { "epoch": 0.050470012470787966, "grad_norm": 2.2341272830963135, "learning_rate": 9.9890293756126e-06, "loss": 0.6433, "step": 1204 }, { "epoch": 0.05051193108579693, "grad_norm": 1.8087660074234009, "learning_rate": 9.988984384503556e-06, "loss": 0.6202, "step": 1205 }, { "epoch": 0.050553849700805883, "grad_norm": 2.634282112121582, "learning_rate": 9.98893930142948e-06, "loss": 0.7098, "step": 1206 }, { "epoch": 0.050595768315814846, "grad_norm": 2.0147223472595215, "learning_rate": 9.988894126391206e-06, "loss": 0.6233, "step": 1207 }, { "epoch": 0.05063768693082381, "grad_norm": 2.055959939956665, "learning_rate": 9.988848859389567e-06, "loss": 0.6284, "step": 1208 }, { "epoch": 0.050679605545832764, "grad_norm": 2.083878755569458, "learning_rate": 9.988803500425393e-06, "loss": 0.651, "step": 1209 }, { "epoch": 0.050721524160841726, "grad_norm": 2.033508777618408, "learning_rate": 9.988758049499525e-06, "loss": 0.643, "step": 1210 }, { "epoch": 0.05076344277585069, "grad_norm": 1.8659179210662842, "learning_rate": 9.988712506612799e-06, "loss": 0.5995, "step": 1211 }, { "epoch": 0.050805361390859644, "grad_norm": 2.097813129425049, "learning_rate": 9.988666871766055e-06, "loss": 0.6555, "step": 1212 }, { "epoch": 0.050847280005868606, "grad_norm": 1.9433560371398926, "learning_rate": 9.988621144960133e-06, "loss": 0.6145, "step": 1213 }, { "epoch": 0.05088919862087757, "grad_norm": 1.951347827911377, "learning_rate": 9.988575326195879e-06, "loss": 0.575, "step": 1214 }, { "epoch": 0.050931117235886524, "grad_norm": 2.17018723487854, "learning_rate": 9.988529415474133e-06, "loss": 0.65, "step": 1215 }, { "epoch": 0.050973035850895486, "grad_norm": 4.9987616539001465, "learning_rate": 9.988483412795745e-06, "loss": 0.6563, "step": 1216 }, { "epoch": 0.05101495446590445, "grad_norm": 1.8470473289489746, "learning_rate": 9.98843731816156e-06, "loss": 0.5685, "step": 1217 }, { "epoch": 0.051056873080913404, "grad_norm": 2.3942973613739014, "learning_rate": 9.988391131572431e-06, "loss": 0.5814, "step": 1218 }, { "epoch": 0.051098791695922366, "grad_norm": 2.0502943992614746, "learning_rate": 9.988344853029208e-06, "loss": 0.597, "step": 1219 }, { "epoch": 0.05114071031093133, "grad_norm": 1.988290786743164, "learning_rate": 9.988298482532742e-06, "loss": 0.6579, "step": 1220 }, { "epoch": 0.051182628925940284, "grad_norm": 2.4023656845092773, "learning_rate": 9.98825202008389e-06, "loss": 0.6611, "step": 1221 }, { "epoch": 0.051224547540949246, "grad_norm": 2.0992112159729004, "learning_rate": 9.98820546568351e-06, "loss": 0.5761, "step": 1222 }, { "epoch": 0.05126646615595821, "grad_norm": 1.7694010734558105, "learning_rate": 9.988158819332456e-06, "loss": 0.5953, "step": 1223 }, { "epoch": 0.05130838477096717, "grad_norm": 1.9513804912567139, "learning_rate": 9.988112081031591e-06, "loss": 0.6569, "step": 1224 }, { "epoch": 0.051350303385976126, "grad_norm": 1.9257766008377075, "learning_rate": 9.988065250781776e-06, "loss": 0.645, "step": 1225 }, { "epoch": 0.05139222200098509, "grad_norm": 2.1598329544067383, "learning_rate": 9.988018328583873e-06, "loss": 0.6243, "step": 1226 }, { "epoch": 0.05143414061599405, "grad_norm": 1.929274320602417, "learning_rate": 9.98797131443875e-06, "loss": 0.5417, "step": 1227 }, { "epoch": 0.051476059231003006, "grad_norm": 1.8553985357284546, "learning_rate": 9.987924208347268e-06, "loss": 0.6279, "step": 1228 }, { "epoch": 0.05151797784601197, "grad_norm": 2.372060775756836, "learning_rate": 9.9878770103103e-06, "loss": 0.6165, "step": 1229 }, { "epoch": 0.05155989646102093, "grad_norm": 2.2535455226898193, "learning_rate": 9.987829720328716e-06, "loss": 0.5687, "step": 1230 }, { "epoch": 0.051601815076029886, "grad_norm": 1.8585152626037598, "learning_rate": 9.987782338403386e-06, "loss": 0.6249, "step": 1231 }, { "epoch": 0.05164373369103885, "grad_norm": 1.9037634134292603, "learning_rate": 9.987734864535185e-06, "loss": 0.5939, "step": 1232 }, { "epoch": 0.05168565230604781, "grad_norm": 2.1330645084381104, "learning_rate": 9.987687298724987e-06, "loss": 0.6142, "step": 1233 }, { "epoch": 0.051727570921056766, "grad_norm": 2.885291576385498, "learning_rate": 9.987639640973666e-06, "loss": 0.6403, "step": 1234 }, { "epoch": 0.05176948953606573, "grad_norm": 1.9965204000473022, "learning_rate": 9.987591891282107e-06, "loss": 0.5995, "step": 1235 }, { "epoch": 0.05181140815107469, "grad_norm": 1.992872714996338, "learning_rate": 9.987544049651185e-06, "loss": 0.6308, "step": 1236 }, { "epoch": 0.051853326766083646, "grad_norm": 3.343101978302002, "learning_rate": 9.987496116081782e-06, "loss": 0.6526, "step": 1237 }, { "epoch": 0.05189524538109261, "grad_norm": 2.065959930419922, "learning_rate": 9.987448090574785e-06, "loss": 0.6216, "step": 1238 }, { "epoch": 0.05193716399610157, "grad_norm": 2.0595076084136963, "learning_rate": 9.987399973131078e-06, "loss": 0.6451, "step": 1239 }, { "epoch": 0.051979082611110526, "grad_norm": 1.9367858171463013, "learning_rate": 9.987351763751546e-06, "loss": 0.6104, "step": 1240 }, { "epoch": 0.05202100122611949, "grad_norm": 1.969705581665039, "learning_rate": 9.987303462437078e-06, "loss": 0.6274, "step": 1241 }, { "epoch": 0.05206291984112845, "grad_norm": 2.4096457958221436, "learning_rate": 9.987255069188569e-06, "loss": 0.6385, "step": 1242 }, { "epoch": 0.052104838456137406, "grad_norm": 2.1456127166748047, "learning_rate": 9.987206584006906e-06, "loss": 0.6081, "step": 1243 }, { "epoch": 0.05214675707114637, "grad_norm": 2.2432804107666016, "learning_rate": 9.987158006892983e-06, "loss": 0.6244, "step": 1244 }, { "epoch": 0.05218867568615533, "grad_norm": 1.9500576257705688, "learning_rate": 9.987109337847696e-06, "loss": 0.606, "step": 1245 }, { "epoch": 0.052230594301164286, "grad_norm": 2.2598683834075928, "learning_rate": 9.987060576871945e-06, "loss": 0.7045, "step": 1246 }, { "epoch": 0.05227251291617325, "grad_norm": 2.444749355316162, "learning_rate": 9.987011723966627e-06, "loss": 0.6264, "step": 1247 }, { "epoch": 0.05231443153118221, "grad_norm": 2.06552791595459, "learning_rate": 9.98696277913264e-06, "loss": 0.632, "step": 1248 }, { "epoch": 0.05235635014619117, "grad_norm": 2.889390468597412, "learning_rate": 9.98691374237089e-06, "loss": 0.6132, "step": 1249 }, { "epoch": 0.05239826876120013, "grad_norm": 1.9861881732940674, "learning_rate": 9.98686461368228e-06, "loss": 0.6352, "step": 1250 }, { "epoch": 0.05244018737620909, "grad_norm": 1.8763314485549927, "learning_rate": 9.986815393067713e-06, "loss": 0.6176, "step": 1251 }, { "epoch": 0.05248210599121805, "grad_norm": 2.4735231399536133, "learning_rate": 9.9867660805281e-06, "loss": 0.6508, "step": 1252 }, { "epoch": 0.05252402460622701, "grad_norm": 2.499662160873413, "learning_rate": 9.986716676064345e-06, "loss": 0.6445, "step": 1253 }, { "epoch": 0.05256594322123597, "grad_norm": 2.2106547355651855, "learning_rate": 9.986667179677366e-06, "loss": 0.6708, "step": 1254 }, { "epoch": 0.05260786183624493, "grad_norm": 2.0983963012695312, "learning_rate": 9.986617591368068e-06, "loss": 0.5835, "step": 1255 }, { "epoch": 0.05264978045125389, "grad_norm": 2.4893620014190674, "learning_rate": 9.98656791113737e-06, "loss": 0.6535, "step": 1256 }, { "epoch": 0.05269169906626285, "grad_norm": 1.746140480041504, "learning_rate": 9.986518138986188e-06, "loss": 0.5857, "step": 1257 }, { "epoch": 0.05273361768127181, "grad_norm": 1.9287320375442505, "learning_rate": 9.986468274915435e-06, "loss": 0.6539, "step": 1258 }, { "epoch": 0.05277553629628077, "grad_norm": 2.1616291999816895, "learning_rate": 9.986418318926035e-06, "loss": 0.6036, "step": 1259 }, { "epoch": 0.05281745491128973, "grad_norm": 1.8560868501663208, "learning_rate": 9.986368271018907e-06, "loss": 0.6463, "step": 1260 }, { "epoch": 0.05285937352629869, "grad_norm": 2.1939847469329834, "learning_rate": 9.986318131194972e-06, "loss": 0.6887, "step": 1261 }, { "epoch": 0.05290129214130765, "grad_norm": 1.8849213123321533, "learning_rate": 9.986267899455155e-06, "loss": 0.595, "step": 1262 }, { "epoch": 0.05294321075631661, "grad_norm": 2.0380661487579346, "learning_rate": 9.986217575800384e-06, "loss": 0.6155, "step": 1263 }, { "epoch": 0.05298512937132557, "grad_norm": 1.8798531293869019, "learning_rate": 9.986167160231585e-06, "loss": 0.6082, "step": 1264 }, { "epoch": 0.05302704798633453, "grad_norm": 2.1892759799957275, "learning_rate": 9.986116652749688e-06, "loss": 0.6866, "step": 1265 }, { "epoch": 0.05306896660134349, "grad_norm": 2.494044065475464, "learning_rate": 9.986066053355623e-06, "loss": 0.638, "step": 1266 }, { "epoch": 0.05311088521635245, "grad_norm": 1.8802580833435059, "learning_rate": 9.986015362050323e-06, "loss": 0.6406, "step": 1267 }, { "epoch": 0.05315280383136141, "grad_norm": 1.7442837953567505, "learning_rate": 9.985964578834724e-06, "loss": 0.5767, "step": 1268 }, { "epoch": 0.05319472244637037, "grad_norm": 1.8914744853973389, "learning_rate": 9.98591370370976e-06, "loss": 0.6387, "step": 1269 }, { "epoch": 0.05323664106137933, "grad_norm": 2.3646955490112305, "learning_rate": 9.98586273667637e-06, "loss": 0.6138, "step": 1270 }, { "epoch": 0.05327855967638829, "grad_norm": 3.05915904045105, "learning_rate": 9.985811677735493e-06, "loss": 0.5898, "step": 1271 }, { "epoch": 0.05332047829139725, "grad_norm": 2.13676381111145, "learning_rate": 9.98576052688807e-06, "loss": 0.6778, "step": 1272 }, { "epoch": 0.05336239690640621, "grad_norm": 2.010139226913452, "learning_rate": 9.985709284135045e-06, "loss": 0.6456, "step": 1273 }, { "epoch": 0.053404315521415176, "grad_norm": 2.2587037086486816, "learning_rate": 9.985657949477361e-06, "loss": 0.6275, "step": 1274 }, { "epoch": 0.05344623413642413, "grad_norm": 2.0190412998199463, "learning_rate": 9.985606522915966e-06, "loss": 0.6139, "step": 1275 }, { "epoch": 0.05348815275143309, "grad_norm": 2.022145986557007, "learning_rate": 9.985555004451805e-06, "loss": 0.6143, "step": 1276 }, { "epoch": 0.053530071366442056, "grad_norm": 3.6481170654296875, "learning_rate": 9.985503394085834e-06, "loss": 0.5626, "step": 1277 }, { "epoch": 0.05357198998145101, "grad_norm": 1.911697506904602, "learning_rate": 9.985451691818997e-06, "loss": 0.6091, "step": 1278 }, { "epoch": 0.053613908596459973, "grad_norm": 2.0157885551452637, "learning_rate": 9.985399897652251e-06, "loss": 0.6145, "step": 1279 }, { "epoch": 0.053655827211468936, "grad_norm": 1.902919888496399, "learning_rate": 9.98534801158655e-06, "loss": 0.6122, "step": 1280 }, { "epoch": 0.05369774582647789, "grad_norm": 2.286977767944336, "learning_rate": 9.98529603362285e-06, "loss": 0.6333, "step": 1281 }, { "epoch": 0.053739664441486854, "grad_norm": 1.9264166355133057, "learning_rate": 9.98524396376211e-06, "loss": 0.5762, "step": 1282 }, { "epoch": 0.053781583056495816, "grad_norm": 1.988459825515747, "learning_rate": 9.98519180200529e-06, "loss": 0.5882, "step": 1283 }, { "epoch": 0.05382350167150477, "grad_norm": 2.1600351333618164, "learning_rate": 9.98513954835335e-06, "loss": 0.609, "step": 1284 }, { "epoch": 0.053865420286513734, "grad_norm": 2.0312626361846924, "learning_rate": 9.985087202807255e-06, "loss": 0.6357, "step": 1285 }, { "epoch": 0.053907338901522696, "grad_norm": 2.403245449066162, "learning_rate": 9.98503476536797e-06, "loss": 0.6188, "step": 1286 }, { "epoch": 0.05394925751653165, "grad_norm": 1.8363150358200073, "learning_rate": 9.984982236036458e-06, "loss": 0.6066, "step": 1287 }, { "epoch": 0.053991176131540614, "grad_norm": 2.041160821914673, "learning_rate": 9.984929614813692e-06, "loss": 0.5718, "step": 1288 }, { "epoch": 0.054033094746549576, "grad_norm": 2.3960676193237305, "learning_rate": 9.98487690170064e-06, "loss": 0.6086, "step": 1289 }, { "epoch": 0.05407501336155853, "grad_norm": 1.9735243320465088, "learning_rate": 9.984824096698273e-06, "loss": 0.6022, "step": 1290 }, { "epoch": 0.054116931976567494, "grad_norm": 2.005246162414551, "learning_rate": 9.984771199807565e-06, "loss": 0.6137, "step": 1291 }, { "epoch": 0.054158850591576456, "grad_norm": 1.9862513542175293, "learning_rate": 9.984718211029492e-06, "loss": 0.5604, "step": 1292 }, { "epoch": 0.05420076920658541, "grad_norm": 1.9234119653701782, "learning_rate": 9.984665130365027e-06, "loss": 0.6161, "step": 1293 }, { "epoch": 0.054242687821594374, "grad_norm": 1.7938486337661743, "learning_rate": 9.984611957815156e-06, "loss": 0.6112, "step": 1294 }, { "epoch": 0.054284606436603336, "grad_norm": 1.7420527935028076, "learning_rate": 9.984558693380851e-06, "loss": 0.5776, "step": 1295 }, { "epoch": 0.05432652505161229, "grad_norm": 2.201335906982422, "learning_rate": 9.9845053370631e-06, "loss": 0.6049, "step": 1296 }, { "epoch": 0.054368443666621254, "grad_norm": 1.9642860889434814, "learning_rate": 9.984451888862883e-06, "loss": 0.5627, "step": 1297 }, { "epoch": 0.054410362281630216, "grad_norm": 1.932640790939331, "learning_rate": 9.984398348781186e-06, "loss": 0.6037, "step": 1298 }, { "epoch": 0.05445228089663918, "grad_norm": 2.041059970855713, "learning_rate": 9.984344716818995e-06, "loss": 0.6457, "step": 1299 }, { "epoch": 0.054494199511648134, "grad_norm": 2.867133617401123, "learning_rate": 9.984290992977301e-06, "loss": 0.6056, "step": 1300 }, { "epoch": 0.054536118126657096, "grad_norm": 2.045234203338623, "learning_rate": 9.984237177257094e-06, "loss": 0.6564, "step": 1301 }, { "epoch": 0.05457803674166606, "grad_norm": 1.9185905456542969, "learning_rate": 9.984183269659363e-06, "loss": 0.5595, "step": 1302 }, { "epoch": 0.054619955356675014, "grad_norm": 2.9808223247528076, "learning_rate": 9.984129270185105e-06, "loss": 0.6004, "step": 1303 }, { "epoch": 0.054661873971683976, "grad_norm": 1.9428937435150146, "learning_rate": 9.984075178835314e-06, "loss": 0.6481, "step": 1304 }, { "epoch": 0.05470379258669294, "grad_norm": 2.08309006690979, "learning_rate": 9.984020995610987e-06, "loss": 0.5992, "step": 1305 }, { "epoch": 0.054745711201701894, "grad_norm": 2.1670565605163574, "learning_rate": 9.983966720513125e-06, "loss": 0.6359, "step": 1306 }, { "epoch": 0.054787629816710856, "grad_norm": 2.1318557262420654, "learning_rate": 9.983912353542727e-06, "loss": 0.6273, "step": 1307 }, { "epoch": 0.05482954843171982, "grad_norm": 1.920535683631897, "learning_rate": 9.983857894700792e-06, "loss": 0.5951, "step": 1308 }, { "epoch": 0.054871467046728774, "grad_norm": 1.8329706192016602, "learning_rate": 9.983803343988327e-06, "loss": 0.6115, "step": 1309 }, { "epoch": 0.054913385661737736, "grad_norm": 1.947283148765564, "learning_rate": 9.983748701406339e-06, "loss": 0.5963, "step": 1310 }, { "epoch": 0.0549553042767467, "grad_norm": 2.468480348587036, "learning_rate": 9.983693966955833e-06, "loss": 0.6419, "step": 1311 }, { "epoch": 0.054997222891755654, "grad_norm": 2.035381317138672, "learning_rate": 9.983639140637817e-06, "loss": 0.6055, "step": 1312 }, { "epoch": 0.055039141506764616, "grad_norm": 1.981642723083496, "learning_rate": 9.983584222453305e-06, "loss": 0.6191, "step": 1313 }, { "epoch": 0.05508106012177358, "grad_norm": 1.9671025276184082, "learning_rate": 9.983529212403306e-06, "loss": 0.611, "step": 1314 }, { "epoch": 0.055122978736782534, "grad_norm": 1.8478256464004517, "learning_rate": 9.983474110488838e-06, "loss": 0.5919, "step": 1315 }, { "epoch": 0.055164897351791496, "grad_norm": 2.3757731914520264, "learning_rate": 9.983418916710912e-06, "loss": 0.6443, "step": 1316 }, { "epoch": 0.05520681596680046, "grad_norm": 2.9753000736236572, "learning_rate": 9.983363631070548e-06, "loss": 0.6094, "step": 1317 }, { "epoch": 0.055248734581809414, "grad_norm": 1.8798216581344604, "learning_rate": 9.983308253568766e-06, "loss": 0.6459, "step": 1318 }, { "epoch": 0.055290653196818376, "grad_norm": 1.8300024271011353, "learning_rate": 9.983252784206582e-06, "loss": 0.5796, "step": 1319 }, { "epoch": 0.05533257181182734, "grad_norm": 1.795448899269104, "learning_rate": 9.983197222985026e-06, "loss": 0.5783, "step": 1320 }, { "epoch": 0.055374490426836294, "grad_norm": 2.1535723209381104, "learning_rate": 9.983141569905117e-06, "loss": 0.5765, "step": 1321 }, { "epoch": 0.055416409041845256, "grad_norm": 2.0039291381835938, "learning_rate": 9.983085824967881e-06, "loss": 0.6091, "step": 1322 }, { "epoch": 0.05545832765685422, "grad_norm": 2.4490580558776855, "learning_rate": 9.983029988174349e-06, "loss": 0.642, "step": 1323 }, { "epoch": 0.05550024627186318, "grad_norm": 1.8516031503677368, "learning_rate": 9.982974059525545e-06, "loss": 0.587, "step": 1324 }, { "epoch": 0.055542164886872136, "grad_norm": 2.451157331466675, "learning_rate": 9.982918039022506e-06, "loss": 0.6314, "step": 1325 }, { "epoch": 0.0555840835018811, "grad_norm": 1.7997502088546753, "learning_rate": 9.982861926666259e-06, "loss": 0.5967, "step": 1326 }, { "epoch": 0.05562600211689006, "grad_norm": 1.8877085447311401, "learning_rate": 9.982805722457842e-06, "loss": 0.5697, "step": 1327 }, { "epoch": 0.055667920731899016, "grad_norm": 1.8941874504089355, "learning_rate": 9.98274942639829e-06, "loss": 0.5717, "step": 1328 }, { "epoch": 0.05570983934690798, "grad_norm": 1.8278114795684814, "learning_rate": 9.982693038488643e-06, "loss": 0.603, "step": 1329 }, { "epoch": 0.05575175796191694, "grad_norm": 2.1739487648010254, "learning_rate": 9.982636558729936e-06, "loss": 0.6517, "step": 1330 }, { "epoch": 0.055793676576925896, "grad_norm": 2.225857973098755, "learning_rate": 9.982579987123212e-06, "loss": 0.5991, "step": 1331 }, { "epoch": 0.05583559519193486, "grad_norm": 1.9420722723007202, "learning_rate": 9.982523323669515e-06, "loss": 0.6171, "step": 1332 }, { "epoch": 0.05587751380694382, "grad_norm": 1.861253261566162, "learning_rate": 9.982466568369889e-06, "loss": 0.6373, "step": 1333 }, { "epoch": 0.055919432421952776, "grad_norm": 2.3274006843566895, "learning_rate": 9.982409721225379e-06, "loss": 0.6335, "step": 1334 }, { "epoch": 0.05596135103696174, "grad_norm": 2.008880138397217, "learning_rate": 9.982352782237032e-06, "loss": 0.6088, "step": 1335 }, { "epoch": 0.0560032696519707, "grad_norm": 1.9226597547531128, "learning_rate": 9.982295751405903e-06, "loss": 0.6345, "step": 1336 }, { "epoch": 0.056045188266979656, "grad_norm": 2.2988739013671875, "learning_rate": 9.982238628733037e-06, "loss": 0.5885, "step": 1337 }, { "epoch": 0.05608710688198862, "grad_norm": 1.9976801872253418, "learning_rate": 9.982181414219491e-06, "loss": 0.6198, "step": 1338 }, { "epoch": 0.05612902549699758, "grad_norm": 1.8364909887313843, "learning_rate": 9.982124107866318e-06, "loss": 0.5919, "step": 1339 }, { "epoch": 0.056170944112006536, "grad_norm": 1.796966314315796, "learning_rate": 9.982066709674574e-06, "loss": 0.5846, "step": 1340 }, { "epoch": 0.0562128627270155, "grad_norm": 1.9829405546188354, "learning_rate": 9.98200921964532e-06, "loss": 0.6729, "step": 1341 }, { "epoch": 0.05625478134202446, "grad_norm": 1.6952950954437256, "learning_rate": 9.98195163777961e-06, "loss": 0.6399, "step": 1342 }, { "epoch": 0.056296699957033416, "grad_norm": 1.7685215473175049, "learning_rate": 9.98189396407851e-06, "loss": 0.5721, "step": 1343 }, { "epoch": 0.05633861857204238, "grad_norm": 2.040578603744507, "learning_rate": 9.981836198543085e-06, "loss": 0.5605, "step": 1344 }, { "epoch": 0.05638053718705134, "grad_norm": 1.9199213981628418, "learning_rate": 9.981778341174393e-06, "loss": 0.611, "step": 1345 }, { "epoch": 0.056422455802060296, "grad_norm": 1.9052711725234985, "learning_rate": 9.981720391973506e-06, "loss": 0.6209, "step": 1346 }, { "epoch": 0.05646437441706926, "grad_norm": 2.16719651222229, "learning_rate": 9.98166235094149e-06, "loss": 0.6137, "step": 1347 }, { "epoch": 0.05650629303207822, "grad_norm": 2.017622232437134, "learning_rate": 9.981604218079417e-06, "loss": 0.586, "step": 1348 }, { "epoch": 0.05654821164708718, "grad_norm": 2.6333634853363037, "learning_rate": 9.981545993388356e-06, "loss": 0.6493, "step": 1349 }, { "epoch": 0.05659013026209614, "grad_norm": 1.9627697467803955, "learning_rate": 9.981487676869382e-06, "loss": 0.6383, "step": 1350 }, { "epoch": 0.0566320488771051, "grad_norm": 1.9662917852401733, "learning_rate": 9.981429268523568e-06, "loss": 0.6799, "step": 1351 }, { "epoch": 0.056673967492114063, "grad_norm": 2.299509048461914, "learning_rate": 9.981370768351993e-06, "loss": 0.6464, "step": 1352 }, { "epoch": 0.05671588610712302, "grad_norm": 2.0262186527252197, "learning_rate": 9.981312176355734e-06, "loss": 0.5925, "step": 1353 }, { "epoch": 0.05675780472213198, "grad_norm": 1.9595824480056763, "learning_rate": 9.981253492535871e-06, "loss": 0.5703, "step": 1354 }, { "epoch": 0.056799723337140944, "grad_norm": 2.1030216217041016, "learning_rate": 9.981194716893488e-06, "loss": 0.6145, "step": 1355 }, { "epoch": 0.0568416419521499, "grad_norm": 1.8685485124588013, "learning_rate": 9.981135849429667e-06, "loss": 0.6374, "step": 1356 }, { "epoch": 0.05688356056715886, "grad_norm": 1.785399317741394, "learning_rate": 9.98107689014549e-06, "loss": 0.596, "step": 1357 }, { "epoch": 0.056925479182167824, "grad_norm": 2.386033773422241, "learning_rate": 9.981017839042048e-06, "loss": 0.6462, "step": 1358 }, { "epoch": 0.05696739779717678, "grad_norm": 1.9790961742401123, "learning_rate": 9.98095869612043e-06, "loss": 0.6121, "step": 1359 }, { "epoch": 0.05700931641218574, "grad_norm": 1.9474191665649414, "learning_rate": 9.980899461381721e-06, "loss": 0.6103, "step": 1360 }, { "epoch": 0.057051235027194704, "grad_norm": 2.3457837104797363, "learning_rate": 9.980840134827019e-06, "loss": 0.5632, "step": 1361 }, { "epoch": 0.05709315364220366, "grad_norm": 2.0120296478271484, "learning_rate": 9.980780716457414e-06, "loss": 0.6595, "step": 1362 }, { "epoch": 0.05713507225721262, "grad_norm": 1.9465663433074951, "learning_rate": 9.980721206274002e-06, "loss": 0.5292, "step": 1363 }, { "epoch": 0.057176990872221584, "grad_norm": 1.9946919679641724, "learning_rate": 9.980661604277881e-06, "loss": 0.6665, "step": 1364 }, { "epoch": 0.05721890948723054, "grad_norm": 1.8847137689590454, "learning_rate": 9.980601910470148e-06, "loss": 0.5803, "step": 1365 }, { "epoch": 0.0572608281022395, "grad_norm": 1.9074627161026, "learning_rate": 9.980542124851905e-06, "loss": 0.5634, "step": 1366 }, { "epoch": 0.057302746717248464, "grad_norm": 1.7877618074417114, "learning_rate": 9.980482247424253e-06, "loss": 0.5622, "step": 1367 }, { "epoch": 0.05734466533225742, "grad_norm": 1.9985162019729614, "learning_rate": 9.980422278188295e-06, "loss": 0.6235, "step": 1368 }, { "epoch": 0.05738658394726638, "grad_norm": 1.8394019603729248, "learning_rate": 9.980362217145139e-06, "loss": 0.6207, "step": 1369 }, { "epoch": 0.057428502562275344, "grad_norm": 1.746224045753479, "learning_rate": 9.98030206429589e-06, "loss": 0.6436, "step": 1370 }, { "epoch": 0.057470421177284306, "grad_norm": 1.71892249584198, "learning_rate": 9.980241819641659e-06, "loss": 0.592, "step": 1371 }, { "epoch": 0.05751233979229326, "grad_norm": 2.322996139526367, "learning_rate": 9.980181483183552e-06, "loss": 0.5803, "step": 1372 }, { "epoch": 0.057554258407302224, "grad_norm": 2.039534091949463, "learning_rate": 9.980121054922687e-06, "loss": 0.6968, "step": 1373 }, { "epoch": 0.057596177022311186, "grad_norm": 1.9519716501235962, "learning_rate": 9.980060534860174e-06, "loss": 0.6282, "step": 1374 }, { "epoch": 0.05763809563732014, "grad_norm": 1.8485933542251587, "learning_rate": 9.97999992299713e-06, "loss": 0.6045, "step": 1375 }, { "epoch": 0.057680014252329104, "grad_norm": 1.775390863418579, "learning_rate": 9.979939219334673e-06, "loss": 0.6212, "step": 1376 }, { "epoch": 0.057721932867338066, "grad_norm": 1.7785335779190063, "learning_rate": 9.979878423873919e-06, "loss": 0.6117, "step": 1377 }, { "epoch": 0.05776385148234702, "grad_norm": 1.9469469785690308, "learning_rate": 9.979817536615993e-06, "loss": 0.5837, "step": 1378 }, { "epoch": 0.057805770097355984, "grad_norm": 2.0754640102386475, "learning_rate": 9.979756557562013e-06, "loss": 0.5756, "step": 1379 }, { "epoch": 0.057847688712364946, "grad_norm": 1.883232831954956, "learning_rate": 9.979695486713105e-06, "loss": 0.6272, "step": 1380 }, { "epoch": 0.0578896073273739, "grad_norm": 1.9015535116195679, "learning_rate": 9.979634324070397e-06, "loss": 0.6278, "step": 1381 }, { "epoch": 0.057931525942382864, "grad_norm": 1.8820173740386963, "learning_rate": 9.979573069635014e-06, "loss": 0.636, "step": 1382 }, { "epoch": 0.057973444557391826, "grad_norm": 2.1800131797790527, "learning_rate": 9.979511723408084e-06, "loss": 0.6462, "step": 1383 }, { "epoch": 0.05801536317240078, "grad_norm": 1.9568400382995605, "learning_rate": 9.979450285390741e-06, "loss": 0.618, "step": 1384 }, { "epoch": 0.058057281787409744, "grad_norm": 1.8350356817245483, "learning_rate": 9.979388755584114e-06, "loss": 0.6177, "step": 1385 }, { "epoch": 0.058099200402418706, "grad_norm": 2.056380271911621, "learning_rate": 9.97932713398934e-06, "loss": 0.6044, "step": 1386 }, { "epoch": 0.05814111901742766, "grad_norm": 1.9604220390319824, "learning_rate": 9.979265420607555e-06, "loss": 0.5929, "step": 1387 }, { "epoch": 0.058183037632436624, "grad_norm": 2.1511707305908203, "learning_rate": 9.979203615439895e-06, "loss": 0.5911, "step": 1388 }, { "epoch": 0.058224956247445586, "grad_norm": 2.0633063316345215, "learning_rate": 9.9791417184875e-06, "loss": 0.6499, "step": 1389 }, { "epoch": 0.05826687486245454, "grad_norm": 2.1070477962493896, "learning_rate": 9.97907972975151e-06, "loss": 0.5976, "step": 1390 }, { "epoch": 0.058308793477463504, "grad_norm": 1.8159520626068115, "learning_rate": 9.97901764923307e-06, "loss": 0.5554, "step": 1391 }, { "epoch": 0.058350712092472466, "grad_norm": 2.3222837448120117, "learning_rate": 9.978955476933321e-06, "loss": 0.6486, "step": 1392 }, { "epoch": 0.05839263070748142, "grad_norm": 1.9340695142745972, "learning_rate": 9.978893212853411e-06, "loss": 0.6525, "step": 1393 }, { "epoch": 0.058434549322490384, "grad_norm": 1.9460103511810303, "learning_rate": 9.97883085699449e-06, "loss": 0.6163, "step": 1394 }, { "epoch": 0.058476467937499346, "grad_norm": 2.0021026134490967, "learning_rate": 9.978768409357704e-06, "loss": 0.6073, "step": 1395 }, { "epoch": 0.05851838655250831, "grad_norm": 1.9436322450637817, "learning_rate": 9.978705869944204e-06, "loss": 0.6274, "step": 1396 }, { "epoch": 0.058560305167517264, "grad_norm": 1.684862732887268, "learning_rate": 9.978643238755146e-06, "loss": 0.5326, "step": 1397 }, { "epoch": 0.058602223782526226, "grad_norm": 1.9728258848190308, "learning_rate": 9.978580515791682e-06, "loss": 0.6159, "step": 1398 }, { "epoch": 0.05864414239753519, "grad_norm": 1.9457786083221436, "learning_rate": 9.978517701054968e-06, "loss": 0.6705, "step": 1399 }, { "epoch": 0.058686061012544144, "grad_norm": 1.7537955045700073, "learning_rate": 9.978454794546165e-06, "loss": 0.5838, "step": 1400 }, { "epoch": 0.058727979627553106, "grad_norm": 16.887022018432617, "learning_rate": 9.978391796266429e-06, "loss": 0.6467, "step": 1401 }, { "epoch": 0.05876989824256207, "grad_norm": 2.056612253189087, "learning_rate": 9.978328706216922e-06, "loss": 0.6623, "step": 1402 }, { "epoch": 0.058811816857571024, "grad_norm": 2.09045672416687, "learning_rate": 9.978265524398809e-06, "loss": 0.6537, "step": 1403 }, { "epoch": 0.058853735472579986, "grad_norm": 2.259674549102783, "learning_rate": 9.978202250813253e-06, "loss": 0.6472, "step": 1404 }, { "epoch": 0.05889565408758895, "grad_norm": 1.8747388124465942, "learning_rate": 9.978138885461418e-06, "loss": 0.6358, "step": 1405 }, { "epoch": 0.058937572702597904, "grad_norm": 2.226490020751953, "learning_rate": 9.978075428344479e-06, "loss": 0.6226, "step": 1406 }, { "epoch": 0.058979491317606866, "grad_norm": 2.598752737045288, "learning_rate": 9.978011879463598e-06, "loss": 0.6548, "step": 1407 }, { "epoch": 0.05902140993261583, "grad_norm": 4.092623233795166, "learning_rate": 9.977948238819953e-06, "loss": 0.6127, "step": 1408 }, { "epoch": 0.059063328547624784, "grad_norm": 1.899196743965149, "learning_rate": 9.977884506414712e-06, "loss": 0.5779, "step": 1409 }, { "epoch": 0.059105247162633746, "grad_norm": 1.7072921991348267, "learning_rate": 9.977820682249051e-06, "loss": 0.6013, "step": 1410 }, { "epoch": 0.05914716577764271, "grad_norm": 1.9644606113433838, "learning_rate": 9.977756766324149e-06, "loss": 0.6768, "step": 1411 }, { "epoch": 0.059189084392651664, "grad_norm": 2.021064519882202, "learning_rate": 9.977692758641182e-06, "loss": 0.6538, "step": 1412 }, { "epoch": 0.059231003007660626, "grad_norm": 1.9015260934829712, "learning_rate": 9.977628659201332e-06, "loss": 0.6425, "step": 1413 }, { "epoch": 0.05927292162266959, "grad_norm": 2.218628168106079, "learning_rate": 9.977564468005777e-06, "loss": 0.6097, "step": 1414 }, { "epoch": 0.059314840237678544, "grad_norm": 2.116910696029663, "learning_rate": 9.977500185055703e-06, "loss": 0.6448, "step": 1415 }, { "epoch": 0.059356758852687506, "grad_norm": 2.6827950477600098, "learning_rate": 9.977435810352293e-06, "loss": 0.6017, "step": 1416 }, { "epoch": 0.05939867746769647, "grad_norm": 2.4983506202697754, "learning_rate": 9.977371343896738e-06, "loss": 0.5887, "step": 1417 }, { "epoch": 0.059440596082705424, "grad_norm": 2.296255111694336, "learning_rate": 9.977306785690221e-06, "loss": 0.6147, "step": 1418 }, { "epoch": 0.059482514697714386, "grad_norm": 1.8953253030776978, "learning_rate": 9.977242135733935e-06, "loss": 0.6877, "step": 1419 }, { "epoch": 0.05952443331272335, "grad_norm": 1.924159049987793, "learning_rate": 9.97717739402907e-06, "loss": 0.6317, "step": 1420 }, { "epoch": 0.05956635192773231, "grad_norm": 2.179140090942383, "learning_rate": 9.977112560576822e-06, "loss": 0.6473, "step": 1421 }, { "epoch": 0.059608270542741267, "grad_norm": 1.8586585521697998, "learning_rate": 9.977047635378384e-06, "loss": 0.6343, "step": 1422 }, { "epoch": 0.05965018915775023, "grad_norm": 2.490114688873291, "learning_rate": 9.976982618434952e-06, "loss": 0.5657, "step": 1423 }, { "epoch": 0.05969210777275919, "grad_norm": 2.078235149383545, "learning_rate": 9.976917509747729e-06, "loss": 0.5633, "step": 1424 }, { "epoch": 0.05973402638776815, "grad_norm": 2.1481244564056396, "learning_rate": 9.976852309317909e-06, "loss": 0.6093, "step": 1425 }, { "epoch": 0.05977594500277711, "grad_norm": 1.8761168718338013, "learning_rate": 9.976787017146698e-06, "loss": 0.6176, "step": 1426 }, { "epoch": 0.05981786361778607, "grad_norm": 1.7204476594924927, "learning_rate": 9.9767216332353e-06, "loss": 0.5956, "step": 1427 }, { "epoch": 0.05985978223279503, "grad_norm": 3.5592687129974365, "learning_rate": 9.976656157584917e-06, "loss": 0.6108, "step": 1428 }, { "epoch": 0.05990170084780399, "grad_norm": 1.9034838676452637, "learning_rate": 9.976590590196758e-06, "loss": 0.6565, "step": 1429 }, { "epoch": 0.05994361946281295, "grad_norm": 1.7582980394363403, "learning_rate": 9.976524931072032e-06, "loss": 0.6173, "step": 1430 }, { "epoch": 0.05998553807782191, "grad_norm": 2.2238662242889404, "learning_rate": 9.976459180211948e-06, "loss": 0.6356, "step": 1431 }, { "epoch": 0.06002745669283087, "grad_norm": 1.7417452335357666, "learning_rate": 9.97639333761772e-06, "loss": 0.6565, "step": 1432 }, { "epoch": 0.06006937530783983, "grad_norm": 2.0009937286376953, "learning_rate": 9.97632740329056e-06, "loss": 0.6107, "step": 1433 }, { "epoch": 0.06011129392284879, "grad_norm": 2.4243884086608887, "learning_rate": 9.976261377231684e-06, "loss": 0.6506, "step": 1434 }, { "epoch": 0.06015321253785775, "grad_norm": 1.9753891229629517, "learning_rate": 9.976195259442308e-06, "loss": 0.6427, "step": 1435 }, { "epoch": 0.06019513115286671, "grad_norm": 2.422389507293701, "learning_rate": 9.976129049923653e-06, "loss": 0.5806, "step": 1436 }, { "epoch": 0.06023704976787567, "grad_norm": 1.8259835243225098, "learning_rate": 9.976062748676937e-06, "loss": 0.618, "step": 1437 }, { "epoch": 0.06027896838288463, "grad_norm": 2.0407299995422363, "learning_rate": 9.975996355703383e-06, "loss": 0.6644, "step": 1438 }, { "epoch": 0.06032088699789359, "grad_norm": 1.869606852531433, "learning_rate": 9.975929871004217e-06, "loss": 0.5838, "step": 1439 }, { "epoch": 0.06036280561290255, "grad_norm": 1.7980517148971558, "learning_rate": 9.975863294580664e-06, "loss": 0.6298, "step": 1440 }, { "epoch": 0.06040472422791151, "grad_norm": 1.9329756498336792, "learning_rate": 9.975796626433947e-06, "loss": 0.5923, "step": 1441 }, { "epoch": 0.06044664284292047, "grad_norm": 2.164726495742798, "learning_rate": 9.9757298665653e-06, "loss": 0.684, "step": 1442 }, { "epoch": 0.06048856145792943, "grad_norm": 1.9109828472137451, "learning_rate": 9.97566301497595e-06, "loss": 0.5568, "step": 1443 }, { "epoch": 0.06053048007293839, "grad_norm": 1.7574269771575928, "learning_rate": 9.975596071667133e-06, "loss": 0.5611, "step": 1444 }, { "epoch": 0.06057239868794735, "grad_norm": 1.802255630493164, "learning_rate": 9.97552903664008e-06, "loss": 0.5995, "step": 1445 }, { "epoch": 0.060614317302956314, "grad_norm": 2.0548758506774902, "learning_rate": 9.975461909896027e-06, "loss": 0.6262, "step": 1446 }, { "epoch": 0.06065623591796527, "grad_norm": 2.458681106567383, "learning_rate": 9.975394691436213e-06, "loss": 0.7421, "step": 1447 }, { "epoch": 0.06069815453297423, "grad_norm": 1.7571147680282593, "learning_rate": 9.975327381261874e-06, "loss": 0.639, "step": 1448 }, { "epoch": 0.060740073147983194, "grad_norm": 1.966943621635437, "learning_rate": 9.975259979374256e-06, "loss": 0.6019, "step": 1449 }, { "epoch": 0.06078199176299215, "grad_norm": 1.882715106010437, "learning_rate": 9.975192485774597e-06, "loss": 0.6018, "step": 1450 }, { "epoch": 0.06082391037800111, "grad_norm": 1.971610426902771, "learning_rate": 9.975124900464142e-06, "loss": 0.5886, "step": 1451 }, { "epoch": 0.060865828993010074, "grad_norm": 2.058687210083008, "learning_rate": 9.975057223444138e-06, "loss": 0.6968, "step": 1452 }, { "epoch": 0.06090774760801903, "grad_norm": 2.0042660236358643, "learning_rate": 9.974989454715833e-06, "loss": 0.681, "step": 1453 }, { "epoch": 0.06094966622302799, "grad_norm": 1.8787566423416138, "learning_rate": 9.974921594280473e-06, "loss": 0.6483, "step": 1454 }, { "epoch": 0.060991584838036954, "grad_norm": 1.9350502490997314, "learning_rate": 9.974853642139313e-06, "loss": 0.6131, "step": 1455 }, { "epoch": 0.06103350345304591, "grad_norm": 1.8818621635437012, "learning_rate": 9.974785598293601e-06, "loss": 0.5828, "step": 1456 }, { "epoch": 0.06107542206805487, "grad_norm": 1.924046277999878, "learning_rate": 9.974717462744597e-06, "loss": 0.5845, "step": 1457 }, { "epoch": 0.061117340683063834, "grad_norm": 1.8216885328292847, "learning_rate": 9.974649235493552e-06, "loss": 0.6179, "step": 1458 }, { "epoch": 0.06115925929807279, "grad_norm": 2.2640228271484375, "learning_rate": 9.974580916541726e-06, "loss": 0.6549, "step": 1459 }, { "epoch": 0.06120117791308175, "grad_norm": 2.4168007373809814, "learning_rate": 9.974512505890377e-06, "loss": 0.654, "step": 1460 }, { "epoch": 0.061243096528090714, "grad_norm": 2.46081280708313, "learning_rate": 9.974444003540768e-06, "loss": 0.6722, "step": 1461 }, { "epoch": 0.06128501514309967, "grad_norm": 1.7982194423675537, "learning_rate": 9.974375409494162e-06, "loss": 0.5604, "step": 1462 }, { "epoch": 0.06132693375810863, "grad_norm": 1.9789844751358032, "learning_rate": 9.974306723751822e-06, "loss": 0.5666, "step": 1463 }, { "epoch": 0.061368852373117594, "grad_norm": 2.1062164306640625, "learning_rate": 9.974237946315014e-06, "loss": 0.6823, "step": 1464 }, { "epoch": 0.06141077098812655, "grad_norm": 1.7374387979507446, "learning_rate": 9.974169077185004e-06, "loss": 0.5903, "step": 1465 }, { "epoch": 0.06145268960313551, "grad_norm": 1.8783818483352661, "learning_rate": 9.974100116363065e-06, "loss": 0.5469, "step": 1466 }, { "epoch": 0.061494608218144474, "grad_norm": 2.178734540939331, "learning_rate": 9.974031063850466e-06, "loss": 0.5872, "step": 1467 }, { "epoch": 0.06153652683315343, "grad_norm": 2.0364181995391846, "learning_rate": 9.973961919648484e-06, "loss": 0.5999, "step": 1468 }, { "epoch": 0.06157844544816239, "grad_norm": 1.8119195699691772, "learning_rate": 9.973892683758386e-06, "loss": 0.6563, "step": 1469 }, { "epoch": 0.061620364063171354, "grad_norm": 1.9790414571762085, "learning_rate": 9.973823356181455e-06, "loss": 0.6195, "step": 1470 }, { "epoch": 0.061662282678180316, "grad_norm": 3.417616367340088, "learning_rate": 9.973753936918966e-06, "loss": 0.5764, "step": 1471 }, { "epoch": 0.06170420129318927, "grad_norm": 2.0145199298858643, "learning_rate": 9.973684425972199e-06, "loss": 0.5779, "step": 1472 }, { "epoch": 0.061746119908198234, "grad_norm": 2.245696544647217, "learning_rate": 9.973614823342436e-06, "loss": 0.6864, "step": 1473 }, { "epoch": 0.061788038523207196, "grad_norm": 1.9435502290725708, "learning_rate": 9.97354512903096e-06, "loss": 0.6166, "step": 1474 }, { "epoch": 0.06182995713821615, "grad_norm": 1.8988313674926758, "learning_rate": 9.973475343039054e-06, "loss": 0.6524, "step": 1475 }, { "epoch": 0.061871875753225114, "grad_norm": 1.8802398443222046, "learning_rate": 9.973405465368006e-06, "loss": 0.5697, "step": 1476 }, { "epoch": 0.061913794368234076, "grad_norm": 1.9600368738174438, "learning_rate": 9.973335496019105e-06, "loss": 0.591, "step": 1477 }, { "epoch": 0.06195571298324303, "grad_norm": 2.24653697013855, "learning_rate": 9.973265434993637e-06, "loss": 0.6294, "step": 1478 }, { "epoch": 0.061997631598251994, "grad_norm": 2.0314157009124756, "learning_rate": 9.973195282292897e-06, "loss": 0.6125, "step": 1479 }, { "epoch": 0.062039550213260956, "grad_norm": 2.1847782135009766, "learning_rate": 9.973125037918178e-06, "loss": 0.6333, "step": 1480 }, { "epoch": 0.06208146882826991, "grad_norm": 2.0074551105499268, "learning_rate": 9.973054701870775e-06, "loss": 0.6087, "step": 1481 }, { "epoch": 0.062123387443278874, "grad_norm": 2.316349744796753, "learning_rate": 9.972984274151982e-06, "loss": 0.6779, "step": 1482 }, { "epoch": 0.062165306058287836, "grad_norm": 2.126746416091919, "learning_rate": 9.9729137547631e-06, "loss": 0.6012, "step": 1483 }, { "epoch": 0.06220722467329679, "grad_norm": 1.8509516716003418, "learning_rate": 9.972843143705428e-06, "loss": 0.611, "step": 1484 }, { "epoch": 0.062249143288305754, "grad_norm": 2.020709276199341, "learning_rate": 9.972772440980266e-06, "loss": 0.6801, "step": 1485 }, { "epoch": 0.062291061903314716, "grad_norm": 1.6932283639907837, "learning_rate": 9.972701646588919e-06, "loss": 0.5789, "step": 1486 }, { "epoch": 0.06233298051832367, "grad_norm": 3.207446336746216, "learning_rate": 9.972630760532693e-06, "loss": 0.6123, "step": 1487 }, { "epoch": 0.062374899133332634, "grad_norm": 1.9380404949188232, "learning_rate": 9.972559782812892e-06, "loss": 0.6114, "step": 1488 }, { "epoch": 0.062416817748341596, "grad_norm": 2.050588607788086, "learning_rate": 9.972488713430825e-06, "loss": 0.6734, "step": 1489 }, { "epoch": 0.06245873636335055, "grad_norm": 1.8303664922714233, "learning_rate": 9.972417552387804e-06, "loss": 0.5602, "step": 1490 }, { "epoch": 0.06250065497835952, "grad_norm": 2.0383758544921875, "learning_rate": 9.972346299685141e-06, "loss": 0.621, "step": 1491 }, { "epoch": 0.06254257359336847, "grad_norm": 2.032069206237793, "learning_rate": 9.972274955324146e-06, "loss": 0.6166, "step": 1492 }, { "epoch": 0.06258449220837743, "grad_norm": 1.9642927646636963, "learning_rate": 9.972203519306136e-06, "loss": 0.5981, "step": 1493 }, { "epoch": 0.0626264108233864, "grad_norm": 1.943442702293396, "learning_rate": 9.972131991632429e-06, "loss": 0.577, "step": 1494 }, { "epoch": 0.06266832943839536, "grad_norm": 2.38413405418396, "learning_rate": 9.972060372304341e-06, "loss": 0.6182, "step": 1495 }, { "epoch": 0.06271024805340432, "grad_norm": 1.8348127603530884, "learning_rate": 9.971988661323195e-06, "loss": 0.6332, "step": 1496 }, { "epoch": 0.06275216666841328, "grad_norm": 2.0237085819244385, "learning_rate": 9.971916858690313e-06, "loss": 0.5506, "step": 1497 }, { "epoch": 0.06279408528342223, "grad_norm": 2.2204232215881348, "learning_rate": 9.971844964407015e-06, "loss": 0.6517, "step": 1498 }, { "epoch": 0.06283600389843119, "grad_norm": 1.777546763420105, "learning_rate": 9.971772978474629e-06, "loss": 0.5654, "step": 1499 }, { "epoch": 0.06287792251344015, "grad_norm": 1.8838952779769897, "learning_rate": 9.971700900894481e-06, "loss": 0.5548, "step": 1500 }, { "epoch": 0.06291984112844912, "grad_norm": 1.9768786430358887, "learning_rate": 9.971628731667901e-06, "loss": 0.5959, "step": 1501 }, { "epoch": 0.06296175974345808, "grad_norm": 1.6919928789138794, "learning_rate": 9.971556470796217e-06, "loss": 0.5626, "step": 1502 }, { "epoch": 0.06300367835846704, "grad_norm": 1.9988470077514648, "learning_rate": 9.971484118280766e-06, "loss": 0.6005, "step": 1503 }, { "epoch": 0.06304559697347599, "grad_norm": 2.054713726043701, "learning_rate": 9.971411674122875e-06, "loss": 0.656, "step": 1504 }, { "epoch": 0.06308751558848495, "grad_norm": 2.00917911529541, "learning_rate": 9.971339138323883e-06, "loss": 0.5633, "step": 1505 }, { "epoch": 0.06312943420349391, "grad_norm": 1.975835919380188, "learning_rate": 9.971266510885126e-06, "loss": 0.5777, "step": 1506 }, { "epoch": 0.06317135281850288, "grad_norm": 2.350722551345825, "learning_rate": 9.971193791807945e-06, "loss": 0.625, "step": 1507 }, { "epoch": 0.06321327143351184, "grad_norm": 2.0642361640930176, "learning_rate": 9.971120981093678e-06, "loss": 0.6164, "step": 1508 }, { "epoch": 0.0632551900485208, "grad_norm": 1.9264179468154907, "learning_rate": 9.971048078743669e-06, "loss": 0.5913, "step": 1509 }, { "epoch": 0.06329710866352976, "grad_norm": 2.1075751781463623, "learning_rate": 9.97097508475926e-06, "loss": 0.6452, "step": 1510 }, { "epoch": 0.06333902727853871, "grad_norm": 1.8642053604125977, "learning_rate": 9.970901999141797e-06, "loss": 0.6328, "step": 1511 }, { "epoch": 0.06338094589354767, "grad_norm": 1.8997160196304321, "learning_rate": 9.97082882189263e-06, "loss": 0.6371, "step": 1512 }, { "epoch": 0.06342286450855664, "grad_norm": 2.0351667404174805, "learning_rate": 9.970755553013101e-06, "loss": 0.6814, "step": 1513 }, { "epoch": 0.0634647831235656, "grad_norm": 1.5075997114181519, "learning_rate": 9.970682192504568e-06, "loss": 0.5491, "step": 1514 }, { "epoch": 0.06350670173857456, "grad_norm": 1.5986747741699219, "learning_rate": 9.970608740368381e-06, "loss": 0.5646, "step": 1515 }, { "epoch": 0.06354862035358352, "grad_norm": 1.7263225317001343, "learning_rate": 9.970535196605894e-06, "loss": 0.5897, "step": 1516 }, { "epoch": 0.06359053896859247, "grad_norm": 1.8060011863708496, "learning_rate": 9.970461561218461e-06, "loss": 0.5934, "step": 1517 }, { "epoch": 0.06363245758360143, "grad_norm": 1.7002872228622437, "learning_rate": 9.970387834207442e-06, "loss": 0.5669, "step": 1518 }, { "epoch": 0.0636743761986104, "grad_norm": 1.614715337753296, "learning_rate": 9.970314015574193e-06, "loss": 0.569, "step": 1519 }, { "epoch": 0.06371629481361936, "grad_norm": 1.8099851608276367, "learning_rate": 9.970240105320078e-06, "loss": 0.5893, "step": 1520 }, { "epoch": 0.06375821342862832, "grad_norm": 1.6587563753128052, "learning_rate": 9.970166103446457e-06, "loss": 0.5689, "step": 1521 }, { "epoch": 0.06380013204363728, "grad_norm": 2.636465311050415, "learning_rate": 9.970092009954695e-06, "loss": 0.647, "step": 1522 }, { "epoch": 0.06384205065864623, "grad_norm": 1.8769365549087524, "learning_rate": 9.970017824846157e-06, "loss": 0.6294, "step": 1523 }, { "epoch": 0.0638839692736552, "grad_norm": 1.8323739767074585, "learning_rate": 9.969943548122212e-06, "loss": 0.565, "step": 1524 }, { "epoch": 0.06392588788866416, "grad_norm": 1.724883794784546, "learning_rate": 9.969869179784228e-06, "loss": 0.5778, "step": 1525 }, { "epoch": 0.06396780650367312, "grad_norm": 2.2568840980529785, "learning_rate": 9.969794719833577e-06, "loss": 0.5904, "step": 1526 }, { "epoch": 0.06400972511868208, "grad_norm": 2.1718692779541016, "learning_rate": 9.96972016827163e-06, "loss": 0.6023, "step": 1527 }, { "epoch": 0.06405164373369104, "grad_norm": 1.644879698753357, "learning_rate": 9.969645525099763e-06, "loss": 0.6267, "step": 1528 }, { "epoch": 0.06409356234869999, "grad_norm": 1.8799303770065308, "learning_rate": 9.96957079031935e-06, "loss": 0.5986, "step": 1529 }, { "epoch": 0.06413548096370895, "grad_norm": 1.892151951789856, "learning_rate": 9.96949596393177e-06, "loss": 0.6023, "step": 1530 }, { "epoch": 0.06417739957871792, "grad_norm": 3.8696506023406982, "learning_rate": 9.969421045938404e-06, "loss": 0.5574, "step": 1531 }, { "epoch": 0.06421931819372688, "grad_norm": 1.7773470878601074, "learning_rate": 9.969346036340628e-06, "loss": 0.685, "step": 1532 }, { "epoch": 0.06426123680873584, "grad_norm": 1.8113553524017334, "learning_rate": 9.96927093513983e-06, "loss": 0.5895, "step": 1533 }, { "epoch": 0.0643031554237448, "grad_norm": 1.7688827514648438, "learning_rate": 9.96919574233739e-06, "loss": 0.6214, "step": 1534 }, { "epoch": 0.06434507403875377, "grad_norm": 1.99970281124115, "learning_rate": 9.969120457934697e-06, "loss": 0.6407, "step": 1535 }, { "epoch": 0.06438699265376271, "grad_norm": 1.757228970527649, "learning_rate": 9.96904508193314e-06, "loss": 0.5446, "step": 1536 }, { "epoch": 0.06442891126877168, "grad_norm": 1.7796037197113037, "learning_rate": 9.968969614334104e-06, "loss": 0.6315, "step": 1537 }, { "epoch": 0.06447082988378064, "grad_norm": 1.6750174760818481, "learning_rate": 9.968894055138984e-06, "loss": 0.5797, "step": 1538 }, { "epoch": 0.0645127484987896, "grad_norm": 2.0458624362945557, "learning_rate": 9.96881840434917e-06, "loss": 0.6555, "step": 1539 }, { "epoch": 0.06455466711379856, "grad_norm": 1.8733863830566406, "learning_rate": 9.96874266196606e-06, "loss": 0.5678, "step": 1540 }, { "epoch": 0.06459658572880753, "grad_norm": 1.7041610479354858, "learning_rate": 9.968666827991045e-06, "loss": 0.5874, "step": 1541 }, { "epoch": 0.06463850434381647, "grad_norm": 2.044769525527954, "learning_rate": 9.968590902425528e-06, "loss": 0.6293, "step": 1542 }, { "epoch": 0.06468042295882544, "grad_norm": 1.744783639907837, "learning_rate": 9.968514885270907e-06, "loss": 0.5627, "step": 1543 }, { "epoch": 0.0647223415738344, "grad_norm": 2.044527292251587, "learning_rate": 9.968438776528581e-06, "loss": 0.594, "step": 1544 }, { "epoch": 0.06476426018884336, "grad_norm": 1.804419994354248, "learning_rate": 9.968362576199956e-06, "loss": 0.6007, "step": 1545 }, { "epoch": 0.06480617880385232, "grad_norm": 1.8001600503921509, "learning_rate": 9.968286284286437e-06, "loss": 0.6213, "step": 1546 }, { "epoch": 0.06484809741886129, "grad_norm": 1.8164454698562622, "learning_rate": 9.968209900789428e-06, "loss": 0.556, "step": 1547 }, { "epoch": 0.06489001603387023, "grad_norm": 1.6610652208328247, "learning_rate": 9.968133425710335e-06, "loss": 0.5525, "step": 1548 }, { "epoch": 0.0649319346488792, "grad_norm": 2.0870296955108643, "learning_rate": 9.968056859050574e-06, "loss": 0.6055, "step": 1549 }, { "epoch": 0.06497385326388816, "grad_norm": 2.3889079093933105, "learning_rate": 9.967980200811551e-06, "loss": 0.6375, "step": 1550 }, { "epoch": 0.06501577187889712, "grad_norm": 1.9462451934814453, "learning_rate": 9.967903450994683e-06, "loss": 0.5105, "step": 1551 }, { "epoch": 0.06505769049390608, "grad_norm": 1.6530399322509766, "learning_rate": 9.967826609601381e-06, "loss": 0.5277, "step": 1552 }, { "epoch": 0.06509960910891505, "grad_norm": 1.9338657855987549, "learning_rate": 9.967749676633063e-06, "loss": 0.6164, "step": 1553 }, { "epoch": 0.065141527723924, "grad_norm": 1.836657166481018, "learning_rate": 9.967672652091148e-06, "loss": 0.6456, "step": 1554 }, { "epoch": 0.06518344633893296, "grad_norm": 1.9397844076156616, "learning_rate": 9.967595535977056e-06, "loss": 0.5881, "step": 1555 }, { "epoch": 0.06522536495394192, "grad_norm": 3.0352206230163574, "learning_rate": 9.967518328292208e-06, "loss": 0.6395, "step": 1556 }, { "epoch": 0.06526728356895088, "grad_norm": 2.021111011505127, "learning_rate": 9.967441029038025e-06, "loss": 0.6162, "step": 1557 }, { "epoch": 0.06530920218395984, "grad_norm": 2.043893814086914, "learning_rate": 9.967363638215935e-06, "loss": 0.6057, "step": 1558 }, { "epoch": 0.0653511207989688, "grad_norm": 1.826809287071228, "learning_rate": 9.967286155827364e-06, "loss": 0.6267, "step": 1559 }, { "epoch": 0.06539303941397777, "grad_norm": 2.2809977531433105, "learning_rate": 9.96720858187374e-06, "loss": 0.6531, "step": 1560 }, { "epoch": 0.06543495802898672, "grad_norm": 2.1746227741241455, "learning_rate": 9.96713091635649e-06, "loss": 0.6081, "step": 1561 }, { "epoch": 0.06547687664399568, "grad_norm": 1.8712083101272583, "learning_rate": 9.96705315927705e-06, "loss": 0.599, "step": 1562 }, { "epoch": 0.06551879525900464, "grad_norm": 1.626783013343811, "learning_rate": 9.966975310636854e-06, "loss": 0.4953, "step": 1563 }, { "epoch": 0.0655607138740136, "grad_norm": 1.7273459434509277, "learning_rate": 9.966897370437333e-06, "loss": 0.5763, "step": 1564 }, { "epoch": 0.06560263248902257, "grad_norm": 1.7428842782974243, "learning_rate": 9.966819338679925e-06, "loss": 0.5756, "step": 1565 }, { "epoch": 0.06564455110403153, "grad_norm": 2.1108145713806152, "learning_rate": 9.96674121536607e-06, "loss": 0.5525, "step": 1566 }, { "epoch": 0.06568646971904048, "grad_norm": 1.9381829500198364, "learning_rate": 9.966663000497206e-06, "loss": 0.6057, "step": 1567 }, { "epoch": 0.06572838833404944, "grad_norm": 1.8819705247879028, "learning_rate": 9.966584694074777e-06, "loss": 0.5791, "step": 1568 }, { "epoch": 0.0657703069490584, "grad_norm": 1.8687307834625244, "learning_rate": 9.966506296100225e-06, "loss": 0.6219, "step": 1569 }, { "epoch": 0.06581222556406736, "grad_norm": 2.5722360610961914, "learning_rate": 9.966427806574995e-06, "loss": 0.6424, "step": 1570 }, { "epoch": 0.06585414417907633, "grad_norm": 2.173868417739868, "learning_rate": 9.966349225500534e-06, "loss": 0.5733, "step": 1571 }, { "epoch": 0.06589606279408529, "grad_norm": 1.9311443567276, "learning_rate": 9.96627055287829e-06, "loss": 0.6154, "step": 1572 }, { "epoch": 0.06593798140909424, "grad_norm": 2.57934308052063, "learning_rate": 9.966191788709716e-06, "loss": 0.6199, "step": 1573 }, { "epoch": 0.0659799000241032, "grad_norm": 1.9724740982055664, "learning_rate": 9.966112932996262e-06, "loss": 0.6095, "step": 1574 }, { "epoch": 0.06602181863911216, "grad_norm": 1.7474085092544556, "learning_rate": 9.96603398573938e-06, "loss": 0.5095, "step": 1575 }, { "epoch": 0.06606373725412112, "grad_norm": 2.0029146671295166, "learning_rate": 9.965954946940526e-06, "loss": 0.6203, "step": 1576 }, { "epoch": 0.06610565586913009, "grad_norm": 2.149632453918457, "learning_rate": 9.96587581660116e-06, "loss": 0.6218, "step": 1577 }, { "epoch": 0.06614757448413905, "grad_norm": 2.2672712802886963, "learning_rate": 9.965796594722736e-06, "loss": 0.7161, "step": 1578 }, { "epoch": 0.066189493099148, "grad_norm": 2.30348539352417, "learning_rate": 9.96571728130672e-06, "loss": 0.5654, "step": 1579 }, { "epoch": 0.06623141171415696, "grad_norm": 1.7360103130340576, "learning_rate": 9.965637876354568e-06, "loss": 0.6084, "step": 1580 }, { "epoch": 0.06627333032916592, "grad_norm": 1.751200556755066, "learning_rate": 9.965558379867748e-06, "loss": 0.5856, "step": 1581 }, { "epoch": 0.06631524894417488, "grad_norm": 2.1457719802856445, "learning_rate": 9.965478791847725e-06, "loss": 0.5974, "step": 1582 }, { "epoch": 0.06635716755918385, "grad_norm": 2.2155494689941406, "learning_rate": 9.965399112295962e-06, "loss": 0.6029, "step": 1583 }, { "epoch": 0.06639908617419281, "grad_norm": 1.8542263507843018, "learning_rate": 9.965319341213932e-06, "loss": 0.6493, "step": 1584 }, { "epoch": 0.06644100478920177, "grad_norm": 1.8525807857513428, "learning_rate": 9.965239478603106e-06, "loss": 0.5755, "step": 1585 }, { "epoch": 0.06648292340421072, "grad_norm": 1.7040561437606812, "learning_rate": 9.965159524464954e-06, "loss": 0.6395, "step": 1586 }, { "epoch": 0.06652484201921968, "grad_norm": 2.0988149642944336, "learning_rate": 9.965079478800948e-06, "loss": 0.6471, "step": 1587 }, { "epoch": 0.06656676063422864, "grad_norm": 1.7646338939666748, "learning_rate": 9.964999341612568e-06, "loss": 0.6187, "step": 1588 }, { "epoch": 0.0666086792492376, "grad_norm": 2.2969839572906494, "learning_rate": 9.964919112901288e-06, "loss": 0.6051, "step": 1589 }, { "epoch": 0.06665059786424657, "grad_norm": 2.080603837966919, "learning_rate": 9.96483879266859e-06, "loss": 0.6225, "step": 1590 }, { "epoch": 0.06669251647925553, "grad_norm": 1.7988615036010742, "learning_rate": 9.96475838091595e-06, "loss": 0.5545, "step": 1591 }, { "epoch": 0.06673443509426448, "grad_norm": 1.6528726816177368, "learning_rate": 9.964677877644854e-06, "loss": 0.5797, "step": 1592 }, { "epoch": 0.06677635370927344, "grad_norm": 1.7879201173782349, "learning_rate": 9.964597282856784e-06, "loss": 0.5971, "step": 1593 }, { "epoch": 0.0668182723242824, "grad_norm": 2.129946708679199, "learning_rate": 9.964516596553226e-06, "loss": 0.6312, "step": 1594 }, { "epoch": 0.06686019093929137, "grad_norm": 1.8413814306259155, "learning_rate": 9.96443581873567e-06, "loss": 0.6115, "step": 1595 }, { "epoch": 0.06690210955430033, "grad_norm": 1.9244859218597412, "learning_rate": 9.964354949405602e-06, "loss": 0.5589, "step": 1596 }, { "epoch": 0.06694402816930929, "grad_norm": 2.0741376876831055, "learning_rate": 9.964273988564512e-06, "loss": 0.611, "step": 1597 }, { "epoch": 0.06698594678431824, "grad_norm": 2.0076310634613037, "learning_rate": 9.964192936213895e-06, "loss": 0.5944, "step": 1598 }, { "epoch": 0.0670278653993272, "grad_norm": 2.421398639678955, "learning_rate": 9.964111792355244e-06, "loss": 0.5984, "step": 1599 }, { "epoch": 0.06706978401433616, "grad_norm": 2.112582206726074, "learning_rate": 9.964030556990054e-06, "loss": 0.6649, "step": 1600 }, { "epoch": 0.06711170262934513, "grad_norm": 2.3441641330718994, "learning_rate": 9.963949230119824e-06, "loss": 0.6558, "step": 1601 }, { "epoch": 0.06715362124435409, "grad_norm": 1.9869433641433716, "learning_rate": 9.963867811746054e-06, "loss": 0.5931, "step": 1602 }, { "epoch": 0.06719553985936305, "grad_norm": 1.996084213256836, "learning_rate": 9.963786301870241e-06, "loss": 0.6453, "step": 1603 }, { "epoch": 0.067237458474372, "grad_norm": 1.731552004814148, "learning_rate": 9.963704700493891e-06, "loss": 0.5395, "step": 1604 }, { "epoch": 0.06727937708938096, "grad_norm": 2.1172618865966797, "learning_rate": 9.963623007618506e-06, "loss": 0.5955, "step": 1605 }, { "epoch": 0.06732129570438992, "grad_norm": 2.063636302947998, "learning_rate": 9.963541223245592e-06, "loss": 0.6425, "step": 1606 }, { "epoch": 0.06736321431939889, "grad_norm": 2.060272216796875, "learning_rate": 9.963459347376658e-06, "loss": 0.6001, "step": 1607 }, { "epoch": 0.06740513293440785, "grad_norm": 3.5145888328552246, "learning_rate": 9.963377380013214e-06, "loss": 0.625, "step": 1608 }, { "epoch": 0.06744705154941681, "grad_norm": 2.282832384109497, "learning_rate": 9.96329532115677e-06, "loss": 0.5632, "step": 1609 }, { "epoch": 0.06748897016442577, "grad_norm": 1.807685375213623, "learning_rate": 9.963213170808836e-06, "loss": 0.5898, "step": 1610 }, { "epoch": 0.06753088877943472, "grad_norm": 1.807379126548767, "learning_rate": 9.963130928970932e-06, "loss": 0.6435, "step": 1611 }, { "epoch": 0.06757280739444368, "grad_norm": 1.8541443347930908, "learning_rate": 9.963048595644567e-06, "loss": 0.5427, "step": 1612 }, { "epoch": 0.06761472600945265, "grad_norm": 2.0661680698394775, "learning_rate": 9.962966170831265e-06, "loss": 0.6223, "step": 1613 }, { "epoch": 0.06765664462446161, "grad_norm": 1.5843942165374756, "learning_rate": 9.96288365453254e-06, "loss": 0.5597, "step": 1614 }, { "epoch": 0.06769856323947057, "grad_norm": 2.0988428592681885, "learning_rate": 9.962801046749918e-06, "loss": 0.6171, "step": 1615 }, { "epoch": 0.06774048185447953, "grad_norm": 1.8806699514389038, "learning_rate": 9.96271834748492e-06, "loss": 0.5948, "step": 1616 }, { "epoch": 0.06778240046948848, "grad_norm": 1.9985541105270386, "learning_rate": 9.962635556739068e-06, "loss": 0.5857, "step": 1617 }, { "epoch": 0.06782431908449744, "grad_norm": 1.9491472244262695, "learning_rate": 9.962552674513892e-06, "loss": 0.6352, "step": 1618 }, { "epoch": 0.06786623769950641, "grad_norm": 1.7968761920928955, "learning_rate": 9.962469700810917e-06, "loss": 0.6155, "step": 1619 }, { "epoch": 0.06790815631451537, "grad_norm": 1.8632582426071167, "learning_rate": 9.962386635631674e-06, "loss": 0.5669, "step": 1620 }, { "epoch": 0.06795007492952433, "grad_norm": 1.95482337474823, "learning_rate": 9.962303478977693e-06, "loss": 0.6627, "step": 1621 }, { "epoch": 0.0679919935445333, "grad_norm": 1.915657877922058, "learning_rate": 9.962220230850507e-06, "loss": 0.6214, "step": 1622 }, { "epoch": 0.06803391215954224, "grad_norm": 1.766395092010498, "learning_rate": 9.962136891251653e-06, "loss": 0.5538, "step": 1623 }, { "epoch": 0.0680758307745512, "grad_norm": 1.9563506841659546, "learning_rate": 9.962053460182664e-06, "loss": 0.6049, "step": 1624 }, { "epoch": 0.06811774938956017, "grad_norm": 2.299910068511963, "learning_rate": 9.96196993764508e-06, "loss": 0.644, "step": 1625 }, { "epoch": 0.06815966800456913, "grad_norm": 1.8048542737960815, "learning_rate": 9.961886323640439e-06, "loss": 0.6757, "step": 1626 }, { "epoch": 0.06820158661957809, "grad_norm": 1.8361070156097412, "learning_rate": 9.961802618170285e-06, "loss": 0.5833, "step": 1627 }, { "epoch": 0.06824350523458705, "grad_norm": 1.936676263809204, "learning_rate": 9.961718821236156e-06, "loss": 0.6234, "step": 1628 }, { "epoch": 0.068285423849596, "grad_norm": 1.8144443035125732, "learning_rate": 9.961634932839603e-06, "loss": 0.6055, "step": 1629 }, { "epoch": 0.06832734246460496, "grad_norm": 1.8409453630447388, "learning_rate": 9.961550952982169e-06, "loss": 0.624, "step": 1630 }, { "epoch": 0.06836926107961393, "grad_norm": 2.417006731033325, "learning_rate": 9.9614668816654e-06, "loss": 0.6629, "step": 1631 }, { "epoch": 0.06841117969462289, "grad_norm": 2.0458288192749023, "learning_rate": 9.96138271889085e-06, "loss": 0.6136, "step": 1632 }, { "epoch": 0.06845309830963185, "grad_norm": 1.925057053565979, "learning_rate": 9.96129846466007e-06, "loss": 0.6295, "step": 1633 }, { "epoch": 0.06849501692464081, "grad_norm": 1.990004539489746, "learning_rate": 9.961214118974607e-06, "loss": 0.6246, "step": 1634 }, { "epoch": 0.06853693553964978, "grad_norm": 4.075019836425781, "learning_rate": 9.961129681836023e-06, "loss": 0.6092, "step": 1635 }, { "epoch": 0.06857885415465872, "grad_norm": 1.977342128753662, "learning_rate": 9.961045153245873e-06, "loss": 0.6361, "step": 1636 }, { "epoch": 0.06862077276966769, "grad_norm": 1.759106159210205, "learning_rate": 9.960960533205713e-06, "loss": 0.5974, "step": 1637 }, { "epoch": 0.06866269138467665, "grad_norm": 1.9170104265213013, "learning_rate": 9.960875821717104e-06, "loss": 0.6618, "step": 1638 }, { "epoch": 0.06870460999968561, "grad_norm": 1.7250962257385254, "learning_rate": 9.960791018781607e-06, "loss": 0.5844, "step": 1639 }, { "epoch": 0.06874652861469457, "grad_norm": 1.8797639608383179, "learning_rate": 9.960706124400785e-06, "loss": 0.625, "step": 1640 }, { "epoch": 0.06878844722970354, "grad_norm": 2.343629837036133, "learning_rate": 9.960621138576205e-06, "loss": 0.6071, "step": 1641 }, { "epoch": 0.06883036584471248, "grad_norm": 1.8154176473617554, "learning_rate": 9.960536061309431e-06, "loss": 0.5502, "step": 1642 }, { "epoch": 0.06887228445972145, "grad_norm": 1.8360791206359863, "learning_rate": 9.960450892602035e-06, "loss": 0.5801, "step": 1643 }, { "epoch": 0.06891420307473041, "grad_norm": 1.9865788221359253, "learning_rate": 9.960365632455581e-06, "loss": 0.631, "step": 1644 }, { "epoch": 0.06895612168973937, "grad_norm": 1.8610825538635254, "learning_rate": 9.960280280871647e-06, "loss": 0.5714, "step": 1645 }, { "epoch": 0.06899804030474833, "grad_norm": 1.8867300748825073, "learning_rate": 9.960194837851801e-06, "loss": 0.5943, "step": 1646 }, { "epoch": 0.0690399589197573, "grad_norm": 1.915864109992981, "learning_rate": 9.960109303397621e-06, "loss": 0.5808, "step": 1647 }, { "epoch": 0.06908187753476625, "grad_norm": 2.016361713409424, "learning_rate": 9.960023677510684e-06, "loss": 0.6001, "step": 1648 }, { "epoch": 0.06912379614977521, "grad_norm": 1.8578771352767944, "learning_rate": 9.959937960192567e-06, "loss": 0.6158, "step": 1649 }, { "epoch": 0.06916571476478417, "grad_norm": 1.9849718809127808, "learning_rate": 9.95985215144485e-06, "loss": 0.582, "step": 1650 }, { "epoch": 0.06920763337979313, "grad_norm": 2.3543643951416016, "learning_rate": 9.959766251269116e-06, "loss": 0.6074, "step": 1651 }, { "epoch": 0.0692495519948021, "grad_norm": 1.8421761989593506, "learning_rate": 9.959680259666948e-06, "loss": 0.57, "step": 1652 }, { "epoch": 0.06929147060981106, "grad_norm": 2.1727492809295654, "learning_rate": 9.959594176639931e-06, "loss": 0.6085, "step": 1653 }, { "epoch": 0.06933338922482, "grad_norm": 2.0250165462493896, "learning_rate": 9.959508002189653e-06, "loss": 0.6867, "step": 1654 }, { "epoch": 0.06937530783982897, "grad_norm": 1.8077000379562378, "learning_rate": 9.9594217363177e-06, "loss": 0.529, "step": 1655 }, { "epoch": 0.06941722645483793, "grad_norm": 1.9095064401626587, "learning_rate": 9.959335379025664e-06, "loss": 0.6132, "step": 1656 }, { "epoch": 0.06945914506984689, "grad_norm": 2.01724910736084, "learning_rate": 9.959248930315133e-06, "loss": 0.6381, "step": 1657 }, { "epoch": 0.06950106368485585, "grad_norm": 1.8363760709762573, "learning_rate": 9.959162390187708e-06, "loss": 0.6375, "step": 1658 }, { "epoch": 0.06954298229986482, "grad_norm": 2.1625990867614746, "learning_rate": 9.959075758644979e-06, "loss": 0.5695, "step": 1659 }, { "epoch": 0.06958490091487378, "grad_norm": 1.9234035015106201, "learning_rate": 9.958989035688545e-06, "loss": 0.617, "step": 1660 }, { "epoch": 0.06962681952988273, "grad_norm": 1.9868340492248535, "learning_rate": 9.958902221320001e-06, "loss": 0.676, "step": 1661 }, { "epoch": 0.06966873814489169, "grad_norm": 2.050377607345581, "learning_rate": 9.958815315540953e-06, "loss": 0.6308, "step": 1662 }, { "epoch": 0.06971065675990065, "grad_norm": 2.534050464630127, "learning_rate": 9.958728318352998e-06, "loss": 0.591, "step": 1663 }, { "epoch": 0.06975257537490961, "grad_norm": 1.6819071769714355, "learning_rate": 9.958641229757741e-06, "loss": 0.6087, "step": 1664 }, { "epoch": 0.06979449398991858, "grad_norm": 1.5610935688018799, "learning_rate": 9.95855404975679e-06, "loss": 0.5505, "step": 1665 }, { "epoch": 0.06983641260492754, "grad_norm": 1.7220449447631836, "learning_rate": 9.95846677835175e-06, "loss": 0.5931, "step": 1666 }, { "epoch": 0.06987833121993649, "grad_norm": 1.9586851596832275, "learning_rate": 9.95837941554423e-06, "loss": 0.6211, "step": 1667 }, { "epoch": 0.06992024983494545, "grad_norm": 2.087082862854004, "learning_rate": 9.95829196133584e-06, "loss": 0.5905, "step": 1668 }, { "epoch": 0.06996216844995441, "grad_norm": 2.1274802684783936, "learning_rate": 9.958204415728192e-06, "loss": 0.5877, "step": 1669 }, { "epoch": 0.07000408706496337, "grad_norm": 1.7863484621047974, "learning_rate": 9.958116778722899e-06, "loss": 0.5821, "step": 1670 }, { "epoch": 0.07004600567997234, "grad_norm": 1.8591324090957642, "learning_rate": 9.95802905032158e-06, "loss": 0.5869, "step": 1671 }, { "epoch": 0.0700879242949813, "grad_norm": 1.573367714881897, "learning_rate": 9.957941230525846e-06, "loss": 0.538, "step": 1672 }, { "epoch": 0.07012984290999025, "grad_norm": 1.8794922828674316, "learning_rate": 9.957853319337323e-06, "loss": 0.6184, "step": 1673 }, { "epoch": 0.07017176152499921, "grad_norm": 2.548551321029663, "learning_rate": 9.957765316757628e-06, "loss": 0.5941, "step": 1674 }, { "epoch": 0.07021368014000817, "grad_norm": 1.6970378160476685, "learning_rate": 9.957677222788383e-06, "loss": 0.5461, "step": 1675 }, { "epoch": 0.07025559875501713, "grad_norm": 1.7674747705459595, "learning_rate": 9.95758903743121e-06, "loss": 0.5689, "step": 1676 }, { "epoch": 0.0702975173700261, "grad_norm": 1.7730556726455688, "learning_rate": 9.957500760687738e-06, "loss": 0.5439, "step": 1677 }, { "epoch": 0.07033943598503506, "grad_norm": 2.086215019226074, "learning_rate": 9.957412392559593e-06, "loss": 0.5831, "step": 1678 }, { "epoch": 0.07038135460004401, "grad_norm": 2.1302177906036377, "learning_rate": 9.957323933048404e-06, "loss": 0.6125, "step": 1679 }, { "epoch": 0.07042327321505297, "grad_norm": 2.090394973754883, "learning_rate": 9.957235382155803e-06, "loss": 0.6617, "step": 1680 }, { "epoch": 0.07046519183006193, "grad_norm": 1.909549355506897, "learning_rate": 9.957146739883419e-06, "loss": 0.5567, "step": 1681 }, { "epoch": 0.0705071104450709, "grad_norm": 2.733907461166382, "learning_rate": 9.957058006232889e-06, "loss": 0.596, "step": 1682 }, { "epoch": 0.07054902906007986, "grad_norm": 1.9229308366775513, "learning_rate": 9.956969181205846e-06, "loss": 0.5861, "step": 1683 }, { "epoch": 0.07059094767508882, "grad_norm": 2.8304293155670166, "learning_rate": 9.95688026480393e-06, "loss": 0.575, "step": 1684 }, { "epoch": 0.07063286629009778, "grad_norm": 1.9943381547927856, "learning_rate": 9.95679125702878e-06, "loss": 0.627, "step": 1685 }, { "epoch": 0.07067478490510673, "grad_norm": 1.8862025737762451, "learning_rate": 9.956702157882035e-06, "loss": 0.5796, "step": 1686 }, { "epoch": 0.07071670352011569, "grad_norm": 2.0174977779388428, "learning_rate": 9.956612967365338e-06, "loss": 0.6439, "step": 1687 }, { "epoch": 0.07075862213512465, "grad_norm": 2.1953275203704834, "learning_rate": 9.956523685480333e-06, "loss": 0.582, "step": 1688 }, { "epoch": 0.07080054075013362, "grad_norm": 1.8827742338180542, "learning_rate": 9.956434312228667e-06, "loss": 0.5898, "step": 1689 }, { "epoch": 0.07084245936514258, "grad_norm": 1.7428003549575806, "learning_rate": 9.956344847611984e-06, "loss": 0.5568, "step": 1690 }, { "epoch": 0.07088437798015154, "grad_norm": 1.765688180923462, "learning_rate": 9.95625529163194e-06, "loss": 0.6284, "step": 1691 }, { "epoch": 0.07092629659516049, "grad_norm": 2.0231540203094482, "learning_rate": 9.956165644290177e-06, "loss": 0.612, "step": 1692 }, { "epoch": 0.07096821521016945, "grad_norm": 8.256461143493652, "learning_rate": 9.956075905588355e-06, "loss": 0.5552, "step": 1693 }, { "epoch": 0.07101013382517841, "grad_norm": 1.7750084400177002, "learning_rate": 9.955986075528125e-06, "loss": 0.6165, "step": 1694 }, { "epoch": 0.07105205244018738, "grad_norm": 1.9044965505599976, "learning_rate": 9.955896154111143e-06, "loss": 0.5208, "step": 1695 }, { "epoch": 0.07109397105519634, "grad_norm": 1.840936541557312, "learning_rate": 9.955806141339065e-06, "loss": 0.5955, "step": 1696 }, { "epoch": 0.0711358896702053, "grad_norm": 1.9363934993743896, "learning_rate": 9.955716037213553e-06, "loss": 0.5694, "step": 1697 }, { "epoch": 0.07117780828521425, "grad_norm": 2.1667850017547607, "learning_rate": 9.955625841736267e-06, "loss": 0.5744, "step": 1698 }, { "epoch": 0.07121972690022321, "grad_norm": 1.8947581052780151, "learning_rate": 9.95553555490887e-06, "loss": 0.572, "step": 1699 }, { "epoch": 0.07126164551523217, "grad_norm": 1.9760385751724243, "learning_rate": 9.955445176733026e-06, "loss": 0.6354, "step": 1700 }, { "epoch": 0.07130356413024114, "grad_norm": 1.823394536972046, "learning_rate": 9.9553547072104e-06, "loss": 0.5274, "step": 1701 }, { "epoch": 0.0713454827452501, "grad_norm": 2.080801486968994, "learning_rate": 9.95526414634266e-06, "loss": 0.6471, "step": 1702 }, { "epoch": 0.07138740136025906, "grad_norm": 1.956143856048584, "learning_rate": 9.955173494131478e-06, "loss": 0.578, "step": 1703 }, { "epoch": 0.07142931997526801, "grad_norm": 1.797631859779358, "learning_rate": 9.955082750578522e-06, "loss": 0.6682, "step": 1704 }, { "epoch": 0.07147123859027697, "grad_norm": 9.178648948669434, "learning_rate": 9.954991915685465e-06, "loss": 0.6087, "step": 1705 }, { "epoch": 0.07151315720528593, "grad_norm": 1.9245249032974243, "learning_rate": 9.954900989453982e-06, "loss": 0.5857, "step": 1706 }, { "epoch": 0.0715550758202949, "grad_norm": 1.8554751873016357, "learning_rate": 9.95480997188575e-06, "loss": 0.61, "step": 1707 }, { "epoch": 0.07159699443530386, "grad_norm": 2.1201090812683105, "learning_rate": 9.954718862982446e-06, "loss": 0.5998, "step": 1708 }, { "epoch": 0.07163891305031282, "grad_norm": 1.7398521900177002, "learning_rate": 9.954627662745749e-06, "loss": 0.5808, "step": 1709 }, { "epoch": 0.07168083166532178, "grad_norm": 1.8780196905136108, "learning_rate": 9.954536371177342e-06, "loss": 0.5796, "step": 1710 }, { "epoch": 0.07172275028033073, "grad_norm": 1.8691776990890503, "learning_rate": 9.954444988278904e-06, "loss": 0.5679, "step": 1711 }, { "epoch": 0.0717646688953397, "grad_norm": 1.9273314476013184, "learning_rate": 9.954353514052124e-06, "loss": 0.7085, "step": 1712 }, { "epoch": 0.07180658751034866, "grad_norm": 1.9086098670959473, "learning_rate": 9.954261948498684e-06, "loss": 0.6017, "step": 1713 }, { "epoch": 0.07184850612535762, "grad_norm": 1.738559365272522, "learning_rate": 9.954170291620276e-06, "loss": 0.6739, "step": 1714 }, { "epoch": 0.07189042474036658, "grad_norm": 1.8054418563842773, "learning_rate": 9.954078543418587e-06, "loss": 0.5849, "step": 1715 }, { "epoch": 0.07193234335537554, "grad_norm": 1.717734456062317, "learning_rate": 9.95398670389531e-06, "loss": 0.6593, "step": 1716 }, { "epoch": 0.07197426197038449, "grad_norm": 2.2027125358581543, "learning_rate": 9.953894773052136e-06, "loss": 0.6895, "step": 1717 }, { "epoch": 0.07201618058539346, "grad_norm": 1.8982250690460205, "learning_rate": 9.953802750890759e-06, "loss": 0.5707, "step": 1718 }, { "epoch": 0.07205809920040242, "grad_norm": 1.908760905265808, "learning_rate": 9.953710637412877e-06, "loss": 0.5555, "step": 1719 }, { "epoch": 0.07210001781541138, "grad_norm": 2.19976806640625, "learning_rate": 9.953618432620189e-06, "loss": 0.6471, "step": 1720 }, { "epoch": 0.07214193643042034, "grad_norm": 1.7898669242858887, "learning_rate": 9.953526136514393e-06, "loss": 0.5723, "step": 1721 }, { "epoch": 0.0721838550454293, "grad_norm": 1.8679451942443848, "learning_rate": 9.95343374909719e-06, "loss": 0.5976, "step": 1722 }, { "epoch": 0.07222577366043825, "grad_norm": 2.0140798091888428, "learning_rate": 9.953341270370285e-06, "loss": 0.5538, "step": 1723 }, { "epoch": 0.07226769227544722, "grad_norm": 1.9277124404907227, "learning_rate": 9.95324870033538e-06, "loss": 0.5844, "step": 1724 }, { "epoch": 0.07230961089045618, "grad_norm": 1.81705904006958, "learning_rate": 9.953156038994184e-06, "loss": 0.6194, "step": 1725 }, { "epoch": 0.07235152950546514, "grad_norm": 1.6600849628448486, "learning_rate": 9.953063286348404e-06, "loss": 0.541, "step": 1726 }, { "epoch": 0.0723934481204741, "grad_norm": 2.1029932498931885, "learning_rate": 9.952970442399749e-06, "loss": 0.6992, "step": 1727 }, { "epoch": 0.07243536673548306, "grad_norm": 1.96725594997406, "learning_rate": 9.952877507149931e-06, "loss": 0.5859, "step": 1728 }, { "epoch": 0.07247728535049201, "grad_norm": 1.7713696956634521, "learning_rate": 9.952784480600665e-06, "loss": 0.5359, "step": 1729 }, { "epoch": 0.07251920396550098, "grad_norm": 2.174532651901245, "learning_rate": 9.952691362753663e-06, "loss": 0.6108, "step": 1730 }, { "epoch": 0.07256112258050994, "grad_norm": 2.070646047592163, "learning_rate": 9.952598153610643e-06, "loss": 0.6234, "step": 1731 }, { "epoch": 0.0726030411955189, "grad_norm": 1.7397512197494507, "learning_rate": 9.952504853173322e-06, "loss": 0.5393, "step": 1732 }, { "epoch": 0.07264495981052786, "grad_norm": 1.8705641031265259, "learning_rate": 9.95241146144342e-06, "loss": 0.559, "step": 1733 }, { "epoch": 0.07268687842553682, "grad_norm": 2.5373990535736084, "learning_rate": 9.952317978422661e-06, "loss": 0.5623, "step": 1734 }, { "epoch": 0.07272879704054579, "grad_norm": 2.016716718673706, "learning_rate": 9.952224404112767e-06, "loss": 0.5855, "step": 1735 }, { "epoch": 0.07277071565555474, "grad_norm": 1.720176339149475, "learning_rate": 9.95213073851546e-06, "loss": 0.6059, "step": 1736 }, { "epoch": 0.0728126342705637, "grad_norm": 1.766685128211975, "learning_rate": 9.952036981632468e-06, "loss": 0.6296, "step": 1737 }, { "epoch": 0.07285455288557266, "grad_norm": 1.819586157798767, "learning_rate": 9.951943133465524e-06, "loss": 0.5659, "step": 1738 }, { "epoch": 0.07289647150058162, "grad_norm": 2.1533145904541016, "learning_rate": 9.951849194016352e-06, "loss": 0.5943, "step": 1739 }, { "epoch": 0.07293839011559058, "grad_norm": 1.7909884452819824, "learning_rate": 9.951755163286687e-06, "loss": 0.5711, "step": 1740 }, { "epoch": 0.07298030873059955, "grad_norm": 1.624725341796875, "learning_rate": 9.95166104127826e-06, "loss": 0.5365, "step": 1741 }, { "epoch": 0.0730222273456085, "grad_norm": 2.02593731880188, "learning_rate": 9.951566827992807e-06, "loss": 0.6599, "step": 1742 }, { "epoch": 0.07306414596061746, "grad_norm": 1.9434278011322021, "learning_rate": 9.951472523432066e-06, "loss": 0.5915, "step": 1743 }, { "epoch": 0.07310606457562642, "grad_norm": 2.1591076850891113, "learning_rate": 9.951378127597774e-06, "loss": 0.6147, "step": 1744 }, { "epoch": 0.07314798319063538, "grad_norm": 2.3564505577087402, "learning_rate": 9.951283640491672e-06, "loss": 0.5882, "step": 1745 }, { "epoch": 0.07318990180564434, "grad_norm": 2.076824903488159, "learning_rate": 9.951189062115498e-06, "loss": 0.6247, "step": 1746 }, { "epoch": 0.0732318204206533, "grad_norm": 2.133509635925293, "learning_rate": 9.951094392471e-06, "loss": 0.6217, "step": 1747 }, { "epoch": 0.07327373903566226, "grad_norm": 2.250396966934204, "learning_rate": 9.950999631559923e-06, "loss": 0.6068, "step": 1748 }, { "epoch": 0.07331565765067122, "grad_norm": 1.861867904663086, "learning_rate": 9.950904779384011e-06, "loss": 0.5901, "step": 1749 }, { "epoch": 0.07335757626568018, "grad_norm": 2.1707303524017334, "learning_rate": 9.950809835945014e-06, "loss": 0.6351, "step": 1750 }, { "epoch": 0.07339949488068914, "grad_norm": 2.348416328430176, "learning_rate": 9.950714801244683e-06, "loss": 0.5715, "step": 1751 }, { "epoch": 0.0734414134956981, "grad_norm": 1.9227237701416016, "learning_rate": 9.950619675284767e-06, "loss": 0.6335, "step": 1752 }, { "epoch": 0.07348333211070707, "grad_norm": 1.6251131296157837, "learning_rate": 9.950524458067023e-06, "loss": 0.5371, "step": 1753 }, { "epoch": 0.07352525072571603, "grad_norm": 1.933607578277588, "learning_rate": 9.950429149593205e-06, "loss": 0.5654, "step": 1754 }, { "epoch": 0.07356716934072498, "grad_norm": 1.8134801387786865, "learning_rate": 9.950333749865067e-06, "loss": 0.5848, "step": 1755 }, { "epoch": 0.07360908795573394, "grad_norm": 1.733901023864746, "learning_rate": 9.95023825888437e-06, "loss": 0.6395, "step": 1756 }, { "epoch": 0.0736510065707429, "grad_norm": 1.9229657649993896, "learning_rate": 9.950142676652876e-06, "loss": 0.5962, "step": 1757 }, { "epoch": 0.07369292518575186, "grad_norm": 1.945077657699585, "learning_rate": 9.950047003172345e-06, "loss": 0.6546, "step": 1758 }, { "epoch": 0.07373484380076083, "grad_norm": 1.982049584388733, "learning_rate": 9.949951238444539e-06, "loss": 0.6303, "step": 1759 }, { "epoch": 0.07377676241576979, "grad_norm": 1.8475878238677979, "learning_rate": 9.949855382471227e-06, "loss": 0.6766, "step": 1760 }, { "epoch": 0.07381868103077874, "grad_norm": 1.8993335962295532, "learning_rate": 9.949759435254174e-06, "loss": 0.5732, "step": 1761 }, { "epoch": 0.0738605996457877, "grad_norm": 1.7874032258987427, "learning_rate": 9.949663396795147e-06, "loss": 0.6623, "step": 1762 }, { "epoch": 0.07390251826079666, "grad_norm": 1.93729829788208, "learning_rate": 9.94956726709592e-06, "loss": 0.675, "step": 1763 }, { "epoch": 0.07394443687580562, "grad_norm": 1.9867476224899292, "learning_rate": 9.949471046158262e-06, "loss": 0.582, "step": 1764 }, { "epoch": 0.07398635549081459, "grad_norm": 1.9457240104675293, "learning_rate": 9.949374733983946e-06, "loss": 0.6075, "step": 1765 }, { "epoch": 0.07402827410582355, "grad_norm": 1.8328204154968262, "learning_rate": 9.949278330574751e-06, "loss": 0.5702, "step": 1766 }, { "epoch": 0.0740701927208325, "grad_norm": 2.1102383136749268, "learning_rate": 9.949181835932451e-06, "loss": 0.5902, "step": 1767 }, { "epoch": 0.07411211133584146, "grad_norm": 1.9573538303375244, "learning_rate": 9.949085250058827e-06, "loss": 0.6358, "step": 1768 }, { "epoch": 0.07415402995085042, "grad_norm": 3.1473276615142822, "learning_rate": 9.948988572955659e-06, "loss": 0.6046, "step": 1769 }, { "epoch": 0.07419594856585938, "grad_norm": 1.8575078248977661, "learning_rate": 9.948891804624726e-06, "loss": 0.5475, "step": 1770 }, { "epoch": 0.07423786718086835, "grad_norm": 1.777667760848999, "learning_rate": 9.948794945067816e-06, "loss": 0.561, "step": 1771 }, { "epoch": 0.07427978579587731, "grad_norm": 1.7125828266143799, "learning_rate": 9.948697994286712e-06, "loss": 0.5328, "step": 1772 }, { "epoch": 0.07432170441088626, "grad_norm": 2.5313851833343506, "learning_rate": 9.948600952283203e-06, "loss": 0.5747, "step": 1773 }, { "epoch": 0.07436362302589522, "grad_norm": 1.8992918729782104, "learning_rate": 9.948503819059076e-06, "loss": 0.5678, "step": 1774 }, { "epoch": 0.07440554164090418, "grad_norm": 7.388280391693115, "learning_rate": 9.948406594616122e-06, "loss": 0.596, "step": 1775 }, { "epoch": 0.07444746025591314, "grad_norm": 1.9778578281402588, "learning_rate": 9.948309278956133e-06, "loss": 0.5571, "step": 1776 }, { "epoch": 0.07448937887092211, "grad_norm": 1.794789433479309, "learning_rate": 9.948211872080903e-06, "loss": 0.563, "step": 1777 }, { "epoch": 0.07453129748593107, "grad_norm": 1.975431203842163, "learning_rate": 9.948114373992228e-06, "loss": 0.5731, "step": 1778 }, { "epoch": 0.07457321610094003, "grad_norm": 1.813735008239746, "learning_rate": 9.948016784691906e-06, "loss": 0.593, "step": 1779 }, { "epoch": 0.07461513471594898, "grad_norm": 1.8546931743621826, "learning_rate": 9.947919104181736e-06, "loss": 0.6073, "step": 1780 }, { "epoch": 0.07465705333095794, "grad_norm": 2.0916292667388916, "learning_rate": 9.947821332463516e-06, "loss": 0.6192, "step": 1781 }, { "epoch": 0.0746989719459669, "grad_norm": 2.035306453704834, "learning_rate": 9.947723469539051e-06, "loss": 0.6236, "step": 1782 }, { "epoch": 0.07474089056097587, "grad_norm": 1.7765226364135742, "learning_rate": 9.947625515410142e-06, "loss": 0.5748, "step": 1783 }, { "epoch": 0.07478280917598483, "grad_norm": 1.9450947046279907, "learning_rate": 9.947527470078598e-06, "loss": 0.5928, "step": 1784 }, { "epoch": 0.07482472779099379, "grad_norm": 1.9230021238327026, "learning_rate": 9.947429333546223e-06, "loss": 0.6379, "step": 1785 }, { "epoch": 0.07486664640600274, "grad_norm": 1.7463710308074951, "learning_rate": 9.947331105814829e-06, "loss": 0.588, "step": 1786 }, { "epoch": 0.0749085650210117, "grad_norm": 1.9544371366500854, "learning_rate": 9.947232786886225e-06, "loss": 0.6781, "step": 1787 }, { "epoch": 0.07495048363602066, "grad_norm": 1.7239230871200562, "learning_rate": 9.947134376762225e-06, "loss": 0.556, "step": 1788 }, { "epoch": 0.07499240225102963, "grad_norm": 3.7905876636505127, "learning_rate": 9.947035875444642e-06, "loss": 0.5948, "step": 1789 }, { "epoch": 0.07503432086603859, "grad_norm": 1.835744023323059, "learning_rate": 9.946937282935292e-06, "loss": 0.6475, "step": 1790 }, { "epoch": 0.07507623948104755, "grad_norm": 1.8536075353622437, "learning_rate": 9.946838599235991e-06, "loss": 0.5958, "step": 1791 }, { "epoch": 0.0751181580960565, "grad_norm": 1.875954270362854, "learning_rate": 9.946739824348559e-06, "loss": 0.6301, "step": 1792 }, { "epoch": 0.07516007671106546, "grad_norm": 2.7739038467407227, "learning_rate": 9.946640958274818e-06, "loss": 0.6127, "step": 1793 }, { "epoch": 0.07520199532607443, "grad_norm": 1.8082829713821411, "learning_rate": 9.946542001016589e-06, "loss": 0.6635, "step": 1794 }, { "epoch": 0.07524391394108339, "grad_norm": 2.076004981994629, "learning_rate": 9.946442952575697e-06, "loss": 0.5875, "step": 1795 }, { "epoch": 0.07528583255609235, "grad_norm": 2.264585256576538, "learning_rate": 9.946343812953967e-06, "loss": 0.6507, "step": 1796 }, { "epoch": 0.07532775117110131, "grad_norm": 1.7309156656265259, "learning_rate": 9.946244582153227e-06, "loss": 0.6086, "step": 1797 }, { "epoch": 0.07536966978611026, "grad_norm": 1.897110104560852, "learning_rate": 9.946145260175307e-06, "loss": 0.6585, "step": 1798 }, { "epoch": 0.07541158840111922, "grad_norm": 1.9393579959869385, "learning_rate": 9.946045847022035e-06, "loss": 0.6315, "step": 1799 }, { "epoch": 0.07545350701612819, "grad_norm": 1.722354531288147, "learning_rate": 9.945946342695248e-06, "loss": 0.583, "step": 1800 }, { "epoch": 0.07549542563113715, "grad_norm": 1.8958659172058105, "learning_rate": 9.945846747196775e-06, "loss": 0.6781, "step": 1801 }, { "epoch": 0.07553734424614611, "grad_norm": 1.8795069456100464, "learning_rate": 9.945747060528457e-06, "loss": 0.5754, "step": 1802 }, { "epoch": 0.07557926286115507, "grad_norm": 1.7911572456359863, "learning_rate": 9.945647282692129e-06, "loss": 0.6362, "step": 1803 }, { "epoch": 0.07562118147616403, "grad_norm": 1.6510968208312988, "learning_rate": 9.94554741368963e-06, "loss": 0.624, "step": 1804 }, { "epoch": 0.07566310009117298, "grad_norm": 1.72357177734375, "learning_rate": 9.9454474535228e-06, "loss": 0.6147, "step": 1805 }, { "epoch": 0.07570501870618195, "grad_norm": 1.7432790994644165, "learning_rate": 9.945347402193484e-06, "loss": 0.6447, "step": 1806 }, { "epoch": 0.07574693732119091, "grad_norm": 1.7929614782333374, "learning_rate": 9.945247259703525e-06, "loss": 0.6271, "step": 1807 }, { "epoch": 0.07578885593619987, "grad_norm": 1.7473565340042114, "learning_rate": 9.945147026054772e-06, "loss": 0.5542, "step": 1808 }, { "epoch": 0.07583077455120883, "grad_norm": 1.6906472444534302, "learning_rate": 9.945046701249066e-06, "loss": 0.6092, "step": 1809 }, { "epoch": 0.0758726931662178, "grad_norm": 1.946958065032959, "learning_rate": 9.944946285288263e-06, "loss": 0.5997, "step": 1810 }, { "epoch": 0.07591461178122674, "grad_norm": 2.056574583053589, "learning_rate": 9.94484577817421e-06, "loss": 0.5433, "step": 1811 }, { "epoch": 0.0759565303962357, "grad_norm": 1.774126410484314, "learning_rate": 9.944745179908764e-06, "loss": 0.5866, "step": 1812 }, { "epoch": 0.07599844901124467, "grad_norm": 1.765708327293396, "learning_rate": 9.944644490493773e-06, "loss": 0.5275, "step": 1813 }, { "epoch": 0.07604036762625363, "grad_norm": 1.9957292079925537, "learning_rate": 9.944543709931099e-06, "loss": 0.6371, "step": 1814 }, { "epoch": 0.07608228624126259, "grad_norm": 2.6685452461242676, "learning_rate": 9.944442838222596e-06, "loss": 0.5703, "step": 1815 }, { "epoch": 0.07612420485627155, "grad_norm": 1.8712198734283447, "learning_rate": 9.944341875370126e-06, "loss": 0.5725, "step": 1816 }, { "epoch": 0.0761661234712805, "grad_norm": 1.6435632705688477, "learning_rate": 9.944240821375547e-06, "loss": 0.5796, "step": 1817 }, { "epoch": 0.07620804208628947, "grad_norm": 1.7421826124191284, "learning_rate": 9.944139676240726e-06, "loss": 0.5619, "step": 1818 }, { "epoch": 0.07624996070129843, "grad_norm": 2.0103957653045654, "learning_rate": 9.944038439967523e-06, "loss": 0.6471, "step": 1819 }, { "epoch": 0.07629187931630739, "grad_norm": 1.808538556098938, "learning_rate": 9.943937112557806e-06, "loss": 0.5378, "step": 1820 }, { "epoch": 0.07633379793131635, "grad_norm": 1.8447339534759521, "learning_rate": 9.943835694013445e-06, "loss": 0.5857, "step": 1821 }, { "epoch": 0.07637571654632531, "grad_norm": 1.8579235076904297, "learning_rate": 9.943734184336306e-06, "loss": 0.5809, "step": 1822 }, { "epoch": 0.07641763516133426, "grad_norm": 1.7397271394729614, "learning_rate": 9.943632583528263e-06, "loss": 0.6621, "step": 1823 }, { "epoch": 0.07645955377634323, "grad_norm": 1.8686870336532593, "learning_rate": 9.943530891591188e-06, "loss": 0.5464, "step": 1824 }, { "epoch": 0.07650147239135219, "grad_norm": 2.478618621826172, "learning_rate": 9.943429108526953e-06, "loss": 0.5881, "step": 1825 }, { "epoch": 0.07654339100636115, "grad_norm": 2.109788179397583, "learning_rate": 9.943327234337439e-06, "loss": 0.6189, "step": 1826 }, { "epoch": 0.07658530962137011, "grad_norm": 1.8669586181640625, "learning_rate": 9.943225269024522e-06, "loss": 0.6144, "step": 1827 }, { "epoch": 0.07662722823637907, "grad_norm": 1.6823439598083496, "learning_rate": 9.943123212590078e-06, "loss": 0.598, "step": 1828 }, { "epoch": 0.07666914685138804, "grad_norm": 1.88450026512146, "learning_rate": 9.943021065035993e-06, "loss": 0.5583, "step": 1829 }, { "epoch": 0.07671106546639699, "grad_norm": 1.8355430364608765, "learning_rate": 9.942918826364148e-06, "loss": 0.6033, "step": 1830 }, { "epoch": 0.07675298408140595, "grad_norm": 1.7839995622634888, "learning_rate": 9.942816496576428e-06, "loss": 0.5685, "step": 1831 }, { "epoch": 0.07679490269641491, "grad_norm": 1.6925464868545532, "learning_rate": 9.942714075674717e-06, "loss": 0.5922, "step": 1832 }, { "epoch": 0.07683682131142387, "grad_norm": 1.8985002040863037, "learning_rate": 9.94261156366091e-06, "loss": 0.6536, "step": 1833 }, { "epoch": 0.07687873992643283, "grad_norm": 1.692430019378662, "learning_rate": 9.942508960536887e-06, "loss": 0.6178, "step": 1834 }, { "epoch": 0.0769206585414418, "grad_norm": 1.6595826148986816, "learning_rate": 9.942406266304544e-06, "loss": 0.5408, "step": 1835 }, { "epoch": 0.07696257715645075, "grad_norm": 2.728257417678833, "learning_rate": 9.942303480965778e-06, "loss": 0.5739, "step": 1836 }, { "epoch": 0.07700449577145971, "grad_norm": 1.793933629989624, "learning_rate": 9.942200604522479e-06, "loss": 0.6016, "step": 1837 }, { "epoch": 0.07704641438646867, "grad_norm": 2.077894449234009, "learning_rate": 9.942097636976541e-06, "loss": 0.5871, "step": 1838 }, { "epoch": 0.07708833300147763, "grad_norm": 1.8580676317214966, "learning_rate": 9.941994578329867e-06, "loss": 0.5424, "step": 1839 }, { "epoch": 0.0771302516164866, "grad_norm": 2.024402141571045, "learning_rate": 9.941891428584356e-06, "loss": 0.6, "step": 1840 }, { "epoch": 0.07717217023149556, "grad_norm": 1.96087646484375, "learning_rate": 9.94178818774191e-06, "loss": 0.6883, "step": 1841 }, { "epoch": 0.0772140888465045, "grad_norm": 1.8747321367263794, "learning_rate": 9.941684855804427e-06, "loss": 0.6045, "step": 1842 }, { "epoch": 0.07725600746151347, "grad_norm": 1.6864689588546753, "learning_rate": 9.941581432773816e-06, "loss": 0.5573, "step": 1843 }, { "epoch": 0.07729792607652243, "grad_norm": 1.6578178405761719, "learning_rate": 9.941477918651985e-06, "loss": 0.5293, "step": 1844 }, { "epoch": 0.07733984469153139, "grad_norm": 2.1586501598358154, "learning_rate": 9.941374313440839e-06, "loss": 0.668, "step": 1845 }, { "epoch": 0.07738176330654035, "grad_norm": 1.7937825918197632, "learning_rate": 9.941270617142287e-06, "loss": 0.568, "step": 1846 }, { "epoch": 0.07742368192154932, "grad_norm": 1.7753318548202515, "learning_rate": 9.941166829758245e-06, "loss": 0.6023, "step": 1847 }, { "epoch": 0.07746560053655827, "grad_norm": 1.7309058904647827, "learning_rate": 9.94106295129062e-06, "loss": 0.5872, "step": 1848 }, { "epoch": 0.07750751915156723, "grad_norm": 1.8145676851272583, "learning_rate": 9.940958981741334e-06, "loss": 0.533, "step": 1849 }, { "epoch": 0.07754943776657619, "grad_norm": 1.9668192863464355, "learning_rate": 9.940854921112296e-06, "loss": 0.6697, "step": 1850 }, { "epoch": 0.07759135638158515, "grad_norm": 3.2555439472198486, "learning_rate": 9.940750769405432e-06, "loss": 0.6054, "step": 1851 }, { "epoch": 0.07763327499659411, "grad_norm": 1.6017757654190063, "learning_rate": 9.940646526622657e-06, "loss": 0.5856, "step": 1852 }, { "epoch": 0.07767519361160308, "grad_norm": 1.8470351696014404, "learning_rate": 9.940542192765892e-06, "loss": 0.6372, "step": 1853 }, { "epoch": 0.07771711222661204, "grad_norm": 4.632908821105957, "learning_rate": 9.940437767837063e-06, "loss": 0.6258, "step": 1854 }, { "epoch": 0.07775903084162099, "grad_norm": 1.6354858875274658, "learning_rate": 9.940333251838092e-06, "loss": 0.5509, "step": 1855 }, { "epoch": 0.07780094945662995, "grad_norm": 1.918757677078247, "learning_rate": 9.94022864477091e-06, "loss": 0.6471, "step": 1856 }, { "epoch": 0.07784286807163891, "grad_norm": 2.084415912628174, "learning_rate": 9.940123946637442e-06, "loss": 0.6472, "step": 1857 }, { "epoch": 0.07788478668664787, "grad_norm": 2.0165774822235107, "learning_rate": 9.940019157439617e-06, "loss": 0.6085, "step": 1858 }, { "epoch": 0.07792670530165684, "grad_norm": 1.8824150562286377, "learning_rate": 9.939914277179368e-06, "loss": 0.5394, "step": 1859 }, { "epoch": 0.0779686239166658, "grad_norm": 1.8135831356048584, "learning_rate": 9.939809305858629e-06, "loss": 0.5906, "step": 1860 }, { "epoch": 0.07801054253167475, "grad_norm": 1.7174360752105713, "learning_rate": 9.939704243479336e-06, "loss": 0.586, "step": 1861 }, { "epoch": 0.07805246114668371, "grad_norm": 2.1712255477905273, "learning_rate": 9.939599090043424e-06, "loss": 0.5977, "step": 1862 }, { "epoch": 0.07809437976169267, "grad_norm": 1.7998179197311401, "learning_rate": 9.93949384555283e-06, "loss": 0.6061, "step": 1863 }, { "epoch": 0.07813629837670164, "grad_norm": 2.1215567588806152, "learning_rate": 9.939388510009495e-06, "loss": 0.6697, "step": 1864 }, { "epoch": 0.0781782169917106, "grad_norm": 1.7240105867385864, "learning_rate": 9.939283083415364e-06, "loss": 0.5951, "step": 1865 }, { "epoch": 0.07822013560671956, "grad_norm": 2.0520570278167725, "learning_rate": 9.939177565772376e-06, "loss": 0.581, "step": 1866 }, { "epoch": 0.07826205422172851, "grad_norm": 1.8491990566253662, "learning_rate": 9.939071957082478e-06, "loss": 0.6098, "step": 1867 }, { "epoch": 0.07830397283673747, "grad_norm": 1.8560086488723755, "learning_rate": 9.938966257347618e-06, "loss": 0.6048, "step": 1868 }, { "epoch": 0.07834589145174643, "grad_norm": 2.0666096210479736, "learning_rate": 9.93886046656974e-06, "loss": 0.6452, "step": 1869 }, { "epoch": 0.0783878100667554, "grad_norm": 1.8893965482711792, "learning_rate": 9.9387545847508e-06, "loss": 0.5706, "step": 1870 }, { "epoch": 0.07842972868176436, "grad_norm": 2.1858532428741455, "learning_rate": 9.938648611892745e-06, "loss": 0.6355, "step": 1871 }, { "epoch": 0.07847164729677332, "grad_norm": 1.9620789289474487, "learning_rate": 9.938542547997531e-06, "loss": 0.5969, "step": 1872 }, { "epoch": 0.07851356591178227, "grad_norm": 2.1478466987609863, "learning_rate": 9.938436393067113e-06, "loss": 0.6425, "step": 1873 }, { "epoch": 0.07855548452679123, "grad_norm": 2.551312208175659, "learning_rate": 9.938330147103446e-06, "loss": 0.6314, "step": 1874 }, { "epoch": 0.07859740314180019, "grad_norm": 1.977199912071228, "learning_rate": 9.938223810108491e-06, "loss": 0.6824, "step": 1875 }, { "epoch": 0.07863932175680916, "grad_norm": 1.7752350568771362, "learning_rate": 9.938117382084206e-06, "loss": 0.6144, "step": 1876 }, { "epoch": 0.07868124037181812, "grad_norm": 1.8303889036178589, "learning_rate": 9.938010863032554e-06, "loss": 0.5765, "step": 1877 }, { "epoch": 0.07872315898682708, "grad_norm": 2.3868703842163086, "learning_rate": 9.9379042529555e-06, "loss": 0.6028, "step": 1878 }, { "epoch": 0.07876507760183604, "grad_norm": 1.9276460409164429, "learning_rate": 9.937797551855004e-06, "loss": 0.6187, "step": 1879 }, { "epoch": 0.07880699621684499, "grad_norm": 1.75641930103302, "learning_rate": 9.937690759733038e-06, "loss": 0.5459, "step": 1880 }, { "epoch": 0.07884891483185395, "grad_norm": 1.8075591325759888, "learning_rate": 9.93758387659157e-06, "loss": 0.5881, "step": 1881 }, { "epoch": 0.07889083344686292, "grad_norm": 1.9989651441574097, "learning_rate": 9.93747690243257e-06, "loss": 0.6641, "step": 1882 }, { "epoch": 0.07893275206187188, "grad_norm": 2.424989700317383, "learning_rate": 9.937369837258007e-06, "loss": 0.6142, "step": 1883 }, { "epoch": 0.07897467067688084, "grad_norm": 1.9266602993011475, "learning_rate": 9.937262681069858e-06, "loss": 0.6211, "step": 1884 }, { "epoch": 0.0790165892918898, "grad_norm": 1.8311864137649536, "learning_rate": 9.937155433870097e-06, "loss": 0.6305, "step": 1885 }, { "epoch": 0.07905850790689875, "grad_norm": 2.07210636138916, "learning_rate": 9.937048095660699e-06, "loss": 0.5564, "step": 1886 }, { "epoch": 0.07910042652190771, "grad_norm": 1.9795689582824707, "learning_rate": 9.936940666443647e-06, "loss": 0.5982, "step": 1887 }, { "epoch": 0.07914234513691668, "grad_norm": 2.4389004707336426, "learning_rate": 9.936833146220919e-06, "loss": 0.6207, "step": 1888 }, { "epoch": 0.07918426375192564, "grad_norm": 1.975395679473877, "learning_rate": 9.936725534994496e-06, "loss": 0.5599, "step": 1889 }, { "epoch": 0.0792261823669346, "grad_norm": 2.895132064819336, "learning_rate": 9.936617832766362e-06, "loss": 0.552, "step": 1890 }, { "epoch": 0.07926810098194356, "grad_norm": 1.926055669784546, "learning_rate": 9.936510039538505e-06, "loss": 0.6221, "step": 1891 }, { "epoch": 0.07931001959695251, "grad_norm": 1.9270720481872559, "learning_rate": 9.936402155312907e-06, "loss": 0.5728, "step": 1892 }, { "epoch": 0.07935193821196147, "grad_norm": 1.9214171171188354, "learning_rate": 9.936294180091561e-06, "loss": 0.6331, "step": 1893 }, { "epoch": 0.07939385682697044, "grad_norm": 1.8932887315750122, "learning_rate": 9.936186113876458e-06, "loss": 0.5666, "step": 1894 }, { "epoch": 0.0794357754419794, "grad_norm": 1.9134489297866821, "learning_rate": 9.936077956669585e-06, "loss": 0.5849, "step": 1895 }, { "epoch": 0.07947769405698836, "grad_norm": 1.9582552909851074, "learning_rate": 9.935969708472941e-06, "loss": 0.5837, "step": 1896 }, { "epoch": 0.07951961267199732, "grad_norm": 2.1954877376556396, "learning_rate": 9.935861369288517e-06, "loss": 0.6588, "step": 1897 }, { "epoch": 0.07956153128700627, "grad_norm": 2.039898633956909, "learning_rate": 9.935752939118315e-06, "loss": 0.5857, "step": 1898 }, { "epoch": 0.07960344990201523, "grad_norm": 2.207066297531128, "learning_rate": 9.935644417964329e-06, "loss": 0.6064, "step": 1899 }, { "epoch": 0.0796453685170242, "grad_norm": 2.140554428100586, "learning_rate": 9.935535805828563e-06, "loss": 0.5664, "step": 1900 }, { "epoch": 0.07968728713203316, "grad_norm": 2.305006265640259, "learning_rate": 9.935427102713016e-06, "loss": 0.5858, "step": 1901 }, { "epoch": 0.07972920574704212, "grad_norm": 2.691394567489624, "learning_rate": 9.935318308619694e-06, "loss": 0.5705, "step": 1902 }, { "epoch": 0.07977112436205108, "grad_norm": 2.1065032482147217, "learning_rate": 9.935209423550603e-06, "loss": 0.5961, "step": 1903 }, { "epoch": 0.07981304297706004, "grad_norm": 1.834181785583496, "learning_rate": 9.935100447507746e-06, "loss": 0.6112, "step": 1904 }, { "epoch": 0.079854961592069, "grad_norm": 1.7658847570419312, "learning_rate": 9.934991380493137e-06, "loss": 0.6474, "step": 1905 }, { "epoch": 0.07989688020707796, "grad_norm": 1.9287869930267334, "learning_rate": 9.934882222508785e-06, "loss": 0.5698, "step": 1906 }, { "epoch": 0.07993879882208692, "grad_norm": 1.9776356220245361, "learning_rate": 9.9347729735567e-06, "loss": 0.6297, "step": 1907 }, { "epoch": 0.07998071743709588, "grad_norm": 2.120420455932617, "learning_rate": 9.934663633638899e-06, "loss": 0.5927, "step": 1908 }, { "epoch": 0.08002263605210484, "grad_norm": 2.3398642539978027, "learning_rate": 9.934554202757394e-06, "loss": 0.5579, "step": 1909 }, { "epoch": 0.0800645546671138, "grad_norm": 1.7240928411483765, "learning_rate": 9.934444680914206e-06, "loss": 0.559, "step": 1910 }, { "epoch": 0.08010647328212275, "grad_norm": 2.7815704345703125, "learning_rate": 9.934335068111349e-06, "loss": 0.6338, "step": 1911 }, { "epoch": 0.08014839189713172, "grad_norm": 2.138206720352173, "learning_rate": 9.934225364350848e-06, "loss": 0.5704, "step": 1912 }, { "epoch": 0.08019031051214068, "grad_norm": 2.1777708530426025, "learning_rate": 9.934115569634725e-06, "loss": 0.62, "step": 1913 }, { "epoch": 0.08023222912714964, "grad_norm": 2.237821578979492, "learning_rate": 9.934005683965002e-06, "loss": 0.5822, "step": 1914 }, { "epoch": 0.0802741477421586, "grad_norm": 1.8078861236572266, "learning_rate": 9.933895707343705e-06, "loss": 0.5951, "step": 1915 }, { "epoch": 0.08031606635716756, "grad_norm": 1.9622104167938232, "learning_rate": 9.933785639772862e-06, "loss": 0.6371, "step": 1916 }, { "epoch": 0.08035798497217651, "grad_norm": 2.2698323726654053, "learning_rate": 9.933675481254501e-06, "loss": 0.6, "step": 1917 }, { "epoch": 0.08039990358718548, "grad_norm": 1.6606658697128296, "learning_rate": 9.933565231790654e-06, "loss": 0.5855, "step": 1918 }, { "epoch": 0.08044182220219444, "grad_norm": 1.742959976196289, "learning_rate": 9.93345489138335e-06, "loss": 0.5592, "step": 1919 }, { "epoch": 0.0804837408172034, "grad_norm": 2.2956509590148926, "learning_rate": 9.933344460034628e-06, "loss": 0.5414, "step": 1920 }, { "epoch": 0.08052565943221236, "grad_norm": 1.9539458751678467, "learning_rate": 9.93323393774652e-06, "loss": 0.5905, "step": 1921 }, { "epoch": 0.08056757804722132, "grad_norm": 1.8262839317321777, "learning_rate": 9.933123324521063e-06, "loss": 0.6634, "step": 1922 }, { "epoch": 0.08060949666223027, "grad_norm": 2.765397071838379, "learning_rate": 9.9330126203603e-06, "loss": 0.5665, "step": 1923 }, { "epoch": 0.08065141527723924, "grad_norm": 2.1857831478118896, "learning_rate": 9.932901825266268e-06, "loss": 0.6167, "step": 1924 }, { "epoch": 0.0806933338922482, "grad_norm": 2.085982322692871, "learning_rate": 9.93279093924101e-06, "loss": 0.5664, "step": 1925 }, { "epoch": 0.08073525250725716, "grad_norm": 1.9315440654754639, "learning_rate": 9.93267996228657e-06, "loss": 0.5976, "step": 1926 }, { "epoch": 0.08077717112226612, "grad_norm": 2.117971181869507, "learning_rate": 9.932568894404994e-06, "loss": 0.642, "step": 1927 }, { "epoch": 0.08081908973727508, "grad_norm": 1.917673110961914, "learning_rate": 9.93245773559833e-06, "loss": 0.5743, "step": 1928 }, { "epoch": 0.08086100835228405, "grad_norm": 2.029207229614258, "learning_rate": 9.932346485868628e-06, "loss": 0.6526, "step": 1929 }, { "epoch": 0.080902926967293, "grad_norm": 1.8571795225143433, "learning_rate": 9.932235145217936e-06, "loss": 0.5881, "step": 1930 }, { "epoch": 0.08094484558230196, "grad_norm": 2.0591635704040527, "learning_rate": 9.932123713648308e-06, "loss": 0.5693, "step": 1931 }, { "epoch": 0.08098676419731092, "grad_norm": 2.496473550796509, "learning_rate": 9.932012191161797e-06, "loss": 0.6112, "step": 1932 }, { "epoch": 0.08102868281231988, "grad_norm": 2.2938413619995117, "learning_rate": 9.93190057776046e-06, "loss": 0.5837, "step": 1933 }, { "epoch": 0.08107060142732885, "grad_norm": 2.3461339473724365, "learning_rate": 9.931788873446356e-06, "loss": 0.6318, "step": 1934 }, { "epoch": 0.08111252004233781, "grad_norm": 1.9540667533874512, "learning_rate": 9.931677078221538e-06, "loss": 0.6228, "step": 1935 }, { "epoch": 0.08115443865734676, "grad_norm": 1.92304265499115, "learning_rate": 9.931565192088074e-06, "loss": 0.6628, "step": 1936 }, { "epoch": 0.08119635727235572, "grad_norm": 2.4215335845947266, "learning_rate": 9.931453215048022e-06, "loss": 0.6024, "step": 1937 }, { "epoch": 0.08123827588736468, "grad_norm": 2.1102101802825928, "learning_rate": 9.931341147103448e-06, "loss": 0.5415, "step": 1938 }, { "epoch": 0.08128019450237364, "grad_norm": 1.887804627418518, "learning_rate": 9.931228988256418e-06, "loss": 0.6386, "step": 1939 }, { "epoch": 0.0813221131173826, "grad_norm": 1.6877409219741821, "learning_rate": 9.931116738508998e-06, "loss": 0.5894, "step": 1940 }, { "epoch": 0.08136403173239157, "grad_norm": 1.7801166772842407, "learning_rate": 9.931004397863257e-06, "loss": 0.5469, "step": 1941 }, { "epoch": 0.08140595034740052, "grad_norm": 1.8565104007720947, "learning_rate": 9.930891966321269e-06, "loss": 0.55, "step": 1942 }, { "epoch": 0.08144786896240948, "grad_norm": 1.8728152513504028, "learning_rate": 9.930779443885103e-06, "loss": 0.5885, "step": 1943 }, { "epoch": 0.08148978757741844, "grad_norm": 1.9645239114761353, "learning_rate": 9.930666830556835e-06, "loss": 0.6305, "step": 1944 }, { "epoch": 0.0815317061924274, "grad_norm": 2.032949924468994, "learning_rate": 9.93055412633854e-06, "loss": 0.5767, "step": 1945 }, { "epoch": 0.08157362480743637, "grad_norm": 1.776950478553772, "learning_rate": 9.930441331232296e-06, "loss": 0.5839, "step": 1946 }, { "epoch": 0.08161554342244533, "grad_norm": 1.946120262145996, "learning_rate": 9.930328445240182e-06, "loss": 0.63, "step": 1947 }, { "epoch": 0.08165746203745428, "grad_norm": 3.180065155029297, "learning_rate": 9.930215468364281e-06, "loss": 0.6222, "step": 1948 }, { "epoch": 0.08169938065246324, "grad_norm": 2.200082540512085, "learning_rate": 9.930102400606672e-06, "loss": 0.5988, "step": 1949 }, { "epoch": 0.0817412992674722, "grad_norm": 1.785510778427124, "learning_rate": 9.92998924196944e-06, "loss": 0.6061, "step": 1950 }, { "epoch": 0.08178321788248116, "grad_norm": 1.6497710943222046, "learning_rate": 9.929875992454673e-06, "loss": 0.5778, "step": 1951 }, { "epoch": 0.08182513649749013, "grad_norm": 2.392869234085083, "learning_rate": 9.929762652064459e-06, "loss": 0.6267, "step": 1952 }, { "epoch": 0.08186705511249909, "grad_norm": 2.0420472621917725, "learning_rate": 9.929649220800884e-06, "loss": 0.6275, "step": 1953 }, { "epoch": 0.08190897372750805, "grad_norm": 1.7320021390914917, "learning_rate": 9.929535698666039e-06, "loss": 0.6079, "step": 1954 }, { "epoch": 0.081950892342517, "grad_norm": 1.811183214187622, "learning_rate": 9.929422085662021e-06, "loss": 0.6028, "step": 1955 }, { "epoch": 0.08199281095752596, "grad_norm": 1.872902512550354, "learning_rate": 9.929308381790921e-06, "loss": 0.64, "step": 1956 }, { "epoch": 0.08203472957253492, "grad_norm": 2.876176357269287, "learning_rate": 9.929194587054836e-06, "loss": 0.6464, "step": 1957 }, { "epoch": 0.08207664818754389, "grad_norm": 2.2728006839752197, "learning_rate": 9.929080701455861e-06, "loss": 0.5775, "step": 1958 }, { "epoch": 0.08211856680255285, "grad_norm": 1.9366264343261719, "learning_rate": 9.928966724996099e-06, "loss": 0.6351, "step": 1959 }, { "epoch": 0.08216048541756181, "grad_norm": 1.932651162147522, "learning_rate": 9.92885265767765e-06, "loss": 0.6374, "step": 1960 }, { "epoch": 0.08220240403257076, "grad_norm": 2.316098928451538, "learning_rate": 9.928738499502614e-06, "loss": 0.6458, "step": 1961 }, { "epoch": 0.08224432264757972, "grad_norm": 1.8509228229522705, "learning_rate": 9.9286242504731e-06, "loss": 0.5473, "step": 1962 }, { "epoch": 0.08228624126258868, "grad_norm": 2.0369985103607178, "learning_rate": 9.92850991059121e-06, "loss": 0.6182, "step": 1963 }, { "epoch": 0.08232815987759765, "grad_norm": 1.8642784357070923, "learning_rate": 9.928395479859055e-06, "loss": 0.5952, "step": 1964 }, { "epoch": 0.08237007849260661, "grad_norm": 1.7405726909637451, "learning_rate": 9.92828095827874e-06, "loss": 0.6135, "step": 1965 }, { "epoch": 0.08241199710761557, "grad_norm": 2.814771890640259, "learning_rate": 9.928166345852381e-06, "loss": 0.5728, "step": 1966 }, { "epoch": 0.08245391572262452, "grad_norm": 1.963122844696045, "learning_rate": 9.928051642582089e-06, "loss": 0.5216, "step": 1967 }, { "epoch": 0.08249583433763348, "grad_norm": 1.8509960174560547, "learning_rate": 9.927936848469975e-06, "loss": 0.5887, "step": 1968 }, { "epoch": 0.08253775295264244, "grad_norm": 1.6779588460922241, "learning_rate": 9.927821963518159e-06, "loss": 0.5743, "step": 1969 }, { "epoch": 0.0825796715676514, "grad_norm": 2.0860252380371094, "learning_rate": 9.927706987728758e-06, "loss": 0.5808, "step": 1970 }, { "epoch": 0.08262159018266037, "grad_norm": 1.7418386936187744, "learning_rate": 9.927591921103891e-06, "loss": 0.5443, "step": 1971 }, { "epoch": 0.08266350879766933, "grad_norm": 1.930419683456421, "learning_rate": 9.92747676364568e-06, "loss": 0.5549, "step": 1972 }, { "epoch": 0.08270542741267828, "grad_norm": 1.881689429283142, "learning_rate": 9.927361515356245e-06, "loss": 0.6429, "step": 1973 }, { "epoch": 0.08274734602768724, "grad_norm": 1.8039093017578125, "learning_rate": 9.927246176237713e-06, "loss": 0.6393, "step": 1974 }, { "epoch": 0.0827892646426962, "grad_norm": 2.205288887023926, "learning_rate": 9.92713074629221e-06, "loss": 0.6256, "step": 1975 }, { "epoch": 0.08283118325770517, "grad_norm": 1.7194586992263794, "learning_rate": 9.927015225521863e-06, "loss": 0.6191, "step": 1976 }, { "epoch": 0.08287310187271413, "grad_norm": 2.0435619354248047, "learning_rate": 9.926899613928803e-06, "loss": 0.5847, "step": 1977 }, { "epoch": 0.08291502048772309, "grad_norm": 1.8983852863311768, "learning_rate": 9.926783911515156e-06, "loss": 0.5892, "step": 1978 }, { "epoch": 0.08295693910273205, "grad_norm": 1.762272834777832, "learning_rate": 9.926668118283062e-06, "loss": 0.5709, "step": 1979 }, { "epoch": 0.082998857717741, "grad_norm": 1.921519160270691, "learning_rate": 9.926552234234652e-06, "loss": 0.6227, "step": 1980 }, { "epoch": 0.08304077633274996, "grad_norm": 1.588403344154358, "learning_rate": 9.926436259372062e-06, "loss": 0.5304, "step": 1981 }, { "epoch": 0.08308269494775893, "grad_norm": 2.1088266372680664, "learning_rate": 9.92632019369743e-06, "loss": 0.5891, "step": 1982 }, { "epoch": 0.08312461356276789, "grad_norm": 1.8924651145935059, "learning_rate": 9.926204037212894e-06, "loss": 0.5716, "step": 1983 }, { "epoch": 0.08316653217777685, "grad_norm": 1.7565178871154785, "learning_rate": 9.9260877899206e-06, "loss": 0.5557, "step": 1984 }, { "epoch": 0.08320845079278581, "grad_norm": 1.8775198459625244, "learning_rate": 9.925971451822685e-06, "loss": 0.6249, "step": 1985 }, { "epoch": 0.08325036940779476, "grad_norm": 2.109475612640381, "learning_rate": 9.925855022921297e-06, "loss": 0.6572, "step": 1986 }, { "epoch": 0.08329228802280372, "grad_norm": 1.9822545051574707, "learning_rate": 9.925738503218582e-06, "loss": 0.581, "step": 1987 }, { "epoch": 0.08333420663781269, "grad_norm": 2.471346616744995, "learning_rate": 9.925621892716688e-06, "loss": 0.5926, "step": 1988 }, { "epoch": 0.08337612525282165, "grad_norm": 1.750967025756836, "learning_rate": 9.925505191417761e-06, "loss": 0.553, "step": 1989 }, { "epoch": 0.08341804386783061, "grad_norm": 2.518498659133911, "learning_rate": 9.92538839932396e-06, "loss": 0.5367, "step": 1990 }, { "epoch": 0.08345996248283957, "grad_norm": 1.9233644008636475, "learning_rate": 9.925271516437428e-06, "loss": 0.5958, "step": 1991 }, { "epoch": 0.08350188109784852, "grad_norm": 2.6409921646118164, "learning_rate": 9.925154542760327e-06, "loss": 0.5801, "step": 1992 }, { "epoch": 0.08354379971285748, "grad_norm": 1.6054893732070923, "learning_rate": 9.92503747829481e-06, "loss": 0.5395, "step": 1993 }, { "epoch": 0.08358571832786645, "grad_norm": 1.8384233713150024, "learning_rate": 9.924920323043037e-06, "loss": 0.5849, "step": 1994 }, { "epoch": 0.08362763694287541, "grad_norm": 1.9611098766326904, "learning_rate": 9.924803077007165e-06, "loss": 0.5185, "step": 1995 }, { "epoch": 0.08366955555788437, "grad_norm": 1.5887959003448486, "learning_rate": 9.924685740189357e-06, "loss": 0.5546, "step": 1996 }, { "epoch": 0.08371147417289333, "grad_norm": 2.406843423843384, "learning_rate": 9.924568312591775e-06, "loss": 0.5915, "step": 1997 }, { "epoch": 0.08375339278790228, "grad_norm": 5.255414962768555, "learning_rate": 9.924450794216585e-06, "loss": 0.566, "step": 1998 }, { "epoch": 0.08379531140291124, "grad_norm": 1.9938455820083618, "learning_rate": 9.924333185065953e-06, "loss": 0.5254, "step": 1999 }, { "epoch": 0.0838372300179202, "grad_norm": 2.1807122230529785, "learning_rate": 9.924215485142046e-06, "loss": 0.5794, "step": 2000 }, { "epoch": 0.08387914863292917, "grad_norm": 2.944096326828003, "learning_rate": 9.924097694447032e-06, "loss": 0.6259, "step": 2001 }, { "epoch": 0.08392106724793813, "grad_norm": 1.7556222677230835, "learning_rate": 9.923979812983086e-06, "loss": 0.6573, "step": 2002 }, { "epoch": 0.08396298586294709, "grad_norm": 1.8467590808868408, "learning_rate": 9.923861840752381e-06, "loss": 0.6725, "step": 2003 }, { "epoch": 0.08400490447795605, "grad_norm": 2.0005128383636475, "learning_rate": 9.92374377775709e-06, "loss": 0.6149, "step": 2004 }, { "epoch": 0.084046823092965, "grad_norm": 1.4826128482818604, "learning_rate": 9.923625623999387e-06, "loss": 0.5121, "step": 2005 }, { "epoch": 0.08408874170797397, "grad_norm": 1.9787875413894653, "learning_rate": 9.923507379481452e-06, "loss": 0.549, "step": 2006 }, { "epoch": 0.08413066032298293, "grad_norm": 1.9140143394470215, "learning_rate": 9.923389044205468e-06, "loss": 0.6179, "step": 2007 }, { "epoch": 0.08417257893799189, "grad_norm": 1.757385015487671, "learning_rate": 9.92327061817361e-06, "loss": 0.614, "step": 2008 }, { "epoch": 0.08421449755300085, "grad_norm": 1.9483387470245361, "learning_rate": 9.923152101388066e-06, "loss": 0.6012, "step": 2009 }, { "epoch": 0.08425641616800982, "grad_norm": 1.7297241687774658, "learning_rate": 9.923033493851018e-06, "loss": 0.6122, "step": 2010 }, { "epoch": 0.08429833478301876, "grad_norm": 1.7738368511199951, "learning_rate": 9.922914795564655e-06, "loss": 0.5959, "step": 2011 }, { "epoch": 0.08434025339802773, "grad_norm": 2.466991901397705, "learning_rate": 9.922796006531163e-06, "loss": 0.6121, "step": 2012 }, { "epoch": 0.08438217201303669, "grad_norm": 1.9744541645050049, "learning_rate": 9.92267712675273e-06, "loss": 0.5745, "step": 2013 }, { "epoch": 0.08442409062804565, "grad_norm": 1.8005977869033813, "learning_rate": 9.922558156231553e-06, "loss": 0.5769, "step": 2014 }, { "epoch": 0.08446600924305461, "grad_norm": 1.811333179473877, "learning_rate": 9.92243909496982e-06, "loss": 0.6254, "step": 2015 }, { "epoch": 0.08450792785806358, "grad_norm": 2.0541229248046875, "learning_rate": 9.922319942969729e-06, "loss": 0.6293, "step": 2016 }, { "epoch": 0.08454984647307252, "grad_norm": 2.0050673484802246, "learning_rate": 9.922200700233474e-06, "loss": 0.6858, "step": 2017 }, { "epoch": 0.08459176508808149, "grad_norm": 1.5404752492904663, "learning_rate": 9.922081366763253e-06, "loss": 0.5277, "step": 2018 }, { "epoch": 0.08463368370309045, "grad_norm": 2.1561837196350098, "learning_rate": 9.921961942561265e-06, "loss": 0.6122, "step": 2019 }, { "epoch": 0.08467560231809941, "grad_norm": 1.8281464576721191, "learning_rate": 9.921842427629715e-06, "loss": 0.6234, "step": 2020 }, { "epoch": 0.08471752093310837, "grad_norm": 1.8158693313598633, "learning_rate": 9.921722821970805e-06, "loss": 0.5377, "step": 2021 }, { "epoch": 0.08475943954811734, "grad_norm": 1.696370005607605, "learning_rate": 9.921603125586736e-06, "loss": 0.6049, "step": 2022 }, { "epoch": 0.08480135816312628, "grad_norm": 1.7229535579681396, "learning_rate": 9.921483338479719e-06, "loss": 0.6061, "step": 2023 }, { "epoch": 0.08484327677813525, "grad_norm": 2.3997809886932373, "learning_rate": 9.92136346065196e-06, "loss": 0.5975, "step": 2024 }, { "epoch": 0.08488519539314421, "grad_norm": 2.00376296043396, "learning_rate": 9.92124349210567e-06, "loss": 0.5601, "step": 2025 }, { "epoch": 0.08492711400815317, "grad_norm": 1.7608205080032349, "learning_rate": 9.921123432843059e-06, "loss": 0.5468, "step": 2026 }, { "epoch": 0.08496903262316213, "grad_norm": 1.8489975929260254, "learning_rate": 9.921003282866341e-06, "loss": 0.6278, "step": 2027 }, { "epoch": 0.0850109512381711, "grad_norm": 1.7699835300445557, "learning_rate": 9.920883042177732e-06, "loss": 0.584, "step": 2028 }, { "epoch": 0.08505286985318006, "grad_norm": 1.8479641675949097, "learning_rate": 9.920762710779444e-06, "loss": 0.5649, "step": 2029 }, { "epoch": 0.085094788468189, "grad_norm": 2.3242616653442383, "learning_rate": 9.920642288673702e-06, "loss": 0.6087, "step": 2030 }, { "epoch": 0.08513670708319797, "grad_norm": 1.9353567361831665, "learning_rate": 9.920521775862717e-06, "loss": 0.5788, "step": 2031 }, { "epoch": 0.08517862569820693, "grad_norm": 1.900292992591858, "learning_rate": 9.92040117234872e-06, "loss": 0.6179, "step": 2032 }, { "epoch": 0.08522054431321589, "grad_norm": 1.7495322227478027, "learning_rate": 9.920280478133926e-06, "loss": 0.5539, "step": 2033 }, { "epoch": 0.08526246292822486, "grad_norm": 2.2173075675964355, "learning_rate": 9.920159693220566e-06, "loss": 0.5968, "step": 2034 }, { "epoch": 0.08530438154323382, "grad_norm": 1.9421931505203247, "learning_rate": 9.920038817610864e-06, "loss": 0.6169, "step": 2035 }, { "epoch": 0.08534630015824277, "grad_norm": 1.7036092281341553, "learning_rate": 9.919917851307046e-06, "loss": 0.6142, "step": 2036 }, { "epoch": 0.08538821877325173, "grad_norm": 3.0450408458709717, "learning_rate": 9.919796794311347e-06, "loss": 0.6229, "step": 2037 }, { "epoch": 0.08543013738826069, "grad_norm": 1.8872671127319336, "learning_rate": 9.919675646625993e-06, "loss": 0.5627, "step": 2038 }, { "epoch": 0.08547205600326965, "grad_norm": 1.9048058986663818, "learning_rate": 9.919554408253221e-06, "loss": 0.6482, "step": 2039 }, { "epoch": 0.08551397461827862, "grad_norm": 1.8489598035812378, "learning_rate": 9.919433079195264e-06, "loss": 0.6431, "step": 2040 }, { "epoch": 0.08555589323328758, "grad_norm": 1.7371931076049805, "learning_rate": 9.919311659454359e-06, "loss": 0.5775, "step": 2041 }, { "epoch": 0.08559781184829653, "grad_norm": 1.7614113092422485, "learning_rate": 9.919190149032744e-06, "loss": 0.5803, "step": 2042 }, { "epoch": 0.08563973046330549, "grad_norm": 1.8100154399871826, "learning_rate": 9.91906854793266e-06, "loss": 0.6053, "step": 2043 }, { "epoch": 0.08568164907831445, "grad_norm": 1.9089694023132324, "learning_rate": 9.918946856156348e-06, "loss": 0.6457, "step": 2044 }, { "epoch": 0.08572356769332341, "grad_norm": 1.80482816696167, "learning_rate": 9.918825073706052e-06, "loss": 0.6038, "step": 2045 }, { "epoch": 0.08576548630833238, "grad_norm": 1.8515828847885132, "learning_rate": 9.918703200584012e-06, "loss": 0.6143, "step": 2046 }, { "epoch": 0.08580740492334134, "grad_norm": 1.862960934638977, "learning_rate": 9.918581236792482e-06, "loss": 0.6185, "step": 2047 }, { "epoch": 0.08584932353835029, "grad_norm": 1.9533194303512573, "learning_rate": 9.918459182333704e-06, "loss": 0.6099, "step": 2048 }, { "epoch": 0.08589124215335925, "grad_norm": 2.134265422821045, "learning_rate": 9.918337037209933e-06, "loss": 0.6609, "step": 2049 }, { "epoch": 0.08593316076836821, "grad_norm": 1.862131953239441, "learning_rate": 9.918214801423416e-06, "loss": 0.5633, "step": 2050 }, { "epoch": 0.08597507938337717, "grad_norm": 1.7068853378295898, "learning_rate": 9.91809247497641e-06, "loss": 0.6034, "step": 2051 }, { "epoch": 0.08601699799838614, "grad_norm": 2.2204341888427734, "learning_rate": 9.917970057871166e-06, "loss": 0.5624, "step": 2052 }, { "epoch": 0.0860589166133951, "grad_norm": 1.8367348909378052, "learning_rate": 9.917847550109946e-06, "loss": 0.5552, "step": 2053 }, { "epoch": 0.08610083522840406, "grad_norm": 2.823199510574341, "learning_rate": 9.917724951695004e-06, "loss": 0.5695, "step": 2054 }, { "epoch": 0.08614275384341301, "grad_norm": 3.737015724182129, "learning_rate": 9.917602262628601e-06, "loss": 0.5461, "step": 2055 }, { "epoch": 0.08618467245842197, "grad_norm": 2.015254259109497, "learning_rate": 9.917479482912998e-06, "loss": 0.564, "step": 2056 }, { "epoch": 0.08622659107343093, "grad_norm": 2.7043981552124023, "learning_rate": 9.917356612550459e-06, "loss": 0.5997, "step": 2057 }, { "epoch": 0.0862685096884399, "grad_norm": 1.9229847192764282, "learning_rate": 9.91723365154325e-06, "loss": 0.6718, "step": 2058 }, { "epoch": 0.08631042830344886, "grad_norm": 1.6708338260650635, "learning_rate": 9.917110599893635e-06, "loss": 0.6116, "step": 2059 }, { "epoch": 0.08635234691845782, "grad_norm": 1.9751836061477661, "learning_rate": 9.916987457603884e-06, "loss": 0.5896, "step": 2060 }, { "epoch": 0.08639426553346677, "grad_norm": 3.8481686115264893, "learning_rate": 9.916864224676267e-06, "loss": 0.5898, "step": 2061 }, { "epoch": 0.08643618414847573, "grad_norm": 1.8013930320739746, "learning_rate": 9.916740901113055e-06, "loss": 0.5376, "step": 2062 }, { "epoch": 0.0864781027634847, "grad_norm": 1.821157455444336, "learning_rate": 9.916617486916523e-06, "loss": 0.5803, "step": 2063 }, { "epoch": 0.08652002137849366, "grad_norm": 1.8168176412582397, "learning_rate": 9.916493982088944e-06, "loss": 0.6369, "step": 2064 }, { "epoch": 0.08656193999350262, "grad_norm": 1.8296902179718018, "learning_rate": 9.916370386632595e-06, "loss": 0.6498, "step": 2065 }, { "epoch": 0.08660385860851158, "grad_norm": 1.6409049034118652, "learning_rate": 9.916246700549754e-06, "loss": 0.5354, "step": 2066 }, { "epoch": 0.08664577722352053, "grad_norm": 1.787764310836792, "learning_rate": 9.916122923842703e-06, "loss": 0.5529, "step": 2067 }, { "epoch": 0.08668769583852949, "grad_norm": 4.099018096923828, "learning_rate": 9.915999056513722e-06, "loss": 0.6031, "step": 2068 }, { "epoch": 0.08672961445353845, "grad_norm": 1.7721142768859863, "learning_rate": 9.915875098565095e-06, "loss": 0.5862, "step": 2069 }, { "epoch": 0.08677153306854742, "grad_norm": 1.777750849723816, "learning_rate": 9.915751049999105e-06, "loss": 0.5075, "step": 2070 }, { "epoch": 0.08681345168355638, "grad_norm": 1.8820496797561646, "learning_rate": 9.915626910818042e-06, "loss": 0.5838, "step": 2071 }, { "epoch": 0.08685537029856534, "grad_norm": 1.592432975769043, "learning_rate": 9.91550268102419e-06, "loss": 0.5752, "step": 2072 }, { "epoch": 0.08689728891357429, "grad_norm": 1.788124918937683, "learning_rate": 9.915378360619842e-06, "loss": 0.5262, "step": 2073 }, { "epoch": 0.08693920752858325, "grad_norm": 1.7705540657043457, "learning_rate": 9.915253949607292e-06, "loss": 0.5618, "step": 2074 }, { "epoch": 0.08698112614359221, "grad_norm": 1.9040625095367432, "learning_rate": 9.91512944798883e-06, "loss": 0.5714, "step": 2075 }, { "epoch": 0.08702304475860118, "grad_norm": 1.7651716470718384, "learning_rate": 9.91500485576675e-06, "loss": 0.5649, "step": 2076 }, { "epoch": 0.08706496337361014, "grad_norm": 2.2230312824249268, "learning_rate": 9.91488017294335e-06, "loss": 0.5948, "step": 2077 }, { "epoch": 0.0871068819886191, "grad_norm": 1.7707092761993408, "learning_rate": 9.91475539952093e-06, "loss": 0.5628, "step": 2078 }, { "epoch": 0.08714880060362806, "grad_norm": 1.7192785739898682, "learning_rate": 9.914630535501788e-06, "loss": 0.5246, "step": 2079 }, { "epoch": 0.08719071921863701, "grad_norm": 2.826127529144287, "learning_rate": 9.914505580888226e-06, "loss": 0.581, "step": 2080 }, { "epoch": 0.08723263783364597, "grad_norm": 1.7078146934509277, "learning_rate": 9.914380535682548e-06, "loss": 0.5353, "step": 2081 }, { "epoch": 0.08727455644865494, "grad_norm": 1.8502501249313354, "learning_rate": 9.914255399887058e-06, "loss": 0.6104, "step": 2082 }, { "epoch": 0.0873164750636639, "grad_norm": 1.9279439449310303, "learning_rate": 9.914130173504065e-06, "loss": 0.4934, "step": 2083 }, { "epoch": 0.08735839367867286, "grad_norm": 1.721403956413269, "learning_rate": 9.914004856535876e-06, "loss": 0.5637, "step": 2084 }, { "epoch": 0.08740031229368182, "grad_norm": 1.8064664602279663, "learning_rate": 9.913879448984799e-06, "loss": 0.6256, "step": 2085 }, { "epoch": 0.08744223090869077, "grad_norm": 2.0715172290802, "learning_rate": 9.913753950853148e-06, "loss": 0.6628, "step": 2086 }, { "epoch": 0.08748414952369973, "grad_norm": 1.5088123083114624, "learning_rate": 9.913628362143237e-06, "loss": 0.5851, "step": 2087 }, { "epoch": 0.0875260681387087, "grad_norm": 1.9578922986984253, "learning_rate": 9.91350268285738e-06, "loss": 0.596, "step": 2088 }, { "epoch": 0.08756798675371766, "grad_norm": 2.0765841007232666, "learning_rate": 9.913376912997893e-06, "loss": 0.5713, "step": 2089 }, { "epoch": 0.08760990536872662, "grad_norm": 1.8716027736663818, "learning_rate": 9.913251052567095e-06, "loss": 0.6265, "step": 2090 }, { "epoch": 0.08765182398373558, "grad_norm": 1.670964002609253, "learning_rate": 9.913125101567307e-06, "loss": 0.5786, "step": 2091 }, { "epoch": 0.08769374259874453, "grad_norm": 2.0192134380340576, "learning_rate": 9.91299906000085e-06, "loss": 0.5691, "step": 2092 }, { "epoch": 0.0877356612137535, "grad_norm": 2.1772713661193848, "learning_rate": 9.912872927870048e-06, "loss": 0.6263, "step": 2093 }, { "epoch": 0.08777757982876246, "grad_norm": 1.7897154092788696, "learning_rate": 9.912746705177222e-06, "loss": 0.5911, "step": 2094 }, { "epoch": 0.08781949844377142, "grad_norm": 1.5461426973342896, "learning_rate": 9.912620391924705e-06, "loss": 0.5924, "step": 2095 }, { "epoch": 0.08786141705878038, "grad_norm": 2.5608367919921875, "learning_rate": 9.912493988114822e-06, "loss": 0.5983, "step": 2096 }, { "epoch": 0.08790333567378934, "grad_norm": 1.9292680025100708, "learning_rate": 9.912367493749904e-06, "loss": 0.6104, "step": 2097 }, { "epoch": 0.08794525428879829, "grad_norm": 1.9132052659988403, "learning_rate": 9.912240908832284e-06, "loss": 0.6366, "step": 2098 }, { "epoch": 0.08798717290380725, "grad_norm": 1.6607879400253296, "learning_rate": 9.912114233364292e-06, "loss": 0.6334, "step": 2099 }, { "epoch": 0.08802909151881622, "grad_norm": 2.025569200515747, "learning_rate": 9.911987467348264e-06, "loss": 0.6351, "step": 2100 }, { "epoch": 0.08807101013382518, "grad_norm": 2.204493761062622, "learning_rate": 9.911860610786539e-06, "loss": 0.6054, "step": 2101 }, { "epoch": 0.08811292874883414, "grad_norm": 1.779425859451294, "learning_rate": 9.911733663681453e-06, "loss": 0.5725, "step": 2102 }, { "epoch": 0.0881548473638431, "grad_norm": 1.823643684387207, "learning_rate": 9.911606626035348e-06, "loss": 0.5371, "step": 2103 }, { "epoch": 0.08819676597885207, "grad_norm": 2.092526435852051, "learning_rate": 9.911479497850565e-06, "loss": 0.5637, "step": 2104 }, { "epoch": 0.08823868459386101, "grad_norm": 1.8237684965133667, "learning_rate": 9.911352279129447e-06, "loss": 0.5947, "step": 2105 }, { "epoch": 0.08828060320886998, "grad_norm": 1.9001063108444214, "learning_rate": 9.91122496987434e-06, "loss": 0.6163, "step": 2106 }, { "epoch": 0.08832252182387894, "grad_norm": 2.0797338485717773, "learning_rate": 9.91109757008759e-06, "loss": 0.625, "step": 2107 }, { "epoch": 0.0883644404388879, "grad_norm": 1.8477039337158203, "learning_rate": 9.910970079771545e-06, "loss": 0.6206, "step": 2108 }, { "epoch": 0.08840635905389686, "grad_norm": 2.1551897525787354, "learning_rate": 9.910842498928557e-06, "loss": 0.6223, "step": 2109 }, { "epoch": 0.08844827766890583, "grad_norm": 1.8500540256500244, "learning_rate": 9.910714827560976e-06, "loss": 0.6003, "step": 2110 }, { "epoch": 0.08849019628391477, "grad_norm": 1.8259968757629395, "learning_rate": 9.910587065671156e-06, "loss": 0.633, "step": 2111 }, { "epoch": 0.08853211489892374, "grad_norm": 2.414053440093994, "learning_rate": 9.910459213261452e-06, "loss": 0.6396, "step": 2112 }, { "epoch": 0.0885740335139327, "grad_norm": 1.8582104444503784, "learning_rate": 9.91033127033422e-06, "loss": 0.4817, "step": 2113 }, { "epoch": 0.08861595212894166, "grad_norm": 1.7930402755737305, "learning_rate": 9.91020323689182e-06, "loss": 0.6231, "step": 2114 }, { "epoch": 0.08865787074395062, "grad_norm": 2.1172537803649902, "learning_rate": 9.910075112936612e-06, "loss": 0.6537, "step": 2115 }, { "epoch": 0.08869978935895959, "grad_norm": 2.0656991004943848, "learning_rate": 9.909946898470958e-06, "loss": 0.6157, "step": 2116 }, { "epoch": 0.08874170797396853, "grad_norm": 1.9354459047317505, "learning_rate": 9.90981859349722e-06, "loss": 0.5772, "step": 2117 }, { "epoch": 0.0887836265889775, "grad_norm": 1.813149333000183, "learning_rate": 9.909690198017764e-06, "loss": 0.5956, "step": 2118 }, { "epoch": 0.08882554520398646, "grad_norm": 2.1451449394226074, "learning_rate": 9.909561712034954e-06, "loss": 0.6314, "step": 2119 }, { "epoch": 0.08886746381899542, "grad_norm": 2.0061025619506836, "learning_rate": 9.909433135551166e-06, "loss": 0.6789, "step": 2120 }, { "epoch": 0.08890938243400438, "grad_norm": 1.827871322631836, "learning_rate": 9.909304468568761e-06, "loss": 0.5776, "step": 2121 }, { "epoch": 0.08895130104901335, "grad_norm": 1.8429229259490967, "learning_rate": 9.909175711090117e-06, "loss": 0.6205, "step": 2122 }, { "epoch": 0.0889932196640223, "grad_norm": 1.9721719026565552, "learning_rate": 9.909046863117605e-06, "loss": 0.6355, "step": 2123 }, { "epoch": 0.08903513827903126, "grad_norm": 2.084131956100464, "learning_rate": 9.9089179246536e-06, "loss": 0.6041, "step": 2124 }, { "epoch": 0.08907705689404022, "grad_norm": 1.891722321510315, "learning_rate": 9.90878889570048e-06, "loss": 0.5827, "step": 2125 }, { "epoch": 0.08911897550904918, "grad_norm": 2.019660472869873, "learning_rate": 9.908659776260626e-06, "loss": 0.6188, "step": 2126 }, { "epoch": 0.08916089412405814, "grad_norm": 2.219763994216919, "learning_rate": 9.90853056633641e-06, "loss": 0.572, "step": 2127 }, { "epoch": 0.0892028127390671, "grad_norm": 2.0681827068328857, "learning_rate": 9.908401265930221e-06, "loss": 0.5705, "step": 2128 }, { "epoch": 0.08924473135407607, "grad_norm": 1.6112126111984253, "learning_rate": 9.908271875044442e-06, "loss": 0.5729, "step": 2129 }, { "epoch": 0.08928664996908502, "grad_norm": 2.6136765480041504, "learning_rate": 9.908142393681453e-06, "loss": 0.6173, "step": 2130 }, { "epoch": 0.08932856858409398, "grad_norm": 2.0766472816467285, "learning_rate": 9.908012821843646e-06, "loss": 0.6097, "step": 2131 }, { "epoch": 0.08937048719910294, "grad_norm": 2.2536239624023438, "learning_rate": 9.90788315953341e-06, "loss": 0.6134, "step": 2132 }, { "epoch": 0.0894124058141119, "grad_norm": 1.9926666021347046, "learning_rate": 9.90775340675313e-06, "loss": 0.5871, "step": 2133 }, { "epoch": 0.08945432442912087, "grad_norm": 1.7968571186065674, "learning_rate": 9.907623563505202e-06, "loss": 0.6229, "step": 2134 }, { "epoch": 0.08949624304412983, "grad_norm": 2.006646156311035, "learning_rate": 9.907493629792017e-06, "loss": 0.5933, "step": 2135 }, { "epoch": 0.08953816165913878, "grad_norm": 2.572939395904541, "learning_rate": 9.907363605615972e-06, "loss": 0.635, "step": 2136 }, { "epoch": 0.08958008027414774, "grad_norm": 2.328796625137329, "learning_rate": 9.907233490979462e-06, "loss": 0.6515, "step": 2137 }, { "epoch": 0.0896219988891567, "grad_norm": 1.6985538005828857, "learning_rate": 9.907103285884889e-06, "loss": 0.5506, "step": 2138 }, { "epoch": 0.08966391750416566, "grad_norm": 1.8278273344039917, "learning_rate": 9.906972990334649e-06, "loss": 0.5766, "step": 2139 }, { "epoch": 0.08970583611917463, "grad_norm": 1.8375288248062134, "learning_rate": 9.906842604331146e-06, "loss": 0.5527, "step": 2140 }, { "epoch": 0.08974775473418359, "grad_norm": 2.2140097618103027, "learning_rate": 9.906712127876783e-06, "loss": 0.5778, "step": 2141 }, { "epoch": 0.08978967334919254, "grad_norm": 2.3071579933166504, "learning_rate": 9.906581560973967e-06, "loss": 0.6267, "step": 2142 }, { "epoch": 0.0898315919642015, "grad_norm": 1.973393440246582, "learning_rate": 9.9064509036251e-06, "loss": 0.6004, "step": 2143 }, { "epoch": 0.08987351057921046, "grad_norm": 1.852741003036499, "learning_rate": 9.906320155832597e-06, "loss": 0.5244, "step": 2144 }, { "epoch": 0.08991542919421942, "grad_norm": 1.9120526313781738, "learning_rate": 9.906189317598863e-06, "loss": 0.5533, "step": 2145 }, { "epoch": 0.08995734780922839, "grad_norm": 2.284674882888794, "learning_rate": 9.90605838892631e-06, "loss": 0.7061, "step": 2146 }, { "epoch": 0.08999926642423735, "grad_norm": 1.9446276426315308, "learning_rate": 9.905927369817355e-06, "loss": 0.6333, "step": 2147 }, { "epoch": 0.0900411850392463, "grad_norm": 2.2206027507781982, "learning_rate": 9.90579626027441e-06, "loss": 0.6286, "step": 2148 }, { "epoch": 0.09008310365425526, "grad_norm": 2.0630757808685303, "learning_rate": 9.905665060299894e-06, "loss": 0.6003, "step": 2149 }, { "epoch": 0.09012502226926422, "grad_norm": 2.049396514892578, "learning_rate": 9.905533769896224e-06, "loss": 0.6037, "step": 2150 }, { "epoch": 0.09016694088427318, "grad_norm": 2.1732661724090576, "learning_rate": 9.905402389065821e-06, "loss": 0.5655, "step": 2151 }, { "epoch": 0.09020885949928215, "grad_norm": 5.561155319213867, "learning_rate": 9.905270917811104e-06, "loss": 0.5643, "step": 2152 }, { "epoch": 0.09025077811429111, "grad_norm": 1.7530308961868286, "learning_rate": 9.905139356134504e-06, "loss": 0.5676, "step": 2153 }, { "epoch": 0.09029269672930007, "grad_norm": 2.162360668182373, "learning_rate": 9.905007704038438e-06, "loss": 0.6455, "step": 2154 }, { "epoch": 0.09033461534430902, "grad_norm": 2.344675302505493, "learning_rate": 9.904875961525336e-06, "loss": 0.5629, "step": 2155 }, { "epoch": 0.09037653395931798, "grad_norm": 1.940867304801941, "learning_rate": 9.904744128597627e-06, "loss": 0.5637, "step": 2156 }, { "epoch": 0.09041845257432694, "grad_norm": 1.7869784832000732, "learning_rate": 9.90461220525774e-06, "loss": 0.5678, "step": 2157 }, { "epoch": 0.0904603711893359, "grad_norm": 1.4992947578430176, "learning_rate": 9.904480191508108e-06, "loss": 0.5549, "step": 2158 }, { "epoch": 0.09050228980434487, "grad_norm": 2.2461767196655273, "learning_rate": 9.904348087351163e-06, "loss": 0.5655, "step": 2159 }, { "epoch": 0.09054420841935383, "grad_norm": 1.852960228919983, "learning_rate": 9.904215892789342e-06, "loss": 0.5727, "step": 2160 }, { "epoch": 0.09058612703436278, "grad_norm": 1.8286482095718384, "learning_rate": 9.90408360782508e-06, "loss": 0.6563, "step": 2161 }, { "epoch": 0.09062804564937174, "grad_norm": 2.54695725440979, "learning_rate": 9.903951232460818e-06, "loss": 0.6101, "step": 2162 }, { "epoch": 0.0906699642643807, "grad_norm": 1.9282547235488892, "learning_rate": 9.903818766698993e-06, "loss": 0.5164, "step": 2163 }, { "epoch": 0.09071188287938967, "grad_norm": 2.1067891120910645, "learning_rate": 9.903686210542048e-06, "loss": 0.5932, "step": 2164 }, { "epoch": 0.09075380149439863, "grad_norm": 1.9701374769210815, "learning_rate": 9.903553563992427e-06, "loss": 0.6252, "step": 2165 }, { "epoch": 0.09079572010940759, "grad_norm": 1.754422903060913, "learning_rate": 9.903420827052577e-06, "loss": 0.6396, "step": 2166 }, { "epoch": 0.09083763872441654, "grad_norm": 1.5336799621582031, "learning_rate": 9.903287999724942e-06, "loss": 0.5617, "step": 2167 }, { "epoch": 0.0908795573394255, "grad_norm": 2.06575345993042, "learning_rate": 9.90315508201197e-06, "loss": 0.601, "step": 2168 }, { "epoch": 0.09092147595443446, "grad_norm": 1.7895638942718506, "learning_rate": 9.903022073916113e-06, "loss": 0.5717, "step": 2169 }, { "epoch": 0.09096339456944343, "grad_norm": 1.8890349864959717, "learning_rate": 9.902888975439824e-06, "loss": 0.5938, "step": 2170 }, { "epoch": 0.09100531318445239, "grad_norm": 1.6285202503204346, "learning_rate": 9.902755786585552e-06, "loss": 0.549, "step": 2171 }, { "epoch": 0.09104723179946135, "grad_norm": 2.436143159866333, "learning_rate": 9.902622507355757e-06, "loss": 0.6629, "step": 2172 }, { "epoch": 0.0910891504144703, "grad_norm": 1.75722336769104, "learning_rate": 9.902489137752892e-06, "loss": 0.6101, "step": 2173 }, { "epoch": 0.09113106902947926, "grad_norm": 1.8744497299194336, "learning_rate": 9.902355677779419e-06, "loss": 0.5799, "step": 2174 }, { "epoch": 0.09117298764448822, "grad_norm": 1.8349382877349854, "learning_rate": 9.902222127437796e-06, "loss": 0.5775, "step": 2175 }, { "epoch": 0.09121490625949719, "grad_norm": 1.9655704498291016, "learning_rate": 9.902088486730485e-06, "loss": 0.5929, "step": 2176 }, { "epoch": 0.09125682487450615, "grad_norm": 1.9767848253250122, "learning_rate": 9.90195475565995e-06, "loss": 0.579, "step": 2177 }, { "epoch": 0.09129874348951511, "grad_norm": 1.7901688814163208, "learning_rate": 9.901820934228654e-06, "loss": 0.6414, "step": 2178 }, { "epoch": 0.09134066210452407, "grad_norm": 1.7850216627120972, "learning_rate": 9.901687022439068e-06, "loss": 0.5553, "step": 2179 }, { "epoch": 0.09138258071953302, "grad_norm": 1.9853707551956177, "learning_rate": 9.901553020293657e-06, "loss": 0.5994, "step": 2180 }, { "epoch": 0.09142449933454198, "grad_norm": 1.921946406364441, "learning_rate": 9.901418927794892e-06, "loss": 0.6577, "step": 2181 }, { "epoch": 0.09146641794955095, "grad_norm": 1.9133824110031128, "learning_rate": 9.901284744945246e-06, "loss": 0.6184, "step": 2182 }, { "epoch": 0.09150833656455991, "grad_norm": 2.0077807903289795, "learning_rate": 9.901150471747191e-06, "loss": 0.5787, "step": 2183 }, { "epoch": 0.09155025517956887, "grad_norm": 1.8595730066299438, "learning_rate": 9.901016108203203e-06, "loss": 0.6335, "step": 2184 }, { "epoch": 0.09159217379457783, "grad_norm": 2.1714859008789062, "learning_rate": 9.900881654315759e-06, "loss": 0.6065, "step": 2185 }, { "epoch": 0.09163409240958678, "grad_norm": 1.9107427597045898, "learning_rate": 9.900747110087336e-06, "loss": 0.6042, "step": 2186 }, { "epoch": 0.09167601102459574, "grad_norm": 2.0308406352996826, "learning_rate": 9.900612475520415e-06, "loss": 0.5523, "step": 2187 }, { "epoch": 0.0917179296396047, "grad_norm": 2.058889627456665, "learning_rate": 9.900477750617477e-06, "loss": 0.6078, "step": 2188 }, { "epoch": 0.09175984825461367, "grad_norm": 1.765581488609314, "learning_rate": 9.900342935381008e-06, "loss": 0.5237, "step": 2189 }, { "epoch": 0.09180176686962263, "grad_norm": 1.9957178831100464, "learning_rate": 9.900208029813492e-06, "loss": 0.6135, "step": 2190 }, { "epoch": 0.0918436854846316, "grad_norm": 1.8451039791107178, "learning_rate": 9.900073033917415e-06, "loss": 0.627, "step": 2191 }, { "epoch": 0.09188560409964054, "grad_norm": 1.8057881593704224, "learning_rate": 9.899937947695264e-06, "loss": 0.5624, "step": 2192 }, { "epoch": 0.0919275227146495, "grad_norm": 2.0334596633911133, "learning_rate": 9.899802771149532e-06, "loss": 0.6151, "step": 2193 }, { "epoch": 0.09196944132965847, "grad_norm": 1.9504154920578003, "learning_rate": 9.899667504282711e-06, "loss": 0.54, "step": 2194 }, { "epoch": 0.09201135994466743, "grad_norm": 1.6502257585525513, "learning_rate": 9.89953214709729e-06, "loss": 0.5727, "step": 2195 }, { "epoch": 0.09205327855967639, "grad_norm": 1.6194103956222534, "learning_rate": 9.89939669959577e-06, "loss": 0.5035, "step": 2196 }, { "epoch": 0.09209519717468535, "grad_norm": 2.0485925674438477, "learning_rate": 9.899261161780644e-06, "loss": 0.5988, "step": 2197 }, { "epoch": 0.0921371157896943, "grad_norm": 1.814632534980774, "learning_rate": 9.899125533654413e-06, "loss": 0.6063, "step": 2198 }, { "epoch": 0.09217903440470326, "grad_norm": 1.673205018043518, "learning_rate": 9.898989815219573e-06, "loss": 0.5596, "step": 2199 }, { "epoch": 0.09222095301971223, "grad_norm": 2.907294511795044, "learning_rate": 9.89885400647863e-06, "loss": 0.5812, "step": 2200 }, { "epoch": 0.09226287163472119, "grad_norm": 2.0127201080322266, "learning_rate": 9.898718107434086e-06, "loss": 0.6389, "step": 2201 }, { "epoch": 0.09230479024973015, "grad_norm": 1.8680412769317627, "learning_rate": 9.898582118088444e-06, "loss": 0.5725, "step": 2202 }, { "epoch": 0.09234670886473911, "grad_norm": 2.2357959747314453, "learning_rate": 9.898446038444216e-06, "loss": 0.5772, "step": 2203 }, { "epoch": 0.09238862747974808, "grad_norm": 1.832002878189087, "learning_rate": 9.898309868503905e-06, "loss": 0.6413, "step": 2204 }, { "epoch": 0.09243054609475702, "grad_norm": 1.8894624710083008, "learning_rate": 9.898173608270023e-06, "loss": 0.5652, "step": 2205 }, { "epoch": 0.09247246470976599, "grad_norm": 2.4744699001312256, "learning_rate": 9.898037257745084e-06, "loss": 0.639, "step": 2206 }, { "epoch": 0.09251438332477495, "grad_norm": 1.9120889902114868, "learning_rate": 9.897900816931597e-06, "loss": 0.6331, "step": 2207 }, { "epoch": 0.09255630193978391, "grad_norm": 2.177194595336914, "learning_rate": 9.897764285832081e-06, "loss": 0.5621, "step": 2208 }, { "epoch": 0.09259822055479287, "grad_norm": 2.0115699768066406, "learning_rate": 9.897627664449051e-06, "loss": 0.667, "step": 2209 }, { "epoch": 0.09264013916980184, "grad_norm": 1.9453437328338623, "learning_rate": 9.897490952785027e-06, "loss": 0.6531, "step": 2210 }, { "epoch": 0.09268205778481078, "grad_norm": 1.8196135759353638, "learning_rate": 9.897354150842526e-06, "loss": 0.5648, "step": 2211 }, { "epoch": 0.09272397639981975, "grad_norm": 1.7325619459152222, "learning_rate": 9.897217258624074e-06, "loss": 0.5634, "step": 2212 }, { "epoch": 0.09276589501482871, "grad_norm": 2.0180468559265137, "learning_rate": 9.89708027613219e-06, "loss": 0.5237, "step": 2213 }, { "epoch": 0.09280781362983767, "grad_norm": 1.7915053367614746, "learning_rate": 9.896943203369403e-06, "loss": 0.6171, "step": 2214 }, { "epoch": 0.09284973224484663, "grad_norm": 1.862127661705017, "learning_rate": 9.896806040338236e-06, "loss": 0.5252, "step": 2215 }, { "epoch": 0.0928916508598556, "grad_norm": 2.475701332092285, "learning_rate": 9.89666878704122e-06, "loss": 0.5784, "step": 2216 }, { "epoch": 0.09293356947486454, "grad_norm": 1.9437659978866577, "learning_rate": 9.896531443480883e-06, "loss": 0.5935, "step": 2217 }, { "epoch": 0.0929754880898735, "grad_norm": 2.030518054962158, "learning_rate": 9.89639400965976e-06, "loss": 0.7174, "step": 2218 }, { "epoch": 0.09301740670488247, "grad_norm": 1.9097528457641602, "learning_rate": 9.896256485580382e-06, "loss": 0.5682, "step": 2219 }, { "epoch": 0.09305932531989143, "grad_norm": 2.395500421524048, "learning_rate": 9.896118871245283e-06, "loss": 0.6894, "step": 2220 }, { "epoch": 0.0931012439349004, "grad_norm": 1.912899136543274, "learning_rate": 9.895981166657003e-06, "loss": 0.5995, "step": 2221 }, { "epoch": 0.09314316254990936, "grad_norm": 1.7511900663375854, "learning_rate": 9.895843371818079e-06, "loss": 0.5675, "step": 2222 }, { "epoch": 0.0931850811649183, "grad_norm": 1.9226981401443481, "learning_rate": 9.89570548673105e-06, "loss": 0.5887, "step": 2223 }, { "epoch": 0.09322699977992727, "grad_norm": 2.4279110431671143, "learning_rate": 9.895567511398458e-06, "loss": 0.6192, "step": 2224 }, { "epoch": 0.09326891839493623, "grad_norm": 2.1236941814422607, "learning_rate": 9.895429445822847e-06, "loss": 0.5534, "step": 2225 }, { "epoch": 0.09331083700994519, "grad_norm": 1.8395193815231323, "learning_rate": 9.895291290006763e-06, "loss": 0.5916, "step": 2226 }, { "epoch": 0.09335275562495415, "grad_norm": 2.0765068531036377, "learning_rate": 9.89515304395275e-06, "loss": 0.5309, "step": 2227 }, { "epoch": 0.09339467423996312, "grad_norm": 1.7649736404418945, "learning_rate": 9.89501470766336e-06, "loss": 0.6457, "step": 2228 }, { "epoch": 0.09343659285497208, "grad_norm": 2.4916577339172363, "learning_rate": 9.89487628114114e-06, "loss": 0.5893, "step": 2229 }, { "epoch": 0.09347851146998103, "grad_norm": 1.7691353559494019, "learning_rate": 9.894737764388642e-06, "loss": 0.5883, "step": 2230 }, { "epoch": 0.09352043008498999, "grad_norm": 1.8300727605819702, "learning_rate": 9.89459915740842e-06, "loss": 0.6002, "step": 2231 }, { "epoch": 0.09356234869999895, "grad_norm": 5.164375305175781, "learning_rate": 9.89446046020303e-06, "loss": 0.5752, "step": 2232 }, { "epoch": 0.09360426731500791, "grad_norm": 1.6642259359359741, "learning_rate": 9.894321672775028e-06, "loss": 0.5824, "step": 2233 }, { "epoch": 0.09364618593001688, "grad_norm": 1.8970779180526733, "learning_rate": 9.894182795126972e-06, "loss": 0.578, "step": 2234 }, { "epoch": 0.09368810454502584, "grad_norm": 1.9741219282150269, "learning_rate": 9.894043827261421e-06, "loss": 0.6203, "step": 2235 }, { "epoch": 0.09373002316003479, "grad_norm": 1.9693771600723267, "learning_rate": 9.893904769180939e-06, "loss": 0.5252, "step": 2236 }, { "epoch": 0.09377194177504375, "grad_norm": 1.9482229948043823, "learning_rate": 9.893765620888087e-06, "loss": 0.5561, "step": 2237 }, { "epoch": 0.09381386039005271, "grad_norm": 1.6174731254577637, "learning_rate": 9.893626382385434e-06, "loss": 0.579, "step": 2238 }, { "epoch": 0.09385577900506167, "grad_norm": 3.649634599685669, "learning_rate": 9.893487053675542e-06, "loss": 0.6155, "step": 2239 }, { "epoch": 0.09389769762007064, "grad_norm": 3.1007161140441895, "learning_rate": 9.89334763476098e-06, "loss": 0.671, "step": 2240 }, { "epoch": 0.0939396162350796, "grad_norm": 2.9658708572387695, "learning_rate": 9.89320812564432e-06, "loss": 0.6002, "step": 2241 }, { "epoch": 0.09398153485008855, "grad_norm": 1.895462989807129, "learning_rate": 9.893068526328133e-06, "loss": 0.5758, "step": 2242 }, { "epoch": 0.09402345346509751, "grad_norm": 1.926766037940979, "learning_rate": 9.892928836814994e-06, "loss": 0.6187, "step": 2243 }, { "epoch": 0.09406537208010647, "grad_norm": 2.3388030529022217, "learning_rate": 9.892789057107474e-06, "loss": 0.5895, "step": 2244 }, { "epoch": 0.09410729069511543, "grad_norm": 1.887125849723816, "learning_rate": 9.892649187208152e-06, "loss": 0.5666, "step": 2245 }, { "epoch": 0.0941492093101244, "grad_norm": 2.5293662548065186, "learning_rate": 9.892509227119606e-06, "loss": 0.6028, "step": 2246 }, { "epoch": 0.09419112792513336, "grad_norm": 1.6563549041748047, "learning_rate": 9.892369176844419e-06, "loss": 0.5424, "step": 2247 }, { "epoch": 0.09423304654014232, "grad_norm": 1.9640777111053467, "learning_rate": 9.892229036385165e-06, "loss": 0.6165, "step": 2248 }, { "epoch": 0.09427496515515127, "grad_norm": 1.795101523399353, "learning_rate": 9.892088805744435e-06, "loss": 0.5855, "step": 2249 }, { "epoch": 0.09431688377016023, "grad_norm": 2.184596061706543, "learning_rate": 9.891948484924811e-06, "loss": 0.559, "step": 2250 }, { "epoch": 0.0943588023851692, "grad_norm": 2.2552855014801025, "learning_rate": 9.891808073928879e-06, "loss": 0.5531, "step": 2251 }, { "epoch": 0.09440072100017816, "grad_norm": 1.6810213327407837, "learning_rate": 9.891667572759229e-06, "loss": 0.5311, "step": 2252 }, { "epoch": 0.09444263961518712, "grad_norm": 1.9337514638900757, "learning_rate": 9.891526981418447e-06, "loss": 0.6021, "step": 2253 }, { "epoch": 0.09448455823019608, "grad_norm": 1.923193335533142, "learning_rate": 9.89138629990913e-06, "loss": 0.6363, "step": 2254 }, { "epoch": 0.09452647684520503, "grad_norm": 2.13315749168396, "learning_rate": 9.891245528233869e-06, "loss": 0.5752, "step": 2255 }, { "epoch": 0.09456839546021399, "grad_norm": 1.6367093324661255, "learning_rate": 9.891104666395257e-06, "loss": 0.5416, "step": 2256 }, { "epoch": 0.09461031407522295, "grad_norm": 1.9678207635879517, "learning_rate": 9.890963714395891e-06, "loss": 0.5437, "step": 2257 }, { "epoch": 0.09465223269023192, "grad_norm": 1.7481168508529663, "learning_rate": 9.890822672238373e-06, "loss": 0.5211, "step": 2258 }, { "epoch": 0.09469415130524088, "grad_norm": 1.9007397890090942, "learning_rate": 9.8906815399253e-06, "loss": 0.5853, "step": 2259 }, { "epoch": 0.09473606992024984, "grad_norm": 2.0106728076934814, "learning_rate": 9.890540317459272e-06, "loss": 0.615, "step": 2260 }, { "epoch": 0.09477798853525879, "grad_norm": 1.8710031509399414, "learning_rate": 9.890399004842897e-06, "loss": 0.5393, "step": 2261 }, { "epoch": 0.09481990715026775, "grad_norm": 1.8337517976760864, "learning_rate": 9.890257602078776e-06, "loss": 0.6367, "step": 2262 }, { "epoch": 0.09486182576527671, "grad_norm": 1.9765931367874146, "learning_rate": 9.890116109169517e-06, "loss": 0.5612, "step": 2263 }, { "epoch": 0.09490374438028568, "grad_norm": 1.700628399848938, "learning_rate": 9.889974526117728e-06, "loss": 0.5482, "step": 2264 }, { "epoch": 0.09494566299529464, "grad_norm": 2.4244611263275146, "learning_rate": 9.889832852926017e-06, "loss": 0.573, "step": 2265 }, { "epoch": 0.0949875816103036, "grad_norm": 1.9237167835235596, "learning_rate": 9.889691089596997e-06, "loss": 0.5935, "step": 2266 }, { "epoch": 0.09502950022531255, "grad_norm": 2.005565881729126, "learning_rate": 9.889549236133285e-06, "loss": 0.5752, "step": 2267 }, { "epoch": 0.09507141884032151, "grad_norm": 1.8591279983520508, "learning_rate": 9.88940729253749e-06, "loss": 0.5379, "step": 2268 }, { "epoch": 0.09511333745533047, "grad_norm": 2.2406258583068848, "learning_rate": 9.889265258812231e-06, "loss": 0.5684, "step": 2269 }, { "epoch": 0.09515525607033944, "grad_norm": 2.099255084991455, "learning_rate": 9.889123134960127e-06, "loss": 0.5623, "step": 2270 }, { "epoch": 0.0951971746853484, "grad_norm": 2.3377134799957275, "learning_rate": 9.888980920983797e-06, "loss": 0.6359, "step": 2271 }, { "epoch": 0.09523909330035736, "grad_norm": 1.9797160625457764, "learning_rate": 9.888838616885862e-06, "loss": 0.6466, "step": 2272 }, { "epoch": 0.09528101191536632, "grad_norm": 1.8796943426132202, "learning_rate": 9.888696222668946e-06, "loss": 0.6011, "step": 2273 }, { "epoch": 0.09532293053037527, "grad_norm": 1.866296648979187, "learning_rate": 9.888553738335673e-06, "loss": 0.5701, "step": 2274 }, { "epoch": 0.09536484914538423, "grad_norm": 1.7408207654953003, "learning_rate": 9.888411163888671e-06, "loss": 0.5838, "step": 2275 }, { "epoch": 0.0954067677603932, "grad_norm": 1.9432488679885864, "learning_rate": 9.888268499330566e-06, "loss": 0.5954, "step": 2276 }, { "epoch": 0.09544868637540216, "grad_norm": 3.982393980026245, "learning_rate": 9.88812574466399e-06, "loss": 0.5893, "step": 2277 }, { "epoch": 0.09549060499041112, "grad_norm": 2.1036272048950195, "learning_rate": 9.887982899891576e-06, "loss": 0.5629, "step": 2278 }, { "epoch": 0.09553252360542008, "grad_norm": 2.1685729026794434, "learning_rate": 9.88783996501595e-06, "loss": 0.5658, "step": 2279 }, { "epoch": 0.09557444222042903, "grad_norm": 2.0024895668029785, "learning_rate": 9.887696940039754e-06, "loss": 0.5873, "step": 2280 }, { "epoch": 0.095616360835438, "grad_norm": 1.9570319652557373, "learning_rate": 9.887553824965621e-06, "loss": 0.5853, "step": 2281 }, { "epoch": 0.09565827945044696, "grad_norm": 1.9199492931365967, "learning_rate": 9.887410619796192e-06, "loss": 0.5905, "step": 2282 }, { "epoch": 0.09570019806545592, "grad_norm": 1.8670483827590942, "learning_rate": 9.887267324534102e-06, "loss": 0.5669, "step": 2283 }, { "epoch": 0.09574211668046488, "grad_norm": 1.7265102863311768, "learning_rate": 9.887123939181997e-06, "loss": 0.5368, "step": 2284 }, { "epoch": 0.09578403529547384, "grad_norm": 2.1351940631866455, "learning_rate": 9.88698046374252e-06, "loss": 0.5625, "step": 2285 }, { "epoch": 0.09582595391048279, "grad_norm": 2.261746883392334, "learning_rate": 9.88683689821831e-06, "loss": 0.559, "step": 2286 }, { "epoch": 0.09586787252549175, "grad_norm": 2.0691282749176025, "learning_rate": 9.88669324261202e-06, "loss": 0.5795, "step": 2287 }, { "epoch": 0.09590979114050072, "grad_norm": 2.3670220375061035, "learning_rate": 9.886549496926297e-06, "loss": 0.5959, "step": 2288 }, { "epoch": 0.09595170975550968, "grad_norm": 1.9804502725601196, "learning_rate": 9.886405661163787e-06, "loss": 0.6369, "step": 2289 }, { "epoch": 0.09599362837051864, "grad_norm": 2.4501562118530273, "learning_rate": 9.886261735327146e-06, "loss": 0.6612, "step": 2290 }, { "epoch": 0.0960355469855276, "grad_norm": 1.877938985824585, "learning_rate": 9.886117719419024e-06, "loss": 0.574, "step": 2291 }, { "epoch": 0.09607746560053655, "grad_norm": 2.5076942443847656, "learning_rate": 9.885973613442076e-06, "loss": 0.573, "step": 2292 }, { "epoch": 0.09611938421554551, "grad_norm": 2.263702392578125, "learning_rate": 9.88582941739896e-06, "loss": 0.6253, "step": 2293 }, { "epoch": 0.09616130283055448, "grad_norm": 2.169928789138794, "learning_rate": 9.885685131292333e-06, "loss": 0.5471, "step": 2294 }, { "epoch": 0.09620322144556344, "grad_norm": 3.015171766281128, "learning_rate": 9.885540755124856e-06, "loss": 0.573, "step": 2295 }, { "epoch": 0.0962451400605724, "grad_norm": 2.2153449058532715, "learning_rate": 9.885396288899187e-06, "loss": 0.6142, "step": 2296 }, { "epoch": 0.09628705867558136, "grad_norm": 1.7146023511886597, "learning_rate": 9.885251732617993e-06, "loss": 0.5288, "step": 2297 }, { "epoch": 0.09632897729059033, "grad_norm": 2.3397417068481445, "learning_rate": 9.885107086283937e-06, "loss": 0.6065, "step": 2298 }, { "epoch": 0.09637089590559927, "grad_norm": 1.897161841392517, "learning_rate": 9.884962349899684e-06, "loss": 0.556, "step": 2299 }, { "epoch": 0.09641281452060824, "grad_norm": 2.190352201461792, "learning_rate": 9.884817523467905e-06, "loss": 0.5392, "step": 2300 }, { "epoch": 0.0964547331356172, "grad_norm": 2.0552854537963867, "learning_rate": 9.884672606991266e-06, "loss": 0.6414, "step": 2301 }, { "epoch": 0.09649665175062616, "grad_norm": 2.006742000579834, "learning_rate": 9.884527600472443e-06, "loss": 0.6535, "step": 2302 }, { "epoch": 0.09653857036563512, "grad_norm": 2.0515525341033936, "learning_rate": 9.884382503914102e-06, "loss": 0.5685, "step": 2303 }, { "epoch": 0.09658048898064409, "grad_norm": 2.2476837635040283, "learning_rate": 9.884237317318926e-06, "loss": 0.6203, "step": 2304 }, { "epoch": 0.09662240759565303, "grad_norm": 2.275414228439331, "learning_rate": 9.884092040689586e-06, "loss": 0.5348, "step": 2305 }, { "epoch": 0.096664326210662, "grad_norm": 1.975216031074524, "learning_rate": 9.883946674028761e-06, "loss": 0.6325, "step": 2306 }, { "epoch": 0.09670624482567096, "grad_norm": 1.8967347145080566, "learning_rate": 9.883801217339131e-06, "loss": 0.6071, "step": 2307 }, { "epoch": 0.09674816344067992, "grad_norm": 1.881808876991272, "learning_rate": 9.883655670623378e-06, "loss": 0.5711, "step": 2308 }, { "epoch": 0.09679008205568888, "grad_norm": 2.4012715816497803, "learning_rate": 9.883510033884183e-06, "loss": 0.6732, "step": 2309 }, { "epoch": 0.09683200067069785, "grad_norm": 2.524492025375366, "learning_rate": 9.883364307124232e-06, "loss": 0.5952, "step": 2310 }, { "epoch": 0.0968739192857068, "grad_norm": 2.2516698837280273, "learning_rate": 9.88321849034621e-06, "loss": 0.599, "step": 2311 }, { "epoch": 0.09691583790071576, "grad_norm": 1.7027032375335693, "learning_rate": 9.883072583552809e-06, "loss": 0.5096, "step": 2312 }, { "epoch": 0.09695775651572472, "grad_norm": 2.0644571781158447, "learning_rate": 9.882926586746715e-06, "loss": 0.5439, "step": 2313 }, { "epoch": 0.09699967513073368, "grad_norm": 2.058392286300659, "learning_rate": 9.882780499930617e-06, "loss": 0.6224, "step": 2314 }, { "epoch": 0.09704159374574264, "grad_norm": 2.1597859859466553, "learning_rate": 9.882634323107213e-06, "loss": 0.5836, "step": 2315 }, { "epoch": 0.0970835123607516, "grad_norm": 2.0810580253601074, "learning_rate": 9.882488056279196e-06, "loss": 0.6309, "step": 2316 }, { "epoch": 0.09712543097576055, "grad_norm": 2.0214362144470215, "learning_rate": 9.88234169944926e-06, "loss": 0.6605, "step": 2317 }, { "epoch": 0.09716734959076952, "grad_norm": 1.8643074035644531, "learning_rate": 9.882195252620104e-06, "loss": 0.5553, "step": 2318 }, { "epoch": 0.09720926820577848, "grad_norm": 2.1324644088745117, "learning_rate": 9.882048715794429e-06, "loss": 0.6182, "step": 2319 }, { "epoch": 0.09725118682078744, "grad_norm": 2.029989242553711, "learning_rate": 9.881902088974936e-06, "loss": 0.6509, "step": 2320 }, { "epoch": 0.0972931054357964, "grad_norm": 1.906558871269226, "learning_rate": 9.881755372164325e-06, "loss": 0.6269, "step": 2321 }, { "epoch": 0.09733502405080537, "grad_norm": 2.0230953693389893, "learning_rate": 9.881608565365304e-06, "loss": 0.5701, "step": 2322 }, { "epoch": 0.09737694266581433, "grad_norm": 1.9596832990646362, "learning_rate": 9.881461668580576e-06, "loss": 0.5826, "step": 2323 }, { "epoch": 0.09741886128082328, "grad_norm": 1.8501468896865845, "learning_rate": 9.881314681812852e-06, "loss": 0.5728, "step": 2324 }, { "epoch": 0.09746077989583224, "grad_norm": 2.2488882541656494, "learning_rate": 9.88116760506484e-06, "loss": 0.6121, "step": 2325 }, { "epoch": 0.0975026985108412, "grad_norm": 1.815647840499878, "learning_rate": 9.88102043833925e-06, "loss": 0.5194, "step": 2326 }, { "epoch": 0.09754461712585016, "grad_norm": 1.886448860168457, "learning_rate": 9.880873181638797e-06, "loss": 0.6247, "step": 2327 }, { "epoch": 0.09758653574085913, "grad_norm": 1.8806421756744385, "learning_rate": 9.880725834966194e-06, "loss": 0.5697, "step": 2328 }, { "epoch": 0.09762845435586809, "grad_norm": 1.8913966417312622, "learning_rate": 9.880578398324158e-06, "loss": 0.5707, "step": 2329 }, { "epoch": 0.09767037297087704, "grad_norm": 2.261909246444702, "learning_rate": 9.880430871715408e-06, "loss": 0.5789, "step": 2330 }, { "epoch": 0.097712291585886, "grad_norm": 2.310274362564087, "learning_rate": 9.880283255142659e-06, "loss": 0.5854, "step": 2331 }, { "epoch": 0.09775421020089496, "grad_norm": 1.8986047506332397, "learning_rate": 9.880135548608637e-06, "loss": 0.52, "step": 2332 }, { "epoch": 0.09779612881590392, "grad_norm": 1.8429330587387085, "learning_rate": 9.879987752116063e-06, "loss": 0.5732, "step": 2333 }, { "epoch": 0.09783804743091289, "grad_norm": 2.1487181186676025, "learning_rate": 9.87983986566766e-06, "loss": 0.5962, "step": 2334 }, { "epoch": 0.09787996604592185, "grad_norm": 2.108668088912964, "learning_rate": 9.879691889266154e-06, "loss": 0.5483, "step": 2335 }, { "epoch": 0.0979218846609308, "grad_norm": 2.881833553314209, "learning_rate": 9.879543822914277e-06, "loss": 0.5195, "step": 2336 }, { "epoch": 0.09796380327593976, "grad_norm": 1.6698205471038818, "learning_rate": 9.879395666614754e-06, "loss": 0.5683, "step": 2337 }, { "epoch": 0.09800572189094872, "grad_norm": 2.1147170066833496, "learning_rate": 9.879247420370318e-06, "loss": 0.5964, "step": 2338 }, { "epoch": 0.09804764050595768, "grad_norm": 1.8800411224365234, "learning_rate": 9.879099084183698e-06, "loss": 0.5721, "step": 2339 }, { "epoch": 0.09808955912096665, "grad_norm": 2.836693286895752, "learning_rate": 9.878950658057635e-06, "loss": 0.5354, "step": 2340 }, { "epoch": 0.09813147773597561, "grad_norm": 2.2790520191192627, "learning_rate": 9.878802141994862e-06, "loss": 0.6029, "step": 2341 }, { "epoch": 0.09817339635098456, "grad_norm": 2.124866485595703, "learning_rate": 9.878653535998113e-06, "loss": 0.6292, "step": 2342 }, { "epoch": 0.09821531496599352, "grad_norm": 1.988811731338501, "learning_rate": 9.878504840070133e-06, "loss": 0.6101, "step": 2343 }, { "epoch": 0.09825723358100248, "grad_norm": 1.8159470558166504, "learning_rate": 9.878356054213661e-06, "loss": 0.5584, "step": 2344 }, { "epoch": 0.09829915219601144, "grad_norm": 1.7610067129135132, "learning_rate": 9.878207178431438e-06, "loss": 0.5142, "step": 2345 }, { "epoch": 0.0983410708110204, "grad_norm": 1.963030457496643, "learning_rate": 9.878058212726209e-06, "loss": 0.5772, "step": 2346 }, { "epoch": 0.09838298942602937, "grad_norm": 2.0287535190582275, "learning_rate": 9.877909157100722e-06, "loss": 0.5386, "step": 2347 }, { "epoch": 0.09842490804103833, "grad_norm": 1.8622887134552002, "learning_rate": 9.877760011557722e-06, "loss": 0.5629, "step": 2348 }, { "epoch": 0.09846682665604728, "grad_norm": 1.667463779449463, "learning_rate": 9.877610776099961e-06, "loss": 0.5592, "step": 2349 }, { "epoch": 0.09850874527105624, "grad_norm": 1.7993773221969604, "learning_rate": 9.877461450730189e-06, "loss": 0.5226, "step": 2350 }, { "epoch": 0.0985506638860652, "grad_norm": 2.8888909816741943, "learning_rate": 9.877312035451153e-06, "loss": 0.5569, "step": 2351 }, { "epoch": 0.09859258250107417, "grad_norm": 3.019322633743286, "learning_rate": 9.877162530265619e-06, "loss": 0.5109, "step": 2352 }, { "epoch": 0.09863450111608313, "grad_norm": 2.0400190353393555, "learning_rate": 9.877012935176332e-06, "loss": 0.6725, "step": 2353 }, { "epoch": 0.09867641973109209, "grad_norm": 1.9262462854385376, "learning_rate": 9.876863250186056e-06, "loss": 0.6237, "step": 2354 }, { "epoch": 0.09871833834610104, "grad_norm": 2.2854201793670654, "learning_rate": 9.876713475297546e-06, "loss": 0.6031, "step": 2355 }, { "epoch": 0.09876025696111, "grad_norm": 1.9625113010406494, "learning_rate": 9.876563610513566e-06, "loss": 0.5936, "step": 2356 }, { "epoch": 0.09880217557611896, "grad_norm": 2.177833318710327, "learning_rate": 9.876413655836879e-06, "loss": 0.5541, "step": 2357 }, { "epoch": 0.09884409419112793, "grad_norm": 1.8978192806243896, "learning_rate": 9.876263611270245e-06, "loss": 0.4953, "step": 2358 }, { "epoch": 0.09888601280613689, "grad_norm": 1.869869351387024, "learning_rate": 9.876113476816434e-06, "loss": 0.6097, "step": 2359 }, { "epoch": 0.09892793142114585, "grad_norm": 1.7298002243041992, "learning_rate": 9.875963252478211e-06, "loss": 0.558, "step": 2360 }, { "epoch": 0.0989698500361548, "grad_norm": 1.8654073476791382, "learning_rate": 9.875812938258347e-06, "loss": 0.558, "step": 2361 }, { "epoch": 0.09901176865116376, "grad_norm": 2.0048959255218506, "learning_rate": 9.875662534159612e-06, "loss": 0.5727, "step": 2362 }, { "epoch": 0.09905368726617272, "grad_norm": 1.895506739616394, "learning_rate": 9.87551204018478e-06, "loss": 0.6297, "step": 2363 }, { "epoch": 0.09909560588118169, "grad_norm": 1.9401657581329346, "learning_rate": 9.875361456336622e-06, "loss": 0.5571, "step": 2364 }, { "epoch": 0.09913752449619065, "grad_norm": 2.0122649669647217, "learning_rate": 9.875210782617915e-06, "loss": 0.5511, "step": 2365 }, { "epoch": 0.09917944311119961, "grad_norm": 1.7768220901489258, "learning_rate": 9.87506001903144e-06, "loss": 0.5283, "step": 2366 }, { "epoch": 0.09922136172620856, "grad_norm": 2.6588852405548096, "learning_rate": 9.87490916557997e-06, "loss": 0.6462, "step": 2367 }, { "epoch": 0.09926328034121752, "grad_norm": 2.1121935844421387, "learning_rate": 9.87475822226629e-06, "loss": 0.6077, "step": 2368 }, { "epoch": 0.09930519895622648, "grad_norm": 2.0539333820343018, "learning_rate": 9.874607189093181e-06, "loss": 0.5739, "step": 2369 }, { "epoch": 0.09934711757123545, "grad_norm": 1.9038231372833252, "learning_rate": 9.874456066063428e-06, "loss": 0.5411, "step": 2370 }, { "epoch": 0.09938903618624441, "grad_norm": 1.8547040224075317, "learning_rate": 9.874304853179815e-06, "loss": 0.5579, "step": 2371 }, { "epoch": 0.09943095480125337, "grad_norm": 2.403979539871216, "learning_rate": 9.874153550445132e-06, "loss": 0.5604, "step": 2372 }, { "epoch": 0.09947287341626233, "grad_norm": 1.772902250289917, "learning_rate": 9.874002157862165e-06, "loss": 0.5407, "step": 2373 }, { "epoch": 0.09951479203127128, "grad_norm": 1.7410002946853638, "learning_rate": 9.87385067543371e-06, "loss": 0.5951, "step": 2374 }, { "epoch": 0.09955671064628024, "grad_norm": 2.1517157554626465, "learning_rate": 9.873699103162552e-06, "loss": 0.6036, "step": 2375 }, { "epoch": 0.0995986292612892, "grad_norm": 1.7313473224639893, "learning_rate": 9.87354744105149e-06, "loss": 0.55, "step": 2376 }, { "epoch": 0.09964054787629817, "grad_norm": 1.8207008838653564, "learning_rate": 9.873395689103318e-06, "loss": 0.5589, "step": 2377 }, { "epoch": 0.09968246649130713, "grad_norm": 3.175940990447998, "learning_rate": 9.873243847320833e-06, "loss": 0.5045, "step": 2378 }, { "epoch": 0.0997243851063161, "grad_norm": 2.1719186305999756, "learning_rate": 9.873091915706835e-06, "loss": 0.5953, "step": 2379 }, { "epoch": 0.09976630372132504, "grad_norm": 1.772848129272461, "learning_rate": 9.872939894264126e-06, "loss": 0.5191, "step": 2380 }, { "epoch": 0.099808222336334, "grad_norm": 1.9063613414764404, "learning_rate": 9.872787782995506e-06, "loss": 0.6016, "step": 2381 }, { "epoch": 0.09985014095134297, "grad_norm": 1.7446368932724, "learning_rate": 9.87263558190378e-06, "loss": 0.5346, "step": 2382 }, { "epoch": 0.09989205956635193, "grad_norm": 2.107907772064209, "learning_rate": 9.872483290991753e-06, "loss": 0.6367, "step": 2383 }, { "epoch": 0.09993397818136089, "grad_norm": 1.9601402282714844, "learning_rate": 9.872330910262231e-06, "loss": 0.5936, "step": 2384 }, { "epoch": 0.09997589679636985, "grad_norm": 2.7780463695526123, "learning_rate": 9.872178439718025e-06, "loss": 0.6332, "step": 2385 }, { "epoch": 0.1000178154113788, "grad_norm": 2.160614490509033, "learning_rate": 9.872025879361947e-06, "loss": 0.524, "step": 2386 }, { "epoch": 0.10005973402638776, "grad_norm": 1.973388671875, "learning_rate": 9.871873229196806e-06, "loss": 0.5778, "step": 2387 }, { "epoch": 0.10010165264139673, "grad_norm": 1.6408636569976807, "learning_rate": 9.871720489225417e-06, "loss": 0.5561, "step": 2388 }, { "epoch": 0.10014357125640569, "grad_norm": 2.1796915531158447, "learning_rate": 9.871567659450596e-06, "loss": 0.5995, "step": 2389 }, { "epoch": 0.10018548987141465, "grad_norm": 2.398256778717041, "learning_rate": 9.87141473987516e-06, "loss": 0.5563, "step": 2390 }, { "epoch": 0.10022740848642361, "grad_norm": 2.3784420490264893, "learning_rate": 9.871261730501927e-06, "loss": 0.5472, "step": 2391 }, { "epoch": 0.10026932710143256, "grad_norm": 2.12707781791687, "learning_rate": 9.871108631333718e-06, "loss": 0.5673, "step": 2392 }, { "epoch": 0.10031124571644152, "grad_norm": 1.9073359966278076, "learning_rate": 9.870955442373358e-06, "loss": 0.5416, "step": 2393 }, { "epoch": 0.10035316433145049, "grad_norm": 2.0812244415283203, "learning_rate": 9.870802163623668e-06, "loss": 0.6234, "step": 2394 }, { "epoch": 0.10039508294645945, "grad_norm": 1.9342825412750244, "learning_rate": 9.870648795087472e-06, "loss": 0.5837, "step": 2395 }, { "epoch": 0.10043700156146841, "grad_norm": 1.6948927640914917, "learning_rate": 9.870495336767599e-06, "loss": 0.5788, "step": 2396 }, { "epoch": 0.10047892017647737, "grad_norm": 2.141526937484741, "learning_rate": 9.870341788666878e-06, "loss": 0.6299, "step": 2397 }, { "epoch": 0.10052083879148634, "grad_norm": 1.651035189628601, "learning_rate": 9.870188150788139e-06, "loss": 0.571, "step": 2398 }, { "epoch": 0.10056275740649528, "grad_norm": 2.3758740425109863, "learning_rate": 9.870034423134214e-06, "loss": 0.5914, "step": 2399 }, { "epoch": 0.10060467602150425, "grad_norm": 1.8547762632369995, "learning_rate": 9.869880605707938e-06, "loss": 0.562, "step": 2400 }, { "epoch": 0.10064659463651321, "grad_norm": 2.0882296562194824, "learning_rate": 9.869726698512145e-06, "loss": 0.5948, "step": 2401 }, { "epoch": 0.10068851325152217, "grad_norm": 1.9541549682617188, "learning_rate": 9.869572701549673e-06, "loss": 0.5734, "step": 2402 }, { "epoch": 0.10073043186653113, "grad_norm": 1.8740531206130981, "learning_rate": 9.86941861482336e-06, "loss": 0.5765, "step": 2403 }, { "epoch": 0.1007723504815401, "grad_norm": 1.7259660959243774, "learning_rate": 9.869264438336046e-06, "loss": 0.5821, "step": 2404 }, { "epoch": 0.10081426909654904, "grad_norm": 1.9355937242507935, "learning_rate": 9.869110172090575e-06, "loss": 0.5223, "step": 2405 }, { "epoch": 0.100856187711558, "grad_norm": 1.7970449924468994, "learning_rate": 9.868955816089789e-06, "loss": 0.5869, "step": 2406 }, { "epoch": 0.10089810632656697, "grad_norm": 1.9428852796554565, "learning_rate": 9.868801370336532e-06, "loss": 0.6173, "step": 2407 }, { "epoch": 0.10094002494157593, "grad_norm": 1.6839606761932373, "learning_rate": 9.868646834833653e-06, "loss": 0.5849, "step": 2408 }, { "epoch": 0.1009819435565849, "grad_norm": 1.6507197618484497, "learning_rate": 9.868492209584002e-06, "loss": 0.6331, "step": 2409 }, { "epoch": 0.10102386217159386, "grad_norm": 1.6580015420913696, "learning_rate": 9.868337494590427e-06, "loss": 0.5575, "step": 2410 }, { "epoch": 0.1010657807866028, "grad_norm": 1.890857458114624, "learning_rate": 9.86818268985578e-06, "loss": 0.5898, "step": 2411 }, { "epoch": 0.10110769940161177, "grad_norm": 1.9480669498443604, "learning_rate": 9.868027795382914e-06, "loss": 0.6141, "step": 2412 }, { "epoch": 0.10114961801662073, "grad_norm": 1.9529714584350586, "learning_rate": 9.867872811174688e-06, "loss": 0.6011, "step": 2413 }, { "epoch": 0.10119153663162969, "grad_norm": 1.961777925491333, "learning_rate": 9.867717737233955e-06, "loss": 0.5238, "step": 2414 }, { "epoch": 0.10123345524663865, "grad_norm": 2.0105831623077393, "learning_rate": 9.867562573563576e-06, "loss": 0.6183, "step": 2415 }, { "epoch": 0.10127537386164762, "grad_norm": 2.142404079437256, "learning_rate": 9.867407320166408e-06, "loss": 0.6184, "step": 2416 }, { "epoch": 0.10131729247665656, "grad_norm": 2.14691424369812, "learning_rate": 9.867251977045317e-06, "loss": 0.5933, "step": 2417 }, { "epoch": 0.10135921109166553, "grad_norm": 2.162038803100586, "learning_rate": 9.867096544203164e-06, "loss": 0.6098, "step": 2418 }, { "epoch": 0.10140112970667449, "grad_norm": 1.633460283279419, "learning_rate": 9.866941021642816e-06, "loss": 0.567, "step": 2419 }, { "epoch": 0.10144304832168345, "grad_norm": 1.9750957489013672, "learning_rate": 9.866785409367136e-06, "loss": 0.6096, "step": 2420 }, { "epoch": 0.10148496693669241, "grad_norm": 2.034585952758789, "learning_rate": 9.866629707378996e-06, "loss": 0.6156, "step": 2421 }, { "epoch": 0.10152688555170138, "grad_norm": 1.7733750343322754, "learning_rate": 9.866473915681267e-06, "loss": 0.5426, "step": 2422 }, { "epoch": 0.10156880416671034, "grad_norm": 2.043238401412964, "learning_rate": 9.866318034276816e-06, "loss": 0.614, "step": 2423 }, { "epoch": 0.10161072278171929, "grad_norm": 1.8537938594818115, "learning_rate": 9.866162063168522e-06, "loss": 0.5532, "step": 2424 }, { "epoch": 0.10165264139672825, "grad_norm": 1.969340443611145, "learning_rate": 9.866006002359256e-06, "loss": 0.6156, "step": 2425 }, { "epoch": 0.10169456001173721, "grad_norm": 2.917506217956543, "learning_rate": 9.8658498518519e-06, "loss": 0.6011, "step": 2426 }, { "epoch": 0.10173647862674617, "grad_norm": 1.8489612340927124, "learning_rate": 9.865693611649324e-06, "loss": 0.5921, "step": 2427 }, { "epoch": 0.10177839724175514, "grad_norm": 2.1512818336486816, "learning_rate": 9.865537281754416e-06, "loss": 0.591, "step": 2428 }, { "epoch": 0.1018203158567641, "grad_norm": 1.7395250797271729, "learning_rate": 9.865380862170054e-06, "loss": 0.5827, "step": 2429 }, { "epoch": 0.10186223447177305, "grad_norm": 1.6445437669754028, "learning_rate": 9.86522435289912e-06, "loss": 0.5273, "step": 2430 }, { "epoch": 0.10190415308678201, "grad_norm": 3.3484339714050293, "learning_rate": 9.865067753944504e-06, "loss": 0.5814, "step": 2431 }, { "epoch": 0.10194607170179097, "grad_norm": 1.6750239133834839, "learning_rate": 9.864911065309086e-06, "loss": 0.5303, "step": 2432 }, { "epoch": 0.10198799031679993, "grad_norm": 1.6903232336044312, "learning_rate": 9.864754286995762e-06, "loss": 0.5956, "step": 2433 }, { "epoch": 0.1020299089318089, "grad_norm": 1.8834190368652344, "learning_rate": 9.864597419007416e-06, "loss": 0.5566, "step": 2434 }, { "epoch": 0.10207182754681786, "grad_norm": 1.7471137046813965, "learning_rate": 9.864440461346943e-06, "loss": 0.6109, "step": 2435 }, { "epoch": 0.10211374616182681, "grad_norm": 1.9476752281188965, "learning_rate": 9.864283414017233e-06, "loss": 0.5768, "step": 2436 }, { "epoch": 0.10215566477683577, "grad_norm": 1.6019541025161743, "learning_rate": 9.864126277021184e-06, "loss": 0.5242, "step": 2437 }, { "epoch": 0.10219758339184473, "grad_norm": 1.9540057182312012, "learning_rate": 9.863969050361692e-06, "loss": 0.6176, "step": 2438 }, { "epoch": 0.1022395020068537, "grad_norm": 1.9286506175994873, "learning_rate": 9.863811734041653e-06, "loss": 0.5813, "step": 2439 }, { "epoch": 0.10228142062186266, "grad_norm": 1.7520314455032349, "learning_rate": 9.863654328063971e-06, "loss": 0.6208, "step": 2440 }, { "epoch": 0.10232333923687162, "grad_norm": 1.7282453775405884, "learning_rate": 9.863496832431545e-06, "loss": 0.6143, "step": 2441 }, { "epoch": 0.10236525785188057, "grad_norm": 1.8782970905303955, "learning_rate": 9.863339247147276e-06, "loss": 0.6011, "step": 2442 }, { "epoch": 0.10240717646688953, "grad_norm": 1.7738603353500366, "learning_rate": 9.863181572214074e-06, "loss": 0.5991, "step": 2443 }, { "epoch": 0.10244909508189849, "grad_norm": 1.7766166925430298, "learning_rate": 9.863023807634841e-06, "loss": 0.574, "step": 2444 }, { "epoch": 0.10249101369690745, "grad_norm": 1.8296078443527222, "learning_rate": 9.862865953412489e-06, "loss": 0.5667, "step": 2445 }, { "epoch": 0.10253293231191642, "grad_norm": 2.038018226623535, "learning_rate": 9.862708009549924e-06, "loss": 0.6057, "step": 2446 }, { "epoch": 0.10257485092692538, "grad_norm": 1.8710365295410156, "learning_rate": 9.862549976050061e-06, "loss": 0.5919, "step": 2447 }, { "epoch": 0.10261676954193434, "grad_norm": 1.9051635265350342, "learning_rate": 9.86239185291581e-06, "loss": 0.5278, "step": 2448 }, { "epoch": 0.10265868815694329, "grad_norm": 2.268164873123169, "learning_rate": 9.862233640150089e-06, "loss": 0.6105, "step": 2449 }, { "epoch": 0.10270060677195225, "grad_norm": 2.8805928230285645, "learning_rate": 9.86207533775581e-06, "loss": 0.6476, "step": 2450 }, { "epoch": 0.10274252538696121, "grad_norm": 1.8878294229507446, "learning_rate": 9.861916945735896e-06, "loss": 0.5485, "step": 2451 }, { "epoch": 0.10278444400197018, "grad_norm": 2.0270376205444336, "learning_rate": 9.861758464093265e-06, "loss": 0.5735, "step": 2452 }, { "epoch": 0.10282636261697914, "grad_norm": 2.2127621173858643, "learning_rate": 9.861599892830836e-06, "loss": 0.6186, "step": 2453 }, { "epoch": 0.1028682812319881, "grad_norm": 2.1757428646087646, "learning_rate": 9.861441231951533e-06, "loss": 0.6183, "step": 2454 }, { "epoch": 0.10291019984699705, "grad_norm": 2.090806484222412, "learning_rate": 9.861282481458285e-06, "loss": 0.6448, "step": 2455 }, { "epoch": 0.10295211846200601, "grad_norm": 2.0660057067871094, "learning_rate": 9.861123641354014e-06, "loss": 0.6078, "step": 2456 }, { "epoch": 0.10299403707701497, "grad_norm": 2.462951421737671, "learning_rate": 9.860964711641647e-06, "loss": 0.5718, "step": 2457 }, { "epoch": 0.10303595569202394, "grad_norm": 2.8360118865966797, "learning_rate": 9.860805692324118e-06, "loss": 0.5935, "step": 2458 }, { "epoch": 0.1030778743070329, "grad_norm": 2.7460620403289795, "learning_rate": 9.860646583404353e-06, "loss": 0.5302, "step": 2459 }, { "epoch": 0.10311979292204186, "grad_norm": 2.119378089904785, "learning_rate": 9.86048738488529e-06, "loss": 0.643, "step": 2460 }, { "epoch": 0.10316171153705081, "grad_norm": 2.4395599365234375, "learning_rate": 9.860328096769863e-06, "loss": 0.6147, "step": 2461 }, { "epoch": 0.10320363015205977, "grad_norm": 1.9907822608947754, "learning_rate": 9.860168719061004e-06, "loss": 0.5884, "step": 2462 }, { "epoch": 0.10324554876706873, "grad_norm": 2.0366485118865967, "learning_rate": 9.860009251761655e-06, "loss": 0.5756, "step": 2463 }, { "epoch": 0.1032874673820777, "grad_norm": 1.789449691772461, "learning_rate": 9.859849694874753e-06, "loss": 0.553, "step": 2464 }, { "epoch": 0.10332938599708666, "grad_norm": 2.0312447547912598, "learning_rate": 9.859690048403241e-06, "loss": 0.5845, "step": 2465 }, { "epoch": 0.10337130461209562, "grad_norm": 1.9125018119812012, "learning_rate": 9.859530312350062e-06, "loss": 0.6101, "step": 2466 }, { "epoch": 0.10341322322710457, "grad_norm": 1.8752843141555786, "learning_rate": 9.859370486718158e-06, "loss": 0.5559, "step": 2467 }, { "epoch": 0.10345514184211353, "grad_norm": 1.9606612920761108, "learning_rate": 9.85921057151048e-06, "loss": 0.6245, "step": 2468 }, { "epoch": 0.1034970604571225, "grad_norm": 2.140011787414551, "learning_rate": 9.85905056672997e-06, "loss": 0.647, "step": 2469 }, { "epoch": 0.10353897907213146, "grad_norm": 1.7725157737731934, "learning_rate": 9.85889047237958e-06, "loss": 0.5702, "step": 2470 }, { "epoch": 0.10358089768714042, "grad_norm": 2.0551440715789795, "learning_rate": 9.858730288462264e-06, "loss": 0.6089, "step": 2471 }, { "epoch": 0.10362281630214938, "grad_norm": 2.3566930294036865, "learning_rate": 9.85857001498097e-06, "loss": 0.5956, "step": 2472 }, { "epoch": 0.10366473491715834, "grad_norm": 1.7630722522735596, "learning_rate": 9.858409651938656e-06, "loss": 0.6168, "step": 2473 }, { "epoch": 0.10370665353216729, "grad_norm": 1.745201587677002, "learning_rate": 9.858249199338276e-06, "loss": 0.5073, "step": 2474 }, { "epoch": 0.10374857214717625, "grad_norm": 1.623641848564148, "learning_rate": 9.858088657182788e-06, "loss": 0.5441, "step": 2475 }, { "epoch": 0.10379049076218522, "grad_norm": 1.8560844659805298, "learning_rate": 9.857928025475152e-06, "loss": 0.5428, "step": 2476 }, { "epoch": 0.10383240937719418, "grad_norm": 4.348439693450928, "learning_rate": 9.85776730421833e-06, "loss": 0.569, "step": 2477 }, { "epoch": 0.10387432799220314, "grad_norm": 1.7610012292861938, "learning_rate": 9.85760649341528e-06, "loss": 0.5474, "step": 2478 }, { "epoch": 0.1039162466072121, "grad_norm": 1.7988249063491821, "learning_rate": 9.857445593068973e-06, "loss": 0.6026, "step": 2479 }, { "epoch": 0.10395816522222105, "grad_norm": 1.5497711896896362, "learning_rate": 9.85728460318237e-06, "loss": 0.5474, "step": 2480 }, { "epoch": 0.10400008383723001, "grad_norm": 1.8236042261123657, "learning_rate": 9.85712352375844e-06, "loss": 0.5834, "step": 2481 }, { "epoch": 0.10404200245223898, "grad_norm": 1.8911840915679932, "learning_rate": 9.856962354800155e-06, "loss": 0.5499, "step": 2482 }, { "epoch": 0.10408392106724794, "grad_norm": 1.8986566066741943, "learning_rate": 9.856801096310482e-06, "loss": 0.6206, "step": 2483 }, { "epoch": 0.1041258396822569, "grad_norm": 1.7365261316299438, "learning_rate": 9.856639748292394e-06, "loss": 0.5544, "step": 2484 }, { "epoch": 0.10416775829726586, "grad_norm": 1.710150957107544, "learning_rate": 9.856478310748868e-06, "loss": 0.576, "step": 2485 }, { "epoch": 0.10420967691227481, "grad_norm": 2.289769172668457, "learning_rate": 9.856316783682878e-06, "loss": 0.5783, "step": 2486 }, { "epoch": 0.10425159552728377, "grad_norm": 2.1984739303588867, "learning_rate": 9.8561551670974e-06, "loss": 0.5939, "step": 2487 }, { "epoch": 0.10429351414229274, "grad_norm": 2.0632741451263428, "learning_rate": 9.855993460995416e-06, "loss": 0.594, "step": 2488 }, { "epoch": 0.1043354327573017, "grad_norm": 2.3552703857421875, "learning_rate": 9.855831665379908e-06, "loss": 0.6005, "step": 2489 }, { "epoch": 0.10437735137231066, "grad_norm": 1.9746352434158325, "learning_rate": 9.855669780253854e-06, "loss": 0.5844, "step": 2490 }, { "epoch": 0.10441926998731962, "grad_norm": 1.82988703250885, "learning_rate": 9.85550780562024e-06, "loss": 0.5633, "step": 2491 }, { "epoch": 0.10446118860232857, "grad_norm": 3.6512136459350586, "learning_rate": 9.855345741482053e-06, "loss": 0.6318, "step": 2492 }, { "epoch": 0.10450310721733753, "grad_norm": 2.1653902530670166, "learning_rate": 9.855183587842279e-06, "loss": 0.5932, "step": 2493 }, { "epoch": 0.1045450258323465, "grad_norm": 1.8068506717681885, "learning_rate": 9.855021344703907e-06, "loss": 0.5814, "step": 2494 }, { "epoch": 0.10458694444735546, "grad_norm": 1.8622801303863525, "learning_rate": 9.85485901206993e-06, "loss": 0.6179, "step": 2495 }, { "epoch": 0.10462886306236442, "grad_norm": 1.8341540098190308, "learning_rate": 9.854696589943338e-06, "loss": 0.5499, "step": 2496 }, { "epoch": 0.10467078167737338, "grad_norm": 1.9058786630630493, "learning_rate": 9.854534078327127e-06, "loss": 0.5786, "step": 2497 }, { "epoch": 0.10471270029238235, "grad_norm": 1.9446393251419067, "learning_rate": 9.854371477224289e-06, "loss": 0.5662, "step": 2498 }, { "epoch": 0.1047546189073913, "grad_norm": 1.8598612546920776, "learning_rate": 9.854208786637826e-06, "loss": 0.563, "step": 2499 }, { "epoch": 0.10479653752240026, "grad_norm": 1.9610663652420044, "learning_rate": 9.854046006570734e-06, "loss": 0.5963, "step": 2500 }, { "epoch": 0.10483845613740922, "grad_norm": 2.2074317932128906, "learning_rate": 9.853883137026015e-06, "loss": 0.5487, "step": 2501 }, { "epoch": 0.10488037475241818, "grad_norm": 1.9605549573898315, "learning_rate": 9.853720178006669e-06, "loss": 0.5935, "step": 2502 }, { "epoch": 0.10492229336742714, "grad_norm": 2.0673744678497314, "learning_rate": 9.853557129515702e-06, "loss": 0.548, "step": 2503 }, { "epoch": 0.1049642119824361, "grad_norm": 1.9527744054794312, "learning_rate": 9.853393991556121e-06, "loss": 0.5977, "step": 2504 }, { "epoch": 0.10500613059744505, "grad_norm": 1.7937383651733398, "learning_rate": 9.85323076413093e-06, "loss": 0.5556, "step": 2505 }, { "epoch": 0.10504804921245402, "grad_norm": 1.7701281309127808, "learning_rate": 9.853067447243139e-06, "loss": 0.5777, "step": 2506 }, { "epoch": 0.10508996782746298, "grad_norm": 1.8738356828689575, "learning_rate": 9.852904040895758e-06, "loss": 0.5871, "step": 2507 }, { "epoch": 0.10513188644247194, "grad_norm": 1.7646781206130981, "learning_rate": 9.852740545091803e-06, "loss": 0.5275, "step": 2508 }, { "epoch": 0.1051738050574809, "grad_norm": 1.783118724822998, "learning_rate": 9.852576959834282e-06, "loss": 0.5551, "step": 2509 }, { "epoch": 0.10521572367248987, "grad_norm": 1.9176993370056152, "learning_rate": 9.852413285126214e-06, "loss": 0.6166, "step": 2510 }, { "epoch": 0.10525764228749881, "grad_norm": 1.911655306816101, "learning_rate": 9.852249520970615e-06, "loss": 0.6131, "step": 2511 }, { "epoch": 0.10529956090250778, "grad_norm": 1.8402708768844604, "learning_rate": 9.852085667370506e-06, "loss": 0.6064, "step": 2512 }, { "epoch": 0.10534147951751674, "grad_norm": 1.7126442193984985, "learning_rate": 9.851921724328904e-06, "loss": 0.5918, "step": 2513 }, { "epoch": 0.1053833981325257, "grad_norm": 1.7882460355758667, "learning_rate": 9.851757691848832e-06, "loss": 0.5635, "step": 2514 }, { "epoch": 0.10542531674753466, "grad_norm": 2.0470659732818604, "learning_rate": 9.851593569933315e-06, "loss": 0.5259, "step": 2515 }, { "epoch": 0.10546723536254363, "grad_norm": 1.8752299547195435, "learning_rate": 9.851429358585379e-06, "loss": 0.576, "step": 2516 }, { "epoch": 0.10550915397755257, "grad_norm": 2.316476821899414, "learning_rate": 9.851265057808046e-06, "loss": 0.5245, "step": 2517 }, { "epoch": 0.10555107259256154, "grad_norm": 2.3399438858032227, "learning_rate": 9.851100667604352e-06, "loss": 0.543, "step": 2518 }, { "epoch": 0.1055929912075705, "grad_norm": 1.8722583055496216, "learning_rate": 9.850936187977322e-06, "loss": 0.5452, "step": 2519 }, { "epoch": 0.10563490982257946, "grad_norm": 1.7642625570297241, "learning_rate": 9.850771618929991e-06, "loss": 0.5636, "step": 2520 }, { "epoch": 0.10567682843758842, "grad_norm": 1.9475517272949219, "learning_rate": 9.85060696046539e-06, "loss": 0.622, "step": 2521 }, { "epoch": 0.10571874705259739, "grad_norm": 2.8699779510498047, "learning_rate": 9.850442212586555e-06, "loss": 0.5335, "step": 2522 }, { "epoch": 0.10576066566760635, "grad_norm": 2.1100714206695557, "learning_rate": 9.850277375296524e-06, "loss": 0.5922, "step": 2523 }, { "epoch": 0.1058025842826153, "grad_norm": 1.777492880821228, "learning_rate": 9.850112448598334e-06, "loss": 0.5639, "step": 2524 }, { "epoch": 0.10584450289762426, "grad_norm": 1.8994704484939575, "learning_rate": 9.849947432495027e-06, "loss": 0.53, "step": 2525 }, { "epoch": 0.10588642151263322, "grad_norm": 1.9182655811309814, "learning_rate": 9.849782326989643e-06, "loss": 0.5974, "step": 2526 }, { "epoch": 0.10592834012764218, "grad_norm": 1.792840600013733, "learning_rate": 9.849617132085228e-06, "loss": 0.5778, "step": 2527 }, { "epoch": 0.10597025874265115, "grad_norm": 1.8761168718338013, "learning_rate": 9.849451847784825e-06, "loss": 0.5923, "step": 2528 }, { "epoch": 0.10601217735766011, "grad_norm": 2.3646512031555176, "learning_rate": 9.849286474091482e-06, "loss": 0.5997, "step": 2529 }, { "epoch": 0.10605409597266906, "grad_norm": 1.9448399543762207, "learning_rate": 9.849121011008245e-06, "loss": 0.5519, "step": 2530 }, { "epoch": 0.10609601458767802, "grad_norm": 2.1624109745025635, "learning_rate": 9.848955458538168e-06, "loss": 0.6098, "step": 2531 }, { "epoch": 0.10613793320268698, "grad_norm": 1.8428140878677368, "learning_rate": 9.848789816684298e-06, "loss": 0.5818, "step": 2532 }, { "epoch": 0.10617985181769594, "grad_norm": 1.7552241086959839, "learning_rate": 9.848624085449692e-06, "loss": 0.612, "step": 2533 }, { "epoch": 0.1062217704327049, "grad_norm": 1.884216070175171, "learning_rate": 9.848458264837405e-06, "loss": 0.539, "step": 2534 }, { "epoch": 0.10626368904771387, "grad_norm": 2.0955076217651367, "learning_rate": 9.848292354850492e-06, "loss": 0.6053, "step": 2535 }, { "epoch": 0.10630560766272282, "grad_norm": 1.856886863708496, "learning_rate": 9.84812635549201e-06, "loss": 0.6354, "step": 2536 }, { "epoch": 0.10634752627773178, "grad_norm": 1.6434437036514282, "learning_rate": 9.847960266765024e-06, "loss": 0.5221, "step": 2537 }, { "epoch": 0.10638944489274074, "grad_norm": 2.1765661239624023, "learning_rate": 9.847794088672594e-06, "loss": 0.616, "step": 2538 }, { "epoch": 0.1064313635077497, "grad_norm": 1.8727631568908691, "learning_rate": 9.847627821217778e-06, "loss": 0.5694, "step": 2539 }, { "epoch": 0.10647328212275867, "grad_norm": 1.9063116312026978, "learning_rate": 9.847461464403647e-06, "loss": 0.5837, "step": 2540 }, { "epoch": 0.10651520073776763, "grad_norm": 2.0151731967926025, "learning_rate": 9.847295018233264e-06, "loss": 0.5606, "step": 2541 }, { "epoch": 0.10655711935277658, "grad_norm": 1.8770321607589722, "learning_rate": 9.847128482709698e-06, "loss": 0.5989, "step": 2542 }, { "epoch": 0.10659903796778554, "grad_norm": 2.4085662364959717, "learning_rate": 9.846961857836021e-06, "loss": 0.5028, "step": 2543 }, { "epoch": 0.1066409565827945, "grad_norm": 1.8255259990692139, "learning_rate": 9.846795143615302e-06, "loss": 0.5686, "step": 2544 }, { "epoch": 0.10668287519780346, "grad_norm": 2.82399845123291, "learning_rate": 9.846628340050618e-06, "loss": 0.5562, "step": 2545 }, { "epoch": 0.10672479381281243, "grad_norm": 3.4226810932159424, "learning_rate": 9.846461447145037e-06, "loss": 0.5852, "step": 2546 }, { "epoch": 0.10676671242782139, "grad_norm": 2.019225597381592, "learning_rate": 9.84629446490164e-06, "loss": 0.5942, "step": 2547 }, { "epoch": 0.10680863104283035, "grad_norm": 2.049131393432617, "learning_rate": 9.846127393323504e-06, "loss": 0.5443, "step": 2548 }, { "epoch": 0.1068505496578393, "grad_norm": 2.025181293487549, "learning_rate": 9.845960232413709e-06, "loss": 0.6091, "step": 2549 }, { "epoch": 0.10689246827284826, "grad_norm": 2.138775587081909, "learning_rate": 9.84579298217534e-06, "loss": 0.5836, "step": 2550 }, { "epoch": 0.10693438688785722, "grad_norm": 1.87035071849823, "learning_rate": 9.845625642611472e-06, "loss": 0.5514, "step": 2551 }, { "epoch": 0.10697630550286619, "grad_norm": 1.7897557020187378, "learning_rate": 9.845458213725196e-06, "loss": 0.5804, "step": 2552 }, { "epoch": 0.10701822411787515, "grad_norm": 2.187969923019409, "learning_rate": 9.845290695519596e-06, "loss": 0.5627, "step": 2553 }, { "epoch": 0.10706014273288411, "grad_norm": 1.949338436126709, "learning_rate": 9.845123087997761e-06, "loss": 0.6722, "step": 2554 }, { "epoch": 0.10710206134789306, "grad_norm": 1.669756531715393, "learning_rate": 9.84495539116278e-06, "loss": 0.5793, "step": 2555 }, { "epoch": 0.10714397996290202, "grad_norm": 1.8581353425979614, "learning_rate": 9.844787605017745e-06, "loss": 0.5775, "step": 2556 }, { "epoch": 0.10718589857791098, "grad_norm": 1.781664252281189, "learning_rate": 9.844619729565748e-06, "loss": 0.62, "step": 2557 }, { "epoch": 0.10722781719291995, "grad_norm": 2.184882640838623, "learning_rate": 9.844451764809882e-06, "loss": 0.6002, "step": 2558 }, { "epoch": 0.10726973580792891, "grad_norm": 1.914078950881958, "learning_rate": 9.844283710753246e-06, "loss": 0.6345, "step": 2559 }, { "epoch": 0.10731165442293787, "grad_norm": 2.074847459793091, "learning_rate": 9.844115567398937e-06, "loss": 0.6276, "step": 2560 }, { "epoch": 0.10735357303794682, "grad_norm": 2.1455204486846924, "learning_rate": 9.843947334750054e-06, "loss": 0.5884, "step": 2561 }, { "epoch": 0.10739549165295578, "grad_norm": 1.6859157085418701, "learning_rate": 9.8437790128097e-06, "loss": 0.5855, "step": 2562 }, { "epoch": 0.10743741026796474, "grad_norm": 1.80985689163208, "learning_rate": 9.843610601580974e-06, "loss": 0.5845, "step": 2563 }, { "epoch": 0.10747932888297371, "grad_norm": 1.9929709434509277, "learning_rate": 9.843442101066983e-06, "loss": 0.6041, "step": 2564 }, { "epoch": 0.10752124749798267, "grad_norm": 1.8622537851333618, "learning_rate": 9.843273511270834e-06, "loss": 0.5482, "step": 2565 }, { "epoch": 0.10756316611299163, "grad_norm": 1.8385565280914307, "learning_rate": 9.843104832195632e-06, "loss": 0.5889, "step": 2566 }, { "epoch": 0.10760508472800058, "grad_norm": 2.056511402130127, "learning_rate": 9.842936063844489e-06, "loss": 0.6151, "step": 2567 }, { "epoch": 0.10764700334300954, "grad_norm": 1.8452346324920654, "learning_rate": 9.842767206220515e-06, "loss": 0.6274, "step": 2568 }, { "epoch": 0.1076889219580185, "grad_norm": 1.623795509338379, "learning_rate": 9.84259825932682e-06, "loss": 0.5646, "step": 2569 }, { "epoch": 0.10773084057302747, "grad_norm": 1.735152244567871, "learning_rate": 9.842429223166524e-06, "loss": 0.5314, "step": 2570 }, { "epoch": 0.10777275918803643, "grad_norm": 2.059155225753784, "learning_rate": 9.842260097742738e-06, "loss": 0.5685, "step": 2571 }, { "epoch": 0.10781467780304539, "grad_norm": 1.8209220170974731, "learning_rate": 9.842090883058581e-06, "loss": 0.636, "step": 2572 }, { "epoch": 0.10785659641805435, "grad_norm": 1.8744640350341797, "learning_rate": 9.841921579117175e-06, "loss": 0.5576, "step": 2573 }, { "epoch": 0.1078985150330633, "grad_norm": 1.9608923196792603, "learning_rate": 9.841752185921636e-06, "loss": 0.6301, "step": 2574 }, { "epoch": 0.10794043364807226, "grad_norm": 2.1463868618011475, "learning_rate": 9.84158270347509e-06, "loss": 0.5874, "step": 2575 }, { "epoch": 0.10798235226308123, "grad_norm": 2.285611152648926, "learning_rate": 9.84141313178066e-06, "loss": 0.6007, "step": 2576 }, { "epoch": 0.10802427087809019, "grad_norm": 1.8609917163848877, "learning_rate": 9.841243470841472e-06, "loss": 0.6135, "step": 2577 }, { "epoch": 0.10806618949309915, "grad_norm": 1.9220952987670898, "learning_rate": 9.841073720660652e-06, "loss": 0.6249, "step": 2578 }, { "epoch": 0.10810810810810811, "grad_norm": 2.200230598449707, "learning_rate": 9.840903881241332e-06, "loss": 0.5443, "step": 2579 }, { "epoch": 0.10815002672311706, "grad_norm": 1.8803908824920654, "learning_rate": 9.840733952586641e-06, "loss": 0.5713, "step": 2580 }, { "epoch": 0.10819194533812602, "grad_norm": 1.7676613330841064, "learning_rate": 9.840563934699711e-06, "loss": 0.5306, "step": 2581 }, { "epoch": 0.10823386395313499, "grad_norm": 1.922881841659546, "learning_rate": 9.840393827583679e-06, "loss": 0.6034, "step": 2582 }, { "epoch": 0.10827578256814395, "grad_norm": 2.3646464347839355, "learning_rate": 9.840223631241677e-06, "loss": 0.5749, "step": 2583 }, { "epoch": 0.10831770118315291, "grad_norm": 2.0960071086883545, "learning_rate": 9.840053345676843e-06, "loss": 0.5986, "step": 2584 }, { "epoch": 0.10835961979816187, "grad_norm": 1.7539423704147339, "learning_rate": 9.839882970892318e-06, "loss": 0.5875, "step": 2585 }, { "epoch": 0.10840153841317082, "grad_norm": 1.6137863397598267, "learning_rate": 9.83971250689124e-06, "loss": 0.5293, "step": 2586 }, { "epoch": 0.10844345702817978, "grad_norm": 1.8158316612243652, "learning_rate": 9.839541953676755e-06, "loss": 0.5979, "step": 2587 }, { "epoch": 0.10848537564318875, "grad_norm": 1.700153112411499, "learning_rate": 9.839371311252002e-06, "loss": 0.5806, "step": 2588 }, { "epoch": 0.10852729425819771, "grad_norm": 2.2595534324645996, "learning_rate": 9.83920057962013e-06, "loss": 0.5725, "step": 2589 }, { "epoch": 0.10856921287320667, "grad_norm": 1.6989885568618774, "learning_rate": 9.839029758784287e-06, "loss": 0.5679, "step": 2590 }, { "epoch": 0.10861113148821563, "grad_norm": 1.771494746208191, "learning_rate": 9.838858848747616e-06, "loss": 0.5316, "step": 2591 }, { "epoch": 0.10865305010322458, "grad_norm": 1.6690858602523804, "learning_rate": 9.838687849513275e-06, "loss": 0.592, "step": 2592 }, { "epoch": 0.10869496871823355, "grad_norm": 1.83194899559021, "learning_rate": 9.838516761084412e-06, "loss": 0.6178, "step": 2593 }, { "epoch": 0.10873688733324251, "grad_norm": 2.6726412773132324, "learning_rate": 9.838345583464182e-06, "loss": 0.6102, "step": 2594 }, { "epoch": 0.10877880594825147, "grad_norm": 1.9783722162246704, "learning_rate": 9.83817431665574e-06, "loss": 0.613, "step": 2595 }, { "epoch": 0.10882072456326043, "grad_norm": 1.9643325805664062, "learning_rate": 9.838002960662243e-06, "loss": 0.5852, "step": 2596 }, { "epoch": 0.1088626431782694, "grad_norm": 1.6394917964935303, "learning_rate": 9.83783151548685e-06, "loss": 0.6087, "step": 2597 }, { "epoch": 0.10890456179327836, "grad_norm": 1.8682359457015991, "learning_rate": 9.837659981132718e-06, "loss": 0.6095, "step": 2598 }, { "epoch": 0.1089464804082873, "grad_norm": 1.9601690769195557, "learning_rate": 9.837488357603015e-06, "loss": 0.5412, "step": 2599 }, { "epoch": 0.10898839902329627, "grad_norm": 1.9817092418670654, "learning_rate": 9.837316644900901e-06, "loss": 0.632, "step": 2600 }, { "epoch": 0.10903031763830523, "grad_norm": 1.8365153074264526, "learning_rate": 9.837144843029542e-06, "loss": 0.5895, "step": 2601 }, { "epoch": 0.10907223625331419, "grad_norm": 4.024829387664795, "learning_rate": 9.836972951992106e-06, "loss": 0.5748, "step": 2602 }, { "epoch": 0.10911415486832315, "grad_norm": 2.8050730228424072, "learning_rate": 9.83680097179176e-06, "loss": 0.5731, "step": 2603 }, { "epoch": 0.10915607348333212, "grad_norm": 2.160048007965088, "learning_rate": 9.836628902431675e-06, "loss": 0.5582, "step": 2604 }, { "epoch": 0.10919799209834107, "grad_norm": 2.0686330795288086, "learning_rate": 9.836456743915022e-06, "loss": 0.5642, "step": 2605 }, { "epoch": 0.10923991071335003, "grad_norm": 1.7639840841293335, "learning_rate": 9.836284496244976e-06, "loss": 0.6054, "step": 2606 }, { "epoch": 0.10928182932835899, "grad_norm": 1.9652316570281982, "learning_rate": 9.836112159424711e-06, "loss": 0.5883, "step": 2607 }, { "epoch": 0.10932374794336795, "grad_norm": 1.7996596097946167, "learning_rate": 9.835939733457403e-06, "loss": 0.5618, "step": 2608 }, { "epoch": 0.10936566655837691, "grad_norm": 1.7851617336273193, "learning_rate": 9.835767218346233e-06, "loss": 0.5799, "step": 2609 }, { "epoch": 0.10940758517338588, "grad_norm": 1.792641043663025, "learning_rate": 9.835594614094379e-06, "loss": 0.5592, "step": 2610 }, { "epoch": 0.10944950378839483, "grad_norm": 1.7845426797866821, "learning_rate": 9.835421920705022e-06, "loss": 0.5573, "step": 2611 }, { "epoch": 0.10949142240340379, "grad_norm": 2.061244010925293, "learning_rate": 9.83524913818135e-06, "loss": 0.5747, "step": 2612 }, { "epoch": 0.10953334101841275, "grad_norm": 2.0017409324645996, "learning_rate": 9.83507626652654e-06, "loss": 0.5823, "step": 2613 }, { "epoch": 0.10957525963342171, "grad_norm": 2.112534761428833, "learning_rate": 9.834903305743786e-06, "loss": 0.5432, "step": 2614 }, { "epoch": 0.10961717824843067, "grad_norm": 1.929187536239624, "learning_rate": 9.834730255836274e-06, "loss": 0.6254, "step": 2615 }, { "epoch": 0.10965909686343964, "grad_norm": 1.8898552656173706, "learning_rate": 9.834557116807193e-06, "loss": 0.598, "step": 2616 }, { "epoch": 0.10970101547844859, "grad_norm": 1.9713921546936035, "learning_rate": 9.834383888659735e-06, "loss": 0.6175, "step": 2617 }, { "epoch": 0.10974293409345755, "grad_norm": 1.8056122064590454, "learning_rate": 9.834210571397093e-06, "loss": 0.633, "step": 2618 }, { "epoch": 0.10978485270846651, "grad_norm": 1.6442679166793823, "learning_rate": 9.834037165022463e-06, "loss": 0.5943, "step": 2619 }, { "epoch": 0.10982677132347547, "grad_norm": 1.7210742235183716, "learning_rate": 9.83386366953904e-06, "loss": 0.5758, "step": 2620 }, { "epoch": 0.10986868993848443, "grad_norm": 1.8585543632507324, "learning_rate": 9.833690084950023e-06, "loss": 0.5635, "step": 2621 }, { "epoch": 0.1099106085534934, "grad_norm": 1.611649990081787, "learning_rate": 9.833516411258613e-06, "loss": 0.5572, "step": 2622 }, { "epoch": 0.10995252716850236, "grad_norm": 1.77712881565094, "learning_rate": 9.833342648468007e-06, "loss": 0.5912, "step": 2623 }, { "epoch": 0.10999444578351131, "grad_norm": 4.447546005249023, "learning_rate": 9.833168796581414e-06, "loss": 0.58, "step": 2624 }, { "epoch": 0.11003636439852027, "grad_norm": 2.0944578647613525, "learning_rate": 9.832994855602035e-06, "loss": 0.5315, "step": 2625 }, { "epoch": 0.11007828301352923, "grad_norm": 1.7626007795333862, "learning_rate": 9.832820825533077e-06, "loss": 0.5957, "step": 2626 }, { "epoch": 0.1101202016285382, "grad_norm": 1.6122740507125854, "learning_rate": 9.832646706377748e-06, "loss": 0.5733, "step": 2627 }, { "epoch": 0.11016212024354716, "grad_norm": 1.851623296737671, "learning_rate": 9.83247249813926e-06, "loss": 0.5877, "step": 2628 }, { "epoch": 0.11020403885855612, "grad_norm": 2.181490182876587, "learning_rate": 9.83229820082082e-06, "loss": 0.5623, "step": 2629 }, { "epoch": 0.11024595747356507, "grad_norm": 2.278165102005005, "learning_rate": 9.832123814425643e-06, "loss": 0.561, "step": 2630 }, { "epoch": 0.11028787608857403, "grad_norm": 2.0535686016082764, "learning_rate": 9.831949338956945e-06, "loss": 0.5745, "step": 2631 }, { "epoch": 0.11032979470358299, "grad_norm": 2.027230739593506, "learning_rate": 9.831774774417941e-06, "loss": 0.5846, "step": 2632 }, { "epoch": 0.11037171331859195, "grad_norm": 1.9407514333724976, "learning_rate": 9.831600120811848e-06, "loss": 0.5834, "step": 2633 }, { "epoch": 0.11041363193360092, "grad_norm": 1.73495352268219, "learning_rate": 9.831425378141885e-06, "loss": 0.5898, "step": 2634 }, { "epoch": 0.11045555054860988, "grad_norm": 1.791089415550232, "learning_rate": 9.831250546411276e-06, "loss": 0.5776, "step": 2635 }, { "epoch": 0.11049746916361883, "grad_norm": 1.7714691162109375, "learning_rate": 9.831075625623242e-06, "loss": 0.5636, "step": 2636 }, { "epoch": 0.11053938777862779, "grad_norm": 1.8992176055908203, "learning_rate": 9.830900615781007e-06, "loss": 0.5263, "step": 2637 }, { "epoch": 0.11058130639363675, "grad_norm": 2.1019535064697266, "learning_rate": 9.8307255168878e-06, "loss": 0.5873, "step": 2638 }, { "epoch": 0.11062322500864571, "grad_norm": 1.6571499109268188, "learning_rate": 9.830550328946845e-06, "loss": 0.53, "step": 2639 }, { "epoch": 0.11066514362365468, "grad_norm": 1.812100887298584, "learning_rate": 9.830375051961372e-06, "loss": 0.6178, "step": 2640 }, { "epoch": 0.11070706223866364, "grad_norm": 1.7793570756912231, "learning_rate": 9.830199685934611e-06, "loss": 0.5971, "step": 2641 }, { "epoch": 0.11074898085367259, "grad_norm": 1.7411296367645264, "learning_rate": 9.830024230869799e-06, "loss": 0.581, "step": 2642 }, { "epoch": 0.11079089946868155, "grad_norm": 1.9747443199157715, "learning_rate": 9.829848686770167e-06, "loss": 0.5729, "step": 2643 }, { "epoch": 0.11083281808369051, "grad_norm": 1.9644166231155396, "learning_rate": 9.829673053638952e-06, "loss": 0.5525, "step": 2644 }, { "epoch": 0.11087473669869947, "grad_norm": 1.6376724243164062, "learning_rate": 9.82949733147939e-06, "loss": 0.4949, "step": 2645 }, { "epoch": 0.11091665531370844, "grad_norm": 1.9437086582183838, "learning_rate": 9.829321520294723e-06, "loss": 0.5484, "step": 2646 }, { "epoch": 0.1109585739287174, "grad_norm": 2.120042085647583, "learning_rate": 9.829145620088188e-06, "loss": 0.5722, "step": 2647 }, { "epoch": 0.11100049254372636, "grad_norm": 1.877417802810669, "learning_rate": 9.82896963086303e-06, "loss": 0.5603, "step": 2648 }, { "epoch": 0.11104241115873531, "grad_norm": 1.9722557067871094, "learning_rate": 9.828793552622493e-06, "loss": 0.5566, "step": 2649 }, { "epoch": 0.11108432977374427, "grad_norm": 2.23000431060791, "learning_rate": 9.828617385369822e-06, "loss": 0.5251, "step": 2650 }, { "epoch": 0.11112624838875323, "grad_norm": 1.912589192390442, "learning_rate": 9.828441129108266e-06, "loss": 0.5314, "step": 2651 }, { "epoch": 0.1111681670037622, "grad_norm": 1.9917210340499878, "learning_rate": 9.828264783841073e-06, "loss": 0.6618, "step": 2652 }, { "epoch": 0.11121008561877116, "grad_norm": 1.7832772731781006, "learning_rate": 9.828088349571494e-06, "loss": 0.5886, "step": 2653 }, { "epoch": 0.11125200423378012, "grad_norm": 2.1418375968933105, "learning_rate": 9.82791182630278e-06, "loss": 0.5754, "step": 2654 }, { "epoch": 0.11129392284878907, "grad_norm": 1.8458868265151978, "learning_rate": 9.827735214038185e-06, "loss": 0.612, "step": 2655 }, { "epoch": 0.11133584146379803, "grad_norm": 1.9249933958053589, "learning_rate": 9.827558512780965e-06, "loss": 0.5736, "step": 2656 }, { "epoch": 0.111377760078807, "grad_norm": 1.820373296737671, "learning_rate": 9.827381722534381e-06, "loss": 0.5447, "step": 2657 }, { "epoch": 0.11141967869381596, "grad_norm": 2.1119213104248047, "learning_rate": 9.827204843301686e-06, "loss": 0.6152, "step": 2658 }, { "epoch": 0.11146159730882492, "grad_norm": 2.2469592094421387, "learning_rate": 9.827027875086143e-06, "loss": 0.5684, "step": 2659 }, { "epoch": 0.11150351592383388, "grad_norm": 1.5827924013137817, "learning_rate": 9.826850817891016e-06, "loss": 0.5317, "step": 2660 }, { "epoch": 0.11154543453884283, "grad_norm": 1.810116171836853, "learning_rate": 9.826673671719567e-06, "loss": 0.5707, "step": 2661 }, { "epoch": 0.11158735315385179, "grad_norm": 2.9625871181488037, "learning_rate": 9.82649643657506e-06, "loss": 0.6137, "step": 2662 }, { "epoch": 0.11162927176886075, "grad_norm": 2.4159061908721924, "learning_rate": 9.826319112460764e-06, "loss": 0.5522, "step": 2663 }, { "epoch": 0.11167119038386972, "grad_norm": 2.24053692817688, "learning_rate": 9.826141699379947e-06, "loss": 0.5719, "step": 2664 }, { "epoch": 0.11171310899887868, "grad_norm": 1.970724105834961, "learning_rate": 9.825964197335882e-06, "loss": 0.6279, "step": 2665 }, { "epoch": 0.11175502761388764, "grad_norm": 1.700913667678833, "learning_rate": 9.825786606331837e-06, "loss": 0.559, "step": 2666 }, { "epoch": 0.11179694622889659, "grad_norm": 1.603604793548584, "learning_rate": 9.825608926371086e-06, "loss": 0.5221, "step": 2667 }, { "epoch": 0.11183886484390555, "grad_norm": 1.767546534538269, "learning_rate": 9.825431157456908e-06, "loss": 0.5294, "step": 2668 }, { "epoch": 0.11188078345891452, "grad_norm": 6.144567966461182, "learning_rate": 9.825253299592576e-06, "loss": 0.5342, "step": 2669 }, { "epoch": 0.11192270207392348, "grad_norm": 1.8695625066757202, "learning_rate": 9.82507535278137e-06, "loss": 0.514, "step": 2670 }, { "epoch": 0.11196462068893244, "grad_norm": 1.9687880277633667, "learning_rate": 9.824897317026571e-06, "loss": 0.6418, "step": 2671 }, { "epoch": 0.1120065393039414, "grad_norm": 1.7120336294174194, "learning_rate": 9.824719192331463e-06, "loss": 0.536, "step": 2672 }, { "epoch": 0.11204845791895036, "grad_norm": 1.8142359256744385, "learning_rate": 9.824540978699321e-06, "loss": 0.5541, "step": 2673 }, { "epoch": 0.11209037653395931, "grad_norm": 1.736407995223999, "learning_rate": 9.82436267613344e-06, "loss": 0.5451, "step": 2674 }, { "epoch": 0.11213229514896828, "grad_norm": 2.0893399715423584, "learning_rate": 9.824184284637102e-06, "loss": 0.6254, "step": 2675 }, { "epoch": 0.11217421376397724, "grad_norm": 1.7639325857162476, "learning_rate": 9.824005804213595e-06, "loss": 0.6211, "step": 2676 }, { "epoch": 0.1122161323789862, "grad_norm": 1.7292104959487915, "learning_rate": 9.82382723486621e-06, "loss": 0.5623, "step": 2677 }, { "epoch": 0.11225805099399516, "grad_norm": 1.872884750366211, "learning_rate": 9.82364857659824e-06, "loss": 0.5974, "step": 2678 }, { "epoch": 0.11229996960900412, "grad_norm": 1.8048725128173828, "learning_rate": 9.823469829412975e-06, "loss": 0.5362, "step": 2679 }, { "epoch": 0.11234188822401307, "grad_norm": 1.93388831615448, "learning_rate": 9.823290993313712e-06, "loss": 0.5989, "step": 2680 }, { "epoch": 0.11238380683902204, "grad_norm": 1.7525447607040405, "learning_rate": 9.823112068303748e-06, "loss": 0.6114, "step": 2681 }, { "epoch": 0.112425725454031, "grad_norm": 1.7815871238708496, "learning_rate": 9.822933054386382e-06, "loss": 0.5928, "step": 2682 }, { "epoch": 0.11246764406903996, "grad_norm": 1.6664369106292725, "learning_rate": 9.822753951564912e-06, "loss": 0.5735, "step": 2683 }, { "epoch": 0.11250956268404892, "grad_norm": 1.716302514076233, "learning_rate": 9.82257475984264e-06, "loss": 0.6393, "step": 2684 }, { "epoch": 0.11255148129905788, "grad_norm": 1.846157193183899, "learning_rate": 9.822395479222868e-06, "loss": 0.6622, "step": 2685 }, { "epoch": 0.11259339991406683, "grad_norm": 2.773103713989258, "learning_rate": 9.822216109708904e-06, "loss": 0.6026, "step": 2686 }, { "epoch": 0.1126353185290758, "grad_norm": 1.6783050298690796, "learning_rate": 9.822036651304052e-06, "loss": 0.5892, "step": 2687 }, { "epoch": 0.11267723714408476, "grad_norm": 1.7015141248703003, "learning_rate": 9.82185710401162e-06, "loss": 0.5819, "step": 2688 }, { "epoch": 0.11271915575909372, "grad_norm": 1.883246898651123, "learning_rate": 9.82167746783492e-06, "loss": 0.5274, "step": 2689 }, { "epoch": 0.11276107437410268, "grad_norm": 1.79868483543396, "learning_rate": 9.82149774277726e-06, "loss": 0.5594, "step": 2690 }, { "epoch": 0.11280299298911164, "grad_norm": 2.3258867263793945, "learning_rate": 9.821317928841954e-06, "loss": 0.5974, "step": 2691 }, { "epoch": 0.11284491160412059, "grad_norm": 1.9987949132919312, "learning_rate": 9.82113802603232e-06, "loss": 0.664, "step": 2692 }, { "epoch": 0.11288683021912956, "grad_norm": 1.9526396989822388, "learning_rate": 9.820958034351668e-06, "loss": 0.649, "step": 2693 }, { "epoch": 0.11292874883413852, "grad_norm": 1.6725151538848877, "learning_rate": 9.820777953803322e-06, "loss": 0.5352, "step": 2694 }, { "epoch": 0.11297066744914748, "grad_norm": 1.8732484579086304, "learning_rate": 9.820597784390599e-06, "loss": 0.6153, "step": 2695 }, { "epoch": 0.11301258606415644, "grad_norm": 1.60910165309906, "learning_rate": 9.820417526116819e-06, "loss": 0.5003, "step": 2696 }, { "epoch": 0.1130545046791654, "grad_norm": 1.6636868715286255, "learning_rate": 9.820237178985306e-06, "loss": 0.5824, "step": 2697 }, { "epoch": 0.11309642329417437, "grad_norm": 2.063746452331543, "learning_rate": 9.820056742999384e-06, "loss": 0.5883, "step": 2698 }, { "epoch": 0.11313834190918332, "grad_norm": 1.8422504663467407, "learning_rate": 9.81987621816238e-06, "loss": 0.5995, "step": 2699 }, { "epoch": 0.11318026052419228, "grad_norm": 1.7153172492980957, "learning_rate": 9.819695604477622e-06, "loss": 0.5999, "step": 2700 }, { "epoch": 0.11322217913920124, "grad_norm": 2.203432321548462, "learning_rate": 9.819514901948438e-06, "loss": 0.6044, "step": 2701 }, { "epoch": 0.1132640977542102, "grad_norm": 2.9176104068756104, "learning_rate": 9.81933411057816e-06, "loss": 0.5508, "step": 2702 }, { "epoch": 0.11330601636921916, "grad_norm": 1.937311053276062, "learning_rate": 9.81915323037012e-06, "loss": 0.608, "step": 2703 }, { "epoch": 0.11334793498422813, "grad_norm": 1.6777985095977783, "learning_rate": 9.81897226132765e-06, "loss": 0.5502, "step": 2704 }, { "epoch": 0.11338985359923708, "grad_norm": 1.723854422569275, "learning_rate": 9.818791203454092e-06, "loss": 0.5494, "step": 2705 }, { "epoch": 0.11343177221424604, "grad_norm": 3.401470184326172, "learning_rate": 9.818610056752778e-06, "loss": 0.6257, "step": 2706 }, { "epoch": 0.113473690829255, "grad_norm": 2.782719373703003, "learning_rate": 9.818428821227049e-06, "loss": 0.59, "step": 2707 }, { "epoch": 0.11351560944426396, "grad_norm": 1.852644681930542, "learning_rate": 9.818247496880246e-06, "loss": 0.5455, "step": 2708 }, { "epoch": 0.11355752805927292, "grad_norm": 4.365888595581055, "learning_rate": 9.818066083715714e-06, "loss": 0.6258, "step": 2709 }, { "epoch": 0.11359944667428189, "grad_norm": 1.8290091753005981, "learning_rate": 9.817884581736791e-06, "loss": 0.5578, "step": 2710 }, { "epoch": 0.11364136528929084, "grad_norm": 1.7986624240875244, "learning_rate": 9.81770299094683e-06, "loss": 0.5919, "step": 2711 }, { "epoch": 0.1136832839042998, "grad_norm": 2.187286376953125, "learning_rate": 9.81752131134917e-06, "loss": 0.6159, "step": 2712 }, { "epoch": 0.11372520251930876, "grad_norm": 1.646606206893921, "learning_rate": 9.817339542947169e-06, "loss": 0.6132, "step": 2713 }, { "epoch": 0.11376712113431772, "grad_norm": 1.6533979177474976, "learning_rate": 9.81715768574417e-06, "loss": 0.5625, "step": 2714 }, { "epoch": 0.11380903974932668, "grad_norm": 2.0618889331817627, "learning_rate": 9.81697573974353e-06, "loss": 0.6024, "step": 2715 }, { "epoch": 0.11385095836433565, "grad_norm": 1.9038589000701904, "learning_rate": 9.816793704948601e-06, "loss": 0.6419, "step": 2716 }, { "epoch": 0.1138928769793446, "grad_norm": 1.6556603908538818, "learning_rate": 9.81661158136274e-06, "loss": 0.6146, "step": 2717 }, { "epoch": 0.11393479559435356, "grad_norm": 1.6577174663543701, "learning_rate": 9.816429368989303e-06, "loss": 0.5001, "step": 2718 }, { "epoch": 0.11397671420936252, "grad_norm": 1.6526949405670166, "learning_rate": 9.816247067831648e-06, "loss": 0.5452, "step": 2719 }, { "epoch": 0.11401863282437148, "grad_norm": 1.7919999361038208, "learning_rate": 9.816064677893138e-06, "loss": 0.5631, "step": 2720 }, { "epoch": 0.11406055143938044, "grad_norm": 2.248457193374634, "learning_rate": 9.815882199177134e-06, "loss": 0.5922, "step": 2721 }, { "epoch": 0.11410247005438941, "grad_norm": 1.7186580896377563, "learning_rate": 9.815699631686996e-06, "loss": 0.59, "step": 2722 }, { "epoch": 0.11414438866939837, "grad_norm": 1.9158979654312134, "learning_rate": 9.815516975426096e-06, "loss": 0.6034, "step": 2723 }, { "epoch": 0.11418630728440732, "grad_norm": 1.6305612325668335, "learning_rate": 9.815334230397797e-06, "loss": 0.5797, "step": 2724 }, { "epoch": 0.11422822589941628, "grad_norm": 1.9002907276153564, "learning_rate": 9.81515139660547e-06, "loss": 0.5729, "step": 2725 }, { "epoch": 0.11427014451442524, "grad_norm": 1.827582597732544, "learning_rate": 9.814968474052483e-06, "loss": 0.58, "step": 2726 }, { "epoch": 0.1143120631294342, "grad_norm": 1.9173872470855713, "learning_rate": 9.814785462742207e-06, "loss": 0.5728, "step": 2727 }, { "epoch": 0.11435398174444317, "grad_norm": 1.8523834943771362, "learning_rate": 9.814602362678018e-06, "loss": 0.5328, "step": 2728 }, { "epoch": 0.11439590035945213, "grad_norm": 1.8472681045532227, "learning_rate": 9.81441917386329e-06, "loss": 0.5565, "step": 2729 }, { "epoch": 0.11443781897446108, "grad_norm": 1.6147063970565796, "learning_rate": 9.8142358963014e-06, "loss": 0.5692, "step": 2730 }, { "epoch": 0.11447973758947004, "grad_norm": 2.0448901653289795, "learning_rate": 9.814052529995729e-06, "loss": 0.627, "step": 2731 }, { "epoch": 0.114521656204479, "grad_norm": 2.154881238937378, "learning_rate": 9.813869074949654e-06, "loss": 0.6243, "step": 2732 }, { "epoch": 0.11456357481948796, "grad_norm": 1.9720797538757324, "learning_rate": 9.813685531166556e-06, "loss": 0.6102, "step": 2733 }, { "epoch": 0.11460549343449693, "grad_norm": 1.8903272151947021, "learning_rate": 9.813501898649822e-06, "loss": 0.5954, "step": 2734 }, { "epoch": 0.11464741204950589, "grad_norm": 2.044787883758545, "learning_rate": 9.813318177402834e-06, "loss": 0.6154, "step": 2735 }, { "epoch": 0.11468933066451484, "grad_norm": 2.1575098037719727, "learning_rate": 9.81313436742898e-06, "loss": 0.5737, "step": 2736 }, { "epoch": 0.1147312492795238, "grad_norm": 1.7406102418899536, "learning_rate": 9.812950468731647e-06, "loss": 0.5649, "step": 2737 }, { "epoch": 0.11477316789453276, "grad_norm": 1.6422377824783325, "learning_rate": 9.812766481314227e-06, "loss": 0.5867, "step": 2738 }, { "epoch": 0.11481508650954173, "grad_norm": 1.9804058074951172, "learning_rate": 9.812582405180108e-06, "loss": 0.5661, "step": 2739 }, { "epoch": 0.11485700512455069, "grad_norm": 2.040597438812256, "learning_rate": 9.812398240332687e-06, "loss": 0.5735, "step": 2740 }, { "epoch": 0.11489892373955965, "grad_norm": 1.9066914319992065, "learning_rate": 9.812213986775358e-06, "loss": 0.6, "step": 2741 }, { "epoch": 0.11494084235456861, "grad_norm": 1.8997405767440796, "learning_rate": 9.812029644511518e-06, "loss": 0.565, "step": 2742 }, { "epoch": 0.11498276096957756, "grad_norm": 1.7038018703460693, "learning_rate": 9.811845213544561e-06, "loss": 0.5782, "step": 2743 }, { "epoch": 0.11502467958458652, "grad_norm": 1.6527129411697388, "learning_rate": 9.811660693877891e-06, "loss": 0.5687, "step": 2744 }, { "epoch": 0.11506659819959549, "grad_norm": 2.030715227127075, "learning_rate": 9.811476085514908e-06, "loss": 0.5632, "step": 2745 }, { "epoch": 0.11510851681460445, "grad_norm": 1.9198111295700073, "learning_rate": 9.811291388459016e-06, "loss": 0.5613, "step": 2746 }, { "epoch": 0.11515043542961341, "grad_norm": 1.8892980813980103, "learning_rate": 9.811106602713617e-06, "loss": 0.6184, "step": 2747 }, { "epoch": 0.11519235404462237, "grad_norm": 1.6812610626220703, "learning_rate": 9.81092172828212e-06, "loss": 0.5836, "step": 2748 }, { "epoch": 0.11523427265963132, "grad_norm": 2.1067638397216797, "learning_rate": 9.810736765167932e-06, "loss": 0.5783, "step": 2749 }, { "epoch": 0.11527619127464028, "grad_norm": 1.812641978263855, "learning_rate": 9.810551713374462e-06, "loss": 0.5951, "step": 2750 }, { "epoch": 0.11531810988964925, "grad_norm": 2.0654869079589844, "learning_rate": 9.81036657290512e-06, "loss": 0.5884, "step": 2751 }, { "epoch": 0.11536002850465821, "grad_norm": 1.8191479444503784, "learning_rate": 9.810181343763324e-06, "loss": 0.5806, "step": 2752 }, { "epoch": 0.11540194711966717, "grad_norm": 1.7818280458450317, "learning_rate": 9.809996025952481e-06, "loss": 0.5438, "step": 2753 }, { "epoch": 0.11544386573467613, "grad_norm": 1.87283456325531, "learning_rate": 9.809810619476012e-06, "loss": 0.5999, "step": 2754 }, { "epoch": 0.11548578434968508, "grad_norm": 1.6971474885940552, "learning_rate": 9.809625124337334e-06, "loss": 0.5617, "step": 2755 }, { "epoch": 0.11552770296469404, "grad_norm": 4.450978755950928, "learning_rate": 9.809439540539867e-06, "loss": 0.5741, "step": 2756 }, { "epoch": 0.115569621579703, "grad_norm": 1.8666092157363892, "learning_rate": 9.80925386808703e-06, "loss": 0.5512, "step": 2757 }, { "epoch": 0.11561154019471197, "grad_norm": 1.7706248760223389, "learning_rate": 9.809068106982246e-06, "loss": 0.5992, "step": 2758 }, { "epoch": 0.11565345880972093, "grad_norm": 1.7425172328948975, "learning_rate": 9.808882257228942e-06, "loss": 0.571, "step": 2759 }, { "epoch": 0.11569537742472989, "grad_norm": 1.938687801361084, "learning_rate": 9.808696318830539e-06, "loss": 0.582, "step": 2760 }, { "epoch": 0.11573729603973884, "grad_norm": 1.827452301979065, "learning_rate": 9.808510291790467e-06, "loss": 0.5627, "step": 2761 }, { "epoch": 0.1157792146547478, "grad_norm": 1.6050926446914673, "learning_rate": 9.808324176112158e-06, "loss": 0.5539, "step": 2762 }, { "epoch": 0.11582113326975677, "grad_norm": 3.539715528488159, "learning_rate": 9.808137971799038e-06, "loss": 0.5156, "step": 2763 }, { "epoch": 0.11586305188476573, "grad_norm": 1.7208247184753418, "learning_rate": 9.807951678854541e-06, "loss": 0.6135, "step": 2764 }, { "epoch": 0.11590497049977469, "grad_norm": 1.8130518198013306, "learning_rate": 9.807765297282104e-06, "loss": 0.5665, "step": 2765 }, { "epoch": 0.11594688911478365, "grad_norm": 1.616891622543335, "learning_rate": 9.80757882708516e-06, "loss": 0.5277, "step": 2766 }, { "epoch": 0.11598880772979261, "grad_norm": 1.8499047756195068, "learning_rate": 9.807392268267144e-06, "loss": 0.6173, "step": 2767 }, { "epoch": 0.11603072634480156, "grad_norm": 1.9105420112609863, "learning_rate": 9.8072056208315e-06, "loss": 0.6129, "step": 2768 }, { "epoch": 0.11607264495981053, "grad_norm": 1.876769781112671, "learning_rate": 9.807018884781665e-06, "loss": 0.5562, "step": 2769 }, { "epoch": 0.11611456357481949, "grad_norm": 3.1101396083831787, "learning_rate": 9.806832060121084e-06, "loss": 0.5953, "step": 2770 }, { "epoch": 0.11615648218982845, "grad_norm": 1.7101248502731323, "learning_rate": 9.806645146853197e-06, "loss": 0.5664, "step": 2771 }, { "epoch": 0.11619840080483741, "grad_norm": 1.6627062559127808, "learning_rate": 9.806458144981453e-06, "loss": 0.5172, "step": 2772 }, { "epoch": 0.11624031941984637, "grad_norm": 1.955999493598938, "learning_rate": 9.806271054509296e-06, "loss": 0.586, "step": 2773 }, { "epoch": 0.11628223803485532, "grad_norm": 1.6712286472320557, "learning_rate": 9.806083875440179e-06, "loss": 0.5721, "step": 2774 }, { "epoch": 0.11632415664986429, "grad_norm": 1.8889118432998657, "learning_rate": 9.805896607777549e-06, "loss": 0.6218, "step": 2775 }, { "epoch": 0.11636607526487325, "grad_norm": 1.8874003887176514, "learning_rate": 9.805709251524858e-06, "loss": 0.6055, "step": 2776 }, { "epoch": 0.11640799387988221, "grad_norm": 1.629642367362976, "learning_rate": 9.805521806685563e-06, "loss": 0.5535, "step": 2777 }, { "epoch": 0.11644991249489117, "grad_norm": 1.867776870727539, "learning_rate": 9.805334273263114e-06, "loss": 0.5682, "step": 2778 }, { "epoch": 0.11649183110990013, "grad_norm": 1.8752492666244507, "learning_rate": 9.805146651260973e-06, "loss": 0.5259, "step": 2779 }, { "epoch": 0.11653374972490908, "grad_norm": 1.7665393352508545, "learning_rate": 9.804958940682596e-06, "loss": 0.5681, "step": 2780 }, { "epoch": 0.11657566833991805, "grad_norm": 1.8581181764602661, "learning_rate": 9.804771141531444e-06, "loss": 0.6011, "step": 2781 }, { "epoch": 0.11661758695492701, "grad_norm": 1.7176246643066406, "learning_rate": 9.804583253810978e-06, "loss": 0.599, "step": 2782 }, { "epoch": 0.11665950556993597, "grad_norm": 2.4582650661468506, "learning_rate": 9.804395277524662e-06, "loss": 0.6375, "step": 2783 }, { "epoch": 0.11670142418494493, "grad_norm": 1.940409779548645, "learning_rate": 9.804207212675959e-06, "loss": 0.6159, "step": 2784 }, { "epoch": 0.1167433427999539, "grad_norm": 1.8447390794754028, "learning_rate": 9.804019059268339e-06, "loss": 0.6281, "step": 2785 }, { "epoch": 0.11678526141496284, "grad_norm": 1.7621290683746338, "learning_rate": 9.80383081730527e-06, "loss": 0.5889, "step": 2786 }, { "epoch": 0.1168271800299718, "grad_norm": 1.6561864614486694, "learning_rate": 9.80364248679022e-06, "loss": 0.623, "step": 2787 }, { "epoch": 0.11686909864498077, "grad_norm": 1.8366433382034302, "learning_rate": 9.803454067726662e-06, "loss": 0.6014, "step": 2788 }, { "epoch": 0.11691101725998973, "grad_norm": 1.6798975467681885, "learning_rate": 9.803265560118069e-06, "loss": 0.5544, "step": 2789 }, { "epoch": 0.11695293587499869, "grad_norm": 1.9219259023666382, "learning_rate": 9.803076963967914e-06, "loss": 0.5504, "step": 2790 }, { "epoch": 0.11699485449000765, "grad_norm": 1.9328763484954834, "learning_rate": 9.802888279279676e-06, "loss": 0.6181, "step": 2791 }, { "epoch": 0.11703677310501662, "grad_norm": 1.8964773416519165, "learning_rate": 9.802699506056835e-06, "loss": 0.534, "step": 2792 }, { "epoch": 0.11707869172002557, "grad_norm": 1.9902743101119995, "learning_rate": 9.802510644302866e-06, "loss": 0.6105, "step": 2793 }, { "epoch": 0.11712061033503453, "grad_norm": 1.5856932401657104, "learning_rate": 9.802321694021251e-06, "loss": 0.5312, "step": 2794 }, { "epoch": 0.11716252895004349, "grad_norm": 1.7985267639160156, "learning_rate": 9.802132655215477e-06, "loss": 0.623, "step": 2795 }, { "epoch": 0.11720444756505245, "grad_norm": 1.8150334358215332, "learning_rate": 9.801943527889026e-06, "loss": 0.5674, "step": 2796 }, { "epoch": 0.11724636618006141, "grad_norm": 2.0760107040405273, "learning_rate": 9.801754312045383e-06, "loss": 0.568, "step": 2797 }, { "epoch": 0.11728828479507038, "grad_norm": 1.6116647720336914, "learning_rate": 9.801565007688039e-06, "loss": 0.5464, "step": 2798 }, { "epoch": 0.11733020341007933, "grad_norm": 1.6991121768951416, "learning_rate": 9.801375614820482e-06, "loss": 0.5425, "step": 2799 }, { "epoch": 0.11737212202508829, "grad_norm": 2.2294113636016846, "learning_rate": 9.801186133446202e-06, "loss": 0.5475, "step": 2800 }, { "epoch": 0.11741404064009725, "grad_norm": 1.8189241886138916, "learning_rate": 9.800996563568695e-06, "loss": 0.5796, "step": 2801 }, { "epoch": 0.11745595925510621, "grad_norm": 1.9643405675888062, "learning_rate": 9.800806905191452e-06, "loss": 0.6327, "step": 2802 }, { "epoch": 0.11749787787011517, "grad_norm": 1.6754090785980225, "learning_rate": 9.800617158317972e-06, "loss": 0.6443, "step": 2803 }, { "epoch": 0.11753979648512414, "grad_norm": 1.8297125101089478, "learning_rate": 9.800427322951749e-06, "loss": 0.5702, "step": 2804 }, { "epoch": 0.11758171510013309, "grad_norm": 1.7106237411499023, "learning_rate": 9.800237399096286e-06, "loss": 0.5249, "step": 2805 }, { "epoch": 0.11762363371514205, "grad_norm": 1.9785083532333374, "learning_rate": 9.800047386755082e-06, "loss": 0.695, "step": 2806 }, { "epoch": 0.11766555233015101, "grad_norm": 1.8322008848190308, "learning_rate": 9.799857285931641e-06, "loss": 0.5581, "step": 2807 }, { "epoch": 0.11770747094515997, "grad_norm": 1.8193824291229248, "learning_rate": 9.799667096629468e-06, "loss": 0.5591, "step": 2808 }, { "epoch": 0.11774938956016893, "grad_norm": 1.6220647096633911, "learning_rate": 9.799476818852066e-06, "loss": 0.567, "step": 2809 }, { "epoch": 0.1177913081751779, "grad_norm": 1.6828410625457764, "learning_rate": 9.799286452602943e-06, "loss": 0.6097, "step": 2810 }, { "epoch": 0.11783322679018685, "grad_norm": 1.6646653413772583, "learning_rate": 9.79909599788561e-06, "loss": 0.5911, "step": 2811 }, { "epoch": 0.11787514540519581, "grad_norm": 1.649328589439392, "learning_rate": 9.798905454703577e-06, "loss": 0.5435, "step": 2812 }, { "epoch": 0.11791706402020477, "grad_norm": 1.8101603984832764, "learning_rate": 9.798714823060356e-06, "loss": 0.6256, "step": 2813 }, { "epoch": 0.11795898263521373, "grad_norm": 2.003643274307251, "learning_rate": 9.798524102959462e-06, "loss": 0.5062, "step": 2814 }, { "epoch": 0.1180009012502227, "grad_norm": 1.734142541885376, "learning_rate": 9.798333294404408e-06, "loss": 0.5936, "step": 2815 }, { "epoch": 0.11804281986523166, "grad_norm": 1.84491765499115, "learning_rate": 9.798142397398714e-06, "loss": 0.5608, "step": 2816 }, { "epoch": 0.11808473848024062, "grad_norm": 1.8329178094863892, "learning_rate": 9.797951411945898e-06, "loss": 0.5521, "step": 2817 }, { "epoch": 0.11812665709524957, "grad_norm": 1.7902995347976685, "learning_rate": 9.797760338049482e-06, "loss": 0.5726, "step": 2818 }, { "epoch": 0.11816857571025853, "grad_norm": 1.6253491640090942, "learning_rate": 9.797569175712986e-06, "loss": 0.4929, "step": 2819 }, { "epoch": 0.11821049432526749, "grad_norm": 1.9976608753204346, "learning_rate": 9.797377924939935e-06, "loss": 0.5864, "step": 2820 }, { "epoch": 0.11825241294027646, "grad_norm": 1.9386621713638306, "learning_rate": 9.797186585733854e-06, "loss": 0.5875, "step": 2821 }, { "epoch": 0.11829433155528542, "grad_norm": 1.6994012594223022, "learning_rate": 9.79699515809827e-06, "loss": 0.5877, "step": 2822 }, { "epoch": 0.11833625017029438, "grad_norm": 1.8107649087905884, "learning_rate": 9.796803642036712e-06, "loss": 0.5671, "step": 2823 }, { "epoch": 0.11837816878530333, "grad_norm": 2.0746936798095703, "learning_rate": 9.796612037552711e-06, "loss": 0.5879, "step": 2824 }, { "epoch": 0.11842008740031229, "grad_norm": 1.9193073511123657, "learning_rate": 9.796420344649796e-06, "loss": 0.5849, "step": 2825 }, { "epoch": 0.11846200601532125, "grad_norm": 1.9519352912902832, "learning_rate": 9.796228563331504e-06, "loss": 0.6025, "step": 2826 }, { "epoch": 0.11850392463033022, "grad_norm": 1.900984525680542, "learning_rate": 9.79603669360137e-06, "loss": 0.6174, "step": 2827 }, { "epoch": 0.11854584324533918, "grad_norm": 1.8655731678009033, "learning_rate": 9.795844735462929e-06, "loss": 0.5614, "step": 2828 }, { "epoch": 0.11858776186034814, "grad_norm": 1.8591279983520508, "learning_rate": 9.795652688919721e-06, "loss": 0.5722, "step": 2829 }, { "epoch": 0.11862968047535709, "grad_norm": 1.987472653388977, "learning_rate": 9.795460553975284e-06, "loss": 0.61, "step": 2830 }, { "epoch": 0.11867159909036605, "grad_norm": 1.8389368057250977, "learning_rate": 9.795268330633163e-06, "loss": 0.5732, "step": 2831 }, { "epoch": 0.11871351770537501, "grad_norm": 1.6704422235488892, "learning_rate": 9.795076018896898e-06, "loss": 0.5684, "step": 2832 }, { "epoch": 0.11875543632038398, "grad_norm": 1.7741771936416626, "learning_rate": 9.794883618770036e-06, "loss": 0.6525, "step": 2833 }, { "epoch": 0.11879735493539294, "grad_norm": 1.8443068265914917, "learning_rate": 9.794691130256124e-06, "loss": 0.5756, "step": 2834 }, { "epoch": 0.1188392735504019, "grad_norm": 1.795312762260437, "learning_rate": 9.794498553358708e-06, "loss": 0.5703, "step": 2835 }, { "epoch": 0.11888119216541085, "grad_norm": 1.6333253383636475, "learning_rate": 9.794305888081342e-06, "loss": 0.5474, "step": 2836 }, { "epoch": 0.11892311078041981, "grad_norm": 1.9176267385482788, "learning_rate": 9.794113134427575e-06, "loss": 0.6235, "step": 2837 }, { "epoch": 0.11896502939542877, "grad_norm": 1.8272185325622559, "learning_rate": 9.793920292400956e-06, "loss": 0.5831, "step": 2838 }, { "epoch": 0.11900694801043774, "grad_norm": 1.9066181182861328, "learning_rate": 9.793727362005047e-06, "loss": 0.5927, "step": 2839 }, { "epoch": 0.1190488666254467, "grad_norm": 1.8769925832748413, "learning_rate": 9.7935343432434e-06, "loss": 0.6263, "step": 2840 }, { "epoch": 0.11909078524045566, "grad_norm": 1.9659017324447632, "learning_rate": 9.793341236119574e-06, "loss": 0.5266, "step": 2841 }, { "epoch": 0.11913270385546462, "grad_norm": 2.089914560317993, "learning_rate": 9.79314804063713e-06, "loss": 0.6335, "step": 2842 }, { "epoch": 0.11917462247047357, "grad_norm": 1.7376819849014282, "learning_rate": 9.792954756799629e-06, "loss": 0.5688, "step": 2843 }, { "epoch": 0.11921654108548253, "grad_norm": 1.8040449619293213, "learning_rate": 9.792761384610631e-06, "loss": 0.6013, "step": 2844 }, { "epoch": 0.1192584597004915, "grad_norm": 1.7143360376358032, "learning_rate": 9.792567924073704e-06, "loss": 0.5584, "step": 2845 }, { "epoch": 0.11930037831550046, "grad_norm": 1.9609497785568237, "learning_rate": 9.792374375192412e-06, "loss": 0.6061, "step": 2846 }, { "epoch": 0.11934229693050942, "grad_norm": 1.8755638599395752, "learning_rate": 9.792180737970324e-06, "loss": 0.591, "step": 2847 }, { "epoch": 0.11938421554551838, "grad_norm": 1.7626473903656006, "learning_rate": 9.791987012411008e-06, "loss": 0.5507, "step": 2848 }, { "epoch": 0.11942613416052733, "grad_norm": 1.6839993000030518, "learning_rate": 9.791793198518039e-06, "loss": 0.5545, "step": 2849 }, { "epoch": 0.1194680527755363, "grad_norm": 2.277898073196411, "learning_rate": 9.791599296294985e-06, "loss": 0.6052, "step": 2850 }, { "epoch": 0.11950997139054526, "grad_norm": 1.8387713432312012, "learning_rate": 9.79140530574542e-06, "loss": 0.598, "step": 2851 }, { "epoch": 0.11955189000555422, "grad_norm": 1.883314847946167, "learning_rate": 9.791211226872925e-06, "loss": 0.5861, "step": 2852 }, { "epoch": 0.11959380862056318, "grad_norm": 1.8585522174835205, "learning_rate": 9.791017059681074e-06, "loss": 0.5919, "step": 2853 }, { "epoch": 0.11963572723557214, "grad_norm": 1.7696863412857056, "learning_rate": 9.790822804173447e-06, "loss": 0.5358, "step": 2854 }, { "epoch": 0.11967764585058109, "grad_norm": 2.942967414855957, "learning_rate": 9.790628460353625e-06, "loss": 0.5565, "step": 2855 }, { "epoch": 0.11971956446559005, "grad_norm": 1.6456780433654785, "learning_rate": 9.79043402822519e-06, "loss": 0.5217, "step": 2856 }, { "epoch": 0.11976148308059902, "grad_norm": 2.316190242767334, "learning_rate": 9.790239507791725e-06, "loss": 0.5823, "step": 2857 }, { "epoch": 0.11980340169560798, "grad_norm": 1.9642689228057861, "learning_rate": 9.790044899056818e-06, "loss": 0.5575, "step": 2858 }, { "epoch": 0.11984532031061694, "grad_norm": 1.796074628829956, "learning_rate": 9.789850202024055e-06, "loss": 0.6464, "step": 2859 }, { "epoch": 0.1198872389256259, "grad_norm": 1.7339011430740356, "learning_rate": 9.789655416697027e-06, "loss": 0.5445, "step": 2860 }, { "epoch": 0.11992915754063485, "grad_norm": 1.748389482498169, "learning_rate": 9.789460543079322e-06, "loss": 0.5781, "step": 2861 }, { "epoch": 0.11997107615564381, "grad_norm": 1.8173837661743164, "learning_rate": 9.789265581174532e-06, "loss": 0.534, "step": 2862 }, { "epoch": 0.12001299477065278, "grad_norm": 1.5432568788528442, "learning_rate": 9.789070530986253e-06, "loss": 0.4987, "step": 2863 }, { "epoch": 0.12005491338566174, "grad_norm": 2.5239975452423096, "learning_rate": 9.788875392518078e-06, "loss": 0.5994, "step": 2864 }, { "epoch": 0.1200968320006707, "grad_norm": 1.92904794216156, "learning_rate": 9.788680165773608e-06, "loss": 0.5602, "step": 2865 }, { "epoch": 0.12013875061567966, "grad_norm": 1.9762123823165894, "learning_rate": 9.788484850756437e-06, "loss": 0.5948, "step": 2866 }, { "epoch": 0.12018066923068862, "grad_norm": 1.7443093061447144, "learning_rate": 9.78828944747017e-06, "loss": 0.5686, "step": 2867 }, { "epoch": 0.12022258784569757, "grad_norm": 1.8715709447860718, "learning_rate": 9.788093955918404e-06, "loss": 0.5918, "step": 2868 }, { "epoch": 0.12026450646070654, "grad_norm": 1.9491428136825562, "learning_rate": 9.787898376104747e-06, "loss": 0.5684, "step": 2869 }, { "epoch": 0.1203064250757155, "grad_norm": 2.3308281898498535, "learning_rate": 9.787702708032803e-06, "loss": 0.5613, "step": 2870 }, { "epoch": 0.12034834369072446, "grad_norm": 3.6675877571105957, "learning_rate": 9.787506951706176e-06, "loss": 0.6337, "step": 2871 }, { "epoch": 0.12039026230573342, "grad_norm": 1.9699063301086426, "learning_rate": 9.787311107128479e-06, "loss": 0.6012, "step": 2872 }, { "epoch": 0.12043218092074238, "grad_norm": 1.8747987747192383, "learning_rate": 9.787115174303318e-06, "loss": 0.6042, "step": 2873 }, { "epoch": 0.12047409953575133, "grad_norm": 1.681703805923462, "learning_rate": 9.78691915323431e-06, "loss": 0.5891, "step": 2874 }, { "epoch": 0.1205160181507603, "grad_norm": 2.116288900375366, "learning_rate": 9.786723043925062e-06, "loss": 0.6007, "step": 2875 }, { "epoch": 0.12055793676576926, "grad_norm": 1.5604157447814941, "learning_rate": 9.786526846379193e-06, "loss": 0.539, "step": 2876 }, { "epoch": 0.12059985538077822, "grad_norm": 1.6882774829864502, "learning_rate": 9.786330560600318e-06, "loss": 0.6018, "step": 2877 }, { "epoch": 0.12064177399578718, "grad_norm": 1.870164394378662, "learning_rate": 9.786134186592059e-06, "loss": 0.6143, "step": 2878 }, { "epoch": 0.12068369261079614, "grad_norm": 2.0806238651275635, "learning_rate": 9.78593772435803e-06, "loss": 0.6834, "step": 2879 }, { "epoch": 0.1207256112258051, "grad_norm": 1.6225700378417969, "learning_rate": 9.785741173901857e-06, "loss": 0.577, "step": 2880 }, { "epoch": 0.12076752984081406, "grad_norm": 1.683260440826416, "learning_rate": 9.78554453522716e-06, "loss": 0.5222, "step": 2881 }, { "epoch": 0.12080944845582302, "grad_norm": 1.7634434700012207, "learning_rate": 9.785347808337566e-06, "loss": 0.5378, "step": 2882 }, { "epoch": 0.12085136707083198, "grad_norm": 2.079617738723755, "learning_rate": 9.785150993236701e-06, "loss": 0.6193, "step": 2883 }, { "epoch": 0.12089328568584094, "grad_norm": 2.3179590702056885, "learning_rate": 9.784954089928193e-06, "loss": 0.6672, "step": 2884 }, { "epoch": 0.1209352043008499, "grad_norm": 1.7474209070205688, "learning_rate": 9.784757098415671e-06, "loss": 0.564, "step": 2885 }, { "epoch": 0.12097712291585885, "grad_norm": 1.8526391983032227, "learning_rate": 9.784560018702767e-06, "loss": 0.5386, "step": 2886 }, { "epoch": 0.12101904153086782, "grad_norm": 1.661253571510315, "learning_rate": 9.784362850793113e-06, "loss": 0.5438, "step": 2887 }, { "epoch": 0.12106096014587678, "grad_norm": 1.8778923749923706, "learning_rate": 9.784165594690344e-06, "loss": 0.5582, "step": 2888 }, { "epoch": 0.12110287876088574, "grad_norm": 2.099252223968506, "learning_rate": 9.783968250398097e-06, "loss": 0.5985, "step": 2889 }, { "epoch": 0.1211447973758947, "grad_norm": 1.6833806037902832, "learning_rate": 9.783770817920009e-06, "loss": 0.6096, "step": 2890 }, { "epoch": 0.12118671599090367, "grad_norm": 1.549032211303711, "learning_rate": 9.783573297259718e-06, "loss": 0.5052, "step": 2891 }, { "epoch": 0.12122863460591263, "grad_norm": 1.777518391609192, "learning_rate": 9.783375688420866e-06, "loss": 0.5686, "step": 2892 }, { "epoch": 0.12127055322092158, "grad_norm": 1.8132013082504272, "learning_rate": 9.783177991407098e-06, "loss": 0.5665, "step": 2893 }, { "epoch": 0.12131247183593054, "grad_norm": 1.8597015142440796, "learning_rate": 9.782980206222055e-06, "loss": 0.6082, "step": 2894 }, { "epoch": 0.1213543904509395, "grad_norm": 1.8288196325302124, "learning_rate": 9.782782332869386e-06, "loss": 0.556, "step": 2895 }, { "epoch": 0.12139630906594846, "grad_norm": 1.6939311027526855, "learning_rate": 9.782584371352734e-06, "loss": 0.5967, "step": 2896 }, { "epoch": 0.12143822768095743, "grad_norm": 1.814015507698059, "learning_rate": 9.782386321675752e-06, "loss": 0.601, "step": 2897 }, { "epoch": 0.12148014629596639, "grad_norm": 1.7643851041793823, "learning_rate": 9.782188183842089e-06, "loss": 0.596, "step": 2898 }, { "epoch": 0.12152206491097534, "grad_norm": 1.7402374744415283, "learning_rate": 9.781989957855398e-06, "loss": 0.582, "step": 2899 }, { "epoch": 0.1215639835259843, "grad_norm": 3.212937116622925, "learning_rate": 9.781791643719335e-06, "loss": 0.6084, "step": 2900 }, { "epoch": 0.12160590214099326, "grad_norm": 2.683952808380127, "learning_rate": 9.781593241437552e-06, "loss": 0.6247, "step": 2901 }, { "epoch": 0.12164782075600222, "grad_norm": 1.831760287284851, "learning_rate": 9.781394751013707e-06, "loss": 0.5698, "step": 2902 }, { "epoch": 0.12168973937101119, "grad_norm": 4.490127086639404, "learning_rate": 9.78119617245146e-06, "loss": 0.6389, "step": 2903 }, { "epoch": 0.12173165798602015, "grad_norm": 1.770747184753418, "learning_rate": 9.780997505754474e-06, "loss": 0.592, "step": 2904 }, { "epoch": 0.1217735766010291, "grad_norm": 1.891520380973816, "learning_rate": 9.780798750926406e-06, "loss": 0.6413, "step": 2905 }, { "epoch": 0.12181549521603806, "grad_norm": 1.903642177581787, "learning_rate": 9.780599907970925e-06, "loss": 0.5399, "step": 2906 }, { "epoch": 0.12185741383104702, "grad_norm": 2.040019989013672, "learning_rate": 9.78040097689169e-06, "loss": 0.5476, "step": 2907 }, { "epoch": 0.12189933244605598, "grad_norm": 1.9075993299484253, "learning_rate": 9.780201957692375e-06, "loss": 0.5039, "step": 2908 }, { "epoch": 0.12194125106106495, "grad_norm": 1.8847938776016235, "learning_rate": 9.780002850376644e-06, "loss": 0.596, "step": 2909 }, { "epoch": 0.12198316967607391, "grad_norm": 1.5887799263000488, "learning_rate": 9.779803654948168e-06, "loss": 0.5936, "step": 2910 }, { "epoch": 0.12202508829108286, "grad_norm": 1.832157015800476, "learning_rate": 9.77960437141062e-06, "loss": 0.5575, "step": 2911 }, { "epoch": 0.12206700690609182, "grad_norm": 1.8159433603286743, "learning_rate": 9.779404999767673e-06, "loss": 0.5529, "step": 2912 }, { "epoch": 0.12210892552110078, "grad_norm": 1.9614264965057373, "learning_rate": 9.779205540023003e-06, "loss": 0.5353, "step": 2913 }, { "epoch": 0.12215084413610974, "grad_norm": 1.863904595375061, "learning_rate": 9.779005992180285e-06, "loss": 0.5858, "step": 2914 }, { "epoch": 0.1221927627511187, "grad_norm": 1.897255301475525, "learning_rate": 9.778806356243198e-06, "loss": 0.5658, "step": 2915 }, { "epoch": 0.12223468136612767, "grad_norm": 1.8787643909454346, "learning_rate": 9.778606632215424e-06, "loss": 0.6264, "step": 2916 }, { "epoch": 0.12227659998113663, "grad_norm": 1.7244065999984741, "learning_rate": 9.778406820100641e-06, "loss": 0.6036, "step": 2917 }, { "epoch": 0.12231851859614558, "grad_norm": 2.199855089187622, "learning_rate": 9.778206919902535e-06, "loss": 0.6806, "step": 2918 }, { "epoch": 0.12236043721115454, "grad_norm": 1.8888771533966064, "learning_rate": 9.778006931624791e-06, "loss": 0.593, "step": 2919 }, { "epoch": 0.1224023558261635, "grad_norm": 2.0137341022491455, "learning_rate": 9.777806855271093e-06, "loss": 0.5772, "step": 2920 }, { "epoch": 0.12244427444117247, "grad_norm": 1.809531569480896, "learning_rate": 9.777606690845131e-06, "loss": 0.5653, "step": 2921 }, { "epoch": 0.12248619305618143, "grad_norm": 1.5785406827926636, "learning_rate": 9.777406438350594e-06, "loss": 0.5438, "step": 2922 }, { "epoch": 0.12252811167119039, "grad_norm": 1.8041632175445557, "learning_rate": 9.777206097791175e-06, "loss": 0.5978, "step": 2923 }, { "epoch": 0.12257003028619934, "grad_norm": 1.7544506788253784, "learning_rate": 9.777005669170567e-06, "loss": 0.5647, "step": 2924 }, { "epoch": 0.1226119489012083, "grad_norm": 2.0825910568237305, "learning_rate": 9.776805152492461e-06, "loss": 0.5919, "step": 2925 }, { "epoch": 0.12265386751621726, "grad_norm": 1.7749460935592651, "learning_rate": 9.776604547760557e-06, "loss": 0.5081, "step": 2926 }, { "epoch": 0.12269578613122623, "grad_norm": 1.69261634349823, "learning_rate": 9.776403854978552e-06, "loss": 0.5853, "step": 2927 }, { "epoch": 0.12273770474623519, "grad_norm": 2.0077478885650635, "learning_rate": 9.776203074150145e-06, "loss": 0.5717, "step": 2928 }, { "epoch": 0.12277962336124415, "grad_norm": 1.726866364479065, "learning_rate": 9.776002205279036e-06, "loss": 0.5915, "step": 2929 }, { "epoch": 0.1228215419762531, "grad_norm": 1.956536889076233, "learning_rate": 9.77580124836893e-06, "loss": 0.5866, "step": 2930 }, { "epoch": 0.12286346059126206, "grad_norm": 1.7888593673706055, "learning_rate": 9.77560020342353e-06, "loss": 0.5644, "step": 2931 }, { "epoch": 0.12290537920627102, "grad_norm": 1.9190303087234497, "learning_rate": 9.775399070446543e-06, "loss": 0.6602, "step": 2932 }, { "epoch": 0.12294729782127999, "grad_norm": 1.8957833051681519, "learning_rate": 9.775197849441676e-06, "loss": 0.5561, "step": 2933 }, { "epoch": 0.12298921643628895, "grad_norm": 1.6199265718460083, "learning_rate": 9.774996540412637e-06, "loss": 0.4932, "step": 2934 }, { "epoch": 0.12303113505129791, "grad_norm": 1.5398176908493042, "learning_rate": 9.774795143363138e-06, "loss": 0.5236, "step": 2935 }, { "epoch": 0.12307305366630686, "grad_norm": 1.7185099124908447, "learning_rate": 9.774593658296892e-06, "loss": 0.5468, "step": 2936 }, { "epoch": 0.12311497228131582, "grad_norm": 1.8397555351257324, "learning_rate": 9.774392085217614e-06, "loss": 0.5664, "step": 2937 }, { "epoch": 0.12315689089632478, "grad_norm": 1.95266854763031, "learning_rate": 9.774190424129017e-06, "loss": 0.6191, "step": 2938 }, { "epoch": 0.12319880951133375, "grad_norm": 1.702089548110962, "learning_rate": 9.773988675034819e-06, "loss": 0.5152, "step": 2939 }, { "epoch": 0.12324072812634271, "grad_norm": 1.5982879400253296, "learning_rate": 9.773786837938742e-06, "loss": 0.5396, "step": 2940 }, { "epoch": 0.12328264674135167, "grad_norm": 1.8060188293457031, "learning_rate": 9.7735849128445e-06, "loss": 0.5409, "step": 2941 }, { "epoch": 0.12332456535636063, "grad_norm": 1.7806740999221802, "learning_rate": 9.773382899755823e-06, "loss": 0.5287, "step": 2942 }, { "epoch": 0.12336648397136958, "grad_norm": 1.6691604852676392, "learning_rate": 9.77318079867643e-06, "loss": 0.6121, "step": 2943 }, { "epoch": 0.12340840258637854, "grad_norm": 1.9172037839889526, "learning_rate": 9.772978609610048e-06, "loss": 0.5986, "step": 2944 }, { "epoch": 0.1234503212013875, "grad_norm": 1.6508002281188965, "learning_rate": 9.772776332560404e-06, "loss": 0.5567, "step": 2945 }, { "epoch": 0.12349223981639647, "grad_norm": 1.9645980596542358, "learning_rate": 9.772573967531226e-06, "loss": 0.5499, "step": 2946 }, { "epoch": 0.12353415843140543, "grad_norm": 1.6353987455368042, "learning_rate": 9.772371514526244e-06, "loss": 0.5321, "step": 2947 }, { "epoch": 0.12357607704641439, "grad_norm": 1.786075472831726, "learning_rate": 9.772168973549192e-06, "loss": 0.5517, "step": 2948 }, { "epoch": 0.12361799566142334, "grad_norm": 2.7968618869781494, "learning_rate": 9.771966344603802e-06, "loss": 0.5267, "step": 2949 }, { "epoch": 0.1236599142764323, "grad_norm": 1.9159047603607178, "learning_rate": 9.771763627693809e-06, "loss": 0.617, "step": 2950 }, { "epoch": 0.12370183289144127, "grad_norm": 2.1825504302978516, "learning_rate": 9.77156082282295e-06, "loss": 0.5798, "step": 2951 }, { "epoch": 0.12374375150645023, "grad_norm": 1.7439532279968262, "learning_rate": 9.771357929994964e-06, "loss": 0.6065, "step": 2952 }, { "epoch": 0.12378567012145919, "grad_norm": 1.8140158653259277, "learning_rate": 9.771154949213593e-06, "loss": 0.633, "step": 2953 }, { "epoch": 0.12382758873646815, "grad_norm": 1.8224557638168335, "learning_rate": 9.770951880482574e-06, "loss": 0.6051, "step": 2954 }, { "epoch": 0.1238695073514771, "grad_norm": 1.8259996175765991, "learning_rate": 9.770748723805655e-06, "loss": 0.6259, "step": 2955 }, { "epoch": 0.12391142596648606, "grad_norm": 1.9937366247177124, "learning_rate": 9.770545479186574e-06, "loss": 0.5273, "step": 2956 }, { "epoch": 0.12395334458149503, "grad_norm": 1.58124577999115, "learning_rate": 9.770342146629086e-06, "loss": 0.5672, "step": 2957 }, { "epoch": 0.12399526319650399, "grad_norm": 1.6788057088851929, "learning_rate": 9.770138726136935e-06, "loss": 0.5302, "step": 2958 }, { "epoch": 0.12403718181151295, "grad_norm": 2.0126192569732666, "learning_rate": 9.769935217713871e-06, "loss": 0.5921, "step": 2959 }, { "epoch": 0.12407910042652191, "grad_norm": 2.1221420764923096, "learning_rate": 9.769731621363645e-06, "loss": 0.609, "step": 2960 }, { "epoch": 0.12412101904153086, "grad_norm": 2.1605546474456787, "learning_rate": 9.769527937090011e-06, "loss": 0.5681, "step": 2961 }, { "epoch": 0.12416293765653982, "grad_norm": 3.1721932888031006, "learning_rate": 9.769324164896722e-06, "loss": 0.5403, "step": 2962 }, { "epoch": 0.12420485627154879, "grad_norm": 1.8233952522277832, "learning_rate": 9.769120304787536e-06, "loss": 0.5912, "step": 2963 }, { "epoch": 0.12424677488655775, "grad_norm": 2.1270697116851807, "learning_rate": 9.768916356766211e-06, "loss": 0.5882, "step": 2964 }, { "epoch": 0.12428869350156671, "grad_norm": 1.798732042312622, "learning_rate": 9.768712320836503e-06, "loss": 0.5405, "step": 2965 }, { "epoch": 0.12433061211657567, "grad_norm": 2.2109906673431396, "learning_rate": 9.76850819700218e-06, "loss": 0.592, "step": 2966 }, { "epoch": 0.12437253073158464, "grad_norm": 3.845005989074707, "learning_rate": 9.768303985266999e-06, "loss": 0.5829, "step": 2967 }, { "epoch": 0.12441444934659358, "grad_norm": 1.9364609718322754, "learning_rate": 9.768099685634725e-06, "loss": 0.5705, "step": 2968 }, { "epoch": 0.12445636796160255, "grad_norm": 1.867772102355957, "learning_rate": 9.767895298109127e-06, "loss": 0.6038, "step": 2969 }, { "epoch": 0.12449828657661151, "grad_norm": 2.081193447113037, "learning_rate": 9.76769082269397e-06, "loss": 0.6125, "step": 2970 }, { "epoch": 0.12454020519162047, "grad_norm": 1.614556908607483, "learning_rate": 9.767486259393023e-06, "loss": 0.5161, "step": 2971 }, { "epoch": 0.12458212380662943, "grad_norm": 1.8030613660812378, "learning_rate": 9.767281608210058e-06, "loss": 0.5478, "step": 2972 }, { "epoch": 0.1246240424216384, "grad_norm": 1.8869339227676392, "learning_rate": 9.767076869148846e-06, "loss": 0.6191, "step": 2973 }, { "epoch": 0.12466596103664734, "grad_norm": 1.724998950958252, "learning_rate": 9.766872042213162e-06, "loss": 0.5471, "step": 2974 }, { "epoch": 0.1247078796516563, "grad_norm": 1.6293150186538696, "learning_rate": 9.766667127406783e-06, "loss": 0.5206, "step": 2975 }, { "epoch": 0.12474979826666527, "grad_norm": 1.941900372505188, "learning_rate": 9.766462124733487e-06, "loss": 0.5833, "step": 2976 }, { "epoch": 0.12479171688167423, "grad_norm": 1.6479380130767822, "learning_rate": 9.766257034197048e-06, "loss": 0.5763, "step": 2977 }, { "epoch": 0.12483363549668319, "grad_norm": 1.9056810140609741, "learning_rate": 9.766051855801253e-06, "loss": 0.5741, "step": 2978 }, { "epoch": 0.12487555411169216, "grad_norm": 1.8463799953460693, "learning_rate": 9.76584658954988e-06, "loss": 0.583, "step": 2979 }, { "epoch": 0.1249174727267011, "grad_norm": 1.6958619356155396, "learning_rate": 9.765641235446712e-06, "loss": 0.5293, "step": 2980 }, { "epoch": 0.12495939134171007, "grad_norm": 2.0312860012054443, "learning_rate": 9.765435793495537e-06, "loss": 0.5532, "step": 2981 }, { "epoch": 0.12500130995671904, "grad_norm": 2.466115951538086, "learning_rate": 9.76523026370014e-06, "loss": 0.5686, "step": 2982 }, { "epoch": 0.12504322857172798, "grad_norm": 1.7178080081939697, "learning_rate": 9.765024646064312e-06, "loss": 0.5601, "step": 2983 }, { "epoch": 0.12508514718673694, "grad_norm": 1.8796842098236084, "learning_rate": 9.764818940591842e-06, "loss": 0.5504, "step": 2984 }, { "epoch": 0.1251270658017459, "grad_norm": 1.712598204612732, "learning_rate": 9.764613147286522e-06, "loss": 0.5712, "step": 2985 }, { "epoch": 0.12516898441675486, "grad_norm": 2.0804009437561035, "learning_rate": 9.764407266152147e-06, "loss": 0.6002, "step": 2986 }, { "epoch": 0.12521090303176383, "grad_norm": 1.9736263751983643, "learning_rate": 9.764201297192509e-06, "loss": 0.5455, "step": 2987 }, { "epoch": 0.1252528216467728, "grad_norm": 1.7705944776535034, "learning_rate": 9.763995240411406e-06, "loss": 0.5617, "step": 2988 }, { "epoch": 0.12529474026178175, "grad_norm": 1.6923612356185913, "learning_rate": 9.763789095812636e-06, "loss": 0.4944, "step": 2989 }, { "epoch": 0.1253366588767907, "grad_norm": 1.5792266130447388, "learning_rate": 9.763582863400002e-06, "loss": 0.5271, "step": 2990 }, { "epoch": 0.12537857749179968, "grad_norm": 1.6836481094360352, "learning_rate": 9.763376543177302e-06, "loss": 0.5281, "step": 2991 }, { "epoch": 0.12542049610680864, "grad_norm": 2.015977382659912, "learning_rate": 9.763170135148341e-06, "loss": 0.5884, "step": 2992 }, { "epoch": 0.1254624147218176, "grad_norm": 1.9512500762939453, "learning_rate": 9.762963639316924e-06, "loss": 0.5834, "step": 2993 }, { "epoch": 0.12550433333682656, "grad_norm": 1.767894983291626, "learning_rate": 9.762757055686856e-06, "loss": 0.5669, "step": 2994 }, { "epoch": 0.12554625195183552, "grad_norm": 1.8647630214691162, "learning_rate": 9.762550384261947e-06, "loss": 0.552, "step": 2995 }, { "epoch": 0.12558817056684446, "grad_norm": 1.7552671432495117, "learning_rate": 9.762343625046004e-06, "loss": 0.5564, "step": 2996 }, { "epoch": 0.12563008918185342, "grad_norm": 2.182681083679199, "learning_rate": 9.762136778042843e-06, "loss": 0.5178, "step": 2997 }, { "epoch": 0.12567200779686238, "grad_norm": 2.0314981937408447, "learning_rate": 9.761929843256271e-06, "loss": 0.5929, "step": 2998 }, { "epoch": 0.12571392641187135, "grad_norm": 1.8440836668014526, "learning_rate": 9.761722820690108e-06, "loss": 0.5418, "step": 2999 }, { "epoch": 0.1257558450268803, "grad_norm": 2.0122969150543213, "learning_rate": 9.761515710348166e-06, "loss": 0.5539, "step": 3000 }, { "epoch": 0.12579776364188927, "grad_norm": 2.167119026184082, "learning_rate": 9.761308512234265e-06, "loss": 0.5785, "step": 3001 }, { "epoch": 0.12583968225689823, "grad_norm": 2.054845094680786, "learning_rate": 9.761101226352223e-06, "loss": 0.602, "step": 3002 }, { "epoch": 0.1258816008719072, "grad_norm": 1.8298927545547485, "learning_rate": 9.760893852705863e-06, "loss": 0.5715, "step": 3003 }, { "epoch": 0.12592351948691616, "grad_norm": 1.7943155765533447, "learning_rate": 9.760686391299006e-06, "loss": 0.5765, "step": 3004 }, { "epoch": 0.12596543810192512, "grad_norm": 1.7926889657974243, "learning_rate": 9.760478842135476e-06, "loss": 0.5954, "step": 3005 }, { "epoch": 0.12600735671693408, "grad_norm": 1.8663461208343506, "learning_rate": 9.760271205219101e-06, "loss": 0.5643, "step": 3006 }, { "epoch": 0.12604927533194304, "grad_norm": 2.033350944519043, "learning_rate": 9.760063480553708e-06, "loss": 0.6059, "step": 3007 }, { "epoch": 0.12609119394695198, "grad_norm": 3.12709641456604, "learning_rate": 9.759855668143124e-06, "loss": 0.5274, "step": 3008 }, { "epoch": 0.12613311256196094, "grad_norm": 1.5462368726730347, "learning_rate": 9.759647767991181e-06, "loss": 0.5315, "step": 3009 }, { "epoch": 0.1261750311769699, "grad_norm": 1.7820570468902588, "learning_rate": 9.759439780101712e-06, "loss": 0.5542, "step": 3010 }, { "epoch": 0.12621694979197887, "grad_norm": 1.6687370538711548, "learning_rate": 9.759231704478549e-06, "loss": 0.5575, "step": 3011 }, { "epoch": 0.12625886840698783, "grad_norm": 1.9887127876281738, "learning_rate": 9.75902354112553e-06, "loss": 0.5242, "step": 3012 }, { "epoch": 0.1263007870219968, "grad_norm": 1.8800910711288452, "learning_rate": 9.758815290046492e-06, "loss": 0.5528, "step": 3013 }, { "epoch": 0.12634270563700575, "grad_norm": 1.75546133518219, "learning_rate": 9.758606951245272e-06, "loss": 0.5847, "step": 3014 }, { "epoch": 0.12638462425201472, "grad_norm": 1.6749520301818848, "learning_rate": 9.758398524725712e-06, "loss": 0.5474, "step": 3015 }, { "epoch": 0.12642654286702368, "grad_norm": 1.9003113508224487, "learning_rate": 9.758190010491653e-06, "loss": 0.581, "step": 3016 }, { "epoch": 0.12646846148203264, "grad_norm": 2.982637405395508, "learning_rate": 9.75798140854694e-06, "loss": 0.5647, "step": 3017 }, { "epoch": 0.1265103800970416, "grad_norm": 1.9625319242477417, "learning_rate": 9.757772718895416e-06, "loss": 0.5699, "step": 3018 }, { "epoch": 0.12655229871205056, "grad_norm": 1.994333267211914, "learning_rate": 9.75756394154093e-06, "loss": 0.6211, "step": 3019 }, { "epoch": 0.12659421732705953, "grad_norm": 2.2234601974487305, "learning_rate": 9.757355076487329e-06, "loss": 0.5359, "step": 3020 }, { "epoch": 0.12663613594206846, "grad_norm": 1.873854398727417, "learning_rate": 9.757146123738465e-06, "loss": 0.5831, "step": 3021 }, { "epoch": 0.12667805455707742, "grad_norm": 1.950947880744934, "learning_rate": 9.75693708329819e-06, "loss": 0.5763, "step": 3022 }, { "epoch": 0.1267199731720864, "grad_norm": 1.7502241134643555, "learning_rate": 9.756727955170354e-06, "loss": 0.5387, "step": 3023 }, { "epoch": 0.12676189178709535, "grad_norm": 2.5357842445373535, "learning_rate": 9.756518739358815e-06, "loss": 0.5371, "step": 3024 }, { "epoch": 0.1268038104021043, "grad_norm": 1.8363227844238281, "learning_rate": 9.75630943586743e-06, "loss": 0.5593, "step": 3025 }, { "epoch": 0.12684572901711327, "grad_norm": 1.9898334741592407, "learning_rate": 9.756100044700054e-06, "loss": 0.5719, "step": 3026 }, { "epoch": 0.12688764763212224, "grad_norm": 2.743854522705078, "learning_rate": 9.755890565860549e-06, "loss": 0.5452, "step": 3027 }, { "epoch": 0.1269295662471312, "grad_norm": 1.8927111625671387, "learning_rate": 9.755680999352777e-06, "loss": 0.6381, "step": 3028 }, { "epoch": 0.12697148486214016, "grad_norm": 1.9090971946716309, "learning_rate": 9.7554713451806e-06, "loss": 0.5829, "step": 3029 }, { "epoch": 0.12701340347714912, "grad_norm": 1.921535849571228, "learning_rate": 9.755261603347883e-06, "loss": 0.5953, "step": 3030 }, { "epoch": 0.12705532209215809, "grad_norm": 1.8371891975402832, "learning_rate": 9.755051773858493e-06, "loss": 0.6043, "step": 3031 }, { "epoch": 0.12709724070716705, "grad_norm": 1.939368724822998, "learning_rate": 9.754841856716295e-06, "loss": 0.5772, "step": 3032 }, { "epoch": 0.12713915932217598, "grad_norm": 1.937261700630188, "learning_rate": 9.754631851925162e-06, "loss": 0.5853, "step": 3033 }, { "epoch": 0.12718107793718494, "grad_norm": 1.945578932762146, "learning_rate": 9.754421759488965e-06, "loss": 0.5969, "step": 3034 }, { "epoch": 0.1272229965521939, "grad_norm": 1.7767049074172974, "learning_rate": 9.754211579411574e-06, "loss": 0.6005, "step": 3035 }, { "epoch": 0.12726491516720287, "grad_norm": 2.2799904346466064, "learning_rate": 9.754001311696867e-06, "loss": 0.5898, "step": 3036 }, { "epoch": 0.12730683378221183, "grad_norm": 1.724784016609192, "learning_rate": 9.753790956348716e-06, "loss": 0.578, "step": 3037 }, { "epoch": 0.1273487523972208, "grad_norm": 2.179413080215454, "learning_rate": 9.753580513371002e-06, "loss": 0.5807, "step": 3038 }, { "epoch": 0.12739067101222976, "grad_norm": 1.6946278810501099, "learning_rate": 9.753369982767603e-06, "loss": 0.5361, "step": 3039 }, { "epoch": 0.12743258962723872, "grad_norm": 1.9024661779403687, "learning_rate": 9.753159364542399e-06, "loss": 0.5734, "step": 3040 }, { "epoch": 0.12747450824224768, "grad_norm": 2.0427496433258057, "learning_rate": 9.752948658699273e-06, "loss": 0.6182, "step": 3041 }, { "epoch": 0.12751642685725664, "grad_norm": 1.9567623138427734, "learning_rate": 9.75273786524211e-06, "loss": 0.5679, "step": 3042 }, { "epoch": 0.1275583454722656, "grad_norm": 1.9805389642715454, "learning_rate": 9.752526984174793e-06, "loss": 0.5633, "step": 3043 }, { "epoch": 0.12760026408727457, "grad_norm": 2.1455044746398926, "learning_rate": 9.752316015501213e-06, "loss": 0.5777, "step": 3044 }, { "epoch": 0.12764218270228353, "grad_norm": 2.081814765930176, "learning_rate": 9.752104959225257e-06, "loss": 0.5737, "step": 3045 }, { "epoch": 0.12768410131729246, "grad_norm": 2.0212385654449463, "learning_rate": 9.751893815350815e-06, "loss": 0.5259, "step": 3046 }, { "epoch": 0.12772601993230143, "grad_norm": 1.7518178224563599, "learning_rate": 9.75168258388178e-06, "loss": 0.5352, "step": 3047 }, { "epoch": 0.1277679385473104, "grad_norm": 1.7387175559997559, "learning_rate": 9.751471264822046e-06, "loss": 0.6163, "step": 3048 }, { "epoch": 0.12780985716231935, "grad_norm": 1.8578071594238281, "learning_rate": 9.751259858175509e-06, "loss": 0.5554, "step": 3049 }, { "epoch": 0.1278517757773283, "grad_norm": 1.606397032737732, "learning_rate": 9.751048363946063e-06, "loss": 0.545, "step": 3050 }, { "epoch": 0.12789369439233728, "grad_norm": 1.9462469816207886, "learning_rate": 9.75083678213761e-06, "loss": 0.6202, "step": 3051 }, { "epoch": 0.12793561300734624, "grad_norm": 1.9537458419799805, "learning_rate": 9.750625112754046e-06, "loss": 0.5755, "step": 3052 }, { "epoch": 0.1279775316223552, "grad_norm": 1.7418991327285767, "learning_rate": 9.750413355799277e-06, "loss": 0.5948, "step": 3053 }, { "epoch": 0.12801945023736416, "grad_norm": 2.4141931533813477, "learning_rate": 9.750201511277206e-06, "loss": 0.5859, "step": 3054 }, { "epoch": 0.12806136885237313, "grad_norm": 1.9121448993682861, "learning_rate": 9.749989579191737e-06, "loss": 0.5855, "step": 3055 }, { "epoch": 0.1281032874673821, "grad_norm": 1.9869208335876465, "learning_rate": 9.749777559546777e-06, "loss": 0.5807, "step": 3056 }, { "epoch": 0.12814520608239105, "grad_norm": 1.7529951333999634, "learning_rate": 9.749565452346234e-06, "loss": 0.5799, "step": 3057 }, { "epoch": 0.12818712469739998, "grad_norm": 2.26420521736145, "learning_rate": 9.749353257594017e-06, "loss": 0.5906, "step": 3058 }, { "epoch": 0.12822904331240895, "grad_norm": 1.8040307760238647, "learning_rate": 9.749140975294038e-06, "loss": 0.5965, "step": 3059 }, { "epoch": 0.1282709619274179, "grad_norm": 1.905185580253601, "learning_rate": 9.748928605450212e-06, "loss": 0.5274, "step": 3060 }, { "epoch": 0.12831288054242687, "grad_norm": 2.268162727355957, "learning_rate": 9.748716148066452e-06, "loss": 0.6109, "step": 3061 }, { "epoch": 0.12835479915743583, "grad_norm": 1.804998517036438, "learning_rate": 9.748503603146674e-06, "loss": 0.5576, "step": 3062 }, { "epoch": 0.1283967177724448, "grad_norm": 2.0147433280944824, "learning_rate": 9.748290970694798e-06, "loss": 0.6246, "step": 3063 }, { "epoch": 0.12843863638745376, "grad_norm": 1.561990737915039, "learning_rate": 9.748078250714741e-06, "loss": 0.5508, "step": 3064 }, { "epoch": 0.12848055500246272, "grad_norm": 1.817747712135315, "learning_rate": 9.747865443210426e-06, "loss": 0.5306, "step": 3065 }, { "epoch": 0.12852247361747168, "grad_norm": 1.8739440441131592, "learning_rate": 9.747652548185775e-06, "loss": 0.5955, "step": 3066 }, { "epoch": 0.12856439223248065, "grad_norm": 1.8404847383499146, "learning_rate": 9.747439565644714e-06, "loss": 0.5766, "step": 3067 }, { "epoch": 0.1286063108474896, "grad_norm": 6.04142427444458, "learning_rate": 9.747226495591166e-06, "loss": 0.5833, "step": 3068 }, { "epoch": 0.12864822946249857, "grad_norm": 1.7120873928070068, "learning_rate": 9.747013338029062e-06, "loss": 0.495, "step": 3069 }, { "epoch": 0.12869014807750753, "grad_norm": 1.9242428541183472, "learning_rate": 9.746800092962328e-06, "loss": 0.5312, "step": 3070 }, { "epoch": 0.12873206669251647, "grad_norm": 1.7029529809951782, "learning_rate": 9.746586760394898e-06, "loss": 0.5854, "step": 3071 }, { "epoch": 0.12877398530752543, "grad_norm": 2.0849061012268066, "learning_rate": 9.746373340330703e-06, "loss": 0.5568, "step": 3072 }, { "epoch": 0.1288159039225344, "grad_norm": 1.8006452322006226, "learning_rate": 9.746159832773676e-06, "loss": 0.6033, "step": 3073 }, { "epoch": 0.12885782253754335, "grad_norm": 2.1427853107452393, "learning_rate": 9.745946237727756e-06, "loss": 0.5683, "step": 3074 }, { "epoch": 0.12889974115255232, "grad_norm": 2.5854432582855225, "learning_rate": 9.745732555196874e-06, "loss": 0.5437, "step": 3075 }, { "epoch": 0.12894165976756128, "grad_norm": 1.827661156654358, "learning_rate": 9.745518785184978e-06, "loss": 0.5424, "step": 3076 }, { "epoch": 0.12898357838257024, "grad_norm": 1.9704066514968872, "learning_rate": 9.745304927696e-06, "loss": 0.5966, "step": 3077 }, { "epoch": 0.1290254969975792, "grad_norm": 2.0461854934692383, "learning_rate": 9.745090982733887e-06, "loss": 0.5535, "step": 3078 }, { "epoch": 0.12906741561258817, "grad_norm": 2.0093460083007812, "learning_rate": 9.744876950302582e-06, "loss": 0.5819, "step": 3079 }, { "epoch": 0.12910933422759713, "grad_norm": 1.7228964567184448, "learning_rate": 9.74466283040603e-06, "loss": 0.5773, "step": 3080 }, { "epoch": 0.1291512528426061, "grad_norm": 1.7502678632736206, "learning_rate": 9.744448623048177e-06, "loss": 0.5774, "step": 3081 }, { "epoch": 0.12919317145761505, "grad_norm": 1.840252161026001, "learning_rate": 9.744234328232974e-06, "loss": 0.5798, "step": 3082 }, { "epoch": 0.129235090072624, "grad_norm": 1.9038398265838623, "learning_rate": 9.74401994596437e-06, "loss": 0.6103, "step": 3083 }, { "epoch": 0.12927700868763295, "grad_norm": 1.6002832651138306, "learning_rate": 9.743805476246316e-06, "loss": 0.599, "step": 3084 }, { "epoch": 0.1293189273026419, "grad_norm": 2.024108648300171, "learning_rate": 9.743590919082766e-06, "loss": 0.5791, "step": 3085 }, { "epoch": 0.12936084591765087, "grad_norm": 2.1411795616149902, "learning_rate": 9.743376274477676e-06, "loss": 0.6462, "step": 3086 }, { "epoch": 0.12940276453265984, "grad_norm": 1.893127202987671, "learning_rate": 9.743161542435002e-06, "loss": 0.5272, "step": 3087 }, { "epoch": 0.1294446831476688, "grad_norm": 2.033695697784424, "learning_rate": 9.742946722958701e-06, "loss": 0.539, "step": 3088 }, { "epoch": 0.12948660176267776, "grad_norm": 1.7275869846343994, "learning_rate": 9.742731816052735e-06, "loss": 0.5407, "step": 3089 }, { "epoch": 0.12952852037768672, "grad_norm": 1.8913222551345825, "learning_rate": 9.742516821721064e-06, "loss": 0.5887, "step": 3090 }, { "epoch": 0.12957043899269569, "grad_norm": 2.122743844985962, "learning_rate": 9.742301739967652e-06, "loss": 0.613, "step": 3091 }, { "epoch": 0.12961235760770465, "grad_norm": 1.8522911071777344, "learning_rate": 9.742086570796463e-06, "loss": 0.5535, "step": 3092 }, { "epoch": 0.1296542762227136, "grad_norm": 1.5357388257980347, "learning_rate": 9.741871314211464e-06, "loss": 0.5084, "step": 3093 }, { "epoch": 0.12969619483772257, "grad_norm": 2.184133768081665, "learning_rate": 9.741655970216624e-06, "loss": 0.6358, "step": 3094 }, { "epoch": 0.12973811345273153, "grad_norm": 1.7724672555923462, "learning_rate": 9.741440538815909e-06, "loss": 0.6194, "step": 3095 }, { "epoch": 0.12978003206774047, "grad_norm": 1.8812785148620605, "learning_rate": 9.741225020013294e-06, "loss": 0.6265, "step": 3096 }, { "epoch": 0.12982195068274943, "grad_norm": 2.477374315261841, "learning_rate": 9.741009413812749e-06, "loss": 0.5791, "step": 3097 }, { "epoch": 0.1298638692977584, "grad_norm": 1.641723871231079, "learning_rate": 9.740793720218252e-06, "loss": 0.5481, "step": 3098 }, { "epoch": 0.12990578791276736, "grad_norm": 2.3796231746673584, "learning_rate": 9.740577939233776e-06, "loss": 0.5643, "step": 3099 }, { "epoch": 0.12994770652777632, "grad_norm": 1.883097529411316, "learning_rate": 9.740362070863298e-06, "loss": 0.5953, "step": 3100 }, { "epoch": 0.12998962514278528, "grad_norm": 3.2092316150665283, "learning_rate": 9.7401461151108e-06, "loss": 0.5858, "step": 3101 }, { "epoch": 0.13003154375779424, "grad_norm": 2.0835890769958496, "learning_rate": 9.739930071980262e-06, "loss": 0.5725, "step": 3102 }, { "epoch": 0.1300734623728032, "grad_norm": 2.671968460083008, "learning_rate": 9.739713941475664e-06, "loss": 0.5328, "step": 3103 }, { "epoch": 0.13011538098781217, "grad_norm": 2.062546730041504, "learning_rate": 9.739497723600993e-06, "loss": 0.6415, "step": 3104 }, { "epoch": 0.13015729960282113, "grad_norm": 1.5196621417999268, "learning_rate": 9.739281418360233e-06, "loss": 0.512, "step": 3105 }, { "epoch": 0.1301992182178301, "grad_norm": 1.8887853622436523, "learning_rate": 9.739065025757373e-06, "loss": 0.6279, "step": 3106 }, { "epoch": 0.13024113683283906, "grad_norm": 1.8137600421905518, "learning_rate": 9.7388485457964e-06, "loss": 0.5838, "step": 3107 }, { "epoch": 0.130283055447848, "grad_norm": 1.7857621908187866, "learning_rate": 9.738631978481305e-06, "loss": 0.5274, "step": 3108 }, { "epoch": 0.13032497406285695, "grad_norm": 2.2911033630371094, "learning_rate": 9.738415323816082e-06, "loss": 0.6076, "step": 3109 }, { "epoch": 0.13036689267786591, "grad_norm": 1.6800456047058105, "learning_rate": 9.738198581804721e-06, "loss": 0.5678, "step": 3110 }, { "epoch": 0.13040881129287488, "grad_norm": 2.1579394340515137, "learning_rate": 9.73798175245122e-06, "loss": 0.609, "step": 3111 }, { "epoch": 0.13045072990788384, "grad_norm": 1.8483662605285645, "learning_rate": 9.737764835759577e-06, "loss": 0.6044, "step": 3112 }, { "epoch": 0.1304926485228928, "grad_norm": 1.8999404907226562, "learning_rate": 9.737547831733787e-06, "loss": 0.6011, "step": 3113 }, { "epoch": 0.13053456713790176, "grad_norm": 2.7425873279571533, "learning_rate": 9.737330740377853e-06, "loss": 0.5478, "step": 3114 }, { "epoch": 0.13057648575291073, "grad_norm": 1.6817551851272583, "learning_rate": 9.737113561695776e-06, "loss": 0.5189, "step": 3115 }, { "epoch": 0.1306184043679197, "grad_norm": 1.7951252460479736, "learning_rate": 9.73689629569156e-06, "loss": 0.6333, "step": 3116 }, { "epoch": 0.13066032298292865, "grad_norm": 1.8861000537872314, "learning_rate": 9.736678942369208e-06, "loss": 0.5917, "step": 3117 }, { "epoch": 0.1307022415979376, "grad_norm": 2.2290239334106445, "learning_rate": 9.73646150173273e-06, "loss": 0.6367, "step": 3118 }, { "epoch": 0.13074416021294658, "grad_norm": 1.6920363903045654, "learning_rate": 9.73624397378613e-06, "loss": 0.5999, "step": 3119 }, { "epoch": 0.13078607882795554, "grad_norm": 2.6090986728668213, "learning_rate": 9.736026358533421e-06, "loss": 0.4941, "step": 3120 }, { "epoch": 0.13082799744296447, "grad_norm": 2.255089044570923, "learning_rate": 9.735808655978614e-06, "loss": 0.6087, "step": 3121 }, { "epoch": 0.13086991605797343, "grad_norm": 1.7641245126724243, "learning_rate": 9.735590866125722e-06, "loss": 0.5895, "step": 3122 }, { "epoch": 0.1309118346729824, "grad_norm": 2.0670769214630127, "learning_rate": 9.735372988978757e-06, "loss": 0.5409, "step": 3123 }, { "epoch": 0.13095375328799136, "grad_norm": 1.6958918571472168, "learning_rate": 9.735155024541738e-06, "loss": 0.5758, "step": 3124 }, { "epoch": 0.13099567190300032, "grad_norm": 2.064870595932007, "learning_rate": 9.734936972818684e-06, "loss": 0.5969, "step": 3125 }, { "epoch": 0.13103759051800928, "grad_norm": 1.7761831283569336, "learning_rate": 9.734718833813612e-06, "loss": 0.575, "step": 3126 }, { "epoch": 0.13107950913301825, "grad_norm": 1.7813069820404053, "learning_rate": 9.734500607530542e-06, "loss": 0.5199, "step": 3127 }, { "epoch": 0.1311214277480272, "grad_norm": 1.936158299446106, "learning_rate": 9.734282293973501e-06, "loss": 0.5693, "step": 3128 }, { "epoch": 0.13116334636303617, "grad_norm": 1.8463352918624878, "learning_rate": 9.73406389314651e-06, "loss": 0.5495, "step": 3129 }, { "epoch": 0.13120526497804513, "grad_norm": 1.7540174722671509, "learning_rate": 9.733845405053594e-06, "loss": 0.6189, "step": 3130 }, { "epoch": 0.1312471835930541, "grad_norm": 1.780552864074707, "learning_rate": 9.733626829698784e-06, "loss": 0.5341, "step": 3131 }, { "epoch": 0.13128910220806306, "grad_norm": 1.8082680702209473, "learning_rate": 9.733408167086108e-06, "loss": 0.5579, "step": 3132 }, { "epoch": 0.131331020823072, "grad_norm": 1.6777417659759521, "learning_rate": 9.733189417219595e-06, "loss": 0.538, "step": 3133 }, { "epoch": 0.13137293943808095, "grad_norm": 1.7971482276916504, "learning_rate": 9.73297058010328e-06, "loss": 0.6084, "step": 3134 }, { "epoch": 0.13141485805308992, "grad_norm": 2.6250452995300293, "learning_rate": 9.732751655741194e-06, "loss": 0.559, "step": 3135 }, { "epoch": 0.13145677666809888, "grad_norm": 2.861032247543335, "learning_rate": 9.732532644137375e-06, "loss": 0.5747, "step": 3136 }, { "epoch": 0.13149869528310784, "grad_norm": 2.0679893493652344, "learning_rate": 9.732313545295858e-06, "loss": 0.6421, "step": 3137 }, { "epoch": 0.1315406138981168, "grad_norm": 2.1196651458740234, "learning_rate": 9.732094359220684e-06, "loss": 0.6398, "step": 3138 }, { "epoch": 0.13158253251312577, "grad_norm": 3.716813564300537, "learning_rate": 9.731875085915891e-06, "loss": 0.5616, "step": 3139 }, { "epoch": 0.13162445112813473, "grad_norm": 1.9741626977920532, "learning_rate": 9.731655725385525e-06, "loss": 0.5594, "step": 3140 }, { "epoch": 0.1316663697431437, "grad_norm": 1.7886987924575806, "learning_rate": 9.731436277633625e-06, "loss": 0.5915, "step": 3141 }, { "epoch": 0.13170828835815265, "grad_norm": 1.5417402982711792, "learning_rate": 9.73121674266424e-06, "loss": 0.5624, "step": 3142 }, { "epoch": 0.13175020697316162, "grad_norm": 3.208010673522949, "learning_rate": 9.730997120481413e-06, "loss": 0.578, "step": 3143 }, { "epoch": 0.13179212558817058, "grad_norm": 1.8483617305755615, "learning_rate": 9.730777411089196e-06, "loss": 0.5095, "step": 3144 }, { "epoch": 0.13183404420317954, "grad_norm": 2.1254093647003174, "learning_rate": 9.730557614491638e-06, "loss": 0.6029, "step": 3145 }, { "epoch": 0.13187596281818847, "grad_norm": 2.0486905574798584, "learning_rate": 9.730337730692788e-06, "loss": 0.5599, "step": 3146 }, { "epoch": 0.13191788143319744, "grad_norm": 1.755137324333191, "learning_rate": 9.730117759696702e-06, "loss": 0.5438, "step": 3147 }, { "epoch": 0.1319598000482064, "grad_norm": 1.8799920082092285, "learning_rate": 9.729897701507435e-06, "loss": 0.5213, "step": 3148 }, { "epoch": 0.13200171866321536, "grad_norm": 2.3146450519561768, "learning_rate": 9.729677556129043e-06, "loss": 0.6079, "step": 3149 }, { "epoch": 0.13204363727822432, "grad_norm": 1.6906994581222534, "learning_rate": 9.729457323565583e-06, "loss": 0.5834, "step": 3150 }, { "epoch": 0.1320855558932333, "grad_norm": 1.5573534965515137, "learning_rate": 9.729237003821118e-06, "loss": 0.5693, "step": 3151 }, { "epoch": 0.13212747450824225, "grad_norm": 1.737776756286621, "learning_rate": 9.729016596899703e-06, "loss": 0.6474, "step": 3152 }, { "epoch": 0.1321693931232512, "grad_norm": 1.7663630247116089, "learning_rate": 9.728796102805407e-06, "loss": 0.6257, "step": 3153 }, { "epoch": 0.13221131173826017, "grad_norm": 1.7841845750808716, "learning_rate": 9.728575521542292e-06, "loss": 0.5695, "step": 3154 }, { "epoch": 0.13225323035326914, "grad_norm": 1.9205989837646484, "learning_rate": 9.728354853114423e-06, "loss": 0.5973, "step": 3155 }, { "epoch": 0.1322951489682781, "grad_norm": 1.9529184103012085, "learning_rate": 9.72813409752587e-06, "loss": 0.5676, "step": 3156 }, { "epoch": 0.13233706758328706, "grad_norm": 2.035876989364624, "learning_rate": 9.727913254780701e-06, "loss": 0.5854, "step": 3157 }, { "epoch": 0.132378986198296, "grad_norm": 2.0465104579925537, "learning_rate": 9.727692324882988e-06, "loss": 0.5551, "step": 3158 }, { "epoch": 0.13242090481330496, "grad_norm": 1.8639068603515625, "learning_rate": 9.727471307836803e-06, "loss": 0.5701, "step": 3159 }, { "epoch": 0.13246282342831392, "grad_norm": 1.8082960844039917, "learning_rate": 9.727250203646218e-06, "loss": 0.5894, "step": 3160 }, { "epoch": 0.13250474204332288, "grad_norm": 1.7550055980682373, "learning_rate": 9.727029012315315e-06, "loss": 0.5731, "step": 3161 }, { "epoch": 0.13254666065833184, "grad_norm": 3.026228904724121, "learning_rate": 9.726807733848162e-06, "loss": 0.5548, "step": 3162 }, { "epoch": 0.1325885792733408, "grad_norm": 1.9155807495117188, "learning_rate": 9.726586368248847e-06, "loss": 0.5236, "step": 3163 }, { "epoch": 0.13263049788834977, "grad_norm": 1.7593166828155518, "learning_rate": 9.726364915521445e-06, "loss": 0.5392, "step": 3164 }, { "epoch": 0.13267241650335873, "grad_norm": 1.9085280895233154, "learning_rate": 9.726143375670038e-06, "loss": 0.5562, "step": 3165 }, { "epoch": 0.1327143351183677, "grad_norm": 1.7419580221176147, "learning_rate": 9.725921748698714e-06, "loss": 0.5668, "step": 3166 }, { "epoch": 0.13275625373337666, "grad_norm": 1.6454713344573975, "learning_rate": 9.725700034611556e-06, "loss": 0.5749, "step": 3167 }, { "epoch": 0.13279817234838562, "grad_norm": 1.6557799577713013, "learning_rate": 9.725478233412651e-06, "loss": 0.5219, "step": 3168 }, { "epoch": 0.13284009096339458, "grad_norm": 2.0821707248687744, "learning_rate": 9.725256345106086e-06, "loss": 0.6074, "step": 3169 }, { "epoch": 0.13288200957840354, "grad_norm": 1.7264612913131714, "learning_rate": 9.725034369695953e-06, "loss": 0.5519, "step": 3170 }, { "epoch": 0.13292392819341248, "grad_norm": 1.961167573928833, "learning_rate": 9.724812307186344e-06, "loss": 0.6113, "step": 3171 }, { "epoch": 0.13296584680842144, "grad_norm": 2.149723768234253, "learning_rate": 9.724590157581353e-06, "loss": 0.6179, "step": 3172 }, { "epoch": 0.1330077654234304, "grad_norm": 1.7029083967208862, "learning_rate": 9.724367920885073e-06, "loss": 0.5095, "step": 3173 }, { "epoch": 0.13304968403843936, "grad_norm": 1.677696943283081, "learning_rate": 9.724145597101602e-06, "loss": 0.5925, "step": 3174 }, { "epoch": 0.13309160265344833, "grad_norm": 1.8255693912506104, "learning_rate": 9.723923186235037e-06, "loss": 0.5724, "step": 3175 }, { "epoch": 0.1331335212684573, "grad_norm": 1.7699748277664185, "learning_rate": 9.723700688289479e-06, "loss": 0.5295, "step": 3176 }, { "epoch": 0.13317543988346625, "grad_norm": 1.8096884489059448, "learning_rate": 9.723478103269028e-06, "loss": 0.6024, "step": 3177 }, { "epoch": 0.1332173584984752, "grad_norm": 1.879753589630127, "learning_rate": 9.723255431177791e-06, "loss": 0.6307, "step": 3178 }, { "epoch": 0.13325927711348418, "grad_norm": 1.9407020807266235, "learning_rate": 9.723032672019868e-06, "loss": 0.5935, "step": 3179 }, { "epoch": 0.13330119572849314, "grad_norm": 2.0479981899261475, "learning_rate": 9.722809825799367e-06, "loss": 0.6334, "step": 3180 }, { "epoch": 0.1333431143435021, "grad_norm": 1.802216649055481, "learning_rate": 9.722586892520398e-06, "loss": 0.5608, "step": 3181 }, { "epoch": 0.13338503295851106, "grad_norm": 1.7455003261566162, "learning_rate": 9.722363872187067e-06, "loss": 0.6539, "step": 3182 }, { "epoch": 0.13342695157352, "grad_norm": 1.9148215055465698, "learning_rate": 9.722140764803484e-06, "loss": 0.5766, "step": 3183 }, { "epoch": 0.13346887018852896, "grad_norm": 1.848048210144043, "learning_rate": 9.721917570373766e-06, "loss": 0.5668, "step": 3184 }, { "epoch": 0.13351078880353792, "grad_norm": 1.6717582941055298, "learning_rate": 9.721694288902027e-06, "loss": 0.5679, "step": 3185 }, { "epoch": 0.13355270741854688, "grad_norm": 2.1917779445648193, "learning_rate": 9.72147092039238e-06, "loss": 0.5957, "step": 3186 }, { "epoch": 0.13359462603355585, "grad_norm": 1.6744781732559204, "learning_rate": 9.721247464848943e-06, "loss": 0.5231, "step": 3187 }, { "epoch": 0.1336365446485648, "grad_norm": 1.882699966430664, "learning_rate": 9.721023922275838e-06, "loss": 0.587, "step": 3188 }, { "epoch": 0.13367846326357377, "grad_norm": 1.705216884613037, "learning_rate": 9.720800292677183e-06, "loss": 0.5845, "step": 3189 }, { "epoch": 0.13372038187858273, "grad_norm": 1.9173693656921387, "learning_rate": 9.720576576057101e-06, "loss": 0.5782, "step": 3190 }, { "epoch": 0.1337623004935917, "grad_norm": 2.1291425228118896, "learning_rate": 9.720352772419714e-06, "loss": 0.6152, "step": 3191 }, { "epoch": 0.13380421910860066, "grad_norm": 1.795709490776062, "learning_rate": 9.720128881769152e-06, "loss": 0.5708, "step": 3192 }, { "epoch": 0.13384613772360962, "grad_norm": 1.7067198753356934, "learning_rate": 9.719904904109538e-06, "loss": 0.5503, "step": 3193 }, { "epoch": 0.13388805633861858, "grad_norm": 1.6736772060394287, "learning_rate": 9.719680839445002e-06, "loss": 0.5431, "step": 3194 }, { "epoch": 0.13392997495362755, "grad_norm": 1.625494122505188, "learning_rate": 9.719456687779674e-06, "loss": 0.5049, "step": 3195 }, { "epoch": 0.13397189356863648, "grad_norm": 2.0819709300994873, "learning_rate": 9.719232449117688e-06, "loss": 0.6239, "step": 3196 }, { "epoch": 0.13401381218364544, "grad_norm": 1.9024322032928467, "learning_rate": 9.719008123463175e-06, "loss": 0.5426, "step": 3197 }, { "epoch": 0.1340557307986544, "grad_norm": 2.058396577835083, "learning_rate": 9.718783710820272e-06, "loss": 0.5287, "step": 3198 }, { "epoch": 0.13409764941366337, "grad_norm": 2.13034725189209, "learning_rate": 9.718559211193113e-06, "loss": 0.5553, "step": 3199 }, { "epoch": 0.13413956802867233, "grad_norm": 1.8949050903320312, "learning_rate": 9.718334624585838e-06, "loss": 0.5732, "step": 3200 }, { "epoch": 0.1341814866436813, "grad_norm": 2.9873640537261963, "learning_rate": 9.71810995100259e-06, "loss": 0.6409, "step": 3201 }, { "epoch": 0.13422340525869025, "grad_norm": 2.040231704711914, "learning_rate": 9.717885190447506e-06, "loss": 0.5213, "step": 3202 }, { "epoch": 0.13426532387369922, "grad_norm": 2.067002058029175, "learning_rate": 9.71766034292473e-06, "loss": 0.5981, "step": 3203 }, { "epoch": 0.13430724248870818, "grad_norm": 2.11075496673584, "learning_rate": 9.71743540843841e-06, "loss": 0.5433, "step": 3204 }, { "epoch": 0.13434916110371714, "grad_norm": 2.106882095336914, "learning_rate": 9.717210386992685e-06, "loss": 0.5302, "step": 3205 }, { "epoch": 0.1343910797187261, "grad_norm": 1.6100594997406006, "learning_rate": 9.716985278591712e-06, "loss": 0.5777, "step": 3206 }, { "epoch": 0.13443299833373507, "grad_norm": 2.3237462043762207, "learning_rate": 9.716760083239636e-06, "loss": 0.4989, "step": 3207 }, { "epoch": 0.134474916948744, "grad_norm": 1.8072261810302734, "learning_rate": 9.716534800940608e-06, "loss": 0.5784, "step": 3208 }, { "epoch": 0.13451683556375296, "grad_norm": 1.735785961151123, "learning_rate": 9.716309431698781e-06, "loss": 0.6082, "step": 3209 }, { "epoch": 0.13455875417876192, "grad_norm": 1.8176679611206055, "learning_rate": 9.716083975518311e-06, "loss": 0.5637, "step": 3210 }, { "epoch": 0.1346006727937709, "grad_norm": 3.7723026275634766, "learning_rate": 9.715858432403352e-06, "loss": 0.6144, "step": 3211 }, { "epoch": 0.13464259140877985, "grad_norm": 1.8252512216567993, "learning_rate": 9.715632802358061e-06, "loss": 0.5467, "step": 3212 }, { "epoch": 0.1346845100237888, "grad_norm": 1.8513869047164917, "learning_rate": 9.7154070853866e-06, "loss": 0.5581, "step": 3213 }, { "epoch": 0.13472642863879777, "grad_norm": 1.8632817268371582, "learning_rate": 9.715181281493127e-06, "loss": 0.5988, "step": 3214 }, { "epoch": 0.13476834725380674, "grad_norm": 1.7245522737503052, "learning_rate": 9.714955390681807e-06, "loss": 0.5855, "step": 3215 }, { "epoch": 0.1348102658688157, "grad_norm": 1.7055659294128418, "learning_rate": 9.714729412956803e-06, "loss": 0.5654, "step": 3216 }, { "epoch": 0.13485218448382466, "grad_norm": 1.5302824974060059, "learning_rate": 9.714503348322279e-06, "loss": 0.5153, "step": 3217 }, { "epoch": 0.13489410309883362, "grad_norm": 1.5859378576278687, "learning_rate": 9.714277196782403e-06, "loss": 0.5342, "step": 3218 }, { "epoch": 0.13493602171384259, "grad_norm": 1.718729019165039, "learning_rate": 9.714050958341344e-06, "loss": 0.6043, "step": 3219 }, { "epoch": 0.13497794032885155, "grad_norm": 1.8786120414733887, "learning_rate": 9.713824633003275e-06, "loss": 0.5493, "step": 3220 }, { "epoch": 0.13501985894386048, "grad_norm": 1.9186183214187622, "learning_rate": 9.713598220772362e-06, "loss": 0.624, "step": 3221 }, { "epoch": 0.13506177755886944, "grad_norm": 3.2291836738586426, "learning_rate": 9.713371721652784e-06, "loss": 0.5604, "step": 3222 }, { "epoch": 0.1351036961738784, "grad_norm": 1.610460638999939, "learning_rate": 9.713145135648716e-06, "loss": 0.5461, "step": 3223 }, { "epoch": 0.13514561478888737, "grad_norm": 1.934914231300354, "learning_rate": 9.712918462764332e-06, "loss": 0.5493, "step": 3224 }, { "epoch": 0.13518753340389633, "grad_norm": 1.5843698978424072, "learning_rate": 9.712691703003809e-06, "loss": 0.5393, "step": 3225 }, { "epoch": 0.1352294520189053, "grad_norm": 1.8042651414871216, "learning_rate": 9.712464856371333e-06, "loss": 0.5312, "step": 3226 }, { "epoch": 0.13527137063391426, "grad_norm": 1.7910174131393433, "learning_rate": 9.71223792287108e-06, "loss": 0.5618, "step": 3227 }, { "epoch": 0.13531328924892322, "grad_norm": 3.1424121856689453, "learning_rate": 9.712010902507236e-06, "loss": 0.5501, "step": 3228 }, { "epoch": 0.13535520786393218, "grad_norm": 1.9077908992767334, "learning_rate": 9.711783795283985e-06, "loss": 0.6126, "step": 3229 }, { "epoch": 0.13539712647894114, "grad_norm": 1.77573561668396, "learning_rate": 9.711556601205514e-06, "loss": 0.6458, "step": 3230 }, { "epoch": 0.1354390450939501, "grad_norm": 1.6631518602371216, "learning_rate": 9.711329320276011e-06, "loss": 0.5418, "step": 3231 }, { "epoch": 0.13548096370895907, "grad_norm": 1.7781294584274292, "learning_rate": 9.711101952499664e-06, "loss": 0.6028, "step": 3232 }, { "epoch": 0.135522882323968, "grad_norm": 2.5717437267303467, "learning_rate": 9.710874497880667e-06, "loss": 0.5664, "step": 3233 }, { "epoch": 0.13556480093897696, "grad_norm": 1.7310222387313843, "learning_rate": 9.710646956423211e-06, "loss": 0.551, "step": 3234 }, { "epoch": 0.13560671955398593, "grad_norm": 1.923030972480774, "learning_rate": 9.710419328131488e-06, "loss": 0.635, "step": 3235 }, { "epoch": 0.1356486381689949, "grad_norm": 2.1201982498168945, "learning_rate": 9.7101916130097e-06, "loss": 0.5713, "step": 3236 }, { "epoch": 0.13569055678400385, "grad_norm": 1.7484320402145386, "learning_rate": 9.70996381106204e-06, "loss": 0.5876, "step": 3237 }, { "epoch": 0.13573247539901281, "grad_norm": 1.7829482555389404, "learning_rate": 9.709735922292708e-06, "loss": 0.498, "step": 3238 }, { "epoch": 0.13577439401402178, "grad_norm": 1.777740240097046, "learning_rate": 9.709507946705904e-06, "loss": 0.5862, "step": 3239 }, { "epoch": 0.13581631262903074, "grad_norm": 1.8401448726654053, "learning_rate": 9.709279884305835e-06, "loss": 0.5365, "step": 3240 }, { "epoch": 0.1358582312440397, "grad_norm": 1.742210865020752, "learning_rate": 9.709051735096699e-06, "loss": 0.5682, "step": 3241 }, { "epoch": 0.13590014985904866, "grad_norm": 1.9685345888137817, "learning_rate": 9.708823499082708e-06, "loss": 0.582, "step": 3242 }, { "epoch": 0.13594206847405763, "grad_norm": 1.7732117176055908, "learning_rate": 9.708595176268061e-06, "loss": 0.6253, "step": 3243 }, { "epoch": 0.1359839870890666, "grad_norm": 1.7973936796188354, "learning_rate": 9.708366766656973e-06, "loss": 0.5701, "step": 3244 }, { "epoch": 0.13602590570407555, "grad_norm": 1.8074055910110474, "learning_rate": 9.708138270253653e-06, "loss": 0.5778, "step": 3245 }, { "epoch": 0.13606782431908448, "grad_norm": 2.2062976360321045, "learning_rate": 9.707909687062314e-06, "loss": 0.5336, "step": 3246 }, { "epoch": 0.13610974293409345, "grad_norm": 2.2119550704956055, "learning_rate": 9.707681017087167e-06, "loss": 0.5845, "step": 3247 }, { "epoch": 0.1361516615491024, "grad_norm": 1.8054003715515137, "learning_rate": 9.707452260332429e-06, "loss": 0.5436, "step": 3248 }, { "epoch": 0.13619358016411137, "grad_norm": 1.9524827003479004, "learning_rate": 9.707223416802317e-06, "loss": 0.5648, "step": 3249 }, { "epoch": 0.13623549877912033, "grad_norm": 1.9937647581100464, "learning_rate": 9.706994486501046e-06, "loss": 0.5452, "step": 3250 }, { "epoch": 0.1362774173941293, "grad_norm": 1.9128153324127197, "learning_rate": 9.706765469432841e-06, "loss": 0.5056, "step": 3251 }, { "epoch": 0.13631933600913826, "grad_norm": 1.7778875827789307, "learning_rate": 9.706536365601922e-06, "loss": 0.468, "step": 3252 }, { "epoch": 0.13636125462414722, "grad_norm": 1.799338459968567, "learning_rate": 9.706307175012512e-06, "loss": 0.5737, "step": 3253 }, { "epoch": 0.13640317323915618, "grad_norm": 2.1910557746887207, "learning_rate": 9.706077897668835e-06, "loss": 0.6032, "step": 3254 }, { "epoch": 0.13644509185416515, "grad_norm": 1.7042335271835327, "learning_rate": 9.705848533575118e-06, "loss": 0.5151, "step": 3255 }, { "epoch": 0.1364870104691741, "grad_norm": 1.8835315704345703, "learning_rate": 9.705619082735587e-06, "loss": 0.6164, "step": 3256 }, { "epoch": 0.13652892908418307, "grad_norm": 1.6321007013320923, "learning_rate": 9.705389545154476e-06, "loss": 0.5325, "step": 3257 }, { "epoch": 0.136570847699192, "grad_norm": 1.6504197120666504, "learning_rate": 9.705159920836013e-06, "loss": 0.5786, "step": 3258 }, { "epoch": 0.13661276631420097, "grad_norm": 2.0049517154693604, "learning_rate": 9.704930209784431e-06, "loss": 0.5392, "step": 3259 }, { "epoch": 0.13665468492920993, "grad_norm": 2.1782610416412354, "learning_rate": 9.704700412003965e-06, "loss": 0.567, "step": 3260 }, { "epoch": 0.1366966035442189, "grad_norm": 1.7781261205673218, "learning_rate": 9.704470527498851e-06, "loss": 0.544, "step": 3261 }, { "epoch": 0.13673852215922785, "grad_norm": 1.9830327033996582, "learning_rate": 9.704240556273329e-06, "loss": 0.5432, "step": 3262 }, { "epoch": 0.13678044077423682, "grad_norm": 1.7458699941635132, "learning_rate": 9.704010498331632e-06, "loss": 0.5941, "step": 3263 }, { "epoch": 0.13682235938924578, "grad_norm": 1.7874451875686646, "learning_rate": 9.703780353678006e-06, "loss": 0.5585, "step": 3264 }, { "epoch": 0.13686427800425474, "grad_norm": 3.1535820960998535, "learning_rate": 9.703550122316692e-06, "loss": 0.528, "step": 3265 }, { "epoch": 0.1369061966192637, "grad_norm": 1.5406235456466675, "learning_rate": 9.703319804251933e-06, "loss": 0.5457, "step": 3266 }, { "epoch": 0.13694811523427267, "grad_norm": 1.7732336521148682, "learning_rate": 9.703089399487979e-06, "loss": 0.5498, "step": 3267 }, { "epoch": 0.13699003384928163, "grad_norm": 1.7706823348999023, "learning_rate": 9.70285890802907e-06, "loss": 0.5332, "step": 3268 }, { "epoch": 0.1370319524642906, "grad_norm": 2.520169973373413, "learning_rate": 9.70262832987946e-06, "loss": 0.595, "step": 3269 }, { "epoch": 0.13707387107929955, "grad_norm": 2.262777328491211, "learning_rate": 9.702397665043398e-06, "loss": 0.5376, "step": 3270 }, { "epoch": 0.1371157896943085, "grad_norm": 2.0497665405273438, "learning_rate": 9.702166913525136e-06, "loss": 0.537, "step": 3271 }, { "epoch": 0.13715770830931745, "grad_norm": 1.6003211736679077, "learning_rate": 9.701936075328927e-06, "loss": 0.5613, "step": 3272 }, { "epoch": 0.1371996269243264, "grad_norm": 1.8617054224014282, "learning_rate": 9.701705150459026e-06, "loss": 0.6222, "step": 3273 }, { "epoch": 0.13724154553933537, "grad_norm": 1.8095499277114868, "learning_rate": 9.701474138919692e-06, "loss": 0.5636, "step": 3274 }, { "epoch": 0.13728346415434434, "grad_norm": 1.686944842338562, "learning_rate": 9.701243040715182e-06, "loss": 0.5132, "step": 3275 }, { "epoch": 0.1373253827693533, "grad_norm": 1.844229817390442, "learning_rate": 9.701011855849756e-06, "loss": 0.5513, "step": 3276 }, { "epoch": 0.13736730138436226, "grad_norm": 1.977428674697876, "learning_rate": 9.700780584327673e-06, "loss": 0.5424, "step": 3277 }, { "epoch": 0.13740921999937122, "grad_norm": 1.5938420295715332, "learning_rate": 9.7005492261532e-06, "loss": 0.5016, "step": 3278 }, { "epoch": 0.13745113861438019, "grad_norm": 1.8635114431381226, "learning_rate": 9.700317781330601e-06, "loss": 0.5729, "step": 3279 }, { "epoch": 0.13749305722938915, "grad_norm": 1.7827939987182617, "learning_rate": 9.70008624986414e-06, "loss": 0.6288, "step": 3280 }, { "epoch": 0.1375349758443981, "grad_norm": 1.7929725646972656, "learning_rate": 9.699854631758088e-06, "loss": 0.5362, "step": 3281 }, { "epoch": 0.13757689445940707, "grad_norm": 1.748603343963623, "learning_rate": 9.699622927016715e-06, "loss": 0.6012, "step": 3282 }, { "epoch": 0.137618813074416, "grad_norm": 2.349348783493042, "learning_rate": 9.699391135644288e-06, "loss": 0.5992, "step": 3283 }, { "epoch": 0.13766073168942497, "grad_norm": 1.7545863389968872, "learning_rate": 9.699159257645081e-06, "loss": 0.5459, "step": 3284 }, { "epoch": 0.13770265030443393, "grad_norm": 1.9772164821624756, "learning_rate": 9.69892729302337e-06, "loss": 0.5896, "step": 3285 }, { "epoch": 0.1377445689194429, "grad_norm": 2.0955986976623535, "learning_rate": 9.698695241783433e-06, "loss": 0.5709, "step": 3286 }, { "epoch": 0.13778648753445186, "grad_norm": 1.7126576900482178, "learning_rate": 9.698463103929542e-06, "loss": 0.5422, "step": 3287 }, { "epoch": 0.13782840614946082, "grad_norm": 1.7039211988449097, "learning_rate": 9.698230879465981e-06, "loss": 0.5256, "step": 3288 }, { "epoch": 0.13787032476446978, "grad_norm": 2.073838472366333, "learning_rate": 9.69799856839703e-06, "loss": 0.6023, "step": 3289 }, { "epoch": 0.13791224337947874, "grad_norm": 2.1183438301086426, "learning_rate": 9.697766170726967e-06, "loss": 0.5489, "step": 3290 }, { "epoch": 0.1379541619944877, "grad_norm": 1.6447057723999023, "learning_rate": 9.697533686460082e-06, "loss": 0.5595, "step": 3291 }, { "epoch": 0.13799608060949667, "grad_norm": 1.7690147161483765, "learning_rate": 9.697301115600655e-06, "loss": 0.5866, "step": 3292 }, { "epoch": 0.13803799922450563, "grad_norm": 2.0104153156280518, "learning_rate": 9.697068458152975e-06, "loss": 0.5278, "step": 3293 }, { "epoch": 0.1380799178395146, "grad_norm": 1.9843510389328003, "learning_rate": 9.696835714121336e-06, "loss": 0.5567, "step": 3294 }, { "epoch": 0.13812183645452356, "grad_norm": 1.794950008392334, "learning_rate": 9.69660288351002e-06, "loss": 0.5932, "step": 3295 }, { "epoch": 0.1381637550695325, "grad_norm": 1.9082129001617432, "learning_rate": 9.696369966323324e-06, "loss": 0.5711, "step": 3296 }, { "epoch": 0.13820567368454145, "grad_norm": 1.7088311910629272, "learning_rate": 9.696136962565541e-06, "loss": 0.6021, "step": 3297 }, { "epoch": 0.13824759229955041, "grad_norm": 2.877138614654541, "learning_rate": 9.695903872240964e-06, "loss": 0.6022, "step": 3298 }, { "epoch": 0.13828951091455938, "grad_norm": 2.0739858150482178, "learning_rate": 9.69567069535389e-06, "loss": 0.5553, "step": 3299 }, { "epoch": 0.13833142952956834, "grad_norm": 1.6777681112289429, "learning_rate": 9.69543743190862e-06, "loss": 0.5453, "step": 3300 }, { "epoch": 0.1383733481445773, "grad_norm": 2.2380971908569336, "learning_rate": 9.695204081909452e-06, "loss": 0.5663, "step": 3301 }, { "epoch": 0.13841526675958626, "grad_norm": 1.8156709671020508, "learning_rate": 9.694970645360688e-06, "loss": 0.586, "step": 3302 }, { "epoch": 0.13845718537459523, "grad_norm": 2.1146433353424072, "learning_rate": 9.694737122266631e-06, "loss": 0.554, "step": 3303 }, { "epoch": 0.1384991039896042, "grad_norm": 1.7023519277572632, "learning_rate": 9.694503512631585e-06, "loss": 0.5051, "step": 3304 }, { "epoch": 0.13854102260461315, "grad_norm": 1.6911852359771729, "learning_rate": 9.694269816459855e-06, "loss": 0.595, "step": 3305 }, { "epoch": 0.1385829412196221, "grad_norm": 1.9216856956481934, "learning_rate": 9.694036033755755e-06, "loss": 0.5088, "step": 3306 }, { "epoch": 0.13862485983463108, "grad_norm": 3.134119987487793, "learning_rate": 9.693802164523584e-06, "loss": 0.5301, "step": 3307 }, { "epoch": 0.13866677844964, "grad_norm": 2.1430089473724365, "learning_rate": 9.693568208767662e-06, "loss": 0.5846, "step": 3308 }, { "epoch": 0.13870869706464897, "grad_norm": 1.843319058418274, "learning_rate": 9.6933341664923e-06, "loss": 0.5465, "step": 3309 }, { "epoch": 0.13875061567965793, "grad_norm": 1.7816420793533325, "learning_rate": 9.69310003770181e-06, "loss": 0.466, "step": 3310 }, { "epoch": 0.1387925342946669, "grad_norm": 1.7367109060287476, "learning_rate": 9.692865822400508e-06, "loss": 0.634, "step": 3311 }, { "epoch": 0.13883445290967586, "grad_norm": 1.870204210281372, "learning_rate": 9.692631520592713e-06, "loss": 0.5827, "step": 3312 }, { "epoch": 0.13887637152468482, "grad_norm": 1.7709780931472778, "learning_rate": 9.692397132282744e-06, "loss": 0.5369, "step": 3313 }, { "epoch": 0.13891829013969378, "grad_norm": 1.7808109521865845, "learning_rate": 9.69216265747492e-06, "loss": 0.5633, "step": 3314 }, { "epoch": 0.13896020875470275, "grad_norm": 1.7876694202423096, "learning_rate": 9.691928096173564e-06, "loss": 0.5393, "step": 3315 }, { "epoch": 0.1390021273697117, "grad_norm": 1.8825676441192627, "learning_rate": 9.691693448383001e-06, "loss": 0.6171, "step": 3316 }, { "epoch": 0.13904404598472067, "grad_norm": 1.7280499935150146, "learning_rate": 9.691458714107554e-06, "loss": 0.5507, "step": 3317 }, { "epoch": 0.13908596459972963, "grad_norm": 1.8566126823425293, "learning_rate": 9.69122389335155e-06, "loss": 0.5336, "step": 3318 }, { "epoch": 0.1391278832147386, "grad_norm": 1.704055905342102, "learning_rate": 9.690988986119323e-06, "loss": 0.5609, "step": 3319 }, { "epoch": 0.13916980182974756, "grad_norm": 1.73953115940094, "learning_rate": 9.690753992415196e-06, "loss": 0.5515, "step": 3320 }, { "epoch": 0.1392117204447565, "grad_norm": 1.8742599487304688, "learning_rate": 9.690518912243506e-06, "loss": 0.549, "step": 3321 }, { "epoch": 0.13925363905976545, "grad_norm": 1.8238372802734375, "learning_rate": 9.690283745608583e-06, "loss": 0.6123, "step": 3322 }, { "epoch": 0.13929555767477442, "grad_norm": 1.8427987098693848, "learning_rate": 9.690048492514763e-06, "loss": 0.5292, "step": 3323 }, { "epoch": 0.13933747628978338, "grad_norm": 1.710721731185913, "learning_rate": 9.689813152966384e-06, "loss": 0.5361, "step": 3324 }, { "epoch": 0.13937939490479234, "grad_norm": 1.8859833478927612, "learning_rate": 9.689577726967782e-06, "loss": 0.5409, "step": 3325 }, { "epoch": 0.1394213135198013, "grad_norm": 1.9331237077713013, "learning_rate": 9.689342214523298e-06, "loss": 0.5553, "step": 3326 }, { "epoch": 0.13946323213481027, "grad_norm": 1.6724435091018677, "learning_rate": 9.689106615637273e-06, "loss": 0.5482, "step": 3327 }, { "epoch": 0.13950515074981923, "grad_norm": 2.539004325866699, "learning_rate": 9.688870930314051e-06, "loss": 0.6293, "step": 3328 }, { "epoch": 0.1395470693648282, "grad_norm": 3.267641305923462, "learning_rate": 9.688635158557976e-06, "loss": 0.5993, "step": 3329 }, { "epoch": 0.13958898797983715, "grad_norm": 2.8938958644866943, "learning_rate": 9.688399300373391e-06, "loss": 0.5454, "step": 3330 }, { "epoch": 0.13963090659484612, "grad_norm": 1.7398037910461426, "learning_rate": 9.688163355764648e-06, "loss": 0.5463, "step": 3331 }, { "epoch": 0.13967282520985508, "grad_norm": 1.6879900693893433, "learning_rate": 9.687927324736096e-06, "loss": 0.5825, "step": 3332 }, { "epoch": 0.139714743824864, "grad_norm": 3.6873912811279297, "learning_rate": 9.687691207292082e-06, "loss": 0.6064, "step": 3333 }, { "epoch": 0.13975666243987298, "grad_norm": 1.8174068927764893, "learning_rate": 9.687455003436963e-06, "loss": 0.557, "step": 3334 }, { "epoch": 0.13979858105488194, "grad_norm": 1.937123417854309, "learning_rate": 9.68721871317509e-06, "loss": 0.6233, "step": 3335 }, { "epoch": 0.1398404996698909, "grad_norm": 1.8606948852539062, "learning_rate": 9.686982336510821e-06, "loss": 0.5322, "step": 3336 }, { "epoch": 0.13988241828489986, "grad_norm": 1.894489049911499, "learning_rate": 9.686745873448511e-06, "loss": 0.5593, "step": 3337 }, { "epoch": 0.13992433689990882, "grad_norm": 1.8264843225479126, "learning_rate": 9.68650932399252e-06, "loss": 0.5435, "step": 3338 }, { "epoch": 0.1399662555149178, "grad_norm": 1.988141417503357, "learning_rate": 9.686272688147211e-06, "loss": 0.5485, "step": 3339 }, { "epoch": 0.14000817412992675, "grad_norm": 1.484878420829773, "learning_rate": 9.686035965916941e-06, "loss": 0.5164, "step": 3340 }, { "epoch": 0.1400500927449357, "grad_norm": 2.0224573612213135, "learning_rate": 9.685799157306077e-06, "loss": 0.6032, "step": 3341 }, { "epoch": 0.14009201135994467, "grad_norm": 1.8011258840560913, "learning_rate": 9.685562262318984e-06, "loss": 0.5883, "step": 3342 }, { "epoch": 0.14013392997495364, "grad_norm": 1.874233603477478, "learning_rate": 9.685325280960028e-06, "loss": 0.5356, "step": 3343 }, { "epoch": 0.1401758485899626, "grad_norm": 2.1995160579681396, "learning_rate": 9.685088213233577e-06, "loss": 0.5349, "step": 3344 }, { "epoch": 0.14021776720497156, "grad_norm": 1.8259581327438354, "learning_rate": 9.684851059144003e-06, "loss": 0.5672, "step": 3345 }, { "epoch": 0.1402596858199805, "grad_norm": 1.9458065032958984, "learning_rate": 9.684613818695674e-06, "loss": 0.4899, "step": 3346 }, { "epoch": 0.14030160443498946, "grad_norm": 1.994893193244934, "learning_rate": 9.684376491892969e-06, "loss": 0.5636, "step": 3347 }, { "epoch": 0.14034352304999842, "grad_norm": 1.7986769676208496, "learning_rate": 9.684139078740257e-06, "loss": 0.5693, "step": 3348 }, { "epoch": 0.14038544166500738, "grad_norm": 1.7440136671066284, "learning_rate": 9.683901579241917e-06, "loss": 0.5608, "step": 3349 }, { "epoch": 0.14042736028001634, "grad_norm": 1.8122841119766235, "learning_rate": 9.683663993402327e-06, "loss": 0.5538, "step": 3350 }, { "epoch": 0.1404692788950253, "grad_norm": 2.027712345123291, "learning_rate": 9.683426321225867e-06, "loss": 0.6147, "step": 3351 }, { "epoch": 0.14051119751003427, "grad_norm": 1.857004165649414, "learning_rate": 9.683188562716916e-06, "loss": 0.5509, "step": 3352 }, { "epoch": 0.14055311612504323, "grad_norm": 1.7979317903518677, "learning_rate": 9.682950717879857e-06, "loss": 0.5731, "step": 3353 }, { "epoch": 0.1405950347400522, "grad_norm": 4.865940093994141, "learning_rate": 9.682712786719078e-06, "loss": 0.5587, "step": 3354 }, { "epoch": 0.14063695335506116, "grad_norm": 2.046025276184082, "learning_rate": 9.682474769238963e-06, "loss": 0.6214, "step": 3355 }, { "epoch": 0.14067887197007012, "grad_norm": 1.7579965591430664, "learning_rate": 9.682236665443897e-06, "loss": 0.5551, "step": 3356 }, { "epoch": 0.14072079058507908, "grad_norm": 1.8393503427505493, "learning_rate": 9.68199847533827e-06, "loss": 0.5922, "step": 3357 }, { "epoch": 0.14076270920008802, "grad_norm": 1.8045613765716553, "learning_rate": 9.681760198926477e-06, "loss": 0.572, "step": 3358 }, { "epoch": 0.14080462781509698, "grad_norm": 1.8726478815078735, "learning_rate": 9.681521836212906e-06, "loss": 0.5741, "step": 3359 }, { "epoch": 0.14084654643010594, "grad_norm": 1.8025504350662231, "learning_rate": 9.68128338720195e-06, "loss": 0.5622, "step": 3360 }, { "epoch": 0.1408884650451149, "grad_norm": 5.3298797607421875, "learning_rate": 9.681044851898008e-06, "loss": 0.5521, "step": 3361 }, { "epoch": 0.14093038366012386, "grad_norm": 2.308681011199951, "learning_rate": 9.680806230305475e-06, "loss": 0.5452, "step": 3362 }, { "epoch": 0.14097230227513283, "grad_norm": 2.0295214653015137, "learning_rate": 9.680567522428753e-06, "loss": 0.5482, "step": 3363 }, { "epoch": 0.1410142208901418, "grad_norm": 1.7534371614456177, "learning_rate": 9.680328728272234e-06, "loss": 0.5943, "step": 3364 }, { "epoch": 0.14105613950515075, "grad_norm": 1.776949167251587, "learning_rate": 9.680089847840331e-06, "loss": 0.5519, "step": 3365 }, { "epoch": 0.14109805812015971, "grad_norm": 1.9026451110839844, "learning_rate": 9.679850881137438e-06, "loss": 0.5375, "step": 3366 }, { "epoch": 0.14113997673516868, "grad_norm": 1.7737637758255005, "learning_rate": 9.679611828167965e-06, "loss": 0.574, "step": 3367 }, { "epoch": 0.14118189535017764, "grad_norm": 3.183068037033081, "learning_rate": 9.679372688936315e-06, "loss": 0.6022, "step": 3368 }, { "epoch": 0.1412238139651866, "grad_norm": 1.7377148866653442, "learning_rate": 9.679133463446902e-06, "loss": 0.5981, "step": 3369 }, { "epoch": 0.14126573258019556, "grad_norm": 1.7039201259613037, "learning_rate": 9.67889415170413e-06, "loss": 0.5597, "step": 3370 }, { "epoch": 0.1413076511952045, "grad_norm": 1.9547739028930664, "learning_rate": 9.678654753712414e-06, "loss": 0.6081, "step": 3371 }, { "epoch": 0.14134956981021346, "grad_norm": 1.8252980709075928, "learning_rate": 9.678415269476164e-06, "loss": 0.5467, "step": 3372 }, { "epoch": 0.14139148842522242, "grad_norm": 1.8613494634628296, "learning_rate": 9.6781756989998e-06, "loss": 0.5854, "step": 3373 }, { "epoch": 0.14143340704023138, "grad_norm": 2.0324041843414307, "learning_rate": 9.67793604228773e-06, "loss": 0.6036, "step": 3374 }, { "epoch": 0.14147532565524035, "grad_norm": 1.5237269401550293, "learning_rate": 9.677696299344377e-06, "loss": 0.5305, "step": 3375 }, { "epoch": 0.1415172442702493, "grad_norm": 1.8799983263015747, "learning_rate": 9.677456470174162e-06, "loss": 0.631, "step": 3376 }, { "epoch": 0.14155916288525827, "grad_norm": 1.9462809562683105, "learning_rate": 9.677216554781501e-06, "loss": 0.5874, "step": 3377 }, { "epoch": 0.14160108150026723, "grad_norm": 1.6475831270217896, "learning_rate": 9.67697655317082e-06, "loss": 0.5129, "step": 3378 }, { "epoch": 0.1416430001152762, "grad_norm": 1.8504962921142578, "learning_rate": 9.676736465346542e-06, "loss": 0.5826, "step": 3379 }, { "epoch": 0.14168491873028516, "grad_norm": 2.106907367706299, "learning_rate": 9.676496291313091e-06, "loss": 0.5464, "step": 3380 }, { "epoch": 0.14172683734529412, "grad_norm": 1.6453380584716797, "learning_rate": 9.676256031074898e-06, "loss": 0.5683, "step": 3381 }, { "epoch": 0.14176875596030308, "grad_norm": 1.8224457502365112, "learning_rate": 9.676015684636388e-06, "loss": 0.6216, "step": 3382 }, { "epoch": 0.14181067457531202, "grad_norm": 1.5697975158691406, "learning_rate": 9.675775252001993e-06, "loss": 0.5233, "step": 3383 }, { "epoch": 0.14185259319032098, "grad_norm": 3.0077760219573975, "learning_rate": 9.675534733176148e-06, "loss": 0.5698, "step": 3384 }, { "epoch": 0.14189451180532994, "grad_norm": 1.7936311960220337, "learning_rate": 9.675294128163282e-06, "loss": 0.5832, "step": 3385 }, { "epoch": 0.1419364304203389, "grad_norm": 2.033831834793091, "learning_rate": 9.675053436967833e-06, "loss": 0.5812, "step": 3386 }, { "epoch": 0.14197834903534787, "grad_norm": 1.7505502700805664, "learning_rate": 9.674812659594236e-06, "loss": 0.5197, "step": 3387 }, { "epoch": 0.14202026765035683, "grad_norm": 1.7949280738830566, "learning_rate": 9.67457179604693e-06, "loss": 0.6236, "step": 3388 }, { "epoch": 0.1420621862653658, "grad_norm": 1.6763081550598145, "learning_rate": 9.674330846330356e-06, "loss": 0.4846, "step": 3389 }, { "epoch": 0.14210410488037475, "grad_norm": 1.745505928993225, "learning_rate": 9.674089810448955e-06, "loss": 0.61, "step": 3390 }, { "epoch": 0.14214602349538372, "grad_norm": 1.5829253196716309, "learning_rate": 9.67384868840717e-06, "loss": 0.5516, "step": 3391 }, { "epoch": 0.14218794211039268, "grad_norm": 1.6481269598007202, "learning_rate": 9.673607480209447e-06, "loss": 0.5447, "step": 3392 }, { "epoch": 0.14222986072540164, "grad_norm": 1.7765100002288818, "learning_rate": 9.67336618586023e-06, "loss": 0.612, "step": 3393 }, { "epoch": 0.1422717793404106, "grad_norm": 1.760526418685913, "learning_rate": 9.673124805363967e-06, "loss": 0.5587, "step": 3394 }, { "epoch": 0.14231369795541957, "grad_norm": 1.919338583946228, "learning_rate": 9.672883338725112e-06, "loss": 0.5517, "step": 3395 }, { "epoch": 0.1423556165704285, "grad_norm": 1.8642282485961914, "learning_rate": 9.67264178594811e-06, "loss": 0.5839, "step": 3396 }, { "epoch": 0.14239753518543746, "grad_norm": 1.6332749128341675, "learning_rate": 9.672400147037416e-06, "loss": 0.5579, "step": 3397 }, { "epoch": 0.14243945380044643, "grad_norm": 2.0286548137664795, "learning_rate": 9.672158421997486e-06, "loss": 0.5762, "step": 3398 }, { "epoch": 0.1424813724154554, "grad_norm": 1.8249307870864868, "learning_rate": 9.671916610832775e-06, "loss": 0.5418, "step": 3399 }, { "epoch": 0.14252329103046435, "grad_norm": 5.349658489227295, "learning_rate": 9.671674713547739e-06, "loss": 0.6063, "step": 3400 }, { "epoch": 0.1425652096454733, "grad_norm": 1.7578271627426147, "learning_rate": 9.671432730146838e-06, "loss": 0.5505, "step": 3401 }, { "epoch": 0.14260712826048227, "grad_norm": 1.7087920904159546, "learning_rate": 9.671190660634532e-06, "loss": 0.5926, "step": 3402 }, { "epoch": 0.14264904687549124, "grad_norm": 1.5656062364578247, "learning_rate": 9.670948505015286e-06, "loss": 0.5177, "step": 3403 }, { "epoch": 0.1426909654905002, "grad_norm": 2.457156181335449, "learning_rate": 9.670706263293561e-06, "loss": 0.5849, "step": 3404 }, { "epoch": 0.14273288410550916, "grad_norm": 1.9770833253860474, "learning_rate": 9.670463935473821e-06, "loss": 0.5537, "step": 3405 }, { "epoch": 0.14277480272051812, "grad_norm": 1.8924264907836914, "learning_rate": 9.670221521560537e-06, "loss": 0.5744, "step": 3406 }, { "epoch": 0.14281672133552709, "grad_norm": 2.2290537357330322, "learning_rate": 9.669979021558176e-06, "loss": 0.6043, "step": 3407 }, { "epoch": 0.14285863995053602, "grad_norm": 2.0515027046203613, "learning_rate": 9.669736435471207e-06, "loss": 0.5966, "step": 3408 }, { "epoch": 0.14290055856554498, "grad_norm": 2.119157314300537, "learning_rate": 9.669493763304103e-06, "loss": 0.5833, "step": 3409 }, { "epoch": 0.14294247718055395, "grad_norm": 1.7953810691833496, "learning_rate": 9.669251005061337e-06, "loss": 0.573, "step": 3410 }, { "epoch": 0.1429843957955629, "grad_norm": 2.1683566570281982, "learning_rate": 9.669008160747382e-06, "loss": 0.5556, "step": 3411 }, { "epoch": 0.14302631441057187, "grad_norm": 2.012903928756714, "learning_rate": 9.66876523036672e-06, "loss": 0.645, "step": 3412 }, { "epoch": 0.14306823302558083, "grad_norm": 1.7457125186920166, "learning_rate": 9.668522213923821e-06, "loss": 0.5288, "step": 3413 }, { "epoch": 0.1431101516405898, "grad_norm": 1.8808157444000244, "learning_rate": 9.66827911142317e-06, "loss": 0.6029, "step": 3414 }, { "epoch": 0.14315207025559876, "grad_norm": 1.9658116102218628, "learning_rate": 9.66803592286925e-06, "loss": 0.5888, "step": 3415 }, { "epoch": 0.14319398887060772, "grad_norm": 2.061899185180664, "learning_rate": 9.667792648266538e-06, "loss": 0.6132, "step": 3416 }, { "epoch": 0.14323590748561668, "grad_norm": 1.7899971008300781, "learning_rate": 9.667549287619525e-06, "loss": 0.5799, "step": 3417 }, { "epoch": 0.14327782610062564, "grad_norm": 2.0810794830322266, "learning_rate": 9.66730584093269e-06, "loss": 0.5484, "step": 3418 }, { "epoch": 0.1433197447156346, "grad_norm": 2.1277339458465576, "learning_rate": 9.667062308210526e-06, "loss": 0.6145, "step": 3419 }, { "epoch": 0.14336166333064357, "grad_norm": 1.6827799081802368, "learning_rate": 9.66681868945752e-06, "loss": 0.5291, "step": 3420 }, { "epoch": 0.1434035819456525, "grad_norm": 1.6877015829086304, "learning_rate": 9.666574984678163e-06, "loss": 0.5742, "step": 3421 }, { "epoch": 0.14344550056066147, "grad_norm": 1.7820061445236206, "learning_rate": 9.666331193876949e-06, "loss": 0.5494, "step": 3422 }, { "epoch": 0.14348741917567043, "grad_norm": 1.7635762691497803, "learning_rate": 9.666087317058367e-06, "loss": 0.5693, "step": 3423 }, { "epoch": 0.1435293377906794, "grad_norm": 2.9195022583007812, "learning_rate": 9.66584335422692e-06, "loss": 0.5362, "step": 3424 }, { "epoch": 0.14357125640568835, "grad_norm": 1.9630411863327026, "learning_rate": 9.665599305387099e-06, "loss": 0.6132, "step": 3425 }, { "epoch": 0.14361317502069731, "grad_norm": 1.7298252582550049, "learning_rate": 9.665355170543405e-06, "loss": 0.5509, "step": 3426 }, { "epoch": 0.14365509363570628, "grad_norm": 1.742079257965088, "learning_rate": 9.665110949700338e-06, "loss": 0.5275, "step": 3427 }, { "epoch": 0.14369701225071524, "grad_norm": 1.8751085996627808, "learning_rate": 9.6648666428624e-06, "loss": 0.5367, "step": 3428 }, { "epoch": 0.1437389308657242, "grad_norm": 1.8112943172454834, "learning_rate": 9.664622250034095e-06, "loss": 0.5865, "step": 3429 }, { "epoch": 0.14378084948073316, "grad_norm": 1.8476240634918213, "learning_rate": 9.664377771219926e-06, "loss": 0.5604, "step": 3430 }, { "epoch": 0.14382276809574213, "grad_norm": 2.8335351943969727, "learning_rate": 9.664133206424402e-06, "loss": 0.5356, "step": 3431 }, { "epoch": 0.1438646867107511, "grad_norm": 2.5594241619110107, "learning_rate": 9.66388855565203e-06, "loss": 0.5682, "step": 3432 }, { "epoch": 0.14390660532576002, "grad_norm": 1.7954436540603638, "learning_rate": 9.663643818907321e-06, "loss": 0.5894, "step": 3433 }, { "epoch": 0.14394852394076899, "grad_norm": 1.6763365268707275, "learning_rate": 9.663398996194784e-06, "loss": 0.5822, "step": 3434 }, { "epoch": 0.14399044255577795, "grad_norm": 1.5884041786193848, "learning_rate": 9.663154087518936e-06, "loss": 0.5831, "step": 3435 }, { "epoch": 0.1440323611707869, "grad_norm": 1.655324101448059, "learning_rate": 9.662909092884288e-06, "loss": 0.5188, "step": 3436 }, { "epoch": 0.14407427978579587, "grad_norm": 1.91289484500885, "learning_rate": 9.662664012295355e-06, "loss": 0.5573, "step": 3437 }, { "epoch": 0.14411619840080483, "grad_norm": 2.047846555709839, "learning_rate": 9.662418845756658e-06, "loss": 0.5761, "step": 3438 }, { "epoch": 0.1441581170158138, "grad_norm": 1.6425552368164062, "learning_rate": 9.662173593272716e-06, "loss": 0.5515, "step": 3439 }, { "epoch": 0.14420003563082276, "grad_norm": 2.1643805503845215, "learning_rate": 9.66192825484805e-06, "loss": 0.5598, "step": 3440 }, { "epoch": 0.14424195424583172, "grad_norm": 1.8243417739868164, "learning_rate": 9.66168283048718e-06, "loss": 0.5502, "step": 3441 }, { "epoch": 0.14428387286084068, "grad_norm": 1.7446379661560059, "learning_rate": 9.66143732019463e-06, "loss": 0.5285, "step": 3442 }, { "epoch": 0.14432579147584965, "grad_norm": 2.031649112701416, "learning_rate": 9.66119172397493e-06, "loss": 0.5809, "step": 3443 }, { "epoch": 0.1443677100908586, "grad_norm": 1.5875380039215088, "learning_rate": 9.660946041832604e-06, "loss": 0.519, "step": 3444 }, { "epoch": 0.14440962870586757, "grad_norm": 1.7789918184280396, "learning_rate": 9.66070027377218e-06, "loss": 0.5368, "step": 3445 }, { "epoch": 0.1444515473208765, "grad_norm": 1.9997907876968384, "learning_rate": 9.660454419798191e-06, "loss": 0.565, "step": 3446 }, { "epoch": 0.14449346593588547, "grad_norm": 2.3361074924468994, "learning_rate": 9.660208479915168e-06, "loss": 0.5605, "step": 3447 }, { "epoch": 0.14453538455089443, "grad_norm": 1.8785983324050903, "learning_rate": 9.659962454127643e-06, "loss": 0.5916, "step": 3448 }, { "epoch": 0.1445773031659034, "grad_norm": 1.9363465309143066, "learning_rate": 9.659716342440152e-06, "loss": 0.5504, "step": 3449 }, { "epoch": 0.14461922178091235, "grad_norm": 1.902037262916565, "learning_rate": 9.659470144857234e-06, "loss": 0.6493, "step": 3450 }, { "epoch": 0.14466114039592132, "grad_norm": 2.1056015491485596, "learning_rate": 9.659223861383424e-06, "loss": 0.5241, "step": 3451 }, { "epoch": 0.14470305901093028, "grad_norm": 1.8223578929901123, "learning_rate": 9.658977492023264e-06, "loss": 0.5245, "step": 3452 }, { "epoch": 0.14474497762593924, "grad_norm": 2.255573034286499, "learning_rate": 9.658731036781295e-06, "loss": 0.5588, "step": 3453 }, { "epoch": 0.1447868962409482, "grad_norm": 1.6882843971252441, "learning_rate": 9.658484495662058e-06, "loss": 0.5966, "step": 3454 }, { "epoch": 0.14482881485595717, "grad_norm": 1.661843180656433, "learning_rate": 9.658237868670102e-06, "loss": 0.5557, "step": 3455 }, { "epoch": 0.14487073347096613, "grad_norm": 1.7638849020004272, "learning_rate": 9.65799115580997e-06, "loss": 0.5113, "step": 3456 }, { "epoch": 0.1449126520859751, "grad_norm": 2.1368234157562256, "learning_rate": 9.657744357086208e-06, "loss": 0.5878, "step": 3457 }, { "epoch": 0.14495457070098403, "grad_norm": 1.6078894138336182, "learning_rate": 9.65749747250337e-06, "loss": 0.5406, "step": 3458 }, { "epoch": 0.144996489315993, "grad_norm": 2.0174946784973145, "learning_rate": 9.657250502066004e-06, "loss": 0.635, "step": 3459 }, { "epoch": 0.14503840793100195, "grad_norm": 2.1543917655944824, "learning_rate": 9.657003445778665e-06, "loss": 0.5246, "step": 3460 }, { "epoch": 0.1450803265460109, "grad_norm": 1.4927952289581299, "learning_rate": 9.656756303645904e-06, "loss": 0.5667, "step": 3461 }, { "epoch": 0.14512224516101987, "grad_norm": 1.5395556688308716, "learning_rate": 9.65650907567228e-06, "loss": 0.5636, "step": 3462 }, { "epoch": 0.14516416377602884, "grad_norm": 1.6133986711502075, "learning_rate": 9.656261761862347e-06, "loss": 0.5631, "step": 3463 }, { "epoch": 0.1452060823910378, "grad_norm": 1.8051496744155884, "learning_rate": 9.656014362220667e-06, "loss": 0.5225, "step": 3464 }, { "epoch": 0.14524800100604676, "grad_norm": 2.738466501235962, "learning_rate": 9.655766876751798e-06, "loss": 0.5806, "step": 3465 }, { "epoch": 0.14528991962105572, "grad_norm": 1.7786576747894287, "learning_rate": 9.655519305460304e-06, "loss": 0.5401, "step": 3466 }, { "epoch": 0.1453318382360647, "grad_norm": 2.125622272491455, "learning_rate": 9.655271648350746e-06, "loss": 0.5248, "step": 3467 }, { "epoch": 0.14537375685107365, "grad_norm": 1.7692515850067139, "learning_rate": 9.655023905427691e-06, "loss": 0.5975, "step": 3468 }, { "epoch": 0.1454156754660826, "grad_norm": 2.060117721557617, "learning_rate": 9.654776076695708e-06, "loss": 0.5827, "step": 3469 }, { "epoch": 0.14545759408109157, "grad_norm": 1.970522165298462, "learning_rate": 9.65452816215936e-06, "loss": 0.5871, "step": 3470 }, { "epoch": 0.1454995126961005, "grad_norm": 2.484266519546509, "learning_rate": 9.654280161823222e-06, "loss": 0.5969, "step": 3471 }, { "epoch": 0.14554143131110947, "grad_norm": 2.256063461303711, "learning_rate": 9.654032075691864e-06, "loss": 0.5857, "step": 3472 }, { "epoch": 0.14558334992611843, "grad_norm": 2.0788989067077637, "learning_rate": 9.653783903769858e-06, "loss": 0.584, "step": 3473 }, { "epoch": 0.1456252685411274, "grad_norm": 1.6760716438293457, "learning_rate": 9.653535646061777e-06, "loss": 0.5356, "step": 3474 }, { "epoch": 0.14566718715613636, "grad_norm": 1.706600546836853, "learning_rate": 9.653287302572202e-06, "loss": 0.5861, "step": 3475 }, { "epoch": 0.14570910577114532, "grad_norm": 1.640931487083435, "learning_rate": 9.65303887330571e-06, "loss": 0.5582, "step": 3476 }, { "epoch": 0.14575102438615428, "grad_norm": 1.9181379079818726, "learning_rate": 9.652790358266877e-06, "loss": 0.6123, "step": 3477 }, { "epoch": 0.14579294300116324, "grad_norm": 1.801936388015747, "learning_rate": 9.652541757460287e-06, "loss": 0.5123, "step": 3478 }, { "epoch": 0.1458348616161722, "grad_norm": 1.685881495475769, "learning_rate": 9.652293070890522e-06, "loss": 0.5062, "step": 3479 }, { "epoch": 0.14587678023118117, "grad_norm": 1.69818115234375, "learning_rate": 9.652044298562164e-06, "loss": 0.592, "step": 3480 }, { "epoch": 0.14591869884619013, "grad_norm": 1.7773641347885132, "learning_rate": 9.6517954404798e-06, "loss": 0.5691, "step": 3481 }, { "epoch": 0.1459606174611991, "grad_norm": 2.2671940326690674, "learning_rate": 9.65154649664802e-06, "loss": 0.6002, "step": 3482 }, { "epoch": 0.14600253607620806, "grad_norm": 1.8432862758636475, "learning_rate": 9.651297467071411e-06, "loss": 0.575, "step": 3483 }, { "epoch": 0.146044454691217, "grad_norm": 1.72111177444458, "learning_rate": 9.651048351754564e-06, "loss": 0.5372, "step": 3484 }, { "epoch": 0.14608637330622595, "grad_norm": 1.9621590375900269, "learning_rate": 9.650799150702069e-06, "loss": 0.5999, "step": 3485 }, { "epoch": 0.14612829192123492, "grad_norm": 1.8334788084030151, "learning_rate": 9.650549863918521e-06, "loss": 0.5511, "step": 3486 }, { "epoch": 0.14617021053624388, "grad_norm": 2.2434144020080566, "learning_rate": 9.650300491408517e-06, "loss": 0.5696, "step": 3487 }, { "epoch": 0.14621212915125284, "grad_norm": 2.93679141998291, "learning_rate": 9.650051033176651e-06, "loss": 0.5749, "step": 3488 }, { "epoch": 0.1462540477662618, "grad_norm": 2.265594720840454, "learning_rate": 9.649801489227522e-06, "loss": 0.6169, "step": 3489 }, { "epoch": 0.14629596638127076, "grad_norm": 2.050053119659424, "learning_rate": 9.649551859565734e-06, "loss": 0.5422, "step": 3490 }, { "epoch": 0.14633788499627973, "grad_norm": 1.953346610069275, "learning_rate": 9.649302144195882e-06, "loss": 0.5488, "step": 3491 }, { "epoch": 0.1463798036112887, "grad_norm": 1.786928415298462, "learning_rate": 9.649052343122573e-06, "loss": 0.6149, "step": 3492 }, { "epoch": 0.14642172222629765, "grad_norm": 1.5312763452529907, "learning_rate": 9.648802456350412e-06, "loss": 0.506, "step": 3493 }, { "epoch": 0.1464636408413066, "grad_norm": 1.826838731765747, "learning_rate": 9.648552483884003e-06, "loss": 0.497, "step": 3494 }, { "epoch": 0.14650555945631558, "grad_norm": 1.8320475816726685, "learning_rate": 9.648302425727957e-06, "loss": 0.5697, "step": 3495 }, { "epoch": 0.1465474780713245, "grad_norm": 2.4757742881774902, "learning_rate": 9.648052281886882e-06, "loss": 0.5614, "step": 3496 }, { "epoch": 0.14658939668633347, "grad_norm": 1.887780785560608, "learning_rate": 9.647802052365388e-06, "loss": 0.5432, "step": 3497 }, { "epoch": 0.14663131530134244, "grad_norm": 2.0618834495544434, "learning_rate": 9.647551737168088e-06, "loss": 0.6257, "step": 3498 }, { "epoch": 0.1466732339163514, "grad_norm": 1.684418797492981, "learning_rate": 9.647301336299598e-06, "loss": 0.6115, "step": 3499 }, { "epoch": 0.14671515253136036, "grad_norm": 1.8850291967391968, "learning_rate": 9.647050849764531e-06, "loss": 0.6081, "step": 3500 }, { "epoch": 0.14675707114636932, "grad_norm": 1.740716576576233, "learning_rate": 9.646800277567508e-06, "loss": 0.6081, "step": 3501 }, { "epoch": 0.14679898976137828, "grad_norm": 1.6097019910812378, "learning_rate": 9.646549619713145e-06, "loss": 0.5939, "step": 3502 }, { "epoch": 0.14684090837638725, "grad_norm": 1.4423134326934814, "learning_rate": 9.646298876206064e-06, "loss": 0.4863, "step": 3503 }, { "epoch": 0.1468828269913962, "grad_norm": 1.6137723922729492, "learning_rate": 9.646048047050886e-06, "loss": 0.5448, "step": 3504 }, { "epoch": 0.14692474560640517, "grad_norm": 1.5509475469589233, "learning_rate": 9.645797132252235e-06, "loss": 0.5087, "step": 3505 }, { "epoch": 0.14696666422141413, "grad_norm": 1.8767855167388916, "learning_rate": 9.645546131814737e-06, "loss": 0.5503, "step": 3506 }, { "epoch": 0.1470085828364231, "grad_norm": 1.9670015573501587, "learning_rate": 9.645295045743018e-06, "loss": 0.5669, "step": 3507 }, { "epoch": 0.14705050145143206, "grad_norm": 1.5218755006790161, "learning_rate": 9.645043874041707e-06, "loss": 0.5153, "step": 3508 }, { "epoch": 0.147092420066441, "grad_norm": 3.9711713790893555, "learning_rate": 9.644792616715432e-06, "loss": 0.5194, "step": 3509 }, { "epoch": 0.14713433868144996, "grad_norm": 1.7729015350341797, "learning_rate": 9.644541273768829e-06, "loss": 0.57, "step": 3510 }, { "epoch": 0.14717625729645892, "grad_norm": 1.9136005640029907, "learning_rate": 9.644289845206527e-06, "loss": 0.6303, "step": 3511 }, { "epoch": 0.14721817591146788, "grad_norm": 1.7838330268859863, "learning_rate": 9.644038331033163e-06, "loss": 0.5989, "step": 3512 }, { "epoch": 0.14726009452647684, "grad_norm": 1.723174810409546, "learning_rate": 9.643786731253373e-06, "loss": 0.5666, "step": 3513 }, { "epoch": 0.1473020131414858, "grad_norm": 1.7206063270568848, "learning_rate": 9.643535045871792e-06, "loss": 0.5946, "step": 3514 }, { "epoch": 0.14734393175649477, "grad_norm": 1.8472816944122314, "learning_rate": 9.643283274893066e-06, "loss": 0.5937, "step": 3515 }, { "epoch": 0.14738585037150373, "grad_norm": 1.7558656930923462, "learning_rate": 9.64303141832183e-06, "loss": 0.5519, "step": 3516 }, { "epoch": 0.1474277689865127, "grad_norm": 2.05037522315979, "learning_rate": 9.642779476162727e-06, "loss": 0.5667, "step": 3517 }, { "epoch": 0.14746968760152165, "grad_norm": 1.622631311416626, "learning_rate": 9.642527448420403e-06, "loss": 0.5615, "step": 3518 }, { "epoch": 0.14751160621653062, "grad_norm": 1.5955398082733154, "learning_rate": 9.642275335099504e-06, "loss": 0.4664, "step": 3519 }, { "epoch": 0.14755352483153958, "grad_norm": 1.6531769037246704, "learning_rate": 9.64202313620468e-06, "loss": 0.6088, "step": 3520 }, { "epoch": 0.1475954434465485, "grad_norm": 1.7414218187332153, "learning_rate": 9.641770851740573e-06, "loss": 0.5572, "step": 3521 }, { "epoch": 0.14763736206155748, "grad_norm": 2.1900417804718018, "learning_rate": 9.641518481711838e-06, "loss": 0.5896, "step": 3522 }, { "epoch": 0.14767928067656644, "grad_norm": 1.7033088207244873, "learning_rate": 9.641266026123129e-06, "loss": 0.6324, "step": 3523 }, { "epoch": 0.1477211992915754, "grad_norm": 1.689660668373108, "learning_rate": 9.641013484979094e-06, "loss": 0.5723, "step": 3524 }, { "epoch": 0.14776311790658436, "grad_norm": 1.7355996370315552, "learning_rate": 9.640760858284392e-06, "loss": 0.572, "step": 3525 }, { "epoch": 0.14780503652159332, "grad_norm": 1.7533501386642456, "learning_rate": 9.640508146043681e-06, "loss": 0.5722, "step": 3526 }, { "epoch": 0.1478469551366023, "grad_norm": 1.7841687202453613, "learning_rate": 9.640255348261616e-06, "loss": 0.5488, "step": 3527 }, { "epoch": 0.14788887375161125, "grad_norm": 1.6333922147750854, "learning_rate": 9.64000246494286e-06, "loss": 0.523, "step": 3528 }, { "epoch": 0.1479307923666202, "grad_norm": 1.8272382020950317, "learning_rate": 9.639749496092072e-06, "loss": 0.5252, "step": 3529 }, { "epoch": 0.14797271098162917, "grad_norm": 1.9946331977844238, "learning_rate": 9.639496441713918e-06, "loss": 0.5594, "step": 3530 }, { "epoch": 0.14801462959663814, "grad_norm": 2.296830415725708, "learning_rate": 9.639243301813059e-06, "loss": 0.6177, "step": 3531 }, { "epoch": 0.1480565482116471, "grad_norm": 1.6278162002563477, "learning_rate": 9.638990076394165e-06, "loss": 0.5316, "step": 3532 }, { "epoch": 0.14809846682665606, "grad_norm": 2.3164074420928955, "learning_rate": 9.6387367654619e-06, "loss": 0.5237, "step": 3533 }, { "epoch": 0.148140385441665, "grad_norm": 2.339543104171753, "learning_rate": 9.638483369020938e-06, "loss": 0.5799, "step": 3534 }, { "epoch": 0.14818230405667396, "grad_norm": 1.8620874881744385, "learning_rate": 9.638229887075947e-06, "loss": 0.5336, "step": 3535 }, { "epoch": 0.14822422267168292, "grad_norm": 1.7794511318206787, "learning_rate": 9.637976319631599e-06, "loss": 0.5932, "step": 3536 }, { "epoch": 0.14826614128669188, "grad_norm": 1.8465381860733032, "learning_rate": 9.637722666692571e-06, "loss": 0.5803, "step": 3537 }, { "epoch": 0.14830805990170084, "grad_norm": 2.0503451824188232, "learning_rate": 9.637468928263535e-06, "loss": 0.5576, "step": 3538 }, { "epoch": 0.1483499785167098, "grad_norm": 1.7015939950942993, "learning_rate": 9.637215104349171e-06, "loss": 0.4793, "step": 3539 }, { "epoch": 0.14839189713171877, "grad_norm": 1.842318058013916, "learning_rate": 9.636961194954158e-06, "loss": 0.4998, "step": 3540 }, { "epoch": 0.14843381574672773, "grad_norm": 1.8150871992111206, "learning_rate": 9.636707200083176e-06, "loss": 0.529, "step": 3541 }, { "epoch": 0.1484757343617367, "grad_norm": 1.801015019416809, "learning_rate": 9.636453119740907e-06, "loss": 0.5649, "step": 3542 }, { "epoch": 0.14851765297674566, "grad_norm": 1.6978837251663208, "learning_rate": 9.636198953932033e-06, "loss": 0.5263, "step": 3543 }, { "epoch": 0.14855957159175462, "grad_norm": 2.375542640686035, "learning_rate": 9.635944702661242e-06, "loss": 0.5652, "step": 3544 }, { "epoch": 0.14860149020676358, "grad_norm": 1.8593968152999878, "learning_rate": 9.63569036593322e-06, "loss": 0.6051, "step": 3545 }, { "epoch": 0.14864340882177252, "grad_norm": 1.943943977355957, "learning_rate": 9.635435943752653e-06, "loss": 0.5215, "step": 3546 }, { "epoch": 0.14868532743678148, "grad_norm": 2.0113823413848877, "learning_rate": 9.635181436124232e-06, "loss": 0.5733, "step": 3547 }, { "epoch": 0.14872724605179044, "grad_norm": 1.7553315162658691, "learning_rate": 9.634926843052652e-06, "loss": 0.5811, "step": 3548 }, { "epoch": 0.1487691646667994, "grad_norm": 1.8374518156051636, "learning_rate": 9.6346721645426e-06, "loss": 0.5502, "step": 3549 }, { "epoch": 0.14881108328180837, "grad_norm": 2.0395307540893555, "learning_rate": 9.634417400598776e-06, "loss": 0.6274, "step": 3550 }, { "epoch": 0.14885300189681733, "grad_norm": 1.8253211975097656, "learning_rate": 9.634162551225873e-06, "loss": 0.5409, "step": 3551 }, { "epoch": 0.1488949205118263, "grad_norm": 1.7950153350830078, "learning_rate": 9.633907616428593e-06, "loss": 0.5687, "step": 3552 }, { "epoch": 0.14893683912683525, "grad_norm": 1.787706732749939, "learning_rate": 9.633652596211629e-06, "loss": 0.5596, "step": 3553 }, { "epoch": 0.14897875774184421, "grad_norm": 1.78297758102417, "learning_rate": 9.633397490579687e-06, "loss": 0.5567, "step": 3554 }, { "epoch": 0.14902067635685318, "grad_norm": 2.079273223876953, "learning_rate": 9.633142299537467e-06, "loss": 0.5403, "step": 3555 }, { "epoch": 0.14906259497186214, "grad_norm": 1.8142184019088745, "learning_rate": 9.632887023089673e-06, "loss": 0.5827, "step": 3556 }, { "epoch": 0.1491045135868711, "grad_norm": 1.917006492614746, "learning_rate": 9.632631661241012e-06, "loss": 0.5749, "step": 3557 }, { "epoch": 0.14914643220188006, "grad_norm": 1.7589094638824463, "learning_rate": 9.632376213996192e-06, "loss": 0.6118, "step": 3558 }, { "epoch": 0.149188350816889, "grad_norm": 1.721602439880371, "learning_rate": 9.632120681359918e-06, "loss": 0.5253, "step": 3559 }, { "epoch": 0.14923026943189796, "grad_norm": 1.7432525157928467, "learning_rate": 9.631865063336905e-06, "loss": 0.6232, "step": 3560 }, { "epoch": 0.14927218804690692, "grad_norm": 1.9726983308792114, "learning_rate": 9.631609359931864e-06, "loss": 0.6162, "step": 3561 }, { "epoch": 0.14931410666191589, "grad_norm": 1.8193279504776, "learning_rate": 9.631353571149504e-06, "loss": 0.5765, "step": 3562 }, { "epoch": 0.14935602527692485, "grad_norm": 1.790437936782837, "learning_rate": 9.631097696994547e-06, "loss": 0.597, "step": 3563 }, { "epoch": 0.1493979438919338, "grad_norm": 2.0921003818511963, "learning_rate": 9.630841737471704e-06, "loss": 0.5416, "step": 3564 }, { "epoch": 0.14943986250694277, "grad_norm": 1.5298768281936646, "learning_rate": 9.630585692585696e-06, "loss": 0.5192, "step": 3565 }, { "epoch": 0.14948178112195173, "grad_norm": 1.9559801816940308, "learning_rate": 9.630329562341243e-06, "loss": 0.6288, "step": 3566 }, { "epoch": 0.1495236997369607, "grad_norm": 1.9109975099563599, "learning_rate": 9.630073346743065e-06, "loss": 0.5778, "step": 3567 }, { "epoch": 0.14956561835196966, "grad_norm": 2.105896234512329, "learning_rate": 9.629817045795887e-06, "loss": 0.5479, "step": 3568 }, { "epoch": 0.14960753696697862, "grad_norm": 1.8092217445373535, "learning_rate": 9.62956065950443e-06, "loss": 0.5391, "step": 3569 }, { "epoch": 0.14964945558198758, "grad_norm": 2.2313241958618164, "learning_rate": 9.629304187873424e-06, "loss": 0.5647, "step": 3570 }, { "epoch": 0.14969137419699652, "grad_norm": 1.8155028820037842, "learning_rate": 9.629047630907596e-06, "loss": 0.5889, "step": 3571 }, { "epoch": 0.14973329281200548, "grad_norm": 1.8711720705032349, "learning_rate": 9.628790988611673e-06, "loss": 0.5981, "step": 3572 }, { "epoch": 0.14977521142701444, "grad_norm": 1.7823346853256226, "learning_rate": 9.628534260990388e-06, "loss": 0.5367, "step": 3573 }, { "epoch": 0.1498171300420234, "grad_norm": 1.5234469175338745, "learning_rate": 9.628277448048472e-06, "loss": 0.4718, "step": 3574 }, { "epoch": 0.14985904865703237, "grad_norm": 1.9413336515426636, "learning_rate": 9.62802054979066e-06, "loss": 0.5699, "step": 3575 }, { "epoch": 0.14990096727204133, "grad_norm": 2.2848567962646484, "learning_rate": 9.627763566221686e-06, "loss": 0.6169, "step": 3576 }, { "epoch": 0.1499428858870503, "grad_norm": 1.7778029441833496, "learning_rate": 9.62750649734629e-06, "loss": 0.5571, "step": 3577 }, { "epoch": 0.14998480450205925, "grad_norm": 1.8253378868103027, "learning_rate": 9.627249343169207e-06, "loss": 0.5878, "step": 3578 }, { "epoch": 0.15002672311706822, "grad_norm": 1.74790620803833, "learning_rate": 9.626992103695182e-06, "loss": 0.5212, "step": 3579 }, { "epoch": 0.15006864173207718, "grad_norm": 1.8089351654052734, "learning_rate": 9.626734778928952e-06, "loss": 0.5529, "step": 3580 }, { "epoch": 0.15011056034708614, "grad_norm": 5.227118968963623, "learning_rate": 9.626477368875265e-06, "loss": 0.5108, "step": 3581 }, { "epoch": 0.1501524789620951, "grad_norm": 1.7353459596633911, "learning_rate": 9.626219873538863e-06, "loss": 0.578, "step": 3582 }, { "epoch": 0.15019439757710407, "grad_norm": 1.6108124256134033, "learning_rate": 9.625962292924491e-06, "loss": 0.599, "step": 3583 }, { "epoch": 0.150236316192113, "grad_norm": 1.733898639678955, "learning_rate": 9.625704627036901e-06, "loss": 0.5583, "step": 3584 }, { "epoch": 0.15027823480712196, "grad_norm": 1.7582900524139404, "learning_rate": 9.62544687588084e-06, "loss": 0.5848, "step": 3585 }, { "epoch": 0.15032015342213093, "grad_norm": 1.8462940454483032, "learning_rate": 9.625189039461063e-06, "loss": 0.5866, "step": 3586 }, { "epoch": 0.1503620720371399, "grad_norm": 1.6015294790267944, "learning_rate": 9.624931117782316e-06, "loss": 0.5111, "step": 3587 }, { "epoch": 0.15040399065214885, "grad_norm": 1.6160014867782593, "learning_rate": 9.624673110849361e-06, "loss": 0.5016, "step": 3588 }, { "epoch": 0.1504459092671578, "grad_norm": 2.6064326763153076, "learning_rate": 9.62441501866695e-06, "loss": 0.558, "step": 3589 }, { "epoch": 0.15048782788216677, "grad_norm": 1.7696563005447388, "learning_rate": 9.62415684123984e-06, "loss": 0.6248, "step": 3590 }, { "epoch": 0.15052974649717574, "grad_norm": 1.8294150829315186, "learning_rate": 9.623898578572792e-06, "loss": 0.54, "step": 3591 }, { "epoch": 0.1505716651121847, "grad_norm": 1.6301625967025757, "learning_rate": 9.623640230670566e-06, "loss": 0.5491, "step": 3592 }, { "epoch": 0.15061358372719366, "grad_norm": 2.1378021240234375, "learning_rate": 9.623381797537925e-06, "loss": 0.5268, "step": 3593 }, { "epoch": 0.15065550234220262, "grad_norm": 1.7214804887771606, "learning_rate": 9.62312327917963e-06, "loss": 0.5619, "step": 3594 }, { "epoch": 0.1506974209572116, "grad_norm": 1.891323447227478, "learning_rate": 9.622864675600452e-06, "loss": 0.5545, "step": 3595 }, { "epoch": 0.15073933957222052, "grad_norm": 1.7987596988677979, "learning_rate": 9.622605986805154e-06, "loss": 0.5437, "step": 3596 }, { "epoch": 0.15078125818722948, "grad_norm": 2.912022590637207, "learning_rate": 9.622347212798503e-06, "loss": 0.5501, "step": 3597 }, { "epoch": 0.15082317680223845, "grad_norm": 1.635650634765625, "learning_rate": 9.622088353585272e-06, "loss": 0.5543, "step": 3598 }, { "epoch": 0.1508650954172474, "grad_norm": 1.7741659879684448, "learning_rate": 9.621829409170233e-06, "loss": 0.6182, "step": 3599 }, { "epoch": 0.15090701403225637, "grad_norm": 1.7816789150238037, "learning_rate": 9.621570379558158e-06, "loss": 0.5387, "step": 3600 }, { "epoch": 0.15094893264726533, "grad_norm": 1.8951542377471924, "learning_rate": 9.621311264753821e-06, "loss": 0.5957, "step": 3601 }, { "epoch": 0.1509908512622743, "grad_norm": 1.899739146232605, "learning_rate": 9.621052064762002e-06, "loss": 0.5863, "step": 3602 }, { "epoch": 0.15103276987728326, "grad_norm": 1.6261546611785889, "learning_rate": 9.620792779587474e-06, "loss": 0.5527, "step": 3603 }, { "epoch": 0.15107468849229222, "grad_norm": 1.6401755809783936, "learning_rate": 9.62053340923502e-06, "loss": 0.5547, "step": 3604 }, { "epoch": 0.15111660710730118, "grad_norm": 1.9522533416748047, "learning_rate": 9.620273953709421e-06, "loss": 0.5655, "step": 3605 }, { "epoch": 0.15115852572231014, "grad_norm": 1.980550765991211, "learning_rate": 9.62001441301546e-06, "loss": 0.5609, "step": 3606 }, { "epoch": 0.1512004443373191, "grad_norm": 1.542944073677063, "learning_rate": 9.619754787157918e-06, "loss": 0.5451, "step": 3607 }, { "epoch": 0.15124236295232807, "grad_norm": 1.521191120147705, "learning_rate": 9.619495076141584e-06, "loss": 0.549, "step": 3608 }, { "epoch": 0.151284281567337, "grad_norm": 1.6236188411712646, "learning_rate": 9.619235279971243e-06, "loss": 0.5396, "step": 3609 }, { "epoch": 0.15132620018234597, "grad_norm": 1.7129281759262085, "learning_rate": 9.618975398651688e-06, "loss": 0.5517, "step": 3610 }, { "epoch": 0.15136811879735493, "grad_norm": 1.8671799898147583, "learning_rate": 9.618715432187707e-06, "loss": 0.5457, "step": 3611 }, { "epoch": 0.1514100374123639, "grad_norm": 1.6414732933044434, "learning_rate": 9.61845538058409e-06, "loss": 0.5375, "step": 3612 }, { "epoch": 0.15145195602737285, "grad_norm": 1.9952919483184814, "learning_rate": 9.618195243845636e-06, "loss": 0.5799, "step": 3613 }, { "epoch": 0.15149387464238181, "grad_norm": 2.1952829360961914, "learning_rate": 9.617935021977134e-06, "loss": 0.5365, "step": 3614 }, { "epoch": 0.15153579325739078, "grad_norm": 1.8865530490875244, "learning_rate": 9.617674714983386e-06, "loss": 0.5977, "step": 3615 }, { "epoch": 0.15157771187239974, "grad_norm": 1.8877601623535156, "learning_rate": 9.617414322869187e-06, "loss": 0.5535, "step": 3616 }, { "epoch": 0.1516196304874087, "grad_norm": 1.7872114181518555, "learning_rate": 9.61715384563934e-06, "loss": 0.6214, "step": 3617 }, { "epoch": 0.15166154910241766, "grad_norm": 1.7594739198684692, "learning_rate": 9.616893283298643e-06, "loss": 0.5185, "step": 3618 }, { "epoch": 0.15170346771742663, "grad_norm": 1.7995809316635132, "learning_rate": 9.616632635851904e-06, "loss": 0.5431, "step": 3619 }, { "epoch": 0.1517453863324356, "grad_norm": 1.8812181949615479, "learning_rate": 9.616371903303922e-06, "loss": 0.5878, "step": 3620 }, { "epoch": 0.15178730494744452, "grad_norm": 2.069096565246582, "learning_rate": 9.616111085659505e-06, "loss": 0.5547, "step": 3621 }, { "epoch": 0.15182922356245349, "grad_norm": 1.7472484111785889, "learning_rate": 9.615850182923464e-06, "loss": 0.5028, "step": 3622 }, { "epoch": 0.15187114217746245, "grad_norm": 2.167663335800171, "learning_rate": 9.615589195100605e-06, "loss": 0.6339, "step": 3623 }, { "epoch": 0.1519130607924714, "grad_norm": 1.8109898567199707, "learning_rate": 9.61532812219574e-06, "loss": 0.5942, "step": 3624 }, { "epoch": 0.15195497940748037, "grad_norm": 2.071075439453125, "learning_rate": 9.615066964213683e-06, "loss": 0.6455, "step": 3625 }, { "epoch": 0.15199689802248934, "grad_norm": 2.239119052886963, "learning_rate": 9.614805721159245e-06, "loss": 0.5277, "step": 3626 }, { "epoch": 0.1520388166374983, "grad_norm": 1.924824595451355, "learning_rate": 9.614544393037244e-06, "loss": 0.6051, "step": 3627 }, { "epoch": 0.15208073525250726, "grad_norm": 1.8111618757247925, "learning_rate": 9.614282979852497e-06, "loss": 0.5832, "step": 3628 }, { "epoch": 0.15212265386751622, "grad_norm": 1.4603947401046753, "learning_rate": 9.61402148160982e-06, "loss": 0.5113, "step": 3629 }, { "epoch": 0.15216457248252518, "grad_norm": 1.7747983932495117, "learning_rate": 9.613759898314038e-06, "loss": 0.5591, "step": 3630 }, { "epoch": 0.15220649109753415, "grad_norm": 1.6680583953857422, "learning_rate": 9.61349822996997e-06, "loss": 0.5669, "step": 3631 }, { "epoch": 0.1522484097125431, "grad_norm": 1.7893919944763184, "learning_rate": 9.613236476582442e-06, "loss": 0.5326, "step": 3632 }, { "epoch": 0.15229032832755207, "grad_norm": 1.804807424545288, "learning_rate": 9.612974638156274e-06, "loss": 0.6183, "step": 3633 }, { "epoch": 0.152332246942561, "grad_norm": 1.7465029954910278, "learning_rate": 9.612712714696297e-06, "loss": 0.5355, "step": 3634 }, { "epoch": 0.15237416555756997, "grad_norm": 1.777010202407837, "learning_rate": 9.61245070620734e-06, "loss": 0.5144, "step": 3635 }, { "epoch": 0.15241608417257893, "grad_norm": 2.187100410461426, "learning_rate": 9.612188612694228e-06, "loss": 0.5859, "step": 3636 }, { "epoch": 0.1524580027875879, "grad_norm": 1.7609097957611084, "learning_rate": 9.611926434161795e-06, "loss": 0.6585, "step": 3637 }, { "epoch": 0.15249992140259686, "grad_norm": 2.010871410369873, "learning_rate": 9.611664170614875e-06, "loss": 0.5861, "step": 3638 }, { "epoch": 0.15254184001760582, "grad_norm": 2.3765652179718018, "learning_rate": 9.611401822058301e-06, "loss": 0.5926, "step": 3639 }, { "epoch": 0.15258375863261478, "grad_norm": 1.953928828239441, "learning_rate": 9.611139388496911e-06, "loss": 0.6042, "step": 3640 }, { "epoch": 0.15262567724762374, "grad_norm": 1.8812962770462036, "learning_rate": 9.61087686993554e-06, "loss": 0.531, "step": 3641 }, { "epoch": 0.1526675958626327, "grad_norm": 1.8337252140045166, "learning_rate": 9.610614266379028e-06, "loss": 0.5649, "step": 3642 }, { "epoch": 0.15270951447764167, "grad_norm": 2.071974039077759, "learning_rate": 9.610351577832215e-06, "loss": 0.5918, "step": 3643 }, { "epoch": 0.15275143309265063, "grad_norm": 4.168229103088379, "learning_rate": 9.610088804299946e-06, "loss": 0.6137, "step": 3644 }, { "epoch": 0.1527933517076596, "grad_norm": 1.9126654863357544, "learning_rate": 9.609825945787062e-06, "loss": 0.6025, "step": 3645 }, { "epoch": 0.15283527032266853, "grad_norm": 2.0201568603515625, "learning_rate": 9.609563002298411e-06, "loss": 0.5348, "step": 3646 }, { "epoch": 0.1528771889376775, "grad_norm": 1.7345962524414062, "learning_rate": 9.609299973838836e-06, "loss": 0.5293, "step": 3647 }, { "epoch": 0.15291910755268645, "grad_norm": 2.187206745147705, "learning_rate": 9.609036860413192e-06, "loss": 0.5633, "step": 3648 }, { "epoch": 0.1529610261676954, "grad_norm": 1.920302152633667, "learning_rate": 9.608773662026323e-06, "loss": 0.5892, "step": 3649 }, { "epoch": 0.15300294478270438, "grad_norm": 2.170102834701538, "learning_rate": 9.608510378683083e-06, "loss": 0.5756, "step": 3650 }, { "epoch": 0.15304486339771334, "grad_norm": 2.087857484817505, "learning_rate": 9.608247010388325e-06, "loss": 0.6326, "step": 3651 }, { "epoch": 0.1530867820127223, "grad_norm": 2.081824541091919, "learning_rate": 9.607983557146906e-06, "loss": 0.4991, "step": 3652 }, { "epoch": 0.15312870062773126, "grad_norm": 1.7784674167633057, "learning_rate": 9.607720018963678e-06, "loss": 0.5657, "step": 3653 }, { "epoch": 0.15317061924274022, "grad_norm": 2.5704336166381836, "learning_rate": 9.607456395843501e-06, "loss": 0.5248, "step": 3654 }, { "epoch": 0.1532125378577492, "grad_norm": 2.4645397663116455, "learning_rate": 9.607192687791238e-06, "loss": 0.5217, "step": 3655 }, { "epoch": 0.15325445647275815, "grad_norm": 1.9106484651565552, "learning_rate": 9.606928894811744e-06, "loss": 0.5421, "step": 3656 }, { "epoch": 0.1532963750877671, "grad_norm": 1.9676308631896973, "learning_rate": 9.606665016909886e-06, "loss": 0.6075, "step": 3657 }, { "epoch": 0.15333829370277607, "grad_norm": 2.425340414047241, "learning_rate": 9.606401054090529e-06, "loss": 0.58, "step": 3658 }, { "epoch": 0.153380212317785, "grad_norm": 1.9042456150054932, "learning_rate": 9.606137006358533e-06, "loss": 0.5705, "step": 3659 }, { "epoch": 0.15342213093279397, "grad_norm": 1.7077627182006836, "learning_rate": 9.605872873718772e-06, "loss": 0.6055, "step": 3660 }, { "epoch": 0.15346404954780293, "grad_norm": 1.6973634958267212, "learning_rate": 9.605608656176111e-06, "loss": 0.5738, "step": 3661 }, { "epoch": 0.1535059681628119, "grad_norm": 1.9001543521881104, "learning_rate": 9.605344353735419e-06, "loss": 0.5228, "step": 3662 }, { "epoch": 0.15354788677782086, "grad_norm": 2.408163070678711, "learning_rate": 9.605079966401573e-06, "loss": 0.5858, "step": 3663 }, { "epoch": 0.15358980539282982, "grad_norm": 1.5027258396148682, "learning_rate": 9.604815494179444e-06, "loss": 0.5188, "step": 3664 }, { "epoch": 0.15363172400783878, "grad_norm": 1.9961168766021729, "learning_rate": 9.60455093707391e-06, "loss": 0.5033, "step": 3665 }, { "epoch": 0.15367364262284774, "grad_norm": 2.232897996902466, "learning_rate": 9.60428629508984e-06, "loss": 0.6092, "step": 3666 }, { "epoch": 0.1537155612378567, "grad_norm": 1.7980942726135254, "learning_rate": 9.604021568232121e-06, "loss": 0.5573, "step": 3667 }, { "epoch": 0.15375747985286567, "grad_norm": 2.003401279449463, "learning_rate": 9.603756756505628e-06, "loss": 0.553, "step": 3668 }, { "epoch": 0.15379939846787463, "grad_norm": 1.766993761062622, "learning_rate": 9.603491859915244e-06, "loss": 0.5904, "step": 3669 }, { "epoch": 0.1538413170828836, "grad_norm": 1.6446119546890259, "learning_rate": 9.603226878465851e-06, "loss": 0.5492, "step": 3670 }, { "epoch": 0.15388323569789253, "grad_norm": 1.802216649055481, "learning_rate": 9.602961812162335e-06, "loss": 0.5089, "step": 3671 }, { "epoch": 0.1539251543129015, "grad_norm": 1.6323007345199585, "learning_rate": 9.60269666100958e-06, "loss": 0.5521, "step": 3672 }, { "epoch": 0.15396707292791045, "grad_norm": 1.7530771493911743, "learning_rate": 9.602431425012477e-06, "loss": 0.5742, "step": 3673 }, { "epoch": 0.15400899154291942, "grad_norm": 2.032590866088867, "learning_rate": 9.602166104175913e-06, "loss": 0.6011, "step": 3674 }, { "epoch": 0.15405091015792838, "grad_norm": 2.0110867023468018, "learning_rate": 9.601900698504778e-06, "loss": 0.5879, "step": 3675 }, { "epoch": 0.15409282877293734, "grad_norm": 1.9774435758590698, "learning_rate": 9.601635208003967e-06, "loss": 0.5684, "step": 3676 }, { "epoch": 0.1541347473879463, "grad_norm": 1.903502106666565, "learning_rate": 9.601369632678373e-06, "loss": 0.5562, "step": 3677 }, { "epoch": 0.15417666600295526, "grad_norm": 1.7092000246047974, "learning_rate": 9.601103972532888e-06, "loss": 0.5577, "step": 3678 }, { "epoch": 0.15421858461796423, "grad_norm": 1.8371819257736206, "learning_rate": 9.600838227572415e-06, "loss": 0.5317, "step": 3679 }, { "epoch": 0.1542605032329732, "grad_norm": 2.160637617111206, "learning_rate": 9.600572397801849e-06, "loss": 0.5902, "step": 3680 }, { "epoch": 0.15430242184798215, "grad_norm": 1.4805123805999756, "learning_rate": 9.60030648322609e-06, "loss": 0.519, "step": 3681 }, { "epoch": 0.15434434046299111, "grad_norm": 2.101639747619629, "learning_rate": 9.600040483850042e-06, "loss": 0.6101, "step": 3682 }, { "epoch": 0.15438625907800008, "grad_norm": 1.6616528034210205, "learning_rate": 9.599774399678605e-06, "loss": 0.541, "step": 3683 }, { "epoch": 0.154428177693009, "grad_norm": 1.7826536893844604, "learning_rate": 9.599508230716688e-06, "loss": 0.5514, "step": 3684 }, { "epoch": 0.15447009630801797, "grad_norm": 1.8635337352752686, "learning_rate": 9.599241976969194e-06, "loss": 0.5595, "step": 3685 }, { "epoch": 0.15451201492302694, "grad_norm": 1.76482355594635, "learning_rate": 9.598975638441034e-06, "loss": 0.54, "step": 3686 }, { "epoch": 0.1545539335380359, "grad_norm": 1.9332510232925415, "learning_rate": 9.598709215137114e-06, "loss": 0.5551, "step": 3687 }, { "epoch": 0.15459585215304486, "grad_norm": 1.7258020639419556, "learning_rate": 9.598442707062348e-06, "loss": 0.5901, "step": 3688 }, { "epoch": 0.15463777076805382, "grad_norm": 1.8747797012329102, "learning_rate": 9.59817611422165e-06, "loss": 0.5449, "step": 3689 }, { "epoch": 0.15467968938306279, "grad_norm": 1.6942987442016602, "learning_rate": 9.59790943661993e-06, "loss": 0.5789, "step": 3690 }, { "epoch": 0.15472160799807175, "grad_norm": 1.756593108177185, "learning_rate": 9.597642674262105e-06, "loss": 0.5376, "step": 3691 }, { "epoch": 0.1547635266130807, "grad_norm": 1.9270482063293457, "learning_rate": 9.597375827153095e-06, "loss": 0.5193, "step": 3692 }, { "epoch": 0.15480544522808967, "grad_norm": 1.7376253604888916, "learning_rate": 9.597108895297818e-06, "loss": 0.569, "step": 3693 }, { "epoch": 0.15484736384309863, "grad_norm": 1.6393158435821533, "learning_rate": 9.596841878701191e-06, "loss": 0.6235, "step": 3694 }, { "epoch": 0.1548892824581076, "grad_norm": 2.002096652984619, "learning_rate": 9.596574777368141e-06, "loss": 0.5828, "step": 3695 }, { "epoch": 0.15493120107311653, "grad_norm": 1.9473057985305786, "learning_rate": 9.59630759130359e-06, "loss": 0.5487, "step": 3696 }, { "epoch": 0.1549731196881255, "grad_norm": 1.7281686067581177, "learning_rate": 9.596040320512462e-06, "loss": 0.5488, "step": 3697 }, { "epoch": 0.15501503830313446, "grad_norm": 1.8763034343719482, "learning_rate": 9.595772964999684e-06, "loss": 0.6226, "step": 3698 }, { "epoch": 0.15505695691814342, "grad_norm": 1.934847354888916, "learning_rate": 9.595505524770186e-06, "loss": 0.6245, "step": 3699 }, { "epoch": 0.15509887553315238, "grad_norm": 1.783992052078247, "learning_rate": 9.595237999828894e-06, "loss": 0.5559, "step": 3700 }, { "epoch": 0.15514079414816134, "grad_norm": 2.011686325073242, "learning_rate": 9.594970390180745e-06, "loss": 0.6097, "step": 3701 }, { "epoch": 0.1551827127631703, "grad_norm": 2.147160768508911, "learning_rate": 9.594702695830668e-06, "loss": 0.4868, "step": 3702 }, { "epoch": 0.15522463137817927, "grad_norm": 1.729063630104065, "learning_rate": 9.594434916783599e-06, "loss": 0.494, "step": 3703 }, { "epoch": 0.15526654999318823, "grad_norm": 1.8774183988571167, "learning_rate": 9.594167053044475e-06, "loss": 0.584, "step": 3704 }, { "epoch": 0.1553084686081972, "grad_norm": 1.6985374689102173, "learning_rate": 9.593899104618232e-06, "loss": 0.5453, "step": 3705 }, { "epoch": 0.15535038722320615, "grad_norm": 2.0370426177978516, "learning_rate": 9.59363107150981e-06, "loss": 0.637, "step": 3706 }, { "epoch": 0.15539230583821512, "grad_norm": 2.93373441696167, "learning_rate": 9.59336295372415e-06, "loss": 0.5718, "step": 3707 }, { "epoch": 0.15543422445322408, "grad_norm": 1.811876893043518, "learning_rate": 9.593094751266193e-06, "loss": 0.5486, "step": 3708 }, { "epoch": 0.155476143068233, "grad_norm": 2.0612661838531494, "learning_rate": 9.592826464140884e-06, "loss": 0.5753, "step": 3709 }, { "epoch": 0.15551806168324198, "grad_norm": 1.5702086687088013, "learning_rate": 9.592558092353168e-06, "loss": 0.4925, "step": 3710 }, { "epoch": 0.15555998029825094, "grad_norm": 2.0217092037200928, "learning_rate": 9.592289635907994e-06, "loss": 0.5959, "step": 3711 }, { "epoch": 0.1556018989132599, "grad_norm": 1.8758327960968018, "learning_rate": 9.592021094810308e-06, "loss": 0.5076, "step": 3712 }, { "epoch": 0.15564381752826886, "grad_norm": 1.9565306901931763, "learning_rate": 9.591752469065063e-06, "loss": 0.5659, "step": 3713 }, { "epoch": 0.15568573614327783, "grad_norm": 2.232433557510376, "learning_rate": 9.591483758677209e-06, "loss": 0.6027, "step": 3714 }, { "epoch": 0.1557276547582868, "grad_norm": 1.984290361404419, "learning_rate": 9.591214963651699e-06, "loss": 0.568, "step": 3715 }, { "epoch": 0.15576957337329575, "grad_norm": 2.0811612606048584, "learning_rate": 9.590946083993487e-06, "loss": 0.5823, "step": 3716 }, { "epoch": 0.1558114919883047, "grad_norm": 1.6891857385635376, "learning_rate": 9.590677119707533e-06, "loss": 0.5026, "step": 3717 }, { "epoch": 0.15585341060331367, "grad_norm": 1.8477412462234497, "learning_rate": 9.59040807079879e-06, "loss": 0.5697, "step": 3718 }, { "epoch": 0.15589532921832264, "grad_norm": 1.722796082496643, "learning_rate": 9.590138937272223e-06, "loss": 0.5097, "step": 3719 }, { "epoch": 0.1559372478333316, "grad_norm": 2.0216245651245117, "learning_rate": 9.58986971913279e-06, "loss": 0.5464, "step": 3720 }, { "epoch": 0.15597916644834053, "grad_norm": 4.510492324829102, "learning_rate": 9.589600416385451e-06, "loss": 0.5483, "step": 3721 }, { "epoch": 0.1560210850633495, "grad_norm": 2.018237590789795, "learning_rate": 9.589331029035176e-06, "loss": 0.634, "step": 3722 }, { "epoch": 0.15606300367835846, "grad_norm": 1.75296950340271, "learning_rate": 9.589061557086928e-06, "loss": 0.576, "step": 3723 }, { "epoch": 0.15610492229336742, "grad_norm": 1.7728112936019897, "learning_rate": 9.588792000545674e-06, "loss": 0.5767, "step": 3724 }, { "epoch": 0.15614684090837638, "grad_norm": 2.0210859775543213, "learning_rate": 9.588522359416383e-06, "loss": 0.5203, "step": 3725 }, { "epoch": 0.15618875952338535, "grad_norm": 1.8556069135665894, "learning_rate": 9.588252633704026e-06, "loss": 0.6073, "step": 3726 }, { "epoch": 0.1562306781383943, "grad_norm": 1.8901137113571167, "learning_rate": 9.587982823413575e-06, "loss": 0.5721, "step": 3727 }, { "epoch": 0.15627259675340327, "grad_norm": 1.956566572189331, "learning_rate": 9.587712928550003e-06, "loss": 0.5889, "step": 3728 }, { "epoch": 0.15631451536841223, "grad_norm": 1.9668744802474976, "learning_rate": 9.587442949118285e-06, "loss": 0.6023, "step": 3729 }, { "epoch": 0.1563564339834212, "grad_norm": 1.6799746751785278, "learning_rate": 9.5871728851234e-06, "loss": 0.5401, "step": 3730 }, { "epoch": 0.15639835259843016, "grad_norm": 1.735190510749817, "learning_rate": 9.586902736570322e-06, "loss": 0.5505, "step": 3731 }, { "epoch": 0.15644027121343912, "grad_norm": 1.7726283073425293, "learning_rate": 9.586632503464036e-06, "loss": 0.5494, "step": 3732 }, { "epoch": 0.15648218982844808, "grad_norm": 2.438145160675049, "learning_rate": 9.586362185809518e-06, "loss": 0.6005, "step": 3733 }, { "epoch": 0.15652410844345702, "grad_norm": 1.7525675296783447, "learning_rate": 9.586091783611753e-06, "loss": 0.5442, "step": 3734 }, { "epoch": 0.15656602705846598, "grad_norm": 1.7292628288269043, "learning_rate": 9.585821296875729e-06, "loss": 0.5053, "step": 3735 }, { "epoch": 0.15660794567347494, "grad_norm": 1.8090280294418335, "learning_rate": 9.585550725606427e-06, "loss": 0.5789, "step": 3736 }, { "epoch": 0.1566498642884839, "grad_norm": 1.865526795387268, "learning_rate": 9.585280069808838e-06, "loss": 0.6335, "step": 3737 }, { "epoch": 0.15669178290349287, "grad_norm": 1.8247452974319458, "learning_rate": 9.58500932948795e-06, "loss": 0.5504, "step": 3738 }, { "epoch": 0.15673370151850183, "grad_norm": 1.9011324644088745, "learning_rate": 9.584738504648753e-06, "loss": 0.5464, "step": 3739 }, { "epoch": 0.1567756201335108, "grad_norm": 1.9716182947158813, "learning_rate": 9.584467595296239e-06, "loss": 0.5946, "step": 3740 }, { "epoch": 0.15681753874851975, "grad_norm": 1.8051491975784302, "learning_rate": 9.584196601435403e-06, "loss": 0.5956, "step": 3741 }, { "epoch": 0.15685945736352871, "grad_norm": 2.1832025051116943, "learning_rate": 9.58392552307124e-06, "loss": 0.6127, "step": 3742 }, { "epoch": 0.15690137597853768, "grad_norm": 1.604078769683838, "learning_rate": 9.583654360208749e-06, "loss": 0.5621, "step": 3743 }, { "epoch": 0.15694329459354664, "grad_norm": 2.445638418197632, "learning_rate": 9.583383112852924e-06, "loss": 0.5017, "step": 3744 }, { "epoch": 0.1569852132085556, "grad_norm": 2.210650682449341, "learning_rate": 9.583111781008769e-06, "loss": 0.5676, "step": 3745 }, { "epoch": 0.15702713182356454, "grad_norm": 2.0637598037719727, "learning_rate": 9.582840364681286e-06, "loss": 0.6303, "step": 3746 }, { "epoch": 0.1570690504385735, "grad_norm": 1.6322659254074097, "learning_rate": 9.582568863875474e-06, "loss": 0.5518, "step": 3747 }, { "epoch": 0.15711096905358246, "grad_norm": 1.6077215671539307, "learning_rate": 9.58229727859634e-06, "loss": 0.5254, "step": 3748 }, { "epoch": 0.15715288766859142, "grad_norm": 1.8936796188354492, "learning_rate": 9.582025608848891e-06, "loss": 0.5682, "step": 3749 }, { "epoch": 0.15719480628360039, "grad_norm": 2.061170816421509, "learning_rate": 9.581753854638136e-06, "loss": 0.5602, "step": 3750 }, { "epoch": 0.15723672489860935, "grad_norm": 1.8159420490264893, "learning_rate": 9.581482015969082e-06, "loss": 0.4958, "step": 3751 }, { "epoch": 0.1572786435136183, "grad_norm": 2.009115219116211, "learning_rate": 9.58121009284674e-06, "loss": 0.5875, "step": 3752 }, { "epoch": 0.15732056212862727, "grad_norm": 2.534414052963257, "learning_rate": 9.580938085276123e-06, "loss": 0.5627, "step": 3753 }, { "epoch": 0.15736248074363623, "grad_norm": 1.9735116958618164, "learning_rate": 9.580665993262246e-06, "loss": 0.576, "step": 3754 }, { "epoch": 0.1574043993586452, "grad_norm": 1.645819902420044, "learning_rate": 9.580393816810124e-06, "loss": 0.5269, "step": 3755 }, { "epoch": 0.15744631797365416, "grad_norm": 1.7131516933441162, "learning_rate": 9.580121555924773e-06, "loss": 0.563, "step": 3756 }, { "epoch": 0.15748823658866312, "grad_norm": 1.999184489250183, "learning_rate": 9.579849210611214e-06, "loss": 0.5615, "step": 3757 }, { "epoch": 0.15753015520367208, "grad_norm": 3.4860620498657227, "learning_rate": 9.579576780874466e-06, "loss": 0.5628, "step": 3758 }, { "epoch": 0.15757207381868102, "grad_norm": 1.7249462604522705, "learning_rate": 9.579304266719553e-06, "loss": 0.5454, "step": 3759 }, { "epoch": 0.15761399243368998, "grad_norm": 2.0669407844543457, "learning_rate": 9.579031668151495e-06, "loss": 0.5753, "step": 3760 }, { "epoch": 0.15765591104869894, "grad_norm": 2.0035126209259033, "learning_rate": 9.578758985175317e-06, "loss": 0.5155, "step": 3761 }, { "epoch": 0.1576978296637079, "grad_norm": 1.727293848991394, "learning_rate": 9.578486217796048e-06, "loss": 0.5737, "step": 3762 }, { "epoch": 0.15773974827871687, "grad_norm": 1.8193250894546509, "learning_rate": 9.578213366018717e-06, "loss": 0.5638, "step": 3763 }, { "epoch": 0.15778166689372583, "grad_norm": 1.886962652206421, "learning_rate": 9.57794042984835e-06, "loss": 0.5664, "step": 3764 }, { "epoch": 0.1578235855087348, "grad_norm": 1.7029160261154175, "learning_rate": 9.57766740928998e-06, "loss": 0.576, "step": 3765 }, { "epoch": 0.15786550412374376, "grad_norm": 1.6017897129058838, "learning_rate": 9.577394304348641e-06, "loss": 0.5586, "step": 3766 }, { "epoch": 0.15790742273875272, "grad_norm": 1.797882318496704, "learning_rate": 9.577121115029364e-06, "loss": 0.5583, "step": 3767 }, { "epoch": 0.15794934135376168, "grad_norm": 2.124929428100586, "learning_rate": 9.576847841337189e-06, "loss": 0.6027, "step": 3768 }, { "epoch": 0.15799125996877064, "grad_norm": 1.7777884006500244, "learning_rate": 9.57657448327715e-06, "loss": 0.5613, "step": 3769 }, { "epoch": 0.1580331785837796, "grad_norm": 1.8161332607269287, "learning_rate": 9.576301040854287e-06, "loss": 0.5514, "step": 3770 }, { "epoch": 0.15807509719878854, "grad_norm": 1.6515154838562012, "learning_rate": 9.576027514073642e-06, "loss": 0.5918, "step": 3771 }, { "epoch": 0.1581170158137975, "grad_norm": 2.0628557205200195, "learning_rate": 9.575753902940256e-06, "loss": 0.6219, "step": 3772 }, { "epoch": 0.15815893442880646, "grad_norm": 1.608535885810852, "learning_rate": 9.575480207459174e-06, "loss": 0.5714, "step": 3773 }, { "epoch": 0.15820085304381543, "grad_norm": 1.62765634059906, "learning_rate": 9.575206427635436e-06, "loss": 0.5988, "step": 3774 }, { "epoch": 0.1582427716588244, "grad_norm": 1.8753851652145386, "learning_rate": 9.574932563474095e-06, "loss": 0.5639, "step": 3775 }, { "epoch": 0.15828469027383335, "grad_norm": 2.111511468887329, "learning_rate": 9.574658614980198e-06, "loss": 0.6061, "step": 3776 }, { "epoch": 0.1583266088888423, "grad_norm": 1.7942333221435547, "learning_rate": 9.574384582158791e-06, "loss": 0.5693, "step": 3777 }, { "epoch": 0.15836852750385128, "grad_norm": 1.8075827360153198, "learning_rate": 9.57411046501493e-06, "loss": 0.5453, "step": 3778 }, { "epoch": 0.15841044611886024, "grad_norm": 1.8172472715377808, "learning_rate": 9.573836263553664e-06, "loss": 0.5814, "step": 3779 }, { "epoch": 0.1584523647338692, "grad_norm": 1.797945499420166, "learning_rate": 9.573561977780051e-06, "loss": 0.5111, "step": 3780 }, { "epoch": 0.15849428334887816, "grad_norm": 1.7086777687072754, "learning_rate": 9.573287607699146e-06, "loss": 0.5483, "step": 3781 }, { "epoch": 0.15853620196388712, "grad_norm": 1.6545969247817993, "learning_rate": 9.573013153316006e-06, "loss": 0.5585, "step": 3782 }, { "epoch": 0.1585781205788961, "grad_norm": 1.780717372894287, "learning_rate": 9.57273861463569e-06, "loss": 0.5627, "step": 3783 }, { "epoch": 0.15862003919390502, "grad_norm": 1.8216012716293335, "learning_rate": 9.57246399166326e-06, "loss": 0.5653, "step": 3784 }, { "epoch": 0.15866195780891398, "grad_norm": 1.7619750499725342, "learning_rate": 9.572189284403776e-06, "loss": 0.6043, "step": 3785 }, { "epoch": 0.15870387642392295, "grad_norm": 1.6225441694259644, "learning_rate": 9.571914492862306e-06, "loss": 0.5421, "step": 3786 }, { "epoch": 0.1587457950389319, "grad_norm": 2.176872730255127, "learning_rate": 9.57163961704391e-06, "loss": 0.5656, "step": 3787 }, { "epoch": 0.15878771365394087, "grad_norm": 1.704017996788025, "learning_rate": 9.571364656953661e-06, "loss": 0.5434, "step": 3788 }, { "epoch": 0.15882963226894983, "grad_norm": 2.0541176795959473, "learning_rate": 9.571089612596622e-06, "loss": 0.5867, "step": 3789 }, { "epoch": 0.1588715508839588, "grad_norm": 1.584069013595581, "learning_rate": 9.570814483977867e-06, "loss": 0.5316, "step": 3790 }, { "epoch": 0.15891346949896776, "grad_norm": 1.6109710931777954, "learning_rate": 9.570539271102464e-06, "loss": 0.5198, "step": 3791 }, { "epoch": 0.15895538811397672, "grad_norm": 2.2152886390686035, "learning_rate": 9.570263973975489e-06, "loss": 0.5623, "step": 3792 }, { "epoch": 0.15899730672898568, "grad_norm": 2.3932957649230957, "learning_rate": 9.569988592602016e-06, "loss": 0.5559, "step": 3793 }, { "epoch": 0.15903922534399464, "grad_norm": 1.7990238666534424, "learning_rate": 9.569713126987122e-06, "loss": 0.5386, "step": 3794 }, { "epoch": 0.1590811439590036, "grad_norm": 1.7115949392318726, "learning_rate": 9.569437577135882e-06, "loss": 0.5925, "step": 3795 }, { "epoch": 0.15912306257401254, "grad_norm": 1.883237600326538, "learning_rate": 9.569161943053378e-06, "loss": 0.5108, "step": 3796 }, { "epoch": 0.1591649811890215, "grad_norm": 1.6521517038345337, "learning_rate": 9.56888622474469e-06, "loss": 0.5049, "step": 3797 }, { "epoch": 0.15920689980403047, "grad_norm": 1.9579368829727173, "learning_rate": 9.568610422214902e-06, "loss": 0.5534, "step": 3798 }, { "epoch": 0.15924881841903943, "grad_norm": 1.9698405265808105, "learning_rate": 9.568334535469095e-06, "loss": 0.5862, "step": 3799 }, { "epoch": 0.1592907370340484, "grad_norm": 2.083745002746582, "learning_rate": 9.568058564512357e-06, "loss": 0.6058, "step": 3800 }, { "epoch": 0.15933265564905735, "grad_norm": 2.050325393676758, "learning_rate": 9.567782509349774e-06, "loss": 0.5631, "step": 3801 }, { "epoch": 0.15937457426406632, "grad_norm": 1.803671956062317, "learning_rate": 9.567506369986437e-06, "loss": 0.5426, "step": 3802 }, { "epoch": 0.15941649287907528, "grad_norm": 1.6849358081817627, "learning_rate": 9.567230146427434e-06, "loss": 0.5117, "step": 3803 }, { "epoch": 0.15945841149408424, "grad_norm": 2.172351121902466, "learning_rate": 9.566953838677856e-06, "loss": 0.5545, "step": 3804 }, { "epoch": 0.1595003301090932, "grad_norm": 2.298459053039551, "learning_rate": 9.5666774467428e-06, "loss": 0.4942, "step": 3805 }, { "epoch": 0.15954224872410216, "grad_norm": 2.2006683349609375, "learning_rate": 9.566400970627357e-06, "loss": 0.5808, "step": 3806 }, { "epoch": 0.15958416733911113, "grad_norm": 1.7031329870224, "learning_rate": 9.566124410336623e-06, "loss": 0.5395, "step": 3807 }, { "epoch": 0.1596260859541201, "grad_norm": 1.8506169319152832, "learning_rate": 9.5658477658757e-06, "loss": 0.574, "step": 3808 }, { "epoch": 0.15966800456912902, "grad_norm": 1.9156692028045654, "learning_rate": 9.565571037249686e-06, "loss": 0.637, "step": 3809 }, { "epoch": 0.159709923184138, "grad_norm": 1.909961462020874, "learning_rate": 9.565294224463682e-06, "loss": 0.5595, "step": 3810 }, { "epoch": 0.15975184179914695, "grad_norm": 2.0740323066711426, "learning_rate": 9.565017327522788e-06, "loss": 0.5397, "step": 3811 }, { "epoch": 0.1597937604141559, "grad_norm": 2.8854317665100098, "learning_rate": 9.564740346432113e-06, "loss": 0.5443, "step": 3812 }, { "epoch": 0.15983567902916487, "grad_norm": 3.0256268978118896, "learning_rate": 9.564463281196758e-06, "loss": 0.5833, "step": 3813 }, { "epoch": 0.15987759764417384, "grad_norm": 1.8254384994506836, "learning_rate": 9.564186131821832e-06, "loss": 0.5579, "step": 3814 }, { "epoch": 0.1599195162591828, "grad_norm": 1.6916579008102417, "learning_rate": 9.56390889831245e-06, "loss": 0.5603, "step": 3815 }, { "epoch": 0.15996143487419176, "grad_norm": 1.9806673526763916, "learning_rate": 9.56363158067371e-06, "loss": 0.5833, "step": 3816 }, { "epoch": 0.16000335348920072, "grad_norm": 2.0736775398254395, "learning_rate": 9.563354178910734e-06, "loss": 0.5702, "step": 3817 }, { "epoch": 0.16004527210420968, "grad_norm": 1.8829221725463867, "learning_rate": 9.563076693028633e-06, "loss": 0.5969, "step": 3818 }, { "epoch": 0.16008719071921865, "grad_norm": 1.7148159742355347, "learning_rate": 9.56279912303252e-06, "loss": 0.5925, "step": 3819 }, { "epoch": 0.1601291093342276, "grad_norm": 2.202219009399414, "learning_rate": 9.562521468927515e-06, "loss": 0.5665, "step": 3820 }, { "epoch": 0.16017102794923654, "grad_norm": 1.9756451845169067, "learning_rate": 9.562243730718733e-06, "loss": 0.5703, "step": 3821 }, { "epoch": 0.1602129465642455, "grad_norm": 2.4870471954345703, "learning_rate": 9.561965908411294e-06, "loss": 0.5776, "step": 3822 }, { "epoch": 0.16025486517925447, "grad_norm": 1.5995686054229736, "learning_rate": 9.561688002010322e-06, "loss": 0.5082, "step": 3823 }, { "epoch": 0.16029678379426343, "grad_norm": 1.9989644289016724, "learning_rate": 9.561410011520936e-06, "loss": 0.4713, "step": 3824 }, { "epoch": 0.1603387024092724, "grad_norm": 1.7842607498168945, "learning_rate": 9.561131936948263e-06, "loss": 0.5815, "step": 3825 }, { "epoch": 0.16038062102428136, "grad_norm": 1.8924006223678589, "learning_rate": 9.56085377829743e-06, "loss": 0.568, "step": 3826 }, { "epoch": 0.16042253963929032, "grad_norm": 4.696589946746826, "learning_rate": 9.56057553557356e-06, "loss": 0.5266, "step": 3827 }, { "epoch": 0.16046445825429928, "grad_norm": 1.6432007551193237, "learning_rate": 9.560297208781787e-06, "loss": 0.528, "step": 3828 }, { "epoch": 0.16050637686930824, "grad_norm": 1.7625523805618286, "learning_rate": 9.560018797927238e-06, "loss": 0.6025, "step": 3829 }, { "epoch": 0.1605482954843172, "grad_norm": 1.670540452003479, "learning_rate": 9.559740303015048e-06, "loss": 0.5589, "step": 3830 }, { "epoch": 0.16059021409932617, "grad_norm": 1.8531676530838013, "learning_rate": 9.559461724050349e-06, "loss": 0.5996, "step": 3831 }, { "epoch": 0.16063213271433513, "grad_norm": 1.6450356245040894, "learning_rate": 9.559183061038275e-06, "loss": 0.5575, "step": 3832 }, { "epoch": 0.1606740513293441, "grad_norm": 1.9540183544158936, "learning_rate": 9.558904313983965e-06, "loss": 0.5945, "step": 3833 }, { "epoch": 0.16071596994435303, "grad_norm": 1.755632758140564, "learning_rate": 9.558625482892557e-06, "loss": 0.5855, "step": 3834 }, { "epoch": 0.160757888559362, "grad_norm": 1.7450734376907349, "learning_rate": 9.55834656776919e-06, "loss": 0.4882, "step": 3835 }, { "epoch": 0.16079980717437095, "grad_norm": 1.9279693365097046, "learning_rate": 9.558067568619005e-06, "loss": 0.575, "step": 3836 }, { "epoch": 0.1608417257893799, "grad_norm": 2.352583646774292, "learning_rate": 9.557788485447146e-06, "loss": 0.6395, "step": 3837 }, { "epoch": 0.16088364440438888, "grad_norm": 1.883034110069275, "learning_rate": 9.557509318258758e-06, "loss": 0.5177, "step": 3838 }, { "epoch": 0.16092556301939784, "grad_norm": 2.1019461154937744, "learning_rate": 9.557230067058984e-06, "loss": 0.6051, "step": 3839 }, { "epoch": 0.1609674816344068, "grad_norm": 1.8095073699951172, "learning_rate": 9.556950731852977e-06, "loss": 0.5669, "step": 3840 }, { "epoch": 0.16100940024941576, "grad_norm": 1.9703325033187866, "learning_rate": 9.556671312645882e-06, "loss": 0.5725, "step": 3841 }, { "epoch": 0.16105131886442473, "grad_norm": 2.2367091178894043, "learning_rate": 9.556391809442849e-06, "loss": 0.5549, "step": 3842 }, { "epoch": 0.1610932374794337, "grad_norm": 1.703377366065979, "learning_rate": 9.556112222249034e-06, "loss": 0.5344, "step": 3843 }, { "epoch": 0.16113515609444265, "grad_norm": 2.0173962116241455, "learning_rate": 9.555832551069588e-06, "loss": 0.5508, "step": 3844 }, { "epoch": 0.1611770747094516, "grad_norm": 1.692087173461914, "learning_rate": 9.555552795909668e-06, "loss": 0.4837, "step": 3845 }, { "epoch": 0.16121899332446055, "grad_norm": 50.822731018066406, "learning_rate": 9.555272956774429e-06, "loss": 0.5573, "step": 3846 }, { "epoch": 0.1612609119394695, "grad_norm": 1.704083800315857, "learning_rate": 9.554993033669031e-06, "loss": 0.5489, "step": 3847 }, { "epoch": 0.16130283055447847, "grad_norm": 1.5825273990631104, "learning_rate": 9.554713026598634e-06, "loss": 0.522, "step": 3848 }, { "epoch": 0.16134474916948743, "grad_norm": 1.7678847312927246, "learning_rate": 9.554432935568399e-06, "loss": 0.5836, "step": 3849 }, { "epoch": 0.1613866677844964, "grad_norm": 1.8840304613113403, "learning_rate": 9.554152760583488e-06, "loss": 0.6448, "step": 3850 }, { "epoch": 0.16142858639950536, "grad_norm": 1.6900873184204102, "learning_rate": 9.553872501649068e-06, "loss": 0.5739, "step": 3851 }, { "epoch": 0.16147050501451432, "grad_norm": 2.490495204925537, "learning_rate": 9.553592158770304e-06, "loss": 0.6141, "step": 3852 }, { "epoch": 0.16151242362952328, "grad_norm": 2.0844154357910156, "learning_rate": 9.553311731952363e-06, "loss": 0.6174, "step": 3853 }, { "epoch": 0.16155434224453225, "grad_norm": 1.5341248512268066, "learning_rate": 9.553031221200414e-06, "loss": 0.609, "step": 3854 }, { "epoch": 0.1615962608595412, "grad_norm": 1.7946189641952515, "learning_rate": 9.552750626519632e-06, "loss": 0.5966, "step": 3855 }, { "epoch": 0.16163817947455017, "grad_norm": 1.7643165588378906, "learning_rate": 9.552469947915183e-06, "loss": 0.6275, "step": 3856 }, { "epoch": 0.16168009808955913, "grad_norm": 1.803971529006958, "learning_rate": 9.552189185392246e-06, "loss": 0.5949, "step": 3857 }, { "epoch": 0.1617220167045681, "grad_norm": 1.8188567161560059, "learning_rate": 9.551908338955993e-06, "loss": 0.5946, "step": 3858 }, { "epoch": 0.16176393531957703, "grad_norm": 1.8284708261489868, "learning_rate": 9.551627408611603e-06, "loss": 0.5839, "step": 3859 }, { "epoch": 0.161805853934586, "grad_norm": 2.058468818664551, "learning_rate": 9.551346394364256e-06, "loss": 0.5949, "step": 3860 }, { "epoch": 0.16184777254959495, "grad_norm": 1.9150406122207642, "learning_rate": 9.55106529621913e-06, "loss": 0.5889, "step": 3861 }, { "epoch": 0.16188969116460392, "grad_norm": 2.658912420272827, "learning_rate": 9.550784114181405e-06, "loss": 0.6003, "step": 3862 }, { "epoch": 0.16193160977961288, "grad_norm": 1.6406817436218262, "learning_rate": 9.550502848256266e-06, "loss": 0.5731, "step": 3863 }, { "epoch": 0.16197352839462184, "grad_norm": 1.9219557046890259, "learning_rate": 9.550221498448899e-06, "loss": 0.6022, "step": 3864 }, { "epoch": 0.1620154470096308, "grad_norm": 1.7378085851669312, "learning_rate": 9.549940064764489e-06, "loss": 0.5963, "step": 3865 }, { "epoch": 0.16205736562463977, "grad_norm": 1.67043936252594, "learning_rate": 9.549658547208223e-06, "loss": 0.5954, "step": 3866 }, { "epoch": 0.16209928423964873, "grad_norm": 2.3495125770568848, "learning_rate": 9.549376945785293e-06, "loss": 0.6058, "step": 3867 }, { "epoch": 0.1621412028546577, "grad_norm": 1.7708333730697632, "learning_rate": 9.549095260500888e-06, "loss": 0.5563, "step": 3868 }, { "epoch": 0.16218312146966665, "grad_norm": 1.779083251953125, "learning_rate": 9.5488134913602e-06, "loss": 0.5928, "step": 3869 }, { "epoch": 0.16222504008467561, "grad_norm": 1.6858950853347778, "learning_rate": 9.548531638368425e-06, "loss": 0.6303, "step": 3870 }, { "epoch": 0.16226695869968455, "grad_norm": 1.5905123949050903, "learning_rate": 9.548249701530756e-06, "loss": 0.5517, "step": 3871 }, { "epoch": 0.1623088773146935, "grad_norm": 1.5732449293136597, "learning_rate": 9.547967680852392e-06, "loss": 0.5344, "step": 3872 }, { "epoch": 0.16235079592970247, "grad_norm": 1.8314744234085083, "learning_rate": 9.547685576338533e-06, "loss": 0.5426, "step": 3873 }, { "epoch": 0.16239271454471144, "grad_norm": 1.705950379371643, "learning_rate": 9.547403387994376e-06, "loss": 0.5146, "step": 3874 }, { "epoch": 0.1624346331597204, "grad_norm": 1.5402175188064575, "learning_rate": 9.547121115825123e-06, "loss": 0.5036, "step": 3875 }, { "epoch": 0.16247655177472936, "grad_norm": 1.513062596321106, "learning_rate": 9.54683875983598e-06, "loss": 0.5349, "step": 3876 }, { "epoch": 0.16251847038973832, "grad_norm": 1.9806139469146729, "learning_rate": 9.54655632003215e-06, "loss": 0.5375, "step": 3877 }, { "epoch": 0.16256038900474729, "grad_norm": 2.117870569229126, "learning_rate": 9.546273796418839e-06, "loss": 0.6051, "step": 3878 }, { "epoch": 0.16260230761975625, "grad_norm": 1.9133844375610352, "learning_rate": 9.545991189001255e-06, "loss": 0.5976, "step": 3879 }, { "epoch": 0.1626442262347652, "grad_norm": 1.721041202545166, "learning_rate": 9.54570849778461e-06, "loss": 0.5999, "step": 3880 }, { "epoch": 0.16268614484977417, "grad_norm": 1.6858690977096558, "learning_rate": 9.545425722774112e-06, "loss": 0.5209, "step": 3881 }, { "epoch": 0.16272806346478313, "grad_norm": 1.838021993637085, "learning_rate": 9.545142863974974e-06, "loss": 0.6037, "step": 3882 }, { "epoch": 0.1627699820797921, "grad_norm": 1.8225175142288208, "learning_rate": 9.544859921392413e-06, "loss": 0.6664, "step": 3883 }, { "epoch": 0.16281190069480103, "grad_norm": 2.5278382301330566, "learning_rate": 9.54457689503164e-06, "loss": 0.5442, "step": 3884 }, { "epoch": 0.16285381930981, "grad_norm": 1.717858076095581, "learning_rate": 9.544293784897875e-06, "loss": 0.5568, "step": 3885 }, { "epoch": 0.16289573792481896, "grad_norm": 1.7673262357711792, "learning_rate": 9.544010590996337e-06, "loss": 0.5632, "step": 3886 }, { "epoch": 0.16293765653982792, "grad_norm": 1.624772071838379, "learning_rate": 9.543727313332248e-06, "loss": 0.5211, "step": 3887 }, { "epoch": 0.16297957515483688, "grad_norm": 1.8260189294815063, "learning_rate": 9.543443951910825e-06, "loss": 0.5482, "step": 3888 }, { "epoch": 0.16302149376984584, "grad_norm": 1.6374313831329346, "learning_rate": 9.543160506737292e-06, "loss": 0.5797, "step": 3889 }, { "epoch": 0.1630634123848548, "grad_norm": 1.757698655128479, "learning_rate": 9.54287697781688e-06, "loss": 0.5426, "step": 3890 }, { "epoch": 0.16310533099986377, "grad_norm": 1.6908364295959473, "learning_rate": 9.542593365154809e-06, "loss": 0.5584, "step": 3891 }, { "epoch": 0.16314724961487273, "grad_norm": 1.8844349384307861, "learning_rate": 9.542309668756308e-06, "loss": 0.5754, "step": 3892 }, { "epoch": 0.1631891682298817, "grad_norm": 1.6193214654922485, "learning_rate": 9.54202588862661e-06, "loss": 0.5978, "step": 3893 }, { "epoch": 0.16323108684489065, "grad_norm": 2.0435688495635986, "learning_rate": 9.541742024770942e-06, "loss": 0.5659, "step": 3894 }, { "epoch": 0.16327300545989962, "grad_norm": 1.4669575691223145, "learning_rate": 9.541458077194541e-06, "loss": 0.5547, "step": 3895 }, { "epoch": 0.16331492407490855, "grad_norm": 1.5109539031982422, "learning_rate": 9.541174045902636e-06, "loss": 0.5552, "step": 3896 }, { "epoch": 0.16335684268991751, "grad_norm": 1.6038033962249756, "learning_rate": 9.540889930900468e-06, "loss": 0.5345, "step": 3897 }, { "epoch": 0.16339876130492648, "grad_norm": 1.6688575744628906, "learning_rate": 9.54060573219327e-06, "loss": 0.6145, "step": 3898 }, { "epoch": 0.16344067991993544, "grad_norm": 1.8421926498413086, "learning_rate": 9.540321449786282e-06, "loss": 0.6057, "step": 3899 }, { "epoch": 0.1634825985349444, "grad_norm": 1.801652431488037, "learning_rate": 9.540037083684747e-06, "loss": 0.6144, "step": 3900 }, { "epoch": 0.16352451714995336, "grad_norm": 1.5844364166259766, "learning_rate": 9.539752633893905e-06, "loss": 0.5403, "step": 3901 }, { "epoch": 0.16356643576496233, "grad_norm": 1.4686857461929321, "learning_rate": 9.539468100418996e-06, "loss": 0.5815, "step": 3902 }, { "epoch": 0.1636083543799713, "grad_norm": 2.1186606884002686, "learning_rate": 9.53918348326527e-06, "loss": 0.619, "step": 3903 }, { "epoch": 0.16365027299498025, "grad_norm": 1.6052534580230713, "learning_rate": 9.538898782437971e-06, "loss": 0.5946, "step": 3904 }, { "epoch": 0.1636921916099892, "grad_norm": 1.5234466791152954, "learning_rate": 9.538613997942349e-06, "loss": 0.5511, "step": 3905 }, { "epoch": 0.16373411022499818, "grad_norm": 1.4575718641281128, "learning_rate": 9.538329129783653e-06, "loss": 0.5472, "step": 3906 }, { "epoch": 0.16377602884000714, "grad_norm": 1.5992494821548462, "learning_rate": 9.538044177967133e-06, "loss": 0.5529, "step": 3907 }, { "epoch": 0.1638179474550161, "grad_norm": 1.527538776397705, "learning_rate": 9.537759142498042e-06, "loss": 0.5459, "step": 3908 }, { "epoch": 0.16385986607002503, "grad_norm": 1.611007571220398, "learning_rate": 9.537474023381633e-06, "loss": 0.5797, "step": 3909 }, { "epoch": 0.163901784685034, "grad_norm": 1.7746349573135376, "learning_rate": 9.537188820623164e-06, "loss": 0.6204, "step": 3910 }, { "epoch": 0.16394370330004296, "grad_norm": 1.6185413599014282, "learning_rate": 9.536903534227893e-06, "loss": 0.6075, "step": 3911 }, { "epoch": 0.16398562191505192, "grad_norm": 1.6562185287475586, "learning_rate": 9.536618164201078e-06, "loss": 0.5045, "step": 3912 }, { "epoch": 0.16402754053006088, "grad_norm": 1.5407596826553345, "learning_rate": 9.536332710547975e-06, "loss": 0.5065, "step": 3913 }, { "epoch": 0.16406945914506985, "grad_norm": 1.6609138250350952, "learning_rate": 9.536047173273852e-06, "loss": 0.5347, "step": 3914 }, { "epoch": 0.1641113777600788, "grad_norm": 1.9054301977157593, "learning_rate": 9.53576155238397e-06, "loss": 0.6034, "step": 3915 }, { "epoch": 0.16415329637508777, "grad_norm": 1.7172237634658813, "learning_rate": 9.535475847883596e-06, "loss": 0.5687, "step": 3916 }, { "epoch": 0.16419521499009673, "grad_norm": 1.680903434753418, "learning_rate": 9.535190059777992e-06, "loss": 0.5656, "step": 3917 }, { "epoch": 0.1642371336051057, "grad_norm": 1.598634958267212, "learning_rate": 9.534904188072431e-06, "loss": 0.5138, "step": 3918 }, { "epoch": 0.16427905222011466, "grad_norm": 1.9375333786010742, "learning_rate": 9.53461823277218e-06, "loss": 0.585, "step": 3919 }, { "epoch": 0.16432097083512362, "grad_norm": 2.5291831493377686, "learning_rate": 9.534332193882511e-06, "loss": 0.6042, "step": 3920 }, { "epoch": 0.16436288945013255, "grad_norm": 1.7855141162872314, "learning_rate": 9.534046071408697e-06, "loss": 0.5704, "step": 3921 }, { "epoch": 0.16440480806514152, "grad_norm": 1.5874501466751099, "learning_rate": 9.533759865356011e-06, "loss": 0.523, "step": 3922 }, { "epoch": 0.16444672668015048, "grad_norm": 1.5464731454849243, "learning_rate": 9.53347357572973e-06, "loss": 0.5382, "step": 3923 }, { "epoch": 0.16448864529515944, "grad_norm": 1.8571645021438599, "learning_rate": 9.53318720253513e-06, "loss": 0.6105, "step": 3924 }, { "epoch": 0.1645305639101684, "grad_norm": 1.8690367937088013, "learning_rate": 9.532900745777492e-06, "loss": 0.5315, "step": 3925 }, { "epoch": 0.16457248252517737, "grad_norm": 1.8079952001571655, "learning_rate": 9.532614205462094e-06, "loss": 0.5454, "step": 3926 }, { "epoch": 0.16461440114018633, "grad_norm": 1.4619544744491577, "learning_rate": 9.53232758159422e-06, "loss": 0.486, "step": 3927 }, { "epoch": 0.1646563197551953, "grad_norm": 1.520977258682251, "learning_rate": 9.532040874179154e-06, "loss": 0.5737, "step": 3928 }, { "epoch": 0.16469823837020425, "grad_norm": 1.647745966911316, "learning_rate": 9.53175408322218e-06, "loss": 0.5532, "step": 3929 }, { "epoch": 0.16474015698521322, "grad_norm": 2.147307872772217, "learning_rate": 9.531467208728583e-06, "loss": 0.6735, "step": 3930 }, { "epoch": 0.16478207560022218, "grad_norm": 1.6010899543762207, "learning_rate": 9.531180250703653e-06, "loss": 0.6314, "step": 3931 }, { "epoch": 0.16482399421523114, "grad_norm": 1.5793850421905518, "learning_rate": 9.530893209152678e-06, "loss": 0.5821, "step": 3932 }, { "epoch": 0.1648659128302401, "grad_norm": 1.6562762260437012, "learning_rate": 9.53060608408095e-06, "loss": 0.5904, "step": 3933 }, { "epoch": 0.16490783144524904, "grad_norm": 1.6667640209197998, "learning_rate": 9.530318875493764e-06, "loss": 0.5875, "step": 3934 }, { "epoch": 0.164949750060258, "grad_norm": 1.6029243469238281, "learning_rate": 9.530031583396412e-06, "loss": 0.5639, "step": 3935 }, { "epoch": 0.16499166867526696, "grad_norm": 2.0428576469421387, "learning_rate": 9.529744207794192e-06, "loss": 0.523, "step": 3936 }, { "epoch": 0.16503358729027592, "grad_norm": 1.994036078453064, "learning_rate": 9.529456748692399e-06, "loss": 0.6048, "step": 3937 }, { "epoch": 0.16507550590528489, "grad_norm": 1.7350592613220215, "learning_rate": 9.529169206096332e-06, "loss": 0.5904, "step": 3938 }, { "epoch": 0.16511742452029385, "grad_norm": 1.8887557983398438, "learning_rate": 9.528881580011292e-06, "loss": 0.5059, "step": 3939 }, { "epoch": 0.1651593431353028, "grad_norm": 1.6731218099594116, "learning_rate": 9.528593870442581e-06, "loss": 0.5563, "step": 3940 }, { "epoch": 0.16520126175031177, "grad_norm": 1.5241636037826538, "learning_rate": 9.528306077395502e-06, "loss": 0.5556, "step": 3941 }, { "epoch": 0.16524318036532074, "grad_norm": 1.722145915031433, "learning_rate": 9.528018200875361e-06, "loss": 0.5831, "step": 3942 }, { "epoch": 0.1652850989803297, "grad_norm": 1.5148299932479858, "learning_rate": 9.527730240887465e-06, "loss": 0.545, "step": 3943 }, { "epoch": 0.16532701759533866, "grad_norm": 1.6771410703659058, "learning_rate": 9.527442197437122e-06, "loss": 0.5553, "step": 3944 }, { "epoch": 0.16536893621034762, "grad_norm": 2.441535234451294, "learning_rate": 9.52715407052964e-06, "loss": 0.5665, "step": 3945 }, { "epoch": 0.16541085482535656, "grad_norm": 1.6628941297531128, "learning_rate": 9.526865860170331e-06, "loss": 0.5525, "step": 3946 }, { "epoch": 0.16545277344036552, "grad_norm": 1.716788649559021, "learning_rate": 9.52657756636451e-06, "loss": 0.5456, "step": 3947 }, { "epoch": 0.16549469205537448, "grad_norm": 1.710870385169983, "learning_rate": 9.526289189117485e-06, "loss": 0.5558, "step": 3948 }, { "epoch": 0.16553661067038344, "grad_norm": 1.6623051166534424, "learning_rate": 9.52600072843458e-06, "loss": 0.5321, "step": 3949 }, { "epoch": 0.1655785292853924, "grad_norm": 1.7342112064361572, "learning_rate": 9.525712184321107e-06, "loss": 0.5522, "step": 3950 }, { "epoch": 0.16562044790040137, "grad_norm": 1.5824306011199951, "learning_rate": 9.525423556782388e-06, "loss": 0.5661, "step": 3951 }, { "epoch": 0.16566236651541033, "grad_norm": 1.6158545017242432, "learning_rate": 9.525134845823741e-06, "loss": 0.5605, "step": 3952 }, { "epoch": 0.1657042851304193, "grad_norm": 1.9446595907211304, "learning_rate": 9.524846051450489e-06, "loss": 0.5621, "step": 3953 }, { "epoch": 0.16574620374542826, "grad_norm": 1.623948097229004, "learning_rate": 9.524557173667957e-06, "loss": 0.5356, "step": 3954 }, { "epoch": 0.16578812236043722, "grad_norm": 1.5228294134140015, "learning_rate": 9.524268212481466e-06, "loss": 0.5789, "step": 3955 }, { "epoch": 0.16583004097544618, "grad_norm": 1.4409551620483398, "learning_rate": 9.523979167896346e-06, "loss": 0.5496, "step": 3956 }, { "epoch": 0.16587195959045514, "grad_norm": 1.8698840141296387, "learning_rate": 9.523690039917924e-06, "loss": 0.513, "step": 3957 }, { "epoch": 0.1659138782054641, "grad_norm": 1.5599713325500488, "learning_rate": 9.52340082855153e-06, "loss": 0.5577, "step": 3958 }, { "epoch": 0.16595579682047304, "grad_norm": 1.7022520303726196, "learning_rate": 9.523111533802497e-06, "loss": 0.6166, "step": 3959 }, { "epoch": 0.165997715435482, "grad_norm": 1.8795865774154663, "learning_rate": 9.522822155676153e-06, "loss": 0.5215, "step": 3960 }, { "epoch": 0.16603963405049096, "grad_norm": 1.6119956970214844, "learning_rate": 9.522532694177837e-06, "loss": 0.6039, "step": 3961 }, { "epoch": 0.16608155266549993, "grad_norm": 1.7499974966049194, "learning_rate": 9.522243149312882e-06, "loss": 0.5262, "step": 3962 }, { "epoch": 0.1661234712805089, "grad_norm": 1.7275538444519043, "learning_rate": 9.521953521086627e-06, "loss": 0.5419, "step": 3963 }, { "epoch": 0.16616538989551785, "grad_norm": 1.7333602905273438, "learning_rate": 9.52166380950441e-06, "loss": 0.5552, "step": 3964 }, { "epoch": 0.1662073085105268, "grad_norm": 1.572391152381897, "learning_rate": 9.521374014571571e-06, "loss": 0.5749, "step": 3965 }, { "epoch": 0.16624922712553578, "grad_norm": 1.577039122581482, "learning_rate": 9.521084136293455e-06, "loss": 0.5556, "step": 3966 }, { "epoch": 0.16629114574054474, "grad_norm": 1.5421100854873657, "learning_rate": 9.520794174675401e-06, "loss": 0.5861, "step": 3967 }, { "epoch": 0.1663330643555537, "grad_norm": 1.6243157386779785, "learning_rate": 9.520504129722756e-06, "loss": 0.5408, "step": 3968 }, { "epoch": 0.16637498297056266, "grad_norm": 1.7298216819763184, "learning_rate": 9.520214001440868e-06, "loss": 0.5992, "step": 3969 }, { "epoch": 0.16641690158557162, "grad_norm": 2.8518173694610596, "learning_rate": 9.519923789835084e-06, "loss": 0.568, "step": 3970 }, { "epoch": 0.16645882020058056, "grad_norm": 7.076637268066406, "learning_rate": 9.519633494910753e-06, "loss": 0.5482, "step": 3971 }, { "epoch": 0.16650073881558952, "grad_norm": 1.7226096391677856, "learning_rate": 9.519343116673227e-06, "loss": 0.6513, "step": 3972 }, { "epoch": 0.16654265743059848, "grad_norm": 2.020914077758789, "learning_rate": 9.519052655127857e-06, "loss": 0.5759, "step": 3973 }, { "epoch": 0.16658457604560745, "grad_norm": 1.7246261835098267, "learning_rate": 9.51876211028e-06, "loss": 0.5294, "step": 3974 }, { "epoch": 0.1666264946606164, "grad_norm": 1.7066333293914795, "learning_rate": 9.51847148213501e-06, "loss": 0.5804, "step": 3975 }, { "epoch": 0.16666841327562537, "grad_norm": 1.5545870065689087, "learning_rate": 9.518180770698247e-06, "loss": 0.5918, "step": 3976 }, { "epoch": 0.16671033189063433, "grad_norm": 1.6846165657043457, "learning_rate": 9.517889975975066e-06, "loss": 0.5631, "step": 3977 }, { "epoch": 0.1667522505056433, "grad_norm": 1.712546467781067, "learning_rate": 9.517599097970828e-06, "loss": 0.5088, "step": 3978 }, { "epoch": 0.16679416912065226, "grad_norm": 1.7562733888626099, "learning_rate": 9.517308136690897e-06, "loss": 0.5294, "step": 3979 }, { "epoch": 0.16683608773566122, "grad_norm": 1.690037727355957, "learning_rate": 9.517017092140634e-06, "loss": 0.5326, "step": 3980 }, { "epoch": 0.16687800635067018, "grad_norm": 1.80409836769104, "learning_rate": 9.516725964325408e-06, "loss": 0.567, "step": 3981 }, { "epoch": 0.16691992496567915, "grad_norm": 1.806148648262024, "learning_rate": 9.516434753250583e-06, "loss": 0.5662, "step": 3982 }, { "epoch": 0.1669618435806881, "grad_norm": 1.6273219585418701, "learning_rate": 9.516143458921525e-06, "loss": 0.5568, "step": 3983 }, { "epoch": 0.16700376219569704, "grad_norm": 1.9014837741851807, "learning_rate": 9.515852081343606e-06, "loss": 0.621, "step": 3984 }, { "epoch": 0.167045680810706, "grad_norm": 2.4708409309387207, "learning_rate": 9.5155606205222e-06, "loss": 0.5636, "step": 3985 }, { "epoch": 0.16708759942571497, "grad_norm": 1.5493369102478027, "learning_rate": 9.515269076462673e-06, "loss": 0.544, "step": 3986 }, { "epoch": 0.16712951804072393, "grad_norm": 1.616898536682129, "learning_rate": 9.514977449170405e-06, "loss": 0.5506, "step": 3987 }, { "epoch": 0.1671714366557329, "grad_norm": 2.206000328063965, "learning_rate": 9.514685738650768e-06, "loss": 0.547, "step": 3988 }, { "epoch": 0.16721335527074185, "grad_norm": 1.8471266031265259, "learning_rate": 9.51439394490914e-06, "loss": 0.5721, "step": 3989 }, { "epoch": 0.16725527388575082, "grad_norm": 2.1560401916503906, "learning_rate": 9.514102067950904e-06, "loss": 0.6248, "step": 3990 }, { "epoch": 0.16729719250075978, "grad_norm": 1.8150615692138672, "learning_rate": 9.513810107781433e-06, "loss": 0.596, "step": 3991 }, { "epoch": 0.16733911111576874, "grad_norm": 1.798609972000122, "learning_rate": 9.513518064406116e-06, "loss": 0.5843, "step": 3992 }, { "epoch": 0.1673810297307777, "grad_norm": 1.6623927354812622, "learning_rate": 9.51322593783033e-06, "loss": 0.5553, "step": 3993 }, { "epoch": 0.16742294834578667, "grad_norm": 1.6960933208465576, "learning_rate": 9.512933728059465e-06, "loss": 0.5053, "step": 3994 }, { "epoch": 0.16746486696079563, "grad_norm": 2.503866195678711, "learning_rate": 9.512641435098905e-06, "loss": 0.5698, "step": 3995 }, { "epoch": 0.16750678557580456, "grad_norm": 1.7783392667770386, "learning_rate": 9.51234905895404e-06, "loss": 0.5645, "step": 3996 }, { "epoch": 0.16754870419081352, "grad_norm": 1.712973952293396, "learning_rate": 9.512056599630257e-06, "loss": 0.5392, "step": 3997 }, { "epoch": 0.1675906228058225, "grad_norm": 1.9385639429092407, "learning_rate": 9.51176405713295e-06, "loss": 0.5973, "step": 3998 }, { "epoch": 0.16763254142083145, "grad_norm": 2.0453238487243652, "learning_rate": 9.51147143146751e-06, "loss": 0.5821, "step": 3999 }, { "epoch": 0.1676744600358404, "grad_norm": 1.702113151550293, "learning_rate": 9.51117872263933e-06, "loss": 0.5931, "step": 4000 }, { "epoch": 0.16771637865084937, "grad_norm": 2.0292458534240723, "learning_rate": 9.510885930653806e-06, "loss": 0.4836, "step": 4001 }, { "epoch": 0.16775829726585834, "grad_norm": 1.672580361366272, "learning_rate": 9.510593055516338e-06, "loss": 0.5445, "step": 4002 }, { "epoch": 0.1678002158808673, "grad_norm": 1.9556745290756226, "learning_rate": 9.510300097232321e-06, "loss": 0.5474, "step": 4003 }, { "epoch": 0.16784213449587626, "grad_norm": 3.345720052719116, "learning_rate": 9.510007055807157e-06, "loss": 0.5737, "step": 4004 }, { "epoch": 0.16788405311088522, "grad_norm": 1.8089849948883057, "learning_rate": 9.509713931246247e-06, "loss": 0.5308, "step": 4005 }, { "epoch": 0.16792597172589419, "grad_norm": 1.791882038116455, "learning_rate": 9.509420723554997e-06, "loss": 0.5544, "step": 4006 }, { "epoch": 0.16796789034090315, "grad_norm": 1.7504711151123047, "learning_rate": 9.50912743273881e-06, "loss": 0.5625, "step": 4007 }, { "epoch": 0.1680098089559121, "grad_norm": 1.779803991317749, "learning_rate": 9.508834058803091e-06, "loss": 0.5799, "step": 4008 }, { "epoch": 0.16805172757092104, "grad_norm": 1.6940441131591797, "learning_rate": 9.508540601753251e-06, "loss": 0.5066, "step": 4009 }, { "epoch": 0.16809364618593, "grad_norm": 1.763575792312622, "learning_rate": 9.508247061594698e-06, "loss": 0.5736, "step": 4010 }, { "epoch": 0.16813556480093897, "grad_norm": 2.1348419189453125, "learning_rate": 9.507953438332841e-06, "loss": 0.6556, "step": 4011 }, { "epoch": 0.16817748341594793, "grad_norm": 2.049924373626709, "learning_rate": 9.507659731973096e-06, "loss": 0.5802, "step": 4012 }, { "epoch": 0.1682194020309569, "grad_norm": 1.7745226621627808, "learning_rate": 9.507365942520875e-06, "loss": 0.5486, "step": 4013 }, { "epoch": 0.16826132064596586, "grad_norm": 2.099625825881958, "learning_rate": 9.507072069981594e-06, "loss": 0.5951, "step": 4014 }, { "epoch": 0.16830323926097482, "grad_norm": 1.8683924674987793, "learning_rate": 9.50677811436067e-06, "loss": 0.5889, "step": 4015 }, { "epoch": 0.16834515787598378, "grad_norm": 1.6916286945343018, "learning_rate": 9.506484075663523e-06, "loss": 0.5083, "step": 4016 }, { "epoch": 0.16838707649099274, "grad_norm": 1.9321045875549316, "learning_rate": 9.50618995389557e-06, "loss": 0.601, "step": 4017 }, { "epoch": 0.1684289951060017, "grad_norm": 1.7246087789535522, "learning_rate": 9.505895749062236e-06, "loss": 0.5986, "step": 4018 }, { "epoch": 0.16847091372101067, "grad_norm": 1.844767451286316, "learning_rate": 9.505601461168943e-06, "loss": 0.6, "step": 4019 }, { "epoch": 0.16851283233601963, "grad_norm": 1.7041199207305908, "learning_rate": 9.505307090221116e-06, "loss": 0.5953, "step": 4020 }, { "epoch": 0.16855475095102856, "grad_norm": 1.7197774648666382, "learning_rate": 9.505012636224183e-06, "loss": 0.588, "step": 4021 }, { "epoch": 0.16859666956603753, "grad_norm": 1.6173204183578491, "learning_rate": 9.504718099183567e-06, "loss": 0.5315, "step": 4022 }, { "epoch": 0.1686385881810465, "grad_norm": 2.1268961429595947, "learning_rate": 9.504423479104701e-06, "loss": 0.5809, "step": 4023 }, { "epoch": 0.16868050679605545, "grad_norm": 1.660476803779602, "learning_rate": 9.504128775993018e-06, "loss": 0.5248, "step": 4024 }, { "epoch": 0.16872242541106441, "grad_norm": 1.7985584735870361, "learning_rate": 9.503833989853946e-06, "loss": 0.6187, "step": 4025 }, { "epoch": 0.16876434402607338, "grad_norm": 1.6182332038879395, "learning_rate": 9.50353912069292e-06, "loss": 0.5459, "step": 4026 }, { "epoch": 0.16880626264108234, "grad_norm": 1.6211329698562622, "learning_rate": 9.503244168515376e-06, "loss": 0.5583, "step": 4027 }, { "epoch": 0.1688481812560913, "grad_norm": 1.7962048053741455, "learning_rate": 9.502949133326753e-06, "loss": 0.5536, "step": 4028 }, { "epoch": 0.16889009987110026, "grad_norm": 1.687767744064331, "learning_rate": 9.502654015132486e-06, "loss": 0.6111, "step": 4029 }, { "epoch": 0.16893201848610923, "grad_norm": 1.6950584650039673, "learning_rate": 9.502358813938019e-06, "loss": 0.5342, "step": 4030 }, { "epoch": 0.1689739371011182, "grad_norm": 1.8178935050964355, "learning_rate": 9.50206352974879e-06, "loss": 0.6188, "step": 4031 }, { "epoch": 0.16901585571612715, "grad_norm": 1.6212096214294434, "learning_rate": 9.501768162570244e-06, "loss": 0.5734, "step": 4032 }, { "epoch": 0.1690577743311361, "grad_norm": 1.6474316120147705, "learning_rate": 9.501472712407827e-06, "loss": 0.482, "step": 4033 }, { "epoch": 0.16909969294614505, "grad_norm": 1.7824865579605103, "learning_rate": 9.501177179266983e-06, "loss": 0.6171, "step": 4034 }, { "epoch": 0.169141611561154, "grad_norm": 2.432032346725464, "learning_rate": 9.500881563153159e-06, "loss": 0.589, "step": 4035 }, { "epoch": 0.16918353017616297, "grad_norm": 1.8206559419631958, "learning_rate": 9.500585864071807e-06, "loss": 0.5398, "step": 4036 }, { "epoch": 0.16922544879117193, "grad_norm": 1.594477891921997, "learning_rate": 9.500290082028375e-06, "loss": 0.5391, "step": 4037 }, { "epoch": 0.1692673674061809, "grad_norm": 1.7598116397857666, "learning_rate": 9.499994217028317e-06, "loss": 0.5302, "step": 4038 }, { "epoch": 0.16930928602118986, "grad_norm": 1.6189522743225098, "learning_rate": 9.49969826907709e-06, "loss": 0.5306, "step": 4039 }, { "epoch": 0.16935120463619882, "grad_norm": 4.081407070159912, "learning_rate": 9.499402238180142e-06, "loss": 0.4768, "step": 4040 }, { "epoch": 0.16939312325120778, "grad_norm": 1.8564001321792603, "learning_rate": 9.499106124342936e-06, "loss": 0.6425, "step": 4041 }, { "epoch": 0.16943504186621675, "grad_norm": 1.5930490493774414, "learning_rate": 9.498809927570926e-06, "loss": 0.5045, "step": 4042 }, { "epoch": 0.1694769604812257, "grad_norm": 2.0204339027404785, "learning_rate": 9.498513647869578e-06, "loss": 0.5519, "step": 4043 }, { "epoch": 0.16951887909623467, "grad_norm": 2.4470901489257812, "learning_rate": 9.498217285244349e-06, "loss": 0.5725, "step": 4044 }, { "epoch": 0.16956079771124363, "grad_norm": 1.6985284090042114, "learning_rate": 9.4979208397007e-06, "loss": 0.5579, "step": 4045 }, { "epoch": 0.16960271632625257, "grad_norm": 1.854931116104126, "learning_rate": 9.4976243112441e-06, "loss": 0.559, "step": 4046 }, { "epoch": 0.16964463494126153, "grad_norm": 1.968806505203247, "learning_rate": 9.497327699880018e-06, "loss": 0.5735, "step": 4047 }, { "epoch": 0.1696865535562705, "grad_norm": 1.7571041584014893, "learning_rate": 9.497031005613911e-06, "loss": 0.5238, "step": 4048 }, { "epoch": 0.16972847217127945, "grad_norm": 1.9552711248397827, "learning_rate": 9.496734228451257e-06, "loss": 0.5972, "step": 4049 }, { "epoch": 0.16977039078628842, "grad_norm": 1.682984471321106, "learning_rate": 9.496437368397525e-06, "loss": 0.5404, "step": 4050 }, { "epoch": 0.16981230940129738, "grad_norm": 1.520551323890686, "learning_rate": 9.496140425458185e-06, "loss": 0.5914, "step": 4051 }, { "epoch": 0.16985422801630634, "grad_norm": 1.8822237253189087, "learning_rate": 9.49584339963871e-06, "loss": 0.5548, "step": 4052 }, { "epoch": 0.1698961466313153, "grad_norm": 1.5970104932785034, "learning_rate": 9.495546290944582e-06, "loss": 0.4956, "step": 4053 }, { "epoch": 0.16993806524632427, "grad_norm": 1.8802253007888794, "learning_rate": 9.49524909938127e-06, "loss": 0.5872, "step": 4054 }, { "epoch": 0.16997998386133323, "grad_norm": 1.6950721740722656, "learning_rate": 9.494951824954255e-06, "loss": 0.536, "step": 4055 }, { "epoch": 0.1700219024763422, "grad_norm": 1.6041009426116943, "learning_rate": 9.494654467669018e-06, "loss": 0.5319, "step": 4056 }, { "epoch": 0.17006382109135115, "grad_norm": 1.790773868560791, "learning_rate": 9.49435702753104e-06, "loss": 0.5398, "step": 4057 }, { "epoch": 0.17010573970636012, "grad_norm": 1.7573174238204956, "learning_rate": 9.494059504545803e-06, "loss": 0.5955, "step": 4058 }, { "epoch": 0.17014765832136905, "grad_norm": 2.499586820602417, "learning_rate": 9.493761898718791e-06, "loss": 0.5507, "step": 4059 }, { "epoch": 0.170189576936378, "grad_norm": 1.7022675275802612, "learning_rate": 9.493464210055492e-06, "loss": 0.5389, "step": 4060 }, { "epoch": 0.17023149555138697, "grad_norm": 1.9612749814987183, "learning_rate": 9.493166438561393e-06, "loss": 0.5857, "step": 4061 }, { "epoch": 0.17027341416639594, "grad_norm": 1.7403377294540405, "learning_rate": 9.492868584241981e-06, "loss": 0.5461, "step": 4062 }, { "epoch": 0.1703153327814049, "grad_norm": 1.7414840459823608, "learning_rate": 9.492570647102747e-06, "loss": 0.5324, "step": 4063 }, { "epoch": 0.17035725139641386, "grad_norm": 1.8038647174835205, "learning_rate": 9.492272627149187e-06, "loss": 0.5882, "step": 4064 }, { "epoch": 0.17039917001142282, "grad_norm": 1.7793632745742798, "learning_rate": 9.491974524386789e-06, "loss": 0.5849, "step": 4065 }, { "epoch": 0.17044108862643179, "grad_norm": 1.5655412673950195, "learning_rate": 9.49167633882105e-06, "loss": 0.5879, "step": 4066 }, { "epoch": 0.17048300724144075, "grad_norm": 1.7584199905395508, "learning_rate": 9.49137807045747e-06, "loss": 0.5621, "step": 4067 }, { "epoch": 0.1705249258564497, "grad_norm": 1.6864396333694458, "learning_rate": 9.491079719301541e-06, "loss": 0.6051, "step": 4068 }, { "epoch": 0.17056684447145867, "grad_norm": 1.84431791305542, "learning_rate": 9.490781285358768e-06, "loss": 0.6381, "step": 4069 }, { "epoch": 0.17060876308646764, "grad_norm": 1.5301125049591064, "learning_rate": 9.490482768634651e-06, "loss": 0.5304, "step": 4070 }, { "epoch": 0.17065068170147657, "grad_norm": 1.5924345254898071, "learning_rate": 9.49018416913469e-06, "loss": 0.5089, "step": 4071 }, { "epoch": 0.17069260031648553, "grad_norm": 1.8778263330459595, "learning_rate": 9.489885486864393e-06, "loss": 0.5621, "step": 4072 }, { "epoch": 0.1707345189314945, "grad_norm": 1.8505027294158936, "learning_rate": 9.489586721829265e-06, "loss": 0.5952, "step": 4073 }, { "epoch": 0.17077643754650346, "grad_norm": 2.462984085083008, "learning_rate": 9.489287874034811e-06, "loss": 0.531, "step": 4074 }, { "epoch": 0.17081835616151242, "grad_norm": 1.7490394115447998, "learning_rate": 9.488988943486541e-06, "loss": 0.5424, "step": 4075 }, { "epoch": 0.17086027477652138, "grad_norm": 1.7581342458724976, "learning_rate": 9.488689930189967e-06, "loss": 0.6385, "step": 4076 }, { "epoch": 0.17090219339153034, "grad_norm": 2.016723394393921, "learning_rate": 9.488390834150597e-06, "loss": 0.5385, "step": 4077 }, { "epoch": 0.1709441120065393, "grad_norm": 1.5532770156860352, "learning_rate": 9.48809165537395e-06, "loss": 0.5171, "step": 4078 }, { "epoch": 0.17098603062154827, "grad_norm": 1.5339629650115967, "learning_rate": 9.487792393865539e-06, "loss": 0.5263, "step": 4079 }, { "epoch": 0.17102794923655723, "grad_norm": 1.7268513441085815, "learning_rate": 9.487493049630875e-06, "loss": 0.5604, "step": 4080 }, { "epoch": 0.1710698678515662, "grad_norm": 1.8400930166244507, "learning_rate": 9.487193622675484e-06, "loss": 0.6705, "step": 4081 }, { "epoch": 0.17111178646657516, "grad_norm": 1.8059179782867432, "learning_rate": 9.48689411300488e-06, "loss": 0.552, "step": 4082 }, { "epoch": 0.17115370508158412, "grad_norm": 4.623781681060791, "learning_rate": 9.486594520624586e-06, "loss": 0.5744, "step": 4083 }, { "epoch": 0.17119562369659305, "grad_norm": 1.9598064422607422, "learning_rate": 9.486294845540126e-06, "loss": 0.5669, "step": 4084 }, { "epoch": 0.17123754231160201, "grad_norm": 3.112347364425659, "learning_rate": 9.485995087757021e-06, "loss": 0.56, "step": 4085 }, { "epoch": 0.17127946092661098, "grad_norm": 1.6720551252365112, "learning_rate": 9.485695247280799e-06, "loss": 0.583, "step": 4086 }, { "epoch": 0.17132137954161994, "grad_norm": 1.527002215385437, "learning_rate": 9.485395324116987e-06, "loss": 0.5283, "step": 4087 }, { "epoch": 0.1713632981566289, "grad_norm": 1.6862847805023193, "learning_rate": 9.485095318271112e-06, "loss": 0.537, "step": 4088 }, { "epoch": 0.17140521677163786, "grad_norm": 1.5471360683441162, "learning_rate": 9.484795229748704e-06, "loss": 0.5062, "step": 4089 }, { "epoch": 0.17144713538664683, "grad_norm": 1.7645634412765503, "learning_rate": 9.484495058555298e-06, "loss": 0.5156, "step": 4090 }, { "epoch": 0.1714890540016558, "grad_norm": 1.5669260025024414, "learning_rate": 9.484194804696425e-06, "loss": 0.5301, "step": 4091 }, { "epoch": 0.17153097261666475, "grad_norm": 1.84481942653656, "learning_rate": 9.48389446817762e-06, "loss": 0.5816, "step": 4092 }, { "epoch": 0.1715728912316737, "grad_norm": 1.6635798215866089, "learning_rate": 9.483594049004419e-06, "loss": 0.5083, "step": 4093 }, { "epoch": 0.17161480984668268, "grad_norm": 1.9589687585830688, "learning_rate": 9.48329354718236e-06, "loss": 0.5736, "step": 4094 }, { "epoch": 0.17165672846169164, "grad_norm": 1.7596700191497803, "learning_rate": 9.482992962716982e-06, "loss": 0.6067, "step": 4095 }, { "epoch": 0.17169864707670057, "grad_norm": 1.6278280019760132, "learning_rate": 9.482692295613827e-06, "loss": 0.5371, "step": 4096 }, { "epoch": 0.17174056569170953, "grad_norm": 2.1768317222595215, "learning_rate": 9.482391545878438e-06, "loss": 0.5287, "step": 4097 }, { "epoch": 0.1717824843067185, "grad_norm": 1.7690577507019043, "learning_rate": 9.482090713516355e-06, "loss": 0.5671, "step": 4098 }, { "epoch": 0.17182440292172746, "grad_norm": 2.087690830230713, "learning_rate": 9.481789798533128e-06, "loss": 0.5369, "step": 4099 }, { "epoch": 0.17186632153673642, "grad_norm": 2.0082812309265137, "learning_rate": 9.4814888009343e-06, "loss": 0.6046, "step": 4100 }, { "epoch": 0.17190824015174538, "grad_norm": 1.7433838844299316, "learning_rate": 9.481187720725423e-06, "loss": 0.5217, "step": 4101 }, { "epoch": 0.17195015876675435, "grad_norm": 1.923892855644226, "learning_rate": 9.480886557912047e-06, "loss": 0.5473, "step": 4102 }, { "epoch": 0.1719920773817633, "grad_norm": 1.6568683385849, "learning_rate": 9.480585312499717e-06, "loss": 0.592, "step": 4103 }, { "epoch": 0.17203399599677227, "grad_norm": 1.775061011314392, "learning_rate": 9.480283984493994e-06, "loss": 0.5447, "step": 4104 }, { "epoch": 0.17207591461178123, "grad_norm": 1.788333773612976, "learning_rate": 9.47998257390043e-06, "loss": 0.5579, "step": 4105 }, { "epoch": 0.1721178332267902, "grad_norm": 1.7010146379470825, "learning_rate": 9.47968108072458e-06, "loss": 0.5335, "step": 4106 }, { "epoch": 0.17215975184179916, "grad_norm": 1.649003267288208, "learning_rate": 9.479379504972002e-06, "loss": 0.5423, "step": 4107 }, { "epoch": 0.17220167045680812, "grad_norm": 1.4320815801620483, "learning_rate": 9.479077846648256e-06, "loss": 0.4907, "step": 4108 }, { "epoch": 0.17224358907181705, "grad_norm": 1.4647090435028076, "learning_rate": 9.478776105758902e-06, "loss": 0.5221, "step": 4109 }, { "epoch": 0.17228550768682602, "grad_norm": 1.632834553718567, "learning_rate": 9.478474282309503e-06, "loss": 0.5264, "step": 4110 }, { "epoch": 0.17232742630183498, "grad_norm": 1.9135249853134155, "learning_rate": 9.47817237630562e-06, "loss": 0.5128, "step": 4111 }, { "epoch": 0.17236934491684394, "grad_norm": 1.5895147323608398, "learning_rate": 9.47787038775282e-06, "loss": 0.5472, "step": 4112 }, { "epoch": 0.1724112635318529, "grad_norm": 2.5632848739624023, "learning_rate": 9.477568316656672e-06, "loss": 0.556, "step": 4113 }, { "epoch": 0.17245318214686187, "grad_norm": 1.8056782484054565, "learning_rate": 9.47726616302274e-06, "loss": 0.5837, "step": 4114 }, { "epoch": 0.17249510076187083, "grad_norm": 1.656386137008667, "learning_rate": 9.476963926856597e-06, "loss": 0.53, "step": 4115 }, { "epoch": 0.1725370193768798, "grad_norm": 1.888274073600769, "learning_rate": 9.476661608163814e-06, "loss": 0.5881, "step": 4116 }, { "epoch": 0.17257893799188875, "grad_norm": 1.8142822980880737, "learning_rate": 9.476359206949962e-06, "loss": 0.5884, "step": 4117 }, { "epoch": 0.17262085660689772, "grad_norm": 1.5722057819366455, "learning_rate": 9.476056723220617e-06, "loss": 0.5529, "step": 4118 }, { "epoch": 0.17266277522190668, "grad_norm": 2.309634208679199, "learning_rate": 9.475754156981354e-06, "loss": 0.5419, "step": 4119 }, { "epoch": 0.17270469383691564, "grad_norm": 1.6522489786148071, "learning_rate": 9.47545150823775e-06, "loss": 0.5063, "step": 4120 }, { "epoch": 0.17274661245192457, "grad_norm": 1.9367735385894775, "learning_rate": 9.475148776995384e-06, "loss": 0.5117, "step": 4121 }, { "epoch": 0.17278853106693354, "grad_norm": 1.5557644367218018, "learning_rate": 9.47484596325984e-06, "loss": 0.5728, "step": 4122 }, { "epoch": 0.1728304496819425, "grad_norm": 1.8167442083358765, "learning_rate": 9.474543067036694e-06, "loss": 0.5449, "step": 4123 }, { "epoch": 0.17287236829695146, "grad_norm": 1.85882568359375, "learning_rate": 9.474240088331534e-06, "loss": 0.5381, "step": 4124 }, { "epoch": 0.17291428691196042, "grad_norm": 1.5812429189682007, "learning_rate": 9.473937027149942e-06, "loss": 0.5352, "step": 4125 }, { "epoch": 0.1729562055269694, "grad_norm": 1.9077056646347046, "learning_rate": 9.473633883497508e-06, "loss": 0.5904, "step": 4126 }, { "epoch": 0.17299812414197835, "grad_norm": 1.771308183670044, "learning_rate": 9.473330657379816e-06, "loss": 0.5716, "step": 4127 }, { "epoch": 0.1730400427569873, "grad_norm": 1.8803128004074097, "learning_rate": 9.473027348802458e-06, "loss": 0.5993, "step": 4128 }, { "epoch": 0.17308196137199627, "grad_norm": 1.6538728475570679, "learning_rate": 9.472723957771026e-06, "loss": 0.5407, "step": 4129 }, { "epoch": 0.17312387998700524, "grad_norm": 1.6664279699325562, "learning_rate": 9.472420484291109e-06, "loss": 0.497, "step": 4130 }, { "epoch": 0.1731657986020142, "grad_norm": 1.7729988098144531, "learning_rate": 9.472116928368304e-06, "loss": 0.5968, "step": 4131 }, { "epoch": 0.17320771721702316, "grad_norm": 1.5611631870269775, "learning_rate": 9.471813290008208e-06, "loss": 0.5205, "step": 4132 }, { "epoch": 0.17324963583203212, "grad_norm": 1.5190483331680298, "learning_rate": 9.471509569216414e-06, "loss": 0.5371, "step": 4133 }, { "epoch": 0.17329155444704106, "grad_norm": 1.5222702026367188, "learning_rate": 9.471205765998524e-06, "loss": 0.5366, "step": 4134 }, { "epoch": 0.17333347306205002, "grad_norm": 1.5341463088989258, "learning_rate": 9.470901880360135e-06, "loss": 0.532, "step": 4135 }, { "epoch": 0.17337539167705898, "grad_norm": 1.5616016387939453, "learning_rate": 9.470597912306852e-06, "loss": 0.5391, "step": 4136 }, { "epoch": 0.17341731029206794, "grad_norm": 1.6101582050323486, "learning_rate": 9.470293861844278e-06, "loss": 0.6243, "step": 4137 }, { "epoch": 0.1734592289070769, "grad_norm": 1.8374935388565063, "learning_rate": 9.469989728978016e-06, "loss": 0.5469, "step": 4138 }, { "epoch": 0.17350114752208587, "grad_norm": 1.797729730606079, "learning_rate": 9.46968551371367e-06, "loss": 0.541, "step": 4139 }, { "epoch": 0.17354306613709483, "grad_norm": 1.6734188795089722, "learning_rate": 9.469381216056854e-06, "loss": 0.5659, "step": 4140 }, { "epoch": 0.1735849847521038, "grad_norm": 1.608069896697998, "learning_rate": 9.469076836013174e-06, "loss": 0.5414, "step": 4141 }, { "epoch": 0.17362690336711276, "grad_norm": 1.631417155265808, "learning_rate": 9.468772373588238e-06, "loss": 0.5513, "step": 4142 }, { "epoch": 0.17366882198212172, "grad_norm": 1.6262640953063965, "learning_rate": 9.468467828787663e-06, "loss": 0.5625, "step": 4143 }, { "epoch": 0.17371074059713068, "grad_norm": 1.5137449502944946, "learning_rate": 9.468163201617063e-06, "loss": 0.5359, "step": 4144 }, { "epoch": 0.17375265921213964, "grad_norm": 1.6043031215667725, "learning_rate": 9.467858492082049e-06, "loss": 0.5702, "step": 4145 }, { "epoch": 0.17379457782714858, "grad_norm": 1.5260822772979736, "learning_rate": 9.467553700188241e-06, "loss": 0.5701, "step": 4146 }, { "epoch": 0.17383649644215754, "grad_norm": 1.665700912475586, "learning_rate": 9.467248825941257e-06, "loss": 0.5922, "step": 4147 }, { "epoch": 0.1738784150571665, "grad_norm": 1.737730860710144, "learning_rate": 9.466943869346716e-06, "loss": 0.5763, "step": 4148 }, { "epoch": 0.17392033367217546, "grad_norm": 1.7674342393875122, "learning_rate": 9.46663883041024e-06, "loss": 0.558, "step": 4149 }, { "epoch": 0.17396225228718443, "grad_norm": 1.593715786933899, "learning_rate": 9.466333709137452e-06, "loss": 0.5828, "step": 4150 }, { "epoch": 0.1740041709021934, "grad_norm": 1.5613399744033813, "learning_rate": 9.466028505533978e-06, "loss": 0.5253, "step": 4151 }, { "epoch": 0.17404608951720235, "grad_norm": 1.8493508100509644, "learning_rate": 9.46572321960544e-06, "loss": 0.5999, "step": 4152 }, { "epoch": 0.1740880081322113, "grad_norm": 2.012580633163452, "learning_rate": 9.465417851357471e-06, "loss": 0.5282, "step": 4153 }, { "epoch": 0.17412992674722028, "grad_norm": 1.686970591545105, "learning_rate": 9.465112400795695e-06, "loss": 0.5797, "step": 4154 }, { "epoch": 0.17417184536222924, "grad_norm": 1.6345033645629883, "learning_rate": 9.464806867925746e-06, "loss": 0.5212, "step": 4155 }, { "epoch": 0.1742137639772382, "grad_norm": 1.9941903352737427, "learning_rate": 9.464501252753256e-06, "loss": 0.5465, "step": 4156 }, { "epoch": 0.17425568259224716, "grad_norm": 1.6041713953018188, "learning_rate": 9.464195555283854e-06, "loss": 0.5636, "step": 4157 }, { "epoch": 0.17429760120725613, "grad_norm": 1.6370561122894287, "learning_rate": 9.463889775523183e-06, "loss": 0.608, "step": 4158 }, { "epoch": 0.17433951982226506, "grad_norm": 1.9433943033218384, "learning_rate": 9.463583913476872e-06, "loss": 0.5364, "step": 4159 }, { "epoch": 0.17438143843727402, "grad_norm": 1.6569839715957642, "learning_rate": 9.463277969150563e-06, "loss": 0.5692, "step": 4160 }, { "epoch": 0.17442335705228298, "grad_norm": 1.601498007774353, "learning_rate": 9.462971942549895e-06, "loss": 0.538, "step": 4161 }, { "epoch": 0.17446527566729195, "grad_norm": 1.7682095766067505, "learning_rate": 9.462665833680509e-06, "loss": 0.5979, "step": 4162 }, { "epoch": 0.1745071942823009, "grad_norm": 1.7757182121276855, "learning_rate": 9.462359642548047e-06, "loss": 0.5259, "step": 4163 }, { "epoch": 0.17454911289730987, "grad_norm": 2.5125269889831543, "learning_rate": 9.462053369158155e-06, "loss": 0.5716, "step": 4164 }, { "epoch": 0.17459103151231883, "grad_norm": 1.5011125802993774, "learning_rate": 9.461747013516477e-06, "loss": 0.5329, "step": 4165 }, { "epoch": 0.1746329501273278, "grad_norm": 1.6033180952072144, "learning_rate": 9.461440575628663e-06, "loss": 0.5624, "step": 4166 }, { "epoch": 0.17467486874233676, "grad_norm": 1.7671473026275635, "learning_rate": 9.461134055500358e-06, "loss": 0.5531, "step": 4167 }, { "epoch": 0.17471678735734572, "grad_norm": 2.2059812545776367, "learning_rate": 9.460827453137214e-06, "loss": 0.6405, "step": 4168 }, { "epoch": 0.17475870597235468, "grad_norm": 1.6523767709732056, "learning_rate": 9.460520768544882e-06, "loss": 0.5524, "step": 4169 }, { "epoch": 0.17480062458736365, "grad_norm": 1.6234432458877563, "learning_rate": 9.460214001729017e-06, "loss": 0.5408, "step": 4170 }, { "epoch": 0.17484254320237258, "grad_norm": 1.6969823837280273, "learning_rate": 9.459907152695273e-06, "loss": 0.5989, "step": 4171 }, { "epoch": 0.17488446181738154, "grad_norm": 1.8531696796417236, "learning_rate": 9.459600221449309e-06, "loss": 0.574, "step": 4172 }, { "epoch": 0.1749263804323905, "grad_norm": 2.3090872764587402, "learning_rate": 9.459293207996775e-06, "loss": 0.5819, "step": 4173 }, { "epoch": 0.17496829904739947, "grad_norm": 1.946921467781067, "learning_rate": 9.458986112343337e-06, "loss": 0.5171, "step": 4174 }, { "epoch": 0.17501021766240843, "grad_norm": 1.5394657850265503, "learning_rate": 9.458678934494657e-06, "loss": 0.5512, "step": 4175 }, { "epoch": 0.1750521362774174, "grad_norm": 1.8463269472122192, "learning_rate": 9.458371674456393e-06, "loss": 0.5696, "step": 4176 }, { "epoch": 0.17509405489242635, "grad_norm": 2.595348834991455, "learning_rate": 9.458064332234211e-06, "loss": 0.5672, "step": 4177 }, { "epoch": 0.17513597350743532, "grad_norm": 2.003321409225464, "learning_rate": 9.457756907833776e-06, "loss": 0.55, "step": 4178 }, { "epoch": 0.17517789212244428, "grad_norm": 2.0433945655822754, "learning_rate": 9.457449401260755e-06, "loss": 0.5749, "step": 4179 }, { "epoch": 0.17521981073745324, "grad_norm": 1.77859365940094, "learning_rate": 9.457141812520817e-06, "loss": 0.5875, "step": 4180 }, { "epoch": 0.1752617293524622, "grad_norm": 2.0772645473480225, "learning_rate": 9.456834141619632e-06, "loss": 0.5825, "step": 4181 }, { "epoch": 0.17530364796747117, "grad_norm": 1.8918888568878174, "learning_rate": 9.456526388562869e-06, "loss": 0.6415, "step": 4182 }, { "epoch": 0.17534556658248013, "grad_norm": 1.6400415897369385, "learning_rate": 9.456218553356205e-06, "loss": 0.5543, "step": 4183 }, { "epoch": 0.17538748519748906, "grad_norm": 1.6922918558120728, "learning_rate": 9.45591063600531e-06, "loss": 0.5255, "step": 4184 }, { "epoch": 0.17542940381249802, "grad_norm": 1.9459041357040405, "learning_rate": 9.455602636515865e-06, "loss": 0.5821, "step": 4185 }, { "epoch": 0.175471322427507, "grad_norm": 1.5934662818908691, "learning_rate": 9.455294554893543e-06, "loss": 0.5405, "step": 4186 }, { "epoch": 0.17551324104251595, "grad_norm": 1.76579749584198, "learning_rate": 9.454986391144028e-06, "loss": 0.5967, "step": 4187 }, { "epoch": 0.1755551596575249, "grad_norm": 1.750468373298645, "learning_rate": 9.454678145272995e-06, "loss": 0.5944, "step": 4188 }, { "epoch": 0.17559707827253387, "grad_norm": 1.8608440160751343, "learning_rate": 9.454369817286132e-06, "loss": 0.5631, "step": 4189 }, { "epoch": 0.17563899688754284, "grad_norm": 1.7094907760620117, "learning_rate": 9.454061407189114e-06, "loss": 0.536, "step": 4190 }, { "epoch": 0.1756809155025518, "grad_norm": 1.8183364868164062, "learning_rate": 9.453752914987635e-06, "loss": 0.592, "step": 4191 }, { "epoch": 0.17572283411756076, "grad_norm": 1.5141836404800415, "learning_rate": 9.453444340687379e-06, "loss": 0.532, "step": 4192 }, { "epoch": 0.17576475273256972, "grad_norm": 1.6259713172912598, "learning_rate": 9.453135684294032e-06, "loss": 0.6448, "step": 4193 }, { "epoch": 0.17580667134757869, "grad_norm": 1.5314102172851562, "learning_rate": 9.452826945813286e-06, "loss": 0.5638, "step": 4194 }, { "epoch": 0.17584858996258765, "grad_norm": 1.9380269050598145, "learning_rate": 9.452518125250829e-06, "loss": 0.5587, "step": 4195 }, { "epoch": 0.17589050857759658, "grad_norm": 1.729406714439392, "learning_rate": 9.452209222612358e-06, "loss": 0.563, "step": 4196 }, { "epoch": 0.17593242719260554, "grad_norm": 1.8287348747253418, "learning_rate": 9.451900237903564e-06, "loss": 0.5781, "step": 4197 }, { "epoch": 0.1759743458076145, "grad_norm": 1.7479116916656494, "learning_rate": 9.451591171130144e-06, "loss": 0.5686, "step": 4198 }, { "epoch": 0.17601626442262347, "grad_norm": 1.584357738494873, "learning_rate": 9.451282022297795e-06, "loss": 0.5952, "step": 4199 }, { "epoch": 0.17605818303763243, "grad_norm": 1.703615427017212, "learning_rate": 9.450972791412214e-06, "loss": 0.5976, "step": 4200 }, { "epoch": 0.1761001016526414, "grad_norm": 2.1241488456726074, "learning_rate": 9.450663478479105e-06, "loss": 0.583, "step": 4201 }, { "epoch": 0.17614202026765036, "grad_norm": 1.593394160270691, "learning_rate": 9.450354083504165e-06, "loss": 0.5921, "step": 4202 }, { "epoch": 0.17618393888265932, "grad_norm": 1.644365668296814, "learning_rate": 9.450044606493103e-06, "loss": 0.5676, "step": 4203 }, { "epoch": 0.17622585749766828, "grad_norm": 1.8174432516098022, "learning_rate": 9.449735047451619e-06, "loss": 0.4741, "step": 4204 }, { "epoch": 0.17626777611267724, "grad_norm": 5.800610065460205, "learning_rate": 9.44942540638542e-06, "loss": 0.5436, "step": 4205 }, { "epoch": 0.1763096947276862, "grad_norm": 3.1635849475860596, "learning_rate": 9.449115683300217e-06, "loss": 0.5951, "step": 4206 }, { "epoch": 0.17635161334269517, "grad_norm": 1.7380269765853882, "learning_rate": 9.448805878201717e-06, "loss": 0.5743, "step": 4207 }, { "epoch": 0.17639353195770413, "grad_norm": 1.7905665636062622, "learning_rate": 9.448495991095629e-06, "loss": 0.5646, "step": 4208 }, { "epoch": 0.17643545057271307, "grad_norm": 1.547707438468933, "learning_rate": 9.448186021987668e-06, "loss": 0.5093, "step": 4209 }, { "epoch": 0.17647736918772203, "grad_norm": 1.6973565816879272, "learning_rate": 9.447875970883548e-06, "loss": 0.5174, "step": 4210 }, { "epoch": 0.176519287802731, "grad_norm": 1.994999647140503, "learning_rate": 9.447565837788984e-06, "loss": 0.5657, "step": 4211 }, { "epoch": 0.17656120641773995, "grad_norm": 1.7339166402816772, "learning_rate": 9.44725562270969e-06, "loss": 0.5314, "step": 4212 }, { "epoch": 0.17660312503274891, "grad_norm": 1.8084725141525269, "learning_rate": 9.446945325651389e-06, "loss": 0.5458, "step": 4213 }, { "epoch": 0.17664504364775788, "grad_norm": 1.8694586753845215, "learning_rate": 9.446634946619797e-06, "loss": 0.5945, "step": 4214 }, { "epoch": 0.17668696226276684, "grad_norm": 1.6437740325927734, "learning_rate": 9.446324485620637e-06, "loss": 0.6303, "step": 4215 }, { "epoch": 0.1767288808777758, "grad_norm": 2.2290055751800537, "learning_rate": 9.446013942659633e-06, "loss": 0.6244, "step": 4216 }, { "epoch": 0.17677079949278476, "grad_norm": 2.926271677017212, "learning_rate": 9.445703317742507e-06, "loss": 0.5719, "step": 4217 }, { "epoch": 0.17681271810779373, "grad_norm": 1.7139618396759033, "learning_rate": 9.445392610874989e-06, "loss": 0.4951, "step": 4218 }, { "epoch": 0.1768546367228027, "grad_norm": 1.938308835029602, "learning_rate": 9.445081822062802e-06, "loss": 0.5996, "step": 4219 }, { "epoch": 0.17689655533781165, "grad_norm": 1.6071958541870117, "learning_rate": 9.444770951311678e-06, "loss": 0.5988, "step": 4220 }, { "epoch": 0.17693847395282059, "grad_norm": 1.730018973350525, "learning_rate": 9.444459998627344e-06, "loss": 0.5649, "step": 4221 }, { "epoch": 0.17698039256782955, "grad_norm": 1.7988669872283936, "learning_rate": 9.444148964015536e-06, "loss": 0.5555, "step": 4222 }, { "epoch": 0.1770223111828385, "grad_norm": 1.8410649299621582, "learning_rate": 9.443837847481985e-06, "loss": 0.5082, "step": 4223 }, { "epoch": 0.17706422979784747, "grad_norm": 1.59534752368927, "learning_rate": 9.443526649032426e-06, "loss": 0.4967, "step": 4224 }, { "epoch": 0.17710614841285643, "grad_norm": 1.4397656917572021, "learning_rate": 9.443215368672597e-06, "loss": 0.5085, "step": 4225 }, { "epoch": 0.1771480670278654, "grad_norm": 1.8158529996871948, "learning_rate": 9.442904006408236e-06, "loss": 0.5665, "step": 4226 }, { "epoch": 0.17718998564287436, "grad_norm": 1.7272917032241821, "learning_rate": 9.442592562245081e-06, "loss": 0.5589, "step": 4227 }, { "epoch": 0.17723190425788332, "grad_norm": 1.4719176292419434, "learning_rate": 9.442281036188874e-06, "loss": 0.5255, "step": 4228 }, { "epoch": 0.17727382287289228, "grad_norm": 2.358241081237793, "learning_rate": 9.441969428245356e-06, "loss": 0.5078, "step": 4229 }, { "epoch": 0.17731574148790125, "grad_norm": 1.7024309635162354, "learning_rate": 9.441657738420273e-06, "loss": 0.6564, "step": 4230 }, { "epoch": 0.1773576601029102, "grad_norm": 1.7256348133087158, "learning_rate": 9.44134596671937e-06, "loss": 0.5923, "step": 4231 }, { "epoch": 0.17739957871791917, "grad_norm": 1.5577399730682373, "learning_rate": 9.441034113148397e-06, "loss": 0.5546, "step": 4232 }, { "epoch": 0.17744149733292813, "grad_norm": 1.6019394397735596, "learning_rate": 9.440722177713097e-06, "loss": 0.5751, "step": 4233 }, { "epoch": 0.17748341594793707, "grad_norm": 1.6956826448440552, "learning_rate": 9.440410160419222e-06, "loss": 0.6096, "step": 4234 }, { "epoch": 0.17752533456294603, "grad_norm": 1.7056381702423096, "learning_rate": 9.440098061272526e-06, "loss": 0.5643, "step": 4235 }, { "epoch": 0.177567253177955, "grad_norm": 1.512692928314209, "learning_rate": 9.43978588027876e-06, "loss": 0.5247, "step": 4236 }, { "epoch": 0.17760917179296395, "grad_norm": 1.697864055633545, "learning_rate": 9.43947361744368e-06, "loss": 0.5622, "step": 4237 }, { "epoch": 0.17765109040797292, "grad_norm": 1.5708444118499756, "learning_rate": 9.43916127277304e-06, "loss": 0.5101, "step": 4238 }, { "epoch": 0.17769300902298188, "grad_norm": 1.746114730834961, "learning_rate": 9.4388488462726e-06, "loss": 0.5873, "step": 4239 }, { "epoch": 0.17773492763799084, "grad_norm": 1.6060422658920288, "learning_rate": 9.438536337948117e-06, "loss": 0.5581, "step": 4240 }, { "epoch": 0.1777768462529998, "grad_norm": 1.7439624071121216, "learning_rate": 9.438223747805353e-06, "loss": 0.5046, "step": 4241 }, { "epoch": 0.17781876486800877, "grad_norm": 1.6305925846099854, "learning_rate": 9.437911075850069e-06, "loss": 0.5581, "step": 4242 }, { "epoch": 0.17786068348301773, "grad_norm": 1.7561964988708496, "learning_rate": 9.437598322088031e-06, "loss": 0.5549, "step": 4243 }, { "epoch": 0.1779026020980267, "grad_norm": 1.917759656906128, "learning_rate": 9.437285486525002e-06, "loss": 0.5782, "step": 4244 }, { "epoch": 0.17794452071303565, "grad_norm": 1.6448519229888916, "learning_rate": 9.43697256916675e-06, "loss": 0.6176, "step": 4245 }, { "epoch": 0.1779864393280446, "grad_norm": 1.5401910543441772, "learning_rate": 9.436659570019042e-06, "loss": 0.5266, "step": 4246 }, { "epoch": 0.17802835794305355, "grad_norm": 1.4996917247772217, "learning_rate": 9.43634648908765e-06, "loss": 0.5473, "step": 4247 }, { "epoch": 0.1780702765580625, "grad_norm": 2.854612350463867, "learning_rate": 9.43603332637834e-06, "loss": 0.5615, "step": 4248 }, { "epoch": 0.17811219517307147, "grad_norm": 1.9661184549331665, "learning_rate": 9.435720081896891e-06, "loss": 0.5365, "step": 4249 }, { "epoch": 0.17815411378808044, "grad_norm": 1.7493510246276855, "learning_rate": 9.435406755649072e-06, "loss": 0.6124, "step": 4250 }, { "epoch": 0.1781960324030894, "grad_norm": 1.5385462045669556, "learning_rate": 9.435093347640665e-06, "loss": 0.5704, "step": 4251 }, { "epoch": 0.17823795101809836, "grad_norm": 1.774080753326416, "learning_rate": 9.434779857877441e-06, "loss": 0.5339, "step": 4252 }, { "epoch": 0.17827986963310732, "grad_norm": 1.8239655494689941, "learning_rate": 9.434466286365181e-06, "loss": 0.5473, "step": 4253 }, { "epoch": 0.1783217882481163, "grad_norm": 1.6719508171081543, "learning_rate": 9.434152633109666e-06, "loss": 0.5545, "step": 4254 }, { "epoch": 0.17836370686312525, "grad_norm": 1.633309006690979, "learning_rate": 9.433838898116678e-06, "loss": 0.5983, "step": 4255 }, { "epoch": 0.1784056254781342, "grad_norm": 1.6322835683822632, "learning_rate": 9.433525081392e-06, "loss": 0.5664, "step": 4256 }, { "epoch": 0.17844754409314317, "grad_norm": 1.7203844785690308, "learning_rate": 9.433211182941413e-06, "loss": 0.5809, "step": 4257 }, { "epoch": 0.17848946270815214, "grad_norm": 1.6702135801315308, "learning_rate": 9.43289720277071e-06, "loss": 0.5753, "step": 4258 }, { "epoch": 0.17853138132316107, "grad_norm": 1.58368718624115, "learning_rate": 9.432583140885674e-06, "loss": 0.6071, "step": 4259 }, { "epoch": 0.17857329993817003, "grad_norm": 2.4710426330566406, "learning_rate": 9.432268997292096e-06, "loss": 0.5299, "step": 4260 }, { "epoch": 0.178615218553179, "grad_norm": 2.520033597946167, "learning_rate": 9.431954771995767e-06, "loss": 0.6255, "step": 4261 }, { "epoch": 0.17865713716818796, "grad_norm": 1.7064237594604492, "learning_rate": 9.431640465002478e-06, "loss": 0.5582, "step": 4262 }, { "epoch": 0.17869905578319692, "grad_norm": 1.7157803773880005, "learning_rate": 9.431326076318024e-06, "loss": 0.5599, "step": 4263 }, { "epoch": 0.17874097439820588, "grad_norm": 1.7533254623413086, "learning_rate": 9.431011605948199e-06, "loss": 0.5688, "step": 4264 }, { "epoch": 0.17878289301321484, "grad_norm": 1.7583317756652832, "learning_rate": 9.430697053898801e-06, "loss": 0.572, "step": 4265 }, { "epoch": 0.1788248116282238, "grad_norm": 1.6284854412078857, "learning_rate": 9.430382420175628e-06, "loss": 0.546, "step": 4266 }, { "epoch": 0.17886673024323277, "grad_norm": 1.572336196899414, "learning_rate": 9.43006770478448e-06, "loss": 0.5279, "step": 4267 }, { "epoch": 0.17890864885824173, "grad_norm": 1.538556456565857, "learning_rate": 9.429752907731158e-06, "loss": 0.5462, "step": 4268 }, { "epoch": 0.1789505674732507, "grad_norm": 1.6223260164260864, "learning_rate": 9.429438029021467e-06, "loss": 0.567, "step": 4269 }, { "epoch": 0.17899248608825966, "grad_norm": 1.5634013414382935, "learning_rate": 9.429123068661208e-06, "loss": 0.5157, "step": 4270 }, { "epoch": 0.1790344047032686, "grad_norm": 2.2047204971313477, "learning_rate": 9.428808026656188e-06, "loss": 0.5697, "step": 4271 }, { "epoch": 0.17907632331827755, "grad_norm": 1.6505508422851562, "learning_rate": 9.428492903012216e-06, "loss": 0.5949, "step": 4272 }, { "epoch": 0.17911824193328651, "grad_norm": 1.9732253551483154, "learning_rate": 9.428177697735097e-06, "loss": 0.5868, "step": 4273 }, { "epoch": 0.17916016054829548, "grad_norm": 1.7827589511871338, "learning_rate": 9.427862410830645e-06, "loss": 0.5574, "step": 4274 }, { "epoch": 0.17920207916330444, "grad_norm": 1.8414908647537231, "learning_rate": 9.427547042304672e-06, "loss": 0.533, "step": 4275 }, { "epoch": 0.1792439977783134, "grad_norm": 1.961470603942871, "learning_rate": 9.427231592162992e-06, "loss": 0.5855, "step": 4276 }, { "epoch": 0.17928591639332236, "grad_norm": 1.633490800857544, "learning_rate": 9.426916060411413e-06, "loss": 0.6028, "step": 4277 }, { "epoch": 0.17932783500833133, "grad_norm": 1.5936483144760132, "learning_rate": 9.42660044705576e-06, "loss": 0.5738, "step": 4278 }, { "epoch": 0.1793697536233403, "grad_norm": 1.6075938940048218, "learning_rate": 9.426284752101846e-06, "loss": 0.5139, "step": 4279 }, { "epoch": 0.17941167223834925, "grad_norm": 1.9486992359161377, "learning_rate": 9.425968975555493e-06, "loss": 0.4953, "step": 4280 }, { "epoch": 0.1794535908533582, "grad_norm": 1.9686468839645386, "learning_rate": 9.425653117422519e-06, "loss": 0.5263, "step": 4281 }, { "epoch": 0.17949550946836718, "grad_norm": 1.6586637496948242, "learning_rate": 9.42533717770875e-06, "loss": 0.4978, "step": 4282 }, { "epoch": 0.17953742808337614, "grad_norm": 1.8512831926345825, "learning_rate": 9.425021156420005e-06, "loss": 0.5547, "step": 4283 }, { "epoch": 0.17957934669838507, "grad_norm": 1.6132255792617798, "learning_rate": 9.424705053562115e-06, "loss": 0.566, "step": 4284 }, { "epoch": 0.17962126531339404, "grad_norm": 1.5320110321044922, "learning_rate": 9.424388869140903e-06, "loss": 0.5032, "step": 4285 }, { "epoch": 0.179663183928403, "grad_norm": 2.0099937915802, "learning_rate": 9.424072603162199e-06, "loss": 0.6402, "step": 4286 }, { "epoch": 0.17970510254341196, "grad_norm": 1.7018965482711792, "learning_rate": 9.423756255631834e-06, "loss": 0.5684, "step": 4287 }, { "epoch": 0.17974702115842092, "grad_norm": 1.9227174520492554, "learning_rate": 9.423439826555636e-06, "loss": 0.59, "step": 4288 }, { "epoch": 0.17978893977342988, "grad_norm": 1.9955183267593384, "learning_rate": 9.423123315939441e-06, "loss": 0.5609, "step": 4289 }, { "epoch": 0.17983085838843885, "grad_norm": 1.795446753501892, "learning_rate": 9.422806723789081e-06, "loss": 0.5684, "step": 4290 }, { "epoch": 0.1798727770034478, "grad_norm": 1.6463240385055542, "learning_rate": 9.422490050110396e-06, "loss": 0.5495, "step": 4291 }, { "epoch": 0.17991469561845677, "grad_norm": 1.6859017610549927, "learning_rate": 9.42217329490922e-06, "loss": 0.5554, "step": 4292 }, { "epoch": 0.17995661423346573, "grad_norm": 1.4463555812835693, "learning_rate": 9.421856458191392e-06, "loss": 0.4918, "step": 4293 }, { "epoch": 0.1799985328484747, "grad_norm": 1.7766633033752441, "learning_rate": 9.421539539962754e-06, "loss": 0.6083, "step": 4294 }, { "epoch": 0.18004045146348366, "grad_norm": 1.9372819662094116, "learning_rate": 9.421222540229148e-06, "loss": 0.5797, "step": 4295 }, { "epoch": 0.1800823700784926, "grad_norm": 2.219200611114502, "learning_rate": 9.420905458996414e-06, "loss": 0.5878, "step": 4296 }, { "epoch": 0.18012428869350156, "grad_norm": 1.58697509765625, "learning_rate": 9.420588296270402e-06, "loss": 0.5856, "step": 4297 }, { "epoch": 0.18016620730851052, "grad_norm": 1.4157140254974365, "learning_rate": 9.420271052056954e-06, "loss": 0.5182, "step": 4298 }, { "epoch": 0.18020812592351948, "grad_norm": 1.8544963598251343, "learning_rate": 9.41995372636192e-06, "loss": 0.5516, "step": 4299 }, { "epoch": 0.18025004453852844, "grad_norm": 1.7737654447555542, "learning_rate": 9.419636319191153e-06, "loss": 0.5403, "step": 4300 }, { "epoch": 0.1802919631535374, "grad_norm": 1.76152503490448, "learning_rate": 9.419318830550497e-06, "loss": 0.5292, "step": 4301 }, { "epoch": 0.18033388176854637, "grad_norm": 1.691458821296692, "learning_rate": 9.419001260445808e-06, "loss": 0.6301, "step": 4302 }, { "epoch": 0.18037580038355533, "grad_norm": 1.622046709060669, "learning_rate": 9.418683608882941e-06, "loss": 0.5463, "step": 4303 }, { "epoch": 0.1804177189985643, "grad_norm": 1.8764479160308838, "learning_rate": 9.41836587586775e-06, "loss": 0.5766, "step": 4304 }, { "epoch": 0.18045963761357325, "grad_norm": 1.5069092512130737, "learning_rate": 9.418048061406092e-06, "loss": 0.5418, "step": 4305 }, { "epoch": 0.18050155622858222, "grad_norm": 2.2896673679351807, "learning_rate": 9.417730165503826e-06, "loss": 0.631, "step": 4306 }, { "epoch": 0.18054347484359118, "grad_norm": 1.8008219003677368, "learning_rate": 9.417412188166813e-06, "loss": 0.5798, "step": 4307 }, { "epoch": 0.18058539345860014, "grad_norm": 1.6182169914245605, "learning_rate": 9.41709412940091e-06, "loss": 0.5691, "step": 4308 }, { "epoch": 0.18062731207360908, "grad_norm": 2.549264430999756, "learning_rate": 9.416775989211985e-06, "loss": 0.5313, "step": 4309 }, { "epoch": 0.18066923068861804, "grad_norm": 1.8992297649383545, "learning_rate": 9.416457767605901e-06, "loss": 0.5234, "step": 4310 }, { "epoch": 0.180711149303627, "grad_norm": 1.6571136713027954, "learning_rate": 9.416139464588525e-06, "loss": 0.5492, "step": 4311 }, { "epoch": 0.18075306791863596, "grad_norm": 2.6271562576293945, "learning_rate": 9.415821080165722e-06, "loss": 0.5483, "step": 4312 }, { "epoch": 0.18079498653364492, "grad_norm": 1.6688838005065918, "learning_rate": 9.415502614343361e-06, "loss": 0.5047, "step": 4313 }, { "epoch": 0.1808369051486539, "grad_norm": 1.9827200174331665, "learning_rate": 9.415184067127314e-06, "loss": 0.5844, "step": 4314 }, { "epoch": 0.18087882376366285, "grad_norm": 1.6747682094573975, "learning_rate": 9.414865438523452e-06, "loss": 0.5671, "step": 4315 }, { "epoch": 0.1809207423786718, "grad_norm": 1.6264394521713257, "learning_rate": 9.41454672853765e-06, "loss": 0.5726, "step": 4316 }, { "epoch": 0.18096266099368077, "grad_norm": 1.6198008060455322, "learning_rate": 9.414227937175782e-06, "loss": 0.5284, "step": 4317 }, { "epoch": 0.18100457960868974, "grad_norm": 1.6471247673034668, "learning_rate": 9.413909064443725e-06, "loss": 0.5099, "step": 4318 }, { "epoch": 0.1810464982236987, "grad_norm": 1.8217085599899292, "learning_rate": 9.413590110347355e-06, "loss": 0.5781, "step": 4319 }, { "epoch": 0.18108841683870766, "grad_norm": 1.5687551498413086, "learning_rate": 9.413271074892553e-06, "loss": 0.5682, "step": 4320 }, { "epoch": 0.1811303354537166, "grad_norm": 1.8531770706176758, "learning_rate": 9.4129519580852e-06, "loss": 0.6338, "step": 4321 }, { "epoch": 0.18117225406872556, "grad_norm": 1.7827072143554688, "learning_rate": 9.41263275993118e-06, "loss": 0.4831, "step": 4322 }, { "epoch": 0.18121417268373452, "grad_norm": 1.6098445653915405, "learning_rate": 9.412313480436373e-06, "loss": 0.5324, "step": 4323 }, { "epoch": 0.18125609129874348, "grad_norm": 1.5906574726104736, "learning_rate": 9.411994119606669e-06, "loss": 0.5263, "step": 4324 }, { "epoch": 0.18129800991375244, "grad_norm": 1.7156389951705933, "learning_rate": 9.411674677447951e-06, "loss": 0.5574, "step": 4325 }, { "epoch": 0.1813399285287614, "grad_norm": 1.7772278785705566, "learning_rate": 9.41135515396611e-06, "loss": 0.5792, "step": 4326 }, { "epoch": 0.18138184714377037, "grad_norm": 1.5904937982559204, "learning_rate": 9.411035549167036e-06, "loss": 0.5646, "step": 4327 }, { "epoch": 0.18142376575877933, "grad_norm": 1.6137112379074097, "learning_rate": 9.410715863056619e-06, "loss": 0.5459, "step": 4328 }, { "epoch": 0.1814656843737883, "grad_norm": 1.7943346500396729, "learning_rate": 9.410396095640753e-06, "loss": 0.6279, "step": 4329 }, { "epoch": 0.18150760298879726, "grad_norm": 1.726824402809143, "learning_rate": 9.41007624692533e-06, "loss": 0.6138, "step": 4330 }, { "epoch": 0.18154952160380622, "grad_norm": 1.7713626623153687, "learning_rate": 9.409756316916251e-06, "loss": 0.509, "step": 4331 }, { "epoch": 0.18159144021881518, "grad_norm": 1.833476185798645, "learning_rate": 9.40943630561941e-06, "loss": 0.5703, "step": 4332 }, { "epoch": 0.18163335883382414, "grad_norm": 1.6796257495880127, "learning_rate": 9.409116213040706e-06, "loss": 0.4903, "step": 4333 }, { "epoch": 0.18167527744883308, "grad_norm": 1.9579803943634033, "learning_rate": 9.40879603918604e-06, "loss": 0.5753, "step": 4334 }, { "epoch": 0.18171719606384204, "grad_norm": 1.8104634284973145, "learning_rate": 9.408475784061314e-06, "loss": 0.5484, "step": 4335 }, { "epoch": 0.181759114678851, "grad_norm": 1.7274014949798584, "learning_rate": 9.408155447672431e-06, "loss": 0.5798, "step": 4336 }, { "epoch": 0.18180103329385996, "grad_norm": 1.766646146774292, "learning_rate": 9.4078350300253e-06, "loss": 0.6159, "step": 4337 }, { "epoch": 0.18184295190886893, "grad_norm": 1.8030884265899658, "learning_rate": 9.407514531125819e-06, "loss": 0.56, "step": 4338 }, { "epoch": 0.1818848705238779, "grad_norm": 1.6265326738357544, "learning_rate": 9.407193950979903e-06, "loss": 0.5609, "step": 4339 }, { "epoch": 0.18192678913888685, "grad_norm": 1.5890108346939087, "learning_rate": 9.406873289593459e-06, "loss": 0.4974, "step": 4340 }, { "epoch": 0.18196870775389581, "grad_norm": 1.6476707458496094, "learning_rate": 9.4065525469724e-06, "loss": 0.5073, "step": 4341 }, { "epoch": 0.18201062636890478, "grad_norm": 1.6799129247665405, "learning_rate": 9.406231723122633e-06, "loss": 0.4953, "step": 4342 }, { "epoch": 0.18205254498391374, "grad_norm": 1.763967752456665, "learning_rate": 9.405910818050078e-06, "loss": 0.5721, "step": 4343 }, { "epoch": 0.1820944635989227, "grad_norm": 2.20866060256958, "learning_rate": 9.405589831760647e-06, "loss": 0.5121, "step": 4344 }, { "epoch": 0.18213638221393166, "grad_norm": 1.9164282083511353, "learning_rate": 9.405268764260258e-06, "loss": 0.577, "step": 4345 }, { "epoch": 0.1821783008289406, "grad_norm": 1.959297776222229, "learning_rate": 9.404947615554832e-06, "loss": 0.5832, "step": 4346 }, { "epoch": 0.18222021944394956, "grad_norm": 1.5434467792510986, "learning_rate": 9.404626385650283e-06, "loss": 0.5187, "step": 4347 }, { "epoch": 0.18226213805895852, "grad_norm": 1.762453317642212, "learning_rate": 9.404305074552538e-06, "loss": 0.5393, "step": 4348 }, { "epoch": 0.18230405667396749, "grad_norm": 1.7852883338928223, "learning_rate": 9.403983682267518e-06, "loss": 0.6325, "step": 4349 }, { "epoch": 0.18234597528897645, "grad_norm": 1.7177097797393799, "learning_rate": 9.403662208801145e-06, "loss": 0.6104, "step": 4350 }, { "epoch": 0.1823878939039854, "grad_norm": 1.611894965171814, "learning_rate": 9.403340654159347e-06, "loss": 0.5639, "step": 4351 }, { "epoch": 0.18242981251899437, "grad_norm": 1.6162039041519165, "learning_rate": 9.403019018348055e-06, "loss": 0.5498, "step": 4352 }, { "epoch": 0.18247173113400333, "grad_norm": 1.5630648136138916, "learning_rate": 9.40269730137319e-06, "loss": 0.5979, "step": 4353 }, { "epoch": 0.1825136497490123, "grad_norm": 1.6335654258728027, "learning_rate": 9.402375503240688e-06, "loss": 0.5703, "step": 4354 }, { "epoch": 0.18255556836402126, "grad_norm": 2.3348610401153564, "learning_rate": 9.402053623956482e-06, "loss": 0.6122, "step": 4355 }, { "epoch": 0.18259748697903022, "grad_norm": 1.4984281063079834, "learning_rate": 9.401731663526502e-06, "loss": 0.4929, "step": 4356 }, { "epoch": 0.18263940559403918, "grad_norm": 1.7171030044555664, "learning_rate": 9.401409621956683e-06, "loss": 0.5408, "step": 4357 }, { "epoch": 0.18268132420904815, "grad_norm": 1.6184743642807007, "learning_rate": 9.401087499252964e-06, "loss": 0.5655, "step": 4358 }, { "epoch": 0.18272324282405708, "grad_norm": 1.8684167861938477, "learning_rate": 9.40076529542128e-06, "loss": 0.6191, "step": 4359 }, { "epoch": 0.18276516143906604, "grad_norm": 1.6437342166900635, "learning_rate": 9.40044301046757e-06, "loss": 0.5942, "step": 4360 }, { "epoch": 0.182807080054075, "grad_norm": 1.6656486988067627, "learning_rate": 9.400120644397778e-06, "loss": 0.57, "step": 4361 }, { "epoch": 0.18284899866908397, "grad_norm": 2.012532949447632, "learning_rate": 9.399798197217846e-06, "loss": 0.5846, "step": 4362 }, { "epoch": 0.18289091728409293, "grad_norm": 1.7135058641433716, "learning_rate": 9.399475668933718e-06, "loss": 0.5917, "step": 4363 }, { "epoch": 0.1829328358991019, "grad_norm": 1.8576165437698364, "learning_rate": 9.399153059551335e-06, "loss": 0.5163, "step": 4364 }, { "epoch": 0.18297475451411085, "grad_norm": 2.0114901065826416, "learning_rate": 9.398830369076647e-06, "loss": 0.5577, "step": 4365 }, { "epoch": 0.18301667312911982, "grad_norm": 1.541257381439209, "learning_rate": 9.398507597515605e-06, "loss": 0.5476, "step": 4366 }, { "epoch": 0.18305859174412878, "grad_norm": 1.5781680345535278, "learning_rate": 9.398184744874156e-06, "loss": 0.5747, "step": 4367 }, { "epoch": 0.18310051035913774, "grad_norm": 1.6490439176559448, "learning_rate": 9.39786181115825e-06, "loss": 0.5335, "step": 4368 }, { "epoch": 0.1831424289741467, "grad_norm": 1.8163208961486816, "learning_rate": 9.397538796373841e-06, "loss": 0.5191, "step": 4369 }, { "epoch": 0.18318434758915567, "grad_norm": 1.7254446744918823, "learning_rate": 9.397215700526884e-06, "loss": 0.5525, "step": 4370 }, { "epoch": 0.1832262662041646, "grad_norm": 1.8997548818588257, "learning_rate": 9.396892523623335e-06, "loss": 0.4819, "step": 4371 }, { "epoch": 0.18326818481917356, "grad_norm": 1.5791425704956055, "learning_rate": 9.39656926566915e-06, "loss": 0.5297, "step": 4372 }, { "epoch": 0.18331010343418253, "grad_norm": 1.603246808052063, "learning_rate": 9.396245926670292e-06, "loss": 0.5032, "step": 4373 }, { "epoch": 0.1833520220491915, "grad_norm": 2.1172475814819336, "learning_rate": 9.395922506632713e-06, "loss": 0.5806, "step": 4374 }, { "epoch": 0.18339394066420045, "grad_norm": 1.9014972448349, "learning_rate": 9.395599005562383e-06, "loss": 0.6103, "step": 4375 }, { "epoch": 0.1834358592792094, "grad_norm": 2.502910614013672, "learning_rate": 9.395275423465262e-06, "loss": 0.6056, "step": 4376 }, { "epoch": 0.18347777789421837, "grad_norm": 1.6361274719238281, "learning_rate": 9.394951760347312e-06, "loss": 0.5415, "step": 4377 }, { "epoch": 0.18351969650922734, "grad_norm": 2.02573823928833, "learning_rate": 9.394628016214506e-06, "loss": 0.6544, "step": 4378 }, { "epoch": 0.1835616151242363, "grad_norm": 1.8972474336624146, "learning_rate": 9.394304191072805e-06, "loss": 0.4887, "step": 4379 }, { "epoch": 0.18360353373924526, "grad_norm": 1.6245241165161133, "learning_rate": 9.393980284928184e-06, "loss": 0.6144, "step": 4380 }, { "epoch": 0.18364545235425422, "grad_norm": 1.8812410831451416, "learning_rate": 9.393656297786609e-06, "loss": 0.553, "step": 4381 }, { "epoch": 0.1836873709692632, "grad_norm": 1.807685375213623, "learning_rate": 9.393332229654054e-06, "loss": 0.5521, "step": 4382 }, { "epoch": 0.18372928958427215, "grad_norm": 1.7606745958328247, "learning_rate": 9.393008080536494e-06, "loss": 0.5333, "step": 4383 }, { "epoch": 0.18377120819928108, "grad_norm": 1.7096360921859741, "learning_rate": 9.392683850439901e-06, "loss": 0.4804, "step": 4384 }, { "epoch": 0.18381312681429005, "grad_norm": 1.7923634052276611, "learning_rate": 9.392359539370257e-06, "loss": 0.5863, "step": 4385 }, { "epoch": 0.183855045429299, "grad_norm": 1.7768882513046265, "learning_rate": 9.392035147333534e-06, "loss": 0.5319, "step": 4386 }, { "epoch": 0.18389696404430797, "grad_norm": 1.7152197360992432, "learning_rate": 9.391710674335717e-06, "loss": 0.6012, "step": 4387 }, { "epoch": 0.18393888265931693, "grad_norm": 1.5157372951507568, "learning_rate": 9.391386120382785e-06, "loss": 0.5328, "step": 4388 }, { "epoch": 0.1839808012743259, "grad_norm": 1.59944486618042, "learning_rate": 9.39106148548072e-06, "loss": 0.5455, "step": 4389 }, { "epoch": 0.18402271988933486, "grad_norm": 1.7402968406677246, "learning_rate": 9.390736769635508e-06, "loss": 0.5453, "step": 4390 }, { "epoch": 0.18406463850434382, "grad_norm": 2.784468412399292, "learning_rate": 9.390411972853133e-06, "loss": 0.5534, "step": 4391 }, { "epoch": 0.18410655711935278, "grad_norm": 1.643625020980835, "learning_rate": 9.390087095139582e-06, "loss": 0.5552, "step": 4392 }, { "epoch": 0.18414847573436174, "grad_norm": 1.6981173753738403, "learning_rate": 9.389762136500846e-06, "loss": 0.5386, "step": 4393 }, { "epoch": 0.1841903943493707, "grad_norm": 1.8483065366744995, "learning_rate": 9.389437096942912e-06, "loss": 0.591, "step": 4394 }, { "epoch": 0.18423231296437967, "grad_norm": 1.7894798517227173, "learning_rate": 9.389111976471775e-06, "loss": 0.5491, "step": 4395 }, { "epoch": 0.1842742315793886, "grad_norm": 2.0815091133117676, "learning_rate": 9.388786775093427e-06, "loss": 0.552, "step": 4396 }, { "epoch": 0.18431615019439757, "grad_norm": 1.7867463827133179, "learning_rate": 9.38846149281386e-06, "loss": 0.5541, "step": 4397 }, { "epoch": 0.18435806880940653, "grad_norm": 1.9754748344421387, "learning_rate": 9.388136129639075e-06, "loss": 0.5903, "step": 4398 }, { "epoch": 0.1843999874244155, "grad_norm": 1.5339020490646362, "learning_rate": 9.387810685575063e-06, "loss": 0.5406, "step": 4399 }, { "epoch": 0.18444190603942445, "grad_norm": 1.7087124586105347, "learning_rate": 9.38748516062783e-06, "loss": 0.5601, "step": 4400 }, { "epoch": 0.18448382465443341, "grad_norm": 2.2339813709259033, "learning_rate": 9.387159554803372e-06, "loss": 0.5242, "step": 4401 }, { "epoch": 0.18452574326944238, "grad_norm": 2.0369014739990234, "learning_rate": 9.386833868107694e-06, "loss": 0.5686, "step": 4402 }, { "epoch": 0.18456766188445134, "grad_norm": 1.943081259727478, "learning_rate": 9.386508100546798e-06, "loss": 0.5706, "step": 4403 }, { "epoch": 0.1846095804994603, "grad_norm": 1.7177271842956543, "learning_rate": 9.386182252126688e-06, "loss": 0.5454, "step": 4404 }, { "epoch": 0.18465149911446926, "grad_norm": 2.2091007232666016, "learning_rate": 9.385856322853373e-06, "loss": 0.5658, "step": 4405 }, { "epoch": 0.18469341772947823, "grad_norm": 1.6720576286315918, "learning_rate": 9.38553031273286e-06, "loss": 0.5589, "step": 4406 }, { "epoch": 0.1847353363444872, "grad_norm": 1.743514895439148, "learning_rate": 9.385204221771158e-06, "loss": 0.5705, "step": 4407 }, { "epoch": 0.18477725495949615, "grad_norm": 1.6629160642623901, "learning_rate": 9.384878049974278e-06, "loss": 0.5544, "step": 4408 }, { "epoch": 0.18481917357450509, "grad_norm": 1.802521824836731, "learning_rate": 9.384551797348235e-06, "loss": 0.5727, "step": 4409 }, { "epoch": 0.18486109218951405, "grad_norm": 1.725213885307312, "learning_rate": 9.38422546389904e-06, "loss": 0.5432, "step": 4410 }, { "epoch": 0.184903010804523, "grad_norm": 1.7133890390396118, "learning_rate": 9.38389904963271e-06, "loss": 0.589, "step": 4411 }, { "epoch": 0.18494492941953197, "grad_norm": 1.7122255563735962, "learning_rate": 9.38357255455526e-06, "loss": 0.533, "step": 4412 }, { "epoch": 0.18498684803454093, "grad_norm": 1.607399344444275, "learning_rate": 9.383245978672711e-06, "loss": 0.549, "step": 4413 }, { "epoch": 0.1850287666495499, "grad_norm": 1.8858169317245483, "learning_rate": 9.382919321991082e-06, "loss": 0.6043, "step": 4414 }, { "epoch": 0.18507068526455886, "grad_norm": 1.7060731649398804, "learning_rate": 9.382592584516395e-06, "loss": 0.5579, "step": 4415 }, { "epoch": 0.18511260387956782, "grad_norm": 1.5952779054641724, "learning_rate": 9.382265766254672e-06, "loss": 0.5745, "step": 4416 }, { "epoch": 0.18515452249457678, "grad_norm": 2.225745916366577, "learning_rate": 9.38193886721194e-06, "loss": 0.5729, "step": 4417 }, { "epoch": 0.18519644110958575, "grad_norm": 1.837511658668518, "learning_rate": 9.38161188739422e-06, "loss": 0.6732, "step": 4418 }, { "epoch": 0.1852383597245947, "grad_norm": 1.7957643270492554, "learning_rate": 9.381284826807543e-06, "loss": 0.545, "step": 4419 }, { "epoch": 0.18528027833960367, "grad_norm": 2.128113269805908, "learning_rate": 9.380957685457936e-06, "loss": 0.5176, "step": 4420 }, { "epoch": 0.1853221969546126, "grad_norm": 1.729100227355957, "learning_rate": 9.380630463351432e-06, "loss": 0.5462, "step": 4421 }, { "epoch": 0.18536411556962157, "grad_norm": 1.5980830192565918, "learning_rate": 9.380303160494062e-06, "loss": 0.5392, "step": 4422 }, { "epoch": 0.18540603418463053, "grad_norm": 1.651346206665039, "learning_rate": 9.379975776891858e-06, "loss": 0.5418, "step": 4423 }, { "epoch": 0.1854479527996395, "grad_norm": 1.9476884603500366, "learning_rate": 9.379648312550857e-06, "loss": 0.55, "step": 4424 }, { "epoch": 0.18548987141464846, "grad_norm": 1.6870070695877075, "learning_rate": 9.379320767477092e-06, "loss": 0.5029, "step": 4425 }, { "epoch": 0.18553179002965742, "grad_norm": 1.6511300802230835, "learning_rate": 9.378993141676605e-06, "loss": 0.5086, "step": 4426 }, { "epoch": 0.18557370864466638, "grad_norm": 2.0460665225982666, "learning_rate": 9.378665435155431e-06, "loss": 0.5467, "step": 4427 }, { "epoch": 0.18561562725967534, "grad_norm": 1.6205178499221802, "learning_rate": 9.378337647919614e-06, "loss": 0.5685, "step": 4428 }, { "epoch": 0.1856575458746843, "grad_norm": 2.2416248321533203, "learning_rate": 9.378009779975194e-06, "loss": 0.5931, "step": 4429 }, { "epoch": 0.18569946448969327, "grad_norm": 2.0378615856170654, "learning_rate": 9.377681831328218e-06, "loss": 0.5688, "step": 4430 }, { "epoch": 0.18574138310470223, "grad_norm": 1.8288599252700806, "learning_rate": 9.37735380198473e-06, "loss": 0.5812, "step": 4431 }, { "epoch": 0.1857833017197112, "grad_norm": 1.4484208822250366, "learning_rate": 9.377025691950774e-06, "loss": 0.5879, "step": 4432 }, { "epoch": 0.18582522033472015, "grad_norm": 1.6758322715759277, "learning_rate": 9.376697501232402e-06, "loss": 0.5628, "step": 4433 }, { "epoch": 0.1858671389497291, "grad_norm": 1.50698721408844, "learning_rate": 9.376369229835662e-06, "loss": 0.5636, "step": 4434 }, { "epoch": 0.18590905756473805, "grad_norm": 1.6338461637496948, "learning_rate": 9.376040877766605e-06, "loss": 0.5967, "step": 4435 }, { "epoch": 0.185950976179747, "grad_norm": 1.8096495866775513, "learning_rate": 9.375712445031283e-06, "loss": 0.5372, "step": 4436 }, { "epoch": 0.18599289479475598, "grad_norm": 1.9332770109176636, "learning_rate": 9.375383931635754e-06, "loss": 0.5713, "step": 4437 }, { "epoch": 0.18603481340976494, "grad_norm": 2.0297789573669434, "learning_rate": 9.375055337586068e-06, "loss": 0.5579, "step": 4438 }, { "epoch": 0.1860767320247739, "grad_norm": 1.7041491270065308, "learning_rate": 9.374726662888287e-06, "loss": 0.5619, "step": 4439 }, { "epoch": 0.18611865063978286, "grad_norm": 2.947340488433838, "learning_rate": 9.374397907548467e-06, "loss": 0.544, "step": 4440 }, { "epoch": 0.18616056925479182, "grad_norm": 1.4338209629058838, "learning_rate": 9.37406907157267e-06, "loss": 0.5358, "step": 4441 }, { "epoch": 0.1862024878698008, "grad_norm": 2.282881259918213, "learning_rate": 9.373740154966955e-06, "loss": 0.5536, "step": 4442 }, { "epoch": 0.18624440648480975, "grad_norm": 1.7073307037353516, "learning_rate": 9.373411157737388e-06, "loss": 0.5423, "step": 4443 }, { "epoch": 0.1862863250998187, "grad_norm": 1.8005257844924927, "learning_rate": 9.373082079890033e-06, "loss": 0.5527, "step": 4444 }, { "epoch": 0.18632824371482767, "grad_norm": 1.868934154510498, "learning_rate": 9.372752921430954e-06, "loss": 0.5189, "step": 4445 }, { "epoch": 0.1863701623298366, "grad_norm": 1.748734474182129, "learning_rate": 9.372423682366222e-06, "loss": 0.5347, "step": 4446 }, { "epoch": 0.18641208094484557, "grad_norm": 4.307859897613525, "learning_rate": 9.372094362701902e-06, "loss": 0.5279, "step": 4447 }, { "epoch": 0.18645399955985453, "grad_norm": 1.6385172605514526, "learning_rate": 9.371764962444068e-06, "loss": 0.5204, "step": 4448 }, { "epoch": 0.1864959181748635, "grad_norm": 1.5428961515426636, "learning_rate": 9.371435481598789e-06, "loss": 0.49, "step": 4449 }, { "epoch": 0.18653783678987246, "grad_norm": 1.7213640213012695, "learning_rate": 9.371105920172143e-06, "loss": 0.6289, "step": 4450 }, { "epoch": 0.18657975540488142, "grad_norm": 2.01719069480896, "learning_rate": 9.3707762781702e-06, "loss": 0.5232, "step": 4451 }, { "epoch": 0.18662167401989038, "grad_norm": 2.314972162246704, "learning_rate": 9.37044655559904e-06, "loss": 0.5852, "step": 4452 }, { "epoch": 0.18666359263489934, "grad_norm": 1.558192491531372, "learning_rate": 9.370116752464738e-06, "loss": 0.5536, "step": 4453 }, { "epoch": 0.1867055112499083, "grad_norm": 2.243940591812134, "learning_rate": 9.369786868773377e-06, "loss": 0.6153, "step": 4454 }, { "epoch": 0.18674742986491727, "grad_norm": 1.7645272016525269, "learning_rate": 9.369456904531036e-06, "loss": 0.6526, "step": 4455 }, { "epoch": 0.18678934847992623, "grad_norm": 1.688520073890686, "learning_rate": 9.369126859743796e-06, "loss": 0.55, "step": 4456 }, { "epoch": 0.1868312670949352, "grad_norm": 1.9593487977981567, "learning_rate": 9.368796734417745e-06, "loss": 0.5493, "step": 4457 }, { "epoch": 0.18687318570994416, "grad_norm": 1.6362496614456177, "learning_rate": 9.368466528558965e-06, "loss": 0.5519, "step": 4458 }, { "epoch": 0.1869151043249531, "grad_norm": 1.899114966392517, "learning_rate": 9.368136242173541e-06, "loss": 0.5704, "step": 4459 }, { "epoch": 0.18695702293996205, "grad_norm": 1.9975225925445557, "learning_rate": 9.367805875267566e-06, "loss": 0.5002, "step": 4460 }, { "epoch": 0.18699894155497102, "grad_norm": 1.7409541606903076, "learning_rate": 9.367475427847127e-06, "loss": 0.5633, "step": 4461 }, { "epoch": 0.18704086016997998, "grad_norm": 1.6759569644927979, "learning_rate": 9.36714489991832e-06, "loss": 0.5379, "step": 4462 }, { "epoch": 0.18708277878498894, "grad_norm": 1.877039909362793, "learning_rate": 9.366814291487232e-06, "loss": 0.5716, "step": 4463 }, { "epoch": 0.1871246973999979, "grad_norm": 1.6084339618682861, "learning_rate": 9.366483602559958e-06, "loss": 0.544, "step": 4464 }, { "epoch": 0.18716661601500686, "grad_norm": 1.6627286672592163, "learning_rate": 9.366152833142597e-06, "loss": 0.5193, "step": 4465 }, { "epoch": 0.18720853463001583, "grad_norm": 1.6629118919372559, "learning_rate": 9.365821983241243e-06, "loss": 0.5408, "step": 4466 }, { "epoch": 0.1872504532450248, "grad_norm": 1.958840012550354, "learning_rate": 9.365491052861997e-06, "loss": 0.5491, "step": 4467 }, { "epoch": 0.18729237186003375, "grad_norm": 1.5955218076705933, "learning_rate": 9.365160042010961e-06, "loss": 0.5374, "step": 4468 }, { "epoch": 0.18733429047504271, "grad_norm": 1.6918312311172485, "learning_rate": 9.36482895069423e-06, "loss": 0.569, "step": 4469 }, { "epoch": 0.18737620909005168, "grad_norm": 1.5600146055221558, "learning_rate": 9.364497778917915e-06, "loss": 0.5325, "step": 4470 }, { "epoch": 0.1874181277050606, "grad_norm": 1.8171240091323853, "learning_rate": 9.364166526688117e-06, "loss": 0.571, "step": 4471 }, { "epoch": 0.18746004632006957, "grad_norm": 1.6140915155410767, "learning_rate": 9.363835194010939e-06, "loss": 0.569, "step": 4472 }, { "epoch": 0.18750196493507854, "grad_norm": 1.6095836162567139, "learning_rate": 9.363503780892496e-06, "loss": 0.5662, "step": 4473 }, { "epoch": 0.1875438835500875, "grad_norm": 1.5969109535217285, "learning_rate": 9.36317228733889e-06, "loss": 0.5431, "step": 4474 }, { "epoch": 0.18758580216509646, "grad_norm": 2.060992479324341, "learning_rate": 9.362840713356236e-06, "loss": 0.5683, "step": 4475 }, { "epoch": 0.18762772078010542, "grad_norm": 1.5572515726089478, "learning_rate": 9.362509058950646e-06, "loss": 0.5694, "step": 4476 }, { "epoch": 0.18766963939511438, "grad_norm": 1.906462550163269, "learning_rate": 9.362177324128232e-06, "loss": 0.5936, "step": 4477 }, { "epoch": 0.18771155801012335, "grad_norm": 1.6850162744522095, "learning_rate": 9.361845508895108e-06, "loss": 0.5768, "step": 4478 }, { "epoch": 0.1877534766251323, "grad_norm": 1.7632980346679688, "learning_rate": 9.361513613257394e-06, "loss": 0.5609, "step": 4479 }, { "epoch": 0.18779539524014127, "grad_norm": 2.4630486965179443, "learning_rate": 9.361181637221205e-06, "loss": 0.5678, "step": 4480 }, { "epoch": 0.18783731385515023, "grad_norm": 1.6010868549346924, "learning_rate": 9.360849580792661e-06, "loss": 0.5259, "step": 4481 }, { "epoch": 0.1878792324701592, "grad_norm": 1.6518034934997559, "learning_rate": 9.360517443977885e-06, "loss": 0.5154, "step": 4482 }, { "epoch": 0.18792115108516816, "grad_norm": 1.7055851221084595, "learning_rate": 9.360185226782998e-06, "loss": 0.5426, "step": 4483 }, { "epoch": 0.1879630697001771, "grad_norm": 1.6149046421051025, "learning_rate": 9.359852929214124e-06, "loss": 0.5184, "step": 4484 }, { "epoch": 0.18800498831518606, "grad_norm": 1.6164747476577759, "learning_rate": 9.35952055127739e-06, "loss": 0.5977, "step": 4485 }, { "epoch": 0.18804690693019502, "grad_norm": 1.7532541751861572, "learning_rate": 9.359188092978919e-06, "loss": 0.5476, "step": 4486 }, { "epoch": 0.18808882554520398, "grad_norm": 2.139286756515503, "learning_rate": 9.358855554324844e-06, "loss": 0.5523, "step": 4487 }, { "epoch": 0.18813074416021294, "grad_norm": 1.8934270143508911, "learning_rate": 9.35852293532129e-06, "loss": 0.5876, "step": 4488 }, { "epoch": 0.1881726627752219, "grad_norm": 1.6644972562789917, "learning_rate": 9.358190235974394e-06, "loss": 0.5451, "step": 4489 }, { "epoch": 0.18821458139023087, "grad_norm": 1.5343823432922363, "learning_rate": 9.357857456290285e-06, "loss": 0.5087, "step": 4490 }, { "epoch": 0.18825650000523983, "grad_norm": 1.6765995025634766, "learning_rate": 9.357524596275098e-06, "loss": 0.5316, "step": 4491 }, { "epoch": 0.1882984186202488, "grad_norm": 1.9086511135101318, "learning_rate": 9.35719165593497e-06, "loss": 0.561, "step": 4492 }, { "epoch": 0.18834033723525775, "grad_norm": 1.7507976293563843, "learning_rate": 9.356858635276035e-06, "loss": 0.5586, "step": 4493 }, { "epoch": 0.18838225585026672, "grad_norm": 1.7510355710983276, "learning_rate": 9.356525534304437e-06, "loss": 0.5439, "step": 4494 }, { "epoch": 0.18842417446527568, "grad_norm": 1.5179351568222046, "learning_rate": 9.356192353026314e-06, "loss": 0.4859, "step": 4495 }, { "epoch": 0.18846609308028464, "grad_norm": 1.7550874948501587, "learning_rate": 9.355859091447806e-06, "loss": 0.5763, "step": 4496 }, { "epoch": 0.18850801169529358, "grad_norm": 1.636663794517517, "learning_rate": 9.355525749575056e-06, "loss": 0.5364, "step": 4497 }, { "epoch": 0.18854993031030254, "grad_norm": 1.7228206396102905, "learning_rate": 9.355192327414212e-06, "loss": 0.5101, "step": 4498 }, { "epoch": 0.1885918489253115, "grad_norm": 1.9729710817337036, "learning_rate": 9.354858824971418e-06, "loss": 0.5593, "step": 4499 }, { "epoch": 0.18863376754032046, "grad_norm": 1.7582571506500244, "learning_rate": 9.35452524225282e-06, "loss": 0.5854, "step": 4500 }, { "epoch": 0.18867568615532943, "grad_norm": 1.7335106134414673, "learning_rate": 9.354191579264572e-06, "loss": 0.5827, "step": 4501 }, { "epoch": 0.1887176047703384, "grad_norm": 1.8219876289367676, "learning_rate": 9.353857836012821e-06, "loss": 0.6059, "step": 4502 }, { "epoch": 0.18875952338534735, "grad_norm": 1.4817811250686646, "learning_rate": 9.35352401250372e-06, "loss": 0.5436, "step": 4503 }, { "epoch": 0.1888014420003563, "grad_norm": 1.6656959056854248, "learning_rate": 9.353190108743421e-06, "loss": 0.5216, "step": 4504 }, { "epoch": 0.18884336061536527, "grad_norm": 1.4805805683135986, "learning_rate": 9.352856124738081e-06, "loss": 0.5935, "step": 4505 }, { "epoch": 0.18888527923037424, "grad_norm": 1.7030978202819824, "learning_rate": 9.352522060493856e-06, "loss": 0.5733, "step": 4506 }, { "epoch": 0.1889271978453832, "grad_norm": 1.7363394498825073, "learning_rate": 9.352187916016904e-06, "loss": 0.5905, "step": 4507 }, { "epoch": 0.18896911646039216, "grad_norm": 1.5803111791610718, "learning_rate": 9.351853691313384e-06, "loss": 0.5776, "step": 4508 }, { "epoch": 0.1890110350754011, "grad_norm": 1.7712212800979614, "learning_rate": 9.35151938638946e-06, "loss": 0.5917, "step": 4509 }, { "epoch": 0.18905295369041006, "grad_norm": 1.725242257118225, "learning_rate": 9.35118500125129e-06, "loss": 0.5891, "step": 4510 }, { "epoch": 0.18909487230541902, "grad_norm": 1.9252078533172607, "learning_rate": 9.350850535905039e-06, "loss": 0.5149, "step": 4511 }, { "epoch": 0.18913679092042798, "grad_norm": 1.5125102996826172, "learning_rate": 9.350515990356874e-06, "loss": 0.576, "step": 4512 }, { "epoch": 0.18917870953543695, "grad_norm": 1.7009432315826416, "learning_rate": 9.350181364612963e-06, "loss": 0.6093, "step": 4513 }, { "epoch": 0.1892206281504459, "grad_norm": 1.8568594455718994, "learning_rate": 9.349846658679468e-06, "loss": 0.6048, "step": 4514 }, { "epoch": 0.18926254676545487, "grad_norm": 2.0467019081115723, "learning_rate": 9.349511872562567e-06, "loss": 0.5743, "step": 4515 }, { "epoch": 0.18930446538046383, "grad_norm": 1.785988688468933, "learning_rate": 9.349177006268427e-06, "loss": 0.5898, "step": 4516 }, { "epoch": 0.1893463839954728, "grad_norm": 4.053733825683594, "learning_rate": 9.348842059803222e-06, "loss": 0.5681, "step": 4517 }, { "epoch": 0.18938830261048176, "grad_norm": 1.7760136127471924, "learning_rate": 9.348507033173125e-06, "loss": 0.5366, "step": 4518 }, { "epoch": 0.18943022122549072, "grad_norm": 1.6588884592056274, "learning_rate": 9.348171926384312e-06, "loss": 0.5139, "step": 4519 }, { "epoch": 0.18947213984049968, "grad_norm": 1.7507978677749634, "learning_rate": 9.347836739442962e-06, "loss": 0.6018, "step": 4520 }, { "epoch": 0.18951405845550864, "grad_norm": 1.6155790090560913, "learning_rate": 9.34750147235525e-06, "loss": 0.5216, "step": 4521 }, { "epoch": 0.18955597707051758, "grad_norm": 1.7779691219329834, "learning_rate": 9.34716612512736e-06, "loss": 0.5748, "step": 4522 }, { "epoch": 0.18959789568552654, "grad_norm": 1.6585345268249512, "learning_rate": 9.346830697765473e-06, "loss": 0.5293, "step": 4523 }, { "epoch": 0.1896398143005355, "grad_norm": 1.8663157224655151, "learning_rate": 9.346495190275771e-06, "loss": 0.5557, "step": 4524 }, { "epoch": 0.18968173291554447, "grad_norm": 1.8660383224487305, "learning_rate": 9.346159602664438e-06, "loss": 0.5408, "step": 4525 }, { "epoch": 0.18972365153055343, "grad_norm": 1.9708828926086426, "learning_rate": 9.345823934937663e-06, "loss": 0.571, "step": 4526 }, { "epoch": 0.1897655701455624, "grad_norm": 1.6852805614471436, "learning_rate": 9.34548818710163e-06, "loss": 0.5195, "step": 4527 }, { "epoch": 0.18980748876057135, "grad_norm": 1.7833538055419922, "learning_rate": 9.34515235916253e-06, "loss": 0.5955, "step": 4528 }, { "epoch": 0.18984940737558031, "grad_norm": 1.4465312957763672, "learning_rate": 9.344816451126552e-06, "loss": 0.5574, "step": 4529 }, { "epoch": 0.18989132599058928, "grad_norm": 1.8185927867889404, "learning_rate": 9.344480462999892e-06, "loss": 0.5996, "step": 4530 }, { "epoch": 0.18993324460559824, "grad_norm": 1.611777424812317, "learning_rate": 9.34414439478874e-06, "loss": 0.5979, "step": 4531 }, { "epoch": 0.1899751632206072, "grad_norm": 1.8824270963668823, "learning_rate": 9.34380824649929e-06, "loss": 0.5494, "step": 4532 }, { "epoch": 0.19001708183561616, "grad_norm": 1.4881521463394165, "learning_rate": 9.343472018137743e-06, "loss": 0.5227, "step": 4533 }, { "epoch": 0.1900590004506251, "grad_norm": 1.5115538835525513, "learning_rate": 9.343135709710292e-06, "loss": 0.5529, "step": 4534 }, { "epoch": 0.19010091906563406, "grad_norm": 1.3312885761260986, "learning_rate": 9.342799321223139e-06, "loss": 0.5026, "step": 4535 }, { "epoch": 0.19014283768064302, "grad_norm": 1.6957412958145142, "learning_rate": 9.342462852682483e-06, "loss": 0.5374, "step": 4536 }, { "epoch": 0.19018475629565199, "grad_norm": 1.9065781831741333, "learning_rate": 9.34212630409453e-06, "loss": 0.5453, "step": 4537 }, { "epoch": 0.19022667491066095, "grad_norm": 1.6619961261749268, "learning_rate": 9.341789675465478e-06, "loss": 0.5203, "step": 4538 }, { "epoch": 0.1902685935256699, "grad_norm": 1.7678821086883545, "learning_rate": 9.341452966801538e-06, "loss": 0.5484, "step": 4539 }, { "epoch": 0.19031051214067887, "grad_norm": 1.6991465091705322, "learning_rate": 9.341116178108914e-06, "loss": 0.6008, "step": 4540 }, { "epoch": 0.19035243075568783, "grad_norm": 1.6870651245117188, "learning_rate": 9.340779309393815e-06, "loss": 0.5476, "step": 4541 }, { "epoch": 0.1903943493706968, "grad_norm": 1.6746138334274292, "learning_rate": 9.340442360662449e-06, "loss": 0.5228, "step": 4542 }, { "epoch": 0.19043626798570576, "grad_norm": 1.6362067461013794, "learning_rate": 9.34010533192103e-06, "loss": 0.5044, "step": 4543 }, { "epoch": 0.19047818660071472, "grad_norm": 1.6004419326782227, "learning_rate": 9.33976822317577e-06, "loss": 0.5531, "step": 4544 }, { "epoch": 0.19052010521572368, "grad_norm": 2.2393195629119873, "learning_rate": 9.339431034432883e-06, "loss": 0.5466, "step": 4545 }, { "epoch": 0.19056202383073265, "grad_norm": 1.6921768188476562, "learning_rate": 9.339093765698581e-06, "loss": 0.559, "step": 4546 }, { "epoch": 0.19060394244574158, "grad_norm": 1.8982435464859009, "learning_rate": 9.338756416979084e-06, "loss": 0.5886, "step": 4547 }, { "epoch": 0.19064586106075054, "grad_norm": 1.5123144388198853, "learning_rate": 9.338418988280614e-06, "loss": 0.5145, "step": 4548 }, { "epoch": 0.1906877796757595, "grad_norm": 1.4942753314971924, "learning_rate": 9.338081479609385e-06, "loss": 0.5092, "step": 4549 }, { "epoch": 0.19072969829076847, "grad_norm": 1.584431767463684, "learning_rate": 9.33774389097162e-06, "loss": 0.5381, "step": 4550 }, { "epoch": 0.19077161690577743, "grad_norm": 1.7544289827346802, "learning_rate": 9.337406222373546e-06, "loss": 0.5335, "step": 4551 }, { "epoch": 0.1908135355207864, "grad_norm": 1.7948074340820312, "learning_rate": 9.337068473821382e-06, "loss": 0.5683, "step": 4552 }, { "epoch": 0.19085545413579535, "grad_norm": 2.0127480030059814, "learning_rate": 9.336730645321358e-06, "loss": 0.5289, "step": 4553 }, { "epoch": 0.19089737275080432, "grad_norm": 1.9304062128067017, "learning_rate": 9.3363927368797e-06, "loss": 0.5247, "step": 4554 }, { "epoch": 0.19093929136581328, "grad_norm": 1.78586745262146, "learning_rate": 9.336054748502637e-06, "loss": 0.594, "step": 4555 }, { "epoch": 0.19098120998082224, "grad_norm": 1.971581220626831, "learning_rate": 9.335716680196398e-06, "loss": 0.5297, "step": 4556 }, { "epoch": 0.1910231285958312, "grad_norm": 3.2224786281585693, "learning_rate": 9.335378531967218e-06, "loss": 0.6003, "step": 4557 }, { "epoch": 0.19106504721084017, "grad_norm": 1.6106750965118408, "learning_rate": 9.335040303821325e-06, "loss": 0.5678, "step": 4558 }, { "epoch": 0.1911069658258491, "grad_norm": 1.8404773473739624, "learning_rate": 9.33470199576496e-06, "loss": 0.5482, "step": 4559 }, { "epoch": 0.19114888444085806, "grad_norm": 1.7925212383270264, "learning_rate": 9.334363607804355e-06, "loss": 0.5517, "step": 4560 }, { "epoch": 0.19119080305586703, "grad_norm": 1.5971819162368774, "learning_rate": 9.334025139945748e-06, "loss": 0.5572, "step": 4561 }, { "epoch": 0.191232721670876, "grad_norm": 1.6255782842636108, "learning_rate": 9.33368659219538e-06, "loss": 0.5268, "step": 4562 }, { "epoch": 0.19127464028588495, "grad_norm": 1.8473684787750244, "learning_rate": 9.333347964559493e-06, "loss": 0.5411, "step": 4563 }, { "epoch": 0.1913165589008939, "grad_norm": 1.6256351470947266, "learning_rate": 9.333009257044325e-06, "loss": 0.5281, "step": 4564 }, { "epoch": 0.19135847751590288, "grad_norm": 1.6591016054153442, "learning_rate": 9.332670469656121e-06, "loss": 0.5478, "step": 4565 }, { "epoch": 0.19140039613091184, "grad_norm": 1.7755409479141235, "learning_rate": 9.332331602401126e-06, "loss": 0.5437, "step": 4566 }, { "epoch": 0.1914423147459208, "grad_norm": 1.8879486322402954, "learning_rate": 9.331992655285589e-06, "loss": 0.55, "step": 4567 }, { "epoch": 0.19148423336092976, "grad_norm": 1.807597279548645, "learning_rate": 9.331653628315756e-06, "loss": 0.5654, "step": 4568 }, { "epoch": 0.19152615197593872, "grad_norm": 1.5974907875061035, "learning_rate": 9.331314521497877e-06, "loss": 0.5429, "step": 4569 }, { "epoch": 0.1915680705909477, "grad_norm": 1.9662894010543823, "learning_rate": 9.330975334838202e-06, "loss": 0.6184, "step": 4570 }, { "epoch": 0.19160998920595665, "grad_norm": 1.8866090774536133, "learning_rate": 9.330636068342982e-06, "loss": 0.5653, "step": 4571 }, { "epoch": 0.19165190782096558, "grad_norm": 2.589963436126709, "learning_rate": 9.330296722018477e-06, "loss": 0.5754, "step": 4572 }, { "epoch": 0.19169382643597455, "grad_norm": 1.9987974166870117, "learning_rate": 9.329957295870936e-06, "loss": 0.5966, "step": 4573 }, { "epoch": 0.1917357450509835, "grad_norm": 1.6883107423782349, "learning_rate": 9.329617789906617e-06, "loss": 0.5208, "step": 4574 }, { "epoch": 0.19177766366599247, "grad_norm": 1.9933022260665894, "learning_rate": 9.329278204131781e-06, "loss": 0.5507, "step": 4575 }, { "epoch": 0.19181958228100143, "grad_norm": 1.9835484027862549, "learning_rate": 9.328938538552688e-06, "loss": 0.6514, "step": 4576 }, { "epoch": 0.1918615008960104, "grad_norm": 1.7961136102676392, "learning_rate": 9.328598793175596e-06, "loss": 0.5518, "step": 4577 }, { "epoch": 0.19190341951101936, "grad_norm": 2.0526485443115234, "learning_rate": 9.328258968006769e-06, "loss": 0.5254, "step": 4578 }, { "epoch": 0.19194533812602832, "grad_norm": 2.0226850509643555, "learning_rate": 9.327919063052473e-06, "loss": 0.5655, "step": 4579 }, { "epoch": 0.19198725674103728, "grad_norm": 1.7611420154571533, "learning_rate": 9.327579078318971e-06, "loss": 0.5865, "step": 4580 }, { "epoch": 0.19202917535604624, "grad_norm": 1.6166925430297852, "learning_rate": 9.327239013812532e-06, "loss": 0.5474, "step": 4581 }, { "epoch": 0.1920710939710552, "grad_norm": 2.0055551528930664, "learning_rate": 9.326898869539424e-06, "loss": 0.6049, "step": 4582 }, { "epoch": 0.19211301258606417, "grad_norm": 1.8622527122497559, "learning_rate": 9.326558645505916e-06, "loss": 0.5347, "step": 4583 }, { "epoch": 0.1921549312010731, "grad_norm": 1.9639166593551636, "learning_rate": 9.326218341718282e-06, "loss": 0.6072, "step": 4584 }, { "epoch": 0.19219684981608207, "grad_norm": 5.799342155456543, "learning_rate": 9.325877958182793e-06, "loss": 0.5346, "step": 4585 }, { "epoch": 0.19223876843109103, "grad_norm": 1.661299705505371, "learning_rate": 9.325537494905724e-06, "loss": 0.5389, "step": 4586 }, { "epoch": 0.1922806870461, "grad_norm": 1.7090030908584595, "learning_rate": 9.325196951893352e-06, "loss": 0.5944, "step": 4587 }, { "epoch": 0.19232260566110895, "grad_norm": 1.7298625707626343, "learning_rate": 9.324856329151954e-06, "loss": 0.5773, "step": 4588 }, { "epoch": 0.19236452427611792, "grad_norm": 2.1563162803649902, "learning_rate": 9.324515626687807e-06, "loss": 0.5436, "step": 4589 }, { "epoch": 0.19240644289112688, "grad_norm": 1.8952248096466064, "learning_rate": 9.324174844507193e-06, "loss": 0.5425, "step": 4590 }, { "epoch": 0.19244836150613584, "grad_norm": 3.206254005432129, "learning_rate": 9.323833982616395e-06, "loss": 0.5148, "step": 4591 }, { "epoch": 0.1924902801211448, "grad_norm": 1.8313385248184204, "learning_rate": 9.323493041021694e-06, "loss": 0.516, "step": 4592 }, { "epoch": 0.19253219873615376, "grad_norm": 1.9199399948120117, "learning_rate": 9.323152019729375e-06, "loss": 0.5728, "step": 4593 }, { "epoch": 0.19257411735116273, "grad_norm": 1.849517583847046, "learning_rate": 9.322810918745727e-06, "loss": 0.5669, "step": 4594 }, { "epoch": 0.1926160359661717, "grad_norm": 1.7479983568191528, "learning_rate": 9.322469738077035e-06, "loss": 0.5207, "step": 4595 }, { "epoch": 0.19265795458118065, "grad_norm": 1.6050182580947876, "learning_rate": 9.32212847772959e-06, "loss": 0.6058, "step": 4596 }, { "epoch": 0.19269987319618959, "grad_norm": 1.8273338079452515, "learning_rate": 9.321787137709679e-06, "loss": 0.5283, "step": 4597 }, { "epoch": 0.19274179181119855, "grad_norm": 1.7296522855758667, "learning_rate": 9.321445718023599e-06, "loss": 0.5356, "step": 4598 }, { "epoch": 0.1927837104262075, "grad_norm": 1.8732116222381592, "learning_rate": 9.321104218677641e-06, "loss": 0.6307, "step": 4599 }, { "epoch": 0.19282562904121647, "grad_norm": 1.7628777027130127, "learning_rate": 9.320762639678101e-06, "loss": 0.4978, "step": 4600 }, { "epoch": 0.19286754765622544, "grad_norm": 2.1463875770568848, "learning_rate": 9.320420981031274e-06, "loss": 0.5835, "step": 4601 }, { "epoch": 0.1929094662712344, "grad_norm": 2.2472221851348877, "learning_rate": 9.320079242743461e-06, "loss": 0.5991, "step": 4602 }, { "epoch": 0.19295138488624336, "grad_norm": 26.331804275512695, "learning_rate": 9.319737424820956e-06, "loss": 0.5113, "step": 4603 }, { "epoch": 0.19299330350125232, "grad_norm": 1.7666015625, "learning_rate": 9.319395527270065e-06, "loss": 0.535, "step": 4604 }, { "epoch": 0.19303522211626128, "grad_norm": 1.657056450843811, "learning_rate": 9.319053550097091e-06, "loss": 0.5336, "step": 4605 }, { "epoch": 0.19307714073127025, "grad_norm": 1.8545128107070923, "learning_rate": 9.318711493308332e-06, "loss": 0.5938, "step": 4606 }, { "epoch": 0.1931190593462792, "grad_norm": 1.443088412284851, "learning_rate": 9.3183693569101e-06, "loss": 0.537, "step": 4607 }, { "epoch": 0.19316097796128817, "grad_norm": 1.78634774684906, "learning_rate": 9.318027140908698e-06, "loss": 0.5967, "step": 4608 }, { "epoch": 0.1932028965762971, "grad_norm": 2.103823184967041, "learning_rate": 9.317684845310435e-06, "loss": 0.5761, "step": 4609 }, { "epoch": 0.19324481519130607, "grad_norm": 1.489309549331665, "learning_rate": 9.31734247012162e-06, "loss": 0.5898, "step": 4610 }, { "epoch": 0.19328673380631503, "grad_norm": 1.8475433588027954, "learning_rate": 9.317000015348568e-06, "loss": 0.5884, "step": 4611 }, { "epoch": 0.193328652421324, "grad_norm": 1.612058162689209, "learning_rate": 9.316657480997585e-06, "loss": 0.5779, "step": 4612 }, { "epoch": 0.19337057103633296, "grad_norm": 1.503296136856079, "learning_rate": 9.316314867074991e-06, "loss": 0.5053, "step": 4613 }, { "epoch": 0.19341248965134192, "grad_norm": 1.6412054300308228, "learning_rate": 9.3159721735871e-06, "loss": 0.5352, "step": 4614 }, { "epoch": 0.19345440826635088, "grad_norm": 2.145853042602539, "learning_rate": 9.315629400540228e-06, "loss": 0.541, "step": 4615 }, { "epoch": 0.19349632688135984, "grad_norm": 3.017043113708496, "learning_rate": 9.315286547940694e-06, "loss": 0.5737, "step": 4616 }, { "epoch": 0.1935382454963688, "grad_norm": 1.893965482711792, "learning_rate": 9.314943615794818e-06, "loss": 0.5653, "step": 4617 }, { "epoch": 0.19358016411137777, "grad_norm": 1.8592458963394165, "learning_rate": 9.314600604108921e-06, "loss": 0.5427, "step": 4618 }, { "epoch": 0.19362208272638673, "grad_norm": 1.6741318702697754, "learning_rate": 9.31425751288933e-06, "loss": 0.5344, "step": 4619 }, { "epoch": 0.1936640013413957, "grad_norm": 1.6721149682998657, "learning_rate": 9.313914342142363e-06, "loss": 0.5551, "step": 4620 }, { "epoch": 0.19370591995640465, "grad_norm": 1.4902420043945312, "learning_rate": 9.31357109187435e-06, "loss": 0.5397, "step": 4621 }, { "epoch": 0.1937478385714136, "grad_norm": 2.052948236465454, "learning_rate": 9.313227762091616e-06, "loss": 0.5484, "step": 4622 }, { "epoch": 0.19378975718642255, "grad_norm": 1.747710943222046, "learning_rate": 9.312884352800491e-06, "loss": 0.503, "step": 4623 }, { "epoch": 0.1938316758014315, "grad_norm": 1.590522050857544, "learning_rate": 9.312540864007307e-06, "loss": 0.5463, "step": 4624 }, { "epoch": 0.19387359441644048, "grad_norm": 1.768412470817566, "learning_rate": 9.312197295718393e-06, "loss": 0.5846, "step": 4625 }, { "epoch": 0.19391551303144944, "grad_norm": 1.6431890726089478, "learning_rate": 9.311853647940084e-06, "loss": 0.5088, "step": 4626 }, { "epoch": 0.1939574316464584, "grad_norm": 1.5356582403182983, "learning_rate": 9.311509920678713e-06, "loss": 0.5331, "step": 4627 }, { "epoch": 0.19399935026146736, "grad_norm": 1.7340275049209595, "learning_rate": 9.311166113940617e-06, "loss": 0.5514, "step": 4628 }, { "epoch": 0.19404126887647632, "grad_norm": 1.7201799154281616, "learning_rate": 9.310822227732134e-06, "loss": 0.5389, "step": 4629 }, { "epoch": 0.1940831874914853, "grad_norm": 7.974297046661377, "learning_rate": 9.310478262059603e-06, "loss": 0.5563, "step": 4630 }, { "epoch": 0.19412510610649425, "grad_norm": 1.7936310768127441, "learning_rate": 9.310134216929365e-06, "loss": 0.6031, "step": 4631 }, { "epoch": 0.1941670247215032, "grad_norm": 1.7766774892807007, "learning_rate": 9.309790092347762e-06, "loss": 0.6018, "step": 4632 }, { "epoch": 0.19420894333651217, "grad_norm": 1.664299726486206, "learning_rate": 9.309445888321134e-06, "loss": 0.4967, "step": 4633 }, { "epoch": 0.1942508619515211, "grad_norm": 2.014047622680664, "learning_rate": 9.309101604855831e-06, "loss": 0.5259, "step": 4634 }, { "epoch": 0.19429278056653007, "grad_norm": 1.7365636825561523, "learning_rate": 9.308757241958196e-06, "loss": 0.5289, "step": 4635 }, { "epoch": 0.19433469918153903, "grad_norm": 1.6220654249191284, "learning_rate": 9.308412799634578e-06, "loss": 0.5122, "step": 4636 }, { "epoch": 0.194376617796548, "grad_norm": 2.2614545822143555, "learning_rate": 9.308068277891326e-06, "loss": 0.5039, "step": 4637 }, { "epoch": 0.19441853641155696, "grad_norm": 1.7729406356811523, "learning_rate": 9.307723676734792e-06, "loss": 0.5307, "step": 4638 }, { "epoch": 0.19446045502656592, "grad_norm": 1.9214478731155396, "learning_rate": 9.307378996171326e-06, "loss": 0.5229, "step": 4639 }, { "epoch": 0.19450237364157488, "grad_norm": 2.330554485321045, "learning_rate": 9.307034236207284e-06, "loss": 0.5831, "step": 4640 }, { "epoch": 0.19454429225658385, "grad_norm": 1.8272851705551147, "learning_rate": 9.30668939684902e-06, "loss": 0.5677, "step": 4641 }, { "epoch": 0.1945862108715928, "grad_norm": 1.7716037034988403, "learning_rate": 9.306344478102892e-06, "loss": 0.527, "step": 4642 }, { "epoch": 0.19462812948660177, "grad_norm": 1.6856111288070679, "learning_rate": 9.305999479975254e-06, "loss": 0.6013, "step": 4643 }, { "epoch": 0.19467004810161073, "grad_norm": 2.2670843601226807, "learning_rate": 9.305654402472469e-06, "loss": 0.5312, "step": 4644 }, { "epoch": 0.1947119667166197, "grad_norm": 1.8266657590866089, "learning_rate": 9.305309245600898e-06, "loss": 0.5961, "step": 4645 }, { "epoch": 0.19475388533162866, "grad_norm": 1.5669139623641968, "learning_rate": 9.304964009366903e-06, "loss": 0.551, "step": 4646 }, { "epoch": 0.1947958039466376, "grad_norm": 1.7037638425827026, "learning_rate": 9.30461869377685e-06, "loss": 0.5915, "step": 4647 }, { "epoch": 0.19483772256164655, "grad_norm": 1.8458967208862305, "learning_rate": 9.3042732988371e-06, "loss": 0.4972, "step": 4648 }, { "epoch": 0.19487964117665552, "grad_norm": 1.7373347282409668, "learning_rate": 9.303927824554022e-06, "loss": 0.5494, "step": 4649 }, { "epoch": 0.19492155979166448, "grad_norm": 1.7879117727279663, "learning_rate": 9.303582270933986e-06, "loss": 0.5519, "step": 4650 }, { "epoch": 0.19496347840667344, "grad_norm": 1.5428576469421387, "learning_rate": 9.303236637983359e-06, "loss": 0.5146, "step": 4651 }, { "epoch": 0.1950053970216824, "grad_norm": 1.795722484588623, "learning_rate": 9.302890925708515e-06, "loss": 0.6057, "step": 4652 }, { "epoch": 0.19504731563669137, "grad_norm": 1.674910068511963, "learning_rate": 9.302545134115825e-06, "loss": 0.5642, "step": 4653 }, { "epoch": 0.19508923425170033, "grad_norm": 1.62185800075531, "learning_rate": 9.302199263211663e-06, "loss": 0.5158, "step": 4654 }, { "epoch": 0.1951311528667093, "grad_norm": 1.7935572862625122, "learning_rate": 9.301853313002406e-06, "loss": 0.6161, "step": 4655 }, { "epoch": 0.19517307148171825, "grad_norm": 1.6358981132507324, "learning_rate": 9.30150728349443e-06, "loss": 0.5652, "step": 4656 }, { "epoch": 0.19521499009672721, "grad_norm": 1.5680930614471436, "learning_rate": 9.301161174694113e-06, "loss": 0.5198, "step": 4657 }, { "epoch": 0.19525690871173618, "grad_norm": 1.7949366569519043, "learning_rate": 9.300814986607837e-06, "loss": 0.5288, "step": 4658 }, { "epoch": 0.1952988273267451, "grad_norm": 2.526865005493164, "learning_rate": 9.300468719241982e-06, "loss": 0.5544, "step": 4659 }, { "epoch": 0.19534074594175407, "grad_norm": 1.8586041927337646, "learning_rate": 9.30012237260293e-06, "loss": 0.4999, "step": 4660 }, { "epoch": 0.19538266455676304, "grad_norm": 1.8197013139724731, "learning_rate": 9.29977594669707e-06, "loss": 0.6196, "step": 4661 }, { "epoch": 0.195424583171772, "grad_norm": 1.59182608127594, "learning_rate": 9.299429441530781e-06, "loss": 0.553, "step": 4662 }, { "epoch": 0.19546650178678096, "grad_norm": 1.614269733428955, "learning_rate": 9.299082857110457e-06, "loss": 0.5496, "step": 4663 }, { "epoch": 0.19550842040178992, "grad_norm": 1.6183472871780396, "learning_rate": 9.298736193442482e-06, "loss": 0.551, "step": 4664 }, { "epoch": 0.19555033901679889, "grad_norm": 1.8003586530685425, "learning_rate": 9.298389450533249e-06, "loss": 0.5166, "step": 4665 }, { "epoch": 0.19559225763180785, "grad_norm": 1.9832384586334229, "learning_rate": 9.298042628389149e-06, "loss": 0.6095, "step": 4666 }, { "epoch": 0.1956341762468168, "grad_norm": 1.7742148637771606, "learning_rate": 9.297695727016573e-06, "loss": 0.5829, "step": 4667 }, { "epoch": 0.19567609486182577, "grad_norm": 1.7086023092269897, "learning_rate": 9.297348746421921e-06, "loss": 0.5937, "step": 4668 }, { "epoch": 0.19571801347683473, "grad_norm": 1.8308372497558594, "learning_rate": 9.297001686611584e-06, "loss": 0.5463, "step": 4669 }, { "epoch": 0.1957599320918437, "grad_norm": 1.7734973430633545, "learning_rate": 9.296654547591962e-06, "loss": 0.5309, "step": 4670 }, { "epoch": 0.19580185070685266, "grad_norm": 3.703270435333252, "learning_rate": 9.296307329369451e-06, "loss": 0.5691, "step": 4671 }, { "epoch": 0.1958437693218616, "grad_norm": 1.7404910326004028, "learning_rate": 9.295960031950456e-06, "loss": 0.5882, "step": 4672 }, { "epoch": 0.19588568793687056, "grad_norm": 1.7567895650863647, "learning_rate": 9.295612655341377e-06, "loss": 0.5656, "step": 4673 }, { "epoch": 0.19592760655187952, "grad_norm": 2.1939334869384766, "learning_rate": 9.295265199548616e-06, "loss": 0.5851, "step": 4674 }, { "epoch": 0.19596952516688848, "grad_norm": 1.7202204465866089, "learning_rate": 9.29491766457858e-06, "loss": 0.5708, "step": 4675 }, { "epoch": 0.19601144378189744, "grad_norm": 1.5894702672958374, "learning_rate": 9.294570050437675e-06, "loss": 0.5461, "step": 4676 }, { "epoch": 0.1960533623969064, "grad_norm": 1.633864402770996, "learning_rate": 9.294222357132308e-06, "loss": 0.5541, "step": 4677 }, { "epoch": 0.19609528101191537, "grad_norm": 1.7331790924072266, "learning_rate": 9.293874584668888e-06, "loss": 0.5297, "step": 4678 }, { "epoch": 0.19613719962692433, "grad_norm": 1.716538429260254, "learning_rate": 9.293526733053823e-06, "loss": 0.5355, "step": 4679 }, { "epoch": 0.1961791182419333, "grad_norm": 1.8365232944488525, "learning_rate": 9.293178802293532e-06, "loss": 0.5622, "step": 4680 }, { "epoch": 0.19622103685694225, "grad_norm": 1.7567198276519775, "learning_rate": 9.292830792394425e-06, "loss": 0.6038, "step": 4681 }, { "epoch": 0.19626295547195122, "grad_norm": 1.9578062295913696, "learning_rate": 9.292482703362916e-06, "loss": 0.591, "step": 4682 }, { "epoch": 0.19630487408696018, "grad_norm": 1.7168713808059692, "learning_rate": 9.292134535205422e-06, "loss": 0.6054, "step": 4683 }, { "epoch": 0.19634679270196911, "grad_norm": 1.9163429737091064, "learning_rate": 9.291786287928362e-06, "loss": 0.5837, "step": 4684 }, { "epoch": 0.19638871131697808, "grad_norm": 1.7157464027404785, "learning_rate": 9.291437961538156e-06, "loss": 0.5949, "step": 4685 }, { "epoch": 0.19643062993198704, "grad_norm": 1.5369154214859009, "learning_rate": 9.291089556041222e-06, "loss": 0.5316, "step": 4686 }, { "epoch": 0.196472548546996, "grad_norm": 1.7741889953613281, "learning_rate": 9.290741071443984e-06, "loss": 0.5324, "step": 4687 }, { "epoch": 0.19651446716200496, "grad_norm": 1.7244207859039307, "learning_rate": 9.290392507752868e-06, "loss": 0.5572, "step": 4688 }, { "epoch": 0.19655638577701393, "grad_norm": 2.9871480464935303, "learning_rate": 9.290043864974298e-06, "loss": 0.6285, "step": 4689 }, { "epoch": 0.1965983043920229, "grad_norm": 1.7810001373291016, "learning_rate": 9.289695143114698e-06, "loss": 0.5066, "step": 4690 }, { "epoch": 0.19664022300703185, "grad_norm": 2.519425392150879, "learning_rate": 9.289346342180501e-06, "loss": 0.557, "step": 4691 }, { "epoch": 0.1966821416220408, "grad_norm": 1.7624591588974, "learning_rate": 9.288997462178133e-06, "loss": 0.5649, "step": 4692 }, { "epoch": 0.19672406023704977, "grad_norm": 1.5558158159255981, "learning_rate": 9.288648503114024e-06, "loss": 0.5385, "step": 4693 }, { "epoch": 0.19676597885205874, "grad_norm": 1.6701842546463013, "learning_rate": 9.288299464994611e-06, "loss": 0.555, "step": 4694 }, { "epoch": 0.1968078974670677, "grad_norm": 1.6251801252365112, "learning_rate": 9.287950347826324e-06, "loss": 0.5225, "step": 4695 }, { "epoch": 0.19684981608207666, "grad_norm": 1.462197184562683, "learning_rate": 9.287601151615601e-06, "loss": 0.5898, "step": 4696 }, { "epoch": 0.1968917346970856, "grad_norm": 1.6523977518081665, "learning_rate": 9.287251876368879e-06, "loss": 0.563, "step": 4697 }, { "epoch": 0.19693365331209456, "grad_norm": 1.6807684898376465, "learning_rate": 9.286902522092596e-06, "loss": 0.5815, "step": 4698 }, { "epoch": 0.19697557192710352, "grad_norm": 1.598879098892212, "learning_rate": 9.286553088793192e-06, "loss": 0.5981, "step": 4699 }, { "epoch": 0.19701749054211248, "grad_norm": 1.7851155996322632, "learning_rate": 9.286203576477104e-06, "loss": 0.513, "step": 4700 }, { "epoch": 0.19705940915712145, "grad_norm": 2.020566940307617, "learning_rate": 9.285853985150782e-06, "loss": 0.5942, "step": 4701 }, { "epoch": 0.1971013277721304, "grad_norm": 1.7282123565673828, "learning_rate": 9.285504314820666e-06, "loss": 0.5516, "step": 4702 }, { "epoch": 0.19714324638713937, "grad_norm": 1.5386440753936768, "learning_rate": 9.285154565493202e-06, "loss": 0.5527, "step": 4703 }, { "epoch": 0.19718516500214833, "grad_norm": 1.9384576082229614, "learning_rate": 9.284804737174838e-06, "loss": 0.5838, "step": 4704 }, { "epoch": 0.1972270836171573, "grad_norm": 2.421567916870117, "learning_rate": 9.284454829872022e-06, "loss": 0.5781, "step": 4705 }, { "epoch": 0.19726900223216626, "grad_norm": 1.8214528560638428, "learning_rate": 9.284104843591204e-06, "loss": 0.5366, "step": 4706 }, { "epoch": 0.19731092084717522, "grad_norm": 1.7746386528015137, "learning_rate": 9.283754778338836e-06, "loss": 0.5741, "step": 4707 }, { "epoch": 0.19735283946218418, "grad_norm": 1.7666665315628052, "learning_rate": 9.283404634121371e-06, "loss": 0.5262, "step": 4708 }, { "epoch": 0.19739475807719312, "grad_norm": 1.569530725479126, "learning_rate": 9.283054410945263e-06, "loss": 0.5154, "step": 4709 }, { "epoch": 0.19743667669220208, "grad_norm": 1.7165998220443726, "learning_rate": 9.282704108816969e-06, "loss": 0.5193, "step": 4710 }, { "epoch": 0.19747859530721104, "grad_norm": 1.7800920009613037, "learning_rate": 9.282353727742943e-06, "loss": 0.5773, "step": 4711 }, { "epoch": 0.19752051392222, "grad_norm": 1.9744951725006104, "learning_rate": 9.282003267729649e-06, "loss": 0.5297, "step": 4712 }, { "epoch": 0.19756243253722897, "grad_norm": 1.5837265253067017, "learning_rate": 9.281652728783543e-06, "loss": 0.5391, "step": 4713 }, { "epoch": 0.19760435115223793, "grad_norm": 1.996087670326233, "learning_rate": 9.281302110911088e-06, "loss": 0.6159, "step": 4714 }, { "epoch": 0.1976462697672469, "grad_norm": 3.224553108215332, "learning_rate": 9.280951414118749e-06, "loss": 0.5299, "step": 4715 }, { "epoch": 0.19768818838225585, "grad_norm": 2.0180983543395996, "learning_rate": 9.280600638412987e-06, "loss": 0.6148, "step": 4716 }, { "epoch": 0.19773010699726482, "grad_norm": 1.9112721681594849, "learning_rate": 9.28024978380027e-06, "loss": 0.5335, "step": 4717 }, { "epoch": 0.19777202561227378, "grad_norm": 1.8947862386703491, "learning_rate": 9.279898850287067e-06, "loss": 0.58, "step": 4718 }, { "epoch": 0.19781394422728274, "grad_norm": 1.667252540588379, "learning_rate": 9.279547837879845e-06, "loss": 0.524, "step": 4719 }, { "epoch": 0.1978558628422917, "grad_norm": 1.7953704595565796, "learning_rate": 9.279196746585073e-06, "loss": 0.5491, "step": 4720 }, { "epoch": 0.19789778145730066, "grad_norm": 1.7299113273620605, "learning_rate": 9.278845576409227e-06, "loss": 0.5352, "step": 4721 }, { "epoch": 0.1979397000723096, "grad_norm": 1.716817855834961, "learning_rate": 9.278494327358777e-06, "loss": 0.5781, "step": 4722 }, { "epoch": 0.19798161868731856, "grad_norm": 1.775706171989441, "learning_rate": 9.2781429994402e-06, "loss": 0.4782, "step": 4723 }, { "epoch": 0.19802353730232752, "grad_norm": 1.9616236686706543, "learning_rate": 9.277791592659971e-06, "loss": 0.5849, "step": 4724 }, { "epoch": 0.19806545591733649, "grad_norm": 1.8742077350616455, "learning_rate": 9.277440107024568e-06, "loss": 0.5183, "step": 4725 }, { "epoch": 0.19810737453234545, "grad_norm": 1.580827236175537, "learning_rate": 9.27708854254047e-06, "loss": 0.5672, "step": 4726 }, { "epoch": 0.1981492931473544, "grad_norm": 1.6790275573730469, "learning_rate": 9.276736899214157e-06, "loss": 0.5594, "step": 4727 }, { "epoch": 0.19819121176236337, "grad_norm": 1.7602473497390747, "learning_rate": 9.27638517705211e-06, "loss": 0.5144, "step": 4728 }, { "epoch": 0.19823313037737234, "grad_norm": 1.777467966079712, "learning_rate": 9.276033376060815e-06, "loss": 0.5717, "step": 4729 }, { "epoch": 0.1982750489923813, "grad_norm": 1.8316650390625, "learning_rate": 9.275681496246757e-06, "loss": 0.5991, "step": 4730 }, { "epoch": 0.19831696760739026, "grad_norm": 1.719696283340454, "learning_rate": 9.275329537616423e-06, "loss": 0.5439, "step": 4731 }, { "epoch": 0.19835888622239922, "grad_norm": 1.8207563161849976, "learning_rate": 9.274977500176298e-06, "loss": 0.5093, "step": 4732 }, { "epoch": 0.19840080483740818, "grad_norm": 1.9969569444656372, "learning_rate": 9.274625383932873e-06, "loss": 0.5933, "step": 4733 }, { "epoch": 0.19844272345241712, "grad_norm": 1.9222573041915894, "learning_rate": 9.274273188892638e-06, "loss": 0.5572, "step": 4734 }, { "epoch": 0.19848464206742608, "grad_norm": 2.0337445735931396, "learning_rate": 9.273920915062088e-06, "loss": 0.5138, "step": 4735 }, { "epoch": 0.19852656068243504, "grad_norm": 1.8621078729629517, "learning_rate": 9.273568562447712e-06, "loss": 0.538, "step": 4736 }, { "epoch": 0.198568479297444, "grad_norm": 1.67344331741333, "learning_rate": 9.273216131056007e-06, "loss": 0.5292, "step": 4737 }, { "epoch": 0.19861039791245297, "grad_norm": 2.0510013103485107, "learning_rate": 9.272863620893471e-06, "loss": 0.5038, "step": 4738 }, { "epoch": 0.19865231652746193, "grad_norm": 1.665215015411377, "learning_rate": 9.272511031966603e-06, "loss": 0.5717, "step": 4739 }, { "epoch": 0.1986942351424709, "grad_norm": 2.259758710861206, "learning_rate": 9.272158364281899e-06, "loss": 0.5674, "step": 4740 }, { "epoch": 0.19873615375747986, "grad_norm": 1.7274765968322754, "learning_rate": 9.271805617845861e-06, "loss": 0.5366, "step": 4741 }, { "epoch": 0.19877807237248882, "grad_norm": 1.42108154296875, "learning_rate": 9.271452792664992e-06, "loss": 0.5206, "step": 4742 }, { "epoch": 0.19881999098749778, "grad_norm": 1.6687172651290894, "learning_rate": 9.271099888745799e-06, "loss": 0.5108, "step": 4743 }, { "epoch": 0.19886190960250674, "grad_norm": 1.6938939094543457, "learning_rate": 9.270746906094781e-06, "loss": 0.5695, "step": 4744 }, { "epoch": 0.1989038282175157, "grad_norm": 1.8808270692825317, "learning_rate": 9.270393844718448e-06, "loss": 0.5733, "step": 4745 }, { "epoch": 0.19894574683252467, "grad_norm": 1.5518288612365723, "learning_rate": 9.270040704623309e-06, "loss": 0.5654, "step": 4746 }, { "epoch": 0.1989876654475336, "grad_norm": 1.4742076396942139, "learning_rate": 9.269687485815872e-06, "loss": 0.5176, "step": 4747 }, { "epoch": 0.19902958406254256, "grad_norm": 2.1550116539001465, "learning_rate": 9.26933418830265e-06, "loss": 0.4967, "step": 4748 }, { "epoch": 0.19907150267755153, "grad_norm": 1.6419118642807007, "learning_rate": 9.268980812090154e-06, "loss": 0.5918, "step": 4749 }, { "epoch": 0.1991134212925605, "grad_norm": 1.6231575012207031, "learning_rate": 9.268627357184899e-06, "loss": 0.5355, "step": 4750 }, { "epoch": 0.19915533990756945, "grad_norm": 1.6154955625534058, "learning_rate": 9.2682738235934e-06, "loss": 0.5113, "step": 4751 }, { "epoch": 0.1991972585225784, "grad_norm": 1.609853982925415, "learning_rate": 9.267920211322172e-06, "loss": 0.5784, "step": 4752 }, { "epoch": 0.19923917713758738, "grad_norm": 1.5648508071899414, "learning_rate": 9.267566520377737e-06, "loss": 0.5903, "step": 4753 }, { "epoch": 0.19928109575259634, "grad_norm": 1.69976806640625, "learning_rate": 9.267212750766612e-06, "loss": 0.5699, "step": 4754 }, { "epoch": 0.1993230143676053, "grad_norm": 1.9319576025009155, "learning_rate": 9.266858902495319e-06, "loss": 0.595, "step": 4755 }, { "epoch": 0.19936493298261426, "grad_norm": 1.5945444107055664, "learning_rate": 9.266504975570381e-06, "loss": 0.5475, "step": 4756 }, { "epoch": 0.19940685159762322, "grad_norm": 1.7201372385025024, "learning_rate": 9.266150969998322e-06, "loss": 0.5865, "step": 4757 }, { "epoch": 0.1994487702126322, "grad_norm": 1.6753246784210205, "learning_rate": 9.265796885785668e-06, "loss": 0.5538, "step": 4758 }, { "epoch": 0.19949068882764112, "grad_norm": 1.8326438665390015, "learning_rate": 9.265442722938947e-06, "loss": 0.5604, "step": 4759 }, { "epoch": 0.19953260744265008, "grad_norm": 1.8158369064331055, "learning_rate": 9.265088481464684e-06, "loss": 0.5272, "step": 4760 }, { "epoch": 0.19957452605765905, "grad_norm": 1.5643633604049683, "learning_rate": 9.264734161369412e-06, "loss": 0.5212, "step": 4761 }, { "epoch": 0.199616444672668, "grad_norm": 1.627730369567871, "learning_rate": 9.264379762659661e-06, "loss": 0.5777, "step": 4762 }, { "epoch": 0.19965836328767697, "grad_norm": 1.7597103118896484, "learning_rate": 9.264025285341964e-06, "loss": 0.5897, "step": 4763 }, { "epoch": 0.19970028190268593, "grad_norm": 1.6288199424743652, "learning_rate": 9.263670729422858e-06, "loss": 0.5314, "step": 4764 }, { "epoch": 0.1997422005176949, "grad_norm": 1.7549561262130737, "learning_rate": 9.263316094908875e-06, "loss": 0.5305, "step": 4765 }, { "epoch": 0.19978411913270386, "grad_norm": 1.6445212364196777, "learning_rate": 9.262961381806553e-06, "loss": 0.5355, "step": 4766 }, { "epoch": 0.19982603774771282, "grad_norm": 1.671890139579773, "learning_rate": 9.262606590122431e-06, "loss": 0.6273, "step": 4767 }, { "epoch": 0.19986795636272178, "grad_norm": 1.8449914455413818, "learning_rate": 9.262251719863051e-06, "loss": 0.5845, "step": 4768 }, { "epoch": 0.19990987497773074, "grad_norm": 1.612329125404358, "learning_rate": 9.261896771034953e-06, "loss": 0.5306, "step": 4769 }, { "epoch": 0.1999517935927397, "grad_norm": 1.4603134393692017, "learning_rate": 9.26154174364468e-06, "loss": 0.5525, "step": 4770 }, { "epoch": 0.19999371220774867, "grad_norm": 4.268134593963623, "learning_rate": 9.261186637698775e-06, "loss": 0.6315, "step": 4771 }, { "epoch": 0.2000356308227576, "grad_norm": 1.7564185857772827, "learning_rate": 9.260831453203785e-06, "loss": 0.6048, "step": 4772 }, { "epoch": 0.20007754943776657, "grad_norm": 1.5636420249938965, "learning_rate": 9.260476190166259e-06, "loss": 0.4943, "step": 4773 }, { "epoch": 0.20011946805277553, "grad_norm": 1.8122166395187378, "learning_rate": 9.260120848592744e-06, "loss": 0.5932, "step": 4774 }, { "epoch": 0.2001613866677845, "grad_norm": 1.760802149772644, "learning_rate": 9.25976542848979e-06, "loss": 0.5125, "step": 4775 }, { "epoch": 0.20020330528279345, "grad_norm": 1.9641119241714478, "learning_rate": 9.25940992986395e-06, "loss": 0.5601, "step": 4776 }, { "epoch": 0.20024522389780242, "grad_norm": 1.657183289527893, "learning_rate": 9.259054352721775e-06, "loss": 0.5439, "step": 4777 }, { "epoch": 0.20028714251281138, "grad_norm": 1.5056099891662598, "learning_rate": 9.258698697069824e-06, "loss": 0.5175, "step": 4778 }, { "epoch": 0.20032906112782034, "grad_norm": 1.771931767463684, "learning_rate": 9.258342962914646e-06, "loss": 0.5357, "step": 4779 }, { "epoch": 0.2003709797428293, "grad_norm": 1.9487190246582031, "learning_rate": 9.257987150262806e-06, "loss": 0.546, "step": 4780 }, { "epoch": 0.20041289835783827, "grad_norm": 1.7248330116271973, "learning_rate": 9.257631259120857e-06, "loss": 0.523, "step": 4781 }, { "epoch": 0.20045481697284723, "grad_norm": 1.5632492303848267, "learning_rate": 9.257275289495362e-06, "loss": 0.5309, "step": 4782 }, { "epoch": 0.2004967355878562, "grad_norm": 1.7176116704940796, "learning_rate": 9.256919241392884e-06, "loss": 0.5374, "step": 4783 }, { "epoch": 0.20053865420286512, "grad_norm": 1.675964593887329, "learning_rate": 9.256563114819984e-06, "loss": 0.5609, "step": 4784 }, { "epoch": 0.2005805728178741, "grad_norm": 1.6542292833328247, "learning_rate": 9.256206909783228e-06, "loss": 0.5632, "step": 4785 }, { "epoch": 0.20062249143288305, "grad_norm": 1.778483510017395, "learning_rate": 9.25585062628918e-06, "loss": 0.5602, "step": 4786 }, { "epoch": 0.200664410047892, "grad_norm": 1.6098171472549438, "learning_rate": 9.25549426434441e-06, "loss": 0.5582, "step": 4787 }, { "epoch": 0.20070632866290097, "grad_norm": 1.7419898509979248, "learning_rate": 9.255137823955487e-06, "loss": 0.5922, "step": 4788 }, { "epoch": 0.20074824727790994, "grad_norm": 1.7363052368164062, "learning_rate": 9.254781305128982e-06, "loss": 0.559, "step": 4789 }, { "epoch": 0.2007901658929189, "grad_norm": 1.8426406383514404, "learning_rate": 9.254424707871463e-06, "loss": 0.5893, "step": 4790 }, { "epoch": 0.20083208450792786, "grad_norm": 1.477474570274353, "learning_rate": 9.254068032189508e-06, "loss": 0.5606, "step": 4791 }, { "epoch": 0.20087400312293682, "grad_norm": 1.7223949432373047, "learning_rate": 9.253711278089688e-06, "loss": 0.5479, "step": 4792 }, { "epoch": 0.20091592173794579, "grad_norm": 1.613073468208313, "learning_rate": 9.253354445578581e-06, "loss": 0.5167, "step": 4793 }, { "epoch": 0.20095784035295475, "grad_norm": 1.7406328916549683, "learning_rate": 9.252997534662768e-06, "loss": 0.5484, "step": 4794 }, { "epoch": 0.2009997589679637, "grad_norm": 1.6713496446609497, "learning_rate": 9.252640545348824e-06, "loss": 0.5536, "step": 4795 }, { "epoch": 0.20104167758297267, "grad_norm": 1.616862416267395, "learning_rate": 9.252283477643329e-06, "loss": 0.5502, "step": 4796 }, { "epoch": 0.2010835961979816, "grad_norm": 2.0908732414245605, "learning_rate": 9.251926331552868e-06, "loss": 0.5474, "step": 4797 }, { "epoch": 0.20112551481299057, "grad_norm": 1.6036204099655151, "learning_rate": 9.251569107084023e-06, "loss": 0.5203, "step": 4798 }, { "epoch": 0.20116743342799953, "grad_norm": 1.6976176500320435, "learning_rate": 9.251211804243378e-06, "loss": 0.6066, "step": 4799 }, { "epoch": 0.2012093520430085, "grad_norm": 1.6116447448730469, "learning_rate": 9.250854423037521e-06, "loss": 0.4971, "step": 4800 }, { "epoch": 0.20125127065801746, "grad_norm": 1.6602485179901123, "learning_rate": 9.25049696347304e-06, "loss": 0.4636, "step": 4801 }, { "epoch": 0.20129318927302642, "grad_norm": 1.7975847721099854, "learning_rate": 9.250139425556523e-06, "loss": 0.5378, "step": 4802 }, { "epoch": 0.20133510788803538, "grad_norm": 1.6196670532226562, "learning_rate": 9.249781809294562e-06, "loss": 0.4975, "step": 4803 }, { "epoch": 0.20137702650304434, "grad_norm": 2.20202898979187, "learning_rate": 9.249424114693749e-06, "loss": 0.556, "step": 4804 }, { "epoch": 0.2014189451180533, "grad_norm": 1.6969560384750366, "learning_rate": 9.249066341760675e-06, "loss": 0.5597, "step": 4805 }, { "epoch": 0.20146086373306227, "grad_norm": 1.6628860235214233, "learning_rate": 9.248708490501938e-06, "loss": 0.5411, "step": 4806 }, { "epoch": 0.20150278234807123, "grad_norm": 1.6275768280029297, "learning_rate": 9.248350560924135e-06, "loss": 0.5632, "step": 4807 }, { "epoch": 0.2015447009630802, "grad_norm": 1.8164656162261963, "learning_rate": 9.247992553033859e-06, "loss": 0.5857, "step": 4808 }, { "epoch": 0.20158661957808913, "grad_norm": 2.246184825897217, "learning_rate": 9.247634466837717e-06, "loss": 0.5866, "step": 4809 }, { "epoch": 0.2016285381930981, "grad_norm": 1.7865544557571411, "learning_rate": 9.247276302342305e-06, "loss": 0.5296, "step": 4810 }, { "epoch": 0.20167045680810705, "grad_norm": 1.6583935022354126, "learning_rate": 9.246918059554225e-06, "loss": 0.5018, "step": 4811 }, { "epoch": 0.201712375423116, "grad_norm": 1.6405644416809082, "learning_rate": 9.246559738480081e-06, "loss": 0.59, "step": 4812 }, { "epoch": 0.20175429403812498, "grad_norm": 2.083516836166382, "learning_rate": 9.246201339126482e-06, "loss": 0.5331, "step": 4813 }, { "epoch": 0.20179621265313394, "grad_norm": 1.7880092859268188, "learning_rate": 9.245842861500028e-06, "loss": 0.5934, "step": 4814 }, { "epoch": 0.2018381312681429, "grad_norm": 1.5669268369674683, "learning_rate": 9.245484305607334e-06, "loss": 0.5154, "step": 4815 }, { "epoch": 0.20188004988315186, "grad_norm": 2.054051637649536, "learning_rate": 9.245125671455002e-06, "loss": 0.5603, "step": 4816 }, { "epoch": 0.20192196849816083, "grad_norm": 1.6363675594329834, "learning_rate": 9.24476695904965e-06, "loss": 0.5376, "step": 4817 }, { "epoch": 0.2019638871131698, "grad_norm": 1.758434534072876, "learning_rate": 9.244408168397886e-06, "loss": 0.5692, "step": 4818 }, { "epoch": 0.20200580572817875, "grad_norm": 1.6572790145874023, "learning_rate": 9.244049299506326e-06, "loss": 0.5649, "step": 4819 }, { "epoch": 0.2020477243431877, "grad_norm": 1.5893120765686035, "learning_rate": 9.243690352381583e-06, "loss": 0.5539, "step": 4820 }, { "epoch": 0.20208964295819667, "grad_norm": 2.003028392791748, "learning_rate": 9.243331327030276e-06, "loss": 0.5604, "step": 4821 }, { "epoch": 0.2021315615732056, "grad_norm": 1.6168354749679565, "learning_rate": 9.242972223459024e-06, "loss": 0.5411, "step": 4822 }, { "epoch": 0.20217348018821457, "grad_norm": 1.9394086599349976, "learning_rate": 9.242613041674442e-06, "loss": 0.5337, "step": 4823 }, { "epoch": 0.20221539880322353, "grad_norm": 1.714266300201416, "learning_rate": 9.242253781683155e-06, "loss": 0.5507, "step": 4824 }, { "epoch": 0.2022573174182325, "grad_norm": 1.922292947769165, "learning_rate": 9.241894443491783e-06, "loss": 0.5142, "step": 4825 }, { "epoch": 0.20229923603324146, "grad_norm": 1.8817002773284912, "learning_rate": 9.241535027106952e-06, "loss": 0.5058, "step": 4826 }, { "epoch": 0.20234115464825042, "grad_norm": 1.7352262735366821, "learning_rate": 9.241175532535288e-06, "loss": 0.5685, "step": 4827 }, { "epoch": 0.20238307326325938, "grad_norm": 1.55848228931427, "learning_rate": 9.240815959783415e-06, "loss": 0.508, "step": 4828 }, { "epoch": 0.20242499187826835, "grad_norm": 1.7445564270019531, "learning_rate": 9.240456308857962e-06, "loss": 0.6002, "step": 4829 }, { "epoch": 0.2024669104932773, "grad_norm": 2.0090765953063965, "learning_rate": 9.24009657976556e-06, "loss": 0.5826, "step": 4830 }, { "epoch": 0.20250882910828627, "grad_norm": 1.8160663843154907, "learning_rate": 9.23973677251284e-06, "loss": 0.6061, "step": 4831 }, { "epoch": 0.20255074772329523, "grad_norm": 1.9073326587677002, "learning_rate": 9.239376887106433e-06, "loss": 0.5378, "step": 4832 }, { "epoch": 0.2025926663383042, "grad_norm": 1.6568000316619873, "learning_rate": 9.239016923552973e-06, "loss": 0.5874, "step": 4833 }, { "epoch": 0.20263458495331313, "grad_norm": 2.303454875946045, "learning_rate": 9.238656881859096e-06, "loss": 0.547, "step": 4834 }, { "epoch": 0.2026765035683221, "grad_norm": 1.755507469177246, "learning_rate": 9.238296762031441e-06, "loss": 0.5951, "step": 4835 }, { "epoch": 0.20271842218333105, "grad_norm": 1.8056526184082031, "learning_rate": 9.237936564076643e-06, "loss": 0.5292, "step": 4836 }, { "epoch": 0.20276034079834002, "grad_norm": 1.5597692728042603, "learning_rate": 9.237576288001345e-06, "loss": 0.5125, "step": 4837 }, { "epoch": 0.20280225941334898, "grad_norm": 1.8607122898101807, "learning_rate": 9.237215933812186e-06, "loss": 0.5467, "step": 4838 }, { "epoch": 0.20284417802835794, "grad_norm": 1.811249852180481, "learning_rate": 9.236855501515807e-06, "loss": 0.6006, "step": 4839 }, { "epoch": 0.2028860966433669, "grad_norm": 1.8802707195281982, "learning_rate": 9.236494991118856e-06, "loss": 0.552, "step": 4840 }, { "epoch": 0.20292801525837587, "grad_norm": 2.1666693687438965, "learning_rate": 9.236134402627976e-06, "loss": 0.5137, "step": 4841 }, { "epoch": 0.20296993387338483, "grad_norm": 1.703199028968811, "learning_rate": 9.235773736049815e-06, "loss": 0.5501, "step": 4842 }, { "epoch": 0.2030118524883938, "grad_norm": 1.9045008420944214, "learning_rate": 9.235412991391022e-06, "loss": 0.5386, "step": 4843 }, { "epoch": 0.20305377110340275, "grad_norm": 1.8252557516098022, "learning_rate": 9.235052168658242e-06, "loss": 0.5244, "step": 4844 }, { "epoch": 0.20309568971841171, "grad_norm": 1.933729648590088, "learning_rate": 9.234691267858134e-06, "loss": 0.5601, "step": 4845 }, { "epoch": 0.20313760833342068, "grad_norm": 1.8387501239776611, "learning_rate": 9.234330288997344e-06, "loss": 0.6218, "step": 4846 }, { "epoch": 0.2031795269484296, "grad_norm": 1.5680128335952759, "learning_rate": 9.233969232082532e-06, "loss": 0.5613, "step": 4847 }, { "epoch": 0.20322144556343857, "grad_norm": 1.5892667770385742, "learning_rate": 9.233608097120348e-06, "loss": 0.5393, "step": 4848 }, { "epoch": 0.20326336417844754, "grad_norm": 1.6090614795684814, "learning_rate": 9.233246884117453e-06, "loss": 0.5404, "step": 4849 }, { "epoch": 0.2033052827934565, "grad_norm": 1.5658637285232544, "learning_rate": 9.232885593080503e-06, "loss": 0.5265, "step": 4850 }, { "epoch": 0.20334720140846546, "grad_norm": 2.018179416656494, "learning_rate": 9.232524224016158e-06, "loss": 0.5462, "step": 4851 }, { "epoch": 0.20338912002347442, "grad_norm": 2.165008306503296, "learning_rate": 9.232162776931081e-06, "loss": 0.6148, "step": 4852 }, { "epoch": 0.20343103863848339, "grad_norm": 1.4457721710205078, "learning_rate": 9.231801251831936e-06, "loss": 0.5601, "step": 4853 }, { "epoch": 0.20347295725349235, "grad_norm": 1.991906762123108, "learning_rate": 9.231439648725381e-06, "loss": 0.5593, "step": 4854 }, { "epoch": 0.2035148758685013, "grad_norm": 1.566906452178955, "learning_rate": 9.23107796761809e-06, "loss": 0.5624, "step": 4855 }, { "epoch": 0.20355679448351027, "grad_norm": 1.6222728490829468, "learning_rate": 9.230716208516723e-06, "loss": 0.5656, "step": 4856 }, { "epoch": 0.20359871309851924, "grad_norm": 1.8103183507919312, "learning_rate": 9.230354371427954e-06, "loss": 0.544, "step": 4857 }, { "epoch": 0.2036406317135282, "grad_norm": 1.804332971572876, "learning_rate": 9.22999245635845e-06, "loss": 0.5319, "step": 4858 }, { "epoch": 0.20368255032853713, "grad_norm": 1.7311357259750366, "learning_rate": 9.22963046331488e-06, "loss": 0.538, "step": 4859 }, { "epoch": 0.2037244689435461, "grad_norm": 1.5036927461624146, "learning_rate": 9.229268392303923e-06, "loss": 0.4628, "step": 4860 }, { "epoch": 0.20376638755855506, "grad_norm": 1.9157493114471436, "learning_rate": 9.228906243332248e-06, "loss": 0.5816, "step": 4861 }, { "epoch": 0.20380830617356402, "grad_norm": 1.8644160032272339, "learning_rate": 9.228544016406534e-06, "loss": 0.5726, "step": 4862 }, { "epoch": 0.20385022478857298, "grad_norm": 1.5677751302719116, "learning_rate": 9.228181711533457e-06, "loss": 0.5175, "step": 4863 }, { "epoch": 0.20389214340358194, "grad_norm": 1.685184121131897, "learning_rate": 9.227819328719694e-06, "loss": 0.605, "step": 4864 }, { "epoch": 0.2039340620185909, "grad_norm": 1.9386314153671265, "learning_rate": 9.227456867971926e-06, "loss": 0.5806, "step": 4865 }, { "epoch": 0.20397598063359987, "grad_norm": 1.9372056722640991, "learning_rate": 9.227094329296837e-06, "loss": 0.5712, "step": 4866 }, { "epoch": 0.20401789924860883, "grad_norm": 1.8694180250167847, "learning_rate": 9.226731712701107e-06, "loss": 0.5254, "step": 4867 }, { "epoch": 0.2040598178636178, "grad_norm": 1.7110141515731812, "learning_rate": 9.22636901819142e-06, "loss": 0.5031, "step": 4868 }, { "epoch": 0.20410173647862676, "grad_norm": 1.663341760635376, "learning_rate": 9.226006245774466e-06, "loss": 0.5313, "step": 4869 }, { "epoch": 0.20414365509363572, "grad_norm": 1.808633804321289, "learning_rate": 9.225643395456926e-06, "loss": 0.5901, "step": 4870 }, { "epoch": 0.20418557370864468, "grad_norm": 1.855664849281311, "learning_rate": 9.225280467245493e-06, "loss": 0.6149, "step": 4871 }, { "epoch": 0.20422749232365361, "grad_norm": 1.588625431060791, "learning_rate": 9.224917461146856e-06, "loss": 0.5279, "step": 4872 }, { "epoch": 0.20426941093866258, "grad_norm": 1.9440184831619263, "learning_rate": 9.224554377167705e-06, "loss": 0.6074, "step": 4873 }, { "epoch": 0.20431132955367154, "grad_norm": 1.8804494142532349, "learning_rate": 9.224191215314738e-06, "loss": 0.5342, "step": 4874 }, { "epoch": 0.2043532481686805, "grad_norm": 1.6536965370178223, "learning_rate": 9.223827975594641e-06, "loss": 0.5888, "step": 4875 }, { "epoch": 0.20439516678368946, "grad_norm": 1.526943564414978, "learning_rate": 9.223464658014119e-06, "loss": 0.613, "step": 4876 }, { "epoch": 0.20443708539869843, "grad_norm": 1.7346528768539429, "learning_rate": 9.22310126257986e-06, "loss": 0.5596, "step": 4877 }, { "epoch": 0.2044790040137074, "grad_norm": 1.7029650211334229, "learning_rate": 9.22273778929857e-06, "loss": 0.5497, "step": 4878 }, { "epoch": 0.20452092262871635, "grad_norm": 1.882148265838623, "learning_rate": 9.222374238176948e-06, "loss": 0.5331, "step": 4879 }, { "epoch": 0.2045628412437253, "grad_norm": 2.2792305946350098, "learning_rate": 9.222010609221693e-06, "loss": 0.5277, "step": 4880 }, { "epoch": 0.20460475985873428, "grad_norm": 3.248774290084839, "learning_rate": 9.221646902439508e-06, "loss": 0.5566, "step": 4881 }, { "epoch": 0.20464667847374324, "grad_norm": 2.3377063274383545, "learning_rate": 9.2212831178371e-06, "loss": 0.5851, "step": 4882 }, { "epoch": 0.2046885970887522, "grad_norm": 2.064182758331299, "learning_rate": 9.220919255421173e-06, "loss": 0.5456, "step": 4883 }, { "epoch": 0.20473051570376113, "grad_norm": 1.7163395881652832, "learning_rate": 9.220555315198435e-06, "loss": 0.6156, "step": 4884 }, { "epoch": 0.2047724343187701, "grad_norm": 1.5644128322601318, "learning_rate": 9.220191297175595e-06, "loss": 0.5031, "step": 4885 }, { "epoch": 0.20481435293377906, "grad_norm": 1.7812517881393433, "learning_rate": 9.219827201359362e-06, "loss": 0.5223, "step": 4886 }, { "epoch": 0.20485627154878802, "grad_norm": 1.792828917503357, "learning_rate": 9.219463027756447e-06, "loss": 0.5394, "step": 4887 }, { "epoch": 0.20489819016379698, "grad_norm": 1.658374309539795, "learning_rate": 9.219098776373566e-06, "loss": 0.5415, "step": 4888 }, { "epoch": 0.20494010877880595, "grad_norm": 1.6101043224334717, "learning_rate": 9.218734447217433e-06, "loss": 0.5308, "step": 4889 }, { "epoch": 0.2049820273938149, "grad_norm": 1.7001346349716187, "learning_rate": 9.21837004029476e-06, "loss": 0.4985, "step": 4890 }, { "epoch": 0.20502394600882387, "grad_norm": 3.357409715652466, "learning_rate": 9.218005555612269e-06, "loss": 0.6102, "step": 4891 }, { "epoch": 0.20506586462383283, "grad_norm": 1.8591028451919556, "learning_rate": 9.217640993176675e-06, "loss": 0.5798, "step": 4892 }, { "epoch": 0.2051077832388418, "grad_norm": 2.0738391876220703, "learning_rate": 9.217276352994701e-06, "loss": 0.5764, "step": 4893 }, { "epoch": 0.20514970185385076, "grad_norm": 3.6695313453674316, "learning_rate": 9.216911635073067e-06, "loss": 0.4885, "step": 4894 }, { "epoch": 0.20519162046885972, "grad_norm": 1.7412753105163574, "learning_rate": 9.216546839418497e-06, "loss": 0.5346, "step": 4895 }, { "epoch": 0.20523353908386868, "grad_norm": 1.678906798362732, "learning_rate": 9.216181966037717e-06, "loss": 0.5915, "step": 4896 }, { "epoch": 0.20527545769887762, "grad_norm": 2.0241036415100098, "learning_rate": 9.215817014937449e-06, "loss": 0.6175, "step": 4897 }, { "epoch": 0.20531737631388658, "grad_norm": 1.732633113861084, "learning_rate": 9.215451986124423e-06, "loss": 0.5676, "step": 4898 }, { "epoch": 0.20535929492889554, "grad_norm": 1.9824124574661255, "learning_rate": 9.215086879605366e-06, "loss": 0.5861, "step": 4899 }, { "epoch": 0.2054012135439045, "grad_norm": 1.7648098468780518, "learning_rate": 9.214721695387012e-06, "loss": 0.5057, "step": 4900 }, { "epoch": 0.20544313215891347, "grad_norm": 1.9374334812164307, "learning_rate": 9.214356433476091e-06, "loss": 0.5247, "step": 4901 }, { "epoch": 0.20548505077392243, "grad_norm": 1.83405339717865, "learning_rate": 9.213991093879334e-06, "loss": 0.5449, "step": 4902 }, { "epoch": 0.2055269693889314, "grad_norm": 1.4963359832763672, "learning_rate": 9.213625676603475e-06, "loss": 0.5799, "step": 4903 }, { "epoch": 0.20556888800394035, "grad_norm": 1.6912009716033936, "learning_rate": 9.213260181655253e-06, "loss": 0.526, "step": 4904 }, { "epoch": 0.20561080661894932, "grad_norm": 1.8832114934921265, "learning_rate": 9.212894609041405e-06, "loss": 0.5315, "step": 4905 }, { "epoch": 0.20565272523395828, "grad_norm": 1.5603489875793457, "learning_rate": 9.212528958768667e-06, "loss": 0.5082, "step": 4906 }, { "epoch": 0.20569464384896724, "grad_norm": 2.8499691486358643, "learning_rate": 9.212163230843785e-06, "loss": 0.5419, "step": 4907 }, { "epoch": 0.2057365624639762, "grad_norm": 1.8368451595306396, "learning_rate": 9.211797425273493e-06, "loss": 0.5737, "step": 4908 }, { "epoch": 0.20577848107898514, "grad_norm": 1.644034743309021, "learning_rate": 9.211431542064541e-06, "loss": 0.5164, "step": 4909 }, { "epoch": 0.2058203996939941, "grad_norm": 2.261192560195923, "learning_rate": 9.211065581223668e-06, "loss": 0.4983, "step": 4910 }, { "epoch": 0.20586231830900306, "grad_norm": 1.8094056844711304, "learning_rate": 9.210699542757625e-06, "loss": 0.5557, "step": 4911 }, { "epoch": 0.20590423692401202, "grad_norm": 1.7475719451904297, "learning_rate": 9.210333426673156e-06, "loss": 0.522, "step": 4912 }, { "epoch": 0.205946155539021, "grad_norm": 1.7588366270065308, "learning_rate": 9.209967232977011e-06, "loss": 0.5252, "step": 4913 }, { "epoch": 0.20598807415402995, "grad_norm": 1.9265762567520142, "learning_rate": 9.20960096167594e-06, "loss": 0.5585, "step": 4914 }, { "epoch": 0.2060299927690389, "grad_norm": 1.6457364559173584, "learning_rate": 9.209234612776697e-06, "loss": 0.5207, "step": 4915 }, { "epoch": 0.20607191138404787, "grad_norm": 1.716248631477356, "learning_rate": 9.20886818628603e-06, "loss": 0.581, "step": 4916 }, { "epoch": 0.20611382999905684, "grad_norm": 1.7085449695587158, "learning_rate": 9.208501682210698e-06, "loss": 0.5441, "step": 4917 }, { "epoch": 0.2061557486140658, "grad_norm": 2.2144861221313477, "learning_rate": 9.208135100557454e-06, "loss": 0.5576, "step": 4918 }, { "epoch": 0.20619766722907476, "grad_norm": 1.7391222715377808, "learning_rate": 9.207768441333059e-06, "loss": 0.5113, "step": 4919 }, { "epoch": 0.20623958584408372, "grad_norm": 1.7112220525741577, "learning_rate": 9.207401704544269e-06, "loss": 0.5581, "step": 4920 }, { "epoch": 0.20628150445909268, "grad_norm": 1.6577116250991821, "learning_rate": 9.207034890197846e-06, "loss": 0.5306, "step": 4921 }, { "epoch": 0.20632342307410162, "grad_norm": 1.8742191791534424, "learning_rate": 9.206667998300549e-06, "loss": 0.55, "step": 4922 }, { "epoch": 0.20636534168911058, "grad_norm": 1.7034831047058105, "learning_rate": 9.206301028859144e-06, "loss": 0.5008, "step": 4923 }, { "epoch": 0.20640726030411954, "grad_norm": 1.8428668975830078, "learning_rate": 9.205933981880396e-06, "loss": 0.5719, "step": 4924 }, { "epoch": 0.2064491789191285, "grad_norm": 1.8244811296463013, "learning_rate": 9.205566857371068e-06, "loss": 0.5722, "step": 4925 }, { "epoch": 0.20649109753413747, "grad_norm": 1.7087091207504272, "learning_rate": 9.20519965533793e-06, "loss": 0.5139, "step": 4926 }, { "epoch": 0.20653301614914643, "grad_norm": 1.623926043510437, "learning_rate": 9.204832375787749e-06, "loss": 0.5611, "step": 4927 }, { "epoch": 0.2065749347641554, "grad_norm": 1.5095982551574707, "learning_rate": 9.204465018727296e-06, "loss": 0.5712, "step": 4928 }, { "epoch": 0.20661685337916436, "grad_norm": 1.8269429206848145, "learning_rate": 9.204097584163342e-06, "loss": 0.5702, "step": 4929 }, { "epoch": 0.20665877199417332, "grad_norm": 1.5176090002059937, "learning_rate": 9.203730072102663e-06, "loss": 0.545, "step": 4930 }, { "epoch": 0.20670069060918228, "grad_norm": 1.6384363174438477, "learning_rate": 9.20336248255203e-06, "loss": 0.5587, "step": 4931 }, { "epoch": 0.20674260922419124, "grad_norm": 1.6858278512954712, "learning_rate": 9.202994815518223e-06, "loss": 0.5227, "step": 4932 }, { "epoch": 0.2067845278392002, "grad_norm": 1.6978960037231445, "learning_rate": 9.202627071008015e-06, "loss": 0.5561, "step": 4933 }, { "epoch": 0.20682644645420914, "grad_norm": 1.501017689704895, "learning_rate": 9.202259249028187e-06, "loss": 0.4876, "step": 4934 }, { "epoch": 0.2068683650692181, "grad_norm": 1.671261191368103, "learning_rate": 9.201891349585518e-06, "loss": 0.5176, "step": 4935 }, { "epoch": 0.20691028368422706, "grad_norm": 2.2490060329437256, "learning_rate": 9.201523372686793e-06, "loss": 0.5585, "step": 4936 }, { "epoch": 0.20695220229923603, "grad_norm": 1.6909011602401733, "learning_rate": 9.201155318338793e-06, "loss": 0.5655, "step": 4937 }, { "epoch": 0.206994120914245, "grad_norm": 1.7248589992523193, "learning_rate": 9.200787186548303e-06, "loss": 0.5938, "step": 4938 }, { "epoch": 0.20703603952925395, "grad_norm": 2.0014069080352783, "learning_rate": 9.200418977322107e-06, "loss": 0.5719, "step": 4939 }, { "epoch": 0.2070779581442629, "grad_norm": 1.6925371885299683, "learning_rate": 9.200050690666994e-06, "loss": 0.5131, "step": 4940 }, { "epoch": 0.20711987675927188, "grad_norm": 1.6341623067855835, "learning_rate": 9.199682326589753e-06, "loss": 0.5697, "step": 4941 }, { "epoch": 0.20716179537428084, "grad_norm": 1.809051275253296, "learning_rate": 9.199313885097177e-06, "loss": 0.5192, "step": 4942 }, { "epoch": 0.2072037139892898, "grad_norm": 1.7542800903320312, "learning_rate": 9.198945366196052e-06, "loss": 0.5566, "step": 4943 }, { "epoch": 0.20724563260429876, "grad_norm": 1.794594168663025, "learning_rate": 9.198576769893175e-06, "loss": 0.5572, "step": 4944 }, { "epoch": 0.20728755121930773, "grad_norm": 1.7451133728027344, "learning_rate": 9.19820809619534e-06, "loss": 0.5209, "step": 4945 }, { "epoch": 0.2073294698343167, "grad_norm": 1.7924542427062988, "learning_rate": 9.197839345109341e-06, "loss": 0.5932, "step": 4946 }, { "epoch": 0.20737138844932562, "grad_norm": 1.9303525686264038, "learning_rate": 9.19747051664198e-06, "loss": 0.5717, "step": 4947 }, { "epoch": 0.20741330706433458, "grad_norm": 1.6143840551376343, "learning_rate": 9.19710161080005e-06, "loss": 0.524, "step": 4948 }, { "epoch": 0.20745522567934355, "grad_norm": 1.859185814857483, "learning_rate": 9.196732627590355e-06, "loss": 0.6113, "step": 4949 }, { "epoch": 0.2074971442943525, "grad_norm": 1.9444730281829834, "learning_rate": 9.196363567019695e-06, "loss": 0.5405, "step": 4950 }, { "epoch": 0.20753906290936147, "grad_norm": 2.7185113430023193, "learning_rate": 9.195994429094875e-06, "loss": 0.4968, "step": 4951 }, { "epoch": 0.20758098152437043, "grad_norm": 1.6446669101715088, "learning_rate": 9.195625213822699e-06, "loss": 0.5629, "step": 4952 }, { "epoch": 0.2076229001393794, "grad_norm": 1.6144309043884277, "learning_rate": 9.19525592120997e-06, "loss": 0.5594, "step": 4953 }, { "epoch": 0.20766481875438836, "grad_norm": 2.6259446144104004, "learning_rate": 9.194886551263502e-06, "loss": 0.5275, "step": 4954 }, { "epoch": 0.20770673736939732, "grad_norm": 1.5761281251907349, "learning_rate": 9.194517103990097e-06, "loss": 0.5387, "step": 4955 }, { "epoch": 0.20774865598440628, "grad_norm": 1.9422399997711182, "learning_rate": 9.194147579396568e-06, "loss": 0.6023, "step": 4956 }, { "epoch": 0.20779057459941525, "grad_norm": 2.358593225479126, "learning_rate": 9.193777977489727e-06, "loss": 0.5743, "step": 4957 }, { "epoch": 0.2078324932144242, "grad_norm": 1.6578404903411865, "learning_rate": 9.193408298276387e-06, "loss": 0.5512, "step": 4958 }, { "epoch": 0.20787441182943314, "grad_norm": 3.6166951656341553, "learning_rate": 9.193038541763361e-06, "loss": 0.5588, "step": 4959 }, { "epoch": 0.2079163304444421, "grad_norm": 1.7030004262924194, "learning_rate": 9.192668707957468e-06, "loss": 0.5181, "step": 4960 }, { "epoch": 0.20795824905945107, "grad_norm": 1.8945035934448242, "learning_rate": 9.192298796865521e-06, "loss": 0.6082, "step": 4961 }, { "epoch": 0.20800016767446003, "grad_norm": 1.7492612600326538, "learning_rate": 9.191928808494343e-06, "loss": 0.5045, "step": 4962 }, { "epoch": 0.208042086289469, "grad_norm": 1.991044044494629, "learning_rate": 9.191558742850753e-06, "loss": 0.5309, "step": 4963 }, { "epoch": 0.20808400490447795, "grad_norm": 1.642580270767212, "learning_rate": 9.191188599941571e-06, "loss": 0.5347, "step": 4964 }, { "epoch": 0.20812592351948692, "grad_norm": 1.6099931001663208, "learning_rate": 9.190818379773623e-06, "loss": 0.4596, "step": 4965 }, { "epoch": 0.20816784213449588, "grad_norm": 1.6406829357147217, "learning_rate": 9.190448082353731e-06, "loss": 0.5385, "step": 4966 }, { "epoch": 0.20820976074950484, "grad_norm": 1.6447210311889648, "learning_rate": 9.190077707688722e-06, "loss": 0.546, "step": 4967 }, { "epoch": 0.2082516793645138, "grad_norm": 1.6152143478393555, "learning_rate": 9.189707255785422e-06, "loss": 0.5292, "step": 4968 }, { "epoch": 0.20829359797952277, "grad_norm": 1.6590255498886108, "learning_rate": 9.189336726650662e-06, "loss": 0.5057, "step": 4969 }, { "epoch": 0.20833551659453173, "grad_norm": 1.9784207344055176, "learning_rate": 9.188966120291269e-06, "loss": 0.5364, "step": 4970 }, { "epoch": 0.2083774352095407, "grad_norm": 1.7853522300720215, "learning_rate": 9.188595436714079e-06, "loss": 0.529, "step": 4971 }, { "epoch": 0.20841935382454962, "grad_norm": 1.6906298398971558, "learning_rate": 9.188224675925922e-06, "loss": 0.535, "step": 4972 }, { "epoch": 0.2084612724395586, "grad_norm": 1.737758994102478, "learning_rate": 9.187853837933633e-06, "loss": 0.5826, "step": 4973 }, { "epoch": 0.20850319105456755, "grad_norm": 1.6241612434387207, "learning_rate": 9.187482922744048e-06, "loss": 0.5872, "step": 4974 }, { "epoch": 0.2085451096695765, "grad_norm": 1.6044169664382935, "learning_rate": 9.187111930364005e-06, "loss": 0.5351, "step": 4975 }, { "epoch": 0.20858702828458547, "grad_norm": 1.8397655487060547, "learning_rate": 9.18674086080034e-06, "loss": 0.5135, "step": 4976 }, { "epoch": 0.20862894689959444, "grad_norm": 1.547715425491333, "learning_rate": 9.186369714059897e-06, "loss": 0.5362, "step": 4977 }, { "epoch": 0.2086708655146034, "grad_norm": 1.630929946899414, "learning_rate": 9.185998490149514e-06, "loss": 0.5405, "step": 4978 }, { "epoch": 0.20871278412961236, "grad_norm": 1.784121036529541, "learning_rate": 9.185627189076039e-06, "loss": 0.5579, "step": 4979 }, { "epoch": 0.20875470274462132, "grad_norm": 1.8820334672927856, "learning_rate": 9.18525581084631e-06, "loss": 0.5715, "step": 4980 }, { "epoch": 0.20879662135963029, "grad_norm": 1.8979380130767822, "learning_rate": 9.184884355467179e-06, "loss": 0.4857, "step": 4981 }, { "epoch": 0.20883853997463925, "grad_norm": 1.6716138124465942, "learning_rate": 9.184512822945488e-06, "loss": 0.5268, "step": 4982 }, { "epoch": 0.2088804585896482, "grad_norm": 1.5907553434371948, "learning_rate": 9.18414121328809e-06, "loss": 0.5753, "step": 4983 }, { "epoch": 0.20892237720465714, "grad_norm": 2.47166109085083, "learning_rate": 9.183769526501831e-06, "loss": 0.5678, "step": 4984 }, { "epoch": 0.2089642958196661, "grad_norm": 1.4136725664138794, "learning_rate": 9.183397762593567e-06, "loss": 0.4968, "step": 4985 }, { "epoch": 0.20900621443467507, "grad_norm": 1.663540244102478, "learning_rate": 9.183025921570148e-06, "loss": 0.5719, "step": 4986 }, { "epoch": 0.20904813304968403, "grad_norm": 1.8049697875976562, "learning_rate": 9.182654003438427e-06, "loss": 0.5478, "step": 4987 }, { "epoch": 0.209090051664693, "grad_norm": 1.8111990690231323, "learning_rate": 9.182282008205263e-06, "loss": 0.579, "step": 4988 }, { "epoch": 0.20913197027970196, "grad_norm": 1.907580852508545, "learning_rate": 9.181909935877512e-06, "loss": 0.6195, "step": 4989 }, { "epoch": 0.20917388889471092, "grad_norm": 2.640205144882202, "learning_rate": 9.181537786462032e-06, "loss": 0.5209, "step": 4990 }, { "epoch": 0.20921580750971988, "grad_norm": 1.8608245849609375, "learning_rate": 9.181165559965686e-06, "loss": 0.5596, "step": 4991 }, { "epoch": 0.20925772612472884, "grad_norm": 1.8774110078811646, "learning_rate": 9.180793256395331e-06, "loss": 0.5371, "step": 4992 }, { "epoch": 0.2092996447397378, "grad_norm": 1.6162973642349243, "learning_rate": 9.180420875757833e-06, "loss": 0.5235, "step": 4993 }, { "epoch": 0.20934156335474677, "grad_norm": 1.973151683807373, "learning_rate": 9.180048418060055e-06, "loss": 0.5692, "step": 4994 }, { "epoch": 0.20938348196975573, "grad_norm": 1.583182692527771, "learning_rate": 9.179675883308862e-06, "loss": 0.5404, "step": 4995 }, { "epoch": 0.2094254005847647, "grad_norm": 1.6820379495620728, "learning_rate": 9.179303271511124e-06, "loss": 0.5455, "step": 4996 }, { "epoch": 0.20946731919977363, "grad_norm": 1.6250813007354736, "learning_rate": 9.178930582673708e-06, "loss": 0.5136, "step": 4997 }, { "epoch": 0.2095092378147826, "grad_norm": 4.865386486053467, "learning_rate": 9.178557816803484e-06, "loss": 0.5549, "step": 4998 }, { "epoch": 0.20955115642979155, "grad_norm": 1.6749402284622192, "learning_rate": 9.178184973907323e-06, "loss": 0.492, "step": 4999 }, { "epoch": 0.20959307504480051, "grad_norm": 1.548056721687317, "learning_rate": 9.177812053992098e-06, "loss": 0.532, "step": 5000 }, { "epoch": 0.20963499365980948, "grad_norm": 1.8546504974365234, "learning_rate": 9.177439057064684e-06, "loss": 0.5622, "step": 5001 }, { "epoch": 0.20967691227481844, "grad_norm": 1.5023034811019897, "learning_rate": 9.177065983131955e-06, "loss": 0.5167, "step": 5002 }, { "epoch": 0.2097188308898274, "grad_norm": 1.5790451765060425, "learning_rate": 9.176692832200788e-06, "loss": 0.5865, "step": 5003 }, { "epoch": 0.20976074950483636, "grad_norm": 1.5580435991287231, "learning_rate": 9.176319604278066e-06, "loss": 0.532, "step": 5004 }, { "epoch": 0.20980266811984533, "grad_norm": 1.876639723777771, "learning_rate": 9.175946299370663e-06, "loss": 0.5517, "step": 5005 }, { "epoch": 0.2098445867348543, "grad_norm": 1.7908222675323486, "learning_rate": 9.175572917485464e-06, "loss": 0.4858, "step": 5006 }, { "epoch": 0.20988650534986325, "grad_norm": 1.6542062759399414, "learning_rate": 9.175199458629352e-06, "loss": 0.5415, "step": 5007 }, { "epoch": 0.2099284239648722, "grad_norm": 2.1628668308258057, "learning_rate": 9.17482592280921e-06, "loss": 0.5417, "step": 5008 }, { "epoch": 0.20997034257988115, "grad_norm": 2.155707597732544, "learning_rate": 9.174452310031921e-06, "loss": 0.5714, "step": 5009 }, { "epoch": 0.2100122611948901, "grad_norm": 1.6631832122802734, "learning_rate": 9.174078620304376e-06, "loss": 0.5947, "step": 5010 }, { "epoch": 0.21005417980989907, "grad_norm": 1.8338117599487305, "learning_rate": 9.173704853633463e-06, "loss": 0.5966, "step": 5011 }, { "epoch": 0.21009609842490803, "grad_norm": 1.9733357429504395, "learning_rate": 9.17333101002607e-06, "loss": 0.5847, "step": 5012 }, { "epoch": 0.210138017039917, "grad_norm": 2.145256280899048, "learning_rate": 9.172957089489091e-06, "loss": 0.5204, "step": 5013 }, { "epoch": 0.21017993565492596, "grad_norm": 1.7434546947479248, "learning_rate": 9.172583092029416e-06, "loss": 0.5782, "step": 5014 }, { "epoch": 0.21022185426993492, "grad_norm": 2.704680919647217, "learning_rate": 9.17220901765394e-06, "loss": 0.5979, "step": 5015 }, { "epoch": 0.21026377288494388, "grad_norm": 2.5162625312805176, "learning_rate": 9.171834866369558e-06, "loss": 0.5963, "step": 5016 }, { "epoch": 0.21030569149995285, "grad_norm": 1.6535491943359375, "learning_rate": 9.171460638183168e-06, "loss": 0.5431, "step": 5017 }, { "epoch": 0.2103476101149618, "grad_norm": 3.526667594909668, "learning_rate": 9.171086333101669e-06, "loss": 0.5574, "step": 5018 }, { "epoch": 0.21038952872997077, "grad_norm": 1.6736130714416504, "learning_rate": 9.170711951131958e-06, "loss": 0.5669, "step": 5019 }, { "epoch": 0.21043144734497973, "grad_norm": 1.734579086303711, "learning_rate": 9.170337492280938e-06, "loss": 0.5494, "step": 5020 }, { "epoch": 0.2104733659599887, "grad_norm": 1.6140100955963135, "learning_rate": 9.169962956555513e-06, "loss": 0.6084, "step": 5021 }, { "epoch": 0.21051528457499763, "grad_norm": 1.581082820892334, "learning_rate": 9.169588343962585e-06, "loss": 0.5621, "step": 5022 }, { "epoch": 0.2105572031900066, "grad_norm": 1.8205780982971191, "learning_rate": 9.169213654509061e-06, "loss": 0.5501, "step": 5023 }, { "epoch": 0.21059912180501555, "grad_norm": 1.8221491575241089, "learning_rate": 9.168838888201847e-06, "loss": 0.584, "step": 5024 }, { "epoch": 0.21064104042002452, "grad_norm": 1.711812138557434, "learning_rate": 9.168464045047852e-06, "loss": 0.5124, "step": 5025 }, { "epoch": 0.21068295903503348, "grad_norm": 1.8034611940383911, "learning_rate": 9.168089125053982e-06, "loss": 0.5523, "step": 5026 }, { "epoch": 0.21072487765004244, "grad_norm": 2.292637825012207, "learning_rate": 9.167714128227155e-06, "loss": 0.5809, "step": 5027 }, { "epoch": 0.2107667962650514, "grad_norm": 1.7536437511444092, "learning_rate": 9.167339054574278e-06, "loss": 0.5688, "step": 5028 }, { "epoch": 0.21080871488006037, "grad_norm": 1.5729812383651733, "learning_rate": 9.166963904102266e-06, "loss": 0.4871, "step": 5029 }, { "epoch": 0.21085063349506933, "grad_norm": 1.6528569459915161, "learning_rate": 9.166588676818036e-06, "loss": 0.5801, "step": 5030 }, { "epoch": 0.2108925521100783, "grad_norm": 2.2565722465515137, "learning_rate": 9.166213372728503e-06, "loss": 0.5382, "step": 5031 }, { "epoch": 0.21093447072508725, "grad_norm": 1.6654446125030518, "learning_rate": 9.165837991840588e-06, "loss": 0.5288, "step": 5032 }, { "epoch": 0.21097638934009622, "grad_norm": 1.697470784187317, "learning_rate": 9.165462534161208e-06, "loss": 0.5459, "step": 5033 }, { "epoch": 0.21101830795510515, "grad_norm": 1.5928113460540771, "learning_rate": 9.165086999697281e-06, "loss": 0.5425, "step": 5034 }, { "epoch": 0.2110602265701141, "grad_norm": 1.6695444583892822, "learning_rate": 9.164711388455738e-06, "loss": 0.5079, "step": 5035 }, { "epoch": 0.21110214518512307, "grad_norm": 1.767033338546753, "learning_rate": 9.164335700443496e-06, "loss": 0.5852, "step": 5036 }, { "epoch": 0.21114406380013204, "grad_norm": 1.6490190029144287, "learning_rate": 9.163959935667483e-06, "loss": 0.5157, "step": 5037 }, { "epoch": 0.211185982415141, "grad_norm": 1.5065780878067017, "learning_rate": 9.163584094134625e-06, "loss": 0.4997, "step": 5038 }, { "epoch": 0.21122790103014996, "grad_norm": 1.7512497901916504, "learning_rate": 9.16320817585185e-06, "loss": 0.568, "step": 5039 }, { "epoch": 0.21126981964515892, "grad_norm": 1.7487246990203857, "learning_rate": 9.162832180826087e-06, "loss": 0.5279, "step": 5040 }, { "epoch": 0.2113117382601679, "grad_norm": 1.6925729513168335, "learning_rate": 9.162456109064268e-06, "loss": 0.5913, "step": 5041 }, { "epoch": 0.21135365687517685, "grad_norm": 1.8725589513778687, "learning_rate": 9.162079960573327e-06, "loss": 0.5355, "step": 5042 }, { "epoch": 0.2113955754901858, "grad_norm": 1.6257635354995728, "learning_rate": 9.161703735360195e-06, "loss": 0.559, "step": 5043 }, { "epoch": 0.21143749410519477, "grad_norm": 1.7480406761169434, "learning_rate": 9.161327433431806e-06, "loss": 0.6191, "step": 5044 }, { "epoch": 0.21147941272020374, "grad_norm": 1.6497312784194946, "learning_rate": 9.1609510547951e-06, "loss": 0.5206, "step": 5045 }, { "epoch": 0.2115213313352127, "grad_norm": 1.469508171081543, "learning_rate": 9.160574599457013e-06, "loss": 0.5009, "step": 5046 }, { "epoch": 0.21156324995022163, "grad_norm": 1.6105250120162964, "learning_rate": 9.160198067424487e-06, "loss": 0.5171, "step": 5047 }, { "epoch": 0.2116051685652306, "grad_norm": 1.548580527305603, "learning_rate": 9.15982145870446e-06, "loss": 0.5264, "step": 5048 }, { "epoch": 0.21164708718023956, "grad_norm": 1.6302992105484009, "learning_rate": 9.159444773303875e-06, "loss": 0.5177, "step": 5049 }, { "epoch": 0.21168900579524852, "grad_norm": 2.1179022789001465, "learning_rate": 9.159068011229674e-06, "loss": 0.5162, "step": 5050 }, { "epoch": 0.21173092441025748, "grad_norm": 1.906962513923645, "learning_rate": 9.158691172488806e-06, "loss": 0.5551, "step": 5051 }, { "epoch": 0.21177284302526644, "grad_norm": 1.7390981912612915, "learning_rate": 9.158314257088216e-06, "loss": 0.5485, "step": 5052 }, { "epoch": 0.2118147616402754, "grad_norm": 1.6463428735733032, "learning_rate": 9.157937265034851e-06, "loss": 0.4979, "step": 5053 }, { "epoch": 0.21185668025528437, "grad_norm": 1.7294666767120361, "learning_rate": 9.15756019633566e-06, "loss": 0.575, "step": 5054 }, { "epoch": 0.21189859887029333, "grad_norm": 1.6292756795883179, "learning_rate": 9.157183050997595e-06, "loss": 0.4872, "step": 5055 }, { "epoch": 0.2119405174853023, "grad_norm": 2.0798609256744385, "learning_rate": 9.156805829027607e-06, "loss": 0.5894, "step": 5056 }, { "epoch": 0.21198243610031126, "grad_norm": 1.6962472200393677, "learning_rate": 9.156428530432651e-06, "loss": 0.5782, "step": 5057 }, { "epoch": 0.21202435471532022, "grad_norm": 1.8085911273956299, "learning_rate": 9.15605115521968e-06, "loss": 0.5643, "step": 5058 }, { "epoch": 0.21206627333032915, "grad_norm": 2.0338103771209717, "learning_rate": 9.155673703395652e-06, "loss": 0.5575, "step": 5059 }, { "epoch": 0.21210819194533811, "grad_norm": 2.466667652130127, "learning_rate": 9.155296174967524e-06, "loss": 0.5145, "step": 5060 }, { "epoch": 0.21215011056034708, "grad_norm": 1.711991310119629, "learning_rate": 9.154918569942257e-06, "loss": 0.5305, "step": 5061 }, { "epoch": 0.21219202917535604, "grad_norm": 1.659472942352295, "learning_rate": 9.154540888326808e-06, "loss": 0.5335, "step": 5062 }, { "epoch": 0.212233947790365, "grad_norm": 1.5518227815628052, "learning_rate": 9.154163130128144e-06, "loss": 0.5471, "step": 5063 }, { "epoch": 0.21227586640537396, "grad_norm": 1.8281304836273193, "learning_rate": 9.153785295353224e-06, "loss": 0.5727, "step": 5064 }, { "epoch": 0.21231778502038293, "grad_norm": 1.495542049407959, "learning_rate": 9.153407384009012e-06, "loss": 0.4818, "step": 5065 }, { "epoch": 0.2123597036353919, "grad_norm": 1.738115668296814, "learning_rate": 9.15302939610248e-06, "loss": 0.4847, "step": 5066 }, { "epoch": 0.21240162225040085, "grad_norm": 1.5490586757659912, "learning_rate": 9.152651331640593e-06, "loss": 0.5064, "step": 5067 }, { "epoch": 0.2124435408654098, "grad_norm": 2.1059091091156006, "learning_rate": 9.152273190630318e-06, "loss": 0.5195, "step": 5068 }, { "epoch": 0.21248545948041878, "grad_norm": 1.8159061670303345, "learning_rate": 9.151894973078627e-06, "loss": 0.5431, "step": 5069 }, { "epoch": 0.21252737809542774, "grad_norm": 1.5353341102600098, "learning_rate": 9.151516678992492e-06, "loss": 0.5012, "step": 5070 }, { "epoch": 0.2125692967104367, "grad_norm": 1.821366310119629, "learning_rate": 9.151138308378887e-06, "loss": 0.5705, "step": 5071 }, { "epoch": 0.21261121532544563, "grad_norm": 1.6057264804840088, "learning_rate": 9.150759861244786e-06, "loss": 0.4942, "step": 5072 }, { "epoch": 0.2126531339404546, "grad_norm": 1.5114452838897705, "learning_rate": 9.150381337597166e-06, "loss": 0.5311, "step": 5073 }, { "epoch": 0.21269505255546356, "grad_norm": 1.9257069826126099, "learning_rate": 9.150002737443002e-06, "loss": 0.5755, "step": 5074 }, { "epoch": 0.21273697117047252, "grad_norm": 1.744123935699463, "learning_rate": 9.149624060789275e-06, "loss": 0.5548, "step": 5075 }, { "epoch": 0.21277888978548148, "grad_norm": 1.6323914527893066, "learning_rate": 9.149245307642966e-06, "loss": 0.5196, "step": 5076 }, { "epoch": 0.21282080840049045, "grad_norm": 1.5974993705749512, "learning_rate": 9.148866478011057e-06, "loss": 0.5394, "step": 5077 }, { "epoch": 0.2128627270154994, "grad_norm": 1.744036078453064, "learning_rate": 9.148487571900527e-06, "loss": 0.5568, "step": 5078 }, { "epoch": 0.21290464563050837, "grad_norm": 1.7937102317810059, "learning_rate": 9.148108589318365e-06, "loss": 0.5493, "step": 5079 }, { "epoch": 0.21294656424551733, "grad_norm": 1.573005199432373, "learning_rate": 9.147729530271556e-06, "loss": 0.5742, "step": 5080 }, { "epoch": 0.2129884828605263, "grad_norm": 1.7749087810516357, "learning_rate": 9.147350394767088e-06, "loss": 0.5394, "step": 5081 }, { "epoch": 0.21303040147553526, "grad_norm": 1.7672090530395508, "learning_rate": 9.146971182811948e-06, "loss": 0.5579, "step": 5082 }, { "epoch": 0.21307232009054422, "grad_norm": 1.5989739894866943, "learning_rate": 9.146591894413128e-06, "loss": 0.5504, "step": 5083 }, { "epoch": 0.21311423870555316, "grad_norm": 1.6979854106903076, "learning_rate": 9.146212529577618e-06, "loss": 0.5677, "step": 5084 }, { "epoch": 0.21315615732056212, "grad_norm": 1.749089002609253, "learning_rate": 9.145833088312412e-06, "loss": 0.5454, "step": 5085 }, { "epoch": 0.21319807593557108, "grad_norm": 1.8643742799758911, "learning_rate": 9.145453570624505e-06, "loss": 0.5758, "step": 5086 }, { "epoch": 0.21323999455058004, "grad_norm": 1.6794376373291016, "learning_rate": 9.145073976520893e-06, "loss": 0.5835, "step": 5087 }, { "epoch": 0.213281913165589, "grad_norm": 1.7966972589492798, "learning_rate": 9.144694306008571e-06, "loss": 0.5776, "step": 5088 }, { "epoch": 0.21332383178059797, "grad_norm": 1.624019742012024, "learning_rate": 9.14431455909454e-06, "loss": 0.5009, "step": 5089 }, { "epoch": 0.21336575039560693, "grad_norm": 1.8682701587677002, "learning_rate": 9.1439347357858e-06, "loss": 0.5706, "step": 5090 }, { "epoch": 0.2134076690106159, "grad_norm": 1.6235169172286987, "learning_rate": 9.143554836089352e-06, "loss": 0.5259, "step": 5091 }, { "epoch": 0.21344958762562485, "grad_norm": 2.132439136505127, "learning_rate": 9.143174860012199e-06, "loss": 0.5779, "step": 5092 }, { "epoch": 0.21349150624063382, "grad_norm": 1.6377538442611694, "learning_rate": 9.142794807561344e-06, "loss": 0.4787, "step": 5093 }, { "epoch": 0.21353342485564278, "grad_norm": 1.872214913368225, "learning_rate": 9.142414678743795e-06, "loss": 0.5809, "step": 5094 }, { "epoch": 0.21357534347065174, "grad_norm": 2.1094722747802734, "learning_rate": 9.142034473566557e-06, "loss": 0.5554, "step": 5095 }, { "epoch": 0.2136172620856607, "grad_norm": 1.930686593055725, "learning_rate": 9.141654192036642e-06, "loss": 0.5082, "step": 5096 }, { "epoch": 0.21365918070066964, "grad_norm": 1.9962172508239746, "learning_rate": 9.141273834161055e-06, "loss": 0.5179, "step": 5097 }, { "epoch": 0.2137010993156786, "grad_norm": 1.7956080436706543, "learning_rate": 9.140893399946811e-06, "loss": 0.5745, "step": 5098 }, { "epoch": 0.21374301793068756, "grad_norm": 1.7152352333068848, "learning_rate": 9.140512889400923e-06, "loss": 0.5892, "step": 5099 }, { "epoch": 0.21378493654569652, "grad_norm": 1.6949503421783447, "learning_rate": 9.140132302530402e-06, "loss": 0.5825, "step": 5100 }, { "epoch": 0.2138268551607055, "grad_norm": 2.130357265472412, "learning_rate": 9.139751639342267e-06, "loss": 0.5613, "step": 5101 }, { "epoch": 0.21386877377571445, "grad_norm": 1.861238956451416, "learning_rate": 9.139370899843533e-06, "loss": 0.5142, "step": 5102 }, { "epoch": 0.2139106923907234, "grad_norm": 1.6810810565948486, "learning_rate": 9.13899008404122e-06, "loss": 0.5306, "step": 5103 }, { "epoch": 0.21395261100573237, "grad_norm": 2.038125514984131, "learning_rate": 9.138609191942345e-06, "loss": 0.5621, "step": 5104 }, { "epoch": 0.21399452962074134, "grad_norm": 1.6850347518920898, "learning_rate": 9.138228223553931e-06, "loss": 0.5784, "step": 5105 }, { "epoch": 0.2140364482357503, "grad_norm": 1.6374000310897827, "learning_rate": 9.137847178883002e-06, "loss": 0.5129, "step": 5106 }, { "epoch": 0.21407836685075926, "grad_norm": 1.9348195791244507, "learning_rate": 9.13746605793658e-06, "loss": 0.5197, "step": 5107 }, { "epoch": 0.21412028546576822, "grad_norm": 2.391627788543701, "learning_rate": 9.137084860721691e-06, "loss": 0.5398, "step": 5108 }, { "epoch": 0.21416220408077716, "grad_norm": 2.0250794887542725, "learning_rate": 9.136703587245362e-06, "loss": 0.5457, "step": 5109 }, { "epoch": 0.21420412269578612, "grad_norm": 3.5103437900543213, "learning_rate": 9.136322237514621e-06, "loss": 0.5514, "step": 5110 }, { "epoch": 0.21424604131079508, "grad_norm": 2.01762318611145, "learning_rate": 9.1359408115365e-06, "loss": 0.5927, "step": 5111 }, { "epoch": 0.21428795992580404, "grad_norm": 1.9464185237884521, "learning_rate": 9.135559309318025e-06, "loss": 0.552, "step": 5112 }, { "epoch": 0.214329878540813, "grad_norm": 1.920516014099121, "learning_rate": 9.135177730866231e-06, "loss": 0.5223, "step": 5113 }, { "epoch": 0.21437179715582197, "grad_norm": 1.749102234840393, "learning_rate": 9.134796076188156e-06, "loss": 0.5404, "step": 5114 }, { "epoch": 0.21441371577083093, "grad_norm": 1.670457124710083, "learning_rate": 9.134414345290828e-06, "loss": 0.5309, "step": 5115 }, { "epoch": 0.2144556343858399, "grad_norm": 1.9560847282409668, "learning_rate": 9.134032538181287e-06, "loss": 0.5769, "step": 5116 }, { "epoch": 0.21449755300084886, "grad_norm": 1.6357383728027344, "learning_rate": 9.133650654866574e-06, "loss": 0.5657, "step": 5117 }, { "epoch": 0.21453947161585782, "grad_norm": 1.8839083909988403, "learning_rate": 9.133268695353726e-06, "loss": 0.56, "step": 5118 }, { "epoch": 0.21458139023086678, "grad_norm": 2.095745086669922, "learning_rate": 9.13288665964978e-06, "loss": 0.5677, "step": 5119 }, { "epoch": 0.21462330884587574, "grad_norm": 2.290882110595703, "learning_rate": 9.132504547761784e-06, "loss": 0.5554, "step": 5120 }, { "epoch": 0.2146652274608847, "grad_norm": 1.4864511489868164, "learning_rate": 9.13212235969678e-06, "loss": 0.5554, "step": 5121 }, { "epoch": 0.21470714607589364, "grad_norm": 1.5726960897445679, "learning_rate": 9.131740095461813e-06, "loss": 0.6018, "step": 5122 }, { "epoch": 0.2147490646909026, "grad_norm": 1.8890377283096313, "learning_rate": 9.131357755063929e-06, "loss": 0.5304, "step": 5123 }, { "epoch": 0.21479098330591156, "grad_norm": 2.1943600177764893, "learning_rate": 9.130975338510176e-06, "loss": 0.627, "step": 5124 }, { "epoch": 0.21483290192092053, "grad_norm": 1.7039562463760376, "learning_rate": 9.130592845807604e-06, "loss": 0.6095, "step": 5125 }, { "epoch": 0.2148748205359295, "grad_norm": 1.5362399816513062, "learning_rate": 9.130210276963264e-06, "loss": 0.5713, "step": 5126 }, { "epoch": 0.21491673915093845, "grad_norm": 1.5477641820907593, "learning_rate": 9.129827631984205e-06, "loss": 0.5955, "step": 5127 }, { "epoch": 0.21495865776594741, "grad_norm": 1.7104198932647705, "learning_rate": 9.129444910877484e-06, "loss": 0.5211, "step": 5128 }, { "epoch": 0.21500057638095638, "grad_norm": 1.5682978630065918, "learning_rate": 9.129062113650156e-06, "loss": 0.4983, "step": 5129 }, { "epoch": 0.21504249499596534, "grad_norm": 1.7631679773330688, "learning_rate": 9.128679240309275e-06, "loss": 0.6059, "step": 5130 }, { "epoch": 0.2150844136109743, "grad_norm": 1.8387739658355713, "learning_rate": 9.1282962908619e-06, "loss": 0.5576, "step": 5131 }, { "epoch": 0.21512633222598326, "grad_norm": 1.7638611793518066, "learning_rate": 9.12791326531509e-06, "loss": 0.5284, "step": 5132 }, { "epoch": 0.21516825084099223, "grad_norm": 1.8532718420028687, "learning_rate": 9.127530163675906e-06, "loss": 0.5389, "step": 5133 }, { "epoch": 0.21521016945600116, "grad_norm": 1.7743581533432007, "learning_rate": 9.12714698595141e-06, "loss": 0.6015, "step": 5134 }, { "epoch": 0.21525208807101012, "grad_norm": 1.8347779512405396, "learning_rate": 9.126763732148663e-06, "loss": 0.5362, "step": 5135 }, { "epoch": 0.21529400668601908, "grad_norm": 1.528139352798462, "learning_rate": 9.126380402274732e-06, "loss": 0.5051, "step": 5136 }, { "epoch": 0.21533592530102805, "grad_norm": 1.625345230102539, "learning_rate": 9.125996996336686e-06, "loss": 0.5194, "step": 5137 }, { "epoch": 0.215377843916037, "grad_norm": 2.169400453567505, "learning_rate": 9.125613514341587e-06, "loss": 0.5751, "step": 5138 }, { "epoch": 0.21541976253104597, "grad_norm": 1.799045443534851, "learning_rate": 9.125229956296506e-06, "loss": 0.5541, "step": 5139 }, { "epoch": 0.21546168114605493, "grad_norm": 1.5975948572158813, "learning_rate": 9.124846322208514e-06, "loss": 0.5036, "step": 5140 }, { "epoch": 0.2155035997610639, "grad_norm": 1.9739187955856323, "learning_rate": 9.124462612084683e-06, "loss": 0.5594, "step": 5141 }, { "epoch": 0.21554551837607286, "grad_norm": 1.8615354299545288, "learning_rate": 9.124078825932085e-06, "loss": 0.5609, "step": 5142 }, { "epoch": 0.21558743699108182, "grad_norm": 1.6974607706069946, "learning_rate": 9.123694963757796e-06, "loss": 0.5365, "step": 5143 }, { "epoch": 0.21562935560609078, "grad_norm": 2.043940305709839, "learning_rate": 9.12331102556889e-06, "loss": 0.6264, "step": 5144 }, { "epoch": 0.21567127422109975, "grad_norm": 1.5898364782333374, "learning_rate": 9.122927011372448e-06, "loss": 0.5757, "step": 5145 }, { "epoch": 0.2157131928361087, "grad_norm": 1.6748065948486328, "learning_rate": 9.122542921175545e-06, "loss": 0.5687, "step": 5146 }, { "epoch": 0.21575511145111764, "grad_norm": 1.7201943397521973, "learning_rate": 9.122158754985263e-06, "loss": 0.4967, "step": 5147 }, { "epoch": 0.2157970300661266, "grad_norm": 1.3828006982803345, "learning_rate": 9.121774512808681e-06, "loss": 0.5267, "step": 5148 }, { "epoch": 0.21583894868113557, "grad_norm": 1.827818512916565, "learning_rate": 9.121390194652886e-06, "loss": 0.5558, "step": 5149 }, { "epoch": 0.21588086729614453, "grad_norm": 1.595979928970337, "learning_rate": 9.121005800524958e-06, "loss": 0.4876, "step": 5150 }, { "epoch": 0.2159227859111535, "grad_norm": 1.535131812095642, "learning_rate": 9.120621330431987e-06, "loss": 0.542, "step": 5151 }, { "epoch": 0.21596470452616245, "grad_norm": 1.3992570638656616, "learning_rate": 9.120236784381059e-06, "loss": 0.5169, "step": 5152 }, { "epoch": 0.21600662314117142, "grad_norm": 1.6073161363601685, "learning_rate": 9.119852162379261e-06, "loss": 0.5817, "step": 5153 }, { "epoch": 0.21604854175618038, "grad_norm": 1.818317174911499, "learning_rate": 9.119467464433684e-06, "loss": 0.581, "step": 5154 }, { "epoch": 0.21609046037118934, "grad_norm": 1.6068240404129028, "learning_rate": 9.11908269055142e-06, "loss": 0.5, "step": 5155 }, { "epoch": 0.2161323789861983, "grad_norm": 1.690317153930664, "learning_rate": 9.11869784073956e-06, "loss": 0.5266, "step": 5156 }, { "epoch": 0.21617429760120727, "grad_norm": 1.5252273082733154, "learning_rate": 9.1183129150052e-06, "loss": 0.5033, "step": 5157 }, { "epoch": 0.21621621621621623, "grad_norm": 2.3804478645324707, "learning_rate": 9.117927913355434e-06, "loss": 0.5185, "step": 5158 }, { "epoch": 0.21625813483122516, "grad_norm": 1.7624136209487915, "learning_rate": 9.11754283579736e-06, "loss": 0.6313, "step": 5159 }, { "epoch": 0.21630005344623413, "grad_norm": 1.6710498332977295, "learning_rate": 9.117157682338077e-06, "loss": 0.495, "step": 5160 }, { "epoch": 0.2163419720612431, "grad_norm": 3.1976475715637207, "learning_rate": 9.116772452984683e-06, "loss": 0.5229, "step": 5161 }, { "epoch": 0.21638389067625205, "grad_norm": 1.4949252605438232, "learning_rate": 9.116387147744282e-06, "loss": 0.5422, "step": 5162 }, { "epoch": 0.216425809291261, "grad_norm": 1.705304503440857, "learning_rate": 9.116001766623972e-06, "loss": 0.5292, "step": 5163 }, { "epoch": 0.21646772790626997, "grad_norm": 1.5637034177780151, "learning_rate": 9.115616309630861e-06, "loss": 0.505, "step": 5164 }, { "epoch": 0.21650964652127894, "grad_norm": 1.6020938158035278, "learning_rate": 9.115230776772053e-06, "loss": 0.5563, "step": 5165 }, { "epoch": 0.2165515651362879, "grad_norm": 1.6251540184020996, "learning_rate": 9.114845168054655e-06, "loss": 0.5336, "step": 5166 }, { "epoch": 0.21659348375129686, "grad_norm": 1.653449296951294, "learning_rate": 9.114459483485773e-06, "loss": 0.5125, "step": 5167 }, { "epoch": 0.21663540236630582, "grad_norm": 8.640624046325684, "learning_rate": 9.11407372307252e-06, "loss": 0.5827, "step": 5168 }, { "epoch": 0.21667732098131479, "grad_norm": 1.5562996864318848, "learning_rate": 9.113687886822006e-06, "loss": 0.5807, "step": 5169 }, { "epoch": 0.21671923959632375, "grad_norm": 2.2193264961242676, "learning_rate": 9.113301974741342e-06, "loss": 0.5138, "step": 5170 }, { "epoch": 0.2167611582113327, "grad_norm": 1.9125628471374512, "learning_rate": 9.112915986837643e-06, "loss": 0.5307, "step": 5171 }, { "epoch": 0.21680307682634165, "grad_norm": 1.3939770460128784, "learning_rate": 9.112529923118023e-06, "loss": 0.5408, "step": 5172 }, { "epoch": 0.2168449954413506, "grad_norm": 1.7257159948349, "learning_rate": 9.1121437835896e-06, "loss": 0.5687, "step": 5173 }, { "epoch": 0.21688691405635957, "grad_norm": 1.6560157537460327, "learning_rate": 9.111757568259492e-06, "loss": 0.5507, "step": 5174 }, { "epoch": 0.21692883267136853, "grad_norm": 1.6824421882629395, "learning_rate": 9.111371277134816e-06, "loss": 0.5484, "step": 5175 }, { "epoch": 0.2169707512863775, "grad_norm": 1.5702269077301025, "learning_rate": 9.110984910222695e-06, "loss": 0.574, "step": 5176 }, { "epoch": 0.21701266990138646, "grad_norm": 1.629880428314209, "learning_rate": 9.110598467530249e-06, "loss": 0.5359, "step": 5177 }, { "epoch": 0.21705458851639542, "grad_norm": 1.7216123342514038, "learning_rate": 9.110211949064605e-06, "loss": 0.5458, "step": 5178 }, { "epoch": 0.21709650713140438, "grad_norm": 1.940519094467163, "learning_rate": 9.109825354832885e-06, "loss": 0.5626, "step": 5179 }, { "epoch": 0.21713842574641334, "grad_norm": 1.7685681581497192, "learning_rate": 9.109438684842216e-06, "loss": 0.5654, "step": 5180 }, { "epoch": 0.2171803443614223, "grad_norm": 1.586480736732483, "learning_rate": 9.109051939099726e-06, "loss": 0.5806, "step": 5181 }, { "epoch": 0.21722226297643127, "grad_norm": 1.8584669828414917, "learning_rate": 9.108665117612545e-06, "loss": 0.5214, "step": 5182 }, { "epoch": 0.21726418159144023, "grad_norm": 1.7629884481430054, "learning_rate": 9.108278220387802e-06, "loss": 0.5492, "step": 5183 }, { "epoch": 0.21730610020644917, "grad_norm": 1.7901405096054077, "learning_rate": 9.10789124743263e-06, "loss": 0.55, "step": 5184 }, { "epoch": 0.21734801882145813, "grad_norm": 5.483756065368652, "learning_rate": 9.107504198754162e-06, "loss": 0.5442, "step": 5185 }, { "epoch": 0.2173899374364671, "grad_norm": 1.6972259283065796, "learning_rate": 9.107117074359533e-06, "loss": 0.4844, "step": 5186 }, { "epoch": 0.21743185605147605, "grad_norm": 1.720901370048523, "learning_rate": 9.106729874255878e-06, "loss": 0.539, "step": 5187 }, { "epoch": 0.21747377466648501, "grad_norm": 1.86781644821167, "learning_rate": 9.106342598450334e-06, "loss": 0.5631, "step": 5188 }, { "epoch": 0.21751569328149398, "grad_norm": 1.759292483329773, "learning_rate": 9.105955246950043e-06, "loss": 0.5446, "step": 5189 }, { "epoch": 0.21755761189650294, "grad_norm": 2.2314088344573975, "learning_rate": 9.105567819762142e-06, "loss": 0.5132, "step": 5190 }, { "epoch": 0.2175995305115119, "grad_norm": 1.926900863647461, "learning_rate": 9.105180316893773e-06, "loss": 0.5607, "step": 5191 }, { "epoch": 0.21764144912652086, "grad_norm": 1.6329660415649414, "learning_rate": 9.104792738352082e-06, "loss": 0.5055, "step": 5192 }, { "epoch": 0.21768336774152983, "grad_norm": 1.8447213172912598, "learning_rate": 9.10440508414421e-06, "loss": 0.5564, "step": 5193 }, { "epoch": 0.2177252863565388, "grad_norm": 1.7680031061172485, "learning_rate": 9.104017354277306e-06, "loss": 0.5685, "step": 5194 }, { "epoch": 0.21776720497154775, "grad_norm": 1.6985923051834106, "learning_rate": 9.103629548758514e-06, "loss": 0.5673, "step": 5195 }, { "epoch": 0.2178091235865567, "grad_norm": 1.8763337135314941, "learning_rate": 9.103241667594986e-06, "loss": 0.5437, "step": 5196 }, { "epoch": 0.21785104220156565, "grad_norm": 1.7487047910690308, "learning_rate": 9.102853710793871e-06, "loss": 0.5582, "step": 5197 }, { "epoch": 0.2178929608165746, "grad_norm": 1.5523325204849243, "learning_rate": 9.102465678362318e-06, "loss": 0.539, "step": 5198 }, { "epoch": 0.21793487943158357, "grad_norm": 2.235501527786255, "learning_rate": 9.102077570307483e-06, "loss": 0.5997, "step": 5199 }, { "epoch": 0.21797679804659253, "grad_norm": 1.8636562824249268, "learning_rate": 9.101689386636518e-06, "loss": 0.5708, "step": 5200 }, { "epoch": 0.2180187166616015, "grad_norm": 1.7027453184127808, "learning_rate": 9.10130112735658e-06, "loss": 0.544, "step": 5201 }, { "epoch": 0.21806063527661046, "grad_norm": 1.4682302474975586, "learning_rate": 9.100912792474826e-06, "loss": 0.5356, "step": 5202 }, { "epoch": 0.21810255389161942, "grad_norm": 1.5870925188064575, "learning_rate": 9.100524381998413e-06, "loss": 0.482, "step": 5203 }, { "epoch": 0.21814447250662838, "grad_norm": 1.9992555379867554, "learning_rate": 9.100135895934502e-06, "loss": 0.4882, "step": 5204 }, { "epoch": 0.21818639112163735, "grad_norm": 1.6494872570037842, "learning_rate": 9.099747334290254e-06, "loss": 0.5578, "step": 5205 }, { "epoch": 0.2182283097366463, "grad_norm": 1.6554747819900513, "learning_rate": 9.099358697072833e-06, "loss": 0.5758, "step": 5206 }, { "epoch": 0.21827022835165527, "grad_norm": 1.571831226348877, "learning_rate": 9.0989699842894e-06, "loss": 0.5268, "step": 5207 }, { "epoch": 0.21831214696666423, "grad_norm": 1.7744624614715576, "learning_rate": 9.098581195947123e-06, "loss": 0.5588, "step": 5208 }, { "epoch": 0.21835406558167317, "grad_norm": 1.62978994846344, "learning_rate": 9.098192332053168e-06, "loss": 0.5769, "step": 5209 }, { "epoch": 0.21839598419668213, "grad_norm": 2.002197504043579, "learning_rate": 9.097803392614702e-06, "loss": 0.5087, "step": 5210 }, { "epoch": 0.2184379028116911, "grad_norm": 1.7455871105194092, "learning_rate": 9.097414377638896e-06, "loss": 0.5207, "step": 5211 }, { "epoch": 0.21847982142670005, "grad_norm": 1.5314805507659912, "learning_rate": 9.09702528713292e-06, "loss": 0.5228, "step": 5212 }, { "epoch": 0.21852174004170902, "grad_norm": 1.7856587171554565, "learning_rate": 9.096636121103948e-06, "loss": 0.5286, "step": 5213 }, { "epoch": 0.21856365865671798, "grad_norm": 1.9781361818313599, "learning_rate": 9.096246879559152e-06, "loss": 0.5858, "step": 5214 }, { "epoch": 0.21860557727172694, "grad_norm": 1.7975749969482422, "learning_rate": 9.095857562505706e-06, "loss": 0.6252, "step": 5215 }, { "epoch": 0.2186474958867359, "grad_norm": 1.7021193504333496, "learning_rate": 9.09546816995079e-06, "loss": 0.5599, "step": 5216 }, { "epoch": 0.21868941450174487, "grad_norm": 1.8171157836914062, "learning_rate": 9.09507870190158e-06, "loss": 0.5002, "step": 5217 }, { "epoch": 0.21873133311675383, "grad_norm": 1.7417255640029907, "learning_rate": 9.094689158365255e-06, "loss": 0.5951, "step": 5218 }, { "epoch": 0.2187732517317628, "grad_norm": 2.7963781356811523, "learning_rate": 9.094299539348996e-06, "loss": 0.4981, "step": 5219 }, { "epoch": 0.21881517034677175, "grad_norm": 1.8466076850891113, "learning_rate": 9.093909844859985e-06, "loss": 0.5534, "step": 5220 }, { "epoch": 0.21885708896178072, "grad_norm": 3.2373597621917725, "learning_rate": 9.093520074905407e-06, "loss": 0.5342, "step": 5221 }, { "epoch": 0.21889900757678965, "grad_norm": 2.3142006397247314, "learning_rate": 9.093130229492443e-06, "loss": 0.5804, "step": 5222 }, { "epoch": 0.2189409261917986, "grad_norm": 1.694006323814392, "learning_rate": 9.092740308628284e-06, "loss": 0.4994, "step": 5223 }, { "epoch": 0.21898284480680758, "grad_norm": 1.6269155740737915, "learning_rate": 9.092350312320115e-06, "loss": 0.5188, "step": 5224 }, { "epoch": 0.21902476342181654, "grad_norm": 1.7069261074066162, "learning_rate": 9.091960240575124e-06, "loss": 0.586, "step": 5225 }, { "epoch": 0.2190666820368255, "grad_norm": 1.57003653049469, "learning_rate": 9.091570093400506e-06, "loss": 0.5441, "step": 5226 }, { "epoch": 0.21910860065183446, "grad_norm": 1.883176326751709, "learning_rate": 9.091179870803448e-06, "loss": 0.5869, "step": 5227 }, { "epoch": 0.21915051926684342, "grad_norm": 1.8546522855758667, "learning_rate": 9.090789572791142e-06, "loss": 0.5653, "step": 5228 }, { "epoch": 0.2191924378818524, "grad_norm": 1.914074420928955, "learning_rate": 9.090399199370789e-06, "loss": 0.5039, "step": 5229 }, { "epoch": 0.21923435649686135, "grad_norm": 1.5585503578186035, "learning_rate": 9.090008750549582e-06, "loss": 0.5517, "step": 5230 }, { "epoch": 0.2192762751118703, "grad_norm": 1.7481236457824707, "learning_rate": 9.089618226334718e-06, "loss": 0.5746, "step": 5231 }, { "epoch": 0.21931819372687927, "grad_norm": 1.5379722118377686, "learning_rate": 9.089227626733394e-06, "loss": 0.4968, "step": 5232 }, { "epoch": 0.21936011234188824, "grad_norm": 1.5961940288543701, "learning_rate": 9.088836951752813e-06, "loss": 0.5174, "step": 5233 }, { "epoch": 0.21940203095689717, "grad_norm": 1.6492544412612915, "learning_rate": 9.088446201400176e-06, "loss": 0.503, "step": 5234 }, { "epoch": 0.21944394957190613, "grad_norm": 2.6588127613067627, "learning_rate": 9.088055375682682e-06, "loss": 0.6134, "step": 5235 }, { "epoch": 0.2194858681869151, "grad_norm": 1.725288987159729, "learning_rate": 9.087664474607542e-06, "loss": 0.5457, "step": 5236 }, { "epoch": 0.21952778680192406, "grad_norm": 1.728535532951355, "learning_rate": 9.087273498181958e-06, "loss": 0.528, "step": 5237 }, { "epoch": 0.21956970541693302, "grad_norm": 1.7531094551086426, "learning_rate": 9.086882446413136e-06, "loss": 0.5126, "step": 5238 }, { "epoch": 0.21961162403194198, "grad_norm": 1.7065157890319824, "learning_rate": 9.086491319308288e-06, "loss": 0.62, "step": 5239 }, { "epoch": 0.21965354264695094, "grad_norm": 2.327319622039795, "learning_rate": 9.086100116874622e-06, "loss": 0.5562, "step": 5240 }, { "epoch": 0.2196954612619599, "grad_norm": 1.6128369569778442, "learning_rate": 9.085708839119347e-06, "loss": 0.5118, "step": 5241 }, { "epoch": 0.21973737987696887, "grad_norm": 2.074037551879883, "learning_rate": 9.085317486049681e-06, "loss": 0.5149, "step": 5242 }, { "epoch": 0.21977929849197783, "grad_norm": 1.7825915813446045, "learning_rate": 9.084926057672831e-06, "loss": 0.5, "step": 5243 }, { "epoch": 0.2198212171069868, "grad_norm": 1.6960281133651733, "learning_rate": 9.084534553996021e-06, "loss": 0.5213, "step": 5244 }, { "epoch": 0.21986313572199576, "grad_norm": 1.8721492290496826, "learning_rate": 9.08414297502646e-06, "loss": 0.6329, "step": 5245 }, { "epoch": 0.21990505433700472, "grad_norm": 1.6234934329986572, "learning_rate": 9.08375132077137e-06, "loss": 0.5362, "step": 5246 }, { "epoch": 0.21994697295201365, "grad_norm": 1.5746803283691406, "learning_rate": 9.08335959123797e-06, "loss": 0.5387, "step": 5247 }, { "epoch": 0.21998889156702262, "grad_norm": 1.7525891065597534, "learning_rate": 9.08296778643348e-06, "loss": 0.5373, "step": 5248 }, { "epoch": 0.22003081018203158, "grad_norm": 4.960565090179443, "learning_rate": 9.082575906365125e-06, "loss": 0.5599, "step": 5249 }, { "epoch": 0.22007272879704054, "grad_norm": 1.7378160953521729, "learning_rate": 9.082183951040127e-06, "loss": 0.5494, "step": 5250 }, { "epoch": 0.2201146474120495, "grad_norm": 1.5952008962631226, "learning_rate": 9.081791920465712e-06, "loss": 0.4943, "step": 5251 }, { "epoch": 0.22015656602705846, "grad_norm": 1.4409153461456299, "learning_rate": 9.081399814649104e-06, "loss": 0.5515, "step": 5252 }, { "epoch": 0.22019848464206743, "grad_norm": 1.6043274402618408, "learning_rate": 9.081007633597533e-06, "loss": 0.5339, "step": 5253 }, { "epoch": 0.2202404032570764, "grad_norm": 1.5936905145645142, "learning_rate": 9.08061537731823e-06, "loss": 0.6265, "step": 5254 }, { "epoch": 0.22028232187208535, "grad_norm": 1.6776821613311768, "learning_rate": 9.080223045818422e-06, "loss": 0.516, "step": 5255 }, { "epoch": 0.2203242404870943, "grad_norm": 1.747164249420166, "learning_rate": 9.079830639105344e-06, "loss": 0.5575, "step": 5256 }, { "epoch": 0.22036615910210328, "grad_norm": 1.5520498752593994, "learning_rate": 9.079438157186227e-06, "loss": 0.5125, "step": 5257 }, { "epoch": 0.22040807771711224, "grad_norm": 2.00256609916687, "learning_rate": 9.079045600068308e-06, "loss": 0.5989, "step": 5258 }, { "epoch": 0.22044999633212117, "grad_norm": 1.636024832725525, "learning_rate": 9.078652967758822e-06, "loss": 0.5514, "step": 5259 }, { "epoch": 0.22049191494713014, "grad_norm": 1.967239499092102, "learning_rate": 9.078260260265008e-06, "loss": 0.5668, "step": 5260 }, { "epoch": 0.2205338335621391, "grad_norm": 1.6595081090927124, "learning_rate": 9.077867477594104e-06, "loss": 0.5812, "step": 5261 }, { "epoch": 0.22057575217714806, "grad_norm": 1.945872187614441, "learning_rate": 9.07747461975335e-06, "loss": 0.5107, "step": 5262 }, { "epoch": 0.22061767079215702, "grad_norm": 1.5675901174545288, "learning_rate": 9.077081686749988e-06, "loss": 0.545, "step": 5263 }, { "epoch": 0.22065958940716598, "grad_norm": 1.733266830444336, "learning_rate": 9.076688678591263e-06, "loss": 0.51, "step": 5264 }, { "epoch": 0.22070150802217495, "grad_norm": 1.7892906665802002, "learning_rate": 9.076295595284416e-06, "loss": 0.545, "step": 5265 }, { "epoch": 0.2207434266371839, "grad_norm": 1.8215327262878418, "learning_rate": 9.075902436836698e-06, "loss": 0.562, "step": 5266 }, { "epoch": 0.22078534525219287, "grad_norm": 1.5805867910385132, "learning_rate": 9.07550920325535e-06, "loss": 0.5522, "step": 5267 }, { "epoch": 0.22082726386720183, "grad_norm": 1.8687702417373657, "learning_rate": 9.075115894547626e-06, "loss": 0.5436, "step": 5268 }, { "epoch": 0.2208691824822108, "grad_norm": 1.9553130865097046, "learning_rate": 9.074722510720775e-06, "loss": 0.5428, "step": 5269 }, { "epoch": 0.22091110109721976, "grad_norm": 1.641377568244934, "learning_rate": 9.074329051782046e-06, "loss": 0.5185, "step": 5270 }, { "epoch": 0.22095301971222872, "grad_norm": 1.6118483543395996, "learning_rate": 9.073935517738695e-06, "loss": 0.5581, "step": 5271 }, { "epoch": 0.22099493832723766, "grad_norm": 1.642230749130249, "learning_rate": 9.073541908597973e-06, "loss": 0.5578, "step": 5272 }, { "epoch": 0.22103685694224662, "grad_norm": 1.6713467836380005, "learning_rate": 9.073148224367138e-06, "loss": 0.5634, "step": 5273 }, { "epoch": 0.22107877555725558, "grad_norm": 1.6351852416992188, "learning_rate": 9.072754465053446e-06, "loss": 0.5403, "step": 5274 }, { "epoch": 0.22112069417226454, "grad_norm": 1.7682411670684814, "learning_rate": 9.072360630664156e-06, "loss": 0.5813, "step": 5275 }, { "epoch": 0.2211626127872735, "grad_norm": 1.7977757453918457, "learning_rate": 9.071966721206529e-06, "loss": 0.5355, "step": 5276 }, { "epoch": 0.22120453140228247, "grad_norm": 1.5706031322479248, "learning_rate": 9.071572736687824e-06, "loss": 0.5466, "step": 5277 }, { "epoch": 0.22124645001729143, "grad_norm": 1.7190648317337036, "learning_rate": 9.071178677115304e-06, "loss": 0.6032, "step": 5278 }, { "epoch": 0.2212883686323004, "grad_norm": 2.128809928894043, "learning_rate": 9.070784542496233e-06, "loss": 0.5213, "step": 5279 }, { "epoch": 0.22133028724730935, "grad_norm": 2.0156362056732178, "learning_rate": 9.070390332837878e-06, "loss": 0.6028, "step": 5280 }, { "epoch": 0.22137220586231832, "grad_norm": 1.6728179454803467, "learning_rate": 9.069996048147503e-06, "loss": 0.4921, "step": 5281 }, { "epoch": 0.22141412447732728, "grad_norm": 1.9030450582504272, "learning_rate": 9.069601688432379e-06, "loss": 0.5815, "step": 5282 }, { "epoch": 0.22145604309233624, "grad_norm": 1.7144123315811157, "learning_rate": 9.069207253699771e-06, "loss": 0.5816, "step": 5283 }, { "epoch": 0.22149796170734518, "grad_norm": 1.6088722944259644, "learning_rate": 9.068812743956955e-06, "loss": 0.5183, "step": 5284 }, { "epoch": 0.22153988032235414, "grad_norm": 1.6912862062454224, "learning_rate": 9.068418159211202e-06, "loss": 0.527, "step": 5285 }, { "epoch": 0.2215817989373631, "grad_norm": 1.5910179615020752, "learning_rate": 9.068023499469782e-06, "loss": 0.5485, "step": 5286 }, { "epoch": 0.22162371755237206, "grad_norm": 1.6700950860977173, "learning_rate": 9.067628764739976e-06, "loss": 0.5536, "step": 5287 }, { "epoch": 0.22166563616738102, "grad_norm": 1.5586026906967163, "learning_rate": 9.067233955029054e-06, "loss": 0.5504, "step": 5288 }, { "epoch": 0.22170755478239, "grad_norm": 1.9898205995559692, "learning_rate": 9.066839070344298e-06, "loss": 0.5372, "step": 5289 }, { "epoch": 0.22174947339739895, "grad_norm": 2.394357919692993, "learning_rate": 9.066444110692987e-06, "loss": 0.4741, "step": 5290 }, { "epoch": 0.2217913920124079, "grad_norm": 2.5104808807373047, "learning_rate": 9.0660490760824e-06, "loss": 0.5008, "step": 5291 }, { "epoch": 0.22183331062741687, "grad_norm": 1.7716188430786133, "learning_rate": 9.065653966519818e-06, "loss": 0.5295, "step": 5292 }, { "epoch": 0.22187522924242584, "grad_norm": 1.7912009954452515, "learning_rate": 9.065258782012526e-06, "loss": 0.5835, "step": 5293 }, { "epoch": 0.2219171478574348, "grad_norm": 1.7112122774124146, "learning_rate": 9.06486352256781e-06, "loss": 0.5241, "step": 5294 }, { "epoch": 0.22195906647244376, "grad_norm": 1.63303804397583, "learning_rate": 9.064468188192954e-06, "loss": 0.4763, "step": 5295 }, { "epoch": 0.22200098508745272, "grad_norm": 1.6284091472625732, "learning_rate": 9.064072778895244e-06, "loss": 0.4957, "step": 5296 }, { "epoch": 0.22204290370246166, "grad_norm": 1.988118290901184, "learning_rate": 9.063677294681971e-06, "loss": 0.5467, "step": 5297 }, { "epoch": 0.22208482231747062, "grad_norm": 1.6726725101470947, "learning_rate": 9.063281735560428e-06, "loss": 0.5095, "step": 5298 }, { "epoch": 0.22212674093247958, "grad_norm": 1.966814637184143, "learning_rate": 9.062886101537901e-06, "loss": 0.5919, "step": 5299 }, { "epoch": 0.22216865954748855, "grad_norm": 2.597898006439209, "learning_rate": 9.062490392621685e-06, "loss": 0.4977, "step": 5300 }, { "epoch": 0.2222105781624975, "grad_norm": 1.5643115043640137, "learning_rate": 9.062094608819076e-06, "loss": 0.4998, "step": 5301 }, { "epoch": 0.22225249677750647, "grad_norm": 1.636375904083252, "learning_rate": 9.061698750137368e-06, "loss": 0.5525, "step": 5302 }, { "epoch": 0.22229441539251543, "grad_norm": 1.5002179145812988, "learning_rate": 9.06130281658386e-06, "loss": 0.5459, "step": 5303 }, { "epoch": 0.2223363340075244, "grad_norm": 1.730696678161621, "learning_rate": 9.060906808165847e-06, "loss": 0.5644, "step": 5304 }, { "epoch": 0.22237825262253336, "grad_norm": 1.7350451946258545, "learning_rate": 9.060510724890635e-06, "loss": 0.4909, "step": 5305 }, { "epoch": 0.22242017123754232, "grad_norm": 1.4583889245986938, "learning_rate": 9.060114566765518e-06, "loss": 0.5094, "step": 5306 }, { "epoch": 0.22246208985255128, "grad_norm": 1.5341380834579468, "learning_rate": 9.059718333797802e-06, "loss": 0.5581, "step": 5307 }, { "epoch": 0.22250400846756024, "grad_norm": 1.7008440494537354, "learning_rate": 9.05932202599479e-06, "loss": 0.5843, "step": 5308 }, { "epoch": 0.22254592708256918, "grad_norm": 1.516749382019043, "learning_rate": 9.058925643363791e-06, "loss": 0.4494, "step": 5309 }, { "epoch": 0.22258784569757814, "grad_norm": 1.5316535234451294, "learning_rate": 9.058529185912108e-06, "loss": 0.5065, "step": 5310 }, { "epoch": 0.2226297643125871, "grad_norm": 1.7049903869628906, "learning_rate": 9.05813265364705e-06, "loss": 0.5654, "step": 5311 }, { "epoch": 0.22267168292759607, "grad_norm": 1.6089980602264404, "learning_rate": 9.057736046575928e-06, "loss": 0.5647, "step": 5312 }, { "epoch": 0.22271360154260503, "grad_norm": 1.9277592897415161, "learning_rate": 9.057339364706048e-06, "loss": 0.5362, "step": 5313 }, { "epoch": 0.222755520157614, "grad_norm": 1.8205089569091797, "learning_rate": 9.05694260804473e-06, "loss": 0.5459, "step": 5314 }, { "epoch": 0.22279743877262295, "grad_norm": 1.5691699981689453, "learning_rate": 9.056545776599281e-06, "loss": 0.535, "step": 5315 }, { "epoch": 0.22283935738763191, "grad_norm": 3.02871036529541, "learning_rate": 9.05614887037702e-06, "loss": 0.5139, "step": 5316 }, { "epoch": 0.22288127600264088, "grad_norm": 1.712282657623291, "learning_rate": 9.05575188938526e-06, "loss": 0.5304, "step": 5317 }, { "epoch": 0.22292319461764984, "grad_norm": 1.9059464931488037, "learning_rate": 9.055354833631323e-06, "loss": 0.5809, "step": 5318 }, { "epoch": 0.2229651132326588, "grad_norm": 1.773678183555603, "learning_rate": 9.054957703122525e-06, "loss": 0.5665, "step": 5319 }, { "epoch": 0.22300703184766776, "grad_norm": 1.9600895643234253, "learning_rate": 9.054560497866188e-06, "loss": 0.5423, "step": 5320 }, { "epoch": 0.22304895046267673, "grad_norm": 2.076404333114624, "learning_rate": 9.054163217869633e-06, "loss": 0.5534, "step": 5321 }, { "epoch": 0.22309086907768566, "grad_norm": 1.582128643989563, "learning_rate": 9.053765863140186e-06, "loss": 0.5625, "step": 5322 }, { "epoch": 0.22313278769269462, "grad_norm": 1.754996657371521, "learning_rate": 9.053368433685166e-06, "loss": 0.5487, "step": 5323 }, { "epoch": 0.22317470630770359, "grad_norm": 1.837766170501709, "learning_rate": 9.052970929511904e-06, "loss": 0.5725, "step": 5324 }, { "epoch": 0.22321662492271255, "grad_norm": 2.0414810180664062, "learning_rate": 9.052573350627726e-06, "loss": 0.4926, "step": 5325 }, { "epoch": 0.2232585435377215, "grad_norm": 1.7131946086883545, "learning_rate": 9.052175697039963e-06, "loss": 0.5701, "step": 5326 }, { "epoch": 0.22330046215273047, "grad_norm": 2.0678653717041016, "learning_rate": 9.05177796875594e-06, "loss": 0.578, "step": 5327 }, { "epoch": 0.22334238076773943, "grad_norm": 2.0569090843200684, "learning_rate": 9.051380165782992e-06, "loss": 0.5123, "step": 5328 }, { "epoch": 0.2233842993827484, "grad_norm": 1.885421872138977, "learning_rate": 9.050982288128451e-06, "loss": 0.5544, "step": 5329 }, { "epoch": 0.22342621799775736, "grad_norm": 1.6110575199127197, "learning_rate": 9.050584335799653e-06, "loss": 0.5692, "step": 5330 }, { "epoch": 0.22346813661276632, "grad_norm": 1.819028377532959, "learning_rate": 9.050186308803933e-06, "loss": 0.5689, "step": 5331 }, { "epoch": 0.22351005522777528, "grad_norm": 1.889149785041809, "learning_rate": 9.049788207148627e-06, "loss": 0.5602, "step": 5332 }, { "epoch": 0.22355197384278425, "grad_norm": 2.1570661067962646, "learning_rate": 9.049390030841074e-06, "loss": 0.5053, "step": 5333 }, { "epoch": 0.22359389245779318, "grad_norm": 2.056922435760498, "learning_rate": 9.048991779888615e-06, "loss": 0.5695, "step": 5334 }, { "epoch": 0.22363581107280214, "grad_norm": 1.7900813817977905, "learning_rate": 9.048593454298589e-06, "loss": 0.507, "step": 5335 }, { "epoch": 0.2236777296878111, "grad_norm": 3.3826165199279785, "learning_rate": 9.04819505407834e-06, "loss": 0.533, "step": 5336 }, { "epoch": 0.22371964830282007, "grad_norm": 1.5176231861114502, "learning_rate": 9.047796579235211e-06, "loss": 0.5188, "step": 5337 }, { "epoch": 0.22376156691782903, "grad_norm": 1.7711890935897827, "learning_rate": 9.047398029776547e-06, "loss": 0.5475, "step": 5338 }, { "epoch": 0.223803485532838, "grad_norm": 1.861695408821106, "learning_rate": 9.046999405709697e-06, "loss": 0.5925, "step": 5339 }, { "epoch": 0.22384540414784695, "grad_norm": 1.722988486289978, "learning_rate": 9.046600707042009e-06, "loss": 0.5402, "step": 5340 }, { "epoch": 0.22388732276285592, "grad_norm": 2.1343109607696533, "learning_rate": 9.04620193378083e-06, "loss": 0.5902, "step": 5341 }, { "epoch": 0.22392924137786488, "grad_norm": 1.796373963356018, "learning_rate": 9.04580308593351e-06, "loss": 0.5609, "step": 5342 }, { "epoch": 0.22397115999287384, "grad_norm": 1.9449753761291504, "learning_rate": 9.045404163507406e-06, "loss": 0.5314, "step": 5343 }, { "epoch": 0.2240130786078828, "grad_norm": 1.5664100646972656, "learning_rate": 9.045005166509868e-06, "loss": 0.5477, "step": 5344 }, { "epoch": 0.22405499722289177, "grad_norm": 1.782931923866272, "learning_rate": 9.044606094948251e-06, "loss": 0.5101, "step": 5345 }, { "epoch": 0.22409691583790073, "grad_norm": 1.8082696199417114, "learning_rate": 9.044206948829914e-06, "loss": 0.5187, "step": 5346 }, { "epoch": 0.22413883445290966, "grad_norm": 1.797568917274475, "learning_rate": 9.043807728162213e-06, "loss": 0.4998, "step": 5347 }, { "epoch": 0.22418075306791863, "grad_norm": 1.7539646625518799, "learning_rate": 9.043408432952504e-06, "loss": 0.5113, "step": 5348 }, { "epoch": 0.2242226716829276, "grad_norm": 1.8581489324569702, "learning_rate": 9.043009063208154e-06, "loss": 0.5334, "step": 5349 }, { "epoch": 0.22426459029793655, "grad_norm": 1.5854939222335815, "learning_rate": 9.042609618936519e-06, "loss": 0.5158, "step": 5350 }, { "epoch": 0.2243065089129455, "grad_norm": 1.99410080909729, "learning_rate": 9.042210100144967e-06, "loss": 0.5721, "step": 5351 }, { "epoch": 0.22434842752795447, "grad_norm": 1.7557806968688965, "learning_rate": 9.041810506840856e-06, "loss": 0.5381, "step": 5352 }, { "epoch": 0.22439034614296344, "grad_norm": 1.5949314832687378, "learning_rate": 9.041410839031557e-06, "loss": 0.5154, "step": 5353 }, { "epoch": 0.2244322647579724, "grad_norm": 1.8060660362243652, "learning_rate": 9.041011096724438e-06, "loss": 0.5477, "step": 5354 }, { "epoch": 0.22447418337298136, "grad_norm": 2.0244081020355225, "learning_rate": 9.040611279926865e-06, "loss": 0.5605, "step": 5355 }, { "epoch": 0.22451610198799032, "grad_norm": 1.5801093578338623, "learning_rate": 9.04021138864621e-06, "loss": 0.5135, "step": 5356 }, { "epoch": 0.2245580206029993, "grad_norm": 1.4557161331176758, "learning_rate": 9.039811422889842e-06, "loss": 0.5023, "step": 5357 }, { "epoch": 0.22459993921800825, "grad_norm": 1.7484184503555298, "learning_rate": 9.039411382665135e-06, "loss": 0.5326, "step": 5358 }, { "epoch": 0.22464185783301718, "grad_norm": 1.7873095273971558, "learning_rate": 9.039011267979465e-06, "loss": 0.5351, "step": 5359 }, { "epoch": 0.22468377644802615, "grad_norm": 1.9061659574508667, "learning_rate": 9.038611078840204e-06, "loss": 0.6194, "step": 5360 }, { "epoch": 0.2247256950630351, "grad_norm": 1.8909717798233032, "learning_rate": 9.038210815254733e-06, "loss": 0.5319, "step": 5361 }, { "epoch": 0.22476761367804407, "grad_norm": 3.3170242309570312, "learning_rate": 9.037810477230428e-06, "loss": 0.5813, "step": 5362 }, { "epoch": 0.22480953229305303, "grad_norm": 1.7666667699813843, "learning_rate": 9.03741006477467e-06, "loss": 0.5562, "step": 5363 }, { "epoch": 0.224851450908062, "grad_norm": 1.5460156202316284, "learning_rate": 9.037009577894836e-06, "loss": 0.5019, "step": 5364 }, { "epoch": 0.22489336952307096, "grad_norm": 1.6232000589370728, "learning_rate": 9.036609016598314e-06, "loss": 0.5287, "step": 5365 }, { "epoch": 0.22493528813807992, "grad_norm": 2.1197924613952637, "learning_rate": 9.036208380892483e-06, "loss": 0.5771, "step": 5366 }, { "epoch": 0.22497720675308888, "grad_norm": 1.865610957145691, "learning_rate": 9.035807670784734e-06, "loss": 0.5116, "step": 5367 }, { "epoch": 0.22501912536809784, "grad_norm": 1.589962124824524, "learning_rate": 9.035406886282445e-06, "loss": 0.4862, "step": 5368 }, { "epoch": 0.2250610439831068, "grad_norm": 1.7577029466629028, "learning_rate": 9.035006027393012e-06, "loss": 0.5644, "step": 5369 }, { "epoch": 0.22510296259811577, "grad_norm": 1.9078391790390015, "learning_rate": 9.034605094123821e-06, "loss": 0.5612, "step": 5370 }, { "epoch": 0.22514488121312473, "grad_norm": 1.8848673105239868, "learning_rate": 9.034204086482262e-06, "loss": 0.5726, "step": 5371 }, { "epoch": 0.22518679982813367, "grad_norm": 1.7235074043273926, "learning_rate": 9.033803004475727e-06, "loss": 0.563, "step": 5372 }, { "epoch": 0.22522871844314263, "grad_norm": 1.5625941753387451, "learning_rate": 9.033401848111612e-06, "loss": 0.5065, "step": 5373 }, { "epoch": 0.2252706370581516, "grad_norm": 1.9409189224243164, "learning_rate": 9.033000617397309e-06, "loss": 0.5606, "step": 5374 }, { "epoch": 0.22531255567316055, "grad_norm": 1.9841450452804565, "learning_rate": 9.032599312340216e-06, "loss": 0.5865, "step": 5375 }, { "epoch": 0.22535447428816952, "grad_norm": 2.156550407409668, "learning_rate": 9.032197932947728e-06, "loss": 0.5756, "step": 5376 }, { "epoch": 0.22539639290317848, "grad_norm": 1.9192513227462769, "learning_rate": 9.031796479227246e-06, "loss": 0.5488, "step": 5377 }, { "epoch": 0.22543831151818744, "grad_norm": 1.794340968132019, "learning_rate": 9.03139495118617e-06, "loss": 0.5668, "step": 5378 }, { "epoch": 0.2254802301331964, "grad_norm": 1.885408639907837, "learning_rate": 9.030993348831903e-06, "loss": 0.5421, "step": 5379 }, { "epoch": 0.22552214874820536, "grad_norm": 1.6090459823608398, "learning_rate": 9.030591672171843e-06, "loss": 0.5124, "step": 5380 }, { "epoch": 0.22556406736321433, "grad_norm": 1.6806933879852295, "learning_rate": 9.0301899212134e-06, "loss": 0.5387, "step": 5381 }, { "epoch": 0.2256059859782233, "grad_norm": 1.5747863054275513, "learning_rate": 9.029788095963976e-06, "loss": 0.4799, "step": 5382 }, { "epoch": 0.22564790459323225, "grad_norm": 1.6807352304458618, "learning_rate": 9.029386196430981e-06, "loss": 0.5137, "step": 5383 }, { "epoch": 0.22568982320824119, "grad_norm": 1.649457335472107, "learning_rate": 9.028984222621821e-06, "loss": 0.5479, "step": 5384 }, { "epoch": 0.22573174182325015, "grad_norm": 1.7404035329818726, "learning_rate": 9.028582174543906e-06, "loss": 0.5704, "step": 5385 }, { "epoch": 0.2257736604382591, "grad_norm": 1.5285066366195679, "learning_rate": 9.02818005220465e-06, "loss": 0.5779, "step": 5386 }, { "epoch": 0.22581557905326807, "grad_norm": 1.814410924911499, "learning_rate": 9.027777855611462e-06, "loss": 0.4836, "step": 5387 }, { "epoch": 0.22585749766827704, "grad_norm": 1.8978596925735474, "learning_rate": 9.027375584771758e-06, "loss": 0.5669, "step": 5388 }, { "epoch": 0.225899416283286, "grad_norm": 1.7628259658813477, "learning_rate": 9.026973239692951e-06, "loss": 0.5111, "step": 5389 }, { "epoch": 0.22594133489829496, "grad_norm": 1.6920185089111328, "learning_rate": 9.02657082038246e-06, "loss": 0.5837, "step": 5390 }, { "epoch": 0.22598325351330392, "grad_norm": 1.9603651762008667, "learning_rate": 9.026168326847704e-06, "loss": 0.582, "step": 5391 }, { "epoch": 0.22602517212831288, "grad_norm": 2.0316739082336426, "learning_rate": 9.025765759096101e-06, "loss": 0.524, "step": 5392 }, { "epoch": 0.22606709074332185, "grad_norm": 1.6945749521255493, "learning_rate": 9.02536311713507e-06, "loss": 0.5804, "step": 5393 }, { "epoch": 0.2261090093583308, "grad_norm": 1.746388554573059, "learning_rate": 9.024960400972038e-06, "loss": 0.5375, "step": 5394 }, { "epoch": 0.22615092797333977, "grad_norm": 1.7871472835540771, "learning_rate": 9.024557610614421e-06, "loss": 0.585, "step": 5395 }, { "epoch": 0.22619284658834873, "grad_norm": 1.9526783227920532, "learning_rate": 9.02415474606965e-06, "loss": 0.5723, "step": 5396 }, { "epoch": 0.22623476520335767, "grad_norm": 2.032522678375244, "learning_rate": 9.023751807345152e-06, "loss": 0.5646, "step": 5397 }, { "epoch": 0.22627668381836663, "grad_norm": 1.7109644412994385, "learning_rate": 9.02334879444835e-06, "loss": 0.5378, "step": 5398 }, { "epoch": 0.2263186024333756, "grad_norm": 1.7053006887435913, "learning_rate": 9.022945707386675e-06, "loss": 0.5301, "step": 5399 }, { "epoch": 0.22636052104838456, "grad_norm": 3.6572351455688477, "learning_rate": 9.022542546167558e-06, "loss": 0.5107, "step": 5400 }, { "epoch": 0.22640243966339352, "grad_norm": 1.6977320909500122, "learning_rate": 9.02213931079843e-06, "loss": 0.4991, "step": 5401 }, { "epoch": 0.22644435827840248, "grad_norm": 1.7599225044250488, "learning_rate": 9.021736001286723e-06, "loss": 0.5661, "step": 5402 }, { "epoch": 0.22648627689341144, "grad_norm": 1.5666677951812744, "learning_rate": 9.021332617639873e-06, "loss": 0.4821, "step": 5403 }, { "epoch": 0.2265281955084204, "grad_norm": 1.651477575302124, "learning_rate": 9.020929159865317e-06, "loss": 0.5764, "step": 5404 }, { "epoch": 0.22657011412342937, "grad_norm": 1.861542820930481, "learning_rate": 9.020525627970489e-06, "loss": 0.5705, "step": 5405 }, { "epoch": 0.22661203273843833, "grad_norm": 1.568177342414856, "learning_rate": 9.020122021962831e-06, "loss": 0.5521, "step": 5406 }, { "epoch": 0.2266539513534473, "grad_norm": 2.3102149963378906, "learning_rate": 9.019718341849779e-06, "loss": 0.5972, "step": 5407 }, { "epoch": 0.22669586996845625, "grad_norm": 1.7273974418640137, "learning_rate": 9.019314587638779e-06, "loss": 0.6107, "step": 5408 }, { "epoch": 0.2267377885834652, "grad_norm": 1.736729383468628, "learning_rate": 9.01891075933727e-06, "loss": 0.5281, "step": 5409 }, { "epoch": 0.22677970719847415, "grad_norm": 1.9610555171966553, "learning_rate": 9.018506856952696e-06, "loss": 0.5413, "step": 5410 }, { "epoch": 0.2268216258134831, "grad_norm": 1.8200780153274536, "learning_rate": 9.018102880492505e-06, "loss": 0.544, "step": 5411 }, { "epoch": 0.22686354442849208, "grad_norm": 1.8188496828079224, "learning_rate": 9.017698829964142e-06, "loss": 0.5556, "step": 5412 }, { "epoch": 0.22690546304350104, "grad_norm": 1.8410395383834839, "learning_rate": 9.017294705375055e-06, "loss": 0.5729, "step": 5413 }, { "epoch": 0.22694738165851, "grad_norm": 1.8892877101898193, "learning_rate": 9.016890506732695e-06, "loss": 0.5484, "step": 5414 }, { "epoch": 0.22698930027351896, "grad_norm": 1.7045975923538208, "learning_rate": 9.01648623404451e-06, "loss": 0.5796, "step": 5415 }, { "epoch": 0.22703121888852792, "grad_norm": 2.0773816108703613, "learning_rate": 9.016081887317954e-06, "loss": 0.6123, "step": 5416 }, { "epoch": 0.2270731375035369, "grad_norm": 1.4720442295074463, "learning_rate": 9.015677466560482e-06, "loss": 0.57, "step": 5417 }, { "epoch": 0.22711505611854585, "grad_norm": 1.7781782150268555, "learning_rate": 9.015272971779545e-06, "loss": 0.5338, "step": 5418 }, { "epoch": 0.2271569747335548, "grad_norm": 2.1429660320281982, "learning_rate": 9.014868402982604e-06, "loss": 0.5362, "step": 5419 }, { "epoch": 0.22719889334856377, "grad_norm": 1.5996452569961548, "learning_rate": 9.014463760177113e-06, "loss": 0.5353, "step": 5420 }, { "epoch": 0.22724081196357274, "grad_norm": 1.832199215888977, "learning_rate": 9.014059043370533e-06, "loss": 0.5869, "step": 5421 }, { "epoch": 0.22728273057858167, "grad_norm": 1.5700397491455078, "learning_rate": 9.013654252570324e-06, "loss": 0.5199, "step": 5422 }, { "epoch": 0.22732464919359063, "grad_norm": 1.7507412433624268, "learning_rate": 9.013249387783948e-06, "loss": 0.5863, "step": 5423 }, { "epoch": 0.2273665678085996, "grad_norm": 1.6960290670394897, "learning_rate": 9.012844449018867e-06, "loss": 0.5127, "step": 5424 }, { "epoch": 0.22740848642360856, "grad_norm": 1.7469027042388916, "learning_rate": 9.012439436282546e-06, "loss": 0.5694, "step": 5425 }, { "epoch": 0.22745040503861752, "grad_norm": 1.5112919807434082, "learning_rate": 9.012034349582451e-06, "loss": 0.4553, "step": 5426 }, { "epoch": 0.22749232365362648, "grad_norm": 1.7171217203140259, "learning_rate": 9.01162918892605e-06, "loss": 0.5824, "step": 5427 }, { "epoch": 0.22753424226863544, "grad_norm": 1.7969026565551758, "learning_rate": 9.01122395432081e-06, "loss": 0.5857, "step": 5428 }, { "epoch": 0.2275761608836444, "grad_norm": 1.6484363079071045, "learning_rate": 9.010818645774202e-06, "loss": 0.4901, "step": 5429 }, { "epoch": 0.22761807949865337, "grad_norm": 1.8123080730438232, "learning_rate": 9.010413263293696e-06, "loss": 0.5687, "step": 5430 }, { "epoch": 0.22765999811366233, "grad_norm": 4.458650588989258, "learning_rate": 9.010007806886768e-06, "loss": 0.5541, "step": 5431 }, { "epoch": 0.2277019167286713, "grad_norm": 2.076903820037842, "learning_rate": 9.009602276560889e-06, "loss": 0.5436, "step": 5432 }, { "epoch": 0.22774383534368026, "grad_norm": 1.6220020055770874, "learning_rate": 9.009196672323534e-06, "loss": 0.5352, "step": 5433 }, { "epoch": 0.2277857539586892, "grad_norm": 1.8986170291900635, "learning_rate": 9.008790994182181e-06, "loss": 0.5995, "step": 5434 }, { "epoch": 0.22782767257369815, "grad_norm": 1.5336329936981201, "learning_rate": 9.00838524214431e-06, "loss": 0.5615, "step": 5435 }, { "epoch": 0.22786959118870712, "grad_norm": 1.7013638019561768, "learning_rate": 9.007979416217396e-06, "loss": 0.5812, "step": 5436 }, { "epoch": 0.22791150980371608, "grad_norm": 1.7744636535644531, "learning_rate": 9.007573516408925e-06, "loss": 0.5552, "step": 5437 }, { "epoch": 0.22795342841872504, "grad_norm": 1.6583735942840576, "learning_rate": 9.007167542726373e-06, "loss": 0.5726, "step": 5438 }, { "epoch": 0.227995347033734, "grad_norm": 1.5825352668762207, "learning_rate": 9.00676149517723e-06, "loss": 0.5191, "step": 5439 }, { "epoch": 0.22803726564874297, "grad_norm": 1.5673890113830566, "learning_rate": 9.006355373768978e-06, "loss": 0.5358, "step": 5440 }, { "epoch": 0.22807918426375193, "grad_norm": 1.4860090017318726, "learning_rate": 9.0059491785091e-06, "loss": 0.5489, "step": 5441 }, { "epoch": 0.2281211028787609, "grad_norm": 1.7115405797958374, "learning_rate": 9.005542909405091e-06, "loss": 0.5654, "step": 5442 }, { "epoch": 0.22816302149376985, "grad_norm": 2.18196964263916, "learning_rate": 9.005136566464434e-06, "loss": 0.5047, "step": 5443 }, { "epoch": 0.22820494010877881, "grad_norm": 1.6091413497924805, "learning_rate": 9.00473014969462e-06, "loss": 0.5293, "step": 5444 }, { "epoch": 0.22824685872378778, "grad_norm": 2.0017271041870117, "learning_rate": 9.004323659103146e-06, "loss": 0.5433, "step": 5445 }, { "epoch": 0.22828877733879674, "grad_norm": 1.5779365301132202, "learning_rate": 9.0039170946975e-06, "loss": 0.5455, "step": 5446 }, { "epoch": 0.22833069595380567, "grad_norm": 1.4509385824203491, "learning_rate": 9.003510456485175e-06, "loss": 0.4788, "step": 5447 }, { "epoch": 0.22837261456881464, "grad_norm": 1.7490534782409668, "learning_rate": 9.003103744473671e-06, "loss": 0.5728, "step": 5448 }, { "epoch": 0.2284145331838236, "grad_norm": 1.7138745784759521, "learning_rate": 9.002696958670482e-06, "loss": 0.561, "step": 5449 }, { "epoch": 0.22845645179883256, "grad_norm": 1.618646502494812, "learning_rate": 9.00229009908311e-06, "loss": 0.5437, "step": 5450 }, { "epoch": 0.22849837041384152, "grad_norm": 2.812309980392456, "learning_rate": 9.001883165719053e-06, "loss": 0.5509, "step": 5451 }, { "epoch": 0.22854028902885049, "grad_norm": 1.6518604755401611, "learning_rate": 9.001476158585812e-06, "loss": 0.5299, "step": 5452 }, { "epoch": 0.22858220764385945, "grad_norm": 1.6873680353164673, "learning_rate": 9.00106907769089e-06, "loss": 0.5757, "step": 5453 }, { "epoch": 0.2286241262588684, "grad_norm": 1.7047374248504639, "learning_rate": 9.000661923041791e-06, "loss": 0.5815, "step": 5454 }, { "epoch": 0.22866604487387737, "grad_norm": 1.7307099103927612, "learning_rate": 9.00025469464602e-06, "loss": 0.5192, "step": 5455 }, { "epoch": 0.22870796348888633, "grad_norm": 1.8782424926757812, "learning_rate": 8.999847392511083e-06, "loss": 0.5282, "step": 5456 }, { "epoch": 0.2287498821038953, "grad_norm": 2.125825881958008, "learning_rate": 8.999440016644489e-06, "loss": 0.5784, "step": 5457 }, { "epoch": 0.22879180071890426, "grad_norm": 1.8556684255599976, "learning_rate": 8.999032567053749e-06, "loss": 0.5472, "step": 5458 }, { "epoch": 0.2288337193339132, "grad_norm": 1.962066650390625, "learning_rate": 8.99862504374637e-06, "loss": 0.4861, "step": 5459 }, { "epoch": 0.22887563794892216, "grad_norm": 1.9649467468261719, "learning_rate": 8.998217446729868e-06, "loss": 0.5166, "step": 5460 }, { "epoch": 0.22891755656393112, "grad_norm": 1.7404032945632935, "learning_rate": 8.997809776011754e-06, "loss": 0.5315, "step": 5461 }, { "epoch": 0.22895947517894008, "grad_norm": 1.4878859519958496, "learning_rate": 8.997402031599545e-06, "loss": 0.4848, "step": 5462 }, { "epoch": 0.22900139379394904, "grad_norm": 1.464295506477356, "learning_rate": 8.996994213500754e-06, "loss": 0.5617, "step": 5463 }, { "epoch": 0.229043312408958, "grad_norm": 1.6922900676727295, "learning_rate": 8.9965863217229e-06, "loss": 0.4859, "step": 5464 }, { "epoch": 0.22908523102396697, "grad_norm": 1.583168864250183, "learning_rate": 8.996178356273505e-06, "loss": 0.5493, "step": 5465 }, { "epoch": 0.22912714963897593, "grad_norm": 1.7181748151779175, "learning_rate": 8.995770317160083e-06, "loss": 0.5919, "step": 5466 }, { "epoch": 0.2291690682539849, "grad_norm": 1.4846588373184204, "learning_rate": 8.995362204390161e-06, "loss": 0.4811, "step": 5467 }, { "epoch": 0.22921098686899385, "grad_norm": 1.7473623752593994, "learning_rate": 8.99495401797126e-06, "loss": 0.5182, "step": 5468 }, { "epoch": 0.22925290548400282, "grad_norm": 1.5124238729476929, "learning_rate": 8.994545757910905e-06, "loss": 0.5039, "step": 5469 }, { "epoch": 0.22929482409901178, "grad_norm": 1.6796292066574097, "learning_rate": 8.994137424216621e-06, "loss": 0.5667, "step": 5470 }, { "epoch": 0.22933674271402074, "grad_norm": 1.4575822353363037, "learning_rate": 8.993729016895935e-06, "loss": 0.4701, "step": 5471 }, { "epoch": 0.22937866132902968, "grad_norm": 1.728867530822754, "learning_rate": 8.993320535956376e-06, "loss": 0.5269, "step": 5472 }, { "epoch": 0.22942057994403864, "grad_norm": 1.8785276412963867, "learning_rate": 8.992911981405473e-06, "loss": 0.5443, "step": 5473 }, { "epoch": 0.2294624985590476, "grad_norm": 1.7226688861846924, "learning_rate": 8.99250335325076e-06, "loss": 0.477, "step": 5474 }, { "epoch": 0.22950441717405656, "grad_norm": 1.6295185089111328, "learning_rate": 8.992094651499764e-06, "loss": 0.5874, "step": 5475 }, { "epoch": 0.22954633578906553, "grad_norm": 1.8866623640060425, "learning_rate": 8.991685876160023e-06, "loss": 0.5391, "step": 5476 }, { "epoch": 0.2295882544040745, "grad_norm": 1.716564416885376, "learning_rate": 8.991277027239072e-06, "loss": 0.4883, "step": 5477 }, { "epoch": 0.22963017301908345, "grad_norm": 1.5807180404663086, "learning_rate": 8.990868104744446e-06, "loss": 0.5818, "step": 5478 }, { "epoch": 0.2296720916340924, "grad_norm": 1.6614609956741333, "learning_rate": 8.990459108683683e-06, "loss": 0.5356, "step": 5479 }, { "epoch": 0.22971401024910137, "grad_norm": 1.5449119806289673, "learning_rate": 8.990050039064325e-06, "loss": 0.5048, "step": 5480 }, { "epoch": 0.22975592886411034, "grad_norm": 1.5608824491500854, "learning_rate": 8.98964089589391e-06, "loss": 0.5685, "step": 5481 }, { "epoch": 0.2297978474791193, "grad_norm": 1.8534330129623413, "learning_rate": 8.989231679179979e-06, "loss": 0.5353, "step": 5482 }, { "epoch": 0.22983976609412826, "grad_norm": 1.94913649559021, "learning_rate": 8.988822388930078e-06, "loss": 0.5595, "step": 5483 }, { "epoch": 0.22988168470913722, "grad_norm": 1.8368827104568481, "learning_rate": 8.988413025151751e-06, "loss": 0.5934, "step": 5484 }, { "epoch": 0.22992360332414616, "grad_norm": 2.0009231567382812, "learning_rate": 8.988003587852543e-06, "loss": 0.5877, "step": 5485 }, { "epoch": 0.22996552193915512, "grad_norm": 5.342905044555664, "learning_rate": 8.987594077040003e-06, "loss": 0.5691, "step": 5486 }, { "epoch": 0.23000744055416408, "grad_norm": 1.4298495054244995, "learning_rate": 8.987184492721678e-06, "loss": 0.4772, "step": 5487 }, { "epoch": 0.23004935916917305, "grad_norm": 1.7932207584381104, "learning_rate": 8.98677483490512e-06, "loss": 0.5439, "step": 5488 }, { "epoch": 0.230091277784182, "grad_norm": 2.4411487579345703, "learning_rate": 8.986365103597878e-06, "loss": 0.5823, "step": 5489 }, { "epoch": 0.23013319639919097, "grad_norm": 1.5503957271575928, "learning_rate": 8.985955298807506e-06, "loss": 0.5456, "step": 5490 }, { "epoch": 0.23017511501419993, "grad_norm": 5.538699626922607, "learning_rate": 8.985545420541562e-06, "loss": 0.5066, "step": 5491 }, { "epoch": 0.2302170336292089, "grad_norm": 1.6613861322402954, "learning_rate": 8.985135468807595e-06, "loss": 0.517, "step": 5492 }, { "epoch": 0.23025895224421786, "grad_norm": 1.803984522819519, "learning_rate": 8.984725443613164e-06, "loss": 0.4901, "step": 5493 }, { "epoch": 0.23030087085922682, "grad_norm": 1.7994270324707031, "learning_rate": 8.984315344965832e-06, "loss": 0.5649, "step": 5494 }, { "epoch": 0.23034278947423578, "grad_norm": 1.8432185649871826, "learning_rate": 8.983905172873152e-06, "loss": 0.545, "step": 5495 }, { "epoch": 0.23038470808924474, "grad_norm": 1.918002724647522, "learning_rate": 8.983494927342689e-06, "loss": 0.5264, "step": 5496 }, { "epoch": 0.23042662670425368, "grad_norm": 1.5410207509994507, "learning_rate": 8.983084608382002e-06, "loss": 0.509, "step": 5497 }, { "epoch": 0.23046854531926264, "grad_norm": 1.6060616970062256, "learning_rate": 8.982674215998658e-06, "loss": 0.5284, "step": 5498 }, { "epoch": 0.2305104639342716, "grad_norm": 1.670082688331604, "learning_rate": 8.982263750200221e-06, "loss": 0.5697, "step": 5499 }, { "epoch": 0.23055238254928057, "grad_norm": 2.107477903366089, "learning_rate": 8.981853210994256e-06, "loss": 0.5603, "step": 5500 }, { "epoch": 0.23059430116428953, "grad_norm": 1.8108062744140625, "learning_rate": 8.981442598388333e-06, "loss": 0.5405, "step": 5501 }, { "epoch": 0.2306362197792985, "grad_norm": 1.9528932571411133, "learning_rate": 8.981031912390019e-06, "loss": 0.5328, "step": 5502 }, { "epoch": 0.23067813839430745, "grad_norm": 1.9470409154891968, "learning_rate": 8.980621153006887e-06, "loss": 0.5628, "step": 5503 }, { "epoch": 0.23072005700931641, "grad_norm": 1.8336032629013062, "learning_rate": 8.980210320246505e-06, "loss": 0.5444, "step": 5504 }, { "epoch": 0.23076197562432538, "grad_norm": 1.5773268938064575, "learning_rate": 8.97979941411645e-06, "loss": 0.5363, "step": 5505 }, { "epoch": 0.23080389423933434, "grad_norm": 1.7274998426437378, "learning_rate": 8.979388434624292e-06, "loss": 0.5117, "step": 5506 }, { "epoch": 0.2308458128543433, "grad_norm": 1.7436575889587402, "learning_rate": 8.978977381777611e-06, "loss": 0.6049, "step": 5507 }, { "epoch": 0.23088773146935226, "grad_norm": 1.5199036598205566, "learning_rate": 8.978566255583985e-06, "loss": 0.5312, "step": 5508 }, { "epoch": 0.23092965008436123, "grad_norm": 1.7572901248931885, "learning_rate": 8.978155056050987e-06, "loss": 0.5789, "step": 5509 }, { "epoch": 0.23097156869937016, "grad_norm": 1.7256196737289429, "learning_rate": 8.977743783186202e-06, "loss": 0.5514, "step": 5510 }, { "epoch": 0.23101348731437912, "grad_norm": 1.9512492418289185, "learning_rate": 8.977332436997208e-06, "loss": 0.5886, "step": 5511 }, { "epoch": 0.23105540592938809, "grad_norm": 1.7439395189285278, "learning_rate": 8.97692101749159e-06, "loss": 0.5227, "step": 5512 }, { "epoch": 0.23109732454439705, "grad_norm": 1.8308416604995728, "learning_rate": 8.976509524676932e-06, "loss": 0.46, "step": 5513 }, { "epoch": 0.231139243159406, "grad_norm": 1.8404786586761475, "learning_rate": 8.976097958560817e-06, "loss": 0.5376, "step": 5514 }, { "epoch": 0.23118116177441497, "grad_norm": 1.6370429992675781, "learning_rate": 8.975686319150833e-06, "loss": 0.5155, "step": 5515 }, { "epoch": 0.23122308038942394, "grad_norm": 1.5548040866851807, "learning_rate": 8.975274606454568e-06, "loss": 0.4926, "step": 5516 }, { "epoch": 0.2312649990044329, "grad_norm": 1.811955451965332, "learning_rate": 8.97486282047961e-06, "loss": 0.5617, "step": 5517 }, { "epoch": 0.23130691761944186, "grad_norm": 1.6564253568649292, "learning_rate": 8.974450961233555e-06, "loss": 0.5833, "step": 5518 }, { "epoch": 0.23134883623445082, "grad_norm": 2.6887400150299072, "learning_rate": 8.974039028723986e-06, "loss": 0.5042, "step": 5519 }, { "epoch": 0.23139075484945978, "grad_norm": 1.8560547828674316, "learning_rate": 8.973627022958505e-06, "loss": 0.5772, "step": 5520 }, { "epoch": 0.23143267346446875, "grad_norm": 1.9316596984863281, "learning_rate": 8.973214943944701e-06, "loss": 0.5484, "step": 5521 }, { "epoch": 0.23147459207947768, "grad_norm": 1.5292742252349854, "learning_rate": 8.972802791690175e-06, "loss": 0.4767, "step": 5522 }, { "epoch": 0.23151651069448664, "grad_norm": 1.9504919052124023, "learning_rate": 8.97239056620252e-06, "loss": 0.499, "step": 5523 }, { "epoch": 0.2315584293094956, "grad_norm": 1.9278465509414673, "learning_rate": 8.971978267489337e-06, "loss": 0.5627, "step": 5524 }, { "epoch": 0.23160034792450457, "grad_norm": 2.0639572143554688, "learning_rate": 8.971565895558225e-06, "loss": 0.5362, "step": 5525 }, { "epoch": 0.23164226653951353, "grad_norm": 1.828550100326538, "learning_rate": 8.971153450416786e-06, "loss": 0.5617, "step": 5526 }, { "epoch": 0.2316841851545225, "grad_norm": 1.9023325443267822, "learning_rate": 8.970740932072624e-06, "loss": 0.5455, "step": 5527 }, { "epoch": 0.23172610376953146, "grad_norm": 1.5449827909469604, "learning_rate": 8.970328340533341e-06, "loss": 0.454, "step": 5528 }, { "epoch": 0.23176802238454042, "grad_norm": 1.71378755569458, "learning_rate": 8.969915675806546e-06, "loss": 0.542, "step": 5529 }, { "epoch": 0.23180994099954938, "grad_norm": 1.6037189960479736, "learning_rate": 8.969502937899841e-06, "loss": 0.4787, "step": 5530 }, { "epoch": 0.23185185961455834, "grad_norm": 2.0479354858398438, "learning_rate": 8.96909012682084e-06, "loss": 0.5025, "step": 5531 }, { "epoch": 0.2318937782295673, "grad_norm": 1.723615288734436, "learning_rate": 8.96867724257715e-06, "loss": 0.5957, "step": 5532 }, { "epoch": 0.23193569684457627, "grad_norm": 1.5964089632034302, "learning_rate": 8.968264285176378e-06, "loss": 0.5922, "step": 5533 }, { "epoch": 0.23197761545958523, "grad_norm": 1.9335525035858154, "learning_rate": 8.967851254626141e-06, "loss": 0.5549, "step": 5534 }, { "epoch": 0.23201953407459416, "grad_norm": 1.552903175354004, "learning_rate": 8.967438150934052e-06, "loss": 0.4858, "step": 5535 }, { "epoch": 0.23206145268960313, "grad_norm": 1.6301645040512085, "learning_rate": 8.967024974107727e-06, "loss": 0.5588, "step": 5536 }, { "epoch": 0.2321033713046121, "grad_norm": 1.7116000652313232, "learning_rate": 8.966611724154779e-06, "loss": 0.5566, "step": 5537 }, { "epoch": 0.23214528991962105, "grad_norm": 1.6746028661727905, "learning_rate": 8.966198401082828e-06, "loss": 0.5351, "step": 5538 }, { "epoch": 0.23218720853463, "grad_norm": 1.7401198148727417, "learning_rate": 8.965785004899491e-06, "loss": 0.4957, "step": 5539 }, { "epoch": 0.23222912714963898, "grad_norm": 1.7424695491790771, "learning_rate": 8.965371535612393e-06, "loss": 0.5082, "step": 5540 }, { "epoch": 0.23227104576464794, "grad_norm": 1.5689375400543213, "learning_rate": 8.964957993229152e-06, "loss": 0.529, "step": 5541 }, { "epoch": 0.2323129643796569, "grad_norm": 1.8745732307434082, "learning_rate": 8.964544377757391e-06, "loss": 0.5768, "step": 5542 }, { "epoch": 0.23235488299466586, "grad_norm": 1.5427268743515015, "learning_rate": 8.964130689204736e-06, "loss": 0.5497, "step": 5543 }, { "epoch": 0.23239680160967482, "grad_norm": 1.5272821187973022, "learning_rate": 8.96371692757881e-06, "loss": 0.5277, "step": 5544 }, { "epoch": 0.2324387202246838, "grad_norm": 3.7389581203460693, "learning_rate": 8.963303092887245e-06, "loss": 0.479, "step": 5545 }, { "epoch": 0.23248063883969275, "grad_norm": 1.8722422122955322, "learning_rate": 8.962889185137663e-06, "loss": 0.6004, "step": 5546 }, { "epoch": 0.23252255745470168, "grad_norm": 1.7856242656707764, "learning_rate": 8.9624752043377e-06, "loss": 0.5183, "step": 5547 }, { "epoch": 0.23256447606971065, "grad_norm": 1.7904438972473145, "learning_rate": 8.962061150494985e-06, "loss": 0.5992, "step": 5548 }, { "epoch": 0.2326063946847196, "grad_norm": 2.332467555999756, "learning_rate": 8.961647023617148e-06, "loss": 0.5512, "step": 5549 }, { "epoch": 0.23264831329972857, "grad_norm": 1.6833604574203491, "learning_rate": 8.961232823711827e-06, "loss": 0.5387, "step": 5550 }, { "epoch": 0.23269023191473753, "grad_norm": 1.5878753662109375, "learning_rate": 8.960818550786653e-06, "loss": 0.5181, "step": 5551 }, { "epoch": 0.2327321505297465, "grad_norm": 1.6388676166534424, "learning_rate": 8.960404204849266e-06, "loss": 0.5061, "step": 5552 }, { "epoch": 0.23277406914475546, "grad_norm": 2.2894043922424316, "learning_rate": 8.959989785907303e-06, "loss": 0.5895, "step": 5553 }, { "epoch": 0.23281598775976442, "grad_norm": 1.8185266256332397, "learning_rate": 8.959575293968402e-06, "loss": 0.5276, "step": 5554 }, { "epoch": 0.23285790637477338, "grad_norm": 1.6399729251861572, "learning_rate": 8.959160729040205e-06, "loss": 0.5606, "step": 5555 }, { "epoch": 0.23289982498978234, "grad_norm": 1.5185624361038208, "learning_rate": 8.958746091130352e-06, "loss": 0.5129, "step": 5556 }, { "epoch": 0.2329417436047913, "grad_norm": 2.015369176864624, "learning_rate": 8.95833138024649e-06, "loss": 0.5545, "step": 5557 }, { "epoch": 0.23298366221980027, "grad_norm": 1.6043096780776978, "learning_rate": 8.957916596396259e-06, "loss": 0.5121, "step": 5558 }, { "epoch": 0.23302558083480923, "grad_norm": 1.712759256362915, "learning_rate": 8.957501739587309e-06, "loss": 0.5433, "step": 5559 }, { "epoch": 0.23306749944981817, "grad_norm": 1.7294739484786987, "learning_rate": 8.957086809827285e-06, "loss": 0.5429, "step": 5560 }, { "epoch": 0.23310941806482713, "grad_norm": 1.9060475826263428, "learning_rate": 8.956671807123834e-06, "loss": 0.5525, "step": 5561 }, { "epoch": 0.2331513366798361, "grad_norm": 1.8145610094070435, "learning_rate": 8.956256731484611e-06, "loss": 0.5631, "step": 5562 }, { "epoch": 0.23319325529484505, "grad_norm": 1.851883888244629, "learning_rate": 8.955841582917263e-06, "loss": 0.576, "step": 5563 }, { "epoch": 0.23323517390985402, "grad_norm": 1.8008874654769897, "learning_rate": 8.955426361429445e-06, "loss": 0.5162, "step": 5564 }, { "epoch": 0.23327709252486298, "grad_norm": 1.7099965810775757, "learning_rate": 8.955011067028809e-06, "loss": 0.5405, "step": 5565 }, { "epoch": 0.23331901113987194, "grad_norm": 1.6791064739227295, "learning_rate": 8.954595699723012e-06, "loss": 0.5733, "step": 5566 }, { "epoch": 0.2333609297548809, "grad_norm": 1.6916959285736084, "learning_rate": 8.954180259519711e-06, "loss": 0.5037, "step": 5567 }, { "epoch": 0.23340284836988986, "grad_norm": 1.748157024383545, "learning_rate": 8.953764746426564e-06, "loss": 0.5764, "step": 5568 }, { "epoch": 0.23344476698489883, "grad_norm": 1.8016959428787231, "learning_rate": 8.953349160451229e-06, "loss": 0.5925, "step": 5569 }, { "epoch": 0.2334866855999078, "grad_norm": 1.4199308156967163, "learning_rate": 8.952933501601367e-06, "loss": 0.5539, "step": 5570 }, { "epoch": 0.23352860421491675, "grad_norm": 1.6426273584365845, "learning_rate": 8.952517769884642e-06, "loss": 0.4861, "step": 5571 }, { "epoch": 0.2335705228299257, "grad_norm": 1.804550051689148, "learning_rate": 8.952101965308714e-06, "loss": 0.5889, "step": 5572 }, { "epoch": 0.23361244144493465, "grad_norm": 1.7930538654327393, "learning_rate": 8.951686087881253e-06, "loss": 0.5591, "step": 5573 }, { "epoch": 0.2336543600599436, "grad_norm": 2.1579818725585938, "learning_rate": 8.95127013760992e-06, "loss": 0.5557, "step": 5574 }, { "epoch": 0.23369627867495257, "grad_norm": 1.5501735210418701, "learning_rate": 8.950854114502385e-06, "loss": 0.5249, "step": 5575 }, { "epoch": 0.23373819728996154, "grad_norm": 1.5460753440856934, "learning_rate": 8.950438018566316e-06, "loss": 0.5913, "step": 5576 }, { "epoch": 0.2337801159049705, "grad_norm": 2.1920430660247803, "learning_rate": 8.950021849809382e-06, "loss": 0.5238, "step": 5577 }, { "epoch": 0.23382203451997946, "grad_norm": 1.9452571868896484, "learning_rate": 8.949605608239259e-06, "loss": 0.5065, "step": 5578 }, { "epoch": 0.23386395313498842, "grad_norm": 2.575004816055298, "learning_rate": 8.949189293863614e-06, "loss": 0.5712, "step": 5579 }, { "epoch": 0.23390587174999738, "grad_norm": 1.6172438859939575, "learning_rate": 8.948772906690126e-06, "loss": 0.5361, "step": 5580 }, { "epoch": 0.23394779036500635, "grad_norm": 1.7097724676132202, "learning_rate": 8.948356446726467e-06, "loss": 0.5145, "step": 5581 }, { "epoch": 0.2339897089800153, "grad_norm": 1.5498751401901245, "learning_rate": 8.947939913980315e-06, "loss": 0.5011, "step": 5582 }, { "epoch": 0.23403162759502427, "grad_norm": 1.999492883682251, "learning_rate": 8.94752330845935e-06, "loss": 0.573, "step": 5583 }, { "epoch": 0.23407354621003323, "grad_norm": 1.740817904472351, "learning_rate": 8.94710663017125e-06, "loss": 0.5306, "step": 5584 }, { "epoch": 0.23411546482504217, "grad_norm": 1.492738127708435, "learning_rate": 8.946689879123695e-06, "loss": 0.5197, "step": 5585 }, { "epoch": 0.23415738344005113, "grad_norm": 1.6449230909347534, "learning_rate": 8.946273055324369e-06, "loss": 0.5058, "step": 5586 }, { "epoch": 0.2341993020550601, "grad_norm": 1.8454716205596924, "learning_rate": 8.945856158780954e-06, "loss": 0.5437, "step": 5587 }, { "epoch": 0.23424122067006906, "grad_norm": 1.6345423460006714, "learning_rate": 8.945439189501137e-06, "loss": 0.6065, "step": 5588 }, { "epoch": 0.23428313928507802, "grad_norm": 1.77861487865448, "learning_rate": 8.945022147492602e-06, "loss": 0.5565, "step": 5589 }, { "epoch": 0.23432505790008698, "grad_norm": 1.9662564992904663, "learning_rate": 8.944605032763039e-06, "loss": 0.53, "step": 5590 }, { "epoch": 0.23436697651509594, "grad_norm": 1.880191445350647, "learning_rate": 8.944187845320134e-06, "loss": 0.5629, "step": 5591 }, { "epoch": 0.2344088951301049, "grad_norm": 1.4898000955581665, "learning_rate": 8.943770585171579e-06, "loss": 0.5373, "step": 5592 }, { "epoch": 0.23445081374511387, "grad_norm": 1.6052361726760864, "learning_rate": 8.943353252325065e-06, "loss": 0.5439, "step": 5593 }, { "epoch": 0.23449273236012283, "grad_norm": 1.6982083320617676, "learning_rate": 8.942935846788285e-06, "loss": 0.5046, "step": 5594 }, { "epoch": 0.2345346509751318, "grad_norm": 1.7862541675567627, "learning_rate": 8.942518368568935e-06, "loss": 0.5676, "step": 5595 }, { "epoch": 0.23457656959014075, "grad_norm": 1.9061752557754517, "learning_rate": 8.942100817674708e-06, "loss": 0.543, "step": 5596 }, { "epoch": 0.2346184882051497, "grad_norm": 2.213665246963501, "learning_rate": 8.941683194113303e-06, "loss": 0.5687, "step": 5597 }, { "epoch": 0.23466040682015865, "grad_norm": 1.8966569900512695, "learning_rate": 8.941265497892417e-06, "loss": 0.5342, "step": 5598 }, { "epoch": 0.2347023254351676, "grad_norm": 1.655647873878479, "learning_rate": 8.940847729019751e-06, "loss": 0.5485, "step": 5599 }, { "epoch": 0.23474424405017658, "grad_norm": 2.430546760559082, "learning_rate": 8.940429887503005e-06, "loss": 0.5349, "step": 5600 }, { "epoch": 0.23478616266518554, "grad_norm": 1.6615886688232422, "learning_rate": 8.94001197334988e-06, "loss": 0.5571, "step": 5601 }, { "epoch": 0.2348280812801945, "grad_norm": 2.590925455093384, "learning_rate": 8.939593986568082e-06, "loss": 0.6267, "step": 5602 }, { "epoch": 0.23486999989520346, "grad_norm": 1.6914396286010742, "learning_rate": 8.939175927165317e-06, "loss": 0.5649, "step": 5603 }, { "epoch": 0.23491191851021243, "grad_norm": 1.827444076538086, "learning_rate": 8.938757795149288e-06, "loss": 0.4804, "step": 5604 }, { "epoch": 0.2349538371252214, "grad_norm": 1.8746649026870728, "learning_rate": 8.938339590527704e-06, "loss": 0.552, "step": 5605 }, { "epoch": 0.23499575574023035, "grad_norm": 1.491266131401062, "learning_rate": 8.937921313308274e-06, "loss": 0.5695, "step": 5606 }, { "epoch": 0.2350376743552393, "grad_norm": 1.81404447555542, "learning_rate": 8.93750296349871e-06, "loss": 0.5563, "step": 5607 }, { "epoch": 0.23507959297024827, "grad_norm": 2.2512991428375244, "learning_rate": 8.937084541106721e-06, "loss": 0.5473, "step": 5608 }, { "epoch": 0.23512151158525724, "grad_norm": 1.7343008518218994, "learning_rate": 8.936666046140022e-06, "loss": 0.5435, "step": 5609 }, { "epoch": 0.23516343020026617, "grad_norm": 1.920077919960022, "learning_rate": 8.936247478606325e-06, "loss": 0.5327, "step": 5610 }, { "epoch": 0.23520534881527513, "grad_norm": 2.0747179985046387, "learning_rate": 8.93582883851335e-06, "loss": 0.5223, "step": 5611 }, { "epoch": 0.2352472674302841, "grad_norm": 1.8331851959228516, "learning_rate": 8.935410125868811e-06, "loss": 0.541, "step": 5612 }, { "epoch": 0.23528918604529306, "grad_norm": 2.397773265838623, "learning_rate": 8.934991340680426e-06, "loss": 0.565, "step": 5613 }, { "epoch": 0.23533110466030202, "grad_norm": 2.0033211708068848, "learning_rate": 8.934572482955916e-06, "loss": 0.516, "step": 5614 }, { "epoch": 0.23537302327531098, "grad_norm": 1.7739217281341553, "learning_rate": 8.934153552703001e-06, "loss": 0.5046, "step": 5615 }, { "epoch": 0.23541494189031995, "grad_norm": 1.7057201862335205, "learning_rate": 8.933734549929406e-06, "loss": 0.542, "step": 5616 }, { "epoch": 0.2354568605053289, "grad_norm": 2.4101672172546387, "learning_rate": 8.93331547464285e-06, "loss": 0.565, "step": 5617 }, { "epoch": 0.23549877912033787, "grad_norm": 1.7249165773391724, "learning_rate": 8.932896326851062e-06, "loss": 0.5451, "step": 5618 }, { "epoch": 0.23554069773534683, "grad_norm": 1.6344636678695679, "learning_rate": 8.932477106561768e-06, "loss": 0.528, "step": 5619 }, { "epoch": 0.2355826163503558, "grad_norm": 1.6275107860565186, "learning_rate": 8.932057813782694e-06, "loss": 0.5248, "step": 5620 }, { "epoch": 0.23562453496536476, "grad_norm": 1.7584350109100342, "learning_rate": 8.93163844852157e-06, "loss": 0.545, "step": 5621 }, { "epoch": 0.2356664535803737, "grad_norm": 1.779869794845581, "learning_rate": 8.931219010786128e-06, "loss": 0.533, "step": 5622 }, { "epoch": 0.23570837219538265, "grad_norm": 1.568975806236267, "learning_rate": 8.930799500584098e-06, "loss": 0.5567, "step": 5623 }, { "epoch": 0.23575029081039162, "grad_norm": 1.5915273427963257, "learning_rate": 8.93037991792321e-06, "loss": 0.5351, "step": 5624 }, { "epoch": 0.23579220942540058, "grad_norm": 1.7308329343795776, "learning_rate": 8.929960262811206e-06, "loss": 0.5422, "step": 5625 }, { "epoch": 0.23583412804040954, "grad_norm": 1.6107062101364136, "learning_rate": 8.929540535255816e-06, "loss": 0.5105, "step": 5626 }, { "epoch": 0.2358760466554185, "grad_norm": 1.8590337038040161, "learning_rate": 8.92912073526478e-06, "loss": 0.5523, "step": 5627 }, { "epoch": 0.23591796527042747, "grad_norm": 1.895532250404358, "learning_rate": 8.928700862845834e-06, "loss": 0.5621, "step": 5628 }, { "epoch": 0.23595988388543643, "grad_norm": 1.6505852937698364, "learning_rate": 8.928280918006718e-06, "loss": 0.4895, "step": 5629 }, { "epoch": 0.2360018025004454, "grad_norm": 1.9504873752593994, "learning_rate": 8.927860900755174e-06, "loss": 0.5415, "step": 5630 }, { "epoch": 0.23604372111545435, "grad_norm": 1.630763053894043, "learning_rate": 8.927440811098946e-06, "loss": 0.5154, "step": 5631 }, { "epoch": 0.23608563973046331, "grad_norm": 1.6764601469039917, "learning_rate": 8.927020649045776e-06, "loss": 0.4875, "step": 5632 }, { "epoch": 0.23612755834547228, "grad_norm": 1.6381239891052246, "learning_rate": 8.926600414603409e-06, "loss": 0.5491, "step": 5633 }, { "epoch": 0.23616947696048124, "grad_norm": 1.623436450958252, "learning_rate": 8.92618010777959e-06, "loss": 0.5405, "step": 5634 }, { "epoch": 0.23621139557549017, "grad_norm": 1.7868518829345703, "learning_rate": 8.92575972858207e-06, "loss": 0.544, "step": 5635 }, { "epoch": 0.23625331419049914, "grad_norm": 1.7836412191390991, "learning_rate": 8.925339277018596e-06, "loss": 0.5407, "step": 5636 }, { "epoch": 0.2362952328055081, "grad_norm": 1.8267415761947632, "learning_rate": 8.924918753096921e-06, "loss": 0.5067, "step": 5637 }, { "epoch": 0.23633715142051706, "grad_norm": 2.0034847259521484, "learning_rate": 8.924498156824793e-06, "loss": 0.5607, "step": 5638 }, { "epoch": 0.23637907003552602, "grad_norm": 2.2839527130126953, "learning_rate": 8.924077488209967e-06, "loss": 0.5555, "step": 5639 }, { "epoch": 0.23642098865053499, "grad_norm": 1.9148329496383667, "learning_rate": 8.923656747260196e-06, "loss": 0.5324, "step": 5640 }, { "epoch": 0.23646290726554395, "grad_norm": 1.8422411680221558, "learning_rate": 8.92323593398324e-06, "loss": 0.5172, "step": 5641 }, { "epoch": 0.2365048258805529, "grad_norm": 1.796950340270996, "learning_rate": 8.922815048386852e-06, "loss": 0.5006, "step": 5642 }, { "epoch": 0.23654674449556187, "grad_norm": 1.699115514755249, "learning_rate": 8.922394090478792e-06, "loss": 0.5426, "step": 5643 }, { "epoch": 0.23658866311057083, "grad_norm": 1.8153263330459595, "learning_rate": 8.92197306026682e-06, "loss": 0.5652, "step": 5644 }, { "epoch": 0.2366305817255798, "grad_norm": 1.4691716432571411, "learning_rate": 8.921551957758694e-06, "loss": 0.4672, "step": 5645 }, { "epoch": 0.23667250034058876, "grad_norm": 1.708858847618103, "learning_rate": 8.921130782962181e-06, "loss": 0.5288, "step": 5646 }, { "epoch": 0.2367144189555977, "grad_norm": 1.8293200731277466, "learning_rate": 8.920709535885044e-06, "loss": 0.5755, "step": 5647 }, { "epoch": 0.23675633757060666, "grad_norm": 1.899106502532959, "learning_rate": 8.920288216535044e-06, "loss": 0.5272, "step": 5648 }, { "epoch": 0.23679825618561562, "grad_norm": 2.0193793773651123, "learning_rate": 8.919866824919952e-06, "loss": 0.6034, "step": 5649 }, { "epoch": 0.23684017480062458, "grad_norm": 1.7868142127990723, "learning_rate": 8.919445361047536e-06, "loss": 0.536, "step": 5650 }, { "epoch": 0.23688209341563354, "grad_norm": 1.6782091856002808, "learning_rate": 8.91902382492556e-06, "loss": 0.4985, "step": 5651 }, { "epoch": 0.2369240120306425, "grad_norm": 1.664678692817688, "learning_rate": 8.9186022165618e-06, "loss": 0.5555, "step": 5652 }, { "epoch": 0.23696593064565147, "grad_norm": 1.9762696027755737, "learning_rate": 8.918180535964023e-06, "loss": 0.5159, "step": 5653 }, { "epoch": 0.23700784926066043, "grad_norm": 1.555351972579956, "learning_rate": 8.917758783140007e-06, "loss": 0.5704, "step": 5654 }, { "epoch": 0.2370497678756694, "grad_norm": 1.658701777458191, "learning_rate": 8.917336958097522e-06, "loss": 0.5217, "step": 5655 }, { "epoch": 0.23709168649067835, "grad_norm": 1.8898457288742065, "learning_rate": 8.916915060844347e-06, "loss": 0.5, "step": 5656 }, { "epoch": 0.23713360510568732, "grad_norm": 2.0661373138427734, "learning_rate": 8.91649309138826e-06, "loss": 0.5406, "step": 5657 }, { "epoch": 0.23717552372069628, "grad_norm": 2.647202968597412, "learning_rate": 8.916071049737033e-06, "loss": 0.6101, "step": 5658 }, { "epoch": 0.23721744233570524, "grad_norm": 1.524696707725525, "learning_rate": 8.915648935898451e-06, "loss": 0.5581, "step": 5659 }, { "epoch": 0.23725936095071418, "grad_norm": 1.9574955701828003, "learning_rate": 8.915226749880294e-06, "loss": 0.5464, "step": 5660 }, { "epoch": 0.23730127956572314, "grad_norm": 2.730168104171753, "learning_rate": 8.914804491690346e-06, "loss": 0.5181, "step": 5661 }, { "epoch": 0.2373431981807321, "grad_norm": 1.723955750465393, "learning_rate": 8.91438216133639e-06, "loss": 0.5901, "step": 5662 }, { "epoch": 0.23738511679574106, "grad_norm": 1.6472212076187134, "learning_rate": 8.913959758826208e-06, "loss": 0.5003, "step": 5663 }, { "epoch": 0.23742703541075003, "grad_norm": 1.849069595336914, "learning_rate": 8.91353728416759e-06, "loss": 0.5807, "step": 5664 }, { "epoch": 0.237468954025759, "grad_norm": 1.7023131847381592, "learning_rate": 8.913114737368323e-06, "loss": 0.5471, "step": 5665 }, { "epoch": 0.23751087264076795, "grad_norm": 1.9198540449142456, "learning_rate": 8.912692118436194e-06, "loss": 0.5774, "step": 5666 }, { "epoch": 0.2375527912557769, "grad_norm": 1.7365010976791382, "learning_rate": 8.912269427378996e-06, "loss": 0.5363, "step": 5667 }, { "epoch": 0.23759470987078588, "grad_norm": 1.5386618375778198, "learning_rate": 8.911846664204518e-06, "loss": 0.4825, "step": 5668 }, { "epoch": 0.23763662848579484, "grad_norm": 1.555417776107788, "learning_rate": 8.911423828920557e-06, "loss": 0.5513, "step": 5669 }, { "epoch": 0.2376785471008038, "grad_norm": 1.611601710319519, "learning_rate": 8.911000921534904e-06, "loss": 0.5069, "step": 5670 }, { "epoch": 0.23772046571581276, "grad_norm": 1.6685541868209839, "learning_rate": 8.910577942055356e-06, "loss": 0.5774, "step": 5671 }, { "epoch": 0.2377623843308217, "grad_norm": 1.5089540481567383, "learning_rate": 8.910154890489709e-06, "loss": 0.4921, "step": 5672 }, { "epoch": 0.23780430294583066, "grad_norm": 1.7375084161758423, "learning_rate": 8.909731766845763e-06, "loss": 0.5293, "step": 5673 }, { "epoch": 0.23784622156083962, "grad_norm": 2.534407615661621, "learning_rate": 8.909308571131317e-06, "loss": 0.5967, "step": 5674 }, { "epoch": 0.23788814017584858, "grad_norm": 1.9772957563400269, "learning_rate": 8.908885303354172e-06, "loss": 0.6629, "step": 5675 }, { "epoch": 0.23793005879085755, "grad_norm": 1.7916309833526611, "learning_rate": 8.90846196352213e-06, "loss": 0.5298, "step": 5676 }, { "epoch": 0.2379719774058665, "grad_norm": 2.0781116485595703, "learning_rate": 8.908038551642996e-06, "loss": 0.5438, "step": 5677 }, { "epoch": 0.23801389602087547, "grad_norm": 2.0222158432006836, "learning_rate": 8.907615067724574e-06, "loss": 0.6158, "step": 5678 }, { "epoch": 0.23805581463588443, "grad_norm": 1.711959719657898, "learning_rate": 8.90719151177467e-06, "loss": 0.6094, "step": 5679 }, { "epoch": 0.2380977332508934, "grad_norm": 1.8187252283096313, "learning_rate": 8.90676788380109e-06, "loss": 0.5549, "step": 5680 }, { "epoch": 0.23813965186590236, "grad_norm": 1.6352951526641846, "learning_rate": 8.906344183811645e-06, "loss": 0.5518, "step": 5681 }, { "epoch": 0.23818157048091132, "grad_norm": 1.6526691913604736, "learning_rate": 8.905920411814148e-06, "loss": 0.5611, "step": 5682 }, { "epoch": 0.23822348909592028, "grad_norm": 1.649233341217041, "learning_rate": 8.905496567816407e-06, "loss": 0.4637, "step": 5683 }, { "epoch": 0.23826540771092924, "grad_norm": 2.6418118476867676, "learning_rate": 8.905072651826235e-06, "loss": 0.5208, "step": 5684 }, { "epoch": 0.23830732632593818, "grad_norm": 1.6943520307540894, "learning_rate": 8.904648663851448e-06, "loss": 0.4945, "step": 5685 }, { "epoch": 0.23834924494094714, "grad_norm": 1.8166457414627075, "learning_rate": 8.904224603899862e-06, "loss": 0.5423, "step": 5686 }, { "epoch": 0.2383911635559561, "grad_norm": 1.762115716934204, "learning_rate": 8.903800471979292e-06, "loss": 0.6022, "step": 5687 }, { "epoch": 0.23843308217096507, "grad_norm": 1.785834789276123, "learning_rate": 8.903376268097557e-06, "loss": 0.5212, "step": 5688 }, { "epoch": 0.23847500078597403, "grad_norm": 1.6968579292297363, "learning_rate": 8.902951992262476e-06, "loss": 0.5798, "step": 5689 }, { "epoch": 0.238516919400983, "grad_norm": 2.054600238800049, "learning_rate": 8.902527644481871e-06, "loss": 0.5979, "step": 5690 }, { "epoch": 0.23855883801599195, "grad_norm": 1.7859200239181519, "learning_rate": 8.902103224763566e-06, "loss": 0.4816, "step": 5691 }, { "epoch": 0.23860075663100092, "grad_norm": 2.156280755996704, "learning_rate": 8.901678733115379e-06, "loss": 0.584, "step": 5692 }, { "epoch": 0.23864267524600988, "grad_norm": 1.805816411972046, "learning_rate": 8.901254169545142e-06, "loss": 0.495, "step": 5693 }, { "epoch": 0.23868459386101884, "grad_norm": 1.7881473302841187, "learning_rate": 8.900829534060676e-06, "loss": 0.5764, "step": 5694 }, { "epoch": 0.2387265124760278, "grad_norm": 2.172227621078491, "learning_rate": 8.900404826669809e-06, "loss": 0.5931, "step": 5695 }, { "epoch": 0.23876843109103676, "grad_norm": 1.694334864616394, "learning_rate": 8.899980047380372e-06, "loss": 0.5666, "step": 5696 }, { "epoch": 0.2388103497060457, "grad_norm": 1.752646565437317, "learning_rate": 8.899555196200195e-06, "loss": 0.5014, "step": 5697 }, { "epoch": 0.23885226832105466, "grad_norm": 1.5573064088821411, "learning_rate": 8.899130273137108e-06, "loss": 0.5206, "step": 5698 }, { "epoch": 0.23889418693606362, "grad_norm": 1.4893826246261597, "learning_rate": 8.898705278198948e-06, "loss": 0.5306, "step": 5699 }, { "epoch": 0.2389361055510726, "grad_norm": 2.234250068664551, "learning_rate": 8.898280211393541e-06, "loss": 0.5219, "step": 5700 }, { "epoch": 0.23897802416608155, "grad_norm": 1.557023525238037, "learning_rate": 8.897855072728731e-06, "loss": 0.5727, "step": 5701 }, { "epoch": 0.2390199427810905, "grad_norm": 1.4419429302215576, "learning_rate": 8.897429862212352e-06, "loss": 0.4883, "step": 5702 }, { "epoch": 0.23906186139609947, "grad_norm": 2.0683233737945557, "learning_rate": 8.897004579852239e-06, "loss": 0.5912, "step": 5703 }, { "epoch": 0.23910378001110844, "grad_norm": 1.7080895900726318, "learning_rate": 8.896579225656235e-06, "loss": 0.538, "step": 5704 }, { "epoch": 0.2391456986261174, "grad_norm": 1.870423436164856, "learning_rate": 8.89615379963218e-06, "loss": 0.5294, "step": 5705 }, { "epoch": 0.23918761724112636, "grad_norm": 1.6936880350112915, "learning_rate": 8.895728301787915e-06, "loss": 0.5656, "step": 5706 }, { "epoch": 0.23922953585613532, "grad_norm": 1.8515321016311646, "learning_rate": 8.895302732131286e-06, "loss": 0.5516, "step": 5707 }, { "epoch": 0.23927145447114428, "grad_norm": 1.5138428211212158, "learning_rate": 8.894877090670136e-06, "loss": 0.5179, "step": 5708 }, { "epoch": 0.23931337308615325, "grad_norm": 1.8422118425369263, "learning_rate": 8.89445137741231e-06, "loss": 0.5944, "step": 5709 }, { "epoch": 0.23935529170116218, "grad_norm": 1.5479406118392944, "learning_rate": 8.894025592365658e-06, "loss": 0.5016, "step": 5710 }, { "epoch": 0.23939721031617114, "grad_norm": 1.8318657875061035, "learning_rate": 8.893599735538027e-06, "loss": 0.5173, "step": 5711 }, { "epoch": 0.2394391289311801, "grad_norm": 1.5615556240081787, "learning_rate": 8.893173806937268e-06, "loss": 0.5517, "step": 5712 }, { "epoch": 0.23948104754618907, "grad_norm": 1.6253242492675781, "learning_rate": 8.892747806571231e-06, "loss": 0.4959, "step": 5713 }, { "epoch": 0.23952296616119803, "grad_norm": 1.7060996294021606, "learning_rate": 8.892321734447771e-06, "loss": 0.6097, "step": 5714 }, { "epoch": 0.239564884776207, "grad_norm": 1.7608078718185425, "learning_rate": 8.891895590574739e-06, "loss": 0.5483, "step": 5715 }, { "epoch": 0.23960680339121596, "grad_norm": 1.500288486480713, "learning_rate": 8.891469374959996e-06, "loss": 0.5575, "step": 5716 }, { "epoch": 0.23964872200622492, "grad_norm": 1.5881198644638062, "learning_rate": 8.89104308761139e-06, "loss": 0.5914, "step": 5717 }, { "epoch": 0.23969064062123388, "grad_norm": 2.0701940059661865, "learning_rate": 8.890616728536787e-06, "loss": 0.5377, "step": 5718 }, { "epoch": 0.23973255923624284, "grad_norm": 1.6629717350006104, "learning_rate": 8.890190297744044e-06, "loss": 0.5379, "step": 5719 }, { "epoch": 0.2397744778512518, "grad_norm": 1.3580495119094849, "learning_rate": 8.889763795241017e-06, "loss": 0.4853, "step": 5720 }, { "epoch": 0.23981639646626077, "grad_norm": 1.524716854095459, "learning_rate": 8.889337221035576e-06, "loss": 0.5094, "step": 5721 }, { "epoch": 0.2398583150812697, "grad_norm": 1.6479321718215942, "learning_rate": 8.888910575135579e-06, "loss": 0.5053, "step": 5722 }, { "epoch": 0.23990023369627866, "grad_norm": 1.6702637672424316, "learning_rate": 8.88848385754889e-06, "loss": 0.536, "step": 5723 }, { "epoch": 0.23994215231128763, "grad_norm": 1.7691116333007812, "learning_rate": 8.888057068283378e-06, "loss": 0.5196, "step": 5724 }, { "epoch": 0.2399840709262966, "grad_norm": 1.6843661069869995, "learning_rate": 8.88763020734691e-06, "loss": 0.5447, "step": 5725 }, { "epoch": 0.24002598954130555, "grad_norm": 1.5331194400787354, "learning_rate": 8.88720327474735e-06, "loss": 0.5231, "step": 5726 }, { "epoch": 0.2400679081563145, "grad_norm": 1.6538488864898682, "learning_rate": 8.886776270492576e-06, "loss": 0.5647, "step": 5727 }, { "epoch": 0.24010982677132348, "grad_norm": 1.6869144439697266, "learning_rate": 8.886349194590453e-06, "loss": 0.5556, "step": 5728 }, { "epoch": 0.24015174538633244, "grad_norm": 1.6704765558242798, "learning_rate": 8.885922047048854e-06, "loss": 0.5124, "step": 5729 }, { "epoch": 0.2401936640013414, "grad_norm": 1.8894845247268677, "learning_rate": 8.885494827875656e-06, "loss": 0.5474, "step": 5730 }, { "epoch": 0.24023558261635036, "grad_norm": 1.658530592918396, "learning_rate": 8.885067537078732e-06, "loss": 0.5216, "step": 5731 }, { "epoch": 0.24027750123135933, "grad_norm": 1.735232949256897, "learning_rate": 8.88464017466596e-06, "loss": 0.5753, "step": 5732 }, { "epoch": 0.2403194198463683, "grad_norm": 1.9495391845703125, "learning_rate": 8.884212740645214e-06, "loss": 0.5545, "step": 5733 }, { "epoch": 0.24036133846137725, "grad_norm": 1.8059688806533813, "learning_rate": 8.883785235024379e-06, "loss": 0.5651, "step": 5734 }, { "epoch": 0.24040325707638618, "grad_norm": 1.8714048862457275, "learning_rate": 8.88335765781133e-06, "loss": 0.6038, "step": 5735 }, { "epoch": 0.24044517569139515, "grad_norm": 1.6069785356521606, "learning_rate": 8.88293000901395e-06, "loss": 0.5571, "step": 5736 }, { "epoch": 0.2404870943064041, "grad_norm": 1.7407153844833374, "learning_rate": 8.882502288640126e-06, "loss": 0.5632, "step": 5737 }, { "epoch": 0.24052901292141307, "grad_norm": 1.5179033279418945, "learning_rate": 8.88207449669774e-06, "loss": 0.5428, "step": 5738 }, { "epoch": 0.24057093153642203, "grad_norm": 2.592343807220459, "learning_rate": 8.881646633194674e-06, "loss": 0.5889, "step": 5739 }, { "epoch": 0.240612850151431, "grad_norm": 1.5240031480789185, "learning_rate": 8.881218698138822e-06, "loss": 0.5338, "step": 5740 }, { "epoch": 0.24065476876643996, "grad_norm": 1.577934741973877, "learning_rate": 8.880790691538068e-06, "loss": 0.5202, "step": 5741 }, { "epoch": 0.24069668738144892, "grad_norm": 1.608219027519226, "learning_rate": 8.880362613400301e-06, "loss": 0.4777, "step": 5742 }, { "epoch": 0.24073860599645788, "grad_norm": 1.926095724105835, "learning_rate": 8.879934463733416e-06, "loss": 0.676, "step": 5743 }, { "epoch": 0.24078052461146685, "grad_norm": 1.5997047424316406, "learning_rate": 8.879506242545303e-06, "loss": 0.533, "step": 5744 }, { "epoch": 0.2408224432264758, "grad_norm": 1.499929666519165, "learning_rate": 8.879077949843855e-06, "loss": 0.5562, "step": 5745 }, { "epoch": 0.24086436184148477, "grad_norm": 1.5865132808685303, "learning_rate": 8.878649585636968e-06, "loss": 0.5815, "step": 5746 }, { "epoch": 0.2409062804564937, "grad_norm": 1.8946492671966553, "learning_rate": 8.878221149932536e-06, "loss": 0.5864, "step": 5747 }, { "epoch": 0.24094819907150267, "grad_norm": 1.6783275604248047, "learning_rate": 8.877792642738458e-06, "loss": 0.5186, "step": 5748 }, { "epoch": 0.24099011768651163, "grad_norm": 1.6944811344146729, "learning_rate": 8.877364064062637e-06, "loss": 0.5199, "step": 5749 }, { "epoch": 0.2410320363015206, "grad_norm": 1.7190443277359009, "learning_rate": 8.876935413912969e-06, "loss": 0.5985, "step": 5750 }, { "epoch": 0.24107395491652955, "grad_norm": 2.597440481185913, "learning_rate": 8.876506692297355e-06, "loss": 0.5178, "step": 5751 }, { "epoch": 0.24111587353153852, "grad_norm": 1.805149793624878, "learning_rate": 8.876077899223699e-06, "loss": 0.5328, "step": 5752 }, { "epoch": 0.24115779214654748, "grad_norm": 1.449561595916748, "learning_rate": 8.875649034699907e-06, "loss": 0.5114, "step": 5753 }, { "epoch": 0.24119971076155644, "grad_norm": 1.4570196866989136, "learning_rate": 8.875220098733881e-06, "loss": 0.5215, "step": 5754 }, { "epoch": 0.2412416293765654, "grad_norm": 1.9042072296142578, "learning_rate": 8.874791091333532e-06, "loss": 0.5629, "step": 5755 }, { "epoch": 0.24128354799157437, "grad_norm": 1.6762804985046387, "learning_rate": 8.874362012506765e-06, "loss": 0.5755, "step": 5756 }, { "epoch": 0.24132546660658333, "grad_norm": 1.7605148553848267, "learning_rate": 8.873932862261489e-06, "loss": 0.5389, "step": 5757 }, { "epoch": 0.2413673852215923, "grad_norm": 1.689829707145691, "learning_rate": 8.873503640605618e-06, "loss": 0.4815, "step": 5758 }, { "epoch": 0.24140930383660125, "grad_norm": 1.6227225065231323, "learning_rate": 8.873074347547063e-06, "loss": 0.5237, "step": 5759 }, { "epoch": 0.2414512224516102, "grad_norm": 1.5810528993606567, "learning_rate": 8.872644983093736e-06, "loss": 0.5444, "step": 5760 }, { "epoch": 0.24149314106661915, "grad_norm": 1.7817301750183105, "learning_rate": 8.872215547253551e-06, "loss": 0.5724, "step": 5761 }, { "epoch": 0.2415350596816281, "grad_norm": 1.5414764881134033, "learning_rate": 8.871786040034426e-06, "loss": 0.5611, "step": 5762 }, { "epoch": 0.24157697829663707, "grad_norm": 1.8285763263702393, "learning_rate": 8.87135646144428e-06, "loss": 0.5212, "step": 5763 }, { "epoch": 0.24161889691164604, "grad_norm": 1.8452893495559692, "learning_rate": 8.87092681149103e-06, "loss": 0.5994, "step": 5764 }, { "epoch": 0.241660815526655, "grad_norm": 1.8052600622177124, "learning_rate": 8.870497090182593e-06, "loss": 0.593, "step": 5765 }, { "epoch": 0.24170273414166396, "grad_norm": 1.6464390754699707, "learning_rate": 8.870067297526894e-06, "loss": 0.5267, "step": 5766 }, { "epoch": 0.24174465275667292, "grad_norm": 1.8497836589813232, "learning_rate": 8.869637433531854e-06, "loss": 0.5744, "step": 5767 }, { "epoch": 0.24178657137168189, "grad_norm": 1.6146433353424072, "learning_rate": 8.869207498205398e-06, "loss": 0.4896, "step": 5768 }, { "epoch": 0.24182848998669085, "grad_norm": 1.587404489517212, "learning_rate": 8.868777491555451e-06, "loss": 0.5535, "step": 5769 }, { "epoch": 0.2418704086016998, "grad_norm": 1.707805871963501, "learning_rate": 8.86834741358994e-06, "loss": 0.5448, "step": 5770 }, { "epoch": 0.24191232721670877, "grad_norm": 1.6707043647766113, "learning_rate": 8.86791726431679e-06, "loss": 0.5393, "step": 5771 }, { "epoch": 0.2419542458317177, "grad_norm": 1.6489808559417725, "learning_rate": 8.867487043743935e-06, "loss": 0.5492, "step": 5772 }, { "epoch": 0.24199616444672667, "grad_norm": 1.6404942274093628, "learning_rate": 8.867056751879302e-06, "loss": 0.591, "step": 5773 }, { "epoch": 0.24203808306173563, "grad_norm": 1.733038067817688, "learning_rate": 8.866626388730823e-06, "loss": 0.5637, "step": 5774 }, { "epoch": 0.2420800016767446, "grad_norm": 1.5586973428726196, "learning_rate": 8.866195954306434e-06, "loss": 0.5643, "step": 5775 }, { "epoch": 0.24212192029175356, "grad_norm": 1.5656007528305054, "learning_rate": 8.865765448614065e-06, "loss": 0.4797, "step": 5776 }, { "epoch": 0.24216383890676252, "grad_norm": 1.7635577917099, "learning_rate": 8.865334871661656e-06, "loss": 0.5676, "step": 5777 }, { "epoch": 0.24220575752177148, "grad_norm": 2.0020344257354736, "learning_rate": 8.864904223457142e-06, "loss": 0.562, "step": 5778 }, { "epoch": 0.24224767613678044, "grad_norm": 1.5983425378799438, "learning_rate": 8.864473504008462e-06, "loss": 0.5187, "step": 5779 }, { "epoch": 0.2422895947517894, "grad_norm": 1.7032063007354736, "learning_rate": 8.864042713323553e-06, "loss": 0.509, "step": 5780 }, { "epoch": 0.24233151336679837, "grad_norm": 3.5711402893066406, "learning_rate": 8.86361185141036e-06, "loss": 0.5185, "step": 5781 }, { "epoch": 0.24237343198180733, "grad_norm": 1.813070297241211, "learning_rate": 8.863180918276826e-06, "loss": 0.5276, "step": 5782 }, { "epoch": 0.2424153505968163, "grad_norm": 1.586844563484192, "learning_rate": 8.862749913930891e-06, "loss": 0.5123, "step": 5783 }, { "epoch": 0.24245726921182525, "grad_norm": 1.8440076112747192, "learning_rate": 8.8623188383805e-06, "loss": 0.5673, "step": 5784 }, { "epoch": 0.2424991878268342, "grad_norm": 2.132859468460083, "learning_rate": 8.861887691633601e-06, "loss": 0.579, "step": 5785 }, { "epoch": 0.24254110644184315, "grad_norm": 2.495044708251953, "learning_rate": 8.861456473698142e-06, "loss": 0.5415, "step": 5786 }, { "epoch": 0.24258302505685211, "grad_norm": 1.7488789558410645, "learning_rate": 8.861025184582073e-06, "loss": 0.5541, "step": 5787 }, { "epoch": 0.24262494367186108, "grad_norm": 1.8226584196090698, "learning_rate": 8.86059382429334e-06, "loss": 0.5338, "step": 5788 }, { "epoch": 0.24266686228687004, "grad_norm": 1.668676733970642, "learning_rate": 8.860162392839898e-06, "loss": 0.5722, "step": 5789 }, { "epoch": 0.242708780901879, "grad_norm": 1.8120466470718384, "learning_rate": 8.859730890229698e-06, "loss": 0.5121, "step": 5790 }, { "epoch": 0.24275069951688796, "grad_norm": 1.5514864921569824, "learning_rate": 8.859299316470695e-06, "loss": 0.5182, "step": 5791 }, { "epoch": 0.24279261813189693, "grad_norm": 1.8387922048568726, "learning_rate": 8.858867671570844e-06, "loss": 0.5222, "step": 5792 }, { "epoch": 0.2428345367469059, "grad_norm": 1.7557380199432373, "learning_rate": 8.858435955538102e-06, "loss": 0.5794, "step": 5793 }, { "epoch": 0.24287645536191485, "grad_norm": 1.5711417198181152, "learning_rate": 8.85800416838043e-06, "loss": 0.5876, "step": 5794 }, { "epoch": 0.2429183739769238, "grad_norm": 1.87690007686615, "learning_rate": 8.857572310105784e-06, "loss": 0.524, "step": 5795 }, { "epoch": 0.24296029259193277, "grad_norm": 2.112942695617676, "learning_rate": 8.857140380722124e-06, "loss": 0.6214, "step": 5796 }, { "epoch": 0.2430022112069417, "grad_norm": 1.5437860488891602, "learning_rate": 8.856708380237413e-06, "loss": 0.484, "step": 5797 }, { "epoch": 0.24304412982195067, "grad_norm": 1.7180386781692505, "learning_rate": 8.856276308659615e-06, "loss": 0.5791, "step": 5798 }, { "epoch": 0.24308604843695963, "grad_norm": 2.1163618564605713, "learning_rate": 8.855844165996695e-06, "loss": 0.5182, "step": 5799 }, { "epoch": 0.2431279670519686, "grad_norm": 1.4903644323349, "learning_rate": 8.855411952256618e-06, "loss": 0.5446, "step": 5800 }, { "epoch": 0.24316988566697756, "grad_norm": 2.3154900074005127, "learning_rate": 8.854979667447353e-06, "loss": 0.55, "step": 5801 }, { "epoch": 0.24321180428198652, "grad_norm": 1.638655185699463, "learning_rate": 8.854547311576865e-06, "loss": 0.5429, "step": 5802 }, { "epoch": 0.24325372289699548, "grad_norm": 1.8520045280456543, "learning_rate": 8.854114884653128e-06, "loss": 0.5862, "step": 5803 }, { "epoch": 0.24329564151200445, "grad_norm": 1.7181788682937622, "learning_rate": 8.85368238668411e-06, "loss": 0.5395, "step": 5804 }, { "epoch": 0.2433375601270134, "grad_norm": 1.652276635169983, "learning_rate": 8.853249817677785e-06, "loss": 0.4983, "step": 5805 }, { "epoch": 0.24337947874202237, "grad_norm": 1.7476145029067993, "learning_rate": 8.852817177642127e-06, "loss": 0.5309, "step": 5806 }, { "epoch": 0.24342139735703133, "grad_norm": 1.712222933769226, "learning_rate": 8.85238446658511e-06, "loss": 0.5742, "step": 5807 }, { "epoch": 0.2434633159720403, "grad_norm": 3.2448744773864746, "learning_rate": 8.851951684514713e-06, "loss": 0.5526, "step": 5808 }, { "epoch": 0.24350523458704926, "grad_norm": 1.644879937171936, "learning_rate": 8.851518831438909e-06, "loss": 0.5557, "step": 5809 }, { "epoch": 0.2435471532020582, "grad_norm": 1.674261450767517, "learning_rate": 8.851085907365682e-06, "loss": 0.5287, "step": 5810 }, { "epoch": 0.24358907181706715, "grad_norm": 1.538538932800293, "learning_rate": 8.850652912303008e-06, "loss": 0.4758, "step": 5811 }, { "epoch": 0.24363099043207612, "grad_norm": 1.9121465682983398, "learning_rate": 8.850219846258873e-06, "loss": 0.5291, "step": 5812 }, { "epoch": 0.24367290904708508, "grad_norm": 1.7323757410049438, "learning_rate": 8.849786709241257e-06, "loss": 0.5612, "step": 5813 }, { "epoch": 0.24371482766209404, "grad_norm": 1.6417702436447144, "learning_rate": 8.849353501258147e-06, "loss": 0.59, "step": 5814 }, { "epoch": 0.243756746277103, "grad_norm": 1.6818751096725464, "learning_rate": 8.848920222317525e-06, "loss": 0.559, "step": 5815 }, { "epoch": 0.24379866489211197, "grad_norm": 1.6081523895263672, "learning_rate": 8.84848687242738e-06, "loss": 0.4993, "step": 5816 }, { "epoch": 0.24384058350712093, "grad_norm": 1.713121771812439, "learning_rate": 8.8480534515957e-06, "loss": 0.5468, "step": 5817 }, { "epoch": 0.2438825021221299, "grad_norm": 1.5177555084228516, "learning_rate": 8.847619959830473e-06, "loss": 0.5045, "step": 5818 }, { "epoch": 0.24392442073713885, "grad_norm": 1.9858450889587402, "learning_rate": 8.847186397139694e-06, "loss": 0.4896, "step": 5819 }, { "epoch": 0.24396633935214782, "grad_norm": 1.7074251174926758, "learning_rate": 8.84675276353135e-06, "loss": 0.5083, "step": 5820 }, { "epoch": 0.24400825796715678, "grad_norm": 1.592391848564148, "learning_rate": 8.846319059013437e-06, "loss": 0.5244, "step": 5821 }, { "epoch": 0.2440501765821657, "grad_norm": 1.9186646938323975, "learning_rate": 8.845885283593951e-06, "loss": 0.5868, "step": 5822 }, { "epoch": 0.24409209519717467, "grad_norm": 1.5470080375671387, "learning_rate": 8.845451437280886e-06, "loss": 0.5555, "step": 5823 }, { "epoch": 0.24413401381218364, "grad_norm": 1.5729812383651733, "learning_rate": 8.84501752008224e-06, "loss": 0.5249, "step": 5824 }, { "epoch": 0.2441759324271926, "grad_norm": 1.530549168586731, "learning_rate": 8.84458353200601e-06, "loss": 0.5549, "step": 5825 }, { "epoch": 0.24421785104220156, "grad_norm": 1.5608060359954834, "learning_rate": 8.8441494730602e-06, "loss": 0.5459, "step": 5826 }, { "epoch": 0.24425976965721052, "grad_norm": 1.4665050506591797, "learning_rate": 8.843715343252806e-06, "loss": 0.4885, "step": 5827 }, { "epoch": 0.24430168827221949, "grad_norm": 1.6715025901794434, "learning_rate": 8.843281142591836e-06, "loss": 0.55, "step": 5828 }, { "epoch": 0.24434360688722845, "grad_norm": 1.6617034673690796, "learning_rate": 8.84284687108529e-06, "loss": 0.5234, "step": 5829 }, { "epoch": 0.2443855255022374, "grad_norm": 1.425238847732544, "learning_rate": 8.842412528741176e-06, "loss": 0.4884, "step": 5830 }, { "epoch": 0.24442744411724637, "grad_norm": 2.2816054821014404, "learning_rate": 8.841978115567497e-06, "loss": 0.602, "step": 5831 }, { "epoch": 0.24446936273225534, "grad_norm": 1.731917381286621, "learning_rate": 8.841543631572264e-06, "loss": 0.561, "step": 5832 }, { "epoch": 0.2445112813472643, "grad_norm": 1.6751577854156494, "learning_rate": 8.841109076763484e-06, "loss": 0.5589, "step": 5833 }, { "epoch": 0.24455319996227326, "grad_norm": 1.7302007675170898, "learning_rate": 8.840674451149169e-06, "loss": 0.5789, "step": 5834 }, { "epoch": 0.2445951185772822, "grad_norm": 1.3795733451843262, "learning_rate": 8.840239754737329e-06, "loss": 0.5156, "step": 5835 }, { "epoch": 0.24463703719229116, "grad_norm": 1.54603111743927, "learning_rate": 8.83980498753598e-06, "loss": 0.5464, "step": 5836 }, { "epoch": 0.24467895580730012, "grad_norm": 1.596683382987976, "learning_rate": 8.839370149553133e-06, "loss": 0.4928, "step": 5837 }, { "epoch": 0.24472087442230908, "grad_norm": 2.2993955612182617, "learning_rate": 8.838935240796807e-06, "loss": 0.5094, "step": 5838 }, { "epoch": 0.24476279303731804, "grad_norm": 1.630143642425537, "learning_rate": 8.838500261275016e-06, "loss": 0.5981, "step": 5839 }, { "epoch": 0.244804711652327, "grad_norm": 1.69083571434021, "learning_rate": 8.838065210995779e-06, "loss": 0.509, "step": 5840 }, { "epoch": 0.24484663026733597, "grad_norm": 1.6852587461471558, "learning_rate": 8.837630089967115e-06, "loss": 0.4881, "step": 5841 }, { "epoch": 0.24488854888234493, "grad_norm": 1.9149932861328125, "learning_rate": 8.837194898197045e-06, "loss": 0.5886, "step": 5842 }, { "epoch": 0.2449304674973539, "grad_norm": 1.4482221603393555, "learning_rate": 8.836759635693595e-06, "loss": 0.5022, "step": 5843 }, { "epoch": 0.24497238611236286, "grad_norm": 1.6406742334365845, "learning_rate": 8.836324302464782e-06, "loss": 0.5492, "step": 5844 }, { "epoch": 0.24501430472737182, "grad_norm": 1.6550958156585693, "learning_rate": 8.835888898518637e-06, "loss": 0.5251, "step": 5845 }, { "epoch": 0.24505622334238078, "grad_norm": 1.7422679662704468, "learning_rate": 8.835453423863181e-06, "loss": 0.5211, "step": 5846 }, { "epoch": 0.24509814195738971, "grad_norm": 1.4465394020080566, "learning_rate": 8.835017878506444e-06, "loss": 0.5203, "step": 5847 }, { "epoch": 0.24514006057239868, "grad_norm": 1.8777059316635132, "learning_rate": 8.834582262456455e-06, "loss": 0.577, "step": 5848 }, { "epoch": 0.24518197918740764, "grad_norm": 1.7527014017105103, "learning_rate": 8.834146575721242e-06, "loss": 0.5399, "step": 5849 }, { "epoch": 0.2452238978024166, "grad_norm": 1.5378400087356567, "learning_rate": 8.833710818308839e-06, "loss": 0.5492, "step": 5850 }, { "epoch": 0.24526581641742556, "grad_norm": 1.7788535356521606, "learning_rate": 8.833274990227275e-06, "loss": 0.5468, "step": 5851 }, { "epoch": 0.24530773503243453, "grad_norm": 1.646271824836731, "learning_rate": 8.832839091484586e-06, "loss": 0.5349, "step": 5852 }, { "epoch": 0.2453496536474435, "grad_norm": 1.74016273021698, "learning_rate": 8.832403122088808e-06, "loss": 0.5468, "step": 5853 }, { "epoch": 0.24539157226245245, "grad_norm": 1.4232980012893677, "learning_rate": 8.831967082047976e-06, "loss": 0.4782, "step": 5854 }, { "epoch": 0.2454334908774614, "grad_norm": 1.5528813600540161, "learning_rate": 8.831530971370129e-06, "loss": 0.5153, "step": 5855 }, { "epoch": 0.24547540949247038, "grad_norm": 1.5579756498336792, "learning_rate": 8.831094790063305e-06, "loss": 0.5069, "step": 5856 }, { "epoch": 0.24551732810747934, "grad_norm": 2.0188181400299072, "learning_rate": 8.830658538135546e-06, "loss": 0.5847, "step": 5857 }, { "epoch": 0.2455592467224883, "grad_norm": 1.6430712938308716, "learning_rate": 8.83022221559489e-06, "loss": 0.4871, "step": 5858 }, { "epoch": 0.24560116533749726, "grad_norm": 1.6278382539749146, "learning_rate": 8.829785822449387e-06, "loss": 0.568, "step": 5859 }, { "epoch": 0.2456430839525062, "grad_norm": 1.5272659063339233, "learning_rate": 8.829349358707072e-06, "loss": 0.4899, "step": 5860 }, { "epoch": 0.24568500256751516, "grad_norm": 1.6579996347427368, "learning_rate": 8.828912824375997e-06, "loss": 0.5135, "step": 5861 }, { "epoch": 0.24572692118252412, "grad_norm": 1.588568925857544, "learning_rate": 8.82847621946421e-06, "loss": 0.5336, "step": 5862 }, { "epoch": 0.24576883979753308, "grad_norm": 1.6210885047912598, "learning_rate": 8.828039543979754e-06, "loss": 0.5688, "step": 5863 }, { "epoch": 0.24581075841254205, "grad_norm": 1.6415884494781494, "learning_rate": 8.827602797930682e-06, "loss": 0.5189, "step": 5864 }, { "epoch": 0.245852677027551, "grad_norm": 1.547438383102417, "learning_rate": 8.827165981325044e-06, "loss": 0.4798, "step": 5865 }, { "epoch": 0.24589459564255997, "grad_norm": 1.6187641620635986, "learning_rate": 8.826729094170891e-06, "loss": 0.5399, "step": 5866 }, { "epoch": 0.24593651425756893, "grad_norm": 1.7955920696258545, "learning_rate": 8.826292136476278e-06, "loss": 0.5133, "step": 5867 }, { "epoch": 0.2459784328725779, "grad_norm": 1.7476822137832642, "learning_rate": 8.825855108249259e-06, "loss": 0.5611, "step": 5868 }, { "epoch": 0.24602035148758686, "grad_norm": 1.822448968887329, "learning_rate": 8.82541800949789e-06, "loss": 0.5817, "step": 5869 }, { "epoch": 0.24606227010259582, "grad_norm": 1.5916563272476196, "learning_rate": 8.824980840230229e-06, "loss": 0.5656, "step": 5870 }, { "epoch": 0.24610418871760478, "grad_norm": 1.6841174364089966, "learning_rate": 8.824543600454335e-06, "loss": 0.5743, "step": 5871 }, { "epoch": 0.24614610733261372, "grad_norm": 1.6940646171569824, "learning_rate": 8.824106290178265e-06, "loss": 0.5038, "step": 5872 }, { "epoch": 0.24618802594762268, "grad_norm": 1.577246904373169, "learning_rate": 8.823668909410084e-06, "loss": 0.4796, "step": 5873 }, { "epoch": 0.24622994456263164, "grad_norm": 1.5359615087509155, "learning_rate": 8.823231458157853e-06, "loss": 0.5262, "step": 5874 }, { "epoch": 0.2462718631776406, "grad_norm": 1.6281628608703613, "learning_rate": 8.822793936429633e-06, "loss": 0.51, "step": 5875 }, { "epoch": 0.24631378179264957, "grad_norm": 1.6936030387878418, "learning_rate": 8.822356344233494e-06, "loss": 0.5525, "step": 5876 }, { "epoch": 0.24635570040765853, "grad_norm": 1.7468814849853516, "learning_rate": 8.821918681577499e-06, "loss": 0.5187, "step": 5877 }, { "epoch": 0.2463976190226675, "grad_norm": 1.630569577217102, "learning_rate": 8.821480948469715e-06, "loss": 0.5382, "step": 5878 }, { "epoch": 0.24643953763767645, "grad_norm": 1.7782167196273804, "learning_rate": 8.821043144918216e-06, "loss": 0.5365, "step": 5879 }, { "epoch": 0.24648145625268542, "grad_norm": 1.815867304801941, "learning_rate": 8.82060527093107e-06, "loss": 0.5006, "step": 5880 }, { "epoch": 0.24652337486769438, "grad_norm": 1.7217741012573242, "learning_rate": 8.820167326516343e-06, "loss": 0.5833, "step": 5881 }, { "epoch": 0.24656529348270334, "grad_norm": 1.5279806852340698, "learning_rate": 8.819729311682116e-06, "loss": 0.5411, "step": 5882 }, { "epoch": 0.2466072120977123, "grad_norm": 1.4190723896026611, "learning_rate": 8.819291226436458e-06, "loss": 0.5287, "step": 5883 }, { "epoch": 0.24664913071272127, "grad_norm": 1.70329749584198, "learning_rate": 8.818853070787447e-06, "loss": 0.5081, "step": 5884 }, { "epoch": 0.2466910493277302, "grad_norm": 1.5898315906524658, "learning_rate": 8.81841484474316e-06, "loss": 0.5213, "step": 5885 }, { "epoch": 0.24673296794273916, "grad_norm": 1.583892822265625, "learning_rate": 8.817976548311673e-06, "loss": 0.5392, "step": 5886 }, { "epoch": 0.24677488655774812, "grad_norm": 1.5481070280075073, "learning_rate": 8.817538181501067e-06, "loss": 0.5579, "step": 5887 }, { "epoch": 0.2468168051727571, "grad_norm": 1.6530184745788574, "learning_rate": 8.817099744319421e-06, "loss": 0.5385, "step": 5888 }, { "epoch": 0.24685872378776605, "grad_norm": 1.5170360803604126, "learning_rate": 8.81666123677482e-06, "loss": 0.5538, "step": 5889 }, { "epoch": 0.246900642402775, "grad_norm": 1.5500571727752686, "learning_rate": 8.816222658875344e-06, "loss": 0.5404, "step": 5890 }, { "epoch": 0.24694256101778397, "grad_norm": 1.4675580263137817, "learning_rate": 8.815784010629078e-06, "loss": 0.5383, "step": 5891 }, { "epoch": 0.24698447963279294, "grad_norm": 1.5840365886688232, "learning_rate": 8.81534529204411e-06, "loss": 0.5748, "step": 5892 }, { "epoch": 0.2470263982478019, "grad_norm": 1.7752305269241333, "learning_rate": 8.814906503128528e-06, "loss": 0.5149, "step": 5893 }, { "epoch": 0.24706831686281086, "grad_norm": 1.9348825216293335, "learning_rate": 8.814467643890418e-06, "loss": 0.5546, "step": 5894 }, { "epoch": 0.24711023547781982, "grad_norm": 1.741826057434082, "learning_rate": 8.81402871433787e-06, "loss": 0.5501, "step": 5895 }, { "epoch": 0.24715215409282879, "grad_norm": 1.7473299503326416, "learning_rate": 8.813589714478974e-06, "loss": 0.6037, "step": 5896 }, { "epoch": 0.24719407270783772, "grad_norm": 1.419248104095459, "learning_rate": 8.813150644321824e-06, "loss": 0.4688, "step": 5897 }, { "epoch": 0.24723599132284668, "grad_norm": 1.613556146621704, "learning_rate": 8.812711503874514e-06, "loss": 0.4764, "step": 5898 }, { "epoch": 0.24727790993785564, "grad_norm": 1.774035096168518, "learning_rate": 8.812272293145139e-06, "loss": 0.55, "step": 5899 }, { "epoch": 0.2473198285528646, "grad_norm": 1.5809541940689087, "learning_rate": 8.811833012141793e-06, "loss": 0.5792, "step": 5900 }, { "epoch": 0.24736174716787357, "grad_norm": 1.6372406482696533, "learning_rate": 8.811393660872577e-06, "loss": 0.584, "step": 5901 }, { "epoch": 0.24740366578288253, "grad_norm": 1.6065857410430908, "learning_rate": 8.810954239345587e-06, "loss": 0.5434, "step": 5902 }, { "epoch": 0.2474455843978915, "grad_norm": 1.6040959358215332, "learning_rate": 8.810514747568924e-06, "loss": 0.4989, "step": 5903 }, { "epoch": 0.24748750301290046, "grad_norm": 1.6703107357025146, "learning_rate": 8.81007518555069e-06, "loss": 0.5055, "step": 5904 }, { "epoch": 0.24752942162790942, "grad_norm": 1.4883959293365479, "learning_rate": 8.809635553298986e-06, "loss": 0.5243, "step": 5905 }, { "epoch": 0.24757134024291838, "grad_norm": 1.5744869709014893, "learning_rate": 8.809195850821919e-06, "loss": 0.5267, "step": 5906 }, { "epoch": 0.24761325885792734, "grad_norm": 1.6335474252700806, "learning_rate": 8.808756078127592e-06, "loss": 0.5327, "step": 5907 }, { "epoch": 0.2476551774729363, "grad_norm": 1.5969696044921875, "learning_rate": 8.808316235224111e-06, "loss": 0.5477, "step": 5908 }, { "epoch": 0.24769709608794527, "grad_norm": 1.7085089683532715, "learning_rate": 8.807876322119586e-06, "loss": 0.5668, "step": 5909 }, { "epoch": 0.2477390147029542, "grad_norm": 1.6363016366958618, "learning_rate": 8.807436338822125e-06, "loss": 0.5493, "step": 5910 }, { "epoch": 0.24778093331796316, "grad_norm": 1.7259591817855835, "learning_rate": 8.80699628533984e-06, "loss": 0.5756, "step": 5911 }, { "epoch": 0.24782285193297213, "grad_norm": 3.2441608905792236, "learning_rate": 8.806556161680839e-06, "loss": 0.5418, "step": 5912 }, { "epoch": 0.2478647705479811, "grad_norm": 2.4144492149353027, "learning_rate": 8.806115967853239e-06, "loss": 0.5478, "step": 5913 }, { "epoch": 0.24790668916299005, "grad_norm": 1.3838448524475098, "learning_rate": 8.805675703865153e-06, "loss": 0.5061, "step": 5914 }, { "epoch": 0.247948607777999, "grad_norm": 1.377578616142273, "learning_rate": 8.805235369724697e-06, "loss": 0.5192, "step": 5915 }, { "epoch": 0.24799052639300798, "grad_norm": 1.5614120960235596, "learning_rate": 8.804794965439987e-06, "loss": 0.5551, "step": 5916 }, { "epoch": 0.24803244500801694, "grad_norm": 1.5911052227020264, "learning_rate": 8.80435449101914e-06, "loss": 0.5675, "step": 5917 }, { "epoch": 0.2480743636230259, "grad_norm": 1.6825014352798462, "learning_rate": 8.80391394647028e-06, "loss": 0.5878, "step": 5918 }, { "epoch": 0.24811628223803486, "grad_norm": 1.7266041040420532, "learning_rate": 8.803473331801524e-06, "loss": 0.5328, "step": 5919 }, { "epoch": 0.24815820085304383, "grad_norm": 1.6599442958831787, "learning_rate": 8.803032647020995e-06, "loss": 0.5318, "step": 5920 }, { "epoch": 0.2482001194680528, "grad_norm": 1.9145481586456299, "learning_rate": 8.802591892136818e-06, "loss": 0.5396, "step": 5921 }, { "epoch": 0.24824203808306172, "grad_norm": 1.7012966871261597, "learning_rate": 8.802151067157115e-06, "loss": 0.6178, "step": 5922 }, { "epoch": 0.24828395669807068, "grad_norm": 1.7518396377563477, "learning_rate": 8.801710172090015e-06, "loss": 0.6067, "step": 5923 }, { "epoch": 0.24832587531307965, "grad_norm": 1.775957703590393, "learning_rate": 8.80126920694364e-06, "loss": 0.5553, "step": 5924 }, { "epoch": 0.2483677939280886, "grad_norm": 1.4547710418701172, "learning_rate": 8.800828171726125e-06, "loss": 0.5163, "step": 5925 }, { "epoch": 0.24840971254309757, "grad_norm": 1.6821562051773071, "learning_rate": 8.800387066445599e-06, "loss": 0.5277, "step": 5926 }, { "epoch": 0.24845163115810653, "grad_norm": 1.5945647954940796, "learning_rate": 8.799945891110189e-06, "loss": 0.5249, "step": 5927 }, { "epoch": 0.2484935497731155, "grad_norm": 1.7048391103744507, "learning_rate": 8.799504645728029e-06, "loss": 0.5345, "step": 5928 }, { "epoch": 0.24853546838812446, "grad_norm": 1.7305892705917358, "learning_rate": 8.799063330307256e-06, "loss": 0.5367, "step": 5929 }, { "epoch": 0.24857738700313342, "grad_norm": 1.593942403793335, "learning_rate": 8.798621944856001e-06, "loss": 0.494, "step": 5930 }, { "epoch": 0.24861930561814238, "grad_norm": 1.6564416885375977, "learning_rate": 8.798180489382401e-06, "loss": 0.5152, "step": 5931 }, { "epoch": 0.24866122423315135, "grad_norm": 1.4974642992019653, "learning_rate": 8.797738963894596e-06, "loss": 0.4907, "step": 5932 }, { "epoch": 0.2487031428481603, "grad_norm": 1.896850347518921, "learning_rate": 8.797297368400724e-06, "loss": 0.5251, "step": 5933 }, { "epoch": 0.24874506146316927, "grad_norm": 1.5588196516036987, "learning_rate": 8.796855702908921e-06, "loss": 0.5669, "step": 5934 }, { "epoch": 0.2487869800781782, "grad_norm": 1.517021894454956, "learning_rate": 8.796413967427337e-06, "loss": 0.5232, "step": 5935 }, { "epoch": 0.24882889869318717, "grad_norm": 1.5716784000396729, "learning_rate": 8.795972161964108e-06, "loss": 0.5247, "step": 5936 }, { "epoch": 0.24887081730819613, "grad_norm": 1.5232353210449219, "learning_rate": 8.795530286527378e-06, "loss": 0.5349, "step": 5937 }, { "epoch": 0.2489127359232051, "grad_norm": 1.8404901027679443, "learning_rate": 8.795088341125296e-06, "loss": 0.5195, "step": 5938 }, { "epoch": 0.24895465453821405, "grad_norm": 1.485647439956665, "learning_rate": 8.794646325766008e-06, "loss": 0.5303, "step": 5939 }, { "epoch": 0.24899657315322302, "grad_norm": 1.4988288879394531, "learning_rate": 8.794204240457658e-06, "loss": 0.5385, "step": 5940 }, { "epoch": 0.24903849176823198, "grad_norm": 1.6794394254684448, "learning_rate": 8.793762085208402e-06, "loss": 0.5172, "step": 5941 }, { "epoch": 0.24908041038324094, "grad_norm": 1.7478148937225342, "learning_rate": 8.793319860026383e-06, "loss": 0.526, "step": 5942 }, { "epoch": 0.2491223289982499, "grad_norm": 1.6239855289459229, "learning_rate": 8.792877564919758e-06, "loss": 0.5756, "step": 5943 }, { "epoch": 0.24916424761325887, "grad_norm": 1.8447821140289307, "learning_rate": 8.79243519989668e-06, "loss": 0.5342, "step": 5944 }, { "epoch": 0.24920616622826783, "grad_norm": 1.6728519201278687, "learning_rate": 8.7919927649653e-06, "loss": 0.5565, "step": 5945 }, { "epoch": 0.2492480848432768, "grad_norm": 1.4748982191085815, "learning_rate": 8.791550260133776e-06, "loss": 0.4791, "step": 5946 }, { "epoch": 0.24929000345828572, "grad_norm": 1.8671244382858276, "learning_rate": 8.791107685410266e-06, "loss": 0.6026, "step": 5947 }, { "epoch": 0.2493319220732947, "grad_norm": 1.527704119682312, "learning_rate": 8.790665040802925e-06, "loss": 0.4588, "step": 5948 }, { "epoch": 0.24937384068830365, "grad_norm": 1.6734755039215088, "learning_rate": 8.790222326319915e-06, "loss": 0.538, "step": 5949 }, { "epoch": 0.2494157593033126, "grad_norm": 1.9110963344573975, "learning_rate": 8.789779541969397e-06, "loss": 0.5667, "step": 5950 }, { "epoch": 0.24945767791832157, "grad_norm": 1.6545321941375732, "learning_rate": 8.78933668775953e-06, "loss": 0.5506, "step": 5951 }, { "epoch": 0.24949959653333054, "grad_norm": 1.7039759159088135, "learning_rate": 8.788893763698482e-06, "loss": 0.5795, "step": 5952 }, { "epoch": 0.2495415151483395, "grad_norm": 1.625809669494629, "learning_rate": 8.788450769794417e-06, "loss": 0.4951, "step": 5953 }, { "epoch": 0.24958343376334846, "grad_norm": 1.5982497930526733, "learning_rate": 8.788007706055496e-06, "loss": 0.5599, "step": 5954 }, { "epoch": 0.24962535237835742, "grad_norm": 1.4802809953689575, "learning_rate": 8.787564572489891e-06, "loss": 0.5441, "step": 5955 }, { "epoch": 0.24966727099336639, "grad_norm": 1.6755675077438354, "learning_rate": 8.78712136910577e-06, "loss": 0.5568, "step": 5956 }, { "epoch": 0.24970918960837535, "grad_norm": 1.6730130910873413, "learning_rate": 8.7866780959113e-06, "loss": 0.5463, "step": 5957 }, { "epoch": 0.2497511082233843, "grad_norm": 1.6703661680221558, "learning_rate": 8.786234752914658e-06, "loss": 0.5515, "step": 5958 }, { "epoch": 0.24979302683839327, "grad_norm": 1.55577552318573, "learning_rate": 8.78579134012401e-06, "loss": 0.5211, "step": 5959 }, { "epoch": 0.2498349454534022, "grad_norm": 1.543796181678772, "learning_rate": 8.785347857547532e-06, "loss": 0.5129, "step": 5960 }, { "epoch": 0.24987686406841117, "grad_norm": 1.613659381866455, "learning_rate": 8.7849043051934e-06, "loss": 0.5316, "step": 5961 }, { "epoch": 0.24991878268342013, "grad_norm": 1.7105307579040527, "learning_rate": 8.78446068306979e-06, "loss": 0.5089, "step": 5962 }, { "epoch": 0.2499607012984291, "grad_norm": 1.6991101503372192, "learning_rate": 8.784016991184878e-06, "loss": 0.5137, "step": 5963 }, { "epoch": 0.2500026199134381, "grad_norm": 5.341670989990234, "learning_rate": 8.783573229546846e-06, "loss": 0.5673, "step": 5964 }, { "epoch": 0.25004453852844705, "grad_norm": 1.5980397462844849, "learning_rate": 8.78312939816387e-06, "loss": 0.5414, "step": 5965 }, { "epoch": 0.25008645714345595, "grad_norm": 1.9034278392791748, "learning_rate": 8.782685497044136e-06, "loss": 0.5901, "step": 5966 }, { "epoch": 0.2501283757584649, "grad_norm": 1.8922743797302246, "learning_rate": 8.782241526195823e-06, "loss": 0.5182, "step": 5967 }, { "epoch": 0.2501702943734739, "grad_norm": 1.6885156631469727, "learning_rate": 8.781797485627116e-06, "loss": 0.5954, "step": 5968 }, { "epoch": 0.25021221298848284, "grad_norm": 1.5979106426239014, "learning_rate": 8.7813533753462e-06, "loss": 0.5502, "step": 5969 }, { "epoch": 0.2502541316034918, "grad_norm": 1.6418455839157104, "learning_rate": 8.780909195361262e-06, "loss": 0.5556, "step": 5970 }, { "epoch": 0.25029605021850077, "grad_norm": 1.8160427808761597, "learning_rate": 8.780464945680491e-06, "loss": 0.5702, "step": 5971 }, { "epoch": 0.2503379688335097, "grad_norm": 1.7890194654464722, "learning_rate": 8.780020626312076e-06, "loss": 0.5403, "step": 5972 }, { "epoch": 0.2503798874485187, "grad_norm": 1.9188748598098755, "learning_rate": 8.779576237264204e-06, "loss": 0.6176, "step": 5973 }, { "epoch": 0.25042180606352765, "grad_norm": 1.8395464420318604, "learning_rate": 8.77913177854507e-06, "loss": 0.5533, "step": 5974 }, { "epoch": 0.2504637246785366, "grad_norm": 1.9184849262237549, "learning_rate": 8.778687250162867e-06, "loss": 0.5728, "step": 5975 }, { "epoch": 0.2505056432935456, "grad_norm": 1.9622427225112915, "learning_rate": 8.778242652125787e-06, "loss": 0.4934, "step": 5976 }, { "epoch": 0.25054756190855454, "grad_norm": 1.6007689237594604, "learning_rate": 8.77779798444203e-06, "loss": 0.5578, "step": 5977 }, { "epoch": 0.2505894805235635, "grad_norm": 1.6061557531356812, "learning_rate": 8.777353247119787e-06, "loss": 0.559, "step": 5978 }, { "epoch": 0.25063139913857246, "grad_norm": 1.8483458757400513, "learning_rate": 8.776908440167258e-06, "loss": 0.605, "step": 5979 }, { "epoch": 0.2506733177535814, "grad_norm": 2.0532147884368896, "learning_rate": 8.776463563592645e-06, "loss": 0.5475, "step": 5980 }, { "epoch": 0.2507152363685904, "grad_norm": 1.4655088186264038, "learning_rate": 8.776018617404148e-06, "loss": 0.5037, "step": 5981 }, { "epoch": 0.25075715498359935, "grad_norm": 1.940627098083496, "learning_rate": 8.775573601609968e-06, "loss": 0.5572, "step": 5982 }, { "epoch": 0.2507990735986083, "grad_norm": 1.8943084478378296, "learning_rate": 8.775128516218306e-06, "loss": 0.5665, "step": 5983 }, { "epoch": 0.2508409922136173, "grad_norm": 1.4757977724075317, "learning_rate": 8.774683361237371e-06, "loss": 0.545, "step": 5984 }, { "epoch": 0.25088291082862624, "grad_norm": 1.7027897834777832, "learning_rate": 8.774238136675365e-06, "loss": 0.5608, "step": 5985 }, { "epoch": 0.2509248294436352, "grad_norm": 1.5579427480697632, "learning_rate": 8.773792842540497e-06, "loss": 0.5434, "step": 5986 }, { "epoch": 0.25096674805864416, "grad_norm": 2.121697187423706, "learning_rate": 8.773347478840974e-06, "loss": 0.5381, "step": 5987 }, { "epoch": 0.2510086666736531, "grad_norm": 1.8350138664245605, "learning_rate": 8.772902045585008e-06, "loss": 0.5532, "step": 5988 }, { "epoch": 0.2510505852886621, "grad_norm": 1.9474287033081055, "learning_rate": 8.772456542780809e-06, "loss": 0.5255, "step": 5989 }, { "epoch": 0.25109250390367105, "grad_norm": 1.8242734670639038, "learning_rate": 8.772010970436589e-06, "loss": 0.5674, "step": 5990 }, { "epoch": 0.25113442251867996, "grad_norm": 1.6611336469650269, "learning_rate": 8.77156532856056e-06, "loss": 0.5218, "step": 5991 }, { "epoch": 0.2511763411336889, "grad_norm": 2.323585033416748, "learning_rate": 8.77111961716094e-06, "loss": 0.5392, "step": 5992 }, { "epoch": 0.2512182597486979, "grad_norm": 2.4660050868988037, "learning_rate": 8.770673836245942e-06, "loss": 0.5506, "step": 5993 }, { "epoch": 0.25126017836370684, "grad_norm": 1.811391830444336, "learning_rate": 8.770227985823787e-06, "loss": 0.5984, "step": 5994 }, { "epoch": 0.2513020969787158, "grad_norm": 1.5318883657455444, "learning_rate": 8.769782065902688e-06, "loss": 0.4915, "step": 5995 }, { "epoch": 0.25134401559372477, "grad_norm": 2.372053861618042, "learning_rate": 8.76933607649087e-06, "loss": 0.5347, "step": 5996 }, { "epoch": 0.25138593420873373, "grad_norm": 2.1363584995269775, "learning_rate": 8.768890017596551e-06, "loss": 0.5598, "step": 5997 }, { "epoch": 0.2514278528237427, "grad_norm": 2.0066311359405518, "learning_rate": 8.768443889227956e-06, "loss": 0.53, "step": 5998 }, { "epoch": 0.25146977143875165, "grad_norm": 2.15578293800354, "learning_rate": 8.767997691393308e-06, "loss": 0.4912, "step": 5999 }, { "epoch": 0.2515116900537606, "grad_norm": 1.6333889961242676, "learning_rate": 8.767551424100831e-06, "loss": 0.5234, "step": 6000 }, { "epoch": 0.2515536086687696, "grad_norm": 1.771182656288147, "learning_rate": 8.767105087358751e-06, "loss": 0.5297, "step": 6001 }, { "epoch": 0.25159552728377854, "grad_norm": 1.8956230878829956, "learning_rate": 8.766658681175297e-06, "loss": 0.5517, "step": 6002 }, { "epoch": 0.2516374458987875, "grad_norm": 1.7526971101760864, "learning_rate": 8.7662122055587e-06, "loss": 0.5431, "step": 6003 }, { "epoch": 0.25167936451379647, "grad_norm": 1.7032123804092407, "learning_rate": 8.765765660517184e-06, "loss": 0.4703, "step": 6004 }, { "epoch": 0.25172128312880543, "grad_norm": 1.6460596323013306, "learning_rate": 8.765319046058986e-06, "loss": 0.516, "step": 6005 }, { "epoch": 0.2517632017438144, "grad_norm": 2.355510711669922, "learning_rate": 8.764872362192337e-06, "loss": 0.5756, "step": 6006 }, { "epoch": 0.25180512035882335, "grad_norm": 2.0039141178131104, "learning_rate": 8.76442560892547e-06, "loss": 0.5607, "step": 6007 }, { "epoch": 0.2518470389738323, "grad_norm": 1.9612418413162231, "learning_rate": 8.763978786266622e-06, "loss": 0.504, "step": 6008 }, { "epoch": 0.2518889575888413, "grad_norm": 1.7071585655212402, "learning_rate": 8.763531894224027e-06, "loss": 0.5327, "step": 6009 }, { "epoch": 0.25193087620385024, "grad_norm": 1.7762106657028198, "learning_rate": 8.763084932805925e-06, "loss": 0.548, "step": 6010 }, { "epoch": 0.2519727948188592, "grad_norm": 1.6932616233825684, "learning_rate": 8.762637902020557e-06, "loss": 0.5524, "step": 6011 }, { "epoch": 0.25201471343386816, "grad_norm": 1.948923945426941, "learning_rate": 8.762190801876159e-06, "loss": 0.5214, "step": 6012 }, { "epoch": 0.2520566320488771, "grad_norm": 1.6573330163955688, "learning_rate": 8.761743632380975e-06, "loss": 0.5226, "step": 6013 }, { "epoch": 0.2520985506638861, "grad_norm": 1.9353742599487305, "learning_rate": 8.761296393543247e-06, "loss": 0.532, "step": 6014 }, { "epoch": 0.25214046927889505, "grad_norm": 1.5884079933166504, "learning_rate": 8.760849085371219e-06, "loss": 0.5617, "step": 6015 }, { "epoch": 0.25218238789390396, "grad_norm": 1.7316521406173706, "learning_rate": 8.760401707873138e-06, "loss": 0.582, "step": 6016 }, { "epoch": 0.2522243065089129, "grad_norm": 1.8487428426742554, "learning_rate": 8.75995426105725e-06, "loss": 0.5495, "step": 6017 }, { "epoch": 0.2522662251239219, "grad_norm": 1.4924215078353882, "learning_rate": 8.759506744931805e-06, "loss": 0.5222, "step": 6018 }, { "epoch": 0.25230814373893085, "grad_norm": 1.6648533344268799, "learning_rate": 8.759059159505047e-06, "loss": 0.4941, "step": 6019 }, { "epoch": 0.2523500623539398, "grad_norm": 1.3910375833511353, "learning_rate": 8.758611504785234e-06, "loss": 0.5185, "step": 6020 }, { "epoch": 0.25239198096894877, "grad_norm": 1.698315978050232, "learning_rate": 8.758163780780613e-06, "loss": 0.4983, "step": 6021 }, { "epoch": 0.25243389958395773, "grad_norm": 3.9705193042755127, "learning_rate": 8.757715987499437e-06, "loss": 0.5719, "step": 6022 }, { "epoch": 0.2524758181989667, "grad_norm": 1.7006504535675049, "learning_rate": 8.757268124949963e-06, "loss": 0.5713, "step": 6023 }, { "epoch": 0.25251773681397566, "grad_norm": 1.639231562614441, "learning_rate": 8.756820193140443e-06, "loss": 0.4915, "step": 6024 }, { "epoch": 0.2525596554289846, "grad_norm": 1.5489228963851929, "learning_rate": 8.756372192079139e-06, "loss": 0.4803, "step": 6025 }, { "epoch": 0.2526015740439936, "grad_norm": 1.871242880821228, "learning_rate": 8.755924121774306e-06, "loss": 0.5129, "step": 6026 }, { "epoch": 0.25264349265900254, "grad_norm": 1.7043439149856567, "learning_rate": 8.755475982234203e-06, "loss": 0.5555, "step": 6027 }, { "epoch": 0.2526854112740115, "grad_norm": 1.615038514137268, "learning_rate": 8.755027773467093e-06, "loss": 0.5128, "step": 6028 }, { "epoch": 0.25272732988902047, "grad_norm": 1.954196572303772, "learning_rate": 8.754579495481237e-06, "loss": 0.5697, "step": 6029 }, { "epoch": 0.25276924850402943, "grad_norm": 1.6276103258132935, "learning_rate": 8.7541311482849e-06, "loss": 0.5421, "step": 6030 }, { "epoch": 0.2528111671190384, "grad_norm": 1.654036521911621, "learning_rate": 8.753682731886341e-06, "loss": 0.4939, "step": 6031 }, { "epoch": 0.25285308573404736, "grad_norm": 2.238093376159668, "learning_rate": 8.753234246293834e-06, "loss": 0.5269, "step": 6032 }, { "epoch": 0.2528950043490563, "grad_norm": 1.4592832326889038, "learning_rate": 8.752785691515642e-06, "loss": 0.5445, "step": 6033 }, { "epoch": 0.2529369229640653, "grad_norm": 2.026315927505493, "learning_rate": 8.752337067560033e-06, "loss": 0.6071, "step": 6034 }, { "epoch": 0.25297884157907424, "grad_norm": 1.6296205520629883, "learning_rate": 8.751888374435277e-06, "loss": 0.5086, "step": 6035 }, { "epoch": 0.2530207601940832, "grad_norm": 1.736077904701233, "learning_rate": 8.751439612149648e-06, "loss": 0.5297, "step": 6036 }, { "epoch": 0.25306267880909217, "grad_norm": 1.6809545755386353, "learning_rate": 8.750990780711413e-06, "loss": 0.4916, "step": 6037 }, { "epoch": 0.25310459742410113, "grad_norm": 1.7587124109268188, "learning_rate": 8.750541880128851e-06, "loss": 0.5944, "step": 6038 }, { "epoch": 0.2531465160391101, "grad_norm": 1.9821414947509766, "learning_rate": 8.750092910410234e-06, "loss": 0.5589, "step": 6039 }, { "epoch": 0.25318843465411905, "grad_norm": 2.440309524536133, "learning_rate": 8.749643871563837e-06, "loss": 0.5066, "step": 6040 }, { "epoch": 0.25323035326912796, "grad_norm": 1.7852330207824707, "learning_rate": 8.74919476359794e-06, "loss": 0.5158, "step": 6041 }, { "epoch": 0.2532722718841369, "grad_norm": 1.864052653312683, "learning_rate": 8.74874558652082e-06, "loss": 0.5502, "step": 6042 }, { "epoch": 0.2533141904991459, "grad_norm": 1.7075767517089844, "learning_rate": 8.748296340340758e-06, "loss": 0.5379, "step": 6043 }, { "epoch": 0.25335610911415485, "grad_norm": 1.793977975845337, "learning_rate": 8.747847025066036e-06, "loss": 0.5697, "step": 6044 }, { "epoch": 0.2533980277291638, "grad_norm": 2.002953290939331, "learning_rate": 8.747397640704934e-06, "loss": 0.5372, "step": 6045 }, { "epoch": 0.2534399463441728, "grad_norm": 2.101422071456909, "learning_rate": 8.746948187265738e-06, "loss": 0.5298, "step": 6046 }, { "epoch": 0.25348186495918174, "grad_norm": 1.7868578433990479, "learning_rate": 8.74649866475673e-06, "loss": 0.5078, "step": 6047 }, { "epoch": 0.2535237835741907, "grad_norm": 1.8420530557632446, "learning_rate": 8.746049073186201e-06, "loss": 0.5516, "step": 6048 }, { "epoch": 0.25356570218919966, "grad_norm": 1.7236571311950684, "learning_rate": 8.745599412562436e-06, "loss": 0.493, "step": 6049 }, { "epoch": 0.2536076208042086, "grad_norm": 1.6013439893722534, "learning_rate": 8.745149682893721e-06, "loss": 0.5104, "step": 6050 }, { "epoch": 0.2536495394192176, "grad_norm": 1.715086579322815, "learning_rate": 8.744699884188352e-06, "loss": 0.5045, "step": 6051 }, { "epoch": 0.25369145803422655, "grad_norm": 1.5324211120605469, "learning_rate": 8.744250016454616e-06, "loss": 0.5445, "step": 6052 }, { "epoch": 0.2537333766492355, "grad_norm": 1.5034120082855225, "learning_rate": 8.743800079700809e-06, "loss": 0.5348, "step": 6053 }, { "epoch": 0.25377529526424447, "grad_norm": 1.8881431818008423, "learning_rate": 8.743350073935222e-06, "loss": 0.5267, "step": 6054 }, { "epoch": 0.25381721387925343, "grad_norm": 1.7735803127288818, "learning_rate": 8.742899999166153e-06, "loss": 0.5282, "step": 6055 }, { "epoch": 0.2538591324942624, "grad_norm": 1.7530497312545776, "learning_rate": 8.742449855401894e-06, "loss": 0.5506, "step": 6056 }, { "epoch": 0.25390105110927136, "grad_norm": 1.4698978662490845, "learning_rate": 8.741999642650747e-06, "loss": 0.5199, "step": 6057 }, { "epoch": 0.2539429697242803, "grad_norm": 1.7071348428726196, "learning_rate": 8.741549360921012e-06, "loss": 0.5748, "step": 6058 }, { "epoch": 0.2539848883392893, "grad_norm": 1.682992935180664, "learning_rate": 8.741099010220983e-06, "loss": 0.5592, "step": 6059 }, { "epoch": 0.25402680695429825, "grad_norm": 1.8100947141647339, "learning_rate": 8.740648590558967e-06, "loss": 0.4996, "step": 6060 }, { "epoch": 0.2540687255693072, "grad_norm": 1.5526244640350342, "learning_rate": 8.740198101943267e-06, "loss": 0.516, "step": 6061 }, { "epoch": 0.25411064418431617, "grad_norm": 1.8298274278640747, "learning_rate": 8.739747544382184e-06, "loss": 0.5655, "step": 6062 }, { "epoch": 0.25415256279932513, "grad_norm": 1.7938910722732544, "learning_rate": 8.739296917884026e-06, "loss": 0.5484, "step": 6063 }, { "epoch": 0.2541944814143341, "grad_norm": 1.807192325592041, "learning_rate": 8.7388462224571e-06, "loss": 0.5277, "step": 6064 }, { "epoch": 0.25423640002934306, "grad_norm": 1.6200166940689087, "learning_rate": 8.738395458109711e-06, "loss": 0.5573, "step": 6065 }, { "epoch": 0.25427831864435196, "grad_norm": 1.5709755420684814, "learning_rate": 8.73794462485017e-06, "loss": 0.564, "step": 6066 }, { "epoch": 0.2543202372593609, "grad_norm": 2.275916814804077, "learning_rate": 8.737493722686788e-06, "loss": 0.5742, "step": 6067 }, { "epoch": 0.2543621558743699, "grad_norm": 1.5411673784255981, "learning_rate": 8.737042751627875e-06, "loss": 0.5704, "step": 6068 }, { "epoch": 0.25440407448937885, "grad_norm": 1.523384690284729, "learning_rate": 8.736591711681747e-06, "loss": 0.5262, "step": 6069 }, { "epoch": 0.2544459931043878, "grad_norm": 1.5923056602478027, "learning_rate": 8.736140602856715e-06, "loss": 0.5284, "step": 6070 }, { "epoch": 0.2544879117193968, "grad_norm": 1.7810444831848145, "learning_rate": 8.735689425161096e-06, "loss": 0.5356, "step": 6071 }, { "epoch": 0.25452983033440574, "grad_norm": 1.6994034051895142, "learning_rate": 8.735238178603209e-06, "loss": 0.5505, "step": 6072 }, { "epoch": 0.2545717489494147, "grad_norm": 1.559814691543579, "learning_rate": 8.734786863191366e-06, "loss": 0.5154, "step": 6073 }, { "epoch": 0.25461366756442366, "grad_norm": 1.4876489639282227, "learning_rate": 8.734335478933893e-06, "loss": 0.5422, "step": 6074 }, { "epoch": 0.2546555861794326, "grad_norm": 2.356470823287964, "learning_rate": 8.733884025839108e-06, "loss": 0.5617, "step": 6075 }, { "epoch": 0.2546975047944416, "grad_norm": 1.6900073289871216, "learning_rate": 8.733432503915333e-06, "loss": 0.523, "step": 6076 }, { "epoch": 0.25473942340945055, "grad_norm": 1.8360666036605835, "learning_rate": 8.73298091317089e-06, "loss": 0.5238, "step": 6077 }, { "epoch": 0.2547813420244595, "grad_norm": 1.8006179332733154, "learning_rate": 8.732529253614106e-06, "loss": 0.5181, "step": 6078 }, { "epoch": 0.2548232606394685, "grad_norm": 2.0928292274475098, "learning_rate": 8.732077525253306e-06, "loss": 0.606, "step": 6079 }, { "epoch": 0.25486517925447744, "grad_norm": 1.8729205131530762, "learning_rate": 8.731625728096814e-06, "loss": 0.5318, "step": 6080 }, { "epoch": 0.2549070978694864, "grad_norm": 2.1613142490386963, "learning_rate": 8.731173862152961e-06, "loss": 0.5104, "step": 6081 }, { "epoch": 0.25494901648449536, "grad_norm": 1.966893196105957, "learning_rate": 8.730721927430077e-06, "loss": 0.563, "step": 6082 }, { "epoch": 0.2549909350995043, "grad_norm": 1.478036642074585, "learning_rate": 8.73026992393649e-06, "loss": 0.5081, "step": 6083 }, { "epoch": 0.2550328537145133, "grad_norm": 1.8447095155715942, "learning_rate": 8.729817851680536e-06, "loss": 0.5361, "step": 6084 }, { "epoch": 0.25507477232952225, "grad_norm": 1.4706097841262817, "learning_rate": 8.729365710670545e-06, "loss": 0.569, "step": 6085 }, { "epoch": 0.2551166909445312, "grad_norm": 1.8081693649291992, "learning_rate": 8.728913500914854e-06, "loss": 0.5484, "step": 6086 }, { "epoch": 0.2551586095595402, "grad_norm": 1.6113402843475342, "learning_rate": 8.728461222421795e-06, "loss": 0.5123, "step": 6087 }, { "epoch": 0.25520052817454913, "grad_norm": 1.7278975248336792, "learning_rate": 8.728008875199712e-06, "loss": 0.5329, "step": 6088 }, { "epoch": 0.2552424467895581, "grad_norm": 1.8454909324645996, "learning_rate": 8.727556459256936e-06, "loss": 0.5502, "step": 6089 }, { "epoch": 0.25528436540456706, "grad_norm": 1.8243627548217773, "learning_rate": 8.727103974601811e-06, "loss": 0.473, "step": 6090 }, { "epoch": 0.25532628401957597, "grad_norm": 1.8219395875930786, "learning_rate": 8.726651421242676e-06, "loss": 0.5204, "step": 6091 }, { "epoch": 0.25536820263458493, "grad_norm": 1.7373454570770264, "learning_rate": 8.726198799187875e-06, "loss": 0.5581, "step": 6092 }, { "epoch": 0.2554101212495939, "grad_norm": 1.7750895023345947, "learning_rate": 8.72574610844575e-06, "loss": 0.595, "step": 6093 }, { "epoch": 0.25545203986460285, "grad_norm": 1.8921302556991577, "learning_rate": 8.725293349024646e-06, "loss": 0.5944, "step": 6094 }, { "epoch": 0.2554939584796118, "grad_norm": 2.1659252643585205, "learning_rate": 8.724840520932909e-06, "loss": 0.509, "step": 6095 }, { "epoch": 0.2555358770946208, "grad_norm": 1.793911337852478, "learning_rate": 8.724387624178887e-06, "loss": 0.5608, "step": 6096 }, { "epoch": 0.25557779570962974, "grad_norm": 2.0293710231781006, "learning_rate": 8.723934658770928e-06, "loss": 0.6124, "step": 6097 }, { "epoch": 0.2556197143246387, "grad_norm": 1.5169905424118042, "learning_rate": 8.723481624717382e-06, "loss": 0.5697, "step": 6098 }, { "epoch": 0.25566163293964767, "grad_norm": 1.8210363388061523, "learning_rate": 8.7230285220266e-06, "loss": 0.5556, "step": 6099 }, { "epoch": 0.2557035515546566, "grad_norm": 1.7356163263320923, "learning_rate": 8.722575350706933e-06, "loss": 0.532, "step": 6100 }, { "epoch": 0.2557454701696656, "grad_norm": 1.4482702016830444, "learning_rate": 8.722122110766735e-06, "loss": 0.5215, "step": 6101 }, { "epoch": 0.25578738878467455, "grad_norm": 1.6096166372299194, "learning_rate": 8.721668802214363e-06, "loss": 0.5399, "step": 6102 }, { "epoch": 0.2558293073996835, "grad_norm": 1.927541732788086, "learning_rate": 8.721215425058172e-06, "loss": 0.5481, "step": 6103 }, { "epoch": 0.2558712260146925, "grad_norm": 1.5579845905303955, "learning_rate": 8.720761979306518e-06, "loss": 0.4927, "step": 6104 }, { "epoch": 0.25591314462970144, "grad_norm": 1.8702768087387085, "learning_rate": 8.720308464967761e-06, "loss": 0.5333, "step": 6105 }, { "epoch": 0.2559550632447104, "grad_norm": 1.9363222122192383, "learning_rate": 8.71985488205026e-06, "loss": 0.5096, "step": 6106 }, { "epoch": 0.25599698185971936, "grad_norm": 2.1768932342529297, "learning_rate": 8.71940123056238e-06, "loss": 0.5638, "step": 6107 }, { "epoch": 0.2560389004747283, "grad_norm": 1.6790727376937866, "learning_rate": 8.718947510512477e-06, "loss": 0.5322, "step": 6108 }, { "epoch": 0.2560808190897373, "grad_norm": 1.6248005628585815, "learning_rate": 8.718493721908919e-06, "loss": 0.5279, "step": 6109 }, { "epoch": 0.25612273770474625, "grad_norm": 1.79585862159729, "learning_rate": 8.718039864760068e-06, "loss": 0.5652, "step": 6110 }, { "epoch": 0.2561646563197552, "grad_norm": 1.5940477848052979, "learning_rate": 8.717585939074295e-06, "loss": 0.544, "step": 6111 }, { "epoch": 0.2562065749347642, "grad_norm": 1.7429296970367432, "learning_rate": 8.717131944859964e-06, "loss": 0.5766, "step": 6112 }, { "epoch": 0.25624849354977314, "grad_norm": 1.6953004598617554, "learning_rate": 8.716677882125442e-06, "loss": 0.5767, "step": 6113 }, { "epoch": 0.2562904121647821, "grad_norm": 1.787447214126587, "learning_rate": 8.716223750879105e-06, "loss": 0.5413, "step": 6114 }, { "epoch": 0.25633233077979106, "grad_norm": 1.7714331150054932, "learning_rate": 8.715769551129318e-06, "loss": 0.5269, "step": 6115 }, { "epoch": 0.25637424939479997, "grad_norm": 1.4886682033538818, "learning_rate": 8.715315282884458e-06, "loss": 0.5556, "step": 6116 }, { "epoch": 0.25641616800980893, "grad_norm": 2.0320611000061035, "learning_rate": 8.714860946152896e-06, "loss": 0.5641, "step": 6117 }, { "epoch": 0.2564580866248179, "grad_norm": 1.6095070838928223, "learning_rate": 8.714406540943008e-06, "loss": 0.5746, "step": 6118 }, { "epoch": 0.25650000523982686, "grad_norm": 1.7167344093322754, "learning_rate": 8.71395206726317e-06, "loss": 0.6069, "step": 6119 }, { "epoch": 0.2565419238548358, "grad_norm": 1.4791451692581177, "learning_rate": 8.713497525121761e-06, "loss": 0.5062, "step": 6120 }, { "epoch": 0.2565838424698448, "grad_norm": 1.499403715133667, "learning_rate": 8.71304291452716e-06, "loss": 0.5412, "step": 6121 }, { "epoch": 0.25662576108485374, "grad_norm": 1.7076725959777832, "learning_rate": 8.712588235487742e-06, "loss": 0.5222, "step": 6122 }, { "epoch": 0.2566676796998627, "grad_norm": 1.8055592775344849, "learning_rate": 8.712133488011896e-06, "loss": 0.5558, "step": 6123 }, { "epoch": 0.25670959831487167, "grad_norm": 1.7419312000274658, "learning_rate": 8.711678672107999e-06, "loss": 0.5413, "step": 6124 }, { "epoch": 0.25675151692988063, "grad_norm": 1.493162751197815, "learning_rate": 8.711223787784438e-06, "loss": 0.5351, "step": 6125 }, { "epoch": 0.2567934355448896, "grad_norm": 1.5702544450759888, "learning_rate": 8.710768835049597e-06, "loss": 0.5124, "step": 6126 }, { "epoch": 0.25683535415989855, "grad_norm": 1.8690941333770752, "learning_rate": 8.710313813911863e-06, "loss": 0.5765, "step": 6127 }, { "epoch": 0.2568772727749075, "grad_norm": 1.4730031490325928, "learning_rate": 8.709858724379623e-06, "loss": 0.5348, "step": 6128 }, { "epoch": 0.2569191913899165, "grad_norm": 1.6943886280059814, "learning_rate": 8.709403566461266e-06, "loss": 0.5458, "step": 6129 }, { "epoch": 0.25696111000492544, "grad_norm": 1.7536532878875732, "learning_rate": 8.708948340165183e-06, "loss": 0.5597, "step": 6130 }, { "epoch": 0.2570030286199344, "grad_norm": 1.5815080404281616, "learning_rate": 8.708493045499761e-06, "loss": 0.5778, "step": 6131 }, { "epoch": 0.25704494723494337, "grad_norm": 1.7399019002914429, "learning_rate": 8.7080376824734e-06, "loss": 0.5768, "step": 6132 }, { "epoch": 0.25708686584995233, "grad_norm": 1.8312116861343384, "learning_rate": 8.707582251094491e-06, "loss": 0.5748, "step": 6133 }, { "epoch": 0.2571287844649613, "grad_norm": 1.7924997806549072, "learning_rate": 8.707126751371428e-06, "loss": 0.5719, "step": 6134 }, { "epoch": 0.25717070307997025, "grad_norm": 1.4943279027938843, "learning_rate": 8.706671183312607e-06, "loss": 0.481, "step": 6135 }, { "epoch": 0.2572126216949792, "grad_norm": 2.0247318744659424, "learning_rate": 8.706215546926429e-06, "loss": 0.5323, "step": 6136 }, { "epoch": 0.2572545403099882, "grad_norm": 1.9269051551818848, "learning_rate": 8.70575984222129e-06, "loss": 0.5551, "step": 6137 }, { "epoch": 0.25729645892499714, "grad_norm": 1.6800415515899658, "learning_rate": 8.70530406920559e-06, "loss": 0.5714, "step": 6138 }, { "epoch": 0.2573383775400061, "grad_norm": 1.6951230764389038, "learning_rate": 8.704848227887734e-06, "loss": 0.571, "step": 6139 }, { "epoch": 0.25738029615501506, "grad_norm": 1.8204632997512817, "learning_rate": 8.704392318276124e-06, "loss": 0.5544, "step": 6140 }, { "epoch": 0.25742221477002397, "grad_norm": 1.635610818862915, "learning_rate": 8.70393634037916e-06, "loss": 0.5014, "step": 6141 }, { "epoch": 0.25746413338503293, "grad_norm": 1.655733585357666, "learning_rate": 8.703480294205252e-06, "loss": 0.5358, "step": 6142 }, { "epoch": 0.2575060520000419, "grad_norm": 1.7949621677398682, "learning_rate": 8.703024179762804e-06, "loss": 0.5231, "step": 6143 }, { "epoch": 0.25754797061505086, "grad_norm": 1.6487385034561157, "learning_rate": 8.702567997060224e-06, "loss": 0.5579, "step": 6144 }, { "epoch": 0.2575898892300598, "grad_norm": 3.9945640563964844, "learning_rate": 8.702111746105922e-06, "loss": 0.5682, "step": 6145 }, { "epoch": 0.2576318078450688, "grad_norm": 2.721414804458618, "learning_rate": 8.701655426908307e-06, "loss": 0.4704, "step": 6146 }, { "epoch": 0.25767372646007775, "grad_norm": 1.5265016555786133, "learning_rate": 8.701199039475794e-06, "loss": 0.5283, "step": 6147 }, { "epoch": 0.2577156450750867, "grad_norm": 1.832080602645874, "learning_rate": 8.700742583816791e-06, "loss": 0.5416, "step": 6148 }, { "epoch": 0.25775756369009567, "grad_norm": 1.5932791233062744, "learning_rate": 8.700286059939716e-06, "loss": 0.5434, "step": 6149 }, { "epoch": 0.25779948230510463, "grad_norm": 1.9207781553268433, "learning_rate": 8.699829467852982e-06, "loss": 0.5534, "step": 6150 }, { "epoch": 0.2578414009201136, "grad_norm": 1.4037779569625854, "learning_rate": 8.699372807565007e-06, "loss": 0.5725, "step": 6151 }, { "epoch": 0.25788331953512256, "grad_norm": 1.8706889152526855, "learning_rate": 8.69891607908421e-06, "loss": 0.534, "step": 6152 }, { "epoch": 0.2579252381501315, "grad_norm": 1.5927692651748657, "learning_rate": 8.698459282419007e-06, "loss": 0.5172, "step": 6153 }, { "epoch": 0.2579671567651405, "grad_norm": 1.7517021894454956, "learning_rate": 8.69800241757782e-06, "loss": 0.5235, "step": 6154 }, { "epoch": 0.25800907538014944, "grad_norm": 2.1779139041900635, "learning_rate": 8.697545484569071e-06, "loss": 0.564, "step": 6155 }, { "epoch": 0.2580509939951584, "grad_norm": 1.6317448616027832, "learning_rate": 8.697088483401184e-06, "loss": 0.5413, "step": 6156 }, { "epoch": 0.25809291261016737, "grad_norm": 1.971082091331482, "learning_rate": 8.69663141408258e-06, "loss": 0.5197, "step": 6157 }, { "epoch": 0.25813483122517633, "grad_norm": 1.6623820066452026, "learning_rate": 8.696174276621688e-06, "loss": 0.5238, "step": 6158 }, { "epoch": 0.2581767498401853, "grad_norm": 1.7158308029174805, "learning_rate": 8.695717071026932e-06, "loss": 0.5396, "step": 6159 }, { "epoch": 0.25821866845519426, "grad_norm": 1.774980068206787, "learning_rate": 8.69525979730674e-06, "loss": 0.5345, "step": 6160 }, { "epoch": 0.2582605870702032, "grad_norm": 1.6010888814926147, "learning_rate": 8.694802455469544e-06, "loss": 0.5189, "step": 6161 }, { "epoch": 0.2583025056852122, "grad_norm": 1.7781065702438354, "learning_rate": 8.694345045523771e-06, "loss": 0.5189, "step": 6162 }, { "epoch": 0.25834442430022114, "grad_norm": 1.863656759262085, "learning_rate": 8.693887567477857e-06, "loss": 0.5623, "step": 6163 }, { "epoch": 0.2583863429152301, "grad_norm": 1.8783260583877563, "learning_rate": 8.69343002134023e-06, "loss": 0.4711, "step": 6164 }, { "epoch": 0.25842826153023907, "grad_norm": 1.5067963600158691, "learning_rate": 8.692972407119327e-06, "loss": 0.515, "step": 6165 }, { "epoch": 0.258470180145248, "grad_norm": 1.6099040508270264, "learning_rate": 8.692514724823583e-06, "loss": 0.5502, "step": 6166 }, { "epoch": 0.25851209876025694, "grad_norm": 1.8539602756500244, "learning_rate": 8.692056974461435e-06, "loss": 0.5564, "step": 6167 }, { "epoch": 0.2585540173752659, "grad_norm": 1.5886443853378296, "learning_rate": 8.691599156041321e-06, "loss": 0.54, "step": 6168 }, { "epoch": 0.25859593599027486, "grad_norm": 1.7857431173324585, "learning_rate": 8.69114126957168e-06, "loss": 0.5476, "step": 6169 }, { "epoch": 0.2586378546052838, "grad_norm": 1.533664584159851, "learning_rate": 8.690683315060952e-06, "loss": 0.5144, "step": 6170 }, { "epoch": 0.2586797732202928, "grad_norm": 1.5497000217437744, "learning_rate": 8.69022529251758e-06, "loss": 0.4859, "step": 6171 }, { "epoch": 0.25872169183530175, "grad_norm": 1.8487660884857178, "learning_rate": 8.689767201950006e-06, "loss": 0.5676, "step": 6172 }, { "epoch": 0.2587636104503107, "grad_norm": 1.4656776189804077, "learning_rate": 8.689309043366675e-06, "loss": 0.5534, "step": 6173 }, { "epoch": 0.2588055290653197, "grad_norm": 1.7214791774749756, "learning_rate": 8.688850816776032e-06, "loss": 0.5455, "step": 6174 }, { "epoch": 0.25884744768032864, "grad_norm": 1.5840940475463867, "learning_rate": 8.688392522186525e-06, "loss": 0.4856, "step": 6175 }, { "epoch": 0.2588893662953376, "grad_norm": 1.7369006872177124, "learning_rate": 8.687934159606596e-06, "loss": 0.5908, "step": 6176 }, { "epoch": 0.25893128491034656, "grad_norm": 2.5670924186706543, "learning_rate": 8.687475729044706e-06, "loss": 0.5762, "step": 6177 }, { "epoch": 0.2589732035253555, "grad_norm": 1.6661322116851807, "learning_rate": 8.687017230509293e-06, "loss": 0.5507, "step": 6178 }, { "epoch": 0.2590151221403645, "grad_norm": 1.6471776962280273, "learning_rate": 8.686558664008817e-06, "loss": 0.568, "step": 6179 }, { "epoch": 0.25905704075537345, "grad_norm": 3.189621686935425, "learning_rate": 8.686100029551729e-06, "loss": 0.5469, "step": 6180 }, { "epoch": 0.2590989593703824, "grad_norm": 2.0243442058563232, "learning_rate": 8.685641327146482e-06, "loss": 0.6111, "step": 6181 }, { "epoch": 0.25914087798539137, "grad_norm": 1.8116989135742188, "learning_rate": 8.685182556801533e-06, "loss": 0.5209, "step": 6182 }, { "epoch": 0.25918279660040033, "grad_norm": 1.4331755638122559, "learning_rate": 8.684723718525338e-06, "loss": 0.5182, "step": 6183 }, { "epoch": 0.2592247152154093, "grad_norm": 1.4540715217590332, "learning_rate": 8.684264812326356e-06, "loss": 0.5445, "step": 6184 }, { "epoch": 0.25926663383041826, "grad_norm": 1.7195003032684326, "learning_rate": 8.683805838213044e-06, "loss": 0.5828, "step": 6185 }, { "epoch": 0.2593085524454272, "grad_norm": 1.7841399908065796, "learning_rate": 8.683346796193865e-06, "loss": 0.5815, "step": 6186 }, { "epoch": 0.2593504710604362, "grad_norm": 1.5535359382629395, "learning_rate": 8.682887686277279e-06, "loss": 0.5346, "step": 6187 }, { "epoch": 0.25939238967544515, "grad_norm": 1.5989042520523071, "learning_rate": 8.68242850847175e-06, "loss": 0.5278, "step": 6188 }, { "epoch": 0.2594343082904541, "grad_norm": 1.9517369270324707, "learning_rate": 8.681969262785744e-06, "loss": 0.4498, "step": 6189 }, { "epoch": 0.25947622690546307, "grad_norm": 1.759149193763733, "learning_rate": 8.681509949227722e-06, "loss": 0.5588, "step": 6190 }, { "epoch": 0.259518145520472, "grad_norm": 2.201176643371582, "learning_rate": 8.681050567806157e-06, "loss": 0.5447, "step": 6191 }, { "epoch": 0.25956006413548094, "grad_norm": 1.5642236471176147, "learning_rate": 8.680591118529511e-06, "loss": 0.5254, "step": 6192 }, { "epoch": 0.2596019827504899, "grad_norm": 1.5616662502288818, "learning_rate": 8.680131601406257e-06, "loss": 0.4962, "step": 6193 }, { "epoch": 0.25964390136549886, "grad_norm": 1.7009871006011963, "learning_rate": 8.679672016444864e-06, "loss": 0.5806, "step": 6194 }, { "epoch": 0.2596858199805078, "grad_norm": 1.5432395935058594, "learning_rate": 8.679212363653805e-06, "loss": 0.5379, "step": 6195 }, { "epoch": 0.2597277385955168, "grad_norm": 1.8074895143508911, "learning_rate": 8.678752643041553e-06, "loss": 0.4924, "step": 6196 }, { "epoch": 0.25976965721052575, "grad_norm": 1.5498006343841553, "learning_rate": 8.678292854616582e-06, "loss": 0.505, "step": 6197 }, { "epoch": 0.2598115758255347, "grad_norm": 1.5988279581069946, "learning_rate": 8.677832998387365e-06, "loss": 0.6029, "step": 6198 }, { "epoch": 0.2598534944405437, "grad_norm": 1.6446118354797363, "learning_rate": 8.677373074362383e-06, "loss": 0.5034, "step": 6199 }, { "epoch": 0.25989541305555264, "grad_norm": 1.8100159168243408, "learning_rate": 8.676913082550112e-06, "loss": 0.5351, "step": 6200 }, { "epoch": 0.2599373316705616, "grad_norm": 1.705471396446228, "learning_rate": 8.67645302295903e-06, "loss": 0.5553, "step": 6201 }, { "epoch": 0.25997925028557056, "grad_norm": 1.554395318031311, "learning_rate": 8.67599289559762e-06, "loss": 0.5409, "step": 6202 }, { "epoch": 0.2600211689005795, "grad_norm": 1.5952379703521729, "learning_rate": 8.675532700474363e-06, "loss": 0.4997, "step": 6203 }, { "epoch": 0.2600630875155885, "grad_norm": 1.8197425603866577, "learning_rate": 8.675072437597742e-06, "loss": 0.5661, "step": 6204 }, { "epoch": 0.26010500613059745, "grad_norm": 1.7583287954330444, "learning_rate": 8.67461210697624e-06, "loss": 0.5478, "step": 6205 }, { "epoch": 0.2601469247456064, "grad_norm": 1.8975943326950073, "learning_rate": 8.674151708618346e-06, "loss": 0.6256, "step": 6206 }, { "epoch": 0.2601888433606154, "grad_norm": 1.679495096206665, "learning_rate": 8.673691242532545e-06, "loss": 0.5156, "step": 6207 }, { "epoch": 0.26023076197562434, "grad_norm": 1.6944161653518677, "learning_rate": 8.673230708727323e-06, "loss": 0.5618, "step": 6208 }, { "epoch": 0.2602726805906333, "grad_norm": 1.7639989852905273, "learning_rate": 8.67277010721117e-06, "loss": 0.4875, "step": 6209 }, { "epoch": 0.26031459920564226, "grad_norm": 1.7498120069503784, "learning_rate": 8.67230943799258e-06, "loss": 0.5637, "step": 6210 }, { "epoch": 0.2603565178206512, "grad_norm": 1.787158489227295, "learning_rate": 8.671848701080041e-06, "loss": 0.5651, "step": 6211 }, { "epoch": 0.2603984364356602, "grad_norm": 1.5987404584884644, "learning_rate": 8.671387896482049e-06, "loss": 0.5189, "step": 6212 }, { "epoch": 0.26044035505066915, "grad_norm": 1.5998843908309937, "learning_rate": 8.670927024207095e-06, "loss": 0.4896, "step": 6213 }, { "epoch": 0.2604822736656781, "grad_norm": 1.7657525539398193, "learning_rate": 8.670466084263676e-06, "loss": 0.5272, "step": 6214 }, { "epoch": 0.2605241922806871, "grad_norm": 2.0713024139404297, "learning_rate": 8.670005076660289e-06, "loss": 0.5512, "step": 6215 }, { "epoch": 0.260566110895696, "grad_norm": 1.791993260383606, "learning_rate": 8.669544001405432e-06, "loss": 0.5974, "step": 6216 }, { "epoch": 0.26060802951070494, "grad_norm": 2.4507405757904053, "learning_rate": 8.669082858507605e-06, "loss": 0.6177, "step": 6217 }, { "epoch": 0.2606499481257139, "grad_norm": 1.7100753784179688, "learning_rate": 8.668621647975308e-06, "loss": 0.5606, "step": 6218 }, { "epoch": 0.26069186674072287, "grad_norm": 1.6625220775604248, "learning_rate": 8.668160369817043e-06, "loss": 0.5093, "step": 6219 }, { "epoch": 0.26073378535573183, "grad_norm": 1.798339605331421, "learning_rate": 8.667699024041311e-06, "loss": 0.531, "step": 6220 }, { "epoch": 0.2607757039707408, "grad_norm": 1.5939337015151978, "learning_rate": 8.667237610656618e-06, "loss": 0.4613, "step": 6221 }, { "epoch": 0.26081762258574975, "grad_norm": 2.1264429092407227, "learning_rate": 8.666776129671472e-06, "loss": 0.5748, "step": 6222 }, { "epoch": 0.2608595412007587, "grad_norm": 1.775396704673767, "learning_rate": 8.666314581094375e-06, "loss": 0.5199, "step": 6223 }, { "epoch": 0.2609014598157677, "grad_norm": 1.829615592956543, "learning_rate": 8.665852964933838e-06, "loss": 0.629, "step": 6224 }, { "epoch": 0.26094337843077664, "grad_norm": 1.7144412994384766, "learning_rate": 8.66539128119837e-06, "loss": 0.5152, "step": 6225 }, { "epoch": 0.2609852970457856, "grad_norm": 2.0949594974517822, "learning_rate": 8.66492952989648e-06, "loss": 0.5353, "step": 6226 }, { "epoch": 0.26102721566079456, "grad_norm": 3.105001926422119, "learning_rate": 8.664467711036681e-06, "loss": 0.5733, "step": 6227 }, { "epoch": 0.2610691342758035, "grad_norm": 1.8263814449310303, "learning_rate": 8.664005824627485e-06, "loss": 0.4946, "step": 6228 }, { "epoch": 0.2611110528908125, "grad_norm": 1.762770414352417, "learning_rate": 8.663543870677408e-06, "loss": 0.5546, "step": 6229 }, { "epoch": 0.26115297150582145, "grad_norm": 2.016653299331665, "learning_rate": 8.663081849194963e-06, "loss": 0.51, "step": 6230 }, { "epoch": 0.2611948901208304, "grad_norm": 1.8883578777313232, "learning_rate": 8.662619760188671e-06, "loss": 0.5666, "step": 6231 }, { "epoch": 0.2612368087358394, "grad_norm": 1.6563427448272705, "learning_rate": 8.662157603667045e-06, "loss": 0.5471, "step": 6232 }, { "epoch": 0.26127872735084834, "grad_norm": 1.6596306562423706, "learning_rate": 8.661695379638607e-06, "loss": 0.5356, "step": 6233 }, { "epoch": 0.2613206459658573, "grad_norm": 1.7064706087112427, "learning_rate": 8.661233088111875e-06, "loss": 0.4892, "step": 6234 }, { "epoch": 0.26136256458086626, "grad_norm": 2.1921331882476807, "learning_rate": 8.660770729095374e-06, "loss": 0.5134, "step": 6235 }, { "epoch": 0.2614044831958752, "grad_norm": 1.6550272703170776, "learning_rate": 8.660308302597627e-06, "loss": 0.5767, "step": 6236 }, { "epoch": 0.2614464018108842, "grad_norm": 2.089930295944214, "learning_rate": 8.659845808627155e-06, "loss": 0.5287, "step": 6237 }, { "epoch": 0.26148832042589315, "grad_norm": 1.4529509544372559, "learning_rate": 8.659383247192485e-06, "loss": 0.4676, "step": 6238 }, { "epoch": 0.2615302390409021, "grad_norm": 1.6890476942062378, "learning_rate": 8.658920618302145e-06, "loss": 0.55, "step": 6239 }, { "epoch": 0.2615721576559111, "grad_norm": 1.5283536911010742, "learning_rate": 8.65845792196466e-06, "loss": 0.4708, "step": 6240 }, { "epoch": 0.26161407627092, "grad_norm": 1.663995623588562, "learning_rate": 8.657995158188562e-06, "loss": 0.5239, "step": 6241 }, { "epoch": 0.26165599488592894, "grad_norm": 1.4845151901245117, "learning_rate": 8.657532326982379e-06, "loss": 0.5252, "step": 6242 }, { "epoch": 0.2616979135009379, "grad_norm": 1.6387044191360474, "learning_rate": 8.657069428354645e-06, "loss": 0.5297, "step": 6243 }, { "epoch": 0.26173983211594687, "grad_norm": 1.6421847343444824, "learning_rate": 8.656606462313892e-06, "loss": 0.5075, "step": 6244 }, { "epoch": 0.26178175073095583, "grad_norm": 1.6324177980422974, "learning_rate": 8.656143428868654e-06, "loss": 0.5871, "step": 6245 }, { "epoch": 0.2618236693459648, "grad_norm": 1.9540553092956543, "learning_rate": 8.655680328027467e-06, "loss": 0.541, "step": 6246 }, { "epoch": 0.26186558796097376, "grad_norm": 1.6943069696426392, "learning_rate": 8.655217159798865e-06, "loss": 0.5127, "step": 6247 }, { "epoch": 0.2619075065759827, "grad_norm": 1.6663399934768677, "learning_rate": 8.65475392419139e-06, "loss": 0.6015, "step": 6248 }, { "epoch": 0.2619494251909917, "grad_norm": 1.7283278703689575, "learning_rate": 8.654290621213578e-06, "loss": 0.5522, "step": 6249 }, { "epoch": 0.26199134380600064, "grad_norm": 1.612264633178711, "learning_rate": 8.65382725087397e-06, "loss": 0.5152, "step": 6250 }, { "epoch": 0.2620332624210096, "grad_norm": 1.7666294574737549, "learning_rate": 8.653363813181107e-06, "loss": 0.5513, "step": 6251 }, { "epoch": 0.26207518103601857, "grad_norm": 1.3925758600234985, "learning_rate": 8.652900308143535e-06, "loss": 0.5437, "step": 6252 }, { "epoch": 0.26211709965102753, "grad_norm": 1.732707142829895, "learning_rate": 8.652436735769793e-06, "loss": 0.4727, "step": 6253 }, { "epoch": 0.2621590182660365, "grad_norm": 1.6793464422225952, "learning_rate": 8.651973096068431e-06, "loss": 0.4919, "step": 6254 }, { "epoch": 0.26220093688104545, "grad_norm": 2.181344985961914, "learning_rate": 8.651509389047994e-06, "loss": 0.5115, "step": 6255 }, { "epoch": 0.2622428554960544, "grad_norm": 1.5713818073272705, "learning_rate": 8.651045614717028e-06, "loss": 0.5365, "step": 6256 }, { "epoch": 0.2622847741110634, "grad_norm": 1.880663514137268, "learning_rate": 8.650581773084085e-06, "loss": 0.5524, "step": 6257 }, { "epoch": 0.26232669272607234, "grad_norm": 1.6311969757080078, "learning_rate": 8.650117864157714e-06, "loss": 0.4879, "step": 6258 }, { "epoch": 0.2623686113410813, "grad_norm": 1.5004721879959106, "learning_rate": 8.649653887946466e-06, "loss": 0.5384, "step": 6259 }, { "epoch": 0.26241052995609027, "grad_norm": 1.4914194345474243, "learning_rate": 8.649189844458892e-06, "loss": 0.535, "step": 6260 }, { "epoch": 0.26245244857109923, "grad_norm": 1.6455167531967163, "learning_rate": 8.648725733703551e-06, "loss": 0.5383, "step": 6261 }, { "epoch": 0.2624943671861082, "grad_norm": 1.5683125257492065, "learning_rate": 8.648261555688992e-06, "loss": 0.4908, "step": 6262 }, { "epoch": 0.26253628580111715, "grad_norm": 1.7488749027252197, "learning_rate": 8.647797310423778e-06, "loss": 0.5508, "step": 6263 }, { "epoch": 0.2625782044161261, "grad_norm": 1.9844046831130981, "learning_rate": 8.647332997916463e-06, "loss": 0.5102, "step": 6264 }, { "epoch": 0.2626201230311351, "grad_norm": 1.7025505304336548, "learning_rate": 8.646868618175605e-06, "loss": 0.5468, "step": 6265 }, { "epoch": 0.262662041646144, "grad_norm": 2.301668167114258, "learning_rate": 8.646404171209767e-06, "loss": 0.5055, "step": 6266 }, { "epoch": 0.26270396026115295, "grad_norm": 1.6585311889648438, "learning_rate": 8.645939657027507e-06, "loss": 0.5428, "step": 6267 }, { "epoch": 0.2627458788761619, "grad_norm": 1.6464154720306396, "learning_rate": 8.645475075637393e-06, "loss": 0.5845, "step": 6268 }, { "epoch": 0.26278779749117087, "grad_norm": 1.784683108329773, "learning_rate": 8.645010427047984e-06, "loss": 0.5559, "step": 6269 }, { "epoch": 0.26282971610617983, "grad_norm": 2.375271797180176, "learning_rate": 8.644545711267847e-06, "loss": 0.5215, "step": 6270 }, { "epoch": 0.2628716347211888, "grad_norm": 1.6733551025390625, "learning_rate": 8.644080928305548e-06, "loss": 0.4839, "step": 6271 }, { "epoch": 0.26291355333619776, "grad_norm": 1.7940155267715454, "learning_rate": 8.643616078169654e-06, "loss": 0.5379, "step": 6272 }, { "epoch": 0.2629554719512067, "grad_norm": 2.6224849224090576, "learning_rate": 8.643151160868736e-06, "loss": 0.4736, "step": 6273 }, { "epoch": 0.2629973905662157, "grad_norm": 1.7166725397109985, "learning_rate": 8.642686176411362e-06, "loss": 0.5538, "step": 6274 }, { "epoch": 0.26303930918122465, "grad_norm": 1.63184654712677, "learning_rate": 8.642221124806104e-06, "loss": 0.5294, "step": 6275 }, { "epoch": 0.2630812277962336, "grad_norm": 1.6909921169281006, "learning_rate": 8.641756006061538e-06, "loss": 0.4872, "step": 6276 }, { "epoch": 0.26312314641124257, "grad_norm": 1.7414942979812622, "learning_rate": 8.64129082018623e-06, "loss": 0.4821, "step": 6277 }, { "epoch": 0.26316506502625153, "grad_norm": 1.9567872285842896, "learning_rate": 8.640825567188763e-06, "loss": 0.5084, "step": 6278 }, { "epoch": 0.2632069836412605, "grad_norm": 1.6816242933273315, "learning_rate": 8.640360247077708e-06, "loss": 0.5414, "step": 6279 }, { "epoch": 0.26324890225626946, "grad_norm": 1.9453452825546265, "learning_rate": 8.639894859861646e-06, "loss": 0.5166, "step": 6280 }, { "epoch": 0.2632908208712784, "grad_norm": 1.7596869468688965, "learning_rate": 8.639429405549153e-06, "loss": 0.5528, "step": 6281 }, { "epoch": 0.2633327394862874, "grad_norm": 1.8387798070907593, "learning_rate": 8.638963884148811e-06, "loss": 0.5258, "step": 6282 }, { "epoch": 0.26337465810129634, "grad_norm": 1.6455355882644653, "learning_rate": 8.6384982956692e-06, "loss": 0.508, "step": 6283 }, { "epoch": 0.2634165767163053, "grad_norm": 1.750296950340271, "learning_rate": 8.638032640118902e-06, "loss": 0.4926, "step": 6284 }, { "epoch": 0.26345849533131427, "grad_norm": 1.6723806858062744, "learning_rate": 8.637566917506504e-06, "loss": 0.5439, "step": 6285 }, { "epoch": 0.26350041394632323, "grad_norm": 1.8870155811309814, "learning_rate": 8.637101127840587e-06, "loss": 0.5427, "step": 6286 }, { "epoch": 0.2635423325613322, "grad_norm": 1.7640670537948608, "learning_rate": 8.636635271129741e-06, "loss": 0.5486, "step": 6287 }, { "epoch": 0.26358425117634116, "grad_norm": 1.6893720626831055, "learning_rate": 8.636169347382551e-06, "loss": 0.5591, "step": 6288 }, { "epoch": 0.2636261697913501, "grad_norm": 1.5906131267547607, "learning_rate": 8.635703356607604e-06, "loss": 0.5127, "step": 6289 }, { "epoch": 0.2636680884063591, "grad_norm": 1.7612277269363403, "learning_rate": 8.635237298813492e-06, "loss": 0.493, "step": 6290 }, { "epoch": 0.263710007021368, "grad_norm": 1.8168542385101318, "learning_rate": 8.634771174008807e-06, "loss": 0.568, "step": 6291 }, { "epoch": 0.26375192563637695, "grad_norm": 1.6265254020690918, "learning_rate": 8.63430498220214e-06, "loss": 0.5328, "step": 6292 }, { "epoch": 0.2637938442513859, "grad_norm": 1.7474298477172852, "learning_rate": 8.633838723402084e-06, "loss": 0.5496, "step": 6293 }, { "epoch": 0.2638357628663949, "grad_norm": 1.5631452798843384, "learning_rate": 8.633372397617236e-06, "loss": 0.5727, "step": 6294 }, { "epoch": 0.26387768148140384, "grad_norm": 1.553151249885559, "learning_rate": 8.63290600485619e-06, "loss": 0.5256, "step": 6295 }, { "epoch": 0.2639196000964128, "grad_norm": 1.508402705192566, "learning_rate": 8.632439545127544e-06, "loss": 0.5274, "step": 6296 }, { "epoch": 0.26396151871142176, "grad_norm": 1.5356354713439941, "learning_rate": 8.631973018439898e-06, "loss": 0.5064, "step": 6297 }, { "epoch": 0.2640034373264307, "grad_norm": 1.6213834285736084, "learning_rate": 8.63150642480185e-06, "loss": 0.5539, "step": 6298 }, { "epoch": 0.2640453559414397, "grad_norm": 1.6036932468414307, "learning_rate": 8.631039764222e-06, "loss": 0.5327, "step": 6299 }, { "epoch": 0.26408727455644865, "grad_norm": 1.551047682762146, "learning_rate": 8.630573036708954e-06, "loss": 0.5152, "step": 6300 }, { "epoch": 0.2641291931714576, "grad_norm": 1.8165172338485718, "learning_rate": 8.630106242271312e-06, "loss": 0.5517, "step": 6301 }, { "epoch": 0.2641711117864666, "grad_norm": 1.6347852945327759, "learning_rate": 8.629639380917679e-06, "loss": 0.5716, "step": 6302 }, { "epoch": 0.26421303040147553, "grad_norm": 1.8749274015426636, "learning_rate": 8.629172452656662e-06, "loss": 0.5694, "step": 6303 }, { "epoch": 0.2642549490164845, "grad_norm": 1.3893245458602905, "learning_rate": 8.62870545749687e-06, "loss": 0.4997, "step": 6304 }, { "epoch": 0.26429686763149346, "grad_norm": 1.8180145025253296, "learning_rate": 8.628238395446908e-06, "loss": 0.5637, "step": 6305 }, { "epoch": 0.2643387862465024, "grad_norm": 1.9910246133804321, "learning_rate": 8.627771266515388e-06, "loss": 0.4753, "step": 6306 }, { "epoch": 0.2643807048615114, "grad_norm": 1.4728916883468628, "learning_rate": 8.627304070710917e-06, "loss": 0.4784, "step": 6307 }, { "epoch": 0.26442262347652035, "grad_norm": 1.4502832889556885, "learning_rate": 8.626836808042113e-06, "loss": 0.5061, "step": 6308 }, { "epoch": 0.2644645420915293, "grad_norm": 1.6390695571899414, "learning_rate": 8.626369478517587e-06, "loss": 0.535, "step": 6309 }, { "epoch": 0.26450646070653827, "grad_norm": 1.65577232837677, "learning_rate": 8.625902082145952e-06, "loss": 0.4948, "step": 6310 }, { "epoch": 0.26454837932154723, "grad_norm": 1.735482096672058, "learning_rate": 8.625434618935824e-06, "loss": 0.586, "step": 6311 }, { "epoch": 0.2645902979365562, "grad_norm": 1.5158555507659912, "learning_rate": 8.62496708889582e-06, "loss": 0.5108, "step": 6312 }, { "epoch": 0.26463221655156516, "grad_norm": 1.5676803588867188, "learning_rate": 8.624499492034562e-06, "loss": 0.5453, "step": 6313 }, { "epoch": 0.2646741351665741, "grad_norm": 1.795926809310913, "learning_rate": 8.624031828360664e-06, "loss": 0.5253, "step": 6314 }, { "epoch": 0.2647160537815831, "grad_norm": 1.576221227645874, "learning_rate": 8.623564097882751e-06, "loss": 0.5316, "step": 6315 }, { "epoch": 0.264757972396592, "grad_norm": 1.5043681859970093, "learning_rate": 8.623096300609441e-06, "loss": 0.4848, "step": 6316 }, { "epoch": 0.26479989101160095, "grad_norm": 1.656893014907837, "learning_rate": 8.62262843654936e-06, "loss": 0.5328, "step": 6317 }, { "epoch": 0.2648418096266099, "grad_norm": 1.5475431680679321, "learning_rate": 8.622160505711135e-06, "loss": 0.5551, "step": 6318 }, { "epoch": 0.2648837282416189, "grad_norm": 1.7337782382965088, "learning_rate": 8.621692508103385e-06, "loss": 0.5385, "step": 6319 }, { "epoch": 0.26492564685662784, "grad_norm": 1.348482370376587, "learning_rate": 8.621224443734742e-06, "loss": 0.503, "step": 6320 }, { "epoch": 0.2649675654716368, "grad_norm": 2.535133123397827, "learning_rate": 8.620756312613833e-06, "loss": 0.5281, "step": 6321 }, { "epoch": 0.26500948408664576, "grad_norm": 1.4844539165496826, "learning_rate": 8.620288114749286e-06, "loss": 0.5325, "step": 6322 }, { "epoch": 0.2650514027016547, "grad_norm": 1.4342520236968994, "learning_rate": 8.619819850149732e-06, "loss": 0.5256, "step": 6323 }, { "epoch": 0.2650933213166637, "grad_norm": 2.0271246433258057, "learning_rate": 8.619351518823806e-06, "loss": 0.5382, "step": 6324 }, { "epoch": 0.26513523993167265, "grad_norm": 1.6457698345184326, "learning_rate": 8.618883120780135e-06, "loss": 0.5079, "step": 6325 }, { "epoch": 0.2651771585466816, "grad_norm": 1.8502109050750732, "learning_rate": 8.618414656027358e-06, "loss": 0.4611, "step": 6326 }, { "epoch": 0.2652190771616906, "grad_norm": 1.5803239345550537, "learning_rate": 8.61794612457411e-06, "loss": 0.5111, "step": 6327 }, { "epoch": 0.26526099577669954, "grad_norm": 1.918705701828003, "learning_rate": 8.617477526429027e-06, "loss": 0.5199, "step": 6328 }, { "epoch": 0.2653029143917085, "grad_norm": 3.396691083908081, "learning_rate": 8.617008861600746e-06, "loss": 0.5552, "step": 6329 }, { "epoch": 0.26534483300671746, "grad_norm": 1.622869849205017, "learning_rate": 8.616540130097905e-06, "loss": 0.561, "step": 6330 }, { "epoch": 0.2653867516217264, "grad_norm": 2.4407074451446533, "learning_rate": 8.61607133192915e-06, "loss": 0.508, "step": 6331 }, { "epoch": 0.2654286702367354, "grad_norm": 1.56794273853302, "learning_rate": 8.615602467103117e-06, "loss": 0.5775, "step": 6332 }, { "epoch": 0.26547058885174435, "grad_norm": 1.646751046180725, "learning_rate": 8.615133535628452e-06, "loss": 0.5319, "step": 6333 }, { "epoch": 0.2655125074667533, "grad_norm": 1.8161252737045288, "learning_rate": 8.614664537513797e-06, "loss": 0.5218, "step": 6334 }, { "epoch": 0.2655544260817623, "grad_norm": 1.5775760412216187, "learning_rate": 8.614195472767798e-06, "loss": 0.5129, "step": 6335 }, { "epoch": 0.26559634469677124, "grad_norm": 2.1288039684295654, "learning_rate": 8.613726341399104e-06, "loss": 0.5715, "step": 6336 }, { "epoch": 0.2656382633117802, "grad_norm": 1.5833678245544434, "learning_rate": 8.61325714341636e-06, "loss": 0.486, "step": 6337 }, { "epoch": 0.26568018192678916, "grad_norm": 1.5321555137634277, "learning_rate": 8.612787878828215e-06, "loss": 0.5129, "step": 6338 }, { "epoch": 0.2657221005417981, "grad_norm": 1.68264639377594, "learning_rate": 8.612318547643319e-06, "loss": 0.5201, "step": 6339 }, { "epoch": 0.2657640191568071, "grad_norm": 1.7791900634765625, "learning_rate": 8.611849149870325e-06, "loss": 0.5458, "step": 6340 }, { "epoch": 0.265805937771816, "grad_norm": 2.00148868560791, "learning_rate": 8.611379685517887e-06, "loss": 0.5724, "step": 6341 }, { "epoch": 0.26584785638682495, "grad_norm": 1.5537580251693726, "learning_rate": 8.610910154594656e-06, "loss": 0.5371, "step": 6342 }, { "epoch": 0.2658897750018339, "grad_norm": 1.8892463445663452, "learning_rate": 8.610440557109288e-06, "loss": 0.5404, "step": 6343 }, { "epoch": 0.2659316936168429, "grad_norm": 1.6676081418991089, "learning_rate": 8.60997089307044e-06, "loss": 0.5144, "step": 6344 }, { "epoch": 0.26597361223185184, "grad_norm": 4.111729145050049, "learning_rate": 8.609501162486768e-06, "loss": 0.5243, "step": 6345 }, { "epoch": 0.2660155308468608, "grad_norm": 1.6006940603256226, "learning_rate": 8.609031365366933e-06, "loss": 0.5219, "step": 6346 }, { "epoch": 0.26605744946186977, "grad_norm": 1.6679043769836426, "learning_rate": 8.608561501719594e-06, "loss": 0.5703, "step": 6347 }, { "epoch": 0.26609936807687873, "grad_norm": 1.6281875371932983, "learning_rate": 8.608091571553415e-06, "loss": 0.4901, "step": 6348 }, { "epoch": 0.2661412866918877, "grad_norm": 2.8677947521209717, "learning_rate": 8.607621574877053e-06, "loss": 0.5424, "step": 6349 }, { "epoch": 0.26618320530689665, "grad_norm": 1.6505979299545288, "learning_rate": 8.607151511699175e-06, "loss": 0.5011, "step": 6350 }, { "epoch": 0.2662251239219056, "grad_norm": 1.833345651626587, "learning_rate": 8.606681382028444e-06, "loss": 0.5272, "step": 6351 }, { "epoch": 0.2662670425369146, "grad_norm": 1.6249881982803345, "learning_rate": 8.606211185873529e-06, "loss": 0.5197, "step": 6352 }, { "epoch": 0.26630896115192354, "grad_norm": 1.7739845514297485, "learning_rate": 8.605740923243098e-06, "loss": 0.5477, "step": 6353 }, { "epoch": 0.2663508797669325, "grad_norm": 1.5424575805664062, "learning_rate": 8.605270594145816e-06, "loss": 0.4972, "step": 6354 }, { "epoch": 0.26639279838194146, "grad_norm": 1.7642533779144287, "learning_rate": 8.604800198590355e-06, "loss": 0.4951, "step": 6355 }, { "epoch": 0.2664347169969504, "grad_norm": 1.6066956520080566, "learning_rate": 8.604329736585386e-06, "loss": 0.501, "step": 6356 }, { "epoch": 0.2664766356119594, "grad_norm": 1.9140654802322388, "learning_rate": 8.603859208139582e-06, "loss": 0.5195, "step": 6357 }, { "epoch": 0.26651855422696835, "grad_norm": 1.7191112041473389, "learning_rate": 8.603388613261615e-06, "loss": 0.5046, "step": 6358 }, { "epoch": 0.2665604728419773, "grad_norm": 2.331679105758667, "learning_rate": 8.602917951960162e-06, "loss": 0.5545, "step": 6359 }, { "epoch": 0.2666023914569863, "grad_norm": 1.5666898488998413, "learning_rate": 8.602447224243895e-06, "loss": 0.5277, "step": 6360 }, { "epoch": 0.26664431007199524, "grad_norm": 1.70162034034729, "learning_rate": 8.601976430121494e-06, "loss": 0.5656, "step": 6361 }, { "epoch": 0.2666862286870042, "grad_norm": 1.5710945129394531, "learning_rate": 8.601505569601638e-06, "loss": 0.5078, "step": 6362 }, { "epoch": 0.26672814730201316, "grad_norm": 1.634114146232605, "learning_rate": 8.601034642693006e-06, "loss": 0.5191, "step": 6363 }, { "epoch": 0.2667700659170221, "grad_norm": 1.6388194561004639, "learning_rate": 8.600563649404276e-06, "loss": 0.5506, "step": 6364 }, { "epoch": 0.2668119845320311, "grad_norm": 1.9112716913223267, "learning_rate": 8.600092589744136e-06, "loss": 0.5455, "step": 6365 }, { "epoch": 0.26685390314704, "grad_norm": 1.6352192163467407, "learning_rate": 8.599621463721264e-06, "loss": 0.519, "step": 6366 }, { "epoch": 0.26689582176204896, "grad_norm": 1.799927830696106, "learning_rate": 8.599150271344347e-06, "loss": 0.5571, "step": 6367 }, { "epoch": 0.2669377403770579, "grad_norm": 1.791700839996338, "learning_rate": 8.598679012622072e-06, "loss": 0.5492, "step": 6368 }, { "epoch": 0.2669796589920669, "grad_norm": 2.143739938735962, "learning_rate": 8.598207687563124e-06, "loss": 0.5815, "step": 6369 }, { "epoch": 0.26702157760707584, "grad_norm": 3.650601863861084, "learning_rate": 8.59773629617619e-06, "loss": 0.504, "step": 6370 }, { "epoch": 0.2670634962220848, "grad_norm": 1.6025148630142212, "learning_rate": 8.597264838469962e-06, "loss": 0.5545, "step": 6371 }, { "epoch": 0.26710541483709377, "grad_norm": 2.0337741374969482, "learning_rate": 8.596793314453131e-06, "loss": 0.6612, "step": 6372 }, { "epoch": 0.26714733345210273, "grad_norm": 1.7454291582107544, "learning_rate": 8.596321724134386e-06, "loss": 0.5225, "step": 6373 }, { "epoch": 0.2671892520671117, "grad_norm": 1.763407826423645, "learning_rate": 8.595850067522422e-06, "loss": 0.5352, "step": 6374 }, { "epoch": 0.26723117068212066, "grad_norm": 1.9147754907608032, "learning_rate": 8.595378344625933e-06, "loss": 0.5466, "step": 6375 }, { "epoch": 0.2672730892971296, "grad_norm": 1.5418691635131836, "learning_rate": 8.594906555453616e-06, "loss": 0.509, "step": 6376 }, { "epoch": 0.2673150079121386, "grad_norm": 3.994637966156006, "learning_rate": 8.594434700014165e-06, "loss": 0.5695, "step": 6377 }, { "epoch": 0.26735692652714754, "grad_norm": 1.800764560699463, "learning_rate": 8.59396277831628e-06, "loss": 0.5238, "step": 6378 }, { "epoch": 0.2673988451421565, "grad_norm": 1.572845697402954, "learning_rate": 8.59349079036866e-06, "loss": 0.518, "step": 6379 }, { "epoch": 0.26744076375716547, "grad_norm": 2.0555968284606934, "learning_rate": 8.593018736180005e-06, "loss": 0.5241, "step": 6380 }, { "epoch": 0.26748268237217443, "grad_norm": 1.994521975517273, "learning_rate": 8.592546615759016e-06, "loss": 0.5683, "step": 6381 }, { "epoch": 0.2675246009871834, "grad_norm": 2.02359938621521, "learning_rate": 8.592074429114398e-06, "loss": 0.5804, "step": 6382 }, { "epoch": 0.26756651960219235, "grad_norm": 1.7308388948440552, "learning_rate": 8.591602176254854e-06, "loss": 0.5162, "step": 6383 }, { "epoch": 0.2676084382172013, "grad_norm": 1.8546489477157593, "learning_rate": 8.591129857189088e-06, "loss": 0.564, "step": 6384 }, { "epoch": 0.2676503568322103, "grad_norm": 2.0874745845794678, "learning_rate": 8.590657471925807e-06, "loss": 0.6022, "step": 6385 }, { "epoch": 0.26769227544721924, "grad_norm": 3.0167312622070312, "learning_rate": 8.590185020473722e-06, "loss": 0.4899, "step": 6386 }, { "epoch": 0.2677341940622282, "grad_norm": 1.8043556213378906, "learning_rate": 8.589712502841536e-06, "loss": 0.5195, "step": 6387 }, { "epoch": 0.26777611267723717, "grad_norm": 1.886093258857727, "learning_rate": 8.589239919037965e-06, "loss": 0.5573, "step": 6388 }, { "epoch": 0.26781803129224613, "grad_norm": 1.82407546043396, "learning_rate": 8.588767269071718e-06, "loss": 0.5975, "step": 6389 }, { "epoch": 0.2678599499072551, "grad_norm": 1.7812175750732422, "learning_rate": 8.588294552951508e-06, "loss": 0.5212, "step": 6390 }, { "epoch": 0.267901868522264, "grad_norm": 1.8151837587356567, "learning_rate": 8.587821770686046e-06, "loss": 0.5596, "step": 6391 }, { "epoch": 0.26794378713727296, "grad_norm": 1.7154126167297363, "learning_rate": 8.587348922284053e-06, "loss": 0.5406, "step": 6392 }, { "epoch": 0.2679857057522819, "grad_norm": 1.7533767223358154, "learning_rate": 8.58687600775424e-06, "loss": 0.5179, "step": 6393 }, { "epoch": 0.2680276243672909, "grad_norm": 1.5648908615112305, "learning_rate": 8.586403027105327e-06, "loss": 0.4636, "step": 6394 }, { "epoch": 0.26806954298229985, "grad_norm": 1.7612848281860352, "learning_rate": 8.585929980346033e-06, "loss": 0.5894, "step": 6395 }, { "epoch": 0.2681114615973088, "grad_norm": 1.6301889419555664, "learning_rate": 8.585456867485075e-06, "loss": 0.4739, "step": 6396 }, { "epoch": 0.26815338021231777, "grad_norm": 1.6036303043365479, "learning_rate": 8.584983688531179e-06, "loss": 0.5385, "step": 6397 }, { "epoch": 0.26819529882732673, "grad_norm": 1.9581608772277832, "learning_rate": 8.584510443493063e-06, "loss": 0.6108, "step": 6398 }, { "epoch": 0.2682372174423357, "grad_norm": 1.7566556930541992, "learning_rate": 8.584037132379453e-06, "loss": 0.6013, "step": 6399 }, { "epoch": 0.26827913605734466, "grad_norm": 1.7773586511611938, "learning_rate": 8.583563755199074e-06, "loss": 0.5277, "step": 6400 }, { "epoch": 0.2683210546723536, "grad_norm": 1.5124157667160034, "learning_rate": 8.583090311960649e-06, "loss": 0.4666, "step": 6401 }, { "epoch": 0.2683629732873626, "grad_norm": 1.8927451372146606, "learning_rate": 8.582616802672909e-06, "loss": 0.5638, "step": 6402 }, { "epoch": 0.26840489190237155, "grad_norm": 1.987288475036621, "learning_rate": 8.582143227344581e-06, "loss": 0.5593, "step": 6403 }, { "epoch": 0.2684468105173805, "grad_norm": 2.1222658157348633, "learning_rate": 8.581669585984396e-06, "loss": 0.5764, "step": 6404 }, { "epoch": 0.26848872913238947, "grad_norm": 1.8130229711532593, "learning_rate": 8.581195878601082e-06, "loss": 0.5698, "step": 6405 }, { "epoch": 0.26853064774739843, "grad_norm": 1.9798282384872437, "learning_rate": 8.580722105203373e-06, "loss": 0.5864, "step": 6406 }, { "epoch": 0.2685725663624074, "grad_norm": 2.015652656555176, "learning_rate": 8.580248265800004e-06, "loss": 0.59, "step": 6407 }, { "epoch": 0.26861448497741636, "grad_norm": 1.6284273862838745, "learning_rate": 8.579774360399706e-06, "loss": 0.5278, "step": 6408 }, { "epoch": 0.2686564035924253, "grad_norm": 1.742936372756958, "learning_rate": 8.579300389011219e-06, "loss": 0.5734, "step": 6409 }, { "epoch": 0.2686983222074343, "grad_norm": 1.5439695119857788, "learning_rate": 8.578826351643274e-06, "loss": 0.5409, "step": 6410 }, { "epoch": 0.26874024082244324, "grad_norm": 1.5878146886825562, "learning_rate": 8.578352248304616e-06, "loss": 0.4998, "step": 6411 }, { "epoch": 0.2687821594374522, "grad_norm": 1.8509266376495361, "learning_rate": 8.577878079003979e-06, "loss": 0.5127, "step": 6412 }, { "epoch": 0.26882407805246117, "grad_norm": 1.4850091934204102, "learning_rate": 8.577403843750105e-06, "loss": 0.4978, "step": 6413 }, { "epoch": 0.26886599666747013, "grad_norm": 1.6369085311889648, "learning_rate": 8.57692954255174e-06, "loss": 0.5663, "step": 6414 }, { "epoch": 0.2689079152824791, "grad_norm": 1.6559475660324097, "learning_rate": 8.57645517541762e-06, "loss": 0.516, "step": 6415 }, { "epoch": 0.268949833897488, "grad_norm": 1.6702251434326172, "learning_rate": 8.575980742356496e-06, "loss": 0.5722, "step": 6416 }, { "epoch": 0.26899175251249696, "grad_norm": 1.8098725080490112, "learning_rate": 8.57550624337711e-06, "loss": 0.566, "step": 6417 }, { "epoch": 0.2690336711275059, "grad_norm": 1.6246155500411987, "learning_rate": 8.575031678488209e-06, "loss": 0.5522, "step": 6418 }, { "epoch": 0.2690755897425149, "grad_norm": 1.94357168674469, "learning_rate": 8.574557047698541e-06, "loss": 0.5606, "step": 6419 }, { "epoch": 0.26911750835752385, "grad_norm": 1.9754977226257324, "learning_rate": 8.574082351016857e-06, "loss": 0.5853, "step": 6420 }, { "epoch": 0.2691594269725328, "grad_norm": 1.812328815460205, "learning_rate": 8.573607588451904e-06, "loss": 0.5436, "step": 6421 }, { "epoch": 0.2692013455875418, "grad_norm": 1.7126719951629639, "learning_rate": 8.573132760012437e-06, "loss": 0.5086, "step": 6422 }, { "epoch": 0.26924326420255074, "grad_norm": 1.541945457458496, "learning_rate": 8.572657865707207e-06, "loss": 0.5511, "step": 6423 }, { "epoch": 0.2692851828175597, "grad_norm": 1.5816137790679932, "learning_rate": 8.572182905544968e-06, "loss": 0.5047, "step": 6424 }, { "epoch": 0.26932710143256866, "grad_norm": 1.751250147819519, "learning_rate": 8.571707879534476e-06, "loss": 0.5157, "step": 6425 }, { "epoch": 0.2693690200475776, "grad_norm": 1.679944634437561, "learning_rate": 8.571232787684487e-06, "loss": 0.5612, "step": 6426 }, { "epoch": 0.2694109386625866, "grad_norm": 2.087174415588379, "learning_rate": 8.570757630003758e-06, "loss": 0.5498, "step": 6427 }, { "epoch": 0.26945285727759555, "grad_norm": 1.8537003993988037, "learning_rate": 8.570282406501048e-06, "loss": 0.5465, "step": 6428 }, { "epoch": 0.2694947758926045, "grad_norm": 1.5825682878494263, "learning_rate": 8.569807117185119e-06, "loss": 0.4897, "step": 6429 }, { "epoch": 0.2695366945076135, "grad_norm": 1.7826449871063232, "learning_rate": 8.56933176206473e-06, "loss": 0.4905, "step": 6430 }, { "epoch": 0.26957861312262243, "grad_norm": 1.7228926420211792, "learning_rate": 8.568856341148643e-06, "loss": 0.526, "step": 6431 }, { "epoch": 0.2696205317376314, "grad_norm": 1.689178228378296, "learning_rate": 8.568380854445627e-06, "loss": 0.5196, "step": 6432 }, { "epoch": 0.26966245035264036, "grad_norm": 1.6509671211242676, "learning_rate": 8.56790530196444e-06, "loss": 0.5416, "step": 6433 }, { "epoch": 0.2697043689676493, "grad_norm": 1.5597914457321167, "learning_rate": 8.567429683713853e-06, "loss": 0.5341, "step": 6434 }, { "epoch": 0.2697462875826583, "grad_norm": 1.8978638648986816, "learning_rate": 8.566953999702629e-06, "loss": 0.5316, "step": 6435 }, { "epoch": 0.26978820619766725, "grad_norm": 2.1538374423980713, "learning_rate": 8.566478249939541e-06, "loss": 0.541, "step": 6436 }, { "epoch": 0.2698301248126762, "grad_norm": 1.811277985572815, "learning_rate": 8.566002434433357e-06, "loss": 0.484, "step": 6437 }, { "epoch": 0.26987204342768517, "grad_norm": 1.6844146251678467, "learning_rate": 8.56552655319285e-06, "loss": 0.4885, "step": 6438 }, { "epoch": 0.26991396204269413, "grad_norm": 2.064981698989868, "learning_rate": 8.565050606226786e-06, "loss": 0.5941, "step": 6439 }, { "epoch": 0.2699558806577031, "grad_norm": 1.9011881351470947, "learning_rate": 8.564574593543945e-06, "loss": 0.5412, "step": 6440 }, { "epoch": 0.269997799272712, "grad_norm": 1.8152074813842773, "learning_rate": 8.564098515153098e-06, "loss": 0.5527, "step": 6441 }, { "epoch": 0.27003971788772096, "grad_norm": 1.547337293624878, "learning_rate": 8.563622371063024e-06, "loss": 0.51, "step": 6442 }, { "epoch": 0.2700816365027299, "grad_norm": 1.7246979475021362, "learning_rate": 8.563146161282497e-06, "loss": 0.505, "step": 6443 }, { "epoch": 0.2701235551177389, "grad_norm": 1.5446079969406128, "learning_rate": 8.562669885820296e-06, "loss": 0.604, "step": 6444 }, { "epoch": 0.27016547373274785, "grad_norm": 1.7183406352996826, "learning_rate": 8.562193544685202e-06, "loss": 0.5795, "step": 6445 }, { "epoch": 0.2702073923477568, "grad_norm": 1.7099114656448364, "learning_rate": 8.561717137885994e-06, "loss": 0.5223, "step": 6446 }, { "epoch": 0.2702493109627658, "grad_norm": 1.9307433366775513, "learning_rate": 8.561240665431455e-06, "loss": 0.539, "step": 6447 }, { "epoch": 0.27029122957777474, "grad_norm": 1.722744107246399, "learning_rate": 8.560764127330366e-06, "loss": 0.5481, "step": 6448 }, { "epoch": 0.2703331481927837, "grad_norm": 1.8700945377349854, "learning_rate": 8.560287523591515e-06, "loss": 0.5686, "step": 6449 }, { "epoch": 0.27037506680779266, "grad_norm": 1.6801196336746216, "learning_rate": 8.559810854223684e-06, "loss": 0.5068, "step": 6450 }, { "epoch": 0.2704169854228016, "grad_norm": 1.6266696453094482, "learning_rate": 8.559334119235662e-06, "loss": 0.5148, "step": 6451 }, { "epoch": 0.2704589040378106, "grad_norm": 1.7214765548706055, "learning_rate": 8.558857318636235e-06, "loss": 0.5367, "step": 6452 }, { "epoch": 0.27050082265281955, "grad_norm": 1.9088207483291626, "learning_rate": 8.558380452434193e-06, "loss": 0.5359, "step": 6453 }, { "epoch": 0.2705427412678285, "grad_norm": 1.5256643295288086, "learning_rate": 8.557903520638328e-06, "loss": 0.5104, "step": 6454 }, { "epoch": 0.2705846598828375, "grad_norm": 1.7328373193740845, "learning_rate": 8.55742652325743e-06, "loss": 0.5632, "step": 6455 }, { "epoch": 0.27062657849784644, "grad_norm": 1.783282995223999, "learning_rate": 8.556949460300294e-06, "loss": 0.5556, "step": 6456 }, { "epoch": 0.2706684971128554, "grad_norm": 2.8252673149108887, "learning_rate": 8.55647233177571e-06, "loss": 0.5411, "step": 6457 }, { "epoch": 0.27071041572786436, "grad_norm": 1.6527341604232788, "learning_rate": 8.555995137692474e-06, "loss": 0.4694, "step": 6458 }, { "epoch": 0.2707523343428733, "grad_norm": 2.0570967197418213, "learning_rate": 8.555517878059386e-06, "loss": 0.4856, "step": 6459 }, { "epoch": 0.2707942529578823, "grad_norm": 1.8975683450698853, "learning_rate": 8.555040552885242e-06, "loss": 0.5178, "step": 6460 }, { "epoch": 0.27083617157289125, "grad_norm": 1.7535994052886963, "learning_rate": 8.554563162178837e-06, "loss": 0.4646, "step": 6461 }, { "epoch": 0.2708780901879002, "grad_norm": 2.393277168273926, "learning_rate": 8.554085705948975e-06, "loss": 0.5796, "step": 6462 }, { "epoch": 0.2709200088029092, "grad_norm": 1.7703171968460083, "learning_rate": 8.553608184204456e-06, "loss": 0.5692, "step": 6463 }, { "epoch": 0.27096192741791814, "grad_norm": 1.7168787717819214, "learning_rate": 8.553130596954084e-06, "loss": 0.5827, "step": 6464 }, { "epoch": 0.2710038460329271, "grad_norm": 1.6195210218429565, "learning_rate": 8.552652944206661e-06, "loss": 0.5104, "step": 6465 }, { "epoch": 0.271045764647936, "grad_norm": 1.9295909404754639, "learning_rate": 8.552175225970993e-06, "loss": 0.5238, "step": 6466 }, { "epoch": 0.27108768326294497, "grad_norm": 1.6224355697631836, "learning_rate": 8.551697442255887e-06, "loss": 0.5193, "step": 6467 }, { "epoch": 0.27112960187795393, "grad_norm": 1.6479296684265137, "learning_rate": 8.551219593070146e-06, "loss": 0.5505, "step": 6468 }, { "epoch": 0.2711715204929629, "grad_norm": 1.7354739904403687, "learning_rate": 8.55074167842258e-06, "loss": 0.5927, "step": 6469 }, { "epoch": 0.27121343910797185, "grad_norm": 1.8711426258087158, "learning_rate": 8.550263698322002e-06, "loss": 0.517, "step": 6470 }, { "epoch": 0.2712553577229808, "grad_norm": 1.6122095584869385, "learning_rate": 8.549785652777221e-06, "loss": 0.5055, "step": 6471 }, { "epoch": 0.2712972763379898, "grad_norm": 1.4626775979995728, "learning_rate": 8.54930754179705e-06, "loss": 0.4957, "step": 6472 }, { "epoch": 0.27133919495299874, "grad_norm": 1.6242728233337402, "learning_rate": 8.548829365390299e-06, "loss": 0.5458, "step": 6473 }, { "epoch": 0.2713811135680077, "grad_norm": 2.2074625492095947, "learning_rate": 8.548351123565785e-06, "loss": 0.5599, "step": 6474 }, { "epoch": 0.27142303218301667, "grad_norm": 1.816361904144287, "learning_rate": 8.547872816332326e-06, "loss": 0.5246, "step": 6475 }, { "epoch": 0.27146495079802563, "grad_norm": 1.8056693077087402, "learning_rate": 8.547394443698734e-06, "loss": 0.5255, "step": 6476 }, { "epoch": 0.2715068694130346, "grad_norm": 1.6881202459335327, "learning_rate": 8.54691600567383e-06, "loss": 0.5855, "step": 6477 }, { "epoch": 0.27154878802804355, "grad_norm": 1.6400716304779053, "learning_rate": 8.546437502266432e-06, "loss": 0.4944, "step": 6478 }, { "epoch": 0.2715907066430525, "grad_norm": 2.028026580810547, "learning_rate": 8.545958933485363e-06, "loss": 0.5373, "step": 6479 }, { "epoch": 0.2716326252580615, "grad_norm": 1.5568963289260864, "learning_rate": 8.545480299339443e-06, "loss": 0.5005, "step": 6480 }, { "epoch": 0.27167454387307044, "grad_norm": 1.6865888833999634, "learning_rate": 8.545001599837494e-06, "loss": 0.5363, "step": 6481 }, { "epoch": 0.2717164624880794, "grad_norm": 1.841353178024292, "learning_rate": 8.544522834988343e-06, "loss": 0.5201, "step": 6482 }, { "epoch": 0.27175838110308836, "grad_norm": 1.8822845220565796, "learning_rate": 8.544044004800814e-06, "loss": 0.553, "step": 6483 }, { "epoch": 0.2718002997180973, "grad_norm": 1.688186526298523, "learning_rate": 8.543565109283732e-06, "loss": 0.5404, "step": 6484 }, { "epoch": 0.2718422183331063, "grad_norm": 1.7328804731369019, "learning_rate": 8.543086148445925e-06, "loss": 0.5032, "step": 6485 }, { "epoch": 0.27188413694811525, "grad_norm": 1.5627217292785645, "learning_rate": 8.542607122296226e-06, "loss": 0.5245, "step": 6486 }, { "epoch": 0.2719260555631242, "grad_norm": 1.5829355716705322, "learning_rate": 8.542128030843459e-06, "loss": 0.5087, "step": 6487 }, { "epoch": 0.2719679741781332, "grad_norm": 1.5452909469604492, "learning_rate": 8.541648874096459e-06, "loss": 0.5408, "step": 6488 }, { "epoch": 0.27200989279314214, "grad_norm": 2.0196218490600586, "learning_rate": 8.54116965206406e-06, "loss": 0.5381, "step": 6489 }, { "epoch": 0.2720518114081511, "grad_norm": 1.7597566843032837, "learning_rate": 8.540690364755093e-06, "loss": 0.4981, "step": 6490 }, { "epoch": 0.27209373002316, "grad_norm": 1.7691859006881714, "learning_rate": 8.540211012178394e-06, "loss": 0.5319, "step": 6491 }, { "epoch": 0.27213564863816897, "grad_norm": 1.6061294078826904, "learning_rate": 8.539731594342799e-06, "loss": 0.5152, "step": 6492 }, { "epoch": 0.27217756725317793, "grad_norm": 1.6171451807022095, "learning_rate": 8.539252111257146e-06, "loss": 0.4888, "step": 6493 }, { "epoch": 0.2722194858681869, "grad_norm": 1.6815212965011597, "learning_rate": 8.538772562930274e-06, "loss": 0.5924, "step": 6494 }, { "epoch": 0.27226140448319586, "grad_norm": 1.7910192012786865, "learning_rate": 8.538292949371021e-06, "loss": 0.4945, "step": 6495 }, { "epoch": 0.2723033230982048, "grad_norm": 1.7947498559951782, "learning_rate": 8.53781327058823e-06, "loss": 0.5081, "step": 6496 }, { "epoch": 0.2723452417132138, "grad_norm": 2.1879348754882812, "learning_rate": 8.53733352659074e-06, "loss": 0.5213, "step": 6497 }, { "epoch": 0.27238716032822274, "grad_norm": 1.3624708652496338, "learning_rate": 8.536853717387398e-06, "loss": 0.5202, "step": 6498 }, { "epoch": 0.2724290789432317, "grad_norm": 1.7973893880844116, "learning_rate": 8.536373842987047e-06, "loss": 0.5554, "step": 6499 }, { "epoch": 0.27247099755824067, "grad_norm": 2.064836025238037, "learning_rate": 8.535893903398533e-06, "loss": 0.5748, "step": 6500 }, { "epoch": 0.27251291617324963, "grad_norm": 1.624122142791748, "learning_rate": 8.535413898630705e-06, "loss": 0.5081, "step": 6501 }, { "epoch": 0.2725548347882586, "grad_norm": 2.0774075984954834, "learning_rate": 8.534933828692406e-06, "loss": 0.5919, "step": 6502 }, { "epoch": 0.27259675340326756, "grad_norm": 2.368540048599243, "learning_rate": 8.534453693592493e-06, "loss": 0.5445, "step": 6503 }, { "epoch": 0.2726386720182765, "grad_norm": 1.514275074005127, "learning_rate": 8.53397349333981e-06, "loss": 0.5168, "step": 6504 }, { "epoch": 0.2726805906332855, "grad_norm": 1.6212290525436401, "learning_rate": 8.53349322794321e-06, "loss": 0.4858, "step": 6505 }, { "epoch": 0.27272250924829444, "grad_norm": 1.7333307266235352, "learning_rate": 8.533012897411547e-06, "loss": 0.5796, "step": 6506 }, { "epoch": 0.2727644278633034, "grad_norm": 1.9093506336212158, "learning_rate": 8.532532501753679e-06, "loss": 0.5668, "step": 6507 }, { "epoch": 0.27280634647831237, "grad_norm": 1.4798239469528198, "learning_rate": 8.532052040978456e-06, "loss": 0.5305, "step": 6508 }, { "epoch": 0.27284826509332133, "grad_norm": 1.9799998998641968, "learning_rate": 8.531571515094735e-06, "loss": 0.5657, "step": 6509 }, { "epoch": 0.2728901837083303, "grad_norm": 1.9570115804672241, "learning_rate": 8.531090924111376e-06, "loss": 0.5551, "step": 6510 }, { "epoch": 0.27293210232333925, "grad_norm": 1.6642229557037354, "learning_rate": 8.530610268037239e-06, "loss": 0.4894, "step": 6511 }, { "epoch": 0.2729740209383482, "grad_norm": 1.8548927307128906, "learning_rate": 8.530129546881184e-06, "loss": 0.5242, "step": 6512 }, { "epoch": 0.2730159395533572, "grad_norm": 1.7283612489700317, "learning_rate": 8.529648760652067e-06, "loss": 0.5694, "step": 6513 }, { "epoch": 0.27305785816836614, "grad_norm": 1.479478359222412, "learning_rate": 8.529167909358757e-06, "loss": 0.5454, "step": 6514 }, { "epoch": 0.2730997767833751, "grad_norm": 1.6374671459197998, "learning_rate": 8.528686993010116e-06, "loss": 0.5212, "step": 6515 }, { "epoch": 0.273141695398384, "grad_norm": 1.8432955741882324, "learning_rate": 8.528206011615006e-06, "loss": 0.5251, "step": 6516 }, { "epoch": 0.273183614013393, "grad_norm": 2.0708155632019043, "learning_rate": 8.527724965182299e-06, "loss": 0.5883, "step": 6517 }, { "epoch": 0.27322553262840193, "grad_norm": 1.5687958002090454, "learning_rate": 8.527243853720856e-06, "loss": 0.5136, "step": 6518 }, { "epoch": 0.2732674512434109, "grad_norm": 1.589406132698059, "learning_rate": 8.52676267723955e-06, "loss": 0.5085, "step": 6519 }, { "epoch": 0.27330936985841986, "grad_norm": 1.7478501796722412, "learning_rate": 8.52628143574725e-06, "loss": 0.5309, "step": 6520 }, { "epoch": 0.2733512884734288, "grad_norm": 1.798455834388733, "learning_rate": 8.525800129252829e-06, "loss": 0.5034, "step": 6521 }, { "epoch": 0.2733932070884378, "grad_norm": 1.6307933330535889, "learning_rate": 8.525318757765153e-06, "loss": 0.5361, "step": 6522 }, { "epoch": 0.27343512570344675, "grad_norm": 1.7728930711746216, "learning_rate": 8.524837321293102e-06, "loss": 0.572, "step": 6523 }, { "epoch": 0.2734770443184557, "grad_norm": 1.7791757583618164, "learning_rate": 8.524355819845547e-06, "loss": 0.5585, "step": 6524 }, { "epoch": 0.27351896293346467, "grad_norm": 1.6589374542236328, "learning_rate": 8.523874253431364e-06, "loss": 0.5828, "step": 6525 }, { "epoch": 0.27356088154847363, "grad_norm": 1.6423618793487549, "learning_rate": 8.523392622059431e-06, "loss": 0.5597, "step": 6526 }, { "epoch": 0.2736028001634826, "grad_norm": 2.050665855407715, "learning_rate": 8.522910925738627e-06, "loss": 0.574, "step": 6527 }, { "epoch": 0.27364471877849156, "grad_norm": 1.5653507709503174, "learning_rate": 8.52242916447783e-06, "loss": 0.5347, "step": 6528 }, { "epoch": 0.2736866373935005, "grad_norm": 1.8765751123428345, "learning_rate": 8.52194733828592e-06, "loss": 0.5497, "step": 6529 }, { "epoch": 0.2737285560085095, "grad_norm": 2.2423558235168457, "learning_rate": 8.52146544717178e-06, "loss": 0.5223, "step": 6530 }, { "epoch": 0.27377047462351844, "grad_norm": 1.7128649950027466, "learning_rate": 8.520983491144293e-06, "loss": 0.5945, "step": 6531 }, { "epoch": 0.2738123932385274, "grad_norm": 3.396420955657959, "learning_rate": 8.520501470212343e-06, "loss": 0.5391, "step": 6532 }, { "epoch": 0.27385431185353637, "grad_norm": 1.5532324314117432, "learning_rate": 8.520019384384817e-06, "loss": 0.474, "step": 6533 }, { "epoch": 0.27389623046854533, "grad_norm": 1.7473421096801758, "learning_rate": 8.519537233670596e-06, "loss": 0.6006, "step": 6534 }, { "epoch": 0.2739381490835543, "grad_norm": 1.5133309364318848, "learning_rate": 8.519055018078575e-06, "loss": 0.5247, "step": 6535 }, { "epoch": 0.27398006769856326, "grad_norm": 1.7095779180526733, "learning_rate": 8.51857273761764e-06, "loss": 0.4951, "step": 6536 }, { "epoch": 0.2740219863135722, "grad_norm": 1.688355565071106, "learning_rate": 8.518090392296679e-06, "loss": 0.516, "step": 6537 }, { "epoch": 0.2740639049285812, "grad_norm": 1.585170865058899, "learning_rate": 8.517607982124586e-06, "loss": 0.5355, "step": 6538 }, { "epoch": 0.27410582354359014, "grad_norm": 1.8908239603042603, "learning_rate": 8.517125507110254e-06, "loss": 0.5692, "step": 6539 }, { "epoch": 0.2741477421585991, "grad_norm": 1.825977087020874, "learning_rate": 8.516642967262575e-06, "loss": 0.5643, "step": 6540 }, { "epoch": 0.274189660773608, "grad_norm": 1.836283564567566, "learning_rate": 8.516160362590441e-06, "loss": 0.5639, "step": 6541 }, { "epoch": 0.274231579388617, "grad_norm": 1.7302272319793701, "learning_rate": 8.515677693102755e-06, "loss": 0.5504, "step": 6542 }, { "epoch": 0.27427349800362594, "grad_norm": 1.7767270803451538, "learning_rate": 8.51519495880841e-06, "loss": 0.4978, "step": 6543 }, { "epoch": 0.2743154166186349, "grad_norm": 1.666104793548584, "learning_rate": 8.514712159716307e-06, "loss": 0.5214, "step": 6544 }, { "epoch": 0.27435733523364386, "grad_norm": 2.0644822120666504, "learning_rate": 8.514229295835342e-06, "loss": 0.5703, "step": 6545 }, { "epoch": 0.2743992538486528, "grad_norm": 1.7675073146820068, "learning_rate": 8.51374636717442e-06, "loss": 0.522, "step": 6546 }, { "epoch": 0.2744411724636618, "grad_norm": 2.085507869720459, "learning_rate": 8.513263373742442e-06, "loss": 0.5174, "step": 6547 }, { "epoch": 0.27448309107867075, "grad_norm": 1.9752752780914307, "learning_rate": 8.512780315548307e-06, "loss": 0.5641, "step": 6548 }, { "epoch": 0.2745250096936797, "grad_norm": 1.6287816762924194, "learning_rate": 8.512297192600926e-06, "loss": 0.5577, "step": 6549 }, { "epoch": 0.2745669283086887, "grad_norm": 1.6610023975372314, "learning_rate": 8.511814004909202e-06, "loss": 0.5304, "step": 6550 }, { "epoch": 0.27460884692369764, "grad_norm": 1.6532349586486816, "learning_rate": 8.51133075248204e-06, "loss": 0.5526, "step": 6551 }, { "epoch": 0.2746507655387066, "grad_norm": 2.1667075157165527, "learning_rate": 8.510847435328353e-06, "loss": 0.5056, "step": 6552 }, { "epoch": 0.27469268415371556, "grad_norm": 1.7736108303070068, "learning_rate": 8.510364053457044e-06, "loss": 0.5258, "step": 6553 }, { "epoch": 0.2747346027687245, "grad_norm": 1.576511025428772, "learning_rate": 8.509880606877028e-06, "loss": 0.5358, "step": 6554 }, { "epoch": 0.2747765213837335, "grad_norm": 1.786139965057373, "learning_rate": 8.509397095597216e-06, "loss": 0.5503, "step": 6555 }, { "epoch": 0.27481843999874245, "grad_norm": 1.527234435081482, "learning_rate": 8.508913519626518e-06, "loss": 0.5651, "step": 6556 }, { "epoch": 0.2748603586137514, "grad_norm": 1.8681299686431885, "learning_rate": 8.508429878973852e-06, "loss": 0.5728, "step": 6557 }, { "epoch": 0.27490227722876037, "grad_norm": 1.731303334236145, "learning_rate": 8.507946173648132e-06, "loss": 0.5161, "step": 6558 }, { "epoch": 0.27494419584376933, "grad_norm": 1.6859499216079712, "learning_rate": 8.507462403658271e-06, "loss": 0.531, "step": 6559 }, { "epoch": 0.2749861144587783, "grad_norm": 1.5694036483764648, "learning_rate": 8.506978569013191e-06, "loss": 0.5376, "step": 6560 }, { "epoch": 0.27502803307378726, "grad_norm": 1.5906462669372559, "learning_rate": 8.506494669721812e-06, "loss": 0.5653, "step": 6561 }, { "epoch": 0.2750699516887962, "grad_norm": 1.5672427415847778, "learning_rate": 8.506010705793046e-06, "loss": 0.5585, "step": 6562 }, { "epoch": 0.2751118703038052, "grad_norm": 1.935591697692871, "learning_rate": 8.505526677235824e-06, "loss": 0.5708, "step": 6563 }, { "epoch": 0.27515378891881415, "grad_norm": 1.6644703149795532, "learning_rate": 8.505042584059062e-06, "loss": 0.4883, "step": 6564 }, { "epoch": 0.2751957075338231, "grad_norm": 1.564639925956726, "learning_rate": 8.504558426271685e-06, "loss": 0.5337, "step": 6565 }, { "epoch": 0.275237626148832, "grad_norm": 1.6672322750091553, "learning_rate": 8.50407420388262e-06, "loss": 0.5335, "step": 6566 }, { "epoch": 0.275279544763841, "grad_norm": 1.9243203401565552, "learning_rate": 8.503589916900791e-06, "loss": 0.5711, "step": 6567 }, { "epoch": 0.27532146337884994, "grad_norm": 1.596651554107666, "learning_rate": 8.503105565335124e-06, "loss": 0.5077, "step": 6568 }, { "epoch": 0.2753633819938589, "grad_norm": 1.6881649494171143, "learning_rate": 8.50262114919455e-06, "loss": 0.5815, "step": 6569 }, { "epoch": 0.27540530060886786, "grad_norm": 1.8239349126815796, "learning_rate": 8.502136668487997e-06, "loss": 0.5576, "step": 6570 }, { "epoch": 0.2754472192238768, "grad_norm": 1.4901424646377563, "learning_rate": 8.501652123224397e-06, "loss": 0.5121, "step": 6571 }, { "epoch": 0.2754891378388858, "grad_norm": 1.5432015657424927, "learning_rate": 8.50116751341268e-06, "loss": 0.5711, "step": 6572 }, { "epoch": 0.27553105645389475, "grad_norm": 1.9355467557907104, "learning_rate": 8.50068283906178e-06, "loss": 0.5397, "step": 6573 }, { "epoch": 0.2755729750689037, "grad_norm": 1.5004466772079468, "learning_rate": 8.500198100180633e-06, "loss": 0.5175, "step": 6574 }, { "epoch": 0.2756148936839127, "grad_norm": 2.3040738105773926, "learning_rate": 8.499713296778171e-06, "loss": 0.5284, "step": 6575 }, { "epoch": 0.27565681229892164, "grad_norm": 1.736295223236084, "learning_rate": 8.499228428863333e-06, "loss": 0.5486, "step": 6576 }, { "epoch": 0.2756987309139306, "grad_norm": 1.7585371732711792, "learning_rate": 8.498743496445056e-06, "loss": 0.5322, "step": 6577 }, { "epoch": 0.27574064952893956, "grad_norm": 1.621402382850647, "learning_rate": 8.498258499532281e-06, "loss": 0.5743, "step": 6578 }, { "epoch": 0.2757825681439485, "grad_norm": 1.4963043928146362, "learning_rate": 8.497773438133946e-06, "loss": 0.4404, "step": 6579 }, { "epoch": 0.2758244867589575, "grad_norm": 1.7577458620071411, "learning_rate": 8.497288312258994e-06, "loss": 0.5033, "step": 6580 }, { "epoch": 0.27586640537396645, "grad_norm": 1.5277496576309204, "learning_rate": 8.496803121916365e-06, "loss": 0.4999, "step": 6581 }, { "epoch": 0.2759083239889754, "grad_norm": 1.6056756973266602, "learning_rate": 8.496317867115006e-06, "loss": 0.4928, "step": 6582 }, { "epoch": 0.2759502426039844, "grad_norm": 1.5294703245162964, "learning_rate": 8.49583254786386e-06, "loss": 0.5657, "step": 6583 }, { "epoch": 0.27599216121899334, "grad_norm": 1.7428056001663208, "learning_rate": 8.495347164171875e-06, "loss": 0.5752, "step": 6584 }, { "epoch": 0.2760340798340023, "grad_norm": 1.56936776638031, "learning_rate": 8.494861716047995e-06, "loss": 0.5443, "step": 6585 }, { "epoch": 0.27607599844901126, "grad_norm": 1.5547806024551392, "learning_rate": 8.494376203501172e-06, "loss": 0.567, "step": 6586 }, { "epoch": 0.2761179170640202, "grad_norm": 1.6771986484527588, "learning_rate": 8.493890626540355e-06, "loss": 0.5415, "step": 6587 }, { "epoch": 0.2761598356790292, "grad_norm": 4.171563148498535, "learning_rate": 8.493404985174493e-06, "loss": 0.5674, "step": 6588 }, { "epoch": 0.27620175429403815, "grad_norm": 1.9065059423446655, "learning_rate": 8.492919279412542e-06, "loss": 0.5505, "step": 6589 }, { "epoch": 0.2762436729090471, "grad_norm": 3.5882551670074463, "learning_rate": 8.49243350926345e-06, "loss": 0.5911, "step": 6590 }, { "epoch": 0.276285591524056, "grad_norm": 1.8346259593963623, "learning_rate": 8.491947674736176e-06, "loss": 0.5652, "step": 6591 }, { "epoch": 0.276327510139065, "grad_norm": 1.5462359189987183, "learning_rate": 8.491461775839675e-06, "loss": 0.5844, "step": 6592 }, { "epoch": 0.27636942875407394, "grad_norm": 1.446979284286499, "learning_rate": 8.490975812582901e-06, "loss": 0.5667, "step": 6593 }, { "epoch": 0.2764113473690829, "grad_norm": 1.7247428894042969, "learning_rate": 8.490489784974815e-06, "loss": 0.5468, "step": 6594 }, { "epoch": 0.27645326598409187, "grad_norm": 1.4554864168167114, "learning_rate": 8.490003693024375e-06, "loss": 0.5124, "step": 6595 }, { "epoch": 0.27649518459910083, "grad_norm": 1.4547996520996094, "learning_rate": 8.489517536740542e-06, "loss": 0.463, "step": 6596 }, { "epoch": 0.2765371032141098, "grad_norm": 1.7155078649520874, "learning_rate": 8.489031316132279e-06, "loss": 0.5035, "step": 6597 }, { "epoch": 0.27657902182911875, "grad_norm": 1.5467517375946045, "learning_rate": 8.488545031208544e-06, "loss": 0.5067, "step": 6598 }, { "epoch": 0.2766209404441277, "grad_norm": 1.995274305343628, "learning_rate": 8.488058681978306e-06, "loss": 0.5675, "step": 6599 }, { "epoch": 0.2766628590591367, "grad_norm": 1.712094783782959, "learning_rate": 8.487572268450528e-06, "loss": 0.5179, "step": 6600 }, { "epoch": 0.27670477767414564, "grad_norm": 2.053330898284912, "learning_rate": 8.487085790634176e-06, "loss": 0.518, "step": 6601 }, { "epoch": 0.2767466962891546, "grad_norm": 1.639882206916809, "learning_rate": 8.486599248538218e-06, "loss": 0.5178, "step": 6602 }, { "epoch": 0.27678861490416357, "grad_norm": 1.6332346200942993, "learning_rate": 8.486112642171625e-06, "loss": 0.5213, "step": 6603 }, { "epoch": 0.27683053351917253, "grad_norm": 1.437493920326233, "learning_rate": 8.485625971543362e-06, "loss": 0.4783, "step": 6604 }, { "epoch": 0.2768724521341815, "grad_norm": 1.55033540725708, "learning_rate": 8.485139236662406e-06, "loss": 0.5195, "step": 6605 }, { "epoch": 0.27691437074919045, "grad_norm": 1.8282772302627563, "learning_rate": 8.484652437537724e-06, "loss": 0.5206, "step": 6606 }, { "epoch": 0.2769562893641994, "grad_norm": 1.8827847242355347, "learning_rate": 8.484165574178292e-06, "loss": 0.485, "step": 6607 }, { "epoch": 0.2769982079792084, "grad_norm": 1.8978010416030884, "learning_rate": 8.483678646593084e-06, "loss": 0.6244, "step": 6608 }, { "epoch": 0.27704012659421734, "grad_norm": 1.7404788732528687, "learning_rate": 8.483191654791077e-06, "loss": 0.5593, "step": 6609 }, { "epoch": 0.2770820452092263, "grad_norm": 2.0481529235839844, "learning_rate": 8.482704598781248e-06, "loss": 0.5529, "step": 6610 }, { "epoch": 0.27712396382423526, "grad_norm": 1.4846091270446777, "learning_rate": 8.482217478572571e-06, "loss": 0.4954, "step": 6611 }, { "epoch": 0.2771658824392442, "grad_norm": 1.7927876710891724, "learning_rate": 8.481730294174034e-06, "loss": 0.5259, "step": 6612 }, { "epoch": 0.2772078010542532, "grad_norm": 1.8115590810775757, "learning_rate": 8.48124304559461e-06, "loss": 0.4945, "step": 6613 }, { "epoch": 0.27724971966926215, "grad_norm": 2.020768880844116, "learning_rate": 8.480755732843282e-06, "loss": 0.5478, "step": 6614 }, { "epoch": 0.2772916382842711, "grad_norm": 1.6666743755340576, "learning_rate": 8.480268355929035e-06, "loss": 0.6142, "step": 6615 }, { "epoch": 0.27733355689928, "grad_norm": 1.5944184064865112, "learning_rate": 8.479780914860852e-06, "loss": 0.5107, "step": 6616 }, { "epoch": 0.277375475514289, "grad_norm": 1.710080862045288, "learning_rate": 8.479293409647718e-06, "loss": 0.5719, "step": 6617 }, { "epoch": 0.27741739412929795, "grad_norm": 1.8323071002960205, "learning_rate": 8.47880584029862e-06, "loss": 0.56, "step": 6618 }, { "epoch": 0.2774593127443069, "grad_norm": 2.0267045497894287, "learning_rate": 8.478318206822546e-06, "loss": 0.559, "step": 6619 }, { "epoch": 0.27750123135931587, "grad_norm": 1.9230042695999146, "learning_rate": 8.477830509228485e-06, "loss": 0.5721, "step": 6620 }, { "epoch": 0.27754314997432483, "grad_norm": 1.640005111694336, "learning_rate": 8.477342747525427e-06, "loss": 0.5692, "step": 6621 }, { "epoch": 0.2775850685893338, "grad_norm": 1.6733840703964233, "learning_rate": 8.476854921722362e-06, "loss": 0.542, "step": 6622 }, { "epoch": 0.27762698720434276, "grad_norm": 1.792150855064392, "learning_rate": 8.476367031828283e-06, "loss": 0.5357, "step": 6623 }, { "epoch": 0.2776689058193517, "grad_norm": 1.5802839994430542, "learning_rate": 8.475879077852183e-06, "loss": 0.5128, "step": 6624 }, { "epoch": 0.2777108244343607, "grad_norm": 1.5597282648086548, "learning_rate": 8.47539105980306e-06, "loss": 0.5337, "step": 6625 }, { "epoch": 0.27775274304936964, "grad_norm": 1.5053786039352417, "learning_rate": 8.474902977689903e-06, "loss": 0.5071, "step": 6626 }, { "epoch": 0.2777946616643786, "grad_norm": 1.5287188291549683, "learning_rate": 8.474414831521717e-06, "loss": 0.4976, "step": 6627 }, { "epoch": 0.27783658027938757, "grad_norm": 1.638315200805664, "learning_rate": 8.473926621307496e-06, "loss": 0.5315, "step": 6628 }, { "epoch": 0.27787849889439653, "grad_norm": 1.7197562456130981, "learning_rate": 8.473438347056239e-06, "loss": 0.5215, "step": 6629 }, { "epoch": 0.2779204175094055, "grad_norm": 1.5929529666900635, "learning_rate": 8.47295000877695e-06, "loss": 0.5345, "step": 6630 }, { "epoch": 0.27796233612441446, "grad_norm": 1.6550804376602173, "learning_rate": 8.472461606478628e-06, "loss": 0.4605, "step": 6631 }, { "epoch": 0.2780042547394234, "grad_norm": 1.673529028892517, "learning_rate": 8.471973140170277e-06, "loss": 0.5419, "step": 6632 }, { "epoch": 0.2780461733544324, "grad_norm": 1.6507974863052368, "learning_rate": 8.471484609860901e-06, "loss": 0.6021, "step": 6633 }, { "epoch": 0.27808809196944134, "grad_norm": 1.6781868934631348, "learning_rate": 8.470996015559505e-06, "loss": 0.5628, "step": 6634 }, { "epoch": 0.2781300105844503, "grad_norm": 1.511526107788086, "learning_rate": 8.470507357275096e-06, "loss": 0.5505, "step": 6635 }, { "epoch": 0.27817192919945927, "grad_norm": 1.6803849935531616, "learning_rate": 8.470018635016682e-06, "loss": 0.5119, "step": 6636 }, { "epoch": 0.27821384781446823, "grad_norm": 2.223773241043091, "learning_rate": 8.469529848793272e-06, "loss": 0.578, "step": 6637 }, { "epoch": 0.2782557664294772, "grad_norm": 1.782196044921875, "learning_rate": 8.469040998613875e-06, "loss": 0.5628, "step": 6638 }, { "epoch": 0.27829768504448615, "grad_norm": 1.5186519622802734, "learning_rate": 8.468552084487502e-06, "loss": 0.496, "step": 6639 }, { "epoch": 0.2783396036594951, "grad_norm": 1.3855211734771729, "learning_rate": 8.468063106423168e-06, "loss": 0.4859, "step": 6640 }, { "epoch": 0.278381522274504, "grad_norm": 1.7228167057037354, "learning_rate": 8.467574064429885e-06, "loss": 0.5559, "step": 6641 }, { "epoch": 0.278423440889513, "grad_norm": 1.4257506132125854, "learning_rate": 8.467084958516668e-06, "loss": 0.5063, "step": 6642 }, { "epoch": 0.27846535950452195, "grad_norm": 2.1139004230499268, "learning_rate": 8.466595788692532e-06, "loss": 0.5156, "step": 6643 }, { "epoch": 0.2785072781195309, "grad_norm": 2.1697423458099365, "learning_rate": 8.466106554966495e-06, "loss": 0.5317, "step": 6644 }, { "epoch": 0.2785491967345399, "grad_norm": 1.4525905847549438, "learning_rate": 8.465617257347577e-06, "loss": 0.5226, "step": 6645 }, { "epoch": 0.27859111534954883, "grad_norm": 1.457842230796814, "learning_rate": 8.465127895844795e-06, "loss": 0.5187, "step": 6646 }, { "epoch": 0.2786330339645578, "grad_norm": 1.784318208694458, "learning_rate": 8.46463847046717e-06, "loss": 0.6009, "step": 6647 }, { "epoch": 0.27867495257956676, "grad_norm": 1.536739468574524, "learning_rate": 8.464148981223726e-06, "loss": 0.5485, "step": 6648 }, { "epoch": 0.2787168711945757, "grad_norm": 1.6147209405899048, "learning_rate": 8.463659428123484e-06, "loss": 0.5549, "step": 6649 }, { "epoch": 0.2787587898095847, "grad_norm": 1.835174560546875, "learning_rate": 8.463169811175468e-06, "loss": 0.5255, "step": 6650 }, { "epoch": 0.27880070842459365, "grad_norm": 1.916274905204773, "learning_rate": 8.462680130388706e-06, "loss": 0.5544, "step": 6651 }, { "epoch": 0.2788426270396026, "grad_norm": 1.5841671228408813, "learning_rate": 8.46219038577222e-06, "loss": 0.6262, "step": 6652 }, { "epoch": 0.27888454565461157, "grad_norm": 1.5080853700637817, "learning_rate": 8.461700577335042e-06, "loss": 0.4579, "step": 6653 }, { "epoch": 0.27892646426962053, "grad_norm": 2.1882073879241943, "learning_rate": 8.461210705086199e-06, "loss": 0.567, "step": 6654 }, { "epoch": 0.2789683828846295, "grad_norm": 1.6553481817245483, "learning_rate": 8.460720769034723e-06, "loss": 0.542, "step": 6655 }, { "epoch": 0.27901030149963846, "grad_norm": 1.6009221076965332, "learning_rate": 8.460230769189643e-06, "loss": 0.5725, "step": 6656 }, { "epoch": 0.2790522201146474, "grad_norm": 1.4532026052474976, "learning_rate": 8.459740705559992e-06, "loss": 0.4627, "step": 6657 }, { "epoch": 0.2790941387296564, "grad_norm": 1.7558965682983398, "learning_rate": 8.459250578154803e-06, "loss": 0.5527, "step": 6658 }, { "epoch": 0.27913605734466534, "grad_norm": 1.5345864295959473, "learning_rate": 8.458760386983116e-06, "loss": 0.5796, "step": 6659 }, { "epoch": 0.2791779759596743, "grad_norm": 1.6033226251602173, "learning_rate": 8.458270132053959e-06, "loss": 0.5733, "step": 6660 }, { "epoch": 0.27921989457468327, "grad_norm": 1.8213653564453125, "learning_rate": 8.457779813376374e-06, "loss": 0.4708, "step": 6661 }, { "epoch": 0.27926181318969223, "grad_norm": 2.1596031188964844, "learning_rate": 8.457289430959399e-06, "loss": 0.557, "step": 6662 }, { "epoch": 0.2793037318047012, "grad_norm": 1.6484224796295166, "learning_rate": 8.456798984812073e-06, "loss": 0.5306, "step": 6663 }, { "epoch": 0.27934565041971016, "grad_norm": 1.7176514863967896, "learning_rate": 8.456308474943436e-06, "loss": 0.4857, "step": 6664 }, { "epoch": 0.2793875690347191, "grad_norm": 1.6443346738815308, "learning_rate": 8.455817901362531e-06, "loss": 0.5187, "step": 6665 }, { "epoch": 0.279429487649728, "grad_norm": 1.4711284637451172, "learning_rate": 8.455327264078399e-06, "loss": 0.465, "step": 6666 }, { "epoch": 0.279471406264737, "grad_norm": 1.837372064590454, "learning_rate": 8.454836563100088e-06, "loss": 0.5308, "step": 6667 }, { "epoch": 0.27951332487974595, "grad_norm": 1.6121004819869995, "learning_rate": 8.454345798436639e-06, "loss": 0.5132, "step": 6668 }, { "epoch": 0.2795552434947549, "grad_norm": 1.5203816890716553, "learning_rate": 8.453854970097103e-06, "loss": 0.4994, "step": 6669 }, { "epoch": 0.2795971621097639, "grad_norm": 1.484857201576233, "learning_rate": 8.453364078090523e-06, "loss": 0.5261, "step": 6670 }, { "epoch": 0.27963908072477284, "grad_norm": 1.6276814937591553, "learning_rate": 8.452873122425952e-06, "loss": 0.5342, "step": 6671 }, { "epoch": 0.2796809993397818, "grad_norm": 1.706744909286499, "learning_rate": 8.452382103112437e-06, "loss": 0.5789, "step": 6672 }, { "epoch": 0.27972291795479076, "grad_norm": 2.0818819999694824, "learning_rate": 8.451891020159033e-06, "loss": 0.5555, "step": 6673 }, { "epoch": 0.2797648365697997, "grad_norm": 1.5930763483047485, "learning_rate": 8.451399873574788e-06, "loss": 0.5676, "step": 6674 }, { "epoch": 0.2798067551848087, "grad_norm": 1.5785291194915771, "learning_rate": 8.450908663368758e-06, "loss": 0.55, "step": 6675 }, { "epoch": 0.27984867379981765, "grad_norm": 1.5722540616989136, "learning_rate": 8.450417389549998e-06, "loss": 0.5622, "step": 6676 }, { "epoch": 0.2798905924148266, "grad_norm": 1.6167428493499756, "learning_rate": 8.449926052127564e-06, "loss": 0.5101, "step": 6677 }, { "epoch": 0.2799325110298356, "grad_norm": 1.5056736469268799, "learning_rate": 8.449434651110514e-06, "loss": 0.4492, "step": 6678 }, { "epoch": 0.27997442964484454, "grad_norm": 1.469366431236267, "learning_rate": 8.4489431865079e-06, "loss": 0.4972, "step": 6679 }, { "epoch": 0.2800163482598535, "grad_norm": 2.586277723312378, "learning_rate": 8.44845165832879e-06, "loss": 0.5639, "step": 6680 }, { "epoch": 0.28005826687486246, "grad_norm": 1.411716103553772, "learning_rate": 8.447960066582241e-06, "loss": 0.4398, "step": 6681 }, { "epoch": 0.2801001854898714, "grad_norm": 1.5490951538085938, "learning_rate": 8.447468411277316e-06, "loss": 0.5096, "step": 6682 }, { "epoch": 0.2801421041048804, "grad_norm": 1.521298885345459, "learning_rate": 8.446976692423075e-06, "loss": 0.582, "step": 6683 }, { "epoch": 0.28018402271988935, "grad_norm": 1.6368969678878784, "learning_rate": 8.446484910028582e-06, "loss": 0.4907, "step": 6684 }, { "epoch": 0.2802259413348983, "grad_norm": 1.5407872200012207, "learning_rate": 8.445993064102907e-06, "loss": 0.5643, "step": 6685 }, { "epoch": 0.28026785994990727, "grad_norm": 1.6148340702056885, "learning_rate": 8.445501154655114e-06, "loss": 0.5958, "step": 6686 }, { "epoch": 0.28030977856491623, "grad_norm": 1.6194413900375366, "learning_rate": 8.44500918169427e-06, "loss": 0.5058, "step": 6687 }, { "epoch": 0.2803516971799252, "grad_norm": 1.6758662462234497, "learning_rate": 8.444517145229443e-06, "loss": 0.6053, "step": 6688 }, { "epoch": 0.28039361579493416, "grad_norm": 1.8090400695800781, "learning_rate": 8.444025045269706e-06, "loss": 0.5718, "step": 6689 }, { "epoch": 0.2804355344099431, "grad_norm": 1.5932105779647827, "learning_rate": 8.443532881824128e-06, "loss": 0.4657, "step": 6690 }, { "epoch": 0.28047745302495203, "grad_norm": 1.6868257522583008, "learning_rate": 8.443040654901782e-06, "loss": 0.5686, "step": 6691 }, { "epoch": 0.280519371639961, "grad_norm": 1.8373644351959229, "learning_rate": 8.44254836451174e-06, "loss": 0.5615, "step": 6692 }, { "epoch": 0.28056129025496995, "grad_norm": 1.5110362768173218, "learning_rate": 8.44205601066308e-06, "loss": 0.529, "step": 6693 }, { "epoch": 0.2806032088699789, "grad_norm": 1.7851132154464722, "learning_rate": 8.441563593364875e-06, "loss": 0.5126, "step": 6694 }, { "epoch": 0.2806451274849879, "grad_norm": 1.6768262386322021, "learning_rate": 8.441071112626202e-06, "loss": 0.5247, "step": 6695 }, { "epoch": 0.28068704609999684, "grad_norm": 1.5863456726074219, "learning_rate": 8.440578568456142e-06, "loss": 0.5133, "step": 6696 }, { "epoch": 0.2807289647150058, "grad_norm": 1.5566165447235107, "learning_rate": 8.440085960863773e-06, "loss": 0.5696, "step": 6697 }, { "epoch": 0.28077088333001476, "grad_norm": 1.5771450996398926, "learning_rate": 8.439593289858172e-06, "loss": 0.5316, "step": 6698 }, { "epoch": 0.2808128019450237, "grad_norm": 1.5105562210083008, "learning_rate": 8.439100555448427e-06, "loss": 0.5229, "step": 6699 }, { "epoch": 0.2808547205600327, "grad_norm": 1.5455827713012695, "learning_rate": 8.438607757643616e-06, "loss": 0.5522, "step": 6700 }, { "epoch": 0.28089663917504165, "grad_norm": 1.8007571697235107, "learning_rate": 8.438114896452825e-06, "loss": 0.5159, "step": 6701 }, { "epoch": 0.2809385577900506, "grad_norm": 1.6510101556777954, "learning_rate": 8.437621971885139e-06, "loss": 0.5216, "step": 6702 }, { "epoch": 0.2809804764050596, "grad_norm": 1.546981692314148, "learning_rate": 8.437128983949646e-06, "loss": 0.582, "step": 6703 }, { "epoch": 0.28102239502006854, "grad_norm": 1.7033416032791138, "learning_rate": 8.436635932655427e-06, "loss": 0.5089, "step": 6704 }, { "epoch": 0.2810643136350775, "grad_norm": 1.640654444694519, "learning_rate": 8.43614281801158e-06, "loss": 0.5694, "step": 6705 }, { "epoch": 0.28110623225008646, "grad_norm": 1.5395663976669312, "learning_rate": 8.43564964002719e-06, "loss": 0.579, "step": 6706 }, { "epoch": 0.2811481508650954, "grad_norm": 2.22558331489563, "learning_rate": 8.435156398711347e-06, "loss": 0.521, "step": 6707 }, { "epoch": 0.2811900694801044, "grad_norm": 1.5603322982788086, "learning_rate": 8.434663094073146e-06, "loss": 0.5306, "step": 6708 }, { "epoch": 0.28123198809511335, "grad_norm": 1.8148844242095947, "learning_rate": 8.434169726121679e-06, "loss": 0.4962, "step": 6709 }, { "epoch": 0.2812739067101223, "grad_norm": 1.8330851793289185, "learning_rate": 8.43367629486604e-06, "loss": 0.5303, "step": 6710 }, { "epoch": 0.2813158253251313, "grad_norm": 1.6245081424713135, "learning_rate": 8.433182800315326e-06, "loss": 0.5753, "step": 6711 }, { "epoch": 0.28135774394014024, "grad_norm": 1.6387070417404175, "learning_rate": 8.432689242478634e-06, "loss": 0.6058, "step": 6712 }, { "epoch": 0.2813996625551492, "grad_norm": 1.4675673246383667, "learning_rate": 8.43219562136506e-06, "loss": 0.5073, "step": 6713 }, { "epoch": 0.28144158117015816, "grad_norm": 1.5615946054458618, "learning_rate": 8.431701936983705e-06, "loss": 0.5569, "step": 6714 }, { "epoch": 0.2814834997851671, "grad_norm": 1.6111187934875488, "learning_rate": 8.43120818934367e-06, "loss": 0.5426, "step": 6715 }, { "epoch": 0.28152541840017603, "grad_norm": 1.672507643699646, "learning_rate": 8.430714378454052e-06, "loss": 0.5675, "step": 6716 }, { "epoch": 0.281567337015185, "grad_norm": 1.7919073104858398, "learning_rate": 8.43022050432396e-06, "loss": 0.5493, "step": 6717 }, { "epoch": 0.28160925563019396, "grad_norm": 1.4647160768508911, "learning_rate": 8.429726566962495e-06, "loss": 0.5387, "step": 6718 }, { "epoch": 0.2816511742452029, "grad_norm": 2.356992721557617, "learning_rate": 8.429232566378763e-06, "loss": 0.5686, "step": 6719 }, { "epoch": 0.2816930928602119, "grad_norm": 1.553465723991394, "learning_rate": 8.42873850258187e-06, "loss": 0.5389, "step": 6720 }, { "epoch": 0.28173501147522084, "grad_norm": 1.6506351232528687, "learning_rate": 8.428244375580921e-06, "loss": 0.5391, "step": 6721 }, { "epoch": 0.2817769300902298, "grad_norm": 1.5483626127243042, "learning_rate": 8.427750185385026e-06, "loss": 0.5516, "step": 6722 }, { "epoch": 0.28181884870523877, "grad_norm": 1.561071515083313, "learning_rate": 8.427255932003296e-06, "loss": 0.5131, "step": 6723 }, { "epoch": 0.28186076732024773, "grad_norm": 1.5288163423538208, "learning_rate": 8.426761615444841e-06, "loss": 0.5416, "step": 6724 }, { "epoch": 0.2819026859352567, "grad_norm": 1.6134644746780396, "learning_rate": 8.426267235718772e-06, "loss": 0.5634, "step": 6725 }, { "epoch": 0.28194460455026565, "grad_norm": 1.5797618627548218, "learning_rate": 8.425772792834204e-06, "loss": 0.5706, "step": 6726 }, { "epoch": 0.2819865231652746, "grad_norm": 1.606568694114685, "learning_rate": 8.425278286800252e-06, "loss": 0.5248, "step": 6727 }, { "epoch": 0.2820284417802836, "grad_norm": 1.4423021078109741, "learning_rate": 8.424783717626028e-06, "loss": 0.5308, "step": 6728 }, { "epoch": 0.28207036039529254, "grad_norm": 1.6918305158615112, "learning_rate": 8.424289085320651e-06, "loss": 0.4891, "step": 6729 }, { "epoch": 0.2821122790103015, "grad_norm": 1.5512197017669678, "learning_rate": 8.423794389893239e-06, "loss": 0.4695, "step": 6730 }, { "epoch": 0.28215419762531047, "grad_norm": 1.6546550989151, "learning_rate": 8.42329963135291e-06, "loss": 0.5422, "step": 6731 }, { "epoch": 0.28219611624031943, "grad_norm": 1.5801653861999512, "learning_rate": 8.422804809708784e-06, "loss": 0.5298, "step": 6732 }, { "epoch": 0.2822380348553284, "grad_norm": 1.4673547744750977, "learning_rate": 8.422309924969986e-06, "loss": 0.5489, "step": 6733 }, { "epoch": 0.28227995347033735, "grad_norm": 1.781057596206665, "learning_rate": 8.421814977145635e-06, "loss": 0.5625, "step": 6734 }, { "epoch": 0.2823218720853463, "grad_norm": 1.6121896505355835, "learning_rate": 8.421319966244854e-06, "loss": 0.5231, "step": 6735 }, { "epoch": 0.2823637907003553, "grad_norm": 1.796220302581787, "learning_rate": 8.420824892276771e-06, "loss": 0.5725, "step": 6736 }, { "epoch": 0.28240570931536424, "grad_norm": 1.5591754913330078, "learning_rate": 8.42032975525051e-06, "loss": 0.4925, "step": 6737 }, { "epoch": 0.2824476279303732, "grad_norm": 1.5963150262832642, "learning_rate": 8.419834555175197e-06, "loss": 0.533, "step": 6738 }, { "epoch": 0.28248954654538216, "grad_norm": 1.5915271043777466, "learning_rate": 8.419339292059963e-06, "loss": 0.563, "step": 6739 }, { "epoch": 0.2825314651603911, "grad_norm": 1.8024308681488037, "learning_rate": 8.418843965913936e-06, "loss": 0.5287, "step": 6740 }, { "epoch": 0.28257338377540003, "grad_norm": 1.534766674041748, "learning_rate": 8.418348576746245e-06, "loss": 0.584, "step": 6741 }, { "epoch": 0.282615302390409, "grad_norm": 1.6241012811660767, "learning_rate": 8.417853124566026e-06, "loss": 0.5375, "step": 6742 }, { "epoch": 0.28265722100541796, "grad_norm": 1.774888515472412, "learning_rate": 8.41735760938241e-06, "loss": 0.5713, "step": 6743 }, { "epoch": 0.2826991396204269, "grad_norm": 1.4211747646331787, "learning_rate": 8.41686203120453e-06, "loss": 0.4636, "step": 6744 }, { "epoch": 0.2827410582354359, "grad_norm": 1.5220303535461426, "learning_rate": 8.416366390041521e-06, "loss": 0.5215, "step": 6745 }, { "epoch": 0.28278297685044484, "grad_norm": 1.724346399307251, "learning_rate": 8.415870685902522e-06, "loss": 0.5563, "step": 6746 }, { "epoch": 0.2828248954654538, "grad_norm": 1.7896478176116943, "learning_rate": 8.415374918796668e-06, "loss": 0.5201, "step": 6747 }, { "epoch": 0.28286681408046277, "grad_norm": 1.5963244438171387, "learning_rate": 8.4148790887331e-06, "loss": 0.5079, "step": 6748 }, { "epoch": 0.28290873269547173, "grad_norm": 1.6344043016433716, "learning_rate": 8.414383195720956e-06, "loss": 0.4838, "step": 6749 }, { "epoch": 0.2829506513104807, "grad_norm": 1.720860481262207, "learning_rate": 8.413887239769381e-06, "loss": 0.4915, "step": 6750 }, { "epoch": 0.28299256992548966, "grad_norm": 1.630658745765686, "learning_rate": 8.41339122088751e-06, "loss": 0.5368, "step": 6751 }, { "epoch": 0.2830344885404986, "grad_norm": 1.6065878868103027, "learning_rate": 8.412895139084494e-06, "loss": 0.5422, "step": 6752 }, { "epoch": 0.2830764071555076, "grad_norm": 1.587157964706421, "learning_rate": 8.412398994369472e-06, "loss": 0.5575, "step": 6753 }, { "epoch": 0.28311832577051654, "grad_norm": 1.5476773977279663, "learning_rate": 8.411902786751593e-06, "loss": 0.4924, "step": 6754 }, { "epoch": 0.2831602443855255, "grad_norm": 1.5547270774841309, "learning_rate": 8.411406516240003e-06, "loss": 0.5095, "step": 6755 }, { "epoch": 0.28320216300053447, "grad_norm": 1.7625505924224854, "learning_rate": 8.41091018284385e-06, "loss": 0.5172, "step": 6756 }, { "epoch": 0.28324408161554343, "grad_norm": 1.962457537651062, "learning_rate": 8.410413786572285e-06, "loss": 0.4608, "step": 6757 }, { "epoch": 0.2832860002305524, "grad_norm": 1.5949373245239258, "learning_rate": 8.409917327434454e-06, "loss": 0.5618, "step": 6758 }, { "epoch": 0.28332791884556136, "grad_norm": 1.5597223043441772, "learning_rate": 8.409420805439512e-06, "loss": 0.5287, "step": 6759 }, { "epoch": 0.2833698374605703, "grad_norm": 1.4785974025726318, "learning_rate": 8.40892422059661e-06, "loss": 0.5106, "step": 6760 }, { "epoch": 0.2834117560755793, "grad_norm": 1.5313410758972168, "learning_rate": 8.408427572914904e-06, "loss": 0.5007, "step": 6761 }, { "epoch": 0.28345367469058824, "grad_norm": 2.03643536567688, "learning_rate": 8.407930862403548e-06, "loss": 0.564, "step": 6762 }, { "epoch": 0.2834955933055972, "grad_norm": 1.6257082223892212, "learning_rate": 8.407434089071698e-06, "loss": 0.53, "step": 6763 }, { "epoch": 0.28353751192060617, "grad_norm": 1.586401343345642, "learning_rate": 8.406937252928509e-06, "loss": 0.5033, "step": 6764 }, { "epoch": 0.28357943053561513, "grad_norm": 1.823806881904602, "learning_rate": 8.406440353983143e-06, "loss": 0.5184, "step": 6765 }, { "epoch": 0.28362134915062404, "grad_norm": 1.7617013454437256, "learning_rate": 8.405943392244759e-06, "loss": 0.5015, "step": 6766 }, { "epoch": 0.283663267765633, "grad_norm": 1.6161149740219116, "learning_rate": 8.405446367722517e-06, "loss": 0.5036, "step": 6767 }, { "epoch": 0.28370518638064196, "grad_norm": 1.616745114326477, "learning_rate": 8.404949280425579e-06, "loss": 0.5655, "step": 6768 }, { "epoch": 0.2837471049956509, "grad_norm": 1.6655657291412354, "learning_rate": 8.404452130363106e-06, "loss": 0.5954, "step": 6769 }, { "epoch": 0.2837890236106599, "grad_norm": 1.7436964511871338, "learning_rate": 8.403954917544266e-06, "loss": 0.5249, "step": 6770 }, { "epoch": 0.28383094222566885, "grad_norm": 1.7409825325012207, "learning_rate": 8.403457641978225e-06, "loss": 0.6068, "step": 6771 }, { "epoch": 0.2838728608406778, "grad_norm": 1.7540735006332397, "learning_rate": 8.402960303674145e-06, "loss": 0.5818, "step": 6772 }, { "epoch": 0.28391477945568677, "grad_norm": 1.6114745140075684, "learning_rate": 8.402462902641198e-06, "loss": 0.4905, "step": 6773 }, { "epoch": 0.28395669807069573, "grad_norm": 1.5523604154586792, "learning_rate": 8.401965438888549e-06, "loss": 0.5244, "step": 6774 }, { "epoch": 0.2839986166857047, "grad_norm": 1.6376736164093018, "learning_rate": 8.401467912425372e-06, "loss": 0.5103, "step": 6775 }, { "epoch": 0.28404053530071366, "grad_norm": 1.5643705129623413, "learning_rate": 8.400970323260835e-06, "loss": 0.5581, "step": 6776 }, { "epoch": 0.2840824539157226, "grad_norm": 1.7642019987106323, "learning_rate": 8.400472671404113e-06, "loss": 0.5413, "step": 6777 }, { "epoch": 0.2841243725307316, "grad_norm": 1.7835185527801514, "learning_rate": 8.399974956864377e-06, "loss": 0.5178, "step": 6778 }, { "epoch": 0.28416629114574055, "grad_norm": 1.5408403873443604, "learning_rate": 8.399477179650804e-06, "loss": 0.5678, "step": 6779 }, { "epoch": 0.2842082097607495, "grad_norm": 1.6266428232192993, "learning_rate": 8.398979339772569e-06, "loss": 0.5365, "step": 6780 }, { "epoch": 0.28425012837575847, "grad_norm": 1.762548804283142, "learning_rate": 8.398481437238849e-06, "loss": 0.5344, "step": 6781 }, { "epoch": 0.28429204699076743, "grad_norm": 1.6179263591766357, "learning_rate": 8.397983472058822e-06, "loss": 0.547, "step": 6782 }, { "epoch": 0.2843339656057764, "grad_norm": 1.8265081644058228, "learning_rate": 8.397485444241667e-06, "loss": 0.5429, "step": 6783 }, { "epoch": 0.28437588422078536, "grad_norm": 1.5159136056900024, "learning_rate": 8.396987353796564e-06, "loss": 0.4823, "step": 6784 }, { "epoch": 0.2844178028357943, "grad_norm": 2.515448570251465, "learning_rate": 8.396489200732695e-06, "loss": 0.4858, "step": 6785 }, { "epoch": 0.2844597214508033, "grad_norm": 1.6706818342208862, "learning_rate": 8.395990985059244e-06, "loss": 0.5204, "step": 6786 }, { "epoch": 0.28450164006581224, "grad_norm": 1.4722096920013428, "learning_rate": 8.395492706785393e-06, "loss": 0.5199, "step": 6787 }, { "epoch": 0.2845435586808212, "grad_norm": 1.5030995607376099, "learning_rate": 8.39499436592033e-06, "loss": 0.4447, "step": 6788 }, { "epoch": 0.28458547729583017, "grad_norm": 4.461399555206299, "learning_rate": 8.394495962473238e-06, "loss": 0.5948, "step": 6789 }, { "epoch": 0.28462739591083913, "grad_norm": 1.7644989490509033, "learning_rate": 8.393997496453305e-06, "loss": 0.5254, "step": 6790 }, { "epoch": 0.28466931452584804, "grad_norm": 1.9478704929351807, "learning_rate": 8.393498967869722e-06, "loss": 0.5659, "step": 6791 }, { "epoch": 0.284711233140857, "grad_norm": 1.5857783555984497, "learning_rate": 8.393000376731677e-06, "loss": 0.5114, "step": 6792 }, { "epoch": 0.28475315175586596, "grad_norm": 1.66830575466156, "learning_rate": 8.392501723048358e-06, "loss": 0.5503, "step": 6793 }, { "epoch": 0.2847950703708749, "grad_norm": 1.6816195249557495, "learning_rate": 8.392003006828963e-06, "loss": 0.5491, "step": 6794 }, { "epoch": 0.2848369889858839, "grad_norm": 1.7626341581344604, "learning_rate": 8.391504228082681e-06, "loss": 0.5642, "step": 6795 }, { "epoch": 0.28487890760089285, "grad_norm": 1.6528615951538086, "learning_rate": 8.391005386818705e-06, "loss": 0.5581, "step": 6796 }, { "epoch": 0.2849208262159018, "grad_norm": 1.9385786056518555, "learning_rate": 8.390506483046234e-06, "loss": 0.5955, "step": 6797 }, { "epoch": 0.2849627448309108, "grad_norm": 1.3690029382705688, "learning_rate": 8.390007516774464e-06, "loss": 0.5185, "step": 6798 }, { "epoch": 0.28500466344591974, "grad_norm": 1.787880301475525, "learning_rate": 8.389508488012591e-06, "loss": 0.6373, "step": 6799 }, { "epoch": 0.2850465820609287, "grad_norm": 1.6914501190185547, "learning_rate": 8.389009396769816e-06, "loss": 0.5999, "step": 6800 }, { "epoch": 0.28508850067593766, "grad_norm": 1.5314080715179443, "learning_rate": 8.388510243055337e-06, "loss": 0.5028, "step": 6801 }, { "epoch": 0.2851304192909466, "grad_norm": 1.5108953714370728, "learning_rate": 8.388011026878358e-06, "loss": 0.5231, "step": 6802 }, { "epoch": 0.2851723379059556, "grad_norm": 1.7005329132080078, "learning_rate": 8.387511748248076e-06, "loss": 0.5015, "step": 6803 }, { "epoch": 0.28521425652096455, "grad_norm": 1.6199309825897217, "learning_rate": 8.3870124071737e-06, "loss": 0.5239, "step": 6804 }, { "epoch": 0.2852561751359735, "grad_norm": 1.7994754314422607, "learning_rate": 8.386513003664435e-06, "loss": 0.5749, "step": 6805 }, { "epoch": 0.2852980937509825, "grad_norm": 1.551240086555481, "learning_rate": 8.386013537729481e-06, "loss": 0.5155, "step": 6806 }, { "epoch": 0.28534001236599144, "grad_norm": 2.8857667446136475, "learning_rate": 8.385514009378048e-06, "loss": 0.5149, "step": 6807 }, { "epoch": 0.2853819309810004, "grad_norm": 1.8240060806274414, "learning_rate": 8.385014418619346e-06, "loss": 0.5074, "step": 6808 }, { "epoch": 0.28542384959600936, "grad_norm": 1.692211627960205, "learning_rate": 8.384514765462584e-06, "loss": 0.5447, "step": 6809 }, { "epoch": 0.2854657682110183, "grad_norm": 1.7300230264663696, "learning_rate": 8.384015049916969e-06, "loss": 0.5144, "step": 6810 }, { "epoch": 0.2855076868260273, "grad_norm": 1.6119139194488525, "learning_rate": 8.383515271991715e-06, "loss": 0.5668, "step": 6811 }, { "epoch": 0.28554960544103625, "grad_norm": 1.8212473392486572, "learning_rate": 8.383015431696037e-06, "loss": 0.5142, "step": 6812 }, { "epoch": 0.2855915240560452, "grad_norm": 1.5788366794586182, "learning_rate": 8.382515529039144e-06, "loss": 0.5424, "step": 6813 }, { "epoch": 0.28563344267105417, "grad_norm": 1.9605352878570557, "learning_rate": 8.382015564030254e-06, "loss": 0.5212, "step": 6814 }, { "epoch": 0.28567536128606313, "grad_norm": 1.677977204322815, "learning_rate": 8.381515536678583e-06, "loss": 0.5326, "step": 6815 }, { "epoch": 0.28571727990107204, "grad_norm": 1.5703229904174805, "learning_rate": 8.381015446993348e-06, "loss": 0.5457, "step": 6816 }, { "epoch": 0.285759198516081, "grad_norm": 1.7072248458862305, "learning_rate": 8.380515294983769e-06, "loss": 0.5469, "step": 6817 }, { "epoch": 0.28580111713108997, "grad_norm": 1.5940965414047241, "learning_rate": 8.38001508065906e-06, "loss": 0.5971, "step": 6818 }, { "epoch": 0.28584303574609893, "grad_norm": 1.5146913528442383, "learning_rate": 8.379514804028448e-06, "loss": 0.5305, "step": 6819 }, { "epoch": 0.2858849543611079, "grad_norm": 1.579283356666565, "learning_rate": 8.379014465101154e-06, "loss": 0.5187, "step": 6820 }, { "epoch": 0.28592687297611685, "grad_norm": 1.5986928939819336, "learning_rate": 8.378514063886397e-06, "loss": 0.4804, "step": 6821 }, { "epoch": 0.2859687915911258, "grad_norm": 1.8123314380645752, "learning_rate": 8.378013600393405e-06, "loss": 0.562, "step": 6822 }, { "epoch": 0.2860107102061348, "grad_norm": 1.665770411491394, "learning_rate": 8.377513074631403e-06, "loss": 0.5616, "step": 6823 }, { "epoch": 0.28605262882114374, "grad_norm": 1.7124536037445068, "learning_rate": 8.377012486609614e-06, "loss": 0.5243, "step": 6824 }, { "epoch": 0.2860945474361527, "grad_norm": 1.7084592580795288, "learning_rate": 8.376511836337271e-06, "loss": 0.5248, "step": 6825 }, { "epoch": 0.28613646605116166, "grad_norm": 1.9220077991485596, "learning_rate": 8.3760111238236e-06, "loss": 0.5597, "step": 6826 }, { "epoch": 0.2861783846661706, "grad_norm": 1.461479902267456, "learning_rate": 8.375510349077831e-06, "loss": 0.4891, "step": 6827 }, { "epoch": 0.2862203032811796, "grad_norm": 2.260643720626831, "learning_rate": 8.375009512109194e-06, "loss": 0.4654, "step": 6828 }, { "epoch": 0.28626222189618855, "grad_norm": 1.6666064262390137, "learning_rate": 8.374508612926923e-06, "loss": 0.4872, "step": 6829 }, { "epoch": 0.2863041405111975, "grad_norm": 1.4397039413452148, "learning_rate": 8.374007651540253e-06, "loss": 0.5238, "step": 6830 }, { "epoch": 0.2863460591262065, "grad_norm": 1.8651481866836548, "learning_rate": 8.373506627958414e-06, "loss": 0.5446, "step": 6831 }, { "epoch": 0.28638797774121544, "grad_norm": 1.4717386960983276, "learning_rate": 8.373005542190644e-06, "loss": 0.4771, "step": 6832 }, { "epoch": 0.2864298963562244, "grad_norm": 1.7668131589889526, "learning_rate": 8.37250439424618e-06, "loss": 0.537, "step": 6833 }, { "epoch": 0.28647181497123336, "grad_norm": 1.7898153066635132, "learning_rate": 8.372003184134262e-06, "loss": 0.5887, "step": 6834 }, { "epoch": 0.2865137335862423, "grad_norm": 2.8665690422058105, "learning_rate": 8.371501911864124e-06, "loss": 0.5404, "step": 6835 }, { "epoch": 0.2865556522012513, "grad_norm": 1.6461055278778076, "learning_rate": 8.37100057744501e-06, "loss": 0.5388, "step": 6836 }, { "epoch": 0.28659757081626025, "grad_norm": 1.5293571949005127, "learning_rate": 8.370499180886162e-06, "loss": 0.5091, "step": 6837 }, { "epoch": 0.2866394894312692, "grad_norm": 1.5031574964523315, "learning_rate": 8.369997722196821e-06, "loss": 0.5172, "step": 6838 }, { "epoch": 0.2866814080462782, "grad_norm": 1.7657865285873413, "learning_rate": 8.36949620138623e-06, "loss": 0.5362, "step": 6839 }, { "epoch": 0.28672332666128714, "grad_norm": 1.7761482000350952, "learning_rate": 8.368994618463633e-06, "loss": 0.5551, "step": 6840 }, { "epoch": 0.28676524527629604, "grad_norm": 1.6420366764068604, "learning_rate": 8.36849297343828e-06, "loss": 0.5199, "step": 6841 }, { "epoch": 0.286807163891305, "grad_norm": 1.9522722959518433, "learning_rate": 8.367991266319416e-06, "loss": 0.5408, "step": 6842 }, { "epoch": 0.28684908250631397, "grad_norm": 1.6802729368209839, "learning_rate": 8.367489497116289e-06, "loss": 0.5, "step": 6843 }, { "epoch": 0.28689100112132293, "grad_norm": 1.5461846590042114, "learning_rate": 8.366987665838147e-06, "loss": 0.5357, "step": 6844 }, { "epoch": 0.2869329197363319, "grad_norm": 1.6813589334487915, "learning_rate": 8.366485772494244e-06, "loss": 0.4919, "step": 6845 }, { "epoch": 0.28697483835134086, "grad_norm": 1.5953726768493652, "learning_rate": 8.365983817093828e-06, "loss": 0.4896, "step": 6846 }, { "epoch": 0.2870167569663498, "grad_norm": 1.6726405620574951, "learning_rate": 8.365481799646156e-06, "loss": 0.5646, "step": 6847 }, { "epoch": 0.2870586755813588, "grad_norm": 1.6833544969558716, "learning_rate": 8.364979720160478e-06, "loss": 0.5335, "step": 6848 }, { "epoch": 0.28710059419636774, "grad_norm": 1.7523850202560425, "learning_rate": 8.364477578646051e-06, "loss": 0.5928, "step": 6849 }, { "epoch": 0.2871425128113767, "grad_norm": 1.6152617931365967, "learning_rate": 8.363975375112132e-06, "loss": 0.5302, "step": 6850 }, { "epoch": 0.28718443142638567, "grad_norm": 1.622237205505371, "learning_rate": 8.363473109567977e-06, "loss": 0.5631, "step": 6851 }, { "epoch": 0.28722635004139463, "grad_norm": 1.6279842853546143, "learning_rate": 8.362970782022844e-06, "loss": 0.5199, "step": 6852 }, { "epoch": 0.2872682686564036, "grad_norm": 1.713444471359253, "learning_rate": 8.362468392485993e-06, "loss": 0.5459, "step": 6853 }, { "epoch": 0.28731018727141255, "grad_norm": 1.7491495609283447, "learning_rate": 8.361965940966688e-06, "loss": 0.5415, "step": 6854 }, { "epoch": 0.2873521058864215, "grad_norm": 1.8892332315444946, "learning_rate": 8.361463427474188e-06, "loss": 0.494, "step": 6855 }, { "epoch": 0.2873940245014305, "grad_norm": 2.049220085144043, "learning_rate": 8.360960852017757e-06, "loss": 0.5148, "step": 6856 }, { "epoch": 0.28743594311643944, "grad_norm": 1.5938044786453247, "learning_rate": 8.360458214606659e-06, "loss": 0.4806, "step": 6857 }, { "epoch": 0.2874778617314484, "grad_norm": 1.953123688697815, "learning_rate": 8.359955515250159e-06, "loss": 0.5675, "step": 6858 }, { "epoch": 0.28751978034645737, "grad_norm": 1.4436687231063843, "learning_rate": 8.359452753957524e-06, "loss": 0.4899, "step": 6859 }, { "epoch": 0.2875616989614663, "grad_norm": 1.7669017314910889, "learning_rate": 8.358949930738023e-06, "loss": 0.5446, "step": 6860 }, { "epoch": 0.2876036175764753, "grad_norm": 1.8213059902191162, "learning_rate": 8.358447045600921e-06, "loss": 0.5248, "step": 6861 }, { "epoch": 0.28764553619148425, "grad_norm": 1.5762916803359985, "learning_rate": 8.357944098555492e-06, "loss": 0.4852, "step": 6862 }, { "epoch": 0.2876874548064932, "grad_norm": 1.6106798648834229, "learning_rate": 8.357441089611008e-06, "loss": 0.5196, "step": 6863 }, { "epoch": 0.2877293734215022, "grad_norm": 1.626772165298462, "learning_rate": 8.356938018776737e-06, "loss": 0.5419, "step": 6864 }, { "epoch": 0.28777129203651114, "grad_norm": 1.5464024543762207, "learning_rate": 8.356434886061955e-06, "loss": 0.4958, "step": 6865 }, { "epoch": 0.28781321065152005, "grad_norm": 1.6402976512908936, "learning_rate": 8.355931691475937e-06, "loss": 0.4792, "step": 6866 }, { "epoch": 0.287855129266529, "grad_norm": 1.5598427057266235, "learning_rate": 8.355428435027954e-06, "loss": 0.5781, "step": 6867 }, { "epoch": 0.28789704788153797, "grad_norm": 4.06840181350708, "learning_rate": 8.35492511672729e-06, "loss": 0.506, "step": 6868 }, { "epoch": 0.28793896649654693, "grad_norm": 1.7799296379089355, "learning_rate": 8.354421736583219e-06, "loss": 0.5492, "step": 6869 }, { "epoch": 0.2879808851115559, "grad_norm": 1.6187031269073486, "learning_rate": 8.353918294605019e-06, "loss": 0.5282, "step": 6870 }, { "epoch": 0.28802280372656486, "grad_norm": 2.1287734508514404, "learning_rate": 8.353414790801972e-06, "loss": 0.4853, "step": 6871 }, { "epoch": 0.2880647223415738, "grad_norm": 1.7474522590637207, "learning_rate": 8.352911225183363e-06, "loss": 0.5233, "step": 6872 }, { "epoch": 0.2881066409565828, "grad_norm": 1.4883140325546265, "learning_rate": 8.352407597758465e-06, "loss": 0.4989, "step": 6873 }, { "epoch": 0.28814855957159174, "grad_norm": 1.5158323049545288, "learning_rate": 8.351903908536572e-06, "loss": 0.5426, "step": 6874 }, { "epoch": 0.2881904781866007, "grad_norm": 1.6148487329483032, "learning_rate": 8.351400157526964e-06, "loss": 0.5357, "step": 6875 }, { "epoch": 0.28823239680160967, "grad_norm": 2.058236837387085, "learning_rate": 8.350896344738926e-06, "loss": 0.5317, "step": 6876 }, { "epoch": 0.28827431541661863, "grad_norm": 1.857457160949707, "learning_rate": 8.350392470181745e-06, "loss": 0.5043, "step": 6877 }, { "epoch": 0.2883162340316276, "grad_norm": 1.8355716466903687, "learning_rate": 8.349888533864712e-06, "loss": 0.5353, "step": 6878 }, { "epoch": 0.28835815264663656, "grad_norm": 1.8345264196395874, "learning_rate": 8.349384535797116e-06, "loss": 0.5541, "step": 6879 }, { "epoch": 0.2884000712616455, "grad_norm": 1.7167671918869019, "learning_rate": 8.348880475988245e-06, "loss": 0.5788, "step": 6880 }, { "epoch": 0.2884419898766545, "grad_norm": 1.6310679912567139, "learning_rate": 8.348376354447393e-06, "loss": 0.4903, "step": 6881 }, { "epoch": 0.28848390849166344, "grad_norm": 1.7910581827163696, "learning_rate": 8.347872171183851e-06, "loss": 0.4938, "step": 6882 }, { "epoch": 0.2885258271066724, "grad_norm": 1.8721741437911987, "learning_rate": 8.347367926206915e-06, "loss": 0.5279, "step": 6883 }, { "epoch": 0.28856774572168137, "grad_norm": 1.851354718208313, "learning_rate": 8.346863619525878e-06, "loss": 0.5071, "step": 6884 }, { "epoch": 0.28860966433669033, "grad_norm": 2.1876847743988037, "learning_rate": 8.346359251150037e-06, "loss": 0.5315, "step": 6885 }, { "epoch": 0.2886515829516993, "grad_norm": 1.5918872356414795, "learning_rate": 8.34585482108869e-06, "loss": 0.4935, "step": 6886 }, { "epoch": 0.28869350156670825, "grad_norm": 1.6971485614776611, "learning_rate": 8.345350329351134e-06, "loss": 0.5628, "step": 6887 }, { "epoch": 0.2887354201817172, "grad_norm": 2.3190035820007324, "learning_rate": 8.344845775946671e-06, "loss": 0.5496, "step": 6888 }, { "epoch": 0.2887773387967262, "grad_norm": 1.7114713191986084, "learning_rate": 8.344341160884599e-06, "loss": 0.5866, "step": 6889 }, { "epoch": 0.28881925741173514, "grad_norm": 1.7621949911117554, "learning_rate": 8.343836484174222e-06, "loss": 0.5711, "step": 6890 }, { "epoch": 0.28886117602674405, "grad_norm": 1.5367164611816406, "learning_rate": 8.343331745824841e-06, "loss": 0.5206, "step": 6891 }, { "epoch": 0.288903094641753, "grad_norm": 1.6437636613845825, "learning_rate": 8.342826945845761e-06, "loss": 0.4667, "step": 6892 }, { "epoch": 0.288945013256762, "grad_norm": 1.5046513080596924, "learning_rate": 8.34232208424629e-06, "loss": 0.4816, "step": 6893 }, { "epoch": 0.28898693187177094, "grad_norm": 1.5600296258926392, "learning_rate": 8.34181716103573e-06, "loss": 0.4772, "step": 6894 }, { "epoch": 0.2890288504867799, "grad_norm": 2.3554928302764893, "learning_rate": 8.34131217622339e-06, "loss": 0.4972, "step": 6895 }, { "epoch": 0.28907076910178886, "grad_norm": 1.644178867340088, "learning_rate": 8.34080712981858e-06, "loss": 0.4513, "step": 6896 }, { "epoch": 0.2891126877167978, "grad_norm": 1.5438151359558105, "learning_rate": 8.340302021830612e-06, "loss": 0.5332, "step": 6897 }, { "epoch": 0.2891546063318068, "grad_norm": 1.5191576480865479, "learning_rate": 8.339796852268788e-06, "loss": 0.519, "step": 6898 }, { "epoch": 0.28919652494681575, "grad_norm": 1.5433855056762695, "learning_rate": 8.339291621142432e-06, "loss": 0.5782, "step": 6899 }, { "epoch": 0.2892384435618247, "grad_norm": 1.7084885835647583, "learning_rate": 8.338786328460848e-06, "loss": 0.5395, "step": 6900 }, { "epoch": 0.28928036217683367, "grad_norm": 1.5257625579833984, "learning_rate": 8.338280974233355e-06, "loss": 0.5028, "step": 6901 }, { "epoch": 0.28932228079184263, "grad_norm": 1.874598503112793, "learning_rate": 8.337775558469266e-06, "loss": 0.4992, "step": 6902 }, { "epoch": 0.2893641994068516, "grad_norm": 1.8576220273971558, "learning_rate": 8.3372700811779e-06, "loss": 0.522, "step": 6903 }, { "epoch": 0.28940611802186056, "grad_norm": 1.8454456329345703, "learning_rate": 8.336764542368574e-06, "loss": 0.5132, "step": 6904 }, { "epoch": 0.2894480366368695, "grad_norm": 1.4864226579666138, "learning_rate": 8.336258942050606e-06, "loss": 0.5622, "step": 6905 }, { "epoch": 0.2894899552518785, "grad_norm": 1.4779444932937622, "learning_rate": 8.335753280233316e-06, "loss": 0.5312, "step": 6906 }, { "epoch": 0.28953187386688745, "grad_norm": 1.5542199611663818, "learning_rate": 8.335247556926028e-06, "loss": 0.4936, "step": 6907 }, { "epoch": 0.2895737924818964, "grad_norm": 1.8307631015777588, "learning_rate": 8.334741772138059e-06, "loss": 0.5493, "step": 6908 }, { "epoch": 0.28961571109690537, "grad_norm": 1.7048345804214478, "learning_rate": 8.33423592587874e-06, "loss": 0.5402, "step": 6909 }, { "epoch": 0.28965762971191433, "grad_norm": 1.6570825576782227, "learning_rate": 8.333730018157386e-06, "loss": 0.5483, "step": 6910 }, { "epoch": 0.2896995483269233, "grad_norm": 2.148314952850342, "learning_rate": 8.33322404898333e-06, "loss": 0.5555, "step": 6911 }, { "epoch": 0.28974146694193226, "grad_norm": 1.5703036785125732, "learning_rate": 8.3327180183659e-06, "loss": 0.6248, "step": 6912 }, { "epoch": 0.2897833855569412, "grad_norm": 1.3990076780319214, "learning_rate": 8.332211926314416e-06, "loss": 0.4957, "step": 6913 }, { "epoch": 0.2898253041719502, "grad_norm": 1.6861348152160645, "learning_rate": 8.331705772838215e-06, "loss": 0.5569, "step": 6914 }, { "epoch": 0.28986722278695914, "grad_norm": 1.5600416660308838, "learning_rate": 8.331199557946621e-06, "loss": 0.5226, "step": 6915 }, { "epoch": 0.28990914140196805, "grad_norm": 1.987480878829956, "learning_rate": 8.33069328164897e-06, "loss": 0.531, "step": 6916 }, { "epoch": 0.289951060016977, "grad_norm": 1.8379359245300293, "learning_rate": 8.330186943954593e-06, "loss": 0.5525, "step": 6917 }, { "epoch": 0.289992978631986, "grad_norm": 1.7927546501159668, "learning_rate": 8.329680544872823e-06, "loss": 0.5786, "step": 6918 }, { "epoch": 0.29003489724699494, "grad_norm": 1.6512037515640259, "learning_rate": 8.329174084412997e-06, "loss": 0.4927, "step": 6919 }, { "epoch": 0.2900768158620039, "grad_norm": 1.624997615814209, "learning_rate": 8.328667562584447e-06, "loss": 0.5684, "step": 6920 }, { "epoch": 0.29011873447701286, "grad_norm": 1.5653271675109863, "learning_rate": 8.328160979396514e-06, "loss": 0.4904, "step": 6921 }, { "epoch": 0.2901606530920218, "grad_norm": 1.418710470199585, "learning_rate": 8.327654334858532e-06, "loss": 0.4891, "step": 6922 }, { "epoch": 0.2902025717070308, "grad_norm": 1.6228889226913452, "learning_rate": 8.327147628979845e-06, "loss": 0.5049, "step": 6923 }, { "epoch": 0.29024449032203975, "grad_norm": 1.7529503107070923, "learning_rate": 8.326640861769791e-06, "loss": 0.5067, "step": 6924 }, { "epoch": 0.2902864089370487, "grad_norm": 1.9572557210922241, "learning_rate": 8.32613403323771e-06, "loss": 0.5658, "step": 6925 }, { "epoch": 0.2903283275520577, "grad_norm": 1.7013453245162964, "learning_rate": 8.325627143392946e-06, "loss": 0.5026, "step": 6926 }, { "epoch": 0.29037024616706664, "grad_norm": 1.9467042684555054, "learning_rate": 8.325120192244845e-06, "loss": 0.5699, "step": 6927 }, { "epoch": 0.2904121647820756, "grad_norm": 1.4755661487579346, "learning_rate": 8.32461317980275e-06, "loss": 0.514, "step": 6928 }, { "epoch": 0.29045408339708456, "grad_norm": 1.7641386985778809, "learning_rate": 8.324106106076006e-06, "loss": 0.5771, "step": 6929 }, { "epoch": 0.2904960020120935, "grad_norm": 1.5001357793807983, "learning_rate": 8.323598971073962e-06, "loss": 0.5211, "step": 6930 }, { "epoch": 0.2905379206271025, "grad_norm": 1.6879833936691284, "learning_rate": 8.323091774805967e-06, "loss": 0.5472, "step": 6931 }, { "epoch": 0.29057983924211145, "grad_norm": 1.5895726680755615, "learning_rate": 8.322584517281368e-06, "loss": 0.5007, "step": 6932 }, { "epoch": 0.2906217578571204, "grad_norm": 1.5867122411727905, "learning_rate": 8.322077198509517e-06, "loss": 0.529, "step": 6933 }, { "epoch": 0.2906636764721294, "grad_norm": 1.704962134361267, "learning_rate": 8.321569818499763e-06, "loss": 0.6113, "step": 6934 }, { "epoch": 0.29070559508713834, "grad_norm": 1.674360752105713, "learning_rate": 8.321062377261465e-06, "loss": 0.559, "step": 6935 }, { "epoch": 0.2907475137021473, "grad_norm": 1.5299601554870605, "learning_rate": 8.320554874803971e-06, "loss": 0.4609, "step": 6936 }, { "epoch": 0.29078943231715626, "grad_norm": 1.6207199096679688, "learning_rate": 8.32004731113664e-06, "loss": 0.5218, "step": 6937 }, { "epoch": 0.2908313509321652, "grad_norm": 1.5530803203582764, "learning_rate": 8.319539686268825e-06, "loss": 0.494, "step": 6938 }, { "epoch": 0.2908732695471742, "grad_norm": 1.610993504524231, "learning_rate": 8.319032000209885e-06, "loss": 0.5602, "step": 6939 }, { "epoch": 0.29091518816218315, "grad_norm": 1.50252103805542, "learning_rate": 8.31852425296918e-06, "loss": 0.44, "step": 6940 }, { "epoch": 0.29095710677719205, "grad_norm": 1.6845630407333374, "learning_rate": 8.318016444556067e-06, "loss": 0.5663, "step": 6941 }, { "epoch": 0.290999025392201, "grad_norm": 1.5476161241531372, "learning_rate": 8.317508574979908e-06, "loss": 0.5641, "step": 6942 }, { "epoch": 0.29104094400721, "grad_norm": 1.6102759838104248, "learning_rate": 8.317000644250067e-06, "loss": 0.5654, "step": 6943 }, { "epoch": 0.29108286262221894, "grad_norm": 1.6612533330917358, "learning_rate": 8.316492652375903e-06, "loss": 0.5419, "step": 6944 }, { "epoch": 0.2911247812372279, "grad_norm": 1.4871913194656372, "learning_rate": 8.31598459936678e-06, "loss": 0.5822, "step": 6945 }, { "epoch": 0.29116669985223687, "grad_norm": 1.7076839208602905, "learning_rate": 8.315476485232068e-06, "loss": 0.5631, "step": 6946 }, { "epoch": 0.29120861846724583, "grad_norm": 1.5012744665145874, "learning_rate": 8.314968309981128e-06, "loss": 0.4702, "step": 6947 }, { "epoch": 0.2912505370822548, "grad_norm": 1.5040804147720337, "learning_rate": 8.314460073623333e-06, "loss": 0.4628, "step": 6948 }, { "epoch": 0.29129245569726375, "grad_norm": 1.5552462339401245, "learning_rate": 8.313951776168046e-06, "loss": 0.5083, "step": 6949 }, { "epoch": 0.2913343743122727, "grad_norm": 1.7292720079421997, "learning_rate": 8.31344341762464e-06, "loss": 0.5182, "step": 6950 }, { "epoch": 0.2913762929272817, "grad_norm": 3.191591501235962, "learning_rate": 8.312934998002487e-06, "loss": 0.462, "step": 6951 }, { "epoch": 0.29141821154229064, "grad_norm": 1.6324788331985474, "learning_rate": 8.312426517310956e-06, "loss": 0.5626, "step": 6952 }, { "epoch": 0.2914601301572996, "grad_norm": 1.6485276222229004, "learning_rate": 8.311917975559422e-06, "loss": 0.4892, "step": 6953 }, { "epoch": 0.29150204877230856, "grad_norm": 1.5971965789794922, "learning_rate": 8.311409372757258e-06, "loss": 0.538, "step": 6954 }, { "epoch": 0.2915439673873175, "grad_norm": 1.7537317276000977, "learning_rate": 8.31090070891384e-06, "loss": 0.5549, "step": 6955 }, { "epoch": 0.2915858860023265, "grad_norm": 1.6965793371200562, "learning_rate": 8.310391984038544e-06, "loss": 0.5204, "step": 6956 }, { "epoch": 0.29162780461733545, "grad_norm": 1.9049341678619385, "learning_rate": 8.309883198140749e-06, "loss": 0.5534, "step": 6957 }, { "epoch": 0.2916697232323444, "grad_norm": 1.6821635961532593, "learning_rate": 8.309374351229832e-06, "loss": 0.5303, "step": 6958 }, { "epoch": 0.2917116418473534, "grad_norm": 1.7915972471237183, "learning_rate": 8.308865443315178e-06, "loss": 0.5332, "step": 6959 }, { "epoch": 0.29175356046236234, "grad_norm": 1.6373094320297241, "learning_rate": 8.308356474406158e-06, "loss": 0.5226, "step": 6960 }, { "epoch": 0.2917954790773713, "grad_norm": 1.5178648233413696, "learning_rate": 8.307847444512164e-06, "loss": 0.525, "step": 6961 }, { "epoch": 0.29183739769238026, "grad_norm": 1.9217352867126465, "learning_rate": 8.307338353642576e-06, "loss": 0.5456, "step": 6962 }, { "epoch": 0.2918793163073892, "grad_norm": 1.511978030204773, "learning_rate": 8.306829201806775e-06, "loss": 0.4852, "step": 6963 }, { "epoch": 0.2919212349223982, "grad_norm": 1.5334428548812866, "learning_rate": 8.30631998901415e-06, "loss": 0.5184, "step": 6964 }, { "epoch": 0.29196315353740715, "grad_norm": 1.7271426916122437, "learning_rate": 8.305810715274087e-06, "loss": 0.5381, "step": 6965 }, { "epoch": 0.2920050721524161, "grad_norm": 1.8543404340744019, "learning_rate": 8.305301380595973e-06, "loss": 0.5422, "step": 6966 }, { "epoch": 0.292046990767425, "grad_norm": 1.783782720565796, "learning_rate": 8.3047919849892e-06, "loss": 0.5467, "step": 6967 }, { "epoch": 0.292088909382434, "grad_norm": 1.7858877182006836, "learning_rate": 8.304282528463153e-06, "loss": 0.5636, "step": 6968 }, { "epoch": 0.29213082799744294, "grad_norm": 1.527155876159668, "learning_rate": 8.303773011027225e-06, "loss": 0.5043, "step": 6969 }, { "epoch": 0.2921727466124519, "grad_norm": 1.457077145576477, "learning_rate": 8.30326343269081e-06, "loss": 0.5162, "step": 6970 }, { "epoch": 0.29221466522746087, "grad_norm": 1.7036566734313965, "learning_rate": 8.302753793463303e-06, "loss": 0.5312, "step": 6971 }, { "epoch": 0.29225658384246983, "grad_norm": 1.681758165359497, "learning_rate": 8.302244093354093e-06, "loss": 0.5295, "step": 6972 }, { "epoch": 0.2922985024574788, "grad_norm": 1.6936672925949097, "learning_rate": 8.301734332372578e-06, "loss": 0.5129, "step": 6973 }, { "epoch": 0.29234042107248775, "grad_norm": 1.493322730064392, "learning_rate": 8.301224510528155e-06, "loss": 0.5107, "step": 6974 }, { "epoch": 0.2923823396874967, "grad_norm": 1.5158123970031738, "learning_rate": 8.300714627830223e-06, "loss": 0.5207, "step": 6975 }, { "epoch": 0.2924242583025057, "grad_norm": 1.4801719188690186, "learning_rate": 8.30020468428818e-06, "loss": 0.5172, "step": 6976 }, { "epoch": 0.29246617691751464, "grad_norm": 1.8944494724273682, "learning_rate": 8.299694679911425e-06, "loss": 0.536, "step": 6977 }, { "epoch": 0.2925080955325236, "grad_norm": 1.6808125972747803, "learning_rate": 8.299184614709362e-06, "loss": 0.5161, "step": 6978 }, { "epoch": 0.29255001414753257, "grad_norm": 1.4079583883285522, "learning_rate": 8.29867448869139e-06, "loss": 0.5014, "step": 6979 }, { "epoch": 0.29259193276254153, "grad_norm": 1.9371696710586548, "learning_rate": 8.298164301866914e-06, "loss": 0.5354, "step": 6980 }, { "epoch": 0.2926338513775505, "grad_norm": 1.7109266519546509, "learning_rate": 8.29765405424534e-06, "loss": 0.557, "step": 6981 }, { "epoch": 0.29267576999255945, "grad_norm": 1.5921627283096313, "learning_rate": 8.297143745836071e-06, "loss": 0.5185, "step": 6982 }, { "epoch": 0.2927176886075684, "grad_norm": 1.672959327697754, "learning_rate": 8.296633376648516e-06, "loss": 0.4654, "step": 6983 }, { "epoch": 0.2927596072225774, "grad_norm": 1.7985007762908936, "learning_rate": 8.296122946692081e-06, "loss": 0.5318, "step": 6984 }, { "epoch": 0.29280152583758634, "grad_norm": 1.6902432441711426, "learning_rate": 8.295612455976177e-06, "loss": 0.5714, "step": 6985 }, { "epoch": 0.2928434444525953, "grad_norm": 1.5003718137741089, "learning_rate": 8.295101904510215e-06, "loss": 0.5428, "step": 6986 }, { "epoch": 0.29288536306760427, "grad_norm": 1.4878644943237305, "learning_rate": 8.294591292303602e-06, "loss": 0.5502, "step": 6987 }, { "epoch": 0.2929272816826132, "grad_norm": 1.7613283395767212, "learning_rate": 8.294080619365754e-06, "loss": 0.5626, "step": 6988 }, { "epoch": 0.2929692002976222, "grad_norm": 1.4944825172424316, "learning_rate": 8.293569885706085e-06, "loss": 0.5304, "step": 6989 }, { "epoch": 0.29301111891263115, "grad_norm": 1.5616298913955688, "learning_rate": 8.293059091334007e-06, "loss": 0.4594, "step": 6990 }, { "epoch": 0.2930530375276401, "grad_norm": 1.5891838073730469, "learning_rate": 8.29254823625894e-06, "loss": 0.52, "step": 6991 }, { "epoch": 0.293094956142649, "grad_norm": 1.5107238292694092, "learning_rate": 8.292037320490293e-06, "loss": 0.5039, "step": 6992 }, { "epoch": 0.293136874757658, "grad_norm": 1.3744803667068481, "learning_rate": 8.291526344037493e-06, "loss": 0.4793, "step": 6993 }, { "epoch": 0.29317879337266695, "grad_norm": 1.5223827362060547, "learning_rate": 8.291015306909954e-06, "loss": 0.5076, "step": 6994 }, { "epoch": 0.2932207119876759, "grad_norm": 1.6229350566864014, "learning_rate": 8.290504209117099e-06, "loss": 0.5459, "step": 6995 }, { "epoch": 0.29326263060268487, "grad_norm": 1.7298487424850464, "learning_rate": 8.289993050668347e-06, "loss": 0.5179, "step": 6996 }, { "epoch": 0.29330454921769383, "grad_norm": 1.7800425291061401, "learning_rate": 8.28948183157312e-06, "loss": 0.5302, "step": 6997 }, { "epoch": 0.2933464678327028, "grad_norm": 1.6654585599899292, "learning_rate": 8.288970551840843e-06, "loss": 0.554, "step": 6998 }, { "epoch": 0.29338838644771176, "grad_norm": 1.4992729425430298, "learning_rate": 8.288459211480942e-06, "loss": 0.4902, "step": 6999 }, { "epoch": 0.2934303050627207, "grad_norm": 1.8133819103240967, "learning_rate": 8.287947810502841e-06, "loss": 0.5827, "step": 7000 }, { "epoch": 0.2934722236777297, "grad_norm": 1.5616344213485718, "learning_rate": 8.28743634891597e-06, "loss": 0.5689, "step": 7001 }, { "epoch": 0.29351414229273864, "grad_norm": 1.9256670475006104, "learning_rate": 8.286924826729751e-06, "loss": 0.5497, "step": 7002 }, { "epoch": 0.2935560609077476, "grad_norm": 1.8953965902328491, "learning_rate": 8.286413243953618e-06, "loss": 0.5244, "step": 7003 }, { "epoch": 0.29359797952275657, "grad_norm": 1.7841331958770752, "learning_rate": 8.285901600597001e-06, "loss": 0.5076, "step": 7004 }, { "epoch": 0.29363989813776553, "grad_norm": 1.857678771018982, "learning_rate": 8.28538989666933e-06, "loss": 0.5062, "step": 7005 }, { "epoch": 0.2936818167527745, "grad_norm": 1.5427966117858887, "learning_rate": 8.28487813218004e-06, "loss": 0.4871, "step": 7006 }, { "epoch": 0.29372373536778346, "grad_norm": 1.7588900327682495, "learning_rate": 8.28436630713856e-06, "loss": 0.5241, "step": 7007 }, { "epoch": 0.2937656539827924, "grad_norm": 1.46497642993927, "learning_rate": 8.28385442155433e-06, "loss": 0.5352, "step": 7008 }, { "epoch": 0.2938075725978014, "grad_norm": 1.8937442302703857, "learning_rate": 8.28334247543678e-06, "loss": 0.5399, "step": 7009 }, { "epoch": 0.29384949121281034, "grad_norm": 1.7497810125350952, "learning_rate": 8.282830468795353e-06, "loss": 0.4796, "step": 7010 }, { "epoch": 0.2938914098278193, "grad_norm": 1.762037754058838, "learning_rate": 8.282318401639486e-06, "loss": 0.527, "step": 7011 }, { "epoch": 0.29393332844282827, "grad_norm": 1.8692286014556885, "learning_rate": 8.281806273978615e-06, "loss": 0.5369, "step": 7012 }, { "epoch": 0.29397524705783723, "grad_norm": 2.0650391578674316, "learning_rate": 8.281294085822183e-06, "loss": 0.4945, "step": 7013 }, { "epoch": 0.2940171656728462, "grad_norm": 2.317591428756714, "learning_rate": 8.28078183717963e-06, "loss": 0.5467, "step": 7014 }, { "epoch": 0.29405908428785515, "grad_norm": 2.4662740230560303, "learning_rate": 8.2802695280604e-06, "loss": 0.5552, "step": 7015 }, { "epoch": 0.2941010029028641, "grad_norm": 1.9734042882919312, "learning_rate": 8.279757158473936e-06, "loss": 0.5849, "step": 7016 }, { "epoch": 0.294142921517873, "grad_norm": 1.5259311199188232, "learning_rate": 8.279244728429684e-06, "loss": 0.5254, "step": 7017 }, { "epoch": 0.294184840132882, "grad_norm": 1.500205397605896, "learning_rate": 8.278732237937088e-06, "loss": 0.5284, "step": 7018 }, { "epoch": 0.29422675874789095, "grad_norm": 1.8270461559295654, "learning_rate": 8.278219687005596e-06, "loss": 0.5556, "step": 7019 }, { "epoch": 0.2942686773628999, "grad_norm": 2.1701297760009766, "learning_rate": 8.277707075644655e-06, "loss": 0.4914, "step": 7020 }, { "epoch": 0.2943105959779089, "grad_norm": 1.6250346899032593, "learning_rate": 8.277194403863715e-06, "loss": 0.5282, "step": 7021 }, { "epoch": 0.29435251459291784, "grad_norm": 1.5879747867584229, "learning_rate": 8.276681671672229e-06, "loss": 0.5052, "step": 7022 }, { "epoch": 0.2943944332079268, "grad_norm": 1.676889419555664, "learning_rate": 8.276168879079644e-06, "loss": 0.5613, "step": 7023 }, { "epoch": 0.29443635182293576, "grad_norm": 1.6892868280410767, "learning_rate": 8.275656026095417e-06, "loss": 0.5498, "step": 7024 }, { "epoch": 0.2944782704379447, "grad_norm": 1.6570580005645752, "learning_rate": 8.275143112728998e-06, "loss": 0.5271, "step": 7025 }, { "epoch": 0.2945201890529537, "grad_norm": 1.5753095149993896, "learning_rate": 8.274630138989845e-06, "loss": 0.56, "step": 7026 }, { "epoch": 0.29456210766796265, "grad_norm": 1.6783733367919922, "learning_rate": 8.274117104887411e-06, "loss": 0.5495, "step": 7027 }, { "epoch": 0.2946040262829716, "grad_norm": 1.5673056840896606, "learning_rate": 8.273604010431156e-06, "loss": 0.5501, "step": 7028 }, { "epoch": 0.29464594489798057, "grad_norm": 1.4230464696884155, "learning_rate": 8.273090855630536e-06, "loss": 0.5236, "step": 7029 }, { "epoch": 0.29468786351298953, "grad_norm": 1.6622713804244995, "learning_rate": 8.27257764049501e-06, "loss": 0.5657, "step": 7030 }, { "epoch": 0.2947297821279985, "grad_norm": 1.7730121612548828, "learning_rate": 8.27206436503404e-06, "loss": 0.5663, "step": 7031 }, { "epoch": 0.29477170074300746, "grad_norm": 1.5746729373931885, "learning_rate": 8.271551029257087e-06, "loss": 0.5067, "step": 7032 }, { "epoch": 0.2948136193580164, "grad_norm": 1.51893949508667, "learning_rate": 8.271037633173614e-06, "loss": 0.5305, "step": 7033 }, { "epoch": 0.2948555379730254, "grad_norm": 1.8580552339553833, "learning_rate": 8.270524176793085e-06, "loss": 0.5405, "step": 7034 }, { "epoch": 0.29489745658803435, "grad_norm": 1.5429308414459229, "learning_rate": 8.270010660124962e-06, "loss": 0.494, "step": 7035 }, { "epoch": 0.2949393752030433, "grad_norm": 1.9758367538452148, "learning_rate": 8.269497083178715e-06, "loss": 0.5237, "step": 7036 }, { "epoch": 0.29498129381805227, "grad_norm": 1.525831699371338, "learning_rate": 8.268983445963807e-06, "loss": 0.5406, "step": 7037 }, { "epoch": 0.29502321243306123, "grad_norm": 1.5153571367263794, "learning_rate": 8.268469748489711e-06, "loss": 0.5454, "step": 7038 }, { "epoch": 0.2950651310480702, "grad_norm": 1.56814444065094, "learning_rate": 8.267955990765893e-06, "loss": 0.5008, "step": 7039 }, { "epoch": 0.29510704966307916, "grad_norm": 1.7123420238494873, "learning_rate": 8.267442172801822e-06, "loss": 0.5684, "step": 7040 }, { "epoch": 0.2951489682780881, "grad_norm": 1.5259768962860107, "learning_rate": 8.266928294606974e-06, "loss": 0.5364, "step": 7041 }, { "epoch": 0.295190886893097, "grad_norm": 1.566023826599121, "learning_rate": 8.26641435619082e-06, "loss": 0.5245, "step": 7042 }, { "epoch": 0.295232805508106, "grad_norm": 1.6888267993927002, "learning_rate": 8.26590035756283e-06, "loss": 0.473, "step": 7043 }, { "epoch": 0.29527472412311495, "grad_norm": 1.9804445505142212, "learning_rate": 8.265386298732483e-06, "loss": 0.5216, "step": 7044 }, { "epoch": 0.2953166427381239, "grad_norm": 1.7014884948730469, "learning_rate": 8.264872179709256e-06, "loss": 0.5144, "step": 7045 }, { "epoch": 0.2953585613531329, "grad_norm": 1.9972151517868042, "learning_rate": 8.264358000502621e-06, "loss": 0.5297, "step": 7046 }, { "epoch": 0.29540047996814184, "grad_norm": 1.6258326768875122, "learning_rate": 8.263843761122061e-06, "loss": 0.5876, "step": 7047 }, { "epoch": 0.2954423985831508, "grad_norm": 1.743669867515564, "learning_rate": 8.263329461577052e-06, "loss": 0.5342, "step": 7048 }, { "epoch": 0.29548431719815976, "grad_norm": 1.6507726907730103, "learning_rate": 8.262815101877077e-06, "loss": 0.5022, "step": 7049 }, { "epoch": 0.2955262358131687, "grad_norm": 1.6532751321792603, "learning_rate": 8.262300682031616e-06, "loss": 0.4933, "step": 7050 }, { "epoch": 0.2955681544281777, "grad_norm": 1.5341224670410156, "learning_rate": 8.261786202050153e-06, "loss": 0.5149, "step": 7051 }, { "epoch": 0.29561007304318665, "grad_norm": 1.7182064056396484, "learning_rate": 8.261271661942168e-06, "loss": 0.5804, "step": 7052 }, { "epoch": 0.2956519916581956, "grad_norm": 1.9288123846054077, "learning_rate": 8.260757061717151e-06, "loss": 0.4886, "step": 7053 }, { "epoch": 0.2956939102732046, "grad_norm": 3.339369058609009, "learning_rate": 8.260242401384584e-06, "loss": 0.4957, "step": 7054 }, { "epoch": 0.29573582888821354, "grad_norm": 1.777793526649475, "learning_rate": 8.259727680953954e-06, "loss": 0.5634, "step": 7055 }, { "epoch": 0.2957777475032225, "grad_norm": 4.185717582702637, "learning_rate": 8.259212900434754e-06, "loss": 0.5096, "step": 7056 }, { "epoch": 0.29581966611823146, "grad_norm": 1.698136568069458, "learning_rate": 8.258698059836468e-06, "loss": 0.4744, "step": 7057 }, { "epoch": 0.2958615847332404, "grad_norm": 2.1616950035095215, "learning_rate": 8.258183159168589e-06, "loss": 0.6149, "step": 7058 }, { "epoch": 0.2959035033482494, "grad_norm": 1.7586932182312012, "learning_rate": 8.257668198440607e-06, "loss": 0.5495, "step": 7059 }, { "epoch": 0.29594542196325835, "grad_norm": 1.7961782217025757, "learning_rate": 8.257153177662016e-06, "loss": 0.5268, "step": 7060 }, { "epoch": 0.2959873405782673, "grad_norm": 1.540325403213501, "learning_rate": 8.256638096842309e-06, "loss": 0.5088, "step": 7061 }, { "epoch": 0.2960292591932763, "grad_norm": 1.6959295272827148, "learning_rate": 8.25612295599098e-06, "loss": 0.5621, "step": 7062 }, { "epoch": 0.29607117780828524, "grad_norm": 1.5002692937850952, "learning_rate": 8.255607755117528e-06, "loss": 0.4955, "step": 7063 }, { "epoch": 0.2961130964232942, "grad_norm": 1.557539463043213, "learning_rate": 8.255092494231446e-06, "loss": 0.5521, "step": 7064 }, { "epoch": 0.29615501503830316, "grad_norm": 1.7429518699645996, "learning_rate": 8.254577173342232e-06, "loss": 0.5654, "step": 7065 }, { "epoch": 0.2961969336533121, "grad_norm": 1.450339436531067, "learning_rate": 8.254061792459391e-06, "loss": 0.5049, "step": 7066 }, { "epoch": 0.29623885226832103, "grad_norm": 1.630880355834961, "learning_rate": 8.253546351592418e-06, "loss": 0.5387, "step": 7067 }, { "epoch": 0.29628077088333, "grad_norm": 1.757493257522583, "learning_rate": 8.253030850750815e-06, "loss": 0.5883, "step": 7068 }, { "epoch": 0.29632268949833895, "grad_norm": 1.5179468393325806, "learning_rate": 8.252515289944086e-06, "loss": 0.5277, "step": 7069 }, { "epoch": 0.2963646081133479, "grad_norm": 1.5267822742462158, "learning_rate": 8.251999669181735e-06, "loss": 0.5042, "step": 7070 }, { "epoch": 0.2964065267283569, "grad_norm": 1.456978678703308, "learning_rate": 8.251483988473264e-06, "loss": 0.4889, "step": 7071 }, { "epoch": 0.29644844534336584, "grad_norm": 1.6841349601745605, "learning_rate": 8.250968247828181e-06, "loss": 0.513, "step": 7072 }, { "epoch": 0.2964903639583748, "grad_norm": 1.4124934673309326, "learning_rate": 8.250452447255994e-06, "loss": 0.485, "step": 7073 }, { "epoch": 0.29653228257338377, "grad_norm": 1.717361330986023, "learning_rate": 8.24993658676621e-06, "loss": 0.5083, "step": 7074 }, { "epoch": 0.2965742011883927, "grad_norm": 1.6744407415390015, "learning_rate": 8.249420666368337e-06, "loss": 0.5215, "step": 7075 }, { "epoch": 0.2966161198034017, "grad_norm": 1.7165310382843018, "learning_rate": 8.248904686071885e-06, "loss": 0.5585, "step": 7076 }, { "epoch": 0.29665803841841065, "grad_norm": 1.7560198307037354, "learning_rate": 8.248388645886369e-06, "loss": 0.5603, "step": 7077 }, { "epoch": 0.2966999570334196, "grad_norm": 1.4434486627578735, "learning_rate": 8.247872545821298e-06, "loss": 0.5248, "step": 7078 }, { "epoch": 0.2967418756484286, "grad_norm": 1.6453893184661865, "learning_rate": 8.247356385886188e-06, "loss": 0.5466, "step": 7079 }, { "epoch": 0.29678379426343754, "grad_norm": 1.5420829057693481, "learning_rate": 8.24684016609055e-06, "loss": 0.4971, "step": 7080 }, { "epoch": 0.2968257128784465, "grad_norm": 1.696214199066162, "learning_rate": 8.246323886443903e-06, "loss": 0.5738, "step": 7081 }, { "epoch": 0.29686763149345546, "grad_norm": 1.6071680784225464, "learning_rate": 8.245807546955763e-06, "loss": 0.5249, "step": 7082 }, { "epoch": 0.2969095501084644, "grad_norm": 2.0973429679870605, "learning_rate": 8.245291147635648e-06, "loss": 0.4945, "step": 7083 }, { "epoch": 0.2969514687234734, "grad_norm": 1.5686863660812378, "learning_rate": 8.244774688493077e-06, "loss": 0.5499, "step": 7084 }, { "epoch": 0.29699338733848235, "grad_norm": 1.7829079627990723, "learning_rate": 8.244258169537571e-06, "loss": 0.5143, "step": 7085 }, { "epoch": 0.2970353059534913, "grad_norm": 1.6558804512023926, "learning_rate": 8.243741590778651e-06, "loss": 0.5867, "step": 7086 }, { "epoch": 0.2970772245685003, "grad_norm": 1.5888583660125732, "learning_rate": 8.243224952225838e-06, "loss": 0.5052, "step": 7087 }, { "epoch": 0.29711914318350924, "grad_norm": 1.6292483806610107, "learning_rate": 8.242708253888658e-06, "loss": 0.4824, "step": 7088 }, { "epoch": 0.2971610617985182, "grad_norm": 1.8459457159042358, "learning_rate": 8.242191495776633e-06, "loss": 0.5536, "step": 7089 }, { "epoch": 0.29720298041352716, "grad_norm": 1.8245190382003784, "learning_rate": 8.24167467789929e-06, "loss": 0.5197, "step": 7090 }, { "epoch": 0.2972448990285361, "grad_norm": 1.8918886184692383, "learning_rate": 8.241157800266157e-06, "loss": 0.5599, "step": 7091 }, { "epoch": 0.29728681764354503, "grad_norm": 2.018813371658325, "learning_rate": 8.24064086288676e-06, "loss": 0.6044, "step": 7092 }, { "epoch": 0.297328736258554, "grad_norm": 1.68244206905365, "learning_rate": 8.240123865770628e-06, "loss": 0.5319, "step": 7093 }, { "epoch": 0.29737065487356296, "grad_norm": 1.5821372270584106, "learning_rate": 8.239606808927294e-06, "loss": 0.5067, "step": 7094 }, { "epoch": 0.2974125734885719, "grad_norm": 1.5653802156448364, "learning_rate": 8.239089692366285e-06, "loss": 0.5547, "step": 7095 }, { "epoch": 0.2974544921035809, "grad_norm": 1.6068273782730103, "learning_rate": 8.238572516097137e-06, "loss": 0.5029, "step": 7096 }, { "epoch": 0.29749641071858984, "grad_norm": 1.469277262687683, "learning_rate": 8.23805528012938e-06, "loss": 0.508, "step": 7097 }, { "epoch": 0.2975383293335988, "grad_norm": 1.6311023235321045, "learning_rate": 8.23753798447255e-06, "loss": 0.5129, "step": 7098 }, { "epoch": 0.29758024794860777, "grad_norm": 1.67640221118927, "learning_rate": 8.237020629136185e-06, "loss": 0.5203, "step": 7099 }, { "epoch": 0.29762216656361673, "grad_norm": 2.8874781131744385, "learning_rate": 8.23650321412982e-06, "loss": 0.6119, "step": 7100 }, { "epoch": 0.2976640851786257, "grad_norm": 1.6658973693847656, "learning_rate": 8.235985739462993e-06, "loss": 0.5151, "step": 7101 }, { "epoch": 0.29770600379363465, "grad_norm": 1.4420640468597412, "learning_rate": 8.23546820514524e-06, "loss": 0.5171, "step": 7102 }, { "epoch": 0.2977479224086436, "grad_norm": 1.6706711053848267, "learning_rate": 8.234950611186106e-06, "loss": 0.5074, "step": 7103 }, { "epoch": 0.2977898410236526, "grad_norm": 1.585992693901062, "learning_rate": 8.234432957595129e-06, "loss": 0.5182, "step": 7104 }, { "epoch": 0.29783175963866154, "grad_norm": 1.7591063976287842, "learning_rate": 8.23391524438185e-06, "loss": 0.5137, "step": 7105 }, { "epoch": 0.2978736782536705, "grad_norm": 2.002218246459961, "learning_rate": 8.233397471555815e-06, "loss": 0.5202, "step": 7106 }, { "epoch": 0.29791559686867947, "grad_norm": 1.8760135173797607, "learning_rate": 8.23287963912657e-06, "loss": 0.5521, "step": 7107 }, { "epoch": 0.29795751548368843, "grad_norm": 2.3265445232391357, "learning_rate": 8.232361747103654e-06, "loss": 0.5334, "step": 7108 }, { "epoch": 0.2979994340986974, "grad_norm": 1.893201231956482, "learning_rate": 8.23184379549662e-06, "loss": 0.5224, "step": 7109 }, { "epoch": 0.29804135271370635, "grad_norm": 2.0270609855651855, "learning_rate": 8.231325784315015e-06, "loss": 0.5598, "step": 7110 }, { "epoch": 0.2980832713287153, "grad_norm": 1.9187419414520264, "learning_rate": 8.230807713568385e-06, "loss": 0.5261, "step": 7111 }, { "epoch": 0.2981251899437243, "grad_norm": 1.7432734966278076, "learning_rate": 8.23028958326628e-06, "loss": 0.5096, "step": 7112 }, { "epoch": 0.29816710855873324, "grad_norm": 1.8350508213043213, "learning_rate": 8.229771393418255e-06, "loss": 0.5536, "step": 7113 }, { "epoch": 0.2982090271737422, "grad_norm": 1.9125583171844482, "learning_rate": 8.229253144033857e-06, "loss": 0.5517, "step": 7114 }, { "epoch": 0.29825094578875117, "grad_norm": 1.7357280254364014, "learning_rate": 8.228734835122642e-06, "loss": 0.5807, "step": 7115 }, { "epoch": 0.2982928644037601, "grad_norm": 1.6662743091583252, "learning_rate": 8.228216466694165e-06, "loss": 0.5346, "step": 7116 }, { "epoch": 0.29833478301876903, "grad_norm": 1.4842939376831055, "learning_rate": 8.22769803875798e-06, "loss": 0.5253, "step": 7117 }, { "epoch": 0.298376701633778, "grad_norm": 1.5134339332580566, "learning_rate": 8.227179551323646e-06, "loss": 0.5183, "step": 7118 }, { "epoch": 0.29841862024878696, "grad_norm": 1.5255558490753174, "learning_rate": 8.226661004400714e-06, "loss": 0.5454, "step": 7119 }, { "epoch": 0.2984605388637959, "grad_norm": 1.7571414709091187, "learning_rate": 8.22614239799875e-06, "loss": 0.515, "step": 7120 }, { "epoch": 0.2985024574788049, "grad_norm": 1.9619356393814087, "learning_rate": 8.22562373212731e-06, "loss": 0.5515, "step": 7121 }, { "epoch": 0.29854437609381385, "grad_norm": 1.6683732271194458, "learning_rate": 8.225105006795954e-06, "loss": 0.5555, "step": 7122 }, { "epoch": 0.2985862947088228, "grad_norm": 1.5646427869796753, "learning_rate": 8.224586222014248e-06, "loss": 0.5307, "step": 7123 }, { "epoch": 0.29862821332383177, "grad_norm": 1.799748420715332, "learning_rate": 8.224067377791754e-06, "loss": 0.5696, "step": 7124 }, { "epoch": 0.29867013193884073, "grad_norm": 1.9679138660430908, "learning_rate": 8.223548474138034e-06, "loss": 0.549, "step": 7125 }, { "epoch": 0.2987120505538497, "grad_norm": 1.757178544998169, "learning_rate": 8.223029511062652e-06, "loss": 0.5744, "step": 7126 }, { "epoch": 0.29875396916885866, "grad_norm": 2.264549970626831, "learning_rate": 8.222510488575179e-06, "loss": 0.5251, "step": 7127 }, { "epoch": 0.2987958877838676, "grad_norm": 2.073390245437622, "learning_rate": 8.221991406685179e-06, "loss": 0.512, "step": 7128 }, { "epoch": 0.2988378063988766, "grad_norm": 1.7150015830993652, "learning_rate": 8.221472265402223e-06, "loss": 0.5279, "step": 7129 }, { "epoch": 0.29887972501388554, "grad_norm": 1.5073838233947754, "learning_rate": 8.220953064735878e-06, "loss": 0.5543, "step": 7130 }, { "epoch": 0.2989216436288945, "grad_norm": 1.5534229278564453, "learning_rate": 8.220433804695718e-06, "loss": 0.509, "step": 7131 }, { "epoch": 0.29896356224390347, "grad_norm": 1.5336533784866333, "learning_rate": 8.21991448529131e-06, "loss": 0.5077, "step": 7132 }, { "epoch": 0.29900548085891243, "grad_norm": 1.5250269174575806, "learning_rate": 8.219395106532233e-06, "loss": 0.4676, "step": 7133 }, { "epoch": 0.2990473994739214, "grad_norm": 1.5248206853866577, "learning_rate": 8.218875668428056e-06, "loss": 0.506, "step": 7134 }, { "epoch": 0.29908931808893036, "grad_norm": 1.6373279094696045, "learning_rate": 8.218356170988354e-06, "loss": 0.5362, "step": 7135 }, { "epoch": 0.2991312367039393, "grad_norm": 1.7803043127059937, "learning_rate": 8.217836614222709e-06, "loss": 0.5414, "step": 7136 }, { "epoch": 0.2991731553189483, "grad_norm": 1.7128756046295166, "learning_rate": 8.217316998140695e-06, "loss": 0.5535, "step": 7137 }, { "epoch": 0.29921507393395724, "grad_norm": 1.5441370010375977, "learning_rate": 8.216797322751887e-06, "loss": 0.4885, "step": 7138 }, { "epoch": 0.2992569925489662, "grad_norm": 1.7737778425216675, "learning_rate": 8.216277588065868e-06, "loss": 0.5593, "step": 7139 }, { "epoch": 0.29929891116397517, "grad_norm": 1.847916841506958, "learning_rate": 8.215757794092218e-06, "loss": 0.5728, "step": 7140 }, { "epoch": 0.29934082977898413, "grad_norm": 1.6177897453308105, "learning_rate": 8.21523794084052e-06, "loss": 0.5082, "step": 7141 }, { "epoch": 0.29938274839399304, "grad_norm": 1.8084394931793213, "learning_rate": 8.214718028320354e-06, "loss": 0.5247, "step": 7142 }, { "epoch": 0.299424667009002, "grad_norm": 1.5999752283096313, "learning_rate": 8.214198056541307e-06, "loss": 0.5494, "step": 7143 }, { "epoch": 0.29946658562401096, "grad_norm": 1.4427411556243896, "learning_rate": 8.213678025512961e-06, "loss": 0.5128, "step": 7144 }, { "epoch": 0.2995085042390199, "grad_norm": 2.179080009460449, "learning_rate": 8.213157935244903e-06, "loss": 0.5058, "step": 7145 }, { "epoch": 0.2995504228540289, "grad_norm": 1.5054538249969482, "learning_rate": 8.212637785746723e-06, "loss": 0.4691, "step": 7146 }, { "epoch": 0.29959234146903785, "grad_norm": 1.7653483152389526, "learning_rate": 8.212117577028004e-06, "loss": 0.5193, "step": 7147 }, { "epoch": 0.2996342600840468, "grad_norm": 1.8234636783599854, "learning_rate": 8.211597309098341e-06, "loss": 0.4881, "step": 7148 }, { "epoch": 0.2996761786990558, "grad_norm": 1.6234620809555054, "learning_rate": 8.211076981967318e-06, "loss": 0.5471, "step": 7149 }, { "epoch": 0.29971809731406474, "grad_norm": 2.0679056644439697, "learning_rate": 8.210556595644533e-06, "loss": 0.4732, "step": 7150 }, { "epoch": 0.2997600159290737, "grad_norm": 1.6891732215881348, "learning_rate": 8.210036150139573e-06, "loss": 0.5293, "step": 7151 }, { "epoch": 0.29980193454408266, "grad_norm": 1.9411282539367676, "learning_rate": 8.209515645462036e-06, "loss": 0.4914, "step": 7152 }, { "epoch": 0.2998438531590916, "grad_norm": 1.6131409406661987, "learning_rate": 8.208995081621516e-06, "loss": 0.5158, "step": 7153 }, { "epoch": 0.2998857717741006, "grad_norm": 1.63996422290802, "learning_rate": 8.208474458627607e-06, "loss": 0.5359, "step": 7154 }, { "epoch": 0.29992769038910955, "grad_norm": 1.636487364768982, "learning_rate": 8.207953776489905e-06, "loss": 0.504, "step": 7155 }, { "epoch": 0.2999696090041185, "grad_norm": 1.7789781093597412, "learning_rate": 8.207433035218013e-06, "loss": 0.5055, "step": 7156 }, { "epoch": 0.30001152761912747, "grad_norm": 1.4915790557861328, "learning_rate": 8.206912234821525e-06, "loss": 0.516, "step": 7157 }, { "epoch": 0.30005344623413643, "grad_norm": 1.5223071575164795, "learning_rate": 8.206391375310045e-06, "loss": 0.5254, "step": 7158 }, { "epoch": 0.3000953648491454, "grad_norm": 1.572163462638855, "learning_rate": 8.205870456693172e-06, "loss": 0.5372, "step": 7159 }, { "epoch": 0.30013728346415436, "grad_norm": 1.6155413389205933, "learning_rate": 8.20534947898051e-06, "loss": 0.5855, "step": 7160 }, { "epoch": 0.3001792020791633, "grad_norm": 1.8402247428894043, "learning_rate": 8.204828442181657e-06, "loss": 0.5787, "step": 7161 }, { "epoch": 0.3002211206941723, "grad_norm": 1.7248437404632568, "learning_rate": 8.204307346306228e-06, "loss": 0.5413, "step": 7162 }, { "epoch": 0.30026303930918125, "grad_norm": 1.8755919933319092, "learning_rate": 8.20378619136382e-06, "loss": 0.5123, "step": 7163 }, { "epoch": 0.3003049579241902, "grad_norm": 1.653073787689209, "learning_rate": 8.203264977364043e-06, "loss": 0.5161, "step": 7164 }, { "epoch": 0.30034687653919917, "grad_norm": 1.7205523252487183, "learning_rate": 8.202743704316506e-06, "loss": 0.5552, "step": 7165 }, { "epoch": 0.30038879515420813, "grad_norm": 2.250609874725342, "learning_rate": 8.202222372230815e-06, "loss": 0.5978, "step": 7166 }, { "epoch": 0.30043071376921704, "grad_norm": 1.4670668840408325, "learning_rate": 8.201700981116582e-06, "loss": 0.4882, "step": 7167 }, { "epoch": 0.300472632384226, "grad_norm": 2.0232937335968018, "learning_rate": 8.20117953098342e-06, "loss": 0.5737, "step": 7168 }, { "epoch": 0.30051455099923496, "grad_norm": 1.8501304388046265, "learning_rate": 8.200658021840934e-06, "loss": 0.5065, "step": 7169 }, { "epoch": 0.3005564696142439, "grad_norm": 1.5190712213516235, "learning_rate": 8.200136453698744e-06, "loss": 0.4946, "step": 7170 }, { "epoch": 0.3005983882292529, "grad_norm": 1.527220606803894, "learning_rate": 8.199614826566465e-06, "loss": 0.5118, "step": 7171 }, { "epoch": 0.30064030684426185, "grad_norm": 1.6530276536941528, "learning_rate": 8.199093140453707e-06, "loss": 0.5162, "step": 7172 }, { "epoch": 0.3006822254592708, "grad_norm": 1.9702554941177368, "learning_rate": 8.19857139537009e-06, "loss": 0.5919, "step": 7173 }, { "epoch": 0.3007241440742798, "grad_norm": 1.6527760028839111, "learning_rate": 8.198049591325233e-06, "loss": 0.5045, "step": 7174 }, { "epoch": 0.30076606268928874, "grad_norm": 1.6738121509552002, "learning_rate": 8.197527728328752e-06, "loss": 0.4877, "step": 7175 }, { "epoch": 0.3008079813042977, "grad_norm": 1.5551995038986206, "learning_rate": 8.197005806390265e-06, "loss": 0.5265, "step": 7176 }, { "epoch": 0.30084989991930666, "grad_norm": 1.7015334367752075, "learning_rate": 8.196483825519398e-06, "loss": 0.5405, "step": 7177 }, { "epoch": 0.3008918185343156, "grad_norm": 1.5575685501098633, "learning_rate": 8.195961785725772e-06, "loss": 0.5104, "step": 7178 }, { "epoch": 0.3009337371493246, "grad_norm": 1.6006312370300293, "learning_rate": 8.195439687019009e-06, "loss": 0.5638, "step": 7179 }, { "epoch": 0.30097565576433355, "grad_norm": 1.4662748575210571, "learning_rate": 8.19491752940873e-06, "loss": 0.484, "step": 7180 }, { "epoch": 0.3010175743793425, "grad_norm": 1.5402543544769287, "learning_rate": 8.194395312904566e-06, "loss": 0.5255, "step": 7181 }, { "epoch": 0.3010594929943515, "grad_norm": 1.4985355138778687, "learning_rate": 8.193873037516139e-06, "loss": 0.5198, "step": 7182 }, { "epoch": 0.30110141160936044, "grad_norm": 1.6731547117233276, "learning_rate": 8.193350703253077e-06, "loss": 0.488, "step": 7183 }, { "epoch": 0.3011433302243694, "grad_norm": 2.0207505226135254, "learning_rate": 8.192828310125012e-06, "loss": 0.5377, "step": 7184 }, { "epoch": 0.30118524883937836, "grad_norm": 1.7483278512954712, "learning_rate": 8.192305858141569e-06, "loss": 0.5322, "step": 7185 }, { "epoch": 0.3012271674543873, "grad_norm": 1.6601039171218872, "learning_rate": 8.191783347312382e-06, "loss": 0.5091, "step": 7186 }, { "epoch": 0.3012690860693963, "grad_norm": 1.6255282163619995, "learning_rate": 8.191260777647083e-06, "loss": 0.5449, "step": 7187 }, { "epoch": 0.30131100468440525, "grad_norm": 2.0322751998901367, "learning_rate": 8.190738149155299e-06, "loss": 0.5672, "step": 7188 }, { "epoch": 0.3013529232994142, "grad_norm": 1.5850163698196411, "learning_rate": 8.19021546184667e-06, "loss": 0.51, "step": 7189 }, { "epoch": 0.3013948419144232, "grad_norm": 1.6411093473434448, "learning_rate": 8.189692715730833e-06, "loss": 0.4882, "step": 7190 }, { "epoch": 0.30143676052943214, "grad_norm": 2.2380969524383545, "learning_rate": 8.189169910817416e-06, "loss": 0.5238, "step": 7191 }, { "epoch": 0.30147867914444104, "grad_norm": 2.2599141597747803, "learning_rate": 8.188647047116062e-06, "loss": 0.5921, "step": 7192 }, { "epoch": 0.30152059775945, "grad_norm": 1.853467583656311, "learning_rate": 8.188124124636408e-06, "loss": 0.5707, "step": 7193 }, { "epoch": 0.30156251637445897, "grad_norm": 1.4392945766448975, "learning_rate": 8.187601143388092e-06, "loss": 0.485, "step": 7194 }, { "epoch": 0.30160443498946793, "grad_norm": 1.6229490041732788, "learning_rate": 8.187078103380757e-06, "loss": 0.5691, "step": 7195 }, { "epoch": 0.3016463536044769, "grad_norm": 1.9316469430923462, "learning_rate": 8.186555004624043e-06, "loss": 0.5809, "step": 7196 }, { "epoch": 0.30168827221948585, "grad_norm": 1.6238356828689575, "learning_rate": 8.186031847127594e-06, "loss": 0.5649, "step": 7197 }, { "epoch": 0.3017301908344948, "grad_norm": 1.7230173349380493, "learning_rate": 8.18550863090105e-06, "loss": 0.579, "step": 7198 }, { "epoch": 0.3017721094495038, "grad_norm": 1.406849980354309, "learning_rate": 8.18498535595406e-06, "loss": 0.5357, "step": 7199 }, { "epoch": 0.30181402806451274, "grad_norm": 2.175023078918457, "learning_rate": 8.184462022296267e-06, "loss": 0.5846, "step": 7200 }, { "epoch": 0.3018559466795217, "grad_norm": 1.7945555448532104, "learning_rate": 8.183938629937319e-06, "loss": 0.536, "step": 7201 }, { "epoch": 0.30189786529453067, "grad_norm": 1.742302417755127, "learning_rate": 8.183415178886865e-06, "loss": 0.5751, "step": 7202 }, { "epoch": 0.3019397839095396, "grad_norm": 1.6686058044433594, "learning_rate": 8.18289166915455e-06, "loss": 0.4625, "step": 7203 }, { "epoch": 0.3019817025245486, "grad_norm": 1.7856248617172241, "learning_rate": 8.182368100750033e-06, "loss": 0.5, "step": 7204 }, { "epoch": 0.30202362113955755, "grad_norm": 1.6122398376464844, "learning_rate": 8.181844473682956e-06, "loss": 0.5124, "step": 7205 }, { "epoch": 0.3020655397545665, "grad_norm": 1.4449498653411865, "learning_rate": 8.181320787962974e-06, "loss": 0.4911, "step": 7206 }, { "epoch": 0.3021074583695755, "grad_norm": 1.7724759578704834, "learning_rate": 8.180797043599743e-06, "loss": 0.5144, "step": 7207 }, { "epoch": 0.30214937698458444, "grad_norm": 1.744871735572815, "learning_rate": 8.180273240602917e-06, "loss": 0.5295, "step": 7208 }, { "epoch": 0.3021912955995934, "grad_norm": 1.5300965309143066, "learning_rate": 8.179749378982147e-06, "loss": 0.4991, "step": 7209 }, { "epoch": 0.30223321421460236, "grad_norm": 1.5763773918151855, "learning_rate": 8.179225458747096e-06, "loss": 0.4806, "step": 7210 }, { "epoch": 0.3022751328296113, "grad_norm": 1.7742140293121338, "learning_rate": 8.178701479907417e-06, "loss": 0.5651, "step": 7211 }, { "epoch": 0.3023170514446203, "grad_norm": 1.5671610832214355, "learning_rate": 8.17817744247277e-06, "loss": 0.5148, "step": 7212 }, { "epoch": 0.30235897005962925, "grad_norm": 1.6850404739379883, "learning_rate": 8.17765334645282e-06, "loss": 0.5618, "step": 7213 }, { "epoch": 0.3024008886746382, "grad_norm": 1.6555074453353882, "learning_rate": 8.177129191857217e-06, "loss": 0.5413, "step": 7214 }, { "epoch": 0.3024428072896472, "grad_norm": 1.890329360961914, "learning_rate": 8.176604978695635e-06, "loss": 0.5397, "step": 7215 }, { "epoch": 0.30248472590465614, "grad_norm": 1.9379827976226807, "learning_rate": 8.17608070697773e-06, "loss": 0.5709, "step": 7216 }, { "epoch": 0.30252664451966504, "grad_norm": 1.874061942100525, "learning_rate": 8.175556376713168e-06, "loss": 0.5545, "step": 7217 }, { "epoch": 0.302568563134674, "grad_norm": 1.6251835823059082, "learning_rate": 8.175031987911613e-06, "loss": 0.5401, "step": 7218 }, { "epoch": 0.30261048174968297, "grad_norm": 1.6924492120742798, "learning_rate": 8.174507540582734e-06, "loss": 0.4848, "step": 7219 }, { "epoch": 0.30265240036469193, "grad_norm": 1.6144907474517822, "learning_rate": 8.173983034736196e-06, "loss": 0.5691, "step": 7220 }, { "epoch": 0.3026943189797009, "grad_norm": 1.7186920642852783, "learning_rate": 8.17345847038167e-06, "loss": 0.5643, "step": 7221 }, { "epoch": 0.30273623759470986, "grad_norm": 1.71539306640625, "learning_rate": 8.172933847528824e-06, "loss": 0.5304, "step": 7222 }, { "epoch": 0.3027781562097188, "grad_norm": 1.9883078336715698, "learning_rate": 8.17240916618733e-06, "loss": 0.5221, "step": 7223 }, { "epoch": 0.3028200748247278, "grad_norm": 1.3726128339767456, "learning_rate": 8.171884426366856e-06, "loss": 0.526, "step": 7224 }, { "epoch": 0.30286199343973674, "grad_norm": 2.206482410430908, "learning_rate": 8.171359628077079e-06, "loss": 0.5138, "step": 7225 }, { "epoch": 0.3029039120547457, "grad_norm": 1.6415514945983887, "learning_rate": 8.170834771327671e-06, "loss": 0.5366, "step": 7226 }, { "epoch": 0.30294583066975467, "grad_norm": 1.481994390487671, "learning_rate": 8.170309856128309e-06, "loss": 0.4756, "step": 7227 }, { "epoch": 0.30298774928476363, "grad_norm": 1.6441112756729126, "learning_rate": 8.169784882488666e-06, "loss": 0.5211, "step": 7228 }, { "epoch": 0.3030296678997726, "grad_norm": 1.6719805002212524, "learning_rate": 8.169259850418422e-06, "loss": 0.5885, "step": 7229 }, { "epoch": 0.30307158651478155, "grad_norm": 1.5935529470443726, "learning_rate": 8.168734759927252e-06, "loss": 0.5468, "step": 7230 }, { "epoch": 0.3031135051297905, "grad_norm": 1.5929688215255737, "learning_rate": 8.16820961102484e-06, "loss": 0.5351, "step": 7231 }, { "epoch": 0.3031554237447995, "grad_norm": 1.677680492401123, "learning_rate": 8.167684403720862e-06, "loss": 0.5883, "step": 7232 }, { "epoch": 0.30319734235980844, "grad_norm": 1.8912694454193115, "learning_rate": 8.167159138025003e-06, "loss": 0.5536, "step": 7233 }, { "epoch": 0.3032392609748174, "grad_norm": 1.5471941232681274, "learning_rate": 8.166633813946942e-06, "loss": 0.5078, "step": 7234 }, { "epoch": 0.30328117958982637, "grad_norm": 1.6612491607666016, "learning_rate": 8.166108431496364e-06, "loss": 0.5321, "step": 7235 }, { "epoch": 0.30332309820483533, "grad_norm": 1.5026731491088867, "learning_rate": 8.165582990682956e-06, "loss": 0.4911, "step": 7236 }, { "epoch": 0.3033650168198443, "grad_norm": 1.7216068506240845, "learning_rate": 8.1650574915164e-06, "loss": 0.5081, "step": 7237 }, { "epoch": 0.30340693543485325, "grad_norm": 1.8692750930786133, "learning_rate": 8.164531934006386e-06, "loss": 0.5308, "step": 7238 }, { "epoch": 0.3034488540498622, "grad_norm": 1.8923134803771973, "learning_rate": 8.1640063181626e-06, "loss": 0.5351, "step": 7239 }, { "epoch": 0.3034907726648712, "grad_norm": 1.8855353593826294, "learning_rate": 8.163480643994733e-06, "loss": 0.5854, "step": 7240 }, { "epoch": 0.30353269127988014, "grad_norm": 1.7305967807769775, "learning_rate": 8.162954911512472e-06, "loss": 0.5319, "step": 7241 }, { "epoch": 0.30357460989488905, "grad_norm": 1.8247166872024536, "learning_rate": 8.16242912072551e-06, "loss": 0.5498, "step": 7242 }, { "epoch": 0.303616528509898, "grad_norm": 1.4595905542373657, "learning_rate": 8.16190327164354e-06, "loss": 0.4901, "step": 7243 }, { "epoch": 0.30365844712490697, "grad_norm": 1.4515999555587769, "learning_rate": 8.161377364276255e-06, "loss": 0.5215, "step": 7244 }, { "epoch": 0.30370036573991593, "grad_norm": 2.0637896060943604, "learning_rate": 8.160851398633347e-06, "loss": 0.5294, "step": 7245 }, { "epoch": 0.3037422843549249, "grad_norm": 1.7703081369400024, "learning_rate": 8.160325374724513e-06, "loss": 0.5585, "step": 7246 }, { "epoch": 0.30378420296993386, "grad_norm": 2.2208175659179688, "learning_rate": 8.159799292559452e-06, "loss": 0.472, "step": 7247 }, { "epoch": 0.3038261215849428, "grad_norm": 1.7261426448822021, "learning_rate": 8.159273152147857e-06, "loss": 0.5293, "step": 7248 }, { "epoch": 0.3038680401999518, "grad_norm": 1.8439475297927856, "learning_rate": 8.15874695349943e-06, "loss": 0.5778, "step": 7249 }, { "epoch": 0.30390995881496075, "grad_norm": 1.6979477405548096, "learning_rate": 8.15822069662387e-06, "loss": 0.5688, "step": 7250 }, { "epoch": 0.3039518774299697, "grad_norm": 1.931257724761963, "learning_rate": 8.157694381530877e-06, "loss": 0.518, "step": 7251 }, { "epoch": 0.30399379604497867, "grad_norm": 1.8829456567764282, "learning_rate": 8.157168008230155e-06, "loss": 0.4979, "step": 7252 }, { "epoch": 0.30403571465998763, "grad_norm": 1.6692315340042114, "learning_rate": 8.156641576731403e-06, "loss": 0.5375, "step": 7253 }, { "epoch": 0.3040776332749966, "grad_norm": 2.407325029373169, "learning_rate": 8.156115087044329e-06, "loss": 0.5437, "step": 7254 }, { "epoch": 0.30411955189000556, "grad_norm": 1.6604138612747192, "learning_rate": 8.155588539178635e-06, "loss": 0.5152, "step": 7255 }, { "epoch": 0.3041614705050145, "grad_norm": 1.8719929456710815, "learning_rate": 8.15506193314403e-06, "loss": 0.5289, "step": 7256 }, { "epoch": 0.3042033891200235, "grad_norm": 1.6123005151748657, "learning_rate": 8.15453526895022e-06, "loss": 0.5305, "step": 7257 }, { "epoch": 0.30424530773503244, "grad_norm": 1.7898468971252441, "learning_rate": 8.154008546606912e-06, "loss": 0.5458, "step": 7258 }, { "epoch": 0.3042872263500414, "grad_norm": 2.0040013790130615, "learning_rate": 8.153481766123819e-06, "loss": 0.559, "step": 7259 }, { "epoch": 0.30432914496505037, "grad_norm": 1.462962031364441, "learning_rate": 8.152954927510648e-06, "loss": 0.4559, "step": 7260 }, { "epoch": 0.30437106358005933, "grad_norm": 1.6974856853485107, "learning_rate": 8.152428030777113e-06, "loss": 0.5815, "step": 7261 }, { "epoch": 0.3044129821950683, "grad_norm": 1.7607147693634033, "learning_rate": 8.151901075932924e-06, "loss": 0.4943, "step": 7262 }, { "epoch": 0.30445490081007726, "grad_norm": 2.502974510192871, "learning_rate": 8.151374062987796e-06, "loss": 0.5041, "step": 7263 }, { "epoch": 0.3044968194250862, "grad_norm": 1.952415943145752, "learning_rate": 8.150846991951444e-06, "loss": 0.5357, "step": 7264 }, { "epoch": 0.3045387380400952, "grad_norm": 1.6453590393066406, "learning_rate": 8.150319862833584e-06, "loss": 0.5416, "step": 7265 }, { "epoch": 0.30458065665510414, "grad_norm": 2.0274577140808105, "learning_rate": 8.149792675643932e-06, "loss": 0.5852, "step": 7266 }, { "epoch": 0.30462257527011305, "grad_norm": 2.062577724456787, "learning_rate": 8.149265430392207e-06, "loss": 0.5758, "step": 7267 }, { "epoch": 0.304664493885122, "grad_norm": 1.5788819789886475, "learning_rate": 8.148738127088127e-06, "loss": 0.4917, "step": 7268 }, { "epoch": 0.304706412500131, "grad_norm": 1.8532828092575073, "learning_rate": 8.148210765741413e-06, "loss": 0.5138, "step": 7269 }, { "epoch": 0.30474833111513994, "grad_norm": 1.8706834316253662, "learning_rate": 8.147683346361786e-06, "loss": 0.515, "step": 7270 }, { "epoch": 0.3047902497301489, "grad_norm": 1.6893616914749146, "learning_rate": 8.147155868958967e-06, "loss": 0.5204, "step": 7271 }, { "epoch": 0.30483216834515786, "grad_norm": 2.0652284622192383, "learning_rate": 8.146628333542682e-06, "loss": 0.5654, "step": 7272 }, { "epoch": 0.3048740869601668, "grad_norm": 1.6214638948440552, "learning_rate": 8.146100740122652e-06, "loss": 0.5664, "step": 7273 }, { "epoch": 0.3049160055751758, "grad_norm": 1.7118887901306152, "learning_rate": 8.145573088708605e-06, "loss": 0.5159, "step": 7274 }, { "epoch": 0.30495792419018475, "grad_norm": 1.9740419387817383, "learning_rate": 8.145045379310269e-06, "loss": 0.536, "step": 7275 }, { "epoch": 0.3049998428051937, "grad_norm": 1.616156816482544, "learning_rate": 8.144517611937367e-06, "loss": 0.5369, "step": 7276 }, { "epoch": 0.3050417614202027, "grad_norm": 1.743239402770996, "learning_rate": 8.143989786599628e-06, "loss": 0.5319, "step": 7277 }, { "epoch": 0.30508368003521164, "grad_norm": 1.6432199478149414, "learning_rate": 8.143461903306786e-06, "loss": 0.4706, "step": 7278 }, { "epoch": 0.3051255986502206, "grad_norm": 1.7639645338058472, "learning_rate": 8.14293396206857e-06, "loss": 0.5985, "step": 7279 }, { "epoch": 0.30516751726522956, "grad_norm": 1.7550286054611206, "learning_rate": 8.14240596289471e-06, "loss": 0.5088, "step": 7280 }, { "epoch": 0.3052094358802385, "grad_norm": 1.6066324710845947, "learning_rate": 8.141877905794941e-06, "loss": 0.5391, "step": 7281 }, { "epoch": 0.3052513544952475, "grad_norm": 1.7104301452636719, "learning_rate": 8.141349790778996e-06, "loss": 0.5789, "step": 7282 }, { "epoch": 0.30529327311025645, "grad_norm": 1.5781042575836182, "learning_rate": 8.140821617856612e-06, "loss": 0.5052, "step": 7283 }, { "epoch": 0.3053351917252654, "grad_norm": 1.4929471015930176, "learning_rate": 8.14029338703752e-06, "loss": 0.5395, "step": 7284 }, { "epoch": 0.30537711034027437, "grad_norm": 1.8109787702560425, "learning_rate": 8.139765098331464e-06, "loss": 0.5701, "step": 7285 }, { "epoch": 0.30541902895528333, "grad_norm": 1.7760556936264038, "learning_rate": 8.139236751748178e-06, "loss": 0.5354, "step": 7286 }, { "epoch": 0.3054609475702923, "grad_norm": 1.7984583377838135, "learning_rate": 8.138708347297404e-06, "loss": 0.5544, "step": 7287 }, { "epoch": 0.30550286618530126, "grad_norm": 1.7801728248596191, "learning_rate": 8.138179884988879e-06, "loss": 0.561, "step": 7288 }, { "epoch": 0.3055447848003102, "grad_norm": 2.1560211181640625, "learning_rate": 8.137651364832346e-06, "loss": 0.5696, "step": 7289 }, { "epoch": 0.3055867034153192, "grad_norm": 1.5812631845474243, "learning_rate": 8.137122786837549e-06, "loss": 0.5321, "step": 7290 }, { "epoch": 0.30562862203032815, "grad_norm": 1.8596898317337036, "learning_rate": 8.136594151014227e-06, "loss": 0.5609, "step": 7291 }, { "epoch": 0.30567054064533705, "grad_norm": 1.9347400665283203, "learning_rate": 8.136065457372132e-06, "loss": 0.5455, "step": 7292 }, { "epoch": 0.305712459260346, "grad_norm": 1.681916356086731, "learning_rate": 8.135536705921003e-06, "loss": 0.511, "step": 7293 }, { "epoch": 0.305754377875355, "grad_norm": 1.9488145112991333, "learning_rate": 8.135007896670592e-06, "loss": 0.5148, "step": 7294 }, { "epoch": 0.30579629649036394, "grad_norm": 1.7591161727905273, "learning_rate": 8.134479029630645e-06, "loss": 0.5358, "step": 7295 }, { "epoch": 0.3058382151053729, "grad_norm": 1.690366268157959, "learning_rate": 8.133950104810907e-06, "loss": 0.5681, "step": 7296 }, { "epoch": 0.30588013372038186, "grad_norm": 1.7369414567947388, "learning_rate": 8.133421122221132e-06, "loss": 0.5063, "step": 7297 }, { "epoch": 0.3059220523353908, "grad_norm": 2.083692789077759, "learning_rate": 8.132892081871073e-06, "loss": 0.5027, "step": 7298 }, { "epoch": 0.3059639709503998, "grad_norm": 7.084941387176514, "learning_rate": 8.132362983770477e-06, "loss": 0.5036, "step": 7299 }, { "epoch": 0.30600588956540875, "grad_norm": 1.8057390451431274, "learning_rate": 8.131833827929101e-06, "loss": 0.5426, "step": 7300 }, { "epoch": 0.3060478081804177, "grad_norm": 1.541200041770935, "learning_rate": 8.131304614356699e-06, "loss": 0.5361, "step": 7301 }, { "epoch": 0.3060897267954267, "grad_norm": 1.6059149503707886, "learning_rate": 8.130775343063023e-06, "loss": 0.5228, "step": 7302 }, { "epoch": 0.30613164541043564, "grad_norm": 1.6574862003326416, "learning_rate": 8.130246014057833e-06, "loss": 0.4802, "step": 7303 }, { "epoch": 0.3061735640254446, "grad_norm": 1.5135747194290161, "learning_rate": 8.129716627350885e-06, "loss": 0.4773, "step": 7304 }, { "epoch": 0.30621548264045356, "grad_norm": 1.914405345916748, "learning_rate": 8.129187182951938e-06, "loss": 0.5143, "step": 7305 }, { "epoch": 0.3062574012554625, "grad_norm": 1.5345796346664429, "learning_rate": 8.128657680870751e-06, "loss": 0.4961, "step": 7306 }, { "epoch": 0.3062993198704715, "grad_norm": 1.788910150527954, "learning_rate": 8.128128121117085e-06, "loss": 0.5842, "step": 7307 }, { "epoch": 0.30634123848548045, "grad_norm": 1.5655269622802734, "learning_rate": 8.127598503700701e-06, "loss": 0.5157, "step": 7308 }, { "epoch": 0.3063831571004894, "grad_norm": 2.003929376602173, "learning_rate": 8.127068828631364e-06, "loss": 0.553, "step": 7309 }, { "epoch": 0.3064250757154984, "grad_norm": 1.8396052122116089, "learning_rate": 8.126539095918835e-06, "loss": 0.4991, "step": 7310 }, { "epoch": 0.30646699433050734, "grad_norm": 1.7042073011398315, "learning_rate": 8.12600930557288e-06, "loss": 0.5015, "step": 7311 }, { "epoch": 0.3065089129455163, "grad_norm": 1.985711932182312, "learning_rate": 8.125479457603264e-06, "loss": 0.5111, "step": 7312 }, { "epoch": 0.30655083156052526, "grad_norm": 1.5960253477096558, "learning_rate": 8.124949552019757e-06, "loss": 0.4737, "step": 7313 }, { "epoch": 0.3065927501755342, "grad_norm": 2.3370542526245117, "learning_rate": 8.124419588832124e-06, "loss": 0.5604, "step": 7314 }, { "epoch": 0.3066346687905432, "grad_norm": 1.6627378463745117, "learning_rate": 8.123889568050136e-06, "loss": 0.5653, "step": 7315 }, { "epoch": 0.30667658740555215, "grad_norm": 1.624083161354065, "learning_rate": 8.123359489683562e-06, "loss": 0.4836, "step": 7316 }, { "epoch": 0.30671850602056105, "grad_norm": 1.7315489053726196, "learning_rate": 8.122829353742172e-06, "loss": 0.5762, "step": 7317 }, { "epoch": 0.30676042463557, "grad_norm": 1.870181679725647, "learning_rate": 8.122299160235742e-06, "loss": 0.5861, "step": 7318 }, { "epoch": 0.306802343250579, "grad_norm": 2.5059545040130615, "learning_rate": 8.121768909174044e-06, "loss": 0.5287, "step": 7319 }, { "epoch": 0.30684426186558794, "grad_norm": 1.5743427276611328, "learning_rate": 8.121238600566851e-06, "loss": 0.4905, "step": 7320 }, { "epoch": 0.3068861804805969, "grad_norm": 1.6012694835662842, "learning_rate": 8.12070823442394e-06, "loss": 0.5033, "step": 7321 }, { "epoch": 0.30692809909560587, "grad_norm": 1.379831075668335, "learning_rate": 8.120177810755086e-06, "loss": 0.4793, "step": 7322 }, { "epoch": 0.30697001771061483, "grad_norm": 1.5071125030517578, "learning_rate": 8.119647329570069e-06, "loss": 0.4599, "step": 7323 }, { "epoch": 0.3070119363256238, "grad_norm": 1.590709924697876, "learning_rate": 8.119116790878665e-06, "loss": 0.4934, "step": 7324 }, { "epoch": 0.30705385494063275, "grad_norm": 1.5596927404403687, "learning_rate": 8.118586194690654e-06, "loss": 0.4648, "step": 7325 }, { "epoch": 0.3070957735556417, "grad_norm": 1.4874119758605957, "learning_rate": 8.11805554101582e-06, "loss": 0.5022, "step": 7326 }, { "epoch": 0.3071376921706507, "grad_norm": 1.5055354833602905, "learning_rate": 8.11752482986394e-06, "loss": 0.5256, "step": 7327 }, { "epoch": 0.30717961078565964, "grad_norm": 1.4433175325393677, "learning_rate": 8.116994061244803e-06, "loss": 0.4603, "step": 7328 }, { "epoch": 0.3072215294006686, "grad_norm": 2.167447090148926, "learning_rate": 8.116463235168188e-06, "loss": 0.4754, "step": 7329 }, { "epoch": 0.30726344801567756, "grad_norm": 4.072625160217285, "learning_rate": 8.11593235164388e-06, "loss": 0.5446, "step": 7330 }, { "epoch": 0.3073053666306865, "grad_norm": 1.605191946029663, "learning_rate": 8.11540141068167e-06, "loss": 0.5183, "step": 7331 }, { "epoch": 0.3073472852456955, "grad_norm": 1.5182808637619019, "learning_rate": 8.114870412291341e-06, "loss": 0.5313, "step": 7332 }, { "epoch": 0.30738920386070445, "grad_norm": 1.6880640983581543, "learning_rate": 8.11433935648268e-06, "loss": 0.5163, "step": 7333 }, { "epoch": 0.3074311224757134, "grad_norm": 1.8730839490890503, "learning_rate": 8.11380824326548e-06, "loss": 0.5217, "step": 7334 }, { "epoch": 0.3074730410907224, "grad_norm": 1.6547539234161377, "learning_rate": 8.113277072649531e-06, "loss": 0.5589, "step": 7335 }, { "epoch": 0.30751495970573134, "grad_norm": 1.793989896774292, "learning_rate": 8.11274584464462e-06, "loss": 0.4885, "step": 7336 }, { "epoch": 0.3075568783207403, "grad_norm": 1.8687633275985718, "learning_rate": 8.112214559260546e-06, "loss": 0.5446, "step": 7337 }, { "epoch": 0.30759879693574926, "grad_norm": 1.3801666498184204, "learning_rate": 8.111683216507096e-06, "loss": 0.4662, "step": 7338 }, { "epoch": 0.3076407155507582, "grad_norm": 2.0831120014190674, "learning_rate": 8.11115181639407e-06, "loss": 0.5203, "step": 7339 }, { "epoch": 0.3076826341657672, "grad_norm": 1.690714955329895, "learning_rate": 8.11062035893126e-06, "loss": 0.5814, "step": 7340 }, { "epoch": 0.30772455278077615, "grad_norm": 2.1881167888641357, "learning_rate": 8.110088844128464e-06, "loss": 0.5114, "step": 7341 }, { "epoch": 0.30776647139578506, "grad_norm": 1.5733327865600586, "learning_rate": 8.10955727199548e-06, "loss": 0.4963, "step": 7342 }, { "epoch": 0.307808390010794, "grad_norm": 1.718052625656128, "learning_rate": 8.109025642542107e-06, "loss": 0.5331, "step": 7343 }, { "epoch": 0.307850308625803, "grad_norm": 2.174992084503174, "learning_rate": 8.108493955778143e-06, "loss": 0.4925, "step": 7344 }, { "epoch": 0.30789222724081194, "grad_norm": 1.5450316667556763, "learning_rate": 8.107962211713392e-06, "loss": 0.5703, "step": 7345 }, { "epoch": 0.3079341458558209, "grad_norm": 1.430412769317627, "learning_rate": 8.107430410357653e-06, "loss": 0.4754, "step": 7346 }, { "epoch": 0.30797606447082987, "grad_norm": 1.7470908164978027, "learning_rate": 8.10689855172073e-06, "loss": 0.5543, "step": 7347 }, { "epoch": 0.30801798308583883, "grad_norm": 1.7573918104171753, "learning_rate": 8.106366635812428e-06, "loss": 0.5043, "step": 7348 }, { "epoch": 0.3080599017008478, "grad_norm": 1.3901866674423218, "learning_rate": 8.105834662642553e-06, "loss": 0.4895, "step": 7349 }, { "epoch": 0.30810182031585676, "grad_norm": 1.5884677171707153, "learning_rate": 8.105302632220908e-06, "loss": 0.5192, "step": 7350 }, { "epoch": 0.3081437389308657, "grad_norm": 1.5212503671646118, "learning_rate": 8.104770544557303e-06, "loss": 0.5148, "step": 7351 }, { "epoch": 0.3081856575458747, "grad_norm": 1.5443413257598877, "learning_rate": 8.104238399661544e-06, "loss": 0.5179, "step": 7352 }, { "epoch": 0.30822757616088364, "grad_norm": 1.4354732036590576, "learning_rate": 8.103706197543441e-06, "loss": 0.5134, "step": 7353 }, { "epoch": 0.3082694947758926, "grad_norm": 1.8469232320785522, "learning_rate": 8.103173938212807e-06, "loss": 0.579, "step": 7354 }, { "epoch": 0.30831141339090157, "grad_norm": 1.4767564535140991, "learning_rate": 8.10264162167945e-06, "loss": 0.485, "step": 7355 }, { "epoch": 0.30835333200591053, "grad_norm": 2.30031418800354, "learning_rate": 8.102109247953184e-06, "loss": 0.5509, "step": 7356 }, { "epoch": 0.3083952506209195, "grad_norm": 1.6221436262130737, "learning_rate": 8.101576817043824e-06, "loss": 0.4852, "step": 7357 }, { "epoch": 0.30843716923592845, "grad_norm": 1.5279903411865234, "learning_rate": 8.101044328961182e-06, "loss": 0.5775, "step": 7358 }, { "epoch": 0.3084790878509374, "grad_norm": 1.82401442527771, "learning_rate": 8.100511783715076e-06, "loss": 0.4879, "step": 7359 }, { "epoch": 0.3085210064659464, "grad_norm": 2.691394567489624, "learning_rate": 8.09997918131532e-06, "loss": 0.48, "step": 7360 }, { "epoch": 0.30856292508095534, "grad_norm": 1.7331629991531372, "learning_rate": 8.099446521771735e-06, "loss": 0.5632, "step": 7361 }, { "epoch": 0.3086048436959643, "grad_norm": 1.5177135467529297, "learning_rate": 8.098913805094137e-06, "loss": 0.5326, "step": 7362 }, { "epoch": 0.30864676231097327, "grad_norm": 1.8524587154388428, "learning_rate": 8.098381031292346e-06, "loss": 0.5356, "step": 7363 }, { "epoch": 0.30868868092598223, "grad_norm": 1.5643068552017212, "learning_rate": 8.097848200376186e-06, "loss": 0.5149, "step": 7364 }, { "epoch": 0.3087305995409912, "grad_norm": 2.006350517272949, "learning_rate": 8.097315312355477e-06, "loss": 0.528, "step": 7365 }, { "epoch": 0.30877251815600015, "grad_norm": 1.625935673713684, "learning_rate": 8.096782367240043e-06, "loss": 0.5682, "step": 7366 }, { "epoch": 0.30881443677100906, "grad_norm": 1.6581597328186035, "learning_rate": 8.096249365039704e-06, "loss": 0.5745, "step": 7367 }, { "epoch": 0.308856355386018, "grad_norm": 1.7447412014007568, "learning_rate": 8.095716305764291e-06, "loss": 0.5875, "step": 7368 }, { "epoch": 0.308898274001027, "grad_norm": 1.8794710636138916, "learning_rate": 8.095183189423628e-06, "loss": 0.5444, "step": 7369 }, { "epoch": 0.30894019261603595, "grad_norm": 1.6992874145507812, "learning_rate": 8.09465001602754e-06, "loss": 0.4836, "step": 7370 }, { "epoch": 0.3089821112310449, "grad_norm": 1.4905693531036377, "learning_rate": 8.094116785585857e-06, "loss": 0.5468, "step": 7371 }, { "epoch": 0.30902402984605387, "grad_norm": 1.5071039199829102, "learning_rate": 8.093583498108411e-06, "loss": 0.5364, "step": 7372 }, { "epoch": 0.30906594846106283, "grad_norm": 1.726900339126587, "learning_rate": 8.093050153605028e-06, "loss": 0.4844, "step": 7373 }, { "epoch": 0.3091078670760718, "grad_norm": 1.658504843711853, "learning_rate": 8.092516752085543e-06, "loss": 0.5287, "step": 7374 }, { "epoch": 0.30914978569108076, "grad_norm": 1.5397905111312866, "learning_rate": 8.091983293559785e-06, "loss": 0.4928, "step": 7375 }, { "epoch": 0.3091917043060897, "grad_norm": 1.696029782295227, "learning_rate": 8.09144977803759e-06, "loss": 0.5208, "step": 7376 }, { "epoch": 0.3092336229210987, "grad_norm": 1.7035871744155884, "learning_rate": 8.090916205528792e-06, "loss": 0.5577, "step": 7377 }, { "epoch": 0.30927554153610765, "grad_norm": 1.4763661623001099, "learning_rate": 8.090382576043228e-06, "loss": 0.4958, "step": 7378 }, { "epoch": 0.3093174601511166, "grad_norm": 2.161125659942627, "learning_rate": 8.089848889590732e-06, "loss": 0.5185, "step": 7379 }, { "epoch": 0.30935937876612557, "grad_norm": 1.5807020664215088, "learning_rate": 8.089315146181144e-06, "loss": 0.5, "step": 7380 }, { "epoch": 0.30940129738113453, "grad_norm": 1.5654515027999878, "learning_rate": 8.088781345824304e-06, "loss": 0.5377, "step": 7381 }, { "epoch": 0.3094432159961435, "grad_norm": 2.1600334644317627, "learning_rate": 8.088247488530047e-06, "loss": 0.5365, "step": 7382 }, { "epoch": 0.30948513461115246, "grad_norm": 1.6190365552902222, "learning_rate": 8.087713574308217e-06, "loss": 0.5506, "step": 7383 }, { "epoch": 0.3095270532261614, "grad_norm": 1.7094154357910156, "learning_rate": 8.087179603168659e-06, "loss": 0.5423, "step": 7384 }, { "epoch": 0.3095689718411704, "grad_norm": 1.5341923236846924, "learning_rate": 8.08664557512121e-06, "loss": 0.5267, "step": 7385 }, { "epoch": 0.30961089045617934, "grad_norm": 1.4771208763122559, "learning_rate": 8.086111490175719e-06, "loss": 0.5409, "step": 7386 }, { "epoch": 0.3096528090711883, "grad_norm": 1.4897470474243164, "learning_rate": 8.085577348342028e-06, "loss": 0.5039, "step": 7387 }, { "epoch": 0.30969472768619727, "grad_norm": 1.5502103567123413, "learning_rate": 8.085043149629984e-06, "loss": 0.5364, "step": 7388 }, { "epoch": 0.30973664630120623, "grad_norm": 1.7017701864242554, "learning_rate": 8.084508894049436e-06, "loss": 0.4876, "step": 7389 }, { "epoch": 0.3097785649162152, "grad_norm": 1.555679202079773, "learning_rate": 8.08397458161023e-06, "loss": 0.5054, "step": 7390 }, { "epoch": 0.30982048353122416, "grad_norm": 1.5005334615707397, "learning_rate": 8.083440212322214e-06, "loss": 0.5571, "step": 7391 }, { "epoch": 0.30986240214623306, "grad_norm": 1.7223023176193237, "learning_rate": 8.082905786195243e-06, "loss": 0.524, "step": 7392 }, { "epoch": 0.309904320761242, "grad_norm": 1.5273535251617432, "learning_rate": 8.082371303239165e-06, "loss": 0.5604, "step": 7393 }, { "epoch": 0.309946239376251, "grad_norm": 1.5614731311798096, "learning_rate": 8.081836763463833e-06, "loss": 0.5207, "step": 7394 }, { "epoch": 0.30998815799125995, "grad_norm": 1.5706273317337036, "learning_rate": 8.081302166879101e-06, "loss": 0.5061, "step": 7395 }, { "epoch": 0.3100300766062689, "grad_norm": 1.4827855825424194, "learning_rate": 8.080767513494822e-06, "loss": 0.5135, "step": 7396 }, { "epoch": 0.3100719952212779, "grad_norm": 1.645719289779663, "learning_rate": 8.080232803320855e-06, "loss": 0.584, "step": 7397 }, { "epoch": 0.31011391383628684, "grad_norm": 1.7558029890060425, "learning_rate": 8.079698036367054e-06, "loss": 0.6316, "step": 7398 }, { "epoch": 0.3101558324512958, "grad_norm": 1.8085956573486328, "learning_rate": 8.079163212643277e-06, "loss": 0.5792, "step": 7399 }, { "epoch": 0.31019775106630476, "grad_norm": 1.6623766422271729, "learning_rate": 8.078628332159384e-06, "loss": 0.5059, "step": 7400 }, { "epoch": 0.3102396696813137, "grad_norm": 1.9097076654434204, "learning_rate": 8.078093394925233e-06, "loss": 0.4482, "step": 7401 }, { "epoch": 0.3102815882963227, "grad_norm": 1.5219465494155884, "learning_rate": 8.077558400950685e-06, "loss": 0.544, "step": 7402 }, { "epoch": 0.31032350691133165, "grad_norm": 1.5548359155654907, "learning_rate": 8.077023350245603e-06, "loss": 0.5412, "step": 7403 }, { "epoch": 0.3103654255263406, "grad_norm": 1.978348970413208, "learning_rate": 8.076488242819848e-06, "loss": 0.4869, "step": 7404 }, { "epoch": 0.3104073441413496, "grad_norm": 1.593103289604187, "learning_rate": 8.075953078683288e-06, "loss": 0.5631, "step": 7405 }, { "epoch": 0.31044926275635853, "grad_norm": 1.5912580490112305, "learning_rate": 8.075417857845784e-06, "loss": 0.6007, "step": 7406 }, { "epoch": 0.3104911813713675, "grad_norm": 4.22111701965332, "learning_rate": 8.074882580317202e-06, "loss": 0.5789, "step": 7407 }, { "epoch": 0.31053309998637646, "grad_norm": 1.6348930597305298, "learning_rate": 8.074347246107412e-06, "loss": 0.555, "step": 7408 }, { "epoch": 0.3105750186013854, "grad_norm": 1.311805009841919, "learning_rate": 8.07381185522628e-06, "loss": 0.481, "step": 7409 }, { "epoch": 0.3106169372163944, "grad_norm": 2.0734052658081055, "learning_rate": 8.073276407683678e-06, "loss": 0.5707, "step": 7410 }, { "epoch": 0.31065885583140335, "grad_norm": 1.9381017684936523, "learning_rate": 8.072740903489473e-06, "loss": 0.5337, "step": 7411 }, { "epoch": 0.3107007744464123, "grad_norm": 1.643236517906189, "learning_rate": 8.072205342653536e-06, "loss": 0.5357, "step": 7412 }, { "epoch": 0.31074269306142127, "grad_norm": 1.7071208953857422, "learning_rate": 8.071669725185743e-06, "loss": 0.5477, "step": 7413 }, { "epoch": 0.31078461167643023, "grad_norm": 1.5584304332733154, "learning_rate": 8.071134051095963e-06, "loss": 0.4765, "step": 7414 }, { "epoch": 0.3108265302914392, "grad_norm": 1.670676589012146, "learning_rate": 8.070598320394075e-06, "loss": 0.5299, "step": 7415 }, { "epoch": 0.31086844890644816, "grad_norm": 1.7427188158035278, "learning_rate": 8.070062533089949e-06, "loss": 0.5391, "step": 7416 }, { "epoch": 0.31091036752145707, "grad_norm": 1.6036865711212158, "learning_rate": 8.069526689193466e-06, "loss": 0.5288, "step": 7417 }, { "epoch": 0.310952286136466, "grad_norm": 1.787373661994934, "learning_rate": 8.068990788714501e-06, "loss": 0.5489, "step": 7418 }, { "epoch": 0.310994204751475, "grad_norm": 1.6159241199493408, "learning_rate": 8.068454831662933e-06, "loss": 0.501, "step": 7419 }, { "epoch": 0.31103612336648395, "grad_norm": 1.5909982919692993, "learning_rate": 8.067918818048644e-06, "loss": 0.5305, "step": 7420 }, { "epoch": 0.3110780419814929, "grad_norm": 1.6448757648468018, "learning_rate": 8.06738274788151e-06, "loss": 0.5901, "step": 7421 }, { "epoch": 0.3111199605965019, "grad_norm": 1.8709499835968018, "learning_rate": 8.06684662117142e-06, "loss": 0.5169, "step": 7422 }, { "epoch": 0.31116187921151084, "grad_norm": 1.8858842849731445, "learning_rate": 8.06631043792825e-06, "loss": 0.5221, "step": 7423 }, { "epoch": 0.3112037978265198, "grad_norm": 1.5359009504318237, "learning_rate": 8.065774198161886e-06, "loss": 0.5102, "step": 7424 }, { "epoch": 0.31124571644152876, "grad_norm": 2.1527180671691895, "learning_rate": 8.065237901882213e-06, "loss": 0.5649, "step": 7425 }, { "epoch": 0.3112876350565377, "grad_norm": 1.597425937652588, "learning_rate": 8.064701549099117e-06, "loss": 0.5622, "step": 7426 }, { "epoch": 0.3113295536715467, "grad_norm": 2.0352840423583984, "learning_rate": 8.064165139822486e-06, "loss": 0.5405, "step": 7427 }, { "epoch": 0.31137147228655565, "grad_norm": 1.8928972482681274, "learning_rate": 8.063628674062204e-06, "loss": 0.4991, "step": 7428 }, { "epoch": 0.3114133909015646, "grad_norm": 1.5602384805679321, "learning_rate": 8.063092151828164e-06, "loss": 0.4758, "step": 7429 }, { "epoch": 0.3114553095165736, "grad_norm": 1.6911725997924805, "learning_rate": 8.062555573130256e-06, "loss": 0.5712, "step": 7430 }, { "epoch": 0.31149722813158254, "grad_norm": 1.833552360534668, "learning_rate": 8.06201893797837e-06, "loss": 0.496, "step": 7431 }, { "epoch": 0.3115391467465915, "grad_norm": 1.7545775175094604, "learning_rate": 8.061482246382396e-06, "loss": 0.5319, "step": 7432 }, { "epoch": 0.31158106536160046, "grad_norm": 1.5328407287597656, "learning_rate": 8.06094549835223e-06, "loss": 0.5276, "step": 7433 }, { "epoch": 0.3116229839766094, "grad_norm": 1.474132776260376, "learning_rate": 8.060408693897766e-06, "loss": 0.5272, "step": 7434 }, { "epoch": 0.3116649025916184, "grad_norm": 2.245870590209961, "learning_rate": 8.059871833028897e-06, "loss": 0.5155, "step": 7435 }, { "epoch": 0.31170682120662735, "grad_norm": 1.6559802293777466, "learning_rate": 8.059334915755523e-06, "loss": 0.5107, "step": 7436 }, { "epoch": 0.3117487398216363, "grad_norm": 1.6727482080459595, "learning_rate": 8.058797942087537e-06, "loss": 0.5636, "step": 7437 }, { "epoch": 0.3117906584366453, "grad_norm": 1.706594705581665, "learning_rate": 8.058260912034841e-06, "loss": 0.5512, "step": 7438 }, { "epoch": 0.31183257705165424, "grad_norm": 1.7761366367340088, "learning_rate": 8.057723825607334e-06, "loss": 0.5874, "step": 7439 }, { "epoch": 0.3118744956666632, "grad_norm": 1.848207712173462, "learning_rate": 8.057186682814914e-06, "loss": 0.565, "step": 7440 }, { "epoch": 0.31191641428167216, "grad_norm": 1.5659843683242798, "learning_rate": 8.056649483667482e-06, "loss": 0.5051, "step": 7441 }, { "epoch": 0.31195833289668107, "grad_norm": 1.8199310302734375, "learning_rate": 8.056112228174947e-06, "loss": 0.5185, "step": 7442 }, { "epoch": 0.31200025151169003, "grad_norm": 1.5098810195922852, "learning_rate": 8.055574916347204e-06, "loss": 0.452, "step": 7443 }, { "epoch": 0.312042170126699, "grad_norm": 1.6172863245010376, "learning_rate": 8.055037548194165e-06, "loss": 0.547, "step": 7444 }, { "epoch": 0.31208408874170795, "grad_norm": 1.6967976093292236, "learning_rate": 8.054500123725729e-06, "loss": 0.5071, "step": 7445 }, { "epoch": 0.3121260073567169, "grad_norm": 1.755767822265625, "learning_rate": 8.053962642951808e-06, "loss": 0.5531, "step": 7446 }, { "epoch": 0.3121679259717259, "grad_norm": 2.072232723236084, "learning_rate": 8.053425105882307e-06, "loss": 0.5487, "step": 7447 }, { "epoch": 0.31220984458673484, "grad_norm": 1.693037748336792, "learning_rate": 8.052887512527136e-06, "loss": 0.554, "step": 7448 }, { "epoch": 0.3122517632017438, "grad_norm": 1.548862099647522, "learning_rate": 8.052349862896204e-06, "loss": 0.4913, "step": 7449 }, { "epoch": 0.31229368181675277, "grad_norm": 1.8326183557510376, "learning_rate": 8.051812156999421e-06, "loss": 0.512, "step": 7450 }, { "epoch": 0.31233560043176173, "grad_norm": 1.764431357383728, "learning_rate": 8.051274394846701e-06, "loss": 0.5703, "step": 7451 }, { "epoch": 0.3123775190467707, "grad_norm": 1.8912714719772339, "learning_rate": 8.050736576447955e-06, "loss": 0.5194, "step": 7452 }, { "epoch": 0.31241943766177965, "grad_norm": 1.52979576587677, "learning_rate": 8.050198701813098e-06, "loss": 0.5473, "step": 7453 }, { "epoch": 0.3124613562767886, "grad_norm": 1.5789419412612915, "learning_rate": 8.049660770952046e-06, "loss": 0.5332, "step": 7454 }, { "epoch": 0.3125032748917976, "grad_norm": 2.0477354526519775, "learning_rate": 8.049122783874712e-06, "loss": 0.6282, "step": 7455 }, { "epoch": 0.31254519350680654, "grad_norm": 1.762180209159851, "learning_rate": 8.048584740591015e-06, "loss": 0.5559, "step": 7456 }, { "epoch": 0.3125871121218155, "grad_norm": 1.8568534851074219, "learning_rate": 8.048046641110873e-06, "loss": 0.4837, "step": 7457 }, { "epoch": 0.31262903073682446, "grad_norm": 1.8205691576004028, "learning_rate": 8.047508485444207e-06, "loss": 0.5095, "step": 7458 }, { "epoch": 0.3126709493518334, "grad_norm": 1.713524580001831, "learning_rate": 8.046970273600931e-06, "loss": 0.5154, "step": 7459 }, { "epoch": 0.3127128679668424, "grad_norm": 1.5211290121078491, "learning_rate": 8.046432005590974e-06, "loss": 0.475, "step": 7460 }, { "epoch": 0.31275478658185135, "grad_norm": 1.5222605466842651, "learning_rate": 8.045893681424253e-06, "loss": 0.4899, "step": 7461 }, { "epoch": 0.3127967051968603, "grad_norm": 1.8505957126617432, "learning_rate": 8.045355301110692e-06, "loss": 0.5736, "step": 7462 }, { "epoch": 0.3128386238118693, "grad_norm": 2.114238977432251, "learning_rate": 8.044816864660218e-06, "loss": 0.5117, "step": 7463 }, { "epoch": 0.31288054242687824, "grad_norm": 1.734485149383545, "learning_rate": 8.044278372082753e-06, "loss": 0.5206, "step": 7464 }, { "epoch": 0.3129224610418872, "grad_norm": 1.7096188068389893, "learning_rate": 8.043739823388226e-06, "loss": 0.4946, "step": 7465 }, { "epoch": 0.31296437965689616, "grad_norm": 1.6204265356063843, "learning_rate": 8.043201218586563e-06, "loss": 0.5336, "step": 7466 }, { "epoch": 0.31300629827190507, "grad_norm": 1.847554326057434, "learning_rate": 8.042662557687691e-06, "loss": 0.587, "step": 7467 }, { "epoch": 0.31304821688691403, "grad_norm": 1.6349554061889648, "learning_rate": 8.042123840701542e-06, "loss": 0.5697, "step": 7468 }, { "epoch": 0.313090135501923, "grad_norm": 1.8518898487091064, "learning_rate": 8.041585067638047e-06, "loss": 0.4798, "step": 7469 }, { "epoch": 0.31313205411693196, "grad_norm": 1.6055060625076294, "learning_rate": 8.041046238507135e-06, "loss": 0.5154, "step": 7470 }, { "epoch": 0.3131739727319409, "grad_norm": 1.563845157623291, "learning_rate": 8.040507353318738e-06, "loss": 0.5372, "step": 7471 }, { "epoch": 0.3132158913469499, "grad_norm": 1.4625509977340698, "learning_rate": 8.039968412082794e-06, "loss": 0.4939, "step": 7472 }, { "epoch": 0.31325780996195884, "grad_norm": 1.985604166984558, "learning_rate": 8.039429414809235e-06, "loss": 0.5387, "step": 7473 }, { "epoch": 0.3132997285769678, "grad_norm": 1.548782467842102, "learning_rate": 8.038890361507993e-06, "loss": 0.5457, "step": 7474 }, { "epoch": 0.31334164719197677, "grad_norm": 1.5676733255386353, "learning_rate": 8.038351252189012e-06, "loss": 0.5301, "step": 7475 }, { "epoch": 0.31338356580698573, "grad_norm": 2.047081470489502, "learning_rate": 8.037812086862225e-06, "loss": 0.5433, "step": 7476 }, { "epoch": 0.3134254844219947, "grad_norm": 1.6652098894119263, "learning_rate": 8.037272865537573e-06, "loss": 0.5129, "step": 7477 }, { "epoch": 0.31346740303700366, "grad_norm": 1.4983571767807007, "learning_rate": 8.036733588224992e-06, "loss": 0.4586, "step": 7478 }, { "epoch": 0.3135093216520126, "grad_norm": 1.6546965837478638, "learning_rate": 8.036194254934428e-06, "loss": 0.5414, "step": 7479 }, { "epoch": 0.3135512402670216, "grad_norm": 1.6052043437957764, "learning_rate": 8.03565486567582e-06, "loss": 0.4895, "step": 7480 }, { "epoch": 0.31359315888203054, "grad_norm": 2.2176079750061035, "learning_rate": 8.03511542045911e-06, "loss": 0.5746, "step": 7481 }, { "epoch": 0.3136350774970395, "grad_norm": 1.817105770111084, "learning_rate": 8.034575919294244e-06, "loss": 0.559, "step": 7482 }, { "epoch": 0.31367699611204847, "grad_norm": 1.931793212890625, "learning_rate": 8.034036362191164e-06, "loss": 0.5898, "step": 7483 }, { "epoch": 0.31371891472705743, "grad_norm": 1.5392687320709229, "learning_rate": 8.03349674915982e-06, "loss": 0.5055, "step": 7484 }, { "epoch": 0.3137608333420664, "grad_norm": 1.5476876497268677, "learning_rate": 8.032957080210158e-06, "loss": 0.5253, "step": 7485 }, { "epoch": 0.31380275195707535, "grad_norm": 2.218508005142212, "learning_rate": 8.032417355352123e-06, "loss": 0.56, "step": 7486 }, { "epoch": 0.3138446705720843, "grad_norm": 1.7968382835388184, "learning_rate": 8.031877574595668e-06, "loss": 0.5176, "step": 7487 }, { "epoch": 0.3138865891870933, "grad_norm": 1.923137903213501, "learning_rate": 8.03133773795074e-06, "loss": 0.4984, "step": 7488 }, { "epoch": 0.31392850780210224, "grad_norm": 1.8406373262405396, "learning_rate": 8.030797845427293e-06, "loss": 0.608, "step": 7489 }, { "epoch": 0.3139704264171112, "grad_norm": 1.6356796026229858, "learning_rate": 8.030257897035276e-06, "loss": 0.5412, "step": 7490 }, { "epoch": 0.31401234503212017, "grad_norm": 1.8455697298049927, "learning_rate": 8.029717892784646e-06, "loss": 0.5863, "step": 7491 }, { "epoch": 0.3140542636471291, "grad_norm": 1.8185371160507202, "learning_rate": 8.029177832685352e-06, "loss": 0.5277, "step": 7492 }, { "epoch": 0.31409618226213804, "grad_norm": 1.7759078741073608, "learning_rate": 8.028637716747355e-06, "loss": 0.4881, "step": 7493 }, { "epoch": 0.314138100877147, "grad_norm": 1.6016323566436768, "learning_rate": 8.028097544980606e-06, "loss": 0.4919, "step": 7494 }, { "epoch": 0.31418001949215596, "grad_norm": 1.6897116899490356, "learning_rate": 8.02755731739507e-06, "loss": 0.5339, "step": 7495 }, { "epoch": 0.3142219381071649, "grad_norm": 1.754416584968567, "learning_rate": 8.027017034000695e-06, "loss": 0.5062, "step": 7496 }, { "epoch": 0.3142638567221739, "grad_norm": 1.6621123552322388, "learning_rate": 8.026476694807448e-06, "loss": 0.5253, "step": 7497 }, { "epoch": 0.31430577533718285, "grad_norm": 1.9631268978118896, "learning_rate": 8.025936299825288e-06, "loss": 0.5006, "step": 7498 }, { "epoch": 0.3143476939521918, "grad_norm": 1.6556965112686157, "learning_rate": 8.025395849064174e-06, "loss": 0.5646, "step": 7499 }, { "epoch": 0.31438961256720077, "grad_norm": 1.6339644193649292, "learning_rate": 8.02485534253407e-06, "loss": 0.4948, "step": 7500 }, { "epoch": 0.31443153118220973, "grad_norm": 1.721774935722351, "learning_rate": 8.02431478024494e-06, "loss": 0.5623, "step": 7501 }, { "epoch": 0.3144734497972187, "grad_norm": 1.7792714834213257, "learning_rate": 8.023774162206748e-06, "loss": 0.5179, "step": 7502 }, { "epoch": 0.31451536841222766, "grad_norm": 1.5609668493270874, "learning_rate": 8.02323348842946e-06, "loss": 0.5345, "step": 7503 }, { "epoch": 0.3145572870272366, "grad_norm": 1.6501749753952026, "learning_rate": 8.022692758923042e-06, "loss": 0.4937, "step": 7504 }, { "epoch": 0.3145992056422456, "grad_norm": 1.4911688566207886, "learning_rate": 8.022151973697462e-06, "loss": 0.5223, "step": 7505 }, { "epoch": 0.31464112425725455, "grad_norm": 1.576934576034546, "learning_rate": 8.021611132762687e-06, "loss": 0.5468, "step": 7506 }, { "epoch": 0.3146830428722635, "grad_norm": 1.684748888015747, "learning_rate": 8.021070236128689e-06, "loss": 0.536, "step": 7507 }, { "epoch": 0.31472496148727247, "grad_norm": 1.5606675148010254, "learning_rate": 8.020529283805438e-06, "loss": 0.5202, "step": 7508 }, { "epoch": 0.31476688010228143, "grad_norm": 1.6717357635498047, "learning_rate": 8.019988275802903e-06, "loss": 0.5577, "step": 7509 }, { "epoch": 0.3148087987172904, "grad_norm": 1.633732557296753, "learning_rate": 8.019447212131062e-06, "loss": 0.5427, "step": 7510 }, { "epoch": 0.31485071733229936, "grad_norm": 1.6942423582077026, "learning_rate": 8.018906092799885e-06, "loss": 0.5061, "step": 7511 }, { "epoch": 0.3148926359473083, "grad_norm": 1.6521265506744385, "learning_rate": 8.018364917819348e-06, "loss": 0.5099, "step": 7512 }, { "epoch": 0.3149345545623173, "grad_norm": 1.8053972721099854, "learning_rate": 8.017823687199427e-06, "loss": 0.4948, "step": 7513 }, { "epoch": 0.31497647317732624, "grad_norm": 1.7895458936691284, "learning_rate": 8.017282400950097e-06, "loss": 0.5202, "step": 7514 }, { "epoch": 0.3150183917923352, "grad_norm": 1.565816044807434, "learning_rate": 8.016741059081337e-06, "loss": 0.5535, "step": 7515 }, { "epoch": 0.31506031040734417, "grad_norm": 1.6284685134887695, "learning_rate": 8.016199661603127e-06, "loss": 0.5855, "step": 7516 }, { "epoch": 0.3151022290223531, "grad_norm": 1.4280774593353271, "learning_rate": 8.015658208525446e-06, "loss": 0.4874, "step": 7517 }, { "epoch": 0.31514414763736204, "grad_norm": 1.5601238012313843, "learning_rate": 8.015116699858273e-06, "loss": 0.4773, "step": 7518 }, { "epoch": 0.315186066252371, "grad_norm": 1.7046977281570435, "learning_rate": 8.014575135611595e-06, "loss": 0.5653, "step": 7519 }, { "epoch": 0.31522798486737996, "grad_norm": 1.7714864015579224, "learning_rate": 8.014033515795392e-06, "loss": 0.514, "step": 7520 }, { "epoch": 0.3152699034823889, "grad_norm": 1.9247748851776123, "learning_rate": 8.013491840419645e-06, "loss": 0.5325, "step": 7521 }, { "epoch": 0.3153118220973979, "grad_norm": 2.588034152984619, "learning_rate": 8.012950109494344e-06, "loss": 0.5172, "step": 7522 }, { "epoch": 0.31535374071240685, "grad_norm": 1.809876561164856, "learning_rate": 8.012408323029472e-06, "loss": 0.5013, "step": 7523 }, { "epoch": 0.3153956593274158, "grad_norm": 1.617050051689148, "learning_rate": 8.011866481035017e-06, "loss": 0.5771, "step": 7524 }, { "epoch": 0.3154375779424248, "grad_norm": 1.65779447555542, "learning_rate": 8.01132458352097e-06, "loss": 0.5626, "step": 7525 }, { "epoch": 0.31547949655743374, "grad_norm": 1.8200186491012573, "learning_rate": 8.010782630497313e-06, "loss": 0.5319, "step": 7526 }, { "epoch": 0.3155214151724427, "grad_norm": 3.246931314468384, "learning_rate": 8.010240621974043e-06, "loss": 0.5347, "step": 7527 }, { "epoch": 0.31556333378745166, "grad_norm": 1.730556845664978, "learning_rate": 8.009698557961148e-06, "loss": 0.5086, "step": 7528 }, { "epoch": 0.3156052524024606, "grad_norm": 1.5121647119522095, "learning_rate": 8.009156438468622e-06, "loss": 0.5317, "step": 7529 }, { "epoch": 0.3156471710174696, "grad_norm": 1.5091462135314941, "learning_rate": 8.008614263506456e-06, "loss": 0.5371, "step": 7530 }, { "epoch": 0.31568908963247855, "grad_norm": 2.5444600582122803, "learning_rate": 8.008072033084647e-06, "loss": 0.5342, "step": 7531 }, { "epoch": 0.3157310082474875, "grad_norm": 1.9645874500274658, "learning_rate": 8.007529747213186e-06, "loss": 0.5294, "step": 7532 }, { "epoch": 0.3157729268624965, "grad_norm": 1.7166273593902588, "learning_rate": 8.006987405902072e-06, "loss": 0.5132, "step": 7533 }, { "epoch": 0.31581484547750543, "grad_norm": 1.5297492742538452, "learning_rate": 8.006445009161305e-06, "loss": 0.5052, "step": 7534 }, { "epoch": 0.3158567640925144, "grad_norm": 2.0663158893585205, "learning_rate": 8.005902557000878e-06, "loss": 0.5295, "step": 7535 }, { "epoch": 0.31589868270752336, "grad_norm": 1.6481897830963135, "learning_rate": 8.005360049430793e-06, "loss": 0.5401, "step": 7536 }, { "epoch": 0.3159406013225323, "grad_norm": 2.1637258529663086, "learning_rate": 8.004817486461052e-06, "loss": 0.4712, "step": 7537 }, { "epoch": 0.3159825199375413, "grad_norm": 2.9817469120025635, "learning_rate": 8.004274868101653e-06, "loss": 0.5529, "step": 7538 }, { "epoch": 0.31602443855255025, "grad_norm": 1.5893336534500122, "learning_rate": 8.003732194362601e-06, "loss": 0.4761, "step": 7539 }, { "epoch": 0.3160663571675592, "grad_norm": 1.6941523551940918, "learning_rate": 8.003189465253898e-06, "loss": 0.5234, "step": 7540 }, { "epoch": 0.31610827578256817, "grad_norm": 1.7416934967041016, "learning_rate": 8.00264668078555e-06, "loss": 0.5356, "step": 7541 }, { "epoch": 0.3161501943975771, "grad_norm": 1.9054195880889893, "learning_rate": 8.00210384096756e-06, "loss": 0.5242, "step": 7542 }, { "epoch": 0.31619211301258604, "grad_norm": 1.7707890272140503, "learning_rate": 8.001560945809937e-06, "loss": 0.5284, "step": 7543 }, { "epoch": 0.316234031627595, "grad_norm": 1.9826836585998535, "learning_rate": 8.001017995322687e-06, "loss": 0.5569, "step": 7544 }, { "epoch": 0.31627595024260396, "grad_norm": 1.759731411933899, "learning_rate": 8.000474989515818e-06, "loss": 0.5064, "step": 7545 }, { "epoch": 0.3163178688576129, "grad_norm": 1.7313164472579956, "learning_rate": 7.999931928399343e-06, "loss": 0.5132, "step": 7546 }, { "epoch": 0.3163597874726219, "grad_norm": 2.026531219482422, "learning_rate": 7.999388811983268e-06, "loss": 0.5493, "step": 7547 }, { "epoch": 0.31640170608763085, "grad_norm": 1.9412060976028442, "learning_rate": 7.998845640277609e-06, "loss": 0.4992, "step": 7548 }, { "epoch": 0.3164436247026398, "grad_norm": 1.674590826034546, "learning_rate": 7.998302413292376e-06, "loss": 0.5242, "step": 7549 }, { "epoch": 0.3164855433176488, "grad_norm": 1.6617817878723145, "learning_rate": 7.997759131037582e-06, "loss": 0.5039, "step": 7550 }, { "epoch": 0.31652746193265774, "grad_norm": 1.897858738899231, "learning_rate": 7.997215793523243e-06, "loss": 0.5519, "step": 7551 }, { "epoch": 0.3165693805476667, "grad_norm": 2.084477424621582, "learning_rate": 7.996672400759373e-06, "loss": 0.5357, "step": 7552 }, { "epoch": 0.31661129916267566, "grad_norm": 1.695433259010315, "learning_rate": 7.996128952755993e-06, "loss": 0.516, "step": 7553 }, { "epoch": 0.3166532177776846, "grad_norm": 1.7100777626037598, "learning_rate": 7.995585449523115e-06, "loss": 0.5162, "step": 7554 }, { "epoch": 0.3166951363926936, "grad_norm": 1.7166560888290405, "learning_rate": 7.995041891070762e-06, "loss": 0.53, "step": 7555 }, { "epoch": 0.31673705500770255, "grad_norm": 1.6304333209991455, "learning_rate": 7.994498277408952e-06, "loss": 0.4998, "step": 7556 }, { "epoch": 0.3167789736227115, "grad_norm": 2.5481655597686768, "learning_rate": 7.993954608547706e-06, "loss": 0.5283, "step": 7557 }, { "epoch": 0.3168208922377205, "grad_norm": 1.7664427757263184, "learning_rate": 7.993410884497046e-06, "loss": 0.553, "step": 7558 }, { "epoch": 0.31686281085272944, "grad_norm": 1.5957238674163818, "learning_rate": 7.992867105266996e-06, "loss": 0.5613, "step": 7559 }, { "epoch": 0.3169047294677384, "grad_norm": 1.7900887727737427, "learning_rate": 7.992323270867575e-06, "loss": 0.5365, "step": 7560 }, { "epoch": 0.31694664808274736, "grad_norm": 1.5526028871536255, "learning_rate": 7.991779381308813e-06, "loss": 0.5091, "step": 7561 }, { "epoch": 0.3169885666977563, "grad_norm": 1.704928994178772, "learning_rate": 7.991235436600734e-06, "loss": 0.5419, "step": 7562 }, { "epoch": 0.3170304853127653, "grad_norm": 1.734326720237732, "learning_rate": 7.990691436753366e-06, "loss": 0.5405, "step": 7563 }, { "epoch": 0.31707240392777425, "grad_norm": 1.6792658567428589, "learning_rate": 7.990147381776733e-06, "loss": 0.5322, "step": 7564 }, { "epoch": 0.3171143225427832, "grad_norm": 1.8069061040878296, "learning_rate": 7.98960327168087e-06, "loss": 0.5219, "step": 7565 }, { "epoch": 0.3171562411577922, "grad_norm": 1.7961534261703491, "learning_rate": 7.989059106475803e-06, "loss": 0.5343, "step": 7566 }, { "epoch": 0.3171981597728011, "grad_norm": 1.6697170734405518, "learning_rate": 7.988514886171563e-06, "loss": 0.505, "step": 7567 }, { "epoch": 0.31724007838781004, "grad_norm": 1.886572003364563, "learning_rate": 7.987970610778182e-06, "loss": 0.5668, "step": 7568 }, { "epoch": 0.317281997002819, "grad_norm": 1.4698009490966797, "learning_rate": 7.987426280305695e-06, "loss": 0.4836, "step": 7569 }, { "epoch": 0.31732391561782797, "grad_norm": 1.8178459405899048, "learning_rate": 7.986881894764134e-06, "loss": 0.5925, "step": 7570 }, { "epoch": 0.31736583423283693, "grad_norm": 1.692115306854248, "learning_rate": 7.986337454163533e-06, "loss": 0.5159, "step": 7571 }, { "epoch": 0.3174077528478459, "grad_norm": 1.94568932056427, "learning_rate": 7.985792958513932e-06, "loss": 0.4674, "step": 7572 }, { "epoch": 0.31744967146285485, "grad_norm": 1.690126657485962, "learning_rate": 7.985248407825363e-06, "loss": 0.5034, "step": 7573 }, { "epoch": 0.3174915900778638, "grad_norm": 1.7347291707992554, "learning_rate": 7.98470380210787e-06, "loss": 0.4634, "step": 7574 }, { "epoch": 0.3175335086928728, "grad_norm": 1.7373842000961304, "learning_rate": 7.984159141371486e-06, "loss": 0.5181, "step": 7575 }, { "epoch": 0.31757542730788174, "grad_norm": 1.6199976205825806, "learning_rate": 7.983614425626254e-06, "loss": 0.5372, "step": 7576 }, { "epoch": 0.3176173459228907, "grad_norm": 1.9347635507583618, "learning_rate": 7.983069654882214e-06, "loss": 0.5162, "step": 7577 }, { "epoch": 0.31765926453789967, "grad_norm": 1.7471671104431152, "learning_rate": 7.982524829149411e-06, "loss": 0.5313, "step": 7578 }, { "epoch": 0.31770118315290863, "grad_norm": 1.8505668640136719, "learning_rate": 7.981979948437884e-06, "loss": 0.4801, "step": 7579 }, { "epoch": 0.3177431017679176, "grad_norm": 1.7323616743087769, "learning_rate": 7.981435012757682e-06, "loss": 0.4934, "step": 7580 }, { "epoch": 0.31778502038292655, "grad_norm": 1.7514712810516357, "learning_rate": 7.980890022118844e-06, "loss": 0.5318, "step": 7581 }, { "epoch": 0.3178269389979355, "grad_norm": 1.7678310871124268, "learning_rate": 7.98034497653142e-06, "loss": 0.5554, "step": 7582 }, { "epoch": 0.3178688576129445, "grad_norm": 1.7803468704223633, "learning_rate": 7.979799876005458e-06, "loss": 0.5288, "step": 7583 }, { "epoch": 0.31791077622795344, "grad_norm": 1.833464503288269, "learning_rate": 7.979254720551003e-06, "loss": 0.5016, "step": 7584 }, { "epoch": 0.3179526948429624, "grad_norm": 1.5117779970169067, "learning_rate": 7.978709510178107e-06, "loss": 0.513, "step": 7585 }, { "epoch": 0.31799461345797136, "grad_norm": 1.5576266050338745, "learning_rate": 7.97816424489682e-06, "loss": 0.5163, "step": 7586 }, { "epoch": 0.3180365320729803, "grad_norm": 2.1995673179626465, "learning_rate": 7.97761892471719e-06, "loss": 0.4758, "step": 7587 }, { "epoch": 0.3180784506879893, "grad_norm": 1.9564841985702515, "learning_rate": 7.977073549649274e-06, "loss": 0.5465, "step": 7588 }, { "epoch": 0.31812036930299825, "grad_norm": 1.7476590871810913, "learning_rate": 7.976528119703122e-06, "loss": 0.5667, "step": 7589 }, { "epoch": 0.3181622879180072, "grad_norm": 1.5896410942077637, "learning_rate": 7.975982634888787e-06, "loss": 0.5114, "step": 7590 }, { "epoch": 0.3182042065330162, "grad_norm": 1.59003746509552, "learning_rate": 7.97543709521633e-06, "loss": 0.5084, "step": 7591 }, { "epoch": 0.3182461251480251, "grad_norm": 1.587135672569275, "learning_rate": 7.974891500695801e-06, "loss": 0.5354, "step": 7592 }, { "epoch": 0.31828804376303405, "grad_norm": 1.6515055894851685, "learning_rate": 7.974345851337261e-06, "loss": 0.5202, "step": 7593 }, { "epoch": 0.318329962378043, "grad_norm": 1.5188994407653809, "learning_rate": 7.973800147150766e-06, "loss": 0.5248, "step": 7594 }, { "epoch": 0.31837188099305197, "grad_norm": 2.289735794067383, "learning_rate": 7.973254388146376e-06, "loss": 0.5404, "step": 7595 }, { "epoch": 0.31841379960806093, "grad_norm": 1.7577836513519287, "learning_rate": 7.972708574334154e-06, "loss": 0.5517, "step": 7596 }, { "epoch": 0.3184557182230699, "grad_norm": 1.8610984086990356, "learning_rate": 7.972162705724157e-06, "loss": 0.5564, "step": 7597 }, { "epoch": 0.31849763683807886, "grad_norm": 1.9155421257019043, "learning_rate": 7.971616782326451e-06, "loss": 0.4968, "step": 7598 }, { "epoch": 0.3185395554530878, "grad_norm": 1.8119350671768188, "learning_rate": 7.971070804151098e-06, "loss": 0.5531, "step": 7599 }, { "epoch": 0.3185814740680968, "grad_norm": 1.6525799036026, "learning_rate": 7.970524771208162e-06, "loss": 0.5609, "step": 7600 }, { "epoch": 0.31862339268310574, "grad_norm": 1.812963604927063, "learning_rate": 7.969978683507707e-06, "loss": 0.5689, "step": 7601 }, { "epoch": 0.3186653112981147, "grad_norm": 1.7844773530960083, "learning_rate": 7.969432541059801e-06, "loss": 0.531, "step": 7602 }, { "epoch": 0.31870722991312367, "grad_norm": 1.5556105375289917, "learning_rate": 7.968886343874512e-06, "loss": 0.5184, "step": 7603 }, { "epoch": 0.31874914852813263, "grad_norm": 1.7827445268630981, "learning_rate": 7.968340091961905e-06, "loss": 0.501, "step": 7604 }, { "epoch": 0.3187910671431416, "grad_norm": 1.742969036102295, "learning_rate": 7.967793785332056e-06, "loss": 0.512, "step": 7605 }, { "epoch": 0.31883298575815056, "grad_norm": 2.1305582523345947, "learning_rate": 7.967247423995031e-06, "loss": 0.5708, "step": 7606 }, { "epoch": 0.3188749043731595, "grad_norm": 1.7535637617111206, "learning_rate": 7.966701007960899e-06, "loss": 0.557, "step": 7607 }, { "epoch": 0.3189168229881685, "grad_norm": 2.0778579711914062, "learning_rate": 7.966154537239736e-06, "loss": 0.5105, "step": 7608 }, { "epoch": 0.31895874160317744, "grad_norm": 1.6896133422851562, "learning_rate": 7.965608011841616e-06, "loss": 0.5144, "step": 7609 }, { "epoch": 0.3190006602181864, "grad_norm": 1.8386281728744507, "learning_rate": 7.965061431776611e-06, "loss": 0.5767, "step": 7610 }, { "epoch": 0.31904257883319537, "grad_norm": 2.00642991065979, "learning_rate": 7.964514797054798e-06, "loss": 0.5333, "step": 7611 }, { "epoch": 0.31908449744820433, "grad_norm": 1.809165358543396, "learning_rate": 7.963968107686256e-06, "loss": 0.5667, "step": 7612 }, { "epoch": 0.3191264160632133, "grad_norm": 1.6419821977615356, "learning_rate": 7.963421363681055e-06, "loss": 0.5248, "step": 7613 }, { "epoch": 0.31916833467822225, "grad_norm": 2.293186902999878, "learning_rate": 7.96287456504928e-06, "loss": 0.4878, "step": 7614 }, { "epoch": 0.3192102532932312, "grad_norm": 1.6066032648086548, "learning_rate": 7.962327711801007e-06, "loss": 0.5223, "step": 7615 }, { "epoch": 0.3192521719082402, "grad_norm": 1.8637279272079468, "learning_rate": 7.96178080394632e-06, "loss": 0.5378, "step": 7616 }, { "epoch": 0.3192940905232491, "grad_norm": 1.798036813735962, "learning_rate": 7.961233841495298e-06, "loss": 0.5503, "step": 7617 }, { "epoch": 0.31933600913825805, "grad_norm": 1.670711874961853, "learning_rate": 7.960686824458023e-06, "loss": 0.5679, "step": 7618 }, { "epoch": 0.319377927753267, "grad_norm": 1.739951252937317, "learning_rate": 7.96013975284458e-06, "loss": 0.5021, "step": 7619 }, { "epoch": 0.319419846368276, "grad_norm": 1.5948318243026733, "learning_rate": 7.959592626665053e-06, "loss": 0.4837, "step": 7620 }, { "epoch": 0.31946176498328493, "grad_norm": 1.8104207515716553, "learning_rate": 7.959045445929527e-06, "loss": 0.551, "step": 7621 }, { "epoch": 0.3195036835982939, "grad_norm": 1.7251014709472656, "learning_rate": 7.95849821064809e-06, "loss": 0.5439, "step": 7622 }, { "epoch": 0.31954560221330286, "grad_norm": 1.7954157590866089, "learning_rate": 7.957950920830827e-06, "loss": 0.5531, "step": 7623 }, { "epoch": 0.3195875208283118, "grad_norm": 1.3822544813156128, "learning_rate": 7.957403576487828e-06, "loss": 0.5306, "step": 7624 }, { "epoch": 0.3196294394433208, "grad_norm": 1.945388674736023, "learning_rate": 7.956856177629183e-06, "loss": 0.5136, "step": 7625 }, { "epoch": 0.31967135805832975, "grad_norm": 1.659151554107666, "learning_rate": 7.956308724264984e-06, "loss": 0.5145, "step": 7626 }, { "epoch": 0.3197132766733387, "grad_norm": 1.5741363763809204, "learning_rate": 7.955761216405318e-06, "loss": 0.4745, "step": 7627 }, { "epoch": 0.31975519528834767, "grad_norm": 1.6327637434005737, "learning_rate": 7.955213654060282e-06, "loss": 0.5281, "step": 7628 }, { "epoch": 0.31979711390335663, "grad_norm": 1.694153904914856, "learning_rate": 7.954666037239967e-06, "loss": 0.53, "step": 7629 }, { "epoch": 0.3198390325183656, "grad_norm": 1.756512999534607, "learning_rate": 7.954118365954469e-06, "loss": 0.5759, "step": 7630 }, { "epoch": 0.31988095113337456, "grad_norm": 1.64560866355896, "learning_rate": 7.95357064021388e-06, "loss": 0.4777, "step": 7631 }, { "epoch": 0.3199228697483835, "grad_norm": 1.7848429679870605, "learning_rate": 7.953022860028301e-06, "loss": 0.5382, "step": 7632 }, { "epoch": 0.3199647883633925, "grad_norm": 10.79824161529541, "learning_rate": 7.95247502540783e-06, "loss": 0.5553, "step": 7633 }, { "epoch": 0.32000670697840145, "grad_norm": 2.0618698596954346, "learning_rate": 7.951927136362563e-06, "loss": 0.5163, "step": 7634 }, { "epoch": 0.3200486255934104, "grad_norm": 1.6657577753067017, "learning_rate": 7.951379192902598e-06, "loss": 0.5406, "step": 7635 }, { "epoch": 0.32009054420841937, "grad_norm": 1.8512192964553833, "learning_rate": 7.950831195038038e-06, "loss": 0.5708, "step": 7636 }, { "epoch": 0.32013246282342833, "grad_norm": 2.5541861057281494, "learning_rate": 7.950283142778987e-06, "loss": 0.529, "step": 7637 }, { "epoch": 0.3201743814384373, "grad_norm": 2.0355749130249023, "learning_rate": 7.949735036135542e-06, "loss": 0.5389, "step": 7638 }, { "epoch": 0.32021630005344626, "grad_norm": 2.538928508758545, "learning_rate": 7.949186875117812e-06, "loss": 0.5231, "step": 7639 }, { "epoch": 0.3202582186684552, "grad_norm": 1.8674317598342896, "learning_rate": 7.948638659735898e-06, "loss": 0.5398, "step": 7640 }, { "epoch": 0.3203001372834642, "grad_norm": 1.7352354526519775, "learning_rate": 7.948090389999905e-06, "loss": 0.4729, "step": 7641 }, { "epoch": 0.3203420558984731, "grad_norm": 1.747909665107727, "learning_rate": 7.947542065919944e-06, "loss": 0.5678, "step": 7642 }, { "epoch": 0.32038397451348205, "grad_norm": 1.7501572370529175, "learning_rate": 7.946993687506119e-06, "loss": 0.5703, "step": 7643 }, { "epoch": 0.320425893128491, "grad_norm": 1.75553560256958, "learning_rate": 7.946445254768537e-06, "loss": 0.5178, "step": 7644 }, { "epoch": 0.3204678117435, "grad_norm": 1.892507791519165, "learning_rate": 7.945896767717314e-06, "loss": 0.5804, "step": 7645 }, { "epoch": 0.32050973035850894, "grad_norm": 1.5037646293640137, "learning_rate": 7.945348226362555e-06, "loss": 0.5291, "step": 7646 }, { "epoch": 0.3205516489735179, "grad_norm": 2.017610549926758, "learning_rate": 7.944799630714373e-06, "loss": 0.5804, "step": 7647 }, { "epoch": 0.32059356758852686, "grad_norm": 3.567481517791748, "learning_rate": 7.944250980782881e-06, "loss": 0.5408, "step": 7648 }, { "epoch": 0.3206354862035358, "grad_norm": 1.6497092247009277, "learning_rate": 7.943702276578192e-06, "loss": 0.4773, "step": 7649 }, { "epoch": 0.3206774048185448, "grad_norm": 1.5666637420654297, "learning_rate": 7.943153518110422e-06, "loss": 0.5474, "step": 7650 }, { "epoch": 0.32071932343355375, "grad_norm": 1.5316921472549438, "learning_rate": 7.942604705389686e-06, "loss": 0.5642, "step": 7651 }, { "epoch": 0.3207612420485627, "grad_norm": 1.6482919454574585, "learning_rate": 7.9420558384261e-06, "loss": 0.5096, "step": 7652 }, { "epoch": 0.3208031606635717, "grad_norm": 1.7298429012298584, "learning_rate": 7.94150691722978e-06, "loss": 0.5569, "step": 7653 }, { "epoch": 0.32084507927858064, "grad_norm": 1.694891095161438, "learning_rate": 7.940957941810847e-06, "loss": 0.5111, "step": 7654 }, { "epoch": 0.3208869978935896, "grad_norm": 1.669197916984558, "learning_rate": 7.940408912179423e-06, "loss": 0.5397, "step": 7655 }, { "epoch": 0.32092891650859856, "grad_norm": 2.2897918224334717, "learning_rate": 7.939859828345621e-06, "loss": 0.5054, "step": 7656 }, { "epoch": 0.3209708351236075, "grad_norm": 1.6730101108551025, "learning_rate": 7.939310690319571e-06, "loss": 0.5149, "step": 7657 }, { "epoch": 0.3210127537386165, "grad_norm": 1.9633283615112305, "learning_rate": 7.938761498111392e-06, "loss": 0.5242, "step": 7658 }, { "epoch": 0.32105467235362545, "grad_norm": 1.6787240505218506, "learning_rate": 7.938212251731205e-06, "loss": 0.5563, "step": 7659 }, { "epoch": 0.3210965909686344, "grad_norm": 1.5296987295150757, "learning_rate": 7.93766295118914e-06, "loss": 0.5268, "step": 7660 }, { "epoch": 0.3211385095836434, "grad_norm": 1.7928863763809204, "learning_rate": 7.937113596495317e-06, "loss": 0.5222, "step": 7661 }, { "epoch": 0.32118042819865233, "grad_norm": 1.4482158422470093, "learning_rate": 7.936564187659868e-06, "loss": 0.4962, "step": 7662 }, { "epoch": 0.3212223468136613, "grad_norm": 1.564987301826477, "learning_rate": 7.936014724692916e-06, "loss": 0.5968, "step": 7663 }, { "epoch": 0.32126426542867026, "grad_norm": 1.7579227685928345, "learning_rate": 7.935465207604592e-06, "loss": 0.5249, "step": 7664 }, { "epoch": 0.3213061840436792, "grad_norm": 1.6678013801574707, "learning_rate": 7.934915636405024e-06, "loss": 0.5722, "step": 7665 }, { "epoch": 0.3213481026586882, "grad_norm": 1.9383323192596436, "learning_rate": 7.934366011104345e-06, "loss": 0.553, "step": 7666 }, { "epoch": 0.3213900212736971, "grad_norm": 1.9633827209472656, "learning_rate": 7.933816331712684e-06, "loss": 0.5551, "step": 7667 }, { "epoch": 0.32143193988870605, "grad_norm": 2.547851800918579, "learning_rate": 7.933266598240176e-06, "loss": 0.5333, "step": 7668 }, { "epoch": 0.321473858503715, "grad_norm": 1.7025984525680542, "learning_rate": 7.932716810696951e-06, "loss": 0.5188, "step": 7669 }, { "epoch": 0.321515777118724, "grad_norm": 1.7765132188796997, "learning_rate": 7.93216696909315e-06, "loss": 0.5096, "step": 7670 }, { "epoch": 0.32155769573373294, "grad_norm": 1.6737792491912842, "learning_rate": 7.931617073438902e-06, "loss": 0.4468, "step": 7671 }, { "epoch": 0.3215996143487419, "grad_norm": 1.7609155178070068, "learning_rate": 7.931067123744347e-06, "loss": 0.4673, "step": 7672 }, { "epoch": 0.32164153296375086, "grad_norm": 1.643344521522522, "learning_rate": 7.930517120019621e-06, "loss": 0.5403, "step": 7673 }, { "epoch": 0.3216834515787598, "grad_norm": 1.6582813262939453, "learning_rate": 7.929967062274862e-06, "loss": 0.5222, "step": 7674 }, { "epoch": 0.3217253701937688, "grad_norm": 1.6917930841445923, "learning_rate": 7.929416950520212e-06, "loss": 0.5116, "step": 7675 }, { "epoch": 0.32176728880877775, "grad_norm": 1.7873485088348389, "learning_rate": 7.92886678476581e-06, "loss": 0.5607, "step": 7676 }, { "epoch": 0.3218092074237867, "grad_norm": 1.814070701599121, "learning_rate": 7.928316565021798e-06, "loss": 0.5941, "step": 7677 }, { "epoch": 0.3218511260387957, "grad_norm": 1.571994423866272, "learning_rate": 7.92776629129832e-06, "loss": 0.48, "step": 7678 }, { "epoch": 0.32189304465380464, "grad_norm": 1.782887578010559, "learning_rate": 7.927215963605516e-06, "loss": 0.544, "step": 7679 }, { "epoch": 0.3219349632688136, "grad_norm": 1.6564754247665405, "learning_rate": 7.926665581953535e-06, "loss": 0.4663, "step": 7680 }, { "epoch": 0.32197688188382256, "grad_norm": 1.6291669607162476, "learning_rate": 7.926115146352517e-06, "loss": 0.5929, "step": 7681 }, { "epoch": 0.3220188004988315, "grad_norm": 1.7231569290161133, "learning_rate": 7.925564656812613e-06, "loss": 0.5512, "step": 7682 }, { "epoch": 0.3220607191138405, "grad_norm": 1.9579286575317383, "learning_rate": 7.925014113343968e-06, "loss": 0.4625, "step": 7683 }, { "epoch": 0.32210263772884945, "grad_norm": 1.784719705581665, "learning_rate": 7.924463515956732e-06, "loss": 0.5171, "step": 7684 }, { "epoch": 0.3221445563438584, "grad_norm": 2.0534844398498535, "learning_rate": 7.923912864661055e-06, "loss": 0.5463, "step": 7685 }, { "epoch": 0.3221864749588674, "grad_norm": 1.6939482688903809, "learning_rate": 7.923362159467085e-06, "loss": 0.5552, "step": 7686 }, { "epoch": 0.32222839357387634, "grad_norm": 1.7529176473617554, "learning_rate": 7.922811400384976e-06, "loss": 0.4877, "step": 7687 }, { "epoch": 0.3222703121888853, "grad_norm": 1.7177947759628296, "learning_rate": 7.92226058742488e-06, "loss": 0.5414, "step": 7688 }, { "epoch": 0.32231223080389426, "grad_norm": 1.5239231586456299, "learning_rate": 7.921709720596948e-06, "loss": 0.5294, "step": 7689 }, { "epoch": 0.3223541494189032, "grad_norm": 2.0624208450317383, "learning_rate": 7.921158799911337e-06, "loss": 0.5679, "step": 7690 }, { "epoch": 0.3223960680339122, "grad_norm": 3.934788227081299, "learning_rate": 7.920607825378204e-06, "loss": 0.5868, "step": 7691 }, { "epoch": 0.3224379866489211, "grad_norm": 1.710099697113037, "learning_rate": 7.920056797007702e-06, "loss": 0.4913, "step": 7692 }, { "epoch": 0.32247990526393006, "grad_norm": 1.8206084966659546, "learning_rate": 7.919505714809989e-06, "loss": 0.5614, "step": 7693 }, { "epoch": 0.322521823878939, "grad_norm": 1.7360973358154297, "learning_rate": 7.918954578795225e-06, "loss": 0.5531, "step": 7694 }, { "epoch": 0.322563742493948, "grad_norm": 1.5991789102554321, "learning_rate": 7.918403388973567e-06, "loss": 0.5765, "step": 7695 }, { "epoch": 0.32260566110895694, "grad_norm": 1.9444890022277832, "learning_rate": 7.91785214535518e-06, "loss": 0.4904, "step": 7696 }, { "epoch": 0.3226475797239659, "grad_norm": 1.7734798192977905, "learning_rate": 7.91730084795022e-06, "loss": 0.4882, "step": 7697 }, { "epoch": 0.32268949833897487, "grad_norm": 1.68620765209198, "learning_rate": 7.916749496768854e-06, "loss": 0.4989, "step": 7698 }, { "epoch": 0.32273141695398383, "grad_norm": 2.233309030532837, "learning_rate": 7.916198091821243e-06, "loss": 0.5669, "step": 7699 }, { "epoch": 0.3227733355689928, "grad_norm": 1.6922686100006104, "learning_rate": 7.915646633117551e-06, "loss": 0.5625, "step": 7700 }, { "epoch": 0.32281525418400175, "grad_norm": 2.1318793296813965, "learning_rate": 7.915095120667944e-06, "loss": 0.5295, "step": 7701 }, { "epoch": 0.3228571727990107, "grad_norm": 1.6293762922286987, "learning_rate": 7.91454355448259e-06, "loss": 0.5381, "step": 7702 }, { "epoch": 0.3228990914140197, "grad_norm": 1.8592931032180786, "learning_rate": 7.913991934571652e-06, "loss": 0.5302, "step": 7703 }, { "epoch": 0.32294101002902864, "grad_norm": 1.9290443658828735, "learning_rate": 7.913440260945303e-06, "loss": 0.5946, "step": 7704 }, { "epoch": 0.3229829286440376, "grad_norm": 1.6731153726577759, "learning_rate": 7.91288853361371e-06, "loss": 0.5506, "step": 7705 }, { "epoch": 0.32302484725904657, "grad_norm": 2.2534351348876953, "learning_rate": 7.912336752587045e-06, "loss": 0.5464, "step": 7706 }, { "epoch": 0.32306676587405553, "grad_norm": 1.6575454473495483, "learning_rate": 7.911784917875475e-06, "loss": 0.5208, "step": 7707 }, { "epoch": 0.3231086844890645, "grad_norm": 1.739306092262268, "learning_rate": 7.91123302948918e-06, "loss": 0.6062, "step": 7708 }, { "epoch": 0.32315060310407345, "grad_norm": 1.814096212387085, "learning_rate": 7.910681087438325e-06, "loss": 0.4989, "step": 7709 }, { "epoch": 0.3231925217190824, "grad_norm": 3.1481518745422363, "learning_rate": 7.91012909173309e-06, "loss": 0.5345, "step": 7710 }, { "epoch": 0.3232344403340914, "grad_norm": 2.685039520263672, "learning_rate": 7.909577042383647e-06, "loss": 0.5555, "step": 7711 }, { "epoch": 0.32327635894910034, "grad_norm": 1.8306032419204712, "learning_rate": 7.909024939400175e-06, "loss": 0.5757, "step": 7712 }, { "epoch": 0.3233182775641093, "grad_norm": 2.3590147495269775, "learning_rate": 7.90847278279285e-06, "loss": 0.5434, "step": 7713 }, { "epoch": 0.32336019617911826, "grad_norm": 1.7663408517837524, "learning_rate": 7.907920572571847e-06, "loss": 0.5756, "step": 7714 }, { "epoch": 0.3234021147941272, "grad_norm": 1.8494051694869995, "learning_rate": 7.907368308747352e-06, "loss": 0.591, "step": 7715 }, { "epoch": 0.3234440334091362, "grad_norm": 1.6019798517227173, "learning_rate": 7.90681599132954e-06, "loss": 0.5025, "step": 7716 }, { "epoch": 0.3234859520241451, "grad_norm": 1.7543989419937134, "learning_rate": 7.906263620328593e-06, "loss": 0.4912, "step": 7717 }, { "epoch": 0.32352787063915406, "grad_norm": 1.9603121280670166, "learning_rate": 7.905711195754696e-06, "loss": 0.5766, "step": 7718 }, { "epoch": 0.323569789254163, "grad_norm": 1.6850720643997192, "learning_rate": 7.90515871761803e-06, "loss": 0.5438, "step": 7719 }, { "epoch": 0.323611707869172, "grad_norm": 2.1100618839263916, "learning_rate": 7.904606185928777e-06, "loss": 0.5191, "step": 7720 }, { "epoch": 0.32365362648418095, "grad_norm": 1.7753117084503174, "learning_rate": 7.904053600697126e-06, "loss": 0.5965, "step": 7721 }, { "epoch": 0.3236955450991899, "grad_norm": 1.5258393287658691, "learning_rate": 7.903500961933261e-06, "loss": 0.5436, "step": 7722 }, { "epoch": 0.32373746371419887, "grad_norm": 1.8700700998306274, "learning_rate": 7.90294826964737e-06, "loss": 0.5653, "step": 7723 }, { "epoch": 0.32377938232920783, "grad_norm": 1.8237957954406738, "learning_rate": 7.902395523849641e-06, "loss": 0.5183, "step": 7724 }, { "epoch": 0.3238213009442168, "grad_norm": 1.566081166267395, "learning_rate": 7.901842724550263e-06, "loss": 0.5324, "step": 7725 }, { "epoch": 0.32386321955922576, "grad_norm": 2.0811312198638916, "learning_rate": 7.901289871759426e-06, "loss": 0.5563, "step": 7726 }, { "epoch": 0.3239051381742347, "grad_norm": 1.639236330986023, "learning_rate": 7.90073696548732e-06, "loss": 0.5755, "step": 7727 }, { "epoch": 0.3239470567892437, "grad_norm": 1.7296074628829956, "learning_rate": 7.900184005744141e-06, "loss": 0.5043, "step": 7728 }, { "epoch": 0.32398897540425264, "grad_norm": 1.639302134513855, "learning_rate": 7.899630992540077e-06, "loss": 0.5346, "step": 7729 }, { "epoch": 0.3240308940192616, "grad_norm": 1.6785917282104492, "learning_rate": 7.899077925885324e-06, "loss": 0.5055, "step": 7730 }, { "epoch": 0.32407281263427057, "grad_norm": 1.6282011270523071, "learning_rate": 7.898524805790078e-06, "loss": 0.5109, "step": 7731 }, { "epoch": 0.32411473124927953, "grad_norm": 1.8220489025115967, "learning_rate": 7.897971632264534e-06, "loss": 0.5226, "step": 7732 }, { "epoch": 0.3241566498642885, "grad_norm": 1.5487565994262695, "learning_rate": 7.897418405318892e-06, "loss": 0.4933, "step": 7733 }, { "epoch": 0.32419856847929746, "grad_norm": 1.8075008392333984, "learning_rate": 7.896865124963341e-06, "loss": 0.5986, "step": 7734 }, { "epoch": 0.3242404870943064, "grad_norm": 1.74941885471344, "learning_rate": 7.896311791208092e-06, "loss": 0.5066, "step": 7735 }, { "epoch": 0.3242824057093154, "grad_norm": 1.9167293310165405, "learning_rate": 7.895758404063337e-06, "loss": 0.5237, "step": 7736 }, { "epoch": 0.32432432432432434, "grad_norm": 1.815810203552246, "learning_rate": 7.895204963539281e-06, "loss": 0.5274, "step": 7737 }, { "epoch": 0.3243662429393333, "grad_norm": 1.7508361339569092, "learning_rate": 7.894651469646122e-06, "loss": 0.5787, "step": 7738 }, { "epoch": 0.32440816155434227, "grad_norm": 1.6154520511627197, "learning_rate": 7.894097922394065e-06, "loss": 0.5286, "step": 7739 }, { "epoch": 0.32445008016935123, "grad_norm": 1.8197615146636963, "learning_rate": 7.893544321793314e-06, "loss": 0.5449, "step": 7740 }, { "epoch": 0.3244919987843602, "grad_norm": 2.0971908569335938, "learning_rate": 7.892990667854072e-06, "loss": 0.5735, "step": 7741 }, { "epoch": 0.3245339173993691, "grad_norm": 1.6321121454238892, "learning_rate": 7.89243696058655e-06, "loss": 0.5137, "step": 7742 }, { "epoch": 0.32457583601437806, "grad_norm": 1.617301106452942, "learning_rate": 7.891883200000948e-06, "loss": 0.5596, "step": 7743 }, { "epoch": 0.324617754629387, "grad_norm": 2.0879313945770264, "learning_rate": 7.89132938610748e-06, "loss": 0.5059, "step": 7744 }, { "epoch": 0.324659673244396, "grad_norm": 1.731560468673706, "learning_rate": 7.89077551891635e-06, "loss": 0.5697, "step": 7745 }, { "epoch": 0.32470159185940495, "grad_norm": 1.6259639263153076, "learning_rate": 7.89022159843777e-06, "loss": 0.516, "step": 7746 }, { "epoch": 0.3247435104744139, "grad_norm": 1.755480170249939, "learning_rate": 7.88966762468195e-06, "loss": 0.5325, "step": 7747 }, { "epoch": 0.3247854290894229, "grad_norm": 1.7509499788284302, "learning_rate": 7.889113597659103e-06, "loss": 0.461, "step": 7748 }, { "epoch": 0.32482734770443183, "grad_norm": 1.7306666374206543, "learning_rate": 7.888559517379441e-06, "loss": 0.5732, "step": 7749 }, { "epoch": 0.3248692663194408, "grad_norm": 1.813716173171997, "learning_rate": 7.888005383853177e-06, "loss": 0.4611, "step": 7750 }, { "epoch": 0.32491118493444976, "grad_norm": 2.2150042057037354, "learning_rate": 7.887451197090526e-06, "loss": 0.5747, "step": 7751 }, { "epoch": 0.3249531035494587, "grad_norm": 6.265244007110596, "learning_rate": 7.886896957101705e-06, "loss": 0.5247, "step": 7752 }, { "epoch": 0.3249950221644677, "grad_norm": 1.5540151596069336, "learning_rate": 7.886342663896929e-06, "loss": 0.5395, "step": 7753 }, { "epoch": 0.32503694077947665, "grad_norm": 1.9506089687347412, "learning_rate": 7.885788317486416e-06, "loss": 0.5464, "step": 7754 }, { "epoch": 0.3250788593944856, "grad_norm": 1.6940128803253174, "learning_rate": 7.885233917880386e-06, "loss": 0.5568, "step": 7755 }, { "epoch": 0.32512077800949457, "grad_norm": 1.6627370119094849, "learning_rate": 7.884679465089056e-06, "loss": 0.5073, "step": 7756 }, { "epoch": 0.32516269662450353, "grad_norm": 2.464963436126709, "learning_rate": 7.88412495912265e-06, "loss": 0.5278, "step": 7757 }, { "epoch": 0.3252046152395125, "grad_norm": 2.073847770690918, "learning_rate": 7.883570399991385e-06, "loss": 0.5227, "step": 7758 }, { "epoch": 0.32524653385452146, "grad_norm": 1.6164829730987549, "learning_rate": 7.883015787705487e-06, "loss": 0.4509, "step": 7759 }, { "epoch": 0.3252884524695304, "grad_norm": 1.8023523092269897, "learning_rate": 7.88246112227518e-06, "loss": 0.5435, "step": 7760 }, { "epoch": 0.3253303710845394, "grad_norm": 1.8777159452438354, "learning_rate": 7.881906403710686e-06, "loss": 0.5184, "step": 7761 }, { "epoch": 0.32537228969954834, "grad_norm": 1.8204662799835205, "learning_rate": 7.881351632022232e-06, "loss": 0.4578, "step": 7762 }, { "epoch": 0.3254142083145573, "grad_norm": 1.5296326875686646, "learning_rate": 7.880796807220044e-06, "loss": 0.4865, "step": 7763 }, { "epoch": 0.32545612692956627, "grad_norm": 1.5838090181350708, "learning_rate": 7.880241929314348e-06, "loss": 0.5269, "step": 7764 }, { "epoch": 0.32549804554457523, "grad_norm": 1.7769681215286255, "learning_rate": 7.879686998315375e-06, "loss": 0.5376, "step": 7765 }, { "epoch": 0.3255399641595842, "grad_norm": 1.908338189125061, "learning_rate": 7.879132014233353e-06, "loss": 0.5751, "step": 7766 }, { "epoch": 0.3255818827745931, "grad_norm": 1.7064634561538696, "learning_rate": 7.878576977078512e-06, "loss": 0.5269, "step": 7767 }, { "epoch": 0.32562380138960206, "grad_norm": 1.5765436887741089, "learning_rate": 7.878021886861086e-06, "loss": 0.5596, "step": 7768 }, { "epoch": 0.325665720004611, "grad_norm": 1.8047629594802856, "learning_rate": 7.877466743591304e-06, "loss": 0.508, "step": 7769 }, { "epoch": 0.32570763861962, "grad_norm": 2.2143030166625977, "learning_rate": 7.8769115472794e-06, "loss": 0.5192, "step": 7770 }, { "epoch": 0.32574955723462895, "grad_norm": 1.6249762773513794, "learning_rate": 7.87635629793561e-06, "loss": 0.5172, "step": 7771 }, { "epoch": 0.3257914758496379, "grad_norm": 2.0472447872161865, "learning_rate": 7.875800995570166e-06, "loss": 0.5242, "step": 7772 }, { "epoch": 0.3258333944646469, "grad_norm": 1.8767549991607666, "learning_rate": 7.875245640193309e-06, "loss": 0.5533, "step": 7773 }, { "epoch": 0.32587531307965584, "grad_norm": 1.6003953218460083, "learning_rate": 7.874690231815273e-06, "loss": 0.5297, "step": 7774 }, { "epoch": 0.3259172316946648, "grad_norm": 2.1502463817596436, "learning_rate": 7.874134770446296e-06, "loss": 0.5306, "step": 7775 }, { "epoch": 0.32595915030967376, "grad_norm": 1.7392504215240479, "learning_rate": 7.873579256096618e-06, "loss": 0.5235, "step": 7776 }, { "epoch": 0.3260010689246827, "grad_norm": 2.3860788345336914, "learning_rate": 7.873023688776478e-06, "loss": 0.5681, "step": 7777 }, { "epoch": 0.3260429875396917, "grad_norm": 3.33986759185791, "learning_rate": 7.87246806849612e-06, "loss": 0.5041, "step": 7778 }, { "epoch": 0.32608490615470065, "grad_norm": 1.7144056558609009, "learning_rate": 7.871912395265781e-06, "loss": 0.5193, "step": 7779 }, { "epoch": 0.3261268247697096, "grad_norm": 1.8681879043579102, "learning_rate": 7.871356669095711e-06, "loss": 0.5029, "step": 7780 }, { "epoch": 0.3261687433847186, "grad_norm": 1.581695556640625, "learning_rate": 7.870800889996148e-06, "loss": 0.5054, "step": 7781 }, { "epoch": 0.32621066199972754, "grad_norm": 1.3207310438156128, "learning_rate": 7.870245057977341e-06, "loss": 0.4706, "step": 7782 }, { "epoch": 0.3262525806147365, "grad_norm": 1.6934574842453003, "learning_rate": 7.869689173049533e-06, "loss": 0.5074, "step": 7783 }, { "epoch": 0.32629449922974546, "grad_norm": 1.5662381649017334, "learning_rate": 7.869133235222973e-06, "loss": 0.5574, "step": 7784 }, { "epoch": 0.3263364178447544, "grad_norm": 1.5253905057907104, "learning_rate": 7.868577244507908e-06, "loss": 0.5133, "step": 7785 }, { "epoch": 0.3263783364597634, "grad_norm": 1.4640766382217407, "learning_rate": 7.868021200914588e-06, "loss": 0.4863, "step": 7786 }, { "epoch": 0.32642025507477235, "grad_norm": 1.902530312538147, "learning_rate": 7.867465104453262e-06, "loss": 0.4796, "step": 7787 }, { "epoch": 0.3264621736897813, "grad_norm": 1.670296549797058, "learning_rate": 7.86690895513418e-06, "loss": 0.5124, "step": 7788 }, { "epoch": 0.32650409230479027, "grad_norm": 1.6418256759643555, "learning_rate": 7.866352752967594e-06, "loss": 0.5442, "step": 7789 }, { "epoch": 0.32654601091979923, "grad_norm": 1.4987423419952393, "learning_rate": 7.86579649796376e-06, "loss": 0.4897, "step": 7790 }, { "epoch": 0.3265879295348082, "grad_norm": 1.6493014097213745, "learning_rate": 7.865240190132928e-06, "loss": 0.5274, "step": 7791 }, { "epoch": 0.3266298481498171, "grad_norm": 1.5208475589752197, "learning_rate": 7.864683829485353e-06, "loss": 0.5293, "step": 7792 }, { "epoch": 0.32667176676482607, "grad_norm": 1.5380579233169556, "learning_rate": 7.864127416031295e-06, "loss": 0.547, "step": 7793 }, { "epoch": 0.32671368537983503, "grad_norm": 1.9361482858657837, "learning_rate": 7.863570949781005e-06, "loss": 0.5678, "step": 7794 }, { "epoch": 0.326755603994844, "grad_norm": 1.6054972410202026, "learning_rate": 7.863014430744744e-06, "loss": 0.5593, "step": 7795 }, { "epoch": 0.32679752260985295, "grad_norm": 1.4845712184906006, "learning_rate": 7.862457858932771e-06, "loss": 0.505, "step": 7796 }, { "epoch": 0.3268394412248619, "grad_norm": 1.6563798189163208, "learning_rate": 7.861901234355344e-06, "loss": 0.5768, "step": 7797 }, { "epoch": 0.3268813598398709, "grad_norm": 1.7778005599975586, "learning_rate": 7.861344557022723e-06, "loss": 0.5199, "step": 7798 }, { "epoch": 0.32692327845487984, "grad_norm": 6.299492835998535, "learning_rate": 7.860787826945173e-06, "loss": 0.4828, "step": 7799 }, { "epoch": 0.3269651970698888, "grad_norm": 1.6190752983093262, "learning_rate": 7.860231044132954e-06, "loss": 0.5381, "step": 7800 }, { "epoch": 0.32700711568489776, "grad_norm": 1.520917296409607, "learning_rate": 7.859674208596328e-06, "loss": 0.4983, "step": 7801 }, { "epoch": 0.3270490342999067, "grad_norm": 1.7365633249282837, "learning_rate": 7.859117320345562e-06, "loss": 0.5104, "step": 7802 }, { "epoch": 0.3270909529149157, "grad_norm": 1.6763367652893066, "learning_rate": 7.858560379390921e-06, "loss": 0.5502, "step": 7803 }, { "epoch": 0.32713287152992465, "grad_norm": 1.934243083000183, "learning_rate": 7.858003385742672e-06, "loss": 0.5558, "step": 7804 }, { "epoch": 0.3271747901449336, "grad_norm": 1.84196937084198, "learning_rate": 7.85744633941108e-06, "loss": 0.4813, "step": 7805 }, { "epoch": 0.3272167087599426, "grad_norm": 1.701556921005249, "learning_rate": 7.856889240406417e-06, "loss": 0.4963, "step": 7806 }, { "epoch": 0.32725862737495154, "grad_norm": 1.7383267879486084, "learning_rate": 7.856332088738948e-06, "loss": 0.5285, "step": 7807 }, { "epoch": 0.3273005459899605, "grad_norm": 1.7802610397338867, "learning_rate": 7.855774884418948e-06, "loss": 0.5115, "step": 7808 }, { "epoch": 0.32734246460496946, "grad_norm": 2.368014335632324, "learning_rate": 7.855217627456687e-06, "loss": 0.5272, "step": 7809 }, { "epoch": 0.3273843832199784, "grad_norm": 1.567126750946045, "learning_rate": 7.854660317862432e-06, "loss": 0.5449, "step": 7810 }, { "epoch": 0.3274263018349874, "grad_norm": 1.7520217895507812, "learning_rate": 7.854102955646463e-06, "loss": 0.5541, "step": 7811 }, { "epoch": 0.32746822044999635, "grad_norm": 1.705245852470398, "learning_rate": 7.853545540819053e-06, "loss": 0.5393, "step": 7812 }, { "epoch": 0.3275101390650053, "grad_norm": 2.4439992904663086, "learning_rate": 7.852988073390473e-06, "loss": 0.5468, "step": 7813 }, { "epoch": 0.3275520576800143, "grad_norm": 1.7060658931732178, "learning_rate": 7.852430553371004e-06, "loss": 0.5433, "step": 7814 }, { "epoch": 0.32759397629502324, "grad_norm": 1.6609151363372803, "learning_rate": 7.85187298077092e-06, "loss": 0.5367, "step": 7815 }, { "epoch": 0.3276358949100322, "grad_norm": 1.5030407905578613, "learning_rate": 7.851315355600501e-06, "loss": 0.5153, "step": 7816 }, { "epoch": 0.3276778135250411, "grad_norm": 1.7136567831039429, "learning_rate": 7.850757677870024e-06, "loss": 0.516, "step": 7817 }, { "epoch": 0.32771973214005007, "grad_norm": 1.7639645338058472, "learning_rate": 7.850199947589774e-06, "loss": 0.5174, "step": 7818 }, { "epoch": 0.32776165075505903, "grad_norm": 1.707804799079895, "learning_rate": 7.849642164770024e-06, "loss": 0.5223, "step": 7819 }, { "epoch": 0.327803569370068, "grad_norm": 1.6727067232131958, "learning_rate": 7.849084329421061e-06, "loss": 0.5121, "step": 7820 }, { "epoch": 0.32784548798507696, "grad_norm": 1.5326234102249146, "learning_rate": 7.848526441553168e-06, "loss": 0.5005, "step": 7821 }, { "epoch": 0.3278874066000859, "grad_norm": 1.7034075260162354, "learning_rate": 7.847968501176628e-06, "loss": 0.4904, "step": 7822 }, { "epoch": 0.3279293252150949, "grad_norm": 1.7597788572311401, "learning_rate": 7.847410508301727e-06, "loss": 0.5245, "step": 7823 }, { "epoch": 0.32797124383010384, "grad_norm": 1.5546437501907349, "learning_rate": 7.846852462938747e-06, "loss": 0.5221, "step": 7824 }, { "epoch": 0.3280131624451128, "grad_norm": 1.6995106935501099, "learning_rate": 7.84629436509798e-06, "loss": 0.5154, "step": 7825 }, { "epoch": 0.32805508106012177, "grad_norm": 2.047633171081543, "learning_rate": 7.84573621478971e-06, "loss": 0.5269, "step": 7826 }, { "epoch": 0.32809699967513073, "grad_norm": 1.5904494524002075, "learning_rate": 7.845178012024228e-06, "loss": 0.5858, "step": 7827 }, { "epoch": 0.3281389182901397, "grad_norm": 1.6256937980651855, "learning_rate": 7.84461975681182e-06, "loss": 0.5162, "step": 7828 }, { "epoch": 0.32818083690514865, "grad_norm": 1.5664360523223877, "learning_rate": 7.844061449162781e-06, "loss": 0.5108, "step": 7829 }, { "epoch": 0.3282227555201576, "grad_norm": 1.7855278253555298, "learning_rate": 7.843503089087404e-06, "loss": 0.5355, "step": 7830 }, { "epoch": 0.3282646741351666, "grad_norm": 1.6213963031768799, "learning_rate": 7.842944676595974e-06, "loss": 0.4956, "step": 7831 }, { "epoch": 0.32830659275017554, "grad_norm": 1.638439655303955, "learning_rate": 7.842386211698792e-06, "loss": 0.5356, "step": 7832 }, { "epoch": 0.3283485113651845, "grad_norm": 1.5318535566329956, "learning_rate": 7.841827694406147e-06, "loss": 0.568, "step": 7833 }, { "epoch": 0.32839042998019347, "grad_norm": 1.6651350259780884, "learning_rate": 7.84126912472834e-06, "loss": 0.5016, "step": 7834 }, { "epoch": 0.32843234859520243, "grad_norm": 2.426368474960327, "learning_rate": 7.840710502675663e-06, "loss": 0.5419, "step": 7835 }, { "epoch": 0.3284742672102114, "grad_norm": 1.6501706838607788, "learning_rate": 7.840151828258415e-06, "loss": 0.4957, "step": 7836 }, { "epoch": 0.32851618582522035, "grad_norm": 1.7713350057601929, "learning_rate": 7.839593101486895e-06, "loss": 0.5012, "step": 7837 }, { "epoch": 0.3285581044402293, "grad_norm": 1.7166519165039062, "learning_rate": 7.839034322371402e-06, "loss": 0.5731, "step": 7838 }, { "epoch": 0.3286000230552383, "grad_norm": 1.453176736831665, "learning_rate": 7.838475490922235e-06, "loss": 0.4764, "step": 7839 }, { "epoch": 0.32864194167024724, "grad_norm": 1.8515337705612183, "learning_rate": 7.837916607149698e-06, "loss": 0.5647, "step": 7840 }, { "epoch": 0.3286838602852562, "grad_norm": 1.9195585250854492, "learning_rate": 7.83735767106409e-06, "loss": 0.5976, "step": 7841 }, { "epoch": 0.3287257789002651, "grad_norm": 1.5785751342773438, "learning_rate": 7.836798682675716e-06, "loss": 0.4665, "step": 7842 }, { "epoch": 0.32876769751527407, "grad_norm": 1.4277437925338745, "learning_rate": 7.836239641994883e-06, "loss": 0.5313, "step": 7843 }, { "epoch": 0.32880961613028303, "grad_norm": 1.6703051328659058, "learning_rate": 7.83568054903189e-06, "loss": 0.5665, "step": 7844 }, { "epoch": 0.328851534745292, "grad_norm": 1.718860387802124, "learning_rate": 7.835121403797047e-06, "loss": 0.513, "step": 7845 }, { "epoch": 0.32889345336030096, "grad_norm": 1.4241788387298584, "learning_rate": 7.834562206300662e-06, "loss": 0.4661, "step": 7846 }, { "epoch": 0.3289353719753099, "grad_norm": 1.5523438453674316, "learning_rate": 7.834002956553039e-06, "loss": 0.5199, "step": 7847 }, { "epoch": 0.3289772905903189, "grad_norm": 1.4772552251815796, "learning_rate": 7.83344365456449e-06, "loss": 0.542, "step": 7848 }, { "epoch": 0.32901920920532784, "grad_norm": 2.286653518676758, "learning_rate": 7.832884300345325e-06, "loss": 0.5631, "step": 7849 }, { "epoch": 0.3290611278203368, "grad_norm": 1.5166579484939575, "learning_rate": 7.832324893905853e-06, "loss": 0.5326, "step": 7850 }, { "epoch": 0.32910304643534577, "grad_norm": 1.6737717390060425, "learning_rate": 7.831765435256389e-06, "loss": 0.5352, "step": 7851 }, { "epoch": 0.32914496505035473, "grad_norm": 1.7856788635253906, "learning_rate": 7.831205924407243e-06, "loss": 0.5649, "step": 7852 }, { "epoch": 0.3291868836653637, "grad_norm": 1.8167301416397095, "learning_rate": 7.83064636136873e-06, "loss": 0.5398, "step": 7853 }, { "epoch": 0.32922880228037266, "grad_norm": 1.863847255706787, "learning_rate": 7.830086746151164e-06, "loss": 0.6269, "step": 7854 }, { "epoch": 0.3292707208953816, "grad_norm": 1.669475793838501, "learning_rate": 7.829527078764863e-06, "loss": 0.5635, "step": 7855 }, { "epoch": 0.3293126395103906, "grad_norm": 1.619794487953186, "learning_rate": 7.828967359220141e-06, "loss": 0.5706, "step": 7856 }, { "epoch": 0.32935455812539954, "grad_norm": 1.4939967393875122, "learning_rate": 7.828407587527316e-06, "loss": 0.5662, "step": 7857 }, { "epoch": 0.3293964767404085, "grad_norm": 1.796439528465271, "learning_rate": 7.827847763696708e-06, "loss": 0.5281, "step": 7858 }, { "epoch": 0.32943839535541747, "grad_norm": 2.61531925201416, "learning_rate": 7.827287887738635e-06, "loss": 0.557, "step": 7859 }, { "epoch": 0.32948031397042643, "grad_norm": 1.5651845932006836, "learning_rate": 7.82672795966342e-06, "loss": 0.5573, "step": 7860 }, { "epoch": 0.3295222325854354, "grad_norm": 1.9476560354232788, "learning_rate": 7.826167979481381e-06, "loss": 0.4809, "step": 7861 }, { "epoch": 0.32956415120044436, "grad_norm": 1.6674667596817017, "learning_rate": 7.825607947202844e-06, "loss": 0.5845, "step": 7862 }, { "epoch": 0.3296060698154533, "grad_norm": 1.6073428392410278, "learning_rate": 7.825047862838129e-06, "loss": 0.5247, "step": 7863 }, { "epoch": 0.3296479884304623, "grad_norm": 1.5653268098831177, "learning_rate": 7.824487726397565e-06, "loss": 0.5727, "step": 7864 }, { "epoch": 0.32968990704547124, "grad_norm": 1.6753641366958618, "learning_rate": 7.823927537891472e-06, "loss": 0.545, "step": 7865 }, { "epoch": 0.3297318256604802, "grad_norm": 1.6416407823562622, "learning_rate": 7.823367297330179e-06, "loss": 0.5676, "step": 7866 }, { "epoch": 0.3297737442754891, "grad_norm": 1.7809624671936035, "learning_rate": 7.822807004724013e-06, "loss": 0.5315, "step": 7867 }, { "epoch": 0.3298156628904981, "grad_norm": 1.5393242835998535, "learning_rate": 7.822246660083303e-06, "loss": 0.5127, "step": 7868 }, { "epoch": 0.32985758150550704, "grad_norm": 1.439406394958496, "learning_rate": 7.821686263418378e-06, "loss": 0.4403, "step": 7869 }, { "epoch": 0.329899500120516, "grad_norm": 1.6480209827423096, "learning_rate": 7.821125814739568e-06, "loss": 0.5832, "step": 7870 }, { "epoch": 0.32994141873552496, "grad_norm": 1.6140925884246826, "learning_rate": 7.820565314057201e-06, "loss": 0.5365, "step": 7871 }, { "epoch": 0.3299833373505339, "grad_norm": 1.506419062614441, "learning_rate": 7.820004761381615e-06, "loss": 0.4929, "step": 7872 }, { "epoch": 0.3300252559655429, "grad_norm": 1.9840878248214722, "learning_rate": 7.819444156723137e-06, "loss": 0.5308, "step": 7873 }, { "epoch": 0.33006717458055185, "grad_norm": 1.8674066066741943, "learning_rate": 7.818883500092105e-06, "loss": 0.502, "step": 7874 }, { "epoch": 0.3301090931955608, "grad_norm": 1.8352993726730347, "learning_rate": 7.818322791498851e-06, "loss": 0.4716, "step": 7875 }, { "epoch": 0.33015101181056977, "grad_norm": 2.523946762084961, "learning_rate": 7.817762030953716e-06, "loss": 0.5434, "step": 7876 }, { "epoch": 0.33019293042557873, "grad_norm": 1.52734375, "learning_rate": 7.81720121846703e-06, "loss": 0.5231, "step": 7877 }, { "epoch": 0.3302348490405877, "grad_norm": 1.5780452489852905, "learning_rate": 7.816640354049137e-06, "loss": 0.4913, "step": 7878 }, { "epoch": 0.33027676765559666, "grad_norm": 2.0580039024353027, "learning_rate": 7.816079437710371e-06, "loss": 0.5518, "step": 7879 }, { "epoch": 0.3303186862706056, "grad_norm": 2.0675058364868164, "learning_rate": 7.815518469461075e-06, "loss": 0.5348, "step": 7880 }, { "epoch": 0.3303606048856146, "grad_norm": 1.7095156908035278, "learning_rate": 7.814957449311586e-06, "loss": 0.5692, "step": 7881 }, { "epoch": 0.33040252350062355, "grad_norm": 1.5390651226043701, "learning_rate": 7.81439637727225e-06, "loss": 0.5423, "step": 7882 }, { "epoch": 0.3304444421156325, "grad_norm": 2.231079578399658, "learning_rate": 7.813835253353408e-06, "loss": 0.5164, "step": 7883 }, { "epoch": 0.33048636073064147, "grad_norm": 1.4606165885925293, "learning_rate": 7.813274077565403e-06, "loss": 0.4946, "step": 7884 }, { "epoch": 0.33052827934565043, "grad_norm": 1.775786280632019, "learning_rate": 7.812712849918579e-06, "loss": 0.5153, "step": 7885 }, { "epoch": 0.3305701979606594, "grad_norm": 1.5743178129196167, "learning_rate": 7.812151570423282e-06, "loss": 0.4912, "step": 7886 }, { "epoch": 0.33061211657566836, "grad_norm": 1.6248793601989746, "learning_rate": 7.81159023908986e-06, "loss": 0.464, "step": 7887 }, { "epoch": 0.3306540351906773, "grad_norm": 1.5548896789550781, "learning_rate": 7.811028855928655e-06, "loss": 0.5173, "step": 7888 }, { "epoch": 0.3306959538056863, "grad_norm": 2.5811195373535156, "learning_rate": 7.810467420950022e-06, "loss": 0.5401, "step": 7889 }, { "epoch": 0.33073787242069524, "grad_norm": 1.646821141242981, "learning_rate": 7.809905934164308e-06, "loss": 0.5409, "step": 7890 }, { "epoch": 0.3307797910357042, "grad_norm": 1.4613009691238403, "learning_rate": 7.809344395581861e-06, "loss": 0.5063, "step": 7891 }, { "epoch": 0.3308217096507131, "grad_norm": 1.5214864015579224, "learning_rate": 7.808782805213035e-06, "loss": 0.5343, "step": 7892 }, { "epoch": 0.3308636282657221, "grad_norm": 1.5974109172821045, "learning_rate": 7.808221163068181e-06, "loss": 0.5167, "step": 7893 }, { "epoch": 0.33090554688073104, "grad_norm": 1.6965022087097168, "learning_rate": 7.80765946915765e-06, "loss": 0.5395, "step": 7894 }, { "epoch": 0.33094746549574, "grad_norm": 1.8594473600387573, "learning_rate": 7.807097723491801e-06, "loss": 0.5367, "step": 7895 }, { "epoch": 0.33098938411074896, "grad_norm": 1.5090216398239136, "learning_rate": 7.806535926080985e-06, "loss": 0.5289, "step": 7896 }, { "epoch": 0.3310313027257579, "grad_norm": 1.62851083278656, "learning_rate": 7.805974076935559e-06, "loss": 0.5425, "step": 7897 }, { "epoch": 0.3310732213407669, "grad_norm": 1.6468226909637451, "learning_rate": 7.805412176065879e-06, "loss": 0.5052, "step": 7898 }, { "epoch": 0.33111513995577585, "grad_norm": 1.5304230451583862, "learning_rate": 7.804850223482305e-06, "loss": 0.5464, "step": 7899 }, { "epoch": 0.3311570585707848, "grad_norm": 1.6205581426620483, "learning_rate": 7.804288219195196e-06, "loss": 0.5755, "step": 7900 }, { "epoch": 0.3311989771857938, "grad_norm": 1.7734581232070923, "learning_rate": 7.803726163214908e-06, "loss": 0.5741, "step": 7901 }, { "epoch": 0.33124089580080274, "grad_norm": 1.4527050256729126, "learning_rate": 7.803164055551807e-06, "loss": 0.5155, "step": 7902 }, { "epoch": 0.3312828144158117, "grad_norm": 1.6450275182724, "learning_rate": 7.802601896216249e-06, "loss": 0.5289, "step": 7903 }, { "epoch": 0.33132473303082066, "grad_norm": 1.7195380926132202, "learning_rate": 7.802039685218601e-06, "loss": 0.5372, "step": 7904 }, { "epoch": 0.3313666516458296, "grad_norm": 1.6295106410980225, "learning_rate": 7.801477422569225e-06, "loss": 0.5227, "step": 7905 }, { "epoch": 0.3314085702608386, "grad_norm": 1.427587866783142, "learning_rate": 7.800915108278486e-06, "loss": 0.496, "step": 7906 }, { "epoch": 0.33145048887584755, "grad_norm": 1.7346699237823486, "learning_rate": 7.800352742356749e-06, "loss": 0.506, "step": 7907 }, { "epoch": 0.3314924074908565, "grad_norm": 1.6757677793502808, "learning_rate": 7.799790324814381e-06, "loss": 0.5345, "step": 7908 }, { "epoch": 0.3315343261058655, "grad_norm": 1.494993805885315, "learning_rate": 7.799227855661747e-06, "loss": 0.5062, "step": 7909 }, { "epoch": 0.33157624472087444, "grad_norm": 1.5438787937164307, "learning_rate": 7.798665334909219e-06, "loss": 0.5715, "step": 7910 }, { "epoch": 0.3316181633358834, "grad_norm": 1.5302423238754272, "learning_rate": 7.798102762567165e-06, "loss": 0.5165, "step": 7911 }, { "epoch": 0.33166008195089236, "grad_norm": 1.6810662746429443, "learning_rate": 7.797540138645953e-06, "loss": 0.6007, "step": 7912 }, { "epoch": 0.3317020005659013, "grad_norm": 1.5955917835235596, "learning_rate": 7.796977463155958e-06, "loss": 0.4991, "step": 7913 }, { "epoch": 0.3317439191809103, "grad_norm": 1.6744312047958374, "learning_rate": 7.796414736107549e-06, "loss": 0.5283, "step": 7914 }, { "epoch": 0.33178583779591925, "grad_norm": 1.6850215196609497, "learning_rate": 7.7958519575111e-06, "loss": 0.5501, "step": 7915 }, { "epoch": 0.3318277564109282, "grad_norm": 1.6605141162872314, "learning_rate": 7.795289127376985e-06, "loss": 0.4928, "step": 7916 }, { "epoch": 0.3318696750259371, "grad_norm": 1.4356263875961304, "learning_rate": 7.79472624571558e-06, "loss": 0.4767, "step": 7917 }, { "epoch": 0.3319115936409461, "grad_norm": 1.4737865924835205, "learning_rate": 7.794163312537263e-06, "loss": 0.5338, "step": 7918 }, { "epoch": 0.33195351225595504, "grad_norm": 1.7453237771987915, "learning_rate": 7.793600327852404e-06, "loss": 0.5341, "step": 7919 }, { "epoch": 0.331995430870964, "grad_norm": 1.4513322114944458, "learning_rate": 7.793037291671387e-06, "loss": 0.5407, "step": 7920 }, { "epoch": 0.33203734948597297, "grad_norm": 2.1083998680114746, "learning_rate": 7.792474204004588e-06, "loss": 0.5713, "step": 7921 }, { "epoch": 0.33207926810098193, "grad_norm": 1.9335545301437378, "learning_rate": 7.791911064862387e-06, "loss": 0.5219, "step": 7922 }, { "epoch": 0.3321211867159909, "grad_norm": 1.5882911682128906, "learning_rate": 7.791347874255166e-06, "loss": 0.5753, "step": 7923 }, { "epoch": 0.33216310533099985, "grad_norm": 1.5859652757644653, "learning_rate": 7.790784632193307e-06, "loss": 0.5386, "step": 7924 }, { "epoch": 0.3322050239460088, "grad_norm": 1.4356565475463867, "learning_rate": 7.790221338687189e-06, "loss": 0.4823, "step": 7925 }, { "epoch": 0.3322469425610178, "grad_norm": 1.6103383302688599, "learning_rate": 7.789657993747199e-06, "loss": 0.4965, "step": 7926 }, { "epoch": 0.33228886117602674, "grad_norm": 1.8565236330032349, "learning_rate": 7.789094597383721e-06, "loss": 0.4761, "step": 7927 }, { "epoch": 0.3323307797910357, "grad_norm": 1.5258792638778687, "learning_rate": 7.78853114960714e-06, "loss": 0.599, "step": 7928 }, { "epoch": 0.33237269840604466, "grad_norm": 1.6123853921890259, "learning_rate": 7.787967650427842e-06, "loss": 0.4906, "step": 7929 }, { "epoch": 0.3324146170210536, "grad_norm": 2.777456283569336, "learning_rate": 7.787404099856215e-06, "loss": 0.5313, "step": 7930 }, { "epoch": 0.3324565356360626, "grad_norm": 1.5907738208770752, "learning_rate": 7.786840497902645e-06, "loss": 0.5064, "step": 7931 }, { "epoch": 0.33249845425107155, "grad_norm": 1.6197090148925781, "learning_rate": 7.786276844577525e-06, "loss": 0.5206, "step": 7932 }, { "epoch": 0.3325403728660805, "grad_norm": 1.8065224885940552, "learning_rate": 7.785713139891245e-06, "loss": 0.56, "step": 7933 }, { "epoch": 0.3325822914810895, "grad_norm": 1.7047655582427979, "learning_rate": 7.785149383854192e-06, "loss": 0.5539, "step": 7934 }, { "epoch": 0.33262421009609844, "grad_norm": 1.6697118282318115, "learning_rate": 7.78458557647676e-06, "loss": 0.5132, "step": 7935 }, { "epoch": 0.3326661287111074, "grad_norm": 1.6209611892700195, "learning_rate": 7.784021717769344e-06, "loss": 0.4885, "step": 7936 }, { "epoch": 0.33270804732611636, "grad_norm": 1.4918382167816162, "learning_rate": 7.783457807742335e-06, "loss": 0.4878, "step": 7937 }, { "epoch": 0.3327499659411253, "grad_norm": 1.5890421867370605, "learning_rate": 7.78289384640613e-06, "loss": 0.5246, "step": 7938 }, { "epoch": 0.3327918845561343, "grad_norm": 1.7199968099594116, "learning_rate": 7.782329833771125e-06, "loss": 0.5481, "step": 7939 }, { "epoch": 0.33283380317114325, "grad_norm": 1.5151865482330322, "learning_rate": 7.781765769847714e-06, "loss": 0.5217, "step": 7940 }, { "epoch": 0.3328757217861522, "grad_norm": 1.8892748355865479, "learning_rate": 7.781201654646298e-06, "loss": 0.53, "step": 7941 }, { "epoch": 0.3329176404011611, "grad_norm": 1.5628582239151, "learning_rate": 7.780637488177274e-06, "loss": 0.5555, "step": 7942 }, { "epoch": 0.3329595590161701, "grad_norm": 1.636765480041504, "learning_rate": 7.780073270451042e-06, "loss": 0.518, "step": 7943 }, { "epoch": 0.33300147763117904, "grad_norm": 1.5124175548553467, "learning_rate": 7.779509001478003e-06, "loss": 0.5096, "step": 7944 }, { "epoch": 0.333043396246188, "grad_norm": 1.5877314805984497, "learning_rate": 7.778944681268559e-06, "loss": 0.4751, "step": 7945 }, { "epoch": 0.33308531486119697, "grad_norm": 1.530936360359192, "learning_rate": 7.778380309833109e-06, "loss": 0.4848, "step": 7946 }, { "epoch": 0.33312723347620593, "grad_norm": 1.9313058853149414, "learning_rate": 7.777815887182061e-06, "loss": 0.5719, "step": 7947 }, { "epoch": 0.3331691520912149, "grad_norm": 1.6336166858673096, "learning_rate": 7.777251413325817e-06, "loss": 0.5285, "step": 7948 }, { "epoch": 0.33321107070622386, "grad_norm": 1.7664436101913452, "learning_rate": 7.776686888274783e-06, "loss": 0.5404, "step": 7949 }, { "epoch": 0.3332529893212328, "grad_norm": 1.5712710618972778, "learning_rate": 7.776122312039364e-06, "loss": 0.5057, "step": 7950 }, { "epoch": 0.3332949079362418, "grad_norm": 1.513602614402771, "learning_rate": 7.775557684629968e-06, "loss": 0.5457, "step": 7951 }, { "epoch": 0.33333682655125074, "grad_norm": 1.7256039381027222, "learning_rate": 7.774993006057004e-06, "loss": 0.577, "step": 7952 }, { "epoch": 0.3333787451662597, "grad_norm": 1.652610421180725, "learning_rate": 7.77442827633088e-06, "loss": 0.49, "step": 7953 }, { "epoch": 0.33342066378126867, "grad_norm": 1.6738020181655884, "learning_rate": 7.773863495462006e-06, "loss": 0.5765, "step": 7954 }, { "epoch": 0.33346258239627763, "grad_norm": 1.4377371072769165, "learning_rate": 7.773298663460792e-06, "loss": 0.4637, "step": 7955 }, { "epoch": 0.3335045010112866, "grad_norm": 1.6408580541610718, "learning_rate": 7.772733780337654e-06, "loss": 0.519, "step": 7956 }, { "epoch": 0.33354641962629555, "grad_norm": 1.596890926361084, "learning_rate": 7.772168846103e-06, "loss": 0.5608, "step": 7957 }, { "epoch": 0.3335883382413045, "grad_norm": 1.4972881078720093, "learning_rate": 7.771603860767244e-06, "loss": 0.5221, "step": 7958 }, { "epoch": 0.3336302568563135, "grad_norm": 1.6061850786209106, "learning_rate": 7.771038824340807e-06, "loss": 0.4459, "step": 7959 }, { "epoch": 0.33367217547132244, "grad_norm": 1.6852211952209473, "learning_rate": 7.770473736834097e-06, "loss": 0.5877, "step": 7960 }, { "epoch": 0.3337140940863314, "grad_norm": 1.4972970485687256, "learning_rate": 7.769908598257534e-06, "loss": 0.5352, "step": 7961 }, { "epoch": 0.33375601270134037, "grad_norm": 1.6713567972183228, "learning_rate": 7.769343408621538e-06, "loss": 0.4841, "step": 7962 }, { "epoch": 0.33379793131634933, "grad_norm": 1.6373047828674316, "learning_rate": 7.768778167936522e-06, "loss": 0.5029, "step": 7963 }, { "epoch": 0.3338398499313583, "grad_norm": 1.561472773551941, "learning_rate": 7.768212876212908e-06, "loss": 0.4813, "step": 7964 }, { "epoch": 0.33388176854636725, "grad_norm": 1.637944221496582, "learning_rate": 7.767647533461115e-06, "loss": 0.5703, "step": 7965 }, { "epoch": 0.3339236871613762, "grad_norm": 1.6271592378616333, "learning_rate": 7.76708213969157e-06, "loss": 0.5176, "step": 7966 }, { "epoch": 0.3339656057763851, "grad_norm": 1.43656587600708, "learning_rate": 7.766516694914688e-06, "loss": 0.5085, "step": 7967 }, { "epoch": 0.3340075243913941, "grad_norm": 1.756969690322876, "learning_rate": 7.765951199140898e-06, "loss": 0.4965, "step": 7968 }, { "epoch": 0.33404944300640305, "grad_norm": 1.45943021774292, "learning_rate": 7.76538565238062e-06, "loss": 0.5474, "step": 7969 }, { "epoch": 0.334091361621412, "grad_norm": 1.713492751121521, "learning_rate": 7.76482005464428e-06, "loss": 0.5432, "step": 7970 }, { "epoch": 0.33413328023642097, "grad_norm": 1.5329558849334717, "learning_rate": 7.764254405942304e-06, "loss": 0.5263, "step": 7971 }, { "epoch": 0.33417519885142993, "grad_norm": 1.6380141973495483, "learning_rate": 7.763688706285122e-06, "loss": 0.5585, "step": 7972 }, { "epoch": 0.3342171174664389, "grad_norm": 1.5405868291854858, "learning_rate": 7.763122955683158e-06, "loss": 0.5189, "step": 7973 }, { "epoch": 0.33425903608144786, "grad_norm": 1.6031919717788696, "learning_rate": 7.762557154146842e-06, "loss": 0.5227, "step": 7974 }, { "epoch": 0.3343009546964568, "grad_norm": 1.7103272676467896, "learning_rate": 7.761991301686603e-06, "loss": 0.5584, "step": 7975 }, { "epoch": 0.3343428733114658, "grad_norm": 1.7163807153701782, "learning_rate": 7.761425398312874e-06, "loss": 0.5196, "step": 7976 }, { "epoch": 0.33438479192647474, "grad_norm": 1.644002914428711, "learning_rate": 7.760859444036086e-06, "loss": 0.5357, "step": 7977 }, { "epoch": 0.3344267105414837, "grad_norm": 1.5233246088027954, "learning_rate": 7.760293438866669e-06, "loss": 0.5114, "step": 7978 }, { "epoch": 0.33446862915649267, "grad_norm": 1.7676446437835693, "learning_rate": 7.759727382815062e-06, "loss": 0.5267, "step": 7979 }, { "epoch": 0.33451054777150163, "grad_norm": 1.504557490348816, "learning_rate": 7.75916127589169e-06, "loss": 0.507, "step": 7980 }, { "epoch": 0.3345524663865106, "grad_norm": 1.5484305620193481, "learning_rate": 7.758595118107e-06, "loss": 0.4737, "step": 7981 }, { "epoch": 0.33459438500151956, "grad_norm": 2.151592254638672, "learning_rate": 7.75802890947142e-06, "loss": 0.5774, "step": 7982 }, { "epoch": 0.3346363036165285, "grad_norm": 1.5661003589630127, "learning_rate": 7.75746264999539e-06, "loss": 0.512, "step": 7983 }, { "epoch": 0.3346782222315375, "grad_norm": 1.6380599737167358, "learning_rate": 7.756896339689347e-06, "loss": 0.5257, "step": 7984 }, { "epoch": 0.33472014084654644, "grad_norm": 1.7349587678909302, "learning_rate": 7.756329978563735e-06, "loss": 0.4964, "step": 7985 }, { "epoch": 0.3347620594615554, "grad_norm": 2.0017738342285156, "learning_rate": 7.755763566628987e-06, "loss": 0.5388, "step": 7986 }, { "epoch": 0.33480397807656437, "grad_norm": 1.3835668563842773, "learning_rate": 7.755197103895549e-06, "loss": 0.5091, "step": 7987 }, { "epoch": 0.33484589669157333, "grad_norm": 1.9132893085479736, "learning_rate": 7.754630590373859e-06, "loss": 0.4918, "step": 7988 }, { "epoch": 0.3348878153065823, "grad_norm": 1.6382179260253906, "learning_rate": 7.754064026074363e-06, "loss": 0.51, "step": 7989 }, { "epoch": 0.33492973392159126, "grad_norm": 2.0173521041870117, "learning_rate": 7.753497411007506e-06, "loss": 0.5851, "step": 7990 }, { "epoch": 0.3349716525366002, "grad_norm": 1.4555882215499878, "learning_rate": 7.75293074518373e-06, "loss": 0.4541, "step": 7991 }, { "epoch": 0.3350135711516091, "grad_norm": 1.4419095516204834, "learning_rate": 7.752364028613481e-06, "loss": 0.4998, "step": 7992 }, { "epoch": 0.3350554897666181, "grad_norm": 1.540397047996521, "learning_rate": 7.751797261307207e-06, "loss": 0.5207, "step": 7993 }, { "epoch": 0.33509740838162705, "grad_norm": 1.67054283618927, "learning_rate": 7.751230443275355e-06, "loss": 0.5867, "step": 7994 }, { "epoch": 0.335139326996636, "grad_norm": 2.068887233734131, "learning_rate": 7.750663574528373e-06, "loss": 0.5059, "step": 7995 }, { "epoch": 0.335181245611645, "grad_norm": 1.5768153667449951, "learning_rate": 7.75009665507671e-06, "loss": 0.5097, "step": 7996 }, { "epoch": 0.33522316422665394, "grad_norm": 1.5306435823440552, "learning_rate": 7.749529684930816e-06, "loss": 0.4939, "step": 7997 }, { "epoch": 0.3352650828416629, "grad_norm": 2.493454694747925, "learning_rate": 7.748962664101145e-06, "loss": 0.5569, "step": 7998 }, { "epoch": 0.33530700145667186, "grad_norm": 1.8167163133621216, "learning_rate": 7.748395592598148e-06, "loss": 0.5246, "step": 7999 }, { "epoch": 0.3353489200716808, "grad_norm": 2.1878650188446045, "learning_rate": 7.747828470432277e-06, "loss": 0.485, "step": 8000 }, { "epoch": 0.3353908386866898, "grad_norm": 1.4533036947250366, "learning_rate": 7.747261297613987e-06, "loss": 0.4873, "step": 8001 }, { "epoch": 0.33543275730169875, "grad_norm": 1.8540654182434082, "learning_rate": 7.746694074153733e-06, "loss": 0.5233, "step": 8002 }, { "epoch": 0.3354746759167077, "grad_norm": 2.3186590671539307, "learning_rate": 7.74612680006197e-06, "loss": 0.523, "step": 8003 }, { "epoch": 0.33551659453171667, "grad_norm": 1.6282657384872437, "learning_rate": 7.745559475349158e-06, "loss": 0.5411, "step": 8004 }, { "epoch": 0.33555851314672563, "grad_norm": 2.264368772506714, "learning_rate": 7.74499210002575e-06, "loss": 0.4979, "step": 8005 }, { "epoch": 0.3356004317617346, "grad_norm": 1.8483229875564575, "learning_rate": 7.744424674102208e-06, "loss": 0.5777, "step": 8006 }, { "epoch": 0.33564235037674356, "grad_norm": 1.6231342554092407, "learning_rate": 7.743857197588993e-06, "loss": 0.5116, "step": 8007 }, { "epoch": 0.3356842689917525, "grad_norm": 1.6805921792984009, "learning_rate": 7.743289670496564e-06, "loss": 0.4998, "step": 8008 }, { "epoch": 0.3357261876067615, "grad_norm": 1.521575689315796, "learning_rate": 7.74272209283538e-06, "loss": 0.5223, "step": 8009 }, { "epoch": 0.33576810622177045, "grad_norm": 1.452742099761963, "learning_rate": 7.74215446461591e-06, "loss": 0.5302, "step": 8010 }, { "epoch": 0.3358100248367794, "grad_norm": 1.740692377090454, "learning_rate": 7.741586785848609e-06, "loss": 0.5353, "step": 8011 }, { "epoch": 0.33585194345178837, "grad_norm": 2.038357734680176, "learning_rate": 7.741019056543947e-06, "loss": 0.5349, "step": 8012 }, { "epoch": 0.33589386206679733, "grad_norm": 1.5421072244644165, "learning_rate": 7.740451276712388e-06, "loss": 0.53, "step": 8013 }, { "epoch": 0.3359357806818063, "grad_norm": 1.4057719707489014, "learning_rate": 7.739883446364398e-06, "loss": 0.5472, "step": 8014 }, { "epoch": 0.33597769929681526, "grad_norm": 1.631494164466858, "learning_rate": 7.739315565510446e-06, "loss": 0.5637, "step": 8015 }, { "epoch": 0.3360196179118242, "grad_norm": 1.5104914903640747, "learning_rate": 7.738747634160997e-06, "loss": 0.5159, "step": 8016 }, { "epoch": 0.3360615365268331, "grad_norm": 1.7815210819244385, "learning_rate": 7.738179652326523e-06, "loss": 0.5488, "step": 8017 }, { "epoch": 0.3361034551418421, "grad_norm": 1.6829679012298584, "learning_rate": 7.737611620017491e-06, "loss": 0.5028, "step": 8018 }, { "epoch": 0.33614537375685105, "grad_norm": 1.7915401458740234, "learning_rate": 7.737043537244375e-06, "loss": 0.502, "step": 8019 }, { "epoch": 0.33618729237186, "grad_norm": 1.6353790760040283, "learning_rate": 7.736475404017643e-06, "loss": 0.4569, "step": 8020 }, { "epoch": 0.336229210986869, "grad_norm": 1.6801018714904785, "learning_rate": 7.735907220347772e-06, "loss": 0.5142, "step": 8021 }, { "epoch": 0.33627112960187794, "grad_norm": 1.5836544036865234, "learning_rate": 7.735338986245234e-06, "loss": 0.5637, "step": 8022 }, { "epoch": 0.3363130482168869, "grad_norm": 1.6545737981796265, "learning_rate": 7.7347707017205e-06, "loss": 0.4821, "step": 8023 }, { "epoch": 0.33635496683189586, "grad_norm": 1.7559001445770264, "learning_rate": 7.734202366784052e-06, "loss": 0.5205, "step": 8024 }, { "epoch": 0.3363968854469048, "grad_norm": 1.499706506729126, "learning_rate": 7.733633981446363e-06, "loss": 0.5166, "step": 8025 }, { "epoch": 0.3364388040619138, "grad_norm": 1.7630032300949097, "learning_rate": 7.73306554571791e-06, "loss": 0.4898, "step": 8026 }, { "epoch": 0.33648072267692275, "grad_norm": 1.5432379245758057, "learning_rate": 7.732497059609172e-06, "loss": 0.5487, "step": 8027 }, { "epoch": 0.3365226412919317, "grad_norm": 1.6100910902023315, "learning_rate": 7.73192852313063e-06, "loss": 0.5543, "step": 8028 }, { "epoch": 0.3365645599069407, "grad_norm": 1.5634435415267944, "learning_rate": 7.731359936292759e-06, "loss": 0.4895, "step": 8029 }, { "epoch": 0.33660647852194964, "grad_norm": 1.7714343070983887, "learning_rate": 7.730791299106046e-06, "loss": 0.4807, "step": 8030 }, { "epoch": 0.3366483971369586, "grad_norm": 1.80374276638031, "learning_rate": 7.730222611580969e-06, "loss": 0.524, "step": 8031 }, { "epoch": 0.33669031575196756, "grad_norm": 1.6129703521728516, "learning_rate": 7.729653873728014e-06, "loss": 0.4825, "step": 8032 }, { "epoch": 0.3367322343669765, "grad_norm": 1.8803606033325195, "learning_rate": 7.729085085557664e-06, "loss": 0.5519, "step": 8033 }, { "epoch": 0.3367741529819855, "grad_norm": 1.683209776878357, "learning_rate": 7.728516247080403e-06, "loss": 0.5363, "step": 8034 }, { "epoch": 0.33681607159699445, "grad_norm": 1.5782893896102905, "learning_rate": 7.727947358306715e-06, "loss": 0.5507, "step": 8035 }, { "epoch": 0.3368579902120034, "grad_norm": 1.7818752527236938, "learning_rate": 7.727378419247088e-06, "loss": 0.5578, "step": 8036 }, { "epoch": 0.3368999088270124, "grad_norm": 1.571522831916809, "learning_rate": 7.726809429912014e-06, "loss": 0.5355, "step": 8037 }, { "epoch": 0.33694182744202134, "grad_norm": 1.557410717010498, "learning_rate": 7.726240390311976e-06, "loss": 0.5143, "step": 8038 }, { "epoch": 0.3369837460570303, "grad_norm": 1.7423030138015747, "learning_rate": 7.725671300457466e-06, "loss": 0.5195, "step": 8039 }, { "epoch": 0.33702566467203926, "grad_norm": 1.7581490278244019, "learning_rate": 7.725102160358974e-06, "loss": 0.5671, "step": 8040 }, { "epoch": 0.3370675832870482, "grad_norm": 1.4726792573928833, "learning_rate": 7.72453297002699e-06, "loss": 0.4959, "step": 8041 }, { "epoch": 0.33710950190205713, "grad_norm": 1.5299949645996094, "learning_rate": 7.723963729472008e-06, "loss": 0.5087, "step": 8042 }, { "epoch": 0.3371514205170661, "grad_norm": 1.4567577838897705, "learning_rate": 7.72339443870452e-06, "loss": 0.5111, "step": 8043 }, { "epoch": 0.33719333913207505, "grad_norm": 1.6573405265808105, "learning_rate": 7.72282509773502e-06, "loss": 0.4535, "step": 8044 }, { "epoch": 0.337235257747084, "grad_norm": 2.005016326904297, "learning_rate": 7.722255706574004e-06, "loss": 0.5404, "step": 8045 }, { "epoch": 0.337277176362093, "grad_norm": 1.8629887104034424, "learning_rate": 7.721686265231966e-06, "loss": 0.5492, "step": 8046 }, { "epoch": 0.33731909497710194, "grad_norm": 1.6195106506347656, "learning_rate": 7.721116773719408e-06, "loss": 0.5156, "step": 8047 }, { "epoch": 0.3373610135921109, "grad_norm": 1.7227429151535034, "learning_rate": 7.72054723204682e-06, "loss": 0.5145, "step": 8048 }, { "epoch": 0.33740293220711987, "grad_norm": 1.7966334819793701, "learning_rate": 7.719977640224708e-06, "loss": 0.505, "step": 8049 }, { "epoch": 0.33744485082212883, "grad_norm": 1.662535309791565, "learning_rate": 7.719407998263567e-06, "loss": 0.5949, "step": 8050 }, { "epoch": 0.3374867694371378, "grad_norm": 1.766169786453247, "learning_rate": 7.7188383061739e-06, "loss": 0.567, "step": 8051 }, { "epoch": 0.33752868805214675, "grad_norm": 1.6130163669586182, "learning_rate": 7.718268563966206e-06, "loss": 0.5139, "step": 8052 }, { "epoch": 0.3375706066671557, "grad_norm": 1.5853736400604248, "learning_rate": 7.71769877165099e-06, "loss": 0.5232, "step": 8053 }, { "epoch": 0.3376125252821647, "grad_norm": 1.7329221963882446, "learning_rate": 7.717128929238753e-06, "loss": 0.5029, "step": 8054 }, { "epoch": 0.33765444389717364, "grad_norm": 1.619358777999878, "learning_rate": 7.716559036740001e-06, "loss": 0.5366, "step": 8055 }, { "epoch": 0.3376963625121826, "grad_norm": 1.5148875713348389, "learning_rate": 7.715989094165239e-06, "loss": 0.5049, "step": 8056 }, { "epoch": 0.33773828112719156, "grad_norm": 2.1671502590179443, "learning_rate": 7.715419101524974e-06, "loss": 0.5475, "step": 8057 }, { "epoch": 0.3377801997422005, "grad_norm": 1.6480873823165894, "learning_rate": 7.71484905882971e-06, "loss": 0.459, "step": 8058 }, { "epoch": 0.3378221183572095, "grad_norm": 1.4955240488052368, "learning_rate": 7.714278966089958e-06, "loss": 0.4793, "step": 8059 }, { "epoch": 0.33786403697221845, "grad_norm": 1.577035903930664, "learning_rate": 7.713708823316224e-06, "loss": 0.5281, "step": 8060 }, { "epoch": 0.3379059555872274, "grad_norm": 1.6376349925994873, "learning_rate": 7.713138630519022e-06, "loss": 0.4457, "step": 8061 }, { "epoch": 0.3379478742022364, "grad_norm": 2.1909728050231934, "learning_rate": 7.71256838770886e-06, "loss": 0.5069, "step": 8062 }, { "epoch": 0.33798979281724534, "grad_norm": 2.0724592208862305, "learning_rate": 7.711998094896247e-06, "loss": 0.5263, "step": 8063 }, { "epoch": 0.3380317114322543, "grad_norm": 1.6586910486221313, "learning_rate": 7.711427752091698e-06, "loss": 0.5052, "step": 8064 }, { "epoch": 0.33807363004726326, "grad_norm": 1.9001777172088623, "learning_rate": 7.71085735930573e-06, "loss": 0.5201, "step": 8065 }, { "epoch": 0.3381155486622722, "grad_norm": 1.666716456413269, "learning_rate": 7.710286916548853e-06, "loss": 0.5317, "step": 8066 }, { "epoch": 0.33815746727728113, "grad_norm": 2.0526602268218994, "learning_rate": 7.709716423831584e-06, "loss": 0.5247, "step": 8067 }, { "epoch": 0.3381993858922901, "grad_norm": 1.739140272140503, "learning_rate": 7.709145881164437e-06, "loss": 0.6019, "step": 8068 }, { "epoch": 0.33824130450729906, "grad_norm": 2.102930784225464, "learning_rate": 7.70857528855793e-06, "loss": 0.5555, "step": 8069 }, { "epoch": 0.338283223122308, "grad_norm": 1.819245457649231, "learning_rate": 7.708004646022587e-06, "loss": 0.567, "step": 8070 }, { "epoch": 0.338325141737317, "grad_norm": 1.7667924165725708, "learning_rate": 7.70743395356892e-06, "loss": 0.5287, "step": 8071 }, { "epoch": 0.33836706035232594, "grad_norm": 1.4700015783309937, "learning_rate": 7.70686321120745e-06, "loss": 0.5304, "step": 8072 }, { "epoch": 0.3384089789673349, "grad_norm": 1.8838518857955933, "learning_rate": 7.706292418948697e-06, "loss": 0.502, "step": 8073 }, { "epoch": 0.33845089758234387, "grad_norm": 1.6131908893585205, "learning_rate": 7.705721576803188e-06, "loss": 0.5021, "step": 8074 }, { "epoch": 0.33849281619735283, "grad_norm": 1.6743569374084473, "learning_rate": 7.705150684781442e-06, "loss": 0.5517, "step": 8075 }, { "epoch": 0.3385347348123618, "grad_norm": 1.992917537689209, "learning_rate": 7.70457974289398e-06, "loss": 0.5384, "step": 8076 }, { "epoch": 0.33857665342737076, "grad_norm": 1.7243403196334839, "learning_rate": 7.704008751151333e-06, "loss": 0.5493, "step": 8077 }, { "epoch": 0.3386185720423797, "grad_norm": 1.6505930423736572, "learning_rate": 7.70343770956402e-06, "loss": 0.5026, "step": 8078 }, { "epoch": 0.3386604906573887, "grad_norm": 1.5364030599594116, "learning_rate": 7.70286661814257e-06, "loss": 0.5907, "step": 8079 }, { "epoch": 0.33870240927239764, "grad_norm": 1.6055785417556763, "learning_rate": 7.702295476897514e-06, "loss": 0.4962, "step": 8080 }, { "epoch": 0.3387443278874066, "grad_norm": 1.9720178842544556, "learning_rate": 7.701724285839374e-06, "loss": 0.5729, "step": 8081 }, { "epoch": 0.33878624650241557, "grad_norm": 1.541481375694275, "learning_rate": 7.701153044978683e-06, "loss": 0.4534, "step": 8082 }, { "epoch": 0.33882816511742453, "grad_norm": 2.100067615509033, "learning_rate": 7.700581754325969e-06, "loss": 0.5986, "step": 8083 }, { "epoch": 0.3388700837324335, "grad_norm": 1.5339171886444092, "learning_rate": 7.700010413891763e-06, "loss": 0.528, "step": 8084 }, { "epoch": 0.33891200234744245, "grad_norm": 1.9118071794509888, "learning_rate": 7.699439023686598e-06, "loss": 0.5445, "step": 8085 }, { "epoch": 0.3389539209624514, "grad_norm": 1.6678893566131592, "learning_rate": 7.698867583721009e-06, "loss": 0.5096, "step": 8086 }, { "epoch": 0.3389958395774604, "grad_norm": 1.6874886751174927, "learning_rate": 7.698296094005524e-06, "loss": 0.4854, "step": 8087 }, { "epoch": 0.33903775819246934, "grad_norm": 1.7103798389434814, "learning_rate": 7.697724554550681e-06, "loss": 0.5463, "step": 8088 }, { "epoch": 0.3390796768074783, "grad_norm": 1.5486356019973755, "learning_rate": 7.697152965367016e-06, "loss": 0.521, "step": 8089 }, { "epoch": 0.33912159542248727, "grad_norm": 1.7923320531845093, "learning_rate": 7.696581326465066e-06, "loss": 0.5464, "step": 8090 }, { "epoch": 0.3391635140374962, "grad_norm": 1.7206754684448242, "learning_rate": 7.696009637855366e-06, "loss": 0.4779, "step": 8091 }, { "epoch": 0.33920543265250513, "grad_norm": 2.44073224067688, "learning_rate": 7.695437899548457e-06, "loss": 0.5759, "step": 8092 }, { "epoch": 0.3392473512675141, "grad_norm": 2.1656110286712646, "learning_rate": 7.694866111554875e-06, "loss": 0.5685, "step": 8093 }, { "epoch": 0.33928926988252306, "grad_norm": 1.7011319398880005, "learning_rate": 7.694294273885161e-06, "loss": 0.505, "step": 8094 }, { "epoch": 0.339331188497532, "grad_norm": 1.740114450454712, "learning_rate": 7.69372238654986e-06, "loss": 0.5146, "step": 8095 }, { "epoch": 0.339373107112541, "grad_norm": 1.6067633628845215, "learning_rate": 7.693150449559507e-06, "loss": 0.4973, "step": 8096 }, { "epoch": 0.33941502572754995, "grad_norm": 1.7519323825836182, "learning_rate": 7.69257846292465e-06, "loss": 0.5505, "step": 8097 }, { "epoch": 0.3394569443425589, "grad_norm": 1.9143027067184448, "learning_rate": 7.692006426655832e-06, "loss": 0.5481, "step": 8098 }, { "epoch": 0.33949886295756787, "grad_norm": 1.7487812042236328, "learning_rate": 7.691434340763598e-06, "loss": 0.5412, "step": 8099 }, { "epoch": 0.33954078157257683, "grad_norm": 2.0471932888031006, "learning_rate": 7.690862205258491e-06, "loss": 0.5374, "step": 8100 }, { "epoch": 0.3395827001875858, "grad_norm": 2.526566505432129, "learning_rate": 7.69029002015106e-06, "loss": 0.5487, "step": 8101 }, { "epoch": 0.33962461880259476, "grad_norm": 1.6637300252914429, "learning_rate": 7.689717785451853e-06, "loss": 0.5046, "step": 8102 }, { "epoch": 0.3396665374176037, "grad_norm": 2.0839529037475586, "learning_rate": 7.689145501171413e-06, "loss": 0.507, "step": 8103 }, { "epoch": 0.3397084560326127, "grad_norm": 2.168111562728882, "learning_rate": 7.688573167320297e-06, "loss": 0.5144, "step": 8104 }, { "epoch": 0.33975037464762164, "grad_norm": 1.6727672815322876, "learning_rate": 7.688000783909052e-06, "loss": 0.4906, "step": 8105 }, { "epoch": 0.3397922932626306, "grad_norm": 1.801802158355713, "learning_rate": 7.687428350948225e-06, "loss": 0.5378, "step": 8106 }, { "epoch": 0.33983421187763957, "grad_norm": 1.6511865854263306, "learning_rate": 7.686855868448376e-06, "loss": 0.5163, "step": 8107 }, { "epoch": 0.33987613049264853, "grad_norm": 1.7647359371185303, "learning_rate": 7.686283336420051e-06, "loss": 0.5092, "step": 8108 }, { "epoch": 0.3399180491076575, "grad_norm": 1.7542005777359009, "learning_rate": 7.685710754873805e-06, "loss": 0.491, "step": 8109 }, { "epoch": 0.33995996772266646, "grad_norm": 1.645634651184082, "learning_rate": 7.685138123820195e-06, "loss": 0.5474, "step": 8110 }, { "epoch": 0.3400018863376754, "grad_norm": 1.774958848953247, "learning_rate": 7.684565443269778e-06, "loss": 0.5379, "step": 8111 }, { "epoch": 0.3400438049526844, "grad_norm": 1.667244791984558, "learning_rate": 7.683992713233107e-06, "loss": 0.4614, "step": 8112 }, { "epoch": 0.34008572356769334, "grad_norm": 1.805387020111084, "learning_rate": 7.68341993372074e-06, "loss": 0.5911, "step": 8113 }, { "epoch": 0.3401276421827023, "grad_norm": 1.7756125926971436, "learning_rate": 7.682847104743238e-06, "loss": 0.5639, "step": 8114 }, { "epoch": 0.34016956079771127, "grad_norm": 1.7866944074630737, "learning_rate": 7.682274226311157e-06, "loss": 0.5344, "step": 8115 }, { "epoch": 0.34021147941272023, "grad_norm": 1.8896299600601196, "learning_rate": 7.681701298435058e-06, "loss": 0.5716, "step": 8116 }, { "epoch": 0.34025339802772914, "grad_norm": 1.7331165075302124, "learning_rate": 7.681128321125505e-06, "loss": 0.5648, "step": 8117 }, { "epoch": 0.3402953166427381, "grad_norm": 1.6787363290786743, "learning_rate": 7.680555294393058e-06, "loss": 0.5727, "step": 8118 }, { "epoch": 0.34033723525774706, "grad_norm": 1.5097347497940063, "learning_rate": 7.679982218248278e-06, "loss": 0.5389, "step": 8119 }, { "epoch": 0.340379153872756, "grad_norm": 1.7485597133636475, "learning_rate": 7.679409092701731e-06, "loss": 0.5795, "step": 8120 }, { "epoch": 0.340421072487765, "grad_norm": 1.6515532732009888, "learning_rate": 7.678835917763982e-06, "loss": 0.572, "step": 8121 }, { "epoch": 0.34046299110277395, "grad_norm": 1.6299329996109009, "learning_rate": 7.678262693445598e-06, "loss": 0.5133, "step": 8122 }, { "epoch": 0.3405049097177829, "grad_norm": 2.2024970054626465, "learning_rate": 7.677689419757142e-06, "loss": 0.5339, "step": 8123 }, { "epoch": 0.3405468283327919, "grad_norm": 2.042250633239746, "learning_rate": 7.677116096709184e-06, "loss": 0.5358, "step": 8124 }, { "epoch": 0.34058874694780084, "grad_norm": 1.7905718088150024, "learning_rate": 7.67654272431229e-06, "loss": 0.546, "step": 8125 }, { "epoch": 0.3406306655628098, "grad_norm": 2.087843418121338, "learning_rate": 7.675969302577032e-06, "loss": 0.5257, "step": 8126 }, { "epoch": 0.34067258417781876, "grad_norm": 1.7127898931503296, "learning_rate": 7.675395831513981e-06, "loss": 0.473, "step": 8127 }, { "epoch": 0.3407145027928277, "grad_norm": 1.8218250274658203, "learning_rate": 7.674822311133706e-06, "loss": 0.5367, "step": 8128 }, { "epoch": 0.3407564214078367, "grad_norm": 1.613520622253418, "learning_rate": 7.67424874144678e-06, "loss": 0.5414, "step": 8129 }, { "epoch": 0.34079834002284565, "grad_norm": 1.501016616821289, "learning_rate": 7.673675122463773e-06, "loss": 0.4655, "step": 8130 }, { "epoch": 0.3408402586378546, "grad_norm": 1.6900699138641357, "learning_rate": 7.673101454195262e-06, "loss": 0.526, "step": 8131 }, { "epoch": 0.34088217725286357, "grad_norm": 1.7063698768615723, "learning_rate": 7.672527736651822e-06, "loss": 0.5412, "step": 8132 }, { "epoch": 0.34092409586787253, "grad_norm": 1.558298945426941, "learning_rate": 7.671953969844027e-06, "loss": 0.5309, "step": 8133 }, { "epoch": 0.3409660144828815, "grad_norm": 1.936102271080017, "learning_rate": 7.671380153782456e-06, "loss": 0.5352, "step": 8134 }, { "epoch": 0.34100793309789046, "grad_norm": 2.134901762008667, "learning_rate": 7.670806288477683e-06, "loss": 0.549, "step": 8135 }, { "epoch": 0.3410498517128994, "grad_norm": 1.5363889932632446, "learning_rate": 7.67023237394029e-06, "loss": 0.4851, "step": 8136 }, { "epoch": 0.3410917703279084, "grad_norm": 1.6326948404312134, "learning_rate": 7.669658410180855e-06, "loss": 0.5326, "step": 8137 }, { "epoch": 0.34113368894291735, "grad_norm": 1.6728448867797852, "learning_rate": 7.669084397209957e-06, "loss": 0.4803, "step": 8138 }, { "epoch": 0.3411756075579263, "grad_norm": 1.5368242263793945, "learning_rate": 7.668510335038177e-06, "loss": 0.5022, "step": 8139 }, { "epoch": 0.34121752617293527, "grad_norm": 1.6162906885147095, "learning_rate": 7.6679362236761e-06, "loss": 0.5297, "step": 8140 }, { "epoch": 0.34125944478794423, "grad_norm": 1.6416070461273193, "learning_rate": 7.667362063134305e-06, "loss": 0.5122, "step": 8141 }, { "epoch": 0.34130136340295314, "grad_norm": 1.545093059539795, "learning_rate": 7.66678785342338e-06, "loss": 0.5191, "step": 8142 }, { "epoch": 0.3413432820179621, "grad_norm": 1.705271601676941, "learning_rate": 7.666213594553907e-06, "loss": 0.5452, "step": 8143 }, { "epoch": 0.34138520063297106, "grad_norm": 1.543188452720642, "learning_rate": 7.665639286536472e-06, "loss": 0.4928, "step": 8144 }, { "epoch": 0.34142711924798, "grad_norm": 1.8165191411972046, "learning_rate": 7.665064929381662e-06, "loss": 0.506, "step": 8145 }, { "epoch": 0.341469037862989, "grad_norm": 1.58920156955719, "learning_rate": 7.664490523100064e-06, "loss": 0.5002, "step": 8146 }, { "epoch": 0.34151095647799795, "grad_norm": 1.6558135747909546, "learning_rate": 7.663916067702266e-06, "loss": 0.5373, "step": 8147 }, { "epoch": 0.3415528750930069, "grad_norm": 1.6292213201522827, "learning_rate": 7.663341563198861e-06, "loss": 0.5029, "step": 8148 }, { "epoch": 0.3415947937080159, "grad_norm": 1.6714553833007812, "learning_rate": 7.662767009600434e-06, "loss": 0.4958, "step": 8149 }, { "epoch": 0.34163671232302484, "grad_norm": 2.113757371902466, "learning_rate": 7.662192406917576e-06, "loss": 0.5565, "step": 8150 }, { "epoch": 0.3416786309380338, "grad_norm": 1.6423627138137817, "learning_rate": 7.661617755160887e-06, "loss": 0.4827, "step": 8151 }, { "epoch": 0.34172054955304276, "grad_norm": 1.771695613861084, "learning_rate": 7.66104305434095e-06, "loss": 0.5862, "step": 8152 }, { "epoch": 0.3417624681680517, "grad_norm": 1.679907202720642, "learning_rate": 7.660468304468364e-06, "loss": 0.5171, "step": 8153 }, { "epoch": 0.3418043867830607, "grad_norm": 1.4701162576675415, "learning_rate": 7.659893505553724e-06, "loss": 0.5102, "step": 8154 }, { "epoch": 0.34184630539806965, "grad_norm": 2.1358273029327393, "learning_rate": 7.659318657607622e-06, "loss": 0.4845, "step": 8155 }, { "epoch": 0.3418882240130786, "grad_norm": 1.751039981842041, "learning_rate": 7.65874376064066e-06, "loss": 0.4523, "step": 8156 }, { "epoch": 0.3419301426280876, "grad_norm": 1.7364919185638428, "learning_rate": 7.65816881466343e-06, "loss": 0.5434, "step": 8157 }, { "epoch": 0.34197206124309654, "grad_norm": 1.8028013706207275, "learning_rate": 7.657593819686534e-06, "loss": 0.5601, "step": 8158 }, { "epoch": 0.3420139798581055, "grad_norm": 1.6291016340255737, "learning_rate": 7.657018775720568e-06, "loss": 0.489, "step": 8159 }, { "epoch": 0.34205589847311446, "grad_norm": 1.565998911857605, "learning_rate": 7.656443682776138e-06, "loss": 0.4992, "step": 8160 }, { "epoch": 0.3420978170881234, "grad_norm": 1.9343012571334839, "learning_rate": 7.655868540863839e-06, "loss": 0.5422, "step": 8161 }, { "epoch": 0.3421397357031324, "grad_norm": 1.7639877796173096, "learning_rate": 7.655293349994273e-06, "loss": 0.4421, "step": 8162 }, { "epoch": 0.34218165431814135, "grad_norm": 1.618535041809082, "learning_rate": 7.654718110178047e-06, "loss": 0.5465, "step": 8163 }, { "epoch": 0.3422235729331503, "grad_norm": 1.4980640411376953, "learning_rate": 7.654142821425762e-06, "loss": 0.5591, "step": 8164 }, { "epoch": 0.3422654915481593, "grad_norm": 1.5591230392456055, "learning_rate": 7.653567483748023e-06, "loss": 0.5066, "step": 8165 }, { "epoch": 0.34230741016316824, "grad_norm": 1.5895209312438965, "learning_rate": 7.652992097155439e-06, "loss": 0.4789, "step": 8166 }, { "epoch": 0.34234932877817714, "grad_norm": 1.6388102769851685, "learning_rate": 7.65241666165861e-06, "loss": 0.5369, "step": 8167 }, { "epoch": 0.3423912473931861, "grad_norm": 1.443947196006775, "learning_rate": 7.651841177268147e-06, "loss": 0.5161, "step": 8168 }, { "epoch": 0.34243316600819507, "grad_norm": 1.8259294033050537, "learning_rate": 7.651265643994659e-06, "loss": 0.5082, "step": 8169 }, { "epoch": 0.34247508462320403, "grad_norm": 1.8698135614395142, "learning_rate": 7.650690061848754e-06, "loss": 0.5181, "step": 8170 }, { "epoch": 0.342517003238213, "grad_norm": 1.7230167388916016, "learning_rate": 7.650114430841042e-06, "loss": 0.5144, "step": 8171 }, { "epoch": 0.34255892185322195, "grad_norm": 1.489868402481079, "learning_rate": 7.649538750982133e-06, "loss": 0.486, "step": 8172 }, { "epoch": 0.3426008404682309, "grad_norm": 1.5387721061706543, "learning_rate": 7.648963022282641e-06, "loss": 0.531, "step": 8173 }, { "epoch": 0.3426427590832399, "grad_norm": 1.563127875328064, "learning_rate": 7.648387244753178e-06, "loss": 0.4821, "step": 8174 }, { "epoch": 0.34268467769824884, "grad_norm": 1.7072159051895142, "learning_rate": 7.647811418404359e-06, "loss": 0.5755, "step": 8175 }, { "epoch": 0.3427265963132578, "grad_norm": 2.1805531978607178, "learning_rate": 7.647235543246794e-06, "loss": 0.5527, "step": 8176 }, { "epoch": 0.34276851492826677, "grad_norm": 1.4554524421691895, "learning_rate": 7.6466596192911e-06, "loss": 0.5222, "step": 8177 }, { "epoch": 0.3428104335432757, "grad_norm": 1.688157558441162, "learning_rate": 7.646083646547899e-06, "loss": 0.5008, "step": 8178 }, { "epoch": 0.3428523521582847, "grad_norm": 2.3956644535064697, "learning_rate": 7.645507625027803e-06, "loss": 0.4723, "step": 8179 }, { "epoch": 0.34289427077329365, "grad_norm": 1.5995193719863892, "learning_rate": 7.64493155474143e-06, "loss": 0.5497, "step": 8180 }, { "epoch": 0.3429361893883026, "grad_norm": 1.8356972932815552, "learning_rate": 7.644355435699402e-06, "loss": 0.5129, "step": 8181 }, { "epoch": 0.3429781080033116, "grad_norm": 1.7068308591842651, "learning_rate": 7.643779267912337e-06, "loss": 0.5282, "step": 8182 }, { "epoch": 0.34302002661832054, "grad_norm": 1.7470356225967407, "learning_rate": 7.643203051390856e-06, "loss": 0.4975, "step": 8183 }, { "epoch": 0.3430619452333295, "grad_norm": 1.7004579305648804, "learning_rate": 7.64262678614558e-06, "loss": 0.4815, "step": 8184 }, { "epoch": 0.34310386384833846, "grad_norm": 1.658503770828247, "learning_rate": 7.642050472187133e-06, "loss": 0.5558, "step": 8185 }, { "epoch": 0.3431457824633474, "grad_norm": 1.6278563737869263, "learning_rate": 7.641474109526137e-06, "loss": 0.4776, "step": 8186 }, { "epoch": 0.3431877010783564, "grad_norm": 1.6068809032440186, "learning_rate": 7.640897698173217e-06, "loss": 0.5106, "step": 8187 }, { "epoch": 0.34322961969336535, "grad_norm": 1.7796247005462646, "learning_rate": 7.640321238139001e-06, "loss": 0.5062, "step": 8188 }, { "epoch": 0.3432715383083743, "grad_norm": 1.536758303642273, "learning_rate": 7.639744729434114e-06, "loss": 0.4885, "step": 8189 }, { "epoch": 0.3433134569233833, "grad_norm": 1.9574337005615234, "learning_rate": 7.63916817206918e-06, "loss": 0.5434, "step": 8190 }, { "epoch": 0.34335537553839224, "grad_norm": 1.506109356880188, "learning_rate": 7.63859156605483e-06, "loss": 0.5265, "step": 8191 }, { "epoch": 0.34339729415340114, "grad_norm": 1.5850582122802734, "learning_rate": 7.638014911401692e-06, "loss": 0.4803, "step": 8192 }, { "epoch": 0.3434392127684101, "grad_norm": 2.276791572570801, "learning_rate": 7.637438208120396e-06, "loss": 0.4875, "step": 8193 }, { "epoch": 0.34348113138341907, "grad_norm": 1.866836428642273, "learning_rate": 7.636861456221574e-06, "loss": 0.5242, "step": 8194 }, { "epoch": 0.34352304999842803, "grad_norm": 2.041116952896118, "learning_rate": 7.636284655715855e-06, "loss": 0.5199, "step": 8195 }, { "epoch": 0.343564968613437, "grad_norm": 1.6752959489822388, "learning_rate": 7.635707806613873e-06, "loss": 0.5595, "step": 8196 }, { "epoch": 0.34360688722844596, "grad_norm": 2.9909486770629883, "learning_rate": 7.635130908926261e-06, "loss": 0.526, "step": 8197 }, { "epoch": 0.3436488058434549, "grad_norm": 1.6836212873458862, "learning_rate": 7.634553962663656e-06, "loss": 0.5317, "step": 8198 }, { "epoch": 0.3436907244584639, "grad_norm": 1.9921103715896606, "learning_rate": 7.633976967836687e-06, "loss": 0.5367, "step": 8199 }, { "epoch": 0.34373264307347284, "grad_norm": 2.1154232025146484, "learning_rate": 7.633399924455998e-06, "loss": 0.5258, "step": 8200 }, { "epoch": 0.3437745616884818, "grad_norm": 1.725731611251831, "learning_rate": 7.63282283253222e-06, "loss": 0.5131, "step": 8201 }, { "epoch": 0.34381648030349077, "grad_norm": 1.584783911705017, "learning_rate": 7.632245692075992e-06, "loss": 0.4931, "step": 8202 }, { "epoch": 0.34385839891849973, "grad_norm": 1.5893152952194214, "learning_rate": 7.631668503097956e-06, "loss": 0.5434, "step": 8203 }, { "epoch": 0.3439003175335087, "grad_norm": 1.470718502998352, "learning_rate": 7.631091265608747e-06, "loss": 0.5135, "step": 8204 }, { "epoch": 0.34394223614851765, "grad_norm": 1.459373116493225, "learning_rate": 7.63051397961901e-06, "loss": 0.5634, "step": 8205 }, { "epoch": 0.3439841547635266, "grad_norm": 1.4058148860931396, "learning_rate": 7.629936645139381e-06, "loss": 0.4397, "step": 8206 }, { "epoch": 0.3440260733785356, "grad_norm": 1.7268927097320557, "learning_rate": 7.629359262180509e-06, "loss": 0.5133, "step": 8207 }, { "epoch": 0.34406799199354454, "grad_norm": 1.9449279308319092, "learning_rate": 7.628781830753033e-06, "loss": 0.5223, "step": 8208 }, { "epoch": 0.3441099106085535, "grad_norm": 1.9476077556610107, "learning_rate": 7.628204350867599e-06, "loss": 0.5038, "step": 8209 }, { "epoch": 0.34415182922356247, "grad_norm": 1.7952779531478882, "learning_rate": 7.627626822534849e-06, "loss": 0.5559, "step": 8210 }, { "epoch": 0.34419374783857143, "grad_norm": 1.6706944704055786, "learning_rate": 7.627049245765432e-06, "loss": 0.4689, "step": 8211 }, { "epoch": 0.3442356664535804, "grad_norm": 1.6826268434524536, "learning_rate": 7.626471620569995e-06, "loss": 0.4773, "step": 8212 }, { "epoch": 0.34427758506858935, "grad_norm": 1.8811285495758057, "learning_rate": 7.625893946959184e-06, "loss": 0.4906, "step": 8213 }, { "epoch": 0.3443195036835983, "grad_norm": 1.7516264915466309, "learning_rate": 7.625316224943647e-06, "loss": 0.575, "step": 8214 }, { "epoch": 0.3443614222986073, "grad_norm": 1.5354266166687012, "learning_rate": 7.6247384545340365e-06, "loss": 0.5284, "step": 8215 }, { "epoch": 0.34440334091361624, "grad_norm": 1.7418016195297241, "learning_rate": 7.624160635741e-06, "loss": 0.4857, "step": 8216 }, { "epoch": 0.34444525952862515, "grad_norm": 1.8782116174697876, "learning_rate": 7.623582768575191e-06, "loss": 0.5587, "step": 8217 }, { "epoch": 0.3444871781436341, "grad_norm": 1.5978645086288452, "learning_rate": 7.62300485304726e-06, "loss": 0.5459, "step": 8218 }, { "epoch": 0.34452909675864307, "grad_norm": 1.4990400075912476, "learning_rate": 7.622426889167862e-06, "loss": 0.5388, "step": 8219 }, { "epoch": 0.34457101537365203, "grad_norm": 1.7224286794662476, "learning_rate": 7.621848876947648e-06, "loss": 0.5186, "step": 8220 }, { "epoch": 0.344612933988661, "grad_norm": 1.9416182041168213, "learning_rate": 7.621270816397276e-06, "loss": 0.5922, "step": 8221 }, { "epoch": 0.34465485260366996, "grad_norm": 1.8957438468933105, "learning_rate": 7.6206927075274e-06, "loss": 0.5194, "step": 8222 }, { "epoch": 0.3446967712186789, "grad_norm": 2.2549076080322266, "learning_rate": 7.620114550348677e-06, "loss": 0.4969, "step": 8223 }, { "epoch": 0.3447386898336879, "grad_norm": 1.6727313995361328, "learning_rate": 7.619536344871763e-06, "loss": 0.5026, "step": 8224 }, { "epoch": 0.34478060844869685, "grad_norm": 1.662007451057434, "learning_rate": 7.618958091107318e-06, "loss": 0.4939, "step": 8225 }, { "epoch": 0.3448225270637058, "grad_norm": 1.8570805788040161, "learning_rate": 7.618379789066004e-06, "loss": 0.4878, "step": 8226 }, { "epoch": 0.34486444567871477, "grad_norm": 2.254488706588745, "learning_rate": 7.617801438758475e-06, "loss": 0.5466, "step": 8227 }, { "epoch": 0.34490636429372373, "grad_norm": 1.8016680479049683, "learning_rate": 7.617223040195397e-06, "loss": 0.5837, "step": 8228 }, { "epoch": 0.3449482829087327, "grad_norm": 2.2219388484954834, "learning_rate": 7.61664459338743e-06, "loss": 0.5593, "step": 8229 }, { "epoch": 0.34499020152374166, "grad_norm": 1.398064374923706, "learning_rate": 7.616066098345238e-06, "loss": 0.4981, "step": 8230 }, { "epoch": 0.3450321201387506, "grad_norm": 1.5754178762435913, "learning_rate": 7.615487555079484e-06, "loss": 0.5465, "step": 8231 }, { "epoch": 0.3450740387537596, "grad_norm": 1.5249338150024414, "learning_rate": 7.614908963600834e-06, "loss": 0.4816, "step": 8232 }, { "epoch": 0.34511595736876854, "grad_norm": 1.566280722618103, "learning_rate": 7.614330323919948e-06, "loss": 0.5546, "step": 8233 }, { "epoch": 0.3451578759837775, "grad_norm": 1.4425058364868164, "learning_rate": 7.613751636047499e-06, "loss": 0.4512, "step": 8234 }, { "epoch": 0.34519979459878647, "grad_norm": 1.8072841167449951, "learning_rate": 7.613172899994154e-06, "loss": 0.5607, "step": 8235 }, { "epoch": 0.34524171321379543, "grad_norm": 1.7263909578323364, "learning_rate": 7.612594115770578e-06, "loss": 0.5352, "step": 8236 }, { "epoch": 0.3452836318288044, "grad_norm": 1.8720471858978271, "learning_rate": 7.61201528338744e-06, "loss": 0.5302, "step": 8237 }, { "epoch": 0.34532555044381336, "grad_norm": 1.6549899578094482, "learning_rate": 7.611436402855412e-06, "loss": 0.5178, "step": 8238 }, { "epoch": 0.3453674690588223, "grad_norm": 1.6235990524291992, "learning_rate": 7.610857474185165e-06, "loss": 0.4986, "step": 8239 }, { "epoch": 0.3454093876738313, "grad_norm": 1.8837950229644775, "learning_rate": 7.6102784973873685e-06, "loss": 0.5763, "step": 8240 }, { "epoch": 0.34545130628884024, "grad_norm": 1.8474103212356567, "learning_rate": 7.609699472472698e-06, "loss": 0.5607, "step": 8241 }, { "epoch": 0.34549322490384915, "grad_norm": 1.7221275568008423, "learning_rate": 7.609120399451824e-06, "loss": 0.5639, "step": 8242 }, { "epoch": 0.3455351435188581, "grad_norm": 1.5661431550979614, "learning_rate": 7.608541278335423e-06, "loss": 0.5529, "step": 8243 }, { "epoch": 0.3455770621338671, "grad_norm": 1.622973084449768, "learning_rate": 7.60796210913417e-06, "loss": 0.5217, "step": 8244 }, { "epoch": 0.34561898074887604, "grad_norm": 1.8717644214630127, "learning_rate": 7.607382891858741e-06, "loss": 0.5479, "step": 8245 }, { "epoch": 0.345660899363885, "grad_norm": 1.9703482389450073, "learning_rate": 7.606803626519812e-06, "loss": 0.5326, "step": 8246 }, { "epoch": 0.34570281797889396, "grad_norm": 1.6422762870788574, "learning_rate": 7.6062243131280635e-06, "loss": 0.537, "step": 8247 }, { "epoch": 0.3457447365939029, "grad_norm": 2.345860719680786, "learning_rate": 7.605644951694171e-06, "loss": 0.5526, "step": 8248 }, { "epoch": 0.3457866552089119, "grad_norm": 1.6762429475784302, "learning_rate": 7.6050655422288175e-06, "loss": 0.5732, "step": 8249 }, { "epoch": 0.34582857382392085, "grad_norm": 1.7173528671264648, "learning_rate": 7.604486084742682e-06, "loss": 0.5044, "step": 8250 }, { "epoch": 0.3458704924389298, "grad_norm": 1.513163685798645, "learning_rate": 7.603906579246445e-06, "loss": 0.5096, "step": 8251 }, { "epoch": 0.3459124110539388, "grad_norm": 1.8993096351623535, "learning_rate": 7.603327025750791e-06, "loss": 0.5504, "step": 8252 }, { "epoch": 0.34595432966894774, "grad_norm": 1.6074726581573486, "learning_rate": 7.602747424266401e-06, "loss": 0.5053, "step": 8253 }, { "epoch": 0.3459962482839567, "grad_norm": 1.7660164833068848, "learning_rate": 7.602167774803962e-06, "loss": 0.4633, "step": 8254 }, { "epoch": 0.34603816689896566, "grad_norm": 1.5861852169036865, "learning_rate": 7.601588077374158e-06, "loss": 0.5273, "step": 8255 }, { "epoch": 0.3460800855139746, "grad_norm": 1.5765345096588135, "learning_rate": 7.601008331987673e-06, "loss": 0.5108, "step": 8256 }, { "epoch": 0.3461220041289836, "grad_norm": 1.562506914138794, "learning_rate": 7.6004285386551945e-06, "loss": 0.4589, "step": 8257 }, { "epoch": 0.34616392274399255, "grad_norm": 1.351585030555725, "learning_rate": 7.599848697387412e-06, "loss": 0.4449, "step": 8258 }, { "epoch": 0.3462058413590015, "grad_norm": 1.709357738494873, "learning_rate": 7.599268808195014e-06, "loss": 0.5925, "step": 8259 }, { "epoch": 0.34624775997401047, "grad_norm": 1.5428142547607422, "learning_rate": 7.598688871088687e-06, "loss": 0.5246, "step": 8260 }, { "epoch": 0.34628967858901943, "grad_norm": 1.6499954462051392, "learning_rate": 7.598108886079123e-06, "loss": 0.5016, "step": 8261 }, { "epoch": 0.3463315972040284, "grad_norm": 1.9237895011901855, "learning_rate": 7.597528853177014e-06, "loss": 0.5155, "step": 8262 }, { "epoch": 0.34637351581903736, "grad_norm": 1.610098123550415, "learning_rate": 7.596948772393053e-06, "loss": 0.5699, "step": 8263 }, { "epoch": 0.3464154344340463, "grad_norm": 1.5314607620239258, "learning_rate": 7.5963686437379305e-06, "loss": 0.5088, "step": 8264 }, { "epoch": 0.3464573530490553, "grad_norm": 1.536771297454834, "learning_rate": 7.595788467222341e-06, "loss": 0.5005, "step": 8265 }, { "epoch": 0.34649927166406425, "grad_norm": 1.6671359539031982, "learning_rate": 7.59520824285698e-06, "loss": 0.5271, "step": 8266 }, { "epoch": 0.34654119027907315, "grad_norm": 1.8025734424591064, "learning_rate": 7.594627970652543e-06, "loss": 0.5456, "step": 8267 }, { "epoch": 0.3465831088940821, "grad_norm": 1.7564284801483154, "learning_rate": 7.594047650619727e-06, "loss": 0.515, "step": 8268 }, { "epoch": 0.3466250275090911, "grad_norm": 2.5334932804107666, "learning_rate": 7.593467282769228e-06, "loss": 0.5676, "step": 8269 }, { "epoch": 0.34666694612410004, "grad_norm": 1.5146286487579346, "learning_rate": 7.592886867111744e-06, "loss": 0.4816, "step": 8270 }, { "epoch": 0.346708864739109, "grad_norm": 1.7225337028503418, "learning_rate": 7.592306403657977e-06, "loss": 0.5378, "step": 8271 }, { "epoch": 0.34675078335411796, "grad_norm": 2.1768081188201904, "learning_rate": 7.591725892418624e-06, "loss": 0.5027, "step": 8272 }, { "epoch": 0.3467927019691269, "grad_norm": 1.6385939121246338, "learning_rate": 7.591145333404388e-06, "loss": 0.5511, "step": 8273 }, { "epoch": 0.3468346205841359, "grad_norm": 1.7549452781677246, "learning_rate": 7.590564726625968e-06, "loss": 0.5617, "step": 8274 }, { "epoch": 0.34687653919914485, "grad_norm": 1.4728903770446777, "learning_rate": 7.589984072094071e-06, "loss": 0.4505, "step": 8275 }, { "epoch": 0.3469184578141538, "grad_norm": 1.6521368026733398, "learning_rate": 7.589403369819397e-06, "loss": 0.5615, "step": 8276 }, { "epoch": 0.3469603764291628, "grad_norm": 2.127333641052246, "learning_rate": 7.588822619812653e-06, "loss": 0.5125, "step": 8277 }, { "epoch": 0.34700229504417174, "grad_norm": 1.7978380918502808, "learning_rate": 7.5882418220845414e-06, "loss": 0.5586, "step": 8278 }, { "epoch": 0.3470442136591807, "grad_norm": 1.7780420780181885, "learning_rate": 7.5876609766457696e-06, "loss": 0.5329, "step": 8279 }, { "epoch": 0.34708613227418966, "grad_norm": 1.6159113645553589, "learning_rate": 7.587080083507045e-06, "loss": 0.521, "step": 8280 }, { "epoch": 0.3471280508891986, "grad_norm": 1.793084740638733, "learning_rate": 7.586499142679077e-06, "loss": 0.5495, "step": 8281 }, { "epoch": 0.3471699695042076, "grad_norm": 2.700270652770996, "learning_rate": 7.585918154172572e-06, "loss": 0.5672, "step": 8282 }, { "epoch": 0.34721188811921655, "grad_norm": 1.7984987497329712, "learning_rate": 7.585337117998242e-06, "loss": 0.529, "step": 8283 }, { "epoch": 0.3472538067342255, "grad_norm": 1.5025495290756226, "learning_rate": 7.5847560341667945e-06, "loss": 0.5152, "step": 8284 }, { "epoch": 0.3472957253492345, "grad_norm": 1.4948604106903076, "learning_rate": 7.5841749026889435e-06, "loss": 0.5251, "step": 8285 }, { "epoch": 0.34733764396424344, "grad_norm": 1.68946373462677, "learning_rate": 7.583593723575401e-06, "loss": 0.5596, "step": 8286 }, { "epoch": 0.3473795625792524, "grad_norm": 1.7686439752578735, "learning_rate": 7.583012496836881e-06, "loss": 0.4851, "step": 8287 }, { "epoch": 0.34742148119426136, "grad_norm": 1.7182822227478027, "learning_rate": 7.582431222484095e-06, "loss": 0.5731, "step": 8288 }, { "epoch": 0.3474633998092703, "grad_norm": 1.3738923072814941, "learning_rate": 7.58184990052776e-06, "loss": 0.5187, "step": 8289 }, { "epoch": 0.3475053184242793, "grad_norm": 1.6856305599212646, "learning_rate": 7.581268530978591e-06, "loss": 0.554, "step": 8290 }, { "epoch": 0.34754723703928825, "grad_norm": 1.7500386238098145, "learning_rate": 7.580687113847305e-06, "loss": 0.5285, "step": 8291 }, { "epoch": 0.34758915565429715, "grad_norm": 1.7740861177444458, "learning_rate": 7.58010564914462e-06, "loss": 0.5695, "step": 8292 }, { "epoch": 0.3476310742693061, "grad_norm": 1.7353230714797974, "learning_rate": 7.579524136881256e-06, "loss": 0.51, "step": 8293 }, { "epoch": 0.3476729928843151, "grad_norm": 1.8985971212387085, "learning_rate": 7.578942577067927e-06, "loss": 0.5205, "step": 8294 }, { "epoch": 0.34771491149932404, "grad_norm": 1.4553675651550293, "learning_rate": 7.5783609697153595e-06, "loss": 0.5075, "step": 8295 }, { "epoch": 0.347756830114333, "grad_norm": 1.8469514846801758, "learning_rate": 7.577779314834273e-06, "loss": 0.5037, "step": 8296 }, { "epoch": 0.34779874872934197, "grad_norm": 1.641555905342102, "learning_rate": 7.577197612435385e-06, "loss": 0.5068, "step": 8297 }, { "epoch": 0.34784066734435093, "grad_norm": 1.8622792959213257, "learning_rate": 7.576615862529424e-06, "loss": 0.5319, "step": 8298 }, { "epoch": 0.3478825859593599, "grad_norm": 1.6249401569366455, "learning_rate": 7.576034065127111e-06, "loss": 0.5288, "step": 8299 }, { "epoch": 0.34792450457436885, "grad_norm": 1.7036020755767822, "learning_rate": 7.575452220239172e-06, "loss": 0.5684, "step": 8300 }, { "epoch": 0.3479664231893778, "grad_norm": 2.1263506412506104, "learning_rate": 7.574870327876332e-06, "loss": 0.5293, "step": 8301 }, { "epoch": 0.3480083418043868, "grad_norm": 1.9294700622558594, "learning_rate": 7.574288388049317e-06, "loss": 0.5306, "step": 8302 }, { "epoch": 0.34805026041939574, "grad_norm": 1.6652268171310425, "learning_rate": 7.573706400768855e-06, "loss": 0.5276, "step": 8303 }, { "epoch": 0.3480921790344047, "grad_norm": 1.6802752017974854, "learning_rate": 7.573124366045671e-06, "loss": 0.5029, "step": 8304 }, { "epoch": 0.34813409764941367, "grad_norm": 1.6346079111099243, "learning_rate": 7.5725422838904985e-06, "loss": 0.5461, "step": 8305 }, { "epoch": 0.3481760162644226, "grad_norm": 1.5534114837646484, "learning_rate": 7.571960154314066e-06, "loss": 0.5213, "step": 8306 }, { "epoch": 0.3482179348794316, "grad_norm": 1.626811146736145, "learning_rate": 7.571377977327101e-06, "loss": 0.4926, "step": 8307 }, { "epoch": 0.34825985349444055, "grad_norm": 1.7724521160125732, "learning_rate": 7.570795752940341e-06, "loss": 0.582, "step": 8308 }, { "epoch": 0.3483017721094495, "grad_norm": 1.8215994834899902, "learning_rate": 7.570213481164513e-06, "loss": 0.5741, "step": 8309 }, { "epoch": 0.3483436907244585, "grad_norm": 1.3890455961227417, "learning_rate": 7.569631162010354e-06, "loss": 0.4997, "step": 8310 }, { "epoch": 0.34838560933946744, "grad_norm": 1.7702305316925049, "learning_rate": 7.569048795488596e-06, "loss": 0.5032, "step": 8311 }, { "epoch": 0.3484275279544764, "grad_norm": 1.9009507894515991, "learning_rate": 7.568466381609976e-06, "loss": 0.5424, "step": 8312 }, { "epoch": 0.34846944656948536, "grad_norm": 1.7055095434188843, "learning_rate": 7.567883920385228e-06, "loss": 0.5262, "step": 8313 }, { "epoch": 0.3485113651844943, "grad_norm": 1.9382191896438599, "learning_rate": 7.567301411825091e-06, "loss": 0.5449, "step": 8314 }, { "epoch": 0.3485532837995033, "grad_norm": 1.711804986000061, "learning_rate": 7.5667188559403025e-06, "loss": 0.4757, "step": 8315 }, { "epoch": 0.34859520241451225, "grad_norm": 1.669567584991455, "learning_rate": 7.566136252741599e-06, "loss": 0.5393, "step": 8316 }, { "epoch": 0.34863712102952116, "grad_norm": 2.190016508102417, "learning_rate": 7.5655536022397206e-06, "loss": 0.5178, "step": 8317 }, { "epoch": 0.3486790396445301, "grad_norm": 1.5856181383132935, "learning_rate": 7.564970904445408e-06, "loss": 0.4972, "step": 8318 }, { "epoch": 0.3487209582595391, "grad_norm": 2.268277645111084, "learning_rate": 7.5643881593694045e-06, "loss": 0.5494, "step": 8319 }, { "epoch": 0.34876287687454804, "grad_norm": 1.7923442125320435, "learning_rate": 7.56380536702245e-06, "loss": 0.5715, "step": 8320 }, { "epoch": 0.348804795489557, "grad_norm": 1.6906561851501465, "learning_rate": 7.563222527415287e-06, "loss": 0.501, "step": 8321 }, { "epoch": 0.34884671410456597, "grad_norm": 1.7036343812942505, "learning_rate": 7.56263964055866e-06, "loss": 0.5144, "step": 8322 }, { "epoch": 0.34888863271957493, "grad_norm": 1.5970648527145386, "learning_rate": 7.562056706463315e-06, "loss": 0.4958, "step": 8323 }, { "epoch": 0.3489305513345839, "grad_norm": 1.6959649324417114, "learning_rate": 7.561473725139997e-06, "loss": 0.5645, "step": 8324 }, { "epoch": 0.34897246994959286, "grad_norm": 1.6533706188201904, "learning_rate": 7.560890696599451e-06, "loss": 0.5734, "step": 8325 }, { "epoch": 0.3490143885646018, "grad_norm": 2.341468095779419, "learning_rate": 7.560307620852425e-06, "loss": 0.4679, "step": 8326 }, { "epoch": 0.3490563071796108, "grad_norm": 2.2828900814056396, "learning_rate": 7.5597244979096685e-06, "loss": 0.5336, "step": 8327 }, { "epoch": 0.34909822579461974, "grad_norm": 1.5347468852996826, "learning_rate": 7.559141327781929e-06, "loss": 0.5283, "step": 8328 }, { "epoch": 0.3491401444096287, "grad_norm": 1.4968644380569458, "learning_rate": 7.558558110479959e-06, "loss": 0.507, "step": 8329 }, { "epoch": 0.34918206302463767, "grad_norm": 1.5850906372070312, "learning_rate": 7.557974846014505e-06, "loss": 0.486, "step": 8330 }, { "epoch": 0.34922398163964663, "grad_norm": 2.872774124145508, "learning_rate": 7.5573915343963215e-06, "loss": 0.5225, "step": 8331 }, { "epoch": 0.3492659002546556, "grad_norm": 1.5932012796401978, "learning_rate": 7.55680817563616e-06, "loss": 0.5499, "step": 8332 }, { "epoch": 0.34930781886966455, "grad_norm": 1.566162109375, "learning_rate": 7.556224769744776e-06, "loss": 0.4784, "step": 8333 }, { "epoch": 0.3493497374846735, "grad_norm": 1.5075414180755615, "learning_rate": 7.555641316732921e-06, "loss": 0.5012, "step": 8334 }, { "epoch": 0.3493916560996825, "grad_norm": 1.7920233011245728, "learning_rate": 7.555057816611352e-06, "loss": 0.5302, "step": 8335 }, { "epoch": 0.34943357471469144, "grad_norm": 1.6601473093032837, "learning_rate": 7.5544742693908235e-06, "loss": 0.5177, "step": 8336 }, { "epoch": 0.3494754933297004, "grad_norm": 1.6333858966827393, "learning_rate": 7.553890675082095e-06, "loss": 0.5397, "step": 8337 }, { "epoch": 0.34951741194470937, "grad_norm": 1.7154399156570435, "learning_rate": 7.553307033695921e-06, "loss": 0.5471, "step": 8338 }, { "epoch": 0.34955933055971833, "grad_norm": 1.9946786165237427, "learning_rate": 7.552723345243063e-06, "loss": 0.5128, "step": 8339 }, { "epoch": 0.3496012491747273, "grad_norm": 1.585091471672058, "learning_rate": 7.552139609734279e-06, "loss": 0.4648, "step": 8340 }, { "epoch": 0.34964316778973625, "grad_norm": 1.8481581211090088, "learning_rate": 7.551555827180328e-06, "loss": 0.4868, "step": 8341 }, { "epoch": 0.34968508640474516, "grad_norm": 1.6488816738128662, "learning_rate": 7.550971997591974e-06, "loss": 0.5133, "step": 8342 }, { "epoch": 0.3497270050197541, "grad_norm": 1.5050957202911377, "learning_rate": 7.550388120979978e-06, "loss": 0.4789, "step": 8343 }, { "epoch": 0.3497689236347631, "grad_norm": 1.7761811017990112, "learning_rate": 7.549804197355102e-06, "loss": 0.5417, "step": 8344 }, { "epoch": 0.34981084224977205, "grad_norm": 1.7279798984527588, "learning_rate": 7.549220226728111e-06, "loss": 0.5536, "step": 8345 }, { "epoch": 0.349852760864781, "grad_norm": 1.8951148986816406, "learning_rate": 7.548636209109769e-06, "loss": 0.5266, "step": 8346 }, { "epoch": 0.34989467947978997, "grad_norm": 1.7443454265594482, "learning_rate": 7.548052144510842e-06, "loss": 0.5593, "step": 8347 }, { "epoch": 0.34993659809479893, "grad_norm": 1.573134183883667, "learning_rate": 7.547468032942096e-06, "loss": 0.5286, "step": 8348 }, { "epoch": 0.3499785167098079, "grad_norm": 1.5610711574554443, "learning_rate": 7.546883874414299e-06, "loss": 0.5222, "step": 8349 }, { "epoch": 0.35002043532481686, "grad_norm": 1.7022508382797241, "learning_rate": 7.546299668938219e-06, "loss": 0.5743, "step": 8350 }, { "epoch": 0.3500623539398258, "grad_norm": 1.674992322921753, "learning_rate": 7.545715416524624e-06, "loss": 0.5647, "step": 8351 }, { "epoch": 0.3501042725548348, "grad_norm": 1.9305956363677979, "learning_rate": 7.545131117184286e-06, "loss": 0.5886, "step": 8352 }, { "epoch": 0.35014619116984375, "grad_norm": 1.882353663444519, "learning_rate": 7.544546770927973e-06, "loss": 0.4931, "step": 8353 }, { "epoch": 0.3501881097848527, "grad_norm": 1.5597444772720337, "learning_rate": 7.543962377766458e-06, "loss": 0.4492, "step": 8354 }, { "epoch": 0.35023002839986167, "grad_norm": 1.7213134765625, "learning_rate": 7.5433779377105145e-06, "loss": 0.5253, "step": 8355 }, { "epoch": 0.35027194701487063, "grad_norm": 1.8179292678833008, "learning_rate": 7.542793450770913e-06, "loss": 0.516, "step": 8356 }, { "epoch": 0.3503138656298796, "grad_norm": 1.5621333122253418, "learning_rate": 7.542208916958433e-06, "loss": 0.5396, "step": 8357 }, { "epoch": 0.35035578424488856, "grad_norm": 1.5031664371490479, "learning_rate": 7.541624336283843e-06, "loss": 0.5314, "step": 8358 }, { "epoch": 0.3503977028598975, "grad_norm": 1.854310154914856, "learning_rate": 7.5410397087579225e-06, "loss": 0.5042, "step": 8359 }, { "epoch": 0.3504396214749065, "grad_norm": 1.7428995370864868, "learning_rate": 7.54045503439145e-06, "loss": 0.5383, "step": 8360 }, { "epoch": 0.35048154008991544, "grad_norm": 1.6764779090881348, "learning_rate": 7.5398703131951985e-06, "loss": 0.5375, "step": 8361 }, { "epoch": 0.3505234587049244, "grad_norm": 2.014472007751465, "learning_rate": 7.539285545179951e-06, "loss": 0.499, "step": 8362 }, { "epoch": 0.35056537731993337, "grad_norm": 1.8003720045089722, "learning_rate": 7.538700730356483e-06, "loss": 0.5262, "step": 8363 }, { "epoch": 0.35060729593494233, "grad_norm": 1.531629204750061, "learning_rate": 7.538115868735578e-06, "loss": 0.4862, "step": 8364 }, { "epoch": 0.3506492145499513, "grad_norm": 1.613426923751831, "learning_rate": 7.537530960328014e-06, "loss": 0.5226, "step": 8365 }, { "epoch": 0.35069113316496026, "grad_norm": 1.7102850675582886, "learning_rate": 7.536946005144577e-06, "loss": 0.5446, "step": 8366 }, { "epoch": 0.35073305177996916, "grad_norm": 1.775202751159668, "learning_rate": 7.536361003196048e-06, "loss": 0.5299, "step": 8367 }, { "epoch": 0.3507749703949781, "grad_norm": 1.8779727220535278, "learning_rate": 7.5357759544932075e-06, "loss": 0.5369, "step": 8368 }, { "epoch": 0.3508168890099871, "grad_norm": 1.6938384771347046, "learning_rate": 7.535190859046845e-06, "loss": 0.5252, "step": 8369 }, { "epoch": 0.35085880762499605, "grad_norm": 1.7612637281417847, "learning_rate": 7.5346057168677436e-06, "loss": 0.492, "step": 8370 }, { "epoch": 0.350900726240005, "grad_norm": 2.0219202041625977, "learning_rate": 7.534020527966689e-06, "loss": 0.5104, "step": 8371 }, { "epoch": 0.350942644855014, "grad_norm": 1.8842922449111938, "learning_rate": 7.533435292354469e-06, "loss": 0.528, "step": 8372 }, { "epoch": 0.35098456347002294, "grad_norm": 1.869471549987793, "learning_rate": 7.532850010041873e-06, "loss": 0.5377, "step": 8373 }, { "epoch": 0.3510264820850319, "grad_norm": 1.7553399801254272, "learning_rate": 7.5322646810396885e-06, "loss": 0.5826, "step": 8374 }, { "epoch": 0.35106840070004086, "grad_norm": 1.8535676002502441, "learning_rate": 7.531679305358706e-06, "loss": 0.5314, "step": 8375 }, { "epoch": 0.3511103193150498, "grad_norm": 1.7530345916748047, "learning_rate": 7.5310938830097135e-06, "loss": 0.5281, "step": 8376 }, { "epoch": 0.3511522379300588, "grad_norm": 1.9540272951126099, "learning_rate": 7.530508414003505e-06, "loss": 0.5091, "step": 8377 }, { "epoch": 0.35119415654506775, "grad_norm": 1.5204054117202759, "learning_rate": 7.529922898350872e-06, "loss": 0.5201, "step": 8378 }, { "epoch": 0.3512360751600767, "grad_norm": 1.60133957862854, "learning_rate": 7.5293373360626076e-06, "loss": 0.4983, "step": 8379 }, { "epoch": 0.3512779937750857, "grad_norm": 1.8018083572387695, "learning_rate": 7.528751727149508e-06, "loss": 0.5237, "step": 8380 }, { "epoch": 0.35131991239009464, "grad_norm": 1.92623770236969, "learning_rate": 7.528166071622365e-06, "loss": 0.5516, "step": 8381 }, { "epoch": 0.3513618310051036, "grad_norm": 1.624111294746399, "learning_rate": 7.527580369491976e-06, "loss": 0.5796, "step": 8382 }, { "epoch": 0.35140374962011256, "grad_norm": 1.6268737316131592, "learning_rate": 7.526994620769137e-06, "loss": 0.5189, "step": 8383 }, { "epoch": 0.3514456682351215, "grad_norm": 2.1459155082702637, "learning_rate": 7.526408825464644e-06, "loss": 0.4909, "step": 8384 }, { "epoch": 0.3514875868501305, "grad_norm": 2.03505539894104, "learning_rate": 7.5258229835892995e-06, "loss": 0.5027, "step": 8385 }, { "epoch": 0.35152950546513945, "grad_norm": 2.204587697982788, "learning_rate": 7.525237095153899e-06, "loss": 0.5353, "step": 8386 }, { "epoch": 0.3515714240801484, "grad_norm": 1.4769781827926636, "learning_rate": 7.524651160169243e-06, "loss": 0.5173, "step": 8387 }, { "epoch": 0.35161334269515737, "grad_norm": 1.9227293729782104, "learning_rate": 7.524065178646134e-06, "loss": 0.4938, "step": 8388 }, { "epoch": 0.35165526131016633, "grad_norm": 2.1857669353485107, "learning_rate": 7.523479150595372e-06, "loss": 0.5597, "step": 8389 }, { "epoch": 0.3516971799251753, "grad_norm": 1.5460219383239746, "learning_rate": 7.522893076027762e-06, "loss": 0.4865, "step": 8390 }, { "epoch": 0.35173909854018426, "grad_norm": 1.663788914680481, "learning_rate": 7.522306954954104e-06, "loss": 0.4595, "step": 8391 }, { "epoch": 0.35178101715519317, "grad_norm": 1.5825958251953125, "learning_rate": 7.521720787385205e-06, "loss": 0.514, "step": 8392 }, { "epoch": 0.3518229357702021, "grad_norm": 1.7748098373413086, "learning_rate": 7.52113457333187e-06, "loss": 0.5006, "step": 8393 }, { "epoch": 0.3518648543852111, "grad_norm": 1.96825110912323, "learning_rate": 7.520548312804905e-06, "loss": 0.5184, "step": 8394 }, { "epoch": 0.35190677300022005, "grad_norm": 1.745904803276062, "learning_rate": 7.519962005815116e-06, "loss": 0.54, "step": 8395 }, { "epoch": 0.351948691615229, "grad_norm": 2.0815696716308594, "learning_rate": 7.51937565237331e-06, "loss": 0.5268, "step": 8396 }, { "epoch": 0.351990610230238, "grad_norm": 1.756770372390747, "learning_rate": 7.518789252490297e-06, "loss": 0.5027, "step": 8397 }, { "epoch": 0.35203252884524694, "grad_norm": 1.6720554828643799, "learning_rate": 7.518202806176887e-06, "loss": 0.4921, "step": 8398 }, { "epoch": 0.3520744474602559, "grad_norm": 3.364011764526367, "learning_rate": 7.517616313443889e-06, "loss": 0.5678, "step": 8399 }, { "epoch": 0.35211636607526486, "grad_norm": 1.83645761013031, "learning_rate": 7.5170297743021155e-06, "loss": 0.5155, "step": 8400 }, { "epoch": 0.3521582846902738, "grad_norm": 1.7706762552261353, "learning_rate": 7.516443188762378e-06, "loss": 0.4907, "step": 8401 }, { "epoch": 0.3522002033052828, "grad_norm": 1.8199044466018677, "learning_rate": 7.515856556835488e-06, "loss": 0.5151, "step": 8402 }, { "epoch": 0.35224212192029175, "grad_norm": 2.655482769012451, "learning_rate": 7.515269878532261e-06, "loss": 0.4923, "step": 8403 }, { "epoch": 0.3522840405353007, "grad_norm": 1.7384055852890015, "learning_rate": 7.514683153863512e-06, "loss": 0.5433, "step": 8404 }, { "epoch": 0.3523259591503097, "grad_norm": 1.894365668296814, "learning_rate": 7.514096382840055e-06, "loss": 0.5157, "step": 8405 }, { "epoch": 0.35236787776531864, "grad_norm": 1.9047847986221313, "learning_rate": 7.513509565472705e-06, "loss": 0.5318, "step": 8406 }, { "epoch": 0.3524097963803276, "grad_norm": 1.8585704565048218, "learning_rate": 7.512922701772283e-06, "loss": 0.5139, "step": 8407 }, { "epoch": 0.35245171499533656, "grad_norm": 2.0066239833831787, "learning_rate": 7.5123357917496055e-06, "loss": 0.5279, "step": 8408 }, { "epoch": 0.3524936336103455, "grad_norm": 1.83695650100708, "learning_rate": 7.511748835415489e-06, "loss": 0.4965, "step": 8409 }, { "epoch": 0.3525355522253545, "grad_norm": 2.5521206855773926, "learning_rate": 7.511161832780756e-06, "loss": 0.5122, "step": 8410 }, { "epoch": 0.35257747084036345, "grad_norm": 1.7746175527572632, "learning_rate": 7.510574783856226e-06, "loss": 0.4765, "step": 8411 }, { "epoch": 0.3526193894553724, "grad_norm": 1.9349075555801392, "learning_rate": 7.50998768865272e-06, "loss": 0.5692, "step": 8412 }, { "epoch": 0.3526613080703814, "grad_norm": 2.0342905521392822, "learning_rate": 7.509400547181063e-06, "loss": 0.4807, "step": 8413 }, { "epoch": 0.35270322668539034, "grad_norm": 1.5098462104797363, "learning_rate": 7.508813359452075e-06, "loss": 0.4999, "step": 8414 }, { "epoch": 0.3527451453003993, "grad_norm": 1.846049189567566, "learning_rate": 7.508226125476581e-06, "loss": 0.5186, "step": 8415 }, { "epoch": 0.35278706391540826, "grad_norm": 1.9642055034637451, "learning_rate": 7.507638845265405e-06, "loss": 0.5316, "step": 8416 }, { "epoch": 0.35282898253041717, "grad_norm": 2.2477126121520996, "learning_rate": 7.507051518829373e-06, "loss": 0.4963, "step": 8417 }, { "epoch": 0.35287090114542613, "grad_norm": 1.8157463073730469, "learning_rate": 7.506464146179315e-06, "loss": 0.51, "step": 8418 }, { "epoch": 0.3529128197604351, "grad_norm": 2.0167055130004883, "learning_rate": 7.505876727326054e-06, "loss": 0.5743, "step": 8419 }, { "epoch": 0.35295473837544405, "grad_norm": 1.6600736379623413, "learning_rate": 7.50528926228042e-06, "loss": 0.506, "step": 8420 }, { "epoch": 0.352996656990453, "grad_norm": 1.7533234357833862, "learning_rate": 7.504701751053241e-06, "loss": 0.5055, "step": 8421 }, { "epoch": 0.353038575605462, "grad_norm": 2.5140156745910645, "learning_rate": 7.504114193655348e-06, "loss": 0.5363, "step": 8422 }, { "epoch": 0.35308049422047094, "grad_norm": 1.8307631015777588, "learning_rate": 7.503526590097574e-06, "loss": 0.5695, "step": 8423 }, { "epoch": 0.3531224128354799, "grad_norm": 2.085214853286743, "learning_rate": 7.502938940390745e-06, "loss": 0.4937, "step": 8424 }, { "epoch": 0.35316433145048887, "grad_norm": 2.009167194366455, "learning_rate": 7.502351244545698e-06, "loss": 0.4844, "step": 8425 }, { "epoch": 0.35320625006549783, "grad_norm": 2.2794878482818604, "learning_rate": 7.5017635025732675e-06, "loss": 0.5537, "step": 8426 }, { "epoch": 0.3532481686805068, "grad_norm": 1.714707374572754, "learning_rate": 7.501175714484283e-06, "loss": 0.5306, "step": 8427 }, { "epoch": 0.35329008729551575, "grad_norm": 1.7317675352096558, "learning_rate": 7.500587880289583e-06, "loss": 0.5355, "step": 8428 }, { "epoch": 0.3533320059105247, "grad_norm": 2.0346999168395996, "learning_rate": 7.500000000000001e-06, "loss": 0.5529, "step": 8429 }, { "epoch": 0.3533739245255337, "grad_norm": 1.7257214784622192, "learning_rate": 7.4994120736263756e-06, "loss": 0.5275, "step": 8430 }, { "epoch": 0.35341584314054264, "grad_norm": 1.8911842107772827, "learning_rate": 7.498824101179545e-06, "loss": 0.5339, "step": 8431 }, { "epoch": 0.3534577617555516, "grad_norm": 1.8839318752288818, "learning_rate": 7.498236082670346e-06, "loss": 0.5093, "step": 8432 }, { "epoch": 0.35349968037056057, "grad_norm": 2.185840368270874, "learning_rate": 7.497648018109618e-06, "loss": 0.5578, "step": 8433 }, { "epoch": 0.3535415989855695, "grad_norm": 1.5750572681427002, "learning_rate": 7.497059907508202e-06, "loss": 0.5472, "step": 8434 }, { "epoch": 0.3535835176005785, "grad_norm": 1.8449749946594238, "learning_rate": 7.4964717508769395e-06, "loss": 0.5257, "step": 8435 }, { "epoch": 0.35362543621558745, "grad_norm": 2.342434883117676, "learning_rate": 7.49588354822667e-06, "loss": 0.5177, "step": 8436 }, { "epoch": 0.3536673548305964, "grad_norm": 1.6959607601165771, "learning_rate": 7.49529529956824e-06, "loss": 0.5326, "step": 8437 }, { "epoch": 0.3537092734456054, "grad_norm": 1.798887014389038, "learning_rate": 7.494707004912489e-06, "loss": 0.5548, "step": 8438 }, { "epoch": 0.35375119206061434, "grad_norm": 1.5647141933441162, "learning_rate": 7.494118664270263e-06, "loss": 0.5233, "step": 8439 }, { "epoch": 0.3537931106756233, "grad_norm": 1.6291677951812744, "learning_rate": 7.493530277652408e-06, "loss": 0.513, "step": 8440 }, { "epoch": 0.35383502929063226, "grad_norm": 1.6030620336532593, "learning_rate": 7.492941845069771e-06, "loss": 0.5132, "step": 8441 }, { "epoch": 0.35387694790564117, "grad_norm": 1.569390058517456, "learning_rate": 7.492353366533195e-06, "loss": 0.5009, "step": 8442 }, { "epoch": 0.35391886652065013, "grad_norm": 1.9229286909103394, "learning_rate": 7.491764842053532e-06, "loss": 0.5496, "step": 8443 }, { "epoch": 0.3539607851356591, "grad_norm": 1.8604069948196411, "learning_rate": 7.491176271641628e-06, "loss": 0.5335, "step": 8444 }, { "epoch": 0.35400270375066806, "grad_norm": 1.6501314640045166, "learning_rate": 7.490587655308333e-06, "loss": 0.5198, "step": 8445 }, { "epoch": 0.354044622365677, "grad_norm": 1.6212624311447144, "learning_rate": 7.489998993064499e-06, "loss": 0.503, "step": 8446 }, { "epoch": 0.354086540980686, "grad_norm": 2.2147023677825928, "learning_rate": 7.489410284920975e-06, "loss": 0.5576, "step": 8447 }, { "epoch": 0.35412845959569494, "grad_norm": 1.855004906654358, "learning_rate": 7.488821530888615e-06, "loss": 0.5626, "step": 8448 }, { "epoch": 0.3541703782107039, "grad_norm": 1.5652443170547485, "learning_rate": 7.4882327309782685e-06, "loss": 0.4339, "step": 8449 }, { "epoch": 0.35421229682571287, "grad_norm": 1.9793199300765991, "learning_rate": 7.4876438852007925e-06, "loss": 0.5064, "step": 8450 }, { "epoch": 0.35425421544072183, "grad_norm": 1.7708367109298706, "learning_rate": 7.487054993567042e-06, "loss": 0.5078, "step": 8451 }, { "epoch": 0.3542961340557308, "grad_norm": 1.6150413751602173, "learning_rate": 7.48646605608787e-06, "loss": 0.5294, "step": 8452 }, { "epoch": 0.35433805267073976, "grad_norm": 1.689150333404541, "learning_rate": 7.4858770727741335e-06, "loss": 0.5386, "step": 8453 }, { "epoch": 0.3543799712857487, "grad_norm": 1.570388674736023, "learning_rate": 7.485288043636689e-06, "loss": 0.4963, "step": 8454 }, { "epoch": 0.3544218899007577, "grad_norm": 1.879534125328064, "learning_rate": 7.484698968686397e-06, "loss": 0.5462, "step": 8455 }, { "epoch": 0.35446380851576664, "grad_norm": 1.953555941581726, "learning_rate": 7.484109847934114e-06, "loss": 0.5481, "step": 8456 }, { "epoch": 0.3545057271307756, "grad_norm": 1.870882511138916, "learning_rate": 7.483520681390699e-06, "loss": 0.5343, "step": 8457 }, { "epoch": 0.35454764574578457, "grad_norm": 1.5889757871627808, "learning_rate": 7.482931469067014e-06, "loss": 0.5068, "step": 8458 }, { "epoch": 0.35458956436079353, "grad_norm": 1.826320767402649, "learning_rate": 7.482342210973921e-06, "loss": 0.5702, "step": 8459 }, { "epoch": 0.3546314829758025, "grad_norm": 1.881026268005371, "learning_rate": 7.481752907122281e-06, "loss": 0.5642, "step": 8460 }, { "epoch": 0.35467340159081145, "grad_norm": 1.5386677980422974, "learning_rate": 7.481163557522956e-06, "loss": 0.4562, "step": 8461 }, { "epoch": 0.3547153202058204, "grad_norm": 1.568254828453064, "learning_rate": 7.4805741621868115e-06, "loss": 0.4999, "step": 8462 }, { "epoch": 0.3547572388208294, "grad_norm": 1.6677501201629639, "learning_rate": 7.4799847211247115e-06, "loss": 0.4968, "step": 8463 }, { "epoch": 0.35479915743583834, "grad_norm": 1.6102367639541626, "learning_rate": 7.479395234347521e-06, "loss": 0.4819, "step": 8464 }, { "epoch": 0.3548410760508473, "grad_norm": 1.7742575407028198, "learning_rate": 7.478805701866108e-06, "loss": 0.4903, "step": 8465 }, { "epoch": 0.35488299466585627, "grad_norm": 1.7037091255187988, "learning_rate": 7.478216123691337e-06, "loss": 0.5126, "step": 8466 }, { "epoch": 0.3549249132808652, "grad_norm": 3.035947561264038, "learning_rate": 7.47762649983408e-06, "loss": 0.5511, "step": 8467 }, { "epoch": 0.35496683189587414, "grad_norm": 1.8149341344833374, "learning_rate": 7.477036830305202e-06, "loss": 0.4972, "step": 8468 }, { "epoch": 0.3550087505108831, "grad_norm": 2.315521717071533, "learning_rate": 7.476447115115576e-06, "loss": 0.5383, "step": 8469 }, { "epoch": 0.35505066912589206, "grad_norm": 1.9016094207763672, "learning_rate": 7.47585735427607e-06, "loss": 0.5513, "step": 8470 }, { "epoch": 0.355092587740901, "grad_norm": 1.8241761922836304, "learning_rate": 7.475267547797556e-06, "loss": 0.5092, "step": 8471 }, { "epoch": 0.35513450635591, "grad_norm": 1.5261861085891724, "learning_rate": 7.4746776956909085e-06, "loss": 0.5085, "step": 8472 }, { "epoch": 0.35517642497091895, "grad_norm": 3.535627841949463, "learning_rate": 7.474087797966997e-06, "loss": 0.5, "step": 8473 }, { "epoch": 0.3552183435859279, "grad_norm": 1.7399557828903198, "learning_rate": 7.473497854636699e-06, "loss": 0.5501, "step": 8474 }, { "epoch": 0.35526026220093687, "grad_norm": 1.6931260824203491, "learning_rate": 7.472907865710887e-06, "loss": 0.5555, "step": 8475 }, { "epoch": 0.35530218081594583, "grad_norm": 1.526708722114563, "learning_rate": 7.4723178312004355e-06, "loss": 0.5268, "step": 8476 }, { "epoch": 0.3553440994309548, "grad_norm": 1.9917023181915283, "learning_rate": 7.4717277511162235e-06, "loss": 0.5427, "step": 8477 }, { "epoch": 0.35538601804596376, "grad_norm": 1.5468136072158813, "learning_rate": 7.471137625469128e-06, "loss": 0.4804, "step": 8478 }, { "epoch": 0.3554279366609727, "grad_norm": 1.4805717468261719, "learning_rate": 7.470547454270028e-06, "loss": 0.4561, "step": 8479 }, { "epoch": 0.3554698552759817, "grad_norm": 1.546186089515686, "learning_rate": 7.469957237529799e-06, "loss": 0.5513, "step": 8480 }, { "epoch": 0.35551177389099065, "grad_norm": 1.5694540739059448, "learning_rate": 7.4693669752593235e-06, "loss": 0.4544, "step": 8481 }, { "epoch": 0.3555536925059996, "grad_norm": 1.4210174083709717, "learning_rate": 7.4687766674694804e-06, "loss": 0.5506, "step": 8482 }, { "epoch": 0.35559561112100857, "grad_norm": 1.8119163513183594, "learning_rate": 7.468186314171155e-06, "loss": 0.5365, "step": 8483 }, { "epoch": 0.35563752973601753, "grad_norm": 2.469650983810425, "learning_rate": 7.467595915375225e-06, "loss": 0.4843, "step": 8484 }, { "epoch": 0.3556794483510265, "grad_norm": 1.8497462272644043, "learning_rate": 7.467005471092576e-06, "loss": 0.5335, "step": 8485 }, { "epoch": 0.35572136696603546, "grad_norm": 1.6833539009094238, "learning_rate": 7.466414981334093e-06, "loss": 0.5111, "step": 8486 }, { "epoch": 0.3557632855810444, "grad_norm": 1.7305735349655151, "learning_rate": 7.465824446110658e-06, "loss": 0.5219, "step": 8487 }, { "epoch": 0.3558052041960534, "grad_norm": 1.9849926233291626, "learning_rate": 7.46523386543316e-06, "loss": 0.5438, "step": 8488 }, { "epoch": 0.35584712281106234, "grad_norm": 1.5413521528244019, "learning_rate": 7.464643239312481e-06, "loss": 0.5422, "step": 8489 }, { "epoch": 0.3558890414260713, "grad_norm": 1.9575291872024536, "learning_rate": 7.464052567759513e-06, "loss": 0.5102, "step": 8490 }, { "epoch": 0.35593096004108027, "grad_norm": 1.9140000343322754, "learning_rate": 7.463461850785143e-06, "loss": 0.5798, "step": 8491 }, { "epoch": 0.3559728786560892, "grad_norm": 1.643080234527588, "learning_rate": 7.462871088400259e-06, "loss": 0.4987, "step": 8492 }, { "epoch": 0.35601479727109814, "grad_norm": 1.4641673564910889, "learning_rate": 7.46228028061575e-06, "loss": 0.4597, "step": 8493 }, { "epoch": 0.3560567158861071, "grad_norm": 1.6221392154693604, "learning_rate": 7.461689427442509e-06, "loss": 0.5, "step": 8494 }, { "epoch": 0.35609863450111606, "grad_norm": 1.506475806236267, "learning_rate": 7.461098528891426e-06, "loss": 0.5004, "step": 8495 }, { "epoch": 0.356140553116125, "grad_norm": 1.494985580444336, "learning_rate": 7.4605075849733946e-06, "loss": 0.5418, "step": 8496 }, { "epoch": 0.356182471731134, "grad_norm": 1.4249857664108276, "learning_rate": 7.459916595699307e-06, "loss": 0.472, "step": 8497 }, { "epoch": 0.35622439034614295, "grad_norm": 1.952214002609253, "learning_rate": 7.459325561080057e-06, "loss": 0.526, "step": 8498 }, { "epoch": 0.3562663089611519, "grad_norm": 1.5456613302230835, "learning_rate": 7.458734481126543e-06, "loss": 0.4554, "step": 8499 }, { "epoch": 0.3563082275761609, "grad_norm": 1.497374415397644, "learning_rate": 7.458143355849655e-06, "loss": 0.4871, "step": 8500 }, { "epoch": 0.35635014619116984, "grad_norm": 1.598807454109192, "learning_rate": 7.4575521852602924e-06, "loss": 0.5977, "step": 8501 }, { "epoch": 0.3563920648061788, "grad_norm": 1.6542829275131226, "learning_rate": 7.456960969369356e-06, "loss": 0.5111, "step": 8502 }, { "epoch": 0.35643398342118776, "grad_norm": 1.7278553247451782, "learning_rate": 7.456369708187738e-06, "loss": 0.5431, "step": 8503 }, { "epoch": 0.3564759020361967, "grad_norm": 2.4303245544433594, "learning_rate": 7.4557784017263415e-06, "loss": 0.5452, "step": 8504 }, { "epoch": 0.3565178206512057, "grad_norm": 1.6086161136627197, "learning_rate": 7.455187049996064e-06, "loss": 0.5366, "step": 8505 }, { "epoch": 0.35655973926621465, "grad_norm": 1.4352431297302246, "learning_rate": 7.454595653007808e-06, "loss": 0.51, "step": 8506 }, { "epoch": 0.3566016578812236, "grad_norm": 1.9033925533294678, "learning_rate": 7.4540042107724745e-06, "loss": 0.498, "step": 8507 }, { "epoch": 0.3566435764962326, "grad_norm": 1.5760482549667358, "learning_rate": 7.453412723300967e-06, "loss": 0.6041, "step": 8508 }, { "epoch": 0.35668549511124154, "grad_norm": 1.4895042181015015, "learning_rate": 7.452821190604186e-06, "loss": 0.5509, "step": 8509 }, { "epoch": 0.3567274137262505, "grad_norm": 1.3762904405593872, "learning_rate": 7.452229612693037e-06, "loss": 0.4701, "step": 8510 }, { "epoch": 0.35676933234125946, "grad_norm": 1.870509147644043, "learning_rate": 7.451637989578427e-06, "loss": 0.5793, "step": 8511 }, { "epoch": 0.3568112509562684, "grad_norm": 1.5008376836776733, "learning_rate": 7.45104632127126e-06, "loss": 0.5071, "step": 8512 }, { "epoch": 0.3568531695712774, "grad_norm": 1.5513367652893066, "learning_rate": 7.45045460778244e-06, "loss": 0.5012, "step": 8513 }, { "epoch": 0.35689508818628635, "grad_norm": 1.6114425659179688, "learning_rate": 7.44986284912288e-06, "loss": 0.5161, "step": 8514 }, { "epoch": 0.3569370068012953, "grad_norm": 1.8249748945236206, "learning_rate": 7.449271045303483e-06, "loss": 0.5435, "step": 8515 }, { "epoch": 0.35697892541630427, "grad_norm": 4.363615036010742, "learning_rate": 7.4486791963351615e-06, "loss": 0.5695, "step": 8516 }, { "epoch": 0.3570208440313132, "grad_norm": 1.5721609592437744, "learning_rate": 7.448087302228823e-06, "loss": 0.577, "step": 8517 }, { "epoch": 0.35706276264632214, "grad_norm": 1.478848934173584, "learning_rate": 7.447495362995382e-06, "loss": 0.4885, "step": 8518 }, { "epoch": 0.3571046812613311, "grad_norm": 1.5464057922363281, "learning_rate": 7.446903378645745e-06, "loss": 0.4986, "step": 8519 }, { "epoch": 0.35714659987634007, "grad_norm": 1.7765628099441528, "learning_rate": 7.446311349190827e-06, "loss": 0.5117, "step": 8520 }, { "epoch": 0.357188518491349, "grad_norm": 1.775508999824524, "learning_rate": 7.445719274641543e-06, "loss": 0.5397, "step": 8521 }, { "epoch": 0.357230437106358, "grad_norm": 1.6054675579071045, "learning_rate": 7.445127155008804e-06, "loss": 0.4936, "step": 8522 }, { "epoch": 0.35727235572136695, "grad_norm": 1.6431087255477905, "learning_rate": 7.4445349903035255e-06, "loss": 0.549, "step": 8523 }, { "epoch": 0.3573142743363759, "grad_norm": 2.0286529064178467, "learning_rate": 7.443942780536626e-06, "loss": 0.521, "step": 8524 }, { "epoch": 0.3573561929513849, "grad_norm": 1.4776570796966553, "learning_rate": 7.443350525719019e-06, "loss": 0.5409, "step": 8525 }, { "epoch": 0.35739811156639384, "grad_norm": 1.5546640157699585, "learning_rate": 7.442758225861623e-06, "loss": 0.5466, "step": 8526 }, { "epoch": 0.3574400301814028, "grad_norm": 1.6087273359298706, "learning_rate": 7.442165880975356e-06, "loss": 0.5156, "step": 8527 }, { "epoch": 0.35748194879641176, "grad_norm": 1.6670209169387817, "learning_rate": 7.441573491071137e-06, "loss": 0.4929, "step": 8528 }, { "epoch": 0.3575238674114207, "grad_norm": 1.5762076377868652, "learning_rate": 7.4409810561598846e-06, "loss": 0.4848, "step": 8529 }, { "epoch": 0.3575657860264297, "grad_norm": 1.7168550491333008, "learning_rate": 7.440388576252524e-06, "loss": 0.6116, "step": 8530 }, { "epoch": 0.35760770464143865, "grad_norm": 1.5644288063049316, "learning_rate": 7.439796051359972e-06, "loss": 0.5506, "step": 8531 }, { "epoch": 0.3576496232564476, "grad_norm": 1.5760948657989502, "learning_rate": 7.439203481493152e-06, "loss": 0.537, "step": 8532 }, { "epoch": 0.3576915418714566, "grad_norm": 1.671749234199524, "learning_rate": 7.43861086666299e-06, "loss": 0.5183, "step": 8533 }, { "epoch": 0.35773346048646554, "grad_norm": 1.6374313831329346, "learning_rate": 7.438018206880406e-06, "loss": 0.5557, "step": 8534 }, { "epoch": 0.3577753791014745, "grad_norm": 1.7872016429901123, "learning_rate": 7.437425502156328e-06, "loss": 0.5775, "step": 8535 }, { "epoch": 0.35781729771648346, "grad_norm": 1.6459167003631592, "learning_rate": 7.4368327525016806e-06, "loss": 0.5174, "step": 8536 }, { "epoch": 0.3578592163314924, "grad_norm": 1.8849847316741943, "learning_rate": 7.436239957927389e-06, "loss": 0.5488, "step": 8537 }, { "epoch": 0.3579011349465014, "grad_norm": 1.5702152252197266, "learning_rate": 7.435647118444383e-06, "loss": 0.531, "step": 8538 }, { "epoch": 0.35794305356151035, "grad_norm": 1.4818432331085205, "learning_rate": 7.435054234063588e-06, "loss": 0.4822, "step": 8539 }, { "epoch": 0.3579849721765193, "grad_norm": 1.8040592670440674, "learning_rate": 7.4344613047959365e-06, "loss": 0.5597, "step": 8540 }, { "epoch": 0.3580268907915283, "grad_norm": 1.4830783605575562, "learning_rate": 7.433868330652355e-06, "loss": 0.4588, "step": 8541 }, { "epoch": 0.3580688094065372, "grad_norm": 1.7090535163879395, "learning_rate": 7.433275311643776e-06, "loss": 0.5382, "step": 8542 }, { "epoch": 0.35811072802154614, "grad_norm": 1.476723074913025, "learning_rate": 7.432682247781131e-06, "loss": 0.5399, "step": 8543 }, { "epoch": 0.3581526466365551, "grad_norm": 1.4061827659606934, "learning_rate": 7.432089139075351e-06, "loss": 0.5318, "step": 8544 }, { "epoch": 0.35819456525156407, "grad_norm": 3.0279934406280518, "learning_rate": 7.431495985537371e-06, "loss": 0.5117, "step": 8545 }, { "epoch": 0.35823648386657303, "grad_norm": 1.692676305770874, "learning_rate": 7.430902787178124e-06, "loss": 0.5415, "step": 8546 }, { "epoch": 0.358278402481582, "grad_norm": 1.446974754333496, "learning_rate": 7.430309544008543e-06, "loss": 0.5361, "step": 8547 }, { "epoch": 0.35832032109659095, "grad_norm": 1.6212751865386963, "learning_rate": 7.4297162560395665e-06, "loss": 0.5931, "step": 8548 }, { "epoch": 0.3583622397115999, "grad_norm": 1.8823214769363403, "learning_rate": 7.4291229232821306e-06, "loss": 0.5779, "step": 8549 }, { "epoch": 0.3584041583266089, "grad_norm": 1.6313749551773071, "learning_rate": 7.42852954574717e-06, "loss": 0.5692, "step": 8550 }, { "epoch": 0.35844607694161784, "grad_norm": 1.7275927066802979, "learning_rate": 7.427936123445626e-06, "loss": 0.5779, "step": 8551 }, { "epoch": 0.3584879955566268, "grad_norm": 1.66798734664917, "learning_rate": 7.4273426563884355e-06, "loss": 0.5793, "step": 8552 }, { "epoch": 0.35852991417163577, "grad_norm": 1.6222347021102905, "learning_rate": 7.426749144586539e-06, "loss": 0.499, "step": 8553 }, { "epoch": 0.35857183278664473, "grad_norm": 1.4925898313522339, "learning_rate": 7.4261555880508774e-06, "loss": 0.4725, "step": 8554 }, { "epoch": 0.3586137514016537, "grad_norm": 1.6809579133987427, "learning_rate": 7.425561986792391e-06, "loss": 0.5569, "step": 8555 }, { "epoch": 0.35865567001666265, "grad_norm": 1.6332786083221436, "learning_rate": 7.424968340822022e-06, "loss": 0.5103, "step": 8556 }, { "epoch": 0.3586975886316716, "grad_norm": 1.6068012714385986, "learning_rate": 7.424374650150714e-06, "loss": 0.5218, "step": 8557 }, { "epoch": 0.3587395072466806, "grad_norm": 1.520749807357788, "learning_rate": 7.4237809147894135e-06, "loss": 0.4885, "step": 8558 }, { "epoch": 0.35878142586168954, "grad_norm": 1.6471043825149536, "learning_rate": 7.423187134749059e-06, "loss": 0.4796, "step": 8559 }, { "epoch": 0.3588233444766985, "grad_norm": 1.7379053831100464, "learning_rate": 7.422593310040602e-06, "loss": 0.5348, "step": 8560 }, { "epoch": 0.35886526309170746, "grad_norm": 1.5877901315689087, "learning_rate": 7.421999440674986e-06, "loss": 0.436, "step": 8561 }, { "epoch": 0.3589071817067164, "grad_norm": 1.6977722644805908, "learning_rate": 7.421405526663159e-06, "loss": 0.5547, "step": 8562 }, { "epoch": 0.3589491003217254, "grad_norm": 2.0706708431243896, "learning_rate": 7.420811568016067e-06, "loss": 0.5244, "step": 8563 }, { "epoch": 0.35899101893673435, "grad_norm": 1.9921889305114746, "learning_rate": 7.4202175647446625e-06, "loss": 0.5071, "step": 8564 }, { "epoch": 0.3590329375517433, "grad_norm": 1.9270226955413818, "learning_rate": 7.419623516859892e-06, "loss": 0.5463, "step": 8565 }, { "epoch": 0.3590748561667523, "grad_norm": 1.7161314487457275, "learning_rate": 7.419029424372706e-06, "loss": 0.5582, "step": 8566 }, { "epoch": 0.3591167747817612, "grad_norm": 1.530977487564087, "learning_rate": 7.418435287294058e-06, "loss": 0.5982, "step": 8567 }, { "epoch": 0.35915869339677015, "grad_norm": 1.7747516632080078, "learning_rate": 7.4178411056349e-06, "loss": 0.5416, "step": 8568 }, { "epoch": 0.3592006120117791, "grad_norm": 1.7534279823303223, "learning_rate": 7.417246879406181e-06, "loss": 0.5394, "step": 8569 }, { "epoch": 0.35924253062678807, "grad_norm": 1.707728624343872, "learning_rate": 7.41665260861886e-06, "loss": 0.4937, "step": 8570 }, { "epoch": 0.35928444924179703, "grad_norm": 1.3894416093826294, "learning_rate": 7.4160582932838876e-06, "loss": 0.4838, "step": 8571 }, { "epoch": 0.359326367856806, "grad_norm": 1.6502333879470825, "learning_rate": 7.415463933412221e-06, "loss": 0.5259, "step": 8572 }, { "epoch": 0.35936828647181496, "grad_norm": 1.4672380685806274, "learning_rate": 7.414869529014816e-06, "loss": 0.5016, "step": 8573 }, { "epoch": 0.3594102050868239, "grad_norm": 1.4929649829864502, "learning_rate": 7.4142750801026305e-06, "loss": 0.562, "step": 8574 }, { "epoch": 0.3594521237018329, "grad_norm": 1.5633388757705688, "learning_rate": 7.4136805866866205e-06, "loss": 0.482, "step": 8575 }, { "epoch": 0.35949404231684184, "grad_norm": 1.4406312704086304, "learning_rate": 7.413086048777745e-06, "loss": 0.5035, "step": 8576 }, { "epoch": 0.3595359609318508, "grad_norm": 1.6265968084335327, "learning_rate": 7.412491466386968e-06, "loss": 0.502, "step": 8577 }, { "epoch": 0.35957787954685977, "grad_norm": 1.7406898736953735, "learning_rate": 7.411896839525241e-06, "loss": 0.5558, "step": 8578 }, { "epoch": 0.35961979816186873, "grad_norm": 1.7046173810958862, "learning_rate": 7.4113021682035335e-06, "loss": 0.5719, "step": 8579 }, { "epoch": 0.3596617167768777, "grad_norm": 1.9614434242248535, "learning_rate": 7.410707452432803e-06, "loss": 0.5778, "step": 8580 }, { "epoch": 0.35970363539188666, "grad_norm": 1.6658166646957397, "learning_rate": 7.4101126922240126e-06, "loss": 0.5134, "step": 8581 }, { "epoch": 0.3597455540068956, "grad_norm": 1.6510040760040283, "learning_rate": 7.409517887588128e-06, "loss": 0.5638, "step": 8582 }, { "epoch": 0.3597874726219046, "grad_norm": 1.5767171382904053, "learning_rate": 7.408923038536111e-06, "loss": 0.5306, "step": 8583 }, { "epoch": 0.35982939123691354, "grad_norm": 1.5880711078643799, "learning_rate": 7.408328145078928e-06, "loss": 0.5106, "step": 8584 }, { "epoch": 0.3598713098519225, "grad_norm": 1.6552716493606567, "learning_rate": 7.407733207227545e-06, "loss": 0.4921, "step": 8585 }, { "epoch": 0.35991322846693147, "grad_norm": 1.720027208328247, "learning_rate": 7.4071382249929305e-06, "loss": 0.5189, "step": 8586 }, { "epoch": 0.35995514708194043, "grad_norm": 1.5010595321655273, "learning_rate": 7.406543198386049e-06, "loss": 0.4891, "step": 8587 }, { "epoch": 0.3599970656969494, "grad_norm": 1.6088354587554932, "learning_rate": 7.405948127417871e-06, "loss": 0.5021, "step": 8588 }, { "epoch": 0.36003898431195835, "grad_norm": 1.7088544368743896, "learning_rate": 7.405353012099366e-06, "loss": 0.5246, "step": 8589 }, { "epoch": 0.3600809029269673, "grad_norm": 1.622462511062622, "learning_rate": 7.404757852441502e-06, "loss": 0.5402, "step": 8590 }, { "epoch": 0.3601228215419763, "grad_norm": 1.7305415868759155, "learning_rate": 7.404162648455253e-06, "loss": 0.5146, "step": 8591 }, { "epoch": 0.3601647401569852, "grad_norm": 1.6993825435638428, "learning_rate": 7.40356740015159e-06, "loss": 0.5771, "step": 8592 }, { "epoch": 0.36020665877199415, "grad_norm": 1.6597626209259033, "learning_rate": 7.402972107541482e-06, "loss": 0.5896, "step": 8593 }, { "epoch": 0.3602485773870031, "grad_norm": 1.9010869264602661, "learning_rate": 7.402376770635907e-06, "loss": 0.5177, "step": 8594 }, { "epoch": 0.3602904960020121, "grad_norm": 1.598465919494629, "learning_rate": 7.401781389445837e-06, "loss": 0.4853, "step": 8595 }, { "epoch": 0.36033241461702104, "grad_norm": 1.6397496461868286, "learning_rate": 7.401185963982248e-06, "loss": 0.5529, "step": 8596 }, { "epoch": 0.36037433323203, "grad_norm": 1.6162887811660767, "learning_rate": 7.400590494256114e-06, "loss": 0.4537, "step": 8597 }, { "epoch": 0.36041625184703896, "grad_norm": 1.541045904159546, "learning_rate": 7.399994980278416e-06, "loss": 0.5818, "step": 8598 }, { "epoch": 0.3604581704620479, "grad_norm": 1.790267825126648, "learning_rate": 7.3993994220601254e-06, "loss": 0.5282, "step": 8599 }, { "epoch": 0.3605000890770569, "grad_norm": 1.9293540716171265, "learning_rate": 7.398803819612225e-06, "loss": 0.5047, "step": 8600 }, { "epoch": 0.36054200769206585, "grad_norm": 1.5934702157974243, "learning_rate": 7.398208172945691e-06, "loss": 0.516, "step": 8601 }, { "epoch": 0.3605839263070748, "grad_norm": 1.8868821859359741, "learning_rate": 7.397612482071506e-06, "loss": 0.5393, "step": 8602 }, { "epoch": 0.36062584492208377, "grad_norm": 1.7661577463150024, "learning_rate": 7.397016747000648e-06, "loss": 0.5102, "step": 8603 }, { "epoch": 0.36066776353709273, "grad_norm": 1.6681338548660278, "learning_rate": 7.396420967744101e-06, "loss": 0.5221, "step": 8604 }, { "epoch": 0.3607096821521017, "grad_norm": 1.6180894374847412, "learning_rate": 7.395825144312848e-06, "loss": 0.5046, "step": 8605 }, { "epoch": 0.36075160076711066, "grad_norm": 1.7532455921173096, "learning_rate": 7.395229276717868e-06, "loss": 0.4944, "step": 8606 }, { "epoch": 0.3607935193821196, "grad_norm": 1.9016811847686768, "learning_rate": 7.39463336497015e-06, "loss": 0.5284, "step": 8607 }, { "epoch": 0.3608354379971286, "grad_norm": 1.5528627634048462, "learning_rate": 7.394037409080674e-06, "loss": 0.4735, "step": 8608 }, { "epoch": 0.36087735661213755, "grad_norm": 1.929351568222046, "learning_rate": 7.39344140906043e-06, "loss": 0.575, "step": 8609 }, { "epoch": 0.3609192752271465, "grad_norm": 1.5092897415161133, "learning_rate": 7.392845364920402e-06, "loss": 0.4611, "step": 8610 }, { "epoch": 0.36096119384215547, "grad_norm": 2.5492780208587646, "learning_rate": 7.392249276671577e-06, "loss": 0.4817, "step": 8611 }, { "epoch": 0.36100311245716443, "grad_norm": 2.0469157695770264, "learning_rate": 7.391653144324944e-06, "loss": 0.5848, "step": 8612 }, { "epoch": 0.3610450310721734, "grad_norm": 1.4941473007202148, "learning_rate": 7.391056967891491e-06, "loss": 0.4918, "step": 8613 }, { "epoch": 0.36108694968718236, "grad_norm": 1.700827717781067, "learning_rate": 7.390460747382209e-06, "loss": 0.499, "step": 8614 }, { "epoch": 0.3611288683021913, "grad_norm": 1.7121275663375854, "learning_rate": 7.389864482808089e-06, "loss": 0.5051, "step": 8615 }, { "epoch": 0.3611707869172003, "grad_norm": 1.9307125806808472, "learning_rate": 7.3892681741801195e-06, "loss": 0.4857, "step": 8616 }, { "epoch": 0.3612127055322092, "grad_norm": 2.0980677604675293, "learning_rate": 7.388671821509295e-06, "loss": 0.4634, "step": 8617 }, { "epoch": 0.36125462414721815, "grad_norm": 3.5231456756591797, "learning_rate": 7.388075424806607e-06, "loss": 0.5324, "step": 8618 }, { "epoch": 0.3612965427622271, "grad_norm": 1.5157277584075928, "learning_rate": 7.387478984083051e-06, "loss": 0.5081, "step": 8619 }, { "epoch": 0.3613384613772361, "grad_norm": 1.5132976770401, "learning_rate": 7.38688249934962e-06, "loss": 0.5281, "step": 8620 }, { "epoch": 0.36138037999224504, "grad_norm": 1.6595158576965332, "learning_rate": 7.38628597061731e-06, "loss": 0.5424, "step": 8621 }, { "epoch": 0.361422298607254, "grad_norm": 2.4109578132629395, "learning_rate": 7.385689397897117e-06, "loss": 0.4926, "step": 8622 }, { "epoch": 0.36146421722226296, "grad_norm": 1.5076656341552734, "learning_rate": 7.385092781200038e-06, "loss": 0.5046, "step": 8623 }, { "epoch": 0.3615061358372719, "grad_norm": 1.7790614366531372, "learning_rate": 7.38449612053707e-06, "loss": 0.5511, "step": 8624 }, { "epoch": 0.3615480544522809, "grad_norm": 1.9318150281906128, "learning_rate": 7.3838994159192135e-06, "loss": 0.5782, "step": 8625 }, { "epoch": 0.36158997306728985, "grad_norm": 1.6157552003860474, "learning_rate": 7.383302667357468e-06, "loss": 0.5245, "step": 8626 }, { "epoch": 0.3616318916822988, "grad_norm": 1.9790836572647095, "learning_rate": 7.38270587486283e-06, "loss": 0.5031, "step": 8627 }, { "epoch": 0.3616738102973078, "grad_norm": 1.8663018941879272, "learning_rate": 7.382109038446306e-06, "loss": 0.4953, "step": 8628 }, { "epoch": 0.36171572891231674, "grad_norm": 1.5821566581726074, "learning_rate": 7.381512158118893e-06, "loss": 0.4741, "step": 8629 }, { "epoch": 0.3617576475273257, "grad_norm": 1.8969340324401855, "learning_rate": 7.380915233891597e-06, "loss": 0.5214, "step": 8630 }, { "epoch": 0.36179956614233466, "grad_norm": 1.6666059494018555, "learning_rate": 7.380318265775419e-06, "loss": 0.5179, "step": 8631 }, { "epoch": 0.3618414847573436, "grad_norm": 1.797742486000061, "learning_rate": 7.379721253781365e-06, "loss": 0.503, "step": 8632 }, { "epoch": 0.3618834033723526, "grad_norm": 1.7145771980285645, "learning_rate": 7.37912419792044e-06, "loss": 0.5322, "step": 8633 }, { "epoch": 0.36192532198736155, "grad_norm": 1.590113639831543, "learning_rate": 7.378527098203649e-06, "loss": 0.4917, "step": 8634 }, { "epoch": 0.3619672406023705, "grad_norm": 1.5140719413757324, "learning_rate": 7.3779299546419995e-06, "loss": 0.4987, "step": 8635 }, { "epoch": 0.3620091592173795, "grad_norm": 1.6019607782363892, "learning_rate": 7.377332767246498e-06, "loss": 0.5216, "step": 8636 }, { "epoch": 0.36205107783238843, "grad_norm": 1.784481167793274, "learning_rate": 7.376735536028153e-06, "loss": 0.4918, "step": 8637 }, { "epoch": 0.3620929964473974, "grad_norm": 1.6782124042510986, "learning_rate": 7.376138260997976e-06, "loss": 0.5269, "step": 8638 }, { "epoch": 0.36213491506240636, "grad_norm": 1.5507731437683105, "learning_rate": 7.375540942166974e-06, "loss": 0.4927, "step": 8639 }, { "epoch": 0.3621768336774153, "grad_norm": 1.973792552947998, "learning_rate": 7.3749435795461585e-06, "loss": 0.5304, "step": 8640 }, { "epoch": 0.3622187522924243, "grad_norm": 2.17708420753479, "learning_rate": 7.374346173146541e-06, "loss": 0.517, "step": 8641 }, { "epoch": 0.3622606709074332, "grad_norm": 1.6846444606781006, "learning_rate": 7.373748722979136e-06, "loss": 0.5729, "step": 8642 }, { "epoch": 0.36230258952244215, "grad_norm": 1.6062825918197632, "learning_rate": 7.373151229054954e-06, "loss": 0.4945, "step": 8643 }, { "epoch": 0.3623445081374511, "grad_norm": 1.5859160423278809, "learning_rate": 7.37255369138501e-06, "loss": 0.5173, "step": 8644 }, { "epoch": 0.3623864267524601, "grad_norm": 1.6363506317138672, "learning_rate": 7.3719561099803185e-06, "loss": 0.5385, "step": 8645 }, { "epoch": 0.36242834536746904, "grad_norm": 1.654512882232666, "learning_rate": 7.371358484851894e-06, "loss": 0.4974, "step": 8646 }, { "epoch": 0.362470263982478, "grad_norm": 1.723818063735962, "learning_rate": 7.370760816010757e-06, "loss": 0.4633, "step": 8647 }, { "epoch": 0.36251218259748696, "grad_norm": 1.8111963272094727, "learning_rate": 7.37016310346792e-06, "loss": 0.5282, "step": 8648 }, { "epoch": 0.3625541012124959, "grad_norm": 1.7204315662384033, "learning_rate": 7.369565347234402e-06, "loss": 0.5672, "step": 8649 }, { "epoch": 0.3625960198275049, "grad_norm": 1.4864978790283203, "learning_rate": 7.368967547321224e-06, "loss": 0.4786, "step": 8650 }, { "epoch": 0.36263793844251385, "grad_norm": 1.5469551086425781, "learning_rate": 7.368369703739404e-06, "loss": 0.5695, "step": 8651 }, { "epoch": 0.3626798570575228, "grad_norm": 1.7117085456848145, "learning_rate": 7.367771816499962e-06, "loss": 0.4947, "step": 8652 }, { "epoch": 0.3627217756725318, "grad_norm": 2.5499460697174072, "learning_rate": 7.3671738856139206e-06, "loss": 0.5143, "step": 8653 }, { "epoch": 0.36276369428754074, "grad_norm": 1.6145182847976685, "learning_rate": 7.3665759110923015e-06, "loss": 0.4791, "step": 8654 }, { "epoch": 0.3628056129025497, "grad_norm": 1.5891183614730835, "learning_rate": 7.365977892946126e-06, "loss": 0.4652, "step": 8655 }, { "epoch": 0.36284753151755866, "grad_norm": 1.5469717979431152, "learning_rate": 7.365379831186421e-06, "loss": 0.4609, "step": 8656 }, { "epoch": 0.3628894501325676, "grad_norm": 1.8347752094268799, "learning_rate": 7.364781725824207e-06, "loss": 0.5407, "step": 8657 }, { "epoch": 0.3629313687475766, "grad_norm": 2.235858201980591, "learning_rate": 7.3641835768705104e-06, "loss": 0.5566, "step": 8658 }, { "epoch": 0.36297328736258555, "grad_norm": 1.6266106367111206, "learning_rate": 7.363585384336361e-06, "loss": 0.5036, "step": 8659 }, { "epoch": 0.3630152059775945, "grad_norm": 1.8501818180084229, "learning_rate": 7.36298714823278e-06, "loss": 0.5491, "step": 8660 }, { "epoch": 0.3630571245926035, "grad_norm": 1.7430075407028198, "learning_rate": 7.362388868570798e-06, "loss": 0.5427, "step": 8661 }, { "epoch": 0.36309904320761244, "grad_norm": 1.687673807144165, "learning_rate": 7.361790545361444e-06, "loss": 0.5275, "step": 8662 }, { "epoch": 0.3631409618226214, "grad_norm": 1.9663002490997314, "learning_rate": 7.361192178615746e-06, "loss": 0.5605, "step": 8663 }, { "epoch": 0.36318288043763036, "grad_norm": 1.8337594270706177, "learning_rate": 7.360593768344733e-06, "loss": 0.5157, "step": 8664 }, { "epoch": 0.3632247990526393, "grad_norm": 1.5214983224868774, "learning_rate": 7.3599953145594386e-06, "loss": 0.4912, "step": 8665 }, { "epoch": 0.3632667176676483, "grad_norm": 1.6514004468917847, "learning_rate": 7.359396817270894e-06, "loss": 0.5196, "step": 8666 }, { "epoch": 0.3633086362826572, "grad_norm": 2.1664717197418213, "learning_rate": 7.358798276490129e-06, "loss": 0.483, "step": 8667 }, { "epoch": 0.36335055489766616, "grad_norm": 1.9421637058258057, "learning_rate": 7.358199692228181e-06, "loss": 0.5477, "step": 8668 }, { "epoch": 0.3633924735126751, "grad_norm": 1.8940142393112183, "learning_rate": 7.35760106449608e-06, "loss": 0.5976, "step": 8669 }, { "epoch": 0.3634343921276841, "grad_norm": 1.8145668506622314, "learning_rate": 7.357002393304863e-06, "loss": 0.5261, "step": 8670 }, { "epoch": 0.36347631074269304, "grad_norm": 1.7162535190582275, "learning_rate": 7.356403678665566e-06, "loss": 0.5503, "step": 8671 }, { "epoch": 0.363518229357702, "grad_norm": 2.9252684116363525, "learning_rate": 7.355804920589225e-06, "loss": 0.5578, "step": 8672 }, { "epoch": 0.36356014797271097, "grad_norm": 1.6021027565002441, "learning_rate": 7.355206119086875e-06, "loss": 0.5031, "step": 8673 }, { "epoch": 0.36360206658771993, "grad_norm": 1.8065770864486694, "learning_rate": 7.354607274169557e-06, "loss": 0.5298, "step": 8674 }, { "epoch": 0.3636439852027289, "grad_norm": 1.6114534139633179, "learning_rate": 7.354008385848312e-06, "loss": 0.5695, "step": 8675 }, { "epoch": 0.36368590381773785, "grad_norm": 2.397514581680298, "learning_rate": 7.353409454134174e-06, "loss": 0.5116, "step": 8676 }, { "epoch": 0.3637278224327468, "grad_norm": 1.6795967817306519, "learning_rate": 7.352810479038185e-06, "loss": 0.469, "step": 8677 }, { "epoch": 0.3637697410477558, "grad_norm": 1.6642282009124756, "learning_rate": 7.35221146057139e-06, "loss": 0.5248, "step": 8678 }, { "epoch": 0.36381165966276474, "grad_norm": 1.7941654920578003, "learning_rate": 7.351612398744828e-06, "loss": 0.5201, "step": 8679 }, { "epoch": 0.3638535782777737, "grad_norm": 1.6842095851898193, "learning_rate": 7.351013293569541e-06, "loss": 0.5121, "step": 8680 }, { "epoch": 0.36389549689278267, "grad_norm": 1.608583688735962, "learning_rate": 7.350414145056576e-06, "loss": 0.535, "step": 8681 }, { "epoch": 0.36393741550779163, "grad_norm": 2.532803535461426, "learning_rate": 7.349814953216974e-06, "loss": 0.5144, "step": 8682 }, { "epoch": 0.3639793341228006, "grad_norm": 1.7096869945526123, "learning_rate": 7.349215718061782e-06, "loss": 0.553, "step": 8683 }, { "epoch": 0.36402125273780955, "grad_norm": 1.886561632156372, "learning_rate": 7.348616439602047e-06, "loss": 0.533, "step": 8684 }, { "epoch": 0.3640631713528185, "grad_norm": 1.6655781269073486, "learning_rate": 7.348017117848816e-06, "loss": 0.4895, "step": 8685 }, { "epoch": 0.3641050899678275, "grad_norm": 1.5516633987426758, "learning_rate": 7.3474177528131315e-06, "loss": 0.5182, "step": 8686 }, { "epoch": 0.36414700858283644, "grad_norm": 1.9770983457565308, "learning_rate": 7.346818344506049e-06, "loss": 0.5093, "step": 8687 }, { "epoch": 0.3641889271978454, "grad_norm": 1.7244964838027954, "learning_rate": 7.346218892938614e-06, "loss": 0.4743, "step": 8688 }, { "epoch": 0.36423084581285436, "grad_norm": 1.6184872388839722, "learning_rate": 7.345619398121877e-06, "loss": 0.5118, "step": 8689 }, { "epoch": 0.3642727644278633, "grad_norm": 1.6393519639968872, "learning_rate": 7.34501986006689e-06, "loss": 0.5083, "step": 8690 }, { "epoch": 0.3643146830428723, "grad_norm": 2.351255178451538, "learning_rate": 7.344420278784703e-06, "loss": 0.5888, "step": 8691 }, { "epoch": 0.3643566016578812, "grad_norm": 1.8146605491638184, "learning_rate": 7.343820654286369e-06, "loss": 0.5184, "step": 8692 }, { "epoch": 0.36439852027289016, "grad_norm": 1.7616965770721436, "learning_rate": 7.343220986582941e-06, "loss": 0.5439, "step": 8693 }, { "epoch": 0.3644404388878991, "grad_norm": 1.9422167539596558, "learning_rate": 7.342621275685475e-06, "loss": 0.5238, "step": 8694 }, { "epoch": 0.3644823575029081, "grad_norm": 1.6270183324813843, "learning_rate": 7.342021521605023e-06, "loss": 0.4524, "step": 8695 }, { "epoch": 0.36452427611791705, "grad_norm": 1.595621943473816, "learning_rate": 7.341421724352643e-06, "loss": 0.5129, "step": 8696 }, { "epoch": 0.364566194732926, "grad_norm": 2.195758819580078, "learning_rate": 7.340821883939391e-06, "loss": 0.5288, "step": 8697 }, { "epoch": 0.36460811334793497, "grad_norm": 1.7654486894607544, "learning_rate": 7.340222000376322e-06, "loss": 0.4918, "step": 8698 }, { "epoch": 0.36465003196294393, "grad_norm": 5.329325199127197, "learning_rate": 7.339622073674497e-06, "loss": 0.5551, "step": 8699 }, { "epoch": 0.3646919505779529, "grad_norm": 1.7257394790649414, "learning_rate": 7.339022103844973e-06, "loss": 0.5359, "step": 8700 }, { "epoch": 0.36473386919296186, "grad_norm": 2.206766366958618, "learning_rate": 7.338422090898809e-06, "loss": 0.5555, "step": 8701 }, { "epoch": 0.3647757878079708, "grad_norm": 2.006028890609741, "learning_rate": 7.337822034847067e-06, "loss": 0.5008, "step": 8702 }, { "epoch": 0.3648177064229798, "grad_norm": 1.605056881904602, "learning_rate": 7.337221935700809e-06, "loss": 0.5194, "step": 8703 }, { "epoch": 0.36485962503798874, "grad_norm": 1.6996870040893555, "learning_rate": 7.336621793471095e-06, "loss": 0.5349, "step": 8704 }, { "epoch": 0.3649015436529977, "grad_norm": 1.5993379354476929, "learning_rate": 7.336021608168988e-06, "loss": 0.5204, "step": 8705 }, { "epoch": 0.36494346226800667, "grad_norm": 1.8001136779785156, "learning_rate": 7.3354213798055544e-06, "loss": 0.5376, "step": 8706 }, { "epoch": 0.36498538088301563, "grad_norm": 1.6710389852523804, "learning_rate": 7.334821108391854e-06, "loss": 0.4991, "step": 8707 }, { "epoch": 0.3650272994980246, "grad_norm": 1.623436450958252, "learning_rate": 7.334220793938954e-06, "loss": 0.5808, "step": 8708 }, { "epoch": 0.36506921811303356, "grad_norm": 1.668707251548767, "learning_rate": 7.3336204364579225e-06, "loss": 0.5305, "step": 8709 }, { "epoch": 0.3651111367280425, "grad_norm": 1.7741553783416748, "learning_rate": 7.333020035959824e-06, "loss": 0.5956, "step": 8710 }, { "epoch": 0.3651530553430515, "grad_norm": 1.655662178993225, "learning_rate": 7.332419592455724e-06, "loss": 0.5661, "step": 8711 }, { "epoch": 0.36519497395806044, "grad_norm": 1.6838186979293823, "learning_rate": 7.331819105956695e-06, "loss": 0.5661, "step": 8712 }, { "epoch": 0.3652368925730694, "grad_norm": 1.6443350315093994, "learning_rate": 7.3312185764738055e-06, "loss": 0.5152, "step": 8713 }, { "epoch": 0.36527881118807837, "grad_norm": 1.9468331336975098, "learning_rate": 7.330618004018122e-06, "loss": 0.5373, "step": 8714 }, { "epoch": 0.36532072980308733, "grad_norm": 1.9538129568099976, "learning_rate": 7.33001738860072e-06, "loss": 0.5472, "step": 8715 }, { "epoch": 0.3653626484180963, "grad_norm": 1.7031594514846802, "learning_rate": 7.329416730232667e-06, "loss": 0.5638, "step": 8716 }, { "epoch": 0.3654045670331052, "grad_norm": 1.412116289138794, "learning_rate": 7.328816028925037e-06, "loss": 0.5277, "step": 8717 }, { "epoch": 0.36544648564811416, "grad_norm": 1.4744325876235962, "learning_rate": 7.328215284688904e-06, "loss": 0.4866, "step": 8718 }, { "epoch": 0.3654884042631231, "grad_norm": 1.8348442316055298, "learning_rate": 7.32761449753534e-06, "loss": 0.5794, "step": 8719 }, { "epoch": 0.3655303228781321, "grad_norm": 1.7836430072784424, "learning_rate": 7.32701366747542e-06, "loss": 0.5144, "step": 8720 }, { "epoch": 0.36557224149314105, "grad_norm": 1.9144049882888794, "learning_rate": 7.326412794520221e-06, "loss": 0.5243, "step": 8721 }, { "epoch": 0.36561416010815, "grad_norm": 1.9127389192581177, "learning_rate": 7.32581187868082e-06, "loss": 0.5772, "step": 8722 }, { "epoch": 0.365656078723159, "grad_norm": 1.9380356073379517, "learning_rate": 7.325210919968288e-06, "loss": 0.5318, "step": 8723 }, { "epoch": 0.36569799733816793, "grad_norm": 1.5565335750579834, "learning_rate": 7.324609918393711e-06, "loss": 0.4859, "step": 8724 }, { "epoch": 0.3657399159531769, "grad_norm": 1.621380090713501, "learning_rate": 7.324008873968163e-06, "loss": 0.5273, "step": 8725 }, { "epoch": 0.36578183456818586, "grad_norm": 1.8463398218154907, "learning_rate": 7.323407786702723e-06, "loss": 0.5614, "step": 8726 }, { "epoch": 0.3658237531831948, "grad_norm": 2.1038193702697754, "learning_rate": 7.322806656608475e-06, "loss": 0.5172, "step": 8727 }, { "epoch": 0.3658656717982038, "grad_norm": 1.731002926826477, "learning_rate": 7.322205483696495e-06, "loss": 0.5393, "step": 8728 }, { "epoch": 0.36590759041321275, "grad_norm": 2.127587080001831, "learning_rate": 7.321604267977866e-06, "loss": 0.557, "step": 8729 }, { "epoch": 0.3659495090282217, "grad_norm": 1.943121314048767, "learning_rate": 7.321003009463675e-06, "loss": 0.559, "step": 8730 }, { "epoch": 0.36599142764323067, "grad_norm": 1.7420940399169922, "learning_rate": 7.320401708165001e-06, "loss": 0.5666, "step": 8731 }, { "epoch": 0.36603334625823963, "grad_norm": 2.2781875133514404, "learning_rate": 7.31980036409293e-06, "loss": 0.5025, "step": 8732 }, { "epoch": 0.3660752648732486, "grad_norm": 1.9253038167953491, "learning_rate": 7.319198977258546e-06, "loss": 0.4416, "step": 8733 }, { "epoch": 0.36611718348825756, "grad_norm": 1.6543004512786865, "learning_rate": 7.318597547672934e-06, "loss": 0.5237, "step": 8734 }, { "epoch": 0.3661591021032665, "grad_norm": 1.7344850301742554, "learning_rate": 7.317996075347182e-06, "loss": 0.5291, "step": 8735 }, { "epoch": 0.3662010207182755, "grad_norm": 1.7973427772521973, "learning_rate": 7.317394560292379e-06, "loss": 0.5357, "step": 8736 }, { "epoch": 0.36624293933328445, "grad_norm": 1.5017491579055786, "learning_rate": 7.316793002519609e-06, "loss": 0.5008, "step": 8737 }, { "epoch": 0.3662848579482934, "grad_norm": 1.7688065767288208, "learning_rate": 7.3161914020399625e-06, "loss": 0.5471, "step": 8738 }, { "epoch": 0.36632677656330237, "grad_norm": 3.903179168701172, "learning_rate": 7.3155897588645295e-06, "loss": 0.5155, "step": 8739 }, { "epoch": 0.36636869517831133, "grad_norm": 1.7634193897247314, "learning_rate": 7.314988073004401e-06, "loss": 0.5447, "step": 8740 }, { "epoch": 0.3664106137933203, "grad_norm": 1.9490221738815308, "learning_rate": 7.3143863444706685e-06, "loss": 0.5339, "step": 8741 }, { "epoch": 0.3664525324083292, "grad_norm": 1.5448261499404907, "learning_rate": 7.313784573274423e-06, "loss": 0.5262, "step": 8742 }, { "epoch": 0.36649445102333816, "grad_norm": 1.5234569311141968, "learning_rate": 7.313182759426758e-06, "loss": 0.496, "step": 8743 }, { "epoch": 0.3665363696383471, "grad_norm": 1.5980933904647827, "learning_rate": 7.312580902938764e-06, "loss": 0.4811, "step": 8744 }, { "epoch": 0.3665782882533561, "grad_norm": 1.8181349039077759, "learning_rate": 7.311979003821541e-06, "loss": 0.4997, "step": 8745 }, { "epoch": 0.36662020686836505, "grad_norm": 1.6983813047409058, "learning_rate": 7.311377062086181e-06, "loss": 0.4888, "step": 8746 }, { "epoch": 0.366662125483374, "grad_norm": 1.755556583404541, "learning_rate": 7.310775077743779e-06, "loss": 0.5479, "step": 8747 }, { "epoch": 0.366704044098383, "grad_norm": 1.423375129699707, "learning_rate": 7.310173050805434e-06, "loss": 0.5186, "step": 8748 }, { "epoch": 0.36674596271339194, "grad_norm": 1.7309271097183228, "learning_rate": 7.3095709812822415e-06, "loss": 0.5463, "step": 8749 }, { "epoch": 0.3667878813284009, "grad_norm": 1.8641462326049805, "learning_rate": 7.308968869185302e-06, "loss": 0.525, "step": 8750 }, { "epoch": 0.36682979994340986, "grad_norm": 1.5442578792572021, "learning_rate": 7.308366714525713e-06, "loss": 0.5027, "step": 8751 }, { "epoch": 0.3668717185584188, "grad_norm": 1.7894541025161743, "learning_rate": 7.307764517314575e-06, "loss": 0.5057, "step": 8752 }, { "epoch": 0.3669136371734278, "grad_norm": 1.5125702619552612, "learning_rate": 7.307162277562988e-06, "loss": 0.4997, "step": 8753 }, { "epoch": 0.36695555578843675, "grad_norm": 1.7641246318817139, "learning_rate": 7.306559995282055e-06, "loss": 0.5499, "step": 8754 }, { "epoch": 0.3669974744034457, "grad_norm": 2.3960607051849365, "learning_rate": 7.305957670482876e-06, "loss": 0.5099, "step": 8755 }, { "epoch": 0.3670393930184547, "grad_norm": 1.7415052652359009, "learning_rate": 7.305355303176555e-06, "loss": 0.4839, "step": 8756 }, { "epoch": 0.36708131163346364, "grad_norm": 1.583151936531067, "learning_rate": 7.304752893374197e-06, "loss": 0.5066, "step": 8757 }, { "epoch": 0.3671232302484726, "grad_norm": 1.8347704410552979, "learning_rate": 7.304150441086906e-06, "loss": 0.5063, "step": 8758 }, { "epoch": 0.36716514886348156, "grad_norm": 1.7746158838272095, "learning_rate": 7.303547946325786e-06, "loss": 0.511, "step": 8759 }, { "epoch": 0.3672070674784905, "grad_norm": 1.6443183422088623, "learning_rate": 7.3029454091019445e-06, "loss": 0.5615, "step": 8760 }, { "epoch": 0.3672489860934995, "grad_norm": 1.8734937906265259, "learning_rate": 7.3023428294264885e-06, "loss": 0.5126, "step": 8761 }, { "epoch": 0.36729090470850845, "grad_norm": 1.6492446660995483, "learning_rate": 7.3017402073105246e-06, "loss": 0.5001, "step": 8762 }, { "epoch": 0.3673328233235174, "grad_norm": 1.8238439559936523, "learning_rate": 7.3011375427651624e-06, "loss": 0.4833, "step": 8763 }, { "epoch": 0.3673747419385264, "grad_norm": 1.4309265613555908, "learning_rate": 7.300534835801512e-06, "loss": 0.5031, "step": 8764 }, { "epoch": 0.36741666055353533, "grad_norm": 1.7694849967956543, "learning_rate": 7.2999320864306814e-06, "loss": 0.5781, "step": 8765 }, { "epoch": 0.3674585791685443, "grad_norm": 1.7094272375106812, "learning_rate": 7.29932929466378e-06, "loss": 0.5095, "step": 8766 }, { "epoch": 0.3675004977835532, "grad_norm": 1.6893783807754517, "learning_rate": 7.298726460511925e-06, "loss": 0.5378, "step": 8767 }, { "epoch": 0.36754241639856217, "grad_norm": 1.649317741394043, "learning_rate": 7.298123583986224e-06, "loss": 0.5528, "step": 8768 }, { "epoch": 0.36758433501357113, "grad_norm": 1.8595126867294312, "learning_rate": 7.297520665097792e-06, "loss": 0.519, "step": 8769 }, { "epoch": 0.3676262536285801, "grad_norm": 1.7091037034988403, "learning_rate": 7.2969177038577444e-06, "loss": 0.5214, "step": 8770 }, { "epoch": 0.36766817224358905, "grad_norm": 1.619302749633789, "learning_rate": 7.296314700277193e-06, "loss": 0.4815, "step": 8771 }, { "epoch": 0.367710090858598, "grad_norm": 1.6398961544036865, "learning_rate": 7.295711654367254e-06, "loss": 0.5389, "step": 8772 }, { "epoch": 0.367752009473607, "grad_norm": 1.5709710121154785, "learning_rate": 7.295108566139045e-06, "loss": 0.4894, "step": 8773 }, { "epoch": 0.36779392808861594, "grad_norm": 1.5594286918640137, "learning_rate": 7.294505435603683e-06, "loss": 0.5181, "step": 8774 }, { "epoch": 0.3678358467036249, "grad_norm": 1.7308025360107422, "learning_rate": 7.293902262772284e-06, "loss": 0.5182, "step": 8775 }, { "epoch": 0.36787776531863386, "grad_norm": 1.5445626974105835, "learning_rate": 7.293299047655968e-06, "loss": 0.4786, "step": 8776 }, { "epoch": 0.3679196839336428, "grad_norm": 1.8616067171096802, "learning_rate": 7.2926957902658555e-06, "loss": 0.5936, "step": 8777 }, { "epoch": 0.3679616025486518, "grad_norm": 1.620985746383667, "learning_rate": 7.292092490613066e-06, "loss": 0.5333, "step": 8778 }, { "epoch": 0.36800352116366075, "grad_norm": 1.7451820373535156, "learning_rate": 7.2914891487087205e-06, "loss": 0.5661, "step": 8779 }, { "epoch": 0.3680454397786697, "grad_norm": 2.2257039546966553, "learning_rate": 7.29088576456394e-06, "loss": 0.5487, "step": 8780 }, { "epoch": 0.3680873583936787, "grad_norm": 1.7126190662384033, "learning_rate": 7.290282338189846e-06, "loss": 0.5772, "step": 8781 }, { "epoch": 0.36812927700868764, "grad_norm": 1.85610830783844, "learning_rate": 7.289678869597564e-06, "loss": 0.5316, "step": 8782 }, { "epoch": 0.3681711956236966, "grad_norm": 1.5840984582901, "learning_rate": 7.2890753587982184e-06, "loss": 0.5506, "step": 8783 }, { "epoch": 0.36821311423870556, "grad_norm": 2.8095908164978027, "learning_rate": 7.288471805802932e-06, "loss": 0.5319, "step": 8784 }, { "epoch": 0.3682550328537145, "grad_norm": 1.8029011487960815, "learning_rate": 7.2878682106228324e-06, "loss": 0.5182, "step": 8785 }, { "epoch": 0.3682969514687235, "grad_norm": 1.8649486303329468, "learning_rate": 7.287264573269045e-06, "loss": 0.4847, "step": 8786 }, { "epoch": 0.36833887008373245, "grad_norm": 1.6639349460601807, "learning_rate": 7.2866608937526976e-06, "loss": 0.4966, "step": 8787 }, { "epoch": 0.3683807886987414, "grad_norm": 1.5431389808654785, "learning_rate": 7.286057172084917e-06, "loss": 0.5583, "step": 8788 }, { "epoch": 0.3684227073137504, "grad_norm": 1.578178882598877, "learning_rate": 7.285453408276834e-06, "loss": 0.4939, "step": 8789 }, { "epoch": 0.36846462592875934, "grad_norm": 1.782157301902771, "learning_rate": 7.284849602339576e-06, "loss": 0.5536, "step": 8790 }, { "epoch": 0.3685065445437683, "grad_norm": 1.5199519395828247, "learning_rate": 7.284245754284273e-06, "loss": 0.4738, "step": 8791 }, { "epoch": 0.3685484631587772, "grad_norm": 1.644939661026001, "learning_rate": 7.283641864122059e-06, "loss": 0.4953, "step": 8792 }, { "epoch": 0.36859038177378617, "grad_norm": 1.545009970664978, "learning_rate": 7.283037931864065e-06, "loss": 0.5098, "step": 8793 }, { "epoch": 0.36863230038879513, "grad_norm": 1.7831451892852783, "learning_rate": 7.282433957521421e-06, "loss": 0.5733, "step": 8794 }, { "epoch": 0.3686742190038041, "grad_norm": 2.3686447143554688, "learning_rate": 7.281829941105264e-06, "loss": 0.554, "step": 8795 }, { "epoch": 0.36871613761881306, "grad_norm": 1.7166798114776611, "learning_rate": 7.281225882626726e-06, "loss": 0.4854, "step": 8796 }, { "epoch": 0.368758056233822, "grad_norm": 1.8500514030456543, "learning_rate": 7.280621782096943e-06, "loss": 0.5471, "step": 8797 }, { "epoch": 0.368799974848831, "grad_norm": 2.3156039714813232, "learning_rate": 7.280017639527049e-06, "loss": 0.5327, "step": 8798 }, { "epoch": 0.36884189346383994, "grad_norm": 1.9477838277816772, "learning_rate": 7.279413454928184e-06, "loss": 0.4726, "step": 8799 }, { "epoch": 0.3688838120788489, "grad_norm": 1.4934624433517456, "learning_rate": 7.278809228311481e-06, "loss": 0.4996, "step": 8800 }, { "epoch": 0.36892573069385787, "grad_norm": 1.7523796558380127, "learning_rate": 7.2782049596880795e-06, "loss": 0.5267, "step": 8801 }, { "epoch": 0.36896764930886683, "grad_norm": 1.8372629880905151, "learning_rate": 7.2776006490691235e-06, "loss": 0.5264, "step": 8802 }, { "epoch": 0.3690095679238758, "grad_norm": 1.6131097078323364, "learning_rate": 7.276996296465744e-06, "loss": 0.4978, "step": 8803 }, { "epoch": 0.36905148653888475, "grad_norm": 1.8710922002792358, "learning_rate": 7.276391901889086e-06, "loss": 0.5546, "step": 8804 }, { "epoch": 0.3690934051538937, "grad_norm": 1.6254223585128784, "learning_rate": 7.275787465350293e-06, "loss": 0.516, "step": 8805 }, { "epoch": 0.3691353237689027, "grad_norm": 1.5980859994888306, "learning_rate": 7.275182986860502e-06, "loss": 0.4968, "step": 8806 }, { "epoch": 0.36917724238391164, "grad_norm": 1.8341788053512573, "learning_rate": 7.274578466430858e-06, "loss": 0.5287, "step": 8807 }, { "epoch": 0.3692191609989206, "grad_norm": 1.9657986164093018, "learning_rate": 7.2739739040725055e-06, "loss": 0.5126, "step": 8808 }, { "epoch": 0.36926107961392957, "grad_norm": 1.6558061838150024, "learning_rate": 7.2733692997965864e-06, "loss": 0.5044, "step": 8809 }, { "epoch": 0.36930299822893853, "grad_norm": 1.563971757888794, "learning_rate": 7.2727646536142484e-06, "loss": 0.4742, "step": 8810 }, { "epoch": 0.3693449168439475, "grad_norm": 1.871967077255249, "learning_rate": 7.272159965536636e-06, "loss": 0.5033, "step": 8811 }, { "epoch": 0.36938683545895645, "grad_norm": 2.7656283378601074, "learning_rate": 7.271555235574895e-06, "loss": 0.5247, "step": 8812 }, { "epoch": 0.3694287540739654, "grad_norm": 2.286233901977539, "learning_rate": 7.270950463740174e-06, "loss": 0.5578, "step": 8813 }, { "epoch": 0.3694706726889744, "grad_norm": 2.209913730621338, "learning_rate": 7.270345650043621e-06, "loss": 0.5142, "step": 8814 }, { "epoch": 0.36951259130398334, "grad_norm": 1.6538052558898926, "learning_rate": 7.269740794496385e-06, "loss": 0.5059, "step": 8815 }, { "epoch": 0.3695545099189923, "grad_norm": 1.8023638725280762, "learning_rate": 7.269135897109613e-06, "loss": 0.5044, "step": 8816 }, { "epoch": 0.3695964285340012, "grad_norm": 1.7015576362609863, "learning_rate": 7.2685309578944595e-06, "loss": 0.4484, "step": 8817 }, { "epoch": 0.36963834714901017, "grad_norm": 1.7200759649276733, "learning_rate": 7.267925976862073e-06, "loss": 0.5278, "step": 8818 }, { "epoch": 0.36968026576401913, "grad_norm": 2.467992067337036, "learning_rate": 7.267320954023608e-06, "loss": 0.4909, "step": 8819 }, { "epoch": 0.3697221843790281, "grad_norm": 1.5984563827514648, "learning_rate": 7.266715889390216e-06, "loss": 0.52, "step": 8820 }, { "epoch": 0.36976410299403706, "grad_norm": 1.6372222900390625, "learning_rate": 7.266110782973048e-06, "loss": 0.5438, "step": 8821 }, { "epoch": 0.369806021609046, "grad_norm": 1.7456746101379395, "learning_rate": 7.2655056347832616e-06, "loss": 0.5216, "step": 8822 }, { "epoch": 0.369847940224055, "grad_norm": 1.5035251379013062, "learning_rate": 7.26490044483201e-06, "loss": 0.4976, "step": 8823 }, { "epoch": 0.36988985883906395, "grad_norm": 1.777757167816162, "learning_rate": 7.264295213130451e-06, "loss": 0.5509, "step": 8824 }, { "epoch": 0.3699317774540729, "grad_norm": 2.1899397373199463, "learning_rate": 7.263689939689739e-06, "loss": 0.5631, "step": 8825 }, { "epoch": 0.36997369606908187, "grad_norm": 1.8684940338134766, "learning_rate": 7.2630846245210355e-06, "loss": 0.5496, "step": 8826 }, { "epoch": 0.37001561468409083, "grad_norm": 1.7853509187698364, "learning_rate": 7.262479267635491e-06, "loss": 0.4573, "step": 8827 }, { "epoch": 0.3700575332990998, "grad_norm": 1.670940637588501, "learning_rate": 7.261873869044272e-06, "loss": 0.581, "step": 8828 }, { "epoch": 0.37009945191410876, "grad_norm": 1.8502131700515747, "learning_rate": 7.261268428758535e-06, "loss": 0.5645, "step": 8829 }, { "epoch": 0.3701413705291177, "grad_norm": 1.9131317138671875, "learning_rate": 7.26066294678944e-06, "loss": 0.5623, "step": 8830 }, { "epoch": 0.3701832891441267, "grad_norm": 1.6148909330368042, "learning_rate": 7.260057423148148e-06, "loss": 0.4946, "step": 8831 }, { "epoch": 0.37022520775913564, "grad_norm": 1.8342866897583008, "learning_rate": 7.259451857845823e-06, "loss": 0.5408, "step": 8832 }, { "epoch": 0.3702671263741446, "grad_norm": 1.6123121976852417, "learning_rate": 7.258846250893625e-06, "loss": 0.4676, "step": 8833 }, { "epoch": 0.37030904498915357, "grad_norm": 1.7781264781951904, "learning_rate": 7.258240602302721e-06, "loss": 0.5658, "step": 8834 }, { "epoch": 0.37035096360416253, "grad_norm": 2.224760055541992, "learning_rate": 7.257634912084272e-06, "loss": 0.479, "step": 8835 }, { "epoch": 0.3703928822191715, "grad_norm": 2.031317949295044, "learning_rate": 7.257029180249447e-06, "loss": 0.5709, "step": 8836 }, { "epoch": 0.37043480083418046, "grad_norm": 1.4976228475570679, "learning_rate": 7.256423406809405e-06, "loss": 0.5353, "step": 8837 }, { "epoch": 0.3704767194491894, "grad_norm": 1.8984365463256836, "learning_rate": 7.255817591775321e-06, "loss": 0.5601, "step": 8838 }, { "epoch": 0.3705186380641984, "grad_norm": 2.0034470558166504, "learning_rate": 7.255211735158357e-06, "loss": 0.541, "step": 8839 }, { "epoch": 0.37056055667920734, "grad_norm": 1.7007535696029663, "learning_rate": 7.254605836969681e-06, "loss": 0.5454, "step": 8840 }, { "epoch": 0.3706024752942163, "grad_norm": 1.5675694942474365, "learning_rate": 7.253999897220465e-06, "loss": 0.5257, "step": 8841 }, { "epoch": 0.3706443939092252, "grad_norm": 1.6729551553726196, "learning_rate": 7.253393915921876e-06, "loss": 0.5398, "step": 8842 }, { "epoch": 0.3706863125242342, "grad_norm": 1.8063056468963623, "learning_rate": 7.252787893085086e-06, "loss": 0.5543, "step": 8843 }, { "epoch": 0.37072823113924314, "grad_norm": 1.9874821901321411, "learning_rate": 7.252181828721265e-06, "loss": 0.5576, "step": 8844 }, { "epoch": 0.3707701497542521, "grad_norm": 1.7547523975372314, "learning_rate": 7.2515757228415865e-06, "loss": 0.5169, "step": 8845 }, { "epoch": 0.37081206836926106, "grad_norm": 1.994198203086853, "learning_rate": 7.250969575457221e-06, "loss": 0.5438, "step": 8846 }, { "epoch": 0.37085398698427, "grad_norm": 1.9643023014068604, "learning_rate": 7.2503633865793435e-06, "loss": 0.5306, "step": 8847 }, { "epoch": 0.370895905599279, "grad_norm": 1.700486183166504, "learning_rate": 7.249757156219128e-06, "loss": 0.532, "step": 8848 }, { "epoch": 0.37093782421428795, "grad_norm": 1.7201299667358398, "learning_rate": 7.24915088438775e-06, "loss": 0.5225, "step": 8849 }, { "epoch": 0.3709797428292969, "grad_norm": 2.036151885986328, "learning_rate": 7.248544571096386e-06, "loss": 0.5443, "step": 8850 }, { "epoch": 0.3710216614443059, "grad_norm": 1.6936135292053223, "learning_rate": 7.24793821635621e-06, "loss": 0.5324, "step": 8851 }, { "epoch": 0.37106358005931483, "grad_norm": 2.151493549346924, "learning_rate": 7.247331820178399e-06, "loss": 0.5132, "step": 8852 }, { "epoch": 0.3711054986743238, "grad_norm": 1.9142537117004395, "learning_rate": 7.246725382574135e-06, "loss": 0.4925, "step": 8853 }, { "epoch": 0.37114741728933276, "grad_norm": 1.637315034866333, "learning_rate": 7.246118903554595e-06, "loss": 0.5226, "step": 8854 }, { "epoch": 0.3711893359043417, "grad_norm": 1.8479875326156616, "learning_rate": 7.245512383130956e-06, "loss": 0.562, "step": 8855 }, { "epoch": 0.3712312545193507, "grad_norm": 1.6923794746398926, "learning_rate": 7.244905821314402e-06, "loss": 0.5052, "step": 8856 }, { "epoch": 0.37127317313435965, "grad_norm": 1.8713661432266235, "learning_rate": 7.2442992181161136e-06, "loss": 0.5791, "step": 8857 }, { "epoch": 0.3713150917493686, "grad_norm": 1.6683663129806519, "learning_rate": 7.2436925735472695e-06, "loss": 0.5113, "step": 8858 }, { "epoch": 0.37135701036437757, "grad_norm": 1.798778772354126, "learning_rate": 7.243085887619056e-06, "loss": 0.5233, "step": 8859 }, { "epoch": 0.37139892897938653, "grad_norm": 1.9379732608795166, "learning_rate": 7.242479160342655e-06, "loss": 0.4897, "step": 8860 }, { "epoch": 0.3714408475943955, "grad_norm": 1.431630253791809, "learning_rate": 7.241872391729251e-06, "loss": 0.5296, "step": 8861 }, { "epoch": 0.37148276620940446, "grad_norm": 2.810035228729248, "learning_rate": 7.241265581790028e-06, "loss": 0.5415, "step": 8862 }, { "epoch": 0.3715246848244134, "grad_norm": 1.909294843673706, "learning_rate": 7.240658730536173e-06, "loss": 0.5035, "step": 8863 }, { "epoch": 0.3715666034394224, "grad_norm": 2.289646863937378, "learning_rate": 7.240051837978872e-06, "loss": 0.5025, "step": 8864 }, { "epoch": 0.37160852205443134, "grad_norm": 2.126948833465576, "learning_rate": 7.239444904129311e-06, "loss": 0.5114, "step": 8865 }, { "epoch": 0.3716504406694403, "grad_norm": 1.8026175498962402, "learning_rate": 7.23883792899868e-06, "loss": 0.4993, "step": 8866 }, { "epoch": 0.3716923592844492, "grad_norm": 1.6841119527816772, "learning_rate": 7.238230912598168e-06, "loss": 0.5265, "step": 8867 }, { "epoch": 0.3717342778994582, "grad_norm": 1.6787912845611572, "learning_rate": 7.237623854938962e-06, "loss": 0.5112, "step": 8868 }, { "epoch": 0.37177619651446714, "grad_norm": 1.7931079864501953, "learning_rate": 7.237016756032254e-06, "loss": 0.5125, "step": 8869 }, { "epoch": 0.3718181151294761, "grad_norm": 1.854407548904419, "learning_rate": 7.236409615889235e-06, "loss": 0.5476, "step": 8870 }, { "epoch": 0.37186003374448506, "grad_norm": 1.8235224485397339, "learning_rate": 7.235802434521095e-06, "loss": 0.5347, "step": 8871 }, { "epoch": 0.371901952359494, "grad_norm": 2.0416927337646484, "learning_rate": 7.23519521193903e-06, "loss": 0.5158, "step": 8872 }, { "epoch": 0.371943870974503, "grad_norm": 1.766646385192871, "learning_rate": 7.23458794815423e-06, "loss": 0.5461, "step": 8873 }, { "epoch": 0.37198578958951195, "grad_norm": 1.5170223712921143, "learning_rate": 7.23398064317789e-06, "loss": 0.5024, "step": 8874 }, { "epoch": 0.3720277082045209, "grad_norm": 2.0759899616241455, "learning_rate": 7.233373297021206e-06, "loss": 0.6157, "step": 8875 }, { "epoch": 0.3720696268195299, "grad_norm": 1.8860749006271362, "learning_rate": 7.232765909695374e-06, "loss": 0.4605, "step": 8876 }, { "epoch": 0.37211154543453884, "grad_norm": 1.8694334030151367, "learning_rate": 7.2321584812115884e-06, "loss": 0.4905, "step": 8877 }, { "epoch": 0.3721534640495478, "grad_norm": 2.0160982608795166, "learning_rate": 7.231551011581046e-06, "loss": 0.4989, "step": 8878 }, { "epoch": 0.37219538266455676, "grad_norm": 2.6414883136749268, "learning_rate": 7.230943500814947e-06, "loss": 0.5251, "step": 8879 }, { "epoch": 0.3722373012795657, "grad_norm": 1.6775752305984497, "learning_rate": 7.230335948924487e-06, "loss": 0.5629, "step": 8880 }, { "epoch": 0.3722792198945747, "grad_norm": 1.7610599994659424, "learning_rate": 7.22972835592087e-06, "loss": 0.538, "step": 8881 }, { "epoch": 0.37232113850958365, "grad_norm": 1.6688907146453857, "learning_rate": 7.229120721815291e-06, "loss": 0.5278, "step": 8882 }, { "epoch": 0.3723630571245926, "grad_norm": 1.7678974866867065, "learning_rate": 7.228513046618954e-06, "loss": 0.516, "step": 8883 }, { "epoch": 0.3724049757396016, "grad_norm": 1.888656735420227, "learning_rate": 7.227905330343059e-06, "loss": 0.5194, "step": 8884 }, { "epoch": 0.37244689435461054, "grad_norm": 1.7014083862304688, "learning_rate": 7.2272975729988106e-06, "loss": 0.5058, "step": 8885 }, { "epoch": 0.3724888129696195, "grad_norm": 1.614975929260254, "learning_rate": 7.22668977459741e-06, "loss": 0.5328, "step": 8886 }, { "epoch": 0.37253073158462846, "grad_norm": 1.77655827999115, "learning_rate": 7.226081935150061e-06, "loss": 0.5433, "step": 8887 }, { "epoch": 0.3725726501996374, "grad_norm": 1.8602834939956665, "learning_rate": 7.225474054667969e-06, "loss": 0.4602, "step": 8888 }, { "epoch": 0.3726145688146464, "grad_norm": 1.6585770845413208, "learning_rate": 7.2248661331623406e-06, "loss": 0.4962, "step": 8889 }, { "epoch": 0.37265648742965535, "grad_norm": 1.627005934715271, "learning_rate": 7.2242581706443804e-06, "loss": 0.4877, "step": 8890 }, { "epoch": 0.3726984060446643, "grad_norm": 1.8555052280426025, "learning_rate": 7.223650167125296e-06, "loss": 0.5462, "step": 8891 }, { "epoch": 0.3727403246596732, "grad_norm": 1.6817373037338257, "learning_rate": 7.223042122616293e-06, "loss": 0.4976, "step": 8892 }, { "epoch": 0.3727822432746822, "grad_norm": 1.5228451490402222, "learning_rate": 7.222434037128583e-06, "loss": 0.4431, "step": 8893 }, { "epoch": 0.37282416188969114, "grad_norm": 1.7073211669921875, "learning_rate": 7.2218259106733745e-06, "loss": 0.4717, "step": 8894 }, { "epoch": 0.3728660805047001, "grad_norm": 1.9785552024841309, "learning_rate": 7.221217743261877e-06, "loss": 0.4451, "step": 8895 }, { "epoch": 0.37290799911970907, "grad_norm": 1.7572225332260132, "learning_rate": 7.2206095349053e-06, "loss": 0.5013, "step": 8896 }, { "epoch": 0.37294991773471803, "grad_norm": 1.7480825185775757, "learning_rate": 7.220001285614857e-06, "loss": 0.5168, "step": 8897 }, { "epoch": 0.372991836349727, "grad_norm": 1.8834153413772583, "learning_rate": 7.2193929954017596e-06, "loss": 0.4683, "step": 8898 }, { "epoch": 0.37303375496473595, "grad_norm": 1.5639773607254028, "learning_rate": 7.218784664277218e-06, "loss": 0.5412, "step": 8899 }, { "epoch": 0.3730756735797449, "grad_norm": 2.059326171875, "learning_rate": 7.218176292252452e-06, "loss": 0.4838, "step": 8900 }, { "epoch": 0.3731175921947539, "grad_norm": 4.546842575073242, "learning_rate": 7.217567879338672e-06, "loss": 0.5336, "step": 8901 }, { "epoch": 0.37315951080976284, "grad_norm": 1.8478842973709106, "learning_rate": 7.216959425547092e-06, "loss": 0.5506, "step": 8902 }, { "epoch": 0.3732014294247718, "grad_norm": 1.7400575876235962, "learning_rate": 7.216350930888931e-06, "loss": 0.4765, "step": 8903 }, { "epoch": 0.37324334803978076, "grad_norm": 1.9395264387130737, "learning_rate": 7.215742395375405e-06, "loss": 0.5084, "step": 8904 }, { "epoch": 0.3732852666547897, "grad_norm": 2.181791067123413, "learning_rate": 7.21513381901773e-06, "loss": 0.5446, "step": 8905 }, { "epoch": 0.3733271852697987, "grad_norm": 1.9331743717193604, "learning_rate": 7.214525201827126e-06, "loss": 0.5058, "step": 8906 }, { "epoch": 0.37336910388480765, "grad_norm": 1.7614613771438599, "learning_rate": 7.213916543814812e-06, "loss": 0.5107, "step": 8907 }, { "epoch": 0.3734110224998166, "grad_norm": 2.017434597015381, "learning_rate": 7.213307844992006e-06, "loss": 0.5032, "step": 8908 }, { "epoch": 0.3734529411148256, "grad_norm": 1.6964110136032104, "learning_rate": 7.2126991053699315e-06, "loss": 0.4961, "step": 8909 }, { "epoch": 0.37349485972983454, "grad_norm": 2.001147985458374, "learning_rate": 7.212090324959806e-06, "loss": 0.536, "step": 8910 }, { "epoch": 0.3735367783448435, "grad_norm": 1.7724623680114746, "learning_rate": 7.211481503772854e-06, "loss": 0.5365, "step": 8911 }, { "epoch": 0.37357869695985246, "grad_norm": 1.635496973991394, "learning_rate": 7.210872641820296e-06, "loss": 0.4982, "step": 8912 }, { "epoch": 0.3736206155748614, "grad_norm": 1.7137880325317383, "learning_rate": 7.210263739113361e-06, "loss": 0.5338, "step": 8913 }, { "epoch": 0.3736625341898704, "grad_norm": 1.746761679649353, "learning_rate": 7.209654795663268e-06, "loss": 0.5726, "step": 8914 }, { "epoch": 0.37370445280487935, "grad_norm": 2.0373668670654297, "learning_rate": 7.209045811481244e-06, "loss": 0.5393, "step": 8915 }, { "epoch": 0.3737463714198883, "grad_norm": 1.7191581726074219, "learning_rate": 7.208436786578512e-06, "loss": 0.5156, "step": 8916 }, { "epoch": 0.3737882900348972, "grad_norm": 1.939695119857788, "learning_rate": 7.207827720966302e-06, "loss": 0.5338, "step": 8917 }, { "epoch": 0.3738302086499062, "grad_norm": 1.8205846548080444, "learning_rate": 7.207218614655841e-06, "loss": 0.502, "step": 8918 }, { "epoch": 0.37387212726491514, "grad_norm": 1.7486841678619385, "learning_rate": 7.2066094676583575e-06, "loss": 0.5407, "step": 8919 }, { "epoch": 0.3739140458799241, "grad_norm": 1.645650863647461, "learning_rate": 7.206000279985077e-06, "loss": 0.4935, "step": 8920 }, { "epoch": 0.37395596449493307, "grad_norm": 1.8072645664215088, "learning_rate": 7.205391051647232e-06, "loss": 0.5091, "step": 8921 }, { "epoch": 0.37399788310994203, "grad_norm": 1.8867677450180054, "learning_rate": 7.204781782656051e-06, "loss": 0.6046, "step": 8922 }, { "epoch": 0.374039801724951, "grad_norm": 1.9170567989349365, "learning_rate": 7.204172473022766e-06, "loss": 0.5593, "step": 8923 }, { "epoch": 0.37408172033995996, "grad_norm": 1.7417699098587036, "learning_rate": 7.203563122758609e-06, "loss": 0.5533, "step": 8924 }, { "epoch": 0.3741236389549689, "grad_norm": 1.9776723384857178, "learning_rate": 7.2029537318748125e-06, "loss": 0.4976, "step": 8925 }, { "epoch": 0.3741655575699779, "grad_norm": 1.7649115324020386, "learning_rate": 7.202344300382609e-06, "loss": 0.4863, "step": 8926 }, { "epoch": 0.37420747618498684, "grad_norm": 2.54068660736084, "learning_rate": 7.201734828293232e-06, "loss": 0.5076, "step": 8927 }, { "epoch": 0.3742493947999958, "grad_norm": 4.162315368652344, "learning_rate": 7.20112531561792e-06, "loss": 0.5274, "step": 8928 }, { "epoch": 0.37429131341500477, "grad_norm": 1.692679524421692, "learning_rate": 7.200515762367902e-06, "loss": 0.5321, "step": 8929 }, { "epoch": 0.37433323203001373, "grad_norm": 2.3418564796447754, "learning_rate": 7.19990616855442e-06, "loss": 0.5742, "step": 8930 }, { "epoch": 0.3743751506450227, "grad_norm": 1.7090353965759277, "learning_rate": 7.19929653418871e-06, "loss": 0.5109, "step": 8931 }, { "epoch": 0.37441706926003165, "grad_norm": 2.0301244258880615, "learning_rate": 7.198686859282008e-06, "loss": 0.5337, "step": 8932 }, { "epoch": 0.3744589878750406, "grad_norm": 1.5098694562911987, "learning_rate": 7.1980771438455535e-06, "loss": 0.5191, "step": 8933 }, { "epoch": 0.3745009064900496, "grad_norm": 4.131714344024658, "learning_rate": 7.197467387890585e-06, "loss": 0.5634, "step": 8934 }, { "epoch": 0.37454282510505854, "grad_norm": 1.942305326461792, "learning_rate": 7.196857591428344e-06, "loss": 0.5427, "step": 8935 }, { "epoch": 0.3745847437200675, "grad_norm": 1.6205312013626099, "learning_rate": 7.1962477544700694e-06, "loss": 0.478, "step": 8936 }, { "epoch": 0.37462666233507647, "grad_norm": 1.7250471115112305, "learning_rate": 7.195637877027005e-06, "loss": 0.5685, "step": 8937 }, { "epoch": 0.37466858095008543, "grad_norm": 2.48717999458313, "learning_rate": 7.195027959110391e-06, "loss": 0.5026, "step": 8938 }, { "epoch": 0.3747104995650944, "grad_norm": 1.5372047424316406, "learning_rate": 7.19441800073147e-06, "loss": 0.5248, "step": 8939 }, { "epoch": 0.37475241818010335, "grad_norm": 1.543784499168396, "learning_rate": 7.193808001901488e-06, "loss": 0.5249, "step": 8940 }, { "epoch": 0.3747943367951123, "grad_norm": 1.6969355344772339, "learning_rate": 7.193197962631688e-06, "loss": 0.5484, "step": 8941 }, { "epoch": 0.3748362554101212, "grad_norm": 2.074091911315918, "learning_rate": 7.192587882933316e-06, "loss": 0.5111, "step": 8942 }, { "epoch": 0.3748781740251302, "grad_norm": 1.916921854019165, "learning_rate": 7.191977762817617e-06, "loss": 0.5073, "step": 8943 }, { "epoch": 0.37492009264013915, "grad_norm": 1.5264781713485718, "learning_rate": 7.191367602295839e-06, "loss": 0.5057, "step": 8944 }, { "epoch": 0.3749620112551481, "grad_norm": 1.7061580419540405, "learning_rate": 7.190757401379227e-06, "loss": 0.5332, "step": 8945 }, { "epoch": 0.37500392987015707, "grad_norm": 2.0327017307281494, "learning_rate": 7.190147160079033e-06, "loss": 0.5551, "step": 8946 }, { "epoch": 0.37504584848516603, "grad_norm": 2.019096851348877, "learning_rate": 7.1895368784065035e-06, "loss": 0.5649, "step": 8947 }, { "epoch": 0.375087767100175, "grad_norm": 2.1966309547424316, "learning_rate": 7.188926556372887e-06, "loss": 0.4836, "step": 8948 }, { "epoch": 0.37512968571518396, "grad_norm": 1.7379577159881592, "learning_rate": 7.188316193989438e-06, "loss": 0.5012, "step": 8949 }, { "epoch": 0.3751716043301929, "grad_norm": 1.7867538928985596, "learning_rate": 7.187705791267403e-06, "loss": 0.5636, "step": 8950 }, { "epoch": 0.3752135229452019, "grad_norm": 2.581295967102051, "learning_rate": 7.187095348218037e-06, "loss": 0.5221, "step": 8951 }, { "epoch": 0.37525544156021085, "grad_norm": 1.6227302551269531, "learning_rate": 7.186484864852592e-06, "loss": 0.4969, "step": 8952 }, { "epoch": 0.3752973601752198, "grad_norm": 1.7693220376968384, "learning_rate": 7.185874341182321e-06, "loss": 0.4935, "step": 8953 }, { "epoch": 0.37533927879022877, "grad_norm": 1.694291353225708, "learning_rate": 7.1852637772184784e-06, "loss": 0.5355, "step": 8954 }, { "epoch": 0.37538119740523773, "grad_norm": 1.9139387607574463, "learning_rate": 7.184653172972319e-06, "loss": 0.4784, "step": 8955 }, { "epoch": 0.3754231160202467, "grad_norm": 2.1532046794891357, "learning_rate": 7.184042528455099e-06, "loss": 0.5216, "step": 8956 }, { "epoch": 0.37546503463525566, "grad_norm": 1.6845301389694214, "learning_rate": 7.1834318436780746e-06, "loss": 0.5564, "step": 8957 }, { "epoch": 0.3755069532502646, "grad_norm": 1.8649064302444458, "learning_rate": 7.182821118652502e-06, "loss": 0.5334, "step": 8958 }, { "epoch": 0.3755488718652736, "grad_norm": 1.899809718132019, "learning_rate": 7.1822103533896404e-06, "loss": 0.5196, "step": 8959 }, { "epoch": 0.37559079048028254, "grad_norm": 1.6619781255722046, "learning_rate": 7.181599547900747e-06, "loss": 0.498, "step": 8960 }, { "epoch": 0.3756327090952915, "grad_norm": 1.7953441143035889, "learning_rate": 7.180988702197083e-06, "loss": 0.5348, "step": 8961 }, { "epoch": 0.37567462771030047, "grad_norm": 1.4699294567108154, "learning_rate": 7.180377816289906e-06, "loss": 0.4688, "step": 8962 }, { "epoch": 0.37571654632530943, "grad_norm": 2.0265612602233887, "learning_rate": 7.1797668901904785e-06, "loss": 0.5342, "step": 8963 }, { "epoch": 0.3757584649403184, "grad_norm": 1.992316484451294, "learning_rate": 7.179155923910062e-06, "loss": 0.4991, "step": 8964 }, { "epoch": 0.37580038355532736, "grad_norm": 1.607159972190857, "learning_rate": 7.178544917459919e-06, "loss": 0.5273, "step": 8965 }, { "epoch": 0.3758423021703363, "grad_norm": 1.909672737121582, "learning_rate": 7.177933870851311e-06, "loss": 0.448, "step": 8966 }, { "epoch": 0.3758842207853453, "grad_norm": 1.807788372039795, "learning_rate": 7.177322784095505e-06, "loss": 0.4806, "step": 8967 }, { "epoch": 0.3759261394003542, "grad_norm": 1.6805914640426636, "learning_rate": 7.176711657203762e-06, "loss": 0.5, "step": 8968 }, { "epoch": 0.37596805801536315, "grad_norm": 1.5970412492752075, "learning_rate": 7.176100490187349e-06, "loss": 0.4739, "step": 8969 }, { "epoch": 0.3760099766303721, "grad_norm": 1.5971497297286987, "learning_rate": 7.1754892830575325e-06, "loss": 0.4981, "step": 8970 }, { "epoch": 0.3760518952453811, "grad_norm": 1.587585687637329, "learning_rate": 7.174878035825579e-06, "loss": 0.4984, "step": 8971 }, { "epoch": 0.37609381386039004, "grad_norm": 1.7030367851257324, "learning_rate": 7.174266748502755e-06, "loss": 0.4966, "step": 8972 }, { "epoch": 0.376135732475399, "grad_norm": 2.3459537029266357, "learning_rate": 7.173655421100328e-06, "loss": 0.5197, "step": 8973 }, { "epoch": 0.37617765109040796, "grad_norm": 1.8178590536117554, "learning_rate": 7.17304405362957e-06, "loss": 0.4993, "step": 8974 }, { "epoch": 0.3762195697054169, "grad_norm": 1.8355079889297485, "learning_rate": 7.172432646101749e-06, "loss": 0.558, "step": 8975 }, { "epoch": 0.3762614883204259, "grad_norm": 1.898769736289978, "learning_rate": 7.171821198528134e-06, "loss": 0.4476, "step": 8976 }, { "epoch": 0.37630340693543485, "grad_norm": 2.481288194656372, "learning_rate": 7.171209710919999e-06, "loss": 0.5572, "step": 8977 }, { "epoch": 0.3763453255504438, "grad_norm": 1.8605167865753174, "learning_rate": 7.170598183288614e-06, "loss": 0.4645, "step": 8978 }, { "epoch": 0.3763872441654528, "grad_norm": 1.9821105003356934, "learning_rate": 7.169986615645252e-06, "loss": 0.5521, "step": 8979 }, { "epoch": 0.37642916278046173, "grad_norm": 1.6574279069900513, "learning_rate": 7.169375008001188e-06, "loss": 0.5209, "step": 8980 }, { "epoch": 0.3764710813954707, "grad_norm": 1.7853769063949585, "learning_rate": 7.168763360367694e-06, "loss": 0.5372, "step": 8981 }, { "epoch": 0.37651300001047966, "grad_norm": 1.9946630001068115, "learning_rate": 7.168151672756044e-06, "loss": 0.4997, "step": 8982 }, { "epoch": 0.3765549186254886, "grad_norm": 1.6772271394729614, "learning_rate": 7.167539945177518e-06, "loss": 0.5075, "step": 8983 }, { "epoch": 0.3765968372404976, "grad_norm": 1.8874282836914062, "learning_rate": 7.166928177643388e-06, "loss": 0.4727, "step": 8984 }, { "epoch": 0.37663875585550655, "grad_norm": 1.647890567779541, "learning_rate": 7.166316370164933e-06, "loss": 0.54, "step": 8985 }, { "epoch": 0.3766806744705155, "grad_norm": 1.66134512424469, "learning_rate": 7.1657045227534315e-06, "loss": 0.4798, "step": 8986 }, { "epoch": 0.37672259308552447, "grad_norm": 1.8002121448516846, "learning_rate": 7.1650926354201596e-06, "loss": 0.5473, "step": 8987 }, { "epoch": 0.37676451170053343, "grad_norm": 1.870100975036621, "learning_rate": 7.1644807081764e-06, "loss": 0.576, "step": 8988 }, { "epoch": 0.3768064303155424, "grad_norm": 1.7035853862762451, "learning_rate": 7.163868741033429e-06, "loss": 0.4807, "step": 8989 }, { "epoch": 0.37684834893055136, "grad_norm": 1.8554643392562866, "learning_rate": 7.1632567340025306e-06, "loss": 0.5421, "step": 8990 }, { "epoch": 0.3768902675455603, "grad_norm": 1.8207639455795288, "learning_rate": 7.1626446870949836e-06, "loss": 0.4866, "step": 8991 }, { "epoch": 0.3769321861605693, "grad_norm": 1.8197945356369019, "learning_rate": 7.162032600322072e-06, "loss": 0.5163, "step": 8992 }, { "epoch": 0.3769741047755782, "grad_norm": 1.7200582027435303, "learning_rate": 7.161420473695078e-06, "loss": 0.5038, "step": 8993 }, { "epoch": 0.37701602339058715, "grad_norm": 1.6995295286178589, "learning_rate": 7.160808307225288e-06, "loss": 0.5351, "step": 8994 }, { "epoch": 0.3770579420055961, "grad_norm": 1.9312167167663574, "learning_rate": 7.160196100923982e-06, "loss": 0.4962, "step": 8995 }, { "epoch": 0.3770998606206051, "grad_norm": 1.744854211807251, "learning_rate": 7.159583854802448e-06, "loss": 0.5389, "step": 8996 }, { "epoch": 0.37714177923561404, "grad_norm": 1.7206172943115234, "learning_rate": 7.1589715688719685e-06, "loss": 0.4596, "step": 8997 }, { "epoch": 0.377183697850623, "grad_norm": 1.7057121992111206, "learning_rate": 7.158359243143838e-06, "loss": 0.5189, "step": 8998 }, { "epoch": 0.37722561646563196, "grad_norm": 1.761602520942688, "learning_rate": 7.157746877629334e-06, "loss": 0.5448, "step": 8999 }, { "epoch": 0.3772675350806409, "grad_norm": 1.8811168670654297, "learning_rate": 7.157134472339751e-06, "loss": 0.5459, "step": 9000 }, { "epoch": 0.3773094536956499, "grad_norm": 1.7461265325546265, "learning_rate": 7.156522027286376e-06, "loss": 0.5203, "step": 9001 }, { "epoch": 0.37735137231065885, "grad_norm": 2.0872960090637207, "learning_rate": 7.155909542480499e-06, "loss": 0.5054, "step": 9002 }, { "epoch": 0.3773932909256678, "grad_norm": 1.9978911876678467, "learning_rate": 7.15529701793341e-06, "loss": 0.5574, "step": 9003 }, { "epoch": 0.3774352095406768, "grad_norm": 1.6540212631225586, "learning_rate": 7.154684453656401e-06, "loss": 0.5116, "step": 9004 }, { "epoch": 0.37747712815568574, "grad_norm": 1.674333095550537, "learning_rate": 7.154071849660762e-06, "loss": 0.5327, "step": 9005 }, { "epoch": 0.3775190467706947, "grad_norm": 2.2087161540985107, "learning_rate": 7.1534592059577845e-06, "loss": 0.5387, "step": 9006 }, { "epoch": 0.37756096538570366, "grad_norm": 1.863542914390564, "learning_rate": 7.152846522558768e-06, "loss": 0.5091, "step": 9007 }, { "epoch": 0.3776028840007126, "grad_norm": 1.9548064470291138, "learning_rate": 7.152233799474999e-06, "loss": 0.4945, "step": 9008 }, { "epoch": 0.3776448026157216, "grad_norm": 1.7924319505691528, "learning_rate": 7.151621036717775e-06, "loss": 0.5029, "step": 9009 }, { "epoch": 0.37768672123073055, "grad_norm": 1.8988429307937622, "learning_rate": 7.151008234298393e-06, "loss": 0.5744, "step": 9010 }, { "epoch": 0.3777286398457395, "grad_norm": 1.7033499479293823, "learning_rate": 7.150395392228146e-06, "loss": 0.5355, "step": 9011 }, { "epoch": 0.3777705584607485, "grad_norm": 1.747502326965332, "learning_rate": 7.149782510518335e-06, "loss": 0.5278, "step": 9012 }, { "epoch": 0.37781247707575744, "grad_norm": 4.224319934844971, "learning_rate": 7.149169589180255e-06, "loss": 0.531, "step": 9013 }, { "epoch": 0.3778543956907664, "grad_norm": 1.7021052837371826, "learning_rate": 7.148556628225204e-06, "loss": 0.5525, "step": 9014 }, { "epoch": 0.37789631430577536, "grad_norm": 1.692541480064392, "learning_rate": 7.147943627664482e-06, "loss": 0.5198, "step": 9015 }, { "epoch": 0.3779382329207843, "grad_norm": 1.9372309446334839, "learning_rate": 7.147330587509388e-06, "loss": 0.5564, "step": 9016 }, { "epoch": 0.3779801515357933, "grad_norm": 2.2412593364715576, "learning_rate": 7.146717507771226e-06, "loss": 0.5129, "step": 9017 }, { "epoch": 0.3780220701508022, "grad_norm": 1.8035753965377808, "learning_rate": 7.146104388461293e-06, "loss": 0.4978, "step": 9018 }, { "epoch": 0.37806398876581115, "grad_norm": 1.6493172645568848, "learning_rate": 7.145491229590891e-06, "loss": 0.546, "step": 9019 }, { "epoch": 0.3781059073808201, "grad_norm": 1.8077409267425537, "learning_rate": 7.144878031171325e-06, "loss": 0.5288, "step": 9020 }, { "epoch": 0.3781478259958291, "grad_norm": 2.016103744506836, "learning_rate": 7.144264793213898e-06, "loss": 0.581, "step": 9021 }, { "epoch": 0.37818974461083804, "grad_norm": 1.7486048936843872, "learning_rate": 7.143651515729913e-06, "loss": 0.4819, "step": 9022 }, { "epoch": 0.378231663225847, "grad_norm": 1.726616382598877, "learning_rate": 7.143038198730678e-06, "loss": 0.5206, "step": 9023 }, { "epoch": 0.37827358184085597, "grad_norm": 1.9415479898452759, "learning_rate": 7.142424842227494e-06, "loss": 0.5281, "step": 9024 }, { "epoch": 0.37831550045586493, "grad_norm": 2.1216630935668945, "learning_rate": 7.1418114462316705e-06, "loss": 0.5405, "step": 9025 }, { "epoch": 0.3783574190708739, "grad_norm": 1.9936307668685913, "learning_rate": 7.1411980107545155e-06, "loss": 0.5822, "step": 9026 }, { "epoch": 0.37839933768588285, "grad_norm": 2.0479254722595215, "learning_rate": 7.140584535807334e-06, "loss": 0.5195, "step": 9027 }, { "epoch": 0.3784412563008918, "grad_norm": 1.7276710271835327, "learning_rate": 7.139971021401435e-06, "loss": 0.494, "step": 9028 }, { "epoch": 0.3784831749159008, "grad_norm": 2.8955607414245605, "learning_rate": 7.13935746754813e-06, "loss": 0.5598, "step": 9029 }, { "epoch": 0.37852509353090974, "grad_norm": 1.9263064861297607, "learning_rate": 7.1387438742587275e-06, "loss": 0.5392, "step": 9030 }, { "epoch": 0.3785670121459187, "grad_norm": 2.299792528152466, "learning_rate": 7.138130241544538e-06, "loss": 0.5639, "step": 9031 }, { "epoch": 0.37860893076092766, "grad_norm": 2.1454758644104004, "learning_rate": 7.137516569416875e-06, "loss": 0.5379, "step": 9032 }, { "epoch": 0.3786508493759366, "grad_norm": 1.7534005641937256, "learning_rate": 7.136902857887047e-06, "loss": 0.5258, "step": 9033 }, { "epoch": 0.3786927679909456, "grad_norm": 1.741087794303894, "learning_rate": 7.1362891069663696e-06, "loss": 0.5061, "step": 9034 }, { "epoch": 0.37873468660595455, "grad_norm": 2.228571891784668, "learning_rate": 7.135675316666155e-06, "loss": 0.4893, "step": 9035 }, { "epoch": 0.3787766052209635, "grad_norm": 1.916062355041504, "learning_rate": 7.135061486997721e-06, "loss": 0.5605, "step": 9036 }, { "epoch": 0.3788185238359725, "grad_norm": 1.9991813898086548, "learning_rate": 7.134447617972378e-06, "loss": 0.5611, "step": 9037 }, { "epoch": 0.37886044245098144, "grad_norm": 1.8783795833587646, "learning_rate": 7.133833709601445e-06, "loss": 0.4563, "step": 9038 }, { "epoch": 0.3789023610659904, "grad_norm": 1.7809001207351685, "learning_rate": 7.133219761896236e-06, "loss": 0.5207, "step": 9039 }, { "epoch": 0.37894427968099936, "grad_norm": 2.6192030906677246, "learning_rate": 7.132605774868071e-06, "loss": 0.5306, "step": 9040 }, { "epoch": 0.3789861982960083, "grad_norm": 2.0246877670288086, "learning_rate": 7.131991748528266e-06, "loss": 0.5204, "step": 9041 }, { "epoch": 0.3790281169110173, "grad_norm": 2.143969774246216, "learning_rate": 7.131377682888142e-06, "loss": 0.4782, "step": 9042 }, { "epoch": 0.3790700355260262, "grad_norm": 1.5927091836929321, "learning_rate": 7.130763577959015e-06, "loss": 0.5116, "step": 9043 }, { "epoch": 0.37911195414103516, "grad_norm": 1.711780309677124, "learning_rate": 7.130149433752208e-06, "loss": 0.5055, "step": 9044 }, { "epoch": 0.3791538727560441, "grad_norm": 1.9016001224517822, "learning_rate": 7.129535250279043e-06, "loss": 0.4847, "step": 9045 }, { "epoch": 0.3791957913710531, "grad_norm": 1.8191198110580444, "learning_rate": 7.128921027550837e-06, "loss": 0.5288, "step": 9046 }, { "epoch": 0.37923770998606204, "grad_norm": 1.8118318319320679, "learning_rate": 7.128306765578915e-06, "loss": 0.5135, "step": 9047 }, { "epoch": 0.379279628601071, "grad_norm": 1.8102445602416992, "learning_rate": 7.127692464374601e-06, "loss": 0.5443, "step": 9048 }, { "epoch": 0.37932154721607997, "grad_norm": 2.0975356101989746, "learning_rate": 7.1270781239492184e-06, "loss": 0.5061, "step": 9049 }, { "epoch": 0.37936346583108893, "grad_norm": 2.1328225135803223, "learning_rate": 7.126463744314091e-06, "loss": 0.5597, "step": 9050 }, { "epoch": 0.3794053844460979, "grad_norm": 1.8607066869735718, "learning_rate": 7.125849325480543e-06, "loss": 0.4777, "step": 9051 }, { "epoch": 0.37944730306110686, "grad_norm": 2.481815814971924, "learning_rate": 7.125234867459903e-06, "loss": 0.5065, "step": 9052 }, { "epoch": 0.3794892216761158, "grad_norm": 1.8653007745742798, "learning_rate": 7.1246203702634945e-06, "loss": 0.5488, "step": 9053 }, { "epoch": 0.3795311402911248, "grad_norm": 1.727284550666809, "learning_rate": 7.1240058339026485e-06, "loss": 0.5597, "step": 9054 }, { "epoch": 0.37957305890613374, "grad_norm": 1.9442273378372192, "learning_rate": 7.123391258388691e-06, "loss": 0.5232, "step": 9055 }, { "epoch": 0.3796149775211427, "grad_norm": 1.7907519340515137, "learning_rate": 7.122776643732949e-06, "loss": 0.535, "step": 9056 }, { "epoch": 0.37965689613615167, "grad_norm": 1.7753325700759888, "learning_rate": 7.122161989946756e-06, "loss": 0.5014, "step": 9057 }, { "epoch": 0.37969881475116063, "grad_norm": 2.006179094314575, "learning_rate": 7.121547297041441e-06, "loss": 0.472, "step": 9058 }, { "epoch": 0.3797407333661696, "grad_norm": 2.21803879737854, "learning_rate": 7.120932565028334e-06, "loss": 0.4948, "step": 9059 }, { "epoch": 0.37978265198117855, "grad_norm": 1.7174477577209473, "learning_rate": 7.120317793918767e-06, "loss": 0.5097, "step": 9060 }, { "epoch": 0.3798245705961875, "grad_norm": 2.098480463027954, "learning_rate": 7.1197029837240735e-06, "loss": 0.5217, "step": 9061 }, { "epoch": 0.3798664892111965, "grad_norm": 1.8619661331176758, "learning_rate": 7.119088134455585e-06, "loss": 0.5161, "step": 9062 }, { "epoch": 0.37990840782620544, "grad_norm": 1.6800932884216309, "learning_rate": 7.118473246124636e-06, "loss": 0.5323, "step": 9063 }, { "epoch": 0.3799503264412144, "grad_norm": 1.671628475189209, "learning_rate": 7.117858318742564e-06, "loss": 0.4836, "step": 9064 }, { "epoch": 0.37999224505622337, "grad_norm": 2.041959762573242, "learning_rate": 7.1172433523206985e-06, "loss": 0.5627, "step": 9065 }, { "epoch": 0.38003416367123233, "grad_norm": 1.6852328777313232, "learning_rate": 7.1166283468703804e-06, "loss": 0.4986, "step": 9066 }, { "epoch": 0.3800760822862413, "grad_norm": 2.1567800045013428, "learning_rate": 7.116013302402944e-06, "loss": 0.5697, "step": 9067 }, { "epoch": 0.3801180009012502, "grad_norm": 1.8408461809158325, "learning_rate": 7.11539821892973e-06, "loss": 0.532, "step": 9068 }, { "epoch": 0.38015991951625916, "grad_norm": 1.6806613206863403, "learning_rate": 7.114783096462072e-06, "loss": 0.5545, "step": 9069 }, { "epoch": 0.3802018381312681, "grad_norm": 1.740361213684082, "learning_rate": 7.114167935011311e-06, "loss": 0.5179, "step": 9070 }, { "epoch": 0.3802437567462771, "grad_norm": 2.1130568981170654, "learning_rate": 7.113552734588788e-06, "loss": 0.5691, "step": 9071 }, { "epoch": 0.38028567536128605, "grad_norm": 1.861262559890747, "learning_rate": 7.112937495205842e-06, "loss": 0.5002, "step": 9072 }, { "epoch": 0.380327593976295, "grad_norm": 1.8587573766708374, "learning_rate": 7.112322216873815e-06, "loss": 0.4899, "step": 9073 }, { "epoch": 0.38036951259130397, "grad_norm": 1.936861276626587, "learning_rate": 7.111706899604046e-06, "loss": 0.5832, "step": 9074 }, { "epoch": 0.38041143120631293, "grad_norm": 1.7832318544387817, "learning_rate": 7.111091543407882e-06, "loss": 0.4923, "step": 9075 }, { "epoch": 0.3804533498213219, "grad_norm": 1.827468752861023, "learning_rate": 7.1104761482966625e-06, "loss": 0.5836, "step": 9076 }, { "epoch": 0.38049526843633086, "grad_norm": 1.7890757322311401, "learning_rate": 7.109860714281733e-06, "loss": 0.5078, "step": 9077 }, { "epoch": 0.3805371870513398, "grad_norm": 2.573279619216919, "learning_rate": 7.109245241374438e-06, "loss": 0.5266, "step": 9078 }, { "epoch": 0.3805791056663488, "grad_norm": 2.117345094680786, "learning_rate": 7.108629729586124e-06, "loss": 0.5282, "step": 9079 }, { "epoch": 0.38062102428135774, "grad_norm": 1.736481785774231, "learning_rate": 7.108014178928135e-06, "loss": 0.56, "step": 9080 }, { "epoch": 0.3806629428963667, "grad_norm": 1.8056342601776123, "learning_rate": 7.107398589411818e-06, "loss": 0.548, "step": 9081 }, { "epoch": 0.38070486151137567, "grad_norm": 1.822862148284912, "learning_rate": 7.106782961048523e-06, "loss": 0.5285, "step": 9082 }, { "epoch": 0.38074678012638463, "grad_norm": 1.6550606489181519, "learning_rate": 7.106167293849597e-06, "loss": 0.5209, "step": 9083 }, { "epoch": 0.3807886987413936, "grad_norm": 1.8285738229751587, "learning_rate": 7.105551587826387e-06, "loss": 0.4992, "step": 9084 }, { "epoch": 0.38083061735640256, "grad_norm": 1.7720638513565063, "learning_rate": 7.104935842990245e-06, "loss": 0.4921, "step": 9085 }, { "epoch": 0.3808725359714115, "grad_norm": 1.8708972930908203, "learning_rate": 7.104320059352521e-06, "loss": 0.5011, "step": 9086 }, { "epoch": 0.3809144545864205, "grad_norm": 1.8348497152328491, "learning_rate": 7.103704236924566e-06, "loss": 0.4998, "step": 9087 }, { "epoch": 0.38095637320142944, "grad_norm": 1.8319097757339478, "learning_rate": 7.1030883757177305e-06, "loss": 0.5342, "step": 9088 }, { "epoch": 0.3809982918164384, "grad_norm": 3.4484269618988037, "learning_rate": 7.102472475743369e-06, "loss": 0.5344, "step": 9089 }, { "epoch": 0.38104021043144737, "grad_norm": 1.7073239088058472, "learning_rate": 7.101856537012833e-06, "loss": 0.5735, "step": 9090 }, { "epoch": 0.38108212904645633, "grad_norm": 1.7120691537857056, "learning_rate": 7.101240559537479e-06, "loss": 0.5045, "step": 9091 }, { "epoch": 0.3811240476614653, "grad_norm": 1.630292296409607, "learning_rate": 7.1006245433286615e-06, "loss": 0.4872, "step": 9092 }, { "epoch": 0.3811659662764742, "grad_norm": 3.8073084354400635, "learning_rate": 7.100008488397732e-06, "loss": 0.526, "step": 9093 }, { "epoch": 0.38120788489148316, "grad_norm": 1.6657510995864868, "learning_rate": 7.099392394756049e-06, "loss": 0.538, "step": 9094 }, { "epoch": 0.3812498035064921, "grad_norm": 1.8841851949691772, "learning_rate": 7.09877626241497e-06, "loss": 0.4692, "step": 9095 }, { "epoch": 0.3812917221215011, "grad_norm": 1.5922385454177856, "learning_rate": 7.0981600913858526e-06, "loss": 0.5063, "step": 9096 }, { "epoch": 0.38133364073651005, "grad_norm": 1.757918357849121, "learning_rate": 7.097543881680055e-06, "loss": 0.5036, "step": 9097 }, { "epoch": 0.381375559351519, "grad_norm": 2.2449536323547363, "learning_rate": 7.096927633308935e-06, "loss": 0.4746, "step": 9098 }, { "epoch": 0.381417477966528, "grad_norm": 1.783363699913025, "learning_rate": 7.096311346283853e-06, "loss": 0.5341, "step": 9099 }, { "epoch": 0.38145939658153694, "grad_norm": 1.8908441066741943, "learning_rate": 7.095695020616169e-06, "loss": 0.5503, "step": 9100 }, { "epoch": 0.3815013151965459, "grad_norm": 1.931659460067749, "learning_rate": 7.095078656317247e-06, "loss": 0.53, "step": 9101 }, { "epoch": 0.38154323381155486, "grad_norm": 1.9193761348724365, "learning_rate": 7.094462253398443e-06, "loss": 0.5385, "step": 9102 }, { "epoch": 0.3815851524265638, "grad_norm": 1.596958041191101, "learning_rate": 7.093845811871125e-06, "loss": 0.5158, "step": 9103 }, { "epoch": 0.3816270710415728, "grad_norm": 1.7464104890823364, "learning_rate": 7.093229331746653e-06, "loss": 0.4901, "step": 9104 }, { "epoch": 0.38166898965658175, "grad_norm": 1.5199048519134521, "learning_rate": 7.092612813036392e-06, "loss": 0.5055, "step": 9105 }, { "epoch": 0.3817109082715907, "grad_norm": 1.7788753509521484, "learning_rate": 7.0919962557517075e-06, "loss": 0.4671, "step": 9106 }, { "epoch": 0.38175282688659967, "grad_norm": 2.2855031490325928, "learning_rate": 7.091379659903964e-06, "loss": 0.5738, "step": 9107 }, { "epoch": 0.38179474550160863, "grad_norm": 1.5307362079620361, "learning_rate": 7.090763025504527e-06, "loss": 0.5521, "step": 9108 }, { "epoch": 0.3818366641166176, "grad_norm": 1.8113062381744385, "learning_rate": 7.090146352564764e-06, "loss": 0.6092, "step": 9109 }, { "epoch": 0.38187858273162656, "grad_norm": 1.9165668487548828, "learning_rate": 7.089529641096043e-06, "loss": 0.5248, "step": 9110 }, { "epoch": 0.3819205013466355, "grad_norm": 1.799187421798706, "learning_rate": 7.0889128911097326e-06, "loss": 0.5385, "step": 9111 }, { "epoch": 0.3819624199616445, "grad_norm": 4.873837471008301, "learning_rate": 7.088296102617201e-06, "loss": 0.5376, "step": 9112 }, { "epoch": 0.38200433857665345, "grad_norm": 2.0129830837249756, "learning_rate": 7.087679275629817e-06, "loss": 0.4696, "step": 9113 }, { "epoch": 0.3820462571916624, "grad_norm": 1.8936690092086792, "learning_rate": 7.087062410158949e-06, "loss": 0.5903, "step": 9114 }, { "epoch": 0.38208817580667137, "grad_norm": 1.7505830526351929, "learning_rate": 7.086445506215976e-06, "loss": 0.513, "step": 9115 }, { "epoch": 0.38213009442168033, "grad_norm": 1.8361375331878662, "learning_rate": 7.08582856381226e-06, "loss": 0.4837, "step": 9116 }, { "epoch": 0.3821720130366893, "grad_norm": 1.7041633129119873, "learning_rate": 7.0852115829591795e-06, "loss": 0.5247, "step": 9117 }, { "epoch": 0.3822139316516982, "grad_norm": 1.6557186841964722, "learning_rate": 7.084594563668107e-06, "loss": 0.5315, "step": 9118 }, { "epoch": 0.38225585026670716, "grad_norm": 1.5943245887756348, "learning_rate": 7.0839775059504145e-06, "loss": 0.4905, "step": 9119 }, { "epoch": 0.3822977688817161, "grad_norm": 1.6201353073120117, "learning_rate": 7.083360409817479e-06, "loss": 0.4913, "step": 9120 }, { "epoch": 0.3823396874967251, "grad_norm": 1.6351386308670044, "learning_rate": 7.082743275280673e-06, "loss": 0.4946, "step": 9121 }, { "epoch": 0.38238160611173405, "grad_norm": 2.0637688636779785, "learning_rate": 7.082126102351375e-06, "loss": 0.4853, "step": 9122 }, { "epoch": 0.382423524726743, "grad_norm": 1.9131648540496826, "learning_rate": 7.081508891040961e-06, "loss": 0.5241, "step": 9123 }, { "epoch": 0.382465443341752, "grad_norm": 1.9267500638961792, "learning_rate": 7.080891641360807e-06, "loss": 0.4792, "step": 9124 }, { "epoch": 0.38250736195676094, "grad_norm": 1.7162290811538696, "learning_rate": 7.080274353322294e-06, "loss": 0.541, "step": 9125 }, { "epoch": 0.3825492805717699, "grad_norm": 1.6372547149658203, "learning_rate": 7.079657026936798e-06, "loss": 0.5181, "step": 9126 }, { "epoch": 0.38259119918677886, "grad_norm": 1.5048362016677856, "learning_rate": 7.079039662215699e-06, "loss": 0.489, "step": 9127 }, { "epoch": 0.3826331178017878, "grad_norm": 1.6137596368789673, "learning_rate": 7.078422259170379e-06, "loss": 0.5332, "step": 9128 }, { "epoch": 0.3826750364167968, "grad_norm": 1.595879077911377, "learning_rate": 7.0778048178122184e-06, "loss": 0.4931, "step": 9129 }, { "epoch": 0.38271695503180575, "grad_norm": 1.877964735031128, "learning_rate": 7.077187338152598e-06, "loss": 0.4638, "step": 9130 }, { "epoch": 0.3827588736468147, "grad_norm": 1.7925605773925781, "learning_rate": 7.076569820202901e-06, "loss": 0.5528, "step": 9131 }, { "epoch": 0.3828007922618237, "grad_norm": 1.997963309288025, "learning_rate": 7.07595226397451e-06, "loss": 0.5435, "step": 9132 }, { "epoch": 0.38284271087683264, "grad_norm": 1.8613641262054443, "learning_rate": 7.075334669478808e-06, "loss": 0.5006, "step": 9133 }, { "epoch": 0.3828846294918416, "grad_norm": 1.5571995973587036, "learning_rate": 7.074717036727183e-06, "loss": 0.4897, "step": 9134 }, { "epoch": 0.38292654810685056, "grad_norm": 1.6130616664886475, "learning_rate": 7.0740993657310156e-06, "loss": 0.4819, "step": 9135 }, { "epoch": 0.3829684667218595, "grad_norm": 1.6729685068130493, "learning_rate": 7.073481656501693e-06, "loss": 0.4999, "step": 9136 }, { "epoch": 0.3830103853368685, "grad_norm": 1.5208137035369873, "learning_rate": 7.072863909050603e-06, "loss": 0.5105, "step": 9137 }, { "epoch": 0.38305230395187745, "grad_norm": 1.870524525642395, "learning_rate": 7.072246123389132e-06, "loss": 0.5648, "step": 9138 }, { "epoch": 0.3830942225668864, "grad_norm": 1.896347165107727, "learning_rate": 7.071628299528671e-06, "loss": 0.5281, "step": 9139 }, { "epoch": 0.3831361411818954, "grad_norm": 1.6572322845458984, "learning_rate": 7.071010437480606e-06, "loss": 0.5384, "step": 9140 }, { "epoch": 0.38317805979690434, "grad_norm": 2.1764602661132812, "learning_rate": 7.0703925372563245e-06, "loss": 0.5382, "step": 9141 }, { "epoch": 0.3832199784119133, "grad_norm": 1.8526731729507446, "learning_rate": 7.069774598867219e-06, "loss": 0.5243, "step": 9142 }, { "epoch": 0.3832618970269222, "grad_norm": 1.569833755493164, "learning_rate": 7.0691566223246825e-06, "loss": 0.5687, "step": 9143 }, { "epoch": 0.38330381564193117, "grad_norm": 1.5099995136260986, "learning_rate": 7.068538607640102e-06, "loss": 0.489, "step": 9144 }, { "epoch": 0.38334573425694013, "grad_norm": 1.8664461374282837, "learning_rate": 7.0679205548248715e-06, "loss": 0.5284, "step": 9145 }, { "epoch": 0.3833876528719491, "grad_norm": 1.7157223224639893, "learning_rate": 7.067302463890385e-06, "loss": 0.4999, "step": 9146 }, { "epoch": 0.38342957148695805, "grad_norm": 1.9647486209869385, "learning_rate": 7.0666843348480364e-06, "loss": 0.4813, "step": 9147 }, { "epoch": 0.383471490101967, "grad_norm": 1.737460970878601, "learning_rate": 7.066066167709219e-06, "loss": 0.4795, "step": 9148 }, { "epoch": 0.383513408716976, "grad_norm": 1.5479236841201782, "learning_rate": 7.065447962485328e-06, "loss": 0.5187, "step": 9149 }, { "epoch": 0.38355532733198494, "grad_norm": 1.7553794384002686, "learning_rate": 7.064829719187759e-06, "loss": 0.537, "step": 9150 }, { "epoch": 0.3835972459469939, "grad_norm": 1.5743966102600098, "learning_rate": 7.064211437827908e-06, "loss": 0.5142, "step": 9151 }, { "epoch": 0.38363916456200287, "grad_norm": 1.6432851552963257, "learning_rate": 7.063593118417175e-06, "loss": 0.4674, "step": 9152 }, { "epoch": 0.38368108317701183, "grad_norm": 1.7010389566421509, "learning_rate": 7.062974760966954e-06, "loss": 0.5109, "step": 9153 }, { "epoch": 0.3837230017920208, "grad_norm": 2.119208574295044, "learning_rate": 7.062356365488644e-06, "loss": 0.5489, "step": 9154 }, { "epoch": 0.38376492040702975, "grad_norm": 1.726905107498169, "learning_rate": 7.061737931993648e-06, "loss": 0.4847, "step": 9155 }, { "epoch": 0.3838068390220387, "grad_norm": 1.7333216667175293, "learning_rate": 7.061119460493362e-06, "loss": 0.4943, "step": 9156 }, { "epoch": 0.3838487576370477, "grad_norm": 1.9630662202835083, "learning_rate": 7.060500950999189e-06, "loss": 0.5392, "step": 9157 }, { "epoch": 0.38389067625205664, "grad_norm": 2.0067758560180664, "learning_rate": 7.0598824035225284e-06, "loss": 0.5558, "step": 9158 }, { "epoch": 0.3839325948670656, "grad_norm": 1.9021501541137695, "learning_rate": 7.059263818074784e-06, "loss": 0.527, "step": 9159 }, { "epoch": 0.38397451348207456, "grad_norm": 2.122234344482422, "learning_rate": 7.058645194667359e-06, "loss": 0.5565, "step": 9160 }, { "epoch": 0.3840164320970835, "grad_norm": 1.643103837966919, "learning_rate": 7.0580265333116535e-06, "loss": 0.5238, "step": 9161 }, { "epoch": 0.3840583507120925, "grad_norm": 1.8861641883850098, "learning_rate": 7.057407834019077e-06, "loss": 0.5289, "step": 9162 }, { "epoch": 0.38410026932710145, "grad_norm": 1.615640640258789, "learning_rate": 7.05678909680103e-06, "loss": 0.5308, "step": 9163 }, { "epoch": 0.3841421879421104, "grad_norm": 2.9307918548583984, "learning_rate": 7.0561703216689185e-06, "loss": 0.4703, "step": 9164 }, { "epoch": 0.3841841065571194, "grad_norm": 1.9695255756378174, "learning_rate": 7.055551508634151e-06, "loss": 0.5544, "step": 9165 }, { "epoch": 0.38422602517212834, "grad_norm": 1.4789766073226929, "learning_rate": 7.054932657708132e-06, "loss": 0.567, "step": 9166 }, { "epoch": 0.3842679437871373, "grad_norm": 2.2132010459899902, "learning_rate": 7.054313768902271e-06, "loss": 0.5384, "step": 9167 }, { "epoch": 0.3843098624021462, "grad_norm": 1.7920031547546387, "learning_rate": 7.053694842227976e-06, "loss": 0.5064, "step": 9168 }, { "epoch": 0.38435178101715517, "grad_norm": 1.6245454549789429, "learning_rate": 7.053075877696655e-06, "loss": 0.525, "step": 9169 }, { "epoch": 0.38439369963216413, "grad_norm": 2.076918125152588, "learning_rate": 7.052456875319718e-06, "loss": 0.5316, "step": 9170 }, { "epoch": 0.3844356182471731, "grad_norm": 1.882864236831665, "learning_rate": 7.051837835108577e-06, "loss": 0.5158, "step": 9171 }, { "epoch": 0.38447753686218206, "grad_norm": 2.0175375938415527, "learning_rate": 7.051218757074643e-06, "loss": 0.5471, "step": 9172 }, { "epoch": 0.384519455477191, "grad_norm": 1.9246891736984253, "learning_rate": 7.050599641229325e-06, "loss": 0.485, "step": 9173 }, { "epoch": 0.3845613740922, "grad_norm": 1.8045437335968018, "learning_rate": 7.049980487584037e-06, "loss": 0.5279, "step": 9174 }, { "epoch": 0.38460329270720894, "grad_norm": 1.7515177726745605, "learning_rate": 7.049361296150194e-06, "loss": 0.5428, "step": 9175 }, { "epoch": 0.3846452113222179, "grad_norm": 1.9315651655197144, "learning_rate": 7.0487420669392095e-06, "loss": 0.5248, "step": 9176 }, { "epoch": 0.38468712993722687, "grad_norm": 1.725338339805603, "learning_rate": 7.048122799962496e-06, "loss": 0.5516, "step": 9177 }, { "epoch": 0.38472904855223583, "grad_norm": 2.036527395248413, "learning_rate": 7.047503495231472e-06, "loss": 0.4843, "step": 9178 }, { "epoch": 0.3847709671672448, "grad_norm": 1.617765188217163, "learning_rate": 7.0468841527575485e-06, "loss": 0.4645, "step": 9179 }, { "epoch": 0.38481288578225376, "grad_norm": 1.9615004062652588, "learning_rate": 7.046264772552147e-06, "loss": 0.5304, "step": 9180 }, { "epoch": 0.3848548043972627, "grad_norm": 1.8088884353637695, "learning_rate": 7.045645354626685e-06, "loss": 0.5151, "step": 9181 }, { "epoch": 0.3848967230122717, "grad_norm": 1.7407886981964111, "learning_rate": 7.045025898992576e-06, "loss": 0.5297, "step": 9182 }, { "epoch": 0.38493864162728064, "grad_norm": 2.030856132507324, "learning_rate": 7.044406405661242e-06, "loss": 0.5508, "step": 9183 }, { "epoch": 0.3849805602422896, "grad_norm": 1.7160810232162476, "learning_rate": 7.043786874644104e-06, "loss": 0.5225, "step": 9184 }, { "epoch": 0.38502247885729857, "grad_norm": 1.9054944515228271, "learning_rate": 7.043167305952579e-06, "loss": 0.5181, "step": 9185 }, { "epoch": 0.38506439747230753, "grad_norm": 2.204662561416626, "learning_rate": 7.04254769959809e-06, "loss": 0.5538, "step": 9186 }, { "epoch": 0.3851063160873165, "grad_norm": 1.7110254764556885, "learning_rate": 7.041928055592059e-06, "loss": 0.5023, "step": 9187 }, { "epoch": 0.38514823470232545, "grad_norm": 1.551329493522644, "learning_rate": 7.041308373945905e-06, "loss": 0.4778, "step": 9188 }, { "epoch": 0.3851901533173344, "grad_norm": 1.6691209077835083, "learning_rate": 7.040688654671054e-06, "loss": 0.508, "step": 9189 }, { "epoch": 0.3852320719323434, "grad_norm": 1.755906581878662, "learning_rate": 7.04006889777893e-06, "loss": 0.5257, "step": 9190 }, { "epoch": 0.38527399054735234, "grad_norm": 1.8428399562835693, "learning_rate": 7.039449103280954e-06, "loss": 0.5052, "step": 9191 }, { "epoch": 0.3853159091623613, "grad_norm": 2.001518964767456, "learning_rate": 7.038829271188554e-06, "loss": 0.5456, "step": 9192 }, { "epoch": 0.3853578277773702, "grad_norm": 1.6606605052947998, "learning_rate": 7.038209401513157e-06, "loss": 0.5405, "step": 9193 }, { "epoch": 0.38539974639237917, "grad_norm": 1.58356773853302, "learning_rate": 7.037589494266185e-06, "loss": 0.4931, "step": 9194 }, { "epoch": 0.38544166500738813, "grad_norm": 1.8208730220794678, "learning_rate": 7.036969549459068e-06, "loss": 0.5076, "step": 9195 }, { "epoch": 0.3854835836223971, "grad_norm": 2.0287938117980957, "learning_rate": 7.0363495671032345e-06, "loss": 0.5118, "step": 9196 }, { "epoch": 0.38552550223740606, "grad_norm": 1.763979434967041, "learning_rate": 7.035729547210111e-06, "loss": 0.5826, "step": 9197 }, { "epoch": 0.385567420852415, "grad_norm": 5.142490863800049, "learning_rate": 7.035109489791128e-06, "loss": 0.4938, "step": 9198 }, { "epoch": 0.385609339467424, "grad_norm": 1.6049760580062866, "learning_rate": 7.0344893948577145e-06, "loss": 0.5148, "step": 9199 }, { "epoch": 0.38565125808243295, "grad_norm": 1.9066044092178345, "learning_rate": 7.0338692624213025e-06, "loss": 0.5538, "step": 9200 }, { "epoch": 0.3856931766974419, "grad_norm": 1.7596203088760376, "learning_rate": 7.033249092493321e-06, "loss": 0.502, "step": 9201 }, { "epoch": 0.38573509531245087, "grad_norm": 1.822078824043274, "learning_rate": 7.032628885085205e-06, "loss": 0.4663, "step": 9202 }, { "epoch": 0.38577701392745983, "grad_norm": 1.9847080707550049, "learning_rate": 7.032008640208383e-06, "loss": 0.5035, "step": 9203 }, { "epoch": 0.3858189325424688, "grad_norm": 1.7227391004562378, "learning_rate": 7.031388357874294e-06, "loss": 0.5481, "step": 9204 }, { "epoch": 0.38586085115747776, "grad_norm": 1.5698342323303223, "learning_rate": 7.030768038094366e-06, "loss": 0.5046, "step": 9205 }, { "epoch": 0.3859027697724867, "grad_norm": 1.9694862365722656, "learning_rate": 7.030147680880039e-06, "loss": 0.5443, "step": 9206 }, { "epoch": 0.3859446883874957, "grad_norm": 1.9346725940704346, "learning_rate": 7.029527286242745e-06, "loss": 0.5047, "step": 9207 }, { "epoch": 0.38598660700250464, "grad_norm": 1.6165688037872314, "learning_rate": 7.028906854193921e-06, "loss": 0.4841, "step": 9208 }, { "epoch": 0.3860285256175136, "grad_norm": 1.6127948760986328, "learning_rate": 7.028286384745006e-06, "loss": 0.4903, "step": 9209 }, { "epoch": 0.38607044423252257, "grad_norm": 1.686574101448059, "learning_rate": 7.027665877907433e-06, "loss": 0.5181, "step": 9210 }, { "epoch": 0.38611236284753153, "grad_norm": 1.5787631273269653, "learning_rate": 7.027045333692643e-06, "loss": 0.5163, "step": 9211 }, { "epoch": 0.3861542814625405, "grad_norm": 1.831336498260498, "learning_rate": 7.026424752112075e-06, "loss": 0.5065, "step": 9212 }, { "epoch": 0.38619620007754946, "grad_norm": 1.7584030628204346, "learning_rate": 7.025804133177169e-06, "loss": 0.5502, "step": 9213 }, { "epoch": 0.3862381186925584, "grad_norm": 1.4662351608276367, "learning_rate": 7.025183476899364e-06, "loss": 0.4923, "step": 9214 }, { "epoch": 0.3862800373075674, "grad_norm": 1.6950627565383911, "learning_rate": 7.024562783290101e-06, "loss": 0.4885, "step": 9215 }, { "epoch": 0.38632195592257634, "grad_norm": 1.8047596216201782, "learning_rate": 7.023942052360821e-06, "loss": 0.4857, "step": 9216 }, { "epoch": 0.3863638745375853, "grad_norm": 2.0733954906463623, "learning_rate": 7.023321284122968e-06, "loss": 0.55, "step": 9217 }, { "epoch": 0.3864057931525942, "grad_norm": 1.6865893602371216, "learning_rate": 7.022700478587987e-06, "loss": 0.4993, "step": 9218 }, { "epoch": 0.3864477117676032, "grad_norm": 1.812738060951233, "learning_rate": 7.022079635767315e-06, "loss": 0.5797, "step": 9219 }, { "epoch": 0.38648963038261214, "grad_norm": 1.7002090215682983, "learning_rate": 7.021458755672403e-06, "loss": 0.5184, "step": 9220 }, { "epoch": 0.3865315489976211, "grad_norm": 1.4310506582260132, "learning_rate": 7.020837838314691e-06, "loss": 0.515, "step": 9221 }, { "epoch": 0.38657346761263006, "grad_norm": 1.6702616214752197, "learning_rate": 7.02021688370563e-06, "loss": 0.4941, "step": 9222 }, { "epoch": 0.386615386227639, "grad_norm": 2.4348583221435547, "learning_rate": 7.019595891856662e-06, "loss": 0.54, "step": 9223 }, { "epoch": 0.386657304842648, "grad_norm": 1.7032655477523804, "learning_rate": 7.018974862779235e-06, "loss": 0.5183, "step": 9224 }, { "epoch": 0.38669922345765695, "grad_norm": 2.787684679031372, "learning_rate": 7.018353796484798e-06, "loss": 0.5213, "step": 9225 }, { "epoch": 0.3867411420726659, "grad_norm": 1.9891754388809204, "learning_rate": 7.0177326929848e-06, "loss": 0.5409, "step": 9226 }, { "epoch": 0.3867830606876749, "grad_norm": 1.7741740942001343, "learning_rate": 7.017111552290689e-06, "loss": 0.5388, "step": 9227 }, { "epoch": 0.38682497930268384, "grad_norm": 1.9287161827087402, "learning_rate": 7.016490374413914e-06, "loss": 0.528, "step": 9228 }, { "epoch": 0.3868668979176928, "grad_norm": 1.730132818222046, "learning_rate": 7.015869159365929e-06, "loss": 0.5253, "step": 9229 }, { "epoch": 0.38690881653270176, "grad_norm": 1.7845853567123413, "learning_rate": 7.015247907158181e-06, "loss": 0.5419, "step": 9230 }, { "epoch": 0.3869507351477107, "grad_norm": 1.6780022382736206, "learning_rate": 7.0146266178021224e-06, "loss": 0.5392, "step": 9231 }, { "epoch": 0.3869926537627197, "grad_norm": 1.57187819480896, "learning_rate": 7.014005291309212e-06, "loss": 0.5345, "step": 9232 }, { "epoch": 0.38703457237772865, "grad_norm": 1.7279472351074219, "learning_rate": 7.013383927690895e-06, "loss": 0.546, "step": 9233 }, { "epoch": 0.3870764909927376, "grad_norm": 1.7297468185424805, "learning_rate": 7.012762526958629e-06, "loss": 0.5212, "step": 9234 }, { "epoch": 0.38711840960774657, "grad_norm": 1.8731846809387207, "learning_rate": 7.012141089123867e-06, "loss": 0.4713, "step": 9235 }, { "epoch": 0.38716032822275553, "grad_norm": 1.9333926439285278, "learning_rate": 7.0115196141980676e-06, "loss": 0.5104, "step": 9236 }, { "epoch": 0.3872022468377645, "grad_norm": 1.6508291959762573, "learning_rate": 7.010898102192686e-06, "loss": 0.5196, "step": 9237 }, { "epoch": 0.38724416545277346, "grad_norm": 1.8952773809432983, "learning_rate": 7.010276553119175e-06, "loss": 0.5983, "step": 9238 }, { "epoch": 0.3872860840677824, "grad_norm": 1.9851070642471313, "learning_rate": 7.009654966988997e-06, "loss": 0.5172, "step": 9239 }, { "epoch": 0.3873280026827914, "grad_norm": 1.5304573774337769, "learning_rate": 7.009033343813607e-06, "loss": 0.568, "step": 9240 }, { "epoch": 0.38736992129780035, "grad_norm": 1.468733787536621, "learning_rate": 7.008411683604465e-06, "loss": 0.4644, "step": 9241 }, { "epoch": 0.3874118399128093, "grad_norm": 1.8908379077911377, "learning_rate": 7.007789986373031e-06, "loss": 0.5356, "step": 9242 }, { "epoch": 0.3874537585278182, "grad_norm": 1.856192708015442, "learning_rate": 7.007168252130764e-06, "loss": 0.5438, "step": 9243 }, { "epoch": 0.3874956771428272, "grad_norm": 1.6537836790084839, "learning_rate": 7.006546480889123e-06, "loss": 0.4969, "step": 9244 }, { "epoch": 0.38753759575783614, "grad_norm": 1.6852020025253296, "learning_rate": 7.005924672659573e-06, "loss": 0.5881, "step": 9245 }, { "epoch": 0.3875795143728451, "grad_norm": 1.5812013149261475, "learning_rate": 7.005302827453575e-06, "loss": 0.5207, "step": 9246 }, { "epoch": 0.38762143298785406, "grad_norm": 1.705620527267456, "learning_rate": 7.004680945282591e-06, "loss": 0.51, "step": 9247 }, { "epoch": 0.387663351602863, "grad_norm": 1.5220668315887451, "learning_rate": 7.004059026158087e-06, "loss": 0.54, "step": 9248 }, { "epoch": 0.387705270217872, "grad_norm": 1.8299922943115234, "learning_rate": 7.003437070091524e-06, "loss": 0.5315, "step": 9249 }, { "epoch": 0.38774718883288095, "grad_norm": 6.385882377624512, "learning_rate": 7.002815077094367e-06, "loss": 0.5619, "step": 9250 }, { "epoch": 0.3877891074478899, "grad_norm": 1.7242189645767212, "learning_rate": 7.002193047178086e-06, "loss": 0.5254, "step": 9251 }, { "epoch": 0.3878310260628989, "grad_norm": 1.9225032329559326, "learning_rate": 7.001570980354142e-06, "loss": 0.4933, "step": 9252 }, { "epoch": 0.38787294467790784, "grad_norm": 1.6483113765716553, "learning_rate": 7.000948876634004e-06, "loss": 0.5387, "step": 9253 }, { "epoch": 0.3879148632929168, "grad_norm": 1.5914040803909302, "learning_rate": 7.00032673602914e-06, "loss": 0.5304, "step": 9254 }, { "epoch": 0.38795678190792576, "grad_norm": 1.6886072158813477, "learning_rate": 6.9997045585510185e-06, "loss": 0.5311, "step": 9255 }, { "epoch": 0.3879987005229347, "grad_norm": 1.6066093444824219, "learning_rate": 6.999082344211108e-06, "loss": 0.516, "step": 9256 }, { "epoch": 0.3880406191379437, "grad_norm": 1.6568117141723633, "learning_rate": 6.998460093020877e-06, "loss": 0.4811, "step": 9257 }, { "epoch": 0.38808253775295265, "grad_norm": 1.654974341392517, "learning_rate": 6.997837804991798e-06, "loss": 0.5471, "step": 9258 }, { "epoch": 0.3881244563679616, "grad_norm": 1.57505202293396, "learning_rate": 6.99721548013534e-06, "loss": 0.4748, "step": 9259 }, { "epoch": 0.3881663749829706, "grad_norm": 1.6363552808761597, "learning_rate": 6.996593118462978e-06, "loss": 0.4888, "step": 9260 }, { "epoch": 0.38820829359797954, "grad_norm": 1.8432049751281738, "learning_rate": 6.995970719986181e-06, "loss": 0.5089, "step": 9261 }, { "epoch": 0.3882502122129885, "grad_norm": 1.5926060676574707, "learning_rate": 6.99534828471642e-06, "loss": 0.5324, "step": 9262 }, { "epoch": 0.38829213082799746, "grad_norm": 1.911421298980713, "learning_rate": 6.994725812665175e-06, "loss": 0.534, "step": 9263 }, { "epoch": 0.3883340494430064, "grad_norm": 1.9748256206512451, "learning_rate": 6.994103303843916e-06, "loss": 0.5119, "step": 9264 }, { "epoch": 0.3883759680580154, "grad_norm": 1.9191615581512451, "learning_rate": 6.993480758264121e-06, "loss": 0.5006, "step": 9265 }, { "epoch": 0.38841788667302435, "grad_norm": 1.57852041721344, "learning_rate": 6.992858175937262e-06, "loss": 0.4803, "step": 9266 }, { "epoch": 0.3884598052880333, "grad_norm": 3.266833543777466, "learning_rate": 6.99223555687482e-06, "loss": 0.5229, "step": 9267 }, { "epoch": 0.3885017239030422, "grad_norm": 1.7922592163085938, "learning_rate": 6.991612901088267e-06, "loss": 0.4951, "step": 9268 }, { "epoch": 0.3885436425180512, "grad_norm": 1.564957857131958, "learning_rate": 6.990990208589086e-06, "loss": 0.5371, "step": 9269 }, { "epoch": 0.38858556113306014, "grad_norm": 1.636455774307251, "learning_rate": 6.990367479388752e-06, "loss": 0.491, "step": 9270 }, { "epoch": 0.3886274797480691, "grad_norm": 1.5755354166030884, "learning_rate": 6.989744713498745e-06, "loss": 0.5499, "step": 9271 }, { "epoch": 0.38866939836307807, "grad_norm": 1.71567964553833, "learning_rate": 6.989121910930544e-06, "loss": 0.5756, "step": 9272 }, { "epoch": 0.38871131697808703, "grad_norm": 1.6570936441421509, "learning_rate": 6.9884990716956315e-06, "loss": 0.5262, "step": 9273 }, { "epoch": 0.388753235593096, "grad_norm": 1.72040855884552, "learning_rate": 6.9878761958054874e-06, "loss": 0.4982, "step": 9274 }, { "epoch": 0.38879515420810495, "grad_norm": 2.57554292678833, "learning_rate": 6.987253283271594e-06, "loss": 0.5177, "step": 9275 }, { "epoch": 0.3888370728231139, "grad_norm": 1.6630711555480957, "learning_rate": 6.986630334105434e-06, "loss": 0.5225, "step": 9276 }, { "epoch": 0.3888789914381229, "grad_norm": 1.6876879930496216, "learning_rate": 6.986007348318491e-06, "loss": 0.4874, "step": 9277 }, { "epoch": 0.38892091005313184, "grad_norm": 1.798668384552002, "learning_rate": 6.985384325922245e-06, "loss": 0.4831, "step": 9278 }, { "epoch": 0.3889628286681408, "grad_norm": 1.7971118688583374, "learning_rate": 6.9847612669281875e-06, "loss": 0.4862, "step": 9279 }, { "epoch": 0.38900474728314977, "grad_norm": 1.7061500549316406, "learning_rate": 6.984138171347797e-06, "loss": 0.5789, "step": 9280 }, { "epoch": 0.38904666589815873, "grad_norm": 1.7014894485473633, "learning_rate": 6.983515039192563e-06, "loss": 0.5053, "step": 9281 }, { "epoch": 0.3890885845131677, "grad_norm": 1.5943655967712402, "learning_rate": 6.982891870473971e-06, "loss": 0.4705, "step": 9282 }, { "epoch": 0.38913050312817665, "grad_norm": 1.5691183805465698, "learning_rate": 6.98226866520351e-06, "loss": 0.5135, "step": 9283 }, { "epoch": 0.3891724217431856, "grad_norm": 1.708746314048767, "learning_rate": 6.981645423392665e-06, "loss": 0.5239, "step": 9284 }, { "epoch": 0.3892143403581946, "grad_norm": 1.5997707843780518, "learning_rate": 6.9810221450529285e-06, "loss": 0.499, "step": 9285 }, { "epoch": 0.38925625897320354, "grad_norm": 1.664949893951416, "learning_rate": 6.980398830195785e-06, "loss": 0.5255, "step": 9286 }, { "epoch": 0.3892981775882125, "grad_norm": 1.6791797876358032, "learning_rate": 6.979775478832726e-06, "loss": 0.4788, "step": 9287 }, { "epoch": 0.38934009620322146, "grad_norm": 1.5772736072540283, "learning_rate": 6.979152090975247e-06, "loss": 0.4832, "step": 9288 }, { "epoch": 0.3893820148182304, "grad_norm": 2.0646915435791016, "learning_rate": 6.978528666634833e-06, "loss": 0.545, "step": 9289 }, { "epoch": 0.3894239334332394, "grad_norm": 2.0641608238220215, "learning_rate": 6.977905205822978e-06, "loss": 0.5188, "step": 9290 }, { "epoch": 0.38946585204824835, "grad_norm": 1.7475143671035767, "learning_rate": 6.977281708551175e-06, "loss": 0.4868, "step": 9291 }, { "epoch": 0.3895077706632573, "grad_norm": 1.731971025466919, "learning_rate": 6.976658174830917e-06, "loss": 0.5314, "step": 9292 }, { "epoch": 0.3895496892782662, "grad_norm": 1.5971866846084595, "learning_rate": 6.976034604673699e-06, "loss": 0.5341, "step": 9293 }, { "epoch": 0.3895916078932752, "grad_norm": 2.5137364864349365, "learning_rate": 6.975410998091014e-06, "loss": 0.5246, "step": 9294 }, { "epoch": 0.38963352650828414, "grad_norm": 1.7284891605377197, "learning_rate": 6.974787355094359e-06, "loss": 0.5965, "step": 9295 }, { "epoch": 0.3896754451232931, "grad_norm": 1.7437481880187988, "learning_rate": 6.974163675695229e-06, "loss": 0.5537, "step": 9296 }, { "epoch": 0.38971736373830207, "grad_norm": 1.5182698965072632, "learning_rate": 6.97353995990512e-06, "loss": 0.4787, "step": 9297 }, { "epoch": 0.38975928235331103, "grad_norm": 2.015638828277588, "learning_rate": 6.972916207735532e-06, "loss": 0.4903, "step": 9298 }, { "epoch": 0.38980120096832, "grad_norm": 1.5178014039993286, "learning_rate": 6.972292419197959e-06, "loss": 0.4826, "step": 9299 }, { "epoch": 0.38984311958332896, "grad_norm": 1.430252194404602, "learning_rate": 6.9716685943039035e-06, "loss": 0.467, "step": 9300 }, { "epoch": 0.3898850381983379, "grad_norm": 1.6579593420028687, "learning_rate": 6.971044733064863e-06, "loss": 0.5017, "step": 9301 }, { "epoch": 0.3899269568133469, "grad_norm": 1.488195538520813, "learning_rate": 6.9704208354923376e-06, "loss": 0.5076, "step": 9302 }, { "epoch": 0.38996887542835584, "grad_norm": 1.578783631324768, "learning_rate": 6.969796901597828e-06, "loss": 0.4866, "step": 9303 }, { "epoch": 0.3900107940433648, "grad_norm": 1.7790892124176025, "learning_rate": 6.9691729313928355e-06, "loss": 0.5721, "step": 9304 }, { "epoch": 0.39005271265837377, "grad_norm": 2.2418875694274902, "learning_rate": 6.968548924888861e-06, "loss": 0.5402, "step": 9305 }, { "epoch": 0.39009463127338273, "grad_norm": 1.947777509689331, "learning_rate": 6.96792488209741e-06, "loss": 0.5159, "step": 9306 }, { "epoch": 0.3901365498883917, "grad_norm": 1.534895658493042, "learning_rate": 6.967300803029986e-06, "loss": 0.5003, "step": 9307 }, { "epoch": 0.39017846850340066, "grad_norm": 1.5704054832458496, "learning_rate": 6.9666766876980885e-06, "loss": 0.5071, "step": 9308 }, { "epoch": 0.3902203871184096, "grad_norm": 1.7764818668365479, "learning_rate": 6.966052536113226e-06, "loss": 0.5745, "step": 9309 }, { "epoch": 0.3902623057334186, "grad_norm": 1.6466846466064453, "learning_rate": 6.965428348286904e-06, "loss": 0.5508, "step": 9310 }, { "epoch": 0.39030422434842754, "grad_norm": 1.662733554840088, "learning_rate": 6.964804124230627e-06, "loss": 0.5096, "step": 9311 }, { "epoch": 0.3903461429634365, "grad_norm": 1.7985451221466064, "learning_rate": 6.964179863955903e-06, "loss": 0.4784, "step": 9312 }, { "epoch": 0.39038806157844547, "grad_norm": 2.115835428237915, "learning_rate": 6.963555567474236e-06, "loss": 0.5284, "step": 9313 }, { "epoch": 0.39042998019345443, "grad_norm": 1.6573518514633179, "learning_rate": 6.962931234797139e-06, "loss": 0.4862, "step": 9314 }, { "epoch": 0.3904718988084634, "grad_norm": 2.020656108856201, "learning_rate": 6.962306865936116e-06, "loss": 0.5327, "step": 9315 }, { "epoch": 0.39051381742347235, "grad_norm": 1.8117396831512451, "learning_rate": 6.961682460902681e-06, "loss": 0.5422, "step": 9316 }, { "epoch": 0.3905557360384813, "grad_norm": 1.541273593902588, "learning_rate": 6.961058019708342e-06, "loss": 0.5591, "step": 9317 }, { "epoch": 0.3905976546534902, "grad_norm": 1.8388735055923462, "learning_rate": 6.960433542364607e-06, "loss": 0.5849, "step": 9318 }, { "epoch": 0.3906395732684992, "grad_norm": 1.6641465425491333, "learning_rate": 6.959809028882992e-06, "loss": 0.5275, "step": 9319 }, { "epoch": 0.39068149188350815, "grad_norm": 2.026340961456299, "learning_rate": 6.959184479275007e-06, "loss": 0.5138, "step": 9320 }, { "epoch": 0.3907234104985171, "grad_norm": 1.6532424688339233, "learning_rate": 6.958559893552164e-06, "loss": 0.5023, "step": 9321 }, { "epoch": 0.39076532911352607, "grad_norm": 1.8013100624084473, "learning_rate": 6.957935271725977e-06, "loss": 0.5975, "step": 9322 }, { "epoch": 0.39080724772853503, "grad_norm": 1.633955478668213, "learning_rate": 6.957310613807961e-06, "loss": 0.5115, "step": 9323 }, { "epoch": 0.390849166343544, "grad_norm": 1.6898791790008545, "learning_rate": 6.956685919809627e-06, "loss": 0.4738, "step": 9324 }, { "epoch": 0.39089108495855296, "grad_norm": 2.5513901710510254, "learning_rate": 6.956061189742495e-06, "loss": 0.4876, "step": 9325 }, { "epoch": 0.3909330035735619, "grad_norm": 1.7960313558578491, "learning_rate": 6.955436423618081e-06, "loss": 0.4896, "step": 9326 }, { "epoch": 0.3909749221885709, "grad_norm": 1.7897673845291138, "learning_rate": 6.954811621447897e-06, "loss": 0.5045, "step": 9327 }, { "epoch": 0.39101684080357985, "grad_norm": 1.549193263053894, "learning_rate": 6.954186783243463e-06, "loss": 0.5187, "step": 9328 }, { "epoch": 0.3910587594185888, "grad_norm": 1.655963659286499, "learning_rate": 6.953561909016299e-06, "loss": 0.5362, "step": 9329 }, { "epoch": 0.39110067803359777, "grad_norm": 2.0007874965667725, "learning_rate": 6.952936998777921e-06, "loss": 0.5666, "step": 9330 }, { "epoch": 0.39114259664860673, "grad_norm": 1.6262483596801758, "learning_rate": 6.952312052539849e-06, "loss": 0.5022, "step": 9331 }, { "epoch": 0.3911845152636157, "grad_norm": 1.6627099514007568, "learning_rate": 6.951687070313604e-06, "loss": 0.5026, "step": 9332 }, { "epoch": 0.39122643387862466, "grad_norm": 1.6063148975372314, "learning_rate": 6.951062052110705e-06, "loss": 0.4856, "step": 9333 }, { "epoch": 0.3912683524936336, "grad_norm": 1.6296292543411255, "learning_rate": 6.9504369979426755e-06, "loss": 0.4738, "step": 9334 }, { "epoch": 0.3913102711086426, "grad_norm": 2.0531044006347656, "learning_rate": 6.949811907821037e-06, "loss": 0.5571, "step": 9335 }, { "epoch": 0.39135218972365154, "grad_norm": 1.7148863077163696, "learning_rate": 6.949186781757309e-06, "loss": 0.4818, "step": 9336 }, { "epoch": 0.3913941083386605, "grad_norm": 1.6922810077667236, "learning_rate": 6.948561619763019e-06, "loss": 0.5251, "step": 9337 }, { "epoch": 0.39143602695366947, "grad_norm": 1.5339006185531616, "learning_rate": 6.94793642184969e-06, "loss": 0.516, "step": 9338 }, { "epoch": 0.39147794556867843, "grad_norm": 1.6569336652755737, "learning_rate": 6.947311188028845e-06, "loss": 0.5222, "step": 9339 }, { "epoch": 0.3915198641836874, "grad_norm": 1.5164536237716675, "learning_rate": 6.94668591831201e-06, "loss": 0.5062, "step": 9340 }, { "epoch": 0.39156178279869636, "grad_norm": 1.5529677867889404, "learning_rate": 6.9460606127107124e-06, "loss": 0.4631, "step": 9341 }, { "epoch": 0.3916037014137053, "grad_norm": 1.8220844268798828, "learning_rate": 6.945435271236476e-06, "loss": 0.5174, "step": 9342 }, { "epoch": 0.3916456200287142, "grad_norm": 1.8494329452514648, "learning_rate": 6.944809893900832e-06, "loss": 0.5147, "step": 9343 }, { "epoch": 0.3916875386437232, "grad_norm": 1.6748791933059692, "learning_rate": 6.944184480715305e-06, "loss": 0.4809, "step": 9344 }, { "epoch": 0.39172945725873215, "grad_norm": 1.8859953880310059, "learning_rate": 6.9435590316914246e-06, "loss": 0.4618, "step": 9345 }, { "epoch": 0.3917713758737411, "grad_norm": 1.5352586507797241, "learning_rate": 6.942933546840721e-06, "loss": 0.4691, "step": 9346 }, { "epoch": 0.3918132944887501, "grad_norm": 1.8261538743972778, "learning_rate": 6.942308026174723e-06, "loss": 0.5458, "step": 9347 }, { "epoch": 0.39185521310375904, "grad_norm": 1.7025880813598633, "learning_rate": 6.9416824697049624e-06, "loss": 0.4881, "step": 9348 }, { "epoch": 0.391897131718768, "grad_norm": 1.693960428237915, "learning_rate": 6.94105687744297e-06, "loss": 0.5714, "step": 9349 }, { "epoch": 0.39193905033377696, "grad_norm": 1.8494974374771118, "learning_rate": 6.940431249400277e-06, "loss": 0.4734, "step": 9350 }, { "epoch": 0.3919809689487859, "grad_norm": 1.8751126527786255, "learning_rate": 6.939805585588416e-06, "loss": 0.5343, "step": 9351 }, { "epoch": 0.3920228875637949, "grad_norm": 2.0714709758758545, "learning_rate": 6.939179886018921e-06, "loss": 0.5241, "step": 9352 }, { "epoch": 0.39206480617880385, "grad_norm": 2.4747636318206787, "learning_rate": 6.938554150703325e-06, "loss": 0.5566, "step": 9353 }, { "epoch": 0.3921067247938128, "grad_norm": 1.725739598274231, "learning_rate": 6.937928379653165e-06, "loss": 0.5098, "step": 9354 }, { "epoch": 0.3921486434088218, "grad_norm": 1.6624484062194824, "learning_rate": 6.9373025728799724e-06, "loss": 0.5251, "step": 9355 }, { "epoch": 0.39219056202383074, "grad_norm": 1.6276538372039795, "learning_rate": 6.936676730395287e-06, "loss": 0.4997, "step": 9356 }, { "epoch": 0.3922324806388397, "grad_norm": 1.7838221788406372, "learning_rate": 6.936050852210643e-06, "loss": 0.4794, "step": 9357 }, { "epoch": 0.39227439925384866, "grad_norm": 1.7671524286270142, "learning_rate": 6.9354249383375775e-06, "loss": 0.4783, "step": 9358 }, { "epoch": 0.3923163178688576, "grad_norm": 1.6447017192840576, "learning_rate": 6.934798988787629e-06, "loss": 0.4537, "step": 9359 }, { "epoch": 0.3923582364838666, "grad_norm": 1.8579332828521729, "learning_rate": 6.934173003572335e-06, "loss": 0.4476, "step": 9360 }, { "epoch": 0.39240015509887555, "grad_norm": 1.5974547863006592, "learning_rate": 6.9335469827032366e-06, "loss": 0.494, "step": 9361 }, { "epoch": 0.3924420737138845, "grad_norm": 2.0129005908966064, "learning_rate": 6.932920926191872e-06, "loss": 0.5336, "step": 9362 }, { "epoch": 0.39248399232889347, "grad_norm": 1.6463494300842285, "learning_rate": 6.932294834049783e-06, "loss": 0.497, "step": 9363 }, { "epoch": 0.39252591094390243, "grad_norm": 1.5172256231307983, "learning_rate": 6.93166870628851e-06, "loss": 0.4978, "step": 9364 }, { "epoch": 0.3925678295589114, "grad_norm": 1.8944132328033447, "learning_rate": 6.931042542919595e-06, "loss": 0.5227, "step": 9365 }, { "epoch": 0.39260974817392036, "grad_norm": 1.6920626163482666, "learning_rate": 6.93041634395458e-06, "loss": 0.5036, "step": 9366 }, { "epoch": 0.3926516667889293, "grad_norm": 1.8698437213897705, "learning_rate": 6.929790109405008e-06, "loss": 0.5579, "step": 9367 }, { "epoch": 0.39269358540393823, "grad_norm": 1.703359842300415, "learning_rate": 6.929163839282425e-06, "loss": 0.537, "step": 9368 }, { "epoch": 0.3927355040189472, "grad_norm": 2.4973654747009277, "learning_rate": 6.9285375335983715e-06, "loss": 0.563, "step": 9369 }, { "epoch": 0.39277742263395615, "grad_norm": 1.869078516960144, "learning_rate": 6.9279111923643946e-06, "loss": 0.5129, "step": 9370 }, { "epoch": 0.3928193412489651, "grad_norm": 1.8020527362823486, "learning_rate": 6.9272848155920415e-06, "loss": 0.4651, "step": 9371 }, { "epoch": 0.3928612598639741, "grad_norm": 1.7610256671905518, "learning_rate": 6.926658403292857e-06, "loss": 0.5157, "step": 9372 }, { "epoch": 0.39290317847898304, "grad_norm": 1.6208702325820923, "learning_rate": 6.9260319554783875e-06, "loss": 0.5087, "step": 9373 }, { "epoch": 0.392945097093992, "grad_norm": 2.2151641845703125, "learning_rate": 6.9254054721601826e-06, "loss": 0.4845, "step": 9374 }, { "epoch": 0.39298701570900096, "grad_norm": 1.8737082481384277, "learning_rate": 6.9247789533497886e-06, "loss": 0.5517, "step": 9375 }, { "epoch": 0.3930289343240099, "grad_norm": 2.865933656692505, "learning_rate": 6.924152399058756e-06, "loss": 0.4362, "step": 9376 }, { "epoch": 0.3930708529390189, "grad_norm": 1.490028738975525, "learning_rate": 6.923525809298635e-06, "loss": 0.5164, "step": 9377 }, { "epoch": 0.39311277155402785, "grad_norm": 1.917353630065918, "learning_rate": 6.922899184080974e-06, "loss": 0.564, "step": 9378 }, { "epoch": 0.3931546901690368, "grad_norm": 1.8252780437469482, "learning_rate": 6.922272523417324e-06, "loss": 0.5192, "step": 9379 }, { "epoch": 0.3931966087840458, "grad_norm": 2.174093723297119, "learning_rate": 6.921645827319238e-06, "loss": 0.4858, "step": 9380 }, { "epoch": 0.39323852739905474, "grad_norm": 1.7414833307266235, "learning_rate": 6.921019095798268e-06, "loss": 0.5102, "step": 9381 }, { "epoch": 0.3932804460140637, "grad_norm": 1.8199149370193481, "learning_rate": 6.920392328865967e-06, "loss": 0.491, "step": 9382 }, { "epoch": 0.39332236462907266, "grad_norm": 1.7693980932235718, "learning_rate": 6.919765526533888e-06, "loss": 0.5001, "step": 9383 }, { "epoch": 0.3933642832440816, "grad_norm": 1.5709954500198364, "learning_rate": 6.919138688813586e-06, "loss": 0.5168, "step": 9384 }, { "epoch": 0.3934062018590906, "grad_norm": 1.9784897565841675, "learning_rate": 6.9185118157166145e-06, "loss": 0.5142, "step": 9385 }, { "epoch": 0.39344812047409955, "grad_norm": 1.9802517890930176, "learning_rate": 6.917884907254531e-06, "loss": 0.5042, "step": 9386 }, { "epoch": 0.3934900390891085, "grad_norm": 1.5357651710510254, "learning_rate": 6.91725796343889e-06, "loss": 0.4822, "step": 9387 }, { "epoch": 0.3935319577041175, "grad_norm": 1.4787859916687012, "learning_rate": 6.9166309842812485e-06, "loss": 0.4791, "step": 9388 }, { "epoch": 0.39357387631912644, "grad_norm": 1.960544228553772, "learning_rate": 6.916003969793165e-06, "loss": 0.5275, "step": 9389 }, { "epoch": 0.3936157949341354, "grad_norm": 2.0882129669189453, "learning_rate": 6.915376919986197e-06, "loss": 0.532, "step": 9390 }, { "epoch": 0.39365771354914436, "grad_norm": 1.6694034337997437, "learning_rate": 6.914749834871904e-06, "loss": 0.4916, "step": 9391 }, { "epoch": 0.3936996321641533, "grad_norm": 1.708152413368225, "learning_rate": 6.914122714461844e-06, "loss": 0.4682, "step": 9392 }, { "epoch": 0.39374155077916223, "grad_norm": 1.7951295375823975, "learning_rate": 6.913495558767578e-06, "loss": 0.5699, "step": 9393 }, { "epoch": 0.3937834693941712, "grad_norm": 1.4676259756088257, "learning_rate": 6.912868367800667e-06, "loss": 0.5315, "step": 9394 }, { "epoch": 0.39382538800918016, "grad_norm": 1.5963680744171143, "learning_rate": 6.912241141572671e-06, "loss": 0.5141, "step": 9395 }, { "epoch": 0.3938673066241891, "grad_norm": 1.4967039823532104, "learning_rate": 6.911613880095156e-06, "loss": 0.476, "step": 9396 }, { "epoch": 0.3939092252391981, "grad_norm": 1.8668233156204224, "learning_rate": 6.91098658337968e-06, "loss": 0.5437, "step": 9397 }, { "epoch": 0.39395114385420704, "grad_norm": 1.8068002462387085, "learning_rate": 6.910359251437807e-06, "loss": 0.5159, "step": 9398 }, { "epoch": 0.393993062469216, "grad_norm": 1.7234132289886475, "learning_rate": 6.909731884281103e-06, "loss": 0.5171, "step": 9399 }, { "epoch": 0.39403498108422497, "grad_norm": 1.9821664094924927, "learning_rate": 6.909104481921133e-06, "loss": 0.4542, "step": 9400 }, { "epoch": 0.39407689969923393, "grad_norm": 2.8754043579101562, "learning_rate": 6.908477044369461e-06, "loss": 0.4972, "step": 9401 }, { "epoch": 0.3941188183142429, "grad_norm": 4.107980251312256, "learning_rate": 6.907849571637653e-06, "loss": 0.5204, "step": 9402 }, { "epoch": 0.39416073692925185, "grad_norm": 1.837775468826294, "learning_rate": 6.907222063737274e-06, "loss": 0.5083, "step": 9403 }, { "epoch": 0.3942026555442608, "grad_norm": 1.7427736520767212, "learning_rate": 6.9065945206798935e-06, "loss": 0.5506, "step": 9404 }, { "epoch": 0.3942445741592698, "grad_norm": 1.9355303049087524, "learning_rate": 6.9059669424770805e-06, "loss": 0.5051, "step": 9405 }, { "epoch": 0.39428649277427874, "grad_norm": 2.2724976539611816, "learning_rate": 6.905339329140401e-06, "loss": 0.459, "step": 9406 }, { "epoch": 0.3943284113892877, "grad_norm": 1.8393968343734741, "learning_rate": 6.904711680681424e-06, "loss": 0.6216, "step": 9407 }, { "epoch": 0.39437033000429667, "grad_norm": 1.9209727048873901, "learning_rate": 6.904083997111721e-06, "loss": 0.5237, "step": 9408 }, { "epoch": 0.3944122486193056, "grad_norm": 1.713167428970337, "learning_rate": 6.903456278442863e-06, "loss": 0.539, "step": 9409 }, { "epoch": 0.3944541672343146, "grad_norm": 1.743743658065796, "learning_rate": 6.902828524686418e-06, "loss": 0.5226, "step": 9410 }, { "epoch": 0.39449608584932355, "grad_norm": 1.841750979423523, "learning_rate": 6.902200735853961e-06, "loss": 0.5593, "step": 9411 }, { "epoch": 0.3945380044643325, "grad_norm": 1.5561131238937378, "learning_rate": 6.901572911957063e-06, "loss": 0.5411, "step": 9412 }, { "epoch": 0.3945799230793415, "grad_norm": 1.8949321508407593, "learning_rate": 6.9009450530072965e-06, "loss": 0.5885, "step": 9413 }, { "epoch": 0.39462184169435044, "grad_norm": 1.804257869720459, "learning_rate": 6.900317159016237e-06, "loss": 0.502, "step": 9414 }, { "epoch": 0.3946637603093594, "grad_norm": 2.1345112323760986, "learning_rate": 6.899689229995458e-06, "loss": 0.5492, "step": 9415 }, { "epoch": 0.39470567892436836, "grad_norm": 1.6954197883605957, "learning_rate": 6.8990612659565324e-06, "loss": 0.4951, "step": 9416 }, { "epoch": 0.3947475975393773, "grad_norm": 1.8810088634490967, "learning_rate": 6.898433266911039e-06, "loss": 0.5364, "step": 9417 }, { "epoch": 0.39478951615438623, "grad_norm": 2.581082582473755, "learning_rate": 6.897805232870552e-06, "loss": 0.4844, "step": 9418 }, { "epoch": 0.3948314347693952, "grad_norm": 1.836748719215393, "learning_rate": 6.89717716384665e-06, "loss": 0.5308, "step": 9419 }, { "epoch": 0.39487335338440416, "grad_norm": 1.862382411956787, "learning_rate": 6.89654905985091e-06, "loss": 0.5482, "step": 9420 }, { "epoch": 0.3949152719994131, "grad_norm": 2.294968843460083, "learning_rate": 6.89592092089491e-06, "loss": 0.5543, "step": 9421 }, { "epoch": 0.3949571906144221, "grad_norm": 1.9492789506912231, "learning_rate": 6.895292746990228e-06, "loss": 0.4904, "step": 9422 }, { "epoch": 0.39499910922943104, "grad_norm": 1.6166484355926514, "learning_rate": 6.8946645381484435e-06, "loss": 0.4981, "step": 9423 }, { "epoch": 0.39504102784444, "grad_norm": 1.766585350036621, "learning_rate": 6.89403629438114e-06, "loss": 0.5153, "step": 9424 }, { "epoch": 0.39508294645944897, "grad_norm": 1.827976942062378, "learning_rate": 6.893408015699895e-06, "loss": 0.4984, "step": 9425 }, { "epoch": 0.39512486507445793, "grad_norm": 1.7179772853851318, "learning_rate": 6.89277970211629e-06, "loss": 0.5253, "step": 9426 }, { "epoch": 0.3951667836894669, "grad_norm": 1.7596105337142944, "learning_rate": 6.8921513536419085e-06, "loss": 0.4818, "step": 9427 }, { "epoch": 0.39520870230447586, "grad_norm": 1.7299880981445312, "learning_rate": 6.891522970288333e-06, "loss": 0.5084, "step": 9428 }, { "epoch": 0.3952506209194848, "grad_norm": 2.1047656536102295, "learning_rate": 6.890894552067144e-06, "loss": 0.5657, "step": 9429 }, { "epoch": 0.3952925395344938, "grad_norm": 1.6480967998504639, "learning_rate": 6.89026609898993e-06, "loss": 0.5045, "step": 9430 }, { "epoch": 0.39533445814950274, "grad_norm": 1.6364244222640991, "learning_rate": 6.889637611068275e-06, "loss": 0.4997, "step": 9431 }, { "epoch": 0.3953763767645117, "grad_norm": 1.5534571409225464, "learning_rate": 6.889009088313761e-06, "loss": 0.5035, "step": 9432 }, { "epoch": 0.39541829537952067, "grad_norm": 1.9479806423187256, "learning_rate": 6.888380530737977e-06, "loss": 0.5626, "step": 9433 }, { "epoch": 0.39546021399452963, "grad_norm": 1.783624529838562, "learning_rate": 6.887751938352508e-06, "loss": 0.5155, "step": 9434 }, { "epoch": 0.3955021326095386, "grad_norm": 1.8281525373458862, "learning_rate": 6.887123311168942e-06, "loss": 0.5502, "step": 9435 }, { "epoch": 0.39554405122454755, "grad_norm": 1.859449863433838, "learning_rate": 6.886494649198866e-06, "loss": 0.5237, "step": 9436 }, { "epoch": 0.3955859698395565, "grad_norm": 1.6709387302398682, "learning_rate": 6.88586595245387e-06, "loss": 0.4676, "step": 9437 }, { "epoch": 0.3956278884545655, "grad_norm": 1.6775366067886353, "learning_rate": 6.885237220945542e-06, "loss": 0.5261, "step": 9438 }, { "epoch": 0.39566980706957444, "grad_norm": 1.835972785949707, "learning_rate": 6.884608454685472e-06, "loss": 0.497, "step": 9439 }, { "epoch": 0.3957117256845834, "grad_norm": 1.5216059684753418, "learning_rate": 6.883979653685251e-06, "loss": 0.432, "step": 9440 }, { "epoch": 0.39575364429959237, "grad_norm": 1.6166589260101318, "learning_rate": 6.883350817956468e-06, "loss": 0.5234, "step": 9441 }, { "epoch": 0.39579556291460133, "grad_norm": 1.681782841682434, "learning_rate": 6.882721947510718e-06, "loss": 0.5492, "step": 9442 }, { "epoch": 0.39583748152961024, "grad_norm": 2.038208484649658, "learning_rate": 6.882093042359591e-06, "loss": 0.5138, "step": 9443 }, { "epoch": 0.3958794001446192, "grad_norm": 1.6056103706359863, "learning_rate": 6.88146410251468e-06, "loss": 0.5067, "step": 9444 }, { "epoch": 0.39592131875962816, "grad_norm": 1.687008023262024, "learning_rate": 6.880835127987581e-06, "loss": 0.5765, "step": 9445 }, { "epoch": 0.3959632373746371, "grad_norm": 1.6476786136627197, "learning_rate": 6.880206118789885e-06, "loss": 0.505, "step": 9446 }, { "epoch": 0.3960051559896461, "grad_norm": 1.6691519021987915, "learning_rate": 6.87957707493319e-06, "loss": 0.4993, "step": 9447 }, { "epoch": 0.39604707460465505, "grad_norm": 1.6486647129058838, "learning_rate": 6.8789479964290895e-06, "loss": 0.5541, "step": 9448 }, { "epoch": 0.396088993219664, "grad_norm": 1.777828335762024, "learning_rate": 6.87831888328918e-06, "loss": 0.4718, "step": 9449 }, { "epoch": 0.39613091183467297, "grad_norm": 1.7451914548873901, "learning_rate": 6.877689735525057e-06, "loss": 0.4747, "step": 9450 }, { "epoch": 0.39617283044968193, "grad_norm": 2.1560990810394287, "learning_rate": 6.8770605531483215e-06, "loss": 0.5167, "step": 9451 }, { "epoch": 0.3962147490646909, "grad_norm": 1.6427793502807617, "learning_rate": 6.8764313361705705e-06, "loss": 0.4578, "step": 9452 }, { "epoch": 0.39625666767969986, "grad_norm": 1.7702021598815918, "learning_rate": 6.8758020846034e-06, "loss": 0.5351, "step": 9453 }, { "epoch": 0.3962985862947088, "grad_norm": 1.863965630531311, "learning_rate": 6.875172798458413e-06, "loss": 0.5404, "step": 9454 }, { "epoch": 0.3963405049097178, "grad_norm": 1.9039589166641235, "learning_rate": 6.874543477747207e-06, "loss": 0.5744, "step": 9455 }, { "epoch": 0.39638242352472675, "grad_norm": 1.9343194961547852, "learning_rate": 6.873914122481382e-06, "loss": 0.4681, "step": 9456 }, { "epoch": 0.3964243421397357, "grad_norm": 1.740952491760254, "learning_rate": 6.8732847326725425e-06, "loss": 0.5157, "step": 9457 }, { "epoch": 0.39646626075474467, "grad_norm": 1.9625269174575806, "learning_rate": 6.872655308332289e-06, "loss": 0.5693, "step": 9458 }, { "epoch": 0.39650817936975363, "grad_norm": 2.1145567893981934, "learning_rate": 6.872025849472222e-06, "loss": 0.5356, "step": 9459 }, { "epoch": 0.3965500979847626, "grad_norm": 1.5955859422683716, "learning_rate": 6.871396356103948e-06, "loss": 0.5101, "step": 9460 }, { "epoch": 0.39659201659977156, "grad_norm": 1.6369608640670776, "learning_rate": 6.870766828239068e-06, "loss": 0.5172, "step": 9461 }, { "epoch": 0.3966339352147805, "grad_norm": 1.760405421257019, "learning_rate": 6.870137265889189e-06, "loss": 0.4567, "step": 9462 }, { "epoch": 0.3966758538297895, "grad_norm": 1.6549086570739746, "learning_rate": 6.869507669065914e-06, "loss": 0.4606, "step": 9463 }, { "epoch": 0.39671777244479844, "grad_norm": 2.295771360397339, "learning_rate": 6.86887803778085e-06, "loss": 0.4931, "step": 9464 }, { "epoch": 0.3967596910598074, "grad_norm": 1.9165284633636475, "learning_rate": 6.868248372045604e-06, "loss": 0.5658, "step": 9465 }, { "epoch": 0.39680160967481637, "grad_norm": 2.0019328594207764, "learning_rate": 6.8676186718717805e-06, "loss": 0.524, "step": 9466 }, { "epoch": 0.39684352828982533, "grad_norm": 1.895755410194397, "learning_rate": 6.866988937270991e-06, "loss": 0.5436, "step": 9467 }, { "epoch": 0.39688544690483424, "grad_norm": 1.6717432737350464, "learning_rate": 6.866359168254839e-06, "loss": 0.4898, "step": 9468 }, { "epoch": 0.3969273655198432, "grad_norm": 1.8899775743484497, "learning_rate": 6.865729364834937e-06, "loss": 0.4852, "step": 9469 }, { "epoch": 0.39696928413485216, "grad_norm": 1.9362132549285889, "learning_rate": 6.865099527022893e-06, "loss": 0.5199, "step": 9470 }, { "epoch": 0.3970112027498611, "grad_norm": 1.4915025234222412, "learning_rate": 6.864469654830318e-06, "loss": 0.4878, "step": 9471 }, { "epoch": 0.3970531213648701, "grad_norm": 1.7769814729690552, "learning_rate": 6.863839748268821e-06, "loss": 0.526, "step": 9472 }, { "epoch": 0.39709503997987905, "grad_norm": 2.996648073196411, "learning_rate": 6.863209807350017e-06, "loss": 0.5397, "step": 9473 }, { "epoch": 0.397136958594888, "grad_norm": 1.6526151895523071, "learning_rate": 6.8625798320855154e-06, "loss": 0.5363, "step": 9474 }, { "epoch": 0.397178877209897, "grad_norm": 1.726401925086975, "learning_rate": 6.86194982248693e-06, "loss": 0.5157, "step": 9475 }, { "epoch": 0.39722079582490594, "grad_norm": 1.6790448427200317, "learning_rate": 6.8613197785658725e-06, "loss": 0.506, "step": 9476 }, { "epoch": 0.3972627144399149, "grad_norm": 1.5552154779434204, "learning_rate": 6.860689700333959e-06, "loss": 0.4984, "step": 9477 }, { "epoch": 0.39730463305492386, "grad_norm": 1.6352338790893555, "learning_rate": 6.860059587802802e-06, "loss": 0.543, "step": 9478 }, { "epoch": 0.3973465516699328, "grad_norm": 1.8803491592407227, "learning_rate": 6.859429440984018e-06, "loss": 0.4613, "step": 9479 }, { "epoch": 0.3973884702849418, "grad_norm": 1.7735120058059692, "learning_rate": 6.858799259889224e-06, "loss": 0.5301, "step": 9480 }, { "epoch": 0.39743038889995075, "grad_norm": 1.9082945585250854, "learning_rate": 6.858169044530035e-06, "loss": 0.5093, "step": 9481 }, { "epoch": 0.3974723075149597, "grad_norm": 1.9352246522903442, "learning_rate": 6.857538794918068e-06, "loss": 0.4802, "step": 9482 }, { "epoch": 0.3975142261299687, "grad_norm": 1.6113998889923096, "learning_rate": 6.8569085110649424e-06, "loss": 0.5259, "step": 9483 }, { "epoch": 0.39755614474497764, "grad_norm": 1.6501227617263794, "learning_rate": 6.856278192982275e-06, "loss": 0.4719, "step": 9484 }, { "epoch": 0.3975980633599866, "grad_norm": 1.7949635982513428, "learning_rate": 6.855647840681685e-06, "loss": 0.4637, "step": 9485 }, { "epoch": 0.39763998197499556, "grad_norm": 1.8609248399734497, "learning_rate": 6.8550174541747925e-06, "loss": 0.5018, "step": 9486 }, { "epoch": 0.3976819005900045, "grad_norm": 1.5857223272323608, "learning_rate": 6.854387033473218e-06, "loss": 0.5095, "step": 9487 }, { "epoch": 0.3977238192050135, "grad_norm": 1.733590841293335, "learning_rate": 6.853756578588582e-06, "loss": 0.5618, "step": 9488 }, { "epoch": 0.39776573782002245, "grad_norm": 2.0038537979125977, "learning_rate": 6.853126089532506e-06, "loss": 0.5691, "step": 9489 }, { "epoch": 0.3978076564350314, "grad_norm": 1.7803682088851929, "learning_rate": 6.852495566316613e-06, "loss": 0.5668, "step": 9490 }, { "epoch": 0.39784957505004037, "grad_norm": 1.6901386976242065, "learning_rate": 6.8518650089525255e-06, "loss": 0.5399, "step": 9491 }, { "epoch": 0.39789149366504933, "grad_norm": 1.907003402709961, "learning_rate": 6.851234417451865e-06, "loss": 0.5189, "step": 9492 }, { "epoch": 0.39793341228005824, "grad_norm": 1.531180739402771, "learning_rate": 6.850603791826258e-06, "loss": 0.517, "step": 9493 }, { "epoch": 0.3979753308950672, "grad_norm": 1.77024245262146, "learning_rate": 6.84997313208733e-06, "loss": 0.538, "step": 9494 }, { "epoch": 0.39801724951007617, "grad_norm": 1.715760350227356, "learning_rate": 6.849342438246703e-06, "loss": 0.5168, "step": 9495 }, { "epoch": 0.3980591681250851, "grad_norm": 1.7241319417953491, "learning_rate": 6.848711710316005e-06, "loss": 0.5512, "step": 9496 }, { "epoch": 0.3981010867400941, "grad_norm": 1.8370710611343384, "learning_rate": 6.8480809483068636e-06, "loss": 0.5446, "step": 9497 }, { "epoch": 0.39814300535510305, "grad_norm": 1.7062264680862427, "learning_rate": 6.847450152230903e-06, "loss": 0.5087, "step": 9498 }, { "epoch": 0.398184923970112, "grad_norm": 1.4040435552597046, "learning_rate": 6.846819322099756e-06, "loss": 0.5132, "step": 9499 }, { "epoch": 0.398226842585121, "grad_norm": 1.5401206016540527, "learning_rate": 6.846188457925044e-06, "loss": 0.5313, "step": 9500 }, { "epoch": 0.39826876120012994, "grad_norm": 2.428520679473877, "learning_rate": 6.845557559718402e-06, "loss": 0.5803, "step": 9501 }, { "epoch": 0.3983106798151389, "grad_norm": 1.7267781496047974, "learning_rate": 6.844926627491458e-06, "loss": 0.549, "step": 9502 }, { "epoch": 0.39835259843014786, "grad_norm": 1.652816653251648, "learning_rate": 6.844295661255842e-06, "loss": 0.5097, "step": 9503 }, { "epoch": 0.3983945170451568, "grad_norm": 1.7498821020126343, "learning_rate": 6.843664661023185e-06, "loss": 0.5045, "step": 9504 }, { "epoch": 0.3984364356601658, "grad_norm": 2.0821175575256348, "learning_rate": 6.8430336268051185e-06, "loss": 0.5211, "step": 9505 }, { "epoch": 0.39847835427517475, "grad_norm": 2.208723545074463, "learning_rate": 6.842402558613275e-06, "loss": 0.5583, "step": 9506 }, { "epoch": 0.3985202728901837, "grad_norm": 1.943469762802124, "learning_rate": 6.841771456459288e-06, "loss": 0.5252, "step": 9507 }, { "epoch": 0.3985621915051927, "grad_norm": 1.7507020235061646, "learning_rate": 6.841140320354788e-06, "loss": 0.5308, "step": 9508 }, { "epoch": 0.39860411012020164, "grad_norm": 1.5579653978347778, "learning_rate": 6.840509150311415e-06, "loss": 0.5072, "step": 9509 }, { "epoch": 0.3986460287352106, "grad_norm": 1.658962368965149, "learning_rate": 6.839877946340798e-06, "loss": 0.4433, "step": 9510 }, { "epoch": 0.39868794735021956, "grad_norm": 1.8266183137893677, "learning_rate": 6.8392467084545764e-06, "loss": 0.512, "step": 9511 }, { "epoch": 0.3987298659652285, "grad_norm": 1.691156029701233, "learning_rate": 6.838615436664382e-06, "loss": 0.5383, "step": 9512 }, { "epoch": 0.3987717845802375, "grad_norm": 1.7955620288848877, "learning_rate": 6.837984130981856e-06, "loss": 0.6149, "step": 9513 }, { "epoch": 0.39881370319524645, "grad_norm": 1.779999017715454, "learning_rate": 6.837352791418634e-06, "loss": 0.5315, "step": 9514 }, { "epoch": 0.3988556218102554, "grad_norm": 1.6347719430923462, "learning_rate": 6.836721417986351e-06, "loss": 0.4917, "step": 9515 }, { "epoch": 0.3988975404252644, "grad_norm": 1.6111948490142822, "learning_rate": 6.836090010696648e-06, "loss": 0.468, "step": 9516 }, { "epoch": 0.39893945904027334, "grad_norm": 1.698393702507019, "learning_rate": 6.835458569561165e-06, "loss": 0.5421, "step": 9517 }, { "epoch": 0.39898137765528224, "grad_norm": 1.6612271070480347, "learning_rate": 6.83482709459154e-06, "loss": 0.5323, "step": 9518 }, { "epoch": 0.3990232962702912, "grad_norm": 2.226226568222046, "learning_rate": 6.834195585799415e-06, "loss": 0.5429, "step": 9519 }, { "epoch": 0.39906521488530017, "grad_norm": 1.4850974082946777, "learning_rate": 6.83356404319643e-06, "loss": 0.5031, "step": 9520 }, { "epoch": 0.39910713350030913, "grad_norm": 1.9937840700149536, "learning_rate": 6.8329324667942265e-06, "loss": 0.4745, "step": 9521 }, { "epoch": 0.3991490521153181, "grad_norm": 1.6753461360931396, "learning_rate": 6.832300856604448e-06, "loss": 0.4686, "step": 9522 }, { "epoch": 0.39919097073032705, "grad_norm": 1.666675090789795, "learning_rate": 6.831669212638734e-06, "loss": 0.4757, "step": 9523 }, { "epoch": 0.399232889345336, "grad_norm": 3.5330328941345215, "learning_rate": 6.831037534908731e-06, "loss": 0.5039, "step": 9524 }, { "epoch": 0.399274807960345, "grad_norm": 1.660827398300171, "learning_rate": 6.8304058234260825e-06, "loss": 0.5082, "step": 9525 }, { "epoch": 0.39931672657535394, "grad_norm": 1.5451394319534302, "learning_rate": 6.829774078202434e-06, "loss": 0.5321, "step": 9526 }, { "epoch": 0.3993586451903629, "grad_norm": 1.9948419332504272, "learning_rate": 6.82914229924943e-06, "loss": 0.4804, "step": 9527 }, { "epoch": 0.39940056380537187, "grad_norm": 1.7538976669311523, "learning_rate": 6.828510486578716e-06, "loss": 0.5062, "step": 9528 }, { "epoch": 0.39944248242038083, "grad_norm": 1.6863892078399658, "learning_rate": 6.8278786402019385e-06, "loss": 0.5261, "step": 9529 }, { "epoch": 0.3994844010353898, "grad_norm": 1.792884349822998, "learning_rate": 6.8272467601307445e-06, "loss": 0.5121, "step": 9530 }, { "epoch": 0.39952631965039875, "grad_norm": 2.632061243057251, "learning_rate": 6.826614846376783e-06, "loss": 0.5276, "step": 9531 }, { "epoch": 0.3995682382654077, "grad_norm": 1.668561577796936, "learning_rate": 6.825982898951704e-06, "loss": 0.5421, "step": 9532 }, { "epoch": 0.3996101568804167, "grad_norm": 1.6757358312606812, "learning_rate": 6.825350917867153e-06, "loss": 0.5207, "step": 9533 }, { "epoch": 0.39965207549542564, "grad_norm": 1.7611883878707886, "learning_rate": 6.824718903134781e-06, "loss": 0.5076, "step": 9534 }, { "epoch": 0.3996939941104346, "grad_norm": 1.8001415729522705, "learning_rate": 6.824086854766238e-06, "loss": 0.5654, "step": 9535 }, { "epoch": 0.39973591272544357, "grad_norm": 1.7613688707351685, "learning_rate": 6.823454772773178e-06, "loss": 0.5181, "step": 9536 }, { "epoch": 0.3997778313404525, "grad_norm": 1.7845454216003418, "learning_rate": 6.82282265716725e-06, "loss": 0.5578, "step": 9537 }, { "epoch": 0.3998197499554615, "grad_norm": 1.3617846965789795, "learning_rate": 6.822190507960105e-06, "loss": 0.5005, "step": 9538 }, { "epoch": 0.39986166857047045, "grad_norm": 1.6571553945541382, "learning_rate": 6.8215583251633964e-06, "loss": 0.5201, "step": 9539 }, { "epoch": 0.3999035871854794, "grad_norm": 2.442183494567871, "learning_rate": 6.820926108788778e-06, "loss": 0.559, "step": 9540 }, { "epoch": 0.3999455058004884, "grad_norm": 1.8022834062576294, "learning_rate": 6.820293858847906e-06, "loss": 0.4891, "step": 9541 }, { "epoch": 0.39998742441549734, "grad_norm": 1.839674472808838, "learning_rate": 6.819661575352433e-06, "loss": 0.5515, "step": 9542 }, { "epoch": 0.40002934303050625, "grad_norm": 1.7624092102050781, "learning_rate": 6.819029258314012e-06, "loss": 0.5226, "step": 9543 }, { "epoch": 0.4000712616455152, "grad_norm": 1.6423286199569702, "learning_rate": 6.8183969077443045e-06, "loss": 0.564, "step": 9544 }, { "epoch": 0.40011318026052417, "grad_norm": 1.7987620830535889, "learning_rate": 6.817764523654963e-06, "loss": 0.4755, "step": 9545 }, { "epoch": 0.40015509887553313, "grad_norm": 1.4905980825424194, "learning_rate": 6.8171321060576445e-06, "loss": 0.4495, "step": 9546 }, { "epoch": 0.4001970174905421, "grad_norm": 1.6631144285202026, "learning_rate": 6.816499654964009e-06, "loss": 0.5026, "step": 9547 }, { "epoch": 0.40023893610555106, "grad_norm": 1.7283704280853271, "learning_rate": 6.815867170385713e-06, "loss": 0.4721, "step": 9548 }, { "epoch": 0.40028085472056, "grad_norm": 1.6236439943313599, "learning_rate": 6.815234652334415e-06, "loss": 0.4974, "step": 9549 }, { "epoch": 0.400322773335569, "grad_norm": 1.9104375839233398, "learning_rate": 6.814602100821778e-06, "loss": 0.6064, "step": 9550 }, { "epoch": 0.40036469195057794, "grad_norm": 1.6029692888259888, "learning_rate": 6.813969515859458e-06, "loss": 0.5407, "step": 9551 }, { "epoch": 0.4004066105655869, "grad_norm": 1.6315442323684692, "learning_rate": 6.813336897459117e-06, "loss": 0.5183, "step": 9552 }, { "epoch": 0.40044852918059587, "grad_norm": 1.851941466331482, "learning_rate": 6.8127042456324186e-06, "loss": 0.5546, "step": 9553 }, { "epoch": 0.40049044779560483, "grad_norm": 1.7551594972610474, "learning_rate": 6.812071560391024e-06, "loss": 0.4788, "step": 9554 }, { "epoch": 0.4005323664106138, "grad_norm": 1.9940516948699951, "learning_rate": 6.811438841746595e-06, "loss": 0.5373, "step": 9555 }, { "epoch": 0.40057428502562276, "grad_norm": 1.7324087619781494, "learning_rate": 6.810806089710794e-06, "loss": 0.5329, "step": 9556 }, { "epoch": 0.4006162036406317, "grad_norm": 1.8248316049575806, "learning_rate": 6.810173304295287e-06, "loss": 0.5543, "step": 9557 }, { "epoch": 0.4006581222556407, "grad_norm": 1.930332064628601, "learning_rate": 6.809540485511737e-06, "loss": 0.5315, "step": 9558 }, { "epoch": 0.40070004087064964, "grad_norm": 2.7603704929351807, "learning_rate": 6.80890763337181e-06, "loss": 0.5465, "step": 9559 }, { "epoch": 0.4007419594856586, "grad_norm": 1.804186224937439, "learning_rate": 6.8082747478871735e-06, "loss": 0.5179, "step": 9560 }, { "epoch": 0.40078387810066757, "grad_norm": 1.722536325454712, "learning_rate": 6.807641829069489e-06, "loss": 0.5052, "step": 9561 }, { "epoch": 0.40082579671567653, "grad_norm": 1.755694031715393, "learning_rate": 6.807008876930427e-06, "loss": 0.5006, "step": 9562 }, { "epoch": 0.4008677153306855, "grad_norm": 1.474846601486206, "learning_rate": 6.806375891481656e-06, "loss": 0.4904, "step": 9563 }, { "epoch": 0.40090963394569445, "grad_norm": 1.8484197854995728, "learning_rate": 6.805742872734842e-06, "loss": 0.5657, "step": 9564 }, { "epoch": 0.4009515525607034, "grad_norm": 1.70122230052948, "learning_rate": 6.805109820701654e-06, "loss": 0.5689, "step": 9565 }, { "epoch": 0.4009934711757124, "grad_norm": 1.6275317668914795, "learning_rate": 6.804476735393762e-06, "loss": 0.4828, "step": 9566 }, { "epoch": 0.40103538979072134, "grad_norm": 1.7349787950515747, "learning_rate": 6.803843616822835e-06, "loss": 0.534, "step": 9567 }, { "epoch": 0.40107730840573025, "grad_norm": 1.830509066581726, "learning_rate": 6.803210465000546e-06, "loss": 0.4928, "step": 9568 }, { "epoch": 0.4011192270207392, "grad_norm": 1.4468492269515991, "learning_rate": 6.802577279938566e-06, "loss": 0.5265, "step": 9569 }, { "epoch": 0.4011611456357482, "grad_norm": 1.5546621084213257, "learning_rate": 6.8019440616485635e-06, "loss": 0.5184, "step": 9570 }, { "epoch": 0.40120306425075714, "grad_norm": 1.7479572296142578, "learning_rate": 6.801310810142214e-06, "loss": 0.4989, "step": 9571 }, { "epoch": 0.4012449828657661, "grad_norm": 1.5801609754562378, "learning_rate": 6.800677525431191e-06, "loss": 0.4605, "step": 9572 }, { "epoch": 0.40128690148077506, "grad_norm": 1.445405125617981, "learning_rate": 6.8000442075271654e-06, "loss": 0.5268, "step": 9573 }, { "epoch": 0.401328820095784, "grad_norm": 1.7373956441879272, "learning_rate": 6.799410856441815e-06, "loss": 0.5033, "step": 9574 }, { "epoch": 0.401370738710793, "grad_norm": 2.005525588989258, "learning_rate": 6.798777472186813e-06, "loss": 0.5738, "step": 9575 }, { "epoch": 0.40141265732580195, "grad_norm": 1.5198731422424316, "learning_rate": 6.798144054773834e-06, "loss": 0.4836, "step": 9576 }, { "epoch": 0.4014545759408109, "grad_norm": 1.8632395267486572, "learning_rate": 6.797510604214556e-06, "loss": 0.5899, "step": 9577 }, { "epoch": 0.40149649455581987, "grad_norm": 1.886540412902832, "learning_rate": 6.796877120520654e-06, "loss": 0.5359, "step": 9578 }, { "epoch": 0.40153841317082883, "grad_norm": 1.7489062547683716, "learning_rate": 6.796243603703808e-06, "loss": 0.5663, "step": 9579 }, { "epoch": 0.4015803317858378, "grad_norm": 1.7287250757217407, "learning_rate": 6.795610053775693e-06, "loss": 0.5257, "step": 9580 }, { "epoch": 0.40162225040084676, "grad_norm": 1.9989466667175293, "learning_rate": 6.794976470747989e-06, "loss": 0.5411, "step": 9581 }, { "epoch": 0.4016641690158557, "grad_norm": 1.671566367149353, "learning_rate": 6.794342854632376e-06, "loss": 0.5288, "step": 9582 }, { "epoch": 0.4017060876308647, "grad_norm": 1.6936851739883423, "learning_rate": 6.793709205440533e-06, "loss": 0.5151, "step": 9583 }, { "epoch": 0.40174800624587365, "grad_norm": 1.7788655757904053, "learning_rate": 6.79307552318414e-06, "loss": 0.5209, "step": 9584 }, { "epoch": 0.4017899248608826, "grad_norm": 1.6781656742095947, "learning_rate": 6.79244180787488e-06, "loss": 0.5068, "step": 9585 }, { "epoch": 0.40183184347589157, "grad_norm": 1.78410804271698, "learning_rate": 6.791808059524432e-06, "loss": 0.5142, "step": 9586 }, { "epoch": 0.40187376209090053, "grad_norm": 1.7478197813034058, "learning_rate": 6.791174278144479e-06, "loss": 0.4935, "step": 9587 }, { "epoch": 0.4019156807059095, "grad_norm": 1.6603974103927612, "learning_rate": 6.7905404637467065e-06, "loss": 0.4462, "step": 9588 }, { "epoch": 0.40195759932091846, "grad_norm": 1.3897161483764648, "learning_rate": 6.789906616342794e-06, "loss": 0.4695, "step": 9589 }, { "epoch": 0.4019995179359274, "grad_norm": 1.6603951454162598, "learning_rate": 6.789272735944428e-06, "loss": 0.4958, "step": 9590 }, { "epoch": 0.4020414365509364, "grad_norm": 1.8642345666885376, "learning_rate": 6.788638822563292e-06, "loss": 0.543, "step": 9591 }, { "epoch": 0.40208335516594534, "grad_norm": 1.9352036714553833, "learning_rate": 6.788004876211074e-06, "loss": 0.5371, "step": 9592 }, { "epoch": 0.40212527378095425, "grad_norm": 2.0383644104003906, "learning_rate": 6.7873708968994566e-06, "loss": 0.5141, "step": 9593 }, { "epoch": 0.4021671923959632, "grad_norm": 1.5239272117614746, "learning_rate": 6.786736884640128e-06, "loss": 0.511, "step": 9594 }, { "epoch": 0.4022091110109722, "grad_norm": 2.0175974369049072, "learning_rate": 6.7861028394447736e-06, "loss": 0.493, "step": 9595 }, { "epoch": 0.40225102962598114, "grad_norm": 1.944506287574768, "learning_rate": 6.785468761325084e-06, "loss": 0.5528, "step": 9596 }, { "epoch": 0.4022929482409901, "grad_norm": 1.6286990642547607, "learning_rate": 6.784834650292746e-06, "loss": 0.4856, "step": 9597 }, { "epoch": 0.40233486685599906, "grad_norm": 1.5941556692123413, "learning_rate": 6.784200506359449e-06, "loss": 0.5012, "step": 9598 }, { "epoch": 0.402376785471008, "grad_norm": 3.2737722396850586, "learning_rate": 6.7835663295368815e-06, "loss": 0.5727, "step": 9599 }, { "epoch": 0.402418704086017, "grad_norm": 1.619913101196289, "learning_rate": 6.782932119836735e-06, "loss": 0.5059, "step": 9600 }, { "epoch": 0.40246062270102595, "grad_norm": 1.815250277519226, "learning_rate": 6.782297877270699e-06, "loss": 0.4892, "step": 9601 }, { "epoch": 0.4025025413160349, "grad_norm": 1.690993070602417, "learning_rate": 6.781663601850467e-06, "loss": 0.4996, "step": 9602 }, { "epoch": 0.4025444599310439, "grad_norm": 1.6803637742996216, "learning_rate": 6.7810292935877284e-06, "loss": 0.5131, "step": 9603 }, { "epoch": 0.40258637854605284, "grad_norm": 1.7023061513900757, "learning_rate": 6.780394952494176e-06, "loss": 0.496, "step": 9604 }, { "epoch": 0.4026282971610618, "grad_norm": 1.5888465642929077, "learning_rate": 6.779760578581506e-06, "loss": 0.4438, "step": 9605 }, { "epoch": 0.40267021577607076, "grad_norm": 1.8037872314453125, "learning_rate": 6.779126171861409e-06, "loss": 0.5349, "step": 9606 }, { "epoch": 0.4027121343910797, "grad_norm": 1.6005535125732422, "learning_rate": 6.7784917323455815e-06, "loss": 0.4889, "step": 9607 }, { "epoch": 0.4027540530060887, "grad_norm": 1.7650583982467651, "learning_rate": 6.777857260045717e-06, "loss": 0.4963, "step": 9608 }, { "epoch": 0.40279597162109765, "grad_norm": 1.6937285661697388, "learning_rate": 6.777222754973511e-06, "loss": 0.4808, "step": 9609 }, { "epoch": 0.4028378902361066, "grad_norm": 1.6386123895645142, "learning_rate": 6.776588217140661e-06, "loss": 0.558, "step": 9610 }, { "epoch": 0.4028798088511156, "grad_norm": 1.7125792503356934, "learning_rate": 6.7759536465588645e-06, "loss": 0.5279, "step": 9611 }, { "epoch": 0.40292172746612454, "grad_norm": 3.347618818283081, "learning_rate": 6.775319043239816e-06, "loss": 0.5486, "step": 9612 }, { "epoch": 0.4029636460811335, "grad_norm": 1.6760786771774292, "learning_rate": 6.774684407195217e-06, "loss": 0.5578, "step": 9613 }, { "epoch": 0.40300556469614246, "grad_norm": 2.0565834045410156, "learning_rate": 6.774049738436762e-06, "loss": 0.5607, "step": 9614 }, { "epoch": 0.4030474833111514, "grad_norm": 1.488777756690979, "learning_rate": 6.773415036976154e-06, "loss": 0.477, "step": 9615 }, { "epoch": 0.4030894019261604, "grad_norm": 1.9207054376602173, "learning_rate": 6.772780302825093e-06, "loss": 0.6077, "step": 9616 }, { "epoch": 0.40313132054116935, "grad_norm": 1.8085777759552002, "learning_rate": 6.772145535995276e-06, "loss": 0.5799, "step": 9617 }, { "epoch": 0.40317323915617825, "grad_norm": 1.8239123821258545, "learning_rate": 6.7715107364984055e-06, "loss": 0.4772, "step": 9618 }, { "epoch": 0.4032151577711872, "grad_norm": 1.5449377298355103, "learning_rate": 6.770875904346185e-06, "loss": 0.5114, "step": 9619 }, { "epoch": 0.4032570763861962, "grad_norm": 1.6925538778305054, "learning_rate": 6.770241039550315e-06, "loss": 0.5418, "step": 9620 }, { "epoch": 0.40329899500120514, "grad_norm": 1.7056317329406738, "learning_rate": 6.769606142122498e-06, "loss": 0.4981, "step": 9621 }, { "epoch": 0.4033409136162141, "grad_norm": 1.3940467834472656, "learning_rate": 6.768971212074439e-06, "loss": 0.5066, "step": 9622 }, { "epoch": 0.40338283223122307, "grad_norm": 1.7107264995574951, "learning_rate": 6.768336249417839e-06, "loss": 0.5048, "step": 9623 }, { "epoch": 0.403424750846232, "grad_norm": 3.1621105670928955, "learning_rate": 6.767701254164408e-06, "loss": 0.5175, "step": 9624 }, { "epoch": 0.403466669461241, "grad_norm": 1.6806941032409668, "learning_rate": 6.767066226325847e-06, "loss": 0.5237, "step": 9625 }, { "epoch": 0.40350858807624995, "grad_norm": 1.832122564315796, "learning_rate": 6.7664311659138626e-06, "loss": 0.5156, "step": 9626 }, { "epoch": 0.4035505066912589, "grad_norm": 1.671156883239746, "learning_rate": 6.765796072940163e-06, "loss": 0.5249, "step": 9627 }, { "epoch": 0.4035924253062679, "grad_norm": 1.8621639013290405, "learning_rate": 6.765160947416452e-06, "loss": 0.5105, "step": 9628 }, { "epoch": 0.40363434392127684, "grad_norm": 1.753538727760315, "learning_rate": 6.7645257893544406e-06, "loss": 0.5664, "step": 9629 }, { "epoch": 0.4036762625362858, "grad_norm": 1.6182444095611572, "learning_rate": 6.763890598765835e-06, "loss": 0.5412, "step": 9630 }, { "epoch": 0.40371818115129476, "grad_norm": 1.6555938720703125, "learning_rate": 6.763255375662346e-06, "loss": 0.5302, "step": 9631 }, { "epoch": 0.4037600997663037, "grad_norm": 2.3374619483947754, "learning_rate": 6.762620120055679e-06, "loss": 0.5356, "step": 9632 }, { "epoch": 0.4038020183813127, "grad_norm": 1.6356661319732666, "learning_rate": 6.761984831957549e-06, "loss": 0.5329, "step": 9633 }, { "epoch": 0.40384393699632165, "grad_norm": 1.9489190578460693, "learning_rate": 6.761349511379664e-06, "loss": 0.5332, "step": 9634 }, { "epoch": 0.4038858556113306, "grad_norm": 2.1083462238311768, "learning_rate": 6.760714158333736e-06, "loss": 0.5906, "step": 9635 }, { "epoch": 0.4039277742263396, "grad_norm": 1.5246354341506958, "learning_rate": 6.760078772831477e-06, "loss": 0.4612, "step": 9636 }, { "epoch": 0.40396969284134854, "grad_norm": 1.7203463315963745, "learning_rate": 6.759443354884598e-06, "loss": 0.5349, "step": 9637 }, { "epoch": 0.4040116114563575, "grad_norm": 1.7564972639083862, "learning_rate": 6.7588079045048124e-06, "loss": 0.4963, "step": 9638 }, { "epoch": 0.40405353007136646, "grad_norm": 2.0011043548583984, "learning_rate": 6.758172421703837e-06, "loss": 0.5162, "step": 9639 }, { "epoch": 0.4040954486863754, "grad_norm": 1.542015790939331, "learning_rate": 6.757536906493383e-06, "loss": 0.5042, "step": 9640 }, { "epoch": 0.4041373673013844, "grad_norm": 1.722630262374878, "learning_rate": 6.756901358885164e-06, "loss": 0.4854, "step": 9641 }, { "epoch": 0.40417928591639335, "grad_norm": 1.6328275203704834, "learning_rate": 6.756265778890899e-06, "loss": 0.5356, "step": 9642 }, { "epoch": 0.40422120453140226, "grad_norm": 1.7634880542755127, "learning_rate": 6.755630166522302e-06, "loss": 0.5048, "step": 9643 }, { "epoch": 0.4042631231464112, "grad_norm": 2.0683138370513916, "learning_rate": 6.7549945217910894e-06, "loss": 0.5081, "step": 9644 }, { "epoch": 0.4043050417614202, "grad_norm": 1.5594918727874756, "learning_rate": 6.7543588447089796e-06, "loss": 0.5346, "step": 9645 }, { "epoch": 0.40434696037642914, "grad_norm": 4.303695201873779, "learning_rate": 6.75372313528769e-06, "loss": 0.5395, "step": 9646 }, { "epoch": 0.4043888789914381, "grad_norm": 1.6413373947143555, "learning_rate": 6.753087393538938e-06, "loss": 0.4953, "step": 9647 }, { "epoch": 0.40443079760644707, "grad_norm": 3.133357286453247, "learning_rate": 6.7524516194744436e-06, "loss": 0.5469, "step": 9648 }, { "epoch": 0.40447271622145603, "grad_norm": 1.4327547550201416, "learning_rate": 6.751815813105927e-06, "loss": 0.5019, "step": 9649 }, { "epoch": 0.404514634836465, "grad_norm": 1.8232640027999878, "learning_rate": 6.751179974445108e-06, "loss": 0.4857, "step": 9650 }, { "epoch": 0.40455655345147395, "grad_norm": 1.624525547027588, "learning_rate": 6.7505441035037055e-06, "loss": 0.5272, "step": 9651 }, { "epoch": 0.4045984720664829, "grad_norm": 1.423700213432312, "learning_rate": 6.749908200293443e-06, "loss": 0.5123, "step": 9652 }, { "epoch": 0.4046403906814919, "grad_norm": 1.7624715566635132, "learning_rate": 6.7492722648260436e-06, "loss": 0.5243, "step": 9653 }, { "epoch": 0.40468230929650084, "grad_norm": 2.3738980293273926, "learning_rate": 6.748636297113227e-06, "loss": 0.5702, "step": 9654 }, { "epoch": 0.4047242279115098, "grad_norm": 1.5070126056671143, "learning_rate": 6.748000297166718e-06, "loss": 0.4972, "step": 9655 }, { "epoch": 0.40476614652651877, "grad_norm": 1.63895583152771, "learning_rate": 6.747364264998239e-06, "loss": 0.5336, "step": 9656 }, { "epoch": 0.40480806514152773, "grad_norm": 1.7146579027175903, "learning_rate": 6.7467282006195164e-06, "loss": 0.5121, "step": 9657 }, { "epoch": 0.4048499837565367, "grad_norm": 1.6509218215942383, "learning_rate": 6.746092104042274e-06, "loss": 0.4866, "step": 9658 }, { "epoch": 0.40489190237154565, "grad_norm": 1.6443990468978882, "learning_rate": 6.745455975278239e-06, "loss": 0.4855, "step": 9659 }, { "epoch": 0.4049338209865546, "grad_norm": 2.2739503383636475, "learning_rate": 6.7448198143391345e-06, "loss": 0.5145, "step": 9660 }, { "epoch": 0.4049757396015636, "grad_norm": 1.915769338607788, "learning_rate": 6.7441836212366885e-06, "loss": 0.5406, "step": 9661 }, { "epoch": 0.40501765821657254, "grad_norm": 2.0457592010498047, "learning_rate": 6.743547395982629e-06, "loss": 0.4738, "step": 9662 }, { "epoch": 0.4050595768315815, "grad_norm": 1.816633939743042, "learning_rate": 6.7429111385886835e-06, "loss": 0.5379, "step": 9663 }, { "epoch": 0.40510149544659046, "grad_norm": 1.7345843315124512, "learning_rate": 6.742274849066582e-06, "loss": 0.4802, "step": 9664 }, { "epoch": 0.4051434140615994, "grad_norm": 2.2792601585388184, "learning_rate": 6.741638527428052e-06, "loss": 0.5043, "step": 9665 }, { "epoch": 0.4051853326766084, "grad_norm": 2.1715192794799805, "learning_rate": 6.7410021736848216e-06, "loss": 0.5496, "step": 9666 }, { "epoch": 0.40522725129161735, "grad_norm": 1.6214426755905151, "learning_rate": 6.740365787848625e-06, "loss": 0.4882, "step": 9667 }, { "epoch": 0.40526916990662626, "grad_norm": 1.6952528953552246, "learning_rate": 6.739729369931191e-06, "loss": 0.5423, "step": 9668 }, { "epoch": 0.4053110885216352, "grad_norm": 1.6608694791793823, "learning_rate": 6.73909291994425e-06, "loss": 0.4837, "step": 9669 }, { "epoch": 0.4053530071366442, "grad_norm": 1.7946734428405762, "learning_rate": 6.738456437899536e-06, "loss": 0.4924, "step": 9670 }, { "epoch": 0.40539492575165315, "grad_norm": 2.074512481689453, "learning_rate": 6.7378199238087806e-06, "loss": 0.4918, "step": 9671 }, { "epoch": 0.4054368443666621, "grad_norm": 1.675769567489624, "learning_rate": 6.737183377683717e-06, "loss": 0.5889, "step": 9672 }, { "epoch": 0.40547876298167107, "grad_norm": 1.720982551574707, "learning_rate": 6.73654679953608e-06, "loss": 0.5588, "step": 9673 }, { "epoch": 0.40552068159668003, "grad_norm": 2.1546053886413574, "learning_rate": 6.735910189377604e-06, "loss": 0.578, "step": 9674 }, { "epoch": 0.405562600211689, "grad_norm": 1.7369071245193481, "learning_rate": 6.7352735472200215e-06, "loss": 0.5619, "step": 9675 }, { "epoch": 0.40560451882669796, "grad_norm": 1.5750868320465088, "learning_rate": 6.734636873075071e-06, "loss": 0.4892, "step": 9676 }, { "epoch": 0.4056464374417069, "grad_norm": 2.0920658111572266, "learning_rate": 6.7340001669544894e-06, "loss": 0.5379, "step": 9677 }, { "epoch": 0.4056883560567159, "grad_norm": 1.9610495567321777, "learning_rate": 6.73336342887001e-06, "loss": 0.542, "step": 9678 }, { "epoch": 0.40573027467172484, "grad_norm": 1.637059211730957, "learning_rate": 6.732726658833373e-06, "loss": 0.5694, "step": 9679 }, { "epoch": 0.4057721932867338, "grad_norm": 1.4975651502609253, "learning_rate": 6.732089856856316e-06, "loss": 0.4594, "step": 9680 }, { "epoch": 0.40581411190174277, "grad_norm": 1.677675485610962, "learning_rate": 6.731453022950576e-06, "loss": 0.5267, "step": 9681 }, { "epoch": 0.40585603051675173, "grad_norm": 1.4781744480133057, "learning_rate": 6.730816157127894e-06, "loss": 0.5118, "step": 9682 }, { "epoch": 0.4058979491317607, "grad_norm": 1.700773000717163, "learning_rate": 6.7301792594000086e-06, "loss": 0.4164, "step": 9683 }, { "epoch": 0.40593986774676966, "grad_norm": 1.5283544063568115, "learning_rate": 6.72954232977866e-06, "loss": 0.4557, "step": 9684 }, { "epoch": 0.4059817863617786, "grad_norm": 1.7654210329055786, "learning_rate": 6.728905368275589e-06, "loss": 0.5682, "step": 9685 }, { "epoch": 0.4060237049767876, "grad_norm": 2.7677979469299316, "learning_rate": 6.72826837490254e-06, "loss": 0.5778, "step": 9686 }, { "epoch": 0.40606562359179654, "grad_norm": 1.8096885681152344, "learning_rate": 6.727631349671251e-06, "loss": 0.5589, "step": 9687 }, { "epoch": 0.4061075422068055, "grad_norm": 1.665595293045044, "learning_rate": 6.726994292593466e-06, "loss": 0.5483, "step": 9688 }, { "epoch": 0.40614946082181447, "grad_norm": 1.9489295482635498, "learning_rate": 6.72635720368093e-06, "loss": 0.5037, "step": 9689 }, { "epoch": 0.40619137943682343, "grad_norm": 1.4934604167938232, "learning_rate": 6.725720082945385e-06, "loss": 0.4488, "step": 9690 }, { "epoch": 0.4062332980518324, "grad_norm": 1.5881192684173584, "learning_rate": 6.725082930398576e-06, "loss": 0.5302, "step": 9691 }, { "epoch": 0.40627521666684135, "grad_norm": 1.5420430898666382, "learning_rate": 6.7244457460522485e-06, "loss": 0.5235, "step": 9692 }, { "epoch": 0.40631713528185026, "grad_norm": 1.6677920818328857, "learning_rate": 6.723808529918148e-06, "loss": 0.5387, "step": 9693 }, { "epoch": 0.4063590538968592, "grad_norm": 1.6079163551330566, "learning_rate": 6.723171282008017e-06, "loss": 0.5741, "step": 9694 }, { "epoch": 0.4064009725118682, "grad_norm": 1.7064096927642822, "learning_rate": 6.72253400233361e-06, "loss": 0.5312, "step": 9695 }, { "epoch": 0.40644289112687715, "grad_norm": 1.7320685386657715, "learning_rate": 6.721896690906669e-06, "loss": 0.5538, "step": 9696 }, { "epoch": 0.4064848097418861, "grad_norm": 2.3013100624084473, "learning_rate": 6.721259347738941e-06, "loss": 0.585, "step": 9697 }, { "epoch": 0.4065267283568951, "grad_norm": 3.9984633922576904, "learning_rate": 6.720621972842177e-06, "loss": 0.5201, "step": 9698 }, { "epoch": 0.40656864697190404, "grad_norm": 1.6705666780471802, "learning_rate": 6.719984566228127e-06, "loss": 0.5132, "step": 9699 }, { "epoch": 0.406610565586913, "grad_norm": 1.8001619577407837, "learning_rate": 6.719347127908539e-06, "loss": 0.5278, "step": 9700 }, { "epoch": 0.40665248420192196, "grad_norm": 1.6541725397109985, "learning_rate": 6.7187096578951626e-06, "loss": 0.5462, "step": 9701 }, { "epoch": 0.4066944028169309, "grad_norm": 1.756857991218567, "learning_rate": 6.718072156199751e-06, "loss": 0.4769, "step": 9702 }, { "epoch": 0.4067363214319399, "grad_norm": 2.0391571521759033, "learning_rate": 6.717434622834053e-06, "loss": 0.4517, "step": 9703 }, { "epoch": 0.40677824004694885, "grad_norm": 1.6183871030807495, "learning_rate": 6.7167970578098215e-06, "loss": 0.5194, "step": 9704 }, { "epoch": 0.4068201586619578, "grad_norm": 1.4483461380004883, "learning_rate": 6.716159461138812e-06, "loss": 0.4973, "step": 9705 }, { "epoch": 0.40686207727696677, "grad_norm": 1.7709486484527588, "learning_rate": 6.715521832832774e-06, "loss": 0.4718, "step": 9706 }, { "epoch": 0.40690399589197573, "grad_norm": 1.915474772453308, "learning_rate": 6.714884172903463e-06, "loss": 0.5343, "step": 9707 }, { "epoch": 0.4069459145069847, "grad_norm": 1.5914181470870972, "learning_rate": 6.714246481362633e-06, "loss": 0.5081, "step": 9708 }, { "epoch": 0.40698783312199366, "grad_norm": 1.6289011240005493, "learning_rate": 6.713608758222039e-06, "loss": 0.5206, "step": 9709 }, { "epoch": 0.4070297517370026, "grad_norm": 1.6133991479873657, "learning_rate": 6.7129710034934355e-06, "loss": 0.5026, "step": 9710 }, { "epoch": 0.4070716703520116, "grad_norm": 1.6878844499588013, "learning_rate": 6.7123332171885815e-06, "loss": 0.4674, "step": 9711 }, { "epoch": 0.40711358896702055, "grad_norm": 1.5541115999221802, "learning_rate": 6.71169539931923e-06, "loss": 0.4646, "step": 9712 }, { "epoch": 0.4071555075820295, "grad_norm": 2.016383409500122, "learning_rate": 6.711057549897141e-06, "loss": 0.5278, "step": 9713 }, { "epoch": 0.40719742619703847, "grad_norm": 1.9522831439971924, "learning_rate": 6.710419668934073e-06, "loss": 0.4651, "step": 9714 }, { "epoch": 0.40723934481204743, "grad_norm": 1.7176127433776855, "learning_rate": 6.709781756441781e-06, "loss": 0.5504, "step": 9715 }, { "epoch": 0.4072812634270564, "grad_norm": 1.7702945470809937, "learning_rate": 6.7091438124320266e-06, "loss": 0.5824, "step": 9716 }, { "epoch": 0.40732318204206536, "grad_norm": 1.7884973287582397, "learning_rate": 6.708505836916569e-06, "loss": 0.5033, "step": 9717 }, { "epoch": 0.40736510065707426, "grad_norm": 1.6153650283813477, "learning_rate": 6.707867829907169e-06, "loss": 0.5068, "step": 9718 }, { "epoch": 0.4074070192720832, "grad_norm": 1.412953495979309, "learning_rate": 6.707229791415585e-06, "loss": 0.4804, "step": 9719 }, { "epoch": 0.4074489378870922, "grad_norm": 1.7164616584777832, "learning_rate": 6.7065917214535804e-06, "loss": 0.5411, "step": 9720 }, { "epoch": 0.40749085650210115, "grad_norm": 1.768784999847412, "learning_rate": 6.705953620032917e-06, "loss": 0.5102, "step": 9721 }, { "epoch": 0.4075327751171101, "grad_norm": 1.5556446313858032, "learning_rate": 6.705315487165356e-06, "loss": 0.5118, "step": 9722 }, { "epoch": 0.4075746937321191, "grad_norm": 1.6472370624542236, "learning_rate": 6.7046773228626625e-06, "loss": 0.498, "step": 9723 }, { "epoch": 0.40761661234712804, "grad_norm": 1.744338870048523, "learning_rate": 6.704039127136601e-06, "loss": 0.4937, "step": 9724 }, { "epoch": 0.407658530962137, "grad_norm": 1.7250041961669922, "learning_rate": 6.7034008999989306e-06, "loss": 0.5782, "step": 9725 }, { "epoch": 0.40770044957714596, "grad_norm": 1.5721590518951416, "learning_rate": 6.702762641461422e-06, "loss": 0.5046, "step": 9726 }, { "epoch": 0.4077423681921549, "grad_norm": 1.859072208404541, "learning_rate": 6.702124351535837e-06, "loss": 0.5133, "step": 9727 }, { "epoch": 0.4077842868071639, "grad_norm": 1.7058812379837036, "learning_rate": 6.701486030233943e-06, "loss": 0.522, "step": 9728 }, { "epoch": 0.40782620542217285, "grad_norm": 1.9457937479019165, "learning_rate": 6.700847677567506e-06, "loss": 0.5316, "step": 9729 }, { "epoch": 0.4078681240371818, "grad_norm": 1.5448721647262573, "learning_rate": 6.700209293548294e-06, "loss": 0.4719, "step": 9730 }, { "epoch": 0.4079100426521908, "grad_norm": 1.577738642692566, "learning_rate": 6.699570878188073e-06, "loss": 0.4993, "step": 9731 }, { "epoch": 0.40795196126719974, "grad_norm": 1.8815748691558838, "learning_rate": 6.698932431498613e-06, "loss": 0.5246, "step": 9732 }, { "epoch": 0.4079938798822087, "grad_norm": 1.7829688787460327, "learning_rate": 6.6982939534916834e-06, "loss": 0.5203, "step": 9733 }, { "epoch": 0.40803579849721766, "grad_norm": 1.755339503288269, "learning_rate": 6.697655444179051e-06, "loss": 0.5145, "step": 9734 }, { "epoch": 0.4080777171122266, "grad_norm": 1.74708092212677, "learning_rate": 6.697016903572489e-06, "loss": 0.5547, "step": 9735 }, { "epoch": 0.4081196357272356, "grad_norm": 1.9359827041625977, "learning_rate": 6.696378331683765e-06, "loss": 0.5269, "step": 9736 }, { "epoch": 0.40816155434224455, "grad_norm": 1.7756879329681396, "learning_rate": 6.6957397285246524e-06, "loss": 0.5434, "step": 9737 }, { "epoch": 0.4082034729572535, "grad_norm": 2.1916916370391846, "learning_rate": 6.695101094106921e-06, "loss": 0.4816, "step": 9738 }, { "epoch": 0.4082453915722625, "grad_norm": 1.612817645072937, "learning_rate": 6.6944624284423465e-06, "loss": 0.5004, "step": 9739 }, { "epoch": 0.40828731018727143, "grad_norm": 1.6929506063461304, "learning_rate": 6.6938237315426965e-06, "loss": 0.4859, "step": 9740 }, { "epoch": 0.4083292288022804, "grad_norm": 1.78228759765625, "learning_rate": 6.693185003419751e-06, "loss": 0.5013, "step": 9741 }, { "epoch": 0.40837114741728936, "grad_norm": 2.91616153717041, "learning_rate": 6.692546244085279e-06, "loss": 0.5133, "step": 9742 }, { "epoch": 0.40841306603229827, "grad_norm": 1.5964237451553345, "learning_rate": 6.691907453551058e-06, "loss": 0.5059, "step": 9743 }, { "epoch": 0.40845498464730723, "grad_norm": 1.6045119762420654, "learning_rate": 6.69126863182886e-06, "loss": 0.5304, "step": 9744 }, { "epoch": 0.4084969032623162, "grad_norm": 1.727583408355713, "learning_rate": 6.690629778930465e-06, "loss": 0.5516, "step": 9745 }, { "epoch": 0.40853882187732515, "grad_norm": 1.6288256645202637, "learning_rate": 6.689990894867645e-06, "loss": 0.5187, "step": 9746 }, { "epoch": 0.4085807404923341, "grad_norm": 1.5280048847198486, "learning_rate": 6.689351979652183e-06, "loss": 0.5134, "step": 9747 }, { "epoch": 0.4086226591073431, "grad_norm": 1.736283540725708, "learning_rate": 6.688713033295849e-06, "loss": 0.5292, "step": 9748 }, { "epoch": 0.40866457772235204, "grad_norm": 1.7881312370300293, "learning_rate": 6.688074055810425e-06, "loss": 0.5443, "step": 9749 }, { "epoch": 0.408706496337361, "grad_norm": 1.817477822303772, "learning_rate": 6.687435047207691e-06, "loss": 0.4763, "step": 9750 }, { "epoch": 0.40874841495236997, "grad_norm": 1.7566055059432983, "learning_rate": 6.6867960074994256e-06, "loss": 0.46, "step": 9751 }, { "epoch": 0.4087903335673789, "grad_norm": 1.6700925827026367, "learning_rate": 6.6861569366974055e-06, "loss": 0.5285, "step": 9752 }, { "epoch": 0.4088322521823879, "grad_norm": 1.5062596797943115, "learning_rate": 6.685517834813415e-06, "loss": 0.5578, "step": 9753 }, { "epoch": 0.40887417079739685, "grad_norm": 1.6488460302352905, "learning_rate": 6.684878701859232e-06, "loss": 0.5205, "step": 9754 }, { "epoch": 0.4089160894124058, "grad_norm": 1.6502512693405151, "learning_rate": 6.68423953784664e-06, "loss": 0.5297, "step": 9755 }, { "epoch": 0.4089580080274148, "grad_norm": 2.0575106143951416, "learning_rate": 6.683600342787422e-06, "loss": 0.5332, "step": 9756 }, { "epoch": 0.40899992664242374, "grad_norm": 1.6582423448562622, "learning_rate": 6.682961116693358e-06, "loss": 0.5453, "step": 9757 }, { "epoch": 0.4090418452574327, "grad_norm": 2.0612382888793945, "learning_rate": 6.682321859576231e-06, "loss": 0.5476, "step": 9758 }, { "epoch": 0.40908376387244166, "grad_norm": 1.8142794370651245, "learning_rate": 6.6816825714478275e-06, "loss": 0.5181, "step": 9759 }, { "epoch": 0.4091256824874506, "grad_norm": 1.7554385662078857, "learning_rate": 6.68104325231993e-06, "loss": 0.5053, "step": 9760 }, { "epoch": 0.4091676011024596, "grad_norm": 1.5327813625335693, "learning_rate": 6.680403902204325e-06, "loss": 0.4934, "step": 9761 }, { "epoch": 0.40920951971746855, "grad_norm": 1.7214016914367676, "learning_rate": 6.679764521112797e-06, "loss": 0.5402, "step": 9762 }, { "epoch": 0.4092514383324775, "grad_norm": 1.6367757320404053, "learning_rate": 6.679125109057132e-06, "loss": 0.5268, "step": 9763 }, { "epoch": 0.4092933569474865, "grad_norm": 1.688653588294983, "learning_rate": 6.678485666049117e-06, "loss": 0.4891, "step": 9764 }, { "epoch": 0.40933527556249544, "grad_norm": 1.5944494009017944, "learning_rate": 6.677846192100539e-06, "loss": 0.6016, "step": 9765 }, { "epoch": 0.4093771941775044, "grad_norm": 1.7817188501358032, "learning_rate": 6.677206687223187e-06, "loss": 0.4967, "step": 9766 }, { "epoch": 0.40941911279251336, "grad_norm": 1.5403472185134888, "learning_rate": 6.676567151428847e-06, "loss": 0.5565, "step": 9767 }, { "epoch": 0.40946103140752227, "grad_norm": 1.5095387697219849, "learning_rate": 6.675927584729309e-06, "loss": 0.4674, "step": 9768 }, { "epoch": 0.40950295002253123, "grad_norm": 1.7904655933380127, "learning_rate": 6.675287987136364e-06, "loss": 0.5046, "step": 9769 }, { "epoch": 0.4095448686375402, "grad_norm": 1.8482180833816528, "learning_rate": 6.674648358661802e-06, "loss": 0.4721, "step": 9770 }, { "epoch": 0.40958678725254916, "grad_norm": 1.7900773286819458, "learning_rate": 6.6740086993174115e-06, "loss": 0.4789, "step": 9771 }, { "epoch": 0.4096287058675581, "grad_norm": 1.5027213096618652, "learning_rate": 6.673369009114985e-06, "loss": 0.4718, "step": 9772 }, { "epoch": 0.4096706244825671, "grad_norm": 1.7376681566238403, "learning_rate": 6.672729288066313e-06, "loss": 0.5094, "step": 9773 }, { "epoch": 0.40971254309757604, "grad_norm": 1.5640805959701538, "learning_rate": 6.67208953618319e-06, "loss": 0.5215, "step": 9774 }, { "epoch": 0.409754461712585, "grad_norm": 1.8038432598114014, "learning_rate": 6.6714497534774095e-06, "loss": 0.5578, "step": 9775 }, { "epoch": 0.40979638032759397, "grad_norm": 1.6524301767349243, "learning_rate": 6.670809939960762e-06, "loss": 0.5256, "step": 9776 }, { "epoch": 0.40983829894260293, "grad_norm": 1.7171120643615723, "learning_rate": 6.670170095645041e-06, "loss": 0.466, "step": 9777 }, { "epoch": 0.4098802175576119, "grad_norm": 1.8035690784454346, "learning_rate": 6.669530220542046e-06, "loss": 0.4454, "step": 9778 }, { "epoch": 0.40992213617262085, "grad_norm": 1.6200331449508667, "learning_rate": 6.66889031466357e-06, "loss": 0.4814, "step": 9779 }, { "epoch": 0.4099640547876298, "grad_norm": 4.9394378662109375, "learning_rate": 6.668250378021406e-06, "loss": 0.5025, "step": 9780 }, { "epoch": 0.4100059734026388, "grad_norm": 1.713623285293579, "learning_rate": 6.667610410627354e-06, "loss": 0.5343, "step": 9781 }, { "epoch": 0.41004789201764774, "grad_norm": 1.4902790784835815, "learning_rate": 6.666970412493209e-06, "loss": 0.4634, "step": 9782 }, { "epoch": 0.4100898106326567, "grad_norm": 1.7857052087783813, "learning_rate": 6.666330383630768e-06, "loss": 0.4996, "step": 9783 }, { "epoch": 0.41013172924766567, "grad_norm": 1.8106141090393066, "learning_rate": 6.665690324051832e-06, "loss": 0.5357, "step": 9784 }, { "epoch": 0.41017364786267463, "grad_norm": 1.7982611656188965, "learning_rate": 6.665050233768198e-06, "loss": 0.5495, "step": 9785 }, { "epoch": 0.4102155664776836, "grad_norm": 1.699447512626648, "learning_rate": 6.664410112791662e-06, "loss": 0.5165, "step": 9786 }, { "epoch": 0.41025748509269255, "grad_norm": 2.2862517833709717, "learning_rate": 6.663769961134028e-06, "loss": 0.523, "step": 9787 }, { "epoch": 0.4102994037077015, "grad_norm": 2.4892852306365967, "learning_rate": 6.6631297788070945e-06, "loss": 0.5185, "step": 9788 }, { "epoch": 0.4103413223227105, "grad_norm": 1.7509225606918335, "learning_rate": 6.662489565822664e-06, "loss": 0.511, "step": 9789 }, { "epoch": 0.41038324093771944, "grad_norm": 2.115933656692505, "learning_rate": 6.661849322192535e-06, "loss": 0.5247, "step": 9790 }, { "epoch": 0.4104251595527284, "grad_norm": 1.9060291051864624, "learning_rate": 6.661209047928513e-06, "loss": 0.5382, "step": 9791 }, { "epoch": 0.41046707816773736, "grad_norm": 1.554975986480713, "learning_rate": 6.660568743042397e-06, "loss": 0.4751, "step": 9792 }, { "epoch": 0.41050899678274627, "grad_norm": 1.6469262838363647, "learning_rate": 6.659928407545992e-06, "loss": 0.4591, "step": 9793 }, { "epoch": 0.41055091539775523, "grad_norm": 1.7184799909591675, "learning_rate": 6.659288041451104e-06, "loss": 0.5661, "step": 9794 }, { "epoch": 0.4105928340127642, "grad_norm": 2.0228066444396973, "learning_rate": 6.658647644769532e-06, "loss": 0.4916, "step": 9795 }, { "epoch": 0.41063475262777316, "grad_norm": 1.5770617723464966, "learning_rate": 6.658007217513085e-06, "loss": 0.5445, "step": 9796 }, { "epoch": 0.4106766712427821, "grad_norm": 1.6645389795303345, "learning_rate": 6.657366759693567e-06, "loss": 0.5315, "step": 9797 }, { "epoch": 0.4107185898577911, "grad_norm": 1.627153992652893, "learning_rate": 6.656726271322783e-06, "loss": 0.5289, "step": 9798 }, { "epoch": 0.41076050847280005, "grad_norm": 1.5494863986968994, "learning_rate": 6.656085752412542e-06, "loss": 0.5411, "step": 9799 }, { "epoch": 0.410802427087809, "grad_norm": 1.8203890323638916, "learning_rate": 6.655445202974649e-06, "loss": 0.5182, "step": 9800 }, { "epoch": 0.41084434570281797, "grad_norm": 3.8911495208740234, "learning_rate": 6.654804623020912e-06, "loss": 0.5526, "step": 9801 }, { "epoch": 0.41088626431782693, "grad_norm": 1.5246814489364624, "learning_rate": 6.654164012563139e-06, "loss": 0.5287, "step": 9802 }, { "epoch": 0.4109281829328359, "grad_norm": 1.8716994524002075, "learning_rate": 6.653523371613142e-06, "loss": 0.556, "step": 9803 }, { "epoch": 0.41097010154784486, "grad_norm": 1.6892794370651245, "learning_rate": 6.652882700182726e-06, "loss": 0.4648, "step": 9804 }, { "epoch": 0.4110120201628538, "grad_norm": 1.5157428979873657, "learning_rate": 6.652241998283701e-06, "loss": 0.5016, "step": 9805 }, { "epoch": 0.4110539387778628, "grad_norm": 1.986281156539917, "learning_rate": 6.6516012659278795e-06, "loss": 0.5703, "step": 9806 }, { "epoch": 0.41109585739287174, "grad_norm": 1.7461658716201782, "learning_rate": 6.650960503127072e-06, "loss": 0.5539, "step": 9807 }, { "epoch": 0.4111377760078807, "grad_norm": 1.9237828254699707, "learning_rate": 6.65031970989309e-06, "loss": 0.5316, "step": 9808 }, { "epoch": 0.41117969462288967, "grad_norm": 1.6415725946426392, "learning_rate": 6.649678886237747e-06, "loss": 0.5626, "step": 9809 }, { "epoch": 0.41122161323789863, "grad_norm": 1.8378798961639404, "learning_rate": 6.649038032172853e-06, "loss": 0.543, "step": 9810 }, { "epoch": 0.4112635318529076, "grad_norm": 1.6124765872955322, "learning_rate": 6.648397147710222e-06, "loss": 0.4932, "step": 9811 }, { "epoch": 0.41130545046791656, "grad_norm": 1.651768684387207, "learning_rate": 6.647756232861669e-06, "loss": 0.5484, "step": 9812 }, { "epoch": 0.4113473690829255, "grad_norm": 1.9144583940505981, "learning_rate": 6.64711528763901e-06, "loss": 0.5162, "step": 9813 }, { "epoch": 0.4113892876979345, "grad_norm": 1.5563538074493408, "learning_rate": 6.646474312054056e-06, "loss": 0.4698, "step": 9814 }, { "epoch": 0.41143120631294344, "grad_norm": 1.481329321861267, "learning_rate": 6.645833306118623e-06, "loss": 0.4879, "step": 9815 }, { "epoch": 0.4114731249279524, "grad_norm": 1.6008774042129517, "learning_rate": 6.64519226984453e-06, "loss": 0.5295, "step": 9816 }, { "epoch": 0.41151504354296137, "grad_norm": 1.605129361152649, "learning_rate": 6.644551203243592e-06, "loss": 0.5639, "step": 9817 }, { "epoch": 0.4115569621579703, "grad_norm": 1.719041109085083, "learning_rate": 6.643910106327626e-06, "loss": 0.5024, "step": 9818 }, { "epoch": 0.41159888077297924, "grad_norm": 1.6528149843215942, "learning_rate": 6.64326897910845e-06, "loss": 0.4976, "step": 9819 }, { "epoch": 0.4116407993879882, "grad_norm": 1.6938797235488892, "learning_rate": 6.642627821597879e-06, "loss": 0.5339, "step": 9820 }, { "epoch": 0.41168271800299716, "grad_norm": 2.0396926403045654, "learning_rate": 6.641986633807738e-06, "loss": 0.4911, "step": 9821 }, { "epoch": 0.4117246366180061, "grad_norm": 1.596168041229248, "learning_rate": 6.641345415749844e-06, "loss": 0.4798, "step": 9822 }, { "epoch": 0.4117665552330151, "grad_norm": 2.0114057064056396, "learning_rate": 6.640704167436014e-06, "loss": 0.5138, "step": 9823 }, { "epoch": 0.41180847384802405, "grad_norm": 1.7770910263061523, "learning_rate": 6.640062888878072e-06, "loss": 0.479, "step": 9824 }, { "epoch": 0.411850392463033, "grad_norm": 2.204766273498535, "learning_rate": 6.6394215800878384e-06, "loss": 0.5226, "step": 9825 }, { "epoch": 0.411892311078042, "grad_norm": 1.7912663221359253, "learning_rate": 6.638780241077134e-06, "loss": 0.5119, "step": 9826 }, { "epoch": 0.41193422969305094, "grad_norm": 2.080540895462036, "learning_rate": 6.638138871857782e-06, "loss": 0.5035, "step": 9827 }, { "epoch": 0.4119761483080599, "grad_norm": 1.5042028427124023, "learning_rate": 6.637497472441603e-06, "loss": 0.4839, "step": 9828 }, { "epoch": 0.41201806692306886, "grad_norm": 1.7519910335540771, "learning_rate": 6.636856042840423e-06, "loss": 0.4823, "step": 9829 }, { "epoch": 0.4120599855380778, "grad_norm": 1.5694448947906494, "learning_rate": 6.636214583066064e-06, "loss": 0.5373, "step": 9830 }, { "epoch": 0.4121019041530868, "grad_norm": 1.59078049659729, "learning_rate": 6.635573093130353e-06, "loss": 0.4732, "step": 9831 }, { "epoch": 0.41214382276809575, "grad_norm": 1.664397120475769, "learning_rate": 6.634931573045111e-06, "loss": 0.5029, "step": 9832 }, { "epoch": 0.4121857413831047, "grad_norm": 1.6021069288253784, "learning_rate": 6.634290022822166e-06, "loss": 0.4955, "step": 9833 }, { "epoch": 0.41222765999811367, "grad_norm": 1.917165756225586, "learning_rate": 6.6336484424733445e-06, "loss": 0.564, "step": 9834 }, { "epoch": 0.41226957861312263, "grad_norm": 1.510884165763855, "learning_rate": 6.633006832010474e-06, "loss": 0.5071, "step": 9835 }, { "epoch": 0.4123114972281316, "grad_norm": 1.8047804832458496, "learning_rate": 6.632365191445378e-06, "loss": 0.4961, "step": 9836 }, { "epoch": 0.41235341584314056, "grad_norm": 1.3826762437820435, "learning_rate": 6.6317235207898855e-06, "loss": 0.4831, "step": 9837 }, { "epoch": 0.4123953344581495, "grad_norm": 1.8685771226882935, "learning_rate": 6.631081820055828e-06, "loss": 0.5388, "step": 9838 }, { "epoch": 0.4124372530731585, "grad_norm": 1.4984732866287231, "learning_rate": 6.63044008925503e-06, "loss": 0.5094, "step": 9839 }, { "epoch": 0.41247917168816745, "grad_norm": 1.6372264623641968, "learning_rate": 6.629798328399324e-06, "loss": 0.4737, "step": 9840 }, { "epoch": 0.4125210903031764, "grad_norm": 2.2893805503845215, "learning_rate": 6.62915653750054e-06, "loss": 0.5013, "step": 9841 }, { "epoch": 0.41256300891818537, "grad_norm": 1.890334129333496, "learning_rate": 6.628514716570505e-06, "loss": 0.5255, "step": 9842 }, { "epoch": 0.4126049275331943, "grad_norm": 1.850669503211975, "learning_rate": 6.6278728656210545e-06, "loss": 0.4912, "step": 9843 }, { "epoch": 0.41264684614820324, "grad_norm": 1.6915724277496338, "learning_rate": 6.627230984664017e-06, "loss": 0.5418, "step": 9844 }, { "epoch": 0.4126887647632122, "grad_norm": 1.9877738952636719, "learning_rate": 6.626589073711227e-06, "loss": 0.508, "step": 9845 }, { "epoch": 0.41273068337822116, "grad_norm": 1.7599962949752808, "learning_rate": 6.625947132774515e-06, "loss": 0.5335, "step": 9846 }, { "epoch": 0.4127726019932301, "grad_norm": 1.6591508388519287, "learning_rate": 6.625305161865716e-06, "loss": 0.501, "step": 9847 }, { "epoch": 0.4128145206082391, "grad_norm": 1.461264967918396, "learning_rate": 6.624663160996663e-06, "loss": 0.4652, "step": 9848 }, { "epoch": 0.41285643922324805, "grad_norm": 1.6313821077346802, "learning_rate": 6.624021130179191e-06, "loss": 0.536, "step": 9849 }, { "epoch": 0.412898357838257, "grad_norm": 1.8450548648834229, "learning_rate": 6.623379069425135e-06, "loss": 0.5125, "step": 9850 }, { "epoch": 0.412940276453266, "grad_norm": 1.5523852109909058, "learning_rate": 6.6227369787463294e-06, "loss": 0.4737, "step": 9851 }, { "epoch": 0.41298219506827494, "grad_norm": 1.9085150957107544, "learning_rate": 6.622094858154611e-06, "loss": 0.493, "step": 9852 }, { "epoch": 0.4130241136832839, "grad_norm": 2.004762649536133, "learning_rate": 6.621452707661817e-06, "loss": 0.5015, "step": 9853 }, { "epoch": 0.41306603229829286, "grad_norm": 1.568422794342041, "learning_rate": 6.620810527279784e-06, "loss": 0.5307, "step": 9854 }, { "epoch": 0.4131079509133018, "grad_norm": 1.626049518585205, "learning_rate": 6.620168317020349e-06, "loss": 0.5518, "step": 9855 }, { "epoch": 0.4131498695283108, "grad_norm": 1.9116984605789185, "learning_rate": 6.61952607689535e-06, "loss": 0.485, "step": 9856 }, { "epoch": 0.41319178814331975, "grad_norm": 1.7665162086486816, "learning_rate": 6.618883806916627e-06, "loss": 0.4988, "step": 9857 }, { "epoch": 0.4132337067583287, "grad_norm": 1.9461417198181152, "learning_rate": 6.618241507096021e-06, "loss": 0.4734, "step": 9858 }, { "epoch": 0.4132756253733377, "grad_norm": 1.5311726331710815, "learning_rate": 6.617599177445369e-06, "loss": 0.4347, "step": 9859 }, { "epoch": 0.41331754398834664, "grad_norm": 1.5665907859802246, "learning_rate": 6.616956817976512e-06, "loss": 0.4912, "step": 9860 }, { "epoch": 0.4133594626033556, "grad_norm": 1.7715060710906982, "learning_rate": 6.6163144287012915e-06, "loss": 0.5064, "step": 9861 }, { "epoch": 0.41340138121836456, "grad_norm": 1.602110743522644, "learning_rate": 6.615672009631548e-06, "loss": 0.5152, "step": 9862 }, { "epoch": 0.4134432998333735, "grad_norm": 1.76734459400177, "learning_rate": 6.615029560779126e-06, "loss": 0.5049, "step": 9863 }, { "epoch": 0.4134852184483825, "grad_norm": 1.741329550743103, "learning_rate": 6.614387082155868e-06, "loss": 0.586, "step": 9864 }, { "epoch": 0.41352713706339145, "grad_norm": 1.4442461729049683, "learning_rate": 6.6137445737736135e-06, "loss": 0.4657, "step": 9865 }, { "epoch": 0.4135690556784004, "grad_norm": 1.6934704780578613, "learning_rate": 6.6131020356442085e-06, "loss": 0.5485, "step": 9866 }, { "epoch": 0.4136109742934094, "grad_norm": 1.4766724109649658, "learning_rate": 6.612459467779499e-06, "loss": 0.5145, "step": 9867 }, { "epoch": 0.4136528929084183, "grad_norm": 1.7345467805862427, "learning_rate": 6.611816870191327e-06, "loss": 0.5238, "step": 9868 }, { "epoch": 0.41369481152342724, "grad_norm": 2.0110554695129395, "learning_rate": 6.611174242891542e-06, "loss": 0.5166, "step": 9869 }, { "epoch": 0.4137367301384362, "grad_norm": 1.7167437076568604, "learning_rate": 6.610531585891985e-06, "loss": 0.4894, "step": 9870 }, { "epoch": 0.41377864875344517, "grad_norm": 1.537941336631775, "learning_rate": 6.609888899204505e-06, "loss": 0.4576, "step": 9871 }, { "epoch": 0.41382056736845413, "grad_norm": 1.758751630783081, "learning_rate": 6.609246182840949e-06, "loss": 0.4879, "step": 9872 }, { "epoch": 0.4138624859834631, "grad_norm": 1.5215418338775635, "learning_rate": 6.608603436813164e-06, "loss": 0.5111, "step": 9873 }, { "epoch": 0.41390440459847205, "grad_norm": 1.8593624830245972, "learning_rate": 6.607960661132999e-06, "loss": 0.5111, "step": 9874 }, { "epoch": 0.413946323213481, "grad_norm": 1.5674360990524292, "learning_rate": 6.607317855812303e-06, "loss": 0.5047, "step": 9875 }, { "epoch": 0.41398824182849, "grad_norm": 1.507245659828186, "learning_rate": 6.606675020862923e-06, "loss": 0.5047, "step": 9876 }, { "epoch": 0.41403016044349894, "grad_norm": 1.7873326539993286, "learning_rate": 6.60603215629671e-06, "loss": 0.5293, "step": 9877 }, { "epoch": 0.4140720790585079, "grad_norm": 1.6403306722640991, "learning_rate": 6.605389262125516e-06, "loss": 0.5013, "step": 9878 }, { "epoch": 0.41411399767351686, "grad_norm": 1.4497406482696533, "learning_rate": 6.604746338361187e-06, "loss": 0.4996, "step": 9879 }, { "epoch": 0.4141559162885258, "grad_norm": 1.5281202793121338, "learning_rate": 6.604103385015581e-06, "loss": 0.5349, "step": 9880 }, { "epoch": 0.4141978349035348, "grad_norm": 2.0344953536987305, "learning_rate": 6.603460402100546e-06, "loss": 0.4383, "step": 9881 }, { "epoch": 0.41423975351854375, "grad_norm": 1.76241934299469, "learning_rate": 6.602817389627936e-06, "loss": 0.5442, "step": 9882 }, { "epoch": 0.4142816721335527, "grad_norm": 1.7835534811019897, "learning_rate": 6.6021743476096014e-06, "loss": 0.4941, "step": 9883 }, { "epoch": 0.4143235907485617, "grad_norm": 1.4111745357513428, "learning_rate": 6.601531276057398e-06, "loss": 0.4727, "step": 9884 }, { "epoch": 0.41436550936357064, "grad_norm": 1.8636447191238403, "learning_rate": 6.60088817498318e-06, "loss": 0.5447, "step": 9885 }, { "epoch": 0.4144074279785796, "grad_norm": 1.8425763845443726, "learning_rate": 6.600245044398801e-06, "loss": 0.524, "step": 9886 }, { "epoch": 0.41444934659358856, "grad_norm": 1.810727596282959, "learning_rate": 6.599601884316118e-06, "loss": 0.5503, "step": 9887 }, { "epoch": 0.4144912652085975, "grad_norm": 1.62720787525177, "learning_rate": 6.598958694746985e-06, "loss": 0.5189, "step": 9888 }, { "epoch": 0.4145331838236065, "grad_norm": 1.7194534540176392, "learning_rate": 6.598315475703259e-06, "loss": 0.5567, "step": 9889 }, { "epoch": 0.41457510243861545, "grad_norm": 1.8554257154464722, "learning_rate": 6.597672227196798e-06, "loss": 0.5517, "step": 9890 }, { "epoch": 0.4146170210536244, "grad_norm": 1.951115369796753, "learning_rate": 6.597028949239456e-06, "loss": 0.5206, "step": 9891 }, { "epoch": 0.4146589396686334, "grad_norm": 1.6670674085617065, "learning_rate": 6.596385641843095e-06, "loss": 0.5252, "step": 9892 }, { "epoch": 0.4147008582836423, "grad_norm": 1.7502458095550537, "learning_rate": 6.595742305019571e-06, "loss": 0.5219, "step": 9893 }, { "epoch": 0.41474277689865124, "grad_norm": 1.51312255859375, "learning_rate": 6.595098938780743e-06, "loss": 0.484, "step": 9894 }, { "epoch": 0.4147846955136602, "grad_norm": 1.5889419317245483, "learning_rate": 6.594455543138472e-06, "loss": 0.4281, "step": 9895 }, { "epoch": 0.41482661412866917, "grad_norm": 1.7206863164901733, "learning_rate": 6.593812118104617e-06, "loss": 0.5393, "step": 9896 }, { "epoch": 0.41486853274367813, "grad_norm": 1.6765599250793457, "learning_rate": 6.5931686636910385e-06, "loss": 0.4997, "step": 9897 }, { "epoch": 0.4149104513586871, "grad_norm": 1.7121660709381104, "learning_rate": 6.592525179909599e-06, "loss": 0.5215, "step": 9898 }, { "epoch": 0.41495236997369606, "grad_norm": 1.9034558534622192, "learning_rate": 6.59188166677216e-06, "loss": 0.4929, "step": 9899 }, { "epoch": 0.414994288588705, "grad_norm": 1.5310903787612915, "learning_rate": 6.591238124290581e-06, "loss": 0.4805, "step": 9900 }, { "epoch": 0.415036207203714, "grad_norm": 1.4559710025787354, "learning_rate": 6.59059455247673e-06, "loss": 0.5116, "step": 9901 }, { "epoch": 0.41507812581872294, "grad_norm": 2.293278455734253, "learning_rate": 6.589950951342466e-06, "loss": 0.49, "step": 9902 }, { "epoch": 0.4151200444337319, "grad_norm": 1.5698127746582031, "learning_rate": 6.5893073208996515e-06, "loss": 0.4853, "step": 9903 }, { "epoch": 0.41516196304874087, "grad_norm": 1.9829072952270508, "learning_rate": 6.588663661160157e-06, "loss": 0.536, "step": 9904 }, { "epoch": 0.41520388166374983, "grad_norm": 1.6088420152664185, "learning_rate": 6.588019972135844e-06, "loss": 0.4881, "step": 9905 }, { "epoch": 0.4152458002787588, "grad_norm": 1.5472564697265625, "learning_rate": 6.587376253838578e-06, "loss": 0.4806, "step": 9906 }, { "epoch": 0.41528771889376775, "grad_norm": 1.591851830482483, "learning_rate": 6.586732506280225e-06, "loss": 0.5247, "step": 9907 }, { "epoch": 0.4153296375087767, "grad_norm": 1.6866904497146606, "learning_rate": 6.586088729472651e-06, "loss": 0.503, "step": 9908 }, { "epoch": 0.4153715561237857, "grad_norm": 1.7859270572662354, "learning_rate": 6.585444923427724e-06, "loss": 0.5195, "step": 9909 }, { "epoch": 0.41541347473879464, "grad_norm": 1.9249119758605957, "learning_rate": 6.584801088157311e-06, "loss": 0.5239, "step": 9910 }, { "epoch": 0.4154553933538036, "grad_norm": 1.6172946691513062, "learning_rate": 6.584157223673284e-06, "loss": 0.521, "step": 9911 }, { "epoch": 0.41549731196881257, "grad_norm": 1.7307381629943848, "learning_rate": 6.583513329987507e-06, "loss": 0.4803, "step": 9912 }, { "epoch": 0.41553923058382153, "grad_norm": 1.6229346990585327, "learning_rate": 6.58286940711185e-06, "loss": 0.5721, "step": 9913 }, { "epoch": 0.4155811491988305, "grad_norm": 1.6275445222854614, "learning_rate": 6.582225455058184e-06, "loss": 0.5055, "step": 9914 }, { "epoch": 0.41562306781383945, "grad_norm": 1.4819875955581665, "learning_rate": 6.581581473838379e-06, "loss": 0.4713, "step": 9915 }, { "epoch": 0.4156649864288484, "grad_norm": 1.5323888063430786, "learning_rate": 6.580937463464306e-06, "loss": 0.4994, "step": 9916 }, { "epoch": 0.4157069050438574, "grad_norm": 1.7190285921096802, "learning_rate": 6.580293423947837e-06, "loss": 0.5338, "step": 9917 }, { "epoch": 0.4157488236588663, "grad_norm": 1.7221217155456543, "learning_rate": 6.579649355300844e-06, "loss": 0.5559, "step": 9918 }, { "epoch": 0.41579074227387525, "grad_norm": 1.6756749153137207, "learning_rate": 6.579005257535197e-06, "loss": 0.5204, "step": 9919 }, { "epoch": 0.4158326608888842, "grad_norm": 1.5016628503799438, "learning_rate": 6.5783611306627734e-06, "loss": 0.507, "step": 9920 }, { "epoch": 0.41587457950389317, "grad_norm": 1.6424280405044556, "learning_rate": 6.577716974695444e-06, "loss": 0.4913, "step": 9921 }, { "epoch": 0.41591649811890213, "grad_norm": 1.6501202583312988, "learning_rate": 6.577072789645081e-06, "loss": 0.4923, "step": 9922 }, { "epoch": 0.4159584167339111, "grad_norm": 1.583823561668396, "learning_rate": 6.576428575523563e-06, "loss": 0.5164, "step": 9923 }, { "epoch": 0.41600033534892006, "grad_norm": 2.0097386837005615, "learning_rate": 6.575784332342764e-06, "loss": 0.5, "step": 9924 }, { "epoch": 0.416042253963929, "grad_norm": 1.518479585647583, "learning_rate": 6.575140060114559e-06, "loss": 0.5338, "step": 9925 }, { "epoch": 0.416084172578938, "grad_norm": 1.6034736633300781, "learning_rate": 6.574495758850824e-06, "loss": 0.5224, "step": 9926 }, { "epoch": 0.41612609119394695, "grad_norm": 1.735177993774414, "learning_rate": 6.5738514285634356e-06, "loss": 0.5422, "step": 9927 }, { "epoch": 0.4161680098089559, "grad_norm": 1.6411716938018799, "learning_rate": 6.573207069264273e-06, "loss": 0.5632, "step": 9928 }, { "epoch": 0.41620992842396487, "grad_norm": 1.4481735229492188, "learning_rate": 6.572562680965213e-06, "loss": 0.4524, "step": 9929 }, { "epoch": 0.41625184703897383, "grad_norm": 1.7415887117385864, "learning_rate": 6.571918263678134e-06, "loss": 0.4958, "step": 9930 }, { "epoch": 0.4162937656539828, "grad_norm": 1.711685061454773, "learning_rate": 6.571273817414915e-06, "loss": 0.5149, "step": 9931 }, { "epoch": 0.41633568426899176, "grad_norm": 2.0050156116485596, "learning_rate": 6.570629342187434e-06, "loss": 0.5768, "step": 9932 }, { "epoch": 0.4163776028840007, "grad_norm": 1.7697232961654663, "learning_rate": 6.569984838007575e-06, "loss": 0.4592, "step": 9933 }, { "epoch": 0.4164195214990097, "grad_norm": 1.671252727508545, "learning_rate": 6.569340304887214e-06, "loss": 0.5319, "step": 9934 }, { "epoch": 0.41646144011401864, "grad_norm": 1.6027358770370483, "learning_rate": 6.568695742838235e-06, "loss": 0.4685, "step": 9935 }, { "epoch": 0.4165033587290276, "grad_norm": 1.6872056722640991, "learning_rate": 6.568051151872519e-06, "loss": 0.5068, "step": 9936 }, { "epoch": 0.41654527734403657, "grad_norm": 2.3435192108154297, "learning_rate": 6.567406532001945e-06, "loss": 0.5224, "step": 9937 }, { "epoch": 0.41658719595904553, "grad_norm": 1.6961201429367065, "learning_rate": 6.5667618832384e-06, "loss": 0.5203, "step": 9938 }, { "epoch": 0.4166291145740545, "grad_norm": 2.030669689178467, "learning_rate": 6.5661172055937686e-06, "loss": 0.479, "step": 9939 }, { "epoch": 0.41667103318906346, "grad_norm": 1.611240267753601, "learning_rate": 6.565472499079928e-06, "loss": 0.456, "step": 9940 }, { "epoch": 0.4167129518040724, "grad_norm": 1.6533536911010742, "learning_rate": 6.564827763708767e-06, "loss": 0.5282, "step": 9941 }, { "epoch": 0.4167548704190814, "grad_norm": 1.7549700736999512, "learning_rate": 6.56418299949217e-06, "loss": 0.4629, "step": 9942 }, { "epoch": 0.4167967890340903, "grad_norm": 1.6430214643478394, "learning_rate": 6.563538206442021e-06, "loss": 0.5015, "step": 9943 }, { "epoch": 0.41683870764909925, "grad_norm": 1.6940191984176636, "learning_rate": 6.562893384570208e-06, "loss": 0.4846, "step": 9944 }, { "epoch": 0.4168806262641082, "grad_norm": 1.7501490116119385, "learning_rate": 6.562248533888616e-06, "loss": 0.5304, "step": 9945 }, { "epoch": 0.4169225448791172, "grad_norm": 1.5591497421264648, "learning_rate": 6.5616036544091295e-06, "loss": 0.4974, "step": 9946 }, { "epoch": 0.41696446349412614, "grad_norm": 1.846045970916748, "learning_rate": 6.5609587461436394e-06, "loss": 0.5407, "step": 9947 }, { "epoch": 0.4170063821091351, "grad_norm": 1.7200837135314941, "learning_rate": 6.560313809104035e-06, "loss": 0.5188, "step": 9948 }, { "epoch": 0.41704830072414406, "grad_norm": 1.8407527208328247, "learning_rate": 6.559668843302202e-06, "loss": 0.5448, "step": 9949 }, { "epoch": 0.417090219339153, "grad_norm": 1.676802158355713, "learning_rate": 6.559023848750028e-06, "loss": 0.4658, "step": 9950 }, { "epoch": 0.417132137954162, "grad_norm": 1.6989610195159912, "learning_rate": 6.5583788254594064e-06, "loss": 0.4625, "step": 9951 }, { "epoch": 0.41717405656917095, "grad_norm": 1.933071494102478, "learning_rate": 6.557733773442225e-06, "loss": 0.5425, "step": 9952 }, { "epoch": 0.4172159751841799, "grad_norm": 1.8418431282043457, "learning_rate": 6.557088692710375e-06, "loss": 0.5253, "step": 9953 }, { "epoch": 0.4172578937991889, "grad_norm": 1.6995418071746826, "learning_rate": 6.556443583275748e-06, "loss": 0.5197, "step": 9954 }, { "epoch": 0.41729981241419783, "grad_norm": 1.6552531719207764, "learning_rate": 6.555798445150234e-06, "loss": 0.5506, "step": 9955 }, { "epoch": 0.4173417310292068, "grad_norm": 1.9923304319381714, "learning_rate": 6.555153278345727e-06, "loss": 0.5371, "step": 9956 }, { "epoch": 0.41738364964421576, "grad_norm": 1.7058688402175903, "learning_rate": 6.55450808287412e-06, "loss": 0.5785, "step": 9957 }, { "epoch": 0.4174255682592247, "grad_norm": 1.8304541110992432, "learning_rate": 6.5538628587473065e-06, "loss": 0.5382, "step": 9958 }, { "epoch": 0.4174674868742337, "grad_norm": 1.7579313516616821, "learning_rate": 6.553217605977177e-06, "loss": 0.5142, "step": 9959 }, { "epoch": 0.41750940548924265, "grad_norm": 1.692914366722107, "learning_rate": 6.55257232457563e-06, "loss": 0.5281, "step": 9960 }, { "epoch": 0.4175513241042516, "grad_norm": 1.474981427192688, "learning_rate": 6.551927014554557e-06, "loss": 0.4485, "step": 9961 }, { "epoch": 0.41759324271926057, "grad_norm": 1.6948708295822144, "learning_rate": 6.551281675925856e-06, "loss": 0.503, "step": 9962 }, { "epoch": 0.41763516133426953, "grad_norm": 1.7842451333999634, "learning_rate": 6.550636308701422e-06, "loss": 0.5308, "step": 9963 }, { "epoch": 0.4176770799492785, "grad_norm": 1.6521507501602173, "learning_rate": 6.549990912893151e-06, "loss": 0.4393, "step": 9964 }, { "epoch": 0.41771899856428746, "grad_norm": 1.5281726121902466, "learning_rate": 6.549345488512939e-06, "loss": 0.5372, "step": 9965 }, { "epoch": 0.4177609171792964, "grad_norm": 1.4408342838287354, "learning_rate": 6.548700035572686e-06, "loss": 0.4676, "step": 9966 }, { "epoch": 0.4178028357943054, "grad_norm": 2.0296072959899902, "learning_rate": 6.548054554084291e-06, "loss": 0.4922, "step": 9967 }, { "epoch": 0.4178447544093143, "grad_norm": 1.81011962890625, "learning_rate": 6.547409044059647e-06, "loss": 0.5534, "step": 9968 }, { "epoch": 0.41788667302432325, "grad_norm": 1.810935378074646, "learning_rate": 6.546763505510658e-06, "loss": 0.5358, "step": 9969 }, { "epoch": 0.4179285916393322, "grad_norm": 1.8649091720581055, "learning_rate": 6.546117938449223e-06, "loss": 0.5187, "step": 9970 }, { "epoch": 0.4179705102543412, "grad_norm": 2.048518180847168, "learning_rate": 6.545472342887239e-06, "loss": 0.5281, "step": 9971 }, { "epoch": 0.41801242886935014, "grad_norm": 1.7821030616760254, "learning_rate": 6.544826718836611e-06, "loss": 0.5182, "step": 9972 }, { "epoch": 0.4180543474843591, "grad_norm": 1.8934130668640137, "learning_rate": 6.544181066309236e-06, "loss": 0.5328, "step": 9973 }, { "epoch": 0.41809626609936806, "grad_norm": 1.7025573253631592, "learning_rate": 6.543535385317019e-06, "loss": 0.4613, "step": 9974 }, { "epoch": 0.418138184714377, "grad_norm": 1.6076428890228271, "learning_rate": 6.542889675871862e-06, "loss": 0.4738, "step": 9975 }, { "epoch": 0.418180103329386, "grad_norm": 1.629765272140503, "learning_rate": 6.542243937985665e-06, "loss": 0.4562, "step": 9976 }, { "epoch": 0.41822202194439495, "grad_norm": 1.6226165294647217, "learning_rate": 6.541598171670335e-06, "loss": 0.4975, "step": 9977 }, { "epoch": 0.4182639405594039, "grad_norm": 1.6565971374511719, "learning_rate": 6.540952376937773e-06, "loss": 0.5135, "step": 9978 }, { "epoch": 0.4183058591744129, "grad_norm": 1.8857003450393677, "learning_rate": 6.540306553799885e-06, "loss": 0.5259, "step": 9979 }, { "epoch": 0.41834777778942184, "grad_norm": 1.5262603759765625, "learning_rate": 6.539660702268572e-06, "loss": 0.4622, "step": 9980 }, { "epoch": 0.4183896964044308, "grad_norm": 1.9560215473175049, "learning_rate": 6.5390148223557455e-06, "loss": 0.5529, "step": 9981 }, { "epoch": 0.41843161501943976, "grad_norm": 2.5359699726104736, "learning_rate": 6.538368914073308e-06, "loss": 0.4795, "step": 9982 }, { "epoch": 0.4184735336344487, "grad_norm": 1.5482120513916016, "learning_rate": 6.537722977433167e-06, "loss": 0.4882, "step": 9983 }, { "epoch": 0.4185154522494577, "grad_norm": 1.6508090496063232, "learning_rate": 6.537077012447226e-06, "loss": 0.5383, "step": 9984 }, { "epoch": 0.41855737086446665, "grad_norm": 1.691001296043396, "learning_rate": 6.536431019127396e-06, "loss": 0.5273, "step": 9985 }, { "epoch": 0.4185992894794756, "grad_norm": 1.6942124366760254, "learning_rate": 6.535784997485587e-06, "loss": 0.551, "step": 9986 }, { "epoch": 0.4186412080944846, "grad_norm": 1.695273995399475, "learning_rate": 6.535138947533701e-06, "loss": 0.5424, "step": 9987 }, { "epoch": 0.41868312670949354, "grad_norm": 1.562483549118042, "learning_rate": 6.534492869283653e-06, "loss": 0.5127, "step": 9988 }, { "epoch": 0.4187250453245025, "grad_norm": 1.4887977838516235, "learning_rate": 6.53384676274735e-06, "loss": 0.4618, "step": 9989 }, { "epoch": 0.41876696393951146, "grad_norm": 1.6222095489501953, "learning_rate": 6.533200627936702e-06, "loss": 0.5042, "step": 9990 }, { "epoch": 0.4188088825545204, "grad_norm": 1.5525381565093994, "learning_rate": 6.532554464863621e-06, "loss": 0.4988, "step": 9991 }, { "epoch": 0.4188508011695294, "grad_norm": 2.107482671737671, "learning_rate": 6.531908273540017e-06, "loss": 0.5453, "step": 9992 }, { "epoch": 0.4188927197845383, "grad_norm": 1.83705472946167, "learning_rate": 6.5312620539778e-06, "loss": 0.5298, "step": 9993 }, { "epoch": 0.41893463839954725, "grad_norm": 1.7944470643997192, "learning_rate": 6.530615806188885e-06, "loss": 0.5374, "step": 9994 }, { "epoch": 0.4189765570145562, "grad_norm": 1.6825140714645386, "learning_rate": 6.5299695301851855e-06, "loss": 0.5528, "step": 9995 }, { "epoch": 0.4190184756295652, "grad_norm": 1.7467738389968872, "learning_rate": 6.5293232259786095e-06, "loss": 0.5111, "step": 9996 }, { "epoch": 0.41906039424457414, "grad_norm": 1.772751808166504, "learning_rate": 6.528676893581076e-06, "loss": 0.5083, "step": 9997 }, { "epoch": 0.4191023128595831, "grad_norm": 1.6217939853668213, "learning_rate": 6.528030533004497e-06, "loss": 0.5087, "step": 9998 }, { "epoch": 0.41914423147459207, "grad_norm": 2.085602283477783, "learning_rate": 6.5273841442607875e-06, "loss": 0.5534, "step": 9999 }, { "epoch": 0.41918615008960103, "grad_norm": 2.41096830368042, "learning_rate": 6.526737727361863e-06, "loss": 0.556, "step": 10000 }, { "epoch": 0.41922806870461, "grad_norm": 1.759181022644043, "learning_rate": 6.52609128231964e-06, "loss": 0.4876, "step": 10001 }, { "epoch": 0.41926998731961895, "grad_norm": 1.622148871421814, "learning_rate": 6.525444809146031e-06, "loss": 0.5512, "step": 10002 }, { "epoch": 0.4193119059346279, "grad_norm": 1.6268603801727295, "learning_rate": 6.5247983078529585e-06, "loss": 0.5141, "step": 10003 }, { "epoch": 0.4193538245496369, "grad_norm": 1.7644435167312622, "learning_rate": 6.524151778452337e-06, "loss": 0.5177, "step": 10004 }, { "epoch": 0.41939574316464584, "grad_norm": 1.5342345237731934, "learning_rate": 6.5235052209560835e-06, "loss": 0.4708, "step": 10005 }, { "epoch": 0.4194376617796548, "grad_norm": 2.0108020305633545, "learning_rate": 6.522858635376118e-06, "loss": 0.5405, "step": 10006 }, { "epoch": 0.41947958039466376, "grad_norm": 1.647542119026184, "learning_rate": 6.522212021724359e-06, "loss": 0.4954, "step": 10007 }, { "epoch": 0.4195214990096727, "grad_norm": 1.7210866212844849, "learning_rate": 6.521565380012725e-06, "loss": 0.5478, "step": 10008 }, { "epoch": 0.4195634176246817, "grad_norm": 1.7115142345428467, "learning_rate": 6.520918710253138e-06, "loss": 0.5635, "step": 10009 }, { "epoch": 0.41960533623969065, "grad_norm": 1.6203409433364868, "learning_rate": 6.520272012457517e-06, "loss": 0.5061, "step": 10010 }, { "epoch": 0.4196472548546996, "grad_norm": 1.6285288333892822, "learning_rate": 6.519625286637781e-06, "loss": 0.5208, "step": 10011 }, { "epoch": 0.4196891734697086, "grad_norm": 1.8714261054992676, "learning_rate": 6.5189785328058555e-06, "loss": 0.4843, "step": 10012 }, { "epoch": 0.41973109208471754, "grad_norm": 2.4248600006103516, "learning_rate": 6.518331750973661e-06, "loss": 0.6133, "step": 10013 }, { "epoch": 0.4197730106997265, "grad_norm": 1.652392029762268, "learning_rate": 6.51768494115312e-06, "loss": 0.5173, "step": 10014 }, { "epoch": 0.41981492931473546, "grad_norm": 1.5255389213562012, "learning_rate": 6.517038103356154e-06, "loss": 0.4533, "step": 10015 }, { "epoch": 0.4198568479297444, "grad_norm": 1.5894333124160767, "learning_rate": 6.516391237594688e-06, "loss": 0.5063, "step": 10016 }, { "epoch": 0.4198987665447534, "grad_norm": 1.9542194604873657, "learning_rate": 6.5157443438806455e-06, "loss": 0.4973, "step": 10017 }, { "epoch": 0.4199406851597623, "grad_norm": 1.6785752773284912, "learning_rate": 6.5150974222259535e-06, "loss": 0.4862, "step": 10018 }, { "epoch": 0.41998260377477126, "grad_norm": 1.7460309267044067, "learning_rate": 6.514450472642533e-06, "loss": 0.5195, "step": 10019 }, { "epoch": 0.4200245223897802, "grad_norm": 1.5965369939804077, "learning_rate": 6.5138034951423116e-06, "loss": 0.5511, "step": 10020 }, { "epoch": 0.4200664410047892, "grad_norm": 1.9124670028686523, "learning_rate": 6.513156489737216e-06, "loss": 0.5284, "step": 10021 }, { "epoch": 0.42010835961979814, "grad_norm": 1.549203634262085, "learning_rate": 6.512509456439173e-06, "loss": 0.4705, "step": 10022 }, { "epoch": 0.4201502782348071, "grad_norm": 1.4672646522521973, "learning_rate": 6.511862395260109e-06, "loss": 0.4926, "step": 10023 }, { "epoch": 0.42019219684981607, "grad_norm": 1.4355577230453491, "learning_rate": 6.511215306211952e-06, "loss": 0.4862, "step": 10024 }, { "epoch": 0.42023411546482503, "grad_norm": 1.645995020866394, "learning_rate": 6.51056818930663e-06, "loss": 0.5497, "step": 10025 }, { "epoch": 0.420276034079834, "grad_norm": 1.5349326133728027, "learning_rate": 6.509921044556071e-06, "loss": 0.4914, "step": 10026 }, { "epoch": 0.42031795269484296, "grad_norm": 1.8848589658737183, "learning_rate": 6.509273871972205e-06, "loss": 0.5521, "step": 10027 }, { "epoch": 0.4203598713098519, "grad_norm": 1.7094498872756958, "learning_rate": 6.508626671566962e-06, "loss": 0.5231, "step": 10028 }, { "epoch": 0.4204017899248609, "grad_norm": 1.6047693490982056, "learning_rate": 6.507979443352273e-06, "loss": 0.4968, "step": 10029 }, { "epoch": 0.42044370853986984, "grad_norm": 1.6660311222076416, "learning_rate": 6.5073321873400654e-06, "loss": 0.5108, "step": 10030 }, { "epoch": 0.4204856271548788, "grad_norm": 1.5364794731140137, "learning_rate": 6.506684903542274e-06, "loss": 0.5248, "step": 10031 }, { "epoch": 0.42052754576988777, "grad_norm": 1.5538978576660156, "learning_rate": 6.50603759197083e-06, "loss": 0.5024, "step": 10032 }, { "epoch": 0.42056946438489673, "grad_norm": 1.6006460189819336, "learning_rate": 6.505390252637664e-06, "loss": 0.5159, "step": 10033 }, { "epoch": 0.4206113829999057, "grad_norm": 1.9057587385177612, "learning_rate": 6.50474288555471e-06, "loss": 0.5334, "step": 10034 }, { "epoch": 0.42065330161491465, "grad_norm": 1.6530368328094482, "learning_rate": 6.504095490733901e-06, "loss": 0.5085, "step": 10035 }, { "epoch": 0.4206952202299236, "grad_norm": 1.6454780101776123, "learning_rate": 6.503448068187171e-06, "loss": 0.4996, "step": 10036 }, { "epoch": 0.4207371388449326, "grad_norm": 2.098057746887207, "learning_rate": 6.502800617926456e-06, "loss": 0.5504, "step": 10037 }, { "epoch": 0.42077905745994154, "grad_norm": 1.4466826915740967, "learning_rate": 6.502153139963687e-06, "loss": 0.4907, "step": 10038 }, { "epoch": 0.4208209760749505, "grad_norm": 1.8291789293289185, "learning_rate": 6.501505634310802e-06, "loss": 0.5306, "step": 10039 }, { "epoch": 0.42086289468995947, "grad_norm": 1.6874533891677856, "learning_rate": 6.500858100979737e-06, "loss": 0.521, "step": 10040 }, { "epoch": 0.42090481330496843, "grad_norm": 1.843862771987915, "learning_rate": 6.500210539982428e-06, "loss": 0.4566, "step": 10041 }, { "epoch": 0.4209467319199774, "grad_norm": 1.5097800493240356, "learning_rate": 6.499562951330811e-06, "loss": 0.5251, "step": 10042 }, { "epoch": 0.4209886505349863, "grad_norm": 2.2482872009277344, "learning_rate": 6.498915335036824e-06, "loss": 0.5101, "step": 10043 }, { "epoch": 0.42103056914999526, "grad_norm": 1.6686352491378784, "learning_rate": 6.498267691112405e-06, "loss": 0.518, "step": 10044 }, { "epoch": 0.4210724877650042, "grad_norm": 1.9256846904754639, "learning_rate": 6.497620019569493e-06, "loss": 0.511, "step": 10045 }, { "epoch": 0.4211144063800132, "grad_norm": 1.7070608139038086, "learning_rate": 6.496972320420026e-06, "loss": 0.5119, "step": 10046 }, { "epoch": 0.42115632499502215, "grad_norm": 1.7393368482589722, "learning_rate": 6.496324593675945e-06, "loss": 0.5601, "step": 10047 }, { "epoch": 0.4211982436100311, "grad_norm": 1.862626314163208, "learning_rate": 6.495676839349185e-06, "loss": 0.4477, "step": 10048 }, { "epoch": 0.42124016222504007, "grad_norm": 1.6692861318588257, "learning_rate": 6.495029057451694e-06, "loss": 0.472, "step": 10049 }, { "epoch": 0.42128208084004903, "grad_norm": 1.424641489982605, "learning_rate": 6.494381247995407e-06, "loss": 0.4674, "step": 10050 }, { "epoch": 0.421323999455058, "grad_norm": 1.5281442403793335, "learning_rate": 6.493733410992269e-06, "loss": 0.4812, "step": 10051 }, { "epoch": 0.42136591807006696, "grad_norm": 1.9366291761398315, "learning_rate": 6.493085546454221e-06, "loss": 0.5085, "step": 10052 }, { "epoch": 0.4214078366850759, "grad_norm": 1.8182724714279175, "learning_rate": 6.492437654393203e-06, "loss": 0.5767, "step": 10053 }, { "epoch": 0.4214497553000849, "grad_norm": 1.902180790901184, "learning_rate": 6.491789734821161e-06, "loss": 0.4811, "step": 10054 }, { "epoch": 0.42149167391509385, "grad_norm": 1.5932854413986206, "learning_rate": 6.491141787750037e-06, "loss": 0.48, "step": 10055 }, { "epoch": 0.4215335925301028, "grad_norm": 1.506941795349121, "learning_rate": 6.490493813191779e-06, "loss": 0.4705, "step": 10056 }, { "epoch": 0.42157551114511177, "grad_norm": 1.5755038261413574, "learning_rate": 6.489845811158324e-06, "loss": 0.5121, "step": 10057 }, { "epoch": 0.42161742976012073, "grad_norm": 1.6341732740402222, "learning_rate": 6.4891977816616225e-06, "loss": 0.5425, "step": 10058 }, { "epoch": 0.4216593483751297, "grad_norm": 1.426882266998291, "learning_rate": 6.488549724713619e-06, "loss": 0.4765, "step": 10059 }, { "epoch": 0.42170126699013866, "grad_norm": 2.320858955383301, "learning_rate": 6.487901640326259e-06, "loss": 0.5366, "step": 10060 }, { "epoch": 0.4217431856051476, "grad_norm": 1.5188742876052856, "learning_rate": 6.487253528511488e-06, "loss": 0.5227, "step": 10061 }, { "epoch": 0.4217851042201566, "grad_norm": 2.897516965866089, "learning_rate": 6.486605389281256e-06, "loss": 0.4946, "step": 10062 }, { "epoch": 0.42182702283516554, "grad_norm": 1.6282788515090942, "learning_rate": 6.485957222647509e-06, "loss": 0.5215, "step": 10063 }, { "epoch": 0.4218689414501745, "grad_norm": 1.6103442907333374, "learning_rate": 6.485309028622192e-06, "loss": 0.5341, "step": 10064 }, { "epoch": 0.42191086006518347, "grad_norm": 1.842398762702942, "learning_rate": 6.48466080721726e-06, "loss": 0.5212, "step": 10065 }, { "epoch": 0.42195277868019243, "grad_norm": 1.8095711469650269, "learning_rate": 6.484012558444657e-06, "loss": 0.5294, "step": 10066 }, { "epoch": 0.4219946972952014, "grad_norm": 2.05334210395813, "learning_rate": 6.483364282316333e-06, "loss": 0.4508, "step": 10067 }, { "epoch": 0.4220366159102103, "grad_norm": 2.455655574798584, "learning_rate": 6.482715978844241e-06, "loss": 0.5526, "step": 10068 }, { "epoch": 0.42207853452521926, "grad_norm": 2.7985403537750244, "learning_rate": 6.482067648040329e-06, "loss": 0.499, "step": 10069 }, { "epoch": 0.4221204531402282, "grad_norm": 1.8219075202941895, "learning_rate": 6.481419289916548e-06, "loss": 0.5454, "step": 10070 }, { "epoch": 0.4221623717552372, "grad_norm": 1.5885449647903442, "learning_rate": 6.480770904484852e-06, "loss": 0.5663, "step": 10071 }, { "epoch": 0.42220429037024615, "grad_norm": 1.6877409219741821, "learning_rate": 6.48012249175719e-06, "loss": 0.5127, "step": 10072 }, { "epoch": 0.4222462089852551, "grad_norm": 1.6975035667419434, "learning_rate": 6.479474051745516e-06, "loss": 0.5194, "step": 10073 }, { "epoch": 0.4222881276002641, "grad_norm": 1.537164330482483, "learning_rate": 6.478825584461784e-06, "loss": 0.5229, "step": 10074 }, { "epoch": 0.42233004621527304, "grad_norm": 1.491857886314392, "learning_rate": 6.478177089917948e-06, "loss": 0.5242, "step": 10075 }, { "epoch": 0.422371964830282, "grad_norm": 1.7155126333236694, "learning_rate": 6.477528568125959e-06, "loss": 0.5155, "step": 10076 }, { "epoch": 0.42241388344529096, "grad_norm": 1.8843250274658203, "learning_rate": 6.476880019097773e-06, "loss": 0.5207, "step": 10077 }, { "epoch": 0.4224558020602999, "grad_norm": 1.734394907951355, "learning_rate": 6.476231442845347e-06, "loss": 0.4761, "step": 10078 }, { "epoch": 0.4224977206753089, "grad_norm": 1.5581105947494507, "learning_rate": 6.475582839380635e-06, "loss": 0.5105, "step": 10079 }, { "epoch": 0.42253963929031785, "grad_norm": 1.8838822841644287, "learning_rate": 6.474934208715593e-06, "loss": 0.5803, "step": 10080 }, { "epoch": 0.4225815579053268, "grad_norm": 1.514931082725525, "learning_rate": 6.474285550862178e-06, "loss": 0.4777, "step": 10081 }, { "epoch": 0.4226234765203358, "grad_norm": 1.5408227443695068, "learning_rate": 6.473636865832347e-06, "loss": 0.5092, "step": 10082 }, { "epoch": 0.42266539513534473, "grad_norm": 1.7642621994018555, "learning_rate": 6.472988153638057e-06, "loss": 0.487, "step": 10083 }, { "epoch": 0.4227073137503537, "grad_norm": 1.7296431064605713, "learning_rate": 6.472339414291269e-06, "loss": 0.5145, "step": 10084 }, { "epoch": 0.42274923236536266, "grad_norm": 1.613874077796936, "learning_rate": 6.471690647803936e-06, "loss": 0.4741, "step": 10085 }, { "epoch": 0.4227911509803716, "grad_norm": 1.8098399639129639, "learning_rate": 6.471041854188022e-06, "loss": 0.4943, "step": 10086 }, { "epoch": 0.4228330695953806, "grad_norm": 1.719495177268982, "learning_rate": 6.4703930334554856e-06, "loss": 0.4865, "step": 10087 }, { "epoch": 0.42287498821038955, "grad_norm": 1.6281218528747559, "learning_rate": 6.4697441856182855e-06, "loss": 0.4849, "step": 10088 }, { "epoch": 0.4229169068253985, "grad_norm": 1.441997766494751, "learning_rate": 6.4690953106883834e-06, "loss": 0.5564, "step": 10089 }, { "epoch": 0.42295882544040747, "grad_norm": 1.5427883863449097, "learning_rate": 6.468446408677741e-06, "loss": 0.4795, "step": 10090 }, { "epoch": 0.42300074405541643, "grad_norm": 1.6107308864593506, "learning_rate": 6.467797479598318e-06, "loss": 0.4833, "step": 10091 }, { "epoch": 0.4230426626704254, "grad_norm": 1.7269090414047241, "learning_rate": 6.467148523462078e-06, "loss": 0.5314, "step": 10092 }, { "epoch": 0.4230845812854343, "grad_norm": 1.7526665925979614, "learning_rate": 6.466499540280984e-06, "loss": 0.5461, "step": 10093 }, { "epoch": 0.42312649990044326, "grad_norm": 1.804932951927185, "learning_rate": 6.465850530066997e-06, "loss": 0.5038, "step": 10094 }, { "epoch": 0.4231684185154522, "grad_norm": 1.973283052444458, "learning_rate": 6.465201492832083e-06, "loss": 0.5063, "step": 10095 }, { "epoch": 0.4232103371304612, "grad_norm": 1.8965975046157837, "learning_rate": 6.464552428588205e-06, "loss": 0.4925, "step": 10096 }, { "epoch": 0.42325225574547015, "grad_norm": 1.8811362981796265, "learning_rate": 6.463903337347328e-06, "loss": 0.5143, "step": 10097 }, { "epoch": 0.4232941743604791, "grad_norm": 1.7988722324371338, "learning_rate": 6.463254219121415e-06, "loss": 0.5639, "step": 10098 }, { "epoch": 0.4233360929754881, "grad_norm": 1.7582110166549683, "learning_rate": 6.462605073922436e-06, "loss": 0.5486, "step": 10099 }, { "epoch": 0.42337801159049704, "grad_norm": 1.9731062650680542, "learning_rate": 6.461955901762353e-06, "loss": 0.4996, "step": 10100 }, { "epoch": 0.423419930205506, "grad_norm": 1.8689370155334473, "learning_rate": 6.461306702653133e-06, "loss": 0.501, "step": 10101 }, { "epoch": 0.42346184882051496, "grad_norm": 1.8951194286346436, "learning_rate": 6.460657476606745e-06, "loss": 0.5357, "step": 10102 }, { "epoch": 0.4235037674355239, "grad_norm": 1.7861449718475342, "learning_rate": 6.460008223635157e-06, "loss": 0.5336, "step": 10103 }, { "epoch": 0.4235456860505329, "grad_norm": 1.5681958198547363, "learning_rate": 6.459358943750333e-06, "loss": 0.4994, "step": 10104 }, { "epoch": 0.42358760466554185, "grad_norm": 1.710913062095642, "learning_rate": 6.4587096369642465e-06, "loss": 0.4976, "step": 10105 }, { "epoch": 0.4236295232805508, "grad_norm": 1.7630388736724854, "learning_rate": 6.458060303288864e-06, "loss": 0.5248, "step": 10106 }, { "epoch": 0.4236714418955598, "grad_norm": 2.3344247341156006, "learning_rate": 6.457410942736157e-06, "loss": 0.5041, "step": 10107 }, { "epoch": 0.42371336051056874, "grad_norm": 1.8720306158065796, "learning_rate": 6.4567615553180915e-06, "loss": 0.5246, "step": 10108 }, { "epoch": 0.4237552791255777, "grad_norm": 2.849421739578247, "learning_rate": 6.456112141046642e-06, "loss": 0.4702, "step": 10109 }, { "epoch": 0.42379719774058666, "grad_norm": 3.0451037883758545, "learning_rate": 6.455462699933777e-06, "loss": 0.546, "step": 10110 }, { "epoch": 0.4238391163555956, "grad_norm": 1.6398295164108276, "learning_rate": 6.454813231991471e-06, "loss": 0.4474, "step": 10111 }, { "epoch": 0.4238810349706046, "grad_norm": 2.353294610977173, "learning_rate": 6.454163737231695e-06, "loss": 0.517, "step": 10112 }, { "epoch": 0.42392295358561355, "grad_norm": 2.127385139465332, "learning_rate": 6.453514215666417e-06, "loss": 0.544, "step": 10113 }, { "epoch": 0.4239648722006225, "grad_norm": 1.8059086799621582, "learning_rate": 6.4528646673076165e-06, "loss": 0.4818, "step": 10114 }, { "epoch": 0.4240067908156315, "grad_norm": 1.9583889245986938, "learning_rate": 6.452215092167264e-06, "loss": 0.4844, "step": 10115 }, { "epoch": 0.42404870943064044, "grad_norm": 1.6365474462509155, "learning_rate": 6.4515654902573355e-06, "loss": 0.5609, "step": 10116 }, { "epoch": 0.4240906280456494, "grad_norm": 1.9609960317611694, "learning_rate": 6.450915861589802e-06, "loss": 0.5208, "step": 10117 }, { "epoch": 0.4241325466606583, "grad_norm": 1.7819342613220215, "learning_rate": 6.450266206176641e-06, "loss": 0.5091, "step": 10118 }, { "epoch": 0.42417446527566727, "grad_norm": 4.818714618682861, "learning_rate": 6.449616524029827e-06, "loss": 0.4997, "step": 10119 }, { "epoch": 0.42421638389067623, "grad_norm": 1.8851566314697266, "learning_rate": 6.448966815161336e-06, "loss": 0.5852, "step": 10120 }, { "epoch": 0.4242583025056852, "grad_norm": 2.2820522785186768, "learning_rate": 6.448317079583146e-06, "loss": 0.5178, "step": 10121 }, { "epoch": 0.42430022112069415, "grad_norm": 1.7949894666671753, "learning_rate": 6.4476673173072324e-06, "loss": 0.5145, "step": 10122 }, { "epoch": 0.4243421397357031, "grad_norm": 1.69815194606781, "learning_rate": 6.447017528345573e-06, "loss": 0.524, "step": 10123 }, { "epoch": 0.4243840583507121, "grad_norm": 1.602746605873108, "learning_rate": 6.446367712710147e-06, "loss": 0.4922, "step": 10124 }, { "epoch": 0.42442597696572104, "grad_norm": 1.713999629020691, "learning_rate": 6.445717870412928e-06, "loss": 0.4855, "step": 10125 }, { "epoch": 0.42446789558073, "grad_norm": 1.7004190683364868, "learning_rate": 6.445068001465904e-06, "loss": 0.4871, "step": 10126 }, { "epoch": 0.42450981419573897, "grad_norm": 1.834908366203308, "learning_rate": 6.444418105881048e-06, "loss": 0.4773, "step": 10127 }, { "epoch": 0.42455173281074793, "grad_norm": 1.6012688875198364, "learning_rate": 6.443768183670338e-06, "loss": 0.5051, "step": 10128 }, { "epoch": 0.4245936514257569, "grad_norm": 2.0516231060028076, "learning_rate": 6.443118234845759e-06, "loss": 0.457, "step": 10129 }, { "epoch": 0.42463557004076585, "grad_norm": 1.8835186958312988, "learning_rate": 6.44246825941929e-06, "loss": 0.5383, "step": 10130 }, { "epoch": 0.4246774886557748, "grad_norm": 1.711086630821228, "learning_rate": 6.4418182574029145e-06, "loss": 0.4981, "step": 10131 }, { "epoch": 0.4247194072707838, "grad_norm": 1.7299245595932007, "learning_rate": 6.441168228808612e-06, "loss": 0.5074, "step": 10132 }, { "epoch": 0.42476132588579274, "grad_norm": 2.11004638671875, "learning_rate": 6.440518173648365e-06, "loss": 0.4918, "step": 10133 }, { "epoch": 0.4248032445008017, "grad_norm": 2.003499746322632, "learning_rate": 6.439868091934156e-06, "loss": 0.5623, "step": 10134 }, { "epoch": 0.42484516311581066, "grad_norm": 1.7182730436325073, "learning_rate": 6.439217983677971e-06, "loss": 0.5299, "step": 10135 }, { "epoch": 0.4248870817308196, "grad_norm": 1.8176847696304321, "learning_rate": 6.438567848891791e-06, "loss": 0.5352, "step": 10136 }, { "epoch": 0.4249290003458286, "grad_norm": 1.7903473377227783, "learning_rate": 6.437917687587603e-06, "loss": 0.4839, "step": 10137 }, { "epoch": 0.42497091896083755, "grad_norm": 1.542264699935913, "learning_rate": 6.437267499777388e-06, "loss": 0.4726, "step": 10138 }, { "epoch": 0.4250128375758465, "grad_norm": 1.7995706796646118, "learning_rate": 6.436617285473134e-06, "loss": 0.496, "step": 10139 }, { "epoch": 0.4250547561908555, "grad_norm": 2.3167872428894043, "learning_rate": 6.435967044686829e-06, "loss": 0.5271, "step": 10140 }, { "epoch": 0.42509667480586444, "grad_norm": 1.6787681579589844, "learning_rate": 6.435316777430455e-06, "loss": 0.5031, "step": 10141 }, { "epoch": 0.4251385934208734, "grad_norm": 1.8997703790664673, "learning_rate": 6.434666483716e-06, "loss": 0.4959, "step": 10142 }, { "epoch": 0.4251805120358823, "grad_norm": 1.9109914302825928, "learning_rate": 6.434016163555452e-06, "loss": 0.5189, "step": 10143 }, { "epoch": 0.42522243065089127, "grad_norm": 1.5806293487548828, "learning_rate": 6.4333658169607986e-06, "loss": 0.5234, "step": 10144 }, { "epoch": 0.42526434926590023, "grad_norm": 1.9247528314590454, "learning_rate": 6.432715443944029e-06, "loss": 0.5528, "step": 10145 }, { "epoch": 0.4253062678809092, "grad_norm": 1.9823464155197144, "learning_rate": 6.432065044517129e-06, "loss": 0.4957, "step": 10146 }, { "epoch": 0.42534818649591816, "grad_norm": 1.5802950859069824, "learning_rate": 6.43141461869209e-06, "loss": 0.4643, "step": 10147 }, { "epoch": 0.4253901051109271, "grad_norm": 1.9572081565856934, "learning_rate": 6.430764166480901e-06, "loss": 0.5086, "step": 10148 }, { "epoch": 0.4254320237259361, "grad_norm": 1.800438642501831, "learning_rate": 6.430113687895554e-06, "loss": 0.5366, "step": 10149 }, { "epoch": 0.42547394234094504, "grad_norm": 2.0167269706726074, "learning_rate": 6.429463182948038e-06, "loss": 0.5405, "step": 10150 }, { "epoch": 0.425515860955954, "grad_norm": 1.9029333591461182, "learning_rate": 6.428812651650345e-06, "loss": 0.5473, "step": 10151 }, { "epoch": 0.42555777957096297, "grad_norm": 1.9950865507125854, "learning_rate": 6.428162094014465e-06, "loss": 0.5314, "step": 10152 }, { "epoch": 0.42559969818597193, "grad_norm": 1.6242886781692505, "learning_rate": 6.42751151005239e-06, "loss": 0.4829, "step": 10153 }, { "epoch": 0.4256416168009809, "grad_norm": 1.8413118124008179, "learning_rate": 6.426860899776116e-06, "loss": 0.5193, "step": 10154 }, { "epoch": 0.42568353541598986, "grad_norm": 1.7168184518814087, "learning_rate": 6.426210263197634e-06, "loss": 0.4375, "step": 10155 }, { "epoch": 0.4257254540309988, "grad_norm": 1.8510847091674805, "learning_rate": 6.425559600328934e-06, "loss": 0.4889, "step": 10156 }, { "epoch": 0.4257673726460078, "grad_norm": 1.717976450920105, "learning_rate": 6.424908911182017e-06, "loss": 0.5795, "step": 10157 }, { "epoch": 0.42580929126101674, "grad_norm": 1.9059927463531494, "learning_rate": 6.4242581957688735e-06, "loss": 0.5843, "step": 10158 }, { "epoch": 0.4258512098760257, "grad_norm": 1.7515031099319458, "learning_rate": 6.423607454101499e-06, "loss": 0.4812, "step": 10159 }, { "epoch": 0.42589312849103467, "grad_norm": 1.8642687797546387, "learning_rate": 6.422956686191889e-06, "loss": 0.5249, "step": 10160 }, { "epoch": 0.42593504710604363, "grad_norm": 1.543716549873352, "learning_rate": 6.42230589205204e-06, "loss": 0.4993, "step": 10161 }, { "epoch": 0.4259769657210526, "grad_norm": 2.0619699954986572, "learning_rate": 6.421655071693946e-06, "loss": 0.4787, "step": 10162 }, { "epoch": 0.42601888433606155, "grad_norm": 1.8558746576309204, "learning_rate": 6.421004225129609e-06, "loss": 0.4921, "step": 10163 }, { "epoch": 0.4260608029510705, "grad_norm": 1.7489993572235107, "learning_rate": 6.420353352371023e-06, "loss": 0.4936, "step": 10164 }, { "epoch": 0.4261027215660795, "grad_norm": 1.9275211095809937, "learning_rate": 6.419702453430184e-06, "loss": 0.5342, "step": 10165 }, { "epoch": 0.42614464018108844, "grad_norm": 1.6608986854553223, "learning_rate": 6.4190515283190955e-06, "loss": 0.5441, "step": 10166 }, { "epoch": 0.4261865587960974, "grad_norm": 1.993091344833374, "learning_rate": 6.418400577049754e-06, "loss": 0.4825, "step": 10167 }, { "epoch": 0.4262284774111063, "grad_norm": 1.7693977355957031, "learning_rate": 6.417749599634158e-06, "loss": 0.5637, "step": 10168 }, { "epoch": 0.4262703960261153, "grad_norm": 2.395324945449829, "learning_rate": 6.417098596084308e-06, "loss": 0.5381, "step": 10169 }, { "epoch": 0.42631231464112423, "grad_norm": 1.7512820959091187, "learning_rate": 6.416447566412205e-06, "loss": 0.5014, "step": 10170 }, { "epoch": 0.4263542332561332, "grad_norm": 1.5206886529922485, "learning_rate": 6.415796510629849e-06, "loss": 0.4327, "step": 10171 }, { "epoch": 0.42639615187114216, "grad_norm": 1.8172190189361572, "learning_rate": 6.415145428749241e-06, "loss": 0.5548, "step": 10172 }, { "epoch": 0.4264380704861511, "grad_norm": 1.65534245967865, "learning_rate": 6.414494320782385e-06, "loss": 0.4884, "step": 10173 }, { "epoch": 0.4264799891011601, "grad_norm": 2.047417163848877, "learning_rate": 6.41384318674128e-06, "loss": 0.5088, "step": 10174 }, { "epoch": 0.42652190771616905, "grad_norm": 1.7915563583374023, "learning_rate": 6.41319202663793e-06, "loss": 0.5161, "step": 10175 }, { "epoch": 0.426563826331178, "grad_norm": 1.750123381614685, "learning_rate": 6.41254084048434e-06, "loss": 0.514, "step": 10176 }, { "epoch": 0.42660574494618697, "grad_norm": 2.1553564071655273, "learning_rate": 6.411889628292512e-06, "loss": 0.5042, "step": 10177 }, { "epoch": 0.42664766356119593, "grad_norm": 1.7856793403625488, "learning_rate": 6.411238390074451e-06, "loss": 0.5087, "step": 10178 }, { "epoch": 0.4266895821762049, "grad_norm": 1.9570858478546143, "learning_rate": 6.410587125842161e-06, "loss": 0.537, "step": 10179 }, { "epoch": 0.42673150079121386, "grad_norm": 1.6871261596679688, "learning_rate": 6.409935835607648e-06, "loss": 0.4611, "step": 10180 }, { "epoch": 0.4267734194062228, "grad_norm": 1.899095058441162, "learning_rate": 6.409284519382915e-06, "loss": 0.5385, "step": 10181 }, { "epoch": 0.4268153380212318, "grad_norm": 1.7240307331085205, "learning_rate": 6.408633177179973e-06, "loss": 0.49, "step": 10182 }, { "epoch": 0.42685725663624074, "grad_norm": 1.8461711406707764, "learning_rate": 6.407981809010823e-06, "loss": 0.4814, "step": 10183 }, { "epoch": 0.4268991752512497, "grad_norm": 1.8780180215835571, "learning_rate": 6.407330414887476e-06, "loss": 0.4788, "step": 10184 }, { "epoch": 0.42694109386625867, "grad_norm": 1.8788508176803589, "learning_rate": 6.4066789948219386e-06, "loss": 0.5227, "step": 10185 }, { "epoch": 0.42698301248126763, "grad_norm": 2.1614043712615967, "learning_rate": 6.406027548826219e-06, "loss": 0.5287, "step": 10186 }, { "epoch": 0.4270249310962766, "grad_norm": 1.6593726873397827, "learning_rate": 6.4053760769123245e-06, "loss": 0.5092, "step": 10187 }, { "epoch": 0.42706684971128556, "grad_norm": 2.3001925945281982, "learning_rate": 6.404724579092265e-06, "loss": 0.5399, "step": 10188 }, { "epoch": 0.4271087683262945, "grad_norm": 1.7051936388015747, "learning_rate": 6.404073055378051e-06, "loss": 0.508, "step": 10189 }, { "epoch": 0.4271506869413035, "grad_norm": 1.6580936908721924, "learning_rate": 6.4034215057816885e-06, "loss": 0.5298, "step": 10190 }, { "epoch": 0.42719260555631244, "grad_norm": 1.6276843547821045, "learning_rate": 6.4027699303151936e-06, "loss": 0.4985, "step": 10191 }, { "epoch": 0.4272345241713214, "grad_norm": 1.9868510961532593, "learning_rate": 6.402118328990574e-06, "loss": 0.4844, "step": 10192 }, { "epoch": 0.4272764427863303, "grad_norm": 2.4959568977355957, "learning_rate": 6.40146670181984e-06, "loss": 0.5252, "step": 10193 }, { "epoch": 0.4273183614013393, "grad_norm": 1.6582084894180298, "learning_rate": 6.400815048815006e-06, "loss": 0.5412, "step": 10194 }, { "epoch": 0.42736028001634824, "grad_norm": 1.7038767337799072, "learning_rate": 6.4001633699880814e-06, "loss": 0.5357, "step": 10195 }, { "epoch": 0.4274021986313572, "grad_norm": 1.5169016122817993, "learning_rate": 6.399511665351082e-06, "loss": 0.4668, "step": 10196 }, { "epoch": 0.42744411724636616, "grad_norm": 1.9366580247879028, "learning_rate": 6.398859934916021e-06, "loss": 0.5397, "step": 10197 }, { "epoch": 0.4274860358613751, "grad_norm": 1.6881160736083984, "learning_rate": 6.398208178694909e-06, "loss": 0.4685, "step": 10198 }, { "epoch": 0.4275279544763841, "grad_norm": 1.7343531847000122, "learning_rate": 6.397556396699762e-06, "loss": 0.5295, "step": 10199 }, { "epoch": 0.42756987309139305, "grad_norm": 1.803816795349121, "learning_rate": 6.396904588942596e-06, "loss": 0.5501, "step": 10200 }, { "epoch": 0.427611791706402, "grad_norm": 1.7437236309051514, "learning_rate": 6.396252755435426e-06, "loss": 0.5159, "step": 10201 }, { "epoch": 0.427653710321411, "grad_norm": 1.7737112045288086, "learning_rate": 6.395600896190265e-06, "loss": 0.5548, "step": 10202 }, { "epoch": 0.42769562893641994, "grad_norm": 1.7253100872039795, "learning_rate": 6.394949011219131e-06, "loss": 0.4984, "step": 10203 }, { "epoch": 0.4277375475514289, "grad_norm": 1.7479199171066284, "learning_rate": 6.3942971005340416e-06, "loss": 0.5165, "step": 10204 }, { "epoch": 0.42777946616643786, "grad_norm": 1.8540642261505127, "learning_rate": 6.393645164147012e-06, "loss": 0.501, "step": 10205 }, { "epoch": 0.4278213847814468, "grad_norm": 1.88332200050354, "learning_rate": 6.39299320207006e-06, "loss": 0.5169, "step": 10206 }, { "epoch": 0.4278633033964558, "grad_norm": 1.739272117614746, "learning_rate": 6.392341214315205e-06, "loss": 0.4935, "step": 10207 }, { "epoch": 0.42790522201146475, "grad_norm": 1.7295247316360474, "learning_rate": 6.391689200894464e-06, "loss": 0.5504, "step": 10208 }, { "epoch": 0.4279471406264737, "grad_norm": 1.9377875328063965, "learning_rate": 6.391037161819855e-06, "loss": 0.5846, "step": 10209 }, { "epoch": 0.42798905924148267, "grad_norm": 2.402900457382202, "learning_rate": 6.390385097103402e-06, "loss": 0.5011, "step": 10210 }, { "epoch": 0.42803097785649163, "grad_norm": 1.8634916543960571, "learning_rate": 6.389733006757122e-06, "loss": 0.5079, "step": 10211 }, { "epoch": 0.4280728964715006, "grad_norm": 1.6015527248382568, "learning_rate": 6.389080890793032e-06, "loss": 0.5033, "step": 10212 }, { "epoch": 0.42811481508650956, "grad_norm": 1.6660054922103882, "learning_rate": 6.388428749223158e-06, "loss": 0.4754, "step": 10213 }, { "epoch": 0.4281567337015185, "grad_norm": 3.639486312866211, "learning_rate": 6.387776582059519e-06, "loss": 0.477, "step": 10214 }, { "epoch": 0.4281986523165275, "grad_norm": 1.9511399269104004, "learning_rate": 6.387124389314139e-06, "loss": 0.4684, "step": 10215 }, { "epoch": 0.42824057093153645, "grad_norm": 2.1416101455688477, "learning_rate": 6.386472170999037e-06, "loss": 0.548, "step": 10216 }, { "epoch": 0.4282824895465454, "grad_norm": 1.8800718784332275, "learning_rate": 6.385819927126237e-06, "loss": 0.5153, "step": 10217 }, { "epoch": 0.4283244081615543, "grad_norm": 2.1140940189361572, "learning_rate": 6.3851676577077614e-06, "loss": 0.5293, "step": 10218 }, { "epoch": 0.4283663267765633, "grad_norm": 1.7213271856307983, "learning_rate": 6.3845153627556365e-06, "loss": 0.559, "step": 10219 }, { "epoch": 0.42840824539157224, "grad_norm": 1.7892751693725586, "learning_rate": 6.383863042281886e-06, "loss": 0.4745, "step": 10220 }, { "epoch": 0.4284501640065812, "grad_norm": 2.0117626190185547, "learning_rate": 6.383210696298532e-06, "loss": 0.4854, "step": 10221 }, { "epoch": 0.42849208262159016, "grad_norm": 1.8860048055648804, "learning_rate": 6.382558324817602e-06, "loss": 0.5683, "step": 10222 }, { "epoch": 0.4285340012365991, "grad_norm": 1.9135841131210327, "learning_rate": 6.38190592785112e-06, "loss": 0.4834, "step": 10223 }, { "epoch": 0.4285759198516081, "grad_norm": 2.7435920238494873, "learning_rate": 6.381253505411111e-06, "loss": 0.5337, "step": 10224 }, { "epoch": 0.42861783846661705, "grad_norm": 2.4168848991394043, "learning_rate": 6.380601057509606e-06, "loss": 0.4782, "step": 10225 }, { "epoch": 0.428659757081626, "grad_norm": 1.638156533241272, "learning_rate": 6.379948584158629e-06, "loss": 0.4636, "step": 10226 }, { "epoch": 0.428701675696635, "grad_norm": 1.8047404289245605, "learning_rate": 6.379296085370204e-06, "loss": 0.5005, "step": 10227 }, { "epoch": 0.42874359431164394, "grad_norm": 1.8075634241104126, "learning_rate": 6.378643561156366e-06, "loss": 0.4906, "step": 10228 }, { "epoch": 0.4287855129266529, "grad_norm": 5.022607326507568, "learning_rate": 6.377991011529138e-06, "loss": 0.537, "step": 10229 }, { "epoch": 0.42882743154166186, "grad_norm": 1.7927502393722534, "learning_rate": 6.377338436500552e-06, "loss": 0.5013, "step": 10230 }, { "epoch": 0.4288693501566708, "grad_norm": 1.7864463329315186, "learning_rate": 6.376685836082636e-06, "loss": 0.5151, "step": 10231 }, { "epoch": 0.4289112687716798, "grad_norm": 1.9683536291122437, "learning_rate": 6.376033210287419e-06, "loss": 0.5339, "step": 10232 }, { "epoch": 0.42895318738668875, "grad_norm": 2.305124521255493, "learning_rate": 6.375380559126932e-06, "loss": 0.5239, "step": 10233 }, { "epoch": 0.4289951060016977, "grad_norm": 2.1162636280059814, "learning_rate": 6.374727882613207e-06, "loss": 0.5367, "step": 10234 }, { "epoch": 0.4290370246167067, "grad_norm": 1.846762776374817, "learning_rate": 6.374075180758273e-06, "loss": 0.5003, "step": 10235 }, { "epoch": 0.42907894323171564, "grad_norm": 1.6876311302185059, "learning_rate": 6.373422453574162e-06, "loss": 0.4876, "step": 10236 }, { "epoch": 0.4291208618467246, "grad_norm": 1.5826759338378906, "learning_rate": 6.372769701072907e-06, "loss": 0.444, "step": 10237 }, { "epoch": 0.42916278046173356, "grad_norm": 1.7903141975402832, "learning_rate": 6.372116923266543e-06, "loss": 0.4828, "step": 10238 }, { "epoch": 0.4292046990767425, "grad_norm": 1.7070388793945312, "learning_rate": 6.371464120167098e-06, "loss": 0.4886, "step": 10239 }, { "epoch": 0.4292466176917515, "grad_norm": 2.0098509788513184, "learning_rate": 6.370811291786609e-06, "loss": 0.5191, "step": 10240 }, { "epoch": 0.42928853630676045, "grad_norm": 1.9130929708480835, "learning_rate": 6.370158438137107e-06, "loss": 0.5037, "step": 10241 }, { "epoch": 0.4293304549217694, "grad_norm": 1.8484902381896973, "learning_rate": 6.3695055592306286e-06, "loss": 0.4994, "step": 10242 }, { "epoch": 0.4293723735367783, "grad_norm": 1.7616145610809326, "learning_rate": 6.368852655079212e-06, "loss": 0.5133, "step": 10243 }, { "epoch": 0.4294142921517873, "grad_norm": 1.6272668838500977, "learning_rate": 6.3681997256948855e-06, "loss": 0.4643, "step": 10244 }, { "epoch": 0.42945621076679624, "grad_norm": 1.9127177000045776, "learning_rate": 6.36754677108969e-06, "loss": 0.4746, "step": 10245 }, { "epoch": 0.4294981293818052, "grad_norm": 1.5136303901672363, "learning_rate": 6.366893791275659e-06, "loss": 0.4499, "step": 10246 }, { "epoch": 0.42954004799681417, "grad_norm": 1.730559229850769, "learning_rate": 6.366240786264832e-06, "loss": 0.5408, "step": 10247 }, { "epoch": 0.42958196661182313, "grad_norm": 1.5952082872390747, "learning_rate": 6.3655877560692455e-06, "loss": 0.4194, "step": 10248 }, { "epoch": 0.4296238852268321, "grad_norm": 1.9868733882904053, "learning_rate": 6.364934700700936e-06, "loss": 0.4975, "step": 10249 }, { "epoch": 0.42966580384184105, "grad_norm": 1.8944942951202393, "learning_rate": 6.364281620171942e-06, "loss": 0.5931, "step": 10250 }, { "epoch": 0.42970772245685, "grad_norm": 1.746077537536621, "learning_rate": 6.363628514494302e-06, "loss": 0.5281, "step": 10251 }, { "epoch": 0.429749641071859, "grad_norm": 1.6984384059906006, "learning_rate": 6.362975383680056e-06, "loss": 0.5572, "step": 10252 }, { "epoch": 0.42979155968686794, "grad_norm": 1.7822961807250977, "learning_rate": 6.362322227741242e-06, "loss": 0.5289, "step": 10253 }, { "epoch": 0.4298334783018769, "grad_norm": 1.932199239730835, "learning_rate": 6.3616690466899026e-06, "loss": 0.5482, "step": 10254 }, { "epoch": 0.42987539691688587, "grad_norm": 2.1138999462127686, "learning_rate": 6.361015840538075e-06, "loss": 0.5239, "step": 10255 }, { "epoch": 0.42991731553189483, "grad_norm": 1.9713298082351685, "learning_rate": 6.3603626092978036e-06, "loss": 0.48, "step": 10256 }, { "epoch": 0.4299592341469038, "grad_norm": 1.615697979927063, "learning_rate": 6.359709352981127e-06, "loss": 0.5768, "step": 10257 }, { "epoch": 0.43000115276191275, "grad_norm": 1.7548866271972656, "learning_rate": 6.359056071600089e-06, "loss": 0.4655, "step": 10258 }, { "epoch": 0.4300430713769217, "grad_norm": 1.3925766944885254, "learning_rate": 6.358402765166731e-06, "loss": 0.4799, "step": 10259 }, { "epoch": 0.4300849899919307, "grad_norm": 1.494431734085083, "learning_rate": 6.357749433693097e-06, "loss": 0.5081, "step": 10260 }, { "epoch": 0.43012690860693964, "grad_norm": 1.9453619718551636, "learning_rate": 6.357096077191227e-06, "loss": 0.5482, "step": 10261 }, { "epoch": 0.4301688272219486, "grad_norm": 1.8397886753082275, "learning_rate": 6.356442695673169e-06, "loss": 0.5321, "step": 10262 }, { "epoch": 0.43021074583695756, "grad_norm": 2.374495029449463, "learning_rate": 6.355789289150964e-06, "loss": 0.5172, "step": 10263 }, { "epoch": 0.4302526644519665, "grad_norm": 1.6985989809036255, "learning_rate": 6.355135857636657e-06, "loss": 0.5067, "step": 10264 }, { "epoch": 0.4302945830669755, "grad_norm": 1.690671443939209, "learning_rate": 6.354482401142295e-06, "loss": 0.5355, "step": 10265 }, { "epoch": 0.43033650168198445, "grad_norm": 1.784326434135437, "learning_rate": 6.353828919679923e-06, "loss": 0.5221, "step": 10266 }, { "epoch": 0.4303784202969934, "grad_norm": 1.781441569328308, "learning_rate": 6.353175413261585e-06, "loss": 0.5365, "step": 10267 }, { "epoch": 0.4304203389120023, "grad_norm": 1.8630824089050293, "learning_rate": 6.35252188189933e-06, "loss": 0.5513, "step": 10268 }, { "epoch": 0.4304622575270113, "grad_norm": 1.6515692472457886, "learning_rate": 6.351868325605204e-06, "loss": 0.5455, "step": 10269 }, { "epoch": 0.43050417614202025, "grad_norm": 1.9025421142578125, "learning_rate": 6.351214744391253e-06, "loss": 0.5247, "step": 10270 }, { "epoch": 0.4305460947570292, "grad_norm": 2.026402235031128, "learning_rate": 6.3505611382695285e-06, "loss": 0.5178, "step": 10271 }, { "epoch": 0.43058801337203817, "grad_norm": 1.596390724182129, "learning_rate": 6.3499075072520755e-06, "loss": 0.5188, "step": 10272 }, { "epoch": 0.43062993198704713, "grad_norm": 1.532160997390747, "learning_rate": 6.349253851350943e-06, "loss": 0.4691, "step": 10273 }, { "epoch": 0.4306718506020561, "grad_norm": 1.7683734893798828, "learning_rate": 6.3486001705781805e-06, "loss": 0.5088, "step": 10274 }, { "epoch": 0.43071376921706506, "grad_norm": 1.8751375675201416, "learning_rate": 6.34794646494584e-06, "loss": 0.5707, "step": 10275 }, { "epoch": 0.430755687832074, "grad_norm": 1.8124585151672363, "learning_rate": 6.347292734465968e-06, "loss": 0.5258, "step": 10276 }, { "epoch": 0.430797606447083, "grad_norm": 1.62732994556427, "learning_rate": 6.346638979150619e-06, "loss": 0.4786, "step": 10277 }, { "epoch": 0.43083952506209194, "grad_norm": 1.5385661125183105, "learning_rate": 6.345985199011842e-06, "loss": 0.5404, "step": 10278 }, { "epoch": 0.4308814436771009, "grad_norm": 2.441532850265503, "learning_rate": 6.345331394061687e-06, "loss": 0.5164, "step": 10279 }, { "epoch": 0.43092336229210987, "grad_norm": 1.6495295763015747, "learning_rate": 6.3446775643122095e-06, "loss": 0.4944, "step": 10280 }, { "epoch": 0.43096528090711883, "grad_norm": 1.8210926055908203, "learning_rate": 6.344023709775459e-06, "loss": 0.5333, "step": 10281 }, { "epoch": 0.4310071995221278, "grad_norm": 2.6885313987731934, "learning_rate": 6.343369830463488e-06, "loss": 0.5183, "step": 10282 }, { "epoch": 0.43104911813713676, "grad_norm": 1.4863123893737793, "learning_rate": 6.342715926388354e-06, "loss": 0.5266, "step": 10283 }, { "epoch": 0.4310910367521457, "grad_norm": 1.9659429788589478, "learning_rate": 6.342061997562107e-06, "loss": 0.5646, "step": 10284 }, { "epoch": 0.4311329553671547, "grad_norm": 1.6683183908462524, "learning_rate": 6.3414080439968015e-06, "loss": 0.5419, "step": 10285 }, { "epoch": 0.43117487398216364, "grad_norm": 2.2374157905578613, "learning_rate": 6.340754065704495e-06, "loss": 0.5106, "step": 10286 }, { "epoch": 0.4312167925971726, "grad_norm": 2.1725313663482666, "learning_rate": 6.3401000626972395e-06, "loss": 0.5224, "step": 10287 }, { "epoch": 0.43125871121218157, "grad_norm": 1.6249315738677979, "learning_rate": 6.339446034987094e-06, "loss": 0.5037, "step": 10288 }, { "epoch": 0.43130062982719053, "grad_norm": 1.5637526512145996, "learning_rate": 6.338791982586109e-06, "loss": 0.4891, "step": 10289 }, { "epoch": 0.4313425484421995, "grad_norm": 2.0767507553100586, "learning_rate": 6.338137905506347e-06, "loss": 0.5347, "step": 10290 }, { "epoch": 0.43138446705720845, "grad_norm": 1.7199063301086426, "learning_rate": 6.337483803759863e-06, "loss": 0.4834, "step": 10291 }, { "epoch": 0.4314263856722174, "grad_norm": 1.6318994760513306, "learning_rate": 6.3368296773587125e-06, "loss": 0.5219, "step": 10292 }, { "epoch": 0.4314683042872263, "grad_norm": 1.8518608808517456, "learning_rate": 6.3361755263149564e-06, "loss": 0.5095, "step": 10293 }, { "epoch": 0.4315102229022353, "grad_norm": 2.0817179679870605, "learning_rate": 6.335521350640651e-06, "loss": 0.5275, "step": 10294 }, { "epoch": 0.43155214151724425, "grad_norm": 1.7847466468811035, "learning_rate": 6.334867150347857e-06, "loss": 0.5362, "step": 10295 }, { "epoch": 0.4315940601322532, "grad_norm": 7.7375078201293945, "learning_rate": 6.334212925448631e-06, "loss": 0.555, "step": 10296 }, { "epoch": 0.4316359787472622, "grad_norm": 1.5950860977172852, "learning_rate": 6.333558675955034e-06, "loss": 0.5123, "step": 10297 }, { "epoch": 0.43167789736227113, "grad_norm": 1.5480574369430542, "learning_rate": 6.332904401879126e-06, "loss": 0.4582, "step": 10298 }, { "epoch": 0.4317198159772801, "grad_norm": 1.5256630182266235, "learning_rate": 6.332250103232971e-06, "loss": 0.5192, "step": 10299 }, { "epoch": 0.43176173459228906, "grad_norm": 1.6405513286590576, "learning_rate": 6.331595780028625e-06, "loss": 0.5057, "step": 10300 }, { "epoch": 0.431803653207298, "grad_norm": 1.6387701034545898, "learning_rate": 6.3309414322781506e-06, "loss": 0.517, "step": 10301 }, { "epoch": 0.431845571822307, "grad_norm": 1.6759029626846313, "learning_rate": 6.330287059993612e-06, "loss": 0.4621, "step": 10302 }, { "epoch": 0.43188749043731595, "grad_norm": 1.6348339319229126, "learning_rate": 6.3296326631870705e-06, "loss": 0.5604, "step": 10303 }, { "epoch": 0.4319294090523249, "grad_norm": 2.1388137340545654, "learning_rate": 6.328978241870589e-06, "loss": 0.4974, "step": 10304 }, { "epoch": 0.43197132766733387, "grad_norm": 1.693396806716919, "learning_rate": 6.3283237960562315e-06, "loss": 0.4947, "step": 10305 }, { "epoch": 0.43201324628234283, "grad_norm": 1.5215226411819458, "learning_rate": 6.327669325756061e-06, "loss": 0.5162, "step": 10306 }, { "epoch": 0.4320551648973518, "grad_norm": 1.6275739669799805, "learning_rate": 6.327014830982141e-06, "loss": 0.5015, "step": 10307 }, { "epoch": 0.43209708351236076, "grad_norm": 2.158628463745117, "learning_rate": 6.326360311746537e-06, "loss": 0.5389, "step": 10308 }, { "epoch": 0.4321390021273697, "grad_norm": 1.6100283861160278, "learning_rate": 6.325705768061317e-06, "loss": 0.5296, "step": 10309 }, { "epoch": 0.4321809207423787, "grad_norm": 1.398552417755127, "learning_rate": 6.325051199938541e-06, "loss": 0.4386, "step": 10310 }, { "epoch": 0.43222283935738764, "grad_norm": 1.6849907636642456, "learning_rate": 6.324396607390278e-06, "loss": 0.5166, "step": 10311 }, { "epoch": 0.4322647579723966, "grad_norm": 1.9059035778045654, "learning_rate": 6.3237419904285955e-06, "loss": 0.535, "step": 10312 }, { "epoch": 0.43230667658740557, "grad_norm": 1.933207631111145, "learning_rate": 6.32308734906556e-06, "loss": 0.45, "step": 10313 }, { "epoch": 0.43234859520241453, "grad_norm": 1.5973167419433594, "learning_rate": 6.322432683313238e-06, "loss": 0.5541, "step": 10314 }, { "epoch": 0.4323905138174235, "grad_norm": 1.8112245798110962, "learning_rate": 6.321777993183697e-06, "loss": 0.5199, "step": 10315 }, { "epoch": 0.43243243243243246, "grad_norm": 1.721734881401062, "learning_rate": 6.321123278689006e-06, "loss": 0.4997, "step": 10316 }, { "epoch": 0.4324743510474414, "grad_norm": 1.7830774784088135, "learning_rate": 6.3204685398412335e-06, "loss": 0.4815, "step": 10317 }, { "epoch": 0.4325162696624503, "grad_norm": 1.8238229751586914, "learning_rate": 6.319813776652452e-06, "loss": 0.5836, "step": 10318 }, { "epoch": 0.4325581882774593, "grad_norm": 2.2655229568481445, "learning_rate": 6.319158989134724e-06, "loss": 0.5559, "step": 10319 }, { "epoch": 0.43260010689246825, "grad_norm": 1.7453278303146362, "learning_rate": 6.318504177300126e-06, "loss": 0.5312, "step": 10320 }, { "epoch": 0.4326420255074772, "grad_norm": 1.5508490800857544, "learning_rate": 6.3178493411607245e-06, "loss": 0.509, "step": 10321 }, { "epoch": 0.4326839441224862, "grad_norm": 1.629941463470459, "learning_rate": 6.317194480728595e-06, "loss": 0.5236, "step": 10322 }, { "epoch": 0.43272586273749514, "grad_norm": 2.230968713760376, "learning_rate": 6.316539596015804e-06, "loss": 0.5153, "step": 10323 }, { "epoch": 0.4327677813525041, "grad_norm": 2.980679512023926, "learning_rate": 6.315884687034426e-06, "loss": 0.5207, "step": 10324 }, { "epoch": 0.43280969996751306, "grad_norm": 1.5525468587875366, "learning_rate": 6.315229753796534e-06, "loss": 0.4824, "step": 10325 }, { "epoch": 0.432851618582522, "grad_norm": 1.8432044982910156, "learning_rate": 6.314574796314197e-06, "loss": 0.5702, "step": 10326 }, { "epoch": 0.432893537197531, "grad_norm": 1.4803926944732666, "learning_rate": 6.313919814599494e-06, "loss": 0.4851, "step": 10327 }, { "epoch": 0.43293545581253995, "grad_norm": 1.5983037948608398, "learning_rate": 6.313264808664494e-06, "loss": 0.4795, "step": 10328 }, { "epoch": 0.4329773744275489, "grad_norm": 1.6857441663742065, "learning_rate": 6.312609778521273e-06, "loss": 0.581, "step": 10329 }, { "epoch": 0.4330192930425579, "grad_norm": 1.520032286643982, "learning_rate": 6.311954724181906e-06, "loss": 0.5153, "step": 10330 }, { "epoch": 0.43306121165756684, "grad_norm": 1.8061867952346802, "learning_rate": 6.311299645658468e-06, "loss": 0.5245, "step": 10331 }, { "epoch": 0.4331031302725758, "grad_norm": 1.574611783027649, "learning_rate": 6.310644542963033e-06, "loss": 0.5006, "step": 10332 }, { "epoch": 0.43314504888758476, "grad_norm": 2.0161783695220947, "learning_rate": 6.309989416107677e-06, "loss": 0.5611, "step": 10333 }, { "epoch": 0.4331869675025937, "grad_norm": 1.5631239414215088, "learning_rate": 6.309334265104479e-06, "loss": 0.4385, "step": 10334 }, { "epoch": 0.4332288861176027, "grad_norm": 1.6742397546768188, "learning_rate": 6.30867908996551e-06, "loss": 0.5157, "step": 10335 }, { "epoch": 0.43327080473261165, "grad_norm": 1.8603171110153198, "learning_rate": 6.308023890702855e-06, "loss": 0.5527, "step": 10336 }, { "epoch": 0.4333127233476206, "grad_norm": 1.726454496383667, "learning_rate": 6.3073686673285885e-06, "loss": 0.5215, "step": 10337 }, { "epoch": 0.43335464196262957, "grad_norm": 1.6123623847961426, "learning_rate": 6.306713419854784e-06, "loss": 0.5259, "step": 10338 }, { "epoch": 0.43339656057763853, "grad_norm": 1.723670244216919, "learning_rate": 6.3060581482935264e-06, "loss": 0.539, "step": 10339 }, { "epoch": 0.4334384791926475, "grad_norm": 1.5551892518997192, "learning_rate": 6.305402852656893e-06, "loss": 0.5265, "step": 10340 }, { "epoch": 0.43348039780765646, "grad_norm": 1.5603983402252197, "learning_rate": 6.304747532956962e-06, "loss": 0.5389, "step": 10341 }, { "epoch": 0.4335223164226654, "grad_norm": 1.689666509628296, "learning_rate": 6.304092189205815e-06, "loss": 0.553, "step": 10342 }, { "epoch": 0.43356423503767433, "grad_norm": 1.6493700742721558, "learning_rate": 6.303436821415529e-06, "loss": 0.4512, "step": 10343 }, { "epoch": 0.4336061536526833, "grad_norm": 1.9456275701522827, "learning_rate": 6.3027814295981885e-06, "loss": 0.4963, "step": 10344 }, { "epoch": 0.43364807226769225, "grad_norm": 1.4714781045913696, "learning_rate": 6.302126013765873e-06, "loss": 0.5139, "step": 10345 }, { "epoch": 0.4336899908827012, "grad_norm": 1.6468267440795898, "learning_rate": 6.301470573930666e-06, "loss": 0.537, "step": 10346 }, { "epoch": 0.4337319094977102, "grad_norm": 1.5999484062194824, "learning_rate": 6.3008151101046465e-06, "loss": 0.5521, "step": 10347 }, { "epoch": 0.43377382811271914, "grad_norm": 1.6398578882217407, "learning_rate": 6.300159622299899e-06, "loss": 0.5021, "step": 10348 }, { "epoch": 0.4338157467277281, "grad_norm": 2.772515058517456, "learning_rate": 6.299504110528507e-06, "loss": 0.5037, "step": 10349 }, { "epoch": 0.43385766534273706, "grad_norm": 1.5081260204315186, "learning_rate": 6.298848574802554e-06, "loss": 0.4887, "step": 10350 }, { "epoch": 0.433899583957746, "grad_norm": 1.5138239860534668, "learning_rate": 6.298193015134122e-06, "loss": 0.5085, "step": 10351 }, { "epoch": 0.433941502572755, "grad_norm": 1.9166854619979858, "learning_rate": 6.297537431535296e-06, "loss": 0.5393, "step": 10352 }, { "epoch": 0.43398342118776395, "grad_norm": 1.8243988752365112, "learning_rate": 6.296881824018162e-06, "loss": 0.5195, "step": 10353 }, { "epoch": 0.4340253398027729, "grad_norm": 1.707634687423706, "learning_rate": 6.296226192594804e-06, "loss": 0.5833, "step": 10354 }, { "epoch": 0.4340672584177819, "grad_norm": 1.530806541442871, "learning_rate": 6.295570537277308e-06, "loss": 0.5045, "step": 10355 }, { "epoch": 0.43410917703279084, "grad_norm": 1.6660960912704468, "learning_rate": 6.294914858077761e-06, "loss": 0.5053, "step": 10356 }, { "epoch": 0.4341510956477998, "grad_norm": 1.9999700784683228, "learning_rate": 6.294259155008249e-06, "loss": 0.6238, "step": 10357 }, { "epoch": 0.43419301426280876, "grad_norm": 1.6427408456802368, "learning_rate": 6.293603428080858e-06, "loss": 0.5173, "step": 10358 }, { "epoch": 0.4342349328778177, "grad_norm": 1.56226646900177, "learning_rate": 6.292947677307676e-06, "loss": 0.5097, "step": 10359 }, { "epoch": 0.4342768514928267, "grad_norm": 1.559769868850708, "learning_rate": 6.292291902700791e-06, "loss": 0.4841, "step": 10360 }, { "epoch": 0.43431877010783565, "grad_norm": 1.7373725175857544, "learning_rate": 6.291636104272292e-06, "loss": 0.5251, "step": 10361 }, { "epoch": 0.4343606887228446, "grad_norm": 1.4779317378997803, "learning_rate": 6.2909802820342665e-06, "loss": 0.5047, "step": 10362 }, { "epoch": 0.4344026073378536, "grad_norm": 1.6360571384429932, "learning_rate": 6.290324435998804e-06, "loss": 0.5775, "step": 10363 }, { "epoch": 0.43444452595286254, "grad_norm": 1.6087802648544312, "learning_rate": 6.2896685661779944e-06, "loss": 0.4982, "step": 10364 }, { "epoch": 0.4344864445678715, "grad_norm": 1.5834463834762573, "learning_rate": 6.28901267258393e-06, "loss": 0.5312, "step": 10365 }, { "epoch": 0.43452836318288046, "grad_norm": 1.6626654863357544, "learning_rate": 6.288356755228695e-06, "loss": 0.4558, "step": 10366 }, { "epoch": 0.4345702817978894, "grad_norm": 2.136723518371582, "learning_rate": 6.2877008141243865e-06, "loss": 0.566, "step": 10367 }, { "epoch": 0.43461220041289833, "grad_norm": 1.591442584991455, "learning_rate": 6.287044849283094e-06, "loss": 0.5323, "step": 10368 }, { "epoch": 0.4346541190279073, "grad_norm": 1.656527042388916, "learning_rate": 6.286388860716908e-06, "loss": 0.5423, "step": 10369 }, { "epoch": 0.43469603764291626, "grad_norm": 1.6613831520080566, "learning_rate": 6.285732848437922e-06, "loss": 0.5028, "step": 10370 }, { "epoch": 0.4347379562579252, "grad_norm": 1.8061177730560303, "learning_rate": 6.285076812458228e-06, "loss": 0.5423, "step": 10371 }, { "epoch": 0.4347798748729342, "grad_norm": 1.5587095022201538, "learning_rate": 6.28442075278992e-06, "loss": 0.4925, "step": 10372 }, { "epoch": 0.43482179348794314, "grad_norm": 1.8756508827209473, "learning_rate": 6.283764669445092e-06, "loss": 0.5711, "step": 10373 }, { "epoch": 0.4348637121029521, "grad_norm": 2.5299174785614014, "learning_rate": 6.283108562435837e-06, "loss": 0.5266, "step": 10374 }, { "epoch": 0.43490563071796107, "grad_norm": 1.5183988809585571, "learning_rate": 6.282452431774248e-06, "loss": 0.5368, "step": 10375 }, { "epoch": 0.43494754933297003, "grad_norm": 1.623874306678772, "learning_rate": 6.281796277472422e-06, "loss": 0.4966, "step": 10376 }, { "epoch": 0.434989467947979, "grad_norm": 1.6499478816986084, "learning_rate": 6.281140099542453e-06, "loss": 0.5233, "step": 10377 }, { "epoch": 0.43503138656298795, "grad_norm": 1.8725498914718628, "learning_rate": 6.280483897996439e-06, "loss": 0.529, "step": 10378 }, { "epoch": 0.4350733051779969, "grad_norm": 1.485474944114685, "learning_rate": 6.279827672846474e-06, "loss": 0.5281, "step": 10379 }, { "epoch": 0.4351152237930059, "grad_norm": 1.6593074798583984, "learning_rate": 6.279171424104656e-06, "loss": 0.534, "step": 10380 }, { "epoch": 0.43515714240801484, "grad_norm": 1.579677939414978, "learning_rate": 6.278515151783077e-06, "loss": 0.5054, "step": 10381 }, { "epoch": 0.4351990610230238, "grad_norm": 1.711161494255066, "learning_rate": 6.277858855893843e-06, "loss": 0.4423, "step": 10382 }, { "epoch": 0.43524097963803277, "grad_norm": 1.7086833715438843, "learning_rate": 6.277202536449047e-06, "loss": 0.5374, "step": 10383 }, { "epoch": 0.43528289825304173, "grad_norm": 1.6840351819992065, "learning_rate": 6.276546193460786e-06, "loss": 0.5019, "step": 10384 }, { "epoch": 0.4353248168680507, "grad_norm": 1.7536579370498657, "learning_rate": 6.2758898269411625e-06, "loss": 0.4727, "step": 10385 }, { "epoch": 0.43536673548305965, "grad_norm": 1.8960355520248413, "learning_rate": 6.275233436902274e-06, "loss": 0.525, "step": 10386 }, { "epoch": 0.4354086540980686, "grad_norm": 2.019113779067993, "learning_rate": 6.274577023356218e-06, "loss": 0.4717, "step": 10387 }, { "epoch": 0.4354505727130776, "grad_norm": 1.6083312034606934, "learning_rate": 6.273920586315099e-06, "loss": 0.4703, "step": 10388 }, { "epoch": 0.43549249132808654, "grad_norm": 1.7801971435546875, "learning_rate": 6.273264125791014e-06, "loss": 0.4832, "step": 10389 }, { "epoch": 0.4355344099430955, "grad_norm": 1.4176288843154907, "learning_rate": 6.272607641796064e-06, "loss": 0.4463, "step": 10390 }, { "epoch": 0.43557632855810446, "grad_norm": 2.3011727333068848, "learning_rate": 6.271951134342353e-06, "loss": 0.493, "step": 10391 }, { "epoch": 0.4356182471731134, "grad_norm": 1.844114899635315, "learning_rate": 6.271294603441982e-06, "loss": 0.5735, "step": 10392 }, { "epoch": 0.43566016578812233, "grad_norm": 1.5128906965255737, "learning_rate": 6.270638049107051e-06, "loss": 0.4767, "step": 10393 }, { "epoch": 0.4357020844031313, "grad_norm": 1.7924606800079346, "learning_rate": 6.269981471349666e-06, "loss": 0.5414, "step": 10394 }, { "epoch": 0.43574400301814026, "grad_norm": 7.297540187835693, "learning_rate": 6.269324870181925e-06, "loss": 0.5195, "step": 10395 }, { "epoch": 0.4357859216331492, "grad_norm": 1.7986308336257935, "learning_rate": 6.2686682456159375e-06, "loss": 0.5267, "step": 10396 }, { "epoch": 0.4358278402481582, "grad_norm": 1.6870579719543457, "learning_rate": 6.268011597663805e-06, "loss": 0.5216, "step": 10397 }, { "epoch": 0.43586975886316714, "grad_norm": 1.645740270614624, "learning_rate": 6.2673549263376295e-06, "loss": 0.4976, "step": 10398 }, { "epoch": 0.4359116774781761, "grad_norm": 1.5905781984329224, "learning_rate": 6.26669823164952e-06, "loss": 0.4863, "step": 10399 }, { "epoch": 0.43595359609318507, "grad_norm": 1.6766194105148315, "learning_rate": 6.266041513611578e-06, "loss": 0.5622, "step": 10400 }, { "epoch": 0.43599551470819403, "grad_norm": 1.788933515548706, "learning_rate": 6.265384772235912e-06, "loss": 0.5157, "step": 10401 }, { "epoch": 0.436037433323203, "grad_norm": 1.632672905921936, "learning_rate": 6.264728007534626e-06, "loss": 0.5363, "step": 10402 }, { "epoch": 0.43607935193821196, "grad_norm": 1.948085069656372, "learning_rate": 6.26407121951983e-06, "loss": 0.5259, "step": 10403 }, { "epoch": 0.4361212705532209, "grad_norm": 1.6226226091384888, "learning_rate": 6.2634144082036265e-06, "loss": 0.4738, "step": 10404 }, { "epoch": 0.4361631891682299, "grad_norm": 1.5955523252487183, "learning_rate": 6.262757573598126e-06, "loss": 0.5065, "step": 10405 }, { "epoch": 0.43620510778323884, "grad_norm": 1.64938485622406, "learning_rate": 6.2621007157154336e-06, "loss": 0.5893, "step": 10406 }, { "epoch": 0.4362470263982478, "grad_norm": 1.960647463798523, "learning_rate": 6.261443834567662e-06, "loss": 0.5744, "step": 10407 }, { "epoch": 0.43628894501325677, "grad_norm": 1.4675683975219727, "learning_rate": 6.2607869301669155e-06, "loss": 0.469, "step": 10408 }, { "epoch": 0.43633086362826573, "grad_norm": 1.6623249053955078, "learning_rate": 6.260130002525304e-06, "loss": 0.4569, "step": 10409 }, { "epoch": 0.4363727822432747, "grad_norm": 1.6578572988510132, "learning_rate": 6.259473051654941e-06, "loss": 0.5403, "step": 10410 }, { "epoch": 0.43641470085828366, "grad_norm": 1.671590805053711, "learning_rate": 6.25881607756793e-06, "loss": 0.5628, "step": 10411 }, { "epoch": 0.4364566194732926, "grad_norm": 1.9696029424667358, "learning_rate": 6.258159080276388e-06, "loss": 0.4907, "step": 10412 }, { "epoch": 0.4364985380883016, "grad_norm": 1.6499228477478027, "learning_rate": 6.257502059792422e-06, "loss": 0.4766, "step": 10413 }, { "epoch": 0.43654045670331054, "grad_norm": 1.6723335981369019, "learning_rate": 6.256845016128142e-06, "loss": 0.4929, "step": 10414 }, { "epoch": 0.4365823753183195, "grad_norm": 2.2085111141204834, "learning_rate": 6.2561879492956625e-06, "loss": 0.5442, "step": 10415 }, { "epoch": 0.43662429393332847, "grad_norm": 1.6674878597259521, "learning_rate": 6.255530859307096e-06, "loss": 0.5116, "step": 10416 }, { "epoch": 0.43666621254833743, "grad_norm": 1.675188660621643, "learning_rate": 6.2548737461745535e-06, "loss": 0.5015, "step": 10417 }, { "epoch": 0.43670813116334634, "grad_norm": 1.5225204229354858, "learning_rate": 6.254216609910148e-06, "loss": 0.5012, "step": 10418 }, { "epoch": 0.4367500497783553, "grad_norm": 1.5045279264450073, "learning_rate": 6.253559450525992e-06, "loss": 0.5047, "step": 10419 }, { "epoch": 0.43679196839336426, "grad_norm": 1.6862339973449707, "learning_rate": 6.252902268034202e-06, "loss": 0.5126, "step": 10420 }, { "epoch": 0.4368338870083732, "grad_norm": 1.692624807357788, "learning_rate": 6.2522450624468914e-06, "loss": 0.5158, "step": 10421 }, { "epoch": 0.4368758056233822, "grad_norm": 1.9355849027633667, "learning_rate": 6.2515878337761725e-06, "loss": 0.4946, "step": 10422 }, { "epoch": 0.43691772423839115, "grad_norm": 1.533266305923462, "learning_rate": 6.2509305820341625e-06, "loss": 0.5429, "step": 10423 }, { "epoch": 0.4369596428534001, "grad_norm": 2.100846290588379, "learning_rate": 6.250273307232977e-06, "loss": 0.4756, "step": 10424 }, { "epoch": 0.43700156146840907, "grad_norm": 5.324635028839111, "learning_rate": 6.249616009384731e-06, "loss": 0.5075, "step": 10425 }, { "epoch": 0.43704348008341803, "grad_norm": 1.7652312517166138, "learning_rate": 6.248958688501543e-06, "loss": 0.4979, "step": 10426 }, { "epoch": 0.437085398698427, "grad_norm": 2.0672407150268555, "learning_rate": 6.248301344595527e-06, "loss": 0.501, "step": 10427 }, { "epoch": 0.43712731731343596, "grad_norm": 1.7034614086151123, "learning_rate": 6.247643977678801e-06, "loss": 0.4702, "step": 10428 }, { "epoch": 0.4371692359284449, "grad_norm": 1.6468340158462524, "learning_rate": 6.2469865877634835e-06, "loss": 0.5693, "step": 10429 }, { "epoch": 0.4372111545434539, "grad_norm": 1.5467209815979004, "learning_rate": 6.2463291748616925e-06, "loss": 0.4868, "step": 10430 }, { "epoch": 0.43725307315846285, "grad_norm": 1.65938138961792, "learning_rate": 6.245671738985546e-06, "loss": 0.5224, "step": 10431 }, { "epoch": 0.4372949917734718, "grad_norm": 1.835119366645813, "learning_rate": 6.245014280147164e-06, "loss": 0.4696, "step": 10432 }, { "epoch": 0.43733691038848077, "grad_norm": 1.8227031230926514, "learning_rate": 6.244356798358663e-06, "loss": 0.5483, "step": 10433 }, { "epoch": 0.43737882900348973, "grad_norm": 1.852766752243042, "learning_rate": 6.243699293632165e-06, "loss": 0.517, "step": 10434 }, { "epoch": 0.4374207476184987, "grad_norm": 1.8652637004852295, "learning_rate": 6.243041765979792e-06, "loss": 0.5277, "step": 10435 }, { "epoch": 0.43746266623350766, "grad_norm": 1.5160808563232422, "learning_rate": 6.24238421541366e-06, "loss": 0.5096, "step": 10436 }, { "epoch": 0.4375045848485166, "grad_norm": 1.9505329132080078, "learning_rate": 6.241726641945893e-06, "loss": 0.6023, "step": 10437 }, { "epoch": 0.4375465034635256, "grad_norm": 1.8096619844436646, "learning_rate": 6.241069045588612e-06, "loss": 0.5905, "step": 10438 }, { "epoch": 0.43758842207853454, "grad_norm": 1.7705987691879272, "learning_rate": 6.2404114263539405e-06, "loss": 0.5684, "step": 10439 }, { "epoch": 0.4376303406935435, "grad_norm": 1.5970841646194458, "learning_rate": 6.239753784253998e-06, "loss": 0.5533, "step": 10440 }, { "epoch": 0.43767225930855247, "grad_norm": 1.8142211437225342, "learning_rate": 6.239096119300909e-06, "loss": 0.5093, "step": 10441 }, { "epoch": 0.43771417792356143, "grad_norm": 1.769691824913025, "learning_rate": 6.2384384315067955e-06, "loss": 0.5346, "step": 10442 }, { "epoch": 0.43775609653857034, "grad_norm": 1.8261481523513794, "learning_rate": 6.237780720883781e-06, "loss": 0.5019, "step": 10443 }, { "epoch": 0.4377980151535793, "grad_norm": 1.7583435773849487, "learning_rate": 6.2371229874439906e-06, "loss": 0.5368, "step": 10444 }, { "epoch": 0.43783993376858826, "grad_norm": 1.885597586631775, "learning_rate": 6.236465231199549e-06, "loss": 0.5563, "step": 10445 }, { "epoch": 0.4378818523835972, "grad_norm": 1.7827249765396118, "learning_rate": 6.235807452162579e-06, "loss": 0.5554, "step": 10446 }, { "epoch": 0.4379237709986062, "grad_norm": 1.5738434791564941, "learning_rate": 6.235149650345208e-06, "loss": 0.5069, "step": 10447 }, { "epoch": 0.43796568961361515, "grad_norm": 2.1791090965270996, "learning_rate": 6.2344918257595595e-06, "loss": 0.4668, "step": 10448 }, { "epoch": 0.4380076082286241, "grad_norm": 1.5286270380020142, "learning_rate": 6.233833978417762e-06, "loss": 0.529, "step": 10449 }, { "epoch": 0.4380495268436331, "grad_norm": 1.8087483644485474, "learning_rate": 6.233176108331941e-06, "loss": 0.505, "step": 10450 }, { "epoch": 0.43809144545864204, "grad_norm": 1.863330364227295, "learning_rate": 6.232518215514222e-06, "loss": 0.4708, "step": 10451 }, { "epoch": 0.438133364073651, "grad_norm": 1.7794369459152222, "learning_rate": 6.2318602999767345e-06, "loss": 0.5593, "step": 10452 }, { "epoch": 0.43817528268865996, "grad_norm": 1.4608772993087769, "learning_rate": 6.231202361731605e-06, "loss": 0.4718, "step": 10453 }, { "epoch": 0.4382172013036689, "grad_norm": 1.6201348304748535, "learning_rate": 6.230544400790964e-06, "loss": 0.4635, "step": 10454 }, { "epoch": 0.4382591199186779, "grad_norm": 1.4488838911056519, "learning_rate": 6.2298864171669345e-06, "loss": 0.4715, "step": 10455 }, { "epoch": 0.43830103853368685, "grad_norm": 1.8194379806518555, "learning_rate": 6.229228410871652e-06, "loss": 0.5283, "step": 10456 }, { "epoch": 0.4383429571486958, "grad_norm": 1.7981010675430298, "learning_rate": 6.228570381917243e-06, "loss": 0.4761, "step": 10457 }, { "epoch": 0.4383848757637048, "grad_norm": 1.562738299369812, "learning_rate": 6.227912330315836e-06, "loss": 0.4618, "step": 10458 }, { "epoch": 0.43842679437871374, "grad_norm": 2.4452860355377197, "learning_rate": 6.227254256079563e-06, "loss": 0.561, "step": 10459 }, { "epoch": 0.4384687129937227, "grad_norm": 1.6768110990524292, "learning_rate": 6.226596159220555e-06, "loss": 0.528, "step": 10460 }, { "epoch": 0.43851063160873166, "grad_norm": 1.6813503503799438, "learning_rate": 6.225938039750942e-06, "loss": 0.5471, "step": 10461 }, { "epoch": 0.4385525502237406, "grad_norm": 1.702208161354065, "learning_rate": 6.2252798976828565e-06, "loss": 0.4785, "step": 10462 }, { "epoch": 0.4385944688387496, "grad_norm": 1.7051562070846558, "learning_rate": 6.224621733028431e-06, "loss": 0.5256, "step": 10463 }, { "epoch": 0.43863638745375855, "grad_norm": 1.5328810214996338, "learning_rate": 6.223963545799795e-06, "loss": 0.4925, "step": 10464 }, { "epoch": 0.4386783060687675, "grad_norm": 1.7072490453720093, "learning_rate": 6.223305336009084e-06, "loss": 0.4324, "step": 10465 }, { "epoch": 0.43872022468377647, "grad_norm": 1.7286320924758911, "learning_rate": 6.22264710366843e-06, "loss": 0.5007, "step": 10466 }, { "epoch": 0.43876214329878543, "grad_norm": 1.7812615633010864, "learning_rate": 6.221988848789967e-06, "loss": 0.5287, "step": 10467 }, { "epoch": 0.43880406191379434, "grad_norm": 1.422300934791565, "learning_rate": 6.221330571385829e-06, "loss": 0.4544, "step": 10468 }, { "epoch": 0.4388459805288033, "grad_norm": 1.661658525466919, "learning_rate": 6.22067227146815e-06, "loss": 0.5018, "step": 10469 }, { "epoch": 0.43888789914381227, "grad_norm": 1.778111219406128, "learning_rate": 6.220013949049066e-06, "loss": 0.5189, "step": 10470 }, { "epoch": 0.43892981775882123, "grad_norm": 1.6728721857070923, "learning_rate": 6.2193556041407095e-06, "loss": 0.5175, "step": 10471 }, { "epoch": 0.4389717363738302, "grad_norm": 1.7871283292770386, "learning_rate": 6.2186972367552205e-06, "loss": 0.4506, "step": 10472 }, { "epoch": 0.43901365498883915, "grad_norm": 1.855818510055542, "learning_rate": 6.218038846904733e-06, "loss": 0.532, "step": 10473 }, { "epoch": 0.4390555736038481, "grad_norm": 1.5770527124404907, "learning_rate": 6.2173804346013815e-06, "loss": 0.532, "step": 10474 }, { "epoch": 0.4390974922188571, "grad_norm": 1.7420332431793213, "learning_rate": 6.216721999857305e-06, "loss": 0.5145, "step": 10475 }, { "epoch": 0.43913941083386604, "grad_norm": 1.7697209119796753, "learning_rate": 6.21606354268464e-06, "loss": 0.5316, "step": 10476 }, { "epoch": 0.439181329448875, "grad_norm": 1.8040987253189087, "learning_rate": 6.215405063095527e-06, "loss": 0.4572, "step": 10477 }, { "epoch": 0.43922324806388396, "grad_norm": 1.5293595790863037, "learning_rate": 6.2147465611020994e-06, "loss": 0.5216, "step": 10478 }, { "epoch": 0.4392651666788929, "grad_norm": 1.6639539003372192, "learning_rate": 6.214088036716499e-06, "loss": 0.4883, "step": 10479 }, { "epoch": 0.4393070852939019, "grad_norm": 1.8678972721099854, "learning_rate": 6.213429489950864e-06, "loss": 0.5167, "step": 10480 }, { "epoch": 0.43934900390891085, "grad_norm": 1.4833611249923706, "learning_rate": 6.212770920817335e-06, "loss": 0.5123, "step": 10481 }, { "epoch": 0.4393909225239198, "grad_norm": 1.9751473665237427, "learning_rate": 6.21211232932805e-06, "loss": 0.5293, "step": 10482 }, { "epoch": 0.4394328411389288, "grad_norm": 8.658839225769043, "learning_rate": 6.21145371549515e-06, "loss": 0.5613, "step": 10483 }, { "epoch": 0.43947475975393774, "grad_norm": 1.7557674646377563, "learning_rate": 6.210795079330774e-06, "loss": 0.5341, "step": 10484 }, { "epoch": 0.4395166783689467, "grad_norm": 1.650598168373108, "learning_rate": 6.210136420847067e-06, "loss": 0.5323, "step": 10485 }, { "epoch": 0.43955859698395566, "grad_norm": 1.518912434577942, "learning_rate": 6.2094777400561665e-06, "loss": 0.5301, "step": 10486 }, { "epoch": 0.4396005155989646, "grad_norm": 1.4493452310562134, "learning_rate": 6.2088190369702176e-06, "loss": 0.4738, "step": 10487 }, { "epoch": 0.4396424342139736, "grad_norm": 1.780374526977539, "learning_rate": 6.2081603116013585e-06, "loss": 0.5821, "step": 10488 }, { "epoch": 0.43968435282898255, "grad_norm": 1.825757384300232, "learning_rate": 6.207501563961735e-06, "loss": 0.5574, "step": 10489 }, { "epoch": 0.4397262714439915, "grad_norm": 1.8575971126556396, "learning_rate": 6.20684279406349e-06, "loss": 0.5383, "step": 10490 }, { "epoch": 0.4397681900590005, "grad_norm": 1.5951849222183228, "learning_rate": 6.206184001918768e-06, "loss": 0.5225, "step": 10491 }, { "epoch": 0.43981010867400944, "grad_norm": 1.647645354270935, "learning_rate": 6.205525187539709e-06, "loss": 0.5434, "step": 10492 }, { "epoch": 0.43985202728901834, "grad_norm": 1.9371494054794312, "learning_rate": 6.204866350938461e-06, "loss": 0.5689, "step": 10493 }, { "epoch": 0.4398939459040273, "grad_norm": 1.4299849271774292, "learning_rate": 6.204207492127167e-06, "loss": 0.4787, "step": 10494 }, { "epoch": 0.43993586451903627, "grad_norm": 1.5260424613952637, "learning_rate": 6.203548611117973e-06, "loss": 0.4918, "step": 10495 }, { "epoch": 0.43997778313404523, "grad_norm": 1.9062552452087402, "learning_rate": 6.202889707923024e-06, "loss": 0.5198, "step": 10496 }, { "epoch": 0.4400197017490542, "grad_norm": 1.618754506111145, "learning_rate": 6.202230782554466e-06, "loss": 0.4996, "step": 10497 }, { "epoch": 0.44006162036406316, "grad_norm": 1.6432842016220093, "learning_rate": 6.201571835024445e-06, "loss": 0.5055, "step": 10498 }, { "epoch": 0.4401035389790721, "grad_norm": 1.8078398704528809, "learning_rate": 6.200912865345109e-06, "loss": 0.5082, "step": 10499 }, { "epoch": 0.4401454575940811, "grad_norm": 1.5090997219085693, "learning_rate": 6.200253873528606e-06, "loss": 0.4675, "step": 10500 }, { "epoch": 0.44018737620909004, "grad_norm": 1.8138121366500854, "learning_rate": 6.19959485958708e-06, "loss": 0.5126, "step": 10501 }, { "epoch": 0.440229294824099, "grad_norm": 1.7709077596664429, "learning_rate": 6.198935823532682e-06, "loss": 0.5203, "step": 10502 }, { "epoch": 0.44027121343910797, "grad_norm": 1.5111278295516968, "learning_rate": 6.19827676537756e-06, "loss": 0.4865, "step": 10503 }, { "epoch": 0.44031313205411693, "grad_norm": 1.6512737274169922, "learning_rate": 6.197617685133861e-06, "loss": 0.5302, "step": 10504 }, { "epoch": 0.4403550506691259, "grad_norm": 1.8753639459609985, "learning_rate": 6.196958582813737e-06, "loss": 0.5056, "step": 10505 }, { "epoch": 0.44039696928413485, "grad_norm": 1.8827623128890991, "learning_rate": 6.196299458429337e-06, "loss": 0.515, "step": 10506 }, { "epoch": 0.4404388878991438, "grad_norm": 2.173604965209961, "learning_rate": 6.195640311992808e-06, "loss": 0.5111, "step": 10507 }, { "epoch": 0.4404808065141528, "grad_norm": 1.9884071350097656, "learning_rate": 6.194981143516303e-06, "loss": 0.5461, "step": 10508 }, { "epoch": 0.44052272512916174, "grad_norm": 1.6212705373764038, "learning_rate": 6.194321953011975e-06, "loss": 0.5153, "step": 10509 }, { "epoch": 0.4405646437441707, "grad_norm": 1.9142060279846191, "learning_rate": 6.193662740491972e-06, "loss": 0.4996, "step": 10510 }, { "epoch": 0.44060656235917967, "grad_norm": 2.0837371349334717, "learning_rate": 6.193003505968444e-06, "loss": 0.5141, "step": 10511 }, { "epoch": 0.4406484809741886, "grad_norm": 1.5316259860992432, "learning_rate": 6.192344249453548e-06, "loss": 0.4646, "step": 10512 }, { "epoch": 0.4406903995891976, "grad_norm": 1.8401329517364502, "learning_rate": 6.191684970959434e-06, "loss": 0.5153, "step": 10513 }, { "epoch": 0.44073231820420655, "grad_norm": 1.7677828073501587, "learning_rate": 6.191025670498255e-06, "loss": 0.5705, "step": 10514 }, { "epoch": 0.4407742368192155, "grad_norm": 1.691869854927063, "learning_rate": 6.190366348082164e-06, "loss": 0.5218, "step": 10515 }, { "epoch": 0.4408161554342245, "grad_norm": 1.6227957010269165, "learning_rate": 6.189707003723316e-06, "loss": 0.5791, "step": 10516 }, { "epoch": 0.44085807404923344, "grad_norm": 1.857151746749878, "learning_rate": 6.189047637433862e-06, "loss": 0.5209, "step": 10517 }, { "epoch": 0.44089999266424235, "grad_norm": 1.6465178728103638, "learning_rate": 6.1883882492259595e-06, "loss": 0.5204, "step": 10518 }, { "epoch": 0.4409419112792513, "grad_norm": 1.613314151763916, "learning_rate": 6.187728839111763e-06, "loss": 0.508, "step": 10519 }, { "epoch": 0.44098382989426027, "grad_norm": 1.662284016609192, "learning_rate": 6.187069407103428e-06, "loss": 0.5504, "step": 10520 }, { "epoch": 0.44102574850926923, "grad_norm": 1.860363245010376, "learning_rate": 6.186409953213108e-06, "loss": 0.4346, "step": 10521 }, { "epoch": 0.4410676671242782, "grad_norm": 1.4317682981491089, "learning_rate": 6.185750477452962e-06, "loss": 0.4556, "step": 10522 }, { "epoch": 0.44110958573928716, "grad_norm": 9.022972106933594, "learning_rate": 6.185090979835143e-06, "loss": 0.4526, "step": 10523 }, { "epoch": 0.4411515043542961, "grad_norm": 1.9758572578430176, "learning_rate": 6.184431460371813e-06, "loss": 0.5318, "step": 10524 }, { "epoch": 0.4411934229693051, "grad_norm": 2.1361141204833984, "learning_rate": 6.183771919075125e-06, "loss": 0.4887, "step": 10525 }, { "epoch": 0.44123534158431404, "grad_norm": 1.587607979774475, "learning_rate": 6.183112355957238e-06, "loss": 0.4543, "step": 10526 }, { "epoch": 0.441277260199323, "grad_norm": 1.7778613567352295, "learning_rate": 6.1824527710303115e-06, "loss": 0.5786, "step": 10527 }, { "epoch": 0.44131917881433197, "grad_norm": 1.6578235626220703, "learning_rate": 6.181793164306502e-06, "loss": 0.5117, "step": 10528 }, { "epoch": 0.44136109742934093, "grad_norm": 1.5372674465179443, "learning_rate": 6.18113353579797e-06, "loss": 0.5283, "step": 10529 }, { "epoch": 0.4414030160443499, "grad_norm": 1.6848597526550293, "learning_rate": 6.1804738855168735e-06, "loss": 0.4702, "step": 10530 }, { "epoch": 0.44144493465935886, "grad_norm": 1.8492062091827393, "learning_rate": 6.179814213475372e-06, "loss": 0.4988, "step": 10531 }, { "epoch": 0.4414868532743678, "grad_norm": 2.1916065216064453, "learning_rate": 6.1791545196856265e-06, "loss": 0.5228, "step": 10532 }, { "epoch": 0.4415287718893768, "grad_norm": 1.614645004272461, "learning_rate": 6.1784948041598e-06, "loss": 0.4885, "step": 10533 }, { "epoch": 0.44157069050438574, "grad_norm": 1.9103147983551025, "learning_rate": 6.177835066910049e-06, "loss": 0.5462, "step": 10534 }, { "epoch": 0.4416126091193947, "grad_norm": 1.6281394958496094, "learning_rate": 6.1771753079485365e-06, "loss": 0.5075, "step": 10535 }, { "epoch": 0.44165452773440367, "grad_norm": 1.7851670980453491, "learning_rate": 6.1765155272874256e-06, "loss": 0.5385, "step": 10536 }, { "epoch": 0.44169644634941263, "grad_norm": 1.492492437362671, "learning_rate": 6.175855724938877e-06, "loss": 0.5055, "step": 10537 }, { "epoch": 0.4417383649644216, "grad_norm": 1.9041155576705933, "learning_rate": 6.1751959009150545e-06, "loss": 0.5714, "step": 10538 }, { "epoch": 0.44178028357943055, "grad_norm": 1.6545785665512085, "learning_rate": 6.1745360552281195e-06, "loss": 0.4903, "step": 10539 }, { "epoch": 0.4418222021944395, "grad_norm": 2.7887871265411377, "learning_rate": 6.173876187890236e-06, "loss": 0.5484, "step": 10540 }, { "epoch": 0.4418641208094485, "grad_norm": 1.5960832834243774, "learning_rate": 6.173216298913568e-06, "loss": 0.4938, "step": 10541 }, { "epoch": 0.44190603942445744, "grad_norm": 1.6472595930099487, "learning_rate": 6.17255638831028e-06, "loss": 0.4801, "step": 10542 }, { "epoch": 0.44194795803946635, "grad_norm": 1.6325925588607788, "learning_rate": 6.171896456092537e-06, "loss": 0.4932, "step": 10543 }, { "epoch": 0.4419898766544753, "grad_norm": 1.7618374824523926, "learning_rate": 6.1712365022725e-06, "loss": 0.5433, "step": 10544 }, { "epoch": 0.4420317952694843, "grad_norm": 2.010972499847412, "learning_rate": 6.1705765268623384e-06, "loss": 0.5344, "step": 10545 }, { "epoch": 0.44207371388449324, "grad_norm": 1.8845168352127075, "learning_rate": 6.169916529874217e-06, "loss": 0.5201, "step": 10546 }, { "epoch": 0.4421156324995022, "grad_norm": 1.755769968032837, "learning_rate": 6.169256511320302e-06, "loss": 0.4848, "step": 10547 }, { "epoch": 0.44215755111451116, "grad_norm": 2.215998411178589, "learning_rate": 6.1685964712127596e-06, "loss": 0.6129, "step": 10548 }, { "epoch": 0.4421994697295201, "grad_norm": 1.5892211198806763, "learning_rate": 6.1679364095637564e-06, "loss": 0.4943, "step": 10549 }, { "epoch": 0.4422413883445291, "grad_norm": 1.647831678390503, "learning_rate": 6.167276326385461e-06, "loss": 0.5696, "step": 10550 }, { "epoch": 0.44228330695953805, "grad_norm": 1.540784239768982, "learning_rate": 6.166616221690039e-06, "loss": 0.52, "step": 10551 }, { "epoch": 0.442325225574547, "grad_norm": 1.7782419919967651, "learning_rate": 6.16595609548966e-06, "loss": 0.5441, "step": 10552 }, { "epoch": 0.44236714418955597, "grad_norm": 1.7560336589813232, "learning_rate": 6.165295947796494e-06, "loss": 0.4912, "step": 10553 }, { "epoch": 0.44240906280456493, "grad_norm": 1.8318216800689697, "learning_rate": 6.164635778622705e-06, "loss": 0.5237, "step": 10554 }, { "epoch": 0.4424509814195739, "grad_norm": 1.5418037176132202, "learning_rate": 6.163975587980467e-06, "loss": 0.4718, "step": 10555 }, { "epoch": 0.44249290003458286, "grad_norm": 1.582916498184204, "learning_rate": 6.16331537588195e-06, "loss": 0.4894, "step": 10556 }, { "epoch": 0.4425348186495918, "grad_norm": 1.861509084701538, "learning_rate": 6.16265514233932e-06, "loss": 0.5012, "step": 10557 }, { "epoch": 0.4425767372646008, "grad_norm": 1.7126946449279785, "learning_rate": 6.16199488736475e-06, "loss": 0.5979, "step": 10558 }, { "epoch": 0.44261865587960975, "grad_norm": 1.531272053718567, "learning_rate": 6.161334610970412e-06, "loss": 0.4485, "step": 10559 }, { "epoch": 0.4426605744946187, "grad_norm": 1.7801588773727417, "learning_rate": 6.1606743131684745e-06, "loss": 0.5357, "step": 10560 }, { "epoch": 0.44270249310962767, "grad_norm": 1.8951787948608398, "learning_rate": 6.160013993971112e-06, "loss": 0.4889, "step": 10561 }, { "epoch": 0.44274441172463663, "grad_norm": 1.620161533355713, "learning_rate": 6.159353653390495e-06, "loss": 0.4862, "step": 10562 }, { "epoch": 0.4427863303396456, "grad_norm": 1.9124130010604858, "learning_rate": 6.158693291438794e-06, "loss": 0.5581, "step": 10563 }, { "epoch": 0.44282824895465456, "grad_norm": 1.9961662292480469, "learning_rate": 6.158032908128186e-06, "loss": 0.5425, "step": 10564 }, { "epoch": 0.4428701675696635, "grad_norm": 1.72080397605896, "learning_rate": 6.157372503470842e-06, "loss": 0.5366, "step": 10565 }, { "epoch": 0.4429120861846725, "grad_norm": 1.7492097616195679, "learning_rate": 6.156712077478936e-06, "loss": 0.474, "step": 10566 }, { "epoch": 0.44295400479968144, "grad_norm": 1.5751887559890747, "learning_rate": 6.156051630164642e-06, "loss": 0.4661, "step": 10567 }, { "epoch": 0.44299592341469035, "grad_norm": 1.677194595336914, "learning_rate": 6.155391161540134e-06, "loss": 0.511, "step": 10568 }, { "epoch": 0.4430378420296993, "grad_norm": 2.3865163326263428, "learning_rate": 6.154730671617586e-06, "loss": 0.5328, "step": 10569 }, { "epoch": 0.4430797606447083, "grad_norm": 2.1487855911254883, "learning_rate": 6.1540701604091765e-06, "loss": 0.5548, "step": 10570 }, { "epoch": 0.44312167925971724, "grad_norm": 1.8996684551239014, "learning_rate": 6.15340962792708e-06, "loss": 0.5334, "step": 10571 }, { "epoch": 0.4431635978747262, "grad_norm": 1.660717248916626, "learning_rate": 6.1527490741834685e-06, "loss": 0.5343, "step": 10572 }, { "epoch": 0.44320551648973516, "grad_norm": 4.068361759185791, "learning_rate": 6.152088499190523e-06, "loss": 0.4918, "step": 10573 }, { "epoch": 0.4432474351047441, "grad_norm": 1.771561861038208, "learning_rate": 6.151427902960418e-06, "loss": 0.5246, "step": 10574 }, { "epoch": 0.4432893537197531, "grad_norm": 3.231579542160034, "learning_rate": 6.1507672855053325e-06, "loss": 0.5309, "step": 10575 }, { "epoch": 0.44333127233476205, "grad_norm": 1.758235216140747, "learning_rate": 6.150106646837442e-06, "loss": 0.4627, "step": 10576 }, { "epoch": 0.443373190949771, "grad_norm": 1.8576228618621826, "learning_rate": 6.1494459869689274e-06, "loss": 0.5017, "step": 10577 }, { "epoch": 0.44341510956478, "grad_norm": 2.0119354724884033, "learning_rate": 6.148785305911962e-06, "loss": 0.5288, "step": 10578 }, { "epoch": 0.44345702817978894, "grad_norm": 1.9866300821304321, "learning_rate": 6.148124603678729e-06, "loss": 0.5143, "step": 10579 }, { "epoch": 0.4434989467947979, "grad_norm": 1.6589019298553467, "learning_rate": 6.147463880281408e-06, "loss": 0.4683, "step": 10580 }, { "epoch": 0.44354086540980686, "grad_norm": 1.476386547088623, "learning_rate": 6.146803135732173e-06, "loss": 0.4852, "step": 10581 }, { "epoch": 0.4435827840248158, "grad_norm": 1.6619311571121216, "learning_rate": 6.146142370043211e-06, "loss": 0.4415, "step": 10582 }, { "epoch": 0.4436247026398248, "grad_norm": 1.8387529850006104, "learning_rate": 6.145481583226698e-06, "loss": 0.4767, "step": 10583 }, { "epoch": 0.44366662125483375, "grad_norm": 1.5641276836395264, "learning_rate": 6.144820775294815e-06, "loss": 0.4658, "step": 10584 }, { "epoch": 0.4437085398698427, "grad_norm": 1.8601796627044678, "learning_rate": 6.144159946259743e-06, "loss": 0.603, "step": 10585 }, { "epoch": 0.4437504584848517, "grad_norm": 1.791912317276001, "learning_rate": 6.143499096133666e-06, "loss": 0.5666, "step": 10586 }, { "epoch": 0.44379237709986064, "grad_norm": 1.6845135688781738, "learning_rate": 6.142838224928764e-06, "loss": 0.5366, "step": 10587 }, { "epoch": 0.4438342957148696, "grad_norm": 1.6296840906143188, "learning_rate": 6.142177332657217e-06, "loss": 0.5219, "step": 10588 }, { "epoch": 0.44387621432987856, "grad_norm": 1.6684738397598267, "learning_rate": 6.14151641933121e-06, "loss": 0.5124, "step": 10589 }, { "epoch": 0.4439181329448875, "grad_norm": 1.5197341442108154, "learning_rate": 6.140855484962929e-06, "loss": 0.5576, "step": 10590 }, { "epoch": 0.4439600515598965, "grad_norm": 4.141867637634277, "learning_rate": 6.140194529564552e-06, "loss": 0.5086, "step": 10591 }, { "epoch": 0.44400197017490545, "grad_norm": 1.6865465641021729, "learning_rate": 6.139533553148265e-06, "loss": 0.4915, "step": 10592 }, { "epoch": 0.44404388878991435, "grad_norm": 1.5043525695800781, "learning_rate": 6.138872555726252e-06, "loss": 0.5086, "step": 10593 }, { "epoch": 0.4440858074049233, "grad_norm": 1.8212645053863525, "learning_rate": 6.138211537310699e-06, "loss": 0.4903, "step": 10594 }, { "epoch": 0.4441277260199323, "grad_norm": 1.6571948528289795, "learning_rate": 6.13755049791379e-06, "loss": 0.5304, "step": 10595 }, { "epoch": 0.44416964463494124, "grad_norm": 2.814484119415283, "learning_rate": 6.1368894375477095e-06, "loss": 0.5206, "step": 10596 }, { "epoch": 0.4442115632499502, "grad_norm": 1.6612623929977417, "learning_rate": 6.136228356224643e-06, "loss": 0.5317, "step": 10597 }, { "epoch": 0.44425348186495917, "grad_norm": 1.6299115419387817, "learning_rate": 6.135567253956779e-06, "loss": 0.5499, "step": 10598 }, { "epoch": 0.44429540047996813, "grad_norm": 1.6979914903640747, "learning_rate": 6.134906130756304e-06, "loss": 0.5526, "step": 10599 }, { "epoch": 0.4443373190949771, "grad_norm": 2.582956552505493, "learning_rate": 6.1342449866354e-06, "loss": 0.4889, "step": 10600 }, { "epoch": 0.44437923770998605, "grad_norm": 1.6341396570205688, "learning_rate": 6.133583821606259e-06, "loss": 0.5052, "step": 10601 }, { "epoch": 0.444421156324995, "grad_norm": 1.7072001695632935, "learning_rate": 6.132922635681069e-06, "loss": 0.5606, "step": 10602 }, { "epoch": 0.444463074940004, "grad_norm": 1.7620047330856323, "learning_rate": 6.132261428872015e-06, "loss": 0.4998, "step": 10603 }, { "epoch": 0.44450499355501294, "grad_norm": 1.6342918872833252, "learning_rate": 6.131600201191286e-06, "loss": 0.5142, "step": 10604 }, { "epoch": 0.4445469121700219, "grad_norm": 1.536424994468689, "learning_rate": 6.1309389526510734e-06, "loss": 0.4431, "step": 10605 }, { "epoch": 0.44458883078503086, "grad_norm": 1.6450965404510498, "learning_rate": 6.130277683263564e-06, "loss": 0.5348, "step": 10606 }, { "epoch": 0.4446307494000398, "grad_norm": 1.608713984489441, "learning_rate": 6.129616393040948e-06, "loss": 0.513, "step": 10607 }, { "epoch": 0.4446726680150488, "grad_norm": 1.6751090288162231, "learning_rate": 6.1289550819954154e-06, "loss": 0.4545, "step": 10608 }, { "epoch": 0.44471458663005775, "grad_norm": 1.6529417037963867, "learning_rate": 6.128293750139157e-06, "loss": 0.5736, "step": 10609 }, { "epoch": 0.4447565052450667, "grad_norm": 1.4485734701156616, "learning_rate": 6.127632397484362e-06, "loss": 0.4602, "step": 10610 }, { "epoch": 0.4447984238600757, "grad_norm": 1.8693573474884033, "learning_rate": 6.126971024043225e-06, "loss": 0.5521, "step": 10611 }, { "epoch": 0.44484034247508464, "grad_norm": 1.6994227170944214, "learning_rate": 6.126309629827933e-06, "loss": 0.5089, "step": 10612 }, { "epoch": 0.4448822610900936, "grad_norm": 2.3504607677459717, "learning_rate": 6.125648214850681e-06, "loss": 0.5497, "step": 10613 }, { "epoch": 0.44492417970510256, "grad_norm": 1.6182175874710083, "learning_rate": 6.124986779123659e-06, "loss": 0.5112, "step": 10614 }, { "epoch": 0.4449660983201115, "grad_norm": 1.650740146636963, "learning_rate": 6.124325322659061e-06, "loss": 0.4868, "step": 10615 }, { "epoch": 0.4450080169351205, "grad_norm": 1.6009674072265625, "learning_rate": 6.123663845469082e-06, "loss": 0.5661, "step": 10616 }, { "epoch": 0.44504993555012945, "grad_norm": 2.3046348094940186, "learning_rate": 6.123002347565913e-06, "loss": 0.4656, "step": 10617 }, { "epoch": 0.44509185416513836, "grad_norm": 1.6188756227493286, "learning_rate": 6.122340828961748e-06, "loss": 0.468, "step": 10618 }, { "epoch": 0.4451337727801473, "grad_norm": 1.8334593772888184, "learning_rate": 6.121679289668781e-06, "loss": 0.4933, "step": 10619 }, { "epoch": 0.4451756913951563, "grad_norm": 1.628055453300476, "learning_rate": 6.121017729699207e-06, "loss": 0.4769, "step": 10620 }, { "epoch": 0.44521761001016524, "grad_norm": 1.78512442111969, "learning_rate": 6.120356149065221e-06, "loss": 0.5026, "step": 10621 }, { "epoch": 0.4452595286251742, "grad_norm": 1.7948044538497925, "learning_rate": 6.119694547779019e-06, "loss": 0.5287, "step": 10622 }, { "epoch": 0.44530144724018317, "grad_norm": 1.6173118352890015, "learning_rate": 6.1190329258527946e-06, "loss": 0.4643, "step": 10623 }, { "epoch": 0.44534336585519213, "grad_norm": 1.93272864818573, "learning_rate": 6.118371283298745e-06, "loss": 0.5265, "step": 10624 }, { "epoch": 0.4453852844702011, "grad_norm": 1.8287495374679565, "learning_rate": 6.117709620129068e-06, "loss": 0.5329, "step": 10625 }, { "epoch": 0.44542720308521006, "grad_norm": 1.7430123090744019, "learning_rate": 6.117047936355959e-06, "loss": 0.5087, "step": 10626 }, { "epoch": 0.445469121700219, "grad_norm": 1.6395621299743652, "learning_rate": 6.1163862319916165e-06, "loss": 0.5229, "step": 10627 }, { "epoch": 0.445511040315228, "grad_norm": 1.8729723691940308, "learning_rate": 6.115724507048235e-06, "loss": 0.4948, "step": 10628 }, { "epoch": 0.44555295893023694, "grad_norm": 1.776129961013794, "learning_rate": 6.115062761538017e-06, "loss": 0.5454, "step": 10629 }, { "epoch": 0.4455948775452459, "grad_norm": 1.6377886533737183, "learning_rate": 6.114400995473157e-06, "loss": 0.5313, "step": 10630 }, { "epoch": 0.44563679616025487, "grad_norm": 1.6074429750442505, "learning_rate": 6.113739208865855e-06, "loss": 0.4864, "step": 10631 }, { "epoch": 0.44567871477526383, "grad_norm": 1.9577184915542603, "learning_rate": 6.113077401728312e-06, "loss": 0.5478, "step": 10632 }, { "epoch": 0.4457206333902728, "grad_norm": 1.9229799509048462, "learning_rate": 6.1124155740727255e-06, "loss": 0.5176, "step": 10633 }, { "epoch": 0.44576255200528175, "grad_norm": 1.9160058498382568, "learning_rate": 6.111753725911293e-06, "loss": 0.5522, "step": 10634 }, { "epoch": 0.4458044706202907, "grad_norm": 1.6888206005096436, "learning_rate": 6.111091857256221e-06, "loss": 0.5336, "step": 10635 }, { "epoch": 0.4458463892352997, "grad_norm": 1.6554456949234009, "learning_rate": 6.110429968119705e-06, "loss": 0.472, "step": 10636 }, { "epoch": 0.44588830785030864, "grad_norm": 1.4692403078079224, "learning_rate": 6.109768058513948e-06, "loss": 0.4878, "step": 10637 }, { "epoch": 0.4459302264653176, "grad_norm": 1.540099024772644, "learning_rate": 6.109106128451151e-06, "loss": 0.4881, "step": 10638 }, { "epoch": 0.44597214508032657, "grad_norm": 1.5893316268920898, "learning_rate": 6.108444177943516e-06, "loss": 0.5079, "step": 10639 }, { "epoch": 0.4460140636953355, "grad_norm": 1.747273325920105, "learning_rate": 6.107782207003245e-06, "loss": 0.5066, "step": 10640 }, { "epoch": 0.4460559823103445, "grad_norm": 1.482652187347412, "learning_rate": 6.10712021564254e-06, "loss": 0.4819, "step": 10641 }, { "epoch": 0.44609790092535345, "grad_norm": 1.6734132766723633, "learning_rate": 6.106458203873605e-06, "loss": 0.512, "step": 10642 }, { "epoch": 0.44613981954036236, "grad_norm": 1.8294620513916016, "learning_rate": 6.105796171708642e-06, "loss": 0.493, "step": 10643 }, { "epoch": 0.4461817381553713, "grad_norm": 1.64866304397583, "learning_rate": 6.1051341191598554e-06, "loss": 0.5104, "step": 10644 }, { "epoch": 0.4462236567703803, "grad_norm": 1.8162833452224731, "learning_rate": 6.104472046239449e-06, "loss": 0.5618, "step": 10645 }, { "epoch": 0.44626557538538925, "grad_norm": 1.6771119832992554, "learning_rate": 6.103809952959629e-06, "loss": 0.4867, "step": 10646 }, { "epoch": 0.4463074940003982, "grad_norm": 1.6148353815078735, "learning_rate": 6.103147839332597e-06, "loss": 0.4923, "step": 10647 }, { "epoch": 0.44634941261540717, "grad_norm": 1.644154667854309, "learning_rate": 6.10248570537056e-06, "loss": 0.5397, "step": 10648 }, { "epoch": 0.44639133123041613, "grad_norm": 1.6518540382385254, "learning_rate": 6.101823551085722e-06, "loss": 0.5126, "step": 10649 }, { "epoch": 0.4464332498454251, "grad_norm": 1.6072105169296265, "learning_rate": 6.101161376490292e-06, "loss": 0.4747, "step": 10650 }, { "epoch": 0.44647516846043406, "grad_norm": 1.7462443113327026, "learning_rate": 6.100499181596473e-06, "loss": 0.5207, "step": 10651 }, { "epoch": 0.446517087075443, "grad_norm": 1.6528054475784302, "learning_rate": 6.099836966416473e-06, "loss": 0.5386, "step": 10652 }, { "epoch": 0.446559005690452, "grad_norm": 1.582110047340393, "learning_rate": 6.099174730962499e-06, "loss": 0.4635, "step": 10653 }, { "epoch": 0.44660092430546094, "grad_norm": 1.843766689300537, "learning_rate": 6.098512475246758e-06, "loss": 0.5605, "step": 10654 }, { "epoch": 0.4466428429204699, "grad_norm": 1.7694367170333862, "learning_rate": 6.09785019928146e-06, "loss": 0.5585, "step": 10655 }, { "epoch": 0.44668476153547887, "grad_norm": 1.7919690608978271, "learning_rate": 6.097187903078811e-06, "loss": 0.5573, "step": 10656 }, { "epoch": 0.44672668015048783, "grad_norm": 2.2727365493774414, "learning_rate": 6.096525586651019e-06, "loss": 0.5031, "step": 10657 }, { "epoch": 0.4467685987654968, "grad_norm": 1.7641185522079468, "learning_rate": 6.095863250010293e-06, "loss": 0.533, "step": 10658 }, { "epoch": 0.44681051738050576, "grad_norm": 1.7392082214355469, "learning_rate": 6.0952008931688444e-06, "loss": 0.521, "step": 10659 }, { "epoch": 0.4468524359955147, "grad_norm": 1.6102168560028076, "learning_rate": 6.094538516138881e-06, "loss": 0.5055, "step": 10660 }, { "epoch": 0.4468943546105237, "grad_norm": 2.2227261066436768, "learning_rate": 6.093876118932612e-06, "loss": 0.534, "step": 10661 }, { "epoch": 0.44693627322553264, "grad_norm": 1.6360324621200562, "learning_rate": 6.0932137015622495e-06, "loss": 0.5009, "step": 10662 }, { "epoch": 0.4469781918405416, "grad_norm": 1.7317265272140503, "learning_rate": 6.092551264040004e-06, "loss": 0.5254, "step": 10663 }, { "epoch": 0.44702011045555057, "grad_norm": 1.554354190826416, "learning_rate": 6.091888806378087e-06, "loss": 0.5024, "step": 10664 }, { "epoch": 0.44706202907055953, "grad_norm": 1.7431910037994385, "learning_rate": 6.091226328588708e-06, "loss": 0.4813, "step": 10665 }, { "epoch": 0.4471039476855685, "grad_norm": 1.773358702659607, "learning_rate": 6.090563830684082e-06, "loss": 0.4756, "step": 10666 }, { "epoch": 0.44714586630057745, "grad_norm": 1.4847391843795776, "learning_rate": 6.089901312676418e-06, "loss": 0.4699, "step": 10667 }, { "epoch": 0.44718778491558636, "grad_norm": 1.795462965965271, "learning_rate": 6.08923877457793e-06, "loss": 0.4882, "step": 10668 }, { "epoch": 0.4472297035305953, "grad_norm": 2.4222729206085205, "learning_rate": 6.0885762164008335e-06, "loss": 0.5244, "step": 10669 }, { "epoch": 0.4472716221456043, "grad_norm": 1.4824975728988647, "learning_rate": 6.087913638157337e-06, "loss": 0.4748, "step": 10670 }, { "epoch": 0.44731354076061325, "grad_norm": 1.812166690826416, "learning_rate": 6.087251039859656e-06, "loss": 0.4897, "step": 10671 }, { "epoch": 0.4473554593756222, "grad_norm": 2.0083065032958984, "learning_rate": 6.086588421520007e-06, "loss": 0.4483, "step": 10672 }, { "epoch": 0.4473973779906312, "grad_norm": 1.747402548789978, "learning_rate": 6.085925783150601e-06, "loss": 0.5079, "step": 10673 }, { "epoch": 0.44743929660564014, "grad_norm": 1.677000641822815, "learning_rate": 6.085263124763657e-06, "loss": 0.5135, "step": 10674 }, { "epoch": 0.4474812152206491, "grad_norm": 1.7133615016937256, "learning_rate": 6.0846004463713844e-06, "loss": 0.5402, "step": 10675 }, { "epoch": 0.44752313383565806, "grad_norm": 1.638821005821228, "learning_rate": 6.083937747986004e-06, "loss": 0.524, "step": 10676 }, { "epoch": 0.447565052450667, "grad_norm": 1.7722991704940796, "learning_rate": 6.083275029619729e-06, "loss": 0.5122, "step": 10677 }, { "epoch": 0.447606971065676, "grad_norm": 1.6292163133621216, "learning_rate": 6.0826122912847765e-06, "loss": 0.4649, "step": 10678 }, { "epoch": 0.44764888968068495, "grad_norm": 1.6493128538131714, "learning_rate": 6.081949532993363e-06, "loss": 0.4997, "step": 10679 }, { "epoch": 0.4476908082956939, "grad_norm": 2.323378086090088, "learning_rate": 6.0812867547577046e-06, "loss": 0.5281, "step": 10680 }, { "epoch": 0.44773272691070287, "grad_norm": 1.787068247795105, "learning_rate": 6.080623956590022e-06, "loss": 0.5558, "step": 10681 }, { "epoch": 0.44777464552571183, "grad_norm": 1.6750433444976807, "learning_rate": 6.079961138502529e-06, "loss": 0.5171, "step": 10682 }, { "epoch": 0.4478165641407208, "grad_norm": 1.8342692852020264, "learning_rate": 6.079298300507446e-06, "loss": 0.5657, "step": 10683 }, { "epoch": 0.44785848275572976, "grad_norm": 1.5433063507080078, "learning_rate": 6.078635442616992e-06, "loss": 0.5101, "step": 10684 }, { "epoch": 0.4479004013707387, "grad_norm": 1.549621343612671, "learning_rate": 6.077972564843383e-06, "loss": 0.4997, "step": 10685 }, { "epoch": 0.4479423199857477, "grad_norm": 1.6419086456298828, "learning_rate": 6.07730966719884e-06, "loss": 0.5575, "step": 10686 }, { "epoch": 0.44798423860075665, "grad_norm": 1.3679410219192505, "learning_rate": 6.076646749695583e-06, "loss": 0.5195, "step": 10687 }, { "epoch": 0.4480261572157656, "grad_norm": 1.9205700159072876, "learning_rate": 6.0759838123458335e-06, "loss": 0.5446, "step": 10688 }, { "epoch": 0.44806807583077457, "grad_norm": 1.74900484085083, "learning_rate": 6.075320855161807e-06, "loss": 0.5684, "step": 10689 }, { "epoch": 0.44810999444578353, "grad_norm": 1.8774079084396362, "learning_rate": 6.074657878155728e-06, "loss": 0.4598, "step": 10690 }, { "epoch": 0.4481519130607925, "grad_norm": 1.5818822383880615, "learning_rate": 6.073994881339818e-06, "loss": 0.5238, "step": 10691 }, { "epoch": 0.44819383167580146, "grad_norm": 1.443477988243103, "learning_rate": 6.073331864726296e-06, "loss": 0.4805, "step": 10692 }, { "epoch": 0.44823575029081036, "grad_norm": 1.937003493309021, "learning_rate": 6.072668828327385e-06, "loss": 0.5449, "step": 10693 }, { "epoch": 0.4482776689058193, "grad_norm": 1.4583699703216553, "learning_rate": 6.0720057721553075e-06, "loss": 0.47, "step": 10694 }, { "epoch": 0.4483195875208283, "grad_norm": 1.5826423168182373, "learning_rate": 6.071342696222286e-06, "loss": 0.4751, "step": 10695 }, { "epoch": 0.44836150613583725, "grad_norm": 1.7065719366073608, "learning_rate": 6.070679600540541e-06, "loss": 0.5296, "step": 10696 }, { "epoch": 0.4484034247508462, "grad_norm": 1.5902959108352661, "learning_rate": 6.0700164851223e-06, "loss": 0.4776, "step": 10697 }, { "epoch": 0.4484453433658552, "grad_norm": 1.7255370616912842, "learning_rate": 6.069353349979783e-06, "loss": 0.4947, "step": 10698 }, { "epoch": 0.44848726198086414, "grad_norm": 1.7603278160095215, "learning_rate": 6.068690195125215e-06, "loss": 0.5397, "step": 10699 }, { "epoch": 0.4485291805958731, "grad_norm": 1.5579478740692139, "learning_rate": 6.068027020570822e-06, "loss": 0.502, "step": 10700 }, { "epoch": 0.44857109921088206, "grad_norm": 1.646144986152649, "learning_rate": 6.067363826328827e-06, "loss": 0.5504, "step": 10701 }, { "epoch": 0.448613017825891, "grad_norm": 3.4189136028289795, "learning_rate": 6.066700612411455e-06, "loss": 0.4951, "step": 10702 }, { "epoch": 0.4486549364409, "grad_norm": 2.075955390930176, "learning_rate": 6.0660373788309324e-06, "loss": 0.564, "step": 10703 }, { "epoch": 0.44869685505590895, "grad_norm": 1.7901480197906494, "learning_rate": 6.065374125599484e-06, "loss": 0.4657, "step": 10704 }, { "epoch": 0.4487387736709179, "grad_norm": 2.1589515209198, "learning_rate": 6.064710852729335e-06, "loss": 0.596, "step": 10705 }, { "epoch": 0.4487806922859269, "grad_norm": 1.7006455659866333, "learning_rate": 6.064047560232717e-06, "loss": 0.5364, "step": 10706 }, { "epoch": 0.44882261090093584, "grad_norm": 1.754886269569397, "learning_rate": 6.063384248121851e-06, "loss": 0.5263, "step": 10707 }, { "epoch": 0.4488645295159448, "grad_norm": 2.52325439453125, "learning_rate": 6.062720916408965e-06, "loss": 0.5305, "step": 10708 }, { "epoch": 0.44890644813095376, "grad_norm": 1.596346139907837, "learning_rate": 6.06205756510629e-06, "loss": 0.5261, "step": 10709 }, { "epoch": 0.4489483667459627, "grad_norm": 1.7594441175460815, "learning_rate": 6.061394194226053e-06, "loss": 0.5471, "step": 10710 }, { "epoch": 0.4489902853609717, "grad_norm": 1.8502482175827026, "learning_rate": 6.060730803780479e-06, "loss": 0.5917, "step": 10711 }, { "epoch": 0.44903220397598065, "grad_norm": 1.6810182332992554, "learning_rate": 6.060067393781801e-06, "loss": 0.5595, "step": 10712 }, { "epoch": 0.4490741225909896, "grad_norm": 1.743784785270691, "learning_rate": 6.059403964242245e-06, "loss": 0.5288, "step": 10713 }, { "epoch": 0.4491160412059986, "grad_norm": 1.5980515480041504, "learning_rate": 6.0587405151740405e-06, "loss": 0.4637, "step": 10714 }, { "epoch": 0.44915795982100754, "grad_norm": 3.876868486404419, "learning_rate": 6.05807704658942e-06, "loss": 0.5295, "step": 10715 }, { "epoch": 0.4491998784360165, "grad_norm": 1.5229965448379517, "learning_rate": 6.057413558500612e-06, "loss": 0.5228, "step": 10716 }, { "epoch": 0.44924179705102546, "grad_norm": 1.7116302251815796, "learning_rate": 6.0567500509198465e-06, "loss": 0.5101, "step": 10717 }, { "epoch": 0.44928371566603437, "grad_norm": 1.7333979606628418, "learning_rate": 6.056086523859354e-06, "loss": 0.5495, "step": 10718 }, { "epoch": 0.44932563428104333, "grad_norm": 1.8207793235778809, "learning_rate": 6.055422977331366e-06, "loss": 0.5229, "step": 10719 }, { "epoch": 0.4493675528960523, "grad_norm": 1.6244055032730103, "learning_rate": 6.0547594113481164e-06, "loss": 0.4833, "step": 10720 }, { "epoch": 0.44940947151106125, "grad_norm": 1.9573532342910767, "learning_rate": 6.054095825921834e-06, "loss": 0.5142, "step": 10721 }, { "epoch": 0.4494513901260702, "grad_norm": 2.078948736190796, "learning_rate": 6.053432221064753e-06, "loss": 0.5413, "step": 10722 }, { "epoch": 0.4494933087410792, "grad_norm": 1.6748976707458496, "learning_rate": 6.052768596789103e-06, "loss": 0.4739, "step": 10723 }, { "epoch": 0.44953522735608814, "grad_norm": 1.465482473373413, "learning_rate": 6.052104953107122e-06, "loss": 0.4796, "step": 10724 }, { "epoch": 0.4495771459710971, "grad_norm": 2.8273346424102783, "learning_rate": 6.051441290031042e-06, "loss": 0.4571, "step": 10725 }, { "epoch": 0.44961906458610607, "grad_norm": 1.6990822553634644, "learning_rate": 6.050777607573092e-06, "loss": 0.5521, "step": 10726 }, { "epoch": 0.449660983201115, "grad_norm": 1.68958580493927, "learning_rate": 6.050113905745511e-06, "loss": 0.5259, "step": 10727 }, { "epoch": 0.449702901816124, "grad_norm": 1.674116849899292, "learning_rate": 6.049450184560532e-06, "loss": 0.5027, "step": 10728 }, { "epoch": 0.44974482043113295, "grad_norm": 1.6994513273239136, "learning_rate": 6.048786444030388e-06, "loss": 0.5874, "step": 10729 }, { "epoch": 0.4497867390461419, "grad_norm": 2.513674736022949, "learning_rate": 6.048122684167318e-06, "loss": 0.5048, "step": 10730 }, { "epoch": 0.4498286576611509, "grad_norm": 2.0008397102355957, "learning_rate": 6.0474589049835545e-06, "loss": 0.5245, "step": 10731 }, { "epoch": 0.44987057627615984, "grad_norm": 3.031130075454712, "learning_rate": 6.046795106491333e-06, "loss": 0.5882, "step": 10732 }, { "epoch": 0.4499124948911688, "grad_norm": 1.5881513357162476, "learning_rate": 6.046131288702892e-06, "loss": 0.4866, "step": 10733 }, { "epoch": 0.44995441350617776, "grad_norm": 1.8382973670959473, "learning_rate": 6.045467451630467e-06, "loss": 0.5034, "step": 10734 }, { "epoch": 0.4499963321211867, "grad_norm": 2.1799323558807373, "learning_rate": 6.044803595286295e-06, "loss": 0.5006, "step": 10735 }, { "epoch": 0.4500382507361957, "grad_norm": 2.024700164794922, "learning_rate": 6.044139719682612e-06, "loss": 0.5601, "step": 10736 }, { "epoch": 0.45008016935120465, "grad_norm": 2.5657799243927, "learning_rate": 6.043475824831659e-06, "loss": 0.5004, "step": 10737 }, { "epoch": 0.4501220879662136, "grad_norm": 1.5544058084487915, "learning_rate": 6.0428119107456696e-06, "loss": 0.4756, "step": 10738 }, { "epoch": 0.4501640065812226, "grad_norm": 1.7143305540084839, "learning_rate": 6.042147977436885e-06, "loss": 0.5053, "step": 10739 }, { "epoch": 0.45020592519623154, "grad_norm": 2.326659917831421, "learning_rate": 6.0414840249175435e-06, "loss": 0.4702, "step": 10740 }, { "epoch": 0.4502478438112405, "grad_norm": 1.6306039094924927, "learning_rate": 6.0408200531998825e-06, "loss": 0.5526, "step": 10741 }, { "epoch": 0.45028976242624946, "grad_norm": 1.7563753128051758, "learning_rate": 6.040156062296143e-06, "loss": 0.5725, "step": 10742 }, { "epoch": 0.45033168104125837, "grad_norm": 1.9139022827148438, "learning_rate": 6.039492052218565e-06, "loss": 0.5066, "step": 10743 }, { "epoch": 0.45037359965626733, "grad_norm": 1.684080958366394, "learning_rate": 6.038828022979389e-06, "loss": 0.5127, "step": 10744 }, { "epoch": 0.4504155182712763, "grad_norm": 1.565782904624939, "learning_rate": 6.038163974590854e-06, "loss": 0.5171, "step": 10745 }, { "epoch": 0.45045743688628526, "grad_norm": 1.8266863822937012, "learning_rate": 6.0374999070652e-06, "loss": 0.536, "step": 10746 }, { "epoch": 0.4504993555012942, "grad_norm": 1.8776450157165527, "learning_rate": 6.036835820414671e-06, "loss": 0.5478, "step": 10747 }, { "epoch": 0.4505412741163032, "grad_norm": 1.8076426982879639, "learning_rate": 6.036171714651507e-06, "loss": 0.5529, "step": 10748 }, { "epoch": 0.45058319273131214, "grad_norm": 1.7350667715072632, "learning_rate": 6.03550758978795e-06, "loss": 0.5103, "step": 10749 }, { "epoch": 0.4506251113463211, "grad_norm": 2.0474867820739746, "learning_rate": 6.034843445836241e-06, "loss": 0.5157, "step": 10750 }, { "epoch": 0.45066702996133007, "grad_norm": 2.2413218021392822, "learning_rate": 6.034179282808623e-06, "loss": 0.4922, "step": 10751 }, { "epoch": 0.45070894857633903, "grad_norm": 1.6942757368087769, "learning_rate": 6.033515100717342e-06, "loss": 0.5514, "step": 10752 }, { "epoch": 0.450750867191348, "grad_norm": 1.6388195753097534, "learning_rate": 6.032850899574638e-06, "loss": 0.4668, "step": 10753 }, { "epoch": 0.45079278580635695, "grad_norm": 1.6037088632583618, "learning_rate": 6.032186679392755e-06, "loss": 0.4974, "step": 10754 }, { "epoch": 0.4508347044213659, "grad_norm": 1.9345426559448242, "learning_rate": 6.031522440183938e-06, "loss": 0.5666, "step": 10755 }, { "epoch": 0.4508766230363749, "grad_norm": 2.2494966983795166, "learning_rate": 6.03085818196043e-06, "loss": 0.5406, "step": 10756 }, { "epoch": 0.45091854165138384, "grad_norm": 2.493164539337158, "learning_rate": 6.030193904734477e-06, "loss": 0.5152, "step": 10757 }, { "epoch": 0.4509604602663928, "grad_norm": 1.830873727798462, "learning_rate": 6.029529608518325e-06, "loss": 0.5078, "step": 10758 }, { "epoch": 0.45100237888140177, "grad_norm": 1.6026573181152344, "learning_rate": 6.028865293324217e-06, "loss": 0.5224, "step": 10759 }, { "epoch": 0.45104429749641073, "grad_norm": 1.726421594619751, "learning_rate": 6.028200959164399e-06, "loss": 0.4562, "step": 10760 }, { "epoch": 0.4510862161114197, "grad_norm": 1.997990369796753, "learning_rate": 6.027536606051119e-06, "loss": 0.5089, "step": 10761 }, { "epoch": 0.45112813472642865, "grad_norm": 1.6523131132125854, "learning_rate": 6.026872233996621e-06, "loss": 0.4397, "step": 10762 }, { "epoch": 0.4511700533414376, "grad_norm": 1.8537958860397339, "learning_rate": 6.026207843013153e-06, "loss": 0.482, "step": 10763 }, { "epoch": 0.4512119719564466, "grad_norm": 1.9671703577041626, "learning_rate": 6.025543433112963e-06, "loss": 0.5179, "step": 10764 }, { "epoch": 0.45125389057145554, "grad_norm": 1.9420392513275146, "learning_rate": 6.024879004308297e-06, "loss": 0.5173, "step": 10765 }, { "epoch": 0.4512958091864645, "grad_norm": 1.9380861520767212, "learning_rate": 6.024214556611403e-06, "loss": 0.521, "step": 10766 }, { "epoch": 0.45133772780147347, "grad_norm": 2.7945170402526855, "learning_rate": 6.023550090034531e-06, "loss": 0.5561, "step": 10767 }, { "epoch": 0.45137964641648237, "grad_norm": 1.9718072414398193, "learning_rate": 6.022885604589927e-06, "loss": 0.4572, "step": 10768 }, { "epoch": 0.45142156503149133, "grad_norm": 1.9169974327087402, "learning_rate": 6.0222211002898416e-06, "loss": 0.565, "step": 10769 }, { "epoch": 0.4514634836465003, "grad_norm": 1.786757230758667, "learning_rate": 6.021556577146523e-06, "loss": 0.51, "step": 10770 }, { "epoch": 0.45150540226150926, "grad_norm": 1.7077685594558716, "learning_rate": 6.020892035172221e-06, "loss": 0.5652, "step": 10771 }, { "epoch": 0.4515473208765182, "grad_norm": 2.2339844703674316, "learning_rate": 6.020227474379187e-06, "loss": 0.5079, "step": 10772 }, { "epoch": 0.4515892394915272, "grad_norm": 2.1187334060668945, "learning_rate": 6.019562894779667e-06, "loss": 0.4579, "step": 10773 }, { "epoch": 0.45163115810653615, "grad_norm": 1.4600528478622437, "learning_rate": 6.0188982963859165e-06, "loss": 0.5177, "step": 10774 }, { "epoch": 0.4516730767215451, "grad_norm": 1.8066680431365967, "learning_rate": 6.018233679210185e-06, "loss": 0.5006, "step": 10775 }, { "epoch": 0.45171499533655407, "grad_norm": 1.8067741394042969, "learning_rate": 6.017569043264723e-06, "loss": 0.5033, "step": 10776 }, { "epoch": 0.45175691395156303, "grad_norm": 1.9096282720565796, "learning_rate": 6.016904388561782e-06, "loss": 0.4625, "step": 10777 }, { "epoch": 0.451798832566572, "grad_norm": 1.6448514461517334, "learning_rate": 6.0162397151136145e-06, "loss": 0.5475, "step": 10778 }, { "epoch": 0.45184075118158096, "grad_norm": 1.9908984899520874, "learning_rate": 6.015575022932471e-06, "loss": 0.5675, "step": 10779 }, { "epoch": 0.4518826697965899, "grad_norm": 1.7601293325424194, "learning_rate": 6.014910312030608e-06, "loss": 0.5068, "step": 10780 }, { "epoch": 0.4519245884115989, "grad_norm": 2.073871612548828, "learning_rate": 6.014245582420276e-06, "loss": 0.56, "step": 10781 }, { "epoch": 0.45196650702660784, "grad_norm": 1.7650867700576782, "learning_rate": 6.0135808341137295e-06, "loss": 0.4925, "step": 10782 }, { "epoch": 0.4520084256416168, "grad_norm": 2.1504552364349365, "learning_rate": 6.012916067123221e-06, "loss": 0.4892, "step": 10783 }, { "epoch": 0.45205034425662577, "grad_norm": 1.9897654056549072, "learning_rate": 6.012251281461006e-06, "loss": 0.5208, "step": 10784 }, { "epoch": 0.45209226287163473, "grad_norm": 2.3986148834228516, "learning_rate": 6.011586477139335e-06, "loss": 0.5401, "step": 10785 }, { "epoch": 0.4521341814866437, "grad_norm": 3.1216485500335693, "learning_rate": 6.01092165417047e-06, "loss": 0.4994, "step": 10786 }, { "epoch": 0.45217610010165266, "grad_norm": 1.7110495567321777, "learning_rate": 6.0102568125666594e-06, "loss": 0.563, "step": 10787 }, { "epoch": 0.4522180187166616, "grad_norm": 1.702291488647461, "learning_rate": 6.0095919523401604e-06, "loss": 0.4522, "step": 10788 }, { "epoch": 0.4522599373316706, "grad_norm": 2.600931167602539, "learning_rate": 6.00892707350323e-06, "loss": 0.5262, "step": 10789 }, { "epoch": 0.45230185594667954, "grad_norm": 1.8975975513458252, "learning_rate": 6.008262176068125e-06, "loss": 0.5411, "step": 10790 }, { "epoch": 0.4523437745616885, "grad_norm": 1.776070237159729, "learning_rate": 6.0075972600471e-06, "loss": 0.5546, "step": 10791 }, { "epoch": 0.45238569317669747, "grad_norm": 1.931267499923706, "learning_rate": 6.0069323254524125e-06, "loss": 0.5659, "step": 10792 }, { "epoch": 0.4524276117917064, "grad_norm": 1.6228896379470825, "learning_rate": 6.00626737229632e-06, "loss": 0.5117, "step": 10793 }, { "epoch": 0.45246953040671534, "grad_norm": 1.5189021825790405, "learning_rate": 6.005602400591078e-06, "loss": 0.4909, "step": 10794 }, { "epoch": 0.4525114490217243, "grad_norm": 2.0568714141845703, "learning_rate": 6.004937410348949e-06, "loss": 0.4751, "step": 10795 }, { "epoch": 0.45255336763673326, "grad_norm": 1.986501932144165, "learning_rate": 6.0042724015821865e-06, "loss": 0.5384, "step": 10796 }, { "epoch": 0.4525952862517422, "grad_norm": 2.054292678833008, "learning_rate": 6.003607374303048e-06, "loss": 0.5637, "step": 10797 }, { "epoch": 0.4526372048667512, "grad_norm": 1.6974220275878906, "learning_rate": 6.002942328523798e-06, "loss": 0.4816, "step": 10798 }, { "epoch": 0.45267912348176015, "grad_norm": 1.9414663314819336, "learning_rate": 6.002277264256693e-06, "loss": 0.5356, "step": 10799 }, { "epoch": 0.4527210420967691, "grad_norm": 1.7994194030761719, "learning_rate": 6.001612181513992e-06, "loss": 0.5465, "step": 10800 }, { "epoch": 0.4527629607117781, "grad_norm": 2.826901435852051, "learning_rate": 6.000947080307954e-06, "loss": 0.5441, "step": 10801 }, { "epoch": 0.45280487932678704, "grad_norm": 1.8090459108352661, "learning_rate": 6.000281960650841e-06, "loss": 0.5486, "step": 10802 }, { "epoch": 0.452846797941796, "grad_norm": 1.6846486330032349, "learning_rate": 5.999616822554912e-06, "loss": 0.4868, "step": 10803 }, { "epoch": 0.45288871655680496, "grad_norm": 1.7505583763122559, "learning_rate": 5.99895166603243e-06, "loss": 0.5016, "step": 10804 }, { "epoch": 0.4529306351718139, "grad_norm": 1.764391303062439, "learning_rate": 5.9982864910956555e-06, "loss": 0.4862, "step": 10805 }, { "epoch": 0.4529725537868229, "grad_norm": 1.7589426040649414, "learning_rate": 5.997621297756847e-06, "loss": 0.5285, "step": 10806 }, { "epoch": 0.45301447240183185, "grad_norm": 2.2658908367156982, "learning_rate": 5.996956086028272e-06, "loss": 0.5613, "step": 10807 }, { "epoch": 0.4530563910168408, "grad_norm": 1.7910088300704956, "learning_rate": 5.996290855922188e-06, "loss": 0.4756, "step": 10808 }, { "epoch": 0.45309830963184977, "grad_norm": 1.671238899230957, "learning_rate": 5.9956256074508605e-06, "loss": 0.5036, "step": 10809 }, { "epoch": 0.45314022824685873, "grad_norm": 1.7602794170379639, "learning_rate": 5.99496034062655e-06, "loss": 0.5269, "step": 10810 }, { "epoch": 0.4531821468618677, "grad_norm": 1.8030447959899902, "learning_rate": 5.994295055461523e-06, "loss": 0.5021, "step": 10811 }, { "epoch": 0.45322406547687666, "grad_norm": 1.712303638458252, "learning_rate": 5.99362975196804e-06, "loss": 0.5113, "step": 10812 }, { "epoch": 0.4532659840918856, "grad_norm": 1.8470994234085083, "learning_rate": 5.9929644301583655e-06, "loss": 0.5026, "step": 10813 }, { "epoch": 0.4533079027068946, "grad_norm": 2.2159836292266846, "learning_rate": 5.992299090044767e-06, "loss": 0.5169, "step": 10814 }, { "epoch": 0.45334982132190355, "grad_norm": 1.807856798171997, "learning_rate": 5.991633731639505e-06, "loss": 0.4996, "step": 10815 }, { "epoch": 0.4533917399369125, "grad_norm": 1.833101749420166, "learning_rate": 5.990968354954845e-06, "loss": 0.4953, "step": 10816 }, { "epoch": 0.45343365855192147, "grad_norm": 1.9040849208831787, "learning_rate": 5.990302960003054e-06, "loss": 0.5502, "step": 10817 }, { "epoch": 0.4534755771669304, "grad_norm": 1.687809944152832, "learning_rate": 5.989637546796396e-06, "loss": 0.5217, "step": 10818 }, { "epoch": 0.45351749578193934, "grad_norm": 1.539696455001831, "learning_rate": 5.988972115347141e-06, "loss": 0.4949, "step": 10819 }, { "epoch": 0.4535594143969483, "grad_norm": 2.0267891883850098, "learning_rate": 5.988306665667549e-06, "loss": 0.4893, "step": 10820 }, { "epoch": 0.45360133301195726, "grad_norm": 1.7560189962387085, "learning_rate": 5.987641197769892e-06, "loss": 0.5074, "step": 10821 }, { "epoch": 0.4536432516269662, "grad_norm": 1.6834663152694702, "learning_rate": 5.986975711666434e-06, "loss": 0.5065, "step": 10822 }, { "epoch": 0.4536851702419752, "grad_norm": 1.712817907333374, "learning_rate": 5.9863102073694445e-06, "loss": 0.5206, "step": 10823 }, { "epoch": 0.45372708885698415, "grad_norm": 2.164672613143921, "learning_rate": 5.9856446848911886e-06, "loss": 0.5288, "step": 10824 }, { "epoch": 0.4537690074719931, "grad_norm": 1.8219892978668213, "learning_rate": 5.984979144243935e-06, "loss": 0.5191, "step": 10825 }, { "epoch": 0.4538109260870021, "grad_norm": 2.0772640705108643, "learning_rate": 5.984313585439954e-06, "loss": 0.5628, "step": 10826 }, { "epoch": 0.45385284470201104, "grad_norm": 1.7870630025863647, "learning_rate": 5.9836480084915115e-06, "loss": 0.5187, "step": 10827 }, { "epoch": 0.45389476331702, "grad_norm": 1.6460862159729004, "learning_rate": 5.982982413410879e-06, "loss": 0.5655, "step": 10828 }, { "epoch": 0.45393668193202896, "grad_norm": 1.8367506265640259, "learning_rate": 5.982316800210324e-06, "loss": 0.5189, "step": 10829 }, { "epoch": 0.4539786005470379, "grad_norm": 1.7080048322677612, "learning_rate": 5.981651168902117e-06, "loss": 0.5301, "step": 10830 }, { "epoch": 0.4540205191620469, "grad_norm": 1.9249720573425293, "learning_rate": 5.980985519498527e-06, "loss": 0.5154, "step": 10831 }, { "epoch": 0.45406243777705585, "grad_norm": 1.9420191049575806, "learning_rate": 5.980319852011827e-06, "loss": 0.493, "step": 10832 }, { "epoch": 0.4541043563920648, "grad_norm": 1.7243479490280151, "learning_rate": 5.9796541664542854e-06, "loss": 0.5314, "step": 10833 }, { "epoch": 0.4541462750070738, "grad_norm": 1.9086487293243408, "learning_rate": 5.978988462838171e-06, "loss": 0.49, "step": 10834 }, { "epoch": 0.45418819362208274, "grad_norm": 1.809960961341858, "learning_rate": 5.9783227411757596e-06, "loss": 0.4327, "step": 10835 }, { "epoch": 0.4542301122370917, "grad_norm": 1.8133641481399536, "learning_rate": 5.977657001479321e-06, "loss": 0.5205, "step": 10836 }, { "epoch": 0.45427203085210066, "grad_norm": 1.9333802461624146, "learning_rate": 5.976991243761128e-06, "loss": 0.519, "step": 10837 }, { "epoch": 0.4543139494671096, "grad_norm": 1.7045116424560547, "learning_rate": 5.9763254680334505e-06, "loss": 0.4821, "step": 10838 }, { "epoch": 0.4543558680821186, "grad_norm": 2.3886232376098633, "learning_rate": 5.975659674308563e-06, "loss": 0.4974, "step": 10839 }, { "epoch": 0.45439778669712755, "grad_norm": 1.5553433895111084, "learning_rate": 5.974993862598738e-06, "loss": 0.4583, "step": 10840 }, { "epoch": 0.4544397053121365, "grad_norm": 1.7827717065811157, "learning_rate": 5.97432803291625e-06, "loss": 0.4723, "step": 10841 }, { "epoch": 0.4544816239271455, "grad_norm": 1.897002935409546, "learning_rate": 5.973662185273372e-06, "loss": 0.4391, "step": 10842 }, { "epoch": 0.4545235425421544, "grad_norm": 2.1631813049316406, "learning_rate": 5.972996319682376e-06, "loss": 0.5631, "step": 10843 }, { "epoch": 0.45456546115716334, "grad_norm": 1.6434117555618286, "learning_rate": 5.972330436155539e-06, "loss": 0.5016, "step": 10844 }, { "epoch": 0.4546073797721723, "grad_norm": 2.146681308746338, "learning_rate": 5.971664534705135e-06, "loss": 0.4903, "step": 10845 }, { "epoch": 0.45464929838718127, "grad_norm": 2.136504650115967, "learning_rate": 5.970998615343439e-06, "loss": 0.5704, "step": 10846 }, { "epoch": 0.45469121700219023, "grad_norm": 1.6552155017852783, "learning_rate": 5.970332678082724e-06, "loss": 0.454, "step": 10847 }, { "epoch": 0.4547331356171992, "grad_norm": 1.7244361639022827, "learning_rate": 5.969666722935269e-06, "loss": 0.4854, "step": 10848 }, { "epoch": 0.45477505423220815, "grad_norm": 1.5472807884216309, "learning_rate": 5.969000749913347e-06, "loss": 0.5134, "step": 10849 }, { "epoch": 0.4548169728472171, "grad_norm": 1.600113034248352, "learning_rate": 5.968334759029236e-06, "loss": 0.5196, "step": 10850 }, { "epoch": 0.4548588914622261, "grad_norm": 1.6253825426101685, "learning_rate": 5.967668750295214e-06, "loss": 0.5288, "step": 10851 }, { "epoch": 0.45490081007723504, "grad_norm": 1.5411736965179443, "learning_rate": 5.967002723723556e-06, "loss": 0.4905, "step": 10852 }, { "epoch": 0.454942728692244, "grad_norm": 1.5297152996063232, "learning_rate": 5.9663366793265385e-06, "loss": 0.486, "step": 10853 }, { "epoch": 0.45498464730725297, "grad_norm": 1.8897136449813843, "learning_rate": 5.9656706171164416e-06, "loss": 0.5247, "step": 10854 }, { "epoch": 0.4550265659222619, "grad_norm": 1.4774669408798218, "learning_rate": 5.965004537105541e-06, "loss": 0.4907, "step": 10855 }, { "epoch": 0.4550684845372709, "grad_norm": 1.6858811378479004, "learning_rate": 5.964338439306117e-06, "loss": 0.4794, "step": 10856 }, { "epoch": 0.45511040315227985, "grad_norm": 2.0192291736602783, "learning_rate": 5.963672323730445e-06, "loss": 0.5083, "step": 10857 }, { "epoch": 0.4551523217672888, "grad_norm": 1.74862539768219, "learning_rate": 5.963006190390808e-06, "loss": 0.5143, "step": 10858 }, { "epoch": 0.4551942403822978, "grad_norm": 1.8471791744232178, "learning_rate": 5.96234003929948e-06, "loss": 0.5358, "step": 10859 }, { "epoch": 0.45523615899730674, "grad_norm": 1.5956850051879883, "learning_rate": 5.9616738704687475e-06, "loss": 0.5079, "step": 10860 }, { "epoch": 0.4552780776123157, "grad_norm": 1.754286766052246, "learning_rate": 5.961007683910886e-06, "loss": 0.5105, "step": 10861 }, { "epoch": 0.45531999622732466, "grad_norm": 1.6948069334030151, "learning_rate": 5.960341479638174e-06, "loss": 0.5026, "step": 10862 }, { "epoch": 0.4553619148423336, "grad_norm": 1.6575835943222046, "learning_rate": 5.959675257662896e-06, "loss": 0.5375, "step": 10863 }, { "epoch": 0.4554038334573426, "grad_norm": 1.7886464595794678, "learning_rate": 5.959009017997331e-06, "loss": 0.5427, "step": 10864 }, { "epoch": 0.45544575207235155, "grad_norm": 1.879755973815918, "learning_rate": 5.9583427606537615e-06, "loss": 0.5273, "step": 10865 }, { "epoch": 0.4554876706873605, "grad_norm": 1.8275580406188965, "learning_rate": 5.957676485644468e-06, "loss": 0.4947, "step": 10866 }, { "epoch": 0.4555295893023695, "grad_norm": 1.5381312370300293, "learning_rate": 5.957010192981731e-06, "loss": 0.5377, "step": 10867 }, { "epoch": 0.4555715079173784, "grad_norm": 1.9346369504928589, "learning_rate": 5.9563438826778345e-06, "loss": 0.5165, "step": 10868 }, { "epoch": 0.45561342653238734, "grad_norm": 2.0641257762908936, "learning_rate": 5.9556775547450604e-06, "loss": 0.48, "step": 10869 }, { "epoch": 0.4556553451473963, "grad_norm": 1.6382378339767456, "learning_rate": 5.9550112091956935e-06, "loss": 0.4976, "step": 10870 }, { "epoch": 0.45569726376240527, "grad_norm": 1.9343849420547485, "learning_rate": 5.954344846042014e-06, "loss": 0.4946, "step": 10871 }, { "epoch": 0.45573918237741423, "grad_norm": 1.6453821659088135, "learning_rate": 5.953678465296306e-06, "loss": 0.513, "step": 10872 }, { "epoch": 0.4557811009924232, "grad_norm": 1.8896609544754028, "learning_rate": 5.9530120669708556e-06, "loss": 0.5035, "step": 10873 }, { "epoch": 0.45582301960743216, "grad_norm": 1.7070293426513672, "learning_rate": 5.952345651077944e-06, "loss": 0.4956, "step": 10874 }, { "epoch": 0.4558649382224411, "grad_norm": 1.643794059753418, "learning_rate": 5.9516792176298575e-06, "loss": 0.5386, "step": 10875 }, { "epoch": 0.4559068568374501, "grad_norm": 1.6268870830535889, "learning_rate": 5.951012766638881e-06, "loss": 0.459, "step": 10876 }, { "epoch": 0.45594877545245904, "grad_norm": 2.069342851638794, "learning_rate": 5.950346298117297e-06, "loss": 0.5377, "step": 10877 }, { "epoch": 0.455990694067468, "grad_norm": 1.6319314241409302, "learning_rate": 5.949679812077393e-06, "loss": 0.5062, "step": 10878 }, { "epoch": 0.45603261268247697, "grad_norm": 1.727251648902893, "learning_rate": 5.949013308531456e-06, "loss": 0.4726, "step": 10879 }, { "epoch": 0.45607453129748593, "grad_norm": 1.750477910041809, "learning_rate": 5.94834678749177e-06, "loss": 0.448, "step": 10880 }, { "epoch": 0.4561164499124949, "grad_norm": 2.0282645225524902, "learning_rate": 5.947680248970622e-06, "loss": 0.5864, "step": 10881 }, { "epoch": 0.45615836852750385, "grad_norm": 2.15922474861145, "learning_rate": 5.947013692980299e-06, "loss": 0.4543, "step": 10882 }, { "epoch": 0.4562002871425128, "grad_norm": 1.8052469491958618, "learning_rate": 5.946347119533087e-06, "loss": 0.5398, "step": 10883 }, { "epoch": 0.4562422057575218, "grad_norm": 2.2815139293670654, "learning_rate": 5.945680528641275e-06, "loss": 0.5245, "step": 10884 }, { "epoch": 0.45628412437253074, "grad_norm": 1.776921033859253, "learning_rate": 5.94501392031715e-06, "loss": 0.5096, "step": 10885 }, { "epoch": 0.4563260429875397, "grad_norm": 1.6758908033370972, "learning_rate": 5.944347294572998e-06, "loss": 0.5014, "step": 10886 }, { "epoch": 0.45636796160254867, "grad_norm": 1.667457103729248, "learning_rate": 5.94368065142111e-06, "loss": 0.5175, "step": 10887 }, { "epoch": 0.45640988021755763, "grad_norm": 1.7647947072982788, "learning_rate": 5.943013990873775e-06, "loss": 0.5096, "step": 10888 }, { "epoch": 0.4564517988325666, "grad_norm": 2.102578639984131, "learning_rate": 5.942347312943281e-06, "loss": 0.5084, "step": 10889 }, { "epoch": 0.45649371744757555, "grad_norm": 1.8910313844680786, "learning_rate": 5.941680617641915e-06, "loss": 0.5057, "step": 10890 }, { "epoch": 0.4565356360625845, "grad_norm": 11.342351913452148, "learning_rate": 5.9410139049819695e-06, "loss": 0.4778, "step": 10891 }, { "epoch": 0.4565775546775935, "grad_norm": 2.505539655685425, "learning_rate": 5.940347174975734e-06, "loss": 0.475, "step": 10892 }, { "epoch": 0.4566194732926024, "grad_norm": 1.752254605293274, "learning_rate": 5.939680427635498e-06, "loss": 0.4943, "step": 10893 }, { "epoch": 0.45666139190761135, "grad_norm": 1.9536044597625732, "learning_rate": 5.939013662973552e-06, "loss": 0.461, "step": 10894 }, { "epoch": 0.4567033105226203, "grad_norm": 1.8435629606246948, "learning_rate": 5.938346881002188e-06, "loss": 0.5015, "step": 10895 }, { "epoch": 0.45674522913762927, "grad_norm": 1.576762080192566, "learning_rate": 5.937680081733696e-06, "loss": 0.5434, "step": 10896 }, { "epoch": 0.45678714775263823, "grad_norm": 2.1572813987731934, "learning_rate": 5.937013265180367e-06, "loss": 0.5819, "step": 10897 }, { "epoch": 0.4568290663676472, "grad_norm": 1.7865036725997925, "learning_rate": 5.936346431354496e-06, "loss": 0.4842, "step": 10898 }, { "epoch": 0.45687098498265616, "grad_norm": 1.957406997680664, "learning_rate": 5.935679580268372e-06, "loss": 0.5588, "step": 10899 }, { "epoch": 0.4569129035976651, "grad_norm": 1.7846715450286865, "learning_rate": 5.935012711934288e-06, "loss": 0.5388, "step": 10900 }, { "epoch": 0.4569548222126741, "grad_norm": 1.8855922222137451, "learning_rate": 5.9343458263645385e-06, "loss": 0.5198, "step": 10901 }, { "epoch": 0.45699674082768305, "grad_norm": 1.773452877998352, "learning_rate": 5.933678923571413e-06, "loss": 0.4579, "step": 10902 }, { "epoch": 0.457038659442692, "grad_norm": 2.2222132682800293, "learning_rate": 5.93301200356721e-06, "loss": 0.5185, "step": 10903 }, { "epoch": 0.45708057805770097, "grad_norm": 1.5579912662506104, "learning_rate": 5.9323450663642205e-06, "loss": 0.5218, "step": 10904 }, { "epoch": 0.45712249667270993, "grad_norm": 1.6795804500579834, "learning_rate": 5.931678111974736e-06, "loss": 0.4706, "step": 10905 }, { "epoch": 0.4571644152877189, "grad_norm": 2.333580255508423, "learning_rate": 5.931011140411055e-06, "loss": 0.5348, "step": 10906 }, { "epoch": 0.45720633390272786, "grad_norm": 2.219932794570923, "learning_rate": 5.930344151685471e-06, "loss": 0.496, "step": 10907 }, { "epoch": 0.4572482525177368, "grad_norm": 2.7644128799438477, "learning_rate": 5.929677145810278e-06, "loss": 0.506, "step": 10908 }, { "epoch": 0.4572901711327458, "grad_norm": 1.8316020965576172, "learning_rate": 5.929010122797773e-06, "loss": 0.4829, "step": 10909 }, { "epoch": 0.45733208974775474, "grad_norm": 1.9618033170700073, "learning_rate": 5.928343082660251e-06, "loss": 0.4806, "step": 10910 }, { "epoch": 0.4573740083627637, "grad_norm": 1.6758567094802856, "learning_rate": 5.927676025410006e-06, "loss": 0.5081, "step": 10911 }, { "epoch": 0.45741592697777267, "grad_norm": 2.674443244934082, "learning_rate": 5.927008951059339e-06, "loss": 0.4746, "step": 10912 }, { "epoch": 0.45745784559278163, "grad_norm": 1.7458446025848389, "learning_rate": 5.926341859620541e-06, "loss": 0.4884, "step": 10913 }, { "epoch": 0.4574997642077906, "grad_norm": 1.683703064918518, "learning_rate": 5.925674751105911e-06, "loss": 0.5007, "step": 10914 }, { "epoch": 0.45754168282279956, "grad_norm": 2.625126600265503, "learning_rate": 5.925007625527748e-06, "loss": 0.4815, "step": 10915 }, { "epoch": 0.4575836014378085, "grad_norm": 1.7099179029464722, "learning_rate": 5.924340482898347e-06, "loss": 0.5691, "step": 10916 }, { "epoch": 0.4576255200528175, "grad_norm": 1.8700644969940186, "learning_rate": 5.923673323230008e-06, "loss": 0.5103, "step": 10917 }, { "epoch": 0.4576674386678264, "grad_norm": 2.4729621410369873, "learning_rate": 5.923006146535028e-06, "loss": 0.5755, "step": 10918 }, { "epoch": 0.45770935728283535, "grad_norm": 1.6325408220291138, "learning_rate": 5.922338952825707e-06, "loss": 0.516, "step": 10919 }, { "epoch": 0.4577512758978443, "grad_norm": 1.5870250463485718, "learning_rate": 5.921671742114341e-06, "loss": 0.4535, "step": 10920 }, { "epoch": 0.4577931945128533, "grad_norm": 1.851684331893921, "learning_rate": 5.92100451441323e-06, "loss": 0.4741, "step": 10921 }, { "epoch": 0.45783511312786224, "grad_norm": 2.2851085662841797, "learning_rate": 5.920337269734676e-06, "loss": 0.4512, "step": 10922 }, { "epoch": 0.4578770317428712, "grad_norm": 2.2558135986328125, "learning_rate": 5.919670008090976e-06, "loss": 0.5138, "step": 10923 }, { "epoch": 0.45791895035788016, "grad_norm": 2.3039355278015137, "learning_rate": 5.91900272949443e-06, "loss": 0.5025, "step": 10924 }, { "epoch": 0.4579608689728891, "grad_norm": 1.5088247060775757, "learning_rate": 5.9183354339573385e-06, "loss": 0.4969, "step": 10925 }, { "epoch": 0.4580027875878981, "grad_norm": 2.241715908050537, "learning_rate": 5.917668121492005e-06, "loss": 0.5141, "step": 10926 }, { "epoch": 0.45804470620290705, "grad_norm": 1.8103687763214111, "learning_rate": 5.917000792110727e-06, "loss": 0.5004, "step": 10927 }, { "epoch": 0.458086624817916, "grad_norm": 2.270285129547119, "learning_rate": 5.9163334458258085e-06, "loss": 0.5302, "step": 10928 }, { "epoch": 0.458128543432925, "grad_norm": 1.6684114933013916, "learning_rate": 5.9156660826495484e-06, "loss": 0.4882, "step": 10929 }, { "epoch": 0.45817046204793394, "grad_norm": 2.000115394592285, "learning_rate": 5.914998702594249e-06, "loss": 0.4887, "step": 10930 }, { "epoch": 0.4582123806629429, "grad_norm": 1.599973440170288, "learning_rate": 5.914331305672215e-06, "loss": 0.4727, "step": 10931 }, { "epoch": 0.45825429927795186, "grad_norm": 1.7606908082962036, "learning_rate": 5.913663891895748e-06, "loss": 0.4848, "step": 10932 }, { "epoch": 0.4582962178929608, "grad_norm": 1.713765263557434, "learning_rate": 5.912996461277148e-06, "loss": 0.4729, "step": 10933 }, { "epoch": 0.4583381365079698, "grad_norm": 1.7883440256118774, "learning_rate": 5.912329013828721e-06, "loss": 0.5004, "step": 10934 }, { "epoch": 0.45838005512297875, "grad_norm": 2.1105058193206787, "learning_rate": 5.911661549562772e-06, "loss": 0.5356, "step": 10935 }, { "epoch": 0.4584219737379877, "grad_norm": 2.0552945137023926, "learning_rate": 5.910994068491601e-06, "loss": 0.5348, "step": 10936 }, { "epoch": 0.45846389235299667, "grad_norm": 3.0336575508117676, "learning_rate": 5.910326570627514e-06, "loss": 0.4645, "step": 10937 }, { "epoch": 0.45850581096800563, "grad_norm": 1.7130988836288452, "learning_rate": 5.909659055982815e-06, "loss": 0.4466, "step": 10938 }, { "epoch": 0.4585477295830146, "grad_norm": 2.6117827892303467, "learning_rate": 5.908991524569808e-06, "loss": 0.4874, "step": 10939 }, { "epoch": 0.45858964819802356, "grad_norm": 1.5614382028579712, "learning_rate": 5.908323976400802e-06, "loss": 0.438, "step": 10940 }, { "epoch": 0.4586315668130325, "grad_norm": 1.820915699005127, "learning_rate": 5.907656411488097e-06, "loss": 0.473, "step": 10941 }, { "epoch": 0.4586734854280415, "grad_norm": 1.8214573860168457, "learning_rate": 5.906988829844e-06, "loss": 0.4666, "step": 10942 }, { "epoch": 0.45871540404305045, "grad_norm": 1.6441457271575928, "learning_rate": 5.906321231480819e-06, "loss": 0.5136, "step": 10943 }, { "epoch": 0.45875732265805935, "grad_norm": 1.94681715965271, "learning_rate": 5.905653616410859e-06, "loss": 0.5427, "step": 10944 }, { "epoch": 0.4587992412730683, "grad_norm": 1.9921821355819702, "learning_rate": 5.904985984646426e-06, "loss": 0.508, "step": 10945 }, { "epoch": 0.4588411598880773, "grad_norm": 1.8940821886062622, "learning_rate": 5.904318336199828e-06, "loss": 0.5378, "step": 10946 }, { "epoch": 0.45888307850308624, "grad_norm": 2.000110387802124, "learning_rate": 5.903650671083372e-06, "loss": 0.5625, "step": 10947 }, { "epoch": 0.4589249971180952, "grad_norm": 1.6410284042358398, "learning_rate": 5.902982989309364e-06, "loss": 0.5268, "step": 10948 }, { "epoch": 0.45896691573310416, "grad_norm": 2.2806036472320557, "learning_rate": 5.9023152908901135e-06, "loss": 0.4637, "step": 10949 }, { "epoch": 0.4590088343481131, "grad_norm": 2.4908082485198975, "learning_rate": 5.90164757583793e-06, "loss": 0.5422, "step": 10950 }, { "epoch": 0.4590507529631221, "grad_norm": 1.719840407371521, "learning_rate": 5.9009798441651164e-06, "loss": 0.441, "step": 10951 }, { "epoch": 0.45909267157813105, "grad_norm": 1.5863771438598633, "learning_rate": 5.900312095883988e-06, "loss": 0.506, "step": 10952 }, { "epoch": 0.45913459019314, "grad_norm": 1.6568959951400757, "learning_rate": 5.8996443310068485e-06, "loss": 0.552, "step": 10953 }, { "epoch": 0.459176508808149, "grad_norm": 1.6504405736923218, "learning_rate": 5.89897654954601e-06, "loss": 0.4638, "step": 10954 }, { "epoch": 0.45921842742315794, "grad_norm": 1.807803988456726, "learning_rate": 5.898308751513783e-06, "loss": 0.4827, "step": 10955 }, { "epoch": 0.4592603460381669, "grad_norm": 1.7686994075775146, "learning_rate": 5.897640936922475e-06, "loss": 0.4717, "step": 10956 }, { "epoch": 0.45930226465317586, "grad_norm": 2.047243356704712, "learning_rate": 5.896973105784398e-06, "loss": 0.52, "step": 10957 }, { "epoch": 0.4593441832681848, "grad_norm": 1.9058791399002075, "learning_rate": 5.89630525811186e-06, "loss": 0.5429, "step": 10958 }, { "epoch": 0.4593861018831938, "grad_norm": 1.5901570320129395, "learning_rate": 5.895637393917176e-06, "loss": 0.5126, "step": 10959 }, { "epoch": 0.45942802049820275, "grad_norm": 1.7912898063659668, "learning_rate": 5.894969513212654e-06, "loss": 0.4646, "step": 10960 }, { "epoch": 0.4594699391132117, "grad_norm": 1.6795417070388794, "learning_rate": 5.894301616010604e-06, "loss": 0.507, "step": 10961 }, { "epoch": 0.4595118577282207, "grad_norm": 1.8088053464889526, "learning_rate": 5.893633702323342e-06, "loss": 0.4632, "step": 10962 }, { "epoch": 0.45955377634322964, "grad_norm": 1.851714015007019, "learning_rate": 5.892965772163179e-06, "loss": 0.5036, "step": 10963 }, { "epoch": 0.4595956949582386, "grad_norm": 2.1505823135375977, "learning_rate": 5.892297825542426e-06, "loss": 0.5452, "step": 10964 }, { "epoch": 0.45963761357324756, "grad_norm": 1.8281668424606323, "learning_rate": 5.8916298624733954e-06, "loss": 0.5042, "step": 10965 }, { "epoch": 0.4596795321882565, "grad_norm": 1.6521580219268799, "learning_rate": 5.890961882968401e-06, "loss": 0.459, "step": 10966 }, { "epoch": 0.4597214508032655, "grad_norm": 1.7442177534103394, "learning_rate": 5.890293887039755e-06, "loss": 0.5397, "step": 10967 }, { "epoch": 0.45976336941827445, "grad_norm": 1.7239428758621216, "learning_rate": 5.889625874699774e-06, "loss": 0.4557, "step": 10968 }, { "epoch": 0.45980528803328335, "grad_norm": 1.8207443952560425, "learning_rate": 5.88895784596077e-06, "loss": 0.4858, "step": 10969 }, { "epoch": 0.4598472066482923, "grad_norm": 2.3514623641967773, "learning_rate": 5.888289800835055e-06, "loss": 0.5299, "step": 10970 }, { "epoch": 0.4598891252633013, "grad_norm": 1.5800151824951172, "learning_rate": 5.887621739334948e-06, "loss": 0.4723, "step": 10971 }, { "epoch": 0.45993104387831024, "grad_norm": 1.6602123975753784, "learning_rate": 5.88695366147276e-06, "loss": 0.5117, "step": 10972 }, { "epoch": 0.4599729624933192, "grad_norm": 1.7456414699554443, "learning_rate": 5.886285567260806e-06, "loss": 0.469, "step": 10973 }, { "epoch": 0.46001488110832817, "grad_norm": 1.9327844381332397, "learning_rate": 5.885617456711406e-06, "loss": 0.4941, "step": 10974 }, { "epoch": 0.46005679972333713, "grad_norm": 1.6432427167892456, "learning_rate": 5.88494932983687e-06, "loss": 0.4775, "step": 10975 }, { "epoch": 0.4600987183383461, "grad_norm": 1.5988943576812744, "learning_rate": 5.884281186649517e-06, "loss": 0.4949, "step": 10976 }, { "epoch": 0.46014063695335505, "grad_norm": 3.278093099594116, "learning_rate": 5.883613027161663e-06, "loss": 0.5065, "step": 10977 }, { "epoch": 0.460182555568364, "grad_norm": 1.7219370603561401, "learning_rate": 5.8829448513856255e-06, "loss": 0.4957, "step": 10978 }, { "epoch": 0.460224474183373, "grad_norm": 1.9154599905014038, "learning_rate": 5.882276659333719e-06, "loss": 0.5329, "step": 10979 }, { "epoch": 0.46026639279838194, "grad_norm": 1.86516273021698, "learning_rate": 5.881608451018261e-06, "loss": 0.5429, "step": 10980 }, { "epoch": 0.4603083114133909, "grad_norm": 2.244011640548706, "learning_rate": 5.880940226451573e-06, "loss": 0.4942, "step": 10981 }, { "epoch": 0.46035023002839986, "grad_norm": 1.9196726083755493, "learning_rate": 5.880271985645967e-06, "loss": 0.4832, "step": 10982 }, { "epoch": 0.4603921486434088, "grad_norm": 1.9355840682983398, "learning_rate": 5.8796037286137656e-06, "loss": 0.5223, "step": 10983 }, { "epoch": 0.4604340672584178, "grad_norm": 1.881309986114502, "learning_rate": 5.878935455367285e-06, "loss": 0.5133, "step": 10984 }, { "epoch": 0.46047598587342675, "grad_norm": 1.6710911989212036, "learning_rate": 5.878267165918843e-06, "loss": 0.4868, "step": 10985 }, { "epoch": 0.4605179044884357, "grad_norm": 3.1384084224700928, "learning_rate": 5.877598860280762e-06, "loss": 0.5216, "step": 10986 }, { "epoch": 0.4605598231034447, "grad_norm": 1.9481422901153564, "learning_rate": 5.87693053846536e-06, "loss": 0.5191, "step": 10987 }, { "epoch": 0.46060174171845364, "grad_norm": 2.1623964309692383, "learning_rate": 5.876262200484953e-06, "loss": 0.5113, "step": 10988 }, { "epoch": 0.4606436603334626, "grad_norm": 12.93856430053711, "learning_rate": 5.875593846351864e-06, "loss": 0.5409, "step": 10989 }, { "epoch": 0.46068557894847156, "grad_norm": 2.3754453659057617, "learning_rate": 5.874925476078415e-06, "loss": 0.5333, "step": 10990 }, { "epoch": 0.4607274975634805, "grad_norm": 1.64780855178833, "learning_rate": 5.874257089676924e-06, "loss": 0.5422, "step": 10991 }, { "epoch": 0.4607694161784895, "grad_norm": 1.843213438987732, "learning_rate": 5.873588687159711e-06, "loss": 0.5185, "step": 10992 }, { "epoch": 0.46081133479349845, "grad_norm": 1.6564573049545288, "learning_rate": 5.872920268539099e-06, "loss": 0.5002, "step": 10993 }, { "epoch": 0.46085325340850736, "grad_norm": 1.7063020467758179, "learning_rate": 5.872251833827408e-06, "loss": 0.5157, "step": 10994 }, { "epoch": 0.4608951720235163, "grad_norm": 2.079540729522705, "learning_rate": 5.87158338303696e-06, "loss": 0.5179, "step": 10995 }, { "epoch": 0.4609370906385253, "grad_norm": 1.641170620918274, "learning_rate": 5.870914916180079e-06, "loss": 0.4642, "step": 10996 }, { "epoch": 0.46097900925353424, "grad_norm": 1.5575040578842163, "learning_rate": 5.870246433269085e-06, "loss": 0.4562, "step": 10997 }, { "epoch": 0.4610209278685432, "grad_norm": 2.8087334632873535, "learning_rate": 5.869577934316299e-06, "loss": 0.5179, "step": 10998 }, { "epoch": 0.46106284648355217, "grad_norm": 1.5498697757720947, "learning_rate": 5.868909419334048e-06, "loss": 0.5313, "step": 10999 }, { "epoch": 0.46110476509856113, "grad_norm": 1.6415334939956665, "learning_rate": 5.8682408883346535e-06, "loss": 0.502, "step": 11000 }, { "epoch": 0.4611466837135701, "grad_norm": 2.8248586654663086, "learning_rate": 5.867572341330437e-06, "loss": 0.5286, "step": 11001 }, { "epoch": 0.46118860232857906, "grad_norm": 2.1667935848236084, "learning_rate": 5.866903778333725e-06, "loss": 0.454, "step": 11002 }, { "epoch": 0.461230520943588, "grad_norm": 1.8264658451080322, "learning_rate": 5.866235199356839e-06, "loss": 0.5154, "step": 11003 }, { "epoch": 0.461272439558597, "grad_norm": 1.765782117843628, "learning_rate": 5.865566604412104e-06, "loss": 0.5296, "step": 11004 }, { "epoch": 0.46131435817360594, "grad_norm": 1.9178433418273926, "learning_rate": 5.864897993511847e-06, "loss": 0.4861, "step": 11005 }, { "epoch": 0.4613562767886149, "grad_norm": 1.927286982536316, "learning_rate": 5.8642293666683926e-06, "loss": 0.5055, "step": 11006 }, { "epoch": 0.46139819540362387, "grad_norm": 1.8772021532058716, "learning_rate": 5.863560723894062e-06, "loss": 0.4857, "step": 11007 }, { "epoch": 0.46144011401863283, "grad_norm": 1.7997627258300781, "learning_rate": 5.862892065201184e-06, "loss": 0.4707, "step": 11008 }, { "epoch": 0.4614820326336418, "grad_norm": 1.8448028564453125, "learning_rate": 5.862223390602083e-06, "loss": 0.6106, "step": 11009 }, { "epoch": 0.46152395124865075, "grad_norm": 1.740034580230713, "learning_rate": 5.861554700109086e-06, "loss": 0.5119, "step": 11010 }, { "epoch": 0.4615658698636597, "grad_norm": 1.8780640363693237, "learning_rate": 5.860885993734521e-06, "loss": 0.5845, "step": 11011 }, { "epoch": 0.4616077884786687, "grad_norm": 1.706851601600647, "learning_rate": 5.8602172714907115e-06, "loss": 0.4977, "step": 11012 }, { "epoch": 0.46164970709367764, "grad_norm": 1.941185474395752, "learning_rate": 5.859548533389985e-06, "loss": 0.5385, "step": 11013 }, { "epoch": 0.4616916257086866, "grad_norm": 2.1418607234954834, "learning_rate": 5.858879779444671e-06, "loss": 0.5143, "step": 11014 }, { "epoch": 0.46173354432369557, "grad_norm": 1.8439509868621826, "learning_rate": 5.858211009667094e-06, "loss": 0.447, "step": 11015 }, { "epoch": 0.46177546293870453, "grad_norm": 1.6258413791656494, "learning_rate": 5.857542224069585e-06, "loss": 0.4785, "step": 11016 }, { "epoch": 0.4618173815537135, "grad_norm": 2.017202854156494, "learning_rate": 5.85687342266447e-06, "loss": 0.5085, "step": 11017 }, { "epoch": 0.46185930016872245, "grad_norm": 1.7400941848754883, "learning_rate": 5.856204605464079e-06, "loss": 0.5014, "step": 11018 }, { "epoch": 0.46190121878373136, "grad_norm": 1.9245363473892212, "learning_rate": 5.855535772480737e-06, "loss": 0.5314, "step": 11019 }, { "epoch": 0.4619431373987403, "grad_norm": 1.9343937635421753, "learning_rate": 5.85486692372678e-06, "loss": 0.5137, "step": 11020 }, { "epoch": 0.4619850560137493, "grad_norm": 2.1938045024871826, "learning_rate": 5.854198059214531e-06, "loss": 0.499, "step": 11021 }, { "epoch": 0.46202697462875825, "grad_norm": 1.7698999643325806, "learning_rate": 5.853529178956319e-06, "loss": 0.5186, "step": 11022 }, { "epoch": 0.4620688932437672, "grad_norm": 1.7378305196762085, "learning_rate": 5.85286028296448e-06, "loss": 0.5397, "step": 11023 }, { "epoch": 0.46211081185877617, "grad_norm": 1.7000843286514282, "learning_rate": 5.85219137125134e-06, "loss": 0.5019, "step": 11024 }, { "epoch": 0.46215273047378513, "grad_norm": 2.106433629989624, "learning_rate": 5.851522443829229e-06, "loss": 0.5255, "step": 11025 }, { "epoch": 0.4621946490887941, "grad_norm": 1.8416599035263062, "learning_rate": 5.850853500710481e-06, "loss": 0.5763, "step": 11026 }, { "epoch": 0.46223656770380306, "grad_norm": 1.677040934562683, "learning_rate": 5.850184541907425e-06, "loss": 0.5099, "step": 11027 }, { "epoch": 0.462278486318812, "grad_norm": 1.870174527168274, "learning_rate": 5.849515567432389e-06, "loss": 0.4608, "step": 11028 }, { "epoch": 0.462320404933821, "grad_norm": 1.8425003290176392, "learning_rate": 5.848846577297711e-06, "loss": 0.4934, "step": 11029 }, { "epoch": 0.46236232354882995, "grad_norm": 1.7293145656585693, "learning_rate": 5.84817757151572e-06, "loss": 0.536, "step": 11030 }, { "epoch": 0.4624042421638389, "grad_norm": 2.1007497310638428, "learning_rate": 5.8475085500987465e-06, "loss": 0.5145, "step": 11031 }, { "epoch": 0.46244616077884787, "grad_norm": 2.3100452423095703, "learning_rate": 5.846839513059124e-06, "loss": 0.4667, "step": 11032 }, { "epoch": 0.46248807939385683, "grad_norm": 1.5991623401641846, "learning_rate": 5.846170460409187e-06, "loss": 0.5271, "step": 11033 }, { "epoch": 0.4625299980088658, "grad_norm": 1.903303623199463, "learning_rate": 5.845501392161268e-06, "loss": 0.5001, "step": 11034 }, { "epoch": 0.46257191662387476, "grad_norm": 1.8529080152511597, "learning_rate": 5.844832308327698e-06, "loss": 0.4622, "step": 11035 }, { "epoch": 0.4626138352388837, "grad_norm": 1.9043015241622925, "learning_rate": 5.844163208920813e-06, "loss": 0.5228, "step": 11036 }, { "epoch": 0.4626557538538927, "grad_norm": 1.654193639755249, "learning_rate": 5.843494093952947e-06, "loss": 0.4771, "step": 11037 }, { "epoch": 0.46269767246890164, "grad_norm": 1.7206413745880127, "learning_rate": 5.842824963436433e-06, "loss": 0.4412, "step": 11038 }, { "epoch": 0.4627395910839106, "grad_norm": 1.8643629550933838, "learning_rate": 5.842155817383607e-06, "loss": 0.4793, "step": 11039 }, { "epoch": 0.46278150969891957, "grad_norm": 1.7296924591064453, "learning_rate": 5.841486655806802e-06, "loss": 0.585, "step": 11040 }, { "epoch": 0.46282342831392853, "grad_norm": 1.674430012702942, "learning_rate": 5.840817478718351e-06, "loss": 0.5199, "step": 11041 }, { "epoch": 0.4628653469289375, "grad_norm": 1.9530495405197144, "learning_rate": 5.840148286130596e-06, "loss": 0.5251, "step": 11042 }, { "epoch": 0.46290726554394646, "grad_norm": 2.606013774871826, "learning_rate": 5.839479078055867e-06, "loss": 0.5055, "step": 11043 }, { "epoch": 0.46294918415895536, "grad_norm": 1.5354562997817993, "learning_rate": 5.838809854506503e-06, "loss": 0.4988, "step": 11044 }, { "epoch": 0.4629911027739643, "grad_norm": 1.7908583879470825, "learning_rate": 5.838140615494838e-06, "loss": 0.5729, "step": 11045 }, { "epoch": 0.4630330213889733, "grad_norm": 1.8106549978256226, "learning_rate": 5.83747136103321e-06, "loss": 0.533, "step": 11046 }, { "epoch": 0.46307494000398225, "grad_norm": 1.8201583623886108, "learning_rate": 5.836802091133954e-06, "loss": 0.4922, "step": 11047 }, { "epoch": 0.4631168586189912, "grad_norm": 1.6803990602493286, "learning_rate": 5.8361328058094105e-06, "loss": 0.5526, "step": 11048 }, { "epoch": 0.4631587772340002, "grad_norm": 1.6393474340438843, "learning_rate": 5.8354635050719135e-06, "loss": 0.5085, "step": 11049 }, { "epoch": 0.46320069584900914, "grad_norm": 1.7524797916412354, "learning_rate": 5.8347941889338e-06, "loss": 0.4929, "step": 11050 }, { "epoch": 0.4632426144640181, "grad_norm": 1.8859132528305054, "learning_rate": 5.8341248574074114e-06, "loss": 0.5041, "step": 11051 }, { "epoch": 0.46328453307902706, "grad_norm": 2.889209032058716, "learning_rate": 5.833455510505084e-06, "loss": 0.479, "step": 11052 }, { "epoch": 0.463326451694036, "grad_norm": 1.822129487991333, "learning_rate": 5.832786148239157e-06, "loss": 0.4779, "step": 11053 }, { "epoch": 0.463368370309045, "grad_norm": 1.7678594589233398, "learning_rate": 5.832116770621969e-06, "loss": 0.4915, "step": 11054 }, { "epoch": 0.46341028892405395, "grad_norm": 2.072977304458618, "learning_rate": 5.831447377665858e-06, "loss": 0.5742, "step": 11055 }, { "epoch": 0.4634522075390629, "grad_norm": 1.4938554763793945, "learning_rate": 5.830777969383162e-06, "loss": 0.4916, "step": 11056 }, { "epoch": 0.4634941261540719, "grad_norm": 3.583691358566284, "learning_rate": 5.8301085457862264e-06, "loss": 0.5146, "step": 11057 }, { "epoch": 0.46353604476908083, "grad_norm": 1.7838966846466064, "learning_rate": 5.8294391068873846e-06, "loss": 0.5177, "step": 11058 }, { "epoch": 0.4635779633840898, "grad_norm": 1.6406748294830322, "learning_rate": 5.8287696526989805e-06, "loss": 0.4681, "step": 11059 }, { "epoch": 0.46361988199909876, "grad_norm": 1.75115966796875, "learning_rate": 5.828100183233353e-06, "loss": 0.4971, "step": 11060 }, { "epoch": 0.4636618006141077, "grad_norm": 1.9366798400878906, "learning_rate": 5.827430698502844e-06, "loss": 0.5073, "step": 11061 }, { "epoch": 0.4637037192291167, "grad_norm": 1.8211252689361572, "learning_rate": 5.826761198519794e-06, "loss": 0.4691, "step": 11062 }, { "epoch": 0.46374563784412565, "grad_norm": 1.8859972953796387, "learning_rate": 5.826091683296544e-06, "loss": 0.4972, "step": 11063 }, { "epoch": 0.4637875564591346, "grad_norm": 1.720743179321289, "learning_rate": 5.825422152845437e-06, "loss": 0.4803, "step": 11064 }, { "epoch": 0.46382947507414357, "grad_norm": 1.9713830947875977, "learning_rate": 5.8247526071788115e-06, "loss": 0.5107, "step": 11065 }, { "epoch": 0.46387139368915253, "grad_norm": 1.8064261674880981, "learning_rate": 5.824083046309013e-06, "loss": 0.5116, "step": 11066 }, { "epoch": 0.4639133123041615, "grad_norm": 1.7330018281936646, "learning_rate": 5.823413470248385e-06, "loss": 0.5161, "step": 11067 }, { "epoch": 0.46395523091917046, "grad_norm": 1.7440377473831177, "learning_rate": 5.822743879009265e-06, "loss": 0.544, "step": 11068 }, { "epoch": 0.46399714953417937, "grad_norm": 1.629321813583374, "learning_rate": 5.822074272604e-06, "loss": 0.5236, "step": 11069 }, { "epoch": 0.4640390681491883, "grad_norm": 1.926373839378357, "learning_rate": 5.821404651044933e-06, "loss": 0.5216, "step": 11070 }, { "epoch": 0.4640809867641973, "grad_norm": 1.8771275281906128, "learning_rate": 5.8207350143444055e-06, "loss": 0.52, "step": 11071 }, { "epoch": 0.46412290537920625, "grad_norm": 1.8874496221542358, "learning_rate": 5.820065362514763e-06, "loss": 0.5933, "step": 11072 }, { "epoch": 0.4641648239942152, "grad_norm": 2.0827341079711914, "learning_rate": 5.81939569556835e-06, "loss": 0.5335, "step": 11073 }, { "epoch": 0.4642067426092242, "grad_norm": 8.140815734863281, "learning_rate": 5.818726013517509e-06, "loss": 0.4975, "step": 11074 }, { "epoch": 0.46424866122423314, "grad_norm": 1.8095979690551758, "learning_rate": 5.8180563163745865e-06, "loss": 0.4928, "step": 11075 }, { "epoch": 0.4642905798392421, "grad_norm": 1.8101129531860352, "learning_rate": 5.817386604151928e-06, "loss": 0.5036, "step": 11076 }, { "epoch": 0.46433249845425106, "grad_norm": 1.685025930404663, "learning_rate": 5.816716876861875e-06, "loss": 0.48, "step": 11077 }, { "epoch": 0.46437441706926, "grad_norm": 1.8919905424118042, "learning_rate": 5.816047134516776e-06, "loss": 0.4896, "step": 11078 }, { "epoch": 0.464416335684269, "grad_norm": 1.8195650577545166, "learning_rate": 5.815377377128977e-06, "loss": 0.5331, "step": 11079 }, { "epoch": 0.46445825429927795, "grad_norm": 1.944394826889038, "learning_rate": 5.814707604710823e-06, "loss": 0.5134, "step": 11080 }, { "epoch": 0.4645001729142869, "grad_norm": 1.9020874500274658, "learning_rate": 5.8140378172746615e-06, "loss": 0.5024, "step": 11081 }, { "epoch": 0.4645420915292959, "grad_norm": 1.634601354598999, "learning_rate": 5.813368014832836e-06, "loss": 0.5018, "step": 11082 }, { "epoch": 0.46458401014430484, "grad_norm": 1.904594898223877, "learning_rate": 5.8126981973976985e-06, "loss": 0.5271, "step": 11083 }, { "epoch": 0.4646259287593138, "grad_norm": 2.245986223220825, "learning_rate": 5.8120283649815905e-06, "loss": 0.4764, "step": 11084 }, { "epoch": 0.46466784737432276, "grad_norm": 1.7501084804534912, "learning_rate": 5.811358517596864e-06, "loss": 0.458, "step": 11085 }, { "epoch": 0.4647097659893317, "grad_norm": 2.1388344764709473, "learning_rate": 5.810688655255865e-06, "loss": 0.5768, "step": 11086 }, { "epoch": 0.4647516846043407, "grad_norm": 1.6604424715042114, "learning_rate": 5.8100187779709396e-06, "loss": 0.4625, "step": 11087 }, { "epoch": 0.46479360321934965, "grad_norm": 1.6811296939849854, "learning_rate": 5.8093488857544385e-06, "loss": 0.4707, "step": 11088 }, { "epoch": 0.4648355218343586, "grad_norm": 1.7922687530517578, "learning_rate": 5.808678978618711e-06, "loss": 0.4973, "step": 11089 }, { "epoch": 0.4648774404493676, "grad_norm": 1.8001761436462402, "learning_rate": 5.808009056576103e-06, "loss": 0.5195, "step": 11090 }, { "epoch": 0.46491935906437654, "grad_norm": 1.8717114925384521, "learning_rate": 5.8073391196389665e-06, "loss": 0.5413, "step": 11091 }, { "epoch": 0.4649612776793855, "grad_norm": 1.7032393217086792, "learning_rate": 5.806669167819649e-06, "loss": 0.4504, "step": 11092 }, { "epoch": 0.46500319629439446, "grad_norm": 2.135709047317505, "learning_rate": 5.805999201130498e-06, "loss": 0.4817, "step": 11093 }, { "epoch": 0.46504511490940337, "grad_norm": 2.1215763092041016, "learning_rate": 5.805329219583869e-06, "loss": 0.5223, "step": 11094 }, { "epoch": 0.46508703352441233, "grad_norm": 1.702573537826538, "learning_rate": 5.8046592231921084e-06, "loss": 0.4952, "step": 11095 }, { "epoch": 0.4651289521394213, "grad_norm": 1.7141139507293701, "learning_rate": 5.803989211967566e-06, "loss": 0.492, "step": 11096 }, { "epoch": 0.46517087075443025, "grad_norm": 1.8121750354766846, "learning_rate": 5.8033191859225955e-06, "loss": 0.5293, "step": 11097 }, { "epoch": 0.4652127893694392, "grad_norm": 1.9457627534866333, "learning_rate": 5.8026491450695465e-06, "loss": 0.4621, "step": 11098 }, { "epoch": 0.4652547079844482, "grad_norm": 2.0674664974212646, "learning_rate": 5.801979089420771e-06, "loss": 0.493, "step": 11099 }, { "epoch": 0.46529662659945714, "grad_norm": 2.1032776832580566, "learning_rate": 5.801309018988618e-06, "loss": 0.4924, "step": 11100 }, { "epoch": 0.4653385452144661, "grad_norm": 2.3147008419036865, "learning_rate": 5.8006389337854415e-06, "loss": 0.5255, "step": 11101 }, { "epoch": 0.46538046382947507, "grad_norm": 1.6487423181533813, "learning_rate": 5.799968833823592e-06, "loss": 0.4689, "step": 11102 }, { "epoch": 0.46542238244448403, "grad_norm": 1.3788888454437256, "learning_rate": 5.799298719115424e-06, "loss": 0.445, "step": 11103 }, { "epoch": 0.465464301059493, "grad_norm": 1.7370567321777344, "learning_rate": 5.7986285896732905e-06, "loss": 0.5376, "step": 11104 }, { "epoch": 0.46550621967450195, "grad_norm": 2.093433141708374, "learning_rate": 5.7979584455095404e-06, "loss": 0.4344, "step": 11105 }, { "epoch": 0.4655481382895109, "grad_norm": 1.7614548206329346, "learning_rate": 5.797288286636531e-06, "loss": 0.4928, "step": 11106 }, { "epoch": 0.4655900569045199, "grad_norm": 1.6990599632263184, "learning_rate": 5.7966181130666146e-06, "loss": 0.4499, "step": 11107 }, { "epoch": 0.46563197551952884, "grad_norm": 1.8870993852615356, "learning_rate": 5.795947924812144e-06, "loss": 0.552, "step": 11108 }, { "epoch": 0.4656738941345378, "grad_norm": 1.67037034034729, "learning_rate": 5.795277721885474e-06, "loss": 0.4775, "step": 11109 }, { "epoch": 0.46571581274954676, "grad_norm": 1.8856606483459473, "learning_rate": 5.794607504298958e-06, "loss": 0.5434, "step": 11110 }, { "epoch": 0.4657577313645557, "grad_norm": 1.554775595664978, "learning_rate": 5.793937272064952e-06, "loss": 0.5064, "step": 11111 }, { "epoch": 0.4657996499795647, "grad_norm": 2.198268413543701, "learning_rate": 5.793267025195808e-06, "loss": 0.5295, "step": 11112 }, { "epoch": 0.46584156859457365, "grad_norm": 2.029025077819824, "learning_rate": 5.792596763703885e-06, "loss": 0.5794, "step": 11113 }, { "epoch": 0.4658834872095826, "grad_norm": 1.617187261581421, "learning_rate": 5.791926487601537e-06, "loss": 0.434, "step": 11114 }, { "epoch": 0.4659254058245916, "grad_norm": 1.8723902702331543, "learning_rate": 5.7912561969011175e-06, "loss": 0.561, "step": 11115 }, { "epoch": 0.46596732443960054, "grad_norm": 2.5423972606658936, "learning_rate": 5.790585891614984e-06, "loss": 0.4823, "step": 11116 }, { "epoch": 0.4660092430546095, "grad_norm": 1.902099370956421, "learning_rate": 5.789915571755493e-06, "loss": 0.5857, "step": 11117 }, { "epoch": 0.46605116166961846, "grad_norm": 1.7037220001220703, "learning_rate": 5.789245237335001e-06, "loss": 0.5135, "step": 11118 }, { "epoch": 0.46609308028462737, "grad_norm": 2.0600292682647705, "learning_rate": 5.7885748883658635e-06, "loss": 0.4473, "step": 11119 }, { "epoch": 0.46613499889963633, "grad_norm": 1.7491766214370728, "learning_rate": 5.787904524860438e-06, "loss": 0.4437, "step": 11120 }, { "epoch": 0.4661769175146453, "grad_norm": 1.8670262098312378, "learning_rate": 5.787234146831081e-06, "loss": 0.4955, "step": 11121 }, { "epoch": 0.46621883612965426, "grad_norm": 1.9724488258361816, "learning_rate": 5.786563754290152e-06, "loss": 0.5224, "step": 11122 }, { "epoch": 0.4662607547446632, "grad_norm": 1.5442029237747192, "learning_rate": 5.785893347250009e-06, "loss": 0.4854, "step": 11123 }, { "epoch": 0.4663026733596722, "grad_norm": 1.7236659526824951, "learning_rate": 5.785222925723006e-06, "loss": 0.5639, "step": 11124 }, { "epoch": 0.46634459197468114, "grad_norm": 2.170487403869629, "learning_rate": 5.784552489721505e-06, "loss": 0.5257, "step": 11125 }, { "epoch": 0.4663865105896901, "grad_norm": 2.0304644107818604, "learning_rate": 5.7838820392578635e-06, "loss": 0.5498, "step": 11126 }, { "epoch": 0.46642842920469907, "grad_norm": 2.0029025077819824, "learning_rate": 5.783211574344441e-06, "loss": 0.4791, "step": 11127 }, { "epoch": 0.46647034781970803, "grad_norm": 2.450782060623169, "learning_rate": 5.7825410949935945e-06, "loss": 0.4987, "step": 11128 }, { "epoch": 0.466512266434717, "grad_norm": 1.4611977338790894, "learning_rate": 5.781870601217685e-06, "loss": 0.4842, "step": 11129 }, { "epoch": 0.46655418504972596, "grad_norm": 1.7767131328582764, "learning_rate": 5.781200093029071e-06, "loss": 0.5014, "step": 11130 }, { "epoch": 0.4665961036647349, "grad_norm": 1.7334011793136597, "learning_rate": 5.780529570440114e-06, "loss": 0.5294, "step": 11131 }, { "epoch": 0.4666380222797439, "grad_norm": 1.7057294845581055, "learning_rate": 5.779859033463174e-06, "loss": 0.5268, "step": 11132 }, { "epoch": 0.46667994089475284, "grad_norm": 1.8255672454833984, "learning_rate": 5.7791884821106115e-06, "loss": 0.5328, "step": 11133 }, { "epoch": 0.4667218595097618, "grad_norm": 1.840610146522522, "learning_rate": 5.778517916394786e-06, "loss": 0.4606, "step": 11134 }, { "epoch": 0.46676377812477077, "grad_norm": 1.957892894744873, "learning_rate": 5.777847336328058e-06, "loss": 0.4732, "step": 11135 }, { "epoch": 0.46680569673977973, "grad_norm": 1.6882920265197754, "learning_rate": 5.77717674192279e-06, "loss": 0.5063, "step": 11136 }, { "epoch": 0.4668476153547887, "grad_norm": 4.479596138000488, "learning_rate": 5.7765061331913445e-06, "loss": 0.51, "step": 11137 }, { "epoch": 0.46688953396979765, "grad_norm": 1.5986226797103882, "learning_rate": 5.775835510146082e-06, "loss": 0.5213, "step": 11138 }, { "epoch": 0.4669314525848066, "grad_norm": 1.6689965724945068, "learning_rate": 5.775164872799362e-06, "loss": 0.5244, "step": 11139 }, { "epoch": 0.4669733711998156, "grad_norm": 1.8162990808486938, "learning_rate": 5.77449422116355e-06, "loss": 0.5088, "step": 11140 }, { "epoch": 0.46701528981482454, "grad_norm": 2.1414878368377686, "learning_rate": 5.773823555251009e-06, "loss": 0.5181, "step": 11141 }, { "epoch": 0.4670572084298335, "grad_norm": 2.0635926723480225, "learning_rate": 5.7731528750741e-06, "loss": 0.5218, "step": 11142 }, { "epoch": 0.46709912704484247, "grad_norm": 1.9858461618423462, "learning_rate": 5.772482180645187e-06, "loss": 0.4882, "step": 11143 }, { "epoch": 0.4671410456598514, "grad_norm": 1.714774489402771, "learning_rate": 5.771811471976632e-06, "loss": 0.5028, "step": 11144 }, { "epoch": 0.46718296427486034, "grad_norm": 2.636641502380371, "learning_rate": 5.7711407490808e-06, "loss": 0.4974, "step": 11145 }, { "epoch": 0.4672248828898693, "grad_norm": 1.9509074687957764, "learning_rate": 5.770470011970054e-06, "loss": 0.5637, "step": 11146 }, { "epoch": 0.46726680150487826, "grad_norm": 1.5612887144088745, "learning_rate": 5.7697992606567585e-06, "loss": 0.4984, "step": 11147 }, { "epoch": 0.4673087201198872, "grad_norm": 1.879496455192566, "learning_rate": 5.769128495153279e-06, "loss": 0.5246, "step": 11148 }, { "epoch": 0.4673506387348962, "grad_norm": 1.658994436264038, "learning_rate": 5.768457715471977e-06, "loss": 0.5111, "step": 11149 }, { "epoch": 0.46739255734990515, "grad_norm": 1.7795811891555786, "learning_rate": 5.76778692162522e-06, "loss": 0.5431, "step": 11150 }, { "epoch": 0.4674344759649141, "grad_norm": 1.7227431535720825, "learning_rate": 5.767116113625375e-06, "loss": 0.494, "step": 11151 }, { "epoch": 0.46747639457992307, "grad_norm": 1.62053644657135, "learning_rate": 5.7664452914848015e-06, "loss": 0.5157, "step": 11152 }, { "epoch": 0.46751831319493203, "grad_norm": 1.5561400651931763, "learning_rate": 5.76577445521587e-06, "loss": 0.4914, "step": 11153 }, { "epoch": 0.467560231809941, "grad_norm": 12.360795021057129, "learning_rate": 5.765103604830946e-06, "loss": 0.5558, "step": 11154 }, { "epoch": 0.46760215042494996, "grad_norm": 2.014545440673828, "learning_rate": 5.764432740342394e-06, "loss": 0.5098, "step": 11155 }, { "epoch": 0.4676440690399589, "grad_norm": 1.9062132835388184, "learning_rate": 5.7637618617625815e-06, "loss": 0.4909, "step": 11156 }, { "epoch": 0.4676859876549679, "grad_norm": 1.7909923791885376, "learning_rate": 5.763090969103874e-06, "loss": 0.5553, "step": 11157 }, { "epoch": 0.46772790626997685, "grad_norm": 1.6811426877975464, "learning_rate": 5.762420062378638e-06, "loss": 0.4964, "step": 11158 }, { "epoch": 0.4677698248849858, "grad_norm": 1.6403913497924805, "learning_rate": 5.761749141599246e-06, "loss": 0.449, "step": 11159 }, { "epoch": 0.46781174349999477, "grad_norm": 1.7131925821304321, "learning_rate": 5.761078206778059e-06, "loss": 0.5159, "step": 11160 }, { "epoch": 0.46785366211500373, "grad_norm": 1.9475308656692505, "learning_rate": 5.760407257927448e-06, "loss": 0.4831, "step": 11161 }, { "epoch": 0.4678955807300127, "grad_norm": 23.150283813476562, "learning_rate": 5.759736295059781e-06, "loss": 0.4871, "step": 11162 }, { "epoch": 0.46793749934502166, "grad_norm": 1.8701937198638916, "learning_rate": 5.759065318187426e-06, "loss": 0.5579, "step": 11163 }, { "epoch": 0.4679794179600306, "grad_norm": 1.986741065979004, "learning_rate": 5.758394327322749e-06, "loss": 0.5602, "step": 11164 }, { "epoch": 0.4680213365750396, "grad_norm": 1.7925755977630615, "learning_rate": 5.7577233224781235e-06, "loss": 0.4743, "step": 11165 }, { "epoch": 0.46806325519004854, "grad_norm": 1.6968417167663574, "learning_rate": 5.757052303665917e-06, "loss": 0.4633, "step": 11166 }, { "epoch": 0.4681051738050575, "grad_norm": 1.6741100549697876, "learning_rate": 5.756381270898494e-06, "loss": 0.4957, "step": 11167 }, { "epoch": 0.46814709242006647, "grad_norm": 2.068181276321411, "learning_rate": 5.75571022418823e-06, "loss": 0.5503, "step": 11168 }, { "epoch": 0.4681890110350754, "grad_norm": 1.6670771837234497, "learning_rate": 5.755039163547493e-06, "loss": 0.4842, "step": 11169 }, { "epoch": 0.46823092965008434, "grad_norm": 1.7509281635284424, "learning_rate": 5.754368088988653e-06, "loss": 0.5198, "step": 11170 }, { "epoch": 0.4682728482650933, "grad_norm": 1.895635962486267, "learning_rate": 5.75369700052408e-06, "loss": 0.5359, "step": 11171 }, { "epoch": 0.46831476688010226, "grad_norm": 1.6964646577835083, "learning_rate": 5.753025898166145e-06, "loss": 0.5177, "step": 11172 }, { "epoch": 0.4683566854951112, "grad_norm": 1.7176055908203125, "learning_rate": 5.752354781927216e-06, "loss": 0.4937, "step": 11173 }, { "epoch": 0.4683986041101202, "grad_norm": 1.5908232927322388, "learning_rate": 5.751683651819671e-06, "loss": 0.5059, "step": 11174 }, { "epoch": 0.46844052272512915, "grad_norm": 1.4113996028900146, "learning_rate": 5.7510125078558745e-06, "loss": 0.4768, "step": 11175 }, { "epoch": 0.4684824413401381, "grad_norm": 1.6422812938690186, "learning_rate": 5.7503413500482e-06, "loss": 0.5244, "step": 11176 }, { "epoch": 0.4685243599551471, "grad_norm": 2.2189462184906006, "learning_rate": 5.749670178409021e-06, "loss": 0.5095, "step": 11177 }, { "epoch": 0.46856627857015604, "grad_norm": 1.7290725708007812, "learning_rate": 5.74899899295071e-06, "loss": 0.5762, "step": 11178 }, { "epoch": 0.468608197185165, "grad_norm": 1.7257872819900513, "learning_rate": 5.748327793685636e-06, "loss": 0.5717, "step": 11179 }, { "epoch": 0.46865011580017396, "grad_norm": 2.159899950027466, "learning_rate": 5.747656580626174e-06, "loss": 0.4653, "step": 11180 }, { "epoch": 0.4686920344151829, "grad_norm": 1.546299695968628, "learning_rate": 5.746985353784697e-06, "loss": 0.5031, "step": 11181 }, { "epoch": 0.4687339530301919, "grad_norm": 1.773864507675171, "learning_rate": 5.746314113173578e-06, "loss": 0.4988, "step": 11182 }, { "epoch": 0.46877587164520085, "grad_norm": 1.845247507095337, "learning_rate": 5.7456428588051894e-06, "loss": 0.492, "step": 11183 }, { "epoch": 0.4688177902602098, "grad_norm": 1.5827105045318604, "learning_rate": 5.744971590691907e-06, "loss": 0.4846, "step": 11184 }, { "epoch": 0.4688597088752188, "grad_norm": 1.859758734703064, "learning_rate": 5.744300308846101e-06, "loss": 0.4956, "step": 11185 }, { "epoch": 0.46890162749022773, "grad_norm": 3.132735252380371, "learning_rate": 5.743629013280149e-06, "loss": 0.4689, "step": 11186 }, { "epoch": 0.4689435461052367, "grad_norm": 1.8251079320907593, "learning_rate": 5.742957704006424e-06, "loss": 0.5429, "step": 11187 }, { "epoch": 0.46898546472024566, "grad_norm": 1.6596219539642334, "learning_rate": 5.7422863810373005e-06, "loss": 0.5372, "step": 11188 }, { "epoch": 0.4690273833352546, "grad_norm": 1.7136660814285278, "learning_rate": 5.741615044385154e-06, "loss": 0.4584, "step": 11189 }, { "epoch": 0.4690693019502636, "grad_norm": 1.6953648328781128, "learning_rate": 5.74094369406236e-06, "loss": 0.5204, "step": 11190 }, { "epoch": 0.46911122056527255, "grad_norm": 3.0963823795318604, "learning_rate": 5.740272330081292e-06, "loss": 0.5428, "step": 11191 }, { "epoch": 0.4691531391802815, "grad_norm": 1.6842421293258667, "learning_rate": 5.7396009524543274e-06, "loss": 0.5232, "step": 11192 }, { "epoch": 0.46919505779529047, "grad_norm": 1.7200874090194702, "learning_rate": 5.738929561193844e-06, "loss": 0.5106, "step": 11193 }, { "epoch": 0.4692369764102994, "grad_norm": 1.6091022491455078, "learning_rate": 5.738258156312213e-06, "loss": 0.535, "step": 11194 }, { "epoch": 0.46927889502530834, "grad_norm": 2.366641044616699, "learning_rate": 5.737586737821813e-06, "loss": 0.4695, "step": 11195 }, { "epoch": 0.4693208136403173, "grad_norm": 1.5977705717086792, "learning_rate": 5.736915305735022e-06, "loss": 0.4764, "step": 11196 }, { "epoch": 0.46936273225532626, "grad_norm": 2.1811161041259766, "learning_rate": 5.7362438600642175e-06, "loss": 0.4792, "step": 11197 }, { "epoch": 0.4694046508703352, "grad_norm": 1.9717375040054321, "learning_rate": 5.735572400821774e-06, "loss": 0.5988, "step": 11198 }, { "epoch": 0.4694465694853442, "grad_norm": 1.6373976469039917, "learning_rate": 5.73490092802007e-06, "loss": 0.4534, "step": 11199 }, { "epoch": 0.46948848810035315, "grad_norm": 1.8030600547790527, "learning_rate": 5.734229441671485e-06, "loss": 0.5114, "step": 11200 }, { "epoch": 0.4695304067153621, "grad_norm": 1.6253961324691772, "learning_rate": 5.733557941788392e-06, "loss": 0.5154, "step": 11201 }, { "epoch": 0.4695723253303711, "grad_norm": 1.6233669519424438, "learning_rate": 5.732886428383175e-06, "loss": 0.4928, "step": 11202 }, { "epoch": 0.46961424394538004, "grad_norm": 1.8153263330459595, "learning_rate": 5.7322149014682115e-06, "loss": 0.5462, "step": 11203 }, { "epoch": 0.469656162560389, "grad_norm": 1.50813627243042, "learning_rate": 5.7315433610558755e-06, "loss": 0.4465, "step": 11204 }, { "epoch": 0.46969808117539796, "grad_norm": 1.7074878215789795, "learning_rate": 5.73087180715855e-06, "loss": 0.4979, "step": 11205 }, { "epoch": 0.4697399997904069, "grad_norm": 1.717069149017334, "learning_rate": 5.7302002397886126e-06, "loss": 0.5061, "step": 11206 }, { "epoch": 0.4697819184054159, "grad_norm": 1.6131372451782227, "learning_rate": 5.729528658958444e-06, "loss": 0.4953, "step": 11207 }, { "epoch": 0.46982383702042485, "grad_norm": 1.7357081174850464, "learning_rate": 5.728857064680423e-06, "loss": 0.5156, "step": 11208 }, { "epoch": 0.4698657556354338, "grad_norm": 1.9788410663604736, "learning_rate": 5.728185456966931e-06, "loss": 0.5382, "step": 11209 }, { "epoch": 0.4699076742504428, "grad_norm": 1.630944848060608, "learning_rate": 5.727513835830343e-06, "loss": 0.4824, "step": 11210 }, { "epoch": 0.46994959286545174, "grad_norm": 1.9567124843597412, "learning_rate": 5.726842201283046e-06, "loss": 0.5496, "step": 11211 }, { "epoch": 0.4699915114804607, "grad_norm": 1.7548190355300903, "learning_rate": 5.726170553337418e-06, "loss": 0.5278, "step": 11212 }, { "epoch": 0.47003343009546966, "grad_norm": 1.4707181453704834, "learning_rate": 5.725498892005839e-06, "loss": 0.4605, "step": 11213 }, { "epoch": 0.4700753487104786, "grad_norm": 2.6304643154144287, "learning_rate": 5.7248272173006905e-06, "loss": 0.4912, "step": 11214 }, { "epoch": 0.4701172673254876, "grad_norm": 1.673279047012329, "learning_rate": 5.724155529234355e-06, "loss": 0.4911, "step": 11215 }, { "epoch": 0.47015918594049655, "grad_norm": 1.6526962518692017, "learning_rate": 5.723483827819213e-06, "loss": 0.4476, "step": 11216 }, { "epoch": 0.4702011045555055, "grad_norm": 1.9104000329971313, "learning_rate": 5.722812113067647e-06, "loss": 0.5377, "step": 11217 }, { "epoch": 0.4702430231705145, "grad_norm": 1.8539659976959229, "learning_rate": 5.72214038499204e-06, "loss": 0.5022, "step": 11218 }, { "epoch": 0.4702849417855234, "grad_norm": 1.6931145191192627, "learning_rate": 5.721468643604771e-06, "loss": 0.5081, "step": 11219 }, { "epoch": 0.47032686040053234, "grad_norm": 1.5531277656555176, "learning_rate": 5.720796888918226e-06, "loss": 0.478, "step": 11220 }, { "epoch": 0.4703687790155413, "grad_norm": 1.5036016702651978, "learning_rate": 5.720125120944788e-06, "loss": 0.4936, "step": 11221 }, { "epoch": 0.47041069763055027, "grad_norm": 1.7755351066589355, "learning_rate": 5.719453339696839e-06, "loss": 0.4782, "step": 11222 }, { "epoch": 0.47045261624555923, "grad_norm": 1.7108317613601685, "learning_rate": 5.71878154518676e-06, "loss": 0.5286, "step": 11223 }, { "epoch": 0.4704945348605682, "grad_norm": 1.478764295578003, "learning_rate": 5.718109737426938e-06, "loss": 0.4812, "step": 11224 }, { "epoch": 0.47053645347557715, "grad_norm": 2.124574899673462, "learning_rate": 5.717437916429757e-06, "loss": 0.4814, "step": 11225 }, { "epoch": 0.4705783720905861, "grad_norm": 1.5622104406356812, "learning_rate": 5.716766082207597e-06, "loss": 0.4813, "step": 11226 }, { "epoch": 0.4706202907055951, "grad_norm": 1.648503065109253, "learning_rate": 5.716094234772848e-06, "loss": 0.5218, "step": 11227 }, { "epoch": 0.47066220932060404, "grad_norm": 1.4020928144454956, "learning_rate": 5.7154223741378915e-06, "loss": 0.4951, "step": 11228 }, { "epoch": 0.470704127935613, "grad_norm": 5.156443119049072, "learning_rate": 5.71475050031511e-06, "loss": 0.493, "step": 11229 }, { "epoch": 0.47074604655062197, "grad_norm": 1.4285597801208496, "learning_rate": 5.714078613316893e-06, "loss": 0.529, "step": 11230 }, { "epoch": 0.47078796516563093, "grad_norm": 1.6314460039138794, "learning_rate": 5.713406713155626e-06, "loss": 0.4722, "step": 11231 }, { "epoch": 0.4708298837806399, "grad_norm": 1.863750696182251, "learning_rate": 5.712734799843689e-06, "loss": 0.459, "step": 11232 }, { "epoch": 0.47087180239564885, "grad_norm": 1.6356676816940308, "learning_rate": 5.7120628733934724e-06, "loss": 0.4835, "step": 11233 }, { "epoch": 0.4709137210106578, "grad_norm": 1.6293103694915771, "learning_rate": 5.711390933817362e-06, "loss": 0.4857, "step": 11234 }, { "epoch": 0.4709556396256668, "grad_norm": 1.5483834743499756, "learning_rate": 5.710718981127743e-06, "loss": 0.4436, "step": 11235 }, { "epoch": 0.47099755824067574, "grad_norm": 1.6849366426467896, "learning_rate": 5.710047015337002e-06, "loss": 0.5362, "step": 11236 }, { "epoch": 0.4710394768556847, "grad_norm": 1.6058861017227173, "learning_rate": 5.709375036457525e-06, "loss": 0.5356, "step": 11237 }, { "epoch": 0.47108139547069366, "grad_norm": 1.4096906185150146, "learning_rate": 5.7087030445017e-06, "loss": 0.5124, "step": 11238 }, { "epoch": 0.4711233140857026, "grad_norm": 1.7636414766311646, "learning_rate": 5.708031039481914e-06, "loss": 0.4911, "step": 11239 }, { "epoch": 0.4711652327007116, "grad_norm": 1.595778465270996, "learning_rate": 5.707359021410557e-06, "loss": 0.4918, "step": 11240 }, { "epoch": 0.47120715131572055, "grad_norm": 1.6251718997955322, "learning_rate": 5.706686990300011e-06, "loss": 0.5366, "step": 11241 }, { "epoch": 0.4712490699307295, "grad_norm": 1.7491871118545532, "learning_rate": 5.706014946162669e-06, "loss": 0.4791, "step": 11242 }, { "epoch": 0.4712909885457385, "grad_norm": 1.9478145837783813, "learning_rate": 5.705342889010918e-06, "loss": 0.5132, "step": 11243 }, { "epoch": 0.4713329071607474, "grad_norm": 1.7514678239822388, "learning_rate": 5.704670818857145e-06, "loss": 0.5635, "step": 11244 }, { "epoch": 0.47137482577575635, "grad_norm": 1.6904453039169312, "learning_rate": 5.7039987357137406e-06, "loss": 0.4543, "step": 11245 }, { "epoch": 0.4714167443907653, "grad_norm": 1.6851423978805542, "learning_rate": 5.703326639593092e-06, "loss": 0.5633, "step": 11246 }, { "epoch": 0.47145866300577427, "grad_norm": 1.5608971118927002, "learning_rate": 5.702654530507588e-06, "loss": 0.5289, "step": 11247 }, { "epoch": 0.47150058162078323, "grad_norm": 1.7002149820327759, "learning_rate": 5.701982408469621e-06, "loss": 0.4976, "step": 11248 }, { "epoch": 0.4715425002357922, "grad_norm": 1.5732276439666748, "learning_rate": 5.701310273491578e-06, "loss": 0.4915, "step": 11249 }, { "epoch": 0.47158441885080116, "grad_norm": 1.3795068264007568, "learning_rate": 5.70063812558585e-06, "loss": 0.4426, "step": 11250 }, { "epoch": 0.4716263374658101, "grad_norm": 1.879286527633667, "learning_rate": 5.699965964764826e-06, "loss": 0.5025, "step": 11251 }, { "epoch": 0.4716682560808191, "grad_norm": 1.6628077030181885, "learning_rate": 5.699293791040899e-06, "loss": 0.4751, "step": 11252 }, { "epoch": 0.47171017469582804, "grad_norm": 1.8016598224639893, "learning_rate": 5.698621604426454e-06, "loss": 0.5152, "step": 11253 }, { "epoch": 0.471752093310837, "grad_norm": 1.5880253314971924, "learning_rate": 5.697949404933889e-06, "loss": 0.5371, "step": 11254 }, { "epoch": 0.47179401192584597, "grad_norm": 1.5570762157440186, "learning_rate": 5.697277192575591e-06, "loss": 0.5566, "step": 11255 }, { "epoch": 0.47183593054085493, "grad_norm": 1.63174569606781, "learning_rate": 5.696604967363952e-06, "loss": 0.5068, "step": 11256 }, { "epoch": 0.4718778491558639, "grad_norm": 1.7734662294387817, "learning_rate": 5.695932729311362e-06, "loss": 0.5095, "step": 11257 }, { "epoch": 0.47191976777087286, "grad_norm": 1.5677088499069214, "learning_rate": 5.695260478430214e-06, "loss": 0.5465, "step": 11258 }, { "epoch": 0.4719616863858818, "grad_norm": 2.027329683303833, "learning_rate": 5.694588214732903e-06, "loss": 0.4942, "step": 11259 }, { "epoch": 0.4720036050008908, "grad_norm": 1.814052700996399, "learning_rate": 5.693915938231817e-06, "loss": 0.5226, "step": 11260 }, { "epoch": 0.47204552361589974, "grad_norm": 2.1006577014923096, "learning_rate": 5.693243648939349e-06, "loss": 0.5224, "step": 11261 }, { "epoch": 0.4720874422309087, "grad_norm": 1.9413882493972778, "learning_rate": 5.692571346867893e-06, "loss": 0.4714, "step": 11262 }, { "epoch": 0.47212936084591767, "grad_norm": 1.794592022895813, "learning_rate": 5.691899032029843e-06, "loss": 0.523, "step": 11263 }, { "epoch": 0.47217127946092663, "grad_norm": 1.4714550971984863, "learning_rate": 5.69122670443759e-06, "loss": 0.5163, "step": 11264 }, { "epoch": 0.4722131980759356, "grad_norm": 1.7213858366012573, "learning_rate": 5.690554364103529e-06, "loss": 0.47, "step": 11265 }, { "epoch": 0.47225511669094455, "grad_norm": 1.7176886796951294, "learning_rate": 5.689882011040051e-06, "loss": 0.4668, "step": 11266 }, { "epoch": 0.4722970353059535, "grad_norm": 1.4439250230789185, "learning_rate": 5.689209645259554e-06, "loss": 0.4991, "step": 11267 }, { "epoch": 0.4723389539209625, "grad_norm": 1.509785771369934, "learning_rate": 5.68853726677443e-06, "loss": 0.4974, "step": 11268 }, { "epoch": 0.4723808725359714, "grad_norm": 2.2975525856018066, "learning_rate": 5.687864875597071e-06, "loss": 0.4574, "step": 11269 }, { "epoch": 0.47242279115098035, "grad_norm": 1.7014269828796387, "learning_rate": 5.6871924717398755e-06, "loss": 0.4923, "step": 11270 }, { "epoch": 0.4724647097659893, "grad_norm": 1.5330111980438232, "learning_rate": 5.686520055215236e-06, "loss": 0.4747, "step": 11271 }, { "epoch": 0.4725066283809983, "grad_norm": 2.325686454772949, "learning_rate": 5.6858476260355496e-06, "loss": 0.4967, "step": 11272 }, { "epoch": 0.47254854699600723, "grad_norm": 1.6454885005950928, "learning_rate": 5.685175184213211e-06, "loss": 0.5306, "step": 11273 }, { "epoch": 0.4725904656110162, "grad_norm": 1.6124311685562134, "learning_rate": 5.684502729760614e-06, "loss": 0.5099, "step": 11274 }, { "epoch": 0.47263238422602516, "grad_norm": 1.5682294368743896, "learning_rate": 5.683830262690154e-06, "loss": 0.5619, "step": 11275 }, { "epoch": 0.4726743028410341, "grad_norm": 2.039116382598877, "learning_rate": 5.68315778301423e-06, "loss": 0.4843, "step": 11276 }, { "epoch": 0.4727162214560431, "grad_norm": 1.873286247253418, "learning_rate": 5.682485290745237e-06, "loss": 0.521, "step": 11277 }, { "epoch": 0.47275814007105205, "grad_norm": 1.554032802581787, "learning_rate": 5.681812785895572e-06, "loss": 0.5081, "step": 11278 }, { "epoch": 0.472800058686061, "grad_norm": 2.9606125354766846, "learning_rate": 5.68114026847763e-06, "loss": 0.5393, "step": 11279 }, { "epoch": 0.47284197730106997, "grad_norm": 1.6604512929916382, "learning_rate": 5.680467738503807e-06, "loss": 0.4949, "step": 11280 }, { "epoch": 0.47288389591607893, "grad_norm": 1.7962850332260132, "learning_rate": 5.6797951959865036e-06, "loss": 0.5356, "step": 11281 }, { "epoch": 0.4729258145310879, "grad_norm": 1.5960029363632202, "learning_rate": 5.6791226409381165e-06, "loss": 0.5203, "step": 11282 }, { "epoch": 0.47296773314609686, "grad_norm": 1.786168098449707, "learning_rate": 5.6784500733710415e-06, "loss": 0.5373, "step": 11283 }, { "epoch": 0.4730096517611058, "grad_norm": 1.8189587593078613, "learning_rate": 5.677777493297676e-06, "loss": 0.5773, "step": 11284 }, { "epoch": 0.4730515703761148, "grad_norm": 2.0625853538513184, "learning_rate": 5.677104900730422e-06, "loss": 0.5857, "step": 11285 }, { "epoch": 0.47309348899112375, "grad_norm": 1.7109720706939697, "learning_rate": 5.676432295681673e-06, "loss": 0.4694, "step": 11286 }, { "epoch": 0.4731354076061327, "grad_norm": 1.6754052639007568, "learning_rate": 5.67575967816383e-06, "loss": 0.4649, "step": 11287 }, { "epoch": 0.47317732622114167, "grad_norm": 1.9108816385269165, "learning_rate": 5.675087048189293e-06, "loss": 0.4884, "step": 11288 }, { "epoch": 0.47321924483615063, "grad_norm": 1.5122301578521729, "learning_rate": 5.6744144057704585e-06, "loss": 0.4606, "step": 11289 }, { "epoch": 0.4732611634511596, "grad_norm": 1.9046740531921387, "learning_rate": 5.673741750919726e-06, "loss": 0.4989, "step": 11290 }, { "epoch": 0.47330308206616856, "grad_norm": 1.930747628211975, "learning_rate": 5.673069083649497e-06, "loss": 0.5322, "step": 11291 }, { "epoch": 0.4733450006811775, "grad_norm": 1.7187891006469727, "learning_rate": 5.67239640397217e-06, "loss": 0.5197, "step": 11292 }, { "epoch": 0.4733869192961865, "grad_norm": 1.592944622039795, "learning_rate": 5.671723711900143e-06, "loss": 0.5118, "step": 11293 }, { "epoch": 0.4734288379111954, "grad_norm": 2.055882453918457, "learning_rate": 5.67105100744582e-06, "loss": 0.5772, "step": 11294 }, { "epoch": 0.47347075652620435, "grad_norm": 1.5754023790359497, "learning_rate": 5.670378290621599e-06, "loss": 0.5519, "step": 11295 }, { "epoch": 0.4735126751412133, "grad_norm": 1.9073899984359741, "learning_rate": 5.6697055614398814e-06, "loss": 0.5317, "step": 11296 }, { "epoch": 0.4735545937562223, "grad_norm": 2.006058692932129, "learning_rate": 5.669032819913068e-06, "loss": 0.5412, "step": 11297 }, { "epoch": 0.47359651237123124, "grad_norm": 1.9916062355041504, "learning_rate": 5.6683600660535575e-06, "loss": 0.5468, "step": 11298 }, { "epoch": 0.4736384309862402, "grad_norm": 2.234086751937866, "learning_rate": 5.667687299873754e-06, "loss": 0.5814, "step": 11299 }, { "epoch": 0.47368034960124916, "grad_norm": 1.7409861087799072, "learning_rate": 5.667014521386057e-06, "loss": 0.5678, "step": 11300 }, { "epoch": 0.4737222682162581, "grad_norm": 1.845420479774475, "learning_rate": 5.666341730602872e-06, "loss": 0.4675, "step": 11301 }, { "epoch": 0.4737641868312671, "grad_norm": 1.712701439857483, "learning_rate": 5.665668927536598e-06, "loss": 0.4915, "step": 11302 }, { "epoch": 0.47380610544627605, "grad_norm": 2.594489097595215, "learning_rate": 5.664996112199636e-06, "loss": 0.4845, "step": 11303 }, { "epoch": 0.473848024061285, "grad_norm": 1.7365456819534302, "learning_rate": 5.6643232846043905e-06, "loss": 0.5393, "step": 11304 }, { "epoch": 0.473889942676294, "grad_norm": 1.8852899074554443, "learning_rate": 5.663650444763265e-06, "loss": 0.5245, "step": 11305 }, { "epoch": 0.47393186129130294, "grad_norm": 1.7345771789550781, "learning_rate": 5.662977592688659e-06, "loss": 0.507, "step": 11306 }, { "epoch": 0.4739737799063119, "grad_norm": 1.4371929168701172, "learning_rate": 5.662304728392979e-06, "loss": 0.4862, "step": 11307 }, { "epoch": 0.47401569852132086, "grad_norm": 1.6499099731445312, "learning_rate": 5.661631851888626e-06, "loss": 0.5683, "step": 11308 }, { "epoch": 0.4740576171363298, "grad_norm": 1.9725819826126099, "learning_rate": 5.660958963188003e-06, "loss": 0.6076, "step": 11309 }, { "epoch": 0.4740995357513388, "grad_norm": 1.4647941589355469, "learning_rate": 5.660286062303519e-06, "loss": 0.5057, "step": 11310 }, { "epoch": 0.47414145436634775, "grad_norm": 1.602736234664917, "learning_rate": 5.6596131492475735e-06, "loss": 0.487, "step": 11311 }, { "epoch": 0.4741833729813567, "grad_norm": 1.769601583480835, "learning_rate": 5.658940224032569e-06, "loss": 0.5024, "step": 11312 }, { "epoch": 0.4742252915963657, "grad_norm": 1.9944504499435425, "learning_rate": 5.658267286670914e-06, "loss": 0.5256, "step": 11313 }, { "epoch": 0.47426721021137463, "grad_norm": 2.129462480545044, "learning_rate": 5.657594337175012e-06, "loss": 0.5157, "step": 11314 }, { "epoch": 0.4743091288263836, "grad_norm": 1.885167121887207, "learning_rate": 5.656921375557267e-06, "loss": 0.4652, "step": 11315 }, { "epoch": 0.47435104744139256, "grad_norm": 2.3579630851745605, "learning_rate": 5.6562484018300845e-06, "loss": 0.567, "step": 11316 }, { "epoch": 0.4743929660564015, "grad_norm": 3.351336717605591, "learning_rate": 5.655575416005869e-06, "loss": 0.5001, "step": 11317 }, { "epoch": 0.4744348846714105, "grad_norm": 2.1503114700317383, "learning_rate": 5.654902418097028e-06, "loss": 0.5423, "step": 11318 }, { "epoch": 0.4744768032864194, "grad_norm": 1.7525149583816528, "learning_rate": 5.654229408115967e-06, "loss": 0.4801, "step": 11319 }, { "epoch": 0.47451872190142835, "grad_norm": 1.583755612373352, "learning_rate": 5.653556386075091e-06, "loss": 0.4969, "step": 11320 }, { "epoch": 0.4745606405164373, "grad_norm": 1.7111159563064575, "learning_rate": 5.652883351986805e-06, "loss": 0.4757, "step": 11321 }, { "epoch": 0.4746025591314463, "grad_norm": 1.7153891324996948, "learning_rate": 5.652210305863519e-06, "loss": 0.5069, "step": 11322 }, { "epoch": 0.47464447774645524, "grad_norm": 1.695471167564392, "learning_rate": 5.651537247717637e-06, "loss": 0.4824, "step": 11323 }, { "epoch": 0.4746863963614642, "grad_norm": 1.6407759189605713, "learning_rate": 5.650864177561565e-06, "loss": 0.5535, "step": 11324 }, { "epoch": 0.47472831497647316, "grad_norm": 1.6539244651794434, "learning_rate": 5.650191095407714e-06, "loss": 0.5137, "step": 11325 }, { "epoch": 0.4747702335914821, "grad_norm": 1.5339282751083374, "learning_rate": 5.649518001268488e-06, "loss": 0.4705, "step": 11326 }, { "epoch": 0.4748121522064911, "grad_norm": 1.5330475568771362, "learning_rate": 5.648844895156296e-06, "loss": 0.5166, "step": 11327 }, { "epoch": 0.47485407082150005, "grad_norm": 1.79531991481781, "learning_rate": 5.6481717770835445e-06, "loss": 0.5411, "step": 11328 }, { "epoch": 0.474895989436509, "grad_norm": 1.7842762470245361, "learning_rate": 5.647498647062644e-06, "loss": 0.4849, "step": 11329 }, { "epoch": 0.474937908051518, "grad_norm": 1.8167043924331665, "learning_rate": 5.646825505106e-06, "loss": 0.4911, "step": 11330 }, { "epoch": 0.47497982666652694, "grad_norm": 1.5402108430862427, "learning_rate": 5.6461523512260205e-06, "loss": 0.5087, "step": 11331 }, { "epoch": 0.4750217452815359, "grad_norm": 1.6608127355575562, "learning_rate": 5.645479185435118e-06, "loss": 0.5618, "step": 11332 }, { "epoch": 0.47506366389654486, "grad_norm": 1.5989435911178589, "learning_rate": 5.644806007745698e-06, "loss": 0.4829, "step": 11333 }, { "epoch": 0.4751055825115538, "grad_norm": 1.6970033645629883, "learning_rate": 5.644132818170172e-06, "loss": 0.509, "step": 11334 }, { "epoch": 0.4751475011265628, "grad_norm": 1.5675557851791382, "learning_rate": 5.6434596167209465e-06, "loss": 0.4935, "step": 11335 }, { "epoch": 0.47518941974157175, "grad_norm": 1.632072925567627, "learning_rate": 5.642786403410433e-06, "loss": 0.5054, "step": 11336 }, { "epoch": 0.4752313383565807, "grad_norm": 1.4095641374588013, "learning_rate": 5.642113178251039e-06, "loss": 0.4856, "step": 11337 }, { "epoch": 0.4752732569715897, "grad_norm": 1.7721059322357178, "learning_rate": 5.641439941255179e-06, "loss": 0.527, "step": 11338 }, { "epoch": 0.47531517558659864, "grad_norm": 1.4698296785354614, "learning_rate": 5.64076669243526e-06, "loss": 0.5151, "step": 11339 }, { "epoch": 0.4753570942016076, "grad_norm": 1.9217708110809326, "learning_rate": 5.640093431803691e-06, "loss": 0.483, "step": 11340 }, { "epoch": 0.47539901281661656, "grad_norm": 1.7704278230667114, "learning_rate": 5.639420159372886e-06, "loss": 0.5456, "step": 11341 }, { "epoch": 0.4754409314316255, "grad_norm": 1.972670078277588, "learning_rate": 5.638746875155254e-06, "loss": 0.4972, "step": 11342 }, { "epoch": 0.4754828500466345, "grad_norm": 2.519864797592163, "learning_rate": 5.6380735791632066e-06, "loss": 0.5228, "step": 11343 }, { "epoch": 0.4755247686616434, "grad_norm": 10.126383781433105, "learning_rate": 5.6374002714091544e-06, "loss": 0.5227, "step": 11344 }, { "epoch": 0.47556668727665236, "grad_norm": 1.363343358039856, "learning_rate": 5.63672695190551e-06, "loss": 0.5076, "step": 11345 }, { "epoch": 0.4756086058916613, "grad_norm": 1.593501329421997, "learning_rate": 5.636053620664683e-06, "loss": 0.5471, "step": 11346 }, { "epoch": 0.4756505245066703, "grad_norm": 2.148231267929077, "learning_rate": 5.635380277699088e-06, "loss": 0.5627, "step": 11347 }, { "epoch": 0.47569244312167924, "grad_norm": 1.5065547227859497, "learning_rate": 5.634706923021138e-06, "loss": 0.4982, "step": 11348 }, { "epoch": 0.4757343617366882, "grad_norm": 1.5178418159484863, "learning_rate": 5.63403355664324e-06, "loss": 0.533, "step": 11349 }, { "epoch": 0.47577628035169717, "grad_norm": 1.7241392135620117, "learning_rate": 5.6333601785778115e-06, "loss": 0.4659, "step": 11350 }, { "epoch": 0.47581819896670613, "grad_norm": 1.502847671508789, "learning_rate": 5.6326867888372635e-06, "loss": 0.5015, "step": 11351 }, { "epoch": 0.4758601175817151, "grad_norm": 1.7463160753250122, "learning_rate": 5.63201338743401e-06, "loss": 0.5539, "step": 11352 }, { "epoch": 0.47590203619672405, "grad_norm": 1.7067935466766357, "learning_rate": 5.631339974380462e-06, "loss": 0.5022, "step": 11353 }, { "epoch": 0.475943954811733, "grad_norm": 1.583660364151001, "learning_rate": 5.630666549689036e-06, "loss": 0.4964, "step": 11354 }, { "epoch": 0.475985873426742, "grad_norm": 1.7578004598617554, "learning_rate": 5.629993113372141e-06, "loss": 0.5294, "step": 11355 }, { "epoch": 0.47602779204175094, "grad_norm": 1.7007657289505005, "learning_rate": 5.629319665442197e-06, "loss": 0.5366, "step": 11356 }, { "epoch": 0.4760697106567599, "grad_norm": 1.6126116514205933, "learning_rate": 5.6286462059116145e-06, "loss": 0.5065, "step": 11357 }, { "epoch": 0.47611162927176887, "grad_norm": 1.5258492231369019, "learning_rate": 5.627972734792806e-06, "loss": 0.5151, "step": 11358 }, { "epoch": 0.47615354788677783, "grad_norm": 1.9050832986831665, "learning_rate": 5.62729925209819e-06, "loss": 0.4817, "step": 11359 }, { "epoch": 0.4761954665017868, "grad_norm": 1.8054990768432617, "learning_rate": 5.62662575784018e-06, "loss": 0.47, "step": 11360 }, { "epoch": 0.47623738511679575, "grad_norm": 1.520810604095459, "learning_rate": 5.62595225203119e-06, "loss": 0.5335, "step": 11361 }, { "epoch": 0.4762793037318047, "grad_norm": 1.8155832290649414, "learning_rate": 5.6252787346836355e-06, "loss": 0.4834, "step": 11362 }, { "epoch": 0.4763212223468137, "grad_norm": 1.7334142923355103, "learning_rate": 5.624605205809932e-06, "loss": 0.4339, "step": 11363 }, { "epoch": 0.47636314096182264, "grad_norm": 1.703338861465454, "learning_rate": 5.623931665422494e-06, "loss": 0.526, "step": 11364 }, { "epoch": 0.4764050595768316, "grad_norm": 1.6206157207489014, "learning_rate": 5.623258113533739e-06, "loss": 0.503, "step": 11365 }, { "epoch": 0.47644697819184056, "grad_norm": 1.6933549642562866, "learning_rate": 5.622584550156083e-06, "loss": 0.5124, "step": 11366 }, { "epoch": 0.4764888968068495, "grad_norm": 1.5579421520233154, "learning_rate": 5.621910975301941e-06, "loss": 0.5506, "step": 11367 }, { "epoch": 0.4765308154218585, "grad_norm": 1.6069670915603638, "learning_rate": 5.62123738898373e-06, "loss": 0.4638, "step": 11368 }, { "epoch": 0.4765727340368674, "grad_norm": 1.6085631847381592, "learning_rate": 5.620563791213867e-06, "loss": 0.4856, "step": 11369 }, { "epoch": 0.47661465265187636, "grad_norm": 1.8768004179000854, "learning_rate": 5.619890182004768e-06, "loss": 0.541, "step": 11370 }, { "epoch": 0.4766565712668853, "grad_norm": 2.1093592643737793, "learning_rate": 5.6192165613688496e-06, "loss": 0.5246, "step": 11371 }, { "epoch": 0.4766984898818943, "grad_norm": 1.692482590675354, "learning_rate": 5.618542929318531e-06, "loss": 0.4888, "step": 11372 }, { "epoch": 0.47674040849690325, "grad_norm": 1.6818045377731323, "learning_rate": 5.617869285866229e-06, "loss": 0.5054, "step": 11373 }, { "epoch": 0.4767823271119122, "grad_norm": 1.8507113456726074, "learning_rate": 5.617195631024359e-06, "loss": 0.5042, "step": 11374 }, { "epoch": 0.47682424572692117, "grad_norm": 1.750128984451294, "learning_rate": 5.616521964805343e-06, "loss": 0.5333, "step": 11375 }, { "epoch": 0.47686616434193013, "grad_norm": 1.7409261465072632, "learning_rate": 5.6158482872215965e-06, "loss": 0.5315, "step": 11376 }, { "epoch": 0.4769080829569391, "grad_norm": 1.6419371366500854, "learning_rate": 5.615174598285537e-06, "loss": 0.5444, "step": 11377 }, { "epoch": 0.47695000157194806, "grad_norm": 1.8676226139068604, "learning_rate": 5.614500898009585e-06, "loss": 0.4908, "step": 11378 }, { "epoch": 0.476991920186957, "grad_norm": 1.6124922037124634, "learning_rate": 5.613827186406158e-06, "loss": 0.5085, "step": 11379 }, { "epoch": 0.477033838801966, "grad_norm": 1.8421266078948975, "learning_rate": 5.613153463487676e-06, "loss": 0.4922, "step": 11380 }, { "epoch": 0.47707575741697494, "grad_norm": 1.7529876232147217, "learning_rate": 5.6124797292665576e-06, "loss": 0.5295, "step": 11381 }, { "epoch": 0.4771176760319839, "grad_norm": 1.8811583518981934, "learning_rate": 5.611805983755222e-06, "loss": 0.506, "step": 11382 }, { "epoch": 0.47715959464699287, "grad_norm": 1.6999280452728271, "learning_rate": 5.611132226966087e-06, "loss": 0.4766, "step": 11383 }, { "epoch": 0.47720151326200183, "grad_norm": 1.4636781215667725, "learning_rate": 5.610458458911577e-06, "loss": 0.4683, "step": 11384 }, { "epoch": 0.4772434318770108, "grad_norm": 2.4833779335021973, "learning_rate": 5.609784679604109e-06, "loss": 0.5107, "step": 11385 }, { "epoch": 0.47728535049201976, "grad_norm": 1.5516425371170044, "learning_rate": 5.609110889056102e-06, "loss": 0.4873, "step": 11386 }, { "epoch": 0.4773272691070287, "grad_norm": 1.8830665349960327, "learning_rate": 5.608437087279979e-06, "loss": 0.4986, "step": 11387 }, { "epoch": 0.4773691877220377, "grad_norm": 1.999741792678833, "learning_rate": 5.607763274288158e-06, "loss": 0.5458, "step": 11388 }, { "epoch": 0.47741110633704664, "grad_norm": 1.6524921655654907, "learning_rate": 5.607089450093063e-06, "loss": 0.5506, "step": 11389 }, { "epoch": 0.4774530249520556, "grad_norm": 1.7902768850326538, "learning_rate": 5.606415614707113e-06, "loss": 0.5167, "step": 11390 }, { "epoch": 0.47749494356706457, "grad_norm": 1.651904582977295, "learning_rate": 5.605741768142729e-06, "loss": 0.501, "step": 11391 }, { "epoch": 0.47753686218207353, "grad_norm": 1.7363593578338623, "learning_rate": 5.60506791041233e-06, "loss": 0.5092, "step": 11392 }, { "epoch": 0.4775787807970825, "grad_norm": 1.4525803327560425, "learning_rate": 5.604394041528345e-06, "loss": 0.4682, "step": 11393 }, { "epoch": 0.4776206994120914, "grad_norm": 1.6803570985794067, "learning_rate": 5.603720161503189e-06, "loss": 0.4786, "step": 11394 }, { "epoch": 0.47766261802710036, "grad_norm": 3.1842687129974365, "learning_rate": 5.603046270349288e-06, "loss": 0.5457, "step": 11395 }, { "epoch": 0.4777045366421093, "grad_norm": 1.7508400678634644, "learning_rate": 5.602372368079062e-06, "loss": 0.5073, "step": 11396 }, { "epoch": 0.4777464552571183, "grad_norm": 1.495104432106018, "learning_rate": 5.601698454704934e-06, "loss": 0.4544, "step": 11397 }, { "epoch": 0.47778837387212725, "grad_norm": 2.281893014907837, "learning_rate": 5.601024530239326e-06, "loss": 0.484, "step": 11398 }, { "epoch": 0.4778302924871362, "grad_norm": 1.8455562591552734, "learning_rate": 5.6003505946946635e-06, "loss": 0.5259, "step": 11399 }, { "epoch": 0.4778722111021452, "grad_norm": 1.8098924160003662, "learning_rate": 5.5996766480833675e-06, "loss": 0.5175, "step": 11400 }, { "epoch": 0.47791412971715413, "grad_norm": 2.162219762802124, "learning_rate": 5.5990026904178595e-06, "loss": 0.5615, "step": 11401 }, { "epoch": 0.4779560483321631, "grad_norm": 2.150259494781494, "learning_rate": 5.598328721710566e-06, "loss": 0.4968, "step": 11402 }, { "epoch": 0.47799796694717206, "grad_norm": 2.229379653930664, "learning_rate": 5.59765474197391e-06, "loss": 0.5024, "step": 11403 }, { "epoch": 0.478039885562181, "grad_norm": 1.6985621452331543, "learning_rate": 5.596980751220316e-06, "loss": 0.5298, "step": 11404 }, { "epoch": 0.47808180417719, "grad_norm": 2.0263454914093018, "learning_rate": 5.5963067494622055e-06, "loss": 0.5827, "step": 11405 }, { "epoch": 0.47812372279219895, "grad_norm": 2.62674617767334, "learning_rate": 5.595632736712005e-06, "loss": 0.5065, "step": 11406 }, { "epoch": 0.4781656414072079, "grad_norm": 1.5851107835769653, "learning_rate": 5.594958712982139e-06, "loss": 0.4574, "step": 11407 }, { "epoch": 0.47820756002221687, "grad_norm": 3.1735422611236572, "learning_rate": 5.594284678285031e-06, "loss": 0.5064, "step": 11408 }, { "epoch": 0.47824947863722583, "grad_norm": 2.469625234603882, "learning_rate": 5.593610632633108e-06, "loss": 0.5072, "step": 11409 }, { "epoch": 0.4782913972522348, "grad_norm": 1.6968812942504883, "learning_rate": 5.592936576038792e-06, "loss": 0.5167, "step": 11410 }, { "epoch": 0.47833331586724376, "grad_norm": 1.8783397674560547, "learning_rate": 5.592262508514509e-06, "loss": 0.5129, "step": 11411 }, { "epoch": 0.4783752344822527, "grad_norm": 1.6781765222549438, "learning_rate": 5.591588430072688e-06, "loss": 0.519, "step": 11412 }, { "epoch": 0.4784171530972617, "grad_norm": 1.7648762464523315, "learning_rate": 5.590914340725751e-06, "loss": 0.5151, "step": 11413 }, { "epoch": 0.47845907171227064, "grad_norm": 1.3676567077636719, "learning_rate": 5.590240240486125e-06, "loss": 0.4749, "step": 11414 }, { "epoch": 0.4785009903272796, "grad_norm": 1.9924324750900269, "learning_rate": 5.589566129366235e-06, "loss": 0.489, "step": 11415 }, { "epoch": 0.47854290894228857, "grad_norm": 1.954546332359314, "learning_rate": 5.5888920073785105e-06, "loss": 0.5101, "step": 11416 }, { "epoch": 0.47858482755729753, "grad_norm": 1.6791764497756958, "learning_rate": 5.5882178745353735e-06, "loss": 0.5377, "step": 11417 }, { "epoch": 0.4786267461723065, "grad_norm": 1.4967187643051147, "learning_rate": 5.587543730849257e-06, "loss": 0.5174, "step": 11418 }, { "epoch": 0.4786686647873154, "grad_norm": 1.724469780921936, "learning_rate": 5.586869576332581e-06, "loss": 0.4649, "step": 11419 }, { "epoch": 0.47871058340232436, "grad_norm": 2.1147024631500244, "learning_rate": 5.586195410997775e-06, "loss": 0.5462, "step": 11420 }, { "epoch": 0.4787525020173333, "grad_norm": 1.5789260864257812, "learning_rate": 5.585521234857268e-06, "loss": 0.5004, "step": 11421 }, { "epoch": 0.4787944206323423, "grad_norm": 1.689700722694397, "learning_rate": 5.584847047923487e-06, "loss": 0.5117, "step": 11422 }, { "epoch": 0.47883633924735125, "grad_norm": 1.5230047702789307, "learning_rate": 5.58417285020886e-06, "loss": 0.5041, "step": 11423 }, { "epoch": 0.4788782578623602, "grad_norm": 1.4908158779144287, "learning_rate": 5.583498641725813e-06, "loss": 0.4989, "step": 11424 }, { "epoch": 0.4789201764773692, "grad_norm": 1.790097951889038, "learning_rate": 5.5828244224867754e-06, "loss": 0.4956, "step": 11425 }, { "epoch": 0.47896209509237814, "grad_norm": 1.5619118213653564, "learning_rate": 5.582150192504175e-06, "loss": 0.4869, "step": 11426 }, { "epoch": 0.4790040137073871, "grad_norm": 1.8122076988220215, "learning_rate": 5.581475951790441e-06, "loss": 0.5236, "step": 11427 }, { "epoch": 0.47904593232239606, "grad_norm": 1.5810837745666504, "learning_rate": 5.5808017003580025e-06, "loss": 0.5171, "step": 11428 }, { "epoch": 0.479087850937405, "grad_norm": 9.759774208068848, "learning_rate": 5.580127438219286e-06, "loss": 0.5234, "step": 11429 }, { "epoch": 0.479129769552414, "grad_norm": 1.6299023628234863, "learning_rate": 5.579453165386722e-06, "loss": 0.4587, "step": 11430 }, { "epoch": 0.47917168816742295, "grad_norm": 1.7650787830352783, "learning_rate": 5.578778881872742e-06, "loss": 0.5403, "step": 11431 }, { "epoch": 0.4792136067824319, "grad_norm": 1.892934799194336, "learning_rate": 5.578104587689772e-06, "loss": 0.517, "step": 11432 }, { "epoch": 0.4792555253974409, "grad_norm": 1.9498380422592163, "learning_rate": 5.577430282850244e-06, "loss": 0.5649, "step": 11433 }, { "epoch": 0.47929744401244984, "grad_norm": 1.7884247303009033, "learning_rate": 5.576755967366586e-06, "loss": 0.4927, "step": 11434 }, { "epoch": 0.4793393626274588, "grad_norm": 1.694785475730896, "learning_rate": 5.576081641251229e-06, "loss": 0.5165, "step": 11435 }, { "epoch": 0.47938128124246776, "grad_norm": 1.6445887088775635, "learning_rate": 5.575407304516605e-06, "loss": 0.5397, "step": 11436 }, { "epoch": 0.4794231998574767, "grad_norm": 1.9297983646392822, "learning_rate": 5.574732957175142e-06, "loss": 0.5732, "step": 11437 }, { "epoch": 0.4794651184724857, "grad_norm": 1.9260220527648926, "learning_rate": 5.574058599239269e-06, "loss": 0.5527, "step": 11438 }, { "epoch": 0.47950703708749465, "grad_norm": 1.7751116752624512, "learning_rate": 5.573384230721423e-06, "loss": 0.5391, "step": 11439 }, { "epoch": 0.4795489557025036, "grad_norm": 1.7646719217300415, "learning_rate": 5.57270985163403e-06, "loss": 0.4851, "step": 11440 }, { "epoch": 0.47959087431751257, "grad_norm": 1.88644278049469, "learning_rate": 5.5720354619895225e-06, "loss": 0.4805, "step": 11441 }, { "epoch": 0.47963279293252153, "grad_norm": 1.9602490663528442, "learning_rate": 5.5713610618003324e-06, "loss": 0.5022, "step": 11442 }, { "epoch": 0.4796747115475305, "grad_norm": 1.5916965007781982, "learning_rate": 5.570686651078891e-06, "loss": 0.5079, "step": 11443 }, { "epoch": 0.4797166301625394, "grad_norm": 1.7311866283416748, "learning_rate": 5.5700122298376294e-06, "loss": 0.5622, "step": 11444 }, { "epoch": 0.47975854877754837, "grad_norm": 1.7434711456298828, "learning_rate": 5.569337798088979e-06, "loss": 0.5733, "step": 11445 }, { "epoch": 0.47980046739255733, "grad_norm": 2.3111040592193604, "learning_rate": 5.568663355845378e-06, "loss": 0.5094, "step": 11446 }, { "epoch": 0.4798423860075663, "grad_norm": 1.4592667818069458, "learning_rate": 5.56798890311925e-06, "loss": 0.5056, "step": 11447 }, { "epoch": 0.47988430462257525, "grad_norm": 1.6516494750976562, "learning_rate": 5.5673144399230326e-06, "loss": 0.4444, "step": 11448 }, { "epoch": 0.4799262232375842, "grad_norm": 1.662302851676941, "learning_rate": 5.566639966269158e-06, "loss": 0.5475, "step": 11449 }, { "epoch": 0.4799681418525932, "grad_norm": 1.6174355745315552, "learning_rate": 5.56596548217006e-06, "loss": 0.4587, "step": 11450 }, { "epoch": 0.48001006046760214, "grad_norm": 1.8834352493286133, "learning_rate": 5.56529098763817e-06, "loss": 0.5639, "step": 11451 }, { "epoch": 0.4800519790826111, "grad_norm": 1.7414085865020752, "learning_rate": 5.5646164826859225e-06, "loss": 0.5178, "step": 11452 }, { "epoch": 0.48009389769762006, "grad_norm": 2.308967351913452, "learning_rate": 5.56394196732575e-06, "loss": 0.4592, "step": 11453 }, { "epoch": 0.480135816312629, "grad_norm": 1.7568674087524414, "learning_rate": 5.563267441570087e-06, "loss": 0.5387, "step": 11454 }, { "epoch": 0.480177734927638, "grad_norm": 1.6200302839279175, "learning_rate": 5.562592905431368e-06, "loss": 0.4544, "step": 11455 }, { "epoch": 0.48021965354264695, "grad_norm": 2.78710675239563, "learning_rate": 5.5619183589220275e-06, "loss": 0.5267, "step": 11456 }, { "epoch": 0.4802615721576559, "grad_norm": 1.7315748929977417, "learning_rate": 5.561243802054497e-06, "loss": 0.5125, "step": 11457 }, { "epoch": 0.4803034907726649, "grad_norm": 1.5363919734954834, "learning_rate": 5.5605692348412136e-06, "loss": 0.4034, "step": 11458 }, { "epoch": 0.48034540938767384, "grad_norm": 1.6203721761703491, "learning_rate": 5.559894657294611e-06, "loss": 0.476, "step": 11459 }, { "epoch": 0.4803873280026828, "grad_norm": 1.7537418603897095, "learning_rate": 5.559220069427125e-06, "loss": 0.5165, "step": 11460 }, { "epoch": 0.48042924661769176, "grad_norm": 1.6534956693649292, "learning_rate": 5.55854547125119e-06, "loss": 0.5178, "step": 11461 }, { "epoch": 0.4804711652327007, "grad_norm": 1.80845046043396, "learning_rate": 5.557870862779241e-06, "loss": 0.5168, "step": 11462 }, { "epoch": 0.4805130838477097, "grad_norm": 1.9674311876296997, "learning_rate": 5.557196244023714e-06, "loss": 0.5077, "step": 11463 }, { "epoch": 0.48055500246271865, "grad_norm": 2.1475417613983154, "learning_rate": 5.556521614997045e-06, "loss": 0.5585, "step": 11464 }, { "epoch": 0.4805969210777276, "grad_norm": 1.8308650255203247, "learning_rate": 5.5558469757116695e-06, "loss": 0.5062, "step": 11465 }, { "epoch": 0.4806388396927366, "grad_norm": 1.644865870475769, "learning_rate": 5.555172326180023e-06, "loss": 0.5317, "step": 11466 }, { "epoch": 0.48068075830774554, "grad_norm": 1.9084627628326416, "learning_rate": 5.554497666414541e-06, "loss": 0.5306, "step": 11467 }, { "epoch": 0.4807226769227545, "grad_norm": 1.6413265466690063, "learning_rate": 5.553822996427664e-06, "loss": 0.4917, "step": 11468 }, { "epoch": 0.4807645955377634, "grad_norm": 1.6010633707046509, "learning_rate": 5.553148316231823e-06, "loss": 0.5099, "step": 11469 }, { "epoch": 0.48080651415277237, "grad_norm": 1.7979764938354492, "learning_rate": 5.552473625839458e-06, "loss": 0.4744, "step": 11470 }, { "epoch": 0.48084843276778133, "grad_norm": 1.605991005897522, "learning_rate": 5.5517989252630064e-06, "loss": 0.5284, "step": 11471 }, { "epoch": 0.4808903513827903, "grad_norm": 1.7044274806976318, "learning_rate": 5.551124214514904e-06, "loss": 0.484, "step": 11472 }, { "epoch": 0.48093226999779926, "grad_norm": 1.5560178756713867, "learning_rate": 5.550449493607588e-06, "loss": 0.4985, "step": 11473 }, { "epoch": 0.4809741886128082, "grad_norm": 1.7152765989303589, "learning_rate": 5.549774762553499e-06, "loss": 0.5284, "step": 11474 }, { "epoch": 0.4810161072278172, "grad_norm": 1.602345585823059, "learning_rate": 5.549100021365069e-06, "loss": 0.5283, "step": 11475 }, { "epoch": 0.48105802584282614, "grad_norm": 2.083078384399414, "learning_rate": 5.548425270054742e-06, "loss": 0.5685, "step": 11476 }, { "epoch": 0.4810999444578351, "grad_norm": 1.5950182676315308, "learning_rate": 5.547750508634951e-06, "loss": 0.4705, "step": 11477 }, { "epoch": 0.48114186307284407, "grad_norm": 1.7434687614440918, "learning_rate": 5.547075737118138e-06, "loss": 0.4943, "step": 11478 }, { "epoch": 0.48118378168785303, "grad_norm": 1.377293586730957, "learning_rate": 5.546400955516739e-06, "loss": 0.4706, "step": 11479 }, { "epoch": 0.481225700302862, "grad_norm": 1.6209943294525146, "learning_rate": 5.545726163843196e-06, "loss": 0.4882, "step": 11480 }, { "epoch": 0.48126761891787095, "grad_norm": 1.5695676803588867, "learning_rate": 5.545051362109943e-06, "loss": 0.5109, "step": 11481 }, { "epoch": 0.4813095375328799, "grad_norm": 2.032052516937256, "learning_rate": 5.544376550329423e-06, "loss": 0.4926, "step": 11482 }, { "epoch": 0.4813514561478889, "grad_norm": 1.8048325777053833, "learning_rate": 5.543701728514074e-06, "loss": 0.5071, "step": 11483 }, { "epoch": 0.48139337476289784, "grad_norm": 1.779581069946289, "learning_rate": 5.543026896676336e-06, "loss": 0.5053, "step": 11484 }, { "epoch": 0.4814352933779068, "grad_norm": 2.660641670227051, "learning_rate": 5.542352054828644e-06, "loss": 0.4781, "step": 11485 }, { "epoch": 0.48147721199291577, "grad_norm": 1.7655282020568848, "learning_rate": 5.541677202983445e-06, "loss": 0.5185, "step": 11486 }, { "epoch": 0.48151913060792473, "grad_norm": 1.7788004875183105, "learning_rate": 5.541002341153176e-06, "loss": 0.4867, "step": 11487 }, { "epoch": 0.4815610492229337, "grad_norm": 2.1052086353302, "learning_rate": 5.540327469350275e-06, "loss": 0.5538, "step": 11488 }, { "epoch": 0.48160296783794265, "grad_norm": 1.623617172241211, "learning_rate": 5.539652587587185e-06, "loss": 0.4597, "step": 11489 }, { "epoch": 0.4816448864529516, "grad_norm": 2.093360424041748, "learning_rate": 5.538977695876346e-06, "loss": 0.5049, "step": 11490 }, { "epoch": 0.4816868050679606, "grad_norm": 1.6385084390640259, "learning_rate": 5.538302794230197e-06, "loss": 0.5304, "step": 11491 }, { "epoch": 0.48172872368296954, "grad_norm": 1.7757636308670044, "learning_rate": 5.537627882661181e-06, "loss": 0.4838, "step": 11492 }, { "epoch": 0.4817706422979785, "grad_norm": 1.4151887893676758, "learning_rate": 5.53695296118174e-06, "loss": 0.5273, "step": 11493 }, { "epoch": 0.4818125609129874, "grad_norm": 1.8111628293991089, "learning_rate": 5.536278029804311e-06, "loss": 0.5336, "step": 11494 }, { "epoch": 0.48185447952799637, "grad_norm": 1.8344433307647705, "learning_rate": 5.535603088541338e-06, "loss": 0.4906, "step": 11495 }, { "epoch": 0.48189639814300533, "grad_norm": 1.6561777591705322, "learning_rate": 5.534928137405263e-06, "loss": 0.4703, "step": 11496 }, { "epoch": 0.4819383167580143, "grad_norm": 1.933488130569458, "learning_rate": 5.5342531764085275e-06, "loss": 0.5015, "step": 11497 }, { "epoch": 0.48198023537302326, "grad_norm": 1.655826210975647, "learning_rate": 5.533578205563574e-06, "loss": 0.5196, "step": 11498 }, { "epoch": 0.4820221539880322, "grad_norm": 1.6439212560653687, "learning_rate": 5.532903224882842e-06, "loss": 0.5001, "step": 11499 }, { "epoch": 0.4820640726030412, "grad_norm": 1.636681318283081, "learning_rate": 5.532228234378776e-06, "loss": 0.5317, "step": 11500 }, { "epoch": 0.48210599121805014, "grad_norm": 2.5887231826782227, "learning_rate": 5.531553234063819e-06, "loss": 0.471, "step": 11501 }, { "epoch": 0.4821479098330591, "grad_norm": 1.7662020921707153, "learning_rate": 5.530878223950414e-06, "loss": 0.5376, "step": 11502 }, { "epoch": 0.48218982844806807, "grad_norm": 1.6289218664169312, "learning_rate": 5.530203204051001e-06, "loss": 0.4904, "step": 11503 }, { "epoch": 0.48223174706307703, "grad_norm": 1.5861934423446655, "learning_rate": 5.5295281743780275e-06, "loss": 0.4916, "step": 11504 }, { "epoch": 0.482273665678086, "grad_norm": 1.2713584899902344, "learning_rate": 5.528853134943932e-06, "loss": 0.4693, "step": 11505 }, { "epoch": 0.48231558429309496, "grad_norm": 1.4887254238128662, "learning_rate": 5.52817808576116e-06, "loss": 0.4418, "step": 11506 }, { "epoch": 0.4823575029081039, "grad_norm": 1.719114065170288, "learning_rate": 5.5275030268421555e-06, "loss": 0.5488, "step": 11507 }, { "epoch": 0.4823994215231129, "grad_norm": 1.5570099353790283, "learning_rate": 5.526827958199364e-06, "loss": 0.5438, "step": 11508 }, { "epoch": 0.48244134013812184, "grad_norm": 1.521195411682129, "learning_rate": 5.526152879845223e-06, "loss": 0.4903, "step": 11509 }, { "epoch": 0.4824832587531308, "grad_norm": 1.690859079360962, "learning_rate": 5.5254777917921845e-06, "loss": 0.4742, "step": 11510 }, { "epoch": 0.48252517736813977, "grad_norm": 1.4328973293304443, "learning_rate": 5.524802694052689e-06, "loss": 0.4888, "step": 11511 }, { "epoch": 0.48256709598314873, "grad_norm": 1.7483078241348267, "learning_rate": 5.5241275866391805e-06, "loss": 0.5613, "step": 11512 }, { "epoch": 0.4826090145981577, "grad_norm": 1.6407010555267334, "learning_rate": 5.523452469564105e-06, "loss": 0.456, "step": 11513 }, { "epoch": 0.48265093321316666, "grad_norm": 1.4060420989990234, "learning_rate": 5.522777342839908e-06, "loss": 0.4692, "step": 11514 }, { "epoch": 0.4826928518281756, "grad_norm": 1.6648316383361816, "learning_rate": 5.522102206479031e-06, "loss": 0.5151, "step": 11515 }, { "epoch": 0.4827347704431846, "grad_norm": 2.1138648986816406, "learning_rate": 5.521427060493923e-06, "loss": 0.5318, "step": 11516 }, { "epoch": 0.48277668905819354, "grad_norm": 1.5846121311187744, "learning_rate": 5.520751904897028e-06, "loss": 0.4796, "step": 11517 }, { "epoch": 0.4828186076732025, "grad_norm": 1.845320463180542, "learning_rate": 5.520076739700791e-06, "loss": 0.5, "step": 11518 }, { "epoch": 0.4828605262882114, "grad_norm": 1.9160301685333252, "learning_rate": 5.519401564917659e-06, "loss": 0.5068, "step": 11519 }, { "epoch": 0.4829024449032204, "grad_norm": 1.4260519742965698, "learning_rate": 5.518726380560077e-06, "loss": 0.4456, "step": 11520 }, { "epoch": 0.48294436351822934, "grad_norm": 1.4848698377609253, "learning_rate": 5.518051186640492e-06, "loss": 0.5196, "step": 11521 }, { "epoch": 0.4829862821332383, "grad_norm": 1.9169100522994995, "learning_rate": 5.517375983171348e-06, "loss": 0.4756, "step": 11522 }, { "epoch": 0.48302820074824726, "grad_norm": 2.065615653991699, "learning_rate": 5.516700770165094e-06, "loss": 0.4625, "step": 11523 }, { "epoch": 0.4830701193632562, "grad_norm": 2.5390563011169434, "learning_rate": 5.516025547634175e-06, "loss": 0.5305, "step": 11524 }, { "epoch": 0.4831120379782652, "grad_norm": 1.8614885807037354, "learning_rate": 5.5153503155910385e-06, "loss": 0.5495, "step": 11525 }, { "epoch": 0.48315395659327415, "grad_norm": 1.8009954690933228, "learning_rate": 5.514675074048132e-06, "loss": 0.5092, "step": 11526 }, { "epoch": 0.4831958752082831, "grad_norm": 1.3766862154006958, "learning_rate": 5.513999823017901e-06, "loss": 0.4825, "step": 11527 }, { "epoch": 0.48323779382329207, "grad_norm": 1.6138246059417725, "learning_rate": 5.513324562512794e-06, "loss": 0.4796, "step": 11528 }, { "epoch": 0.48327971243830103, "grad_norm": 1.8157215118408203, "learning_rate": 5.512649292545259e-06, "loss": 0.5139, "step": 11529 }, { "epoch": 0.48332163105331, "grad_norm": 1.724649429321289, "learning_rate": 5.511974013127743e-06, "loss": 0.4737, "step": 11530 }, { "epoch": 0.48336354966831896, "grad_norm": 1.6018379926681519, "learning_rate": 5.511298724272693e-06, "loss": 0.5432, "step": 11531 }, { "epoch": 0.4834054682833279, "grad_norm": 1.6779940128326416, "learning_rate": 5.510623425992559e-06, "loss": 0.519, "step": 11532 }, { "epoch": 0.4834473868983369, "grad_norm": 1.8401323556900024, "learning_rate": 5.5099481182997885e-06, "loss": 0.5046, "step": 11533 }, { "epoch": 0.48348930551334585, "grad_norm": 1.9305537939071655, "learning_rate": 5.5092728012068275e-06, "loss": 0.5219, "step": 11534 }, { "epoch": 0.4835312241283548, "grad_norm": 1.7925699949264526, "learning_rate": 5.508597474726128e-06, "loss": 0.5069, "step": 11535 }, { "epoch": 0.48357314274336377, "grad_norm": 1.398495078086853, "learning_rate": 5.507922138870137e-06, "loss": 0.4783, "step": 11536 }, { "epoch": 0.48361506135837273, "grad_norm": 1.7507563829421997, "learning_rate": 5.507246793651302e-06, "loss": 0.5325, "step": 11537 }, { "epoch": 0.4836569799733817, "grad_norm": 1.9027695655822754, "learning_rate": 5.506571439082076e-06, "loss": 0.524, "step": 11538 }, { "epoch": 0.48369889858839066, "grad_norm": 1.5703760385513306, "learning_rate": 5.505896075174903e-06, "loss": 0.4965, "step": 11539 }, { "epoch": 0.4837408172033996, "grad_norm": 1.6949517726898193, "learning_rate": 5.505220701942238e-06, "loss": 0.5122, "step": 11540 }, { "epoch": 0.4837827358184086, "grad_norm": 1.7093056440353394, "learning_rate": 5.504545319396527e-06, "loss": 0.4635, "step": 11541 }, { "epoch": 0.48382465443341754, "grad_norm": 1.7738115787506104, "learning_rate": 5.5038699275502205e-06, "loss": 0.5114, "step": 11542 }, { "epoch": 0.4838665730484265, "grad_norm": 1.680562973022461, "learning_rate": 5.503194526415767e-06, "loss": 0.4992, "step": 11543 }, { "epoch": 0.4839084916634354, "grad_norm": 1.855246663093567, "learning_rate": 5.50251911600562e-06, "loss": 0.5309, "step": 11544 }, { "epoch": 0.4839504102784444, "grad_norm": 1.565548300743103, "learning_rate": 5.501843696332227e-06, "loss": 0.4631, "step": 11545 }, { "epoch": 0.48399232889345334, "grad_norm": 1.608700156211853, "learning_rate": 5.501168267408039e-06, "loss": 0.5071, "step": 11546 }, { "epoch": 0.4840342475084623, "grad_norm": 1.4987558126449585, "learning_rate": 5.500492829245508e-06, "loss": 0.5413, "step": 11547 }, { "epoch": 0.48407616612347126, "grad_norm": 1.5541253089904785, "learning_rate": 5.499817381857082e-06, "loss": 0.5325, "step": 11548 }, { "epoch": 0.4841180847384802, "grad_norm": 1.7890677452087402, "learning_rate": 5.499141925255214e-06, "loss": 0.4926, "step": 11549 }, { "epoch": 0.4841600033534892, "grad_norm": 1.8703362941741943, "learning_rate": 5.498466459452355e-06, "loss": 0.494, "step": 11550 }, { "epoch": 0.48420192196849815, "grad_norm": 1.4910545349121094, "learning_rate": 5.497790984460957e-06, "loss": 0.518, "step": 11551 }, { "epoch": 0.4842438405835071, "grad_norm": 2.0078582763671875, "learning_rate": 5.4971155002934674e-06, "loss": 0.5055, "step": 11552 }, { "epoch": 0.4842857591985161, "grad_norm": 1.667839765548706, "learning_rate": 5.496440006962344e-06, "loss": 0.492, "step": 11553 }, { "epoch": 0.48432767781352504, "grad_norm": 1.6713799238204956, "learning_rate": 5.495764504480035e-06, "loss": 0.5417, "step": 11554 }, { "epoch": 0.484369596428534, "grad_norm": 1.621360421180725, "learning_rate": 5.4950889928589905e-06, "loss": 0.4566, "step": 11555 }, { "epoch": 0.48441151504354296, "grad_norm": 2.3675825595855713, "learning_rate": 5.494413472111666e-06, "loss": 0.5369, "step": 11556 }, { "epoch": 0.4844534336585519, "grad_norm": 1.5413203239440918, "learning_rate": 5.493737942250512e-06, "loss": 0.5323, "step": 11557 }, { "epoch": 0.4844953522735609, "grad_norm": 1.6305608749389648, "learning_rate": 5.493062403287983e-06, "loss": 0.4702, "step": 11558 }, { "epoch": 0.48453727088856985, "grad_norm": 1.9988312721252441, "learning_rate": 5.49238685523653e-06, "loss": 0.4923, "step": 11559 }, { "epoch": 0.4845791895035788, "grad_norm": 1.6352826356887817, "learning_rate": 5.491711298108605e-06, "loss": 0.5153, "step": 11560 }, { "epoch": 0.4846211081185878, "grad_norm": 1.8360209465026855, "learning_rate": 5.491035731916663e-06, "loss": 0.5364, "step": 11561 }, { "epoch": 0.48466302673359674, "grad_norm": 1.8429701328277588, "learning_rate": 5.490360156673155e-06, "loss": 0.495, "step": 11562 }, { "epoch": 0.4847049453486057, "grad_norm": 1.8265745639801025, "learning_rate": 5.489684572390537e-06, "loss": 0.5272, "step": 11563 }, { "epoch": 0.48474686396361466, "grad_norm": 1.65957772731781, "learning_rate": 5.489008979081261e-06, "loss": 0.5098, "step": 11564 }, { "epoch": 0.4847887825786236, "grad_norm": 1.7723500728607178, "learning_rate": 5.488333376757777e-06, "loss": 0.5167, "step": 11565 }, { "epoch": 0.4848307011936326, "grad_norm": 1.7075053453445435, "learning_rate": 5.487657765432546e-06, "loss": 0.4967, "step": 11566 }, { "epoch": 0.48487261980864155, "grad_norm": 1.9344953298568726, "learning_rate": 5.486982145118017e-06, "loss": 0.4421, "step": 11567 }, { "epoch": 0.4849145384236505, "grad_norm": 1.93052339553833, "learning_rate": 5.486306515826645e-06, "loss": 0.5592, "step": 11568 }, { "epoch": 0.4849564570386594, "grad_norm": 1.7659094333648682, "learning_rate": 5.4856308775708855e-06, "loss": 0.5073, "step": 11569 }, { "epoch": 0.4849983756536684, "grad_norm": 2.3185203075408936, "learning_rate": 5.484955230363193e-06, "loss": 0.4864, "step": 11570 }, { "epoch": 0.48504029426867734, "grad_norm": 1.7199617624282837, "learning_rate": 5.484279574216018e-06, "loss": 0.4607, "step": 11571 }, { "epoch": 0.4850822128836863, "grad_norm": 1.8140791654586792, "learning_rate": 5.483603909141821e-06, "loss": 0.5388, "step": 11572 }, { "epoch": 0.48512413149869527, "grad_norm": 1.82850182056427, "learning_rate": 5.482928235153055e-06, "loss": 0.5426, "step": 11573 }, { "epoch": 0.48516605011370423, "grad_norm": 1.8063241243362427, "learning_rate": 5.482252552262173e-06, "loss": 0.5374, "step": 11574 }, { "epoch": 0.4852079687287132, "grad_norm": 1.8319011926651, "learning_rate": 5.481576860481632e-06, "loss": 0.5457, "step": 11575 }, { "epoch": 0.48524988734372215, "grad_norm": 1.5930956602096558, "learning_rate": 5.480901159823888e-06, "loss": 0.5447, "step": 11576 }, { "epoch": 0.4852918059587311, "grad_norm": 1.5057610273361206, "learning_rate": 5.480225450301397e-06, "loss": 0.4411, "step": 11577 }, { "epoch": 0.4853337245737401, "grad_norm": 1.7144801616668701, "learning_rate": 5.4795497319266124e-06, "loss": 0.5189, "step": 11578 }, { "epoch": 0.48537564318874904, "grad_norm": 1.9019070863723755, "learning_rate": 5.478874004711992e-06, "loss": 0.5073, "step": 11579 }, { "epoch": 0.485417561803758, "grad_norm": 1.5546716451644897, "learning_rate": 5.47819826866999e-06, "loss": 0.5425, "step": 11580 }, { "epoch": 0.48545948041876696, "grad_norm": 1.4504897594451904, "learning_rate": 5.4775225238130645e-06, "loss": 0.456, "step": 11581 }, { "epoch": 0.4855013990337759, "grad_norm": 2.2277345657348633, "learning_rate": 5.476846770153673e-06, "loss": 0.5234, "step": 11582 }, { "epoch": 0.4855433176487849, "grad_norm": 1.7829792499542236, "learning_rate": 5.476171007704269e-06, "loss": 0.4905, "step": 11583 }, { "epoch": 0.48558523626379385, "grad_norm": 1.5989770889282227, "learning_rate": 5.47549523647731e-06, "loss": 0.4724, "step": 11584 }, { "epoch": 0.4856271548788028, "grad_norm": 1.7489315271377563, "learning_rate": 5.474819456485257e-06, "loss": 0.4998, "step": 11585 }, { "epoch": 0.4856690734938118, "grad_norm": 1.9444364309310913, "learning_rate": 5.474143667740562e-06, "loss": 0.5374, "step": 11586 }, { "epoch": 0.48571099210882074, "grad_norm": 1.6790847778320312, "learning_rate": 5.473467870255682e-06, "loss": 0.4606, "step": 11587 }, { "epoch": 0.4857529107238297, "grad_norm": 2.152864933013916, "learning_rate": 5.472792064043079e-06, "loss": 0.4883, "step": 11588 }, { "epoch": 0.48579482933883866, "grad_norm": 1.942949652671814, "learning_rate": 5.4721162491152054e-06, "loss": 0.4939, "step": 11589 }, { "epoch": 0.4858367479538476, "grad_norm": 2.4581925868988037, "learning_rate": 5.471440425484523e-06, "loss": 0.5739, "step": 11590 }, { "epoch": 0.4858786665688566, "grad_norm": 1.8978800773620605, "learning_rate": 5.470764593163489e-06, "loss": 0.5804, "step": 11591 }, { "epoch": 0.48592058518386555, "grad_norm": 3.754869222640991, "learning_rate": 5.470088752164558e-06, "loss": 0.5124, "step": 11592 }, { "epoch": 0.4859625037988745, "grad_norm": 2.7619338035583496, "learning_rate": 5.469412902500193e-06, "loss": 0.5108, "step": 11593 }, { "epoch": 0.4860044224138834, "grad_norm": 1.7805061340332031, "learning_rate": 5.468737044182849e-06, "loss": 0.4469, "step": 11594 }, { "epoch": 0.4860463410288924, "grad_norm": 1.6712146997451782, "learning_rate": 5.468061177224986e-06, "loss": 0.4822, "step": 11595 }, { "epoch": 0.48608825964390134, "grad_norm": 1.6683502197265625, "learning_rate": 5.467385301639063e-06, "loss": 0.4575, "step": 11596 }, { "epoch": 0.4861301782589103, "grad_norm": 1.8659497499465942, "learning_rate": 5.466709417437537e-06, "loss": 0.5183, "step": 11597 }, { "epoch": 0.48617209687391927, "grad_norm": 1.8650798797607422, "learning_rate": 5.466033524632869e-06, "loss": 0.5508, "step": 11598 }, { "epoch": 0.48621401548892823, "grad_norm": 1.7183080911636353, "learning_rate": 5.465357623237515e-06, "loss": 0.5048, "step": 11599 }, { "epoch": 0.4862559341039372, "grad_norm": 1.6901882886886597, "learning_rate": 5.464681713263939e-06, "loss": 0.457, "step": 11600 }, { "epoch": 0.48629785271894616, "grad_norm": 2.0821032524108887, "learning_rate": 5.464005794724599e-06, "loss": 0.4753, "step": 11601 }, { "epoch": 0.4863397713339551, "grad_norm": 1.7551754713058472, "learning_rate": 5.46332986763195e-06, "loss": 0.5056, "step": 11602 }, { "epoch": 0.4863816899489641, "grad_norm": 1.770509123802185, "learning_rate": 5.462653931998457e-06, "loss": 0.5161, "step": 11603 }, { "epoch": 0.48642360856397304, "grad_norm": 1.9086768627166748, "learning_rate": 5.461977987836578e-06, "loss": 0.5292, "step": 11604 }, { "epoch": 0.486465527178982, "grad_norm": 2.0336525440216064, "learning_rate": 5.461302035158775e-06, "loss": 0.4715, "step": 11605 }, { "epoch": 0.48650744579399097, "grad_norm": 1.9515845775604248, "learning_rate": 5.460626073977506e-06, "loss": 0.5247, "step": 11606 }, { "epoch": 0.48654936440899993, "grad_norm": 1.7059354782104492, "learning_rate": 5.459950104305233e-06, "loss": 0.5348, "step": 11607 }, { "epoch": 0.4865912830240089, "grad_norm": 1.91908597946167, "learning_rate": 5.459274126154412e-06, "loss": 0.4803, "step": 11608 }, { "epoch": 0.48663320163901785, "grad_norm": 1.9666712284088135, "learning_rate": 5.458598139537511e-06, "loss": 0.5003, "step": 11609 }, { "epoch": 0.4866751202540268, "grad_norm": 4.516389846801758, "learning_rate": 5.457922144466987e-06, "loss": 0.4886, "step": 11610 }, { "epoch": 0.4867170388690358, "grad_norm": 2.047346830368042, "learning_rate": 5.4572461409553e-06, "loss": 0.5316, "step": 11611 }, { "epoch": 0.48675895748404474, "grad_norm": 1.7784106731414795, "learning_rate": 5.456570129014913e-06, "loss": 0.4854, "step": 11612 }, { "epoch": 0.4868008760990537, "grad_norm": 1.8492587804794312, "learning_rate": 5.455894108658288e-06, "loss": 0.5747, "step": 11613 }, { "epoch": 0.48684279471406267, "grad_norm": 1.93460214138031, "learning_rate": 5.455218079897884e-06, "loss": 0.5286, "step": 11614 }, { "epoch": 0.48688471332907163, "grad_norm": 1.85172438621521, "learning_rate": 5.454542042746165e-06, "loss": 0.551, "step": 11615 }, { "epoch": 0.4869266319440806, "grad_norm": 1.8269069194793701, "learning_rate": 5.453865997215592e-06, "loss": 0.4865, "step": 11616 }, { "epoch": 0.48696855055908955, "grad_norm": 1.6508878469467163, "learning_rate": 5.453189943318624e-06, "loss": 0.4999, "step": 11617 }, { "epoch": 0.4870104691740985, "grad_norm": 1.9414563179016113, "learning_rate": 5.452513881067728e-06, "loss": 0.5011, "step": 11618 }, { "epoch": 0.4870523877891074, "grad_norm": 2.4049556255340576, "learning_rate": 5.451837810475366e-06, "loss": 0.5386, "step": 11619 }, { "epoch": 0.4870943064041164, "grad_norm": 1.7782206535339355, "learning_rate": 5.451161731553996e-06, "loss": 0.5647, "step": 11620 }, { "epoch": 0.48713622501912535, "grad_norm": 2.1894443035125732, "learning_rate": 5.450485644316084e-06, "loss": 0.5402, "step": 11621 }, { "epoch": 0.4871781436341343, "grad_norm": 2.0159645080566406, "learning_rate": 5.449809548774091e-06, "loss": 0.5653, "step": 11622 }, { "epoch": 0.48722006224914327, "grad_norm": 1.805952787399292, "learning_rate": 5.449133444940482e-06, "loss": 0.4634, "step": 11623 }, { "epoch": 0.48726198086415223, "grad_norm": 1.859015941619873, "learning_rate": 5.448457332827719e-06, "loss": 0.441, "step": 11624 }, { "epoch": 0.4873038994791612, "grad_norm": 2.5523600578308105, "learning_rate": 5.447781212448264e-06, "loss": 0.4859, "step": 11625 }, { "epoch": 0.48734581809417016, "grad_norm": 2.433640241622925, "learning_rate": 5.447105083814581e-06, "loss": 0.5217, "step": 11626 }, { "epoch": 0.4873877367091791, "grad_norm": 2.5294864177703857, "learning_rate": 5.446428946939135e-06, "loss": 0.5444, "step": 11627 }, { "epoch": 0.4874296553241881, "grad_norm": 1.7369656562805176, "learning_rate": 5.445752801834388e-06, "loss": 0.5007, "step": 11628 }, { "epoch": 0.48747157393919704, "grad_norm": 3.3132262229919434, "learning_rate": 5.445076648512806e-06, "loss": 0.4881, "step": 11629 }, { "epoch": 0.487513492554206, "grad_norm": 1.8607414960861206, "learning_rate": 5.4444004869868495e-06, "loss": 0.5474, "step": 11630 }, { "epoch": 0.48755541116921497, "grad_norm": 1.5890969038009644, "learning_rate": 5.4437243172689856e-06, "loss": 0.4798, "step": 11631 }, { "epoch": 0.48759732978422393, "grad_norm": 4.358987331390381, "learning_rate": 5.443048139371676e-06, "loss": 0.4986, "step": 11632 }, { "epoch": 0.4876392483992329, "grad_norm": 1.7976430654525757, "learning_rate": 5.442371953307387e-06, "loss": 0.5254, "step": 11633 }, { "epoch": 0.48768116701424186, "grad_norm": 1.8080432415008545, "learning_rate": 5.441695759088583e-06, "loss": 0.4703, "step": 11634 }, { "epoch": 0.4877230856292508, "grad_norm": 1.9238537549972534, "learning_rate": 5.441019556727728e-06, "loss": 0.5074, "step": 11635 }, { "epoch": 0.4877650042442598, "grad_norm": 1.7702785730361938, "learning_rate": 5.440343346237287e-06, "loss": 0.5292, "step": 11636 }, { "epoch": 0.48780692285926874, "grad_norm": 1.6285605430603027, "learning_rate": 5.4396671276297254e-06, "loss": 0.4734, "step": 11637 }, { "epoch": 0.4878488414742777, "grad_norm": 2.235856294631958, "learning_rate": 5.438990900917509e-06, "loss": 0.4627, "step": 11638 }, { "epoch": 0.48789076008928667, "grad_norm": 1.793733835220337, "learning_rate": 5.438314666113101e-06, "loss": 0.4911, "step": 11639 }, { "epoch": 0.48793267870429563, "grad_norm": 1.6269856691360474, "learning_rate": 5.437638423228968e-06, "loss": 0.472, "step": 11640 }, { "epoch": 0.4879745973193046, "grad_norm": 1.5306012630462646, "learning_rate": 5.4369621722775754e-06, "loss": 0.5199, "step": 11641 }, { "epoch": 0.48801651593431356, "grad_norm": 1.5738039016723633, "learning_rate": 5.436285913271391e-06, "loss": 0.4675, "step": 11642 }, { "epoch": 0.4880584345493225, "grad_norm": 2.4605917930603027, "learning_rate": 5.435609646222877e-06, "loss": 0.5181, "step": 11643 }, { "epoch": 0.4881003531643314, "grad_norm": 2.0224108695983887, "learning_rate": 5.434933371144503e-06, "loss": 0.5074, "step": 11644 }, { "epoch": 0.4881422717793404, "grad_norm": 2.099513292312622, "learning_rate": 5.4342570880487315e-06, "loss": 0.5044, "step": 11645 }, { "epoch": 0.48818419039434935, "grad_norm": 2.1530792713165283, "learning_rate": 5.4335807969480325e-06, "loss": 0.5007, "step": 11646 }, { "epoch": 0.4882261090093583, "grad_norm": 1.7752817869186401, "learning_rate": 5.43290449785487e-06, "loss": 0.4898, "step": 11647 }, { "epoch": 0.4882680276243673, "grad_norm": 1.8062094449996948, "learning_rate": 5.432228190781712e-06, "loss": 0.5123, "step": 11648 }, { "epoch": 0.48830994623937624, "grad_norm": 2.174349069595337, "learning_rate": 5.431551875741026e-06, "loss": 0.5282, "step": 11649 }, { "epoch": 0.4883518648543852, "grad_norm": 1.7476428747177124, "learning_rate": 5.430875552745276e-06, "loss": 0.5569, "step": 11650 }, { "epoch": 0.48839378346939416, "grad_norm": 2.291496992111206, "learning_rate": 5.4301992218069324e-06, "loss": 0.4557, "step": 11651 }, { "epoch": 0.4884357020844031, "grad_norm": 1.7414140701293945, "learning_rate": 5.429522882938461e-06, "loss": 0.4426, "step": 11652 }, { "epoch": 0.4884776206994121, "grad_norm": 1.5635048151016235, "learning_rate": 5.428846536152327e-06, "loss": 0.4474, "step": 11653 }, { "epoch": 0.48851953931442105, "grad_norm": 1.6564289331436157, "learning_rate": 5.428170181461e-06, "loss": 0.4797, "step": 11654 }, { "epoch": 0.48856145792943, "grad_norm": 1.7670009136199951, "learning_rate": 5.427493818876949e-06, "loss": 0.489, "step": 11655 }, { "epoch": 0.48860337654443897, "grad_norm": 4.1503190994262695, "learning_rate": 5.42681744841264e-06, "loss": 0.4913, "step": 11656 }, { "epoch": 0.48864529515944793, "grad_norm": 1.8338279724121094, "learning_rate": 5.426141070080542e-06, "loss": 0.5712, "step": 11657 }, { "epoch": 0.4886872137744569, "grad_norm": 1.861071228981018, "learning_rate": 5.4254646838931224e-06, "loss": 0.4798, "step": 11658 }, { "epoch": 0.48872913238946586, "grad_norm": 2.01648211479187, "learning_rate": 5.424788289862849e-06, "loss": 0.5202, "step": 11659 }, { "epoch": 0.4887710510044748, "grad_norm": 1.9545786380767822, "learning_rate": 5.424111888002191e-06, "loss": 0.5179, "step": 11660 }, { "epoch": 0.4888129696194838, "grad_norm": 2.006664991378784, "learning_rate": 5.423435478323618e-06, "loss": 0.4762, "step": 11661 }, { "epoch": 0.48885488823449275, "grad_norm": 1.8701282739639282, "learning_rate": 5.422759060839595e-06, "loss": 0.5009, "step": 11662 }, { "epoch": 0.4888968068495017, "grad_norm": 1.9570326805114746, "learning_rate": 5.422082635562593e-06, "loss": 0.4933, "step": 11663 }, { "epoch": 0.48893872546451067, "grad_norm": 1.74179208278656, "learning_rate": 5.421406202505083e-06, "loss": 0.5057, "step": 11664 }, { "epoch": 0.48898064407951963, "grad_norm": 2.1219208240509033, "learning_rate": 5.420729761679532e-06, "loss": 0.5805, "step": 11665 }, { "epoch": 0.4890225626945286, "grad_norm": 1.840804934501648, "learning_rate": 5.420053313098408e-06, "loss": 0.5677, "step": 11666 }, { "epoch": 0.48906448130953756, "grad_norm": 2.0126516819000244, "learning_rate": 5.419376856774183e-06, "loss": 0.5213, "step": 11667 }, { "epoch": 0.4891063999245465, "grad_norm": 2.1379690170288086, "learning_rate": 5.418700392719326e-06, "loss": 0.4877, "step": 11668 }, { "epoch": 0.4891483185395554, "grad_norm": 1.6351665258407593, "learning_rate": 5.418023920946305e-06, "loss": 0.4933, "step": 11669 }, { "epoch": 0.4891902371545644, "grad_norm": 2.057762384414673, "learning_rate": 5.417347441467592e-06, "loss": 0.5461, "step": 11670 }, { "epoch": 0.48923215576957335, "grad_norm": 1.9426838159561157, "learning_rate": 5.416670954295656e-06, "loss": 0.5512, "step": 11671 }, { "epoch": 0.4892740743845823, "grad_norm": 1.6574389934539795, "learning_rate": 5.415994459442965e-06, "loss": 0.4621, "step": 11672 }, { "epoch": 0.4893159929995913, "grad_norm": 1.6702126264572144, "learning_rate": 5.415317956921992e-06, "loss": 0.4947, "step": 11673 }, { "epoch": 0.48935791161460024, "grad_norm": 1.9107062816619873, "learning_rate": 5.4146414467452065e-06, "loss": 0.5159, "step": 11674 }, { "epoch": 0.4893998302296092, "grad_norm": 1.5245882272720337, "learning_rate": 5.413964928925079e-06, "loss": 0.4847, "step": 11675 }, { "epoch": 0.48944174884461816, "grad_norm": 2.0486268997192383, "learning_rate": 5.413288403474082e-06, "loss": 0.4755, "step": 11676 }, { "epoch": 0.4894836674596271, "grad_norm": 1.8436837196350098, "learning_rate": 5.412611870404684e-06, "loss": 0.4857, "step": 11677 }, { "epoch": 0.4895255860746361, "grad_norm": 1.828706979751587, "learning_rate": 5.411935329729357e-06, "loss": 0.4824, "step": 11678 }, { "epoch": 0.48956750468964505, "grad_norm": 1.5931181907653809, "learning_rate": 5.411258781460569e-06, "loss": 0.4524, "step": 11679 }, { "epoch": 0.489609423304654, "grad_norm": 1.8254948854446411, "learning_rate": 5.410582225610797e-06, "loss": 0.5359, "step": 11680 }, { "epoch": 0.489651341919663, "grad_norm": 2.031507968902588, "learning_rate": 5.409905662192509e-06, "loss": 0.5146, "step": 11681 }, { "epoch": 0.48969326053467194, "grad_norm": 1.8914778232574463, "learning_rate": 5.409229091218173e-06, "loss": 0.5611, "step": 11682 }, { "epoch": 0.4897351791496809, "grad_norm": 4.761115550994873, "learning_rate": 5.408552512700268e-06, "loss": 0.5143, "step": 11683 }, { "epoch": 0.48977709776468986, "grad_norm": 2.7921407222747803, "learning_rate": 5.407875926651261e-06, "loss": 0.5001, "step": 11684 }, { "epoch": 0.4898190163796988, "grad_norm": 1.68576979637146, "learning_rate": 5.407199333083627e-06, "loss": 0.5515, "step": 11685 }, { "epoch": 0.4898609349947078, "grad_norm": 1.9923583269119263, "learning_rate": 5.406522732009835e-06, "loss": 0.5125, "step": 11686 }, { "epoch": 0.48990285360971675, "grad_norm": 2.6821677684783936, "learning_rate": 5.405846123442358e-06, "loss": 0.4919, "step": 11687 }, { "epoch": 0.4899447722247257, "grad_norm": 1.7626910209655762, "learning_rate": 5.405169507393669e-06, "loss": 0.4943, "step": 11688 }, { "epoch": 0.4899866908397347, "grad_norm": 1.8192476034164429, "learning_rate": 5.404492883876241e-06, "loss": 0.5066, "step": 11689 }, { "epoch": 0.49002860945474364, "grad_norm": 2.2562952041625977, "learning_rate": 5.403816252902545e-06, "loss": 0.5212, "step": 11690 }, { "epoch": 0.4900705280697526, "grad_norm": 1.7446479797363281, "learning_rate": 5.403139614485054e-06, "loss": 0.4744, "step": 11691 }, { "epoch": 0.49011244668476156, "grad_norm": 2.8805294036865234, "learning_rate": 5.402462968636243e-06, "loss": 0.5763, "step": 11692 }, { "epoch": 0.4901543652997705, "grad_norm": 7.519794940948486, "learning_rate": 5.401786315368583e-06, "loss": 0.4441, "step": 11693 }, { "epoch": 0.49019628391477943, "grad_norm": 1.6291159391403198, "learning_rate": 5.401109654694549e-06, "loss": 0.5497, "step": 11694 }, { "epoch": 0.4902382025297884, "grad_norm": 1.9088079929351807, "learning_rate": 5.400432986626611e-06, "loss": 0.5059, "step": 11695 }, { "epoch": 0.49028012114479735, "grad_norm": 1.5698878765106201, "learning_rate": 5.399756311177246e-06, "loss": 0.5155, "step": 11696 }, { "epoch": 0.4903220397598063, "grad_norm": 1.7934249639511108, "learning_rate": 5.399079628358924e-06, "loss": 0.5432, "step": 11697 }, { "epoch": 0.4903639583748153, "grad_norm": 1.7013062238693237, "learning_rate": 5.398402938184122e-06, "loss": 0.459, "step": 11698 }, { "epoch": 0.49040587698982424, "grad_norm": 1.7888474464416504, "learning_rate": 5.397726240665314e-06, "loss": 0.5381, "step": 11699 }, { "epoch": 0.4904477956048332, "grad_norm": 1.993380069732666, "learning_rate": 5.39704953581497e-06, "loss": 0.5187, "step": 11700 }, { "epoch": 0.49048971421984217, "grad_norm": 2.0345394611358643, "learning_rate": 5.396372823645568e-06, "loss": 0.4998, "step": 11701 }, { "epoch": 0.49053163283485113, "grad_norm": 1.739517331123352, "learning_rate": 5.395696104169581e-06, "loss": 0.5207, "step": 11702 }, { "epoch": 0.4905735514498601, "grad_norm": 1.8051315546035767, "learning_rate": 5.395019377399484e-06, "loss": 0.4863, "step": 11703 }, { "epoch": 0.49061547006486905, "grad_norm": 2.197404623031616, "learning_rate": 5.394342643347749e-06, "loss": 0.5105, "step": 11704 }, { "epoch": 0.490657388679878, "grad_norm": 1.9302430152893066, "learning_rate": 5.393665902026854e-06, "loss": 0.5033, "step": 11705 }, { "epoch": 0.490699307294887, "grad_norm": 1.817189335823059, "learning_rate": 5.392989153449273e-06, "loss": 0.4888, "step": 11706 }, { "epoch": 0.49074122590989594, "grad_norm": 1.9040672779083252, "learning_rate": 5.392312397627479e-06, "loss": 0.4706, "step": 11707 }, { "epoch": 0.4907831445249049, "grad_norm": 1.8243309259414673, "learning_rate": 5.391635634573949e-06, "loss": 0.4306, "step": 11708 }, { "epoch": 0.49082506313991386, "grad_norm": 1.993849515914917, "learning_rate": 5.390958864301157e-06, "loss": 0.5796, "step": 11709 }, { "epoch": 0.4908669817549228, "grad_norm": 1.55641770362854, "learning_rate": 5.390282086821577e-06, "loss": 0.4912, "step": 11710 }, { "epoch": 0.4909089003699318, "grad_norm": 2.2466673851013184, "learning_rate": 5.389605302147689e-06, "loss": 0.4662, "step": 11711 }, { "epoch": 0.49095081898494075, "grad_norm": 2.482008457183838, "learning_rate": 5.388928510291965e-06, "loss": 0.5332, "step": 11712 }, { "epoch": 0.4909927375999497, "grad_norm": 1.6447542905807495, "learning_rate": 5.388251711266882e-06, "loss": 0.5383, "step": 11713 }, { "epoch": 0.4910346562149587, "grad_norm": 1.901390790939331, "learning_rate": 5.387574905084916e-06, "loss": 0.4626, "step": 11714 }, { "epoch": 0.49107657482996764, "grad_norm": 1.7902239561080933, "learning_rate": 5.386898091758541e-06, "loss": 0.4769, "step": 11715 }, { "epoch": 0.4911184934449766, "grad_norm": 2.102581024169922, "learning_rate": 5.386221271300233e-06, "loss": 0.4965, "step": 11716 }, { "epoch": 0.49116041205998556, "grad_norm": 1.912703275680542, "learning_rate": 5.385544443722475e-06, "loss": 0.5203, "step": 11717 }, { "epoch": 0.4912023306749945, "grad_norm": 2.049281120300293, "learning_rate": 5.384867609037735e-06, "loss": 0.496, "step": 11718 }, { "epoch": 0.49124424929000343, "grad_norm": 4.374125957489014, "learning_rate": 5.384190767258492e-06, "loss": 0.5036, "step": 11719 }, { "epoch": 0.4912861679050124, "grad_norm": 1.7558475732803345, "learning_rate": 5.383513918397224e-06, "loss": 0.5112, "step": 11720 }, { "epoch": 0.49132808652002136, "grad_norm": 2.0586273670196533, "learning_rate": 5.382837062466407e-06, "loss": 0.5397, "step": 11721 }, { "epoch": 0.4913700051350303, "grad_norm": 1.6873211860656738, "learning_rate": 5.382160199478518e-06, "loss": 0.4644, "step": 11722 }, { "epoch": 0.4914119237500393, "grad_norm": 2.2053582668304443, "learning_rate": 5.381483329446035e-06, "loss": 0.4897, "step": 11723 }, { "epoch": 0.49145384236504824, "grad_norm": 1.7032475471496582, "learning_rate": 5.380806452381433e-06, "loss": 0.542, "step": 11724 }, { "epoch": 0.4914957609800572, "grad_norm": 1.9171967506408691, "learning_rate": 5.38012956829719e-06, "loss": 0.5386, "step": 11725 }, { "epoch": 0.49153767959506617, "grad_norm": 1.66884183883667, "learning_rate": 5.379452677205785e-06, "loss": 0.5015, "step": 11726 }, { "epoch": 0.49157959821007513, "grad_norm": 1.6540602445602417, "learning_rate": 5.378775779119696e-06, "loss": 0.4919, "step": 11727 }, { "epoch": 0.4916215168250841, "grad_norm": 1.5936657190322876, "learning_rate": 5.3780988740513964e-06, "loss": 0.5105, "step": 11728 }, { "epoch": 0.49166343544009306, "grad_norm": 1.678377628326416, "learning_rate": 5.377421962013367e-06, "loss": 0.5147, "step": 11729 }, { "epoch": 0.491705354055102, "grad_norm": 1.7078609466552734, "learning_rate": 5.376745043018086e-06, "loss": 0.502, "step": 11730 }, { "epoch": 0.491747272670111, "grad_norm": 1.87528657913208, "learning_rate": 5.376068117078032e-06, "loss": 0.4947, "step": 11731 }, { "epoch": 0.49178919128511994, "grad_norm": 2.5247068405151367, "learning_rate": 5.375391184205681e-06, "loss": 0.5062, "step": 11732 }, { "epoch": 0.4918311099001289, "grad_norm": 1.8127013444900513, "learning_rate": 5.374714244413514e-06, "loss": 0.5333, "step": 11733 }, { "epoch": 0.49187302851513787, "grad_norm": 2.193240165710449, "learning_rate": 5.374037297714006e-06, "loss": 0.542, "step": 11734 }, { "epoch": 0.49191494713014683, "grad_norm": 1.7644468545913696, "learning_rate": 5.373360344119638e-06, "loss": 0.5238, "step": 11735 }, { "epoch": 0.4919568657451558, "grad_norm": 1.771351933479309, "learning_rate": 5.37268338364289e-06, "loss": 0.4899, "step": 11736 }, { "epoch": 0.49199878436016475, "grad_norm": 1.6367334127426147, "learning_rate": 5.372006416296237e-06, "loss": 0.5276, "step": 11737 }, { "epoch": 0.4920407029751737, "grad_norm": 2.023267984390259, "learning_rate": 5.371329442092161e-06, "loss": 0.5484, "step": 11738 }, { "epoch": 0.4920826215901827, "grad_norm": 1.8313250541687012, "learning_rate": 5.37065246104314e-06, "loss": 0.4795, "step": 11739 }, { "epoch": 0.49212454020519164, "grad_norm": 2.7180917263031006, "learning_rate": 5.3699754731616525e-06, "loss": 0.5249, "step": 11740 }, { "epoch": 0.4921664588202006, "grad_norm": 2.2571020126342773, "learning_rate": 5.36929847846018e-06, "loss": 0.478, "step": 11741 }, { "epoch": 0.49220837743520957, "grad_norm": 1.6353517770767212, "learning_rate": 5.368621476951199e-06, "loss": 0.4843, "step": 11742 }, { "epoch": 0.4922502960502185, "grad_norm": 1.7667728662490845, "learning_rate": 5.3679444686471914e-06, "loss": 0.5242, "step": 11743 }, { "epoch": 0.49229221466522743, "grad_norm": 1.7206828594207764, "learning_rate": 5.3672674535606365e-06, "loss": 0.5277, "step": 11744 }, { "epoch": 0.4923341332802364, "grad_norm": 1.837562084197998, "learning_rate": 5.366590431704013e-06, "loss": 0.4964, "step": 11745 }, { "epoch": 0.49237605189524536, "grad_norm": 1.6467595100402832, "learning_rate": 5.365913403089803e-06, "loss": 0.461, "step": 11746 }, { "epoch": 0.4924179705102543, "grad_norm": 1.7665501832962036, "learning_rate": 5.365236367730484e-06, "loss": 0.489, "step": 11747 }, { "epoch": 0.4924598891252633, "grad_norm": 5.180134296417236, "learning_rate": 5.3645593256385375e-06, "loss": 0.5126, "step": 11748 }, { "epoch": 0.49250180774027225, "grad_norm": 2.0870611667633057, "learning_rate": 5.363882276826445e-06, "loss": 0.4931, "step": 11749 }, { "epoch": 0.4925437263552812, "grad_norm": 1.8830218315124512, "learning_rate": 5.363205221306685e-06, "loss": 0.489, "step": 11750 }, { "epoch": 0.49258564497029017, "grad_norm": 2.5857818126678467, "learning_rate": 5.362528159091739e-06, "loss": 0.4903, "step": 11751 }, { "epoch": 0.49262756358529913, "grad_norm": 1.7256073951721191, "learning_rate": 5.361851090194088e-06, "loss": 0.4604, "step": 11752 }, { "epoch": 0.4926694822003081, "grad_norm": 1.6175678968429565, "learning_rate": 5.36117401462621e-06, "loss": 0.4885, "step": 11753 }, { "epoch": 0.49271140081531706, "grad_norm": 2.3485758304595947, "learning_rate": 5.360496932400591e-06, "loss": 0.5489, "step": 11754 }, { "epoch": 0.492753319430326, "grad_norm": 1.625577688217163, "learning_rate": 5.35981984352971e-06, "loss": 0.5415, "step": 11755 }, { "epoch": 0.492795238045335, "grad_norm": 2.1545329093933105, "learning_rate": 5.359142748026045e-06, "loss": 0.497, "step": 11756 }, { "epoch": 0.49283715666034394, "grad_norm": 1.8503751754760742, "learning_rate": 5.358465645902081e-06, "loss": 0.5005, "step": 11757 }, { "epoch": 0.4928790752753529, "grad_norm": 1.528557300567627, "learning_rate": 5.357788537170299e-06, "loss": 0.4355, "step": 11758 }, { "epoch": 0.49292099389036187, "grad_norm": 4.261449813842773, "learning_rate": 5.3571114218431795e-06, "loss": 0.5201, "step": 11759 }, { "epoch": 0.49296291250537083, "grad_norm": 2.86156964302063, "learning_rate": 5.356434299933205e-06, "loss": 0.5211, "step": 11760 }, { "epoch": 0.4930048311203798, "grad_norm": 1.7187011241912842, "learning_rate": 5.355757171452857e-06, "loss": 0.4755, "step": 11761 }, { "epoch": 0.49304674973538876, "grad_norm": 2.050933361053467, "learning_rate": 5.355080036414616e-06, "loss": 0.5131, "step": 11762 }, { "epoch": 0.4930886683503977, "grad_norm": 1.7827889919281006, "learning_rate": 5.354402894830967e-06, "loss": 0.4947, "step": 11763 }, { "epoch": 0.4931305869654067, "grad_norm": 1.6849827766418457, "learning_rate": 5.353725746714391e-06, "loss": 0.5292, "step": 11764 }, { "epoch": 0.49317250558041564, "grad_norm": 1.562822937965393, "learning_rate": 5.3530485920773675e-06, "loss": 0.4908, "step": 11765 }, { "epoch": 0.4932144241954246, "grad_norm": 1.6551761627197266, "learning_rate": 5.352371430932383e-06, "loss": 0.5256, "step": 11766 }, { "epoch": 0.49325634281043357, "grad_norm": 1.5948814153671265, "learning_rate": 5.351694263291918e-06, "loss": 0.5143, "step": 11767 }, { "epoch": 0.49329826142544253, "grad_norm": 1.99015212059021, "learning_rate": 5.351017089168455e-06, "loss": 0.4732, "step": 11768 }, { "epoch": 0.49334018004045144, "grad_norm": 1.8447011709213257, "learning_rate": 5.350339908574478e-06, "loss": 0.5135, "step": 11769 }, { "epoch": 0.4933820986554604, "grad_norm": 1.5055657625198364, "learning_rate": 5.34966272152247e-06, "loss": 0.492, "step": 11770 }, { "epoch": 0.49342401727046936, "grad_norm": 2.003556966781616, "learning_rate": 5.34898552802491e-06, "loss": 0.4995, "step": 11771 }, { "epoch": 0.4934659358854783, "grad_norm": 1.5318204164505005, "learning_rate": 5.348308328094286e-06, "loss": 0.5054, "step": 11772 }, { "epoch": 0.4935078545004873, "grad_norm": 1.9336214065551758, "learning_rate": 5.34763112174308e-06, "loss": 0.5085, "step": 11773 }, { "epoch": 0.49354977311549625, "grad_norm": 1.9726356267929077, "learning_rate": 5.3469539089837744e-06, "loss": 0.5171, "step": 11774 }, { "epoch": 0.4935916917305052, "grad_norm": 1.6779597997665405, "learning_rate": 5.3462766898288535e-06, "loss": 0.5121, "step": 11775 }, { "epoch": 0.4936336103455142, "grad_norm": 2.020587921142578, "learning_rate": 5.345599464290801e-06, "loss": 0.5079, "step": 11776 }, { "epoch": 0.49367552896052314, "grad_norm": 1.5596671104431152, "learning_rate": 5.344922232382098e-06, "loss": 0.4819, "step": 11777 }, { "epoch": 0.4937174475755321, "grad_norm": 1.6428166627883911, "learning_rate": 5.344244994115233e-06, "loss": 0.522, "step": 11778 }, { "epoch": 0.49375936619054106, "grad_norm": 2.103330135345459, "learning_rate": 5.343567749502686e-06, "loss": 0.4769, "step": 11779 }, { "epoch": 0.49380128480555, "grad_norm": 1.5519410371780396, "learning_rate": 5.342890498556942e-06, "loss": 0.472, "step": 11780 }, { "epoch": 0.493843203420559, "grad_norm": 1.6871662139892578, "learning_rate": 5.342213241290487e-06, "loss": 0.5454, "step": 11781 }, { "epoch": 0.49388512203556795, "grad_norm": 1.8601977825164795, "learning_rate": 5.341535977715804e-06, "loss": 0.4882, "step": 11782 }, { "epoch": 0.4939270406505769, "grad_norm": 1.5744775533676147, "learning_rate": 5.340858707845378e-06, "loss": 0.4497, "step": 11783 }, { "epoch": 0.49396895926558587, "grad_norm": 1.5762325525283813, "learning_rate": 5.34018143169169e-06, "loss": 0.5088, "step": 11784 }, { "epoch": 0.49401087788059483, "grad_norm": 1.8469387292861938, "learning_rate": 5.3395041492672296e-06, "loss": 0.5011, "step": 11785 }, { "epoch": 0.4940527964956038, "grad_norm": 1.6626074314117432, "learning_rate": 5.33882686058448e-06, "loss": 0.4776, "step": 11786 }, { "epoch": 0.49409471511061276, "grad_norm": 1.6192625761032104, "learning_rate": 5.338149565655924e-06, "loss": 0.5084, "step": 11787 }, { "epoch": 0.4941366337256217, "grad_norm": 1.4761176109313965, "learning_rate": 5.33747226449405e-06, "loss": 0.4846, "step": 11788 }, { "epoch": 0.4941785523406307, "grad_norm": 1.7057137489318848, "learning_rate": 5.3367949571113405e-06, "loss": 0.4837, "step": 11789 }, { "epoch": 0.49422047095563965, "grad_norm": 1.8605540990829468, "learning_rate": 5.336117643520281e-06, "loss": 0.4552, "step": 11790 }, { "epoch": 0.4942623895706486, "grad_norm": 1.8559162616729736, "learning_rate": 5.335440323733358e-06, "loss": 0.5069, "step": 11791 }, { "epoch": 0.49430430818565757, "grad_norm": 1.849738359451294, "learning_rate": 5.334762997763057e-06, "loss": 0.5161, "step": 11792 }, { "epoch": 0.49434622680066653, "grad_norm": 1.7496908903121948, "learning_rate": 5.334085665621863e-06, "loss": 0.5146, "step": 11793 }, { "epoch": 0.49438814541567544, "grad_norm": 1.984166145324707, "learning_rate": 5.33340832732226e-06, "loss": 0.5456, "step": 11794 }, { "epoch": 0.4944300640306844, "grad_norm": 1.5149871110916138, "learning_rate": 5.332730982876737e-06, "loss": 0.5056, "step": 11795 }, { "epoch": 0.49447198264569336, "grad_norm": 2.302992582321167, "learning_rate": 5.332053632297777e-06, "loss": 0.5214, "step": 11796 }, { "epoch": 0.4945139012607023, "grad_norm": 1.7810478210449219, "learning_rate": 5.33137627559787e-06, "loss": 0.5216, "step": 11797 }, { "epoch": 0.4945558198757113, "grad_norm": 1.9261761903762817, "learning_rate": 5.330698912789497e-06, "loss": 0.5047, "step": 11798 }, { "epoch": 0.49459773849072025, "grad_norm": 2.0338878631591797, "learning_rate": 5.330021543885147e-06, "loss": 0.5493, "step": 11799 }, { "epoch": 0.4946396571057292, "grad_norm": 1.5051780939102173, "learning_rate": 5.329344168897306e-06, "loss": 0.4519, "step": 11800 }, { "epoch": 0.4946815757207382, "grad_norm": 1.701248049736023, "learning_rate": 5.328666787838461e-06, "loss": 0.4944, "step": 11801 }, { "epoch": 0.49472349433574714, "grad_norm": 1.9787925481796265, "learning_rate": 5.327989400721099e-06, "loss": 0.5713, "step": 11802 }, { "epoch": 0.4947654129507561, "grad_norm": 1.8726942539215088, "learning_rate": 5.327312007557704e-06, "loss": 0.5154, "step": 11803 }, { "epoch": 0.49480733156576506, "grad_norm": 1.9245489835739136, "learning_rate": 5.326634608360766e-06, "loss": 0.5087, "step": 11804 }, { "epoch": 0.494849250180774, "grad_norm": 1.703505516052246, "learning_rate": 5.3259572031427685e-06, "loss": 0.4387, "step": 11805 }, { "epoch": 0.494891168795783, "grad_norm": 1.4685876369476318, "learning_rate": 5.325279791916204e-06, "loss": 0.4874, "step": 11806 }, { "epoch": 0.49493308741079195, "grad_norm": 1.8249719142913818, "learning_rate": 5.324602374693554e-06, "loss": 0.5606, "step": 11807 }, { "epoch": 0.4949750060258009, "grad_norm": 1.6593016386032104, "learning_rate": 5.323924951487307e-06, "loss": 0.5749, "step": 11808 }, { "epoch": 0.4950169246408099, "grad_norm": 1.8642491102218628, "learning_rate": 5.3232475223099534e-06, "loss": 0.4865, "step": 11809 }, { "epoch": 0.49505884325581884, "grad_norm": 1.782719612121582, "learning_rate": 5.322570087173978e-06, "loss": 0.5577, "step": 11810 }, { "epoch": 0.4951007618708278, "grad_norm": 1.763983964920044, "learning_rate": 5.321892646091869e-06, "loss": 0.5358, "step": 11811 }, { "epoch": 0.49514268048583676, "grad_norm": 2.5700485706329346, "learning_rate": 5.321215199076114e-06, "loss": 0.5195, "step": 11812 }, { "epoch": 0.4951845991008457, "grad_norm": 1.8396254777908325, "learning_rate": 5.3205377461392e-06, "loss": 0.4786, "step": 11813 }, { "epoch": 0.4952265177158547, "grad_norm": 1.9367482662200928, "learning_rate": 5.3198602872936165e-06, "loss": 0.5142, "step": 11814 }, { "epoch": 0.49526843633086365, "grad_norm": 1.6605806350708008, "learning_rate": 5.319182822551852e-06, "loss": 0.5092, "step": 11815 }, { "epoch": 0.4953103549458726, "grad_norm": 1.6526786088943481, "learning_rate": 5.318505351926393e-06, "loss": 0.4931, "step": 11816 }, { "epoch": 0.4953522735608816, "grad_norm": 1.6367849111557007, "learning_rate": 5.317827875429727e-06, "loss": 0.5174, "step": 11817 }, { "epoch": 0.49539419217589054, "grad_norm": 1.5329489707946777, "learning_rate": 5.317150393074343e-06, "loss": 0.4867, "step": 11818 }, { "epoch": 0.49543611079089944, "grad_norm": 1.4864542484283447, "learning_rate": 5.316472904872732e-06, "loss": 0.5437, "step": 11819 }, { "epoch": 0.4954780294059084, "grad_norm": 1.677970051765442, "learning_rate": 5.3157954108373785e-06, "loss": 0.5003, "step": 11820 }, { "epoch": 0.49551994802091737, "grad_norm": 2.077186346054077, "learning_rate": 5.3151179109807745e-06, "loss": 0.5352, "step": 11821 }, { "epoch": 0.49556186663592633, "grad_norm": 1.6154764890670776, "learning_rate": 5.314440405315407e-06, "loss": 0.5244, "step": 11822 }, { "epoch": 0.4956037852509353, "grad_norm": 1.7839457988739014, "learning_rate": 5.313762893853766e-06, "loss": 0.5444, "step": 11823 }, { "epoch": 0.49564570386594425, "grad_norm": 4.607604026794434, "learning_rate": 5.313085376608338e-06, "loss": 0.4834, "step": 11824 }, { "epoch": 0.4956876224809532, "grad_norm": 1.5846024751663208, "learning_rate": 5.312407853591616e-06, "loss": 0.4606, "step": 11825 }, { "epoch": 0.4957295410959622, "grad_norm": 1.6273032426834106, "learning_rate": 5.311730324816086e-06, "loss": 0.5132, "step": 11826 }, { "epoch": 0.49577145971097114, "grad_norm": 2.058011770248413, "learning_rate": 5.311052790294238e-06, "loss": 0.4577, "step": 11827 }, { "epoch": 0.4958133783259801, "grad_norm": 1.4638190269470215, "learning_rate": 5.310375250038563e-06, "loss": 0.4711, "step": 11828 }, { "epoch": 0.49585529694098907, "grad_norm": 1.7588684558868408, "learning_rate": 5.309697704061548e-06, "loss": 0.5382, "step": 11829 }, { "epoch": 0.495897215555998, "grad_norm": 1.5614547729492188, "learning_rate": 5.309020152375686e-06, "loss": 0.4898, "step": 11830 }, { "epoch": 0.495939134171007, "grad_norm": 1.5035426616668701, "learning_rate": 5.308342594993463e-06, "loss": 0.4512, "step": 11831 }, { "epoch": 0.49598105278601595, "grad_norm": 1.8841524124145508, "learning_rate": 5.307665031927371e-06, "loss": 0.4858, "step": 11832 }, { "epoch": 0.4960229714010249, "grad_norm": 1.5393505096435547, "learning_rate": 5.306987463189899e-06, "loss": 0.5067, "step": 11833 }, { "epoch": 0.4960648900160339, "grad_norm": 1.6577181816101074, "learning_rate": 5.306309888793539e-06, "loss": 0.4788, "step": 11834 }, { "epoch": 0.49610680863104284, "grad_norm": 1.593801736831665, "learning_rate": 5.305632308750779e-06, "loss": 0.5032, "step": 11835 }, { "epoch": 0.4961487272460518, "grad_norm": 1.6188467741012573, "learning_rate": 5.304954723074108e-06, "loss": 0.5216, "step": 11836 }, { "epoch": 0.49619064586106076, "grad_norm": 1.751652479171753, "learning_rate": 5.30427713177602e-06, "loss": 0.4638, "step": 11837 }, { "epoch": 0.4962325644760697, "grad_norm": 1.705855131149292, "learning_rate": 5.303599534869004e-06, "loss": 0.5096, "step": 11838 }, { "epoch": 0.4962744830910787, "grad_norm": 1.804930329322815, "learning_rate": 5.30292193236555e-06, "loss": 0.4815, "step": 11839 }, { "epoch": 0.49631640170608765, "grad_norm": 1.5684683322906494, "learning_rate": 5.302244324278149e-06, "loss": 0.5338, "step": 11840 }, { "epoch": 0.4963583203210966, "grad_norm": 1.873637080192566, "learning_rate": 5.301566710619291e-06, "loss": 0.5171, "step": 11841 }, { "epoch": 0.4964002389361056, "grad_norm": 1.6351662874221802, "learning_rate": 5.300889091401468e-06, "loss": 0.4935, "step": 11842 }, { "epoch": 0.49644215755111454, "grad_norm": 1.5475112199783325, "learning_rate": 5.300211466637171e-06, "loss": 0.5311, "step": 11843 }, { "epoch": 0.49648407616612344, "grad_norm": 1.7565127611160278, "learning_rate": 5.2995338363388904e-06, "loss": 0.4929, "step": 11844 }, { "epoch": 0.4965259947811324, "grad_norm": 1.7357803583145142, "learning_rate": 5.2988562005191166e-06, "loss": 0.5437, "step": 11845 }, { "epoch": 0.49656791339614137, "grad_norm": 1.8730432987213135, "learning_rate": 5.298178559190344e-06, "loss": 0.4584, "step": 11846 }, { "epoch": 0.49660983201115033, "grad_norm": 1.8517937660217285, "learning_rate": 5.29750091236506e-06, "loss": 0.5205, "step": 11847 }, { "epoch": 0.4966517506261593, "grad_norm": 1.6920093297958374, "learning_rate": 5.296823260055759e-06, "loss": 0.4853, "step": 11848 }, { "epoch": 0.49669366924116826, "grad_norm": 1.7416518926620483, "learning_rate": 5.296145602274932e-06, "loss": 0.4969, "step": 11849 }, { "epoch": 0.4967355878561772, "grad_norm": 2.0571627616882324, "learning_rate": 5.295467939035069e-06, "loss": 0.4894, "step": 11850 }, { "epoch": 0.4967775064711862, "grad_norm": 2.0542523860931396, "learning_rate": 5.294790270348662e-06, "loss": 0.5332, "step": 11851 }, { "epoch": 0.49681942508619514, "grad_norm": 2.2758641242980957, "learning_rate": 5.294112596228206e-06, "loss": 0.4909, "step": 11852 }, { "epoch": 0.4968613437012041, "grad_norm": 1.5622667074203491, "learning_rate": 5.293434916686191e-06, "loss": 0.4441, "step": 11853 }, { "epoch": 0.49690326231621307, "grad_norm": 1.771256685256958, "learning_rate": 5.292757231735108e-06, "loss": 0.4855, "step": 11854 }, { "epoch": 0.49694518093122203, "grad_norm": 1.8812450170516968, "learning_rate": 5.2920795413874506e-06, "loss": 0.4766, "step": 11855 }, { "epoch": 0.496987099546231, "grad_norm": 1.5467233657836914, "learning_rate": 5.291401845655711e-06, "loss": 0.4923, "step": 11856 }, { "epoch": 0.49702901816123995, "grad_norm": 2.427628755569458, "learning_rate": 5.290724144552379e-06, "loss": 0.4931, "step": 11857 }, { "epoch": 0.4970709367762489, "grad_norm": 1.8146549463272095, "learning_rate": 5.290046438089952e-06, "loss": 0.472, "step": 11858 }, { "epoch": 0.4971128553912579, "grad_norm": 1.639697551727295, "learning_rate": 5.289368726280919e-06, "loss": 0.4548, "step": 11859 }, { "epoch": 0.49715477400626684, "grad_norm": 1.8461401462554932, "learning_rate": 5.288691009137774e-06, "loss": 0.5501, "step": 11860 }, { "epoch": 0.4971966926212758, "grad_norm": 1.72470223903656, "learning_rate": 5.288013286673007e-06, "loss": 0.5263, "step": 11861 }, { "epoch": 0.49723861123628477, "grad_norm": 1.7275439500808716, "learning_rate": 5.287335558899116e-06, "loss": 0.4963, "step": 11862 }, { "epoch": 0.49728052985129373, "grad_norm": 2.080357551574707, "learning_rate": 5.286657825828591e-06, "loss": 0.5428, "step": 11863 }, { "epoch": 0.4973224484663027, "grad_norm": 1.5874114036560059, "learning_rate": 5.285980087473923e-06, "loss": 0.5184, "step": 11864 }, { "epoch": 0.49736436708131165, "grad_norm": 1.9062716960906982, "learning_rate": 5.2853023438476095e-06, "loss": 0.4854, "step": 11865 }, { "epoch": 0.4974062856963206, "grad_norm": 1.5449676513671875, "learning_rate": 5.284624594962141e-06, "loss": 0.4885, "step": 11866 }, { "epoch": 0.4974482043113296, "grad_norm": 1.5853550434112549, "learning_rate": 5.283946840830012e-06, "loss": 0.5198, "step": 11867 }, { "epoch": 0.49749012292633854, "grad_norm": 2.3991074562072754, "learning_rate": 5.283269081463716e-06, "loss": 0.5296, "step": 11868 }, { "epoch": 0.49753204154134745, "grad_norm": 1.776947259902954, "learning_rate": 5.282591316875745e-06, "loss": 0.5119, "step": 11869 }, { "epoch": 0.4975739601563564, "grad_norm": 1.779350757598877, "learning_rate": 5.281913547078593e-06, "loss": 0.4865, "step": 11870 }, { "epoch": 0.49761587877136537, "grad_norm": 1.825608730316162, "learning_rate": 5.281235772084756e-06, "loss": 0.5775, "step": 11871 }, { "epoch": 0.49765779738637433, "grad_norm": 1.5396275520324707, "learning_rate": 5.280557991906727e-06, "loss": 0.4931, "step": 11872 }, { "epoch": 0.4976997160013833, "grad_norm": 1.7660101652145386, "learning_rate": 5.279880206556997e-06, "loss": 0.514, "step": 11873 }, { "epoch": 0.49774163461639226, "grad_norm": 1.5841518640518188, "learning_rate": 5.279202416048064e-06, "loss": 0.4771, "step": 11874 }, { "epoch": 0.4977835532314012, "grad_norm": 2.049530029296875, "learning_rate": 5.27852462039242e-06, "loss": 0.4706, "step": 11875 }, { "epoch": 0.4978254718464102, "grad_norm": 1.7148518562316895, "learning_rate": 5.277846819602559e-06, "loss": 0.46, "step": 11876 }, { "epoch": 0.49786739046141915, "grad_norm": 1.6788358688354492, "learning_rate": 5.277169013690977e-06, "loss": 0.461, "step": 11877 }, { "epoch": 0.4979093090764281, "grad_norm": 1.8779735565185547, "learning_rate": 5.276491202670168e-06, "loss": 0.4437, "step": 11878 }, { "epoch": 0.49795122769143707, "grad_norm": 1.7611322402954102, "learning_rate": 5.2758133865526225e-06, "loss": 0.5065, "step": 11879 }, { "epoch": 0.49799314630644603, "grad_norm": 2.967869281768799, "learning_rate": 5.275135565350842e-06, "loss": 0.5352, "step": 11880 }, { "epoch": 0.498035064921455, "grad_norm": 1.7666484117507935, "learning_rate": 5.274457739077317e-06, "loss": 0.4524, "step": 11881 }, { "epoch": 0.49807698353646396, "grad_norm": 1.7247365713119507, "learning_rate": 5.273779907744543e-06, "loss": 0.4784, "step": 11882 }, { "epoch": 0.4981189021514729, "grad_norm": 1.862870693206787, "learning_rate": 5.2731020713650136e-06, "loss": 0.421, "step": 11883 }, { "epoch": 0.4981608207664819, "grad_norm": 1.7049932479858398, "learning_rate": 5.272424229951226e-06, "loss": 0.5131, "step": 11884 }, { "epoch": 0.49820273938149084, "grad_norm": 1.8765370845794678, "learning_rate": 5.2717463835156754e-06, "loss": 0.438, "step": 11885 }, { "epoch": 0.4982446579964998, "grad_norm": 1.848286509513855, "learning_rate": 5.271068532070854e-06, "loss": 0.4847, "step": 11886 }, { "epoch": 0.49828657661150877, "grad_norm": 2.1029253005981445, "learning_rate": 5.270390675629261e-06, "loss": 0.5548, "step": 11887 }, { "epoch": 0.49832849522651773, "grad_norm": 1.7977464199066162, "learning_rate": 5.269712814203387e-06, "loss": 0.5678, "step": 11888 }, { "epoch": 0.4983704138415267, "grad_norm": 2.005959987640381, "learning_rate": 5.269034947805732e-06, "loss": 0.5166, "step": 11889 }, { "epoch": 0.49841233245653566, "grad_norm": 1.4575691223144531, "learning_rate": 5.268357076448789e-06, "loss": 0.4934, "step": 11890 }, { "epoch": 0.4984542510715446, "grad_norm": 1.5399757623672485, "learning_rate": 5.267679200145055e-06, "loss": 0.4633, "step": 11891 }, { "epoch": 0.4984961696865536, "grad_norm": 2.142423391342163, "learning_rate": 5.2670013189070245e-06, "loss": 0.4508, "step": 11892 }, { "epoch": 0.49853808830156254, "grad_norm": 1.8425203561782837, "learning_rate": 5.266323432747194e-06, "loss": 0.5326, "step": 11893 }, { "epoch": 0.49858000691657145, "grad_norm": 1.7258250713348389, "learning_rate": 5.265645541678059e-06, "loss": 0.4681, "step": 11894 }, { "epoch": 0.4986219255315804, "grad_norm": 2.1109445095062256, "learning_rate": 5.2649676457121165e-06, "loss": 0.511, "step": 11895 }, { "epoch": 0.4986638441465894, "grad_norm": 2.170419216156006, "learning_rate": 5.264289744861861e-06, "loss": 0.5184, "step": 11896 }, { "epoch": 0.49870576276159834, "grad_norm": 1.6346238851547241, "learning_rate": 5.26361183913979e-06, "loss": 0.4912, "step": 11897 }, { "epoch": 0.4987476813766073, "grad_norm": 1.8268725872039795, "learning_rate": 5.262933928558398e-06, "loss": 0.4823, "step": 11898 }, { "epoch": 0.49878959999161626, "grad_norm": 1.8791335821151733, "learning_rate": 5.262256013130182e-06, "loss": 0.5285, "step": 11899 }, { "epoch": 0.4988315186066252, "grad_norm": 1.6728332042694092, "learning_rate": 5.261578092867643e-06, "loss": 0.5035, "step": 11900 }, { "epoch": 0.4988734372216342, "grad_norm": 1.9902185201644897, "learning_rate": 5.260900167783269e-06, "loss": 0.5066, "step": 11901 }, { "epoch": 0.49891535583664315, "grad_norm": 2.0131983757019043, "learning_rate": 5.260222237889563e-06, "loss": 0.5657, "step": 11902 }, { "epoch": 0.4989572744516521, "grad_norm": 2.2785515785217285, "learning_rate": 5.259544303199019e-06, "loss": 0.4717, "step": 11903 }, { "epoch": 0.4989991930666611, "grad_norm": 1.674078345298767, "learning_rate": 5.258866363724136e-06, "loss": 0.5108, "step": 11904 }, { "epoch": 0.49904111168167004, "grad_norm": 1.9031445980072021, "learning_rate": 5.258188419477408e-06, "loss": 0.4773, "step": 11905 }, { "epoch": 0.499083030296679, "grad_norm": 1.7473191022872925, "learning_rate": 5.257510470471334e-06, "loss": 0.4829, "step": 11906 }, { "epoch": 0.49912494891168796, "grad_norm": 1.654068946838379, "learning_rate": 5.25683251671841e-06, "loss": 0.4679, "step": 11907 }, { "epoch": 0.4991668675266969, "grad_norm": 1.9282338619232178, "learning_rate": 5.256154558231134e-06, "loss": 0.5331, "step": 11908 }, { "epoch": 0.4992087861417059, "grad_norm": 2.41667103767395, "learning_rate": 5.255476595022003e-06, "loss": 0.4937, "step": 11909 }, { "epoch": 0.49925070475671485, "grad_norm": 1.8078101873397827, "learning_rate": 5.2547986271035145e-06, "loss": 0.5616, "step": 11910 }, { "epoch": 0.4992926233717238, "grad_norm": 1.7574070692062378, "learning_rate": 5.254120654488165e-06, "loss": 0.5292, "step": 11911 }, { "epoch": 0.49933454198673277, "grad_norm": 1.7499516010284424, "learning_rate": 5.253442677188453e-06, "loss": 0.4946, "step": 11912 }, { "epoch": 0.49937646060174173, "grad_norm": 1.8400015830993652, "learning_rate": 5.252764695216875e-06, "loss": 0.5325, "step": 11913 }, { "epoch": 0.4994183792167507, "grad_norm": 1.6963025331497192, "learning_rate": 5.25208670858593e-06, "loss": 0.5588, "step": 11914 }, { "epoch": 0.49946029783175966, "grad_norm": 1.7693421840667725, "learning_rate": 5.251408717308114e-06, "loss": 0.5093, "step": 11915 }, { "epoch": 0.4995022164467686, "grad_norm": 1.6914094686508179, "learning_rate": 5.2507307213959256e-06, "loss": 0.5945, "step": 11916 }, { "epoch": 0.4995441350617776, "grad_norm": 1.6655621528625488, "learning_rate": 5.250052720861865e-06, "loss": 0.498, "step": 11917 }, { "epoch": 0.49958605367678655, "grad_norm": 3.290961742401123, "learning_rate": 5.249374715718426e-06, "loss": 0.5082, "step": 11918 }, { "epoch": 0.49962797229179545, "grad_norm": 1.6523596048355103, "learning_rate": 5.24869670597811e-06, "loss": 0.5354, "step": 11919 }, { "epoch": 0.4996698909068044, "grad_norm": 1.9276143312454224, "learning_rate": 5.2480186916534145e-06, "loss": 0.4925, "step": 11920 }, { "epoch": 0.4997118095218134, "grad_norm": 1.5930757522583008, "learning_rate": 5.2473406727568365e-06, "loss": 0.5043, "step": 11921 }, { "epoch": 0.49975372813682234, "grad_norm": 2.0487051010131836, "learning_rate": 5.246662649300874e-06, "loss": 0.5345, "step": 11922 }, { "epoch": 0.4997956467518313, "grad_norm": 1.9557191133499146, "learning_rate": 5.245984621298029e-06, "loss": 0.5205, "step": 11923 }, { "epoch": 0.49983756536684026, "grad_norm": 1.6437627077102661, "learning_rate": 5.245306588760797e-06, "loss": 0.4526, "step": 11924 }, { "epoch": 0.4998794839818492, "grad_norm": 2.779473304748535, "learning_rate": 5.244628551701676e-06, "loss": 0.5022, "step": 11925 }, { "epoch": 0.4999214025968582, "grad_norm": 1.5828973054885864, "learning_rate": 5.243950510133167e-06, "loss": 0.4842, "step": 11926 }, { "epoch": 0.49996332121186715, "grad_norm": 1.9283214807510376, "learning_rate": 5.243272464067767e-06, "loss": 0.4966, "step": 11927 }, { "epoch": 0.5000052398268762, "grad_norm": 1.9998466968536377, "learning_rate": 5.2425944135179754e-06, "loss": 0.4952, "step": 11928 }, { "epoch": 0.5000471584418851, "grad_norm": 5.250613212585449, "learning_rate": 5.241916358496291e-06, "loss": 0.4706, "step": 11929 }, { "epoch": 0.5000890770568941, "grad_norm": 1.78824782371521, "learning_rate": 5.2412382990152136e-06, "loss": 0.4972, "step": 11930 }, { "epoch": 0.500130995671903, "grad_norm": 1.6970125436782837, "learning_rate": 5.240560235087241e-06, "loss": 0.5354, "step": 11931 }, { "epoch": 0.5001729142869119, "grad_norm": 1.6552129983901978, "learning_rate": 5.239882166724873e-06, "loss": 0.4498, "step": 11932 }, { "epoch": 0.5002148329019209, "grad_norm": 1.5652347803115845, "learning_rate": 5.23920409394061e-06, "loss": 0.5107, "step": 11933 }, { "epoch": 0.5002567515169298, "grad_norm": 1.819624900817871, "learning_rate": 5.2385260167469495e-06, "loss": 0.521, "step": 11934 }, { "epoch": 0.5002986701319388, "grad_norm": 1.9511616230010986, "learning_rate": 5.23784793515639e-06, "loss": 0.5234, "step": 11935 }, { "epoch": 0.5003405887469478, "grad_norm": 1.68449866771698, "learning_rate": 5.237169849181435e-06, "loss": 0.5084, "step": 11936 }, { "epoch": 0.5003825073619568, "grad_norm": 1.8076881170272827, "learning_rate": 5.236491758834581e-06, "loss": 0.5321, "step": 11937 }, { "epoch": 0.5004244259769657, "grad_norm": 1.7915680408477783, "learning_rate": 5.235813664128328e-06, "loss": 0.4865, "step": 11938 }, { "epoch": 0.5004663445919747, "grad_norm": 1.665475845336914, "learning_rate": 5.235135565075177e-06, "loss": 0.4734, "step": 11939 }, { "epoch": 0.5005082632069836, "grad_norm": 2.660050868988037, "learning_rate": 5.234457461687627e-06, "loss": 0.482, "step": 11940 }, { "epoch": 0.5005501818219926, "grad_norm": 1.6394627094268799, "learning_rate": 5.233779353978175e-06, "loss": 0.5014, "step": 11941 }, { "epoch": 0.5005921004370015, "grad_norm": 1.8645694255828857, "learning_rate": 5.233101241959327e-06, "loss": 0.5152, "step": 11942 }, { "epoch": 0.5006340190520105, "grad_norm": 3.658003807067871, "learning_rate": 5.23242312564358e-06, "loss": 0.5514, "step": 11943 }, { "epoch": 0.5006759376670195, "grad_norm": 1.6882264614105225, "learning_rate": 5.2317450050434315e-06, "loss": 0.4926, "step": 11944 }, { "epoch": 0.5007178562820285, "grad_norm": 1.7572355270385742, "learning_rate": 5.231066880171385e-06, "loss": 0.5204, "step": 11945 }, { "epoch": 0.5007597748970374, "grad_norm": 1.7286804914474487, "learning_rate": 5.230388751039941e-06, "loss": 0.4816, "step": 11946 }, { "epoch": 0.5008016935120464, "grad_norm": 1.7566561698913574, "learning_rate": 5.229710617661599e-06, "loss": 0.4843, "step": 11947 }, { "epoch": 0.5008436121270553, "grad_norm": 1.7904928922653198, "learning_rate": 5.229032480048858e-06, "loss": 0.4931, "step": 11948 }, { "epoch": 0.5008855307420643, "grad_norm": 1.7826162576675415, "learning_rate": 5.228354338214221e-06, "loss": 0.5294, "step": 11949 }, { "epoch": 0.5009274493570732, "grad_norm": 1.841388463973999, "learning_rate": 5.227676192170186e-06, "loss": 0.5272, "step": 11950 }, { "epoch": 0.5009693679720822, "grad_norm": 1.7870222330093384, "learning_rate": 5.226998041929258e-06, "loss": 0.4962, "step": 11951 }, { "epoch": 0.5010112865870912, "grad_norm": 1.8048286437988281, "learning_rate": 5.226319887503932e-06, "loss": 0.5155, "step": 11952 }, { "epoch": 0.5010532052021002, "grad_norm": 2.010921001434326, "learning_rate": 5.225641728906712e-06, "loss": 0.5054, "step": 11953 }, { "epoch": 0.5010951238171091, "grad_norm": 1.5543792247772217, "learning_rate": 5.2249635661500995e-06, "loss": 0.506, "step": 11954 }, { "epoch": 0.5011370424321181, "grad_norm": 1.7021470069885254, "learning_rate": 5.224285399246594e-06, "loss": 0.4949, "step": 11955 }, { "epoch": 0.501178961047127, "grad_norm": 1.7689133882522583, "learning_rate": 5.223607228208699e-06, "loss": 0.4916, "step": 11956 }, { "epoch": 0.5012208796621359, "grad_norm": 1.556581974029541, "learning_rate": 5.222929053048913e-06, "loss": 0.4745, "step": 11957 }, { "epoch": 0.5012627982771449, "grad_norm": 1.733903408050537, "learning_rate": 5.222250873779739e-06, "loss": 0.5233, "step": 11958 }, { "epoch": 0.5013047168921538, "grad_norm": 1.9342824220657349, "learning_rate": 5.221572690413675e-06, "loss": 0.4955, "step": 11959 }, { "epoch": 0.5013466355071629, "grad_norm": 1.7805087566375732, "learning_rate": 5.2208945029632256e-06, "loss": 0.5017, "step": 11960 }, { "epoch": 0.5013885541221718, "grad_norm": 1.8287171125411987, "learning_rate": 5.220216311440893e-06, "loss": 0.5226, "step": 11961 }, { "epoch": 0.5014304727371808, "grad_norm": 2.352678060531616, "learning_rate": 5.219538115859175e-06, "loss": 0.5523, "step": 11962 }, { "epoch": 0.5014723913521897, "grad_norm": 2.428537130355835, "learning_rate": 5.218859916230577e-06, "loss": 0.488, "step": 11963 }, { "epoch": 0.5015143099671987, "grad_norm": 1.754288911819458, "learning_rate": 5.2181817125675985e-06, "loss": 0.5456, "step": 11964 }, { "epoch": 0.5015562285822076, "grad_norm": 1.7232451438903809, "learning_rate": 5.21750350488274e-06, "loss": 0.5128, "step": 11965 }, { "epoch": 0.5015981471972166, "grad_norm": 1.5443381071090698, "learning_rate": 5.216825293188508e-06, "loss": 0.4796, "step": 11966 }, { "epoch": 0.5016400658122255, "grad_norm": 1.8648937940597534, "learning_rate": 5.2161470774974e-06, "loss": 0.5271, "step": 11967 }, { "epoch": 0.5016819844272346, "grad_norm": 1.6305360794067383, "learning_rate": 5.2154688578219196e-06, "loss": 0.4643, "step": 11968 }, { "epoch": 0.5017239030422435, "grad_norm": 1.6020625829696655, "learning_rate": 5.214790634174568e-06, "loss": 0.5097, "step": 11969 }, { "epoch": 0.5017658216572525, "grad_norm": 1.795945167541504, "learning_rate": 5.214112406567849e-06, "loss": 0.5022, "step": 11970 }, { "epoch": 0.5018077402722614, "grad_norm": 2.0079007148742676, "learning_rate": 5.213434175014264e-06, "loss": 0.5647, "step": 11971 }, { "epoch": 0.5018496588872704, "grad_norm": 3.5963165760040283, "learning_rate": 5.212755939526314e-06, "loss": 0.5347, "step": 11972 }, { "epoch": 0.5018915775022793, "grad_norm": 1.6976500749588013, "learning_rate": 5.212077700116502e-06, "loss": 0.4962, "step": 11973 }, { "epoch": 0.5019334961172883, "grad_norm": 1.6737133264541626, "learning_rate": 5.211399456797333e-06, "loss": 0.4918, "step": 11974 }, { "epoch": 0.5019754147322972, "grad_norm": 1.6332621574401855, "learning_rate": 5.210721209581305e-06, "loss": 0.4577, "step": 11975 }, { "epoch": 0.5020173333473062, "grad_norm": 1.7096433639526367, "learning_rate": 5.210042958480924e-06, "loss": 0.4643, "step": 11976 }, { "epoch": 0.5020592519623152, "grad_norm": 1.699782133102417, "learning_rate": 5.209364703508692e-06, "loss": 0.4663, "step": 11977 }, { "epoch": 0.5021011705773242, "grad_norm": 2.025392532348633, "learning_rate": 5.208686444677109e-06, "loss": 0.5079, "step": 11978 }, { "epoch": 0.5021430891923331, "grad_norm": 2.0525028705596924, "learning_rate": 5.2080081819986805e-06, "loss": 0.5455, "step": 11979 }, { "epoch": 0.5021850078073421, "grad_norm": 1.6692506074905396, "learning_rate": 5.207329915485912e-06, "loss": 0.4859, "step": 11980 }, { "epoch": 0.502226926422351, "grad_norm": 1.6391676664352417, "learning_rate": 5.206651645151298e-06, "loss": 0.5077, "step": 11981 }, { "epoch": 0.5022688450373599, "grad_norm": 6.985421180725098, "learning_rate": 5.20597337100735e-06, "loss": 0.5034, "step": 11982 }, { "epoch": 0.5023107636523689, "grad_norm": 1.7516169548034668, "learning_rate": 5.205295093066567e-06, "loss": 0.5227, "step": 11983 }, { "epoch": 0.5023526822673778, "grad_norm": 1.5817553997039795, "learning_rate": 5.204616811341452e-06, "loss": 0.3951, "step": 11984 }, { "epoch": 0.5023946008823869, "grad_norm": 1.6121984720230103, "learning_rate": 5.20393852584451e-06, "loss": 0.4658, "step": 11985 }, { "epoch": 0.5024365194973958, "grad_norm": 1.7095493078231812, "learning_rate": 5.203260236588243e-06, "loss": 0.5873, "step": 11986 }, { "epoch": 0.5024784381124048, "grad_norm": 1.849190354347229, "learning_rate": 5.2025819435851534e-06, "loss": 0.512, "step": 11987 }, { "epoch": 0.5025203567274137, "grad_norm": 1.5381165742874146, "learning_rate": 5.201903646847747e-06, "loss": 0.5073, "step": 11988 }, { "epoch": 0.5025622753424227, "grad_norm": 1.6433041095733643, "learning_rate": 5.201225346388527e-06, "loss": 0.4856, "step": 11989 }, { "epoch": 0.5026041939574316, "grad_norm": 1.8900821208953857, "learning_rate": 5.200547042219994e-06, "loss": 0.4742, "step": 11990 }, { "epoch": 0.5026461125724406, "grad_norm": 1.7336957454681396, "learning_rate": 5.199868734354653e-06, "loss": 0.5539, "step": 11991 }, { "epoch": 0.5026880311874495, "grad_norm": 1.5939725637435913, "learning_rate": 5.1991904228050094e-06, "loss": 0.4647, "step": 11992 }, { "epoch": 0.5027299498024586, "grad_norm": 1.7893482446670532, "learning_rate": 5.198512107583565e-06, "loss": 0.4657, "step": 11993 }, { "epoch": 0.5027718684174675, "grad_norm": 1.9031237363815308, "learning_rate": 5.197833788702826e-06, "loss": 0.4858, "step": 11994 }, { "epoch": 0.5028137870324765, "grad_norm": 1.5862127542495728, "learning_rate": 5.1971554661752935e-06, "loss": 0.5069, "step": 11995 }, { "epoch": 0.5028557056474854, "grad_norm": 1.6259222030639648, "learning_rate": 5.196477140013472e-06, "loss": 0.4828, "step": 11996 }, { "epoch": 0.5028976242624944, "grad_norm": 1.4993318319320679, "learning_rate": 5.195798810229866e-06, "loss": 0.503, "step": 11997 }, { "epoch": 0.5029395428775033, "grad_norm": 1.6034421920776367, "learning_rate": 5.195120476836981e-06, "loss": 0.4799, "step": 11998 }, { "epoch": 0.5029814614925123, "grad_norm": 1.8445945978164673, "learning_rate": 5.194442139847319e-06, "loss": 0.4799, "step": 11999 }, { "epoch": 0.5030233801075212, "grad_norm": 1.851670503616333, "learning_rate": 5.193763799273385e-06, "loss": 0.5232, "step": 12000 }, { "epoch": 0.5030652987225303, "grad_norm": 1.7824255228042603, "learning_rate": 5.193085455127683e-06, "loss": 0.5339, "step": 12001 }, { "epoch": 0.5031072173375392, "grad_norm": 1.7657389640808105, "learning_rate": 5.192407107422718e-06, "loss": 0.5058, "step": 12002 }, { "epoch": 0.5031491359525482, "grad_norm": 1.6035239696502686, "learning_rate": 5.191728756170994e-06, "loss": 0.4777, "step": 12003 }, { "epoch": 0.5031910545675571, "grad_norm": 1.6638902425765991, "learning_rate": 5.191050401385015e-06, "loss": 0.4886, "step": 12004 }, { "epoch": 0.5032329731825661, "grad_norm": 2.2475574016571045, "learning_rate": 5.1903720430772865e-06, "loss": 0.4877, "step": 12005 }, { "epoch": 0.503274891797575, "grad_norm": 1.7429531812667847, "learning_rate": 5.18969368126031e-06, "loss": 0.5026, "step": 12006 }, { "epoch": 0.5033168104125839, "grad_norm": 1.7228758335113525, "learning_rate": 5.189015315946596e-06, "loss": 0.5082, "step": 12007 }, { "epoch": 0.5033587290275929, "grad_norm": 1.6083471775054932, "learning_rate": 5.188336947148645e-06, "loss": 0.46, "step": 12008 }, { "epoch": 0.5034006476426018, "grad_norm": 1.7583563327789307, "learning_rate": 5.187658574878962e-06, "loss": 0.4596, "step": 12009 }, { "epoch": 0.5034425662576109, "grad_norm": 4.81679105758667, "learning_rate": 5.186980199150053e-06, "loss": 0.4974, "step": 12010 }, { "epoch": 0.5034844848726198, "grad_norm": 1.8236943483352661, "learning_rate": 5.186301819974424e-06, "loss": 0.5145, "step": 12011 }, { "epoch": 0.5035264034876288, "grad_norm": 1.6004446744918823, "learning_rate": 5.1856234373645775e-06, "loss": 0.5247, "step": 12012 }, { "epoch": 0.5035683221026377, "grad_norm": 1.7765120267868042, "learning_rate": 5.184945051333019e-06, "loss": 0.4988, "step": 12013 }, { "epoch": 0.5036102407176467, "grad_norm": 1.9628067016601562, "learning_rate": 5.184266661892254e-06, "loss": 0.4574, "step": 12014 }, { "epoch": 0.5036521593326556, "grad_norm": 1.7869257926940918, "learning_rate": 5.183588269054788e-06, "loss": 0.4447, "step": 12015 }, { "epoch": 0.5036940779476646, "grad_norm": 1.6152864694595337, "learning_rate": 5.1829098728331266e-06, "loss": 0.5149, "step": 12016 }, { "epoch": 0.5037359965626735, "grad_norm": 1.85918390750885, "learning_rate": 5.182231473239775e-06, "loss": 0.5468, "step": 12017 }, { "epoch": 0.5037779151776826, "grad_norm": 1.9285550117492676, "learning_rate": 5.181553070287236e-06, "loss": 0.5565, "step": 12018 }, { "epoch": 0.5038198337926915, "grad_norm": 1.9654186964035034, "learning_rate": 5.180874663988019e-06, "loss": 0.5613, "step": 12019 }, { "epoch": 0.5038617524077005, "grad_norm": 1.947475552558899, "learning_rate": 5.180196254354627e-06, "loss": 0.4643, "step": 12020 }, { "epoch": 0.5039036710227094, "grad_norm": 1.7047464847564697, "learning_rate": 5.179517841399567e-06, "loss": 0.5196, "step": 12021 }, { "epoch": 0.5039455896377184, "grad_norm": 2.3856186866760254, "learning_rate": 5.178839425135344e-06, "loss": 0.5072, "step": 12022 }, { "epoch": 0.5039875082527273, "grad_norm": 3.1332645416259766, "learning_rate": 5.178161005574462e-06, "loss": 0.5348, "step": 12023 }, { "epoch": 0.5040294268677363, "grad_norm": 1.5299197435379028, "learning_rate": 5.177482582729427e-06, "loss": 0.4761, "step": 12024 }, { "epoch": 0.5040713454827452, "grad_norm": 1.6053266525268555, "learning_rate": 5.176804156612748e-06, "loss": 0.4935, "step": 12025 }, { "epoch": 0.5041132640977543, "grad_norm": 2.2541346549987793, "learning_rate": 5.176125727236929e-06, "loss": 0.4747, "step": 12026 }, { "epoch": 0.5041551827127632, "grad_norm": 1.8380324840545654, "learning_rate": 5.175447294614475e-06, "loss": 0.5298, "step": 12027 }, { "epoch": 0.5041971013277722, "grad_norm": 1.851418375968933, "learning_rate": 5.174768858757892e-06, "loss": 0.5027, "step": 12028 }, { "epoch": 0.5042390199427811, "grad_norm": 1.6045894622802734, "learning_rate": 5.174090419679688e-06, "loss": 0.4509, "step": 12029 }, { "epoch": 0.5042809385577901, "grad_norm": 1.772584319114685, "learning_rate": 5.173411977392368e-06, "loss": 0.5011, "step": 12030 }, { "epoch": 0.504322857172799, "grad_norm": 1.7008790969848633, "learning_rate": 5.172733531908437e-06, "loss": 0.5201, "step": 12031 }, { "epoch": 0.5043647757878079, "grad_norm": 1.5332516431808472, "learning_rate": 5.172055083240403e-06, "loss": 0.5194, "step": 12032 }, { "epoch": 0.5044066944028169, "grad_norm": 1.4893031120300293, "learning_rate": 5.17137663140077e-06, "loss": 0.4483, "step": 12033 }, { "epoch": 0.5044486130178258, "grad_norm": 1.852039098739624, "learning_rate": 5.1706981764020455e-06, "loss": 0.5325, "step": 12034 }, { "epoch": 0.5044905316328349, "grad_norm": 1.6570605039596558, "learning_rate": 5.170019718256738e-06, "loss": 0.4972, "step": 12035 }, { "epoch": 0.5045324502478438, "grad_norm": 1.6916351318359375, "learning_rate": 5.169341256977352e-06, "loss": 0.5174, "step": 12036 }, { "epoch": 0.5045743688628528, "grad_norm": 1.7646182775497437, "learning_rate": 5.168662792576392e-06, "loss": 0.5846, "step": 12037 }, { "epoch": 0.5046162874778617, "grad_norm": 1.4614039659500122, "learning_rate": 5.1679843250663685e-06, "loss": 0.4714, "step": 12038 }, { "epoch": 0.5046582060928707, "grad_norm": 2.0946128368377686, "learning_rate": 5.167305854459785e-06, "loss": 0.4875, "step": 12039 }, { "epoch": 0.5047001247078796, "grad_norm": 1.6899259090423584, "learning_rate": 5.16662738076915e-06, "loss": 0.4855, "step": 12040 }, { "epoch": 0.5047420433228886, "grad_norm": 1.8553365468978882, "learning_rate": 5.16594890400697e-06, "loss": 0.5018, "step": 12041 }, { "epoch": 0.5047839619378975, "grad_norm": 1.7155836820602417, "learning_rate": 5.16527042418575e-06, "loss": 0.4927, "step": 12042 }, { "epoch": 0.5048258805529066, "grad_norm": 1.673756718635559, "learning_rate": 5.164591941317999e-06, "loss": 0.4984, "step": 12043 }, { "epoch": 0.5048677991679155, "grad_norm": 1.9937175512313843, "learning_rate": 5.163913455416223e-06, "loss": 0.5577, "step": 12044 }, { "epoch": 0.5049097177829245, "grad_norm": 1.6326433420181274, "learning_rate": 5.163234966492931e-06, "loss": 0.5298, "step": 12045 }, { "epoch": 0.5049516363979334, "grad_norm": 1.650891661643982, "learning_rate": 5.162556474560625e-06, "loss": 0.503, "step": 12046 }, { "epoch": 0.5049935550129424, "grad_norm": 1.876058578491211, "learning_rate": 5.161877979631817e-06, "loss": 0.4844, "step": 12047 }, { "epoch": 0.5050354736279513, "grad_norm": 1.9639525413513184, "learning_rate": 5.1611994817190115e-06, "loss": 0.5238, "step": 12048 }, { "epoch": 0.5050773922429603, "grad_norm": 1.7761259078979492, "learning_rate": 5.160520980834716e-06, "loss": 0.4767, "step": 12049 }, { "epoch": 0.5051193108579692, "grad_norm": 1.6215518712997437, "learning_rate": 5.159842476991439e-06, "loss": 0.5328, "step": 12050 }, { "epoch": 0.5051612294729783, "grad_norm": 1.7296208143234253, "learning_rate": 5.159163970201686e-06, "loss": 0.5115, "step": 12051 }, { "epoch": 0.5052031480879872, "grad_norm": 1.8375821113586426, "learning_rate": 5.1584854604779655e-06, "loss": 0.5324, "step": 12052 }, { "epoch": 0.5052450667029962, "grad_norm": 1.7328194379806519, "learning_rate": 5.157806947832784e-06, "loss": 0.4749, "step": 12053 }, { "epoch": 0.5052869853180051, "grad_norm": 2.009242057800293, "learning_rate": 5.157128432278651e-06, "loss": 0.4863, "step": 12054 }, { "epoch": 0.5053289039330141, "grad_norm": 1.755702018737793, "learning_rate": 5.1564499138280725e-06, "loss": 0.4995, "step": 12055 }, { "epoch": 0.505370822548023, "grad_norm": 1.9399818181991577, "learning_rate": 5.155771392493556e-06, "loss": 0.4975, "step": 12056 }, { "epoch": 0.5054127411630319, "grad_norm": 1.8450367450714111, "learning_rate": 5.155092868287609e-06, "loss": 0.5538, "step": 12057 }, { "epoch": 0.5054546597780409, "grad_norm": 1.738968014717102, "learning_rate": 5.154414341222738e-06, "loss": 0.5322, "step": 12058 }, { "epoch": 0.5054965783930498, "grad_norm": 1.4734361171722412, "learning_rate": 5.153735811311455e-06, "loss": 0.4819, "step": 12059 }, { "epoch": 0.5055384970080589, "grad_norm": 1.796870231628418, "learning_rate": 5.153057278566263e-06, "loss": 0.4763, "step": 12060 }, { "epoch": 0.5055804156230678, "grad_norm": 2.0130832195281982, "learning_rate": 5.152378742999671e-06, "loss": 0.5133, "step": 12061 }, { "epoch": 0.5056223342380768, "grad_norm": 1.6142010688781738, "learning_rate": 5.151700204624188e-06, "loss": 0.5668, "step": 12062 }, { "epoch": 0.5056642528530857, "grad_norm": 1.9396623373031616, "learning_rate": 5.151021663452322e-06, "loss": 0.4594, "step": 12063 }, { "epoch": 0.5057061714680947, "grad_norm": 1.8174595832824707, "learning_rate": 5.1503431194965806e-06, "loss": 0.5417, "step": 12064 }, { "epoch": 0.5057480900831036, "grad_norm": 1.914004921913147, "learning_rate": 5.149664572769472e-06, "loss": 0.5149, "step": 12065 }, { "epoch": 0.5057900086981126, "grad_norm": 1.4921696186065674, "learning_rate": 5.148986023283503e-06, "loss": 0.4399, "step": 12066 }, { "epoch": 0.5058319273131215, "grad_norm": 1.9716806411743164, "learning_rate": 5.148307471051182e-06, "loss": 0.531, "step": 12067 }, { "epoch": 0.5058738459281306, "grad_norm": 1.5435707569122314, "learning_rate": 5.14762891608502e-06, "loss": 0.4869, "step": 12068 }, { "epoch": 0.5059157645431395, "grad_norm": 1.864661455154419, "learning_rate": 5.146950358397521e-06, "loss": 0.5095, "step": 12069 }, { "epoch": 0.5059576831581485, "grad_norm": 1.705358624458313, "learning_rate": 5.146271798001196e-06, "loss": 0.5172, "step": 12070 }, { "epoch": 0.5059996017731574, "grad_norm": 8.621943473815918, "learning_rate": 5.145593234908552e-06, "loss": 0.4486, "step": 12071 }, { "epoch": 0.5060415203881664, "grad_norm": 1.685938835144043, "learning_rate": 5.144914669132099e-06, "loss": 0.564, "step": 12072 }, { "epoch": 0.5060834390031753, "grad_norm": 1.752822995185852, "learning_rate": 5.144236100684344e-06, "loss": 0.4968, "step": 12073 }, { "epoch": 0.5061253576181843, "grad_norm": 1.7630150318145752, "learning_rate": 5.1435575295777965e-06, "loss": 0.5094, "step": 12074 }, { "epoch": 0.5061672762331932, "grad_norm": 1.6938108205795288, "learning_rate": 5.142878955824964e-06, "loss": 0.4944, "step": 12075 }, { "epoch": 0.5062091948482023, "grad_norm": 2.1013450622558594, "learning_rate": 5.142200379438353e-06, "loss": 0.4447, "step": 12076 }, { "epoch": 0.5062511134632112, "grad_norm": 1.8194546699523926, "learning_rate": 5.1415218004304775e-06, "loss": 0.4966, "step": 12077 }, { "epoch": 0.5062930320782202, "grad_norm": 1.9238121509552002, "learning_rate": 5.1408432188138435e-06, "loss": 0.5391, "step": 12078 }, { "epoch": 0.5063349506932291, "grad_norm": 1.6388765573501587, "learning_rate": 5.140164634600958e-06, "loss": 0.5518, "step": 12079 }, { "epoch": 0.5063768693082381, "grad_norm": 1.505212426185608, "learning_rate": 5.139486047804332e-06, "loss": 0.4963, "step": 12080 }, { "epoch": 0.506418787923247, "grad_norm": 1.6429152488708496, "learning_rate": 5.138807458436473e-06, "loss": 0.4999, "step": 12081 }, { "epoch": 0.5064607065382559, "grad_norm": 2.614192008972168, "learning_rate": 5.138128866509891e-06, "loss": 0.4914, "step": 12082 }, { "epoch": 0.5065026251532649, "grad_norm": 1.9387521743774414, "learning_rate": 5.1374502720370945e-06, "loss": 0.5588, "step": 12083 }, { "epoch": 0.5065445437682738, "grad_norm": 1.7698746919631958, "learning_rate": 5.136771675030592e-06, "loss": 0.5567, "step": 12084 }, { "epoch": 0.5065864623832829, "grad_norm": 3.0429797172546387, "learning_rate": 5.136093075502893e-06, "loss": 0.4789, "step": 12085 }, { "epoch": 0.5066283809982918, "grad_norm": 1.6729469299316406, "learning_rate": 5.1354144734665045e-06, "loss": 0.4793, "step": 12086 }, { "epoch": 0.5066702996133008, "grad_norm": 1.5349828004837036, "learning_rate": 5.13473586893394e-06, "loss": 0.4563, "step": 12087 }, { "epoch": 0.5067122182283097, "grad_norm": 2.545682668685913, "learning_rate": 5.1340572619177055e-06, "loss": 0.4974, "step": 12088 }, { "epoch": 0.5067541368433187, "grad_norm": 2.11393141746521, "learning_rate": 5.133378652430309e-06, "loss": 0.5176, "step": 12089 }, { "epoch": 0.5067960554583276, "grad_norm": 1.712823510169983, "learning_rate": 5.132700040484262e-06, "loss": 0.4977, "step": 12090 }, { "epoch": 0.5068379740733366, "grad_norm": 2.035980463027954, "learning_rate": 5.132021426092073e-06, "loss": 0.523, "step": 12091 }, { "epoch": 0.5068798926883455, "grad_norm": 2.210231065750122, "learning_rate": 5.131342809266253e-06, "loss": 0.4935, "step": 12092 }, { "epoch": 0.5069218113033546, "grad_norm": 2.3238017559051514, "learning_rate": 5.130664190019309e-06, "loss": 0.5149, "step": 12093 }, { "epoch": 0.5069637299183635, "grad_norm": 1.7364861965179443, "learning_rate": 5.129985568363751e-06, "loss": 0.5367, "step": 12094 }, { "epoch": 0.5070056485333725, "grad_norm": 1.6323670148849487, "learning_rate": 5.129306944312089e-06, "loss": 0.4725, "step": 12095 }, { "epoch": 0.5070475671483814, "grad_norm": 2.0068724155426025, "learning_rate": 5.128628317876833e-06, "loss": 0.5298, "step": 12096 }, { "epoch": 0.5070894857633904, "grad_norm": 1.672166347503662, "learning_rate": 5.1279496890704916e-06, "loss": 0.5021, "step": 12097 }, { "epoch": 0.5071314043783993, "grad_norm": 1.8256272077560425, "learning_rate": 5.127271057905573e-06, "loss": 0.4904, "step": 12098 }, { "epoch": 0.5071733229934083, "grad_norm": 1.9095234870910645, "learning_rate": 5.12659242439459e-06, "loss": 0.4833, "step": 12099 }, { "epoch": 0.5072152416084172, "grad_norm": 2.1441006660461426, "learning_rate": 5.12591378855005e-06, "loss": 0.5212, "step": 12100 }, { "epoch": 0.5072571602234263, "grad_norm": 1.8411669731140137, "learning_rate": 5.1252351503844635e-06, "loss": 0.517, "step": 12101 }, { "epoch": 0.5072990788384352, "grad_norm": 1.9520471096038818, "learning_rate": 5.12455650991034e-06, "loss": 0.5492, "step": 12102 }, { "epoch": 0.5073409974534442, "grad_norm": 2.5115301609039307, "learning_rate": 5.12387786714019e-06, "loss": 0.4768, "step": 12103 }, { "epoch": 0.5073829160684531, "grad_norm": 2.025862216949463, "learning_rate": 5.123199222086521e-06, "loss": 0.4777, "step": 12104 }, { "epoch": 0.5074248346834621, "grad_norm": 4.1878252029418945, "learning_rate": 5.122520574761846e-06, "loss": 0.5628, "step": 12105 }, { "epoch": 0.507466753298471, "grad_norm": 2.348184585571289, "learning_rate": 5.121841925178674e-06, "loss": 0.4691, "step": 12106 }, { "epoch": 0.5075086719134799, "grad_norm": 1.8073351383209229, "learning_rate": 5.121163273349514e-06, "loss": 0.4537, "step": 12107 }, { "epoch": 0.5075505905284889, "grad_norm": 1.9541990756988525, "learning_rate": 5.120484619286877e-06, "loss": 0.4923, "step": 12108 }, { "epoch": 0.5075925091434978, "grad_norm": 1.7596323490142822, "learning_rate": 5.119805963003272e-06, "loss": 0.4974, "step": 12109 }, { "epoch": 0.5076344277585069, "grad_norm": 1.9455734491348267, "learning_rate": 5.1191273045112096e-06, "loss": 0.5123, "step": 12110 }, { "epoch": 0.5076763463735158, "grad_norm": 2.301745653152466, "learning_rate": 5.118448643823202e-06, "loss": 0.5174, "step": 12111 }, { "epoch": 0.5077182649885248, "grad_norm": 1.5642086267471313, "learning_rate": 5.117769980951755e-06, "loss": 0.5078, "step": 12112 }, { "epoch": 0.5077601836035337, "grad_norm": 1.8080521821975708, "learning_rate": 5.117091315909382e-06, "loss": 0.5356, "step": 12113 }, { "epoch": 0.5078021022185427, "grad_norm": 1.8976551294326782, "learning_rate": 5.116412648708592e-06, "loss": 0.4734, "step": 12114 }, { "epoch": 0.5078440208335516, "grad_norm": 2.0247957706451416, "learning_rate": 5.115733979361898e-06, "loss": 0.5556, "step": 12115 }, { "epoch": 0.5078859394485606, "grad_norm": 1.5876290798187256, "learning_rate": 5.115055307881805e-06, "loss": 0.52, "step": 12116 }, { "epoch": 0.5079278580635695, "grad_norm": 2.1486682891845703, "learning_rate": 5.114376634280827e-06, "loss": 0.468, "step": 12117 }, { "epoch": 0.5079697766785786, "grad_norm": 3.405596971511841, "learning_rate": 5.113697958571476e-06, "loss": 0.5528, "step": 12118 }, { "epoch": 0.5080116952935875, "grad_norm": 2.0711848735809326, "learning_rate": 5.113019280766258e-06, "loss": 0.5103, "step": 12119 }, { "epoch": 0.5080536139085965, "grad_norm": 1.8365991115570068, "learning_rate": 5.112340600877688e-06, "loss": 0.5544, "step": 12120 }, { "epoch": 0.5080955325236054, "grad_norm": 1.7712410688400269, "learning_rate": 5.1116619189182725e-06, "loss": 0.504, "step": 12121 }, { "epoch": 0.5081374511386144, "grad_norm": 1.6591970920562744, "learning_rate": 5.1109832349005254e-06, "loss": 0.4585, "step": 12122 }, { "epoch": 0.5081793697536233, "grad_norm": 2.1576030254364014, "learning_rate": 5.110304548836954e-06, "loss": 0.4733, "step": 12123 }, { "epoch": 0.5082212883686323, "grad_norm": 1.9561353921890259, "learning_rate": 5.109625860740072e-06, "loss": 0.5206, "step": 12124 }, { "epoch": 0.5082632069836412, "grad_norm": 1.9117090702056885, "learning_rate": 5.108947170622389e-06, "loss": 0.468, "step": 12125 }, { "epoch": 0.5083051255986503, "grad_norm": 2.0603415966033936, "learning_rate": 5.108268478496414e-06, "loss": 0.5163, "step": 12126 }, { "epoch": 0.5083470442136592, "grad_norm": 2.2640299797058105, "learning_rate": 5.10758978437466e-06, "loss": 0.4713, "step": 12127 }, { "epoch": 0.5083889628286682, "grad_norm": 1.7702428102493286, "learning_rate": 5.106911088269638e-06, "loss": 0.4695, "step": 12128 }, { "epoch": 0.5084308814436771, "grad_norm": 2.118906259536743, "learning_rate": 5.106232390193857e-06, "loss": 0.4956, "step": 12129 }, { "epoch": 0.5084728000586861, "grad_norm": 2.077941417694092, "learning_rate": 5.105553690159829e-06, "loss": 0.5225, "step": 12130 }, { "epoch": 0.508514718673695, "grad_norm": 1.8597252368927002, "learning_rate": 5.104874988180065e-06, "loss": 0.5495, "step": 12131 }, { "epoch": 0.5085566372887039, "grad_norm": 2.0864968299865723, "learning_rate": 5.104196284267074e-06, "loss": 0.4997, "step": 12132 }, { "epoch": 0.508598555903713, "grad_norm": 1.9438542127609253, "learning_rate": 5.10351757843337e-06, "loss": 0.4614, "step": 12133 }, { "epoch": 0.5086404745187219, "grad_norm": 1.9084389209747314, "learning_rate": 5.102838870691463e-06, "loss": 0.4854, "step": 12134 }, { "epoch": 0.5086823931337309, "grad_norm": 1.8428269624710083, "learning_rate": 5.102160161053863e-06, "loss": 0.5269, "step": 12135 }, { "epoch": 0.5087243117487398, "grad_norm": 1.7431714534759521, "learning_rate": 5.101481449533081e-06, "loss": 0.4845, "step": 12136 }, { "epoch": 0.5087662303637488, "grad_norm": 1.8627301454544067, "learning_rate": 5.100802736141629e-06, "loss": 0.465, "step": 12137 }, { "epoch": 0.5088081489787577, "grad_norm": 2.4404523372650146, "learning_rate": 5.100124020892019e-06, "loss": 0.5102, "step": 12138 }, { "epoch": 0.5088500675937667, "grad_norm": 1.8151363134384155, "learning_rate": 5.09944530379676e-06, "loss": 0.5439, "step": 12139 }, { "epoch": 0.5088919862087756, "grad_norm": 1.98479163646698, "learning_rate": 5.098766584868366e-06, "loss": 0.5171, "step": 12140 }, { "epoch": 0.5089339048237846, "grad_norm": 1.6318752765655518, "learning_rate": 5.098087864119343e-06, "loss": 0.5022, "step": 12141 }, { "epoch": 0.5089758234387936, "grad_norm": 2.4528303146362305, "learning_rate": 5.0974091415622085e-06, "loss": 0.4749, "step": 12142 }, { "epoch": 0.5090177420538026, "grad_norm": 1.8425188064575195, "learning_rate": 5.09673041720947e-06, "loss": 0.5122, "step": 12143 }, { "epoch": 0.5090596606688115, "grad_norm": 1.7507455348968506, "learning_rate": 5.0960516910736415e-06, "loss": 0.4469, "step": 12144 }, { "epoch": 0.5091015792838205, "grad_norm": 1.5907007455825806, "learning_rate": 5.095372963167232e-06, "loss": 0.4464, "step": 12145 }, { "epoch": 0.5091434978988294, "grad_norm": 1.993834376335144, "learning_rate": 5.094694233502754e-06, "loss": 0.4334, "step": 12146 }, { "epoch": 0.5091854165138384, "grad_norm": 1.698928952217102, "learning_rate": 5.094015502092718e-06, "loss": 0.4902, "step": 12147 }, { "epoch": 0.5092273351288473, "grad_norm": 1.9986058473587036, "learning_rate": 5.093336768949637e-06, "loss": 0.5206, "step": 12148 }, { "epoch": 0.5092692537438563, "grad_norm": 1.5910961627960205, "learning_rate": 5.092658034086022e-06, "loss": 0.5357, "step": 12149 }, { "epoch": 0.5093111723588652, "grad_norm": 1.5834109783172607, "learning_rate": 5.091979297514382e-06, "loss": 0.5029, "step": 12150 }, { "epoch": 0.5093530909738743, "grad_norm": 1.6882528066635132, "learning_rate": 5.091300559247233e-06, "loss": 0.4916, "step": 12151 }, { "epoch": 0.5093950095888832, "grad_norm": 1.6034331321716309, "learning_rate": 5.090621819297084e-06, "loss": 0.4758, "step": 12152 }, { "epoch": 0.5094369282038922, "grad_norm": 1.5563639402389526, "learning_rate": 5.089943077676447e-06, "loss": 0.4829, "step": 12153 }, { "epoch": 0.5094788468189011, "grad_norm": 1.5836012363433838, "learning_rate": 5.089264334397833e-06, "loss": 0.4901, "step": 12154 }, { "epoch": 0.5095207654339101, "grad_norm": 1.6978296041488647, "learning_rate": 5.088585589473756e-06, "loss": 0.4731, "step": 12155 }, { "epoch": 0.509562684048919, "grad_norm": 3.1614468097686768, "learning_rate": 5.087906842916724e-06, "loss": 0.5042, "step": 12156 }, { "epoch": 0.5096046026639279, "grad_norm": 1.5709139108657837, "learning_rate": 5.087228094739253e-06, "loss": 0.4614, "step": 12157 }, { "epoch": 0.509646521278937, "grad_norm": 2.347775459289551, "learning_rate": 5.086549344953852e-06, "loss": 0.4904, "step": 12158 }, { "epoch": 0.5096884398939459, "grad_norm": 1.9762142896652222, "learning_rate": 5.0858705935730324e-06, "loss": 0.4891, "step": 12159 }, { "epoch": 0.5097303585089549, "grad_norm": 1.826913595199585, "learning_rate": 5.085191840609307e-06, "loss": 0.4951, "step": 12160 }, { "epoch": 0.5097722771239638, "grad_norm": 1.7389167547225952, "learning_rate": 5.084513086075189e-06, "loss": 0.505, "step": 12161 }, { "epoch": 0.5098141957389728, "grad_norm": 1.4849265813827515, "learning_rate": 5.083834329983189e-06, "loss": 0.4843, "step": 12162 }, { "epoch": 0.5098561143539817, "grad_norm": 1.6381185054779053, "learning_rate": 5.083155572345818e-06, "loss": 0.4799, "step": 12163 }, { "epoch": 0.5098980329689907, "grad_norm": 1.5858137607574463, "learning_rate": 5.0824768131755905e-06, "loss": 0.4927, "step": 12164 }, { "epoch": 0.5099399515839996, "grad_norm": 1.6148672103881836, "learning_rate": 5.0817980524850165e-06, "loss": 0.534, "step": 12165 }, { "epoch": 0.5099818701990086, "grad_norm": 1.461706519126892, "learning_rate": 5.081119290286608e-06, "loss": 0.5352, "step": 12166 }, { "epoch": 0.5100237888140176, "grad_norm": 4.241180419921875, "learning_rate": 5.080440526592878e-06, "loss": 0.5003, "step": 12167 }, { "epoch": 0.5100657074290266, "grad_norm": 2.065126895904541, "learning_rate": 5.079761761416338e-06, "loss": 0.5064, "step": 12168 }, { "epoch": 0.5101076260440355, "grad_norm": 1.8690178394317627, "learning_rate": 5.079082994769498e-06, "loss": 0.4492, "step": 12169 }, { "epoch": 0.5101495446590445, "grad_norm": 1.6600157022476196, "learning_rate": 5.0784042266648745e-06, "loss": 0.5299, "step": 12170 }, { "epoch": 0.5101914632740534, "grad_norm": 1.892921805381775, "learning_rate": 5.077725457114978e-06, "loss": 0.5035, "step": 12171 }, { "epoch": 0.5102333818890624, "grad_norm": 1.8202874660491943, "learning_rate": 5.07704668613232e-06, "loss": 0.5233, "step": 12172 }, { "epoch": 0.5102753005040713, "grad_norm": 1.8614273071289062, "learning_rate": 5.076367913729412e-06, "loss": 0.5183, "step": 12173 }, { "epoch": 0.5103172191190803, "grad_norm": 3.305490016937256, "learning_rate": 5.0756891399187665e-06, "loss": 0.4369, "step": 12174 }, { "epoch": 0.5103591377340893, "grad_norm": 1.732622742652893, "learning_rate": 5.075010364712896e-06, "loss": 0.5402, "step": 12175 }, { "epoch": 0.5104010563490983, "grad_norm": 1.692163348197937, "learning_rate": 5.074331588124317e-06, "loss": 0.4603, "step": 12176 }, { "epoch": 0.5104429749641072, "grad_norm": 1.6049190759658813, "learning_rate": 5.073652810165535e-06, "loss": 0.517, "step": 12177 }, { "epoch": 0.5104848935791162, "grad_norm": 1.7362339496612549, "learning_rate": 5.0729740308490645e-06, "loss": 0.4976, "step": 12178 }, { "epoch": 0.5105268121941251, "grad_norm": 1.6950874328613281, "learning_rate": 5.0722952501874205e-06, "loss": 0.5151, "step": 12179 }, { "epoch": 0.5105687308091341, "grad_norm": 2.0353012084960938, "learning_rate": 5.071616468193112e-06, "loss": 0.5343, "step": 12180 }, { "epoch": 0.510610649424143, "grad_norm": 1.6138302087783813, "learning_rate": 5.070937684878654e-06, "loss": 0.4851, "step": 12181 }, { "epoch": 0.5106525680391519, "grad_norm": 1.4709891080856323, "learning_rate": 5.070258900256557e-06, "loss": 0.4203, "step": 12182 }, { "epoch": 0.510694486654161, "grad_norm": 1.6187199354171753, "learning_rate": 5.069580114339335e-06, "loss": 0.4865, "step": 12183 }, { "epoch": 0.5107364052691699, "grad_norm": 1.8198719024658203, "learning_rate": 5.0689013271395e-06, "loss": 0.5187, "step": 12184 }, { "epoch": 0.5107783238841789, "grad_norm": 1.5222266912460327, "learning_rate": 5.068222538669564e-06, "loss": 0.4747, "step": 12185 }, { "epoch": 0.5108202424991878, "grad_norm": 1.6654356718063354, "learning_rate": 5.06754374894204e-06, "loss": 0.4843, "step": 12186 }, { "epoch": 0.5108621611141968, "grad_norm": 1.6899553537368774, "learning_rate": 5.066864957969438e-06, "loss": 0.5283, "step": 12187 }, { "epoch": 0.5109040797292057, "grad_norm": 2.0589122772216797, "learning_rate": 5.066186165764275e-06, "loss": 0.5047, "step": 12188 }, { "epoch": 0.5109459983442147, "grad_norm": 1.9834136962890625, "learning_rate": 5.065507372339062e-06, "loss": 0.4589, "step": 12189 }, { "epoch": 0.5109879169592236, "grad_norm": 2.243290662765503, "learning_rate": 5.064828577706311e-06, "loss": 0.4464, "step": 12190 }, { "epoch": 0.5110298355742326, "grad_norm": 1.7880079746246338, "learning_rate": 5.0641497818785345e-06, "loss": 0.4682, "step": 12191 }, { "epoch": 0.5110717541892416, "grad_norm": 2.021878957748413, "learning_rate": 5.063470984868245e-06, "loss": 0.5238, "step": 12192 }, { "epoch": 0.5111136728042506, "grad_norm": 1.8249744176864624, "learning_rate": 5.062792186687957e-06, "loss": 0.5022, "step": 12193 }, { "epoch": 0.5111555914192595, "grad_norm": 1.8300021886825562, "learning_rate": 5.062113387350179e-06, "loss": 0.508, "step": 12194 }, { "epoch": 0.5111975100342685, "grad_norm": 1.7391186952590942, "learning_rate": 5.06143458686743e-06, "loss": 0.5059, "step": 12195 }, { "epoch": 0.5112394286492774, "grad_norm": 1.888684630393982, "learning_rate": 5.060755785252218e-06, "loss": 0.5178, "step": 12196 }, { "epoch": 0.5112813472642864, "grad_norm": 1.7399544715881348, "learning_rate": 5.060076982517056e-06, "loss": 0.5087, "step": 12197 }, { "epoch": 0.5113232658792953, "grad_norm": 1.7673330307006836, "learning_rate": 5.059398178674459e-06, "loss": 0.5064, "step": 12198 }, { "epoch": 0.5113651844943043, "grad_norm": 1.8299801349639893, "learning_rate": 5.0587193737369375e-06, "loss": 0.4948, "step": 12199 }, { "epoch": 0.5114071031093133, "grad_norm": 2.0389249324798584, "learning_rate": 5.0580405677170074e-06, "loss": 0.567, "step": 12200 }, { "epoch": 0.5114490217243223, "grad_norm": 1.7117221355438232, "learning_rate": 5.057361760627178e-06, "loss": 0.4743, "step": 12201 }, { "epoch": 0.5114909403393312, "grad_norm": 1.6636180877685547, "learning_rate": 5.0566829524799655e-06, "loss": 0.4627, "step": 12202 }, { "epoch": 0.5115328589543402, "grad_norm": 1.8697751760482788, "learning_rate": 5.056004143287879e-06, "loss": 0.4988, "step": 12203 }, { "epoch": 0.5115747775693491, "grad_norm": 1.8522028923034668, "learning_rate": 5.055325333063437e-06, "loss": 0.5412, "step": 12204 }, { "epoch": 0.5116166961843581, "grad_norm": 1.6981264352798462, "learning_rate": 5.054646521819146e-06, "loss": 0.4498, "step": 12205 }, { "epoch": 0.511658614799367, "grad_norm": 2.1972365379333496, "learning_rate": 5.053967709567521e-06, "loss": 0.5447, "step": 12206 }, { "epoch": 0.5117005334143759, "grad_norm": 1.5182515382766724, "learning_rate": 5.053288896321078e-06, "loss": 0.4909, "step": 12207 }, { "epoch": 0.511742452029385, "grad_norm": 1.7172915935516357, "learning_rate": 5.052610082092327e-06, "loss": 0.4785, "step": 12208 }, { "epoch": 0.5117843706443939, "grad_norm": 1.5603512525558472, "learning_rate": 5.051931266893783e-06, "loss": 0.4307, "step": 12209 }, { "epoch": 0.5118262892594029, "grad_norm": 2.239481210708618, "learning_rate": 5.051252450737957e-06, "loss": 0.4437, "step": 12210 }, { "epoch": 0.5118682078744118, "grad_norm": 2.1827902793884277, "learning_rate": 5.0505736336373634e-06, "loss": 0.5087, "step": 12211 }, { "epoch": 0.5119101264894208, "grad_norm": 2.163808822631836, "learning_rate": 5.049894815604514e-06, "loss": 0.5385, "step": 12212 }, { "epoch": 0.5119520451044297, "grad_norm": 2.097888946533203, "learning_rate": 5.049215996651924e-06, "loss": 0.5014, "step": 12213 }, { "epoch": 0.5119939637194387, "grad_norm": 2.1738944053649902, "learning_rate": 5.048537176792104e-06, "loss": 0.5263, "step": 12214 }, { "epoch": 0.5120358823344476, "grad_norm": 1.99886953830719, "learning_rate": 5.047858356037567e-06, "loss": 0.4993, "step": 12215 }, { "epoch": 0.5120778009494567, "grad_norm": 2.0858378410339355, "learning_rate": 5.047179534400829e-06, "loss": 0.5163, "step": 12216 }, { "epoch": 0.5121197195644656, "grad_norm": 1.7571115493774414, "learning_rate": 5.046500711894402e-06, "loss": 0.5407, "step": 12217 }, { "epoch": 0.5121616381794746, "grad_norm": 1.7925817966461182, "learning_rate": 5.045821888530799e-06, "loss": 0.5076, "step": 12218 }, { "epoch": 0.5122035567944835, "grad_norm": 1.4534120559692383, "learning_rate": 5.04514306432253e-06, "loss": 0.5029, "step": 12219 }, { "epoch": 0.5122454754094925, "grad_norm": 1.708398699760437, "learning_rate": 5.044464239282114e-06, "loss": 0.5807, "step": 12220 }, { "epoch": 0.5122873940245014, "grad_norm": 1.812915325164795, "learning_rate": 5.043785413422059e-06, "loss": 0.4463, "step": 12221 }, { "epoch": 0.5123293126395104, "grad_norm": 1.6945172548294067, "learning_rate": 5.043106586754881e-06, "loss": 0.4831, "step": 12222 }, { "epoch": 0.5123712312545193, "grad_norm": 1.5830814838409424, "learning_rate": 5.042427759293095e-06, "loss": 0.498, "step": 12223 }, { "epoch": 0.5124131498695284, "grad_norm": 1.8446675539016724, "learning_rate": 5.041748931049209e-06, "loss": 0.4616, "step": 12224 }, { "epoch": 0.5124550684845373, "grad_norm": 1.7755273580551147, "learning_rate": 5.04107010203574e-06, "loss": 0.5009, "step": 12225 }, { "epoch": 0.5124969870995463, "grad_norm": 1.6441055536270142, "learning_rate": 5.0403912722652015e-06, "loss": 0.4926, "step": 12226 }, { "epoch": 0.5125389057145552, "grad_norm": 1.7182023525238037, "learning_rate": 5.039712441750105e-06, "loss": 0.5016, "step": 12227 }, { "epoch": 0.5125808243295642, "grad_norm": 2.3714170455932617, "learning_rate": 5.039033610502965e-06, "loss": 0.5036, "step": 12228 }, { "epoch": 0.5126227429445731, "grad_norm": 1.978092074394226, "learning_rate": 5.038354778536293e-06, "loss": 0.5169, "step": 12229 }, { "epoch": 0.5126646615595821, "grad_norm": 1.50960111618042, "learning_rate": 5.037675945862604e-06, "loss": 0.4609, "step": 12230 }, { "epoch": 0.512706580174591, "grad_norm": 2.138939142227173, "learning_rate": 5.036997112494411e-06, "loss": 0.5078, "step": 12231 }, { "epoch": 0.5127484987895999, "grad_norm": 1.6959404945373535, "learning_rate": 5.03631827844423e-06, "loss": 0.4831, "step": 12232 }, { "epoch": 0.512790417404609, "grad_norm": 3.380202293395996, "learning_rate": 5.0356394437245695e-06, "loss": 0.4688, "step": 12233 }, { "epoch": 0.5128323360196179, "grad_norm": 1.7175383567810059, "learning_rate": 5.034960608347944e-06, "loss": 0.521, "step": 12234 }, { "epoch": 0.5128742546346269, "grad_norm": 1.9008723497390747, "learning_rate": 5.03428177232687e-06, "loss": 0.469, "step": 12235 }, { "epoch": 0.5129161732496358, "grad_norm": 2.040695905685425, "learning_rate": 5.0336029356738585e-06, "loss": 0.5232, "step": 12236 }, { "epoch": 0.5129580918646448, "grad_norm": 1.793784260749817, "learning_rate": 5.032924098401423e-06, "loss": 0.4759, "step": 12237 }, { "epoch": 0.5130000104796537, "grad_norm": 1.8241463899612427, "learning_rate": 5.032245260522077e-06, "loss": 0.4984, "step": 12238 }, { "epoch": 0.5130419290946627, "grad_norm": 1.7653675079345703, "learning_rate": 5.031566422048335e-06, "loss": 0.5788, "step": 12239 }, { "epoch": 0.5130838477096716, "grad_norm": 1.7296106815338135, "learning_rate": 5.030887582992709e-06, "loss": 0.451, "step": 12240 }, { "epoch": 0.5131257663246807, "grad_norm": 2.122926712036133, "learning_rate": 5.030208743367713e-06, "loss": 0.4623, "step": 12241 }, { "epoch": 0.5131676849396896, "grad_norm": 2.101661205291748, "learning_rate": 5.0295299031858615e-06, "loss": 0.5149, "step": 12242 }, { "epoch": 0.5132096035546986, "grad_norm": 1.6055529117584229, "learning_rate": 5.028851062459665e-06, "loss": 0.4871, "step": 12243 }, { "epoch": 0.5132515221697075, "grad_norm": 1.8539034128189087, "learning_rate": 5.028172221201641e-06, "loss": 0.5155, "step": 12244 }, { "epoch": 0.5132934407847165, "grad_norm": 1.7199649810791016, "learning_rate": 5.027493379424301e-06, "loss": 0.4741, "step": 12245 }, { "epoch": 0.5133353593997254, "grad_norm": 1.4517394304275513, "learning_rate": 5.026814537140158e-06, "loss": 0.5256, "step": 12246 }, { "epoch": 0.5133772780147344, "grad_norm": 2.0662853717803955, "learning_rate": 5.0261356943617256e-06, "loss": 0.5195, "step": 12247 }, { "epoch": 0.5134191966297433, "grad_norm": 1.7400944232940674, "learning_rate": 5.025456851101519e-06, "loss": 0.4864, "step": 12248 }, { "epoch": 0.5134611152447524, "grad_norm": 1.9017844200134277, "learning_rate": 5.024778007372048e-06, "loss": 0.5146, "step": 12249 }, { "epoch": 0.5135030338597613, "grad_norm": 1.5876429080963135, "learning_rate": 5.024099163185832e-06, "loss": 0.5364, "step": 12250 }, { "epoch": 0.5135449524747703, "grad_norm": 1.798435926437378, "learning_rate": 5.023420318555381e-06, "loss": 0.5499, "step": 12251 }, { "epoch": 0.5135868710897792, "grad_norm": 1.6328015327453613, "learning_rate": 5.0227414734932055e-06, "loss": 0.5426, "step": 12252 }, { "epoch": 0.5136287897047882, "grad_norm": 1.901868462562561, "learning_rate": 5.0220626280118255e-06, "loss": 0.5061, "step": 12253 }, { "epoch": 0.5136707083197971, "grad_norm": 3.132024049758911, "learning_rate": 5.0213837821237495e-06, "loss": 0.497, "step": 12254 }, { "epoch": 0.5137126269348061, "grad_norm": 1.5769191980361938, "learning_rate": 5.020704935841495e-06, "loss": 0.4724, "step": 12255 }, { "epoch": 0.513754545549815, "grad_norm": 1.5160744190216064, "learning_rate": 5.020026089177572e-06, "loss": 0.4818, "step": 12256 }, { "epoch": 0.5137964641648239, "grad_norm": 1.5920768976211548, "learning_rate": 5.019347242144497e-06, "loss": 0.4985, "step": 12257 }, { "epoch": 0.513838382779833, "grad_norm": 1.696092963218689, "learning_rate": 5.018668394754781e-06, "loss": 0.4926, "step": 12258 }, { "epoch": 0.5138803013948419, "grad_norm": 1.587821364402771, "learning_rate": 5.017989547020939e-06, "loss": 0.4884, "step": 12259 }, { "epoch": 0.5139222200098509, "grad_norm": 1.5326974391937256, "learning_rate": 5.017310698955487e-06, "loss": 0.4647, "step": 12260 }, { "epoch": 0.5139641386248598, "grad_norm": 1.732748031616211, "learning_rate": 5.016631850570933e-06, "loss": 0.4928, "step": 12261 }, { "epoch": 0.5140060572398688, "grad_norm": 1.7620569467544556, "learning_rate": 5.015953001879795e-06, "loss": 0.4938, "step": 12262 }, { "epoch": 0.5140479758548777, "grad_norm": 2.256847858428955, "learning_rate": 5.0152741528945866e-06, "loss": 0.5708, "step": 12263 }, { "epoch": 0.5140898944698867, "grad_norm": 1.9005308151245117, "learning_rate": 5.01459530362782e-06, "loss": 0.473, "step": 12264 }, { "epoch": 0.5141318130848956, "grad_norm": 1.5786267518997192, "learning_rate": 5.013916454092008e-06, "loss": 0.4856, "step": 12265 }, { "epoch": 0.5141737316999047, "grad_norm": 1.7543612718582153, "learning_rate": 5.013237604299667e-06, "loss": 0.551, "step": 12266 }, { "epoch": 0.5142156503149136, "grad_norm": 1.7124996185302734, "learning_rate": 5.012558754263307e-06, "loss": 0.5548, "step": 12267 }, { "epoch": 0.5142575689299226, "grad_norm": 1.621178150177002, "learning_rate": 5.011879903995445e-06, "loss": 0.4863, "step": 12268 }, { "epoch": 0.5142994875449315, "grad_norm": 1.7250391244888306, "learning_rate": 5.011201053508592e-06, "loss": 0.5579, "step": 12269 }, { "epoch": 0.5143414061599405, "grad_norm": 2.0576231479644775, "learning_rate": 5.010522202815265e-06, "loss": 0.5023, "step": 12270 }, { "epoch": 0.5143833247749494, "grad_norm": 1.9219331741333008, "learning_rate": 5.009843351927975e-06, "loss": 0.4806, "step": 12271 }, { "epoch": 0.5144252433899584, "grad_norm": 1.53122878074646, "learning_rate": 5.009164500859236e-06, "loss": 0.4563, "step": 12272 }, { "epoch": 0.5144671620049673, "grad_norm": 1.844587802886963, "learning_rate": 5.0084856496215624e-06, "loss": 0.5572, "step": 12273 }, { "epoch": 0.5145090806199764, "grad_norm": 1.7707818746566772, "learning_rate": 5.007806798227468e-06, "loss": 0.5357, "step": 12274 }, { "epoch": 0.5145509992349853, "grad_norm": 2.1227164268493652, "learning_rate": 5.0071279466894664e-06, "loss": 0.5424, "step": 12275 }, { "epoch": 0.5145929178499943, "grad_norm": 1.5163729190826416, "learning_rate": 5.00644909502007e-06, "loss": 0.4551, "step": 12276 }, { "epoch": 0.5146348364650032, "grad_norm": 1.78486967086792, "learning_rate": 5.005770243231793e-06, "loss": 0.4989, "step": 12277 }, { "epoch": 0.5146767550800122, "grad_norm": 1.845657229423523, "learning_rate": 5.00509139133715e-06, "loss": 0.474, "step": 12278 }, { "epoch": 0.5147186736950211, "grad_norm": 1.531947374343872, "learning_rate": 5.004412539348656e-06, "loss": 0.4874, "step": 12279 }, { "epoch": 0.5147605923100301, "grad_norm": 1.631838321685791, "learning_rate": 5.00373368727882e-06, "loss": 0.5212, "step": 12280 }, { "epoch": 0.514802510925039, "grad_norm": 2.2149698734283447, "learning_rate": 5.003054835140161e-06, "loss": 0.4732, "step": 12281 }, { "epoch": 0.5148444295400479, "grad_norm": 2.249950647354126, "learning_rate": 5.002375982945189e-06, "loss": 0.5057, "step": 12282 }, { "epoch": 0.514886348155057, "grad_norm": 1.9011266231536865, "learning_rate": 5.00169713070642e-06, "loss": 0.4699, "step": 12283 }, { "epoch": 0.5149282667700659, "grad_norm": 1.599523901939392, "learning_rate": 5.001018278436366e-06, "loss": 0.5618, "step": 12284 }, { "epoch": 0.5149701853850749, "grad_norm": 1.5966461896896362, "learning_rate": 5.000339426147541e-06, "loss": 0.4536, "step": 12285 }, { "epoch": 0.5150121040000838, "grad_norm": 1.7400087118148804, "learning_rate": 4.99966057385246e-06, "loss": 0.4956, "step": 12286 }, { "epoch": 0.5150540226150928, "grad_norm": 1.5016969442367554, "learning_rate": 4.998981721563637e-06, "loss": 0.494, "step": 12287 }, { "epoch": 0.5150959412301017, "grad_norm": 1.7192268371582031, "learning_rate": 4.998302869293582e-06, "loss": 0.4846, "step": 12288 }, { "epoch": 0.5151378598451107, "grad_norm": 1.6734027862548828, "learning_rate": 4.997624017054813e-06, "loss": 0.4548, "step": 12289 }, { "epoch": 0.5151797784601196, "grad_norm": 1.63032066822052, "learning_rate": 4.996945164859842e-06, "loss": 0.4751, "step": 12290 }, { "epoch": 0.5152216970751287, "grad_norm": 1.990181565284729, "learning_rate": 4.9962663127211805e-06, "loss": 0.471, "step": 12291 }, { "epoch": 0.5152636156901376, "grad_norm": 1.8345779180526733, "learning_rate": 4.995587460651345e-06, "loss": 0.4503, "step": 12292 }, { "epoch": 0.5153055343051466, "grad_norm": 1.4906368255615234, "learning_rate": 4.994908608662851e-06, "loss": 0.4968, "step": 12293 }, { "epoch": 0.5153474529201555, "grad_norm": 1.6049211025238037, "learning_rate": 4.994229756768208e-06, "loss": 0.5192, "step": 12294 }, { "epoch": 0.5153893715351645, "grad_norm": 1.5292428731918335, "learning_rate": 4.993550904979931e-06, "loss": 0.4484, "step": 12295 }, { "epoch": 0.5154312901501734, "grad_norm": 3.2760419845581055, "learning_rate": 4.992872053310536e-06, "loss": 0.4933, "step": 12296 }, { "epoch": 0.5154732087651824, "grad_norm": 1.6445276737213135, "learning_rate": 4.9921932017725335e-06, "loss": 0.5414, "step": 12297 }, { "epoch": 0.5155151273801913, "grad_norm": 1.5708245038986206, "learning_rate": 4.991514350378438e-06, "loss": 0.4341, "step": 12298 }, { "epoch": 0.5155570459952004, "grad_norm": 1.569518804550171, "learning_rate": 4.9908354991407666e-06, "loss": 0.5087, "step": 12299 }, { "epoch": 0.5155989646102093, "grad_norm": 1.6044977903366089, "learning_rate": 4.990156648072026e-06, "loss": 0.4752, "step": 12300 }, { "epoch": 0.5156408832252183, "grad_norm": 1.7031404972076416, "learning_rate": 4.989477797184735e-06, "loss": 0.4944, "step": 12301 }, { "epoch": 0.5156828018402272, "grad_norm": 1.822169303894043, "learning_rate": 4.9887989464914085e-06, "loss": 0.4597, "step": 12302 }, { "epoch": 0.5157247204552362, "grad_norm": 1.6391385793685913, "learning_rate": 4.988120096004558e-06, "loss": 0.4545, "step": 12303 }, { "epoch": 0.5157666390702451, "grad_norm": 1.699237585067749, "learning_rate": 4.987441245736694e-06, "loss": 0.5008, "step": 12304 }, { "epoch": 0.5158085576852541, "grad_norm": 1.5098352432250977, "learning_rate": 4.9867623957003365e-06, "loss": 0.4767, "step": 12305 }, { "epoch": 0.515850476300263, "grad_norm": 2.114792823791504, "learning_rate": 4.986083545907993e-06, "loss": 0.5103, "step": 12306 }, { "epoch": 0.515892394915272, "grad_norm": 1.6608242988586426, "learning_rate": 4.985404696372182e-06, "loss": 0.5018, "step": 12307 }, { "epoch": 0.515934313530281, "grad_norm": 1.661421298980713, "learning_rate": 4.984725847105416e-06, "loss": 0.4895, "step": 12308 }, { "epoch": 0.5159762321452899, "grad_norm": 1.5245723724365234, "learning_rate": 4.9840469981202055e-06, "loss": 0.493, "step": 12309 }, { "epoch": 0.5160181507602989, "grad_norm": 1.6950329542160034, "learning_rate": 4.983368149429067e-06, "loss": 0.4613, "step": 12310 }, { "epoch": 0.5160600693753078, "grad_norm": 2.176706314086914, "learning_rate": 4.982689301044516e-06, "loss": 0.583, "step": 12311 }, { "epoch": 0.5161019879903168, "grad_norm": 1.6762826442718506, "learning_rate": 4.982010452979062e-06, "loss": 0.5321, "step": 12312 }, { "epoch": 0.5161439066053257, "grad_norm": 1.8187556266784668, "learning_rate": 4.98133160524522e-06, "loss": 0.4883, "step": 12313 }, { "epoch": 0.5161858252203347, "grad_norm": 1.7595032453536987, "learning_rate": 4.980652757855505e-06, "loss": 0.5224, "step": 12314 }, { "epoch": 0.5162277438353436, "grad_norm": 1.9831575155258179, "learning_rate": 4.9799739108224285e-06, "loss": 0.4977, "step": 12315 }, { "epoch": 0.5162696624503527, "grad_norm": 1.794463038444519, "learning_rate": 4.979295064158507e-06, "loss": 0.4438, "step": 12316 }, { "epoch": 0.5163115810653616, "grad_norm": 1.8266417980194092, "learning_rate": 4.978616217876253e-06, "loss": 0.4939, "step": 12317 }, { "epoch": 0.5163534996803706, "grad_norm": 1.7899079322814941, "learning_rate": 4.977937371988176e-06, "loss": 0.4893, "step": 12318 }, { "epoch": 0.5163954182953795, "grad_norm": 1.670578122138977, "learning_rate": 4.977258526506794e-06, "loss": 0.5318, "step": 12319 }, { "epoch": 0.5164373369103885, "grad_norm": 1.5854737758636475, "learning_rate": 4.976579681444622e-06, "loss": 0.4758, "step": 12320 }, { "epoch": 0.5164792555253974, "grad_norm": 1.525292992591858, "learning_rate": 4.975900836814171e-06, "loss": 0.4408, "step": 12321 }, { "epoch": 0.5165211741404064, "grad_norm": 2.0637049674987793, "learning_rate": 4.975221992627952e-06, "loss": 0.489, "step": 12322 }, { "epoch": 0.5165630927554153, "grad_norm": 1.5431753396987915, "learning_rate": 4.974543148898483e-06, "loss": 0.4934, "step": 12323 }, { "epoch": 0.5166050113704244, "grad_norm": 2.500732421875, "learning_rate": 4.973864305638276e-06, "loss": 0.451, "step": 12324 }, { "epoch": 0.5166469299854333, "grad_norm": 1.6072406768798828, "learning_rate": 4.9731854628598436e-06, "loss": 0.5128, "step": 12325 }, { "epoch": 0.5166888486004423, "grad_norm": 1.7290605306625366, "learning_rate": 4.972506620575702e-06, "loss": 0.4822, "step": 12326 }, { "epoch": 0.5167307672154512, "grad_norm": 1.4920589923858643, "learning_rate": 4.97182777879836e-06, "loss": 0.4482, "step": 12327 }, { "epoch": 0.5167726858304602, "grad_norm": 1.9992986917495728, "learning_rate": 4.971148937540335e-06, "loss": 0.4779, "step": 12328 }, { "epoch": 0.5168146044454691, "grad_norm": 1.4955168962478638, "learning_rate": 4.970470096814141e-06, "loss": 0.4456, "step": 12329 }, { "epoch": 0.5168565230604781, "grad_norm": 2.058178424835205, "learning_rate": 4.96979125663229e-06, "loss": 0.4862, "step": 12330 }, { "epoch": 0.516898441675487, "grad_norm": 1.6787902116775513, "learning_rate": 4.969112417007292e-06, "loss": 0.4757, "step": 12331 }, { "epoch": 0.516940360290496, "grad_norm": 1.6765248775482178, "learning_rate": 4.968433577951668e-06, "loss": 0.456, "step": 12332 }, { "epoch": 0.516982278905505, "grad_norm": 1.753046989440918, "learning_rate": 4.967754739477924e-06, "loss": 0.5635, "step": 12333 }, { "epoch": 0.5170241975205139, "grad_norm": 1.6904537677764893, "learning_rate": 4.967075901598579e-06, "loss": 0.5452, "step": 12334 }, { "epoch": 0.5170661161355229, "grad_norm": 1.7331633567810059, "learning_rate": 4.966397064326143e-06, "loss": 0.538, "step": 12335 }, { "epoch": 0.5171080347505318, "grad_norm": 1.6974009275436401, "learning_rate": 4.965718227673132e-06, "loss": 0.4759, "step": 12336 }, { "epoch": 0.5171499533655408, "grad_norm": 1.786873459815979, "learning_rate": 4.965039391652056e-06, "loss": 0.5133, "step": 12337 }, { "epoch": 0.5171918719805497, "grad_norm": 1.6477023363113403, "learning_rate": 4.964360556275433e-06, "loss": 0.4874, "step": 12338 }, { "epoch": 0.5172337905955587, "grad_norm": 1.5594953298568726, "learning_rate": 4.963681721555772e-06, "loss": 0.4451, "step": 12339 }, { "epoch": 0.5172757092105676, "grad_norm": 1.8399869203567505, "learning_rate": 4.963002887505589e-06, "loss": 0.4971, "step": 12340 }, { "epoch": 0.5173176278255767, "grad_norm": 1.713197112083435, "learning_rate": 4.962324054137397e-06, "loss": 0.4675, "step": 12341 }, { "epoch": 0.5173595464405856, "grad_norm": 2.0411267280578613, "learning_rate": 4.961645221463708e-06, "loss": 0.5391, "step": 12342 }, { "epoch": 0.5174014650555946, "grad_norm": 1.8307371139526367, "learning_rate": 4.960966389497036e-06, "loss": 0.5286, "step": 12343 }, { "epoch": 0.5174433836706035, "grad_norm": 1.6708327531814575, "learning_rate": 4.9602875582498965e-06, "loss": 0.4976, "step": 12344 }, { "epoch": 0.5174853022856125, "grad_norm": 1.9062933921813965, "learning_rate": 4.9596087277348e-06, "loss": 0.4976, "step": 12345 }, { "epoch": 0.5175272209006214, "grad_norm": 1.5611952543258667, "learning_rate": 4.958929897964259e-06, "loss": 0.5041, "step": 12346 }, { "epoch": 0.5175691395156304, "grad_norm": 1.7453840970993042, "learning_rate": 4.958251068950792e-06, "loss": 0.5051, "step": 12347 }, { "epoch": 0.5176110581306393, "grad_norm": 2.0865092277526855, "learning_rate": 4.9575722407069065e-06, "loss": 0.5474, "step": 12348 }, { "epoch": 0.5176529767456484, "grad_norm": 1.5509092807769775, "learning_rate": 4.956893413245118e-06, "loss": 0.4629, "step": 12349 }, { "epoch": 0.5176948953606573, "grad_norm": 2.1443910598754883, "learning_rate": 4.956214586577943e-06, "loss": 0.4999, "step": 12350 }, { "epoch": 0.5177368139756663, "grad_norm": 1.689455270767212, "learning_rate": 4.955535760717887e-06, "loss": 0.4678, "step": 12351 }, { "epoch": 0.5177787325906752, "grad_norm": 5.087302207946777, "learning_rate": 4.95485693567747e-06, "loss": 0.4968, "step": 12352 }, { "epoch": 0.5178206512056842, "grad_norm": 1.61978280544281, "learning_rate": 4.9541781114692036e-06, "loss": 0.4568, "step": 12353 }, { "epoch": 0.5178625698206931, "grad_norm": 1.828094244003296, "learning_rate": 4.9534992881056e-06, "loss": 0.4871, "step": 12354 }, { "epoch": 0.5179044884357021, "grad_norm": 1.5140810012817383, "learning_rate": 4.952820465599171e-06, "loss": 0.483, "step": 12355 }, { "epoch": 0.517946407050711, "grad_norm": 1.8351659774780273, "learning_rate": 4.9521416439624335e-06, "loss": 0.496, "step": 12356 }, { "epoch": 0.51798832566572, "grad_norm": 1.7276867628097534, "learning_rate": 4.951462823207898e-06, "loss": 0.509, "step": 12357 }, { "epoch": 0.518030244280729, "grad_norm": 1.7742838859558105, "learning_rate": 4.950784003348078e-06, "loss": 0.5597, "step": 12358 }, { "epoch": 0.5180721628957379, "grad_norm": 1.682252287864685, "learning_rate": 4.950105184395488e-06, "loss": 0.4844, "step": 12359 }, { "epoch": 0.5181140815107469, "grad_norm": 1.9254536628723145, "learning_rate": 4.949426366362638e-06, "loss": 0.5221, "step": 12360 }, { "epoch": 0.5181560001257558, "grad_norm": 1.7116669416427612, "learning_rate": 4.948747549262045e-06, "loss": 0.4679, "step": 12361 }, { "epoch": 0.5181979187407648, "grad_norm": 1.8554757833480835, "learning_rate": 4.948068733106218e-06, "loss": 0.4625, "step": 12362 }, { "epoch": 0.5182398373557737, "grad_norm": 1.64598548412323, "learning_rate": 4.947389917907674e-06, "loss": 0.473, "step": 12363 }, { "epoch": 0.5182817559707827, "grad_norm": 1.6959290504455566, "learning_rate": 4.946711103678925e-06, "loss": 0.4982, "step": 12364 }, { "epoch": 0.5183236745857916, "grad_norm": 1.6894954442977905, "learning_rate": 4.9460322904324795e-06, "loss": 0.4479, "step": 12365 }, { "epoch": 0.5183655932008007, "grad_norm": 1.7510862350463867, "learning_rate": 4.945353478180855e-06, "loss": 0.5006, "step": 12366 }, { "epoch": 0.5184075118158096, "grad_norm": 1.5634992122650146, "learning_rate": 4.9446746669365665e-06, "loss": 0.5205, "step": 12367 }, { "epoch": 0.5184494304308186, "grad_norm": 1.764404535293579, "learning_rate": 4.9439958567121224e-06, "loss": 0.5437, "step": 12368 }, { "epoch": 0.5184913490458275, "grad_norm": 1.7669651508331299, "learning_rate": 4.943317047520036e-06, "loss": 0.5117, "step": 12369 }, { "epoch": 0.5185332676608365, "grad_norm": 2.1320486068725586, "learning_rate": 4.942638239372824e-06, "loss": 0.4818, "step": 12370 }, { "epoch": 0.5185751862758454, "grad_norm": 1.852733850479126, "learning_rate": 4.941959432282994e-06, "loss": 0.5046, "step": 12371 }, { "epoch": 0.5186171048908544, "grad_norm": 1.4763809442520142, "learning_rate": 4.941280626263063e-06, "loss": 0.4934, "step": 12372 }, { "epoch": 0.5186590235058633, "grad_norm": 1.758466362953186, "learning_rate": 4.940601821325544e-06, "loss": 0.4962, "step": 12373 }, { "epoch": 0.5187009421208724, "grad_norm": 1.4587267637252808, "learning_rate": 4.939923017482946e-06, "loss": 0.5025, "step": 12374 }, { "epoch": 0.5187428607358813, "grad_norm": 2.2150628566741943, "learning_rate": 4.939244214747783e-06, "loss": 0.5009, "step": 12375 }, { "epoch": 0.5187847793508903, "grad_norm": 1.7276954650878906, "learning_rate": 4.938565413132572e-06, "loss": 0.5353, "step": 12376 }, { "epoch": 0.5188266979658992, "grad_norm": 1.5869008302688599, "learning_rate": 4.937886612649822e-06, "loss": 0.4745, "step": 12377 }, { "epoch": 0.5188686165809082, "grad_norm": 2.427433729171753, "learning_rate": 4.937207813312045e-06, "loss": 0.4953, "step": 12378 }, { "epoch": 0.5189105351959171, "grad_norm": 1.7866042852401733, "learning_rate": 4.936529015131757e-06, "loss": 0.5559, "step": 12379 }, { "epoch": 0.5189524538109261, "grad_norm": 1.6455591917037964, "learning_rate": 4.935850218121467e-06, "loss": 0.5006, "step": 12380 }, { "epoch": 0.518994372425935, "grad_norm": 1.7078616619110107, "learning_rate": 4.9351714222936905e-06, "loss": 0.4722, "step": 12381 }, { "epoch": 0.519036291040944, "grad_norm": 1.576859951019287, "learning_rate": 4.934492627660939e-06, "loss": 0.468, "step": 12382 }, { "epoch": 0.519078209655953, "grad_norm": 2.042217493057251, "learning_rate": 4.933813834235726e-06, "loss": 0.54, "step": 12383 }, { "epoch": 0.5191201282709619, "grad_norm": 1.7124524116516113, "learning_rate": 4.933135042030562e-06, "loss": 0.5193, "step": 12384 }, { "epoch": 0.5191620468859709, "grad_norm": 3.0515947341918945, "learning_rate": 4.932456251057963e-06, "loss": 0.5476, "step": 12385 }, { "epoch": 0.5192039655009798, "grad_norm": 1.5613352060317993, "learning_rate": 4.931777461330438e-06, "loss": 0.4569, "step": 12386 }, { "epoch": 0.5192458841159888, "grad_norm": 1.5161688327789307, "learning_rate": 4.931098672860501e-06, "loss": 0.4751, "step": 12387 }, { "epoch": 0.5192878027309977, "grad_norm": 1.7473074197769165, "learning_rate": 4.930419885660667e-06, "loss": 0.4657, "step": 12388 }, { "epoch": 0.5193297213460067, "grad_norm": 1.8151661157608032, "learning_rate": 4.929741099743444e-06, "loss": 0.494, "step": 12389 }, { "epoch": 0.5193716399610157, "grad_norm": 2.0484297275543213, "learning_rate": 4.9290623151213465e-06, "loss": 0.5021, "step": 12390 }, { "epoch": 0.5194135585760247, "grad_norm": 1.7451039552688599, "learning_rate": 4.928383531806889e-06, "loss": 0.5578, "step": 12391 }, { "epoch": 0.5194554771910336, "grad_norm": 1.8499913215637207, "learning_rate": 4.927704749812581e-06, "loss": 0.4987, "step": 12392 }, { "epoch": 0.5194973958060426, "grad_norm": 1.6520222425460815, "learning_rate": 4.927025969150936e-06, "loss": 0.4889, "step": 12393 }, { "epoch": 0.5195393144210515, "grad_norm": 1.7519750595092773, "learning_rate": 4.926347189834467e-06, "loss": 0.4785, "step": 12394 }, { "epoch": 0.5195812330360605, "grad_norm": 1.6391886472702026, "learning_rate": 4.925668411875685e-06, "loss": 0.4673, "step": 12395 }, { "epoch": 0.5196231516510694, "grad_norm": 1.4521019458770752, "learning_rate": 4.924989635287103e-06, "loss": 0.4468, "step": 12396 }, { "epoch": 0.5196650702660784, "grad_norm": 2.5171217918395996, "learning_rate": 4.924310860081235e-06, "loss": 0.5178, "step": 12397 }, { "epoch": 0.5197069888810874, "grad_norm": 1.5928430557250977, "learning_rate": 4.923632086270589e-06, "loss": 0.4762, "step": 12398 }, { "epoch": 0.5197489074960964, "grad_norm": 1.6431379318237305, "learning_rate": 4.922953313867681e-06, "loss": 0.5107, "step": 12399 }, { "epoch": 0.5197908261111053, "grad_norm": 1.572811484336853, "learning_rate": 4.922274542885023e-06, "loss": 0.4506, "step": 12400 }, { "epoch": 0.5198327447261143, "grad_norm": 1.7019574642181396, "learning_rate": 4.921595773335126e-06, "loss": 0.4831, "step": 12401 }, { "epoch": 0.5198746633411232, "grad_norm": 1.810787558555603, "learning_rate": 4.920917005230502e-06, "loss": 0.5291, "step": 12402 }, { "epoch": 0.5199165819561322, "grad_norm": 2.0706472396850586, "learning_rate": 4.920238238583665e-06, "loss": 0.4902, "step": 12403 }, { "epoch": 0.5199585005711411, "grad_norm": 1.6691787242889404, "learning_rate": 4.919559473407124e-06, "loss": 0.5236, "step": 12404 }, { "epoch": 0.5200004191861501, "grad_norm": 1.832823395729065, "learning_rate": 4.918880709713393e-06, "loss": 0.4181, "step": 12405 }, { "epoch": 0.520042337801159, "grad_norm": 1.414800763130188, "learning_rate": 4.918201947514987e-06, "loss": 0.4651, "step": 12406 }, { "epoch": 0.520084256416168, "grad_norm": 1.55849289894104, "learning_rate": 4.917523186824411e-06, "loss": 0.5682, "step": 12407 }, { "epoch": 0.520126175031177, "grad_norm": 1.5722723007202148, "learning_rate": 4.916844427654182e-06, "loss": 0.5059, "step": 12408 }, { "epoch": 0.5201680936461859, "grad_norm": 1.7533389329910278, "learning_rate": 4.9161656700168126e-06, "loss": 0.5628, "step": 12409 }, { "epoch": 0.5202100122611949, "grad_norm": 1.710349202156067, "learning_rate": 4.915486913924813e-06, "loss": 0.5008, "step": 12410 }, { "epoch": 0.5202519308762038, "grad_norm": 1.776035189628601, "learning_rate": 4.9148081593906935e-06, "loss": 0.4702, "step": 12411 }, { "epoch": 0.5202938494912128, "grad_norm": 1.6284939050674438, "learning_rate": 4.91412940642697e-06, "loss": 0.4608, "step": 12412 }, { "epoch": 0.5203357681062217, "grad_norm": 1.570407509803772, "learning_rate": 4.9134506550461505e-06, "loss": 0.5019, "step": 12413 }, { "epoch": 0.5203776867212307, "grad_norm": 2.064829111099243, "learning_rate": 4.912771905260749e-06, "loss": 0.5221, "step": 12414 }, { "epoch": 0.5204196053362397, "grad_norm": 1.8100087642669678, "learning_rate": 4.912093157083278e-06, "loss": 0.4889, "step": 12415 }, { "epoch": 0.5204615239512487, "grad_norm": 2.4018280506134033, "learning_rate": 4.911414410526246e-06, "loss": 0.5194, "step": 12416 }, { "epoch": 0.5205034425662576, "grad_norm": 1.7709510326385498, "learning_rate": 4.910735665602167e-06, "loss": 0.5315, "step": 12417 }, { "epoch": 0.5205453611812666, "grad_norm": 1.7696198225021362, "learning_rate": 4.910056922323555e-06, "loss": 0.5046, "step": 12418 }, { "epoch": 0.5205872797962755, "grad_norm": 1.5256521701812744, "learning_rate": 4.909378180702918e-06, "loss": 0.4879, "step": 12419 }, { "epoch": 0.5206291984112845, "grad_norm": 2.2433722019195557, "learning_rate": 4.908699440752768e-06, "loss": 0.5183, "step": 12420 }, { "epoch": 0.5206711170262934, "grad_norm": 1.642586350440979, "learning_rate": 4.9080207024856194e-06, "loss": 0.5135, "step": 12421 }, { "epoch": 0.5207130356413024, "grad_norm": 1.8435518741607666, "learning_rate": 4.90734196591398e-06, "loss": 0.5139, "step": 12422 }, { "epoch": 0.5207549542563114, "grad_norm": 1.648476243019104, "learning_rate": 4.906663231050364e-06, "loss": 0.4603, "step": 12423 }, { "epoch": 0.5207968728713204, "grad_norm": 3.1711971759796143, "learning_rate": 4.9059844979072825e-06, "loss": 0.4907, "step": 12424 }, { "epoch": 0.5208387914863293, "grad_norm": 1.5369048118591309, "learning_rate": 4.9053057664972475e-06, "loss": 0.521, "step": 12425 }, { "epoch": 0.5208807101013383, "grad_norm": 1.5484997034072876, "learning_rate": 4.904627036832769e-06, "loss": 0.4649, "step": 12426 }, { "epoch": 0.5209226287163472, "grad_norm": 1.6668453216552734, "learning_rate": 4.90394830892636e-06, "loss": 0.5023, "step": 12427 }, { "epoch": 0.5209645473313562, "grad_norm": 1.5801191329956055, "learning_rate": 4.903269582790531e-06, "loss": 0.5151, "step": 12428 }, { "epoch": 0.5210064659463651, "grad_norm": 1.6748404502868652, "learning_rate": 4.9025908584377915e-06, "loss": 0.5089, "step": 12429 }, { "epoch": 0.5210483845613741, "grad_norm": 1.6774710416793823, "learning_rate": 4.9019121358806586e-06, "loss": 0.5279, "step": 12430 }, { "epoch": 0.521090303176383, "grad_norm": 1.7159382104873657, "learning_rate": 4.901233415131637e-06, "loss": 0.4985, "step": 12431 }, { "epoch": 0.521132221791392, "grad_norm": 1.5308997631072998, "learning_rate": 4.9005546962032414e-06, "loss": 0.4685, "step": 12432 }, { "epoch": 0.521174140406401, "grad_norm": 1.5220297574996948, "learning_rate": 4.899875979107983e-06, "loss": 0.4712, "step": 12433 }, { "epoch": 0.5212160590214099, "grad_norm": 1.538900375366211, "learning_rate": 4.899197263858372e-06, "loss": 0.4798, "step": 12434 }, { "epoch": 0.5212579776364189, "grad_norm": 1.6216557025909424, "learning_rate": 4.898518550466919e-06, "loss": 0.461, "step": 12435 }, { "epoch": 0.5212998962514278, "grad_norm": 1.6247929334640503, "learning_rate": 4.897839838946139e-06, "loss": 0.5012, "step": 12436 }, { "epoch": 0.5213418148664368, "grad_norm": 1.8068181276321411, "learning_rate": 4.897161129308538e-06, "loss": 0.4766, "step": 12437 }, { "epoch": 0.5213837334814457, "grad_norm": 1.620928406715393, "learning_rate": 4.896482421566631e-06, "loss": 0.4982, "step": 12438 }, { "epoch": 0.5214256520964548, "grad_norm": 1.5360326766967773, "learning_rate": 4.895803715732927e-06, "loss": 0.5065, "step": 12439 }, { "epoch": 0.5214675707114637, "grad_norm": 1.586369276046753, "learning_rate": 4.8951250118199365e-06, "loss": 0.5525, "step": 12440 }, { "epoch": 0.5215094893264727, "grad_norm": 1.641353964805603, "learning_rate": 4.894446309840173e-06, "loss": 0.4889, "step": 12441 }, { "epoch": 0.5215514079414816, "grad_norm": 2.4605634212493896, "learning_rate": 4.893767609806144e-06, "loss": 0.5167, "step": 12442 }, { "epoch": 0.5215933265564906, "grad_norm": 1.6995081901550293, "learning_rate": 4.893088911730363e-06, "loss": 0.4815, "step": 12443 }, { "epoch": 0.5216352451714995, "grad_norm": 1.7221415042877197, "learning_rate": 4.892410215625342e-06, "loss": 0.5288, "step": 12444 }, { "epoch": 0.5216771637865085, "grad_norm": 1.6824078559875488, "learning_rate": 4.891731521503588e-06, "loss": 0.5232, "step": 12445 }, { "epoch": 0.5217190824015174, "grad_norm": 1.7499074935913086, "learning_rate": 4.891052829377612e-06, "loss": 0.4814, "step": 12446 }, { "epoch": 0.5217610010165264, "grad_norm": 1.552907109260559, "learning_rate": 4.89037413925993e-06, "loss": 0.4833, "step": 12447 }, { "epoch": 0.5218029196315354, "grad_norm": 1.7768371105194092, "learning_rate": 4.8896954511630476e-06, "loss": 0.5173, "step": 12448 }, { "epoch": 0.5218448382465444, "grad_norm": 1.8022500276565552, "learning_rate": 4.889016765099476e-06, "loss": 0.5125, "step": 12449 }, { "epoch": 0.5218867568615533, "grad_norm": 2.047614812850952, "learning_rate": 4.88833808108173e-06, "loss": 0.4739, "step": 12450 }, { "epoch": 0.5219286754765623, "grad_norm": 2.0263614654541016, "learning_rate": 4.887659399122314e-06, "loss": 0.4802, "step": 12451 }, { "epoch": 0.5219705940915712, "grad_norm": 2.7828376293182373, "learning_rate": 4.886980719233743e-06, "loss": 0.5196, "step": 12452 }, { "epoch": 0.5220125127065802, "grad_norm": 1.9262727499008179, "learning_rate": 4.886302041428527e-06, "loss": 0.5251, "step": 12453 }, { "epoch": 0.5220544313215891, "grad_norm": 1.7272183895111084, "learning_rate": 4.8856233657191736e-06, "loss": 0.5081, "step": 12454 }, { "epoch": 0.5220963499365981, "grad_norm": 1.7826266288757324, "learning_rate": 4.884944692118195e-06, "loss": 0.531, "step": 12455 }, { "epoch": 0.522138268551607, "grad_norm": 2.5976362228393555, "learning_rate": 4.884266020638105e-06, "loss": 0.5187, "step": 12456 }, { "epoch": 0.522180187166616, "grad_norm": 1.5931860208511353, "learning_rate": 4.8835873512914095e-06, "loss": 0.505, "step": 12457 }, { "epoch": 0.522222105781625, "grad_norm": 1.9170446395874023, "learning_rate": 4.882908684090619e-06, "loss": 0.5104, "step": 12458 }, { "epoch": 0.5222640243966339, "grad_norm": 2.3546957969665527, "learning_rate": 4.882230019048247e-06, "loss": 0.4831, "step": 12459 }, { "epoch": 0.5223059430116429, "grad_norm": 1.4862064123153687, "learning_rate": 4.8815513561768005e-06, "loss": 0.4611, "step": 12460 }, { "epoch": 0.5223478616266518, "grad_norm": 1.4628283977508545, "learning_rate": 4.880872695488791e-06, "loss": 0.4591, "step": 12461 }, { "epoch": 0.5223897802416608, "grad_norm": 2.0866591930389404, "learning_rate": 4.88019403699673e-06, "loss": 0.5386, "step": 12462 }, { "epoch": 0.5224316988566697, "grad_norm": 1.9644702672958374, "learning_rate": 4.879515380713125e-06, "loss": 0.5246, "step": 12463 }, { "epoch": 0.5224736174716788, "grad_norm": 1.8196659088134766, "learning_rate": 4.878836726650486e-06, "loss": 0.4285, "step": 12464 }, { "epoch": 0.5225155360866877, "grad_norm": 1.9830083847045898, "learning_rate": 4.8781580748213275e-06, "loss": 0.547, "step": 12465 }, { "epoch": 0.5225574547016967, "grad_norm": 1.919579267501831, "learning_rate": 4.877479425238155e-06, "loss": 0.4327, "step": 12466 }, { "epoch": 0.5225993733167056, "grad_norm": 1.730913758277893, "learning_rate": 4.8768007779134795e-06, "loss": 0.4876, "step": 12467 }, { "epoch": 0.5226412919317146, "grad_norm": 2.07967209815979, "learning_rate": 4.876122132859813e-06, "loss": 0.4868, "step": 12468 }, { "epoch": 0.5226832105467235, "grad_norm": 1.708114504814148, "learning_rate": 4.875443490089661e-06, "loss": 0.4726, "step": 12469 }, { "epoch": 0.5227251291617325, "grad_norm": 1.7405232191085815, "learning_rate": 4.874764849615538e-06, "loss": 0.5153, "step": 12470 }, { "epoch": 0.5227670477767414, "grad_norm": 1.7675809860229492, "learning_rate": 4.874086211449952e-06, "loss": 0.4818, "step": 12471 }, { "epoch": 0.5228089663917505, "grad_norm": 1.5835644006729126, "learning_rate": 4.873407575605412e-06, "loss": 0.5045, "step": 12472 }, { "epoch": 0.5228508850067594, "grad_norm": 1.6067254543304443, "learning_rate": 4.872728942094427e-06, "loss": 0.4874, "step": 12473 }, { "epoch": 0.5228928036217684, "grad_norm": 1.776391863822937, "learning_rate": 4.872050310929512e-06, "loss": 0.5431, "step": 12474 }, { "epoch": 0.5229347222367773, "grad_norm": 1.911441683769226, "learning_rate": 4.871371682123168e-06, "loss": 0.5351, "step": 12475 }, { "epoch": 0.5229766408517863, "grad_norm": 1.7789177894592285, "learning_rate": 4.870693055687912e-06, "loss": 0.5166, "step": 12476 }, { "epoch": 0.5230185594667952, "grad_norm": 1.7074599266052246, "learning_rate": 4.870014431636252e-06, "loss": 0.5466, "step": 12477 }, { "epoch": 0.5230604780818042, "grad_norm": 1.7986505031585693, "learning_rate": 4.869335809980693e-06, "loss": 0.497, "step": 12478 }, { "epoch": 0.5231023966968131, "grad_norm": 1.887317180633545, "learning_rate": 4.868657190733748e-06, "loss": 0.5338, "step": 12479 }, { "epoch": 0.5231443153118222, "grad_norm": 2.153245210647583, "learning_rate": 4.867978573907928e-06, "loss": 0.5445, "step": 12480 }, { "epoch": 0.5231862339268311, "grad_norm": 2.0457849502563477, "learning_rate": 4.867299959515739e-06, "loss": 0.4556, "step": 12481 }, { "epoch": 0.52322815254184, "grad_norm": 1.7395159006118774, "learning_rate": 4.866621347569692e-06, "loss": 0.4954, "step": 12482 }, { "epoch": 0.523270071156849, "grad_norm": 1.7210912704467773, "learning_rate": 4.865942738082298e-06, "loss": 0.5727, "step": 12483 }, { "epoch": 0.5233119897718579, "grad_norm": 1.8853694200515747, "learning_rate": 4.865264131066061e-06, "loss": 0.4577, "step": 12484 }, { "epoch": 0.5233539083868669, "grad_norm": 1.7288044691085815, "learning_rate": 4.8645855265334954e-06, "loss": 0.5009, "step": 12485 }, { "epoch": 0.5233958270018758, "grad_norm": 1.8189806938171387, "learning_rate": 4.8639069244971095e-06, "loss": 0.498, "step": 12486 }, { "epoch": 0.5234377456168848, "grad_norm": 1.7100887298583984, "learning_rate": 4.86322832496941e-06, "loss": 0.5043, "step": 12487 }, { "epoch": 0.5234796642318937, "grad_norm": 1.8281713724136353, "learning_rate": 4.8625497279629055e-06, "loss": 0.4762, "step": 12488 }, { "epoch": 0.5235215828469028, "grad_norm": 1.7821097373962402, "learning_rate": 4.86187113349011e-06, "loss": 0.5285, "step": 12489 }, { "epoch": 0.5235635014619117, "grad_norm": 1.6387954950332642, "learning_rate": 4.861192541563528e-06, "loss": 0.452, "step": 12490 }, { "epoch": 0.5236054200769207, "grad_norm": 2.015448570251465, "learning_rate": 4.860513952195668e-06, "loss": 0.5092, "step": 12491 }, { "epoch": 0.5236473386919296, "grad_norm": 1.7029560804367065, "learning_rate": 4.859835365399043e-06, "loss": 0.5845, "step": 12492 }, { "epoch": 0.5236892573069386, "grad_norm": 1.5297892093658447, "learning_rate": 4.859156781186157e-06, "loss": 0.4756, "step": 12493 }, { "epoch": 0.5237311759219475, "grad_norm": 1.721099853515625, "learning_rate": 4.8584781995695225e-06, "loss": 0.5665, "step": 12494 }, { "epoch": 0.5237730945369565, "grad_norm": 1.5742229223251343, "learning_rate": 4.857799620561648e-06, "loss": 0.4867, "step": 12495 }, { "epoch": 0.5238150131519654, "grad_norm": 1.7278333902359009, "learning_rate": 4.857121044175038e-06, "loss": 0.5096, "step": 12496 }, { "epoch": 0.5238569317669745, "grad_norm": 1.466888666152954, "learning_rate": 4.856442470422204e-06, "loss": 0.5291, "step": 12497 }, { "epoch": 0.5238988503819834, "grad_norm": 1.7586002349853516, "learning_rate": 4.855763899315658e-06, "loss": 0.4677, "step": 12498 }, { "epoch": 0.5239407689969924, "grad_norm": 1.7248945236206055, "learning_rate": 4.855085330867902e-06, "loss": 0.5147, "step": 12499 }, { "epoch": 0.5239826876120013, "grad_norm": 1.530383825302124, "learning_rate": 4.854406765091448e-06, "loss": 0.473, "step": 12500 }, { "epoch": 0.5240246062270103, "grad_norm": 1.7464635372161865, "learning_rate": 4.853728201998806e-06, "loss": 0.4577, "step": 12501 }, { "epoch": 0.5240665248420192, "grad_norm": 1.6449573040008545, "learning_rate": 4.85304964160248e-06, "loss": 0.5033, "step": 12502 }, { "epoch": 0.5241084434570282, "grad_norm": 1.7579692602157593, "learning_rate": 4.852371083914982e-06, "loss": 0.5196, "step": 12503 }, { "epoch": 0.5241503620720371, "grad_norm": 1.5384769439697266, "learning_rate": 4.85169252894882e-06, "loss": 0.5339, "step": 12504 }, { "epoch": 0.5241922806870462, "grad_norm": 1.6691300868988037, "learning_rate": 4.851013976716498e-06, "loss": 0.5435, "step": 12505 }, { "epoch": 0.5242341993020551, "grad_norm": 1.5055285692214966, "learning_rate": 4.850335427230528e-06, "loss": 0.4631, "step": 12506 }, { "epoch": 0.524276117917064, "grad_norm": 1.6725263595581055, "learning_rate": 4.84965688050342e-06, "loss": 0.4963, "step": 12507 }, { "epoch": 0.524318036532073, "grad_norm": 1.8955941200256348, "learning_rate": 4.848978336547679e-06, "loss": 0.5126, "step": 12508 }, { "epoch": 0.5243599551470819, "grad_norm": 1.7001521587371826, "learning_rate": 4.8482997953758124e-06, "loss": 0.5455, "step": 12509 }, { "epoch": 0.5244018737620909, "grad_norm": 1.6143598556518555, "learning_rate": 4.847621257000331e-06, "loss": 0.4706, "step": 12510 }, { "epoch": 0.5244437923770998, "grad_norm": 1.619430422782898, "learning_rate": 4.846942721433738e-06, "loss": 0.4736, "step": 12511 }, { "epoch": 0.5244857109921088, "grad_norm": 2.584833860397339, "learning_rate": 4.846264188688548e-06, "loss": 0.5119, "step": 12512 }, { "epoch": 0.5245276296071177, "grad_norm": 1.7119275331497192, "learning_rate": 4.845585658777263e-06, "loss": 0.53, "step": 12513 }, { "epoch": 0.5245695482221268, "grad_norm": 1.8038403987884521, "learning_rate": 4.8449071317123925e-06, "loss": 0.4908, "step": 12514 }, { "epoch": 0.5246114668371357, "grad_norm": 1.8967245817184448, "learning_rate": 4.844228607506447e-06, "loss": 0.5129, "step": 12515 }, { "epoch": 0.5246533854521447, "grad_norm": 1.8535436391830444, "learning_rate": 4.843550086171929e-06, "loss": 0.544, "step": 12516 }, { "epoch": 0.5246953040671536, "grad_norm": 1.711838960647583, "learning_rate": 4.84287156772135e-06, "loss": 0.5489, "step": 12517 }, { "epoch": 0.5247372226821626, "grad_norm": 1.722387433052063, "learning_rate": 4.842193052167217e-06, "loss": 0.5029, "step": 12518 }, { "epoch": 0.5247791412971715, "grad_norm": 1.6201938390731812, "learning_rate": 4.841514539522036e-06, "loss": 0.5232, "step": 12519 }, { "epoch": 0.5248210599121805, "grad_norm": 1.747510552406311, "learning_rate": 4.840836029798315e-06, "loss": 0.5159, "step": 12520 }, { "epoch": 0.5248629785271894, "grad_norm": 1.8520358800888062, "learning_rate": 4.840157523008564e-06, "loss": 0.5342, "step": 12521 }, { "epoch": 0.5249048971421985, "grad_norm": 1.7341506481170654, "learning_rate": 4.839479019165285e-06, "loss": 0.5668, "step": 12522 }, { "epoch": 0.5249468157572074, "grad_norm": 1.7327914237976074, "learning_rate": 4.83880051828099e-06, "loss": 0.5453, "step": 12523 }, { "epoch": 0.5249887343722164, "grad_norm": 1.7206575870513916, "learning_rate": 4.838122020368186e-06, "loss": 0.4799, "step": 12524 }, { "epoch": 0.5250306529872253, "grad_norm": 1.7406537532806396, "learning_rate": 4.837443525439376e-06, "loss": 0.4407, "step": 12525 }, { "epoch": 0.5250725716022343, "grad_norm": 1.7266396284103394, "learning_rate": 4.836765033507071e-06, "loss": 0.4983, "step": 12526 }, { "epoch": 0.5251144902172432, "grad_norm": 1.6440664529800415, "learning_rate": 4.836086544583778e-06, "loss": 0.5014, "step": 12527 }, { "epoch": 0.5251564088322522, "grad_norm": 1.7416813373565674, "learning_rate": 4.835408058682003e-06, "loss": 0.4871, "step": 12528 }, { "epoch": 0.5251983274472611, "grad_norm": 2.2021484375, "learning_rate": 4.83472957581425e-06, "loss": 0.5854, "step": 12529 }, { "epoch": 0.5252402460622702, "grad_norm": 1.7389017343521118, "learning_rate": 4.834051095993033e-06, "loss": 0.5406, "step": 12530 }, { "epoch": 0.5252821646772791, "grad_norm": 1.926190972328186, "learning_rate": 4.833372619230851e-06, "loss": 0.4887, "step": 12531 }, { "epoch": 0.525324083292288, "grad_norm": 1.975886344909668, "learning_rate": 4.832694145540216e-06, "loss": 0.5031, "step": 12532 }, { "epoch": 0.525366001907297, "grad_norm": 1.7505507469177246, "learning_rate": 4.832015674933634e-06, "loss": 0.5552, "step": 12533 }, { "epoch": 0.5254079205223059, "grad_norm": 1.9456311464309692, "learning_rate": 4.831337207423609e-06, "loss": 0.5542, "step": 12534 }, { "epoch": 0.5254498391373149, "grad_norm": 1.7023791074752808, "learning_rate": 4.830658743022649e-06, "loss": 0.4765, "step": 12535 }, { "epoch": 0.5254917577523238, "grad_norm": 1.6679221391677856, "learning_rate": 4.829980281743263e-06, "loss": 0.4576, "step": 12536 }, { "epoch": 0.5255336763673328, "grad_norm": 1.784698724746704, "learning_rate": 4.829301823597955e-06, "loss": 0.5157, "step": 12537 }, { "epoch": 0.5255755949823417, "grad_norm": 2.5354738235473633, "learning_rate": 4.828623368599231e-06, "loss": 0.4739, "step": 12538 }, { "epoch": 0.5256175135973508, "grad_norm": 1.615088939666748, "learning_rate": 4.8279449167595995e-06, "loss": 0.5117, "step": 12539 }, { "epoch": 0.5256594322123597, "grad_norm": 2.496683120727539, "learning_rate": 4.827266468091564e-06, "loss": 0.5316, "step": 12540 }, { "epoch": 0.5257013508273687, "grad_norm": 1.865936040878296, "learning_rate": 4.826588022607633e-06, "loss": 0.4766, "step": 12541 }, { "epoch": 0.5257432694423776, "grad_norm": 1.8844754695892334, "learning_rate": 4.825909580320313e-06, "loss": 0.5075, "step": 12542 }, { "epoch": 0.5257851880573866, "grad_norm": 1.6872529983520508, "learning_rate": 4.8252311412421085e-06, "loss": 0.5003, "step": 12543 }, { "epoch": 0.5258271066723955, "grad_norm": 1.697681188583374, "learning_rate": 4.824552705385525e-06, "loss": 0.4853, "step": 12544 }, { "epoch": 0.5258690252874045, "grad_norm": 1.8902287483215332, "learning_rate": 4.823874272763073e-06, "loss": 0.501, "step": 12545 }, { "epoch": 0.5259109439024134, "grad_norm": 1.8935742378234863, "learning_rate": 4.823195843387253e-06, "loss": 0.5312, "step": 12546 }, { "epoch": 0.5259528625174225, "grad_norm": 2.2447566986083984, "learning_rate": 4.822517417270573e-06, "loss": 0.5405, "step": 12547 }, { "epoch": 0.5259947811324314, "grad_norm": 1.714138388633728, "learning_rate": 4.821838994425541e-06, "loss": 0.5149, "step": 12548 }, { "epoch": 0.5260366997474404, "grad_norm": 2.077540636062622, "learning_rate": 4.821160574864658e-06, "loss": 0.5558, "step": 12549 }, { "epoch": 0.5260786183624493, "grad_norm": 1.7084661722183228, "learning_rate": 4.820482158600434e-06, "loss": 0.46, "step": 12550 }, { "epoch": 0.5261205369774583, "grad_norm": 1.962821364402771, "learning_rate": 4.819803745645375e-06, "loss": 0.5548, "step": 12551 }, { "epoch": 0.5261624555924672, "grad_norm": 1.835936427116394, "learning_rate": 4.819125336011982e-06, "loss": 0.5338, "step": 12552 }, { "epoch": 0.5262043742074762, "grad_norm": 1.8039846420288086, "learning_rate": 4.818446929712764e-06, "loss": 0.5049, "step": 12553 }, { "epoch": 0.5262462928224851, "grad_norm": 1.6960865259170532, "learning_rate": 4.8177685267602276e-06, "loss": 0.4786, "step": 12554 }, { "epoch": 0.5262882114374942, "grad_norm": 2.139575242996216, "learning_rate": 4.817090127166875e-06, "loss": 0.5292, "step": 12555 }, { "epoch": 0.5263301300525031, "grad_norm": 1.940937876701355, "learning_rate": 4.816411730945213e-06, "loss": 0.4762, "step": 12556 }, { "epoch": 0.526372048667512, "grad_norm": 2.1249871253967285, "learning_rate": 4.8157333381077484e-06, "loss": 0.4873, "step": 12557 }, { "epoch": 0.526413967282521, "grad_norm": 1.8846863508224487, "learning_rate": 4.815054948666982e-06, "loss": 0.5138, "step": 12558 }, { "epoch": 0.5264558858975299, "grad_norm": 1.7489407062530518, "learning_rate": 4.814376562635424e-06, "loss": 0.5019, "step": 12559 }, { "epoch": 0.5264978045125389, "grad_norm": 1.710193395614624, "learning_rate": 4.813698180025579e-06, "loss": 0.4861, "step": 12560 }, { "epoch": 0.5265397231275478, "grad_norm": 1.875300645828247, "learning_rate": 4.813019800849948e-06, "loss": 0.4797, "step": 12561 }, { "epoch": 0.5265816417425568, "grad_norm": 1.8405709266662598, "learning_rate": 4.812341425121038e-06, "loss": 0.4756, "step": 12562 }, { "epoch": 0.5266235603575657, "grad_norm": 1.609065294265747, "learning_rate": 4.811663052851357e-06, "loss": 0.5037, "step": 12563 }, { "epoch": 0.5266654789725748, "grad_norm": 1.7339457273483276, "learning_rate": 4.8109846840534055e-06, "loss": 0.4278, "step": 12564 }, { "epoch": 0.5267073975875837, "grad_norm": 1.7548038959503174, "learning_rate": 4.81030631873969e-06, "loss": 0.4684, "step": 12565 }, { "epoch": 0.5267493162025927, "grad_norm": 1.891700029373169, "learning_rate": 4.809627956922717e-06, "loss": 0.5099, "step": 12566 }, { "epoch": 0.5267912348176016, "grad_norm": 2.3660953044891357, "learning_rate": 4.808949598614986e-06, "loss": 0.5277, "step": 12567 }, { "epoch": 0.5268331534326106, "grad_norm": 2.0062618255615234, "learning_rate": 4.808271243829008e-06, "loss": 0.4849, "step": 12568 }, { "epoch": 0.5268750720476195, "grad_norm": 1.8371139764785767, "learning_rate": 4.807592892577284e-06, "loss": 0.5084, "step": 12569 }, { "epoch": 0.5269169906626285, "grad_norm": 1.684533715248108, "learning_rate": 4.8069145448723185e-06, "loss": 0.5188, "step": 12570 }, { "epoch": 0.5269589092776374, "grad_norm": 1.9294277429580688, "learning_rate": 4.806236200726615e-06, "loss": 0.5981, "step": 12571 }, { "epoch": 0.5270008278926465, "grad_norm": 1.9075130224227905, "learning_rate": 4.805557860152682e-06, "loss": 0.5067, "step": 12572 }, { "epoch": 0.5270427465076554, "grad_norm": 2.2248928546905518, "learning_rate": 4.80487952316302e-06, "loss": 0.551, "step": 12573 }, { "epoch": 0.5270846651226644, "grad_norm": 1.7620521783828735, "learning_rate": 4.804201189770134e-06, "loss": 0.4799, "step": 12574 }, { "epoch": 0.5271265837376733, "grad_norm": 2.052032947540283, "learning_rate": 4.80352285998653e-06, "loss": 0.4605, "step": 12575 }, { "epoch": 0.5271685023526823, "grad_norm": 2.9253427982330322, "learning_rate": 4.802844533824708e-06, "loss": 0.5014, "step": 12576 }, { "epoch": 0.5272104209676912, "grad_norm": 2.4024901390075684, "learning_rate": 4.802166211297175e-06, "loss": 0.4722, "step": 12577 }, { "epoch": 0.5272523395827002, "grad_norm": 1.8168867826461792, "learning_rate": 4.8014878924164355e-06, "loss": 0.485, "step": 12578 }, { "epoch": 0.5272942581977091, "grad_norm": 1.9458611011505127, "learning_rate": 4.800809577194992e-06, "loss": 0.5078, "step": 12579 }, { "epoch": 0.5273361768127182, "grad_norm": 1.63385009765625, "learning_rate": 4.800131265645347e-06, "loss": 0.4983, "step": 12580 }, { "epoch": 0.5273780954277271, "grad_norm": 2.3594040870666504, "learning_rate": 4.799452957780009e-06, "loss": 0.4898, "step": 12581 }, { "epoch": 0.527420014042736, "grad_norm": 1.7973089218139648, "learning_rate": 4.798774653611475e-06, "loss": 0.485, "step": 12582 }, { "epoch": 0.527461932657745, "grad_norm": 2.0314905643463135, "learning_rate": 4.798096353152254e-06, "loss": 0.5374, "step": 12583 }, { "epoch": 0.5275038512727539, "grad_norm": 2.2020211219787598, "learning_rate": 4.797418056414848e-06, "loss": 0.4743, "step": 12584 }, { "epoch": 0.5275457698877629, "grad_norm": 1.8608709573745728, "learning_rate": 4.7967397634117585e-06, "loss": 0.5138, "step": 12585 }, { "epoch": 0.5275876885027718, "grad_norm": 1.7884633541107178, "learning_rate": 4.796061474155492e-06, "loss": 0.5304, "step": 12586 }, { "epoch": 0.5276296071177808, "grad_norm": 1.8187484741210938, "learning_rate": 4.795383188658549e-06, "loss": 0.4797, "step": 12587 }, { "epoch": 0.5276715257327897, "grad_norm": 4.445209980010986, "learning_rate": 4.794704906933434e-06, "loss": 0.4846, "step": 12588 }, { "epoch": 0.5277134443477988, "grad_norm": 1.6464855670928955, "learning_rate": 4.794026628992653e-06, "loss": 0.53, "step": 12589 }, { "epoch": 0.5277553629628077, "grad_norm": 2.1945834159851074, "learning_rate": 4.7933483548487035e-06, "loss": 0.474, "step": 12590 }, { "epoch": 0.5277972815778167, "grad_norm": 1.6873725652694702, "learning_rate": 4.792670084514091e-06, "loss": 0.5046, "step": 12591 }, { "epoch": 0.5278392001928256, "grad_norm": 1.574795126914978, "learning_rate": 4.79199181800132e-06, "loss": 0.4919, "step": 12592 }, { "epoch": 0.5278811188078346, "grad_norm": 1.7431328296661377, "learning_rate": 4.791313555322893e-06, "loss": 0.52, "step": 12593 }, { "epoch": 0.5279230374228435, "grad_norm": 1.6835049390792847, "learning_rate": 4.79063529649131e-06, "loss": 0.4971, "step": 12594 }, { "epoch": 0.5279649560378525, "grad_norm": 2.599034309387207, "learning_rate": 4.789957041519078e-06, "loss": 0.5149, "step": 12595 }, { "epoch": 0.5280068746528614, "grad_norm": 1.9659985303878784, "learning_rate": 4.789278790418696e-06, "loss": 0.5023, "step": 12596 }, { "epoch": 0.5280487932678705, "grad_norm": 1.914428472518921, "learning_rate": 4.78860054320267e-06, "loss": 0.5293, "step": 12597 }, { "epoch": 0.5280907118828794, "grad_norm": 2.3110671043395996, "learning_rate": 4.7879222998835e-06, "loss": 0.4811, "step": 12598 }, { "epoch": 0.5281326304978884, "grad_norm": 1.8109586238861084, "learning_rate": 4.787244060473688e-06, "loss": 0.4686, "step": 12599 }, { "epoch": 0.5281745491128973, "grad_norm": 1.7778284549713135, "learning_rate": 4.786565824985737e-06, "loss": 0.5214, "step": 12600 }, { "epoch": 0.5282164677279063, "grad_norm": 1.7833292484283447, "learning_rate": 4.785887593432153e-06, "loss": 0.4885, "step": 12601 }, { "epoch": 0.5282583863429152, "grad_norm": 2.0567262172698975, "learning_rate": 4.785209365825434e-06, "loss": 0.4737, "step": 12602 }, { "epoch": 0.5283003049579242, "grad_norm": 1.8698891401290894, "learning_rate": 4.784531142178082e-06, "loss": 0.4627, "step": 12603 }, { "epoch": 0.5283422235729331, "grad_norm": 2.311955690383911, "learning_rate": 4.783852922502602e-06, "loss": 0.4664, "step": 12604 }, { "epoch": 0.5283841421879422, "grad_norm": 1.9626461267471313, "learning_rate": 4.783174706811494e-06, "loss": 0.529, "step": 12605 }, { "epoch": 0.5284260608029511, "grad_norm": 2.187823534011841, "learning_rate": 4.782496495117261e-06, "loss": 0.5224, "step": 12606 }, { "epoch": 0.52846797941796, "grad_norm": 1.857643723487854, "learning_rate": 4.781818287432405e-06, "loss": 0.4701, "step": 12607 }, { "epoch": 0.528509898032969, "grad_norm": 1.7453227043151855, "learning_rate": 4.781140083769426e-06, "loss": 0.4685, "step": 12608 }, { "epoch": 0.5285518166479779, "grad_norm": 1.7732974290847778, "learning_rate": 4.780461884140826e-06, "loss": 0.5173, "step": 12609 }, { "epoch": 0.5285937352629869, "grad_norm": 2.0537269115448, "learning_rate": 4.77978368855911e-06, "loss": 0.5259, "step": 12610 }, { "epoch": 0.5286356538779958, "grad_norm": 1.6328158378601074, "learning_rate": 4.779105497036776e-06, "loss": 0.516, "step": 12611 }, { "epoch": 0.5286775724930048, "grad_norm": 1.7955026626586914, "learning_rate": 4.778427309586326e-06, "loss": 0.5325, "step": 12612 }, { "epoch": 0.5287194911080138, "grad_norm": 1.9268568754196167, "learning_rate": 4.777749126220265e-06, "loss": 0.4945, "step": 12613 }, { "epoch": 0.5287614097230228, "grad_norm": 1.7303978204727173, "learning_rate": 4.777070946951088e-06, "loss": 0.496, "step": 12614 }, { "epoch": 0.5288033283380317, "grad_norm": 1.8212873935699463, "learning_rate": 4.776392771791302e-06, "loss": 0.5377, "step": 12615 }, { "epoch": 0.5288452469530407, "grad_norm": 1.770748496055603, "learning_rate": 4.7757146007534065e-06, "loss": 0.437, "step": 12616 }, { "epoch": 0.5288871655680496, "grad_norm": 1.5585646629333496, "learning_rate": 4.775036433849901e-06, "loss": 0.4694, "step": 12617 }, { "epoch": 0.5289290841830586, "grad_norm": 1.9361492395401, "learning_rate": 4.7743582710932875e-06, "loss": 0.4932, "step": 12618 }, { "epoch": 0.5289710027980675, "grad_norm": 1.6619447469711304, "learning_rate": 4.77368011249607e-06, "loss": 0.554, "step": 12619 }, { "epoch": 0.5290129214130765, "grad_norm": 1.7306894063949585, "learning_rate": 4.773001958070745e-06, "loss": 0.5168, "step": 12620 }, { "epoch": 0.5290548400280854, "grad_norm": 1.9227983951568604, "learning_rate": 4.772323807829814e-06, "loss": 0.5279, "step": 12621 }, { "epoch": 0.5290967586430945, "grad_norm": 1.5146650075912476, "learning_rate": 4.771645661785781e-06, "loss": 0.5127, "step": 12622 }, { "epoch": 0.5291386772581034, "grad_norm": 1.7431378364562988, "learning_rate": 4.770967519951144e-06, "loss": 0.5228, "step": 12623 }, { "epoch": 0.5291805958731124, "grad_norm": 2.6862869262695312, "learning_rate": 4.770289382338402e-06, "loss": 0.5189, "step": 12624 }, { "epoch": 0.5292225144881213, "grad_norm": 1.9098883867263794, "learning_rate": 4.76961124896006e-06, "loss": 0.5291, "step": 12625 }, { "epoch": 0.5292644331031303, "grad_norm": 1.564769983291626, "learning_rate": 4.768933119828616e-06, "loss": 0.5048, "step": 12626 }, { "epoch": 0.5293063517181392, "grad_norm": 12.852276802062988, "learning_rate": 4.768254994956569e-06, "loss": 0.4936, "step": 12627 }, { "epoch": 0.5293482703331482, "grad_norm": 1.565663456916809, "learning_rate": 4.7675768743564235e-06, "loss": 0.4427, "step": 12628 }, { "epoch": 0.5293901889481571, "grad_norm": 1.450931429862976, "learning_rate": 4.766898758040674e-06, "loss": 0.4702, "step": 12629 }, { "epoch": 0.5294321075631662, "grad_norm": 1.9091413021087646, "learning_rate": 4.766220646021825e-06, "loss": 0.4588, "step": 12630 }, { "epoch": 0.5294740261781751, "grad_norm": 1.6949450969696045, "learning_rate": 4.765542538312376e-06, "loss": 0.5162, "step": 12631 }, { "epoch": 0.529515944793184, "grad_norm": 1.5326157808303833, "learning_rate": 4.764864434924825e-06, "loss": 0.4712, "step": 12632 }, { "epoch": 0.529557863408193, "grad_norm": 6.029474258422852, "learning_rate": 4.764186335871672e-06, "loss": 0.5342, "step": 12633 }, { "epoch": 0.5295997820232019, "grad_norm": 1.6792579889297485, "learning_rate": 4.76350824116542e-06, "loss": 0.4248, "step": 12634 }, { "epoch": 0.5296417006382109, "grad_norm": 1.6489861011505127, "learning_rate": 4.7628301508185666e-06, "loss": 0.4488, "step": 12635 }, { "epoch": 0.5296836192532198, "grad_norm": 1.7549574375152588, "learning_rate": 4.76215206484361e-06, "loss": 0.5023, "step": 12636 }, { "epoch": 0.5297255378682288, "grad_norm": 1.7857351303100586, "learning_rate": 4.761473983253053e-06, "loss": 0.5427, "step": 12637 }, { "epoch": 0.5297674564832378, "grad_norm": 2.06050705909729, "learning_rate": 4.760795906059391e-06, "loss": 0.5631, "step": 12638 }, { "epoch": 0.5298093750982468, "grad_norm": 1.8244162797927856, "learning_rate": 4.7601178332751275e-06, "loss": 0.4905, "step": 12639 }, { "epoch": 0.5298512937132557, "grad_norm": 1.7380893230438232, "learning_rate": 4.759439764912761e-06, "loss": 0.5253, "step": 12640 }, { "epoch": 0.5298932123282647, "grad_norm": 1.784403920173645, "learning_rate": 4.758761700984788e-06, "loss": 0.4645, "step": 12641 }, { "epoch": 0.5299351309432736, "grad_norm": 1.783103346824646, "learning_rate": 4.758083641503709e-06, "loss": 0.499, "step": 12642 }, { "epoch": 0.5299770495582826, "grad_norm": 1.6529496908187866, "learning_rate": 4.757405586482026e-06, "loss": 0.5103, "step": 12643 }, { "epoch": 0.5300189681732915, "grad_norm": 1.8219468593597412, "learning_rate": 4.7567275359322345e-06, "loss": 0.5162, "step": 12644 }, { "epoch": 0.5300608867883005, "grad_norm": 2.365276336669922, "learning_rate": 4.756049489866834e-06, "loss": 0.5327, "step": 12645 }, { "epoch": 0.5301028054033095, "grad_norm": 2.1459121704101562, "learning_rate": 4.755371448298325e-06, "loss": 0.5118, "step": 12646 }, { "epoch": 0.5301447240183185, "grad_norm": 1.7419403791427612, "learning_rate": 4.754693411239204e-06, "loss": 0.5104, "step": 12647 }, { "epoch": 0.5301866426333274, "grad_norm": 1.6521047353744507, "learning_rate": 4.754015378701972e-06, "loss": 0.456, "step": 12648 }, { "epoch": 0.5302285612483364, "grad_norm": 1.6370948553085327, "learning_rate": 4.753337350699127e-06, "loss": 0.5282, "step": 12649 }, { "epoch": 0.5302704798633453, "grad_norm": 2.5670108795166016, "learning_rate": 4.752659327243165e-06, "loss": 0.501, "step": 12650 }, { "epoch": 0.5303123984783543, "grad_norm": 2.27115535736084, "learning_rate": 4.751981308346586e-06, "loss": 0.4881, "step": 12651 }, { "epoch": 0.5303543170933632, "grad_norm": 1.4563453197479248, "learning_rate": 4.751303294021891e-06, "loss": 0.459, "step": 12652 }, { "epoch": 0.5303962357083722, "grad_norm": 1.5885573625564575, "learning_rate": 4.750625284281575e-06, "loss": 0.4475, "step": 12653 }, { "epoch": 0.5304381543233812, "grad_norm": 1.7420426607131958, "learning_rate": 4.749947279138136e-06, "loss": 0.4637, "step": 12654 }, { "epoch": 0.5304800729383902, "grad_norm": 1.9409427642822266, "learning_rate": 4.749269278604075e-06, "loss": 0.5331, "step": 12655 }, { "epoch": 0.5305219915533991, "grad_norm": 1.9201945066452026, "learning_rate": 4.748591282691887e-06, "loss": 0.4758, "step": 12656 }, { "epoch": 0.530563910168408, "grad_norm": 2.3288309574127197, "learning_rate": 4.747913291414072e-06, "loss": 0.489, "step": 12657 }, { "epoch": 0.530605828783417, "grad_norm": 1.9411163330078125, "learning_rate": 4.747235304783128e-06, "loss": 0.5027, "step": 12658 }, { "epoch": 0.5306477473984259, "grad_norm": 2.377960205078125, "learning_rate": 4.746557322811549e-06, "loss": 0.452, "step": 12659 }, { "epoch": 0.5306896660134349, "grad_norm": 2.5302534103393555, "learning_rate": 4.745879345511837e-06, "loss": 0.506, "step": 12660 }, { "epoch": 0.5307315846284438, "grad_norm": 1.7473978996276855, "learning_rate": 4.745201372896487e-06, "loss": 0.4548, "step": 12661 }, { "epoch": 0.5307735032434528, "grad_norm": 2.054351806640625, "learning_rate": 4.7445234049779985e-06, "loss": 0.542, "step": 12662 }, { "epoch": 0.5308154218584618, "grad_norm": 1.800539493560791, "learning_rate": 4.743845441768867e-06, "loss": 0.4895, "step": 12663 }, { "epoch": 0.5308573404734708, "grad_norm": 1.7425940036773682, "learning_rate": 4.7431674832815915e-06, "loss": 0.5353, "step": 12664 }, { "epoch": 0.5308992590884797, "grad_norm": 2.7333898544311523, "learning_rate": 4.7424895295286675e-06, "loss": 0.5229, "step": 12665 }, { "epoch": 0.5309411777034887, "grad_norm": 1.8439728021621704, "learning_rate": 4.741811580522594e-06, "loss": 0.4945, "step": 12666 }, { "epoch": 0.5309830963184976, "grad_norm": 2.3537585735321045, "learning_rate": 4.741133636275866e-06, "loss": 0.5244, "step": 12667 }, { "epoch": 0.5310250149335066, "grad_norm": 1.9447376728057861, "learning_rate": 4.7404556968009814e-06, "loss": 0.5383, "step": 12668 }, { "epoch": 0.5310669335485155, "grad_norm": 2.1624178886413574, "learning_rate": 4.73977776211044e-06, "loss": 0.536, "step": 12669 }, { "epoch": 0.5311088521635245, "grad_norm": 1.9898768663406372, "learning_rate": 4.739099832216732e-06, "loss": 0.4867, "step": 12670 }, { "epoch": 0.5311507707785335, "grad_norm": 3.839714527130127, "learning_rate": 4.738421907132359e-06, "loss": 0.5117, "step": 12671 }, { "epoch": 0.5311926893935425, "grad_norm": 2.036914110183716, "learning_rate": 4.737743986869819e-06, "loss": 0.5446, "step": 12672 }, { "epoch": 0.5312346080085514, "grad_norm": 3.1993136405944824, "learning_rate": 4.737066071441604e-06, "loss": 0.5175, "step": 12673 }, { "epoch": 0.5312765266235604, "grad_norm": 1.832879900932312, "learning_rate": 4.736388160860211e-06, "loss": 0.4572, "step": 12674 }, { "epoch": 0.5313184452385693, "grad_norm": 1.6418389081954956, "learning_rate": 4.735710255138142e-06, "loss": 0.5341, "step": 12675 }, { "epoch": 0.5313603638535783, "grad_norm": 1.7454986572265625, "learning_rate": 4.735032354287884e-06, "loss": 0.5193, "step": 12676 }, { "epoch": 0.5314022824685872, "grad_norm": 1.9142879247665405, "learning_rate": 4.734354458321942e-06, "loss": 0.4891, "step": 12677 }, { "epoch": 0.5314442010835962, "grad_norm": 1.716832160949707, "learning_rate": 4.7336765672528075e-06, "loss": 0.5107, "step": 12678 }, { "epoch": 0.5314861196986052, "grad_norm": 2.1383321285247803, "learning_rate": 4.732998681092976e-06, "loss": 0.5188, "step": 12679 }, { "epoch": 0.5315280383136142, "grad_norm": 1.855259656906128, "learning_rate": 4.732320799854944e-06, "loss": 0.5482, "step": 12680 }, { "epoch": 0.5315699569286231, "grad_norm": 1.6634912490844727, "learning_rate": 4.7316429235512115e-06, "loss": 0.5519, "step": 12681 }, { "epoch": 0.531611875543632, "grad_norm": 1.6367944478988647, "learning_rate": 4.730965052194269e-06, "loss": 0.5059, "step": 12682 }, { "epoch": 0.531653794158641, "grad_norm": 1.9114995002746582, "learning_rate": 4.730287185796613e-06, "loss": 0.5308, "step": 12683 }, { "epoch": 0.5316957127736499, "grad_norm": 1.5994375944137573, "learning_rate": 4.729609324370742e-06, "loss": 0.5048, "step": 12684 }, { "epoch": 0.5317376313886589, "grad_norm": 1.9217491149902344, "learning_rate": 4.7289314679291465e-06, "loss": 0.5073, "step": 12685 }, { "epoch": 0.5317795500036678, "grad_norm": 1.8951562643051147, "learning_rate": 4.728253616484326e-06, "loss": 0.4752, "step": 12686 }, { "epoch": 0.5318214686186769, "grad_norm": 2.066423177719116, "learning_rate": 4.727575770048776e-06, "loss": 0.5124, "step": 12687 }, { "epoch": 0.5318633872336858, "grad_norm": 1.6331592798233032, "learning_rate": 4.726897928634987e-06, "loss": 0.4761, "step": 12688 }, { "epoch": 0.5319053058486948, "grad_norm": 1.5999336242675781, "learning_rate": 4.726220092255458e-06, "loss": 0.4714, "step": 12689 }, { "epoch": 0.5319472244637037, "grad_norm": 2.212569236755371, "learning_rate": 4.725542260922685e-06, "loss": 0.5247, "step": 12690 }, { "epoch": 0.5319891430787127, "grad_norm": 1.909672737121582, "learning_rate": 4.72486443464916e-06, "loss": 0.5782, "step": 12691 }, { "epoch": 0.5320310616937216, "grad_norm": 1.9078545570373535, "learning_rate": 4.724186613447377e-06, "loss": 0.4722, "step": 12692 }, { "epoch": 0.5320729803087306, "grad_norm": 1.858201503753662, "learning_rate": 4.723508797329835e-06, "loss": 0.5663, "step": 12693 }, { "epoch": 0.5321148989237395, "grad_norm": 1.6704161167144775, "learning_rate": 4.722830986309024e-06, "loss": 0.4523, "step": 12694 }, { "epoch": 0.5321568175387485, "grad_norm": 1.5513166189193726, "learning_rate": 4.722153180397442e-06, "loss": 0.4491, "step": 12695 }, { "epoch": 0.5321987361537575, "grad_norm": 1.6502189636230469, "learning_rate": 4.721475379607583e-06, "loss": 0.5011, "step": 12696 }, { "epoch": 0.5322406547687665, "grad_norm": 1.799518346786499, "learning_rate": 4.720797583951938e-06, "loss": 0.5207, "step": 12697 }, { "epoch": 0.5322825733837754, "grad_norm": 1.814866304397583, "learning_rate": 4.720119793443003e-06, "loss": 0.4835, "step": 12698 }, { "epoch": 0.5323244919987844, "grad_norm": 1.740046501159668, "learning_rate": 4.719442008093275e-06, "loss": 0.4912, "step": 12699 }, { "epoch": 0.5323664106137933, "grad_norm": 1.6937724351882935, "learning_rate": 4.718764227915246e-06, "loss": 0.4895, "step": 12700 }, { "epoch": 0.5324083292288023, "grad_norm": 1.8986163139343262, "learning_rate": 4.718086452921407e-06, "loss": 0.4866, "step": 12701 }, { "epoch": 0.5324502478438112, "grad_norm": 1.803285002708435, "learning_rate": 4.717408683124257e-06, "loss": 0.4562, "step": 12702 }, { "epoch": 0.5324921664588202, "grad_norm": 1.5410966873168945, "learning_rate": 4.716730918536286e-06, "loss": 0.5003, "step": 12703 }, { "epoch": 0.5325340850738292, "grad_norm": 1.7909470796585083, "learning_rate": 4.716053159169989e-06, "loss": 0.4796, "step": 12704 }, { "epoch": 0.5325760036888382, "grad_norm": 1.5829441547393799, "learning_rate": 4.715375405037862e-06, "loss": 0.4585, "step": 12705 }, { "epoch": 0.5326179223038471, "grad_norm": 1.664760947227478, "learning_rate": 4.714697656152392e-06, "loss": 0.5099, "step": 12706 }, { "epoch": 0.532659840918856, "grad_norm": 2.158926248550415, "learning_rate": 4.714019912526077e-06, "loss": 0.5471, "step": 12707 }, { "epoch": 0.532701759533865, "grad_norm": 1.8244550228118896, "learning_rate": 4.713342174171412e-06, "loss": 0.5237, "step": 12708 }, { "epoch": 0.5327436781488739, "grad_norm": 1.8177860975265503, "learning_rate": 4.712664441100886e-06, "loss": 0.5357, "step": 12709 }, { "epoch": 0.5327855967638829, "grad_norm": 2.1730175018310547, "learning_rate": 4.711986713326993e-06, "loss": 0.4563, "step": 12710 }, { "epoch": 0.5328275153788918, "grad_norm": 1.5460385084152222, "learning_rate": 4.711308990862229e-06, "loss": 0.4909, "step": 12711 }, { "epoch": 0.5328694339939009, "grad_norm": 2.2355082035064697, "learning_rate": 4.710631273719082e-06, "loss": 0.5418, "step": 12712 }, { "epoch": 0.5329113526089098, "grad_norm": 1.7444993257522583, "learning_rate": 4.70995356191005e-06, "loss": 0.4484, "step": 12713 }, { "epoch": 0.5329532712239188, "grad_norm": 1.9111939668655396, "learning_rate": 4.7092758554476215e-06, "loss": 0.4767, "step": 12714 }, { "epoch": 0.5329951898389277, "grad_norm": 1.6319984197616577, "learning_rate": 4.708598154344292e-06, "loss": 0.492, "step": 12715 }, { "epoch": 0.5330371084539367, "grad_norm": 1.5269503593444824, "learning_rate": 4.70792045861255e-06, "loss": 0.508, "step": 12716 }, { "epoch": 0.5330790270689456, "grad_norm": 3.5350611209869385, "learning_rate": 4.707242768264893e-06, "loss": 0.5264, "step": 12717 }, { "epoch": 0.5331209456839546, "grad_norm": 1.7311651706695557, "learning_rate": 4.70656508331381e-06, "loss": 0.4567, "step": 12718 }, { "epoch": 0.5331628642989635, "grad_norm": 1.9715781211853027, "learning_rate": 4.705887403771795e-06, "loss": 0.4576, "step": 12719 }, { "epoch": 0.5332047829139726, "grad_norm": 1.9719505310058594, "learning_rate": 4.705209729651339e-06, "loss": 0.5492, "step": 12720 }, { "epoch": 0.5332467015289815, "grad_norm": 6.405797958374023, "learning_rate": 4.704532060964933e-06, "loss": 0.4862, "step": 12721 }, { "epoch": 0.5332886201439905, "grad_norm": 1.7673074007034302, "learning_rate": 4.703854397725069e-06, "loss": 0.4943, "step": 12722 }, { "epoch": 0.5333305387589994, "grad_norm": 1.8643920421600342, "learning_rate": 4.703176739944242e-06, "loss": 0.466, "step": 12723 }, { "epoch": 0.5333724573740084, "grad_norm": 1.8102591037750244, "learning_rate": 4.702499087634941e-06, "loss": 0.484, "step": 12724 }, { "epoch": 0.5334143759890173, "grad_norm": 2.077955722808838, "learning_rate": 4.701821440809657e-06, "loss": 0.5189, "step": 12725 }, { "epoch": 0.5334562946040263, "grad_norm": 1.9854736328125, "learning_rate": 4.701143799480884e-06, "loss": 0.4979, "step": 12726 }, { "epoch": 0.5334982132190352, "grad_norm": 2.065463066101074, "learning_rate": 4.70046616366111e-06, "loss": 0.5314, "step": 12727 }, { "epoch": 0.5335401318340443, "grad_norm": 1.9047541618347168, "learning_rate": 4.69978853336283e-06, "loss": 0.4334, "step": 12728 }, { "epoch": 0.5335820504490532, "grad_norm": 2.537848949432373, "learning_rate": 4.699110908598534e-06, "loss": 0.4864, "step": 12729 }, { "epoch": 0.5336239690640622, "grad_norm": 3.8078534603118896, "learning_rate": 4.69843328938071e-06, "loss": 0.5259, "step": 12730 }, { "epoch": 0.5336658876790711, "grad_norm": 1.614353895187378, "learning_rate": 4.697755675721852e-06, "loss": 0.4916, "step": 12731 }, { "epoch": 0.53370780629408, "grad_norm": 1.7659056186676025, "learning_rate": 4.697078067634452e-06, "loss": 0.5418, "step": 12732 }, { "epoch": 0.533749724909089, "grad_norm": 1.910522222518921, "learning_rate": 4.696400465130998e-06, "loss": 0.4819, "step": 12733 }, { "epoch": 0.5337916435240979, "grad_norm": 1.8561123609542847, "learning_rate": 4.69572286822398e-06, "loss": 0.4508, "step": 12734 }, { "epoch": 0.5338335621391069, "grad_norm": 2.1983394622802734, "learning_rate": 4.695045276925893e-06, "loss": 0.5179, "step": 12735 }, { "epoch": 0.5338754807541158, "grad_norm": 1.6119712591171265, "learning_rate": 4.694367691249223e-06, "loss": 0.5292, "step": 12736 }, { "epoch": 0.5339173993691249, "grad_norm": 1.5967730283737183, "learning_rate": 4.693690111206463e-06, "loss": 0.4856, "step": 12737 }, { "epoch": 0.5339593179841338, "grad_norm": 2.1889476776123047, "learning_rate": 4.693012536810103e-06, "loss": 0.5359, "step": 12738 }, { "epoch": 0.5340012365991428, "grad_norm": 1.5859112739562988, "learning_rate": 4.69233496807263e-06, "loss": 0.4583, "step": 12739 }, { "epoch": 0.5340431552141517, "grad_norm": 1.4971424341201782, "learning_rate": 4.69165740500654e-06, "loss": 0.4902, "step": 12740 }, { "epoch": 0.5340850738291607, "grad_norm": 1.530650019645691, "learning_rate": 4.6909798476243166e-06, "loss": 0.5015, "step": 12741 }, { "epoch": 0.5341269924441696, "grad_norm": 1.6434420347213745, "learning_rate": 4.690302295938453e-06, "loss": 0.4905, "step": 12742 }, { "epoch": 0.5341689110591786, "grad_norm": 1.8914742469787598, "learning_rate": 4.68962474996144e-06, "loss": 0.5198, "step": 12743 }, { "epoch": 0.5342108296741875, "grad_norm": 1.9851233959197998, "learning_rate": 4.688947209705763e-06, "loss": 0.4997, "step": 12744 }, { "epoch": 0.5342527482891966, "grad_norm": 1.7485241889953613, "learning_rate": 4.688269675183915e-06, "loss": 0.5297, "step": 12745 }, { "epoch": 0.5342946669042055, "grad_norm": 1.7181259393692017, "learning_rate": 4.687592146408386e-06, "loss": 0.5306, "step": 12746 }, { "epoch": 0.5343365855192145, "grad_norm": 1.7445200681686401, "learning_rate": 4.686914623391663e-06, "loss": 0.4614, "step": 12747 }, { "epoch": 0.5343785041342234, "grad_norm": 1.7712842226028442, "learning_rate": 4.686237106146236e-06, "loss": 0.4948, "step": 12748 }, { "epoch": 0.5344204227492324, "grad_norm": 1.7162445783615112, "learning_rate": 4.685559594684595e-06, "loss": 0.5303, "step": 12749 }, { "epoch": 0.5344623413642413, "grad_norm": 1.8048752546310425, "learning_rate": 4.684882089019227e-06, "loss": 0.5084, "step": 12750 }, { "epoch": 0.5345042599792503, "grad_norm": 1.7685942649841309, "learning_rate": 4.684204589162622e-06, "loss": 0.5168, "step": 12751 }, { "epoch": 0.5345461785942592, "grad_norm": 1.871564269065857, "learning_rate": 4.683527095127271e-06, "loss": 0.4607, "step": 12752 }, { "epoch": 0.5345880972092683, "grad_norm": 1.6633621454238892, "learning_rate": 4.682849606925659e-06, "loss": 0.5617, "step": 12753 }, { "epoch": 0.5346300158242772, "grad_norm": 1.6270294189453125, "learning_rate": 4.682172124570274e-06, "loss": 0.452, "step": 12754 }, { "epoch": 0.5346719344392862, "grad_norm": 1.8367217779159546, "learning_rate": 4.68149464807361e-06, "loss": 0.4999, "step": 12755 }, { "epoch": 0.5347138530542951, "grad_norm": 2.415727376937866, "learning_rate": 4.680817177448149e-06, "loss": 0.4711, "step": 12756 }, { "epoch": 0.534755771669304, "grad_norm": 1.767454743385315, "learning_rate": 4.6801397127063835e-06, "loss": 0.5095, "step": 12757 }, { "epoch": 0.534797690284313, "grad_norm": 1.633158802986145, "learning_rate": 4.6794622538608015e-06, "loss": 0.4592, "step": 12758 }, { "epoch": 0.5348396088993219, "grad_norm": 1.6037641763687134, "learning_rate": 4.6787848009238875e-06, "loss": 0.4156, "step": 12759 }, { "epoch": 0.5348815275143309, "grad_norm": 2.0646603107452393, "learning_rate": 4.678107353908132e-06, "loss": 0.5082, "step": 12760 }, { "epoch": 0.5349234461293398, "grad_norm": 1.5599030256271362, "learning_rate": 4.6774299128260235e-06, "loss": 0.4636, "step": 12761 }, { "epoch": 0.5349653647443489, "grad_norm": 2.262355089187622, "learning_rate": 4.676752477690048e-06, "loss": 0.4381, "step": 12762 }, { "epoch": 0.5350072833593578, "grad_norm": 2.0949795246124268, "learning_rate": 4.676075048512693e-06, "loss": 0.4594, "step": 12763 }, { "epoch": 0.5350492019743668, "grad_norm": 1.8617781400680542, "learning_rate": 4.675397625306448e-06, "loss": 0.4909, "step": 12764 }, { "epoch": 0.5350911205893757, "grad_norm": 2.603468894958496, "learning_rate": 4.674720208083798e-06, "loss": 0.5628, "step": 12765 }, { "epoch": 0.5351330392043847, "grad_norm": 1.6265487670898438, "learning_rate": 4.6740427968572315e-06, "loss": 0.4721, "step": 12766 }, { "epoch": 0.5351749578193936, "grad_norm": 1.760691523551941, "learning_rate": 4.673365391639237e-06, "loss": 0.4667, "step": 12767 }, { "epoch": 0.5352168764344026, "grad_norm": 1.6049309968948364, "learning_rate": 4.6726879924422975e-06, "loss": 0.4753, "step": 12768 }, { "epoch": 0.5352587950494115, "grad_norm": 1.7180439233779907, "learning_rate": 4.672010599278902e-06, "loss": 0.5667, "step": 12769 }, { "epoch": 0.5353007136644206, "grad_norm": 1.6637927293777466, "learning_rate": 4.6713332121615405e-06, "loss": 0.5177, "step": 12770 }, { "epoch": 0.5353426322794295, "grad_norm": 1.9032528400421143, "learning_rate": 4.6706558311026955e-06, "loss": 0.4984, "step": 12771 }, { "epoch": 0.5353845508944385, "grad_norm": 1.826474666595459, "learning_rate": 4.669978456114853e-06, "loss": 0.5302, "step": 12772 }, { "epoch": 0.5354264695094474, "grad_norm": 1.6971542835235596, "learning_rate": 4.669301087210505e-06, "loss": 0.5118, "step": 12773 }, { "epoch": 0.5354683881244564, "grad_norm": 2.0806987285614014, "learning_rate": 4.668623724402132e-06, "loss": 0.5153, "step": 12774 }, { "epoch": 0.5355103067394653, "grad_norm": 2.2972891330718994, "learning_rate": 4.667946367702223e-06, "loss": 0.4814, "step": 12775 }, { "epoch": 0.5355522253544743, "grad_norm": 2.042593479156494, "learning_rate": 4.667269017123265e-06, "loss": 0.4802, "step": 12776 }, { "epoch": 0.5355941439694832, "grad_norm": 1.7910605669021606, "learning_rate": 4.6665916726777405e-06, "loss": 0.5061, "step": 12777 }, { "epoch": 0.5356360625844923, "grad_norm": 1.793792486190796, "learning_rate": 4.665914334378138e-06, "loss": 0.4927, "step": 12778 }, { "epoch": 0.5356779811995012, "grad_norm": 1.901161551475525, "learning_rate": 4.665237002236944e-06, "loss": 0.5692, "step": 12779 }, { "epoch": 0.5357198998145102, "grad_norm": 1.8137530088424683, "learning_rate": 4.664559676266643e-06, "loss": 0.5051, "step": 12780 }, { "epoch": 0.5357618184295191, "grad_norm": 1.8910043239593506, "learning_rate": 4.663882356479718e-06, "loss": 0.4792, "step": 12781 }, { "epoch": 0.535803737044528, "grad_norm": 1.940076470375061, "learning_rate": 4.663205042888661e-06, "loss": 0.4556, "step": 12782 }, { "epoch": 0.535845655659537, "grad_norm": 1.634446620941162, "learning_rate": 4.662527735505952e-06, "loss": 0.4862, "step": 12783 }, { "epoch": 0.5358875742745459, "grad_norm": 1.8231931924819946, "learning_rate": 4.661850434344076e-06, "loss": 0.5583, "step": 12784 }, { "epoch": 0.5359294928895549, "grad_norm": 1.838924527168274, "learning_rate": 4.661173139415523e-06, "loss": 0.4684, "step": 12785 }, { "epoch": 0.5359714115045638, "grad_norm": 1.782175064086914, "learning_rate": 4.660495850732771e-06, "loss": 0.4362, "step": 12786 }, { "epoch": 0.5360133301195729, "grad_norm": 2.059556484222412, "learning_rate": 4.65981856830831e-06, "loss": 0.5142, "step": 12787 }, { "epoch": 0.5360552487345818, "grad_norm": 1.8216253519058228, "learning_rate": 4.659141292154625e-06, "loss": 0.5598, "step": 12788 }, { "epoch": 0.5360971673495908, "grad_norm": 3.5879640579223633, "learning_rate": 4.658464022284198e-06, "loss": 0.4724, "step": 12789 }, { "epoch": 0.5361390859645997, "grad_norm": 1.6773210763931274, "learning_rate": 4.657786758709514e-06, "loss": 0.5183, "step": 12790 }, { "epoch": 0.5361810045796087, "grad_norm": 1.7120766639709473, "learning_rate": 4.6571095014430595e-06, "loss": 0.4771, "step": 12791 }, { "epoch": 0.5362229231946176, "grad_norm": 2.159636974334717, "learning_rate": 4.656432250497315e-06, "loss": 0.5727, "step": 12792 }, { "epoch": 0.5362648418096266, "grad_norm": 1.6535212993621826, "learning_rate": 4.655755005884769e-06, "loss": 0.5398, "step": 12793 }, { "epoch": 0.5363067604246355, "grad_norm": 1.8051515817642212, "learning_rate": 4.655077767617904e-06, "loss": 0.5387, "step": 12794 }, { "epoch": 0.5363486790396446, "grad_norm": 1.820841670036316, "learning_rate": 4.654400535709201e-06, "loss": 0.5441, "step": 12795 }, { "epoch": 0.5363905976546535, "grad_norm": 2.055495023727417, "learning_rate": 4.653723310171147e-06, "loss": 0.4971, "step": 12796 }, { "epoch": 0.5364325162696625, "grad_norm": 1.9306347370147705, "learning_rate": 4.653046091016226e-06, "loss": 0.4103, "step": 12797 }, { "epoch": 0.5364744348846714, "grad_norm": 2.0654091835021973, "learning_rate": 4.652368878256922e-06, "loss": 0.4988, "step": 12798 }, { "epoch": 0.5365163534996804, "grad_norm": 1.719234585762024, "learning_rate": 4.651691671905714e-06, "loss": 0.4654, "step": 12799 }, { "epoch": 0.5365582721146893, "grad_norm": 1.686360239982605, "learning_rate": 4.651014471975091e-06, "loss": 0.4642, "step": 12800 }, { "epoch": 0.5366001907296983, "grad_norm": 2.009342670440674, "learning_rate": 4.650337278477533e-06, "loss": 0.5317, "step": 12801 }, { "epoch": 0.5366421093447072, "grad_norm": 1.8301103115081787, "learning_rate": 4.649660091425523e-06, "loss": 0.558, "step": 12802 }, { "epoch": 0.5366840279597163, "grad_norm": 1.9428837299346924, "learning_rate": 4.648982910831546e-06, "loss": 0.5545, "step": 12803 }, { "epoch": 0.5367259465747252, "grad_norm": 2.6628172397613525, "learning_rate": 4.648305736708084e-06, "loss": 0.4882, "step": 12804 }, { "epoch": 0.5367678651897342, "grad_norm": 1.7967579364776611, "learning_rate": 4.647628569067618e-06, "loss": 0.4941, "step": 12805 }, { "epoch": 0.5368097838047431, "grad_norm": 1.7292214632034302, "learning_rate": 4.646951407922634e-06, "loss": 0.4861, "step": 12806 }, { "epoch": 0.536851702419752, "grad_norm": 2.8450005054473877, "learning_rate": 4.646274253285611e-06, "loss": 0.511, "step": 12807 }, { "epoch": 0.536893621034761, "grad_norm": 1.769261121749878, "learning_rate": 4.645597105169034e-06, "loss": 0.4666, "step": 12808 }, { "epoch": 0.5369355396497699, "grad_norm": 2.0250916481018066, "learning_rate": 4.644919963585385e-06, "loss": 0.5136, "step": 12809 }, { "epoch": 0.5369774582647789, "grad_norm": 1.9886726140975952, "learning_rate": 4.644242828547144e-06, "loss": 0.472, "step": 12810 }, { "epoch": 0.5370193768797878, "grad_norm": 1.9622900485992432, "learning_rate": 4.6435657000667974e-06, "loss": 0.4782, "step": 12811 }, { "epoch": 0.5370612954947969, "grad_norm": 2.210176467895508, "learning_rate": 4.642888578156821e-06, "loss": 0.4788, "step": 12812 }, { "epoch": 0.5371032141098058, "grad_norm": 1.882601022720337, "learning_rate": 4.642211462829702e-06, "loss": 0.538, "step": 12813 }, { "epoch": 0.5371451327248148, "grad_norm": 1.5648918151855469, "learning_rate": 4.641534354097921e-06, "loss": 0.4845, "step": 12814 }, { "epoch": 0.5371870513398237, "grad_norm": 2.35758376121521, "learning_rate": 4.6408572519739565e-06, "loss": 0.4878, "step": 12815 }, { "epoch": 0.5372289699548327, "grad_norm": 1.7056413888931274, "learning_rate": 4.640180156470292e-06, "loss": 0.4659, "step": 12816 }, { "epoch": 0.5372708885698416, "grad_norm": 2.4939045906066895, "learning_rate": 4.63950306759941e-06, "loss": 0.4935, "step": 12817 }, { "epoch": 0.5373128071848506, "grad_norm": 1.8650649785995483, "learning_rate": 4.638825985373791e-06, "loss": 0.5287, "step": 12818 }, { "epoch": 0.5373547257998595, "grad_norm": 1.5335928201675415, "learning_rate": 4.638148909805914e-06, "loss": 0.5352, "step": 12819 }, { "epoch": 0.5373966444148686, "grad_norm": 1.6812245845794678, "learning_rate": 4.637471840908264e-06, "loss": 0.4722, "step": 12820 }, { "epoch": 0.5374385630298775, "grad_norm": 1.7610688209533691, "learning_rate": 4.6367947786933165e-06, "loss": 0.4477, "step": 12821 }, { "epoch": 0.5374804816448865, "grad_norm": 2.2609801292419434, "learning_rate": 4.636117723173557e-06, "loss": 0.4977, "step": 12822 }, { "epoch": 0.5375224002598954, "grad_norm": 2.0048751831054688, "learning_rate": 4.635440674361464e-06, "loss": 0.4974, "step": 12823 }, { "epoch": 0.5375643188749044, "grad_norm": 3.599680185317993, "learning_rate": 4.634763632269518e-06, "loss": 0.5446, "step": 12824 }, { "epoch": 0.5376062374899133, "grad_norm": 1.5821173191070557, "learning_rate": 4.634086596910198e-06, "loss": 0.4805, "step": 12825 }, { "epoch": 0.5376481561049223, "grad_norm": 1.7210606336593628, "learning_rate": 4.633409568295988e-06, "loss": 0.527, "step": 12826 }, { "epoch": 0.5376900747199312, "grad_norm": 2.0484771728515625, "learning_rate": 4.632732546439366e-06, "loss": 0.4819, "step": 12827 }, { "epoch": 0.5377319933349403, "grad_norm": 1.8031350374221802, "learning_rate": 4.632055531352809e-06, "loss": 0.5187, "step": 12828 }, { "epoch": 0.5377739119499492, "grad_norm": 1.7482539415359497, "learning_rate": 4.631378523048803e-06, "loss": 0.5408, "step": 12829 }, { "epoch": 0.5378158305649582, "grad_norm": 1.7878692150115967, "learning_rate": 4.630701521539822e-06, "loss": 0.5069, "step": 12830 }, { "epoch": 0.5378577491799671, "grad_norm": 1.5741056203842163, "learning_rate": 4.630024526838348e-06, "loss": 0.4636, "step": 12831 }, { "epoch": 0.537899667794976, "grad_norm": 2.1007819175720215, "learning_rate": 4.629347538956863e-06, "loss": 0.4865, "step": 12832 }, { "epoch": 0.537941586409985, "grad_norm": 1.7160229682922363, "learning_rate": 4.628670557907841e-06, "loss": 0.4697, "step": 12833 }, { "epoch": 0.5379835050249939, "grad_norm": 1.9413058757781982, "learning_rate": 4.627993583703763e-06, "loss": 0.4654, "step": 12834 }, { "epoch": 0.5380254236400029, "grad_norm": 1.9841485023498535, "learning_rate": 4.627316616357112e-06, "loss": 0.4892, "step": 12835 }, { "epoch": 0.5380673422550118, "grad_norm": 1.472414493560791, "learning_rate": 4.626639655880363e-06, "loss": 0.4372, "step": 12836 }, { "epoch": 0.5381092608700209, "grad_norm": 2.030273675918579, "learning_rate": 4.625962702285995e-06, "loss": 0.4452, "step": 12837 }, { "epoch": 0.5381511794850298, "grad_norm": 1.7422186136245728, "learning_rate": 4.625285755586489e-06, "loss": 0.5128, "step": 12838 }, { "epoch": 0.5381930981000388, "grad_norm": 1.7312772274017334, "learning_rate": 4.62460881579432e-06, "loss": 0.4885, "step": 12839 }, { "epoch": 0.5382350167150477, "grad_norm": 1.8147964477539062, "learning_rate": 4.623931882921969e-06, "loss": 0.5017, "step": 12840 }, { "epoch": 0.5382769353300567, "grad_norm": 2.260652780532837, "learning_rate": 4.623254956981916e-06, "loss": 0.5024, "step": 12841 }, { "epoch": 0.5383188539450656, "grad_norm": 1.7451720237731934, "learning_rate": 4.6225780379866336e-06, "loss": 0.4863, "step": 12842 }, { "epoch": 0.5383607725600746, "grad_norm": 1.645328402519226, "learning_rate": 4.621901125948604e-06, "loss": 0.5242, "step": 12843 }, { "epoch": 0.5384026911750835, "grad_norm": 1.824460744857788, "learning_rate": 4.621224220880307e-06, "loss": 0.5146, "step": 12844 }, { "epoch": 0.5384446097900926, "grad_norm": 1.6159292459487915, "learning_rate": 4.620547322794217e-06, "loss": 0.5009, "step": 12845 }, { "epoch": 0.5384865284051015, "grad_norm": 3.6669912338256836, "learning_rate": 4.619870431702811e-06, "loss": 0.5258, "step": 12846 }, { "epoch": 0.5385284470201105, "grad_norm": 1.668001651763916, "learning_rate": 4.61919354761857e-06, "loss": 0.4907, "step": 12847 }, { "epoch": 0.5385703656351194, "grad_norm": 1.6472933292388916, "learning_rate": 4.618516670553967e-06, "loss": 0.4528, "step": 12848 }, { "epoch": 0.5386122842501284, "grad_norm": 1.7158808708190918, "learning_rate": 4.617839800521483e-06, "loss": 0.4597, "step": 12849 }, { "epoch": 0.5386542028651373, "grad_norm": 1.7169950008392334, "learning_rate": 4.6171629375335945e-06, "loss": 0.5224, "step": 12850 }, { "epoch": 0.5386961214801463, "grad_norm": 2.0593838691711426, "learning_rate": 4.616486081602778e-06, "loss": 0.5376, "step": 12851 }, { "epoch": 0.5387380400951552, "grad_norm": 1.7607688903808594, "learning_rate": 4.615809232741509e-06, "loss": 0.5085, "step": 12852 }, { "epoch": 0.5387799587101643, "grad_norm": 1.5394073724746704, "learning_rate": 4.6151323909622684e-06, "loss": 0.4561, "step": 12853 }, { "epoch": 0.5388218773251732, "grad_norm": 2.2856061458587646, "learning_rate": 4.6144555562775276e-06, "loss": 0.5025, "step": 12854 }, { "epoch": 0.5388637959401822, "grad_norm": 1.7679768800735474, "learning_rate": 4.613778728699766e-06, "loss": 0.4783, "step": 12855 }, { "epoch": 0.5389057145551911, "grad_norm": 1.816490888595581, "learning_rate": 4.613101908241462e-06, "loss": 0.4439, "step": 12856 }, { "epoch": 0.5389476331702, "grad_norm": 1.593259572982788, "learning_rate": 4.612425094915086e-06, "loss": 0.5269, "step": 12857 }, { "epoch": 0.538989551785209, "grad_norm": 1.7663521766662598, "learning_rate": 4.611748288733118e-06, "loss": 0.4633, "step": 12858 }, { "epoch": 0.5390314704002179, "grad_norm": 1.6238657236099243, "learning_rate": 4.611071489708036e-06, "loss": 0.4612, "step": 12859 }, { "epoch": 0.539073389015227, "grad_norm": 1.6244938373565674, "learning_rate": 4.610394697852312e-06, "loss": 0.5128, "step": 12860 }, { "epoch": 0.5391153076302359, "grad_norm": 1.6844741106033325, "learning_rate": 4.609717913178422e-06, "loss": 0.4409, "step": 12861 }, { "epoch": 0.5391572262452449, "grad_norm": 1.5963679552078247, "learning_rate": 4.6090411356988455e-06, "loss": 0.476, "step": 12862 }, { "epoch": 0.5391991448602538, "grad_norm": 1.6475402116775513, "learning_rate": 4.6083643654260525e-06, "loss": 0.5294, "step": 12863 }, { "epoch": 0.5392410634752628, "grad_norm": 2.540869951248169, "learning_rate": 4.607687602372522e-06, "loss": 0.5168, "step": 12864 }, { "epoch": 0.5392829820902717, "grad_norm": 1.678884744644165, "learning_rate": 4.60701084655073e-06, "loss": 0.487, "step": 12865 }, { "epoch": 0.5393249007052807, "grad_norm": 1.6617439985275269, "learning_rate": 4.606334097973147e-06, "loss": 0.4931, "step": 12866 }, { "epoch": 0.5393668193202896, "grad_norm": 1.749139666557312, "learning_rate": 4.605657356652251e-06, "loss": 0.4418, "step": 12867 }, { "epoch": 0.5394087379352986, "grad_norm": 1.6484413146972656, "learning_rate": 4.604980622600518e-06, "loss": 0.4797, "step": 12868 }, { "epoch": 0.5394506565503076, "grad_norm": 1.853072166442871, "learning_rate": 4.6043038958304205e-06, "loss": 0.5914, "step": 12869 }, { "epoch": 0.5394925751653166, "grad_norm": 1.96409010887146, "learning_rate": 4.603627176354432e-06, "loss": 0.5279, "step": 12870 }, { "epoch": 0.5395344937803255, "grad_norm": 1.9786708354949951, "learning_rate": 4.6029504641850315e-06, "loss": 0.5251, "step": 12871 }, { "epoch": 0.5395764123953345, "grad_norm": 1.739968180656433, "learning_rate": 4.6022737593346875e-06, "loss": 0.5644, "step": 12872 }, { "epoch": 0.5396183310103434, "grad_norm": 1.794335126876831, "learning_rate": 4.601597061815879e-06, "loss": 0.4775, "step": 12873 }, { "epoch": 0.5396602496253524, "grad_norm": 4.996282577514648, "learning_rate": 4.6009203716410775e-06, "loss": 0.5378, "step": 12874 }, { "epoch": 0.5397021682403613, "grad_norm": 1.4619933366775513, "learning_rate": 4.600243688822756e-06, "loss": 0.4944, "step": 12875 }, { "epoch": 0.5397440868553703, "grad_norm": 1.9512832164764404, "learning_rate": 4.5995670133733896e-06, "loss": 0.5479, "step": 12876 }, { "epoch": 0.5397860054703792, "grad_norm": 1.9043720960617065, "learning_rate": 4.598890345305454e-06, "loss": 0.5337, "step": 12877 }, { "epoch": 0.5398279240853883, "grad_norm": 1.7369683980941772, "learning_rate": 4.5982136846314185e-06, "loss": 0.4978, "step": 12878 }, { "epoch": 0.5398698427003972, "grad_norm": 1.7493603229522705, "learning_rate": 4.597537031363757e-06, "loss": 0.4739, "step": 12879 }, { "epoch": 0.5399117613154062, "grad_norm": 2.06640625, "learning_rate": 4.596860385514947e-06, "loss": 0.4497, "step": 12880 }, { "epoch": 0.5399536799304151, "grad_norm": 2.100797176361084, "learning_rate": 4.596183747097456e-06, "loss": 0.4812, "step": 12881 }, { "epoch": 0.539995598545424, "grad_norm": 1.6242198944091797, "learning_rate": 4.59550711612376e-06, "loss": 0.4913, "step": 12882 }, { "epoch": 0.540037517160433, "grad_norm": 1.8042134046554565, "learning_rate": 4.594830492606333e-06, "loss": 0.5025, "step": 12883 }, { "epoch": 0.5400794357754419, "grad_norm": 1.936673641204834, "learning_rate": 4.594153876557643e-06, "loss": 0.4936, "step": 12884 }, { "epoch": 0.540121354390451, "grad_norm": 1.9241021871566772, "learning_rate": 4.593477267990168e-06, "loss": 0.5181, "step": 12885 }, { "epoch": 0.5401632730054599, "grad_norm": 2.1344494819641113, "learning_rate": 4.592800666916375e-06, "loss": 0.5096, "step": 12886 }, { "epoch": 0.5402051916204689, "grad_norm": 2.253939151763916, "learning_rate": 4.59212407334874e-06, "loss": 0.4613, "step": 12887 }, { "epoch": 0.5402471102354778, "grad_norm": 1.5918670892715454, "learning_rate": 4.591447487299734e-06, "loss": 0.443, "step": 12888 }, { "epoch": 0.5402890288504868, "grad_norm": 2.0701358318328857, "learning_rate": 4.590770908781828e-06, "loss": 0.5917, "step": 12889 }, { "epoch": 0.5403309474654957, "grad_norm": 1.5540282726287842, "learning_rate": 4.590094337807494e-06, "loss": 0.4693, "step": 12890 }, { "epoch": 0.5403728660805047, "grad_norm": 2.0235164165496826, "learning_rate": 4.589417774389206e-06, "loss": 0.4637, "step": 12891 }, { "epoch": 0.5404147846955136, "grad_norm": 1.8606069087982178, "learning_rate": 4.588741218539433e-06, "loss": 0.4626, "step": 12892 }, { "epoch": 0.5404567033105226, "grad_norm": 1.6747961044311523, "learning_rate": 4.588064670270646e-06, "loss": 0.4982, "step": 12893 }, { "epoch": 0.5404986219255316, "grad_norm": 1.7194994688034058, "learning_rate": 4.587388129595318e-06, "loss": 0.497, "step": 12894 }, { "epoch": 0.5405405405405406, "grad_norm": 1.9990196228027344, "learning_rate": 4.58671159652592e-06, "loss": 0.5238, "step": 12895 }, { "epoch": 0.5405824591555495, "grad_norm": 1.892733097076416, "learning_rate": 4.586035071074922e-06, "loss": 0.5593, "step": 12896 }, { "epoch": 0.5406243777705585, "grad_norm": 1.6705303192138672, "learning_rate": 4.585358553254795e-06, "loss": 0.4432, "step": 12897 }, { "epoch": 0.5406662963855674, "grad_norm": 1.7241209745407104, "learning_rate": 4.58468204307801e-06, "loss": 0.5017, "step": 12898 }, { "epoch": 0.5407082150005764, "grad_norm": 2.344217300415039, "learning_rate": 4.584005540557037e-06, "loss": 0.4924, "step": 12899 }, { "epoch": 0.5407501336155853, "grad_norm": 2.3869221210479736, "learning_rate": 4.583329045704348e-06, "loss": 0.5465, "step": 12900 }, { "epoch": 0.5407920522305943, "grad_norm": 2.002117156982422, "learning_rate": 4.58265255853241e-06, "loss": 0.4809, "step": 12901 }, { "epoch": 0.5408339708456033, "grad_norm": 1.8843815326690674, "learning_rate": 4.581976079053696e-06, "loss": 0.5401, "step": 12902 }, { "epoch": 0.5408758894606123, "grad_norm": 1.997326135635376, "learning_rate": 4.581299607280677e-06, "loss": 0.5281, "step": 12903 }, { "epoch": 0.5409178080756212, "grad_norm": 1.792761206626892, "learning_rate": 4.580623143225818e-06, "loss": 0.4991, "step": 12904 }, { "epoch": 0.5409597266906302, "grad_norm": 1.9335626363754272, "learning_rate": 4.579946686901592e-06, "loss": 0.4716, "step": 12905 }, { "epoch": 0.5410016453056391, "grad_norm": 1.813483476638794, "learning_rate": 4.57927023832047e-06, "loss": 0.4751, "step": 12906 }, { "epoch": 0.541043563920648, "grad_norm": 1.6913741827011108, "learning_rate": 4.578593797494919e-06, "loss": 0.4904, "step": 12907 }, { "epoch": 0.541085482535657, "grad_norm": 1.8313857316970825, "learning_rate": 4.577917364437407e-06, "loss": 0.5015, "step": 12908 }, { "epoch": 0.5411274011506659, "grad_norm": 1.892490029335022, "learning_rate": 4.577240939160407e-06, "loss": 0.5147, "step": 12909 }, { "epoch": 0.541169319765675, "grad_norm": 1.5919691324234009, "learning_rate": 4.5765645216763845e-06, "loss": 0.4973, "step": 12910 }, { "epoch": 0.5412112383806839, "grad_norm": 1.957202434539795, "learning_rate": 4.57588811199781e-06, "loss": 0.5119, "step": 12911 }, { "epoch": 0.5412531569956929, "grad_norm": 1.8421610593795776, "learning_rate": 4.575211710137153e-06, "loss": 0.4981, "step": 12912 }, { "epoch": 0.5412950756107018, "grad_norm": 3.487422227859497, "learning_rate": 4.574535316106879e-06, "loss": 0.4445, "step": 12913 }, { "epoch": 0.5413369942257108, "grad_norm": 2.0195751190185547, "learning_rate": 4.5738589299194584e-06, "loss": 0.5307, "step": 12914 }, { "epoch": 0.5413789128407197, "grad_norm": 1.6348907947540283, "learning_rate": 4.573182551587361e-06, "loss": 0.4577, "step": 12915 }, { "epoch": 0.5414208314557287, "grad_norm": 2.205078601837158, "learning_rate": 4.572506181123052e-06, "loss": 0.4993, "step": 12916 }, { "epoch": 0.5414627500707376, "grad_norm": 2.031698226928711, "learning_rate": 4.571829818539e-06, "loss": 0.5027, "step": 12917 }, { "epoch": 0.5415046686857466, "grad_norm": 1.923401117324829, "learning_rate": 4.571153463847675e-06, "loss": 0.4918, "step": 12918 }, { "epoch": 0.5415465873007556, "grad_norm": 1.9690040349960327, "learning_rate": 4.570477117061542e-06, "loss": 0.5027, "step": 12919 }, { "epoch": 0.5415885059157646, "grad_norm": 1.6820942163467407, "learning_rate": 4.569800778193069e-06, "loss": 0.4864, "step": 12920 }, { "epoch": 0.5416304245307735, "grad_norm": 1.6974989175796509, "learning_rate": 4.569124447254726e-06, "loss": 0.4904, "step": 12921 }, { "epoch": 0.5416723431457825, "grad_norm": 1.6712430715560913, "learning_rate": 4.568448124258976e-06, "loss": 0.5151, "step": 12922 }, { "epoch": 0.5417142617607914, "grad_norm": 2.3073389530181885, "learning_rate": 4.567771809218287e-06, "loss": 0.4799, "step": 12923 }, { "epoch": 0.5417561803758004, "grad_norm": 1.8395929336547852, "learning_rate": 4.567095502145131e-06, "loss": 0.4874, "step": 12924 }, { "epoch": 0.5417980989908093, "grad_norm": 1.9357229471206665, "learning_rate": 4.566419203051969e-06, "loss": 0.537, "step": 12925 }, { "epoch": 0.5418400176058183, "grad_norm": 1.9867080450057983, "learning_rate": 4.565742911951269e-06, "loss": 0.5383, "step": 12926 }, { "epoch": 0.5418819362208273, "grad_norm": 1.783508539199829, "learning_rate": 4.5650666288555e-06, "loss": 0.4644, "step": 12927 }, { "epoch": 0.5419238548358363, "grad_norm": 2.1877152919769287, "learning_rate": 4.564390353777125e-06, "loss": 0.5184, "step": 12928 }, { "epoch": 0.5419657734508452, "grad_norm": 1.7529791593551636, "learning_rate": 4.5637140867286115e-06, "loss": 0.5283, "step": 12929 }, { "epoch": 0.5420076920658542, "grad_norm": 1.764847755432129, "learning_rate": 4.563037827722427e-06, "loss": 0.4923, "step": 12930 }, { "epoch": 0.5420496106808631, "grad_norm": 1.8098790645599365, "learning_rate": 4.562361576771034e-06, "loss": 0.4724, "step": 12931 }, { "epoch": 0.542091529295872, "grad_norm": 1.7908499240875244, "learning_rate": 4.5616853338869e-06, "loss": 0.5218, "step": 12932 }, { "epoch": 0.542133447910881, "grad_norm": 2.1419363021850586, "learning_rate": 4.5610090990824935e-06, "loss": 0.4673, "step": 12933 }, { "epoch": 0.5421753665258899, "grad_norm": 1.7142037153244019, "learning_rate": 4.560332872370277e-06, "loss": 0.4344, "step": 12934 }, { "epoch": 0.542217285140899, "grad_norm": 2.060720920562744, "learning_rate": 4.559656653762713e-06, "loss": 0.4655, "step": 12935 }, { "epoch": 0.5422592037559079, "grad_norm": 1.7378185987472534, "learning_rate": 4.558980443272274e-06, "loss": 0.4822, "step": 12936 }, { "epoch": 0.5423011223709169, "grad_norm": 1.939782977104187, "learning_rate": 4.558304240911418e-06, "loss": 0.5175, "step": 12937 }, { "epoch": 0.5423430409859258, "grad_norm": 1.6345901489257812, "learning_rate": 4.557628046692613e-06, "loss": 0.4714, "step": 12938 }, { "epoch": 0.5423849596009348, "grad_norm": 2.043121814727783, "learning_rate": 4.556951860628326e-06, "loss": 0.517, "step": 12939 }, { "epoch": 0.5424268782159437, "grad_norm": 1.774446725845337, "learning_rate": 4.556275682731016e-06, "loss": 0.5504, "step": 12940 }, { "epoch": 0.5424687968309527, "grad_norm": 1.8185378313064575, "learning_rate": 4.5555995130131505e-06, "loss": 0.4817, "step": 12941 }, { "epoch": 0.5425107154459616, "grad_norm": 1.8300279378890991, "learning_rate": 4.554923351487196e-06, "loss": 0.5406, "step": 12942 }, { "epoch": 0.5425526340609707, "grad_norm": 2.4965925216674805, "learning_rate": 4.554247198165613e-06, "loss": 0.5019, "step": 12943 }, { "epoch": 0.5425945526759796, "grad_norm": 1.938368558883667, "learning_rate": 4.553571053060865e-06, "loss": 0.5526, "step": 12944 }, { "epoch": 0.5426364712909886, "grad_norm": 1.6755640506744385, "learning_rate": 4.55289491618542e-06, "loss": 0.4678, "step": 12945 }, { "epoch": 0.5426783899059975, "grad_norm": 3.132143974304199, "learning_rate": 4.552218787551737e-06, "loss": 0.5376, "step": 12946 }, { "epoch": 0.5427203085210065, "grad_norm": 1.7642320394515991, "learning_rate": 4.5515426671722825e-06, "loss": 0.4239, "step": 12947 }, { "epoch": 0.5427622271360154, "grad_norm": 3.086120128631592, "learning_rate": 4.550866555059519e-06, "loss": 0.4808, "step": 12948 }, { "epoch": 0.5428041457510244, "grad_norm": 3.445746421813965, "learning_rate": 4.55019045122591e-06, "loss": 0.4893, "step": 12949 }, { "epoch": 0.5428460643660333, "grad_norm": 1.6829335689544678, "learning_rate": 4.549514355683917e-06, "loss": 0.4622, "step": 12950 }, { "epoch": 0.5428879829810423, "grad_norm": 5.032920837402344, "learning_rate": 4.548838268446006e-06, "loss": 0.5689, "step": 12951 }, { "epoch": 0.5429299015960513, "grad_norm": 1.7101712226867676, "learning_rate": 4.548162189524636e-06, "loss": 0.5022, "step": 12952 }, { "epoch": 0.5429718202110603, "grad_norm": 1.8037936687469482, "learning_rate": 4.547486118932272e-06, "loss": 0.4358, "step": 12953 }, { "epoch": 0.5430137388260692, "grad_norm": 1.7306874990463257, "learning_rate": 4.546810056681377e-06, "loss": 0.5008, "step": 12954 }, { "epoch": 0.5430556574410782, "grad_norm": 1.8294011354446411, "learning_rate": 4.54613400278441e-06, "loss": 0.4951, "step": 12955 }, { "epoch": 0.5430975760560871, "grad_norm": 1.6685737371444702, "learning_rate": 4.545457957253835e-06, "loss": 0.425, "step": 12956 }, { "epoch": 0.543139494671096, "grad_norm": 1.775489330291748, "learning_rate": 4.544781920102117e-06, "loss": 0.5045, "step": 12957 }, { "epoch": 0.543181413286105, "grad_norm": 1.9958674907684326, "learning_rate": 4.544105891341713e-06, "loss": 0.5235, "step": 12958 }, { "epoch": 0.5432233319011139, "grad_norm": 2.1090786457061768, "learning_rate": 4.543429870985086e-06, "loss": 0.5273, "step": 12959 }, { "epoch": 0.543265250516123, "grad_norm": 1.6375755071640015, "learning_rate": 4.542753859044701e-06, "loss": 0.4717, "step": 12960 }, { "epoch": 0.5433071691311319, "grad_norm": 1.8839210271835327, "learning_rate": 4.542077855533014e-06, "loss": 0.5363, "step": 12961 }, { "epoch": 0.5433490877461409, "grad_norm": 1.7058449983596802, "learning_rate": 4.541401860462491e-06, "loss": 0.5224, "step": 12962 }, { "epoch": 0.5433910063611498, "grad_norm": 2.123283624649048, "learning_rate": 4.5407258738455885e-06, "loss": 0.5052, "step": 12963 }, { "epoch": 0.5434329249761588, "grad_norm": 1.9547390937805176, "learning_rate": 4.5400498956947694e-06, "loss": 0.5662, "step": 12964 }, { "epoch": 0.5434748435911677, "grad_norm": 1.8567256927490234, "learning_rate": 4.539373926022496e-06, "loss": 0.4792, "step": 12965 }, { "epoch": 0.5435167622061767, "grad_norm": 2.154116630554199, "learning_rate": 4.538697964841226e-06, "loss": 0.4948, "step": 12966 }, { "epoch": 0.5435586808211856, "grad_norm": 2.099184036254883, "learning_rate": 4.5380220121634225e-06, "loss": 0.5605, "step": 12967 }, { "epoch": 0.5436005994361947, "grad_norm": 1.9041006565093994, "learning_rate": 4.537346068001545e-06, "loss": 0.4972, "step": 12968 }, { "epoch": 0.5436425180512036, "grad_norm": 1.9831920862197876, "learning_rate": 4.536670132368052e-06, "loss": 0.4983, "step": 12969 }, { "epoch": 0.5436844366662126, "grad_norm": 1.6091628074645996, "learning_rate": 4.535994205275404e-06, "loss": 0.4992, "step": 12970 }, { "epoch": 0.5437263552812215, "grad_norm": 1.7858582735061646, "learning_rate": 4.535318286736063e-06, "loss": 0.543, "step": 12971 }, { "epoch": 0.5437682738962305, "grad_norm": 2.202176809310913, "learning_rate": 4.534642376762486e-06, "loss": 0.4541, "step": 12972 }, { "epoch": 0.5438101925112394, "grad_norm": 3.8004000186920166, "learning_rate": 4.5339664753671335e-06, "loss": 0.5331, "step": 12973 }, { "epoch": 0.5438521111262484, "grad_norm": 1.787100076675415, "learning_rate": 4.533290582562465e-06, "loss": 0.4899, "step": 12974 }, { "epoch": 0.5438940297412573, "grad_norm": 2.0522515773773193, "learning_rate": 4.532614698360939e-06, "loss": 0.5617, "step": 12975 }, { "epoch": 0.5439359483562664, "grad_norm": 1.8440512418746948, "learning_rate": 4.531938822775015e-06, "loss": 0.4582, "step": 12976 }, { "epoch": 0.5439778669712753, "grad_norm": 2.7390503883361816, "learning_rate": 4.5312629558171536e-06, "loss": 0.4984, "step": 12977 }, { "epoch": 0.5440197855862843, "grad_norm": 1.7002984285354614, "learning_rate": 4.530587097499808e-06, "loss": 0.463, "step": 12978 }, { "epoch": 0.5440617042012932, "grad_norm": 2.102195978164673, "learning_rate": 4.529911247835442e-06, "loss": 0.519, "step": 12979 }, { "epoch": 0.5441036228163022, "grad_norm": 1.594806432723999, "learning_rate": 4.529235406836513e-06, "loss": 0.4398, "step": 12980 }, { "epoch": 0.5441455414313111, "grad_norm": 1.6805129051208496, "learning_rate": 4.528559574515479e-06, "loss": 0.4905, "step": 12981 }, { "epoch": 0.54418746004632, "grad_norm": 2.044701099395752, "learning_rate": 4.527883750884795e-06, "loss": 0.4447, "step": 12982 }, { "epoch": 0.544229378661329, "grad_norm": 1.9155665636062622, "learning_rate": 4.527207935956925e-06, "loss": 0.5143, "step": 12983 }, { "epoch": 0.5442712972763379, "grad_norm": 2.0839273929595947, "learning_rate": 4.526532129744319e-06, "loss": 0.4961, "step": 12984 }, { "epoch": 0.544313215891347, "grad_norm": 1.8883039951324463, "learning_rate": 4.525856332259441e-06, "loss": 0.4964, "step": 12985 }, { "epoch": 0.5443551345063559, "grad_norm": 1.8853962421417236, "learning_rate": 4.525180543514747e-06, "loss": 0.4674, "step": 12986 }, { "epoch": 0.5443970531213649, "grad_norm": 2.6351287364959717, "learning_rate": 4.5245047635226905e-06, "loss": 0.5286, "step": 12987 }, { "epoch": 0.5444389717363738, "grad_norm": 1.5047190189361572, "learning_rate": 4.523828992295732e-06, "loss": 0.4616, "step": 12988 }, { "epoch": 0.5444808903513828, "grad_norm": 1.8234180212020874, "learning_rate": 4.52315322984633e-06, "loss": 0.5444, "step": 12989 }, { "epoch": 0.5445228089663917, "grad_norm": 2.603363275527954, "learning_rate": 4.522477476186936e-06, "loss": 0.5326, "step": 12990 }, { "epoch": 0.5445647275814007, "grad_norm": 5.567237377166748, "learning_rate": 4.521801731330011e-06, "loss": 0.4925, "step": 12991 }, { "epoch": 0.5446066461964096, "grad_norm": 1.5995961427688599, "learning_rate": 4.521125995288011e-06, "loss": 0.4464, "step": 12992 }, { "epoch": 0.5446485648114187, "grad_norm": 1.710754156112671, "learning_rate": 4.520450268073389e-06, "loss": 0.4354, "step": 12993 }, { "epoch": 0.5446904834264276, "grad_norm": 1.7580382823944092, "learning_rate": 4.519774549698605e-06, "loss": 0.4724, "step": 12994 }, { "epoch": 0.5447324020414366, "grad_norm": 2.3860092163085938, "learning_rate": 4.519098840176113e-06, "loss": 0.4636, "step": 12995 }, { "epoch": 0.5447743206564455, "grad_norm": 1.975708246231079, "learning_rate": 4.518423139518369e-06, "loss": 0.4977, "step": 12996 }, { "epoch": 0.5448162392714545, "grad_norm": 1.8846521377563477, "learning_rate": 4.517747447737827e-06, "loss": 0.5412, "step": 12997 }, { "epoch": 0.5448581578864634, "grad_norm": 1.914135456085205, "learning_rate": 4.517071764846947e-06, "loss": 0.4639, "step": 12998 }, { "epoch": 0.5449000765014724, "grad_norm": 8.884424209594727, "learning_rate": 4.5163960908581796e-06, "loss": 0.4935, "step": 12999 }, { "epoch": 0.5449419951164813, "grad_norm": 1.628261685371399, "learning_rate": 4.515720425783982e-06, "loss": 0.5074, "step": 13000 }, { "epoch": 0.5449839137314904, "grad_norm": 2.051283836364746, "learning_rate": 4.5150447696368104e-06, "loss": 0.5172, "step": 13001 }, { "epoch": 0.5450258323464993, "grad_norm": 1.9092576503753662, "learning_rate": 4.514369122429116e-06, "loss": 0.5416, "step": 13002 }, { "epoch": 0.5450677509615083, "grad_norm": 1.6979601383209229, "learning_rate": 4.513693484173355e-06, "loss": 0.4398, "step": 13003 }, { "epoch": 0.5451096695765172, "grad_norm": 1.727494239807129, "learning_rate": 4.5130178548819845e-06, "loss": 0.5037, "step": 13004 }, { "epoch": 0.5451515881915262, "grad_norm": 1.817773461341858, "learning_rate": 4.512342234567456e-06, "loss": 0.478, "step": 13005 }, { "epoch": 0.5451935068065351, "grad_norm": 1.7890545129776, "learning_rate": 4.511666623242223e-06, "loss": 0.4825, "step": 13006 }, { "epoch": 0.545235425421544, "grad_norm": 2.187366008758545, "learning_rate": 4.5109910209187425e-06, "loss": 0.5599, "step": 13007 }, { "epoch": 0.545277344036553, "grad_norm": 1.8889240026474, "learning_rate": 4.510315427609464e-06, "loss": 0.4363, "step": 13008 }, { "epoch": 0.5453192626515619, "grad_norm": 1.5782421827316284, "learning_rate": 4.509639843326846e-06, "loss": 0.4755, "step": 13009 }, { "epoch": 0.545361181266571, "grad_norm": 1.6462805271148682, "learning_rate": 4.5089642680833395e-06, "loss": 0.4493, "step": 13010 }, { "epoch": 0.5454030998815799, "grad_norm": 1.9361096620559692, "learning_rate": 4.508288701891396e-06, "loss": 0.5611, "step": 13011 }, { "epoch": 0.5454450184965889, "grad_norm": 1.7300282716751099, "learning_rate": 4.507613144763471e-06, "loss": 0.4925, "step": 13012 }, { "epoch": 0.5454869371115978, "grad_norm": 1.8141111135482788, "learning_rate": 4.506937596712019e-06, "loss": 0.5195, "step": 13013 }, { "epoch": 0.5455288557266068, "grad_norm": 1.8224387168884277, "learning_rate": 4.506262057749489e-06, "loss": 0.5516, "step": 13014 }, { "epoch": 0.5455707743416157, "grad_norm": 1.6895931959152222, "learning_rate": 4.505586527888335e-06, "loss": 0.5143, "step": 13015 }, { "epoch": 0.5456126929566247, "grad_norm": 1.6088082790374756, "learning_rate": 4.504911007141012e-06, "loss": 0.4789, "step": 13016 }, { "epoch": 0.5456546115716336, "grad_norm": 1.8073610067367554, "learning_rate": 4.504235495519968e-06, "loss": 0.5077, "step": 13017 }, { "epoch": 0.5456965301866427, "grad_norm": 1.8763703107833862, "learning_rate": 4.503559993037657e-06, "loss": 0.4627, "step": 13018 }, { "epoch": 0.5457384488016516, "grad_norm": 1.8080233335494995, "learning_rate": 4.502884499706534e-06, "loss": 0.5145, "step": 13019 }, { "epoch": 0.5457803674166606, "grad_norm": 1.8287816047668457, "learning_rate": 4.502209015539045e-06, "loss": 0.4925, "step": 13020 }, { "epoch": 0.5458222860316695, "grad_norm": 1.6337605714797974, "learning_rate": 4.501533540547645e-06, "loss": 0.4569, "step": 13021 }, { "epoch": 0.5458642046466785, "grad_norm": 2.1482253074645996, "learning_rate": 4.500858074744787e-06, "loss": 0.4713, "step": 13022 }, { "epoch": 0.5459061232616874, "grad_norm": 1.5899049043655396, "learning_rate": 4.50018261814292e-06, "loss": 0.5087, "step": 13023 }, { "epoch": 0.5459480418766964, "grad_norm": 1.75591242313385, "learning_rate": 4.499507170754493e-06, "loss": 0.4737, "step": 13024 }, { "epoch": 0.5459899604917053, "grad_norm": 1.5440908670425415, "learning_rate": 4.498831732591963e-06, "loss": 0.5477, "step": 13025 }, { "epoch": 0.5460318791067144, "grad_norm": 1.7216166257858276, "learning_rate": 4.498156303667774e-06, "loss": 0.523, "step": 13026 }, { "epoch": 0.5460737977217233, "grad_norm": 1.7213006019592285, "learning_rate": 4.497480883994381e-06, "loss": 0.5437, "step": 13027 }, { "epoch": 0.5461157163367323, "grad_norm": 1.541095495223999, "learning_rate": 4.496805473584235e-06, "loss": 0.4574, "step": 13028 }, { "epoch": 0.5461576349517412, "grad_norm": 2.0594639778137207, "learning_rate": 4.496130072449781e-06, "loss": 0.4748, "step": 13029 }, { "epoch": 0.5461995535667502, "grad_norm": 1.488981008529663, "learning_rate": 4.495454680603474e-06, "loss": 0.5027, "step": 13030 }, { "epoch": 0.5462414721817591, "grad_norm": 1.5963472127914429, "learning_rate": 4.494779298057763e-06, "loss": 0.5666, "step": 13031 }, { "epoch": 0.546283390796768, "grad_norm": 1.5894702672958374, "learning_rate": 4.4941039248250975e-06, "loss": 0.5026, "step": 13032 }, { "epoch": 0.546325309411777, "grad_norm": 1.6546854972839355, "learning_rate": 4.493428560917925e-06, "loss": 0.5296, "step": 13033 }, { "epoch": 0.546367228026786, "grad_norm": 1.8592616319656372, "learning_rate": 4.492753206348699e-06, "loss": 0.5613, "step": 13034 }, { "epoch": 0.546409146641795, "grad_norm": 1.7959355115890503, "learning_rate": 4.492077861129864e-06, "loss": 0.4912, "step": 13035 }, { "epoch": 0.5464510652568039, "grad_norm": 1.588929533958435, "learning_rate": 4.491402525273874e-06, "loss": 0.5129, "step": 13036 }, { "epoch": 0.5464929838718129, "grad_norm": 2.0807981491088867, "learning_rate": 4.490727198793174e-06, "loss": 0.5128, "step": 13037 }, { "epoch": 0.5465349024868218, "grad_norm": 2.0329947471618652, "learning_rate": 4.490051881700213e-06, "loss": 0.4457, "step": 13038 }, { "epoch": 0.5465768211018308, "grad_norm": 2.2134313583374023, "learning_rate": 4.4893765740074426e-06, "loss": 0.4775, "step": 13039 }, { "epoch": 0.5466187397168397, "grad_norm": 2.123572587966919, "learning_rate": 4.488701275727308e-06, "loss": 0.5092, "step": 13040 }, { "epoch": 0.5466606583318487, "grad_norm": 1.685006022453308, "learning_rate": 4.4880259868722585e-06, "loss": 0.4624, "step": 13041 }, { "epoch": 0.5467025769468576, "grad_norm": 1.7126842737197876, "learning_rate": 4.487350707454742e-06, "loss": 0.4878, "step": 13042 }, { "epoch": 0.5467444955618667, "grad_norm": 1.6701784133911133, "learning_rate": 4.486675437487208e-06, "loss": 0.506, "step": 13043 }, { "epoch": 0.5467864141768756, "grad_norm": 2.3515191078186035, "learning_rate": 4.4860001769821e-06, "loss": 0.5067, "step": 13044 }, { "epoch": 0.5468283327918846, "grad_norm": 1.6020936965942383, "learning_rate": 4.485324925951871e-06, "loss": 0.4802, "step": 13045 }, { "epoch": 0.5468702514068935, "grad_norm": 1.459431529045105, "learning_rate": 4.484649684408962e-06, "loss": 0.4572, "step": 13046 }, { "epoch": 0.5469121700219025, "grad_norm": 1.7097216844558716, "learning_rate": 4.483974452365827e-06, "loss": 0.5015, "step": 13047 }, { "epoch": 0.5469540886369114, "grad_norm": 2.036034107208252, "learning_rate": 4.483299229834909e-06, "loss": 0.4932, "step": 13048 }, { "epoch": 0.5469960072519204, "grad_norm": 1.6403560638427734, "learning_rate": 4.482624016828654e-06, "loss": 0.4361, "step": 13049 }, { "epoch": 0.5470379258669293, "grad_norm": 1.9856330156326294, "learning_rate": 4.48194881335951e-06, "loss": 0.4715, "step": 13050 }, { "epoch": 0.5470798444819384, "grad_norm": 1.8263907432556152, "learning_rate": 4.481273619439925e-06, "loss": 0.5005, "step": 13051 }, { "epoch": 0.5471217630969473, "grad_norm": 1.9882017374038696, "learning_rate": 4.480598435082343e-06, "loss": 0.5143, "step": 13052 }, { "epoch": 0.5471636817119563, "grad_norm": 1.8199893236160278, "learning_rate": 4.4799232602992095e-06, "loss": 0.5316, "step": 13053 }, { "epoch": 0.5472056003269652, "grad_norm": 1.5415066480636597, "learning_rate": 4.479248095102974e-06, "loss": 0.5482, "step": 13054 }, { "epoch": 0.5472475189419742, "grad_norm": 1.7519681453704834, "learning_rate": 4.478572939506078e-06, "loss": 0.5204, "step": 13055 }, { "epoch": 0.5472894375569831, "grad_norm": 1.8435391187667847, "learning_rate": 4.47789779352097e-06, "loss": 0.4863, "step": 13056 }, { "epoch": 0.547331356171992, "grad_norm": 1.8992723226547241, "learning_rate": 4.477222657160096e-06, "loss": 0.5122, "step": 13057 }, { "epoch": 0.547373274787001, "grad_norm": 1.6801689863204956, "learning_rate": 4.476547530435896e-06, "loss": 0.4551, "step": 13058 }, { "epoch": 0.54741519340201, "grad_norm": 4.965147018432617, "learning_rate": 4.4758724133608195e-06, "loss": 0.5229, "step": 13059 }, { "epoch": 0.547457112017019, "grad_norm": 3.292344331741333, "learning_rate": 4.475197305947312e-06, "loss": 0.4818, "step": 13060 }, { "epoch": 0.5474990306320279, "grad_norm": 2.183128833770752, "learning_rate": 4.474522208207816e-06, "loss": 0.4567, "step": 13061 }, { "epoch": 0.5475409492470369, "grad_norm": 1.9032737016677856, "learning_rate": 4.473847120154776e-06, "loss": 0.5493, "step": 13062 }, { "epoch": 0.5475828678620458, "grad_norm": 1.6538258790969849, "learning_rate": 4.473172041800639e-06, "loss": 0.4784, "step": 13063 }, { "epoch": 0.5476247864770548, "grad_norm": 1.8854080438613892, "learning_rate": 4.472496973157845e-06, "loss": 0.5048, "step": 13064 }, { "epoch": 0.5476667050920637, "grad_norm": 2.035046100616455, "learning_rate": 4.471821914238841e-06, "loss": 0.4585, "step": 13065 }, { "epoch": 0.5477086237070727, "grad_norm": 1.8252084255218506, "learning_rate": 4.4711468650560705e-06, "loss": 0.515, "step": 13066 }, { "epoch": 0.5477505423220816, "grad_norm": 1.8756723403930664, "learning_rate": 4.470471825621975e-06, "loss": 0.4912, "step": 13067 }, { "epoch": 0.5477924609370907, "grad_norm": 2.2329630851745605, "learning_rate": 4.469796795948998e-06, "loss": 0.4463, "step": 13068 }, { "epoch": 0.5478343795520996, "grad_norm": 1.9842724800109863, "learning_rate": 4.469121776049587e-06, "loss": 0.4668, "step": 13069 }, { "epoch": 0.5478762981671086, "grad_norm": 1.6985429525375366, "learning_rate": 4.468446765936182e-06, "loss": 0.5015, "step": 13070 }, { "epoch": 0.5479182167821175, "grad_norm": 2.5495994091033936, "learning_rate": 4.467771765621224e-06, "loss": 0.4873, "step": 13071 }, { "epoch": 0.5479601353971265, "grad_norm": 1.6854230165481567, "learning_rate": 4.46709677511716e-06, "loss": 0.5286, "step": 13072 }, { "epoch": 0.5480020540121354, "grad_norm": 1.8065935373306274, "learning_rate": 4.466421794436429e-06, "loss": 0.4621, "step": 13073 }, { "epoch": 0.5480439726271444, "grad_norm": 1.8938367366790771, "learning_rate": 4.465746823591474e-06, "loss": 0.4758, "step": 13074 }, { "epoch": 0.5480858912421533, "grad_norm": 1.772397756576538, "learning_rate": 4.4650718625947396e-06, "loss": 0.5421, "step": 13075 }, { "epoch": 0.5481278098571624, "grad_norm": 1.6059191226959229, "learning_rate": 4.4643969114586636e-06, "loss": 0.4996, "step": 13076 }, { "epoch": 0.5481697284721713, "grad_norm": 1.6387096643447876, "learning_rate": 4.46372197019569e-06, "loss": 0.4526, "step": 13077 }, { "epoch": 0.5482116470871803, "grad_norm": 2.006793975830078, "learning_rate": 4.463047038818263e-06, "loss": 0.5554, "step": 13078 }, { "epoch": 0.5482535657021892, "grad_norm": 2.072960138320923, "learning_rate": 4.46237211733882e-06, "loss": 0.4908, "step": 13079 }, { "epoch": 0.5482954843171982, "grad_norm": 1.8357256650924683, "learning_rate": 4.461697205769803e-06, "loss": 0.456, "step": 13080 }, { "epoch": 0.5483374029322071, "grad_norm": 1.7812905311584473, "learning_rate": 4.461022304123656e-06, "loss": 0.497, "step": 13081 }, { "epoch": 0.548379321547216, "grad_norm": 1.5430790185928345, "learning_rate": 4.460347412412816e-06, "loss": 0.4696, "step": 13082 }, { "epoch": 0.548421240162225, "grad_norm": 1.9015344381332397, "learning_rate": 4.459672530649726e-06, "loss": 0.512, "step": 13083 }, { "epoch": 0.548463158777234, "grad_norm": 2.113086223602295, "learning_rate": 4.458997658846826e-06, "loss": 0.4696, "step": 13084 }, { "epoch": 0.548505077392243, "grad_norm": 1.7540662288665771, "learning_rate": 4.458322797016556e-06, "loss": 0.5284, "step": 13085 }, { "epoch": 0.5485469960072519, "grad_norm": 2.0337607860565186, "learning_rate": 4.457647945171355e-06, "loss": 0.4884, "step": 13086 }, { "epoch": 0.5485889146222609, "grad_norm": 2.043287754058838, "learning_rate": 4.456973103323668e-06, "loss": 0.5474, "step": 13087 }, { "epoch": 0.5486308332372698, "grad_norm": 1.7157773971557617, "learning_rate": 4.456298271485927e-06, "loss": 0.5294, "step": 13088 }, { "epoch": 0.5486727518522788, "grad_norm": 1.617796778678894, "learning_rate": 4.455623449670578e-06, "loss": 0.4799, "step": 13089 }, { "epoch": 0.5487146704672877, "grad_norm": 1.8763267993927002, "learning_rate": 4.454948637890059e-06, "loss": 0.4715, "step": 13090 }, { "epoch": 0.5487565890822967, "grad_norm": 2.295413017272949, "learning_rate": 4.454273836156806e-06, "loss": 0.4273, "step": 13091 }, { "epoch": 0.5487985076973056, "grad_norm": 1.6073256731033325, "learning_rate": 4.4535990444832615e-06, "loss": 0.5034, "step": 13092 }, { "epoch": 0.5488404263123147, "grad_norm": 1.721110224723816, "learning_rate": 4.452924262881863e-06, "loss": 0.4755, "step": 13093 }, { "epoch": 0.5488823449273236, "grad_norm": 8.05040454864502, "learning_rate": 4.45224949136505e-06, "loss": 0.5587, "step": 13094 }, { "epoch": 0.5489242635423326, "grad_norm": 1.8542296886444092, "learning_rate": 4.451574729945259e-06, "loss": 0.4941, "step": 13095 }, { "epoch": 0.5489661821573415, "grad_norm": 1.7592591047286987, "learning_rate": 4.450899978634932e-06, "loss": 0.4882, "step": 13096 }, { "epoch": 0.5490081007723505, "grad_norm": 1.8096520900726318, "learning_rate": 4.4502252374465026e-06, "loss": 0.5521, "step": 13097 }, { "epoch": 0.5490500193873594, "grad_norm": 1.6066484451293945, "learning_rate": 4.449550506392412e-06, "loss": 0.4508, "step": 13098 }, { "epoch": 0.5490919380023684, "grad_norm": 1.8325179815292358, "learning_rate": 4.448875785485098e-06, "loss": 0.4627, "step": 13099 }, { "epoch": 0.5491338566173773, "grad_norm": 2.172694206237793, "learning_rate": 4.448201074736994e-06, "loss": 0.493, "step": 13100 }, { "epoch": 0.5491757752323864, "grad_norm": 1.6721266508102417, "learning_rate": 4.447526374160542e-06, "loss": 0.4306, "step": 13101 }, { "epoch": 0.5492176938473953, "grad_norm": 1.7563265562057495, "learning_rate": 4.446851683768178e-06, "loss": 0.4931, "step": 13102 }, { "epoch": 0.5492596124624043, "grad_norm": 2.0741562843322754, "learning_rate": 4.446177003572338e-06, "loss": 0.4762, "step": 13103 }, { "epoch": 0.5493015310774132, "grad_norm": 1.7565633058547974, "learning_rate": 4.445502333585459e-06, "loss": 0.5119, "step": 13104 }, { "epoch": 0.5493434496924222, "grad_norm": 1.9147765636444092, "learning_rate": 4.444827673819979e-06, "loss": 0.4818, "step": 13105 }, { "epoch": 0.5493853683074311, "grad_norm": 1.7003246545791626, "learning_rate": 4.444153024288332e-06, "loss": 0.482, "step": 13106 }, { "epoch": 0.54942728692244, "grad_norm": 1.6219972372055054, "learning_rate": 4.443478385002956e-06, "loss": 0.4407, "step": 13107 }, { "epoch": 0.549469205537449, "grad_norm": 1.6397390365600586, "learning_rate": 4.442803755976288e-06, "loss": 0.4887, "step": 13108 }, { "epoch": 0.549511124152458, "grad_norm": 1.7405673265457153, "learning_rate": 4.44212913722076e-06, "loss": 0.4321, "step": 13109 }, { "epoch": 0.549553042767467, "grad_norm": 1.8370221853256226, "learning_rate": 4.441454528748813e-06, "loss": 0.5383, "step": 13110 }, { "epoch": 0.5495949613824759, "grad_norm": 1.546294093132019, "learning_rate": 4.440779930572877e-06, "loss": 0.4676, "step": 13111 }, { "epoch": 0.5496368799974849, "grad_norm": 1.9499027729034424, "learning_rate": 4.4401053427053906e-06, "loss": 0.539, "step": 13112 }, { "epoch": 0.5496787986124938, "grad_norm": 1.7779077291488647, "learning_rate": 4.439430765158789e-06, "loss": 0.5183, "step": 13113 }, { "epoch": 0.5497207172275028, "grad_norm": 1.7380468845367432, "learning_rate": 4.438756197945505e-06, "loss": 0.512, "step": 13114 }, { "epoch": 0.5497626358425117, "grad_norm": 1.7287986278533936, "learning_rate": 4.438081641077974e-06, "loss": 0.5789, "step": 13115 }, { "epoch": 0.5498045544575207, "grad_norm": 1.7592177391052246, "learning_rate": 4.437407094568633e-06, "loss": 0.4013, "step": 13116 }, { "epoch": 0.5498464730725297, "grad_norm": 1.5742954015731812, "learning_rate": 4.436732558429915e-06, "loss": 0.4957, "step": 13117 }, { "epoch": 0.5498883916875387, "grad_norm": 1.4650813341140747, "learning_rate": 4.436058032674251e-06, "loss": 0.4461, "step": 13118 }, { "epoch": 0.5499303103025476, "grad_norm": 1.664433240890503, "learning_rate": 4.43538351731408e-06, "loss": 0.5281, "step": 13119 }, { "epoch": 0.5499722289175566, "grad_norm": 1.735544204711914, "learning_rate": 4.4347090123618315e-06, "loss": 0.4689, "step": 13120 }, { "epoch": 0.5500141475325655, "grad_norm": 1.6068743467330933, "learning_rate": 4.4340345178299415e-06, "loss": 0.4409, "step": 13121 }, { "epoch": 0.5500560661475745, "grad_norm": 1.6153732538223267, "learning_rate": 4.433360033730844e-06, "loss": 0.4742, "step": 13122 }, { "epoch": 0.5500979847625834, "grad_norm": 1.5596450567245483, "learning_rate": 4.432685560076968e-06, "loss": 0.5043, "step": 13123 }, { "epoch": 0.5501399033775924, "grad_norm": 1.6587984561920166, "learning_rate": 4.432011096880751e-06, "loss": 0.4976, "step": 13124 }, { "epoch": 0.5501818219926013, "grad_norm": 1.4695364236831665, "learning_rate": 4.431336644154626e-06, "loss": 0.4708, "step": 13125 }, { "epoch": 0.5502237406076104, "grad_norm": 1.5063691139221191, "learning_rate": 4.430662201911021e-06, "loss": 0.4802, "step": 13126 }, { "epoch": 0.5502656592226193, "grad_norm": 1.8391252756118774, "learning_rate": 4.429987770162372e-06, "loss": 0.5216, "step": 13127 }, { "epoch": 0.5503075778376283, "grad_norm": 1.774377465248108, "learning_rate": 4.429313348921112e-06, "loss": 0.5173, "step": 13128 }, { "epoch": 0.5503494964526372, "grad_norm": 1.7815062999725342, "learning_rate": 4.42863893819967e-06, "loss": 0.5034, "step": 13129 }, { "epoch": 0.5503914150676462, "grad_norm": 1.7769780158996582, "learning_rate": 4.42796453801048e-06, "loss": 0.4527, "step": 13130 }, { "epoch": 0.5504333336826551, "grad_norm": 1.634919285774231, "learning_rate": 4.427290148365972e-06, "loss": 0.513, "step": 13131 }, { "epoch": 0.550475252297664, "grad_norm": 1.6981658935546875, "learning_rate": 4.426615769278579e-06, "loss": 0.5169, "step": 13132 }, { "epoch": 0.550517170912673, "grad_norm": 2.1260316371917725, "learning_rate": 4.42594140076073e-06, "loss": 0.5139, "step": 13133 }, { "epoch": 0.550559089527682, "grad_norm": 1.6818329095840454, "learning_rate": 4.4252670428248605e-06, "loss": 0.512, "step": 13134 }, { "epoch": 0.550601008142691, "grad_norm": 1.8509671688079834, "learning_rate": 4.424592695483397e-06, "loss": 0.5085, "step": 13135 }, { "epoch": 0.5506429267576999, "grad_norm": 1.8910518884658813, "learning_rate": 4.423918358748772e-06, "loss": 0.4787, "step": 13136 }, { "epoch": 0.5506848453727089, "grad_norm": 1.6055097579956055, "learning_rate": 4.423244032633416e-06, "loss": 0.4102, "step": 13137 }, { "epoch": 0.5507267639877178, "grad_norm": 1.8748834133148193, "learning_rate": 4.422569717149758e-06, "loss": 0.4859, "step": 13138 }, { "epoch": 0.5507686826027268, "grad_norm": 2.122995376586914, "learning_rate": 4.421895412310228e-06, "loss": 0.5402, "step": 13139 }, { "epoch": 0.5508106012177357, "grad_norm": 1.9026645421981812, "learning_rate": 4.42122111812726e-06, "loss": 0.4919, "step": 13140 }, { "epoch": 0.5508525198327447, "grad_norm": 1.8835347890853882, "learning_rate": 4.420546834613279e-06, "loss": 0.5231, "step": 13141 }, { "epoch": 0.5508944384477537, "grad_norm": 3.3365535736083984, "learning_rate": 4.419872561780715e-06, "loss": 0.5171, "step": 13142 }, { "epoch": 0.5509363570627627, "grad_norm": 1.6080467700958252, "learning_rate": 4.419198299642e-06, "loss": 0.441, "step": 13143 }, { "epoch": 0.5509782756777716, "grad_norm": 1.4769459962844849, "learning_rate": 4.4185240482095595e-06, "loss": 0.4355, "step": 13144 }, { "epoch": 0.5510201942927806, "grad_norm": 1.4306445121765137, "learning_rate": 4.417849807495825e-06, "loss": 0.4923, "step": 13145 }, { "epoch": 0.5510621129077895, "grad_norm": 1.8334182500839233, "learning_rate": 4.417175577513227e-06, "loss": 0.5199, "step": 13146 }, { "epoch": 0.5511040315227985, "grad_norm": 1.6129628419876099, "learning_rate": 4.416501358274188e-06, "loss": 0.4188, "step": 13147 }, { "epoch": 0.5511459501378074, "grad_norm": 1.6437357664108276, "learning_rate": 4.41582714979114e-06, "loss": 0.5408, "step": 13148 }, { "epoch": 0.5511878687528164, "grad_norm": 1.721449613571167, "learning_rate": 4.4151529520765134e-06, "loss": 0.5223, "step": 13149 }, { "epoch": 0.5512297873678254, "grad_norm": 2.051738739013672, "learning_rate": 4.414478765142733e-06, "loss": 0.4784, "step": 13150 }, { "epoch": 0.5512717059828344, "grad_norm": 1.5158153772354126, "learning_rate": 4.413804589002226e-06, "loss": 0.429, "step": 13151 }, { "epoch": 0.5513136245978433, "grad_norm": 1.807446002960205, "learning_rate": 4.413130423667421e-06, "loss": 0.5212, "step": 13152 }, { "epoch": 0.5513555432128523, "grad_norm": 1.7363749742507935, "learning_rate": 4.412456269150746e-06, "loss": 0.504, "step": 13153 }, { "epoch": 0.5513974618278612, "grad_norm": 1.514298677444458, "learning_rate": 4.4117821254646265e-06, "loss": 0.4985, "step": 13154 }, { "epoch": 0.5514393804428702, "grad_norm": 1.9473912715911865, "learning_rate": 4.411107992621492e-06, "loss": 0.5521, "step": 13155 }, { "epoch": 0.5514812990578791, "grad_norm": 1.6033324003219604, "learning_rate": 4.410433870633766e-06, "loss": 0.4749, "step": 13156 }, { "epoch": 0.551523217672888, "grad_norm": 1.769068956375122, "learning_rate": 4.409759759513876e-06, "loss": 0.5078, "step": 13157 }, { "epoch": 0.551565136287897, "grad_norm": 1.8732149600982666, "learning_rate": 4.409085659274251e-06, "loss": 0.4271, "step": 13158 }, { "epoch": 0.551607054902906, "grad_norm": 1.7717972993850708, "learning_rate": 4.408411569927315e-06, "loss": 0.5158, "step": 13159 }, { "epoch": 0.551648973517915, "grad_norm": 1.7836285829544067, "learning_rate": 4.407737491485491e-06, "loss": 0.5188, "step": 13160 }, { "epoch": 0.5516908921329239, "grad_norm": 1.6600693464279175, "learning_rate": 4.40706342396121e-06, "loss": 0.497, "step": 13161 }, { "epoch": 0.5517328107479329, "grad_norm": 1.7613829374313354, "learning_rate": 4.406389367366894e-06, "loss": 0.5044, "step": 13162 }, { "epoch": 0.5517747293629418, "grad_norm": 1.6224355697631836, "learning_rate": 4.4057153217149695e-06, "loss": 0.4398, "step": 13163 }, { "epoch": 0.5518166479779508, "grad_norm": 1.7327479124069214, "learning_rate": 4.405041287017864e-06, "loss": 0.4442, "step": 13164 }, { "epoch": 0.5518585665929597, "grad_norm": 1.9554499387741089, "learning_rate": 4.4043672632879955e-06, "loss": 0.5139, "step": 13165 }, { "epoch": 0.5519004852079687, "grad_norm": 1.7608749866485596, "learning_rate": 4.4036932505377945e-06, "loss": 0.4744, "step": 13166 }, { "epoch": 0.5519424038229777, "grad_norm": 1.5262603759765625, "learning_rate": 4.403019248779686e-06, "loss": 0.4923, "step": 13167 }, { "epoch": 0.5519843224379867, "grad_norm": 1.4959661960601807, "learning_rate": 4.402345258026091e-06, "loss": 0.4987, "step": 13168 }, { "epoch": 0.5520262410529956, "grad_norm": 1.7709319591522217, "learning_rate": 4.401671278289434e-06, "loss": 0.5279, "step": 13169 }, { "epoch": 0.5520681596680046, "grad_norm": 2.7197368144989014, "learning_rate": 4.400997309582142e-06, "loss": 0.4988, "step": 13170 }, { "epoch": 0.5521100782830135, "grad_norm": 1.4400296211242676, "learning_rate": 4.400323351916634e-06, "loss": 0.4533, "step": 13171 }, { "epoch": 0.5521519968980225, "grad_norm": 1.3929953575134277, "learning_rate": 4.399649405305338e-06, "loss": 0.4391, "step": 13172 }, { "epoch": 0.5521939155130314, "grad_norm": 1.5571056604385376, "learning_rate": 4.398975469760676e-06, "loss": 0.5095, "step": 13173 }, { "epoch": 0.5522358341280404, "grad_norm": 2.4020354747772217, "learning_rate": 4.398301545295067e-06, "loss": 0.5059, "step": 13174 }, { "epoch": 0.5522777527430494, "grad_norm": 1.4945634603500366, "learning_rate": 4.397627631920939e-06, "loss": 0.5092, "step": 13175 }, { "epoch": 0.5523196713580584, "grad_norm": 1.699838399887085, "learning_rate": 4.396953729650713e-06, "loss": 0.494, "step": 13176 }, { "epoch": 0.5523615899730673, "grad_norm": 1.691150188446045, "learning_rate": 4.396279838496812e-06, "loss": 0.4822, "step": 13177 }, { "epoch": 0.5524035085880763, "grad_norm": 1.8535752296447754, "learning_rate": 4.395605958471656e-06, "loss": 0.4942, "step": 13178 }, { "epoch": 0.5524454272030852, "grad_norm": 1.716016411781311, "learning_rate": 4.3949320895876705e-06, "loss": 0.4382, "step": 13179 }, { "epoch": 0.5524873458180942, "grad_norm": 1.7898049354553223, "learning_rate": 4.394258231857274e-06, "loss": 0.4992, "step": 13180 }, { "epoch": 0.5525292644331031, "grad_norm": 1.79570472240448, "learning_rate": 4.39358438529289e-06, "loss": 0.5213, "step": 13181 }, { "epoch": 0.552571183048112, "grad_norm": 2.0255775451660156, "learning_rate": 4.392910549906939e-06, "loss": 0.5317, "step": 13182 }, { "epoch": 0.552613101663121, "grad_norm": 1.749367117881775, "learning_rate": 4.392236725711843e-06, "loss": 0.4471, "step": 13183 }, { "epoch": 0.55265502027813, "grad_norm": 1.6053240299224854, "learning_rate": 4.391562912720022e-06, "loss": 0.4643, "step": 13184 }, { "epoch": 0.552696938893139, "grad_norm": 1.728859305381775, "learning_rate": 4.3908891109439e-06, "loss": 0.4751, "step": 13185 }, { "epoch": 0.5527388575081479, "grad_norm": 1.7272073030471802, "learning_rate": 4.390215320395893e-06, "loss": 0.5435, "step": 13186 }, { "epoch": 0.5527807761231569, "grad_norm": 2.1150903701782227, "learning_rate": 4.389541541088424e-06, "loss": 0.5552, "step": 13187 }, { "epoch": 0.5528226947381658, "grad_norm": 6.415891170501709, "learning_rate": 4.388867773033914e-06, "loss": 0.5424, "step": 13188 }, { "epoch": 0.5528646133531748, "grad_norm": 1.4801735877990723, "learning_rate": 4.38819401624478e-06, "loss": 0.4757, "step": 13189 }, { "epoch": 0.5529065319681837, "grad_norm": 8.779144287109375, "learning_rate": 4.387520270733445e-06, "loss": 0.4261, "step": 13190 }, { "epoch": 0.5529484505831928, "grad_norm": 1.6101746559143066, "learning_rate": 4.386846536512326e-06, "loss": 0.4491, "step": 13191 }, { "epoch": 0.5529903691982017, "grad_norm": 3.140242099761963, "learning_rate": 4.386172813593843e-06, "loss": 0.5269, "step": 13192 }, { "epoch": 0.5530322878132107, "grad_norm": 1.901631236076355, "learning_rate": 4.3854991019904185e-06, "loss": 0.5063, "step": 13193 }, { "epoch": 0.5530742064282196, "grad_norm": 1.7241307497024536, "learning_rate": 4.384825401714465e-06, "loss": 0.5559, "step": 13194 }, { "epoch": 0.5531161250432286, "grad_norm": 2.034520387649536, "learning_rate": 4.384151712778405e-06, "loss": 0.4871, "step": 13195 }, { "epoch": 0.5531580436582375, "grad_norm": 1.6310096979141235, "learning_rate": 4.38347803519466e-06, "loss": 0.4562, "step": 13196 }, { "epoch": 0.5531999622732465, "grad_norm": 1.4634281396865845, "learning_rate": 4.3828043689756426e-06, "loss": 0.449, "step": 13197 }, { "epoch": 0.5532418808882554, "grad_norm": 1.6379213333129883, "learning_rate": 4.382130714133773e-06, "loss": 0.446, "step": 13198 }, { "epoch": 0.5532837995032645, "grad_norm": 1.7684245109558105, "learning_rate": 4.381457070681471e-06, "loss": 0.5126, "step": 13199 }, { "epoch": 0.5533257181182734, "grad_norm": 1.9030107259750366, "learning_rate": 4.380783438631151e-06, "loss": 0.5248, "step": 13200 }, { "epoch": 0.5533676367332824, "grad_norm": 1.745710015296936, "learning_rate": 4.380109817995234e-06, "loss": 0.4675, "step": 13201 }, { "epoch": 0.5534095553482913, "grad_norm": 1.718253493309021, "learning_rate": 4.3794362087861355e-06, "loss": 0.4875, "step": 13202 }, { "epoch": 0.5534514739633003, "grad_norm": 1.884289264678955, "learning_rate": 4.378762611016271e-06, "loss": 0.52, "step": 13203 }, { "epoch": 0.5534933925783092, "grad_norm": 1.9654186964035034, "learning_rate": 4.37808902469806e-06, "loss": 0.551, "step": 13204 }, { "epoch": 0.5535353111933182, "grad_norm": 1.6337023973464966, "learning_rate": 4.377415449843919e-06, "loss": 0.5451, "step": 13205 }, { "epoch": 0.5535772298083271, "grad_norm": 2.231804609298706, "learning_rate": 4.3767418864662625e-06, "loss": 0.4716, "step": 13206 }, { "epoch": 0.553619148423336, "grad_norm": 1.6145652532577515, "learning_rate": 4.376068334577506e-06, "loss": 0.5121, "step": 13207 }, { "epoch": 0.5536610670383451, "grad_norm": 1.6846579313278198, "learning_rate": 4.37539479419007e-06, "loss": 0.5214, "step": 13208 }, { "epoch": 0.553702985653354, "grad_norm": 1.872926115989685, "learning_rate": 4.374721265316366e-06, "loss": 0.5641, "step": 13209 }, { "epoch": 0.553744904268363, "grad_norm": 1.7663440704345703, "learning_rate": 4.374047747968811e-06, "loss": 0.483, "step": 13210 }, { "epoch": 0.5537868228833719, "grad_norm": 1.5752575397491455, "learning_rate": 4.373374242159822e-06, "loss": 0.502, "step": 13211 }, { "epoch": 0.5538287414983809, "grad_norm": 1.745006799697876, "learning_rate": 4.3727007479018105e-06, "loss": 0.5211, "step": 13212 }, { "epoch": 0.5538706601133898, "grad_norm": 2.549570322036743, "learning_rate": 4.3720272652071935e-06, "loss": 0.537, "step": 13213 }, { "epoch": 0.5539125787283988, "grad_norm": 1.578230857849121, "learning_rate": 4.371353794088388e-06, "loss": 0.5072, "step": 13214 }, { "epoch": 0.5539544973434077, "grad_norm": 1.86330246925354, "learning_rate": 4.3706803345578055e-06, "loss": 0.5325, "step": 13215 }, { "epoch": 0.5539964159584168, "grad_norm": 1.741714596748352, "learning_rate": 4.370006886627859e-06, "loss": 0.4865, "step": 13216 }, { "epoch": 0.5540383345734257, "grad_norm": 1.7966153621673584, "learning_rate": 4.369333450310968e-06, "loss": 0.5508, "step": 13217 }, { "epoch": 0.5540802531884347, "grad_norm": 1.5787749290466309, "learning_rate": 4.36866002561954e-06, "loss": 0.5076, "step": 13218 }, { "epoch": 0.5541221718034436, "grad_norm": 1.693611979484558, "learning_rate": 4.367986612565992e-06, "loss": 0.5551, "step": 13219 }, { "epoch": 0.5541640904184526, "grad_norm": 1.5808618068695068, "learning_rate": 4.367313211162739e-06, "loss": 0.4713, "step": 13220 }, { "epoch": 0.5542060090334615, "grad_norm": 1.6039519309997559, "learning_rate": 4.366639821422189e-06, "loss": 0.4949, "step": 13221 }, { "epoch": 0.5542479276484705, "grad_norm": 1.5585521459579468, "learning_rate": 4.36596644335676e-06, "loss": 0.5131, "step": 13222 }, { "epoch": 0.5542898462634794, "grad_norm": 2.6350948810577393, "learning_rate": 4.3652930769788645e-06, "loss": 0.4899, "step": 13223 }, { "epoch": 0.5543317648784885, "grad_norm": 1.8031034469604492, "learning_rate": 4.3646197223009126e-06, "loss": 0.4831, "step": 13224 }, { "epoch": 0.5543736834934974, "grad_norm": 1.5527245998382568, "learning_rate": 4.363946379335316e-06, "loss": 0.4969, "step": 13225 }, { "epoch": 0.5544156021085064, "grad_norm": 1.890942931175232, "learning_rate": 4.3632730480944915e-06, "loss": 0.5218, "step": 13226 }, { "epoch": 0.5544575207235153, "grad_norm": 1.8356584310531616, "learning_rate": 4.362599728590846e-06, "loss": 0.5146, "step": 13227 }, { "epoch": 0.5544994393385243, "grad_norm": 1.4448554515838623, "learning_rate": 4.361926420836794e-06, "loss": 0.5109, "step": 13228 }, { "epoch": 0.5545413579535332, "grad_norm": 1.6497406959533691, "learning_rate": 4.361253124844747e-06, "loss": 0.4902, "step": 13229 }, { "epoch": 0.5545832765685422, "grad_norm": 1.6634902954101562, "learning_rate": 4.360579840627115e-06, "loss": 0.4827, "step": 13230 }, { "epoch": 0.5546251951835511, "grad_norm": 1.8512706756591797, "learning_rate": 4.35990656819631e-06, "loss": 0.4771, "step": 13231 }, { "epoch": 0.55466711379856, "grad_norm": 2.1014983654022217, "learning_rate": 4.359233307564743e-06, "loss": 0.5229, "step": 13232 }, { "epoch": 0.5547090324135691, "grad_norm": 1.5203430652618408, "learning_rate": 4.358560058744822e-06, "loss": 0.549, "step": 13233 }, { "epoch": 0.554750951028578, "grad_norm": 2.0589683055877686, "learning_rate": 4.3578868217489615e-06, "loss": 0.5263, "step": 13234 }, { "epoch": 0.554792869643587, "grad_norm": 1.7652901411056519, "learning_rate": 4.35721359658957e-06, "loss": 0.5272, "step": 13235 }, { "epoch": 0.5548347882585959, "grad_norm": 2.0390799045562744, "learning_rate": 4.356540383279055e-06, "loss": 0.4669, "step": 13236 }, { "epoch": 0.5548767068736049, "grad_norm": 1.5923876762390137, "learning_rate": 4.355867181829829e-06, "loss": 0.4747, "step": 13237 }, { "epoch": 0.5549186254886138, "grad_norm": 1.417489767074585, "learning_rate": 4.355193992254303e-06, "loss": 0.4492, "step": 13238 }, { "epoch": 0.5549605441036228, "grad_norm": 3.1746580600738525, "learning_rate": 4.354520814564884e-06, "loss": 0.5272, "step": 13239 }, { "epoch": 0.5550024627186317, "grad_norm": 1.6669944524765015, "learning_rate": 4.3538476487739795e-06, "loss": 0.5189, "step": 13240 }, { "epoch": 0.5550443813336408, "grad_norm": 1.5393848419189453, "learning_rate": 4.353174494894003e-06, "loss": 0.4834, "step": 13241 }, { "epoch": 0.5550862999486497, "grad_norm": 1.6131373643875122, "learning_rate": 4.352501352937357e-06, "loss": 0.5095, "step": 13242 }, { "epoch": 0.5551282185636587, "grad_norm": 1.7541834115982056, "learning_rate": 4.351828222916456e-06, "loss": 0.5778, "step": 13243 }, { "epoch": 0.5551701371786676, "grad_norm": 1.540083885192871, "learning_rate": 4.351155104843706e-06, "loss": 0.4188, "step": 13244 }, { "epoch": 0.5552120557936766, "grad_norm": 2.424184560775757, "learning_rate": 4.350481998731513e-06, "loss": 0.5398, "step": 13245 }, { "epoch": 0.5552539744086855, "grad_norm": 1.5749988555908203, "learning_rate": 4.349808904592286e-06, "loss": 0.4703, "step": 13246 }, { "epoch": 0.5552958930236945, "grad_norm": 1.5152332782745361, "learning_rate": 4.349135822438435e-06, "loss": 0.4648, "step": 13247 }, { "epoch": 0.5553378116387034, "grad_norm": 2.051534652709961, "learning_rate": 4.348462752282365e-06, "loss": 0.5266, "step": 13248 }, { "epoch": 0.5553797302537125, "grad_norm": 1.8856412172317505, "learning_rate": 4.347789694136482e-06, "loss": 0.4777, "step": 13249 }, { "epoch": 0.5554216488687214, "grad_norm": 1.9635363817214966, "learning_rate": 4.347116648013196e-06, "loss": 0.5277, "step": 13250 }, { "epoch": 0.5554635674837304, "grad_norm": 1.5618826150894165, "learning_rate": 4.34644361392491e-06, "loss": 0.5023, "step": 13251 }, { "epoch": 0.5555054860987393, "grad_norm": 1.7041808366775513, "learning_rate": 4.345770591884034e-06, "loss": 0.5112, "step": 13252 }, { "epoch": 0.5555474047137483, "grad_norm": 1.6867249011993408, "learning_rate": 4.345097581902973e-06, "loss": 0.4705, "step": 13253 }, { "epoch": 0.5555893233287572, "grad_norm": 1.6499675512313843, "learning_rate": 4.3444245839941314e-06, "loss": 0.4683, "step": 13254 }, { "epoch": 0.5556312419437662, "grad_norm": 1.6806455850601196, "learning_rate": 4.343751598169916e-06, "loss": 0.4843, "step": 13255 }, { "epoch": 0.5556731605587751, "grad_norm": 2.002480983734131, "learning_rate": 4.343078624442735e-06, "loss": 0.4774, "step": 13256 }, { "epoch": 0.555715079173784, "grad_norm": 1.9828780889511108, "learning_rate": 4.34240566282499e-06, "loss": 0.5007, "step": 13257 }, { "epoch": 0.5557569977887931, "grad_norm": 1.5715301036834717, "learning_rate": 4.3417327133290865e-06, "loss": 0.4786, "step": 13258 }, { "epoch": 0.555798916403802, "grad_norm": 2.141594171524048, "learning_rate": 4.341059775967432e-06, "loss": 0.5429, "step": 13259 }, { "epoch": 0.555840835018811, "grad_norm": 1.6669907569885254, "learning_rate": 4.340386850752428e-06, "loss": 0.549, "step": 13260 }, { "epoch": 0.5558827536338199, "grad_norm": 1.591509461402893, "learning_rate": 4.339713937696483e-06, "loss": 0.4326, "step": 13261 }, { "epoch": 0.5559246722488289, "grad_norm": 1.7872676849365234, "learning_rate": 4.339041036811998e-06, "loss": 0.5044, "step": 13262 }, { "epoch": 0.5559665908638378, "grad_norm": 1.6930595636367798, "learning_rate": 4.3383681481113754e-06, "loss": 0.4748, "step": 13263 }, { "epoch": 0.5560085094788468, "grad_norm": 1.924591302871704, "learning_rate": 4.337695271607024e-06, "loss": 0.491, "step": 13264 }, { "epoch": 0.5560504280938557, "grad_norm": 2.3174335956573486, "learning_rate": 4.337022407311343e-06, "loss": 0.5049, "step": 13265 }, { "epoch": 0.5560923467088648, "grad_norm": 1.6328229904174805, "learning_rate": 4.336349555236737e-06, "loss": 0.4811, "step": 13266 }, { "epoch": 0.5561342653238737, "grad_norm": 1.7359097003936768, "learning_rate": 4.335676715395612e-06, "loss": 0.4584, "step": 13267 }, { "epoch": 0.5561761839388827, "grad_norm": 2.0439674854278564, "learning_rate": 4.335003887800366e-06, "loss": 0.4483, "step": 13268 }, { "epoch": 0.5562181025538916, "grad_norm": 1.7358912229537964, "learning_rate": 4.3343310724634035e-06, "loss": 0.4584, "step": 13269 }, { "epoch": 0.5562600211689006, "grad_norm": 1.7468690872192383, "learning_rate": 4.3336582693971295e-06, "loss": 0.4887, "step": 13270 }, { "epoch": 0.5563019397839095, "grad_norm": 2.026609420776367, "learning_rate": 4.332985478613944e-06, "loss": 0.5163, "step": 13271 }, { "epoch": 0.5563438583989185, "grad_norm": 1.8521323204040527, "learning_rate": 4.332312700126247e-06, "loss": 0.5154, "step": 13272 }, { "epoch": 0.5563857770139274, "grad_norm": 1.6558841466903687, "learning_rate": 4.331639933946445e-06, "loss": 0.4819, "step": 13273 }, { "epoch": 0.5564276956289365, "grad_norm": 2.3122036457061768, "learning_rate": 4.330967180086935e-06, "loss": 0.4562, "step": 13274 }, { "epoch": 0.5564696142439454, "grad_norm": 1.86676824092865, "learning_rate": 4.33029443856012e-06, "loss": 0.4416, "step": 13275 }, { "epoch": 0.5565115328589544, "grad_norm": 1.8612018823623657, "learning_rate": 4.329621709378402e-06, "loss": 0.5144, "step": 13276 }, { "epoch": 0.5565534514739633, "grad_norm": 1.600775957107544, "learning_rate": 4.3289489925541815e-06, "loss": 0.4871, "step": 13277 }, { "epoch": 0.5565953700889723, "grad_norm": 1.9177852869033813, "learning_rate": 4.328276288099856e-06, "loss": 0.5418, "step": 13278 }, { "epoch": 0.5566372887039812, "grad_norm": 1.629281997680664, "learning_rate": 4.3276035960278324e-06, "loss": 0.5247, "step": 13279 }, { "epoch": 0.5566792073189902, "grad_norm": 3.1342437267303467, "learning_rate": 4.326930916350504e-06, "loss": 0.4421, "step": 13280 }, { "epoch": 0.5567211259339991, "grad_norm": 3.5672013759613037, "learning_rate": 4.326258249080274e-06, "loss": 0.4999, "step": 13281 }, { "epoch": 0.556763044549008, "grad_norm": 1.7622694969177246, "learning_rate": 4.325585594229544e-06, "loss": 0.4683, "step": 13282 }, { "epoch": 0.5568049631640171, "grad_norm": 1.7931036949157715, "learning_rate": 4.324912951810709e-06, "loss": 0.4931, "step": 13283 }, { "epoch": 0.556846881779026, "grad_norm": 1.7523577213287354, "learning_rate": 4.32424032183617e-06, "loss": 0.4887, "step": 13284 }, { "epoch": 0.556888800394035, "grad_norm": 1.6025011539459229, "learning_rate": 4.323567704318329e-06, "loss": 0.5413, "step": 13285 }, { "epoch": 0.5569307190090439, "grad_norm": 1.7129161357879639, "learning_rate": 4.322895099269581e-06, "loss": 0.4823, "step": 13286 }, { "epoch": 0.5569726376240529, "grad_norm": 1.5067193508148193, "learning_rate": 4.322222506702324e-06, "loss": 0.4631, "step": 13287 }, { "epoch": 0.5570145562390618, "grad_norm": 3.8946847915649414, "learning_rate": 4.321549926628961e-06, "loss": 0.4858, "step": 13288 }, { "epoch": 0.5570564748540708, "grad_norm": 1.8048628568649292, "learning_rate": 4.320877359061884e-06, "loss": 0.5645, "step": 13289 }, { "epoch": 0.5570983934690797, "grad_norm": 1.6205774545669556, "learning_rate": 4.320204804013497e-06, "loss": 0.5006, "step": 13290 }, { "epoch": 0.5571403120840888, "grad_norm": 1.4710415601730347, "learning_rate": 4.319532261496194e-06, "loss": 0.4103, "step": 13291 }, { "epoch": 0.5571822306990977, "grad_norm": 1.9063915014266968, "learning_rate": 4.318859731522372e-06, "loss": 0.4667, "step": 13292 }, { "epoch": 0.5572241493141067, "grad_norm": 1.6942685842514038, "learning_rate": 4.318187214104429e-06, "loss": 0.4602, "step": 13293 }, { "epoch": 0.5572660679291156, "grad_norm": 1.5940061807632446, "learning_rate": 4.3175147092547634e-06, "loss": 0.5282, "step": 13294 }, { "epoch": 0.5573079865441246, "grad_norm": 1.6804956197738647, "learning_rate": 4.316842216985771e-06, "loss": 0.5255, "step": 13295 }, { "epoch": 0.5573499051591335, "grad_norm": 1.6990445852279663, "learning_rate": 4.316169737309846e-06, "loss": 0.4812, "step": 13296 }, { "epoch": 0.5573918237741425, "grad_norm": 2.3824143409729004, "learning_rate": 4.3154972702393885e-06, "loss": 0.4898, "step": 13297 }, { "epoch": 0.5574337423891514, "grad_norm": 1.9192883968353271, "learning_rate": 4.314824815786791e-06, "loss": 0.4822, "step": 13298 }, { "epoch": 0.5574756610041605, "grad_norm": 1.5986424684524536, "learning_rate": 4.314152373964451e-06, "loss": 0.4921, "step": 13299 }, { "epoch": 0.5575175796191694, "grad_norm": 1.8179482221603394, "learning_rate": 4.313479944784765e-06, "loss": 0.474, "step": 13300 }, { "epoch": 0.5575594982341784, "grad_norm": 1.8048657178878784, "learning_rate": 4.312807528260125e-06, "loss": 0.4995, "step": 13301 }, { "epoch": 0.5576014168491873, "grad_norm": 2.2248451709747314, "learning_rate": 4.312135124402929e-06, "loss": 0.5295, "step": 13302 }, { "epoch": 0.5576433354641963, "grad_norm": 1.7267839908599854, "learning_rate": 4.311462733225573e-06, "loss": 0.4929, "step": 13303 }, { "epoch": 0.5576852540792052, "grad_norm": 1.9092944860458374, "learning_rate": 4.310790354740448e-06, "loss": 0.4737, "step": 13304 }, { "epoch": 0.5577271726942142, "grad_norm": 2.2357866764068604, "learning_rate": 4.310117988959949e-06, "loss": 0.5187, "step": 13305 }, { "epoch": 0.5577690913092231, "grad_norm": 1.8757479190826416, "learning_rate": 4.309445635896474e-06, "loss": 0.5325, "step": 13306 }, { "epoch": 0.557811009924232, "grad_norm": 1.6602100133895874, "learning_rate": 4.308773295562412e-06, "loss": 0.4444, "step": 13307 }, { "epoch": 0.5578529285392411, "grad_norm": 1.995849609375, "learning_rate": 4.308100967970159e-06, "loss": 0.4945, "step": 13308 }, { "epoch": 0.55789484715425, "grad_norm": 2.2241671085357666, "learning_rate": 4.3074286531321085e-06, "loss": 0.4836, "step": 13309 }, { "epoch": 0.557936765769259, "grad_norm": 1.6195755004882812, "learning_rate": 4.306756351060652e-06, "loss": 0.5302, "step": 13310 }, { "epoch": 0.5579786843842679, "grad_norm": 1.7350436449050903, "learning_rate": 4.306084061768184e-06, "loss": 0.5063, "step": 13311 }, { "epoch": 0.5580206029992769, "grad_norm": 2.155956506729126, "learning_rate": 4.305411785267099e-06, "loss": 0.4855, "step": 13312 }, { "epoch": 0.5580625216142858, "grad_norm": 1.7252180576324463, "learning_rate": 4.3047395215697865e-06, "loss": 0.476, "step": 13313 }, { "epoch": 0.5581044402292948, "grad_norm": 1.8878445625305176, "learning_rate": 4.304067270688639e-06, "loss": 0.4987, "step": 13314 }, { "epoch": 0.5581463588443037, "grad_norm": 1.734419584274292, "learning_rate": 4.3033950326360505e-06, "loss": 0.5393, "step": 13315 }, { "epoch": 0.5581882774593128, "grad_norm": 1.7201707363128662, "learning_rate": 4.302722807424411e-06, "loss": 0.479, "step": 13316 }, { "epoch": 0.5582301960743217, "grad_norm": 1.791774034500122, "learning_rate": 4.302050595066112e-06, "loss": 0.5004, "step": 13317 }, { "epoch": 0.5582721146893307, "grad_norm": 1.6191290616989136, "learning_rate": 4.301378395573547e-06, "loss": 0.4664, "step": 13318 }, { "epoch": 0.5583140333043396, "grad_norm": 1.6259491443634033, "learning_rate": 4.300706208959104e-06, "loss": 0.5033, "step": 13319 }, { "epoch": 0.5583559519193486, "grad_norm": 1.7181870937347412, "learning_rate": 4.300034035235174e-06, "loss": 0.4498, "step": 13320 }, { "epoch": 0.5583978705343575, "grad_norm": 2.503319501876831, "learning_rate": 4.299361874414152e-06, "loss": 0.4864, "step": 13321 }, { "epoch": 0.5584397891493665, "grad_norm": 1.5953561067581177, "learning_rate": 4.298689726508424e-06, "loss": 0.4458, "step": 13322 }, { "epoch": 0.5584817077643754, "grad_norm": 2.6374034881591797, "learning_rate": 4.298017591530379e-06, "loss": 0.5028, "step": 13323 }, { "epoch": 0.5585236263793845, "grad_norm": 1.9078797101974487, "learning_rate": 4.297345469492413e-06, "loss": 0.5207, "step": 13324 }, { "epoch": 0.5585655449943934, "grad_norm": 2.038447380065918, "learning_rate": 4.296673360406909e-06, "loss": 0.5366, "step": 13325 }, { "epoch": 0.5586074636094024, "grad_norm": 1.4698337316513062, "learning_rate": 4.296001264286261e-06, "loss": 0.5, "step": 13326 }, { "epoch": 0.5586493822244113, "grad_norm": 1.7460180521011353, "learning_rate": 4.295329181142856e-06, "loss": 0.4935, "step": 13327 }, { "epoch": 0.5586913008394203, "grad_norm": 1.7934683561325073, "learning_rate": 4.294657110989083e-06, "loss": 0.4714, "step": 13328 }, { "epoch": 0.5587332194544292, "grad_norm": 2.392183303833008, "learning_rate": 4.293985053837331e-06, "loss": 0.4841, "step": 13329 }, { "epoch": 0.5587751380694382, "grad_norm": 2.489898443222046, "learning_rate": 4.29331300969999e-06, "loss": 0.5247, "step": 13330 }, { "epoch": 0.5588170566844471, "grad_norm": 4.753463268280029, "learning_rate": 4.292640978589445e-06, "loss": 0.4992, "step": 13331 }, { "epoch": 0.558858975299456, "grad_norm": 3.8835408687591553, "learning_rate": 4.291968960518086e-06, "loss": 0.4981, "step": 13332 }, { "epoch": 0.5589008939144651, "grad_norm": 1.6748408079147339, "learning_rate": 4.291296955498302e-06, "loss": 0.5068, "step": 13333 }, { "epoch": 0.558942812529474, "grad_norm": 1.7456390857696533, "learning_rate": 4.290624963542476e-06, "loss": 0.5329, "step": 13334 }, { "epoch": 0.558984731144483, "grad_norm": 1.757698655128479, "learning_rate": 4.289952984663001e-06, "loss": 0.4462, "step": 13335 }, { "epoch": 0.5590266497594919, "grad_norm": 1.6342540979385376, "learning_rate": 4.2892810188722586e-06, "loss": 0.4443, "step": 13336 }, { "epoch": 0.5590685683745009, "grad_norm": 1.9798295497894287, "learning_rate": 4.288609066182639e-06, "loss": 0.516, "step": 13337 }, { "epoch": 0.5591104869895098, "grad_norm": 1.772290825843811, "learning_rate": 4.287937126606529e-06, "loss": 0.4607, "step": 13338 }, { "epoch": 0.5591524056045188, "grad_norm": 2.2370383739471436, "learning_rate": 4.287265200156313e-06, "loss": 0.4446, "step": 13339 }, { "epoch": 0.5591943242195277, "grad_norm": 1.6368207931518555, "learning_rate": 4.286593286844376e-06, "loss": 0.5117, "step": 13340 }, { "epoch": 0.5592362428345368, "grad_norm": 1.8642271757125854, "learning_rate": 4.285921386683108e-06, "loss": 0.5223, "step": 13341 }, { "epoch": 0.5592781614495457, "grad_norm": 1.8632892370224, "learning_rate": 4.285249499684891e-06, "loss": 0.4802, "step": 13342 }, { "epoch": 0.5593200800645547, "grad_norm": 1.7102867364883423, "learning_rate": 4.28457762586211e-06, "loss": 0.4783, "step": 13343 }, { "epoch": 0.5593619986795636, "grad_norm": 1.9989789724349976, "learning_rate": 4.283905765227154e-06, "loss": 0.5486, "step": 13344 }, { "epoch": 0.5594039172945726, "grad_norm": 1.7414884567260742, "learning_rate": 4.283233917792404e-06, "loss": 0.5048, "step": 13345 }, { "epoch": 0.5594458359095815, "grad_norm": 1.603126883506775, "learning_rate": 4.2825620835702455e-06, "loss": 0.4531, "step": 13346 }, { "epoch": 0.5594877545245905, "grad_norm": 1.6869651079177856, "learning_rate": 4.2818902625730644e-06, "loss": 0.5161, "step": 13347 }, { "epoch": 0.5595296731395994, "grad_norm": 1.908068299293518, "learning_rate": 4.281218454813241e-06, "loss": 0.4686, "step": 13348 }, { "epoch": 0.5595715917546085, "grad_norm": 1.6908336877822876, "learning_rate": 4.2805466603031635e-06, "loss": 0.486, "step": 13349 }, { "epoch": 0.5596135103696174, "grad_norm": 3.215810775756836, "learning_rate": 4.279874879055214e-06, "loss": 0.4605, "step": 13350 }, { "epoch": 0.5596554289846264, "grad_norm": 1.7213037014007568, "learning_rate": 4.2792031110817755e-06, "loss": 0.5026, "step": 13351 }, { "epoch": 0.5596973475996353, "grad_norm": 2.0442299842834473, "learning_rate": 4.278531356395229e-06, "loss": 0.5199, "step": 13352 }, { "epoch": 0.5597392662146443, "grad_norm": 2.2281887531280518, "learning_rate": 4.277859615007963e-06, "loss": 0.4527, "step": 13353 }, { "epoch": 0.5597811848296532, "grad_norm": 2.258732795715332, "learning_rate": 4.277187886932354e-06, "loss": 0.4994, "step": 13354 }, { "epoch": 0.5598231034446622, "grad_norm": 2.038241386413574, "learning_rate": 4.276516172180788e-06, "loss": 0.4872, "step": 13355 }, { "epoch": 0.5598650220596711, "grad_norm": 1.5208168029785156, "learning_rate": 4.275844470765648e-06, "loss": 0.4691, "step": 13356 }, { "epoch": 0.55990694067468, "grad_norm": 2.7411155700683594, "learning_rate": 4.275172782699311e-06, "loss": 0.5234, "step": 13357 }, { "epoch": 0.5599488592896891, "grad_norm": 1.7261037826538086, "learning_rate": 4.274501107994162e-06, "loss": 0.4546, "step": 13358 }, { "epoch": 0.559990777904698, "grad_norm": 1.9479577541351318, "learning_rate": 4.273829446662584e-06, "loss": 0.5107, "step": 13359 }, { "epoch": 0.560032696519707, "grad_norm": 1.633506178855896, "learning_rate": 4.273157798716956e-06, "loss": 0.512, "step": 13360 }, { "epoch": 0.5600746151347159, "grad_norm": 4.3300676345825195, "learning_rate": 4.272486164169657e-06, "loss": 0.5224, "step": 13361 }, { "epoch": 0.5601165337497249, "grad_norm": 1.6792455911636353, "learning_rate": 4.271814543033073e-06, "loss": 0.5345, "step": 13362 }, { "epoch": 0.5601584523647338, "grad_norm": 1.830784797668457, "learning_rate": 4.271142935319578e-06, "loss": 0.4967, "step": 13363 }, { "epoch": 0.5602003709797428, "grad_norm": 1.8656693696975708, "learning_rate": 4.270471341041557e-06, "loss": 0.4878, "step": 13364 }, { "epoch": 0.5602422895947518, "grad_norm": 1.8920003175735474, "learning_rate": 4.269799760211388e-06, "loss": 0.5161, "step": 13365 }, { "epoch": 0.5602842082097608, "grad_norm": 1.6140002012252808, "learning_rate": 4.269128192841452e-06, "loss": 0.4918, "step": 13366 }, { "epoch": 0.5603261268247697, "grad_norm": 1.7441614866256714, "learning_rate": 4.268456638944125e-06, "loss": 0.4947, "step": 13367 }, { "epoch": 0.5603680454397787, "grad_norm": 5.066776275634766, "learning_rate": 4.267785098531792e-06, "loss": 0.4428, "step": 13368 }, { "epoch": 0.5604099640547876, "grad_norm": 1.9764310121536255, "learning_rate": 4.267113571616826e-06, "loss": 0.4761, "step": 13369 }, { "epoch": 0.5604518826697966, "grad_norm": 2.1727259159088135, "learning_rate": 4.266442058211608e-06, "loss": 0.5016, "step": 13370 }, { "epoch": 0.5604938012848055, "grad_norm": 1.8194923400878906, "learning_rate": 4.265770558328518e-06, "loss": 0.5355, "step": 13371 }, { "epoch": 0.5605357198998145, "grad_norm": 1.5671244859695435, "learning_rate": 4.2650990719799305e-06, "loss": 0.5102, "step": 13372 }, { "epoch": 0.5605776385148235, "grad_norm": 1.725490927696228, "learning_rate": 4.264427599178227e-06, "loss": 0.4613, "step": 13373 }, { "epoch": 0.5606195571298325, "grad_norm": 1.584019422531128, "learning_rate": 4.263756139935784e-06, "loss": 0.5248, "step": 13374 }, { "epoch": 0.5606614757448414, "grad_norm": 1.799149751663208, "learning_rate": 4.263084694264979e-06, "loss": 0.5184, "step": 13375 }, { "epoch": 0.5607033943598504, "grad_norm": 5.931929588317871, "learning_rate": 4.262413262178187e-06, "loss": 0.4923, "step": 13376 }, { "epoch": 0.5607453129748593, "grad_norm": 1.7144484519958496, "learning_rate": 4.2617418436877895e-06, "loss": 0.4731, "step": 13377 }, { "epoch": 0.5607872315898683, "grad_norm": 1.6521551609039307, "learning_rate": 4.261070438806159e-06, "loss": 0.4926, "step": 13378 }, { "epoch": 0.5608291502048772, "grad_norm": 1.836776852607727, "learning_rate": 4.260399047545673e-06, "loss": 0.5635, "step": 13379 }, { "epoch": 0.5608710688198862, "grad_norm": 2.100519895553589, "learning_rate": 4.25972766991871e-06, "loss": 0.4759, "step": 13380 }, { "epoch": 0.5609129874348951, "grad_norm": 1.7969251871109009, "learning_rate": 4.2590563059376416e-06, "loss": 0.5203, "step": 13381 }, { "epoch": 0.5609549060499041, "grad_norm": 2.813976526260376, "learning_rate": 4.258384955614846e-06, "loss": 0.5213, "step": 13382 }, { "epoch": 0.5609968246649131, "grad_norm": 1.6515862941741943, "learning_rate": 4.257713618962701e-06, "loss": 0.5024, "step": 13383 }, { "epoch": 0.561038743279922, "grad_norm": 1.782731056213379, "learning_rate": 4.257042295993578e-06, "loss": 0.4892, "step": 13384 }, { "epoch": 0.561080661894931, "grad_norm": 1.788008213043213, "learning_rate": 4.256370986719852e-06, "loss": 0.4952, "step": 13385 }, { "epoch": 0.5611225805099399, "grad_norm": 1.8303070068359375, "learning_rate": 4.2556996911539e-06, "loss": 0.4819, "step": 13386 }, { "epoch": 0.5611644991249489, "grad_norm": 1.889269471168518, "learning_rate": 4.2550284093080955e-06, "loss": 0.4823, "step": 13387 }, { "epoch": 0.5612064177399578, "grad_norm": 2.390979528427124, "learning_rate": 4.254357141194811e-06, "loss": 0.4871, "step": 13388 }, { "epoch": 0.5612483363549668, "grad_norm": 2.982835531234741, "learning_rate": 4.253685886826424e-06, "loss": 0.5031, "step": 13389 }, { "epoch": 0.5612902549699758, "grad_norm": 1.709765076637268, "learning_rate": 4.253014646215304e-06, "loss": 0.525, "step": 13390 }, { "epoch": 0.5613321735849848, "grad_norm": 3.5475516319274902, "learning_rate": 4.252343419373825e-06, "loss": 0.4978, "step": 13391 }, { "epoch": 0.5613740921999937, "grad_norm": 1.6807546615600586, "learning_rate": 4.251672206314366e-06, "loss": 0.509, "step": 13392 }, { "epoch": 0.5614160108150027, "grad_norm": 1.6250967979431152, "learning_rate": 4.251001007049293e-06, "loss": 0.4678, "step": 13393 }, { "epoch": 0.5614579294300116, "grad_norm": 1.7102057933807373, "learning_rate": 4.2503298215909785e-06, "loss": 0.5138, "step": 13394 }, { "epoch": 0.5614998480450206, "grad_norm": 1.8516165018081665, "learning_rate": 4.249658649951801e-06, "loss": 0.488, "step": 13395 }, { "epoch": 0.5615417666600295, "grad_norm": 1.8340033292770386, "learning_rate": 4.248987492144126e-06, "loss": 0.4952, "step": 13396 }, { "epoch": 0.5615836852750385, "grad_norm": 2.069741725921631, "learning_rate": 4.248316348180331e-06, "loss": 0.4765, "step": 13397 }, { "epoch": 0.5616256038900475, "grad_norm": 1.683626651763916, "learning_rate": 4.247645218072785e-06, "loss": 0.5264, "step": 13398 }, { "epoch": 0.5616675225050565, "grad_norm": 1.5646835565567017, "learning_rate": 4.246974101833857e-06, "loss": 0.4731, "step": 13399 }, { "epoch": 0.5617094411200654, "grad_norm": 1.7212154865264893, "learning_rate": 4.246302999475921e-06, "loss": 0.5488, "step": 13400 }, { "epoch": 0.5617513597350744, "grad_norm": 1.6257299184799194, "learning_rate": 4.2456319110113484e-06, "loss": 0.496, "step": 13401 }, { "epoch": 0.5617932783500833, "grad_norm": 1.6375994682312012, "learning_rate": 4.244960836452508e-06, "loss": 0.4962, "step": 13402 }, { "epoch": 0.5618351969650923, "grad_norm": 1.5987811088562012, "learning_rate": 4.244289775811769e-06, "loss": 0.5096, "step": 13403 }, { "epoch": 0.5618771155801012, "grad_norm": 1.6984705924987793, "learning_rate": 4.2436187291015076e-06, "loss": 0.4873, "step": 13404 }, { "epoch": 0.5619190341951102, "grad_norm": 1.6310497522354126, "learning_rate": 4.242947696334086e-06, "loss": 0.5061, "step": 13405 }, { "epoch": 0.5619609528101192, "grad_norm": 1.5223249197006226, "learning_rate": 4.242276677521877e-06, "loss": 0.464, "step": 13406 }, { "epoch": 0.5620028714251281, "grad_norm": 1.7574175596237183, "learning_rate": 4.241605672677252e-06, "loss": 0.5057, "step": 13407 }, { "epoch": 0.5620447900401371, "grad_norm": 1.6519036293029785, "learning_rate": 4.240934681812576e-06, "loss": 0.4824, "step": 13408 }, { "epoch": 0.562086708655146, "grad_norm": 1.9260375499725342, "learning_rate": 4.240263704940221e-06, "loss": 0.4972, "step": 13409 }, { "epoch": 0.562128627270155, "grad_norm": 1.9253857135772705, "learning_rate": 4.239592742072553e-06, "loss": 0.4989, "step": 13410 }, { "epoch": 0.5621705458851639, "grad_norm": 1.6350990533828735, "learning_rate": 4.238921793221942e-06, "loss": 0.4935, "step": 13411 }, { "epoch": 0.5622124645001729, "grad_norm": 1.5658537149429321, "learning_rate": 4.238250858400757e-06, "loss": 0.4985, "step": 13412 }, { "epoch": 0.5622543831151818, "grad_norm": 1.5947290658950806, "learning_rate": 4.2375799376213625e-06, "loss": 0.4621, "step": 13413 }, { "epoch": 0.5622963017301909, "grad_norm": 1.594697117805481, "learning_rate": 4.236909030896128e-06, "loss": 0.4503, "step": 13414 }, { "epoch": 0.5623382203451998, "grad_norm": 1.539076328277588, "learning_rate": 4.236238138237422e-06, "loss": 0.4905, "step": 13415 }, { "epoch": 0.5623801389602088, "grad_norm": 2.678173303604126, "learning_rate": 4.235567259657608e-06, "loss": 0.5104, "step": 13416 }, { "epoch": 0.5624220575752177, "grad_norm": 1.8590614795684814, "learning_rate": 4.2348963951690555e-06, "loss": 0.5109, "step": 13417 }, { "epoch": 0.5624639761902267, "grad_norm": 1.6636691093444824, "learning_rate": 4.234225544784133e-06, "loss": 0.4656, "step": 13418 }, { "epoch": 0.5625058948052356, "grad_norm": 2.0159473419189453, "learning_rate": 4.233554708515199e-06, "loss": 0.5813, "step": 13419 }, { "epoch": 0.5625478134202446, "grad_norm": 1.7572664022445679, "learning_rate": 4.2328838863746265e-06, "loss": 0.5253, "step": 13420 }, { "epoch": 0.5625897320352535, "grad_norm": 1.664898157119751, "learning_rate": 4.232213078374781e-06, "loss": 0.5155, "step": 13421 }, { "epoch": 0.5626316506502625, "grad_norm": 1.7449980974197388, "learning_rate": 4.231542284528025e-06, "loss": 0.5353, "step": 13422 }, { "epoch": 0.5626735692652715, "grad_norm": 1.7379140853881836, "learning_rate": 4.2308715048467225e-06, "loss": 0.5063, "step": 13423 }, { "epoch": 0.5627154878802805, "grad_norm": 1.4754668474197388, "learning_rate": 4.230200739343243e-06, "loss": 0.4502, "step": 13424 }, { "epoch": 0.5627574064952894, "grad_norm": 1.8024063110351562, "learning_rate": 4.2295299880299475e-06, "loss": 0.4889, "step": 13425 }, { "epoch": 0.5627993251102984, "grad_norm": 1.6431077718734741, "learning_rate": 4.2288592509192015e-06, "loss": 0.4606, "step": 13426 }, { "epoch": 0.5628412437253073, "grad_norm": 1.6928952932357788, "learning_rate": 4.22818852802337e-06, "loss": 0.5586, "step": 13427 }, { "epoch": 0.5628831623403163, "grad_norm": 4.728835105895996, "learning_rate": 4.227517819354815e-06, "loss": 0.4861, "step": 13428 }, { "epoch": 0.5629250809553252, "grad_norm": 2.1797351837158203, "learning_rate": 4.2268471249259e-06, "loss": 0.554, "step": 13429 }, { "epoch": 0.5629669995703342, "grad_norm": 1.5913383960723877, "learning_rate": 4.226176444748993e-06, "loss": 0.493, "step": 13430 }, { "epoch": 0.5630089181853432, "grad_norm": 1.7199429273605347, "learning_rate": 4.225505778836451e-06, "loss": 0.4312, "step": 13431 }, { "epoch": 0.5630508368003521, "grad_norm": 1.8313015699386597, "learning_rate": 4.224835127200639e-06, "loss": 0.5508, "step": 13432 }, { "epoch": 0.5630927554153611, "grad_norm": 1.7204647064208984, "learning_rate": 4.224164489853921e-06, "loss": 0.4893, "step": 13433 }, { "epoch": 0.56313467403037, "grad_norm": 1.7309945821762085, "learning_rate": 4.223493866808656e-06, "loss": 0.4693, "step": 13434 }, { "epoch": 0.563176592645379, "grad_norm": 2.0969314575195312, "learning_rate": 4.22282325807721e-06, "loss": 0.4938, "step": 13435 }, { "epoch": 0.5632185112603879, "grad_norm": 1.8281664848327637, "learning_rate": 4.222152663671944e-06, "loss": 0.4682, "step": 13436 }, { "epoch": 0.5632604298753969, "grad_norm": 3.518787384033203, "learning_rate": 4.221482083605216e-06, "loss": 0.467, "step": 13437 }, { "epoch": 0.5633023484904058, "grad_norm": 1.848792314529419, "learning_rate": 4.2208115178893885e-06, "loss": 0.4912, "step": 13438 }, { "epoch": 0.5633442671054149, "grad_norm": 1.6307893991470337, "learning_rate": 4.220140966536827e-06, "loss": 0.4465, "step": 13439 }, { "epoch": 0.5633861857204238, "grad_norm": 2.1003992557525635, "learning_rate": 4.2194704295598864e-06, "loss": 0.5359, "step": 13440 }, { "epoch": 0.5634281043354328, "grad_norm": 3.05061936378479, "learning_rate": 4.218799906970929e-06, "loss": 0.5111, "step": 13441 }, { "epoch": 0.5634700229504417, "grad_norm": 1.65850830078125, "learning_rate": 4.2181293987823175e-06, "loss": 0.4854, "step": 13442 }, { "epoch": 0.5635119415654507, "grad_norm": 1.7726161479949951, "learning_rate": 4.217458905006407e-06, "loss": 0.4607, "step": 13443 }, { "epoch": 0.5635538601804596, "grad_norm": 1.7982734441757202, "learning_rate": 4.216788425655561e-06, "loss": 0.5886, "step": 13444 }, { "epoch": 0.5635957787954686, "grad_norm": 1.9038653373718262, "learning_rate": 4.216117960742139e-06, "loss": 0.5211, "step": 13445 }, { "epoch": 0.5636376974104775, "grad_norm": 1.8403960466384888, "learning_rate": 4.215447510278496e-06, "loss": 0.4927, "step": 13446 }, { "epoch": 0.5636796160254866, "grad_norm": 2.144618511199951, "learning_rate": 4.2147770742769945e-06, "loss": 0.5405, "step": 13447 }, { "epoch": 0.5637215346404955, "grad_norm": 1.7365658283233643, "learning_rate": 4.2141066527499935e-06, "loss": 0.5122, "step": 13448 }, { "epoch": 0.5637634532555045, "grad_norm": 1.9205347299575806, "learning_rate": 4.2134362457098494e-06, "loss": 0.5324, "step": 13449 }, { "epoch": 0.5638053718705134, "grad_norm": 1.9617384672164917, "learning_rate": 4.212765853168919e-06, "loss": 0.5617, "step": 13450 }, { "epoch": 0.5638472904855224, "grad_norm": 1.6235958337783813, "learning_rate": 4.2120954751395644e-06, "loss": 0.5148, "step": 13451 }, { "epoch": 0.5638892091005313, "grad_norm": 2.8351027965545654, "learning_rate": 4.211425111634137e-06, "loss": 0.4568, "step": 13452 }, { "epoch": 0.5639311277155403, "grad_norm": 1.4543793201446533, "learning_rate": 4.210754762665e-06, "loss": 0.4452, "step": 13453 }, { "epoch": 0.5639730463305492, "grad_norm": 1.5189038515090942, "learning_rate": 4.210084428244509e-06, "loss": 0.4674, "step": 13454 }, { "epoch": 0.5640149649455583, "grad_norm": 1.9736131429672241, "learning_rate": 4.209414108385017e-06, "loss": 0.5364, "step": 13455 }, { "epoch": 0.5640568835605672, "grad_norm": 2.151937484741211, "learning_rate": 4.208743803098883e-06, "loss": 0.5273, "step": 13456 }, { "epoch": 0.5640988021755761, "grad_norm": 5.510616779327393, "learning_rate": 4.208073512398465e-06, "loss": 0.4899, "step": 13457 }, { "epoch": 0.5641407207905851, "grad_norm": 1.7206213474273682, "learning_rate": 4.207403236296117e-06, "loss": 0.5402, "step": 13458 }, { "epoch": 0.564182639405594, "grad_norm": 1.7941076755523682, "learning_rate": 4.206732974804192e-06, "loss": 0.4413, "step": 13459 }, { "epoch": 0.564224558020603, "grad_norm": 1.7818268537521362, "learning_rate": 4.206062727935051e-06, "loss": 0.4787, "step": 13460 }, { "epoch": 0.5642664766356119, "grad_norm": 1.8360342979431152, "learning_rate": 4.205392495701043e-06, "loss": 0.4764, "step": 13461 }, { "epoch": 0.5643083952506209, "grad_norm": 1.7079331874847412, "learning_rate": 4.204722278114528e-06, "loss": 0.501, "step": 13462 }, { "epoch": 0.5643503138656298, "grad_norm": 2.015592098236084, "learning_rate": 4.2040520751878575e-06, "loss": 0.4643, "step": 13463 }, { "epoch": 0.5643922324806389, "grad_norm": 2.5575332641601562, "learning_rate": 4.203381886933387e-06, "loss": 0.5347, "step": 13464 }, { "epoch": 0.5644341510956478, "grad_norm": 2.7097177505493164, "learning_rate": 4.2027117133634695e-06, "loss": 0.5167, "step": 13465 }, { "epoch": 0.5644760697106568, "grad_norm": 1.5395655632019043, "learning_rate": 4.202041554490461e-06, "loss": 0.5093, "step": 13466 }, { "epoch": 0.5645179883256657, "grad_norm": 2.093737840652466, "learning_rate": 4.201371410326711e-06, "loss": 0.5239, "step": 13467 }, { "epoch": 0.5645599069406747, "grad_norm": 1.7020477056503296, "learning_rate": 4.200701280884576e-06, "loss": 0.5023, "step": 13468 }, { "epoch": 0.5646018255556836, "grad_norm": 1.6732020378112793, "learning_rate": 4.200031166176409e-06, "loss": 0.4565, "step": 13469 }, { "epoch": 0.5646437441706926, "grad_norm": 3.8324027061462402, "learning_rate": 4.19936106621456e-06, "loss": 0.5245, "step": 13470 }, { "epoch": 0.5646856627857015, "grad_norm": 2.316240072250366, "learning_rate": 4.198690981011383e-06, "loss": 0.4707, "step": 13471 }, { "epoch": 0.5647275814007106, "grad_norm": 1.9466646909713745, "learning_rate": 4.198020910579232e-06, "loss": 0.4618, "step": 13472 }, { "epoch": 0.5647695000157195, "grad_norm": 1.7522341012954712, "learning_rate": 4.197350854930454e-06, "loss": 0.5038, "step": 13473 }, { "epoch": 0.5648114186307285, "grad_norm": 1.7860549688339233, "learning_rate": 4.196680814077404e-06, "loss": 0.5275, "step": 13474 }, { "epoch": 0.5648533372457374, "grad_norm": 1.4920018911361694, "learning_rate": 4.1960107880324345e-06, "loss": 0.4166, "step": 13475 }, { "epoch": 0.5648952558607464, "grad_norm": 1.621281385421753, "learning_rate": 4.195340776807892e-06, "loss": 0.5057, "step": 13476 }, { "epoch": 0.5649371744757553, "grad_norm": 1.722206711769104, "learning_rate": 4.194670780416132e-06, "loss": 0.4674, "step": 13477 }, { "epoch": 0.5649790930907643, "grad_norm": 1.7177906036376953, "learning_rate": 4.194000798869503e-06, "loss": 0.455, "step": 13478 }, { "epoch": 0.5650210117057732, "grad_norm": 1.8124279975891113, "learning_rate": 4.193330832180354e-06, "loss": 0.52, "step": 13479 }, { "epoch": 0.5650629303207823, "grad_norm": 1.784026026725769, "learning_rate": 4.192660880361034e-06, "loss": 0.5193, "step": 13480 }, { "epoch": 0.5651048489357912, "grad_norm": 1.985954999923706, "learning_rate": 4.191990943423898e-06, "loss": 0.4612, "step": 13481 }, { "epoch": 0.5651467675508001, "grad_norm": 2.648098945617676, "learning_rate": 4.19132102138129e-06, "loss": 0.5164, "step": 13482 }, { "epoch": 0.5651886861658091, "grad_norm": 1.6669448614120483, "learning_rate": 4.1906511142455615e-06, "loss": 0.4794, "step": 13483 }, { "epoch": 0.565230604780818, "grad_norm": 1.8011295795440674, "learning_rate": 4.189981222029062e-06, "loss": 0.5177, "step": 13484 }, { "epoch": 0.565272523395827, "grad_norm": 1.9837288856506348, "learning_rate": 4.189311344744137e-06, "loss": 0.5491, "step": 13485 }, { "epoch": 0.5653144420108359, "grad_norm": 1.7018659114837646, "learning_rate": 4.188641482403138e-06, "loss": 0.5031, "step": 13486 }, { "epoch": 0.5653563606258449, "grad_norm": 1.6121559143066406, "learning_rate": 4.187971635018411e-06, "loss": 0.4708, "step": 13487 }, { "epoch": 0.5653982792408538, "grad_norm": 1.7782812118530273, "learning_rate": 4.187301802602303e-06, "loss": 0.5383, "step": 13488 }, { "epoch": 0.5654401978558629, "grad_norm": 1.7467955350875854, "learning_rate": 4.1866319851671655e-06, "loss": 0.4844, "step": 13489 }, { "epoch": 0.5654821164708718, "grad_norm": 1.8216172456741333, "learning_rate": 4.18596218272534e-06, "loss": 0.5076, "step": 13490 }, { "epoch": 0.5655240350858808, "grad_norm": 1.7461570501327515, "learning_rate": 4.185292395289178e-06, "loss": 0.4675, "step": 13491 }, { "epoch": 0.5655659537008897, "grad_norm": 1.6575126647949219, "learning_rate": 4.184622622871025e-06, "loss": 0.4822, "step": 13492 }, { "epoch": 0.5656078723158987, "grad_norm": 1.749014973640442, "learning_rate": 4.183952865483225e-06, "loss": 0.4793, "step": 13493 }, { "epoch": 0.5656497909309076, "grad_norm": 1.8648884296417236, "learning_rate": 4.183283123138126e-06, "loss": 0.5228, "step": 13494 }, { "epoch": 0.5656917095459166, "grad_norm": 1.890565276145935, "learning_rate": 4.182613395848076e-06, "loss": 0.4669, "step": 13495 }, { "epoch": 0.5657336281609255, "grad_norm": 1.6971766948699951, "learning_rate": 4.181943683625416e-06, "loss": 0.5069, "step": 13496 }, { "epoch": 0.5657755467759346, "grad_norm": 1.73418128490448, "learning_rate": 4.1812739864824915e-06, "loss": 0.5331, "step": 13497 }, { "epoch": 0.5658174653909435, "grad_norm": 2.157658815383911, "learning_rate": 4.180604304431653e-06, "loss": 0.5466, "step": 13498 }, { "epoch": 0.5658593840059525, "grad_norm": 1.5350122451782227, "learning_rate": 4.179934637485238e-06, "loss": 0.4977, "step": 13499 }, { "epoch": 0.5659013026209614, "grad_norm": 1.6670101881027222, "learning_rate": 4.179264985655595e-06, "loss": 0.543, "step": 13500 }, { "epoch": 0.5659432212359704, "grad_norm": 1.9021661281585693, "learning_rate": 4.17859534895507e-06, "loss": 0.4906, "step": 13501 }, { "epoch": 0.5659851398509793, "grad_norm": 1.6046956777572632, "learning_rate": 4.1779257273960016e-06, "loss": 0.507, "step": 13502 }, { "epoch": 0.5660270584659883, "grad_norm": 1.5015462636947632, "learning_rate": 4.177256120990736e-06, "loss": 0.5104, "step": 13503 }, { "epoch": 0.5660689770809972, "grad_norm": 1.6116687059402466, "learning_rate": 4.176586529751618e-06, "loss": 0.5248, "step": 13504 }, { "epoch": 0.5661108956960063, "grad_norm": 1.9620330333709717, "learning_rate": 4.1759169536909876e-06, "loss": 0.4882, "step": 13505 }, { "epoch": 0.5661528143110152, "grad_norm": 1.8652429580688477, "learning_rate": 4.1752473928211885e-06, "loss": 0.5074, "step": 13506 }, { "epoch": 0.5661947329260241, "grad_norm": 1.6972893476486206, "learning_rate": 4.174577847154565e-06, "loss": 0.4756, "step": 13507 }, { "epoch": 0.5662366515410331, "grad_norm": 1.8164618015289307, "learning_rate": 4.173908316703457e-06, "loss": 0.5111, "step": 13508 }, { "epoch": 0.566278570156042, "grad_norm": 1.886207938194275, "learning_rate": 4.173238801480207e-06, "loss": 0.4652, "step": 13509 }, { "epoch": 0.566320488771051, "grad_norm": 1.7389862537384033, "learning_rate": 4.172569301497157e-06, "loss": 0.4806, "step": 13510 }, { "epoch": 0.5663624073860599, "grad_norm": 4.032443046569824, "learning_rate": 4.171899816766648e-06, "loss": 0.4996, "step": 13511 }, { "epoch": 0.5664043260010689, "grad_norm": 2.0409648418426514, "learning_rate": 4.17123034730102e-06, "loss": 0.5444, "step": 13512 }, { "epoch": 0.5664462446160778, "grad_norm": 1.9194332361221313, "learning_rate": 4.170560893112617e-06, "loss": 0.4998, "step": 13513 }, { "epoch": 0.5664881632310869, "grad_norm": 1.9681727886199951, "learning_rate": 4.169891454213776e-06, "loss": 0.4125, "step": 13514 }, { "epoch": 0.5665300818460958, "grad_norm": 1.847852110862732, "learning_rate": 4.1692220306168385e-06, "loss": 0.4721, "step": 13515 }, { "epoch": 0.5665720004611048, "grad_norm": 5.996267318725586, "learning_rate": 4.168552622334146e-06, "loss": 0.4949, "step": 13516 }, { "epoch": 0.5666139190761137, "grad_norm": 8.082171440124512, "learning_rate": 4.167883229378034e-06, "loss": 0.4518, "step": 13517 }, { "epoch": 0.5666558376911227, "grad_norm": 1.8221312761306763, "learning_rate": 4.167213851760843e-06, "loss": 0.4544, "step": 13518 }, { "epoch": 0.5666977563061316, "grad_norm": 2.1003947257995605, "learning_rate": 4.166544489494917e-06, "loss": 0.5067, "step": 13519 }, { "epoch": 0.5667396749211406, "grad_norm": 1.8408079147338867, "learning_rate": 4.16587514259259e-06, "loss": 0.5216, "step": 13520 }, { "epoch": 0.5667815935361495, "grad_norm": 1.901093602180481, "learning_rate": 4.1652058110662e-06, "loss": 0.542, "step": 13521 }, { "epoch": 0.5668235121511586, "grad_norm": 1.8337090015411377, "learning_rate": 4.16453649492809e-06, "loss": 0.4752, "step": 13522 }, { "epoch": 0.5668654307661675, "grad_norm": 1.9196263551712036, "learning_rate": 4.163867194190592e-06, "loss": 0.5135, "step": 13523 }, { "epoch": 0.5669073493811765, "grad_norm": 1.765282154083252, "learning_rate": 4.163197908866047e-06, "loss": 0.5622, "step": 13524 }, { "epoch": 0.5669492679961854, "grad_norm": 2.460829496383667, "learning_rate": 4.162528638966793e-06, "loss": 0.489, "step": 13525 }, { "epoch": 0.5669911866111944, "grad_norm": 1.664656162261963, "learning_rate": 4.161859384505164e-06, "loss": 0.5037, "step": 13526 }, { "epoch": 0.5670331052262033, "grad_norm": 2.2301182746887207, "learning_rate": 4.161190145493498e-06, "loss": 0.5402, "step": 13527 }, { "epoch": 0.5670750238412123, "grad_norm": 1.8446216583251953, "learning_rate": 4.1605209219441345e-06, "loss": 0.5029, "step": 13528 }, { "epoch": 0.5671169424562212, "grad_norm": 1.9942761659622192, "learning_rate": 4.159851713869406e-06, "loss": 0.5303, "step": 13529 }, { "epoch": 0.5671588610712303, "grad_norm": 1.644755482673645, "learning_rate": 4.159182521281648e-06, "loss": 0.4719, "step": 13530 }, { "epoch": 0.5672007796862392, "grad_norm": 2.300570487976074, "learning_rate": 4.158513344193202e-06, "loss": 0.4955, "step": 13531 }, { "epoch": 0.5672426983012481, "grad_norm": 1.6514724493026733, "learning_rate": 4.157844182616395e-06, "loss": 0.4525, "step": 13532 }, { "epoch": 0.5672846169162571, "grad_norm": 1.5139356851577759, "learning_rate": 4.157175036563568e-06, "loss": 0.4706, "step": 13533 }, { "epoch": 0.567326535531266, "grad_norm": 2.0444540977478027, "learning_rate": 4.156505906047055e-06, "loss": 0.4811, "step": 13534 }, { "epoch": 0.567368454146275, "grad_norm": 1.665917992591858, "learning_rate": 4.155836791079188e-06, "loss": 0.5153, "step": 13535 }, { "epoch": 0.5674103727612839, "grad_norm": 1.7077914476394653, "learning_rate": 4.155167691672301e-06, "loss": 0.5068, "step": 13536 }, { "epoch": 0.5674522913762929, "grad_norm": 1.6292465925216675, "learning_rate": 4.154498607838734e-06, "loss": 0.4836, "step": 13537 }, { "epoch": 0.5674942099913018, "grad_norm": 1.7452466487884521, "learning_rate": 4.153829539590814e-06, "loss": 0.5136, "step": 13538 }, { "epoch": 0.5675361286063109, "grad_norm": 1.7227689027786255, "learning_rate": 4.153160486940876e-06, "loss": 0.5443, "step": 13539 }, { "epoch": 0.5675780472213198, "grad_norm": 1.5776208639144897, "learning_rate": 4.152491449901256e-06, "loss": 0.47, "step": 13540 }, { "epoch": 0.5676199658363288, "grad_norm": 1.6023764610290527, "learning_rate": 4.151822428484282e-06, "loss": 0.4576, "step": 13541 }, { "epoch": 0.5676618844513377, "grad_norm": 1.7429078817367554, "learning_rate": 4.15115342270229e-06, "loss": 0.5124, "step": 13542 }, { "epoch": 0.5677038030663467, "grad_norm": 2.0322821140289307, "learning_rate": 4.150484432567613e-06, "loss": 0.5267, "step": 13543 }, { "epoch": 0.5677457216813556, "grad_norm": 2.028149366378784, "learning_rate": 4.149815458092578e-06, "loss": 0.4811, "step": 13544 }, { "epoch": 0.5677876402963646, "grad_norm": 1.7228314876556396, "learning_rate": 4.149146499289519e-06, "loss": 0.5601, "step": 13545 }, { "epoch": 0.5678295589113735, "grad_norm": 1.906232476234436, "learning_rate": 4.148477556170772e-06, "loss": 0.4832, "step": 13546 }, { "epoch": 0.5678714775263826, "grad_norm": 1.8117440938949585, "learning_rate": 4.147808628748662e-06, "loss": 0.5326, "step": 13547 }, { "epoch": 0.5679133961413915, "grad_norm": 1.75026273727417, "learning_rate": 4.147139717035521e-06, "loss": 0.4164, "step": 13548 }, { "epoch": 0.5679553147564005, "grad_norm": 2.176123857498169, "learning_rate": 4.1464708210436814e-06, "loss": 0.5545, "step": 13549 }, { "epoch": 0.5679972333714094, "grad_norm": 1.880239725112915, "learning_rate": 4.1458019407854715e-06, "loss": 0.4998, "step": 13550 }, { "epoch": 0.5680391519864184, "grad_norm": 1.9930757284164429, "learning_rate": 4.145133076273222e-06, "loss": 0.5083, "step": 13551 }, { "epoch": 0.5680810706014273, "grad_norm": 1.462550401687622, "learning_rate": 4.144464227519264e-06, "loss": 0.482, "step": 13552 }, { "epoch": 0.5681229892164363, "grad_norm": 2.0025556087493896, "learning_rate": 4.143795394535923e-06, "loss": 0.4995, "step": 13553 }, { "epoch": 0.5681649078314452, "grad_norm": 1.6985136270523071, "learning_rate": 4.1431265773355304e-06, "loss": 0.4949, "step": 13554 }, { "epoch": 0.5682068264464543, "grad_norm": 1.9000413417816162, "learning_rate": 4.142457775930416e-06, "loss": 0.5342, "step": 13555 }, { "epoch": 0.5682487450614632, "grad_norm": 1.6486637592315674, "learning_rate": 4.141788990332908e-06, "loss": 0.4676, "step": 13556 }, { "epoch": 0.5682906636764721, "grad_norm": 2.0402724742889404, "learning_rate": 4.14112022055533e-06, "loss": 0.4918, "step": 13557 }, { "epoch": 0.5683325822914811, "grad_norm": 2.0040838718414307, "learning_rate": 4.140451466610017e-06, "loss": 0.5167, "step": 13558 }, { "epoch": 0.56837450090649, "grad_norm": 1.6516252756118774, "learning_rate": 4.13978272850929e-06, "loss": 0.4544, "step": 13559 }, { "epoch": 0.568416419521499, "grad_norm": 2.0447568893432617, "learning_rate": 4.139114006265481e-06, "loss": 0.4895, "step": 13560 }, { "epoch": 0.5684583381365079, "grad_norm": 1.838876724243164, "learning_rate": 4.1384452998909145e-06, "loss": 0.5022, "step": 13561 }, { "epoch": 0.5685002567515169, "grad_norm": 1.594863772392273, "learning_rate": 4.137776609397918e-06, "loss": 0.4924, "step": 13562 }, { "epoch": 0.5685421753665258, "grad_norm": 1.8872877359390259, "learning_rate": 4.137107934798819e-06, "loss": 0.5215, "step": 13563 }, { "epoch": 0.5685840939815349, "grad_norm": 1.5714448690414429, "learning_rate": 4.136439276105941e-06, "loss": 0.5012, "step": 13564 }, { "epoch": 0.5686260125965438, "grad_norm": 1.685923457145691, "learning_rate": 4.135770633331609e-06, "loss": 0.4997, "step": 13565 }, { "epoch": 0.5686679312115528, "grad_norm": 1.8839796781539917, "learning_rate": 4.135102006488154e-06, "loss": 0.4766, "step": 13566 }, { "epoch": 0.5687098498265617, "grad_norm": 1.7739742994308472, "learning_rate": 4.134433395587897e-06, "loss": 0.54, "step": 13567 }, { "epoch": 0.5687517684415707, "grad_norm": 1.7603578567504883, "learning_rate": 4.133764800643163e-06, "loss": 0.5272, "step": 13568 }, { "epoch": 0.5687936870565796, "grad_norm": 1.8170719146728516, "learning_rate": 4.133096221666279e-06, "loss": 0.4931, "step": 13569 }, { "epoch": 0.5688356056715886, "grad_norm": 1.817444920539856, "learning_rate": 4.132427658669564e-06, "loss": 0.4831, "step": 13570 }, { "epoch": 0.5688775242865975, "grad_norm": 1.497835636138916, "learning_rate": 4.131759111665349e-06, "loss": 0.417, "step": 13571 }, { "epoch": 0.5689194429016066, "grad_norm": 1.6378192901611328, "learning_rate": 4.1310905806659545e-06, "loss": 0.5058, "step": 13572 }, { "epoch": 0.5689613615166155, "grad_norm": 1.8657969236373901, "learning_rate": 4.130422065683702e-06, "loss": 0.5085, "step": 13573 }, { "epoch": 0.5690032801316245, "grad_norm": 1.5558863878250122, "learning_rate": 4.129753566730917e-06, "loss": 0.476, "step": 13574 }, { "epoch": 0.5690451987466334, "grad_norm": 1.669381856918335, "learning_rate": 4.129085083819924e-06, "loss": 0.5005, "step": 13575 }, { "epoch": 0.5690871173616424, "grad_norm": 1.484395980834961, "learning_rate": 4.128416616963041e-06, "loss": 0.4939, "step": 13576 }, { "epoch": 0.5691290359766513, "grad_norm": 1.691875696182251, "learning_rate": 4.127748166172593e-06, "loss": 0.4567, "step": 13577 }, { "epoch": 0.5691709545916603, "grad_norm": 1.840302586555481, "learning_rate": 4.127079731460904e-06, "loss": 0.4955, "step": 13578 }, { "epoch": 0.5692128732066692, "grad_norm": 1.5936888456344604, "learning_rate": 4.12641131284029e-06, "loss": 0.4995, "step": 13579 }, { "epoch": 0.5692547918216783, "grad_norm": 1.7941499948501587, "learning_rate": 4.125742910323078e-06, "loss": 0.4625, "step": 13580 }, { "epoch": 0.5692967104366872, "grad_norm": 1.6270110607147217, "learning_rate": 4.125074523921588e-06, "loss": 0.4776, "step": 13581 }, { "epoch": 0.5693386290516961, "grad_norm": 1.751931071281433, "learning_rate": 4.1244061536481365e-06, "loss": 0.5145, "step": 13582 }, { "epoch": 0.5693805476667051, "grad_norm": 1.6086244583129883, "learning_rate": 4.123737799515048e-06, "loss": 0.4784, "step": 13583 }, { "epoch": 0.569422466281714, "grad_norm": 1.6777634620666504, "learning_rate": 4.123069461534643e-06, "loss": 0.4629, "step": 13584 }, { "epoch": 0.569464384896723, "grad_norm": 1.9048871994018555, "learning_rate": 4.12240113971924e-06, "loss": 0.4831, "step": 13585 }, { "epoch": 0.5695063035117319, "grad_norm": 1.8934528827667236, "learning_rate": 4.121732834081157e-06, "loss": 0.527, "step": 13586 }, { "epoch": 0.5695482221267409, "grad_norm": 1.6529992818832397, "learning_rate": 4.121064544632717e-06, "loss": 0.4162, "step": 13587 }, { "epoch": 0.5695901407417499, "grad_norm": 1.8190699815750122, "learning_rate": 4.120396271386235e-06, "loss": 0.5275, "step": 13588 }, { "epoch": 0.5696320593567589, "grad_norm": 1.5819740295410156, "learning_rate": 4.119728014354034e-06, "loss": 0.4711, "step": 13589 }, { "epoch": 0.5696739779717678, "grad_norm": 1.7774629592895508, "learning_rate": 4.11905977354843e-06, "loss": 0.5457, "step": 13590 }, { "epoch": 0.5697158965867768, "grad_norm": 1.8862884044647217, "learning_rate": 4.11839154898174e-06, "loss": 0.5287, "step": 13591 }, { "epoch": 0.5697578152017857, "grad_norm": 1.8696483373641968, "learning_rate": 4.117723340666282e-06, "loss": 0.5425, "step": 13592 }, { "epoch": 0.5697997338167947, "grad_norm": 1.5296729803085327, "learning_rate": 4.117055148614377e-06, "loss": 0.4496, "step": 13593 }, { "epoch": 0.5698416524318036, "grad_norm": 1.6385334730148315, "learning_rate": 4.116386972838339e-06, "loss": 0.5432, "step": 13594 }, { "epoch": 0.5698835710468126, "grad_norm": 1.6822386980056763, "learning_rate": 4.115718813350484e-06, "loss": 0.461, "step": 13595 }, { "epoch": 0.5699254896618215, "grad_norm": 1.611228585243225, "learning_rate": 4.115050670163132e-06, "loss": 0.5315, "step": 13596 }, { "epoch": 0.5699674082768306, "grad_norm": 1.6262315511703491, "learning_rate": 4.114382543288596e-06, "loss": 0.4503, "step": 13597 }, { "epoch": 0.5700093268918395, "grad_norm": 1.8891632556915283, "learning_rate": 4.113714432739195e-06, "loss": 0.4965, "step": 13598 }, { "epoch": 0.5700512455068485, "grad_norm": 1.8228224515914917, "learning_rate": 4.113046338527244e-06, "loss": 0.5236, "step": 13599 }, { "epoch": 0.5700931641218574, "grad_norm": 1.6213759183883667, "learning_rate": 4.112378260665054e-06, "loss": 0.5112, "step": 13600 }, { "epoch": 0.5701350827368664, "grad_norm": 1.674833059310913, "learning_rate": 4.111710199164945e-06, "loss": 0.4991, "step": 13601 }, { "epoch": 0.5701770013518753, "grad_norm": 1.4964988231658936, "learning_rate": 4.111042154039232e-06, "loss": 0.4848, "step": 13602 }, { "epoch": 0.5702189199668843, "grad_norm": 1.6840729713439941, "learning_rate": 4.110374125300227e-06, "loss": 0.5066, "step": 13603 }, { "epoch": 0.5702608385818932, "grad_norm": 1.6898770332336426, "learning_rate": 4.109706112960245e-06, "loss": 0.4578, "step": 13604 }, { "epoch": 0.5703027571969023, "grad_norm": 1.6848770380020142, "learning_rate": 4.109038117031601e-06, "loss": 0.445, "step": 13605 }, { "epoch": 0.5703446758119112, "grad_norm": 1.3417071104049683, "learning_rate": 4.108370137526606e-06, "loss": 0.4439, "step": 13606 }, { "epoch": 0.5703865944269201, "grad_norm": 1.522199273109436, "learning_rate": 4.107702174457577e-06, "loss": 0.5249, "step": 13607 }, { "epoch": 0.5704285130419291, "grad_norm": 1.8509223461151123, "learning_rate": 4.107034227836822e-06, "loss": 0.4721, "step": 13608 }, { "epoch": 0.570470431656938, "grad_norm": 1.6150765419006348, "learning_rate": 4.106366297676658e-06, "loss": 0.5335, "step": 13609 }, { "epoch": 0.570512350271947, "grad_norm": 1.6374893188476562, "learning_rate": 4.105698383989396e-06, "loss": 0.501, "step": 13610 }, { "epoch": 0.5705542688869559, "grad_norm": 1.5197222232818604, "learning_rate": 4.105030486787349e-06, "loss": 0.4789, "step": 13611 }, { "epoch": 0.570596187501965, "grad_norm": 1.3381744623184204, "learning_rate": 4.1043626060828256e-06, "loss": 0.4633, "step": 13612 }, { "epoch": 0.5706381061169739, "grad_norm": 1.5394586324691772, "learning_rate": 4.103694741888141e-06, "loss": 0.5097, "step": 13613 }, { "epoch": 0.5706800247319829, "grad_norm": 1.7329097986221313, "learning_rate": 4.1030268942156045e-06, "loss": 0.4509, "step": 13614 }, { "epoch": 0.5707219433469918, "grad_norm": 1.708640456199646, "learning_rate": 4.1023590630775265e-06, "loss": 0.4421, "step": 13615 }, { "epoch": 0.5707638619620008, "grad_norm": 1.5187885761260986, "learning_rate": 4.101691248486217e-06, "loss": 0.5027, "step": 13616 }, { "epoch": 0.5708057805770097, "grad_norm": 1.7961381673812866, "learning_rate": 4.1010234504539905e-06, "loss": 0.4648, "step": 13617 }, { "epoch": 0.5708476991920187, "grad_norm": 1.7152050733566284, "learning_rate": 4.100355668993152e-06, "loss": 0.5277, "step": 13618 }, { "epoch": 0.5708896178070276, "grad_norm": 1.6252386569976807, "learning_rate": 4.099687904116013e-06, "loss": 0.5217, "step": 13619 }, { "epoch": 0.5709315364220366, "grad_norm": 1.9776884317398071, "learning_rate": 4.099020155834884e-06, "loss": 0.4928, "step": 13620 }, { "epoch": 0.5709734550370456, "grad_norm": 1.3681583404541016, "learning_rate": 4.098352424162071e-06, "loss": 0.4546, "step": 13621 }, { "epoch": 0.5710153736520546, "grad_norm": 2.1412999629974365, "learning_rate": 4.0976847091098865e-06, "loss": 0.4816, "step": 13622 }, { "epoch": 0.5710572922670635, "grad_norm": 1.5390639305114746, "learning_rate": 4.097017010690637e-06, "loss": 0.503, "step": 13623 }, { "epoch": 0.5710992108820725, "grad_norm": 1.4530335664749146, "learning_rate": 4.096349328916629e-06, "loss": 0.4535, "step": 13624 }, { "epoch": 0.5711411294970814, "grad_norm": 1.891366720199585, "learning_rate": 4.095681663800171e-06, "loss": 0.4565, "step": 13625 }, { "epoch": 0.5711830481120904, "grad_norm": 1.5931551456451416, "learning_rate": 4.095014015353575e-06, "loss": 0.4759, "step": 13626 }, { "epoch": 0.5712249667270993, "grad_norm": 1.7417418956756592, "learning_rate": 4.094346383589143e-06, "loss": 0.4895, "step": 13627 }, { "epoch": 0.5712668853421083, "grad_norm": 1.6554677486419678, "learning_rate": 4.093678768519182e-06, "loss": 0.4715, "step": 13628 }, { "epoch": 0.5713088039571173, "grad_norm": 1.7663452625274658, "learning_rate": 4.0930111701560015e-06, "loss": 0.5236, "step": 13629 }, { "epoch": 0.5713507225721263, "grad_norm": 1.4169223308563232, "learning_rate": 4.092343588511904e-06, "loss": 0.4121, "step": 13630 }, { "epoch": 0.5713926411871352, "grad_norm": 1.669232726097107, "learning_rate": 4.0916760235992e-06, "loss": 0.5177, "step": 13631 }, { "epoch": 0.5714345598021441, "grad_norm": 1.7855674028396606, "learning_rate": 4.091008475430193e-06, "loss": 0.4952, "step": 13632 }, { "epoch": 0.5714764784171531, "grad_norm": 1.9308397769927979, "learning_rate": 4.090340944017186e-06, "loss": 0.4879, "step": 13633 }, { "epoch": 0.571518397032162, "grad_norm": 1.5181883573532104, "learning_rate": 4.089673429372488e-06, "loss": 0.5218, "step": 13634 }, { "epoch": 0.571560315647171, "grad_norm": 2.3677597045898438, "learning_rate": 4.089005931508401e-06, "loss": 0.4682, "step": 13635 }, { "epoch": 0.5716022342621799, "grad_norm": 1.5460280179977417, "learning_rate": 4.08833845043723e-06, "loss": 0.4935, "step": 13636 }, { "epoch": 0.571644152877189, "grad_norm": 1.7221678495407104, "learning_rate": 4.0876709861712805e-06, "loss": 0.4904, "step": 13637 }, { "epoch": 0.5716860714921979, "grad_norm": 1.7012598514556885, "learning_rate": 4.087003538722854e-06, "loss": 0.4893, "step": 13638 }, { "epoch": 0.5717279901072069, "grad_norm": 1.939502477645874, "learning_rate": 4.086336108104254e-06, "loss": 0.5342, "step": 13639 }, { "epoch": 0.5717699087222158, "grad_norm": 1.6001828908920288, "learning_rate": 4.085668694327787e-06, "loss": 0.4887, "step": 13640 }, { "epoch": 0.5718118273372248, "grad_norm": 2.0698788166046143, "learning_rate": 4.085001297405753e-06, "loss": 0.4863, "step": 13641 }, { "epoch": 0.5718537459522337, "grad_norm": 1.7268973588943481, "learning_rate": 4.084333917350454e-06, "loss": 0.5189, "step": 13642 }, { "epoch": 0.5718956645672427, "grad_norm": 1.7345435619354248, "learning_rate": 4.083666554174195e-06, "loss": 0.5015, "step": 13643 }, { "epoch": 0.5719375831822516, "grad_norm": 3.965744972229004, "learning_rate": 4.082999207889274e-06, "loss": 0.4948, "step": 13644 }, { "epoch": 0.5719795017972606, "grad_norm": 1.6880035400390625, "learning_rate": 4.082331878507997e-06, "loss": 0.4662, "step": 13645 }, { "epoch": 0.5720214204122696, "grad_norm": 1.744616985321045, "learning_rate": 4.081664566042663e-06, "loss": 0.512, "step": 13646 }, { "epoch": 0.5720633390272786, "grad_norm": 1.6379314661026, "learning_rate": 4.080997270505572e-06, "loss": 0.4601, "step": 13647 }, { "epoch": 0.5721052576422875, "grad_norm": 2.144773483276367, "learning_rate": 4.080329991909025e-06, "loss": 0.4609, "step": 13648 }, { "epoch": 0.5721471762572965, "grad_norm": 1.795669436454773, "learning_rate": 4.0796627302653265e-06, "loss": 0.4714, "step": 13649 }, { "epoch": 0.5721890948723054, "grad_norm": 1.619411587715149, "learning_rate": 4.078995485586772e-06, "loss": 0.4691, "step": 13650 }, { "epoch": 0.5722310134873144, "grad_norm": 1.8553060293197632, "learning_rate": 4.07832825788566e-06, "loss": 0.5765, "step": 13651 }, { "epoch": 0.5722729321023233, "grad_norm": 1.805384635925293, "learning_rate": 4.077661047174296e-06, "loss": 0.501, "step": 13652 }, { "epoch": 0.5723148507173323, "grad_norm": 1.4820325374603271, "learning_rate": 4.0769938534649725e-06, "loss": 0.4711, "step": 13653 }, { "epoch": 0.5723567693323413, "grad_norm": 1.8547652959823608, "learning_rate": 4.076326676769993e-06, "loss": 0.5121, "step": 13654 }, { "epoch": 0.5723986879473503, "grad_norm": 2.0996756553649902, "learning_rate": 4.0756595171016535e-06, "loss": 0.5368, "step": 13655 }, { "epoch": 0.5724406065623592, "grad_norm": 2.575077533721924, "learning_rate": 4.074992374472254e-06, "loss": 0.5537, "step": 13656 }, { "epoch": 0.5724825251773681, "grad_norm": 1.8684911727905273, "learning_rate": 4.07432524889409e-06, "loss": 0.5299, "step": 13657 }, { "epoch": 0.5725244437923771, "grad_norm": 1.9152748584747314, "learning_rate": 4.073658140379462e-06, "loss": 0.5316, "step": 13658 }, { "epoch": 0.572566362407386, "grad_norm": 1.8068420886993408, "learning_rate": 4.0729910489406635e-06, "loss": 0.4944, "step": 13659 }, { "epoch": 0.572608281022395, "grad_norm": 1.7703684568405151, "learning_rate": 4.072323974589995e-06, "loss": 0.4601, "step": 13660 }, { "epoch": 0.5726501996374039, "grad_norm": 1.5098966360092163, "learning_rate": 4.0716569173397515e-06, "loss": 0.4706, "step": 13661 }, { "epoch": 0.572692118252413, "grad_norm": 1.626545786857605, "learning_rate": 4.0709898772022275e-06, "loss": 0.4665, "step": 13662 }, { "epoch": 0.5727340368674219, "grad_norm": 2.264599561691284, "learning_rate": 4.070322854189721e-06, "loss": 0.518, "step": 13663 }, { "epoch": 0.5727759554824309, "grad_norm": 1.7766287326812744, "learning_rate": 4.06965584831453e-06, "loss": 0.4999, "step": 13664 }, { "epoch": 0.5728178740974398, "grad_norm": 1.5312206745147705, "learning_rate": 4.068988859588946e-06, "loss": 0.5057, "step": 13665 }, { "epoch": 0.5728597927124488, "grad_norm": 2.028177261352539, "learning_rate": 4.068321888025264e-06, "loss": 0.4899, "step": 13666 }, { "epoch": 0.5729017113274577, "grad_norm": 1.7731029987335205, "learning_rate": 4.067654933635783e-06, "loss": 0.4992, "step": 13667 }, { "epoch": 0.5729436299424667, "grad_norm": 1.7386010885238647, "learning_rate": 4.066987996432791e-06, "loss": 0.5193, "step": 13668 }, { "epoch": 0.5729855485574756, "grad_norm": 1.5766520500183105, "learning_rate": 4.066321076428587e-06, "loss": 0.4675, "step": 13669 }, { "epoch": 0.5730274671724847, "grad_norm": 1.6364872455596924, "learning_rate": 4.065654173635464e-06, "loss": 0.4733, "step": 13670 }, { "epoch": 0.5730693857874936, "grad_norm": 2.4951648712158203, "learning_rate": 4.064987288065713e-06, "loss": 0.5158, "step": 13671 }, { "epoch": 0.5731113044025026, "grad_norm": 1.7633254528045654, "learning_rate": 4.064320419731629e-06, "loss": 0.4905, "step": 13672 }, { "epoch": 0.5731532230175115, "grad_norm": 2.743227243423462, "learning_rate": 4.063653568645506e-06, "loss": 0.4995, "step": 13673 }, { "epoch": 0.5731951416325205, "grad_norm": 1.9657009840011597, "learning_rate": 4.062986734819634e-06, "loss": 0.5535, "step": 13674 }, { "epoch": 0.5732370602475294, "grad_norm": 1.9649639129638672, "learning_rate": 4.062319918266305e-06, "loss": 0.4438, "step": 13675 }, { "epoch": 0.5732789788625384, "grad_norm": 1.8365478515625, "learning_rate": 4.061653118997814e-06, "loss": 0.4708, "step": 13676 }, { "epoch": 0.5733208974775473, "grad_norm": 1.692610263824463, "learning_rate": 4.060986337026449e-06, "loss": 0.5488, "step": 13677 }, { "epoch": 0.5733628160925563, "grad_norm": 1.6163240671157837, "learning_rate": 4.060319572364504e-06, "loss": 0.4802, "step": 13678 }, { "epoch": 0.5734047347075653, "grad_norm": 1.5980061292648315, "learning_rate": 4.059652825024269e-06, "loss": 0.4748, "step": 13679 }, { "epoch": 0.5734466533225743, "grad_norm": 1.7160829305648804, "learning_rate": 4.058986095018032e-06, "loss": 0.523, "step": 13680 }, { "epoch": 0.5734885719375832, "grad_norm": 1.587267279624939, "learning_rate": 4.058319382358086e-06, "loss": 0.5202, "step": 13681 }, { "epoch": 0.5735304905525921, "grad_norm": 1.602037787437439, "learning_rate": 4.057652687056722e-06, "loss": 0.4616, "step": 13682 }, { "epoch": 0.5735724091676011, "grad_norm": 1.693047046661377, "learning_rate": 4.056986009126227e-06, "loss": 0.5392, "step": 13683 }, { "epoch": 0.57361432778261, "grad_norm": 3.256122589111328, "learning_rate": 4.05631934857889e-06, "loss": 0.4871, "step": 13684 }, { "epoch": 0.573656246397619, "grad_norm": 1.7555938959121704, "learning_rate": 4.0556527054270035e-06, "loss": 0.4754, "step": 13685 }, { "epoch": 0.5736981650126279, "grad_norm": 1.804050326347351, "learning_rate": 4.054986079682852e-06, "loss": 0.4933, "step": 13686 }, { "epoch": 0.573740083627637, "grad_norm": 1.7759864330291748, "learning_rate": 4.054319471358726e-06, "loss": 0.4918, "step": 13687 }, { "epoch": 0.5737820022426459, "grad_norm": 1.5932484865188599, "learning_rate": 4.053652880466915e-06, "loss": 0.4885, "step": 13688 }, { "epoch": 0.5738239208576549, "grad_norm": 1.8570568561553955, "learning_rate": 4.052986307019703e-06, "loss": 0.4962, "step": 13689 }, { "epoch": 0.5738658394726638, "grad_norm": 1.9401910305023193, "learning_rate": 4.052319751029378e-06, "loss": 0.4932, "step": 13690 }, { "epoch": 0.5739077580876728, "grad_norm": 2.2909884452819824, "learning_rate": 4.051653212508231e-06, "loss": 0.4881, "step": 13691 }, { "epoch": 0.5739496767026817, "grad_norm": 3.462679624557495, "learning_rate": 4.0509866914685455e-06, "loss": 0.5035, "step": 13692 }, { "epoch": 0.5739915953176907, "grad_norm": 2.2144908905029297, "learning_rate": 4.0503201879226066e-06, "loss": 0.489, "step": 13693 }, { "epoch": 0.5740335139326996, "grad_norm": 6.2090301513671875, "learning_rate": 4.049653701882705e-06, "loss": 0.5034, "step": 13694 }, { "epoch": 0.5740754325477087, "grad_norm": 1.7592436075210571, "learning_rate": 4.048987233361121e-06, "loss": 0.4483, "step": 13695 }, { "epoch": 0.5741173511627176, "grad_norm": 2.216843843460083, "learning_rate": 4.048320782370143e-06, "loss": 0.4988, "step": 13696 }, { "epoch": 0.5741592697777266, "grad_norm": 1.617705225944519, "learning_rate": 4.0476543489220566e-06, "loss": 0.4517, "step": 13697 }, { "epoch": 0.5742011883927355, "grad_norm": 1.8434174060821533, "learning_rate": 4.046987933029145e-06, "loss": 0.4388, "step": 13698 }, { "epoch": 0.5742431070077445, "grad_norm": 1.8090250492095947, "learning_rate": 4.046321534703693e-06, "loss": 0.5001, "step": 13699 }, { "epoch": 0.5742850256227534, "grad_norm": 1.7551692724227905, "learning_rate": 4.045655153957988e-06, "loss": 0.545, "step": 13700 }, { "epoch": 0.5743269442377624, "grad_norm": 1.7396092414855957, "learning_rate": 4.044988790804308e-06, "loss": 0.5599, "step": 13701 }, { "epoch": 0.5743688628527713, "grad_norm": 1.521331787109375, "learning_rate": 4.044322445254939e-06, "loss": 0.5093, "step": 13702 }, { "epoch": 0.5744107814677804, "grad_norm": 1.7096935510635376, "learning_rate": 4.043656117322166e-06, "loss": 0.5014, "step": 13703 }, { "epoch": 0.5744527000827893, "grad_norm": 2.9898149967193604, "learning_rate": 4.04298980701827e-06, "loss": 0.4671, "step": 13704 }, { "epoch": 0.5744946186977983, "grad_norm": 1.9269665479660034, "learning_rate": 4.042323514355534e-06, "loss": 0.4934, "step": 13705 }, { "epoch": 0.5745365373128072, "grad_norm": 2.179203510284424, "learning_rate": 4.04165723934624e-06, "loss": 0.5042, "step": 13706 }, { "epoch": 0.5745784559278161, "grad_norm": 2.665609359741211, "learning_rate": 4.04099098200267e-06, "loss": 0.4942, "step": 13707 }, { "epoch": 0.5746203745428251, "grad_norm": 1.7742217779159546, "learning_rate": 4.040324742337104e-06, "loss": 0.5299, "step": 13708 }, { "epoch": 0.574662293157834, "grad_norm": 1.9835386276245117, "learning_rate": 4.039658520361828e-06, "loss": 0.5439, "step": 13709 }, { "epoch": 0.574704211772843, "grad_norm": 1.9206629991531372, "learning_rate": 4.038992316089116e-06, "loss": 0.5075, "step": 13710 }, { "epoch": 0.5747461303878519, "grad_norm": 1.6932462453842163, "learning_rate": 4.038326129531255e-06, "loss": 0.5055, "step": 13711 }, { "epoch": 0.574788049002861, "grad_norm": 1.9576817750930786, "learning_rate": 4.037659960700521e-06, "loss": 0.4883, "step": 13712 }, { "epoch": 0.5748299676178699, "grad_norm": 1.827183723449707, "learning_rate": 4.036993809609194e-06, "loss": 0.4684, "step": 13713 }, { "epoch": 0.5748718862328789, "grad_norm": 1.9113357067108154, "learning_rate": 4.036327676269557e-06, "loss": 0.4545, "step": 13714 }, { "epoch": 0.5749138048478878, "grad_norm": 1.827970266342163, "learning_rate": 4.035661560693886e-06, "loss": 0.4887, "step": 13715 }, { "epoch": 0.5749557234628968, "grad_norm": 1.8419890403747559, "learning_rate": 4.03499546289446e-06, "loss": 0.432, "step": 13716 }, { "epoch": 0.5749976420779057, "grad_norm": 2.19152569770813, "learning_rate": 4.034329382883561e-06, "loss": 0.5419, "step": 13717 }, { "epoch": 0.5750395606929147, "grad_norm": 2.3861382007598877, "learning_rate": 4.033663320673462e-06, "loss": 0.5203, "step": 13718 }, { "epoch": 0.5750814793079236, "grad_norm": 2.3409605026245117, "learning_rate": 4.0329972762764445e-06, "loss": 0.5441, "step": 13719 }, { "epoch": 0.5751233979229327, "grad_norm": 1.766408920288086, "learning_rate": 4.032331249704787e-06, "loss": 0.4462, "step": 13720 }, { "epoch": 0.5751653165379416, "grad_norm": 1.678147554397583, "learning_rate": 4.031665240970765e-06, "loss": 0.4659, "step": 13721 }, { "epoch": 0.5752072351529506, "grad_norm": 1.6055186986923218, "learning_rate": 4.030999250086654e-06, "loss": 0.4528, "step": 13722 }, { "epoch": 0.5752491537679595, "grad_norm": 1.5212174654006958, "learning_rate": 4.030333277064733e-06, "loss": 0.4877, "step": 13723 }, { "epoch": 0.5752910723829685, "grad_norm": 1.5076992511749268, "learning_rate": 4.0296673219172775e-06, "loss": 0.5123, "step": 13724 }, { "epoch": 0.5753329909979774, "grad_norm": 1.7986572980880737, "learning_rate": 4.029001384656563e-06, "loss": 0.5306, "step": 13725 }, { "epoch": 0.5753749096129864, "grad_norm": 1.7576886415481567, "learning_rate": 4.028335465294868e-06, "loss": 0.5082, "step": 13726 }, { "epoch": 0.5754168282279953, "grad_norm": 1.888170838356018, "learning_rate": 4.027669563844462e-06, "loss": 0.5222, "step": 13727 }, { "epoch": 0.5754587468430044, "grad_norm": 1.8426605463027954, "learning_rate": 4.027003680317624e-06, "loss": 0.4892, "step": 13728 }, { "epoch": 0.5755006654580133, "grad_norm": 1.6028372049331665, "learning_rate": 4.02633781472663e-06, "loss": 0.5288, "step": 13729 }, { "epoch": 0.5755425840730223, "grad_norm": 1.6079115867614746, "learning_rate": 4.025671967083752e-06, "loss": 0.5031, "step": 13730 }, { "epoch": 0.5755845026880312, "grad_norm": 1.9575798511505127, "learning_rate": 4.0250061374012625e-06, "loss": 0.5203, "step": 13731 }, { "epoch": 0.5756264213030401, "grad_norm": 1.846923828125, "learning_rate": 4.024340325691439e-06, "loss": 0.5045, "step": 13732 }, { "epoch": 0.5756683399180491, "grad_norm": 2.2925500869750977, "learning_rate": 4.02367453196655e-06, "loss": 0.5041, "step": 13733 }, { "epoch": 0.575710258533058, "grad_norm": 1.4189261198043823, "learning_rate": 4.023008756238875e-06, "loss": 0.4043, "step": 13734 }, { "epoch": 0.575752177148067, "grad_norm": 1.8417831659317017, "learning_rate": 4.022342998520681e-06, "loss": 0.5198, "step": 13735 }, { "epoch": 0.5757940957630759, "grad_norm": 1.7497224807739258, "learning_rate": 4.021677258824241e-06, "loss": 0.5052, "step": 13736 }, { "epoch": 0.575836014378085, "grad_norm": 2.0480473041534424, "learning_rate": 4.021011537161829e-06, "loss": 0.5537, "step": 13737 }, { "epoch": 0.5758779329930939, "grad_norm": 1.9276047945022583, "learning_rate": 4.020345833545717e-06, "loss": 0.4636, "step": 13738 }, { "epoch": 0.5759198516081029, "grad_norm": 1.8839939832687378, "learning_rate": 4.019680147988176e-06, "loss": 0.503, "step": 13739 }, { "epoch": 0.5759617702231118, "grad_norm": 1.6471917629241943, "learning_rate": 4.019014480501473e-06, "loss": 0.4288, "step": 13740 }, { "epoch": 0.5760036888381208, "grad_norm": 1.964679479598999, "learning_rate": 4.018348831097886e-06, "loss": 0.4626, "step": 13741 }, { "epoch": 0.5760456074531297, "grad_norm": 2.4272420406341553, "learning_rate": 4.017683199789677e-06, "loss": 0.5003, "step": 13742 }, { "epoch": 0.5760875260681387, "grad_norm": 1.6492993831634521, "learning_rate": 4.017017586589122e-06, "loss": 0.4748, "step": 13743 }, { "epoch": 0.5761294446831476, "grad_norm": 1.574897050857544, "learning_rate": 4.016351991508489e-06, "loss": 0.4906, "step": 13744 }, { "epoch": 0.5761713632981567, "grad_norm": 1.9068676233291626, "learning_rate": 4.015686414560048e-06, "loss": 0.5216, "step": 13745 }, { "epoch": 0.5762132819131656, "grad_norm": 1.7007712125778198, "learning_rate": 4.0150208557560655e-06, "loss": 0.4708, "step": 13746 }, { "epoch": 0.5762552005281746, "grad_norm": 1.6856430768966675, "learning_rate": 4.014355315108814e-06, "loss": 0.4595, "step": 13747 }, { "epoch": 0.5762971191431835, "grad_norm": 1.8616546392440796, "learning_rate": 4.013689792630557e-06, "loss": 0.4932, "step": 13748 }, { "epoch": 0.5763390377581925, "grad_norm": 1.9015742540359497, "learning_rate": 4.013024288333567e-06, "loss": 0.5406, "step": 13749 }, { "epoch": 0.5763809563732014, "grad_norm": 1.8639373779296875, "learning_rate": 4.01235880223011e-06, "loss": 0.4749, "step": 13750 }, { "epoch": 0.5764228749882104, "grad_norm": 2.019766330718994, "learning_rate": 4.011693334332452e-06, "loss": 0.4871, "step": 13751 }, { "epoch": 0.5764647936032193, "grad_norm": 1.440506100654602, "learning_rate": 4.01102788465286e-06, "loss": 0.5049, "step": 13752 }, { "epoch": 0.5765067122182284, "grad_norm": 2.100733757019043, "learning_rate": 4.0103624532036045e-06, "loss": 0.539, "step": 13753 }, { "epoch": 0.5765486308332373, "grad_norm": 1.776873230934143, "learning_rate": 4.009697039996948e-06, "loss": 0.4921, "step": 13754 }, { "epoch": 0.5765905494482463, "grad_norm": 1.557265043258667, "learning_rate": 4.009031645045156e-06, "loss": 0.489, "step": 13755 }, { "epoch": 0.5766324680632552, "grad_norm": 1.5734068155288696, "learning_rate": 4.0083662683604986e-06, "loss": 0.4464, "step": 13756 }, { "epoch": 0.5766743866782641, "grad_norm": 1.7439854145050049, "learning_rate": 4.007700909955236e-06, "loss": 0.4827, "step": 13757 }, { "epoch": 0.5767163052932731, "grad_norm": 1.9733622074127197, "learning_rate": 4.007035569841635e-06, "loss": 0.4901, "step": 13758 }, { "epoch": 0.576758223908282, "grad_norm": 1.747161626815796, "learning_rate": 4.006370248031963e-06, "loss": 0.4543, "step": 13759 }, { "epoch": 0.576800142523291, "grad_norm": 1.7543669939041138, "learning_rate": 4.005704944538479e-06, "loss": 0.5247, "step": 13760 }, { "epoch": 0.5768420611382999, "grad_norm": 2.110978603363037, "learning_rate": 4.00503965937345e-06, "loss": 0.5413, "step": 13761 }, { "epoch": 0.576883979753309, "grad_norm": 1.718286395072937, "learning_rate": 4.004374392549141e-06, "loss": 0.5233, "step": 13762 }, { "epoch": 0.5769258983683179, "grad_norm": 2.2218949794769287, "learning_rate": 4.003709144077813e-06, "loss": 0.4544, "step": 13763 }, { "epoch": 0.5769678169833269, "grad_norm": 1.9320834875106812, "learning_rate": 4.003043913971729e-06, "loss": 0.5227, "step": 13764 }, { "epoch": 0.5770097355983358, "grad_norm": 1.588202953338623, "learning_rate": 4.0023787022431535e-06, "loss": 0.514, "step": 13765 }, { "epoch": 0.5770516542133448, "grad_norm": 1.585089921951294, "learning_rate": 4.001713508904346e-06, "loss": 0.4735, "step": 13766 }, { "epoch": 0.5770935728283537, "grad_norm": 2.265594244003296, "learning_rate": 4.001048333967571e-06, "loss": 0.4514, "step": 13767 }, { "epoch": 0.5771354914433627, "grad_norm": 1.660334587097168, "learning_rate": 4.00038317744509e-06, "loss": 0.5095, "step": 13768 }, { "epoch": 0.5771774100583716, "grad_norm": 1.8177047967910767, "learning_rate": 3.9997180393491604e-06, "loss": 0.5104, "step": 13769 }, { "epoch": 0.5772193286733807, "grad_norm": 2.181898593902588, "learning_rate": 3.999052919692047e-06, "loss": 0.5245, "step": 13770 }, { "epoch": 0.5772612472883896, "grad_norm": 2.2042453289031982, "learning_rate": 3.99838781848601e-06, "loss": 0.4931, "step": 13771 }, { "epoch": 0.5773031659033986, "grad_norm": 1.6177939176559448, "learning_rate": 3.9977227357433085e-06, "loss": 0.497, "step": 13772 }, { "epoch": 0.5773450845184075, "grad_norm": 1.9317249059677124, "learning_rate": 3.997057671476202e-06, "loss": 0.4718, "step": 13773 }, { "epoch": 0.5773870031334165, "grad_norm": 1.8944288492202759, "learning_rate": 3.9963926256969525e-06, "loss": 0.4982, "step": 13774 }, { "epoch": 0.5774289217484254, "grad_norm": 1.7311768531799316, "learning_rate": 3.995727598417815e-06, "loss": 0.4826, "step": 13775 }, { "epoch": 0.5774708403634344, "grad_norm": 1.66413414478302, "learning_rate": 3.995062589651052e-06, "loss": 0.4681, "step": 13776 }, { "epoch": 0.5775127589784433, "grad_norm": 1.8347084522247314, "learning_rate": 3.994397599408923e-06, "loss": 0.5297, "step": 13777 }, { "epoch": 0.5775546775934524, "grad_norm": 1.6192346811294556, "learning_rate": 3.993732627703681e-06, "loss": 0.4821, "step": 13778 }, { "epoch": 0.5775965962084613, "grad_norm": 2.0144312381744385, "learning_rate": 3.9930676745475875e-06, "loss": 0.4952, "step": 13779 }, { "epoch": 0.5776385148234703, "grad_norm": 1.9372698068618774, "learning_rate": 3.992402739952901e-06, "loss": 0.507, "step": 13780 }, { "epoch": 0.5776804334384792, "grad_norm": 2.5186057090759277, "learning_rate": 3.991737823931876e-06, "loss": 0.4341, "step": 13781 }, { "epoch": 0.5777223520534881, "grad_norm": 1.647491693496704, "learning_rate": 3.99107292649677e-06, "loss": 0.4751, "step": 13782 }, { "epoch": 0.5777642706684971, "grad_norm": 2.7659385204315186, "learning_rate": 3.990408047659841e-06, "loss": 0.4817, "step": 13783 }, { "epoch": 0.577806189283506, "grad_norm": 1.8578119277954102, "learning_rate": 3.989743187433342e-06, "loss": 0.4877, "step": 13784 }, { "epoch": 0.577848107898515, "grad_norm": 1.9640825986862183, "learning_rate": 3.9890783458295326e-06, "loss": 0.4874, "step": 13785 }, { "epoch": 0.577890026513524, "grad_norm": 1.9064421653747559, "learning_rate": 3.988413522860666e-06, "loss": 0.476, "step": 13786 }, { "epoch": 0.577931945128533, "grad_norm": 1.6806583404541016, "learning_rate": 3.9877487185389966e-06, "loss": 0.5447, "step": 13787 }, { "epoch": 0.5779738637435419, "grad_norm": 1.5523028373718262, "learning_rate": 3.987083932876782e-06, "loss": 0.4785, "step": 13788 }, { "epoch": 0.5780157823585509, "grad_norm": 1.6423909664154053, "learning_rate": 3.986419165886272e-06, "loss": 0.5218, "step": 13789 }, { "epoch": 0.5780577009735598, "grad_norm": 1.8327782154083252, "learning_rate": 3.985754417579726e-06, "loss": 0.5334, "step": 13790 }, { "epoch": 0.5780996195885688, "grad_norm": 1.5662413835525513, "learning_rate": 3.985089687969393e-06, "loss": 0.4652, "step": 13791 }, { "epoch": 0.5781415382035777, "grad_norm": 1.9634464979171753, "learning_rate": 3.98442497706753e-06, "loss": 0.5342, "step": 13792 }, { "epoch": 0.5781834568185867, "grad_norm": 1.7078999280929565, "learning_rate": 3.983760284886388e-06, "loss": 0.4839, "step": 13793 }, { "epoch": 0.5782253754335956, "grad_norm": 1.8005633354187012, "learning_rate": 3.983095611438221e-06, "loss": 0.5028, "step": 13794 }, { "epoch": 0.5782672940486047, "grad_norm": 1.723183512687683, "learning_rate": 3.982430956735279e-06, "loss": 0.5062, "step": 13795 }, { "epoch": 0.5783092126636136, "grad_norm": 1.7564181089401245, "learning_rate": 3.981766320789816e-06, "loss": 0.5114, "step": 13796 }, { "epoch": 0.5783511312786226, "grad_norm": 1.913967490196228, "learning_rate": 3.981101703614086e-06, "loss": 0.5653, "step": 13797 }, { "epoch": 0.5783930498936315, "grad_norm": 1.502542495727539, "learning_rate": 3.980437105220334e-06, "loss": 0.4312, "step": 13798 }, { "epoch": 0.5784349685086405, "grad_norm": 2.658177137374878, "learning_rate": 3.979772525620815e-06, "loss": 0.4657, "step": 13799 }, { "epoch": 0.5784768871236494, "grad_norm": 1.7172503471374512, "learning_rate": 3.979107964827781e-06, "loss": 0.5362, "step": 13800 }, { "epoch": 0.5785188057386584, "grad_norm": 7.316105365753174, "learning_rate": 3.978443422853479e-06, "loss": 0.5337, "step": 13801 }, { "epoch": 0.5785607243536673, "grad_norm": 1.8008265495300293, "learning_rate": 3.977778899710159e-06, "loss": 0.559, "step": 13802 }, { "epoch": 0.5786026429686764, "grad_norm": 2.013359308242798, "learning_rate": 3.977114395410075e-06, "loss": 0.4798, "step": 13803 }, { "epoch": 0.5786445615836853, "grad_norm": 1.7302361726760864, "learning_rate": 3.9764499099654704e-06, "loss": 0.4758, "step": 13804 }, { "epoch": 0.5786864801986943, "grad_norm": 1.598093867301941, "learning_rate": 3.9757854433885975e-06, "loss": 0.5276, "step": 13805 }, { "epoch": 0.5787283988137032, "grad_norm": 1.491171956062317, "learning_rate": 3.975120995691705e-06, "loss": 0.488, "step": 13806 }, { "epoch": 0.5787703174287121, "grad_norm": 6.876087188720703, "learning_rate": 3.974456566887038e-06, "loss": 0.521, "step": 13807 }, { "epoch": 0.5788122360437211, "grad_norm": 1.731788992881775, "learning_rate": 3.973792156986847e-06, "loss": 0.4338, "step": 13808 }, { "epoch": 0.57885415465873, "grad_norm": 2.2327005863189697, "learning_rate": 3.973127766003381e-06, "loss": 0.503, "step": 13809 }, { "epoch": 0.578896073273739, "grad_norm": 1.5902291536331177, "learning_rate": 3.972463393948883e-06, "loss": 0.484, "step": 13810 }, { "epoch": 0.578937991888748, "grad_norm": 2.8337185382843018, "learning_rate": 3.971799040835601e-06, "loss": 0.4952, "step": 13811 }, { "epoch": 0.578979910503757, "grad_norm": 2.0564990043640137, "learning_rate": 3.971134706675785e-06, "loss": 0.5285, "step": 13812 }, { "epoch": 0.5790218291187659, "grad_norm": 1.9854925870895386, "learning_rate": 3.970470391481676e-06, "loss": 0.4934, "step": 13813 }, { "epoch": 0.5790637477337749, "grad_norm": 1.608895182609558, "learning_rate": 3.969806095265522e-06, "loss": 0.5161, "step": 13814 }, { "epoch": 0.5791056663487838, "grad_norm": 2.328148365020752, "learning_rate": 3.969141818039571e-06, "loss": 0.5147, "step": 13815 }, { "epoch": 0.5791475849637928, "grad_norm": 2.6488704681396484, "learning_rate": 3.968477559816063e-06, "loss": 0.4819, "step": 13816 }, { "epoch": 0.5791895035788017, "grad_norm": 3.408989429473877, "learning_rate": 3.967813320607245e-06, "loss": 0.4824, "step": 13817 }, { "epoch": 0.5792314221938107, "grad_norm": 2.7034921646118164, "learning_rate": 3.967149100425363e-06, "loss": 0.4652, "step": 13818 }, { "epoch": 0.5792733408088196, "grad_norm": 2.0047097206115723, "learning_rate": 3.9664848992826596e-06, "loss": 0.5036, "step": 13819 }, { "epoch": 0.5793152594238287, "grad_norm": 1.7755703926086426, "learning_rate": 3.965820717191377e-06, "loss": 0.4644, "step": 13820 }, { "epoch": 0.5793571780388376, "grad_norm": 2.3148820400238037, "learning_rate": 3.9651565541637614e-06, "loss": 0.5336, "step": 13821 }, { "epoch": 0.5793990966538466, "grad_norm": 1.8231630325317383, "learning_rate": 3.964492410212052e-06, "loss": 0.4973, "step": 13822 }, { "epoch": 0.5794410152688555, "grad_norm": 1.9866803884506226, "learning_rate": 3.963828285348494e-06, "loss": 0.5071, "step": 13823 }, { "epoch": 0.5794829338838645, "grad_norm": 1.754602313041687, "learning_rate": 3.963164179585331e-06, "loss": 0.4826, "step": 13824 }, { "epoch": 0.5795248524988734, "grad_norm": 1.7302091121673584, "learning_rate": 3.9625000929348005e-06, "loss": 0.4962, "step": 13825 }, { "epoch": 0.5795667711138824, "grad_norm": 1.658151388168335, "learning_rate": 3.961836025409147e-06, "loss": 0.5077, "step": 13826 }, { "epoch": 0.5796086897288913, "grad_norm": 1.629920244216919, "learning_rate": 3.961171977020612e-06, "loss": 0.4616, "step": 13827 }, { "epoch": 0.5796506083439004, "grad_norm": 2.1492269039154053, "learning_rate": 3.960507947781436e-06, "loss": 0.5028, "step": 13828 }, { "epoch": 0.5796925269589093, "grad_norm": 2.1099884510040283, "learning_rate": 3.9598439377038575e-06, "loss": 0.5316, "step": 13829 }, { "epoch": 0.5797344455739183, "grad_norm": 2.287158966064453, "learning_rate": 3.95917994680012e-06, "loss": 0.4623, "step": 13830 }, { "epoch": 0.5797763641889272, "grad_norm": 1.8456193208694458, "learning_rate": 3.958515975082459e-06, "loss": 0.4759, "step": 13831 }, { "epoch": 0.5798182828039361, "grad_norm": 2.2578649520874023, "learning_rate": 3.9578520225631165e-06, "loss": 0.5056, "step": 13832 }, { "epoch": 0.5798602014189451, "grad_norm": 2.230369806289673, "learning_rate": 3.957188089254334e-06, "loss": 0.5231, "step": 13833 }, { "epoch": 0.579902120033954, "grad_norm": 1.46173894405365, "learning_rate": 3.956524175168344e-06, "loss": 0.4582, "step": 13834 }, { "epoch": 0.579944038648963, "grad_norm": 2.120753288269043, "learning_rate": 3.955860280317388e-06, "loss": 0.4652, "step": 13835 }, { "epoch": 0.579985957263972, "grad_norm": 2.1112937927246094, "learning_rate": 3.955196404713707e-06, "loss": 0.486, "step": 13836 }, { "epoch": 0.580027875878981, "grad_norm": 1.8563255071640015, "learning_rate": 3.954532548369535e-06, "loss": 0.523, "step": 13837 }, { "epoch": 0.5800697944939899, "grad_norm": 1.823169469833374, "learning_rate": 3.953868711297109e-06, "loss": 0.4727, "step": 13838 }, { "epoch": 0.5801117131089989, "grad_norm": 2.5958313941955566, "learning_rate": 3.953204893508668e-06, "loss": 0.4921, "step": 13839 }, { "epoch": 0.5801536317240078, "grad_norm": 1.7214678525924683, "learning_rate": 3.952541095016447e-06, "loss": 0.4093, "step": 13840 }, { "epoch": 0.5801955503390168, "grad_norm": 2.482264518737793, "learning_rate": 3.9518773158326836e-06, "loss": 0.4845, "step": 13841 }, { "epoch": 0.5802374689540257, "grad_norm": 2.5942955017089844, "learning_rate": 3.951213555969613e-06, "loss": 0.4452, "step": 13842 }, { "epoch": 0.5802793875690347, "grad_norm": 1.6473509073257446, "learning_rate": 3.950549815439471e-06, "loss": 0.4682, "step": 13843 }, { "epoch": 0.5803213061840437, "grad_norm": 1.554904580116272, "learning_rate": 3.94988609425449e-06, "loss": 0.4738, "step": 13844 }, { "epoch": 0.5803632247990527, "grad_norm": 1.614803671836853, "learning_rate": 3.94922239242691e-06, "loss": 0.4592, "step": 13845 }, { "epoch": 0.5804051434140616, "grad_norm": 1.8967500925064087, "learning_rate": 3.948558709968961e-06, "loss": 0.4738, "step": 13846 }, { "epoch": 0.5804470620290706, "grad_norm": 1.6876672506332397, "learning_rate": 3.9478950468928786e-06, "loss": 0.4826, "step": 13847 }, { "epoch": 0.5804889806440795, "grad_norm": 1.7323957681655884, "learning_rate": 3.947231403210897e-06, "loss": 0.438, "step": 13848 }, { "epoch": 0.5805308992590885, "grad_norm": 1.559510588645935, "learning_rate": 3.946567778935248e-06, "loss": 0.4929, "step": 13849 }, { "epoch": 0.5805728178740974, "grad_norm": 1.5772061347961426, "learning_rate": 3.945904174078166e-06, "loss": 0.4817, "step": 13850 }, { "epoch": 0.5806147364891064, "grad_norm": 2.022087335586548, "learning_rate": 3.945240588651885e-06, "loss": 0.5529, "step": 13851 }, { "epoch": 0.5806566551041153, "grad_norm": 1.662729263305664, "learning_rate": 3.9445770226686344e-06, "loss": 0.4404, "step": 13852 }, { "epoch": 0.5806985737191244, "grad_norm": 1.908665418624878, "learning_rate": 3.943913476140647e-06, "loss": 0.4943, "step": 13853 }, { "epoch": 0.5807404923341333, "grad_norm": 1.9575048685073853, "learning_rate": 3.943249949080156e-06, "loss": 0.4697, "step": 13854 }, { "epoch": 0.5807824109491423, "grad_norm": 1.8073594570159912, "learning_rate": 3.942586441499389e-06, "loss": 0.4685, "step": 13855 }, { "epoch": 0.5808243295641512, "grad_norm": 2.187312364578247, "learning_rate": 3.941922953410581e-06, "loss": 0.5111, "step": 13856 }, { "epoch": 0.5808662481791601, "grad_norm": 1.766825795173645, "learning_rate": 3.941259484825961e-06, "loss": 0.4567, "step": 13857 }, { "epoch": 0.5809081667941691, "grad_norm": 1.8074452877044678, "learning_rate": 3.940596035757757e-06, "loss": 0.5015, "step": 13858 }, { "epoch": 0.580950085409178, "grad_norm": 2.250577688217163, "learning_rate": 3.939932606218202e-06, "loss": 0.53, "step": 13859 }, { "epoch": 0.580992004024187, "grad_norm": 1.6400352716445923, "learning_rate": 3.939269196219522e-06, "loss": 0.4543, "step": 13860 }, { "epoch": 0.581033922639196, "grad_norm": 1.801223635673523, "learning_rate": 3.93860580577395e-06, "loss": 0.5792, "step": 13861 }, { "epoch": 0.581075841254205, "grad_norm": 1.6804804801940918, "learning_rate": 3.9379424348937125e-06, "loss": 0.4607, "step": 13862 }, { "epoch": 0.5811177598692139, "grad_norm": 1.8428231477737427, "learning_rate": 3.937279083591036e-06, "loss": 0.5191, "step": 13863 }, { "epoch": 0.5811596784842229, "grad_norm": 1.5064901113510132, "learning_rate": 3.93661575187815e-06, "loss": 0.4267, "step": 13864 }, { "epoch": 0.5812015970992318, "grad_norm": 1.81699538230896, "learning_rate": 3.935952439767286e-06, "loss": 0.4758, "step": 13865 }, { "epoch": 0.5812435157142408, "grad_norm": 1.9799214601516724, "learning_rate": 3.935289147270666e-06, "loss": 0.4973, "step": 13866 }, { "epoch": 0.5812854343292497, "grad_norm": 1.5876954793930054, "learning_rate": 3.934625874400518e-06, "loss": 0.4229, "step": 13867 }, { "epoch": 0.5813273529442587, "grad_norm": 1.6825764179229736, "learning_rate": 3.933962621169071e-06, "loss": 0.4969, "step": 13868 }, { "epoch": 0.5813692715592677, "grad_norm": 1.8558030128479004, "learning_rate": 3.933299387588547e-06, "loss": 0.4799, "step": 13869 }, { "epoch": 0.5814111901742767, "grad_norm": 1.6511759757995605, "learning_rate": 3.932636173671175e-06, "loss": 0.479, "step": 13870 }, { "epoch": 0.5814531087892856, "grad_norm": 1.824205994606018, "learning_rate": 3.931972979429181e-06, "loss": 0.5381, "step": 13871 }, { "epoch": 0.5814950274042946, "grad_norm": 1.693419337272644, "learning_rate": 3.931309804874786e-06, "loss": 0.4671, "step": 13872 }, { "epoch": 0.5815369460193035, "grad_norm": 1.8214424848556519, "learning_rate": 3.9306466500202175e-06, "loss": 0.5303, "step": 13873 }, { "epoch": 0.5815788646343125, "grad_norm": 1.9526907205581665, "learning_rate": 3.929983514877703e-06, "loss": 0.5562, "step": 13874 }, { "epoch": 0.5816207832493214, "grad_norm": 1.595999836921692, "learning_rate": 3.929320399459461e-06, "loss": 0.4639, "step": 13875 }, { "epoch": 0.5816627018643304, "grad_norm": 1.8328062295913696, "learning_rate": 3.928657303777716e-06, "loss": 0.5022, "step": 13876 }, { "epoch": 0.5817046204793394, "grad_norm": 1.6572346687316895, "learning_rate": 3.927994227844695e-06, "loss": 0.5124, "step": 13877 }, { "epoch": 0.5817465390943484, "grad_norm": 1.5849837064743042, "learning_rate": 3.9273311716726155e-06, "loss": 0.5229, "step": 13878 }, { "epoch": 0.5817884577093573, "grad_norm": 1.7465256452560425, "learning_rate": 3.926668135273705e-06, "loss": 0.4829, "step": 13879 }, { "epoch": 0.5818303763243663, "grad_norm": 1.7454004287719727, "learning_rate": 3.926005118660185e-06, "loss": 0.4807, "step": 13880 }, { "epoch": 0.5818722949393752, "grad_norm": 3.408903121948242, "learning_rate": 3.925342121844273e-06, "loss": 0.5038, "step": 13881 }, { "epoch": 0.5819142135543841, "grad_norm": 1.9649901390075684, "learning_rate": 3.924679144838193e-06, "loss": 0.512, "step": 13882 }, { "epoch": 0.5819561321693931, "grad_norm": 1.770744800567627, "learning_rate": 3.924016187654169e-06, "loss": 0.4921, "step": 13883 }, { "epoch": 0.581998050784402, "grad_norm": 1.7437477111816406, "learning_rate": 3.923353250304418e-06, "loss": 0.4553, "step": 13884 }, { "epoch": 0.582039969399411, "grad_norm": 1.546209454536438, "learning_rate": 3.922690332801161e-06, "loss": 0.4494, "step": 13885 }, { "epoch": 0.58208188801442, "grad_norm": 1.5914018154144287, "learning_rate": 3.922027435156619e-06, "loss": 0.4658, "step": 13886 }, { "epoch": 0.582123806629429, "grad_norm": 2.1614909172058105, "learning_rate": 3.92136455738301e-06, "loss": 0.5377, "step": 13887 }, { "epoch": 0.5821657252444379, "grad_norm": 1.6306359767913818, "learning_rate": 3.920701699492555e-06, "loss": 0.5066, "step": 13888 }, { "epoch": 0.5822076438594469, "grad_norm": 1.5205509662628174, "learning_rate": 3.920038861497472e-06, "loss": 0.4597, "step": 13889 }, { "epoch": 0.5822495624744558, "grad_norm": 1.6533184051513672, "learning_rate": 3.91937604340998e-06, "loss": 0.462, "step": 13890 }, { "epoch": 0.5822914810894648, "grad_norm": 1.4957133531570435, "learning_rate": 3.9187132452422946e-06, "loss": 0.4923, "step": 13891 }, { "epoch": 0.5823333997044737, "grad_norm": 1.6268943548202515, "learning_rate": 3.918050467006639e-06, "loss": 0.5135, "step": 13892 }, { "epoch": 0.5823753183194827, "grad_norm": 2.0467092990875244, "learning_rate": 3.917387708715224e-06, "loss": 0.525, "step": 13893 }, { "epoch": 0.5824172369344917, "grad_norm": 1.809283971786499, "learning_rate": 3.916724970380272e-06, "loss": 0.4782, "step": 13894 }, { "epoch": 0.5824591555495007, "grad_norm": 2.1022496223449707, "learning_rate": 3.916062252013999e-06, "loss": 0.4687, "step": 13895 }, { "epoch": 0.5825010741645096, "grad_norm": 1.739231824874878, "learning_rate": 3.915399553628616e-06, "loss": 0.4217, "step": 13896 }, { "epoch": 0.5825429927795186, "grad_norm": 1.8695226907730103, "learning_rate": 3.914736875236344e-06, "loss": 0.4543, "step": 13897 }, { "epoch": 0.5825849113945275, "grad_norm": 1.981643557548523, "learning_rate": 3.9140742168494e-06, "loss": 0.5454, "step": 13898 }, { "epoch": 0.5826268300095365, "grad_norm": 1.6320922374725342, "learning_rate": 3.913411578479995e-06, "loss": 0.4686, "step": 13899 }, { "epoch": 0.5826687486245454, "grad_norm": 1.7045223712921143, "learning_rate": 3.912748960140344e-06, "loss": 0.4753, "step": 13900 }, { "epoch": 0.5827106672395544, "grad_norm": 1.7820852994918823, "learning_rate": 3.912086361842665e-06, "loss": 0.5281, "step": 13901 }, { "epoch": 0.5827525858545634, "grad_norm": 1.8803231716156006, "learning_rate": 3.911423783599168e-06, "loss": 0.5017, "step": 13902 }, { "epoch": 0.5827945044695724, "grad_norm": 1.9425737857818604, "learning_rate": 3.91076122542207e-06, "loss": 0.4706, "step": 13903 }, { "epoch": 0.5828364230845813, "grad_norm": 1.9401848316192627, "learning_rate": 3.910098687323584e-06, "loss": 0.4708, "step": 13904 }, { "epoch": 0.5828783416995903, "grad_norm": 1.9998092651367188, "learning_rate": 3.909436169315919e-06, "loss": 0.5024, "step": 13905 }, { "epoch": 0.5829202603145992, "grad_norm": 1.7208460569381714, "learning_rate": 3.908773671411291e-06, "loss": 0.4639, "step": 13906 }, { "epoch": 0.5829621789296081, "grad_norm": 1.8611057996749878, "learning_rate": 3.908111193621914e-06, "loss": 0.4872, "step": 13907 }, { "epoch": 0.5830040975446171, "grad_norm": 2.159332275390625, "learning_rate": 3.907448735959997e-06, "loss": 0.5516, "step": 13908 }, { "epoch": 0.583046016159626, "grad_norm": 1.8803329467773438, "learning_rate": 3.9067862984377505e-06, "loss": 0.4703, "step": 13909 }, { "epoch": 0.583087934774635, "grad_norm": 1.6506961584091187, "learning_rate": 3.906123881067389e-06, "loss": 0.4786, "step": 13910 }, { "epoch": 0.583129853389644, "grad_norm": 1.4947067499160767, "learning_rate": 3.90546148386112e-06, "loss": 0.4475, "step": 13911 }, { "epoch": 0.583171772004653, "grad_norm": 2.105166435241699, "learning_rate": 3.904799106831157e-06, "loss": 0.5375, "step": 13912 }, { "epoch": 0.5832136906196619, "grad_norm": 2.0248661041259766, "learning_rate": 3.9041367499897095e-06, "loss": 0.5406, "step": 13913 }, { "epoch": 0.5832556092346709, "grad_norm": 1.6843332052230835, "learning_rate": 3.903474413348983e-06, "loss": 0.4796, "step": 13914 }, { "epoch": 0.5832975278496798, "grad_norm": 1.7150397300720215, "learning_rate": 3.90281209692119e-06, "loss": 0.5178, "step": 13915 }, { "epoch": 0.5833394464646888, "grad_norm": 1.919516682624817, "learning_rate": 3.902149800718542e-06, "loss": 0.5196, "step": 13916 }, { "epoch": 0.5833813650796977, "grad_norm": 1.6721019744873047, "learning_rate": 3.9014875247532426e-06, "loss": 0.4716, "step": 13917 }, { "epoch": 0.5834232836947068, "grad_norm": 1.8094955682754517, "learning_rate": 3.900825269037502e-06, "loss": 0.5247, "step": 13918 }, { "epoch": 0.5834652023097157, "grad_norm": 1.982019066810608, "learning_rate": 3.900163033583529e-06, "loss": 0.4273, "step": 13919 }, { "epoch": 0.5835071209247247, "grad_norm": 1.7270872592926025, "learning_rate": 3.899500818403528e-06, "loss": 0.4966, "step": 13920 }, { "epoch": 0.5835490395397336, "grad_norm": 1.6091246604919434, "learning_rate": 3.8988386235097096e-06, "loss": 0.4803, "step": 13921 }, { "epoch": 0.5835909581547426, "grad_norm": 1.6692562103271484, "learning_rate": 3.89817644891428e-06, "loss": 0.5335, "step": 13922 }, { "epoch": 0.5836328767697515, "grad_norm": 1.7812634706497192, "learning_rate": 3.897514294629442e-06, "loss": 0.4681, "step": 13923 }, { "epoch": 0.5836747953847605, "grad_norm": 1.8393452167510986, "learning_rate": 3.8968521606674035e-06, "loss": 0.4821, "step": 13924 }, { "epoch": 0.5837167139997694, "grad_norm": 2.274698257446289, "learning_rate": 3.896190047040373e-06, "loss": 0.5069, "step": 13925 }, { "epoch": 0.5837586326147784, "grad_norm": 1.7115881443023682, "learning_rate": 3.895527953760552e-06, "loss": 0.502, "step": 13926 }, { "epoch": 0.5838005512297874, "grad_norm": 2.2920405864715576, "learning_rate": 3.8948658808401445e-06, "loss": 0.5343, "step": 13927 }, { "epoch": 0.5838424698447964, "grad_norm": 1.8552597761154175, "learning_rate": 3.89420382829136e-06, "loss": 0.5009, "step": 13928 }, { "epoch": 0.5838843884598053, "grad_norm": 1.834137201309204, "learning_rate": 3.8935417961263964e-06, "loss": 0.4842, "step": 13929 }, { "epoch": 0.5839263070748143, "grad_norm": 1.7833003997802734, "learning_rate": 3.8928797843574606e-06, "loss": 0.4508, "step": 13930 }, { "epoch": 0.5839682256898232, "grad_norm": 1.737820029258728, "learning_rate": 3.892217792996758e-06, "loss": 0.4442, "step": 13931 }, { "epoch": 0.5840101443048322, "grad_norm": 1.5495527982711792, "learning_rate": 3.8915558220564854e-06, "loss": 0.4508, "step": 13932 }, { "epoch": 0.5840520629198411, "grad_norm": 2.3072032928466797, "learning_rate": 3.890893871548852e-06, "loss": 0.4833, "step": 13933 }, { "epoch": 0.58409398153485, "grad_norm": 1.7812559604644775, "learning_rate": 3.890231941486053e-06, "loss": 0.4844, "step": 13934 }, { "epoch": 0.584135900149859, "grad_norm": 1.837844967842102, "learning_rate": 3.889570031880296e-06, "loss": 0.5504, "step": 13935 }, { "epoch": 0.584177818764868, "grad_norm": 2.1990737915039062, "learning_rate": 3.888908142743782e-06, "loss": 0.5564, "step": 13936 }, { "epoch": 0.584219737379877, "grad_norm": 1.6088201999664307, "learning_rate": 3.8882462740887075e-06, "loss": 0.5396, "step": 13937 }, { "epoch": 0.5842616559948859, "grad_norm": 1.942000150680542, "learning_rate": 3.887584425927276e-06, "loss": 0.4954, "step": 13938 }, { "epoch": 0.5843035746098949, "grad_norm": 1.6430819034576416, "learning_rate": 3.886922598271691e-06, "loss": 0.5256, "step": 13939 }, { "epoch": 0.5843454932249038, "grad_norm": 2.140345573425293, "learning_rate": 3.886260791134146e-06, "loss": 0.5167, "step": 13940 }, { "epoch": 0.5843874118399128, "grad_norm": 1.583509922027588, "learning_rate": 3.885599004526845e-06, "loss": 0.4779, "step": 13941 }, { "epoch": 0.5844293304549217, "grad_norm": 1.4792611598968506, "learning_rate": 3.884937238461987e-06, "loss": 0.4902, "step": 13942 }, { "epoch": 0.5844712490699308, "grad_norm": 1.7755959033966064, "learning_rate": 3.884275492951766e-06, "loss": 0.4865, "step": 13943 }, { "epoch": 0.5845131676849397, "grad_norm": 1.6534639596939087, "learning_rate": 3.883613768008385e-06, "loss": 0.4884, "step": 13944 }, { "epoch": 0.5845550862999487, "grad_norm": 1.5685169696807861, "learning_rate": 3.8829520636440426e-06, "loss": 0.5338, "step": 13945 }, { "epoch": 0.5845970049149576, "grad_norm": 1.6946014165878296, "learning_rate": 3.882290379870934e-06, "loss": 0.5084, "step": 13946 }, { "epoch": 0.5846389235299666, "grad_norm": 1.7408764362335205, "learning_rate": 3.881628716701256e-06, "loss": 0.4688, "step": 13947 }, { "epoch": 0.5846808421449755, "grad_norm": 1.7865774631500244, "learning_rate": 3.880967074147208e-06, "loss": 0.4779, "step": 13948 }, { "epoch": 0.5847227607599845, "grad_norm": 4.746894359588623, "learning_rate": 3.880305452220983e-06, "loss": 0.4753, "step": 13949 }, { "epoch": 0.5847646793749934, "grad_norm": 1.656231164932251, "learning_rate": 3.87964385093478e-06, "loss": 0.5162, "step": 13950 }, { "epoch": 0.5848065979900025, "grad_norm": 1.9117640256881714, "learning_rate": 3.878982270300796e-06, "loss": 0.5203, "step": 13951 }, { "epoch": 0.5848485166050114, "grad_norm": 1.7174360752105713, "learning_rate": 3.878320710331221e-06, "loss": 0.4794, "step": 13952 }, { "epoch": 0.5848904352200204, "grad_norm": 1.8423885107040405, "learning_rate": 3.877659171038252e-06, "loss": 0.4185, "step": 13953 }, { "epoch": 0.5849323538350293, "grad_norm": 1.7938250303268433, "learning_rate": 3.876997652434088e-06, "loss": 0.5407, "step": 13954 }, { "epoch": 0.5849742724500383, "grad_norm": 1.7235101461410522, "learning_rate": 3.8763361545309195e-06, "loss": 0.4724, "step": 13955 }, { "epoch": 0.5850161910650472, "grad_norm": 1.482199788093567, "learning_rate": 3.875674677340938e-06, "loss": 0.4589, "step": 13956 }, { "epoch": 0.5850581096800562, "grad_norm": 1.554215431213379, "learning_rate": 3.875013220876342e-06, "loss": 0.4399, "step": 13957 }, { "epoch": 0.5851000282950651, "grad_norm": 1.6294312477111816, "learning_rate": 3.8743517851493205e-06, "loss": 0.4916, "step": 13958 }, { "epoch": 0.585141946910074, "grad_norm": 1.654642939567566, "learning_rate": 3.873690370172068e-06, "loss": 0.5164, "step": 13959 }, { "epoch": 0.5851838655250831, "grad_norm": 1.732265830039978, "learning_rate": 3.8730289759567785e-06, "loss": 0.4698, "step": 13960 }, { "epoch": 0.585225784140092, "grad_norm": 2.3623836040496826, "learning_rate": 3.8723676025156385e-06, "loss": 0.4922, "step": 13961 }, { "epoch": 0.585267702755101, "grad_norm": 1.8095625638961792, "learning_rate": 3.871706249860844e-06, "loss": 0.5228, "step": 13962 }, { "epoch": 0.5853096213701099, "grad_norm": 1.8033583164215088, "learning_rate": 3.871044918004586e-06, "loss": 0.5189, "step": 13963 }, { "epoch": 0.5853515399851189, "grad_norm": 1.9570597410202026, "learning_rate": 3.870383606959054e-06, "loss": 0.5095, "step": 13964 }, { "epoch": 0.5853934586001278, "grad_norm": 1.6131997108459473, "learning_rate": 3.869722316736437e-06, "loss": 0.4937, "step": 13965 }, { "epoch": 0.5854353772151368, "grad_norm": 2.669628143310547, "learning_rate": 3.869061047348929e-06, "loss": 0.4929, "step": 13966 }, { "epoch": 0.5854772958301457, "grad_norm": 1.7162774801254272, "learning_rate": 3.8683997988087144e-06, "loss": 0.4804, "step": 13967 }, { "epoch": 0.5855192144451548, "grad_norm": 1.6381042003631592, "learning_rate": 3.867738571127987e-06, "loss": 0.4828, "step": 13968 }, { "epoch": 0.5855611330601637, "grad_norm": 1.5657907724380493, "learning_rate": 3.867077364318934e-06, "loss": 0.478, "step": 13969 }, { "epoch": 0.5856030516751727, "grad_norm": 2.515899181365967, "learning_rate": 3.8664161783937415e-06, "loss": 0.5178, "step": 13970 }, { "epoch": 0.5856449702901816, "grad_norm": 1.9101136922836304, "learning_rate": 3.865755013364601e-06, "loss": 0.4744, "step": 13971 }, { "epoch": 0.5856868889051906, "grad_norm": 1.672728180885315, "learning_rate": 3.865093869243699e-06, "loss": 0.4806, "step": 13972 }, { "epoch": 0.5857288075201995, "grad_norm": 1.965901494026184, "learning_rate": 3.864432746043222e-06, "loss": 0.4672, "step": 13973 }, { "epoch": 0.5857707261352085, "grad_norm": 1.9464219808578491, "learning_rate": 3.863771643775357e-06, "loss": 0.5057, "step": 13974 }, { "epoch": 0.5858126447502174, "grad_norm": 1.7591267824172974, "learning_rate": 3.863110562452293e-06, "loss": 0.4845, "step": 13975 }, { "epoch": 0.5858545633652265, "grad_norm": 1.8278816938400269, "learning_rate": 3.862449502086212e-06, "loss": 0.5538, "step": 13976 }, { "epoch": 0.5858964819802354, "grad_norm": 1.9113759994506836, "learning_rate": 3.861788462689302e-06, "loss": 0.478, "step": 13977 }, { "epoch": 0.5859384005952444, "grad_norm": 1.965996503829956, "learning_rate": 3.8611274442737486e-06, "loss": 0.4501, "step": 13978 }, { "epoch": 0.5859803192102533, "grad_norm": 1.7035601139068604, "learning_rate": 3.860466446851737e-06, "loss": 0.4494, "step": 13979 }, { "epoch": 0.5860222378252623, "grad_norm": 2.1714460849761963, "learning_rate": 3.859805470435449e-06, "loss": 0.5541, "step": 13980 }, { "epoch": 0.5860641564402712, "grad_norm": 2.107065439224243, "learning_rate": 3.859144515037074e-06, "loss": 0.5039, "step": 13981 }, { "epoch": 0.5861060750552802, "grad_norm": 1.8445419073104858, "learning_rate": 3.858483580668791e-06, "loss": 0.4991, "step": 13982 }, { "epoch": 0.5861479936702891, "grad_norm": 1.882513403892517, "learning_rate": 3.857822667342784e-06, "loss": 0.4961, "step": 13983 }, { "epoch": 0.586189912285298, "grad_norm": 2.7005388736724854, "learning_rate": 3.85716177507124e-06, "loss": 0.5132, "step": 13984 }, { "epoch": 0.5862318309003071, "grad_norm": 1.6949090957641602, "learning_rate": 3.856500903866336e-06, "loss": 0.5472, "step": 13985 }, { "epoch": 0.586273749515316, "grad_norm": 1.9698987007141113, "learning_rate": 3.855840053740258e-06, "loss": 0.512, "step": 13986 }, { "epoch": 0.586315668130325, "grad_norm": 2.694011688232422, "learning_rate": 3.855179224705187e-06, "loss": 0.4786, "step": 13987 }, { "epoch": 0.5863575867453339, "grad_norm": 1.9490162134170532, "learning_rate": 3.854518416773304e-06, "loss": 0.5177, "step": 13988 }, { "epoch": 0.5863995053603429, "grad_norm": 1.990309715270996, "learning_rate": 3.85385762995679e-06, "loss": 0.4821, "step": 13989 }, { "epoch": 0.5864414239753518, "grad_norm": 1.9562698602676392, "learning_rate": 3.853196864267828e-06, "loss": 0.5286, "step": 13990 }, { "epoch": 0.5864833425903608, "grad_norm": 2.0099589824676514, "learning_rate": 3.852536119718594e-06, "loss": 0.4667, "step": 13991 }, { "epoch": 0.5865252612053697, "grad_norm": 1.8817803859710693, "learning_rate": 3.851875396321272e-06, "loss": 0.5401, "step": 13992 }, { "epoch": 0.5865671798203788, "grad_norm": 1.91103196144104, "learning_rate": 3.8512146940880396e-06, "loss": 0.544, "step": 13993 }, { "epoch": 0.5866090984353877, "grad_norm": 2.174916982650757, "learning_rate": 3.850554013031075e-06, "loss": 0.5156, "step": 13994 }, { "epoch": 0.5866510170503967, "grad_norm": 1.9578195810317993, "learning_rate": 3.849893353162558e-06, "loss": 0.4719, "step": 13995 }, { "epoch": 0.5866929356654056, "grad_norm": 1.710908055305481, "learning_rate": 3.849232714494669e-06, "loss": 0.453, "step": 13996 }, { "epoch": 0.5867348542804146, "grad_norm": 1.7209581136703491, "learning_rate": 3.848572097039583e-06, "loss": 0.5231, "step": 13997 }, { "epoch": 0.5867767728954235, "grad_norm": 1.7582552433013916, "learning_rate": 3.847911500809478e-06, "loss": 0.4337, "step": 13998 }, { "epoch": 0.5868186915104325, "grad_norm": 1.9808809757232666, "learning_rate": 3.847250925816533e-06, "loss": 0.5007, "step": 13999 }, { "epoch": 0.5868606101254414, "grad_norm": 1.611243724822998, "learning_rate": 3.846590372072922e-06, "loss": 0.4669, "step": 14000 }, { "epoch": 0.5869025287404505, "grad_norm": 2.0670037269592285, "learning_rate": 3.845929839590824e-06, "loss": 0.4619, "step": 14001 }, { "epoch": 0.5869444473554594, "grad_norm": 2.3048079013824463, "learning_rate": 3.845269328382415e-06, "loss": 0.477, "step": 14002 }, { "epoch": 0.5869863659704684, "grad_norm": 2.4319634437561035, "learning_rate": 3.844608838459867e-06, "loss": 0.4808, "step": 14003 }, { "epoch": 0.5870282845854773, "grad_norm": 2.2767856121063232, "learning_rate": 3.8439483698353585e-06, "loss": 0.512, "step": 14004 }, { "epoch": 0.5870702032004863, "grad_norm": 1.6761822700500488, "learning_rate": 3.843287922521066e-06, "loss": 0.547, "step": 14005 }, { "epoch": 0.5871121218154952, "grad_norm": 2.110887289047241, "learning_rate": 3.842627496529159e-06, "loss": 0.502, "step": 14006 }, { "epoch": 0.5871540404305042, "grad_norm": 1.807735562324524, "learning_rate": 3.8419670918718146e-06, "loss": 0.5391, "step": 14007 }, { "epoch": 0.5871959590455131, "grad_norm": 1.842809796333313, "learning_rate": 3.841306708561207e-06, "loss": 0.4709, "step": 14008 }, { "epoch": 0.587237877660522, "grad_norm": 2.0182459354400635, "learning_rate": 3.840646346609507e-06, "loss": 0.5002, "step": 14009 }, { "epoch": 0.5872797962755311, "grad_norm": 2.0686850547790527, "learning_rate": 3.8399860060288905e-06, "loss": 0.4815, "step": 14010 }, { "epoch": 0.58732171489054, "grad_norm": 2.524101734161377, "learning_rate": 3.839325686831527e-06, "loss": 0.4722, "step": 14011 }, { "epoch": 0.587363633505549, "grad_norm": 1.6884479522705078, "learning_rate": 3.838665389029589e-06, "loss": 0.5074, "step": 14012 }, { "epoch": 0.5874055521205579, "grad_norm": 2.373789072036743, "learning_rate": 3.838005112635252e-06, "loss": 0.4681, "step": 14013 }, { "epoch": 0.5874474707355669, "grad_norm": 3.068854808807373, "learning_rate": 3.8373448576606815e-06, "loss": 0.5141, "step": 14014 }, { "epoch": 0.5874893893505758, "grad_norm": 2.2636115550994873, "learning_rate": 3.836684624118052e-06, "loss": 0.4543, "step": 14015 }, { "epoch": 0.5875313079655848, "grad_norm": 2.4683408737182617, "learning_rate": 3.8360244120195345e-06, "loss": 0.4862, "step": 14016 }, { "epoch": 0.5875732265805937, "grad_norm": 2.4957258701324463, "learning_rate": 3.835364221377296e-06, "loss": 0.5175, "step": 14017 }, { "epoch": 0.5876151451956028, "grad_norm": 2.2843668460845947, "learning_rate": 3.834704052203508e-06, "loss": 0.4977, "step": 14018 }, { "epoch": 0.5876570638106117, "grad_norm": 2.1849329471588135, "learning_rate": 3.834043904510342e-06, "loss": 0.4975, "step": 14019 }, { "epoch": 0.5876989824256207, "grad_norm": 1.8112231492996216, "learning_rate": 3.833383778309963e-06, "loss": 0.4738, "step": 14020 }, { "epoch": 0.5877409010406296, "grad_norm": 1.740668773651123, "learning_rate": 3.8327236736145406e-06, "loss": 0.5037, "step": 14021 }, { "epoch": 0.5877828196556386, "grad_norm": 2.845914363861084, "learning_rate": 3.832063590436246e-06, "loss": 0.4692, "step": 14022 }, { "epoch": 0.5878247382706475, "grad_norm": 1.6752605438232422, "learning_rate": 3.831403528787242e-06, "loss": 0.4677, "step": 14023 }, { "epoch": 0.5878666568856565, "grad_norm": 1.8070601224899292, "learning_rate": 3.8307434886797e-06, "loss": 0.4678, "step": 14024 }, { "epoch": 0.5879085755006654, "grad_norm": 1.6528922319412231, "learning_rate": 3.830083470125785e-06, "loss": 0.451, "step": 14025 }, { "epoch": 0.5879504941156745, "grad_norm": 1.9519010782241821, "learning_rate": 3.829423473137663e-06, "loss": 0.4787, "step": 14026 }, { "epoch": 0.5879924127306834, "grad_norm": 1.5443272590637207, "learning_rate": 3.828763497727501e-06, "loss": 0.4783, "step": 14027 }, { "epoch": 0.5880343313456924, "grad_norm": 2.1888020038604736, "learning_rate": 3.828103543907467e-06, "loss": 0.4994, "step": 14028 }, { "epoch": 0.5880762499607013, "grad_norm": 1.7803153991699219, "learning_rate": 3.827443611689721e-06, "loss": 0.4463, "step": 14029 }, { "epoch": 0.5881181685757103, "grad_norm": 1.8332277536392212, "learning_rate": 3.8267837010864325e-06, "loss": 0.5068, "step": 14030 }, { "epoch": 0.5881600871907192, "grad_norm": 1.7521812915802002, "learning_rate": 3.826123812109766e-06, "loss": 0.4354, "step": 14031 }, { "epoch": 0.5882020058057282, "grad_norm": 2.6904475688934326, "learning_rate": 3.825463944771881e-06, "loss": 0.4734, "step": 14032 }, { "epoch": 0.5882439244207371, "grad_norm": 1.693677544593811, "learning_rate": 3.8248040990849455e-06, "loss": 0.5201, "step": 14033 }, { "epoch": 0.588285843035746, "grad_norm": 1.8914893865585327, "learning_rate": 3.824144275061124e-06, "loss": 0.4465, "step": 14034 }, { "epoch": 0.5883277616507551, "grad_norm": 1.6788010597229004, "learning_rate": 3.823484472712575e-06, "loss": 0.4931, "step": 14035 }, { "epoch": 0.588369680265764, "grad_norm": 1.583319067955017, "learning_rate": 3.8228246920514634e-06, "loss": 0.5018, "step": 14036 }, { "epoch": 0.588411598880773, "grad_norm": 3.1314139366149902, "learning_rate": 3.8221649330899525e-06, "loss": 0.4832, "step": 14037 }, { "epoch": 0.5884535174957819, "grad_norm": 1.9602469205856323, "learning_rate": 3.821505195840202e-06, "loss": 0.5168, "step": 14038 }, { "epoch": 0.5884954361107909, "grad_norm": 1.7572691440582275, "learning_rate": 3.8208454803143735e-06, "loss": 0.5056, "step": 14039 }, { "epoch": 0.5885373547257998, "grad_norm": 2.077211618423462, "learning_rate": 3.82018578652463e-06, "loss": 0.5306, "step": 14040 }, { "epoch": 0.5885792733408088, "grad_norm": 2.104067087173462, "learning_rate": 3.819526114483128e-06, "loss": 0.5576, "step": 14041 }, { "epoch": 0.5886211919558177, "grad_norm": 1.7975479364395142, "learning_rate": 3.818866464202031e-06, "loss": 0.4786, "step": 14042 }, { "epoch": 0.5886631105708268, "grad_norm": 1.6958215236663818, "learning_rate": 3.8182068356935e-06, "loss": 0.5051, "step": 14043 }, { "epoch": 0.5887050291858357, "grad_norm": 1.7658774852752686, "learning_rate": 3.81754722896969e-06, "loss": 0.4936, "step": 14044 }, { "epoch": 0.5887469478008447, "grad_norm": 1.6527771949768066, "learning_rate": 3.816887644042762e-06, "loss": 0.4924, "step": 14045 }, { "epoch": 0.5887888664158536, "grad_norm": 2.247000217437744, "learning_rate": 3.816228080924876e-06, "loss": 0.5319, "step": 14046 }, { "epoch": 0.5888307850308626, "grad_norm": 2.191347599029541, "learning_rate": 3.8155685396281874e-06, "loss": 0.4643, "step": 14047 }, { "epoch": 0.5888727036458715, "grad_norm": 1.8225975036621094, "learning_rate": 3.814909020164856e-06, "loss": 0.5023, "step": 14048 }, { "epoch": 0.5889146222608805, "grad_norm": 1.7612967491149902, "learning_rate": 3.8142495225470407e-06, "loss": 0.4591, "step": 14049 }, { "epoch": 0.5889565408758894, "grad_norm": 1.5551735162734985, "learning_rate": 3.813590046786893e-06, "loss": 0.4515, "step": 14050 }, { "epoch": 0.5889984594908985, "grad_norm": 1.8895686864852905, "learning_rate": 3.8129305928965725e-06, "loss": 0.5363, "step": 14051 }, { "epoch": 0.5890403781059074, "grad_norm": 1.780521273612976, "learning_rate": 3.812271160888238e-06, "loss": 0.511, "step": 14052 }, { "epoch": 0.5890822967209164, "grad_norm": 2.2673823833465576, "learning_rate": 3.8116117507740413e-06, "loss": 0.5108, "step": 14053 }, { "epoch": 0.5891242153359253, "grad_norm": 1.8628482818603516, "learning_rate": 3.8109523625661384e-06, "loss": 0.4907, "step": 14054 }, { "epoch": 0.5891661339509343, "grad_norm": 2.0813653469085693, "learning_rate": 3.8102929962766865e-06, "loss": 0.5387, "step": 14055 }, { "epoch": 0.5892080525659432, "grad_norm": 1.6093733310699463, "learning_rate": 3.8096336519178365e-06, "loss": 0.5266, "step": 14056 }, { "epoch": 0.5892499711809522, "grad_norm": 1.9522120952606201, "learning_rate": 3.808974329501746e-06, "loss": 0.5148, "step": 14057 }, { "epoch": 0.5892918897959611, "grad_norm": 4.540215015411377, "learning_rate": 3.808315029040568e-06, "loss": 0.5052, "step": 14058 }, { "epoch": 0.58933380841097, "grad_norm": 2.039928436279297, "learning_rate": 3.807655750546453e-06, "loss": 0.5065, "step": 14059 }, { "epoch": 0.5893757270259791, "grad_norm": 2.1102283000946045, "learning_rate": 3.8069964940315557e-06, "loss": 0.5143, "step": 14060 }, { "epoch": 0.589417645640988, "grad_norm": 1.9635335206985474, "learning_rate": 3.8063372595080305e-06, "loss": 0.4776, "step": 14061 }, { "epoch": 0.589459564255997, "grad_norm": 2.46325421333313, "learning_rate": 3.805678046988027e-06, "loss": 0.5188, "step": 14062 }, { "epoch": 0.5895014828710059, "grad_norm": 1.6986199617385864, "learning_rate": 3.8050188564836967e-06, "loss": 0.519, "step": 14063 }, { "epoch": 0.5895434014860149, "grad_norm": 1.7182698249816895, "learning_rate": 3.804359688007194e-06, "loss": 0.4869, "step": 14064 }, { "epoch": 0.5895853201010238, "grad_norm": 1.6652592420578003, "learning_rate": 3.803700541570665e-06, "loss": 0.4971, "step": 14065 }, { "epoch": 0.5896272387160328, "grad_norm": 2.1149275302886963, "learning_rate": 3.803041417186264e-06, "loss": 0.5011, "step": 14066 }, { "epoch": 0.5896691573310417, "grad_norm": 1.9925177097320557, "learning_rate": 3.802382314866141e-06, "loss": 0.5152, "step": 14067 }, { "epoch": 0.5897110759460508, "grad_norm": 2.102508306503296, "learning_rate": 3.801723234622442e-06, "loss": 0.5053, "step": 14068 }, { "epoch": 0.5897529945610597, "grad_norm": 1.6292699575424194, "learning_rate": 3.8010641764673182e-06, "loss": 0.4921, "step": 14069 }, { "epoch": 0.5897949131760687, "grad_norm": 1.6508103609085083, "learning_rate": 3.8004051404129214e-06, "loss": 0.4902, "step": 14070 }, { "epoch": 0.5898368317910776, "grad_norm": 2.110546350479126, "learning_rate": 3.7997461264713963e-06, "loss": 0.4896, "step": 14071 }, { "epoch": 0.5898787504060866, "grad_norm": 3.242966413497925, "learning_rate": 3.7990871346548906e-06, "loss": 0.4702, "step": 14072 }, { "epoch": 0.5899206690210955, "grad_norm": 1.9139716625213623, "learning_rate": 3.7984281649755562e-06, "loss": 0.5371, "step": 14073 }, { "epoch": 0.5899625876361045, "grad_norm": 2.1393580436706543, "learning_rate": 3.797769217445535e-06, "loss": 0.493, "step": 14074 }, { "epoch": 0.5900045062511134, "grad_norm": 1.8706177473068237, "learning_rate": 3.7971102920769776e-06, "loss": 0.4843, "step": 14075 }, { "epoch": 0.5900464248661225, "grad_norm": 2.07897686958313, "learning_rate": 3.796451388882029e-06, "loss": 0.52, "step": 14076 }, { "epoch": 0.5900883434811314, "grad_norm": 2.1496329307556152, "learning_rate": 3.795792507872834e-06, "loss": 0.4816, "step": 14077 }, { "epoch": 0.5901302620961404, "grad_norm": 1.7882089614868164, "learning_rate": 3.7951336490615398e-06, "loss": 0.511, "step": 14078 }, { "epoch": 0.5901721807111493, "grad_norm": 1.8524831533432007, "learning_rate": 3.7944748124602925e-06, "loss": 0.5141, "step": 14079 }, { "epoch": 0.5902140993261583, "grad_norm": 1.9300198554992676, "learning_rate": 3.7938159980812332e-06, "loss": 0.5164, "step": 14080 }, { "epoch": 0.5902560179411672, "grad_norm": 1.8409115076065063, "learning_rate": 3.79315720593651e-06, "loss": 0.4907, "step": 14081 }, { "epoch": 0.5902979365561762, "grad_norm": 1.8810319900512695, "learning_rate": 3.792498436038266e-06, "loss": 0.4784, "step": 14082 }, { "epoch": 0.5903398551711851, "grad_norm": 1.8044360876083374, "learning_rate": 3.791839688398642e-06, "loss": 0.467, "step": 14083 }, { "epoch": 0.590381773786194, "grad_norm": 2.5039660930633545, "learning_rate": 3.791180963029786e-06, "loss": 0.5079, "step": 14084 }, { "epoch": 0.5904236924012031, "grad_norm": 2.2604739665985107, "learning_rate": 3.7905222599438348e-06, "loss": 0.5377, "step": 14085 }, { "epoch": 0.590465611016212, "grad_norm": 2.009162187576294, "learning_rate": 3.789863579152935e-06, "loss": 0.5011, "step": 14086 }, { "epoch": 0.590507529631221, "grad_norm": 4.842443466186523, "learning_rate": 3.7892049206692277e-06, "loss": 0.461, "step": 14087 }, { "epoch": 0.5905494482462299, "grad_norm": 2.157266855239868, "learning_rate": 3.788546284504852e-06, "loss": 0.4847, "step": 14088 }, { "epoch": 0.5905913668612389, "grad_norm": 1.689813256263733, "learning_rate": 3.787887670671951e-06, "loss": 0.4744, "step": 14089 }, { "epoch": 0.5906332854762478, "grad_norm": 2.119542360305786, "learning_rate": 3.787229079182667e-06, "loss": 0.5412, "step": 14090 }, { "epoch": 0.5906752040912568, "grad_norm": 2.2042527198791504, "learning_rate": 3.7865705100491375e-06, "loss": 0.4956, "step": 14091 }, { "epoch": 0.5907171227062658, "grad_norm": 1.7685205936431885, "learning_rate": 3.785911963283502e-06, "loss": 0.4982, "step": 14092 }, { "epoch": 0.5907590413212748, "grad_norm": 1.8812012672424316, "learning_rate": 3.785253438897903e-06, "loss": 0.5069, "step": 14093 }, { "epoch": 0.5908009599362837, "grad_norm": 1.8602937459945679, "learning_rate": 3.7845949369044753e-06, "loss": 0.4863, "step": 14094 }, { "epoch": 0.5908428785512927, "grad_norm": 1.9743456840515137, "learning_rate": 3.783936457315361e-06, "loss": 0.5259, "step": 14095 }, { "epoch": 0.5908847971663016, "grad_norm": 1.844374418258667, "learning_rate": 3.7832780001426976e-06, "loss": 0.5349, "step": 14096 }, { "epoch": 0.5909267157813106, "grad_norm": 1.6552563905715942, "learning_rate": 3.782619565398621e-06, "loss": 0.4514, "step": 14097 }, { "epoch": 0.5909686343963195, "grad_norm": 1.7140370607376099, "learning_rate": 3.7819611530952687e-06, "loss": 0.5, "step": 14098 }, { "epoch": 0.5910105530113285, "grad_norm": 8.166871070861816, "learning_rate": 3.781302763244781e-06, "loss": 0.4965, "step": 14099 }, { "epoch": 0.5910524716263375, "grad_norm": 2.0086119174957275, "learning_rate": 3.7806443958592913e-06, "loss": 0.4935, "step": 14100 }, { "epoch": 0.5910943902413465, "grad_norm": 1.771751880645752, "learning_rate": 3.7799860509509358e-06, "loss": 0.4498, "step": 14101 }, { "epoch": 0.5911363088563554, "grad_norm": 1.945280909538269, "learning_rate": 3.7793277285318517e-06, "loss": 0.4349, "step": 14102 }, { "epoch": 0.5911782274713644, "grad_norm": 2.122053861618042, "learning_rate": 3.778669428614172e-06, "loss": 0.5502, "step": 14103 }, { "epoch": 0.5912201460863733, "grad_norm": 1.9460936784744263, "learning_rate": 3.778011151210035e-06, "loss": 0.5387, "step": 14104 }, { "epoch": 0.5912620647013823, "grad_norm": 1.565618872642517, "learning_rate": 3.777352896331573e-06, "loss": 0.4666, "step": 14105 }, { "epoch": 0.5913039833163912, "grad_norm": 1.6455210447311401, "learning_rate": 3.776694663990918e-06, "loss": 0.5024, "step": 14106 }, { "epoch": 0.5913459019314002, "grad_norm": 1.7412729263305664, "learning_rate": 3.776036454200206e-06, "loss": 0.4206, "step": 14107 }, { "epoch": 0.5913878205464091, "grad_norm": 1.5390427112579346, "learning_rate": 3.775378266971572e-06, "loss": 0.4616, "step": 14108 }, { "epoch": 0.591429739161418, "grad_norm": 1.7338299751281738, "learning_rate": 3.7747201023171456e-06, "loss": 0.4434, "step": 14109 }, { "epoch": 0.5914716577764271, "grad_norm": 3.2773330211639404, "learning_rate": 3.7740619602490587e-06, "loss": 0.4842, "step": 14110 }, { "epoch": 0.591513576391436, "grad_norm": 1.7734659910202026, "learning_rate": 3.773403840779447e-06, "loss": 0.4684, "step": 14111 }, { "epoch": 0.591555495006445, "grad_norm": 1.8059430122375488, "learning_rate": 3.772745743920438e-06, "loss": 0.5112, "step": 14112 }, { "epoch": 0.5915974136214539, "grad_norm": 1.7913708686828613, "learning_rate": 3.7720876696841658e-06, "loss": 0.5379, "step": 14113 }, { "epoch": 0.5916393322364629, "grad_norm": 1.630046010017395, "learning_rate": 3.7714296180827603e-06, "loss": 0.4799, "step": 14114 }, { "epoch": 0.5916812508514718, "grad_norm": 1.8210021257400513, "learning_rate": 3.7707715891283498e-06, "loss": 0.4835, "step": 14115 }, { "epoch": 0.5917231694664808, "grad_norm": 2.1138784885406494, "learning_rate": 3.7701135828330654e-06, "loss": 0.536, "step": 14116 }, { "epoch": 0.5917650880814898, "grad_norm": 1.8217114210128784, "learning_rate": 3.769455599209039e-06, "loss": 0.4594, "step": 14117 }, { "epoch": 0.5918070066964988, "grad_norm": 1.9458814859390259, "learning_rate": 3.768797638268396e-06, "loss": 0.5018, "step": 14118 }, { "epoch": 0.5918489253115077, "grad_norm": 2.0612940788269043, "learning_rate": 3.7681397000232663e-06, "loss": 0.4795, "step": 14119 }, { "epoch": 0.5918908439265167, "grad_norm": 2.06561541557312, "learning_rate": 3.7674817844857797e-06, "loss": 0.4821, "step": 14120 }, { "epoch": 0.5919327625415256, "grad_norm": 1.9159964323043823, "learning_rate": 3.7668238916680604e-06, "loss": 0.4642, "step": 14121 }, { "epoch": 0.5919746811565346, "grad_norm": 1.6149101257324219, "learning_rate": 3.7661660215822394e-06, "loss": 0.4698, "step": 14122 }, { "epoch": 0.5920165997715435, "grad_norm": 2.0948309898376465, "learning_rate": 3.7655081742404413e-06, "loss": 0.4704, "step": 14123 }, { "epoch": 0.5920585183865525, "grad_norm": 1.8730272054672241, "learning_rate": 3.764850349654794e-06, "loss": 0.5036, "step": 14124 }, { "epoch": 0.5921004370015615, "grad_norm": 3.5264337062835693, "learning_rate": 3.7641925478374213e-06, "loss": 0.5085, "step": 14125 }, { "epoch": 0.5921423556165705, "grad_norm": 1.7599692344665527, "learning_rate": 3.7635347688004533e-06, "loss": 0.4736, "step": 14126 }, { "epoch": 0.5921842742315794, "grad_norm": 1.6774380207061768, "learning_rate": 3.76287701255601e-06, "loss": 0.4993, "step": 14127 }, { "epoch": 0.5922261928465884, "grad_norm": 1.9286147356033325, "learning_rate": 3.76221927911622e-06, "loss": 0.5206, "step": 14128 }, { "epoch": 0.5922681114615973, "grad_norm": 1.746678113937378, "learning_rate": 3.761561568493207e-06, "loss": 0.5449, "step": 14129 }, { "epoch": 0.5923100300766063, "grad_norm": 1.8446741104125977, "learning_rate": 3.7609038806990927e-06, "loss": 0.5112, "step": 14130 }, { "epoch": 0.5923519486916152, "grad_norm": 1.7132220268249512, "learning_rate": 3.7602462157460023e-06, "loss": 0.4839, "step": 14131 }, { "epoch": 0.5923938673066242, "grad_norm": 1.8970552682876587, "learning_rate": 3.7595885736460607e-06, "loss": 0.5257, "step": 14132 }, { "epoch": 0.5924357859216332, "grad_norm": 1.8654658794403076, "learning_rate": 3.758930954411388e-06, "loss": 0.5266, "step": 14133 }, { "epoch": 0.5924777045366421, "grad_norm": 16.46829605102539, "learning_rate": 3.758273358054107e-06, "loss": 0.5157, "step": 14134 }, { "epoch": 0.5925196231516511, "grad_norm": 2.048403739929199, "learning_rate": 3.7576157845863415e-06, "loss": 0.4472, "step": 14135 }, { "epoch": 0.59256154176666, "grad_norm": 1.6262127161026, "learning_rate": 3.75695823402021e-06, "loss": 0.4858, "step": 14136 }, { "epoch": 0.592603460381669, "grad_norm": 1.8122072219848633, "learning_rate": 3.7563007063678357e-06, "loss": 0.5077, "step": 14137 }, { "epoch": 0.5926453789966779, "grad_norm": 1.8352560997009277, "learning_rate": 3.755643201641339e-06, "loss": 0.511, "step": 14138 }, { "epoch": 0.5926872976116869, "grad_norm": 1.8413842916488647, "learning_rate": 3.7549857198528385e-06, "loss": 0.4597, "step": 14139 }, { "epoch": 0.5927292162266958, "grad_norm": 1.4361000061035156, "learning_rate": 3.7543282610144543e-06, "loss": 0.5194, "step": 14140 }, { "epoch": 0.5927711348417048, "grad_norm": 1.8625575304031372, "learning_rate": 3.753670825138309e-06, "loss": 0.4963, "step": 14141 }, { "epoch": 0.5928130534567138, "grad_norm": 1.6071968078613281, "learning_rate": 3.753013412236518e-06, "loss": 0.4508, "step": 14142 }, { "epoch": 0.5928549720717228, "grad_norm": 1.539943814277649, "learning_rate": 3.7523560223211998e-06, "loss": 0.4565, "step": 14143 }, { "epoch": 0.5928968906867317, "grad_norm": 1.863222599029541, "learning_rate": 3.7516986554044755e-06, "loss": 0.5268, "step": 14144 }, { "epoch": 0.5929388093017407, "grad_norm": 1.9476635456085205, "learning_rate": 3.7510413114984583e-06, "loss": 0.5634, "step": 14145 }, { "epoch": 0.5929807279167496, "grad_norm": 1.818443775177002, "learning_rate": 3.7503839906152693e-06, "loss": 0.5187, "step": 14146 }, { "epoch": 0.5930226465317586, "grad_norm": 1.993219017982483, "learning_rate": 3.749726692767025e-06, "loss": 0.525, "step": 14147 }, { "epoch": 0.5930645651467675, "grad_norm": 1.9029054641723633, "learning_rate": 3.7490694179658384e-06, "loss": 0.5426, "step": 14148 }, { "epoch": 0.5931064837617765, "grad_norm": 1.9833065271377563, "learning_rate": 3.7484121662238283e-06, "loss": 0.5011, "step": 14149 }, { "epoch": 0.5931484023767855, "grad_norm": 2.2327001094818115, "learning_rate": 3.7477549375531107e-06, "loss": 0.4678, "step": 14150 }, { "epoch": 0.5931903209917945, "grad_norm": 3.366273880004883, "learning_rate": 3.7470977319657997e-06, "loss": 0.4692, "step": 14151 }, { "epoch": 0.5932322396068034, "grad_norm": 2.5096254348754883, "learning_rate": 3.746440549474008e-06, "loss": 0.563, "step": 14152 }, { "epoch": 0.5932741582218124, "grad_norm": 1.607502818107605, "learning_rate": 3.7457833900898544e-06, "loss": 0.454, "step": 14153 }, { "epoch": 0.5933160768368213, "grad_norm": 2.033024549484253, "learning_rate": 3.7451262538254474e-06, "loss": 0.4498, "step": 14154 }, { "epoch": 0.5933579954518303, "grad_norm": 1.7764511108398438, "learning_rate": 3.7444691406929045e-06, "loss": 0.4748, "step": 14155 }, { "epoch": 0.5933999140668392, "grad_norm": 1.7628443241119385, "learning_rate": 3.7438120507043383e-06, "loss": 0.5063, "step": 14156 }, { "epoch": 0.5934418326818482, "grad_norm": 2.229525327682495, "learning_rate": 3.7431549838718585e-06, "loss": 0.5453, "step": 14157 }, { "epoch": 0.5934837512968572, "grad_norm": 1.8747096061706543, "learning_rate": 3.742497940207581e-06, "loss": 0.4458, "step": 14158 }, { "epoch": 0.5935256699118661, "grad_norm": 2.053321361541748, "learning_rate": 3.7418409197236137e-06, "loss": 0.5314, "step": 14159 }, { "epoch": 0.5935675885268751, "grad_norm": 1.6006944179534912, "learning_rate": 3.74118392243207e-06, "loss": 0.4216, "step": 14160 }, { "epoch": 0.593609507141884, "grad_norm": 1.967962384223938, "learning_rate": 3.7405269483450625e-06, "loss": 0.4973, "step": 14161 }, { "epoch": 0.593651425756893, "grad_norm": 1.886644721031189, "learning_rate": 3.7398699974746966e-06, "loss": 0.5455, "step": 14162 }, { "epoch": 0.5936933443719019, "grad_norm": 3.172342538833618, "learning_rate": 3.7392130698330853e-06, "loss": 0.5154, "step": 14163 }, { "epoch": 0.5937352629869109, "grad_norm": 1.8510149717330933, "learning_rate": 3.7385561654323407e-06, "loss": 0.46, "step": 14164 }, { "epoch": 0.5937771816019198, "grad_norm": 2.014798879623413, "learning_rate": 3.7378992842845673e-06, "loss": 0.4954, "step": 14165 }, { "epoch": 0.5938191002169289, "grad_norm": 3.3371050357818604, "learning_rate": 3.737242426401876e-06, "loss": 0.5003, "step": 14166 }, { "epoch": 0.5938610188319378, "grad_norm": 2.1230063438415527, "learning_rate": 3.7365855917963756e-06, "loss": 0.5385, "step": 14167 }, { "epoch": 0.5939029374469468, "grad_norm": 1.986851692199707, "learning_rate": 3.735928780480172e-06, "loss": 0.4757, "step": 14168 }, { "epoch": 0.5939448560619557, "grad_norm": 1.7016738653182983, "learning_rate": 3.735271992465375e-06, "loss": 0.481, "step": 14169 }, { "epoch": 0.5939867746769647, "grad_norm": 2.1105196475982666, "learning_rate": 3.7346152277640894e-06, "loss": 0.4886, "step": 14170 }, { "epoch": 0.5940286932919736, "grad_norm": 1.5557916164398193, "learning_rate": 3.7339584863884236e-06, "loss": 0.4951, "step": 14171 }, { "epoch": 0.5940706119069826, "grad_norm": 2.213979721069336, "learning_rate": 3.7333017683504814e-06, "loss": 0.5207, "step": 14172 }, { "epoch": 0.5941125305219915, "grad_norm": 1.7977591753005981, "learning_rate": 3.732645073662372e-06, "loss": 0.5671, "step": 14173 }, { "epoch": 0.5941544491370006, "grad_norm": 1.6977665424346924, "learning_rate": 3.7319884023361976e-06, "loss": 0.5162, "step": 14174 }, { "epoch": 0.5941963677520095, "grad_norm": 2.3054580688476562, "learning_rate": 3.7313317543840633e-06, "loss": 0.4648, "step": 14175 }, { "epoch": 0.5942382863670185, "grad_norm": 1.9077699184417725, "learning_rate": 3.7306751298180767e-06, "loss": 0.4953, "step": 14176 }, { "epoch": 0.5942802049820274, "grad_norm": 2.0727992057800293, "learning_rate": 3.7300185286503364e-06, "loss": 0.5628, "step": 14177 }, { "epoch": 0.5943221235970364, "grad_norm": 1.5832480192184448, "learning_rate": 3.729361950892949e-06, "loss": 0.486, "step": 14178 }, { "epoch": 0.5943640422120453, "grad_norm": 2.0845015048980713, "learning_rate": 3.72870539655802e-06, "loss": 0.5387, "step": 14179 }, { "epoch": 0.5944059608270543, "grad_norm": 1.6826772689819336, "learning_rate": 3.7280488656576476e-06, "loss": 0.5235, "step": 14180 }, { "epoch": 0.5944478794420632, "grad_norm": 1.7952133417129517, "learning_rate": 3.727392358203935e-06, "loss": 0.5058, "step": 14181 }, { "epoch": 0.5944897980570722, "grad_norm": 1.7064039707183838, "learning_rate": 3.7267358742089877e-06, "loss": 0.4681, "step": 14182 }, { "epoch": 0.5945317166720812, "grad_norm": 2.114521026611328, "learning_rate": 3.726079413684902e-06, "loss": 0.4981, "step": 14183 }, { "epoch": 0.5945736352870901, "grad_norm": 1.8977528810501099, "learning_rate": 3.725422976643782e-06, "loss": 0.5296, "step": 14184 }, { "epoch": 0.5946155539020991, "grad_norm": 1.9554113149642944, "learning_rate": 3.724766563097729e-06, "loss": 0.5028, "step": 14185 }, { "epoch": 0.594657472517108, "grad_norm": 2.4229540824890137, "learning_rate": 3.7241101730588387e-06, "loss": 0.4794, "step": 14186 }, { "epoch": 0.594699391132117, "grad_norm": 1.8938771486282349, "learning_rate": 3.7234538065392135e-06, "loss": 0.4791, "step": 14187 }, { "epoch": 0.5947413097471259, "grad_norm": 1.7030738592147827, "learning_rate": 3.722797463550955e-06, "loss": 0.4994, "step": 14188 }, { "epoch": 0.5947832283621349, "grad_norm": 1.6065053939819336, "learning_rate": 3.7221411441061583e-06, "loss": 0.4314, "step": 14189 }, { "epoch": 0.5948251469771438, "grad_norm": 1.6083319187164307, "learning_rate": 3.7214848482169217e-06, "loss": 0.4782, "step": 14190 }, { "epoch": 0.5948670655921529, "grad_norm": 1.8110322952270508, "learning_rate": 3.7208285758953477e-06, "loss": 0.5241, "step": 14191 }, { "epoch": 0.5949089842071618, "grad_norm": 2.8569858074188232, "learning_rate": 3.7201723271535273e-06, "loss": 0.4644, "step": 14192 }, { "epoch": 0.5949509028221708, "grad_norm": 1.6844607591629028, "learning_rate": 3.719516102003562e-06, "loss": 0.471, "step": 14193 }, { "epoch": 0.5949928214371797, "grad_norm": 1.5949786901474, "learning_rate": 3.718859900457549e-06, "loss": 0.4499, "step": 14194 }, { "epoch": 0.5950347400521887, "grad_norm": 2.6897671222686768, "learning_rate": 3.718203722527579e-06, "loss": 0.4586, "step": 14195 }, { "epoch": 0.5950766586671976, "grad_norm": 2.021881341934204, "learning_rate": 3.7175475682257527e-06, "loss": 0.483, "step": 14196 }, { "epoch": 0.5951185772822066, "grad_norm": 2.117176055908203, "learning_rate": 3.7168914375641656e-06, "loss": 0.5184, "step": 14197 }, { "epoch": 0.5951604958972155, "grad_norm": 2.3797898292541504, "learning_rate": 3.7162353305549105e-06, "loss": 0.4582, "step": 14198 }, { "epoch": 0.5952024145122246, "grad_norm": 2.522005319595337, "learning_rate": 3.7155792472100805e-06, "loss": 0.4384, "step": 14199 }, { "epoch": 0.5952443331272335, "grad_norm": 1.8372734785079956, "learning_rate": 3.7149231875417734e-06, "loss": 0.4625, "step": 14200 }, { "epoch": 0.5952862517422425, "grad_norm": 2.5704071521759033, "learning_rate": 3.7142671515620793e-06, "loss": 0.4944, "step": 14201 }, { "epoch": 0.5953281703572514, "grad_norm": 1.7879446744918823, "learning_rate": 3.713611139283093e-06, "loss": 0.4747, "step": 14202 }, { "epoch": 0.5953700889722604, "grad_norm": 1.5980937480926514, "learning_rate": 3.712955150716909e-06, "loss": 0.4999, "step": 14203 }, { "epoch": 0.5954120075872693, "grad_norm": 1.4903180599212646, "learning_rate": 3.712299185875615e-06, "loss": 0.4682, "step": 14204 }, { "epoch": 0.5954539262022783, "grad_norm": 2.256498336791992, "learning_rate": 3.7116432447713056e-06, "loss": 0.4949, "step": 14205 }, { "epoch": 0.5954958448172872, "grad_norm": 1.987391710281372, "learning_rate": 3.7109873274160736e-06, "loss": 0.5021, "step": 14206 }, { "epoch": 0.5955377634322963, "grad_norm": 1.9887832403182983, "learning_rate": 3.7103314338220076e-06, "loss": 0.507, "step": 14207 }, { "epoch": 0.5955796820473052, "grad_norm": 2.1245596408843994, "learning_rate": 3.7096755640011974e-06, "loss": 0.5067, "step": 14208 }, { "epoch": 0.5956216006623141, "grad_norm": 2.2576794624328613, "learning_rate": 3.709019717965736e-06, "loss": 0.4806, "step": 14209 }, { "epoch": 0.5956635192773231, "grad_norm": 1.6737650632858276, "learning_rate": 3.7083638957277097e-06, "loss": 0.5096, "step": 14210 }, { "epoch": 0.595705437892332, "grad_norm": 1.9534578323364258, "learning_rate": 3.7077080972992103e-06, "loss": 0.4729, "step": 14211 }, { "epoch": 0.595747356507341, "grad_norm": 1.6504565477371216, "learning_rate": 3.7070523226923267e-06, "loss": 0.5389, "step": 14212 }, { "epoch": 0.5957892751223499, "grad_norm": 5.740548133850098, "learning_rate": 3.706396571919144e-06, "loss": 0.4847, "step": 14213 }, { "epoch": 0.5958311937373589, "grad_norm": 2.0288283824920654, "learning_rate": 3.705740844991752e-06, "loss": 0.5516, "step": 14214 }, { "epoch": 0.5958731123523678, "grad_norm": 2.853410482406616, "learning_rate": 3.7050851419222407e-06, "loss": 0.492, "step": 14215 }, { "epoch": 0.5959150309673769, "grad_norm": 2.6410439014434814, "learning_rate": 3.7044294627226927e-06, "loss": 0.4938, "step": 14216 }, { "epoch": 0.5959569495823858, "grad_norm": 1.8206377029418945, "learning_rate": 3.7037738074051964e-06, "loss": 0.5252, "step": 14217 }, { "epoch": 0.5959988681973948, "grad_norm": 2.332383632659912, "learning_rate": 3.70311817598184e-06, "loss": 0.5291, "step": 14218 }, { "epoch": 0.5960407868124037, "grad_norm": 1.9782371520996094, "learning_rate": 3.7024625684647054e-06, "loss": 0.4979, "step": 14219 }, { "epoch": 0.5960827054274127, "grad_norm": 2.1449739933013916, "learning_rate": 3.70180698486588e-06, "loss": 0.5401, "step": 14220 }, { "epoch": 0.5961246240424216, "grad_norm": 1.733422040939331, "learning_rate": 3.7011514251974477e-06, "loss": 0.5728, "step": 14221 }, { "epoch": 0.5961665426574306, "grad_norm": 7.546350479125977, "learning_rate": 3.700495889471494e-06, "loss": 0.5271, "step": 14222 }, { "epoch": 0.5962084612724395, "grad_norm": 2.3818109035491943, "learning_rate": 3.6998403777001002e-06, "loss": 0.4943, "step": 14223 }, { "epoch": 0.5962503798874486, "grad_norm": 1.9561586380004883, "learning_rate": 3.6991848898953547e-06, "loss": 0.4833, "step": 14224 }, { "epoch": 0.5962922985024575, "grad_norm": 2.1435534954071045, "learning_rate": 3.6985294260693348e-06, "loss": 0.5541, "step": 14225 }, { "epoch": 0.5963342171174665, "grad_norm": 2.2080390453338623, "learning_rate": 3.697873986234127e-06, "loss": 0.4706, "step": 14226 }, { "epoch": 0.5963761357324754, "grad_norm": 2.2333083152770996, "learning_rate": 3.6972185704018127e-06, "loss": 0.512, "step": 14227 }, { "epoch": 0.5964180543474844, "grad_norm": 2.084533214569092, "learning_rate": 3.6965631785844713e-06, "loss": 0.5416, "step": 14228 }, { "epoch": 0.5964599729624933, "grad_norm": 2.3538031578063965, "learning_rate": 3.695907810794186e-06, "loss": 0.5229, "step": 14229 }, { "epoch": 0.5965018915775023, "grad_norm": 1.7543469667434692, "learning_rate": 3.695252467043039e-06, "loss": 0.4837, "step": 14230 }, { "epoch": 0.5965438101925112, "grad_norm": 2.618314743041992, "learning_rate": 3.6945971473431086e-06, "loss": 0.4796, "step": 14231 }, { "epoch": 0.5965857288075203, "grad_norm": 2.3251125812530518, "learning_rate": 3.6939418517064727e-06, "loss": 0.4807, "step": 14232 }, { "epoch": 0.5966276474225292, "grad_norm": 1.8465490341186523, "learning_rate": 3.6932865801452165e-06, "loss": 0.4787, "step": 14233 }, { "epoch": 0.5966695660375381, "grad_norm": 1.823809027671814, "learning_rate": 3.6926313326714135e-06, "loss": 0.4819, "step": 14234 }, { "epoch": 0.5967114846525471, "grad_norm": 2.173557758331299, "learning_rate": 3.6919761092971464e-06, "loss": 0.5394, "step": 14235 }, { "epoch": 0.596753403267556, "grad_norm": 1.7016165256500244, "learning_rate": 3.6913209100344903e-06, "loss": 0.4867, "step": 14236 }, { "epoch": 0.596795321882565, "grad_norm": 2.314819097518921, "learning_rate": 3.690665734895523e-06, "loss": 0.5082, "step": 14237 }, { "epoch": 0.5968372404975739, "grad_norm": 1.6489988565444946, "learning_rate": 3.690010583892325e-06, "loss": 0.4701, "step": 14238 }, { "epoch": 0.5968791591125829, "grad_norm": 1.7876417636871338, "learning_rate": 3.6893554570369695e-06, "loss": 0.5296, "step": 14239 }, { "epoch": 0.5969210777275918, "grad_norm": 1.8356757164001465, "learning_rate": 3.6887003543415344e-06, "loss": 0.481, "step": 14240 }, { "epoch": 0.5969629963426009, "grad_norm": 1.7233740091323853, "learning_rate": 3.6880452758180963e-06, "loss": 0.5151, "step": 14241 }, { "epoch": 0.5970049149576098, "grad_norm": 1.812569499015808, "learning_rate": 3.6873902214787282e-06, "loss": 0.4894, "step": 14242 }, { "epoch": 0.5970468335726188, "grad_norm": 1.6659175157546997, "learning_rate": 3.6867351913355064e-06, "loss": 0.4854, "step": 14243 }, { "epoch": 0.5970887521876277, "grad_norm": 1.7617188692092896, "learning_rate": 3.686080185400508e-06, "loss": 0.5385, "step": 14244 }, { "epoch": 0.5971306708026367, "grad_norm": 3.87324857711792, "learning_rate": 3.685425203685804e-06, "loss": 0.5044, "step": 14245 }, { "epoch": 0.5971725894176456, "grad_norm": 2.0859670639038086, "learning_rate": 3.6847702462034685e-06, "loss": 0.4748, "step": 14246 }, { "epoch": 0.5972145080326546, "grad_norm": 2.4174411296844482, "learning_rate": 3.6841153129655766e-06, "loss": 0.5004, "step": 14247 }, { "epoch": 0.5972564266476635, "grad_norm": 1.8020812273025513, "learning_rate": 3.683460403984198e-06, "loss": 0.5053, "step": 14248 }, { "epoch": 0.5972983452626726, "grad_norm": 1.4910720586776733, "learning_rate": 3.6828055192714075e-06, "loss": 0.473, "step": 14249 }, { "epoch": 0.5973402638776815, "grad_norm": 1.5759896039962769, "learning_rate": 3.6821506588392775e-06, "loss": 0.4801, "step": 14250 }, { "epoch": 0.5973821824926905, "grad_norm": 2.2158005237579346, "learning_rate": 3.6814958226998763e-06, "loss": 0.5069, "step": 14251 }, { "epoch": 0.5974241011076994, "grad_norm": 3.53937029838562, "learning_rate": 3.680841010865277e-06, "loss": 0.4608, "step": 14252 }, { "epoch": 0.5974660197227084, "grad_norm": 1.9425287246704102, "learning_rate": 3.6801862233475516e-06, "loss": 0.5488, "step": 14253 }, { "epoch": 0.5975079383377173, "grad_norm": 2.1063857078552246, "learning_rate": 3.6795314601587673e-06, "loss": 0.5254, "step": 14254 }, { "epoch": 0.5975498569527263, "grad_norm": 1.5608183145523071, "learning_rate": 3.6788767213109943e-06, "loss": 0.4456, "step": 14255 }, { "epoch": 0.5975917755677352, "grad_norm": 1.9216538667678833, "learning_rate": 3.6782220068163053e-06, "loss": 0.4684, "step": 14256 }, { "epoch": 0.5976336941827443, "grad_norm": 1.8752151727676392, "learning_rate": 3.677567316686763e-06, "loss": 0.4692, "step": 14257 }, { "epoch": 0.5976756127977532, "grad_norm": 1.4108262062072754, "learning_rate": 3.676912650934441e-06, "loss": 0.5022, "step": 14258 }, { "epoch": 0.5977175314127621, "grad_norm": 2.0115723609924316, "learning_rate": 3.676258009571406e-06, "loss": 0.5059, "step": 14259 }, { "epoch": 0.5977594500277711, "grad_norm": 2.0303993225097656, "learning_rate": 3.6756033926097223e-06, "loss": 0.4832, "step": 14260 }, { "epoch": 0.59780136864278, "grad_norm": 1.76766037940979, "learning_rate": 3.6749488000614596e-06, "loss": 0.5487, "step": 14261 }, { "epoch": 0.597843287257789, "grad_norm": 1.6812009811401367, "learning_rate": 3.6742942319386854e-06, "loss": 0.4628, "step": 14262 }, { "epoch": 0.5978852058727979, "grad_norm": 1.4669162034988403, "learning_rate": 3.6736396882534643e-06, "loss": 0.4746, "step": 14263 }, { "epoch": 0.5979271244878069, "grad_norm": 1.7334120273590088, "learning_rate": 3.67298516901786e-06, "loss": 0.4765, "step": 14264 }, { "epoch": 0.5979690431028158, "grad_norm": 1.8818751573562622, "learning_rate": 3.6723306742439413e-06, "loss": 0.5066, "step": 14265 }, { "epoch": 0.5980109617178249, "grad_norm": 1.7501213550567627, "learning_rate": 3.6716762039437693e-06, "loss": 0.5387, "step": 14266 }, { "epoch": 0.5980528803328338, "grad_norm": 1.7926899194717407, "learning_rate": 3.671021758129412e-06, "loss": 0.491, "step": 14267 }, { "epoch": 0.5980947989478428, "grad_norm": 1.610935926437378, "learning_rate": 3.6703673368129307e-06, "loss": 0.5226, "step": 14268 }, { "epoch": 0.5981367175628517, "grad_norm": 1.6215403079986572, "learning_rate": 3.6697129400063893e-06, "loss": 0.4904, "step": 14269 }, { "epoch": 0.5981786361778607, "grad_norm": 1.5122007131576538, "learning_rate": 3.66905856772185e-06, "loss": 0.4932, "step": 14270 }, { "epoch": 0.5982205547928696, "grad_norm": 1.6592326164245605, "learning_rate": 3.6684042199713774e-06, "loss": 0.5052, "step": 14271 }, { "epoch": 0.5982624734078786, "grad_norm": 1.67000412940979, "learning_rate": 3.667749896767031e-06, "loss": 0.4218, "step": 14272 }, { "epoch": 0.5983043920228875, "grad_norm": 1.8035271167755127, "learning_rate": 3.6670955981208745e-06, "loss": 0.5069, "step": 14273 }, { "epoch": 0.5983463106378966, "grad_norm": 2.0797200202941895, "learning_rate": 3.6664413240449683e-06, "loss": 0.4876, "step": 14274 }, { "epoch": 0.5983882292529055, "grad_norm": 1.5540581941604614, "learning_rate": 3.6657870745513715e-06, "loss": 0.5006, "step": 14275 }, { "epoch": 0.5984301478679145, "grad_norm": 1.7664413452148438, "learning_rate": 3.665132849652144e-06, "loss": 0.5274, "step": 14276 }, { "epoch": 0.5984720664829234, "grad_norm": 1.8935339450836182, "learning_rate": 3.664478649359351e-06, "loss": 0.4897, "step": 14277 }, { "epoch": 0.5985139850979324, "grad_norm": 2.146028518676758, "learning_rate": 3.6638244736850452e-06, "loss": 0.4758, "step": 14278 }, { "epoch": 0.5985559037129413, "grad_norm": 2.2577691078186035, "learning_rate": 3.663170322641288e-06, "loss": 0.5048, "step": 14279 }, { "epoch": 0.5985978223279503, "grad_norm": 2.106360673904419, "learning_rate": 3.662516196240139e-06, "loss": 0.4902, "step": 14280 }, { "epoch": 0.5986397409429592, "grad_norm": 1.8328438997268677, "learning_rate": 3.6618620944936533e-06, "loss": 0.5084, "step": 14281 }, { "epoch": 0.5986816595579683, "grad_norm": 1.6891528367996216, "learning_rate": 3.6612080174138916e-06, "loss": 0.5218, "step": 14282 }, { "epoch": 0.5987235781729772, "grad_norm": 1.62271249294281, "learning_rate": 3.6605539650129096e-06, "loss": 0.5044, "step": 14283 }, { "epoch": 0.5987654967879861, "grad_norm": 1.7358403205871582, "learning_rate": 3.659899937302761e-06, "loss": 0.5093, "step": 14284 }, { "epoch": 0.5988074154029951, "grad_norm": 2.0580599308013916, "learning_rate": 3.6592459342955057e-06, "loss": 0.4978, "step": 14285 }, { "epoch": 0.598849334018004, "grad_norm": 1.821413278579712, "learning_rate": 3.658591956003199e-06, "loss": 0.4972, "step": 14286 }, { "epoch": 0.598891252633013, "grad_norm": 1.7002670764923096, "learning_rate": 3.6579380024378946e-06, "loss": 0.475, "step": 14287 }, { "epoch": 0.5989331712480219, "grad_norm": 2.0396862030029297, "learning_rate": 3.6572840736116466e-06, "loss": 0.5113, "step": 14288 }, { "epoch": 0.5989750898630309, "grad_norm": 1.9877961874008179, "learning_rate": 3.6566301695365127e-06, "loss": 0.5225, "step": 14289 }, { "epoch": 0.5990170084780398, "grad_norm": 1.6610641479492188, "learning_rate": 3.655976290224542e-06, "loss": 0.5138, "step": 14290 }, { "epoch": 0.5990589270930489, "grad_norm": 1.6949518918991089, "learning_rate": 3.655322435687792e-06, "loss": 0.513, "step": 14291 }, { "epoch": 0.5991008457080578, "grad_norm": 1.6332305669784546, "learning_rate": 3.6546686059383146e-06, "loss": 0.4456, "step": 14292 }, { "epoch": 0.5991427643230668, "grad_norm": 1.7655282020568848, "learning_rate": 3.654014800988159e-06, "loss": 0.5169, "step": 14293 }, { "epoch": 0.5991846829380757, "grad_norm": 1.7943828105926514, "learning_rate": 3.653361020849381e-06, "loss": 0.4764, "step": 14294 }, { "epoch": 0.5992266015530847, "grad_norm": 1.7216087579727173, "learning_rate": 3.6527072655340324e-06, "loss": 0.4965, "step": 14295 }, { "epoch": 0.5992685201680936, "grad_norm": 2.4500370025634766, "learning_rate": 3.652053535054162e-06, "loss": 0.4816, "step": 14296 }, { "epoch": 0.5993104387831026, "grad_norm": 1.7460460662841797, "learning_rate": 3.6513998294218195e-06, "loss": 0.5013, "step": 14297 }, { "epoch": 0.5993523573981115, "grad_norm": 2.3488047122955322, "learning_rate": 3.650746148649059e-06, "loss": 0.5039, "step": 14298 }, { "epoch": 0.5993942760131206, "grad_norm": 1.8644804954528809, "learning_rate": 3.6500924927479266e-06, "loss": 0.533, "step": 14299 }, { "epoch": 0.5994361946281295, "grad_norm": 1.6883331537246704, "learning_rate": 3.649438861730473e-06, "loss": 0.4642, "step": 14300 }, { "epoch": 0.5994781132431385, "grad_norm": 1.6485791206359863, "learning_rate": 3.6487852556087483e-06, "loss": 0.4814, "step": 14301 }, { "epoch": 0.5995200318581474, "grad_norm": 1.8888176679611206, "learning_rate": 3.648131674394797e-06, "loss": 0.5209, "step": 14302 }, { "epoch": 0.5995619504731564, "grad_norm": 1.7604613304138184, "learning_rate": 3.64747811810067e-06, "loss": 0.5145, "step": 14303 }, { "epoch": 0.5996038690881653, "grad_norm": 2.133099317550659, "learning_rate": 3.6468245867384162e-06, "loss": 0.5321, "step": 14304 }, { "epoch": 0.5996457877031743, "grad_norm": 1.7730011940002441, "learning_rate": 3.646171080320079e-06, "loss": 0.483, "step": 14305 }, { "epoch": 0.5996877063181832, "grad_norm": 1.6780503988265991, "learning_rate": 3.6455175988577053e-06, "loss": 0.4614, "step": 14306 }, { "epoch": 0.5997296249331923, "grad_norm": 1.9981931447982788, "learning_rate": 3.6448641423633447e-06, "loss": 0.5496, "step": 14307 }, { "epoch": 0.5997715435482012, "grad_norm": 1.8418145179748535, "learning_rate": 3.6442107108490377e-06, "loss": 0.5281, "step": 14308 }, { "epoch": 0.5998134621632101, "grad_norm": 5.298699855804443, "learning_rate": 3.6435573043268334e-06, "loss": 0.505, "step": 14309 }, { "epoch": 0.5998553807782191, "grad_norm": 1.8253144025802612, "learning_rate": 3.6429039228087737e-06, "loss": 0.5061, "step": 14310 }, { "epoch": 0.599897299393228, "grad_norm": 1.5562177896499634, "learning_rate": 3.6422505663069053e-06, "loss": 0.48, "step": 14311 }, { "epoch": 0.599939218008237, "grad_norm": 2.9748830795288086, "learning_rate": 3.641597234833271e-06, "loss": 0.4863, "step": 14312 }, { "epoch": 0.5999811366232459, "grad_norm": 1.7930231094360352, "learning_rate": 3.6409439283999125e-06, "loss": 0.4953, "step": 14313 }, { "epoch": 0.6000230552382549, "grad_norm": 1.500656008720398, "learning_rate": 3.640290647018874e-06, "loss": 0.4863, "step": 14314 }, { "epoch": 0.6000649738532638, "grad_norm": 1.767349362373352, "learning_rate": 3.639637390702198e-06, "loss": 0.5404, "step": 14315 }, { "epoch": 0.6001068924682729, "grad_norm": 1.8870799541473389, "learning_rate": 3.6389841594619265e-06, "loss": 0.4697, "step": 14316 }, { "epoch": 0.6001488110832818, "grad_norm": 1.9286236763000488, "learning_rate": 3.6383309533100987e-06, "loss": 0.4321, "step": 14317 }, { "epoch": 0.6001907296982908, "grad_norm": 2.3959624767303467, "learning_rate": 3.63767777225876e-06, "loss": 0.4372, "step": 14318 }, { "epoch": 0.6002326483132997, "grad_norm": 1.9012380838394165, "learning_rate": 3.6370246163199464e-06, "loss": 0.523, "step": 14319 }, { "epoch": 0.6002745669283087, "grad_norm": 2.9148521423339844, "learning_rate": 3.6363714855056997e-06, "loss": 0.484, "step": 14320 }, { "epoch": 0.6003164855433176, "grad_norm": 1.884635090827942, "learning_rate": 3.6357183798280616e-06, "loss": 0.4351, "step": 14321 }, { "epoch": 0.6003584041583266, "grad_norm": 2.743957042694092, "learning_rate": 3.6350652992990663e-06, "loss": 0.4567, "step": 14322 }, { "epoch": 0.6004003227733355, "grad_norm": 1.7033541202545166, "learning_rate": 3.6344122439307557e-06, "loss": 0.4617, "step": 14323 }, { "epoch": 0.6004422413883446, "grad_norm": 2.6017396450042725, "learning_rate": 3.633759213735169e-06, "loss": 0.4821, "step": 14324 }, { "epoch": 0.6004841600033535, "grad_norm": 1.7141183614730835, "learning_rate": 3.633106208724342e-06, "loss": 0.4714, "step": 14325 }, { "epoch": 0.6005260786183625, "grad_norm": 1.6630301475524902, "learning_rate": 3.632453228910311e-06, "loss": 0.4856, "step": 14326 }, { "epoch": 0.6005679972333714, "grad_norm": 2.6935782432556152, "learning_rate": 3.631800274305116e-06, "loss": 0.499, "step": 14327 }, { "epoch": 0.6006099158483804, "grad_norm": 1.6927692890167236, "learning_rate": 3.63114734492079e-06, "loss": 0.47, "step": 14328 }, { "epoch": 0.6006518344633893, "grad_norm": 2.4207210540771484, "learning_rate": 3.630494440769371e-06, "loss": 0.4921, "step": 14329 }, { "epoch": 0.6006937530783983, "grad_norm": 1.838593602180481, "learning_rate": 3.6298415618628948e-06, "loss": 0.4661, "step": 14330 }, { "epoch": 0.6007356716934072, "grad_norm": 2.6528806686401367, "learning_rate": 3.6291887082133935e-06, "loss": 0.4773, "step": 14331 }, { "epoch": 0.6007775903084163, "grad_norm": 1.5435199737548828, "learning_rate": 3.628535879832903e-06, "loss": 0.507, "step": 14332 }, { "epoch": 0.6008195089234252, "grad_norm": 1.649591326713562, "learning_rate": 3.6278830767334594e-06, "loss": 0.5211, "step": 14333 }, { "epoch": 0.6008614275384341, "grad_norm": 1.7059502601623535, "learning_rate": 3.6272302989270934e-06, "loss": 0.5238, "step": 14334 }, { "epoch": 0.6009033461534431, "grad_norm": 1.8410124778747559, "learning_rate": 3.626577546425838e-06, "loss": 0.4322, "step": 14335 }, { "epoch": 0.600945264768452, "grad_norm": 1.7322503328323364, "learning_rate": 3.6259248192417295e-06, "loss": 0.5161, "step": 14336 }, { "epoch": 0.600987183383461, "grad_norm": 1.9193215370178223, "learning_rate": 3.6252721173867944e-06, "loss": 0.5183, "step": 14337 }, { "epoch": 0.6010291019984699, "grad_norm": 1.8145725727081299, "learning_rate": 3.624619440873069e-06, "loss": 0.4584, "step": 14338 }, { "epoch": 0.601071020613479, "grad_norm": 1.931533694267273, "learning_rate": 3.6239667897125835e-06, "loss": 0.508, "step": 14339 }, { "epoch": 0.6011129392284879, "grad_norm": 1.8496068716049194, "learning_rate": 3.6233141639173662e-06, "loss": 0.4868, "step": 14340 }, { "epoch": 0.6011548578434969, "grad_norm": 2.1878576278686523, "learning_rate": 3.622661563499449e-06, "loss": 0.5513, "step": 14341 }, { "epoch": 0.6011967764585058, "grad_norm": 1.8475764989852905, "learning_rate": 3.6220089884708633e-06, "loss": 0.5202, "step": 14342 }, { "epoch": 0.6012386950735148, "grad_norm": 1.522956132888794, "learning_rate": 3.6213564388436363e-06, "loss": 0.4258, "step": 14343 }, { "epoch": 0.6012806136885237, "grad_norm": 1.6012989282608032, "learning_rate": 3.620703914629796e-06, "loss": 0.4888, "step": 14344 }, { "epoch": 0.6013225323035327, "grad_norm": 2.2919089794158936, "learning_rate": 3.6200514158413747e-06, "loss": 0.4839, "step": 14345 }, { "epoch": 0.6013644509185416, "grad_norm": 1.8720513582229614, "learning_rate": 3.6193989424903954e-06, "loss": 0.5213, "step": 14346 }, { "epoch": 0.6014063695335506, "grad_norm": 1.7451974153518677, "learning_rate": 3.6187464945888896e-06, "loss": 0.4852, "step": 14347 }, { "epoch": 0.6014482881485596, "grad_norm": 1.9611237049102783, "learning_rate": 3.6180940721488832e-06, "loss": 0.4442, "step": 14348 }, { "epoch": 0.6014902067635686, "grad_norm": 1.568954586982727, "learning_rate": 3.6174416751824005e-06, "loss": 0.4923, "step": 14349 }, { "epoch": 0.6015321253785775, "grad_norm": 1.7306524515151978, "learning_rate": 3.616789303701469e-06, "loss": 0.5388, "step": 14350 }, { "epoch": 0.6015740439935865, "grad_norm": 1.8192527294158936, "learning_rate": 3.616136957718116e-06, "loss": 0.5523, "step": 14351 }, { "epoch": 0.6016159626085954, "grad_norm": 1.901645302772522, "learning_rate": 3.6154846372443644e-06, "loss": 0.456, "step": 14352 }, { "epoch": 0.6016578812236044, "grad_norm": 1.6461210250854492, "learning_rate": 3.614832342292238e-06, "loss": 0.4517, "step": 14353 }, { "epoch": 0.6016997998386133, "grad_norm": 2.159125328063965, "learning_rate": 3.614180072873765e-06, "loss": 0.4646, "step": 14354 }, { "epoch": 0.6017417184536223, "grad_norm": 2.0923824310302734, "learning_rate": 3.613527829000965e-06, "loss": 0.5044, "step": 14355 }, { "epoch": 0.6017836370686312, "grad_norm": 1.7640759944915771, "learning_rate": 3.6128756106858633e-06, "loss": 0.4733, "step": 14356 }, { "epoch": 0.6018255556836403, "grad_norm": 2.7705183029174805, "learning_rate": 3.612223417940482e-06, "loss": 0.4829, "step": 14357 }, { "epoch": 0.6018674742986492, "grad_norm": 1.6936031579971313, "learning_rate": 3.611571250776843e-06, "loss": 0.4642, "step": 14358 }, { "epoch": 0.6019093929136581, "grad_norm": 1.7454628944396973, "learning_rate": 3.6109191092069685e-06, "loss": 0.4609, "step": 14359 }, { "epoch": 0.6019513115286671, "grad_norm": 2.278646230697632, "learning_rate": 3.610266993242881e-06, "loss": 0.4786, "step": 14360 }, { "epoch": 0.601993230143676, "grad_norm": 1.7794859409332275, "learning_rate": 3.609614902896599e-06, "loss": 0.5137, "step": 14361 }, { "epoch": 0.602035148758685, "grad_norm": 1.930009365081787, "learning_rate": 3.6089628381801446e-06, "loss": 0.5128, "step": 14362 }, { "epoch": 0.6020770673736939, "grad_norm": 2.2446303367614746, "learning_rate": 3.6083107991055383e-06, "loss": 0.4701, "step": 14363 }, { "epoch": 0.602118985988703, "grad_norm": 1.6616179943084717, "learning_rate": 3.6076587856847966e-06, "loss": 0.4381, "step": 14364 }, { "epoch": 0.6021609046037119, "grad_norm": 1.8292202949523926, "learning_rate": 3.60700679792994e-06, "loss": 0.4723, "step": 14365 }, { "epoch": 0.6022028232187209, "grad_norm": 1.9139189720153809, "learning_rate": 3.6063548358529898e-06, "loss": 0.4711, "step": 14366 }, { "epoch": 0.6022447418337298, "grad_norm": 1.8416768312454224, "learning_rate": 3.6057028994659605e-06, "loss": 0.467, "step": 14367 }, { "epoch": 0.6022866604487388, "grad_norm": 1.7412621974945068, "learning_rate": 3.605050988780869e-06, "loss": 0.4867, "step": 14368 }, { "epoch": 0.6023285790637477, "grad_norm": 1.9683884382247925, "learning_rate": 3.6043991038097364e-06, "loss": 0.5309, "step": 14369 }, { "epoch": 0.6023704976787567, "grad_norm": 1.8663585186004639, "learning_rate": 3.6037472445645747e-06, "loss": 0.5042, "step": 14370 }, { "epoch": 0.6024124162937656, "grad_norm": 2.6771798133850098, "learning_rate": 3.6030954110574045e-06, "loss": 0.4799, "step": 14371 }, { "epoch": 0.6024543349087746, "grad_norm": 1.8074930906295776, "learning_rate": 3.602443603300239e-06, "loss": 0.4706, "step": 14372 }, { "epoch": 0.6024962535237836, "grad_norm": 1.7057775259017944, "learning_rate": 3.601791821305092e-06, "loss": 0.4897, "step": 14373 }, { "epoch": 0.6025381721387926, "grad_norm": 1.8806772232055664, "learning_rate": 3.60114006508398e-06, "loss": 0.475, "step": 14374 }, { "epoch": 0.6025800907538015, "grad_norm": 1.6713310480117798, "learning_rate": 3.600488334648919e-06, "loss": 0.4712, "step": 14375 }, { "epoch": 0.6026220093688105, "grad_norm": 2.686950445175171, "learning_rate": 3.5998366300119194e-06, "loss": 0.5546, "step": 14376 }, { "epoch": 0.6026639279838194, "grad_norm": 2.7175188064575195, "learning_rate": 3.5991849511849953e-06, "loss": 0.5443, "step": 14377 }, { "epoch": 0.6027058465988284, "grad_norm": 1.6305011510849, "learning_rate": 3.598533298180162e-06, "loss": 0.4681, "step": 14378 }, { "epoch": 0.6027477652138373, "grad_norm": 2.506804943084717, "learning_rate": 3.597881671009428e-06, "loss": 0.5503, "step": 14379 }, { "epoch": 0.6027896838288463, "grad_norm": 1.71471107006073, "learning_rate": 3.5972300696848073e-06, "loss": 0.4649, "step": 14380 }, { "epoch": 0.6028316024438553, "grad_norm": 1.7305891513824463, "learning_rate": 3.5965784942183123e-06, "loss": 0.4962, "step": 14381 }, { "epoch": 0.6028735210588643, "grad_norm": 2.09332013130188, "learning_rate": 3.5959269446219514e-06, "loss": 0.5266, "step": 14382 }, { "epoch": 0.6029154396738732, "grad_norm": 1.4275729656219482, "learning_rate": 3.595275420907737e-06, "loss": 0.3956, "step": 14383 }, { "epoch": 0.6029573582888821, "grad_norm": 2.2441506385803223, "learning_rate": 3.5946239230876767e-06, "loss": 0.4875, "step": 14384 }, { "epoch": 0.6029992769038911, "grad_norm": 1.7396999597549438, "learning_rate": 3.5939724511737827e-06, "loss": 0.4652, "step": 14385 }, { "epoch": 0.6030411955189, "grad_norm": 1.8485788106918335, "learning_rate": 3.5933210051780644e-06, "loss": 0.5484, "step": 14386 }, { "epoch": 0.603083114133909, "grad_norm": 2.0722780227661133, "learning_rate": 3.5926695851125254e-06, "loss": 0.4499, "step": 14387 }, { "epoch": 0.6031250327489179, "grad_norm": 1.666833758354187, "learning_rate": 3.5920181909891776e-06, "loss": 0.518, "step": 14388 }, { "epoch": 0.603166951363927, "grad_norm": 2.038055658340454, "learning_rate": 3.5913668228200297e-06, "loss": 0.5032, "step": 14389 }, { "epoch": 0.6032088699789359, "grad_norm": 1.7712868452072144, "learning_rate": 3.5907154806170864e-06, "loss": 0.4895, "step": 14390 }, { "epoch": 0.6032507885939449, "grad_norm": 1.77834951877594, "learning_rate": 3.590064164392354e-06, "loss": 0.4982, "step": 14391 }, { "epoch": 0.6032927072089538, "grad_norm": 1.8054900169372559, "learning_rate": 3.5894128741578414e-06, "loss": 0.4803, "step": 14392 }, { "epoch": 0.6033346258239628, "grad_norm": 1.9674509763717651, "learning_rate": 3.5887616099255506e-06, "loss": 0.4956, "step": 14393 }, { "epoch": 0.6033765444389717, "grad_norm": 1.851741075515747, "learning_rate": 3.5881103717074894e-06, "loss": 0.519, "step": 14394 }, { "epoch": 0.6034184630539807, "grad_norm": 2.0580670833587646, "learning_rate": 3.587459159515663e-06, "loss": 0.4624, "step": 14395 }, { "epoch": 0.6034603816689896, "grad_norm": 2.0661680698394775, "learning_rate": 3.5868079733620707e-06, "loss": 0.497, "step": 14396 }, { "epoch": 0.6035023002839986, "grad_norm": 1.6914268732070923, "learning_rate": 3.586156813258721e-06, "loss": 0.4863, "step": 14397 }, { "epoch": 0.6035442188990076, "grad_norm": 1.7739408016204834, "learning_rate": 3.5855056792176178e-06, "loss": 0.5591, "step": 14398 }, { "epoch": 0.6035861375140166, "grad_norm": 1.852427363395691, "learning_rate": 3.584854571250761e-06, "loss": 0.522, "step": 14399 }, { "epoch": 0.6036280561290255, "grad_norm": 1.6883416175842285, "learning_rate": 3.5842034893701527e-06, "loss": 0.5433, "step": 14400 }, { "epoch": 0.6036699747440345, "grad_norm": 1.6001285314559937, "learning_rate": 3.583552433587798e-06, "loss": 0.4609, "step": 14401 }, { "epoch": 0.6037118933590434, "grad_norm": 2.4386751651763916, "learning_rate": 3.5829014039156936e-06, "loss": 0.4689, "step": 14402 }, { "epoch": 0.6037538119740524, "grad_norm": 1.9555963277816772, "learning_rate": 3.5822504003658433e-06, "loss": 0.4906, "step": 14403 }, { "epoch": 0.6037957305890613, "grad_norm": 2.15608811378479, "learning_rate": 3.5815994229502477e-06, "loss": 0.5477, "step": 14404 }, { "epoch": 0.6038376492040703, "grad_norm": 1.8033723831176758, "learning_rate": 3.5809484716809057e-06, "loss": 0.5247, "step": 14405 }, { "epoch": 0.6038795678190793, "grad_norm": 1.91927170753479, "learning_rate": 3.580297546569815e-06, "loss": 0.5012, "step": 14406 }, { "epoch": 0.6039214864340883, "grad_norm": 1.8958921432495117, "learning_rate": 3.5796466476289795e-06, "loss": 0.475, "step": 14407 }, { "epoch": 0.6039634050490972, "grad_norm": 1.6449476480484009, "learning_rate": 3.5789957748703917e-06, "loss": 0.4767, "step": 14408 }, { "epoch": 0.6040053236641061, "grad_norm": 1.882460117340088, "learning_rate": 3.5783449283060546e-06, "loss": 0.4897, "step": 14409 }, { "epoch": 0.6040472422791151, "grad_norm": 1.6505424976348877, "learning_rate": 3.5776941079479632e-06, "loss": 0.5163, "step": 14410 }, { "epoch": 0.604089160894124, "grad_norm": 1.7564767599105835, "learning_rate": 3.5770433138081127e-06, "loss": 0.5011, "step": 14411 }, { "epoch": 0.604131079509133, "grad_norm": 1.72688889503479, "learning_rate": 3.5763925458985018e-06, "loss": 0.5226, "step": 14412 }, { "epoch": 0.6041729981241419, "grad_norm": 1.5274162292480469, "learning_rate": 3.575741804231128e-06, "loss": 0.4783, "step": 14413 }, { "epoch": 0.604214916739151, "grad_norm": 2.0212178230285645, "learning_rate": 3.5750910888179845e-06, "loss": 0.5238, "step": 14414 }, { "epoch": 0.6042568353541599, "grad_norm": 2.0574281215667725, "learning_rate": 3.5744403996710654e-06, "loss": 0.4528, "step": 14415 }, { "epoch": 0.6042987539691689, "grad_norm": 2.1473052501678467, "learning_rate": 3.573789736802369e-06, "loss": 0.512, "step": 14416 }, { "epoch": 0.6043406725841778, "grad_norm": 1.8933234214782715, "learning_rate": 3.5731391002238845e-06, "loss": 0.4802, "step": 14417 }, { "epoch": 0.6043825911991868, "grad_norm": 1.768310546875, "learning_rate": 3.57248848994761e-06, "loss": 0.5696, "step": 14418 }, { "epoch": 0.6044245098141957, "grad_norm": 1.9777159690856934, "learning_rate": 3.571837905985538e-06, "loss": 0.5244, "step": 14419 }, { "epoch": 0.6044664284292047, "grad_norm": 1.743851661682129, "learning_rate": 3.571187348349657e-06, "loss": 0.542, "step": 14420 }, { "epoch": 0.6045083470442136, "grad_norm": 1.6516309976577759, "learning_rate": 3.5705368170519616e-06, "loss": 0.488, "step": 14421 }, { "epoch": 0.6045502656592227, "grad_norm": 2.0239546298980713, "learning_rate": 3.569886312104447e-06, "loss": 0.4278, "step": 14422 }, { "epoch": 0.6045921842742316, "grad_norm": 1.9183579683303833, "learning_rate": 3.569235833519099e-06, "loss": 0.4892, "step": 14423 }, { "epoch": 0.6046341028892406, "grad_norm": 1.915952205657959, "learning_rate": 3.5685853813079103e-06, "loss": 0.5006, "step": 14424 }, { "epoch": 0.6046760215042495, "grad_norm": 1.7768386602401733, "learning_rate": 3.5679349554828728e-06, "loss": 0.4411, "step": 14425 }, { "epoch": 0.6047179401192585, "grad_norm": 2.341740846633911, "learning_rate": 3.5672845560559732e-06, "loss": 0.4267, "step": 14426 }, { "epoch": 0.6047598587342674, "grad_norm": 2.0071277618408203, "learning_rate": 3.5666341830392027e-06, "loss": 0.4766, "step": 14427 }, { "epoch": 0.6048017773492764, "grad_norm": 1.4634425640106201, "learning_rate": 3.5659838364445505e-06, "loss": 0.4444, "step": 14428 }, { "epoch": 0.6048436959642853, "grad_norm": 1.7393368482589722, "learning_rate": 3.5653335162840013e-06, "loss": 0.533, "step": 14429 }, { "epoch": 0.6048856145792944, "grad_norm": 2.04732346534729, "learning_rate": 3.564683222569546e-06, "loss": 0.513, "step": 14430 }, { "epoch": 0.6049275331943033, "grad_norm": 1.496978998184204, "learning_rate": 3.564032955313173e-06, "loss": 0.4859, "step": 14431 }, { "epoch": 0.6049694518093123, "grad_norm": 1.8836902379989624, "learning_rate": 3.563382714526866e-06, "loss": 0.574, "step": 14432 }, { "epoch": 0.6050113704243212, "grad_norm": 2.094254732131958, "learning_rate": 3.562732500222612e-06, "loss": 0.5267, "step": 14433 }, { "epoch": 0.6050532890393301, "grad_norm": 2.3267016410827637, "learning_rate": 3.562082312412399e-06, "loss": 0.4999, "step": 14434 }, { "epoch": 0.6050952076543391, "grad_norm": 1.459376335144043, "learning_rate": 3.5614321511082095e-06, "loss": 0.4525, "step": 14435 }, { "epoch": 0.605137126269348, "grad_norm": 1.4594193696975708, "learning_rate": 3.5607820163220303e-06, "loss": 0.4657, "step": 14436 }, { "epoch": 0.605179044884357, "grad_norm": 1.5780521631240845, "learning_rate": 3.560131908065846e-06, "loss": 0.5206, "step": 14437 }, { "epoch": 0.6052209634993659, "grad_norm": 1.7629433870315552, "learning_rate": 3.559481826351637e-06, "loss": 0.4535, "step": 14438 }, { "epoch": 0.605262882114375, "grad_norm": 1.684579849243164, "learning_rate": 3.558831771191389e-06, "loss": 0.56, "step": 14439 }, { "epoch": 0.6053048007293839, "grad_norm": 1.9227551221847534, "learning_rate": 3.5581817425970876e-06, "loss": 0.5095, "step": 14440 }, { "epoch": 0.6053467193443929, "grad_norm": 2.6497747898101807, "learning_rate": 3.5575317405807107e-06, "loss": 0.4992, "step": 14441 }, { "epoch": 0.6053886379594018, "grad_norm": 1.9035536050796509, "learning_rate": 3.5568817651542413e-06, "loss": 0.5024, "step": 14442 }, { "epoch": 0.6054305565744108, "grad_norm": 1.7939740419387817, "learning_rate": 3.5562318163296637e-06, "loss": 0.5012, "step": 14443 }, { "epoch": 0.6054724751894197, "grad_norm": 1.7229773998260498, "learning_rate": 3.5555818941189544e-06, "loss": 0.49, "step": 14444 }, { "epoch": 0.6055143938044287, "grad_norm": 1.9160560369491577, "learning_rate": 3.5549319985340973e-06, "loss": 0.4849, "step": 14445 }, { "epoch": 0.6055563124194376, "grad_norm": 2.1715359687805176, "learning_rate": 3.554282129587072e-06, "loss": 0.475, "step": 14446 }, { "epoch": 0.6055982310344467, "grad_norm": 2.0890049934387207, "learning_rate": 3.5536322872898556e-06, "loss": 0.4696, "step": 14447 }, { "epoch": 0.6056401496494556, "grad_norm": 1.6663901805877686, "learning_rate": 3.552982471654427e-06, "loss": 0.4447, "step": 14448 }, { "epoch": 0.6056820682644646, "grad_norm": 2.310224771499634, "learning_rate": 3.5523326826927696e-06, "loss": 0.4271, "step": 14449 }, { "epoch": 0.6057239868794735, "grad_norm": 2.0595598220825195, "learning_rate": 3.5516829204168555e-06, "loss": 0.465, "step": 14450 }, { "epoch": 0.6057659054944825, "grad_norm": 2.3224635124206543, "learning_rate": 3.5510331848386643e-06, "loss": 0.5007, "step": 14451 }, { "epoch": 0.6058078241094914, "grad_norm": 1.925246238708496, "learning_rate": 3.550383475970175e-06, "loss": 0.4691, "step": 14452 }, { "epoch": 0.6058497427245004, "grad_norm": 1.6936001777648926, "learning_rate": 3.5497337938233602e-06, "loss": 0.5094, "step": 14453 }, { "epoch": 0.6058916613395093, "grad_norm": 1.3780418634414673, "learning_rate": 3.5490841384101993e-06, "loss": 0.4582, "step": 14454 }, { "epoch": 0.6059335799545184, "grad_norm": 3.4854254722595215, "learning_rate": 3.5484345097426665e-06, "loss": 0.5515, "step": 14455 }, { "epoch": 0.6059754985695273, "grad_norm": 1.6596349477767944, "learning_rate": 3.5477849078327364e-06, "loss": 0.5232, "step": 14456 }, { "epoch": 0.6060174171845363, "grad_norm": 1.5321485996246338, "learning_rate": 3.547135332692383e-06, "loss": 0.4456, "step": 14457 }, { "epoch": 0.6060593357995452, "grad_norm": 1.8047997951507568, "learning_rate": 3.546485784333583e-06, "loss": 0.5663, "step": 14458 }, { "epoch": 0.6061012544145541, "grad_norm": 1.6030678749084473, "learning_rate": 3.5458362627683068e-06, "loss": 0.5143, "step": 14459 }, { "epoch": 0.6061431730295631, "grad_norm": 1.9480987787246704, "learning_rate": 3.5451867680085307e-06, "loss": 0.4943, "step": 14460 }, { "epoch": 0.606185091644572, "grad_norm": 1.9249380826950073, "learning_rate": 3.5445373000662235e-06, "loss": 0.4887, "step": 14461 }, { "epoch": 0.606227010259581, "grad_norm": 1.8763718605041504, "learning_rate": 3.5438878589533587e-06, "loss": 0.4768, "step": 14462 }, { "epoch": 0.6062689288745899, "grad_norm": 1.7762444019317627, "learning_rate": 3.54323844468191e-06, "loss": 0.5113, "step": 14463 }, { "epoch": 0.606310847489599, "grad_norm": 2.6328988075256348, "learning_rate": 3.542589057263845e-06, "loss": 0.4717, "step": 14464 }, { "epoch": 0.6063527661046079, "grad_norm": 1.6766020059585571, "learning_rate": 3.541939696711136e-06, "loss": 0.5082, "step": 14465 }, { "epoch": 0.6063946847196169, "grad_norm": 1.8182764053344727, "learning_rate": 3.5412903630357556e-06, "loss": 0.4807, "step": 14466 }, { "epoch": 0.6064366033346258, "grad_norm": 2.131040334701538, "learning_rate": 3.5406410562496675e-06, "loss": 0.4975, "step": 14467 }, { "epoch": 0.6064785219496348, "grad_norm": 1.8758234977722168, "learning_rate": 3.539991776364844e-06, "loss": 0.5075, "step": 14468 }, { "epoch": 0.6065204405646437, "grad_norm": 2.439650058746338, "learning_rate": 3.5393425233932566e-06, "loss": 0.5302, "step": 14469 }, { "epoch": 0.6065623591796527, "grad_norm": 3.1094515323638916, "learning_rate": 3.5386932973468687e-06, "loss": 0.4486, "step": 14470 }, { "epoch": 0.6066042777946616, "grad_norm": 1.995247483253479, "learning_rate": 3.5380440982376486e-06, "loss": 0.4848, "step": 14471 }, { "epoch": 0.6066461964096707, "grad_norm": 1.8232783079147339, "learning_rate": 3.537394926077567e-06, "loss": 0.4201, "step": 14472 }, { "epoch": 0.6066881150246796, "grad_norm": 2.6295313835144043, "learning_rate": 3.5367457808785853e-06, "loss": 0.5056, "step": 14473 }, { "epoch": 0.6067300336396886, "grad_norm": 1.7411649227142334, "learning_rate": 3.536096662652674e-06, "loss": 0.4443, "step": 14474 }, { "epoch": 0.6067719522546975, "grad_norm": 1.8842014074325562, "learning_rate": 3.5354475714117977e-06, "loss": 0.4404, "step": 14475 }, { "epoch": 0.6068138708697065, "grad_norm": 1.920903205871582, "learning_rate": 3.534798507167918e-06, "loss": 0.4952, "step": 14476 }, { "epoch": 0.6068557894847154, "grad_norm": 1.8736881017684937, "learning_rate": 3.534149469933003e-06, "loss": 0.5462, "step": 14477 }, { "epoch": 0.6068977080997244, "grad_norm": 1.9433133602142334, "learning_rate": 3.533500459719018e-06, "loss": 0.4637, "step": 14478 }, { "epoch": 0.6069396267147333, "grad_norm": 2.0999770164489746, "learning_rate": 3.5328514765379233e-06, "loss": 0.5515, "step": 14479 }, { "epoch": 0.6069815453297424, "grad_norm": 2.1331663131713867, "learning_rate": 3.532202520401683e-06, "loss": 0.5181, "step": 14480 }, { "epoch": 0.6070234639447513, "grad_norm": 1.691792607307434, "learning_rate": 3.5315535913222614e-06, "loss": 0.5443, "step": 14481 }, { "epoch": 0.6070653825597603, "grad_norm": 1.7473714351654053, "learning_rate": 3.5309046893116174e-06, "loss": 0.4976, "step": 14482 }, { "epoch": 0.6071073011747692, "grad_norm": 1.7599425315856934, "learning_rate": 3.5302558143817157e-06, "loss": 0.5196, "step": 14483 }, { "epoch": 0.6071492197897781, "grad_norm": 1.8876279592514038, "learning_rate": 3.529606966544518e-06, "loss": 0.5062, "step": 14484 }, { "epoch": 0.6071911384047871, "grad_norm": 1.5219398736953735, "learning_rate": 3.5289581458119795e-06, "loss": 0.4436, "step": 14485 }, { "epoch": 0.607233057019796, "grad_norm": 1.995198369026184, "learning_rate": 3.528309352196064e-06, "loss": 0.4979, "step": 14486 }, { "epoch": 0.607274975634805, "grad_norm": 2.3286242485046387, "learning_rate": 3.5276605857087343e-06, "loss": 0.4346, "step": 14487 }, { "epoch": 0.6073168942498139, "grad_norm": 1.7223505973815918, "learning_rate": 3.5270118463619446e-06, "loss": 0.5122, "step": 14488 }, { "epoch": 0.607358812864823, "grad_norm": 1.9361166954040527, "learning_rate": 3.526363134167654e-06, "loss": 0.4481, "step": 14489 }, { "epoch": 0.6074007314798319, "grad_norm": 2.0493061542510986, "learning_rate": 3.5257144491378243e-06, "loss": 0.5109, "step": 14490 }, { "epoch": 0.6074426500948409, "grad_norm": 1.9248998165130615, "learning_rate": 3.5250657912844083e-06, "loss": 0.4422, "step": 14491 }, { "epoch": 0.6074845687098498, "grad_norm": 2.131120443344116, "learning_rate": 3.5244171606193667e-06, "loss": 0.4788, "step": 14492 }, { "epoch": 0.6075264873248588, "grad_norm": 1.7855864763259888, "learning_rate": 3.5237685571546553e-06, "loss": 0.4967, "step": 14493 }, { "epoch": 0.6075684059398677, "grad_norm": 1.729468822479248, "learning_rate": 3.5231199809022276e-06, "loss": 0.5033, "step": 14494 }, { "epoch": 0.6076103245548767, "grad_norm": 1.9294500350952148, "learning_rate": 3.5224714318740414e-06, "loss": 0.4515, "step": 14495 }, { "epoch": 0.6076522431698856, "grad_norm": 1.9786802530288696, "learning_rate": 3.5218229100820544e-06, "loss": 0.5359, "step": 14496 }, { "epoch": 0.6076941617848947, "grad_norm": 1.8947511911392212, "learning_rate": 3.5211744155382173e-06, "loss": 0.4733, "step": 14497 }, { "epoch": 0.6077360803999036, "grad_norm": 1.8919895887374878, "learning_rate": 3.520525948254484e-06, "loss": 0.456, "step": 14498 }, { "epoch": 0.6077779990149126, "grad_norm": 1.9075983762741089, "learning_rate": 3.5198775082428114e-06, "loss": 0.4992, "step": 14499 }, { "epoch": 0.6078199176299215, "grad_norm": 2.5036168098449707, "learning_rate": 3.5192290955151497e-06, "loss": 0.5316, "step": 14500 }, { "epoch": 0.6078618362449305, "grad_norm": 2.1739022731781006, "learning_rate": 3.518580710083453e-06, "loss": 0.451, "step": 14501 }, { "epoch": 0.6079037548599394, "grad_norm": 2.1562445163726807, "learning_rate": 3.517932351959673e-06, "loss": 0.4651, "step": 14502 }, { "epoch": 0.6079456734749484, "grad_norm": 1.6652357578277588, "learning_rate": 3.5172840211557606e-06, "loss": 0.4549, "step": 14503 }, { "epoch": 0.6079875920899573, "grad_norm": 1.758846640586853, "learning_rate": 3.5166357176836676e-06, "loss": 0.5108, "step": 14504 }, { "epoch": 0.6080295107049664, "grad_norm": 1.8285531997680664, "learning_rate": 3.5159874415553452e-06, "loss": 0.5334, "step": 14505 }, { "epoch": 0.6080714293199753, "grad_norm": 1.8671859502792358, "learning_rate": 3.5153391927827414e-06, "loss": 0.4036, "step": 14506 }, { "epoch": 0.6081133479349843, "grad_norm": 1.7275571823120117, "learning_rate": 3.5146909713778076e-06, "loss": 0.4653, "step": 14507 }, { "epoch": 0.6081552665499932, "grad_norm": 1.8863648176193237, "learning_rate": 3.5140427773524942e-06, "loss": 0.496, "step": 14508 }, { "epoch": 0.6081971851650021, "grad_norm": 1.8633828163146973, "learning_rate": 3.5133946107187454e-06, "loss": 0.4789, "step": 14509 }, { "epoch": 0.6082391037800111, "grad_norm": 2.89823317527771, "learning_rate": 3.512746471488512e-06, "loss": 0.4905, "step": 14510 }, { "epoch": 0.60828102239502, "grad_norm": 1.8356084823608398, "learning_rate": 3.5120983596737434e-06, "loss": 0.5044, "step": 14511 }, { "epoch": 0.608322941010029, "grad_norm": 1.9293800592422485, "learning_rate": 3.5114502752863834e-06, "loss": 0.4412, "step": 14512 }, { "epoch": 0.608364859625038, "grad_norm": 1.8469116687774658, "learning_rate": 3.5108022183383784e-06, "loss": 0.4708, "step": 14513 }, { "epoch": 0.608406778240047, "grad_norm": 2.018829822540283, "learning_rate": 3.5101541888416777e-06, "loss": 0.4409, "step": 14514 }, { "epoch": 0.6084486968550559, "grad_norm": 1.9717388153076172, "learning_rate": 3.509506186808224e-06, "loss": 0.5285, "step": 14515 }, { "epoch": 0.6084906154700649, "grad_norm": 2.283444881439209, "learning_rate": 3.5088582122499624e-06, "loss": 0.5076, "step": 14516 }, { "epoch": 0.6085325340850738, "grad_norm": 1.758771538734436, "learning_rate": 3.508210265178841e-06, "loss": 0.5326, "step": 14517 }, { "epoch": 0.6085744527000828, "grad_norm": 2.272601366043091, "learning_rate": 3.507562345606798e-06, "loss": 0.5028, "step": 14518 }, { "epoch": 0.6086163713150917, "grad_norm": 1.8215993642807007, "learning_rate": 3.50691445354578e-06, "loss": 0.4511, "step": 14519 }, { "epoch": 0.6086582899301007, "grad_norm": 2.0133895874023438, "learning_rate": 3.5062665890077326e-06, "loss": 0.5146, "step": 14520 }, { "epoch": 0.6087002085451096, "grad_norm": 1.7863026857376099, "learning_rate": 3.5056187520045937e-06, "loss": 0.5273, "step": 14521 }, { "epoch": 0.6087421271601187, "grad_norm": 1.6504607200622559, "learning_rate": 3.504970942548307e-06, "loss": 0.4675, "step": 14522 }, { "epoch": 0.6087840457751276, "grad_norm": 1.7224887609481812, "learning_rate": 3.5043231606508154e-06, "loss": 0.5176, "step": 14523 }, { "epoch": 0.6088259643901366, "grad_norm": 1.788879156112671, "learning_rate": 3.5036754063240574e-06, "loss": 0.4845, "step": 14524 }, { "epoch": 0.6088678830051455, "grad_norm": 1.7686916589736938, "learning_rate": 3.503027679579975e-06, "loss": 0.4759, "step": 14525 }, { "epoch": 0.6089098016201545, "grad_norm": 1.904590368270874, "learning_rate": 3.502379980430509e-06, "loss": 0.5168, "step": 14526 }, { "epoch": 0.6089517202351634, "grad_norm": 1.9014356136322021, "learning_rate": 3.501732308887596e-06, "loss": 0.4777, "step": 14527 }, { "epoch": 0.6089936388501724, "grad_norm": 1.8549251556396484, "learning_rate": 3.5010846649631757e-06, "loss": 0.522, "step": 14528 }, { "epoch": 0.6090355574651813, "grad_norm": 1.8396207094192505, "learning_rate": 3.500437048669191e-06, "loss": 0.51, "step": 14529 }, { "epoch": 0.6090774760801904, "grad_norm": 1.7554965019226074, "learning_rate": 3.499789460017574e-06, "loss": 0.5013, "step": 14530 }, { "epoch": 0.6091193946951993, "grad_norm": 1.6582145690917969, "learning_rate": 3.499141899020263e-06, "loss": 0.4386, "step": 14531 }, { "epoch": 0.6091613133102083, "grad_norm": 1.7536180019378662, "learning_rate": 3.4984943656891988e-06, "loss": 0.4769, "step": 14532 }, { "epoch": 0.6092032319252172, "grad_norm": 1.6974703073501587, "learning_rate": 3.4978468600363135e-06, "loss": 0.4879, "step": 14533 }, { "epoch": 0.6092451505402261, "grad_norm": 1.8907053470611572, "learning_rate": 3.497199382073547e-06, "loss": 0.5119, "step": 14534 }, { "epoch": 0.6092870691552351, "grad_norm": 1.7456711530685425, "learning_rate": 3.4965519318128305e-06, "loss": 0.5247, "step": 14535 }, { "epoch": 0.609328987770244, "grad_norm": 1.6262778043746948, "learning_rate": 3.4959045092660996e-06, "loss": 0.4632, "step": 14536 }, { "epoch": 0.609370906385253, "grad_norm": 3.098085641860962, "learning_rate": 3.495257114445292e-06, "loss": 0.53, "step": 14537 }, { "epoch": 0.609412825000262, "grad_norm": 1.8653725385665894, "learning_rate": 3.4946097473623377e-06, "loss": 0.4755, "step": 14538 }, { "epoch": 0.609454743615271, "grad_norm": 1.9255306720733643, "learning_rate": 3.4939624080291717e-06, "loss": 0.4762, "step": 14539 }, { "epoch": 0.6094966622302799, "grad_norm": 2.526275634765625, "learning_rate": 3.4933150964577283e-06, "loss": 0.4823, "step": 14540 }, { "epoch": 0.6095385808452889, "grad_norm": 1.888055443763733, "learning_rate": 3.4926678126599354e-06, "loss": 0.5086, "step": 14541 }, { "epoch": 0.6095804994602978, "grad_norm": 1.5091674327850342, "learning_rate": 3.4920205566477284e-06, "loss": 0.5241, "step": 14542 }, { "epoch": 0.6096224180753068, "grad_norm": 3.784684658050537, "learning_rate": 3.4913733284330396e-06, "loss": 0.5036, "step": 14543 }, { "epoch": 0.6096643366903157, "grad_norm": 2.1399152278900146, "learning_rate": 3.4907261280277973e-06, "loss": 0.4766, "step": 14544 }, { "epoch": 0.6097062553053247, "grad_norm": 1.6067765951156616, "learning_rate": 3.4900789554439307e-06, "loss": 0.4091, "step": 14545 }, { "epoch": 0.6097481739203336, "grad_norm": 4.104739189147949, "learning_rate": 3.4894318106933735e-06, "loss": 0.5089, "step": 14546 }, { "epoch": 0.6097900925353427, "grad_norm": 1.8181263208389282, "learning_rate": 3.48878469378805e-06, "loss": 0.4892, "step": 14547 }, { "epoch": 0.6098320111503516, "grad_norm": 1.8795639276504517, "learning_rate": 3.4881376047398923e-06, "loss": 0.5057, "step": 14548 }, { "epoch": 0.6098739297653606, "grad_norm": 1.5035163164138794, "learning_rate": 3.4874905435608287e-06, "loss": 0.4482, "step": 14549 }, { "epoch": 0.6099158483803695, "grad_norm": 2.2503623962402344, "learning_rate": 3.486843510262785e-06, "loss": 0.4792, "step": 14550 }, { "epoch": 0.6099577669953785, "grad_norm": 3.0896525382995605, "learning_rate": 3.486196504857689e-06, "loss": 0.5016, "step": 14551 }, { "epoch": 0.6099996856103874, "grad_norm": 2.739441394805908, "learning_rate": 3.485549527357469e-06, "loss": 0.49, "step": 14552 }, { "epoch": 0.6100416042253964, "grad_norm": 1.8369992971420288, "learning_rate": 3.4849025777740486e-06, "loss": 0.4955, "step": 14553 }, { "epoch": 0.6100835228404053, "grad_norm": 3.3743438720703125, "learning_rate": 3.4842556561193553e-06, "loss": 0.4672, "step": 14554 }, { "epoch": 0.6101254414554144, "grad_norm": 2.3285508155822754, "learning_rate": 3.4836087624053144e-06, "loss": 0.4942, "step": 14555 }, { "epoch": 0.6101673600704233, "grad_norm": 1.81049644947052, "learning_rate": 3.4829618966438477e-06, "loss": 0.4749, "step": 14556 }, { "epoch": 0.6102092786854323, "grad_norm": 1.7603721618652344, "learning_rate": 3.482315058846881e-06, "loss": 0.5135, "step": 14557 }, { "epoch": 0.6102511973004412, "grad_norm": 1.9012585878372192, "learning_rate": 3.481668249026341e-06, "loss": 0.4748, "step": 14558 }, { "epoch": 0.6102931159154501, "grad_norm": 1.7686916589736938, "learning_rate": 3.4810214671941457e-06, "loss": 0.472, "step": 14559 }, { "epoch": 0.6103350345304591, "grad_norm": 1.6828173398971558, "learning_rate": 3.480374713362219e-06, "loss": 0.5223, "step": 14560 }, { "epoch": 0.610376953145468, "grad_norm": 1.9936468601226807, "learning_rate": 3.479727987542486e-06, "loss": 0.4382, "step": 14561 }, { "epoch": 0.610418871760477, "grad_norm": 1.7593097686767578, "learning_rate": 3.479081289746863e-06, "loss": 0.4484, "step": 14562 }, { "epoch": 0.610460790375486, "grad_norm": 1.8903663158416748, "learning_rate": 3.4784346199872753e-06, "loss": 0.4781, "step": 14563 }, { "epoch": 0.610502708990495, "grad_norm": 1.7141045331954956, "learning_rate": 3.4777879782756434e-06, "loss": 0.4823, "step": 14564 }, { "epoch": 0.6105446276055039, "grad_norm": 3.8648011684417725, "learning_rate": 3.4771413646238832e-06, "loss": 0.5219, "step": 14565 }, { "epoch": 0.6105865462205129, "grad_norm": 1.8544961214065552, "learning_rate": 3.4764947790439164e-06, "loss": 0.483, "step": 14566 }, { "epoch": 0.6106284648355218, "grad_norm": 2.101916790008545, "learning_rate": 3.4758482215476654e-06, "loss": 0.5416, "step": 14567 }, { "epoch": 0.6106703834505308, "grad_norm": 1.579346776008606, "learning_rate": 3.4752016921470432e-06, "loss": 0.507, "step": 14568 }, { "epoch": 0.6107123020655397, "grad_norm": 1.6561683416366577, "learning_rate": 3.474555190853969e-06, "loss": 0.4753, "step": 14569 }, { "epoch": 0.6107542206805487, "grad_norm": 2.2772912979125977, "learning_rate": 3.473908717680363e-06, "loss": 0.535, "step": 14570 }, { "epoch": 0.6107961392955576, "grad_norm": 3.2646353244781494, "learning_rate": 3.4732622726381383e-06, "loss": 0.4845, "step": 14571 }, { "epoch": 0.6108380579105667, "grad_norm": 1.8941677808761597, "learning_rate": 3.4726158557392133e-06, "loss": 0.4592, "step": 14572 }, { "epoch": 0.6108799765255756, "grad_norm": 1.7918897867202759, "learning_rate": 3.471969466995505e-06, "loss": 0.5248, "step": 14573 }, { "epoch": 0.6109218951405846, "grad_norm": 1.676213264465332, "learning_rate": 3.471323106418925e-06, "loss": 0.4585, "step": 14574 }, { "epoch": 0.6109638137555935, "grad_norm": 2.0910940170288086, "learning_rate": 3.4706767740213904e-06, "loss": 0.5116, "step": 14575 }, { "epoch": 0.6110057323706025, "grad_norm": 1.6496039628982544, "learning_rate": 3.470030469814817e-06, "loss": 0.4888, "step": 14576 }, { "epoch": 0.6110476509856114, "grad_norm": 1.795583724975586, "learning_rate": 3.469384193811116e-06, "loss": 0.5237, "step": 14577 }, { "epoch": 0.6110895696006204, "grad_norm": 1.850273609161377, "learning_rate": 3.4687379460222005e-06, "loss": 0.5267, "step": 14578 }, { "epoch": 0.6111314882156293, "grad_norm": 1.9361798763275146, "learning_rate": 3.4680917264599856e-06, "loss": 0.4941, "step": 14579 }, { "epoch": 0.6111734068306384, "grad_norm": 2.311511516571045, "learning_rate": 3.46744553513638e-06, "loss": 0.4118, "step": 14580 }, { "epoch": 0.6112153254456473, "grad_norm": 2.317399501800537, "learning_rate": 3.4667993720632986e-06, "loss": 0.5049, "step": 14581 }, { "epoch": 0.6112572440606563, "grad_norm": 1.8345047235488892, "learning_rate": 3.4661532372526518e-06, "loss": 0.4793, "step": 14582 }, { "epoch": 0.6112991626756652, "grad_norm": 2.4959423542022705, "learning_rate": 3.465507130716348e-06, "loss": 0.508, "step": 14583 }, { "epoch": 0.6113410812906741, "grad_norm": 1.8304003477096558, "learning_rate": 3.4648610524662986e-06, "loss": 0.4849, "step": 14584 }, { "epoch": 0.6113829999056831, "grad_norm": 1.6284887790679932, "learning_rate": 3.4642150025144156e-06, "loss": 0.4774, "step": 14585 }, { "epoch": 0.611424918520692, "grad_norm": 1.8161227703094482, "learning_rate": 3.4635689808726048e-06, "loss": 0.5213, "step": 14586 }, { "epoch": 0.611466837135701, "grad_norm": 1.6629068851470947, "learning_rate": 3.462922987552775e-06, "loss": 0.4475, "step": 14587 }, { "epoch": 0.61150875575071, "grad_norm": 1.7681244611740112, "learning_rate": 3.4622770225668367e-06, "loss": 0.4715, "step": 14588 }, { "epoch": 0.611550674365719, "grad_norm": 1.801107406616211, "learning_rate": 3.4616310859266935e-06, "loss": 0.4904, "step": 14589 }, { "epoch": 0.6115925929807279, "grad_norm": 1.500996708869934, "learning_rate": 3.4609851776442553e-06, "loss": 0.4507, "step": 14590 }, { "epoch": 0.6116345115957369, "grad_norm": 1.9773083925247192, "learning_rate": 3.4603392977314293e-06, "loss": 0.5131, "step": 14591 }, { "epoch": 0.6116764302107458, "grad_norm": 1.6188859939575195, "learning_rate": 3.4596934462001173e-06, "loss": 0.4743, "step": 14592 }, { "epoch": 0.6117183488257548, "grad_norm": 1.8364613056182861, "learning_rate": 3.4590476230622276e-06, "loss": 0.466, "step": 14593 }, { "epoch": 0.6117602674407637, "grad_norm": 1.8402150869369507, "learning_rate": 3.4584018283296674e-06, "loss": 0.536, "step": 14594 }, { "epoch": 0.6118021860557727, "grad_norm": 2.025733232498169, "learning_rate": 3.457756062014336e-06, "loss": 0.5752, "step": 14595 }, { "epoch": 0.6118441046707817, "grad_norm": 1.8184725046157837, "learning_rate": 3.457110324128139e-06, "loss": 0.4855, "step": 14596 }, { "epoch": 0.6118860232857907, "grad_norm": 2.0389466285705566, "learning_rate": 3.456464614682982e-06, "loss": 0.5141, "step": 14597 }, { "epoch": 0.6119279419007996, "grad_norm": 1.6149052381515503, "learning_rate": 3.4558189336907642e-06, "loss": 0.4582, "step": 14598 }, { "epoch": 0.6119698605158086, "grad_norm": 2.182659149169922, "learning_rate": 3.455173281163391e-06, "loss": 0.4832, "step": 14599 }, { "epoch": 0.6120117791308175, "grad_norm": 1.9075415134429932, "learning_rate": 3.4545276571127617e-06, "loss": 0.5083, "step": 14600 }, { "epoch": 0.6120536977458265, "grad_norm": 3.1105806827545166, "learning_rate": 3.4538820615507795e-06, "loss": 0.5096, "step": 14601 }, { "epoch": 0.6120956163608354, "grad_norm": 1.7467167377471924, "learning_rate": 3.4532364944893426e-06, "loss": 0.4835, "step": 14602 }, { "epoch": 0.6121375349758444, "grad_norm": 1.4078954458236694, "learning_rate": 3.452590955940355e-06, "loss": 0.4322, "step": 14603 }, { "epoch": 0.6121794535908534, "grad_norm": 1.8346936702728271, "learning_rate": 3.4519454459157107e-06, "loss": 0.4645, "step": 14604 }, { "epoch": 0.6122213722058624, "grad_norm": 1.830230712890625, "learning_rate": 3.451299964427314e-06, "loss": 0.4198, "step": 14605 }, { "epoch": 0.6122632908208713, "grad_norm": 1.9483082294464111, "learning_rate": 3.4506545114870623e-06, "loss": 0.4939, "step": 14606 }, { "epoch": 0.6123052094358803, "grad_norm": 1.762022614479065, "learning_rate": 3.4500090871068503e-06, "loss": 0.5118, "step": 14607 }, { "epoch": 0.6123471280508892, "grad_norm": 1.8787709474563599, "learning_rate": 3.4493636912985805e-06, "loss": 0.5224, "step": 14608 }, { "epoch": 0.6123890466658981, "grad_norm": 1.6967335939407349, "learning_rate": 3.448718324074145e-06, "loss": 0.4594, "step": 14609 }, { "epoch": 0.6124309652809071, "grad_norm": 1.9125581979751587, "learning_rate": 3.4480729854454443e-06, "loss": 0.4695, "step": 14610 }, { "epoch": 0.612472883895916, "grad_norm": 1.741499662399292, "learning_rate": 3.4474276754243728e-06, "loss": 0.4683, "step": 14611 }, { "epoch": 0.612514802510925, "grad_norm": 1.5454834699630737, "learning_rate": 3.4467823940228238e-06, "loss": 0.4358, "step": 14612 }, { "epoch": 0.612556721125934, "grad_norm": 2.0181665420532227, "learning_rate": 3.4461371412526944e-06, "loss": 0.4979, "step": 14613 }, { "epoch": 0.612598639740943, "grad_norm": 1.6278365850448608, "learning_rate": 3.4454919171258815e-06, "loss": 0.5977, "step": 14614 }, { "epoch": 0.6126405583559519, "grad_norm": 1.7078468799591064, "learning_rate": 3.4448467216542737e-06, "loss": 0.4368, "step": 14615 }, { "epoch": 0.6126824769709609, "grad_norm": 1.8421850204467773, "learning_rate": 3.4442015548497667e-06, "loss": 0.4994, "step": 14616 }, { "epoch": 0.6127243955859698, "grad_norm": 1.50448477268219, "learning_rate": 3.443556416724254e-06, "loss": 0.491, "step": 14617 }, { "epoch": 0.6127663142009788, "grad_norm": 1.9406906366348267, "learning_rate": 3.4429113072896257e-06, "loss": 0.5325, "step": 14618 }, { "epoch": 0.6128082328159877, "grad_norm": 1.6350497007369995, "learning_rate": 3.442266226557777e-06, "loss": 0.4824, "step": 14619 }, { "epoch": 0.6128501514309967, "grad_norm": 1.771070122718811, "learning_rate": 3.4416211745405956e-06, "loss": 0.478, "step": 14620 }, { "epoch": 0.6128920700460057, "grad_norm": 2.0666821002960205, "learning_rate": 3.4409761512499727e-06, "loss": 0.504, "step": 14621 }, { "epoch": 0.6129339886610147, "grad_norm": 1.58277428150177, "learning_rate": 3.440331156697799e-06, "loss": 0.5515, "step": 14622 }, { "epoch": 0.6129759072760236, "grad_norm": 1.811973214149475, "learning_rate": 3.439686190895967e-06, "loss": 0.4676, "step": 14623 }, { "epoch": 0.6130178258910326, "grad_norm": 1.6383166313171387, "learning_rate": 3.439041253856361e-06, "loss": 0.4986, "step": 14624 }, { "epoch": 0.6130597445060415, "grad_norm": 2.158522605895996, "learning_rate": 3.438396345590871e-06, "loss": 0.4393, "step": 14625 }, { "epoch": 0.6131016631210505, "grad_norm": 2.8229753971099854, "learning_rate": 3.4377514661113876e-06, "loss": 0.4648, "step": 14626 }, { "epoch": 0.6131435817360594, "grad_norm": 13.277711868286133, "learning_rate": 3.437106615429794e-06, "loss": 0.5024, "step": 14627 }, { "epoch": 0.6131855003510684, "grad_norm": 1.5589343309402466, "learning_rate": 3.4364617935579802e-06, "loss": 0.4689, "step": 14628 }, { "epoch": 0.6132274189660774, "grad_norm": 1.788002848625183, "learning_rate": 3.4358170005078327e-06, "loss": 0.4931, "step": 14629 }, { "epoch": 0.6132693375810864, "grad_norm": 2.037759304046631, "learning_rate": 3.4351722362912343e-06, "loss": 0.4806, "step": 14630 }, { "epoch": 0.6133112561960953, "grad_norm": 1.7856343984603882, "learning_rate": 3.4345275009200727e-06, "loss": 0.5227, "step": 14631 }, { "epoch": 0.6133531748111043, "grad_norm": 1.8564653396606445, "learning_rate": 3.433882794406235e-06, "loss": 0.4795, "step": 14632 }, { "epoch": 0.6133950934261132, "grad_norm": 1.5160351991653442, "learning_rate": 3.433238116761601e-06, "loss": 0.472, "step": 14633 }, { "epoch": 0.6134370120411221, "grad_norm": 1.658227801322937, "learning_rate": 3.4325934679980556e-06, "loss": 0.4806, "step": 14634 }, { "epoch": 0.6134789306561311, "grad_norm": 2.5283312797546387, "learning_rate": 3.4319488481274843e-06, "loss": 0.522, "step": 14635 }, { "epoch": 0.61352084927114, "grad_norm": 1.7424280643463135, "learning_rate": 3.4313042571617665e-06, "loss": 0.4969, "step": 14636 }, { "epoch": 0.613562767886149, "grad_norm": 2.1706807613372803, "learning_rate": 3.4306596951127867e-06, "loss": 0.5014, "step": 14637 }, { "epoch": 0.613604686501158, "grad_norm": 2.0633726119995117, "learning_rate": 3.430015161992427e-06, "loss": 0.5069, "step": 14638 }, { "epoch": 0.613646605116167, "grad_norm": 1.5318596363067627, "learning_rate": 3.4293706578125662e-06, "loss": 0.4979, "step": 14639 }, { "epoch": 0.6136885237311759, "grad_norm": 1.579113245010376, "learning_rate": 3.428726182585086e-06, "loss": 0.5002, "step": 14640 }, { "epoch": 0.6137304423461849, "grad_norm": 1.6907970905303955, "learning_rate": 3.4280817363218675e-06, "loss": 0.4697, "step": 14641 }, { "epoch": 0.6137723609611938, "grad_norm": 1.9424229860305786, "learning_rate": 3.4274373190347876e-06, "loss": 0.5256, "step": 14642 }, { "epoch": 0.6138142795762028, "grad_norm": 2.009381055831909, "learning_rate": 3.4267929307357272e-06, "loss": 0.531, "step": 14643 }, { "epoch": 0.6138561981912117, "grad_norm": 1.9013335704803467, "learning_rate": 3.426148571436566e-06, "loss": 0.5178, "step": 14644 }, { "epoch": 0.6138981168062208, "grad_norm": 1.6349103450775146, "learning_rate": 3.425504241149178e-06, "loss": 0.4767, "step": 14645 }, { "epoch": 0.6139400354212297, "grad_norm": 2.0565555095672607, "learning_rate": 3.424859939885443e-06, "loss": 0.4439, "step": 14646 }, { "epoch": 0.6139819540362387, "grad_norm": 1.6752253770828247, "learning_rate": 3.424215667657238e-06, "loss": 0.5119, "step": 14647 }, { "epoch": 0.6140238726512476, "grad_norm": 1.7495688199996948, "learning_rate": 3.4235714244764385e-06, "loss": 0.5189, "step": 14648 }, { "epoch": 0.6140657912662566, "grad_norm": 1.874057650566101, "learning_rate": 3.422927210354919e-06, "loss": 0.491, "step": 14649 }, { "epoch": 0.6141077098812655, "grad_norm": 1.9816285371780396, "learning_rate": 3.422283025304559e-06, "loss": 0.5204, "step": 14650 }, { "epoch": 0.6141496284962745, "grad_norm": 1.7138934135437012, "learning_rate": 3.421638869337228e-06, "loss": 0.4305, "step": 14651 }, { "epoch": 0.6141915471112834, "grad_norm": 1.7248234748840332, "learning_rate": 3.4209947424648027e-06, "loss": 0.4977, "step": 14652 }, { "epoch": 0.6142334657262924, "grad_norm": 1.779731273651123, "learning_rate": 3.4203506446991584e-06, "loss": 0.4663, "step": 14653 }, { "epoch": 0.6142753843413014, "grad_norm": 1.7246603965759277, "learning_rate": 3.4197065760521634e-06, "loss": 0.4772, "step": 14654 }, { "epoch": 0.6143173029563104, "grad_norm": 1.8191540241241455, "learning_rate": 3.4190625365356933e-06, "loss": 0.4505, "step": 14655 }, { "epoch": 0.6143592215713193, "grad_norm": 2.1235220432281494, "learning_rate": 3.418418526161622e-06, "loss": 0.5556, "step": 14656 }, { "epoch": 0.6144011401863283, "grad_norm": 1.7608009576797485, "learning_rate": 3.4177745449418177e-06, "loss": 0.45, "step": 14657 }, { "epoch": 0.6144430588013372, "grad_norm": 1.7096079587936401, "learning_rate": 3.4171305928881516e-06, "loss": 0.4725, "step": 14658 }, { "epoch": 0.6144849774163461, "grad_norm": 1.7913143634796143, "learning_rate": 3.416486670012496e-06, "loss": 0.4696, "step": 14659 }, { "epoch": 0.6145268960313551, "grad_norm": 1.6869120597839355, "learning_rate": 3.4158427763267177e-06, "loss": 0.5062, "step": 14660 }, { "epoch": 0.614568814646364, "grad_norm": 1.7045570611953735, "learning_rate": 3.415198911842689e-06, "loss": 0.5344, "step": 14661 }, { "epoch": 0.614610733261373, "grad_norm": 1.5532358884811401, "learning_rate": 3.414555076572278e-06, "loss": 0.4621, "step": 14662 }, { "epoch": 0.614652651876382, "grad_norm": 2.2405686378479004, "learning_rate": 3.413911270527351e-06, "loss": 0.4775, "step": 14663 }, { "epoch": 0.614694570491391, "grad_norm": 1.7052351236343384, "learning_rate": 3.413267493719776e-06, "loss": 0.5082, "step": 14664 }, { "epoch": 0.6147364891063999, "grad_norm": 2.2346348762512207, "learning_rate": 3.4126237461614243e-06, "loss": 0.4922, "step": 14665 }, { "epoch": 0.6147784077214089, "grad_norm": 1.8640824556350708, "learning_rate": 3.411980027864158e-06, "loss": 0.5497, "step": 14666 }, { "epoch": 0.6148203263364178, "grad_norm": 2.262399911880493, "learning_rate": 3.4113363388398435e-06, "loss": 0.4703, "step": 14667 }, { "epoch": 0.6148622449514268, "grad_norm": 2.0637147426605225, "learning_rate": 3.4106926791003493e-06, "loss": 0.4991, "step": 14668 }, { "epoch": 0.6149041635664357, "grad_norm": 1.8910425901412964, "learning_rate": 3.410049048657536e-06, "loss": 0.5071, "step": 14669 }, { "epoch": 0.6149460821814448, "grad_norm": 1.862060546875, "learning_rate": 3.409405447523272e-06, "loss": 0.5102, "step": 14670 }, { "epoch": 0.6149880007964537, "grad_norm": 1.621540904045105, "learning_rate": 3.4087618757094203e-06, "loss": 0.4754, "step": 14671 }, { "epoch": 0.6150299194114627, "grad_norm": 2.612861156463623, "learning_rate": 3.408118333227842e-06, "loss": 0.5098, "step": 14672 }, { "epoch": 0.6150718380264716, "grad_norm": 1.5751296281814575, "learning_rate": 3.4074748200904014e-06, "loss": 0.4844, "step": 14673 }, { "epoch": 0.6151137566414806, "grad_norm": 1.6621063947677612, "learning_rate": 3.4068313363089624e-06, "loss": 0.4263, "step": 14674 }, { "epoch": 0.6151556752564895, "grad_norm": 1.8540990352630615, "learning_rate": 3.4061878818953846e-06, "loss": 0.5136, "step": 14675 }, { "epoch": 0.6151975938714985, "grad_norm": 1.933419942855835, "learning_rate": 3.4055444568615292e-06, "loss": 0.5135, "step": 14676 }, { "epoch": 0.6152395124865074, "grad_norm": 1.7649774551391602, "learning_rate": 3.404901061219259e-06, "loss": 0.4788, "step": 14677 }, { "epoch": 0.6152814311015165, "grad_norm": 2.056431531906128, "learning_rate": 3.4042576949804307e-06, "loss": 0.5128, "step": 14678 }, { "epoch": 0.6153233497165254, "grad_norm": 1.8356820344924927, "learning_rate": 3.4036143581569066e-06, "loss": 0.4653, "step": 14679 }, { "epoch": 0.6153652683315344, "grad_norm": 1.5561528205871582, "learning_rate": 3.4029710507605463e-06, "loss": 0.4344, "step": 14680 }, { "epoch": 0.6154071869465433, "grad_norm": 2.164363145828247, "learning_rate": 3.4023277728032045e-06, "loss": 0.5257, "step": 14681 }, { "epoch": 0.6154491055615523, "grad_norm": 1.85269296169281, "learning_rate": 3.4016845242967433e-06, "loss": 0.4971, "step": 14682 }, { "epoch": 0.6154910241765612, "grad_norm": 1.6882792711257935, "learning_rate": 3.4010413052530167e-06, "loss": 0.5387, "step": 14683 }, { "epoch": 0.6155329427915701, "grad_norm": 2.2514634132385254, "learning_rate": 3.400398115683884e-06, "loss": 0.4657, "step": 14684 }, { "epoch": 0.6155748614065791, "grad_norm": 1.6415297985076904, "learning_rate": 3.3997549556012e-06, "loss": 0.4997, "step": 14685 }, { "epoch": 0.615616780021588, "grad_norm": 1.7576264142990112, "learning_rate": 3.399111825016822e-06, "loss": 0.4347, "step": 14686 }, { "epoch": 0.6156586986365971, "grad_norm": 1.6089242696762085, "learning_rate": 3.3984687239426027e-06, "loss": 0.5142, "step": 14687 }, { "epoch": 0.615700617251606, "grad_norm": 2.36189341545105, "learning_rate": 3.3978256523904006e-06, "loss": 0.501, "step": 14688 }, { "epoch": 0.615742535866615, "grad_norm": 1.5034540891647339, "learning_rate": 3.397182610372066e-06, "loss": 0.4537, "step": 14689 }, { "epoch": 0.6157844544816239, "grad_norm": 1.7832746505737305, "learning_rate": 3.3965395978994554e-06, "loss": 0.5441, "step": 14690 }, { "epoch": 0.6158263730966329, "grad_norm": 1.59839928150177, "learning_rate": 3.3958966149844214e-06, "loss": 0.5055, "step": 14691 }, { "epoch": 0.6158682917116418, "grad_norm": 1.8799066543579102, "learning_rate": 3.395253661638813e-06, "loss": 0.4833, "step": 14692 }, { "epoch": 0.6159102103266508, "grad_norm": 1.7553508281707764, "learning_rate": 3.3946107378744852e-06, "loss": 0.5042, "step": 14693 }, { "epoch": 0.6159521289416597, "grad_norm": 1.5799447298049927, "learning_rate": 3.393967843703292e-06, "loss": 0.509, "step": 14694 }, { "epoch": 0.6159940475566688, "grad_norm": 1.9584574699401855, "learning_rate": 3.3933249791370794e-06, "loss": 0.5288, "step": 14695 }, { "epoch": 0.6160359661716777, "grad_norm": 2.228342056274414, "learning_rate": 3.392682144187699e-06, "loss": 0.4658, "step": 14696 }, { "epoch": 0.6160778847866867, "grad_norm": 2.354857921600342, "learning_rate": 3.3920393388670027e-06, "loss": 0.4609, "step": 14697 }, { "epoch": 0.6161198034016956, "grad_norm": 1.6300700902938843, "learning_rate": 3.391396563186836e-06, "loss": 0.4718, "step": 14698 }, { "epoch": 0.6161617220167046, "grad_norm": 1.8013139963150024, "learning_rate": 3.3907538171590515e-06, "loss": 0.478, "step": 14699 }, { "epoch": 0.6162036406317135, "grad_norm": 1.9546798467636108, "learning_rate": 3.390111100795497e-06, "loss": 0.4384, "step": 14700 }, { "epoch": 0.6162455592467225, "grad_norm": 1.694750189781189, "learning_rate": 3.389468414108016e-06, "loss": 0.5353, "step": 14701 }, { "epoch": 0.6162874778617314, "grad_norm": 1.7376033067703247, "learning_rate": 3.3888257571084586e-06, "loss": 0.5238, "step": 14702 }, { "epoch": 0.6163293964767405, "grad_norm": 1.5581309795379639, "learning_rate": 3.388183129808673e-06, "loss": 0.4558, "step": 14703 }, { "epoch": 0.6163713150917494, "grad_norm": 1.4508787393569946, "learning_rate": 3.3875405322205024e-06, "loss": 0.48, "step": 14704 }, { "epoch": 0.6164132337067584, "grad_norm": 1.6097373962402344, "learning_rate": 3.3868979643557915e-06, "loss": 0.5045, "step": 14705 }, { "epoch": 0.6164551523217673, "grad_norm": 1.8804399967193604, "learning_rate": 3.386255426226389e-06, "loss": 0.4937, "step": 14706 }, { "epoch": 0.6164970709367763, "grad_norm": 1.5997676849365234, "learning_rate": 3.3856129178441344e-06, "loss": 0.4605, "step": 14707 }, { "epoch": 0.6165389895517852, "grad_norm": 1.6561325788497925, "learning_rate": 3.384970439220875e-06, "loss": 0.5004, "step": 14708 }, { "epoch": 0.6165809081667941, "grad_norm": 1.6090776920318604, "learning_rate": 3.3843279903684534e-06, "loss": 0.4444, "step": 14709 }, { "epoch": 0.6166228267818031, "grad_norm": 2.076923131942749, "learning_rate": 3.3836855712987106e-06, "loss": 0.5506, "step": 14710 }, { "epoch": 0.616664745396812, "grad_norm": 1.7031594514846802, "learning_rate": 3.383043182023489e-06, "loss": 0.4906, "step": 14711 }, { "epoch": 0.6167066640118211, "grad_norm": 1.604828119277954, "learning_rate": 3.3824008225546336e-06, "loss": 0.4827, "step": 14712 }, { "epoch": 0.61674858262683, "grad_norm": 1.5870269536972046, "learning_rate": 3.381758492903981e-06, "loss": 0.4332, "step": 14713 }, { "epoch": 0.616790501241839, "grad_norm": 1.3093959093093872, "learning_rate": 3.3811161930833725e-06, "loss": 0.457, "step": 14714 }, { "epoch": 0.6168324198568479, "grad_norm": 2.051828145980835, "learning_rate": 3.380473923104652e-06, "loss": 0.5165, "step": 14715 }, { "epoch": 0.6168743384718569, "grad_norm": 1.9197827577590942, "learning_rate": 3.379831682979653e-06, "loss": 0.5324, "step": 14716 }, { "epoch": 0.6169162570868658, "grad_norm": 2.4222235679626465, "learning_rate": 3.379189472720218e-06, "loss": 0.521, "step": 14717 }, { "epoch": 0.6169581757018748, "grad_norm": 1.4894944429397583, "learning_rate": 3.3785472923381855e-06, "loss": 0.4327, "step": 14718 }, { "epoch": 0.6170000943168837, "grad_norm": 1.8604810237884521, "learning_rate": 3.37790514184539e-06, "loss": 0.5287, "step": 14719 }, { "epoch": 0.6170420129318928, "grad_norm": 1.6060808897018433, "learning_rate": 3.377263021253671e-06, "loss": 0.4232, "step": 14720 }, { "epoch": 0.6170839315469017, "grad_norm": 1.9345512390136719, "learning_rate": 3.3766209305748663e-06, "loss": 0.4999, "step": 14721 }, { "epoch": 0.6171258501619107, "grad_norm": 1.8782713413238525, "learning_rate": 3.3759788698208108e-06, "loss": 0.4819, "step": 14722 }, { "epoch": 0.6171677687769196, "grad_norm": 2.0429396629333496, "learning_rate": 3.3753368390033373e-06, "loss": 0.4984, "step": 14723 }, { "epoch": 0.6172096873919286, "grad_norm": 2.1399035453796387, "learning_rate": 3.374694838134286e-06, "loss": 0.4888, "step": 14724 }, { "epoch": 0.6172516060069375, "grad_norm": 1.7033668756484985, "learning_rate": 3.374052867225486e-06, "loss": 0.5225, "step": 14725 }, { "epoch": 0.6172935246219465, "grad_norm": 1.6052483320236206, "learning_rate": 3.373410926288775e-06, "loss": 0.4759, "step": 14726 }, { "epoch": 0.6173354432369554, "grad_norm": 1.7252740859985352, "learning_rate": 3.372769015335985e-06, "loss": 0.5382, "step": 14727 }, { "epoch": 0.6173773618519645, "grad_norm": 1.69936203956604, "learning_rate": 3.3721271343789475e-06, "loss": 0.4736, "step": 14728 }, { "epoch": 0.6174192804669734, "grad_norm": 1.6313332319259644, "learning_rate": 3.371485283429495e-06, "loss": 0.4816, "step": 14729 }, { "epoch": 0.6174611990819824, "grad_norm": 1.657236099243164, "learning_rate": 3.370843462499462e-06, "loss": 0.4828, "step": 14730 }, { "epoch": 0.6175031176969913, "grad_norm": 1.711800217628479, "learning_rate": 3.370201671600678e-06, "loss": 0.4476, "step": 14731 }, { "epoch": 0.6175450363120003, "grad_norm": 1.6302025318145752, "learning_rate": 3.3695599107449705e-06, "loss": 0.4603, "step": 14732 }, { "epoch": 0.6175869549270092, "grad_norm": 1.6828906536102295, "learning_rate": 3.368918179944175e-06, "loss": 0.5023, "step": 14733 }, { "epoch": 0.6176288735420181, "grad_norm": 1.6050560474395752, "learning_rate": 3.3682764792101153e-06, "loss": 0.5149, "step": 14734 }, { "epoch": 0.6176707921570271, "grad_norm": 1.7213315963745117, "learning_rate": 3.367634808554624e-06, "loss": 0.5051, "step": 14735 }, { "epoch": 0.617712710772036, "grad_norm": 1.568631887435913, "learning_rate": 3.3669931679895285e-06, "loss": 0.5203, "step": 14736 }, { "epoch": 0.6177546293870451, "grad_norm": 1.4785698652267456, "learning_rate": 3.3663515575266563e-06, "loss": 0.4356, "step": 14737 }, { "epoch": 0.617796548002054, "grad_norm": 1.6319812536239624, "learning_rate": 3.3657099771778336e-06, "loss": 0.4776, "step": 14738 }, { "epoch": 0.617838466617063, "grad_norm": 1.753831386566162, "learning_rate": 3.3650684269548905e-06, "loss": 0.446, "step": 14739 }, { "epoch": 0.6178803852320719, "grad_norm": 1.8680428266525269, "learning_rate": 3.364426906869648e-06, "loss": 0.5177, "step": 14740 }, { "epoch": 0.6179223038470809, "grad_norm": 1.7900633811950684, "learning_rate": 3.3637854169339363e-06, "loss": 0.4317, "step": 14741 }, { "epoch": 0.6179642224620898, "grad_norm": 1.985991358757019, "learning_rate": 3.363143957159579e-06, "loss": 0.5046, "step": 14742 }, { "epoch": 0.6180061410770988, "grad_norm": 1.6798741817474365, "learning_rate": 3.362502527558398e-06, "loss": 0.5119, "step": 14743 }, { "epoch": 0.6180480596921077, "grad_norm": 2.0255136489868164, "learning_rate": 3.3618611281422185e-06, "loss": 0.4789, "step": 14744 }, { "epoch": 0.6180899783071168, "grad_norm": 1.6685703992843628, "learning_rate": 3.3612197589228674e-06, "loss": 0.4773, "step": 14745 }, { "epoch": 0.6181318969221257, "grad_norm": 1.6723636388778687, "learning_rate": 3.360578419912163e-06, "loss": 0.5403, "step": 14746 }, { "epoch": 0.6181738155371347, "grad_norm": 1.6238296031951904, "learning_rate": 3.359937111121928e-06, "loss": 0.4956, "step": 14747 }, { "epoch": 0.6182157341521436, "grad_norm": 1.725005865097046, "learning_rate": 3.3592958325639875e-06, "loss": 0.5025, "step": 14748 }, { "epoch": 0.6182576527671526, "grad_norm": 1.619922399520874, "learning_rate": 3.3586545842501573e-06, "loss": 0.4804, "step": 14749 }, { "epoch": 0.6182995713821615, "grad_norm": 1.8366438150405884, "learning_rate": 3.3580133661922628e-06, "loss": 0.4821, "step": 14750 }, { "epoch": 0.6183414899971705, "grad_norm": 2.2014670372009277, "learning_rate": 3.357372178402122e-06, "loss": 0.4427, "step": 14751 }, { "epoch": 0.6183834086121794, "grad_norm": 1.6001044511795044, "learning_rate": 3.356731020891553e-06, "loss": 0.4821, "step": 14752 }, { "epoch": 0.6184253272271885, "grad_norm": 1.6115649938583374, "learning_rate": 3.356089893672375e-06, "loss": 0.4957, "step": 14753 }, { "epoch": 0.6184672458421974, "grad_norm": 2.3779938220977783, "learning_rate": 3.35544879675641e-06, "loss": 0.5178, "step": 14754 }, { "epoch": 0.6185091644572064, "grad_norm": 2.9068894386291504, "learning_rate": 3.354807730155471e-06, "loss": 0.4884, "step": 14755 }, { "epoch": 0.6185510830722153, "grad_norm": 1.8566508293151855, "learning_rate": 3.3541666938813765e-06, "loss": 0.5381, "step": 14756 }, { "epoch": 0.6185930016872243, "grad_norm": 1.6137773990631104, "learning_rate": 3.353525687945946e-06, "loss": 0.4566, "step": 14757 }, { "epoch": 0.6186349203022332, "grad_norm": 2.3941562175750732, "learning_rate": 3.3528847123609908e-06, "loss": 0.4545, "step": 14758 }, { "epoch": 0.6186768389172421, "grad_norm": 2.1975581645965576, "learning_rate": 3.3522437671383314e-06, "loss": 0.5766, "step": 14759 }, { "epoch": 0.6187187575322511, "grad_norm": 1.7680004835128784, "learning_rate": 3.3516028522897793e-06, "loss": 0.5151, "step": 14760 }, { "epoch": 0.61876067614726, "grad_norm": 1.6549469232559204, "learning_rate": 3.3509619678271476e-06, "loss": 0.4829, "step": 14761 }, { "epoch": 0.6188025947622691, "grad_norm": 1.7648605108261108, "learning_rate": 3.3503211137622553e-06, "loss": 0.4774, "step": 14762 }, { "epoch": 0.618844513377278, "grad_norm": 3.5209591388702393, "learning_rate": 3.34968029010691e-06, "loss": 0.4434, "step": 14763 }, { "epoch": 0.618886431992287, "grad_norm": 2.2715904712677, "learning_rate": 3.349039496872929e-06, "loss": 0.4618, "step": 14764 }, { "epoch": 0.6189283506072959, "grad_norm": 1.7829090356826782, "learning_rate": 3.3483987340721234e-06, "loss": 0.4732, "step": 14765 }, { "epoch": 0.6189702692223049, "grad_norm": 1.6996442079544067, "learning_rate": 3.3477580017163014e-06, "loss": 0.5066, "step": 14766 }, { "epoch": 0.6190121878373138, "grad_norm": 2.3124542236328125, "learning_rate": 3.347117299817276e-06, "loss": 0.4925, "step": 14767 }, { "epoch": 0.6190541064523228, "grad_norm": 1.5901707410812378, "learning_rate": 3.346476628386861e-06, "loss": 0.4738, "step": 14768 }, { "epoch": 0.6190960250673317, "grad_norm": 1.8273158073425293, "learning_rate": 3.345835987436862e-06, "loss": 0.4798, "step": 14769 }, { "epoch": 0.6191379436823408, "grad_norm": 2.0164847373962402, "learning_rate": 3.345195376979089e-06, "loss": 0.5004, "step": 14770 }, { "epoch": 0.6191798622973497, "grad_norm": 1.91616952419281, "learning_rate": 3.3445547970253533e-06, "loss": 0.456, "step": 14771 }, { "epoch": 0.6192217809123587, "grad_norm": 2.191096305847168, "learning_rate": 3.3439142475874597e-06, "loss": 0.5386, "step": 14772 }, { "epoch": 0.6192636995273676, "grad_norm": 2.052128791809082, "learning_rate": 3.3432737286772175e-06, "loss": 0.474, "step": 14773 }, { "epoch": 0.6193056181423766, "grad_norm": 1.5223027467727661, "learning_rate": 3.342633240306436e-06, "loss": 0.4666, "step": 14774 }, { "epoch": 0.6193475367573855, "grad_norm": 1.6597192287445068, "learning_rate": 3.341992782486917e-06, "loss": 0.5494, "step": 14775 }, { "epoch": 0.6193894553723945, "grad_norm": 1.6471543312072754, "learning_rate": 3.3413523552304683e-06, "loss": 0.4976, "step": 14776 }, { "epoch": 0.6194313739874034, "grad_norm": 2.698498249053955, "learning_rate": 3.340711958548899e-06, "loss": 0.5107, "step": 14777 }, { "epoch": 0.6194732926024125, "grad_norm": 1.6236745119094849, "learning_rate": 3.3400715924540095e-06, "loss": 0.4729, "step": 14778 }, { "epoch": 0.6195152112174214, "grad_norm": 1.5122613906860352, "learning_rate": 3.339431256957604e-06, "loss": 0.4163, "step": 14779 }, { "epoch": 0.6195571298324304, "grad_norm": 1.712312936782837, "learning_rate": 3.33879095207149e-06, "loss": 0.4818, "step": 14780 }, { "epoch": 0.6195990484474393, "grad_norm": 2.4244656562805176, "learning_rate": 3.3381506778074657e-06, "loss": 0.4525, "step": 14781 }, { "epoch": 0.6196409670624483, "grad_norm": 1.9749702215194702, "learning_rate": 3.337510434177338e-06, "loss": 0.4839, "step": 14782 }, { "epoch": 0.6196828856774572, "grad_norm": 2.808551073074341, "learning_rate": 3.3368702211929067e-06, "loss": 0.4464, "step": 14783 }, { "epoch": 0.6197248042924661, "grad_norm": 1.8383210897445679, "learning_rate": 3.3362300388659737e-06, "loss": 0.4907, "step": 14784 }, { "epoch": 0.6197667229074751, "grad_norm": 1.7401663064956665, "learning_rate": 3.3355898872083385e-06, "loss": 0.5237, "step": 14785 }, { "epoch": 0.619808641522484, "grad_norm": 1.8811497688293457, "learning_rate": 3.3349497662318053e-06, "loss": 0.5103, "step": 14786 }, { "epoch": 0.6198505601374931, "grad_norm": 1.6281887292861938, "learning_rate": 3.3343096759481687e-06, "loss": 0.4849, "step": 14787 }, { "epoch": 0.619892478752502, "grad_norm": 1.8889397382736206, "learning_rate": 3.3336696163692317e-06, "loss": 0.4778, "step": 14788 }, { "epoch": 0.619934397367511, "grad_norm": 1.9061707258224487, "learning_rate": 3.333029587506793e-06, "loss": 0.5303, "step": 14789 }, { "epoch": 0.6199763159825199, "grad_norm": 1.5616962909698486, "learning_rate": 3.332389589372648e-06, "loss": 0.4782, "step": 14790 }, { "epoch": 0.6200182345975289, "grad_norm": 1.6494113206863403, "learning_rate": 3.331749621978594e-06, "loss": 0.4708, "step": 14791 }, { "epoch": 0.6200601532125378, "grad_norm": 2.148707628250122, "learning_rate": 3.3311096853364322e-06, "loss": 0.5836, "step": 14792 }, { "epoch": 0.6201020718275468, "grad_norm": 1.5679315328598022, "learning_rate": 3.3304697794579554e-06, "loss": 0.4479, "step": 14793 }, { "epoch": 0.6201439904425557, "grad_norm": 1.9360108375549316, "learning_rate": 3.329829904354959e-06, "loss": 0.4613, "step": 14794 }, { "epoch": 0.6201859090575648, "grad_norm": 1.6025118827819824, "learning_rate": 3.329190060039241e-06, "loss": 0.4212, "step": 14795 }, { "epoch": 0.6202278276725737, "grad_norm": 2.129054069519043, "learning_rate": 3.3285502465225926e-06, "loss": 0.5027, "step": 14796 }, { "epoch": 0.6202697462875827, "grad_norm": 2.056631326675415, "learning_rate": 3.3279104638168107e-06, "loss": 0.4511, "step": 14797 }, { "epoch": 0.6203116649025916, "grad_norm": 1.817148208618164, "learning_rate": 3.3272707119336887e-06, "loss": 0.5281, "step": 14798 }, { "epoch": 0.6203535835176006, "grad_norm": 1.5064167976379395, "learning_rate": 3.326630990885017e-06, "loss": 0.4748, "step": 14799 }, { "epoch": 0.6203955021326095, "grad_norm": 1.7217068672180176, "learning_rate": 3.3259913006825893e-06, "loss": 0.4451, "step": 14800 }, { "epoch": 0.6204374207476185, "grad_norm": 1.8035482168197632, "learning_rate": 3.3253516413382e-06, "loss": 0.5022, "step": 14801 }, { "epoch": 0.6204793393626274, "grad_norm": 1.7617688179016113, "learning_rate": 3.3247120128636366e-06, "loss": 0.5018, "step": 14802 }, { "epoch": 0.6205212579776365, "grad_norm": 1.796427845954895, "learning_rate": 3.3240724152706906e-06, "loss": 0.5037, "step": 14803 }, { "epoch": 0.6205631765926454, "grad_norm": 1.4955830574035645, "learning_rate": 3.3234328485711555e-06, "loss": 0.4603, "step": 14804 }, { "epoch": 0.6206050952076544, "grad_norm": 1.721657395362854, "learning_rate": 3.322793312776815e-06, "loss": 0.4875, "step": 14805 }, { "epoch": 0.6206470138226633, "grad_norm": 1.7053405046463013, "learning_rate": 3.322153807899462e-06, "loss": 0.4705, "step": 14806 }, { "epoch": 0.6206889324376723, "grad_norm": 1.8738179206848145, "learning_rate": 3.321514333950885e-06, "loss": 0.4956, "step": 14807 }, { "epoch": 0.6207308510526812, "grad_norm": 1.7163082361221313, "learning_rate": 3.3208748909428697e-06, "loss": 0.4861, "step": 14808 }, { "epoch": 0.6207727696676901, "grad_norm": 1.8042323589324951, "learning_rate": 3.3202354788872036e-06, "loss": 0.5011, "step": 14809 }, { "epoch": 0.6208146882826991, "grad_norm": 1.7930560111999512, "learning_rate": 3.3195960977956766e-06, "loss": 0.4685, "step": 14810 }, { "epoch": 0.620856606897708, "grad_norm": 1.6063419580459595, "learning_rate": 3.318956747680071e-06, "loss": 0.4458, "step": 14811 }, { "epoch": 0.6208985255127171, "grad_norm": 1.7473161220550537, "learning_rate": 3.3183174285521733e-06, "loss": 0.5015, "step": 14812 }, { "epoch": 0.620940444127726, "grad_norm": 2.1602225303649902, "learning_rate": 3.3176781404237713e-06, "loss": 0.4785, "step": 14813 }, { "epoch": 0.620982362742735, "grad_norm": 1.7604904174804688, "learning_rate": 3.317038883306645e-06, "loss": 0.4498, "step": 14814 }, { "epoch": 0.6210242813577439, "grad_norm": 1.7169792652130127, "learning_rate": 3.31639965721258e-06, "loss": 0.4619, "step": 14815 }, { "epoch": 0.6210661999727529, "grad_norm": 1.5744301080703735, "learning_rate": 3.3157604621533617e-06, "loss": 0.4613, "step": 14816 }, { "epoch": 0.6211081185877618, "grad_norm": 1.6604708433151245, "learning_rate": 3.3151212981407686e-06, "loss": 0.4913, "step": 14817 }, { "epoch": 0.6211500372027708, "grad_norm": 1.9378013610839844, "learning_rate": 3.3144821651865855e-06, "loss": 0.446, "step": 14818 }, { "epoch": 0.6211919558177798, "grad_norm": 1.829755425453186, "learning_rate": 3.3138430633025953e-06, "loss": 0.5478, "step": 14819 }, { "epoch": 0.6212338744327888, "grad_norm": 1.578457236289978, "learning_rate": 3.3132039925005765e-06, "loss": 0.4996, "step": 14820 }, { "epoch": 0.6212757930477977, "grad_norm": 1.7879084348678589, "learning_rate": 3.312564952792309e-06, "loss": 0.5303, "step": 14821 }, { "epoch": 0.6213177116628067, "grad_norm": 2.0932998657226562, "learning_rate": 3.311925944189576e-06, "loss": 0.5015, "step": 14822 }, { "epoch": 0.6213596302778156, "grad_norm": 1.596673607826233, "learning_rate": 3.311286966704152e-06, "loss": 0.4701, "step": 14823 }, { "epoch": 0.6214015488928246, "grad_norm": 2.511594772338867, "learning_rate": 3.310648020347819e-06, "loss": 0.5243, "step": 14824 }, { "epoch": 0.6214434675078335, "grad_norm": 2.1290814876556396, "learning_rate": 3.3100091051323556e-06, "loss": 0.4735, "step": 14825 }, { "epoch": 0.6214853861228425, "grad_norm": 1.8238396644592285, "learning_rate": 3.3093702210695366e-06, "loss": 0.5244, "step": 14826 }, { "epoch": 0.6215273047378514, "grad_norm": 1.707495093345642, "learning_rate": 3.3087313681711397e-06, "loss": 0.4672, "step": 14827 }, { "epoch": 0.6215692233528605, "grad_norm": 1.6211376190185547, "learning_rate": 3.3080925464489444e-06, "loss": 0.4735, "step": 14828 }, { "epoch": 0.6216111419678694, "grad_norm": 1.6597567796707153, "learning_rate": 3.3074537559147225e-06, "loss": 0.5031, "step": 14829 }, { "epoch": 0.6216530605828784, "grad_norm": 1.8310211896896362, "learning_rate": 3.3068149965802498e-06, "loss": 0.4712, "step": 14830 }, { "epoch": 0.6216949791978873, "grad_norm": 1.844882607460022, "learning_rate": 3.306176268457304e-06, "loss": 0.559, "step": 14831 }, { "epoch": 0.6217368978128963, "grad_norm": 1.7363553047180176, "learning_rate": 3.3055375715576555e-06, "loss": 0.4421, "step": 14832 }, { "epoch": 0.6217788164279052, "grad_norm": 1.927266240119934, "learning_rate": 3.3048989058930803e-06, "loss": 0.4509, "step": 14833 }, { "epoch": 0.6218207350429141, "grad_norm": 2.1833512783050537, "learning_rate": 3.3042602714753492e-06, "loss": 0.4931, "step": 14834 }, { "epoch": 0.6218626536579231, "grad_norm": 2.0197019577026367, "learning_rate": 3.3036216683162366e-06, "loss": 0.5124, "step": 14835 }, { "epoch": 0.621904572272932, "grad_norm": 1.4599888324737549, "learning_rate": 3.3029830964275145e-06, "loss": 0.4818, "step": 14836 }, { "epoch": 0.6219464908879411, "grad_norm": 1.7216007709503174, "learning_rate": 3.30234455582095e-06, "loss": 0.4919, "step": 14837 }, { "epoch": 0.62198840950295, "grad_norm": 1.645950436592102, "learning_rate": 3.3017060465083174e-06, "loss": 0.4602, "step": 14838 }, { "epoch": 0.622030328117959, "grad_norm": 1.7486631870269775, "learning_rate": 3.301067568501388e-06, "loss": 0.4666, "step": 14839 }, { "epoch": 0.6220722467329679, "grad_norm": 1.6494438648223877, "learning_rate": 3.3004291218119288e-06, "loss": 0.4424, "step": 14840 }, { "epoch": 0.6221141653479769, "grad_norm": 1.771181583404541, "learning_rate": 3.2997907064517075e-06, "loss": 0.4657, "step": 14841 }, { "epoch": 0.6221560839629858, "grad_norm": 1.7158575057983398, "learning_rate": 3.299152322432496e-06, "loss": 0.4695, "step": 14842 }, { "epoch": 0.6221980025779948, "grad_norm": 1.824041485786438, "learning_rate": 3.2985139697660583e-06, "loss": 0.4376, "step": 14843 }, { "epoch": 0.6222399211930038, "grad_norm": 1.9115376472473145, "learning_rate": 3.297875648464164e-06, "loss": 0.4888, "step": 14844 }, { "epoch": 0.6222818398080128, "grad_norm": 1.9321413040161133, "learning_rate": 3.297237358538581e-06, "loss": 0.4985, "step": 14845 }, { "epoch": 0.6223237584230217, "grad_norm": 1.930191993713379, "learning_rate": 3.29659910000107e-06, "loss": 0.493, "step": 14846 }, { "epoch": 0.6223656770380307, "grad_norm": 1.9591628313064575, "learning_rate": 3.295960872863401e-06, "loss": 0.4935, "step": 14847 }, { "epoch": 0.6224075956530396, "grad_norm": 1.6618077754974365, "learning_rate": 3.295322677137338e-06, "loss": 0.4996, "step": 14848 }, { "epoch": 0.6224495142680486, "grad_norm": 2.1466312408447266, "learning_rate": 3.294684512834645e-06, "loss": 0.4815, "step": 14849 }, { "epoch": 0.6224914328830575, "grad_norm": 1.8193024396896362, "learning_rate": 3.294046379967084e-06, "loss": 0.4998, "step": 14850 }, { "epoch": 0.6225333514980665, "grad_norm": 1.9032217264175415, "learning_rate": 3.2934082785464216e-06, "loss": 0.4894, "step": 14851 }, { "epoch": 0.6225752701130755, "grad_norm": 1.764507532119751, "learning_rate": 3.2927702085844167e-06, "loss": 0.4591, "step": 14852 }, { "epoch": 0.6226171887280845, "grad_norm": 1.7202752828598022, "learning_rate": 3.2921321700928333e-06, "loss": 0.5275, "step": 14853 }, { "epoch": 0.6226591073430934, "grad_norm": 1.8668019771575928, "learning_rate": 3.2914941630834334e-06, "loss": 0.5389, "step": 14854 }, { "epoch": 0.6227010259581024, "grad_norm": 1.7879434823989868, "learning_rate": 3.290856187567975e-06, "loss": 0.4753, "step": 14855 }, { "epoch": 0.6227429445731113, "grad_norm": 2.9261975288391113, "learning_rate": 3.2902182435582198e-06, "loss": 0.4653, "step": 14856 }, { "epoch": 0.6227848631881203, "grad_norm": 1.6256006956100464, "learning_rate": 3.289580331065929e-06, "loss": 0.494, "step": 14857 }, { "epoch": 0.6228267818031292, "grad_norm": 1.7122368812561035, "learning_rate": 3.28894245010286e-06, "loss": 0.5032, "step": 14858 }, { "epoch": 0.6228687004181381, "grad_norm": 2.18571400642395, "learning_rate": 3.2883046006807706e-06, "loss": 0.4545, "step": 14859 }, { "epoch": 0.6229106190331472, "grad_norm": 1.679681420326233, "learning_rate": 3.2876667828114206e-06, "loss": 0.4448, "step": 14860 }, { "epoch": 0.6229525376481561, "grad_norm": 1.7671996355056763, "learning_rate": 3.287028996506565e-06, "loss": 0.472, "step": 14861 }, { "epoch": 0.6229944562631651, "grad_norm": 1.5490325689315796, "learning_rate": 3.2863912417779624e-06, "loss": 0.4176, "step": 14862 }, { "epoch": 0.623036374878174, "grad_norm": 1.832641839981079, "learning_rate": 3.2857535186373695e-06, "loss": 0.4841, "step": 14863 }, { "epoch": 0.623078293493183, "grad_norm": 1.854887843132019, "learning_rate": 3.285115827096539e-06, "loss": 0.5084, "step": 14864 }, { "epoch": 0.6231202121081919, "grad_norm": 1.97562837600708, "learning_rate": 3.2844781671672265e-06, "loss": 0.5612, "step": 14865 }, { "epoch": 0.6231621307232009, "grad_norm": 2.073779821395874, "learning_rate": 3.2838405388611895e-06, "loss": 0.5309, "step": 14866 }, { "epoch": 0.6232040493382098, "grad_norm": 1.71927011013031, "learning_rate": 3.283202942190179e-06, "loss": 0.4648, "step": 14867 }, { "epoch": 0.6232459679532188, "grad_norm": 1.6857757568359375, "learning_rate": 3.2825653771659484e-06, "loss": 0.4622, "step": 14868 }, { "epoch": 0.6232878865682278, "grad_norm": 1.8363401889801025, "learning_rate": 3.2819278438002523e-06, "loss": 0.4796, "step": 14869 }, { "epoch": 0.6233298051832368, "grad_norm": 1.8303059339523315, "learning_rate": 3.2812903421048387e-06, "loss": 0.5389, "step": 14870 }, { "epoch": 0.6233717237982457, "grad_norm": 1.7602012157440186, "learning_rate": 3.2806528720914634e-06, "loss": 0.5022, "step": 14871 }, { "epoch": 0.6234136424132547, "grad_norm": 1.6675078868865967, "learning_rate": 3.280015433771876e-06, "loss": 0.5004, "step": 14872 }, { "epoch": 0.6234555610282636, "grad_norm": 1.730658769607544, "learning_rate": 3.279378027157824e-06, "loss": 0.5381, "step": 14873 }, { "epoch": 0.6234974796432726, "grad_norm": 1.813947081565857, "learning_rate": 3.27874065226106e-06, "loss": 0.5216, "step": 14874 }, { "epoch": 0.6235393982582815, "grad_norm": 2.3363068103790283, "learning_rate": 3.2781033090933338e-06, "loss": 0.4719, "step": 14875 }, { "epoch": 0.6235813168732905, "grad_norm": 1.5298372507095337, "learning_rate": 3.2774659976663923e-06, "loss": 0.4984, "step": 14876 }, { "epoch": 0.6236232354882995, "grad_norm": 1.615425944328308, "learning_rate": 3.2768287179919827e-06, "loss": 0.5019, "step": 14877 }, { "epoch": 0.6236651541033085, "grad_norm": 1.6812233924865723, "learning_rate": 3.2761914700818554e-06, "loss": 0.4893, "step": 14878 }, { "epoch": 0.6237070727183174, "grad_norm": 1.5675877332687378, "learning_rate": 3.275554253947753e-06, "loss": 0.4928, "step": 14879 }, { "epoch": 0.6237489913333264, "grad_norm": 1.767248272895813, "learning_rate": 3.2749170696014253e-06, "loss": 0.5014, "step": 14880 }, { "epoch": 0.6237909099483353, "grad_norm": 1.869977593421936, "learning_rate": 3.2742799170546167e-06, "loss": 0.4821, "step": 14881 }, { "epoch": 0.6238328285633443, "grad_norm": 2.209545135498047, "learning_rate": 3.273642796319072e-06, "loss": 0.5027, "step": 14882 }, { "epoch": 0.6238747471783532, "grad_norm": 1.7924507856369019, "learning_rate": 3.273005707406534e-06, "loss": 0.5206, "step": 14883 }, { "epoch": 0.6239166657933621, "grad_norm": 1.713914394378662, "learning_rate": 3.272368650328751e-06, "loss": 0.5193, "step": 14884 }, { "epoch": 0.6239585844083712, "grad_norm": 1.7218254804611206, "learning_rate": 3.271731625097462e-06, "loss": 0.4795, "step": 14885 }, { "epoch": 0.6240005030233801, "grad_norm": 1.822386622428894, "learning_rate": 3.271094631724411e-06, "loss": 0.4693, "step": 14886 }, { "epoch": 0.6240424216383891, "grad_norm": 1.7876365184783936, "learning_rate": 3.270457670221342e-06, "loss": 0.4893, "step": 14887 }, { "epoch": 0.624084340253398, "grad_norm": 2.0513360500335693, "learning_rate": 3.2698207405999927e-06, "loss": 0.49, "step": 14888 }, { "epoch": 0.624126258868407, "grad_norm": 1.7298583984375, "learning_rate": 3.269183842872107e-06, "loss": 0.4902, "step": 14889 }, { "epoch": 0.6241681774834159, "grad_norm": 1.7011586427688599, "learning_rate": 3.268546977049425e-06, "loss": 0.4896, "step": 14890 }, { "epoch": 0.6242100960984249, "grad_norm": 1.5333045721054077, "learning_rate": 3.2679101431436853e-06, "loss": 0.5331, "step": 14891 }, { "epoch": 0.6242520147134338, "grad_norm": 2.633633613586426, "learning_rate": 3.267273341166627e-06, "loss": 0.4427, "step": 14892 }, { "epoch": 0.6242939333284429, "grad_norm": 1.9955819845199585, "learning_rate": 3.2666365711299907e-06, "loss": 0.4887, "step": 14893 }, { "epoch": 0.6243358519434518, "grad_norm": 1.6589055061340332, "learning_rate": 3.265999833045511e-06, "loss": 0.4793, "step": 14894 }, { "epoch": 0.6243777705584608, "grad_norm": 1.885614037513733, "learning_rate": 3.2653631269249286e-06, "loss": 0.5028, "step": 14895 }, { "epoch": 0.6244196891734697, "grad_norm": 1.9052084684371948, "learning_rate": 3.2647264527799798e-06, "loss": 0.4972, "step": 14896 }, { "epoch": 0.6244616077884787, "grad_norm": 2.1788322925567627, "learning_rate": 3.264089810622398e-06, "loss": 0.4488, "step": 14897 }, { "epoch": 0.6245035264034876, "grad_norm": 2.31736159324646, "learning_rate": 3.26345320046392e-06, "loss": 0.4771, "step": 14898 }, { "epoch": 0.6245454450184966, "grad_norm": 1.5567784309387207, "learning_rate": 3.2628166223162837e-06, "loss": 0.4506, "step": 14899 }, { "epoch": 0.6245873636335055, "grad_norm": 1.8324605226516724, "learning_rate": 3.2621800761912215e-06, "loss": 0.4702, "step": 14900 }, { "epoch": 0.6246292822485146, "grad_norm": 1.9567444324493408, "learning_rate": 3.2615435621004647e-06, "loss": 0.4675, "step": 14901 }, { "epoch": 0.6246712008635235, "grad_norm": 1.7639037370681763, "learning_rate": 3.260907080055752e-06, "loss": 0.4797, "step": 14902 }, { "epoch": 0.6247131194785325, "grad_norm": 2.0240395069122314, "learning_rate": 3.2602706300688103e-06, "loss": 0.4881, "step": 14903 }, { "epoch": 0.6247550380935414, "grad_norm": 1.8547637462615967, "learning_rate": 3.259634212151376e-06, "loss": 0.5092, "step": 14904 }, { "epoch": 0.6247969567085504, "grad_norm": 1.7486505508422852, "learning_rate": 3.25899782631518e-06, "loss": 0.5012, "step": 14905 }, { "epoch": 0.6248388753235593, "grad_norm": 1.807291030883789, "learning_rate": 3.2583614725719497e-06, "loss": 0.4777, "step": 14906 }, { "epoch": 0.6248807939385683, "grad_norm": 2.163968563079834, "learning_rate": 3.2577251509334207e-06, "loss": 0.4726, "step": 14907 }, { "epoch": 0.6249227125535772, "grad_norm": 1.787177562713623, "learning_rate": 3.2570888614113173e-06, "loss": 0.4972, "step": 14908 }, { "epoch": 0.6249646311685861, "grad_norm": 2.2176594734191895, "learning_rate": 3.2564526040173723e-06, "loss": 0.4645, "step": 14909 }, { "epoch": 0.6250065497835952, "grad_norm": 1.6922063827514648, "learning_rate": 3.2558163787633145e-06, "loss": 0.5156, "step": 14910 }, { "epoch": 0.6250484683986041, "grad_norm": 1.610294222831726, "learning_rate": 3.2551801856608676e-06, "loss": 0.5047, "step": 14911 }, { "epoch": 0.6250903870136131, "grad_norm": 1.623203158378601, "learning_rate": 3.254544024721763e-06, "loss": 0.5037, "step": 14912 }, { "epoch": 0.625132305628622, "grad_norm": 1.7367515563964844, "learning_rate": 3.2539078959577277e-06, "loss": 0.5046, "step": 14913 }, { "epoch": 0.625174224243631, "grad_norm": 1.5437339544296265, "learning_rate": 3.2532717993804852e-06, "loss": 0.4945, "step": 14914 }, { "epoch": 0.6252161428586399, "grad_norm": 1.5138757228851318, "learning_rate": 3.252635735001762e-06, "loss": 0.4721, "step": 14915 }, { "epoch": 0.6252580614736489, "grad_norm": 1.6766210794448853, "learning_rate": 3.251999702833285e-06, "loss": 0.5393, "step": 14916 }, { "epoch": 0.6252999800886578, "grad_norm": 1.8890798091888428, "learning_rate": 3.2513637028867746e-06, "loss": 0.5178, "step": 14917 }, { "epoch": 0.6253418987036669, "grad_norm": 1.6445019245147705, "learning_rate": 3.2507277351739585e-06, "loss": 0.4438, "step": 14918 }, { "epoch": 0.6253838173186758, "grad_norm": 2.2232797145843506, "learning_rate": 3.250091799706558e-06, "loss": 0.5351, "step": 14919 }, { "epoch": 0.6254257359336848, "grad_norm": 2.381117820739746, "learning_rate": 3.2494558964962957e-06, "loss": 0.5424, "step": 14920 }, { "epoch": 0.6254676545486937, "grad_norm": 2.2349565029144287, "learning_rate": 3.248820025554893e-06, "loss": 0.5093, "step": 14921 }, { "epoch": 0.6255095731637027, "grad_norm": 1.9470362663269043, "learning_rate": 3.2481841868940743e-06, "loss": 0.475, "step": 14922 }, { "epoch": 0.6255514917787116, "grad_norm": 2.125880718231201, "learning_rate": 3.247548380525558e-06, "loss": 0.4374, "step": 14923 }, { "epoch": 0.6255934103937206, "grad_norm": 1.6125084161758423, "learning_rate": 3.246912606461063e-06, "loss": 0.5178, "step": 14924 }, { "epoch": 0.6256353290087295, "grad_norm": 1.7485156059265137, "learning_rate": 3.246276864712312e-06, "loss": 0.51, "step": 14925 }, { "epoch": 0.6256772476237386, "grad_norm": 1.7914481163024902, "learning_rate": 3.2456411552910217e-06, "loss": 0.5176, "step": 14926 }, { "epoch": 0.6257191662387475, "grad_norm": 1.8334522247314453, "learning_rate": 3.2450054782089114e-06, "loss": 0.4919, "step": 14927 }, { "epoch": 0.6257610848537565, "grad_norm": 2.3499410152435303, "learning_rate": 3.2443698334776994e-06, "loss": 0.4475, "step": 14928 }, { "epoch": 0.6258030034687654, "grad_norm": 1.7258387804031372, "learning_rate": 3.243734221109103e-06, "loss": 0.5047, "step": 14929 }, { "epoch": 0.6258449220837744, "grad_norm": 2.1743462085723877, "learning_rate": 3.2430986411148356e-06, "loss": 0.4279, "step": 14930 }, { "epoch": 0.6258868406987833, "grad_norm": 1.5211552381515503, "learning_rate": 3.2424630935066194e-06, "loss": 0.5225, "step": 14931 }, { "epoch": 0.6259287593137923, "grad_norm": 1.4470853805541992, "learning_rate": 3.2418275782961637e-06, "loss": 0.4839, "step": 14932 }, { "epoch": 0.6259706779288012, "grad_norm": 1.7169921398162842, "learning_rate": 3.241192095495187e-06, "loss": 0.4879, "step": 14933 }, { "epoch": 0.6260125965438101, "grad_norm": 1.9320268630981445, "learning_rate": 3.2405566451154037e-06, "loss": 0.4807, "step": 14934 }, { "epoch": 0.6260545151588192, "grad_norm": 2.092188596725464, "learning_rate": 3.239921227168524e-06, "loss": 0.5169, "step": 14935 }, { "epoch": 0.6260964337738281, "grad_norm": 1.78296959400177, "learning_rate": 3.239285841666264e-06, "loss": 0.5562, "step": 14936 }, { "epoch": 0.6261383523888371, "grad_norm": 3.3527441024780273, "learning_rate": 3.238650488620337e-06, "loss": 0.5173, "step": 14937 }, { "epoch": 0.626180271003846, "grad_norm": 1.8573259115219116, "learning_rate": 3.238015168042452e-06, "loss": 0.5086, "step": 14938 }, { "epoch": 0.626222189618855, "grad_norm": 1.7683833837509155, "learning_rate": 3.2373798799443207e-06, "loss": 0.4873, "step": 14939 }, { "epoch": 0.6262641082338639, "grad_norm": 1.7158714532852173, "learning_rate": 3.2367446243376565e-06, "loss": 0.4739, "step": 14940 }, { "epoch": 0.6263060268488729, "grad_norm": 1.8630344867706299, "learning_rate": 3.2361094012341656e-06, "loss": 0.4631, "step": 14941 }, { "epoch": 0.6263479454638818, "grad_norm": 1.6372560262680054, "learning_rate": 3.2354742106455607e-06, "loss": 0.4614, "step": 14942 }, { "epoch": 0.6263898640788909, "grad_norm": 1.5367622375488281, "learning_rate": 3.2348390525835495e-06, "loss": 0.4748, "step": 14943 }, { "epoch": 0.6264317826938998, "grad_norm": 1.463765025138855, "learning_rate": 3.234203927059839e-06, "loss": 0.4804, "step": 14944 }, { "epoch": 0.6264737013089088, "grad_norm": 1.6696784496307373, "learning_rate": 3.233568834086137e-06, "loss": 0.5013, "step": 14945 }, { "epoch": 0.6265156199239177, "grad_norm": 1.9430577754974365, "learning_rate": 3.232933773674154e-06, "loss": 0.4835, "step": 14946 }, { "epoch": 0.6265575385389267, "grad_norm": 1.7450522184371948, "learning_rate": 3.2322987458355937e-06, "loss": 0.4604, "step": 14947 }, { "epoch": 0.6265994571539356, "grad_norm": 1.632967233657837, "learning_rate": 3.23166375058216e-06, "loss": 0.4637, "step": 14948 }, { "epoch": 0.6266413757689446, "grad_norm": 1.584488868713379, "learning_rate": 3.231028787925563e-06, "loss": 0.5321, "step": 14949 }, { "epoch": 0.6266832943839535, "grad_norm": 1.9772204160690308, "learning_rate": 3.230393857877503e-06, "loss": 0.4631, "step": 14950 }, { "epoch": 0.6267252129989626, "grad_norm": 1.8486937284469604, "learning_rate": 3.229758960449687e-06, "loss": 0.443, "step": 14951 }, { "epoch": 0.6267671316139715, "grad_norm": 1.7387709617614746, "learning_rate": 3.229124095653817e-06, "loss": 0.511, "step": 14952 }, { "epoch": 0.6268090502289805, "grad_norm": 1.9672051668167114, "learning_rate": 3.2284892635015953e-06, "loss": 0.4649, "step": 14953 }, { "epoch": 0.6268509688439894, "grad_norm": 1.6412535905838013, "learning_rate": 3.2278544640047248e-06, "loss": 0.4873, "step": 14954 }, { "epoch": 0.6268928874589984, "grad_norm": 1.8614208698272705, "learning_rate": 3.2272196971749093e-06, "loss": 0.5025, "step": 14955 }, { "epoch": 0.6269348060740073, "grad_norm": 1.7131098508834839, "learning_rate": 3.2265849630238464e-06, "loss": 0.4895, "step": 14956 }, { "epoch": 0.6269767246890163, "grad_norm": 1.769938349723816, "learning_rate": 3.2259502615632375e-06, "loss": 0.4988, "step": 14957 }, { "epoch": 0.6270186433040252, "grad_norm": 1.9295611381530762, "learning_rate": 3.225315592804785e-06, "loss": 0.498, "step": 14958 }, { "epoch": 0.6270605619190341, "grad_norm": 1.9319937229156494, "learning_rate": 3.2246809567601843e-06, "loss": 0.4931, "step": 14959 }, { "epoch": 0.6271024805340432, "grad_norm": 1.7109378576278687, "learning_rate": 3.224046353441137e-06, "loss": 0.463, "step": 14960 }, { "epoch": 0.6271443991490521, "grad_norm": 1.980225682258606, "learning_rate": 3.2234117828593404e-06, "loss": 0.5214, "step": 14961 }, { "epoch": 0.6271863177640611, "grad_norm": 2.0014822483062744, "learning_rate": 3.2227772450264895e-06, "loss": 0.4854, "step": 14962 }, { "epoch": 0.62722823637907, "grad_norm": 1.994009256362915, "learning_rate": 3.2221427399542836e-06, "loss": 0.517, "step": 14963 }, { "epoch": 0.627270154994079, "grad_norm": 1.9639617204666138, "learning_rate": 3.22150826765442e-06, "loss": 0.4846, "step": 14964 }, { "epoch": 0.6273120736090879, "grad_norm": 2.1493937969207764, "learning_rate": 3.2208738281385922e-06, "loss": 0.5004, "step": 14965 }, { "epoch": 0.6273539922240969, "grad_norm": 1.681551456451416, "learning_rate": 3.220239421418495e-06, "loss": 0.4456, "step": 14966 }, { "epoch": 0.6273959108391058, "grad_norm": 1.584591269493103, "learning_rate": 3.219605047505825e-06, "loss": 0.4988, "step": 14967 }, { "epoch": 0.6274378294541149, "grad_norm": 1.732503056526184, "learning_rate": 3.2189707064122732e-06, "loss": 0.467, "step": 14968 }, { "epoch": 0.6274797480691238, "grad_norm": 1.8745551109313965, "learning_rate": 3.2183363981495346e-06, "loss": 0.513, "step": 14969 }, { "epoch": 0.6275216666841328, "grad_norm": 1.6798253059387207, "learning_rate": 3.2177021227293013e-06, "loss": 0.5117, "step": 14970 }, { "epoch": 0.6275635852991417, "grad_norm": 2.5031509399414062, "learning_rate": 3.2170678801632665e-06, "loss": 0.5022, "step": 14971 }, { "epoch": 0.6276055039141507, "grad_norm": 1.7440177202224731, "learning_rate": 3.216433670463119e-06, "loss": 0.4993, "step": 14972 }, { "epoch": 0.6276474225291596, "grad_norm": 1.7178698778152466, "learning_rate": 3.215799493640553e-06, "loss": 0.4775, "step": 14973 }, { "epoch": 0.6276893411441686, "grad_norm": 2.1576061248779297, "learning_rate": 3.2151653497072544e-06, "loss": 0.511, "step": 14974 }, { "epoch": 0.6277312597591775, "grad_norm": 1.6204214096069336, "learning_rate": 3.214531238674916e-06, "loss": 0.4635, "step": 14975 }, { "epoch": 0.6277731783741866, "grad_norm": 1.6685179471969604, "learning_rate": 3.2138971605552277e-06, "loss": 0.44, "step": 14976 }, { "epoch": 0.6278150969891955, "grad_norm": 1.762355089187622, "learning_rate": 3.2132631153598737e-06, "loss": 0.4841, "step": 14977 }, { "epoch": 0.6278570156042045, "grad_norm": 1.5786274671554565, "learning_rate": 3.212629103100545e-06, "loss": 0.4752, "step": 14978 }, { "epoch": 0.6278989342192134, "grad_norm": 1.5516777038574219, "learning_rate": 3.211995123788928e-06, "loss": 0.4957, "step": 14979 }, { "epoch": 0.6279408528342224, "grad_norm": 1.7641264200210571, "learning_rate": 3.2113611774367085e-06, "loss": 0.4861, "step": 14980 }, { "epoch": 0.6279827714492313, "grad_norm": 2.2697997093200684, "learning_rate": 3.2107272640555724e-06, "loss": 0.503, "step": 14981 }, { "epoch": 0.6280246900642403, "grad_norm": 2.0499019622802734, "learning_rate": 3.210093383657208e-06, "loss": 0.4616, "step": 14982 }, { "epoch": 0.6280666086792492, "grad_norm": 2.315021514892578, "learning_rate": 3.2094595362532947e-06, "loss": 0.4896, "step": 14983 }, { "epoch": 0.6281085272942581, "grad_norm": 1.746817708015442, "learning_rate": 3.208825721855522e-06, "loss": 0.5222, "step": 14984 }, { "epoch": 0.6281504459092672, "grad_norm": 1.6804577112197876, "learning_rate": 3.20819194047557e-06, "loss": 0.4481, "step": 14985 }, { "epoch": 0.6281923645242761, "grad_norm": 1.5356556177139282, "learning_rate": 3.2075581921251216e-06, "loss": 0.4861, "step": 14986 }, { "epoch": 0.6282342831392851, "grad_norm": 2.403608798980713, "learning_rate": 3.2069244768158613e-06, "loss": 0.4785, "step": 14987 }, { "epoch": 0.628276201754294, "grad_norm": 1.9388707876205444, "learning_rate": 3.2062907945594678e-06, "loss": 0.4711, "step": 14988 }, { "epoch": 0.628318120369303, "grad_norm": 1.8242480754852295, "learning_rate": 3.205657145367625e-06, "loss": 0.4891, "step": 14989 }, { "epoch": 0.6283600389843119, "grad_norm": 1.6550551652908325, "learning_rate": 3.2050235292520126e-06, "loss": 0.459, "step": 14990 }, { "epoch": 0.6284019575993209, "grad_norm": 2.0208828449249268, "learning_rate": 3.2043899462243082e-06, "loss": 0.4865, "step": 14991 }, { "epoch": 0.6284438762143298, "grad_norm": 1.9020774364471436, "learning_rate": 3.2037563962961925e-06, "loss": 0.5091, "step": 14992 }, { "epoch": 0.6284857948293389, "grad_norm": 1.8980627059936523, "learning_rate": 3.2031228794793473e-06, "loss": 0.4789, "step": 14993 }, { "epoch": 0.6285277134443478, "grad_norm": 2.152446746826172, "learning_rate": 3.2024893957854465e-06, "loss": 0.4785, "step": 14994 }, { "epoch": 0.6285696320593568, "grad_norm": 1.6896867752075195, "learning_rate": 3.2018559452261667e-06, "loss": 0.5098, "step": 14995 }, { "epoch": 0.6286115506743657, "grad_norm": 1.795357346534729, "learning_rate": 3.2012225278131894e-06, "loss": 0.4673, "step": 14996 }, { "epoch": 0.6286534692893747, "grad_norm": 1.9354480504989624, "learning_rate": 3.200589143558186e-06, "loss": 0.5562, "step": 14997 }, { "epoch": 0.6286953879043836, "grad_norm": 1.761894702911377, "learning_rate": 3.1999557924728354e-06, "loss": 0.5455, "step": 14998 }, { "epoch": 0.6287373065193926, "grad_norm": 1.6296298503875732, "learning_rate": 3.1993224745688123e-06, "loss": 0.4588, "step": 14999 }, { "epoch": 0.6287792251344015, "grad_norm": 1.6583060026168823, "learning_rate": 3.1986891898577878e-06, "loss": 0.5254, "step": 15000 }, { "epoch": 0.6288211437494106, "grad_norm": 1.8596436977386475, "learning_rate": 3.1980559383514377e-06, "loss": 0.4595, "step": 15001 }, { "epoch": 0.6288630623644195, "grad_norm": 1.6822834014892578, "learning_rate": 3.1974227200614373e-06, "loss": 0.4518, "step": 15002 }, { "epoch": 0.6289049809794285, "grad_norm": 1.7298215627670288, "learning_rate": 3.196789534999456e-06, "loss": 0.4918, "step": 15003 }, { "epoch": 0.6289468995944374, "grad_norm": 2.1730377674102783, "learning_rate": 3.1961563831771657e-06, "loss": 0.4932, "step": 15004 }, { "epoch": 0.6289888182094464, "grad_norm": 1.6442091464996338, "learning_rate": 3.195523264606241e-06, "loss": 0.4772, "step": 15005 }, { "epoch": 0.6290307368244553, "grad_norm": 1.7724652290344238, "learning_rate": 3.1948901792983476e-06, "loss": 0.4989, "step": 15006 }, { "epoch": 0.6290726554394643, "grad_norm": 1.6193746328353882, "learning_rate": 3.19425712726516e-06, "loss": 0.5269, "step": 15007 }, { "epoch": 0.6291145740544732, "grad_norm": 1.6827796697616577, "learning_rate": 3.1936241085183473e-06, "loss": 0.4732, "step": 15008 }, { "epoch": 0.6291564926694821, "grad_norm": 1.8163048028945923, "learning_rate": 3.1929911230695743e-06, "loss": 0.5183, "step": 15009 }, { "epoch": 0.6291984112844912, "grad_norm": 1.52720308303833, "learning_rate": 3.1923581709305117e-06, "loss": 0.4683, "step": 15010 }, { "epoch": 0.6292403298995001, "grad_norm": 1.7501722574234009, "learning_rate": 3.1917252521128294e-06, "loss": 0.4661, "step": 15011 }, { "epoch": 0.6292822485145091, "grad_norm": 1.9395391941070557, "learning_rate": 3.1910923666281913e-06, "loss": 0.5075, "step": 15012 }, { "epoch": 0.629324167129518, "grad_norm": 1.9405019283294678, "learning_rate": 3.1904595144882633e-06, "loss": 0.4873, "step": 15013 }, { "epoch": 0.629366085744527, "grad_norm": 2.161517858505249, "learning_rate": 3.189826695704715e-06, "loss": 0.4782, "step": 15014 }, { "epoch": 0.6294080043595359, "grad_norm": 2.8091275691986084, "learning_rate": 3.1891939102892067e-06, "loss": 0.4621, "step": 15015 }, { "epoch": 0.6294499229745449, "grad_norm": 1.6654391288757324, "learning_rate": 3.188561158253407e-06, "loss": 0.5153, "step": 15016 }, { "epoch": 0.6294918415895538, "grad_norm": 1.6415914297103882, "learning_rate": 3.1879284396089772e-06, "loss": 0.4835, "step": 15017 }, { "epoch": 0.6295337602045629, "grad_norm": 1.8307974338531494, "learning_rate": 3.187295754367582e-06, "loss": 0.4362, "step": 15018 }, { "epoch": 0.6295756788195718, "grad_norm": 1.6142932176589966, "learning_rate": 3.1866631025408827e-06, "loss": 0.4862, "step": 15019 }, { "epoch": 0.6296175974345808, "grad_norm": 2.2245407104492188, "learning_rate": 3.1860304841405445e-06, "loss": 0.4859, "step": 15020 }, { "epoch": 0.6296595160495897, "grad_norm": 1.558793067932129, "learning_rate": 3.1853978991782235e-06, "loss": 0.4814, "step": 15021 }, { "epoch": 0.6297014346645987, "grad_norm": 2.34055757522583, "learning_rate": 3.1847653476655855e-06, "loss": 0.5319, "step": 15022 }, { "epoch": 0.6297433532796076, "grad_norm": 1.6133079528808594, "learning_rate": 3.184132829614289e-06, "loss": 0.4134, "step": 15023 }, { "epoch": 0.6297852718946166, "grad_norm": 2.077028751373291, "learning_rate": 3.1835003450359925e-06, "loss": 0.518, "step": 15024 }, { "epoch": 0.6298271905096255, "grad_norm": 13.370808601379395, "learning_rate": 3.182867893942355e-06, "loss": 0.4738, "step": 15025 }, { "epoch": 0.6298691091246346, "grad_norm": 1.9995156526565552, "learning_rate": 3.182235476345039e-06, "loss": 0.4844, "step": 15026 }, { "epoch": 0.6299110277396435, "grad_norm": 1.6057382822036743, "learning_rate": 3.181603092255697e-06, "loss": 0.5232, "step": 15027 }, { "epoch": 0.6299529463546525, "grad_norm": 2.6388325691223145, "learning_rate": 3.1809707416859868e-06, "loss": 0.4535, "step": 15028 }, { "epoch": 0.6299948649696614, "grad_norm": 1.8767393827438354, "learning_rate": 3.1803384246475693e-06, "loss": 0.4916, "step": 15029 }, { "epoch": 0.6300367835846704, "grad_norm": 1.6942296028137207, "learning_rate": 3.1797061411520945e-06, "loss": 0.5012, "step": 15030 }, { "epoch": 0.6300787021996793, "grad_norm": 1.980770468711853, "learning_rate": 3.1790738912112217e-06, "loss": 0.473, "step": 15031 }, { "epoch": 0.6301206208146883, "grad_norm": 1.7950057983398438, "learning_rate": 3.178441674836606e-06, "loss": 0.4861, "step": 15032 }, { "epoch": 0.6301625394296972, "grad_norm": 1.8357715606689453, "learning_rate": 3.1778094920398974e-06, "loss": 0.512, "step": 15033 }, { "epoch": 0.6302044580447062, "grad_norm": 1.512966275215149, "learning_rate": 3.177177342832751e-06, "loss": 0.4306, "step": 15034 }, { "epoch": 0.6302463766597152, "grad_norm": 1.9695966243743896, "learning_rate": 3.1765452272268234e-06, "loss": 0.5436, "step": 15035 }, { "epoch": 0.6302882952747241, "grad_norm": 1.8470255136489868, "learning_rate": 3.175913145233762e-06, "loss": 0.4136, "step": 15036 }, { "epoch": 0.6303302138897331, "grad_norm": 2.1347250938415527, "learning_rate": 3.175281096865219e-06, "loss": 0.4463, "step": 15037 }, { "epoch": 0.630372132504742, "grad_norm": 2.3428564071655273, "learning_rate": 3.174649082132849e-06, "loss": 0.5329, "step": 15038 }, { "epoch": 0.630414051119751, "grad_norm": 1.831871509552002, "learning_rate": 3.1740171010482975e-06, "loss": 0.4722, "step": 15039 }, { "epoch": 0.6304559697347599, "grad_norm": 1.8118523359298706, "learning_rate": 3.173385153623217e-06, "loss": 0.4953, "step": 15040 }, { "epoch": 0.6304978883497689, "grad_norm": 1.5431809425354004, "learning_rate": 3.1727532398692572e-06, "loss": 0.4349, "step": 15041 }, { "epoch": 0.6305398069647778, "grad_norm": 1.6762988567352295, "learning_rate": 3.172121359798063e-06, "loss": 0.4665, "step": 15042 }, { "epoch": 0.6305817255797869, "grad_norm": 2.0177602767944336, "learning_rate": 3.1714895134212854e-06, "loss": 0.4628, "step": 15043 }, { "epoch": 0.6306236441947958, "grad_norm": 2.243267059326172, "learning_rate": 3.170857700750572e-06, "loss": 0.5628, "step": 15044 }, { "epoch": 0.6306655628098048, "grad_norm": 1.7979861497879028, "learning_rate": 3.170225921797567e-06, "loss": 0.4442, "step": 15045 }, { "epoch": 0.6307074814248137, "grad_norm": 1.577256441116333, "learning_rate": 3.1695941765739175e-06, "loss": 0.459, "step": 15046 }, { "epoch": 0.6307494000398227, "grad_norm": 2.413647413253784, "learning_rate": 3.1689624650912705e-06, "loss": 0.5092, "step": 15047 }, { "epoch": 0.6307913186548316, "grad_norm": 1.8659257888793945, "learning_rate": 3.168330787361267e-06, "loss": 0.5289, "step": 15048 }, { "epoch": 0.6308332372698406, "grad_norm": 1.8949989080429077, "learning_rate": 3.167699143395554e-06, "loss": 0.4768, "step": 15049 }, { "epoch": 0.6308751558848495, "grad_norm": 2.103358030319214, "learning_rate": 3.1670675332057752e-06, "loss": 0.4854, "step": 15050 }, { "epoch": 0.6309170744998586, "grad_norm": 1.5315805673599243, "learning_rate": 3.1664359568035707e-06, "loss": 0.4164, "step": 15051 }, { "epoch": 0.6309589931148675, "grad_norm": 1.773581862449646, "learning_rate": 3.165804414200585e-06, "loss": 0.5216, "step": 15052 }, { "epoch": 0.6310009117298765, "grad_norm": 2.2993857860565186, "learning_rate": 3.16517290540846e-06, "loss": 0.4829, "step": 15053 }, { "epoch": 0.6310428303448854, "grad_norm": 1.6834745407104492, "learning_rate": 3.164541430438836e-06, "loss": 0.5166, "step": 15054 }, { "epoch": 0.6310847489598944, "grad_norm": 1.615968942642212, "learning_rate": 3.163909989303352e-06, "loss": 0.4526, "step": 15055 }, { "epoch": 0.6311266675749033, "grad_norm": 1.7397125959396362, "learning_rate": 3.1632785820136515e-06, "loss": 0.4855, "step": 15056 }, { "epoch": 0.6311685861899123, "grad_norm": 1.8778985738754272, "learning_rate": 3.162647208581368e-06, "loss": 0.4821, "step": 15057 }, { "epoch": 0.6312105048049212, "grad_norm": 1.6596879959106445, "learning_rate": 3.162015869018146e-06, "loss": 0.4992, "step": 15058 }, { "epoch": 0.6312524234199302, "grad_norm": 1.675534725189209, "learning_rate": 3.161384563335619e-06, "loss": 0.499, "step": 15059 }, { "epoch": 0.6312943420349392, "grad_norm": 1.8523823022842407, "learning_rate": 3.1607532915454252e-06, "loss": 0.5382, "step": 15060 }, { "epoch": 0.6313362606499481, "grad_norm": 2.537548303604126, "learning_rate": 3.1601220536592037e-06, "loss": 0.5205, "step": 15061 }, { "epoch": 0.6313781792649571, "grad_norm": 1.7782418727874756, "learning_rate": 3.1594908496885867e-06, "loss": 0.4532, "step": 15062 }, { "epoch": 0.631420097879966, "grad_norm": 1.8821686506271362, "learning_rate": 3.158859679645212e-06, "loss": 0.4684, "step": 15063 }, { "epoch": 0.631462016494975, "grad_norm": 1.740275502204895, "learning_rate": 3.1582285435407145e-06, "loss": 0.5141, "step": 15064 }, { "epoch": 0.6315039351099839, "grad_norm": 1.7505921125411987, "learning_rate": 3.1575974413867273e-06, "loss": 0.48, "step": 15065 }, { "epoch": 0.631545853724993, "grad_norm": 1.5990538597106934, "learning_rate": 3.156966373194883e-06, "loss": 0.4936, "step": 15066 }, { "epoch": 0.6315877723400019, "grad_norm": 1.8204532861709595, "learning_rate": 3.1563353389768177e-06, "loss": 0.5524, "step": 15067 }, { "epoch": 0.6316296909550109, "grad_norm": 1.6665477752685547, "learning_rate": 3.15570433874416e-06, "loss": 0.5063, "step": 15068 }, { "epoch": 0.6316716095700198, "grad_norm": 1.9000459909439087, "learning_rate": 3.155073372508544e-06, "loss": 0.4532, "step": 15069 }, { "epoch": 0.6317135281850288, "grad_norm": 1.863911509513855, "learning_rate": 3.1544424402816e-06, "loss": 0.4936, "step": 15070 }, { "epoch": 0.6317554468000377, "grad_norm": 1.4446203708648682, "learning_rate": 3.1538115420749573e-06, "loss": 0.469, "step": 15071 }, { "epoch": 0.6317973654150467, "grad_norm": 3.2456610202789307, "learning_rate": 3.153180677900246e-06, "loss": 0.5209, "step": 15072 }, { "epoch": 0.6318392840300556, "grad_norm": 1.8231239318847656, "learning_rate": 3.152549847769098e-06, "loss": 0.5141, "step": 15073 }, { "epoch": 0.6318812026450646, "grad_norm": 1.7547013759613037, "learning_rate": 3.1519190516931385e-06, "loss": 0.4914, "step": 15074 }, { "epoch": 0.6319231212600736, "grad_norm": 2.033534049987793, "learning_rate": 3.1512882896839956e-06, "loss": 0.568, "step": 15075 }, { "epoch": 0.6319650398750826, "grad_norm": 1.434867262840271, "learning_rate": 3.1506575617532984e-06, "loss": 0.4613, "step": 15076 }, { "epoch": 0.6320069584900915, "grad_norm": 1.8991541862487793, "learning_rate": 3.1500268679126717e-06, "loss": 0.5081, "step": 15077 }, { "epoch": 0.6320488771051005, "grad_norm": 1.698839783668518, "learning_rate": 3.1493962081737426e-06, "loss": 0.4899, "step": 15078 }, { "epoch": 0.6320907957201094, "grad_norm": 1.9036356210708618, "learning_rate": 3.1487655825481367e-06, "loss": 0.5041, "step": 15079 }, { "epoch": 0.6321327143351184, "grad_norm": 2.0392837524414062, "learning_rate": 3.1481349910474766e-06, "loss": 0.4992, "step": 15080 }, { "epoch": 0.6321746329501273, "grad_norm": 1.790602445602417, "learning_rate": 3.1475044336833873e-06, "loss": 0.4736, "step": 15081 }, { "epoch": 0.6322165515651363, "grad_norm": 1.6853067874908447, "learning_rate": 3.146873910467495e-06, "loss": 0.4809, "step": 15082 }, { "epoch": 0.6322584701801452, "grad_norm": 1.8103771209716797, "learning_rate": 3.1462434214114195e-06, "loss": 0.4911, "step": 15083 }, { "epoch": 0.6323003887951542, "grad_norm": 2.9699513912200928, "learning_rate": 3.1456129665267827e-06, "loss": 0.4465, "step": 15084 }, { "epoch": 0.6323423074101632, "grad_norm": 1.551753282546997, "learning_rate": 3.144982545825209e-06, "loss": 0.5017, "step": 15085 }, { "epoch": 0.6323842260251721, "grad_norm": 2.1478114128112793, "learning_rate": 3.1443521593183155e-06, "loss": 0.4589, "step": 15086 }, { "epoch": 0.6324261446401811, "grad_norm": 2.0494754314422607, "learning_rate": 3.143721807017726e-06, "loss": 0.483, "step": 15087 }, { "epoch": 0.63246806325519, "grad_norm": 2.088914632797241, "learning_rate": 3.1430914889350596e-06, "loss": 0.4941, "step": 15088 }, { "epoch": 0.632509981870199, "grad_norm": 1.6493068933486938, "learning_rate": 3.1424612050819324e-06, "loss": 0.4185, "step": 15089 }, { "epoch": 0.6325519004852079, "grad_norm": 1.5906460285186768, "learning_rate": 3.141830955469965e-06, "loss": 0.508, "step": 15090 }, { "epoch": 0.632593819100217, "grad_norm": 1.5549578666687012, "learning_rate": 3.141200740110777e-06, "loss": 0.51, "step": 15091 }, { "epoch": 0.6326357377152259, "grad_norm": 1.6133487224578857, "learning_rate": 3.1405705590159828e-06, "loss": 0.4041, "step": 15092 }, { "epoch": 0.6326776563302349, "grad_norm": 1.7861552238464355, "learning_rate": 3.139940412197199e-06, "loss": 0.4793, "step": 15093 }, { "epoch": 0.6327195749452438, "grad_norm": 1.8926085233688354, "learning_rate": 3.139310299666044e-06, "loss": 0.4806, "step": 15094 }, { "epoch": 0.6327614935602528, "grad_norm": 4.270185947418213, "learning_rate": 3.138680221434129e-06, "loss": 0.5293, "step": 15095 }, { "epoch": 0.6328034121752617, "grad_norm": 1.7095144987106323, "learning_rate": 3.138050177513072e-06, "loss": 0.4993, "step": 15096 }, { "epoch": 0.6328453307902707, "grad_norm": 1.9832146167755127, "learning_rate": 3.137420167914487e-06, "loss": 0.4585, "step": 15097 }, { "epoch": 0.6328872494052796, "grad_norm": 1.709026575088501, "learning_rate": 3.1367901926499845e-06, "loss": 0.4828, "step": 15098 }, { "epoch": 0.6329291680202886, "grad_norm": 3.657870292663574, "learning_rate": 3.1361602517311784e-06, "loss": 0.4644, "step": 15099 }, { "epoch": 0.6329710866352976, "grad_norm": 1.7376065254211426, "learning_rate": 3.1355303451696837e-06, "loss": 0.4368, "step": 15100 }, { "epoch": 0.6330130052503066, "grad_norm": 2.0115771293640137, "learning_rate": 3.134900472977109e-06, "loss": 0.5381, "step": 15101 }, { "epoch": 0.6330549238653155, "grad_norm": 1.8169045448303223, "learning_rate": 3.1342706351650644e-06, "loss": 0.4653, "step": 15102 }, { "epoch": 0.6330968424803245, "grad_norm": 2.0320966243743896, "learning_rate": 3.133640831745163e-06, "loss": 0.408, "step": 15103 }, { "epoch": 0.6331387610953334, "grad_norm": 1.7583988904953003, "learning_rate": 3.133011062729011e-06, "loss": 0.5356, "step": 15104 }, { "epoch": 0.6331806797103424, "grad_norm": 2.0100138187408447, "learning_rate": 3.13238132812822e-06, "loss": 0.4554, "step": 15105 }, { "epoch": 0.6332225983253513, "grad_norm": 1.6421478986740112, "learning_rate": 3.131751627954398e-06, "loss": 0.5394, "step": 15106 }, { "epoch": 0.6332645169403603, "grad_norm": 1.897538423538208, "learning_rate": 3.13112196221915e-06, "loss": 0.4648, "step": 15107 }, { "epoch": 0.6333064355553693, "grad_norm": 1.852393627166748, "learning_rate": 3.1304923309340858e-06, "loss": 0.4895, "step": 15108 }, { "epoch": 0.6333483541703782, "grad_norm": 1.7417733669281006, "learning_rate": 3.1298627341108123e-06, "loss": 0.4889, "step": 15109 }, { "epoch": 0.6333902727853872, "grad_norm": 1.7347317934036255, "learning_rate": 3.129233171760933e-06, "loss": 0.4358, "step": 15110 }, { "epoch": 0.6334321914003961, "grad_norm": 2.0274293422698975, "learning_rate": 3.128603643896053e-06, "loss": 0.5478, "step": 15111 }, { "epoch": 0.6334741100154051, "grad_norm": 1.5434504747390747, "learning_rate": 3.1279741505277794e-06, "loss": 0.4522, "step": 15112 }, { "epoch": 0.633516028630414, "grad_norm": 1.8578122854232788, "learning_rate": 3.1273446916677132e-06, "loss": 0.456, "step": 15113 }, { "epoch": 0.633557947245423, "grad_norm": 1.8373230695724487, "learning_rate": 3.126715267327458e-06, "loss": 0.4657, "step": 15114 }, { "epoch": 0.6335998658604319, "grad_norm": 1.571093201637268, "learning_rate": 3.1260858775186187e-06, "loss": 0.4411, "step": 15115 }, { "epoch": 0.633641784475441, "grad_norm": 1.6577651500701904, "learning_rate": 3.1254565222527953e-06, "loss": 0.4824, "step": 15116 }, { "epoch": 0.6336837030904499, "grad_norm": 1.5303890705108643, "learning_rate": 3.1248272015415882e-06, "loss": 0.4959, "step": 15117 }, { "epoch": 0.6337256217054589, "grad_norm": 1.7242964506149292, "learning_rate": 3.1241979153966014e-06, "loss": 0.492, "step": 15118 }, { "epoch": 0.6337675403204678, "grad_norm": 1.8851746320724487, "learning_rate": 3.1235686638294303e-06, "loss": 0.5289, "step": 15119 }, { "epoch": 0.6338094589354768, "grad_norm": 2.058337688446045, "learning_rate": 3.1229394468516784e-06, "loss": 0.5391, "step": 15120 }, { "epoch": 0.6338513775504857, "grad_norm": 1.6754056215286255, "learning_rate": 3.1223102644749438e-06, "loss": 0.4602, "step": 15121 }, { "epoch": 0.6338932961654947, "grad_norm": 2.265352964401245, "learning_rate": 3.121681116710822e-06, "loss": 0.5214, "step": 15122 }, { "epoch": 0.6339352147805036, "grad_norm": 1.8597584962844849, "learning_rate": 3.1210520035709113e-06, "loss": 0.5616, "step": 15123 }, { "epoch": 0.6339771333955126, "grad_norm": 1.7957179546356201, "learning_rate": 3.1204229250668117e-06, "loss": 0.5269, "step": 15124 }, { "epoch": 0.6340190520105216, "grad_norm": 1.5705479383468628, "learning_rate": 3.1197938812101157e-06, "loss": 0.45, "step": 15125 }, { "epoch": 0.6340609706255306, "grad_norm": 1.6560829877853394, "learning_rate": 3.1191648720124202e-06, "loss": 0.5213, "step": 15126 }, { "epoch": 0.6341028892405395, "grad_norm": 2.603205442428589, "learning_rate": 3.1185358974853206e-06, "loss": 0.5154, "step": 15127 }, { "epoch": 0.6341448078555485, "grad_norm": 1.9098091125488281, "learning_rate": 3.11790695764041e-06, "loss": 0.545, "step": 15128 }, { "epoch": 0.6341867264705574, "grad_norm": 1.9913798570632935, "learning_rate": 3.1172780524892833e-06, "loss": 0.5506, "step": 15129 }, { "epoch": 0.6342286450855664, "grad_norm": 1.8152562379837036, "learning_rate": 3.116649182043534e-06, "loss": 0.5227, "step": 15130 }, { "epoch": 0.6342705637005753, "grad_norm": 2.1116864681243896, "learning_rate": 3.116020346314751e-06, "loss": 0.4781, "step": 15131 }, { "epoch": 0.6343124823155843, "grad_norm": 1.5427391529083252, "learning_rate": 3.1153915453145304e-06, "loss": 0.4662, "step": 15132 }, { "epoch": 0.6343544009305933, "grad_norm": 1.8018715381622314, "learning_rate": 3.11476277905446e-06, "loss": 0.4585, "step": 15133 }, { "epoch": 0.6343963195456022, "grad_norm": 1.9987627267837524, "learning_rate": 3.1141340475461316e-06, "loss": 0.5922, "step": 15134 }, { "epoch": 0.6344382381606112, "grad_norm": 1.7392041683197021, "learning_rate": 3.1135053508011365e-06, "loss": 0.5352, "step": 15135 }, { "epoch": 0.6344801567756201, "grad_norm": 1.8646166324615479, "learning_rate": 3.11287668883106e-06, "loss": 0.4843, "step": 15136 }, { "epoch": 0.6345220753906291, "grad_norm": 1.9111567735671997, "learning_rate": 3.1122480616474925e-06, "loss": 0.5486, "step": 15137 }, { "epoch": 0.634563994005638, "grad_norm": 1.4711644649505615, "learning_rate": 3.1116194692620248e-06, "loss": 0.4649, "step": 15138 }, { "epoch": 0.634605912620647, "grad_norm": 2.950359582901001, "learning_rate": 3.110990911686241e-06, "loss": 0.457, "step": 15139 }, { "epoch": 0.6346478312356559, "grad_norm": 1.9341158866882324, "learning_rate": 3.1103623889317263e-06, "loss": 0.4725, "step": 15140 }, { "epoch": 0.634689749850665, "grad_norm": 1.936536192893982, "learning_rate": 3.1097339010100713e-06, "loss": 0.5348, "step": 15141 }, { "epoch": 0.6347316684656739, "grad_norm": 1.540769100189209, "learning_rate": 3.1091054479328564e-06, "loss": 0.4878, "step": 15142 }, { "epoch": 0.6347735870806829, "grad_norm": 1.84333336353302, "learning_rate": 3.1084770297116694e-06, "loss": 0.4745, "step": 15143 }, { "epoch": 0.6348155056956918, "grad_norm": 2.1297848224639893, "learning_rate": 3.107848646358095e-06, "loss": 0.4863, "step": 15144 }, { "epoch": 0.6348574243107008, "grad_norm": 1.4695545434951782, "learning_rate": 3.107220297883712e-06, "loss": 0.4031, "step": 15145 }, { "epoch": 0.6348993429257097, "grad_norm": 1.7280442714691162, "learning_rate": 3.1065919843001065e-06, "loss": 0.4527, "step": 15146 }, { "epoch": 0.6349412615407187, "grad_norm": 1.4800986051559448, "learning_rate": 3.1059637056188617e-06, "loss": 0.4904, "step": 15147 }, { "epoch": 0.6349831801557276, "grad_norm": 1.810453176498413, "learning_rate": 3.1053354618515573e-06, "loss": 0.4569, "step": 15148 }, { "epoch": 0.6350250987707367, "grad_norm": 2.780010461807251, "learning_rate": 3.104707253009773e-06, "loss": 0.505, "step": 15149 }, { "epoch": 0.6350670173857456, "grad_norm": 2.0536129474639893, "learning_rate": 3.1040790791050927e-06, "loss": 0.497, "step": 15150 }, { "epoch": 0.6351089360007546, "grad_norm": 1.8571186065673828, "learning_rate": 3.103450940149092e-06, "loss": 0.5219, "step": 15151 }, { "epoch": 0.6351508546157635, "grad_norm": 2.0345818996429443, "learning_rate": 3.1028228361533515e-06, "loss": 0.5398, "step": 15152 }, { "epoch": 0.6351927732307725, "grad_norm": 1.6361833810806274, "learning_rate": 3.1021947671294507e-06, "loss": 0.4836, "step": 15153 }, { "epoch": 0.6352346918457814, "grad_norm": 1.6845203638076782, "learning_rate": 3.101566733088963e-06, "loss": 0.5292, "step": 15154 }, { "epoch": 0.6352766104607904, "grad_norm": 1.833971381187439, "learning_rate": 3.1009387340434684e-06, "loss": 0.4824, "step": 15155 }, { "epoch": 0.6353185290757993, "grad_norm": 1.5594881772994995, "learning_rate": 3.1003107700045453e-06, "loss": 0.4899, "step": 15156 }, { "epoch": 0.6353604476908083, "grad_norm": 1.7567493915557861, "learning_rate": 3.0996828409837654e-06, "loss": 0.4807, "step": 15157 }, { "epoch": 0.6354023663058173, "grad_norm": 1.9927682876586914, "learning_rate": 3.0990549469927043e-06, "loss": 0.4971, "step": 15158 }, { "epoch": 0.6354442849208262, "grad_norm": 1.7596977949142456, "learning_rate": 3.0984270880429398e-06, "loss": 0.5606, "step": 15159 }, { "epoch": 0.6354862035358352, "grad_norm": 1.7797081470489502, "learning_rate": 3.0977992641460404e-06, "loss": 0.5205, "step": 15160 }, { "epoch": 0.6355281221508441, "grad_norm": 2.529088020324707, "learning_rate": 3.0971714753135828e-06, "loss": 0.5175, "step": 15161 }, { "epoch": 0.6355700407658531, "grad_norm": 1.8061225414276123, "learning_rate": 3.0965437215571393e-06, "loss": 0.5168, "step": 15162 }, { "epoch": 0.635611959380862, "grad_norm": 1.571179986000061, "learning_rate": 3.09591600288828e-06, "loss": 0.4604, "step": 15163 }, { "epoch": 0.635653877995871, "grad_norm": 1.872184157371521, "learning_rate": 3.095288319318576e-06, "loss": 0.4636, "step": 15164 }, { "epoch": 0.6356957966108799, "grad_norm": 1.6508114337921143, "learning_rate": 3.0946606708596013e-06, "loss": 0.4713, "step": 15165 }, { "epoch": 0.635737715225889, "grad_norm": 1.9917021989822388, "learning_rate": 3.0940330575229203e-06, "loss": 0.4949, "step": 15166 }, { "epoch": 0.6357796338408979, "grad_norm": 1.5064020156860352, "learning_rate": 3.0934054793201064e-06, "loss": 0.4609, "step": 15167 }, { "epoch": 0.6358215524559069, "grad_norm": 1.6376943588256836, "learning_rate": 3.0927779362627275e-06, "loss": 0.4883, "step": 15168 }, { "epoch": 0.6358634710709158, "grad_norm": 1.7151234149932861, "learning_rate": 3.0921504283623492e-06, "loss": 0.4446, "step": 15169 }, { "epoch": 0.6359053896859248, "grad_norm": 2.1445634365081787, "learning_rate": 3.0915229556305393e-06, "loss": 0.5381, "step": 15170 }, { "epoch": 0.6359473083009337, "grad_norm": 2.1786460876464844, "learning_rate": 3.090895518078868e-06, "loss": 0.5202, "step": 15171 }, { "epoch": 0.6359892269159427, "grad_norm": 1.7281932830810547, "learning_rate": 3.0902681157188976e-06, "loss": 0.5008, "step": 15172 }, { "epoch": 0.6360311455309516, "grad_norm": 1.573872447013855, "learning_rate": 3.089640748562193e-06, "loss": 0.4425, "step": 15173 }, { "epoch": 0.6360730641459607, "grad_norm": 1.6368608474731445, "learning_rate": 3.0890134166203224e-06, "loss": 0.5331, "step": 15174 }, { "epoch": 0.6361149827609696, "grad_norm": 1.6550159454345703, "learning_rate": 3.088386119904846e-06, "loss": 0.4564, "step": 15175 }, { "epoch": 0.6361569013759786, "grad_norm": 1.551748275756836, "learning_rate": 3.087758858427329e-06, "loss": 0.4981, "step": 15176 }, { "epoch": 0.6361988199909875, "grad_norm": 1.924797773361206, "learning_rate": 3.087131632199335e-06, "loss": 0.4606, "step": 15177 }, { "epoch": 0.6362407386059965, "grad_norm": 1.874786138534546, "learning_rate": 3.0865044412324236e-06, "loss": 0.4928, "step": 15178 }, { "epoch": 0.6362826572210054, "grad_norm": 1.55496084690094, "learning_rate": 3.085877285538157e-06, "loss": 0.4328, "step": 15179 }, { "epoch": 0.6363245758360144, "grad_norm": 1.9425028562545776, "learning_rate": 3.085250165128099e-06, "loss": 0.4341, "step": 15180 }, { "epoch": 0.6363664944510233, "grad_norm": 1.5478640794754028, "learning_rate": 3.084623080013805e-06, "loss": 0.4662, "step": 15181 }, { "epoch": 0.6364084130660324, "grad_norm": 1.6760433912277222, "learning_rate": 3.083996030206836e-06, "loss": 0.5139, "step": 15182 }, { "epoch": 0.6364503316810413, "grad_norm": 1.7146023511886597, "learning_rate": 3.0833690157187535e-06, "loss": 0.5108, "step": 15183 }, { "epoch": 0.6364922502960502, "grad_norm": 1.8647948503494263, "learning_rate": 3.0827420365611116e-06, "loss": 0.5109, "step": 15184 }, { "epoch": 0.6365341689110592, "grad_norm": 1.7261744737625122, "learning_rate": 3.0821150927454705e-06, "loss": 0.4684, "step": 15185 }, { "epoch": 0.6365760875260681, "grad_norm": 1.6361078023910522, "learning_rate": 3.0814881842833875e-06, "loss": 0.4502, "step": 15186 }, { "epoch": 0.6366180061410771, "grad_norm": 1.7043025493621826, "learning_rate": 3.0808613111864162e-06, "loss": 0.4954, "step": 15187 }, { "epoch": 0.636659924756086, "grad_norm": 2.0548739433288574, "learning_rate": 3.0802344734661127e-06, "loss": 0.5324, "step": 15188 }, { "epoch": 0.636701843371095, "grad_norm": 1.9021942615509033, "learning_rate": 3.079607671134035e-06, "loss": 0.4987, "step": 15189 }, { "epoch": 0.6367437619861039, "grad_norm": 2.0091497898101807, "learning_rate": 3.078980904201734e-06, "loss": 0.5355, "step": 15190 }, { "epoch": 0.636785680601113, "grad_norm": 1.8879692554473877, "learning_rate": 3.0783541726807627e-06, "loss": 0.5004, "step": 15191 }, { "epoch": 0.6368275992161219, "grad_norm": 1.9788647890090942, "learning_rate": 3.077727476582678e-06, "loss": 0.4955, "step": 15192 }, { "epoch": 0.6368695178311309, "grad_norm": 1.5998483896255493, "learning_rate": 3.077100815919028e-06, "loss": 0.4979, "step": 15193 }, { "epoch": 0.6369114364461398, "grad_norm": 1.6867079734802246, "learning_rate": 3.0764741907013673e-06, "loss": 0.4511, "step": 15194 }, { "epoch": 0.6369533550611488, "grad_norm": 3.0998523235321045, "learning_rate": 3.0758476009412464e-06, "loss": 0.5108, "step": 15195 }, { "epoch": 0.6369952736761577, "grad_norm": 2.2236363887786865, "learning_rate": 3.0752210466502127e-06, "loss": 0.4707, "step": 15196 }, { "epoch": 0.6370371922911667, "grad_norm": 1.5846049785614014, "learning_rate": 3.0745945278398183e-06, "loss": 0.514, "step": 15197 }, { "epoch": 0.6370791109061756, "grad_norm": 1.9495997428894043, "learning_rate": 3.0739680445216137e-06, "loss": 0.4623, "step": 15198 }, { "epoch": 0.6371210295211847, "grad_norm": 1.6844481229782104, "learning_rate": 3.073341596707145e-06, "loss": 0.4395, "step": 15199 }, { "epoch": 0.6371629481361936, "grad_norm": 1.8903331756591797, "learning_rate": 3.0727151844079593e-06, "loss": 0.4655, "step": 15200 }, { "epoch": 0.6372048667512026, "grad_norm": 1.56187903881073, "learning_rate": 3.072088807635607e-06, "loss": 0.5035, "step": 15201 }, { "epoch": 0.6372467853662115, "grad_norm": 3.969640016555786, "learning_rate": 3.07146246640163e-06, "loss": 0.4618, "step": 15202 }, { "epoch": 0.6372887039812205, "grad_norm": 1.645208716392517, "learning_rate": 3.0708361607175773e-06, "loss": 0.4817, "step": 15203 }, { "epoch": 0.6373306225962294, "grad_norm": 2.5241284370422363, "learning_rate": 3.070209890594994e-06, "loss": 0.4657, "step": 15204 }, { "epoch": 0.6373725412112384, "grad_norm": 1.8164807558059692, "learning_rate": 3.069583656045422e-06, "loss": 0.4943, "step": 15205 }, { "epoch": 0.6374144598262473, "grad_norm": 1.9556167125701904, "learning_rate": 3.0689574570804075e-06, "loss": 0.4509, "step": 15206 }, { "epoch": 0.6374563784412564, "grad_norm": 2.471795082092285, "learning_rate": 3.068331293711492e-06, "loss": 0.522, "step": 15207 }, { "epoch": 0.6374982970562653, "grad_norm": 2.135995626449585, "learning_rate": 3.0677051659502184e-06, "loss": 0.4752, "step": 15208 }, { "epoch": 0.6375402156712742, "grad_norm": 1.800854206085205, "learning_rate": 3.067079073808129e-06, "loss": 0.4887, "step": 15209 }, { "epoch": 0.6375821342862832, "grad_norm": 1.8455564975738525, "learning_rate": 3.0664530172967655e-06, "loss": 0.4881, "step": 15210 }, { "epoch": 0.6376240529012921, "grad_norm": 1.9169912338256836, "learning_rate": 3.065826996427666e-06, "loss": 0.4974, "step": 15211 }, { "epoch": 0.6376659715163011, "grad_norm": 1.8236714601516724, "learning_rate": 3.0652010112123743e-06, "loss": 0.5011, "step": 15212 }, { "epoch": 0.63770789013131, "grad_norm": 1.8541895151138306, "learning_rate": 3.064575061662425e-06, "loss": 0.5408, "step": 15213 }, { "epoch": 0.637749808746319, "grad_norm": 2.0815987586975098, "learning_rate": 3.063949147789359e-06, "loss": 0.5026, "step": 15214 }, { "epoch": 0.6377917273613279, "grad_norm": 1.7265050411224365, "learning_rate": 3.0633232696047154e-06, "loss": 0.43, "step": 15215 }, { "epoch": 0.637833645976337, "grad_norm": 1.803307294845581, "learning_rate": 3.0626974271200284e-06, "loss": 0.5408, "step": 15216 }, { "epoch": 0.6378755645913459, "grad_norm": 1.8883278369903564, "learning_rate": 3.062071620346836e-06, "loss": 0.4892, "step": 15217 }, { "epoch": 0.6379174832063549, "grad_norm": 1.616906762123108, "learning_rate": 3.061445849296676e-06, "loss": 0.4446, "step": 15218 }, { "epoch": 0.6379594018213638, "grad_norm": 1.7279715538024902, "learning_rate": 3.060820113981081e-06, "loss": 0.4962, "step": 15219 }, { "epoch": 0.6380013204363728, "grad_norm": 1.5800811052322388, "learning_rate": 3.0601944144115845e-06, "loss": 0.4727, "step": 15220 }, { "epoch": 0.6380432390513817, "grad_norm": 1.759667992591858, "learning_rate": 3.059568750599725e-06, "loss": 0.5204, "step": 15221 }, { "epoch": 0.6380851576663907, "grad_norm": 2.940377712249756, "learning_rate": 3.0589431225570317e-06, "loss": 0.4862, "step": 15222 }, { "epoch": 0.6381270762813996, "grad_norm": 2.0910351276397705, "learning_rate": 3.0583175302950384e-06, "loss": 0.4454, "step": 15223 }, { "epoch": 0.6381689948964087, "grad_norm": 1.6140334606170654, "learning_rate": 3.0576919738252786e-06, "loss": 0.4449, "step": 15224 }, { "epoch": 0.6382109135114176, "grad_norm": 2.2103652954101562, "learning_rate": 3.0570664531592796e-06, "loss": 0.5134, "step": 15225 }, { "epoch": 0.6382528321264266, "grad_norm": 1.5882501602172852, "learning_rate": 3.0564409683085754e-06, "loss": 0.4983, "step": 15226 }, { "epoch": 0.6382947507414355, "grad_norm": 1.957087516784668, "learning_rate": 3.0558155192846968e-06, "loss": 0.5186, "step": 15227 }, { "epoch": 0.6383366693564445, "grad_norm": 2.116978168487549, "learning_rate": 3.0551901060991697e-06, "loss": 0.5197, "step": 15228 }, { "epoch": 0.6383785879714534, "grad_norm": 1.731237530708313, "learning_rate": 3.054564728763524e-06, "loss": 0.4607, "step": 15229 }, { "epoch": 0.6384205065864624, "grad_norm": 1.8891807794570923, "learning_rate": 3.0539393872892896e-06, "loss": 0.5323, "step": 15230 }, { "epoch": 0.6384624252014713, "grad_norm": 2.6382694244384766, "learning_rate": 3.053314081687991e-06, "loss": 0.4796, "step": 15231 }, { "epoch": 0.6385043438164804, "grad_norm": 2.7988874912261963, "learning_rate": 3.052688811971156e-06, "loss": 0.5146, "step": 15232 }, { "epoch": 0.6385462624314893, "grad_norm": 2.1218292713165283, "learning_rate": 3.0520635781503125e-06, "loss": 0.4747, "step": 15233 }, { "epoch": 0.6385881810464982, "grad_norm": 1.568444848060608, "learning_rate": 3.0514383802369816e-06, "loss": 0.5184, "step": 15234 }, { "epoch": 0.6386300996615072, "grad_norm": 1.8026736974716187, "learning_rate": 3.050813218242691e-06, "loss": 0.5089, "step": 15235 }, { "epoch": 0.6386720182765161, "grad_norm": 1.670105218887329, "learning_rate": 3.050188092178965e-06, "loss": 0.5045, "step": 15236 }, { "epoch": 0.6387139368915251, "grad_norm": 1.589318871498108, "learning_rate": 3.049563002057326e-06, "loss": 0.4589, "step": 15237 }, { "epoch": 0.638755855506534, "grad_norm": 1.7756520509719849, "learning_rate": 3.0489379478892955e-06, "loss": 0.5002, "step": 15238 }, { "epoch": 0.638797774121543, "grad_norm": 1.8619989156723022, "learning_rate": 3.0483129296863977e-06, "loss": 0.5565, "step": 15239 }, { "epoch": 0.638839692736552, "grad_norm": 1.6049658060073853, "learning_rate": 3.0476879474601516e-06, "loss": 0.4426, "step": 15240 }, { "epoch": 0.638881611351561, "grad_norm": 1.8557029962539673, "learning_rate": 3.04706300122208e-06, "loss": 0.5058, "step": 15241 }, { "epoch": 0.6389235299665699, "grad_norm": 1.7231884002685547, "learning_rate": 3.046438090983704e-06, "loss": 0.4899, "step": 15242 }, { "epoch": 0.6389654485815789, "grad_norm": 1.7481414079666138, "learning_rate": 3.045813216756538e-06, "loss": 0.4858, "step": 15243 }, { "epoch": 0.6390073671965878, "grad_norm": 2.903275966644287, "learning_rate": 3.045188378552104e-06, "loss": 0.4949, "step": 15244 }, { "epoch": 0.6390492858115968, "grad_norm": 1.9792839288711548, "learning_rate": 3.0445635763819225e-06, "loss": 0.5193, "step": 15245 }, { "epoch": 0.6390912044266057, "grad_norm": 1.9323673248291016, "learning_rate": 3.0439388102575062e-06, "loss": 0.4691, "step": 15246 }, { "epoch": 0.6391331230416147, "grad_norm": 1.8328906297683716, "learning_rate": 3.0433140801903736e-06, "loss": 0.5223, "step": 15247 }, { "epoch": 0.6391750416566236, "grad_norm": 1.9289565086364746, "learning_rate": 3.0426893861920426e-06, "loss": 0.5138, "step": 15248 }, { "epoch": 0.6392169602716327, "grad_norm": 1.913586974143982, "learning_rate": 3.0420647282740245e-06, "loss": 0.5416, "step": 15249 }, { "epoch": 0.6392588788866416, "grad_norm": 2.1621387004852295, "learning_rate": 3.041440106447837e-06, "loss": 0.4733, "step": 15250 }, { "epoch": 0.6393007975016506, "grad_norm": 2.5638744831085205, "learning_rate": 3.0408155207249947e-06, "loss": 0.4343, "step": 15251 }, { "epoch": 0.6393427161166595, "grad_norm": 1.8766599893569946, "learning_rate": 3.0401909711170092e-06, "loss": 0.4915, "step": 15252 }, { "epoch": 0.6393846347316685, "grad_norm": 1.9394434690475464, "learning_rate": 3.0395664576353923e-06, "loss": 0.4501, "step": 15253 }, { "epoch": 0.6394265533466774, "grad_norm": 1.8844939470291138, "learning_rate": 3.03894198029166e-06, "loss": 0.4532, "step": 15254 }, { "epoch": 0.6394684719616864, "grad_norm": 1.8131357431411743, "learning_rate": 3.0383175390973207e-06, "loss": 0.4439, "step": 15255 }, { "epoch": 0.6395103905766953, "grad_norm": 1.8547582626342773, "learning_rate": 3.0376931340638837e-06, "loss": 0.5266, "step": 15256 }, { "epoch": 0.6395523091917044, "grad_norm": 1.6724803447723389, "learning_rate": 3.037068765202864e-06, "loss": 0.489, "step": 15257 }, { "epoch": 0.6395942278067133, "grad_norm": 2.0972723960876465, "learning_rate": 3.0364444325257653e-06, "loss": 0.5162, "step": 15258 }, { "epoch": 0.6396361464217222, "grad_norm": 2.1467576026916504, "learning_rate": 3.0358201360441e-06, "loss": 0.4902, "step": 15259 }, { "epoch": 0.6396780650367312, "grad_norm": 1.6645286083221436, "learning_rate": 3.035195875769375e-06, "loss": 0.4498, "step": 15260 }, { "epoch": 0.6397199836517401, "grad_norm": 2.1392099857330322, "learning_rate": 3.034571651713098e-06, "loss": 0.532, "step": 15261 }, { "epoch": 0.6397619022667491, "grad_norm": 1.9577953815460205, "learning_rate": 3.033947463886774e-06, "loss": 0.56, "step": 15262 }, { "epoch": 0.639803820881758, "grad_norm": 2.073768138885498, "learning_rate": 3.0333233123019136e-06, "loss": 0.5159, "step": 15263 }, { "epoch": 0.639845739496767, "grad_norm": 1.9633065462112427, "learning_rate": 3.032699196970016e-06, "loss": 0.4933, "step": 15264 }, { "epoch": 0.639887658111776, "grad_norm": 1.9864686727523804, "learning_rate": 3.0320751179025903e-06, "loss": 0.4654, "step": 15265 }, { "epoch": 0.639929576726785, "grad_norm": 2.08784556388855, "learning_rate": 3.0314510751111402e-06, "loss": 0.457, "step": 15266 }, { "epoch": 0.6399714953417939, "grad_norm": 1.788484811782837, "learning_rate": 3.030827068607166e-06, "loss": 0.4882, "step": 15267 }, { "epoch": 0.6400134139568029, "grad_norm": 2.0081846714019775, "learning_rate": 3.0302030984021723e-06, "loss": 0.4567, "step": 15268 }, { "epoch": 0.6400553325718118, "grad_norm": 1.9204487800598145, "learning_rate": 3.0295791645076645e-06, "loss": 0.5527, "step": 15269 }, { "epoch": 0.6400972511868208, "grad_norm": 2.5946154594421387, "learning_rate": 3.0289552669351386e-06, "loss": 0.4675, "step": 15270 }, { "epoch": 0.6401391698018297, "grad_norm": 2.1405582427978516, "learning_rate": 3.0283314056960965e-06, "loss": 0.479, "step": 15271 }, { "epoch": 0.6401810884168387, "grad_norm": 1.7156462669372559, "learning_rate": 3.0277075808020417e-06, "loss": 0.4735, "step": 15272 }, { "epoch": 0.6402230070318476, "grad_norm": 1.8294914960861206, "learning_rate": 3.027083792264469e-06, "loss": 0.5217, "step": 15273 }, { "epoch": 0.6402649256468567, "grad_norm": 1.7920796871185303, "learning_rate": 3.02646004009488e-06, "loss": 0.5228, "step": 15274 }, { "epoch": 0.6403068442618656, "grad_norm": 1.9624027013778687, "learning_rate": 3.025836324304773e-06, "loss": 0.496, "step": 15275 }, { "epoch": 0.6403487628768746, "grad_norm": 1.8314220905303955, "learning_rate": 3.0252126449056417e-06, "loss": 0.4562, "step": 15276 }, { "epoch": 0.6403906814918835, "grad_norm": 3.974862813949585, "learning_rate": 3.0245890019089854e-06, "loss": 0.4533, "step": 15277 }, { "epoch": 0.6404326001068925, "grad_norm": 1.6484469175338745, "learning_rate": 3.023965395326302e-06, "loss": 0.4942, "step": 15278 }, { "epoch": 0.6404745187219014, "grad_norm": 1.747052788734436, "learning_rate": 3.0233418251690838e-06, "loss": 0.499, "step": 15279 }, { "epoch": 0.6405164373369104, "grad_norm": 2.699464797973633, "learning_rate": 3.0227182914488252e-06, "loss": 0.4881, "step": 15280 }, { "epoch": 0.6405583559519193, "grad_norm": 2.8774073123931885, "learning_rate": 3.022094794177024e-06, "loss": 0.5042, "step": 15281 }, { "epoch": 0.6406002745669284, "grad_norm": 2.0040013790130615, "learning_rate": 3.021471333365168e-06, "loss": 0.4856, "step": 15282 }, { "epoch": 0.6406421931819373, "grad_norm": 1.6229380369186401, "learning_rate": 3.0208479090247554e-06, "loss": 0.4895, "step": 15283 }, { "epoch": 0.6406841117969462, "grad_norm": 1.752104640007019, "learning_rate": 3.0202245211672744e-06, "loss": 0.4325, "step": 15284 }, { "epoch": 0.6407260304119552, "grad_norm": 1.9242967367172241, "learning_rate": 3.019601169804216e-06, "loss": 0.5357, "step": 15285 }, { "epoch": 0.6407679490269641, "grad_norm": 1.5816303491592407, "learning_rate": 3.018977854947075e-06, "loss": 0.4921, "step": 15286 }, { "epoch": 0.6408098676419731, "grad_norm": 1.7925633192062378, "learning_rate": 3.018354576607336e-06, "loss": 0.4562, "step": 15287 }, { "epoch": 0.640851786256982, "grad_norm": 1.5981501340866089, "learning_rate": 3.017731334796492e-06, "loss": 0.4776, "step": 15288 }, { "epoch": 0.640893704871991, "grad_norm": 1.8483389616012573, "learning_rate": 3.017108129526031e-06, "loss": 0.5199, "step": 15289 }, { "epoch": 0.640935623487, "grad_norm": 1.6231157779693604, "learning_rate": 3.0164849608074387e-06, "loss": 0.4775, "step": 15290 }, { "epoch": 0.640977542102009, "grad_norm": 1.7014150619506836, "learning_rate": 3.015861828652204e-06, "loss": 0.43, "step": 15291 }, { "epoch": 0.6410194607170179, "grad_norm": 1.7436339855194092, "learning_rate": 3.015238733071816e-06, "loss": 0.4767, "step": 15292 }, { "epoch": 0.6410613793320269, "grad_norm": 1.821264386177063, "learning_rate": 3.0146156740777566e-06, "loss": 0.5303, "step": 15293 }, { "epoch": 0.6411032979470358, "grad_norm": 1.9978660345077515, "learning_rate": 3.013992651681511e-06, "loss": 0.5031, "step": 15294 }, { "epoch": 0.6411452165620448, "grad_norm": 1.6850595474243164, "learning_rate": 3.013369665894568e-06, "loss": 0.5027, "step": 15295 }, { "epoch": 0.6411871351770537, "grad_norm": 1.648443579673767, "learning_rate": 3.0127467167284064e-06, "loss": 0.5221, "step": 15296 }, { "epoch": 0.6412290537920627, "grad_norm": 1.4690133333206177, "learning_rate": 3.0121238041945134e-06, "loss": 0.4395, "step": 15297 }, { "epoch": 0.6412709724070716, "grad_norm": 1.690834403038025, "learning_rate": 3.0115009283043693e-06, "loss": 0.4967, "step": 15298 }, { "epoch": 0.6413128910220807, "grad_norm": 1.7304432392120361, "learning_rate": 3.0108780890694566e-06, "loss": 0.5217, "step": 15299 }, { "epoch": 0.6413548096370896, "grad_norm": 2.1244494915008545, "learning_rate": 3.010255286501256e-06, "loss": 0.5162, "step": 15300 }, { "epoch": 0.6413967282520986, "grad_norm": 2.1914758682250977, "learning_rate": 3.0096325206112496e-06, "loss": 0.4813, "step": 15301 }, { "epoch": 0.6414386468671075, "grad_norm": 2.004333019256592, "learning_rate": 3.009009791410915e-06, "loss": 0.4958, "step": 15302 }, { "epoch": 0.6414805654821165, "grad_norm": 1.945855975151062, "learning_rate": 3.008387098911733e-06, "loss": 0.5026, "step": 15303 }, { "epoch": 0.6415224840971254, "grad_norm": 1.8623268604278564, "learning_rate": 3.0077644431251827e-06, "loss": 0.4657, "step": 15304 }, { "epoch": 0.6415644027121344, "grad_norm": 2.5623505115509033, "learning_rate": 3.007141824062738e-06, "loss": 0.5036, "step": 15305 }, { "epoch": 0.6416063213271433, "grad_norm": 1.7476966381072998, "learning_rate": 3.0065192417358795e-06, "loss": 0.4509, "step": 15306 }, { "epoch": 0.6416482399421524, "grad_norm": 1.7053841352462769, "learning_rate": 3.0058966961560846e-06, "loss": 0.5269, "step": 15307 }, { "epoch": 0.6416901585571613, "grad_norm": 1.6765174865722656, "learning_rate": 3.005274187334827e-06, "loss": 0.4766, "step": 15308 }, { "epoch": 0.6417320771721702, "grad_norm": 1.7520942687988281, "learning_rate": 3.0046517152835795e-06, "loss": 0.5132, "step": 15309 }, { "epoch": 0.6417739957871792, "grad_norm": 2.3090908527374268, "learning_rate": 3.0040292800138226e-06, "loss": 0.4852, "step": 15310 }, { "epoch": 0.6418159144021881, "grad_norm": 1.7549428939819336, "learning_rate": 3.003406881537024e-06, "loss": 0.5147, "step": 15311 }, { "epoch": 0.6418578330171971, "grad_norm": 1.4329240322113037, "learning_rate": 3.002784519864661e-06, "loss": 0.4369, "step": 15312 }, { "epoch": 0.641899751632206, "grad_norm": 2.0538434982299805, "learning_rate": 3.002162195008205e-06, "loss": 0.5073, "step": 15313 }, { "epoch": 0.641941670247215, "grad_norm": 1.7783851623535156, "learning_rate": 3.001539906979124e-06, "loss": 0.4616, "step": 15314 }, { "epoch": 0.641983588862224, "grad_norm": 1.6564828157424927, "learning_rate": 3.000917655788893e-06, "loss": 0.4279, "step": 15315 }, { "epoch": 0.642025507477233, "grad_norm": 1.812177300453186, "learning_rate": 3.000295441448983e-06, "loss": 0.4709, "step": 15316 }, { "epoch": 0.6420674260922419, "grad_norm": 1.900107502937317, "learning_rate": 2.9996732639708615e-06, "loss": 0.4972, "step": 15317 }, { "epoch": 0.6421093447072509, "grad_norm": 1.748963475227356, "learning_rate": 2.999051123365997e-06, "loss": 0.4833, "step": 15318 }, { "epoch": 0.6421512633222598, "grad_norm": 1.570786476135254, "learning_rate": 2.99842901964586e-06, "loss": 0.4531, "step": 15319 }, { "epoch": 0.6421931819372688, "grad_norm": 1.9860601425170898, "learning_rate": 2.9978069528219156e-06, "loss": 0.4238, "step": 15320 }, { "epoch": 0.6422351005522777, "grad_norm": 1.7111682891845703, "learning_rate": 2.997184922905633e-06, "loss": 0.5227, "step": 15321 }, { "epoch": 0.6422770191672867, "grad_norm": 1.829986572265625, "learning_rate": 2.9965629299084788e-06, "loss": 0.4501, "step": 15322 }, { "epoch": 0.6423189377822957, "grad_norm": 2.961852550506592, "learning_rate": 2.995940973841914e-06, "loss": 0.4961, "step": 15323 }, { "epoch": 0.6423608563973047, "grad_norm": 1.6058111190795898, "learning_rate": 2.9953190547174083e-06, "loss": 0.4798, "step": 15324 }, { "epoch": 0.6424027750123136, "grad_norm": 1.7422196865081787, "learning_rate": 2.994697172546426e-06, "loss": 0.4671, "step": 15325 }, { "epoch": 0.6424446936273226, "grad_norm": 1.8477826118469238, "learning_rate": 2.994075327340428e-06, "loss": 0.5438, "step": 15326 }, { "epoch": 0.6424866122423315, "grad_norm": 1.4887678623199463, "learning_rate": 2.9934535191108773e-06, "loss": 0.4303, "step": 15327 }, { "epoch": 0.6425285308573405, "grad_norm": 1.4113906621932983, "learning_rate": 2.9928317478692392e-06, "loss": 0.4581, "step": 15328 }, { "epoch": 0.6425704494723494, "grad_norm": 1.7308019399642944, "learning_rate": 2.9922100136269705e-06, "loss": 0.4662, "step": 15329 }, { "epoch": 0.6426123680873584, "grad_norm": 1.8152531385421753, "learning_rate": 2.9915883163955354e-06, "loss": 0.4854, "step": 15330 }, { "epoch": 0.6426542867023673, "grad_norm": 1.5898551940917969, "learning_rate": 2.9909666561863947e-06, "loss": 0.4532, "step": 15331 }, { "epoch": 0.6426962053173764, "grad_norm": 1.6678571701049805, "learning_rate": 2.990345033011004e-06, "loss": 0.4639, "step": 15332 }, { "epoch": 0.6427381239323853, "grad_norm": 1.8296055793762207, "learning_rate": 2.9897234468808246e-06, "loss": 0.5139, "step": 15333 }, { "epoch": 0.6427800425473942, "grad_norm": 1.658876895904541, "learning_rate": 2.989101897807316e-06, "loss": 0.5116, "step": 15334 }, { "epoch": 0.6428219611624032, "grad_norm": 1.6515837907791138, "learning_rate": 2.9884803858019333e-06, "loss": 0.5278, "step": 15335 }, { "epoch": 0.6428638797774121, "grad_norm": 1.8874136209487915, "learning_rate": 2.9878589108761324e-06, "loss": 0.5027, "step": 15336 }, { "epoch": 0.6429057983924211, "grad_norm": 1.8071128129959106, "learning_rate": 2.987237473041373e-06, "loss": 0.5136, "step": 15337 }, { "epoch": 0.64294771700743, "grad_norm": 1.5864619016647339, "learning_rate": 2.9866160723091062e-06, "loss": 0.425, "step": 15338 }, { "epoch": 0.642989635622439, "grad_norm": 1.7622381448745728, "learning_rate": 2.9859947086907903e-06, "loss": 0.4843, "step": 15339 }, { "epoch": 0.643031554237448, "grad_norm": 1.770789384841919, "learning_rate": 2.985373382197878e-06, "loss": 0.5009, "step": 15340 }, { "epoch": 0.643073472852457, "grad_norm": 1.8143163919448853, "learning_rate": 2.9847520928418204e-06, "loss": 0.4939, "step": 15341 }, { "epoch": 0.6431153914674659, "grad_norm": 1.699279546737671, "learning_rate": 2.984130840634072e-06, "loss": 0.4608, "step": 15342 }, { "epoch": 0.6431573100824749, "grad_norm": 1.8104815483093262, "learning_rate": 2.9835096255860862e-06, "loss": 0.4768, "step": 15343 }, { "epoch": 0.6431992286974838, "grad_norm": 1.6180777549743652, "learning_rate": 2.982888447709312e-06, "loss": 0.4735, "step": 15344 }, { "epoch": 0.6432411473124928, "grad_norm": 1.608960509300232, "learning_rate": 2.9822673070152008e-06, "loss": 0.474, "step": 15345 }, { "epoch": 0.6432830659275017, "grad_norm": 1.8228812217712402, "learning_rate": 2.9816462035152027e-06, "loss": 0.5046, "step": 15346 }, { "epoch": 0.6433249845425107, "grad_norm": 2.0956549644470215, "learning_rate": 2.9810251372207655e-06, "loss": 0.5172, "step": 15347 }, { "epoch": 0.6433669031575197, "grad_norm": 1.7973768711090088, "learning_rate": 2.98040410814334e-06, "loss": 0.5274, "step": 15348 }, { "epoch": 0.6434088217725287, "grad_norm": 1.8846197128295898, "learning_rate": 2.979783116294372e-06, "loss": 0.4608, "step": 15349 }, { "epoch": 0.6434507403875376, "grad_norm": 1.7532190084457397, "learning_rate": 2.9791621616853094e-06, "loss": 0.4495, "step": 15350 }, { "epoch": 0.6434926590025466, "grad_norm": 2.2170569896698, "learning_rate": 2.9785412443275985e-06, "loss": 0.4688, "step": 15351 }, { "epoch": 0.6435345776175555, "grad_norm": 1.9169570207595825, "learning_rate": 2.977920364232686e-06, "loss": 0.5555, "step": 15352 }, { "epoch": 0.6435764962325645, "grad_norm": 1.6704045534133911, "learning_rate": 2.9772995214120147e-06, "loss": 0.4698, "step": 15353 }, { "epoch": 0.6436184148475734, "grad_norm": 2.7908926010131836, "learning_rate": 2.9766787158770317e-06, "loss": 0.4768, "step": 15354 }, { "epoch": 0.6436603334625824, "grad_norm": 2.633164167404175, "learning_rate": 2.9760579476391806e-06, "loss": 0.4973, "step": 15355 }, { "epoch": 0.6437022520775914, "grad_norm": 1.7463809251785278, "learning_rate": 2.9754372167099007e-06, "loss": 0.5128, "step": 15356 }, { "epoch": 0.6437441706926004, "grad_norm": 1.6501168012619019, "learning_rate": 2.9748165231006387e-06, "loss": 0.4556, "step": 15357 }, { "epoch": 0.6437860893076093, "grad_norm": 1.6716673374176025, "learning_rate": 2.9741958668228322e-06, "loss": 0.4789, "step": 15358 }, { "epoch": 0.6438280079226182, "grad_norm": 1.6828545331954956, "learning_rate": 2.973575247887926e-06, "loss": 0.4699, "step": 15359 }, { "epoch": 0.6438699265376272, "grad_norm": 2.254659414291382, "learning_rate": 2.9729546663073593e-06, "loss": 0.5311, "step": 15360 }, { "epoch": 0.6439118451526361, "grad_norm": 1.4628247022628784, "learning_rate": 2.9723341220925695e-06, "loss": 0.4928, "step": 15361 }, { "epoch": 0.6439537637676451, "grad_norm": 1.624295711517334, "learning_rate": 2.9717136152549965e-06, "loss": 0.4823, "step": 15362 }, { "epoch": 0.643995682382654, "grad_norm": 1.9435170888900757, "learning_rate": 2.971093145806081e-06, "loss": 0.4762, "step": 15363 }, { "epoch": 0.644037600997663, "grad_norm": 2.0257129669189453, "learning_rate": 2.9704727137572574e-06, "loss": 0.4573, "step": 15364 }, { "epoch": 0.644079519612672, "grad_norm": 1.556050419807434, "learning_rate": 2.9698523191199624e-06, "loss": 0.4662, "step": 15365 }, { "epoch": 0.644121438227681, "grad_norm": 1.5912896394729614, "learning_rate": 2.9692319619056354e-06, "loss": 0.4683, "step": 15366 }, { "epoch": 0.6441633568426899, "grad_norm": 1.59990394115448, "learning_rate": 2.968611642125708e-06, "loss": 0.4611, "step": 15367 }, { "epoch": 0.6442052754576989, "grad_norm": 2.418041706085205, "learning_rate": 2.967991359791618e-06, "loss": 0.4488, "step": 15368 }, { "epoch": 0.6442471940727078, "grad_norm": 1.7214314937591553, "learning_rate": 2.9673711149147988e-06, "loss": 0.4923, "step": 15369 }, { "epoch": 0.6442891126877168, "grad_norm": 2.4941728115081787, "learning_rate": 2.9667509075066807e-06, "loss": 0.4965, "step": 15370 }, { "epoch": 0.6443310313027257, "grad_norm": 2.0000157356262207, "learning_rate": 2.966130737578699e-06, "loss": 0.5395, "step": 15371 }, { "epoch": 0.6443729499177347, "grad_norm": 1.8904476165771484, "learning_rate": 2.9655106051422876e-06, "loss": 0.4989, "step": 15372 }, { "epoch": 0.6444148685327437, "grad_norm": 1.665281891822815, "learning_rate": 2.9648905102088744e-06, "loss": 0.4856, "step": 15373 }, { "epoch": 0.6444567871477527, "grad_norm": 1.744871735572815, "learning_rate": 2.96427045278989e-06, "loss": 0.483, "step": 15374 }, { "epoch": 0.6444987057627616, "grad_norm": 1.610905647277832, "learning_rate": 2.963650432896768e-06, "loss": 0.4395, "step": 15375 }, { "epoch": 0.6445406243777706, "grad_norm": 2.2405452728271484, "learning_rate": 2.9630304505409326e-06, "loss": 0.4649, "step": 15376 }, { "epoch": 0.6445825429927795, "grad_norm": 2.3328113555908203, "learning_rate": 2.962410505733816e-06, "loss": 0.451, "step": 15377 }, { "epoch": 0.6446244616077885, "grad_norm": 1.6930241584777832, "learning_rate": 2.9617905984868463e-06, "loss": 0.5218, "step": 15378 }, { "epoch": 0.6446663802227974, "grad_norm": 1.9378852844238281, "learning_rate": 2.961170728811447e-06, "loss": 0.4824, "step": 15379 }, { "epoch": 0.6447082988378064, "grad_norm": 1.4383317232131958, "learning_rate": 2.960550896719046e-06, "loss": 0.474, "step": 15380 }, { "epoch": 0.6447502174528154, "grad_norm": 1.584168791770935, "learning_rate": 2.959931102221072e-06, "loss": 0.4473, "step": 15381 }, { "epoch": 0.6447921360678244, "grad_norm": 1.959553599357605, "learning_rate": 2.9593113453289475e-06, "loss": 0.5056, "step": 15382 }, { "epoch": 0.6448340546828333, "grad_norm": 1.7488127946853638, "learning_rate": 2.9586916260540954e-06, "loss": 0.4815, "step": 15383 }, { "epoch": 0.6448759732978422, "grad_norm": 1.7840080261230469, "learning_rate": 2.9580719444079437e-06, "loss": 0.5099, "step": 15384 }, { "epoch": 0.6449178919128512, "grad_norm": 1.458752155303955, "learning_rate": 2.9574523004019106e-06, "loss": 0.4584, "step": 15385 }, { "epoch": 0.6449598105278601, "grad_norm": 1.6132924556732178, "learning_rate": 2.956832694047421e-06, "loss": 0.477, "step": 15386 }, { "epoch": 0.6450017291428691, "grad_norm": 2.0199851989746094, "learning_rate": 2.956213125355898e-06, "loss": 0.502, "step": 15387 }, { "epoch": 0.645043647757878, "grad_norm": 1.5122113227844238, "learning_rate": 2.9555935943387585e-06, "loss": 0.4594, "step": 15388 }, { "epoch": 0.645085566372887, "grad_norm": 1.84884512424469, "learning_rate": 2.9549741010074247e-06, "loss": 0.5011, "step": 15389 }, { "epoch": 0.645127484987896, "grad_norm": 1.6608144044876099, "learning_rate": 2.954354645373318e-06, "loss": 0.5167, "step": 15390 }, { "epoch": 0.645169403602905, "grad_norm": 1.6309171915054321, "learning_rate": 2.953735227447855e-06, "loss": 0.4855, "step": 15391 }, { "epoch": 0.6452113222179139, "grad_norm": 1.573901653289795, "learning_rate": 2.953115847242452e-06, "loss": 0.47, "step": 15392 }, { "epoch": 0.6452532408329229, "grad_norm": 1.946297287940979, "learning_rate": 2.9524965047685318e-06, "loss": 0.5214, "step": 15393 }, { "epoch": 0.6452951594479318, "grad_norm": 1.555212140083313, "learning_rate": 2.9518772000375053e-06, "loss": 0.4705, "step": 15394 }, { "epoch": 0.6453370780629408, "grad_norm": 1.7078701257705688, "learning_rate": 2.951257933060792e-06, "loss": 0.4586, "step": 15395 }, { "epoch": 0.6453789966779497, "grad_norm": 1.6192384958267212, "learning_rate": 2.950638703849806e-06, "loss": 0.4888, "step": 15396 }, { "epoch": 0.6454209152929588, "grad_norm": 1.622552514076233, "learning_rate": 2.9500195124159637e-06, "loss": 0.4808, "step": 15397 }, { "epoch": 0.6454628339079677, "grad_norm": 2.0514113903045654, "learning_rate": 2.949400358770676e-06, "loss": 0.494, "step": 15398 }, { "epoch": 0.6455047525229767, "grad_norm": 1.559736728668213, "learning_rate": 2.9487812429253597e-06, "loss": 0.4928, "step": 15399 }, { "epoch": 0.6455466711379856, "grad_norm": 1.697375774383545, "learning_rate": 2.948162164891424e-06, "loss": 0.4554, "step": 15400 }, { "epoch": 0.6455885897529946, "grad_norm": 1.7068753242492676, "learning_rate": 2.9475431246802825e-06, "loss": 0.4619, "step": 15401 }, { "epoch": 0.6456305083680035, "grad_norm": 1.5722171068191528, "learning_rate": 2.9469241223033472e-06, "loss": 0.4303, "step": 15402 }, { "epoch": 0.6456724269830125, "grad_norm": 1.628816843032837, "learning_rate": 2.9463051577720257e-06, "loss": 0.5167, "step": 15403 }, { "epoch": 0.6457143455980214, "grad_norm": 1.8962061405181885, "learning_rate": 2.9456862310977295e-06, "loss": 0.4633, "step": 15404 }, { "epoch": 0.6457562642130305, "grad_norm": 4.27614164352417, "learning_rate": 2.9450673422918696e-06, "loss": 0.435, "step": 15405 }, { "epoch": 0.6457981828280394, "grad_norm": 1.6714746952056885, "learning_rate": 2.9444484913658513e-06, "loss": 0.4998, "step": 15406 }, { "epoch": 0.6458401014430484, "grad_norm": 1.8960026502609253, "learning_rate": 2.9438296783310823e-06, "loss": 0.4819, "step": 15407 }, { "epoch": 0.6458820200580573, "grad_norm": 1.8299602270126343, "learning_rate": 2.9432109031989726e-06, "loss": 0.4887, "step": 15408 }, { "epoch": 0.6459239386730662, "grad_norm": 2.0159358978271484, "learning_rate": 2.942592165980924e-06, "loss": 0.5294, "step": 15409 }, { "epoch": 0.6459658572880752, "grad_norm": 1.8812501430511475, "learning_rate": 2.9419734666883465e-06, "loss": 0.4844, "step": 15410 }, { "epoch": 0.6460077759030841, "grad_norm": 1.9009413719177246, "learning_rate": 2.9413548053326436e-06, "loss": 0.4839, "step": 15411 }, { "epoch": 0.6460496945180931, "grad_norm": 1.6621829271316528, "learning_rate": 2.9407361819252165e-06, "loss": 0.4633, "step": 15412 }, { "epoch": 0.646091613133102, "grad_norm": 1.7169361114501953, "learning_rate": 2.9401175964774715e-06, "loss": 0.4845, "step": 15413 }, { "epoch": 0.6461335317481111, "grad_norm": 1.7559407949447632, "learning_rate": 2.9394990490008125e-06, "loss": 0.4984, "step": 15414 }, { "epoch": 0.64617545036312, "grad_norm": 1.5228970050811768, "learning_rate": 2.9388805395066394e-06, "loss": 0.5013, "step": 15415 }, { "epoch": 0.646217368978129, "grad_norm": 1.709645390510559, "learning_rate": 2.9382620680063535e-06, "loss": 0.4978, "step": 15416 }, { "epoch": 0.6462592875931379, "grad_norm": 1.728033185005188, "learning_rate": 2.937643634511357e-06, "loss": 0.4648, "step": 15417 }, { "epoch": 0.6463012062081469, "grad_norm": 2.2816290855407715, "learning_rate": 2.937025239033048e-06, "loss": 0.4968, "step": 15418 }, { "epoch": 0.6463431248231558, "grad_norm": 1.8572815656661987, "learning_rate": 2.936406881582827e-06, "loss": 0.4621, "step": 15419 }, { "epoch": 0.6463850434381648, "grad_norm": 1.7758500576019287, "learning_rate": 2.935788562172094e-06, "loss": 0.556, "step": 15420 }, { "epoch": 0.6464269620531737, "grad_norm": 1.7009029388427734, "learning_rate": 2.9351702808122424e-06, "loss": 0.4358, "step": 15421 }, { "epoch": 0.6464688806681828, "grad_norm": 1.7787823677062988, "learning_rate": 2.9345520375146723e-06, "loss": 0.4651, "step": 15422 }, { "epoch": 0.6465107992831917, "grad_norm": 1.5358726978302002, "learning_rate": 2.9339338322907825e-06, "loss": 0.4809, "step": 15423 }, { "epoch": 0.6465527178982007, "grad_norm": 2.2706618309020996, "learning_rate": 2.9333156651519644e-06, "loss": 0.471, "step": 15424 }, { "epoch": 0.6465946365132096, "grad_norm": 1.6663620471954346, "learning_rate": 2.932697536109615e-06, "loss": 0.4768, "step": 15425 }, { "epoch": 0.6466365551282186, "grad_norm": 2.2313926219940186, "learning_rate": 2.93207944517513e-06, "loss": 0.4666, "step": 15426 }, { "epoch": 0.6466784737432275, "grad_norm": 1.7069402933120728, "learning_rate": 2.9314613923598996e-06, "loss": 0.4983, "step": 15427 }, { "epoch": 0.6467203923582365, "grad_norm": 1.6241906881332397, "learning_rate": 2.93084337767532e-06, "loss": 0.4812, "step": 15428 }, { "epoch": 0.6467623109732454, "grad_norm": 1.9250192642211914, "learning_rate": 2.9302254011327826e-06, "loss": 0.4912, "step": 15429 }, { "epoch": 0.6468042295882545, "grad_norm": 1.6523491144180298, "learning_rate": 2.929607462743677e-06, "loss": 0.5479, "step": 15430 }, { "epoch": 0.6468461482032634, "grad_norm": 1.929934024810791, "learning_rate": 2.9289895625193975e-06, "loss": 0.4696, "step": 15431 }, { "epoch": 0.6468880668182724, "grad_norm": 1.5864914655685425, "learning_rate": 2.9283717004713307e-06, "loss": 0.4976, "step": 15432 }, { "epoch": 0.6469299854332813, "grad_norm": 1.6864467859268188, "learning_rate": 2.927753876610868e-06, "loss": 0.4995, "step": 15433 }, { "epoch": 0.6469719040482902, "grad_norm": 1.6976234912872314, "learning_rate": 2.9271360909493996e-06, "loss": 0.4742, "step": 15434 }, { "epoch": 0.6470138226632992, "grad_norm": 1.9975855350494385, "learning_rate": 2.926518343498309e-06, "loss": 0.5142, "step": 15435 }, { "epoch": 0.6470557412783081, "grad_norm": 2.0355875492095947, "learning_rate": 2.9259006342689865e-06, "loss": 0.4719, "step": 15436 }, { "epoch": 0.6470976598933171, "grad_norm": 1.6951195001602173, "learning_rate": 2.9252829632728203e-06, "loss": 0.4305, "step": 15437 }, { "epoch": 0.647139578508326, "grad_norm": 1.5742284059524536, "learning_rate": 2.9246653305211937e-06, "loss": 0.5454, "step": 15438 }, { "epoch": 0.6471814971233351, "grad_norm": 1.999169111251831, "learning_rate": 2.9240477360254917e-06, "loss": 0.4597, "step": 15439 }, { "epoch": 0.647223415738344, "grad_norm": 1.7665507793426514, "learning_rate": 2.923430179797101e-06, "loss": 0.4932, "step": 15440 }, { "epoch": 0.647265334353353, "grad_norm": 1.7152624130249023, "learning_rate": 2.922812661847403e-06, "loss": 0.4832, "step": 15441 }, { "epoch": 0.6473072529683619, "grad_norm": 1.5952626466751099, "learning_rate": 2.9221951821877824e-06, "loss": 0.5151, "step": 15442 }, { "epoch": 0.6473491715833709, "grad_norm": 1.8593770265579224, "learning_rate": 2.921577740829622e-06, "loss": 0.5455, "step": 15443 }, { "epoch": 0.6473910901983798, "grad_norm": 1.5366218090057373, "learning_rate": 2.920960337784302e-06, "loss": 0.4515, "step": 15444 }, { "epoch": 0.6474330088133888, "grad_norm": 1.6394343376159668, "learning_rate": 2.920342973063203e-06, "loss": 0.4709, "step": 15445 }, { "epoch": 0.6474749274283977, "grad_norm": 1.5145306587219238, "learning_rate": 2.9197256466777083e-06, "loss": 0.4887, "step": 15446 }, { "epoch": 0.6475168460434068, "grad_norm": 1.8263936042785645, "learning_rate": 2.9191083586391945e-06, "loss": 0.5611, "step": 15447 }, { "epoch": 0.6475587646584157, "grad_norm": 1.5455527305603027, "learning_rate": 2.918491108959041e-06, "loss": 0.4772, "step": 15448 }, { "epoch": 0.6476006832734247, "grad_norm": 1.5119489431381226, "learning_rate": 2.917873897648627e-06, "loss": 0.4566, "step": 15449 }, { "epoch": 0.6476426018884336, "grad_norm": 1.8517701625823975, "learning_rate": 2.917256724719328e-06, "loss": 0.4529, "step": 15450 }, { "epoch": 0.6476845205034426, "grad_norm": 1.8866921663284302, "learning_rate": 2.9166395901825216e-06, "loss": 0.4753, "step": 15451 }, { "epoch": 0.6477264391184515, "grad_norm": 1.909549355506897, "learning_rate": 2.9160224940495863e-06, "loss": 0.5402, "step": 15452 }, { "epoch": 0.6477683577334605, "grad_norm": 2.001009225845337, "learning_rate": 2.915405436331894e-06, "loss": 0.5345, "step": 15453 }, { "epoch": 0.6478102763484694, "grad_norm": 1.6589858531951904, "learning_rate": 2.9147884170408213e-06, "loss": 0.4933, "step": 15454 }, { "epoch": 0.6478521949634785, "grad_norm": 1.489476203918457, "learning_rate": 2.914171436187743e-06, "loss": 0.4522, "step": 15455 }, { "epoch": 0.6478941135784874, "grad_norm": 1.9330228567123413, "learning_rate": 2.9135544937840276e-06, "loss": 0.4785, "step": 15456 }, { "epoch": 0.6479360321934964, "grad_norm": 2.792938709259033, "learning_rate": 2.9129375898410507e-06, "loss": 0.4646, "step": 15457 }, { "epoch": 0.6479779508085053, "grad_norm": 1.7530219554901123, "learning_rate": 2.9123207243701867e-06, "loss": 0.5123, "step": 15458 }, { "epoch": 0.6480198694235142, "grad_norm": 1.5629862546920776, "learning_rate": 2.9117038973828018e-06, "loss": 0.4741, "step": 15459 }, { "epoch": 0.6480617880385232, "grad_norm": 1.6479970216751099, "learning_rate": 2.9110871088902683e-06, "loss": 0.4939, "step": 15460 }, { "epoch": 0.6481037066535321, "grad_norm": 1.6766139268875122, "learning_rate": 2.910470358903958e-06, "loss": 0.4646, "step": 15461 }, { "epoch": 0.6481456252685411, "grad_norm": 1.8361681699752808, "learning_rate": 2.9098536474352356e-06, "loss": 0.5073, "step": 15462 }, { "epoch": 0.64818754388355, "grad_norm": 1.9688127040863037, "learning_rate": 2.9092369744954746e-06, "loss": 0.446, "step": 15463 }, { "epoch": 0.6482294624985591, "grad_norm": 1.9668614864349365, "learning_rate": 2.908620340096037e-06, "loss": 0.4965, "step": 15464 }, { "epoch": 0.648271381113568, "grad_norm": 1.531886339187622, "learning_rate": 2.9080037442482945e-06, "loss": 0.4531, "step": 15465 }, { "epoch": 0.648313299728577, "grad_norm": 1.4518450498580933, "learning_rate": 2.907387186963609e-06, "loss": 0.4834, "step": 15466 }, { "epoch": 0.6483552183435859, "grad_norm": 1.6609379053115845, "learning_rate": 2.906770668253349e-06, "loss": 0.4871, "step": 15467 }, { "epoch": 0.6483971369585949, "grad_norm": 1.6553199291229248, "learning_rate": 2.906154188128877e-06, "loss": 0.5228, "step": 15468 }, { "epoch": 0.6484390555736038, "grad_norm": 1.7376233339309692, "learning_rate": 2.905537746601557e-06, "loss": 0.4949, "step": 15469 }, { "epoch": 0.6484809741886128, "grad_norm": 2.3621418476104736, "learning_rate": 2.9049213436827555e-06, "loss": 0.4938, "step": 15470 }, { "epoch": 0.6485228928036217, "grad_norm": 1.9276096820831299, "learning_rate": 2.9043049793838307e-06, "loss": 0.4566, "step": 15471 }, { "epoch": 0.6485648114186308, "grad_norm": 2.1585919857025146, "learning_rate": 2.903688653716147e-06, "loss": 0.4684, "step": 15472 }, { "epoch": 0.6486067300336397, "grad_norm": 1.6651579141616821, "learning_rate": 2.9030723666910655e-06, "loss": 0.4829, "step": 15473 }, { "epoch": 0.6486486486486487, "grad_norm": 1.6520771980285645, "learning_rate": 2.9024561183199472e-06, "loss": 0.4526, "step": 15474 }, { "epoch": 0.6486905672636576, "grad_norm": 1.7922779321670532, "learning_rate": 2.901839908614148e-06, "loss": 0.5061, "step": 15475 }, { "epoch": 0.6487324858786666, "grad_norm": 3.2030811309814453, "learning_rate": 2.9012237375850316e-06, "loss": 0.4678, "step": 15476 }, { "epoch": 0.6487744044936755, "grad_norm": 1.5956391096115112, "learning_rate": 2.900607605243952e-06, "loss": 0.4206, "step": 15477 }, { "epoch": 0.6488163231086845, "grad_norm": 1.697942852973938, "learning_rate": 2.8999915116022692e-06, "loss": 0.5398, "step": 15478 }, { "epoch": 0.6488582417236934, "grad_norm": 1.553255558013916, "learning_rate": 2.899375456671342e-06, "loss": 0.4496, "step": 15479 }, { "epoch": 0.6489001603387025, "grad_norm": 2.4631145000457764, "learning_rate": 2.8987594404625206e-06, "loss": 0.5072, "step": 15480 }, { "epoch": 0.6489420789537114, "grad_norm": 1.6923596858978271, "learning_rate": 2.8981434629871654e-06, "loss": 0.4746, "step": 15481 }, { "epoch": 0.6489839975687204, "grad_norm": 1.8949882984161377, "learning_rate": 2.8975275242566316e-06, "loss": 0.5475, "step": 15482 }, { "epoch": 0.6490259161837293, "grad_norm": 1.9070576429367065, "learning_rate": 2.8969116242822716e-06, "loss": 0.5422, "step": 15483 }, { "epoch": 0.6490678347987382, "grad_norm": 1.7497810125350952, "learning_rate": 2.8962957630754356e-06, "loss": 0.5139, "step": 15484 }, { "epoch": 0.6491097534137472, "grad_norm": 1.8752055168151855, "learning_rate": 2.895679940647481e-06, "loss": 0.5207, "step": 15485 }, { "epoch": 0.6491516720287561, "grad_norm": 1.864601492881775, "learning_rate": 2.895064157009756e-06, "loss": 0.5282, "step": 15486 }, { "epoch": 0.6491935906437651, "grad_norm": 1.7333967685699463, "learning_rate": 2.894448412173614e-06, "loss": 0.4849, "step": 15487 }, { "epoch": 0.649235509258774, "grad_norm": 1.6881850957870483, "learning_rate": 2.8938327061504055e-06, "loss": 0.433, "step": 15488 }, { "epoch": 0.6492774278737831, "grad_norm": 1.8547883033752441, "learning_rate": 2.8932170389514775e-06, "loss": 0.5157, "step": 15489 }, { "epoch": 0.649319346488792, "grad_norm": 1.5360187292099, "learning_rate": 2.8926014105881812e-06, "loss": 0.4612, "step": 15490 }, { "epoch": 0.649361265103801, "grad_norm": 1.4624847173690796, "learning_rate": 2.8919858210718666e-06, "loss": 0.4347, "step": 15491 }, { "epoch": 0.6494031837188099, "grad_norm": 1.795164704322815, "learning_rate": 2.8913702704138784e-06, "loss": 0.4502, "step": 15492 }, { "epoch": 0.6494451023338189, "grad_norm": 1.8606762886047363, "learning_rate": 2.8907547586255625e-06, "loss": 0.5339, "step": 15493 }, { "epoch": 0.6494870209488278, "grad_norm": 1.7755616903305054, "learning_rate": 2.8901392857182693e-06, "loss": 0.5179, "step": 15494 }, { "epoch": 0.6495289395638368, "grad_norm": 1.7444651126861572, "learning_rate": 2.8895238517033387e-06, "loss": 0.5173, "step": 15495 }, { "epoch": 0.6495708581788457, "grad_norm": 1.7095965147018433, "learning_rate": 2.888908456592119e-06, "loss": 0.534, "step": 15496 }, { "epoch": 0.6496127767938548, "grad_norm": 1.5640041828155518, "learning_rate": 2.8882931003959548e-06, "loss": 0.4232, "step": 15497 }, { "epoch": 0.6496546954088637, "grad_norm": 1.6773544549942017, "learning_rate": 2.8876777831261866e-06, "loss": 0.4904, "step": 15498 }, { "epoch": 0.6496966140238727, "grad_norm": 1.6041916608810425, "learning_rate": 2.887062504794158e-06, "loss": 0.444, "step": 15499 }, { "epoch": 0.6497385326388816, "grad_norm": 1.7858123779296875, "learning_rate": 2.886447265411213e-06, "loss": 0.4968, "step": 15500 }, { "epoch": 0.6497804512538906, "grad_norm": 1.5794248580932617, "learning_rate": 2.8858320649886905e-06, "loss": 0.4398, "step": 15501 }, { "epoch": 0.6498223698688995, "grad_norm": 1.8392211198806763, "learning_rate": 2.88521690353793e-06, "loss": 0.5261, "step": 15502 }, { "epoch": 0.6498642884839085, "grad_norm": 1.795541763305664, "learning_rate": 2.884601781070273e-06, "loss": 0.4985, "step": 15503 }, { "epoch": 0.6499062070989174, "grad_norm": 1.509706735610962, "learning_rate": 2.8839866975970563e-06, "loss": 0.5088, "step": 15504 }, { "epoch": 0.6499481257139265, "grad_norm": 1.8338944911956787, "learning_rate": 2.8833716531296195e-06, "loss": 0.4972, "step": 15505 }, { "epoch": 0.6499900443289354, "grad_norm": 1.637107253074646, "learning_rate": 2.8827566476793023e-06, "loss": 0.4772, "step": 15506 }, { "epoch": 0.6500319629439444, "grad_norm": 1.5851044654846191, "learning_rate": 2.8821416812574377e-06, "loss": 0.4456, "step": 15507 }, { "epoch": 0.6500738815589533, "grad_norm": 2.050861120223999, "learning_rate": 2.8815267538753643e-06, "loss": 0.5109, "step": 15508 }, { "epoch": 0.6501158001739622, "grad_norm": 1.4422553777694702, "learning_rate": 2.8809118655444156e-06, "loss": 0.4428, "step": 15509 }, { "epoch": 0.6501577187889712, "grad_norm": 1.638628602027893, "learning_rate": 2.880297016275928e-06, "loss": 0.474, "step": 15510 }, { "epoch": 0.6501996374039801, "grad_norm": 1.9004687070846558, "learning_rate": 2.8796822060812336e-06, "loss": 0.5422, "step": 15511 }, { "epoch": 0.6502415560189891, "grad_norm": 1.54917311668396, "learning_rate": 2.8790674349716674e-06, "loss": 0.453, "step": 15512 }, { "epoch": 0.650283474633998, "grad_norm": 1.6054965257644653, "learning_rate": 2.8784527029585595e-06, "loss": 0.4267, "step": 15513 }, { "epoch": 0.6503253932490071, "grad_norm": 2.0583865642547607, "learning_rate": 2.8778380100532456e-06, "loss": 0.4716, "step": 15514 }, { "epoch": 0.650367311864016, "grad_norm": 1.8536306619644165, "learning_rate": 2.8772233562670515e-06, "loss": 0.4821, "step": 15515 }, { "epoch": 0.650409230479025, "grad_norm": 1.9165922403335571, "learning_rate": 2.8766087416113104e-06, "loss": 0.5534, "step": 15516 }, { "epoch": 0.6504511490940339, "grad_norm": 1.8114867210388184, "learning_rate": 2.875994166097353e-06, "loss": 0.4813, "step": 15517 }, { "epoch": 0.6504930677090429, "grad_norm": 1.978650689125061, "learning_rate": 2.8753796297365055e-06, "loss": 0.5201, "step": 15518 }, { "epoch": 0.6505349863240518, "grad_norm": 1.772779107093811, "learning_rate": 2.874765132540097e-06, "loss": 0.4648, "step": 15519 }, { "epoch": 0.6505769049390608, "grad_norm": 1.8049055337905884, "learning_rate": 2.874150674519458e-06, "loss": 0.5064, "step": 15520 }, { "epoch": 0.6506188235540697, "grad_norm": 1.8549330234527588, "learning_rate": 2.8735362556859116e-06, "loss": 0.4769, "step": 15521 }, { "epoch": 0.6506607421690788, "grad_norm": 1.6578010320663452, "learning_rate": 2.8729218760507832e-06, "loss": 0.5139, "step": 15522 }, { "epoch": 0.6507026607840877, "grad_norm": 1.500585913658142, "learning_rate": 2.872307535625401e-06, "loss": 0.5199, "step": 15523 }, { "epoch": 0.6507445793990967, "grad_norm": 1.6973097324371338, "learning_rate": 2.8716932344210857e-06, "loss": 0.4631, "step": 15524 }, { "epoch": 0.6507864980141056, "grad_norm": 2.3301422595977783, "learning_rate": 2.871078972449164e-06, "loss": 0.4245, "step": 15525 }, { "epoch": 0.6508284166291146, "grad_norm": 1.8587987422943115, "learning_rate": 2.87046474972096e-06, "loss": 0.4926, "step": 15526 }, { "epoch": 0.6508703352441235, "grad_norm": 1.5297998189926147, "learning_rate": 2.8698505662477917e-06, "loss": 0.4657, "step": 15527 }, { "epoch": 0.6509122538591325, "grad_norm": 1.6831800937652588, "learning_rate": 2.869236422040984e-06, "loss": 0.4772, "step": 15528 }, { "epoch": 0.6509541724741414, "grad_norm": 1.7622122764587402, "learning_rate": 2.8686223171118595e-06, "loss": 0.5363, "step": 15529 }, { "epoch": 0.6509960910891505, "grad_norm": 1.5581425428390503, "learning_rate": 2.8680082514717354e-06, "loss": 0.4823, "step": 15530 }, { "epoch": 0.6510380097041594, "grad_norm": 1.6403422355651855, "learning_rate": 2.8673942251319296e-06, "loss": 0.4997, "step": 15531 }, { "epoch": 0.6510799283191684, "grad_norm": 1.924804925918579, "learning_rate": 2.8667802381037657e-06, "loss": 0.4974, "step": 15532 }, { "epoch": 0.6511218469341773, "grad_norm": 1.7793089151382446, "learning_rate": 2.8661662903985566e-06, "loss": 0.529, "step": 15533 }, { "epoch": 0.6511637655491862, "grad_norm": 1.8838025331497192, "learning_rate": 2.865552382027622e-06, "loss": 0.5571, "step": 15534 }, { "epoch": 0.6512056841641952, "grad_norm": 1.8707002401351929, "learning_rate": 2.864938513002281e-06, "loss": 0.4801, "step": 15535 }, { "epoch": 0.6512476027792041, "grad_norm": 1.5798479318618774, "learning_rate": 2.8643246833338446e-06, "loss": 0.5325, "step": 15536 }, { "epoch": 0.6512895213942131, "grad_norm": 1.852709174156189, "learning_rate": 2.86371089303363e-06, "loss": 0.5092, "step": 15537 }, { "epoch": 0.651331440009222, "grad_norm": 2.0244686603546143, "learning_rate": 2.8630971421129537e-06, "loss": 0.5183, "step": 15538 }, { "epoch": 0.6513733586242311, "grad_norm": 2.0064852237701416, "learning_rate": 2.862483430583128e-06, "loss": 0.5056, "step": 15539 }, { "epoch": 0.65141527723924, "grad_norm": 1.7965426445007324, "learning_rate": 2.861869758455462e-06, "loss": 0.4955, "step": 15540 }, { "epoch": 0.651457195854249, "grad_norm": 1.8405619859695435, "learning_rate": 2.861256125741274e-06, "loss": 0.5254, "step": 15541 }, { "epoch": 0.6514991144692579, "grad_norm": 1.726507306098938, "learning_rate": 2.8606425324518704e-06, "loss": 0.5343, "step": 15542 }, { "epoch": 0.6515410330842669, "grad_norm": 1.9594175815582275, "learning_rate": 2.8600289785985647e-06, "loss": 0.5244, "step": 15543 }, { "epoch": 0.6515829516992758, "grad_norm": 1.673552393913269, "learning_rate": 2.859415464192668e-06, "loss": 0.5139, "step": 15544 }, { "epoch": 0.6516248703142848, "grad_norm": 2.0837056636810303, "learning_rate": 2.8588019892454854e-06, "loss": 0.4275, "step": 15545 }, { "epoch": 0.6516667889292937, "grad_norm": 1.8965275287628174, "learning_rate": 2.8581885537683286e-06, "loss": 0.4771, "step": 15546 }, { "epoch": 0.6517087075443028, "grad_norm": 1.6180416345596313, "learning_rate": 2.857575157772507e-06, "loss": 0.5092, "step": 15547 }, { "epoch": 0.6517506261593117, "grad_norm": 1.573063611984253, "learning_rate": 2.8569618012693253e-06, "loss": 0.4052, "step": 15548 }, { "epoch": 0.6517925447743207, "grad_norm": 1.934844732284546, "learning_rate": 2.8563484842700877e-06, "loss": 0.4739, "step": 15549 }, { "epoch": 0.6518344633893296, "grad_norm": 1.6864871978759766, "learning_rate": 2.855735206786105e-06, "loss": 0.4685, "step": 15550 }, { "epoch": 0.6518763820043386, "grad_norm": 1.9103624820709229, "learning_rate": 2.855121968828677e-06, "loss": 0.4957, "step": 15551 }, { "epoch": 0.6519183006193475, "grad_norm": 2.1427009105682373, "learning_rate": 2.85450877040911e-06, "loss": 0.5236, "step": 15552 }, { "epoch": 0.6519602192343565, "grad_norm": 1.7068318128585815, "learning_rate": 2.8538956115387106e-06, "loss": 0.4701, "step": 15553 }, { "epoch": 0.6520021378493654, "grad_norm": 1.8240599632263184, "learning_rate": 2.8532824922287763e-06, "loss": 0.5444, "step": 15554 }, { "epoch": 0.6520440564643745, "grad_norm": 2.115325689315796, "learning_rate": 2.852669412490611e-06, "loss": 0.4601, "step": 15555 }, { "epoch": 0.6520859750793834, "grad_norm": 2.1320576667785645, "learning_rate": 2.852056372335519e-06, "loss": 0.5069, "step": 15556 }, { "epoch": 0.6521278936943924, "grad_norm": 1.7454692125320435, "learning_rate": 2.851443371774798e-06, "loss": 0.4993, "step": 15557 }, { "epoch": 0.6521698123094013, "grad_norm": 1.7733795642852783, "learning_rate": 2.850830410819747e-06, "loss": 0.4613, "step": 15558 }, { "epoch": 0.6522117309244102, "grad_norm": 1.8420008420944214, "learning_rate": 2.8502174894816675e-06, "loss": 0.5175, "step": 15559 }, { "epoch": 0.6522536495394192, "grad_norm": 2.421971082687378, "learning_rate": 2.8496046077718546e-06, "loss": 0.4921, "step": 15560 }, { "epoch": 0.6522955681544281, "grad_norm": 1.8100082874298096, "learning_rate": 2.8489917657016082e-06, "loss": 0.4583, "step": 15561 }, { "epoch": 0.6523374867694371, "grad_norm": 2.4507129192352295, "learning_rate": 2.848378963282227e-06, "loss": 0.5959, "step": 15562 }, { "epoch": 0.652379405384446, "grad_norm": 1.7669020891189575, "learning_rate": 2.8477662005250027e-06, "loss": 0.4594, "step": 15563 }, { "epoch": 0.6524213239994551, "grad_norm": 1.5793951749801636, "learning_rate": 2.8471534774412336e-06, "loss": 0.5057, "step": 15564 }, { "epoch": 0.652463242614464, "grad_norm": 1.6714202165603638, "learning_rate": 2.846540794042215e-06, "loss": 0.4512, "step": 15565 }, { "epoch": 0.652505161229473, "grad_norm": 1.6305254697799683, "learning_rate": 2.845928150339239e-06, "loss": 0.4314, "step": 15566 }, { "epoch": 0.6525470798444819, "grad_norm": 1.7072266340255737, "learning_rate": 2.845315546343601e-06, "loss": 0.4799, "step": 15567 }, { "epoch": 0.6525889984594909, "grad_norm": 1.8237684965133667, "learning_rate": 2.8447029820665917e-06, "loss": 0.4933, "step": 15568 }, { "epoch": 0.6526309170744998, "grad_norm": 2.230212688446045, "learning_rate": 2.844090457519502e-06, "loss": 0.5346, "step": 15569 }, { "epoch": 0.6526728356895088, "grad_norm": 1.8403561115264893, "learning_rate": 2.843477972713624e-06, "loss": 0.4881, "step": 15570 }, { "epoch": 0.6527147543045178, "grad_norm": 1.6833523511886597, "learning_rate": 2.8428655276602502e-06, "loss": 0.4697, "step": 15571 }, { "epoch": 0.6527566729195268, "grad_norm": 2.3293375968933105, "learning_rate": 2.842253122370666e-06, "loss": 0.5089, "step": 15572 }, { "epoch": 0.6527985915345357, "grad_norm": 1.5988572835922241, "learning_rate": 2.841640756856163e-06, "loss": 0.4713, "step": 15573 }, { "epoch": 0.6528405101495447, "grad_norm": 1.9090279340744019, "learning_rate": 2.841028431128031e-06, "loss": 0.4987, "step": 15574 }, { "epoch": 0.6528824287645536, "grad_norm": 1.7724249362945557, "learning_rate": 2.840416145197553e-06, "loss": 0.4641, "step": 15575 }, { "epoch": 0.6529243473795626, "grad_norm": 1.8015986680984497, "learning_rate": 2.8398038990760197e-06, "loss": 0.4683, "step": 15576 }, { "epoch": 0.6529662659945715, "grad_norm": 1.9618116617202759, "learning_rate": 2.8391916927747156e-06, "loss": 0.5055, "step": 15577 }, { "epoch": 0.6530081846095805, "grad_norm": 1.8886492252349854, "learning_rate": 2.8385795263049224e-06, "loss": 0.4588, "step": 15578 }, { "epoch": 0.6530501032245895, "grad_norm": 1.6721867322921753, "learning_rate": 2.8379673996779277e-06, "loss": 0.4938, "step": 15579 }, { "epoch": 0.6530920218395985, "grad_norm": 1.6682308912277222, "learning_rate": 2.837355312905018e-06, "loss": 0.4942, "step": 15580 }, { "epoch": 0.6531339404546074, "grad_norm": 1.7790509462356567, "learning_rate": 2.8367432659974707e-06, "loss": 0.5033, "step": 15581 }, { "epoch": 0.6531758590696164, "grad_norm": 1.6596177816390991, "learning_rate": 2.836131258966573e-06, "loss": 0.4691, "step": 15582 }, { "epoch": 0.6532177776846253, "grad_norm": 1.8499360084533691, "learning_rate": 2.835519291823602e-06, "loss": 0.4674, "step": 15583 }, { "epoch": 0.6532596962996342, "grad_norm": 1.703783631324768, "learning_rate": 2.8349073645798404e-06, "loss": 0.4746, "step": 15584 }, { "epoch": 0.6533016149146432, "grad_norm": 1.757399559020996, "learning_rate": 2.83429547724657e-06, "loss": 0.5033, "step": 15585 }, { "epoch": 0.6533435335296521, "grad_norm": 1.85019052028656, "learning_rate": 2.833683629835069e-06, "loss": 0.5011, "step": 15586 }, { "epoch": 0.6533854521446611, "grad_norm": 1.6769407987594604, "learning_rate": 2.8330718223566132e-06, "loss": 0.4795, "step": 15587 }, { "epoch": 0.6534273707596701, "grad_norm": 1.7930870056152344, "learning_rate": 2.8324600548224852e-06, "loss": 0.5052, "step": 15588 }, { "epoch": 0.6534692893746791, "grad_norm": 2.2561638355255127, "learning_rate": 2.8318483272439567e-06, "loss": 0.4982, "step": 15589 }, { "epoch": 0.653511207989688, "grad_norm": 6.0958404541015625, "learning_rate": 2.8312366396323077e-06, "loss": 0.4897, "step": 15590 }, { "epoch": 0.653553126604697, "grad_norm": 1.662348747253418, "learning_rate": 2.830624991998815e-06, "loss": 0.4973, "step": 15591 }, { "epoch": 0.6535950452197059, "grad_norm": 1.8264192342758179, "learning_rate": 2.8300133843547483e-06, "loss": 0.5161, "step": 15592 }, { "epoch": 0.6536369638347149, "grad_norm": 2.0833687782287598, "learning_rate": 2.8294018167113863e-06, "loss": 0.5423, "step": 15593 }, { "epoch": 0.6536788824497238, "grad_norm": 1.7103242874145508, "learning_rate": 2.8287902890800024e-06, "loss": 0.5002, "step": 15594 }, { "epoch": 0.6537208010647328, "grad_norm": 1.8904085159301758, "learning_rate": 2.828178801471868e-06, "loss": 0.459, "step": 15595 }, { "epoch": 0.6537627196797418, "grad_norm": 3.0472686290740967, "learning_rate": 2.827567353898253e-06, "loss": 0.475, "step": 15596 }, { "epoch": 0.6538046382947508, "grad_norm": 1.6188325881958008, "learning_rate": 2.8269559463704328e-06, "loss": 0.4947, "step": 15597 }, { "epoch": 0.6538465569097597, "grad_norm": 2.1977713108062744, "learning_rate": 2.8263445788996736e-06, "loss": 0.5334, "step": 15598 }, { "epoch": 0.6538884755247687, "grad_norm": 2.2123987674713135, "learning_rate": 2.825733251497247e-06, "loss": 0.5048, "step": 15599 }, { "epoch": 0.6539303941397776, "grad_norm": 1.624894618988037, "learning_rate": 2.825121964174423e-06, "loss": 0.4156, "step": 15600 }, { "epoch": 0.6539723127547866, "grad_norm": 1.648083209991455, "learning_rate": 2.8245107169424683e-06, "loss": 0.4973, "step": 15601 }, { "epoch": 0.6540142313697955, "grad_norm": 1.9564448595046997, "learning_rate": 2.823899509812651e-06, "loss": 0.5131, "step": 15602 }, { "epoch": 0.6540561499848045, "grad_norm": 1.5963257551193237, "learning_rate": 2.823288342796239e-06, "loss": 0.4309, "step": 15603 }, { "epoch": 0.6540980685998135, "grad_norm": 2.1309330463409424, "learning_rate": 2.822677215904498e-06, "loss": 0.4768, "step": 15604 }, { "epoch": 0.6541399872148225, "grad_norm": 1.7719013690948486, "learning_rate": 2.8220661291486893e-06, "loss": 0.5037, "step": 15605 }, { "epoch": 0.6541819058298314, "grad_norm": 1.9382414817810059, "learning_rate": 2.8214550825400838e-06, "loss": 0.4634, "step": 15606 }, { "epoch": 0.6542238244448404, "grad_norm": 1.712152361869812, "learning_rate": 2.8208440760899396e-06, "loss": 0.4677, "step": 15607 }, { "epoch": 0.6542657430598493, "grad_norm": 1.765098214149475, "learning_rate": 2.8202331098095228e-06, "loss": 0.4501, "step": 15608 }, { "epoch": 0.6543076616748582, "grad_norm": 1.773819923400879, "learning_rate": 2.8196221837100963e-06, "loss": 0.4693, "step": 15609 }, { "epoch": 0.6543495802898672, "grad_norm": 1.6357557773590088, "learning_rate": 2.819011297802919e-06, "loss": 0.499, "step": 15610 }, { "epoch": 0.6543914989048761, "grad_norm": 2.446556568145752, "learning_rate": 2.8184004520992536e-06, "loss": 0.4993, "step": 15611 }, { "epoch": 0.6544334175198852, "grad_norm": 5.988809108734131, "learning_rate": 2.8177896466103616e-06, "loss": 0.4964, "step": 15612 }, { "epoch": 0.6544753361348941, "grad_norm": 1.6020126342773438, "learning_rate": 2.817178881347498e-06, "loss": 0.4347, "step": 15613 }, { "epoch": 0.6545172547499031, "grad_norm": 3.2815356254577637, "learning_rate": 2.816568156321927e-06, "loss": 0.5274, "step": 15614 }, { "epoch": 0.654559173364912, "grad_norm": 1.9124294519424438, "learning_rate": 2.8159574715449033e-06, "loss": 0.4891, "step": 15615 }, { "epoch": 0.654601091979921, "grad_norm": 1.9214246273040771, "learning_rate": 2.815346827027682e-06, "loss": 0.485, "step": 15616 }, { "epoch": 0.6546430105949299, "grad_norm": 1.8322070837020874, "learning_rate": 2.8147362227815224e-06, "loss": 0.4744, "step": 15617 }, { "epoch": 0.6546849292099389, "grad_norm": 1.8007993698120117, "learning_rate": 2.8141256588176807e-06, "loss": 0.5089, "step": 15618 }, { "epoch": 0.6547268478249478, "grad_norm": 1.9340227842330933, "learning_rate": 2.813515135147409e-06, "loss": 0.5173, "step": 15619 }, { "epoch": 0.6547687664399569, "grad_norm": 1.7816157341003418, "learning_rate": 2.8129046517819624e-06, "loss": 0.4576, "step": 15620 }, { "epoch": 0.6548106850549658, "grad_norm": 2.1089553833007812, "learning_rate": 2.812294208732598e-06, "loss": 0.5043, "step": 15621 }, { "epoch": 0.6548526036699748, "grad_norm": 1.9544689655303955, "learning_rate": 2.8116838060105634e-06, "loss": 0.485, "step": 15622 }, { "epoch": 0.6548945222849837, "grad_norm": 1.7107633352279663, "learning_rate": 2.8110734436271135e-06, "loss": 0.4507, "step": 15623 }, { "epoch": 0.6549364408999927, "grad_norm": 2.1534202098846436, "learning_rate": 2.8104631215935e-06, "loss": 0.5068, "step": 15624 }, { "epoch": 0.6549783595150016, "grad_norm": 2.015345335006714, "learning_rate": 2.809852839920969e-06, "loss": 0.49, "step": 15625 }, { "epoch": 0.6550202781300106, "grad_norm": 1.713701605796814, "learning_rate": 2.809242598620773e-06, "loss": 0.5083, "step": 15626 }, { "epoch": 0.6550621967450195, "grad_norm": 1.6738908290863037, "learning_rate": 2.8086323977041637e-06, "loss": 0.4932, "step": 15627 }, { "epoch": 0.6551041153600285, "grad_norm": 1.838126540184021, "learning_rate": 2.808022237182384e-06, "loss": 0.4774, "step": 15628 }, { "epoch": 0.6551460339750375, "grad_norm": 1.653658390045166, "learning_rate": 2.8074121170666846e-06, "loss": 0.4841, "step": 15629 }, { "epoch": 0.6551879525900465, "grad_norm": 1.8979727029800415, "learning_rate": 2.8068020373683137e-06, "loss": 0.4913, "step": 15630 }, { "epoch": 0.6552298712050554, "grad_norm": 1.9045958518981934, "learning_rate": 2.8061919980985124e-06, "loss": 0.4831, "step": 15631 }, { "epoch": 0.6552717898200644, "grad_norm": 1.6337329149246216, "learning_rate": 2.8055819992685314e-06, "loss": 0.4702, "step": 15632 }, { "epoch": 0.6553137084350733, "grad_norm": 1.6654983758926392, "learning_rate": 2.8049720408896125e-06, "loss": 0.4811, "step": 15633 }, { "epoch": 0.6553556270500822, "grad_norm": 1.8828588724136353, "learning_rate": 2.804362122972997e-06, "loss": 0.4962, "step": 15634 }, { "epoch": 0.6553975456650912, "grad_norm": 1.7420620918273926, "learning_rate": 2.8037522455299314e-06, "loss": 0.4694, "step": 15635 }, { "epoch": 0.6554394642801001, "grad_norm": 1.672683835029602, "learning_rate": 2.8031424085716587e-06, "loss": 0.5082, "step": 15636 }, { "epoch": 0.6554813828951092, "grad_norm": 2.0135247707366943, "learning_rate": 2.802532612109416e-06, "loss": 0.5284, "step": 15637 }, { "epoch": 0.6555233015101181, "grad_norm": 1.558758020401001, "learning_rate": 2.8019228561544464e-06, "loss": 0.45, "step": 15638 }, { "epoch": 0.6555652201251271, "grad_norm": 1.7585079669952393, "learning_rate": 2.8013131407179934e-06, "loss": 0.4562, "step": 15639 }, { "epoch": 0.655607138740136, "grad_norm": 1.6260119676589966, "learning_rate": 2.80070346581129e-06, "loss": 0.4724, "step": 15640 }, { "epoch": 0.655649057355145, "grad_norm": 2.1503732204437256, "learning_rate": 2.8000938314455807e-06, "loss": 0.4685, "step": 15641 }, { "epoch": 0.6556909759701539, "grad_norm": 2.0007619857788086, "learning_rate": 2.7994842376320995e-06, "loss": 0.4603, "step": 15642 }, { "epoch": 0.6557328945851629, "grad_norm": 1.9303728342056274, "learning_rate": 2.7988746843820822e-06, "loss": 0.4503, "step": 15643 }, { "epoch": 0.6557748132001718, "grad_norm": 2.029602527618408, "learning_rate": 2.7982651717067685e-06, "loss": 0.461, "step": 15644 }, { "epoch": 0.6558167318151809, "grad_norm": 1.8350293636322021, "learning_rate": 2.7976556996173933e-06, "loss": 0.4629, "step": 15645 }, { "epoch": 0.6558586504301898, "grad_norm": 1.7614198923110962, "learning_rate": 2.7970462681251887e-06, "loss": 0.4759, "step": 15646 }, { "epoch": 0.6559005690451988, "grad_norm": 2.1335976123809814, "learning_rate": 2.7964368772413907e-06, "loss": 0.4677, "step": 15647 }, { "epoch": 0.6559424876602077, "grad_norm": 1.7211261987686157, "learning_rate": 2.795827526977235e-06, "loss": 0.4223, "step": 15648 }, { "epoch": 0.6559844062752167, "grad_norm": 2.1485538482666016, "learning_rate": 2.7952182173439494e-06, "loss": 0.4591, "step": 15649 }, { "epoch": 0.6560263248902256, "grad_norm": 1.859695315361023, "learning_rate": 2.79460894835277e-06, "loss": 0.5376, "step": 15650 }, { "epoch": 0.6560682435052346, "grad_norm": 1.897114634513855, "learning_rate": 2.793999720014926e-06, "loss": 0.4589, "step": 15651 }, { "epoch": 0.6561101621202435, "grad_norm": 1.8658499717712402, "learning_rate": 2.793390532341645e-06, "loss": 0.4877, "step": 15652 }, { "epoch": 0.6561520807352526, "grad_norm": 1.884204626083374, "learning_rate": 2.7927813853441587e-06, "loss": 0.5215, "step": 15653 }, { "epoch": 0.6561939993502615, "grad_norm": 1.8219959735870361, "learning_rate": 2.7921722790336993e-06, "loss": 0.5095, "step": 15654 }, { "epoch": 0.6562359179652705, "grad_norm": 1.9165637493133545, "learning_rate": 2.7915632134214886e-06, "loss": 0.5014, "step": 15655 }, { "epoch": 0.6562778365802794, "grad_norm": 1.9576750993728638, "learning_rate": 2.7909541885187597e-06, "loss": 0.5645, "step": 15656 }, { "epoch": 0.6563197551952884, "grad_norm": 1.9390487670898438, "learning_rate": 2.790345204336734e-06, "loss": 0.5145, "step": 15657 }, { "epoch": 0.6563616738102973, "grad_norm": 1.9418714046478271, "learning_rate": 2.7897362608866398e-06, "loss": 0.4882, "step": 15658 }, { "epoch": 0.6564035924253062, "grad_norm": 1.746083378791809, "learning_rate": 2.789127358179704e-06, "loss": 0.4927, "step": 15659 }, { "epoch": 0.6564455110403152, "grad_norm": 4.396975517272949, "learning_rate": 2.7885184962271474e-06, "loss": 0.4287, "step": 15660 }, { "epoch": 0.6564874296553241, "grad_norm": 1.776552677154541, "learning_rate": 2.7879096750401964e-06, "loss": 0.4313, "step": 15661 }, { "epoch": 0.6565293482703332, "grad_norm": 3.8256125450134277, "learning_rate": 2.7873008946300727e-06, "loss": 0.4466, "step": 15662 }, { "epoch": 0.6565712668853421, "grad_norm": 1.6893999576568604, "learning_rate": 2.7866921550079964e-06, "loss": 0.4912, "step": 15663 }, { "epoch": 0.6566131855003511, "grad_norm": 2.0917141437530518, "learning_rate": 2.786083456185189e-06, "loss": 0.4565, "step": 15664 }, { "epoch": 0.65665510411536, "grad_norm": 1.9470378160476685, "learning_rate": 2.7854747981728757e-06, "loss": 0.5227, "step": 15665 }, { "epoch": 0.656697022730369, "grad_norm": 2.50195574760437, "learning_rate": 2.7848661809822706e-06, "loss": 0.4952, "step": 15666 }, { "epoch": 0.6567389413453779, "grad_norm": 1.6567872762680054, "learning_rate": 2.7842576046245954e-06, "loss": 0.4892, "step": 15667 }, { "epoch": 0.6567808599603869, "grad_norm": 1.6332014799118042, "learning_rate": 2.78364906911107e-06, "loss": 0.4771, "step": 15668 }, { "epoch": 0.6568227785753958, "grad_norm": 1.8133293390274048, "learning_rate": 2.7830405744529084e-06, "loss": 0.4672, "step": 15669 }, { "epoch": 0.6568646971904049, "grad_norm": 2.047213554382324, "learning_rate": 2.78243212066133e-06, "loss": 0.4988, "step": 15670 }, { "epoch": 0.6569066158054138, "grad_norm": 2.4135208129882812, "learning_rate": 2.7818237077475505e-06, "loss": 0.5231, "step": 15671 }, { "epoch": 0.6569485344204228, "grad_norm": 2.1351609230041504, "learning_rate": 2.781215335722782e-06, "loss": 0.5096, "step": 15672 }, { "epoch": 0.6569904530354317, "grad_norm": 1.8746970891952515, "learning_rate": 2.7806070045982413e-06, "loss": 0.5194, "step": 15673 }, { "epoch": 0.6570323716504407, "grad_norm": 1.539628505706787, "learning_rate": 2.779998714385145e-06, "loss": 0.4625, "step": 15674 }, { "epoch": 0.6570742902654496, "grad_norm": 1.5698329210281372, "learning_rate": 2.7793904650947013e-06, "loss": 0.4504, "step": 15675 }, { "epoch": 0.6571162088804586, "grad_norm": 1.6827834844589233, "learning_rate": 2.778782256738124e-06, "loss": 0.493, "step": 15676 }, { "epoch": 0.6571581274954675, "grad_norm": 2.9588561058044434, "learning_rate": 2.7781740893266268e-06, "loss": 0.505, "step": 15677 }, { "epoch": 0.6572000461104766, "grad_norm": 1.7679708003997803, "learning_rate": 2.7775659628714168e-06, "loss": 0.4577, "step": 15678 }, { "epoch": 0.6572419647254855, "grad_norm": 1.6366171836853027, "learning_rate": 2.776957877383708e-06, "loss": 0.5015, "step": 15679 }, { "epoch": 0.6572838833404945, "grad_norm": 1.8361200094223022, "learning_rate": 2.7763498328747076e-06, "loss": 0.508, "step": 15680 }, { "epoch": 0.6573258019555034, "grad_norm": 1.769083857536316, "learning_rate": 2.7757418293556217e-06, "loss": 0.4996, "step": 15681 }, { "epoch": 0.6573677205705124, "grad_norm": 2.4601428508758545, "learning_rate": 2.7751338668376603e-06, "loss": 0.5227, "step": 15682 }, { "epoch": 0.6574096391855213, "grad_norm": 1.7711646556854248, "learning_rate": 2.774525945332033e-06, "loss": 0.5224, "step": 15683 }, { "epoch": 0.6574515578005302, "grad_norm": 1.8895727396011353, "learning_rate": 2.7739180648499405e-06, "loss": 0.4948, "step": 15684 }, { "epoch": 0.6574934764155392, "grad_norm": 1.8012266159057617, "learning_rate": 2.7733102254025913e-06, "loss": 0.4342, "step": 15685 }, { "epoch": 0.6575353950305481, "grad_norm": 1.9885144233703613, "learning_rate": 2.7727024270011915e-06, "loss": 0.5262, "step": 15686 }, { "epoch": 0.6575773136455572, "grad_norm": 1.7026715278625488, "learning_rate": 2.772094669656941e-06, "loss": 0.4661, "step": 15687 }, { "epoch": 0.6576192322605661, "grad_norm": 2.4624266624450684, "learning_rate": 2.771486953381048e-06, "loss": 0.4964, "step": 15688 }, { "epoch": 0.6576611508755751, "grad_norm": 2.0267703533172607, "learning_rate": 2.7708792781847116e-06, "loss": 0.4979, "step": 15689 }, { "epoch": 0.657703069490584, "grad_norm": 2.121866464614868, "learning_rate": 2.7702716440791324e-06, "loss": 0.5464, "step": 15690 }, { "epoch": 0.657744988105593, "grad_norm": 1.94377601146698, "learning_rate": 2.769664051075513e-06, "loss": 0.5004, "step": 15691 }, { "epoch": 0.6577869067206019, "grad_norm": 1.7092790603637695, "learning_rate": 2.7690564991850557e-06, "loss": 0.4809, "step": 15692 }, { "epoch": 0.6578288253356109, "grad_norm": 2.148545265197754, "learning_rate": 2.7684489884189546e-06, "loss": 0.56, "step": 15693 }, { "epoch": 0.6578707439506198, "grad_norm": 1.793596625328064, "learning_rate": 2.7678415187884124e-06, "loss": 0.4651, "step": 15694 }, { "epoch": 0.6579126625656289, "grad_norm": 1.7125126123428345, "learning_rate": 2.7672340903046275e-06, "loss": 0.4898, "step": 15695 }, { "epoch": 0.6579545811806378, "grad_norm": 1.630582332611084, "learning_rate": 2.7666267029787936e-06, "loss": 0.4775, "step": 15696 }, { "epoch": 0.6579964997956468, "grad_norm": 1.7299857139587402, "learning_rate": 2.7660193568221105e-06, "loss": 0.5365, "step": 15697 }, { "epoch": 0.6580384184106557, "grad_norm": 2.117798328399658, "learning_rate": 2.7654120518457705e-06, "loss": 0.4987, "step": 15698 }, { "epoch": 0.6580803370256647, "grad_norm": 1.6990503072738647, "learning_rate": 2.7648047880609726e-06, "loss": 0.4527, "step": 15699 }, { "epoch": 0.6581222556406736, "grad_norm": 1.765352487564087, "learning_rate": 2.7641975654789053e-06, "loss": 0.4777, "step": 15700 }, { "epoch": 0.6581641742556826, "grad_norm": 1.5735955238342285, "learning_rate": 2.763590384110768e-06, "loss": 0.5324, "step": 15701 }, { "epoch": 0.6582060928706915, "grad_norm": 1.7700603008270264, "learning_rate": 2.7629832439677474e-06, "loss": 0.5107, "step": 15702 }, { "epoch": 0.6582480114857006, "grad_norm": 2.5004045963287354, "learning_rate": 2.7623761450610387e-06, "loss": 0.4936, "step": 15703 }, { "epoch": 0.6582899301007095, "grad_norm": 1.7397249937057495, "learning_rate": 2.7617690874018343e-06, "loss": 0.4268, "step": 15704 }, { "epoch": 0.6583318487157185, "grad_norm": 1.7265379428863525, "learning_rate": 2.7611620710013198e-06, "loss": 0.477, "step": 15705 }, { "epoch": 0.6583737673307274, "grad_norm": 2.2678606510162354, "learning_rate": 2.7605550958706907e-06, "loss": 0.513, "step": 15706 }, { "epoch": 0.6584156859457364, "grad_norm": 2.2322914600372314, "learning_rate": 2.7599481620211293e-06, "loss": 0.4497, "step": 15707 }, { "epoch": 0.6584576045607453, "grad_norm": 1.6008880138397217, "learning_rate": 2.759341269463829e-06, "loss": 0.3991, "step": 15708 }, { "epoch": 0.6584995231757542, "grad_norm": 1.564838171005249, "learning_rate": 2.7587344182099728e-06, "loss": 0.503, "step": 15709 }, { "epoch": 0.6585414417907632, "grad_norm": 1.8092259168624878, "learning_rate": 2.758127608270752e-06, "loss": 0.5053, "step": 15710 }, { "epoch": 0.6585833604057721, "grad_norm": 1.7141859531402588, "learning_rate": 2.757520839657346e-06, "loss": 0.4648, "step": 15711 }, { "epoch": 0.6586252790207812, "grad_norm": 2.218221664428711, "learning_rate": 2.756914112380944e-06, "loss": 0.4623, "step": 15712 }, { "epoch": 0.6586671976357901, "grad_norm": 1.9108456373214722, "learning_rate": 2.7563074264527317e-06, "loss": 0.5241, "step": 15713 }, { "epoch": 0.6587091162507991, "grad_norm": 2.115361213684082, "learning_rate": 2.7557007818838877e-06, "loss": 0.5371, "step": 15714 }, { "epoch": 0.658751034865808, "grad_norm": 1.6152989864349365, "learning_rate": 2.7550941786855973e-06, "loss": 0.4886, "step": 15715 }, { "epoch": 0.658792953480817, "grad_norm": 1.8427317142486572, "learning_rate": 2.7544876168690435e-06, "loss": 0.467, "step": 15716 }, { "epoch": 0.6588348720958259, "grad_norm": 1.6676194667816162, "learning_rate": 2.7538810964454077e-06, "loss": 0.4521, "step": 15717 }, { "epoch": 0.6588767907108349, "grad_norm": 1.6756457090377808, "learning_rate": 2.753274617425865e-06, "loss": 0.4626, "step": 15718 }, { "epoch": 0.6589187093258438, "grad_norm": 1.729913592338562, "learning_rate": 2.7526681798216023e-06, "loss": 0.5321, "step": 15719 }, { "epoch": 0.6589606279408529, "grad_norm": 1.973576307296753, "learning_rate": 2.752061783643792e-06, "loss": 0.4678, "step": 15720 }, { "epoch": 0.6590025465558618, "grad_norm": 2.7624285221099854, "learning_rate": 2.751455428903615e-06, "loss": 0.4967, "step": 15721 }, { "epoch": 0.6590444651708708, "grad_norm": 1.6202290058135986, "learning_rate": 2.7508491156122506e-06, "loss": 0.5094, "step": 15722 }, { "epoch": 0.6590863837858797, "grad_norm": 1.5703431367874146, "learning_rate": 2.7502428437808717e-06, "loss": 0.4932, "step": 15723 }, { "epoch": 0.6591283024008887, "grad_norm": 2.06508469581604, "learning_rate": 2.7496366134206552e-06, "loss": 0.4945, "step": 15724 }, { "epoch": 0.6591702210158976, "grad_norm": 1.8628960847854614, "learning_rate": 2.7490304245427797e-06, "loss": 0.4938, "step": 15725 }, { "epoch": 0.6592121396309066, "grad_norm": 1.5088505744934082, "learning_rate": 2.7484242771584156e-06, "loss": 0.4828, "step": 15726 }, { "epoch": 0.6592540582459155, "grad_norm": 2.201078414916992, "learning_rate": 2.7478181712787354e-06, "loss": 0.4495, "step": 15727 }, { "epoch": 0.6592959768609246, "grad_norm": 1.9020931720733643, "learning_rate": 2.747212106914916e-06, "loss": 0.4604, "step": 15728 }, { "epoch": 0.6593378954759335, "grad_norm": 1.8386998176574707, "learning_rate": 2.746606084078125e-06, "loss": 0.5204, "step": 15729 }, { "epoch": 0.6593798140909425, "grad_norm": 2.0623016357421875, "learning_rate": 2.7460001027795356e-06, "loss": 0.5184, "step": 15730 }, { "epoch": 0.6594217327059514, "grad_norm": 1.962545394897461, "learning_rate": 2.74539416303032e-06, "loss": 0.5304, "step": 15731 }, { "epoch": 0.6594636513209604, "grad_norm": 1.7720460891723633, "learning_rate": 2.7447882648416447e-06, "loss": 0.4627, "step": 15732 }, { "epoch": 0.6595055699359693, "grad_norm": 1.7632296085357666, "learning_rate": 2.744182408224681e-06, "loss": 0.5406, "step": 15733 }, { "epoch": 0.6595474885509782, "grad_norm": 1.645767092704773, "learning_rate": 2.7435765931905943e-06, "loss": 0.4634, "step": 15734 }, { "epoch": 0.6595894071659872, "grad_norm": 1.601854681968689, "learning_rate": 2.7429708197505563e-06, "loss": 0.5092, "step": 15735 }, { "epoch": 0.6596313257809961, "grad_norm": 1.5238553285598755, "learning_rate": 2.7423650879157306e-06, "loss": 0.4506, "step": 15736 }, { "epoch": 0.6596732443960052, "grad_norm": 1.6659773588180542, "learning_rate": 2.741759397697281e-06, "loss": 0.5118, "step": 15737 }, { "epoch": 0.6597151630110141, "grad_norm": 1.731827735900879, "learning_rate": 2.7411537491063754e-06, "loss": 0.4874, "step": 15738 }, { "epoch": 0.6597570816260231, "grad_norm": 1.6430021524429321, "learning_rate": 2.74054814215418e-06, "loss": 0.4941, "step": 15739 }, { "epoch": 0.659799000241032, "grad_norm": 2.6951818466186523, "learning_rate": 2.739942576851854e-06, "loss": 0.478, "step": 15740 }, { "epoch": 0.659840918856041, "grad_norm": 2.3219947814941406, "learning_rate": 2.7393370532105616e-06, "loss": 0.456, "step": 15741 }, { "epoch": 0.6598828374710499, "grad_norm": 1.7652345895767212, "learning_rate": 2.7387315712414675e-06, "loss": 0.4612, "step": 15742 }, { "epoch": 0.6599247560860589, "grad_norm": 1.5414087772369385, "learning_rate": 2.7381261309557287e-06, "loss": 0.4413, "step": 15743 }, { "epoch": 0.6599666747010678, "grad_norm": 1.8018531799316406, "learning_rate": 2.7375207323645103e-06, "loss": 0.5195, "step": 15744 }, { "epoch": 0.6600085933160769, "grad_norm": 1.8026267290115356, "learning_rate": 2.7369153754789666e-06, "loss": 0.4553, "step": 15745 }, { "epoch": 0.6600505119310858, "grad_norm": 2.618229389190674, "learning_rate": 2.7363100603102623e-06, "loss": 0.532, "step": 15746 }, { "epoch": 0.6600924305460948, "grad_norm": 1.7625038623809814, "learning_rate": 2.7357047868695497e-06, "loss": 0.4808, "step": 15747 }, { "epoch": 0.6601343491611037, "grad_norm": 1.822464108467102, "learning_rate": 2.7350995551679917e-06, "loss": 0.4795, "step": 15748 }, { "epoch": 0.6601762677761127, "grad_norm": 1.8121991157531738, "learning_rate": 2.73449436521674e-06, "loss": 0.5172, "step": 15749 }, { "epoch": 0.6602181863911216, "grad_norm": 1.751409888267517, "learning_rate": 2.7338892170269525e-06, "loss": 0.5034, "step": 15750 }, { "epoch": 0.6602601050061306, "grad_norm": 1.833011507987976, "learning_rate": 2.733284110609786e-06, "loss": 0.4728, "step": 15751 }, { "epoch": 0.6603020236211395, "grad_norm": 2.076258420944214, "learning_rate": 2.7326790459763925e-06, "loss": 0.4991, "step": 15752 }, { "epoch": 0.6603439422361486, "grad_norm": 1.6198163032531738, "learning_rate": 2.732074023137926e-06, "loss": 0.4389, "step": 15753 }, { "epoch": 0.6603858608511575, "grad_norm": 1.9091194868087769, "learning_rate": 2.7314690421055413e-06, "loss": 0.4941, "step": 15754 }, { "epoch": 0.6604277794661665, "grad_norm": 2.303708553314209, "learning_rate": 2.7308641028903883e-06, "loss": 0.4444, "step": 15755 }, { "epoch": 0.6604696980811754, "grad_norm": 1.7767162322998047, "learning_rate": 2.730259205503617e-06, "loss": 0.5139, "step": 15756 }, { "epoch": 0.6605116166961844, "grad_norm": 1.9604531526565552, "learning_rate": 2.7296543499563817e-06, "loss": 0.5229, "step": 15757 }, { "epoch": 0.6605535353111933, "grad_norm": 1.7717705965042114, "learning_rate": 2.7290495362598267e-06, "loss": 0.4614, "step": 15758 }, { "epoch": 0.6605954539262022, "grad_norm": 1.9921462535858154, "learning_rate": 2.7284447644251057e-06, "loss": 0.4973, "step": 15759 }, { "epoch": 0.6606373725412112, "grad_norm": 1.892340898513794, "learning_rate": 2.727840034463366e-06, "loss": 0.4844, "step": 15760 }, { "epoch": 0.6606792911562201, "grad_norm": 1.819940209388733, "learning_rate": 2.727235346385752e-06, "loss": 0.5106, "step": 15761 }, { "epoch": 0.6607212097712292, "grad_norm": 1.7694716453552246, "learning_rate": 2.726630700203413e-06, "loss": 0.5235, "step": 15762 }, { "epoch": 0.6607631283862381, "grad_norm": 1.7278410196304321, "learning_rate": 2.7260260959274953e-06, "loss": 0.5042, "step": 15763 }, { "epoch": 0.6608050470012471, "grad_norm": 1.605568528175354, "learning_rate": 2.7254215335691437e-06, "loss": 0.4968, "step": 15764 }, { "epoch": 0.660846965616256, "grad_norm": 1.7693604230880737, "learning_rate": 2.7248170131394995e-06, "loss": 0.4862, "step": 15765 }, { "epoch": 0.660888884231265, "grad_norm": 1.7658686637878418, "learning_rate": 2.7242125346497107e-06, "loss": 0.4674, "step": 15766 }, { "epoch": 0.6609308028462739, "grad_norm": 1.9425179958343506, "learning_rate": 2.7236080981109147e-06, "loss": 0.5025, "step": 15767 }, { "epoch": 0.6609727214612829, "grad_norm": 1.7352131605148315, "learning_rate": 2.723003703534257e-06, "loss": 0.4345, "step": 15768 }, { "epoch": 0.6610146400762918, "grad_norm": 2.23470139503479, "learning_rate": 2.72239935093088e-06, "loss": 0.4562, "step": 15769 }, { "epoch": 0.6610565586913009, "grad_norm": 1.98737370967865, "learning_rate": 2.7217950403119196e-06, "loss": 0.4837, "step": 15770 }, { "epoch": 0.6610984773063098, "grad_norm": 1.8315072059631348, "learning_rate": 2.7211907716885187e-06, "loss": 0.4785, "step": 15771 }, { "epoch": 0.6611403959213188, "grad_norm": 2.942559003829956, "learning_rate": 2.7205865450718183e-06, "loss": 0.4639, "step": 15772 }, { "epoch": 0.6611823145363277, "grad_norm": 1.741012454032898, "learning_rate": 2.719982360472953e-06, "loss": 0.4615, "step": 15773 }, { "epoch": 0.6612242331513367, "grad_norm": 1.8754031658172607, "learning_rate": 2.7193782179030593e-06, "loss": 0.4838, "step": 15774 }, { "epoch": 0.6612661517663456, "grad_norm": 2.9279463291168213, "learning_rate": 2.718774117373276e-06, "loss": 0.4857, "step": 15775 }, { "epoch": 0.6613080703813546, "grad_norm": 1.6675466299057007, "learning_rate": 2.718170058894738e-06, "loss": 0.4921, "step": 15776 }, { "epoch": 0.6613499889963635, "grad_norm": 1.652462124824524, "learning_rate": 2.717566042478579e-06, "loss": 0.4782, "step": 15777 }, { "epoch": 0.6613919076113726, "grad_norm": 1.7627367973327637, "learning_rate": 2.716962068135937e-06, "loss": 0.4668, "step": 15778 }, { "epoch": 0.6614338262263815, "grad_norm": 1.8100578784942627, "learning_rate": 2.716358135877941e-06, "loss": 0.4818, "step": 15779 }, { "epoch": 0.6614757448413905, "grad_norm": 1.6214576959609985, "learning_rate": 2.715754245715726e-06, "loss": 0.4521, "step": 15780 }, { "epoch": 0.6615176634563994, "grad_norm": 2.7531278133392334, "learning_rate": 2.715150397660426e-06, "loss": 0.5336, "step": 15781 }, { "epoch": 0.6615595820714084, "grad_norm": 1.8977036476135254, "learning_rate": 2.7145465917231693e-06, "loss": 0.5713, "step": 15782 }, { "epoch": 0.6616015006864173, "grad_norm": 2.048222541809082, "learning_rate": 2.713942827915085e-06, "loss": 0.463, "step": 15783 }, { "epoch": 0.6616434193014262, "grad_norm": 1.7961256504058838, "learning_rate": 2.7133391062473054e-06, "loss": 0.501, "step": 15784 }, { "epoch": 0.6616853379164352, "grad_norm": 1.668588399887085, "learning_rate": 2.712735426730957e-06, "loss": 0.4652, "step": 15785 }, { "epoch": 0.6617272565314442, "grad_norm": 1.7983819246292114, "learning_rate": 2.712131789377168e-06, "loss": 0.4955, "step": 15786 }, { "epoch": 0.6617691751464532, "grad_norm": 2.052449941635132, "learning_rate": 2.7115281941970696e-06, "loss": 0.4945, "step": 15787 }, { "epoch": 0.6618110937614621, "grad_norm": 2.1453070640563965, "learning_rate": 2.7109246412017824e-06, "loss": 0.519, "step": 15788 }, { "epoch": 0.6618530123764711, "grad_norm": 1.9554938077926636, "learning_rate": 2.710321130402436e-06, "loss": 0.4899, "step": 15789 }, { "epoch": 0.66189493099148, "grad_norm": 2.080838441848755, "learning_rate": 2.709717661810155e-06, "loss": 0.4553, "step": 15790 }, { "epoch": 0.661936849606489, "grad_norm": 1.8800278902053833, "learning_rate": 2.7091142354360633e-06, "loss": 0.4957, "step": 15791 }, { "epoch": 0.6619787682214979, "grad_norm": 1.8328495025634766, "learning_rate": 2.708510851291281e-06, "loss": 0.4839, "step": 15792 }, { "epoch": 0.6620206868365069, "grad_norm": 2.2357866764068604, "learning_rate": 2.7079075093869356e-06, "loss": 0.49, "step": 15793 }, { "epoch": 0.6620626054515159, "grad_norm": 1.892092227935791, "learning_rate": 2.7073042097341453e-06, "loss": 0.5162, "step": 15794 }, { "epoch": 0.6621045240665249, "grad_norm": 1.9523848295211792, "learning_rate": 2.7067009523440312e-06, "loss": 0.5795, "step": 15795 }, { "epoch": 0.6621464426815338, "grad_norm": 1.6901357173919678, "learning_rate": 2.7060977372277176e-06, "loss": 0.4962, "step": 15796 }, { "epoch": 0.6621883612965428, "grad_norm": 1.7235809564590454, "learning_rate": 2.705494564396318e-06, "loss": 0.5404, "step": 15797 }, { "epoch": 0.6622302799115517, "grad_norm": 1.623225212097168, "learning_rate": 2.704891433860955e-06, "loss": 0.4783, "step": 15798 }, { "epoch": 0.6622721985265607, "grad_norm": 3.3433241844177246, "learning_rate": 2.7042883456327472e-06, "loss": 0.5723, "step": 15799 }, { "epoch": 0.6623141171415696, "grad_norm": 2.5829107761383057, "learning_rate": 2.703685299722808e-06, "loss": 0.5387, "step": 15800 }, { "epoch": 0.6623560357565786, "grad_norm": 1.5547336339950562, "learning_rate": 2.7030822961422576e-06, "loss": 0.4776, "step": 15801 }, { "epoch": 0.6623979543715875, "grad_norm": 1.5612071752548218, "learning_rate": 2.7024793349022093e-06, "loss": 0.4449, "step": 15802 }, { "epoch": 0.6624398729865966, "grad_norm": 1.8191277980804443, "learning_rate": 2.701876416013777e-06, "loss": 0.4757, "step": 15803 }, { "epoch": 0.6624817916016055, "grad_norm": 1.7431097030639648, "learning_rate": 2.7012735394880758e-06, "loss": 0.5059, "step": 15804 }, { "epoch": 0.6625237102166145, "grad_norm": 1.8988384008407593, "learning_rate": 2.7006707053362206e-06, "loss": 0.4905, "step": 15805 }, { "epoch": 0.6625656288316234, "grad_norm": 1.5390739440917969, "learning_rate": 2.7000679135693207e-06, "loss": 0.5102, "step": 15806 }, { "epoch": 0.6626075474466324, "grad_norm": 1.833784818649292, "learning_rate": 2.6994651641984903e-06, "loss": 0.5136, "step": 15807 }, { "epoch": 0.6626494660616413, "grad_norm": 1.8480007648468018, "learning_rate": 2.698862457234839e-06, "loss": 0.4259, "step": 15808 }, { "epoch": 0.6626913846766502, "grad_norm": 1.6966662406921387, "learning_rate": 2.698259792689475e-06, "loss": 0.4704, "step": 15809 }, { "epoch": 0.6627333032916592, "grad_norm": 1.9943066835403442, "learning_rate": 2.697657170573513e-06, "loss": 0.5132, "step": 15810 }, { "epoch": 0.6627752219066682, "grad_norm": 1.6449122428894043, "learning_rate": 2.6970545908980576e-06, "loss": 0.4646, "step": 15811 }, { "epoch": 0.6628171405216772, "grad_norm": 1.7799946069717407, "learning_rate": 2.6964520536742157e-06, "loss": 0.5152, "step": 15812 }, { "epoch": 0.6628590591366861, "grad_norm": 1.8768290281295776, "learning_rate": 2.6958495589130974e-06, "loss": 0.4726, "step": 15813 }, { "epoch": 0.6629009777516951, "grad_norm": 1.6509953737258911, "learning_rate": 2.6952471066258045e-06, "loss": 0.5089, "step": 15814 }, { "epoch": 0.662942896366704, "grad_norm": 1.6028376817703247, "learning_rate": 2.694644696823445e-06, "loss": 0.4543, "step": 15815 }, { "epoch": 0.662984814981713, "grad_norm": 1.837530493736267, "learning_rate": 2.694042329517126e-06, "loss": 0.4684, "step": 15816 }, { "epoch": 0.6630267335967219, "grad_norm": 2.7722020149230957, "learning_rate": 2.6934400047179466e-06, "loss": 0.4727, "step": 15817 }, { "epoch": 0.663068652211731, "grad_norm": 1.688675880432129, "learning_rate": 2.6928377224370115e-06, "loss": 0.4524, "step": 15818 }, { "epoch": 0.6631105708267399, "grad_norm": 1.7884498834609985, "learning_rate": 2.6922354826854263e-06, "loss": 0.5144, "step": 15819 }, { "epoch": 0.6631524894417489, "grad_norm": 1.7649198770523071, "learning_rate": 2.6916332854742895e-06, "loss": 0.5127, "step": 15820 }, { "epoch": 0.6631944080567578, "grad_norm": 1.8761531114578247, "learning_rate": 2.6910311308147e-06, "loss": 0.4513, "step": 15821 }, { "epoch": 0.6632363266717668, "grad_norm": 1.7654587030410767, "learning_rate": 2.6904290187177606e-06, "loss": 0.4852, "step": 15822 }, { "epoch": 0.6632782452867757, "grad_norm": 2.206480026245117, "learning_rate": 2.6898269491945676e-06, "loss": 0.4827, "step": 15823 }, { "epoch": 0.6633201639017847, "grad_norm": 1.9261293411254883, "learning_rate": 2.6892249222562215e-06, "loss": 0.513, "step": 15824 }, { "epoch": 0.6633620825167936, "grad_norm": 3.4632437229156494, "learning_rate": 2.688622937913821e-06, "loss": 0.4757, "step": 15825 }, { "epoch": 0.6634040011318026, "grad_norm": 1.7571781873703003, "learning_rate": 2.68802099617846e-06, "loss": 0.4747, "step": 15826 }, { "epoch": 0.6634459197468116, "grad_norm": 1.646996021270752, "learning_rate": 2.6874190970612346e-06, "loss": 0.4512, "step": 15827 }, { "epoch": 0.6634878383618206, "grad_norm": 1.629557490348816, "learning_rate": 2.686817240573244e-06, "loss": 0.4652, "step": 15828 }, { "epoch": 0.6635297569768295, "grad_norm": 2.331118106842041, "learning_rate": 2.686215426725579e-06, "loss": 0.4817, "step": 15829 }, { "epoch": 0.6635716755918385, "grad_norm": 10.591339111328125, "learning_rate": 2.6856136555293327e-06, "loss": 0.4216, "step": 15830 }, { "epoch": 0.6636135942068474, "grad_norm": 1.7249990701675415, "learning_rate": 2.685011926995601e-06, "loss": 0.5634, "step": 15831 }, { "epoch": 0.6636555128218564, "grad_norm": 1.9577653408050537, "learning_rate": 2.6844102411354718e-06, "loss": 0.4537, "step": 15832 }, { "epoch": 0.6636974314368653, "grad_norm": 1.7637399435043335, "learning_rate": 2.683808597960038e-06, "loss": 0.4889, "step": 15833 }, { "epoch": 0.6637393500518742, "grad_norm": 1.8848048448562622, "learning_rate": 2.6832069974803933e-06, "loss": 0.4717, "step": 15834 }, { "epoch": 0.6637812686668833, "grad_norm": 1.6061140298843384, "learning_rate": 2.682605439707623e-06, "loss": 0.4547, "step": 15835 }, { "epoch": 0.6638231872818922, "grad_norm": 2.0773634910583496, "learning_rate": 2.6820039246528173e-06, "loss": 0.4872, "step": 15836 }, { "epoch": 0.6638651058969012, "grad_norm": 1.9843584299087524, "learning_rate": 2.6814024523270664e-06, "loss": 0.4319, "step": 15837 }, { "epoch": 0.6639070245119101, "grad_norm": 1.7502775192260742, "learning_rate": 2.6808010227414566e-06, "loss": 0.4867, "step": 15838 }, { "epoch": 0.6639489431269191, "grad_norm": 1.998113989830017, "learning_rate": 2.6801996359070713e-06, "loss": 0.4967, "step": 15839 }, { "epoch": 0.663990861741928, "grad_norm": 1.6590546369552612, "learning_rate": 2.6795982918350012e-06, "loss": 0.5149, "step": 15840 }, { "epoch": 0.664032780356937, "grad_norm": 2.0359771251678467, "learning_rate": 2.678996990536327e-06, "loss": 0.4953, "step": 15841 }, { "epoch": 0.6640746989719459, "grad_norm": 2.157705783843994, "learning_rate": 2.6783957320221332e-06, "loss": 0.5054, "step": 15842 }, { "epoch": 0.664116617586955, "grad_norm": 2.8443000316619873, "learning_rate": 2.6777945163035076e-06, "loss": 0.4876, "step": 15843 }, { "epoch": 0.6641585362019639, "grad_norm": 2.0534627437591553, "learning_rate": 2.6771933433915275e-06, "loss": 0.4879, "step": 15844 }, { "epoch": 0.6642004548169729, "grad_norm": 1.7595851421356201, "learning_rate": 2.6765922132972767e-06, "loss": 0.5022, "step": 15845 }, { "epoch": 0.6642423734319818, "grad_norm": 1.642661452293396, "learning_rate": 2.6759911260318384e-06, "loss": 0.5546, "step": 15846 }, { "epoch": 0.6642842920469908, "grad_norm": 1.9075615406036377, "learning_rate": 2.6753900816062894e-06, "loss": 0.4559, "step": 15847 }, { "epoch": 0.6643262106619997, "grad_norm": 1.767184853553772, "learning_rate": 2.674789080031712e-06, "loss": 0.4494, "step": 15848 }, { "epoch": 0.6643681292770087, "grad_norm": 1.7285921573638916, "learning_rate": 2.6741881213191834e-06, "loss": 0.5011, "step": 15849 }, { "epoch": 0.6644100478920176, "grad_norm": 1.833156704902649, "learning_rate": 2.67358720547978e-06, "loss": 0.5373, "step": 15850 }, { "epoch": 0.6644519665070266, "grad_norm": 1.9587723016738892, "learning_rate": 2.67298633252458e-06, "loss": 0.5026, "step": 15851 }, { "epoch": 0.6644938851220356, "grad_norm": 2.0228750705718994, "learning_rate": 2.6723855024646617e-06, "loss": 0.534, "step": 15852 }, { "epoch": 0.6645358037370446, "grad_norm": 1.5763881206512451, "learning_rate": 2.6717847153110975e-06, "loss": 0.4477, "step": 15853 }, { "epoch": 0.6645777223520535, "grad_norm": 2.3030712604522705, "learning_rate": 2.671183971074963e-06, "loss": 0.4902, "step": 15854 }, { "epoch": 0.6646196409670625, "grad_norm": 2.833099126815796, "learning_rate": 2.670583269767334e-06, "loss": 0.5396, "step": 15855 }, { "epoch": 0.6646615595820714, "grad_norm": 2.0440480709075928, "learning_rate": 2.669982611399281e-06, "loss": 0.463, "step": 15856 }, { "epoch": 0.6647034781970804, "grad_norm": 1.7453827857971191, "learning_rate": 2.669381995981879e-06, "loss": 0.5259, "step": 15857 }, { "epoch": 0.6647453968120893, "grad_norm": 4.288143157958984, "learning_rate": 2.668781423526198e-06, "loss": 0.5577, "step": 15858 }, { "epoch": 0.6647873154270982, "grad_norm": 2.1123642921447754, "learning_rate": 2.6681808940433063e-06, "loss": 0.4828, "step": 15859 }, { "epoch": 0.6648292340421073, "grad_norm": 1.7092124223709106, "learning_rate": 2.667580407544277e-06, "loss": 0.4692, "step": 15860 }, { "epoch": 0.6648711526571162, "grad_norm": 1.685520052909851, "learning_rate": 2.6669799640401796e-06, "loss": 0.4594, "step": 15861 }, { "epoch": 0.6649130712721252, "grad_norm": 1.8287557363510132, "learning_rate": 2.6663795635420796e-06, "loss": 0.4592, "step": 15862 }, { "epoch": 0.6649549898871341, "grad_norm": 1.6095610857009888, "learning_rate": 2.665779206061046e-06, "loss": 0.5019, "step": 15863 }, { "epoch": 0.6649969085021431, "grad_norm": 1.9175710678100586, "learning_rate": 2.6651788916081477e-06, "loss": 0.5067, "step": 15864 }, { "epoch": 0.665038827117152, "grad_norm": 2.9937219619750977, "learning_rate": 2.664578620194447e-06, "loss": 0.5143, "step": 15865 }, { "epoch": 0.665080745732161, "grad_norm": 1.8011143207550049, "learning_rate": 2.6639783918310125e-06, "loss": 0.4359, "step": 15866 }, { "epoch": 0.6651226643471699, "grad_norm": 1.7899531126022339, "learning_rate": 2.663378206528907e-06, "loss": 0.502, "step": 15867 }, { "epoch": 0.665164582962179, "grad_norm": 1.5601446628570557, "learning_rate": 2.662778064299193e-06, "loss": 0.4506, "step": 15868 }, { "epoch": 0.6652065015771879, "grad_norm": 2.416351556777954, "learning_rate": 2.662177965152933e-06, "loss": 0.4414, "step": 15869 }, { "epoch": 0.6652484201921969, "grad_norm": 1.627978801727295, "learning_rate": 2.6615779091011928e-06, "loss": 0.4491, "step": 15870 }, { "epoch": 0.6652903388072058, "grad_norm": 1.9015346765518188, "learning_rate": 2.6609778961550285e-06, "loss": 0.5034, "step": 15871 }, { "epoch": 0.6653322574222148, "grad_norm": 1.6841617822647095, "learning_rate": 2.6603779263255035e-06, "loss": 0.5001, "step": 15872 }, { "epoch": 0.6653741760372237, "grad_norm": 2.0490214824676514, "learning_rate": 2.6597779996236795e-06, "loss": 0.4787, "step": 15873 }, { "epoch": 0.6654160946522327, "grad_norm": 1.9584622383117676, "learning_rate": 2.65917811606061e-06, "loss": 0.4348, "step": 15874 }, { "epoch": 0.6654580132672416, "grad_norm": 1.722413182258606, "learning_rate": 2.6585782756473584e-06, "loss": 0.4664, "step": 15875 }, { "epoch": 0.6654999318822507, "grad_norm": 1.8365819454193115, "learning_rate": 2.657978478394979e-06, "loss": 0.4949, "step": 15876 }, { "epoch": 0.6655418504972596, "grad_norm": 1.8143996000289917, "learning_rate": 2.6573787243145267e-06, "loss": 0.4837, "step": 15877 }, { "epoch": 0.6655837691122686, "grad_norm": 1.9230997562408447, "learning_rate": 2.656779013417059e-06, "loss": 0.5005, "step": 15878 }, { "epoch": 0.6656256877272775, "grad_norm": 2.6126797199249268, "learning_rate": 2.6561793457136333e-06, "loss": 0.4758, "step": 15879 }, { "epoch": 0.6656676063422865, "grad_norm": 1.7993602752685547, "learning_rate": 2.655579721215299e-06, "loss": 0.5004, "step": 15880 }, { "epoch": 0.6657095249572954, "grad_norm": 1.9143564701080322, "learning_rate": 2.6549801399331127e-06, "loss": 0.4613, "step": 15881 }, { "epoch": 0.6657514435723044, "grad_norm": 1.6818139553070068, "learning_rate": 2.654380601878124e-06, "loss": 0.5311, "step": 15882 }, { "epoch": 0.6657933621873133, "grad_norm": 1.8955907821655273, "learning_rate": 2.6537811070613863e-06, "loss": 0.591, "step": 15883 }, { "epoch": 0.6658352808023222, "grad_norm": 2.020106792449951, "learning_rate": 2.6531816554939527e-06, "loss": 0.5024, "step": 15884 }, { "epoch": 0.6658771994173313, "grad_norm": 1.5080674886703491, "learning_rate": 2.65258224718687e-06, "loss": 0.4541, "step": 15885 }, { "epoch": 0.6659191180323402, "grad_norm": 2.0484371185302734, "learning_rate": 2.6519828821511875e-06, "loss": 0.4785, "step": 15886 }, { "epoch": 0.6659610366473492, "grad_norm": 1.8256340026855469, "learning_rate": 2.6513835603979554e-06, "loss": 0.5297, "step": 15887 }, { "epoch": 0.6660029552623581, "grad_norm": 1.5942984819412231, "learning_rate": 2.650784281938219e-06, "loss": 0.4515, "step": 15888 }, { "epoch": 0.6660448738773671, "grad_norm": 3.599663019180298, "learning_rate": 2.650185046783027e-06, "loss": 0.4621, "step": 15889 }, { "epoch": 0.666086792492376, "grad_norm": 1.8690038919448853, "learning_rate": 2.6495858549434267e-06, "loss": 0.4976, "step": 15890 }, { "epoch": 0.666128711107385, "grad_norm": 2.407766580581665, "learning_rate": 2.64898670643046e-06, "loss": 0.5159, "step": 15891 }, { "epoch": 0.6661706297223939, "grad_norm": 2.203554391860962, "learning_rate": 2.6483876012551735e-06, "loss": 0.5351, "step": 15892 }, { "epoch": 0.666212548337403, "grad_norm": 1.7415491342544556, "learning_rate": 2.6477885394286117e-06, "loss": 0.4717, "step": 15893 }, { "epoch": 0.6662544669524119, "grad_norm": 2.904093027114868, "learning_rate": 2.6471895209618147e-06, "loss": 0.4563, "step": 15894 }, { "epoch": 0.6662963855674209, "grad_norm": 1.6433300971984863, "learning_rate": 2.6465905458658287e-06, "loss": 0.4809, "step": 15895 }, { "epoch": 0.6663383041824298, "grad_norm": 1.7708672285079956, "learning_rate": 2.645991614151692e-06, "loss": 0.4499, "step": 15896 }, { "epoch": 0.6663802227974388, "grad_norm": 1.6205796003341675, "learning_rate": 2.645392725830444e-06, "loss": 0.5096, "step": 15897 }, { "epoch": 0.6664221414124477, "grad_norm": 2.9544036388397217, "learning_rate": 2.6447938809131256e-06, "loss": 0.4336, "step": 15898 }, { "epoch": 0.6664640600274567, "grad_norm": 1.5501024723052979, "learning_rate": 2.6441950794107784e-06, "loss": 0.4186, "step": 15899 }, { "epoch": 0.6665059786424656, "grad_norm": 1.6997885704040527, "learning_rate": 2.643596321334436e-06, "loss": 0.4757, "step": 15900 }, { "epoch": 0.6665478972574747, "grad_norm": 1.726485252380371, "learning_rate": 2.642997606695137e-06, "loss": 0.4468, "step": 15901 }, { "epoch": 0.6665898158724836, "grad_norm": 1.7779487371444702, "learning_rate": 2.6423989355039214e-06, "loss": 0.44, "step": 15902 }, { "epoch": 0.6666317344874926, "grad_norm": 2.068476915359497, "learning_rate": 2.64180030777182e-06, "loss": 0.4737, "step": 15903 }, { "epoch": 0.6666736531025015, "grad_norm": 1.7861806154251099, "learning_rate": 2.6412017235098717e-06, "loss": 0.4697, "step": 15904 }, { "epoch": 0.6667155717175105, "grad_norm": 1.6835241317749023, "learning_rate": 2.640603182729109e-06, "loss": 0.4168, "step": 15905 }, { "epoch": 0.6667574903325194, "grad_norm": 2.4933242797851562, "learning_rate": 2.6400046854405627e-06, "loss": 0.4643, "step": 15906 }, { "epoch": 0.6667994089475284, "grad_norm": 1.7166129350662231, "learning_rate": 2.639406231655267e-06, "loss": 0.4669, "step": 15907 }, { "epoch": 0.6668413275625373, "grad_norm": 1.7971094846725464, "learning_rate": 2.6388078213842573e-06, "loss": 0.486, "step": 15908 }, { "epoch": 0.6668832461775462, "grad_norm": 1.6960467100143433, "learning_rate": 2.638209454638558e-06, "loss": 0.5203, "step": 15909 }, { "epoch": 0.6669251647925553, "grad_norm": 2.0682320594787598, "learning_rate": 2.6376111314292024e-06, "loss": 0.5023, "step": 15910 }, { "epoch": 0.6669670834075642, "grad_norm": 1.818707823753357, "learning_rate": 2.637012851767222e-06, "loss": 0.5144, "step": 15911 }, { "epoch": 0.6670090020225732, "grad_norm": 1.4178091287612915, "learning_rate": 2.6364146156636406e-06, "loss": 0.4776, "step": 15912 }, { "epoch": 0.6670509206375821, "grad_norm": 1.8149758577346802, "learning_rate": 2.6358164231294904e-06, "loss": 0.4529, "step": 15913 }, { "epoch": 0.6670928392525911, "grad_norm": 1.6387262344360352, "learning_rate": 2.635218274175796e-06, "loss": 0.4722, "step": 15914 }, { "epoch": 0.6671347578676, "grad_norm": 1.570873737335205, "learning_rate": 2.634620168813582e-06, "loss": 0.5036, "step": 15915 }, { "epoch": 0.667176676482609, "grad_norm": 1.5108267068862915, "learning_rate": 2.6340221070538746e-06, "loss": 0.419, "step": 15916 }, { "epoch": 0.6672185950976179, "grad_norm": 1.8656220436096191, "learning_rate": 2.633424088907701e-06, "loss": 0.4864, "step": 15917 }, { "epoch": 0.667260513712627, "grad_norm": 3.7205088138580322, "learning_rate": 2.6328261143860802e-06, "loss": 0.4503, "step": 15918 }, { "epoch": 0.6673024323276359, "grad_norm": 1.8847332000732422, "learning_rate": 2.632228183500038e-06, "loss": 0.4478, "step": 15919 }, { "epoch": 0.6673443509426449, "grad_norm": 1.6342015266418457, "learning_rate": 2.631630296260598e-06, "loss": 0.5006, "step": 15920 }, { "epoch": 0.6673862695576538, "grad_norm": 1.722564935684204, "learning_rate": 2.631032452678777e-06, "loss": 0.5171, "step": 15921 }, { "epoch": 0.6674281881726628, "grad_norm": 1.927829623222351, "learning_rate": 2.6304346527655994e-06, "loss": 0.4758, "step": 15922 }, { "epoch": 0.6674701067876717, "grad_norm": 1.6224002838134766, "learning_rate": 2.629836896532083e-06, "loss": 0.4761, "step": 15923 }, { "epoch": 0.6675120254026807, "grad_norm": 1.6128318309783936, "learning_rate": 2.629239183989246e-06, "loss": 0.4715, "step": 15924 }, { "epoch": 0.6675539440176896, "grad_norm": 1.8134105205535889, "learning_rate": 2.628641515148106e-06, "loss": 0.4727, "step": 15925 }, { "epoch": 0.6675958626326987, "grad_norm": 2.0313634872436523, "learning_rate": 2.6280438900196836e-06, "loss": 0.4972, "step": 15926 }, { "epoch": 0.6676377812477076, "grad_norm": 1.5922725200653076, "learning_rate": 2.6274463086149915e-06, "loss": 0.4595, "step": 15927 }, { "epoch": 0.6676796998627166, "grad_norm": 2.6515889167785645, "learning_rate": 2.626848770945046e-06, "loss": 0.5257, "step": 15928 }, { "epoch": 0.6677216184777255, "grad_norm": 1.8931692838668823, "learning_rate": 2.6262512770208653e-06, "loss": 0.4655, "step": 15929 }, { "epoch": 0.6677635370927345, "grad_norm": 2.1812801361083984, "learning_rate": 2.6256538268534582e-06, "loss": 0.5414, "step": 15930 }, { "epoch": 0.6678054557077434, "grad_norm": 1.6597816944122314, "learning_rate": 2.625056420453843e-06, "loss": 0.5217, "step": 15931 }, { "epoch": 0.6678473743227524, "grad_norm": 1.7676854133605957, "learning_rate": 2.6244590578330286e-06, "loss": 0.4772, "step": 15932 }, { "epoch": 0.6678892929377613, "grad_norm": 1.9530378580093384, "learning_rate": 2.623861739002026e-06, "loss": 0.4844, "step": 15933 }, { "epoch": 0.6679312115527702, "grad_norm": 1.5490885972976685, "learning_rate": 2.6232644639718474e-06, "loss": 0.4805, "step": 15934 }, { "epoch": 0.6679731301677793, "grad_norm": 1.629682183265686, "learning_rate": 2.622667232753504e-06, "loss": 0.5335, "step": 15935 }, { "epoch": 0.6680150487827882, "grad_norm": 1.8840076923370361, "learning_rate": 2.622070045358002e-06, "loss": 0.4541, "step": 15936 }, { "epoch": 0.6680569673977972, "grad_norm": 1.5935864448547363, "learning_rate": 2.6214729017963512e-06, "loss": 0.4737, "step": 15937 }, { "epoch": 0.6680988860128061, "grad_norm": 2.736758232116699, "learning_rate": 2.6208758020795617e-06, "loss": 0.5101, "step": 15938 }, { "epoch": 0.6681408046278151, "grad_norm": 1.7738518714904785, "learning_rate": 2.620278746218635e-06, "loss": 0.4682, "step": 15939 }, { "epoch": 0.668182723242824, "grad_norm": 1.8634206056594849, "learning_rate": 2.619681734224583e-06, "loss": 0.4421, "step": 15940 }, { "epoch": 0.668224641857833, "grad_norm": 1.7246001958847046, "learning_rate": 2.619084766108404e-06, "loss": 0.5139, "step": 15941 }, { "epoch": 0.6682665604728419, "grad_norm": 1.8024848699569702, "learning_rate": 2.618487841881109e-06, "loss": 0.4797, "step": 15942 }, { "epoch": 0.668308479087851, "grad_norm": 2.417890787124634, "learning_rate": 2.617890961553695e-06, "loss": 0.5099, "step": 15943 }, { "epoch": 0.6683503977028599, "grad_norm": 1.7460018396377563, "learning_rate": 2.6172941251371707e-06, "loss": 0.4819, "step": 15944 }, { "epoch": 0.6683923163178689, "grad_norm": 1.926205039024353, "learning_rate": 2.6166973326425337e-06, "loss": 0.5024, "step": 15945 }, { "epoch": 0.6684342349328778, "grad_norm": 1.614038348197937, "learning_rate": 2.6161005840807864e-06, "loss": 0.4898, "step": 15946 }, { "epoch": 0.6684761535478868, "grad_norm": 1.871177315711975, "learning_rate": 2.6155038794629306e-06, "loss": 0.4954, "step": 15947 }, { "epoch": 0.6685180721628957, "grad_norm": 1.8534013032913208, "learning_rate": 2.6149072187999623e-06, "loss": 0.4857, "step": 15948 }, { "epoch": 0.6685599907779047, "grad_norm": 1.6726698875427246, "learning_rate": 2.614310602102883e-06, "loss": 0.4724, "step": 15949 }, { "epoch": 0.6686019093929136, "grad_norm": 1.8518750667572021, "learning_rate": 2.613714029382691e-06, "loss": 0.4613, "step": 15950 }, { "epoch": 0.6686438280079227, "grad_norm": 1.8210904598236084, "learning_rate": 2.6131175006503815e-06, "loss": 0.4622, "step": 15951 }, { "epoch": 0.6686857466229316, "grad_norm": 1.9976023435592651, "learning_rate": 2.61252101591695e-06, "loss": 0.5418, "step": 15952 }, { "epoch": 0.6687276652379406, "grad_norm": 1.890822410583496, "learning_rate": 2.611924575193395e-06, "loss": 0.4593, "step": 15953 }, { "epoch": 0.6687695838529495, "grad_norm": 1.5221608877182007, "learning_rate": 2.6113281784907063e-06, "loss": 0.476, "step": 15954 }, { "epoch": 0.6688115024679585, "grad_norm": 1.6156995296478271, "learning_rate": 2.6107318258198834e-06, "loss": 0.4744, "step": 15955 }, { "epoch": 0.6688534210829674, "grad_norm": 2.319661855697632, "learning_rate": 2.610135517191913e-06, "loss": 0.4931, "step": 15956 }, { "epoch": 0.6688953396979764, "grad_norm": 1.7850947380065918, "learning_rate": 2.609539252617791e-06, "loss": 0.4465, "step": 15957 }, { "epoch": 0.6689372583129853, "grad_norm": 1.626223087310791, "learning_rate": 2.6089430321085104e-06, "loss": 0.4951, "step": 15958 }, { "epoch": 0.6689791769279942, "grad_norm": 1.8834434747695923, "learning_rate": 2.6083468556750568e-06, "loss": 0.5059, "step": 15959 }, { "epoch": 0.6690210955430033, "grad_norm": 1.7369688749313354, "learning_rate": 2.607750723328425e-06, "loss": 0.4707, "step": 15960 }, { "epoch": 0.6690630141580122, "grad_norm": 1.5905498266220093, "learning_rate": 2.6071546350796017e-06, "loss": 0.4828, "step": 15961 }, { "epoch": 0.6691049327730212, "grad_norm": 1.6677613258361816, "learning_rate": 2.6065585909395722e-06, "loss": 0.4809, "step": 15962 }, { "epoch": 0.6691468513880301, "grad_norm": 1.744099736213684, "learning_rate": 2.6059625909193265e-06, "loss": 0.5126, "step": 15963 }, { "epoch": 0.6691887700030391, "grad_norm": 1.7571300268173218, "learning_rate": 2.6053666350298527e-06, "loss": 0.5086, "step": 15964 }, { "epoch": 0.669230688618048, "grad_norm": 1.88534677028656, "learning_rate": 2.604770723282133e-06, "loss": 0.5243, "step": 15965 }, { "epoch": 0.669272607233057, "grad_norm": 1.8684355020523071, "learning_rate": 2.6041748556871533e-06, "loss": 0.5179, "step": 15966 }, { "epoch": 0.6693145258480659, "grad_norm": 2.581465482711792, "learning_rate": 2.6035790322558998e-06, "loss": 0.5629, "step": 15967 }, { "epoch": 0.669356444463075, "grad_norm": 1.8775283098220825, "learning_rate": 2.602983252999352e-06, "loss": 0.5089, "step": 15968 }, { "epoch": 0.6693983630780839, "grad_norm": 1.8211816549301147, "learning_rate": 2.602387517928496e-06, "loss": 0.5151, "step": 15969 }, { "epoch": 0.6694402816930929, "grad_norm": 1.8759804964065552, "learning_rate": 2.601791827054312e-06, "loss": 0.5262, "step": 15970 }, { "epoch": 0.6694822003081018, "grad_norm": 1.8243156671524048, "learning_rate": 2.601196180387778e-06, "loss": 0.4579, "step": 15971 }, { "epoch": 0.6695241189231108, "grad_norm": 1.794249176979065, "learning_rate": 2.600600577939876e-06, "loss": 0.4813, "step": 15972 }, { "epoch": 0.6695660375381197, "grad_norm": 1.7649600505828857, "learning_rate": 2.6000050197215876e-06, "loss": 0.5116, "step": 15973 }, { "epoch": 0.6696079561531287, "grad_norm": 1.6683382987976074, "learning_rate": 2.599409505743886e-06, "loss": 0.4901, "step": 15974 }, { "epoch": 0.6696498747681376, "grad_norm": 1.7121329307556152, "learning_rate": 2.5988140360177526e-06, "loss": 0.4849, "step": 15975 }, { "epoch": 0.6696917933831467, "grad_norm": 1.6269152164459229, "learning_rate": 2.598218610554164e-06, "loss": 0.4579, "step": 15976 }, { "epoch": 0.6697337119981556, "grad_norm": 1.7587172985076904, "learning_rate": 2.5976232293640935e-06, "loss": 0.4865, "step": 15977 }, { "epoch": 0.6697756306131646, "grad_norm": 2.0868725776672363, "learning_rate": 2.597027892458519e-06, "loss": 0.4522, "step": 15978 }, { "epoch": 0.6698175492281735, "grad_norm": 2.1593375205993652, "learning_rate": 2.596432599848414e-06, "loss": 0.474, "step": 15979 }, { "epoch": 0.6698594678431825, "grad_norm": 1.7051712274551392, "learning_rate": 2.5958373515447488e-06, "loss": 0.4165, "step": 15980 }, { "epoch": 0.6699013864581914, "grad_norm": 1.9739078283309937, "learning_rate": 2.5952421475584986e-06, "loss": 0.4582, "step": 15981 }, { "epoch": 0.6699433050732004, "grad_norm": 3.6347594261169434, "learning_rate": 2.5946469879006366e-06, "loss": 0.4404, "step": 15982 }, { "epoch": 0.6699852236882093, "grad_norm": 1.6422739028930664, "learning_rate": 2.59405187258213e-06, "loss": 0.4474, "step": 15983 }, { "epoch": 0.6700271423032182, "grad_norm": 1.800872802734375, "learning_rate": 2.5934568016139516e-06, "loss": 0.5163, "step": 15984 }, { "epoch": 0.6700690609182273, "grad_norm": 1.6383228302001953, "learning_rate": 2.5928617750070716e-06, "loss": 0.4868, "step": 15985 }, { "epoch": 0.6701109795332362, "grad_norm": 1.8327751159667969, "learning_rate": 2.5922667927724555e-06, "loss": 0.5043, "step": 15986 }, { "epoch": 0.6701528981482452, "grad_norm": 1.9959074258804321, "learning_rate": 2.591671854921074e-06, "loss": 0.4397, "step": 15987 }, { "epoch": 0.6701948167632541, "grad_norm": 1.7625590562820435, "learning_rate": 2.59107696146389e-06, "loss": 0.4637, "step": 15988 }, { "epoch": 0.6702367353782631, "grad_norm": 1.8421965837478638, "learning_rate": 2.590482112411875e-06, "loss": 0.5052, "step": 15989 }, { "epoch": 0.670278653993272, "grad_norm": 7.543492794036865, "learning_rate": 2.589887307775989e-06, "loss": 0.5087, "step": 15990 }, { "epoch": 0.670320572608281, "grad_norm": 1.7305786609649658, "learning_rate": 2.5892925475672e-06, "loss": 0.476, "step": 15991 }, { "epoch": 0.67036249122329, "grad_norm": 1.8066850900650024, "learning_rate": 2.588697831796468e-06, "loss": 0.4219, "step": 15992 }, { "epoch": 0.670404409838299, "grad_norm": 1.559830904006958, "learning_rate": 2.5881031604747584e-06, "loss": 0.4625, "step": 15993 }, { "epoch": 0.6704463284533079, "grad_norm": 2.0717852115631104, "learning_rate": 2.5875085336130356e-06, "loss": 0.5068, "step": 15994 }, { "epoch": 0.6704882470683169, "grad_norm": 1.6869581937789917, "learning_rate": 2.5869139512222547e-06, "loss": 0.5262, "step": 15995 }, { "epoch": 0.6705301656833258, "grad_norm": 2.118271589279175, "learning_rate": 2.5863194133133795e-06, "loss": 0.4541, "step": 15996 }, { "epoch": 0.6705720842983348, "grad_norm": 1.9281108379364014, "learning_rate": 2.5857249198973712e-06, "loss": 0.4767, "step": 15997 }, { "epoch": 0.6706140029133437, "grad_norm": 1.6189526319503784, "learning_rate": 2.5851304709851855e-06, "loss": 0.4515, "step": 15998 }, { "epoch": 0.6706559215283527, "grad_norm": 2.1767289638519287, "learning_rate": 2.58453606658778e-06, "loss": 0.4895, "step": 15999 }, { "epoch": 0.6706978401433616, "grad_norm": 1.5767440795898438, "learning_rate": 2.5839417067161145e-06, "loss": 0.4401, "step": 16000 }, { "epoch": 0.6707397587583707, "grad_norm": 1.6690679788589478, "learning_rate": 2.583347391381142e-06, "loss": 0.5165, "step": 16001 }, { "epoch": 0.6707816773733796, "grad_norm": 1.9271414279937744, "learning_rate": 2.582753120593819e-06, "loss": 0.4914, "step": 16002 }, { "epoch": 0.6708235959883886, "grad_norm": 2.0071825981140137, "learning_rate": 2.5821588943651023e-06, "loss": 0.4227, "step": 16003 }, { "epoch": 0.6708655146033975, "grad_norm": 1.5617884397506714, "learning_rate": 2.581564712705942e-06, "loss": 0.4811, "step": 16004 }, { "epoch": 0.6709074332184065, "grad_norm": 2.032348155975342, "learning_rate": 2.5809705756272927e-06, "loss": 0.4299, "step": 16005 }, { "epoch": 0.6709493518334154, "grad_norm": 1.8337346315383911, "learning_rate": 2.5803764831401092e-06, "loss": 0.4441, "step": 16006 }, { "epoch": 0.6709912704484244, "grad_norm": 1.5775424242019653, "learning_rate": 2.5797824352553404e-06, "loss": 0.5012, "step": 16007 }, { "epoch": 0.6710331890634333, "grad_norm": 1.7400099039077759, "learning_rate": 2.5791884319839332e-06, "loss": 0.485, "step": 16008 }, { "epoch": 0.6710751076784423, "grad_norm": 1.7881975173950195, "learning_rate": 2.578594473336844e-06, "loss": 0.5124, "step": 16009 }, { "epoch": 0.6711170262934513, "grad_norm": 1.8302981853485107, "learning_rate": 2.578000559325015e-06, "loss": 0.4972, "step": 16010 }, { "epoch": 0.6711589449084602, "grad_norm": 1.9650145769119263, "learning_rate": 2.5774066899593988e-06, "loss": 0.5053, "step": 16011 }, { "epoch": 0.6712008635234692, "grad_norm": 2.357060194015503, "learning_rate": 2.5768128652509416e-06, "loss": 0.4798, "step": 16012 }, { "epoch": 0.6712427821384781, "grad_norm": 1.6187657117843628, "learning_rate": 2.5762190852105886e-06, "loss": 0.472, "step": 16013 }, { "epoch": 0.6712847007534871, "grad_norm": 1.7914766073226929, "learning_rate": 2.5756253498492847e-06, "loss": 0.5163, "step": 16014 }, { "epoch": 0.671326619368496, "grad_norm": 1.8033725023269653, "learning_rate": 2.575031659177979e-06, "loss": 0.4587, "step": 16015 }, { "epoch": 0.671368537983505, "grad_norm": 1.9292774200439453, "learning_rate": 2.574438013207611e-06, "loss": 0.4546, "step": 16016 }, { "epoch": 0.671410456598514, "grad_norm": 2.2847111225128174, "learning_rate": 2.5738444119491242e-06, "loss": 0.5316, "step": 16017 }, { "epoch": 0.671452375213523, "grad_norm": 1.7182590961456299, "learning_rate": 2.5732508554134627e-06, "loss": 0.4743, "step": 16018 }, { "epoch": 0.6714942938285319, "grad_norm": 1.9735451936721802, "learning_rate": 2.5726573436115653e-06, "loss": 0.5183, "step": 16019 }, { "epoch": 0.6715362124435409, "grad_norm": 2.043297052383423, "learning_rate": 2.5720638765543743e-06, "loss": 0.5481, "step": 16020 }, { "epoch": 0.6715781310585498, "grad_norm": 1.822648525238037, "learning_rate": 2.5714704542528313e-06, "loss": 0.4967, "step": 16021 }, { "epoch": 0.6716200496735588, "grad_norm": 1.837651014328003, "learning_rate": 2.5708770767178703e-06, "loss": 0.4402, "step": 16022 }, { "epoch": 0.6716619682885677, "grad_norm": 3.9280178546905518, "learning_rate": 2.5702837439604326e-06, "loss": 0.4759, "step": 16023 }, { "epoch": 0.6717038869035767, "grad_norm": 1.7571510076522827, "learning_rate": 2.569690455991457e-06, "loss": 0.5058, "step": 16024 }, { "epoch": 0.6717458055185856, "grad_norm": 1.603417992591858, "learning_rate": 2.5690972128218783e-06, "loss": 0.4432, "step": 16025 }, { "epoch": 0.6717877241335947, "grad_norm": 1.9206929206848145, "learning_rate": 2.5685040144626303e-06, "loss": 0.468, "step": 16026 }, { "epoch": 0.6718296427486036, "grad_norm": 1.9658299684524536, "learning_rate": 2.5679108609246504e-06, "loss": 0.5042, "step": 16027 }, { "epoch": 0.6718715613636126, "grad_norm": 1.7377983331680298, "learning_rate": 2.5673177522188703e-06, "loss": 0.5331, "step": 16028 }, { "epoch": 0.6719134799786215, "grad_norm": 2.14131236076355, "learning_rate": 2.566724688356224e-06, "loss": 0.4314, "step": 16029 }, { "epoch": 0.6719553985936305, "grad_norm": 1.6460330486297607, "learning_rate": 2.5661316693476463e-06, "loss": 0.4733, "step": 16030 }, { "epoch": 0.6719973172086394, "grad_norm": 1.7665212154388428, "learning_rate": 2.565538695204064e-06, "loss": 0.4714, "step": 16031 }, { "epoch": 0.6720392358236484, "grad_norm": 1.9599254131317139, "learning_rate": 2.564945765936413e-06, "loss": 0.5279, "step": 16032 }, { "epoch": 0.6720811544386573, "grad_norm": 1.7584865093231201, "learning_rate": 2.5643528815556185e-06, "loss": 0.4954, "step": 16033 }, { "epoch": 0.6721230730536663, "grad_norm": 4.297409534454346, "learning_rate": 2.5637600420726128e-06, "loss": 0.4541, "step": 16034 }, { "epoch": 0.6721649916686753, "grad_norm": 3.0309600830078125, "learning_rate": 2.5631672474983215e-06, "loss": 0.5364, "step": 16035 }, { "epoch": 0.6722069102836842, "grad_norm": 1.6710169315338135, "learning_rate": 2.5625744978436744e-06, "loss": 0.4511, "step": 16036 }, { "epoch": 0.6722488288986932, "grad_norm": 1.9922895431518555, "learning_rate": 2.5619817931195952e-06, "loss": 0.4939, "step": 16037 }, { "epoch": 0.6722907475137021, "grad_norm": 1.7282770872116089, "learning_rate": 2.5613891333370133e-06, "loss": 0.4974, "step": 16038 }, { "epoch": 0.6723326661287111, "grad_norm": 1.8342225551605225, "learning_rate": 2.560796518506849e-06, "loss": 0.4928, "step": 16039 }, { "epoch": 0.67237458474372, "grad_norm": 2.8782460689544678, "learning_rate": 2.5602039486400287e-06, "loss": 0.4862, "step": 16040 }, { "epoch": 0.672416503358729, "grad_norm": 1.8207834959030151, "learning_rate": 2.5596114237474783e-06, "loss": 0.4975, "step": 16041 }, { "epoch": 0.672458421973738, "grad_norm": 1.895416021347046, "learning_rate": 2.559018943840116e-06, "loss": 0.542, "step": 16042 }, { "epoch": 0.672500340588747, "grad_norm": 1.9153571128845215, "learning_rate": 2.558426508928864e-06, "loss": 0.5135, "step": 16043 }, { "epoch": 0.6725422592037559, "grad_norm": 1.9691052436828613, "learning_rate": 2.5578341190246458e-06, "loss": 0.5389, "step": 16044 }, { "epoch": 0.6725841778187649, "grad_norm": 2.1079564094543457, "learning_rate": 2.5572417741383797e-06, "loss": 0.4717, "step": 16045 }, { "epoch": 0.6726260964337738, "grad_norm": 2.110452890396118, "learning_rate": 2.5566494742809828e-06, "loss": 0.5838, "step": 16046 }, { "epoch": 0.6726680150487828, "grad_norm": 2.401027202606201, "learning_rate": 2.556057219463377e-06, "loss": 0.5285, "step": 16047 }, { "epoch": 0.6727099336637917, "grad_norm": 2.2450695037841797, "learning_rate": 2.555465009696475e-06, "loss": 0.4962, "step": 16048 }, { "epoch": 0.6727518522788007, "grad_norm": 1.8027966022491455, "learning_rate": 2.5548728449911963e-06, "loss": 0.5053, "step": 16049 }, { "epoch": 0.6727937708938097, "grad_norm": 1.6365288496017456, "learning_rate": 2.554280725358459e-06, "loss": 0.4926, "step": 16050 }, { "epoch": 0.6728356895088187, "grad_norm": 1.5901139974594116, "learning_rate": 2.553688650809173e-06, "loss": 0.5011, "step": 16051 }, { "epoch": 0.6728776081238276, "grad_norm": 1.6818596124649048, "learning_rate": 2.553096621354255e-06, "loss": 0.4734, "step": 16052 }, { "epoch": 0.6729195267388366, "grad_norm": 3.0321569442749023, "learning_rate": 2.5525046370046205e-06, "loss": 0.4321, "step": 16053 }, { "epoch": 0.6729614453538455, "grad_norm": 1.827843189239502, "learning_rate": 2.5519126977711785e-06, "loss": 0.505, "step": 16054 }, { "epoch": 0.6730033639688545, "grad_norm": 1.7179853916168213, "learning_rate": 2.5513208036648398e-06, "loss": 0.482, "step": 16055 }, { "epoch": 0.6730452825838634, "grad_norm": 2.6601998805999756, "learning_rate": 2.550728954696519e-06, "loss": 0.4999, "step": 16056 }, { "epoch": 0.6730872011988724, "grad_norm": 2.027419328689575, "learning_rate": 2.5501371508771223e-06, "loss": 0.5103, "step": 16057 }, { "epoch": 0.6731291198138813, "grad_norm": 1.6750191450119019, "learning_rate": 2.5495453922175594e-06, "loss": 0.5251, "step": 16058 }, { "epoch": 0.6731710384288903, "grad_norm": 2.212144613265991, "learning_rate": 2.5489536787287426e-06, "loss": 0.4822, "step": 16059 }, { "epoch": 0.6732129570438993, "grad_norm": 1.6918710470199585, "learning_rate": 2.5483620104215735e-06, "loss": 0.4286, "step": 16060 }, { "epoch": 0.6732548756589082, "grad_norm": 1.8568509817123413, "learning_rate": 2.547770387306962e-06, "loss": 0.4925, "step": 16061 }, { "epoch": 0.6732967942739172, "grad_norm": 1.7906533479690552, "learning_rate": 2.547178809395815e-06, "loss": 0.5082, "step": 16062 }, { "epoch": 0.6733387128889261, "grad_norm": 1.7975436449050903, "learning_rate": 2.5465872766990363e-06, "loss": 0.4486, "step": 16063 }, { "epoch": 0.6733806315039351, "grad_norm": 1.9969109296798706, "learning_rate": 2.5459957892275267e-06, "loss": 0.4521, "step": 16064 }, { "epoch": 0.673422550118944, "grad_norm": 1.5897456407546997, "learning_rate": 2.545404346992194e-06, "loss": 0.5086, "step": 16065 }, { "epoch": 0.673464468733953, "grad_norm": 1.675615668296814, "learning_rate": 2.5448129500039374e-06, "loss": 0.492, "step": 16066 }, { "epoch": 0.673506387348962, "grad_norm": 1.6281352043151855, "learning_rate": 2.5442215982736597e-06, "loss": 0.4375, "step": 16067 }, { "epoch": 0.673548305963971, "grad_norm": 1.852745771408081, "learning_rate": 2.5436302918122635e-06, "loss": 0.5, "step": 16068 }, { "epoch": 0.6735902245789799, "grad_norm": 1.955029845237732, "learning_rate": 2.5430390306306452e-06, "loss": 0.5052, "step": 16069 }, { "epoch": 0.6736321431939889, "grad_norm": 1.5345559120178223, "learning_rate": 2.542447814739706e-06, "loss": 0.4417, "step": 16070 }, { "epoch": 0.6736740618089978, "grad_norm": 1.8679944276809692, "learning_rate": 2.5418566441503454e-06, "loss": 0.5071, "step": 16071 }, { "epoch": 0.6737159804240068, "grad_norm": 1.7266581058502197, "learning_rate": 2.54126551887346e-06, "loss": 0.4853, "step": 16072 }, { "epoch": 0.6737578990390157, "grad_norm": 1.918400764465332, "learning_rate": 2.5406744389199433e-06, "loss": 0.4641, "step": 16073 }, { "epoch": 0.6737998176540247, "grad_norm": 1.748205542564392, "learning_rate": 2.540083404300695e-06, "loss": 0.4668, "step": 16074 }, { "epoch": 0.6738417362690337, "grad_norm": 3.2069222927093506, "learning_rate": 2.539492415026607e-06, "loss": 0.4617, "step": 16075 }, { "epoch": 0.6738836548840427, "grad_norm": 1.8448988199234009, "learning_rate": 2.5389014711085747e-06, "loss": 0.5179, "step": 16076 }, { "epoch": 0.6739255734990516, "grad_norm": 1.4567482471466064, "learning_rate": 2.538310572557493e-06, "loss": 0.4138, "step": 16077 }, { "epoch": 0.6739674921140606, "grad_norm": 2.019551992416382, "learning_rate": 2.5377197193842513e-06, "loss": 0.465, "step": 16078 }, { "epoch": 0.6740094107290695, "grad_norm": 1.9405752420425415, "learning_rate": 2.5371289115997423e-06, "loss": 0.5239, "step": 16079 }, { "epoch": 0.6740513293440785, "grad_norm": 1.6204380989074707, "learning_rate": 2.5365381492148588e-06, "loss": 0.5056, "step": 16080 }, { "epoch": 0.6740932479590874, "grad_norm": 3.5078630447387695, "learning_rate": 2.535947432240487e-06, "loss": 0.5599, "step": 16081 }, { "epoch": 0.6741351665740964, "grad_norm": 2.148634910583496, "learning_rate": 2.5353567606875195e-06, "loss": 0.5066, "step": 16082 }, { "epoch": 0.6741770851891054, "grad_norm": 1.644196629524231, "learning_rate": 2.5347661345668434e-06, "loss": 0.4882, "step": 16083 }, { "epoch": 0.6742190038041143, "grad_norm": 6.582376480102539, "learning_rate": 2.534175553889343e-06, "loss": 0.5066, "step": 16084 }, { "epoch": 0.6742609224191233, "grad_norm": 1.6876330375671387, "learning_rate": 2.5335850186659083e-06, "loss": 0.4811, "step": 16085 }, { "epoch": 0.6743028410341322, "grad_norm": 1.8540822267532349, "learning_rate": 2.5329945289074253e-06, "loss": 0.4517, "step": 16086 }, { "epoch": 0.6743447596491412, "grad_norm": 1.6733386516571045, "learning_rate": 2.5324040846247754e-06, "loss": 0.4857, "step": 16087 }, { "epoch": 0.6743866782641501, "grad_norm": 1.7862352132797241, "learning_rate": 2.5318136858288455e-06, "loss": 0.496, "step": 16088 }, { "epoch": 0.6744285968791591, "grad_norm": 1.7806553840637207, "learning_rate": 2.5312233325305204e-06, "loss": 0.4915, "step": 16089 }, { "epoch": 0.674470515494168, "grad_norm": 1.7179617881774902, "learning_rate": 2.5306330247406773e-06, "loss": 0.4694, "step": 16090 }, { "epoch": 0.674512434109177, "grad_norm": 1.570029854774475, "learning_rate": 2.530042762470203e-06, "loss": 0.51, "step": 16091 }, { "epoch": 0.674554352724186, "grad_norm": 2.0321128368377686, "learning_rate": 2.5294525457299757e-06, "loss": 0.506, "step": 16092 }, { "epoch": 0.674596271339195, "grad_norm": 2.175286293029785, "learning_rate": 2.5288623745308734e-06, "loss": 0.4927, "step": 16093 }, { "epoch": 0.6746381899542039, "grad_norm": 1.8365286588668823, "learning_rate": 2.5282722488837765e-06, "loss": 0.4994, "step": 16094 }, { "epoch": 0.6746801085692129, "grad_norm": 3.270824670791626, "learning_rate": 2.527682168799566e-06, "loss": 0.4647, "step": 16095 }, { "epoch": 0.6747220271842218, "grad_norm": 1.6988366842269897, "learning_rate": 2.527092134289115e-06, "loss": 0.4958, "step": 16096 }, { "epoch": 0.6747639457992308, "grad_norm": 1.7561091184616089, "learning_rate": 2.526502145363302e-06, "loss": 0.508, "step": 16097 }, { "epoch": 0.6748058644142397, "grad_norm": 1.8200057744979858, "learning_rate": 2.525912202033004e-06, "loss": 0.4991, "step": 16098 }, { "epoch": 0.6748477830292487, "grad_norm": 1.5963420867919922, "learning_rate": 2.5253223043090923e-06, "loss": 0.4603, "step": 16099 }, { "epoch": 0.6748897016442577, "grad_norm": 2.3709890842437744, "learning_rate": 2.524732452202445e-06, "loss": 0.471, "step": 16100 }, { "epoch": 0.6749316202592667, "grad_norm": 1.8435325622558594, "learning_rate": 2.5241426457239325e-06, "loss": 0.4803, "step": 16101 }, { "epoch": 0.6749735388742756, "grad_norm": 1.971924066543579, "learning_rate": 2.523552884884426e-06, "loss": 0.5627, "step": 16102 }, { "epoch": 0.6750154574892846, "grad_norm": 2.136354446411133, "learning_rate": 2.522963169694798e-06, "loss": 0.4631, "step": 16103 }, { "epoch": 0.6750573761042935, "grad_norm": 1.8182973861694336, "learning_rate": 2.522373500165922e-06, "loss": 0.4472, "step": 16104 }, { "epoch": 0.6750992947193025, "grad_norm": 1.754966378211975, "learning_rate": 2.521783876308663e-06, "loss": 0.4507, "step": 16105 }, { "epoch": 0.6751412133343114, "grad_norm": 1.7459611892700195, "learning_rate": 2.5211942981338943e-06, "loss": 0.5026, "step": 16106 }, { "epoch": 0.6751831319493204, "grad_norm": 2.1461691856384277, "learning_rate": 2.52060476565248e-06, "loss": 0.4908, "step": 16107 }, { "epoch": 0.6752250505643294, "grad_norm": 1.8662238121032715, "learning_rate": 2.5200152788752897e-06, "loss": 0.5342, "step": 16108 }, { "epoch": 0.6752669691793383, "grad_norm": 1.5479127168655396, "learning_rate": 2.519425837813191e-06, "loss": 0.4252, "step": 16109 }, { "epoch": 0.6753088877943473, "grad_norm": 2.8578264713287354, "learning_rate": 2.5188364424770475e-06, "loss": 0.5052, "step": 16110 }, { "epoch": 0.6753508064093562, "grad_norm": 1.7192424535751343, "learning_rate": 2.5182470928777214e-06, "loss": 0.5016, "step": 16111 }, { "epoch": 0.6753927250243652, "grad_norm": 1.7277637720108032, "learning_rate": 2.5176577890260824e-06, "loss": 0.5009, "step": 16112 }, { "epoch": 0.6754346436393741, "grad_norm": 1.7139203548431396, "learning_rate": 2.5170685309329873e-06, "loss": 0.4984, "step": 16113 }, { "epoch": 0.6754765622543831, "grad_norm": 2.1850900650024414, "learning_rate": 2.516479318609302e-06, "loss": 0.4899, "step": 16114 }, { "epoch": 0.675518480869392, "grad_norm": 1.5600616931915283, "learning_rate": 2.5158901520658883e-06, "loss": 0.5176, "step": 16115 }, { "epoch": 0.675560399484401, "grad_norm": 1.9930691719055176, "learning_rate": 2.515301031313604e-06, "loss": 0.524, "step": 16116 }, { "epoch": 0.67560231809941, "grad_norm": 1.8059223890304565, "learning_rate": 2.5147119563633106e-06, "loss": 0.4794, "step": 16117 }, { "epoch": 0.675644236714419, "grad_norm": 1.6358948945999146, "learning_rate": 2.514122927225868e-06, "loss": 0.4817, "step": 16118 }, { "epoch": 0.6756861553294279, "grad_norm": 1.6651298999786377, "learning_rate": 2.5135339439121327e-06, "loss": 0.4787, "step": 16119 }, { "epoch": 0.6757280739444369, "grad_norm": 1.8665305376052856, "learning_rate": 2.5129450064329597e-06, "loss": 0.4735, "step": 16120 }, { "epoch": 0.6757699925594458, "grad_norm": 1.8309128284454346, "learning_rate": 2.512356114799209e-06, "loss": 0.5022, "step": 16121 }, { "epoch": 0.6758119111744548, "grad_norm": 1.7559226751327515, "learning_rate": 2.5117672690217327e-06, "loss": 0.4473, "step": 16122 }, { "epoch": 0.6758538297894637, "grad_norm": 1.800361156463623, "learning_rate": 2.5111784691113864e-06, "loss": 0.4882, "step": 16123 }, { "epoch": 0.6758957484044728, "grad_norm": 2.050964832305908, "learning_rate": 2.510589715079027e-06, "loss": 0.4572, "step": 16124 }, { "epoch": 0.6759376670194817, "grad_norm": 1.9454878568649292, "learning_rate": 2.5100010069355023e-06, "loss": 0.4994, "step": 16125 }, { "epoch": 0.6759795856344907, "grad_norm": 2.450437068939209, "learning_rate": 2.5094123446916667e-06, "loss": 0.5415, "step": 16126 }, { "epoch": 0.6760215042494996, "grad_norm": 1.912253499031067, "learning_rate": 2.508823728358374e-06, "loss": 0.4804, "step": 16127 }, { "epoch": 0.6760634228645086, "grad_norm": 1.8845672607421875, "learning_rate": 2.508235157946469e-06, "loss": 0.4793, "step": 16128 }, { "epoch": 0.6761053414795175, "grad_norm": 2.089606523513794, "learning_rate": 2.5076466334668066e-06, "loss": 0.4974, "step": 16129 }, { "epoch": 0.6761472600945265, "grad_norm": 1.714218258857727, "learning_rate": 2.5070581549302325e-06, "loss": 0.4719, "step": 16130 }, { "epoch": 0.6761891787095354, "grad_norm": 2.3502326011657715, "learning_rate": 2.5064697223475932e-06, "loss": 0.4986, "step": 16131 }, { "epoch": 0.6762310973245445, "grad_norm": 1.5614771842956543, "learning_rate": 2.5058813357297373e-06, "loss": 0.4494, "step": 16132 }, { "epoch": 0.6762730159395534, "grad_norm": 2.734717845916748, "learning_rate": 2.5052929950875134e-06, "loss": 0.5056, "step": 16133 }, { "epoch": 0.6763149345545623, "grad_norm": 2.4987540245056152, "learning_rate": 2.5047047004317615e-06, "loss": 0.454, "step": 16134 }, { "epoch": 0.6763568531695713, "grad_norm": 1.4575592279434204, "learning_rate": 2.5041164517733296e-06, "loss": 0.516, "step": 16135 }, { "epoch": 0.6763987717845802, "grad_norm": 1.7068837881088257, "learning_rate": 2.503528249123062e-06, "loss": 0.4659, "step": 16136 }, { "epoch": 0.6764406903995892, "grad_norm": 1.7840049266815186, "learning_rate": 2.502940092491798e-06, "loss": 0.5245, "step": 16137 }, { "epoch": 0.6764826090145981, "grad_norm": 1.6424928903579712, "learning_rate": 2.5023519818903834e-06, "loss": 0.425, "step": 16138 }, { "epoch": 0.6765245276296071, "grad_norm": 1.8198412656784058, "learning_rate": 2.501763917329657e-06, "loss": 0.4866, "step": 16139 }, { "epoch": 0.676566446244616, "grad_norm": 1.6579029560089111, "learning_rate": 2.501175898820457e-06, "loss": 0.5143, "step": 16140 }, { "epoch": 0.676608364859625, "grad_norm": 2.8275833129882812, "learning_rate": 2.5005879263736257e-06, "loss": 0.475, "step": 16141 }, { "epoch": 0.676650283474634, "grad_norm": 1.8584387302398682, "learning_rate": 2.5000000000000015e-06, "loss": 0.4961, "step": 16142 }, { "epoch": 0.676692202089643, "grad_norm": 3.8012003898620605, "learning_rate": 2.4994121197104194e-06, "loss": 0.4457, "step": 16143 }, { "epoch": 0.6767341207046519, "grad_norm": 1.7242233753204346, "learning_rate": 2.498824285515718e-06, "loss": 0.4936, "step": 16144 }, { "epoch": 0.6767760393196609, "grad_norm": 1.61961829662323, "learning_rate": 2.498236497426735e-06, "loss": 0.4548, "step": 16145 }, { "epoch": 0.6768179579346698, "grad_norm": 1.675132155418396, "learning_rate": 2.4976487554543014e-06, "loss": 0.5182, "step": 16146 }, { "epoch": 0.6768598765496788, "grad_norm": 1.6209206581115723, "learning_rate": 2.497061059609256e-06, "loss": 0.459, "step": 16147 }, { "epoch": 0.6769017951646877, "grad_norm": 1.8410173654556274, "learning_rate": 2.4964734099024295e-06, "loss": 0.4776, "step": 16148 }, { "epoch": 0.6769437137796968, "grad_norm": 2.0143344402313232, "learning_rate": 2.4958858063446535e-06, "loss": 0.4651, "step": 16149 }, { "epoch": 0.6769856323947057, "grad_norm": 1.5325605869293213, "learning_rate": 2.4952982489467597e-06, "loss": 0.4617, "step": 16150 }, { "epoch": 0.6770275510097147, "grad_norm": 2.7613277435302734, "learning_rate": 2.4947107377195828e-06, "loss": 0.4819, "step": 16151 }, { "epoch": 0.6770694696247236, "grad_norm": 1.8218435049057007, "learning_rate": 2.4941232726739474e-06, "loss": 0.4559, "step": 16152 }, { "epoch": 0.6771113882397326, "grad_norm": 1.5667716264724731, "learning_rate": 2.4935358538206854e-06, "loss": 0.5113, "step": 16153 }, { "epoch": 0.6771533068547415, "grad_norm": 1.826664686203003, "learning_rate": 2.4929484811706275e-06, "loss": 0.4863, "step": 16154 }, { "epoch": 0.6771952254697505, "grad_norm": 1.6993921995162964, "learning_rate": 2.4923611547345956e-06, "loss": 0.4826, "step": 16155 }, { "epoch": 0.6772371440847594, "grad_norm": 1.6904202699661255, "learning_rate": 2.491773874523421e-06, "loss": 0.4365, "step": 16156 }, { "epoch": 0.6772790626997685, "grad_norm": 1.8575477600097656, "learning_rate": 2.491186640547928e-06, "loss": 0.5194, "step": 16157 }, { "epoch": 0.6773209813147774, "grad_norm": 4.8565192222595215, "learning_rate": 2.490599452818939e-06, "loss": 0.5115, "step": 16158 }, { "epoch": 0.6773628999297863, "grad_norm": 1.5573564767837524, "learning_rate": 2.4900123113472794e-06, "loss": 0.4588, "step": 16159 }, { "epoch": 0.6774048185447953, "grad_norm": 2.069960594177246, "learning_rate": 2.4894252161437756e-06, "loss": 0.5239, "step": 16160 }, { "epoch": 0.6774467371598042, "grad_norm": 1.8622218370437622, "learning_rate": 2.488838167219245e-06, "loss": 0.483, "step": 16161 }, { "epoch": 0.6774886557748132, "grad_norm": 1.9437335729599, "learning_rate": 2.488251164584511e-06, "loss": 0.506, "step": 16162 }, { "epoch": 0.6775305743898221, "grad_norm": 1.5048776865005493, "learning_rate": 2.4876642082503966e-06, "loss": 0.4551, "step": 16163 }, { "epoch": 0.6775724930048311, "grad_norm": 2.2383387088775635, "learning_rate": 2.487077298227717e-06, "loss": 0.4883, "step": 16164 }, { "epoch": 0.67761441161984, "grad_norm": 1.8548866510391235, "learning_rate": 2.4864904345272954e-06, "loss": 0.5331, "step": 16165 }, { "epoch": 0.6776563302348491, "grad_norm": 1.9236102104187012, "learning_rate": 2.4859036171599483e-06, "loss": 0.4899, "step": 16166 }, { "epoch": 0.677698248849858, "grad_norm": 1.9642820358276367, "learning_rate": 2.48531684613649e-06, "loss": 0.4894, "step": 16167 }, { "epoch": 0.677740167464867, "grad_norm": 1.5972826480865479, "learning_rate": 2.4847301214677392e-06, "loss": 0.4793, "step": 16168 }, { "epoch": 0.6777820860798759, "grad_norm": 1.742393136024475, "learning_rate": 2.4841434431645135e-06, "loss": 0.4722, "step": 16169 }, { "epoch": 0.6778240046948849, "grad_norm": 2.781653642654419, "learning_rate": 2.4835568112376236e-06, "loss": 0.5098, "step": 16170 }, { "epoch": 0.6778659233098938, "grad_norm": 1.6643530130386353, "learning_rate": 2.4829702256978845e-06, "loss": 0.4322, "step": 16171 }, { "epoch": 0.6779078419249028, "grad_norm": 1.7810618877410889, "learning_rate": 2.482383686556112e-06, "loss": 0.5148, "step": 16172 }, { "epoch": 0.6779497605399117, "grad_norm": 1.6682497262954712, "learning_rate": 2.481797193823113e-06, "loss": 0.4657, "step": 16173 }, { "epoch": 0.6779916791549208, "grad_norm": 2.000626802444458, "learning_rate": 2.481210747509704e-06, "loss": 0.4417, "step": 16174 }, { "epoch": 0.6780335977699297, "grad_norm": 1.6918487548828125, "learning_rate": 2.4806243476266904e-06, "loss": 0.4733, "step": 16175 }, { "epoch": 0.6780755163849387, "grad_norm": 1.7691850662231445, "learning_rate": 2.480037994184887e-06, "loss": 0.5101, "step": 16176 }, { "epoch": 0.6781174349999476, "grad_norm": 1.8743716478347778, "learning_rate": 2.479451687195096e-06, "loss": 0.5626, "step": 16177 }, { "epoch": 0.6781593536149566, "grad_norm": 2.5233397483825684, "learning_rate": 2.478865426668132e-06, "loss": 0.4507, "step": 16178 }, { "epoch": 0.6782012722299655, "grad_norm": 2.041583299636841, "learning_rate": 2.478279212614796e-06, "loss": 0.5163, "step": 16179 }, { "epoch": 0.6782431908449745, "grad_norm": 1.7084263563156128, "learning_rate": 2.477693045045898e-06, "loss": 0.427, "step": 16180 }, { "epoch": 0.6782851094599834, "grad_norm": 1.509926199913025, "learning_rate": 2.47710692397224e-06, "loss": 0.4294, "step": 16181 }, { "epoch": 0.6783270280749925, "grad_norm": 1.662312626838684, "learning_rate": 2.4765208494046285e-06, "loss": 0.4613, "step": 16182 }, { "epoch": 0.6783689466900014, "grad_norm": 1.7368483543395996, "learning_rate": 2.4759348213538682e-06, "loss": 0.521, "step": 16183 }, { "epoch": 0.6784108653050103, "grad_norm": 2.2028088569641113, "learning_rate": 2.4753488398307577e-06, "loss": 0.4758, "step": 16184 }, { "epoch": 0.6784527839200193, "grad_norm": 1.7986526489257812, "learning_rate": 2.4747629048461034e-06, "loss": 0.5103, "step": 16185 }, { "epoch": 0.6784947025350282, "grad_norm": 1.667487621307373, "learning_rate": 2.4741770164107043e-06, "loss": 0.48, "step": 16186 }, { "epoch": 0.6785366211500372, "grad_norm": 1.739171028137207, "learning_rate": 2.4735911745353573e-06, "loss": 0.5612, "step": 16187 }, { "epoch": 0.6785785397650461, "grad_norm": 1.754135012626648, "learning_rate": 2.473005379230865e-06, "loss": 0.456, "step": 16188 }, { "epoch": 0.6786204583800551, "grad_norm": 1.883955478668213, "learning_rate": 2.472419630508027e-06, "loss": 0.4525, "step": 16189 }, { "epoch": 0.678662376995064, "grad_norm": 1.5459109544754028, "learning_rate": 2.4718339283776365e-06, "loss": 0.4567, "step": 16190 }, { "epoch": 0.6787042956100731, "grad_norm": 1.9248785972595215, "learning_rate": 2.4712482728504926e-06, "loss": 0.4702, "step": 16191 }, { "epoch": 0.678746214225082, "grad_norm": 1.5879367589950562, "learning_rate": 2.4706626639373933e-06, "loss": 0.492, "step": 16192 }, { "epoch": 0.678788132840091, "grad_norm": 2.2236032485961914, "learning_rate": 2.470077101649129e-06, "loss": 0.5098, "step": 16193 }, { "epoch": 0.6788300514550999, "grad_norm": 1.743332862854004, "learning_rate": 2.4694915859964967e-06, "loss": 0.4359, "step": 16194 }, { "epoch": 0.6788719700701089, "grad_norm": 1.6279714107513428, "learning_rate": 2.46890611699029e-06, "loss": 0.4709, "step": 16195 }, { "epoch": 0.6789138886851178, "grad_norm": 2.1639137268066406, "learning_rate": 2.4683206946412976e-06, "loss": 0.4997, "step": 16196 }, { "epoch": 0.6789558073001268, "grad_norm": 2.0928664207458496, "learning_rate": 2.4677353189603127e-06, "loss": 0.4927, "step": 16197 }, { "epoch": 0.6789977259151357, "grad_norm": 1.6837456226348877, "learning_rate": 2.467149989958129e-06, "loss": 0.533, "step": 16198 }, { "epoch": 0.6790396445301448, "grad_norm": 1.9201143980026245, "learning_rate": 2.4665647076455312e-06, "loss": 0.543, "step": 16199 }, { "epoch": 0.6790815631451537, "grad_norm": 1.5852365493774414, "learning_rate": 2.4659794720333115e-06, "loss": 0.4599, "step": 16200 }, { "epoch": 0.6791234817601627, "grad_norm": 1.6573011875152588, "learning_rate": 2.465394283132258e-06, "loss": 0.463, "step": 16201 }, { "epoch": 0.6791654003751716, "grad_norm": 1.8247284889221191, "learning_rate": 2.4648091409531554e-06, "loss": 0.5017, "step": 16202 }, { "epoch": 0.6792073189901806, "grad_norm": 1.9557236433029175, "learning_rate": 2.464224045506794e-06, "loss": 0.5104, "step": 16203 }, { "epoch": 0.6792492376051895, "grad_norm": 1.6827237606048584, "learning_rate": 2.463638996803956e-06, "loss": 0.4661, "step": 16204 }, { "epoch": 0.6792911562201985, "grad_norm": 1.8269602060317993, "learning_rate": 2.463053994855425e-06, "loss": 0.5054, "step": 16205 }, { "epoch": 0.6793330748352074, "grad_norm": 1.7796549797058105, "learning_rate": 2.4624690396719857e-06, "loss": 0.4568, "step": 16206 }, { "epoch": 0.6793749934502165, "grad_norm": 1.7058088779449463, "learning_rate": 2.4618841312644244e-06, "loss": 0.4685, "step": 16207 }, { "epoch": 0.6794169120652254, "grad_norm": 1.7311509847640991, "learning_rate": 2.4612992696435183e-06, "loss": 0.5302, "step": 16208 }, { "epoch": 0.6794588306802343, "grad_norm": 1.727785348892212, "learning_rate": 2.46071445482005e-06, "loss": 0.4867, "step": 16209 }, { "epoch": 0.6795007492952433, "grad_norm": 2.105717182159424, "learning_rate": 2.460129686804803e-06, "loss": 0.5212, "step": 16210 }, { "epoch": 0.6795426679102522, "grad_norm": 1.6573905944824219, "learning_rate": 2.459544965608551e-06, "loss": 0.502, "step": 16211 }, { "epoch": 0.6795845865252612, "grad_norm": 1.7839525938034058, "learning_rate": 2.4589602912420784e-06, "loss": 0.5276, "step": 16212 }, { "epoch": 0.6796265051402701, "grad_norm": 1.6776083707809448, "learning_rate": 2.458375663716159e-06, "loss": 0.4552, "step": 16213 }, { "epoch": 0.6796684237552791, "grad_norm": 2.1589395999908447, "learning_rate": 2.4577910830415698e-06, "loss": 0.5267, "step": 16214 }, { "epoch": 0.679710342370288, "grad_norm": 2.3836069107055664, "learning_rate": 2.4572065492290863e-06, "loss": 0.4903, "step": 16215 }, { "epoch": 0.6797522609852971, "grad_norm": 2.0013458728790283, "learning_rate": 2.4566220622894876e-06, "loss": 0.4898, "step": 16216 }, { "epoch": 0.679794179600306, "grad_norm": 1.7530105113983154, "learning_rate": 2.456037622233543e-06, "loss": 0.4961, "step": 16217 }, { "epoch": 0.679836098215315, "grad_norm": 1.8440572023391724, "learning_rate": 2.455453229072027e-06, "loss": 0.4512, "step": 16218 }, { "epoch": 0.6798780168303239, "grad_norm": 1.7040659189224243, "learning_rate": 2.4548688828157156e-06, "loss": 0.5147, "step": 16219 }, { "epoch": 0.6799199354453329, "grad_norm": 1.379530906677246, "learning_rate": 2.4542845834753765e-06, "loss": 0.4428, "step": 16220 }, { "epoch": 0.6799618540603418, "grad_norm": 1.7663923501968384, "learning_rate": 2.453700331061783e-06, "loss": 0.4456, "step": 16221 }, { "epoch": 0.6800037726753508, "grad_norm": 1.932861089706421, "learning_rate": 2.4531161255857015e-06, "loss": 0.5061, "step": 16222 }, { "epoch": 0.6800456912903597, "grad_norm": 1.4840699434280396, "learning_rate": 2.4525319670579055e-06, "loss": 0.4336, "step": 16223 }, { "epoch": 0.6800876099053688, "grad_norm": 1.8137521743774414, "learning_rate": 2.4519478554891597e-06, "loss": 0.5273, "step": 16224 }, { "epoch": 0.6801295285203777, "grad_norm": 1.5631487369537354, "learning_rate": 2.451363790890234e-06, "loss": 0.4396, "step": 16225 }, { "epoch": 0.6801714471353867, "grad_norm": 1.7605047225952148, "learning_rate": 2.450779773271891e-06, "loss": 0.4726, "step": 16226 }, { "epoch": 0.6802133657503956, "grad_norm": 6.570701599121094, "learning_rate": 2.450195802644899e-06, "loss": 0.5169, "step": 16227 }, { "epoch": 0.6802552843654046, "grad_norm": 1.7641640901565552, "learning_rate": 2.4496118790200245e-06, "loss": 0.4825, "step": 16228 }, { "epoch": 0.6802972029804135, "grad_norm": 1.7161755561828613, "learning_rate": 2.449028002408027e-06, "loss": 0.5191, "step": 16229 }, { "epoch": 0.6803391215954225, "grad_norm": 2.1459665298461914, "learning_rate": 2.4484441728196715e-06, "loss": 0.5175, "step": 16230 }, { "epoch": 0.6803810402104314, "grad_norm": 1.7349830865859985, "learning_rate": 2.4478603902657227e-06, "loss": 0.4985, "step": 16231 }, { "epoch": 0.6804229588254405, "grad_norm": 1.7295619249343872, "learning_rate": 2.447276654756939e-06, "loss": 0.4937, "step": 16232 }, { "epoch": 0.6804648774404494, "grad_norm": 1.5115454196929932, "learning_rate": 2.44669296630408e-06, "loss": 0.4292, "step": 16233 }, { "epoch": 0.6805067960554583, "grad_norm": 2.7612931728363037, "learning_rate": 2.4461093249179073e-06, "loss": 0.4981, "step": 16234 }, { "epoch": 0.6805487146704673, "grad_norm": 1.6252316236495972, "learning_rate": 2.4455257306091773e-06, "loss": 0.5038, "step": 16235 }, { "epoch": 0.6805906332854762, "grad_norm": 1.5640573501586914, "learning_rate": 2.4449421833886483e-06, "loss": 0.4849, "step": 16236 }, { "epoch": 0.6806325519004852, "grad_norm": 1.77211594581604, "learning_rate": 2.44435868326708e-06, "loss": 0.4624, "step": 16237 }, { "epoch": 0.6806744705154941, "grad_norm": 1.697316288948059, "learning_rate": 2.443775230255225e-06, "loss": 0.4808, "step": 16238 }, { "epoch": 0.6807163891305031, "grad_norm": 1.772916555404663, "learning_rate": 2.4431918243638394e-06, "loss": 0.4593, "step": 16239 }, { "epoch": 0.680758307745512, "grad_norm": 2.190613031387329, "learning_rate": 2.4426084656036798e-06, "loss": 0.492, "step": 16240 }, { "epoch": 0.6808002263605211, "grad_norm": 1.8865834474563599, "learning_rate": 2.442025153985497e-06, "loss": 0.5237, "step": 16241 }, { "epoch": 0.68084214497553, "grad_norm": 1.6727386713027954, "learning_rate": 2.441441889520043e-06, "loss": 0.4824, "step": 16242 }, { "epoch": 0.680884063590539, "grad_norm": 1.899506688117981, "learning_rate": 2.4408586722180727e-06, "loss": 0.4804, "step": 16243 }, { "epoch": 0.6809259822055479, "grad_norm": 1.5917109251022339, "learning_rate": 2.4402755020903323e-06, "loss": 0.5369, "step": 16244 }, { "epoch": 0.6809679008205569, "grad_norm": 1.9215130805969238, "learning_rate": 2.4396923791475744e-06, "loss": 0.4976, "step": 16245 }, { "epoch": 0.6810098194355658, "grad_norm": 1.975412130355835, "learning_rate": 2.4391093034005505e-06, "loss": 0.5293, "step": 16246 }, { "epoch": 0.6810517380505748, "grad_norm": 2.924314498901367, "learning_rate": 2.4385262748600042e-06, "loss": 0.45, "step": 16247 }, { "epoch": 0.6810936566655837, "grad_norm": 1.8423224687576294, "learning_rate": 2.437943293536684e-06, "loss": 0.4209, "step": 16248 }, { "epoch": 0.6811355752805928, "grad_norm": 1.9816954135894775, "learning_rate": 2.43736035944134e-06, "loss": 0.4472, "step": 16249 }, { "epoch": 0.6811774938956017, "grad_norm": 1.884200096130371, "learning_rate": 2.436777472584715e-06, "loss": 0.474, "step": 16250 }, { "epoch": 0.6812194125106107, "grad_norm": 1.8314039707183838, "learning_rate": 2.436194632977552e-06, "loss": 0.4536, "step": 16251 }, { "epoch": 0.6812613311256196, "grad_norm": 1.8623301982879639, "learning_rate": 2.435611840630598e-06, "loss": 0.5198, "step": 16252 }, { "epoch": 0.6813032497406286, "grad_norm": 1.5012481212615967, "learning_rate": 2.435029095554593e-06, "loss": 0.4716, "step": 16253 }, { "epoch": 0.6813451683556375, "grad_norm": 1.5945113897323608, "learning_rate": 2.4344463977602807e-06, "loss": 0.4684, "step": 16254 }, { "epoch": 0.6813870869706465, "grad_norm": 1.5930280685424805, "learning_rate": 2.4338637472584035e-06, "loss": 0.4917, "step": 16255 }, { "epoch": 0.6814290055856554, "grad_norm": 4.361103534698486, "learning_rate": 2.433281144059699e-06, "loss": 0.511, "step": 16256 }, { "epoch": 0.6814709242006645, "grad_norm": 1.7892056703567505, "learning_rate": 2.4326985881749106e-06, "loss": 0.5139, "step": 16257 }, { "epoch": 0.6815128428156734, "grad_norm": 1.9092297554016113, "learning_rate": 2.4321160796147727e-06, "loss": 0.5269, "step": 16258 }, { "epoch": 0.6815547614306823, "grad_norm": 2.2332186698913574, "learning_rate": 2.431533618390026e-06, "loss": 0.5232, "step": 16259 }, { "epoch": 0.6815966800456913, "grad_norm": 1.7326488494873047, "learning_rate": 2.4309512045114064e-06, "loss": 0.5244, "step": 16260 }, { "epoch": 0.6816385986607002, "grad_norm": 1.6296964883804321, "learning_rate": 2.4303688379896483e-06, "loss": 0.4799, "step": 16261 }, { "epoch": 0.6816805172757092, "grad_norm": 1.8677724599838257, "learning_rate": 2.429786518835488e-06, "loss": 0.5203, "step": 16262 }, { "epoch": 0.6817224358907181, "grad_norm": 1.6621077060699463, "learning_rate": 2.4292042470596623e-06, "loss": 0.4252, "step": 16263 }, { "epoch": 0.6817643545057271, "grad_norm": 1.6028249263763428, "learning_rate": 2.4286220226728995e-06, "loss": 0.4624, "step": 16264 }, { "epoch": 0.681806273120736, "grad_norm": 1.7100740671157837, "learning_rate": 2.4280398456859356e-06, "loss": 0.5536, "step": 16265 }, { "epoch": 0.6818481917357451, "grad_norm": 1.6536173820495605, "learning_rate": 2.427457716109503e-06, "loss": 0.4732, "step": 16266 }, { "epoch": 0.681890110350754, "grad_norm": 1.883782982826233, "learning_rate": 2.426875633954329e-06, "loss": 0.4659, "step": 16267 }, { "epoch": 0.681932028965763, "grad_norm": 1.8479712009429932, "learning_rate": 2.4262935992311482e-06, "loss": 0.4386, "step": 16268 }, { "epoch": 0.6819739475807719, "grad_norm": 1.7109789848327637, "learning_rate": 2.4257116119506846e-06, "loss": 0.5026, "step": 16269 }, { "epoch": 0.6820158661957809, "grad_norm": 1.7863777875900269, "learning_rate": 2.42512967212367e-06, "loss": 0.4904, "step": 16270 }, { "epoch": 0.6820577848107898, "grad_norm": 1.5017359256744385, "learning_rate": 2.4245477797608287e-06, "loss": 0.4492, "step": 16271 }, { "epoch": 0.6820997034257988, "grad_norm": 1.6856005191802979, "learning_rate": 2.4239659348728906e-06, "loss": 0.471, "step": 16272 }, { "epoch": 0.6821416220408077, "grad_norm": 1.7214062213897705, "learning_rate": 2.4233841374705767e-06, "loss": 0.5366, "step": 16273 }, { "epoch": 0.6821835406558168, "grad_norm": 2.2295124530792236, "learning_rate": 2.422802387564615e-06, "loss": 0.478, "step": 16274 }, { "epoch": 0.6822254592708257, "grad_norm": 1.7731337547302246, "learning_rate": 2.42222068516573e-06, "loss": 0.4787, "step": 16275 }, { "epoch": 0.6822673778858347, "grad_norm": 1.9006367921829224, "learning_rate": 2.421639030284641e-06, "loss": 0.5236, "step": 16276 }, { "epoch": 0.6823092965008436, "grad_norm": 1.9376448392868042, "learning_rate": 2.421057422932072e-06, "loss": 0.5291, "step": 16277 }, { "epoch": 0.6823512151158526, "grad_norm": 2.38995623588562, "learning_rate": 2.4204758631187466e-06, "loss": 0.4482, "step": 16278 }, { "epoch": 0.6823931337308615, "grad_norm": 1.8926180601119995, "learning_rate": 2.4198943508553817e-06, "loss": 0.4908, "step": 16279 }, { "epoch": 0.6824350523458705, "grad_norm": 1.7800014019012451, "learning_rate": 2.419312886152696e-06, "loss": 0.4975, "step": 16280 }, { "epoch": 0.6824769709608794, "grad_norm": 1.5653034448623657, "learning_rate": 2.4187314690214112e-06, "loss": 0.4336, "step": 16281 }, { "epoch": 0.6825188895758885, "grad_norm": 1.9155492782592773, "learning_rate": 2.418150099472242e-06, "loss": 0.5286, "step": 16282 }, { "epoch": 0.6825608081908974, "grad_norm": 1.6220347881317139, "learning_rate": 2.4175687775159062e-06, "loss": 0.4616, "step": 16283 }, { "epoch": 0.6826027268059063, "grad_norm": 1.6900770664215088, "learning_rate": 2.416987503163121e-06, "loss": 0.5082, "step": 16284 }, { "epoch": 0.6826446454209153, "grad_norm": 1.6902217864990234, "learning_rate": 2.416406276424599e-06, "loss": 0.4639, "step": 16285 }, { "epoch": 0.6826865640359242, "grad_norm": 2.3109192848205566, "learning_rate": 2.415825097311056e-06, "loss": 0.5481, "step": 16286 }, { "epoch": 0.6827284826509332, "grad_norm": 2.06697154045105, "learning_rate": 2.4152439658332067e-06, "loss": 0.5197, "step": 16287 }, { "epoch": 0.6827704012659421, "grad_norm": 2.1059889793395996, "learning_rate": 2.4146628820017607e-06, "loss": 0.5415, "step": 16288 }, { "epoch": 0.6828123198809511, "grad_norm": 1.700476050376892, "learning_rate": 2.414081845827429e-06, "loss": 0.4905, "step": 16289 }, { "epoch": 0.68285423849596, "grad_norm": 1.5708919763565063, "learning_rate": 2.413500857320926e-06, "loss": 0.4743, "step": 16290 }, { "epoch": 0.6828961571109691, "grad_norm": 1.7726579904556274, "learning_rate": 2.4129199164929563e-06, "loss": 0.4803, "step": 16291 }, { "epoch": 0.682938075725978, "grad_norm": 1.9959999322891235, "learning_rate": 2.412339023354231e-06, "loss": 0.4714, "step": 16292 }, { "epoch": 0.682979994340987, "grad_norm": 1.565497875213623, "learning_rate": 2.4117581779154615e-06, "loss": 0.4164, "step": 16293 }, { "epoch": 0.6830219129559959, "grad_norm": 1.7359683513641357, "learning_rate": 2.411177380187349e-06, "loss": 0.4525, "step": 16294 }, { "epoch": 0.6830638315710049, "grad_norm": 1.869056224822998, "learning_rate": 2.4105966301806028e-06, "loss": 0.5059, "step": 16295 }, { "epoch": 0.6831057501860138, "grad_norm": 1.6050986051559448, "learning_rate": 2.4100159279059306e-06, "loss": 0.4765, "step": 16296 }, { "epoch": 0.6831476688010228, "grad_norm": 1.8057698011398315, "learning_rate": 2.4094352733740334e-06, "loss": 0.4661, "step": 16297 }, { "epoch": 0.6831895874160318, "grad_norm": 1.7077162265777588, "learning_rate": 2.4088546665956137e-06, "loss": 0.4117, "step": 16298 }, { "epoch": 0.6832315060310408, "grad_norm": 2.0995404720306396, "learning_rate": 2.408274107581378e-06, "loss": 0.5445, "step": 16299 }, { "epoch": 0.6832734246460497, "grad_norm": 1.902785301208496, "learning_rate": 2.407693596342025e-06, "loss": 0.4795, "step": 16300 }, { "epoch": 0.6833153432610587, "grad_norm": 1.6840513944625854, "learning_rate": 2.4071131328882563e-06, "loss": 0.4884, "step": 16301 }, { "epoch": 0.6833572618760676, "grad_norm": 1.7985975742340088, "learning_rate": 2.406532717230774e-06, "loss": 0.5197, "step": 16302 }, { "epoch": 0.6833991804910766, "grad_norm": 2.4847233295440674, "learning_rate": 2.4059523493802743e-06, "loss": 0.466, "step": 16303 }, { "epoch": 0.6834410991060855, "grad_norm": 2.081238269805908, "learning_rate": 2.4053720293474563e-06, "loss": 0.4879, "step": 16304 }, { "epoch": 0.6834830177210945, "grad_norm": 1.6773592233657837, "learning_rate": 2.4047917571430206e-06, "loss": 0.4197, "step": 16305 }, { "epoch": 0.6835249363361035, "grad_norm": 1.7764256000518799, "learning_rate": 2.4042115327776606e-06, "loss": 0.5005, "step": 16306 }, { "epoch": 0.6835668549511125, "grad_norm": 2.0110983848571777, "learning_rate": 2.4036313562620707e-06, "loss": 0.4486, "step": 16307 }, { "epoch": 0.6836087735661214, "grad_norm": 1.8075941801071167, "learning_rate": 2.4030512276069496e-06, "loss": 0.4954, "step": 16308 }, { "epoch": 0.6836506921811303, "grad_norm": 2.0637364387512207, "learning_rate": 2.4024711468229865e-06, "loss": 0.4804, "step": 16309 }, { "epoch": 0.6836926107961393, "grad_norm": 1.4591706991195679, "learning_rate": 2.4018911139208765e-06, "loss": 0.5275, "step": 16310 }, { "epoch": 0.6837345294111482, "grad_norm": 1.9619916677474976, "learning_rate": 2.401311128911315e-06, "loss": 0.4564, "step": 16311 }, { "epoch": 0.6837764480261572, "grad_norm": 1.7472273111343384, "learning_rate": 2.4007311918049874e-06, "loss": 0.5046, "step": 16312 }, { "epoch": 0.6838183666411661, "grad_norm": 1.8180975914001465, "learning_rate": 2.400151302612588e-06, "loss": 0.5298, "step": 16313 }, { "epoch": 0.6838602852561751, "grad_norm": 1.7809070348739624, "learning_rate": 2.3995714613448064e-06, "loss": 0.5022, "step": 16314 }, { "epoch": 0.6839022038711841, "grad_norm": 1.7687095403671265, "learning_rate": 2.3989916680123295e-06, "loss": 0.4283, "step": 16315 }, { "epoch": 0.6839441224861931, "grad_norm": 1.721570372581482, "learning_rate": 2.398411922625844e-06, "loss": 0.5362, "step": 16316 }, { "epoch": 0.683986041101202, "grad_norm": 1.6944513320922852, "learning_rate": 2.3978322251960395e-06, "loss": 0.4879, "step": 16317 }, { "epoch": 0.684027959716211, "grad_norm": 1.6352124214172363, "learning_rate": 2.3972525757335995e-06, "loss": 0.4399, "step": 16318 }, { "epoch": 0.6840698783312199, "grad_norm": 1.7455123662948608, "learning_rate": 2.3966729742492097e-06, "loss": 0.4745, "step": 16319 }, { "epoch": 0.6841117969462289, "grad_norm": 1.9003859758377075, "learning_rate": 2.396093420753556e-06, "loss": 0.5116, "step": 16320 }, { "epoch": 0.6841537155612378, "grad_norm": 1.765731930732727, "learning_rate": 2.3955139152573192e-06, "loss": 0.4997, "step": 16321 }, { "epoch": 0.6841956341762468, "grad_norm": 1.7747807502746582, "learning_rate": 2.3949344577711825e-06, "loss": 0.4511, "step": 16322 }, { "epoch": 0.6842375527912558, "grad_norm": 2.017124652862549, "learning_rate": 2.394355048305829e-06, "loss": 0.496, "step": 16323 }, { "epoch": 0.6842794714062648, "grad_norm": 2.0082457065582275, "learning_rate": 2.393775686871937e-06, "loss": 0.5244, "step": 16324 }, { "epoch": 0.6843213900212737, "grad_norm": 1.7897920608520508, "learning_rate": 2.393196373480189e-06, "loss": 0.4888, "step": 16325 }, { "epoch": 0.6843633086362827, "grad_norm": 2.1096854209899902, "learning_rate": 2.3926171081412614e-06, "loss": 0.4933, "step": 16326 }, { "epoch": 0.6844052272512916, "grad_norm": 2.0026981830596924, "learning_rate": 2.3920378908658313e-06, "loss": 0.4997, "step": 16327 }, { "epoch": 0.6844471458663006, "grad_norm": 1.9274461269378662, "learning_rate": 2.3914587216645774e-06, "loss": 0.5183, "step": 16328 }, { "epoch": 0.6844890644813095, "grad_norm": 1.8723022937774658, "learning_rate": 2.3908796005481777e-06, "loss": 0.5469, "step": 16329 }, { "epoch": 0.6845309830963185, "grad_norm": 1.6174845695495605, "learning_rate": 2.3903005275273035e-06, "loss": 0.4117, "step": 16330 }, { "epoch": 0.6845729017113275, "grad_norm": 1.9166170358657837, "learning_rate": 2.389721502612633e-06, "loss": 0.5056, "step": 16331 }, { "epoch": 0.6846148203263365, "grad_norm": 1.8810592889785767, "learning_rate": 2.3891425258148365e-06, "loss": 0.4976, "step": 16332 }, { "epoch": 0.6846567389413454, "grad_norm": 1.571653962135315, "learning_rate": 2.3885635971445875e-06, "loss": 0.4363, "step": 16333 }, { "epoch": 0.6846986575563543, "grad_norm": 1.7640457153320312, "learning_rate": 2.387984716612561e-06, "loss": 0.5167, "step": 16334 }, { "epoch": 0.6847405761713633, "grad_norm": 1.7611700296401978, "learning_rate": 2.387405884229425e-06, "loss": 0.4772, "step": 16335 }, { "epoch": 0.6847824947863722, "grad_norm": 1.7232894897460938, "learning_rate": 2.3868271000058473e-06, "loss": 0.4646, "step": 16336 }, { "epoch": 0.6848244134013812, "grad_norm": 2.022801160812378, "learning_rate": 2.386248363952502e-06, "loss": 0.5315, "step": 16337 }, { "epoch": 0.6848663320163901, "grad_norm": 1.912062406539917, "learning_rate": 2.3856696760800522e-06, "loss": 0.5094, "step": 16338 }, { "epoch": 0.6849082506313992, "grad_norm": 1.8263925313949585, "learning_rate": 2.3850910363991684e-06, "loss": 0.4108, "step": 16339 }, { "epoch": 0.6849501692464081, "grad_norm": 1.6894474029541016, "learning_rate": 2.384512444920518e-06, "loss": 0.5031, "step": 16340 }, { "epoch": 0.6849920878614171, "grad_norm": 2.210358142852783, "learning_rate": 2.383933901654763e-06, "loss": 0.5075, "step": 16341 }, { "epoch": 0.685034006476426, "grad_norm": 1.7606650590896606, "learning_rate": 2.3833554066125697e-06, "loss": 0.454, "step": 16342 }, { "epoch": 0.685075925091435, "grad_norm": 1.767945408821106, "learning_rate": 2.382776959804604e-06, "loss": 0.5499, "step": 16343 }, { "epoch": 0.6851178437064439, "grad_norm": 1.8265588283538818, "learning_rate": 2.3821985612415272e-06, "loss": 0.4874, "step": 16344 }, { "epoch": 0.6851597623214529, "grad_norm": 1.5599637031555176, "learning_rate": 2.3816202109339987e-06, "loss": 0.489, "step": 16345 }, { "epoch": 0.6852016809364618, "grad_norm": 2.2199654579162598, "learning_rate": 2.381041908892684e-06, "loss": 0.4823, "step": 16346 }, { "epoch": 0.6852435995514708, "grad_norm": 2.0374348163604736, "learning_rate": 2.380463655128239e-06, "loss": 0.4605, "step": 16347 }, { "epoch": 0.6852855181664798, "grad_norm": 1.6349012851715088, "learning_rate": 2.379885449651325e-06, "loss": 0.4626, "step": 16348 }, { "epoch": 0.6853274367814888, "grad_norm": 1.9235455989837646, "learning_rate": 2.3793072924726025e-06, "loss": 0.4836, "step": 16349 }, { "epoch": 0.6853693553964977, "grad_norm": 1.6987167596817017, "learning_rate": 2.3787291836027252e-06, "loss": 0.5255, "step": 16350 }, { "epoch": 0.6854112740115067, "grad_norm": 1.676611304283142, "learning_rate": 2.378151123052352e-06, "loss": 0.4496, "step": 16351 }, { "epoch": 0.6854531926265156, "grad_norm": 1.7604873180389404, "learning_rate": 2.377573110832139e-06, "loss": 0.49, "step": 16352 }, { "epoch": 0.6854951112415246, "grad_norm": 1.7416328191757202, "learning_rate": 2.3769951469527412e-06, "loss": 0.4549, "step": 16353 }, { "epoch": 0.6855370298565335, "grad_norm": 1.9781782627105713, "learning_rate": 2.37641723142481e-06, "loss": 0.5365, "step": 16354 }, { "epoch": 0.6855789484715425, "grad_norm": 1.70148503780365, "learning_rate": 2.3758393642590018e-06, "loss": 0.5112, "step": 16355 }, { "epoch": 0.6856208670865515, "grad_norm": 2.421699285507202, "learning_rate": 2.3752615454659648e-06, "loss": 0.4715, "step": 16356 }, { "epoch": 0.6856627857015605, "grad_norm": 1.7187079191207886, "learning_rate": 2.3746837750563527e-06, "loss": 0.5073, "step": 16357 }, { "epoch": 0.6857047043165694, "grad_norm": 1.891663670539856, "learning_rate": 2.374106053040818e-06, "loss": 0.4733, "step": 16358 }, { "epoch": 0.6857466229315783, "grad_norm": 2.0947678089141846, "learning_rate": 2.3735283794300064e-06, "loss": 0.4656, "step": 16359 }, { "epoch": 0.6857885415465873, "grad_norm": 1.6358447074890137, "learning_rate": 2.3729507542345674e-06, "loss": 0.4885, "step": 16360 }, { "epoch": 0.6858304601615962, "grad_norm": 1.8826627731323242, "learning_rate": 2.372373177465152e-06, "loss": 0.5146, "step": 16361 }, { "epoch": 0.6858723787766052, "grad_norm": 1.7793419361114502, "learning_rate": 2.3717956491324024e-06, "loss": 0.5062, "step": 16362 }, { "epoch": 0.6859142973916141, "grad_norm": 2.372941255569458, "learning_rate": 2.371218169246969e-06, "loss": 0.4231, "step": 16363 }, { "epoch": 0.6859562160066232, "grad_norm": 1.6876651048660278, "learning_rate": 2.3706407378194934e-06, "loss": 0.4755, "step": 16364 }, { "epoch": 0.6859981346216321, "grad_norm": 1.821205735206604, "learning_rate": 2.3700633548606196e-06, "loss": 0.4601, "step": 16365 }, { "epoch": 0.6860400532366411, "grad_norm": 1.8339765071868896, "learning_rate": 2.3694860203809916e-06, "loss": 0.4869, "step": 16366 }, { "epoch": 0.68608197185165, "grad_norm": 2.0961291790008545, "learning_rate": 2.3689087343912547e-06, "loss": 0.4635, "step": 16367 }, { "epoch": 0.686123890466659, "grad_norm": 6.04636287689209, "learning_rate": 2.368331496902046e-06, "loss": 0.4672, "step": 16368 }, { "epoch": 0.6861658090816679, "grad_norm": 1.7432687282562256, "learning_rate": 2.367754307924008e-06, "loss": 0.5293, "step": 16369 }, { "epoch": 0.6862077276966769, "grad_norm": 3.175204277038574, "learning_rate": 2.367177167467782e-06, "loss": 0.4687, "step": 16370 }, { "epoch": 0.6862496463116858, "grad_norm": 2.053459405899048, "learning_rate": 2.3666000755440026e-06, "loss": 0.5225, "step": 16371 }, { "epoch": 0.6862915649266949, "grad_norm": 1.5922410488128662, "learning_rate": 2.366023032163313e-06, "loss": 0.4848, "step": 16372 }, { "epoch": 0.6863334835417038, "grad_norm": 1.7722580432891846, "learning_rate": 2.3654460373363473e-06, "loss": 0.497, "step": 16373 }, { "epoch": 0.6863754021567128, "grad_norm": 1.6535590887069702, "learning_rate": 2.36486909107374e-06, "loss": 0.5128, "step": 16374 }, { "epoch": 0.6864173207717217, "grad_norm": 5.8099589347839355, "learning_rate": 2.364292193386128e-06, "loss": 0.4544, "step": 16375 }, { "epoch": 0.6864592393867307, "grad_norm": 1.7199275493621826, "learning_rate": 2.363715344284147e-06, "loss": 0.4898, "step": 16376 }, { "epoch": 0.6865011580017396, "grad_norm": 1.8725807666778564, "learning_rate": 2.3631385437784275e-06, "loss": 0.51, "step": 16377 }, { "epoch": 0.6865430766167486, "grad_norm": 1.6021443605422974, "learning_rate": 2.362561791879604e-06, "loss": 0.4713, "step": 16378 }, { "epoch": 0.6865849952317575, "grad_norm": 1.8643033504486084, "learning_rate": 2.361985088598309e-06, "loss": 0.4973, "step": 16379 }, { "epoch": 0.6866269138467666, "grad_norm": 1.6632065773010254, "learning_rate": 2.3614084339451706e-06, "loss": 0.5025, "step": 16380 }, { "epoch": 0.6866688324617755, "grad_norm": 1.774289846420288, "learning_rate": 2.3608318279308213e-06, "loss": 0.5104, "step": 16381 }, { "epoch": 0.6867107510767845, "grad_norm": 2.148627758026123, "learning_rate": 2.3602552705658894e-06, "loss": 0.4843, "step": 16382 }, { "epoch": 0.6867526696917934, "grad_norm": 1.6152129173278809, "learning_rate": 2.3596787618609996e-06, "loss": 0.4856, "step": 16383 }, { "epoch": 0.6867945883068023, "grad_norm": 2.131476879119873, "learning_rate": 2.3591023018267823e-06, "loss": 0.4979, "step": 16384 }, { "epoch": 0.6868365069218113, "grad_norm": 1.6639735698699951, "learning_rate": 2.3585258904738646e-06, "loss": 0.482, "step": 16385 }, { "epoch": 0.6868784255368202, "grad_norm": 1.7917019128799438, "learning_rate": 2.357949527812869e-06, "loss": 0.4707, "step": 16386 }, { "epoch": 0.6869203441518292, "grad_norm": 1.8698844909667969, "learning_rate": 2.357373213854421e-06, "loss": 0.4657, "step": 16387 }, { "epoch": 0.6869622627668381, "grad_norm": 2.077497959136963, "learning_rate": 2.3567969486091463e-06, "loss": 0.4654, "step": 16388 }, { "epoch": 0.6870041813818472, "grad_norm": 1.6995747089385986, "learning_rate": 2.356220732087664e-06, "loss": 0.434, "step": 16389 }, { "epoch": 0.6870460999968561, "grad_norm": 2.304807186126709, "learning_rate": 2.355644564300599e-06, "loss": 0.4511, "step": 16390 }, { "epoch": 0.6870880186118651, "grad_norm": 1.7620769739151, "learning_rate": 2.3550684452585715e-06, "loss": 0.4886, "step": 16391 }, { "epoch": 0.687129937226874, "grad_norm": 1.7001526355743408, "learning_rate": 2.3544923749721983e-06, "loss": 0.4835, "step": 16392 }, { "epoch": 0.687171855841883, "grad_norm": 2.126883029937744, "learning_rate": 2.353916353452101e-06, "loss": 0.4946, "step": 16393 }, { "epoch": 0.6872137744568919, "grad_norm": 1.992248773574829, "learning_rate": 2.3533403807088996e-06, "loss": 0.4617, "step": 16394 }, { "epoch": 0.6872556930719009, "grad_norm": 2.9186160564422607, "learning_rate": 2.3527644567532075e-06, "loss": 0.4729, "step": 16395 }, { "epoch": 0.6872976116869098, "grad_norm": 2.980788469314575, "learning_rate": 2.352188581595643e-06, "loss": 0.5149, "step": 16396 }, { "epoch": 0.6873395303019189, "grad_norm": 1.94413423538208, "learning_rate": 2.351612755246823e-06, "loss": 0.5197, "step": 16397 }, { "epoch": 0.6873814489169278, "grad_norm": 1.785504698753357, "learning_rate": 2.351036977717359e-06, "loss": 0.4404, "step": 16398 }, { "epoch": 0.6874233675319368, "grad_norm": 1.8233400583267212, "learning_rate": 2.350461249017868e-06, "loss": 0.5017, "step": 16399 }, { "epoch": 0.6874652861469457, "grad_norm": 1.711167573928833, "learning_rate": 2.349885569158961e-06, "loss": 0.4763, "step": 16400 }, { "epoch": 0.6875072047619547, "grad_norm": 1.9719146490097046, "learning_rate": 2.349309938151248e-06, "loss": 0.4895, "step": 16401 }, { "epoch": 0.6875491233769636, "grad_norm": 1.902206540107727, "learning_rate": 2.348734356005342e-06, "loss": 0.506, "step": 16402 }, { "epoch": 0.6875910419919726, "grad_norm": 2.218442440032959, "learning_rate": 2.348158822731855e-06, "loss": 0.4798, "step": 16403 }, { "epoch": 0.6876329606069815, "grad_norm": 1.9149705171585083, "learning_rate": 2.3475833383413916e-06, "loss": 0.4462, "step": 16404 }, { "epoch": 0.6876748792219906, "grad_norm": 1.7081835269927979, "learning_rate": 2.3470079028445645e-06, "loss": 0.4838, "step": 16405 }, { "epoch": 0.6877167978369995, "grad_norm": 2.134960412979126, "learning_rate": 2.3464325162519774e-06, "loss": 0.538, "step": 16406 }, { "epoch": 0.6877587164520085, "grad_norm": 1.6798713207244873, "learning_rate": 2.345857178574238e-06, "loss": 0.4173, "step": 16407 }, { "epoch": 0.6878006350670174, "grad_norm": 2.0499515533447266, "learning_rate": 2.345281889821955e-06, "loss": 0.4318, "step": 16408 }, { "epoch": 0.6878425536820263, "grad_norm": 2.1010560989379883, "learning_rate": 2.3447066500057276e-06, "loss": 0.4631, "step": 16409 }, { "epoch": 0.6878844722970353, "grad_norm": 1.9332269430160522, "learning_rate": 2.344131459136164e-06, "loss": 0.5345, "step": 16410 }, { "epoch": 0.6879263909120442, "grad_norm": 2.2911550998687744, "learning_rate": 2.3435563172238663e-06, "loss": 0.4874, "step": 16411 }, { "epoch": 0.6879683095270532, "grad_norm": 1.515244483947754, "learning_rate": 2.3429812242794327e-06, "loss": 0.4547, "step": 16412 }, { "epoch": 0.6880102281420621, "grad_norm": 1.7924836874008179, "learning_rate": 2.342406180313467e-06, "loss": 0.4753, "step": 16413 }, { "epoch": 0.6880521467570712, "grad_norm": 1.6409443616867065, "learning_rate": 2.341831185336572e-06, "loss": 0.5131, "step": 16414 }, { "epoch": 0.6880940653720801, "grad_norm": 1.698625922203064, "learning_rate": 2.341256239359342e-06, "loss": 0.4685, "step": 16415 }, { "epoch": 0.6881359839870891, "grad_norm": 1.9222118854522705, "learning_rate": 2.3406813423923776e-06, "loss": 0.5053, "step": 16416 }, { "epoch": 0.688177902602098, "grad_norm": 1.6851880550384521, "learning_rate": 2.3401064944462783e-06, "loss": 0.4533, "step": 16417 }, { "epoch": 0.688219821217107, "grad_norm": 1.6884617805480957, "learning_rate": 2.3395316955316365e-06, "loss": 0.5058, "step": 16418 }, { "epoch": 0.6882617398321159, "grad_norm": 1.9334077835083008, "learning_rate": 2.3389569456590517e-06, "loss": 0.4842, "step": 16419 }, { "epoch": 0.6883036584471249, "grad_norm": 1.9633042812347412, "learning_rate": 2.338382244839117e-06, "loss": 0.5034, "step": 16420 }, { "epoch": 0.6883455770621338, "grad_norm": 1.8924245834350586, "learning_rate": 2.337807593082424e-06, "loss": 0.4658, "step": 16421 }, { "epoch": 0.6883874956771429, "grad_norm": 2.096604108810425, "learning_rate": 2.3372329903995677e-06, "loss": 0.4783, "step": 16422 }, { "epoch": 0.6884294142921518, "grad_norm": 2.1307754516601562, "learning_rate": 2.3366584368011424e-06, "loss": 0.4848, "step": 16423 }, { "epoch": 0.6884713329071608, "grad_norm": 2.169191360473633, "learning_rate": 2.3360839322977343e-06, "loss": 0.457, "step": 16424 }, { "epoch": 0.6885132515221697, "grad_norm": 3.001807451248169, "learning_rate": 2.3355094768999365e-06, "loss": 0.4698, "step": 16425 }, { "epoch": 0.6885551701371787, "grad_norm": 1.6460334062576294, "learning_rate": 2.33493507061834e-06, "loss": 0.5032, "step": 16426 }, { "epoch": 0.6885970887521876, "grad_norm": 1.8374922275543213, "learning_rate": 2.3343607134635288e-06, "loss": 0.4883, "step": 16427 }, { "epoch": 0.6886390073671966, "grad_norm": 1.6198830604553223, "learning_rate": 2.3337864054460947e-06, "loss": 0.4732, "step": 16428 }, { "epoch": 0.6886809259822055, "grad_norm": 1.9160019159317017, "learning_rate": 2.333212146576623e-06, "loss": 0.486, "step": 16429 }, { "epoch": 0.6887228445972146, "grad_norm": 1.9580801725387573, "learning_rate": 2.332637936865696e-06, "loss": 0.5239, "step": 16430 }, { "epoch": 0.6887647632122235, "grad_norm": 1.8081088066101074, "learning_rate": 2.332063776323901e-06, "loss": 0.5797, "step": 16431 }, { "epoch": 0.6888066818272325, "grad_norm": 1.785813331604004, "learning_rate": 2.331489664961824e-06, "loss": 0.5066, "step": 16432 }, { "epoch": 0.6888486004422414, "grad_norm": 1.627916932106018, "learning_rate": 2.330915602790044e-06, "loss": 0.4446, "step": 16433 }, { "epoch": 0.6888905190572503, "grad_norm": 1.734134316444397, "learning_rate": 2.3303415898191456e-06, "loss": 0.4925, "step": 16434 }, { "epoch": 0.6889324376722593, "grad_norm": 1.7548210620880127, "learning_rate": 2.3297676260597105e-06, "loss": 0.4706, "step": 16435 }, { "epoch": 0.6889743562872682, "grad_norm": 1.7786105871200562, "learning_rate": 2.3291937115223163e-06, "loss": 0.4168, "step": 16436 }, { "epoch": 0.6890162749022772, "grad_norm": 1.688385248184204, "learning_rate": 2.3286198462175454e-06, "loss": 0.4874, "step": 16437 }, { "epoch": 0.6890581935172861, "grad_norm": 2.4665510654449463, "learning_rate": 2.3280460301559742e-06, "loss": 0.5351, "step": 16438 }, { "epoch": 0.6891001121322952, "grad_norm": 2.064272880554199, "learning_rate": 2.3274722633481793e-06, "loss": 0.5499, "step": 16439 }, { "epoch": 0.6891420307473041, "grad_norm": 2.020604372024536, "learning_rate": 2.3268985458047383e-06, "loss": 0.4987, "step": 16440 }, { "epoch": 0.6891839493623131, "grad_norm": 1.5660425424575806, "learning_rate": 2.326324877536229e-06, "loss": 0.4841, "step": 16441 }, { "epoch": 0.689225867977322, "grad_norm": 1.9297531843185425, "learning_rate": 2.325751258553223e-06, "loss": 0.5086, "step": 16442 }, { "epoch": 0.689267786592331, "grad_norm": 1.9127241373062134, "learning_rate": 2.3251776888662946e-06, "loss": 0.4867, "step": 16443 }, { "epoch": 0.6893097052073399, "grad_norm": 1.5549368858337402, "learning_rate": 2.32460416848602e-06, "loss": 0.4895, "step": 16444 }, { "epoch": 0.6893516238223489, "grad_norm": 1.8433843851089478, "learning_rate": 2.324030697422967e-06, "loss": 0.5261, "step": 16445 }, { "epoch": 0.6893935424373578, "grad_norm": 1.890932321548462, "learning_rate": 2.3234572756877107e-06, "loss": 0.4685, "step": 16446 }, { "epoch": 0.6894354610523669, "grad_norm": 2.1412577629089355, "learning_rate": 2.322883903290819e-06, "loss": 0.4278, "step": 16447 }, { "epoch": 0.6894773796673758, "grad_norm": 1.9356141090393066, "learning_rate": 2.3223105802428605e-06, "loss": 0.5122, "step": 16448 }, { "epoch": 0.6895192982823848, "grad_norm": 1.98673677444458, "learning_rate": 2.3217373065544036e-06, "loss": 0.5204, "step": 16449 }, { "epoch": 0.6895612168973937, "grad_norm": 1.8061389923095703, "learning_rate": 2.3211640822360192e-06, "loss": 0.4511, "step": 16450 }, { "epoch": 0.6896031355124027, "grad_norm": 2.482567310333252, "learning_rate": 2.3205909072982695e-06, "loss": 0.4593, "step": 16451 }, { "epoch": 0.6896450541274116, "grad_norm": 1.7617485523223877, "learning_rate": 2.320017781751722e-06, "loss": 0.5203, "step": 16452 }, { "epoch": 0.6896869727424206, "grad_norm": 2.8510141372680664, "learning_rate": 2.319444705606945e-06, "loss": 0.4818, "step": 16453 }, { "epoch": 0.6897288913574295, "grad_norm": 1.9695438146591187, "learning_rate": 2.3188716788744958e-06, "loss": 0.4785, "step": 16454 }, { "epoch": 0.6897708099724386, "grad_norm": 1.7028920650482178, "learning_rate": 2.3182987015649428e-06, "loss": 0.3984, "step": 16455 }, { "epoch": 0.6898127285874475, "grad_norm": 1.6789108514785767, "learning_rate": 2.317725773688844e-06, "loss": 0.4129, "step": 16456 }, { "epoch": 0.6898546472024565, "grad_norm": 2.0095415115356445, "learning_rate": 2.3171528952567644e-06, "loss": 0.4688, "step": 16457 }, { "epoch": 0.6898965658174654, "grad_norm": 1.5862221717834473, "learning_rate": 2.3165800662792607e-06, "loss": 0.4279, "step": 16458 }, { "epoch": 0.6899384844324743, "grad_norm": 1.6618316173553467, "learning_rate": 2.3160072867668955e-06, "loss": 0.4637, "step": 16459 }, { "epoch": 0.6899804030474833, "grad_norm": 1.6642223596572876, "learning_rate": 2.315434556730223e-06, "loss": 0.4628, "step": 16460 }, { "epoch": 0.6900223216624922, "grad_norm": 1.8861016035079956, "learning_rate": 2.314861876179804e-06, "loss": 0.4446, "step": 16461 }, { "epoch": 0.6900642402775012, "grad_norm": 1.5290788412094116, "learning_rate": 2.3142892451261955e-06, "loss": 0.4854, "step": 16462 }, { "epoch": 0.6901061588925101, "grad_norm": 1.7701411247253418, "learning_rate": 2.3137166635799507e-06, "loss": 0.4321, "step": 16463 }, { "epoch": 0.6901480775075192, "grad_norm": 1.9548044204711914, "learning_rate": 2.313144131551625e-06, "loss": 0.4771, "step": 16464 }, { "epoch": 0.6901899961225281, "grad_norm": 2.1968135833740234, "learning_rate": 2.3125716490517745e-06, "loss": 0.5555, "step": 16465 }, { "epoch": 0.6902319147375371, "grad_norm": 1.8289281129837036, "learning_rate": 2.3119992160909503e-06, "loss": 0.4499, "step": 16466 }, { "epoch": 0.690273833352546, "grad_norm": 1.9411596059799194, "learning_rate": 2.3114268326797036e-06, "loss": 0.4425, "step": 16467 }, { "epoch": 0.690315751967555, "grad_norm": 1.8297837972640991, "learning_rate": 2.3108544988285876e-06, "loss": 0.4671, "step": 16468 }, { "epoch": 0.6903576705825639, "grad_norm": 1.8964356184005737, "learning_rate": 2.310282214548149e-06, "loss": 0.4654, "step": 16469 }, { "epoch": 0.6903995891975729, "grad_norm": 1.9848862886428833, "learning_rate": 2.3097099798489396e-06, "loss": 0.5092, "step": 16470 }, { "epoch": 0.6904415078125818, "grad_norm": 1.785353183746338, "learning_rate": 2.30913779474151e-06, "loss": 0.4345, "step": 16471 }, { "epoch": 0.6904834264275909, "grad_norm": 3.213999032974243, "learning_rate": 2.3085656592364025e-06, "loss": 0.5467, "step": 16472 }, { "epoch": 0.6905253450425998, "grad_norm": 2.5708303451538086, "learning_rate": 2.3079935733441665e-06, "loss": 0.4678, "step": 16473 }, { "epoch": 0.6905672636576088, "grad_norm": 2.020962953567505, "learning_rate": 2.3074215370753493e-06, "loss": 0.4639, "step": 16474 }, { "epoch": 0.6906091822726177, "grad_norm": 2.0329806804656982, "learning_rate": 2.306849550440494e-06, "loss": 0.5248, "step": 16475 }, { "epoch": 0.6906511008876267, "grad_norm": 1.5541186332702637, "learning_rate": 2.306277613450142e-06, "loss": 0.4503, "step": 16476 }, { "epoch": 0.6906930195026356, "grad_norm": 1.823136806488037, "learning_rate": 2.3057057261148403e-06, "loss": 0.4823, "step": 16477 }, { "epoch": 0.6907349381176446, "grad_norm": 1.8752248287200928, "learning_rate": 2.3051338884451263e-06, "loss": 0.4844, "step": 16478 }, { "epoch": 0.6907768567326535, "grad_norm": 2.2598352432250977, "learning_rate": 2.3045621004515457e-06, "loss": 0.4871, "step": 16479 }, { "epoch": 0.6908187753476626, "grad_norm": 1.7878285646438599, "learning_rate": 2.3039903621446345e-06, "loss": 0.4757, "step": 16480 }, { "epoch": 0.6908606939626715, "grad_norm": 1.7743022441864014, "learning_rate": 2.303418673534934e-06, "loss": 0.4668, "step": 16481 }, { "epoch": 0.6909026125776805, "grad_norm": 1.75819730758667, "learning_rate": 2.3028470346329847e-06, "loss": 0.5164, "step": 16482 }, { "epoch": 0.6909445311926894, "grad_norm": 2.044119119644165, "learning_rate": 2.302275445449319e-06, "loss": 0.5486, "step": 16483 }, { "epoch": 0.6909864498076983, "grad_norm": 3.505645990371704, "learning_rate": 2.301703905994478e-06, "loss": 0.5129, "step": 16484 }, { "epoch": 0.6910283684227073, "grad_norm": 1.7363892793655396, "learning_rate": 2.3011324162789954e-06, "loss": 0.4602, "step": 16485 }, { "epoch": 0.6910702870377162, "grad_norm": 1.7796796560287476, "learning_rate": 2.3005609763134034e-06, "loss": 0.5, "step": 16486 }, { "epoch": 0.6911122056527252, "grad_norm": 1.6315617561340332, "learning_rate": 2.2999895861082382e-06, "loss": 0.4977, "step": 16487 }, { "epoch": 0.6911541242677341, "grad_norm": 1.8151113986968994, "learning_rate": 2.2994182456740345e-06, "loss": 0.5022, "step": 16488 }, { "epoch": 0.6911960428827432, "grad_norm": 2.967460870742798, "learning_rate": 2.2988469550213195e-06, "loss": 0.4917, "step": 16489 }, { "epoch": 0.6912379614977521, "grad_norm": 1.9635910987854004, "learning_rate": 2.298275714160627e-06, "loss": 0.4518, "step": 16490 }, { "epoch": 0.6912798801127611, "grad_norm": 1.893892765045166, "learning_rate": 2.2977045231024885e-06, "loss": 0.4858, "step": 16491 }, { "epoch": 0.69132179872777, "grad_norm": 1.9507025480270386, "learning_rate": 2.2971333818574297e-06, "loss": 0.4678, "step": 16492 }, { "epoch": 0.691363717342779, "grad_norm": 1.8346624374389648, "learning_rate": 2.296562290435982e-06, "loss": 0.5193, "step": 16493 }, { "epoch": 0.6914056359577879, "grad_norm": 1.7297189235687256, "learning_rate": 2.2959912488486712e-06, "loss": 0.4637, "step": 16494 }, { "epoch": 0.6914475545727969, "grad_norm": 2.0602304935455322, "learning_rate": 2.295420257106021e-06, "loss": 0.4503, "step": 16495 }, { "epoch": 0.6914894731878058, "grad_norm": 3.1892824172973633, "learning_rate": 2.2948493152185602e-06, "loss": 0.5195, "step": 16496 }, { "epoch": 0.6915313918028149, "grad_norm": 1.5925204753875732, "learning_rate": 2.2942784231968144e-06, "loss": 0.4769, "step": 16497 }, { "epoch": 0.6915733104178238, "grad_norm": 2.0175342559814453, "learning_rate": 2.293707581051303e-06, "loss": 0.4764, "step": 16498 }, { "epoch": 0.6916152290328328, "grad_norm": 1.596413254737854, "learning_rate": 2.2931367887925515e-06, "loss": 0.434, "step": 16499 }, { "epoch": 0.6916571476478417, "grad_norm": 1.7317817211151123, "learning_rate": 2.2925660464310828e-06, "loss": 0.5281, "step": 16500 }, { "epoch": 0.6916990662628507, "grad_norm": 1.826751470565796, "learning_rate": 2.2919953539774144e-06, "loss": 0.5063, "step": 16501 }, { "epoch": 0.6917409848778596, "grad_norm": 2.0229506492614746, "learning_rate": 2.2914247114420696e-06, "loss": 0.4934, "step": 16502 }, { "epoch": 0.6917829034928686, "grad_norm": 2.005533456802368, "learning_rate": 2.290854118835564e-06, "loss": 0.5099, "step": 16503 }, { "epoch": 0.6918248221078775, "grad_norm": 1.71453857421875, "learning_rate": 2.290283576168419e-06, "loss": 0.5225, "step": 16504 }, { "epoch": 0.6918667407228866, "grad_norm": 5.6684346199035645, "learning_rate": 2.289713083451149e-06, "loss": 0.4574, "step": 16505 }, { "epoch": 0.6919086593378955, "grad_norm": 1.5812416076660156, "learning_rate": 2.2891426406942726e-06, "loss": 0.4509, "step": 16506 }, { "epoch": 0.6919505779529045, "grad_norm": 1.6624565124511719, "learning_rate": 2.288572247908302e-06, "loss": 0.4618, "step": 16507 }, { "epoch": 0.6919924965679134, "grad_norm": 1.5985195636749268, "learning_rate": 2.2880019051037543e-06, "loss": 0.4897, "step": 16508 }, { "epoch": 0.6920344151829223, "grad_norm": 1.6810739040374756, "learning_rate": 2.2874316122911437e-06, "loss": 0.5505, "step": 16509 }, { "epoch": 0.6920763337979313, "grad_norm": 1.7406728267669678, "learning_rate": 2.286861369480979e-06, "loss": 0.4701, "step": 16510 }, { "epoch": 0.6921182524129402, "grad_norm": 2.6509549617767334, "learning_rate": 2.286291176683775e-06, "loss": 0.5186, "step": 16511 }, { "epoch": 0.6921601710279492, "grad_norm": 1.7031124830245972, "learning_rate": 2.2857210339100432e-06, "loss": 0.4804, "step": 16512 }, { "epoch": 0.6922020896429582, "grad_norm": 1.833728551864624, "learning_rate": 2.2851509411702917e-06, "loss": 0.4889, "step": 16513 }, { "epoch": 0.6922440082579672, "grad_norm": 1.834878921508789, "learning_rate": 2.284580898475028e-06, "loss": 0.5273, "step": 16514 }, { "epoch": 0.6922859268729761, "grad_norm": 2.0084168910980225, "learning_rate": 2.2840109058347625e-06, "loss": 0.5481, "step": 16515 }, { "epoch": 0.6923278454879851, "grad_norm": 1.81930410861969, "learning_rate": 2.28344096326e-06, "loss": 0.487, "step": 16516 }, { "epoch": 0.692369764102994, "grad_norm": 2.1558260917663574, "learning_rate": 2.2828710707612472e-06, "loss": 0.4724, "step": 16517 }, { "epoch": 0.692411682718003, "grad_norm": 1.5366791486740112, "learning_rate": 2.282301228349012e-06, "loss": 0.4944, "step": 16518 }, { "epoch": 0.6924536013330119, "grad_norm": 2.291930913925171, "learning_rate": 2.281731436033795e-06, "loss": 0.4886, "step": 16519 }, { "epoch": 0.6924955199480209, "grad_norm": 1.8630236387252808, "learning_rate": 2.281161693826101e-06, "loss": 0.4754, "step": 16520 }, { "epoch": 0.6925374385630299, "grad_norm": 2.87697696685791, "learning_rate": 2.280592001736434e-06, "loss": 0.5011, "step": 16521 }, { "epoch": 0.6925793571780389, "grad_norm": 2.1802330017089844, "learning_rate": 2.280022359775294e-06, "loss": 0.4777, "step": 16522 }, { "epoch": 0.6926212757930478, "grad_norm": 1.5571715831756592, "learning_rate": 2.2794527679531804e-06, "loss": 0.5073, "step": 16523 }, { "epoch": 0.6926631944080568, "grad_norm": 1.7134058475494385, "learning_rate": 2.2788832262805947e-06, "loss": 0.4752, "step": 16524 }, { "epoch": 0.6927051130230657, "grad_norm": 1.739422082901001, "learning_rate": 2.278313734768034e-06, "loss": 0.4856, "step": 16525 }, { "epoch": 0.6927470316380747, "grad_norm": 1.8196709156036377, "learning_rate": 2.277744293425997e-06, "loss": 0.4549, "step": 16526 }, { "epoch": 0.6927889502530836, "grad_norm": 2.0970377922058105, "learning_rate": 2.2771749022649818e-06, "loss": 0.5283, "step": 16527 }, { "epoch": 0.6928308688680926, "grad_norm": 1.8839638233184814, "learning_rate": 2.276605561295481e-06, "loss": 0.4438, "step": 16528 }, { "epoch": 0.6928727874831015, "grad_norm": 1.5760092735290527, "learning_rate": 2.2760362705279924e-06, "loss": 0.4205, "step": 16529 }, { "epoch": 0.6929147060981106, "grad_norm": 1.898353934288025, "learning_rate": 2.2754670299730115e-06, "loss": 0.4892, "step": 16530 }, { "epoch": 0.6929566247131195, "grad_norm": 1.712449312210083, "learning_rate": 2.2748978396410285e-06, "loss": 0.5004, "step": 16531 }, { "epoch": 0.6929985433281285, "grad_norm": 2.1346428394317627, "learning_rate": 2.274328699542535e-06, "loss": 0.4991, "step": 16532 }, { "epoch": 0.6930404619431374, "grad_norm": 1.538938045501709, "learning_rate": 2.2737596096880254e-06, "loss": 0.4652, "step": 16533 }, { "epoch": 0.6930823805581463, "grad_norm": 1.7783797979354858, "learning_rate": 2.273190570087987e-06, "loss": 0.4716, "step": 16534 }, { "epoch": 0.6931242991731553, "grad_norm": 1.9332019090652466, "learning_rate": 2.2726215807529106e-06, "loss": 0.4786, "step": 16535 }, { "epoch": 0.6931662177881642, "grad_norm": 1.8031128644943237, "learning_rate": 2.272052641693287e-06, "loss": 0.4531, "step": 16536 }, { "epoch": 0.6932081364031732, "grad_norm": 1.9239001274108887, "learning_rate": 2.2714837529195992e-06, "loss": 0.4372, "step": 16537 }, { "epoch": 0.6932500550181822, "grad_norm": 1.5904619693756104, "learning_rate": 2.270914914442337e-06, "loss": 0.4133, "step": 16538 }, { "epoch": 0.6932919736331912, "grad_norm": 2.12890625, "learning_rate": 2.270346126271987e-06, "loss": 0.4556, "step": 16539 }, { "epoch": 0.6933338922482001, "grad_norm": 1.810502290725708, "learning_rate": 2.2697773884190327e-06, "loss": 0.4851, "step": 16540 }, { "epoch": 0.6933758108632091, "grad_norm": 1.845729112625122, "learning_rate": 2.2692087008939556e-06, "loss": 0.5439, "step": 16541 }, { "epoch": 0.693417729478218, "grad_norm": 1.958133578300476, "learning_rate": 2.2686400637072432e-06, "loss": 0.4838, "step": 16542 }, { "epoch": 0.693459648093227, "grad_norm": 1.826317310333252, "learning_rate": 2.268071476869373e-06, "loss": 0.4648, "step": 16543 }, { "epoch": 0.6935015667082359, "grad_norm": 1.8710613250732422, "learning_rate": 2.267502940390829e-06, "loss": 0.512, "step": 16544 }, { "epoch": 0.693543485323245, "grad_norm": 1.7967439889907837, "learning_rate": 2.266934454282092e-06, "loss": 0.4919, "step": 16545 }, { "epoch": 0.6935854039382539, "grad_norm": 1.63448166847229, "learning_rate": 2.266366018553638e-06, "loss": 0.4363, "step": 16546 }, { "epoch": 0.6936273225532629, "grad_norm": 2.849565267562866, "learning_rate": 2.2657976332159474e-06, "loss": 0.5341, "step": 16547 }, { "epoch": 0.6936692411682718, "grad_norm": 1.9308879375457764, "learning_rate": 2.2652292982795e-06, "loss": 0.5236, "step": 16548 }, { "epoch": 0.6937111597832808, "grad_norm": 2.2637901306152344, "learning_rate": 2.264661013754769e-06, "loss": 0.463, "step": 16549 }, { "epoch": 0.6937530783982897, "grad_norm": 1.655425786972046, "learning_rate": 2.2640927796522296e-06, "loss": 0.4664, "step": 16550 }, { "epoch": 0.6937949970132987, "grad_norm": 2.65371036529541, "learning_rate": 2.2635245959823587e-06, "loss": 0.5244, "step": 16551 }, { "epoch": 0.6938369156283076, "grad_norm": 1.7521464824676514, "learning_rate": 2.262956462755627e-06, "loss": 0.4761, "step": 16552 }, { "epoch": 0.6938788342433166, "grad_norm": 1.7756884098052979, "learning_rate": 2.262388379982509e-06, "loss": 0.4752, "step": 16553 }, { "epoch": 0.6939207528583256, "grad_norm": 1.815452218055725, "learning_rate": 2.261820347673479e-06, "loss": 0.5005, "step": 16554 }, { "epoch": 0.6939626714733346, "grad_norm": 1.4306421279907227, "learning_rate": 2.2612523658390032e-06, "loss": 0.4593, "step": 16555 }, { "epoch": 0.6940045900883435, "grad_norm": 2.397122383117676, "learning_rate": 2.260684434489556e-06, "loss": 0.4977, "step": 16556 }, { "epoch": 0.6940465087033525, "grad_norm": 1.8080172538757324, "learning_rate": 2.2601165536356025e-06, "loss": 0.512, "step": 16557 }, { "epoch": 0.6940884273183614, "grad_norm": 1.6047600507736206, "learning_rate": 2.2595487232876117e-06, "loss": 0.4528, "step": 16558 }, { "epoch": 0.6941303459333703, "grad_norm": 1.925586462020874, "learning_rate": 2.2589809434560543e-06, "loss": 0.4928, "step": 16559 }, { "epoch": 0.6941722645483793, "grad_norm": 2.0772969722747803, "learning_rate": 2.2584132141513937e-06, "loss": 0.5207, "step": 16560 }, { "epoch": 0.6942141831633882, "grad_norm": 1.7599804401397705, "learning_rate": 2.2578455353840934e-06, "loss": 0.5139, "step": 16561 }, { "epoch": 0.6942561017783972, "grad_norm": 1.72782301902771, "learning_rate": 2.2572779071646217e-06, "loss": 0.515, "step": 16562 }, { "epoch": 0.6942980203934062, "grad_norm": 1.8930200338363647, "learning_rate": 2.2567103295034383e-06, "loss": 0.526, "step": 16563 }, { "epoch": 0.6943399390084152, "grad_norm": 1.689746618270874, "learning_rate": 2.256142802411007e-06, "loss": 0.4957, "step": 16564 }, { "epoch": 0.6943818576234241, "grad_norm": 1.969151258468628, "learning_rate": 2.255575325897792e-06, "loss": 0.4757, "step": 16565 }, { "epoch": 0.6944237762384331, "grad_norm": 1.8462791442871094, "learning_rate": 2.2550078999742505e-06, "loss": 0.4567, "step": 16566 }, { "epoch": 0.694465694853442, "grad_norm": 1.5812569856643677, "learning_rate": 2.2544405246508428e-06, "loss": 0.4637, "step": 16567 }, { "epoch": 0.694507613468451, "grad_norm": 1.8604952096939087, "learning_rate": 2.25387319993803e-06, "loss": 0.5438, "step": 16568 }, { "epoch": 0.6945495320834599, "grad_norm": 2.2956011295318604, "learning_rate": 2.2533059258462692e-06, "loss": 0.4755, "step": 16569 }, { "epoch": 0.694591450698469, "grad_norm": 2.041572093963623, "learning_rate": 2.252738702386014e-06, "loss": 0.438, "step": 16570 }, { "epoch": 0.6946333693134779, "grad_norm": 2.071621894836426, "learning_rate": 2.2521715295677253e-06, "loss": 0.4621, "step": 16571 }, { "epoch": 0.6946752879284869, "grad_norm": 1.9190279245376587, "learning_rate": 2.2516044074018534e-06, "loss": 0.5116, "step": 16572 }, { "epoch": 0.6947172065434958, "grad_norm": 2.2046165466308594, "learning_rate": 2.251037335898855e-06, "loss": 0.5525, "step": 16573 }, { "epoch": 0.6947591251585048, "grad_norm": 1.9092674255371094, "learning_rate": 2.2504703150691847e-06, "loss": 0.4568, "step": 16574 }, { "epoch": 0.6948010437735137, "grad_norm": 1.861490249633789, "learning_rate": 2.249903344923291e-06, "loss": 0.4971, "step": 16575 }, { "epoch": 0.6948429623885227, "grad_norm": 2.116396427154541, "learning_rate": 2.2493364254716276e-06, "loss": 0.5188, "step": 16576 }, { "epoch": 0.6948848810035316, "grad_norm": 1.961286187171936, "learning_rate": 2.248769556724646e-06, "loss": 0.5395, "step": 16577 }, { "epoch": 0.6949267996185406, "grad_norm": 1.8732562065124512, "learning_rate": 2.2482027386927945e-06, "loss": 0.5094, "step": 16578 }, { "epoch": 0.6949687182335496, "grad_norm": 1.9742141962051392, "learning_rate": 2.2476359713865194e-06, "loss": 0.4807, "step": 16579 }, { "epoch": 0.6950106368485586, "grad_norm": 1.8798096179962158, "learning_rate": 2.2470692548162714e-06, "loss": 0.4958, "step": 16580 }, { "epoch": 0.6950525554635675, "grad_norm": 1.4894773960113525, "learning_rate": 2.2465025889924944e-06, "loss": 0.4717, "step": 16581 }, { "epoch": 0.6950944740785765, "grad_norm": 1.4968531131744385, "learning_rate": 2.245935973925636e-06, "loss": 0.4705, "step": 16582 }, { "epoch": 0.6951363926935854, "grad_norm": 1.6788386106491089, "learning_rate": 2.245369409626142e-06, "loss": 0.466, "step": 16583 }, { "epoch": 0.6951783113085943, "grad_norm": 1.8370441198349, "learning_rate": 2.2448028961044527e-06, "loss": 0.5381, "step": 16584 }, { "epoch": 0.6952202299236033, "grad_norm": 1.7806460857391357, "learning_rate": 2.244236433371013e-06, "loss": 0.5238, "step": 16585 }, { "epoch": 0.6952621485386122, "grad_norm": 1.5878942012786865, "learning_rate": 2.243670021436267e-06, "loss": 0.4831, "step": 16586 }, { "epoch": 0.6953040671536213, "grad_norm": 1.9462369680404663, "learning_rate": 2.2431036603106543e-06, "loss": 0.5399, "step": 16587 }, { "epoch": 0.6953459857686302, "grad_norm": 2.032632350921631, "learning_rate": 2.2425373500046115e-06, "loss": 0.4959, "step": 16588 }, { "epoch": 0.6953879043836392, "grad_norm": 2.0341522693634033, "learning_rate": 2.241971090528583e-06, "loss": 0.502, "step": 16589 }, { "epoch": 0.6954298229986481, "grad_norm": 1.7852269411087036, "learning_rate": 2.2414048818930023e-06, "loss": 0.4892, "step": 16590 }, { "epoch": 0.6954717416136571, "grad_norm": 2.1181843280792236, "learning_rate": 2.2408387241083098e-06, "loss": 0.5117, "step": 16591 }, { "epoch": 0.695513660228666, "grad_norm": 1.8739941120147705, "learning_rate": 2.240272617184942e-06, "loss": 0.5168, "step": 16592 }, { "epoch": 0.695555578843675, "grad_norm": 3.822512149810791, "learning_rate": 2.2397065611333318e-06, "loss": 0.4714, "step": 16593 }, { "epoch": 0.6955974974586839, "grad_norm": 1.7885569334030151, "learning_rate": 2.239140555963915e-06, "loss": 0.4891, "step": 16594 }, { "epoch": 0.695639416073693, "grad_norm": 1.7546848058700562, "learning_rate": 2.238574601687127e-06, "loss": 0.5551, "step": 16595 }, { "epoch": 0.6956813346887019, "grad_norm": 1.9005956649780273, "learning_rate": 2.238008698313399e-06, "loss": 0.5427, "step": 16596 }, { "epoch": 0.6957232533037109, "grad_norm": 1.9939483404159546, "learning_rate": 2.23744284585316e-06, "loss": 0.4409, "step": 16597 }, { "epoch": 0.6957651719187198, "grad_norm": 1.8849018812179565, "learning_rate": 2.2368770443168452e-06, "loss": 0.5066, "step": 16598 }, { "epoch": 0.6958070905337288, "grad_norm": 1.9147915840148926, "learning_rate": 2.236311293714881e-06, "loss": 0.5335, "step": 16599 }, { "epoch": 0.6958490091487377, "grad_norm": 1.8141121864318848, "learning_rate": 2.235745594057696e-06, "loss": 0.4333, "step": 16600 }, { "epoch": 0.6958909277637467, "grad_norm": 2.1067450046539307, "learning_rate": 2.2351799453557223e-06, "loss": 0.5211, "step": 16601 }, { "epoch": 0.6959328463787556, "grad_norm": 1.7263400554656982, "learning_rate": 2.234614347619382e-06, "loss": 0.4362, "step": 16602 }, { "epoch": 0.6959747649937646, "grad_norm": 5.275597095489502, "learning_rate": 2.2340488008591033e-06, "loss": 0.5674, "step": 16603 }, { "epoch": 0.6960166836087736, "grad_norm": 1.8704135417938232, "learning_rate": 2.233483305085313e-06, "loss": 0.4434, "step": 16604 }, { "epoch": 0.6960586022237826, "grad_norm": 1.6804873943328857, "learning_rate": 2.232917860308431e-06, "loss": 0.4812, "step": 16605 }, { "epoch": 0.6961005208387915, "grad_norm": 4.571805477142334, "learning_rate": 2.232352466538885e-06, "loss": 0.4926, "step": 16606 }, { "epoch": 0.6961424394538005, "grad_norm": 1.8450002670288086, "learning_rate": 2.231787123787095e-06, "loss": 0.4804, "step": 16607 }, { "epoch": 0.6961843580688094, "grad_norm": 1.818721890449524, "learning_rate": 2.231221832063481e-06, "loss": 0.4958, "step": 16608 }, { "epoch": 0.6962262766838183, "grad_norm": 1.45090651512146, "learning_rate": 2.2306565913784643e-06, "loss": 0.4204, "step": 16609 }, { "epoch": 0.6962681952988273, "grad_norm": 2.065354108810425, "learning_rate": 2.2300914017424672e-06, "loss": 0.5352, "step": 16610 }, { "epoch": 0.6963101139138362, "grad_norm": 1.7655926942825317, "learning_rate": 2.2295262631659038e-06, "loss": 0.443, "step": 16611 }, { "epoch": 0.6963520325288453, "grad_norm": 1.803741693496704, "learning_rate": 2.2289611756591937e-06, "loss": 0.4991, "step": 16612 }, { "epoch": 0.6963939511438542, "grad_norm": 2.0348269939422607, "learning_rate": 2.228396139232755e-06, "loss": 0.5066, "step": 16613 }, { "epoch": 0.6964358697588632, "grad_norm": 1.8123698234558105, "learning_rate": 2.2278311538970006e-06, "loss": 0.496, "step": 16614 }, { "epoch": 0.6964777883738721, "grad_norm": 1.6656121015548706, "learning_rate": 2.227266219662348e-06, "loss": 0.4776, "step": 16615 }, { "epoch": 0.6965197069888811, "grad_norm": 1.821157693862915, "learning_rate": 2.2267013365392093e-06, "loss": 0.5139, "step": 16616 }, { "epoch": 0.69656162560389, "grad_norm": 2.00408673286438, "learning_rate": 2.2261365045379955e-06, "loss": 0.4569, "step": 16617 }, { "epoch": 0.696603544218899, "grad_norm": 1.6887792348861694, "learning_rate": 2.225571723669121e-06, "loss": 0.4132, "step": 16618 }, { "epoch": 0.6966454628339079, "grad_norm": 1.8817591667175293, "learning_rate": 2.225006993942998e-06, "loss": 0.4488, "step": 16619 }, { "epoch": 0.696687381448917, "grad_norm": 1.9561091661453247, "learning_rate": 2.224442315370033e-06, "loss": 0.4299, "step": 16620 }, { "epoch": 0.6967293000639259, "grad_norm": 1.743079662322998, "learning_rate": 2.2238776879606362e-06, "loss": 0.4952, "step": 16621 }, { "epoch": 0.6967712186789349, "grad_norm": 2.650953769683838, "learning_rate": 2.2233131117252188e-06, "loss": 0.5228, "step": 16622 }, { "epoch": 0.6968131372939438, "grad_norm": 1.804583191871643, "learning_rate": 2.2227485866741834e-06, "loss": 0.5035, "step": 16623 }, { "epoch": 0.6968550559089528, "grad_norm": 1.9043534994125366, "learning_rate": 2.22218411281794e-06, "loss": 0.5282, "step": 16624 }, { "epoch": 0.6968969745239617, "grad_norm": 2.2194933891296387, "learning_rate": 2.221619690166893e-06, "loss": 0.4726, "step": 16625 }, { "epoch": 0.6969388931389707, "grad_norm": 1.7581806182861328, "learning_rate": 2.221055318731443e-06, "loss": 0.4993, "step": 16626 }, { "epoch": 0.6969808117539796, "grad_norm": 1.5506399869918823, "learning_rate": 2.2204909985219975e-06, "loss": 0.467, "step": 16627 }, { "epoch": 0.6970227303689887, "grad_norm": 1.7744979858398438, "learning_rate": 2.2199267295489595e-06, "loss": 0.5212, "step": 16628 }, { "epoch": 0.6970646489839976, "grad_norm": 2.423011064529419, "learning_rate": 2.219362511822727e-06, "loss": 0.441, "step": 16629 }, { "epoch": 0.6971065675990066, "grad_norm": 1.9427862167358398, "learning_rate": 2.2187983453537045e-06, "loss": 0.4841, "step": 16630 }, { "epoch": 0.6971484862140155, "grad_norm": 1.7740944623947144, "learning_rate": 2.218234230152287e-06, "loss": 0.4848, "step": 16631 }, { "epoch": 0.6971904048290245, "grad_norm": 2.1038739681243896, "learning_rate": 2.2176701662288763e-06, "loss": 0.5464, "step": 16632 }, { "epoch": 0.6972323234440334, "grad_norm": 1.8782914876937866, "learning_rate": 2.2171061535938714e-06, "loss": 0.4559, "step": 16633 }, { "epoch": 0.6972742420590423, "grad_norm": 1.759207844734192, "learning_rate": 2.2165421922576675e-06, "loss": 0.5148, "step": 16634 }, { "epoch": 0.6973161606740513, "grad_norm": 1.6448622941970825, "learning_rate": 2.2159782822306588e-06, "loss": 0.4452, "step": 16635 }, { "epoch": 0.6973580792890602, "grad_norm": 1.6463462114334106, "learning_rate": 2.2154144235232427e-06, "loss": 0.4933, "step": 16636 }, { "epoch": 0.6973999979040693, "grad_norm": 1.8512517213821411, "learning_rate": 2.2148506161458105e-06, "loss": 0.4958, "step": 16637 }, { "epoch": 0.6974419165190782, "grad_norm": 3.9116735458374023, "learning_rate": 2.214286860108757e-06, "loss": 0.4809, "step": 16638 }, { "epoch": 0.6974838351340872, "grad_norm": 1.8283973932266235, "learning_rate": 2.213723155422476e-06, "loss": 0.493, "step": 16639 }, { "epoch": 0.6975257537490961, "grad_norm": 1.9543087482452393, "learning_rate": 2.2131595020973547e-06, "loss": 0.4811, "step": 16640 }, { "epoch": 0.6975676723641051, "grad_norm": 2.3423640727996826, "learning_rate": 2.2125959001437853e-06, "loss": 0.4397, "step": 16641 }, { "epoch": 0.697609590979114, "grad_norm": 2.2696268558502197, "learning_rate": 2.212032349572159e-06, "loss": 0.4599, "step": 16642 }, { "epoch": 0.697651509594123, "grad_norm": 1.7133756875991821, "learning_rate": 2.211468850392862e-06, "loss": 0.4732, "step": 16643 }, { "epoch": 0.6976934282091319, "grad_norm": 1.8691331148147583, "learning_rate": 2.2109054026162798e-06, "loss": 0.5075, "step": 16644 }, { "epoch": 0.697735346824141, "grad_norm": 1.806989073753357, "learning_rate": 2.2103420062528024e-06, "loss": 0.4456, "step": 16645 }, { "epoch": 0.6977772654391499, "grad_norm": 2.0133121013641357, "learning_rate": 2.2097786613128126e-06, "loss": 0.4958, "step": 16646 }, { "epoch": 0.6978191840541589, "grad_norm": 1.9921157360076904, "learning_rate": 2.2092153678066946e-06, "loss": 0.4474, "step": 16647 }, { "epoch": 0.6978611026691678, "grad_norm": 1.805110216140747, "learning_rate": 2.208652125744836e-06, "loss": 0.4655, "step": 16648 }, { "epoch": 0.6979030212841768, "grad_norm": 1.9846115112304688, "learning_rate": 2.208088935137614e-06, "loss": 0.5225, "step": 16649 }, { "epoch": 0.6979449398991857, "grad_norm": 1.5178163051605225, "learning_rate": 2.207525795995413e-06, "loss": 0.5058, "step": 16650 }, { "epoch": 0.6979868585141947, "grad_norm": 1.7950828075408936, "learning_rate": 2.206962708328615e-06, "loss": 0.4946, "step": 16651 }, { "epoch": 0.6980287771292036, "grad_norm": 2.0134878158569336, "learning_rate": 2.2063996721475966e-06, "loss": 0.4898, "step": 16652 }, { "epoch": 0.6980706957442127, "grad_norm": 1.9980324506759644, "learning_rate": 2.2058366874627404e-06, "loss": 0.4947, "step": 16653 }, { "epoch": 0.6981126143592216, "grad_norm": 1.5339939594268799, "learning_rate": 2.2052737542844216e-06, "loss": 0.4591, "step": 16654 }, { "epoch": 0.6981545329742306, "grad_norm": 2.1881446838378906, "learning_rate": 2.204710872623016e-06, "loss": 0.5093, "step": 16655 }, { "epoch": 0.6981964515892395, "grad_norm": 1.731752872467041, "learning_rate": 2.2041480424889013e-06, "loss": 0.4689, "step": 16656 }, { "epoch": 0.6982383702042485, "grad_norm": 1.6749639511108398, "learning_rate": 2.2035852638924534e-06, "loss": 0.4834, "step": 16657 }, { "epoch": 0.6982802888192574, "grad_norm": 1.6884150505065918, "learning_rate": 2.2030225368440437e-06, "loss": 0.4327, "step": 16658 }, { "epoch": 0.6983222074342663, "grad_norm": 1.7678849697113037, "learning_rate": 2.202459861354047e-06, "loss": 0.4944, "step": 16659 }, { "epoch": 0.6983641260492753, "grad_norm": 1.715132713317871, "learning_rate": 2.2018972374328373e-06, "loss": 0.466, "step": 16660 }, { "epoch": 0.6984060446642842, "grad_norm": 1.5813167095184326, "learning_rate": 2.2013346650907815e-06, "loss": 0.4839, "step": 16661 }, { "epoch": 0.6984479632792933, "grad_norm": 1.7668564319610596, "learning_rate": 2.200772144338254e-06, "loss": 0.4165, "step": 16662 }, { "epoch": 0.6984898818943022, "grad_norm": 1.7317529916763306, "learning_rate": 2.2002096751856224e-06, "loss": 0.5081, "step": 16663 }, { "epoch": 0.6985318005093112, "grad_norm": 1.8467639684677124, "learning_rate": 2.1996472576432532e-06, "loss": 0.4925, "step": 16664 }, { "epoch": 0.6985737191243201, "grad_norm": 2.2741994857788086, "learning_rate": 2.1990848917215147e-06, "loss": 0.4901, "step": 16665 }, { "epoch": 0.6986156377393291, "grad_norm": 2.1354012489318848, "learning_rate": 2.1985225774307767e-06, "loss": 0.4873, "step": 16666 }, { "epoch": 0.698657556354338, "grad_norm": 2.438943862915039, "learning_rate": 2.1979603147814e-06, "loss": 0.5134, "step": 16667 }, { "epoch": 0.698699474969347, "grad_norm": 1.8396363258361816, "learning_rate": 2.1973981037837506e-06, "loss": 0.4983, "step": 16668 }, { "epoch": 0.6987413935843559, "grad_norm": 1.7786269187927246, "learning_rate": 2.196835944448195e-06, "loss": 0.5045, "step": 16669 }, { "epoch": 0.698783312199365, "grad_norm": 6.096492767333984, "learning_rate": 2.1962738367850915e-06, "loss": 0.4562, "step": 16670 }, { "epoch": 0.6988252308143739, "grad_norm": 1.934525728225708, "learning_rate": 2.195711780804806e-06, "loss": 0.4543, "step": 16671 }, { "epoch": 0.6988671494293829, "grad_norm": 1.7727515697479248, "learning_rate": 2.1951497765176973e-06, "loss": 0.5189, "step": 16672 }, { "epoch": 0.6989090680443918, "grad_norm": 1.8646845817565918, "learning_rate": 2.1945878239341224e-06, "loss": 0.4499, "step": 16673 }, { "epoch": 0.6989509866594008, "grad_norm": 1.6404780149459839, "learning_rate": 2.1940259230644425e-06, "loss": 0.4912, "step": 16674 }, { "epoch": 0.6989929052744097, "grad_norm": 1.7443174123764038, "learning_rate": 2.193464073919018e-06, "loss": 0.559, "step": 16675 }, { "epoch": 0.6990348238894187, "grad_norm": 1.8885186910629272, "learning_rate": 2.1929022765082007e-06, "loss": 0.4737, "step": 16676 }, { "epoch": 0.6990767425044276, "grad_norm": 1.8948395252227783, "learning_rate": 2.1923405308423497e-06, "loss": 0.4985, "step": 16677 }, { "epoch": 0.6991186611194367, "grad_norm": 2.06471848487854, "learning_rate": 2.1917788369318216e-06, "loss": 0.5303, "step": 16678 }, { "epoch": 0.6991605797344456, "grad_norm": 1.6033624410629272, "learning_rate": 2.191217194786966e-06, "loss": 0.4602, "step": 16679 }, { "epoch": 0.6992024983494546, "grad_norm": 2.062310218811035, "learning_rate": 2.190655604418141e-06, "loss": 0.5072, "step": 16680 }, { "epoch": 0.6992444169644635, "grad_norm": 1.9498798847198486, "learning_rate": 2.190094065835696e-06, "loss": 0.4533, "step": 16681 }, { "epoch": 0.6992863355794725, "grad_norm": 1.5978482961654663, "learning_rate": 2.1895325790499793e-06, "loss": 0.4483, "step": 16682 }, { "epoch": 0.6993282541944814, "grad_norm": 1.7202599048614502, "learning_rate": 2.1889711440713452e-06, "loss": 0.4797, "step": 16683 }, { "epoch": 0.6993701728094903, "grad_norm": 1.7040261030197144, "learning_rate": 2.188409760910144e-06, "loss": 0.5124, "step": 16684 }, { "epoch": 0.6994120914244993, "grad_norm": 1.8172640800476074, "learning_rate": 2.1878484295767196e-06, "loss": 0.5414, "step": 16685 }, { "epoch": 0.6994540100395082, "grad_norm": 1.6473743915557861, "learning_rate": 2.187287150081422e-06, "loss": 0.4588, "step": 16686 }, { "epoch": 0.6994959286545173, "grad_norm": 1.7803386449813843, "learning_rate": 2.1867259224345986e-06, "loss": 0.5147, "step": 16687 }, { "epoch": 0.6995378472695262, "grad_norm": 1.880175232887268, "learning_rate": 2.1861647466465925e-06, "loss": 0.5223, "step": 16688 }, { "epoch": 0.6995797658845352, "grad_norm": 2.643040418624878, "learning_rate": 2.1856036227277505e-06, "loss": 0.5014, "step": 16689 }, { "epoch": 0.6996216844995441, "grad_norm": 1.8819668292999268, "learning_rate": 2.185042550688413e-06, "loss": 0.5191, "step": 16690 }, { "epoch": 0.6996636031145531, "grad_norm": 2.938178539276123, "learning_rate": 2.1844815305389274e-06, "loss": 0.5216, "step": 16691 }, { "epoch": 0.699705521729562, "grad_norm": 1.9470696449279785, "learning_rate": 2.1839205622896296e-06, "loss": 0.5221, "step": 16692 }, { "epoch": 0.699747440344571, "grad_norm": 2.2204878330230713, "learning_rate": 2.1833596459508654e-06, "loss": 0.4665, "step": 16693 }, { "epoch": 0.6997893589595799, "grad_norm": 1.8975375890731812, "learning_rate": 2.18279878153297e-06, "loss": 0.5054, "step": 16694 }, { "epoch": 0.699831277574589, "grad_norm": 1.6527806520462036, "learning_rate": 2.182237969046285e-06, "loss": 0.4763, "step": 16695 }, { "epoch": 0.6998731961895979, "grad_norm": 2.1628262996673584, "learning_rate": 2.181677208501149e-06, "loss": 0.5354, "step": 16696 }, { "epoch": 0.6999151148046069, "grad_norm": 2.134239673614502, "learning_rate": 2.1811164999078952e-06, "loss": 0.5704, "step": 16697 }, { "epoch": 0.6999570334196158, "grad_norm": 2.0157861709594727, "learning_rate": 2.180555843276865e-06, "loss": 0.5212, "step": 16698 }, { "epoch": 0.6999989520346248, "grad_norm": 1.821472406387329, "learning_rate": 2.179995238618387e-06, "loss": 0.4691, "step": 16699 }, { "epoch": 0.7000408706496337, "grad_norm": 1.9083476066589355, "learning_rate": 2.1794346859428e-06, "loss": 0.5056, "step": 16700 }, { "epoch": 0.7000827892646427, "grad_norm": 1.7742263078689575, "learning_rate": 2.178874185260434e-06, "loss": 0.4773, "step": 16701 }, { "epoch": 0.7001247078796516, "grad_norm": 2.3110439777374268, "learning_rate": 2.1783137365816243e-06, "loss": 0.5065, "step": 16702 }, { "epoch": 0.7001666264946607, "grad_norm": 5.62224817276001, "learning_rate": 2.177753339916698e-06, "loss": 0.4645, "step": 16703 }, { "epoch": 0.7002085451096696, "grad_norm": 2.0978050231933594, "learning_rate": 2.1771929952759885e-06, "loss": 0.5407, "step": 16704 }, { "epoch": 0.7002504637246786, "grad_norm": 1.7794373035430908, "learning_rate": 2.176632702669822e-06, "loss": 0.5083, "step": 16705 }, { "epoch": 0.7002923823396875, "grad_norm": 1.6732165813446045, "learning_rate": 2.1760724621085287e-06, "loss": 0.4304, "step": 16706 }, { "epoch": 0.7003343009546965, "grad_norm": 1.906073808670044, "learning_rate": 2.1755122736024375e-06, "loss": 0.4942, "step": 16707 }, { "epoch": 0.7003762195697054, "grad_norm": 2.576326847076416, "learning_rate": 2.174952137161871e-06, "loss": 0.4469, "step": 16708 }, { "epoch": 0.7004181381847143, "grad_norm": 1.7149449586868286, "learning_rate": 2.174392052797158e-06, "loss": 0.4558, "step": 16709 }, { "epoch": 0.7004600567997233, "grad_norm": 6.435934066772461, "learning_rate": 2.1738320205186213e-06, "loss": 0.4766, "step": 16710 }, { "epoch": 0.7005019754147322, "grad_norm": 1.7367205619812012, "learning_rate": 2.1732720403365827e-06, "loss": 0.4729, "step": 16711 }, { "epoch": 0.7005438940297413, "grad_norm": 2.1814231872558594, "learning_rate": 2.172712112261366e-06, "loss": 0.5274, "step": 16712 }, { "epoch": 0.7005858126447502, "grad_norm": 2.169543743133545, "learning_rate": 2.1721522363032942e-06, "loss": 0.5266, "step": 16713 }, { "epoch": 0.7006277312597592, "grad_norm": 1.7335740327835083, "learning_rate": 2.1715924124726856e-06, "loss": 0.4783, "step": 16714 }, { "epoch": 0.7006696498747681, "grad_norm": 2.0358948707580566, "learning_rate": 2.1710326407798597e-06, "loss": 0.4992, "step": 16715 }, { "epoch": 0.7007115684897771, "grad_norm": 1.9443597793579102, "learning_rate": 2.170472921235139e-06, "loss": 0.4869, "step": 16716 }, { "epoch": 0.700753487104786, "grad_norm": 1.881137728691101, "learning_rate": 2.169913253848836e-06, "loss": 0.4317, "step": 16717 }, { "epoch": 0.700795405719795, "grad_norm": 1.9733473062515259, "learning_rate": 2.1693536386312713e-06, "loss": 0.5277, "step": 16718 }, { "epoch": 0.700837324334804, "grad_norm": 1.7515000104904175, "learning_rate": 2.1687940755927593e-06, "loss": 0.4903, "step": 16719 }, { "epoch": 0.700879242949813, "grad_norm": 2.1590330600738525, "learning_rate": 2.1682345647436126e-06, "loss": 0.5516, "step": 16720 }, { "epoch": 0.7009211615648219, "grad_norm": 2.1692075729370117, "learning_rate": 2.167675106094147e-06, "loss": 0.4871, "step": 16721 }, { "epoch": 0.7009630801798309, "grad_norm": 1.8659976720809937, "learning_rate": 2.167115699654677e-06, "loss": 0.5213, "step": 16722 }, { "epoch": 0.7010049987948398, "grad_norm": 1.9705063104629517, "learning_rate": 2.166556345435511e-06, "loss": 0.4666, "step": 16723 }, { "epoch": 0.7010469174098488, "grad_norm": 2.248577117919922, "learning_rate": 2.165997043446961e-06, "loss": 0.5158, "step": 16724 }, { "epoch": 0.7010888360248577, "grad_norm": 2.087381362915039, "learning_rate": 2.1654377936993403e-06, "loss": 0.5394, "step": 16725 }, { "epoch": 0.7011307546398667, "grad_norm": 2.07902455329895, "learning_rate": 2.1648785962029534e-06, "loss": 0.4741, "step": 16726 }, { "epoch": 0.7011726732548756, "grad_norm": 1.740891456604004, "learning_rate": 2.1643194509681114e-06, "loss": 0.4584, "step": 16727 }, { "epoch": 0.7012145918698847, "grad_norm": 1.7276365756988525, "learning_rate": 2.1637603580051202e-06, "loss": 0.5034, "step": 16728 }, { "epoch": 0.7012565104848936, "grad_norm": 2.9843180179595947, "learning_rate": 2.1632013173242843e-06, "loss": 0.4965, "step": 16729 }, { "epoch": 0.7012984290999026, "grad_norm": 2.673015594482422, "learning_rate": 2.1626423289359105e-06, "loss": 0.5006, "step": 16730 }, { "epoch": 0.7013403477149115, "grad_norm": 3.4547271728515625, "learning_rate": 2.1620833928503045e-06, "loss": 0.5279, "step": 16731 }, { "epoch": 0.7013822663299205, "grad_norm": 1.840647578239441, "learning_rate": 2.1615245090777655e-06, "loss": 0.505, "step": 16732 }, { "epoch": 0.7014241849449294, "grad_norm": 1.8209911584854126, "learning_rate": 2.160965677628599e-06, "loss": 0.4604, "step": 16733 }, { "epoch": 0.7014661035599383, "grad_norm": 1.7404886484146118, "learning_rate": 2.1604068985131062e-06, "loss": 0.4944, "step": 16734 }, { "epoch": 0.7015080221749473, "grad_norm": 2.0289089679718018, "learning_rate": 2.1598481717415854e-06, "loss": 0.4936, "step": 16735 }, { "epoch": 0.7015499407899563, "grad_norm": 2.127363681793213, "learning_rate": 2.159289497324339e-06, "loss": 0.5342, "step": 16736 }, { "epoch": 0.7015918594049653, "grad_norm": 2.059767484664917, "learning_rate": 2.1587308752716613e-06, "loss": 0.47, "step": 16737 }, { "epoch": 0.7016337780199742, "grad_norm": 2.303772211074829, "learning_rate": 2.158172305593854e-06, "loss": 0.4981, "step": 16738 }, { "epoch": 0.7016756966349832, "grad_norm": 1.7335307598114014, "learning_rate": 2.15761378830121e-06, "loss": 0.4885, "step": 16739 }, { "epoch": 0.7017176152499921, "grad_norm": 2.0355424880981445, "learning_rate": 2.1570553234040283e-06, "loss": 0.4556, "step": 16740 }, { "epoch": 0.7017595338650011, "grad_norm": 1.9086530208587646, "learning_rate": 2.156496910912599e-06, "loss": 0.5175, "step": 16741 }, { "epoch": 0.70180145248001, "grad_norm": 1.896264910697937, "learning_rate": 2.1559385508372182e-06, "loss": 0.4357, "step": 16742 }, { "epoch": 0.701843371095019, "grad_norm": 1.8332902193069458, "learning_rate": 2.1553802431881804e-06, "loss": 0.5117, "step": 16743 }, { "epoch": 0.701885289710028, "grad_norm": 2.0300400257110596, "learning_rate": 2.154821987975774e-06, "loss": 0.4677, "step": 16744 }, { "epoch": 0.701927208325037, "grad_norm": 1.9513269662857056, "learning_rate": 2.15426378521029e-06, "loss": 0.5218, "step": 16745 }, { "epoch": 0.7019691269400459, "grad_norm": 1.9728790521621704, "learning_rate": 2.153705634902021e-06, "loss": 0.4813, "step": 16746 }, { "epoch": 0.7020110455550549, "grad_norm": 3.8877837657928467, "learning_rate": 2.153147537061254e-06, "loss": 0.4797, "step": 16747 }, { "epoch": 0.7020529641700638, "grad_norm": 2.031062602996826, "learning_rate": 2.152589491698275e-06, "loss": 0.4953, "step": 16748 }, { "epoch": 0.7020948827850728, "grad_norm": 2.0065102577209473, "learning_rate": 2.1520314988233736e-06, "loss": 0.5334, "step": 16749 }, { "epoch": 0.7021368014000817, "grad_norm": 2.1128621101379395, "learning_rate": 2.151473558446833e-06, "loss": 0.4946, "step": 16750 }, { "epoch": 0.7021787200150907, "grad_norm": 2.2651190757751465, "learning_rate": 2.1509156705789385e-06, "loss": 0.485, "step": 16751 }, { "epoch": 0.7022206386300996, "grad_norm": 1.8341730833053589, "learning_rate": 2.1503578352299777e-06, "loss": 0.5069, "step": 16752 }, { "epoch": 0.7022625572451087, "grad_norm": 1.804317593574524, "learning_rate": 2.149800052410228e-06, "loss": 0.5296, "step": 16753 }, { "epoch": 0.7023044758601176, "grad_norm": 2.111598014831543, "learning_rate": 2.1492423221299748e-06, "loss": 0.4589, "step": 16754 }, { "epoch": 0.7023463944751266, "grad_norm": 1.6769894361495972, "learning_rate": 2.1486846443994998e-06, "loss": 0.4568, "step": 16755 }, { "epoch": 0.7023883130901355, "grad_norm": 1.9447362422943115, "learning_rate": 2.1481270192290814e-06, "loss": 0.4437, "step": 16756 }, { "epoch": 0.7024302317051445, "grad_norm": 1.6925510168075562, "learning_rate": 2.147569446628997e-06, "loss": 0.3981, "step": 16757 }, { "epoch": 0.7024721503201534, "grad_norm": 2.007380723953247, "learning_rate": 2.1470119266095285e-06, "loss": 0.5104, "step": 16758 }, { "epoch": 0.7025140689351623, "grad_norm": 1.7526500225067139, "learning_rate": 2.1464544591809493e-06, "loss": 0.4524, "step": 16759 }, { "epoch": 0.7025559875501713, "grad_norm": 1.674448013305664, "learning_rate": 2.145897044353537e-06, "loss": 0.4268, "step": 16760 }, { "epoch": 0.7025979061651803, "grad_norm": 1.4794291257858276, "learning_rate": 2.145339682137569e-06, "loss": 0.4471, "step": 16761 }, { "epoch": 0.7026398247801893, "grad_norm": 2.6141884326934814, "learning_rate": 2.1447823725433155e-06, "loss": 0.4388, "step": 16762 }, { "epoch": 0.7026817433951982, "grad_norm": 1.7006150484085083, "learning_rate": 2.1442251155810518e-06, "loss": 0.4548, "step": 16763 }, { "epoch": 0.7027236620102072, "grad_norm": 1.7912501096725464, "learning_rate": 2.143667911261052e-06, "loss": 0.4879, "step": 16764 }, { "epoch": 0.7027655806252161, "grad_norm": 2.0450849533081055, "learning_rate": 2.143110759593586e-06, "loss": 0.4846, "step": 16765 }, { "epoch": 0.7028074992402251, "grad_norm": 1.6972849369049072, "learning_rate": 2.142553660588921e-06, "loss": 0.5397, "step": 16766 }, { "epoch": 0.702849417855234, "grad_norm": 1.962876319885254, "learning_rate": 2.1419966142573305e-06, "loss": 0.4467, "step": 16767 }, { "epoch": 0.702891336470243, "grad_norm": 2.106959819793701, "learning_rate": 2.1414396206090803e-06, "loss": 0.4981, "step": 16768 }, { "epoch": 0.702933255085252, "grad_norm": 1.8785362243652344, "learning_rate": 2.1408826796544385e-06, "loss": 0.5491, "step": 16769 }, { "epoch": 0.702975173700261, "grad_norm": 2.710977792739868, "learning_rate": 2.140325791403674e-06, "loss": 0.4662, "step": 16770 }, { "epoch": 0.7030170923152699, "grad_norm": 2.2388663291931152, "learning_rate": 2.1397689558670483e-06, "loss": 0.4928, "step": 16771 }, { "epoch": 0.7030590109302789, "grad_norm": 2.064631938934326, "learning_rate": 2.139212173054828e-06, "loss": 0.495, "step": 16772 }, { "epoch": 0.7031009295452878, "grad_norm": 1.861941933631897, "learning_rate": 2.1386554429772773e-06, "loss": 0.404, "step": 16773 }, { "epoch": 0.7031428481602968, "grad_norm": 1.6477587223052979, "learning_rate": 2.1380987656446585e-06, "loss": 0.5098, "step": 16774 }, { "epoch": 0.7031847667753057, "grad_norm": 3.142225980758667, "learning_rate": 2.1375421410672304e-06, "loss": 0.4968, "step": 16775 }, { "epoch": 0.7032266853903147, "grad_norm": 1.786342978477478, "learning_rate": 2.1369855692552577e-06, "loss": 0.4499, "step": 16776 }, { "epoch": 0.7032686040053236, "grad_norm": 1.9922822713851929, "learning_rate": 2.136429050218996e-06, "loss": 0.4689, "step": 16777 }, { "epoch": 0.7033105226203327, "grad_norm": 1.6538169384002686, "learning_rate": 2.135872583968706e-06, "loss": 0.4708, "step": 16778 }, { "epoch": 0.7033524412353416, "grad_norm": 1.513492465019226, "learning_rate": 2.1353161705146476e-06, "loss": 0.4494, "step": 16779 }, { "epoch": 0.7033943598503506, "grad_norm": 2.3951022624969482, "learning_rate": 2.134759809867073e-06, "loss": 0.4942, "step": 16780 }, { "epoch": 0.7034362784653595, "grad_norm": 1.923731803894043, "learning_rate": 2.1342035020362425e-06, "loss": 0.508, "step": 16781 }, { "epoch": 0.7034781970803685, "grad_norm": 1.9219251871109009, "learning_rate": 2.133647247032406e-06, "loss": 0.4639, "step": 16782 }, { "epoch": 0.7035201156953774, "grad_norm": 2.060267925262451, "learning_rate": 2.133091044865822e-06, "loss": 0.4447, "step": 16783 }, { "epoch": 0.7035620343103863, "grad_norm": 1.9969216585159302, "learning_rate": 2.1325348955467403e-06, "loss": 0.4355, "step": 16784 }, { "epoch": 0.7036039529253953, "grad_norm": 1.6598234176635742, "learning_rate": 2.1319787990854147e-06, "loss": 0.4814, "step": 16785 }, { "epoch": 0.7036458715404043, "grad_norm": 1.9645605087280273, "learning_rate": 2.131422755492093e-06, "loss": 0.4867, "step": 16786 }, { "epoch": 0.7036877901554133, "grad_norm": 1.806856393814087, "learning_rate": 2.1308667647770294e-06, "loss": 0.4402, "step": 16787 }, { "epoch": 0.7037297087704222, "grad_norm": 1.8121122121810913, "learning_rate": 2.130310826950468e-06, "loss": 0.5237, "step": 16788 }, { "epoch": 0.7037716273854312, "grad_norm": 1.5672497749328613, "learning_rate": 2.12975494202266e-06, "loss": 0.4703, "step": 16789 }, { "epoch": 0.7038135460004401, "grad_norm": 1.9505592584609985, "learning_rate": 2.1291991100038534e-06, "loss": 0.5098, "step": 16790 }, { "epoch": 0.7038554646154491, "grad_norm": 1.6817617416381836, "learning_rate": 2.1286433309042904e-06, "loss": 0.4381, "step": 16791 }, { "epoch": 0.703897383230458, "grad_norm": 1.7428613901138306, "learning_rate": 2.1280876047342183e-06, "loss": 0.4804, "step": 16792 }, { "epoch": 0.703939301845467, "grad_norm": 1.8947786092758179, "learning_rate": 2.1275319315038824e-06, "loss": 0.4341, "step": 16793 }, { "epoch": 0.703981220460476, "grad_norm": 3.374570846557617, "learning_rate": 2.1269763112235242e-06, "loss": 0.4643, "step": 16794 }, { "epoch": 0.704023139075485, "grad_norm": 2.6013848781585693, "learning_rate": 2.1264207439033844e-06, "loss": 0.4591, "step": 16795 }, { "epoch": 0.7040650576904939, "grad_norm": 1.9618488550186157, "learning_rate": 2.1258652295537074e-06, "loss": 0.5147, "step": 16796 }, { "epoch": 0.7041069763055029, "grad_norm": 1.5483839511871338, "learning_rate": 2.1253097681847295e-06, "loss": 0.451, "step": 16797 }, { "epoch": 0.7041488949205118, "grad_norm": 1.7844085693359375, "learning_rate": 2.1247543598066923e-06, "loss": 0.4339, "step": 16798 }, { "epoch": 0.7041908135355208, "grad_norm": 2.1682567596435547, "learning_rate": 2.124199004429835e-06, "loss": 0.4961, "step": 16799 }, { "epoch": 0.7042327321505297, "grad_norm": 2.9533190727233887, "learning_rate": 2.1236437020643912e-06, "loss": 0.5479, "step": 16800 }, { "epoch": 0.7042746507655387, "grad_norm": 1.7331035137176514, "learning_rate": 2.1230884527206002e-06, "loss": 0.4912, "step": 16801 }, { "epoch": 0.7043165693805477, "grad_norm": 1.5337467193603516, "learning_rate": 2.1225332564086976e-06, "loss": 0.4819, "step": 16802 }, { "epoch": 0.7043584879955567, "grad_norm": 2.2758517265319824, "learning_rate": 2.1219781131389167e-06, "loss": 0.4983, "step": 16803 }, { "epoch": 0.7044004066105656, "grad_norm": 1.6262966394424438, "learning_rate": 2.1214230229214884e-06, "loss": 0.4935, "step": 16804 }, { "epoch": 0.7044423252255746, "grad_norm": 2.2064976692199707, "learning_rate": 2.120867985766649e-06, "loss": 0.4765, "step": 16805 }, { "epoch": 0.7044842438405835, "grad_norm": 1.8787884712219238, "learning_rate": 2.120313001684626e-06, "loss": 0.4992, "step": 16806 }, { "epoch": 0.7045261624555925, "grad_norm": 1.7143608331680298, "learning_rate": 2.1197580706856526e-06, "loss": 0.4495, "step": 16807 }, { "epoch": 0.7045680810706014, "grad_norm": 1.95984947681427, "learning_rate": 2.1192031927799582e-06, "loss": 0.4944, "step": 16808 }, { "epoch": 0.7046099996856103, "grad_norm": 1.620081901550293, "learning_rate": 2.1186483679777692e-06, "loss": 0.4462, "step": 16809 }, { "epoch": 0.7046519183006194, "grad_norm": 1.8967384099960327, "learning_rate": 2.118093596289314e-06, "loss": 0.4254, "step": 16810 }, { "epoch": 0.7046938369156283, "grad_norm": 1.8942762613296509, "learning_rate": 2.1175388777248213e-06, "loss": 0.4908, "step": 16811 }, { "epoch": 0.7047357555306373, "grad_norm": 1.9388768672943115, "learning_rate": 2.1169842122945143e-06, "loss": 0.4682, "step": 16812 }, { "epoch": 0.7047776741456462, "grad_norm": 1.8309770822525024, "learning_rate": 2.116429600008616e-06, "loss": 0.4873, "step": 16813 }, { "epoch": 0.7048195927606552, "grad_norm": 1.6898770332336426, "learning_rate": 2.1158750408773534e-06, "loss": 0.5248, "step": 16814 }, { "epoch": 0.7048615113756641, "grad_norm": 1.502691626548767, "learning_rate": 2.115320534910945e-06, "loss": 0.4199, "step": 16815 }, { "epoch": 0.7049034299906731, "grad_norm": 1.9753955602645874, "learning_rate": 2.1147660821196147e-06, "loss": 0.478, "step": 16816 }, { "epoch": 0.704945348605682, "grad_norm": 1.674033522605896, "learning_rate": 2.114211682513585e-06, "loss": 0.4568, "step": 16817 }, { "epoch": 0.704987267220691, "grad_norm": 1.497267484664917, "learning_rate": 2.1136573361030716e-06, "loss": 0.4646, "step": 16818 }, { "epoch": 0.7050291858357, "grad_norm": 1.8222777843475342, "learning_rate": 2.1131030428982953e-06, "loss": 0.5366, "step": 16819 }, { "epoch": 0.705071104450709, "grad_norm": 1.9375804662704468, "learning_rate": 2.1125488029094744e-06, "loss": 0.4843, "step": 16820 }, { "epoch": 0.7051130230657179, "grad_norm": 1.6057993173599243, "learning_rate": 2.1119946161468253e-06, "loss": 0.4621, "step": 16821 }, { "epoch": 0.7051549416807269, "grad_norm": 1.6678383350372314, "learning_rate": 2.1114404826205604e-06, "loss": 0.4655, "step": 16822 }, { "epoch": 0.7051968602957358, "grad_norm": 1.8667454719543457, "learning_rate": 2.1108864023408994e-06, "loss": 0.4836, "step": 16823 }, { "epoch": 0.7052387789107448, "grad_norm": 2.0119001865386963, "learning_rate": 2.110332375318051e-06, "loss": 0.4952, "step": 16824 }, { "epoch": 0.7052806975257537, "grad_norm": 2.2956361770629883, "learning_rate": 2.109778401562231e-06, "loss": 0.4583, "step": 16825 }, { "epoch": 0.7053226161407627, "grad_norm": 1.8688359260559082, "learning_rate": 2.109224481083652e-06, "loss": 0.4787, "step": 16826 }, { "epoch": 0.7053645347557717, "grad_norm": 2.3143908977508545, "learning_rate": 2.108670613892522e-06, "loss": 0.528, "step": 16827 }, { "epoch": 0.7054064533707807, "grad_norm": 1.682181477546692, "learning_rate": 2.1081167999990515e-06, "loss": 0.5108, "step": 16828 }, { "epoch": 0.7054483719857896, "grad_norm": 1.895921230316162, "learning_rate": 2.1075630394134523e-06, "loss": 0.5269, "step": 16829 }, { "epoch": 0.7054902906007986, "grad_norm": 1.552056908607483, "learning_rate": 2.107009332145929e-06, "loss": 0.4738, "step": 16830 }, { "epoch": 0.7055322092158075, "grad_norm": 2.3841583728790283, "learning_rate": 2.106455678206688e-06, "loss": 0.5003, "step": 16831 }, { "epoch": 0.7055741278308165, "grad_norm": 1.5723762512207031, "learning_rate": 2.105902077605938e-06, "loss": 0.4713, "step": 16832 }, { "epoch": 0.7056160464458254, "grad_norm": 1.9309560060501099, "learning_rate": 2.1053485303538807e-06, "loss": 0.4952, "step": 16833 }, { "epoch": 0.7056579650608343, "grad_norm": 1.849054217338562, "learning_rate": 2.104795036460721e-06, "loss": 0.5164, "step": 16834 }, { "epoch": 0.7056998836758434, "grad_norm": 2.0408122539520264, "learning_rate": 2.104241595936664e-06, "loss": 0.4528, "step": 16835 }, { "epoch": 0.7057418022908523, "grad_norm": 1.6201385259628296, "learning_rate": 2.1036882087919088e-06, "loss": 0.466, "step": 16836 }, { "epoch": 0.7057837209058613, "grad_norm": 1.7844692468643188, "learning_rate": 2.103134875036657e-06, "loss": 0.5014, "step": 16837 }, { "epoch": 0.7058256395208702, "grad_norm": 2.1973958015441895, "learning_rate": 2.1025815946811106e-06, "loss": 0.4903, "step": 16838 }, { "epoch": 0.7058675581358792, "grad_norm": 1.9460508823394775, "learning_rate": 2.1020283677354656e-06, "loss": 0.5201, "step": 16839 }, { "epoch": 0.7059094767508881, "grad_norm": 1.760578989982605, "learning_rate": 2.1014751942099233e-06, "loss": 0.5022, "step": 16840 }, { "epoch": 0.7059513953658971, "grad_norm": 1.7679994106292725, "learning_rate": 2.100922074114678e-06, "loss": 0.4393, "step": 16841 }, { "epoch": 0.705993313980906, "grad_norm": 2.206116199493408, "learning_rate": 2.1003690074599255e-06, "loss": 0.4939, "step": 16842 }, { "epoch": 0.706035232595915, "grad_norm": 1.6886228322982788, "learning_rate": 2.099815994255861e-06, "loss": 0.439, "step": 16843 }, { "epoch": 0.706077151210924, "grad_norm": 2.069490432739258, "learning_rate": 2.0992630345126806e-06, "loss": 0.5069, "step": 16844 }, { "epoch": 0.706119069825933, "grad_norm": 2.737734794616699, "learning_rate": 2.0987101282405754e-06, "loss": 0.4799, "step": 16845 }, { "epoch": 0.7061609884409419, "grad_norm": 1.7938092947006226, "learning_rate": 2.0981572754497374e-06, "loss": 0.4852, "step": 16846 }, { "epoch": 0.7062029070559509, "grad_norm": 1.7016260623931885, "learning_rate": 2.09760447615036e-06, "loss": 0.4613, "step": 16847 }, { "epoch": 0.7062448256709598, "grad_norm": 3.3721327781677246, "learning_rate": 2.0970517303526306e-06, "loss": 0.558, "step": 16848 }, { "epoch": 0.7062867442859688, "grad_norm": 1.9333034753799438, "learning_rate": 2.0964990380667405e-06, "loss": 0.5076, "step": 16849 }, { "epoch": 0.7063286629009777, "grad_norm": 2.0336761474609375, "learning_rate": 2.0959463993028765e-06, "loss": 0.4853, "step": 16850 }, { "epoch": 0.7063705815159868, "grad_norm": 1.9327775239944458, "learning_rate": 2.095393814071225e-06, "loss": 0.4656, "step": 16851 }, { "epoch": 0.7064125001309957, "grad_norm": 1.8007936477661133, "learning_rate": 2.094841282381972e-06, "loss": 0.4413, "step": 16852 }, { "epoch": 0.7064544187460047, "grad_norm": 1.67472243309021, "learning_rate": 2.0942888042453063e-06, "loss": 0.5609, "step": 16853 }, { "epoch": 0.7064963373610136, "grad_norm": 1.7979334592819214, "learning_rate": 2.0937363796714073e-06, "loss": 0.4831, "step": 16854 }, { "epoch": 0.7065382559760226, "grad_norm": 1.6348334550857544, "learning_rate": 2.093184008670462e-06, "loss": 0.486, "step": 16855 }, { "epoch": 0.7065801745910315, "grad_norm": 2.660518169403076, "learning_rate": 2.0926316912526496e-06, "loss": 0.4916, "step": 16856 }, { "epoch": 0.7066220932060405, "grad_norm": 2.396703004837036, "learning_rate": 2.092079427428152e-06, "loss": 0.5179, "step": 16857 }, { "epoch": 0.7066640118210494, "grad_norm": 1.83467698097229, "learning_rate": 2.0915272172071522e-06, "loss": 0.4785, "step": 16858 }, { "epoch": 0.7067059304360583, "grad_norm": 1.9532485008239746, "learning_rate": 2.0909750605998274e-06, "loss": 0.4787, "step": 16859 }, { "epoch": 0.7067478490510674, "grad_norm": 1.7202874422073364, "learning_rate": 2.090422957616354e-06, "loss": 0.479, "step": 16860 }, { "epoch": 0.7067897676660763, "grad_norm": 1.950175404548645, "learning_rate": 2.089870908266913e-06, "loss": 0.4531, "step": 16861 }, { "epoch": 0.7068316862810853, "grad_norm": 1.9420851469039917, "learning_rate": 2.0893189125616765e-06, "loss": 0.4828, "step": 16862 }, { "epoch": 0.7068736048960942, "grad_norm": 1.7533012628555298, "learning_rate": 2.088766970510822e-06, "loss": 0.4456, "step": 16863 }, { "epoch": 0.7069155235111032, "grad_norm": 1.7798794507980347, "learning_rate": 2.0882150821245257e-06, "loss": 0.4918, "step": 16864 }, { "epoch": 0.7069574421261121, "grad_norm": 1.849310278892517, "learning_rate": 2.0876632474129566e-06, "loss": 0.469, "step": 16865 }, { "epoch": 0.7069993607411211, "grad_norm": 1.8609669208526611, "learning_rate": 2.08711146638629e-06, "loss": 0.5304, "step": 16866 }, { "epoch": 0.70704127935613, "grad_norm": 2.533724784851074, "learning_rate": 2.0865597390546982e-06, "loss": 0.5276, "step": 16867 }, { "epoch": 0.707083197971139, "grad_norm": 1.838275671005249, "learning_rate": 2.08600806542835e-06, "loss": 0.499, "step": 16868 }, { "epoch": 0.707125116586148, "grad_norm": 2.156874418258667, "learning_rate": 2.0854564455174127e-06, "loss": 0.4961, "step": 16869 }, { "epoch": 0.707167035201157, "grad_norm": 1.6681411266326904, "learning_rate": 2.0849048793320583e-06, "loss": 0.4922, "step": 16870 }, { "epoch": 0.7072089538161659, "grad_norm": 1.7559510469436646, "learning_rate": 2.0843533668824507e-06, "loss": 0.4552, "step": 16871 }, { "epoch": 0.7072508724311749, "grad_norm": 1.8826098442077637, "learning_rate": 2.0838019081787583e-06, "loss": 0.5131, "step": 16872 }, { "epoch": 0.7072927910461838, "grad_norm": 1.8177071809768677, "learning_rate": 2.0832505032311474e-06, "loss": 0.4867, "step": 16873 }, { "epoch": 0.7073347096611928, "grad_norm": 1.8797143697738647, "learning_rate": 2.0826991520497796e-06, "loss": 0.5399, "step": 16874 }, { "epoch": 0.7073766282762017, "grad_norm": 2.5230040550231934, "learning_rate": 2.0821478546448204e-06, "loss": 0.4675, "step": 16875 }, { "epoch": 0.7074185468912108, "grad_norm": 2.389190435409546, "learning_rate": 2.081596611026433e-06, "loss": 0.5375, "step": 16876 }, { "epoch": 0.7074604655062197, "grad_norm": 2.003627300262451, "learning_rate": 2.0810454212047774e-06, "loss": 0.4398, "step": 16877 }, { "epoch": 0.7075023841212287, "grad_norm": 1.7812418937683105, "learning_rate": 2.080494285190013e-06, "loss": 0.4801, "step": 16878 }, { "epoch": 0.7075443027362376, "grad_norm": 2.3981168270111084, "learning_rate": 2.0799432029923012e-06, "loss": 0.4414, "step": 16879 }, { "epoch": 0.7075862213512466, "grad_norm": 1.9130293130874634, "learning_rate": 2.079392174621798e-06, "loss": 0.4799, "step": 16880 }, { "epoch": 0.7076281399662555, "grad_norm": 2.0897269248962402, "learning_rate": 2.078841200088663e-06, "loss": 0.5071, "step": 16881 }, { "epoch": 0.7076700585812645, "grad_norm": 1.8970633745193481, "learning_rate": 2.0782902794030535e-06, "loss": 0.4738, "step": 16882 }, { "epoch": 0.7077119771962734, "grad_norm": 2.1737513542175293, "learning_rate": 2.077739412575122e-06, "loss": 0.4588, "step": 16883 }, { "epoch": 0.7077538958112823, "grad_norm": 1.765180230140686, "learning_rate": 2.0771885996150247e-06, "loss": 0.5313, "step": 16884 }, { "epoch": 0.7077958144262914, "grad_norm": 1.9034357070922852, "learning_rate": 2.0766378405329165e-06, "loss": 0.4644, "step": 16885 }, { "epoch": 0.7078377330413003, "grad_norm": 1.9465950727462769, "learning_rate": 2.076087135338946e-06, "loss": 0.4913, "step": 16886 }, { "epoch": 0.7078796516563093, "grad_norm": 1.6572413444519043, "learning_rate": 2.0755364840432695e-06, "loss": 0.4454, "step": 16887 }, { "epoch": 0.7079215702713182, "grad_norm": 2.2778871059417725, "learning_rate": 2.0749858866560345e-06, "loss": 0.4313, "step": 16888 }, { "epoch": 0.7079634888863272, "grad_norm": 1.8464444875717163, "learning_rate": 2.074435343187389e-06, "loss": 0.4738, "step": 16889 }, { "epoch": 0.7080054075013361, "grad_norm": 1.8163176774978638, "learning_rate": 2.0738848536474844e-06, "loss": 0.496, "step": 16890 }, { "epoch": 0.7080473261163451, "grad_norm": 2.062622547149658, "learning_rate": 2.073334418046468e-06, "loss": 0.5231, "step": 16891 }, { "epoch": 0.708089244731354, "grad_norm": 2.0630321502685547, "learning_rate": 2.0727840363944844e-06, "loss": 0.4875, "step": 16892 }, { "epoch": 0.7081311633463631, "grad_norm": 2.246678590774536, "learning_rate": 2.07223370870168e-06, "loss": 0.494, "step": 16893 }, { "epoch": 0.708173081961372, "grad_norm": 2.063830852508545, "learning_rate": 2.0716834349782018e-06, "loss": 0.4817, "step": 16894 }, { "epoch": 0.708215000576381, "grad_norm": 2.1072676181793213, "learning_rate": 2.0711332152341896e-06, "loss": 0.4565, "step": 16895 }, { "epoch": 0.7082569191913899, "grad_norm": 2.221588134765625, "learning_rate": 2.070583049479789e-06, "loss": 0.5018, "step": 16896 }, { "epoch": 0.7082988378063989, "grad_norm": 2.3740334510803223, "learning_rate": 2.07003293772514e-06, "loss": 0.4818, "step": 16897 }, { "epoch": 0.7083407564214078, "grad_norm": 2.6569972038269043, "learning_rate": 2.069482879980381e-06, "loss": 0.4388, "step": 16898 }, { "epoch": 0.7083826750364168, "grad_norm": 2.09073543548584, "learning_rate": 2.068932876255655e-06, "loss": 0.5101, "step": 16899 }, { "epoch": 0.7084245936514257, "grad_norm": 1.9594963788986206, "learning_rate": 2.0683829265611005e-06, "loss": 0.4755, "step": 16900 }, { "epoch": 0.7084665122664348, "grad_norm": 1.9664090871810913, "learning_rate": 2.0678330309068518e-06, "loss": 0.5641, "step": 16901 }, { "epoch": 0.7085084308814437, "grad_norm": 2.0396201610565186, "learning_rate": 2.067283189303048e-06, "loss": 0.5464, "step": 16902 }, { "epoch": 0.7085503494964527, "grad_norm": 2.4791266918182373, "learning_rate": 2.066733401759826e-06, "loss": 0.4327, "step": 16903 }, { "epoch": 0.7085922681114616, "grad_norm": 2.441535234451294, "learning_rate": 2.0661836682873164e-06, "loss": 0.4934, "step": 16904 }, { "epoch": 0.7086341867264706, "grad_norm": 1.9338819980621338, "learning_rate": 2.065633988895657e-06, "loss": 0.4849, "step": 16905 }, { "epoch": 0.7086761053414795, "grad_norm": 2.0443344116210938, "learning_rate": 2.0650843635949785e-06, "loss": 0.5462, "step": 16906 }, { "epoch": 0.7087180239564885, "grad_norm": 3.563512086868286, "learning_rate": 2.0645347923954105e-06, "loss": 0.4757, "step": 16907 }, { "epoch": 0.7087599425714974, "grad_norm": 1.9404393434524536, "learning_rate": 2.063985275307086e-06, "loss": 0.5314, "step": 16908 }, { "epoch": 0.7088018611865063, "grad_norm": 2.3020505905151367, "learning_rate": 2.0634358123401353e-06, "loss": 0.5104, "step": 16909 }, { "epoch": 0.7088437798015154, "grad_norm": 1.7332624197006226, "learning_rate": 2.0628864035046835e-06, "loss": 0.4432, "step": 16910 }, { "epoch": 0.7088856984165243, "grad_norm": 2.3742027282714844, "learning_rate": 2.062337048810861e-06, "loss": 0.4593, "step": 16911 }, { "epoch": 0.7089276170315333, "grad_norm": 2.0642917156219482, "learning_rate": 2.061787748268796e-06, "loss": 0.4603, "step": 16912 }, { "epoch": 0.7089695356465422, "grad_norm": 1.9518271684646606, "learning_rate": 2.061238501888609e-06, "loss": 0.5288, "step": 16913 }, { "epoch": 0.7090114542615512, "grad_norm": 1.7017531394958496, "learning_rate": 2.0606893096804306e-06, "loss": 0.4401, "step": 16914 }, { "epoch": 0.7090533728765601, "grad_norm": 1.939149260520935, "learning_rate": 2.06014017165438e-06, "loss": 0.4594, "step": 16915 }, { "epoch": 0.7090952914915691, "grad_norm": 1.9454923868179321, "learning_rate": 2.05959108782058e-06, "loss": 0.4937, "step": 16916 }, { "epoch": 0.709137210106578, "grad_norm": 1.8759324550628662, "learning_rate": 2.0590420581891535e-06, "loss": 0.4671, "step": 16917 }, { "epoch": 0.7091791287215871, "grad_norm": 1.9422922134399414, "learning_rate": 2.0584930827702223e-06, "loss": 0.4309, "step": 16918 }, { "epoch": 0.709221047336596, "grad_norm": 1.8772944211959839, "learning_rate": 2.0579441615739027e-06, "loss": 0.4719, "step": 16919 }, { "epoch": 0.709262965951605, "grad_norm": 1.7295507192611694, "learning_rate": 2.0573952946103153e-06, "loss": 0.5368, "step": 16920 }, { "epoch": 0.7093048845666139, "grad_norm": 2.0858805179595947, "learning_rate": 2.0568464818895793e-06, "loss": 0.4525, "step": 16921 }, { "epoch": 0.7093468031816229, "grad_norm": 2.486081123352051, "learning_rate": 2.0562977234218084e-06, "loss": 0.4485, "step": 16922 }, { "epoch": 0.7093887217966318, "grad_norm": 1.6924092769622803, "learning_rate": 2.055749019217121e-06, "loss": 0.4198, "step": 16923 }, { "epoch": 0.7094306404116408, "grad_norm": 1.999572515487671, "learning_rate": 2.05520036928563e-06, "loss": 0.5006, "step": 16924 }, { "epoch": 0.7094725590266497, "grad_norm": 2.087491273880005, "learning_rate": 2.054651773637447e-06, "loss": 0.4597, "step": 16925 }, { "epoch": 0.7095144776416588, "grad_norm": 2.5916428565979004, "learning_rate": 2.054103232282687e-06, "loss": 0.4265, "step": 16926 }, { "epoch": 0.7095563962566677, "grad_norm": 1.7217445373535156, "learning_rate": 2.053554745231464e-06, "loss": 0.4617, "step": 16927 }, { "epoch": 0.7095983148716767, "grad_norm": 5.756962299346924, "learning_rate": 2.0530063124938826e-06, "loss": 0.5123, "step": 16928 }, { "epoch": 0.7096402334866856, "grad_norm": 2.1064929962158203, "learning_rate": 2.052457934080058e-06, "loss": 0.4628, "step": 16929 }, { "epoch": 0.7096821521016946, "grad_norm": 1.9181017875671387, "learning_rate": 2.051909610000095e-06, "loss": 0.422, "step": 16930 }, { "epoch": 0.7097240707167035, "grad_norm": 2.2417519092559814, "learning_rate": 2.051361340264103e-06, "loss": 0.4648, "step": 16931 }, { "epoch": 0.7097659893317125, "grad_norm": 2.757875919342041, "learning_rate": 2.05081312488219e-06, "loss": 0.4917, "step": 16932 }, { "epoch": 0.7098079079467214, "grad_norm": 2.150007486343384, "learning_rate": 2.050264963864458e-06, "loss": 0.4598, "step": 16933 }, { "epoch": 0.7098498265617303, "grad_norm": 2.2296535968780518, "learning_rate": 2.049716857221015e-06, "loss": 0.5177, "step": 16934 }, { "epoch": 0.7098917451767394, "grad_norm": 2.1532998085021973, "learning_rate": 2.0491688049619636e-06, "loss": 0.5052, "step": 16935 }, { "epoch": 0.7099336637917483, "grad_norm": 1.528725504875183, "learning_rate": 2.0486208070974032e-06, "loss": 0.4802, "step": 16936 }, { "epoch": 0.7099755824067573, "grad_norm": 1.7636774778366089, "learning_rate": 2.048072863637439e-06, "loss": 0.5024, "step": 16937 }, { "epoch": 0.7100175010217662, "grad_norm": 1.912631630897522, "learning_rate": 2.0475249745921722e-06, "loss": 0.5033, "step": 16938 }, { "epoch": 0.7100594196367752, "grad_norm": 1.4090858697891235, "learning_rate": 2.046977139971699e-06, "loss": 0.4745, "step": 16939 }, { "epoch": 0.7101013382517841, "grad_norm": 1.8915541172027588, "learning_rate": 2.0464293597861197e-06, "loss": 0.4729, "step": 16940 }, { "epoch": 0.7101432568667931, "grad_norm": 2.4266750812530518, "learning_rate": 2.045881634045534e-06, "loss": 0.4957, "step": 16941 }, { "epoch": 0.710185175481802, "grad_norm": 3.133906602859497, "learning_rate": 2.0453339627600345e-06, "loss": 0.4543, "step": 16942 }, { "epoch": 0.7102270940968111, "grad_norm": 1.70814847946167, "learning_rate": 2.0447863459397206e-06, "loss": 0.4098, "step": 16943 }, { "epoch": 0.71026901271182, "grad_norm": 2.1239819526672363, "learning_rate": 2.044238783594684e-06, "loss": 0.4679, "step": 16944 }, { "epoch": 0.710310931326829, "grad_norm": 1.8774288892745972, "learning_rate": 2.0436912757350186e-06, "loss": 0.4786, "step": 16945 }, { "epoch": 0.7103528499418379, "grad_norm": 1.9134825468063354, "learning_rate": 2.0431438223708174e-06, "loss": 0.4646, "step": 16946 }, { "epoch": 0.7103947685568469, "grad_norm": 1.591239333152771, "learning_rate": 2.0425964235121733e-06, "loss": 0.4783, "step": 16947 }, { "epoch": 0.7104366871718558, "grad_norm": 1.882190227508545, "learning_rate": 2.0420490791691744e-06, "loss": 0.51, "step": 16948 }, { "epoch": 0.7104786057868648, "grad_norm": 2.16538667678833, "learning_rate": 2.041501789351911e-06, "loss": 0.4749, "step": 16949 }, { "epoch": 0.7105205244018737, "grad_norm": 1.6387407779693604, "learning_rate": 2.0409545540704745e-06, "loss": 0.4782, "step": 16950 }, { "epoch": 0.7105624430168828, "grad_norm": 2.851632595062256, "learning_rate": 2.0404073733349477e-06, "loss": 0.4667, "step": 16951 }, { "epoch": 0.7106043616318917, "grad_norm": 1.8413540124893188, "learning_rate": 2.0398602471554213e-06, "loss": 0.4149, "step": 16952 }, { "epoch": 0.7106462802469007, "grad_norm": 1.6318528652191162, "learning_rate": 2.0393131755419793e-06, "loss": 0.4598, "step": 16953 }, { "epoch": 0.7106881988619096, "grad_norm": 2.3871021270751953, "learning_rate": 2.0387661585047043e-06, "loss": 0.4893, "step": 16954 }, { "epoch": 0.7107301174769186, "grad_norm": 1.6913717985153198, "learning_rate": 2.0382191960536807e-06, "loss": 0.5178, "step": 16955 }, { "epoch": 0.7107720360919275, "grad_norm": 1.9684348106384277, "learning_rate": 2.0376722881989937e-06, "loss": 0.4722, "step": 16956 }, { "epoch": 0.7108139547069365, "grad_norm": 1.733089804649353, "learning_rate": 2.0371254349507213e-06, "loss": 0.4929, "step": 16957 }, { "epoch": 0.7108558733219454, "grad_norm": 1.7318804264068604, "learning_rate": 2.036578636318945e-06, "loss": 0.5285, "step": 16958 }, { "epoch": 0.7108977919369543, "grad_norm": 1.730720043182373, "learning_rate": 2.0360318923137467e-06, "loss": 0.4336, "step": 16959 }, { "epoch": 0.7109397105519634, "grad_norm": 1.846696138381958, "learning_rate": 2.0354852029452017e-06, "loss": 0.5097, "step": 16960 }, { "epoch": 0.7109816291669723, "grad_norm": 2.192516565322876, "learning_rate": 2.0349385682233896e-06, "loss": 0.5214, "step": 16961 }, { "epoch": 0.7110235477819813, "grad_norm": 2.0444889068603516, "learning_rate": 2.0343919881583864e-06, "loss": 0.5272, "step": 16962 }, { "epoch": 0.7110654663969902, "grad_norm": 1.90697181224823, "learning_rate": 2.033845462760265e-06, "loss": 0.4829, "step": 16963 }, { "epoch": 0.7111073850119992, "grad_norm": 1.7905045747756958, "learning_rate": 2.0332989920391022e-06, "loss": 0.5061, "step": 16964 }, { "epoch": 0.7111493036270081, "grad_norm": 2.3592989444732666, "learning_rate": 2.0327525760049727e-06, "loss": 0.4944, "step": 16965 }, { "epoch": 0.7111912222420171, "grad_norm": 2.112992286682129, "learning_rate": 2.0322062146679457e-06, "loss": 0.5694, "step": 16966 }, { "epoch": 0.711233140857026, "grad_norm": 2.0934407711029053, "learning_rate": 2.031659908038094e-06, "loss": 0.5223, "step": 16967 }, { "epoch": 0.7112750594720351, "grad_norm": 2.669294595718384, "learning_rate": 2.0311136561254904e-06, "loss": 0.5066, "step": 16968 }, { "epoch": 0.711316978087044, "grad_norm": 1.789565920829773, "learning_rate": 2.0305674589402002e-06, "loss": 0.4898, "step": 16969 }, { "epoch": 0.711358896702053, "grad_norm": 1.8384448289871216, "learning_rate": 2.0300213164922954e-06, "loss": 0.4925, "step": 16970 }, { "epoch": 0.7114008153170619, "grad_norm": 1.939509630203247, "learning_rate": 2.029475228791841e-06, "loss": 0.4985, "step": 16971 }, { "epoch": 0.7114427339320709, "grad_norm": 1.8522833585739136, "learning_rate": 2.0289291958489043e-06, "loss": 0.5057, "step": 16972 }, { "epoch": 0.7114846525470798, "grad_norm": 2.1808254718780518, "learning_rate": 2.0283832176735495e-06, "loss": 0.49, "step": 16973 }, { "epoch": 0.7115265711620888, "grad_norm": 2.3821258544921875, "learning_rate": 2.027837294275844e-06, "loss": 0.4816, "step": 16974 }, { "epoch": 0.7115684897770977, "grad_norm": 2.129272222518921, "learning_rate": 2.0272914256658473e-06, "loss": 0.4879, "step": 16975 }, { "epoch": 0.7116104083921068, "grad_norm": 1.8041473627090454, "learning_rate": 2.0267456118536233e-06, "loss": 0.4929, "step": 16976 }, { "epoch": 0.7116523270071157, "grad_norm": 1.775555968284607, "learning_rate": 2.0261998528492353e-06, "loss": 0.4662, "step": 16977 }, { "epoch": 0.7116942456221247, "grad_norm": 2.094040632247925, "learning_rate": 2.0256541486627402e-06, "loss": 0.4513, "step": 16978 }, { "epoch": 0.7117361642371336, "grad_norm": 1.9102039337158203, "learning_rate": 2.0251084993042004e-06, "loss": 0.458, "step": 16979 }, { "epoch": 0.7117780828521426, "grad_norm": 2.108987331390381, "learning_rate": 2.0245629047836712e-06, "loss": 0.4682, "step": 16980 }, { "epoch": 0.7118200014671515, "grad_norm": 2.0184273719787598, "learning_rate": 2.024017365111213e-06, "loss": 0.4949, "step": 16981 }, { "epoch": 0.7118619200821605, "grad_norm": 2.7773537635803223, "learning_rate": 2.0234718802968794e-06, "loss": 0.4703, "step": 16982 }, { "epoch": 0.7119038386971694, "grad_norm": 2.6122522354125977, "learning_rate": 2.0229264503507277e-06, "loss": 0.5201, "step": 16983 }, { "epoch": 0.7119457573121784, "grad_norm": 3.931224822998047, "learning_rate": 2.02238107528281e-06, "loss": 0.476, "step": 16984 }, { "epoch": 0.7119876759271874, "grad_norm": 1.992903709411621, "learning_rate": 2.021835755103181e-06, "loss": 0.4745, "step": 16985 }, { "epoch": 0.7120295945421963, "grad_norm": 2.0802090167999268, "learning_rate": 2.021290489821894e-06, "loss": 0.5255, "step": 16986 }, { "epoch": 0.7120715131572053, "grad_norm": 1.9868499040603638, "learning_rate": 2.020745279448997e-06, "loss": 0.5126, "step": 16987 }, { "epoch": 0.7121134317722142, "grad_norm": 1.8246581554412842, "learning_rate": 2.020200123994542e-06, "loss": 0.4434, "step": 16988 }, { "epoch": 0.7121553503872232, "grad_norm": 3.430440902709961, "learning_rate": 2.0196550234685802e-06, "loss": 0.4829, "step": 16989 }, { "epoch": 0.7121972690022321, "grad_norm": 5.474377632141113, "learning_rate": 2.0191099778811577e-06, "loss": 0.5099, "step": 16990 }, { "epoch": 0.7122391876172411, "grad_norm": 2.2944324016571045, "learning_rate": 2.0185649872423202e-06, "loss": 0.5132, "step": 16991 }, { "epoch": 0.71228110623225, "grad_norm": 2.8534064292907715, "learning_rate": 2.0180200515621173e-06, "loss": 0.5096, "step": 16992 }, { "epoch": 0.7123230248472591, "grad_norm": 1.5999746322631836, "learning_rate": 2.0174751708505906e-06, "loss": 0.508, "step": 16993 }, { "epoch": 0.712364943462268, "grad_norm": 1.7030823230743408, "learning_rate": 2.016930345117786e-06, "loss": 0.4721, "step": 16994 }, { "epoch": 0.712406862077277, "grad_norm": 2.127436637878418, "learning_rate": 2.0163855743737476e-06, "loss": 0.5301, "step": 16995 }, { "epoch": 0.7124487806922859, "grad_norm": 2.220693349838257, "learning_rate": 2.015840858628515e-06, "loss": 0.4932, "step": 16996 }, { "epoch": 0.7124906993072949, "grad_norm": 2.3110151290893555, "learning_rate": 2.0152961978921308e-06, "loss": 0.4427, "step": 16997 }, { "epoch": 0.7125326179223038, "grad_norm": 2.212726354598999, "learning_rate": 2.014751592174637e-06, "loss": 0.4337, "step": 16998 }, { "epoch": 0.7125745365373128, "grad_norm": 2.1934964656829834, "learning_rate": 2.0142070414860704e-06, "loss": 0.4654, "step": 16999 }, { "epoch": 0.7126164551523217, "grad_norm": 2.7501437664031982, "learning_rate": 2.0136625458364673e-06, "loss": 0.4728, "step": 17000 }, { "epoch": 0.7126583737673308, "grad_norm": 2.0760498046875, "learning_rate": 2.0131181052358684e-06, "loss": 0.5098, "step": 17001 }, { "epoch": 0.7127002923823397, "grad_norm": 2.4012293815612793, "learning_rate": 2.012573719694306e-06, "loss": 0.5249, "step": 17002 }, { "epoch": 0.7127422109973487, "grad_norm": 1.9491770267486572, "learning_rate": 2.0120293892218178e-06, "loss": 0.5397, "step": 17003 }, { "epoch": 0.7127841296123576, "grad_norm": 1.8491970300674438, "learning_rate": 2.0114851138284387e-06, "loss": 0.4457, "step": 17004 }, { "epoch": 0.7128260482273666, "grad_norm": 1.7296767234802246, "learning_rate": 2.010940893524198e-06, "loss": 0.442, "step": 17005 }, { "epoch": 0.7128679668423755, "grad_norm": 2.1124329566955566, "learning_rate": 2.0103967283191318e-06, "loss": 0.4866, "step": 17006 }, { "epoch": 0.7129098854573845, "grad_norm": 1.8907206058502197, "learning_rate": 2.0098526182232663e-06, "loss": 0.4416, "step": 17007 }, { "epoch": 0.7129518040723934, "grad_norm": 1.6575478315353394, "learning_rate": 2.0093085632466363e-06, "loss": 0.4983, "step": 17008 }, { "epoch": 0.7129937226874024, "grad_norm": 2.172804594039917, "learning_rate": 2.0087645633992685e-06, "loss": 0.5098, "step": 17009 }, { "epoch": 0.7130356413024114, "grad_norm": 1.731003999710083, "learning_rate": 2.008220618691189e-06, "loss": 0.4705, "step": 17010 }, { "epoch": 0.7130775599174203, "grad_norm": 1.9015462398529053, "learning_rate": 2.007676729132426e-06, "loss": 0.5307, "step": 17011 }, { "epoch": 0.7131194785324293, "grad_norm": 1.870053768157959, "learning_rate": 2.0071328947330082e-06, "loss": 0.5058, "step": 17012 }, { "epoch": 0.7131613971474382, "grad_norm": 1.7320476770401, "learning_rate": 2.006589115502956e-06, "loss": 0.4766, "step": 17013 }, { "epoch": 0.7132033157624472, "grad_norm": 1.8941534757614136, "learning_rate": 2.006045391452295e-06, "loss": 0.4689, "step": 17014 }, { "epoch": 0.7132452343774561, "grad_norm": 1.9780120849609375, "learning_rate": 2.0055017225910496e-06, "loss": 0.4828, "step": 17015 }, { "epoch": 0.7132871529924651, "grad_norm": 1.9677789211273193, "learning_rate": 2.0049581089292387e-06, "loss": 0.5008, "step": 17016 }, { "epoch": 0.713329071607474, "grad_norm": 2.653608798980713, "learning_rate": 2.004414550476886e-06, "loss": 0.4237, "step": 17017 }, { "epoch": 0.7133709902224831, "grad_norm": 2.073333978652954, "learning_rate": 2.0038710472440087e-06, "loss": 0.5211, "step": 17018 }, { "epoch": 0.713412908837492, "grad_norm": 1.7859410047531128, "learning_rate": 2.0033275992406283e-06, "loss": 0.5076, "step": 17019 }, { "epoch": 0.713454827452501, "grad_norm": 2.074104070663452, "learning_rate": 2.002784206476759e-06, "loss": 0.4511, "step": 17020 }, { "epoch": 0.7134967460675099, "grad_norm": 2.8313632011413574, "learning_rate": 2.0022408689624205e-06, "loss": 0.4878, "step": 17021 }, { "epoch": 0.7135386646825189, "grad_norm": 1.8989789485931396, "learning_rate": 2.001697586707626e-06, "loss": 0.5008, "step": 17022 }, { "epoch": 0.7135805832975278, "grad_norm": 2.0022056102752686, "learning_rate": 2.0011543597223913e-06, "loss": 0.4769, "step": 17023 }, { "epoch": 0.7136225019125368, "grad_norm": 1.7673842906951904, "learning_rate": 2.000611188016732e-06, "loss": 0.4751, "step": 17024 }, { "epoch": 0.7136644205275458, "grad_norm": 1.6933263540267944, "learning_rate": 2.0000680716006578e-06, "loss": 0.4518, "step": 17025 }, { "epoch": 0.7137063391425548, "grad_norm": 2.4072229862213135, "learning_rate": 1.9995250104841805e-06, "loss": 0.4751, "step": 17026 }, { "epoch": 0.7137482577575637, "grad_norm": 3.7752907276153564, "learning_rate": 1.998982004677314e-06, "loss": 0.4696, "step": 17027 }, { "epoch": 0.7137901763725727, "grad_norm": 1.9013537168502808, "learning_rate": 1.998439054190065e-06, "loss": 0.5401, "step": 17028 }, { "epoch": 0.7138320949875816, "grad_norm": 2.1558401584625244, "learning_rate": 1.9978961590324406e-06, "loss": 0.501, "step": 17029 }, { "epoch": 0.7138740136025906, "grad_norm": 1.8662269115447998, "learning_rate": 1.997353319214453e-06, "loss": 0.4248, "step": 17030 }, { "epoch": 0.7139159322175995, "grad_norm": 1.8690274953842163, "learning_rate": 1.996810534746103e-06, "loss": 0.4864, "step": 17031 }, { "epoch": 0.7139578508326085, "grad_norm": 1.8115384578704834, "learning_rate": 1.9962678056373998e-06, "loss": 0.493, "step": 17032 }, { "epoch": 0.7139997694476174, "grad_norm": 1.478309988975525, "learning_rate": 1.9957251318983483e-06, "loss": 0.4705, "step": 17033 }, { "epoch": 0.7140416880626264, "grad_norm": 1.6859060525894165, "learning_rate": 1.9951825135389492e-06, "loss": 0.4127, "step": 17034 }, { "epoch": 0.7140836066776354, "grad_norm": 3.9623451232910156, "learning_rate": 1.9946399505692064e-06, "loss": 0.453, "step": 17035 }, { "epoch": 0.7141255252926443, "grad_norm": 2.248319149017334, "learning_rate": 1.994097442999123e-06, "loss": 0.4478, "step": 17036 }, { "epoch": 0.7141674439076533, "grad_norm": 2.7071802616119385, "learning_rate": 1.993554990838698e-06, "loss": 0.4903, "step": 17037 }, { "epoch": 0.7142093625226622, "grad_norm": 1.9525810480117798, "learning_rate": 1.9930125940979284e-06, "loss": 0.5146, "step": 17038 }, { "epoch": 0.7142512811376712, "grad_norm": 2.1241657733917236, "learning_rate": 1.9924702527868166e-06, "loss": 0.4707, "step": 17039 }, { "epoch": 0.7142931997526801, "grad_norm": 1.9109625816345215, "learning_rate": 1.991927966915356e-06, "loss": 0.5194, "step": 17040 }, { "epoch": 0.7143351183676891, "grad_norm": 2.039299249649048, "learning_rate": 1.9913857364935445e-06, "loss": 0.4954, "step": 17041 }, { "epoch": 0.714377036982698, "grad_norm": 1.7828930616378784, "learning_rate": 1.9908435615313798e-06, "loss": 0.4609, "step": 17042 }, { "epoch": 0.7144189555977071, "grad_norm": 1.9757180213928223, "learning_rate": 1.9903014420388523e-06, "loss": 0.4892, "step": 17043 }, { "epoch": 0.714460874212716, "grad_norm": 2.445976495742798, "learning_rate": 1.9897593780259567e-06, "loss": 0.5019, "step": 17044 }, { "epoch": 0.714502792827725, "grad_norm": 1.8648066520690918, "learning_rate": 1.989217369502687e-06, "loss": 0.4865, "step": 17045 }, { "epoch": 0.7145447114427339, "grad_norm": 2.191756010055542, "learning_rate": 1.9886754164790334e-06, "loss": 0.493, "step": 17046 }, { "epoch": 0.7145866300577429, "grad_norm": 1.8560502529144287, "learning_rate": 1.988133518964983e-06, "loss": 0.4946, "step": 17047 }, { "epoch": 0.7146285486727518, "grad_norm": 4.572960376739502, "learning_rate": 1.98759167697053e-06, "loss": 0.4965, "step": 17048 }, { "epoch": 0.7146704672877608, "grad_norm": 4.61828088760376, "learning_rate": 1.987049890505657e-06, "loss": 0.5018, "step": 17049 }, { "epoch": 0.7147123859027698, "grad_norm": 3.763618230819702, "learning_rate": 1.9865081595803552e-06, "loss": 0.4512, "step": 17050 }, { "epoch": 0.7147543045177788, "grad_norm": 4.2397918701171875, "learning_rate": 1.9859664842046106e-06, "loss": 0.5481, "step": 17051 }, { "epoch": 0.7147962231327877, "grad_norm": 2.9282021522521973, "learning_rate": 1.9854248643884053e-06, "loss": 0.406, "step": 17052 }, { "epoch": 0.7148381417477967, "grad_norm": 1.731276035308838, "learning_rate": 1.984883300141726e-06, "loss": 0.4669, "step": 17053 }, { "epoch": 0.7148800603628056, "grad_norm": 1.752480149269104, "learning_rate": 1.984341791474555e-06, "loss": 0.4819, "step": 17054 }, { "epoch": 0.7149219789778146, "grad_norm": 2.1207973957061768, "learning_rate": 1.9838003383968753e-06, "loss": 0.5321, "step": 17055 }, { "epoch": 0.7149638975928235, "grad_norm": 2.0764997005462646, "learning_rate": 1.9832589409186643e-06, "loss": 0.4567, "step": 17056 }, { "epoch": 0.7150058162078325, "grad_norm": 1.840937852859497, "learning_rate": 1.982717599049906e-06, "loss": 0.4749, "step": 17057 }, { "epoch": 0.7150477348228415, "grad_norm": 2.0202767848968506, "learning_rate": 1.9821763128005756e-06, "loss": 0.5147, "step": 17058 }, { "epoch": 0.7150896534378504, "grad_norm": 1.7967708110809326, "learning_rate": 1.981635082180653e-06, "loss": 0.4663, "step": 17059 }, { "epoch": 0.7151315720528594, "grad_norm": 2.103318214416504, "learning_rate": 1.9810939072001166e-06, "loss": 0.4469, "step": 17060 }, { "epoch": 0.7151734906678683, "grad_norm": 1.9607298374176025, "learning_rate": 1.9805527878689386e-06, "loss": 0.4631, "step": 17061 }, { "epoch": 0.7152154092828773, "grad_norm": 1.5480518341064453, "learning_rate": 1.9800117241970964e-06, "loss": 0.417, "step": 17062 }, { "epoch": 0.7152573278978862, "grad_norm": 2.1536927223205566, "learning_rate": 1.9794707161945642e-06, "loss": 0.4641, "step": 17063 }, { "epoch": 0.7152992465128952, "grad_norm": 1.9512158632278442, "learning_rate": 1.9789297638713135e-06, "loss": 0.4963, "step": 17064 }, { "epoch": 0.7153411651279041, "grad_norm": 2.1620728969573975, "learning_rate": 1.978388867237315e-06, "loss": 0.5746, "step": 17065 }, { "epoch": 0.7153830837429132, "grad_norm": 2.3518667221069336, "learning_rate": 1.9778480263025414e-06, "loss": 0.5086, "step": 17066 }, { "epoch": 0.7154250023579221, "grad_norm": 1.7876405715942383, "learning_rate": 1.97730724107696e-06, "loss": 0.4882, "step": 17067 }, { "epoch": 0.7154669209729311, "grad_norm": 3.7470617294311523, "learning_rate": 1.976766511570541e-06, "loss": 0.4886, "step": 17068 }, { "epoch": 0.71550883958794, "grad_norm": 1.968501091003418, "learning_rate": 1.976225837793254e-06, "loss": 0.4999, "step": 17069 }, { "epoch": 0.715550758202949, "grad_norm": 1.7716968059539795, "learning_rate": 1.9756852197550607e-06, "loss": 0.4999, "step": 17070 }, { "epoch": 0.7155926768179579, "grad_norm": 1.8156259059906006, "learning_rate": 1.97514465746593e-06, "loss": 0.5127, "step": 17071 }, { "epoch": 0.7156345954329669, "grad_norm": 2.807562828063965, "learning_rate": 1.974604150935827e-06, "loss": 0.471, "step": 17072 }, { "epoch": 0.7156765140479758, "grad_norm": 1.9077885150909424, "learning_rate": 1.974063700174713e-06, "loss": 0.5282, "step": 17073 }, { "epoch": 0.7157184326629848, "grad_norm": 2.2871310710906982, "learning_rate": 1.973523305192553e-06, "loss": 0.4596, "step": 17074 }, { "epoch": 0.7157603512779938, "grad_norm": 2.2051360607147217, "learning_rate": 1.9729829659993065e-06, "loss": 0.4628, "step": 17075 }, { "epoch": 0.7158022698930028, "grad_norm": 2.0789449214935303, "learning_rate": 1.972442682604933e-06, "loss": 0.4949, "step": 17076 }, { "epoch": 0.7158441885080117, "grad_norm": 2.280284881591797, "learning_rate": 1.971902455019393e-06, "loss": 0.4036, "step": 17077 }, { "epoch": 0.7158861071230207, "grad_norm": 2.128973960876465, "learning_rate": 1.971362283252647e-06, "loss": 0.4491, "step": 17078 }, { "epoch": 0.7159280257380296, "grad_norm": 2.4210216999053955, "learning_rate": 1.970822167314648e-06, "loss": 0.5447, "step": 17079 }, { "epoch": 0.7159699443530386, "grad_norm": 1.976449966430664, "learning_rate": 1.9702821072153573e-06, "loss": 0.5235, "step": 17080 }, { "epoch": 0.7160118629680475, "grad_norm": 2.04990553855896, "learning_rate": 1.9697421029647252e-06, "loss": 0.4607, "step": 17081 }, { "epoch": 0.7160537815830565, "grad_norm": 2.043470859527588, "learning_rate": 1.969202154572708e-06, "loss": 0.4908, "step": 17082 }, { "epoch": 0.7160957001980655, "grad_norm": 2.7415835857391357, "learning_rate": 1.9686622620492613e-06, "loss": 0.5343, "step": 17083 }, { "epoch": 0.7161376188130744, "grad_norm": 1.8987517356872559, "learning_rate": 1.9681224254043348e-06, "loss": 0.4634, "step": 17084 }, { "epoch": 0.7161795374280834, "grad_norm": 3.5113308429718018, "learning_rate": 1.9675826446478787e-06, "loss": 0.506, "step": 17085 }, { "epoch": 0.7162214560430923, "grad_norm": 1.791461706161499, "learning_rate": 1.9670429197898454e-06, "loss": 0.5026, "step": 17086 }, { "epoch": 0.7162633746581013, "grad_norm": 1.8959052562713623, "learning_rate": 1.9665032508401816e-06, "loss": 0.5192, "step": 17087 }, { "epoch": 0.7163052932731102, "grad_norm": 2.117640495300293, "learning_rate": 1.9659636378088364e-06, "loss": 0.527, "step": 17088 }, { "epoch": 0.7163472118881192, "grad_norm": 1.804038405418396, "learning_rate": 1.965424080705759e-06, "loss": 0.4891, "step": 17089 }, { "epoch": 0.7163891305031281, "grad_norm": 1.93666672706604, "learning_rate": 1.964884579540892e-06, "loss": 0.4844, "step": 17090 }, { "epoch": 0.7164310491181372, "grad_norm": 1.7368757724761963, "learning_rate": 1.9643451343241815e-06, "loss": 0.4533, "step": 17091 }, { "epoch": 0.7164729677331461, "grad_norm": 3.1256251335144043, "learning_rate": 1.963805745065574e-06, "loss": 0.5281, "step": 17092 }, { "epoch": 0.7165148863481551, "grad_norm": 1.9158055782318115, "learning_rate": 1.9632664117750095e-06, "loss": 0.4853, "step": 17093 }, { "epoch": 0.716556804963164, "grad_norm": 1.9296215772628784, "learning_rate": 1.9627271344624294e-06, "loss": 0.4915, "step": 17094 }, { "epoch": 0.716598723578173, "grad_norm": 1.8560739755630493, "learning_rate": 1.962187913137777e-06, "loss": 0.4375, "step": 17095 }, { "epoch": 0.7166406421931819, "grad_norm": 1.9319144487380981, "learning_rate": 1.961648747810989e-06, "loss": 0.4673, "step": 17096 }, { "epoch": 0.7166825608081909, "grad_norm": 2.197874069213867, "learning_rate": 1.9611096384920065e-06, "loss": 0.4124, "step": 17097 }, { "epoch": 0.7167244794231998, "grad_norm": 2.7908496856689453, "learning_rate": 1.960570585190768e-06, "loss": 0.5514, "step": 17098 }, { "epoch": 0.7167663980382089, "grad_norm": 3.430623769760132, "learning_rate": 1.9600315879172073e-06, "loss": 0.483, "step": 17099 }, { "epoch": 0.7168083166532178, "grad_norm": 2.0434536933898926, "learning_rate": 1.9594926466812612e-06, "loss": 0.4584, "step": 17100 }, { "epoch": 0.7168502352682268, "grad_norm": 1.932708740234375, "learning_rate": 1.9589537614928674e-06, "loss": 0.4956, "step": 17101 }, { "epoch": 0.7168921538832357, "grad_norm": 3.4857680797576904, "learning_rate": 1.958414932361956e-06, "loss": 0.4481, "step": 17102 }, { "epoch": 0.7169340724982447, "grad_norm": 2.3144922256469727, "learning_rate": 1.957876159298459e-06, "loss": 0.4901, "step": 17103 }, { "epoch": 0.7169759911132536, "grad_norm": 1.7777695655822754, "learning_rate": 1.9573374423123116e-06, "loss": 0.4571, "step": 17104 }, { "epoch": 0.7170179097282626, "grad_norm": 2.8098866939544678, "learning_rate": 1.95679878141344e-06, "loss": 0.466, "step": 17105 }, { "epoch": 0.7170598283432715, "grad_norm": 2.5267207622528076, "learning_rate": 1.9562601766117757e-06, "loss": 0.4433, "step": 17106 }, { "epoch": 0.7171017469582806, "grad_norm": 2.124739408493042, "learning_rate": 1.9557216279172486e-06, "loss": 0.4878, "step": 17107 }, { "epoch": 0.7171436655732895, "grad_norm": 1.7650947570800781, "learning_rate": 1.9551831353397833e-06, "loss": 0.445, "step": 17108 }, { "epoch": 0.7171855841882984, "grad_norm": 1.9412908554077148, "learning_rate": 1.9546446988893075e-06, "loss": 0.5164, "step": 17109 }, { "epoch": 0.7172275028033074, "grad_norm": 1.950510025024414, "learning_rate": 1.9541063185757485e-06, "loss": 0.5122, "step": 17110 }, { "epoch": 0.7172694214183163, "grad_norm": 2.143314838409424, "learning_rate": 1.9535679944090284e-06, "loss": 0.4399, "step": 17111 }, { "epoch": 0.7173113400333253, "grad_norm": 2.110752582550049, "learning_rate": 1.9530297263990692e-06, "loss": 0.454, "step": 17112 }, { "epoch": 0.7173532586483342, "grad_norm": 1.7698533535003662, "learning_rate": 1.9524915145557966e-06, "loss": 0.4604, "step": 17113 }, { "epoch": 0.7173951772633432, "grad_norm": 2.094590663909912, "learning_rate": 1.951953358889128e-06, "loss": 0.4914, "step": 17114 }, { "epoch": 0.7174370958783521, "grad_norm": 1.7390273809432983, "learning_rate": 1.9514152594089857e-06, "loss": 0.4372, "step": 17115 }, { "epoch": 0.7174790144933612, "grad_norm": 1.850557804107666, "learning_rate": 1.95087721612529e-06, "loss": 0.5079, "step": 17116 }, { "epoch": 0.7175209331083701, "grad_norm": 2.5862929821014404, "learning_rate": 1.9503392290479557e-06, "loss": 0.4469, "step": 17117 }, { "epoch": 0.7175628517233791, "grad_norm": 1.9327514171600342, "learning_rate": 1.949801298186902e-06, "loss": 0.4934, "step": 17118 }, { "epoch": 0.717604770338388, "grad_norm": 1.799569845199585, "learning_rate": 1.9492634235520458e-06, "loss": 0.4634, "step": 17119 }, { "epoch": 0.717646688953397, "grad_norm": 2.3282277584075928, "learning_rate": 1.9487256051533e-06, "loss": 0.5109, "step": 17120 }, { "epoch": 0.7176886075684059, "grad_norm": 2.212116241455078, "learning_rate": 1.94818784300058e-06, "loss": 0.4554, "step": 17121 }, { "epoch": 0.7177305261834149, "grad_norm": 1.7385575771331787, "learning_rate": 1.947650137103799e-06, "loss": 0.5317, "step": 17122 }, { "epoch": 0.7177724447984238, "grad_norm": 1.9189128875732422, "learning_rate": 1.947112487472866e-06, "loss": 0.4582, "step": 17123 }, { "epoch": 0.7178143634134329, "grad_norm": 1.7227814197540283, "learning_rate": 1.946574894117694e-06, "loss": 0.5197, "step": 17124 }, { "epoch": 0.7178562820284418, "grad_norm": 2.3999671936035156, "learning_rate": 1.946037357048194e-06, "loss": 0.5022, "step": 17125 }, { "epoch": 0.7178982006434508, "grad_norm": 2.8339967727661133, "learning_rate": 1.9454998762742716e-06, "loss": 0.4693, "step": 17126 }, { "epoch": 0.7179401192584597, "grad_norm": 2.669086456298828, "learning_rate": 1.9449624518058365e-06, "loss": 0.5144, "step": 17127 }, { "epoch": 0.7179820378734687, "grad_norm": 1.6174930334091187, "learning_rate": 1.944425083652797e-06, "loss": 0.454, "step": 17128 }, { "epoch": 0.7180239564884776, "grad_norm": 2.0860955715179443, "learning_rate": 1.9438877718250545e-06, "loss": 0.5486, "step": 17129 }, { "epoch": 0.7180658751034866, "grad_norm": 2.3131251335144043, "learning_rate": 1.943350516332518e-06, "loss": 0.4943, "step": 17130 }, { "epoch": 0.7181077937184955, "grad_norm": 1.7184381484985352, "learning_rate": 1.942813317185089e-06, "loss": 0.5147, "step": 17131 }, { "epoch": 0.7181497123335046, "grad_norm": 2.3970766067504883, "learning_rate": 1.942276174392669e-06, "loss": 0.5092, "step": 17132 }, { "epoch": 0.7181916309485135, "grad_norm": 3.198755979537964, "learning_rate": 1.9417390879651597e-06, "loss": 0.4688, "step": 17133 }, { "epoch": 0.7182335495635224, "grad_norm": 1.7185887098312378, "learning_rate": 1.9412020579124644e-06, "loss": 0.4669, "step": 17134 }, { "epoch": 0.7182754681785314, "grad_norm": 2.107445001602173, "learning_rate": 1.9406650842444786e-06, "loss": 0.509, "step": 17135 }, { "epoch": 0.7183173867935403, "grad_norm": 2.0208072662353516, "learning_rate": 1.940128166971103e-06, "loss": 0.5296, "step": 17136 }, { "epoch": 0.7183593054085493, "grad_norm": 2.19565749168396, "learning_rate": 1.939591306102236e-06, "loss": 0.4695, "step": 17137 }, { "epoch": 0.7184012240235582, "grad_norm": 2.195946455001831, "learning_rate": 1.9390545016477708e-06, "loss": 0.5053, "step": 17138 }, { "epoch": 0.7184431426385672, "grad_norm": 2.5079123973846436, "learning_rate": 1.9385177536176057e-06, "loss": 0.4317, "step": 17139 }, { "epoch": 0.7184850612535761, "grad_norm": 1.628757357597351, "learning_rate": 1.9379810620216333e-06, "loss": 0.4762, "step": 17140 }, { "epoch": 0.7185269798685852, "grad_norm": 2.3240113258361816, "learning_rate": 1.9374444268697456e-06, "loss": 0.4984, "step": 17141 }, { "epoch": 0.7185688984835941, "grad_norm": 1.9310047626495361, "learning_rate": 1.9369078481718356e-06, "loss": 0.5211, "step": 17142 }, { "epoch": 0.7186108170986031, "grad_norm": 1.922409176826477, "learning_rate": 1.936371325937797e-06, "loss": 0.4891, "step": 17143 }, { "epoch": 0.718652735713612, "grad_norm": 1.8964189291000366, "learning_rate": 1.9358348601775157e-06, "loss": 0.4722, "step": 17144 }, { "epoch": 0.718694654328621, "grad_norm": 1.5603395700454712, "learning_rate": 1.9352984509008827e-06, "loss": 0.4716, "step": 17145 }, { "epoch": 0.7187365729436299, "grad_norm": 4.128653526306152, "learning_rate": 1.934762098117788e-06, "loss": 0.5039, "step": 17146 }, { "epoch": 0.7187784915586389, "grad_norm": 2.8534975051879883, "learning_rate": 1.9342258018381145e-06, "loss": 0.459, "step": 17147 }, { "epoch": 0.7188204101736478, "grad_norm": 1.7855172157287598, "learning_rate": 1.9336895620717516e-06, "loss": 0.456, "step": 17148 }, { "epoch": 0.7188623287886569, "grad_norm": 2.651592493057251, "learning_rate": 1.933153378828583e-06, "loss": 0.5254, "step": 17149 }, { "epoch": 0.7189042474036658, "grad_norm": 2.1354763507843018, "learning_rate": 1.9326172521184896e-06, "loss": 0.5157, "step": 17150 }, { "epoch": 0.7189461660186748, "grad_norm": 1.6205533742904663, "learning_rate": 1.932081181951357e-06, "loss": 0.4981, "step": 17151 }, { "epoch": 0.7189880846336837, "grad_norm": 3.7185943126678467, "learning_rate": 1.9315451683370683e-06, "loss": 0.517, "step": 17152 }, { "epoch": 0.7190300032486927, "grad_norm": 2.1053597927093506, "learning_rate": 1.9310092112855004e-06, "loss": 0.5156, "step": 17153 }, { "epoch": 0.7190719218637016, "grad_norm": 1.8508708477020264, "learning_rate": 1.930473310806537e-06, "loss": 0.4704, "step": 17154 }, { "epoch": 0.7191138404787106, "grad_norm": 2.4173905849456787, "learning_rate": 1.9299374669100527e-06, "loss": 0.4835, "step": 17155 }, { "epoch": 0.7191557590937195, "grad_norm": 2.7478697299957275, "learning_rate": 1.929401679605927e-06, "loss": 0.478, "step": 17156 }, { "epoch": 0.7191976777087286, "grad_norm": 3.3174917697906494, "learning_rate": 1.928865948904039e-06, "loss": 0.4843, "step": 17157 }, { "epoch": 0.7192395963237375, "grad_norm": 2.0121257305145264, "learning_rate": 1.9283302748142606e-06, "loss": 0.4827, "step": 17158 }, { "epoch": 0.7192815149387464, "grad_norm": 2.26318097114563, "learning_rate": 1.9277946573464657e-06, "loss": 0.4385, "step": 17159 }, { "epoch": 0.7193234335537554, "grad_norm": 1.7973421812057495, "learning_rate": 1.9272590965105305e-06, "loss": 0.4503, "step": 17160 }, { "epoch": 0.7193653521687643, "grad_norm": 2.4671645164489746, "learning_rate": 1.926723592316324e-06, "loss": 0.5025, "step": 17161 }, { "epoch": 0.7194072707837733, "grad_norm": 1.6900767087936401, "learning_rate": 1.9261881447737198e-06, "loss": 0.4931, "step": 17162 }, { "epoch": 0.7194491893987822, "grad_norm": 1.8849984407424927, "learning_rate": 1.9256527538925894e-06, "loss": 0.4807, "step": 17163 }, { "epoch": 0.7194911080137912, "grad_norm": 1.8420917987823486, "learning_rate": 1.925117419682799e-06, "loss": 0.4418, "step": 17164 }, { "epoch": 0.7195330266288001, "grad_norm": 1.9590438604354858, "learning_rate": 1.924582142154217e-06, "loss": 0.4486, "step": 17165 }, { "epoch": 0.7195749452438092, "grad_norm": 2.047645330429077, "learning_rate": 1.9240469213167145e-06, "loss": 0.5061, "step": 17166 }, { "epoch": 0.7196168638588181, "grad_norm": 1.755964994430542, "learning_rate": 1.923511757180152e-06, "loss": 0.519, "step": 17167 }, { "epoch": 0.7196587824738271, "grad_norm": 1.9185000658035278, "learning_rate": 1.9229766497543993e-06, "loss": 0.5063, "step": 17168 }, { "epoch": 0.719700701088836, "grad_norm": 1.9362469911575317, "learning_rate": 1.9224415990493183e-06, "loss": 0.4777, "step": 17169 }, { "epoch": 0.719742619703845, "grad_norm": 2.653545379638672, "learning_rate": 1.9219066050747696e-06, "loss": 0.4415, "step": 17170 }, { "epoch": 0.7197845383188539, "grad_norm": 1.6443185806274414, "learning_rate": 1.921371667840618e-06, "loss": 0.4948, "step": 17171 }, { "epoch": 0.7198264569338629, "grad_norm": 1.9915978908538818, "learning_rate": 1.9208367873567246e-06, "loss": 0.4442, "step": 17172 }, { "epoch": 0.7198683755488718, "grad_norm": 2.075380563735962, "learning_rate": 1.920301963632947e-06, "loss": 0.5089, "step": 17173 }, { "epoch": 0.7199102941638809, "grad_norm": 1.7258909940719604, "learning_rate": 1.9197671966791453e-06, "loss": 0.4704, "step": 17174 }, { "epoch": 0.7199522127788898, "grad_norm": 1.6151666641235352, "learning_rate": 1.919232486505178e-06, "loss": 0.4655, "step": 17175 }, { "epoch": 0.7199941313938988, "grad_norm": 1.793295979499817, "learning_rate": 1.9186978331209e-06, "loss": 0.5111, "step": 17176 }, { "epoch": 0.7200360500089077, "grad_norm": 2.0040998458862305, "learning_rate": 1.9181632365361686e-06, "loss": 0.4933, "step": 17177 }, { "epoch": 0.7200779686239167, "grad_norm": 1.788356900215149, "learning_rate": 1.917628696760838e-06, "loss": 0.4921, "step": 17178 }, { "epoch": 0.7201198872389256, "grad_norm": 1.9701460599899292, "learning_rate": 1.9170942138047587e-06, "loss": 0.5016, "step": 17179 }, { "epoch": 0.7201618058539346, "grad_norm": 1.7114561796188354, "learning_rate": 1.9165597876777864e-06, "loss": 0.4527, "step": 17180 }, { "epoch": 0.7202037244689435, "grad_norm": 2.3742496967315674, "learning_rate": 1.9160254183897733e-06, "loss": 0.5172, "step": 17181 }, { "epoch": 0.7202456430839526, "grad_norm": 1.7967149019241333, "learning_rate": 1.915491105950566e-06, "loss": 0.4626, "step": 17182 }, { "epoch": 0.7202875616989615, "grad_norm": 2.250643253326416, "learning_rate": 1.914956850370016e-06, "loss": 0.4768, "step": 17183 }, { "epoch": 0.7203294803139704, "grad_norm": 1.7529739141464233, "learning_rate": 1.9144226516579734e-06, "loss": 0.4759, "step": 17184 }, { "epoch": 0.7203713989289794, "grad_norm": 1.9526753425598145, "learning_rate": 1.9138885098242814e-06, "loss": 0.5327, "step": 17185 }, { "epoch": 0.7204133175439883, "grad_norm": 1.7239447832107544, "learning_rate": 1.9133544248787904e-06, "loss": 0.4684, "step": 17186 }, { "epoch": 0.7204552361589973, "grad_norm": 2.314579486846924, "learning_rate": 1.9128203968313438e-06, "loss": 0.4539, "step": 17187 }, { "epoch": 0.7204971547740062, "grad_norm": 2.039257287979126, "learning_rate": 1.912286425691783e-06, "loss": 0.5535, "step": 17188 }, { "epoch": 0.7205390733890152, "grad_norm": 2.04620361328125, "learning_rate": 1.9117525114699536e-06, "loss": 0.4878, "step": 17189 }, { "epoch": 0.7205809920040241, "grad_norm": 2.589866876602173, "learning_rate": 1.911218654175699e-06, "loss": 0.5385, "step": 17190 }, { "epoch": 0.7206229106190332, "grad_norm": 2.7026991844177246, "learning_rate": 1.9106848538188566e-06, "loss": 0.5208, "step": 17191 }, { "epoch": 0.7206648292340421, "grad_norm": 2.254187822341919, "learning_rate": 1.910151110409268e-06, "loss": 0.4761, "step": 17192 }, { "epoch": 0.7207067478490511, "grad_norm": 2.5474045276641846, "learning_rate": 1.9096174239567738e-06, "loss": 0.4334, "step": 17193 }, { "epoch": 0.72074866646406, "grad_norm": 1.9174425601959229, "learning_rate": 1.909083794471208e-06, "loss": 0.5766, "step": 17194 }, { "epoch": 0.720790585079069, "grad_norm": 2.3287954330444336, "learning_rate": 1.908550221962412e-06, "loss": 0.4335, "step": 17195 }, { "epoch": 0.7208325036940779, "grad_norm": 1.876413106918335, "learning_rate": 1.9080167064402183e-06, "loss": 0.4937, "step": 17196 }, { "epoch": 0.7208744223090869, "grad_norm": 2.0128657817840576, "learning_rate": 1.90748324791446e-06, "loss": 0.5195, "step": 17197 }, { "epoch": 0.7209163409240958, "grad_norm": 3.0096397399902344, "learning_rate": 1.9069498463949732e-06, "loss": 0.5079, "step": 17198 }, { "epoch": 0.7209582595391049, "grad_norm": 2.2344632148742676, "learning_rate": 1.9064165018915914e-06, "loss": 0.4987, "step": 17199 }, { "epoch": 0.7210001781541138, "grad_norm": 2.030806541442871, "learning_rate": 1.9058832144141438e-06, "loss": 0.4748, "step": 17200 }, { "epoch": 0.7210420967691228, "grad_norm": 1.801802396774292, "learning_rate": 1.9053499839724605e-06, "loss": 0.4761, "step": 17201 }, { "epoch": 0.7210840153841317, "grad_norm": 1.7005293369293213, "learning_rate": 1.9048168105763743e-06, "loss": 0.4992, "step": 17202 }, { "epoch": 0.7211259339991407, "grad_norm": 1.764397382736206, "learning_rate": 1.9042836942357096e-06, "loss": 0.5085, "step": 17203 }, { "epoch": 0.7211678526141496, "grad_norm": 2.422839403152466, "learning_rate": 1.9037506349602969e-06, "loss": 0.4751, "step": 17204 }, { "epoch": 0.7212097712291586, "grad_norm": 1.8342493772506714, "learning_rate": 1.9032176327599606e-06, "loss": 0.4656, "step": 17205 }, { "epoch": 0.7212516898441675, "grad_norm": 2.143653392791748, "learning_rate": 1.9026846876445248e-06, "loss": 0.489, "step": 17206 }, { "epoch": 0.7212936084591766, "grad_norm": 2.162208318710327, "learning_rate": 1.9021517996238143e-06, "loss": 0.4604, "step": 17207 }, { "epoch": 0.7213355270741855, "grad_norm": 1.8318880796432495, "learning_rate": 1.9016189687076546e-06, "loss": 0.4876, "step": 17208 }, { "epoch": 0.7213774456891944, "grad_norm": 2.0970964431762695, "learning_rate": 1.9010861949058645e-06, "loss": 0.49, "step": 17209 }, { "epoch": 0.7214193643042034, "grad_norm": 2.222086191177368, "learning_rate": 1.900553478228266e-06, "loss": 0.4503, "step": 17210 }, { "epoch": 0.7214612829192123, "grad_norm": 2.127159357070923, "learning_rate": 1.9000208186846814e-06, "loss": 0.5151, "step": 17211 }, { "epoch": 0.7215032015342213, "grad_norm": 2.593048095703125, "learning_rate": 1.899488216284925e-06, "loss": 0.4755, "step": 17212 }, { "epoch": 0.7215451201492302, "grad_norm": 2.1636195182800293, "learning_rate": 1.898955671038819e-06, "loss": 0.4592, "step": 17213 }, { "epoch": 0.7215870387642392, "grad_norm": 3.240546703338623, "learning_rate": 1.8984231829561767e-06, "loss": 0.5317, "step": 17214 }, { "epoch": 0.7216289573792481, "grad_norm": 1.8630297183990479, "learning_rate": 1.8978907520468166e-06, "loss": 0.5035, "step": 17215 }, { "epoch": 0.7216708759942572, "grad_norm": 1.6762752532958984, "learning_rate": 1.8973583783205501e-06, "loss": 0.4563, "step": 17216 }, { "epoch": 0.7217127946092661, "grad_norm": 5.8951005935668945, "learning_rate": 1.8968260617871947e-06, "loss": 0.4943, "step": 17217 }, { "epoch": 0.7217547132242751, "grad_norm": 2.8625216484069824, "learning_rate": 1.8962938024565591e-06, "loss": 0.4469, "step": 17218 }, { "epoch": 0.721796631839284, "grad_norm": 1.9223567247390747, "learning_rate": 1.8957616003384565e-06, "loss": 0.4585, "step": 17219 }, { "epoch": 0.721838550454293, "grad_norm": 1.9209040403366089, "learning_rate": 1.8952294554426991e-06, "loss": 0.485, "step": 17220 }, { "epoch": 0.7218804690693019, "grad_norm": 2.3182685375213623, "learning_rate": 1.8946973677790926e-06, "loss": 0.4878, "step": 17221 }, { "epoch": 0.7219223876843109, "grad_norm": 1.80973482131958, "learning_rate": 1.894165337357447e-06, "loss": 0.4533, "step": 17222 }, { "epoch": 0.7219643062993198, "grad_norm": 1.9364808797836304, "learning_rate": 1.8936333641875715e-06, "loss": 0.5073, "step": 17223 }, { "epoch": 0.7220062249143289, "grad_norm": 2.3219971656799316, "learning_rate": 1.8931014482792709e-06, "loss": 0.5267, "step": 17224 }, { "epoch": 0.7220481435293378, "grad_norm": 2.1161954402923584, "learning_rate": 1.8925695896423479e-06, "loss": 0.4992, "step": 17225 }, { "epoch": 0.7220900621443468, "grad_norm": 2.219501495361328, "learning_rate": 1.8920377882866102e-06, "loss": 0.5223, "step": 17226 }, { "epoch": 0.7221319807593557, "grad_norm": 2.932074785232544, "learning_rate": 1.891506044221858e-06, "loss": 0.4843, "step": 17227 }, { "epoch": 0.7221738993743647, "grad_norm": 1.9508692026138306, "learning_rate": 1.8909743574578959e-06, "loss": 0.5335, "step": 17228 }, { "epoch": 0.7222158179893736, "grad_norm": 2.230323076248169, "learning_rate": 1.8904427280045218e-06, "loss": 0.5765, "step": 17229 }, { "epoch": 0.7222577366043826, "grad_norm": 5.8318681716918945, "learning_rate": 1.8899111558715366e-06, "loss": 0.4965, "step": 17230 }, { "epoch": 0.7222996552193915, "grad_norm": 1.7575037479400635, "learning_rate": 1.889379641068742e-06, "loss": 0.4722, "step": 17231 }, { "epoch": 0.7223415738344006, "grad_norm": 1.6632236242294312, "learning_rate": 1.888848183605932e-06, "loss": 0.4679, "step": 17232 }, { "epoch": 0.7223834924494095, "grad_norm": 1.8845391273498535, "learning_rate": 1.8883167834929056e-06, "loss": 0.4416, "step": 17233 }, { "epoch": 0.7224254110644184, "grad_norm": 1.8798917531967163, "learning_rate": 1.887785440739458e-06, "loss": 0.4785, "step": 17234 }, { "epoch": 0.7224673296794274, "grad_norm": 1.7539341449737549, "learning_rate": 1.887254155355382e-06, "loss": 0.4783, "step": 17235 }, { "epoch": 0.7225092482944363, "grad_norm": 2.2398006916046143, "learning_rate": 1.8867229273504717e-06, "loss": 0.5603, "step": 17236 }, { "epoch": 0.7225511669094453, "grad_norm": 2.221312999725342, "learning_rate": 1.8861917567345222e-06, "loss": 0.5471, "step": 17237 }, { "epoch": 0.7225930855244542, "grad_norm": 1.7416400909423828, "learning_rate": 1.8856606435173214e-06, "loss": 0.4914, "step": 17238 }, { "epoch": 0.7226350041394632, "grad_norm": 2.399430751800537, "learning_rate": 1.8851295877086606e-06, "loss": 0.4676, "step": 17239 }, { "epoch": 0.7226769227544722, "grad_norm": 1.77982497215271, "learning_rate": 1.8845985893183321e-06, "loss": 0.4963, "step": 17240 }, { "epoch": 0.7227188413694812, "grad_norm": 2.351165771484375, "learning_rate": 1.8840676483561194e-06, "loss": 0.5298, "step": 17241 }, { "epoch": 0.7227607599844901, "grad_norm": 1.9304969310760498, "learning_rate": 1.8835367648318143e-06, "loss": 0.5432, "step": 17242 }, { "epoch": 0.7228026785994991, "grad_norm": 1.9670884609222412, "learning_rate": 1.8830059387552001e-06, "loss": 0.506, "step": 17243 }, { "epoch": 0.722844597214508, "grad_norm": 1.8322789669036865, "learning_rate": 1.8824751701360605e-06, "loss": 0.4139, "step": 17244 }, { "epoch": 0.722886515829517, "grad_norm": 1.8404386043548584, "learning_rate": 1.8819444589841818e-06, "loss": 0.499, "step": 17245 }, { "epoch": 0.7229284344445259, "grad_norm": 1.6320850849151611, "learning_rate": 1.8814138053093478e-06, "loss": 0.4816, "step": 17246 }, { "epoch": 0.7229703530595349, "grad_norm": 1.8321166038513184, "learning_rate": 1.8808832091213369e-06, "loss": 0.4807, "step": 17247 }, { "epoch": 0.7230122716745438, "grad_norm": 2.3360984325408936, "learning_rate": 1.8803526704299324e-06, "loss": 0.4341, "step": 17248 }, { "epoch": 0.7230541902895529, "grad_norm": 1.9243649244308472, "learning_rate": 1.8798221892449152e-06, "loss": 0.4887, "step": 17249 }, { "epoch": 0.7230961089045618, "grad_norm": 1.9405343532562256, "learning_rate": 1.8792917655760612e-06, "loss": 0.532, "step": 17250 }, { "epoch": 0.7231380275195708, "grad_norm": 2.3642091751098633, "learning_rate": 1.8787613994331506e-06, "loss": 0.4622, "step": 17251 }, { "epoch": 0.7231799461345797, "grad_norm": 1.7323142290115356, "learning_rate": 1.8782310908259588e-06, "loss": 0.4597, "step": 17252 }, { "epoch": 0.7232218647495887, "grad_norm": 1.5625481605529785, "learning_rate": 1.8777008397642594e-06, "loss": 0.4848, "step": 17253 }, { "epoch": 0.7232637833645976, "grad_norm": 1.6777557134628296, "learning_rate": 1.8771706462578282e-06, "loss": 0.4542, "step": 17254 }, { "epoch": 0.7233057019796066, "grad_norm": 1.9460641145706177, "learning_rate": 1.8766405103164409e-06, "loss": 0.4565, "step": 17255 }, { "epoch": 0.7233476205946155, "grad_norm": 1.7951974868774414, "learning_rate": 1.8761104319498663e-06, "loss": 0.4466, "step": 17256 }, { "epoch": 0.7233895392096246, "grad_norm": 1.6931670904159546, "learning_rate": 1.8755804111678765e-06, "loss": 0.4257, "step": 17257 }, { "epoch": 0.7234314578246335, "grad_norm": 2.035914897918701, "learning_rate": 1.875050447980245e-06, "loss": 0.4726, "step": 17258 }, { "epoch": 0.7234733764396424, "grad_norm": 1.8819661140441895, "learning_rate": 1.8745205423967367e-06, "loss": 0.5191, "step": 17259 }, { "epoch": 0.7235152950546514, "grad_norm": 4.9372358322143555, "learning_rate": 1.8739906944271224e-06, "loss": 0.4938, "step": 17260 }, { "epoch": 0.7235572136696603, "grad_norm": 2.0291380882263184, "learning_rate": 1.873460904081167e-06, "loss": 0.5511, "step": 17261 }, { "epoch": 0.7235991322846693, "grad_norm": 1.8487426042556763, "learning_rate": 1.8729311713686387e-06, "loss": 0.4751, "step": 17262 }, { "epoch": 0.7236410508996782, "grad_norm": 2.4545722007751465, "learning_rate": 1.8724014962992997e-06, "loss": 0.4933, "step": 17263 }, { "epoch": 0.7236829695146872, "grad_norm": 3.3551392555236816, "learning_rate": 1.8718718788829172e-06, "loss": 0.4659, "step": 17264 }, { "epoch": 0.7237248881296962, "grad_norm": 2.156975746154785, "learning_rate": 1.8713423191292507e-06, "loss": 0.4863, "step": 17265 }, { "epoch": 0.7237668067447052, "grad_norm": 1.8528027534484863, "learning_rate": 1.8708128170480627e-06, "loss": 0.4813, "step": 17266 }, { "epoch": 0.7238087253597141, "grad_norm": 1.7913827896118164, "learning_rate": 1.8702833726491164e-06, "loss": 0.477, "step": 17267 }, { "epoch": 0.7238506439747231, "grad_norm": 1.7494118213653564, "learning_rate": 1.8697539859421676e-06, "loss": 0.4344, "step": 17268 }, { "epoch": 0.723892562589732, "grad_norm": 1.629348635673523, "learning_rate": 1.8692246569369765e-06, "loss": 0.4903, "step": 17269 }, { "epoch": 0.723934481204741, "grad_norm": 1.858197808265686, "learning_rate": 1.8686953856433026e-06, "loss": 0.4677, "step": 17270 }, { "epoch": 0.7239763998197499, "grad_norm": 1.894033432006836, "learning_rate": 1.8681661720709005e-06, "loss": 0.4787, "step": 17271 }, { "epoch": 0.724018318434759, "grad_norm": 2.359215021133423, "learning_rate": 1.8676370162295238e-06, "loss": 0.4329, "step": 17272 }, { "epoch": 0.7240602370497679, "grad_norm": 1.9374186992645264, "learning_rate": 1.8671079181289299e-06, "loss": 0.4863, "step": 17273 }, { "epoch": 0.7241021556647769, "grad_norm": 2.0412251949310303, "learning_rate": 1.8665788777788686e-06, "loss": 0.4862, "step": 17274 }, { "epoch": 0.7241440742797858, "grad_norm": 2.0725183486938477, "learning_rate": 1.8660498951890938e-06, "loss": 0.5264, "step": 17275 }, { "epoch": 0.7241859928947948, "grad_norm": 1.840206503868103, "learning_rate": 1.8655209703693583e-06, "loss": 0.479, "step": 17276 }, { "epoch": 0.7242279115098037, "grad_norm": 1.7493821382522583, "learning_rate": 1.8649921033294088e-06, "loss": 0.4877, "step": 17277 }, { "epoch": 0.7242698301248127, "grad_norm": 2.8893935680389404, "learning_rate": 1.864463294078996e-06, "loss": 0.4546, "step": 17278 }, { "epoch": 0.7243117487398216, "grad_norm": 1.9359874725341797, "learning_rate": 1.8639345426278693e-06, "loss": 0.4401, "step": 17279 }, { "epoch": 0.7243536673548306, "grad_norm": 14.860313415527344, "learning_rate": 1.8634058489857736e-06, "loss": 0.4829, "step": 17280 }, { "epoch": 0.7243955859698396, "grad_norm": 1.9069551229476929, "learning_rate": 1.8628772131624533e-06, "loss": 0.5131, "step": 17281 }, { "epoch": 0.7244375045848486, "grad_norm": 5.83101749420166, "learning_rate": 1.8623486351676567e-06, "loss": 0.4334, "step": 17282 }, { "epoch": 0.7244794231998575, "grad_norm": 2.780940532684326, "learning_rate": 1.8618201150111232e-06, "loss": 0.4943, "step": 17283 }, { "epoch": 0.7245213418148664, "grad_norm": 2.584653615951538, "learning_rate": 1.8612916527025976e-06, "loss": 0.4838, "step": 17284 }, { "epoch": 0.7245632604298754, "grad_norm": 1.9159488677978516, "learning_rate": 1.8607632482518228e-06, "loss": 0.4246, "step": 17285 }, { "epoch": 0.7246051790448843, "grad_norm": 2.865325927734375, "learning_rate": 1.860234901668536e-06, "loss": 0.4452, "step": 17286 }, { "epoch": 0.7246470976598933, "grad_norm": 1.766890287399292, "learning_rate": 1.8597066129624785e-06, "loss": 0.4808, "step": 17287 }, { "epoch": 0.7246890162749022, "grad_norm": 2.1730990409851074, "learning_rate": 1.8591783821433896e-06, "loss": 0.4765, "step": 17288 }, { "epoch": 0.7247309348899112, "grad_norm": 1.8142993450164795, "learning_rate": 1.858650209221005e-06, "loss": 0.4758, "step": 17289 }, { "epoch": 0.7247728535049202, "grad_norm": 1.8424917459487915, "learning_rate": 1.8581220942050598e-06, "loss": 0.462, "step": 17290 }, { "epoch": 0.7248147721199292, "grad_norm": 2.08625864982605, "learning_rate": 1.857594037105292e-06, "loss": 0.4889, "step": 17291 }, { "epoch": 0.7248566907349381, "grad_norm": 2.7325778007507324, "learning_rate": 1.857066037931432e-06, "loss": 0.464, "step": 17292 }, { "epoch": 0.7248986093499471, "grad_norm": 2.146643877029419, "learning_rate": 1.8565380966932146e-06, "loss": 0.5039, "step": 17293 }, { "epoch": 0.724940527964956, "grad_norm": 2.132479429244995, "learning_rate": 1.8560102134003732e-06, "loss": 0.4871, "step": 17294 }, { "epoch": 0.724982446579965, "grad_norm": 1.9222551584243774, "learning_rate": 1.8554823880626355e-06, "loss": 0.4702, "step": 17295 }, { "epoch": 0.7250243651949739, "grad_norm": 2.3576571941375732, "learning_rate": 1.854954620689733e-06, "loss": 0.45, "step": 17296 }, { "epoch": 0.725066283809983, "grad_norm": 2.007500648498535, "learning_rate": 1.8544269112913953e-06, "loss": 0.442, "step": 17297 }, { "epoch": 0.7251082024249919, "grad_norm": 1.9199246168136597, "learning_rate": 1.8538992598773499e-06, "loss": 0.4672, "step": 17298 }, { "epoch": 0.7251501210400009, "grad_norm": 1.8700224161148071, "learning_rate": 1.85337166645732e-06, "loss": 0.4997, "step": 17299 }, { "epoch": 0.7251920396550098, "grad_norm": 2.088158369064331, "learning_rate": 1.852844131041035e-06, "loss": 0.5065, "step": 17300 }, { "epoch": 0.7252339582700188, "grad_norm": 2.082887887954712, "learning_rate": 1.8523166536382158e-06, "loss": 0.4185, "step": 17301 }, { "epoch": 0.7252758768850277, "grad_norm": 2.1780009269714355, "learning_rate": 1.8517892342585874e-06, "loss": 0.4234, "step": 17302 }, { "epoch": 0.7253177955000367, "grad_norm": 1.8901548385620117, "learning_rate": 1.8512618729118743e-06, "loss": 0.4814, "step": 17303 }, { "epoch": 0.7253597141150456, "grad_norm": 2.0116045475006104, "learning_rate": 1.8507345696077938e-06, "loss": 0.4983, "step": 17304 }, { "epoch": 0.7254016327300546, "grad_norm": 1.8533930778503418, "learning_rate": 1.8502073243560698e-06, "loss": 0.4727, "step": 17305 }, { "epoch": 0.7254435513450636, "grad_norm": 1.8275210857391357, "learning_rate": 1.8496801371664169e-06, "loss": 0.4539, "step": 17306 }, { "epoch": 0.7254854699600726, "grad_norm": 2.0761070251464844, "learning_rate": 1.849153008048556e-06, "loss": 0.457, "step": 17307 }, { "epoch": 0.7255273885750815, "grad_norm": 4.239376068115234, "learning_rate": 1.848625937012205e-06, "loss": 0.5091, "step": 17308 }, { "epoch": 0.7255693071900904, "grad_norm": 1.7636843919754028, "learning_rate": 1.8480989240670783e-06, "loss": 0.459, "step": 17309 }, { "epoch": 0.7256112258050994, "grad_norm": 1.9105654954910278, "learning_rate": 1.8475719692228888e-06, "loss": 0.5157, "step": 17310 }, { "epoch": 0.7256531444201083, "grad_norm": 4.065323829650879, "learning_rate": 1.8470450724893535e-06, "loss": 0.4809, "step": 17311 }, { "epoch": 0.7256950630351173, "grad_norm": 1.9056346416473389, "learning_rate": 1.8465182338761823e-06, "loss": 0.4787, "step": 17312 }, { "epoch": 0.7257369816501262, "grad_norm": 1.7581472396850586, "learning_rate": 1.8459914533930872e-06, "loss": 0.4492, "step": 17313 }, { "epoch": 0.7257789002651353, "grad_norm": 2.1995689868927, "learning_rate": 1.8454647310497814e-06, "loss": 0.4574, "step": 17314 }, { "epoch": 0.7258208188801442, "grad_norm": 1.9863988161087036, "learning_rate": 1.8449380668559708e-06, "loss": 0.4739, "step": 17315 }, { "epoch": 0.7258627374951532, "grad_norm": 4.760396957397461, "learning_rate": 1.8444114608213648e-06, "loss": 0.5108, "step": 17316 }, { "epoch": 0.7259046561101621, "grad_norm": 2.195758819580078, "learning_rate": 1.843884912955673e-06, "loss": 0.5093, "step": 17317 }, { "epoch": 0.7259465747251711, "grad_norm": 2.1141107082366943, "learning_rate": 1.843358423268599e-06, "loss": 0.4469, "step": 17318 }, { "epoch": 0.72598849334018, "grad_norm": 2.1225807666778564, "learning_rate": 1.8428319917698478e-06, "loss": 0.4934, "step": 17319 }, { "epoch": 0.726030411955189, "grad_norm": 4.103355884552002, "learning_rate": 1.8423056184691258e-06, "loss": 0.4478, "step": 17320 }, { "epoch": 0.7260723305701979, "grad_norm": 2.383129358291626, "learning_rate": 1.8417793033761323e-06, "loss": 0.485, "step": 17321 }, { "epoch": 0.726114249185207, "grad_norm": 1.8684800863265991, "learning_rate": 1.8412530465005706e-06, "loss": 0.5403, "step": 17322 }, { "epoch": 0.7261561678002159, "grad_norm": 1.9534679651260376, "learning_rate": 1.8407268478521445e-06, "loss": 0.5315, "step": 17323 }, { "epoch": 0.7261980864152249, "grad_norm": 2.115171194076538, "learning_rate": 1.8402007074405498e-06, "loss": 0.5063, "step": 17324 }, { "epoch": 0.7262400050302338, "grad_norm": 2.961127758026123, "learning_rate": 1.8396746252754866e-06, "loss": 0.4383, "step": 17325 }, { "epoch": 0.7262819236452428, "grad_norm": 1.8897390365600586, "learning_rate": 1.8391486013666542e-06, "loss": 0.4313, "step": 17326 }, { "epoch": 0.7263238422602517, "grad_norm": 1.5803861618041992, "learning_rate": 1.838622635723748e-06, "loss": 0.4023, "step": 17327 }, { "epoch": 0.7263657608752607, "grad_norm": 1.7745062112808228, "learning_rate": 1.8380967283564611e-06, "loss": 0.4056, "step": 17328 }, { "epoch": 0.7264076794902696, "grad_norm": 2.2803354263305664, "learning_rate": 1.8375708792744917e-06, "loss": 0.4659, "step": 17329 }, { "epoch": 0.7264495981052786, "grad_norm": 1.630673885345459, "learning_rate": 1.8370450884875295e-06, "loss": 0.3832, "step": 17330 }, { "epoch": 0.7264915167202876, "grad_norm": 2.5612375736236572, "learning_rate": 1.836519356005268e-06, "loss": 0.5144, "step": 17331 }, { "epoch": 0.7265334353352966, "grad_norm": 2.1482322216033936, "learning_rate": 1.8359936818374008e-06, "loss": 0.496, "step": 17332 }, { "epoch": 0.7265753539503055, "grad_norm": 2.0047760009765625, "learning_rate": 1.8354680659936147e-06, "loss": 0.4623, "step": 17333 }, { "epoch": 0.7266172725653144, "grad_norm": 2.6306707859039307, "learning_rate": 1.8349425084835993e-06, "loss": 0.4767, "step": 17334 }, { "epoch": 0.7266591911803234, "grad_norm": 2.1380255222320557, "learning_rate": 1.834417009317045e-06, "loss": 0.4912, "step": 17335 }, { "epoch": 0.7267011097953323, "grad_norm": 2.189481019973755, "learning_rate": 1.8338915685036367e-06, "loss": 0.4979, "step": 17336 }, { "epoch": 0.7267430284103413, "grad_norm": 2.953550338745117, "learning_rate": 1.8333661860530595e-06, "loss": 0.5158, "step": 17337 }, { "epoch": 0.7267849470253502, "grad_norm": 2.524420738220215, "learning_rate": 1.8328408619749999e-06, "loss": 0.4879, "step": 17338 }, { "epoch": 0.7268268656403593, "grad_norm": 2.3546082973480225, "learning_rate": 1.832315596279139e-06, "loss": 0.4566, "step": 17339 }, { "epoch": 0.7268687842553682, "grad_norm": 2.448845624923706, "learning_rate": 1.831790388975161e-06, "loss": 0.4973, "step": 17340 }, { "epoch": 0.7269107028703772, "grad_norm": 1.9647397994995117, "learning_rate": 1.8312652400727481e-06, "loss": 0.5372, "step": 17341 }, { "epoch": 0.7269526214853861, "grad_norm": 2.29095458984375, "learning_rate": 1.8307401495815791e-06, "loss": 0.4921, "step": 17342 }, { "epoch": 0.7269945401003951, "grad_norm": 2.407510280609131, "learning_rate": 1.8302151175113343e-06, "loss": 0.4757, "step": 17343 }, { "epoch": 0.727036458715404, "grad_norm": 1.979101538658142, "learning_rate": 1.8296901438716924e-06, "loss": 0.5186, "step": 17344 }, { "epoch": 0.727078377330413, "grad_norm": 2.903646230697632, "learning_rate": 1.8291652286723306e-06, "loss": 0.496, "step": 17345 }, { "epoch": 0.7271202959454219, "grad_norm": 2.740189552307129, "learning_rate": 1.8286403719229224e-06, "loss": 0.5276, "step": 17346 }, { "epoch": 0.727162214560431, "grad_norm": 2.07826566696167, "learning_rate": 1.8281155736331463e-06, "loss": 0.5022, "step": 17347 }, { "epoch": 0.7272041331754399, "grad_norm": 1.8234678506851196, "learning_rate": 1.827590833812673e-06, "loss": 0.4539, "step": 17348 }, { "epoch": 0.7272460517904489, "grad_norm": 2.7976889610290527, "learning_rate": 1.827066152471177e-06, "loss": 0.5379, "step": 17349 }, { "epoch": 0.7272879704054578, "grad_norm": 2.1260592937469482, "learning_rate": 1.8265415296183315e-06, "loss": 0.4918, "step": 17350 }, { "epoch": 0.7273298890204668, "grad_norm": 2.276353597640991, "learning_rate": 1.8260169652638044e-06, "loss": 0.4734, "step": 17351 }, { "epoch": 0.7273718076354757, "grad_norm": 2.1953423023223877, "learning_rate": 1.8254924594172662e-06, "loss": 0.5149, "step": 17352 }, { "epoch": 0.7274137262504847, "grad_norm": 2.078516960144043, "learning_rate": 1.8249680120883878e-06, "loss": 0.4917, "step": 17353 }, { "epoch": 0.7274556448654936, "grad_norm": 1.7327170372009277, "learning_rate": 1.824443623286833e-06, "loss": 0.4633, "step": 17354 }, { "epoch": 0.7274975634805027, "grad_norm": 1.9809002876281738, "learning_rate": 1.8239192930222715e-06, "loss": 0.4865, "step": 17355 }, { "epoch": 0.7275394820955116, "grad_norm": 2.0015835762023926, "learning_rate": 1.8233950213043671e-06, "loss": 0.5307, "step": 17356 }, { "epoch": 0.7275814007105206, "grad_norm": 2.051281213760376, "learning_rate": 1.8228708081427826e-06, "loss": 0.4834, "step": 17357 }, { "epoch": 0.7276233193255295, "grad_norm": 2.297178030014038, "learning_rate": 1.8223466535471823e-06, "loss": 0.4784, "step": 17358 }, { "epoch": 0.7276652379405384, "grad_norm": 1.9925161600112915, "learning_rate": 1.8218225575272302e-06, "loss": 0.5134, "step": 17359 }, { "epoch": 0.7277071565555474, "grad_norm": 1.7451610565185547, "learning_rate": 1.8212985200925836e-06, "loss": 0.4884, "step": 17360 }, { "epoch": 0.7277490751705563, "grad_norm": 1.8296139240264893, "learning_rate": 1.8207745412529048e-06, "loss": 0.5024, "step": 17361 }, { "epoch": 0.7277909937855653, "grad_norm": 2.9071905612945557, "learning_rate": 1.8202506210178534e-06, "loss": 0.4792, "step": 17362 }, { "epoch": 0.7278329124005742, "grad_norm": 1.8587543964385986, "learning_rate": 1.8197267593970847e-06, "loss": 0.4995, "step": 17363 }, { "epoch": 0.7278748310155833, "grad_norm": 2.5549697875976562, "learning_rate": 1.8192029564002578e-06, "loss": 0.5363, "step": 17364 }, { "epoch": 0.7279167496305922, "grad_norm": 1.9460701942443848, "learning_rate": 1.8186792120370279e-06, "loss": 0.5205, "step": 17365 }, { "epoch": 0.7279586682456012, "grad_norm": 1.8720077276229858, "learning_rate": 1.8181555263170457e-06, "loss": 0.4426, "step": 17366 }, { "epoch": 0.7280005868606101, "grad_norm": 2.4570882320404053, "learning_rate": 1.8176318992499687e-06, "loss": 0.5118, "step": 17367 }, { "epoch": 0.7280425054756191, "grad_norm": 1.7692705392837524, "learning_rate": 1.8171083308454495e-06, "loss": 0.4307, "step": 17368 }, { "epoch": 0.728084424090628, "grad_norm": 1.6805315017700195, "learning_rate": 1.8165848211131359e-06, "loss": 0.4022, "step": 17369 }, { "epoch": 0.728126342705637, "grad_norm": 2.1692636013031006, "learning_rate": 1.8160613700626806e-06, "loss": 0.5199, "step": 17370 }, { "epoch": 0.7281682613206459, "grad_norm": 1.721313238143921, "learning_rate": 1.815537977703734e-06, "loss": 0.4335, "step": 17371 }, { "epoch": 0.728210179935655, "grad_norm": 1.9394155740737915, "learning_rate": 1.815014644045941e-06, "loss": 0.4868, "step": 17372 }, { "epoch": 0.7282520985506639, "grad_norm": 1.8342779874801636, "learning_rate": 1.8144913690989513e-06, "loss": 0.5287, "step": 17373 }, { "epoch": 0.7282940171656729, "grad_norm": 2.6853599548339844, "learning_rate": 1.813968152872409e-06, "loss": 0.4819, "step": 17374 }, { "epoch": 0.7283359357806818, "grad_norm": 2.7562098503112793, "learning_rate": 1.813444995375958e-06, "loss": 0.4077, "step": 17375 }, { "epoch": 0.7283778543956908, "grad_norm": 3.1240153312683105, "learning_rate": 1.812921896619243e-06, "loss": 0.4852, "step": 17376 }, { "epoch": 0.7284197730106997, "grad_norm": 2.001075506210327, "learning_rate": 1.8123988566119088e-06, "loss": 0.5474, "step": 17377 }, { "epoch": 0.7284616916257087, "grad_norm": 2.0361382961273193, "learning_rate": 1.811875875363593e-06, "loss": 0.4725, "step": 17378 }, { "epoch": 0.7285036102407176, "grad_norm": 2.0798394680023193, "learning_rate": 1.8113529528839396e-06, "loss": 0.4724, "step": 17379 }, { "epoch": 0.7285455288557267, "grad_norm": 1.7067558765411377, "learning_rate": 1.8108300891825852e-06, "loss": 0.4253, "step": 17380 }, { "epoch": 0.7285874474707356, "grad_norm": 1.938232660293579, "learning_rate": 1.8103072842691689e-06, "loss": 0.4722, "step": 17381 }, { "epoch": 0.7286293660857446, "grad_norm": 1.91451096534729, "learning_rate": 1.8097845381533297e-06, "loss": 0.5171, "step": 17382 }, { "epoch": 0.7286712847007535, "grad_norm": 2.1062004566192627, "learning_rate": 1.8092618508447024e-06, "loss": 0.5001, "step": 17383 }, { "epoch": 0.7287132033157624, "grad_norm": 1.9113147258758545, "learning_rate": 1.8087392223529198e-06, "loss": 0.4877, "step": 17384 }, { "epoch": 0.7287551219307714, "grad_norm": 2.6123762130737305, "learning_rate": 1.8082166526876198e-06, "loss": 0.4968, "step": 17385 }, { "epoch": 0.7287970405457803, "grad_norm": 1.8750100135803223, "learning_rate": 1.8076941418584316e-06, "loss": 0.4361, "step": 17386 }, { "epoch": 0.7288389591607893, "grad_norm": 4.734814643859863, "learning_rate": 1.8071716898749886e-06, "loss": 0.5533, "step": 17387 }, { "epoch": 0.7288808777757982, "grad_norm": 2.6760294437408447, "learning_rate": 1.8066492967469235e-06, "loss": 0.4617, "step": 17388 }, { "epoch": 0.7289227963908073, "grad_norm": 1.549178123474121, "learning_rate": 1.8061269624838618e-06, "loss": 0.4467, "step": 17389 }, { "epoch": 0.7289647150058162, "grad_norm": 2.4517290592193604, "learning_rate": 1.8056046870954346e-06, "loss": 0.4454, "step": 17390 }, { "epoch": 0.7290066336208252, "grad_norm": 2.0263328552246094, "learning_rate": 1.80508247059127e-06, "loss": 0.4841, "step": 17391 }, { "epoch": 0.7290485522358341, "grad_norm": 1.7273411750793457, "learning_rate": 1.804560312980994e-06, "loss": 0.4772, "step": 17392 }, { "epoch": 0.7290904708508431, "grad_norm": 2.4625918865203857, "learning_rate": 1.8040382142742292e-06, "loss": 0.4969, "step": 17393 }, { "epoch": 0.729132389465852, "grad_norm": 4.039811611175537, "learning_rate": 1.8035161744806029e-06, "loss": 0.4726, "step": 17394 }, { "epoch": 0.729174308080861, "grad_norm": 1.7778681516647339, "learning_rate": 1.8029941936097351e-06, "loss": 0.4592, "step": 17395 }, { "epoch": 0.7292162266958699, "grad_norm": 2.4566056728363037, "learning_rate": 1.80247227167125e-06, "loss": 0.5537, "step": 17396 }, { "epoch": 0.729258145310879, "grad_norm": 1.6906325817108154, "learning_rate": 1.8019504086747697e-06, "loss": 0.4559, "step": 17397 }, { "epoch": 0.7293000639258879, "grad_norm": 1.9137041568756104, "learning_rate": 1.8014286046299107e-06, "loss": 0.4885, "step": 17398 }, { "epoch": 0.7293419825408969, "grad_norm": 1.881880521774292, "learning_rate": 1.8009068595462937e-06, "loss": 0.4559, "step": 17399 }, { "epoch": 0.7293839011559058, "grad_norm": 2.1967532634735107, "learning_rate": 1.8003851734335376e-06, "loss": 0.5036, "step": 17400 }, { "epoch": 0.7294258197709148, "grad_norm": 2.070793867111206, "learning_rate": 1.799863546301256e-06, "loss": 0.4677, "step": 17401 }, { "epoch": 0.7294677383859237, "grad_norm": 2.316226005554199, "learning_rate": 1.7993419781590676e-06, "loss": 0.5024, "step": 17402 }, { "epoch": 0.7295096570009327, "grad_norm": 2.2918732166290283, "learning_rate": 1.798820469016585e-06, "loss": 0.4939, "step": 17403 }, { "epoch": 0.7295515756159416, "grad_norm": 2.141770124435425, "learning_rate": 1.7982990188834192e-06, "loss": 0.4392, "step": 17404 }, { "epoch": 0.7295934942309507, "grad_norm": 2.052881956100464, "learning_rate": 1.7977776277691856e-06, "loss": 0.4785, "step": 17405 }, { "epoch": 0.7296354128459596, "grad_norm": 3.4780519008636475, "learning_rate": 1.797256295683496e-06, "loss": 0.4964, "step": 17406 }, { "epoch": 0.7296773314609686, "grad_norm": 1.9827076196670532, "learning_rate": 1.796735022635957e-06, "loss": 0.4926, "step": 17407 }, { "epoch": 0.7297192500759775, "grad_norm": 1.7738057374954224, "learning_rate": 1.7962138086361797e-06, "loss": 0.4601, "step": 17408 }, { "epoch": 0.7297611686909864, "grad_norm": 2.251721143722534, "learning_rate": 1.7956926536937735e-06, "loss": 0.5078, "step": 17409 }, { "epoch": 0.7298030873059954, "grad_norm": 1.8702186346054077, "learning_rate": 1.7951715578183414e-06, "loss": 0.4825, "step": 17410 }, { "epoch": 0.7298450059210043, "grad_norm": 1.7714213132858276, "learning_rate": 1.7946505210194931e-06, "loss": 0.4845, "step": 17411 }, { "epoch": 0.7298869245360133, "grad_norm": 1.888914942741394, "learning_rate": 1.7941295433068311e-06, "loss": 0.5397, "step": 17412 }, { "epoch": 0.7299288431510222, "grad_norm": 2.082775115966797, "learning_rate": 1.793608624689957e-06, "loss": 0.5155, "step": 17413 }, { "epoch": 0.7299707617660313, "grad_norm": 2.0267131328582764, "learning_rate": 1.7930877651784756e-06, "loss": 0.5031, "step": 17414 }, { "epoch": 0.7300126803810402, "grad_norm": 1.8967974185943604, "learning_rate": 1.7925669647819893e-06, "loss": 0.469, "step": 17415 }, { "epoch": 0.7300545989960492, "grad_norm": 1.692001223564148, "learning_rate": 1.7920462235100955e-06, "loss": 0.4302, "step": 17416 }, { "epoch": 0.7300965176110581, "grad_norm": 1.8441543579101562, "learning_rate": 1.7915255413723941e-06, "loss": 0.4987, "step": 17417 }, { "epoch": 0.7301384362260671, "grad_norm": 1.6663702726364136, "learning_rate": 1.791004918378486e-06, "loss": 0.4792, "step": 17418 }, { "epoch": 0.730180354841076, "grad_norm": 1.7871352434158325, "learning_rate": 1.790484354537964e-06, "loss": 0.4814, "step": 17419 }, { "epoch": 0.730222273456085, "grad_norm": 2.2193377017974854, "learning_rate": 1.7899638498604278e-06, "loss": 0.4933, "step": 17420 }, { "epoch": 0.7302641920710939, "grad_norm": 1.9260430335998535, "learning_rate": 1.7894434043554697e-06, "loss": 0.5189, "step": 17421 }, { "epoch": 0.730306110686103, "grad_norm": 3.4966342449188232, "learning_rate": 1.7889230180326828e-06, "loss": 0.4171, "step": 17422 }, { "epoch": 0.7303480293011119, "grad_norm": 1.7049674987792969, "learning_rate": 1.7884026909016606e-06, "loss": 0.4872, "step": 17423 }, { "epoch": 0.7303899479161209, "grad_norm": 2.1407864093780518, "learning_rate": 1.787882422971997e-06, "loss": 0.4968, "step": 17424 }, { "epoch": 0.7304318665311298, "grad_norm": 1.9971929788589478, "learning_rate": 1.787362214253278e-06, "loss": 0.5093, "step": 17425 }, { "epoch": 0.7304737851461388, "grad_norm": 1.9371957778930664, "learning_rate": 1.786842064755096e-06, "loss": 0.4333, "step": 17426 }, { "epoch": 0.7305157037611477, "grad_norm": 1.6514760255813599, "learning_rate": 1.7863219744870397e-06, "loss": 0.3974, "step": 17427 }, { "epoch": 0.7305576223761567, "grad_norm": 2.6394147872924805, "learning_rate": 1.7858019434586932e-06, "loss": 0.5135, "step": 17428 }, { "epoch": 0.7305995409911656, "grad_norm": 2.446592330932617, "learning_rate": 1.7852819716796465e-06, "loss": 0.4835, "step": 17429 }, { "epoch": 0.7306414596061747, "grad_norm": 1.6956183910369873, "learning_rate": 1.7847620591594823e-06, "loss": 0.468, "step": 17430 }, { "epoch": 0.7306833782211836, "grad_norm": 1.6668729782104492, "learning_rate": 1.7842422059077829e-06, "loss": 0.4619, "step": 17431 }, { "epoch": 0.7307252968361926, "grad_norm": 2.392542839050293, "learning_rate": 1.783722411934133e-06, "loss": 0.4732, "step": 17432 }, { "epoch": 0.7307672154512015, "grad_norm": 4.11683988571167, "learning_rate": 1.7832026772481154e-06, "loss": 0.4358, "step": 17433 }, { "epoch": 0.7308091340662104, "grad_norm": 1.8832981586456299, "learning_rate": 1.7826830018593073e-06, "loss": 0.5118, "step": 17434 }, { "epoch": 0.7308510526812194, "grad_norm": 1.704691767692566, "learning_rate": 1.7821633857772907e-06, "loss": 0.4743, "step": 17435 }, { "epoch": 0.7308929712962283, "grad_norm": 3.1184732913970947, "learning_rate": 1.7816438290116455e-06, "loss": 0.4634, "step": 17436 }, { "epoch": 0.7309348899112373, "grad_norm": 2.1489861011505127, "learning_rate": 1.781124331571945e-06, "loss": 0.4653, "step": 17437 }, { "epoch": 0.7309768085262462, "grad_norm": 1.8620340824127197, "learning_rate": 1.7806048934677695e-06, "loss": 0.4816, "step": 17438 }, { "epoch": 0.7310187271412553, "grad_norm": 1.7227914333343506, "learning_rate": 1.7800855147086916e-06, "loss": 0.4746, "step": 17439 }, { "epoch": 0.7310606457562642, "grad_norm": 1.6011765003204346, "learning_rate": 1.7795661953042847e-06, "loss": 0.4492, "step": 17440 }, { "epoch": 0.7311025643712732, "grad_norm": 1.8086233139038086, "learning_rate": 1.7790469352641226e-06, "loss": 0.4751, "step": 17441 }, { "epoch": 0.7311444829862821, "grad_norm": 2.899554491043091, "learning_rate": 1.7785277345977791e-06, "loss": 0.5288, "step": 17442 }, { "epoch": 0.7311864016012911, "grad_norm": 2.4055392742156982, "learning_rate": 1.7780085933148217e-06, "loss": 0.4803, "step": 17443 }, { "epoch": 0.7312283202163, "grad_norm": 1.9991798400878906, "learning_rate": 1.7774895114248214e-06, "loss": 0.516, "step": 17444 }, { "epoch": 0.731270238831309, "grad_norm": 1.8287909030914307, "learning_rate": 1.776970488937349e-06, "loss": 0.4982, "step": 17445 }, { "epoch": 0.731312157446318, "grad_norm": 2.976149082183838, "learning_rate": 1.776451525861968e-06, "loss": 0.4793, "step": 17446 }, { "epoch": 0.731354076061327, "grad_norm": 1.8454270362854004, "learning_rate": 1.775932622208248e-06, "loss": 0.4377, "step": 17447 }, { "epoch": 0.7313959946763359, "grad_norm": 1.8334619998931885, "learning_rate": 1.7754137779857516e-06, "loss": 0.5156, "step": 17448 }, { "epoch": 0.7314379132913449, "grad_norm": 1.7524334192276, "learning_rate": 1.7748949932040465e-06, "loss": 0.4338, "step": 17449 }, { "epoch": 0.7314798319063538, "grad_norm": 2.066155433654785, "learning_rate": 1.7743762678726916e-06, "loss": 0.4735, "step": 17450 }, { "epoch": 0.7315217505213628, "grad_norm": 2.494713068008423, "learning_rate": 1.7738576020012528e-06, "loss": 0.4713, "step": 17451 }, { "epoch": 0.7315636691363717, "grad_norm": 2.010904550552368, "learning_rate": 1.773338995599287e-06, "loss": 0.5473, "step": 17452 }, { "epoch": 0.7316055877513807, "grad_norm": 2.1232309341430664, "learning_rate": 1.7728204486763578e-06, "loss": 0.476, "step": 17453 }, { "epoch": 0.7316475063663896, "grad_norm": 2.0807132720947266, "learning_rate": 1.7723019612420206e-06, "loss": 0.513, "step": 17454 }, { "epoch": 0.7316894249813987, "grad_norm": 1.6880335807800293, "learning_rate": 1.7717835333058348e-06, "loss": 0.402, "step": 17455 }, { "epoch": 0.7317313435964076, "grad_norm": 1.7288001775741577, "learning_rate": 1.7712651648773581e-06, "loss": 0.4306, "step": 17456 }, { "epoch": 0.7317732622114166, "grad_norm": 1.6814099550247192, "learning_rate": 1.7707468559661428e-06, "loss": 0.4612, "step": 17457 }, { "epoch": 0.7318151808264255, "grad_norm": 2.189866542816162, "learning_rate": 1.770228606581747e-06, "loss": 0.4955, "step": 17458 }, { "epoch": 0.7318570994414344, "grad_norm": 1.7959390878677368, "learning_rate": 1.7697104167337214e-06, "loss": 0.4506, "step": 17459 }, { "epoch": 0.7318990180564434, "grad_norm": 1.8621116876602173, "learning_rate": 1.769192286431617e-06, "loss": 0.4869, "step": 17460 }, { "epoch": 0.7319409366714523, "grad_norm": 1.6715244054794312, "learning_rate": 1.7686742156849863e-06, "loss": 0.4533, "step": 17461 }, { "epoch": 0.7319828552864613, "grad_norm": 1.9339324235916138, "learning_rate": 1.7681562045033807e-06, "loss": 0.4736, "step": 17462 }, { "epoch": 0.7320247739014702, "grad_norm": 3.252393960952759, "learning_rate": 1.7676382528963464e-06, "loss": 0.4591, "step": 17463 }, { "epoch": 0.7320666925164793, "grad_norm": 1.7992864847183228, "learning_rate": 1.7671203608734317e-06, "loss": 0.4922, "step": 17464 }, { "epoch": 0.7321086111314882, "grad_norm": 1.831546664237976, "learning_rate": 1.7666025284441862e-06, "loss": 0.4686, "step": 17465 }, { "epoch": 0.7321505297464972, "grad_norm": 2.2566561698913574, "learning_rate": 1.7660847556181509e-06, "loss": 0.477, "step": 17466 }, { "epoch": 0.7321924483615061, "grad_norm": 2.290818929672241, "learning_rate": 1.7655670424048743e-06, "loss": 0.5181, "step": 17467 }, { "epoch": 0.7322343669765151, "grad_norm": 2.0056724548339844, "learning_rate": 1.7650493888138975e-06, "loss": 0.4739, "step": 17468 }, { "epoch": 0.732276285591524, "grad_norm": 1.8039205074310303, "learning_rate": 1.7645317948547614e-06, "loss": 0.5711, "step": 17469 }, { "epoch": 0.732318204206533, "grad_norm": 1.9783624410629272, "learning_rate": 1.7640142605370092e-06, "loss": 0.526, "step": 17470 }, { "epoch": 0.732360122821542, "grad_norm": 2.302844285964966, "learning_rate": 1.7634967858701817e-06, "loss": 0.4177, "step": 17471 }, { "epoch": 0.732402041436551, "grad_norm": 2.0039920806884766, "learning_rate": 1.7629793708638148e-06, "loss": 0.4945, "step": 17472 }, { "epoch": 0.7324439600515599, "grad_norm": 2.3159756660461426, "learning_rate": 1.7624620155274485e-06, "loss": 0.4902, "step": 17473 }, { "epoch": 0.7324858786665689, "grad_norm": 2.199751853942871, "learning_rate": 1.7619447198706213e-06, "loss": 0.4342, "step": 17474 }, { "epoch": 0.7325277972815778, "grad_norm": 1.777522325515747, "learning_rate": 1.7614274839028645e-06, "loss": 0.4347, "step": 17475 }, { "epoch": 0.7325697158965868, "grad_norm": 1.899364948272705, "learning_rate": 1.7609103076337163e-06, "loss": 0.4529, "step": 17476 }, { "epoch": 0.7326116345115957, "grad_norm": 2.3028619289398193, "learning_rate": 1.760393191072709e-06, "loss": 0.5281, "step": 17477 }, { "epoch": 0.7326535531266047, "grad_norm": 1.608788251876831, "learning_rate": 1.759876134229373e-06, "loss": 0.4607, "step": 17478 }, { "epoch": 0.7326954717416136, "grad_norm": 2.411220073699951, "learning_rate": 1.7593591371132406e-06, "loss": 0.4528, "step": 17479 }, { "epoch": 0.7327373903566227, "grad_norm": 1.8777967691421509, "learning_rate": 1.758842199733845e-06, "loss": 0.5042, "step": 17480 }, { "epoch": 0.7327793089716316, "grad_norm": 3.4121081829071045, "learning_rate": 1.7583253221007102e-06, "loss": 0.5042, "step": 17481 }, { "epoch": 0.7328212275866406, "grad_norm": 1.6654797792434692, "learning_rate": 1.7578085042233673e-06, "loss": 0.4272, "step": 17482 }, { "epoch": 0.7328631462016495, "grad_norm": 2.0724644660949707, "learning_rate": 1.7572917461113438e-06, "loss": 0.5013, "step": 17483 }, { "epoch": 0.7329050648166584, "grad_norm": 2.0156397819519043, "learning_rate": 1.756775047774162e-06, "loss": 0.4708, "step": 17484 }, { "epoch": 0.7329469834316674, "grad_norm": 3.3895113468170166, "learning_rate": 1.756258409221351e-06, "loss": 0.4868, "step": 17485 }, { "epoch": 0.7329889020466763, "grad_norm": 1.9344098567962646, "learning_rate": 1.7557418304624314e-06, "loss": 0.4265, "step": 17486 }, { "epoch": 0.7330308206616853, "grad_norm": 1.623752236366272, "learning_rate": 1.755225311506924e-06, "loss": 0.4953, "step": 17487 }, { "epoch": 0.7330727392766943, "grad_norm": 1.872359037399292, "learning_rate": 1.754708852364353e-06, "loss": 0.4763, "step": 17488 }, { "epoch": 0.7331146578917033, "grad_norm": 2.44708251953125, "learning_rate": 1.754192453044239e-06, "loss": 0.4709, "step": 17489 }, { "epoch": 0.7331565765067122, "grad_norm": 2.418193817138672, "learning_rate": 1.7536761135560986e-06, "loss": 0.475, "step": 17490 }, { "epoch": 0.7331984951217212, "grad_norm": 1.9708729982376099, "learning_rate": 1.7531598339094507e-06, "loss": 0.5212, "step": 17491 }, { "epoch": 0.7332404137367301, "grad_norm": 1.8811832666397095, "learning_rate": 1.7526436141138147e-06, "loss": 0.45, "step": 17492 }, { "epoch": 0.7332823323517391, "grad_norm": 1.8438786268234253, "learning_rate": 1.7521274541787025e-06, "loss": 0.4818, "step": 17493 }, { "epoch": 0.733324250966748, "grad_norm": 1.926062822341919, "learning_rate": 1.7516113541136326e-06, "loss": 0.5239, "step": 17494 }, { "epoch": 0.733366169581757, "grad_norm": 1.812072992324829, "learning_rate": 1.7510953139281144e-06, "loss": 0.4341, "step": 17495 }, { "epoch": 0.733408088196766, "grad_norm": 1.950877070426941, "learning_rate": 1.7505793336316652e-06, "loss": 0.4629, "step": 17496 }, { "epoch": 0.733450006811775, "grad_norm": 1.802384853363037, "learning_rate": 1.7500634132337912e-06, "loss": 0.4717, "step": 17497 }, { "epoch": 0.7334919254267839, "grad_norm": 1.8372962474822998, "learning_rate": 1.7495475527440075e-06, "loss": 0.5509, "step": 17498 }, { "epoch": 0.7335338440417929, "grad_norm": 1.8666715621948242, "learning_rate": 1.7490317521718192e-06, "loss": 0.5151, "step": 17499 }, { "epoch": 0.7335757626568018, "grad_norm": 2.605708599090576, "learning_rate": 1.7485160115267357e-06, "loss": 0.5148, "step": 17500 }, { "epoch": 0.7336176812718108, "grad_norm": 2.5525074005126953, "learning_rate": 1.7480003308182668e-06, "loss": 0.4725, "step": 17501 }, { "epoch": 0.7336595998868197, "grad_norm": 2.1060214042663574, "learning_rate": 1.7474847100559146e-06, "loss": 0.4719, "step": 17502 }, { "epoch": 0.7337015185018287, "grad_norm": 1.7450056076049805, "learning_rate": 1.7469691492491848e-06, "loss": 0.4467, "step": 17503 }, { "epoch": 0.7337434371168376, "grad_norm": 2.4275288581848145, "learning_rate": 1.7464536484075833e-06, "loss": 0.4894, "step": 17504 }, { "epoch": 0.7337853557318467, "grad_norm": 2.1293797492980957, "learning_rate": 1.7459382075406112e-06, "loss": 0.4347, "step": 17505 }, { "epoch": 0.7338272743468556, "grad_norm": 2.0580685138702393, "learning_rate": 1.745422826657767e-06, "loss": 0.4737, "step": 17506 }, { "epoch": 0.7338691929618646, "grad_norm": 1.898400902748108, "learning_rate": 1.7449075057685566e-06, "loss": 0.4788, "step": 17507 }, { "epoch": 0.7339111115768735, "grad_norm": 2.183722496032715, "learning_rate": 1.744392244882474e-06, "loss": 0.4895, "step": 17508 }, { "epoch": 0.7339530301918824, "grad_norm": 1.9400153160095215, "learning_rate": 1.7438770440090197e-06, "loss": 0.4087, "step": 17509 }, { "epoch": 0.7339949488068914, "grad_norm": 1.949691891670227, "learning_rate": 1.743361903157692e-06, "loss": 0.5354, "step": 17510 }, { "epoch": 0.7340368674219003, "grad_norm": 1.7080928087234497, "learning_rate": 1.7428468223379847e-06, "loss": 0.4647, "step": 17511 }, { "epoch": 0.7340787860369093, "grad_norm": 1.6488455533981323, "learning_rate": 1.7423318015593927e-06, "loss": 0.4884, "step": 17512 }, { "epoch": 0.7341207046519183, "grad_norm": 2.8736674785614014, "learning_rate": 1.7418168408314118e-06, "loss": 0.4923, "step": 17513 }, { "epoch": 0.7341626232669273, "grad_norm": 1.7874009609222412, "learning_rate": 1.7413019401635335e-06, "loss": 0.4341, "step": 17514 }, { "epoch": 0.7342045418819362, "grad_norm": 1.7931642532348633, "learning_rate": 1.7407870995652471e-06, "loss": 0.4623, "step": 17515 }, { "epoch": 0.7342464604969452, "grad_norm": 2.644226551055908, "learning_rate": 1.7402723190460464e-06, "loss": 0.4996, "step": 17516 }, { "epoch": 0.7342883791119541, "grad_norm": 1.9505889415740967, "learning_rate": 1.7397575986154175e-06, "loss": 0.5108, "step": 17517 }, { "epoch": 0.7343302977269631, "grad_norm": 2.1115341186523438, "learning_rate": 1.7392429382828503e-06, "loss": 0.573, "step": 17518 }, { "epoch": 0.734372216341972, "grad_norm": 1.8225480318069458, "learning_rate": 1.738728338057833e-06, "loss": 0.5427, "step": 17519 }, { "epoch": 0.734414134956981, "grad_norm": 1.644424557685852, "learning_rate": 1.7382137979498487e-06, "loss": 0.4202, "step": 17520 }, { "epoch": 0.73445605357199, "grad_norm": 1.6455435752868652, "learning_rate": 1.7376993179683838e-06, "loss": 0.4589, "step": 17521 }, { "epoch": 0.734497972186999, "grad_norm": 2.3298914432525635, "learning_rate": 1.7371848981229232e-06, "loss": 0.4647, "step": 17522 }, { "epoch": 0.7345398908020079, "grad_norm": 2.402024745941162, "learning_rate": 1.7366705384229493e-06, "loss": 0.5207, "step": 17523 }, { "epoch": 0.7345818094170169, "grad_norm": 2.0287539958953857, "learning_rate": 1.7361562388779402e-06, "loss": 0.5537, "step": 17524 }, { "epoch": 0.7346237280320258, "grad_norm": 1.8529856204986572, "learning_rate": 1.7356419994973807e-06, "loss": 0.4408, "step": 17525 }, { "epoch": 0.7346656466470348, "grad_norm": 2.3235113620758057, "learning_rate": 1.7351278202907462e-06, "loss": 0.4869, "step": 17526 }, { "epoch": 0.7347075652620437, "grad_norm": 2.464935541152954, "learning_rate": 1.7346137012675169e-06, "loss": 0.4927, "step": 17527 }, { "epoch": 0.7347494838770527, "grad_norm": 1.9626027345657349, "learning_rate": 1.7340996424371709e-06, "loss": 0.5308, "step": 17528 }, { "epoch": 0.7347914024920617, "grad_norm": 2.013936758041382, "learning_rate": 1.733585643809182e-06, "loss": 0.464, "step": 17529 }, { "epoch": 0.7348333211070707, "grad_norm": 3.2816569805145264, "learning_rate": 1.7330717053930275e-06, "loss": 0.4865, "step": 17530 }, { "epoch": 0.7348752397220796, "grad_norm": 1.8508244752883911, "learning_rate": 1.732557827198178e-06, "loss": 0.4897, "step": 17531 }, { "epoch": 0.7349171583370886, "grad_norm": 1.4757130146026611, "learning_rate": 1.7320440092341095e-06, "loss": 0.4329, "step": 17532 }, { "epoch": 0.7349590769520975, "grad_norm": 1.8841201066970825, "learning_rate": 1.7315302515102923e-06, "loss": 0.4813, "step": 17533 }, { "epoch": 0.7350009955671064, "grad_norm": 2.1582000255584717, "learning_rate": 1.7310165540361945e-06, "loss": 0.4991, "step": 17534 }, { "epoch": 0.7350429141821154, "grad_norm": 1.7265210151672363, "learning_rate": 1.7305029168212871e-06, "loss": 0.4629, "step": 17535 }, { "epoch": 0.7350848327971243, "grad_norm": 1.907392144203186, "learning_rate": 1.72998933987504e-06, "loss": 0.4759, "step": 17536 }, { "epoch": 0.7351267514121334, "grad_norm": 2.234945058822632, "learning_rate": 1.7294758232069174e-06, "loss": 0.4984, "step": 17537 }, { "epoch": 0.7351686700271423, "grad_norm": 1.612807273864746, "learning_rate": 1.7289623668263867e-06, "loss": 0.4541, "step": 17538 }, { "epoch": 0.7352105886421513, "grad_norm": 1.9687025547027588, "learning_rate": 1.7284489707429142e-06, "loss": 0.5191, "step": 17539 }, { "epoch": 0.7352525072571602, "grad_norm": 1.9584276676177979, "learning_rate": 1.7279356349659605e-06, "loss": 0.4626, "step": 17540 }, { "epoch": 0.7352944258721692, "grad_norm": 1.816489815711975, "learning_rate": 1.7274223595049916e-06, "loss": 0.4819, "step": 17541 }, { "epoch": 0.7353363444871781, "grad_norm": 1.8296277523040771, "learning_rate": 1.7269091443694657e-06, "loss": 0.4408, "step": 17542 }, { "epoch": 0.7353782631021871, "grad_norm": 1.7966859340667725, "learning_rate": 1.7263959895688464e-06, "loss": 0.45, "step": 17543 }, { "epoch": 0.735420181717196, "grad_norm": 1.7699335813522339, "learning_rate": 1.72588289511259e-06, "loss": 0.4689, "step": 17544 }, { "epoch": 0.735462100332205, "grad_norm": 1.934822916984558, "learning_rate": 1.725369861010157e-06, "loss": 0.4861, "step": 17545 }, { "epoch": 0.735504018947214, "grad_norm": 2.2457127571105957, "learning_rate": 1.7248568872710025e-06, "loss": 0.5025, "step": 17546 }, { "epoch": 0.735545937562223, "grad_norm": 6.71703577041626, "learning_rate": 1.7243439739045837e-06, "loss": 0.4716, "step": 17547 }, { "epoch": 0.7355878561772319, "grad_norm": 1.8787933588027954, "learning_rate": 1.7238311209203567e-06, "loss": 0.4682, "step": 17548 }, { "epoch": 0.7356297747922409, "grad_norm": 1.9321726560592651, "learning_rate": 1.7233183283277722e-06, "loss": 0.4779, "step": 17549 }, { "epoch": 0.7356716934072498, "grad_norm": 2.033653736114502, "learning_rate": 1.7228055961362844e-06, "loss": 0.4501, "step": 17550 }, { "epoch": 0.7357136120222588, "grad_norm": 1.8546392917633057, "learning_rate": 1.722292924355346e-06, "loss": 0.5162, "step": 17551 }, { "epoch": 0.7357555306372677, "grad_norm": 1.7471882104873657, "learning_rate": 1.7217803129944073e-06, "loss": 0.5098, "step": 17552 }, { "epoch": 0.7357974492522767, "grad_norm": 2.1737685203552246, "learning_rate": 1.7212677620629137e-06, "loss": 0.4323, "step": 17553 }, { "epoch": 0.7358393678672857, "grad_norm": 2.4077718257904053, "learning_rate": 1.7207552715703185e-06, "loss": 0.4228, "step": 17554 }, { "epoch": 0.7358812864822947, "grad_norm": 1.695142388343811, "learning_rate": 1.7202428415260648e-06, "loss": 0.5091, "step": 17555 }, { "epoch": 0.7359232050973036, "grad_norm": 1.7439045906066895, "learning_rate": 1.7197304719396002e-06, "loss": 0.5265, "step": 17556 }, { "epoch": 0.7359651237123126, "grad_norm": 1.7190264463424683, "learning_rate": 1.7192181628203713e-06, "loss": 0.4748, "step": 17557 }, { "epoch": 0.7360070423273215, "grad_norm": 1.7342818975448608, "learning_rate": 1.7187059141778179e-06, "loss": 0.4834, "step": 17558 }, { "epoch": 0.7360489609423304, "grad_norm": 1.9869579076766968, "learning_rate": 1.7181937260213849e-06, "loss": 0.4749, "step": 17559 }, { "epoch": 0.7360908795573394, "grad_norm": 2.1682724952697754, "learning_rate": 1.7176815983605155e-06, "loss": 0.5077, "step": 17560 }, { "epoch": 0.7361327981723483, "grad_norm": 4.051933288574219, "learning_rate": 1.7171695312046482e-06, "loss": 0.4523, "step": 17561 }, { "epoch": 0.7361747167873574, "grad_norm": 2.0361242294311523, "learning_rate": 1.71665752456322e-06, "loss": 0.4269, "step": 17562 }, { "epoch": 0.7362166354023663, "grad_norm": 4.241717338562012, "learning_rate": 1.7161455784456737e-06, "loss": 0.4422, "step": 17563 }, { "epoch": 0.7362585540173753, "grad_norm": 1.888037919998169, "learning_rate": 1.7156336928614414e-06, "loss": 0.4757, "step": 17564 }, { "epoch": 0.7363004726323842, "grad_norm": 2.5953774452209473, "learning_rate": 1.7151218678199621e-06, "loss": 0.4916, "step": 17565 }, { "epoch": 0.7363423912473932, "grad_norm": 1.6394606828689575, "learning_rate": 1.714610103330671e-06, "loss": 0.4649, "step": 17566 }, { "epoch": 0.7363843098624021, "grad_norm": 1.6511261463165283, "learning_rate": 1.7140983994029997e-06, "loss": 0.5202, "step": 17567 }, { "epoch": 0.7364262284774111, "grad_norm": 1.8133914470672607, "learning_rate": 1.7135867560463814e-06, "loss": 0.4485, "step": 17568 }, { "epoch": 0.73646814709242, "grad_norm": 2.0216939449310303, "learning_rate": 1.7130751732702499e-06, "loss": 0.4616, "step": 17569 }, { "epoch": 0.736510065707429, "grad_norm": 1.6788547039031982, "learning_rate": 1.7125636510840332e-06, "loss": 0.4521, "step": 17570 }, { "epoch": 0.736551984322438, "grad_norm": 1.7858688831329346, "learning_rate": 1.7120521894971592e-06, "loss": 0.4233, "step": 17571 }, { "epoch": 0.736593902937447, "grad_norm": 3.4939627647399902, "learning_rate": 1.7115407885190599e-06, "loss": 0.4857, "step": 17572 }, { "epoch": 0.7366358215524559, "grad_norm": 1.6336158514022827, "learning_rate": 1.7110294481591578e-06, "loss": 0.506, "step": 17573 }, { "epoch": 0.7366777401674649, "grad_norm": 1.8252860307693481, "learning_rate": 1.7105181684268813e-06, "loss": 0.4966, "step": 17574 }, { "epoch": 0.7367196587824738, "grad_norm": 2.218226432800293, "learning_rate": 1.7100069493316556e-06, "loss": 0.4581, "step": 17575 }, { "epoch": 0.7367615773974828, "grad_norm": 1.7624506950378418, "learning_rate": 1.709495790882903e-06, "loss": 0.4838, "step": 17576 }, { "epoch": 0.7368034960124917, "grad_norm": 1.8790653944015503, "learning_rate": 1.7089846930900455e-06, "loss": 0.4605, "step": 17577 }, { "epoch": 0.7368454146275007, "grad_norm": 1.8044612407684326, "learning_rate": 1.7084736559625075e-06, "loss": 0.4403, "step": 17578 }, { "epoch": 0.7368873332425097, "grad_norm": 2.37113094329834, "learning_rate": 1.7079626795097076e-06, "loss": 0.4815, "step": 17579 }, { "epoch": 0.7369292518575187, "grad_norm": 2.1172540187835693, "learning_rate": 1.7074517637410626e-06, "loss": 0.4226, "step": 17580 }, { "epoch": 0.7369711704725276, "grad_norm": 1.6962133646011353, "learning_rate": 1.7069409086659943e-06, "loss": 0.4606, "step": 17581 }, { "epoch": 0.7370130890875366, "grad_norm": 1.7233699560165405, "learning_rate": 1.706430114293916e-06, "loss": 0.4931, "step": 17582 }, { "epoch": 0.7370550077025455, "grad_norm": 1.6899454593658447, "learning_rate": 1.7059193806342456e-06, "loss": 0.4408, "step": 17583 }, { "epoch": 0.7370969263175544, "grad_norm": 2.038008213043213, "learning_rate": 1.7054087076963993e-06, "loss": 0.4307, "step": 17584 }, { "epoch": 0.7371388449325634, "grad_norm": 1.8116700649261475, "learning_rate": 1.7048980954897864e-06, "loss": 0.4291, "step": 17585 }, { "epoch": 0.7371807635475723, "grad_norm": 1.8658983707427979, "learning_rate": 1.7043875440238229e-06, "loss": 0.501, "step": 17586 }, { "epoch": 0.7372226821625814, "grad_norm": 1.6443605422973633, "learning_rate": 1.7038770533079195e-06, "loss": 0.4469, "step": 17587 }, { "epoch": 0.7372646007775903, "grad_norm": 1.9202699661254883, "learning_rate": 1.7033666233514867e-06, "loss": 0.5074, "step": 17588 }, { "epoch": 0.7373065193925993, "grad_norm": 3.7193119525909424, "learning_rate": 1.7028562541639305e-06, "loss": 0.4546, "step": 17589 }, { "epoch": 0.7373484380076082, "grad_norm": 1.6692166328430176, "learning_rate": 1.7023459457546633e-06, "loss": 0.4577, "step": 17590 }, { "epoch": 0.7373903566226172, "grad_norm": 1.7673263549804688, "learning_rate": 1.7018356981330874e-06, "loss": 0.4846, "step": 17591 }, { "epoch": 0.7374322752376261, "grad_norm": 1.840797781944275, "learning_rate": 1.7013255113086114e-06, "loss": 0.4976, "step": 17592 }, { "epoch": 0.7374741938526351, "grad_norm": 1.8312668800354004, "learning_rate": 1.7008153852906406e-06, "loss": 0.5178, "step": 17593 }, { "epoch": 0.737516112467644, "grad_norm": 1.7560955286026, "learning_rate": 1.700305320088576e-06, "loss": 0.4996, "step": 17594 }, { "epoch": 0.737558031082653, "grad_norm": 1.8859782218933105, "learning_rate": 1.6997953157118207e-06, "loss": 0.5122, "step": 17595 }, { "epoch": 0.737599949697662, "grad_norm": 1.7521770000457764, "learning_rate": 1.6992853721697783e-06, "loss": 0.4422, "step": 17596 }, { "epoch": 0.737641868312671, "grad_norm": 1.9565708637237549, "learning_rate": 1.698775489471845e-06, "loss": 0.4963, "step": 17597 }, { "epoch": 0.7376837869276799, "grad_norm": 2.103330135345459, "learning_rate": 1.6982656676274233e-06, "loss": 0.5078, "step": 17598 }, { "epoch": 0.7377257055426889, "grad_norm": 1.7811133861541748, "learning_rate": 1.6977559066459098e-06, "loss": 0.458, "step": 17599 }, { "epoch": 0.7377676241576978, "grad_norm": 2.114593982696533, "learning_rate": 1.6972462065366995e-06, "loss": 0.5029, "step": 17600 }, { "epoch": 0.7378095427727068, "grad_norm": 2.9082887172698975, "learning_rate": 1.696736567309189e-06, "loss": 0.4967, "step": 17601 }, { "epoch": 0.7378514613877157, "grad_norm": 2.018038034439087, "learning_rate": 1.696226988972775e-06, "loss": 0.4486, "step": 17602 }, { "epoch": 0.7378933800027248, "grad_norm": 1.9020198583602905, "learning_rate": 1.6957174715368479e-06, "loss": 0.5318, "step": 17603 }, { "epoch": 0.7379352986177337, "grad_norm": 2.093414068222046, "learning_rate": 1.6952080150108025e-06, "loss": 0.4618, "step": 17604 }, { "epoch": 0.7379772172327427, "grad_norm": 1.8853791952133179, "learning_rate": 1.6946986194040271e-06, "loss": 0.4988, "step": 17605 }, { "epoch": 0.7380191358477516, "grad_norm": 1.9367057085037231, "learning_rate": 1.6941892847259128e-06, "loss": 0.4467, "step": 17606 }, { "epoch": 0.7380610544627606, "grad_norm": 1.6816924810409546, "learning_rate": 1.693680010985851e-06, "loss": 0.4606, "step": 17607 }, { "epoch": 0.7381029730777695, "grad_norm": 1.8082592487335205, "learning_rate": 1.6931707981932272e-06, "loss": 0.4647, "step": 17608 }, { "epoch": 0.7381448916927784, "grad_norm": 2.121196746826172, "learning_rate": 1.6926616463574264e-06, "loss": 0.4846, "step": 17609 }, { "epoch": 0.7381868103077874, "grad_norm": 1.6256968975067139, "learning_rate": 1.6921525554878376e-06, "loss": 0.4453, "step": 17610 }, { "epoch": 0.7382287289227963, "grad_norm": 1.9955241680145264, "learning_rate": 1.6916435255938417e-06, "loss": 0.4586, "step": 17611 }, { "epoch": 0.7382706475378054, "grad_norm": 1.7786576747894287, "learning_rate": 1.6911345566848242e-06, "loss": 0.4267, "step": 17612 }, { "epoch": 0.7383125661528143, "grad_norm": 2.452322483062744, "learning_rate": 1.6906256487701683e-06, "loss": 0.464, "step": 17613 }, { "epoch": 0.7383544847678233, "grad_norm": 1.7658706903457642, "learning_rate": 1.6901168018592518e-06, "loss": 0.4701, "step": 17614 }, { "epoch": 0.7383964033828322, "grad_norm": 1.9648653268814087, "learning_rate": 1.689608015961456e-06, "loss": 0.4922, "step": 17615 }, { "epoch": 0.7384383219978412, "grad_norm": 1.881316065788269, "learning_rate": 1.6890992910861614e-06, "loss": 0.4142, "step": 17616 }, { "epoch": 0.7384802406128501, "grad_norm": 3.205920457839966, "learning_rate": 1.6885906272427448e-06, "loss": 0.4998, "step": 17617 }, { "epoch": 0.7385221592278591, "grad_norm": 1.7734293937683105, "learning_rate": 1.6880820244405804e-06, "loss": 0.4365, "step": 17618 }, { "epoch": 0.738564077842868, "grad_norm": 1.7087879180908203, "learning_rate": 1.6875734826890466e-06, "loss": 0.4714, "step": 17619 }, { "epoch": 0.7386059964578771, "grad_norm": 2.3062984943389893, "learning_rate": 1.6870650019975148e-06, "loss": 0.4949, "step": 17620 }, { "epoch": 0.738647915072886, "grad_norm": 1.7437515258789062, "learning_rate": 1.6865565823753598e-06, "loss": 0.4619, "step": 17621 }, { "epoch": 0.738689833687895, "grad_norm": 1.8660188913345337, "learning_rate": 1.6860482238319553e-06, "loss": 0.4797, "step": 17622 }, { "epoch": 0.7387317523029039, "grad_norm": 2.506758689880371, "learning_rate": 1.6855399263766681e-06, "loss": 0.495, "step": 17623 }, { "epoch": 0.7387736709179129, "grad_norm": 2.2129967212677, "learning_rate": 1.6850316900188708e-06, "loss": 0.5448, "step": 17624 }, { "epoch": 0.7388155895329218, "grad_norm": 1.5356024503707886, "learning_rate": 1.6845235147679335e-06, "loss": 0.4149, "step": 17625 }, { "epoch": 0.7388575081479308, "grad_norm": 2.4524741172790527, "learning_rate": 1.684015400633221e-06, "loss": 0.4337, "step": 17626 }, { "epoch": 0.7388994267629397, "grad_norm": 1.858046054840088, "learning_rate": 1.6835073476240987e-06, "loss": 0.4996, "step": 17627 }, { "epoch": 0.7389413453779488, "grad_norm": 2.1302685737609863, "learning_rate": 1.6829993557499358e-06, "loss": 0.5092, "step": 17628 }, { "epoch": 0.7389832639929577, "grad_norm": 2.3915576934814453, "learning_rate": 1.682491425020092e-06, "loss": 0.4915, "step": 17629 }, { "epoch": 0.7390251826079667, "grad_norm": 1.79460608959198, "learning_rate": 1.6819835554439324e-06, "loss": 0.469, "step": 17630 }, { "epoch": 0.7390671012229756, "grad_norm": 2.0426011085510254, "learning_rate": 1.6814757470308207e-06, "loss": 0.5103, "step": 17631 }, { "epoch": 0.7391090198379846, "grad_norm": 1.7761582136154175, "learning_rate": 1.6809679997901146e-06, "loss": 0.4733, "step": 17632 }, { "epoch": 0.7391509384529935, "grad_norm": 1.9638309478759766, "learning_rate": 1.6804603137311747e-06, "loss": 0.5291, "step": 17633 }, { "epoch": 0.7391928570680024, "grad_norm": 1.6995445489883423, "learning_rate": 1.6799526888633615e-06, "loss": 0.4313, "step": 17634 }, { "epoch": 0.7392347756830114, "grad_norm": 1.9245458841323853, "learning_rate": 1.6794451251960291e-06, "loss": 0.5716, "step": 17635 }, { "epoch": 0.7392766942980203, "grad_norm": 3.2656614780426025, "learning_rate": 1.6789376227385367e-06, "loss": 0.4448, "step": 17636 }, { "epoch": 0.7393186129130294, "grad_norm": 1.8282004594802856, "learning_rate": 1.678430181500238e-06, "loss": 0.4478, "step": 17637 }, { "epoch": 0.7393605315280383, "grad_norm": 2.1974551677703857, "learning_rate": 1.6779228014904853e-06, "loss": 0.4375, "step": 17638 }, { "epoch": 0.7394024501430473, "grad_norm": 1.8824540376663208, "learning_rate": 1.6774154827186334e-06, "loss": 0.4769, "step": 17639 }, { "epoch": 0.7394443687580562, "grad_norm": 1.5980629920959473, "learning_rate": 1.6769082251940354e-06, "loss": 0.4467, "step": 17640 }, { "epoch": 0.7394862873730652, "grad_norm": 1.9482301473617554, "learning_rate": 1.6764010289260385e-06, "loss": 0.4777, "step": 17641 }, { "epoch": 0.7395282059880741, "grad_norm": 1.6365329027175903, "learning_rate": 1.6758938939239938e-06, "loss": 0.4338, "step": 17642 }, { "epoch": 0.7395701246030831, "grad_norm": 2.3618218898773193, "learning_rate": 1.6753868201972512e-06, "loss": 0.4511, "step": 17643 }, { "epoch": 0.739612043218092, "grad_norm": 1.8598517179489136, "learning_rate": 1.6748798077551548e-06, "loss": 0.4582, "step": 17644 }, { "epoch": 0.7396539618331011, "grad_norm": 1.8281326293945312, "learning_rate": 1.6743728566070543e-06, "loss": 0.4986, "step": 17645 }, { "epoch": 0.73969588044811, "grad_norm": 2.254594564437866, "learning_rate": 1.6738659667622926e-06, "loss": 0.5221, "step": 17646 }, { "epoch": 0.739737799063119, "grad_norm": 1.6567541360855103, "learning_rate": 1.6733591382302116e-06, "loss": 0.5035, "step": 17647 }, { "epoch": 0.7397797176781279, "grad_norm": 1.9092698097229004, "learning_rate": 1.672852371020156e-06, "loss": 0.5053, "step": 17648 }, { "epoch": 0.7398216362931369, "grad_norm": 1.7416576147079468, "learning_rate": 1.6723456651414693e-06, "loss": 0.428, "step": 17649 }, { "epoch": 0.7398635549081458, "grad_norm": 1.7828508615493774, "learning_rate": 1.6718390206034878e-06, "loss": 0.445, "step": 17650 }, { "epoch": 0.7399054735231548, "grad_norm": 2.036890745162964, "learning_rate": 1.6713324374155531e-06, "loss": 0.5043, "step": 17651 }, { "epoch": 0.7399473921381637, "grad_norm": 1.8551691770553589, "learning_rate": 1.6708259155870049e-06, "loss": 0.4671, "step": 17652 }, { "epoch": 0.7399893107531728, "grad_norm": 1.4799102544784546, "learning_rate": 1.6703194551271772e-06, "loss": 0.4605, "step": 17653 }, { "epoch": 0.7400312293681817, "grad_norm": 1.59275484085083, "learning_rate": 1.6698130560454084e-06, "loss": 0.4416, "step": 17654 }, { "epoch": 0.7400731479831907, "grad_norm": 2.284489631652832, "learning_rate": 1.669306718351032e-06, "loss": 0.5112, "step": 17655 }, { "epoch": 0.7401150665981996, "grad_norm": 1.5660231113433838, "learning_rate": 1.6688004420533804e-06, "loss": 0.4787, "step": 17656 }, { "epoch": 0.7401569852132086, "grad_norm": 1.7565791606903076, "learning_rate": 1.668294227161787e-06, "loss": 0.4583, "step": 17657 }, { "epoch": 0.7401989038282175, "grad_norm": 1.5867124795913696, "learning_rate": 1.6677880736855857e-06, "loss": 0.4915, "step": 17658 }, { "epoch": 0.7402408224432264, "grad_norm": 1.6352187395095825, "learning_rate": 1.6672819816341023e-06, "loss": 0.5138, "step": 17659 }, { "epoch": 0.7402827410582354, "grad_norm": 1.680163025856018, "learning_rate": 1.6667759510166687e-06, "loss": 0.4426, "step": 17660 }, { "epoch": 0.7403246596732443, "grad_norm": 1.9960778951644897, "learning_rate": 1.6662699818426143e-06, "loss": 0.5578, "step": 17661 }, { "epoch": 0.7403665782882534, "grad_norm": 1.7604405879974365, "learning_rate": 1.6657640741212616e-06, "loss": 0.49, "step": 17662 }, { "epoch": 0.7404084969032623, "grad_norm": 1.7072299718856812, "learning_rate": 1.665258227861941e-06, "loss": 0.4598, "step": 17663 }, { "epoch": 0.7404504155182713, "grad_norm": 1.6513245105743408, "learning_rate": 1.6647524430739747e-06, "loss": 0.4616, "step": 17664 }, { "epoch": 0.7404923341332802, "grad_norm": 2.0598037242889404, "learning_rate": 1.6642467197666845e-06, "loss": 0.4584, "step": 17665 }, { "epoch": 0.7405342527482892, "grad_norm": 1.6771496534347534, "learning_rate": 1.6637410579493946e-06, "loss": 0.4945, "step": 17666 }, { "epoch": 0.7405761713632981, "grad_norm": 2.6093850135803223, "learning_rate": 1.6632354576314281e-06, "loss": 0.4896, "step": 17667 }, { "epoch": 0.7406180899783071, "grad_norm": 1.8535120487213135, "learning_rate": 1.6627299188221008e-06, "loss": 0.4664, "step": 17668 }, { "epoch": 0.740660008593316, "grad_norm": 1.6470332145690918, "learning_rate": 1.6622244415307337e-06, "loss": 0.5361, "step": 17669 }, { "epoch": 0.7407019272083251, "grad_norm": 1.7528635263442993, "learning_rate": 1.6617190257666465e-06, "loss": 0.5147, "step": 17670 }, { "epoch": 0.740743845823334, "grad_norm": 2.1938323974609375, "learning_rate": 1.6612136715391526e-06, "loss": 0.465, "step": 17671 }, { "epoch": 0.740785764438343, "grad_norm": 1.740464210510254, "learning_rate": 1.6607083788575706e-06, "loss": 0.4843, "step": 17672 }, { "epoch": 0.7408276830533519, "grad_norm": 2.260637044906616, "learning_rate": 1.6602031477312131e-06, "loss": 0.4861, "step": 17673 }, { "epoch": 0.7408696016683609, "grad_norm": 1.6863043308258057, "learning_rate": 1.6596979781693918e-06, "loss": 0.4986, "step": 17674 }, { "epoch": 0.7409115202833698, "grad_norm": 1.8055486679077148, "learning_rate": 1.6591928701814204e-06, "loss": 0.5232, "step": 17675 }, { "epoch": 0.7409534388983788, "grad_norm": 2.033137798309326, "learning_rate": 1.6586878237766113e-06, "loss": 0.4678, "step": 17676 }, { "epoch": 0.7409953575133877, "grad_norm": 1.6163408756256104, "learning_rate": 1.6581828389642717e-06, "loss": 0.4659, "step": 17677 }, { "epoch": 0.7410372761283968, "grad_norm": 2.5253329277038574, "learning_rate": 1.6576779157537133e-06, "loss": 0.5015, "step": 17678 }, { "epoch": 0.7410791947434057, "grad_norm": 1.9571349620819092, "learning_rate": 1.6571730541542397e-06, "loss": 0.5264, "step": 17679 }, { "epoch": 0.7411211133584147, "grad_norm": 2.207008123397827, "learning_rate": 1.6566682541751594e-06, "loss": 0.4709, "step": 17680 }, { "epoch": 0.7411630319734236, "grad_norm": 1.729436993598938, "learning_rate": 1.6561635158257799e-06, "loss": 0.4902, "step": 17681 }, { "epoch": 0.7412049505884326, "grad_norm": 1.9272973537445068, "learning_rate": 1.6556588391154016e-06, "loss": 0.4932, "step": 17682 }, { "epoch": 0.7412468692034415, "grad_norm": 1.6147894859313965, "learning_rate": 1.655154224053331e-06, "loss": 0.45, "step": 17683 }, { "epoch": 0.7412887878184504, "grad_norm": 2.6721584796905518, "learning_rate": 1.6546496706488679e-06, "loss": 0.4634, "step": 17684 }, { "epoch": 0.7413307064334594, "grad_norm": 1.8879719972610474, "learning_rate": 1.6541451789113121e-06, "loss": 0.49, "step": 17685 }, { "epoch": 0.7413726250484683, "grad_norm": 1.7776967287063599, "learning_rate": 1.653640748849964e-06, "loss": 0.5889, "step": 17686 }, { "epoch": 0.7414145436634774, "grad_norm": 1.927202582359314, "learning_rate": 1.653136380474124e-06, "loss": 0.465, "step": 17687 }, { "epoch": 0.7414564622784863, "grad_norm": 1.9063643217086792, "learning_rate": 1.6526320737930867e-06, "loss": 0.5061, "step": 17688 }, { "epoch": 0.7414983808934953, "grad_norm": 1.6693620681762695, "learning_rate": 1.652127828816149e-06, "loss": 0.4143, "step": 17689 }, { "epoch": 0.7415402995085042, "grad_norm": 1.7072826623916626, "learning_rate": 1.6516236455526086e-06, "loss": 0.4745, "step": 17690 }, { "epoch": 0.7415822181235132, "grad_norm": 1.7730612754821777, "learning_rate": 1.6511195240117556e-06, "loss": 0.5081, "step": 17691 }, { "epoch": 0.7416241367385221, "grad_norm": 1.7997313737869263, "learning_rate": 1.6506154642028859e-06, "loss": 0.4582, "step": 17692 }, { "epoch": 0.7416660553535311, "grad_norm": 2.173414468765259, "learning_rate": 1.6501114661352902e-06, "loss": 0.4385, "step": 17693 }, { "epoch": 0.74170797396854, "grad_norm": 1.664766788482666, "learning_rate": 1.6496075298182562e-06, "loss": 0.4685, "step": 17694 }, { "epoch": 0.7417498925835491, "grad_norm": 1.7408512830734253, "learning_rate": 1.6491036552610761e-06, "loss": 0.5165, "step": 17695 }, { "epoch": 0.741791811198558, "grad_norm": 1.7548449039459229, "learning_rate": 1.6485998424730394e-06, "loss": 0.3995, "step": 17696 }, { "epoch": 0.741833729813567, "grad_norm": 1.70966637134552, "learning_rate": 1.6480960914634291e-06, "loss": 0.4745, "step": 17697 }, { "epoch": 0.7418756484285759, "grad_norm": 1.5478124618530273, "learning_rate": 1.647592402241534e-06, "loss": 0.4739, "step": 17698 }, { "epoch": 0.7419175670435849, "grad_norm": 2.0483760833740234, "learning_rate": 1.6470887748166404e-06, "loss": 0.4585, "step": 17699 }, { "epoch": 0.7419594856585938, "grad_norm": 1.7926892042160034, "learning_rate": 1.6465852091980277e-06, "loss": 0.5194, "step": 17700 }, { "epoch": 0.7420014042736028, "grad_norm": 1.7252159118652344, "learning_rate": 1.6460817053949823e-06, "loss": 0.4645, "step": 17701 }, { "epoch": 0.7420433228886117, "grad_norm": 2.808720827102661, "learning_rate": 1.6455782634167845e-06, "loss": 0.4728, "step": 17702 }, { "epoch": 0.7420852415036208, "grad_norm": 1.9853992462158203, "learning_rate": 1.645074883272712e-06, "loss": 0.4894, "step": 17703 }, { "epoch": 0.7421271601186297, "grad_norm": 1.871640920639038, "learning_rate": 1.6445715649720462e-06, "loss": 0.4031, "step": 17704 }, { "epoch": 0.7421690787336387, "grad_norm": 1.9012906551361084, "learning_rate": 1.6440683085240666e-06, "loss": 0.4692, "step": 17705 }, { "epoch": 0.7422109973486476, "grad_norm": 1.8801541328430176, "learning_rate": 1.6435651139380464e-06, "loss": 0.5064, "step": 17706 }, { "epoch": 0.7422529159636566, "grad_norm": 1.713871717453003, "learning_rate": 1.6430619812232635e-06, "loss": 0.4586, "step": 17707 }, { "epoch": 0.7422948345786655, "grad_norm": 1.6720731258392334, "learning_rate": 1.642558910388994e-06, "loss": 0.5182, "step": 17708 }, { "epoch": 0.7423367531936744, "grad_norm": 1.885581612586975, "learning_rate": 1.642055901444507e-06, "loss": 0.4359, "step": 17709 }, { "epoch": 0.7423786718086834, "grad_norm": 1.706183671951294, "learning_rate": 1.6415529543990793e-06, "loss": 0.5224, "step": 17710 }, { "epoch": 0.7424205904236924, "grad_norm": 1.9405901432037354, "learning_rate": 1.6410500692619802e-06, "loss": 0.5164, "step": 17711 }, { "epoch": 0.7424625090387014, "grad_norm": 1.6999144554138184, "learning_rate": 1.6405472460424775e-06, "loss": 0.5379, "step": 17712 }, { "epoch": 0.7425044276537103, "grad_norm": 1.5838721990585327, "learning_rate": 1.640044484749842e-06, "loss": 0.4827, "step": 17713 }, { "epoch": 0.7425463462687193, "grad_norm": 2.2449605464935303, "learning_rate": 1.6395417853933432e-06, "loss": 0.4703, "step": 17714 }, { "epoch": 0.7425882648837282, "grad_norm": 1.6865668296813965, "learning_rate": 1.6390391479822443e-06, "loss": 0.4154, "step": 17715 }, { "epoch": 0.7426301834987372, "grad_norm": 1.7857608795166016, "learning_rate": 1.6385365725258124e-06, "loss": 0.4923, "step": 17716 }, { "epoch": 0.7426721021137461, "grad_norm": 1.7145673036575317, "learning_rate": 1.6380340590333132e-06, "loss": 0.4759, "step": 17717 }, { "epoch": 0.7427140207287551, "grad_norm": 1.7661501169204712, "learning_rate": 1.6375316075140069e-06, "loss": 0.4695, "step": 17718 }, { "epoch": 0.742755939343764, "grad_norm": 1.854152798652649, "learning_rate": 1.637029217977158e-06, "loss": 0.4718, "step": 17719 }, { "epoch": 0.7427978579587731, "grad_norm": 1.935254693031311, "learning_rate": 1.6365268904320269e-06, "loss": 0.4612, "step": 17720 }, { "epoch": 0.742839776573782, "grad_norm": 1.6674842834472656, "learning_rate": 1.6360246248878708e-06, "loss": 0.4417, "step": 17721 }, { "epoch": 0.742881695188791, "grad_norm": 3.1484463214874268, "learning_rate": 1.6355224213539505e-06, "loss": 0.4432, "step": 17722 }, { "epoch": 0.7429236138037999, "grad_norm": 1.742936611175537, "learning_rate": 1.6350202798395242e-06, "loss": 0.4746, "step": 17723 }, { "epoch": 0.7429655324188089, "grad_norm": 1.6005067825317383, "learning_rate": 1.6345182003538456e-06, "loss": 0.4753, "step": 17724 }, { "epoch": 0.7430074510338178, "grad_norm": 4.059077262878418, "learning_rate": 1.6340161829061717e-06, "loss": 0.5147, "step": 17725 }, { "epoch": 0.7430493696488268, "grad_norm": 1.647829532623291, "learning_rate": 1.6335142275057576e-06, "loss": 0.4999, "step": 17726 }, { "epoch": 0.7430912882638357, "grad_norm": 1.8220593929290771, "learning_rate": 1.633012334161853e-06, "loss": 0.4896, "step": 17727 }, { "epoch": 0.7431332068788448, "grad_norm": 1.631274700164795, "learning_rate": 1.6325105028837129e-06, "loss": 0.4631, "step": 17728 }, { "epoch": 0.7431751254938537, "grad_norm": 2.3715248107910156, "learning_rate": 1.6320087336805846e-06, "loss": 0.5015, "step": 17729 }, { "epoch": 0.7432170441088627, "grad_norm": 1.7432527542114258, "learning_rate": 1.6315070265617212e-06, "loss": 0.4956, "step": 17730 }, { "epoch": 0.7432589627238716, "grad_norm": 1.8703380823135376, "learning_rate": 1.631005381536367e-06, "loss": 0.5109, "step": 17731 }, { "epoch": 0.7433008813388806, "grad_norm": 1.7670326232910156, "learning_rate": 1.6305037986137729e-06, "loss": 0.4867, "step": 17732 }, { "epoch": 0.7433427999538895, "grad_norm": 1.7441880702972412, "learning_rate": 1.630002277803181e-06, "loss": 0.4692, "step": 17733 }, { "epoch": 0.7433847185688984, "grad_norm": 1.8837066888809204, "learning_rate": 1.6295008191138384e-06, "loss": 0.4965, "step": 17734 }, { "epoch": 0.7434266371839074, "grad_norm": 1.6281074285507202, "learning_rate": 1.6289994225549904e-06, "loss": 0.5192, "step": 17735 }, { "epoch": 0.7434685557989164, "grad_norm": 5.119332790374756, "learning_rate": 1.6284980881358764e-06, "loss": 0.4475, "step": 17736 }, { "epoch": 0.7435104744139254, "grad_norm": 1.9516727924346924, "learning_rate": 1.627996815865739e-06, "loss": 0.4598, "step": 17737 }, { "epoch": 0.7435523930289343, "grad_norm": 2.698688268661499, "learning_rate": 1.6274956057538199e-06, "loss": 0.4974, "step": 17738 }, { "epoch": 0.7435943116439433, "grad_norm": 2.0494091510772705, "learning_rate": 1.6269944578093577e-06, "loss": 0.4941, "step": 17739 }, { "epoch": 0.7436362302589522, "grad_norm": 1.8537874221801758, "learning_rate": 1.626493372041587e-06, "loss": 0.4594, "step": 17740 }, { "epoch": 0.7436781488739612, "grad_norm": 1.614240050315857, "learning_rate": 1.6259923484597495e-06, "loss": 0.4725, "step": 17741 }, { "epoch": 0.7437200674889701, "grad_norm": 1.5519973039627075, "learning_rate": 1.625491387073077e-06, "loss": 0.4709, "step": 17742 }, { "epoch": 0.7437619861039791, "grad_norm": 1.614683985710144, "learning_rate": 1.6249904878908057e-06, "loss": 0.5492, "step": 17743 }, { "epoch": 0.743803904718988, "grad_norm": 1.749647855758667, "learning_rate": 1.6244896509221707e-06, "loss": 0.4806, "step": 17744 }, { "epoch": 0.7438458233339971, "grad_norm": 1.9273282289505005, "learning_rate": 1.6239888761764005e-06, "loss": 0.4614, "step": 17745 }, { "epoch": 0.743887741949006, "grad_norm": 2.109066963195801, "learning_rate": 1.623488163662728e-06, "loss": 0.4875, "step": 17746 }, { "epoch": 0.743929660564015, "grad_norm": 1.7076460123062134, "learning_rate": 1.6229875133903855e-06, "loss": 0.4757, "step": 17747 }, { "epoch": 0.7439715791790239, "grad_norm": 1.997628927230835, "learning_rate": 1.6224869253685993e-06, "loss": 0.457, "step": 17748 }, { "epoch": 0.7440134977940329, "grad_norm": 1.7603204250335693, "learning_rate": 1.621986399606596e-06, "loss": 0.4543, "step": 17749 }, { "epoch": 0.7440554164090418, "grad_norm": 1.7027019262313843, "learning_rate": 1.6214859361136048e-06, "loss": 0.5095, "step": 17750 }, { "epoch": 0.7440973350240508, "grad_norm": 1.688051700592041, "learning_rate": 1.6209855348988486e-06, "loss": 0.4696, "step": 17751 }, { "epoch": 0.7441392536390598, "grad_norm": 2.0236117839813232, "learning_rate": 1.6204851959715533e-06, "loss": 0.4911, "step": 17752 }, { "epoch": 0.7441811722540688, "grad_norm": 2.0040225982666016, "learning_rate": 1.6199849193409407e-06, "loss": 0.4455, "step": 17753 }, { "epoch": 0.7442230908690777, "grad_norm": 2.0996077060699463, "learning_rate": 1.6194847050162328e-06, "loss": 0.4313, "step": 17754 }, { "epoch": 0.7442650094840867, "grad_norm": 1.7299236059188843, "learning_rate": 1.6189845530066529e-06, "loss": 0.4619, "step": 17755 }, { "epoch": 0.7443069280990956, "grad_norm": 1.7656373977661133, "learning_rate": 1.618484463321417e-06, "loss": 0.5017, "step": 17756 }, { "epoch": 0.7443488467141046, "grad_norm": 1.9209052324295044, "learning_rate": 1.6179844359697466e-06, "loss": 0.4547, "step": 17757 }, { "epoch": 0.7443907653291135, "grad_norm": 2.1178929805755615, "learning_rate": 1.6174844709608578e-06, "loss": 0.4902, "step": 17758 }, { "epoch": 0.7444326839441224, "grad_norm": 2.2626636028289795, "learning_rate": 1.6169845683039653e-06, "loss": 0.4447, "step": 17759 }, { "epoch": 0.7444746025591314, "grad_norm": 1.770277976989746, "learning_rate": 1.6164847280082846e-06, "loss": 0.4649, "step": 17760 }, { "epoch": 0.7445165211741404, "grad_norm": 2.752483606338501, "learning_rate": 1.6159849500830326e-06, "loss": 0.4814, "step": 17761 }, { "epoch": 0.7445584397891494, "grad_norm": 2.2686891555786133, "learning_rate": 1.6154852345374183e-06, "loss": 0.5254, "step": 17762 }, { "epoch": 0.7446003584041583, "grad_norm": 1.9588910341262817, "learning_rate": 1.6149855813806542e-06, "loss": 0.4233, "step": 17763 }, { "epoch": 0.7446422770191673, "grad_norm": 1.9562431573867798, "learning_rate": 1.6144859906219534e-06, "loss": 0.5064, "step": 17764 }, { "epoch": 0.7446841956341762, "grad_norm": 1.6760386228561401, "learning_rate": 1.6139864622705209e-06, "loss": 0.4777, "step": 17765 }, { "epoch": 0.7447261142491852, "grad_norm": 1.9731796979904175, "learning_rate": 1.6134869963355688e-06, "loss": 0.4578, "step": 17766 }, { "epoch": 0.7447680328641941, "grad_norm": 1.7238225936889648, "learning_rate": 1.6129875928263022e-06, "loss": 0.4776, "step": 17767 }, { "epoch": 0.7448099514792031, "grad_norm": 2.747838020324707, "learning_rate": 1.6124882517519252e-06, "loss": 0.5172, "step": 17768 }, { "epoch": 0.744851870094212, "grad_norm": 1.7187155485153198, "learning_rate": 1.6119889731216448e-06, "loss": 0.5241, "step": 17769 }, { "epoch": 0.7448937887092211, "grad_norm": 1.758542776107788, "learning_rate": 1.6114897569446648e-06, "loss": 0.4986, "step": 17770 }, { "epoch": 0.74493570732423, "grad_norm": 1.638658881187439, "learning_rate": 1.6109906032301852e-06, "loss": 0.4845, "step": 17771 }, { "epoch": 0.744977625939239, "grad_norm": 1.7434905767440796, "learning_rate": 1.6104915119874086e-06, "loss": 0.4679, "step": 17772 }, { "epoch": 0.7450195445542479, "grad_norm": 1.6409391164779663, "learning_rate": 1.6099924832255371e-06, "loss": 0.4182, "step": 17773 }, { "epoch": 0.7450614631692569, "grad_norm": 1.8141688108444214, "learning_rate": 1.6094935169537657e-06, "loss": 0.47, "step": 17774 }, { "epoch": 0.7451033817842658, "grad_norm": 1.784175157546997, "learning_rate": 1.6089946131812962e-06, "loss": 0.4908, "step": 17775 }, { "epoch": 0.7451453003992748, "grad_norm": 2.04750657081604, "learning_rate": 1.6084957719173212e-06, "loss": 0.5416, "step": 17776 }, { "epoch": 0.7451872190142838, "grad_norm": 1.8733537197113037, "learning_rate": 1.6079969931710393e-06, "loss": 0.4852, "step": 17777 }, { "epoch": 0.7452291376292928, "grad_norm": 1.8378838300704956, "learning_rate": 1.6074982769516417e-06, "loss": 0.4654, "step": 17778 }, { "epoch": 0.7452710562443017, "grad_norm": 1.671662449836731, "learning_rate": 1.6069996232683255e-06, "loss": 0.4534, "step": 17779 }, { "epoch": 0.7453129748593107, "grad_norm": 1.7580286264419556, "learning_rate": 1.6065010321302787e-06, "loss": 0.5072, "step": 17780 }, { "epoch": 0.7453548934743196, "grad_norm": 2.130448579788208, "learning_rate": 1.6060025035466942e-06, "loss": 0.5214, "step": 17781 }, { "epoch": 0.7453968120893286, "grad_norm": 1.7429478168487549, "learning_rate": 1.6055040375267633e-06, "loss": 0.4641, "step": 17782 }, { "epoch": 0.7454387307043375, "grad_norm": 2.636367082595825, "learning_rate": 1.605005634079671e-06, "loss": 0.499, "step": 17783 }, { "epoch": 0.7454806493193464, "grad_norm": 1.6049188375473022, "learning_rate": 1.6045072932146061e-06, "loss": 0.4671, "step": 17784 }, { "epoch": 0.7455225679343555, "grad_norm": 1.8807077407836914, "learning_rate": 1.6040090149407566e-06, "loss": 0.441, "step": 17785 }, { "epoch": 0.7455644865493644, "grad_norm": 2.3088324069976807, "learning_rate": 1.6035107992673065e-06, "loss": 0.5021, "step": 17786 }, { "epoch": 0.7456064051643734, "grad_norm": 1.9759594202041626, "learning_rate": 1.603012646203438e-06, "loss": 0.4709, "step": 17787 }, { "epoch": 0.7456483237793823, "grad_norm": 1.7283806800842285, "learning_rate": 1.602514555758336e-06, "loss": 0.4642, "step": 17788 }, { "epoch": 0.7456902423943913, "grad_norm": 5.574119567871094, "learning_rate": 1.6020165279411804e-06, "loss": 0.484, "step": 17789 }, { "epoch": 0.7457321610094002, "grad_norm": 1.9626381397247314, "learning_rate": 1.6015185627611519e-06, "loss": 0.531, "step": 17790 }, { "epoch": 0.7457740796244092, "grad_norm": 2.1630797386169434, "learning_rate": 1.6010206602274326e-06, "loss": 0.4698, "step": 17791 }, { "epoch": 0.7458159982394181, "grad_norm": 2.2738473415374756, "learning_rate": 1.6005228203491963e-06, "loss": 0.511, "step": 17792 }, { "epoch": 0.7458579168544271, "grad_norm": 1.683493733406067, "learning_rate": 1.6000250431356223e-06, "loss": 0.4853, "step": 17793 }, { "epoch": 0.7458998354694361, "grad_norm": 1.5135103464126587, "learning_rate": 1.5995273285958884e-06, "loss": 0.4425, "step": 17794 }, { "epoch": 0.7459417540844451, "grad_norm": 1.8283361196517944, "learning_rate": 1.5990296767391666e-06, "loss": 0.4676, "step": 17795 }, { "epoch": 0.745983672699454, "grad_norm": 2.59732723236084, "learning_rate": 1.5985320875746295e-06, "loss": 0.497, "step": 17796 }, { "epoch": 0.746025591314463, "grad_norm": 2.298262119293213, "learning_rate": 1.598034561111453e-06, "loss": 0.4673, "step": 17797 }, { "epoch": 0.7460675099294719, "grad_norm": 1.9566084146499634, "learning_rate": 1.5975370973588044e-06, "loss": 0.5042, "step": 17798 }, { "epoch": 0.7461094285444809, "grad_norm": 2.070573091506958, "learning_rate": 1.5970396963258556e-06, "loss": 0.463, "step": 17799 }, { "epoch": 0.7461513471594898, "grad_norm": 1.9354822635650635, "learning_rate": 1.5965423580217776e-06, "loss": 0.451, "step": 17800 }, { "epoch": 0.7461932657744988, "grad_norm": 1.6283493041992188, "learning_rate": 1.596045082455734e-06, "loss": 0.4243, "step": 17801 }, { "epoch": 0.7462351843895078, "grad_norm": 1.8949546813964844, "learning_rate": 1.5955478696368937e-06, "loss": 0.4875, "step": 17802 }, { "epoch": 0.7462771030045168, "grad_norm": 1.8904577493667603, "learning_rate": 1.595050719574423e-06, "loss": 0.4765, "step": 17803 }, { "epoch": 0.7463190216195257, "grad_norm": 2.126721143722534, "learning_rate": 1.5945536322774858e-06, "loss": 0.4765, "step": 17804 }, { "epoch": 0.7463609402345347, "grad_norm": 1.9216618537902832, "learning_rate": 1.5940566077552423e-06, "loss": 0.4519, "step": 17805 }, { "epoch": 0.7464028588495436, "grad_norm": 2.0407793521881104, "learning_rate": 1.5935596460168585e-06, "loss": 0.4593, "step": 17806 }, { "epoch": 0.7464447774645526, "grad_norm": 2.2311036586761475, "learning_rate": 1.593062747071492e-06, "loss": 0.4958, "step": 17807 }, { "epoch": 0.7464866960795615, "grad_norm": 2.1750118732452393, "learning_rate": 1.5925659109283037e-06, "loss": 0.5253, "step": 17808 }, { "epoch": 0.7465286146945704, "grad_norm": 1.5783642530441284, "learning_rate": 1.5920691375964536e-06, "loss": 0.4306, "step": 17809 }, { "epoch": 0.7465705333095795, "grad_norm": 2.344038486480713, "learning_rate": 1.5915724270850963e-06, "loss": 0.4535, "step": 17810 }, { "epoch": 0.7466124519245884, "grad_norm": 2.1201400756835938, "learning_rate": 1.5910757794033887e-06, "loss": 0.4427, "step": 17811 }, { "epoch": 0.7466543705395974, "grad_norm": 1.679945945739746, "learning_rate": 1.5905791945604887e-06, "loss": 0.4772, "step": 17812 }, { "epoch": 0.7466962891546063, "grad_norm": 1.7434064149856567, "learning_rate": 1.5900826725655477e-06, "loss": 0.482, "step": 17813 }, { "epoch": 0.7467382077696153, "grad_norm": 2.0999464988708496, "learning_rate": 1.5895862134277168e-06, "loss": 0.486, "step": 17814 }, { "epoch": 0.7467801263846242, "grad_norm": 1.8951829671859741, "learning_rate": 1.589089817156151e-06, "loss": 0.5227, "step": 17815 }, { "epoch": 0.7468220449996332, "grad_norm": 1.9689218997955322, "learning_rate": 1.5885934837599977e-06, "loss": 0.4623, "step": 17816 }, { "epoch": 0.7468639636146421, "grad_norm": 1.6984853744506836, "learning_rate": 1.5880972132484069e-06, "loss": 0.4523, "step": 17817 }, { "epoch": 0.7469058822296512, "grad_norm": 2.654668092727661, "learning_rate": 1.5876010056305286e-06, "loss": 0.5271, "step": 17818 }, { "epoch": 0.7469478008446601, "grad_norm": 1.7542134523391724, "learning_rate": 1.5871048609155072e-06, "loss": 0.4662, "step": 17819 }, { "epoch": 0.7469897194596691, "grad_norm": 2.0181543827056885, "learning_rate": 1.5866087791124891e-06, "loss": 0.5239, "step": 17820 }, { "epoch": 0.747031638074678, "grad_norm": 1.9575767517089844, "learning_rate": 1.5861127602306209e-06, "loss": 0.524, "step": 17821 }, { "epoch": 0.747073556689687, "grad_norm": 1.7039659023284912, "learning_rate": 1.585616804279045e-06, "loss": 0.4712, "step": 17822 }, { "epoch": 0.7471154753046959, "grad_norm": 1.8344258069992065, "learning_rate": 1.5851209112669007e-06, "loss": 0.5025, "step": 17823 }, { "epoch": 0.7471573939197049, "grad_norm": 1.9607290029525757, "learning_rate": 1.5846250812033337e-06, "loss": 0.4552, "step": 17824 }, { "epoch": 0.7471993125347138, "grad_norm": 1.7497667074203491, "learning_rate": 1.5841293140974796e-06, "loss": 0.4761, "step": 17825 }, { "epoch": 0.7472412311497229, "grad_norm": 1.9036632776260376, "learning_rate": 1.5836336099584792e-06, "loss": 0.487, "step": 17826 }, { "epoch": 0.7472831497647318, "grad_norm": 1.8046094179153442, "learning_rate": 1.5831379687954724e-06, "loss": 0.5138, "step": 17827 }, { "epoch": 0.7473250683797408, "grad_norm": 1.7424274682998657, "learning_rate": 1.5826423906175913e-06, "loss": 0.4507, "step": 17828 }, { "epoch": 0.7473669869947497, "grad_norm": 2.148118734359741, "learning_rate": 1.5821468754339758e-06, "loss": 0.5103, "step": 17829 }, { "epoch": 0.7474089056097587, "grad_norm": 14.372140884399414, "learning_rate": 1.5816514232537555e-06, "loss": 0.4447, "step": 17830 }, { "epoch": 0.7474508242247676, "grad_norm": 1.9214606285095215, "learning_rate": 1.5811560340860655e-06, "loss": 0.4724, "step": 17831 }, { "epoch": 0.7474927428397766, "grad_norm": 2.007632255554199, "learning_rate": 1.5806607079400387e-06, "loss": 0.4816, "step": 17832 }, { "epoch": 0.7475346614547855, "grad_norm": 2.0135624408721924, "learning_rate": 1.5801654448248055e-06, "loss": 0.4879, "step": 17833 }, { "epoch": 0.7475765800697944, "grad_norm": 2.057934522628784, "learning_rate": 1.5796702447494928e-06, "loss": 0.4696, "step": 17834 }, { "epoch": 0.7476184986848035, "grad_norm": 1.8573623895645142, "learning_rate": 1.579175107723232e-06, "loss": 0.513, "step": 17835 }, { "epoch": 0.7476604172998124, "grad_norm": 1.7193043231964111, "learning_rate": 1.5786800337551471e-06, "loss": 0.4531, "step": 17836 }, { "epoch": 0.7477023359148214, "grad_norm": 1.6202328205108643, "learning_rate": 1.5781850228543666e-06, "loss": 0.4467, "step": 17837 }, { "epoch": 0.7477442545298303, "grad_norm": 2.0565648078918457, "learning_rate": 1.5776900750300155e-06, "loss": 0.5491, "step": 17838 }, { "epoch": 0.7477861731448393, "grad_norm": 1.8950448036193848, "learning_rate": 1.5771951902912159e-06, "loss": 0.463, "step": 17839 }, { "epoch": 0.7478280917598482, "grad_norm": 2.110851526260376, "learning_rate": 1.5767003686470905e-06, "loss": 0.4843, "step": 17840 }, { "epoch": 0.7478700103748572, "grad_norm": 2.042088508605957, "learning_rate": 1.576205610106763e-06, "loss": 0.4281, "step": 17841 }, { "epoch": 0.7479119289898661, "grad_norm": 1.6993316411972046, "learning_rate": 1.575710914679352e-06, "loss": 0.4641, "step": 17842 }, { "epoch": 0.7479538476048752, "grad_norm": 1.5977948904037476, "learning_rate": 1.575216282373974e-06, "loss": 0.4249, "step": 17843 }, { "epoch": 0.7479957662198841, "grad_norm": 1.8507604598999023, "learning_rate": 1.5747217131997516e-06, "loss": 0.4734, "step": 17844 }, { "epoch": 0.7480376848348931, "grad_norm": 1.6985886096954346, "learning_rate": 1.5742272071657976e-06, "loss": 0.4875, "step": 17845 }, { "epoch": 0.748079603449902, "grad_norm": 1.8215795755386353, "learning_rate": 1.5737327642812284e-06, "loss": 0.4696, "step": 17846 }, { "epoch": 0.748121522064911, "grad_norm": 1.9695487022399902, "learning_rate": 1.5732383845551612e-06, "loss": 0.4756, "step": 17847 }, { "epoch": 0.7481634406799199, "grad_norm": 1.6055861711502075, "learning_rate": 1.5727440679967054e-06, "loss": 0.4297, "step": 17848 }, { "epoch": 0.7482053592949289, "grad_norm": 1.6845368146896362, "learning_rate": 1.5722498146149745e-06, "loss": 0.4498, "step": 17849 }, { "epoch": 0.7482472779099378, "grad_norm": 1.7078969478607178, "learning_rate": 1.5717556244190807e-06, "loss": 0.4412, "step": 17850 }, { "epoch": 0.7482891965249469, "grad_norm": 1.805996060371399, "learning_rate": 1.5712614974181335e-06, "loss": 0.5142, "step": 17851 }, { "epoch": 0.7483311151399558, "grad_norm": 1.671226978302002, "learning_rate": 1.570767433621238e-06, "loss": 0.4428, "step": 17852 }, { "epoch": 0.7483730337549648, "grad_norm": 1.6770460605621338, "learning_rate": 1.5702734330375063e-06, "loss": 0.5063, "step": 17853 }, { "epoch": 0.7484149523699737, "grad_norm": 1.9549355506896973, "learning_rate": 1.5697794956760403e-06, "loss": 0.4691, "step": 17854 }, { "epoch": 0.7484568709849827, "grad_norm": 1.80462646484375, "learning_rate": 1.5692856215459472e-06, "loss": 0.4992, "step": 17855 }, { "epoch": 0.7484987895999916, "grad_norm": 1.9827324151992798, "learning_rate": 1.5687918106563326e-06, "loss": 0.4467, "step": 17856 }, { "epoch": 0.7485407082150006, "grad_norm": 1.8868255615234375, "learning_rate": 1.5682980630162964e-06, "loss": 0.4754, "step": 17857 }, { "epoch": 0.7485826268300095, "grad_norm": 1.889480471611023, "learning_rate": 1.5678043786349406e-06, "loss": 0.4779, "step": 17858 }, { "epoch": 0.7486245454450184, "grad_norm": 1.646131157875061, "learning_rate": 1.567310757521368e-06, "loss": 0.4897, "step": 17859 }, { "epoch": 0.7486664640600275, "grad_norm": 1.6884902715682983, "learning_rate": 1.566817199684676e-06, "loss": 0.5107, "step": 17860 }, { "epoch": 0.7487083826750364, "grad_norm": 1.8274120092391968, "learning_rate": 1.5663237051339613e-06, "loss": 0.5166, "step": 17861 }, { "epoch": 0.7487503012900454, "grad_norm": 1.9539544582366943, "learning_rate": 1.5658302738783231e-06, "loss": 0.493, "step": 17862 }, { "epoch": 0.7487922199050543, "grad_norm": 2.0748374462127686, "learning_rate": 1.5653369059268552e-06, "loss": 0.4792, "step": 17863 }, { "epoch": 0.7488341385200633, "grad_norm": 2.059156656265259, "learning_rate": 1.564843601288653e-06, "loss": 0.5434, "step": 17864 }, { "epoch": 0.7488760571350722, "grad_norm": 2.9418466091156006, "learning_rate": 1.564350359972812e-06, "loss": 0.4261, "step": 17865 }, { "epoch": 0.7489179757500812, "grad_norm": 3.2851498126983643, "learning_rate": 1.5638571819884207e-06, "loss": 0.52, "step": 17866 }, { "epoch": 0.7489598943650901, "grad_norm": 1.6291306018829346, "learning_rate": 1.5633640673445717e-06, "loss": 0.4731, "step": 17867 }, { "epoch": 0.7490018129800992, "grad_norm": 1.737988829612732, "learning_rate": 1.562871016050357e-06, "loss": 0.4865, "step": 17868 }, { "epoch": 0.7490437315951081, "grad_norm": 1.707391619682312, "learning_rate": 1.5623780281148632e-06, "loss": 0.4795, "step": 17869 }, { "epoch": 0.7490856502101171, "grad_norm": 2.126323938369751, "learning_rate": 1.5618851035471766e-06, "loss": 0.4591, "step": 17870 }, { "epoch": 0.749127568825126, "grad_norm": 1.8489617109298706, "learning_rate": 1.5613922423563866e-06, "loss": 0.4821, "step": 17871 }, { "epoch": 0.749169487440135, "grad_norm": 2.1818182468414307, "learning_rate": 1.5608994445515752e-06, "loss": 0.4765, "step": 17872 }, { "epoch": 0.7492114060551439, "grad_norm": 2.124091386795044, "learning_rate": 1.560406710141828e-06, "loss": 0.4911, "step": 17873 }, { "epoch": 0.7492533246701529, "grad_norm": 1.737199068069458, "learning_rate": 1.5599140391362299e-06, "loss": 0.4877, "step": 17874 }, { "epoch": 0.7492952432851618, "grad_norm": 1.8107762336730957, "learning_rate": 1.5594214315438589e-06, "loss": 0.481, "step": 17875 }, { "epoch": 0.7493371619001709, "grad_norm": 1.712576150894165, "learning_rate": 1.5589288873737974e-06, "loss": 0.5296, "step": 17876 }, { "epoch": 0.7493790805151798, "grad_norm": 1.8166755437850952, "learning_rate": 1.5584364066351266e-06, "loss": 0.5545, "step": 17877 }, { "epoch": 0.7494209991301888, "grad_norm": 1.7971664667129517, "learning_rate": 1.5579439893369209e-06, "loss": 0.4829, "step": 17878 }, { "epoch": 0.7494629177451977, "grad_norm": 2.0379650592803955, "learning_rate": 1.5574516354882608e-06, "loss": 0.5131, "step": 17879 }, { "epoch": 0.7495048363602067, "grad_norm": 1.880576252937317, "learning_rate": 1.5569593450982208e-06, "loss": 0.4557, "step": 17880 }, { "epoch": 0.7495467549752156, "grad_norm": 2.03133225440979, "learning_rate": 1.5564671181758739e-06, "loss": 0.4851, "step": 17881 }, { "epoch": 0.7495886735902246, "grad_norm": 2.059492826461792, "learning_rate": 1.5559749547302949e-06, "loss": 0.4535, "step": 17882 }, { "epoch": 0.7496305922052335, "grad_norm": 1.8119661808013916, "learning_rate": 1.555482854770558e-06, "loss": 0.4681, "step": 17883 }, { "epoch": 0.7496725108202424, "grad_norm": 1.8481863737106323, "learning_rate": 1.5549908183057315e-06, "loss": 0.4928, "step": 17884 }, { "epoch": 0.7497144294352515, "grad_norm": 1.8607538938522339, "learning_rate": 1.5544988453448867e-06, "loss": 0.489, "step": 17885 }, { "epoch": 0.7497563480502604, "grad_norm": 1.6605281829833984, "learning_rate": 1.5540069358970934e-06, "loss": 0.4602, "step": 17886 }, { "epoch": 0.7497982666652694, "grad_norm": 1.8957821130752563, "learning_rate": 1.5535150899714174e-06, "loss": 0.5321, "step": 17887 }, { "epoch": 0.7498401852802783, "grad_norm": 1.6740479469299316, "learning_rate": 1.5530233075769274e-06, "loss": 0.4321, "step": 17888 }, { "epoch": 0.7498821038952873, "grad_norm": 1.6966347694396973, "learning_rate": 1.5525315887226872e-06, "loss": 0.5028, "step": 17889 }, { "epoch": 0.7499240225102962, "grad_norm": 2.27241587638855, "learning_rate": 1.5520399334177599e-06, "loss": 0.4909, "step": 17890 }, { "epoch": 0.7499659411253052, "grad_norm": 2.140631914138794, "learning_rate": 1.5515483416712097e-06, "loss": 0.4981, "step": 17891 }, { "epoch": 0.7500078597403141, "grad_norm": 1.6133097410202026, "learning_rate": 1.5510568134921e-06, "loss": 0.4733, "step": 17892 }, { "epoch": 0.7500497783553232, "grad_norm": 1.8103950023651123, "learning_rate": 1.5505653488894885e-06, "loss": 0.4726, "step": 17893 }, { "epoch": 0.7500916969703321, "grad_norm": 2.5923519134521484, "learning_rate": 1.5500739478724357e-06, "loss": 0.5072, "step": 17894 }, { "epoch": 0.7501336155853411, "grad_norm": 1.9558380842208862, "learning_rate": 1.5495826104500023e-06, "loss": 0.4802, "step": 17895 }, { "epoch": 0.75017553420035, "grad_norm": 2.0070841312408447, "learning_rate": 1.5490913366312416e-06, "loss": 0.4973, "step": 17896 }, { "epoch": 0.750217452815359, "grad_norm": 2.110546827316284, "learning_rate": 1.5486001264252132e-06, "loss": 0.4884, "step": 17897 }, { "epoch": 0.7502593714303679, "grad_norm": 2.044698476791382, "learning_rate": 1.5481089798409698e-06, "loss": 0.4746, "step": 17898 }, { "epoch": 0.7503012900453769, "grad_norm": 1.687894582748413, "learning_rate": 1.5476178968875632e-06, "loss": 0.5001, "step": 17899 }, { "epoch": 0.7503432086603858, "grad_norm": 2.0019450187683105, "learning_rate": 1.5471268775740488e-06, "loss": 0.4734, "step": 17900 }, { "epoch": 0.7503851272753949, "grad_norm": 3.441828727722168, "learning_rate": 1.5466359219094785e-06, "loss": 0.4928, "step": 17901 }, { "epoch": 0.7504270458904038, "grad_norm": 1.914366364479065, "learning_rate": 1.5461450299028985e-06, "loss": 0.5388, "step": 17902 }, { "epoch": 0.7504689645054128, "grad_norm": 1.786861538887024, "learning_rate": 1.5456542015633625e-06, "loss": 0.4458, "step": 17903 }, { "epoch": 0.7505108831204217, "grad_norm": 1.7660611867904663, "learning_rate": 1.5451634368999136e-06, "loss": 0.4643, "step": 17904 }, { "epoch": 0.7505528017354307, "grad_norm": 1.7831978797912598, "learning_rate": 1.5446727359216007e-06, "loss": 0.4845, "step": 17905 }, { "epoch": 0.7505947203504396, "grad_norm": 1.5280512571334839, "learning_rate": 1.5441820986374707e-06, "loss": 0.5027, "step": 17906 }, { "epoch": 0.7506366389654486, "grad_norm": 1.7398240566253662, "learning_rate": 1.5436915250565666e-06, "loss": 0.4452, "step": 17907 }, { "epoch": 0.7506785575804575, "grad_norm": 1.7044557332992554, "learning_rate": 1.543201015187929e-06, "loss": 0.4614, "step": 17908 }, { "epoch": 0.7507204761954666, "grad_norm": 1.7391939163208008, "learning_rate": 1.542710569040603e-06, "loss": 0.4563, "step": 17909 }, { "epoch": 0.7507623948104755, "grad_norm": 1.6666916608810425, "learning_rate": 1.542220186623627e-06, "loss": 0.4601, "step": 17910 }, { "epoch": 0.7508043134254844, "grad_norm": 1.8177481889724731, "learning_rate": 1.5417298679460418e-06, "loss": 0.4365, "step": 17911 }, { "epoch": 0.7508462320404934, "grad_norm": 1.813983678817749, "learning_rate": 1.5412396130168873e-06, "loss": 0.5555, "step": 17912 }, { "epoch": 0.7508881506555023, "grad_norm": 1.652848243713379, "learning_rate": 1.5407494218451967e-06, "loss": 0.4929, "step": 17913 }, { "epoch": 0.7509300692705113, "grad_norm": 1.602914810180664, "learning_rate": 1.5402592944400085e-06, "loss": 0.5002, "step": 17914 }, { "epoch": 0.7509719878855202, "grad_norm": 1.750798225402832, "learning_rate": 1.539769230810359e-06, "loss": 0.5283, "step": 17915 }, { "epoch": 0.7510139065005292, "grad_norm": 1.9525682926177979, "learning_rate": 1.5392792309652782e-06, "loss": 0.4512, "step": 17916 }, { "epoch": 0.7510558251155381, "grad_norm": 2.1637449264526367, "learning_rate": 1.5387892949138023e-06, "loss": 0.5152, "step": 17917 }, { "epoch": 0.7510977437305472, "grad_norm": 2.4916422367095947, "learning_rate": 1.5382994226649606e-06, "loss": 0.447, "step": 17918 }, { "epoch": 0.7511396623455561, "grad_norm": 1.657853603363037, "learning_rate": 1.537809614227782e-06, "loss": 0.4637, "step": 17919 }, { "epoch": 0.7511815809605651, "grad_norm": 1.8256858587265015, "learning_rate": 1.5373198696112967e-06, "loss": 0.469, "step": 17920 }, { "epoch": 0.751223499575574, "grad_norm": 2.6037094593048096, "learning_rate": 1.5368301888245346e-06, "loss": 0.4653, "step": 17921 }, { "epoch": 0.751265418190583, "grad_norm": 1.6783907413482666, "learning_rate": 1.5363405718765185e-06, "loss": 0.4324, "step": 17922 }, { "epoch": 0.7513073368055919, "grad_norm": 2.149048089981079, "learning_rate": 1.5358510187762747e-06, "loss": 0.5123, "step": 17923 }, { "epoch": 0.7513492554206009, "grad_norm": 1.6488010883331299, "learning_rate": 1.5353615295328312e-06, "loss": 0.4404, "step": 17924 }, { "epoch": 0.7513911740356098, "grad_norm": 1.5030217170715332, "learning_rate": 1.5348721041552056e-06, "loss": 0.42, "step": 17925 }, { "epoch": 0.7514330926506189, "grad_norm": 2.496856927871704, "learning_rate": 1.5343827426524243e-06, "loss": 0.5098, "step": 17926 }, { "epoch": 0.7514750112656278, "grad_norm": 1.7619346380233765, "learning_rate": 1.5338934450335064e-06, "loss": 0.5188, "step": 17927 }, { "epoch": 0.7515169298806368, "grad_norm": 1.8549333810806274, "learning_rate": 1.5334042113074693e-06, "loss": 0.4227, "step": 17928 }, { "epoch": 0.7515588484956457, "grad_norm": 1.9379079341888428, "learning_rate": 1.5329150414833333e-06, "loss": 0.509, "step": 17929 }, { "epoch": 0.7516007671106547, "grad_norm": 1.8991899490356445, "learning_rate": 1.5324259355701166e-06, "loss": 0.4632, "step": 17930 }, { "epoch": 0.7516426857256636, "grad_norm": 1.555614948272705, "learning_rate": 1.5319368935768326e-06, "loss": 0.4605, "step": 17931 }, { "epoch": 0.7516846043406726, "grad_norm": 1.8843464851379395, "learning_rate": 1.5314479155124978e-06, "loss": 0.4617, "step": 17932 }, { "epoch": 0.7517265229556815, "grad_norm": 2.3160603046417236, "learning_rate": 1.5309590013861264e-06, "loss": 0.449, "step": 17933 }, { "epoch": 0.7517684415706906, "grad_norm": 2.177973508834839, "learning_rate": 1.5304701512067294e-06, "loss": 0.4926, "step": 17934 }, { "epoch": 0.7518103601856995, "grad_norm": 1.7836787700653076, "learning_rate": 1.5299813649833194e-06, "loss": 0.4876, "step": 17935 }, { "epoch": 0.7518522788007084, "grad_norm": 2.0978658199310303, "learning_rate": 1.5294926427249062e-06, "loss": 0.4812, "step": 17936 }, { "epoch": 0.7518941974157174, "grad_norm": 1.834763526916504, "learning_rate": 1.5290039844404969e-06, "loss": 0.5112, "step": 17937 }, { "epoch": 0.7519361160307263, "grad_norm": 2.119682788848877, "learning_rate": 1.5285153901391004e-06, "loss": 0.5218, "step": 17938 }, { "epoch": 0.7519780346457353, "grad_norm": 1.8143110275268555, "learning_rate": 1.5280268598297248e-06, "loss": 0.4645, "step": 17939 }, { "epoch": 0.7520199532607442, "grad_norm": 1.7846729755401611, "learning_rate": 1.5275383935213727e-06, "loss": 0.4838, "step": 17940 }, { "epoch": 0.7520618718757532, "grad_norm": 2.0435523986816406, "learning_rate": 1.5270499912230502e-06, "loss": 0.5217, "step": 17941 }, { "epoch": 0.7521037904907621, "grad_norm": 1.599205732345581, "learning_rate": 1.526561652943761e-06, "loss": 0.4745, "step": 17942 }, { "epoch": 0.7521457091057712, "grad_norm": 1.8201370239257812, "learning_rate": 1.5260733786925048e-06, "loss": 0.4704, "step": 17943 }, { "epoch": 0.7521876277207801, "grad_norm": 1.7541813850402832, "learning_rate": 1.5255851684782846e-06, "loss": 0.4798, "step": 17944 }, { "epoch": 0.7522295463357891, "grad_norm": 1.655961275100708, "learning_rate": 1.5250970223100986e-06, "loss": 0.4621, "step": 17945 }, { "epoch": 0.752271464950798, "grad_norm": 2.0150673389434814, "learning_rate": 1.5246089401969432e-06, "loss": 0.4797, "step": 17946 }, { "epoch": 0.752313383565807, "grad_norm": 1.6835404634475708, "learning_rate": 1.5241209221478177e-06, "loss": 0.4621, "step": 17947 }, { "epoch": 0.7523553021808159, "grad_norm": 1.9239304065704346, "learning_rate": 1.5236329681717193e-06, "loss": 0.4669, "step": 17948 }, { "epoch": 0.7523972207958249, "grad_norm": 2.3865041732788086, "learning_rate": 1.5231450782776397e-06, "loss": 0.4812, "step": 17949 }, { "epoch": 0.7524391394108338, "grad_norm": 1.9475913047790527, "learning_rate": 1.5226572524745737e-06, "loss": 0.512, "step": 17950 }, { "epoch": 0.7524810580258429, "grad_norm": 2.62503719329834, "learning_rate": 1.5221694907715156e-06, "loss": 0.4781, "step": 17951 }, { "epoch": 0.7525229766408518, "grad_norm": 1.7931950092315674, "learning_rate": 1.5216817931774535e-06, "loss": 0.4301, "step": 17952 }, { "epoch": 0.7525648952558608, "grad_norm": 2.532428503036499, "learning_rate": 1.5211941597013808e-06, "loss": 0.4627, "step": 17953 }, { "epoch": 0.7526068138708697, "grad_norm": 2.2017033100128174, "learning_rate": 1.5207065903522838e-06, "loss": 0.453, "step": 17954 }, { "epoch": 0.7526487324858787, "grad_norm": 1.9118578433990479, "learning_rate": 1.52021908513915e-06, "loss": 0.5141, "step": 17955 }, { "epoch": 0.7526906511008876, "grad_norm": 1.6434638500213623, "learning_rate": 1.5197316440709664e-06, "loss": 0.4498, "step": 17956 }, { "epoch": 0.7527325697158966, "grad_norm": 1.8095784187316895, "learning_rate": 1.51924426715672e-06, "loss": 0.4927, "step": 17957 }, { "epoch": 0.7527744883309055, "grad_norm": 1.8733164072036743, "learning_rate": 1.518756954405392e-06, "loss": 0.4899, "step": 17958 }, { "epoch": 0.7528164069459146, "grad_norm": 1.8956074714660645, "learning_rate": 1.518269705825967e-06, "loss": 0.4138, "step": 17959 }, { "epoch": 0.7528583255609235, "grad_norm": 2.6528258323669434, "learning_rate": 1.5177825214274284e-06, "loss": 0.5198, "step": 17960 }, { "epoch": 0.7529002441759324, "grad_norm": 1.7725520133972168, "learning_rate": 1.517295401218753e-06, "loss": 0.4823, "step": 17961 }, { "epoch": 0.7529421627909414, "grad_norm": 1.8643736839294434, "learning_rate": 1.5168083452089238e-06, "loss": 0.4505, "step": 17962 }, { "epoch": 0.7529840814059503, "grad_norm": 1.8761004209518433, "learning_rate": 1.516321353406916e-06, "loss": 0.4928, "step": 17963 }, { "epoch": 0.7530260000209593, "grad_norm": 1.9975974559783936, "learning_rate": 1.5158344258217094e-06, "loss": 0.4262, "step": 17964 }, { "epoch": 0.7530679186359682, "grad_norm": 2.4224753379821777, "learning_rate": 1.5153475624622772e-06, "loss": 0.4749, "step": 17965 }, { "epoch": 0.7531098372509772, "grad_norm": 1.835712194442749, "learning_rate": 1.5148607633375966e-06, "loss": 0.4813, "step": 17966 }, { "epoch": 0.7531517558659861, "grad_norm": 1.661219596862793, "learning_rate": 1.5143740284566383e-06, "loss": 0.4723, "step": 17967 }, { "epoch": 0.7531936744809952, "grad_norm": 2.0109145641326904, "learning_rate": 1.5138873578283764e-06, "loss": 0.4573, "step": 17968 }, { "epoch": 0.7532355930960041, "grad_norm": 1.7471483945846558, "learning_rate": 1.5134007514617827e-06, "loss": 0.4381, "step": 17969 }, { "epoch": 0.7532775117110131, "grad_norm": 1.8482204675674438, "learning_rate": 1.5129142093658244e-06, "loss": 0.4691, "step": 17970 }, { "epoch": 0.753319430326022, "grad_norm": 1.9353731870651245, "learning_rate": 1.5124277315494724e-06, "loss": 0.4972, "step": 17971 }, { "epoch": 0.753361348941031, "grad_norm": 1.7169735431671143, "learning_rate": 1.511941318021695e-06, "loss": 0.464, "step": 17972 }, { "epoch": 0.7534032675560399, "grad_norm": 1.690649151802063, "learning_rate": 1.5114549687914576e-06, "loss": 0.5167, "step": 17973 }, { "epoch": 0.7534451861710489, "grad_norm": 1.6861622333526611, "learning_rate": 1.5109686838677228e-06, "loss": 0.3952, "step": 17974 }, { "epoch": 0.7534871047860578, "grad_norm": 2.0127201080322266, "learning_rate": 1.510482463259459e-06, "loss": 0.4684, "step": 17975 }, { "epoch": 0.7535290234010669, "grad_norm": 1.8552290201187134, "learning_rate": 1.5099963069756251e-06, "loss": 0.4616, "step": 17976 }, { "epoch": 0.7535709420160758, "grad_norm": 2.109698534011841, "learning_rate": 1.5095102150251867e-06, "loss": 0.5265, "step": 17977 }, { "epoch": 0.7536128606310848, "grad_norm": 3.738579750061035, "learning_rate": 1.5090241874170996e-06, "loss": 0.4747, "step": 17978 }, { "epoch": 0.7536547792460937, "grad_norm": 2.410832405090332, "learning_rate": 1.5085382241603258e-06, "loss": 0.5429, "step": 17979 }, { "epoch": 0.7536966978611027, "grad_norm": 1.756840467453003, "learning_rate": 1.5080523252638247e-06, "loss": 0.4856, "step": 17980 }, { "epoch": 0.7537386164761116, "grad_norm": 1.9496599435806274, "learning_rate": 1.50756649073655e-06, "loss": 0.4476, "step": 17981 }, { "epoch": 0.7537805350911206, "grad_norm": 1.6101948022842407, "learning_rate": 1.5070807205874604e-06, "loss": 0.5036, "step": 17982 }, { "epoch": 0.7538224537061295, "grad_norm": 1.7498953342437744, "learning_rate": 1.5065950148255088e-06, "loss": 0.4263, "step": 17983 }, { "epoch": 0.7538643723211386, "grad_norm": 2.181547164916992, "learning_rate": 1.5061093734596472e-06, "loss": 0.5003, "step": 17984 }, { "epoch": 0.7539062909361475, "grad_norm": 2.0260579586029053, "learning_rate": 1.505623796498829e-06, "loss": 0.5087, "step": 17985 }, { "epoch": 0.7539482095511564, "grad_norm": 2.380458354949951, "learning_rate": 1.5051382839520069e-06, "loss": 0.4558, "step": 17986 }, { "epoch": 0.7539901281661654, "grad_norm": 1.572733998298645, "learning_rate": 1.5046528358281277e-06, "loss": 0.462, "step": 17987 }, { "epoch": 0.7540320467811743, "grad_norm": 2.154284954071045, "learning_rate": 1.5041674521361405e-06, "loss": 0.461, "step": 17988 }, { "epoch": 0.7540739653961833, "grad_norm": 1.9357868432998657, "learning_rate": 1.503682132884996e-06, "loss": 0.4713, "step": 17989 }, { "epoch": 0.7541158840111922, "grad_norm": 2.041240930557251, "learning_rate": 1.503196878083636e-06, "loss": 0.4098, "step": 17990 }, { "epoch": 0.7541578026262012, "grad_norm": 2.2484488487243652, "learning_rate": 1.5027116877410085e-06, "loss": 0.4989, "step": 17991 }, { "epoch": 0.7541997212412102, "grad_norm": 1.6959220170974731, "learning_rate": 1.5022265618660563e-06, "loss": 0.4022, "step": 17992 }, { "epoch": 0.7542416398562192, "grad_norm": 1.952150821685791, "learning_rate": 1.5017415004677212e-06, "loss": 0.5487, "step": 17993 }, { "epoch": 0.7542835584712281, "grad_norm": 1.7697492837905884, "learning_rate": 1.5012565035549442e-06, "loss": 0.4661, "step": 17994 }, { "epoch": 0.7543254770862371, "grad_norm": 2.1116697788238525, "learning_rate": 1.5007715711366688e-06, "loss": 0.4818, "step": 17995 }, { "epoch": 0.754367395701246, "grad_norm": 1.9464551210403442, "learning_rate": 1.5002867032218305e-06, "loss": 0.4659, "step": 17996 }, { "epoch": 0.754409314316255, "grad_norm": 1.6533632278442383, "learning_rate": 1.499801899819368e-06, "loss": 0.4846, "step": 17997 }, { "epoch": 0.7544512329312639, "grad_norm": 1.8002334833145142, "learning_rate": 1.4993171609382207e-06, "loss": 0.4829, "step": 17998 }, { "epoch": 0.7544931515462729, "grad_norm": 2.0762596130371094, "learning_rate": 1.4988324865873204e-06, "loss": 0.5163, "step": 17999 }, { "epoch": 0.7545350701612819, "grad_norm": 1.8877016305923462, "learning_rate": 1.498347876775605e-06, "loss": 0.4702, "step": 18000 }, { "epoch": 0.7545769887762909, "grad_norm": 3.653977870941162, "learning_rate": 1.4978633315120045e-06, "loss": 0.4578, "step": 18001 }, { "epoch": 0.7546189073912998, "grad_norm": 2.088702917098999, "learning_rate": 1.497378850805451e-06, "loss": 0.5389, "step": 18002 }, { "epoch": 0.7546608260063088, "grad_norm": 1.7970056533813477, "learning_rate": 1.4968944346648761e-06, "loss": 0.4437, "step": 18003 }, { "epoch": 0.7547027446213177, "grad_norm": 1.7692333459854126, "learning_rate": 1.4964100830992113e-06, "loss": 0.4382, "step": 18004 }, { "epoch": 0.7547446632363267, "grad_norm": 1.7374173402786255, "learning_rate": 1.4959257961173811e-06, "loss": 0.4674, "step": 18005 }, { "epoch": 0.7547865818513356, "grad_norm": 1.7096498012542725, "learning_rate": 1.4954415737283145e-06, "loss": 0.4674, "step": 18006 }, { "epoch": 0.7548285004663446, "grad_norm": 1.9554986953735352, "learning_rate": 1.4949574159409391e-06, "loss": 0.443, "step": 18007 }, { "epoch": 0.7548704190813535, "grad_norm": 3.2466490268707275, "learning_rate": 1.4944733227641773e-06, "loss": 0.5357, "step": 18008 }, { "epoch": 0.7549123376963626, "grad_norm": 1.9655892848968506, "learning_rate": 1.4939892942069544e-06, "loss": 0.5412, "step": 18009 }, { "epoch": 0.7549542563113715, "grad_norm": 1.8936222791671753, "learning_rate": 1.4935053302781904e-06, "loss": 0.4412, "step": 18010 }, { "epoch": 0.7549961749263804, "grad_norm": 2.289231061935425, "learning_rate": 1.4930214309868103e-06, "loss": 0.5311, "step": 18011 }, { "epoch": 0.7550380935413894, "grad_norm": 1.877968192100525, "learning_rate": 1.4925375963417294e-06, "loss": 0.4433, "step": 18012 }, { "epoch": 0.7550800121563983, "grad_norm": 1.8075286149978638, "learning_rate": 1.4920538263518713e-06, "loss": 0.4556, "step": 18013 }, { "epoch": 0.7551219307714073, "grad_norm": 1.855995535850525, "learning_rate": 1.4915701210261491e-06, "loss": 0.5262, "step": 18014 }, { "epoch": 0.7551638493864162, "grad_norm": 1.7204360961914062, "learning_rate": 1.4910864803734821e-06, "loss": 0.4766, "step": 18015 }, { "epoch": 0.7552057680014252, "grad_norm": 3.41631817817688, "learning_rate": 1.4906029044027859e-06, "loss": 0.4818, "step": 18016 }, { "epoch": 0.7552476866164342, "grad_norm": 1.9520021677017212, "learning_rate": 1.4901193931229723e-06, "loss": 0.4754, "step": 18017 }, { "epoch": 0.7552896052314432, "grad_norm": 1.5932666063308716, "learning_rate": 1.4896359465429556e-06, "loss": 0.462, "step": 18018 }, { "epoch": 0.7553315238464521, "grad_norm": 1.8629118204116821, "learning_rate": 1.4891525646716488e-06, "loss": 0.4385, "step": 18019 }, { "epoch": 0.7553734424614611, "grad_norm": 1.5435413122177124, "learning_rate": 1.4886692475179604e-06, "loss": 0.4914, "step": 18020 }, { "epoch": 0.75541536107647, "grad_norm": 2.0396034717559814, "learning_rate": 1.4881859950907989e-06, "loss": 0.4355, "step": 18021 }, { "epoch": 0.755457279691479, "grad_norm": 1.8456875085830688, "learning_rate": 1.4877028073990752e-06, "loss": 0.5246, "step": 18022 }, { "epoch": 0.7554991983064879, "grad_norm": 1.8046159744262695, "learning_rate": 1.4872196844516934e-06, "loss": 0.54, "step": 18023 }, { "epoch": 0.755541116921497, "grad_norm": 2.015094757080078, "learning_rate": 1.48673662625756e-06, "loss": 0.5561, "step": 18024 }, { "epoch": 0.7555830355365059, "grad_norm": 2.3780808448791504, "learning_rate": 1.4862536328255816e-06, "loss": 0.4382, "step": 18025 }, { "epoch": 0.7556249541515149, "grad_norm": 1.793455719947815, "learning_rate": 1.4857707041646586e-06, "loss": 0.4822, "step": 18026 }, { "epoch": 0.7556668727665238, "grad_norm": 1.8160587549209595, "learning_rate": 1.485287840283694e-06, "loss": 0.5321, "step": 18027 }, { "epoch": 0.7557087913815328, "grad_norm": 1.8033910989761353, "learning_rate": 1.484805041191591e-06, "loss": 0.4096, "step": 18028 }, { "epoch": 0.7557507099965417, "grad_norm": 2.0386009216308594, "learning_rate": 1.4843223068972467e-06, "loss": 0.5129, "step": 18029 }, { "epoch": 0.7557926286115507, "grad_norm": 1.7938746213912964, "learning_rate": 1.48383963740956e-06, "loss": 0.4812, "step": 18030 }, { "epoch": 0.7558345472265596, "grad_norm": 1.787822961807251, "learning_rate": 1.483357032737429e-06, "loss": 0.4756, "step": 18031 }, { "epoch": 0.7558764658415686, "grad_norm": 1.6482068300247192, "learning_rate": 1.4828744928897486e-06, "loss": 0.3988, "step": 18032 }, { "epoch": 0.7559183844565776, "grad_norm": 1.8557597398757935, "learning_rate": 1.4823920178754148e-06, "loss": 0.4706, "step": 18033 }, { "epoch": 0.7559603030715866, "grad_norm": 2.018857002258301, "learning_rate": 1.4819096077033223e-06, "loss": 0.4686, "step": 18034 }, { "epoch": 0.7560022216865955, "grad_norm": 2.06592059135437, "learning_rate": 1.4814272623823616e-06, "loss": 0.4394, "step": 18035 }, { "epoch": 0.7560441403016044, "grad_norm": 2.0684943199157715, "learning_rate": 1.4809449819214243e-06, "loss": 0.4951, "step": 18036 }, { "epoch": 0.7560860589166134, "grad_norm": 1.8257330656051636, "learning_rate": 1.4804627663294036e-06, "loss": 0.4886, "step": 18037 }, { "epoch": 0.7561279775316223, "grad_norm": 2.2789251804351807, "learning_rate": 1.4799806156151857e-06, "loss": 0.4554, "step": 18038 }, { "epoch": 0.7561698961466313, "grad_norm": 1.8499740362167358, "learning_rate": 1.4794985297876575e-06, "loss": 0.4493, "step": 18039 }, { "epoch": 0.7562118147616402, "grad_norm": 2.3037376403808594, "learning_rate": 1.4790165088557084e-06, "loss": 0.5172, "step": 18040 }, { "epoch": 0.7562537333766493, "grad_norm": 4.070533752441406, "learning_rate": 1.4785345528282213e-06, "loss": 0.5152, "step": 18041 }, { "epoch": 0.7562956519916582, "grad_norm": 1.9742281436920166, "learning_rate": 1.4780526617140805e-06, "loss": 0.4662, "step": 18042 }, { "epoch": 0.7563375706066672, "grad_norm": 2.5125539302825928, "learning_rate": 1.4775708355221723e-06, "loss": 0.4844, "step": 18043 }, { "epoch": 0.7563794892216761, "grad_norm": 2.4254491329193115, "learning_rate": 1.4770890742613742e-06, "loss": 0.4665, "step": 18044 }, { "epoch": 0.7564214078366851, "grad_norm": 2.0126798152923584, "learning_rate": 1.4766073779405688e-06, "loss": 0.464, "step": 18045 }, { "epoch": 0.756463326451694, "grad_norm": 1.7883813381195068, "learning_rate": 1.4761257465686373e-06, "loss": 0.4368, "step": 18046 }, { "epoch": 0.756505245066703, "grad_norm": 1.9172313213348389, "learning_rate": 1.4756441801544557e-06, "loss": 0.497, "step": 18047 }, { "epoch": 0.7565471636817119, "grad_norm": 1.7941272258758545, "learning_rate": 1.4751626787068996e-06, "loss": 0.4768, "step": 18048 }, { "epoch": 0.756589082296721, "grad_norm": 2.35363507270813, "learning_rate": 1.474681242234849e-06, "loss": 0.4656, "step": 18049 }, { "epoch": 0.7566310009117299, "grad_norm": 2.0151188373565674, "learning_rate": 1.4741998707471732e-06, "loss": 0.5108, "step": 18050 }, { "epoch": 0.7566729195267389, "grad_norm": 2.6411120891571045, "learning_rate": 1.4737185642527495e-06, "loss": 0.5521, "step": 18051 }, { "epoch": 0.7567148381417478, "grad_norm": 2.33780574798584, "learning_rate": 1.4732373227604507e-06, "loss": 0.4589, "step": 18052 }, { "epoch": 0.7567567567567568, "grad_norm": 1.675826907157898, "learning_rate": 1.4727561462791446e-06, "loss": 0.4522, "step": 18053 }, { "epoch": 0.7567986753717657, "grad_norm": 2.3695454597473145, "learning_rate": 1.472275034817704e-06, "loss": 0.3995, "step": 18054 }, { "epoch": 0.7568405939867747, "grad_norm": 1.657683253288269, "learning_rate": 1.4717939883849947e-06, "loss": 0.5155, "step": 18055 }, { "epoch": 0.7568825126017836, "grad_norm": 2.2386274337768555, "learning_rate": 1.4713130069898867e-06, "loss": 0.5069, "step": 18056 }, { "epoch": 0.7569244312167926, "grad_norm": 1.8396433591842651, "learning_rate": 1.4708320906412443e-06, "loss": 0.4937, "step": 18057 }, { "epoch": 0.7569663498318016, "grad_norm": 2.545102119445801, "learning_rate": 1.4703512393479346e-06, "loss": 0.4723, "step": 18058 }, { "epoch": 0.7570082684468106, "grad_norm": 2.309913396835327, "learning_rate": 1.469870453118819e-06, "loss": 0.5053, "step": 18059 }, { "epoch": 0.7570501870618195, "grad_norm": 2.00697660446167, "learning_rate": 1.4693897319627626e-06, "loss": 0.5076, "step": 18060 }, { "epoch": 0.7570921056768284, "grad_norm": 1.7153712511062622, "learning_rate": 1.4689090758886244e-06, "loss": 0.4454, "step": 18061 }, { "epoch": 0.7571340242918374, "grad_norm": 1.8140454292297363, "learning_rate": 1.4684284849052654e-06, "loss": 0.5058, "step": 18062 }, { "epoch": 0.7571759429068463, "grad_norm": 2.0021533966064453, "learning_rate": 1.4679479590215468e-06, "loss": 0.4463, "step": 18063 }, { "epoch": 0.7572178615218553, "grad_norm": 1.682137370109558, "learning_rate": 1.4674674982463227e-06, "loss": 0.4663, "step": 18064 }, { "epoch": 0.7572597801368642, "grad_norm": 1.8039214611053467, "learning_rate": 1.466987102588452e-06, "loss": 0.5208, "step": 18065 }, { "epoch": 0.7573016987518733, "grad_norm": 1.5925935506820679, "learning_rate": 1.4665067720567915e-06, "loss": 0.4938, "step": 18066 }, { "epoch": 0.7573436173668822, "grad_norm": 2.139225959777832, "learning_rate": 1.4660265066601936e-06, "loss": 0.4839, "step": 18067 }, { "epoch": 0.7573855359818912, "grad_norm": 1.8500298261642456, "learning_rate": 1.46554630640751e-06, "loss": 0.4693, "step": 18068 }, { "epoch": 0.7574274545969001, "grad_norm": 2.382877826690674, "learning_rate": 1.4650661713075953e-06, "loss": 0.4829, "step": 18069 }, { "epoch": 0.7574693732119091, "grad_norm": 1.6818856000900269, "learning_rate": 1.4645861013692975e-06, "loss": 0.4221, "step": 18070 }, { "epoch": 0.757511291826918, "grad_norm": 1.9635450839996338, "learning_rate": 1.464106096601467e-06, "loss": 0.468, "step": 18071 }, { "epoch": 0.757553210441927, "grad_norm": 1.8085753917694092, "learning_rate": 1.4636261570129545e-06, "loss": 0.4666, "step": 18072 }, { "epoch": 0.7575951290569359, "grad_norm": 2.2067456245422363, "learning_rate": 1.463146282612603e-06, "loss": 0.5316, "step": 18073 }, { "epoch": 0.757637047671945, "grad_norm": 1.9521682262420654, "learning_rate": 1.4626664734092605e-06, "loss": 0.4626, "step": 18074 }, { "epoch": 0.7576789662869539, "grad_norm": 1.890811800956726, "learning_rate": 1.4621867294117726e-06, "loss": 0.5037, "step": 18075 }, { "epoch": 0.7577208849019629, "grad_norm": 1.7238503694534302, "learning_rate": 1.4617070506289815e-06, "loss": 0.4343, "step": 18076 }, { "epoch": 0.7577628035169718, "grad_norm": 2.222538709640503, "learning_rate": 1.4612274370697282e-06, "loss": 0.4762, "step": 18077 }, { "epoch": 0.7578047221319808, "grad_norm": 2.0020854473114014, "learning_rate": 1.460747888742856e-06, "loss": 0.4959, "step": 18078 }, { "epoch": 0.7578466407469897, "grad_norm": 1.578396201133728, "learning_rate": 1.460268405657202e-06, "loss": 0.4327, "step": 18079 }, { "epoch": 0.7578885593619987, "grad_norm": 2.1564595699310303, "learning_rate": 1.459788987821607e-06, "loss": 0.5489, "step": 18080 }, { "epoch": 0.7579304779770076, "grad_norm": 1.7356616258621216, "learning_rate": 1.4593096352449088e-06, "loss": 0.4372, "step": 18081 }, { "epoch": 0.7579723965920167, "grad_norm": 1.5997426509857178, "learning_rate": 1.4588303479359411e-06, "loss": 0.4428, "step": 18082 }, { "epoch": 0.7580143152070256, "grad_norm": 1.8441526889801025, "learning_rate": 1.4583511259035404e-06, "loss": 0.47, "step": 18083 }, { "epoch": 0.7580562338220346, "grad_norm": 3.206393003463745, "learning_rate": 1.4578719691565424e-06, "loss": 0.4657, "step": 18084 }, { "epoch": 0.7580981524370435, "grad_norm": 4.384908676147461, "learning_rate": 1.4573928777037778e-06, "loss": 0.4954, "step": 18085 }, { "epoch": 0.7581400710520524, "grad_norm": 2.1912529468536377, "learning_rate": 1.4569138515540766e-06, "loss": 0.4923, "step": 18086 }, { "epoch": 0.7581819896670614, "grad_norm": 1.5705622434616089, "learning_rate": 1.4564348907162713e-06, "loss": 0.5037, "step": 18087 }, { "epoch": 0.7582239082820703, "grad_norm": 2.7584311962127686, "learning_rate": 1.4559559951991885e-06, "loss": 0.4168, "step": 18088 }, { "epoch": 0.7582658268970793, "grad_norm": 2.0031988620758057, "learning_rate": 1.4554771650116579e-06, "loss": 0.4909, "step": 18089 }, { "epoch": 0.7583077455120882, "grad_norm": 1.9447717666625977, "learning_rate": 1.4549984001625066e-06, "loss": 0.4861, "step": 18090 }, { "epoch": 0.7583496641270973, "grad_norm": 2.191890239715576, "learning_rate": 1.4545197006605577e-06, "loss": 0.5736, "step": 18091 }, { "epoch": 0.7583915827421062, "grad_norm": 1.8581713438034058, "learning_rate": 1.454041066514637e-06, "loss": 0.4713, "step": 18092 }, { "epoch": 0.7584335013571152, "grad_norm": 1.8168543577194214, "learning_rate": 1.4535624977335682e-06, "loss": 0.5259, "step": 18093 }, { "epoch": 0.7584754199721241, "grad_norm": 1.8595126867294312, "learning_rate": 1.4530839943261721e-06, "loss": 0.4805, "step": 18094 }, { "epoch": 0.7585173385871331, "grad_norm": 1.8267253637313843, "learning_rate": 1.4526055563012675e-06, "loss": 0.5089, "step": 18095 }, { "epoch": 0.758559257202142, "grad_norm": 1.7893409729003906, "learning_rate": 1.452127183667677e-06, "loss": 0.4742, "step": 18096 }, { "epoch": 0.758601175817151, "grad_norm": 2.2386341094970703, "learning_rate": 1.4516488764342152e-06, "loss": 0.5295, "step": 18097 }, { "epoch": 0.7586430944321599, "grad_norm": 2.027827024459839, "learning_rate": 1.4511706346097015e-06, "loss": 0.5281, "step": 18098 }, { "epoch": 0.758685013047169, "grad_norm": 2.058647394180298, "learning_rate": 1.450692458202952e-06, "loss": 0.4548, "step": 18099 }, { "epoch": 0.7587269316621779, "grad_norm": 2.1152281761169434, "learning_rate": 1.4502143472227791e-06, "loss": 0.4226, "step": 18100 }, { "epoch": 0.7587688502771869, "grad_norm": 1.655518889427185, "learning_rate": 1.4497363016779969e-06, "loss": 0.442, "step": 18101 }, { "epoch": 0.7588107688921958, "grad_norm": 1.9617938995361328, "learning_rate": 1.4492583215774198e-06, "loss": 0.4325, "step": 18102 }, { "epoch": 0.7588526875072048, "grad_norm": 1.8148572444915771, "learning_rate": 1.4487804069298568e-06, "loss": 0.4425, "step": 18103 }, { "epoch": 0.7588946061222137, "grad_norm": 2.001014471054077, "learning_rate": 1.4483025577441157e-06, "loss": 0.488, "step": 18104 }, { "epoch": 0.7589365247372227, "grad_norm": 1.6752883195877075, "learning_rate": 1.4478247740290086e-06, "loss": 0.4891, "step": 18105 }, { "epoch": 0.7589784433522316, "grad_norm": 2.0915815830230713, "learning_rate": 1.4473470557933395e-06, "loss": 0.5131, "step": 18106 }, { "epoch": 0.7590203619672407, "grad_norm": 1.9518861770629883, "learning_rate": 1.4468694030459163e-06, "loss": 0.4668, "step": 18107 }, { "epoch": 0.7590622805822496, "grad_norm": 2.0348727703094482, "learning_rate": 1.4463918157955449e-06, "loss": 0.5228, "step": 18108 }, { "epoch": 0.7591041991972586, "grad_norm": 1.8781354427337646, "learning_rate": 1.4459142940510263e-06, "loss": 0.4301, "step": 18109 }, { "epoch": 0.7591461178122675, "grad_norm": 1.7382868528366089, "learning_rate": 1.4454368378211636e-06, "loss": 0.5095, "step": 18110 }, { "epoch": 0.7591880364272764, "grad_norm": 2.400799512863159, "learning_rate": 1.444959447114761e-06, "loss": 0.4461, "step": 18111 }, { "epoch": 0.7592299550422854, "grad_norm": 6.901335716247559, "learning_rate": 1.4444821219406146e-06, "loss": 0.4268, "step": 18112 }, { "epoch": 0.7592718736572943, "grad_norm": 1.7413856983184814, "learning_rate": 1.4440048623075264e-06, "loss": 0.4798, "step": 18113 }, { "epoch": 0.7593137922723033, "grad_norm": 1.9425870180130005, "learning_rate": 1.4435276682242927e-06, "loss": 0.5403, "step": 18114 }, { "epoch": 0.7593557108873122, "grad_norm": 2.1922032833099365, "learning_rate": 1.443050539699708e-06, "loss": 0.5045, "step": 18115 }, { "epoch": 0.7593976295023213, "grad_norm": 4.7830657958984375, "learning_rate": 1.4425734767425697e-06, "loss": 0.4612, "step": 18116 }, { "epoch": 0.7594395481173302, "grad_norm": 1.9538946151733398, "learning_rate": 1.4420964793616727e-06, "loss": 0.4629, "step": 18117 }, { "epoch": 0.7594814667323392, "grad_norm": 1.7178417444229126, "learning_rate": 1.4416195475658063e-06, "loss": 0.4564, "step": 18118 }, { "epoch": 0.7595233853473481, "grad_norm": 1.9618852138519287, "learning_rate": 1.4411426813637646e-06, "loss": 0.4827, "step": 18119 }, { "epoch": 0.7595653039623571, "grad_norm": 2.2890470027923584, "learning_rate": 1.4406658807643398e-06, "loss": 0.5042, "step": 18120 }, { "epoch": 0.759607222577366, "grad_norm": 2.1707534790039062, "learning_rate": 1.4401891457763168e-06, "loss": 0.5015, "step": 18121 }, { "epoch": 0.759649141192375, "grad_norm": 2.0551722049713135, "learning_rate": 1.4397124764084868e-06, "loss": 0.4921, "step": 18122 }, { "epoch": 0.7596910598073839, "grad_norm": 2.2603652477264404, "learning_rate": 1.439235872669636e-06, "loss": 0.5599, "step": 18123 }, { "epoch": 0.759732978422393, "grad_norm": 1.811165690422058, "learning_rate": 1.4387593345685475e-06, "loss": 0.4854, "step": 18124 }, { "epoch": 0.7597748970374019, "grad_norm": 1.754988670349121, "learning_rate": 1.4382828621140072e-06, "loss": 0.5169, "step": 18125 }, { "epoch": 0.7598168156524109, "grad_norm": 1.7368252277374268, "learning_rate": 1.4378064553147996e-06, "loss": 0.465, "step": 18126 }, { "epoch": 0.7598587342674198, "grad_norm": 1.8968391418457031, "learning_rate": 1.4373301141797048e-06, "loss": 0.4732, "step": 18127 }, { "epoch": 0.7599006528824288, "grad_norm": 2.031317949295044, "learning_rate": 1.436853838717505e-06, "loss": 0.4504, "step": 18128 }, { "epoch": 0.7599425714974377, "grad_norm": 2.038971424102783, "learning_rate": 1.4363776289369773e-06, "loss": 0.4673, "step": 18129 }, { "epoch": 0.7599844901124467, "grad_norm": 2.48712420463562, "learning_rate": 1.4359014848469016e-06, "loss": 0.4977, "step": 18130 }, { "epoch": 0.7600264087274556, "grad_norm": 2.261615037918091, "learning_rate": 1.4354254064560564e-06, "loss": 0.4514, "step": 18131 }, { "epoch": 0.7600683273424647, "grad_norm": 1.7542613744735718, "learning_rate": 1.4349493937732157e-06, "loss": 0.4582, "step": 18132 }, { "epoch": 0.7601102459574736, "grad_norm": 1.921301245689392, "learning_rate": 1.434473446807153e-06, "loss": 0.4787, "step": 18133 }, { "epoch": 0.7601521645724826, "grad_norm": 2.522662401199341, "learning_rate": 1.4339975655666448e-06, "loss": 0.4943, "step": 18134 }, { "epoch": 0.7601940831874915, "grad_norm": 1.6567283868789673, "learning_rate": 1.4335217500604598e-06, "loss": 0.5222, "step": 18135 }, { "epoch": 0.7602360018025004, "grad_norm": 2.1357903480529785, "learning_rate": 1.4330460002973711e-06, "loss": 0.5658, "step": 18136 }, { "epoch": 0.7602779204175094, "grad_norm": 2.3316333293914795, "learning_rate": 1.43257031628615e-06, "loss": 0.4453, "step": 18137 }, { "epoch": 0.7603198390325183, "grad_norm": 1.729269027709961, "learning_rate": 1.432094698035561e-06, "loss": 0.464, "step": 18138 }, { "epoch": 0.7603617576475273, "grad_norm": 2.013442039489746, "learning_rate": 1.4316191455543742e-06, "loss": 0.5191, "step": 18139 }, { "epoch": 0.7604036762625362, "grad_norm": 1.9118080139160156, "learning_rate": 1.4311436588513572e-06, "loss": 0.473, "step": 18140 }, { "epoch": 0.7604455948775453, "grad_norm": 1.6551979780197144, "learning_rate": 1.4306682379352726e-06, "loss": 0.4445, "step": 18141 }, { "epoch": 0.7604875134925542, "grad_norm": 2.0318918228149414, "learning_rate": 1.430192882814883e-06, "loss": 0.5197, "step": 18142 }, { "epoch": 0.7605294321075632, "grad_norm": 1.958431363105774, "learning_rate": 1.4297175934989543e-06, "loss": 0.4895, "step": 18143 }, { "epoch": 0.7605713507225721, "grad_norm": 2.3040497303009033, "learning_rate": 1.4292423699962442e-06, "loss": 0.4859, "step": 18144 }, { "epoch": 0.7606132693375811, "grad_norm": 2.365633726119995, "learning_rate": 1.4287672123155143e-06, "loss": 0.5357, "step": 18145 }, { "epoch": 0.76065518795259, "grad_norm": 2.031118154525757, "learning_rate": 1.4282921204655259e-06, "loss": 0.5169, "step": 18146 }, { "epoch": 0.760697106567599, "grad_norm": 2.0145041942596436, "learning_rate": 1.427817094455033e-06, "loss": 0.4479, "step": 18147 }, { "epoch": 0.7607390251826079, "grad_norm": 2.9566919803619385, "learning_rate": 1.4273421342927935e-06, "loss": 0.4719, "step": 18148 }, { "epoch": 0.760780943797617, "grad_norm": 1.789311170578003, "learning_rate": 1.4268672399875639e-06, "loss": 0.5016, "step": 18149 }, { "epoch": 0.7608228624126259, "grad_norm": 2.030374050140381, "learning_rate": 1.4263924115480971e-06, "loss": 0.5016, "step": 18150 }, { "epoch": 0.7608647810276349, "grad_norm": 1.6165046691894531, "learning_rate": 1.4259176489831444e-06, "loss": 0.4537, "step": 18151 }, { "epoch": 0.7609066996426438, "grad_norm": 1.715909481048584, "learning_rate": 1.4254429523014606e-06, "loss": 0.5008, "step": 18152 }, { "epoch": 0.7609486182576528, "grad_norm": 2.034388780593872, "learning_rate": 1.424968321511792e-06, "loss": 0.4048, "step": 18153 }, { "epoch": 0.7609905368726617, "grad_norm": 1.5107836723327637, "learning_rate": 1.4244937566228912e-06, "loss": 0.4172, "step": 18154 }, { "epoch": 0.7610324554876707, "grad_norm": 1.8222593069076538, "learning_rate": 1.4240192576435059e-06, "loss": 0.5232, "step": 18155 }, { "epoch": 0.7610743741026796, "grad_norm": 1.8297797441482544, "learning_rate": 1.4235448245823802e-06, "loss": 0.4464, "step": 18156 }, { "epoch": 0.7611162927176887, "grad_norm": 2.027771234512329, "learning_rate": 1.4230704574482618e-06, "loss": 0.4988, "step": 18157 }, { "epoch": 0.7611582113326976, "grad_norm": 1.5632582902908325, "learning_rate": 1.4225961562498959e-06, "loss": 0.4202, "step": 18158 }, { "epoch": 0.7612001299477066, "grad_norm": 1.7857117652893066, "learning_rate": 1.4221219209960225e-06, "loss": 0.4532, "step": 18159 }, { "epoch": 0.7612420485627155, "grad_norm": 1.9776172637939453, "learning_rate": 1.421647751695387e-06, "loss": 0.4994, "step": 18160 }, { "epoch": 0.7612839671777244, "grad_norm": 1.8802833557128906, "learning_rate": 1.4211736483567279e-06, "loss": 0.4986, "step": 18161 }, { "epoch": 0.7613258857927334, "grad_norm": 1.9304418563842773, "learning_rate": 1.4206996109887838e-06, "loss": 0.4768, "step": 18162 }, { "epoch": 0.7613678044077423, "grad_norm": 2.374056816101074, "learning_rate": 1.4202256396002945e-06, "loss": 0.4853, "step": 18163 }, { "epoch": 0.7614097230227513, "grad_norm": 1.7503560781478882, "learning_rate": 1.4197517341999978e-06, "loss": 0.4726, "step": 18164 }, { "epoch": 0.7614516416377602, "grad_norm": 1.6845735311508179, "learning_rate": 1.4192778947966268e-06, "loss": 0.4231, "step": 18165 }, { "epoch": 0.7614935602527693, "grad_norm": 1.972002387046814, "learning_rate": 1.4188041213989173e-06, "loss": 0.4784, "step": 18166 }, { "epoch": 0.7615354788677782, "grad_norm": 1.6588294506072998, "learning_rate": 1.4183304140156051e-06, "loss": 0.4744, "step": 18167 }, { "epoch": 0.7615773974827872, "grad_norm": 1.8359973430633545, "learning_rate": 1.4178567726554181e-06, "loss": 0.5155, "step": 18168 }, { "epoch": 0.7616193160977961, "grad_norm": 1.6906399726867676, "learning_rate": 1.4173831973270912e-06, "loss": 0.4899, "step": 18169 }, { "epoch": 0.7616612347128051, "grad_norm": 1.909520149230957, "learning_rate": 1.4169096880393524e-06, "loss": 0.4581, "step": 18170 }, { "epoch": 0.761703153327814, "grad_norm": 1.9892228841781616, "learning_rate": 1.4164362448009284e-06, "loss": 0.4613, "step": 18171 }, { "epoch": 0.761745071942823, "grad_norm": 1.7742525339126587, "learning_rate": 1.4159628676205477e-06, "loss": 0.5233, "step": 18172 }, { "epoch": 0.761786990557832, "grad_norm": 1.9236482381820679, "learning_rate": 1.4154895565069387e-06, "loss": 0.444, "step": 18173 }, { "epoch": 0.761828909172841, "grad_norm": 2.2985148429870605, "learning_rate": 1.4150163114688226e-06, "loss": 0.4611, "step": 18174 }, { "epoch": 0.7618708277878499, "grad_norm": 1.6750209331512451, "learning_rate": 1.414543132514925e-06, "loss": 0.4866, "step": 18175 }, { "epoch": 0.7619127464028589, "grad_norm": 1.7878518104553223, "learning_rate": 1.414070019653969e-06, "loss": 0.4647, "step": 18176 }, { "epoch": 0.7619546650178678, "grad_norm": 1.6954150199890137, "learning_rate": 1.4135969728946736e-06, "loss": 0.4653, "step": 18177 }, { "epoch": 0.7619965836328768, "grad_norm": 2.090283155441284, "learning_rate": 1.4131239922457612e-06, "loss": 0.4127, "step": 18178 }, { "epoch": 0.7620385022478857, "grad_norm": 3.7891783714294434, "learning_rate": 1.4126510777159497e-06, "loss": 0.5018, "step": 18179 }, { "epoch": 0.7620804208628947, "grad_norm": 2.258354663848877, "learning_rate": 1.4121782293139546e-06, "loss": 0.4629, "step": 18180 }, { "epoch": 0.7621223394779036, "grad_norm": 1.762879490852356, "learning_rate": 1.4117054470484936e-06, "loss": 0.4945, "step": 18181 }, { "epoch": 0.7621642580929127, "grad_norm": 2.9146950244903564, "learning_rate": 1.4112327309282836e-06, "loss": 0.5361, "step": 18182 }, { "epoch": 0.7622061767079216, "grad_norm": 3.195727586746216, "learning_rate": 1.4107600809620358e-06, "loss": 0.4183, "step": 18183 }, { "epoch": 0.7622480953229306, "grad_norm": 1.8412531614303589, "learning_rate": 1.4102874971584635e-06, "loss": 0.5154, "step": 18184 }, { "epoch": 0.7622900139379395, "grad_norm": 1.9000952243804932, "learning_rate": 1.40981497952628e-06, "loss": 0.4601, "step": 18185 }, { "epoch": 0.7623319325529484, "grad_norm": 2.2429306507110596, "learning_rate": 1.409342528074193e-06, "loss": 0.5312, "step": 18186 }, { "epoch": 0.7623738511679574, "grad_norm": 1.7744427919387817, "learning_rate": 1.4088701428109136e-06, "loss": 0.4809, "step": 18187 }, { "epoch": 0.7624157697829663, "grad_norm": 1.7401947975158691, "learning_rate": 1.408397823745149e-06, "loss": 0.516, "step": 18188 }, { "epoch": 0.7624576883979753, "grad_norm": 2.48685359954834, "learning_rate": 1.4079255708856031e-06, "loss": 0.4582, "step": 18189 }, { "epoch": 0.7624996070129842, "grad_norm": 2.038363456726074, "learning_rate": 1.4074533842409838e-06, "loss": 0.4695, "step": 18190 }, { "epoch": 0.7625415256279933, "grad_norm": 1.578703761100769, "learning_rate": 1.4069812638199964e-06, "loss": 0.4284, "step": 18191 }, { "epoch": 0.7625834442430022, "grad_norm": 2.2071950435638428, "learning_rate": 1.4065092096313405e-06, "loss": 0.5048, "step": 18192 }, { "epoch": 0.7626253628580112, "grad_norm": 2.153773784637451, "learning_rate": 1.40603722168372e-06, "loss": 0.5025, "step": 18193 }, { "epoch": 0.7626672814730201, "grad_norm": 2.0303521156311035, "learning_rate": 1.4055652999858355e-06, "loss": 0.5302, "step": 18194 }, { "epoch": 0.7627092000880291, "grad_norm": 2.5637118816375732, "learning_rate": 1.4050934445463844e-06, "loss": 0.4882, "step": 18195 }, { "epoch": 0.762751118703038, "grad_norm": 1.8014646768569946, "learning_rate": 1.4046216553740677e-06, "loss": 0.4343, "step": 18196 }, { "epoch": 0.762793037318047, "grad_norm": 2.000208854675293, "learning_rate": 1.4041499324775797e-06, "loss": 0.5151, "step": 18197 }, { "epoch": 0.762834955933056, "grad_norm": 1.6612290143966675, "learning_rate": 1.4036782758656152e-06, "loss": 0.4514, "step": 18198 }, { "epoch": 0.762876874548065, "grad_norm": 2.094619035720825, "learning_rate": 1.40320668554687e-06, "loss": 0.4984, "step": 18199 }, { "epoch": 0.7629187931630739, "grad_norm": 2.073138952255249, "learning_rate": 1.4027351615300388e-06, "loss": 0.5065, "step": 18200 }, { "epoch": 0.7629607117780829, "grad_norm": 1.8712162971496582, "learning_rate": 1.4022637038238107e-06, "loss": 0.5246, "step": 18201 }, { "epoch": 0.7630026303930918, "grad_norm": 1.8340362310409546, "learning_rate": 1.4017923124368781e-06, "loss": 0.4634, "step": 18202 }, { "epoch": 0.7630445490081008, "grad_norm": 2.868757963180542, "learning_rate": 1.4013209873779288e-06, "loss": 0.4797, "step": 18203 }, { "epoch": 0.7630864676231097, "grad_norm": 1.9196043014526367, "learning_rate": 1.4008497286556522e-06, "loss": 0.5304, "step": 18204 }, { "epoch": 0.7631283862381187, "grad_norm": 1.804999828338623, "learning_rate": 1.4003785362787363e-06, "loss": 0.4615, "step": 18205 }, { "epoch": 0.7631703048531276, "grad_norm": 1.7335609197616577, "learning_rate": 1.3999074102558646e-06, "loss": 0.4792, "step": 18206 }, { "epoch": 0.7632122234681367, "grad_norm": 2.185889482498169, "learning_rate": 1.3994363505957242e-06, "loss": 0.4855, "step": 18207 }, { "epoch": 0.7632541420831456, "grad_norm": 1.7061045169830322, "learning_rate": 1.3989653573069972e-06, "loss": 0.4885, "step": 18208 }, { "epoch": 0.7632960606981546, "grad_norm": 1.598589301109314, "learning_rate": 1.3984944303983638e-06, "loss": 0.4547, "step": 18209 }, { "epoch": 0.7633379793131635, "grad_norm": 2.518885374069214, "learning_rate": 1.3980235698785073e-06, "loss": 0.5201, "step": 18210 }, { "epoch": 0.7633798979281724, "grad_norm": 1.7426416873931885, "learning_rate": 1.3975527757561074e-06, "loss": 0.4874, "step": 18211 }, { "epoch": 0.7634218165431814, "grad_norm": 2.420944929122925, "learning_rate": 1.3970820480398406e-06, "loss": 0.4833, "step": 18212 }, { "epoch": 0.7634637351581903, "grad_norm": 2.136011838912964, "learning_rate": 1.3966113867383852e-06, "loss": 0.465, "step": 18213 }, { "epoch": 0.7635056537731993, "grad_norm": 1.5883522033691406, "learning_rate": 1.3961407918604192e-06, "loss": 0.474, "step": 18214 }, { "epoch": 0.7635475723882083, "grad_norm": 1.707960605621338, "learning_rate": 1.395670263414614e-06, "loss": 0.4738, "step": 18215 }, { "epoch": 0.7635894910032173, "grad_norm": 1.8637497425079346, "learning_rate": 1.3951998014096457e-06, "loss": 0.4696, "step": 18216 }, { "epoch": 0.7636314096182262, "grad_norm": 2.731238603591919, "learning_rate": 1.394729405854186e-06, "loss": 0.4591, "step": 18217 }, { "epoch": 0.7636733282332352, "grad_norm": 1.5415494441986084, "learning_rate": 1.3942590767569041e-06, "loss": 0.4517, "step": 18218 }, { "epoch": 0.7637152468482441, "grad_norm": 1.7181390523910522, "learning_rate": 1.393788814126471e-06, "loss": 0.449, "step": 18219 }, { "epoch": 0.7637571654632531, "grad_norm": 1.9931674003601074, "learning_rate": 1.3933186179715575e-06, "loss": 0.4977, "step": 18220 }, { "epoch": 0.763799084078262, "grad_norm": 1.93375825881958, "learning_rate": 1.3928484883008275e-06, "loss": 0.4778, "step": 18221 }, { "epoch": 0.763841002693271, "grad_norm": 2.1153016090393066, "learning_rate": 1.392378425122949e-06, "loss": 0.4497, "step": 18222 }, { "epoch": 0.76388292130828, "grad_norm": 1.6868336200714111, "learning_rate": 1.3919084284465878e-06, "loss": 0.5554, "step": 18223 }, { "epoch": 0.763924839923289, "grad_norm": 1.8943854570388794, "learning_rate": 1.3914384982804058e-06, "loss": 0.4949, "step": 18224 }, { "epoch": 0.7639667585382979, "grad_norm": 2.023756504058838, "learning_rate": 1.390968634633068e-06, "loss": 0.4928, "step": 18225 }, { "epoch": 0.7640086771533069, "grad_norm": 1.7822890281677246, "learning_rate": 1.390498837513234e-06, "loss": 0.4375, "step": 18226 }, { "epoch": 0.7640505957683158, "grad_norm": 1.9526804685592651, "learning_rate": 1.3900291069295619e-06, "loss": 0.4986, "step": 18227 }, { "epoch": 0.7640925143833248, "grad_norm": 1.7391358613967896, "learning_rate": 1.3895594428907128e-06, "loss": 0.4851, "step": 18228 }, { "epoch": 0.7641344329983337, "grad_norm": 2.416041612625122, "learning_rate": 1.3890898454053458e-06, "loss": 0.4807, "step": 18229 }, { "epoch": 0.7641763516133427, "grad_norm": 1.6427100896835327, "learning_rate": 1.3886203144821141e-06, "loss": 0.4198, "step": 18230 }, { "epoch": 0.7642182702283516, "grad_norm": 1.8920726776123047, "learning_rate": 1.3881508501296743e-06, "loss": 0.5093, "step": 18231 }, { "epoch": 0.7642601888433607, "grad_norm": 1.8641338348388672, "learning_rate": 1.3876814523566817e-06, "loss": 0.457, "step": 18232 }, { "epoch": 0.7643021074583696, "grad_norm": 1.8658051490783691, "learning_rate": 1.3872121211717864e-06, "loss": 0.4669, "step": 18233 }, { "epoch": 0.7643440260733786, "grad_norm": 1.8752326965332031, "learning_rate": 1.3867428565836421e-06, "loss": 0.5015, "step": 18234 }, { "epoch": 0.7643859446883875, "grad_norm": 1.6622624397277832, "learning_rate": 1.3862736586008984e-06, "loss": 0.4467, "step": 18235 }, { "epoch": 0.7644278633033964, "grad_norm": 2.122097969055176, "learning_rate": 1.385804527232203e-06, "loss": 0.5622, "step": 18236 }, { "epoch": 0.7644697819184054, "grad_norm": 1.8770804405212402, "learning_rate": 1.3853354624862037e-06, "loss": 0.49, "step": 18237 }, { "epoch": 0.7645117005334143, "grad_norm": 2.21454119682312, "learning_rate": 1.38486646437155e-06, "loss": 0.5023, "step": 18238 }, { "epoch": 0.7645536191484233, "grad_norm": 1.8317012786865234, "learning_rate": 1.384397532896884e-06, "loss": 0.4222, "step": 18239 }, { "epoch": 0.7645955377634323, "grad_norm": 2.4782683849334717, "learning_rate": 1.3839286680708507e-06, "loss": 0.4799, "step": 18240 }, { "epoch": 0.7646374563784413, "grad_norm": 2.017103672027588, "learning_rate": 1.3834598699020945e-06, "loss": 0.4812, "step": 18241 }, { "epoch": 0.7646793749934502, "grad_norm": 1.7147517204284668, "learning_rate": 1.382991138399255e-06, "loss": 0.5199, "step": 18242 }, { "epoch": 0.7647212936084592, "grad_norm": 1.6029983758926392, "learning_rate": 1.3825224735709747e-06, "loss": 0.4461, "step": 18243 }, { "epoch": 0.7647632122234681, "grad_norm": 1.64287531375885, "learning_rate": 1.3820538754258904e-06, "loss": 0.4793, "step": 18244 }, { "epoch": 0.7648051308384771, "grad_norm": 2.073148250579834, "learning_rate": 1.3815853439726423e-06, "loss": 0.4965, "step": 18245 }, { "epoch": 0.764847049453486, "grad_norm": 1.7176748514175415, "learning_rate": 1.381116879219865e-06, "loss": 0.451, "step": 18246 }, { "epoch": 0.764888968068495, "grad_norm": 2.2535810470581055, "learning_rate": 1.3806484811761966e-06, "loss": 0.4713, "step": 18247 }, { "epoch": 0.764930886683504, "grad_norm": 2.001138687133789, "learning_rate": 1.3801801498502682e-06, "loss": 0.4495, "step": 18248 }, { "epoch": 0.764972805298513, "grad_norm": 1.700187087059021, "learning_rate": 1.379711885250714e-06, "loss": 0.4832, "step": 18249 }, { "epoch": 0.7650147239135219, "grad_norm": 2.50156831741333, "learning_rate": 1.379243687386168e-06, "loss": 0.4479, "step": 18250 }, { "epoch": 0.7650566425285309, "grad_norm": 1.9842381477355957, "learning_rate": 1.3787755562652582e-06, "loss": 0.4535, "step": 18251 }, { "epoch": 0.7650985611435398, "grad_norm": 1.7602343559265137, "learning_rate": 1.378307491896616e-06, "loss": 0.5097, "step": 18252 }, { "epoch": 0.7651404797585488, "grad_norm": 1.8445457220077515, "learning_rate": 1.3778394942888667e-06, "loss": 0.4717, "step": 18253 }, { "epoch": 0.7651823983735577, "grad_norm": 1.8344027996063232, "learning_rate": 1.3773715634506402e-06, "loss": 0.4816, "step": 18254 }, { "epoch": 0.7652243169885667, "grad_norm": 1.9459391832351685, "learning_rate": 1.3769036993905593e-06, "loss": 0.4581, "step": 18255 }, { "epoch": 0.7652662356035757, "grad_norm": 1.6602964401245117, "learning_rate": 1.3764359021172519e-06, "loss": 0.4678, "step": 18256 }, { "epoch": 0.7653081542185847, "grad_norm": 1.6438153982162476, "learning_rate": 1.3759681716393374e-06, "loss": 0.4418, "step": 18257 }, { "epoch": 0.7653500728335936, "grad_norm": 1.7180685997009277, "learning_rate": 1.37550050796544e-06, "loss": 0.4642, "step": 18258 }, { "epoch": 0.7653919914486026, "grad_norm": 1.5653955936431885, "learning_rate": 1.375032911104181e-06, "loss": 0.4393, "step": 18259 }, { "epoch": 0.7654339100636115, "grad_norm": 1.7739474773406982, "learning_rate": 1.3745653810641774e-06, "loss": 0.4273, "step": 18260 }, { "epoch": 0.7654758286786204, "grad_norm": 2.1986639499664307, "learning_rate": 1.3740979178540492e-06, "loss": 0.4829, "step": 18261 }, { "epoch": 0.7655177472936294, "grad_norm": 2.268324136734009, "learning_rate": 1.3736305214824146e-06, "loss": 0.4768, "step": 18262 }, { "epoch": 0.7655596659086383, "grad_norm": 1.9718531370162964, "learning_rate": 1.373163191957888e-06, "loss": 0.5228, "step": 18263 }, { "epoch": 0.7656015845236473, "grad_norm": 2.152543544769287, "learning_rate": 1.3726959292890823e-06, "loss": 0.5048, "step": 18264 }, { "epoch": 0.7656435031386563, "grad_norm": 1.9913954734802246, "learning_rate": 1.3722287334846146e-06, "loss": 0.4842, "step": 18265 }, { "epoch": 0.7656854217536653, "grad_norm": 2.0486834049224854, "learning_rate": 1.3717616045530934e-06, "loss": 0.5381, "step": 18266 }, { "epoch": 0.7657273403686742, "grad_norm": 2.0926032066345215, "learning_rate": 1.371294542503131e-06, "loss": 0.4552, "step": 18267 }, { "epoch": 0.7657692589836832, "grad_norm": 1.7503310441970825, "learning_rate": 1.3708275473433386e-06, "loss": 0.4748, "step": 18268 }, { "epoch": 0.7658111775986921, "grad_norm": 1.8597493171691895, "learning_rate": 1.3703606190823215e-06, "loss": 0.4863, "step": 18269 }, { "epoch": 0.7658530962137011, "grad_norm": 1.7833786010742188, "learning_rate": 1.3698937577286886e-06, "loss": 0.422, "step": 18270 }, { "epoch": 0.76589501482871, "grad_norm": 1.4756698608398438, "learning_rate": 1.3694269632910473e-06, "loss": 0.4663, "step": 18271 }, { "epoch": 0.765936933443719, "grad_norm": 1.7540009021759033, "learning_rate": 1.3689602357780012e-06, "loss": 0.4587, "step": 18272 }, { "epoch": 0.765978852058728, "grad_norm": 1.9815223217010498, "learning_rate": 1.3684935751981515e-06, "loss": 0.4633, "step": 18273 }, { "epoch": 0.766020770673737, "grad_norm": 1.7331457138061523, "learning_rate": 1.3680269815601044e-06, "loss": 0.5143, "step": 18274 }, { "epoch": 0.7660626892887459, "grad_norm": 1.427176833152771, "learning_rate": 1.367560454872457e-06, "loss": 0.4424, "step": 18275 }, { "epoch": 0.7661046079037549, "grad_norm": 1.641241431236267, "learning_rate": 1.3670939951438107e-06, "loss": 0.4345, "step": 18276 }, { "epoch": 0.7661465265187638, "grad_norm": 2.1090049743652344, "learning_rate": 1.3666276023827657e-06, "loss": 0.5028, "step": 18277 }, { "epoch": 0.7661884451337728, "grad_norm": 1.791218876838684, "learning_rate": 1.3661612765979166e-06, "loss": 0.4808, "step": 18278 }, { "epoch": 0.7662303637487817, "grad_norm": 1.4969629049301147, "learning_rate": 1.3656950177978622e-06, "loss": 0.4455, "step": 18279 }, { "epoch": 0.7662722823637907, "grad_norm": 1.7742284536361694, "learning_rate": 1.3652288259911945e-06, "loss": 0.4931, "step": 18280 }, { "epoch": 0.7663142009787997, "grad_norm": 1.97783362865448, "learning_rate": 1.36476270118651e-06, "loss": 0.4958, "step": 18281 }, { "epoch": 0.7663561195938087, "grad_norm": 1.7609158754348755, "learning_rate": 1.3642966433923992e-06, "loss": 0.446, "step": 18282 }, { "epoch": 0.7663980382088176, "grad_norm": 1.9332976341247559, "learning_rate": 1.3638306526174523e-06, "loss": 0.4762, "step": 18283 }, { "epoch": 0.7664399568238266, "grad_norm": 1.9887055158615112, "learning_rate": 1.3633647288702606e-06, "loss": 0.4275, "step": 18284 }, { "epoch": 0.7664818754388355, "grad_norm": 2.396620988845825, "learning_rate": 1.3628988721594138e-06, "loss": 0.542, "step": 18285 }, { "epoch": 0.7665237940538444, "grad_norm": 1.9099136590957642, "learning_rate": 1.3624330824934967e-06, "loss": 0.5063, "step": 18286 }, { "epoch": 0.7665657126688534, "grad_norm": 1.6196725368499756, "learning_rate": 1.3619673598810973e-06, "loss": 0.4756, "step": 18287 }, { "epoch": 0.7666076312838623, "grad_norm": 1.895580768585205, "learning_rate": 1.361501704330801e-06, "loss": 0.462, "step": 18288 }, { "epoch": 0.7666495498988714, "grad_norm": 2.2688064575195312, "learning_rate": 1.3610361158511898e-06, "loss": 0.5347, "step": 18289 }, { "epoch": 0.7666914685138803, "grad_norm": 1.7808833122253418, "learning_rate": 1.3605705944508485e-06, "loss": 0.4135, "step": 18290 }, { "epoch": 0.7667333871288893, "grad_norm": 1.7917929887771606, "learning_rate": 1.3601051401383553e-06, "loss": 0.4806, "step": 18291 }, { "epoch": 0.7667753057438982, "grad_norm": 2.31252121925354, "learning_rate": 1.3596397529222932e-06, "loss": 0.513, "step": 18292 }, { "epoch": 0.7668172243589072, "grad_norm": 2.601196765899658, "learning_rate": 1.3591744328112382e-06, "loss": 0.4865, "step": 18293 }, { "epoch": 0.7668591429739161, "grad_norm": 2.059843063354492, "learning_rate": 1.3587091798137714e-06, "loss": 0.5125, "step": 18294 }, { "epoch": 0.7669010615889251, "grad_norm": 1.6515114307403564, "learning_rate": 1.3582439939384644e-06, "loss": 0.5118, "step": 18295 }, { "epoch": 0.766942980203934, "grad_norm": 2.0993802547454834, "learning_rate": 1.3577788751938952e-06, "loss": 0.4796, "step": 18296 }, { "epoch": 0.766984898818943, "grad_norm": 1.695239543914795, "learning_rate": 1.3573138235886392e-06, "loss": 0.5255, "step": 18297 }, { "epoch": 0.767026817433952, "grad_norm": 1.8314732313156128, "learning_rate": 1.356848839131265e-06, "loss": 0.4759, "step": 18298 }, { "epoch": 0.767068736048961, "grad_norm": 2.1400444507598877, "learning_rate": 1.356383921830346e-06, "loss": 0.4326, "step": 18299 }, { "epoch": 0.7671106546639699, "grad_norm": 5.183689117431641, "learning_rate": 1.3559190716944538e-06, "loss": 0.4829, "step": 18300 }, { "epoch": 0.7671525732789789, "grad_norm": 1.524951457977295, "learning_rate": 1.355454288732156e-06, "loss": 0.4155, "step": 18301 }, { "epoch": 0.7671944918939878, "grad_norm": 1.8377273082733154, "learning_rate": 1.354989572952018e-06, "loss": 0.4388, "step": 18302 }, { "epoch": 0.7672364105089968, "grad_norm": 2.14518141746521, "learning_rate": 1.3545249243626096e-06, "loss": 0.4603, "step": 18303 }, { "epoch": 0.7672783291240057, "grad_norm": 1.7987233400344849, "learning_rate": 1.3540603429724936e-06, "loss": 0.4568, "step": 18304 }, { "epoch": 0.7673202477390147, "grad_norm": 3.928494930267334, "learning_rate": 1.353595828790234e-06, "loss": 0.4674, "step": 18305 }, { "epoch": 0.7673621663540237, "grad_norm": 1.8773285150527954, "learning_rate": 1.3531313818243962e-06, "loss": 0.4484, "step": 18306 }, { "epoch": 0.7674040849690327, "grad_norm": 1.7782917022705078, "learning_rate": 1.3526670020835386e-06, "loss": 0.4915, "step": 18307 }, { "epoch": 0.7674460035840416, "grad_norm": 1.6355563402175903, "learning_rate": 1.352202689576222e-06, "loss": 0.4982, "step": 18308 }, { "epoch": 0.7674879221990506, "grad_norm": 1.7478715181350708, "learning_rate": 1.3517384443110077e-06, "loss": 0.5406, "step": 18309 }, { "epoch": 0.7675298408140595, "grad_norm": 1.6888597011566162, "learning_rate": 1.3512742662964513e-06, "loss": 0.5582, "step": 18310 }, { "epoch": 0.7675717594290684, "grad_norm": 1.7429598569869995, "learning_rate": 1.3508101555411085e-06, "loss": 0.4659, "step": 18311 }, { "epoch": 0.7676136780440774, "grad_norm": 1.8351163864135742, "learning_rate": 1.3503461120535365e-06, "loss": 0.4665, "step": 18312 }, { "epoch": 0.7676555966590863, "grad_norm": 2.161550998687744, "learning_rate": 1.3498821358422875e-06, "loss": 0.5052, "step": 18313 }, { "epoch": 0.7676975152740954, "grad_norm": 1.812988519668579, "learning_rate": 1.3494182269159151e-06, "loss": 0.4583, "step": 18314 }, { "epoch": 0.7677394338891043, "grad_norm": 1.631005883216858, "learning_rate": 1.3489543852829723e-06, "loss": 0.4917, "step": 18315 }, { "epoch": 0.7677813525041133, "grad_norm": 1.7311949729919434, "learning_rate": 1.348490610952007e-06, "loss": 0.4663, "step": 18316 }, { "epoch": 0.7678232711191222, "grad_norm": 1.6261752843856812, "learning_rate": 1.3480269039315684e-06, "loss": 0.4898, "step": 18317 }, { "epoch": 0.7678651897341312, "grad_norm": 1.8569120168685913, "learning_rate": 1.3475632642302066e-06, "loss": 0.4923, "step": 18318 }, { "epoch": 0.7679071083491401, "grad_norm": 1.8390487432479858, "learning_rate": 1.3470996918564672e-06, "loss": 0.4678, "step": 18319 }, { "epoch": 0.7679490269641491, "grad_norm": 1.7038508653640747, "learning_rate": 1.3466361868188938e-06, "loss": 0.4862, "step": 18320 }, { "epoch": 0.767990945579158, "grad_norm": 1.716658592224121, "learning_rate": 1.3461727491260324e-06, "loss": 0.499, "step": 18321 }, { "epoch": 0.768032864194167, "grad_norm": 1.7673624753952026, "learning_rate": 1.345709378786424e-06, "loss": 0.4548, "step": 18322 }, { "epoch": 0.768074782809176, "grad_norm": 1.6999956369400024, "learning_rate": 1.3452460758086111e-06, "loss": 0.4587, "step": 18323 }, { "epoch": 0.768116701424185, "grad_norm": 2.3218376636505127, "learning_rate": 1.3447828402011358e-06, "loss": 0.4938, "step": 18324 }, { "epoch": 0.7681586200391939, "grad_norm": 1.9096862077713013, "learning_rate": 1.3443196719725343e-06, "loss": 0.5003, "step": 18325 }, { "epoch": 0.7682005386542029, "grad_norm": 1.9578596353530884, "learning_rate": 1.3438565711313457e-06, "loss": 0.4751, "step": 18326 }, { "epoch": 0.7682424572692118, "grad_norm": 1.9241294860839844, "learning_rate": 1.3433935376861084e-06, "loss": 0.4867, "step": 18327 }, { "epoch": 0.7682843758842208, "grad_norm": 1.7771618366241455, "learning_rate": 1.342930571645356e-06, "loss": 0.4371, "step": 18328 }, { "epoch": 0.7683262944992297, "grad_norm": 2.1710097789764404, "learning_rate": 1.3424676730176213e-06, "loss": 0.5036, "step": 18329 }, { "epoch": 0.7683682131142388, "grad_norm": 1.4703233242034912, "learning_rate": 1.34200484181144e-06, "loss": 0.4366, "step": 18330 }, { "epoch": 0.7684101317292477, "grad_norm": 2.1121490001678467, "learning_rate": 1.3415420780353411e-06, "loss": 0.514, "step": 18331 }, { "epoch": 0.7684520503442567, "grad_norm": 1.6666778326034546, "learning_rate": 1.3410793816978563e-06, "loss": 0.4656, "step": 18332 }, { "epoch": 0.7684939689592656, "grad_norm": 2.0836262702941895, "learning_rate": 1.3406167528075164e-06, "loss": 0.44, "step": 18333 }, { "epoch": 0.7685358875742746, "grad_norm": 2.0438568592071533, "learning_rate": 1.3401541913728462e-06, "loss": 0.5009, "step": 18334 }, { "epoch": 0.7685778061892835, "grad_norm": 1.5306241512298584, "learning_rate": 1.3396916974023738e-06, "loss": 0.4921, "step": 18335 }, { "epoch": 0.7686197248042924, "grad_norm": 1.8591358661651611, "learning_rate": 1.3392292709046262e-06, "loss": 0.4808, "step": 18336 }, { "epoch": 0.7686616434193014, "grad_norm": 1.968424677848816, "learning_rate": 1.3387669118881258e-06, "loss": 0.4575, "step": 18337 }, { "epoch": 0.7687035620343103, "grad_norm": 1.556674599647522, "learning_rate": 1.3383046203613946e-06, "loss": 0.4626, "step": 18338 }, { "epoch": 0.7687454806493194, "grad_norm": 1.9015312194824219, "learning_rate": 1.3378423963329578e-06, "loss": 0.5083, "step": 18339 }, { "epoch": 0.7687873992643283, "grad_norm": 2.254866600036621, "learning_rate": 1.337380239811331e-06, "loss": 0.4324, "step": 18340 }, { "epoch": 0.7688293178793373, "grad_norm": 1.8794147968292236, "learning_rate": 1.336918150805037e-06, "loss": 0.4729, "step": 18341 }, { "epoch": 0.7688712364943462, "grad_norm": 1.989688754081726, "learning_rate": 1.3364561293225942e-06, "loss": 0.5197, "step": 18342 }, { "epoch": 0.7689131551093552, "grad_norm": 1.6960622072219849, "learning_rate": 1.335994175372516e-06, "loss": 0.4555, "step": 18343 }, { "epoch": 0.7689550737243641, "grad_norm": 1.693853735923767, "learning_rate": 1.33553228896332e-06, "loss": 0.4985, "step": 18344 }, { "epoch": 0.7689969923393731, "grad_norm": 1.7419179677963257, "learning_rate": 1.3350704701035216e-06, "loss": 0.4535, "step": 18345 }, { "epoch": 0.769038910954382, "grad_norm": 2.291027545928955, "learning_rate": 1.3346087188016315e-06, "loss": 0.5291, "step": 18346 }, { "epoch": 0.7690808295693911, "grad_norm": 1.6061118841171265, "learning_rate": 1.3341470350661633e-06, "loss": 0.4566, "step": 18347 }, { "epoch": 0.7691227481844, "grad_norm": 1.5438082218170166, "learning_rate": 1.333685418905627e-06, "loss": 0.4553, "step": 18348 }, { "epoch": 0.769164666799409, "grad_norm": 1.6991500854492188, "learning_rate": 1.33322387032853e-06, "loss": 0.4679, "step": 18349 }, { "epoch": 0.7692065854144179, "grad_norm": 1.6216486692428589, "learning_rate": 1.3327623893433817e-06, "loss": 0.4937, "step": 18350 }, { "epoch": 0.7692485040294269, "grad_norm": 1.6962974071502686, "learning_rate": 1.3323009759586903e-06, "loss": 0.4807, "step": 18351 }, { "epoch": 0.7692904226444358, "grad_norm": 1.7818008661270142, "learning_rate": 1.331839630182959e-06, "loss": 0.4775, "step": 18352 }, { "epoch": 0.7693323412594448, "grad_norm": 1.8700354099273682, "learning_rate": 1.3313783520246937e-06, "loss": 0.4527, "step": 18353 }, { "epoch": 0.7693742598744537, "grad_norm": 1.5728713274002075, "learning_rate": 1.330917141492396e-06, "loss": 0.4627, "step": 18354 }, { "epoch": 0.7694161784894628, "grad_norm": 1.7034655809402466, "learning_rate": 1.330455998594568e-06, "loss": 0.4823, "step": 18355 }, { "epoch": 0.7694580971044717, "grad_norm": 1.713243007659912, "learning_rate": 1.3299949233397124e-06, "loss": 0.438, "step": 18356 }, { "epoch": 0.7695000157194807, "grad_norm": 1.800279974937439, "learning_rate": 1.3295339157363268e-06, "loss": 0.4445, "step": 18357 }, { "epoch": 0.7695419343344896, "grad_norm": 3.2579853534698486, "learning_rate": 1.3290729757929071e-06, "loss": 0.5554, "step": 18358 }, { "epoch": 0.7695838529494986, "grad_norm": 2.146193504333496, "learning_rate": 1.3286121035179545e-06, "loss": 0.429, "step": 18359 }, { "epoch": 0.7696257715645075, "grad_norm": 1.6452316045761108, "learning_rate": 1.3281512989199601e-06, "loss": 0.4299, "step": 18360 }, { "epoch": 0.7696676901795164, "grad_norm": 1.8029125928878784, "learning_rate": 1.327690562007421e-06, "loss": 0.4334, "step": 18361 }, { "epoch": 0.7697096087945254, "grad_norm": 3.0710766315460205, "learning_rate": 1.3272298927888305e-06, "loss": 0.5169, "step": 18362 }, { "epoch": 0.7697515274095343, "grad_norm": 2.232128143310547, "learning_rate": 1.3267692912726782e-06, "loss": 0.4848, "step": 18363 }, { "epoch": 0.7697934460245434, "grad_norm": 1.884507179260254, "learning_rate": 1.3263087574674561e-06, "loss": 0.5089, "step": 18364 }, { "epoch": 0.7698353646395523, "grad_norm": 2.7226462364196777, "learning_rate": 1.3258482913816544e-06, "loss": 0.4865, "step": 18365 }, { "epoch": 0.7698772832545613, "grad_norm": 2.417736530303955, "learning_rate": 1.3253878930237602e-06, "loss": 0.4739, "step": 18366 }, { "epoch": 0.7699192018695702, "grad_norm": 1.5219154357910156, "learning_rate": 1.324927562402259e-06, "loss": 0.4008, "step": 18367 }, { "epoch": 0.7699611204845792, "grad_norm": 1.9054325819015503, "learning_rate": 1.3244672995256386e-06, "loss": 0.3812, "step": 18368 }, { "epoch": 0.7700030390995881, "grad_norm": 2.0320076942443848, "learning_rate": 1.3240071044023812e-06, "loss": 0.4328, "step": 18369 }, { "epoch": 0.7700449577145971, "grad_norm": 1.6860209703445435, "learning_rate": 1.323546977040971e-06, "loss": 0.4629, "step": 18370 }, { "epoch": 0.770086876329606, "grad_norm": 1.7053035497665405, "learning_rate": 1.3230869174498905e-06, "loss": 0.4521, "step": 18371 }, { "epoch": 0.7701287949446151, "grad_norm": 2.042499303817749, "learning_rate": 1.3226269256376184e-06, "loss": 0.4718, "step": 18372 }, { "epoch": 0.770170713559624, "grad_norm": 2.4495487213134766, "learning_rate": 1.3221670016126353e-06, "loss": 0.434, "step": 18373 }, { "epoch": 0.770212632174633, "grad_norm": 1.8038042783737183, "learning_rate": 1.3217071453834206e-06, "loss": 0.4756, "step": 18374 }, { "epoch": 0.7702545507896419, "grad_norm": 1.8193156719207764, "learning_rate": 1.3212473569584493e-06, "loss": 0.4581, "step": 18375 }, { "epoch": 0.7702964694046509, "grad_norm": 1.583516240119934, "learning_rate": 1.3207876363461959e-06, "loss": 0.4254, "step": 18376 }, { "epoch": 0.7703383880196598, "grad_norm": 1.8675336837768555, "learning_rate": 1.3203279835551374e-06, "loss": 0.4775, "step": 18377 }, { "epoch": 0.7703803066346688, "grad_norm": 1.8469244241714478, "learning_rate": 1.3198683985937445e-06, "loss": 0.5084, "step": 18378 }, { "epoch": 0.7704222252496777, "grad_norm": 3.174766778945923, "learning_rate": 1.3194088814704897e-06, "loss": 0.5294, "step": 18379 }, { "epoch": 0.7704641438646868, "grad_norm": 1.6910510063171387, "learning_rate": 1.3189494321938455e-06, "loss": 0.4618, "step": 18380 }, { "epoch": 0.7705060624796957, "grad_norm": 1.991301417350769, "learning_rate": 1.3184900507722781e-06, "loss": 0.472, "step": 18381 }, { "epoch": 0.7705479810947047, "grad_norm": 1.813232421875, "learning_rate": 1.318030737214257e-06, "loss": 0.4374, "step": 18382 }, { "epoch": 0.7705898997097136, "grad_norm": 1.6808403730392456, "learning_rate": 1.3175714915282505e-06, "loss": 0.4553, "step": 18383 }, { "epoch": 0.7706318183247226, "grad_norm": 1.8452966213226318, "learning_rate": 1.3171123137227225e-06, "loss": 0.5271, "step": 18384 }, { "epoch": 0.7706737369397315, "grad_norm": 1.8597959280014038, "learning_rate": 1.316653203806137e-06, "loss": 0.4894, "step": 18385 }, { "epoch": 0.7707156555547404, "grad_norm": 1.8272536993026733, "learning_rate": 1.3161941617869583e-06, "loss": 0.5172, "step": 18386 }, { "epoch": 0.7707575741697494, "grad_norm": 1.8045769929885864, "learning_rate": 1.3157351876736464e-06, "loss": 0.4371, "step": 18387 }, { "epoch": 0.7707994927847583, "grad_norm": 3.4803524017333984, "learning_rate": 1.3152762814746628e-06, "loss": 0.5007, "step": 18388 }, { "epoch": 0.7708414113997674, "grad_norm": 1.7999134063720703, "learning_rate": 1.3148174431984685e-06, "loss": 0.4799, "step": 18389 }, { "epoch": 0.7708833300147763, "grad_norm": 1.6694746017456055, "learning_rate": 1.3143586728535185e-06, "loss": 0.447, "step": 18390 }, { "epoch": 0.7709252486297853, "grad_norm": 1.9426136016845703, "learning_rate": 1.3138999704482714e-06, "loss": 0.4228, "step": 18391 }, { "epoch": 0.7709671672447942, "grad_norm": 1.99558687210083, "learning_rate": 1.3134413359911835e-06, "loss": 0.513, "step": 18392 }, { "epoch": 0.7710090858598032, "grad_norm": 1.6056159734725952, "learning_rate": 1.3129827694907065e-06, "loss": 0.4633, "step": 18393 }, { "epoch": 0.7710510044748121, "grad_norm": 1.8210793733596802, "learning_rate": 1.3125242709552972e-06, "loss": 0.5242, "step": 18394 }, { "epoch": 0.7710929230898211, "grad_norm": 1.708537220954895, "learning_rate": 1.3120658403934045e-06, "loss": 0.4761, "step": 18395 }, { "epoch": 0.77113484170483, "grad_norm": 1.563151240348816, "learning_rate": 1.3116074778134786e-06, "loss": 0.4415, "step": 18396 }, { "epoch": 0.7711767603198391, "grad_norm": 1.7322598695755005, "learning_rate": 1.3111491832239693e-06, "loss": 0.4933, "step": 18397 }, { "epoch": 0.771218678934848, "grad_norm": 2.2207224369049072, "learning_rate": 1.310690956633327e-06, "loss": 0.4593, "step": 18398 }, { "epoch": 0.771260597549857, "grad_norm": 1.8360470533370972, "learning_rate": 1.3102327980499946e-06, "loss": 0.4991, "step": 18399 }, { "epoch": 0.7713025161648659, "grad_norm": 2.1786787509918213, "learning_rate": 1.3097747074824202e-06, "loss": 0.4753, "step": 18400 }, { "epoch": 0.7713444347798749, "grad_norm": 1.9578320980072021, "learning_rate": 1.3093166849390487e-06, "loss": 0.504, "step": 18401 }, { "epoch": 0.7713863533948838, "grad_norm": 1.9971568584442139, "learning_rate": 1.3088587304283207e-06, "loss": 0.5004, "step": 18402 }, { "epoch": 0.7714282720098928, "grad_norm": 5.816951274871826, "learning_rate": 1.3084008439586804e-06, "loss": 0.456, "step": 18403 }, { "epoch": 0.7714701906249017, "grad_norm": 1.9036433696746826, "learning_rate": 1.3079430255385672e-06, "loss": 0.4989, "step": 18404 }, { "epoch": 0.7715121092399108, "grad_norm": 1.9441510438919067, "learning_rate": 1.3074852751764184e-06, "loss": 0.5093, "step": 18405 }, { "epoch": 0.7715540278549197, "grad_norm": 2.028623580932617, "learning_rate": 1.3070275928806742e-06, "loss": 0.4902, "step": 18406 }, { "epoch": 0.7715959464699287, "grad_norm": 1.6694037914276123, "learning_rate": 1.3065699786597718e-06, "loss": 0.4805, "step": 18407 }, { "epoch": 0.7716378650849376, "grad_norm": 1.8065780401229858, "learning_rate": 1.3061124325221453e-06, "loss": 0.4933, "step": 18408 }, { "epoch": 0.7716797836999466, "grad_norm": 1.9825754165649414, "learning_rate": 1.3056549544762287e-06, "loss": 0.4888, "step": 18409 }, { "epoch": 0.7717217023149555, "grad_norm": 2.2236499786376953, "learning_rate": 1.3051975445304576e-06, "loss": 0.4588, "step": 18410 }, { "epoch": 0.7717636209299644, "grad_norm": 1.7615331411361694, "learning_rate": 1.30474020269326e-06, "loss": 0.4504, "step": 18411 }, { "epoch": 0.7718055395449734, "grad_norm": 1.8474839925765991, "learning_rate": 1.3042829289730702e-06, "loss": 0.4723, "step": 18412 }, { "epoch": 0.7718474581599823, "grad_norm": 1.752909541130066, "learning_rate": 1.3038257233783153e-06, "loss": 0.5387, "step": 18413 }, { "epoch": 0.7718893767749914, "grad_norm": 1.7504295110702515, "learning_rate": 1.3033685859174218e-06, "loss": 0.4934, "step": 18414 }, { "epoch": 0.7719312953900003, "grad_norm": 1.9556373357772827, "learning_rate": 1.3029115165988177e-06, "loss": 0.4839, "step": 18415 }, { "epoch": 0.7719732140050093, "grad_norm": 2.1674206256866455, "learning_rate": 1.3024545154309304e-06, "loss": 0.4492, "step": 18416 }, { "epoch": 0.7720151326200182, "grad_norm": 1.7981717586517334, "learning_rate": 1.3019975824221808e-06, "loss": 0.4794, "step": 18417 }, { "epoch": 0.7720570512350272, "grad_norm": 2.8896963596343994, "learning_rate": 1.3015407175809936e-06, "loss": 0.4918, "step": 18418 }, { "epoch": 0.7720989698500361, "grad_norm": 1.7598174810409546, "learning_rate": 1.3010839209157922e-06, "loss": 0.4435, "step": 18419 }, { "epoch": 0.7721408884650451, "grad_norm": 2.0360329151153564, "learning_rate": 1.3006271924349928e-06, "loss": 0.4872, "step": 18420 }, { "epoch": 0.772182807080054, "grad_norm": 1.9873548746109009, "learning_rate": 1.3001705321470192e-06, "loss": 0.5216, "step": 18421 }, { "epoch": 0.7722247256950631, "grad_norm": 1.8031913042068481, "learning_rate": 1.2997139400602865e-06, "loss": 0.4937, "step": 18422 }, { "epoch": 0.772266644310072, "grad_norm": 1.9227194786071777, "learning_rate": 1.2992574161832106e-06, "loss": 0.4524, "step": 18423 }, { "epoch": 0.772308562925081, "grad_norm": 1.9320333003997803, "learning_rate": 1.2988009605242074e-06, "loss": 0.4995, "step": 18424 }, { "epoch": 0.7723504815400899, "grad_norm": 1.9431523084640503, "learning_rate": 1.2983445730916938e-06, "loss": 0.4956, "step": 18425 }, { "epoch": 0.7723924001550989, "grad_norm": 1.5433529615402222, "learning_rate": 1.297888253894079e-06, "loss": 0.4949, "step": 18426 }, { "epoch": 0.7724343187701078, "grad_norm": 3.6039047241210938, "learning_rate": 1.297432002939778e-06, "loss": 0.534, "step": 18427 }, { "epoch": 0.7724762373851168, "grad_norm": 1.8745770454406738, "learning_rate": 1.2969758202371974e-06, "loss": 0.4517, "step": 18428 }, { "epoch": 0.7725181560001257, "grad_norm": 1.827720284461975, "learning_rate": 1.2965197057947488e-06, "loss": 0.4909, "step": 18429 }, { "epoch": 0.7725600746151348, "grad_norm": 3.1199371814727783, "learning_rate": 1.2960636596208408e-06, "loss": 0.4935, "step": 18430 }, { "epoch": 0.7726019932301437, "grad_norm": 2.1361162662506104, "learning_rate": 1.2956076817238787e-06, "loss": 0.5363, "step": 18431 }, { "epoch": 0.7726439118451527, "grad_norm": 1.8556374311447144, "learning_rate": 1.2951517721122665e-06, "loss": 0.4491, "step": 18432 }, { "epoch": 0.7726858304601616, "grad_norm": 1.9586262702941895, "learning_rate": 1.2946959307944107e-06, "loss": 0.4906, "step": 18433 }, { "epoch": 0.7727277490751706, "grad_norm": 1.6843947172164917, "learning_rate": 1.2942401577787116e-06, "loss": 0.4529, "step": 18434 }, { "epoch": 0.7727696676901795, "grad_norm": 1.5588452816009521, "learning_rate": 1.293784453073572e-06, "loss": 0.4566, "step": 18435 }, { "epoch": 0.7728115863051884, "grad_norm": 1.9283921718597412, "learning_rate": 1.2933288166873942e-06, "loss": 0.5028, "step": 18436 }, { "epoch": 0.7728535049201974, "grad_norm": 2.15838885307312, "learning_rate": 1.2928732486285733e-06, "loss": 0.4599, "step": 18437 }, { "epoch": 0.7728954235352063, "grad_norm": 2.4637765884399414, "learning_rate": 1.2924177489055095e-06, "loss": 0.4899, "step": 18438 }, { "epoch": 0.7729373421502154, "grad_norm": 1.8680598735809326, "learning_rate": 1.2919623175266005e-06, "loss": 0.5001, "step": 18439 }, { "epoch": 0.7729792607652243, "grad_norm": 2.480651617050171, "learning_rate": 1.2915069545002385e-06, "loss": 0.4836, "step": 18440 }, { "epoch": 0.7730211793802333, "grad_norm": 1.8907887935638428, "learning_rate": 1.2910516598348199e-06, "loss": 0.4126, "step": 18441 }, { "epoch": 0.7730630979952422, "grad_norm": 2.314929485321045, "learning_rate": 1.2905964335387372e-06, "loss": 0.5111, "step": 18442 }, { "epoch": 0.7731050166102512, "grad_norm": 1.758271336555481, "learning_rate": 1.290141275620379e-06, "loss": 0.5616, "step": 18443 }, { "epoch": 0.7731469352252601, "grad_norm": 1.7281599044799805, "learning_rate": 1.2896861860881382e-06, "loss": 0.5041, "step": 18444 }, { "epoch": 0.7731888538402691, "grad_norm": 1.8639692068099976, "learning_rate": 1.2892311649504047e-06, "loss": 0.5365, "step": 18445 }, { "epoch": 0.773230772455278, "grad_norm": 2.0877280235290527, "learning_rate": 1.2887762122155628e-06, "loss": 0.466, "step": 18446 }, { "epoch": 0.7732726910702871, "grad_norm": 1.7440454959869385, "learning_rate": 1.2883213278920008e-06, "loss": 0.4713, "step": 18447 }, { "epoch": 0.773314609685296, "grad_norm": 1.7050824165344238, "learning_rate": 1.2878665119881056e-06, "loss": 0.4669, "step": 18448 }, { "epoch": 0.773356528300305, "grad_norm": 2.5598268508911133, "learning_rate": 1.287411764512258e-06, "loss": 0.5159, "step": 18449 }, { "epoch": 0.7733984469153139, "grad_norm": 1.858585000038147, "learning_rate": 1.2869570854728436e-06, "loss": 0.4863, "step": 18450 }, { "epoch": 0.7734403655303229, "grad_norm": 1.9922322034835815, "learning_rate": 1.2865024748782418e-06, "loss": 0.5266, "step": 18451 }, { "epoch": 0.7734822841453318, "grad_norm": 2.8906772136688232, "learning_rate": 1.286047932736832e-06, "loss": 0.443, "step": 18452 }, { "epoch": 0.7735242027603408, "grad_norm": 1.7410920858383179, "learning_rate": 1.2855934590569936e-06, "loss": 0.5125, "step": 18453 }, { "epoch": 0.7735661213753497, "grad_norm": 1.882189393043518, "learning_rate": 1.2851390538471065e-06, "loss": 0.5232, "step": 18454 }, { "epoch": 0.7736080399903588, "grad_norm": 2.32576584815979, "learning_rate": 1.284684717115544e-06, "loss": 0.4964, "step": 18455 }, { "epoch": 0.7736499586053677, "grad_norm": 2.211045265197754, "learning_rate": 1.2842304488706826e-06, "loss": 0.5036, "step": 18456 }, { "epoch": 0.7736918772203767, "grad_norm": 2.0437278747558594, "learning_rate": 1.2837762491208972e-06, "loss": 0.5163, "step": 18457 }, { "epoch": 0.7737337958353856, "grad_norm": 2.130143880844116, "learning_rate": 1.283322117874558e-06, "loss": 0.4604, "step": 18458 }, { "epoch": 0.7737757144503946, "grad_norm": 2.215120792388916, "learning_rate": 1.2828680551400386e-06, "loss": 0.425, "step": 18459 }, { "epoch": 0.7738176330654035, "grad_norm": 1.481955647468567, "learning_rate": 1.2824140609257074e-06, "loss": 0.4466, "step": 18460 }, { "epoch": 0.7738595516804124, "grad_norm": 1.882848858833313, "learning_rate": 1.281960135239933e-06, "loss": 0.4506, "step": 18461 }, { "epoch": 0.7739014702954214, "grad_norm": 1.8507893085479736, "learning_rate": 1.2815062780910826e-06, "loss": 0.446, "step": 18462 }, { "epoch": 0.7739433889104304, "grad_norm": 2.292322874069214, "learning_rate": 1.2810524894875253e-06, "loss": 0.5153, "step": 18463 }, { "epoch": 0.7739853075254394, "grad_norm": 2.0771265029907227, "learning_rate": 1.2805987694376227e-06, "loss": 0.4208, "step": 18464 }, { "epoch": 0.7740272261404483, "grad_norm": 1.7640405893325806, "learning_rate": 1.2801451179497392e-06, "loss": 0.521, "step": 18465 }, { "epoch": 0.7740691447554573, "grad_norm": 1.594667673110962, "learning_rate": 1.2796915350322403e-06, "loss": 0.4756, "step": 18466 }, { "epoch": 0.7741110633704662, "grad_norm": 1.8778759241104126, "learning_rate": 1.2792380206934824e-06, "loss": 0.5139, "step": 18467 }, { "epoch": 0.7741529819854752, "grad_norm": 1.5952032804489136, "learning_rate": 1.2787845749418298e-06, "loss": 0.5008, "step": 18468 }, { "epoch": 0.7741949006004841, "grad_norm": 1.5801109075546265, "learning_rate": 1.2783311977856388e-06, "loss": 0.4861, "step": 18469 }, { "epoch": 0.7742368192154931, "grad_norm": 4.760839939117432, "learning_rate": 1.2778778892332661e-06, "loss": 0.4829, "step": 18470 }, { "epoch": 0.774278737830502, "grad_norm": 2.845803737640381, "learning_rate": 1.2774246492930687e-06, "loss": 0.4418, "step": 18471 }, { "epoch": 0.7743206564455111, "grad_norm": 1.8774479627609253, "learning_rate": 1.2769714779734026e-06, "loss": 0.5148, "step": 18472 }, { "epoch": 0.77436257506052, "grad_norm": 1.9290326833724976, "learning_rate": 1.2765183752826194e-06, "loss": 0.4956, "step": 18473 }, { "epoch": 0.774404493675529, "grad_norm": 1.6895538568496704, "learning_rate": 1.2760653412290723e-06, "loss": 0.4411, "step": 18474 }, { "epoch": 0.7744464122905379, "grad_norm": 2.3131918907165527, "learning_rate": 1.2756123758211142e-06, "loss": 0.5379, "step": 18475 }, { "epoch": 0.7744883309055469, "grad_norm": 1.6905065774917603, "learning_rate": 1.2751594790670913e-06, "loss": 0.4137, "step": 18476 }, { "epoch": 0.7745302495205558, "grad_norm": 1.9805362224578857, "learning_rate": 1.2747066509753553e-06, "loss": 0.4928, "step": 18477 }, { "epoch": 0.7745721681355648, "grad_norm": 1.8964016437530518, "learning_rate": 1.2742538915542524e-06, "loss": 0.4813, "step": 18478 }, { "epoch": 0.7746140867505737, "grad_norm": 1.606059193611145, "learning_rate": 1.2738012008121264e-06, "loss": 0.4172, "step": 18479 }, { "epoch": 0.7746560053655828, "grad_norm": 1.7937893867492676, "learning_rate": 1.2733485787573247e-06, "loss": 0.4691, "step": 18480 }, { "epoch": 0.7746979239805917, "grad_norm": 2.0039350986480713, "learning_rate": 1.2728960253981908e-06, "loss": 0.5116, "step": 18481 }, { "epoch": 0.7747398425956007, "grad_norm": 1.9892818927764893, "learning_rate": 1.2724435407430652e-06, "loss": 0.4864, "step": 18482 }, { "epoch": 0.7747817612106096, "grad_norm": 1.798521637916565, "learning_rate": 1.2719911248002893e-06, "loss": 0.4681, "step": 18483 }, { "epoch": 0.7748236798256186, "grad_norm": 2.4777727127075195, "learning_rate": 1.2715387775782046e-06, "loss": 0.4849, "step": 18484 }, { "epoch": 0.7748655984406275, "grad_norm": 1.838233470916748, "learning_rate": 1.2710864990851468e-06, "loss": 0.4563, "step": 18485 }, { "epoch": 0.7749075170556364, "grad_norm": 1.928139567375183, "learning_rate": 1.270634289329456e-06, "loss": 0.4688, "step": 18486 }, { "epoch": 0.7749494356706454, "grad_norm": 1.7140958309173584, "learning_rate": 1.2701821483194648e-06, "loss": 0.508, "step": 18487 }, { "epoch": 0.7749913542856544, "grad_norm": 1.966025471687317, "learning_rate": 1.2697300760635107e-06, "loss": 0.5064, "step": 18488 }, { "epoch": 0.7750332729006634, "grad_norm": 1.877813696861267, "learning_rate": 1.2692780725699244e-06, "loss": 0.5042, "step": 18489 }, { "epoch": 0.7750751915156723, "grad_norm": 2.712704658508301, "learning_rate": 1.2688261378470407e-06, "loss": 0.5125, "step": 18490 }, { "epoch": 0.7751171101306813, "grad_norm": 1.5400031805038452, "learning_rate": 1.2683742719031872e-06, "loss": 0.4694, "step": 18491 }, { "epoch": 0.7751590287456902, "grad_norm": 1.927774429321289, "learning_rate": 1.2679224747466955e-06, "loss": 0.5375, "step": 18492 }, { "epoch": 0.7752009473606992, "grad_norm": 2.4353251457214355, "learning_rate": 1.2674707463858948e-06, "loss": 0.4765, "step": 18493 }, { "epoch": 0.7752428659757081, "grad_norm": 2.1308047771453857, "learning_rate": 1.2670190868291094e-06, "loss": 0.5081, "step": 18494 }, { "epoch": 0.7752847845907171, "grad_norm": 1.7712074518203735, "learning_rate": 1.266567496084667e-06, "loss": 0.4837, "step": 18495 }, { "epoch": 0.775326703205726, "grad_norm": 2.9184460639953613, "learning_rate": 1.266115974160892e-06, "loss": 0.4622, "step": 18496 }, { "epoch": 0.7753686218207351, "grad_norm": 2.121838331222534, "learning_rate": 1.265664521066108e-06, "loss": 0.5016, "step": 18497 }, { "epoch": 0.775410540435744, "grad_norm": 1.9397562742233276, "learning_rate": 1.265213136808634e-06, "loss": 0.5581, "step": 18498 }, { "epoch": 0.775452459050753, "grad_norm": 3.575937271118164, "learning_rate": 1.264761821396794e-06, "loss": 0.4454, "step": 18499 }, { "epoch": 0.7754943776657619, "grad_norm": 1.793557047843933, "learning_rate": 1.2643105748389045e-06, "loss": 0.4173, "step": 18500 }, { "epoch": 0.7755362962807709, "grad_norm": 1.9930267333984375, "learning_rate": 1.2638593971432871e-06, "loss": 0.4876, "step": 18501 }, { "epoch": 0.7755782148957798, "grad_norm": 2.0124664306640625, "learning_rate": 1.2634082883182547e-06, "loss": 0.452, "step": 18502 }, { "epoch": 0.7756201335107888, "grad_norm": 1.6218334436416626, "learning_rate": 1.2629572483721252e-06, "loss": 0.4613, "step": 18503 }, { "epoch": 0.7756620521257978, "grad_norm": 1.9670661687850952, "learning_rate": 1.2625062773132135e-06, "loss": 0.4576, "step": 18504 }, { "epoch": 0.7757039707408068, "grad_norm": 1.8503040075302124, "learning_rate": 1.2620553751498304e-06, "loss": 0.4844, "step": 18505 }, { "epoch": 0.7757458893558157, "grad_norm": 2.033649206161499, "learning_rate": 1.2616045418902907e-06, "loss": 0.4839, "step": 18506 }, { "epoch": 0.7757878079708247, "grad_norm": 1.7839655876159668, "learning_rate": 1.2611537775429028e-06, "loss": 0.4954, "step": 18507 }, { "epoch": 0.7758297265858336, "grad_norm": 1.9412305355072021, "learning_rate": 1.260703082115975e-06, "loss": 0.4859, "step": 18508 }, { "epoch": 0.7758716452008426, "grad_norm": 3.356255054473877, "learning_rate": 1.2602524556178163e-06, "loss": 0.472, "step": 18509 }, { "epoch": 0.7759135638158515, "grad_norm": 1.7260103225708008, "learning_rate": 1.259801898056735e-06, "loss": 0.4869, "step": 18510 }, { "epoch": 0.7759554824308604, "grad_norm": 1.7370589971542358, "learning_rate": 1.259351409441033e-06, "loss": 0.5094, "step": 18511 }, { "epoch": 0.7759974010458695, "grad_norm": 2.372159719467163, "learning_rate": 1.2589009897790173e-06, "loss": 0.4525, "step": 18512 }, { "epoch": 0.7760393196608784, "grad_norm": 1.8872928619384766, "learning_rate": 1.2584506390789914e-06, "loss": 0.4594, "step": 18513 }, { "epoch": 0.7760812382758874, "grad_norm": 2.04891300201416, "learning_rate": 1.2580003573492533e-06, "loss": 0.4867, "step": 18514 }, { "epoch": 0.7761231568908963, "grad_norm": 1.6951302289962769, "learning_rate": 1.2575501445981076e-06, "loss": 0.4472, "step": 18515 }, { "epoch": 0.7761650755059053, "grad_norm": 1.8506404161453247, "learning_rate": 1.2571000008338508e-06, "loss": 0.4711, "step": 18516 }, { "epoch": 0.7762069941209142, "grad_norm": 6.154794692993164, "learning_rate": 1.2566499260647796e-06, "loss": 0.4846, "step": 18517 }, { "epoch": 0.7762489127359232, "grad_norm": 2.107635498046875, "learning_rate": 1.256199920299192e-06, "loss": 0.5578, "step": 18518 }, { "epoch": 0.7762908313509321, "grad_norm": 1.900172233581543, "learning_rate": 1.255749983545385e-06, "loss": 0.4565, "step": 18519 }, { "epoch": 0.7763327499659411, "grad_norm": 2.1731083393096924, "learning_rate": 1.255300115811649e-06, "loss": 0.4556, "step": 18520 }, { "epoch": 0.7763746685809501, "grad_norm": 2.0079333782196045, "learning_rate": 1.2548503171062786e-06, "loss": 0.4264, "step": 18521 }, { "epoch": 0.7764165871959591, "grad_norm": 1.7776153087615967, "learning_rate": 1.2544005874375663e-06, "loss": 0.4853, "step": 18522 }, { "epoch": 0.776458505810968, "grad_norm": 1.9791489839553833, "learning_rate": 1.2539509268138e-06, "loss": 0.5432, "step": 18523 }, { "epoch": 0.776500424425977, "grad_norm": 2.0026373863220215, "learning_rate": 1.2535013352432706e-06, "loss": 0.5307, "step": 18524 }, { "epoch": 0.7765423430409859, "grad_norm": 2.5967516899108887, "learning_rate": 1.2530518127342639e-06, "loss": 0.5075, "step": 18525 }, { "epoch": 0.7765842616559949, "grad_norm": 2.10653018951416, "learning_rate": 1.2526023592950682e-06, "loss": 0.4611, "step": 18526 }, { "epoch": 0.7766261802710038, "grad_norm": 1.9169723987579346, "learning_rate": 1.2521529749339656e-06, "loss": 0.4589, "step": 18527 }, { "epoch": 0.7766680988860128, "grad_norm": 1.8800662755966187, "learning_rate": 1.2517036596592435e-06, "loss": 0.4305, "step": 18528 }, { "epoch": 0.7767100175010218, "grad_norm": 2.7859113216400146, "learning_rate": 1.251254413479181e-06, "loss": 0.4536, "step": 18529 }, { "epoch": 0.7767519361160308, "grad_norm": 2.307804822921753, "learning_rate": 1.2508052364020606e-06, "loss": 0.5068, "step": 18530 }, { "epoch": 0.7767938547310397, "grad_norm": 2.761531114578247, "learning_rate": 1.2503561284361642e-06, "loss": 0.5452, "step": 18531 }, { "epoch": 0.7768357733460487, "grad_norm": 1.717212438583374, "learning_rate": 1.2499070895897675e-06, "loss": 0.4848, "step": 18532 }, { "epoch": 0.7768776919610576, "grad_norm": 1.893140196800232, "learning_rate": 1.2494581198711508e-06, "loss": 0.4972, "step": 18533 }, { "epoch": 0.7769196105760666, "grad_norm": 1.7933276891708374, "learning_rate": 1.2490092192885872e-06, "loss": 0.4541, "step": 18534 }, { "epoch": 0.7769615291910755, "grad_norm": 1.8832974433898926, "learning_rate": 1.2485603878503544e-06, "loss": 0.4662, "step": 18535 }, { "epoch": 0.7770034478060844, "grad_norm": 4.05194091796875, "learning_rate": 1.2481116255647235e-06, "loss": 0.4763, "step": 18536 }, { "epoch": 0.7770453664210935, "grad_norm": 1.7963870763778687, "learning_rate": 1.2476629324399692e-06, "loss": 0.465, "step": 18537 }, { "epoch": 0.7770872850361024, "grad_norm": 2.0021257400512695, "learning_rate": 1.2472143084843596e-06, "loss": 0.4739, "step": 18538 }, { "epoch": 0.7771292036511114, "grad_norm": 1.7548694610595703, "learning_rate": 1.2467657537061661e-06, "loss": 0.4282, "step": 18539 }, { "epoch": 0.7771711222661203, "grad_norm": 2.0424022674560547, "learning_rate": 1.2463172681136587e-06, "loss": 0.449, "step": 18540 }, { "epoch": 0.7772130408811293, "grad_norm": 1.8826103210449219, "learning_rate": 1.2458688517151025e-06, "loss": 0.4835, "step": 18541 }, { "epoch": 0.7772549594961382, "grad_norm": 1.9382497072219849, "learning_rate": 1.2454205045187634e-06, "loss": 0.467, "step": 18542 }, { "epoch": 0.7772968781111472, "grad_norm": 1.9778600931167603, "learning_rate": 1.2449722265329078e-06, "loss": 0.5148, "step": 18543 }, { "epoch": 0.7773387967261561, "grad_norm": 1.861434817314148, "learning_rate": 1.2445240177657986e-06, "loss": 0.489, "step": 18544 }, { "epoch": 0.7773807153411652, "grad_norm": 2.583632469177246, "learning_rate": 1.2440758782256962e-06, "loss": 0.5206, "step": 18545 }, { "epoch": 0.7774226339561741, "grad_norm": 1.9381957054138184, "learning_rate": 1.2436278079208635e-06, "loss": 0.4654, "step": 18546 }, { "epoch": 0.7774645525711831, "grad_norm": 1.7732754945755005, "learning_rate": 1.243179806859558e-06, "loss": 0.5053, "step": 18547 }, { "epoch": 0.777506471186192, "grad_norm": 2.214634418487549, "learning_rate": 1.242731875050039e-06, "loss": 0.4756, "step": 18548 }, { "epoch": 0.777548389801201, "grad_norm": 2.113180160522461, "learning_rate": 1.2422840125005647e-06, "loss": 0.4898, "step": 18549 }, { "epoch": 0.7775903084162099, "grad_norm": 1.685482144355774, "learning_rate": 1.2418362192193888e-06, "loss": 0.4962, "step": 18550 }, { "epoch": 0.7776322270312189, "grad_norm": 1.8945980072021484, "learning_rate": 1.2413884952147664e-06, "loss": 0.4574, "step": 18551 }, { "epoch": 0.7776741456462278, "grad_norm": 1.7754544019699097, "learning_rate": 1.2409408404949524e-06, "loss": 0.5403, "step": 18552 }, { "epoch": 0.7777160642612369, "grad_norm": 2.062422752380371, "learning_rate": 1.2404932550681976e-06, "loss": 0.4915, "step": 18553 }, { "epoch": 0.7777579828762458, "grad_norm": 1.653709888458252, "learning_rate": 1.2400457389427505e-06, "loss": 0.4371, "step": 18554 }, { "epoch": 0.7777999014912548, "grad_norm": 2.263490676879883, "learning_rate": 1.2395982921268634e-06, "loss": 0.4612, "step": 18555 }, { "epoch": 0.7778418201062637, "grad_norm": 1.8867743015289307, "learning_rate": 1.2391509146287822e-06, "loss": 0.5006, "step": 18556 }, { "epoch": 0.7778837387212727, "grad_norm": 2.458057165145874, "learning_rate": 1.2387036064567547e-06, "loss": 0.5271, "step": 18557 }, { "epoch": 0.7779256573362816, "grad_norm": 1.6291106939315796, "learning_rate": 1.2382563676190279e-06, "loss": 0.3983, "step": 18558 }, { "epoch": 0.7779675759512906, "grad_norm": 1.9714233875274658, "learning_rate": 1.2378091981238427e-06, "loss": 0.4776, "step": 18559 }, { "epoch": 0.7780094945662995, "grad_norm": 1.7627865076065063, "learning_rate": 1.2373620979794444e-06, "loss": 0.4656, "step": 18560 }, { "epoch": 0.7780514131813084, "grad_norm": 2.0193123817443848, "learning_rate": 1.236915067194075e-06, "loss": 0.4743, "step": 18561 }, { "epoch": 0.7780933317963175, "grad_norm": 2.99232816696167, "learning_rate": 1.2364681057759743e-06, "loss": 0.4566, "step": 18562 }, { "epoch": 0.7781352504113264, "grad_norm": 1.5209240913391113, "learning_rate": 1.2360212137333794e-06, "loss": 0.4235, "step": 18563 }, { "epoch": 0.7781771690263354, "grad_norm": 2.116908550262451, "learning_rate": 1.2355743910745315e-06, "loss": 0.4896, "step": 18564 }, { "epoch": 0.7782190876413443, "grad_norm": 1.8093245029449463, "learning_rate": 1.2351276378076644e-06, "loss": 0.4878, "step": 18565 }, { "epoch": 0.7782610062563533, "grad_norm": 2.4196724891662598, "learning_rate": 1.2346809539410143e-06, "loss": 0.5091, "step": 18566 }, { "epoch": 0.7783029248713622, "grad_norm": 1.7375458478927612, "learning_rate": 1.2342343394828165e-06, "loss": 0.4732, "step": 18567 }, { "epoch": 0.7783448434863712, "grad_norm": 1.7080672979354858, "learning_rate": 1.2337877944413018e-06, "loss": 0.4145, "step": 18568 }, { "epoch": 0.7783867621013801, "grad_norm": 1.866598129272461, "learning_rate": 1.2333413188247023e-06, "loss": 0.4921, "step": 18569 }, { "epoch": 0.7784286807163892, "grad_norm": 1.630624532699585, "learning_rate": 1.2328949126412493e-06, "loss": 0.4814, "step": 18570 }, { "epoch": 0.7784705993313981, "grad_norm": 1.8337297439575195, "learning_rate": 1.2324485758991712e-06, "loss": 0.4592, "step": 18571 }, { "epoch": 0.7785125179464071, "grad_norm": 1.9575859308242798, "learning_rate": 1.2320023086066934e-06, "loss": 0.4904, "step": 18572 }, { "epoch": 0.778554436561416, "grad_norm": 2.313516855239868, "learning_rate": 1.2315561107720458e-06, "loss": 0.5067, "step": 18573 }, { "epoch": 0.778596355176425, "grad_norm": 1.9730195999145508, "learning_rate": 1.23110998240345e-06, "loss": 0.4532, "step": 18574 }, { "epoch": 0.7786382737914339, "grad_norm": 2.042146921157837, "learning_rate": 1.2306639235091312e-06, "loss": 0.4633, "step": 18575 }, { "epoch": 0.7786801924064429, "grad_norm": 2.800715923309326, "learning_rate": 1.2302179340973136e-06, "loss": 0.4578, "step": 18576 }, { "epoch": 0.7787221110214518, "grad_norm": 1.8017503023147583, "learning_rate": 1.2297720141762148e-06, "loss": 0.4794, "step": 18577 }, { "epoch": 0.7787640296364609, "grad_norm": 1.8325934410095215, "learning_rate": 1.2293261637540588e-06, "loss": 0.425, "step": 18578 }, { "epoch": 0.7788059482514698, "grad_norm": 2.3152804374694824, "learning_rate": 1.2288803828390604e-06, "loss": 0.4826, "step": 18579 }, { "epoch": 0.7788478668664788, "grad_norm": 1.9702744483947754, "learning_rate": 1.228434671439439e-06, "loss": 0.4617, "step": 18580 }, { "epoch": 0.7788897854814877, "grad_norm": 1.9281193017959595, "learning_rate": 1.2279890295634123e-06, "loss": 0.5151, "step": 18581 }, { "epoch": 0.7789317040964967, "grad_norm": 1.8727831840515137, "learning_rate": 1.2275434572191925e-06, "loss": 0.4793, "step": 18582 }, { "epoch": 0.7789736227115056, "grad_norm": 1.8676252365112305, "learning_rate": 1.2270979544149924e-06, "loss": 0.4953, "step": 18583 }, { "epoch": 0.7790155413265146, "grad_norm": 1.7252315282821655, "learning_rate": 1.226652521159027e-06, "loss": 0.4619, "step": 18584 }, { "epoch": 0.7790574599415235, "grad_norm": 1.735671043395996, "learning_rate": 1.2262071574595046e-06, "loss": 0.4497, "step": 18585 }, { "epoch": 0.7790993785565324, "grad_norm": 1.920853853225708, "learning_rate": 1.225761863324636e-06, "loss": 0.5085, "step": 18586 }, { "epoch": 0.7791412971715415, "grad_norm": 1.8721225261688232, "learning_rate": 1.2253166387626314e-06, "loss": 0.4556, "step": 18587 }, { "epoch": 0.7791832157865504, "grad_norm": 1.9198193550109863, "learning_rate": 1.224871483781695e-06, "loss": 0.5273, "step": 18588 }, { "epoch": 0.7792251344015594, "grad_norm": 1.787862777709961, "learning_rate": 1.2244263983900334e-06, "loss": 0.5204, "step": 18589 }, { "epoch": 0.7792670530165683, "grad_norm": 1.8306249380111694, "learning_rate": 1.223981382595853e-06, "loss": 0.4927, "step": 18590 }, { "epoch": 0.7793089716315773, "grad_norm": 1.9324027299880981, "learning_rate": 1.223536436407356e-06, "loss": 0.5117, "step": 18591 }, { "epoch": 0.7793508902465862, "grad_norm": 2.2686588764190674, "learning_rate": 1.223091559832742e-06, "loss": 0.4878, "step": 18592 }, { "epoch": 0.7793928088615952, "grad_norm": 1.9134187698364258, "learning_rate": 1.2226467528802156e-06, "loss": 0.4233, "step": 18593 }, { "epoch": 0.7794347274766041, "grad_norm": 2.150569200515747, "learning_rate": 1.2222020155579727e-06, "loss": 0.4842, "step": 18594 }, { "epoch": 0.7794766460916132, "grad_norm": 1.5855294466018677, "learning_rate": 1.2217573478742128e-06, "loss": 0.4532, "step": 18595 }, { "epoch": 0.7795185647066221, "grad_norm": 1.69194495677948, "learning_rate": 1.2213127498371346e-06, "loss": 0.4842, "step": 18596 }, { "epoch": 0.7795604833216311, "grad_norm": 1.711597204208374, "learning_rate": 1.2208682214549307e-06, "loss": 0.5497, "step": 18597 }, { "epoch": 0.77960240193664, "grad_norm": 1.5995197296142578, "learning_rate": 1.2204237627357961e-06, "loss": 0.5052, "step": 18598 }, { "epoch": 0.779644320551649, "grad_norm": 1.733463168144226, "learning_rate": 1.2199793736879263e-06, "loss": 0.4773, "step": 18599 }, { "epoch": 0.7796862391666579, "grad_norm": 1.7665756940841675, "learning_rate": 1.2195350543195106e-06, "loss": 0.5248, "step": 18600 }, { "epoch": 0.7797281577816669, "grad_norm": 1.9232878684997559, "learning_rate": 1.219090804638739e-06, "loss": 0.4545, "step": 18601 }, { "epoch": 0.7797700763966758, "grad_norm": 2.2974624633789062, "learning_rate": 1.218646624653802e-06, "loss": 0.4772, "step": 18602 }, { "epoch": 0.7798119950116849, "grad_norm": 2.0544190406799316, "learning_rate": 1.218202514372886e-06, "loss": 0.4914, "step": 18603 }, { "epoch": 0.7798539136266938, "grad_norm": 2.461414337158203, "learning_rate": 1.2177584738041786e-06, "loss": 0.4246, "step": 18604 }, { "epoch": 0.7798958322417028, "grad_norm": 30.383089065551758, "learning_rate": 1.2173145029558664e-06, "loss": 0.4746, "step": 18605 }, { "epoch": 0.7799377508567117, "grad_norm": 1.804591178894043, "learning_rate": 1.2168706018361303e-06, "loss": 0.4556, "step": 18606 }, { "epoch": 0.7799796694717207, "grad_norm": 3.140697479248047, "learning_rate": 1.2164267704531546e-06, "loss": 0.459, "step": 18607 }, { "epoch": 0.7800215880867296, "grad_norm": 1.929683804512024, "learning_rate": 1.215983008815122e-06, "loss": 0.519, "step": 18608 }, { "epoch": 0.7800635067017386, "grad_norm": 2.0480239391326904, "learning_rate": 1.2155393169302115e-06, "loss": 0.4994, "step": 18609 }, { "epoch": 0.7801054253167475, "grad_norm": 1.773314356803894, "learning_rate": 1.2150956948066005e-06, "loss": 0.4446, "step": 18610 }, { "epoch": 0.7801473439317564, "grad_norm": 1.7182948589324951, "learning_rate": 1.2146521424524698e-06, "loss": 0.4526, "step": 18611 }, { "epoch": 0.7801892625467655, "grad_norm": 1.9303686618804932, "learning_rate": 1.2142086598759917e-06, "loss": 0.4536, "step": 18612 }, { "epoch": 0.7802311811617744, "grad_norm": 2.0066490173339844, "learning_rate": 1.2137652470853434e-06, "loss": 0.4563, "step": 18613 }, { "epoch": 0.7802730997767834, "grad_norm": 2.2103636264801025, "learning_rate": 1.2133219040887002e-06, "loss": 0.445, "step": 18614 }, { "epoch": 0.7803150183917923, "grad_norm": 2.014112710952759, "learning_rate": 1.212878630894231e-06, "loss": 0.4827, "step": 18615 }, { "epoch": 0.7803569370068013, "grad_norm": 1.9190338850021362, "learning_rate": 1.2124354275101086e-06, "loss": 0.4591, "step": 18616 }, { "epoch": 0.7803988556218102, "grad_norm": 2.0366766452789307, "learning_rate": 1.2119922939445045e-06, "loss": 0.5516, "step": 18617 }, { "epoch": 0.7804407742368192, "grad_norm": 1.7305954694747925, "learning_rate": 1.211549230205586e-06, "loss": 0.4208, "step": 18618 }, { "epoch": 0.7804826928518281, "grad_norm": 1.62615168094635, "learning_rate": 1.2111062363015186e-06, "loss": 0.4423, "step": 18619 }, { "epoch": 0.7805246114668372, "grad_norm": 1.8885040283203125, "learning_rate": 1.210663312240471e-06, "loss": 0.4206, "step": 18620 }, { "epoch": 0.7805665300818461, "grad_norm": 1.6759939193725586, "learning_rate": 1.2102204580306047e-06, "loss": 0.4468, "step": 18621 }, { "epoch": 0.7806084486968551, "grad_norm": 1.989296555519104, "learning_rate": 1.209777673680086e-06, "loss": 0.4412, "step": 18622 }, { "epoch": 0.780650367311864, "grad_norm": 2.095677375793457, "learning_rate": 1.2093349591970765e-06, "loss": 0.4523, "step": 18623 }, { "epoch": 0.780692285926873, "grad_norm": 1.7187763452529907, "learning_rate": 1.2088923145897357e-06, "loss": 0.4825, "step": 18624 }, { "epoch": 0.7807342045418819, "grad_norm": 1.6407475471496582, "learning_rate": 1.2084497398662242e-06, "loss": 0.4352, "step": 18625 }, { "epoch": 0.7807761231568909, "grad_norm": 1.6700048446655273, "learning_rate": 1.208007235034701e-06, "loss": 0.4311, "step": 18626 }, { "epoch": 0.7808180417718998, "grad_norm": 2.286808729171753, "learning_rate": 1.2075648001033214e-06, "loss": 0.5062, "step": 18627 }, { "epoch": 0.7808599603869089, "grad_norm": 2.174196243286133, "learning_rate": 1.207122435080243e-06, "loss": 0.4775, "step": 18628 }, { "epoch": 0.7809018790019178, "grad_norm": 2.0672097206115723, "learning_rate": 1.206680139973619e-06, "loss": 0.4897, "step": 18629 }, { "epoch": 0.7809437976169268, "grad_norm": 1.8141144514083862, "learning_rate": 1.2062379147916004e-06, "loss": 0.5058, "step": 18630 }, { "epoch": 0.7809857162319357, "grad_norm": 1.6272944211959839, "learning_rate": 1.205795759542342e-06, "loss": 0.489, "step": 18631 }, { "epoch": 0.7810276348469447, "grad_norm": 1.881438136100769, "learning_rate": 1.2053536742339944e-06, "loss": 0.4801, "step": 18632 }, { "epoch": 0.7810695534619536, "grad_norm": 1.718485951423645, "learning_rate": 1.2049116588747045e-06, "loss": 0.4712, "step": 18633 }, { "epoch": 0.7811114720769626, "grad_norm": 2.907627820968628, "learning_rate": 1.204469713472622e-06, "loss": 0.5206, "step": 18634 }, { "epoch": 0.7811533906919715, "grad_norm": 1.8292953968048096, "learning_rate": 1.2040278380358943e-06, "loss": 0.491, "step": 18635 }, { "epoch": 0.7811953093069804, "grad_norm": 1.6417255401611328, "learning_rate": 1.2035860325726639e-06, "loss": 0.4436, "step": 18636 }, { "epoch": 0.7812372279219895, "grad_norm": 1.6561731100082397, "learning_rate": 1.203144297091079e-06, "loss": 0.4786, "step": 18637 }, { "epoch": 0.7812791465369984, "grad_norm": 2.0556271076202393, "learning_rate": 1.202702631599279e-06, "loss": 0.514, "step": 18638 }, { "epoch": 0.7813210651520074, "grad_norm": 2.2100579738616943, "learning_rate": 1.2022610361054054e-06, "loss": 0.4899, "step": 18639 }, { "epoch": 0.7813629837670163, "grad_norm": 1.6226806640625, "learning_rate": 1.2018195106175996e-06, "loss": 0.4493, "step": 18640 }, { "epoch": 0.7814049023820253, "grad_norm": 2.2407567501068115, "learning_rate": 1.2013780551440013e-06, "loss": 0.4943, "step": 18641 }, { "epoch": 0.7814468209970342, "grad_norm": 2.2629148960113525, "learning_rate": 1.2009366696927454e-06, "loss": 0.5082, "step": 18642 }, { "epoch": 0.7814887396120432, "grad_norm": 10.532854080200195, "learning_rate": 1.2004953542719705e-06, "loss": 0.439, "step": 18643 }, { "epoch": 0.7815306582270521, "grad_norm": 1.451075792312622, "learning_rate": 1.200054108889812e-06, "loss": 0.4589, "step": 18644 }, { "epoch": 0.7815725768420612, "grad_norm": 1.8144913911819458, "learning_rate": 1.199612933554402e-06, "loss": 0.4962, "step": 18645 }, { "epoch": 0.7816144954570701, "grad_norm": 2.2286717891693115, "learning_rate": 1.1991718282738745e-06, "loss": 0.4436, "step": 18646 }, { "epoch": 0.7816564140720791, "grad_norm": 1.5738505125045776, "learning_rate": 1.1987307930563608e-06, "loss": 0.472, "step": 18647 }, { "epoch": 0.781698332687088, "grad_norm": 1.83292555809021, "learning_rate": 1.1982898279099875e-06, "loss": 0.4815, "step": 18648 }, { "epoch": 0.781740251302097, "grad_norm": 1.5159385204315186, "learning_rate": 1.1978489328428855e-06, "loss": 0.4289, "step": 18649 }, { "epoch": 0.7817821699171059, "grad_norm": 2.0891811847686768, "learning_rate": 1.1974081078631838e-06, "loss": 0.5246, "step": 18650 }, { "epoch": 0.7818240885321149, "grad_norm": 1.6395595073699951, "learning_rate": 1.1969673529790054e-06, "loss": 0.4656, "step": 18651 }, { "epoch": 0.7818660071471238, "grad_norm": 1.716305136680603, "learning_rate": 1.1965266681984777e-06, "loss": 0.4329, "step": 18652 }, { "epoch": 0.7819079257621329, "grad_norm": 2.84765362739563, "learning_rate": 1.1960860535297215e-06, "loss": 0.4866, "step": 18653 }, { "epoch": 0.7819498443771418, "grad_norm": 2.6250784397125244, "learning_rate": 1.1956455089808594e-06, "loss": 0.5036, "step": 18654 }, { "epoch": 0.7819917629921508, "grad_norm": 1.740851640701294, "learning_rate": 1.1952050345600153e-06, "loss": 0.4485, "step": 18655 }, { "epoch": 0.7820336816071597, "grad_norm": 1.8481905460357666, "learning_rate": 1.1947646302753058e-06, "loss": 0.417, "step": 18656 }, { "epoch": 0.7820756002221687, "grad_norm": 1.8606414794921875, "learning_rate": 1.1943242961348483e-06, "loss": 0.5224, "step": 18657 }, { "epoch": 0.7821175188371776, "grad_norm": 1.7609082460403442, "learning_rate": 1.193884032146763e-06, "loss": 0.4508, "step": 18658 }, { "epoch": 0.7821594374521866, "grad_norm": 2.029660940170288, "learning_rate": 1.193443838319162e-06, "loss": 0.4571, "step": 18659 }, { "epoch": 0.7822013560671955, "grad_norm": 1.7588798999786377, "learning_rate": 1.193003714660162e-06, "loss": 0.4932, "step": 18660 }, { "epoch": 0.7822432746822044, "grad_norm": 1.900282621383667, "learning_rate": 1.1925636611778767e-06, "loss": 0.4739, "step": 18661 }, { "epoch": 0.7822851932972135, "grad_norm": 3.8561205863952637, "learning_rate": 1.1921236778804152e-06, "loss": 0.4666, "step": 18662 }, { "epoch": 0.7823271119122224, "grad_norm": 1.732369303703308, "learning_rate": 1.1916837647758895e-06, "loss": 0.434, "step": 18663 }, { "epoch": 0.7823690305272314, "grad_norm": 4.715036869049072, "learning_rate": 1.1912439218724097e-06, "loss": 0.4609, "step": 18664 }, { "epoch": 0.7824109491422403, "grad_norm": 1.7778525352478027, "learning_rate": 1.190804149178083e-06, "loss": 0.5021, "step": 18665 }, { "epoch": 0.7824528677572493, "grad_norm": 1.8428716659545898, "learning_rate": 1.1903644467010145e-06, "loss": 0.4279, "step": 18666 }, { "epoch": 0.7824947863722582, "grad_norm": 1.7405248880386353, "learning_rate": 1.1899248144493125e-06, "loss": 0.4933, "step": 18667 }, { "epoch": 0.7825367049872672, "grad_norm": 1.6913926601409912, "learning_rate": 1.1894852524310774e-06, "loss": 0.4268, "step": 18668 }, { "epoch": 0.7825786236022761, "grad_norm": 1.8684743642807007, "learning_rate": 1.1890457606544138e-06, "loss": 0.4794, "step": 18669 }, { "epoch": 0.7826205422172852, "grad_norm": 2.3296380043029785, "learning_rate": 1.188606339127425e-06, "loss": 0.4887, "step": 18670 }, { "epoch": 0.7826624608322941, "grad_norm": 1.8655234575271606, "learning_rate": 1.188166987858207e-06, "loss": 0.4538, "step": 18671 }, { "epoch": 0.7827043794473031, "grad_norm": 1.7483245134353638, "learning_rate": 1.1877277068548614e-06, "loss": 0.4159, "step": 18672 }, { "epoch": 0.782746298062312, "grad_norm": 1.5780470371246338, "learning_rate": 1.1872884961254865e-06, "loss": 0.457, "step": 18673 }, { "epoch": 0.782788216677321, "grad_norm": 2.1354355812072754, "learning_rate": 1.1868493556781763e-06, "loss": 0.5145, "step": 18674 }, { "epoch": 0.7828301352923299, "grad_norm": 3.0443382263183594, "learning_rate": 1.1864102855210275e-06, "loss": 0.4615, "step": 18675 }, { "epoch": 0.7828720539073389, "grad_norm": 1.6853309869766235, "learning_rate": 1.1859712856621331e-06, "loss": 0.4076, "step": 18676 }, { "epoch": 0.7829139725223478, "grad_norm": 1.6572129726409912, "learning_rate": 1.1855323561095838e-06, "loss": 0.4443, "step": 18677 }, { "epoch": 0.7829558911373569, "grad_norm": 1.8182047605514526, "learning_rate": 1.1850934968714728e-06, "loss": 0.441, "step": 18678 }, { "epoch": 0.7829978097523658, "grad_norm": 1.801632285118103, "learning_rate": 1.18465470795589e-06, "loss": 0.5102, "step": 18679 }, { "epoch": 0.7830397283673748, "grad_norm": 1.6322518587112427, "learning_rate": 1.184215989370922e-06, "loss": 0.4733, "step": 18680 }, { "epoch": 0.7830816469823837, "grad_norm": 1.8757684230804443, "learning_rate": 1.1837773411246566e-06, "loss": 0.4976, "step": 18681 }, { "epoch": 0.7831235655973927, "grad_norm": 2.160898447036743, "learning_rate": 1.1833387632251825e-06, "loss": 0.5086, "step": 18682 }, { "epoch": 0.7831654842124016, "grad_norm": 1.7453627586364746, "learning_rate": 1.1829002556805797e-06, "loss": 0.5268, "step": 18683 }, { "epoch": 0.7832074028274106, "grad_norm": 1.7729939222335815, "learning_rate": 1.1824618184989355e-06, "loss": 0.5467, "step": 18684 }, { "epoch": 0.7832493214424195, "grad_norm": 1.9726239442825317, "learning_rate": 1.18202345168833e-06, "loss": 0.514, "step": 18685 }, { "epoch": 0.7832912400574285, "grad_norm": 2.2061588764190674, "learning_rate": 1.1815851552568425e-06, "loss": 0.5674, "step": 18686 }, { "epoch": 0.7833331586724375, "grad_norm": 1.713705062866211, "learning_rate": 1.181146929212554e-06, "loss": 0.4841, "step": 18687 }, { "epoch": 0.7833750772874464, "grad_norm": 1.9126900434494019, "learning_rate": 1.1807087735635437e-06, "loss": 0.5075, "step": 18688 }, { "epoch": 0.7834169959024554, "grad_norm": 1.732574701309204, "learning_rate": 1.1802706883178855e-06, "loss": 0.4736, "step": 18689 }, { "epoch": 0.7834589145174643, "grad_norm": 1.9368587732315063, "learning_rate": 1.1798326734836568e-06, "loss": 0.4736, "step": 18690 }, { "epoch": 0.7835008331324733, "grad_norm": 1.7192182540893555, "learning_rate": 1.1793947290689327e-06, "loss": 0.5068, "step": 18691 }, { "epoch": 0.7835427517474822, "grad_norm": 1.554837703704834, "learning_rate": 1.1789568550817844e-06, "loss": 0.4887, "step": 18692 }, { "epoch": 0.7835846703624912, "grad_norm": 1.9969384670257568, "learning_rate": 1.1785190515302848e-06, "loss": 0.4523, "step": 18693 }, { "epoch": 0.7836265889775001, "grad_norm": 1.737345814704895, "learning_rate": 1.1780813184225032e-06, "loss": 0.446, "step": 18694 }, { "epoch": 0.7836685075925092, "grad_norm": 1.983569860458374, "learning_rate": 1.1776436557665077e-06, "loss": 0.5279, "step": 18695 }, { "epoch": 0.7837104262075181, "grad_norm": 1.8933345079421997, "learning_rate": 1.1772060635703674e-06, "loss": 0.4219, "step": 18696 }, { "epoch": 0.7837523448225271, "grad_norm": 1.896459937095642, "learning_rate": 1.1767685418421498e-06, "loss": 0.4915, "step": 18697 }, { "epoch": 0.783794263437536, "grad_norm": 1.6947168111801147, "learning_rate": 1.1763310905899171e-06, "loss": 0.4961, "step": 18698 }, { "epoch": 0.783836182052545, "grad_norm": 1.8099113702774048, "learning_rate": 1.1758937098217344e-06, "loss": 0.5005, "step": 18699 }, { "epoch": 0.7838781006675539, "grad_norm": 1.8213601112365723, "learning_rate": 1.1754563995456664e-06, "loss": 0.5105, "step": 18700 }, { "epoch": 0.7839200192825629, "grad_norm": 2.3235678672790527, "learning_rate": 1.1750191597697707e-06, "loss": 0.4651, "step": 18701 }, { "epoch": 0.7839619378975718, "grad_norm": 1.852807879447937, "learning_rate": 1.1745819905021106e-06, "loss": 0.4458, "step": 18702 }, { "epoch": 0.7840038565125809, "grad_norm": 1.966672420501709, "learning_rate": 1.174144891750743e-06, "loss": 0.4968, "step": 18703 }, { "epoch": 0.7840457751275898, "grad_norm": 1.9086253643035889, "learning_rate": 1.1737078635237238e-06, "loss": 0.5181, "step": 18704 }, { "epoch": 0.7840876937425988, "grad_norm": 1.839823842048645, "learning_rate": 1.1732709058291098e-06, "loss": 0.4566, "step": 18705 }, { "epoch": 0.7841296123576077, "grad_norm": 1.9239109754562378, "learning_rate": 1.1728340186749588e-06, "loss": 0.4355, "step": 18706 }, { "epoch": 0.7841715309726167, "grad_norm": 3.441415309906006, "learning_rate": 1.1723972020693197e-06, "loss": 0.5213, "step": 18707 }, { "epoch": 0.7842134495876256, "grad_norm": 2.0149171352386475, "learning_rate": 1.171960456020247e-06, "loss": 0.4918, "step": 18708 }, { "epoch": 0.7842553682026346, "grad_norm": 1.793703317642212, "learning_rate": 1.1715237805357922e-06, "loss": 0.4658, "step": 18709 }, { "epoch": 0.7842972868176435, "grad_norm": 1.706516146659851, "learning_rate": 1.1710871756240027e-06, "loss": 0.4958, "step": 18710 }, { "epoch": 0.7843392054326525, "grad_norm": 1.6952011585235596, "learning_rate": 1.1706506412929293e-06, "loss": 0.4815, "step": 18711 }, { "epoch": 0.7843811240476615, "grad_norm": 1.5767168998718262, "learning_rate": 1.170214177550617e-06, "loss": 0.4919, "step": 18712 }, { "epoch": 0.7844230426626704, "grad_norm": 1.8544719219207764, "learning_rate": 1.1697777844051105e-06, "loss": 0.4482, "step": 18713 }, { "epoch": 0.7844649612776794, "grad_norm": 1.788593053817749, "learning_rate": 1.1693414618644555e-06, "loss": 0.4829, "step": 18714 }, { "epoch": 0.7845068798926883, "grad_norm": 2.406498908996582, "learning_rate": 1.1689052099366966e-06, "loss": 0.4798, "step": 18715 }, { "epoch": 0.7845487985076973, "grad_norm": 1.8837740421295166, "learning_rate": 1.1684690286298721e-06, "loss": 0.544, "step": 18716 }, { "epoch": 0.7845907171227062, "grad_norm": 1.6207300424575806, "learning_rate": 1.1680329179520245e-06, "loss": 0.4776, "step": 18717 }, { "epoch": 0.7846326357377152, "grad_norm": 1.6914759874343872, "learning_rate": 1.1675968779111934e-06, "loss": 0.4532, "step": 18718 }, { "epoch": 0.7846745543527242, "grad_norm": 2.080974817276001, "learning_rate": 1.1671609085154146e-06, "loss": 0.514, "step": 18719 }, { "epoch": 0.7847164729677332, "grad_norm": 2.1774041652679443, "learning_rate": 1.1667250097727274e-06, "loss": 0.4682, "step": 18720 }, { "epoch": 0.7847583915827421, "grad_norm": 1.7139948606491089, "learning_rate": 1.1662891816911632e-06, "loss": 0.4931, "step": 18721 }, { "epoch": 0.7848003101977511, "grad_norm": 2.23793625831604, "learning_rate": 1.16585342427876e-06, "loss": 0.4642, "step": 18722 }, { "epoch": 0.78484222881276, "grad_norm": 2.203707218170166, "learning_rate": 1.1654177375435465e-06, "loss": 0.4687, "step": 18723 }, { "epoch": 0.784884147427769, "grad_norm": 2.2630293369293213, "learning_rate": 1.1649821214935575e-06, "loss": 0.5121, "step": 18724 }, { "epoch": 0.7849260660427779, "grad_norm": 1.7252748012542725, "learning_rate": 1.1645465761368202e-06, "loss": 0.5024, "step": 18725 }, { "epoch": 0.7849679846577869, "grad_norm": 1.7246358394622803, "learning_rate": 1.1641111014813656e-06, "loss": 0.4796, "step": 18726 }, { "epoch": 0.7850099032727959, "grad_norm": 1.7790683507919312, "learning_rate": 1.1636756975352181e-06, "loss": 0.474, "step": 18727 }, { "epoch": 0.7850518218878049, "grad_norm": 2.005557060241699, "learning_rate": 1.1632403643064062e-06, "loss": 0.5056, "step": 18728 }, { "epoch": 0.7850937405028138, "grad_norm": 1.9107929468154907, "learning_rate": 1.1628051018029551e-06, "loss": 0.5286, "step": 18729 }, { "epoch": 0.7851356591178228, "grad_norm": 2.038756847381592, "learning_rate": 1.1623699100328856e-06, "loss": 0.4341, "step": 18730 }, { "epoch": 0.7851775777328317, "grad_norm": 2.0883467197418213, "learning_rate": 1.1619347890042232e-06, "loss": 0.4538, "step": 18731 }, { "epoch": 0.7852194963478407, "grad_norm": 2.22194504737854, "learning_rate": 1.161499738724987e-06, "loss": 0.4755, "step": 18732 }, { "epoch": 0.7852614149628496, "grad_norm": 1.8983319997787476, "learning_rate": 1.1610647592031949e-06, "loss": 0.5136, "step": 18733 }, { "epoch": 0.7853033335778586, "grad_norm": 6.474451541900635, "learning_rate": 1.160629850446867e-06, "loss": 0.4833, "step": 18734 }, { "epoch": 0.7853452521928675, "grad_norm": 1.7948124408721924, "learning_rate": 1.1601950124640216e-06, "loss": 0.4582, "step": 18735 }, { "epoch": 0.7853871708078765, "grad_norm": 2.1225316524505615, "learning_rate": 1.159760245262671e-06, "loss": 0.4503, "step": 18736 }, { "epoch": 0.7854290894228855, "grad_norm": 1.6164034605026245, "learning_rate": 1.1593255488508315e-06, "loss": 0.4541, "step": 18737 }, { "epoch": 0.7854710080378944, "grad_norm": 1.845589280128479, "learning_rate": 1.158890923236517e-06, "loss": 0.4848, "step": 18738 }, { "epoch": 0.7855129266529034, "grad_norm": 1.6574445962905884, "learning_rate": 1.1584563684277372e-06, "loss": 0.483, "step": 18739 }, { "epoch": 0.7855548452679123, "grad_norm": 1.8380330801010132, "learning_rate": 1.1580218844325047e-06, "loss": 0.4867, "step": 18740 }, { "epoch": 0.7855967638829213, "grad_norm": 2.8636984825134277, "learning_rate": 1.157587471258827e-06, "loss": 0.503, "step": 18741 }, { "epoch": 0.7856386824979302, "grad_norm": 1.7981013059616089, "learning_rate": 1.1571531289147114e-06, "loss": 0.4815, "step": 18742 }, { "epoch": 0.7856806011129392, "grad_norm": 1.467255711555481, "learning_rate": 1.156718857408165e-06, "loss": 0.4469, "step": 18743 }, { "epoch": 0.7857225197279482, "grad_norm": 1.908179521560669, "learning_rate": 1.1562846567471952e-06, "loss": 0.5118, "step": 18744 }, { "epoch": 0.7857644383429572, "grad_norm": 1.9675108194351196, "learning_rate": 1.1558505269398018e-06, "loss": 0.4938, "step": 18745 }, { "epoch": 0.7858063569579661, "grad_norm": 1.9084264039993286, "learning_rate": 1.1554164679939899e-06, "loss": 0.4914, "step": 18746 }, { "epoch": 0.7858482755729751, "grad_norm": 1.780293583869934, "learning_rate": 1.1549824799177623e-06, "loss": 0.5057, "step": 18747 }, { "epoch": 0.785890194187984, "grad_norm": 2.1104252338409424, "learning_rate": 1.1545485627191156e-06, "loss": 0.5302, "step": 18748 }, { "epoch": 0.785932112802993, "grad_norm": 1.6218717098236084, "learning_rate": 1.1541147164060506e-06, "loss": 0.5557, "step": 18749 }, { "epoch": 0.7859740314180019, "grad_norm": 1.682836651802063, "learning_rate": 1.1536809409865646e-06, "loss": 0.4819, "step": 18750 }, { "epoch": 0.786015950033011, "grad_norm": 1.915837287902832, "learning_rate": 1.1532472364686515e-06, "loss": 0.4726, "step": 18751 }, { "epoch": 0.7860578686480199, "grad_norm": 1.9892386198043823, "learning_rate": 1.1528136028603076e-06, "loss": 0.4591, "step": 18752 }, { "epoch": 0.7860997872630289, "grad_norm": 1.8893629312515259, "learning_rate": 1.1523800401695283e-06, "loss": 0.4493, "step": 18753 }, { "epoch": 0.7861417058780378, "grad_norm": 1.724250316619873, "learning_rate": 1.151946548404302e-06, "loss": 0.4409, "step": 18754 }, { "epoch": 0.7861836244930468, "grad_norm": 2.457838296890259, "learning_rate": 1.151513127572621e-06, "loss": 0.5638, "step": 18755 }, { "epoch": 0.7862255431080557, "grad_norm": 1.7650097608566284, "learning_rate": 1.1510797776824768e-06, "loss": 0.4927, "step": 18756 }, { "epoch": 0.7862674617230647, "grad_norm": 2.014439344406128, "learning_rate": 1.1506464987418547e-06, "loss": 0.454, "step": 18757 }, { "epoch": 0.7863093803380736, "grad_norm": 1.590558409690857, "learning_rate": 1.1502132907587437e-06, "loss": 0.4758, "step": 18758 }, { "epoch": 0.7863512989530826, "grad_norm": 1.8609974384307861, "learning_rate": 1.1497801537411291e-06, "loss": 0.4718, "step": 18759 }, { "epoch": 0.7863932175680916, "grad_norm": 2.3344738483428955, "learning_rate": 1.149347087696993e-06, "loss": 0.4565, "step": 18760 }, { "epoch": 0.7864351361831005, "grad_norm": 1.8577123880386353, "learning_rate": 1.1489140926343195e-06, "loss": 0.4633, "step": 18761 }, { "epoch": 0.7864770547981095, "grad_norm": 2.3962597846984863, "learning_rate": 1.148481168561093e-06, "loss": 0.4423, "step": 18762 }, { "epoch": 0.7865189734131184, "grad_norm": 2.3684511184692383, "learning_rate": 1.1480483154852894e-06, "loss": 0.5506, "step": 18763 }, { "epoch": 0.7865608920281274, "grad_norm": 1.8545571565628052, "learning_rate": 1.1476155334148902e-06, "loss": 0.4824, "step": 18764 }, { "epoch": 0.7866028106431363, "grad_norm": 2.1145944595336914, "learning_rate": 1.1471828223578745e-06, "loss": 0.5183, "step": 18765 }, { "epoch": 0.7866447292581453, "grad_norm": 1.7321377992630005, "learning_rate": 1.1467501823222155e-06, "loss": 0.4487, "step": 18766 }, { "epoch": 0.7866866478731542, "grad_norm": 1.903153896331787, "learning_rate": 1.1463176133158916e-06, "loss": 0.4582, "step": 18767 }, { "epoch": 0.7867285664881633, "grad_norm": 1.9441202878952026, "learning_rate": 1.1458851153468731e-06, "loss": 0.5459, "step": 18768 }, { "epoch": 0.7867704851031722, "grad_norm": 1.7814199924468994, "learning_rate": 1.145452688423136e-06, "loss": 0.4357, "step": 18769 }, { "epoch": 0.7868124037181812, "grad_norm": 1.7916334867477417, "learning_rate": 1.1450203325526489e-06, "loss": 0.4413, "step": 18770 }, { "epoch": 0.7868543223331901, "grad_norm": 2.61326265335083, "learning_rate": 1.144588047743383e-06, "loss": 0.4195, "step": 18771 }, { "epoch": 0.7868962409481991, "grad_norm": 2.3024423122406006, "learning_rate": 1.144155834003306e-06, "loss": 0.5029, "step": 18772 }, { "epoch": 0.786938159563208, "grad_norm": 1.7852667570114136, "learning_rate": 1.1437236913403849e-06, "loss": 0.4438, "step": 18773 }, { "epoch": 0.786980078178217, "grad_norm": 1.9047987461090088, "learning_rate": 1.1432916197625881e-06, "loss": 0.5105, "step": 18774 }, { "epoch": 0.7870219967932259, "grad_norm": 1.7327059507369995, "learning_rate": 1.1428596192778775e-06, "loss": 0.5241, "step": 18775 }, { "epoch": 0.787063915408235, "grad_norm": 2.0062103271484375, "learning_rate": 1.1424276898942172e-06, "loss": 0.5308, "step": 18776 }, { "epoch": 0.7871058340232439, "grad_norm": 1.815486192703247, "learning_rate": 1.1419958316195712e-06, "loss": 0.5069, "step": 18777 }, { "epoch": 0.7871477526382529, "grad_norm": 1.7404881715774536, "learning_rate": 1.141564044461898e-06, "loss": 0.4619, "step": 18778 }, { "epoch": 0.7871896712532618, "grad_norm": 1.7581675052642822, "learning_rate": 1.1411323284291564e-06, "loss": 0.4871, "step": 18779 }, { "epoch": 0.7872315898682708, "grad_norm": 1.9491996765136719, "learning_rate": 1.1407006835293073e-06, "loss": 0.4252, "step": 18780 }, { "epoch": 0.7872735084832797, "grad_norm": 1.688870906829834, "learning_rate": 1.140269109770304e-06, "loss": 0.4752, "step": 18781 }, { "epoch": 0.7873154270982887, "grad_norm": 1.856532096862793, "learning_rate": 1.1398376071601036e-06, "loss": 0.4731, "step": 18782 }, { "epoch": 0.7873573457132976, "grad_norm": 1.7483478784561157, "learning_rate": 1.1394061757066621e-06, "loss": 0.5334, "step": 18783 }, { "epoch": 0.7873992643283066, "grad_norm": 2.03438663482666, "learning_rate": 1.1389748154179286e-06, "loss": 0.5291, "step": 18784 }, { "epoch": 0.7874411829433156, "grad_norm": 2.150294303894043, "learning_rate": 1.1385435263018575e-06, "loss": 0.4598, "step": 18785 }, { "epoch": 0.7874831015583245, "grad_norm": 2.580796957015991, "learning_rate": 1.1381123083663993e-06, "loss": 0.5302, "step": 18786 }, { "epoch": 0.7875250201733335, "grad_norm": 1.7519348859786987, "learning_rate": 1.1376811616195016e-06, "loss": 0.5126, "step": 18787 }, { "epoch": 0.7875669387883424, "grad_norm": 1.8249300718307495, "learning_rate": 1.137250086069111e-06, "loss": 0.4813, "step": 18788 }, { "epoch": 0.7876088574033514, "grad_norm": 1.9928535223007202, "learning_rate": 1.1368190817231768e-06, "loss": 0.4519, "step": 18789 }, { "epoch": 0.7876507760183603, "grad_norm": 1.910698652267456, "learning_rate": 1.13638814858964e-06, "loss": 0.473, "step": 18790 }, { "epoch": 0.7876926946333693, "grad_norm": 2.4301445484161377, "learning_rate": 1.1359572866764467e-06, "loss": 0.4254, "step": 18791 }, { "epoch": 0.7877346132483782, "grad_norm": 1.8917062282562256, "learning_rate": 1.1355264959915402e-06, "loss": 0.4311, "step": 18792 }, { "epoch": 0.7877765318633873, "grad_norm": 1.5702375173568726, "learning_rate": 1.1350957765428588e-06, "loss": 0.5129, "step": 18793 }, { "epoch": 0.7878184504783962, "grad_norm": 1.9732118844985962, "learning_rate": 1.1346651283383442e-06, "loss": 0.4934, "step": 18794 }, { "epoch": 0.7878603690934052, "grad_norm": 1.9756150245666504, "learning_rate": 1.1342345513859355e-06, "loss": 0.4854, "step": 18795 }, { "epoch": 0.7879022877084141, "grad_norm": 2.0789921283721924, "learning_rate": 1.1338040456935684e-06, "loss": 0.4323, "step": 18796 }, { "epoch": 0.7879442063234231, "grad_norm": 1.9473488330841064, "learning_rate": 1.1333736112691772e-06, "loss": 0.5023, "step": 18797 }, { "epoch": 0.787986124938432, "grad_norm": 2.1943819522857666, "learning_rate": 1.1329432481206998e-06, "loss": 0.5548, "step": 18798 }, { "epoch": 0.788028043553441, "grad_norm": 2.6954360008239746, "learning_rate": 1.132512956256066e-06, "loss": 0.5058, "step": 18799 }, { "epoch": 0.7880699621684499, "grad_norm": 1.961176872253418, "learning_rate": 1.1320827356832092e-06, "loss": 0.4506, "step": 18800 }, { "epoch": 0.788111880783459, "grad_norm": 1.9080780744552612, "learning_rate": 1.1316525864100618e-06, "loss": 0.516, "step": 18801 }, { "epoch": 0.7881537993984679, "grad_norm": 1.8092937469482422, "learning_rate": 1.1312225084445493e-06, "loss": 0.5502, "step": 18802 }, { "epoch": 0.7881957180134769, "grad_norm": 1.897192358970642, "learning_rate": 1.1307925017946031e-06, "loss": 0.5109, "step": 18803 }, { "epoch": 0.7882376366284858, "grad_norm": 2.223818302154541, "learning_rate": 1.1303625664681466e-06, "loss": 0.5104, "step": 18804 }, { "epoch": 0.7882795552434948, "grad_norm": 2.0477917194366455, "learning_rate": 1.1299327024731077e-06, "loss": 0.4911, "step": 18805 }, { "epoch": 0.7883214738585037, "grad_norm": 1.8321104049682617, "learning_rate": 1.1295029098174093e-06, "loss": 0.5113, "step": 18806 }, { "epoch": 0.7883633924735127, "grad_norm": 1.7415896654129028, "learning_rate": 1.1290731885089728e-06, "loss": 0.4383, "step": 18807 }, { "epoch": 0.7884053110885216, "grad_norm": 2.272874355316162, "learning_rate": 1.128643538555721e-06, "loss": 0.4658, "step": 18808 }, { "epoch": 0.7884472297035306, "grad_norm": 1.7248188257217407, "learning_rate": 1.1282139599655746e-06, "loss": 0.4965, "step": 18809 }, { "epoch": 0.7884891483185396, "grad_norm": 2.177825450897217, "learning_rate": 1.1277844527464492e-06, "loss": 0.5423, "step": 18810 }, { "epoch": 0.7885310669335485, "grad_norm": 1.9577908515930176, "learning_rate": 1.1273550169062652e-06, "loss": 0.4166, "step": 18811 }, { "epoch": 0.7885729855485575, "grad_norm": 1.5786181688308716, "learning_rate": 1.1269256524529386e-06, "loss": 0.4829, "step": 18812 }, { "epoch": 0.7886149041635664, "grad_norm": 1.9005203247070312, "learning_rate": 1.1264963593943823e-06, "loss": 0.4276, "step": 18813 }, { "epoch": 0.7886568227785754, "grad_norm": 1.7630244493484497, "learning_rate": 1.1260671377385113e-06, "loss": 0.4803, "step": 18814 }, { "epoch": 0.7886987413935843, "grad_norm": 1.6682605743408203, "learning_rate": 1.125637987493236e-06, "loss": 0.4362, "step": 18815 }, { "epoch": 0.7887406600085933, "grad_norm": 1.999840497970581, "learning_rate": 1.1252089086664703e-06, "loss": 0.5201, "step": 18816 }, { "epoch": 0.7887825786236022, "grad_norm": 1.9219229221343994, "learning_rate": 1.1247799012661192e-06, "loss": 0.4689, "step": 18817 }, { "epoch": 0.7888244972386113, "grad_norm": 1.7249802350997925, "learning_rate": 1.1243509653000956e-06, "loss": 0.4684, "step": 18818 }, { "epoch": 0.7888664158536202, "grad_norm": 1.8164076805114746, "learning_rate": 1.1239221007763017e-06, "loss": 0.4759, "step": 18819 }, { "epoch": 0.7889083344686292, "grad_norm": 1.8424839973449707, "learning_rate": 1.1234933077026461e-06, "loss": 0.4657, "step": 18820 }, { "epoch": 0.7889502530836381, "grad_norm": 1.843476414680481, "learning_rate": 1.1230645860870337e-06, "loss": 0.4707, "step": 18821 }, { "epoch": 0.7889921716986471, "grad_norm": 1.9040241241455078, "learning_rate": 1.122635935937364e-06, "loss": 0.4218, "step": 18822 }, { "epoch": 0.789034090313656, "grad_norm": 1.7940040826797485, "learning_rate": 1.122207357261541e-06, "loss": 0.4504, "step": 18823 }, { "epoch": 0.789076008928665, "grad_norm": 1.779849648475647, "learning_rate": 1.1217788500674653e-06, "loss": 0.498, "step": 18824 }, { "epoch": 0.7891179275436739, "grad_norm": 1.7908234596252441, "learning_rate": 1.1213504143630354e-06, "loss": 0.4433, "step": 18825 }, { "epoch": 0.789159846158683, "grad_norm": 1.6691542863845825, "learning_rate": 1.1209220501561474e-06, "loss": 0.4615, "step": 18826 }, { "epoch": 0.7892017647736919, "grad_norm": 4.308655261993408, "learning_rate": 1.1204937574546998e-06, "loss": 0.4463, "step": 18827 }, { "epoch": 0.7892436833887009, "grad_norm": 2.0660135746002197, "learning_rate": 1.1200655362665852e-06, "loss": 0.5183, "step": 18828 }, { "epoch": 0.7892856020037098, "grad_norm": 1.7155722379684448, "learning_rate": 1.1196373865996984e-06, "loss": 0.5058, "step": 18829 }, { "epoch": 0.7893275206187188, "grad_norm": 1.803228497505188, "learning_rate": 1.1192093084619338e-06, "loss": 0.4581, "step": 18830 }, { "epoch": 0.7893694392337277, "grad_norm": 1.6917848587036133, "learning_rate": 1.1187813018611786e-06, "loss": 0.5057, "step": 18831 }, { "epoch": 0.7894113578487367, "grad_norm": 1.7518150806427002, "learning_rate": 1.1183533668053248e-06, "loss": 0.5256, "step": 18832 }, { "epoch": 0.7894532764637456, "grad_norm": 1.7312395572662354, "learning_rate": 1.1179255033022624e-06, "loss": 0.5165, "step": 18833 }, { "epoch": 0.7894951950787547, "grad_norm": 1.6842702627182007, "learning_rate": 1.1174977113598757e-06, "loss": 0.4917, "step": 18834 }, { "epoch": 0.7895371136937636, "grad_norm": 1.7067545652389526, "learning_rate": 1.11706999098605e-06, "loss": 0.4743, "step": 18835 }, { "epoch": 0.7895790323087725, "grad_norm": 1.541534662246704, "learning_rate": 1.1166423421886729e-06, "loss": 0.4705, "step": 18836 }, { "epoch": 0.7896209509237815, "grad_norm": 1.655028223991394, "learning_rate": 1.116214764975624e-06, "loss": 0.454, "step": 18837 }, { "epoch": 0.7896628695387904, "grad_norm": 1.9448468685150146, "learning_rate": 1.1157872593547863e-06, "loss": 0.4779, "step": 18838 }, { "epoch": 0.7897047881537994, "grad_norm": 1.591043472290039, "learning_rate": 1.1153598253340425e-06, "loss": 0.3846, "step": 18839 }, { "epoch": 0.7897467067688083, "grad_norm": 1.5430234670639038, "learning_rate": 1.1149324629212687e-06, "loss": 0.4887, "step": 18840 }, { "epoch": 0.7897886253838173, "grad_norm": 1.6336078643798828, "learning_rate": 1.1145051721243438e-06, "loss": 0.4944, "step": 18841 }, { "epoch": 0.7898305439988262, "grad_norm": 1.6848855018615723, "learning_rate": 1.114077952951146e-06, "loss": 0.5009, "step": 18842 }, { "epoch": 0.7898724626138353, "grad_norm": 1.6857993602752686, "learning_rate": 1.113650805409549e-06, "loss": 0.4637, "step": 18843 }, { "epoch": 0.7899143812288442, "grad_norm": 1.8965141773223877, "learning_rate": 1.1132237295074254e-06, "loss": 0.475, "step": 18844 }, { "epoch": 0.7899562998438532, "grad_norm": 1.7695355415344238, "learning_rate": 1.11279672525265e-06, "loss": 0.4747, "step": 18845 }, { "epoch": 0.7899982184588621, "grad_norm": 2.377767324447632, "learning_rate": 1.1123697926530925e-06, "loss": 0.5607, "step": 18846 }, { "epoch": 0.7900401370738711, "grad_norm": 1.7135926485061646, "learning_rate": 1.1119429317166226e-06, "loss": 0.495, "step": 18847 }, { "epoch": 0.79008205568888, "grad_norm": 1.8452988862991333, "learning_rate": 1.111516142451111e-06, "loss": 0.5356, "step": 18848 }, { "epoch": 0.790123974303889, "grad_norm": 1.9961130619049072, "learning_rate": 1.1110894248644228e-06, "loss": 0.4934, "step": 18849 }, { "epoch": 0.7901658929188979, "grad_norm": 1.7915352582931519, "learning_rate": 1.1106627789644248e-06, "loss": 0.4311, "step": 18850 }, { "epoch": 0.790207811533907, "grad_norm": 1.8360332250595093, "learning_rate": 1.1102362047589826e-06, "loss": 0.4304, "step": 18851 }, { "epoch": 0.7902497301489159, "grad_norm": 1.774373173713684, "learning_rate": 1.109809702255959e-06, "loss": 0.5018, "step": 18852 }, { "epoch": 0.7902916487639249, "grad_norm": 1.8528022766113281, "learning_rate": 1.1093832714632136e-06, "loss": 0.4761, "step": 18853 }, { "epoch": 0.7903335673789338, "grad_norm": 1.635568380355835, "learning_rate": 1.1089569123886107e-06, "loss": 0.4566, "step": 18854 }, { "epoch": 0.7903754859939428, "grad_norm": 1.911122441291809, "learning_rate": 1.1085306250400063e-06, "loss": 0.4721, "step": 18855 }, { "epoch": 0.7904174046089517, "grad_norm": 2.04795241355896, "learning_rate": 1.1081044094252601e-06, "loss": 0.5248, "step": 18856 }, { "epoch": 0.7904593232239607, "grad_norm": 2.0124149322509766, "learning_rate": 1.1076782655522306e-06, "loss": 0.48, "step": 18857 }, { "epoch": 0.7905012418389696, "grad_norm": 2.195876121520996, "learning_rate": 1.1072521934287694e-06, "loss": 0.4809, "step": 18858 }, { "epoch": 0.7905431604539787, "grad_norm": 1.915514588356018, "learning_rate": 1.1068261930627323e-06, "loss": 0.48, "step": 18859 }, { "epoch": 0.7905850790689876, "grad_norm": 1.914156198501587, "learning_rate": 1.1064002644619737e-06, "loss": 0.5291, "step": 18860 }, { "epoch": 0.7906269976839965, "grad_norm": 1.6852480173110962, "learning_rate": 1.1059744076343437e-06, "loss": 0.4415, "step": 18861 }, { "epoch": 0.7906689162990055, "grad_norm": 2.0780134201049805, "learning_rate": 1.105548622587691e-06, "loss": 0.4888, "step": 18862 }, { "epoch": 0.7907108349140144, "grad_norm": 1.9563391208648682, "learning_rate": 1.1051229093298666e-06, "loss": 0.5019, "step": 18863 }, { "epoch": 0.7907527535290234, "grad_norm": 1.7496099472045898, "learning_rate": 1.1046972678687152e-06, "loss": 0.4644, "step": 18864 }, { "epoch": 0.7907946721440323, "grad_norm": 2.099665880203247, "learning_rate": 1.104271698212085e-06, "loss": 0.4921, "step": 18865 }, { "epoch": 0.7908365907590413, "grad_norm": 1.7477056980133057, "learning_rate": 1.1038462003678213e-06, "loss": 0.4585, "step": 18866 }, { "epoch": 0.7908785093740502, "grad_norm": 1.7047348022460938, "learning_rate": 1.1034207743437659e-06, "loss": 0.4749, "step": 18867 }, { "epoch": 0.7909204279890593, "grad_norm": 3.8448190689086914, "learning_rate": 1.102995420147761e-06, "loss": 0.5244, "step": 18868 }, { "epoch": 0.7909623466040682, "grad_norm": 2.050731658935547, "learning_rate": 1.1025701377876502e-06, "loss": 0.4811, "step": 18869 }, { "epoch": 0.7910042652190772, "grad_norm": 1.6294664144515991, "learning_rate": 1.1021449272712687e-06, "loss": 0.4513, "step": 18870 }, { "epoch": 0.7910461838340861, "grad_norm": 1.8572381734848022, "learning_rate": 1.1017197886064584e-06, "loss": 0.4995, "step": 18871 }, { "epoch": 0.7910881024490951, "grad_norm": 2.2302613258361816, "learning_rate": 1.1012947218010551e-06, "loss": 0.5144, "step": 18872 }, { "epoch": 0.791130021064104, "grad_norm": 1.8531336784362793, "learning_rate": 1.1008697268628922e-06, "loss": 0.5084, "step": 18873 }, { "epoch": 0.791171939679113, "grad_norm": 3.302983522415161, "learning_rate": 1.100444803799805e-06, "loss": 0.5084, "step": 18874 }, { "epoch": 0.7912138582941219, "grad_norm": 2.049010992050171, "learning_rate": 1.1000199526196288e-06, "loss": 0.5815, "step": 18875 }, { "epoch": 0.791255776909131, "grad_norm": 2.203226327896118, "learning_rate": 1.0995951733301919e-06, "loss": 0.4305, "step": 18876 }, { "epoch": 0.7912976955241399, "grad_norm": 1.9451643228530884, "learning_rate": 1.099170465939327e-06, "loss": 0.4823, "step": 18877 }, { "epoch": 0.7913396141391489, "grad_norm": 3.2666544914245605, "learning_rate": 1.0987458304548605e-06, "loss": 0.4188, "step": 18878 }, { "epoch": 0.7913815327541578, "grad_norm": 1.6678591966629028, "learning_rate": 1.098321266884621e-06, "loss": 0.4841, "step": 18879 }, { "epoch": 0.7914234513691668, "grad_norm": 1.7852225303649902, "learning_rate": 1.0978967752364367e-06, "loss": 0.5078, "step": 18880 }, { "epoch": 0.7914653699841757, "grad_norm": 1.8878509998321533, "learning_rate": 1.0974723555181305e-06, "loss": 0.5151, "step": 18881 }, { "epoch": 0.7915072885991847, "grad_norm": 2.065664768218994, "learning_rate": 1.0970480077375256e-06, "loss": 0.4689, "step": 18882 }, { "epoch": 0.7915492072141936, "grad_norm": 1.8520549535751343, "learning_rate": 1.0966237319024459e-06, "loss": 0.5137, "step": 18883 }, { "epoch": 0.7915911258292027, "grad_norm": 2.609755516052246, "learning_rate": 1.09619952802071e-06, "loss": 0.5134, "step": 18884 }, { "epoch": 0.7916330444442116, "grad_norm": 1.897033452987671, "learning_rate": 1.0957753961001395e-06, "loss": 0.4476, "step": 18885 }, { "epoch": 0.7916749630592205, "grad_norm": 1.918880820274353, "learning_rate": 1.0953513361485529e-06, "loss": 0.5, "step": 18886 }, { "epoch": 0.7917168816742295, "grad_norm": 1.864707589149475, "learning_rate": 1.0949273481737654e-06, "loss": 0.4705, "step": 18887 }, { "epoch": 0.7917588002892384, "grad_norm": 1.8124523162841797, "learning_rate": 1.0945034321835934e-06, "loss": 0.4382, "step": 18888 }, { "epoch": 0.7918007189042474, "grad_norm": 2.219088077545166, "learning_rate": 1.0940795881858535e-06, "loss": 0.4717, "step": 18889 }, { "epoch": 0.7918426375192563, "grad_norm": 1.633101463317871, "learning_rate": 1.093655816188356e-06, "loss": 0.4989, "step": 18890 }, { "epoch": 0.7918845561342653, "grad_norm": 1.5525120496749878, "learning_rate": 1.0932321161989117e-06, "loss": 0.4213, "step": 18891 }, { "epoch": 0.7919264747492742, "grad_norm": 1.7296772003173828, "learning_rate": 1.0928084882253337e-06, "loss": 0.5016, "step": 18892 }, { "epoch": 0.7919683933642833, "grad_norm": 1.6426154375076294, "learning_rate": 1.0923849322754287e-06, "loss": 0.4419, "step": 18893 }, { "epoch": 0.7920103119792922, "grad_norm": 1.72688889503479, "learning_rate": 1.0919614483570051e-06, "loss": 0.4734, "step": 18894 }, { "epoch": 0.7920522305943012, "grad_norm": 2.547590732574463, "learning_rate": 1.091538036477871e-06, "loss": 0.4573, "step": 18895 }, { "epoch": 0.7920941492093101, "grad_norm": 1.5279837846755981, "learning_rate": 1.0911146966458285e-06, "loss": 0.4418, "step": 18896 }, { "epoch": 0.7921360678243191, "grad_norm": 2.1205039024353027, "learning_rate": 1.0906914288686831e-06, "loss": 0.464, "step": 18897 }, { "epoch": 0.792177986439328, "grad_norm": 1.9947006702423096, "learning_rate": 1.0902682331542375e-06, "loss": 0.5128, "step": 18898 }, { "epoch": 0.792219905054337, "grad_norm": 2.0577938556671143, "learning_rate": 1.0898451095102924e-06, "loss": 0.4953, "step": 18899 }, { "epoch": 0.7922618236693459, "grad_norm": 1.668103575706482, "learning_rate": 1.0894220579446456e-06, "loss": 0.4926, "step": 18900 }, { "epoch": 0.792303742284355, "grad_norm": 1.6907060146331787, "learning_rate": 1.0889990784650983e-06, "loss": 0.5232, "step": 18901 }, { "epoch": 0.7923456608993639, "grad_norm": 1.832960605621338, "learning_rate": 1.0885761710794446e-06, "loss": 0.5012, "step": 18902 }, { "epoch": 0.7923875795143729, "grad_norm": 2.3246426582336426, "learning_rate": 1.0881533357954821e-06, "loss": 0.5291, "step": 18903 }, { "epoch": 0.7924294981293818, "grad_norm": 1.7612582445144653, "learning_rate": 1.0877305726210064e-06, "loss": 0.4678, "step": 18904 }, { "epoch": 0.7924714167443908, "grad_norm": 2.1142168045043945, "learning_rate": 1.0873078815638072e-06, "loss": 0.4926, "step": 18905 }, { "epoch": 0.7925133353593997, "grad_norm": 2.6094863414764404, "learning_rate": 1.0868852626316785e-06, "loss": 0.4992, "step": 18906 }, { "epoch": 0.7925552539744087, "grad_norm": 2.487424373626709, "learning_rate": 1.0864627158324109e-06, "loss": 0.4791, "step": 18907 }, { "epoch": 0.7925971725894176, "grad_norm": 1.8575551509857178, "learning_rate": 1.086040241173792e-06, "loss": 0.4265, "step": 18908 }, { "epoch": 0.7926390912044267, "grad_norm": 2.040168285369873, "learning_rate": 1.0856178386636117e-06, "loss": 0.5145, "step": 18909 }, { "epoch": 0.7926810098194356, "grad_norm": 1.8078439235687256, "learning_rate": 1.0851955083096556e-06, "loss": 0.5017, "step": 18910 }, { "epoch": 0.7927229284344445, "grad_norm": 1.7520688772201538, "learning_rate": 1.0847732501197062e-06, "loss": 0.4916, "step": 18911 }, { "epoch": 0.7927648470494535, "grad_norm": 2.035996437072754, "learning_rate": 1.0843510641015492e-06, "loss": 0.5174, "step": 18912 }, { "epoch": 0.7928067656644624, "grad_norm": 2.032205820083618, "learning_rate": 1.083928950262969e-06, "loss": 0.4458, "step": 18913 }, { "epoch": 0.7928486842794714, "grad_norm": 1.6786799430847168, "learning_rate": 1.0835069086117434e-06, "loss": 0.4846, "step": 18914 }, { "epoch": 0.7928906028944803, "grad_norm": 1.7334339618682861, "learning_rate": 1.083084939155653e-06, "loss": 0.4831, "step": 18915 }, { "epoch": 0.7929325215094893, "grad_norm": 5.354379177093506, "learning_rate": 1.0826630419024786e-06, "loss": 0.4367, "step": 18916 }, { "epoch": 0.7929744401244982, "grad_norm": 1.7953046560287476, "learning_rate": 1.0822412168599938e-06, "loss": 0.4758, "step": 18917 }, { "epoch": 0.7930163587395073, "grad_norm": 3.0745270252227783, "learning_rate": 1.0818194640359776e-06, "loss": 0.4713, "step": 18918 }, { "epoch": 0.7930582773545162, "grad_norm": 2.343068838119507, "learning_rate": 1.0813977834382028e-06, "loss": 0.4741, "step": 18919 }, { "epoch": 0.7931001959695252, "grad_norm": 1.9490797519683838, "learning_rate": 1.0809761750744413e-06, "loss": 0.4826, "step": 18920 }, { "epoch": 0.7931421145845341, "grad_norm": 1.9170398712158203, "learning_rate": 1.0805546389524657e-06, "loss": 0.4528, "step": 18921 }, { "epoch": 0.7931840331995431, "grad_norm": 2.142176628112793, "learning_rate": 1.0801331750800486e-06, "loss": 0.4883, "step": 18922 }, { "epoch": 0.793225951814552, "grad_norm": 1.650801658630371, "learning_rate": 1.0797117834649562e-06, "loss": 0.5161, "step": 18923 }, { "epoch": 0.793267870429561, "grad_norm": 2.0428683757781982, "learning_rate": 1.079290464114957e-06, "loss": 0.4406, "step": 18924 }, { "epoch": 0.79330978904457, "grad_norm": 4.005202770233154, "learning_rate": 1.0788692170378196e-06, "loss": 0.5229, "step": 18925 }, { "epoch": 0.793351707659579, "grad_norm": 1.7602105140686035, "learning_rate": 1.0784480422413056e-06, "loss": 0.4853, "step": 18926 }, { "epoch": 0.7933936262745879, "grad_norm": 1.929434061050415, "learning_rate": 1.0780269397331827e-06, "loss": 0.489, "step": 18927 }, { "epoch": 0.7934355448895969, "grad_norm": 2.1334125995635986, "learning_rate": 1.0776059095212105e-06, "loss": 0.4661, "step": 18928 }, { "epoch": 0.7934774635046058, "grad_norm": 3.1909823417663574, "learning_rate": 1.0771849516131493e-06, "loss": 0.5028, "step": 18929 }, { "epoch": 0.7935193821196148, "grad_norm": 1.8459166288375854, "learning_rate": 1.0767640660167612e-06, "loss": 0.4438, "step": 18930 }, { "epoch": 0.7935613007346237, "grad_norm": 1.8255689144134521, "learning_rate": 1.0763432527398042e-06, "loss": 0.4198, "step": 18931 }, { "epoch": 0.7936032193496327, "grad_norm": 2.3277976512908936, "learning_rate": 1.0759225117900346e-06, "loss": 0.5015, "step": 18932 }, { "epoch": 0.7936451379646416, "grad_norm": 1.8634037971496582, "learning_rate": 1.0755018431752078e-06, "loss": 0.4747, "step": 18933 }, { "epoch": 0.7936870565796507, "grad_norm": 1.8970998525619507, "learning_rate": 1.0750812469030814e-06, "loss": 0.4826, "step": 18934 }, { "epoch": 0.7937289751946596, "grad_norm": 2.271265745162964, "learning_rate": 1.0746607229814038e-06, "loss": 0.4916, "step": 18935 }, { "epoch": 0.7937708938096685, "grad_norm": 1.787083387374878, "learning_rate": 1.0742402714179312e-06, "loss": 0.5073, "step": 18936 }, { "epoch": 0.7938128124246775, "grad_norm": 2.505208969116211, "learning_rate": 1.0738198922204119e-06, "loss": 0.4986, "step": 18937 }, { "epoch": 0.7938547310396864, "grad_norm": 1.8464628458023071, "learning_rate": 1.0733995853965934e-06, "loss": 0.4436, "step": 18938 }, { "epoch": 0.7938966496546954, "grad_norm": 1.7333133220672607, "learning_rate": 1.0729793509542252e-06, "loss": 0.5763, "step": 18939 }, { "epoch": 0.7939385682697043, "grad_norm": 2.113129138946533, "learning_rate": 1.0725591889010555e-06, "loss": 0.4941, "step": 18940 }, { "epoch": 0.7939804868847133, "grad_norm": 1.6811723709106445, "learning_rate": 1.0721390992448261e-06, "loss": 0.4458, "step": 18941 }, { "epoch": 0.7940224054997223, "grad_norm": 2.886328935623169, "learning_rate": 1.0717190819932822e-06, "loss": 0.4831, "step": 18942 }, { "epoch": 0.7940643241147313, "grad_norm": 2.3675270080566406, "learning_rate": 1.0712991371541676e-06, "loss": 0.4735, "step": 18943 }, { "epoch": 0.7941062427297402, "grad_norm": 1.901224970817566, "learning_rate": 1.0708792647352212e-06, "loss": 0.4783, "step": 18944 }, { "epoch": 0.7941481613447492, "grad_norm": 1.7032712697982788, "learning_rate": 1.0704594647441846e-06, "loss": 0.4581, "step": 18945 }, { "epoch": 0.7941900799597581, "grad_norm": 1.6790629625320435, "learning_rate": 1.0700397371887956e-06, "loss": 0.4457, "step": 18946 }, { "epoch": 0.7942319985747671, "grad_norm": 1.7710288763046265, "learning_rate": 1.0696200820767894e-06, "loss": 0.5085, "step": 18947 }, { "epoch": 0.794273917189776, "grad_norm": 1.9328759908676147, "learning_rate": 1.0692004994159038e-06, "loss": 0.4492, "step": 18948 }, { "epoch": 0.794315835804785, "grad_norm": 1.6471061706542969, "learning_rate": 1.068780989213874e-06, "loss": 0.4568, "step": 18949 }, { "epoch": 0.794357754419794, "grad_norm": 2.366009473800659, "learning_rate": 1.0683615514784307e-06, "loss": 0.4679, "step": 18950 }, { "epoch": 0.794399673034803, "grad_norm": 1.8685232400894165, "learning_rate": 1.067942186217308e-06, "loss": 0.4976, "step": 18951 }, { "epoch": 0.7944415916498119, "grad_norm": 1.8121039867401123, "learning_rate": 1.0675228934382338e-06, "loss": 0.472, "step": 18952 }, { "epoch": 0.7944835102648209, "grad_norm": 4.450352668762207, "learning_rate": 1.0671036731489387e-06, "loss": 0.4658, "step": 18953 }, { "epoch": 0.7945254288798298, "grad_norm": 2.0102503299713135, "learning_rate": 1.0666845253571516e-06, "loss": 0.4883, "step": 18954 }, { "epoch": 0.7945673474948388, "grad_norm": 1.7634267807006836, "learning_rate": 1.0662654500705965e-06, "loss": 0.4529, "step": 18955 }, { "epoch": 0.7946092661098477, "grad_norm": 2.4343864917755127, "learning_rate": 1.065846447297001e-06, "loss": 0.5269, "step": 18956 }, { "epoch": 0.7946511847248567, "grad_norm": 1.7504736185073853, "learning_rate": 1.065427517044087e-06, "loss": 0.4348, "step": 18957 }, { "epoch": 0.7946931033398656, "grad_norm": 1.7695635557174683, "learning_rate": 1.0650086593195764e-06, "loss": 0.4869, "step": 18958 }, { "epoch": 0.7947350219548747, "grad_norm": 1.814453125, "learning_rate": 1.064589874131191e-06, "loss": 0.4995, "step": 18959 }, { "epoch": 0.7947769405698836, "grad_norm": 2.7504920959472656, "learning_rate": 1.0641711614866518e-06, "loss": 0.4948, "step": 18960 }, { "epoch": 0.7948188591848925, "grad_norm": 2.402071475982666, "learning_rate": 1.063752521393675e-06, "loss": 0.5191, "step": 18961 }, { "epoch": 0.7948607777999015, "grad_norm": 2.1979100704193115, "learning_rate": 1.063333953859979e-06, "loss": 0.5065, "step": 18962 }, { "epoch": 0.7949026964149104, "grad_norm": 2.7470579147338867, "learning_rate": 1.0629154588932805e-06, "loss": 0.437, "step": 18963 }, { "epoch": 0.7949446150299194, "grad_norm": 2.136554718017578, "learning_rate": 1.0624970365012915e-06, "loss": 0.4878, "step": 18964 }, { "epoch": 0.7949865336449283, "grad_norm": 1.6880300045013428, "learning_rate": 1.062078686691727e-06, "loss": 0.4459, "step": 18965 }, { "epoch": 0.7950284522599373, "grad_norm": 1.7014793157577515, "learning_rate": 1.0616604094722982e-06, "loss": 0.4652, "step": 18966 }, { "epoch": 0.7950703708749463, "grad_norm": 1.9314733743667603, "learning_rate": 1.0612422048507136e-06, "loss": 0.4712, "step": 18967 }, { "epoch": 0.7951122894899553, "grad_norm": 1.8008944988250732, "learning_rate": 1.0608240728346842e-06, "loss": 0.5374, "step": 18968 }, { "epoch": 0.7951542081049642, "grad_norm": 1.6676143407821655, "learning_rate": 1.0604060134319188e-06, "loss": 0.4461, "step": 18969 }, { "epoch": 0.7951961267199732, "grad_norm": 1.7081385850906372, "learning_rate": 1.0599880266501206e-06, "loss": 0.4772, "step": 18970 }, { "epoch": 0.7952380453349821, "grad_norm": 1.8631997108459473, "learning_rate": 1.059570112496996e-06, "loss": 0.4733, "step": 18971 }, { "epoch": 0.7952799639499911, "grad_norm": 2.029985189437866, "learning_rate": 1.0591522709802505e-06, "loss": 0.5294, "step": 18972 }, { "epoch": 0.795321882565, "grad_norm": 1.834768295288086, "learning_rate": 1.0587345021075828e-06, "loss": 0.4493, "step": 18973 }, { "epoch": 0.795363801180009, "grad_norm": 1.8394362926483154, "learning_rate": 1.0583168058866982e-06, "loss": 0.4579, "step": 18974 }, { "epoch": 0.795405719795018, "grad_norm": 1.9655964374542236, "learning_rate": 1.0578991823252938e-06, "loss": 0.4908, "step": 18975 }, { "epoch": 0.795447638410027, "grad_norm": 1.920845866203308, "learning_rate": 1.0574816314310665e-06, "loss": 0.4025, "step": 18976 }, { "epoch": 0.7954895570250359, "grad_norm": 2.338744640350342, "learning_rate": 1.0570641532117154e-06, "loss": 0.4578, "step": 18977 }, { "epoch": 0.7955314756400449, "grad_norm": 1.6579945087432861, "learning_rate": 1.0566467476749364e-06, "loss": 0.4908, "step": 18978 }, { "epoch": 0.7955733942550538, "grad_norm": 1.687130331993103, "learning_rate": 1.0562294148284225e-06, "loss": 0.4519, "step": 18979 }, { "epoch": 0.7956153128700628, "grad_norm": 1.9260269403457642, "learning_rate": 1.0558121546798667e-06, "loss": 0.4985, "step": 18980 }, { "epoch": 0.7956572314850717, "grad_norm": 1.7700732946395874, "learning_rate": 1.055394967236963e-06, "loss": 0.4932, "step": 18981 }, { "epoch": 0.7956991501000807, "grad_norm": 1.735804557800293, "learning_rate": 1.054977852507398e-06, "loss": 0.4916, "step": 18982 }, { "epoch": 0.7957410687150896, "grad_norm": 1.9750573635101318, "learning_rate": 1.0545608104988643e-06, "loss": 0.5131, "step": 18983 }, { "epoch": 0.7957829873300987, "grad_norm": 1.8075697422027588, "learning_rate": 1.0541438412190468e-06, "loss": 0.4506, "step": 18984 }, { "epoch": 0.7958249059451076, "grad_norm": 1.8454434871673584, "learning_rate": 1.053726944675632e-06, "loss": 0.4679, "step": 18985 }, { "epoch": 0.7958668245601165, "grad_norm": 1.7526893615722656, "learning_rate": 1.0533101208763051e-06, "loss": 0.4517, "step": 18986 }, { "epoch": 0.7959087431751255, "grad_norm": 1.574100136756897, "learning_rate": 1.0528933698287513e-06, "loss": 0.4639, "step": 18987 }, { "epoch": 0.7959506617901344, "grad_norm": 1.6258636713027954, "learning_rate": 1.0524766915406503e-06, "loss": 0.4664, "step": 18988 }, { "epoch": 0.7959925804051434, "grad_norm": 2.164750099182129, "learning_rate": 1.0520600860196845e-06, "loss": 0.4941, "step": 18989 }, { "epoch": 0.7960344990201523, "grad_norm": 1.8871310949325562, "learning_rate": 1.0516435532735342e-06, "loss": 0.4494, "step": 18990 }, { "epoch": 0.7960764176351613, "grad_norm": 1.66608726978302, "learning_rate": 1.051227093309875e-06, "loss": 0.4537, "step": 18991 }, { "epoch": 0.7961183362501703, "grad_norm": 2.0961170196533203, "learning_rate": 1.050810706136387e-06, "loss": 0.4554, "step": 18992 }, { "epoch": 0.7961602548651793, "grad_norm": 2.589672327041626, "learning_rate": 1.050394391760744e-06, "loss": 0.5359, "step": 18993 }, { "epoch": 0.7962021734801882, "grad_norm": 1.9403108358383179, "learning_rate": 1.0499781501906187e-06, "loss": 0.5061, "step": 18994 }, { "epoch": 0.7962440920951972, "grad_norm": 1.9652130603790283, "learning_rate": 1.049561981433686e-06, "loss": 0.4699, "step": 18995 }, { "epoch": 0.7962860107102061, "grad_norm": 1.6180801391601562, "learning_rate": 1.0491458854976178e-06, "loss": 0.4334, "step": 18996 }, { "epoch": 0.7963279293252151, "grad_norm": 1.5565598011016846, "learning_rate": 1.0487298623900815e-06, "loss": 0.4663, "step": 18997 }, { "epoch": 0.796369847940224, "grad_norm": 1.5019501447677612, "learning_rate": 1.0483139121187486e-06, "loss": 0.4168, "step": 18998 }, { "epoch": 0.796411766555233, "grad_norm": 1.5869677066802979, "learning_rate": 1.0478980346912865e-06, "loss": 0.4256, "step": 18999 }, { "epoch": 0.796453685170242, "grad_norm": 1.9163161516189575, "learning_rate": 1.0474822301153592e-06, "loss": 0.4795, "step": 19000 }, { "epoch": 0.796495603785251, "grad_norm": 2.2736945152282715, "learning_rate": 1.047066498398634e-06, "loss": 0.5309, "step": 19001 }, { "epoch": 0.7965375224002599, "grad_norm": 2.0184264183044434, "learning_rate": 1.0466508395487718e-06, "loss": 0.4956, "step": 19002 }, { "epoch": 0.7965794410152689, "grad_norm": 1.7838610410690308, "learning_rate": 1.0462352535734376e-06, "loss": 0.4932, "step": 19003 }, { "epoch": 0.7966213596302778, "grad_norm": 1.8812280893325806, "learning_rate": 1.0458197404802895e-06, "loss": 0.4672, "step": 19004 }, { "epoch": 0.7966632782452868, "grad_norm": 1.6009470224380493, "learning_rate": 1.0454043002769892e-06, "loss": 0.4472, "step": 19005 }, { "epoch": 0.7967051968602957, "grad_norm": 1.8925470113754272, "learning_rate": 1.0449889329711921e-06, "loss": 0.4819, "step": 19006 }, { "epoch": 0.7967471154753047, "grad_norm": 1.9660450220108032, "learning_rate": 1.0445736385705563e-06, "loss": 0.5062, "step": 19007 }, { "epoch": 0.7967890340903137, "grad_norm": 1.8590331077575684, "learning_rate": 1.0441584170827385e-06, "loss": 0.4546, "step": 19008 }, { "epoch": 0.7968309527053227, "grad_norm": 3.019827127456665, "learning_rate": 1.0437432685153902e-06, "loss": 0.4359, "step": 19009 }, { "epoch": 0.7968728713203316, "grad_norm": 2.009315013885498, "learning_rate": 1.0433281928761658e-06, "loss": 0.5119, "step": 19010 }, { "epoch": 0.7969147899353405, "grad_norm": 2.1485846042633057, "learning_rate": 1.0429131901727168e-06, "loss": 0.4922, "step": 19011 }, { "epoch": 0.7969567085503495, "grad_norm": 1.7738397121429443, "learning_rate": 1.042498260412693e-06, "loss": 0.5063, "step": 19012 }, { "epoch": 0.7969986271653584, "grad_norm": 2.062220811843872, "learning_rate": 1.0420834036037413e-06, "loss": 0.4999, "step": 19013 }, { "epoch": 0.7970405457803674, "grad_norm": 1.9117324352264404, "learning_rate": 1.041668619753512e-06, "loss": 0.4964, "step": 19014 }, { "epoch": 0.7970824643953763, "grad_norm": 1.8155407905578613, "learning_rate": 1.0412539088696483e-06, "loss": 0.4915, "step": 19015 }, { "epoch": 0.7971243830103854, "grad_norm": 1.788620114326477, "learning_rate": 1.0408392709597953e-06, "loss": 0.4587, "step": 19016 }, { "epoch": 0.7971663016253943, "grad_norm": 1.9217867851257324, "learning_rate": 1.0404247060315993e-06, "loss": 0.4797, "step": 19017 }, { "epoch": 0.7972082202404033, "grad_norm": 1.918312668800354, "learning_rate": 1.0400102140926976e-06, "loss": 0.4442, "step": 19018 }, { "epoch": 0.7972501388554122, "grad_norm": 1.7210938930511475, "learning_rate": 1.0395957951507336e-06, "loss": 0.441, "step": 19019 }, { "epoch": 0.7972920574704212, "grad_norm": 1.4576102495193481, "learning_rate": 1.0391814492133467e-06, "loss": 0.4515, "step": 19020 }, { "epoch": 0.7973339760854301, "grad_norm": 1.6912879943847656, "learning_rate": 1.038767176288175e-06, "loss": 0.5098, "step": 19021 }, { "epoch": 0.7973758947004391, "grad_norm": 3.0451149940490723, "learning_rate": 1.0383529763828525e-06, "loss": 0.4807, "step": 19022 }, { "epoch": 0.797417813315448, "grad_norm": 1.6585017442703247, "learning_rate": 1.0379388495050174e-06, "loss": 0.4762, "step": 19023 }, { "epoch": 0.797459731930457, "grad_norm": 1.9568166732788086, "learning_rate": 1.0375247956623009e-06, "loss": 0.5152, "step": 19024 }, { "epoch": 0.797501650545466, "grad_norm": 1.799176573753357, "learning_rate": 1.0371108148623365e-06, "loss": 0.4615, "step": 19025 }, { "epoch": 0.797543569160475, "grad_norm": 2.404005289077759, "learning_rate": 1.0366969071127569e-06, "loss": 0.4662, "step": 19026 }, { "epoch": 0.7975854877754839, "grad_norm": 1.8767269849777222, "learning_rate": 1.0362830724211902e-06, "loss": 0.5408, "step": 19027 }, { "epoch": 0.7976274063904929, "grad_norm": 1.7773804664611816, "learning_rate": 1.035869310795266e-06, "loss": 0.4789, "step": 19028 }, { "epoch": 0.7976693250055018, "grad_norm": 1.573854684829712, "learning_rate": 1.03545562224261e-06, "loss": 0.45, "step": 19029 }, { "epoch": 0.7977112436205108, "grad_norm": 1.6632256507873535, "learning_rate": 1.03504200677085e-06, "loss": 0.4346, "step": 19030 }, { "epoch": 0.7977531622355197, "grad_norm": 1.799338698387146, "learning_rate": 1.0346284643876088e-06, "loss": 0.462, "step": 19031 }, { "epoch": 0.7977950808505287, "grad_norm": 2.1615419387817383, "learning_rate": 1.0342149951005088e-06, "loss": 0.4755, "step": 19032 }, { "epoch": 0.7978369994655377, "grad_norm": 2.088452100753784, "learning_rate": 1.0338015989171728e-06, "loss": 0.4645, "step": 19033 }, { "epoch": 0.7978789180805467, "grad_norm": 1.7383272647857666, "learning_rate": 1.033388275845223e-06, "loss": 0.4593, "step": 19034 }, { "epoch": 0.7979208366955556, "grad_norm": 2.0103940963745117, "learning_rate": 1.032975025892275e-06, "loss": 0.5293, "step": 19035 }, { "epoch": 0.7979627553105645, "grad_norm": 1.710835337638855, "learning_rate": 1.0325618490659477e-06, "loss": 0.4485, "step": 19036 }, { "epoch": 0.7980046739255735, "grad_norm": 4.106670379638672, "learning_rate": 1.03214874537386e-06, "loss": 0.498, "step": 19037 }, { "epoch": 0.7980465925405824, "grad_norm": 1.7399418354034424, "learning_rate": 1.0317357148236228e-06, "loss": 0.5068, "step": 19038 }, { "epoch": 0.7980885111555914, "grad_norm": 1.8610846996307373, "learning_rate": 1.0313227574228535e-06, "loss": 0.4409, "step": 19039 }, { "epoch": 0.7981304297706003, "grad_norm": 1.756898045539856, "learning_rate": 1.0309098731791628e-06, "loss": 0.5221, "step": 19040 }, { "epoch": 0.7981723483856094, "grad_norm": 1.9951589107513428, "learning_rate": 1.0304970621001598e-06, "loss": 0.4774, "step": 19041 }, { "epoch": 0.7982142670006183, "grad_norm": 2.0607357025146484, "learning_rate": 1.0300843241934555e-06, "loss": 0.5235, "step": 19042 }, { "epoch": 0.7982561856156273, "grad_norm": 1.8327076435089111, "learning_rate": 1.0296716594666606e-06, "loss": 0.4819, "step": 19043 }, { "epoch": 0.7982981042306362, "grad_norm": 1.8346281051635742, "learning_rate": 1.029259067927378e-06, "loss": 0.4776, "step": 19044 }, { "epoch": 0.7983400228456452, "grad_norm": 2.3942832946777344, "learning_rate": 1.028846549583215e-06, "loss": 0.4944, "step": 19045 }, { "epoch": 0.7983819414606541, "grad_norm": 1.8055188655853271, "learning_rate": 1.0284341044417772e-06, "loss": 0.4536, "step": 19046 }, { "epoch": 0.7984238600756631, "grad_norm": 1.683530330657959, "learning_rate": 1.0280217325106652e-06, "loss": 0.4487, "step": 19047 }, { "epoch": 0.798465778690672, "grad_norm": 3.0775907039642334, "learning_rate": 1.0276094337974823e-06, "loss": 0.4463, "step": 19048 }, { "epoch": 0.798507697305681, "grad_norm": 3.654977798461914, "learning_rate": 1.027197208309827e-06, "loss": 0.4445, "step": 19049 }, { "epoch": 0.79854961592069, "grad_norm": 1.660445213317871, "learning_rate": 1.0267850560553e-06, "loss": 0.5294, "step": 19050 }, { "epoch": 0.798591534535699, "grad_norm": 1.77147376537323, "learning_rate": 1.0263729770414965e-06, "loss": 0.5211, "step": 19051 }, { "epoch": 0.7986334531507079, "grad_norm": 1.663713812828064, "learning_rate": 1.025960971276015e-06, "loss": 0.4391, "step": 19052 }, { "epoch": 0.7986753717657169, "grad_norm": 1.7382304668426514, "learning_rate": 1.0255490387664475e-06, "loss": 0.5308, "step": 19053 }, { "epoch": 0.7987172903807258, "grad_norm": 1.9346879720687866, "learning_rate": 1.0251371795203895e-06, "loss": 0.4183, "step": 19054 }, { "epoch": 0.7987592089957348, "grad_norm": 1.7593141794204712, "learning_rate": 1.0247253935454331e-06, "loss": 0.4977, "step": 19055 }, { "epoch": 0.7988011276107437, "grad_norm": 1.6766735315322876, "learning_rate": 1.0243136808491676e-06, "loss": 0.5175, "step": 19056 }, { "epoch": 0.7988430462257528, "grad_norm": 1.8466951847076416, "learning_rate": 1.0239020414391831e-06, "loss": 0.5194, "step": 19057 }, { "epoch": 0.7988849648407617, "grad_norm": 1.3895634412765503, "learning_rate": 1.023490475323069e-06, "loss": 0.4288, "step": 19058 }, { "epoch": 0.7989268834557707, "grad_norm": 1.8050270080566406, "learning_rate": 1.023078982508411e-06, "loss": 0.5316, "step": 19059 }, { "epoch": 0.7989688020707796, "grad_norm": 3.043383836746216, "learning_rate": 1.0226675630027922e-06, "loss": 0.4376, "step": 19060 }, { "epoch": 0.7990107206857885, "grad_norm": 1.6052320003509521, "learning_rate": 1.0222562168138e-06, "loss": 0.4728, "step": 19061 }, { "epoch": 0.7990526393007975, "grad_norm": 1.9444139003753662, "learning_rate": 1.0218449439490136e-06, "loss": 0.4432, "step": 19062 }, { "epoch": 0.7990945579158064, "grad_norm": 1.9504737854003906, "learning_rate": 1.0214337444160165e-06, "loss": 0.4798, "step": 19063 }, { "epoch": 0.7991364765308154, "grad_norm": 1.9311726093292236, "learning_rate": 1.0210226182223893e-06, "loss": 0.4607, "step": 19064 }, { "epoch": 0.7991783951458243, "grad_norm": 2.3623409271240234, "learning_rate": 1.0206115653757082e-06, "loss": 0.4456, "step": 19065 }, { "epoch": 0.7992203137608334, "grad_norm": 1.9701677560806274, "learning_rate": 1.0202005858835512e-06, "loss": 0.4679, "step": 19066 }, { "epoch": 0.7992622323758423, "grad_norm": 2.169684648513794, "learning_rate": 1.0197896797534956e-06, "loss": 0.53, "step": 19067 }, { "epoch": 0.7993041509908513, "grad_norm": 1.7149533033370972, "learning_rate": 1.0193788469931155e-06, "loss": 0.4594, "step": 19068 }, { "epoch": 0.7993460696058602, "grad_norm": 1.6110228300094604, "learning_rate": 1.0189680876099817e-06, "loss": 0.4305, "step": 19069 }, { "epoch": 0.7993879882208692, "grad_norm": 1.9476943016052246, "learning_rate": 1.018557401611669e-06, "loss": 0.5287, "step": 19070 }, { "epoch": 0.7994299068358781, "grad_norm": 1.6779118776321411, "learning_rate": 1.0181467890057444e-06, "loss": 0.4185, "step": 19071 }, { "epoch": 0.7994718254508871, "grad_norm": 2.155207872390747, "learning_rate": 1.01773624979978e-06, "loss": 0.5078, "step": 19072 }, { "epoch": 0.799513744065896, "grad_norm": 3.556304454803467, "learning_rate": 1.0173257840013433e-06, "loss": 0.468, "step": 19073 }, { "epoch": 0.799555662680905, "grad_norm": 1.6888480186462402, "learning_rate": 1.0169153916179986e-06, "loss": 0.4867, "step": 19074 }, { "epoch": 0.799597581295914, "grad_norm": 1.798616886138916, "learning_rate": 1.0165050726573123e-06, "loss": 0.4511, "step": 19075 }, { "epoch": 0.799639499910923, "grad_norm": 1.8829309940338135, "learning_rate": 1.016094827126849e-06, "loss": 0.4591, "step": 19076 }, { "epoch": 0.7996814185259319, "grad_norm": 1.8198474645614624, "learning_rate": 1.0156846550341704e-06, "loss": 0.4408, "step": 19077 }, { "epoch": 0.7997233371409409, "grad_norm": 2.7309601306915283, "learning_rate": 1.0152745563868356e-06, "loss": 0.51, "step": 19078 }, { "epoch": 0.7997652557559498, "grad_norm": 1.7760850191116333, "learning_rate": 1.0148645311924065e-06, "loss": 0.4674, "step": 19079 }, { "epoch": 0.7998071743709588, "grad_norm": 1.704298496246338, "learning_rate": 1.01445457945844e-06, "loss": 0.4872, "step": 19080 }, { "epoch": 0.7998490929859677, "grad_norm": 1.8484476804733276, "learning_rate": 1.0140447011924931e-06, "loss": 0.4389, "step": 19081 }, { "epoch": 0.7998910116009768, "grad_norm": 1.7861276865005493, "learning_rate": 1.013634896402123e-06, "loss": 0.4912, "step": 19082 }, { "epoch": 0.7999329302159857, "grad_norm": 2.2769064903259277, "learning_rate": 1.013225165094881e-06, "loss": 0.4942, "step": 19083 }, { "epoch": 0.7999748488309947, "grad_norm": 1.8221505880355835, "learning_rate": 1.012815507278322e-06, "loss": 0.4723, "step": 19084 }, { "epoch": 0.8000167674460036, "grad_norm": 3.5523929595947266, "learning_rate": 1.012405922959998e-06, "loss": 0.4493, "step": 19085 }, { "epoch": 0.8000586860610125, "grad_norm": 2.0210118293762207, "learning_rate": 1.0119964121474584e-06, "loss": 0.3969, "step": 19086 }, { "epoch": 0.8001006046760215, "grad_norm": 1.6792926788330078, "learning_rate": 1.0115869748482499e-06, "loss": 0.5404, "step": 19087 }, { "epoch": 0.8001425232910304, "grad_norm": 1.9288896322250366, "learning_rate": 1.0111776110699234e-06, "loss": 0.4844, "step": 19088 }, { "epoch": 0.8001844419060394, "grad_norm": 2.0418899059295654, "learning_rate": 1.010768320820022e-06, "loss": 0.4643, "step": 19089 }, { "epoch": 0.8002263605210483, "grad_norm": 2.0368852615356445, "learning_rate": 1.0103591041060917e-06, "loss": 0.4551, "step": 19090 }, { "epoch": 0.8002682791360574, "grad_norm": 1.5435796976089478, "learning_rate": 1.0099499609356767e-06, "loss": 0.4565, "step": 19091 }, { "epoch": 0.8003101977510663, "grad_norm": 1.862259030342102, "learning_rate": 1.0095408913163169e-06, "loss": 0.4878, "step": 19092 }, { "epoch": 0.8003521163660753, "grad_norm": 1.6522417068481445, "learning_rate": 1.0091318952555545e-06, "loss": 0.402, "step": 19093 }, { "epoch": 0.8003940349810842, "grad_norm": 1.8620702028274536, "learning_rate": 1.0087229727609294e-06, "loss": 0.5061, "step": 19094 }, { "epoch": 0.8004359535960932, "grad_norm": 1.7141025066375732, "learning_rate": 1.0083141238399785e-06, "loss": 0.4517, "step": 19095 }, { "epoch": 0.8004778722111021, "grad_norm": 1.6555243730545044, "learning_rate": 1.0079053485002372e-06, "loss": 0.4254, "step": 19096 }, { "epoch": 0.8005197908261111, "grad_norm": 1.6509976387023926, "learning_rate": 1.007496646749243e-06, "loss": 0.4693, "step": 19097 }, { "epoch": 0.80056170944112, "grad_norm": 2.175758123397827, "learning_rate": 1.0070880185945276e-06, "loss": 0.4714, "step": 19098 }, { "epoch": 0.8006036280561291, "grad_norm": 2.0063769817352295, "learning_rate": 1.0066794640436244e-06, "loss": 0.4645, "step": 19099 }, { "epoch": 0.800645546671138, "grad_norm": 1.7631710767745972, "learning_rate": 1.0062709831040661e-06, "loss": 0.4895, "step": 19100 }, { "epoch": 0.800687465286147, "grad_norm": 2.029247760772705, "learning_rate": 1.0058625757833795e-06, "loss": 0.4281, "step": 19101 }, { "epoch": 0.8007293839011559, "grad_norm": 2.1416501998901367, "learning_rate": 1.0054542420890962e-06, "loss": 0.4986, "step": 19102 }, { "epoch": 0.8007713025161649, "grad_norm": 2.084137439727783, "learning_rate": 1.0050459820287405e-06, "loss": 0.4512, "step": 19103 }, { "epoch": 0.8008132211311738, "grad_norm": 1.7098346948623657, "learning_rate": 1.004637795609839e-06, "loss": 0.4182, "step": 19104 }, { "epoch": 0.8008551397461828, "grad_norm": 1.8761701583862305, "learning_rate": 1.0042296828399172e-06, "loss": 0.4859, "step": 19105 }, { "epoch": 0.8008970583611917, "grad_norm": 1.8448845148086548, "learning_rate": 1.0038216437264981e-06, "loss": 0.4584, "step": 19106 }, { "epoch": 0.8009389769762008, "grad_norm": 2.466907262802124, "learning_rate": 1.0034136782771003e-06, "loss": 0.4335, "step": 19107 }, { "epoch": 0.8009808955912097, "grad_norm": 1.7398664951324463, "learning_rate": 1.0030057864992482e-06, "loss": 0.5158, "step": 19108 }, { "epoch": 0.8010228142062187, "grad_norm": 3.33376145362854, "learning_rate": 1.0025979684004573e-06, "loss": 0.4486, "step": 19109 }, { "epoch": 0.8010647328212276, "grad_norm": 1.6634806394577026, "learning_rate": 1.0021902239882464e-06, "loss": 0.4606, "step": 19110 }, { "epoch": 0.8011066514362365, "grad_norm": 1.9554479122161865, "learning_rate": 1.001782553270133e-06, "loss": 0.5141, "step": 19111 }, { "epoch": 0.8011485700512455, "grad_norm": 2.971341371536255, "learning_rate": 1.00137495625363e-06, "loss": 0.472, "step": 19112 }, { "epoch": 0.8011904886662544, "grad_norm": 1.9066829681396484, "learning_rate": 1.0009674329462515e-06, "loss": 0.5667, "step": 19113 }, { "epoch": 0.8012324072812634, "grad_norm": 1.8383413553237915, "learning_rate": 1.0005599833555114e-06, "loss": 0.4231, "step": 19114 }, { "epoch": 0.8012743258962723, "grad_norm": 2.7947707176208496, "learning_rate": 1.000152607488919e-06, "loss": 0.5043, "step": 19115 }, { "epoch": 0.8013162445112814, "grad_norm": 2.0765256881713867, "learning_rate": 9.99745305353982e-07, "loss": 0.4462, "step": 19116 }, { "epoch": 0.8013581631262903, "grad_norm": 1.6573728322982788, "learning_rate": 9.993380769582113e-07, "loss": 0.5104, "step": 19117 }, { "epoch": 0.8014000817412993, "grad_norm": 2.443497657775879, "learning_rate": 9.989309223091116e-07, "loss": 0.4416, "step": 19118 }, { "epoch": 0.8014420003563082, "grad_norm": 1.9920011758804321, "learning_rate": 9.985238414141886e-07, "loss": 0.4459, "step": 19119 }, { "epoch": 0.8014839189713172, "grad_norm": 2.117764472961426, "learning_rate": 9.981168342809482e-07, "loss": 0.4324, "step": 19120 }, { "epoch": 0.8015258375863261, "grad_norm": 1.7163890600204468, "learning_rate": 9.977099009168906e-07, "loss": 0.4396, "step": 19121 }, { "epoch": 0.8015677562013351, "grad_norm": 2.355905294418335, "learning_rate": 9.973030413295175e-07, "loss": 0.479, "step": 19122 }, { "epoch": 0.801609674816344, "grad_norm": 1.8215830326080322, "learning_rate": 9.968962555263306e-07, "loss": 0.4956, "step": 19123 }, { "epoch": 0.8016515934313531, "grad_norm": 1.9482626914978027, "learning_rate": 9.96489543514827e-07, "loss": 0.4884, "step": 19124 }, { "epoch": 0.801693512046362, "grad_norm": 1.9655239582061768, "learning_rate": 9.960829053025028e-07, "loss": 0.476, "step": 19125 }, { "epoch": 0.801735430661371, "grad_norm": 1.7723100185394287, "learning_rate": 9.956763408968567e-07, "loss": 0.4463, "step": 19126 }, { "epoch": 0.8017773492763799, "grad_norm": 1.9947192668914795, "learning_rate": 9.952698503053799e-07, "loss": 0.4723, "step": 19127 }, { "epoch": 0.8018192678913889, "grad_norm": 1.9970303773880005, "learning_rate": 9.948634335355667e-07, "loss": 0.5301, "step": 19128 }, { "epoch": 0.8018611865063978, "grad_norm": 2.027989387512207, "learning_rate": 9.944570905949108e-07, "loss": 0.4786, "step": 19129 }, { "epoch": 0.8019031051214068, "grad_norm": 1.7978911399841309, "learning_rate": 9.940508214908996e-07, "loss": 0.4534, "step": 19130 }, { "epoch": 0.8019450237364157, "grad_norm": 2.6484029293060303, "learning_rate": 9.936446262310234e-07, "loss": 0.4648, "step": 19131 }, { "epoch": 0.8019869423514248, "grad_norm": 2.236750841140747, "learning_rate": 9.932385048227712e-07, "loss": 0.4577, "step": 19132 }, { "epoch": 0.8020288609664337, "grad_norm": 1.7334098815917969, "learning_rate": 9.92832457273628e-07, "loss": 0.4784, "step": 19133 }, { "epoch": 0.8020707795814427, "grad_norm": 1.552003264427185, "learning_rate": 9.924264835910774e-07, "loss": 0.4633, "step": 19134 }, { "epoch": 0.8021126981964516, "grad_norm": 1.8751877546310425, "learning_rate": 9.920205837826053e-07, "loss": 0.4658, "step": 19135 }, { "epoch": 0.8021546168114605, "grad_norm": 2.6656911373138428, "learning_rate": 9.916147578556918e-07, "loss": 0.4848, "step": 19136 }, { "epoch": 0.8021965354264695, "grad_norm": 2.034294843673706, "learning_rate": 9.912090058178187e-07, "loss": 0.4748, "step": 19137 }, { "epoch": 0.8022384540414784, "grad_norm": 1.6675304174423218, "learning_rate": 9.908033276764672e-07, "loss": 0.4901, "step": 19138 }, { "epoch": 0.8022803726564874, "grad_norm": 1.8934789896011353, "learning_rate": 9.90397723439112e-07, "loss": 0.4753, "step": 19139 }, { "epoch": 0.8023222912714963, "grad_norm": 1.737507700920105, "learning_rate": 9.899921931132322e-07, "loss": 0.4345, "step": 19140 }, { "epoch": 0.8023642098865054, "grad_norm": 1.970853328704834, "learning_rate": 9.89586736706304e-07, "loss": 0.4751, "step": 19141 }, { "epoch": 0.8024061285015143, "grad_norm": 1.7745156288146973, "learning_rate": 9.891813542258e-07, "loss": 0.4873, "step": 19142 }, { "epoch": 0.8024480471165233, "grad_norm": 2.130913019180298, "learning_rate": 9.887760456791912e-07, "loss": 0.502, "step": 19143 }, { "epoch": 0.8024899657315322, "grad_norm": 2.5011069774627686, "learning_rate": 9.883708110739521e-07, "loss": 0.4877, "step": 19144 }, { "epoch": 0.8025318843465412, "grad_norm": 2.097520351409912, "learning_rate": 9.8796565041755e-07, "loss": 0.4446, "step": 19145 }, { "epoch": 0.8025738029615501, "grad_norm": 2.1822566986083984, "learning_rate": 9.875605637174545e-07, "loss": 0.4635, "step": 19146 }, { "epoch": 0.8026157215765591, "grad_norm": 2.0040993690490723, "learning_rate": 9.871555509811347e-07, "loss": 0.5274, "step": 19147 }, { "epoch": 0.802657640191568, "grad_norm": 1.7102265357971191, "learning_rate": 9.867506122160531e-07, "loss": 0.4293, "step": 19148 }, { "epoch": 0.8026995588065771, "grad_norm": 2.0612950325012207, "learning_rate": 9.863457474296761e-07, "loss": 0.4902, "step": 19149 }, { "epoch": 0.802741477421586, "grad_norm": 2.1909096240997314, "learning_rate": 9.859409566294675e-07, "loss": 0.4328, "step": 19150 }, { "epoch": 0.802783396036595, "grad_norm": 1.841975450515747, "learning_rate": 9.855362398228868e-07, "loss": 0.4866, "step": 19151 }, { "epoch": 0.8028253146516039, "grad_norm": 1.8052462339401245, "learning_rate": 9.851315970173968e-07, "loss": 0.4767, "step": 19152 }, { "epoch": 0.8028672332666129, "grad_norm": 1.7879736423492432, "learning_rate": 9.847270282204563e-07, "loss": 0.4681, "step": 19153 }, { "epoch": 0.8029091518816218, "grad_norm": 1.8045759201049805, "learning_rate": 9.8432253343952e-07, "loss": 0.5126, "step": 19154 }, { "epoch": 0.8029510704966308, "grad_norm": 1.8690088987350464, "learning_rate": 9.839181126820462e-07, "loss": 0.5024, "step": 19155 }, { "epoch": 0.8029929891116397, "grad_norm": 2.0880582332611084, "learning_rate": 9.835137659554915e-07, "loss": 0.4412, "step": 19156 }, { "epoch": 0.8030349077266488, "grad_norm": 1.8178898096084595, "learning_rate": 9.831094932673068e-07, "loss": 0.4921, "step": 19157 }, { "epoch": 0.8030768263416577, "grad_norm": 2.6274800300598145, "learning_rate": 9.827052946249454e-07, "loss": 0.4813, "step": 19158 }, { "epoch": 0.8031187449566667, "grad_norm": 1.9242169857025146, "learning_rate": 9.823011700358593e-07, "loss": 0.5701, "step": 19159 }, { "epoch": 0.8031606635716756, "grad_norm": 1.9363411664962769, "learning_rate": 9.818971195074955e-07, "loss": 0.5346, "step": 19160 }, { "epoch": 0.8032025821866845, "grad_norm": 1.7808681726455688, "learning_rate": 9.814931430473045e-07, "loss": 0.4753, "step": 19161 }, { "epoch": 0.8032445008016935, "grad_norm": 1.6525566577911377, "learning_rate": 9.810892406627325e-07, "loss": 0.479, "step": 19162 }, { "epoch": 0.8032864194167024, "grad_norm": 1.827711582183838, "learning_rate": 9.806854123612225e-07, "loss": 0.5374, "step": 19163 }, { "epoch": 0.8033283380317114, "grad_norm": 1.76055109500885, "learning_rate": 9.802816581502212e-07, "loss": 0.4784, "step": 19164 }, { "epoch": 0.8033702566467203, "grad_norm": 1.6644796133041382, "learning_rate": 9.798779780371709e-07, "loss": 0.4787, "step": 19165 }, { "epoch": 0.8034121752617294, "grad_norm": 2.154517412185669, "learning_rate": 9.794743720295114e-07, "loss": 0.4904, "step": 19166 }, { "epoch": 0.8034540938767383, "grad_norm": 2.0641579627990723, "learning_rate": 9.790708401346837e-07, "loss": 0.4549, "step": 19167 }, { "epoch": 0.8034960124917473, "grad_norm": 1.9032663106918335, "learning_rate": 9.786673823601278e-07, "loss": 0.4678, "step": 19168 }, { "epoch": 0.8035379311067562, "grad_norm": 2.05590558052063, "learning_rate": 9.782639987132774e-07, "loss": 0.4402, "step": 19169 }, { "epoch": 0.8035798497217652, "grad_norm": 1.932997465133667, "learning_rate": 9.778606892015725e-07, "loss": 0.5627, "step": 19170 }, { "epoch": 0.8036217683367741, "grad_norm": 2.0899460315704346, "learning_rate": 9.77457453832445e-07, "loss": 0.4523, "step": 19171 }, { "epoch": 0.8036636869517831, "grad_norm": 2.926795482635498, "learning_rate": 9.77054292613327e-07, "loss": 0.488, "step": 19172 }, { "epoch": 0.803705605566792, "grad_norm": 1.7865487337112427, "learning_rate": 9.766512055516514e-07, "loss": 0.5239, "step": 19173 }, { "epoch": 0.8037475241818011, "grad_norm": 1.9971342086791992, "learning_rate": 9.7624819265485e-07, "loss": 0.4407, "step": 19174 }, { "epoch": 0.80378944279681, "grad_norm": 1.579485535621643, "learning_rate": 9.758452539303492e-07, "loss": 0.5114, "step": 19175 }, { "epoch": 0.803831361411819, "grad_norm": 1.8358099460601807, "learning_rate": 9.754423893855797e-07, "loss": 0.4896, "step": 19176 }, { "epoch": 0.8038732800268279, "grad_norm": 1.901849389076233, "learning_rate": 9.750395990279643e-07, "loss": 0.5241, "step": 19177 }, { "epoch": 0.8039151986418369, "grad_norm": 1.6493972539901733, "learning_rate": 9.746368828649294e-07, "loss": 0.4919, "step": 19178 }, { "epoch": 0.8039571172568458, "grad_norm": 1.970658779144287, "learning_rate": 9.742342409039002e-07, "loss": 0.4929, "step": 19179 }, { "epoch": 0.8039990358718548, "grad_norm": 2.2078909873962402, "learning_rate": 9.738316731522973e-07, "loss": 0.4638, "step": 19180 }, { "epoch": 0.8040409544868637, "grad_norm": 1.8857141733169556, "learning_rate": 9.734291796175399e-07, "loss": 0.4983, "step": 19181 }, { "epoch": 0.8040828731018728, "grad_norm": 1.5375796556472778, "learning_rate": 9.730267603070508e-07, "loss": 0.4475, "step": 19182 }, { "epoch": 0.8041247917168817, "grad_norm": 1.9268419742584229, "learning_rate": 9.726244152282448e-07, "loss": 0.4641, "step": 19183 }, { "epoch": 0.8041667103318907, "grad_norm": 2.1289682388305664, "learning_rate": 9.7222214438854e-07, "loss": 0.4614, "step": 19184 }, { "epoch": 0.8042086289468996, "grad_norm": 1.5662941932678223, "learning_rate": 9.718199477953527e-07, "loss": 0.4981, "step": 19185 }, { "epoch": 0.8042505475619085, "grad_norm": 1.7070173025131226, "learning_rate": 9.714178254560947e-07, "loss": 0.4432, "step": 19186 }, { "epoch": 0.8042924661769175, "grad_norm": 1.695554256439209, "learning_rate": 9.710157773781804e-07, "loss": 0.4739, "step": 19187 }, { "epoch": 0.8043343847919264, "grad_norm": 1.7532099485397339, "learning_rate": 9.70613803569021e-07, "loss": 0.4968, "step": 19188 }, { "epoch": 0.8043763034069354, "grad_norm": 2.3851590156555176, "learning_rate": 9.702119040360241e-07, "loss": 0.491, "step": 19189 }, { "epoch": 0.8044182220219444, "grad_norm": 2.7870593070983887, "learning_rate": 9.698100787866015e-07, "loss": 0.4543, "step": 19190 }, { "epoch": 0.8044601406369534, "grad_norm": 1.8458905220031738, "learning_rate": 9.694083278281585e-07, "loss": 0.4392, "step": 19191 }, { "epoch": 0.8045020592519623, "grad_norm": 1.9782729148864746, "learning_rate": 9.690066511680995e-07, "loss": 0.4634, "step": 19192 }, { "epoch": 0.8045439778669713, "grad_norm": 1.8160911798477173, "learning_rate": 9.686050488138303e-07, "loss": 0.4549, "step": 19193 }, { "epoch": 0.8045858964819802, "grad_norm": 2.090630531311035, "learning_rate": 9.682035207727547e-07, "loss": 0.4037, "step": 19194 }, { "epoch": 0.8046278150969892, "grad_norm": 1.9854707717895508, "learning_rate": 9.678020670522726e-07, "loss": 0.4571, "step": 19195 }, { "epoch": 0.8046697337119981, "grad_norm": 2.5346360206604004, "learning_rate": 9.674006876597848e-07, "loss": 0.505, "step": 19196 }, { "epoch": 0.8047116523270071, "grad_norm": 1.994789481163025, "learning_rate": 9.669993826026918e-07, "loss": 0.4604, "step": 19197 }, { "epoch": 0.804753570942016, "grad_norm": 1.9701956510543823, "learning_rate": 9.665981518883882e-07, "loss": 0.5101, "step": 19198 }, { "epoch": 0.8047954895570251, "grad_norm": 1.7004485130310059, "learning_rate": 9.661969955242733e-07, "loss": 0.4387, "step": 19199 }, { "epoch": 0.804837408172034, "grad_norm": 2.0401320457458496, "learning_rate": 9.657959135177402e-07, "loss": 0.4626, "step": 19200 }, { "epoch": 0.804879326787043, "grad_norm": 1.7569706439971924, "learning_rate": 9.653949058761808e-07, "loss": 0.4444, "step": 19201 }, { "epoch": 0.8049212454020519, "grad_norm": 1.772599220275879, "learning_rate": 9.64993972606989e-07, "loss": 0.483, "step": 19202 }, { "epoch": 0.8049631640170609, "grad_norm": 1.7398704290390015, "learning_rate": 9.645931137175556e-07, "loss": 0.5418, "step": 19203 }, { "epoch": 0.8050050826320698, "grad_norm": 2.294606924057007, "learning_rate": 9.641923292152688e-07, "loss": 0.5262, "step": 19204 }, { "epoch": 0.8050470012470788, "grad_norm": 1.8613615036010742, "learning_rate": 9.63791619107517e-07, "loss": 0.5119, "step": 19205 }, { "epoch": 0.8050889198620877, "grad_norm": 1.7674959897994995, "learning_rate": 9.63390983401688e-07, "loss": 0.4575, "step": 19206 }, { "epoch": 0.8051308384770968, "grad_norm": 1.6582167148590088, "learning_rate": 9.629904221051646e-07, "loss": 0.5051, "step": 19207 }, { "epoch": 0.8051727570921057, "grad_norm": 1.8915330171585083, "learning_rate": 9.625899352253327e-07, "loss": 0.507, "step": 19208 }, { "epoch": 0.8052146757071147, "grad_norm": 2.3540306091308594, "learning_rate": 9.621895227695739e-07, "loss": 0.451, "step": 19209 }, { "epoch": 0.8052565943221236, "grad_norm": 4.435093402862549, "learning_rate": 9.61789184745268e-07, "loss": 0.5, "step": 19210 }, { "epoch": 0.8052985129371325, "grad_norm": 3.2969324588775635, "learning_rate": 9.613889211597955e-07, "loss": 0.5177, "step": 19211 }, { "epoch": 0.8053404315521415, "grad_norm": 1.8734797239303589, "learning_rate": 9.60988732020537e-07, "loss": 0.4172, "step": 19212 }, { "epoch": 0.8053823501671504, "grad_norm": 1.7236706018447876, "learning_rate": 9.605886173348654e-07, "loss": 0.4583, "step": 19213 }, { "epoch": 0.8054242687821594, "grad_norm": 2.7786166667938232, "learning_rate": 9.601885771101589e-07, "loss": 0.4936, "step": 19214 }, { "epoch": 0.8054661873971684, "grad_norm": 1.7494233846664429, "learning_rate": 9.597886113537919e-07, "loss": 0.4374, "step": 19215 }, { "epoch": 0.8055081060121774, "grad_norm": 2.3551487922668457, "learning_rate": 9.593887200731356e-07, "loss": 0.4468, "step": 19216 }, { "epoch": 0.8055500246271863, "grad_norm": 2.5659096240997314, "learning_rate": 9.589889032755635e-07, "loss": 0.4953, "step": 19217 }, { "epoch": 0.8055919432421953, "grad_norm": 1.8245707750320435, "learning_rate": 9.58589160968444e-07, "loss": 0.4666, "step": 19218 }, { "epoch": 0.8056338618572042, "grad_norm": 5.833681106567383, "learning_rate": 9.58189493159145e-07, "loss": 0.4651, "step": 19219 }, { "epoch": 0.8056757804722132, "grad_norm": 2.409064769744873, "learning_rate": 9.577898998550356e-07, "loss": 0.5027, "step": 19220 }, { "epoch": 0.8057176990872221, "grad_norm": 9.631321907043457, "learning_rate": 9.573903810634822e-07, "loss": 0.5106, "step": 19221 }, { "epoch": 0.8057596177022311, "grad_norm": 1.92827570438385, "learning_rate": 9.569909367918473e-07, "loss": 0.4904, "step": 19222 }, { "epoch": 0.80580153631724, "grad_norm": 1.7047176361083984, "learning_rate": 9.565915670474951e-07, "loss": 0.4911, "step": 19223 }, { "epoch": 0.8058434549322491, "grad_norm": 2.0286905765533447, "learning_rate": 9.561922718377887e-07, "loss": 0.5095, "step": 19224 }, { "epoch": 0.805885373547258, "grad_norm": 1.590463399887085, "learning_rate": 9.557930511700858e-07, "loss": 0.4896, "step": 19225 }, { "epoch": 0.805927292162267, "grad_norm": 2.0308005809783936, "learning_rate": 9.553939050517486e-07, "loss": 0.4954, "step": 19226 }, { "epoch": 0.8059692107772759, "grad_norm": 2.0182008743286133, "learning_rate": 9.549948334901332e-07, "loss": 0.4667, "step": 19227 }, { "epoch": 0.8060111293922849, "grad_norm": 2.0428969860076904, "learning_rate": 9.54595836492595e-07, "loss": 0.4899, "step": 19228 }, { "epoch": 0.8060530480072938, "grad_norm": 1.8059570789337158, "learning_rate": 9.541969140664897e-07, "loss": 0.4495, "step": 19229 }, { "epoch": 0.8060949666223028, "grad_norm": 2.6172244548797607, "learning_rate": 9.537980662191726e-07, "loss": 0.4354, "step": 19230 }, { "epoch": 0.8061368852373118, "grad_norm": 1.6587822437286377, "learning_rate": 9.533992929579927e-07, "loss": 0.5021, "step": 19231 }, { "epoch": 0.8061788038523208, "grad_norm": 2.2599408626556396, "learning_rate": 9.530005942903031e-07, "loss": 0.4855, "step": 19232 }, { "epoch": 0.8062207224673297, "grad_norm": 2.90374493598938, "learning_rate": 9.526019702234535e-07, "loss": 0.4874, "step": 19233 }, { "epoch": 0.8062626410823387, "grad_norm": 1.6522263288497925, "learning_rate": 9.522034207647907e-07, "loss": 0.4684, "step": 19234 }, { "epoch": 0.8063045596973476, "grad_norm": 1.9515854120254517, "learning_rate": 9.518049459216628e-07, "loss": 0.5296, "step": 19235 }, { "epoch": 0.8063464783123565, "grad_norm": 1.607835292816162, "learning_rate": 9.514065457014127e-07, "loss": 0.5051, "step": 19236 }, { "epoch": 0.8063883969273655, "grad_norm": 2.3006227016448975, "learning_rate": 9.510082201113874e-07, "loss": 0.5183, "step": 19237 }, { "epoch": 0.8064303155423744, "grad_norm": 2.0440456867218018, "learning_rate": 9.506099691589266e-07, "loss": 0.4529, "step": 19238 }, { "epoch": 0.8064722341573834, "grad_norm": 1.9827146530151367, "learning_rate": 9.502117928513743e-07, "loss": 0.5146, "step": 19239 }, { "epoch": 0.8065141527723924, "grad_norm": 1.8858649730682373, "learning_rate": 9.498136911960676e-07, "loss": 0.448, "step": 19240 }, { "epoch": 0.8065560713874014, "grad_norm": 1.9128788709640503, "learning_rate": 9.494156642003466e-07, "loss": 0.5043, "step": 19241 }, { "epoch": 0.8065979900024103, "grad_norm": 2.25046706199646, "learning_rate": 9.490177118715494e-07, "loss": 0.507, "step": 19242 }, { "epoch": 0.8066399086174193, "grad_norm": 1.9008269309997559, "learning_rate": 9.486198342170083e-07, "loss": 0.4574, "step": 19243 }, { "epoch": 0.8066818272324282, "grad_norm": 1.7959861755371094, "learning_rate": 9.482220312440604e-07, "loss": 0.4798, "step": 19244 }, { "epoch": 0.8067237458474372, "grad_norm": 1.7729729413986206, "learning_rate": 9.478243029600392e-07, "loss": 0.4276, "step": 19245 }, { "epoch": 0.8067656644624461, "grad_norm": 2.0160725116729736, "learning_rate": 9.474266493722745e-07, "loss": 0.3929, "step": 19246 }, { "epoch": 0.8068075830774551, "grad_norm": 1.869387149810791, "learning_rate": 9.470290704880964e-07, "loss": 0.4161, "step": 19247 }, { "epoch": 0.806849501692464, "grad_norm": 2.8212034702301025, "learning_rate": 9.466315663148351e-07, "loss": 0.4556, "step": 19248 }, { "epoch": 0.8068914203074731, "grad_norm": 2.5588293075561523, "learning_rate": 9.46234136859816e-07, "loss": 0.4403, "step": 19249 }, { "epoch": 0.806933338922482, "grad_norm": 2.1305336952209473, "learning_rate": 9.458367821303682e-07, "loss": 0.4737, "step": 19250 }, { "epoch": 0.806975257537491, "grad_norm": 1.9254812002182007, "learning_rate": 9.454395021338125e-07, "loss": 0.4687, "step": 19251 }, { "epoch": 0.8070171761524999, "grad_norm": 1.810335636138916, "learning_rate": 9.450422968774753e-07, "loss": 0.4523, "step": 19252 }, { "epoch": 0.8070590947675089, "grad_norm": 1.940767526626587, "learning_rate": 9.446451663686778e-07, "loss": 0.4502, "step": 19253 }, { "epoch": 0.8071010133825178, "grad_norm": 1.996809959411621, "learning_rate": 9.4424811061474e-07, "loss": 0.5472, "step": 19254 }, { "epoch": 0.8071429319975268, "grad_norm": 1.9733079671859741, "learning_rate": 9.438511296229819e-07, "loss": 0.4726, "step": 19255 }, { "epoch": 0.8071848506125358, "grad_norm": 1.7327996492385864, "learning_rate": 9.43454223400721e-07, "loss": 0.4983, "step": 19256 }, { "epoch": 0.8072267692275448, "grad_norm": 1.9958555698394775, "learning_rate": 9.430573919552722e-07, "loss": 0.3882, "step": 19257 }, { "epoch": 0.8072686878425537, "grad_norm": 1.7888457775115967, "learning_rate": 9.426606352939521e-07, "loss": 0.509, "step": 19258 }, { "epoch": 0.8073106064575627, "grad_norm": 1.9204976558685303, "learning_rate": 9.422639534240752e-07, "loss": 0.508, "step": 19259 }, { "epoch": 0.8073525250725716, "grad_norm": 1.6623187065124512, "learning_rate": 9.418673463529515e-07, "loss": 0.468, "step": 19260 }, { "epoch": 0.8073944436875805, "grad_norm": 1.7994223833084106, "learning_rate": 9.414708140878925e-07, "loss": 0.512, "step": 19261 }, { "epoch": 0.8074363623025895, "grad_norm": 2.9616024494171143, "learning_rate": 9.410743566362102e-07, "loss": 0.4549, "step": 19262 }, { "epoch": 0.8074782809175984, "grad_norm": 2.0836474895477295, "learning_rate": 9.406779740052091e-07, "loss": 0.4882, "step": 19263 }, { "epoch": 0.8075201995326075, "grad_norm": 1.7796669006347656, "learning_rate": 9.402816662021996e-07, "loss": 0.4408, "step": 19264 }, { "epoch": 0.8075621181476164, "grad_norm": 2.2453081607818604, "learning_rate": 9.398854332344848e-07, "loss": 0.5463, "step": 19265 }, { "epoch": 0.8076040367626254, "grad_norm": 1.9694585800170898, "learning_rate": 9.394892751093676e-07, "loss": 0.471, "step": 19266 }, { "epoch": 0.8076459553776343, "grad_norm": 1.8527379035949707, "learning_rate": 9.390931918341522e-07, "loss": 0.4628, "step": 19267 }, { "epoch": 0.8076878739926433, "grad_norm": 1.8815417289733887, "learning_rate": 9.386971834161412e-07, "loss": 0.4902, "step": 19268 }, { "epoch": 0.8077297926076522, "grad_norm": 2.6216816902160645, "learning_rate": 9.383012498626321e-07, "loss": 0.4547, "step": 19269 }, { "epoch": 0.8077717112226612, "grad_norm": 1.6918723583221436, "learning_rate": 9.379053911809238e-07, "loss": 0.4782, "step": 19270 }, { "epoch": 0.8078136298376701, "grad_norm": 2.0156781673431396, "learning_rate": 9.375096073783152e-07, "loss": 0.5243, "step": 19271 }, { "epoch": 0.8078555484526792, "grad_norm": 1.9231971502304077, "learning_rate": 9.371138984620998e-07, "loss": 0.491, "step": 19272 }, { "epoch": 0.8078974670676881, "grad_norm": 1.7595478296279907, "learning_rate": 9.367182644395745e-07, "loss": 0.508, "step": 19273 }, { "epoch": 0.8079393856826971, "grad_norm": 2.038947820663452, "learning_rate": 9.363227053180302e-07, "loss": 0.443, "step": 19274 }, { "epoch": 0.807981304297706, "grad_norm": 2.085663080215454, "learning_rate": 9.359272211047576e-07, "loss": 0.4788, "step": 19275 }, { "epoch": 0.808023222912715, "grad_norm": 1.5036460161209106, "learning_rate": 9.355318118070483e-07, "loss": 0.4456, "step": 19276 }, { "epoch": 0.8080651415277239, "grad_norm": 2.0177974700927734, "learning_rate": 9.351364774321924e-07, "loss": 0.4793, "step": 19277 }, { "epoch": 0.8081070601427329, "grad_norm": 1.5984903573989868, "learning_rate": 9.347412179874749e-07, "loss": 0.4275, "step": 19278 }, { "epoch": 0.8081489787577418, "grad_norm": 2.192631721496582, "learning_rate": 9.34346033480183e-07, "loss": 0.4659, "step": 19279 }, { "epoch": 0.8081908973727508, "grad_norm": 2.2787654399871826, "learning_rate": 9.339509239176026e-07, "loss": 0.4806, "step": 19280 }, { "epoch": 0.8082328159877598, "grad_norm": 1.78542959690094, "learning_rate": 9.335558893070145e-07, "loss": 0.4312, "step": 19281 }, { "epoch": 0.8082747346027688, "grad_norm": 2.029719591140747, "learning_rate": 9.331609296557032e-07, "loss": 0.4707, "step": 19282 }, { "epoch": 0.8083166532177777, "grad_norm": 1.8522770404815674, "learning_rate": 9.327660449709464e-07, "loss": 0.4269, "step": 19283 }, { "epoch": 0.8083585718327867, "grad_norm": 1.7296371459960938, "learning_rate": 9.323712352600261e-07, "loss": 0.4955, "step": 19284 }, { "epoch": 0.8084004904477956, "grad_norm": 1.6939982175827026, "learning_rate": 9.319765005302179e-07, "loss": 0.4367, "step": 19285 }, { "epoch": 0.8084424090628045, "grad_norm": 1.826377272605896, "learning_rate": 9.315818407887999e-07, "loss": 0.4112, "step": 19286 }, { "epoch": 0.8084843276778135, "grad_norm": 1.7328277826309204, "learning_rate": 9.311872560430452e-07, "loss": 0.4591, "step": 19287 }, { "epoch": 0.8085262462928224, "grad_norm": 2.087495803833008, "learning_rate": 9.307927463002286e-07, "loss": 0.4145, "step": 19288 }, { "epoch": 0.8085681649078315, "grad_norm": 1.9325891733169556, "learning_rate": 9.30398311567623e-07, "loss": 0.4351, "step": 19289 }, { "epoch": 0.8086100835228404, "grad_norm": 2.0023114681243896, "learning_rate": 9.300039518524978e-07, "loss": 0.4704, "step": 19290 }, { "epoch": 0.8086520021378494, "grad_norm": 2.0141685009002686, "learning_rate": 9.296096671621224e-07, "loss": 0.5119, "step": 19291 }, { "epoch": 0.8086939207528583, "grad_norm": 2.713022232055664, "learning_rate": 9.292154575037676e-07, "loss": 0.4757, "step": 19292 }, { "epoch": 0.8087358393678673, "grad_norm": 1.9557698965072632, "learning_rate": 9.288213228846976e-07, "loss": 0.4805, "step": 19293 }, { "epoch": 0.8087777579828762, "grad_norm": 1.8144536018371582, "learning_rate": 9.284272633121777e-07, "loss": 0.4482, "step": 19294 }, { "epoch": 0.8088196765978852, "grad_norm": 1.7709790468215942, "learning_rate": 9.280332787934731e-07, "loss": 0.4884, "step": 19295 }, { "epoch": 0.8088615952128941, "grad_norm": 2.4552628993988037, "learning_rate": 9.276393693358443e-07, "loss": 0.4759, "step": 19296 }, { "epoch": 0.8089035138279032, "grad_norm": 2.1402459144592285, "learning_rate": 9.272455349465542e-07, "loss": 0.4822, "step": 19297 }, { "epoch": 0.8089454324429121, "grad_norm": 1.8854070901870728, "learning_rate": 9.268517756328638e-07, "loss": 0.4875, "step": 19298 }, { "epoch": 0.8089873510579211, "grad_norm": 2.8638126850128174, "learning_rate": 9.264580914020283e-07, "loss": 0.4455, "step": 19299 }, { "epoch": 0.80902926967293, "grad_norm": 1.8117921352386475, "learning_rate": 9.260644822613063e-07, "loss": 0.477, "step": 19300 }, { "epoch": 0.809071188287939, "grad_norm": 2.000553846359253, "learning_rate": 9.256709482179549e-07, "loss": 0.4779, "step": 19301 }, { "epoch": 0.8091131069029479, "grad_norm": 2.160956621170044, "learning_rate": 9.252774892792271e-07, "loss": 0.4968, "step": 19302 }, { "epoch": 0.8091550255179569, "grad_norm": 1.6106387376785278, "learning_rate": 9.248841054523744e-07, "loss": 0.4425, "step": 19303 }, { "epoch": 0.8091969441329658, "grad_norm": 2.303847551345825, "learning_rate": 9.244907967446503e-07, "loss": 0.4755, "step": 19304 }, { "epoch": 0.8092388627479749, "grad_norm": 1.8633419275283813, "learning_rate": 9.240975631633036e-07, "loss": 0.434, "step": 19305 }, { "epoch": 0.8092807813629838, "grad_norm": 2.0985476970672607, "learning_rate": 9.23704404715583e-07, "loss": 0.3976, "step": 19306 }, { "epoch": 0.8093226999779928, "grad_norm": 1.6447217464447021, "learning_rate": 9.23311321408738e-07, "loss": 0.4472, "step": 19307 }, { "epoch": 0.8093646185930017, "grad_norm": 2.4672183990478516, "learning_rate": 9.229183132500113e-07, "loss": 0.4811, "step": 19308 }, { "epoch": 0.8094065372080107, "grad_norm": 1.9020105600357056, "learning_rate": 9.225253802466494e-07, "loss": 0.4733, "step": 19309 }, { "epoch": 0.8094484558230196, "grad_norm": 1.8041714429855347, "learning_rate": 9.221325224058963e-07, "loss": 0.4628, "step": 19310 }, { "epoch": 0.8094903744380285, "grad_norm": 1.7777624130249023, "learning_rate": 9.217397397349931e-07, "loss": 0.4718, "step": 19311 }, { "epoch": 0.8095322930530375, "grad_norm": 1.8931164741516113, "learning_rate": 9.213470322411782e-07, "loss": 0.4439, "step": 19312 }, { "epoch": 0.8095742116680464, "grad_norm": 1.679611325263977, "learning_rate": 9.209543999316934e-07, "loss": 0.47, "step": 19313 }, { "epoch": 0.8096161302830555, "grad_norm": 2.013868808746338, "learning_rate": 9.205618428137736e-07, "loss": 0.465, "step": 19314 }, { "epoch": 0.8096580488980644, "grad_norm": 2.180830955505371, "learning_rate": 9.201693608946572e-07, "loss": 0.4745, "step": 19315 }, { "epoch": 0.8096999675130734, "grad_norm": 1.8085802793502808, "learning_rate": 9.197769541815793e-07, "loss": 0.4924, "step": 19316 }, { "epoch": 0.8097418861280823, "grad_norm": 2.3716657161712646, "learning_rate": 9.193846226817709e-07, "loss": 0.523, "step": 19317 }, { "epoch": 0.8097838047430913, "grad_norm": 1.781914472579956, "learning_rate": 9.189923664024664e-07, "loss": 0.4418, "step": 19318 }, { "epoch": 0.8098257233581002, "grad_norm": 3.027405261993408, "learning_rate": 9.186001853508969e-07, "loss": 0.4676, "step": 19319 }, { "epoch": 0.8098676419731092, "grad_norm": 2.6591148376464844, "learning_rate": 9.182080795342901e-07, "loss": 0.485, "step": 19320 }, { "epoch": 0.8099095605881181, "grad_norm": 1.6884069442749023, "learning_rate": 9.178160489598736e-07, "loss": 0.471, "step": 19321 }, { "epoch": 0.8099514792031272, "grad_norm": 2.2460267543792725, "learning_rate": 9.17424093634876e-07, "loss": 0.5197, "step": 19322 }, { "epoch": 0.8099933978181361, "grad_norm": 2.0142300128936768, "learning_rate": 9.170322135665199e-07, "loss": 0.4471, "step": 19323 }, { "epoch": 0.8100353164331451, "grad_norm": 1.8616840839385986, "learning_rate": 9.166404087620306e-07, "loss": 0.5072, "step": 19324 }, { "epoch": 0.810077235048154, "grad_norm": 3.734241485595703, "learning_rate": 9.162486792286319e-07, "loss": 0.4538, "step": 19325 }, { "epoch": 0.810119153663163, "grad_norm": 1.8978127241134644, "learning_rate": 9.15857024973541e-07, "loss": 0.4709, "step": 19326 }, { "epoch": 0.8101610722781719, "grad_norm": 2.053097724914551, "learning_rate": 9.154654460039813e-07, "loss": 0.4895, "step": 19327 }, { "epoch": 0.8102029908931809, "grad_norm": 2.471240520477295, "learning_rate": 9.150739423271688e-07, "loss": 0.551, "step": 19328 }, { "epoch": 0.8102449095081898, "grad_norm": 1.8486636877059937, "learning_rate": 9.146825139503213e-07, "loss": 0.4997, "step": 19329 }, { "epoch": 0.8102868281231989, "grad_norm": 2.834493398666382, "learning_rate": 9.142911608806532e-07, "loss": 0.4848, "step": 19330 }, { "epoch": 0.8103287467382078, "grad_norm": 2.290773868560791, "learning_rate": 9.138998831253798e-07, "loss": 0.4702, "step": 19331 }, { "epoch": 0.8103706653532168, "grad_norm": 2.215867757797241, "learning_rate": 9.135086806917126e-07, "loss": 0.5141, "step": 19332 }, { "epoch": 0.8104125839682257, "grad_norm": 1.9953181743621826, "learning_rate": 9.131175535868647e-07, "loss": 0.454, "step": 19333 }, { "epoch": 0.8104545025832347, "grad_norm": 2.7782487869262695, "learning_rate": 9.127265018180431e-07, "loss": 0.4921, "step": 19334 }, { "epoch": 0.8104964211982436, "grad_norm": 1.8151084184646606, "learning_rate": 9.123355253924582e-07, "loss": 0.4613, "step": 19335 }, { "epoch": 0.8105383398132525, "grad_norm": 1.6022882461547852, "learning_rate": 9.11944624317318e-07, "loss": 0.4415, "step": 19336 }, { "epoch": 0.8105802584282615, "grad_norm": 1.8920543193817139, "learning_rate": 9.115537985998257e-07, "loss": 0.4959, "step": 19337 }, { "epoch": 0.8106221770432704, "grad_norm": 2.29636549949646, "learning_rate": 9.111630482471872e-07, "loss": 0.4812, "step": 19338 }, { "epoch": 0.8106640956582795, "grad_norm": 1.8568675518035889, "learning_rate": 9.107723732666069e-07, "loss": 0.4884, "step": 19339 }, { "epoch": 0.8107060142732884, "grad_norm": 1.6500325202941895, "learning_rate": 9.103817736652842e-07, "loss": 0.4627, "step": 19340 }, { "epoch": 0.8107479328882974, "grad_norm": 1.9918354749679565, "learning_rate": 9.099912494504188e-07, "loss": 0.4762, "step": 19341 }, { "epoch": 0.8107898515033063, "grad_norm": 1.8594354391098022, "learning_rate": 9.096008006292117e-07, "loss": 0.4812, "step": 19342 }, { "epoch": 0.8108317701183153, "grad_norm": 3.232527732849121, "learning_rate": 9.092104272088575e-07, "loss": 0.442, "step": 19343 }, { "epoch": 0.8108736887333242, "grad_norm": 1.7359784841537476, "learning_rate": 9.088201291965537e-07, "loss": 0.474, "step": 19344 }, { "epoch": 0.8109156073483332, "grad_norm": 1.7983746528625488, "learning_rate": 9.084299065994967e-07, "loss": 0.4435, "step": 19345 }, { "epoch": 0.8109575259633421, "grad_norm": 2.17960524559021, "learning_rate": 9.080397594248763e-07, "loss": 0.5343, "step": 19346 }, { "epoch": 0.8109994445783512, "grad_norm": 1.7713725566864014, "learning_rate": 9.076496876798857e-07, "loss": 0.4931, "step": 19347 }, { "epoch": 0.8110413631933601, "grad_norm": 1.7734382152557373, "learning_rate": 9.072596913717174e-07, "loss": 0.4841, "step": 19348 }, { "epoch": 0.8110832818083691, "grad_norm": 1.868538498878479, "learning_rate": 9.068697705075585e-07, "loss": 0.4676, "step": 19349 }, { "epoch": 0.811125200423378, "grad_norm": 1.8415333032608032, "learning_rate": 9.064799250945955e-07, "loss": 0.4767, "step": 19350 }, { "epoch": 0.811167119038387, "grad_norm": 1.7979679107666016, "learning_rate": 9.060901551400169e-07, "loss": 0.5121, "step": 19351 }, { "epoch": 0.8112090376533959, "grad_norm": 1.9137049913406372, "learning_rate": 9.057004606510056e-07, "loss": 0.4959, "step": 19352 }, { "epoch": 0.8112509562684049, "grad_norm": 1.7040231227874756, "learning_rate": 9.053108416347461e-07, "loss": 0.4729, "step": 19353 }, { "epoch": 0.8112928748834138, "grad_norm": 1.8785645961761475, "learning_rate": 9.049212980984223e-07, "loss": 0.459, "step": 19354 }, { "epoch": 0.8113347934984229, "grad_norm": 1.7923345565795898, "learning_rate": 9.045318300492112e-07, "loss": 0.459, "step": 19355 }, { "epoch": 0.8113767121134318, "grad_norm": 1.9199107885360718, "learning_rate": 9.041424374942942e-07, "loss": 0.498, "step": 19356 }, { "epoch": 0.8114186307284408, "grad_norm": 1.8247555494308472, "learning_rate": 9.037531204408501e-07, "loss": 0.4704, "step": 19357 }, { "epoch": 0.8114605493434497, "grad_norm": 2.6387779712677, "learning_rate": 9.033638788960547e-07, "loss": 0.5208, "step": 19358 }, { "epoch": 0.8115024679584587, "grad_norm": 1.728724718093872, "learning_rate": 9.029747128670813e-07, "loss": 0.4334, "step": 19359 }, { "epoch": 0.8115443865734676, "grad_norm": 2.233818292617798, "learning_rate": 9.025856223611062e-07, "loss": 0.5022, "step": 19360 }, { "epoch": 0.8115863051884765, "grad_norm": 2.049882173538208, "learning_rate": 9.021966073852995e-07, "loss": 0.4559, "step": 19361 }, { "epoch": 0.8116282238034855, "grad_norm": 1.6896408796310425, "learning_rate": 9.01807667946833e-07, "loss": 0.4474, "step": 19362 }, { "epoch": 0.8116701424184944, "grad_norm": 1.6896016597747803, "learning_rate": 9.014188040528781e-07, "loss": 0.4321, "step": 19363 }, { "epoch": 0.8117120610335035, "grad_norm": 1.7576855421066284, "learning_rate": 9.010300157106006e-07, "loss": 0.5193, "step": 19364 }, { "epoch": 0.8117539796485124, "grad_norm": 1.8826210498809814, "learning_rate": 9.006413029271677e-07, "loss": 0.507, "step": 19365 }, { "epoch": 0.8117958982635214, "grad_norm": 1.700250267982483, "learning_rate": 9.00252665709746e-07, "loss": 0.4483, "step": 19366 }, { "epoch": 0.8118378168785303, "grad_norm": 1.8689959049224854, "learning_rate": 8.998641040654993e-07, "loss": 0.4559, "step": 19367 }, { "epoch": 0.8118797354935393, "grad_norm": 1.6410423517227173, "learning_rate": 8.994756180015879e-07, "loss": 0.4394, "step": 19368 }, { "epoch": 0.8119216541085482, "grad_norm": 1.7798751592636108, "learning_rate": 8.990872075251766e-07, "loss": 0.5125, "step": 19369 }, { "epoch": 0.8119635727235572, "grad_norm": 1.8711236715316772, "learning_rate": 8.986988726434215e-07, "loss": 0.4807, "step": 19370 }, { "epoch": 0.8120054913385661, "grad_norm": 2.0525059700012207, "learning_rate": 8.983106133634828e-07, "loss": 0.4828, "step": 19371 }, { "epoch": 0.8120474099535752, "grad_norm": 1.8130035400390625, "learning_rate": 8.979224296925187e-07, "loss": 0.4856, "step": 19372 }, { "epoch": 0.8120893285685841, "grad_norm": 2.1039414405822754, "learning_rate": 8.975343216376831e-07, "loss": 0.4471, "step": 19373 }, { "epoch": 0.8121312471835931, "grad_norm": 2.040579080581665, "learning_rate": 8.9714628920613e-07, "loss": 0.5218, "step": 19374 }, { "epoch": 0.812173165798602, "grad_norm": 1.8324952125549316, "learning_rate": 8.967583324050144e-07, "loss": 0.4875, "step": 19375 }, { "epoch": 0.812215084413611, "grad_norm": 2.0315442085266113, "learning_rate": 8.963704512414867e-07, "loss": 0.4951, "step": 19376 }, { "epoch": 0.8122570030286199, "grad_norm": 1.6114323139190674, "learning_rate": 8.95982645722695e-07, "loss": 0.4695, "step": 19377 }, { "epoch": 0.8122989216436289, "grad_norm": 1.68329656124115, "learning_rate": 8.955949158557914e-07, "loss": 0.4624, "step": 19378 }, { "epoch": 0.8123408402586378, "grad_norm": 1.9193980693817139, "learning_rate": 8.952072616479196e-07, "loss": 0.4781, "step": 19379 }, { "epoch": 0.8123827588736469, "grad_norm": 1.913495421409607, "learning_rate": 8.948196831062278e-07, "loss": 0.4423, "step": 19380 }, { "epoch": 0.8124246774886558, "grad_norm": 1.6846669912338257, "learning_rate": 8.944321802378603e-07, "loss": 0.475, "step": 19381 }, { "epoch": 0.8124665961036648, "grad_norm": 1.8548121452331543, "learning_rate": 8.940447530499591e-07, "loss": 0.431, "step": 19382 }, { "epoch": 0.8125085147186737, "grad_norm": 1.8255749940872192, "learning_rate": 8.936574015496663e-07, "loss": 0.4669, "step": 19383 }, { "epoch": 0.8125504333336827, "grad_norm": 1.5809162855148315, "learning_rate": 8.93270125744124e-07, "loss": 0.4634, "step": 19384 }, { "epoch": 0.8125923519486916, "grad_norm": 1.7402433156967163, "learning_rate": 8.92882925640468e-07, "loss": 0.4716, "step": 19385 }, { "epoch": 0.8126342705637005, "grad_norm": 1.7010815143585205, "learning_rate": 8.92495801245839e-07, "loss": 0.4851, "step": 19386 }, { "epoch": 0.8126761891787095, "grad_norm": 1.9460936784744263, "learning_rate": 8.92108752567371e-07, "loss": 0.5577, "step": 19387 }, { "epoch": 0.8127181077937184, "grad_norm": 1.948914647102356, "learning_rate": 8.917217796121986e-07, "loss": 0.4113, "step": 19388 }, { "epoch": 0.8127600264087275, "grad_norm": 1.7653915882110596, "learning_rate": 8.913348823874552e-07, "loss": 0.4253, "step": 19389 }, { "epoch": 0.8128019450237364, "grad_norm": 1.5811536312103271, "learning_rate": 8.909480609002741e-07, "loss": 0.5114, "step": 19390 }, { "epoch": 0.8128438636387454, "grad_norm": 2.0003151893615723, "learning_rate": 8.905613151577841e-07, "loss": 0.4445, "step": 19391 }, { "epoch": 0.8128857822537543, "grad_norm": 1.956624984741211, "learning_rate": 8.90174645167115e-07, "loss": 0.5128, "step": 19392 }, { "epoch": 0.8129277008687633, "grad_norm": 1.749282717704773, "learning_rate": 8.897880509353956e-07, "loss": 0.4889, "step": 19393 }, { "epoch": 0.8129696194837722, "grad_norm": 2.0196290016174316, "learning_rate": 8.894015324697502e-07, "loss": 0.485, "step": 19394 }, { "epoch": 0.8130115380987812, "grad_norm": 2.0010933876037598, "learning_rate": 8.890150897773064e-07, "loss": 0.4746, "step": 19395 }, { "epoch": 0.8130534567137901, "grad_norm": 1.7165813446044922, "learning_rate": 8.886287228651863e-07, "loss": 0.4384, "step": 19396 }, { "epoch": 0.8130953753287992, "grad_norm": 1.8623113632202148, "learning_rate": 8.882424317405103e-07, "loss": 0.4886, "step": 19397 }, { "epoch": 0.8131372939438081, "grad_norm": 2.384765386581421, "learning_rate": 8.878562164104005e-07, "loss": 0.4447, "step": 19398 }, { "epoch": 0.8131792125588171, "grad_norm": 1.879811406135559, "learning_rate": 8.874700768819778e-07, "loss": 0.4701, "step": 19399 }, { "epoch": 0.813221131173826, "grad_norm": 1.5671567916870117, "learning_rate": 8.870840131623582e-07, "loss": 0.3889, "step": 19400 }, { "epoch": 0.813263049788835, "grad_norm": 2.067695379257202, "learning_rate": 8.866980252586594e-07, "loss": 0.5211, "step": 19401 }, { "epoch": 0.8133049684038439, "grad_norm": 1.6959232091903687, "learning_rate": 8.863121131779951e-07, "loss": 0.5107, "step": 19402 }, { "epoch": 0.8133468870188529, "grad_norm": 1.705968976020813, "learning_rate": 8.859262769274801e-07, "loss": 0.4768, "step": 19403 }, { "epoch": 0.8133888056338618, "grad_norm": 2.151031494140625, "learning_rate": 8.855405165142272e-07, "loss": 0.4763, "step": 19404 }, { "epoch": 0.8134307242488709, "grad_norm": 4.548779487609863, "learning_rate": 8.851548319453479e-07, "loss": 0.4811, "step": 19405 }, { "epoch": 0.8134726428638798, "grad_norm": 2.0421462059020996, "learning_rate": 8.847692232279487e-07, "loss": 0.4474, "step": 19406 }, { "epoch": 0.8135145614788888, "grad_norm": 1.695726990699768, "learning_rate": 8.843836903691411e-07, "loss": 0.4266, "step": 19407 }, { "epoch": 0.8135564800938977, "grad_norm": 1.8312746286392212, "learning_rate": 8.839982333760294e-07, "loss": 0.5068, "step": 19408 }, { "epoch": 0.8135983987089067, "grad_norm": 2.067196846008301, "learning_rate": 8.836128522557197e-07, "loss": 0.4695, "step": 19409 }, { "epoch": 0.8136403173239156, "grad_norm": 6.727606296539307, "learning_rate": 8.832275470153179e-07, "loss": 0.4185, "step": 19410 }, { "epoch": 0.8136822359389245, "grad_norm": 1.639891266822815, "learning_rate": 8.828423176619239e-07, "loss": 0.4269, "step": 19411 }, { "epoch": 0.8137241545539335, "grad_norm": 1.936939001083374, "learning_rate": 8.824571642026397e-07, "loss": 0.4335, "step": 19412 }, { "epoch": 0.8137660731689424, "grad_norm": 2.326988458633423, "learning_rate": 8.82072086644567e-07, "loss": 0.4997, "step": 19413 }, { "epoch": 0.8138079917839515, "grad_norm": 1.9928548336029053, "learning_rate": 8.816870849948017e-07, "loss": 0.4613, "step": 19414 }, { "epoch": 0.8138499103989604, "grad_norm": 1.8211876153945923, "learning_rate": 8.81302159260441e-07, "loss": 0.4947, "step": 19415 }, { "epoch": 0.8138918290139694, "grad_norm": 1.7115075588226318, "learning_rate": 8.809173094485823e-07, "loss": 0.4742, "step": 19416 }, { "epoch": 0.8139337476289783, "grad_norm": 2.0954973697662354, "learning_rate": 8.805325355663169e-07, "loss": 0.4553, "step": 19417 }, { "epoch": 0.8139756662439873, "grad_norm": 1.9850449562072754, "learning_rate": 8.801478376207396e-07, "loss": 0.4511, "step": 19418 }, { "epoch": 0.8140175848589962, "grad_norm": 1.9322322607040405, "learning_rate": 8.797632156189429e-07, "loss": 0.466, "step": 19419 }, { "epoch": 0.8140595034740052, "grad_norm": 1.7737370729446411, "learning_rate": 8.793786695680134e-07, "loss": 0.4345, "step": 19420 }, { "epoch": 0.8141014220890141, "grad_norm": 2.2130062580108643, "learning_rate": 8.789941994750417e-07, "loss": 0.4488, "step": 19421 }, { "epoch": 0.8141433407040232, "grad_norm": 1.995403528213501, "learning_rate": 8.786098053471159e-07, "loss": 0.5792, "step": 19422 }, { "epoch": 0.8141852593190321, "grad_norm": 2.0119659900665283, "learning_rate": 8.78225487191321e-07, "loss": 0.4288, "step": 19423 }, { "epoch": 0.8142271779340411, "grad_norm": 1.77647066116333, "learning_rate": 8.7784124501474e-07, "loss": 0.4859, "step": 19424 }, { "epoch": 0.81426909654905, "grad_norm": 1.9467637538909912, "learning_rate": 8.774570788244579e-07, "loss": 0.437, "step": 19425 }, { "epoch": 0.814311015164059, "grad_norm": 2.296149492263794, "learning_rate": 8.770729886275537e-07, "loss": 0.5261, "step": 19426 }, { "epoch": 0.8143529337790679, "grad_norm": 2.151524305343628, "learning_rate": 8.766889744311096e-07, "loss": 0.4288, "step": 19427 }, { "epoch": 0.8143948523940769, "grad_norm": 1.8221628665924072, "learning_rate": 8.763050362422054e-07, "loss": 0.4761, "step": 19428 }, { "epoch": 0.8144367710090858, "grad_norm": 2.4478793144226074, "learning_rate": 8.759211740679152e-07, "loss": 0.4774, "step": 19429 }, { "epoch": 0.8144786896240949, "grad_norm": 1.9134221076965332, "learning_rate": 8.755373879153173e-07, "loss": 0.4547, "step": 19430 }, { "epoch": 0.8145206082391038, "grad_norm": 2.2046682834625244, "learning_rate": 8.751536777914865e-07, "loss": 0.4597, "step": 19431 }, { "epoch": 0.8145625268541128, "grad_norm": 2.0426127910614014, "learning_rate": 8.747700437034945e-07, "loss": 0.4901, "step": 19432 }, { "epoch": 0.8146044454691217, "grad_norm": 1.833182454109192, "learning_rate": 8.74386485658415e-07, "loss": 0.509, "step": 19433 }, { "epoch": 0.8146463640841307, "grad_norm": 1.8877440690994263, "learning_rate": 8.740030036633168e-07, "loss": 0.5117, "step": 19434 }, { "epoch": 0.8146882826991396, "grad_norm": 1.7082513570785522, "learning_rate": 8.736195977252682e-07, "loss": 0.4316, "step": 19435 }, { "epoch": 0.8147302013141485, "grad_norm": 1.9657387733459473, "learning_rate": 8.732362678513373e-07, "loss": 0.5063, "step": 19436 }, { "epoch": 0.8147721199291575, "grad_norm": 2.015789747238159, "learning_rate": 8.728530140485924e-07, "loss": 0.4802, "step": 19437 }, { "epoch": 0.8148140385441665, "grad_norm": 2.118358850479126, "learning_rate": 8.724698363240952e-07, "loss": 0.5044, "step": 19438 }, { "epoch": 0.8148559571591755, "grad_norm": 1.6940600872039795, "learning_rate": 8.72086734684911e-07, "loss": 0.4547, "step": 19439 }, { "epoch": 0.8148978757741844, "grad_norm": 1.890645980834961, "learning_rate": 8.717037091381015e-07, "loss": 0.4903, "step": 19440 }, { "epoch": 0.8149397943891934, "grad_norm": 1.7287925481796265, "learning_rate": 8.713207596907264e-07, "loss": 0.454, "step": 19441 }, { "epoch": 0.8149817130042023, "grad_norm": 2.095319986343384, "learning_rate": 8.709378863498463e-07, "loss": 0.4308, "step": 19442 }, { "epoch": 0.8150236316192113, "grad_norm": 1.8705230951309204, "learning_rate": 8.705550891225178e-07, "loss": 0.4543, "step": 19443 }, { "epoch": 0.8150655502342202, "grad_norm": 1.747689127922058, "learning_rate": 8.701723680157964e-07, "loss": 0.451, "step": 19444 }, { "epoch": 0.8151074688492292, "grad_norm": 2.0573673248291016, "learning_rate": 8.69789723036738e-07, "loss": 0.5296, "step": 19445 }, { "epoch": 0.8151493874642382, "grad_norm": 1.8543580770492554, "learning_rate": 8.694071541923976e-07, "loss": 0.4566, "step": 19446 }, { "epoch": 0.8151913060792472, "grad_norm": 1.772984266281128, "learning_rate": 8.690246614898251e-07, "loss": 0.5334, "step": 19447 }, { "epoch": 0.8152332246942561, "grad_norm": 1.70995032787323, "learning_rate": 8.686422449360715e-07, "loss": 0.4283, "step": 19448 }, { "epoch": 0.8152751433092651, "grad_norm": 2.739125967025757, "learning_rate": 8.682599045381878e-07, "loss": 0.5143, "step": 19449 }, { "epoch": 0.815317061924274, "grad_norm": 2.04731822013855, "learning_rate": 8.678776403032202e-07, "loss": 0.4533, "step": 19450 }, { "epoch": 0.815358980539283, "grad_norm": 2.001673936843872, "learning_rate": 8.674954522382168e-07, "loss": 0.4656, "step": 19451 }, { "epoch": 0.8154008991542919, "grad_norm": 1.5881896018981934, "learning_rate": 8.671133403502214e-07, "loss": 0.4194, "step": 19452 }, { "epoch": 0.8154428177693009, "grad_norm": 2.28147292137146, "learning_rate": 8.667313046462767e-07, "loss": 0.5175, "step": 19453 }, { "epoch": 0.8154847363843098, "grad_norm": 1.9797916412353516, "learning_rate": 8.663493451334265e-07, "loss": 0.5039, "step": 19454 }, { "epoch": 0.8155266549993189, "grad_norm": 1.7113827466964722, "learning_rate": 8.65967461818713e-07, "loss": 0.4558, "step": 19455 }, { "epoch": 0.8155685736143278, "grad_norm": 1.8812872171401978, "learning_rate": 8.655856547091729e-07, "loss": 0.4416, "step": 19456 }, { "epoch": 0.8156104922293368, "grad_norm": 2.1071114540100098, "learning_rate": 8.652039238118453e-07, "loss": 0.4292, "step": 19457 }, { "epoch": 0.8156524108443457, "grad_norm": 2.1374869346618652, "learning_rate": 8.648222691337688e-07, "loss": 0.463, "step": 19458 }, { "epoch": 0.8156943294593547, "grad_norm": 2.014514207839966, "learning_rate": 8.644406906819758e-07, "loss": 0.4895, "step": 19459 }, { "epoch": 0.8157362480743636, "grad_norm": 1.8155077695846558, "learning_rate": 8.640591884635025e-07, "loss": 0.5056, "step": 19460 }, { "epoch": 0.8157781666893725, "grad_norm": 1.69242525100708, "learning_rate": 8.636777624853804e-07, "loss": 0.4402, "step": 19461 }, { "epoch": 0.8158200853043815, "grad_norm": 2.0165789127349854, "learning_rate": 8.63296412754639e-07, "loss": 0.5083, "step": 19462 }, { "epoch": 0.8158620039193905, "grad_norm": 1.8258308172225952, "learning_rate": 8.629151392783097e-07, "loss": 0.4436, "step": 19463 }, { "epoch": 0.8159039225343995, "grad_norm": 3.066532611846924, "learning_rate": 8.625339420634216e-07, "loss": 0.5163, "step": 19464 }, { "epoch": 0.8159458411494084, "grad_norm": 2.4300479888916016, "learning_rate": 8.621528211169988e-07, "loss": 0.5283, "step": 19465 }, { "epoch": 0.8159877597644174, "grad_norm": 1.6571146249771118, "learning_rate": 8.617717764460687e-07, "loss": 0.4817, "step": 19466 }, { "epoch": 0.8160296783794263, "grad_norm": 1.8804905414581299, "learning_rate": 8.613908080576561e-07, "loss": 0.4353, "step": 19467 }, { "epoch": 0.8160715969944353, "grad_norm": 2.3636605739593506, "learning_rate": 8.610099159587814e-07, "loss": 0.4687, "step": 19468 }, { "epoch": 0.8161135156094442, "grad_norm": 1.7001466751098633, "learning_rate": 8.606291001564681e-07, "loss": 0.4331, "step": 19469 }, { "epoch": 0.8161554342244532, "grad_norm": 1.8972188234329224, "learning_rate": 8.602483606577339e-07, "loss": 0.4578, "step": 19470 }, { "epoch": 0.8161973528394622, "grad_norm": 2.0164873600006104, "learning_rate": 8.598676974695991e-07, "loss": 0.47, "step": 19471 }, { "epoch": 0.8162392714544712, "grad_norm": 2.421377182006836, "learning_rate": 8.594871105990782e-07, "loss": 0.5419, "step": 19472 }, { "epoch": 0.8162811900694801, "grad_norm": 1.929441213607788, "learning_rate": 8.5910660005319e-07, "loss": 0.502, "step": 19473 }, { "epoch": 0.8163231086844891, "grad_norm": 2.1553876399993896, "learning_rate": 8.587261658389457e-07, "loss": 0.5173, "step": 19474 }, { "epoch": 0.816365027299498, "grad_norm": 2.1017284393310547, "learning_rate": 8.583458079633599e-07, "loss": 0.5002, "step": 19475 }, { "epoch": 0.816406945914507, "grad_norm": 2.0462422370910645, "learning_rate": 8.57965526433443e-07, "loss": 0.5256, "step": 19476 }, { "epoch": 0.8164488645295159, "grad_norm": 1.673499584197998, "learning_rate": 8.575853212562057e-07, "loss": 0.5268, "step": 19477 }, { "epoch": 0.816490783144525, "grad_norm": 1.7181355953216553, "learning_rate": 8.572051924386571e-07, "loss": 0.473, "step": 19478 }, { "epoch": 0.8165327017595339, "grad_norm": 1.8976227045059204, "learning_rate": 8.568251399878019e-07, "loss": 0.4636, "step": 19479 }, { "epoch": 0.8165746203745429, "grad_norm": 1.889829397201538, "learning_rate": 8.564451639106491e-07, "loss": 0.5136, "step": 19480 }, { "epoch": 0.8166165389895518, "grad_norm": 2.3947911262512207, "learning_rate": 8.560652642142014e-07, "loss": 0.4752, "step": 19481 }, { "epoch": 0.8166584576045608, "grad_norm": 2.1265411376953125, "learning_rate": 8.55685440905461e-07, "loss": 0.4668, "step": 19482 }, { "epoch": 0.8167003762195697, "grad_norm": 1.8343678712844849, "learning_rate": 8.553056939914295e-07, "loss": 0.5066, "step": 19483 }, { "epoch": 0.8167422948345787, "grad_norm": 1.9607151746749878, "learning_rate": 8.549260234791091e-07, "loss": 0.4439, "step": 19484 }, { "epoch": 0.8167842134495876, "grad_norm": 1.8838930130004883, "learning_rate": 8.545464293754957e-07, "loss": 0.5047, "step": 19485 }, { "epoch": 0.8168261320645965, "grad_norm": 1.7720930576324463, "learning_rate": 8.541669116875878e-07, "loss": 0.4719, "step": 19486 }, { "epoch": 0.8168680506796056, "grad_norm": 1.8615282773971558, "learning_rate": 8.537874704223831e-07, "loss": 0.4777, "step": 19487 }, { "epoch": 0.8169099692946145, "grad_norm": 1.7871437072753906, "learning_rate": 8.534081055868732e-07, "loss": 0.5032, "step": 19488 }, { "epoch": 0.8169518879096235, "grad_norm": 1.8365328311920166, "learning_rate": 8.530288171880535e-07, "loss": 0.4603, "step": 19489 }, { "epoch": 0.8169938065246324, "grad_norm": 2.0466952323913574, "learning_rate": 8.526496052329142e-07, "loss": 0.5326, "step": 19490 }, { "epoch": 0.8170357251396414, "grad_norm": 1.9244898557662964, "learning_rate": 8.522704697284451e-07, "loss": 0.4241, "step": 19491 }, { "epoch": 0.8170776437546503, "grad_norm": 2.332590341567993, "learning_rate": 8.518914106816356e-07, "loss": 0.5055, "step": 19492 }, { "epoch": 0.8171195623696593, "grad_norm": 1.9857535362243652, "learning_rate": 8.515124280994747e-07, "loss": 0.4595, "step": 19493 }, { "epoch": 0.8171614809846682, "grad_norm": 1.9107871055603027, "learning_rate": 8.51133521988946e-07, "loss": 0.4717, "step": 19494 }, { "epoch": 0.8172033995996772, "grad_norm": 2.813908100128174, "learning_rate": 8.507546923570348e-07, "loss": 0.4421, "step": 19495 }, { "epoch": 0.8172453182146862, "grad_norm": 1.9203437566757202, "learning_rate": 8.503759392107258e-07, "loss": 0.4238, "step": 19496 }, { "epoch": 0.8172872368296952, "grad_norm": 1.911688208580017, "learning_rate": 8.499972625569991e-07, "loss": 0.4507, "step": 19497 }, { "epoch": 0.8173291554447041, "grad_norm": 2.2664875984191895, "learning_rate": 8.496186624028363e-07, "loss": 0.4555, "step": 19498 }, { "epoch": 0.8173710740597131, "grad_norm": 2.287752628326416, "learning_rate": 8.492401387552162e-07, "loss": 0.4797, "step": 19499 }, { "epoch": 0.817412992674722, "grad_norm": 1.606762409210205, "learning_rate": 8.488616916211145e-07, "loss": 0.455, "step": 19500 }, { "epoch": 0.817454911289731, "grad_norm": 1.74209725856781, "learning_rate": 8.484833210075083e-07, "loss": 0.4151, "step": 19501 }, { "epoch": 0.8174968299047399, "grad_norm": 1.9688609838485718, "learning_rate": 8.481050269213748e-07, "loss": 0.5071, "step": 19502 }, { "epoch": 0.817538748519749, "grad_norm": 1.909798502922058, "learning_rate": 8.477268093696833e-07, "loss": 0.4831, "step": 19503 }, { "epoch": 0.8175806671347579, "grad_norm": 1.9218579530715942, "learning_rate": 8.473486683594079e-07, "loss": 0.4673, "step": 19504 }, { "epoch": 0.8176225857497669, "grad_norm": 2.7255029678344727, "learning_rate": 8.469706038975206e-07, "loss": 0.5065, "step": 19505 }, { "epoch": 0.8176645043647758, "grad_norm": 2.140686273574829, "learning_rate": 8.465926159909871e-07, "loss": 0.463, "step": 19506 }, { "epoch": 0.8177064229797848, "grad_norm": 1.8431717157363892, "learning_rate": 8.462147046467784e-07, "loss": 0.4522, "step": 19507 }, { "epoch": 0.8177483415947937, "grad_norm": 2.2590091228485107, "learning_rate": 8.458368698718589e-07, "loss": 0.4513, "step": 19508 }, { "epoch": 0.8177902602098027, "grad_norm": 2.3919169902801514, "learning_rate": 8.454591116731931e-07, "loss": 0.4758, "step": 19509 }, { "epoch": 0.8178321788248116, "grad_norm": 1.6713148355484009, "learning_rate": 8.450814300577442e-07, "loss": 0.4558, "step": 19510 }, { "epoch": 0.8178740974398205, "grad_norm": 1.7998987436294556, "learning_rate": 8.447038250324769e-07, "loss": 0.4608, "step": 19511 }, { "epoch": 0.8179160160548296, "grad_norm": 2.201723337173462, "learning_rate": 8.443262966043486e-07, "loss": 0.4585, "step": 19512 }, { "epoch": 0.8179579346698385, "grad_norm": 2.2200512886047363, "learning_rate": 8.439488447803207e-07, "loss": 0.595, "step": 19513 }, { "epoch": 0.8179998532848475, "grad_norm": 1.7053296566009521, "learning_rate": 8.435714695673508e-07, "loss": 0.4011, "step": 19514 }, { "epoch": 0.8180417718998564, "grad_norm": 1.892763614654541, "learning_rate": 8.431941709723935e-07, "loss": 0.4657, "step": 19515 }, { "epoch": 0.8180836905148654, "grad_norm": 2.199631452560425, "learning_rate": 8.428169490024063e-07, "loss": 0.4354, "step": 19516 }, { "epoch": 0.8181256091298743, "grad_norm": 2.4752657413482666, "learning_rate": 8.424398036643405e-07, "loss": 0.5169, "step": 19517 }, { "epoch": 0.8181675277448833, "grad_norm": 1.7753297090530396, "learning_rate": 8.420627349651506e-07, "loss": 0.4861, "step": 19518 }, { "epoch": 0.8182094463598922, "grad_norm": 2.0023117065429688, "learning_rate": 8.416857429117848e-07, "loss": 0.4556, "step": 19519 }, { "epoch": 0.8182513649749013, "grad_norm": 1.9505391120910645, "learning_rate": 8.413088275111947e-07, "loss": 0.4775, "step": 19520 }, { "epoch": 0.8182932835899102, "grad_norm": 1.786434531211853, "learning_rate": 8.409319887703255e-07, "loss": 0.4503, "step": 19521 }, { "epoch": 0.8183352022049192, "grad_norm": 1.856136441230774, "learning_rate": 8.405552266961259e-07, "loss": 0.4445, "step": 19522 }, { "epoch": 0.8183771208199281, "grad_norm": 1.6936224699020386, "learning_rate": 8.401785412955415e-07, "loss": 0.4631, "step": 19523 }, { "epoch": 0.8184190394349371, "grad_norm": 1.6341663599014282, "learning_rate": 8.398019325755136e-07, "loss": 0.5051, "step": 19524 }, { "epoch": 0.818460958049946, "grad_norm": 3.586078643798828, "learning_rate": 8.394254005429858e-07, "loss": 0.5515, "step": 19525 }, { "epoch": 0.818502876664955, "grad_norm": 1.9307711124420166, "learning_rate": 8.390489452048999e-07, "loss": 0.5305, "step": 19526 }, { "epoch": 0.8185447952799639, "grad_norm": 1.785474419593811, "learning_rate": 8.386725665681949e-07, "loss": 0.4657, "step": 19527 }, { "epoch": 0.818586713894973, "grad_norm": 2.087797164916992, "learning_rate": 8.382962646398068e-07, "loss": 0.4942, "step": 19528 }, { "epoch": 0.8186286325099819, "grad_norm": 2.4217658042907715, "learning_rate": 8.379200394266751e-07, "loss": 0.5498, "step": 19529 }, { "epoch": 0.8186705511249909, "grad_norm": 2.1349265575408936, "learning_rate": 8.37543890935732e-07, "loss": 0.4433, "step": 19530 }, { "epoch": 0.8187124697399998, "grad_norm": 1.9963409900665283, "learning_rate": 8.37167819173913e-07, "loss": 0.4545, "step": 19531 }, { "epoch": 0.8187543883550088, "grad_norm": 3.0462350845336914, "learning_rate": 8.367918241481515e-07, "loss": 0.5104, "step": 19532 }, { "epoch": 0.8187963069700177, "grad_norm": 1.6674659252166748, "learning_rate": 8.364159058653759e-07, "loss": 0.476, "step": 19533 }, { "epoch": 0.8188382255850267, "grad_norm": 1.7181967496871948, "learning_rate": 8.360400643325173e-07, "loss": 0.5399, "step": 19534 }, { "epoch": 0.8188801442000356, "grad_norm": 1.7769827842712402, "learning_rate": 8.356642995565051e-07, "loss": 0.4949, "step": 19535 }, { "epoch": 0.8189220628150445, "grad_norm": 2.7305855751037598, "learning_rate": 8.352886115442644e-07, "loss": 0.475, "step": 19536 }, { "epoch": 0.8189639814300536, "grad_norm": 1.8319761753082275, "learning_rate": 8.34913000302719e-07, "loss": 0.4453, "step": 19537 }, { "epoch": 0.8190059000450625, "grad_norm": 1.8165323734283447, "learning_rate": 8.345374658387956e-07, "loss": 0.4511, "step": 19538 }, { "epoch": 0.8190478186600715, "grad_norm": 2.2678732872009277, "learning_rate": 8.341620081594143e-07, "loss": 0.4763, "step": 19539 }, { "epoch": 0.8190897372750804, "grad_norm": 1.862087607383728, "learning_rate": 8.337866272714978e-07, "loss": 0.4789, "step": 19540 }, { "epoch": 0.8191316558900894, "grad_norm": 1.6806588172912598, "learning_rate": 8.334113231819657e-07, "loss": 0.4343, "step": 19541 }, { "epoch": 0.8191735745050983, "grad_norm": 2.4552876949310303, "learning_rate": 8.330360958977351e-07, "loss": 0.4565, "step": 19542 }, { "epoch": 0.8192154931201073, "grad_norm": 1.9736526012420654, "learning_rate": 8.326609454257229e-07, "loss": 0.4593, "step": 19543 }, { "epoch": 0.8192574117351162, "grad_norm": 1.6550825834274292, "learning_rate": 8.322858717728466e-07, "loss": 0.5236, "step": 19544 }, { "epoch": 0.8192993303501253, "grad_norm": 1.923747181892395, "learning_rate": 8.31910874946018e-07, "loss": 0.4718, "step": 19545 }, { "epoch": 0.8193412489651342, "grad_norm": 1.8090335130691528, "learning_rate": 8.315359549521496e-07, "loss": 0.4747, "step": 19546 }, { "epoch": 0.8193831675801432, "grad_norm": 2.223813533782959, "learning_rate": 8.311611117981544e-07, "loss": 0.5042, "step": 19547 }, { "epoch": 0.8194250861951521, "grad_norm": 2.3083701133728027, "learning_rate": 8.307863454909393e-07, "loss": 0.4743, "step": 19548 }, { "epoch": 0.8194670048101611, "grad_norm": 1.8440592288970947, "learning_rate": 8.304116560374142e-07, "loss": 0.4623, "step": 19549 }, { "epoch": 0.81950892342517, "grad_norm": 1.9218109846115112, "learning_rate": 8.300370434444871e-07, "loss": 0.5071, "step": 19550 }, { "epoch": 0.819550842040179, "grad_norm": 2.210052728652954, "learning_rate": 8.296625077190612e-07, "loss": 0.5051, "step": 19551 }, { "epoch": 0.8195927606551879, "grad_norm": 1.810174584388733, "learning_rate": 8.292880488680433e-07, "loss": 0.476, "step": 19552 }, { "epoch": 0.819634679270197, "grad_norm": 1.5958377122879028, "learning_rate": 8.289136668983328e-07, "loss": 0.4985, "step": 19553 }, { "epoch": 0.8196765978852059, "grad_norm": 1.5302762985229492, "learning_rate": 8.285393618168336e-07, "loss": 0.4386, "step": 19554 }, { "epoch": 0.8197185165002149, "grad_norm": 2.0808112621307373, "learning_rate": 8.281651336304442e-07, "loss": 0.4855, "step": 19555 }, { "epoch": 0.8197604351152238, "grad_norm": 1.9406965970993042, "learning_rate": 8.277909823460622e-07, "loss": 0.4565, "step": 19556 }, { "epoch": 0.8198023537302328, "grad_norm": 2.2047348022460938, "learning_rate": 8.274169079705858e-07, "loss": 0.4974, "step": 19557 }, { "epoch": 0.8198442723452417, "grad_norm": 1.932798981666565, "learning_rate": 8.270429105109112e-07, "loss": 0.4741, "step": 19558 }, { "epoch": 0.8198861909602507, "grad_norm": 1.7725660800933838, "learning_rate": 8.266689899739299e-07, "loss": 0.5028, "step": 19559 }, { "epoch": 0.8199281095752596, "grad_norm": 1.6644458770751953, "learning_rate": 8.26295146366537e-07, "loss": 0.4721, "step": 19560 }, { "epoch": 0.8199700281902685, "grad_norm": 1.7555185556411743, "learning_rate": 8.259213796956245e-07, "loss": 0.4222, "step": 19561 }, { "epoch": 0.8200119468052776, "grad_norm": 1.6732455492019653, "learning_rate": 8.255476899680792e-07, "loss": 0.4187, "step": 19562 }, { "epoch": 0.8200538654202865, "grad_norm": 2.146207094192505, "learning_rate": 8.251740771907923e-07, "loss": 0.4601, "step": 19563 }, { "epoch": 0.8200957840352955, "grad_norm": 2.100008964538574, "learning_rate": 8.248005413706489e-07, "loss": 0.4313, "step": 19564 }, { "epoch": 0.8201377026503044, "grad_norm": 1.9995723962783813, "learning_rate": 8.244270825145367e-07, "loss": 0.5356, "step": 19565 }, { "epoch": 0.8201796212653134, "grad_norm": 2.5707032680511475, "learning_rate": 8.240537006293376e-07, "loss": 0.4597, "step": 19566 }, { "epoch": 0.8202215398803223, "grad_norm": 1.9809118509292603, "learning_rate": 8.236803957219363e-07, "loss": 0.51, "step": 19567 }, { "epoch": 0.8202634584953313, "grad_norm": 1.6355009078979492, "learning_rate": 8.233071677992122e-07, "loss": 0.443, "step": 19568 }, { "epoch": 0.8203053771103402, "grad_norm": 1.673344612121582, "learning_rate": 8.229340168680461e-07, "loss": 0.4466, "step": 19569 }, { "epoch": 0.8203472957253493, "grad_norm": 2.1315536499023438, "learning_rate": 8.225609429353187e-07, "loss": 0.4489, "step": 19570 }, { "epoch": 0.8203892143403582, "grad_norm": 1.8980209827423096, "learning_rate": 8.221879460079035e-07, "loss": 0.4445, "step": 19571 }, { "epoch": 0.8204311329553672, "grad_norm": 3.756352424621582, "learning_rate": 8.218150260926777e-07, "loss": 0.4796, "step": 19572 }, { "epoch": 0.8204730515703761, "grad_norm": 1.7758173942565918, "learning_rate": 8.214421831965169e-07, "loss": 0.4676, "step": 19573 }, { "epoch": 0.8205149701853851, "grad_norm": 2.135772228240967, "learning_rate": 8.210694173262934e-07, "loss": 0.4933, "step": 19574 }, { "epoch": 0.820556888800394, "grad_norm": 1.8592453002929688, "learning_rate": 8.206967284888762e-07, "loss": 0.4908, "step": 19575 }, { "epoch": 0.820598807415403, "grad_norm": 1.7875256538391113, "learning_rate": 8.203241166911386e-07, "loss": 0.4305, "step": 19576 }, { "epoch": 0.8206407260304119, "grad_norm": 1.617148995399475, "learning_rate": 8.199515819399462e-07, "loss": 0.4559, "step": 19577 }, { "epoch": 0.820682644645421, "grad_norm": 2.14924955368042, "learning_rate": 8.19579124242168e-07, "loss": 0.4876, "step": 19578 }, { "epoch": 0.8207245632604299, "grad_norm": 1.7736413478851318, "learning_rate": 8.192067436046703e-07, "loss": 0.4472, "step": 19579 }, { "epoch": 0.8207664818754389, "grad_norm": 1.8883593082427979, "learning_rate": 8.188344400343156e-07, "loss": 0.5029, "step": 19580 }, { "epoch": 0.8208084004904478, "grad_norm": 2.242258310317993, "learning_rate": 8.184622135379677e-07, "loss": 0.495, "step": 19581 }, { "epoch": 0.8208503191054568, "grad_norm": 1.7662605047225952, "learning_rate": 8.180900641224887e-07, "loss": 0.4328, "step": 19582 }, { "epoch": 0.8208922377204657, "grad_norm": 2.220484733581543, "learning_rate": 8.177179917947387e-07, "loss": 0.4676, "step": 19583 }, { "epoch": 0.8209341563354747, "grad_norm": 1.8066208362579346, "learning_rate": 8.173459965615744e-07, "loss": 0.4568, "step": 19584 }, { "epoch": 0.8209760749504836, "grad_norm": 1.627248764038086, "learning_rate": 8.169740784298552e-07, "loss": 0.4345, "step": 19585 }, { "epoch": 0.8210179935654925, "grad_norm": 1.6129744052886963, "learning_rate": 8.166022374064353e-07, "loss": 0.4752, "step": 19586 }, { "epoch": 0.8210599121805016, "grad_norm": 3.3605542182922363, "learning_rate": 8.162304734981691e-07, "loss": 0.4534, "step": 19587 }, { "epoch": 0.8211018307955105, "grad_norm": 1.8068106174468994, "learning_rate": 8.158587867119117e-07, "loss": 0.4663, "step": 19588 }, { "epoch": 0.8211437494105195, "grad_norm": 1.8157581090927124, "learning_rate": 8.154871770545125e-07, "loss": 0.4337, "step": 19589 }, { "epoch": 0.8211856680255284, "grad_norm": 1.8882579803466797, "learning_rate": 8.151156445328218e-07, "loss": 0.4989, "step": 19590 }, { "epoch": 0.8212275866405374, "grad_norm": 1.552520751953125, "learning_rate": 8.1474418915369e-07, "loss": 0.4096, "step": 19591 }, { "epoch": 0.8212695052555463, "grad_norm": 1.7610965967178345, "learning_rate": 8.14372810923963e-07, "loss": 0.4378, "step": 19592 }, { "epoch": 0.8213114238705553, "grad_norm": 1.9501436948776245, "learning_rate": 8.140015098504856e-07, "loss": 0.4576, "step": 19593 }, { "epoch": 0.8213533424855642, "grad_norm": 1.7815271615982056, "learning_rate": 8.136302859401046e-07, "loss": 0.4479, "step": 19594 }, { "epoch": 0.8213952611005733, "grad_norm": 1.9085636138916016, "learning_rate": 8.132591391996608e-07, "loss": 0.4753, "step": 19595 }, { "epoch": 0.8214371797155822, "grad_norm": 1.7943624258041382, "learning_rate": 8.128880696359964e-07, "loss": 0.501, "step": 19596 }, { "epoch": 0.8214790983305912, "grad_norm": 2.203003168106079, "learning_rate": 8.125170772559537e-07, "loss": 0.4644, "step": 19597 }, { "epoch": 0.8215210169456001, "grad_norm": 1.5929523706436157, "learning_rate": 8.121461620663679e-07, "loss": 0.4796, "step": 19598 }, { "epoch": 0.8215629355606091, "grad_norm": 1.7690799236297607, "learning_rate": 8.117753240740783e-07, "loss": 0.4737, "step": 19599 }, { "epoch": 0.821604854175618, "grad_norm": 1.5768780708312988, "learning_rate": 8.114045632859219e-07, "loss": 0.4094, "step": 19600 }, { "epoch": 0.821646772790627, "grad_norm": 2.125831365585327, "learning_rate": 8.11033879708732e-07, "loss": 0.4747, "step": 19601 }, { "epoch": 0.8216886914056359, "grad_norm": 1.7427442073822021, "learning_rate": 8.106632733493397e-07, "loss": 0.4275, "step": 19602 }, { "epoch": 0.821730610020645, "grad_norm": 2.569612503051758, "learning_rate": 8.102927442145797e-07, "loss": 0.4937, "step": 19603 }, { "epoch": 0.8217725286356539, "grad_norm": 1.9736570119857788, "learning_rate": 8.099222923112798e-07, "loss": 0.5133, "step": 19604 }, { "epoch": 0.8218144472506629, "grad_norm": 2.3806145191192627, "learning_rate": 8.095519176462702e-07, "loss": 0.5012, "step": 19605 }, { "epoch": 0.8218563658656718, "grad_norm": 1.9557822942733765, "learning_rate": 8.091816202263786e-07, "loss": 0.4376, "step": 19606 }, { "epoch": 0.8218982844806808, "grad_norm": 2.029987096786499, "learning_rate": 8.08811400058429e-07, "loss": 0.4494, "step": 19607 }, { "epoch": 0.8219402030956897, "grad_norm": 2.6413114070892334, "learning_rate": 8.084412571492472e-07, "loss": 0.489, "step": 19608 }, { "epoch": 0.8219821217106987, "grad_norm": 2.1436665058135986, "learning_rate": 8.080711915056571e-07, "loss": 0.4495, "step": 19609 }, { "epoch": 0.8220240403257076, "grad_norm": 2.5999186038970947, "learning_rate": 8.0770120313448e-07, "loss": 0.4962, "step": 19610 }, { "epoch": 0.8220659589407165, "grad_norm": 1.7364546060562134, "learning_rate": 8.073312920425342e-07, "loss": 0.4849, "step": 19611 }, { "epoch": 0.8221078775557256, "grad_norm": 1.8041757345199585, "learning_rate": 8.069614582366408e-07, "loss": 0.5129, "step": 19612 }, { "epoch": 0.8221497961707345, "grad_norm": 2.0195767879486084, "learning_rate": 8.065917017236146e-07, "loss": 0.4853, "step": 19613 }, { "epoch": 0.8221917147857435, "grad_norm": 1.9905896186828613, "learning_rate": 8.06222022510274e-07, "loss": 0.5554, "step": 19614 }, { "epoch": 0.8222336334007524, "grad_norm": 1.9708296060562134, "learning_rate": 8.058524206034334e-07, "loss": 0.5011, "step": 19615 }, { "epoch": 0.8222755520157614, "grad_norm": 2.0117406845092773, "learning_rate": 8.054828960099043e-07, "loss": 0.5101, "step": 19616 }, { "epoch": 0.8223174706307703, "grad_norm": 1.881069302558899, "learning_rate": 8.051134487364986e-07, "loss": 0.483, "step": 19617 }, { "epoch": 0.8223593892457793, "grad_norm": 2.091158390045166, "learning_rate": 8.04744078790029e-07, "loss": 0.4594, "step": 19618 }, { "epoch": 0.8224013078607882, "grad_norm": 1.7279223203659058, "learning_rate": 8.043747861773016e-07, "loss": 0.418, "step": 19619 }, { "epoch": 0.8224432264757973, "grad_norm": 1.659368872642517, "learning_rate": 8.040055709051254e-07, "loss": 0.4986, "step": 19620 }, { "epoch": 0.8224851450908062, "grad_norm": 2.4672303199768066, "learning_rate": 8.036364329803059e-07, "loss": 0.4372, "step": 19621 }, { "epoch": 0.8225270637058152, "grad_norm": 1.6484616994857788, "learning_rate": 8.032673724096462e-07, "loss": 0.4559, "step": 19622 }, { "epoch": 0.8225689823208241, "grad_norm": 2.0767321586608887, "learning_rate": 8.028983891999509e-07, "loss": 0.4922, "step": 19623 }, { "epoch": 0.8226109009358331, "grad_norm": 1.7897634506225586, "learning_rate": 8.025294833580221e-07, "loss": 0.4518, "step": 19624 }, { "epoch": 0.822652819550842, "grad_norm": 2.0220706462860107, "learning_rate": 8.021606548906591e-07, "loss": 0.4989, "step": 19625 }, { "epoch": 0.822694738165851, "grad_norm": 1.962369441986084, "learning_rate": 8.017919038046618e-07, "loss": 0.423, "step": 19626 }, { "epoch": 0.8227366567808599, "grad_norm": 2.261080503463745, "learning_rate": 8.014232301068258e-07, "loss": 0.5086, "step": 19627 }, { "epoch": 0.822778575395869, "grad_norm": 1.992367148399353, "learning_rate": 8.010546338039482e-07, "loss": 0.5006, "step": 19628 }, { "epoch": 0.8228204940108779, "grad_norm": 1.7991276979446411, "learning_rate": 8.006861149028244e-07, "loss": 0.4461, "step": 19629 }, { "epoch": 0.8228624126258869, "grad_norm": 1.8830115795135498, "learning_rate": 8.003176734102469e-07, "loss": 0.49, "step": 19630 }, { "epoch": 0.8229043312408958, "grad_norm": 2.9085686206817627, "learning_rate": 7.999493093330063e-07, "loss": 0.4778, "step": 19631 }, { "epoch": 0.8229462498559048, "grad_norm": 1.6489276885986328, "learning_rate": 7.99581022677895e-07, "loss": 0.5035, "step": 19632 }, { "epoch": 0.8229881684709137, "grad_norm": 2.0615310668945312, "learning_rate": 7.992128134516991e-07, "loss": 0.4737, "step": 19633 }, { "epoch": 0.8230300870859227, "grad_norm": 2.175248622894287, "learning_rate": 7.988446816612077e-07, "loss": 0.4643, "step": 19634 }, { "epoch": 0.8230720057009316, "grad_norm": 1.67628014087677, "learning_rate": 7.984766273132078e-07, "loss": 0.4624, "step": 19635 }, { "epoch": 0.8231139243159405, "grad_norm": 4.1547417640686035, "learning_rate": 7.98108650414482e-07, "loss": 0.5108, "step": 19636 }, { "epoch": 0.8231558429309496, "grad_norm": 1.6622135639190674, "learning_rate": 7.977407509718138e-07, "loss": 0.4931, "step": 19637 }, { "epoch": 0.8231977615459585, "grad_norm": 2.014230966567993, "learning_rate": 7.973729289919868e-07, "loss": 0.5072, "step": 19638 }, { "epoch": 0.8232396801609675, "grad_norm": 2.2175545692443848, "learning_rate": 7.9700518448178e-07, "loss": 0.4892, "step": 19639 }, { "epoch": 0.8232815987759764, "grad_norm": 2.0333335399627686, "learning_rate": 7.966375174479707e-07, "loss": 0.4646, "step": 19640 }, { "epoch": 0.8233235173909854, "grad_norm": 1.84213125705719, "learning_rate": 7.96269927897339e-07, "loss": 0.503, "step": 19641 }, { "epoch": 0.8233654360059943, "grad_norm": 1.7742747068405151, "learning_rate": 7.959024158366585e-07, "loss": 0.4685, "step": 19642 }, { "epoch": 0.8234073546210033, "grad_norm": 1.7841755151748657, "learning_rate": 7.955349812727053e-07, "loss": 0.5085, "step": 19643 }, { "epoch": 0.8234492732360122, "grad_norm": 1.9594953060150146, "learning_rate": 7.951676242122535e-07, "loss": 0.4626, "step": 19644 }, { "epoch": 0.8234911918510213, "grad_norm": 2.146462917327881, "learning_rate": 7.948003446620716e-07, "loss": 0.467, "step": 19645 }, { "epoch": 0.8235331104660302, "grad_norm": 3.053485155105591, "learning_rate": 7.944331426289326e-07, "loss": 0.5191, "step": 19646 }, { "epoch": 0.8235750290810392, "grad_norm": 1.9995540380477905, "learning_rate": 7.940660181196058e-07, "loss": 0.4818, "step": 19647 }, { "epoch": 0.8236169476960481, "grad_norm": 1.6406457424163818, "learning_rate": 7.936989711408571e-07, "loss": 0.4835, "step": 19648 }, { "epoch": 0.8236588663110571, "grad_norm": 2.4593887329101562, "learning_rate": 7.933320016994517e-07, "loss": 0.5132, "step": 19649 }, { "epoch": 0.823700784926066, "grad_norm": 1.9088654518127441, "learning_rate": 7.929651098021562e-07, "loss": 0.4618, "step": 19650 }, { "epoch": 0.823742703541075, "grad_norm": 3.5295984745025635, "learning_rate": 7.925982954557321e-07, "loss": 0.5129, "step": 19651 }, { "epoch": 0.823784622156084, "grad_norm": 2.651876211166382, "learning_rate": 7.922315586669416e-07, "loss": 0.5059, "step": 19652 }, { "epoch": 0.823826540771093, "grad_norm": 1.7284619808197021, "learning_rate": 7.918648994425465e-07, "loss": 0.5037, "step": 19653 }, { "epoch": 0.8238684593861019, "grad_norm": 2.8390889167785645, "learning_rate": 7.914983177893032e-07, "loss": 0.4457, "step": 19654 }, { "epoch": 0.8239103780011109, "grad_norm": 2.203308343887329, "learning_rate": 7.911318137139707e-07, "loss": 0.4639, "step": 19655 }, { "epoch": 0.8239522966161198, "grad_norm": 1.6809755563735962, "learning_rate": 7.907653872233051e-07, "loss": 0.4546, "step": 19656 }, { "epoch": 0.8239942152311288, "grad_norm": 1.9757646322250366, "learning_rate": 7.903990383240611e-07, "loss": 0.4083, "step": 19657 }, { "epoch": 0.8240361338461377, "grad_norm": 1.8174477815628052, "learning_rate": 7.900327670229896e-07, "loss": 0.4622, "step": 19658 }, { "epoch": 0.8240780524611467, "grad_norm": 1.811355471611023, "learning_rate": 7.896665733268455e-07, "loss": 0.5153, "step": 19659 }, { "epoch": 0.8241199710761556, "grad_norm": 2.038723945617676, "learning_rate": 7.893004572423762e-07, "loss": 0.4865, "step": 19660 }, { "epoch": 0.8241618896911646, "grad_norm": 1.828215479850769, "learning_rate": 7.889344187763321e-07, "loss": 0.4751, "step": 19661 }, { "epoch": 0.8242038083061736, "grad_norm": 1.930497407913208, "learning_rate": 7.885684579354607e-07, "loss": 0.492, "step": 19662 }, { "epoch": 0.8242457269211825, "grad_norm": 1.80362868309021, "learning_rate": 7.882025747265076e-07, "loss": 0.4909, "step": 19663 }, { "epoch": 0.8242876455361915, "grad_norm": 2.736063241958618, "learning_rate": 7.878367691562167e-07, "loss": 0.4504, "step": 19664 }, { "epoch": 0.8243295641512004, "grad_norm": 2.478797197341919, "learning_rate": 7.87471041231333e-07, "loss": 0.4757, "step": 19665 }, { "epoch": 0.8243714827662094, "grad_norm": 2.4280130863189697, "learning_rate": 7.871053909585963e-07, "loss": 0.4917, "step": 19666 }, { "epoch": 0.8244134013812183, "grad_norm": 1.84312903881073, "learning_rate": 7.867398183447483e-07, "loss": 0.4825, "step": 19667 }, { "epoch": 0.8244553199962273, "grad_norm": 1.86553156375885, "learning_rate": 7.863743233965271e-07, "loss": 0.4662, "step": 19668 }, { "epoch": 0.8244972386112362, "grad_norm": 1.8454798460006714, "learning_rate": 7.860089061206688e-07, "loss": 0.4811, "step": 19669 }, { "epoch": 0.8245391572262453, "grad_norm": 1.7334721088409424, "learning_rate": 7.85643566523911e-07, "loss": 0.5143, "step": 19670 }, { "epoch": 0.8245810758412542, "grad_norm": 2.178868293762207, "learning_rate": 7.852783046129892e-07, "loss": 0.4831, "step": 19671 }, { "epoch": 0.8246229944562632, "grad_norm": 1.8494316339492798, "learning_rate": 7.849131203946337e-07, "loss": 0.4841, "step": 19672 }, { "epoch": 0.8246649130712721, "grad_norm": 2.333892583847046, "learning_rate": 7.845480138755773e-07, "loss": 0.4992, "step": 19673 }, { "epoch": 0.8247068316862811, "grad_norm": 2.3143482208251953, "learning_rate": 7.841829850625526e-07, "loss": 0.4624, "step": 19674 }, { "epoch": 0.82474875030129, "grad_norm": 1.9977738857269287, "learning_rate": 7.838180339622847e-07, "loss": 0.4827, "step": 19675 }, { "epoch": 0.824790668916299, "grad_norm": 1.7812117338180542, "learning_rate": 7.834531605815038e-07, "loss": 0.4679, "step": 19676 }, { "epoch": 0.824832587531308, "grad_norm": 1.957580327987671, "learning_rate": 7.830883649269344e-07, "loss": 0.4874, "step": 19677 }, { "epoch": 0.824874506146317, "grad_norm": 15.124992370605469, "learning_rate": 7.827236470053007e-07, "loss": 0.4738, "step": 19678 }, { "epoch": 0.8249164247613259, "grad_norm": 1.7941668033599854, "learning_rate": 7.823590068233262e-07, "loss": 0.5025, "step": 19679 }, { "epoch": 0.8249583433763349, "grad_norm": 2.2566640377044678, "learning_rate": 7.819944443877336e-07, "loss": 0.4837, "step": 19680 }, { "epoch": 0.8250002619913438, "grad_norm": 1.9903901815414429, "learning_rate": 7.816299597052412e-07, "loss": 0.4636, "step": 19681 }, { "epoch": 0.8250421806063528, "grad_norm": 1.980044960975647, "learning_rate": 7.812655527825686e-07, "loss": 0.4659, "step": 19682 }, { "epoch": 0.8250840992213617, "grad_norm": 2.5408833026885986, "learning_rate": 7.809012236264347e-07, "loss": 0.443, "step": 19683 }, { "epoch": 0.8251260178363707, "grad_norm": 2.018299102783203, "learning_rate": 7.805369722435529e-07, "loss": 0.48, "step": 19684 }, { "epoch": 0.8251679364513796, "grad_norm": 1.9163424968719482, "learning_rate": 7.801727986406399e-07, "loss": 0.4789, "step": 19685 }, { "epoch": 0.8252098550663886, "grad_norm": 1.9597491025924683, "learning_rate": 7.798087028244078e-07, "loss": 0.494, "step": 19686 }, { "epoch": 0.8252517736813976, "grad_norm": 2.6024587154388428, "learning_rate": 7.794446848015663e-07, "loss": 0.414, "step": 19687 }, { "epoch": 0.8252936922964065, "grad_norm": 2.9364943504333496, "learning_rate": 7.790807445788279e-07, "loss": 0.4784, "step": 19688 }, { "epoch": 0.8253356109114155, "grad_norm": 1.9897576570510864, "learning_rate": 7.787168821629016e-07, "loss": 0.5156, "step": 19689 }, { "epoch": 0.8253775295264244, "grad_norm": 2.063988447189331, "learning_rate": 7.783530975604924e-07, "loss": 0.4781, "step": 19690 }, { "epoch": 0.8254194481414334, "grad_norm": 3.4700071811676025, "learning_rate": 7.779893907783081e-07, "loss": 0.4801, "step": 19691 }, { "epoch": 0.8254613667564423, "grad_norm": 2.0157790184020996, "learning_rate": 7.776257618230532e-07, "loss": 0.4987, "step": 19692 }, { "epoch": 0.8255032853714513, "grad_norm": 3.6429498195648193, "learning_rate": 7.772622107014293e-07, "loss": 0.4316, "step": 19693 }, { "epoch": 0.8255452039864603, "grad_norm": 1.9664887189865112, "learning_rate": 7.768987374201403e-07, "loss": 0.497, "step": 19694 }, { "epoch": 0.8255871226014693, "grad_norm": 2.0536811351776123, "learning_rate": 7.765353419858845e-07, "loss": 0.5296, "step": 19695 }, { "epoch": 0.8256290412164782, "grad_norm": 1.655397891998291, "learning_rate": 7.761720244053594e-07, "loss": 0.4765, "step": 19696 }, { "epoch": 0.8256709598314872, "grad_norm": 1.8574014902114868, "learning_rate": 7.758087846852641e-07, "loss": 0.4926, "step": 19697 }, { "epoch": 0.8257128784464961, "grad_norm": 2.3426566123962402, "learning_rate": 7.754456228322954e-07, "loss": 0.4431, "step": 19698 }, { "epoch": 0.8257547970615051, "grad_norm": 2.3089535236358643, "learning_rate": 7.750825388531452e-07, "loss": 0.4703, "step": 19699 }, { "epoch": 0.825796715676514, "grad_norm": 3.111124277114868, "learning_rate": 7.747195327545088e-07, "loss": 0.4526, "step": 19700 }, { "epoch": 0.825838634291523, "grad_norm": 2.0953285694122314, "learning_rate": 7.74356604543075e-07, "loss": 0.5453, "step": 19701 }, { "epoch": 0.825880552906532, "grad_norm": 1.8125957250595093, "learning_rate": 7.739937542255355e-07, "loss": 0.4445, "step": 19702 }, { "epoch": 0.825922471521541, "grad_norm": 1.6615554094314575, "learning_rate": 7.736309818085802e-07, "loss": 0.4716, "step": 19703 }, { "epoch": 0.8259643901365499, "grad_norm": 1.9030214548110962, "learning_rate": 7.732682872988951e-07, "loss": 0.4281, "step": 19704 }, { "epoch": 0.8260063087515589, "grad_norm": 3.717804193496704, "learning_rate": 7.72905670703164e-07, "loss": 0.4805, "step": 19705 }, { "epoch": 0.8260482273665678, "grad_norm": 3.8900229930877686, "learning_rate": 7.725431320280752e-07, "loss": 0.4529, "step": 19706 }, { "epoch": 0.8260901459815768, "grad_norm": 2.55326247215271, "learning_rate": 7.721806712803076e-07, "loss": 0.4765, "step": 19707 }, { "epoch": 0.8261320645965857, "grad_norm": 3.247833013534546, "learning_rate": 7.718182884665448e-07, "loss": 0.4414, "step": 19708 }, { "epoch": 0.8261739832115947, "grad_norm": 2.5565531253814697, "learning_rate": 7.714559835934676e-07, "loss": 0.4353, "step": 19709 }, { "epoch": 0.8262159018266036, "grad_norm": 1.999268889427185, "learning_rate": 7.710937566677528e-07, "loss": 0.507, "step": 19710 }, { "epoch": 0.8262578204416126, "grad_norm": 3.49883770942688, "learning_rate": 7.70731607696078e-07, "loss": 0.4193, "step": 19711 }, { "epoch": 0.8262997390566216, "grad_norm": 1.9526393413543701, "learning_rate": 7.703695366851205e-07, "loss": 0.4357, "step": 19712 }, { "epoch": 0.8263416576716305, "grad_norm": 2.214252233505249, "learning_rate": 7.700075436415516e-07, "loss": 0.5298, "step": 19713 }, { "epoch": 0.8263835762866395, "grad_norm": 2.132040500640869, "learning_rate": 7.696456285720478e-07, "loss": 0.5097, "step": 19714 }, { "epoch": 0.8264254949016484, "grad_norm": 2.3652241230010986, "learning_rate": 7.692837914832779e-07, "loss": 0.4763, "step": 19715 }, { "epoch": 0.8264674135166574, "grad_norm": 2.1740317344665527, "learning_rate": 7.689220323819118e-07, "loss": 0.4608, "step": 19716 }, { "epoch": 0.8265093321316663, "grad_norm": 2.082382917404175, "learning_rate": 7.685603512746187e-07, "loss": 0.4617, "step": 19717 }, { "epoch": 0.8265512507466753, "grad_norm": 1.8321936130523682, "learning_rate": 7.681987481680669e-07, "loss": 0.4667, "step": 19718 }, { "epoch": 0.8265931693616843, "grad_norm": 3.183361530303955, "learning_rate": 7.678372230689196e-07, "loss": 0.4702, "step": 19719 }, { "epoch": 0.8266350879766933, "grad_norm": 1.8674627542495728, "learning_rate": 7.674757759838425e-07, "loss": 0.5201, "step": 19720 }, { "epoch": 0.8266770065917022, "grad_norm": 2.0058133602142334, "learning_rate": 7.671144069194991e-07, "loss": 0.487, "step": 19721 }, { "epoch": 0.8267189252067112, "grad_norm": 1.859818935394287, "learning_rate": 7.667531158825486e-07, "loss": 0.4769, "step": 19722 }, { "epoch": 0.8267608438217201, "grad_norm": 2.994260787963867, "learning_rate": 7.663919028796535e-07, "loss": 0.5303, "step": 19723 }, { "epoch": 0.8268027624367291, "grad_norm": 3.1584975719451904, "learning_rate": 7.660307679174706e-07, "loss": 0.5256, "step": 19724 }, { "epoch": 0.826844681051738, "grad_norm": 1.5982959270477295, "learning_rate": 7.656697110026561e-07, "loss": 0.435, "step": 19725 }, { "epoch": 0.826886599666747, "grad_norm": 1.5697154998779297, "learning_rate": 7.653087321418673e-07, "loss": 0.375, "step": 19726 }, { "epoch": 0.826928518281756, "grad_norm": 1.944961667060852, "learning_rate": 7.649478313417579e-07, "loss": 0.4321, "step": 19727 }, { "epoch": 0.826970436896765, "grad_norm": 2.3608767986297607, "learning_rate": 7.645870086089802e-07, "loss": 0.5048, "step": 19728 }, { "epoch": 0.8270123555117739, "grad_norm": 1.8380324840545654, "learning_rate": 7.642262639501852e-07, "loss": 0.5024, "step": 19729 }, { "epoch": 0.8270542741267829, "grad_norm": 1.6755295991897583, "learning_rate": 7.638655973720243e-07, "loss": 0.4986, "step": 19730 }, { "epoch": 0.8270961927417918, "grad_norm": 2.707479238510132, "learning_rate": 7.635050088811441e-07, "loss": 0.5074, "step": 19731 }, { "epoch": 0.8271381113568008, "grad_norm": 1.7894796133041382, "learning_rate": 7.631444984841935e-07, "loss": 0.4214, "step": 19732 }, { "epoch": 0.8271800299718097, "grad_norm": 1.664589285850525, "learning_rate": 7.62784066187816e-07, "loss": 0.521, "step": 19733 }, { "epoch": 0.8272219485868187, "grad_norm": 2.0938007831573486, "learning_rate": 7.62423711998656e-07, "loss": 0.5268, "step": 19734 }, { "epoch": 0.8272638672018277, "grad_norm": 1.9959986209869385, "learning_rate": 7.620634359233564e-07, "loss": 0.4405, "step": 19735 }, { "epoch": 0.8273057858168366, "grad_norm": 2.5031166076660156, "learning_rate": 7.617032379685602e-07, "loss": 0.4521, "step": 19736 }, { "epoch": 0.8273477044318456, "grad_norm": 2.201284408569336, "learning_rate": 7.613431181409037e-07, "loss": 0.4942, "step": 19737 }, { "epoch": 0.8273896230468545, "grad_norm": 1.9016177654266357, "learning_rate": 7.609830764470272e-07, "loss": 0.4953, "step": 19738 }, { "epoch": 0.8274315416618635, "grad_norm": 3.155712842941284, "learning_rate": 7.606231128935687e-07, "loss": 0.4305, "step": 19739 }, { "epoch": 0.8274734602768724, "grad_norm": 1.940287709236145, "learning_rate": 7.602632274871613e-07, "loss": 0.4997, "step": 19740 }, { "epoch": 0.8275153788918814, "grad_norm": 1.8758820295333862, "learning_rate": 7.599034202344413e-07, "loss": 0.5037, "step": 19741 }, { "epoch": 0.8275572975068903, "grad_norm": 2.616590738296509, "learning_rate": 7.595436911420396e-07, "loss": 0.4745, "step": 19742 }, { "epoch": 0.8275992161218994, "grad_norm": 1.9965269565582275, "learning_rate": 7.591840402165867e-07, "loss": 0.459, "step": 19743 }, { "epoch": 0.8276411347369083, "grad_norm": 1.701290488243103, "learning_rate": 7.588244674647133e-07, "loss": 0.4469, "step": 19744 }, { "epoch": 0.8276830533519173, "grad_norm": 3.8836777210235596, "learning_rate": 7.584649728930488e-07, "loss": 0.5084, "step": 19745 }, { "epoch": 0.8277249719669262, "grad_norm": 1.919998288154602, "learning_rate": 7.581055565082174e-07, "loss": 0.4833, "step": 19746 }, { "epoch": 0.8277668905819352, "grad_norm": 2.0776844024658203, "learning_rate": 7.57746218316846e-07, "loss": 0.4118, "step": 19747 }, { "epoch": 0.8278088091969441, "grad_norm": 1.9996497631072998, "learning_rate": 7.573869583255594e-07, "loss": 0.4814, "step": 19748 }, { "epoch": 0.8278507278119531, "grad_norm": 2.108229160308838, "learning_rate": 7.570277765409778e-07, "loss": 0.5486, "step": 19749 }, { "epoch": 0.827892646426962, "grad_norm": 6.176340579986572, "learning_rate": 7.566686729697248e-07, "loss": 0.4574, "step": 19750 }, { "epoch": 0.827934565041971, "grad_norm": 1.5224905014038086, "learning_rate": 7.563096476184185e-07, "loss": 0.4598, "step": 19751 }, { "epoch": 0.82797648365698, "grad_norm": 1.6544991731643677, "learning_rate": 7.559507004936761e-07, "loss": 0.4541, "step": 19752 }, { "epoch": 0.828018402271989, "grad_norm": 1.809086799621582, "learning_rate": 7.555918316021149e-07, "loss": 0.4716, "step": 19753 }, { "epoch": 0.8280603208869979, "grad_norm": 1.895955204963684, "learning_rate": 7.552330409503522e-07, "loss": 0.5077, "step": 19754 }, { "epoch": 0.8281022395020069, "grad_norm": 2.045137643814087, "learning_rate": 7.548743285449983e-07, "loss": 0.5006, "step": 19755 }, { "epoch": 0.8281441581170158, "grad_norm": 1.8269052505493164, "learning_rate": 7.54515694392668e-07, "loss": 0.5051, "step": 19756 }, { "epoch": 0.8281860767320248, "grad_norm": 2.2015106678009033, "learning_rate": 7.541571384999729e-07, "loss": 0.4811, "step": 19757 }, { "epoch": 0.8282279953470337, "grad_norm": 2.273501396179199, "learning_rate": 7.537986608735199e-07, "loss": 0.4757, "step": 19758 }, { "epoch": 0.8282699139620427, "grad_norm": 1.9591456651687622, "learning_rate": 7.534402615199193e-07, "loss": 0.4943, "step": 19759 }, { "epoch": 0.8283118325770517, "grad_norm": 1.8392176628112793, "learning_rate": 7.530819404457762e-07, "loss": 0.4608, "step": 19760 }, { "epoch": 0.8283537511920606, "grad_norm": 2.562345027923584, "learning_rate": 7.527236976576974e-07, "loss": 0.482, "step": 19761 }, { "epoch": 0.8283956698070696, "grad_norm": 1.6317577362060547, "learning_rate": 7.523655331622837e-07, "loss": 0.4778, "step": 19762 }, { "epoch": 0.8284375884220785, "grad_norm": 1.7843128442764282, "learning_rate": 7.520074469661409e-07, "loss": 0.434, "step": 19763 }, { "epoch": 0.8284795070370875, "grad_norm": 2.448887348175049, "learning_rate": 7.516494390758672e-07, "loss": 0.5371, "step": 19764 }, { "epoch": 0.8285214256520964, "grad_norm": 1.842946171760559, "learning_rate": 7.512915094980627e-07, "loss": 0.5107, "step": 19765 }, { "epoch": 0.8285633442671054, "grad_norm": 1.6493149995803833, "learning_rate": 7.509336582393267e-07, "loss": 0.4809, "step": 19766 }, { "epoch": 0.8286052628821143, "grad_norm": 1.746689796447754, "learning_rate": 7.505758853062534e-07, "loss": 0.4723, "step": 19767 }, { "epoch": 0.8286471814971234, "grad_norm": 2.288350820541382, "learning_rate": 7.50218190705439e-07, "loss": 0.4804, "step": 19768 }, { "epoch": 0.8286891001121323, "grad_norm": 1.912717580795288, "learning_rate": 7.498605744434778e-07, "loss": 0.4436, "step": 19769 }, { "epoch": 0.8287310187271413, "grad_norm": 1.9334716796875, "learning_rate": 7.495030365269617e-07, "loss": 0.4404, "step": 19770 }, { "epoch": 0.8287729373421502, "grad_norm": 1.8066673278808594, "learning_rate": 7.4914557696248e-07, "loss": 0.4625, "step": 19771 }, { "epoch": 0.8288148559571592, "grad_norm": 1.7720404863357544, "learning_rate": 7.48788195756624e-07, "loss": 0.478, "step": 19772 }, { "epoch": 0.8288567745721681, "grad_norm": 1.8444699048995972, "learning_rate": 7.484308929159795e-07, "loss": 0.511, "step": 19773 }, { "epoch": 0.8288986931871771, "grad_norm": 2.1421966552734375, "learning_rate": 7.480736684471346e-07, "loss": 0.4562, "step": 19774 }, { "epoch": 0.828940611802186, "grad_norm": 1.9634398221969604, "learning_rate": 7.477165223566729e-07, "loss": 0.4801, "step": 19775 }, { "epoch": 0.828982530417195, "grad_norm": 1.9966813325881958, "learning_rate": 7.473594546511781e-07, "loss": 0.4557, "step": 19776 }, { "epoch": 0.829024449032204, "grad_norm": 1.6576871871948242, "learning_rate": 7.470024653372337e-07, "loss": 0.4436, "step": 19777 }, { "epoch": 0.829066367647213, "grad_norm": 1.7884132862091064, "learning_rate": 7.466455544214185e-07, "loss": 0.4848, "step": 19778 }, { "epoch": 0.8291082862622219, "grad_norm": 1.8759584426879883, "learning_rate": 7.462887219103132e-07, "loss": 0.4416, "step": 19779 }, { "epoch": 0.8291502048772309, "grad_norm": 2.1897034645080566, "learning_rate": 7.459319678104949e-07, "loss": 0.4709, "step": 19780 }, { "epoch": 0.8291921234922398, "grad_norm": 1.853955626487732, "learning_rate": 7.455752921285386e-07, "loss": 0.4781, "step": 19781 }, { "epoch": 0.8292340421072488, "grad_norm": 2.143069267272949, "learning_rate": 7.4521869487102e-07, "loss": 0.4677, "step": 19782 }, { "epoch": 0.8292759607222577, "grad_norm": 2.2990362644195557, "learning_rate": 7.448621760445141e-07, "loss": 0.4658, "step": 19783 }, { "epoch": 0.8293178793372668, "grad_norm": 1.5974092483520508, "learning_rate": 7.445057356555901e-07, "loss": 0.4536, "step": 19784 }, { "epoch": 0.8293597979522757, "grad_norm": 1.547823190689087, "learning_rate": 7.441493737108202e-07, "loss": 0.4527, "step": 19785 }, { "epoch": 0.8294017165672846, "grad_norm": 1.7402268648147583, "learning_rate": 7.437930902167739e-07, "loss": 0.4582, "step": 19786 }, { "epoch": 0.8294436351822936, "grad_norm": 1.8725149631500244, "learning_rate": 7.434368851800167e-07, "loss": 0.4927, "step": 19787 }, { "epoch": 0.8294855537973025, "grad_norm": 4.874673366546631, "learning_rate": 7.43080758607117e-07, "loss": 0.531, "step": 19788 }, { "epoch": 0.8295274724123115, "grad_norm": 2.4401493072509766, "learning_rate": 7.427247105046392e-07, "loss": 0.5249, "step": 19789 }, { "epoch": 0.8295693910273204, "grad_norm": 2.5099523067474365, "learning_rate": 7.423687408791442e-07, "loss": 0.4826, "step": 19790 }, { "epoch": 0.8296113096423294, "grad_norm": 1.9559310674667358, "learning_rate": 7.420128497371953e-07, "loss": 0.4539, "step": 19791 }, { "epoch": 0.8296532282573383, "grad_norm": 2.1805951595306396, "learning_rate": 7.416570370853548e-07, "loss": 0.4848, "step": 19792 }, { "epoch": 0.8296951468723474, "grad_norm": 2.0238378047943115, "learning_rate": 7.41301302930178e-07, "loss": 0.4777, "step": 19793 }, { "epoch": 0.8297370654873563, "grad_norm": 2.766493082046509, "learning_rate": 7.409456472782244e-07, "loss": 0.5325, "step": 19794 }, { "epoch": 0.8297789841023653, "grad_norm": 1.8051048517227173, "learning_rate": 7.405900701360513e-07, "loss": 0.4886, "step": 19795 }, { "epoch": 0.8298209027173742, "grad_norm": 1.7009520530700684, "learning_rate": 7.402345715102105e-07, "loss": 0.5223, "step": 19796 }, { "epoch": 0.8298628213323832, "grad_norm": 2.5400068759918213, "learning_rate": 7.398791514072573e-07, "loss": 0.4819, "step": 19797 }, { "epoch": 0.8299047399473921, "grad_norm": 2.006666421890259, "learning_rate": 7.395238098337415e-07, "loss": 0.4803, "step": 19798 }, { "epoch": 0.8299466585624011, "grad_norm": 1.7977066040039062, "learning_rate": 7.391685467962156e-07, "loss": 0.4588, "step": 19799 }, { "epoch": 0.82998857717741, "grad_norm": 2.0712971687316895, "learning_rate": 7.388133623012261e-07, "loss": 0.4868, "step": 19800 }, { "epoch": 0.830030495792419, "grad_norm": 2.012972593307495, "learning_rate": 7.384582563553222e-07, "loss": 0.4319, "step": 19801 }, { "epoch": 0.830072414407428, "grad_norm": 1.5892422199249268, "learning_rate": 7.381032289650481e-07, "loss": 0.4401, "step": 19802 }, { "epoch": 0.830114333022437, "grad_norm": 2.6135201454162598, "learning_rate": 7.37748280136949e-07, "loss": 0.4421, "step": 19803 }, { "epoch": 0.8301562516374459, "grad_norm": 1.5539360046386719, "learning_rate": 7.373934098775686e-07, "loss": 0.499, "step": 19804 }, { "epoch": 0.8301981702524549, "grad_norm": 2.583106517791748, "learning_rate": 7.370386181934475e-07, "loss": 0.4579, "step": 19805 }, { "epoch": 0.8302400888674638, "grad_norm": 1.6025978326797485, "learning_rate": 7.366839050911268e-07, "loss": 0.4373, "step": 19806 }, { "epoch": 0.8302820074824728, "grad_norm": 1.6483333110809326, "learning_rate": 7.36329270577143e-07, "loss": 0.4857, "step": 19807 }, { "epoch": 0.8303239260974817, "grad_norm": 2.374803066253662, "learning_rate": 7.359747146580365e-07, "loss": 0.462, "step": 19808 }, { "epoch": 0.8303658447124908, "grad_norm": 3.0544707775115967, "learning_rate": 7.356202373403398e-07, "loss": 0.4615, "step": 19809 }, { "epoch": 0.8304077633274997, "grad_norm": 1.7565112113952637, "learning_rate": 7.3526583863059e-07, "loss": 0.4824, "step": 19810 }, { "epoch": 0.8304496819425086, "grad_norm": 1.9099382162094116, "learning_rate": 7.349115185353173e-07, "loss": 0.5121, "step": 19811 }, { "epoch": 0.8304916005575176, "grad_norm": 1.6991841793060303, "learning_rate": 7.345572770610548e-07, "loss": 0.4582, "step": 19812 }, { "epoch": 0.8305335191725265, "grad_norm": 1.8590909242630005, "learning_rate": 7.34203114214333e-07, "loss": 0.4931, "step": 19813 }, { "epoch": 0.8305754377875355, "grad_norm": 2.048046588897705, "learning_rate": 7.338490300016781e-07, "loss": 0.5318, "step": 19814 }, { "epoch": 0.8306173564025444, "grad_norm": 1.8735318183898926, "learning_rate": 7.33495024429619e-07, "loss": 0.5235, "step": 19815 }, { "epoch": 0.8306592750175534, "grad_norm": 2.248386859893799, "learning_rate": 7.33141097504682e-07, "loss": 0.4725, "step": 19816 }, { "epoch": 0.8307011936325623, "grad_norm": 2.67100191116333, "learning_rate": 7.327872492333899e-07, "loss": 0.541, "step": 19817 }, { "epoch": 0.8307431122475714, "grad_norm": 2.794682264328003, "learning_rate": 7.324334796222643e-07, "loss": 0.4884, "step": 19818 }, { "epoch": 0.8307850308625803, "grad_norm": 2.186131477355957, "learning_rate": 7.320797886778291e-07, "loss": 0.5232, "step": 19819 }, { "epoch": 0.8308269494775893, "grad_norm": 2.0784454345703125, "learning_rate": 7.31726176406602e-07, "loss": 0.4672, "step": 19820 }, { "epoch": 0.8308688680925982, "grad_norm": 1.771507740020752, "learning_rate": 7.313726428151014e-07, "loss": 0.4631, "step": 19821 }, { "epoch": 0.8309107867076072, "grad_norm": 2.149435043334961, "learning_rate": 7.310191879098466e-07, "loss": 0.5188, "step": 19822 }, { "epoch": 0.8309527053226161, "grad_norm": 1.7882606983184814, "learning_rate": 7.306658116973503e-07, "loss": 0.4655, "step": 19823 }, { "epoch": 0.8309946239376251, "grad_norm": 1.7440111637115479, "learning_rate": 7.303125141841272e-07, "loss": 0.4547, "step": 19824 }, { "epoch": 0.831036542552634, "grad_norm": 1.9730064868927002, "learning_rate": 7.299592953766915e-07, "loss": 0.4716, "step": 19825 }, { "epoch": 0.8310784611676431, "grad_norm": 1.7357243299484253, "learning_rate": 7.296061552815531e-07, "loss": 0.4723, "step": 19826 }, { "epoch": 0.831120379782652, "grad_norm": 2.0104331970214844, "learning_rate": 7.292530939052206e-07, "loss": 0.5341, "step": 19827 }, { "epoch": 0.831162298397661, "grad_norm": 1.8858124017715454, "learning_rate": 7.289001112542043e-07, "loss": 0.5188, "step": 19828 }, { "epoch": 0.8312042170126699, "grad_norm": 1.9993562698364258, "learning_rate": 7.285472073350087e-07, "loss": 0.4909, "step": 19829 }, { "epoch": 0.8312461356276789, "grad_norm": 2.724785089492798, "learning_rate": 7.281943821541398e-07, "loss": 0.5301, "step": 19830 }, { "epoch": 0.8312880542426878, "grad_norm": 1.7969560623168945, "learning_rate": 7.278416357181034e-07, "loss": 0.4438, "step": 19831 }, { "epoch": 0.8313299728576968, "grad_norm": 2.466167688369751, "learning_rate": 7.274889680333991e-07, "loss": 0.5281, "step": 19832 }, { "epoch": 0.8313718914727057, "grad_norm": 1.876574158668518, "learning_rate": 7.271363791065289e-07, "loss": 0.4153, "step": 19833 }, { "epoch": 0.8314138100877148, "grad_norm": 1.992518663406372, "learning_rate": 7.267838689439938e-07, "loss": 0.5139, "step": 19834 }, { "epoch": 0.8314557287027237, "grad_norm": 1.9072853326797485, "learning_rate": 7.264314375522907e-07, "loss": 0.4642, "step": 19835 }, { "epoch": 0.8314976473177326, "grad_norm": 2.0832066535949707, "learning_rate": 7.260790849379146e-07, "loss": 0.4777, "step": 19836 }, { "epoch": 0.8315395659327416, "grad_norm": 1.5514256954193115, "learning_rate": 7.257268111073634e-07, "loss": 0.4579, "step": 19837 }, { "epoch": 0.8315814845477505, "grad_norm": 8.883255958557129, "learning_rate": 7.253746160671282e-07, "loss": 0.5069, "step": 19838 }, { "epoch": 0.8316234031627595, "grad_norm": 2.0297489166259766, "learning_rate": 7.250224998237026e-07, "loss": 0.4485, "step": 19839 }, { "epoch": 0.8316653217777684, "grad_norm": 1.8093904256820679, "learning_rate": 7.246704623835782e-07, "loss": 0.4595, "step": 19840 }, { "epoch": 0.8317072403927774, "grad_norm": 2.323556661605835, "learning_rate": 7.243185037532425e-07, "loss": 0.4175, "step": 19841 }, { "epoch": 0.8317491590077863, "grad_norm": 1.862260103225708, "learning_rate": 7.23966623939184e-07, "loss": 0.4816, "step": 19842 }, { "epoch": 0.8317910776227954, "grad_norm": 1.674770474433899, "learning_rate": 7.236148229478907e-07, "loss": 0.4651, "step": 19843 }, { "epoch": 0.8318329962378043, "grad_norm": 2.175837993621826, "learning_rate": 7.232631007858459e-07, "loss": 0.4825, "step": 19844 }, { "epoch": 0.8318749148528133, "grad_norm": 2.194258689880371, "learning_rate": 7.229114574595325e-07, "loss": 0.4628, "step": 19845 }, { "epoch": 0.8319168334678222, "grad_norm": 1.7435193061828613, "learning_rate": 7.225598929754346e-07, "loss": 0.5076, "step": 19846 }, { "epoch": 0.8319587520828312, "grad_norm": 1.581869125366211, "learning_rate": 7.222084073400304e-07, "loss": 0.4831, "step": 19847 }, { "epoch": 0.8320006706978401, "grad_norm": 2.7472047805786133, "learning_rate": 7.218570005598003e-07, "loss": 0.4773, "step": 19848 }, { "epoch": 0.8320425893128491, "grad_norm": 1.739780306816101, "learning_rate": 7.215056726412233e-07, "loss": 0.4869, "step": 19849 }, { "epoch": 0.832084507927858, "grad_norm": 1.6609617471694946, "learning_rate": 7.211544235907736e-07, "loss": 0.4329, "step": 19850 }, { "epoch": 0.8321264265428671, "grad_norm": 1.6824387311935425, "learning_rate": 7.208032534149273e-07, "loss": 0.5182, "step": 19851 }, { "epoch": 0.832168345157876, "grad_norm": 2.0955917835235596, "learning_rate": 7.204521621201565e-07, "loss": 0.4764, "step": 19852 }, { "epoch": 0.832210263772885, "grad_norm": 1.973292589187622, "learning_rate": 7.201011497129335e-07, "loss": 0.5006, "step": 19853 }, { "epoch": 0.8322521823878939, "grad_norm": 2.52423095703125, "learning_rate": 7.197502161997305e-07, "loss": 0.4948, "step": 19854 }, { "epoch": 0.8322941010029029, "grad_norm": 2.3150360584259033, "learning_rate": 7.193993615870149e-07, "loss": 0.4861, "step": 19855 }, { "epoch": 0.8323360196179118, "grad_norm": 2.278684377670288, "learning_rate": 7.190485858812529e-07, "loss": 0.4673, "step": 19856 }, { "epoch": 0.8323779382329208, "grad_norm": 2.1218159198760986, "learning_rate": 7.186978890889134e-07, "loss": 0.5166, "step": 19857 }, { "epoch": 0.8324198568479297, "grad_norm": 1.979129433631897, "learning_rate": 7.18347271216458e-07, "loss": 0.4932, "step": 19858 }, { "epoch": 0.8324617754629388, "grad_norm": 1.8618701696395874, "learning_rate": 7.179967322703518e-07, "loss": 0.4883, "step": 19859 }, { "epoch": 0.8325036940779477, "grad_norm": 2.0458297729492188, "learning_rate": 7.176462722570576e-07, "loss": 0.4977, "step": 19860 }, { "epoch": 0.8325456126929566, "grad_norm": 1.931450605392456, "learning_rate": 7.172958911830324e-07, "loss": 0.5264, "step": 19861 }, { "epoch": 0.8325875313079656, "grad_norm": 2.250474691390991, "learning_rate": 7.169455890547372e-07, "loss": 0.4944, "step": 19862 }, { "epoch": 0.8326294499229745, "grad_norm": 1.7410869598388672, "learning_rate": 7.165953658786301e-07, "loss": 0.4542, "step": 19863 }, { "epoch": 0.8326713685379835, "grad_norm": 1.5042238235473633, "learning_rate": 7.162452216611659e-07, "loss": 0.4441, "step": 19864 }, { "epoch": 0.8327132871529924, "grad_norm": 1.9836549758911133, "learning_rate": 7.158951564087974e-07, "loss": 0.5075, "step": 19865 }, { "epoch": 0.8327552057680014, "grad_norm": 1.8602430820465088, "learning_rate": 7.155451701279804e-07, "loss": 0.4938, "step": 19866 }, { "epoch": 0.8327971243830103, "grad_norm": 3.7809274196624756, "learning_rate": 7.151952628251641e-07, "loss": 0.4057, "step": 19867 }, { "epoch": 0.8328390429980194, "grad_norm": 1.7453280687332153, "learning_rate": 7.148454345067996e-07, "loss": 0.4584, "step": 19868 }, { "epoch": 0.8328809616130283, "grad_norm": 1.715083122253418, "learning_rate": 7.144956851793361e-07, "loss": 0.4838, "step": 19869 }, { "epoch": 0.8329228802280373, "grad_norm": 1.8397451639175415, "learning_rate": 7.14146014849219e-07, "loss": 0.4418, "step": 19870 }, { "epoch": 0.8329647988430462, "grad_norm": 1.898840069770813, "learning_rate": 7.137964235228956e-07, "loss": 0.4695, "step": 19871 }, { "epoch": 0.8330067174580552, "grad_norm": 2.4299726486206055, "learning_rate": 7.13446911206811e-07, "loss": 0.5093, "step": 19872 }, { "epoch": 0.8330486360730641, "grad_norm": 2.0728559494018555, "learning_rate": 7.130974779074062e-07, "loss": 0.4914, "step": 19873 }, { "epoch": 0.8330905546880731, "grad_norm": 2.413752317428589, "learning_rate": 7.127481236311218e-07, "loss": 0.4604, "step": 19874 }, { "epoch": 0.833132473303082, "grad_norm": 1.7864943742752075, "learning_rate": 7.123988483843996e-07, "loss": 0.4973, "step": 19875 }, { "epoch": 0.8331743919180911, "grad_norm": 1.6793129444122314, "learning_rate": 7.120496521736764e-07, "loss": 0.4659, "step": 19876 }, { "epoch": 0.8332163105331, "grad_norm": 2.030956268310547, "learning_rate": 7.117005350053901e-07, "loss": 0.4991, "step": 19877 }, { "epoch": 0.833258229148109, "grad_norm": 1.9466567039489746, "learning_rate": 7.11351496885977e-07, "loss": 0.5282, "step": 19878 }, { "epoch": 0.8333001477631179, "grad_norm": 1.879141092300415, "learning_rate": 7.110025378218688e-07, "loss": 0.4856, "step": 19879 }, { "epoch": 0.8333420663781269, "grad_norm": 1.985688328742981, "learning_rate": 7.106536578195e-07, "loss": 0.4297, "step": 19880 }, { "epoch": 0.8333839849931358, "grad_norm": 1.7389096021652222, "learning_rate": 7.103048568853021e-07, "loss": 0.4509, "step": 19881 }, { "epoch": 0.8334259036081448, "grad_norm": 1.8284618854522705, "learning_rate": 7.099561350257034e-07, "loss": 0.5004, "step": 19882 }, { "epoch": 0.8334678222231537, "grad_norm": 2.0046119689941406, "learning_rate": 7.096074922471318e-07, "loss": 0.5314, "step": 19883 }, { "epoch": 0.8335097408381628, "grad_norm": 1.6066936254501343, "learning_rate": 7.092589285560158e-07, "loss": 0.4879, "step": 19884 }, { "epoch": 0.8335516594531717, "grad_norm": 2.1018259525299072, "learning_rate": 7.08910443958779e-07, "loss": 0.4864, "step": 19885 }, { "epoch": 0.8335935780681806, "grad_norm": 1.7027530670166016, "learning_rate": 7.085620384618453e-07, "loss": 0.4969, "step": 19886 }, { "epoch": 0.8336354966831896, "grad_norm": 2.244468927383423, "learning_rate": 7.082137120716392e-07, "loss": 0.4959, "step": 19887 }, { "epoch": 0.8336774152981985, "grad_norm": 2.1394906044006348, "learning_rate": 7.078654647945787e-07, "loss": 0.452, "step": 19888 }, { "epoch": 0.8337193339132075, "grad_norm": 1.6329541206359863, "learning_rate": 7.075172966370847e-07, "loss": 0.4464, "step": 19889 }, { "epoch": 0.8337612525282164, "grad_norm": 2.056729793548584, "learning_rate": 7.071692076055763e-07, "loss": 0.4657, "step": 19890 }, { "epoch": 0.8338031711432254, "grad_norm": 1.941598892211914, "learning_rate": 7.068211977064693e-07, "loss": 0.4288, "step": 19891 }, { "epoch": 0.8338450897582343, "grad_norm": 2.962061882019043, "learning_rate": 7.064732669461771e-07, "loss": 0.426, "step": 19892 }, { "epoch": 0.8338870083732434, "grad_norm": 3.3125011920928955, "learning_rate": 7.061254153311153e-07, "loss": 0.5256, "step": 19893 }, { "epoch": 0.8339289269882523, "grad_norm": 1.8185997009277344, "learning_rate": 7.057776428676944e-07, "loss": 0.4795, "step": 19894 }, { "epoch": 0.8339708456032613, "grad_norm": 1.7342517375946045, "learning_rate": 7.054299495623263e-07, "loss": 0.4535, "step": 19895 }, { "epoch": 0.8340127642182702, "grad_norm": 2.3322925567626953, "learning_rate": 7.050823354214209e-07, "loss": 0.4682, "step": 19896 }, { "epoch": 0.8340546828332792, "grad_norm": 2.037640333175659, "learning_rate": 7.047348004513843e-07, "loss": 0.4967, "step": 19897 }, { "epoch": 0.8340966014482881, "grad_norm": 1.727847695350647, "learning_rate": 7.043873446586236e-07, "loss": 0.4961, "step": 19898 }, { "epoch": 0.8341385200632971, "grad_norm": 2.1567628383636475, "learning_rate": 7.040399680495441e-07, "loss": 0.5185, "step": 19899 }, { "epoch": 0.834180438678306, "grad_norm": 1.6985459327697754, "learning_rate": 7.036926706305486e-07, "loss": 0.4431, "step": 19900 }, { "epoch": 0.8342223572933151, "grad_norm": 2.1872782707214355, "learning_rate": 7.033454524080396e-07, "loss": 0.4307, "step": 19901 }, { "epoch": 0.834264275908324, "grad_norm": 1.7292202711105347, "learning_rate": 7.029983133884178e-07, "loss": 0.5196, "step": 19902 }, { "epoch": 0.834306194523333, "grad_norm": 2.8328511714935303, "learning_rate": 7.026512535780805e-07, "loss": 0.4862, "step": 19903 }, { "epoch": 0.8343481131383419, "grad_norm": 1.8633793592453003, "learning_rate": 7.023042729834262e-07, "loss": 0.4013, "step": 19904 }, { "epoch": 0.8343900317533509, "grad_norm": 1.824820637702942, "learning_rate": 7.019573716108524e-07, "loss": 0.5031, "step": 19905 }, { "epoch": 0.8344319503683598, "grad_norm": 2.2370340824127197, "learning_rate": 7.016105494667513e-07, "loss": 0.4368, "step": 19906 }, { "epoch": 0.8344738689833688, "grad_norm": 1.9589272737503052, "learning_rate": 7.012638065575178e-07, "loss": 0.4614, "step": 19907 }, { "epoch": 0.8345157875983777, "grad_norm": 2.063941717147827, "learning_rate": 7.009171428895445e-07, "loss": 0.4844, "step": 19908 }, { "epoch": 0.8345577062133868, "grad_norm": 1.6552993059158325, "learning_rate": 7.005705584692185e-07, "loss": 0.4161, "step": 19909 }, { "epoch": 0.8345996248283957, "grad_norm": 1.9383471012115479, "learning_rate": 7.00224053302932e-07, "loss": 0.4824, "step": 19910 }, { "epoch": 0.8346415434434046, "grad_norm": 2.686471462249756, "learning_rate": 6.998776273970708e-07, "loss": 0.4504, "step": 19911 }, { "epoch": 0.8346834620584136, "grad_norm": 2.5409293174743652, "learning_rate": 6.995312807580195e-07, "loss": 0.4831, "step": 19912 }, { "epoch": 0.8347253806734225, "grad_norm": 1.8339107036590576, "learning_rate": 6.99185013392164e-07, "loss": 0.4391, "step": 19913 }, { "epoch": 0.8347672992884315, "grad_norm": 1.7025381326675415, "learning_rate": 6.988388253058887e-07, "loss": 0.4366, "step": 19914 }, { "epoch": 0.8348092179034404, "grad_norm": 2.2606213092803955, "learning_rate": 6.984927165055716e-07, "loss": 0.5132, "step": 19915 }, { "epoch": 0.8348511365184494, "grad_norm": 1.9495042562484741, "learning_rate": 6.98146686997595e-07, "loss": 0.4724, "step": 19916 }, { "epoch": 0.8348930551334584, "grad_norm": 1.6575963497161865, "learning_rate": 6.978007367883377e-07, "loss": 0.4959, "step": 19917 }, { "epoch": 0.8349349737484674, "grad_norm": 3.064236879348755, "learning_rate": 6.974548658841756e-07, "loss": 0.4965, "step": 19918 }, { "epoch": 0.8349768923634763, "grad_norm": 1.8970202207565308, "learning_rate": 6.97109074291486e-07, "loss": 0.4485, "step": 19919 }, { "epoch": 0.8350188109784853, "grad_norm": 2.1677017211914062, "learning_rate": 6.96763362016642e-07, "loss": 0.5062, "step": 19920 }, { "epoch": 0.8350607295934942, "grad_norm": 2.1757872104644775, "learning_rate": 6.964177290660152e-07, "loss": 0.4435, "step": 19921 }, { "epoch": 0.8351026482085032, "grad_norm": 2.044790744781494, "learning_rate": 6.960721754459782e-07, "loss": 0.5112, "step": 19922 }, { "epoch": 0.8351445668235121, "grad_norm": 1.8222320079803467, "learning_rate": 6.957267011629015e-07, "loss": 0.4969, "step": 19923 }, { "epoch": 0.8351864854385211, "grad_norm": 2.976396083831787, "learning_rate": 6.953813062231512e-07, "loss": 0.4452, "step": 19924 }, { "epoch": 0.83522840405353, "grad_norm": 1.6235496997833252, "learning_rate": 6.950359906330972e-07, "loss": 0.4566, "step": 19925 }, { "epoch": 0.8352703226685391, "grad_norm": 1.9640675783157349, "learning_rate": 6.946907543991021e-07, "loss": 0.4731, "step": 19926 }, { "epoch": 0.835312241283548, "grad_norm": 1.7550972700119019, "learning_rate": 6.943455975275309e-07, "loss": 0.4997, "step": 19927 }, { "epoch": 0.835354159898557, "grad_norm": 1.6552002429962158, "learning_rate": 6.940005200247474e-07, "loss": 0.469, "step": 19928 }, { "epoch": 0.8353960785135659, "grad_norm": 1.6280912160873413, "learning_rate": 6.936555218971119e-07, "loss": 0.4754, "step": 19929 }, { "epoch": 0.8354379971285749, "grad_norm": 1.8740299940109253, "learning_rate": 6.933106031509817e-07, "loss": 0.4874, "step": 19930 }, { "epoch": 0.8354799157435838, "grad_norm": 1.862906575202942, "learning_rate": 6.929657637927179e-07, "loss": 0.4768, "step": 19931 }, { "epoch": 0.8355218343585928, "grad_norm": 2.349717617034912, "learning_rate": 6.926210038286751e-07, "loss": 0.4811, "step": 19932 }, { "epoch": 0.8355637529736017, "grad_norm": 1.732203722000122, "learning_rate": 6.922763232652091e-07, "loss": 0.4564, "step": 19933 }, { "epoch": 0.8356056715886108, "grad_norm": 1.7774882316589355, "learning_rate": 6.919317221086753e-07, "loss": 0.4161, "step": 19934 }, { "epoch": 0.8356475902036197, "grad_norm": 2.3528506755828857, "learning_rate": 6.915872003654234e-07, "loss": 0.4894, "step": 19935 }, { "epoch": 0.8356895088186286, "grad_norm": 1.8290375471115112, "learning_rate": 6.912427580418046e-07, "loss": 0.4267, "step": 19936 }, { "epoch": 0.8357314274336376, "grad_norm": 2.00864577293396, "learning_rate": 6.908983951441705e-07, "loss": 0.473, "step": 19937 }, { "epoch": 0.8357733460486465, "grad_norm": 2.2929859161376953, "learning_rate": 6.905541116788672e-07, "loss": 0.4379, "step": 19938 }, { "epoch": 0.8358152646636555, "grad_norm": 1.9719141721725464, "learning_rate": 6.902099076522401e-07, "loss": 0.4678, "step": 19939 }, { "epoch": 0.8358571832786644, "grad_norm": 1.683797836303711, "learning_rate": 6.898657830706367e-07, "loss": 0.4644, "step": 19940 }, { "epoch": 0.8358991018936734, "grad_norm": 1.750909686088562, "learning_rate": 6.895217379403974e-07, "loss": 0.4589, "step": 19941 }, { "epoch": 0.8359410205086824, "grad_norm": 2.262606620788574, "learning_rate": 6.891777722678661e-07, "loss": 0.4522, "step": 19942 }, { "epoch": 0.8359829391236914, "grad_norm": 2.0580124855041504, "learning_rate": 6.888338860593841e-07, "loss": 0.5527, "step": 19943 }, { "epoch": 0.8360248577387003, "grad_norm": 1.7827671766281128, "learning_rate": 6.884900793212878e-07, "loss": 0.4304, "step": 19944 }, { "epoch": 0.8360667763537093, "grad_norm": 1.811539649963379, "learning_rate": 6.881463520599168e-07, "loss": 0.4619, "step": 19945 }, { "epoch": 0.8361086949687182, "grad_norm": 1.9675352573394775, "learning_rate": 6.878027042816082e-07, "loss": 0.4772, "step": 19946 }, { "epoch": 0.8361506135837272, "grad_norm": 1.8330780267715454, "learning_rate": 6.874591359926941e-07, "loss": 0.5255, "step": 19947 }, { "epoch": 0.8361925321987361, "grad_norm": 1.9475733041763306, "learning_rate": 6.8711564719951e-07, "loss": 0.4607, "step": 19948 }, { "epoch": 0.8362344508137451, "grad_norm": 2.3627939224243164, "learning_rate": 6.867722379083863e-07, "loss": 0.4687, "step": 19949 }, { "epoch": 0.836276369428754, "grad_norm": 2.213244676589966, "learning_rate": 6.864289081256525e-07, "loss": 0.4988, "step": 19950 }, { "epoch": 0.8363182880437631, "grad_norm": 2.6124825477600098, "learning_rate": 6.860856578576386e-07, "loss": 0.5331, "step": 19951 }, { "epoch": 0.836360206658772, "grad_norm": 1.7785946130752563, "learning_rate": 6.857424871106727e-07, "loss": 0.4837, "step": 19952 }, { "epoch": 0.836402125273781, "grad_norm": 1.9051350355148315, "learning_rate": 6.853993958910793e-07, "loss": 0.4724, "step": 19953 }, { "epoch": 0.8364440438887899, "grad_norm": 1.8424967527389526, "learning_rate": 6.850563842051827e-07, "loss": 0.4794, "step": 19954 }, { "epoch": 0.8364859625037989, "grad_norm": 1.8740686178207397, "learning_rate": 6.847134520593074e-07, "loss": 0.5161, "step": 19955 }, { "epoch": 0.8365278811188078, "grad_norm": 2.414367437362671, "learning_rate": 6.843705994597732e-07, "loss": 0.4844, "step": 19956 }, { "epoch": 0.8365697997338168, "grad_norm": 1.7757518291473389, "learning_rate": 6.840278264129018e-07, "loss": 0.4801, "step": 19957 }, { "epoch": 0.8366117183488258, "grad_norm": 1.8541475534439087, "learning_rate": 6.836851329250105e-07, "loss": 0.5111, "step": 19958 }, { "epoch": 0.8366536369638348, "grad_norm": 1.960744023323059, "learning_rate": 6.833425190024157e-07, "loss": 0.4402, "step": 19959 }, { "epoch": 0.8366955555788437, "grad_norm": 1.6970189809799194, "learning_rate": 6.82999984651434e-07, "loss": 0.416, "step": 19960 }, { "epoch": 0.8367374741938526, "grad_norm": 2.2012264728546143, "learning_rate": 6.826575298783805e-07, "loss": 0.4713, "step": 19961 }, { "epoch": 0.8367793928088616, "grad_norm": 2.0248451232910156, "learning_rate": 6.823151546895662e-07, "loss": 0.4361, "step": 19962 }, { "epoch": 0.8368213114238705, "grad_norm": 1.9116476774215698, "learning_rate": 6.819728590913027e-07, "loss": 0.4933, "step": 19963 }, { "epoch": 0.8368632300388795, "grad_norm": 2.286325454711914, "learning_rate": 6.816306430899012e-07, "loss": 0.4388, "step": 19964 }, { "epoch": 0.8369051486538884, "grad_norm": 1.950516939163208, "learning_rate": 6.812885066916675e-07, "loss": 0.518, "step": 19965 }, { "epoch": 0.8369470672688974, "grad_norm": 3.062164783477783, "learning_rate": 6.809464499029111e-07, "loss": 0.4836, "step": 19966 }, { "epoch": 0.8369889858839064, "grad_norm": 1.94637131690979, "learning_rate": 6.806044727299361e-07, "loss": 0.4894, "step": 19967 }, { "epoch": 0.8370309044989154, "grad_norm": 1.7039802074432373, "learning_rate": 6.802625751790448e-07, "loss": 0.4687, "step": 19968 }, { "epoch": 0.8370728231139243, "grad_norm": 1.834748387336731, "learning_rate": 6.799207572565414e-07, "loss": 0.491, "step": 19969 }, { "epoch": 0.8371147417289333, "grad_norm": 1.9532612562179565, "learning_rate": 6.795790189687274e-07, "loss": 0.5035, "step": 19970 }, { "epoch": 0.8371566603439422, "grad_norm": 1.7321113348007202, "learning_rate": 6.792373603219005e-07, "loss": 0.4628, "step": 19971 }, { "epoch": 0.8371985789589512, "grad_norm": 2.1007120609283447, "learning_rate": 6.788957813223596e-07, "loss": 0.4875, "step": 19972 }, { "epoch": 0.8372404975739601, "grad_norm": 1.7954550981521606, "learning_rate": 6.785542819764018e-07, "loss": 0.4411, "step": 19973 }, { "epoch": 0.8372824161889691, "grad_norm": 1.6831315755844116, "learning_rate": 6.782128622903212e-07, "loss": 0.4035, "step": 19974 }, { "epoch": 0.837324334803978, "grad_norm": 2.8877053260803223, "learning_rate": 6.778715222704124e-07, "loss": 0.5123, "step": 19975 }, { "epoch": 0.8373662534189871, "grad_norm": 1.9334936141967773, "learning_rate": 6.775302619229668e-07, "loss": 0.495, "step": 19976 }, { "epoch": 0.837408172033996, "grad_norm": 2.9865031242370605, "learning_rate": 6.771890812542748e-07, "loss": 0.5021, "step": 19977 }, { "epoch": 0.837450090649005, "grad_norm": 1.6698707342147827, "learning_rate": 6.768479802706251e-07, "loss": 0.4054, "step": 19978 }, { "epoch": 0.8374920092640139, "grad_norm": 1.873456597328186, "learning_rate": 6.76506958978308e-07, "loss": 0.4473, "step": 19979 }, { "epoch": 0.8375339278790229, "grad_norm": 1.8858270645141602, "learning_rate": 6.761660173836066e-07, "loss": 0.5024, "step": 19980 }, { "epoch": 0.8375758464940318, "grad_norm": 2.0686869621276855, "learning_rate": 6.758251554928075e-07, "loss": 0.4791, "step": 19981 }, { "epoch": 0.8376177651090408, "grad_norm": 1.9172606468200684, "learning_rate": 6.754843733121946e-07, "loss": 0.4995, "step": 19982 }, { "epoch": 0.8376596837240498, "grad_norm": 2.7518818378448486, "learning_rate": 6.75143670848048e-07, "loss": 0.4626, "step": 19983 }, { "epoch": 0.8377016023390588, "grad_norm": 2.031919240951538, "learning_rate": 6.748030481066492e-07, "loss": 0.4627, "step": 19984 }, { "epoch": 0.8377435209540677, "grad_norm": 1.6680278778076172, "learning_rate": 6.744625050942777e-07, "loss": 0.4256, "step": 19985 }, { "epoch": 0.8377854395690766, "grad_norm": 1.6623646020889282, "learning_rate": 6.741220418172084e-07, "loss": 0.4183, "step": 19986 }, { "epoch": 0.8378273581840856, "grad_norm": 2.6300203800201416, "learning_rate": 6.737816582817191e-07, "loss": 0.4492, "step": 19987 }, { "epoch": 0.8378692767990945, "grad_norm": 2.2269344329833984, "learning_rate": 6.734413544940854e-07, "loss": 0.5194, "step": 19988 }, { "epoch": 0.8379111954141035, "grad_norm": 1.6007283926010132, "learning_rate": 6.731011304605778e-07, "loss": 0.4504, "step": 19989 }, { "epoch": 0.8379531140291124, "grad_norm": 1.9357209205627441, "learning_rate": 6.72760986187469e-07, "loss": 0.4725, "step": 19990 }, { "epoch": 0.8379950326441215, "grad_norm": 1.6856012344360352, "learning_rate": 6.724209216810301e-07, "loss": 0.4475, "step": 19991 }, { "epoch": 0.8380369512591304, "grad_norm": 1.9875560998916626, "learning_rate": 6.720809369475279e-07, "loss": 0.4792, "step": 19992 }, { "epoch": 0.8380788698741394, "grad_norm": 1.8388638496398926, "learning_rate": 6.717410319932316e-07, "loss": 0.4536, "step": 19993 }, { "epoch": 0.8381207884891483, "grad_norm": 2.1039657592773438, "learning_rate": 6.714012068244046e-07, "loss": 0.4518, "step": 19994 }, { "epoch": 0.8381627071041573, "grad_norm": 1.7786418199539185, "learning_rate": 6.710614614473132e-07, "loss": 0.4885, "step": 19995 }, { "epoch": 0.8382046257191662, "grad_norm": 2.938971757888794, "learning_rate": 6.707217958682183e-07, "loss": 0.5072, "step": 19996 }, { "epoch": 0.8382465443341752, "grad_norm": 1.8543076515197754, "learning_rate": 6.703822100933832e-07, "loss": 0.4671, "step": 19997 }, { "epoch": 0.8382884629491841, "grad_norm": 1.8634064197540283, "learning_rate": 6.700427041290652e-07, "loss": 0.521, "step": 19998 }, { "epoch": 0.8383303815641932, "grad_norm": 1.7580379247665405, "learning_rate": 6.697032779815254e-07, "loss": 0.4846, "step": 19999 }, { "epoch": 0.8383723001792021, "grad_norm": 1.7444368600845337, "learning_rate": 6.693639316570178e-07, "loss": 0.5003, "step": 20000 }, { "epoch": 0.8384142187942111, "grad_norm": 1.852308750152588, "learning_rate": 6.690246651617994e-07, "loss": 0.4835, "step": 20001 }, { "epoch": 0.83845613740922, "grad_norm": 1.9299973249435425, "learning_rate": 6.686854785021252e-07, "loss": 0.4444, "step": 20002 }, { "epoch": 0.838498056024229, "grad_norm": 2.061157464981079, "learning_rate": 6.683463716842447e-07, "loss": 0.4624, "step": 20003 }, { "epoch": 0.8385399746392379, "grad_norm": 2.0348269939422607, "learning_rate": 6.68007344714412e-07, "loss": 0.4609, "step": 20004 }, { "epoch": 0.8385818932542469, "grad_norm": 1.6766185760498047, "learning_rate": 6.676683975988752e-07, "loss": 0.4844, "step": 20005 }, { "epoch": 0.8386238118692558, "grad_norm": 1.820902943611145, "learning_rate": 6.673295303438804e-07, "loss": 0.5218, "step": 20006 }, { "epoch": 0.8386657304842648, "grad_norm": 1.8458596467971802, "learning_rate": 6.669907429556766e-07, "loss": 0.4562, "step": 20007 }, { "epoch": 0.8387076490992738, "grad_norm": 2.3182523250579834, "learning_rate": 6.666520354405093e-07, "loss": 0.4063, "step": 20008 }, { "epoch": 0.8387495677142828, "grad_norm": 2.0200424194335938, "learning_rate": 6.663134078046196e-07, "loss": 0.4552, "step": 20009 }, { "epoch": 0.8387914863292917, "grad_norm": 2.038465976715088, "learning_rate": 6.659748600542515e-07, "loss": 0.4989, "step": 20010 }, { "epoch": 0.8388334049443006, "grad_norm": 1.74176824092865, "learning_rate": 6.65636392195646e-07, "loss": 0.5106, "step": 20011 }, { "epoch": 0.8388753235593096, "grad_norm": 1.8199690580368042, "learning_rate": 6.65298004235041e-07, "loss": 0.4716, "step": 20012 }, { "epoch": 0.8389172421743185, "grad_norm": 1.789178490638733, "learning_rate": 6.649596961786752e-07, "loss": 0.507, "step": 20013 }, { "epoch": 0.8389591607893275, "grad_norm": 1.941552996635437, "learning_rate": 6.646214680327851e-07, "loss": 0.4579, "step": 20014 }, { "epoch": 0.8390010794043364, "grad_norm": 1.8026385307312012, "learning_rate": 6.64283319803603e-07, "loss": 0.4252, "step": 20015 }, { "epoch": 0.8390429980193455, "grad_norm": 1.7963309288024902, "learning_rate": 6.639452514973638e-07, "loss": 0.453, "step": 20016 }, { "epoch": 0.8390849166343544, "grad_norm": 2.3122360706329346, "learning_rate": 6.636072631203011e-07, "loss": 0.4782, "step": 20017 }, { "epoch": 0.8391268352493634, "grad_norm": 1.933811902999878, "learning_rate": 6.632693546786422e-07, "loss": 0.5055, "step": 20018 }, { "epoch": 0.8391687538643723, "grad_norm": 5.569144248962402, "learning_rate": 6.629315261786179e-07, "loss": 0.4682, "step": 20019 }, { "epoch": 0.8392106724793813, "grad_norm": 1.6751291751861572, "learning_rate": 6.625937776264557e-07, "loss": 0.4935, "step": 20020 }, { "epoch": 0.8392525910943902, "grad_norm": 1.8264518976211548, "learning_rate": 6.622561090283802e-07, "loss": 0.5292, "step": 20021 }, { "epoch": 0.8392945097093992, "grad_norm": 2.050428628921509, "learning_rate": 6.619185203906175e-07, "loss": 0.4764, "step": 20022 }, { "epoch": 0.8393364283244081, "grad_norm": 1.9167835712432861, "learning_rate": 6.615810117193894e-07, "loss": 0.4666, "step": 20023 }, { "epoch": 0.8393783469394172, "grad_norm": 1.9940526485443115, "learning_rate": 6.612435830209169e-07, "loss": 0.4535, "step": 20024 }, { "epoch": 0.8394202655544261, "grad_norm": 1.5683913230895996, "learning_rate": 6.609062343014205e-07, "loss": 0.4322, "step": 20025 }, { "epoch": 0.8394621841694351, "grad_norm": 2.2476143836975098, "learning_rate": 6.605689655671205e-07, "loss": 0.4384, "step": 20026 }, { "epoch": 0.839504102784444, "grad_norm": 2.1740803718566895, "learning_rate": 6.602317768242312e-07, "loss": 0.4411, "step": 20027 }, { "epoch": 0.839546021399453, "grad_norm": 1.8722928762435913, "learning_rate": 6.598946680789697e-07, "loss": 0.4518, "step": 20028 }, { "epoch": 0.8395879400144619, "grad_norm": 1.899167537689209, "learning_rate": 6.595576393375508e-07, "loss": 0.4635, "step": 20029 }, { "epoch": 0.8396298586294709, "grad_norm": 4.403110504150391, "learning_rate": 6.592206906061854e-07, "loss": 0.4731, "step": 20030 }, { "epoch": 0.8396717772444798, "grad_norm": 1.5484464168548584, "learning_rate": 6.58883821891087e-07, "loss": 0.4651, "step": 20031 }, { "epoch": 0.8397136958594889, "grad_norm": 1.635703444480896, "learning_rate": 6.585470331984634e-07, "loss": 0.4519, "step": 20032 }, { "epoch": 0.8397556144744978, "grad_norm": 2.0271172523498535, "learning_rate": 6.582103245345229e-07, "loss": 0.4488, "step": 20033 }, { "epoch": 0.8397975330895068, "grad_norm": 1.705955147743225, "learning_rate": 6.578736959054721e-07, "loss": 0.4278, "step": 20034 }, { "epoch": 0.8398394517045157, "grad_norm": 2.3540265560150146, "learning_rate": 6.575371473175185e-07, "loss": 0.4882, "step": 20035 }, { "epoch": 0.8398813703195246, "grad_norm": 1.7781018018722534, "learning_rate": 6.572006787768626e-07, "loss": 0.4484, "step": 20036 }, { "epoch": 0.8399232889345336, "grad_norm": 4.3063273429870605, "learning_rate": 6.568642902897087e-07, "loss": 0.52, "step": 20037 }, { "epoch": 0.8399652075495425, "grad_norm": 1.7038308382034302, "learning_rate": 6.565279818622588e-07, "loss": 0.4656, "step": 20038 }, { "epoch": 0.8400071261645515, "grad_norm": 1.795184850692749, "learning_rate": 6.561917535007096e-07, "loss": 0.4825, "step": 20039 }, { "epoch": 0.8400490447795604, "grad_norm": 1.8515251874923706, "learning_rate": 6.558556052112614e-07, "loss": 0.4955, "step": 20040 }, { "epoch": 0.8400909633945695, "grad_norm": 3.1223666667938232, "learning_rate": 6.555195370001083e-07, "loss": 0.5116, "step": 20041 }, { "epoch": 0.8401328820095784, "grad_norm": 1.6668444871902466, "learning_rate": 6.551835488734481e-07, "loss": 0.4489, "step": 20042 }, { "epoch": 0.8401748006245874, "grad_norm": 1.8490837812423706, "learning_rate": 6.54847640837471e-07, "loss": 0.4936, "step": 20043 }, { "epoch": 0.8402167192395963, "grad_norm": 2.6538398265838623, "learning_rate": 6.545118128983718e-07, "loss": 0.4923, "step": 20044 }, { "epoch": 0.8402586378546053, "grad_norm": 2.3143248558044434, "learning_rate": 6.541760650623391e-07, "loss": 0.5445, "step": 20045 }, { "epoch": 0.8403005564696142, "grad_norm": 1.898390293121338, "learning_rate": 6.538403973355623e-07, "loss": 0.472, "step": 20046 }, { "epoch": 0.8403424750846232, "grad_norm": 2.5175087451934814, "learning_rate": 6.535048097242308e-07, "loss": 0.4856, "step": 20047 }, { "epoch": 0.8403843936996321, "grad_norm": 1.9486223459243774, "learning_rate": 6.531693022345282e-07, "loss": 0.4648, "step": 20048 }, { "epoch": 0.8404263123146412, "grad_norm": 1.9949020147323608, "learning_rate": 6.528338748726398e-07, "loss": 0.5163, "step": 20049 }, { "epoch": 0.8404682309296501, "grad_norm": 1.999633550643921, "learning_rate": 6.524985276447504e-07, "loss": 0.517, "step": 20050 }, { "epoch": 0.8405101495446591, "grad_norm": 2.662790298461914, "learning_rate": 6.521632605570405e-07, "loss": 0.5145, "step": 20051 }, { "epoch": 0.840552068159668, "grad_norm": 1.9844228029251099, "learning_rate": 6.51828073615689e-07, "loss": 0.4546, "step": 20052 }, { "epoch": 0.840593986774677, "grad_norm": 2.2110848426818848, "learning_rate": 6.514929668268771e-07, "loss": 0.5232, "step": 20053 }, { "epoch": 0.8406359053896859, "grad_norm": 1.9984419345855713, "learning_rate": 6.511579401967793e-07, "loss": 0.4767, "step": 20054 }, { "epoch": 0.8406778240046949, "grad_norm": 1.4627864360809326, "learning_rate": 6.508229937315735e-07, "loss": 0.4384, "step": 20055 }, { "epoch": 0.8407197426197038, "grad_norm": 2.1476430892944336, "learning_rate": 6.504881274374336e-07, "loss": 0.4387, "step": 20056 }, { "epoch": 0.8407616612347129, "grad_norm": 2.023428201675415, "learning_rate": 6.501533413205313e-07, "loss": 0.451, "step": 20057 }, { "epoch": 0.8408035798497218, "grad_norm": 1.9425121545791626, "learning_rate": 6.498186353870384e-07, "loss": 0.4765, "step": 20058 }, { "epoch": 0.8408454984647308, "grad_norm": 2.0198559761047363, "learning_rate": 6.494840096431265e-07, "loss": 0.5115, "step": 20059 }, { "epoch": 0.8408874170797397, "grad_norm": 2.022489070892334, "learning_rate": 6.49149464094962e-07, "loss": 0.4674, "step": 20060 }, { "epoch": 0.8409293356947486, "grad_norm": 2.0006158351898193, "learning_rate": 6.488149987487114e-07, "loss": 0.4141, "step": 20061 }, { "epoch": 0.8409712543097576, "grad_norm": 2.43792462348938, "learning_rate": 6.484806136105421e-07, "loss": 0.4331, "step": 20062 }, { "epoch": 0.8410131729247665, "grad_norm": 1.587036371231079, "learning_rate": 6.481463086866158e-07, "loss": 0.408, "step": 20063 }, { "epoch": 0.8410550915397755, "grad_norm": 2.2136728763580322, "learning_rate": 6.478120839830959e-07, "loss": 0.4783, "step": 20064 }, { "epoch": 0.8410970101547844, "grad_norm": 2.166867256164551, "learning_rate": 6.474779395061448e-07, "loss": 0.4231, "step": 20065 }, { "epoch": 0.8411389287697935, "grad_norm": 1.7407357692718506, "learning_rate": 6.471438752619197e-07, "loss": 0.4428, "step": 20066 }, { "epoch": 0.8411808473848024, "grad_norm": 1.7787340879440308, "learning_rate": 6.468098912565796e-07, "loss": 0.3966, "step": 20067 }, { "epoch": 0.8412227659998114, "grad_norm": 2.171618700027466, "learning_rate": 6.464759874962817e-07, "loss": 0.4673, "step": 20068 }, { "epoch": 0.8412646846148203, "grad_norm": 1.8875221014022827, "learning_rate": 6.461421639871807e-07, "loss": 0.4971, "step": 20069 }, { "epoch": 0.8413066032298293, "grad_norm": 2.3673462867736816, "learning_rate": 6.458084207354287e-07, "loss": 0.4211, "step": 20070 }, { "epoch": 0.8413485218448382, "grad_norm": 2.0251479148864746, "learning_rate": 6.454747577471804e-07, "loss": 0.531, "step": 20071 }, { "epoch": 0.8413904404598472, "grad_norm": 2.1837525367736816, "learning_rate": 6.451411750285835e-07, "loss": 0.4692, "step": 20072 }, { "epoch": 0.8414323590748561, "grad_norm": 1.9141305685043335, "learning_rate": 6.448076725857888e-07, "loss": 0.4801, "step": 20073 }, { "epoch": 0.8414742776898652, "grad_norm": 3.7293922901153564, "learning_rate": 6.444742504249446e-07, "loss": 0.5689, "step": 20074 }, { "epoch": 0.8415161963048741, "grad_norm": 1.9252909421920776, "learning_rate": 6.441409085521955e-07, "loss": 0.441, "step": 20075 }, { "epoch": 0.8415581149198831, "grad_norm": 1.761357307434082, "learning_rate": 6.438076469736882e-07, "loss": 0.4536, "step": 20076 }, { "epoch": 0.841600033534892, "grad_norm": 1.9292746782302856, "learning_rate": 6.434744656955633e-07, "loss": 0.4896, "step": 20077 }, { "epoch": 0.841641952149901, "grad_norm": 1.9770182371139526, "learning_rate": 6.431413647239648e-07, "loss": 0.4922, "step": 20078 }, { "epoch": 0.8416838707649099, "grad_norm": 1.9084813594818115, "learning_rate": 6.428083440650313e-07, "loss": 0.4531, "step": 20079 }, { "epoch": 0.8417257893799189, "grad_norm": 1.7687987089157104, "learning_rate": 6.424754037249037e-07, "loss": 0.4018, "step": 20080 }, { "epoch": 0.8417677079949278, "grad_norm": 2.4501707553863525, "learning_rate": 6.421425437097162e-07, "loss": 0.4535, "step": 20081 }, { "epoch": 0.8418096266099369, "grad_norm": 2.5736467838287354, "learning_rate": 6.418097640256077e-07, "loss": 0.4902, "step": 20082 }, { "epoch": 0.8418515452249458, "grad_norm": 1.98057222366333, "learning_rate": 6.414770646787105e-07, "loss": 0.4793, "step": 20083 }, { "epoch": 0.8418934638399548, "grad_norm": 1.7279564142227173, "learning_rate": 6.411444456751575e-07, "loss": 0.4546, "step": 20084 }, { "epoch": 0.8419353824549637, "grad_norm": 2.1195461750030518, "learning_rate": 6.408119070210822e-07, "loss": 0.4807, "step": 20085 }, { "epoch": 0.8419773010699726, "grad_norm": 2.0414764881134033, "learning_rate": 6.404794487226118e-07, "loss": 0.5192, "step": 20086 }, { "epoch": 0.8420192196849816, "grad_norm": 1.7865290641784668, "learning_rate": 6.40147070785877e-07, "loss": 0.442, "step": 20087 }, { "epoch": 0.8420611382999905, "grad_norm": 1.7421478033065796, "learning_rate": 6.398147732170023e-07, "loss": 0.4617, "step": 20088 }, { "epoch": 0.8421030569149995, "grad_norm": 1.8717973232269287, "learning_rate": 6.394825560221157e-07, "loss": 0.5267, "step": 20089 }, { "epoch": 0.8421449755300084, "grad_norm": 1.8089009523391724, "learning_rate": 6.391504192073389e-07, "loss": 0.4775, "step": 20090 }, { "epoch": 0.8421868941450175, "grad_norm": 1.7474472522735596, "learning_rate": 6.38818362778797e-07, "loss": 0.5144, "step": 20091 }, { "epoch": 0.8422288127600264, "grad_norm": 2.0997307300567627, "learning_rate": 6.384863867426077e-07, "loss": 0.4747, "step": 20092 }, { "epoch": 0.8422707313750354, "grad_norm": 1.9464293718338013, "learning_rate": 6.381544911048926e-07, "loss": 0.4343, "step": 20093 }, { "epoch": 0.8423126499900443, "grad_norm": 1.8959380388259888, "learning_rate": 6.378226758717698e-07, "loss": 0.4114, "step": 20094 }, { "epoch": 0.8423545686050533, "grad_norm": 2.095080614089966, "learning_rate": 6.374909410493551e-07, "loss": 0.4707, "step": 20095 }, { "epoch": 0.8423964872200622, "grad_norm": 2.3188655376434326, "learning_rate": 6.371592866437637e-07, "loss": 0.4531, "step": 20096 }, { "epoch": 0.8424384058350712, "grad_norm": 1.7355433702468872, "learning_rate": 6.368277126611105e-07, "loss": 0.4702, "step": 20097 }, { "epoch": 0.8424803244500801, "grad_norm": 1.871444582939148, "learning_rate": 6.364962191075064e-07, "loss": 0.4502, "step": 20098 }, { "epoch": 0.8425222430650892, "grad_norm": 2.4608516693115234, "learning_rate": 6.361648059890613e-07, "loss": 0.459, "step": 20099 }, { "epoch": 0.8425641616800981, "grad_norm": 1.9994884729385376, "learning_rate": 6.358334733118859e-07, "loss": 0.4824, "step": 20100 }, { "epoch": 0.8426060802951071, "grad_norm": 2.273216724395752, "learning_rate": 6.355022210820866e-07, "loss": 0.4348, "step": 20101 }, { "epoch": 0.842647998910116, "grad_norm": 1.871793270111084, "learning_rate": 6.351710493057695e-07, "loss": 0.4897, "step": 20102 }, { "epoch": 0.842689917525125, "grad_norm": 2.013331174850464, "learning_rate": 6.348399579890413e-07, "loss": 0.4761, "step": 20103 }, { "epoch": 0.8427318361401339, "grad_norm": 4.8798441886901855, "learning_rate": 6.345089471380028e-07, "loss": 0.4407, "step": 20104 }, { "epoch": 0.8427737547551429, "grad_norm": 1.7059600353240967, "learning_rate": 6.341780167587563e-07, "loss": 0.4533, "step": 20105 }, { "epoch": 0.8428156733701518, "grad_norm": 1.8426384925842285, "learning_rate": 6.338471668574042e-07, "loss": 0.4168, "step": 20106 }, { "epoch": 0.8428575919851609, "grad_norm": 1.8618470430374146, "learning_rate": 6.335163974400433e-07, "loss": 0.4594, "step": 20107 }, { "epoch": 0.8428995106001698, "grad_norm": 2.049375534057617, "learning_rate": 6.331857085127702e-07, "loss": 0.4783, "step": 20108 }, { "epoch": 0.8429414292151788, "grad_norm": 1.8538132905960083, "learning_rate": 6.328551000816824e-07, "loss": 0.4329, "step": 20109 }, { "epoch": 0.8429833478301877, "grad_norm": 1.932103157043457, "learning_rate": 6.325245721528723e-07, "loss": 0.4511, "step": 20110 }, { "epoch": 0.8430252664451966, "grad_norm": 1.8060575723648071, "learning_rate": 6.321941247324343e-07, "loss": 0.4435, "step": 20111 }, { "epoch": 0.8430671850602056, "grad_norm": 1.707252025604248, "learning_rate": 6.318637578264603e-07, "loss": 0.4565, "step": 20112 }, { "epoch": 0.8431091036752145, "grad_norm": 1.6189216375350952, "learning_rate": 6.315334714410377e-07, "loss": 0.4717, "step": 20113 }, { "epoch": 0.8431510222902235, "grad_norm": 1.7407243251800537, "learning_rate": 6.312032655822565e-07, "loss": 0.4587, "step": 20114 }, { "epoch": 0.8431929409052324, "grad_norm": 1.9215186834335327, "learning_rate": 6.308731402562046e-07, "loss": 0.5147, "step": 20115 }, { "epoch": 0.8432348595202415, "grad_norm": 1.7725197076797485, "learning_rate": 6.305430954689662e-07, "loss": 0.4426, "step": 20116 }, { "epoch": 0.8432767781352504, "grad_norm": 1.6700356006622314, "learning_rate": 6.302131312266241e-07, "loss": 0.4445, "step": 20117 }, { "epoch": 0.8433186967502594, "grad_norm": 1.8073492050170898, "learning_rate": 6.298832475352629e-07, "loss": 0.4649, "step": 20118 }, { "epoch": 0.8433606153652683, "grad_norm": 2.518355131149292, "learning_rate": 6.295534444009615e-07, "loss": 0.4243, "step": 20119 }, { "epoch": 0.8434025339802773, "grad_norm": 2.754992961883545, "learning_rate": 6.292237218298003e-07, "loss": 0.5164, "step": 20120 }, { "epoch": 0.8434444525952862, "grad_norm": 1.9383959770202637, "learning_rate": 6.288940798278587e-07, "loss": 0.4275, "step": 20121 }, { "epoch": 0.8434863712102952, "grad_norm": 3.0308802127838135, "learning_rate": 6.285645184012106e-07, "loss": 0.5189, "step": 20122 }, { "epoch": 0.8435282898253041, "grad_norm": 1.9536913633346558, "learning_rate": 6.282350375559326e-07, "loss": 0.5009, "step": 20123 }, { "epoch": 0.8435702084403132, "grad_norm": 1.7067909240722656, "learning_rate": 6.279056372980985e-07, "loss": 0.4461, "step": 20124 }, { "epoch": 0.8436121270553221, "grad_norm": 1.7873986959457397, "learning_rate": 6.275763176337801e-07, "loss": 0.4908, "step": 20125 }, { "epoch": 0.8436540456703311, "grad_norm": 2.467097759246826, "learning_rate": 6.272470785690459e-07, "loss": 0.4517, "step": 20126 }, { "epoch": 0.84369596428534, "grad_norm": 1.9829739332199097, "learning_rate": 6.269179201099685e-07, "loss": 0.488, "step": 20127 }, { "epoch": 0.843737882900349, "grad_norm": 2.2454588413238525, "learning_rate": 6.265888422626121e-07, "loss": 0.4884, "step": 20128 }, { "epoch": 0.8437798015153579, "grad_norm": 1.6845736503601074, "learning_rate": 6.262598450330448e-07, "loss": 0.4019, "step": 20129 }, { "epoch": 0.8438217201303669, "grad_norm": 1.7857104539871216, "learning_rate": 6.259309284273313e-07, "loss": 0.5091, "step": 20130 }, { "epoch": 0.8438636387453758, "grad_norm": 1.8620142936706543, "learning_rate": 6.256020924515332e-07, "loss": 0.4597, "step": 20131 }, { "epoch": 0.8439055573603849, "grad_norm": 1.898272156715393, "learning_rate": 6.252733371117131e-07, "loss": 0.4905, "step": 20132 }, { "epoch": 0.8439474759753938, "grad_norm": 2.0950236320495605, "learning_rate": 6.249446624139327e-07, "loss": 0.4976, "step": 20133 }, { "epoch": 0.8439893945904028, "grad_norm": 2.269235610961914, "learning_rate": 6.246160683642477e-07, "loss": 0.4612, "step": 20134 }, { "epoch": 0.8440313132054117, "grad_norm": 1.831635594367981, "learning_rate": 6.242875549687178e-07, "loss": 0.4841, "step": 20135 }, { "epoch": 0.8440732318204206, "grad_norm": 1.794899821281433, "learning_rate": 6.239591222333974e-07, "loss": 0.463, "step": 20136 }, { "epoch": 0.8441151504354296, "grad_norm": 1.8782436847686768, "learning_rate": 6.236307701643396e-07, "loss": 0.4734, "step": 20137 }, { "epoch": 0.8441570690504385, "grad_norm": 2.797853946685791, "learning_rate": 6.233024987675989e-07, "loss": 0.519, "step": 20138 }, { "epoch": 0.8441989876654475, "grad_norm": 4.100302696228027, "learning_rate": 6.229743080492268e-07, "loss": 0.4278, "step": 20139 }, { "epoch": 0.8442409062804564, "grad_norm": 1.6160495281219482, "learning_rate": 6.226461980152715e-07, "loss": 0.4429, "step": 20140 }, { "epoch": 0.8442828248954655, "grad_norm": 1.7059894800186157, "learning_rate": 6.223181686717816e-07, "loss": 0.4633, "step": 20141 }, { "epoch": 0.8443247435104744, "grad_norm": 1.9410285949707031, "learning_rate": 6.219902200248057e-07, "loss": 0.4502, "step": 20142 }, { "epoch": 0.8443666621254834, "grad_norm": 2.4101226329803467, "learning_rate": 6.216623520803866e-07, "loss": 0.471, "step": 20143 }, { "epoch": 0.8444085807404923, "grad_norm": 1.8233777284622192, "learning_rate": 6.213345648445701e-07, "loss": 0.4424, "step": 20144 }, { "epoch": 0.8444504993555013, "grad_norm": 1.8430553674697876, "learning_rate": 6.21006858323398e-07, "loss": 0.4624, "step": 20145 }, { "epoch": 0.8444924179705102, "grad_norm": 1.8814468383789062, "learning_rate": 6.206792325229094e-07, "loss": 0.4733, "step": 20146 }, { "epoch": 0.8445343365855192, "grad_norm": 2.0830419063568115, "learning_rate": 6.203516874491444e-07, "loss": 0.4687, "step": 20147 }, { "epoch": 0.8445762552005281, "grad_norm": 1.6927516460418701, "learning_rate": 6.200242231081432e-07, "loss": 0.4274, "step": 20148 }, { "epoch": 0.8446181738155372, "grad_norm": 2.3442161083221436, "learning_rate": 6.196968395059389e-07, "loss": 0.4664, "step": 20149 }, { "epoch": 0.8446600924305461, "grad_norm": 2.0058112144470215, "learning_rate": 6.193695366485686e-07, "loss": 0.5221, "step": 20150 }, { "epoch": 0.8447020110455551, "grad_norm": 2.161175489425659, "learning_rate": 6.190423145420638e-07, "loss": 0.4634, "step": 20151 }, { "epoch": 0.844743929660564, "grad_norm": 2.036881923675537, "learning_rate": 6.187151731924578e-07, "loss": 0.5507, "step": 20152 }, { "epoch": 0.844785848275573, "grad_norm": 3.6302545070648193, "learning_rate": 6.183881126057817e-07, "loss": 0.4909, "step": 20153 }, { "epoch": 0.8448277668905819, "grad_norm": 1.684396505355835, "learning_rate": 6.180611327880631e-07, "loss": 0.4643, "step": 20154 }, { "epoch": 0.8448696855055909, "grad_norm": 2.5763444900512695, "learning_rate": 6.177342337453284e-07, "loss": 0.4728, "step": 20155 }, { "epoch": 0.8449116041205998, "grad_norm": 9.337459564208984, "learning_rate": 6.174074154836063e-07, "loss": 0.4455, "step": 20156 }, { "epoch": 0.8449535227356089, "grad_norm": 2.0109474658966064, "learning_rate": 6.170806780089189e-07, "loss": 0.4455, "step": 20157 }, { "epoch": 0.8449954413506178, "grad_norm": 2.2249271869659424, "learning_rate": 6.167540213272893e-07, "loss": 0.5628, "step": 20158 }, { "epoch": 0.8450373599656268, "grad_norm": 1.9705499410629272, "learning_rate": 6.164274454447411e-07, "loss": 0.4848, "step": 20159 }, { "epoch": 0.8450792785806357, "grad_norm": 2.067579746246338, "learning_rate": 6.161009503672916e-07, "loss": 0.4707, "step": 20160 }, { "epoch": 0.8451211971956446, "grad_norm": 1.7612650394439697, "learning_rate": 6.15774536100961e-07, "loss": 0.4983, "step": 20161 }, { "epoch": 0.8451631158106536, "grad_norm": 1.566614031791687, "learning_rate": 6.154482026517667e-07, "loss": 0.421, "step": 20162 }, { "epoch": 0.8452050344256625, "grad_norm": 2.300333023071289, "learning_rate": 6.151219500257227e-07, "loss": 0.5061, "step": 20163 }, { "epoch": 0.8452469530406715, "grad_norm": 2.811537265777588, "learning_rate": 6.147957782288433e-07, "loss": 0.509, "step": 20164 }, { "epoch": 0.8452888716556805, "grad_norm": 1.8145253658294678, "learning_rate": 6.144696872671419e-07, "loss": 0.3998, "step": 20165 }, { "epoch": 0.8453307902706895, "grad_norm": 3.027263641357422, "learning_rate": 6.141436771466286e-07, "loss": 0.4926, "step": 20166 }, { "epoch": 0.8453727088856984, "grad_norm": 1.9658873081207275, "learning_rate": 6.138177478733126e-07, "loss": 0.5156, "step": 20167 }, { "epoch": 0.8454146275007074, "grad_norm": 2.0226776599884033, "learning_rate": 6.134918994532041e-07, "loss": 0.4444, "step": 20168 }, { "epoch": 0.8454565461157163, "grad_norm": 1.490716814994812, "learning_rate": 6.131661318923071e-07, "loss": 0.403, "step": 20169 }, { "epoch": 0.8454984647307253, "grad_norm": 1.6802729368209839, "learning_rate": 6.12840445196628e-07, "loss": 0.4781, "step": 20170 }, { "epoch": 0.8455403833457342, "grad_norm": 2.1955957412719727, "learning_rate": 6.125148393721713e-07, "loss": 0.4736, "step": 20171 }, { "epoch": 0.8455823019607432, "grad_norm": 1.9147406816482544, "learning_rate": 6.12189314424938e-07, "loss": 0.495, "step": 20172 }, { "epoch": 0.8456242205757522, "grad_norm": 1.814805507659912, "learning_rate": 6.118638703609276e-07, "loss": 0.4618, "step": 20173 }, { "epoch": 0.8456661391907612, "grad_norm": 2.010437488555908, "learning_rate": 6.115385071861412e-07, "loss": 0.4292, "step": 20174 }, { "epoch": 0.8457080578057701, "grad_norm": 2.0887153148651123, "learning_rate": 6.11213224906575e-07, "loss": 0.5312, "step": 20175 }, { "epoch": 0.8457499764207791, "grad_norm": 8.553238868713379, "learning_rate": 6.108880235282255e-07, "loss": 0.4785, "step": 20176 }, { "epoch": 0.845791895035788, "grad_norm": 1.7578157186508179, "learning_rate": 6.105629030570887e-07, "loss": 0.4484, "step": 20177 }, { "epoch": 0.845833813650797, "grad_norm": 1.9400769472122192, "learning_rate": 6.102378634991552e-07, "loss": 0.4856, "step": 20178 }, { "epoch": 0.8458757322658059, "grad_norm": 3.0709939002990723, "learning_rate": 6.09912904860418e-07, "loss": 0.4851, "step": 20179 }, { "epoch": 0.8459176508808149, "grad_norm": 2.213726043701172, "learning_rate": 6.095880271468685e-07, "loss": 0.4899, "step": 20180 }, { "epoch": 0.8459595694958238, "grad_norm": 2.4508612155914307, "learning_rate": 6.09263230364493e-07, "loss": 0.4694, "step": 20181 }, { "epoch": 0.8460014881108329, "grad_norm": 1.9928463697433472, "learning_rate": 6.089385145192811e-07, "loss": 0.4558, "step": 20182 }, { "epoch": 0.8460434067258418, "grad_norm": 1.9835422039031982, "learning_rate": 6.08613879617217e-07, "loss": 0.4835, "step": 20183 }, { "epoch": 0.8460853253408508, "grad_norm": 1.8660988807678223, "learning_rate": 6.082893256642841e-07, "loss": 0.4999, "step": 20184 }, { "epoch": 0.8461272439558597, "grad_norm": 2.607818841934204, "learning_rate": 6.079648526664661e-07, "loss": 0.4435, "step": 20185 }, { "epoch": 0.8461691625708686, "grad_norm": 1.8291138410568237, "learning_rate": 6.076404606297453e-07, "loss": 0.5086, "step": 20186 }, { "epoch": 0.8462110811858776, "grad_norm": 1.9798502922058105, "learning_rate": 6.073161495600993e-07, "loss": 0.5327, "step": 20187 }, { "epoch": 0.8462529998008865, "grad_norm": 2.2543866634368896, "learning_rate": 6.069919194635076e-07, "loss": 0.4925, "step": 20188 }, { "epoch": 0.8462949184158955, "grad_norm": 2.050359010696411, "learning_rate": 6.066677703459472e-07, "loss": 0.5188, "step": 20189 }, { "epoch": 0.8463368370309045, "grad_norm": 2.2270617485046387, "learning_rate": 6.063437022133922e-07, "loss": 0.4444, "step": 20190 }, { "epoch": 0.8463787556459135, "grad_norm": 1.976125717163086, "learning_rate": 6.060197150718184e-07, "loss": 0.5438, "step": 20191 }, { "epoch": 0.8464206742609224, "grad_norm": 1.9603980779647827, "learning_rate": 6.056958089271964e-07, "loss": 0.4478, "step": 20192 }, { "epoch": 0.8464625928759314, "grad_norm": 1.8705604076385498, "learning_rate": 6.05371983785496e-07, "loss": 0.5156, "step": 20193 }, { "epoch": 0.8465045114909403, "grad_norm": 1.7818195819854736, "learning_rate": 6.050482396526875e-07, "loss": 0.477, "step": 20194 }, { "epoch": 0.8465464301059493, "grad_norm": 1.8706302642822266, "learning_rate": 6.047245765347404e-07, "loss": 0.4849, "step": 20195 }, { "epoch": 0.8465883487209582, "grad_norm": 2.0008583068847656, "learning_rate": 6.044009944376184e-07, "loss": 0.5136, "step": 20196 }, { "epoch": 0.8466302673359672, "grad_norm": 1.9802950620651245, "learning_rate": 6.040774933672866e-07, "loss": 0.5557, "step": 20197 }, { "epoch": 0.8466721859509762, "grad_norm": 1.9744701385498047, "learning_rate": 6.03754073329711e-07, "loss": 0.4732, "step": 20198 }, { "epoch": 0.8467141045659852, "grad_norm": 2.1350364685058594, "learning_rate": 6.034307343308499e-07, "loss": 0.5193, "step": 20199 }, { "epoch": 0.8467560231809941, "grad_norm": 1.9390902519226074, "learning_rate": 6.031074763766659e-07, "loss": 0.4632, "step": 20200 }, { "epoch": 0.8467979417960031, "grad_norm": 1.9101076126098633, "learning_rate": 6.027842994731175e-07, "loss": 0.4543, "step": 20201 }, { "epoch": 0.846839860411012, "grad_norm": 2.535000801086426, "learning_rate": 6.024612036261607e-07, "loss": 0.491, "step": 20202 }, { "epoch": 0.846881779026021, "grad_norm": 1.8233486413955688, "learning_rate": 6.021381888417515e-07, "loss": 0.4981, "step": 20203 }, { "epoch": 0.8469236976410299, "grad_norm": 2.087524890899658, "learning_rate": 6.018152551258466e-07, "loss": 0.4616, "step": 20204 }, { "epoch": 0.846965616256039, "grad_norm": 1.9942017793655396, "learning_rate": 6.01492402484396e-07, "loss": 0.5537, "step": 20205 }, { "epoch": 0.8470075348710479, "grad_norm": 2.191577672958374, "learning_rate": 6.011696309233522e-07, "loss": 0.4761, "step": 20206 }, { "epoch": 0.8470494534860569, "grad_norm": 1.942590355873108, "learning_rate": 6.008469404486661e-07, "loss": 0.5077, "step": 20207 }, { "epoch": 0.8470913721010658, "grad_norm": 1.9609627723693848, "learning_rate": 6.005243310662839e-07, "loss": 0.4719, "step": 20208 }, { "epoch": 0.8471332907160748, "grad_norm": 1.6797194480895996, "learning_rate": 6.002018027821543e-07, "loss": 0.4327, "step": 20209 }, { "epoch": 0.8471752093310837, "grad_norm": 1.7938687801361084, "learning_rate": 5.998793556022226e-07, "loss": 0.4826, "step": 20210 }, { "epoch": 0.8472171279460926, "grad_norm": 1.9033105373382568, "learning_rate": 5.995569895324305e-07, "loss": 0.4816, "step": 20211 }, { "epoch": 0.8472590465611016, "grad_norm": 1.7675669193267822, "learning_rate": 5.992347045787217e-07, "loss": 0.5128, "step": 20212 }, { "epoch": 0.8473009651761105, "grad_norm": 2.047494411468506, "learning_rate": 5.989125007470386e-07, "loss": 0.4865, "step": 20213 }, { "epoch": 0.8473428837911195, "grad_norm": 2.0150632858276367, "learning_rate": 5.985903780433183e-07, "loss": 0.4964, "step": 20214 }, { "epoch": 0.8473848024061285, "grad_norm": 2.7013278007507324, "learning_rate": 5.982683364734992e-07, "loss": 0.4441, "step": 20215 }, { "epoch": 0.8474267210211375, "grad_norm": 1.9256932735443115, "learning_rate": 5.979463760435195e-07, "loss": 0.4855, "step": 20216 }, { "epoch": 0.8474686396361464, "grad_norm": 1.898369550704956, "learning_rate": 5.976244967593115e-07, "loss": 0.5435, "step": 20217 }, { "epoch": 0.8475105582511554, "grad_norm": 1.8319275379180908, "learning_rate": 5.97302698626811e-07, "loss": 0.4187, "step": 20218 }, { "epoch": 0.8475524768661643, "grad_norm": 2.3544399738311768, "learning_rate": 5.969809816519484e-07, "loss": 0.4647, "step": 20219 }, { "epoch": 0.8475943954811733, "grad_norm": 2.2246720790863037, "learning_rate": 5.966593458406534e-07, "loss": 0.4757, "step": 20220 }, { "epoch": 0.8476363140961822, "grad_norm": 1.8499705791473389, "learning_rate": 5.96337791198856e-07, "loss": 0.4591, "step": 20221 }, { "epoch": 0.8476782327111912, "grad_norm": 1.8780145645141602, "learning_rate": 5.960163177324852e-07, "loss": 0.4625, "step": 20222 }, { "epoch": 0.8477201513262002, "grad_norm": 2.8701493740081787, "learning_rate": 5.956949254474637e-07, "loss": 0.4731, "step": 20223 }, { "epoch": 0.8477620699412092, "grad_norm": 1.6921366453170776, "learning_rate": 5.953736143497179e-07, "loss": 0.419, "step": 20224 }, { "epoch": 0.8478039885562181, "grad_norm": 1.8552277088165283, "learning_rate": 5.9505238444517e-07, "loss": 0.4634, "step": 20225 }, { "epoch": 0.8478459071712271, "grad_norm": 1.7753123044967651, "learning_rate": 5.947312357397416e-07, "loss": 0.4684, "step": 20226 }, { "epoch": 0.847887825786236, "grad_norm": 2.823385715484619, "learning_rate": 5.944101682393538e-07, "loss": 0.4924, "step": 20227 }, { "epoch": 0.847929744401245, "grad_norm": 1.520349144935608, "learning_rate": 5.940891819499234e-07, "loss": 0.4753, "step": 20228 }, { "epoch": 0.8479716630162539, "grad_norm": 1.732678771018982, "learning_rate": 5.937682768773684e-07, "loss": 0.4958, "step": 20229 }, { "epoch": 0.848013581631263, "grad_norm": 1.746498703956604, "learning_rate": 5.934474530276035e-07, "loss": 0.4465, "step": 20230 }, { "epoch": 0.8480555002462719, "grad_norm": 1.8172606229782104, "learning_rate": 5.931267104065425e-07, "loss": 0.4432, "step": 20231 }, { "epoch": 0.8480974188612809, "grad_norm": 1.700111746788025, "learning_rate": 5.928060490200982e-07, "loss": 0.4656, "step": 20232 }, { "epoch": 0.8481393374762898, "grad_norm": 1.8507122993469238, "learning_rate": 5.924854688741826e-07, "loss": 0.4863, "step": 20233 }, { "epoch": 0.8481812560912988, "grad_norm": 1.851986050605774, "learning_rate": 5.921649699747028e-07, "loss": 0.5216, "step": 20234 }, { "epoch": 0.8482231747063077, "grad_norm": 2.043133497238159, "learning_rate": 5.918445523275684e-07, "loss": 0.4942, "step": 20235 }, { "epoch": 0.8482650933213166, "grad_norm": 1.9161032438278198, "learning_rate": 5.915242159386869e-07, "loss": 0.5165, "step": 20236 }, { "epoch": 0.8483070119363256, "grad_norm": 1.822424054145813, "learning_rate": 5.912039608139608e-07, "loss": 0.4948, "step": 20237 }, { "epoch": 0.8483489305513345, "grad_norm": 2.3273065090179443, "learning_rate": 5.908837869592954e-07, "loss": 0.4824, "step": 20238 }, { "epoch": 0.8483908491663436, "grad_norm": 1.925247073173523, "learning_rate": 5.90563694380592e-07, "loss": 0.4774, "step": 20239 }, { "epoch": 0.8484327677813525, "grad_norm": 2.6369810104370117, "learning_rate": 5.902436830837504e-07, "loss": 0.5028, "step": 20240 }, { "epoch": 0.8484746863963615, "grad_norm": 1.7154291868209839, "learning_rate": 5.899237530746699e-07, "loss": 0.4922, "step": 20241 }, { "epoch": 0.8485166050113704, "grad_norm": 1.9389811754226685, "learning_rate": 5.896039043592494e-07, "loss": 0.5087, "step": 20242 }, { "epoch": 0.8485585236263794, "grad_norm": 3.2408535480499268, "learning_rate": 5.892841369433827e-07, "loss": 0.4786, "step": 20243 }, { "epoch": 0.8486004422413883, "grad_norm": 1.9251912832260132, "learning_rate": 5.889644508329651e-07, "loss": 0.3935, "step": 20244 }, { "epoch": 0.8486423608563973, "grad_norm": 1.7655726671218872, "learning_rate": 5.886448460338912e-07, "loss": 0.5134, "step": 20245 }, { "epoch": 0.8486842794714062, "grad_norm": 1.8282748460769653, "learning_rate": 5.883253225520497e-07, "loss": 0.4691, "step": 20246 }, { "epoch": 0.8487261980864153, "grad_norm": 2.520932674407959, "learning_rate": 5.880058803933331e-07, "loss": 0.4985, "step": 20247 }, { "epoch": 0.8487681167014242, "grad_norm": 1.8707815408706665, "learning_rate": 5.876865195636283e-07, "loss": 0.4468, "step": 20248 }, { "epoch": 0.8488100353164332, "grad_norm": 2.13834810256958, "learning_rate": 5.873672400688224e-07, "loss": 0.4941, "step": 20249 }, { "epoch": 0.8488519539314421, "grad_norm": 1.9177098274230957, "learning_rate": 5.870480419148006e-07, "loss": 0.4602, "step": 20250 }, { "epoch": 0.8488938725464511, "grad_norm": 1.7926784753799438, "learning_rate": 5.867289251074487e-07, "loss": 0.5244, "step": 20251 }, { "epoch": 0.84893579116146, "grad_norm": 2.665449619293213, "learning_rate": 5.864098896526466e-07, "loss": 0.5067, "step": 20252 }, { "epoch": 0.848977709776469, "grad_norm": 1.9904569387435913, "learning_rate": 5.860909355562766e-07, "loss": 0.4738, "step": 20253 }, { "epoch": 0.8490196283914779, "grad_norm": 2.217773914337158, "learning_rate": 5.857720628242191e-07, "loss": 0.5012, "step": 20254 }, { "epoch": 0.849061547006487, "grad_norm": 2.463996171951294, "learning_rate": 5.854532714623507e-07, "loss": 0.4484, "step": 20255 }, { "epoch": 0.8491034656214959, "grad_norm": 2.377588987350464, "learning_rate": 5.851345614765486e-07, "loss": 0.4466, "step": 20256 }, { "epoch": 0.8491453842365049, "grad_norm": 1.7895398139953613, "learning_rate": 5.848159328726882e-07, "loss": 0.5066, "step": 20257 }, { "epoch": 0.8491873028515138, "grad_norm": 2.196549892425537, "learning_rate": 5.844973856566405e-07, "loss": 0.4878, "step": 20258 }, { "epoch": 0.8492292214665228, "grad_norm": 2.2615127563476562, "learning_rate": 5.8417891983428e-07, "loss": 0.5254, "step": 20259 }, { "epoch": 0.8492711400815317, "grad_norm": 1.568990707397461, "learning_rate": 5.838605354114768e-07, "loss": 0.4413, "step": 20260 }, { "epoch": 0.8493130586965406, "grad_norm": 1.743314504623413, "learning_rate": 5.835422323940987e-07, "loss": 0.4432, "step": 20261 }, { "epoch": 0.8493549773115496, "grad_norm": 2.2236390113830566, "learning_rate": 5.832240107880144e-07, "loss": 0.4926, "step": 20262 }, { "epoch": 0.8493968959265585, "grad_norm": 2.223325729370117, "learning_rate": 5.829058705990898e-07, "loss": 0.4757, "step": 20263 }, { "epoch": 0.8494388145415676, "grad_norm": 5.318516254425049, "learning_rate": 5.825878118331885e-07, "loss": 0.4249, "step": 20264 }, { "epoch": 0.8494807331565765, "grad_norm": 2.2978222370147705, "learning_rate": 5.822698344961747e-07, "loss": 0.5048, "step": 20265 }, { "epoch": 0.8495226517715855, "grad_norm": 1.644738793373108, "learning_rate": 5.819519385939093e-07, "loss": 0.4891, "step": 20266 }, { "epoch": 0.8495645703865944, "grad_norm": 1.821885585784912, "learning_rate": 5.816341241322509e-07, "loss": 0.4498, "step": 20267 }, { "epoch": 0.8496064890016034, "grad_norm": 3.240677833557129, "learning_rate": 5.8131639111706e-07, "loss": 0.476, "step": 20268 }, { "epoch": 0.8496484076166123, "grad_norm": 1.7259315252304077, "learning_rate": 5.809987395541932e-07, "loss": 0.4947, "step": 20269 }, { "epoch": 0.8496903262316213, "grad_norm": 2.1622960567474365, "learning_rate": 5.806811694495046e-07, "loss": 0.4985, "step": 20270 }, { "epoch": 0.8497322448466302, "grad_norm": 2.415431499481201, "learning_rate": 5.803636808088486e-07, "loss": 0.4317, "step": 20271 }, { "epoch": 0.8497741634616393, "grad_norm": 1.8993613719940186, "learning_rate": 5.800462736380802e-07, "loss": 0.4665, "step": 20272 }, { "epoch": 0.8498160820766482, "grad_norm": 1.9827966690063477, "learning_rate": 5.797289479430468e-07, "loss": 0.5071, "step": 20273 }, { "epoch": 0.8498580006916572, "grad_norm": 1.9180071353912354, "learning_rate": 5.794117037296004e-07, "loss": 0.4606, "step": 20274 }, { "epoch": 0.8498999193066661, "grad_norm": 1.6247974634170532, "learning_rate": 5.79094541003587e-07, "loss": 0.4769, "step": 20275 }, { "epoch": 0.8499418379216751, "grad_norm": 1.6882739067077637, "learning_rate": 5.787774597708551e-07, "loss": 0.4239, "step": 20276 }, { "epoch": 0.849983756536684, "grad_norm": 1.9256330728530884, "learning_rate": 5.784604600372473e-07, "loss": 0.4827, "step": 20277 }, { "epoch": 0.850025675151693, "grad_norm": 2.003671646118164, "learning_rate": 5.781435418086095e-07, "loss": 0.4469, "step": 20278 }, { "epoch": 0.8500675937667019, "grad_norm": 2.240762233734131, "learning_rate": 5.778267050907815e-07, "loss": 0.4855, "step": 20279 }, { "epoch": 0.850109512381711, "grad_norm": 1.884847640991211, "learning_rate": 5.775099498896048e-07, "loss": 0.509, "step": 20280 }, { "epoch": 0.8501514309967199, "grad_norm": 2.130392074584961, "learning_rate": 5.77193276210919e-07, "loss": 0.4948, "step": 20281 }, { "epoch": 0.8501933496117289, "grad_norm": 1.9455020427703857, "learning_rate": 5.7687668406056e-07, "loss": 0.5318, "step": 20282 }, { "epoch": 0.8502352682267378, "grad_norm": 2.147966146469116, "learning_rate": 5.765601734443644e-07, "loss": 0.5047, "step": 20283 }, { "epoch": 0.8502771868417468, "grad_norm": 2.2024574279785156, "learning_rate": 5.762437443681679e-07, "loss": 0.4222, "step": 20284 }, { "epoch": 0.8503191054567557, "grad_norm": 1.9738526344299316, "learning_rate": 5.759273968378026e-07, "loss": 0.4945, "step": 20285 }, { "epoch": 0.8503610240717646, "grad_norm": 2.3220152854919434, "learning_rate": 5.756111308590984e-07, "loss": 0.4372, "step": 20286 }, { "epoch": 0.8504029426867736, "grad_norm": 1.7680482864379883, "learning_rate": 5.75294946437887e-07, "loss": 0.4504, "step": 20287 }, { "epoch": 0.8504448613017825, "grad_norm": 1.7720941305160522, "learning_rate": 5.749788435799958e-07, "loss": 0.5394, "step": 20288 }, { "epoch": 0.8504867799167916, "grad_norm": 1.8008534908294678, "learning_rate": 5.746628222912521e-07, "loss": 0.4525, "step": 20289 }, { "epoch": 0.8505286985318005, "grad_norm": 1.8909465074539185, "learning_rate": 5.743468825774823e-07, "loss": 0.4758, "step": 20290 }, { "epoch": 0.8505706171468095, "grad_norm": 2.365420341491699, "learning_rate": 5.740310244445085e-07, "loss": 0.4749, "step": 20291 }, { "epoch": 0.8506125357618184, "grad_norm": 2.6000871658325195, "learning_rate": 5.737152478981545e-07, "loss": 0.5069, "step": 20292 }, { "epoch": 0.8506544543768274, "grad_norm": 2.218358278274536, "learning_rate": 5.733995529442416e-07, "loss": 0.4674, "step": 20293 }, { "epoch": 0.8506963729918363, "grad_norm": 1.9380204677581787, "learning_rate": 5.730839395885879e-07, "loss": 0.4926, "step": 20294 }, { "epoch": 0.8507382916068453, "grad_norm": 2.446140766143799, "learning_rate": 5.727684078370111e-07, "loss": 0.4572, "step": 20295 }, { "epoch": 0.8507802102218542, "grad_norm": 2.091403007507324, "learning_rate": 5.724529576953292e-07, "loss": 0.4907, "step": 20296 }, { "epoch": 0.8508221288368633, "grad_norm": 1.8565244674682617, "learning_rate": 5.721375891693548e-07, "loss": 0.4513, "step": 20297 }, { "epoch": 0.8508640474518722, "grad_norm": 2.2205753326416016, "learning_rate": 5.718223022649028e-07, "loss": 0.473, "step": 20298 }, { "epoch": 0.8509059660668812, "grad_norm": 1.7565913200378418, "learning_rate": 5.715070969877862e-07, "loss": 0.5363, "step": 20299 }, { "epoch": 0.8509478846818901, "grad_norm": 2.792964220046997, "learning_rate": 5.711919733438126e-07, "loss": 0.4601, "step": 20300 }, { "epoch": 0.8509898032968991, "grad_norm": 1.697209358215332, "learning_rate": 5.708769313387936e-07, "loss": 0.4341, "step": 20301 }, { "epoch": 0.851031721911908, "grad_norm": 2.3682773113250732, "learning_rate": 5.705619709785343e-07, "loss": 0.4992, "step": 20302 }, { "epoch": 0.851073640526917, "grad_norm": 1.7984369993209839, "learning_rate": 5.702470922688424e-07, "loss": 0.4652, "step": 20303 }, { "epoch": 0.8511155591419259, "grad_norm": 3.4206981658935547, "learning_rate": 5.699322952155212e-07, "loss": 0.4485, "step": 20304 }, { "epoch": 0.851157477756935, "grad_norm": 1.9184931516647339, "learning_rate": 5.696175798243731e-07, "loss": 0.4667, "step": 20305 }, { "epoch": 0.8511993963719439, "grad_norm": 1.551801323890686, "learning_rate": 5.693029461012001e-07, "loss": 0.4528, "step": 20306 }, { "epoch": 0.8512413149869529, "grad_norm": 1.8920327425003052, "learning_rate": 5.689883940518026e-07, "loss": 0.4566, "step": 20307 }, { "epoch": 0.8512832336019618, "grad_norm": 1.571916103363037, "learning_rate": 5.686739236819777e-07, "loss": 0.486, "step": 20308 }, { "epoch": 0.8513251522169708, "grad_norm": 1.7546340227127075, "learning_rate": 5.68359534997523e-07, "loss": 0.4423, "step": 20309 }, { "epoch": 0.8513670708319797, "grad_norm": 1.938881278038025, "learning_rate": 5.680452280042348e-07, "loss": 0.4409, "step": 20310 }, { "epoch": 0.8514089894469886, "grad_norm": 2.1093180179595947, "learning_rate": 5.677310027079047e-07, "loss": 0.4661, "step": 20311 }, { "epoch": 0.8514509080619976, "grad_norm": 2.4179160594940186, "learning_rate": 5.67416859114327e-07, "loss": 0.4234, "step": 20312 }, { "epoch": 0.8514928266770065, "grad_norm": 1.8053412437438965, "learning_rate": 5.67102797229292e-07, "loss": 0.4913, "step": 20313 }, { "epoch": 0.8515347452920156, "grad_norm": 1.8632526397705078, "learning_rate": 5.667888170585873e-07, "loss": 0.4879, "step": 20314 }, { "epoch": 0.8515766639070245, "grad_norm": 1.9798007011413574, "learning_rate": 5.664749186080021e-07, "loss": 0.5173, "step": 20315 }, { "epoch": 0.8516185825220335, "grad_norm": 2.0426769256591797, "learning_rate": 5.661611018833236e-07, "loss": 0.4303, "step": 20316 }, { "epoch": 0.8516605011370424, "grad_norm": 1.9704058170318604, "learning_rate": 5.658473668903341e-07, "loss": 0.5468, "step": 20317 }, { "epoch": 0.8517024197520514, "grad_norm": 1.8369777202606201, "learning_rate": 5.655337136348194e-07, "loss": 0.4594, "step": 20318 }, { "epoch": 0.8517443383670603, "grad_norm": 1.9587634801864624, "learning_rate": 5.652201421225606e-07, "loss": 0.4778, "step": 20319 }, { "epoch": 0.8517862569820693, "grad_norm": 1.7481228113174438, "learning_rate": 5.649066523593367e-07, "loss": 0.4598, "step": 20320 }, { "epoch": 0.8518281755970782, "grad_norm": 1.8407601118087769, "learning_rate": 5.645932443509283e-07, "loss": 0.4582, "step": 20321 }, { "epoch": 0.8518700942120873, "grad_norm": 2.4627797603607178, "learning_rate": 5.642799181031105e-07, "loss": 0.4954, "step": 20322 }, { "epoch": 0.8519120128270962, "grad_norm": 1.6059201955795288, "learning_rate": 5.639666736216615e-07, "loss": 0.4712, "step": 20323 }, { "epoch": 0.8519539314421052, "grad_norm": 1.7582170963287354, "learning_rate": 5.636535109123531e-07, "loss": 0.4685, "step": 20324 }, { "epoch": 0.8519958500571141, "grad_norm": 1.9592605829238892, "learning_rate": 5.633404299809597e-07, "loss": 0.4879, "step": 20325 }, { "epoch": 0.8520377686721231, "grad_norm": 2.872612476348877, "learning_rate": 5.630274308332517e-07, "loss": 0.5127, "step": 20326 }, { "epoch": 0.852079687287132, "grad_norm": 2.381974220275879, "learning_rate": 5.62714513474999e-07, "loss": 0.432, "step": 20327 }, { "epoch": 0.852121605902141, "grad_norm": 1.7175979614257812, "learning_rate": 5.624016779119707e-07, "loss": 0.5089, "step": 20328 }, { "epoch": 0.8521635245171499, "grad_norm": 2.1851186752319336, "learning_rate": 5.620889241499316e-07, "loss": 0.5444, "step": 20329 }, { "epoch": 0.852205443132159, "grad_norm": 2.1664202213287354, "learning_rate": 5.61776252194648e-07, "loss": 0.4798, "step": 20330 }, { "epoch": 0.8522473617471679, "grad_norm": 3.7473251819610596, "learning_rate": 5.614636620518843e-07, "loss": 0.5256, "step": 20331 }, { "epoch": 0.8522892803621769, "grad_norm": 6.492257595062256, "learning_rate": 5.611511537274022e-07, "loss": 0.4706, "step": 20332 }, { "epoch": 0.8523311989771858, "grad_norm": 1.9550141096115112, "learning_rate": 5.608387272269611e-07, "loss": 0.464, "step": 20333 }, { "epoch": 0.8523731175921948, "grad_norm": 1.8505191802978516, "learning_rate": 5.605263825563223e-07, "loss": 0.5387, "step": 20334 }, { "epoch": 0.8524150362072037, "grad_norm": 2.2767062187194824, "learning_rate": 5.602141197212407e-07, "loss": 0.5163, "step": 20335 }, { "epoch": 0.8524569548222126, "grad_norm": 2.2621073722839355, "learning_rate": 5.599019387274745e-07, "loss": 0.515, "step": 20336 }, { "epoch": 0.8524988734372216, "grad_norm": 1.776537537574768, "learning_rate": 5.595898395807786e-07, "loss": 0.4668, "step": 20337 }, { "epoch": 0.8525407920522305, "grad_norm": 1.7178839445114136, "learning_rate": 5.592778222869045e-07, "loss": 0.4357, "step": 20338 }, { "epoch": 0.8525827106672396, "grad_norm": 1.8952970504760742, "learning_rate": 5.589658868516041e-07, "loss": 0.5321, "step": 20339 }, { "epoch": 0.8526246292822485, "grad_norm": 1.837751865386963, "learning_rate": 5.586540332806295e-07, "loss": 0.4973, "step": 20340 }, { "epoch": 0.8526665478972575, "grad_norm": 2.166736364364624, "learning_rate": 5.583422615797274e-07, "loss": 0.4953, "step": 20341 }, { "epoch": 0.8527084665122664, "grad_norm": 1.8440542221069336, "learning_rate": 5.580305717546447e-07, "loss": 0.4796, "step": 20342 }, { "epoch": 0.8527503851272754, "grad_norm": 3.5930936336517334, "learning_rate": 5.577189638111285e-07, "loss": 0.499, "step": 20343 }, { "epoch": 0.8527923037422843, "grad_norm": 2.212810754776001, "learning_rate": 5.57407437754921e-07, "loss": 0.4983, "step": 20344 }, { "epoch": 0.8528342223572933, "grad_norm": 1.740793228149414, "learning_rate": 5.570959935917653e-07, "loss": 0.436, "step": 20345 }, { "epoch": 0.8528761409723022, "grad_norm": 1.873205304145813, "learning_rate": 5.567846313274039e-07, "loss": 0.5265, "step": 20346 }, { "epoch": 0.8529180595873113, "grad_norm": 2.442131996154785, "learning_rate": 5.564733509675746e-07, "loss": 0.5226, "step": 20347 }, { "epoch": 0.8529599782023202, "grad_norm": 2.2830705642700195, "learning_rate": 5.561621525180161e-07, "loss": 0.4523, "step": 20348 }, { "epoch": 0.8530018968173292, "grad_norm": 1.6482187509536743, "learning_rate": 5.558510359844654e-07, "loss": 0.4298, "step": 20349 }, { "epoch": 0.8530438154323381, "grad_norm": 2.169891119003296, "learning_rate": 5.555400013726575e-07, "loss": 0.4134, "step": 20350 }, { "epoch": 0.8530857340473471, "grad_norm": 1.961201786994934, "learning_rate": 5.552290486883244e-07, "loss": 0.4539, "step": 20351 }, { "epoch": 0.853127652662356, "grad_norm": 1.6012868881225586, "learning_rate": 5.549181779371999e-07, "loss": 0.4071, "step": 20352 }, { "epoch": 0.853169571277365, "grad_norm": 1.8240867853164673, "learning_rate": 5.546073891250125e-07, "loss": 0.4592, "step": 20353 }, { "epoch": 0.8532114898923739, "grad_norm": 2.9101247787475586, "learning_rate": 5.542966822574925e-07, "loss": 0.4657, "step": 20354 }, { "epoch": 0.853253408507383, "grad_norm": 2.161036968231201, "learning_rate": 5.539860573403683e-07, "loss": 0.4751, "step": 20355 }, { "epoch": 0.8532953271223919, "grad_norm": 2.661985158920288, "learning_rate": 5.536755143793632e-07, "loss": 0.4943, "step": 20356 }, { "epoch": 0.8533372457374009, "grad_norm": 2.054111957550049, "learning_rate": 5.533650533802037e-07, "loss": 0.4465, "step": 20357 }, { "epoch": 0.8533791643524098, "grad_norm": 1.8341389894485474, "learning_rate": 5.530546743486132e-07, "loss": 0.4651, "step": 20358 }, { "epoch": 0.8534210829674188, "grad_norm": 1.9271565675735474, "learning_rate": 5.527443772903113e-07, "loss": 0.4502, "step": 20359 }, { "epoch": 0.8534630015824277, "grad_norm": 2.000960111618042, "learning_rate": 5.524341622110185e-07, "loss": 0.491, "step": 20360 }, { "epoch": 0.8535049201974366, "grad_norm": 1.9929403066635132, "learning_rate": 5.521240291164537e-07, "loss": 0.5227, "step": 20361 }, { "epoch": 0.8535468388124456, "grad_norm": 1.806086778640747, "learning_rate": 5.518139780123327e-07, "loss": 0.4759, "step": 20362 }, { "epoch": 0.8535887574274545, "grad_norm": 2.2378451824188232, "learning_rate": 5.515040089043716e-07, "loss": 0.4747, "step": 20363 }, { "epoch": 0.8536306760424636, "grad_norm": 2.6561667919158936, "learning_rate": 5.511941217982852e-07, "loss": 0.4753, "step": 20364 }, { "epoch": 0.8536725946574725, "grad_norm": 3.1084132194519043, "learning_rate": 5.508843166997841e-07, "loss": 0.4414, "step": 20365 }, { "epoch": 0.8537145132724815, "grad_norm": 1.787210464477539, "learning_rate": 5.505745936145795e-07, "loss": 0.4675, "step": 20366 }, { "epoch": 0.8537564318874904, "grad_norm": 1.9668309688568115, "learning_rate": 5.502649525483822e-07, "loss": 0.4615, "step": 20367 }, { "epoch": 0.8537983505024994, "grad_norm": 1.709344744682312, "learning_rate": 5.49955393506899e-07, "loss": 0.4651, "step": 20368 }, { "epoch": 0.8538402691175083, "grad_norm": 2.047708511352539, "learning_rate": 5.496459164958351e-07, "loss": 0.4713, "step": 20369 }, { "epoch": 0.8538821877325173, "grad_norm": 1.8946269750595093, "learning_rate": 5.493365215208973e-07, "loss": 0.5089, "step": 20370 }, { "epoch": 0.8539241063475262, "grad_norm": 2.13723087310791, "learning_rate": 5.490272085877874e-07, "loss": 0.5159, "step": 20371 }, { "epoch": 0.8539660249625353, "grad_norm": 1.8561028242111206, "learning_rate": 5.487179777022066e-07, "loss": 0.5111, "step": 20372 }, { "epoch": 0.8540079435775442, "grad_norm": 2.100940704345703, "learning_rate": 5.48408828869858e-07, "loss": 0.4919, "step": 20373 }, { "epoch": 0.8540498621925532, "grad_norm": 2.08805775642395, "learning_rate": 5.480997620964373e-07, "loss": 0.4445, "step": 20374 }, { "epoch": 0.8540917808075621, "grad_norm": 1.8921551704406738, "learning_rate": 5.47790777387644e-07, "loss": 0.4742, "step": 20375 }, { "epoch": 0.8541336994225711, "grad_norm": 2.0879409313201904, "learning_rate": 5.474818747491712e-07, "loss": 0.5104, "step": 20376 }, { "epoch": 0.85417561803758, "grad_norm": 2.009453773498535, "learning_rate": 5.471730541867154e-07, "loss": 0.467, "step": 20377 }, { "epoch": 0.854217536652589, "grad_norm": 1.6125826835632324, "learning_rate": 5.468643157059689e-07, "loss": 0.4456, "step": 20378 }, { "epoch": 0.854259455267598, "grad_norm": 1.7523243427276611, "learning_rate": 5.465556593126226e-07, "loss": 0.4573, "step": 20379 }, { "epoch": 0.854301373882607, "grad_norm": 1.881807804107666, "learning_rate": 5.462470850123653e-07, "loss": 0.418, "step": 20380 }, { "epoch": 0.8543432924976159, "grad_norm": 2.2323153018951416, "learning_rate": 5.459385928108862e-07, "loss": 0.4559, "step": 20381 }, { "epoch": 0.8543852111126249, "grad_norm": 1.9628134965896606, "learning_rate": 5.45630182713871e-07, "loss": 0.5066, "step": 20382 }, { "epoch": 0.8544271297276338, "grad_norm": 2.017657518386841, "learning_rate": 5.453218547270056e-07, "loss": 0.5073, "step": 20383 }, { "epoch": 0.8544690483426428, "grad_norm": 1.7833547592163086, "learning_rate": 5.450136088559743e-07, "loss": 0.4511, "step": 20384 }, { "epoch": 0.8545109669576517, "grad_norm": 1.7833809852600098, "learning_rate": 5.44705445106457e-07, "loss": 0.4464, "step": 20385 }, { "epoch": 0.8545528855726606, "grad_norm": 2.554395914077759, "learning_rate": 5.44397363484136e-07, "loss": 0.5088, "step": 20386 }, { "epoch": 0.8545948041876696, "grad_norm": 2.074772357940674, "learning_rate": 5.440893639946903e-07, "loss": 0.4758, "step": 20387 }, { "epoch": 0.8546367228026786, "grad_norm": 2.0660417079925537, "learning_rate": 5.437814466437974e-07, "loss": 0.4801, "step": 20388 }, { "epoch": 0.8546786414176876, "grad_norm": 1.8939965963363647, "learning_rate": 5.434736114371325e-07, "loss": 0.4639, "step": 20389 }, { "epoch": 0.8547205600326965, "grad_norm": 1.9011927843093872, "learning_rate": 5.431658583803706e-07, "loss": 0.4708, "step": 20390 }, { "epoch": 0.8547624786477055, "grad_norm": 2.147998809814453, "learning_rate": 5.428581874791844e-07, "loss": 0.4523, "step": 20391 }, { "epoch": 0.8548043972627144, "grad_norm": 1.730385422706604, "learning_rate": 5.425505987392454e-07, "loss": 0.4467, "step": 20392 }, { "epoch": 0.8548463158777234, "grad_norm": 1.9186650514602661, "learning_rate": 5.422430921662252e-07, "loss": 0.4932, "step": 20393 }, { "epoch": 0.8548882344927323, "grad_norm": 1.6090478897094727, "learning_rate": 5.419356677657899e-07, "loss": 0.453, "step": 20394 }, { "epoch": 0.8549301531077413, "grad_norm": 1.765721321105957, "learning_rate": 5.416283255436072e-07, "loss": 0.4324, "step": 20395 }, { "epoch": 0.8549720717227502, "grad_norm": 1.8886799812316895, "learning_rate": 5.413210655053442e-07, "loss": 0.4757, "step": 20396 }, { "epoch": 0.8550139903377593, "grad_norm": 1.7145131826400757, "learning_rate": 5.410138876566629e-07, "loss": 0.477, "step": 20397 }, { "epoch": 0.8550559089527682, "grad_norm": 1.93833327293396, "learning_rate": 5.407067920032255e-07, "loss": 0.4951, "step": 20398 }, { "epoch": 0.8550978275677772, "grad_norm": 2.2394564151763916, "learning_rate": 5.403997785506949e-07, "loss": 0.4107, "step": 20399 }, { "epoch": 0.8551397461827861, "grad_norm": 1.7269225120544434, "learning_rate": 5.400928473047273e-07, "loss": 0.4397, "step": 20400 }, { "epoch": 0.8551816647977951, "grad_norm": 2.022841215133667, "learning_rate": 5.39785998270983e-07, "loss": 0.4689, "step": 20401 }, { "epoch": 0.855223583412804, "grad_norm": 2.448988437652588, "learning_rate": 5.39479231455119e-07, "loss": 0.4207, "step": 20402 }, { "epoch": 0.855265502027813, "grad_norm": 1.9887298345565796, "learning_rate": 5.391725468627873e-07, "loss": 0.5149, "step": 20403 }, { "epoch": 0.855307420642822, "grad_norm": 1.99930739402771, "learning_rate": 5.388659444996436e-07, "loss": 0.468, "step": 20404 }, { "epoch": 0.855349339257831, "grad_norm": 2.0396337509155273, "learning_rate": 5.385594243713394e-07, "loss": 0.5397, "step": 20405 }, { "epoch": 0.8553912578728399, "grad_norm": 1.5926947593688965, "learning_rate": 5.38252986483524e-07, "loss": 0.4434, "step": 20406 }, { "epoch": 0.8554331764878489, "grad_norm": 1.7407915592193604, "learning_rate": 5.379466308418463e-07, "loss": 0.4522, "step": 20407 }, { "epoch": 0.8554750951028578, "grad_norm": 2.0146968364715576, "learning_rate": 5.376403574519546e-07, "loss": 0.4714, "step": 20408 }, { "epoch": 0.8555170137178668, "grad_norm": 1.835121750831604, "learning_rate": 5.373341663194925e-07, "loss": 0.4992, "step": 20409 }, { "epoch": 0.8555589323328757, "grad_norm": 1.8973867893218994, "learning_rate": 5.370280574501064e-07, "loss": 0.4677, "step": 20410 }, { "epoch": 0.8556008509478846, "grad_norm": 2.898041248321533, "learning_rate": 5.367220308494386e-07, "loss": 0.4803, "step": 20411 }, { "epoch": 0.8556427695628936, "grad_norm": 2.085045337677002, "learning_rate": 5.364160865231288e-07, "loss": 0.5129, "step": 20412 }, { "epoch": 0.8556846881779026, "grad_norm": 1.9539374113082886, "learning_rate": 5.361102244768185e-07, "loss": 0.5173, "step": 20413 }, { "epoch": 0.8557266067929116, "grad_norm": 1.8090057373046875, "learning_rate": 5.358044447161453e-07, "loss": 0.4827, "step": 20414 }, { "epoch": 0.8557685254079205, "grad_norm": 1.670097827911377, "learning_rate": 5.35498747246746e-07, "loss": 0.4404, "step": 20415 }, { "epoch": 0.8558104440229295, "grad_norm": 1.8889433145523071, "learning_rate": 5.351931320742537e-07, "loss": 0.4208, "step": 20416 }, { "epoch": 0.8558523626379384, "grad_norm": 2.141364097595215, "learning_rate": 5.348875992043052e-07, "loss": 0.5006, "step": 20417 }, { "epoch": 0.8558942812529474, "grad_norm": 1.7916110754013062, "learning_rate": 5.345821486425302e-07, "loss": 0.4924, "step": 20418 }, { "epoch": 0.8559361998679563, "grad_norm": 2.6717209815979004, "learning_rate": 5.342767803945592e-07, "loss": 0.5141, "step": 20419 }, { "epoch": 0.8559781184829653, "grad_norm": 1.874018669128418, "learning_rate": 5.339714944660235e-07, "loss": 0.4824, "step": 20420 }, { "epoch": 0.8560200370979743, "grad_norm": 1.7833517789840698, "learning_rate": 5.336662908625484e-07, "loss": 0.4967, "step": 20421 }, { "epoch": 0.8560619557129833, "grad_norm": 1.6505346298217773, "learning_rate": 5.333611695897606e-07, "loss": 0.5374, "step": 20422 }, { "epoch": 0.8561038743279922, "grad_norm": 2.0130538940429688, "learning_rate": 5.330561306532855e-07, "loss": 0.4505, "step": 20423 }, { "epoch": 0.8561457929430012, "grad_norm": 1.6909271478652954, "learning_rate": 5.327511740587443e-07, "loss": 0.4402, "step": 20424 }, { "epoch": 0.8561877115580101, "grad_norm": 2.0305426120758057, "learning_rate": 5.324462998117608e-07, "loss": 0.4899, "step": 20425 }, { "epoch": 0.8562296301730191, "grad_norm": 1.7653558254241943, "learning_rate": 5.321415079179532e-07, "loss": 0.4361, "step": 20426 }, { "epoch": 0.856271548788028, "grad_norm": 2.5640852451324463, "learning_rate": 5.318367983829393e-07, "loss": 0.4831, "step": 20427 }, { "epoch": 0.856313467403037, "grad_norm": 1.6148029565811157, "learning_rate": 5.315321712123367e-07, "loss": 0.4506, "step": 20428 }, { "epoch": 0.856355386018046, "grad_norm": 1.8210151195526123, "learning_rate": 5.312276264117628e-07, "loss": 0.5503, "step": 20429 }, { "epoch": 0.856397304633055, "grad_norm": 3.190363645553589, "learning_rate": 5.309231639868278e-07, "loss": 0.4921, "step": 20430 }, { "epoch": 0.8564392232480639, "grad_norm": 2.1389598846435547, "learning_rate": 5.306187839431465e-07, "loss": 0.5405, "step": 20431 }, { "epoch": 0.8564811418630729, "grad_norm": 1.9929929971694946, "learning_rate": 5.303144862863302e-07, "loss": 0.4569, "step": 20432 }, { "epoch": 0.8565230604780818, "grad_norm": 1.8511390686035156, "learning_rate": 5.300102710219862e-07, "loss": 0.4877, "step": 20433 }, { "epoch": 0.8565649790930908, "grad_norm": 2.1319870948791504, "learning_rate": 5.29706138155724e-07, "loss": 0.5748, "step": 20434 }, { "epoch": 0.8566068977080997, "grad_norm": 1.5907766819000244, "learning_rate": 5.294020876931494e-07, "loss": 0.442, "step": 20435 }, { "epoch": 0.8566488163231086, "grad_norm": 1.7057137489318848, "learning_rate": 5.290981196398659e-07, "loss": 0.5013, "step": 20436 }, { "epoch": 0.8566907349381176, "grad_norm": 3.614452362060547, "learning_rate": 5.287942340014779e-07, "loss": 0.4814, "step": 20437 }, { "epoch": 0.8567326535531266, "grad_norm": 1.8360071182250977, "learning_rate": 5.284904307835875e-07, "loss": 0.4487, "step": 20438 }, { "epoch": 0.8567745721681356, "grad_norm": 1.723222017288208, "learning_rate": 5.281867099917937e-07, "loss": 0.4507, "step": 20439 }, { "epoch": 0.8568164907831445, "grad_norm": 3.8096587657928467, "learning_rate": 5.278830716316957e-07, "loss": 0.4399, "step": 20440 }, { "epoch": 0.8568584093981535, "grad_norm": 2.0544140338897705, "learning_rate": 5.275795157088915e-07, "loss": 0.4927, "step": 20441 }, { "epoch": 0.8569003280131624, "grad_norm": 2.194823741912842, "learning_rate": 5.272760422289752e-07, "loss": 0.4925, "step": 20442 }, { "epoch": 0.8569422466281714, "grad_norm": 1.6687564849853516, "learning_rate": 5.269726511975426e-07, "loss": 0.4662, "step": 20443 }, { "epoch": 0.8569841652431803, "grad_norm": 1.898376703262329, "learning_rate": 5.266693426201858e-07, "loss": 0.4419, "step": 20444 }, { "epoch": 0.8570260838581893, "grad_norm": 1.6431652307510376, "learning_rate": 5.263661165024941e-07, "loss": 0.464, "step": 20445 }, { "epoch": 0.8570680024731983, "grad_norm": 1.8077653646469116, "learning_rate": 5.260629728500582e-07, "loss": 0.4825, "step": 20446 }, { "epoch": 0.8571099210882073, "grad_norm": 3.183640718460083, "learning_rate": 5.257599116684676e-07, "loss": 0.4716, "step": 20447 }, { "epoch": 0.8571518397032162, "grad_norm": 1.8277455568313599, "learning_rate": 5.254569329633069e-07, "loss": 0.4506, "step": 20448 }, { "epoch": 0.8571937583182252, "grad_norm": 2.0916082859039307, "learning_rate": 5.251540367401625e-07, "loss": 0.3937, "step": 20449 }, { "epoch": 0.8572356769332341, "grad_norm": 3.289271354675293, "learning_rate": 5.24851223004616e-07, "loss": 0.4332, "step": 20450 }, { "epoch": 0.8572775955482431, "grad_norm": 2.1054275035858154, "learning_rate": 5.245484917622506e-07, "loss": 0.4933, "step": 20451 }, { "epoch": 0.857319514163252, "grad_norm": 1.7632966041564941, "learning_rate": 5.242458430186476e-07, "loss": 0.4686, "step": 20452 }, { "epoch": 0.857361432778261, "grad_norm": 2.054180383682251, "learning_rate": 5.23943276779385e-07, "loss": 0.5228, "step": 20453 }, { "epoch": 0.85740335139327, "grad_norm": 1.365748405456543, "learning_rate": 5.236407930500392e-07, "loss": 0.4805, "step": 20454 }, { "epoch": 0.857445270008279, "grad_norm": 1.6550887823104858, "learning_rate": 5.233383918361878e-07, "loss": 0.4517, "step": 20455 }, { "epoch": 0.8574871886232879, "grad_norm": 1.8415374755859375, "learning_rate": 5.230360731434036e-07, "loss": 0.4684, "step": 20456 }, { "epoch": 0.8575291072382969, "grad_norm": 2.065861701965332, "learning_rate": 5.2273383697726e-07, "loss": 0.4427, "step": 20457 }, { "epoch": 0.8575710258533058, "grad_norm": 2.1170432567596436, "learning_rate": 5.224316833433296e-07, "loss": 0.4234, "step": 20458 }, { "epoch": 0.8576129444683148, "grad_norm": 2.0109100341796875, "learning_rate": 5.221296122471803e-07, "loss": 0.4893, "step": 20459 }, { "epoch": 0.8576548630833237, "grad_norm": 2.1348719596862793, "learning_rate": 5.218276236943808e-07, "loss": 0.4869, "step": 20460 }, { "epoch": 0.8576967816983326, "grad_norm": 2.1860499382019043, "learning_rate": 5.215257176904992e-07, "loss": 0.5264, "step": 20461 }, { "epoch": 0.8577387003133417, "grad_norm": 2.4382519721984863, "learning_rate": 5.212238942410986e-07, "loss": 0.4711, "step": 20462 }, { "epoch": 0.8577806189283506, "grad_norm": 1.6841918230056763, "learning_rate": 5.20922153351745e-07, "loss": 0.4588, "step": 20463 }, { "epoch": 0.8578225375433596, "grad_norm": 2.031715154647827, "learning_rate": 5.206204950279992e-07, "loss": 0.5091, "step": 20464 }, { "epoch": 0.8578644561583685, "grad_norm": 2.6938886642456055, "learning_rate": 5.20318919275421e-07, "loss": 0.4506, "step": 20465 }, { "epoch": 0.8579063747733775, "grad_norm": 1.7702327966690063, "learning_rate": 5.200174260995705e-07, "loss": 0.5241, "step": 20466 }, { "epoch": 0.8579482933883864, "grad_norm": 1.8740397691726685, "learning_rate": 5.197160155060066e-07, "loss": 0.4667, "step": 20467 }, { "epoch": 0.8579902120033954, "grad_norm": 1.9072778224945068, "learning_rate": 5.194146875002831e-07, "loss": 0.517, "step": 20468 }, { "epoch": 0.8580321306184043, "grad_norm": 1.9704840183258057, "learning_rate": 5.191134420879557e-07, "loss": 0.5035, "step": 20469 }, { "epoch": 0.8580740492334133, "grad_norm": 2.410569906234741, "learning_rate": 5.188122792745781e-07, "loss": 0.5225, "step": 20470 }, { "epoch": 0.8581159678484223, "grad_norm": 1.9460777044296265, "learning_rate": 5.185111990657e-07, "loss": 0.487, "step": 20471 }, { "epoch": 0.8581578864634313, "grad_norm": 1.9494798183441162, "learning_rate": 5.182102014668739e-07, "loss": 0.5051, "step": 20472 }, { "epoch": 0.8581998050784402, "grad_norm": 1.876712441444397, "learning_rate": 5.179092864836466e-07, "loss": 0.5221, "step": 20473 }, { "epoch": 0.8582417236934492, "grad_norm": 2.1337296962738037, "learning_rate": 5.176084541215643e-07, "loss": 0.4968, "step": 20474 }, { "epoch": 0.8582836423084581, "grad_norm": 1.7949193716049194, "learning_rate": 5.173077043861729e-07, "loss": 0.4475, "step": 20475 }, { "epoch": 0.8583255609234671, "grad_norm": 1.9702852964401245, "learning_rate": 5.170070372830188e-07, "loss": 0.4986, "step": 20476 }, { "epoch": 0.858367479538476, "grad_norm": 1.6152453422546387, "learning_rate": 5.167064528176408e-07, "loss": 0.5148, "step": 20477 }, { "epoch": 0.858409398153485, "grad_norm": 2.032151460647583, "learning_rate": 5.164059509955816e-07, "loss": 0.5311, "step": 20478 }, { "epoch": 0.858451316768494, "grad_norm": 1.940498948097229, "learning_rate": 5.161055318223812e-07, "loss": 0.5076, "step": 20479 }, { "epoch": 0.858493235383503, "grad_norm": 1.559847116470337, "learning_rate": 5.158051953035753e-07, "loss": 0.4316, "step": 20480 }, { "epoch": 0.8585351539985119, "grad_norm": 2.355506420135498, "learning_rate": 5.155049414447028e-07, "loss": 0.4556, "step": 20481 }, { "epoch": 0.8585770726135209, "grad_norm": 1.9089139699935913, "learning_rate": 5.152047702512969e-07, "loss": 0.4063, "step": 20482 }, { "epoch": 0.8586189912285298, "grad_norm": 2.4612486362457275, "learning_rate": 5.149046817288906e-07, "loss": 0.4145, "step": 20483 }, { "epoch": 0.8586609098435388, "grad_norm": 1.7562792301177979, "learning_rate": 5.146046758830154e-07, "loss": 0.4771, "step": 20484 }, { "epoch": 0.8587028284585477, "grad_norm": 2.2351315021514893, "learning_rate": 5.143047527192029e-07, "loss": 0.4497, "step": 20485 }, { "epoch": 0.8587447470735566, "grad_norm": 1.7997057437896729, "learning_rate": 5.140049122429802e-07, "loss": 0.4456, "step": 20486 }, { "epoch": 0.8587866656885657, "grad_norm": 3.6702053546905518, "learning_rate": 5.137051544598759e-07, "loss": 0.5095, "step": 20487 }, { "epoch": 0.8588285843035746, "grad_norm": 1.6941519975662231, "learning_rate": 5.134054793754151e-07, "loss": 0.4976, "step": 20488 }, { "epoch": 0.8588705029185836, "grad_norm": 1.9267160892486572, "learning_rate": 5.131058869951211e-07, "loss": 0.4559, "step": 20489 }, { "epoch": 0.8589124215335925, "grad_norm": 1.665431022644043, "learning_rate": 5.128063773245185e-07, "loss": 0.4359, "step": 20490 }, { "epoch": 0.8589543401486015, "grad_norm": 1.7226715087890625, "learning_rate": 5.12506950369126e-07, "loss": 0.4364, "step": 20491 }, { "epoch": 0.8589962587636104, "grad_norm": 1.5750495195388794, "learning_rate": 5.122076061344638e-07, "loss": 0.4679, "step": 20492 }, { "epoch": 0.8590381773786194, "grad_norm": 2.67008900642395, "learning_rate": 5.119083446260503e-07, "loss": 0.491, "step": 20493 }, { "epoch": 0.8590800959936283, "grad_norm": 1.8209633827209473, "learning_rate": 5.116091658494026e-07, "loss": 0.41, "step": 20494 }, { "epoch": 0.8591220146086374, "grad_norm": 1.7166666984558105, "learning_rate": 5.113100698100337e-07, "loss": 0.3946, "step": 20495 }, { "epoch": 0.8591639332236463, "grad_norm": 3.7503767013549805, "learning_rate": 5.11011056513459e-07, "loss": 0.4861, "step": 20496 }, { "epoch": 0.8592058518386553, "grad_norm": 4.336763381958008, "learning_rate": 5.107121259651898e-07, "loss": 0.4964, "step": 20497 }, { "epoch": 0.8592477704536642, "grad_norm": 1.8941835165023804, "learning_rate": 5.10413278170736e-07, "loss": 0.4842, "step": 20498 }, { "epoch": 0.8592896890686732, "grad_norm": 2.015648365020752, "learning_rate": 5.101145131356072e-07, "loss": 0.4563, "step": 20499 }, { "epoch": 0.8593316076836821, "grad_norm": 1.6108614206314087, "learning_rate": 5.098158308653107e-07, "loss": 0.4673, "step": 20500 }, { "epoch": 0.8593735262986911, "grad_norm": 2.0132033824920654, "learning_rate": 5.095172313653502e-07, "loss": 0.4973, "step": 20501 }, { "epoch": 0.8594154449137, "grad_norm": 1.78738534450531, "learning_rate": 5.092187146412319e-07, "loss": 0.5125, "step": 20502 }, { "epoch": 0.859457363528709, "grad_norm": 1.931630253791809, "learning_rate": 5.089202806984594e-07, "loss": 0.4637, "step": 20503 }, { "epoch": 0.859499282143718, "grad_norm": 2.1699767112731934, "learning_rate": 5.086219295425321e-07, "loss": 0.4964, "step": 20504 }, { "epoch": 0.859541200758727, "grad_norm": 1.8701328039169312, "learning_rate": 5.083236611789499e-07, "loss": 0.4838, "step": 20505 }, { "epoch": 0.8595831193737359, "grad_norm": 2.0992586612701416, "learning_rate": 5.08025475613213e-07, "loss": 0.4902, "step": 20506 }, { "epoch": 0.8596250379887449, "grad_norm": 1.7380642890930176, "learning_rate": 5.077273728508153e-07, "loss": 0.4418, "step": 20507 }, { "epoch": 0.8596669566037538, "grad_norm": 1.724155068397522, "learning_rate": 5.074293528972535e-07, "loss": 0.4853, "step": 20508 }, { "epoch": 0.8597088752187628, "grad_norm": 1.6676431894302368, "learning_rate": 5.071314157580204e-07, "loss": 0.4373, "step": 20509 }, { "epoch": 0.8597507938337717, "grad_norm": 1.8620030879974365, "learning_rate": 5.068335614386094e-07, "loss": 0.4252, "step": 20510 }, { "epoch": 0.8597927124487806, "grad_norm": 1.9369593858718872, "learning_rate": 5.065357899445089e-07, "loss": 0.38, "step": 20511 }, { "epoch": 0.8598346310637897, "grad_norm": 1.9001144170761108, "learning_rate": 5.062381012812101e-07, "loss": 0.457, "step": 20512 }, { "epoch": 0.8598765496787986, "grad_norm": 1.8769659996032715, "learning_rate": 5.059404954541986e-07, "loss": 0.4329, "step": 20513 }, { "epoch": 0.8599184682938076, "grad_norm": 4.2406487464904785, "learning_rate": 5.05642972468961e-07, "loss": 0.5018, "step": 20514 }, { "epoch": 0.8599603869088165, "grad_norm": 2.2565841674804688, "learning_rate": 5.05345532330983e-07, "loss": 0.4834, "step": 20515 }, { "epoch": 0.8600023055238255, "grad_norm": 1.7829084396362305, "learning_rate": 5.050481750457459e-07, "loss": 0.518, "step": 20516 }, { "epoch": 0.8600442241388344, "grad_norm": 1.8456063270568848, "learning_rate": 5.047509006187312e-07, "loss": 0.5074, "step": 20517 }, { "epoch": 0.8600861427538434, "grad_norm": 5.241970539093018, "learning_rate": 5.044537090554202e-07, "loss": 0.4471, "step": 20518 }, { "epoch": 0.8601280613688523, "grad_norm": 1.7848339080810547, "learning_rate": 5.041566003612897e-07, "loss": 0.4205, "step": 20519 }, { "epoch": 0.8601699799838614, "grad_norm": 2.2377407550811768, "learning_rate": 5.038595745418168e-07, "loss": 0.4909, "step": 20520 }, { "epoch": 0.8602118985988703, "grad_norm": 9.718987464904785, "learning_rate": 5.035626316024772e-07, "loss": 0.4289, "step": 20521 }, { "epoch": 0.8602538172138793, "grad_norm": 1.5930204391479492, "learning_rate": 5.032657715487438e-07, "loss": 0.4452, "step": 20522 }, { "epoch": 0.8602957358288882, "grad_norm": 1.8940540552139282, "learning_rate": 5.0296899438609e-07, "loss": 0.4541, "step": 20523 }, { "epoch": 0.8603376544438972, "grad_norm": 1.972105622291565, "learning_rate": 5.026723001199851e-07, "loss": 0.4714, "step": 20524 }, { "epoch": 0.8603795730589061, "grad_norm": 1.6402630805969238, "learning_rate": 5.023756887558989e-07, "loss": 0.5102, "step": 20525 }, { "epoch": 0.8604214916739151, "grad_norm": 1.7454346418380737, "learning_rate": 5.020791602993003e-07, "loss": 0.4676, "step": 20526 }, { "epoch": 0.860463410288924, "grad_norm": 2.143990993499756, "learning_rate": 5.017827147556531e-07, "loss": 0.4414, "step": 20527 }, { "epoch": 0.860505328903933, "grad_norm": 1.8255294561386108, "learning_rate": 5.014863521304236e-07, "loss": 0.4051, "step": 20528 }, { "epoch": 0.860547247518942, "grad_norm": 2.153536796569824, "learning_rate": 5.011900724290747e-07, "loss": 0.4772, "step": 20529 }, { "epoch": 0.860589166133951, "grad_norm": 2.0860419273376465, "learning_rate": 5.008938756570659e-07, "loss": 0.4623, "step": 20530 }, { "epoch": 0.8606310847489599, "grad_norm": 2.5872974395751953, "learning_rate": 5.005977618198587e-07, "loss": 0.4884, "step": 20531 }, { "epoch": 0.8606730033639689, "grad_norm": 2.036432981491089, "learning_rate": 5.00301730922913e-07, "loss": 0.4895, "step": 20532 }, { "epoch": 0.8607149219789778, "grad_norm": 2.5892386436462402, "learning_rate": 5.000057829716832e-07, "loss": 0.4895, "step": 20533 }, { "epoch": 0.8607568405939868, "grad_norm": 2.0333871841430664, "learning_rate": 4.997099179716253e-07, "loss": 0.5238, "step": 20534 }, { "epoch": 0.8607987592089957, "grad_norm": 1.9117558002471924, "learning_rate": 4.994141359281951e-07, "loss": 0.4526, "step": 20535 }, { "epoch": 0.8608406778240046, "grad_norm": 2.117326259613037, "learning_rate": 4.991184368468421e-07, "loss": 0.5445, "step": 20536 }, { "epoch": 0.8608825964390137, "grad_norm": 1.8215608596801758, "learning_rate": 4.988228207330192e-07, "loss": 0.5179, "step": 20537 }, { "epoch": 0.8609245150540226, "grad_norm": 1.896913766860962, "learning_rate": 4.985272875921754e-07, "loss": 0.4946, "step": 20538 }, { "epoch": 0.8609664336690316, "grad_norm": 1.659896731376648, "learning_rate": 4.982318374297568e-07, "loss": 0.4586, "step": 20539 }, { "epoch": 0.8610083522840405, "grad_norm": 1.9707746505737305, "learning_rate": 4.979364702512107e-07, "loss": 0.4681, "step": 20540 }, { "epoch": 0.8610502708990495, "grad_norm": 2.3788247108459473, "learning_rate": 4.976411860619828e-07, "loss": 0.5313, "step": 20541 }, { "epoch": 0.8610921895140584, "grad_norm": 2.2000365257263184, "learning_rate": 4.973459848675144e-07, "loss": 0.465, "step": 20542 }, { "epoch": 0.8611341081290674, "grad_norm": 1.9083709716796875, "learning_rate": 4.970508666732477e-07, "loss": 0.4634, "step": 20543 }, { "epoch": 0.8611760267440763, "grad_norm": 1.7044981718063354, "learning_rate": 4.96755831484625e-07, "loss": 0.4948, "step": 20544 }, { "epoch": 0.8612179453590854, "grad_norm": 1.839485764503479, "learning_rate": 4.964608793070813e-07, "loss": 0.432, "step": 20545 }, { "epoch": 0.8612598639740943, "grad_norm": 1.8831455707550049, "learning_rate": 4.961660101460564e-07, "loss": 0.4626, "step": 20546 }, { "epoch": 0.8613017825891033, "grad_norm": 2.264559268951416, "learning_rate": 4.958712240069846e-07, "loss": 0.5324, "step": 20547 }, { "epoch": 0.8613437012041122, "grad_norm": 1.8498607873916626, "learning_rate": 4.955765208952996e-07, "loss": 0.4258, "step": 20548 }, { "epoch": 0.8613856198191212, "grad_norm": 1.7619102001190186, "learning_rate": 4.952819008164339e-07, "loss": 0.4461, "step": 20549 }, { "epoch": 0.8614275384341301, "grad_norm": 1.9309190511703491, "learning_rate": 4.949873637758195e-07, "loss": 0.4803, "step": 20550 }, { "epoch": 0.8614694570491391, "grad_norm": 3.0987257957458496, "learning_rate": 4.946929097788844e-07, "loss": 0.4298, "step": 20551 }, { "epoch": 0.861511375664148, "grad_norm": 1.609725832939148, "learning_rate": 4.943985388310568e-07, "loss": 0.4141, "step": 20552 }, { "epoch": 0.8615532942791571, "grad_norm": 1.7152057886123657, "learning_rate": 4.941042509377647e-07, "loss": 0.4772, "step": 20553 }, { "epoch": 0.861595212894166, "grad_norm": 2.06061053276062, "learning_rate": 4.938100461044304e-07, "loss": 0.4571, "step": 20554 }, { "epoch": 0.861637131509175, "grad_norm": 2.113743305206299, "learning_rate": 4.935159243364796e-07, "loss": 0.5452, "step": 20555 }, { "epoch": 0.8616790501241839, "grad_norm": 1.812792420387268, "learning_rate": 4.932218856393311e-07, "loss": 0.4804, "step": 20556 }, { "epoch": 0.8617209687391929, "grad_norm": 1.9498575925827026, "learning_rate": 4.929279300184081e-07, "loss": 0.4855, "step": 20557 }, { "epoch": 0.8617628873542018, "grad_norm": 1.737647294998169, "learning_rate": 4.926340574791267e-07, "loss": 0.4804, "step": 20558 }, { "epoch": 0.8618048059692108, "grad_norm": 1.6897294521331787, "learning_rate": 4.923402680269063e-07, "loss": 0.4671, "step": 20559 }, { "epoch": 0.8618467245842197, "grad_norm": 1.96532142162323, "learning_rate": 4.920465616671599e-07, "loss": 0.4432, "step": 20560 }, { "epoch": 0.8618886431992286, "grad_norm": 2.386031150817871, "learning_rate": 4.917529384053038e-07, "loss": 0.4562, "step": 20561 }, { "epoch": 0.8619305618142377, "grad_norm": 4.65814733505249, "learning_rate": 4.914593982467508e-07, "loss": 0.4648, "step": 20562 }, { "epoch": 0.8619724804292466, "grad_norm": 1.6258273124694824, "learning_rate": 4.911659411969094e-07, "loss": 0.4579, "step": 20563 }, { "epoch": 0.8620143990442556, "grad_norm": 1.6306570768356323, "learning_rate": 4.908725672611908e-07, "loss": 0.4715, "step": 20564 }, { "epoch": 0.8620563176592645, "grad_norm": 1.9145355224609375, "learning_rate": 4.90579276445004e-07, "loss": 0.4872, "step": 20565 }, { "epoch": 0.8620982362742735, "grad_norm": 2.1007745265960693, "learning_rate": 4.902860687537536e-07, "loss": 0.5041, "step": 20566 }, { "epoch": 0.8621401548892824, "grad_norm": 2.592541456222534, "learning_rate": 4.899929441928447e-07, "loss": 0.5125, "step": 20567 }, { "epoch": 0.8621820735042914, "grad_norm": 2.9794139862060547, "learning_rate": 4.896999027676819e-07, "loss": 0.4523, "step": 20568 }, { "epoch": 0.8622239921193003, "grad_norm": 1.589786410331726, "learning_rate": 4.894069444836646e-07, "loss": 0.4046, "step": 20569 }, { "epoch": 0.8622659107343094, "grad_norm": 2.0685064792633057, "learning_rate": 4.891140693461949e-07, "loss": 0.457, "step": 20570 }, { "epoch": 0.8623078293493183, "grad_norm": 1.891926646232605, "learning_rate": 4.888212773606721e-07, "loss": 0.4619, "step": 20571 }, { "epoch": 0.8623497479643273, "grad_norm": 2.0861101150512695, "learning_rate": 4.885285685324914e-07, "loss": 0.4435, "step": 20572 }, { "epoch": 0.8623916665793362, "grad_norm": 1.813094139099121, "learning_rate": 4.882359428670502e-07, "loss": 0.4521, "step": 20573 }, { "epoch": 0.8624335851943452, "grad_norm": 1.7021214962005615, "learning_rate": 4.879434003697425e-07, "loss": 0.4349, "step": 20574 }, { "epoch": 0.8624755038093541, "grad_norm": 1.7109670639038086, "learning_rate": 4.876509410459607e-07, "loss": 0.4698, "step": 20575 }, { "epoch": 0.8625174224243631, "grad_norm": 1.9498411417007446, "learning_rate": 4.873585649010942e-07, "loss": 0.5014, "step": 20576 }, { "epoch": 0.862559341039372, "grad_norm": 2.0369443893432617, "learning_rate": 4.870662719405355e-07, "loss": 0.4041, "step": 20577 }, { "epoch": 0.8626012596543811, "grad_norm": 1.7966904640197754, "learning_rate": 4.867740621696704e-07, "loss": 0.4314, "step": 20578 }, { "epoch": 0.86264317826939, "grad_norm": 2.6974477767944336, "learning_rate": 4.864819355938854e-07, "loss": 0.4506, "step": 20579 }, { "epoch": 0.862685096884399, "grad_norm": 1.652639389038086, "learning_rate": 4.861898922185676e-07, "loss": 0.446, "step": 20580 }, { "epoch": 0.8627270154994079, "grad_norm": 2.325899839401245, "learning_rate": 4.85897932049098e-07, "loss": 0.4554, "step": 20581 }, { "epoch": 0.8627689341144169, "grad_norm": 2.048996686935425, "learning_rate": 4.856060550908597e-07, "loss": 0.4574, "step": 20582 }, { "epoch": 0.8628108527294258, "grad_norm": 1.939887523651123, "learning_rate": 4.853142613492335e-07, "loss": 0.4679, "step": 20583 }, { "epoch": 0.8628527713444348, "grad_norm": 3.2987167835235596, "learning_rate": 4.850225508295975e-07, "loss": 0.5128, "step": 20584 }, { "epoch": 0.8628946899594437, "grad_norm": 2.030062437057495, "learning_rate": 4.847309235373277e-07, "loss": 0.4543, "step": 20585 }, { "epoch": 0.8629366085744526, "grad_norm": 2.269568920135498, "learning_rate": 4.844393794778024e-07, "loss": 0.485, "step": 20586 }, { "epoch": 0.8629785271894617, "grad_norm": 1.8625431060791016, "learning_rate": 4.841479186563941e-07, "loss": 0.5131, "step": 20587 }, { "epoch": 0.8630204458044706, "grad_norm": 2.104173421859741, "learning_rate": 4.838565410784751e-07, "loss": 0.5292, "step": 20588 }, { "epoch": 0.8630623644194796, "grad_norm": 1.7885148525238037, "learning_rate": 4.835652467494189e-07, "loss": 0.4902, "step": 20589 }, { "epoch": 0.8631042830344885, "grad_norm": 2.5729713439941406, "learning_rate": 4.832740356745924e-07, "loss": 0.4897, "step": 20590 }, { "epoch": 0.8631462016494975, "grad_norm": 2.445896863937378, "learning_rate": 4.829829078593651e-07, "loss": 0.4775, "step": 20591 }, { "epoch": 0.8631881202645064, "grad_norm": 2.338840961456299, "learning_rate": 4.826918633091043e-07, "loss": 0.4761, "step": 20592 }, { "epoch": 0.8632300388795154, "grad_norm": 1.959416389465332, "learning_rate": 4.824009020291736e-07, "loss": 0.4744, "step": 20593 }, { "epoch": 0.8632719574945243, "grad_norm": 1.9763747453689575, "learning_rate": 4.82110024024936e-07, "loss": 0.5087, "step": 20594 }, { "epoch": 0.8633138761095334, "grad_norm": 1.8025474548339844, "learning_rate": 4.818192293017553e-07, "loss": 0.5101, "step": 20595 }, { "epoch": 0.8633557947245423, "grad_norm": 1.8895318508148193, "learning_rate": 4.815285178649903e-07, "loss": 0.4995, "step": 20596 }, { "epoch": 0.8633977133395513, "grad_norm": 2.38423228263855, "learning_rate": 4.812378897200004e-07, "loss": 0.4677, "step": 20597 }, { "epoch": 0.8634396319545602, "grad_norm": 2.663846969604492, "learning_rate": 4.809473448721436e-07, "loss": 0.4911, "step": 20598 }, { "epoch": 0.8634815505695692, "grad_norm": 2.211902618408203, "learning_rate": 4.806568833267744e-07, "loss": 0.5206, "step": 20599 }, { "epoch": 0.8635234691845781, "grad_norm": 2.104484796524048, "learning_rate": 4.803665050892486e-07, "loss": 0.5271, "step": 20600 }, { "epoch": 0.8635653877995871, "grad_norm": 2.0415737628936768, "learning_rate": 4.800762101649171e-07, "loss": 0.4382, "step": 20601 }, { "epoch": 0.863607306414596, "grad_norm": 1.7640244960784912, "learning_rate": 4.79785998559133e-07, "loss": 0.4583, "step": 20602 }, { "epoch": 0.8636492250296051, "grad_norm": 2.9239609241485596, "learning_rate": 4.794958702772446e-07, "loss": 0.4833, "step": 20603 }, { "epoch": 0.863691143644614, "grad_norm": 2.376633644104004, "learning_rate": 4.792058253246007e-07, "loss": 0.5012, "step": 20604 }, { "epoch": 0.863733062259623, "grad_norm": 1.773597002029419, "learning_rate": 4.789158637065466e-07, "loss": 0.4948, "step": 20605 }, { "epoch": 0.8637749808746319, "grad_norm": 3.4048893451690674, "learning_rate": 4.786259854284298e-07, "loss": 0.5175, "step": 20606 }, { "epoch": 0.8638168994896409, "grad_norm": 2.0351343154907227, "learning_rate": 4.783361904955913e-07, "loss": 0.4696, "step": 20607 }, { "epoch": 0.8638588181046498, "grad_norm": 1.8136740922927856, "learning_rate": 4.780464789133738e-07, "loss": 0.4682, "step": 20608 }, { "epoch": 0.8639007367196588, "grad_norm": 1.8714896440505981, "learning_rate": 4.777568506871188e-07, "loss": 0.4593, "step": 20609 }, { "epoch": 0.8639426553346677, "grad_norm": 1.6782549619674683, "learning_rate": 4.77467305822164e-07, "loss": 0.484, "step": 20610 }, { "epoch": 0.8639845739496766, "grad_norm": 1.9112032651901245, "learning_rate": 4.771778443238473e-07, "loss": 0.4703, "step": 20611 }, { "epoch": 0.8640264925646857, "grad_norm": 1.8121914863586426, "learning_rate": 4.768884661975048e-07, "loss": 0.4815, "step": 20612 }, { "epoch": 0.8640684111796946, "grad_norm": 2.2583506107330322, "learning_rate": 4.765991714484708e-07, "loss": 0.4761, "step": 20613 }, { "epoch": 0.8641103297947036, "grad_norm": 1.9493577480316162, "learning_rate": 4.76309960082077e-07, "loss": 0.4777, "step": 20614 }, { "epoch": 0.8641522484097125, "grad_norm": 1.720883846282959, "learning_rate": 4.760208321036558e-07, "loss": 0.4496, "step": 20615 }, { "epoch": 0.8641941670247215, "grad_norm": 1.8142750263214111, "learning_rate": 4.757317875185352e-07, "loss": 0.4336, "step": 20616 }, { "epoch": 0.8642360856397304, "grad_norm": 1.7496552467346191, "learning_rate": 4.7544282633204486e-07, "loss": 0.441, "step": 20617 }, { "epoch": 0.8642780042547394, "grad_norm": 1.9987345933914185, "learning_rate": 4.7515394854951214e-07, "loss": 0.4836, "step": 20618 }, { "epoch": 0.8643199228697483, "grad_norm": 2.0399394035339355, "learning_rate": 4.7486515417625947e-07, "loss": 0.4673, "step": 20619 }, { "epoch": 0.8643618414847574, "grad_norm": 2.135204553604126, "learning_rate": 4.74576443217612e-07, "loss": 0.4954, "step": 20620 }, { "epoch": 0.8644037600997663, "grad_norm": 1.904606819152832, "learning_rate": 4.742878156788927e-07, "loss": 0.4936, "step": 20621 }, { "epoch": 0.8644456787147753, "grad_norm": 3.677143096923828, "learning_rate": 4.7399927156542127e-07, "loss": 0.5075, "step": 20622 }, { "epoch": 0.8644875973297842, "grad_norm": 2.078446626663208, "learning_rate": 4.7371081088251447e-07, "loss": 0.4797, "step": 20623 }, { "epoch": 0.8645295159447932, "grad_norm": 2.782961845397949, "learning_rate": 4.7342243363549314e-07, "loss": 0.5049, "step": 20624 }, { "epoch": 0.8645714345598021, "grad_norm": 1.9861485958099365, "learning_rate": 4.731341398296702e-07, "loss": 0.4506, "step": 20625 }, { "epoch": 0.8646133531748111, "grad_norm": 2.0397017002105713, "learning_rate": 4.728459294703613e-07, "loss": 0.3782, "step": 20626 }, { "epoch": 0.86465527178982, "grad_norm": 2.089592218399048, "learning_rate": 4.7255780256287966e-07, "loss": 0.5002, "step": 20627 }, { "epoch": 0.8646971904048291, "grad_norm": 1.792880892753601, "learning_rate": 4.7226975911253527e-07, "loss": 0.4752, "step": 20628 }, { "epoch": 0.864739109019838, "grad_norm": 1.8642725944519043, "learning_rate": 4.719817991246384e-07, "loss": 0.4965, "step": 20629 }, { "epoch": 0.864781027634847, "grad_norm": 2.2266182899475098, "learning_rate": 4.7169392260449866e-07, "loss": 0.4595, "step": 20630 }, { "epoch": 0.8648229462498559, "grad_norm": 2.500181198120117, "learning_rate": 4.7140612955742073e-07, "loss": 0.4555, "step": 20631 }, { "epoch": 0.8648648648648649, "grad_norm": 1.9873790740966797, "learning_rate": 4.711184199887092e-07, "loss": 0.4728, "step": 20632 }, { "epoch": 0.8649067834798738, "grad_norm": 1.7623592615127563, "learning_rate": 4.708307939036699e-07, "loss": 0.4546, "step": 20633 }, { "epoch": 0.8649487020948828, "grad_norm": 2.63006329536438, "learning_rate": 4.7054325130760247e-07, "loss": 0.439, "step": 20634 }, { "epoch": 0.8649906207098917, "grad_norm": 1.6647895574569702, "learning_rate": 4.702557922058082e-07, "loss": 0.4567, "step": 20635 }, { "epoch": 0.8650325393249007, "grad_norm": 1.7586034536361694, "learning_rate": 4.6996841660358725e-07, "loss": 0.4693, "step": 20636 }, { "epoch": 0.8650744579399097, "grad_norm": 2.5856921672821045, "learning_rate": 4.696811245062355e-07, "loss": 0.4714, "step": 20637 }, { "epoch": 0.8651163765549186, "grad_norm": 1.8600436449050903, "learning_rate": 4.693939159190486e-07, "loss": 0.4612, "step": 20638 }, { "epoch": 0.8651582951699276, "grad_norm": 1.7675158977508545, "learning_rate": 4.691067908473229e-07, "loss": 0.4617, "step": 20639 }, { "epoch": 0.8652002137849365, "grad_norm": 3.072758197784424, "learning_rate": 4.688197492963492e-07, "loss": 0.5129, "step": 20640 }, { "epoch": 0.8652421323999455, "grad_norm": 1.973281741142273, "learning_rate": 4.6853279127141883e-07, "loss": 0.4721, "step": 20641 }, { "epoch": 0.8652840510149544, "grad_norm": 1.7582461833953857, "learning_rate": 4.682459167778225e-07, "loss": 0.4776, "step": 20642 }, { "epoch": 0.8653259696299634, "grad_norm": 1.6653553247451782, "learning_rate": 4.679591258208471e-07, "loss": 0.5006, "step": 20643 }, { "epoch": 0.8653678882449723, "grad_norm": 1.909204363822937, "learning_rate": 4.676724184057796e-07, "loss": 0.4658, "step": 20644 }, { "epoch": 0.8654098068599814, "grad_norm": 1.8128501176834106, "learning_rate": 4.673857945379062e-07, "loss": 0.4873, "step": 20645 }, { "epoch": 0.8654517254749903, "grad_norm": 2.416830062866211, "learning_rate": 4.670992542225089e-07, "loss": 0.4228, "step": 20646 }, { "epoch": 0.8654936440899993, "grad_norm": 1.6098424196243286, "learning_rate": 4.6681279746487007e-07, "loss": 0.4254, "step": 20647 }, { "epoch": 0.8655355627050082, "grad_norm": 2.000229597091675, "learning_rate": 4.6652642427027104e-07, "loss": 0.4861, "step": 20648 }, { "epoch": 0.8655774813200172, "grad_norm": 1.6875287294387817, "learning_rate": 4.6624013464399045e-07, "loss": 0.4281, "step": 20649 }, { "epoch": 0.8656193999350261, "grad_norm": 1.9541336297988892, "learning_rate": 4.659539285913045e-07, "loss": 0.4646, "step": 20650 }, { "epoch": 0.8656613185500351, "grad_norm": 1.8482497930526733, "learning_rate": 4.656678061174902e-07, "loss": 0.469, "step": 20651 }, { "epoch": 0.865703237165044, "grad_norm": 2.513458490371704, "learning_rate": 4.653817672278205e-07, "loss": 0.5107, "step": 20652 }, { "epoch": 0.8657451557800531, "grad_norm": 1.7383415699005127, "learning_rate": 4.650958119275695e-07, "loss": 0.4558, "step": 20653 }, { "epoch": 0.865787074395062, "grad_norm": 2.1806113719940186, "learning_rate": 4.64809940222008e-07, "loss": 0.5235, "step": 20654 }, { "epoch": 0.865828993010071, "grad_norm": 2.8927221298217773, "learning_rate": 4.645241521164051e-07, "loss": 0.5145, "step": 20655 }, { "epoch": 0.8658709116250799, "grad_norm": 1.7595243453979492, "learning_rate": 4.6423844761602944e-07, "loss": 0.4748, "step": 20656 }, { "epoch": 0.8659128302400889, "grad_norm": 1.7664566040039062, "learning_rate": 4.639528267261484e-07, "loss": 0.4807, "step": 20657 }, { "epoch": 0.8659547488550978, "grad_norm": 2.1028683185577393, "learning_rate": 4.63667289452025e-07, "loss": 0.4289, "step": 20658 }, { "epoch": 0.8659966674701068, "grad_norm": 2.197509765625, "learning_rate": 4.63381835798925e-07, "loss": 0.5136, "step": 20659 }, { "epoch": 0.8660385860851157, "grad_norm": 1.909241795539856, "learning_rate": 4.6309646577210875e-07, "loss": 0.4429, "step": 20660 }, { "epoch": 0.8660805047001247, "grad_norm": 1.799893856048584, "learning_rate": 4.6281117937683637e-07, "loss": 0.4855, "step": 20661 }, { "epoch": 0.8661224233151337, "grad_norm": 1.7452096939086914, "learning_rate": 4.6252597661836763e-07, "loss": 0.5212, "step": 20662 }, { "epoch": 0.8661643419301426, "grad_norm": 2.056201219558716, "learning_rate": 4.622408575019605e-07, "loss": 0.541, "step": 20663 }, { "epoch": 0.8662062605451516, "grad_norm": 1.924530029296875, "learning_rate": 4.619558220328685e-07, "loss": 0.4885, "step": 20664 }, { "epoch": 0.8662481791601605, "grad_norm": 2.0689618587493896, "learning_rate": 4.6167087021634747e-07, "loss": 0.4647, "step": 20665 }, { "epoch": 0.8662900977751695, "grad_norm": 1.9435069561004639, "learning_rate": 4.6138600205765105e-07, "loss": 0.497, "step": 20666 }, { "epoch": 0.8663320163901784, "grad_norm": 17.865427017211914, "learning_rate": 4.611012175620283e-07, "loss": 0.4159, "step": 20667 }, { "epoch": 0.8663739350051874, "grad_norm": 2.1922976970672607, "learning_rate": 4.608165167347306e-07, "loss": 0.4476, "step": 20668 }, { "epoch": 0.8664158536201964, "grad_norm": 1.875264048576355, "learning_rate": 4.6053189958100543e-07, "loss": 0.4794, "step": 20669 }, { "epoch": 0.8664577722352054, "grad_norm": 1.9038894176483154, "learning_rate": 4.60247366106098e-07, "loss": 0.4537, "step": 20670 }, { "epoch": 0.8664996908502143, "grad_norm": 1.5469111204147339, "learning_rate": 4.5996291631525416e-07, "loss": 0.4236, "step": 20671 }, { "epoch": 0.8665416094652233, "grad_norm": 1.7749748229980469, "learning_rate": 4.5967855021371857e-07, "loss": 0.4585, "step": 20672 }, { "epoch": 0.8665835280802322, "grad_norm": 1.8410307168960571, "learning_rate": 4.593942678067309e-07, "loss": 0.5072, "step": 20673 }, { "epoch": 0.8666254466952412, "grad_norm": 1.4732975959777832, "learning_rate": 4.591100690995343e-07, "loss": 0.4459, "step": 20674 }, { "epoch": 0.8666673653102501, "grad_norm": 1.7844494581222534, "learning_rate": 4.588259540973644e-07, "loss": 0.4395, "step": 20675 }, { "epoch": 0.8667092839252591, "grad_norm": 1.6420087814331055, "learning_rate": 4.5854192280546053e-07, "loss": 0.4762, "step": 20676 }, { "epoch": 0.866751202540268, "grad_norm": 1.902658462524414, "learning_rate": 4.5825797522905834e-07, "loss": 0.4899, "step": 20677 }, { "epoch": 0.8667931211552771, "grad_norm": 1.954925775527954, "learning_rate": 4.579741113733921e-07, "loss": 0.5082, "step": 20678 }, { "epoch": 0.866835039770286, "grad_norm": 2.055772304534912, "learning_rate": 4.576903312436931e-07, "loss": 0.4528, "step": 20679 }, { "epoch": 0.866876958385295, "grad_norm": 2.5546724796295166, "learning_rate": 4.5740663484519386e-07, "loss": 0.4753, "step": 20680 }, { "epoch": 0.8669188770003039, "grad_norm": 1.9192084074020386, "learning_rate": 4.5712302218312245e-07, "loss": 0.4583, "step": 20681 }, { "epoch": 0.8669607956153129, "grad_norm": 2.070667266845703, "learning_rate": 4.568394932627079e-07, "loss": 0.4546, "step": 20682 }, { "epoch": 0.8670027142303218, "grad_norm": 1.800680160522461, "learning_rate": 4.565560480891773e-07, "loss": 0.5041, "step": 20683 }, { "epoch": 0.8670446328453308, "grad_norm": 2.29529070854187, "learning_rate": 4.5627268666775413e-07, "loss": 0.4549, "step": 20684 }, { "epoch": 0.8670865514603397, "grad_norm": 2.0324037075042725, "learning_rate": 4.5598940900366265e-07, "loss": 0.5079, "step": 20685 }, { "epoch": 0.8671284700753487, "grad_norm": 2.003281593322754, "learning_rate": 4.557062151021247e-07, "loss": 0.4844, "step": 20686 }, { "epoch": 0.8671703886903577, "grad_norm": 1.7578020095825195, "learning_rate": 4.5542310496836107e-07, "loss": 0.4714, "step": 20687 }, { "epoch": 0.8672123073053666, "grad_norm": 2.253422975540161, "learning_rate": 4.551400786075888e-07, "loss": 0.5018, "step": 20688 }, { "epoch": 0.8672542259203756, "grad_norm": 1.9322421550750732, "learning_rate": 4.54857136025027e-07, "loss": 0.5017, "step": 20689 }, { "epoch": 0.8672961445353845, "grad_norm": 2.171823740005493, "learning_rate": 4.545742772258893e-07, "loss": 0.4552, "step": 20690 }, { "epoch": 0.8673380631503935, "grad_norm": 1.8112635612487793, "learning_rate": 4.5429150221539097e-07, "loss": 0.4171, "step": 20691 }, { "epoch": 0.8673799817654024, "grad_norm": 1.9893443584442139, "learning_rate": 4.540088109987456e-07, "loss": 0.4736, "step": 20692 }, { "epoch": 0.8674219003804114, "grad_norm": 1.5192840099334717, "learning_rate": 4.537262035811618e-07, "loss": 0.408, "step": 20693 }, { "epoch": 0.8674638189954204, "grad_norm": 2.0567965507507324, "learning_rate": 4.53443679967851e-07, "loss": 0.4936, "step": 20694 }, { "epoch": 0.8675057376104294, "grad_norm": 1.7766331434249878, "learning_rate": 4.5316124016402117e-07, "loss": 0.488, "step": 20695 }, { "epoch": 0.8675476562254383, "grad_norm": 1.6680151224136353, "learning_rate": 4.5287888417487825e-07, "loss": 0.441, "step": 20696 }, { "epoch": 0.8675895748404473, "grad_norm": 1.9899154901504517, "learning_rate": 4.525966120056258e-07, "loss": 0.5037, "step": 20697 }, { "epoch": 0.8676314934554562, "grad_norm": 1.8288832902908325, "learning_rate": 4.5231442366146905e-07, "loss": 0.4412, "step": 20698 }, { "epoch": 0.8676734120704652, "grad_norm": 2.006197929382324, "learning_rate": 4.5203231914760836e-07, "loss": 0.5242, "step": 20699 }, { "epoch": 0.8677153306854741, "grad_norm": 2.511906385421753, "learning_rate": 4.51750298469244e-07, "loss": 0.4405, "step": 20700 }, { "epoch": 0.8677572493004831, "grad_norm": 2.1279959678649902, "learning_rate": 4.5146836163157624e-07, "loss": 0.5047, "step": 20701 }, { "epoch": 0.867799167915492, "grad_norm": 1.9699668884277344, "learning_rate": 4.511865086398004e-07, "loss": 0.4573, "step": 20702 }, { "epoch": 0.8678410865305011, "grad_norm": 1.9949363470077515, "learning_rate": 4.509047394991123e-07, "loss": 0.448, "step": 20703 }, { "epoch": 0.86788300514551, "grad_norm": 1.5778244733810425, "learning_rate": 4.5062305421470777e-07, "loss": 0.4514, "step": 20704 }, { "epoch": 0.867924923760519, "grad_norm": 1.746419906616211, "learning_rate": 4.5034145279177656e-07, "loss": 0.4634, "step": 20705 }, { "epoch": 0.8679668423755279, "grad_norm": 1.9532350301742554, "learning_rate": 4.500599352355123e-07, "loss": 0.4864, "step": 20706 }, { "epoch": 0.8680087609905369, "grad_norm": 1.6543018817901611, "learning_rate": 4.4977850155110304e-07, "loss": 0.4922, "step": 20707 }, { "epoch": 0.8680506796055458, "grad_norm": 1.9289480447769165, "learning_rate": 4.494971517437352e-07, "loss": 0.462, "step": 20708 }, { "epoch": 0.8680925982205548, "grad_norm": 1.8128248453140259, "learning_rate": 4.492158858185969e-07, "loss": 0.5332, "step": 20709 }, { "epoch": 0.8681345168355638, "grad_norm": 1.8134429454803467, "learning_rate": 4.489347037808733e-07, "loss": 0.5085, "step": 20710 }, { "epoch": 0.8681764354505727, "grad_norm": 1.7972981929779053, "learning_rate": 4.4865360563574543e-07, "loss": 0.4796, "step": 20711 }, { "epoch": 0.8682183540655817, "grad_norm": 1.7704740762710571, "learning_rate": 4.4837259138839686e-07, "loss": 0.4733, "step": 20712 }, { "epoch": 0.8682602726805906, "grad_norm": 1.882925271987915, "learning_rate": 4.480916610440078e-07, "loss": 0.4499, "step": 20713 }, { "epoch": 0.8683021912955996, "grad_norm": 1.856562852859497, "learning_rate": 4.478108146077553e-07, "loss": 0.4783, "step": 20714 }, { "epoch": 0.8683441099106085, "grad_norm": 1.9113914966583252, "learning_rate": 4.4753005208481803e-07, "loss": 0.5089, "step": 20715 }, { "epoch": 0.8683860285256175, "grad_norm": 2.0129458904266357, "learning_rate": 4.472493734803707e-07, "loss": 0.5241, "step": 20716 }, { "epoch": 0.8684279471406264, "grad_norm": 1.6390671730041504, "learning_rate": 4.4696877879958634e-07, "loss": 0.44, "step": 20717 }, { "epoch": 0.8684698657556355, "grad_norm": 1.8113130331039429, "learning_rate": 4.466882680476381e-07, "loss": 0.4388, "step": 20718 }, { "epoch": 0.8685117843706444, "grad_norm": 1.6164125204086304, "learning_rate": 4.4640784122969796e-07, "loss": 0.4086, "step": 20719 }, { "epoch": 0.8685537029856534, "grad_norm": 2.192150831222534, "learning_rate": 4.461274983509328e-07, "loss": 0.5407, "step": 20720 }, { "epoch": 0.8685956216006623, "grad_norm": 1.7339688539505005, "learning_rate": 4.4584723941651196e-07, "loss": 0.4945, "step": 20721 }, { "epoch": 0.8686375402156713, "grad_norm": 2.5648417472839355, "learning_rate": 4.4556706443160233e-07, "loss": 0.5159, "step": 20722 }, { "epoch": 0.8686794588306802, "grad_norm": 1.7382320165634155, "learning_rate": 4.452869734013665e-07, "loss": 0.4567, "step": 20723 }, { "epoch": 0.8687213774456892, "grad_norm": 1.6617964506149292, "learning_rate": 4.4500696633096975e-07, "loss": 0.4485, "step": 20724 }, { "epoch": 0.8687632960606981, "grad_norm": 1.8973350524902344, "learning_rate": 4.4472704322557234e-07, "loss": 0.4476, "step": 20725 }, { "epoch": 0.8688052146757071, "grad_norm": 2.28969407081604, "learning_rate": 4.4444720409033305e-07, "loss": 0.4652, "step": 20726 }, { "epoch": 0.8688471332907161, "grad_norm": 1.9601260423660278, "learning_rate": 4.4416744893041206e-07, "loss": 0.4788, "step": 20727 }, { "epoch": 0.8688890519057251, "grad_norm": 1.9559297561645508, "learning_rate": 4.4388777775096705e-07, "loss": 0.4818, "step": 20728 }, { "epoch": 0.868930970520734, "grad_norm": 1.9819891452789307, "learning_rate": 4.43608190557151e-07, "loss": 0.4449, "step": 20729 }, { "epoch": 0.868972889135743, "grad_norm": 2.253697156906128, "learning_rate": 4.4332868735411926e-07, "loss": 0.4499, "step": 20730 }, { "epoch": 0.8690148077507519, "grad_norm": 2.025205373764038, "learning_rate": 4.430492681470244e-07, "loss": 0.4557, "step": 20731 }, { "epoch": 0.8690567263657609, "grad_norm": 2.3530197143554688, "learning_rate": 4.4276993294101565e-07, "loss": 0.4548, "step": 20732 }, { "epoch": 0.8690986449807698, "grad_norm": 1.7181860208511353, "learning_rate": 4.424906817412439e-07, "loss": 0.4449, "step": 20733 }, { "epoch": 0.8691405635957788, "grad_norm": 1.8223850727081299, "learning_rate": 4.4221151455285547e-07, "loss": 0.443, "step": 20734 }, { "epoch": 0.8691824822107878, "grad_norm": 2.8918848037719727, "learning_rate": 4.419324313809964e-07, "loss": 0.5273, "step": 20735 }, { "epoch": 0.8692244008257967, "grad_norm": 1.5907542705535889, "learning_rate": 4.4165343223081134e-07, "loss": 0.436, "step": 20736 }, { "epoch": 0.8692663194408057, "grad_norm": 4.2657318115234375, "learning_rate": 4.413745171074446e-07, "loss": 0.5434, "step": 20737 }, { "epoch": 0.8693082380558146, "grad_norm": 4.695685386657715, "learning_rate": 4.4109568601603594e-07, "loss": 0.4768, "step": 20738 }, { "epoch": 0.8693501566708236, "grad_norm": 1.7067407369613647, "learning_rate": 4.408169389617251e-07, "loss": 0.4562, "step": 20739 }, { "epoch": 0.8693920752858325, "grad_norm": 1.9058562517166138, "learning_rate": 4.4053827594965235e-07, "loss": 0.5006, "step": 20740 }, { "epoch": 0.8694339939008415, "grad_norm": 3.228879451751709, "learning_rate": 4.4025969698495263e-07, "loss": 0.4724, "step": 20741 }, { "epoch": 0.8694759125158504, "grad_norm": 1.8623530864715576, "learning_rate": 4.3998120207276227e-07, "loss": 0.4209, "step": 20742 }, { "epoch": 0.8695178311308595, "grad_norm": 1.9700008630752563, "learning_rate": 4.397027912182139e-07, "loss": 0.4813, "step": 20743 }, { "epoch": 0.8695597497458684, "grad_norm": 2.4935367107391357, "learning_rate": 4.394244644264406e-07, "loss": 0.4654, "step": 20744 }, { "epoch": 0.8696016683608774, "grad_norm": 2.007610321044922, "learning_rate": 4.391462217025716e-07, "loss": 0.4873, "step": 20745 }, { "epoch": 0.8696435869758863, "grad_norm": 1.7167646884918213, "learning_rate": 4.3886806305173833e-07, "loss": 0.4659, "step": 20746 }, { "epoch": 0.8696855055908953, "grad_norm": 1.9680224657058716, "learning_rate": 4.385899884790651e-07, "loss": 0.4811, "step": 20747 }, { "epoch": 0.8697274242059042, "grad_norm": 2.282062530517578, "learning_rate": 4.383119979896805e-07, "loss": 0.4728, "step": 20748 }, { "epoch": 0.8697693428209132, "grad_norm": 2.0015270709991455, "learning_rate": 4.3803409158870714e-07, "loss": 0.5358, "step": 20749 }, { "epoch": 0.8698112614359221, "grad_norm": 2.374844551086426, "learning_rate": 4.3775626928126866e-07, "loss": 0.4284, "step": 20750 }, { "epoch": 0.8698531800509312, "grad_norm": 1.7047971487045288, "learning_rate": 4.3747853107248716e-07, "loss": 0.4806, "step": 20751 }, { "epoch": 0.8698950986659401, "grad_norm": 2.322864294052124, "learning_rate": 4.3720087696748016e-07, "loss": 0.4703, "step": 20752 }, { "epoch": 0.8699370172809491, "grad_norm": 1.783773422241211, "learning_rate": 4.3692330697136854e-07, "loss": 0.4663, "step": 20753 }, { "epoch": 0.869978935895958, "grad_norm": 1.9185009002685547, "learning_rate": 4.366458210892671e-07, "loss": 0.5221, "step": 20754 }, { "epoch": 0.870020854510967, "grad_norm": 2.066876173019409, "learning_rate": 4.363684193262907e-07, "loss": 0.4594, "step": 20755 }, { "epoch": 0.8700627731259759, "grad_norm": 1.8991831541061401, "learning_rate": 4.3609110168755354e-07, "loss": 0.4691, "step": 20756 }, { "epoch": 0.8701046917409849, "grad_norm": 2.583763360977173, "learning_rate": 4.358138681781682e-07, "loss": 0.5367, "step": 20757 }, { "epoch": 0.8701466103559938, "grad_norm": 1.7781108617782593, "learning_rate": 4.3553671880324336e-07, "loss": 0.4963, "step": 20758 }, { "epoch": 0.8701885289710029, "grad_norm": 2.249575138092041, "learning_rate": 4.352596535678888e-07, "loss": 0.5188, "step": 20759 }, { "epoch": 0.8702304475860118, "grad_norm": 2.09664249420166, "learning_rate": 4.349826724772133e-07, "loss": 0.4212, "step": 20760 }, { "epoch": 0.8702723662010207, "grad_norm": 1.7436015605926514, "learning_rate": 4.3470577553631985e-07, "loss": 0.4575, "step": 20761 }, { "epoch": 0.8703142848160297, "grad_norm": 2.8965554237365723, "learning_rate": 4.3442896275031565e-07, "loss": 0.44, "step": 20762 }, { "epoch": 0.8703562034310386, "grad_norm": 1.8517194986343384, "learning_rate": 4.341522341243015e-07, "loss": 0.4802, "step": 20763 }, { "epoch": 0.8703981220460476, "grad_norm": 2.4237897396087646, "learning_rate": 4.3387558966337784e-07, "loss": 0.4546, "step": 20764 }, { "epoch": 0.8704400406610565, "grad_norm": 1.9265543222427368, "learning_rate": 4.33599029372645e-07, "loss": 0.5185, "step": 20765 }, { "epoch": 0.8704819592760655, "grad_norm": 2.0005133152008057, "learning_rate": 4.3332255325720273e-07, "loss": 0.4395, "step": 20766 }, { "epoch": 0.8705238778910744, "grad_norm": 1.821792483329773, "learning_rate": 4.330461613221443e-07, "loss": 0.4552, "step": 20767 }, { "epoch": 0.8705657965060835, "grad_norm": 2.1186869144439697, "learning_rate": 4.327698535725666e-07, "loss": 0.4863, "step": 20768 }, { "epoch": 0.8706077151210924, "grad_norm": 2.601750612258911, "learning_rate": 4.3249363001356403e-07, "loss": 0.4947, "step": 20769 }, { "epoch": 0.8706496337361014, "grad_norm": 2.0551116466522217, "learning_rate": 4.322174906502252e-07, "loss": 0.5212, "step": 20770 }, { "epoch": 0.8706915523511103, "grad_norm": 1.9490001201629639, "learning_rate": 4.319414354876439e-07, "loss": 0.4199, "step": 20771 }, { "epoch": 0.8707334709661193, "grad_norm": 1.9417425394058228, "learning_rate": 4.316654645309065e-07, "loss": 0.4846, "step": 20772 }, { "epoch": 0.8707753895811282, "grad_norm": 2.145296335220337, "learning_rate": 4.313895777851002e-07, "loss": 0.5014, "step": 20773 }, { "epoch": 0.8708173081961372, "grad_norm": 1.6521086692810059, "learning_rate": 4.3111377525531075e-07, "loss": 0.4545, "step": 20774 }, { "epoch": 0.8708592268111461, "grad_norm": 2.094097137451172, "learning_rate": 4.3083805694662364e-07, "loss": 0.4865, "step": 20775 }, { "epoch": 0.8709011454261552, "grad_norm": 1.9766106605529785, "learning_rate": 4.3056242286411973e-07, "loss": 0.5168, "step": 20776 }, { "epoch": 0.8709430640411641, "grad_norm": 2.212798595428467, "learning_rate": 4.302868730128801e-07, "loss": 0.4524, "step": 20777 }, { "epoch": 0.8709849826561731, "grad_norm": 2.023780345916748, "learning_rate": 4.300114073979855e-07, "loss": 0.5005, "step": 20778 }, { "epoch": 0.871026901271182, "grad_norm": 2.368157386779785, "learning_rate": 4.2973602602451146e-07, "loss": 0.5239, "step": 20779 }, { "epoch": 0.871068819886191, "grad_norm": 2.1497108936309814, "learning_rate": 4.294607288975372e-07, "loss": 0.4498, "step": 20780 }, { "epoch": 0.8711107385011999, "grad_norm": 1.9974124431610107, "learning_rate": 4.2918551602213533e-07, "loss": 0.4008, "step": 20781 }, { "epoch": 0.8711526571162089, "grad_norm": 2.224693775177002, "learning_rate": 4.28910387403379e-07, "loss": 0.4851, "step": 20782 }, { "epoch": 0.8711945757312178, "grad_norm": 1.978895902633667, "learning_rate": 4.2863534304634027e-07, "loss": 0.4955, "step": 20783 }, { "epoch": 0.8712364943462269, "grad_norm": 2.374040126800537, "learning_rate": 4.2836038295609004e-07, "loss": 0.4828, "step": 20784 }, { "epoch": 0.8712784129612358, "grad_norm": 1.7613122463226318, "learning_rate": 4.280855071376949e-07, "loss": 0.489, "step": 20785 }, { "epoch": 0.8713203315762447, "grad_norm": 2.3116326332092285, "learning_rate": 4.27810715596223e-07, "loss": 0.4751, "step": 20786 }, { "epoch": 0.8713622501912537, "grad_norm": 1.7363580465316772, "learning_rate": 4.275360083367408e-07, "loss": 0.4437, "step": 20787 }, { "epoch": 0.8714041688062626, "grad_norm": 1.547186255455017, "learning_rate": 4.272613853643104e-07, "loss": 0.4527, "step": 20788 }, { "epoch": 0.8714460874212716, "grad_norm": 1.8697309494018555, "learning_rate": 4.269868466839955e-07, "loss": 0.4551, "step": 20789 }, { "epoch": 0.8714880060362805, "grad_norm": 1.7608288526535034, "learning_rate": 4.267123923008548e-07, "loss": 0.4657, "step": 20790 }, { "epoch": 0.8715299246512895, "grad_norm": 1.7629514932632446, "learning_rate": 4.264380222199499e-07, "loss": 0.4584, "step": 20791 }, { "epoch": 0.8715718432662984, "grad_norm": 1.6726889610290527, "learning_rate": 4.261637364463367e-07, "loss": 0.4887, "step": 20792 }, { "epoch": 0.8716137618813075, "grad_norm": 1.7907860279083252, "learning_rate": 4.2588953498507223e-07, "loss": 0.4149, "step": 20793 }, { "epoch": 0.8716556804963164, "grad_norm": 2.141606569290161, "learning_rate": 4.256154178412103e-07, "loss": 0.4914, "step": 20794 }, { "epoch": 0.8716975991113254, "grad_norm": 1.717076063156128, "learning_rate": 4.25341385019804e-07, "loss": 0.4449, "step": 20795 }, { "epoch": 0.8717395177263343, "grad_norm": 2.010707139968872, "learning_rate": 4.2506743652590555e-07, "loss": 0.478, "step": 20796 }, { "epoch": 0.8717814363413433, "grad_norm": 2.290315866470337, "learning_rate": 4.247935723645641e-07, "loss": 0.4755, "step": 20797 }, { "epoch": 0.8718233549563522, "grad_norm": 1.9207708835601807, "learning_rate": 4.2451979254082734e-07, "loss": 0.4429, "step": 20798 }, { "epoch": 0.8718652735713612, "grad_norm": 2.190638542175293, "learning_rate": 4.2424609705974397e-07, "loss": 0.4944, "step": 20799 }, { "epoch": 0.8719071921863701, "grad_norm": 2.4024641513824463, "learning_rate": 4.239724859263583e-07, "loss": 0.4215, "step": 20800 }, { "epoch": 0.8719491108013792, "grad_norm": 2.0281453132629395, "learning_rate": 4.236989591457125e-07, "loss": 0.4608, "step": 20801 }, { "epoch": 0.8719910294163881, "grad_norm": 2.0429935455322266, "learning_rate": 4.2342551672285073e-07, "loss": 0.4692, "step": 20802 }, { "epoch": 0.8720329480313971, "grad_norm": 2.000143527984619, "learning_rate": 4.2315215866281177e-07, "loss": 0.4486, "step": 20803 }, { "epoch": 0.872074866646406, "grad_norm": 1.8750524520874023, "learning_rate": 4.2287888497063556e-07, "loss": 0.503, "step": 20804 }, { "epoch": 0.872116785261415, "grad_norm": 2.1702277660369873, "learning_rate": 4.2260569565136023e-07, "loss": 0.5149, "step": 20805 }, { "epoch": 0.8721587038764239, "grad_norm": 1.963923692703247, "learning_rate": 4.223325907100201e-07, "loss": 0.4605, "step": 20806 }, { "epoch": 0.8722006224914329, "grad_norm": 3.1945621967315674, "learning_rate": 4.2205957015165e-07, "loss": 0.4682, "step": 20807 }, { "epoch": 0.8722425411064418, "grad_norm": 1.5875588655471802, "learning_rate": 4.2178663398128426e-07, "loss": 0.4773, "step": 20808 }, { "epoch": 0.8722844597214509, "grad_norm": 1.8171343803405762, "learning_rate": 4.215137822039522e-07, "loss": 0.4944, "step": 20809 }, { "epoch": 0.8723263783364598, "grad_norm": 1.8394604921340942, "learning_rate": 4.2124101482468314e-07, "loss": 0.4772, "step": 20810 }, { "epoch": 0.8723682969514687, "grad_norm": 2.0360634326934814, "learning_rate": 4.209683318485069e-07, "loss": 0.5046, "step": 20811 }, { "epoch": 0.8724102155664777, "grad_norm": 1.781977653503418, "learning_rate": 4.2069573328044846e-07, "loss": 0.5159, "step": 20812 }, { "epoch": 0.8724521341814866, "grad_norm": 1.8677438497543335, "learning_rate": 4.2042321912553365e-07, "loss": 0.4916, "step": 20813 }, { "epoch": 0.8724940527964956, "grad_norm": 1.8090667724609375, "learning_rate": 4.201507893887863e-07, "loss": 0.4406, "step": 20814 }, { "epoch": 0.8725359714115045, "grad_norm": 1.5525082349777222, "learning_rate": 4.1987844407522684e-07, "loss": 0.4628, "step": 20815 }, { "epoch": 0.8725778900265135, "grad_norm": 1.6983656883239746, "learning_rate": 4.196061831898768e-07, "loss": 0.4567, "step": 20816 }, { "epoch": 0.8726198086415224, "grad_norm": 1.737795114517212, "learning_rate": 4.193340067377549e-07, "loss": 0.414, "step": 20817 }, { "epoch": 0.8726617272565315, "grad_norm": 1.9380526542663574, "learning_rate": 4.190619147238784e-07, "loss": 0.4977, "step": 20818 }, { "epoch": 0.8727036458715404, "grad_norm": 1.7275141477584839, "learning_rate": 4.187899071532614e-07, "loss": 0.475, "step": 20819 }, { "epoch": 0.8727455644865494, "grad_norm": 1.6934804916381836, "learning_rate": 4.1851798403092005e-07, "loss": 0.4982, "step": 20820 }, { "epoch": 0.8727874831015583, "grad_norm": 1.6364493370056152, "learning_rate": 4.1824614536186526e-07, "loss": 0.4826, "step": 20821 }, { "epoch": 0.8728294017165673, "grad_norm": 1.9588532447814941, "learning_rate": 4.179743911511086e-07, "loss": 0.4977, "step": 20822 }, { "epoch": 0.8728713203315762, "grad_norm": 1.6127222776412964, "learning_rate": 4.1770272140366055e-07, "loss": 0.4402, "step": 20823 }, { "epoch": 0.8729132389465852, "grad_norm": 2.6036155223846436, "learning_rate": 4.17431136124527e-07, "loss": 0.5205, "step": 20824 }, { "epoch": 0.8729551575615941, "grad_norm": 2.419698476791382, "learning_rate": 4.171596353187163e-07, "loss": 0.4372, "step": 20825 }, { "epoch": 0.8729970761766032, "grad_norm": 1.737602710723877, "learning_rate": 4.16888218991231e-07, "loss": 0.5173, "step": 20826 }, { "epoch": 0.8730389947916121, "grad_norm": 2.142946481704712, "learning_rate": 4.166168871470766e-07, "loss": 0.4686, "step": 20827 }, { "epoch": 0.8730809134066211, "grad_norm": 2.228740930557251, "learning_rate": 4.1634563979125297e-07, "loss": 0.4405, "step": 20828 }, { "epoch": 0.87312283202163, "grad_norm": 2.3453304767608643, "learning_rate": 4.160744769287606e-07, "loss": 0.4831, "step": 20829 }, { "epoch": 0.873164750636639, "grad_norm": 1.9947324991226196, "learning_rate": 4.1580339856459764e-07, "loss": 0.4808, "step": 20830 }, { "epoch": 0.8732066692516479, "grad_norm": 1.8759537935256958, "learning_rate": 4.155324047037629e-07, "loss": 0.4344, "step": 20831 }, { "epoch": 0.8732485878666569, "grad_norm": 1.9469938278198242, "learning_rate": 4.1526149535124905e-07, "loss": 0.5275, "step": 20832 }, { "epoch": 0.8732905064816658, "grad_norm": 2.164585828781128, "learning_rate": 4.1499067051205157e-07, "loss": 0.5206, "step": 20833 }, { "epoch": 0.8733324250966749, "grad_norm": 1.8868309259414673, "learning_rate": 4.1471993019116364e-07, "loss": 0.4343, "step": 20834 }, { "epoch": 0.8733743437116838, "grad_norm": 2.103874683380127, "learning_rate": 4.144492743935735e-07, "loss": 0.449, "step": 20835 }, { "epoch": 0.8734162623266927, "grad_norm": 1.7099297046661377, "learning_rate": 4.141787031242728e-07, "loss": 0.481, "step": 20836 }, { "epoch": 0.8734581809417017, "grad_norm": 1.9474616050720215, "learning_rate": 4.1390821638824683e-07, "loss": 0.4761, "step": 20837 }, { "epoch": 0.8735000995567106, "grad_norm": 4.003272533416748, "learning_rate": 4.13637814190484e-07, "loss": 0.4702, "step": 20838 }, { "epoch": 0.8735420181717196, "grad_norm": 2.2556817531585693, "learning_rate": 4.133674965359663e-07, "loss": 0.4474, "step": 20839 }, { "epoch": 0.8735839367867285, "grad_norm": 1.7443617582321167, "learning_rate": 4.1309726342967926e-07, "loss": 0.4599, "step": 20840 }, { "epoch": 0.8736258554017375, "grad_norm": 2.103550672531128, "learning_rate": 4.128271148766022e-07, "loss": 0.4831, "step": 20841 }, { "epoch": 0.8736677740167464, "grad_norm": 1.6579070091247559, "learning_rate": 4.125570508817156e-07, "loss": 0.442, "step": 20842 }, { "epoch": 0.8737096926317555, "grad_norm": 2.124011993408203, "learning_rate": 4.1228707144999823e-07, "loss": 0.4701, "step": 20843 }, { "epoch": 0.8737516112467644, "grad_norm": 1.7993091344833374, "learning_rate": 4.1201717658642614e-07, "loss": 0.488, "step": 20844 }, { "epoch": 0.8737935298617734, "grad_norm": 2.019651412963867, "learning_rate": 4.117473662959742e-07, "loss": 0.4779, "step": 20845 }, { "epoch": 0.8738354484767823, "grad_norm": 3.264139175415039, "learning_rate": 4.114776405836179e-07, "loss": 0.4602, "step": 20846 }, { "epoch": 0.8738773670917913, "grad_norm": 2.371699094772339, "learning_rate": 4.112079994543272e-07, "loss": 0.4701, "step": 20847 }, { "epoch": 0.8739192857068002, "grad_norm": 1.8714662790298462, "learning_rate": 4.1093844291307305e-07, "loss": 0.4771, "step": 20848 }, { "epoch": 0.8739612043218092, "grad_norm": 1.7745195627212524, "learning_rate": 4.106689709648248e-07, "loss": 0.4394, "step": 20849 }, { "epoch": 0.8740031229368181, "grad_norm": 1.6376253366470337, "learning_rate": 4.1039958361454913e-07, "loss": 0.4517, "step": 20850 }, { "epoch": 0.8740450415518272, "grad_norm": 1.774584412574768, "learning_rate": 4.1013028086721196e-07, "loss": 0.4438, "step": 20851 }, { "epoch": 0.8740869601668361, "grad_norm": 1.9987608194351196, "learning_rate": 4.0986106272777826e-07, "loss": 0.4455, "step": 20852 }, { "epoch": 0.8741288787818451, "grad_norm": 1.7968623638153076, "learning_rate": 4.095919292012096e-07, "loss": 0.4767, "step": 20853 }, { "epoch": 0.874170797396854, "grad_norm": 2.038618326187134, "learning_rate": 4.093228802924681e-07, "loss": 0.4659, "step": 20854 }, { "epoch": 0.874212716011863, "grad_norm": 1.920681118965149, "learning_rate": 4.0905391600651325e-07, "loss": 0.4698, "step": 20855 }, { "epoch": 0.8742546346268719, "grad_norm": 1.7327162027359009, "learning_rate": 4.087850363483031e-07, "loss": 0.4732, "step": 20856 }, { "epoch": 0.8742965532418809, "grad_norm": 2.025482654571533, "learning_rate": 4.085162413227922e-07, "loss": 0.4745, "step": 20857 }, { "epoch": 0.8743384718568898, "grad_norm": 2.0063297748565674, "learning_rate": 4.082475309349382e-07, "loss": 0.483, "step": 20858 }, { "epoch": 0.8743803904718989, "grad_norm": 2.3321609497070312, "learning_rate": 4.0797890518969205e-07, "loss": 0.5142, "step": 20859 }, { "epoch": 0.8744223090869078, "grad_norm": 1.726741075515747, "learning_rate": 4.077103640920066e-07, "loss": 0.4793, "step": 20860 }, { "epoch": 0.8744642277019167, "grad_norm": 1.6365599632263184, "learning_rate": 4.0744190764683225e-07, "loss": 0.4638, "step": 20861 }, { "epoch": 0.8745061463169257, "grad_norm": 1.911630392074585, "learning_rate": 4.071735358591172e-07, "loss": 0.4699, "step": 20862 }, { "epoch": 0.8745480649319346, "grad_norm": 2.0474510192871094, "learning_rate": 4.069052487338082e-07, "loss": 0.5462, "step": 20863 }, { "epoch": 0.8745899835469436, "grad_norm": 2.0561492443084717, "learning_rate": 4.0663704627585176e-07, "loss": 0.4462, "step": 20864 }, { "epoch": 0.8746319021619525, "grad_norm": 1.896866798400879, "learning_rate": 4.063689284901917e-07, "loss": 0.4354, "step": 20865 }, { "epoch": 0.8746738207769615, "grad_norm": 2.494147777557373, "learning_rate": 4.061008953817691e-07, "loss": 0.4817, "step": 20866 }, { "epoch": 0.8747157393919704, "grad_norm": 5.549930095672607, "learning_rate": 4.0583294695552667e-07, "loss": 0.5082, "step": 20867 }, { "epoch": 0.8747576580069795, "grad_norm": 1.9200373888015747, "learning_rate": 4.05565083216401e-07, "loss": 0.4527, "step": 20868 }, { "epoch": 0.8747995766219884, "grad_norm": 2.092275381088257, "learning_rate": 4.05297304169332e-07, "loss": 0.4713, "step": 20869 }, { "epoch": 0.8748414952369974, "grad_norm": 2.57698655128479, "learning_rate": 4.0502960981925575e-07, "loss": 0.4594, "step": 20870 }, { "epoch": 0.8748834138520063, "grad_norm": 1.7309644222259521, "learning_rate": 4.0476200017110614e-07, "loss": 0.4596, "step": 20871 }, { "epoch": 0.8749253324670153, "grad_norm": 1.9329661130905151, "learning_rate": 4.044944752298152e-07, "loss": 0.4711, "step": 20872 }, { "epoch": 0.8749672510820242, "grad_norm": 2.940791130065918, "learning_rate": 4.042270350003175e-07, "loss": 0.4537, "step": 20873 }, { "epoch": 0.8750091696970332, "grad_norm": 2.1188175678253174, "learning_rate": 4.0395967948754e-07, "loss": 0.5005, "step": 20874 }, { "epoch": 0.8750510883120421, "grad_norm": 1.8889027833938599, "learning_rate": 4.036924086964117e-07, "loss": 0.4734, "step": 20875 }, { "epoch": 0.8750930069270512, "grad_norm": 1.9744118452072144, "learning_rate": 4.0342522263186025e-07, "loss": 0.5012, "step": 20876 }, { "epoch": 0.8751349255420601, "grad_norm": 1.7774920463562012, "learning_rate": 4.031581212988095e-07, "loss": 0.4891, "step": 20877 }, { "epoch": 0.8751768441570691, "grad_norm": 1.7241836786270142, "learning_rate": 4.0289110470218387e-07, "loss": 0.4635, "step": 20878 }, { "epoch": 0.875218762772078, "grad_norm": 2.096113681793213, "learning_rate": 4.0262417284690603e-07, "loss": 0.4923, "step": 20879 }, { "epoch": 0.875260681387087, "grad_norm": 2.0849361419677734, "learning_rate": 4.0235732573789545e-07, "loss": 0.4821, "step": 20880 }, { "epoch": 0.8753026000020959, "grad_norm": 1.846625566482544, "learning_rate": 4.020905633800709e-07, "loss": 0.4679, "step": 20881 }, { "epoch": 0.8753445186171049, "grad_norm": 1.9060876369476318, "learning_rate": 4.0182388577835185e-07, "loss": 0.5285, "step": 20882 }, { "epoch": 0.8753864372321138, "grad_norm": 1.8923906087875366, "learning_rate": 4.015572929376527e-07, "loss": 0.4555, "step": 20883 }, { "epoch": 0.8754283558471229, "grad_norm": 1.9703248739242554, "learning_rate": 4.0129078486288665e-07, "loss": 0.4866, "step": 20884 }, { "epoch": 0.8754702744621318, "grad_norm": 1.7701547145843506, "learning_rate": 4.010243615589682e-07, "loss": 0.4495, "step": 20885 }, { "epoch": 0.8755121930771407, "grad_norm": 1.8658982515335083, "learning_rate": 4.0075802303080667e-07, "loss": 0.4915, "step": 20886 }, { "epoch": 0.8755541116921497, "grad_norm": 1.8269649744033813, "learning_rate": 4.004917692833132e-07, "loss": 0.479, "step": 20887 }, { "epoch": 0.8755960303071586, "grad_norm": 1.715566873550415, "learning_rate": 4.002256003213961e-07, "loss": 0.445, "step": 20888 }, { "epoch": 0.8756379489221676, "grad_norm": 1.8444517850875854, "learning_rate": 3.9995951614995975e-07, "loss": 0.4631, "step": 20889 }, { "epoch": 0.8756798675371765, "grad_norm": 1.9318276643753052, "learning_rate": 3.9969351677391076e-07, "loss": 0.4077, "step": 20890 }, { "epoch": 0.8757217861521855, "grad_norm": 2.94209361076355, "learning_rate": 3.9942760219815304e-07, "loss": 0.4424, "step": 20891 }, { "epoch": 0.8757637047671945, "grad_norm": 2.028170585632324, "learning_rate": 3.9916177242758647e-07, "loss": 0.4034, "step": 20892 }, { "epoch": 0.8758056233822035, "grad_norm": 2.390277624130249, "learning_rate": 3.988960274671128e-07, "loss": 0.4575, "step": 20893 }, { "epoch": 0.8758475419972124, "grad_norm": 1.8251967430114746, "learning_rate": 3.986303673216302e-07, "loss": 0.4849, "step": 20894 }, { "epoch": 0.8758894606122214, "grad_norm": 1.9316816329956055, "learning_rate": 3.983647919960343e-07, "loss": 0.4774, "step": 20895 }, { "epoch": 0.8759313792272303, "grad_norm": 1.8602194786071777, "learning_rate": 3.9809930149522226e-07, "loss": 0.4715, "step": 20896 }, { "epoch": 0.8759732978422393, "grad_norm": 1.8408123254776, "learning_rate": 3.9783389582408847e-07, "loss": 0.4505, "step": 20897 }, { "epoch": 0.8760152164572482, "grad_norm": 2.1095309257507324, "learning_rate": 3.9756857498752357e-07, "loss": 0.4382, "step": 20898 }, { "epoch": 0.8760571350722572, "grad_norm": 2.0894811153411865, "learning_rate": 3.973033389904202e-07, "loss": 0.4604, "step": 20899 }, { "epoch": 0.8760990536872661, "grad_norm": 1.9130291938781738, "learning_rate": 3.9703818783766614e-07, "loss": 0.4476, "step": 20900 }, { "epoch": 0.8761409723022752, "grad_norm": 1.9024105072021484, "learning_rate": 3.9677312153414916e-07, "loss": 0.4703, "step": 20901 }, { "epoch": 0.8761828909172841, "grad_norm": 2.002840518951416, "learning_rate": 3.965081400847576e-07, "loss": 0.4979, "step": 20902 }, { "epoch": 0.8762248095322931, "grad_norm": 1.886824131011963, "learning_rate": 3.962432434943736e-07, "loss": 0.45, "step": 20903 }, { "epoch": 0.876266728147302, "grad_norm": 1.8770256042480469, "learning_rate": 3.959784317678805e-07, "loss": 0.451, "step": 20904 }, { "epoch": 0.876308646762311, "grad_norm": 2.7132937908172607, "learning_rate": 3.957137049101606e-07, "loss": 0.4657, "step": 20905 }, { "epoch": 0.8763505653773199, "grad_norm": 1.9391978979110718, "learning_rate": 3.954490629260926e-07, "loss": 0.4898, "step": 20906 }, { "epoch": 0.8763924839923289, "grad_norm": 7.927518367767334, "learning_rate": 3.9518450582055555e-07, "loss": 0.4862, "step": 20907 }, { "epoch": 0.8764344026073378, "grad_norm": 1.959398865699768, "learning_rate": 3.9492003359842714e-07, "loss": 0.4838, "step": 20908 }, { "epoch": 0.8764763212223469, "grad_norm": 2.1977503299713135, "learning_rate": 3.9465564626458064e-07, "loss": 0.4842, "step": 20909 }, { "epoch": 0.8765182398373558, "grad_norm": 2.2311923503875732, "learning_rate": 3.9439134382389055e-07, "loss": 0.4576, "step": 20910 }, { "epoch": 0.8765601584523647, "grad_norm": 1.9476019144058228, "learning_rate": 3.941271262812296e-07, "loss": 0.4758, "step": 20911 }, { "epoch": 0.8766020770673737, "grad_norm": 1.9764838218688965, "learning_rate": 3.9386299364146786e-07, "loss": 0.4706, "step": 20912 }, { "epoch": 0.8766439956823826, "grad_norm": 2.437054395675659, "learning_rate": 3.935989459094736e-07, "loss": 0.4983, "step": 20913 }, { "epoch": 0.8766859142973916, "grad_norm": 1.9232145547866821, "learning_rate": 3.933349830901151e-07, "loss": 0.4882, "step": 20914 }, { "epoch": 0.8767278329124005, "grad_norm": 2.552150011062622, "learning_rate": 3.930711051882569e-07, "loss": 0.4583, "step": 20915 }, { "epoch": 0.8767697515274095, "grad_norm": 1.591103434562683, "learning_rate": 3.928073122087639e-07, "loss": 0.448, "step": 20916 }, { "epoch": 0.8768116701424185, "grad_norm": 1.8827489614486694, "learning_rate": 3.925436041564995e-07, "loss": 0.4832, "step": 20917 }, { "epoch": 0.8768535887574275, "grad_norm": 1.9947223663330078, "learning_rate": 3.922799810363237e-07, "loss": 0.4894, "step": 20918 }, { "epoch": 0.8768955073724364, "grad_norm": 2.0185439586639404, "learning_rate": 3.920164428530959e-07, "loss": 0.4398, "step": 20919 }, { "epoch": 0.8769374259874454, "grad_norm": 1.7165333032608032, "learning_rate": 3.917529896116762e-07, "loss": 0.4839, "step": 20920 }, { "epoch": 0.8769793446024543, "grad_norm": 1.8185151815414429, "learning_rate": 3.9148962131691893e-07, "loss": 0.4576, "step": 20921 }, { "epoch": 0.8770212632174633, "grad_norm": 1.7913472652435303, "learning_rate": 3.912263379736786e-07, "loss": 0.5013, "step": 20922 }, { "epoch": 0.8770631818324722, "grad_norm": 1.9876147508621216, "learning_rate": 3.909631395868102e-07, "loss": 0.4966, "step": 20923 }, { "epoch": 0.8771051004474812, "grad_norm": 1.8941624164581299, "learning_rate": 3.9070002616116375e-07, "loss": 0.5007, "step": 20924 }, { "epoch": 0.8771470190624902, "grad_norm": 1.5447732210159302, "learning_rate": 3.904369977015898e-07, "loss": 0.4255, "step": 20925 }, { "epoch": 0.8771889376774992, "grad_norm": 1.8780734539031982, "learning_rate": 3.9017405421293894e-07, "loss": 0.4859, "step": 20926 }, { "epoch": 0.8772308562925081, "grad_norm": 1.9069290161132812, "learning_rate": 3.8991119570005507e-07, "loss": 0.4956, "step": 20927 }, { "epoch": 0.8772727749075171, "grad_norm": 2.2718539237976074, "learning_rate": 3.896484221677849e-07, "loss": 0.4603, "step": 20928 }, { "epoch": 0.877314693522526, "grad_norm": 1.7812715768814087, "learning_rate": 3.893857336209733e-07, "loss": 0.474, "step": 20929 }, { "epoch": 0.877356612137535, "grad_norm": 1.794501543045044, "learning_rate": 3.891231300644621e-07, "loss": 0.4961, "step": 20930 }, { "epoch": 0.8773985307525439, "grad_norm": 2.18493390083313, "learning_rate": 3.888606115030907e-07, "loss": 0.4898, "step": 20931 }, { "epoch": 0.8774404493675529, "grad_norm": 3.0185976028442383, "learning_rate": 3.885981779417003e-07, "loss": 0.5267, "step": 20932 }, { "epoch": 0.8774823679825619, "grad_norm": 1.839419960975647, "learning_rate": 3.883358293851258e-07, "loss": 0.4381, "step": 20933 }, { "epoch": 0.8775242865975709, "grad_norm": 1.8199548721313477, "learning_rate": 3.8807356583820567e-07, "loss": 0.4863, "step": 20934 }, { "epoch": 0.8775662052125798, "grad_norm": 2.084751605987549, "learning_rate": 3.8781138730577374e-07, "loss": 0.4801, "step": 20935 }, { "epoch": 0.8776081238275887, "grad_norm": 1.9058514833450317, "learning_rate": 3.8754929379266225e-07, "loss": 0.4515, "step": 20936 }, { "epoch": 0.8776500424425977, "grad_norm": 2.0724241733551025, "learning_rate": 3.8728728530370294e-07, "loss": 0.4571, "step": 20937 }, { "epoch": 0.8776919610576066, "grad_norm": 1.789437174797058, "learning_rate": 3.870253618437264e-07, "loss": 0.4798, "step": 20938 }, { "epoch": 0.8777338796726156, "grad_norm": 1.8573451042175293, "learning_rate": 3.867635234175593e-07, "loss": 0.491, "step": 20939 }, { "epoch": 0.8777757982876245, "grad_norm": 2.530299425125122, "learning_rate": 3.8650177003002995e-07, "loss": 0.4737, "step": 20940 }, { "epoch": 0.8778177169026335, "grad_norm": 2.401737928390503, "learning_rate": 3.8624010168596237e-07, "loss": 0.48, "step": 20941 }, { "epoch": 0.8778596355176425, "grad_norm": 2.1232423782348633, "learning_rate": 3.8597851839017984e-07, "loss": 0.4725, "step": 20942 }, { "epoch": 0.8779015541326515, "grad_norm": 1.9325673580169678, "learning_rate": 3.857170201475041e-07, "loss": 0.4285, "step": 20943 }, { "epoch": 0.8779434727476604, "grad_norm": 4.760945796966553, "learning_rate": 3.8545560696275687e-07, "loss": 0.4752, "step": 20944 }, { "epoch": 0.8779853913626694, "grad_norm": 2.3583574295043945, "learning_rate": 3.8519427884075535e-07, "loss": 0.4982, "step": 20945 }, { "epoch": 0.8780273099776783, "grad_norm": 2.0813677310943604, "learning_rate": 3.8493303578631793e-07, "loss": 0.4513, "step": 20946 }, { "epoch": 0.8780692285926873, "grad_norm": 1.7781975269317627, "learning_rate": 3.8467187780426016e-07, "loss": 0.4235, "step": 20947 }, { "epoch": 0.8781111472076962, "grad_norm": 1.8589597940444946, "learning_rate": 3.84410804899395e-07, "loss": 0.5141, "step": 20948 }, { "epoch": 0.8781530658227052, "grad_norm": 2.314368963241577, "learning_rate": 3.8414981707653677e-07, "loss": 0.4879, "step": 20949 }, { "epoch": 0.8781949844377142, "grad_norm": 1.7935779094696045, "learning_rate": 3.8388891434049555e-07, "loss": 0.4216, "step": 20950 }, { "epoch": 0.8782369030527232, "grad_norm": 1.7189183235168457, "learning_rate": 3.836280966960798e-07, "loss": 0.4505, "step": 20951 }, { "epoch": 0.8782788216677321, "grad_norm": 1.9509823322296143, "learning_rate": 3.833673641480978e-07, "loss": 0.4535, "step": 20952 }, { "epoch": 0.8783207402827411, "grad_norm": 1.9204376935958862, "learning_rate": 3.8310671670135746e-07, "loss": 0.4498, "step": 20953 }, { "epoch": 0.87836265889775, "grad_norm": 2.390127658843994, "learning_rate": 3.8284615436066095e-07, "loss": 0.503, "step": 20954 }, { "epoch": 0.878404577512759, "grad_norm": 1.723808765411377, "learning_rate": 3.825856771308123e-07, "loss": 0.4369, "step": 20955 }, { "epoch": 0.8784464961277679, "grad_norm": 1.799194574356079, "learning_rate": 3.823252850166148e-07, "loss": 0.4932, "step": 20956 }, { "epoch": 0.878488414742777, "grad_norm": 2.198089599609375, "learning_rate": 3.820649780228658e-07, "loss": 0.4679, "step": 20957 }, { "epoch": 0.8785303333577859, "grad_norm": 2.1616852283477783, "learning_rate": 3.8180475615436586e-07, "loss": 0.4961, "step": 20958 }, { "epoch": 0.8785722519727949, "grad_norm": 2.1243772506713867, "learning_rate": 3.815446194159106e-07, "loss": 0.5041, "step": 20959 }, { "epoch": 0.8786141705878038, "grad_norm": 2.4216926097869873, "learning_rate": 3.812845678122945e-07, "loss": 0.479, "step": 20960 }, { "epoch": 0.8786560892028127, "grad_norm": 1.6988149881362915, "learning_rate": 3.810246013483121e-07, "loss": 0.4348, "step": 20961 }, { "epoch": 0.8786980078178217, "grad_norm": 2.297713279724121, "learning_rate": 3.8076472002875677e-07, "loss": 0.47, "step": 20962 }, { "epoch": 0.8787399264328306, "grad_norm": 2.40753173828125, "learning_rate": 3.805049238584169e-07, "loss": 0.479, "step": 20963 }, { "epoch": 0.8787818450478396, "grad_norm": 1.9440351724624634, "learning_rate": 3.802452128420825e-07, "loss": 0.4845, "step": 20964 }, { "epoch": 0.8788237636628485, "grad_norm": 1.91280996799469, "learning_rate": 3.79985586984542e-07, "loss": 0.4807, "step": 20965 }, { "epoch": 0.8788656822778576, "grad_norm": 1.5133650302886963, "learning_rate": 3.797260462905794e-07, "loss": 0.4342, "step": 20966 }, { "epoch": 0.8789076008928665, "grad_norm": 2.0397872924804688, "learning_rate": 3.7946659076498024e-07, "loss": 0.4753, "step": 20967 }, { "epoch": 0.8789495195078755, "grad_norm": 1.8492542505264282, "learning_rate": 3.7920722041252743e-07, "loss": 0.4194, "step": 20968 }, { "epoch": 0.8789914381228844, "grad_norm": 2.022653579711914, "learning_rate": 3.7894793523799986e-07, "loss": 0.5165, "step": 20969 }, { "epoch": 0.8790333567378934, "grad_norm": 1.7574647665023804, "learning_rate": 3.7868873524617935e-07, "loss": 0.4854, "step": 20970 }, { "epoch": 0.8790752753529023, "grad_norm": 1.790940761566162, "learning_rate": 3.784296204418436e-07, "loss": 0.5242, "step": 20971 }, { "epoch": 0.8791171939679113, "grad_norm": 3.7434449195861816, "learning_rate": 3.781705908297684e-07, "loss": 0.4521, "step": 20972 }, { "epoch": 0.8791591125829202, "grad_norm": 2.2671566009521484, "learning_rate": 3.779116464147292e-07, "loss": 0.5061, "step": 20973 }, { "epoch": 0.8792010311979293, "grad_norm": 1.9530662298202515, "learning_rate": 3.776527872014979e-07, "loss": 0.4726, "step": 20974 }, { "epoch": 0.8792429498129382, "grad_norm": 1.9922317266464233, "learning_rate": 3.773940131948478e-07, "loss": 0.4375, "step": 20975 }, { "epoch": 0.8792848684279472, "grad_norm": 1.982163429260254, "learning_rate": 3.77135324399549e-07, "loss": 0.4487, "step": 20976 }, { "epoch": 0.8793267870429561, "grad_norm": 1.8641139268875122, "learning_rate": 3.7687672082037e-07, "loss": 0.459, "step": 20977 }, { "epoch": 0.8793687056579651, "grad_norm": 1.8954403400421143, "learning_rate": 3.7661820246207626e-07, "loss": 0.4681, "step": 20978 }, { "epoch": 0.879410624272974, "grad_norm": 3.757434844970703, "learning_rate": 3.7635976932943517e-07, "loss": 0.405, "step": 20979 }, { "epoch": 0.879452542887983, "grad_norm": 2.189577102661133, "learning_rate": 3.7610142142720907e-07, "loss": 0.4534, "step": 20980 }, { "epoch": 0.8794944615029919, "grad_norm": 2.018312931060791, "learning_rate": 3.7584315876016075e-07, "loss": 0.496, "step": 20981 }, { "epoch": 0.879536380118001, "grad_norm": 1.7334825992584229, "learning_rate": 3.75584981333052e-07, "loss": 0.4258, "step": 20982 }, { "epoch": 0.8795782987330099, "grad_norm": 1.8412449359893799, "learning_rate": 3.753268891506401e-07, "loss": 0.4632, "step": 20983 }, { "epoch": 0.8796202173480189, "grad_norm": 1.7540870904922485, "learning_rate": 3.7506888221768345e-07, "loss": 0.4984, "step": 20984 }, { "epoch": 0.8796621359630278, "grad_norm": 2.0513861179351807, "learning_rate": 3.7481096053893937e-07, "loss": 0.4441, "step": 20985 }, { "epoch": 0.8797040545780367, "grad_norm": 1.8453245162963867, "learning_rate": 3.745531241191602e-07, "loss": 0.4105, "step": 20986 }, { "epoch": 0.8797459731930457, "grad_norm": 1.8197790384292603, "learning_rate": 3.742953729631005e-07, "loss": 0.4568, "step": 20987 }, { "epoch": 0.8797878918080546, "grad_norm": 3.2198550701141357, "learning_rate": 3.740377070755108e-07, "loss": 0.4583, "step": 20988 }, { "epoch": 0.8798298104230636, "grad_norm": 1.9313422441482544, "learning_rate": 3.737801264611396e-07, "loss": 0.4713, "step": 20989 }, { "epoch": 0.8798717290380725, "grad_norm": 2.6564762592315674, "learning_rate": 3.735226311247364e-07, "loss": 0.4979, "step": 20990 }, { "epoch": 0.8799136476530816, "grad_norm": 4.707306861877441, "learning_rate": 3.732652210710486e-07, "loss": 0.5007, "step": 20991 }, { "epoch": 0.8799555662680905, "grad_norm": 1.6772139072418213, "learning_rate": 3.730078963048189e-07, "loss": 0.4506, "step": 20992 }, { "epoch": 0.8799974848830995, "grad_norm": 1.9154045581817627, "learning_rate": 3.7275065683079257e-07, "loss": 0.4371, "step": 20993 }, { "epoch": 0.8800394034981084, "grad_norm": 2.077352285385132, "learning_rate": 3.724935026537113e-07, "loss": 0.4905, "step": 20994 }, { "epoch": 0.8800813221131174, "grad_norm": 2.049713373184204, "learning_rate": 3.7223643377831463e-07, "loss": 0.4537, "step": 20995 }, { "epoch": 0.8801232407281263, "grad_norm": 2.0843217372894287, "learning_rate": 3.719794502093421e-07, "loss": 0.4567, "step": 20996 }, { "epoch": 0.8801651593431353, "grad_norm": 2.1459689140319824, "learning_rate": 3.717225519515305e-07, "loss": 0.4432, "step": 20997 }, { "epoch": 0.8802070779581442, "grad_norm": 1.9399211406707764, "learning_rate": 3.7146573900961433e-07, "loss": 0.462, "step": 20998 }, { "epoch": 0.8802489965731533, "grad_norm": 2.0415892601013184, "learning_rate": 3.712090113883282e-07, "loss": 0.4878, "step": 20999 }, { "epoch": 0.8802909151881622, "grad_norm": 1.8361608982086182, "learning_rate": 3.7095236909240604e-07, "loss": 0.5208, "step": 21000 }, { "epoch": 0.8803328338031712, "grad_norm": 2.351299524307251, "learning_rate": 3.706958121265764e-07, "loss": 0.4278, "step": 21001 }, { "epoch": 0.8803747524181801, "grad_norm": 1.7517640590667725, "learning_rate": 3.704393404955697e-07, "loss": 0.435, "step": 21002 }, { "epoch": 0.8804166710331891, "grad_norm": 2.064545154571533, "learning_rate": 3.7018295420411463e-07, "loss": 0.4671, "step": 21003 }, { "epoch": 0.880458589648198, "grad_norm": 1.8258134126663208, "learning_rate": 3.6992665325693565e-07, "loss": 0.4347, "step": 21004 }, { "epoch": 0.880500508263207, "grad_norm": 2.7706687450408936, "learning_rate": 3.6967043765875843e-07, "loss": 0.4958, "step": 21005 }, { "epoch": 0.8805424268782159, "grad_norm": 3.1668641567230225, "learning_rate": 3.694143074143053e-07, "loss": 0.5095, "step": 21006 }, { "epoch": 0.880584345493225, "grad_norm": 1.9925826787948608, "learning_rate": 3.6915826252829747e-07, "loss": 0.4942, "step": 21007 }, { "epoch": 0.8806262641082339, "grad_norm": 1.6357702016830444, "learning_rate": 3.689023030054545e-07, "loss": 0.493, "step": 21008 }, { "epoch": 0.8806681827232429, "grad_norm": 1.9805971384048462, "learning_rate": 3.686464288504965e-07, "loss": 0.5015, "step": 21009 }, { "epoch": 0.8807101013382518, "grad_norm": 2.317173957824707, "learning_rate": 3.6839064006813805e-07, "loss": 0.4534, "step": 21010 }, { "epoch": 0.8807520199532607, "grad_norm": 2.1104576587677, "learning_rate": 3.6813493666309483e-07, "loss": 0.5138, "step": 21011 }, { "epoch": 0.8807939385682697, "grad_norm": 1.9551560878753662, "learning_rate": 3.67879318640082e-07, "loss": 0.5011, "step": 21012 }, { "epoch": 0.8808358571832786, "grad_norm": 2.0619938373565674, "learning_rate": 3.67623786003809e-07, "loss": 0.5027, "step": 21013 }, { "epoch": 0.8808777757982876, "grad_norm": 1.7668613195419312, "learning_rate": 3.673683387589888e-07, "loss": 0.5104, "step": 21014 }, { "epoch": 0.8809196944132965, "grad_norm": 2.418835401535034, "learning_rate": 3.6711297691032823e-07, "loss": 0.512, "step": 21015 }, { "epoch": 0.8809616130283056, "grad_norm": 1.9723899364471436, "learning_rate": 3.6685770046253453e-07, "loss": 0.5029, "step": 21016 }, { "epoch": 0.8810035316433145, "grad_norm": 1.8377214670181274, "learning_rate": 3.666025094203146e-07, "loss": 0.4756, "step": 21017 }, { "epoch": 0.8810454502583235, "grad_norm": 2.0384819507598877, "learning_rate": 3.6634740378837184e-07, "loss": 0.499, "step": 21018 }, { "epoch": 0.8810873688733324, "grad_norm": 1.7928727865219116, "learning_rate": 3.660923835714086e-07, "loss": 0.4577, "step": 21019 }, { "epoch": 0.8811292874883414, "grad_norm": 1.7656888961791992, "learning_rate": 3.6583744877412565e-07, "loss": 0.4739, "step": 21020 }, { "epoch": 0.8811712061033503, "grad_norm": 1.9544117450714111, "learning_rate": 3.655825994012241e-07, "loss": 0.4341, "step": 21021 }, { "epoch": 0.8812131247183593, "grad_norm": 2.644129753112793, "learning_rate": 3.6532783545739915e-07, "loss": 0.4704, "step": 21022 }, { "epoch": 0.8812550433333682, "grad_norm": 1.7842479944229126, "learning_rate": 3.650731569473498e-07, "loss": 0.4134, "step": 21023 }, { "epoch": 0.8812969619483773, "grad_norm": 1.8150070905685425, "learning_rate": 3.648185638757679e-07, "loss": 0.5302, "step": 21024 }, { "epoch": 0.8813388805633862, "grad_norm": 2.8687684535980225, "learning_rate": 3.6456405624734847e-07, "loss": 0.5072, "step": 21025 }, { "epoch": 0.8813807991783952, "grad_norm": 1.6900078058242798, "learning_rate": 3.643096340667818e-07, "loss": 0.4381, "step": 21026 }, { "epoch": 0.8814227177934041, "grad_norm": 1.795593023300171, "learning_rate": 3.6405529733875953e-07, "loss": 0.5221, "step": 21027 }, { "epoch": 0.8814646364084131, "grad_norm": 1.8725252151489258, "learning_rate": 3.638010460679675e-07, "loss": 0.4436, "step": 21028 }, { "epoch": 0.881506555023422, "grad_norm": 1.8771413564682007, "learning_rate": 3.6354688025909415e-07, "loss": 0.445, "step": 21029 }, { "epoch": 0.881548473638431, "grad_norm": 1.7505292892456055, "learning_rate": 3.632927999168251e-07, "loss": 0.5103, "step": 21030 }, { "epoch": 0.8815903922534399, "grad_norm": 2.893019676208496, "learning_rate": 3.630388050458422e-07, "loss": 0.446, "step": 21031 }, { "epoch": 0.881632310868449, "grad_norm": 2.0198845863342285, "learning_rate": 3.627848956508295e-07, "loss": 0.4526, "step": 21032 }, { "epoch": 0.8816742294834579, "grad_norm": 2.3598692417144775, "learning_rate": 3.625310717364655e-07, "loss": 0.5013, "step": 21033 }, { "epoch": 0.8817161480984669, "grad_norm": 1.8952581882476807, "learning_rate": 3.622773333074314e-07, "loss": 0.498, "step": 21034 }, { "epoch": 0.8817580667134758, "grad_norm": 6.8617424964904785, "learning_rate": 3.6202368036840186e-07, "loss": 0.4861, "step": 21035 }, { "epoch": 0.8817999853284847, "grad_norm": 1.9492132663726807, "learning_rate": 3.6177011292405474e-07, "loss": 0.4746, "step": 21036 }, { "epoch": 0.8818419039434937, "grad_norm": 1.9447335004806519, "learning_rate": 3.6151663097906297e-07, "loss": 0.4718, "step": 21037 }, { "epoch": 0.8818838225585026, "grad_norm": 2.1408615112304688, "learning_rate": 3.612632345380995e-07, "loss": 0.4928, "step": 21038 }, { "epoch": 0.8819257411735116, "grad_norm": 1.5310630798339844, "learning_rate": 3.610099236058362e-07, "loss": 0.4594, "step": 21039 }, { "epoch": 0.8819676597885205, "grad_norm": 2.1815803050994873, "learning_rate": 3.607566981869409e-07, "loss": 0.4896, "step": 21040 }, { "epoch": 0.8820095784035296, "grad_norm": 2.4202165603637695, "learning_rate": 3.6050355828608275e-07, "loss": 0.4779, "step": 21041 }, { "epoch": 0.8820514970185385, "grad_norm": 1.659162998199463, "learning_rate": 3.6025050390792793e-07, "loss": 0.4457, "step": 21042 }, { "epoch": 0.8820934156335475, "grad_norm": 2.452791213989258, "learning_rate": 3.599975350571405e-07, "loss": 0.4404, "step": 21043 }, { "epoch": 0.8821353342485564, "grad_norm": 4.074539661407471, "learning_rate": 3.5974465173838403e-07, "loss": 0.4802, "step": 21044 }, { "epoch": 0.8821772528635654, "grad_norm": 1.928544521331787, "learning_rate": 3.5949185395632026e-07, "loss": 0.4296, "step": 21045 }, { "epoch": 0.8822191714785743, "grad_norm": 1.7854751348495483, "learning_rate": 3.5923914171560767e-07, "loss": 0.4719, "step": 21046 }, { "epoch": 0.8822610900935833, "grad_norm": 2.4388856887817383, "learning_rate": 3.589865150209071e-07, "loss": 0.5023, "step": 21047 }, { "epoch": 0.8823030087085922, "grad_norm": 1.768734097480774, "learning_rate": 3.58733973876873e-07, "loss": 0.4904, "step": 21048 }, { "epoch": 0.8823449273236013, "grad_norm": 2.0853195190429688, "learning_rate": 3.584815182881618e-07, "loss": 0.459, "step": 21049 }, { "epoch": 0.8823868459386102, "grad_norm": 2.1414389610290527, "learning_rate": 3.58229148259428e-07, "loss": 0.4757, "step": 21050 }, { "epoch": 0.8824287645536192, "grad_norm": 1.6887550354003906, "learning_rate": 3.5797686379532236e-07, "loss": 0.4437, "step": 21051 }, { "epoch": 0.8824706831686281, "grad_norm": 1.8251391649246216, "learning_rate": 3.5772466490049617e-07, "loss": 0.4857, "step": 21052 }, { "epoch": 0.8825126017836371, "grad_norm": 1.8498413562774658, "learning_rate": 3.5747255157959847e-07, "loss": 0.4855, "step": 21053 }, { "epoch": 0.882554520398646, "grad_norm": 1.8886144161224365, "learning_rate": 3.5722052383727503e-07, "loss": 0.4553, "step": 21054 }, { "epoch": 0.882596439013655, "grad_norm": 1.7410986423492432, "learning_rate": 3.5696858167817263e-07, "loss": 0.446, "step": 21055 }, { "epoch": 0.8826383576286639, "grad_norm": 1.8157635927200317, "learning_rate": 3.5671672510693644e-07, "loss": 0.515, "step": 21056 }, { "epoch": 0.882680276243673, "grad_norm": 2.010869026184082, "learning_rate": 3.564649541282078e-07, "loss": 0.483, "step": 21057 }, { "epoch": 0.8827221948586819, "grad_norm": 2.3947982788085938, "learning_rate": 3.5621326874662797e-07, "loss": 0.3942, "step": 21058 }, { "epoch": 0.8827641134736909, "grad_norm": 2.2083616256713867, "learning_rate": 3.559616689668377e-07, "loss": 0.4996, "step": 21059 }, { "epoch": 0.8828060320886998, "grad_norm": 1.7775111198425293, "learning_rate": 3.557101547934733e-07, "loss": 0.4438, "step": 21060 }, { "epoch": 0.8828479507037087, "grad_norm": 1.7639907598495483, "learning_rate": 3.5545872623117204e-07, "loss": 0.4898, "step": 21061 }, { "epoch": 0.8828898693187177, "grad_norm": 1.9213333129882812, "learning_rate": 3.552073832845682e-07, "loss": 0.5064, "step": 21062 }, { "epoch": 0.8829317879337266, "grad_norm": 1.8230377435684204, "learning_rate": 3.5495612595829456e-07, "loss": 0.4935, "step": 21063 }, { "epoch": 0.8829737065487356, "grad_norm": 2.013577461242676, "learning_rate": 3.54704954256983e-07, "loss": 0.4735, "step": 21064 }, { "epoch": 0.8830156251637445, "grad_norm": 1.8888388872146606, "learning_rate": 3.5445386818526437e-07, "loss": 0.4969, "step": 21065 }, { "epoch": 0.8830575437787536, "grad_norm": 1.9741239547729492, "learning_rate": 3.5420286774776546e-07, "loss": 0.4749, "step": 21066 }, { "epoch": 0.8830994623937625, "grad_norm": 3.6665823459625244, "learning_rate": 3.5395195294911477e-07, "loss": 0.4685, "step": 21067 }, { "epoch": 0.8831413810087715, "grad_norm": 2.340162754058838, "learning_rate": 3.5370112379393695e-07, "loss": 0.4724, "step": 21068 }, { "epoch": 0.8831832996237804, "grad_norm": 3.698524236679077, "learning_rate": 3.5345038028685494e-07, "loss": 0.4958, "step": 21069 }, { "epoch": 0.8832252182387894, "grad_norm": 3.0356359481811523, "learning_rate": 3.5319972243249233e-07, "loss": 0.4749, "step": 21070 }, { "epoch": 0.8832671368537983, "grad_norm": 1.7361748218536377, "learning_rate": 3.5294915023546815e-07, "loss": 0.4777, "step": 21071 }, { "epoch": 0.8833090554688073, "grad_norm": 1.8616963624954224, "learning_rate": 3.526986637004032e-07, "loss": 0.498, "step": 21072 }, { "epoch": 0.8833509740838162, "grad_norm": 1.7221704721450806, "learning_rate": 3.5244826283191204e-07, "loss": 0.4426, "step": 21073 }, { "epoch": 0.8833928926988253, "grad_norm": 1.791109561920166, "learning_rate": 3.5219794763461324e-07, "loss": 0.4513, "step": 21074 }, { "epoch": 0.8834348113138342, "grad_norm": 1.9285292625427246, "learning_rate": 3.5194771811311924e-07, "loss": 0.4298, "step": 21075 }, { "epoch": 0.8834767299288432, "grad_norm": 1.7497800588607788, "learning_rate": 3.516975742720435e-07, "loss": 0.4644, "step": 21076 }, { "epoch": 0.8835186485438521, "grad_norm": 1.8756675720214844, "learning_rate": 3.514475161159975e-07, "loss": 0.4365, "step": 21077 }, { "epoch": 0.8835605671588611, "grad_norm": 2.100163221359253, "learning_rate": 3.5119754364958904e-07, "loss": 0.4892, "step": 21078 }, { "epoch": 0.88360248577387, "grad_norm": 2.0085771083831787, "learning_rate": 3.5094765687742727e-07, "loss": 0.4763, "step": 21079 }, { "epoch": 0.883644404388879, "grad_norm": 2.1139891147613525, "learning_rate": 3.5069785580411907e-07, "loss": 0.4302, "step": 21080 }, { "epoch": 0.8836863230038879, "grad_norm": 2.0888590812683105, "learning_rate": 3.5044814043426855e-07, "loss": 0.4189, "step": 21081 }, { "epoch": 0.883728241618897, "grad_norm": 1.9142972230911255, "learning_rate": 3.501985107724781e-07, "loss": 0.4648, "step": 21082 }, { "epoch": 0.8837701602339059, "grad_norm": 2.103247880935669, "learning_rate": 3.499489668233508e-07, "loss": 0.5218, "step": 21083 }, { "epoch": 0.8838120788489149, "grad_norm": 1.6626336574554443, "learning_rate": 3.496995085914845e-07, "loss": 0.451, "step": 21084 }, { "epoch": 0.8838539974639238, "grad_norm": 1.6187996864318848, "learning_rate": 3.494501360814795e-07, "loss": 0.4037, "step": 21085 }, { "epoch": 0.8838959160789327, "grad_norm": 1.7130687236785889, "learning_rate": 3.4920084929793207e-07, "loss": 0.4528, "step": 21086 }, { "epoch": 0.8839378346939417, "grad_norm": 2.181899070739746, "learning_rate": 3.489516482454375e-07, "loss": 0.4848, "step": 21087 }, { "epoch": 0.8839797533089506, "grad_norm": 2.2498013973236084, "learning_rate": 3.487025329285887e-07, "loss": 0.5197, "step": 21088 }, { "epoch": 0.8840216719239596, "grad_norm": 1.9257200956344604, "learning_rate": 3.484535033519798e-07, "loss": 0.4232, "step": 21089 }, { "epoch": 0.8840635905389685, "grad_norm": 1.9285874366760254, "learning_rate": 3.482045595201999e-07, "loss": 0.5042, "step": 21090 }, { "epoch": 0.8841055091539776, "grad_norm": 1.842008113861084, "learning_rate": 3.4795570143783707e-07, "loss": 0.4352, "step": 21091 }, { "epoch": 0.8841474277689865, "grad_norm": 2.4711012840270996, "learning_rate": 3.4770692910948034e-07, "loss": 0.443, "step": 21092 }, { "epoch": 0.8841893463839955, "grad_norm": 2.054774045944214, "learning_rate": 3.474582425397138e-07, "loss": 0.5096, "step": 21093 }, { "epoch": 0.8842312649990044, "grad_norm": 1.8457566499710083, "learning_rate": 3.4720964173312334e-07, "loss": 0.4282, "step": 21094 }, { "epoch": 0.8842731836140134, "grad_norm": 1.6035114526748657, "learning_rate": 3.469611266942913e-07, "loss": 0.4144, "step": 21095 }, { "epoch": 0.8843151022290223, "grad_norm": 1.7829018831253052, "learning_rate": 3.4671269742779733e-07, "loss": 0.4958, "step": 21096 }, { "epoch": 0.8843570208440313, "grad_norm": 1.7821733951568604, "learning_rate": 3.4646435393822177e-07, "loss": 0.406, "step": 21097 }, { "epoch": 0.8843989394590402, "grad_norm": 2.0489401817321777, "learning_rate": 3.462160962301431e-07, "loss": 0.5325, "step": 21098 }, { "epoch": 0.8844408580740493, "grad_norm": 2.0877158641815186, "learning_rate": 3.459679243081376e-07, "loss": 0.4705, "step": 21099 }, { "epoch": 0.8844827766890582, "grad_norm": 2.459331750869751, "learning_rate": 3.4571983817677847e-07, "loss": 0.429, "step": 21100 }, { "epoch": 0.8845246953040672, "grad_norm": 1.9657769203186035, "learning_rate": 3.454718378406402e-07, "loss": 0.5198, "step": 21101 }, { "epoch": 0.8845666139190761, "grad_norm": 2.059067487716675, "learning_rate": 3.452239233042937e-07, "loss": 0.4729, "step": 21102 }, { "epoch": 0.8846085325340851, "grad_norm": 1.8225347995758057, "learning_rate": 3.4497609457230853e-07, "loss": 0.4924, "step": 21103 }, { "epoch": 0.884650451149094, "grad_norm": 1.936270833015442, "learning_rate": 3.44728351649255e-07, "loss": 0.4706, "step": 21104 }, { "epoch": 0.884692369764103, "grad_norm": 1.7033032178878784, "learning_rate": 3.4448069453969726e-07, "loss": 0.4455, "step": 21105 }, { "epoch": 0.8847342883791119, "grad_norm": 1.7846646308898926, "learning_rate": 3.4423312324820215e-07, "loss": 0.5019, "step": 21106 }, { "epoch": 0.884776206994121, "grad_norm": 4.623114585876465, "learning_rate": 3.439856377793338e-07, "loss": 0.4953, "step": 21107 }, { "epoch": 0.8848181256091299, "grad_norm": 1.7691459655761719, "learning_rate": 3.437382381376536e-07, "loss": 0.4675, "step": 21108 }, { "epoch": 0.8848600442241389, "grad_norm": 2.0995852947235107, "learning_rate": 3.434909243277207e-07, "loss": 0.4759, "step": 21109 }, { "epoch": 0.8849019628391478, "grad_norm": 1.6848222017288208, "learning_rate": 3.4324369635409636e-07, "loss": 0.4566, "step": 21110 }, { "epoch": 0.8849438814541567, "grad_norm": 1.9700584411621094, "learning_rate": 3.4299655422133594e-07, "loss": 0.4708, "step": 21111 }, { "epoch": 0.8849858000691657, "grad_norm": 1.81571364402771, "learning_rate": 3.427494979339957e-07, "loss": 0.4718, "step": 21112 }, { "epoch": 0.8850277186841746, "grad_norm": 2.077112913131714, "learning_rate": 3.4250252749663094e-07, "loss": 0.497, "step": 21113 }, { "epoch": 0.8850696372991836, "grad_norm": 1.8530153036117554, "learning_rate": 3.4225564291379245e-07, "loss": 0.4494, "step": 21114 }, { "epoch": 0.8851115559141925, "grad_norm": 1.9158650636672974, "learning_rate": 3.420088441900321e-07, "loss": 0.4617, "step": 21115 }, { "epoch": 0.8851534745292016, "grad_norm": 1.6049691438674927, "learning_rate": 3.4176213132989965e-07, "loss": 0.4354, "step": 21116 }, { "epoch": 0.8851953931442105, "grad_norm": 1.936937689781189, "learning_rate": 3.4151550433794253e-07, "loss": 0.495, "step": 21117 }, { "epoch": 0.8852373117592195, "grad_norm": 1.94093656539917, "learning_rate": 3.4126896321870664e-07, "loss": 0.4238, "step": 21118 }, { "epoch": 0.8852792303742284, "grad_norm": 1.794788122177124, "learning_rate": 3.4102250797673763e-07, "loss": 0.4667, "step": 21119 }, { "epoch": 0.8853211489892374, "grad_norm": 1.9877300262451172, "learning_rate": 3.4077613861657697e-07, "loss": 0.4595, "step": 21120 }, { "epoch": 0.8853630676042463, "grad_norm": 2.222231149673462, "learning_rate": 3.405298551427666e-07, "loss": 0.4871, "step": 21121 }, { "epoch": 0.8854049862192553, "grad_norm": 1.6807634830474854, "learning_rate": 3.402836575598478e-07, "loss": 0.4917, "step": 21122 }, { "epoch": 0.8854469048342642, "grad_norm": 1.820353627204895, "learning_rate": 3.400375458723576e-07, "loss": 0.4963, "step": 21123 }, { "epoch": 0.8854888234492733, "grad_norm": 2.0508670806884766, "learning_rate": 3.3979152008483343e-07, "loss": 0.5333, "step": 21124 }, { "epoch": 0.8855307420642822, "grad_norm": 1.646036982536316, "learning_rate": 3.395455802018094e-07, "loss": 0.469, "step": 21125 }, { "epoch": 0.8855726606792912, "grad_norm": 1.7698136568069458, "learning_rate": 3.3929972622781913e-07, "loss": 0.4658, "step": 21126 }, { "epoch": 0.8856145792943001, "grad_norm": 1.831931233406067, "learning_rate": 3.390539581673968e-07, "loss": 0.4329, "step": 21127 }, { "epoch": 0.8856564979093091, "grad_norm": 1.9843266010284424, "learning_rate": 3.388082760250705e-07, "loss": 0.4141, "step": 21128 }, { "epoch": 0.885698416524318, "grad_norm": 1.7370954751968384, "learning_rate": 3.3856267980536926e-07, "loss": 0.4743, "step": 21129 }, { "epoch": 0.885740335139327, "grad_norm": 1.9684199094772339, "learning_rate": 3.383171695128218e-07, "loss": 0.5624, "step": 21130 }, { "epoch": 0.885782253754336, "grad_norm": 1.909469723701477, "learning_rate": 3.380717451519516e-07, "loss": 0.4576, "step": 21131 }, { "epoch": 0.885824172369345, "grad_norm": 2.0757946968078613, "learning_rate": 3.378264067272841e-07, "loss": 0.4815, "step": 21132 }, { "epoch": 0.8858660909843539, "grad_norm": 2.026790142059326, "learning_rate": 3.375811542433421e-07, "loss": 0.5097, "step": 21133 }, { "epoch": 0.8859080095993629, "grad_norm": 1.9837300777435303, "learning_rate": 3.373359877046456e-07, "loss": 0.5115, "step": 21134 }, { "epoch": 0.8859499282143718, "grad_norm": 3.1615095138549805, "learning_rate": 3.3709090711571356e-07, "loss": 0.4854, "step": 21135 }, { "epoch": 0.8859918468293807, "grad_norm": 2.663029193878174, "learning_rate": 3.3684591248106526e-07, "loss": 0.4386, "step": 21136 }, { "epoch": 0.8860337654443897, "grad_norm": 2.4320695400238037, "learning_rate": 3.3660100380521643e-07, "loss": 0.4248, "step": 21137 }, { "epoch": 0.8860756840593986, "grad_norm": 2.313115119934082, "learning_rate": 3.363561810926802e-07, "loss": 0.4416, "step": 21138 }, { "epoch": 0.8861176026744076, "grad_norm": 1.9528387784957886, "learning_rate": 3.361114443479707e-07, "loss": 0.4837, "step": 21139 }, { "epoch": 0.8861595212894166, "grad_norm": 1.7801772356033325, "learning_rate": 3.358667935755988e-07, "loss": 0.4725, "step": 21140 }, { "epoch": 0.8862014399044256, "grad_norm": 1.9885705709457397, "learning_rate": 3.356222287800748e-07, "loss": 0.491, "step": 21141 }, { "epoch": 0.8862433585194345, "grad_norm": 2.650017738342285, "learning_rate": 3.353777499659072e-07, "loss": 0.4645, "step": 21142 }, { "epoch": 0.8862852771344435, "grad_norm": 1.9452497959136963, "learning_rate": 3.351333571376014e-07, "loss": 0.4997, "step": 21143 }, { "epoch": 0.8863271957494524, "grad_norm": 1.9655860662460327, "learning_rate": 3.348890502996632e-07, "loss": 0.4926, "step": 21144 }, { "epoch": 0.8863691143644614, "grad_norm": 2.079009771347046, "learning_rate": 3.346448294565968e-07, "loss": 0.4746, "step": 21145 }, { "epoch": 0.8864110329794703, "grad_norm": 1.6974155902862549, "learning_rate": 3.3440069461290304e-07, "loss": 0.4675, "step": 21146 }, { "epoch": 0.8864529515944793, "grad_norm": 1.9739259481430054, "learning_rate": 3.341566457730822e-07, "loss": 0.5083, "step": 21147 }, { "epoch": 0.8864948702094883, "grad_norm": 2.1091673374176025, "learning_rate": 3.339126829416334e-07, "loss": 0.4506, "step": 21148 }, { "epoch": 0.8865367888244973, "grad_norm": 2.1169052124023438, "learning_rate": 3.336688061230531e-07, "loss": 0.4395, "step": 21149 }, { "epoch": 0.8865787074395062, "grad_norm": 1.9437732696533203, "learning_rate": 3.334250153218377e-07, "loss": 0.4696, "step": 21150 }, { "epoch": 0.8866206260545152, "grad_norm": 1.9782984256744385, "learning_rate": 3.3318131054248136e-07, "loss": 0.4409, "step": 21151 }, { "epoch": 0.8866625446695241, "grad_norm": 5.793848037719727, "learning_rate": 3.32937691789475e-07, "loss": 0.5128, "step": 21152 }, { "epoch": 0.8867044632845331, "grad_norm": 1.6783514022827148, "learning_rate": 3.3269415906731053e-07, "loss": 0.4132, "step": 21153 }, { "epoch": 0.886746381899542, "grad_norm": 1.7008968591690063, "learning_rate": 3.324507123804771e-07, "loss": 0.412, "step": 21154 }, { "epoch": 0.886788300514551, "grad_norm": 1.8328031301498413, "learning_rate": 3.322073517334623e-07, "loss": 0.4603, "step": 21155 }, { "epoch": 0.88683021912956, "grad_norm": 1.9803240299224854, "learning_rate": 3.3196407713075143e-07, "loss": 0.4853, "step": 21156 }, { "epoch": 0.886872137744569, "grad_norm": 3.571556568145752, "learning_rate": 3.317208885768303e-07, "loss": 0.4745, "step": 21157 }, { "epoch": 0.8869140563595779, "grad_norm": 1.7507249116897583, "learning_rate": 3.314777860761798e-07, "loss": 0.4328, "step": 21158 }, { "epoch": 0.8869559749745869, "grad_norm": 1.8969060182571411, "learning_rate": 3.312347696332824e-07, "loss": 0.4817, "step": 21159 }, { "epoch": 0.8869978935895958, "grad_norm": 4.698701858520508, "learning_rate": 3.3099183925261845e-07, "loss": 0.4953, "step": 21160 }, { "epoch": 0.8870398122046047, "grad_norm": 1.796102523803711, "learning_rate": 3.307489949386644e-07, "loss": 0.4909, "step": 21161 }, { "epoch": 0.8870817308196137, "grad_norm": 4.257130146026611, "learning_rate": 3.305062366958978e-07, "loss": 0.4905, "step": 21162 }, { "epoch": 0.8871236494346226, "grad_norm": 10.466378211975098, "learning_rate": 3.3026356452879383e-07, "loss": 0.5596, "step": 21163 }, { "epoch": 0.8871655680496316, "grad_norm": 2.2548635005950928, "learning_rate": 3.3002097844182513e-07, "loss": 0.4498, "step": 21164 }, { "epoch": 0.8872074866646406, "grad_norm": 2.029869556427002, "learning_rate": 3.297784784394631e-07, "loss": 0.5115, "step": 21165 }, { "epoch": 0.8872494052796496, "grad_norm": 1.582155704498291, "learning_rate": 3.295360645261797e-07, "loss": 0.4102, "step": 21166 }, { "epoch": 0.8872913238946585, "grad_norm": 2.899716377258301, "learning_rate": 3.2929373670644084e-07, "loss": 0.4564, "step": 21167 }, { "epoch": 0.8873332425096675, "grad_norm": 1.9548026323318481, "learning_rate": 3.290514949847151e-07, "loss": 0.4468, "step": 21168 }, { "epoch": 0.8873751611246764, "grad_norm": 2.0246949195861816, "learning_rate": 3.2880933936546843e-07, "loss": 0.478, "step": 21169 }, { "epoch": 0.8874170797396854, "grad_norm": 1.9826704263687134, "learning_rate": 3.285672698531628e-07, "loss": 0.4644, "step": 21170 }, { "epoch": 0.8874589983546943, "grad_norm": 2.3467509746551514, "learning_rate": 3.283252864522618e-07, "loss": 0.4544, "step": 21171 }, { "epoch": 0.8875009169697033, "grad_norm": 1.9765428304672241, "learning_rate": 3.2808338916722635e-07, "loss": 0.4539, "step": 21172 }, { "epoch": 0.8875428355847123, "grad_norm": 2.5757579803466797, "learning_rate": 3.278415780025146e-07, "loss": 0.5062, "step": 21173 }, { "epoch": 0.8875847541997213, "grad_norm": 3.210411787033081, "learning_rate": 3.275998529625846e-07, "loss": 0.5022, "step": 21174 }, { "epoch": 0.8876266728147302, "grad_norm": 2.3594298362731934, "learning_rate": 3.2735821405189217e-07, "loss": 0.5267, "step": 21175 }, { "epoch": 0.8876685914297392, "grad_norm": 1.943419337272644, "learning_rate": 3.271166612748899e-07, "loss": 0.5227, "step": 21176 }, { "epoch": 0.8877105100447481, "grad_norm": 1.7000994682312012, "learning_rate": 3.2687519463603266e-07, "loss": 0.4749, "step": 21177 }, { "epoch": 0.8877524286597571, "grad_norm": 2.661038637161255, "learning_rate": 3.2663381413977126e-07, "loss": 0.4681, "step": 21178 }, { "epoch": 0.887794347274766, "grad_norm": 2.7974860668182373, "learning_rate": 3.263925197905543e-07, "loss": 0.4236, "step": 21179 }, { "epoch": 0.887836265889775, "grad_norm": 1.6031488180160522, "learning_rate": 3.2615131159283e-07, "loss": 0.4687, "step": 21180 }, { "epoch": 0.887878184504784, "grad_norm": 1.6389483213424683, "learning_rate": 3.2591018955104526e-07, "loss": 0.4685, "step": 21181 }, { "epoch": 0.887920103119793, "grad_norm": 2.36492657661438, "learning_rate": 3.256691536696438e-07, "loss": 0.4517, "step": 21182 }, { "epoch": 0.8879620217348019, "grad_norm": 1.7144101858139038, "learning_rate": 3.254282039530704e-07, "loss": 0.4687, "step": 21183 }, { "epoch": 0.8880039403498109, "grad_norm": 2.0220680236816406, "learning_rate": 3.2518734040576595e-07, "loss": 0.4768, "step": 21184 }, { "epoch": 0.8880458589648198, "grad_norm": 2.009237051010132, "learning_rate": 3.249465630321691e-07, "loss": 0.4736, "step": 21185 }, { "epoch": 0.8880877775798287, "grad_norm": 1.592868447303772, "learning_rate": 3.247058718367191e-07, "loss": 0.3882, "step": 21186 }, { "epoch": 0.8881296961948377, "grad_norm": 1.6628344058990479, "learning_rate": 3.24465266823854e-07, "loss": 0.489, "step": 21187 }, { "epoch": 0.8881716148098466, "grad_norm": 1.6713788509368896, "learning_rate": 3.242247479980071e-07, "loss": 0.47, "step": 21188 }, { "epoch": 0.8882135334248557, "grad_norm": 3.140488624572754, "learning_rate": 3.239843153636124e-07, "loss": 0.5234, "step": 21189 }, { "epoch": 0.8882554520398646, "grad_norm": 2.259244203567505, "learning_rate": 3.237439689251037e-07, "loss": 0.4488, "step": 21190 }, { "epoch": 0.8882973706548736, "grad_norm": 1.8824466466903687, "learning_rate": 3.2350370868690974e-07, "loss": 0.4434, "step": 21191 }, { "epoch": 0.8883392892698825, "grad_norm": 2.5569159984588623, "learning_rate": 3.2326353465346014e-07, "loss": 0.453, "step": 21192 }, { "epoch": 0.8883812078848915, "grad_norm": 2.089076519012451, "learning_rate": 3.230234468291821e-07, "loss": 0.468, "step": 21193 }, { "epoch": 0.8884231264999004, "grad_norm": 1.9576842784881592, "learning_rate": 3.2278344521850023e-07, "loss": 0.4128, "step": 21194 }, { "epoch": 0.8884650451149094, "grad_norm": 1.7835568189620972, "learning_rate": 3.2254352982583946e-07, "loss": 0.4155, "step": 21195 }, { "epoch": 0.8885069637299183, "grad_norm": 3.7098963260650635, "learning_rate": 3.2230370065562346e-07, "loss": 0.4806, "step": 21196 }, { "epoch": 0.8885488823449273, "grad_norm": 1.7502570152282715, "learning_rate": 3.220639577122708e-07, "loss": 0.449, "step": 21197 }, { "epoch": 0.8885908009599363, "grad_norm": 1.8740650415420532, "learning_rate": 3.218243010002031e-07, "loss": 0.4497, "step": 21198 }, { "epoch": 0.8886327195749453, "grad_norm": 1.8821278810501099, "learning_rate": 3.215847305238362e-07, "loss": 0.4989, "step": 21199 }, { "epoch": 0.8886746381899542, "grad_norm": 2.324197292327881, "learning_rate": 3.2134524628758655e-07, "loss": 0.4662, "step": 21200 }, { "epoch": 0.8887165568049632, "grad_norm": 3.113186836242676, "learning_rate": 3.2110584829587064e-07, "loss": 0.4888, "step": 21201 }, { "epoch": 0.8887584754199721, "grad_norm": 2.0721354484558105, "learning_rate": 3.2086653655309994e-07, "loss": 0.4959, "step": 21202 }, { "epoch": 0.8888003940349811, "grad_norm": 2.0415120124816895, "learning_rate": 3.2062731106368483e-07, "loss": 0.5073, "step": 21203 }, { "epoch": 0.88884231264999, "grad_norm": 1.9461771249771118, "learning_rate": 3.203881718320373e-07, "loss": 0.4808, "step": 21204 }, { "epoch": 0.888884231264999, "grad_norm": 1.8501754999160767, "learning_rate": 3.201491188625638e-07, "loss": 0.4737, "step": 21205 }, { "epoch": 0.888926149880008, "grad_norm": 1.9288725852966309, "learning_rate": 3.1991015215967093e-07, "loss": 0.5229, "step": 21206 }, { "epoch": 0.888968068495017, "grad_norm": 1.8823601007461548, "learning_rate": 3.196712717277661e-07, "loss": 0.4442, "step": 21207 }, { "epoch": 0.8890099871100259, "grad_norm": 1.6692529916763306, "learning_rate": 3.1943247757124927e-07, "loss": 0.4523, "step": 21208 }, { "epoch": 0.8890519057250349, "grad_norm": 2.5039334297180176, "learning_rate": 3.1919376969452467e-07, "loss": 0.4951, "step": 21209 }, { "epoch": 0.8890938243400438, "grad_norm": 2.077526092529297, "learning_rate": 3.189551481019926e-07, "loss": 0.4396, "step": 21210 }, { "epoch": 0.8891357429550527, "grad_norm": 1.748194694519043, "learning_rate": 3.1871661279805065e-07, "loss": 0.4553, "step": 21211 }, { "epoch": 0.8891776615700617, "grad_norm": 1.821988821029663, "learning_rate": 3.1847816378709593e-07, "loss": 0.4879, "step": 21212 }, { "epoch": 0.8892195801850706, "grad_norm": 1.9335734844207764, "learning_rate": 3.1823980107352494e-07, "loss": 0.4934, "step": 21213 }, { "epoch": 0.8892614988000797, "grad_norm": 2.6253068447113037, "learning_rate": 3.180015246617302e-07, "loss": 0.4233, "step": 21214 }, { "epoch": 0.8893034174150886, "grad_norm": 1.969411015510559, "learning_rate": 3.177633345561043e-07, "loss": 0.4907, "step": 21215 }, { "epoch": 0.8893453360300976, "grad_norm": 2.88158917427063, "learning_rate": 3.175252307610394e-07, "loss": 0.5126, "step": 21216 }, { "epoch": 0.8893872546451065, "grad_norm": 1.953775405883789, "learning_rate": 3.172872132809224e-07, "loss": 0.4866, "step": 21217 }, { "epoch": 0.8894291732601155, "grad_norm": 1.7952638864517212, "learning_rate": 3.170492821201426e-07, "loss": 0.462, "step": 21218 }, { "epoch": 0.8894710918751244, "grad_norm": 4.602819919586182, "learning_rate": 3.168114372830855e-07, "loss": 0.4802, "step": 21219 }, { "epoch": 0.8895130104901334, "grad_norm": 1.727277398109436, "learning_rate": 3.165736787741347e-07, "loss": 0.4891, "step": 21220 }, { "epoch": 0.8895549291051423, "grad_norm": 2.175774335861206, "learning_rate": 3.163360065976745e-07, "loss": 0.4348, "step": 21221 }, { "epoch": 0.8895968477201514, "grad_norm": 1.7723368406295776, "learning_rate": 3.160984207580847e-07, "loss": 0.4707, "step": 21222 }, { "epoch": 0.8896387663351603, "grad_norm": 2.021606206893921, "learning_rate": 3.1586092125974464e-07, "loss": 0.5, "step": 21223 }, { "epoch": 0.8896806849501693, "grad_norm": 3.3216135501861572, "learning_rate": 3.1562350810703244e-07, "loss": 0.4888, "step": 21224 }, { "epoch": 0.8897226035651782, "grad_norm": 1.970598816871643, "learning_rate": 3.153861813043263e-07, "loss": 0.5339, "step": 21225 }, { "epoch": 0.8897645221801872, "grad_norm": 1.7584822177886963, "learning_rate": 3.1514894085599824e-07, "loss": 0.4776, "step": 21226 }, { "epoch": 0.8898064407951961, "grad_norm": 7.404303073883057, "learning_rate": 3.1491178676642365e-07, "loss": 0.5104, "step": 21227 }, { "epoch": 0.8898483594102051, "grad_norm": 1.8322677612304688, "learning_rate": 3.1467471903997346e-07, "loss": 0.4725, "step": 21228 }, { "epoch": 0.889890278025214, "grad_norm": 1.8541368246078491, "learning_rate": 3.1443773768101703e-07, "loss": 0.4557, "step": 21229 }, { "epoch": 0.889932196640223, "grad_norm": 1.9167710542678833, "learning_rate": 3.1420084269392357e-07, "loss": 0.4985, "step": 21230 }, { "epoch": 0.889974115255232, "grad_norm": 1.8376905918121338, "learning_rate": 3.139640340830602e-07, "loss": 0.4124, "step": 21231 }, { "epoch": 0.890016033870241, "grad_norm": 1.7969954013824463, "learning_rate": 3.1372731185279116e-07, "loss": 0.4334, "step": 21232 }, { "epoch": 0.8900579524852499, "grad_norm": 1.8178389072418213, "learning_rate": 3.134906760074796e-07, "loss": 0.4835, "step": 21233 }, { "epoch": 0.8900998711002589, "grad_norm": 1.7538233995437622, "learning_rate": 3.132541265514899e-07, "loss": 0.484, "step": 21234 }, { "epoch": 0.8901417897152678, "grad_norm": 2.1128852367401123, "learning_rate": 3.130176634891802e-07, "loss": 0.4293, "step": 21235 }, { "epoch": 0.8901837083302767, "grad_norm": 2.2351417541503906, "learning_rate": 3.127812868249108e-07, "loss": 0.4466, "step": 21236 }, { "epoch": 0.8902256269452857, "grad_norm": 2.637434244155884, "learning_rate": 3.1254499656303893e-07, "loss": 0.4762, "step": 21237 }, { "epoch": 0.8902675455602946, "grad_norm": 2.062980890274048, "learning_rate": 3.1230879270791883e-07, "loss": 0.5101, "step": 21238 }, { "epoch": 0.8903094641753037, "grad_norm": 2.1896371841430664, "learning_rate": 3.120726752639064e-07, "loss": 0.5048, "step": 21239 }, { "epoch": 0.8903513827903126, "grad_norm": 2.1340842247009277, "learning_rate": 3.1183664423535374e-07, "loss": 0.5179, "step": 21240 }, { "epoch": 0.8903933014053216, "grad_norm": 2.080953359603882, "learning_rate": 3.116006996266102e-07, "loss": 0.4369, "step": 21241 }, { "epoch": 0.8904352200203305, "grad_norm": 1.8647301197052002, "learning_rate": 3.1136484144202614e-07, "loss": 0.4671, "step": 21242 }, { "epoch": 0.8904771386353395, "grad_norm": 2.582719087600708, "learning_rate": 3.111290696859504e-07, "loss": 0.4448, "step": 21243 }, { "epoch": 0.8905190572503484, "grad_norm": 2.218438148498535, "learning_rate": 3.108933843627271e-07, "loss": 0.4728, "step": 21244 }, { "epoch": 0.8905609758653574, "grad_norm": 2.537766456604004, "learning_rate": 3.106577854767023e-07, "loss": 0.5073, "step": 21245 }, { "epoch": 0.8906028944803663, "grad_norm": 5.049331188201904, "learning_rate": 3.1042227303221873e-07, "loss": 0.4358, "step": 21246 }, { "epoch": 0.8906448130953754, "grad_norm": 1.9627246856689453, "learning_rate": 3.1018684703361667e-07, "loss": 0.5128, "step": 21247 }, { "epoch": 0.8906867317103843, "grad_norm": 1.6953063011169434, "learning_rate": 3.099515074852377e-07, "loss": 0.4829, "step": 21248 }, { "epoch": 0.8907286503253933, "grad_norm": 2.171922206878662, "learning_rate": 3.097162543914184e-07, "loss": 0.4235, "step": 21249 }, { "epoch": 0.8907705689404022, "grad_norm": 2.9580962657928467, "learning_rate": 3.094810877564952e-07, "loss": 0.4537, "step": 21250 }, { "epoch": 0.8908124875554112, "grad_norm": 1.8762381076812744, "learning_rate": 3.0924600758480416e-07, "loss": 0.4786, "step": 21251 }, { "epoch": 0.8908544061704201, "grad_norm": 1.9523422718048096, "learning_rate": 3.0901101388067843e-07, "loss": 0.4783, "step": 21252 }, { "epoch": 0.8908963247854291, "grad_norm": 2.196751356124878, "learning_rate": 3.0877610664844904e-07, "loss": 0.4334, "step": 21253 }, { "epoch": 0.890938243400438, "grad_norm": 1.498874545097351, "learning_rate": 3.0854128589244634e-07, "loss": 0.4484, "step": 21254 }, { "epoch": 0.890980162015447, "grad_norm": 1.9458264112472534, "learning_rate": 3.083065516170003e-07, "loss": 0.4515, "step": 21255 }, { "epoch": 0.891022080630456, "grad_norm": 2.173591136932373, "learning_rate": 3.080719038264363e-07, "loss": 0.4233, "step": 21256 }, { "epoch": 0.891063999245465, "grad_norm": 1.5888456106185913, "learning_rate": 3.0783734252508136e-07, "loss": 0.3923, "step": 21257 }, { "epoch": 0.8911059178604739, "grad_norm": 1.953719973564148, "learning_rate": 3.0760286771725765e-07, "loss": 0.3864, "step": 21258 }, { "epoch": 0.8911478364754829, "grad_norm": 1.6776726245880127, "learning_rate": 3.073684794072873e-07, "loss": 0.4602, "step": 21259 }, { "epoch": 0.8911897550904918, "grad_norm": 1.8172929286956787, "learning_rate": 3.0713417759949237e-07, "loss": 0.4927, "step": 21260 }, { "epoch": 0.8912316737055007, "grad_norm": 1.9635624885559082, "learning_rate": 3.0689996229819106e-07, "loss": 0.5511, "step": 21261 }, { "epoch": 0.8912735923205097, "grad_norm": 2.1493828296661377, "learning_rate": 3.0666583350770107e-07, "loss": 0.4586, "step": 21262 }, { "epoch": 0.8913155109355186, "grad_norm": 4.400426387786865, "learning_rate": 3.0643179123233723e-07, "loss": 0.4873, "step": 21263 }, { "epoch": 0.8913574295505277, "grad_norm": 1.8284991979599, "learning_rate": 3.0619783547641613e-07, "loss": 0.4626, "step": 21264 }, { "epoch": 0.8913993481655366, "grad_norm": 1.9561511278152466, "learning_rate": 3.059639662442476e-07, "loss": 0.4366, "step": 21265 }, { "epoch": 0.8914412667805456, "grad_norm": 1.8611959218978882, "learning_rate": 3.057301835401449e-07, "loss": 0.5175, "step": 21266 }, { "epoch": 0.8914831853955545, "grad_norm": 2.4013659954071045, "learning_rate": 3.054964873684163e-07, "loss": 0.4292, "step": 21267 }, { "epoch": 0.8915251040105635, "grad_norm": 2.177530288696289, "learning_rate": 3.0526287773337047e-07, "loss": 0.4688, "step": 21268 }, { "epoch": 0.8915670226255724, "grad_norm": 1.8407907485961914, "learning_rate": 3.0502935463931237e-07, "loss": 0.444, "step": 21269 }, { "epoch": 0.8916089412405814, "grad_norm": 1.9803047180175781, "learning_rate": 3.0479591809054856e-07, "loss": 0.4833, "step": 21270 }, { "epoch": 0.8916508598555903, "grad_norm": 1.805623173713684, "learning_rate": 3.0456256809137995e-07, "loss": 0.4648, "step": 21271 }, { "epoch": 0.8916927784705994, "grad_norm": 1.8515671491622925, "learning_rate": 3.043293046461104e-07, "loss": 0.5159, "step": 21272 }, { "epoch": 0.8917346970856083, "grad_norm": 2.0698866844177246, "learning_rate": 3.040961277590376e-07, "loss": 0.4941, "step": 21273 }, { "epoch": 0.8917766157006173, "grad_norm": 2.0959513187408447, "learning_rate": 3.038630374344603e-07, "loss": 0.4704, "step": 21274 }, { "epoch": 0.8918185343156262, "grad_norm": 1.8181850910186768, "learning_rate": 3.036300336766773e-07, "loss": 0.4794, "step": 21275 }, { "epoch": 0.8918604529306352, "grad_norm": 1.9558926820755005, "learning_rate": 3.033971164899807e-07, "loss": 0.4848, "step": 21276 }, { "epoch": 0.8919023715456441, "grad_norm": 2.0127766132354736, "learning_rate": 3.031642858786665e-07, "loss": 0.5276, "step": 21277 }, { "epoch": 0.8919442901606531, "grad_norm": 1.6712067127227783, "learning_rate": 3.029315418470252e-07, "loss": 0.4599, "step": 21278 }, { "epoch": 0.891986208775662, "grad_norm": 2.0587985515594482, "learning_rate": 3.026988843993467e-07, "loss": 0.4569, "step": 21279 }, { "epoch": 0.892028127390671, "grad_norm": 1.7695006132125854, "learning_rate": 3.024663135399203e-07, "loss": 0.4871, "step": 21280 }, { "epoch": 0.89207004600568, "grad_norm": 2.4817898273468018, "learning_rate": 3.022338292730348e-07, "loss": 0.4845, "step": 21281 }, { "epoch": 0.892111964620689, "grad_norm": 1.8571951389312744, "learning_rate": 3.0200143160297235e-07, "loss": 0.5302, "step": 21282 }, { "epoch": 0.8921538832356979, "grad_norm": 2.8431448936462402, "learning_rate": 3.017691205340195e-07, "loss": 0.4972, "step": 21283 }, { "epoch": 0.8921958018507069, "grad_norm": 2.064563035964966, "learning_rate": 3.015368960704584e-07, "loss": 0.45, "step": 21284 }, { "epoch": 0.8922377204657158, "grad_norm": 1.9583697319030762, "learning_rate": 3.0130475821656846e-07, "loss": 0.5024, "step": 21285 }, { "epoch": 0.8922796390807247, "grad_norm": 1.9426155090332031, "learning_rate": 3.0107270697663005e-07, "loss": 0.4725, "step": 21286 }, { "epoch": 0.8923215576957337, "grad_norm": 1.8281067609786987, "learning_rate": 3.0084074235492034e-07, "loss": 0.4942, "step": 21287 }, { "epoch": 0.8923634763107426, "grad_norm": 2.1628551483154297, "learning_rate": 3.0060886435571477e-07, "loss": 0.4702, "step": 21288 }, { "epoch": 0.8924053949257517, "grad_norm": 2.20304012298584, "learning_rate": 3.003770729832872e-07, "loss": 0.4998, "step": 21289 }, { "epoch": 0.8924473135407606, "grad_norm": 1.642530918121338, "learning_rate": 3.001453682419131e-07, "loss": 0.4455, "step": 21290 }, { "epoch": 0.8924892321557696, "grad_norm": 1.745131015777588, "learning_rate": 2.999137501358601e-07, "loss": 0.4771, "step": 21291 }, { "epoch": 0.8925311507707785, "grad_norm": 1.7385988235473633, "learning_rate": 2.9968221866939986e-07, "loss": 0.4672, "step": 21292 }, { "epoch": 0.8925730693857875, "grad_norm": 1.7865618467330933, "learning_rate": 2.9945077384680054e-07, "loss": 0.4664, "step": 21293 }, { "epoch": 0.8926149880007964, "grad_norm": 1.7912054061889648, "learning_rate": 2.9921941567232716e-07, "loss": 0.4564, "step": 21294 }, { "epoch": 0.8926569066158054, "grad_norm": 2.2867531776428223, "learning_rate": 2.989881441502462e-07, "loss": 0.4493, "step": 21295 }, { "epoch": 0.8926988252308143, "grad_norm": 2.2504043579101562, "learning_rate": 2.9875695928481995e-07, "loss": 0.4318, "step": 21296 }, { "epoch": 0.8927407438458234, "grad_norm": 2.8365297317504883, "learning_rate": 2.9852586108030876e-07, "loss": 0.4227, "step": 21297 }, { "epoch": 0.8927826624608323, "grad_norm": 1.9315659999847412, "learning_rate": 2.9829484954097376e-07, "loss": 0.4443, "step": 21298 }, { "epoch": 0.8928245810758413, "grad_norm": 1.9984650611877441, "learning_rate": 2.9806392467107424e-07, "loss": 0.483, "step": 21299 }, { "epoch": 0.8928664996908502, "grad_norm": 1.8690563440322876, "learning_rate": 2.978330864748652e-07, "loss": 0.4786, "step": 21300 }, { "epoch": 0.8929084183058592, "grad_norm": 1.7914055585861206, "learning_rate": 2.976023349566026e-07, "loss": 0.4388, "step": 21301 }, { "epoch": 0.8929503369208681, "grad_norm": 1.582801103591919, "learning_rate": 2.9737167012054093e-07, "loss": 0.4833, "step": 21302 }, { "epoch": 0.8929922555358771, "grad_norm": 2.548039197921753, "learning_rate": 2.9714109197093053e-07, "loss": 0.5061, "step": 21303 }, { "epoch": 0.893034174150886, "grad_norm": 2.052264451980591, "learning_rate": 2.9691060051202314e-07, "loss": 0.4604, "step": 21304 }, { "epoch": 0.8930760927658951, "grad_norm": 2.0141348838806152, "learning_rate": 2.9668019574806694e-07, "loss": 0.4362, "step": 21305 }, { "epoch": 0.893118011380904, "grad_norm": 2.0219247341156006, "learning_rate": 2.964498776833091e-07, "loss": 0.5333, "step": 21306 }, { "epoch": 0.893159929995913, "grad_norm": 2.089650869369507, "learning_rate": 2.962196463219946e-07, "loss": 0.4721, "step": 21307 }, { "epoch": 0.8932018486109219, "grad_norm": 5.196156024932861, "learning_rate": 2.959895016683695e-07, "loss": 0.4713, "step": 21308 }, { "epoch": 0.8932437672259309, "grad_norm": 2.0483407974243164, "learning_rate": 2.9575944372667363e-07, "loss": 0.5028, "step": 21309 }, { "epoch": 0.8932856858409398, "grad_norm": 1.6754857301712036, "learning_rate": 2.9552947250114926e-07, "loss": 0.4458, "step": 21310 }, { "epoch": 0.8933276044559487, "grad_norm": 1.9173561334609985, "learning_rate": 2.9529958799603577e-07, "loss": 0.4174, "step": 21311 }, { "epoch": 0.8933695230709577, "grad_norm": 2.1701011657714844, "learning_rate": 2.950697902155697e-07, "loss": 0.4534, "step": 21312 }, { "epoch": 0.8934114416859666, "grad_norm": 3.398216962814331, "learning_rate": 2.9484007916398884e-07, "loss": 0.4577, "step": 21313 }, { "epoch": 0.8934533603009757, "grad_norm": 1.7837634086608887, "learning_rate": 2.946104548455253e-07, "loss": 0.4546, "step": 21314 }, { "epoch": 0.8934952789159846, "grad_norm": 2.2471394538879395, "learning_rate": 2.943809172644135e-07, "loss": 0.5249, "step": 21315 }, { "epoch": 0.8935371975309936, "grad_norm": 1.9027800559997559, "learning_rate": 2.94151466424884e-07, "loss": 0.4647, "step": 21316 }, { "epoch": 0.8935791161460025, "grad_norm": 2.0402889251708984, "learning_rate": 2.9392210233116667e-07, "loss": 0.5404, "step": 21317 }, { "epoch": 0.8936210347610115, "grad_norm": 1.7807462215423584, "learning_rate": 2.936928249874893e-07, "loss": 0.5194, "step": 21318 }, { "epoch": 0.8936629533760204, "grad_norm": 1.8272619247436523, "learning_rate": 2.934636343980779e-07, "loss": 0.4636, "step": 21319 }, { "epoch": 0.8937048719910294, "grad_norm": 6.819090843200684, "learning_rate": 2.9323453056715913e-07, "loss": 0.5128, "step": 21320 }, { "epoch": 0.8937467906060383, "grad_norm": 2.0051114559173584, "learning_rate": 2.93005513498954e-07, "loss": 0.4522, "step": 21321 }, { "epoch": 0.8937887092210474, "grad_norm": 1.9669792652130127, "learning_rate": 2.927765831976848e-07, "loss": 0.4412, "step": 21322 }, { "epoch": 0.8938306278360563, "grad_norm": 1.8288520574569702, "learning_rate": 2.9254773966757246e-07, "loss": 0.508, "step": 21323 }, { "epoch": 0.8938725464510653, "grad_norm": 2.4231226444244385, "learning_rate": 2.923189829128348e-07, "loss": 0.4709, "step": 21324 }, { "epoch": 0.8939144650660742, "grad_norm": 1.8408328294754028, "learning_rate": 2.920903129376873e-07, "loss": 0.4653, "step": 21325 }, { "epoch": 0.8939563836810832, "grad_norm": 2.2005105018615723, "learning_rate": 2.918617297463477e-07, "loss": 0.5303, "step": 21326 }, { "epoch": 0.8939983022960921, "grad_norm": 1.8827601671218872, "learning_rate": 2.9163323334302775e-07, "loss": 0.4411, "step": 21327 }, { "epoch": 0.8940402209111011, "grad_norm": 1.8248379230499268, "learning_rate": 2.9140482373193943e-07, "loss": 0.4478, "step": 21328 }, { "epoch": 0.89408213952611, "grad_norm": 1.6286730766296387, "learning_rate": 2.9117650091729454e-07, "loss": 0.433, "step": 21329 }, { "epoch": 0.8941240581411191, "grad_norm": 1.8555811643600464, "learning_rate": 2.9094826490330076e-07, "loss": 0.475, "step": 21330 }, { "epoch": 0.894165976756128, "grad_norm": 2.7239654064178467, "learning_rate": 2.907201156941658e-07, "loss": 0.519, "step": 21331 }, { "epoch": 0.894207895371137, "grad_norm": 1.8705503940582275, "learning_rate": 2.904920532940952e-07, "loss": 0.4666, "step": 21332 }, { "epoch": 0.8942498139861459, "grad_norm": 4.879483699798584, "learning_rate": 2.902640777072935e-07, "loss": 0.4704, "step": 21333 }, { "epoch": 0.8942917326011549, "grad_norm": 2.470128297805786, "learning_rate": 2.9003618893796105e-07, "loss": 0.4975, "step": 21334 }, { "epoch": 0.8943336512161638, "grad_norm": 3.4765784740448, "learning_rate": 2.898083869903018e-07, "loss": 0.4928, "step": 21335 }, { "epoch": 0.8943755698311727, "grad_norm": 2.115279197692871, "learning_rate": 2.8958067186851177e-07, "loss": 0.4862, "step": 21336 }, { "epoch": 0.8944174884461817, "grad_norm": 2.191467523574829, "learning_rate": 2.893530435767905e-07, "loss": 0.4931, "step": 21337 }, { "epoch": 0.8944594070611906, "grad_norm": 2.2959237098693848, "learning_rate": 2.89125502119334e-07, "loss": 0.4265, "step": 21338 }, { "epoch": 0.8945013256761997, "grad_norm": 1.7184863090515137, "learning_rate": 2.888980475003361e-07, "loss": 0.4672, "step": 21339 }, { "epoch": 0.8945432442912086, "grad_norm": 1.7892907857894897, "learning_rate": 2.886706797239891e-07, "loss": 0.4482, "step": 21340 }, { "epoch": 0.8945851629062176, "grad_norm": 1.9704468250274658, "learning_rate": 2.8844339879448625e-07, "loss": 0.4676, "step": 21341 }, { "epoch": 0.8946270815212265, "grad_norm": 1.8895373344421387, "learning_rate": 2.882162047160153e-07, "loss": 0.4862, "step": 21342 }, { "epoch": 0.8946690001362355, "grad_norm": 1.7108540534973145, "learning_rate": 2.8798909749276463e-07, "loss": 0.4865, "step": 21343 }, { "epoch": 0.8947109187512444, "grad_norm": 1.9426376819610596, "learning_rate": 2.877620771289213e-07, "loss": 0.4962, "step": 21344 }, { "epoch": 0.8947528373662534, "grad_norm": 3.1873135566711426, "learning_rate": 2.875351436286683e-07, "loss": 0.4765, "step": 21345 }, { "epoch": 0.8947947559812623, "grad_norm": 3.6589486598968506, "learning_rate": 2.87308296996191e-07, "loss": 0.427, "step": 21346 }, { "epoch": 0.8948366745962714, "grad_norm": 1.8110713958740234, "learning_rate": 2.870815372356706e-07, "loss": 0.4859, "step": 21347 }, { "epoch": 0.8948785932112803, "grad_norm": 2.0005385875701904, "learning_rate": 2.868548643512853e-07, "loss": 0.5378, "step": 21348 }, { "epoch": 0.8949205118262893, "grad_norm": 1.5035607814788818, "learning_rate": 2.866282783472163e-07, "loss": 0.4506, "step": 21349 }, { "epoch": 0.8949624304412982, "grad_norm": 2.8386871814727783, "learning_rate": 2.86401779227638e-07, "loss": 0.4426, "step": 21350 }, { "epoch": 0.8950043490563072, "grad_norm": 1.7968919277191162, "learning_rate": 2.861753669967271e-07, "loss": 0.4915, "step": 21351 }, { "epoch": 0.8950462676713161, "grad_norm": 2.034792184829712, "learning_rate": 2.859490416586563e-07, "loss": 0.5263, "step": 21352 }, { "epoch": 0.8950881862863251, "grad_norm": 1.7392988204956055, "learning_rate": 2.8572280321759847e-07, "loss": 0.453, "step": 21353 }, { "epoch": 0.895130104901334, "grad_norm": 1.9354135990142822, "learning_rate": 2.8549665167772243e-07, "loss": 0.4656, "step": 21354 }, { "epoch": 0.8951720235163431, "grad_norm": 1.8088171482086182, "learning_rate": 2.8527058704319934e-07, "loss": 0.4139, "step": 21355 }, { "epoch": 0.895213942131352, "grad_norm": 1.988835334777832, "learning_rate": 2.8504460931819413e-07, "loss": 0.4924, "step": 21356 }, { "epoch": 0.895255860746361, "grad_norm": 2.2271013259887695, "learning_rate": 2.848187185068729e-07, "loss": 0.511, "step": 21357 }, { "epoch": 0.8952977793613699, "grad_norm": 1.8898500204086304, "learning_rate": 2.845929146134013e-07, "loss": 0.4395, "step": 21358 }, { "epoch": 0.8953396979763789, "grad_norm": 1.8756828308105469, "learning_rate": 2.843671976419393e-07, "loss": 0.4237, "step": 21359 }, { "epoch": 0.8953816165913878, "grad_norm": 1.8413448333740234, "learning_rate": 2.8414156759664966e-07, "loss": 0.465, "step": 21360 }, { "epoch": 0.8954235352063967, "grad_norm": 2.078073024749756, "learning_rate": 2.839160244816902e-07, "loss": 0.4902, "step": 21361 }, { "epoch": 0.8954654538214057, "grad_norm": 1.7821393013000488, "learning_rate": 2.8369056830121976e-07, "loss": 0.5081, "step": 21362 }, { "epoch": 0.8955073724364147, "grad_norm": 10.380471229553223, "learning_rate": 2.834651990593928e-07, "loss": 0.4948, "step": 21363 }, { "epoch": 0.8955492910514237, "grad_norm": 1.7866750955581665, "learning_rate": 2.8323991676036554e-07, "loss": 0.4483, "step": 21364 }, { "epoch": 0.8955912096664326, "grad_norm": 2.0207536220550537, "learning_rate": 2.830147214082884e-07, "loss": 0.4183, "step": 21365 }, { "epoch": 0.8956331282814416, "grad_norm": 1.779451608657837, "learning_rate": 2.827896130073143e-07, "loss": 0.5231, "step": 21366 }, { "epoch": 0.8956750468964505, "grad_norm": 1.8765509128570557, "learning_rate": 2.825645915615927e-07, "loss": 0.5136, "step": 21367 }, { "epoch": 0.8957169655114595, "grad_norm": 2.049609422683716, "learning_rate": 2.823396570752707e-07, "loss": 0.4761, "step": 21368 }, { "epoch": 0.8957588841264684, "grad_norm": 1.8713043928146362, "learning_rate": 2.8211480955249503e-07, "loss": 0.4424, "step": 21369 }, { "epoch": 0.8958008027414774, "grad_norm": 2.1790523529052734, "learning_rate": 2.8189004899741137e-07, "loss": 0.4173, "step": 21370 }, { "epoch": 0.8958427213564863, "grad_norm": 1.9414573907852173, "learning_rate": 2.8166537541416185e-07, "loss": 0.4712, "step": 21371 }, { "epoch": 0.8958846399714954, "grad_norm": 2.7298996448516846, "learning_rate": 2.814407888068876e-07, "loss": 0.5288, "step": 21372 }, { "epoch": 0.8959265585865043, "grad_norm": 1.6935489177703857, "learning_rate": 2.8121628917972987e-07, "loss": 0.4468, "step": 21373 }, { "epoch": 0.8959684772015133, "grad_norm": 2.1532487869262695, "learning_rate": 2.809918765368258e-07, "loss": 0.4251, "step": 21374 }, { "epoch": 0.8960103958165222, "grad_norm": 3.0705060958862305, "learning_rate": 2.8076755088231266e-07, "loss": 0.5171, "step": 21375 }, { "epoch": 0.8960523144315312, "grad_norm": 1.6548970937728882, "learning_rate": 2.805433122203266e-07, "loss": 0.481, "step": 21376 }, { "epoch": 0.8960942330465401, "grad_norm": 2.286364793777466, "learning_rate": 2.8031916055499873e-07, "loss": 0.4496, "step": 21377 }, { "epoch": 0.8961361516615491, "grad_norm": 2.1017708778381348, "learning_rate": 2.8009509589046247e-07, "loss": 0.4109, "step": 21378 }, { "epoch": 0.896178070276558, "grad_norm": 1.8660253286361694, "learning_rate": 2.7987111823084947e-07, "loss": 0.5024, "step": 21379 }, { "epoch": 0.8962199888915671, "grad_norm": 1.934253454208374, "learning_rate": 2.796472275802864e-07, "loss": 0.3983, "step": 21380 }, { "epoch": 0.896261907506576, "grad_norm": 2.0084776878356934, "learning_rate": 2.794234239429e-07, "loss": 0.4966, "step": 21381 }, { "epoch": 0.896303826121585, "grad_norm": 1.8559635877609253, "learning_rate": 2.791997073228181e-07, "loss": 0.4558, "step": 21382 }, { "epoch": 0.8963457447365939, "grad_norm": 2.1490509510040283, "learning_rate": 2.7897607772416233e-07, "loss": 0.4935, "step": 21383 }, { "epoch": 0.8963876633516029, "grad_norm": 1.646772027015686, "learning_rate": 2.7875253515105617e-07, "loss": 0.4594, "step": 21384 }, { "epoch": 0.8964295819666118, "grad_norm": 1.821739673614502, "learning_rate": 2.7852907960762066e-07, "loss": 0.4283, "step": 21385 }, { "epoch": 0.8964715005816207, "grad_norm": 1.8560137748718262, "learning_rate": 2.7830571109797367e-07, "loss": 0.4987, "step": 21386 }, { "epoch": 0.8965134191966297, "grad_norm": 2.046701192855835, "learning_rate": 2.78082429626233e-07, "loss": 0.5105, "step": 21387 }, { "epoch": 0.8965553378116387, "grad_norm": 1.9712612628936768, "learning_rate": 2.7785923519651594e-07, "loss": 0.5167, "step": 21388 }, { "epoch": 0.8965972564266477, "grad_norm": 1.9346959590911865, "learning_rate": 2.776361278129358e-07, "loss": 0.5299, "step": 21389 }, { "epoch": 0.8966391750416566, "grad_norm": 1.7300844192504883, "learning_rate": 2.774131074796038e-07, "loss": 0.4268, "step": 21390 }, { "epoch": 0.8966810936566656, "grad_norm": 2.0140507221221924, "learning_rate": 2.771901742006339e-07, "loss": 0.4366, "step": 21391 }, { "epoch": 0.8967230122716745, "grad_norm": 1.9131025075912476, "learning_rate": 2.769673279801327e-07, "loss": 0.4645, "step": 21392 }, { "epoch": 0.8967649308866835, "grad_norm": 1.9337860345840454, "learning_rate": 2.767445688222098e-07, "loss": 0.4722, "step": 21393 }, { "epoch": 0.8968068495016924, "grad_norm": 1.854324221611023, "learning_rate": 2.7652189673097184e-07, "loss": 0.4519, "step": 21394 }, { "epoch": 0.8968487681167014, "grad_norm": 1.8519375324249268, "learning_rate": 2.762993117105217e-07, "loss": 0.4658, "step": 21395 }, { "epoch": 0.8968906867317104, "grad_norm": 2.3460750579833984, "learning_rate": 2.7607681376496387e-07, "loss": 0.4855, "step": 21396 }, { "epoch": 0.8969326053467194, "grad_norm": 1.811585783958435, "learning_rate": 2.7585440289839947e-07, "loss": 0.5337, "step": 21397 }, { "epoch": 0.8969745239617283, "grad_norm": 1.7943074703216553, "learning_rate": 2.756320791149286e-07, "loss": 0.5097, "step": 21398 }, { "epoch": 0.8970164425767373, "grad_norm": 2.308873176574707, "learning_rate": 2.7540984241864853e-07, "loss": 0.48, "step": 21399 }, { "epoch": 0.8970583611917462, "grad_norm": 2.142836809158325, "learning_rate": 2.751876928136571e-07, "loss": 0.4891, "step": 21400 }, { "epoch": 0.8971002798067552, "grad_norm": 2.3660542964935303, "learning_rate": 2.749656303040477e-07, "loss": 0.5112, "step": 21401 }, { "epoch": 0.8971421984217641, "grad_norm": 1.872650384902954, "learning_rate": 2.7474365489391474e-07, "loss": 0.5088, "step": 21402 }, { "epoch": 0.8971841170367731, "grad_norm": 2.140373706817627, "learning_rate": 2.745217665873512e-07, "loss": 0.459, "step": 21403 }, { "epoch": 0.897226035651782, "grad_norm": 1.7259294986724854, "learning_rate": 2.7429996538844486e-07, "loss": 0.4511, "step": 21404 }, { "epoch": 0.8972679542667911, "grad_norm": 1.8302831649780273, "learning_rate": 2.7407825130128574e-07, "loss": 0.5081, "step": 21405 }, { "epoch": 0.8973098728818, "grad_norm": 2.122525215148926, "learning_rate": 2.7385662432996177e-07, "loss": 0.4788, "step": 21406 }, { "epoch": 0.897351791496809, "grad_norm": 2.9337475299835205, "learning_rate": 2.7363508447855627e-07, "loss": 0.5462, "step": 21407 }, { "epoch": 0.8973937101118179, "grad_norm": 2.5449886322021484, "learning_rate": 2.734136317511543e-07, "loss": 0.4768, "step": 21408 }, { "epoch": 0.8974356287268269, "grad_norm": 1.8982781171798706, "learning_rate": 2.7319226615183827e-07, "loss": 0.4738, "step": 21409 }, { "epoch": 0.8974775473418358, "grad_norm": 2.001323699951172, "learning_rate": 2.72970987684687e-07, "loss": 0.47, "step": 21410 }, { "epoch": 0.8975194659568447, "grad_norm": 4.8840861320495605, "learning_rate": 2.7274979635378116e-07, "loss": 0.4213, "step": 21411 }, { "epoch": 0.8975613845718537, "grad_norm": 3.5160398483276367, "learning_rate": 2.725286921631981e-07, "loss": 0.503, "step": 21412 }, { "epoch": 0.8976033031868627, "grad_norm": 1.810010552406311, "learning_rate": 2.7230767511701227e-07, "loss": 0.5388, "step": 21413 }, { "epoch": 0.8976452218018717, "grad_norm": 1.6163101196289062, "learning_rate": 2.720867452192988e-07, "loss": 0.4985, "step": 21414 }, { "epoch": 0.8976871404168806, "grad_norm": 1.640400767326355, "learning_rate": 2.7186590247413047e-07, "loss": 0.4622, "step": 21415 }, { "epoch": 0.8977290590318896, "grad_norm": 1.7491216659545898, "learning_rate": 2.7164514688557744e-07, "loss": 0.4899, "step": 21416 }, { "epoch": 0.8977709776468985, "grad_norm": 2.096306085586548, "learning_rate": 2.714244784577097e-07, "loss": 0.4512, "step": 21417 }, { "epoch": 0.8978128962619075, "grad_norm": 2.0464117527008057, "learning_rate": 2.712038971945946e-07, "loss": 0.4759, "step": 21418 }, { "epoch": 0.8978548148769164, "grad_norm": 2.6644654273986816, "learning_rate": 2.709834031002978e-07, "loss": 0.4521, "step": 21419 }, { "epoch": 0.8978967334919254, "grad_norm": 2.466003179550171, "learning_rate": 2.7076299617888434e-07, "loss": 0.4392, "step": 21420 }, { "epoch": 0.8979386521069344, "grad_norm": 1.875690221786499, "learning_rate": 2.705426764344177e-07, "loss": 0.5054, "step": 21421 }, { "epoch": 0.8979805707219434, "grad_norm": 2.5833680629730225, "learning_rate": 2.7032244387095784e-07, "loss": 0.5254, "step": 21422 }, { "epoch": 0.8980224893369523, "grad_norm": 1.8379571437835693, "learning_rate": 2.7010229849256606e-07, "loss": 0.4671, "step": 21423 }, { "epoch": 0.8980644079519613, "grad_norm": 1.8573052883148193, "learning_rate": 2.698822403032986e-07, "loss": 0.4311, "step": 21424 }, { "epoch": 0.8981063265669702, "grad_norm": 1.6215561628341675, "learning_rate": 2.696622693072126e-07, "loss": 0.4379, "step": 21425 }, { "epoch": 0.8981482451819792, "grad_norm": 2.0938632488250732, "learning_rate": 2.6944238550836443e-07, "loss": 0.504, "step": 21426 }, { "epoch": 0.8981901637969881, "grad_norm": 1.8825849294662476, "learning_rate": 2.692225889108058e-07, "loss": 0.5706, "step": 21427 }, { "epoch": 0.8982320824119971, "grad_norm": 1.9794063568115234, "learning_rate": 2.690028795185878e-07, "loss": 0.4798, "step": 21428 }, { "epoch": 0.898274001027006, "grad_norm": 2.158552885055542, "learning_rate": 2.687832573357624e-07, "loss": 0.5095, "step": 21429 }, { "epoch": 0.8983159196420151, "grad_norm": 1.8812251091003418, "learning_rate": 2.685637223663756e-07, "loss": 0.4773, "step": 21430 }, { "epoch": 0.898357838257024, "grad_norm": 1.9806019067764282, "learning_rate": 2.68344274614476e-07, "loss": 0.517, "step": 21431 }, { "epoch": 0.898399756872033, "grad_norm": 1.8967835903167725, "learning_rate": 2.6812491408410913e-07, "loss": 0.4947, "step": 21432 }, { "epoch": 0.8984416754870419, "grad_norm": 3.098639965057373, "learning_rate": 2.679056407793168e-07, "loss": 0.4462, "step": 21433 }, { "epoch": 0.8984835941020509, "grad_norm": 3.0025057792663574, "learning_rate": 2.6768645470414244e-07, "loss": 0.4944, "step": 21434 }, { "epoch": 0.8985255127170598, "grad_norm": 1.8462709188461304, "learning_rate": 2.6746735586262676e-07, "loss": 0.5013, "step": 21435 }, { "epoch": 0.8985674313320687, "grad_norm": 5.0174431800842285, "learning_rate": 2.672483442588075e-07, "loss": 0.5065, "step": 21436 }, { "epoch": 0.8986093499470778, "grad_norm": 1.801288366317749, "learning_rate": 2.670294198967216e-07, "loss": 0.52, "step": 21437 }, { "epoch": 0.8986512685620867, "grad_norm": 2.1612017154693604, "learning_rate": 2.668105827804063e-07, "loss": 0.4971, "step": 21438 }, { "epoch": 0.8986931871770957, "grad_norm": 2.0759713649749756, "learning_rate": 2.6659183291389337e-07, "loss": 0.4623, "step": 21439 }, { "epoch": 0.8987351057921046, "grad_norm": 2.734757900238037, "learning_rate": 2.663731703012162e-07, "loss": 0.4945, "step": 21440 }, { "epoch": 0.8987770244071136, "grad_norm": 1.9497026205062866, "learning_rate": 2.661545949464067e-07, "loss": 0.4946, "step": 21441 }, { "epoch": 0.8988189430221225, "grad_norm": 1.956663966178894, "learning_rate": 2.659361068534916e-07, "loss": 0.4506, "step": 21442 }, { "epoch": 0.8988608616371315, "grad_norm": 2.269803524017334, "learning_rate": 2.657177060265004e-07, "loss": 0.4634, "step": 21443 }, { "epoch": 0.8989027802521404, "grad_norm": 1.7879726886749268, "learning_rate": 2.6549939246945875e-07, "loss": 0.5009, "step": 21444 }, { "epoch": 0.8989446988671494, "grad_norm": 2.245424270629883, "learning_rate": 2.6528116618639023e-07, "loss": 0.4406, "step": 21445 }, { "epoch": 0.8989866174821584, "grad_norm": 2.286518096923828, "learning_rate": 2.650630271813176e-07, "loss": 0.4914, "step": 21446 }, { "epoch": 0.8990285360971674, "grad_norm": 1.881154179573059, "learning_rate": 2.648449754582627e-07, "loss": 0.4392, "step": 21447 }, { "epoch": 0.8990704547121763, "grad_norm": 1.8646717071533203, "learning_rate": 2.64627011021244e-07, "loss": 0.4727, "step": 21448 }, { "epoch": 0.8991123733271853, "grad_norm": 1.6779906749725342, "learning_rate": 2.6440913387427935e-07, "loss": 0.4229, "step": 21449 }, { "epoch": 0.8991542919421942, "grad_norm": 2.146165132522583, "learning_rate": 2.6419134402138667e-07, "loss": 0.5144, "step": 21450 }, { "epoch": 0.8991962105572032, "grad_norm": 2.0350992679595947, "learning_rate": 2.639736414665789e-07, "loss": 0.4744, "step": 21451 }, { "epoch": 0.8992381291722121, "grad_norm": 2.0044188499450684, "learning_rate": 2.637560262138694e-07, "loss": 0.5071, "step": 21452 }, { "epoch": 0.8992800477872211, "grad_norm": 3.7025575637817383, "learning_rate": 2.635384982672712e-07, "loss": 0.463, "step": 21453 }, { "epoch": 0.89932196640223, "grad_norm": 1.9252939224243164, "learning_rate": 2.633210576307915e-07, "loss": 0.4775, "step": 21454 }, { "epoch": 0.8993638850172391, "grad_norm": 2.346092462539673, "learning_rate": 2.6310370430844055e-07, "loss": 0.4665, "step": 21455 }, { "epoch": 0.899405803632248, "grad_norm": 3.2776384353637695, "learning_rate": 2.628864383042246e-07, "loss": 0.5502, "step": 21456 }, { "epoch": 0.899447722247257, "grad_norm": 1.9229722023010254, "learning_rate": 2.626692596221475e-07, "loss": 0.442, "step": 21457 }, { "epoch": 0.8994896408622659, "grad_norm": 2.3043739795684814, "learning_rate": 2.6245216826621346e-07, "loss": 0.4543, "step": 21458 }, { "epoch": 0.8995315594772749, "grad_norm": 2.0658152103424072, "learning_rate": 2.622351642404247e-07, "loss": 0.4773, "step": 21459 }, { "epoch": 0.8995734780922838, "grad_norm": 1.7804099321365356, "learning_rate": 2.6201824754878037e-07, "loss": 0.4798, "step": 21460 }, { "epoch": 0.8996153967072927, "grad_norm": 2.0765559673309326, "learning_rate": 2.6180141819527994e-07, "loss": 0.5118, "step": 21461 }, { "epoch": 0.8996573153223018, "grad_norm": 1.5918595790863037, "learning_rate": 2.6158467618392025e-07, "loss": 0.4366, "step": 21462 }, { "epoch": 0.8996992339373107, "grad_norm": 1.7526874542236328, "learning_rate": 2.613680215186959e-07, "loss": 0.5192, "step": 21463 }, { "epoch": 0.8997411525523197, "grad_norm": 2.3926734924316406, "learning_rate": 2.6115145420360143e-07, "loss": 0.4846, "step": 21464 }, { "epoch": 0.8997830711673286, "grad_norm": 1.9507384300231934, "learning_rate": 2.609349742426287e-07, "loss": 0.4807, "step": 21465 }, { "epoch": 0.8998249897823376, "grad_norm": 1.7433598041534424, "learning_rate": 2.607185816397678e-07, "loss": 0.5337, "step": 21466 }, { "epoch": 0.8998669083973465, "grad_norm": 2.1259727478027344, "learning_rate": 2.605022763990078e-07, "loss": 0.4622, "step": 21467 }, { "epoch": 0.8999088270123555, "grad_norm": 2.375704765319824, "learning_rate": 2.6028605852433717e-07, "loss": 0.4756, "step": 21468 }, { "epoch": 0.8999507456273644, "grad_norm": 2.0028698444366455, "learning_rate": 2.600699280197394e-07, "loss": 0.4838, "step": 21469 }, { "epoch": 0.8999926642423735, "grad_norm": 1.881184458732605, "learning_rate": 2.5985388488920015e-07, "loss": 0.4832, "step": 21470 }, { "epoch": 0.9000345828573824, "grad_norm": 3.9941353797912598, "learning_rate": 2.596379291367024e-07, "loss": 0.4566, "step": 21471 }, { "epoch": 0.9000765014723914, "grad_norm": 1.9673269987106323, "learning_rate": 2.5942206076622456e-07, "loss": 0.4486, "step": 21472 }, { "epoch": 0.9001184200874003, "grad_norm": 2.334933280944824, "learning_rate": 2.592062797817485e-07, "loss": 0.4429, "step": 21473 }, { "epoch": 0.9001603387024093, "grad_norm": 1.9681267738342285, "learning_rate": 2.5899058618725106e-07, "loss": 0.4801, "step": 21474 }, { "epoch": 0.9002022573174182, "grad_norm": 2.3456695079803467, "learning_rate": 2.5877497998670676e-07, "loss": 0.4179, "step": 21475 }, { "epoch": 0.9002441759324272, "grad_norm": 1.9125367403030396, "learning_rate": 2.5855946118409135e-07, "loss": 0.4341, "step": 21476 }, { "epoch": 0.9002860945474361, "grad_norm": 1.9524188041687012, "learning_rate": 2.583440297833778e-07, "loss": 0.4529, "step": 21477 }, { "epoch": 0.9003280131624452, "grad_norm": 1.9483784437179565, "learning_rate": 2.581286857885368e-07, "loss": 0.4856, "step": 21478 }, { "epoch": 0.9003699317774541, "grad_norm": 2.0928823947906494, "learning_rate": 2.579134292035379e-07, "loss": 0.4726, "step": 21479 }, { "epoch": 0.9004118503924631, "grad_norm": 1.8362834453582764, "learning_rate": 2.5769826003234964e-07, "loss": 0.4917, "step": 21480 }, { "epoch": 0.900453769007472, "grad_norm": 1.947160243988037, "learning_rate": 2.574831782789372e-07, "loss": 0.5129, "step": 21481 }, { "epoch": 0.900495687622481, "grad_norm": 1.6853519678115845, "learning_rate": 2.572681839472668e-07, "loss": 0.4634, "step": 21482 }, { "epoch": 0.9005376062374899, "grad_norm": 1.9338958263397217, "learning_rate": 2.570532770413009e-07, "loss": 0.4325, "step": 21483 }, { "epoch": 0.9005795248524989, "grad_norm": 2.118544101715088, "learning_rate": 2.568384575650001e-07, "loss": 0.4691, "step": 21484 }, { "epoch": 0.9006214434675078, "grad_norm": 2.0283732414245605, "learning_rate": 2.566237255223253e-07, "loss": 0.4994, "step": 21485 }, { "epoch": 0.9006633620825167, "grad_norm": 2.096952199935913, "learning_rate": 2.5640908091723483e-07, "loss": 0.486, "step": 21486 }, { "epoch": 0.9007052806975258, "grad_norm": 2.750791072845459, "learning_rate": 2.56194523753685e-07, "loss": 0.4941, "step": 21487 }, { "epoch": 0.9007471993125347, "grad_norm": 1.834673523902893, "learning_rate": 2.559800540356305e-07, "loss": 0.4416, "step": 21488 }, { "epoch": 0.9007891179275437, "grad_norm": 1.7888466119766235, "learning_rate": 2.557656717670265e-07, "loss": 0.4334, "step": 21489 }, { "epoch": 0.9008310365425526, "grad_norm": 1.892209529876709, "learning_rate": 2.555513769518225e-07, "loss": 0.4852, "step": 21490 }, { "epoch": 0.9008729551575616, "grad_norm": 1.865105152130127, "learning_rate": 2.55337169593971e-07, "loss": 0.5099, "step": 21491 }, { "epoch": 0.9009148737725705, "grad_norm": 2.024198532104492, "learning_rate": 2.551230496974194e-07, "loss": 0.5164, "step": 21492 }, { "epoch": 0.9009567923875795, "grad_norm": 1.714229702949524, "learning_rate": 2.5490901726611396e-07, "loss": 0.5301, "step": 21493 }, { "epoch": 0.9009987110025884, "grad_norm": 2.219006061553955, "learning_rate": 2.546950723040009e-07, "loss": 0.4705, "step": 21494 }, { "epoch": 0.9010406296175975, "grad_norm": 1.9877203702926636, "learning_rate": 2.544812148150244e-07, "loss": 0.4882, "step": 21495 }, { "epoch": 0.9010825482326064, "grad_norm": 1.6939483880996704, "learning_rate": 2.5426744480312573e-07, "loss": 0.4982, "step": 21496 }, { "epoch": 0.9011244668476154, "grad_norm": 2.063260078430176, "learning_rate": 2.540537622722467e-07, "loss": 0.4965, "step": 21497 }, { "epoch": 0.9011663854626243, "grad_norm": 2.1533267498016357, "learning_rate": 2.5384016722632463e-07, "loss": 0.4502, "step": 21498 }, { "epoch": 0.9012083040776333, "grad_norm": 1.8953901529312134, "learning_rate": 2.536266596692982e-07, "loss": 0.5121, "step": 21499 }, { "epoch": 0.9012502226926422, "grad_norm": 1.7009780406951904, "learning_rate": 2.534132396051031e-07, "loss": 0.454, "step": 21500 }, { "epoch": 0.9012921413076512, "grad_norm": 1.8466485738754272, "learning_rate": 2.5319990703767226e-07, "loss": 0.4731, "step": 21501 }, { "epoch": 0.9013340599226601, "grad_norm": 2.641434669494629, "learning_rate": 2.5298666197093926e-07, "loss": 0.4931, "step": 21502 }, { "epoch": 0.9013759785376692, "grad_norm": 2.241941452026367, "learning_rate": 2.5277350440883476e-07, "loss": 0.4987, "step": 21503 }, { "epoch": 0.9014178971526781, "grad_norm": 2.5129716396331787, "learning_rate": 2.525604343552873e-07, "loss": 0.4531, "step": 21504 }, { "epoch": 0.9014598157676871, "grad_norm": 1.831192135810852, "learning_rate": 2.5234745181422495e-07, "loss": 0.5026, "step": 21505 }, { "epoch": 0.901501734382696, "grad_norm": 2.504894971847534, "learning_rate": 2.5213455678957445e-07, "loss": 0.4257, "step": 21506 }, { "epoch": 0.901543652997705, "grad_norm": 1.8712583780288696, "learning_rate": 2.5192174928525935e-07, "loss": 0.4583, "step": 21507 }, { "epoch": 0.9015855716127139, "grad_norm": 1.795533299446106, "learning_rate": 2.517090293052027e-07, "loss": 0.4465, "step": 21508 }, { "epoch": 0.9016274902277229, "grad_norm": 2.1036200523376465, "learning_rate": 2.514963968533263e-07, "loss": 0.4756, "step": 21509 }, { "epoch": 0.9016694088427318, "grad_norm": 1.795445442199707, "learning_rate": 2.5128385193354863e-07, "loss": 0.4115, "step": 21510 }, { "epoch": 0.9017113274577407, "grad_norm": 1.8064712285995483, "learning_rate": 2.5107139454978893e-07, "loss": 0.4849, "step": 21511 }, { "epoch": 0.9017532460727498, "grad_norm": 1.969978928565979, "learning_rate": 2.508590247059628e-07, "loss": 0.4527, "step": 21512 }, { "epoch": 0.9017951646877587, "grad_norm": 1.8564590215682983, "learning_rate": 2.506467424059844e-07, "loss": 0.4419, "step": 21513 }, { "epoch": 0.9018370833027677, "grad_norm": 1.8982634544372559, "learning_rate": 2.5043454765376785e-07, "loss": 0.4787, "step": 21514 }, { "epoch": 0.9018790019177766, "grad_norm": 2.0965075492858887, "learning_rate": 2.502224404532244e-07, "loss": 0.4724, "step": 21515 }, { "epoch": 0.9019209205327856, "grad_norm": 1.7393885850906372, "learning_rate": 2.5001042080826376e-07, "loss": 0.4744, "step": 21516 }, { "epoch": 0.9019628391477945, "grad_norm": 2.525388479232788, "learning_rate": 2.497984887227939e-07, "loss": 0.4438, "step": 21517 }, { "epoch": 0.9020047577628035, "grad_norm": 1.8052573204040527, "learning_rate": 2.495866442007228e-07, "loss": 0.4072, "step": 21518 }, { "epoch": 0.9020466763778124, "grad_norm": 2.0966789722442627, "learning_rate": 2.493748872459539e-07, "loss": 0.4345, "step": 21519 }, { "epoch": 0.9020885949928215, "grad_norm": 2.019672393798828, "learning_rate": 2.4916321786239263e-07, "loss": 0.4594, "step": 21520 }, { "epoch": 0.9021305136078304, "grad_norm": 4.006419658660889, "learning_rate": 2.48951636053939e-07, "loss": 0.456, "step": 21521 }, { "epoch": 0.9021724322228394, "grad_norm": 2.465939998626709, "learning_rate": 2.4874014182449337e-07, "loss": 0.4511, "step": 21522 }, { "epoch": 0.9022143508378483, "grad_norm": 1.6715707778930664, "learning_rate": 2.485287351779547e-07, "loss": 0.4796, "step": 21523 }, { "epoch": 0.9022562694528573, "grad_norm": 2.2882301807403564, "learning_rate": 2.483174161182211e-07, "loss": 0.4564, "step": 21524 }, { "epoch": 0.9022981880678662, "grad_norm": 1.8072224855422974, "learning_rate": 2.481061846491856e-07, "loss": 0.4381, "step": 21525 }, { "epoch": 0.9023401066828752, "grad_norm": 1.6744580268859863, "learning_rate": 2.478950407747438e-07, "loss": 0.4368, "step": 21526 }, { "epoch": 0.9023820252978841, "grad_norm": 1.8196452856063843, "learning_rate": 2.4768398449878827e-07, "loss": 0.5013, "step": 21527 }, { "epoch": 0.9024239439128932, "grad_norm": 2.2590274810791016, "learning_rate": 2.4747301582520757e-07, "loss": 0.5071, "step": 21528 }, { "epoch": 0.9024658625279021, "grad_norm": 2.106574535369873, "learning_rate": 2.4726213475789185e-07, "loss": 0.4901, "step": 21529 }, { "epoch": 0.9025077811429111, "grad_norm": 3.0076940059661865, "learning_rate": 2.470513413007286e-07, "loss": 0.4918, "step": 21530 }, { "epoch": 0.90254969975792, "grad_norm": 2.5418028831481934, "learning_rate": 2.4684063545760307e-07, "loss": 0.5319, "step": 21531 }, { "epoch": 0.902591618372929, "grad_norm": 1.6814048290252686, "learning_rate": 2.466300172323988e-07, "loss": 0.4572, "step": 21532 }, { "epoch": 0.9026335369879379, "grad_norm": 1.9303196668624878, "learning_rate": 2.464194866289993e-07, "loss": 0.4902, "step": 21533 }, { "epoch": 0.9026754556029469, "grad_norm": 1.8668553829193115, "learning_rate": 2.4620904365128427e-07, "loss": 0.4913, "step": 21534 }, { "epoch": 0.9027173742179558, "grad_norm": 1.8002156019210815, "learning_rate": 2.45998688303134e-07, "loss": 0.4163, "step": 21535 }, { "epoch": 0.9027592928329647, "grad_norm": 1.9164525270462036, "learning_rate": 2.4578842058842645e-07, "loss": 0.459, "step": 21536 }, { "epoch": 0.9028012114479738, "grad_norm": 1.799050211906433, "learning_rate": 2.4557824051103576e-07, "loss": 0.4511, "step": 21537 }, { "epoch": 0.9028431300629827, "grad_norm": 1.8184847831726074, "learning_rate": 2.453681480748382e-07, "loss": 0.4456, "step": 21538 }, { "epoch": 0.9028850486779917, "grad_norm": 1.7402414083480835, "learning_rate": 2.451581432837058e-07, "loss": 0.5029, "step": 21539 }, { "epoch": 0.9029269672930006, "grad_norm": 1.8913545608520508, "learning_rate": 2.4494822614150925e-07, "loss": 0.4674, "step": 21540 }, { "epoch": 0.9029688859080096, "grad_norm": 1.9887274503707886, "learning_rate": 2.447383966521183e-07, "loss": 0.5111, "step": 21541 }, { "epoch": 0.9030108045230185, "grad_norm": 2.2810914516448975, "learning_rate": 2.445286548194015e-07, "loss": 0.4604, "step": 21542 }, { "epoch": 0.9030527231380275, "grad_norm": 1.809494137763977, "learning_rate": 2.4431900064722405e-07, "loss": 0.4275, "step": 21543 }, { "epoch": 0.9030946417530364, "grad_norm": 2.013806104660034, "learning_rate": 2.4410943413945174e-07, "loss": 0.4757, "step": 21544 }, { "epoch": 0.9031365603680455, "grad_norm": 3.212716579437256, "learning_rate": 2.4389995529994713e-07, "loss": 0.5147, "step": 21545 }, { "epoch": 0.9031784789830544, "grad_norm": 2.0019750595092773, "learning_rate": 2.436905641325715e-07, "loss": 0.4429, "step": 21546 }, { "epoch": 0.9032203975980634, "grad_norm": 2.506713390350342, "learning_rate": 2.434812606411857e-07, "loss": 0.4589, "step": 21547 }, { "epoch": 0.9032623162130723, "grad_norm": 2.031566858291626, "learning_rate": 2.4327204482964663e-07, "loss": 0.4663, "step": 21548 }, { "epoch": 0.9033042348280813, "grad_norm": 2.105851650238037, "learning_rate": 2.430629167018117e-07, "loss": 0.4298, "step": 21549 }, { "epoch": 0.9033461534430902, "grad_norm": 2.263702154159546, "learning_rate": 2.428538762615351e-07, "loss": 0.4705, "step": 21550 }, { "epoch": 0.9033880720580992, "grad_norm": 1.7635586261749268, "learning_rate": 2.4264492351267153e-07, "loss": 0.5049, "step": 21551 }, { "epoch": 0.9034299906731081, "grad_norm": 1.6175000667572021, "learning_rate": 2.4243605845907125e-07, "loss": 0.4564, "step": 21552 }, { "epoch": 0.9034719092881172, "grad_norm": 2.94677734375, "learning_rate": 2.4222728110458507e-07, "loss": 0.5416, "step": 21553 }, { "epoch": 0.9035138279031261, "grad_norm": 1.8377617597579956, "learning_rate": 2.4201859145306206e-07, "loss": 0.5195, "step": 21554 }, { "epoch": 0.9035557465181351, "grad_norm": 1.9689284563064575, "learning_rate": 2.418099895083481e-07, "loss": 0.4879, "step": 21555 }, { "epoch": 0.903597665133144, "grad_norm": 1.7780104875564575, "learning_rate": 2.41601475274289e-07, "loss": 0.4808, "step": 21556 }, { "epoch": 0.903639583748153, "grad_norm": 2.2112269401550293, "learning_rate": 2.413930487547289e-07, "loss": 0.4682, "step": 21557 }, { "epoch": 0.9036815023631619, "grad_norm": 2.164797306060791, "learning_rate": 2.411847099535092e-07, "loss": 0.4939, "step": 21558 }, { "epoch": 0.9037234209781709, "grad_norm": 1.8200387954711914, "learning_rate": 2.409764588744701e-07, "loss": 0.4796, "step": 21559 }, { "epoch": 0.9037653395931798, "grad_norm": 1.9001721143722534, "learning_rate": 2.4076829552145187e-07, "loss": 0.4834, "step": 21560 }, { "epoch": 0.9038072582081887, "grad_norm": 2.089055061340332, "learning_rate": 2.4056021989828927e-07, "loss": 0.4776, "step": 21561 }, { "epoch": 0.9038491768231978, "grad_norm": 1.9478392601013184, "learning_rate": 2.4035223200881973e-07, "loss": 0.4668, "step": 21562 }, { "epoch": 0.9038910954382067, "grad_norm": 2.4175426959991455, "learning_rate": 2.4014433185687745e-07, "loss": 0.4768, "step": 21563 }, { "epoch": 0.9039330140532157, "grad_norm": 2.1163220405578613, "learning_rate": 2.399365194462933e-07, "loss": 0.4347, "step": 21564 }, { "epoch": 0.9039749326682246, "grad_norm": 2.135089159011841, "learning_rate": 2.3972879478089906e-07, "loss": 0.4831, "step": 21565 }, { "epoch": 0.9040168512832336, "grad_norm": 2.155715227127075, "learning_rate": 2.39521157864524e-07, "loss": 0.5012, "step": 21566 }, { "epoch": 0.9040587698982425, "grad_norm": 2.0971126556396484, "learning_rate": 2.3931360870099563e-07, "loss": 0.4346, "step": 21567 }, { "epoch": 0.9041006885132515, "grad_norm": 1.6636661291122437, "learning_rate": 2.3910614729413804e-07, "loss": 0.4538, "step": 21568 }, { "epoch": 0.9041426071282604, "grad_norm": 2.1395211219787598, "learning_rate": 2.3889877364777824e-07, "loss": 0.5402, "step": 21569 }, { "epoch": 0.9041845257432695, "grad_norm": 1.7075377702713013, "learning_rate": 2.386914877657365e-07, "loss": 0.4618, "step": 21570 }, { "epoch": 0.9042264443582784, "grad_norm": 2.0192086696624756, "learning_rate": 2.384842896518352e-07, "loss": 0.4965, "step": 21571 }, { "epoch": 0.9042683629732874, "grad_norm": 2.0182042121887207, "learning_rate": 2.382771793098937e-07, "loss": 0.5057, "step": 21572 }, { "epoch": 0.9043102815882963, "grad_norm": 2.1498236656188965, "learning_rate": 2.3807015674372936e-07, "loss": 0.4732, "step": 21573 }, { "epoch": 0.9043522002033053, "grad_norm": 1.7648046016693115, "learning_rate": 2.378632219571586e-07, "loss": 0.4196, "step": 21574 }, { "epoch": 0.9043941188183142, "grad_norm": 1.96664559841156, "learning_rate": 2.3765637495399563e-07, "loss": 0.5734, "step": 21575 }, { "epoch": 0.9044360374333232, "grad_norm": 1.942560076713562, "learning_rate": 2.3744961573805404e-07, "loss": 0.4337, "step": 21576 }, { "epoch": 0.9044779560483321, "grad_norm": 2.351813793182373, "learning_rate": 2.3724294431314465e-07, "loss": 0.4465, "step": 21577 }, { "epoch": 0.9045198746633412, "grad_norm": 1.859272837638855, "learning_rate": 2.3703636068307666e-07, "loss": 0.4987, "step": 21578 }, { "epoch": 0.9045617932783501, "grad_norm": 1.700304627418518, "learning_rate": 2.3682986485165928e-07, "loss": 0.4664, "step": 21579 }, { "epoch": 0.9046037118933591, "grad_norm": 2.6360602378845215, "learning_rate": 2.3662345682269883e-07, "loss": 0.4715, "step": 21580 }, { "epoch": 0.904645630508368, "grad_norm": 2.2387657165527344, "learning_rate": 2.3641713659999897e-07, "loss": 0.5569, "step": 21581 }, { "epoch": 0.904687549123377, "grad_norm": 2.143679141998291, "learning_rate": 2.3621090418736392e-07, "loss": 0.4838, "step": 21582 }, { "epoch": 0.9047294677383859, "grad_norm": 2.54610538482666, "learning_rate": 2.360047595885956e-07, "loss": 0.4277, "step": 21583 }, { "epoch": 0.9047713863533949, "grad_norm": 1.8419007062911987, "learning_rate": 2.357987028074926e-07, "loss": 0.4852, "step": 21584 }, { "epoch": 0.9048133049684038, "grad_norm": 4.989404678344727, "learning_rate": 2.355927338478553e-07, "loss": 0.4195, "step": 21585 }, { "epoch": 0.9048552235834127, "grad_norm": 1.972730278968811, "learning_rate": 2.353868527134795e-07, "loss": 0.4606, "step": 21586 }, { "epoch": 0.9048971421984218, "grad_norm": 1.8334250450134277, "learning_rate": 2.3518105940815884e-07, "loss": 0.4133, "step": 21587 }, { "epoch": 0.9049390608134307, "grad_norm": 2.1813361644744873, "learning_rate": 2.349753539356886e-07, "loss": 0.4367, "step": 21588 }, { "epoch": 0.9049809794284397, "grad_norm": 1.9911819696426392, "learning_rate": 2.3476973629986134e-07, "loss": 0.4756, "step": 21589 }, { "epoch": 0.9050228980434486, "grad_norm": 3.052917242050171, "learning_rate": 2.345642065044651e-07, "loss": 0.4909, "step": 21590 }, { "epoch": 0.9050648166584576, "grad_norm": 2.9030230045318604, "learning_rate": 2.3435876455328966e-07, "loss": 0.5012, "step": 21591 }, { "epoch": 0.9051067352734665, "grad_norm": 3.3486485481262207, "learning_rate": 2.3415341045012252e-07, "loss": 0.4916, "step": 21592 }, { "epoch": 0.9051486538884755, "grad_norm": 2.1817238330841064, "learning_rate": 2.3394814419874844e-07, "loss": 0.4619, "step": 21593 }, { "epoch": 0.9051905725034844, "grad_norm": 2.8345401287078857, "learning_rate": 2.337429658029522e-07, "loss": 0.489, "step": 21594 }, { "epoch": 0.9052324911184935, "grad_norm": 2.015860080718994, "learning_rate": 2.335378752665146e-07, "loss": 0.5078, "step": 21595 }, { "epoch": 0.9052744097335024, "grad_norm": 1.993025779724121, "learning_rate": 2.3333287259321712e-07, "loss": 0.4963, "step": 21596 }, { "epoch": 0.9053163283485114, "grad_norm": 2.0204226970672607, "learning_rate": 2.3312795778683783e-07, "loss": 0.4643, "step": 21597 }, { "epoch": 0.9053582469635203, "grad_norm": 1.877914309501648, "learning_rate": 2.3292313085115537e-07, "loss": 0.4379, "step": 21598 }, { "epoch": 0.9054001655785293, "grad_norm": 2.502796173095703, "learning_rate": 2.3271839178994448e-07, "loss": 0.524, "step": 21599 }, { "epoch": 0.9054420841935382, "grad_norm": 1.8161137104034424, "learning_rate": 2.3251374060697886e-07, "loss": 0.448, "step": 21600 }, { "epoch": 0.9054840028085472, "grad_norm": 1.6338993310928345, "learning_rate": 2.3230917730603263e-07, "loss": 0.4572, "step": 21601 }, { "epoch": 0.9055259214235561, "grad_norm": 3.514728546142578, "learning_rate": 2.3210470189087453e-07, "loss": 0.4959, "step": 21602 }, { "epoch": 0.9055678400385652, "grad_norm": 1.880471110343933, "learning_rate": 2.3190031436527484e-07, "loss": 0.4571, "step": 21603 }, { "epoch": 0.9056097586535741, "grad_norm": 1.896449089050293, "learning_rate": 2.3169601473300218e-07, "loss": 0.4866, "step": 21604 }, { "epoch": 0.9056516772685831, "grad_norm": 1.9468880891799927, "learning_rate": 2.3149180299782137e-07, "loss": 0.4314, "step": 21605 }, { "epoch": 0.905693595883592, "grad_norm": 2.131005048751831, "learning_rate": 2.3128767916349605e-07, "loss": 0.464, "step": 21606 }, { "epoch": 0.905735514498601, "grad_norm": 2.2711634635925293, "learning_rate": 2.3108364323379095e-07, "loss": 0.4892, "step": 21607 }, { "epoch": 0.9057774331136099, "grad_norm": 2.2154226303100586, "learning_rate": 2.308796952124648e-07, "loss": 0.4764, "step": 21608 }, { "epoch": 0.9058193517286189, "grad_norm": 1.9995783567428589, "learning_rate": 2.3067583510327894e-07, "loss": 0.4363, "step": 21609 }, { "epoch": 0.9058612703436278, "grad_norm": 2.001405715942383, "learning_rate": 2.3047206290999101e-07, "loss": 0.439, "step": 21610 }, { "epoch": 0.9059031889586368, "grad_norm": 1.580808162689209, "learning_rate": 2.3026837863635632e-07, "loss": 0.4603, "step": 21611 }, { "epoch": 0.9059451075736458, "grad_norm": 1.9480500221252441, "learning_rate": 2.3006478228613016e-07, "loss": 0.4697, "step": 21612 }, { "epoch": 0.9059870261886547, "grad_norm": 2.166278600692749, "learning_rate": 2.2986127386306567e-07, "loss": 0.4411, "step": 21613 }, { "epoch": 0.9060289448036637, "grad_norm": 1.9459692239761353, "learning_rate": 2.2965785337091483e-07, "loss": 0.4598, "step": 21614 }, { "epoch": 0.9060708634186726, "grad_norm": 1.8889093399047852, "learning_rate": 2.294545208134258e-07, "loss": 0.4943, "step": 21615 }, { "epoch": 0.9061127820336816, "grad_norm": 1.8727431297302246, "learning_rate": 2.2925127619434773e-07, "loss": 0.5374, "step": 21616 }, { "epoch": 0.9061547006486905, "grad_norm": 2.3850173950195312, "learning_rate": 2.290481195174271e-07, "loss": 0.4602, "step": 21617 }, { "epoch": 0.9061966192636995, "grad_norm": 2.1077446937561035, "learning_rate": 2.2884505078640818e-07, "loss": 0.4967, "step": 21618 }, { "epoch": 0.9062385378787085, "grad_norm": 1.940143346786499, "learning_rate": 2.286420700050357e-07, "loss": 0.5094, "step": 21619 }, { "epoch": 0.9062804564937175, "grad_norm": 1.8974926471710205, "learning_rate": 2.2843917717705e-07, "loss": 0.4931, "step": 21620 }, { "epoch": 0.9063223751087264, "grad_norm": 1.7280806303024292, "learning_rate": 2.2823637230619146e-07, "loss": 0.411, "step": 21621 }, { "epoch": 0.9063642937237354, "grad_norm": 1.9331270456314087, "learning_rate": 2.2803365539619927e-07, "loss": 0.4611, "step": 21622 }, { "epoch": 0.9064062123387443, "grad_norm": 1.882190227508545, "learning_rate": 2.2783102645080934e-07, "loss": 0.4727, "step": 21623 }, { "epoch": 0.9064481309537533, "grad_norm": 2.644714832305908, "learning_rate": 2.2762848547375648e-07, "loss": 0.4884, "step": 21624 }, { "epoch": 0.9064900495687622, "grad_norm": 1.9869624376296997, "learning_rate": 2.2742603246877604e-07, "loss": 0.4705, "step": 21625 }, { "epoch": 0.9065319681837712, "grad_norm": 1.8889436721801758, "learning_rate": 2.2722366743959777e-07, "loss": 0.4903, "step": 21626 }, { "epoch": 0.9065738867987801, "grad_norm": 1.7169346809387207, "learning_rate": 2.2702139038995318e-07, "loss": 0.437, "step": 21627 }, { "epoch": 0.9066158054137892, "grad_norm": 3.018000602722168, "learning_rate": 2.2681920132357093e-07, "loss": 0.512, "step": 21628 }, { "epoch": 0.9066577240287981, "grad_norm": 2.4020962715148926, "learning_rate": 2.2661710024417804e-07, "loss": 0.5266, "step": 21629 }, { "epoch": 0.9066996426438071, "grad_norm": 2.43698787689209, "learning_rate": 2.2641508715549987e-07, "loss": 0.4658, "step": 21630 }, { "epoch": 0.906741561258816, "grad_norm": 1.9050925970077515, "learning_rate": 2.2621316206126065e-07, "loss": 0.4793, "step": 21631 }, { "epoch": 0.906783479873825, "grad_norm": 3.092880964279175, "learning_rate": 2.2601132496518185e-07, "loss": 0.4629, "step": 21632 }, { "epoch": 0.9068253984888339, "grad_norm": 2.4084665775299072, "learning_rate": 2.258095758709844e-07, "loss": 0.4922, "step": 21633 }, { "epoch": 0.9068673171038429, "grad_norm": 2.502965211868286, "learning_rate": 2.2560791478238752e-07, "loss": 0.4934, "step": 21634 }, { "epoch": 0.9069092357188518, "grad_norm": 1.7528208494186401, "learning_rate": 2.2540634170310828e-07, "loss": 0.4242, "step": 21635 }, { "epoch": 0.9069511543338608, "grad_norm": 2.058021306991577, "learning_rate": 2.25204856636862e-07, "loss": 0.4592, "step": 21636 }, { "epoch": 0.9069930729488698, "grad_norm": 2.0437209606170654, "learning_rate": 2.2500345958736403e-07, "loss": 0.4302, "step": 21637 }, { "epoch": 0.9070349915638787, "grad_norm": 1.6568541526794434, "learning_rate": 2.2480215055832533e-07, "loss": 0.4516, "step": 21638 }, { "epoch": 0.9070769101788877, "grad_norm": 2.073335886001587, "learning_rate": 2.2460092955345736e-07, "loss": 0.4769, "step": 21639 }, { "epoch": 0.9071188287938966, "grad_norm": 1.8777704238891602, "learning_rate": 2.2439979657647048e-07, "loss": 0.4748, "step": 21640 }, { "epoch": 0.9071607474089056, "grad_norm": 1.721787452697754, "learning_rate": 2.2419875163107063e-07, "loss": 0.48, "step": 21641 }, { "epoch": 0.9072026660239145, "grad_norm": 1.8854196071624756, "learning_rate": 2.2399779472096428e-07, "loss": 0.4427, "step": 21642 }, { "epoch": 0.9072445846389235, "grad_norm": 1.8317192792892456, "learning_rate": 2.237969258498568e-07, "loss": 0.4599, "step": 21643 }, { "epoch": 0.9072865032539325, "grad_norm": 2.223738431930542, "learning_rate": 2.235961450214491e-07, "loss": 0.5006, "step": 21644 }, { "epoch": 0.9073284218689415, "grad_norm": 1.8682576417922974, "learning_rate": 2.2339545223944327e-07, "loss": 0.4654, "step": 21645 }, { "epoch": 0.9073703404839504, "grad_norm": 1.9483962059020996, "learning_rate": 2.231948475075396e-07, "loss": 0.4968, "step": 21646 }, { "epoch": 0.9074122590989594, "grad_norm": 1.732767105102539, "learning_rate": 2.2299433082943467e-07, "loss": 0.4643, "step": 21647 }, { "epoch": 0.9074541777139683, "grad_norm": 2.7248449325561523, "learning_rate": 2.2279390220882546e-07, "loss": 0.4398, "step": 21648 }, { "epoch": 0.9074960963289773, "grad_norm": 1.8423922061920166, "learning_rate": 2.2259356164940627e-07, "loss": 0.5038, "step": 21649 }, { "epoch": 0.9075380149439862, "grad_norm": 2.284282922744751, "learning_rate": 2.2239330915486969e-07, "loss": 0.5057, "step": 21650 }, { "epoch": 0.9075799335589952, "grad_norm": 2.0146591663360596, "learning_rate": 2.221931447289083e-07, "loss": 0.4644, "step": 21651 }, { "epoch": 0.9076218521740042, "grad_norm": 1.793040156364441, "learning_rate": 2.219930683752114e-07, "loss": 0.4918, "step": 21652 }, { "epoch": 0.9076637707890132, "grad_norm": 4.013707637786865, "learning_rate": 2.2179308009746658e-07, "loss": 0.39, "step": 21653 }, { "epoch": 0.9077056894040221, "grad_norm": 2.1660706996917725, "learning_rate": 2.215931798993609e-07, "loss": 0.5092, "step": 21654 }, { "epoch": 0.9077476080190311, "grad_norm": 1.7440476417541504, "learning_rate": 2.213933677845781e-07, "loss": 0.5008, "step": 21655 }, { "epoch": 0.90778952663404, "grad_norm": 2.0574140548706055, "learning_rate": 2.211936437568024e-07, "loss": 0.4584, "step": 21656 }, { "epoch": 0.907831445249049, "grad_norm": 5.449752330780029, "learning_rate": 2.2099400781971648e-07, "loss": 0.4791, "step": 21657 }, { "epoch": 0.9078733638640579, "grad_norm": 1.930094838142395, "learning_rate": 2.2079445997699844e-07, "loss": 0.4807, "step": 21658 }, { "epoch": 0.9079152824790669, "grad_norm": 1.8197976350784302, "learning_rate": 2.2059500023232705e-07, "loss": 0.4879, "step": 21659 }, { "epoch": 0.9079572010940758, "grad_norm": 2.126030683517456, "learning_rate": 2.2039562858938046e-07, "loss": 0.4361, "step": 21660 }, { "epoch": 0.9079991197090848, "grad_norm": 1.97755765914917, "learning_rate": 2.2019634505183297e-07, "loss": 0.4631, "step": 21661 }, { "epoch": 0.9080410383240938, "grad_norm": 2.1708953380584717, "learning_rate": 2.1999714962335716e-07, "loss": 0.5437, "step": 21662 }, { "epoch": 0.9080829569391027, "grad_norm": 1.7226688861846924, "learning_rate": 2.197980423076268e-07, "loss": 0.4359, "step": 21663 }, { "epoch": 0.9081248755541117, "grad_norm": 2.2709147930145264, "learning_rate": 2.1959902310831006e-07, "loss": 0.48, "step": 21664 }, { "epoch": 0.9081667941691206, "grad_norm": 1.6253585815429688, "learning_rate": 2.1940009202907674e-07, "loss": 0.4984, "step": 21665 }, { "epoch": 0.9082087127841296, "grad_norm": 1.955033779144287, "learning_rate": 2.1920124907359396e-07, "loss": 0.4974, "step": 21666 }, { "epoch": 0.9082506313991385, "grad_norm": 2.0917632579803467, "learning_rate": 2.1900249424552657e-07, "loss": 0.51, "step": 21667 }, { "epoch": 0.9082925500141475, "grad_norm": 2.757782220840454, "learning_rate": 2.188038275485388e-07, "loss": 0.4944, "step": 21668 }, { "epoch": 0.9083344686291565, "grad_norm": 2.370696783065796, "learning_rate": 2.186052489862933e-07, "loss": 0.5367, "step": 21669 }, { "epoch": 0.9083763872441655, "grad_norm": 2.2876060009002686, "learning_rate": 2.1840675856244997e-07, "loss": 0.4874, "step": 21670 }, { "epoch": 0.9084183058591744, "grad_norm": 2.0159780979156494, "learning_rate": 2.1820835628066695e-07, "loss": 0.453, "step": 21671 }, { "epoch": 0.9084602244741834, "grad_norm": 1.7199454307556152, "learning_rate": 2.1801004214460297e-07, "loss": 0.4926, "step": 21672 }, { "epoch": 0.9085021430891923, "grad_norm": 1.8348708152770996, "learning_rate": 2.1781181615791237e-07, "loss": 0.4593, "step": 21673 }, { "epoch": 0.9085440617042013, "grad_norm": 1.9361772537231445, "learning_rate": 2.176136783242494e-07, "loss": 0.5037, "step": 21674 }, { "epoch": 0.9085859803192102, "grad_norm": 2.2840025424957275, "learning_rate": 2.1741562864726784e-07, "loss": 0.4947, "step": 21675 }, { "epoch": 0.9086278989342192, "grad_norm": 2.1008589267730713, "learning_rate": 2.1721766713061644e-07, "loss": 0.5081, "step": 21676 }, { "epoch": 0.9086698175492282, "grad_norm": 2.183544874191284, "learning_rate": 2.1701979377794556e-07, "loss": 0.4168, "step": 21677 }, { "epoch": 0.9087117361642372, "grad_norm": 1.8157190084457397, "learning_rate": 2.1682200859290349e-07, "loss": 0.4766, "step": 21678 }, { "epoch": 0.9087536547792461, "grad_norm": 2.106111764907837, "learning_rate": 2.1662431157913444e-07, "loss": 0.4786, "step": 21679 }, { "epoch": 0.9087955733942551, "grad_norm": 1.7443090677261353, "learning_rate": 2.1642670274028333e-07, "loss": 0.4771, "step": 21680 }, { "epoch": 0.908837492009264, "grad_norm": 1.8023508787155151, "learning_rate": 2.1622918207999334e-07, "loss": 0.4527, "step": 21681 }, { "epoch": 0.908879410624273, "grad_norm": 1.8384045362472534, "learning_rate": 2.1603174960190431e-07, "loss": 0.4393, "step": 21682 }, { "epoch": 0.9089213292392819, "grad_norm": 2.747809648513794, "learning_rate": 2.1583440530965615e-07, "loss": 0.4714, "step": 21683 }, { "epoch": 0.908963247854291, "grad_norm": 2.284106492996216, "learning_rate": 2.156371492068876e-07, "loss": 0.4912, "step": 21684 }, { "epoch": 0.9090051664692999, "grad_norm": 2.42753529548645, "learning_rate": 2.1543998129723354e-07, "loss": 0.4944, "step": 21685 }, { "epoch": 0.9090470850843088, "grad_norm": 1.668925404548645, "learning_rate": 2.152429015843288e-07, "loss": 0.4594, "step": 21686 }, { "epoch": 0.9090890036993178, "grad_norm": 2.132641553878784, "learning_rate": 2.150459100718072e-07, "loss": 0.5121, "step": 21687 }, { "epoch": 0.9091309223143267, "grad_norm": 2.078383207321167, "learning_rate": 2.1484900676329857e-07, "loss": 0.5505, "step": 21688 }, { "epoch": 0.9091728409293357, "grad_norm": 1.811691403388977, "learning_rate": 2.1465219166243388e-07, "loss": 0.4277, "step": 21689 }, { "epoch": 0.9092147595443446, "grad_norm": 1.8946179151535034, "learning_rate": 2.1445546477284084e-07, "loss": 0.4227, "step": 21690 }, { "epoch": 0.9092566781593536, "grad_norm": 2.123103380203247, "learning_rate": 2.142588260981443e-07, "loss": 0.4913, "step": 21691 }, { "epoch": 0.9092985967743625, "grad_norm": 1.794787883758545, "learning_rate": 2.1406227564197023e-07, "loss": 0.4679, "step": 21692 }, { "epoch": 0.9093405153893716, "grad_norm": 2.0992870330810547, "learning_rate": 2.1386581340794298e-07, "loss": 0.4237, "step": 21693 }, { "epoch": 0.9093824340043805, "grad_norm": 3.6960887908935547, "learning_rate": 2.1366943939968132e-07, "loss": 0.4573, "step": 21694 }, { "epoch": 0.9094243526193895, "grad_norm": 1.7555829286575317, "learning_rate": 2.1347315362080733e-07, "loss": 0.4773, "step": 21695 }, { "epoch": 0.9094662712343984, "grad_norm": 2.055832862854004, "learning_rate": 2.132769560749387e-07, "loss": 0.4719, "step": 21696 }, { "epoch": 0.9095081898494074, "grad_norm": 1.9663339853286743, "learning_rate": 2.1308084676569197e-07, "loss": 0.462, "step": 21697 }, { "epoch": 0.9095501084644163, "grad_norm": 1.9537012577056885, "learning_rate": 2.12884825696682e-07, "loss": 0.5265, "step": 21698 }, { "epoch": 0.9095920270794253, "grad_norm": 1.9527661800384521, "learning_rate": 2.1268889287152261e-07, "loss": 0.4941, "step": 21699 }, { "epoch": 0.9096339456944342, "grad_norm": 1.753604531288147, "learning_rate": 2.1249304829382478e-07, "loss": 0.436, "step": 21700 }, { "epoch": 0.9096758643094432, "grad_norm": 1.7757574319839478, "learning_rate": 2.1229729196719895e-07, "loss": 0.4499, "step": 21701 }, { "epoch": 0.9097177829244522, "grad_norm": 2.5035812854766846, "learning_rate": 2.1210162389525446e-07, "loss": 0.4619, "step": 21702 }, { "epoch": 0.9097597015394612, "grad_norm": 1.756069540977478, "learning_rate": 2.1190604408159677e-07, "loss": 0.4725, "step": 21703 }, { "epoch": 0.9098016201544701, "grad_norm": 1.7966831922531128, "learning_rate": 2.117105525298313e-07, "loss": 0.4893, "step": 21704 }, { "epoch": 0.9098435387694791, "grad_norm": 2.06107234954834, "learning_rate": 2.1151514924356354e-07, "loss": 0.4734, "step": 21705 }, { "epoch": 0.909885457384488, "grad_norm": 2.7092158794403076, "learning_rate": 2.1131983422639335e-07, "loss": 0.5131, "step": 21706 }, { "epoch": 0.909927375999497, "grad_norm": 1.7510899305343628, "learning_rate": 2.1112460748192232e-07, "loss": 0.4553, "step": 21707 }, { "epoch": 0.9099692946145059, "grad_norm": 1.7848849296569824, "learning_rate": 2.1092946901374867e-07, "loss": 0.4711, "step": 21708 }, { "epoch": 0.910011213229515, "grad_norm": 1.7986421585083008, "learning_rate": 2.1073441882546896e-07, "loss": 0.5053, "step": 21709 }, { "epoch": 0.9100531318445239, "grad_norm": 1.8156883716583252, "learning_rate": 2.1053945692067922e-07, "loss": 0.5371, "step": 21710 }, { "epoch": 0.9100950504595328, "grad_norm": 1.7973706722259521, "learning_rate": 2.1034458330297436e-07, "loss": 0.4878, "step": 21711 }, { "epoch": 0.9101369690745418, "grad_norm": 1.9818342924118042, "learning_rate": 2.1014979797594481e-07, "loss": 0.4872, "step": 21712 }, { "epoch": 0.9101788876895507, "grad_norm": 2.5640642642974854, "learning_rate": 2.099551009431816e-07, "loss": 0.5081, "step": 21713 }, { "epoch": 0.9102208063045597, "grad_norm": 1.734667181968689, "learning_rate": 2.0976049220827522e-07, "loss": 0.478, "step": 21714 }, { "epoch": 0.9102627249195686, "grad_norm": 1.8906522989273071, "learning_rate": 2.095659717748111e-07, "loss": 0.506, "step": 21715 }, { "epoch": 0.9103046435345776, "grad_norm": 1.8592169284820557, "learning_rate": 2.0937153964637635e-07, "loss": 0.4442, "step": 21716 }, { "epoch": 0.9103465621495865, "grad_norm": 1.965209722518921, "learning_rate": 2.0917719582655427e-07, "loss": 0.4987, "step": 21717 }, { "epoch": 0.9103884807645956, "grad_norm": 1.7969890832901, "learning_rate": 2.0898294031892641e-07, "loss": 0.4194, "step": 21718 }, { "epoch": 0.9104303993796045, "grad_norm": 2.0580835342407227, "learning_rate": 2.0878877312707546e-07, "loss": 0.4759, "step": 21719 }, { "epoch": 0.9104723179946135, "grad_norm": 1.7574081420898438, "learning_rate": 2.0859469425458022e-07, "loss": 0.4495, "step": 21720 }, { "epoch": 0.9105142366096224, "grad_norm": 2.0004899501800537, "learning_rate": 2.0840070370501675e-07, "loss": 0.4803, "step": 21721 }, { "epoch": 0.9105561552246314, "grad_norm": 3.6954214572906494, "learning_rate": 2.0820680148196325e-07, "loss": 0.4865, "step": 21722 }, { "epoch": 0.9105980738396403, "grad_norm": 2.16656231880188, "learning_rate": 2.0801298758899246e-07, "loss": 0.4814, "step": 21723 }, { "epoch": 0.9106399924546493, "grad_norm": 2.3264687061309814, "learning_rate": 2.0781926202967706e-07, "loss": 0.4237, "step": 21724 }, { "epoch": 0.9106819110696582, "grad_norm": 2.1246676445007324, "learning_rate": 2.0762562480758974e-07, "loss": 0.5087, "step": 21725 }, { "epoch": 0.9107238296846673, "grad_norm": 1.8185511827468872, "learning_rate": 2.0743207592629822e-07, "loss": 0.5092, "step": 21726 }, { "epoch": 0.9107657482996762, "grad_norm": 2.24528431892395, "learning_rate": 2.0723861538937017e-07, "loss": 0.5032, "step": 21727 }, { "epoch": 0.9108076669146852, "grad_norm": 1.9434616565704346, "learning_rate": 2.0704524320037334e-07, "loss": 0.4806, "step": 21728 }, { "epoch": 0.9108495855296941, "grad_norm": 2.021601915359497, "learning_rate": 2.0685195936287094e-07, "loss": 0.5124, "step": 21729 }, { "epoch": 0.9108915041447031, "grad_norm": 1.88805091381073, "learning_rate": 2.0665876388042628e-07, "loss": 0.4084, "step": 21730 }, { "epoch": 0.910933422759712, "grad_norm": 1.9177395105361938, "learning_rate": 2.0646565675660092e-07, "loss": 0.517, "step": 21731 }, { "epoch": 0.910975341374721, "grad_norm": 2.0350847244262695, "learning_rate": 2.0627263799495422e-07, "loss": 0.5282, "step": 21732 }, { "epoch": 0.9110172599897299, "grad_norm": 2.710219383239746, "learning_rate": 2.0607970759904393e-07, "loss": 0.4549, "step": 21733 }, { "epoch": 0.911059178604739, "grad_norm": 1.8957597017288208, "learning_rate": 2.0588686557242776e-07, "loss": 0.4877, "step": 21734 }, { "epoch": 0.9111010972197479, "grad_norm": 2.2526516914367676, "learning_rate": 2.0569411191865895e-07, "loss": 0.4547, "step": 21735 }, { "epoch": 0.9111430158347568, "grad_norm": 2.156019926071167, "learning_rate": 2.0550144664129136e-07, "loss": 0.5269, "step": 21736 }, { "epoch": 0.9111849344497658, "grad_norm": 2.2927515506744385, "learning_rate": 2.053088697438771e-07, "loss": 0.4558, "step": 21737 }, { "epoch": 0.9112268530647747, "grad_norm": 1.9131548404693604, "learning_rate": 2.0511638122996391e-07, "loss": 0.5159, "step": 21738 }, { "epoch": 0.9112687716797837, "grad_norm": 1.78616201877594, "learning_rate": 2.049239811031023e-07, "loss": 0.4623, "step": 21739 }, { "epoch": 0.9113106902947926, "grad_norm": 1.996537685394287, "learning_rate": 2.047316693668383e-07, "loss": 0.5188, "step": 21740 }, { "epoch": 0.9113526089098016, "grad_norm": 1.7287791967391968, "learning_rate": 2.0453944602471632e-07, "loss": 0.481, "step": 21741 }, { "epoch": 0.9113945275248105, "grad_norm": 2.018409490585327, "learning_rate": 2.0434731108028015e-07, "loss": 0.432, "step": 21742 }, { "epoch": 0.9114364461398196, "grad_norm": 2.0274546146392822, "learning_rate": 2.04155264537072e-07, "loss": 0.4783, "step": 21743 }, { "epoch": 0.9114783647548285, "grad_norm": 1.3743103742599487, "learning_rate": 2.039633063986307e-07, "loss": 0.3992, "step": 21744 }, { "epoch": 0.9115202833698375, "grad_norm": 1.9395384788513184, "learning_rate": 2.037714366684962e-07, "loss": 0.4752, "step": 21745 }, { "epoch": 0.9115622019848464, "grad_norm": 1.7381654977798462, "learning_rate": 2.0357965535020508e-07, "loss": 0.4909, "step": 21746 }, { "epoch": 0.9116041205998554, "grad_norm": 1.675973892211914, "learning_rate": 2.033879624472912e-07, "loss": 0.4358, "step": 21747 }, { "epoch": 0.9116460392148643, "grad_norm": 1.7823834419250488, "learning_rate": 2.0319635796328895e-07, "loss": 0.4814, "step": 21748 }, { "epoch": 0.9116879578298733, "grad_norm": 2.312871217727661, "learning_rate": 2.0300484190173165e-07, "loss": 0.5321, "step": 21749 }, { "epoch": 0.9117298764448822, "grad_norm": 2.3883895874023438, "learning_rate": 2.02813414266147e-07, "loss": 0.492, "step": 21750 }, { "epoch": 0.9117717950598913, "grad_norm": 1.8631800413131714, "learning_rate": 2.0262207506006602e-07, "loss": 0.4729, "step": 21751 }, { "epoch": 0.9118137136749002, "grad_norm": 1.9720432758331299, "learning_rate": 2.0243082428701488e-07, "loss": 0.4676, "step": 21752 }, { "epoch": 0.9118556322899092, "grad_norm": 2.3331756591796875, "learning_rate": 2.0223966195051902e-07, "loss": 0.5055, "step": 21753 }, { "epoch": 0.9118975509049181, "grad_norm": 1.7755506038665771, "learning_rate": 2.0204858805410232e-07, "loss": 0.4922, "step": 21754 }, { "epoch": 0.9119394695199271, "grad_norm": 1.7871335744857788, "learning_rate": 2.0185760260128695e-07, "loss": 0.4919, "step": 21755 }, { "epoch": 0.911981388134936, "grad_norm": 1.8466604948043823, "learning_rate": 2.0166670559559343e-07, "loss": 0.4785, "step": 21756 }, { "epoch": 0.912023306749945, "grad_norm": 1.767537236213684, "learning_rate": 2.0147589704054005e-07, "loss": 0.4886, "step": 21757 }, { "epoch": 0.9120652253649539, "grad_norm": 1.880072832107544, "learning_rate": 2.0128517693964567e-07, "loss": 0.445, "step": 21758 }, { "epoch": 0.912107143979963, "grad_norm": 1.931490182876587, "learning_rate": 2.010945452964247e-07, "loss": 0.4724, "step": 21759 }, { "epoch": 0.9121490625949719, "grad_norm": 1.9476807117462158, "learning_rate": 2.00904002114391e-07, "loss": 0.4765, "step": 21760 }, { "epoch": 0.9121909812099808, "grad_norm": 1.772918939590454, "learning_rate": 2.007135473970584e-07, "loss": 0.4831, "step": 21761 }, { "epoch": 0.9122328998249898, "grad_norm": 1.799064040184021, "learning_rate": 2.0052318114793578e-07, "loss": 0.4783, "step": 21762 }, { "epoch": 0.9122748184399987, "grad_norm": 2.372951030731201, "learning_rate": 2.0033290337053424e-07, "loss": 0.4748, "step": 21763 }, { "epoch": 0.9123167370550077, "grad_norm": 1.639140248298645, "learning_rate": 2.0014271406835984e-07, "loss": 0.5198, "step": 21764 }, { "epoch": 0.9123586556700166, "grad_norm": 1.733924150466919, "learning_rate": 1.9995261324491867e-07, "loss": 0.54, "step": 21765 }, { "epoch": 0.9124005742850256, "grad_norm": 1.9412294626235962, "learning_rate": 1.9976260090371514e-07, "loss": 0.4502, "step": 21766 }, { "epoch": 0.9124424929000345, "grad_norm": 1.667534589767456, "learning_rate": 1.9957267704825256e-07, "loss": 0.4811, "step": 21767 }, { "epoch": 0.9124844115150436, "grad_norm": 1.5540109872817993, "learning_rate": 1.9938284168203036e-07, "loss": 0.4427, "step": 21768 }, { "epoch": 0.9125263301300525, "grad_norm": 2.539119005203247, "learning_rate": 1.991930948085491e-07, "loss": 0.4408, "step": 21769 }, { "epoch": 0.9125682487450615, "grad_norm": 1.9617676734924316, "learning_rate": 1.9900343643130705e-07, "loss": 0.3708, "step": 21770 }, { "epoch": 0.9126101673600704, "grad_norm": 2.234205961227417, "learning_rate": 1.9881386655379865e-07, "loss": 0.4713, "step": 21771 }, { "epoch": 0.9126520859750794, "grad_norm": 2.0147252082824707, "learning_rate": 1.9862438517951943e-07, "loss": 0.4435, "step": 21772 }, { "epoch": 0.9126940045900883, "grad_norm": 1.657863736152649, "learning_rate": 1.9843499231196272e-07, "loss": 0.4297, "step": 21773 }, { "epoch": 0.9127359232050973, "grad_norm": 1.6969066858291626, "learning_rate": 1.9824568795461795e-07, "loss": 0.479, "step": 21774 }, { "epoch": 0.9127778418201062, "grad_norm": 4.860257148742676, "learning_rate": 1.9805647211097566e-07, "loss": 0.4653, "step": 21775 }, { "epoch": 0.9128197604351153, "grad_norm": 1.776990294456482, "learning_rate": 1.9786734478452418e-07, "loss": 0.3779, "step": 21776 }, { "epoch": 0.9128616790501242, "grad_norm": 1.9767208099365234, "learning_rate": 1.9767830597874903e-07, "loss": 0.4465, "step": 21777 }, { "epoch": 0.9129035976651332, "grad_norm": 2.0257532596588135, "learning_rate": 1.9748935569713524e-07, "loss": 0.5033, "step": 21778 }, { "epoch": 0.9129455162801421, "grad_norm": 1.8464665412902832, "learning_rate": 1.9730049394316665e-07, "loss": 0.5175, "step": 21779 }, { "epoch": 0.9129874348951511, "grad_norm": 1.9947357177734375, "learning_rate": 1.971117207203238e-07, "loss": 0.5078, "step": 21780 }, { "epoch": 0.91302935351016, "grad_norm": 1.901716709136963, "learning_rate": 1.9692303603208618e-07, "loss": 0.4732, "step": 21781 }, { "epoch": 0.913071272125169, "grad_norm": 2.271848440170288, "learning_rate": 1.9673443988193263e-07, "loss": 0.4362, "step": 21782 }, { "epoch": 0.9131131907401779, "grad_norm": 2.0914833545684814, "learning_rate": 1.965459322733393e-07, "loss": 0.4408, "step": 21783 }, { "epoch": 0.913155109355187, "grad_norm": 1.8424646854400635, "learning_rate": 1.9635751320978113e-07, "loss": 0.4647, "step": 21784 }, { "epoch": 0.9131970279701959, "grad_norm": 1.8351184129714966, "learning_rate": 1.961691826947315e-07, "loss": 0.5202, "step": 21785 }, { "epoch": 0.9132389465852048, "grad_norm": 1.922704815864563, "learning_rate": 1.959809407316615e-07, "loss": 0.488, "step": 21786 }, { "epoch": 0.9132808652002138, "grad_norm": 1.8831095695495605, "learning_rate": 1.9579278732404117e-07, "loss": 0.4566, "step": 21787 }, { "epoch": 0.9133227838152227, "grad_norm": 2.032492160797119, "learning_rate": 1.9560472247534047e-07, "loss": 0.4259, "step": 21788 }, { "epoch": 0.9133647024302317, "grad_norm": 3.204075574874878, "learning_rate": 1.9541674618902383e-07, "loss": 0.5212, "step": 21789 }, { "epoch": 0.9134066210452406, "grad_norm": 1.7508958578109741, "learning_rate": 1.9522885846855745e-07, "loss": 0.4679, "step": 21790 }, { "epoch": 0.9134485396602496, "grad_norm": 2.006096124649048, "learning_rate": 1.9504105931740514e-07, "loss": 0.4811, "step": 21791 }, { "epoch": 0.9134904582752585, "grad_norm": 1.793396234512329, "learning_rate": 1.9485334873902806e-07, "loss": 0.4732, "step": 21792 }, { "epoch": 0.9135323768902676, "grad_norm": 1.7408950328826904, "learning_rate": 1.9466572673688623e-07, "loss": 0.5223, "step": 21793 }, { "epoch": 0.9135742955052765, "grad_norm": 2.12697696685791, "learning_rate": 1.9447819331443908e-07, "loss": 0.4723, "step": 21794 }, { "epoch": 0.9136162141202855, "grad_norm": 1.81796395778656, "learning_rate": 1.9429074847514217e-07, "loss": 0.4642, "step": 21795 }, { "epoch": 0.9136581327352944, "grad_norm": 1.93450129032135, "learning_rate": 1.941033922224528e-07, "loss": 0.4689, "step": 21796 }, { "epoch": 0.9137000513503034, "grad_norm": 2.079047441482544, "learning_rate": 1.9391612455982255e-07, "loss": 0.5012, "step": 21797 }, { "epoch": 0.9137419699653123, "grad_norm": 2.141085624694824, "learning_rate": 1.937289454907043e-07, "loss": 0.4746, "step": 21798 }, { "epoch": 0.9137838885803213, "grad_norm": 1.7877839803695679, "learning_rate": 1.935418550185486e-07, "loss": 0.4125, "step": 21799 }, { "epoch": 0.9138258071953302, "grad_norm": 2.1867034435272217, "learning_rate": 1.9335485314680436e-07, "loss": 0.5181, "step": 21800 }, { "epoch": 0.9138677258103393, "grad_norm": 3.123744487762451, "learning_rate": 1.9316793987891825e-07, "loss": 0.4999, "step": 21801 }, { "epoch": 0.9139096444253482, "grad_norm": 1.8499935865402222, "learning_rate": 1.929811152183364e-07, "loss": 0.4642, "step": 21802 }, { "epoch": 0.9139515630403572, "grad_norm": 2.1305384635925293, "learning_rate": 1.9279437916850164e-07, "loss": 0.502, "step": 21803 }, { "epoch": 0.9139934816553661, "grad_norm": 1.7483649253845215, "learning_rate": 1.926077317328562e-07, "loss": 0.4996, "step": 21804 }, { "epoch": 0.9140354002703751, "grad_norm": 2.8243396282196045, "learning_rate": 1.924211729148423e-07, "loss": 0.4416, "step": 21805 }, { "epoch": 0.914077318885384, "grad_norm": 1.7831178903579712, "learning_rate": 1.9223470271789725e-07, "loss": 0.4321, "step": 21806 }, { "epoch": 0.914119237500393, "grad_norm": 1.8270014524459839, "learning_rate": 1.920483211454588e-07, "loss": 0.4883, "step": 21807 }, { "epoch": 0.9141611561154019, "grad_norm": 1.9624080657958984, "learning_rate": 1.9186202820096312e-07, "loss": 0.5137, "step": 21808 }, { "epoch": 0.914203074730411, "grad_norm": 2.8245229721069336, "learning_rate": 1.9167582388784357e-07, "loss": 0.4458, "step": 21809 }, { "epoch": 0.9142449933454199, "grad_norm": 2.360708236694336, "learning_rate": 1.914897082095335e-07, "loss": 0.4672, "step": 21810 }, { "epoch": 0.9142869119604288, "grad_norm": 1.966388463973999, "learning_rate": 1.9130368116946295e-07, "loss": 0.488, "step": 21811 }, { "epoch": 0.9143288305754378, "grad_norm": 2.103442668914795, "learning_rate": 1.9111774277106033e-07, "loss": 0.4848, "step": 21812 }, { "epoch": 0.9143707491904467, "grad_norm": 1.9914720058441162, "learning_rate": 1.9093189301775449e-07, "loss": 0.4962, "step": 21813 }, { "epoch": 0.9144126678054557, "grad_norm": 1.6883548498153687, "learning_rate": 1.9074613191297108e-07, "loss": 0.4948, "step": 21814 }, { "epoch": 0.9144545864204646, "grad_norm": 1.6963763236999512, "learning_rate": 1.9056045946013403e-07, "loss": 0.4678, "step": 21815 }, { "epoch": 0.9144965050354736, "grad_norm": 2.347022294998169, "learning_rate": 1.9037487566266554e-07, "loss": 0.4788, "step": 21816 }, { "epoch": 0.9145384236504825, "grad_norm": 1.7698163986206055, "learning_rate": 1.9018938052398793e-07, "loss": 0.5202, "step": 21817 }, { "epoch": 0.9145803422654916, "grad_norm": 1.747215986251831, "learning_rate": 1.9000397404751958e-07, "loss": 0.4531, "step": 21818 }, { "epoch": 0.9146222608805005, "grad_norm": 1.7456680536270142, "learning_rate": 1.8981865623667827e-07, "loss": 0.4781, "step": 21819 }, { "epoch": 0.9146641794955095, "grad_norm": 1.6339848041534424, "learning_rate": 1.8963342709488074e-07, "loss": 0.4832, "step": 21820 }, { "epoch": 0.9147060981105184, "grad_norm": 1.9106574058532715, "learning_rate": 1.8944828662553984e-07, "loss": 0.4856, "step": 21821 }, { "epoch": 0.9147480167255274, "grad_norm": 1.6424294710159302, "learning_rate": 1.8926323483206944e-07, "loss": 0.4453, "step": 21822 }, { "epoch": 0.9147899353405363, "grad_norm": 1.9156785011291504, "learning_rate": 1.8907827171788128e-07, "loss": 0.4581, "step": 21823 }, { "epoch": 0.9148318539555453, "grad_norm": 1.6881523132324219, "learning_rate": 1.8889339728638378e-07, "loss": 0.4309, "step": 21824 }, { "epoch": 0.9148737725705542, "grad_norm": 1.800531029701233, "learning_rate": 1.887086115409853e-07, "loss": 0.4284, "step": 21825 }, { "epoch": 0.9149156911855633, "grad_norm": 2.1152539253234863, "learning_rate": 1.8852391448509255e-07, "loss": 0.4551, "step": 21826 }, { "epoch": 0.9149576098005722, "grad_norm": 2.1769003868103027, "learning_rate": 1.883393061221095e-07, "loss": 0.4917, "step": 21827 }, { "epoch": 0.9149995284155812, "grad_norm": 1.9440611600875854, "learning_rate": 1.8815478645544006e-07, "loss": 0.4614, "step": 21828 }, { "epoch": 0.9150414470305901, "grad_norm": 1.9885401725769043, "learning_rate": 1.8797035548848374e-07, "loss": 0.4722, "step": 21829 }, { "epoch": 0.9150833656455991, "grad_norm": 1.5214961767196655, "learning_rate": 1.8778601322464286e-07, "loss": 0.4593, "step": 21830 }, { "epoch": 0.915125284260608, "grad_norm": 1.9235374927520752, "learning_rate": 1.8760175966731298e-07, "loss": 0.4917, "step": 21831 }, { "epoch": 0.915167202875617, "grad_norm": 1.9336063861846924, "learning_rate": 1.8741759481989253e-07, "loss": 0.5019, "step": 21832 }, { "epoch": 0.9152091214906259, "grad_norm": 1.9767042398452759, "learning_rate": 1.872335186857749e-07, "loss": 0.4631, "step": 21833 }, { "epoch": 0.915251040105635, "grad_norm": 1.9491603374481201, "learning_rate": 1.87049531268354e-07, "loss": 0.4114, "step": 21834 }, { "epoch": 0.9152929587206439, "grad_norm": 2.305774688720703, "learning_rate": 1.868656325710222e-07, "loss": 0.4787, "step": 21835 }, { "epoch": 0.9153348773356528, "grad_norm": 1.8667529821395874, "learning_rate": 1.8668182259716727e-07, "loss": 0.4973, "step": 21836 }, { "epoch": 0.9153767959506618, "grad_norm": 1.835091471672058, "learning_rate": 1.8649810135017875e-07, "loss": 0.4867, "step": 21837 }, { "epoch": 0.9154187145656707, "grad_norm": 1.7947032451629639, "learning_rate": 1.8631446883344451e-07, "loss": 0.4591, "step": 21838 }, { "epoch": 0.9154606331806797, "grad_norm": 2.0582027435302734, "learning_rate": 1.861309250503479e-07, "loss": 0.4681, "step": 21839 }, { "epoch": 0.9155025517956886, "grad_norm": 2.380260944366455, "learning_rate": 1.8594747000427183e-07, "loss": 0.5055, "step": 21840 }, { "epoch": 0.9155444704106976, "grad_norm": 1.8854888677597046, "learning_rate": 1.8576410369859966e-07, "loss": 0.499, "step": 21841 }, { "epoch": 0.9155863890257065, "grad_norm": 1.8017340898513794, "learning_rate": 1.8558082613671035e-07, "loss": 0.4806, "step": 21842 }, { "epoch": 0.9156283076407156, "grad_norm": 2.087545394897461, "learning_rate": 1.853976373219829e-07, "loss": 0.4993, "step": 21843 }, { "epoch": 0.9156702262557245, "grad_norm": 1.538559079170227, "learning_rate": 1.8521453725779458e-07, "loss": 0.4669, "step": 21844 }, { "epoch": 0.9157121448707335, "grad_norm": 1.6265053749084473, "learning_rate": 1.850315259475194e-07, "loss": 0.4217, "step": 21845 }, { "epoch": 0.9157540634857424, "grad_norm": 2.0379462242126465, "learning_rate": 1.8484860339453125e-07, "loss": 0.4882, "step": 21846 }, { "epoch": 0.9157959821007514, "grad_norm": 1.913557529449463, "learning_rate": 1.8466576960220305e-07, "loss": 0.4921, "step": 21847 }, { "epoch": 0.9158379007157603, "grad_norm": 1.870491862297058, "learning_rate": 1.8448302457390432e-07, "loss": 0.4558, "step": 21848 }, { "epoch": 0.9158798193307693, "grad_norm": 1.7576806545257568, "learning_rate": 1.8430036831300346e-07, "loss": 0.3979, "step": 21849 }, { "epoch": 0.9159217379457782, "grad_norm": 2.7061269283294678, "learning_rate": 1.8411780082286834e-07, "loss": 0.4329, "step": 21850 }, { "epoch": 0.9159636565607873, "grad_norm": 2.214543581008911, "learning_rate": 1.839353221068635e-07, "loss": 0.47, "step": 21851 }, { "epoch": 0.9160055751757962, "grad_norm": 2.8391661643981934, "learning_rate": 1.8375293216835233e-07, "loss": 0.4947, "step": 21852 }, { "epoch": 0.9160474937908052, "grad_norm": 2.1782093048095703, "learning_rate": 1.8357063101069827e-07, "loss": 0.505, "step": 21853 }, { "epoch": 0.9160894124058141, "grad_norm": 1.918781042098999, "learning_rate": 1.8338841863726088e-07, "loss": 0.5246, "step": 21854 }, { "epoch": 0.9161313310208231, "grad_norm": 1.7893943786621094, "learning_rate": 1.832062950513991e-07, "loss": 0.4456, "step": 21855 }, { "epoch": 0.916173249635832, "grad_norm": 1.800390601158142, "learning_rate": 1.830242602564708e-07, "loss": 0.4352, "step": 21856 }, { "epoch": 0.916215168250841, "grad_norm": 2.23048734664917, "learning_rate": 1.828423142558311e-07, "loss": 0.4482, "step": 21857 }, { "epoch": 0.91625708686585, "grad_norm": 1.9699329137802124, "learning_rate": 1.8266045705283343e-07, "loss": 0.4832, "step": 21858 }, { "epoch": 0.916299005480859, "grad_norm": 1.8758677244186401, "learning_rate": 1.8247868865083064e-07, "loss": 0.4688, "step": 21859 }, { "epoch": 0.9163409240958679, "grad_norm": 2.3071653842926025, "learning_rate": 1.8229700905317282e-07, "loss": 0.4426, "step": 21860 }, { "epoch": 0.9163828427108768, "grad_norm": 1.9069745540618896, "learning_rate": 1.82115418263209e-07, "loss": 0.5484, "step": 21861 }, { "epoch": 0.9164247613258858, "grad_norm": 1.960137128829956, "learning_rate": 1.8193391628428814e-07, "loss": 0.5088, "step": 21862 }, { "epoch": 0.9164666799408947, "grad_norm": 1.8609426021575928, "learning_rate": 1.8175250311975367e-07, "loss": 0.4521, "step": 21863 }, { "epoch": 0.9165085985559037, "grad_norm": 1.9412072896957397, "learning_rate": 1.8157117877295072e-07, "loss": 0.5481, "step": 21864 }, { "epoch": 0.9165505171709126, "grad_norm": 2.0453829765319824, "learning_rate": 1.813899432472227e-07, "loss": 0.4922, "step": 21865 }, { "epoch": 0.9165924357859216, "grad_norm": 1.7698239088058472, "learning_rate": 1.8120879654590974e-07, "loss": 0.4543, "step": 21866 }, { "epoch": 0.9166343544009306, "grad_norm": 2.0676865577697754, "learning_rate": 1.810277386723497e-07, "loss": 0.4747, "step": 21867 }, { "epoch": 0.9166762730159396, "grad_norm": 2.072838544845581, "learning_rate": 1.8084676962988212e-07, "loss": 0.4627, "step": 21868 }, { "epoch": 0.9167181916309485, "grad_norm": 2.4048731327056885, "learning_rate": 1.806658894218416e-07, "loss": 0.5287, "step": 21869 }, { "epoch": 0.9167601102459575, "grad_norm": 2.1249876022338867, "learning_rate": 1.8048509805156266e-07, "loss": 0.4593, "step": 21870 }, { "epoch": 0.9168020288609664, "grad_norm": 1.8048584461212158, "learning_rate": 1.8030439552237877e-07, "loss": 0.4868, "step": 21871 }, { "epoch": 0.9168439474759754, "grad_norm": 1.9841418266296387, "learning_rate": 1.8012378183762003e-07, "loss": 0.4519, "step": 21872 }, { "epoch": 0.9168858660909843, "grad_norm": 1.772665023803711, "learning_rate": 1.7994325700061655e-07, "loss": 0.4558, "step": 21873 }, { "epoch": 0.9169277847059933, "grad_norm": 1.933056116104126, "learning_rate": 1.7976282101469454e-07, "loss": 0.537, "step": 21874 }, { "epoch": 0.9169697033210022, "grad_norm": 2.106368064880371, "learning_rate": 1.795824738831825e-07, "loss": 0.5006, "step": 21875 }, { "epoch": 0.9170116219360113, "grad_norm": 1.9478347301483154, "learning_rate": 1.794022156094022e-07, "loss": 0.4916, "step": 21876 }, { "epoch": 0.9170535405510202, "grad_norm": 1.988649606704712, "learning_rate": 1.7922204619667872e-07, "loss": 0.4464, "step": 21877 }, { "epoch": 0.9170954591660292, "grad_norm": 2.1481785774230957, "learning_rate": 1.7904196564833165e-07, "loss": 0.4984, "step": 21878 }, { "epoch": 0.9171373777810381, "grad_norm": 1.8877698183059692, "learning_rate": 1.7886197396768224e-07, "loss": 0.5355, "step": 21879 }, { "epoch": 0.9171792963960471, "grad_norm": 2.0626158714294434, "learning_rate": 1.7868207115804613e-07, "loss": 0.4551, "step": 21880 }, { "epoch": 0.917221215011056, "grad_norm": 1.8892182111740112, "learning_rate": 1.7850225722274073e-07, "loss": 0.4401, "step": 21881 }, { "epoch": 0.917263133626065, "grad_norm": 3.355910539627075, "learning_rate": 1.7832253216508166e-07, "loss": 0.4931, "step": 21882 }, { "epoch": 0.917305052241074, "grad_norm": 1.9325122833251953, "learning_rate": 1.781428959883802e-07, "loss": 0.5078, "step": 21883 }, { "epoch": 0.917346970856083, "grad_norm": 4.31589937210083, "learning_rate": 1.7796334869594867e-07, "loss": 0.4768, "step": 21884 }, { "epoch": 0.9173888894710919, "grad_norm": 2.1867923736572266, "learning_rate": 1.7778389029109666e-07, "loss": 0.4619, "step": 21885 }, { "epoch": 0.9174308080861009, "grad_norm": 2.001014232635498, "learning_rate": 1.776045207771321e-07, "loss": 0.4731, "step": 21886 }, { "epoch": 0.9174727267011098, "grad_norm": 2.092175006866455, "learning_rate": 1.7742524015736118e-07, "loss": 0.5294, "step": 21887 }, { "epoch": 0.9175146453161187, "grad_norm": 1.8953348398208618, "learning_rate": 1.7724604843508963e-07, "loss": 0.4609, "step": 21888 }, { "epoch": 0.9175565639311277, "grad_norm": 1.9534740447998047, "learning_rate": 1.7706694561361925e-07, "loss": 0.5125, "step": 21889 }, { "epoch": 0.9175984825461366, "grad_norm": 2.0648319721221924, "learning_rate": 1.7688793169625185e-07, "loss": 0.4621, "step": 21890 }, { "epoch": 0.9176404011611456, "grad_norm": 1.9211461544036865, "learning_rate": 1.7670900668628867e-07, "loss": 0.412, "step": 21891 }, { "epoch": 0.9176823197761546, "grad_norm": 2.843254804611206, "learning_rate": 1.7653017058702593e-07, "loss": 0.4177, "step": 21892 }, { "epoch": 0.9177242383911636, "grad_norm": 2.026960849761963, "learning_rate": 1.7635142340176158e-07, "loss": 0.4811, "step": 21893 }, { "epoch": 0.9177661570061725, "grad_norm": 1.7860668897628784, "learning_rate": 1.7617276513379078e-07, "loss": 0.4341, "step": 21894 }, { "epoch": 0.9178080756211815, "grad_norm": 1.8818022012710571, "learning_rate": 1.7599419578640643e-07, "loss": 0.4653, "step": 21895 }, { "epoch": 0.9178499942361904, "grad_norm": 2.295227289199829, "learning_rate": 1.7581571536289977e-07, "loss": 0.492, "step": 21896 }, { "epoch": 0.9178919128511994, "grad_norm": 2.0878360271453857, "learning_rate": 1.7563732386656095e-07, "loss": 0.4844, "step": 21897 }, { "epoch": 0.9179338314662083, "grad_norm": 2.137617349624634, "learning_rate": 1.754590213006785e-07, "loss": 0.5076, "step": 21898 }, { "epoch": 0.9179757500812173, "grad_norm": 1.6919389963150024, "learning_rate": 1.7528080766853916e-07, "loss": 0.4867, "step": 21899 }, { "epoch": 0.9180176686962263, "grad_norm": 2.660494089126587, "learning_rate": 1.751026829734287e-07, "loss": 0.4386, "step": 21900 }, { "epoch": 0.9180595873112353, "grad_norm": 2.019798994064331, "learning_rate": 1.7492464721863e-07, "loss": 0.4998, "step": 21901 }, { "epoch": 0.9181015059262442, "grad_norm": 2.0422635078430176, "learning_rate": 1.7474670040742438e-07, "loss": 0.487, "step": 21902 }, { "epoch": 0.9181434245412532, "grad_norm": 2.0677759647369385, "learning_rate": 1.7456884254309303e-07, "loss": 0.5427, "step": 21903 }, { "epoch": 0.9181853431562621, "grad_norm": 1.8935754299163818, "learning_rate": 1.743910736289145e-07, "loss": 0.452, "step": 21904 }, { "epoch": 0.9182272617712711, "grad_norm": 2.016430377960205, "learning_rate": 1.7421339366816502e-07, "loss": 0.4792, "step": 21905 }, { "epoch": 0.91826918038628, "grad_norm": 4.4176506996154785, "learning_rate": 1.7403580266412035e-07, "loss": 0.5139, "step": 21906 }, { "epoch": 0.918311099001289, "grad_norm": 2.179831027984619, "learning_rate": 1.7385830062005337e-07, "loss": 0.4712, "step": 21907 }, { "epoch": 0.918353017616298, "grad_norm": 2.074669599533081, "learning_rate": 1.7368088753923652e-07, "loss": 0.4498, "step": 21908 }, { "epoch": 0.918394936231307, "grad_norm": 1.968383550643921, "learning_rate": 1.7350356342494102e-07, "loss": 0.4812, "step": 21909 }, { "epoch": 0.9184368548463159, "grad_norm": 1.895100712776184, "learning_rate": 1.7332632828043482e-07, "loss": 0.4851, "step": 21910 }, { "epoch": 0.9184787734613249, "grad_norm": 2.6939969062805176, "learning_rate": 1.731491821089848e-07, "loss": 0.4774, "step": 21911 }, { "epoch": 0.9185206920763338, "grad_norm": 2.2670812606811523, "learning_rate": 1.7297212491385717e-07, "loss": 0.4997, "step": 21912 }, { "epoch": 0.9185626106913427, "grad_norm": 1.8186002969741821, "learning_rate": 1.7279515669831492e-07, "loss": 0.4807, "step": 21913 }, { "epoch": 0.9186045293063517, "grad_norm": 1.8314344882965088, "learning_rate": 1.7261827746562043e-07, "loss": 0.4576, "step": 21914 }, { "epoch": 0.9186464479213606, "grad_norm": 1.9091943502426147, "learning_rate": 1.7244148721903443e-07, "loss": 0.5124, "step": 21915 }, { "epoch": 0.9186883665363696, "grad_norm": 2.4221253395080566, "learning_rate": 1.7226478596181595e-07, "loss": 0.4975, "step": 21916 }, { "epoch": 0.9187302851513786, "grad_norm": 1.7626028060913086, "learning_rate": 1.7208817369722132e-07, "loss": 0.4231, "step": 21917 }, { "epoch": 0.9187722037663876, "grad_norm": 1.8407115936279297, "learning_rate": 1.7191165042850734e-07, "loss": 0.4802, "step": 21918 }, { "epoch": 0.9188141223813965, "grad_norm": 1.9385608434677124, "learning_rate": 1.7173521615892753e-07, "loss": 0.5007, "step": 21919 }, { "epoch": 0.9188560409964055, "grad_norm": 1.8961615562438965, "learning_rate": 1.7155887089173372e-07, "loss": 0.4886, "step": 21920 }, { "epoch": 0.9188979596114144, "grad_norm": 1.8192729949951172, "learning_rate": 1.7138261463017724e-07, "loss": 0.4051, "step": 21921 }, { "epoch": 0.9189398782264234, "grad_norm": 1.6952868700027466, "learning_rate": 1.712064473775077e-07, "loss": 0.4798, "step": 21922 }, { "epoch": 0.9189817968414323, "grad_norm": 2.1103408336639404, "learning_rate": 1.7103036913697024e-07, "loss": 0.4118, "step": 21923 }, { "epoch": 0.9190237154564413, "grad_norm": 1.9664175510406494, "learning_rate": 1.7085437991181342e-07, "loss": 0.4486, "step": 21924 }, { "epoch": 0.9190656340714503, "grad_norm": 1.9108694791793823, "learning_rate": 1.7067847970527908e-07, "loss": 0.5101, "step": 21925 }, { "epoch": 0.9191075526864593, "grad_norm": 1.809901237487793, "learning_rate": 1.705026685206107e-07, "loss": 0.5062, "step": 21926 }, { "epoch": 0.9191494713014682, "grad_norm": 2.055570125579834, "learning_rate": 1.7032694636104962e-07, "loss": 0.4745, "step": 21927 }, { "epoch": 0.9191913899164772, "grad_norm": 2.07936429977417, "learning_rate": 1.701513132298338e-07, "loss": 0.497, "step": 21928 }, { "epoch": 0.9192333085314861, "grad_norm": 1.9480518102645874, "learning_rate": 1.699757691302012e-07, "loss": 0.4544, "step": 21929 }, { "epoch": 0.9192752271464951, "grad_norm": 2.144789218902588, "learning_rate": 1.6980031406538922e-07, "loss": 0.521, "step": 21930 }, { "epoch": 0.919317145761504, "grad_norm": 1.8258213996887207, "learning_rate": 1.6962494803862971e-07, "loss": 0.4703, "step": 21931 }, { "epoch": 0.919359064376513, "grad_norm": 1.884203553199768, "learning_rate": 1.6944967105315735e-07, "loss": 0.4571, "step": 21932 }, { "epoch": 0.919400982991522, "grad_norm": 1.869002103805542, "learning_rate": 1.692744831122023e-07, "loss": 0.4923, "step": 21933 }, { "epoch": 0.919442901606531, "grad_norm": 1.9513403177261353, "learning_rate": 1.6909938421899308e-07, "loss": 0.4836, "step": 21934 }, { "epoch": 0.9194848202215399, "grad_norm": 2.241919994354248, "learning_rate": 1.6892437437675823e-07, "loss": 0.4699, "step": 21935 }, { "epoch": 0.9195267388365489, "grad_norm": 2.1757380962371826, "learning_rate": 1.6874945358872464e-07, "loss": 0.5175, "step": 21936 }, { "epoch": 0.9195686574515578, "grad_norm": 2.1854350566864014, "learning_rate": 1.6857462185811525e-07, "loss": 0.4957, "step": 21937 }, { "epoch": 0.9196105760665667, "grad_norm": 2.190063953399658, "learning_rate": 1.6839987918815303e-07, "loss": 0.4913, "step": 21938 }, { "epoch": 0.9196524946815757, "grad_norm": 1.8192130327224731, "learning_rate": 1.6822522558206046e-07, "loss": 0.4657, "step": 21939 }, { "epoch": 0.9196944132965846, "grad_norm": 1.9951692819595337, "learning_rate": 1.68050661043056e-07, "loss": 0.4636, "step": 21940 }, { "epoch": 0.9197363319115937, "grad_norm": 2.0530335903167725, "learning_rate": 1.6787618557435769e-07, "loss": 0.4736, "step": 21941 }, { "epoch": 0.9197782505266026, "grad_norm": 1.7825427055358887, "learning_rate": 1.6770179917918128e-07, "loss": 0.4309, "step": 21942 }, { "epoch": 0.9198201691416116, "grad_norm": 1.8368573188781738, "learning_rate": 1.6752750186074197e-07, "loss": 0.4666, "step": 21943 }, { "epoch": 0.9198620877566205, "grad_norm": 1.6061652898788452, "learning_rate": 1.6735329362225217e-07, "loss": 0.453, "step": 21944 }, { "epoch": 0.9199040063716295, "grad_norm": 1.9234259128570557, "learning_rate": 1.6717917446692378e-07, "loss": 0.5111, "step": 21945 }, { "epoch": 0.9199459249866384, "grad_norm": 1.8377642631530762, "learning_rate": 1.6700514439796588e-07, "loss": 0.4432, "step": 21946 }, { "epoch": 0.9199878436016474, "grad_norm": 1.7920286655426025, "learning_rate": 1.6683120341858705e-07, "loss": 0.4862, "step": 21947 }, { "epoch": 0.9200297622166563, "grad_norm": 1.779144525527954, "learning_rate": 1.6665735153199303e-07, "loss": 0.4168, "step": 21948 }, { "epoch": 0.9200716808316654, "grad_norm": 1.8509941101074219, "learning_rate": 1.6648358874138847e-07, "loss": 0.4192, "step": 21949 }, { "epoch": 0.9201135994466743, "grad_norm": 1.918771743774414, "learning_rate": 1.6630991504997806e-07, "loss": 0.4517, "step": 21950 }, { "epoch": 0.9201555180616833, "grad_norm": 2.1029791831970215, "learning_rate": 1.6613633046096145e-07, "loss": 0.4296, "step": 21951 }, { "epoch": 0.9201974366766922, "grad_norm": 6.056175231933594, "learning_rate": 1.659628349775383e-07, "loss": 0.4577, "step": 21952 }, { "epoch": 0.9202393552917012, "grad_norm": 1.8759498596191406, "learning_rate": 1.6578942860290825e-07, "loss": 0.4194, "step": 21953 }, { "epoch": 0.9202812739067101, "grad_norm": 2.527822256088257, "learning_rate": 1.6561611134026655e-07, "loss": 0.4392, "step": 21954 }, { "epoch": 0.9203231925217191, "grad_norm": 1.8209364414215088, "learning_rate": 1.6544288319280842e-07, "loss": 0.4853, "step": 21955 }, { "epoch": 0.920365111136728, "grad_norm": 2.974937677383423, "learning_rate": 1.6526974416372744e-07, "loss": 0.4361, "step": 21956 }, { "epoch": 0.920407029751737, "grad_norm": 1.8277013301849365, "learning_rate": 1.6509669425621488e-07, "loss": 0.4792, "step": 21957 }, { "epoch": 0.920448948366746, "grad_norm": 1.901731014251709, "learning_rate": 1.649237334734599e-07, "loss": 0.4673, "step": 21958 }, { "epoch": 0.920490866981755, "grad_norm": 1.8003696203231812, "learning_rate": 1.6475086181865275e-07, "loss": 0.4239, "step": 21959 }, { "epoch": 0.9205327855967639, "grad_norm": 1.7433873414993286, "learning_rate": 1.6457807929497915e-07, "loss": 0.5519, "step": 21960 }, { "epoch": 0.9205747042117729, "grad_norm": 1.8700960874557495, "learning_rate": 1.644053859056227e-07, "loss": 0.4605, "step": 21961 }, { "epoch": 0.9206166228267818, "grad_norm": 2.060642957687378, "learning_rate": 1.642327816537692e-07, "loss": 0.4645, "step": 21962 }, { "epoch": 0.9206585414417907, "grad_norm": 1.9606887102127075, "learning_rate": 1.6406026654259778e-07, "loss": 0.5061, "step": 21963 }, { "epoch": 0.9207004600567997, "grad_norm": 1.9505914449691772, "learning_rate": 1.638878405752903e-07, "loss": 0.5079, "step": 21964 }, { "epoch": 0.9207423786718086, "grad_norm": 1.807307243347168, "learning_rate": 1.6371550375502532e-07, "loss": 0.4213, "step": 21965 }, { "epoch": 0.9207842972868177, "grad_norm": 2.273468255996704, "learning_rate": 1.6354325608497869e-07, "loss": 0.4682, "step": 21966 }, { "epoch": 0.9208262159018266, "grad_norm": 1.7188531160354614, "learning_rate": 1.633710975683256e-07, "loss": 0.4557, "step": 21967 }, { "epoch": 0.9208681345168356, "grad_norm": 1.7926615476608276, "learning_rate": 1.6319902820824075e-07, "loss": 0.4096, "step": 21968 }, { "epoch": 0.9209100531318445, "grad_norm": 1.715788722038269, "learning_rate": 1.6302704800789494e-07, "loss": 0.4143, "step": 21969 }, { "epoch": 0.9209519717468535, "grad_norm": 2.3828237056732178, "learning_rate": 1.6285515697045785e-07, "loss": 0.4801, "step": 21970 }, { "epoch": 0.9209938903618624, "grad_norm": 1.5650707483291626, "learning_rate": 1.6268335509909972e-07, "loss": 0.4436, "step": 21971 }, { "epoch": 0.9210358089768714, "grad_norm": 1.7505074739456177, "learning_rate": 1.625116423969858e-07, "loss": 0.4743, "step": 21972 }, { "epoch": 0.9210777275918803, "grad_norm": 1.8851855993270874, "learning_rate": 1.623400188672819e-07, "loss": 0.4891, "step": 21973 }, { "epoch": 0.9211196462068894, "grad_norm": 1.7794138193130493, "learning_rate": 1.621684845131527e-07, "loss": 0.48, "step": 21974 }, { "epoch": 0.9211615648218983, "grad_norm": 1.824893593788147, "learning_rate": 1.6199703933775902e-07, "loss": 0.5066, "step": 21975 }, { "epoch": 0.9212034834369073, "grad_norm": 2.054368495941162, "learning_rate": 1.6182568334426108e-07, "loss": 0.4866, "step": 21976 }, { "epoch": 0.9212454020519162, "grad_norm": 1.682499885559082, "learning_rate": 1.6165441653581915e-07, "loss": 0.5341, "step": 21977 }, { "epoch": 0.9212873206669252, "grad_norm": 1.8991771936416626, "learning_rate": 1.614832389155885e-07, "loss": 0.4393, "step": 21978 }, { "epoch": 0.9213292392819341, "grad_norm": 2.15661358833313, "learning_rate": 1.6131215048672543e-07, "loss": 0.4609, "step": 21979 }, { "epoch": 0.9213711578969431, "grad_norm": 1.9160996675491333, "learning_rate": 1.6114115125238415e-07, "loss": 0.4589, "step": 21980 }, { "epoch": 0.921413076511952, "grad_norm": 1.8471266031265259, "learning_rate": 1.6097024121571546e-07, "loss": 0.493, "step": 21981 }, { "epoch": 0.921454995126961, "grad_norm": 1.7326838970184326, "learning_rate": 1.6079942037987074e-07, "loss": 0.463, "step": 21982 }, { "epoch": 0.92149691374197, "grad_norm": 1.9833253622055054, "learning_rate": 1.6062868874799908e-07, "loss": 0.5035, "step": 21983 }, { "epoch": 0.921538832356979, "grad_norm": 1.733850359916687, "learning_rate": 1.604580463232469e-07, "loss": 0.4285, "step": 21984 }, { "epoch": 0.9215807509719879, "grad_norm": 1.9960520267486572, "learning_rate": 1.6028749310876e-07, "loss": 0.4659, "step": 21985 }, { "epoch": 0.9216226695869969, "grad_norm": 2.321087121963501, "learning_rate": 1.6011702910768312e-07, "loss": 0.4479, "step": 21986 }, { "epoch": 0.9216645882020058, "grad_norm": 1.7602729797363281, "learning_rate": 1.5994665432315759e-07, "loss": 0.4675, "step": 21987 }, { "epoch": 0.9217065068170147, "grad_norm": 1.6437817811965942, "learning_rate": 1.5977636875832424e-07, "loss": 0.4368, "step": 21988 }, { "epoch": 0.9217484254320237, "grad_norm": 1.811158537864685, "learning_rate": 1.596061724163228e-07, "loss": 0.4263, "step": 21989 }, { "epoch": 0.9217903440470326, "grad_norm": 1.7434452772140503, "learning_rate": 1.594360653002891e-07, "loss": 0.4906, "step": 21990 }, { "epoch": 0.9218322626620417, "grad_norm": 2.4589195251464844, "learning_rate": 1.592660474133595e-07, "loss": 0.514, "step": 21991 }, { "epoch": 0.9218741812770506, "grad_norm": 1.9473202228546143, "learning_rate": 1.590961187586687e-07, "loss": 0.4514, "step": 21992 }, { "epoch": 0.9219160998920596, "grad_norm": 3.507120132446289, "learning_rate": 1.5892627933934812e-07, "loss": 0.4674, "step": 21993 }, { "epoch": 0.9219580185070685, "grad_norm": 2.3741655349731445, "learning_rate": 1.5875652915852913e-07, "loss": 0.4753, "step": 21994 }, { "epoch": 0.9219999371220775, "grad_norm": 4.30784273147583, "learning_rate": 1.5858686821934145e-07, "loss": 0.4443, "step": 21995 }, { "epoch": 0.9220418557370864, "grad_norm": 1.9706283807754517, "learning_rate": 1.584172965249109e-07, "loss": 0.5214, "step": 21996 }, { "epoch": 0.9220837743520954, "grad_norm": 1.6087300777435303, "learning_rate": 1.58247814078365e-07, "loss": 0.4238, "step": 21997 }, { "epoch": 0.9221256929671043, "grad_norm": 2.7285478115081787, "learning_rate": 1.5807842088282677e-07, "loss": 0.4711, "step": 21998 }, { "epoch": 0.9221676115821134, "grad_norm": 1.7338335514068604, "learning_rate": 1.5790911694141874e-07, "loss": 0.4589, "step": 21999 }, { "epoch": 0.9222095301971223, "grad_norm": 2.026001214981079, "learning_rate": 1.5773990225726233e-07, "loss": 0.478, "step": 22000 }, { "epoch": 0.9222514488121313, "grad_norm": 2.2398128509521484, "learning_rate": 1.575707768334772e-07, "loss": 0.4753, "step": 22001 }, { "epoch": 0.9222933674271402, "grad_norm": 2.0948517322540283, "learning_rate": 1.5740174067317925e-07, "loss": 0.4404, "step": 22002 }, { "epoch": 0.9223352860421492, "grad_norm": 1.933497428894043, "learning_rate": 1.5723279377948597e-07, "loss": 0.4888, "step": 22003 }, { "epoch": 0.9223772046571581, "grad_norm": 2.212697744369507, "learning_rate": 1.5706393615551153e-07, "loss": 0.4411, "step": 22004 }, { "epoch": 0.9224191232721671, "grad_norm": 2.6365084648132324, "learning_rate": 1.5689516780436787e-07, "loss": 0.5135, "step": 22005 }, { "epoch": 0.922461041887176, "grad_norm": 2.5048539638519287, "learning_rate": 1.5672648872916696e-07, "loss": 0.461, "step": 22006 }, { "epoch": 0.922502960502185, "grad_norm": 1.8605157136917114, "learning_rate": 1.5655789893301798e-07, "loss": 0.469, "step": 22007 }, { "epoch": 0.922544879117194, "grad_norm": 1.9182451963424683, "learning_rate": 1.5638939841902733e-07, "loss": 0.4132, "step": 22008 }, { "epoch": 0.922586797732203, "grad_norm": 3.208561420440674, "learning_rate": 1.5622098719030198e-07, "loss": 0.4818, "step": 22009 }, { "epoch": 0.9226287163472119, "grad_norm": 1.9432083368301392, "learning_rate": 1.5605266524994667e-07, "loss": 0.4287, "step": 22010 }, { "epoch": 0.9226706349622209, "grad_norm": 2.12825345993042, "learning_rate": 1.558844326010639e-07, "loss": 0.4733, "step": 22011 }, { "epoch": 0.9227125535772298, "grad_norm": 2.0424063205718994, "learning_rate": 1.5571628924675452e-07, "loss": 0.4813, "step": 22012 }, { "epoch": 0.9227544721922387, "grad_norm": 2.3333675861358643, "learning_rate": 1.5554823519011886e-07, "loss": 0.5003, "step": 22013 }, { "epoch": 0.9227963908072477, "grad_norm": 1.966382384300232, "learning_rate": 1.5538027043425387e-07, "loss": 0.4551, "step": 22014 }, { "epoch": 0.9228383094222566, "grad_norm": 2.2040462493896484, "learning_rate": 1.552123949822565e-07, "loss": 0.4665, "step": 22015 }, { "epoch": 0.9228802280372657, "grad_norm": 2.49749755859375, "learning_rate": 1.550446088372204e-07, "loss": 0.5098, "step": 22016 }, { "epoch": 0.9229221466522746, "grad_norm": 1.7630913257598877, "learning_rate": 1.5487691200223976e-07, "loss": 0.4805, "step": 22017 }, { "epoch": 0.9229640652672836, "grad_norm": 1.8652305603027344, "learning_rate": 1.5470930448040432e-07, "loss": 0.5099, "step": 22018 }, { "epoch": 0.9230059838822925, "grad_norm": 1.7494375705718994, "learning_rate": 1.54541786274805e-07, "loss": 0.4703, "step": 22019 }, { "epoch": 0.9230479024973015, "grad_norm": 1.9336462020874023, "learning_rate": 1.543743573885287e-07, "loss": 0.4973, "step": 22020 }, { "epoch": 0.9230898211123104, "grad_norm": 4.585105895996094, "learning_rate": 1.5420701782466297e-07, "loss": 0.4464, "step": 22021 }, { "epoch": 0.9231317397273194, "grad_norm": 10.699079513549805, "learning_rate": 1.5403976758629092e-07, "loss": 0.491, "step": 22022 }, { "epoch": 0.9231736583423283, "grad_norm": 1.9409847259521484, "learning_rate": 1.5387260667649617e-07, "loss": 0.5133, "step": 22023 }, { "epoch": 0.9232155769573374, "grad_norm": 2.2191524505615234, "learning_rate": 1.5370553509836127e-07, "loss": 0.4975, "step": 22024 }, { "epoch": 0.9232574955723463, "grad_norm": 1.7200289964675903, "learning_rate": 1.5353855285496432e-07, "loss": 0.4448, "step": 22025 }, { "epoch": 0.9232994141873553, "grad_norm": 1.6828488111495972, "learning_rate": 1.5337165994938509e-07, "loss": 0.4847, "step": 22026 }, { "epoch": 0.9233413328023642, "grad_norm": 1.9662359952926636, "learning_rate": 1.5320485638469884e-07, "loss": 0.4875, "step": 22027 }, { "epoch": 0.9233832514173732, "grad_norm": 2.4275057315826416, "learning_rate": 1.5303814216397982e-07, "loss": 0.514, "step": 22028 }, { "epoch": 0.9234251700323821, "grad_norm": 2.1926984786987305, "learning_rate": 1.5287151729030169e-07, "loss": 0.4994, "step": 22029 }, { "epoch": 0.9234670886473911, "grad_norm": 2.0549213886260986, "learning_rate": 1.5270498176673753e-07, "loss": 0.4884, "step": 22030 }, { "epoch": 0.9235090072624, "grad_norm": 1.7252082824707031, "learning_rate": 1.5253853559635433e-07, "loss": 0.4665, "step": 22031 }, { "epoch": 0.9235509258774091, "grad_norm": 2.07033634185791, "learning_rate": 1.5237217878222243e-07, "loss": 0.5308, "step": 22032 }, { "epoch": 0.923592844492418, "grad_norm": 2.127084970474243, "learning_rate": 1.5220591132740825e-07, "loss": 0.4771, "step": 22033 }, { "epoch": 0.923634763107427, "grad_norm": 2.372227430343628, "learning_rate": 1.5203973323497545e-07, "loss": 0.489, "step": 22034 }, { "epoch": 0.9236766817224359, "grad_norm": 1.9918694496154785, "learning_rate": 1.518736445079888e-07, "loss": 0.4713, "step": 22035 }, { "epoch": 0.9237186003374449, "grad_norm": 2.1103086471557617, "learning_rate": 1.517076451495092e-07, "loss": 0.4869, "step": 22036 }, { "epoch": 0.9237605189524538, "grad_norm": 2.051225423812866, "learning_rate": 1.515417351625964e-07, "loss": 0.4417, "step": 22037 }, { "epoch": 0.9238024375674627, "grad_norm": 1.8734246492385864, "learning_rate": 1.5137591455030854e-07, "loss": 0.4653, "step": 22038 }, { "epoch": 0.9238443561824717, "grad_norm": 1.897574543952942, "learning_rate": 1.5121018331570315e-07, "loss": 0.4893, "step": 22039 }, { "epoch": 0.9238862747974806, "grad_norm": 1.6233489513397217, "learning_rate": 1.5104454146183446e-07, "loss": 0.4602, "step": 22040 }, { "epoch": 0.9239281934124897, "grad_norm": 2.1048247814178467, "learning_rate": 1.5087898899175556e-07, "loss": 0.4529, "step": 22041 }, { "epoch": 0.9239701120274986, "grad_norm": 1.928188681602478, "learning_rate": 1.507135259085202e-07, "loss": 0.4535, "step": 22042 }, { "epoch": 0.9240120306425076, "grad_norm": 1.8264740705490112, "learning_rate": 1.5054815221517583e-07, "loss": 0.4443, "step": 22043 }, { "epoch": 0.9240539492575165, "grad_norm": 1.979318618774414, "learning_rate": 1.5038286791477284e-07, "loss": 0.4615, "step": 22044 }, { "epoch": 0.9240958678725255, "grad_norm": 2.4310364723205566, "learning_rate": 1.5021767301035716e-07, "loss": 0.4666, "step": 22045 }, { "epoch": 0.9241377864875344, "grad_norm": 1.7932888269424438, "learning_rate": 1.500525675049741e-07, "loss": 0.4645, "step": 22046 }, { "epoch": 0.9241797051025434, "grad_norm": 1.9441649913787842, "learning_rate": 1.4988755140166677e-07, "loss": 0.4647, "step": 22047 }, { "epoch": 0.9242216237175523, "grad_norm": 1.6571779251098633, "learning_rate": 1.4972262470347776e-07, "loss": 0.4555, "step": 22048 }, { "epoch": 0.9242635423325614, "grad_norm": 1.893912434577942, "learning_rate": 1.4955778741344627e-07, "loss": 0.4941, "step": 22049 }, { "epoch": 0.9243054609475703, "grad_norm": 2.0651259422302246, "learning_rate": 1.493930395346116e-07, "loss": 0.4805, "step": 22050 }, { "epoch": 0.9243473795625793, "grad_norm": 2.266855239868164, "learning_rate": 1.4922838107001068e-07, "loss": 0.564, "step": 22051 }, { "epoch": 0.9243892981775882, "grad_norm": 2.0969197750091553, "learning_rate": 1.4906381202267838e-07, "loss": 0.4914, "step": 22052 }, { "epoch": 0.9244312167925972, "grad_norm": 1.9909029006958008, "learning_rate": 1.488993323956489e-07, "loss": 0.4727, "step": 22053 }, { "epoch": 0.9244731354076061, "grad_norm": 2.1686909198760986, "learning_rate": 1.4873494219195428e-07, "loss": 0.4771, "step": 22054 }, { "epoch": 0.9245150540226151, "grad_norm": 1.8572285175323486, "learning_rate": 1.485706414146232e-07, "loss": 0.4426, "step": 22055 }, { "epoch": 0.924556972637624, "grad_norm": 1.8569164276123047, "learning_rate": 1.4840643006668543e-07, "loss": 0.455, "step": 22056 }, { "epoch": 0.9245988912526331, "grad_norm": 2.019577980041504, "learning_rate": 1.4824230815116914e-07, "loss": 0.4633, "step": 22057 }, { "epoch": 0.924640809867642, "grad_norm": 2.104729652404785, "learning_rate": 1.4807827567109746e-07, "loss": 0.4221, "step": 22058 }, { "epoch": 0.924682728482651, "grad_norm": 1.9842712879180908, "learning_rate": 1.4791433262949516e-07, "loss": 0.5102, "step": 22059 }, { "epoch": 0.9247246470976599, "grad_norm": 1.7894293069839478, "learning_rate": 1.477504790293849e-07, "loss": 0.4668, "step": 22060 }, { "epoch": 0.9247665657126689, "grad_norm": 1.6855945587158203, "learning_rate": 1.475867148737864e-07, "loss": 0.4514, "step": 22061 }, { "epoch": 0.9248084843276778, "grad_norm": 1.9131015539169312, "learning_rate": 1.474230401657184e-07, "loss": 0.4958, "step": 22062 }, { "epoch": 0.9248504029426867, "grad_norm": 1.8283756971359253, "learning_rate": 1.4725945490819848e-07, "loss": 0.4766, "step": 22063 }, { "epoch": 0.9248923215576957, "grad_norm": 2.2231297492980957, "learning_rate": 1.4709595910424145e-07, "loss": 0.5122, "step": 22064 }, { "epoch": 0.9249342401727046, "grad_norm": 1.7471888065338135, "learning_rate": 1.4693255275686158e-07, "loss": 0.4522, "step": 22065 }, { "epoch": 0.9249761587877137, "grad_norm": 2.799304246902466, "learning_rate": 1.4676923586907144e-07, "loss": 0.4895, "step": 22066 }, { "epoch": 0.9250180774027226, "grad_norm": 2.216782331466675, "learning_rate": 1.466060084438803e-07, "loss": 0.4759, "step": 22067 }, { "epoch": 0.9250599960177316, "grad_norm": 1.9291127920150757, "learning_rate": 1.464428704842974e-07, "loss": 0.4411, "step": 22068 }, { "epoch": 0.9251019146327405, "grad_norm": 2.000542402267456, "learning_rate": 1.4627982199333147e-07, "loss": 0.4713, "step": 22069 }, { "epoch": 0.9251438332477495, "grad_norm": 2.0733067989349365, "learning_rate": 1.461168629739862e-07, "loss": 0.4387, "step": 22070 }, { "epoch": 0.9251857518627584, "grad_norm": 2.9637694358825684, "learning_rate": 1.459539934292664e-07, "loss": 0.5065, "step": 22071 }, { "epoch": 0.9252276704777674, "grad_norm": 2.5405333042144775, "learning_rate": 1.4579121336217473e-07, "loss": 0.4795, "step": 22072 }, { "epoch": 0.9252695890927763, "grad_norm": 1.8403356075286865, "learning_rate": 1.4562852277571095e-07, "loss": 0.4684, "step": 22073 }, { "epoch": 0.9253115077077854, "grad_norm": 2.0800423622131348, "learning_rate": 1.454659216728743e-07, "loss": 0.4814, "step": 22074 }, { "epoch": 0.9253534263227943, "grad_norm": 2.022451877593994, "learning_rate": 1.4530341005666247e-07, "loss": 0.4537, "step": 22075 }, { "epoch": 0.9253953449378033, "grad_norm": 2.744051456451416, "learning_rate": 1.4514098793007026e-07, "loss": 0.49, "step": 22076 }, { "epoch": 0.9254372635528122, "grad_norm": 3.8099615573883057, "learning_rate": 1.4497865529609245e-07, "loss": 0.4622, "step": 22077 }, { "epoch": 0.9254791821678212, "grad_norm": 1.889776349067688, "learning_rate": 1.448164121577217e-07, "loss": 0.4295, "step": 22078 }, { "epoch": 0.9255211007828301, "grad_norm": 1.6755300760269165, "learning_rate": 1.4465425851794834e-07, "loss": 0.4092, "step": 22079 }, { "epoch": 0.9255630193978391, "grad_norm": 1.9266753196716309, "learning_rate": 1.444921943797606e-07, "loss": 0.4798, "step": 22080 }, { "epoch": 0.925604938012848, "grad_norm": 3.1131958961486816, "learning_rate": 1.443302197461477e-07, "loss": 0.48, "step": 22081 }, { "epoch": 0.9256468566278571, "grad_norm": 1.8607908487319946, "learning_rate": 1.441683346200945e-07, "loss": 0.4551, "step": 22082 }, { "epoch": 0.925688775242866, "grad_norm": 2.215562105178833, "learning_rate": 1.4400653900458416e-07, "loss": 0.4629, "step": 22083 }, { "epoch": 0.925730693857875, "grad_norm": 1.521142601966858, "learning_rate": 1.4384483290260098e-07, "loss": 0.4446, "step": 22084 }, { "epoch": 0.9257726124728839, "grad_norm": 2.0279805660247803, "learning_rate": 1.4368321631712368e-07, "loss": 0.4962, "step": 22085 }, { "epoch": 0.9258145310878929, "grad_norm": 1.661797285079956, "learning_rate": 1.435216892511332e-07, "loss": 0.4991, "step": 22086 }, { "epoch": 0.9258564497029018, "grad_norm": 1.7016422748565674, "learning_rate": 1.4336025170760713e-07, "loss": 0.4757, "step": 22087 }, { "epoch": 0.9258983683179107, "grad_norm": 1.9175280332565308, "learning_rate": 1.4319890368951984e-07, "loss": 0.5022, "step": 22088 }, { "epoch": 0.9259402869329197, "grad_norm": 1.6778169870376587, "learning_rate": 1.4303764519984665e-07, "loss": 0.438, "step": 22089 }, { "epoch": 0.9259822055479286, "grad_norm": 1.6053459644317627, "learning_rate": 1.428764762415602e-07, "loss": 0.4576, "step": 22090 }, { "epoch": 0.9260241241629377, "grad_norm": 2.1711504459381104, "learning_rate": 1.4271539681763148e-07, "loss": 0.4391, "step": 22091 }, { "epoch": 0.9260660427779466, "grad_norm": 2.342050075531006, "learning_rate": 1.425544069310286e-07, "loss": 0.5014, "step": 22092 }, { "epoch": 0.9261079613929556, "grad_norm": 1.9237531423568726, "learning_rate": 1.4239350658472096e-07, "loss": 0.4945, "step": 22093 }, { "epoch": 0.9261498800079645, "grad_norm": 2.005350112915039, "learning_rate": 1.4223269578167277e-07, "loss": 0.4462, "step": 22094 }, { "epoch": 0.9261917986229735, "grad_norm": 1.7565048933029175, "learning_rate": 1.420719745248489e-07, "loss": 0.4427, "step": 22095 }, { "epoch": 0.9262337172379824, "grad_norm": 2.09305477142334, "learning_rate": 1.4191134281721308e-07, "loss": 0.4537, "step": 22096 }, { "epoch": 0.9262756358529914, "grad_norm": 2.03770112991333, "learning_rate": 1.417508006617252e-07, "loss": 0.4607, "step": 22097 }, { "epoch": 0.9263175544680003, "grad_norm": 1.7804901599884033, "learning_rate": 1.415903480613451e-07, "loss": 0.5213, "step": 22098 }, { "epoch": 0.9263594730830094, "grad_norm": 1.855928897857666, "learning_rate": 1.4142998501903038e-07, "loss": 0.445, "step": 22099 }, { "epoch": 0.9264013916980183, "grad_norm": 2.146279811859131, "learning_rate": 1.4126971153773704e-07, "loss": 0.4713, "step": 22100 }, { "epoch": 0.9264433103130273, "grad_norm": 2.005216360092163, "learning_rate": 1.4110952762041995e-07, "loss": 0.4842, "step": 22101 }, { "epoch": 0.9264852289280362, "grad_norm": 1.9990288019180298, "learning_rate": 1.409494332700312e-07, "loss": 0.4846, "step": 22102 }, { "epoch": 0.9265271475430452, "grad_norm": 1.874213457107544, "learning_rate": 1.407894284895217e-07, "loss": 0.4774, "step": 22103 }, { "epoch": 0.9265690661580541, "grad_norm": 1.669797420501709, "learning_rate": 1.406295132818425e-07, "loss": 0.4817, "step": 22104 }, { "epoch": 0.9266109847730631, "grad_norm": 2.922478199005127, "learning_rate": 1.4046968764993951e-07, "loss": 0.5487, "step": 22105 }, { "epoch": 0.926652903388072, "grad_norm": 1.930740475654602, "learning_rate": 1.403099515967593e-07, "loss": 0.4612, "step": 22106 }, { "epoch": 0.9266948220030811, "grad_norm": 1.7237884998321533, "learning_rate": 1.4015030512524787e-07, "loss": 0.42, "step": 22107 }, { "epoch": 0.92673674061809, "grad_norm": 2.5037646293640137, "learning_rate": 1.399907482383467e-07, "loss": 0.4073, "step": 22108 }, { "epoch": 0.926778659233099, "grad_norm": 2.1223742961883545, "learning_rate": 1.3983128093899734e-07, "loss": 0.4482, "step": 22109 }, { "epoch": 0.9268205778481079, "grad_norm": 2.114854335784912, "learning_rate": 1.3967190323013912e-07, "loss": 0.4459, "step": 22110 }, { "epoch": 0.9268624964631169, "grad_norm": 1.6747517585754395, "learning_rate": 1.3951261511471025e-07, "loss": 0.4963, "step": 22111 }, { "epoch": 0.9269044150781258, "grad_norm": 1.8391450643539429, "learning_rate": 1.3935341659564671e-07, "loss": 0.4754, "step": 22112 }, { "epoch": 0.9269463336931347, "grad_norm": 1.617884635925293, "learning_rate": 1.3919430767588393e-07, "loss": 0.4367, "step": 22113 }, { "epoch": 0.9269882523081437, "grad_norm": 1.97555673122406, "learning_rate": 1.3903528835835345e-07, "loss": 0.4647, "step": 22114 }, { "epoch": 0.9270301709231527, "grad_norm": 1.7391008138656616, "learning_rate": 1.388763586459879e-07, "loss": 0.4649, "step": 22115 }, { "epoch": 0.9270720895381617, "grad_norm": 3.5431649684906006, "learning_rate": 1.3871751854171612e-07, "loss": 0.5258, "step": 22116 }, { "epoch": 0.9271140081531706, "grad_norm": 3.6264195442199707, "learning_rate": 1.3855876804846624e-07, "loss": 0.4992, "step": 22117 }, { "epoch": 0.9271559267681796, "grad_norm": 1.6900442838668823, "learning_rate": 1.3840010716916485e-07, "loss": 0.4325, "step": 22118 }, { "epoch": 0.9271978453831885, "grad_norm": 1.8950104713439941, "learning_rate": 1.3824153590673684e-07, "loss": 0.4575, "step": 22119 }, { "epoch": 0.9272397639981975, "grad_norm": 1.941360592842102, "learning_rate": 1.3808305426410484e-07, "loss": 0.4367, "step": 22120 }, { "epoch": 0.9272816826132064, "grad_norm": 1.7895407676696777, "learning_rate": 1.379246622441899e-07, "loss": 0.486, "step": 22121 }, { "epoch": 0.9273236012282154, "grad_norm": 1.7028405666351318, "learning_rate": 1.3776635984991293e-07, "loss": 0.4429, "step": 22122 }, { "epoch": 0.9273655198432244, "grad_norm": 1.6475043296813965, "learning_rate": 1.3760814708419058e-07, "loss": 0.4938, "step": 22123 }, { "epoch": 0.9274074384582334, "grad_norm": 1.7425987720489502, "learning_rate": 1.374500239499399e-07, "loss": 0.4677, "step": 22124 }, { "epoch": 0.9274493570732423, "grad_norm": 3.0302939414978027, "learning_rate": 1.3729199045007636e-07, "loss": 0.5044, "step": 22125 }, { "epoch": 0.9274912756882513, "grad_norm": 1.9948967695236206, "learning_rate": 1.3713404658751205e-07, "loss": 0.4891, "step": 22126 }, { "epoch": 0.9275331943032602, "grad_norm": 2.0066936016082764, "learning_rate": 1.3697619236515914e-07, "loss": 0.4703, "step": 22127 }, { "epoch": 0.9275751129182692, "grad_norm": 1.7751848697662354, "learning_rate": 1.3681842778592692e-07, "loss": 0.4355, "step": 22128 }, { "epoch": 0.9276170315332781, "grad_norm": 1.9114781618118286, "learning_rate": 1.3666075285272473e-07, "loss": 0.501, "step": 22129 }, { "epoch": 0.9276589501482871, "grad_norm": 1.7410274744033813, "learning_rate": 1.3650316756845695e-07, "loss": 0.4703, "step": 22130 }, { "epoch": 0.927700868763296, "grad_norm": 1.752708077430725, "learning_rate": 1.3634567193603067e-07, "loss": 0.4705, "step": 22131 }, { "epoch": 0.9277427873783051, "grad_norm": 1.8990222215652466, "learning_rate": 1.3618826595834743e-07, "loss": 0.4768, "step": 22132 }, { "epoch": 0.927784705993314, "grad_norm": 3.11108136177063, "learning_rate": 1.3603094963830943e-07, "loss": 0.435, "step": 22133 }, { "epoch": 0.927826624608323, "grad_norm": 1.9651871919631958, "learning_rate": 1.3587372297881707e-07, "loss": 0.4784, "step": 22134 }, { "epoch": 0.9278685432233319, "grad_norm": 1.705791711807251, "learning_rate": 1.3571658598276805e-07, "loss": 0.4232, "step": 22135 }, { "epoch": 0.9279104618383409, "grad_norm": 2.564178705215454, "learning_rate": 1.3555953865305893e-07, "loss": 0.4806, "step": 22136 }, { "epoch": 0.9279523804533498, "grad_norm": 3.4666287899017334, "learning_rate": 1.3540258099258518e-07, "loss": 0.5637, "step": 22137 }, { "epoch": 0.9279942990683587, "grad_norm": 1.848646879196167, "learning_rate": 1.3524571300424005e-07, "loss": 0.4299, "step": 22138 }, { "epoch": 0.9280362176833677, "grad_norm": 2.260986328125, "learning_rate": 1.35088934690914e-07, "loss": 0.4847, "step": 22139 }, { "epoch": 0.9280781362983767, "grad_norm": 2.4800431728363037, "learning_rate": 1.349322460554986e-07, "loss": 0.4371, "step": 22140 }, { "epoch": 0.9281200549133857, "grad_norm": 1.9062219858169556, "learning_rate": 1.3477564710088097e-07, "loss": 0.4957, "step": 22141 }, { "epoch": 0.9281619735283946, "grad_norm": 1.8834115266799927, "learning_rate": 1.3461913782994828e-07, "loss": 0.4999, "step": 22142 }, { "epoch": 0.9282038921434036, "grad_norm": 1.8066096305847168, "learning_rate": 1.34462718245586e-07, "loss": 0.4446, "step": 22143 }, { "epoch": 0.9282458107584125, "grad_norm": 1.89940345287323, "learning_rate": 1.3430638835067677e-07, "loss": 0.4768, "step": 22144 }, { "epoch": 0.9282877293734215, "grad_norm": 1.813169002532959, "learning_rate": 1.3415014814810279e-07, "loss": 0.4618, "step": 22145 }, { "epoch": 0.9283296479884304, "grad_norm": 2.160814046859741, "learning_rate": 1.339939976407445e-07, "loss": 0.4123, "step": 22146 }, { "epoch": 0.9283715666034394, "grad_norm": 4.076356887817383, "learning_rate": 1.338379368314796e-07, "loss": 0.469, "step": 22147 }, { "epoch": 0.9284134852184484, "grad_norm": 1.838712453842163, "learning_rate": 1.3368196572318415e-07, "loss": 0.4712, "step": 22148 }, { "epoch": 0.9284554038334574, "grad_norm": 2.1505327224731445, "learning_rate": 1.3352608431873526e-07, "loss": 0.5021, "step": 22149 }, { "epoch": 0.9284973224484663, "grad_norm": 2.1431477069854736, "learning_rate": 1.3337029262100454e-07, "loss": 0.4482, "step": 22150 }, { "epoch": 0.9285392410634753, "grad_norm": 1.7391331195831299, "learning_rate": 1.3321459063286468e-07, "loss": 0.4979, "step": 22151 }, { "epoch": 0.9285811596784842, "grad_norm": 1.8202942609786987, "learning_rate": 1.3305897835718673e-07, "loss": 0.4759, "step": 22152 }, { "epoch": 0.9286230782934932, "grad_norm": 2.136697292327881, "learning_rate": 1.3290345579683673e-07, "loss": 0.4078, "step": 22153 }, { "epoch": 0.9286649969085021, "grad_norm": 1.6237123012542725, "learning_rate": 1.3274802295468347e-07, "loss": 0.4558, "step": 22154 }, { "epoch": 0.9287069155235111, "grad_norm": 2.281050443649292, "learning_rate": 1.325926798335919e-07, "loss": 0.4372, "step": 22155 }, { "epoch": 0.92874883413852, "grad_norm": 2.057187080383301, "learning_rate": 1.3243742643642588e-07, "loss": 0.4578, "step": 22156 }, { "epoch": 0.9287907527535291, "grad_norm": 2.2026703357696533, "learning_rate": 1.3228226276604584e-07, "loss": 0.4579, "step": 22157 }, { "epoch": 0.928832671368538, "grad_norm": 1.6921591758728027, "learning_rate": 1.3212718882531285e-07, "loss": 0.4834, "step": 22158 }, { "epoch": 0.928874589983547, "grad_norm": 2.3397629261016846, "learning_rate": 1.3197220461708572e-07, "loss": 0.5191, "step": 22159 }, { "epoch": 0.9289165085985559, "grad_norm": 1.9032963514328003, "learning_rate": 1.3181731014422105e-07, "loss": 0.4519, "step": 22160 }, { "epoch": 0.9289584272135649, "grad_norm": 2.5126259326934814, "learning_rate": 1.3166250540957437e-07, "loss": 0.5244, "step": 22161 }, { "epoch": 0.9290003458285738, "grad_norm": 1.9662550687789917, "learning_rate": 1.3150779041599892e-07, "loss": 0.4745, "step": 22162 }, { "epoch": 0.9290422644435827, "grad_norm": 1.8654567003250122, "learning_rate": 1.3135316516634633e-07, "loss": 0.4589, "step": 22163 }, { "epoch": 0.9290841830585918, "grad_norm": 1.8162987232208252, "learning_rate": 1.3119862966346874e-07, "loss": 0.5257, "step": 22164 }, { "epoch": 0.9291261016736007, "grad_norm": 1.9946925640106201, "learning_rate": 1.3104418391021223e-07, "loss": 0.4627, "step": 22165 }, { "epoch": 0.9291680202886097, "grad_norm": 1.8555054664611816, "learning_rate": 1.3088982790942617e-07, "loss": 0.4523, "step": 22166 }, { "epoch": 0.9292099389036186, "grad_norm": 1.9751983880996704, "learning_rate": 1.307355616639544e-07, "loss": 0.5011, "step": 22167 }, { "epoch": 0.9292518575186276, "grad_norm": 2.4453155994415283, "learning_rate": 1.3058138517664077e-07, "loss": 0.4438, "step": 22168 }, { "epoch": 0.9292937761336365, "grad_norm": 1.806841254234314, "learning_rate": 1.3042729845032743e-07, "loss": 0.4452, "step": 22169 }, { "epoch": 0.9293356947486455, "grad_norm": 1.8052467107772827, "learning_rate": 1.3027330148785545e-07, "loss": 0.4629, "step": 22170 }, { "epoch": 0.9293776133636544, "grad_norm": 1.9311686754226685, "learning_rate": 1.3011939429206255e-07, "loss": 0.4542, "step": 22171 }, { "epoch": 0.9294195319786634, "grad_norm": 2.176999807357788, "learning_rate": 1.2996557686578648e-07, "loss": 0.4923, "step": 22172 }, { "epoch": 0.9294614505936724, "grad_norm": 2.9981889724731445, "learning_rate": 1.2981184921186163e-07, "loss": 0.4628, "step": 22173 }, { "epoch": 0.9295033692086814, "grad_norm": 2.235130786895752, "learning_rate": 1.2965821133312295e-07, "loss": 0.5045, "step": 22174 }, { "epoch": 0.9295452878236903, "grad_norm": 2.317988395690918, "learning_rate": 1.2950466323240262e-07, "loss": 0.4868, "step": 22175 }, { "epoch": 0.9295872064386993, "grad_norm": 2.8237571716308594, "learning_rate": 1.2935120491253006e-07, "loss": 0.535, "step": 22176 }, { "epoch": 0.9296291250537082, "grad_norm": 1.941616177558899, "learning_rate": 1.2919783637633465e-07, "loss": 0.422, "step": 22177 }, { "epoch": 0.9296710436687172, "grad_norm": 2.4598162174224854, "learning_rate": 1.2904455762664359e-07, "loss": 0.5476, "step": 22178 }, { "epoch": 0.9297129622837261, "grad_norm": 2.1903059482574463, "learning_rate": 1.2889136866628183e-07, "loss": 0.4851, "step": 22179 }, { "epoch": 0.9297548808987351, "grad_norm": 2.111255168914795, "learning_rate": 1.287382694980732e-07, "loss": 0.5083, "step": 22180 }, { "epoch": 0.929796799513744, "grad_norm": 2.1473004817962646, "learning_rate": 1.2858526012484162e-07, "loss": 0.4523, "step": 22181 }, { "epoch": 0.9298387181287531, "grad_norm": 3.7979888916015625, "learning_rate": 1.284323405494048e-07, "loss": 0.4737, "step": 22182 }, { "epoch": 0.929880636743762, "grad_norm": 1.9921132326126099, "learning_rate": 1.282795107745838e-07, "loss": 0.4941, "step": 22183 }, { "epoch": 0.929922555358771, "grad_norm": 1.8274447917938232, "learning_rate": 1.2812677080319524e-07, "loss": 0.4949, "step": 22184 }, { "epoch": 0.9299644739737799, "grad_norm": 3.9890568256378174, "learning_rate": 1.279741206380547e-07, "loss": 0.5205, "step": 22185 }, { "epoch": 0.9300063925887889, "grad_norm": 2.138503313064575, "learning_rate": 1.278215602819749e-07, "loss": 0.4437, "step": 22186 }, { "epoch": 0.9300483112037978, "grad_norm": 2.681218147277832, "learning_rate": 1.276690897377697e-07, "loss": 0.439, "step": 22187 }, { "epoch": 0.9300902298188067, "grad_norm": 2.4890592098236084, "learning_rate": 1.275167090082491e-07, "loss": 0.5162, "step": 22188 }, { "epoch": 0.9301321484338158, "grad_norm": 1.761307954788208, "learning_rate": 1.2736441809622136e-07, "loss": 0.4666, "step": 22189 }, { "epoch": 0.9301740670488247, "grad_norm": 1.7917635440826416, "learning_rate": 1.2721221700449481e-07, "loss": 0.4416, "step": 22190 }, { "epoch": 0.9302159856638337, "grad_norm": 1.8445652723312378, "learning_rate": 1.2706010573587447e-07, "loss": 0.4388, "step": 22191 }, { "epoch": 0.9302579042788426, "grad_norm": 2.5656843185424805, "learning_rate": 1.2690808429316414e-07, "loss": 0.4312, "step": 22192 }, { "epoch": 0.9302998228938516, "grad_norm": 1.9480167627334595, "learning_rate": 1.2675615267916719e-07, "loss": 0.4794, "step": 22193 }, { "epoch": 0.9303417415088605, "grad_norm": 3.897926092147827, "learning_rate": 1.2660431089668356e-07, "loss": 0.4679, "step": 22194 }, { "epoch": 0.9303836601238695, "grad_norm": 1.900024652481079, "learning_rate": 1.264525589485116e-07, "loss": 0.4832, "step": 22195 }, { "epoch": 0.9304255787388784, "grad_norm": 1.9772884845733643, "learning_rate": 1.2630089683744962e-07, "loss": 0.4551, "step": 22196 }, { "epoch": 0.9304674973538875, "grad_norm": 2.349416971206665, "learning_rate": 1.261493245662926e-07, "loss": 0.4554, "step": 22197 }, { "epoch": 0.9305094159688964, "grad_norm": 2.14754581451416, "learning_rate": 1.2599784213783439e-07, "loss": 0.4371, "step": 22198 }, { "epoch": 0.9305513345839054, "grad_norm": 5.144147872924805, "learning_rate": 1.258464495548689e-07, "loss": 0.555, "step": 22199 }, { "epoch": 0.9305932531989143, "grad_norm": 1.6914310455322266, "learning_rate": 1.25695146820185e-07, "loss": 0.4594, "step": 22200 }, { "epoch": 0.9306351718139233, "grad_norm": 1.860097885131836, "learning_rate": 1.2554393393657272e-07, "loss": 0.4546, "step": 22201 }, { "epoch": 0.9306770904289322, "grad_norm": 1.9536770582199097, "learning_rate": 1.2539281090681975e-07, "loss": 0.5382, "step": 22202 }, { "epoch": 0.9307190090439412, "grad_norm": 2.5860965251922607, "learning_rate": 1.2524177773371115e-07, "loss": 0.455, "step": 22203 }, { "epoch": 0.9307609276589501, "grad_norm": 2.1483349800109863, "learning_rate": 1.2509083442003133e-07, "loss": 0.5465, "step": 22204 }, { "epoch": 0.9308028462739592, "grad_norm": 1.8100475072860718, "learning_rate": 1.2493998096856252e-07, "loss": 0.4523, "step": 22205 }, { "epoch": 0.9308447648889681, "grad_norm": 1.8633875846862793, "learning_rate": 1.2478921738208527e-07, "loss": 0.4697, "step": 22206 }, { "epoch": 0.9308866835039771, "grad_norm": 2.4898831844329834, "learning_rate": 1.2463854366337902e-07, "loss": 0.4497, "step": 22207 }, { "epoch": 0.930928602118986, "grad_norm": 6.189417839050293, "learning_rate": 1.244879598152221e-07, "loss": 0.454, "step": 22208 }, { "epoch": 0.930970520733995, "grad_norm": 2.090216636657715, "learning_rate": 1.2433746584038843e-07, "loss": 0.5234, "step": 22209 }, { "epoch": 0.9310124393490039, "grad_norm": 1.8648512363433838, "learning_rate": 1.241870617416535e-07, "loss": 0.47, "step": 22210 }, { "epoch": 0.9310543579640129, "grad_norm": 2.004114866256714, "learning_rate": 1.2403674752178963e-07, "loss": 0.5082, "step": 22211 }, { "epoch": 0.9310962765790218, "grad_norm": 2.776050090789795, "learning_rate": 1.238865231835673e-07, "loss": 0.4748, "step": 22212 }, { "epoch": 0.9311381951940307, "grad_norm": 2.4146056175231934, "learning_rate": 1.2373638872975602e-07, "loss": 0.47, "step": 22213 }, { "epoch": 0.9311801138090398, "grad_norm": 1.6455780267715454, "learning_rate": 1.2358634416312355e-07, "loss": 0.4498, "step": 22214 }, { "epoch": 0.9312220324240487, "grad_norm": 1.9757298231124878, "learning_rate": 1.2343638948643488e-07, "loss": 0.4632, "step": 22215 }, { "epoch": 0.9312639510390577, "grad_norm": 2.100449323654175, "learning_rate": 1.232865247024545e-07, "loss": 0.482, "step": 22216 }, { "epoch": 0.9313058696540666, "grad_norm": 2.597877264022827, "learning_rate": 1.2313674981394574e-07, "loss": 0.4872, "step": 22217 }, { "epoch": 0.9313477882690756, "grad_norm": 1.871809720993042, "learning_rate": 1.229870648236686e-07, "loss": 0.4802, "step": 22218 }, { "epoch": 0.9313897068840845, "grad_norm": 2.1896135807037354, "learning_rate": 1.228374697343826e-07, "loss": 0.4481, "step": 22219 }, { "epoch": 0.9314316254990935, "grad_norm": 1.871091365814209, "learning_rate": 1.22687964548846e-07, "loss": 0.4889, "step": 22220 }, { "epoch": 0.9314735441141024, "grad_norm": 2.056030035018921, "learning_rate": 1.2253854926981333e-07, "loss": 0.4212, "step": 22221 }, { "epoch": 0.9315154627291115, "grad_norm": 2.1886887550354004, "learning_rate": 1.2238922390004015e-07, "loss": 0.5033, "step": 22222 }, { "epoch": 0.9315573813441204, "grad_norm": 1.7116756439208984, "learning_rate": 1.222399884422787e-07, "loss": 0.4702, "step": 22223 }, { "epoch": 0.9315992999591294, "grad_norm": 2.223334312438965, "learning_rate": 1.2209084289927896e-07, "loss": 0.5088, "step": 22224 }, { "epoch": 0.9316412185741383, "grad_norm": 2.2875258922576904, "learning_rate": 1.2194178727379158e-07, "loss": 0.5239, "step": 22225 }, { "epoch": 0.9316831371891473, "grad_norm": 1.9659056663513184, "learning_rate": 1.2179282156856377e-07, "loss": 0.4823, "step": 22226 }, { "epoch": 0.9317250558041562, "grad_norm": 1.8310232162475586, "learning_rate": 1.2164394578634053e-07, "loss": 0.4465, "step": 22227 }, { "epoch": 0.9317669744191652, "grad_norm": 2.0006158351898193, "learning_rate": 1.214951599298675e-07, "loss": 0.5091, "step": 22228 }, { "epoch": 0.9318088930341741, "grad_norm": 1.965194582939148, "learning_rate": 1.2134646400188742e-07, "loss": 0.442, "step": 22229 }, { "epoch": 0.9318508116491832, "grad_norm": 2.145960807800293, "learning_rate": 1.2119785800513983e-07, "loss": 0.4829, "step": 22230 }, { "epoch": 0.9318927302641921, "grad_norm": 1.7516216039657593, "learning_rate": 1.2104934194236585e-07, "loss": 0.4986, "step": 22231 }, { "epoch": 0.9319346488792011, "grad_norm": 2.462218761444092, "learning_rate": 1.2090091581630216e-07, "loss": 0.4876, "step": 22232 }, { "epoch": 0.93197656749421, "grad_norm": 2.264103651046753, "learning_rate": 1.2075257962968435e-07, "loss": 0.4569, "step": 22233 }, { "epoch": 0.932018486109219, "grad_norm": 2.1811976432800293, "learning_rate": 1.2060433338524747e-07, "loss": 0.4852, "step": 22234 }, { "epoch": 0.9320604047242279, "grad_norm": 9.307158470153809, "learning_rate": 1.2045617708572488e-07, "loss": 0.4274, "step": 22235 }, { "epoch": 0.9321023233392369, "grad_norm": 2.1278433799743652, "learning_rate": 1.2030811073384607e-07, "loss": 0.4366, "step": 22236 }, { "epoch": 0.9321442419542458, "grad_norm": 1.9875961542129517, "learning_rate": 1.2016013433234107e-07, "loss": 0.4582, "step": 22237 }, { "epoch": 0.9321861605692547, "grad_norm": 2.5206587314605713, "learning_rate": 1.2001224788393883e-07, "loss": 0.5389, "step": 22238 }, { "epoch": 0.9322280791842638, "grad_norm": 1.9276132583618164, "learning_rate": 1.1986445139136383e-07, "loss": 0.475, "step": 22239 }, { "epoch": 0.9322699977992727, "grad_norm": 1.6287130117416382, "learning_rate": 1.1971674485734164e-07, "loss": 0.4834, "step": 22240 }, { "epoch": 0.9323119164142817, "grad_norm": 1.9952083826065063, "learning_rate": 1.1956912828459398e-07, "loss": 0.5217, "step": 22241 }, { "epoch": 0.9323538350292906, "grad_norm": 2.124885320663452, "learning_rate": 1.1942160167584203e-07, "loss": 0.463, "step": 22242 }, { "epoch": 0.9323957536442996, "grad_norm": 1.7011313438415527, "learning_rate": 1.192741650338064e-07, "loss": 0.4814, "step": 22243 }, { "epoch": 0.9324376722593085, "grad_norm": 1.8396698236465454, "learning_rate": 1.1912681836120376e-07, "loss": 0.5148, "step": 22244 }, { "epoch": 0.9324795908743175, "grad_norm": 1.8758536577224731, "learning_rate": 1.189795616607503e-07, "loss": 0.4909, "step": 22245 }, { "epoch": 0.9325215094893264, "grad_norm": 8.344972610473633, "learning_rate": 1.1883239493516163e-07, "loss": 0.4718, "step": 22246 }, { "epoch": 0.9325634281043355, "grad_norm": 1.4531540870666504, "learning_rate": 1.1868531818714945e-07, "loss": 0.3543, "step": 22247 }, { "epoch": 0.9326053467193444, "grad_norm": 1.9369384050369263, "learning_rate": 1.1853833141942439e-07, "loss": 0.4557, "step": 22248 }, { "epoch": 0.9326472653343534, "grad_norm": 1.8172603845596313, "learning_rate": 1.1839143463469815e-07, "loss": 0.5145, "step": 22249 }, { "epoch": 0.9326891839493623, "grad_norm": 1.892041563987732, "learning_rate": 1.1824462783567636e-07, "loss": 0.4683, "step": 22250 }, { "epoch": 0.9327311025643713, "grad_norm": 2.1354501247406006, "learning_rate": 1.180979110250663e-07, "loss": 0.4812, "step": 22251 }, { "epoch": 0.9327730211793802, "grad_norm": 1.6244525909423828, "learning_rate": 1.1795128420557244e-07, "loss": 0.4522, "step": 22252 }, { "epoch": 0.9328149397943892, "grad_norm": 2.4488718509674072, "learning_rate": 1.1780474737989655e-07, "loss": 0.4692, "step": 22253 }, { "epoch": 0.9328568584093981, "grad_norm": 1.803415060043335, "learning_rate": 1.1765830055074145e-07, "loss": 0.4498, "step": 22254 }, { "epoch": 0.9328987770244072, "grad_norm": 2.5124588012695312, "learning_rate": 1.1751194372080555e-07, "loss": 0.4919, "step": 22255 }, { "epoch": 0.9329406956394161, "grad_norm": 1.89309561252594, "learning_rate": 1.1736567689278721e-07, "loss": 0.5037, "step": 22256 }, { "epoch": 0.9329826142544251, "grad_norm": 1.7694154977798462, "learning_rate": 1.1721950006938265e-07, "loss": 0.4542, "step": 22257 }, { "epoch": 0.933024532869434, "grad_norm": 1.9683237075805664, "learning_rate": 1.1707341325328692e-07, "loss": 0.4719, "step": 22258 }, { "epoch": 0.933066451484443, "grad_norm": 2.539951801300049, "learning_rate": 1.1692741644719119e-07, "loss": 0.4694, "step": 22259 }, { "epoch": 0.9331083700994519, "grad_norm": 2.0255777835845947, "learning_rate": 1.1678150965378888e-07, "loss": 0.4359, "step": 22260 }, { "epoch": 0.9331502887144609, "grad_norm": 2.409263849258423, "learning_rate": 1.1663569287576892e-07, "loss": 0.4911, "step": 22261 }, { "epoch": 0.9331922073294698, "grad_norm": 2.1640167236328125, "learning_rate": 1.1648996611581809e-07, "loss": 0.462, "step": 22262 }, { "epoch": 0.9332341259444787, "grad_norm": 5.179112911224365, "learning_rate": 1.163443293766231e-07, "loss": 0.5178, "step": 22263 }, { "epoch": 0.9332760445594878, "grad_norm": 1.720829725265503, "learning_rate": 1.1619878266087015e-07, "loss": 0.4997, "step": 22264 }, { "epoch": 0.9333179631744967, "grad_norm": 1.7365267276763916, "learning_rate": 1.1605332597123986e-07, "loss": 0.4305, "step": 22265 }, { "epoch": 0.9333598817895057, "grad_norm": 1.8497756719589233, "learning_rate": 1.159079593104151e-07, "loss": 0.4447, "step": 22266 }, { "epoch": 0.9334018004045146, "grad_norm": 1.7814971208572388, "learning_rate": 1.1576268268107538e-07, "loss": 0.4755, "step": 22267 }, { "epoch": 0.9334437190195236, "grad_norm": 2.4367692470550537, "learning_rate": 1.1561749608589745e-07, "loss": 0.5649, "step": 22268 }, { "epoch": 0.9334856376345325, "grad_norm": 2.585484027862549, "learning_rate": 1.1547239952755974e-07, "loss": 0.4919, "step": 22269 }, { "epoch": 0.9335275562495415, "grad_norm": 1.7335100173950195, "learning_rate": 1.1532739300873507e-07, "loss": 0.4568, "step": 22270 }, { "epoch": 0.9335694748645504, "grad_norm": 2.243345260620117, "learning_rate": 1.1518247653209691e-07, "loss": 0.52, "step": 22271 }, { "epoch": 0.9336113934795595, "grad_norm": 1.9814051389694214, "learning_rate": 1.1503765010031697e-07, "loss": 0.4235, "step": 22272 }, { "epoch": 0.9336533120945684, "grad_norm": 2.2723867893218994, "learning_rate": 1.1489291371606482e-07, "loss": 0.4838, "step": 22273 }, { "epoch": 0.9336952307095774, "grad_norm": 1.5005497932434082, "learning_rate": 1.1474826738200772e-07, "loss": 0.4288, "step": 22274 }, { "epoch": 0.9337371493245863, "grad_norm": 1.9847352504730225, "learning_rate": 1.1460371110081304e-07, "loss": 0.4559, "step": 22275 }, { "epoch": 0.9337790679395953, "grad_norm": 1.9103713035583496, "learning_rate": 1.1445924487514526e-07, "loss": 0.4586, "step": 22276 }, { "epoch": 0.9338209865546042, "grad_norm": 1.9277491569519043, "learning_rate": 1.1431486870766728e-07, "loss": 0.4525, "step": 22277 }, { "epoch": 0.9338629051696132, "grad_norm": 1.945132851600647, "learning_rate": 1.141705826010403e-07, "loss": 0.4642, "step": 22278 }, { "epoch": 0.9339048237846221, "grad_norm": 1.597049355506897, "learning_rate": 1.1402638655792442e-07, "loss": 0.4526, "step": 22279 }, { "epoch": 0.9339467423996312, "grad_norm": 2.4828813076019287, "learning_rate": 1.1388228058097694e-07, "loss": 0.4799, "step": 22280 }, { "epoch": 0.9339886610146401, "grad_norm": 2.707526683807373, "learning_rate": 1.1373826467285465e-07, "loss": 0.4625, "step": 22281 }, { "epoch": 0.9340305796296491, "grad_norm": 1.9959123134613037, "learning_rate": 1.1359433883621319e-07, "loss": 0.4568, "step": 22282 }, { "epoch": 0.934072498244658, "grad_norm": 1.9496710300445557, "learning_rate": 1.134505030737043e-07, "loss": 0.4336, "step": 22283 }, { "epoch": 0.934114416859667, "grad_norm": 1.937265396118164, "learning_rate": 1.133067573879798e-07, "loss": 0.451, "step": 22284 }, { "epoch": 0.9341563354746759, "grad_norm": 1.6025787591934204, "learning_rate": 1.1316310178168977e-07, "loss": 0.4794, "step": 22285 }, { "epoch": 0.9341982540896849, "grad_norm": 2.25495982170105, "learning_rate": 1.1301953625748208e-07, "loss": 0.4527, "step": 22286 }, { "epoch": 0.9342401727046938, "grad_norm": 1.68916654586792, "learning_rate": 1.128760608180035e-07, "loss": 0.4386, "step": 22287 }, { "epoch": 0.9342820913197027, "grad_norm": 2.384172201156616, "learning_rate": 1.1273267546589861e-07, "loss": 0.4585, "step": 22288 }, { "epoch": 0.9343240099347118, "grad_norm": 1.7668808698654175, "learning_rate": 1.1258938020380971e-07, "loss": 0.4741, "step": 22289 }, { "epoch": 0.9343659285497207, "grad_norm": 2.0836410522460938, "learning_rate": 1.1244617503437916e-07, "loss": 0.4833, "step": 22290 }, { "epoch": 0.9344078471647297, "grad_norm": 1.798777461051941, "learning_rate": 1.1230305996024704e-07, "loss": 0.4839, "step": 22291 }, { "epoch": 0.9344497657797386, "grad_norm": 3.770719528198242, "learning_rate": 1.1216003498405015e-07, "loss": 0.4616, "step": 22292 }, { "epoch": 0.9344916843947476, "grad_norm": 1.9700918197631836, "learning_rate": 1.1201710010842581e-07, "loss": 0.451, "step": 22293 }, { "epoch": 0.9345336030097565, "grad_norm": 2.099724292755127, "learning_rate": 1.1187425533600971e-07, "loss": 0.4631, "step": 22294 }, { "epoch": 0.9345755216247655, "grad_norm": 1.7345951795578003, "learning_rate": 1.1173150066943306e-07, "loss": 0.4797, "step": 22295 }, { "epoch": 0.9346174402397744, "grad_norm": 3.271690607070923, "learning_rate": 1.1158883611132931e-07, "loss": 0.4661, "step": 22296 }, { "epoch": 0.9346593588547835, "grad_norm": 1.9298241138458252, "learning_rate": 1.1144626166432693e-07, "loss": 0.4364, "step": 22297 }, { "epoch": 0.9347012774697924, "grad_norm": 1.9022029638290405, "learning_rate": 1.1130377733105436e-07, "loss": 0.4638, "step": 22298 }, { "epoch": 0.9347431960848014, "grad_norm": 1.7237515449523926, "learning_rate": 1.1116138311413838e-07, "loss": 0.4469, "step": 22299 }, { "epoch": 0.9347851146998103, "grad_norm": 1.9561853408813477, "learning_rate": 1.1101907901620413e-07, "loss": 0.47, "step": 22300 }, { "epoch": 0.9348270333148193, "grad_norm": 1.8926244974136353, "learning_rate": 1.1087686503987338e-07, "loss": 0.5218, "step": 22301 }, { "epoch": 0.9348689519298282, "grad_norm": 2.353804588317871, "learning_rate": 1.1073474118776905e-07, "loss": 0.4391, "step": 22302 }, { "epoch": 0.9349108705448372, "grad_norm": 1.8300812244415283, "learning_rate": 1.1059270746251072e-07, "loss": 0.4576, "step": 22303 }, { "epoch": 0.9349527891598461, "grad_norm": 2.1364734172821045, "learning_rate": 1.1045076386671627e-07, "loss": 0.4396, "step": 22304 }, { "epoch": 0.9349947077748552, "grad_norm": 2.053377389907837, "learning_rate": 1.1030891040300252e-07, "loss": 0.4603, "step": 22305 }, { "epoch": 0.9350366263898641, "grad_norm": 1.8172354698181152, "learning_rate": 1.1016714707398401e-07, "loss": 0.4358, "step": 22306 }, { "epoch": 0.9350785450048731, "grad_norm": 2.7288758754730225, "learning_rate": 1.1002547388227425e-07, "loss": 0.5162, "step": 22307 }, { "epoch": 0.935120463619882, "grad_norm": 2.5425381660461426, "learning_rate": 1.0988389083048446e-07, "loss": 0.4899, "step": 22308 }, { "epoch": 0.935162382234891, "grad_norm": 2.1353886127471924, "learning_rate": 1.0974239792122532e-07, "loss": 0.4531, "step": 22309 }, { "epoch": 0.9352043008498999, "grad_norm": 2.335071086883545, "learning_rate": 1.096009951571042e-07, "loss": 0.5129, "step": 22310 }, { "epoch": 0.9352462194649089, "grad_norm": 2.0297224521636963, "learning_rate": 1.0945968254072791e-07, "loss": 0.4891, "step": 22311 }, { "epoch": 0.9352881380799178, "grad_norm": 1.951003074645996, "learning_rate": 1.093184600747016e-07, "loss": 0.4473, "step": 22312 }, { "epoch": 0.9353300566949267, "grad_norm": 2.0471925735473633, "learning_rate": 1.0917732776162759e-07, "loss": 0.4554, "step": 22313 }, { "epoch": 0.9353719753099358, "grad_norm": 1.6216659545898438, "learning_rate": 1.0903628560410883e-07, "loss": 0.3956, "step": 22314 }, { "epoch": 0.9354138939249447, "grad_norm": 2.4793920516967773, "learning_rate": 1.0889533360474492e-07, "loss": 0.5314, "step": 22315 }, { "epoch": 0.9354558125399537, "grad_norm": 2.193223714828491, "learning_rate": 1.0875447176613318e-07, "loss": 0.5347, "step": 22316 }, { "epoch": 0.9354977311549626, "grad_norm": 2.1277213096618652, "learning_rate": 1.0861370009087102e-07, "loss": 0.4585, "step": 22317 }, { "epoch": 0.9355396497699716, "grad_norm": 2.3109450340270996, "learning_rate": 1.0847301858155357e-07, "loss": 0.4914, "step": 22318 }, { "epoch": 0.9355815683849805, "grad_norm": 1.6629116535186768, "learning_rate": 1.0833242724077264e-07, "loss": 0.4735, "step": 22319 }, { "epoch": 0.9356234869999895, "grad_norm": 1.8495076894760132, "learning_rate": 1.0819192607112172e-07, "loss": 0.4337, "step": 22320 }, { "epoch": 0.9356654056149984, "grad_norm": 2.2965896129608154, "learning_rate": 1.0805151507518985e-07, "loss": 0.4804, "step": 22321 }, { "epoch": 0.9357073242300075, "grad_norm": 1.9046461582183838, "learning_rate": 1.0791119425556495e-07, "loss": 0.5171, "step": 22322 }, { "epoch": 0.9357492428450164, "grad_norm": 2.0417866706848145, "learning_rate": 1.0777096361483441e-07, "loss": 0.4881, "step": 22323 }, { "epoch": 0.9357911614600254, "grad_norm": 1.7465087175369263, "learning_rate": 1.0763082315558283e-07, "loss": 0.4999, "step": 22324 }, { "epoch": 0.9358330800750343, "grad_norm": 1.7816544771194458, "learning_rate": 1.0749077288039367e-07, "loss": 0.4667, "step": 22325 }, { "epoch": 0.9358749986900433, "grad_norm": 1.848368525505066, "learning_rate": 1.0735081279184878e-07, "loss": 0.5284, "step": 22326 }, { "epoch": 0.9359169173050522, "grad_norm": 2.0026302337646484, "learning_rate": 1.0721094289252664e-07, "loss": 0.4649, "step": 22327 }, { "epoch": 0.9359588359200612, "grad_norm": 2.7944929599761963, "learning_rate": 1.0707116318500743e-07, "loss": 0.5359, "step": 22328 }, { "epoch": 0.9360007545350701, "grad_norm": 9.346301078796387, "learning_rate": 1.0693147367186685e-07, "loss": 0.4423, "step": 22329 }, { "epoch": 0.9360426731500792, "grad_norm": 1.7905054092407227, "learning_rate": 1.0679187435568005e-07, "loss": 0.5085, "step": 22330 }, { "epoch": 0.9360845917650881, "grad_norm": 1.8516767024993896, "learning_rate": 1.0665236523902056e-07, "loss": 0.4883, "step": 22331 }, { "epoch": 0.9361265103800971, "grad_norm": 4.171413898468018, "learning_rate": 1.0651294632445963e-07, "loss": 0.4478, "step": 22332 }, { "epoch": 0.936168428995106, "grad_norm": 1.9357011318206787, "learning_rate": 1.0637361761456744e-07, "loss": 0.4655, "step": 22333 }, { "epoch": 0.936210347610115, "grad_norm": 2.107961416244507, "learning_rate": 1.0623437911191303e-07, "loss": 0.4785, "step": 22334 }, { "epoch": 0.9362522662251239, "grad_norm": 1.8752694129943848, "learning_rate": 1.0609523081906214e-07, "loss": 0.4994, "step": 22335 }, { "epoch": 0.9362941848401329, "grad_norm": 1.9087997674942017, "learning_rate": 1.0595617273857995e-07, "loss": 0.4588, "step": 22336 }, { "epoch": 0.9363361034551418, "grad_norm": 1.771095633506775, "learning_rate": 1.0581720487302938e-07, "loss": 0.4528, "step": 22337 }, { "epoch": 0.9363780220701508, "grad_norm": 2.456454038619995, "learning_rate": 1.056783272249734e-07, "loss": 0.5081, "step": 22338 }, { "epoch": 0.9364199406851598, "grad_norm": 2.5544543266296387, "learning_rate": 1.0553953979697106e-07, "loss": 0.4576, "step": 22339 }, { "epoch": 0.9364618593001687, "grad_norm": 1.6768321990966797, "learning_rate": 1.0540084259158035e-07, "loss": 0.4211, "step": 22340 }, { "epoch": 0.9365037779151777, "grad_norm": 1.9776133298873901, "learning_rate": 1.0526223561135917e-07, "loss": 0.4911, "step": 22341 }, { "epoch": 0.9365456965301866, "grad_norm": 1.7642353773117065, "learning_rate": 1.0512371885886163e-07, "loss": 0.4137, "step": 22342 }, { "epoch": 0.9365876151451956, "grad_norm": 1.6045973300933838, "learning_rate": 1.0498529233664178e-07, "loss": 0.5285, "step": 22343 }, { "epoch": 0.9366295337602045, "grad_norm": 2.048157215118408, "learning_rate": 1.0484695604725037e-07, "loss": 0.4539, "step": 22344 }, { "epoch": 0.9366714523752135, "grad_norm": 1.6792285442352295, "learning_rate": 1.0470870999323867e-07, "loss": 0.4409, "step": 22345 }, { "epoch": 0.9367133709902224, "grad_norm": 3.981337070465088, "learning_rate": 1.0457055417715357e-07, "loss": 0.4628, "step": 22346 }, { "epoch": 0.9367552896052315, "grad_norm": 2.3330397605895996, "learning_rate": 1.0443248860154298e-07, "loss": 0.4891, "step": 22347 }, { "epoch": 0.9367972082202404, "grad_norm": 1.8951159715652466, "learning_rate": 1.0429451326895157e-07, "loss": 0.478, "step": 22348 }, { "epoch": 0.9368391268352494, "grad_norm": 2.0880019664764404, "learning_rate": 1.041566281819223e-07, "loss": 0.5077, "step": 22349 }, { "epoch": 0.9368810454502583, "grad_norm": 3.3943073749542236, "learning_rate": 1.0401883334299756e-07, "loss": 0.5275, "step": 22350 }, { "epoch": 0.9369229640652673, "grad_norm": 1.6898396015167236, "learning_rate": 1.0388112875471701e-07, "loss": 0.4631, "step": 22351 }, { "epoch": 0.9369648826802762, "grad_norm": 2.100015878677368, "learning_rate": 1.0374351441961916e-07, "loss": 0.5428, "step": 22352 }, { "epoch": 0.9370068012952852, "grad_norm": 2.1456496715545654, "learning_rate": 1.0360599034024143e-07, "loss": 0.3986, "step": 22353 }, { "epoch": 0.9370487199102941, "grad_norm": 1.7532039880752563, "learning_rate": 1.0346855651911791e-07, "loss": 0.476, "step": 22354 }, { "epoch": 0.9370906385253032, "grad_norm": 1.7877355813980103, "learning_rate": 1.0333121295878157e-07, "loss": 0.4962, "step": 22355 }, { "epoch": 0.9371325571403121, "grad_norm": 2.320741653442383, "learning_rate": 1.0319395966176593e-07, "loss": 0.5003, "step": 22356 }, { "epoch": 0.9371744757553211, "grad_norm": 2.0971076488494873, "learning_rate": 1.0305679663059898e-07, "loss": 0.4999, "step": 22357 }, { "epoch": 0.93721639437033, "grad_norm": 2.390725612640381, "learning_rate": 1.0291972386781035e-07, "loss": 0.4945, "step": 22358 }, { "epoch": 0.937258312985339, "grad_norm": 2.0552382469177246, "learning_rate": 1.0278274137592748e-07, "loss": 0.4392, "step": 22359 }, { "epoch": 0.9373002316003479, "grad_norm": 1.7218085527420044, "learning_rate": 1.0264584915747389e-07, "loss": 0.4498, "step": 22360 }, { "epoch": 0.9373421502153569, "grad_norm": 1.8779929876327515, "learning_rate": 1.0250904721497312e-07, "loss": 0.5106, "step": 22361 }, { "epoch": 0.9373840688303658, "grad_norm": 1.761600136756897, "learning_rate": 1.0237233555094873e-07, "loss": 0.427, "step": 22362 }, { "epoch": 0.9374259874453748, "grad_norm": 1.9363471269607544, "learning_rate": 1.0223571416791922e-07, "loss": 0.4659, "step": 22363 }, { "epoch": 0.9374679060603838, "grad_norm": 3.3735837936401367, "learning_rate": 1.0209918306840316e-07, "loss": 0.4774, "step": 22364 }, { "epoch": 0.9375098246753927, "grad_norm": 2.164095878601074, "learning_rate": 1.0196274225491743e-07, "loss": 0.5029, "step": 22365 }, { "epoch": 0.9375517432904017, "grad_norm": 1.7633721828460693, "learning_rate": 1.0182639172997722e-07, "loss": 0.4456, "step": 22366 }, { "epoch": 0.9375936619054106, "grad_norm": 1.7218273878097534, "learning_rate": 1.0169013149609553e-07, "loss": 0.4216, "step": 22367 }, { "epoch": 0.9376355805204196, "grad_norm": 1.7437171936035156, "learning_rate": 1.0155396155578534e-07, "loss": 0.4994, "step": 22368 }, { "epoch": 0.9376774991354285, "grad_norm": 2.0847318172454834, "learning_rate": 1.0141788191155577e-07, "loss": 0.5184, "step": 22369 }, { "epoch": 0.9377194177504375, "grad_norm": 1.739876389503479, "learning_rate": 1.0128189256591481e-07, "loss": 0.4866, "step": 22370 }, { "epoch": 0.9377613363654465, "grad_norm": 1.9002304077148438, "learning_rate": 1.01145993521371e-07, "loss": 0.4207, "step": 22371 }, { "epoch": 0.9378032549804555, "grad_norm": 1.9211879968643188, "learning_rate": 1.010101847804279e-07, "loss": 0.5225, "step": 22372 }, { "epoch": 0.9378451735954644, "grad_norm": 2.7121589183807373, "learning_rate": 1.0087446634558906e-07, "loss": 0.5046, "step": 22373 }, { "epoch": 0.9378870922104734, "grad_norm": 2.025566339492798, "learning_rate": 1.0073883821935692e-07, "loss": 0.439, "step": 22374 }, { "epoch": 0.9379290108254823, "grad_norm": 1.8534059524536133, "learning_rate": 1.0060330040423117e-07, "loss": 0.4642, "step": 22375 }, { "epoch": 0.9379709294404913, "grad_norm": 2.1556854248046875, "learning_rate": 1.0046785290270978e-07, "loss": 0.4726, "step": 22376 }, { "epoch": 0.9380128480555002, "grad_norm": 2.028817892074585, "learning_rate": 1.003324957172913e-07, "loss": 0.5065, "step": 22377 }, { "epoch": 0.9380547666705092, "grad_norm": 1.7134767770767212, "learning_rate": 1.0019722885046879e-07, "loss": 0.4786, "step": 22378 }, { "epoch": 0.9380966852855182, "grad_norm": 1.71021568775177, "learning_rate": 1.0006205230473686e-07, "loss": 0.456, "step": 22379 }, { "epoch": 0.9381386039005272, "grad_norm": 2.16548228263855, "learning_rate": 9.992696608258689e-08, "loss": 0.487, "step": 22380 }, { "epoch": 0.9381805225155361, "grad_norm": 1.926582932472229, "learning_rate": 9.979197018650966e-08, "loss": 0.4796, "step": 22381 }, { "epoch": 0.9382224411305451, "grad_norm": 2.0378830432891846, "learning_rate": 9.96570646189926e-08, "loss": 0.5185, "step": 22382 }, { "epoch": 0.938264359745554, "grad_norm": 2.34066104888916, "learning_rate": 9.952224938252319e-08, "loss": 0.4713, "step": 22383 }, { "epoch": 0.938306278360563, "grad_norm": 2.2349886894226074, "learning_rate": 9.938752447958666e-08, "loss": 0.4635, "step": 22384 }, { "epoch": 0.9383481969755719, "grad_norm": 1.8954373598098755, "learning_rate": 9.925288991266546e-08, "loss": 0.4256, "step": 22385 }, { "epoch": 0.9383901155905809, "grad_norm": 1.800683617591858, "learning_rate": 9.911834568424316e-08, "loss": 0.4625, "step": 22386 }, { "epoch": 0.9384320342055898, "grad_norm": 2.2465600967407227, "learning_rate": 9.898389179679834e-08, "loss": 0.4365, "step": 22387 }, { "epoch": 0.9384739528205988, "grad_norm": 1.7975813150405884, "learning_rate": 9.884952825280958e-08, "loss": 0.4842, "step": 22388 }, { "epoch": 0.9385158714356078, "grad_norm": 1.602163314819336, "learning_rate": 9.871525505475488e-08, "loss": 0.4339, "step": 22389 }, { "epoch": 0.9385577900506167, "grad_norm": 2.3179495334625244, "learning_rate": 9.858107220510893e-08, "loss": 0.5305, "step": 22390 }, { "epoch": 0.9385997086656257, "grad_norm": 3.3952441215515137, "learning_rate": 9.844697970634364e-08, "loss": 0.4812, "step": 22391 }, { "epoch": 0.9386416272806346, "grad_norm": 1.745378851890564, "learning_rate": 9.831297756093317e-08, "loss": 0.4754, "step": 22392 }, { "epoch": 0.9386835458956436, "grad_norm": 1.9741506576538086, "learning_rate": 9.817906577134606e-08, "loss": 0.462, "step": 22393 }, { "epoch": 0.9387254645106525, "grad_norm": 2.4124934673309326, "learning_rate": 9.804524434005148e-08, "loss": 0.4699, "step": 22394 }, { "epoch": 0.9387673831256615, "grad_norm": 2.955536365509033, "learning_rate": 9.79115132695163e-08, "loss": 0.4619, "step": 22395 }, { "epoch": 0.9388093017406705, "grad_norm": 1.6572095155715942, "learning_rate": 9.777787256220473e-08, "loss": 0.476, "step": 22396 }, { "epoch": 0.9388512203556795, "grad_norm": 1.8825459480285645, "learning_rate": 9.764432222058195e-08, "loss": 0.4523, "step": 22397 }, { "epoch": 0.9388931389706884, "grad_norm": 2.3652360439300537, "learning_rate": 9.751086224710771e-08, "loss": 0.4868, "step": 22398 }, { "epoch": 0.9389350575856974, "grad_norm": 1.7370657920837402, "learning_rate": 9.73774926442439e-08, "loss": 0.5206, "step": 22399 }, { "epoch": 0.9389769762007063, "grad_norm": 1.6992216110229492, "learning_rate": 9.724421341444856e-08, "loss": 0.4441, "step": 22400 }, { "epoch": 0.9390188948157153, "grad_norm": 2.148531675338745, "learning_rate": 9.711102456017806e-08, "loss": 0.5087, "step": 22401 }, { "epoch": 0.9390608134307242, "grad_norm": 1.9577205181121826, "learning_rate": 9.697792608388767e-08, "loss": 0.5188, "step": 22402 }, { "epoch": 0.9391027320457332, "grad_norm": 1.8962007761001587, "learning_rate": 9.684491798803097e-08, "loss": 0.5028, "step": 22403 }, { "epoch": 0.9391446506607422, "grad_norm": 2.813725471496582, "learning_rate": 9.67120002750599e-08, "loss": 0.4685, "step": 22404 }, { "epoch": 0.9391865692757512, "grad_norm": 1.8787612915039062, "learning_rate": 9.657917294742414e-08, "loss": 0.4696, "step": 22405 }, { "epoch": 0.9392284878907601, "grad_norm": 1.9573395252227783, "learning_rate": 9.644643600757287e-08, "loss": 0.5001, "step": 22406 }, { "epoch": 0.9392704065057691, "grad_norm": 2.0471136569976807, "learning_rate": 9.631378945795244e-08, "loss": 0.4753, "step": 22407 }, { "epoch": 0.939312325120778, "grad_norm": 2.644253730773926, "learning_rate": 9.618123330100814e-08, "loss": 0.4768, "step": 22408 }, { "epoch": 0.939354243735787, "grad_norm": 1.6951773166656494, "learning_rate": 9.604876753918413e-08, "loss": 0.4612, "step": 22409 }, { "epoch": 0.9393961623507959, "grad_norm": 1.8480913639068604, "learning_rate": 9.591639217492121e-08, "loss": 0.467, "step": 22410 }, { "epoch": 0.939438080965805, "grad_norm": 2.195004463195801, "learning_rate": 9.578410721065967e-08, "loss": 0.4825, "step": 22411 }, { "epoch": 0.9394799995808139, "grad_norm": 2.3945255279541016, "learning_rate": 9.56519126488381e-08, "loss": 0.4621, "step": 22412 }, { "epoch": 0.9395219181958228, "grad_norm": 1.814554214477539, "learning_rate": 9.551980849189402e-08, "loss": 0.5055, "step": 22413 }, { "epoch": 0.9395638368108318, "grad_norm": 1.814776062965393, "learning_rate": 9.538779474226156e-08, "loss": 0.4629, "step": 22414 }, { "epoch": 0.9396057554258407, "grad_norm": 2.0609323978424072, "learning_rate": 9.525587140237491e-08, "loss": 0.5177, "step": 22415 }, { "epoch": 0.9396476740408497, "grad_norm": 1.5494318008422852, "learning_rate": 9.512403847466545e-08, "loss": 0.466, "step": 22416 }, { "epoch": 0.9396895926558586, "grad_norm": 2.0392379760742188, "learning_rate": 9.499229596156345e-08, "loss": 0.5154, "step": 22417 }, { "epoch": 0.9397315112708676, "grad_norm": 1.8093552589416504, "learning_rate": 9.486064386549809e-08, "loss": 0.5081, "step": 22418 }, { "epoch": 0.9397734298858765, "grad_norm": 2.0653443336486816, "learning_rate": 9.472908218889521e-08, "loss": 0.4568, "step": 22419 }, { "epoch": 0.9398153485008856, "grad_norm": 1.851523995399475, "learning_rate": 9.459761093418063e-08, "loss": 0.4666, "step": 22420 }, { "epoch": 0.9398572671158945, "grad_norm": 1.857313632965088, "learning_rate": 9.446623010377743e-08, "loss": 0.4776, "step": 22421 }, { "epoch": 0.9398991857309035, "grad_norm": 2.3131375312805176, "learning_rate": 9.433493970010699e-08, "loss": 0.4696, "step": 22422 }, { "epoch": 0.9399411043459124, "grad_norm": 2.002322196960449, "learning_rate": 9.420373972559072e-08, "loss": 0.427, "step": 22423 }, { "epoch": 0.9399830229609214, "grad_norm": 2.6569883823394775, "learning_rate": 9.407263018264667e-08, "loss": 0.446, "step": 22424 }, { "epoch": 0.9400249415759303, "grad_norm": 1.6726728677749634, "learning_rate": 9.39416110736907e-08, "loss": 0.4544, "step": 22425 }, { "epoch": 0.9400668601909393, "grad_norm": 1.6934716701507568, "learning_rate": 9.381068240113922e-08, "loss": 0.4732, "step": 22426 }, { "epoch": 0.9401087788059482, "grad_norm": 2.091825485229492, "learning_rate": 9.367984416740528e-08, "loss": 0.4814, "step": 22427 }, { "epoch": 0.9401506974209572, "grad_norm": 2.3720293045043945, "learning_rate": 9.354909637490028e-08, "loss": 0.4836, "step": 22428 }, { "epoch": 0.9401926160359662, "grad_norm": 1.8803802728652954, "learning_rate": 9.34184390260351e-08, "loss": 0.4731, "step": 22429 }, { "epoch": 0.9402345346509752, "grad_norm": 1.9264332056045532, "learning_rate": 9.328787212321777e-08, "loss": 0.4476, "step": 22430 }, { "epoch": 0.9402764532659841, "grad_norm": 2.0860848426818848, "learning_rate": 9.315739566885473e-08, "loss": 0.4166, "step": 22431 }, { "epoch": 0.9403183718809931, "grad_norm": 2.987797498703003, "learning_rate": 9.302700966535183e-08, "loss": 0.4554, "step": 22432 }, { "epoch": 0.940360290496002, "grad_norm": 1.637690782546997, "learning_rate": 9.28967141151127e-08, "loss": 0.4408, "step": 22433 }, { "epoch": 0.940402209111011, "grad_norm": 1.7024585008621216, "learning_rate": 9.276650902053818e-08, "loss": 0.436, "step": 22434 }, { "epoch": 0.9404441277260199, "grad_norm": 3.5394279956817627, "learning_rate": 9.263639438402916e-08, "loss": 0.4695, "step": 22435 }, { "epoch": 0.940486046341029, "grad_norm": 2.2045371532440186, "learning_rate": 9.250637020798426e-08, "loss": 0.4562, "step": 22436 }, { "epoch": 0.9405279649560379, "grad_norm": 2.202676296234131, "learning_rate": 9.237643649480044e-08, "loss": 0.4708, "step": 22437 }, { "epoch": 0.9405698835710468, "grad_norm": 2.2607100009918213, "learning_rate": 9.224659324687135e-08, "loss": 0.4855, "step": 22438 }, { "epoch": 0.9406118021860558, "grad_norm": 4.34979772567749, "learning_rate": 9.21168404665923e-08, "loss": 0.5205, "step": 22439 }, { "epoch": 0.9406537208010647, "grad_norm": 4.303403854370117, "learning_rate": 9.198717815635417e-08, "loss": 0.4394, "step": 22440 }, { "epoch": 0.9406956394160737, "grad_norm": 2.1900970935821533, "learning_rate": 9.185760631854723e-08, "loss": 0.4832, "step": 22441 }, { "epoch": 0.9407375580310826, "grad_norm": 2.6562612056732178, "learning_rate": 9.172812495556015e-08, "loss": 0.4817, "step": 22442 }, { "epoch": 0.9407794766460916, "grad_norm": 1.996277928352356, "learning_rate": 9.159873406977937e-08, "loss": 0.5253, "step": 22443 }, { "epoch": 0.9408213952611005, "grad_norm": 2.1428141593933105, "learning_rate": 9.146943366359018e-08, "loss": 0.5268, "step": 22444 }, { "epoch": 0.9408633138761096, "grad_norm": 3.0237107276916504, "learning_rate": 9.134022373937624e-08, "loss": 0.4578, "step": 22445 }, { "epoch": 0.9409052324911185, "grad_norm": 1.7444183826446533, "learning_rate": 9.121110429951952e-08, "loss": 0.4051, "step": 22446 }, { "epoch": 0.9409471511061275, "grad_norm": 1.878196120262146, "learning_rate": 9.108207534639979e-08, "loss": 0.4646, "step": 22447 }, { "epoch": 0.9409890697211364, "grad_norm": 1.8434953689575195, "learning_rate": 9.095313688239571e-08, "loss": 0.4626, "step": 22448 }, { "epoch": 0.9410309883361454, "grad_norm": 3.8610570430755615, "learning_rate": 9.082428890988371e-08, "loss": 0.4793, "step": 22449 }, { "epoch": 0.9410729069511543, "grad_norm": 1.529139518737793, "learning_rate": 9.06955314312391e-08, "loss": 0.4535, "step": 22450 }, { "epoch": 0.9411148255661633, "grad_norm": 1.833460807800293, "learning_rate": 9.056686444883612e-08, "loss": 0.5122, "step": 22451 }, { "epoch": 0.9411567441811722, "grad_norm": 2.147862434387207, "learning_rate": 9.043828796504562e-08, "loss": 0.4611, "step": 22452 }, { "epoch": 0.9411986627961813, "grad_norm": 2.0631163120269775, "learning_rate": 9.030980198223738e-08, "loss": 0.4806, "step": 22453 }, { "epoch": 0.9412405814111902, "grad_norm": 2.1923868656158447, "learning_rate": 9.018140650278118e-08, "loss": 0.4462, "step": 22454 }, { "epoch": 0.9412825000261992, "grad_norm": 1.7279709577560425, "learning_rate": 9.00531015290429e-08, "loss": 0.4666, "step": 22455 }, { "epoch": 0.9413244186412081, "grad_norm": 2.1173925399780273, "learning_rate": 8.992488706338842e-08, "loss": 0.4523, "step": 22456 }, { "epoch": 0.9413663372562171, "grad_norm": 1.6793383359909058, "learning_rate": 8.97967631081803e-08, "loss": 0.4777, "step": 22457 }, { "epoch": 0.941408255871226, "grad_norm": 1.562360167503357, "learning_rate": 8.966872966578055e-08, "loss": 0.4812, "step": 22458 }, { "epoch": 0.941450174486235, "grad_norm": 1.9621130228042603, "learning_rate": 8.954078673854949e-08, "loss": 0.4809, "step": 22459 }, { "epoch": 0.9414920931012439, "grad_norm": 2.035226345062256, "learning_rate": 8.941293432884579e-08, "loss": 0.4499, "step": 22460 }, { "epoch": 0.941534011716253, "grad_norm": 1.9116133451461792, "learning_rate": 8.928517243902535e-08, "loss": 0.4704, "step": 22461 }, { "epoch": 0.9415759303312619, "grad_norm": 1.6948904991149902, "learning_rate": 8.915750107144405e-08, "loss": 0.4509, "step": 22462 }, { "epoch": 0.9416178489462708, "grad_norm": 2.009950637817383, "learning_rate": 8.902992022845558e-08, "loss": 0.4691, "step": 22463 }, { "epoch": 0.9416597675612798, "grad_norm": 2.3997654914855957, "learning_rate": 8.890242991241082e-08, "loss": 0.4908, "step": 22464 }, { "epoch": 0.9417016861762887, "grad_norm": 1.9602733850479126, "learning_rate": 8.877503012566124e-08, "loss": 0.4623, "step": 22465 }, { "epoch": 0.9417436047912977, "grad_norm": 1.8442691564559937, "learning_rate": 8.864772087055384e-08, "loss": 0.452, "step": 22466 }, { "epoch": 0.9417855234063066, "grad_norm": 2.0205233097076416, "learning_rate": 8.85205021494362e-08, "loss": 0.4675, "step": 22467 }, { "epoch": 0.9418274420213156, "grad_norm": 2.973325490951538, "learning_rate": 8.839337396465253e-08, "loss": 0.4007, "step": 22468 }, { "epoch": 0.9418693606363245, "grad_norm": 1.7183852195739746, "learning_rate": 8.826633631854764e-08, "loss": 0.4425, "step": 22469 }, { "epoch": 0.9419112792513336, "grad_norm": 1.9563356637954712, "learning_rate": 8.813938921346187e-08, "loss": 0.5218, "step": 22470 }, { "epoch": 0.9419531978663425, "grad_norm": 1.942633867263794, "learning_rate": 8.801253265173671e-08, "loss": 0.4509, "step": 22471 }, { "epoch": 0.9419951164813515, "grad_norm": 1.9117529392242432, "learning_rate": 8.78857666357097e-08, "loss": 0.475, "step": 22472 }, { "epoch": 0.9420370350963604, "grad_norm": 2.796454906463623, "learning_rate": 8.77590911677173e-08, "loss": 0.5171, "step": 22473 }, { "epoch": 0.9420789537113694, "grad_norm": 2.2222964763641357, "learning_rate": 8.7632506250096e-08, "loss": 0.4708, "step": 22474 }, { "epoch": 0.9421208723263783, "grad_norm": 5.568567752838135, "learning_rate": 8.750601188517782e-08, "loss": 0.459, "step": 22475 }, { "epoch": 0.9421627909413873, "grad_norm": 1.6968162059783936, "learning_rate": 8.737960807529477e-08, "loss": 0.5013, "step": 22476 }, { "epoch": 0.9422047095563962, "grad_norm": 1.8976045846939087, "learning_rate": 8.725329482277778e-08, "loss": 0.5245, "step": 22477 }, { "epoch": 0.9422466281714053, "grad_norm": 1.795719861984253, "learning_rate": 8.712707212995387e-08, "loss": 0.4775, "step": 22478 }, { "epoch": 0.9422885467864142, "grad_norm": 2.250793218612671, "learning_rate": 8.70009399991506e-08, "loss": 0.4709, "step": 22479 }, { "epoch": 0.9423304654014232, "grad_norm": 2.1717910766601562, "learning_rate": 8.687489843269392e-08, "loss": 0.4883, "step": 22480 }, { "epoch": 0.9423723840164321, "grad_norm": 1.7055219411849976, "learning_rate": 8.674894743290529e-08, "loss": 0.4421, "step": 22481 }, { "epoch": 0.9424143026314411, "grad_norm": 3.132999897003174, "learning_rate": 8.66230870021073e-08, "loss": 0.4642, "step": 22482 }, { "epoch": 0.94245622124645, "grad_norm": 1.9164481163024902, "learning_rate": 8.649731714262089e-08, "loss": 0.4678, "step": 22483 }, { "epoch": 0.942498139861459, "grad_norm": 1.9106159210205078, "learning_rate": 8.637163785676417e-08, "loss": 0.4889, "step": 22484 }, { "epoch": 0.9425400584764679, "grad_norm": 2.957301139831543, "learning_rate": 8.624604914685253e-08, "loss": 0.5038, "step": 22485 }, { "epoch": 0.942581977091477, "grad_norm": 2.3571457862854004, "learning_rate": 8.612055101520245e-08, "loss": 0.4338, "step": 22486 }, { "epoch": 0.9426238957064859, "grad_norm": 2.192725419998169, "learning_rate": 8.599514346412596e-08, "loss": 0.4135, "step": 22487 }, { "epoch": 0.9426658143214948, "grad_norm": 1.7627159357070923, "learning_rate": 8.586982649593566e-08, "loss": 0.506, "step": 22488 }, { "epoch": 0.9427077329365038, "grad_norm": 2.1516032218933105, "learning_rate": 8.57446001129425e-08, "loss": 0.4914, "step": 22489 }, { "epoch": 0.9427496515515127, "grad_norm": 1.9903138875961304, "learning_rate": 8.561946431745294e-08, "loss": 0.4303, "step": 22490 }, { "epoch": 0.9427915701665217, "grad_norm": 2.0811846256256104, "learning_rate": 8.549441911177514e-08, "loss": 0.4623, "step": 22491 }, { "epoch": 0.9428334887815306, "grad_norm": 2.0352213382720947, "learning_rate": 8.536946449821338e-08, "loss": 0.4269, "step": 22492 }, { "epoch": 0.9428754073965396, "grad_norm": 1.9305744171142578, "learning_rate": 8.524460047907135e-08, "loss": 0.5452, "step": 22493 }, { "epoch": 0.9429173260115485, "grad_norm": 1.9856359958648682, "learning_rate": 8.511982705665112e-08, "loss": 0.454, "step": 22494 }, { "epoch": 0.9429592446265576, "grad_norm": 2.335545063018799, "learning_rate": 8.499514423325194e-08, "loss": 0.4573, "step": 22495 }, { "epoch": 0.9430011632415665, "grad_norm": 1.858832597732544, "learning_rate": 8.4870552011172e-08, "loss": 0.5167, "step": 22496 }, { "epoch": 0.9430430818565755, "grad_norm": 2.4298532009124756, "learning_rate": 8.474605039270889e-08, "loss": 0.5037, "step": 22497 }, { "epoch": 0.9430850004715844, "grad_norm": 3.4076316356658936, "learning_rate": 8.462163938015744e-08, "loss": 0.5032, "step": 22498 }, { "epoch": 0.9431269190865934, "grad_norm": 1.9321272373199463, "learning_rate": 8.449731897581026e-08, "loss": 0.4886, "step": 22499 }, { "epoch": 0.9431688377016023, "grad_norm": 2.1817471981048584, "learning_rate": 8.437308918195996e-08, "loss": 0.477, "step": 22500 }, { "epoch": 0.9432107563166113, "grad_norm": 3.4160726070404053, "learning_rate": 8.424895000089584e-08, "loss": 0.4809, "step": 22501 }, { "epoch": 0.9432526749316202, "grad_norm": 2.5134589672088623, "learning_rate": 8.41249014349066e-08, "loss": 0.5007, "step": 22502 }, { "epoch": 0.9432945935466293, "grad_norm": 1.8262251615524292, "learning_rate": 8.400094348627875e-08, "loss": 0.4603, "step": 22503 }, { "epoch": 0.9433365121616382, "grad_norm": 1.7614535093307495, "learning_rate": 8.387707615729767e-08, "loss": 0.4344, "step": 22504 }, { "epoch": 0.9433784307766472, "grad_norm": 1.8284010887145996, "learning_rate": 8.375329945024602e-08, "loss": 0.4624, "step": 22505 }, { "epoch": 0.9434203493916561, "grad_norm": 2.1920719146728516, "learning_rate": 8.362961336740582e-08, "loss": 0.4582, "step": 22506 }, { "epoch": 0.9434622680066651, "grad_norm": 2.01946759223938, "learning_rate": 8.350601791105694e-08, "loss": 0.5359, "step": 22507 }, { "epoch": 0.943504186621674, "grad_norm": 1.8810099363327026, "learning_rate": 8.338251308347755e-08, "loss": 0.5128, "step": 22508 }, { "epoch": 0.943546105236683, "grad_norm": 2.034651041030884, "learning_rate": 8.325909888694472e-08, "loss": 0.4786, "step": 22509 }, { "epoch": 0.9435880238516919, "grad_norm": 1.9277050495147705, "learning_rate": 8.313577532373385e-08, "loss": 0.4498, "step": 22510 }, { "epoch": 0.943629942466701, "grad_norm": 2.0313456058502197, "learning_rate": 8.3012542396117e-08, "loss": 0.5087, "step": 22511 }, { "epoch": 0.9436718610817099, "grad_norm": 1.8018214702606201, "learning_rate": 8.288940010636626e-08, "loss": 0.5453, "step": 22512 }, { "epoch": 0.9437137796967188, "grad_norm": 1.8561123609542847, "learning_rate": 8.276634845675257e-08, "loss": 0.4291, "step": 22513 }, { "epoch": 0.9437556983117278, "grad_norm": 1.954662799835205, "learning_rate": 8.264338744954248e-08, "loss": 0.4732, "step": 22514 }, { "epoch": 0.9437976169267367, "grad_norm": 2.5847344398498535, "learning_rate": 8.25205170870036e-08, "loss": 0.4661, "step": 22515 }, { "epoch": 0.9438395355417457, "grad_norm": 1.7596567869186401, "learning_rate": 8.239773737140078e-08, "loss": 0.4628, "step": 22516 }, { "epoch": 0.9438814541567546, "grad_norm": 1.925991415977478, "learning_rate": 8.227504830499777e-08, "loss": 0.4958, "step": 22517 }, { "epoch": 0.9439233727717636, "grad_norm": 1.6675482988357544, "learning_rate": 8.2152449890055e-08, "loss": 0.4524, "step": 22518 }, { "epoch": 0.9439652913867725, "grad_norm": 1.7370437383651733, "learning_rate": 8.202994212883342e-08, "loss": 0.4338, "step": 22519 }, { "epoch": 0.9440072100017816, "grad_norm": 2.139338254928589, "learning_rate": 8.190752502359122e-08, "loss": 0.4716, "step": 22520 }, { "epoch": 0.9440491286167905, "grad_norm": 2.5136172771453857, "learning_rate": 8.17851985765844e-08, "loss": 0.5097, "step": 22521 }, { "epoch": 0.9440910472317995, "grad_norm": 1.9199451208114624, "learning_rate": 8.166296279006891e-08, "loss": 0.4672, "step": 22522 }, { "epoch": 0.9441329658468084, "grad_norm": 2.3013339042663574, "learning_rate": 8.154081766629629e-08, "loss": 0.4574, "step": 22523 }, { "epoch": 0.9441748844618174, "grad_norm": 2.032396078109741, "learning_rate": 8.141876320751973e-08, "loss": 0.4636, "step": 22524 }, { "epoch": 0.9442168030768263, "grad_norm": 1.7316417694091797, "learning_rate": 8.129679941598856e-08, "loss": 0.4891, "step": 22525 }, { "epoch": 0.9442587216918353, "grad_norm": 1.584686517715454, "learning_rate": 8.11749262939504e-08, "loss": 0.411, "step": 22526 }, { "epoch": 0.9443006403068442, "grad_norm": 2.073456287384033, "learning_rate": 8.105314384365292e-08, "loss": 0.5012, "step": 22527 }, { "epoch": 0.9443425589218533, "grad_norm": 1.7944581508636475, "learning_rate": 8.093145206734043e-08, "loss": 0.4983, "step": 22528 }, { "epoch": 0.9443844775368622, "grad_norm": 1.6532862186431885, "learning_rate": 8.080985096725614e-08, "loss": 0.418, "step": 22529 }, { "epoch": 0.9444263961518712, "grad_norm": 1.7184319496154785, "learning_rate": 8.068834054564212e-08, "loss": 0.5041, "step": 22530 }, { "epoch": 0.9444683147668801, "grad_norm": 2.132373809814453, "learning_rate": 8.056692080473716e-08, "loss": 0.5334, "step": 22531 }, { "epoch": 0.9445102333818891, "grad_norm": 2.6912806034088135, "learning_rate": 8.044559174678057e-08, "loss": 0.4872, "step": 22532 }, { "epoch": 0.944552151996898, "grad_norm": 1.6800868511199951, "learning_rate": 8.032435337400779e-08, "loss": 0.4598, "step": 22533 }, { "epoch": 0.944594070611907, "grad_norm": 2.0001487731933594, "learning_rate": 8.02032056886548e-08, "loss": 0.4634, "step": 22534 }, { "epoch": 0.9446359892269159, "grad_norm": 2.8622114658355713, "learning_rate": 8.008214869295428e-08, "loss": 0.4643, "step": 22535 }, { "epoch": 0.944677907841925, "grad_norm": 1.8652015924453735, "learning_rate": 7.996118238913719e-08, "loss": 0.4717, "step": 22536 }, { "epoch": 0.9447198264569339, "grad_norm": 1.9913491010665894, "learning_rate": 7.98403067794351e-08, "loss": 0.5017, "step": 22537 }, { "epoch": 0.9447617450719428, "grad_norm": 1.8702423572540283, "learning_rate": 7.971952186607401e-08, "loss": 0.4193, "step": 22538 }, { "epoch": 0.9448036636869518, "grad_norm": 14.935493469238281, "learning_rate": 7.959882765128213e-08, "loss": 0.4302, "step": 22539 }, { "epoch": 0.9448455823019607, "grad_norm": 2.063774585723877, "learning_rate": 7.947822413728323e-08, "loss": 0.5144, "step": 22540 }, { "epoch": 0.9448875009169697, "grad_norm": 2.00264573097229, "learning_rate": 7.93577113263011e-08, "loss": 0.4433, "step": 22541 }, { "epoch": 0.9449294195319786, "grad_norm": 1.852858543395996, "learning_rate": 7.923728922055674e-08, "loss": 0.4851, "step": 22542 }, { "epoch": 0.9449713381469876, "grad_norm": 2.189326763153076, "learning_rate": 7.911695782227003e-08, "loss": 0.499, "step": 22543 }, { "epoch": 0.9450132567619965, "grad_norm": 3.2433629035949707, "learning_rate": 7.899671713365976e-08, "loss": 0.4145, "step": 22544 }, { "epoch": 0.9450551753770056, "grad_norm": 2.283933162689209, "learning_rate": 7.887656715694192e-08, "loss": 0.4888, "step": 22545 }, { "epoch": 0.9450970939920145, "grad_norm": 2.0235252380371094, "learning_rate": 7.875650789433086e-08, "loss": 0.4841, "step": 22546 }, { "epoch": 0.9451390126070235, "grad_norm": 2.0292065143585205, "learning_rate": 7.863653934803983e-08, "loss": 0.439, "step": 22547 }, { "epoch": 0.9451809312220324, "grad_norm": 2.0455055236816406, "learning_rate": 7.851666152028148e-08, "loss": 0.4674, "step": 22548 }, { "epoch": 0.9452228498370414, "grad_norm": 2.4450130462646484, "learning_rate": 7.83968744132646e-08, "loss": 0.4066, "step": 22549 }, { "epoch": 0.9452647684520503, "grad_norm": 2.1775100231170654, "learning_rate": 7.827717802919743e-08, "loss": 0.4907, "step": 22550 }, { "epoch": 0.9453066870670593, "grad_norm": 1.7346234321594238, "learning_rate": 7.815757237028654e-08, "loss": 0.517, "step": 22551 }, { "epoch": 0.9453486056820682, "grad_norm": 4.11234188079834, "learning_rate": 7.803805743873572e-08, "loss": 0.5103, "step": 22552 }, { "epoch": 0.9453905242970773, "grad_norm": 2.0319244861602783, "learning_rate": 7.791863323674931e-08, "loss": 0.4913, "step": 22553 }, { "epoch": 0.9454324429120862, "grad_norm": 2.1173720359802246, "learning_rate": 7.779929976652889e-08, "loss": 0.4913, "step": 22554 }, { "epoch": 0.9454743615270952, "grad_norm": 4.936809062957764, "learning_rate": 7.76800570302727e-08, "loss": 0.4634, "step": 22555 }, { "epoch": 0.9455162801421041, "grad_norm": 1.9461315870285034, "learning_rate": 7.75609050301801e-08, "loss": 0.497, "step": 22556 }, { "epoch": 0.9455581987571131, "grad_norm": 1.9818017482757568, "learning_rate": 7.744184376844765e-08, "loss": 0.4987, "step": 22557 }, { "epoch": 0.945600117372122, "grad_norm": 1.9274234771728516, "learning_rate": 7.732287324726917e-08, "loss": 0.4493, "step": 22558 }, { "epoch": 0.945642035987131, "grad_norm": 1.803601622581482, "learning_rate": 7.720399346883845e-08, "loss": 0.4356, "step": 22559 }, { "epoch": 0.9456839546021399, "grad_norm": 1.7355619668960571, "learning_rate": 7.708520443534651e-08, "loss": 0.4793, "step": 22560 }, { "epoch": 0.945725873217149, "grad_norm": 1.808945655822754, "learning_rate": 7.696650614898272e-08, "loss": 0.4596, "step": 22561 }, { "epoch": 0.9457677918321579, "grad_norm": 1.9570581912994385, "learning_rate": 7.684789861193531e-08, "loss": 0.4909, "step": 22562 }, { "epoch": 0.9458097104471668, "grad_norm": 2.27431321144104, "learning_rate": 7.672938182639145e-08, "loss": 0.4246, "step": 22563 }, { "epoch": 0.9458516290621758, "grad_norm": 1.7079583406448364, "learning_rate": 7.661095579453437e-08, "loss": 0.4074, "step": 22564 }, { "epoch": 0.9458935476771847, "grad_norm": 1.7237582206726074, "learning_rate": 7.649262051854844e-08, "loss": 0.4866, "step": 22565 }, { "epoch": 0.9459354662921937, "grad_norm": 2.309046506881714, "learning_rate": 7.63743760006147e-08, "loss": 0.5233, "step": 22566 }, { "epoch": 0.9459773849072026, "grad_norm": 1.7400985956192017, "learning_rate": 7.625622224291251e-08, "loss": 0.4747, "step": 22567 }, { "epoch": 0.9460193035222116, "grad_norm": 1.860258936882019, "learning_rate": 7.613815924762014e-08, "loss": 0.459, "step": 22568 }, { "epoch": 0.9460612221372205, "grad_norm": 1.9309581518173218, "learning_rate": 7.602018701691415e-08, "loss": 0.4508, "step": 22569 }, { "epoch": 0.9461031407522296, "grad_norm": 1.9140645265579224, "learning_rate": 7.590230555296785e-08, "loss": 0.5283, "step": 22570 }, { "epoch": 0.9461450593672385, "grad_norm": 1.912340760231018, "learning_rate": 7.578451485795557e-08, "loss": 0.4469, "step": 22571 }, { "epoch": 0.9461869779822475, "grad_norm": 1.6826298236846924, "learning_rate": 7.566681493404838e-08, "loss": 0.4003, "step": 22572 }, { "epoch": 0.9462288965972564, "grad_norm": 1.6746139526367188, "learning_rate": 7.554920578341562e-08, "loss": 0.4638, "step": 22573 }, { "epoch": 0.9462708152122654, "grad_norm": 1.8469492197036743, "learning_rate": 7.543168740822504e-08, "loss": 0.4956, "step": 22574 }, { "epoch": 0.9463127338272743, "grad_norm": 1.9337759017944336, "learning_rate": 7.531425981064433e-08, "loss": 0.5157, "step": 22575 }, { "epoch": 0.9463546524422833, "grad_norm": 2.8972837924957275, "learning_rate": 7.519692299283621e-08, "loss": 0.4647, "step": 22576 }, { "epoch": 0.9463965710572922, "grad_norm": 1.8693432807922363, "learning_rate": 7.507967695696505e-08, "loss": 0.4784, "step": 22577 }, { "epoch": 0.9464384896723013, "grad_norm": 2.5997135639190674, "learning_rate": 7.49625217051908e-08, "loss": 0.5068, "step": 22578 }, { "epoch": 0.9464804082873102, "grad_norm": 1.7359263896942139, "learning_rate": 7.484545723967451e-08, "loss": 0.4449, "step": 22579 }, { "epoch": 0.9465223269023192, "grad_norm": 1.9504599571228027, "learning_rate": 7.472848356257278e-08, "loss": 0.4571, "step": 22580 }, { "epoch": 0.9465642455173281, "grad_norm": 2.3926467895507812, "learning_rate": 7.461160067604334e-08, "loss": 0.4681, "step": 22581 }, { "epoch": 0.9466061641323371, "grad_norm": 1.945132851600647, "learning_rate": 7.449480858223889e-08, "loss": 0.4468, "step": 22582 }, { "epoch": 0.946648082747346, "grad_norm": 1.7457326650619507, "learning_rate": 7.437810728331384e-08, "loss": 0.5456, "step": 22583 }, { "epoch": 0.946690001362355, "grad_norm": 1.8544795513153076, "learning_rate": 7.42614967814187e-08, "loss": 0.4864, "step": 22584 }, { "epoch": 0.946731919977364, "grad_norm": 2.250539779663086, "learning_rate": 7.414497707870338e-08, "loss": 0.442, "step": 22585 }, { "epoch": 0.946773838592373, "grad_norm": 2.0424063205718994, "learning_rate": 7.402854817731564e-08, "loss": 0.4913, "step": 22586 }, { "epoch": 0.9468157572073819, "grad_norm": 2.017641305923462, "learning_rate": 7.391221007940153e-08, "loss": 0.483, "step": 22587 }, { "epoch": 0.9468576758223908, "grad_norm": 1.7124090194702148, "learning_rate": 7.379596278710655e-08, "loss": 0.4008, "step": 22588 }, { "epoch": 0.9468995944373998, "grad_norm": 1.9062365293502808, "learning_rate": 7.367980630257177e-08, "loss": 0.5052, "step": 22589 }, { "epoch": 0.9469415130524087, "grad_norm": 1.814876675605774, "learning_rate": 7.356374062793936e-08, "loss": 0.51, "step": 22590 }, { "epoch": 0.9469834316674177, "grad_norm": 1.8543647527694702, "learning_rate": 7.344776576534929e-08, "loss": 0.4644, "step": 22591 }, { "epoch": 0.9470253502824266, "grad_norm": 1.8241785764694214, "learning_rate": 7.333188171693817e-08, "loss": 0.5255, "step": 22592 }, { "epoch": 0.9470672688974356, "grad_norm": 2.0663843154907227, "learning_rate": 7.321608848484374e-08, "loss": 0.5293, "step": 22593 }, { "epoch": 0.9471091875124446, "grad_norm": 1.7167878150939941, "learning_rate": 7.31003860711993e-08, "loss": 0.4337, "step": 22594 }, { "epoch": 0.9471511061274536, "grad_norm": 1.8738915920257568, "learning_rate": 7.298477447813757e-08, "loss": 0.4405, "step": 22595 }, { "epoch": 0.9471930247424625, "grad_norm": 2.0005273818969727, "learning_rate": 7.286925370779074e-08, "loss": 0.4733, "step": 22596 }, { "epoch": 0.9472349433574715, "grad_norm": 2.089034080505371, "learning_rate": 7.275382376228768e-08, "loss": 0.44, "step": 22597 }, { "epoch": 0.9472768619724804, "grad_norm": 1.8876805305480957, "learning_rate": 7.263848464375556e-08, "loss": 0.5041, "step": 22598 }, { "epoch": 0.9473187805874894, "grad_norm": 2.117943286895752, "learning_rate": 7.252323635432213e-08, "loss": 0.4546, "step": 22599 }, { "epoch": 0.9473606992024983, "grad_norm": 1.9823801517486572, "learning_rate": 7.240807889610957e-08, "loss": 0.4698, "step": 22600 }, { "epoch": 0.9474026178175073, "grad_norm": 2.022171974182129, "learning_rate": 7.22930122712423e-08, "loss": 0.5189, "step": 22601 }, { "epoch": 0.9474445364325162, "grad_norm": 2.2611238956451416, "learning_rate": 7.217803648184141e-08, "loss": 0.479, "step": 22602 }, { "epoch": 0.9474864550475253, "grad_norm": 1.7327911853790283, "learning_rate": 7.206315153002574e-08, "loss": 0.5187, "step": 22603 }, { "epoch": 0.9475283736625342, "grad_norm": 1.987001895904541, "learning_rate": 7.194835741791306e-08, "loss": 0.5002, "step": 22604 }, { "epoch": 0.9475702922775432, "grad_norm": 1.7942276000976562, "learning_rate": 7.183365414761944e-08, "loss": 0.478, "step": 22605 }, { "epoch": 0.9476122108925521, "grad_norm": 3.061373472213745, "learning_rate": 7.171904172125987e-08, "loss": 0.4794, "step": 22606 }, { "epoch": 0.9476541295075611, "grad_norm": 8.368473052978516, "learning_rate": 7.160452014094654e-08, "loss": 0.464, "step": 22607 }, { "epoch": 0.94769604812257, "grad_norm": 2.1904265880584717, "learning_rate": 7.149008940879055e-08, "loss": 0.5155, "step": 22608 }, { "epoch": 0.947737966737579, "grad_norm": 1.7728551626205444, "learning_rate": 7.137574952690074e-08, "loss": 0.448, "step": 22609 }, { "epoch": 0.947779885352588, "grad_norm": 2.2181472778320312, "learning_rate": 7.126150049738601e-08, "loss": 0.4802, "step": 22610 }, { "epoch": 0.947821803967597, "grad_norm": 2.0352492332458496, "learning_rate": 7.114734232235188e-08, "loss": 0.4419, "step": 22611 }, { "epoch": 0.9478637225826059, "grad_norm": 1.7614014148712158, "learning_rate": 7.103327500390168e-08, "loss": 0.4765, "step": 22612 }, { "epoch": 0.9479056411976148, "grad_norm": 1.9136526584625244, "learning_rate": 7.091929854413926e-08, "loss": 0.5344, "step": 22613 }, { "epoch": 0.9479475598126238, "grad_norm": 1.8586491346359253, "learning_rate": 7.080541294516574e-08, "loss": 0.5084, "step": 22614 }, { "epoch": 0.9479894784276327, "grad_norm": 1.8501536846160889, "learning_rate": 7.069161820908e-08, "loss": 0.4637, "step": 22615 }, { "epoch": 0.9480313970426417, "grad_norm": 2.1460185050964355, "learning_rate": 7.057791433797978e-08, "loss": 0.4727, "step": 22616 }, { "epoch": 0.9480733156576506, "grad_norm": 2.245983839035034, "learning_rate": 7.046430133396065e-08, "loss": 0.433, "step": 22617 }, { "epoch": 0.9481152342726596, "grad_norm": 1.905297875404358, "learning_rate": 7.035077919911759e-08, "loss": 0.4523, "step": 22618 }, { "epoch": 0.9481571528876686, "grad_norm": 1.8849859237670898, "learning_rate": 7.023734793554226e-08, "loss": 0.4967, "step": 22619 }, { "epoch": 0.9481990715026776, "grad_norm": 1.9731688499450684, "learning_rate": 7.012400754532689e-08, "loss": 0.4811, "step": 22620 }, { "epoch": 0.9482409901176865, "grad_norm": 1.9866878986358643, "learning_rate": 7.001075803055978e-08, "loss": 0.5088, "step": 22621 }, { "epoch": 0.9482829087326955, "grad_norm": 2.872619390487671, "learning_rate": 6.989759939332874e-08, "loss": 0.4971, "step": 22622 }, { "epoch": 0.9483248273477044, "grad_norm": 1.864209771156311, "learning_rate": 6.978453163571986e-08, "loss": 0.5214, "step": 22623 }, { "epoch": 0.9483667459627134, "grad_norm": 1.94536554813385, "learning_rate": 6.967155475981758e-08, "loss": 0.466, "step": 22624 }, { "epoch": 0.9484086645777223, "grad_norm": 1.758325457572937, "learning_rate": 6.955866876770412e-08, "loss": 0.4348, "step": 22625 }, { "epoch": 0.9484505831927313, "grad_norm": 1.9415247440338135, "learning_rate": 6.94458736614606e-08, "loss": 0.4218, "step": 22626 }, { "epoch": 0.9484925018077403, "grad_norm": 3.2993052005767822, "learning_rate": 6.933316944316592e-08, "loss": 0.4764, "step": 22627 }, { "epoch": 0.9485344204227493, "grad_norm": 1.886743187904358, "learning_rate": 6.922055611489786e-08, "loss": 0.4691, "step": 22628 }, { "epoch": 0.9485763390377582, "grad_norm": 1.9607986211776733, "learning_rate": 6.910803367873197e-08, "loss": 0.4565, "step": 22629 }, { "epoch": 0.9486182576527672, "grad_norm": 1.7688325643539429, "learning_rate": 6.899560213674273e-08, "loss": 0.4694, "step": 22630 }, { "epoch": 0.9486601762677761, "grad_norm": 2.1972172260284424, "learning_rate": 6.888326149100289e-08, "loss": 0.4842, "step": 22631 }, { "epoch": 0.9487020948827851, "grad_norm": 1.710889220237732, "learning_rate": 6.877101174358303e-08, "loss": 0.4864, "step": 22632 }, { "epoch": 0.948744013497794, "grad_norm": 1.6149319410324097, "learning_rate": 6.865885289655205e-08, "loss": 0.4161, "step": 22633 }, { "epoch": 0.948785932112803, "grad_norm": 1.9265791177749634, "learning_rate": 6.854678495197831e-08, "loss": 0.4526, "step": 22634 }, { "epoch": 0.948827850727812, "grad_norm": 2.131740093231201, "learning_rate": 6.843480791192736e-08, "loss": 0.4812, "step": 22635 }, { "epoch": 0.948869769342821, "grad_norm": 2.0906779766082764, "learning_rate": 6.8322921778462e-08, "loss": 0.4323, "step": 22636 }, { "epoch": 0.9489116879578299, "grad_norm": 2.6914777755737305, "learning_rate": 6.82111265536467e-08, "loss": 0.4577, "step": 22637 }, { "epoch": 0.9489536065728388, "grad_norm": 3.9042913913726807, "learning_rate": 6.809942223954036e-08, "loss": 0.5169, "step": 22638 }, { "epoch": 0.9489955251878478, "grad_norm": 2.040966033935547, "learning_rate": 6.798780883820355e-08, "loss": 0.5004, "step": 22639 }, { "epoch": 0.9490374438028567, "grad_norm": 2.428373336791992, "learning_rate": 6.787628635169296e-08, "loss": 0.4747, "step": 22640 }, { "epoch": 0.9490793624178657, "grad_norm": 2.11376953125, "learning_rate": 6.776485478206474e-08, "loss": 0.4105, "step": 22641 }, { "epoch": 0.9491212810328746, "grad_norm": 1.952586054801941, "learning_rate": 6.765351413137277e-08, "loss": 0.5232, "step": 22642 }, { "epoch": 0.9491631996478836, "grad_norm": 4.888535499572754, "learning_rate": 6.754226440166934e-08, "loss": 0.4776, "step": 22643 }, { "epoch": 0.9492051182628926, "grad_norm": 1.886183500289917, "learning_rate": 6.74311055950061e-08, "loss": 0.46, "step": 22644 }, { "epoch": 0.9492470368779016, "grad_norm": 1.5468788146972656, "learning_rate": 6.732003771343032e-08, "loss": 0.4172, "step": 22645 }, { "epoch": 0.9492889554929105, "grad_norm": 2.4329681396484375, "learning_rate": 6.720906075899148e-08, "loss": 0.4555, "step": 22646 }, { "epoch": 0.9493308741079195, "grad_norm": 1.7524495124816895, "learning_rate": 6.709817473373348e-08, "loss": 0.5376, "step": 22647 }, { "epoch": 0.9493727927229284, "grad_norm": 1.9760602712631226, "learning_rate": 6.698737963970082e-08, "loss": 0.4019, "step": 22648 }, { "epoch": 0.9494147113379374, "grad_norm": 1.7648528814315796, "learning_rate": 6.687667547893684e-08, "loss": 0.5262, "step": 22649 }, { "epoch": 0.9494566299529463, "grad_norm": 1.7426257133483887, "learning_rate": 6.676606225348103e-08, "loss": 0.5192, "step": 22650 }, { "epoch": 0.9494985485679553, "grad_norm": 2.1855697631835938, "learning_rate": 6.665553996537288e-08, "loss": 0.5291, "step": 22651 }, { "epoch": 0.9495404671829643, "grad_norm": 2.801754951477051, "learning_rate": 6.65451086166502e-08, "loss": 0.5249, "step": 22652 }, { "epoch": 0.9495823857979733, "grad_norm": 2.483616828918457, "learning_rate": 6.643476820934802e-08, "loss": 0.5359, "step": 22653 }, { "epoch": 0.9496243044129822, "grad_norm": 1.9278299808502197, "learning_rate": 6.632451874550028e-08, "loss": 0.4208, "step": 22654 }, { "epoch": 0.9496662230279912, "grad_norm": 1.8640650510787964, "learning_rate": 6.621436022713923e-08, "loss": 0.5102, "step": 22655 }, { "epoch": 0.9497081416430001, "grad_norm": 1.8081196546554565, "learning_rate": 6.610429265629604e-08, "loss": 0.485, "step": 22656 }, { "epoch": 0.9497500602580091, "grad_norm": 1.9541913270950317, "learning_rate": 6.599431603499907e-08, "loss": 0.4701, "step": 22657 }, { "epoch": 0.949791978873018, "grad_norm": 16.76724624633789, "learning_rate": 6.588443036527558e-08, "loss": 0.4781, "step": 22658 }, { "epoch": 0.949833897488027, "grad_norm": 1.9142848253250122, "learning_rate": 6.577463564915176e-08, "loss": 0.4531, "step": 22659 }, { "epoch": 0.949875816103036, "grad_norm": 2.01627779006958, "learning_rate": 6.566493188865097e-08, "loss": 0.4932, "step": 22660 }, { "epoch": 0.949917734718045, "grad_norm": 2.107501745223999, "learning_rate": 6.555531908579605e-08, "loss": 0.4912, "step": 22661 }, { "epoch": 0.9499596533330539, "grad_norm": 1.880171298980713, "learning_rate": 6.544579724260703e-08, "loss": 0.5406, "step": 22662 }, { "epoch": 0.9500015719480628, "grad_norm": 3.110931873321533, "learning_rate": 6.533636636110285e-08, "loss": 0.5194, "step": 22663 }, { "epoch": 0.9500434905630718, "grad_norm": 2.1003568172454834, "learning_rate": 6.522702644330081e-08, "loss": 0.4494, "step": 22664 }, { "epoch": 0.9500854091780807, "grad_norm": 2.2666425704956055, "learning_rate": 6.511777749121651e-08, "loss": 0.4548, "step": 22665 }, { "epoch": 0.9501273277930897, "grad_norm": 2.0065295696258545, "learning_rate": 6.500861950686333e-08, "loss": 0.4861, "step": 22666 }, { "epoch": 0.9501692464080986, "grad_norm": 2.022322416305542, "learning_rate": 6.48995524922541e-08, "loss": 0.4789, "step": 22667 }, { "epoch": 0.9502111650231077, "grad_norm": 1.7074925899505615, "learning_rate": 6.47905764493989e-08, "loss": 0.4805, "step": 22668 }, { "epoch": 0.9502530836381166, "grad_norm": 2.3908934593200684, "learning_rate": 6.468169138030666e-08, "loss": 0.4579, "step": 22669 }, { "epoch": 0.9502950022531256, "grad_norm": 2.2322630882263184, "learning_rate": 6.457289728698468e-08, "loss": 0.4787, "step": 22670 }, { "epoch": 0.9503369208681345, "grad_norm": 3.061173677444458, "learning_rate": 6.446419417143857e-08, "loss": 0.4851, "step": 22671 }, { "epoch": 0.9503788394831435, "grad_norm": 2.049588918685913, "learning_rate": 6.435558203567171e-08, "loss": 0.4926, "step": 22672 }, { "epoch": 0.9504207580981524, "grad_norm": 2.1439547538757324, "learning_rate": 6.424706088168641e-08, "loss": 0.4582, "step": 22673 }, { "epoch": 0.9504626767131614, "grad_norm": 2.490588665008545, "learning_rate": 6.413863071148274e-08, "loss": 0.4883, "step": 22674 }, { "epoch": 0.9505045953281703, "grad_norm": 2.0197129249572754, "learning_rate": 6.40302915270602e-08, "loss": 0.4445, "step": 22675 }, { "epoch": 0.9505465139431793, "grad_norm": 1.6704277992248535, "learning_rate": 6.392204333041551e-08, "loss": 0.4578, "step": 22676 }, { "epoch": 0.9505884325581883, "grad_norm": 1.8169163465499878, "learning_rate": 6.381388612354322e-08, "loss": 0.4443, "step": 22677 }, { "epoch": 0.9506303511731973, "grad_norm": 1.6910368204116821, "learning_rate": 6.370581990843838e-08, "loss": 0.4877, "step": 22678 }, { "epoch": 0.9506722697882062, "grad_norm": 1.9787853956222534, "learning_rate": 6.359784468709273e-08, "loss": 0.4676, "step": 22679 }, { "epoch": 0.9507141884032152, "grad_norm": 1.9637115001678467, "learning_rate": 6.348996046149635e-08, "loss": 0.4897, "step": 22680 }, { "epoch": 0.9507561070182241, "grad_norm": 1.871485948562622, "learning_rate": 6.338216723363821e-08, "loss": 0.4584, "step": 22681 }, { "epoch": 0.9507980256332331, "grad_norm": 2.0658979415893555, "learning_rate": 6.327446500550505e-08, "loss": 0.4691, "step": 22682 }, { "epoch": 0.950839944248242, "grad_norm": 1.9685660600662231, "learning_rate": 6.316685377908193e-08, "loss": 0.4412, "step": 22683 }, { "epoch": 0.950881862863251, "grad_norm": 1.9728821516036987, "learning_rate": 6.305933355635341e-08, "loss": 0.4718, "step": 22684 }, { "epoch": 0.95092378147826, "grad_norm": 1.8760021924972534, "learning_rate": 6.295190433930065e-08, "loss": 0.4879, "step": 22685 }, { "epoch": 0.950965700093269, "grad_norm": 2.03012752532959, "learning_rate": 6.28445661299043e-08, "loss": 0.4599, "step": 22686 }, { "epoch": 0.9510076187082779, "grad_norm": 2.3813812732696533, "learning_rate": 6.273731893014279e-08, "loss": 0.4939, "step": 22687 }, { "epoch": 0.9510495373232868, "grad_norm": 1.9634572267532349, "learning_rate": 6.26301627419934e-08, "loss": 0.4275, "step": 22688 }, { "epoch": 0.9510914559382958, "grad_norm": 5.0489630699157715, "learning_rate": 6.252309756743124e-08, "loss": 0.4585, "step": 22689 }, { "epoch": 0.9511333745533047, "grad_norm": 2.382084608078003, "learning_rate": 6.241612340843028e-08, "loss": 0.4394, "step": 22690 }, { "epoch": 0.9511752931683137, "grad_norm": 1.9587185382843018, "learning_rate": 6.230924026696172e-08, "loss": 0.5097, "step": 22691 }, { "epoch": 0.9512172117833226, "grad_norm": 2.1499569416046143, "learning_rate": 6.220244814499621e-08, "loss": 0.4645, "step": 22692 }, { "epoch": 0.9512591303983317, "grad_norm": 1.8375928401947021, "learning_rate": 6.209574704450217e-08, "loss": 0.4288, "step": 22693 }, { "epoch": 0.9513010490133406, "grad_norm": 1.7936776876449585, "learning_rate": 6.19891369674469e-08, "loss": 0.4811, "step": 22694 }, { "epoch": 0.9513429676283496, "grad_norm": 2.496884822845459, "learning_rate": 6.188261791579498e-08, "loss": 0.4715, "step": 22695 }, { "epoch": 0.9513848862433585, "grad_norm": 1.6711204051971436, "learning_rate": 6.177618989151035e-08, "loss": 0.4198, "step": 22696 }, { "epoch": 0.9514268048583675, "grad_norm": 2.9264369010925293, "learning_rate": 6.166985289655481e-08, "loss": 0.5538, "step": 22697 }, { "epoch": 0.9514687234733764, "grad_norm": 2.409790515899658, "learning_rate": 6.156360693288787e-08, "loss": 0.4844, "step": 22698 }, { "epoch": 0.9515106420883854, "grad_norm": 2.123059034347534, "learning_rate": 6.145745200246966e-08, "loss": 0.4618, "step": 22699 }, { "epoch": 0.9515525607033943, "grad_norm": 2.1574912071228027, "learning_rate": 6.135138810725582e-08, "loss": 0.4957, "step": 22700 }, { "epoch": 0.9515944793184034, "grad_norm": 1.7872322797775269, "learning_rate": 6.124541524920147e-08, "loss": 0.4082, "step": 22701 }, { "epoch": 0.9516363979334123, "grad_norm": 1.8214300870895386, "learning_rate": 6.113953343026002e-08, "loss": 0.4444, "step": 22702 }, { "epoch": 0.9516783165484213, "grad_norm": 2.044208288192749, "learning_rate": 6.103374265238382e-08, "loss": 0.4477, "step": 22703 }, { "epoch": 0.9517202351634302, "grad_norm": 2.2471609115600586, "learning_rate": 6.092804291752242e-08, "loss": 0.469, "step": 22704 }, { "epoch": 0.9517621537784392, "grad_norm": 2.220200777053833, "learning_rate": 6.082243422762479e-08, "loss": 0.4929, "step": 22705 }, { "epoch": 0.9518040723934481, "grad_norm": 1.8559751510620117, "learning_rate": 6.071691658463663e-08, "loss": 0.4888, "step": 22706 }, { "epoch": 0.9518459910084571, "grad_norm": 2.8551456928253174, "learning_rate": 6.061148999050414e-08, "loss": 0.5182, "step": 22707 }, { "epoch": 0.951887909623466, "grad_norm": 1.9810348749160767, "learning_rate": 6.050615444717078e-08, "loss": 0.4617, "step": 22708 }, { "epoch": 0.951929828238475, "grad_norm": 1.840046763420105, "learning_rate": 6.040090995657777e-08, "loss": 0.4532, "step": 22709 }, { "epoch": 0.951971746853484, "grad_norm": 1.7988219261169434, "learning_rate": 6.029575652066466e-08, "loss": 0.4115, "step": 22710 }, { "epoch": 0.952013665468493, "grad_norm": 2.3527753353118896, "learning_rate": 6.019069414137102e-08, "loss": 0.5, "step": 22711 }, { "epoch": 0.9520555840835019, "grad_norm": 1.9494693279266357, "learning_rate": 6.008572282063252e-08, "loss": 0.5064, "step": 22712 }, { "epoch": 0.9520975026985108, "grad_norm": 3.300339937210083, "learning_rate": 5.998084256038427e-08, "loss": 0.5451, "step": 22713 }, { "epoch": 0.9521394213135198, "grad_norm": 2.6083550453186035, "learning_rate": 5.987605336256031e-08, "loss": 0.4245, "step": 22714 }, { "epoch": 0.9521813399285287, "grad_norm": 1.9459118843078613, "learning_rate": 5.977135522909127e-08, "loss": 0.4889, "step": 22715 }, { "epoch": 0.9522232585435377, "grad_norm": 1.826108694076538, "learning_rate": 5.966674816190732e-08, "loss": 0.4754, "step": 22716 }, { "epoch": 0.9522651771585466, "grad_norm": 1.89769446849823, "learning_rate": 5.956223216293799e-08, "loss": 0.4919, "step": 22717 }, { "epoch": 0.9523070957735557, "grad_norm": 1.835681676864624, "learning_rate": 5.945780723410899e-08, "loss": 0.4629, "step": 22718 }, { "epoch": 0.9523490143885646, "grad_norm": 1.8490997552871704, "learning_rate": 5.9353473377344315e-08, "loss": 0.4856, "step": 22719 }, { "epoch": 0.9523909330035736, "grad_norm": 2.4190804958343506, "learning_rate": 5.924923059456911e-08, "loss": 0.4265, "step": 22720 }, { "epoch": 0.9524328516185825, "grad_norm": 1.8525866270065308, "learning_rate": 5.914507888770293e-08, "loss": 0.4396, "step": 22721 }, { "epoch": 0.9524747702335915, "grad_norm": 3.3555784225463867, "learning_rate": 5.904101825866704e-08, "loss": 0.443, "step": 22722 }, { "epoch": 0.9525166888486004, "grad_norm": 2.321521520614624, "learning_rate": 5.8937048709379885e-08, "loss": 0.4678, "step": 22723 }, { "epoch": 0.9525586074636094, "grad_norm": 1.8661413192749023, "learning_rate": 5.883317024175661e-08, "loss": 0.4574, "step": 22724 }, { "epoch": 0.9526005260786183, "grad_norm": 2.469216823577881, "learning_rate": 5.8729382857713456e-08, "loss": 0.4613, "step": 22725 }, { "epoch": 0.9526424446936274, "grad_norm": 2.7467727661132812, "learning_rate": 5.862568655916279e-08, "loss": 0.5241, "step": 22726 }, { "epoch": 0.9526843633086363, "grad_norm": 2.7448692321777344, "learning_rate": 5.85220813480164e-08, "loss": 0.503, "step": 22727 }, { "epoch": 0.9527262819236453, "grad_norm": 1.8676263093948364, "learning_rate": 5.8418567226184445e-08, "loss": 0.4888, "step": 22728 }, { "epoch": 0.9527682005386542, "grad_norm": 2.315880060195923, "learning_rate": 5.831514419557427e-08, "loss": 0.4821, "step": 22729 }, { "epoch": 0.9528101191536632, "grad_norm": 1.826635718345642, "learning_rate": 5.821181225809269e-08, "loss": 0.5172, "step": 22730 }, { "epoch": 0.9528520377686721, "grad_norm": 2.0123682022094727, "learning_rate": 5.81085714156443e-08, "loss": 0.5401, "step": 22731 }, { "epoch": 0.9528939563836811, "grad_norm": 2.0150105953216553, "learning_rate": 5.800542167013312e-08, "loss": 0.3939, "step": 22732 }, { "epoch": 0.95293587499869, "grad_norm": 2.3833088874816895, "learning_rate": 5.790236302345931e-08, "loss": 0.4965, "step": 22733 }, { "epoch": 0.952977793613699, "grad_norm": 1.8680589199066162, "learning_rate": 5.7799395477523e-08, "loss": 0.4988, "step": 22734 }, { "epoch": 0.953019712228708, "grad_norm": 2.8808720111846924, "learning_rate": 5.769651903422324e-08, "loss": 0.4933, "step": 22735 }, { "epoch": 0.953061630843717, "grad_norm": 2.0141611099243164, "learning_rate": 5.759373369545518e-08, "loss": 0.4457, "step": 22736 }, { "epoch": 0.9531035494587259, "grad_norm": 2.54978346824646, "learning_rate": 5.7491039463113965e-08, "loss": 0.4499, "step": 22737 }, { "epoch": 0.9531454680737348, "grad_norm": 2.313028573989868, "learning_rate": 5.738843633909308e-08, "loss": 0.4624, "step": 22738 }, { "epoch": 0.9531873866887438, "grad_norm": 2.155168294906616, "learning_rate": 5.7285924325282684e-08, "loss": 0.5336, "step": 22739 }, { "epoch": 0.9532293053037527, "grad_norm": 2.1031932830810547, "learning_rate": 5.7183503423573484e-08, "loss": 0.4494, "step": 22740 }, { "epoch": 0.9532712239187617, "grad_norm": 1.9306355714797974, "learning_rate": 5.708117363585286e-08, "loss": 0.4604, "step": 22741 }, { "epoch": 0.9533131425337706, "grad_norm": 2.0137224197387695, "learning_rate": 5.697893496400764e-08, "loss": 0.4457, "step": 22742 }, { "epoch": 0.9533550611487797, "grad_norm": 1.6827675104141235, "learning_rate": 5.687678740992242e-08, "loss": 0.4924, "step": 22743 }, { "epoch": 0.9533969797637886, "grad_norm": 2.005437135696411, "learning_rate": 5.6774730975480144e-08, "loss": 0.4895, "step": 22744 }, { "epoch": 0.9534388983787976, "grad_norm": 1.9730496406555176, "learning_rate": 5.667276566256152e-08, "loss": 0.4345, "step": 22745 }, { "epoch": 0.9534808169938065, "grad_norm": 1.8029544353485107, "learning_rate": 5.657089147304673e-08, "loss": 0.4678, "step": 22746 }, { "epoch": 0.9535227356088155, "grad_norm": 2.0421509742736816, "learning_rate": 5.646910840881314e-08, "loss": 0.4504, "step": 22747 }, { "epoch": 0.9535646542238244, "grad_norm": 1.9198004007339478, "learning_rate": 5.63674164717376e-08, "loss": 0.4048, "step": 22748 }, { "epoch": 0.9536065728388334, "grad_norm": 2.9637668132781982, "learning_rate": 5.6265815663694154e-08, "loss": 0.5187, "step": 22749 }, { "epoch": 0.9536484914538423, "grad_norm": 1.9901235103607178, "learning_rate": 5.6164305986556314e-08, "loss": 0.4281, "step": 22750 }, { "epoch": 0.9536904100688514, "grad_norm": 1.8302197456359863, "learning_rate": 5.6062887442194236e-08, "loss": 0.416, "step": 22751 }, { "epoch": 0.9537323286838603, "grad_norm": 2.0076849460601807, "learning_rate": 5.59615600324781e-08, "loss": 0.4731, "step": 22752 }, { "epoch": 0.9537742472988693, "grad_norm": 1.831446886062622, "learning_rate": 5.5860323759276415e-08, "loss": 0.4977, "step": 22753 }, { "epoch": 0.9538161659138782, "grad_norm": 2.0884644985198975, "learning_rate": 5.575917862445379e-08, "loss": 0.5053, "step": 22754 }, { "epoch": 0.9538580845288872, "grad_norm": 2.4024319648742676, "learning_rate": 5.565812462987596e-08, "loss": 0.5035, "step": 22755 }, { "epoch": 0.9539000031438961, "grad_norm": 2.271062135696411, "learning_rate": 5.555716177740533e-08, "loss": 0.4653, "step": 22756 }, { "epoch": 0.9539419217589051, "grad_norm": 1.9113761186599731, "learning_rate": 5.5456290068902627e-08, "loss": 0.464, "step": 22757 }, { "epoch": 0.953983840373914, "grad_norm": 3.1144182682037354, "learning_rate": 5.5355509506227476e-08, "loss": 0.5506, "step": 22758 }, { "epoch": 0.9540257589889231, "grad_norm": 2.340679168701172, "learning_rate": 5.5254820091238395e-08, "loss": 0.4977, "step": 22759 }, { "epoch": 0.954067677603932, "grad_norm": 1.9150863885879517, "learning_rate": 5.515422182579e-08, "loss": 0.449, "step": 22760 }, { "epoch": 0.954109596218941, "grad_norm": 2.0225605964660645, "learning_rate": 5.5053714711737486e-08, "loss": 0.4856, "step": 22761 }, { "epoch": 0.9541515148339499, "grad_norm": 2.316084384918213, "learning_rate": 5.495329875093436e-08, "loss": 0.4793, "step": 22762 }, { "epoch": 0.9541934334489588, "grad_norm": 1.702192783355713, "learning_rate": 5.485297394523026e-08, "loss": 0.487, "step": 22763 }, { "epoch": 0.9542353520639678, "grad_norm": 1.8664054870605469, "learning_rate": 5.475274029647537e-08, "loss": 0.4496, "step": 22764 }, { "epoch": 0.9542772706789767, "grad_norm": 2.0521347522735596, "learning_rate": 5.4652597806517105e-08, "loss": 0.5077, "step": 22765 }, { "epoch": 0.9543191892939857, "grad_norm": 1.8478937149047852, "learning_rate": 5.455254647720121e-08, "loss": 0.4714, "step": 22766 }, { "epoch": 0.9543611079089946, "grad_norm": 2.2473576068878174, "learning_rate": 5.445258631037231e-08, "loss": 0.4836, "step": 22767 }, { "epoch": 0.9544030265240037, "grad_norm": 1.982924461364746, "learning_rate": 5.435271730787339e-08, "loss": 0.4493, "step": 22768 }, { "epoch": 0.9544449451390126, "grad_norm": 2.43186092376709, "learning_rate": 5.425293947154464e-08, "loss": 0.4842, "step": 22769 }, { "epoch": 0.9544868637540216, "grad_norm": 1.9880789518356323, "learning_rate": 5.4153252803225696e-08, "loss": 0.4511, "step": 22770 }, { "epoch": 0.9545287823690305, "grad_norm": 2.220787763595581, "learning_rate": 5.405365730475398e-08, "loss": 0.4591, "step": 22771 }, { "epoch": 0.9545707009840395, "grad_norm": 1.776859998703003, "learning_rate": 5.3954152977965246e-08, "loss": 0.4564, "step": 22772 }, { "epoch": 0.9546126195990484, "grad_norm": 1.8797684907913208, "learning_rate": 5.385473982469469e-08, "loss": 0.4811, "step": 22773 }, { "epoch": 0.9546545382140574, "grad_norm": 2.2695939540863037, "learning_rate": 5.375541784677363e-08, "loss": 0.4989, "step": 22774 }, { "epoch": 0.9546964568290663, "grad_norm": 1.8397711515426636, "learning_rate": 5.365618704603392e-08, "loss": 0.5343, "step": 22775 }, { "epoch": 0.9547383754440754, "grad_norm": 1.9041776657104492, "learning_rate": 5.355704742430412e-08, "loss": 0.443, "step": 22776 }, { "epoch": 0.9547802940590843, "grad_norm": 1.8352606296539307, "learning_rate": 5.345799898341164e-08, "loss": 0.4851, "step": 22777 }, { "epoch": 0.9548222126740933, "grad_norm": 1.8106517791748047, "learning_rate": 5.33590417251828e-08, "loss": 0.4697, "step": 22778 }, { "epoch": 0.9548641312891022, "grad_norm": 1.6117204427719116, "learning_rate": 5.3260175651441705e-08, "loss": 0.4253, "step": 22779 }, { "epoch": 0.9549060499041112, "grad_norm": 2.192380428314209, "learning_rate": 5.3161400764010216e-08, "loss": 0.4962, "step": 22780 }, { "epoch": 0.9549479685191201, "grad_norm": 2.478158473968506, "learning_rate": 5.306271706470911e-08, "loss": 0.4603, "step": 22781 }, { "epoch": 0.9549898871341291, "grad_norm": 1.9209446907043457, "learning_rate": 5.29641245553586e-08, "loss": 0.462, "step": 22782 }, { "epoch": 0.955031805749138, "grad_norm": 6.109324932098389, "learning_rate": 5.2865623237775556e-08, "loss": 0.4724, "step": 22783 }, { "epoch": 0.9550737243641471, "grad_norm": 2.3687188625335693, "learning_rate": 5.2767213113775196e-08, "loss": 0.4891, "step": 22784 }, { "epoch": 0.955115642979156, "grad_norm": 1.8528728485107422, "learning_rate": 5.266889418517218e-08, "loss": 0.471, "step": 22785 }, { "epoch": 0.955157561594165, "grad_norm": 1.728931188583374, "learning_rate": 5.257066645377784e-08, "loss": 0.4673, "step": 22786 }, { "epoch": 0.9551994802091739, "grad_norm": 2.01082706451416, "learning_rate": 5.2472529921404056e-08, "loss": 0.4727, "step": 22787 }, { "epoch": 0.9552413988241828, "grad_norm": 1.7607707977294922, "learning_rate": 5.237448458985994e-08, "loss": 0.4171, "step": 22788 }, { "epoch": 0.9552833174391918, "grad_norm": 1.8654307126998901, "learning_rate": 5.2276530460951824e-08, "loss": 0.4612, "step": 22789 }, { "epoch": 0.9553252360542007, "grad_norm": 1.9224183559417725, "learning_rate": 5.2178667536485486e-08, "loss": 0.5224, "step": 22790 }, { "epoch": 0.9553671546692097, "grad_norm": 2.6664133071899414, "learning_rate": 5.2080895818266145e-08, "loss": 0.5279, "step": 22791 }, { "epoch": 0.9554090732842186, "grad_norm": 1.8324748277664185, "learning_rate": 5.198321530809458e-08, "loss": 0.4198, "step": 22792 }, { "epoch": 0.9554509918992277, "grad_norm": 2.4285945892333984, "learning_rate": 5.188562600777214e-08, "loss": 0.5096, "step": 22793 }, { "epoch": 0.9554929105142366, "grad_norm": 2.0819077491760254, "learning_rate": 5.178812791909793e-08, "loss": 0.455, "step": 22794 }, { "epoch": 0.9555348291292456, "grad_norm": 1.6884421110153198, "learning_rate": 5.1690721043868296e-08, "loss": 0.4883, "step": 22795 }, { "epoch": 0.9555767477442545, "grad_norm": 2.4210455417633057, "learning_rate": 5.159340538387958e-08, "loss": 0.4946, "step": 22796 }, { "epoch": 0.9556186663592635, "grad_norm": 9.04871940612793, "learning_rate": 5.14961809409259e-08, "loss": 0.4689, "step": 22797 }, { "epoch": 0.9556605849742724, "grad_norm": 1.9260125160217285, "learning_rate": 5.13990477167986e-08, "loss": 0.4414, "step": 22798 }, { "epoch": 0.9557025035892814, "grad_norm": 2.049678325653076, "learning_rate": 5.130200571328847e-08, "loss": 0.4457, "step": 22799 }, { "epoch": 0.9557444222042903, "grad_norm": 2.1083600521087646, "learning_rate": 5.1205054932185194e-08, "loss": 0.4785, "step": 22800 }, { "epoch": 0.9557863408192994, "grad_norm": 2.567174196243286, "learning_rate": 5.110819537527456e-08, "loss": 0.5117, "step": 22801 }, { "epoch": 0.9558282594343083, "grad_norm": 2.381356954574585, "learning_rate": 5.101142704434292e-08, "loss": 0.4707, "step": 22802 }, { "epoch": 0.9558701780493173, "grad_norm": 1.9391019344329834, "learning_rate": 5.091474994117385e-08, "loss": 0.4927, "step": 22803 }, { "epoch": 0.9559120966643262, "grad_norm": 2.0047073364257812, "learning_rate": 5.0818164067549804e-08, "loss": 0.4888, "step": 22804 }, { "epoch": 0.9559540152793352, "grad_norm": 1.8063877820968628, "learning_rate": 5.072166942524992e-08, "loss": 0.4249, "step": 22805 }, { "epoch": 0.9559959338943441, "grad_norm": 1.73622465133667, "learning_rate": 5.0625266016055e-08, "loss": 0.465, "step": 22806 }, { "epoch": 0.9560378525093531, "grad_norm": 2.0174310207366943, "learning_rate": 5.052895384174028e-08, "loss": 0.4935, "step": 22807 }, { "epoch": 0.956079771124362, "grad_norm": 1.721543312072754, "learning_rate": 5.0432732904082125e-08, "loss": 0.4359, "step": 22808 }, { "epoch": 0.9561216897393711, "grad_norm": 2.2381794452667236, "learning_rate": 5.0336603204854116e-08, "loss": 0.5345, "step": 22809 }, { "epoch": 0.95616360835438, "grad_norm": 2.031680107116699, "learning_rate": 5.024056474582761e-08, "loss": 0.4513, "step": 22810 }, { "epoch": 0.956205526969389, "grad_norm": 2.714224100112915, "learning_rate": 5.014461752877397e-08, "loss": 0.489, "step": 22811 }, { "epoch": 0.9562474455843979, "grad_norm": 1.9133654832839966, "learning_rate": 5.004876155546123e-08, "loss": 0.5003, "step": 22812 }, { "epoch": 0.9562893641994068, "grad_norm": 2.2303237915039062, "learning_rate": 4.995299682765631e-08, "loss": 0.4752, "step": 22813 }, { "epoch": 0.9563312828144158, "grad_norm": 1.631529450416565, "learning_rate": 4.985732334712445e-08, "loss": 0.4428, "step": 22814 }, { "epoch": 0.9563732014294247, "grad_norm": 3.2804739475250244, "learning_rate": 4.976174111562981e-08, "loss": 0.4516, "step": 22815 }, { "epoch": 0.9564151200444337, "grad_norm": 1.7411842346191406, "learning_rate": 4.9666250134933754e-08, "loss": 0.4547, "step": 22816 }, { "epoch": 0.9564570386594426, "grad_norm": 1.7354121208190918, "learning_rate": 4.95708504067971e-08, "loss": 0.4405, "step": 22817 }, { "epoch": 0.9564989572744517, "grad_norm": 1.9795231819152832, "learning_rate": 4.9475541932977875e-08, "loss": 0.4645, "step": 22818 }, { "epoch": 0.9565408758894606, "grad_norm": 2.00266432762146, "learning_rate": 4.938032471523302e-08, "loss": 0.4677, "step": 22819 }, { "epoch": 0.9565827945044696, "grad_norm": 2.04199481010437, "learning_rate": 4.9285198755317786e-08, "loss": 0.4804, "step": 22820 }, { "epoch": 0.9566247131194785, "grad_norm": 2.5628817081451416, "learning_rate": 4.919016405498578e-08, "loss": 0.4803, "step": 22821 }, { "epoch": 0.9566666317344875, "grad_norm": 2.0020036697387695, "learning_rate": 4.909522061598948e-08, "loss": 0.5226, "step": 22822 }, { "epoch": 0.9567085503494964, "grad_norm": 1.7272058725357056, "learning_rate": 4.9000368440077497e-08, "loss": 0.4569, "step": 22823 }, { "epoch": 0.9567504689645054, "grad_norm": 1.8903738260269165, "learning_rate": 4.890560752899953e-08, "loss": 0.4365, "step": 22824 }, { "epoch": 0.9567923875795143, "grad_norm": 2.880056142807007, "learning_rate": 4.8810937884501975e-08, "loss": 0.4246, "step": 22825 }, { "epoch": 0.9568343061945234, "grad_norm": 1.9962692260742188, "learning_rate": 4.871635950833009e-08, "loss": 0.4827, "step": 22826 }, { "epoch": 0.9568762248095323, "grad_norm": 1.7465094327926636, "learning_rate": 4.862187240222693e-08, "loss": 0.4372, "step": 22827 }, { "epoch": 0.9569181434245413, "grad_norm": 1.8221704959869385, "learning_rate": 4.852747656793444e-08, "loss": 0.5124, "step": 22828 }, { "epoch": 0.9569600620395502, "grad_norm": 1.830152988433838, "learning_rate": 4.843317200719289e-08, "loss": 0.4424, "step": 22829 }, { "epoch": 0.9570019806545592, "grad_norm": 1.984379768371582, "learning_rate": 4.833895872174088e-08, "loss": 0.4996, "step": 22830 }, { "epoch": 0.9570438992695681, "grad_norm": 1.6974849700927734, "learning_rate": 4.824483671331426e-08, "loss": 0.4418, "step": 22831 }, { "epoch": 0.9570858178845771, "grad_norm": 1.78248929977417, "learning_rate": 4.81508059836483e-08, "loss": 0.4387, "step": 22832 }, { "epoch": 0.957127736499586, "grad_norm": 1.6816956996917725, "learning_rate": 4.805686653447716e-08, "loss": 0.4427, "step": 22833 }, { "epoch": 0.9571696551145951, "grad_norm": 1.8056308031082153, "learning_rate": 4.7963018367531146e-08, "loss": 0.4728, "step": 22834 }, { "epoch": 0.957211573729604, "grad_norm": 1.7530946731567383, "learning_rate": 4.786926148454107e-08, "loss": 0.4742, "step": 22835 }, { "epoch": 0.957253492344613, "grad_norm": 1.9093899726867676, "learning_rate": 4.7775595887235014e-08, "loss": 0.4901, "step": 22836 }, { "epoch": 0.9572954109596219, "grad_norm": 2.0572805404663086, "learning_rate": 4.768202157733992e-08, "loss": 0.4776, "step": 22837 }, { "epoch": 0.9573373295746308, "grad_norm": 2.307734251022339, "learning_rate": 4.7588538556579965e-08, "loss": 0.4373, "step": 22838 }, { "epoch": 0.9573792481896398, "grad_norm": 2.0186479091644287, "learning_rate": 4.7495146826678775e-08, "loss": 0.5271, "step": 22839 }, { "epoch": 0.9574211668046487, "grad_norm": 1.9568750858306885, "learning_rate": 4.7401846389358296e-08, "loss": 0.4298, "step": 22840 }, { "epoch": 0.9574630854196577, "grad_norm": 2.005427837371826, "learning_rate": 4.730863724633772e-08, "loss": 0.4926, "step": 22841 }, { "epoch": 0.9575050040346667, "grad_norm": 1.9764628410339355, "learning_rate": 4.721551939933622e-08, "loss": 0.4619, "step": 22842 }, { "epoch": 0.9575469226496757, "grad_norm": 2.3068296909332275, "learning_rate": 4.7122492850068536e-08, "loss": 0.455, "step": 22843 }, { "epoch": 0.9575888412646846, "grad_norm": 1.8649468421936035, "learning_rate": 4.7029557600251075e-08, "loss": 0.4172, "step": 22844 }, { "epoch": 0.9576307598796936, "grad_norm": 1.7383830547332764, "learning_rate": 4.693671365159691e-08, "loss": 0.4857, "step": 22845 }, { "epoch": 0.9576726784947025, "grad_norm": 1.9593733549118042, "learning_rate": 4.684396100581634e-08, "loss": 0.4868, "step": 22846 }, { "epoch": 0.9577145971097115, "grad_norm": 2.4541375637054443, "learning_rate": 4.675129966462022e-08, "loss": 0.4471, "step": 22847 }, { "epoch": 0.9577565157247204, "grad_norm": 1.5895966291427612, "learning_rate": 4.665872962971552e-08, "loss": 0.4957, "step": 22848 }, { "epoch": 0.9577984343397294, "grad_norm": 1.84056556224823, "learning_rate": 4.656625090281031e-08, "loss": 0.474, "step": 22849 }, { "epoch": 0.9578403529547383, "grad_norm": 1.8987977504730225, "learning_rate": 4.6473863485608226e-08, "loss": 0.5048, "step": 22850 }, { "epoch": 0.9578822715697474, "grad_norm": 1.9741278886795044, "learning_rate": 4.638156737981181e-08, "loss": 0.5195, "step": 22851 }, { "epoch": 0.9579241901847563, "grad_norm": 2.218050241470337, "learning_rate": 4.628936258712302e-08, "loss": 0.4091, "step": 22852 }, { "epoch": 0.9579661087997653, "grad_norm": 1.9181441068649292, "learning_rate": 4.6197249109242173e-08, "loss": 0.46, "step": 22853 }, { "epoch": 0.9580080274147742, "grad_norm": 2.3229331970214844, "learning_rate": 4.610522694786623e-08, "loss": 0.4847, "step": 22854 }, { "epoch": 0.9580499460297832, "grad_norm": 2.299290418624878, "learning_rate": 4.601329610469163e-08, "loss": 0.5488, "step": 22855 }, { "epoch": 0.9580918646447921, "grad_norm": 2.0476417541503906, "learning_rate": 4.592145658141367e-08, "loss": 0.4494, "step": 22856 }, { "epoch": 0.9581337832598011, "grad_norm": 2.104675769805908, "learning_rate": 4.5829708379724335e-08, "loss": 0.5533, "step": 22857 }, { "epoch": 0.95817570187481, "grad_norm": 1.9868197441101074, "learning_rate": 4.573805150131616e-08, "loss": 0.5271, "step": 22858 }, { "epoch": 0.9582176204898191, "grad_norm": 1.9095616340637207, "learning_rate": 4.5646485947877794e-08, "loss": 0.4947, "step": 22859 }, { "epoch": 0.958259539104828, "grad_norm": 2.3953425884246826, "learning_rate": 4.555501172109678e-08, "loss": 0.4679, "step": 22860 }, { "epoch": 0.958301457719837, "grad_norm": 2.2026069164276123, "learning_rate": 4.546362882266009e-08, "loss": 0.4908, "step": 22861 }, { "epoch": 0.9583433763348459, "grad_norm": 1.7941181659698486, "learning_rate": 4.5372337254251943e-08, "loss": 0.4349, "step": 22862 }, { "epoch": 0.9583852949498548, "grad_norm": 2.340120553970337, "learning_rate": 4.528113701755488e-08, "loss": 0.4358, "step": 22863 }, { "epoch": 0.9584272135648638, "grad_norm": 2.588787317276001, "learning_rate": 4.5190028114250886e-08, "loss": 0.4648, "step": 22864 }, { "epoch": 0.9584691321798727, "grad_norm": 2.076467275619507, "learning_rate": 4.5099010546018616e-08, "loss": 0.4984, "step": 22865 }, { "epoch": 0.9585110507948817, "grad_norm": 2.146554946899414, "learning_rate": 4.5008084314536184e-08, "loss": 0.4884, "step": 22866 }, { "epoch": 0.9585529694098907, "grad_norm": 2.157512903213501, "learning_rate": 4.491724942147946e-08, "loss": 0.4282, "step": 22867 }, { "epoch": 0.9585948880248997, "grad_norm": 2.545283555984497, "learning_rate": 4.482650586852322e-08, "loss": 0.5251, "step": 22868 }, { "epoch": 0.9586368066399086, "grad_norm": 1.9260993003845215, "learning_rate": 4.473585365734057e-08, "loss": 0.4803, "step": 22869 }, { "epoch": 0.9586787252549176, "grad_norm": 2.1378560066223145, "learning_rate": 4.464529278960128e-08, "loss": 0.5181, "step": 22870 }, { "epoch": 0.9587206438699265, "grad_norm": 2.162036895751953, "learning_rate": 4.455482326697569e-08, "loss": 0.4696, "step": 22871 }, { "epoch": 0.9587625624849355, "grad_norm": 1.8605808019638062, "learning_rate": 4.446444509113079e-08, "loss": 0.4398, "step": 22872 }, { "epoch": 0.9588044810999444, "grad_norm": 1.8189878463745117, "learning_rate": 4.437415826373359e-08, "loss": 0.5093, "step": 22873 }, { "epoch": 0.9588463997149534, "grad_norm": 1.8737988471984863, "learning_rate": 4.4283962786447753e-08, "loss": 0.4565, "step": 22874 }, { "epoch": 0.9588883183299624, "grad_norm": 2.54370379447937, "learning_rate": 4.419385866093584e-08, "loss": 0.4639, "step": 22875 }, { "epoch": 0.9589302369449714, "grad_norm": 2.438232898712158, "learning_rate": 4.410384588885874e-08, "loss": 0.5058, "step": 22876 }, { "epoch": 0.9589721555599803, "grad_norm": 1.827422857284546, "learning_rate": 4.4013924471876247e-08, "loss": 0.5089, "step": 22877 }, { "epoch": 0.9590140741749893, "grad_norm": 2.1141457557678223, "learning_rate": 4.392409441164591e-08, "loss": 0.4548, "step": 22878 }, { "epoch": 0.9590559927899982, "grad_norm": 1.8087350130081177, "learning_rate": 4.3834355709823084e-08, "loss": 0.4888, "step": 22879 }, { "epoch": 0.9590979114050072, "grad_norm": 1.60758638381958, "learning_rate": 4.374470836806199e-08, "loss": 0.4374, "step": 22880 }, { "epoch": 0.9591398300200161, "grad_norm": 2.144988536834717, "learning_rate": 4.365515238801521e-08, "loss": 0.5229, "step": 22881 }, { "epoch": 0.9591817486350251, "grad_norm": 2.6030359268188477, "learning_rate": 4.3565687771334186e-08, "loss": 0.5065, "step": 22882 }, { "epoch": 0.959223667250034, "grad_norm": 2.272559881210327, "learning_rate": 4.34763145196676e-08, "loss": 0.4796, "step": 22883 }, { "epoch": 0.9592655858650431, "grad_norm": 2.2801015377044678, "learning_rate": 4.338703263466304e-08, "loss": 0.5377, "step": 22884 }, { "epoch": 0.959307504480052, "grad_norm": 1.9794683456420898, "learning_rate": 4.329784211796584e-08, "loss": 0.496, "step": 22885 }, { "epoch": 0.959349423095061, "grad_norm": 1.870983362197876, "learning_rate": 4.32087429712208e-08, "loss": 0.4821, "step": 22886 }, { "epoch": 0.9593913417100699, "grad_norm": 1.9949877262115479, "learning_rate": 4.311973519606993e-08, "loss": 0.499, "step": 22887 }, { "epoch": 0.9594332603250788, "grad_norm": 2.0474369525909424, "learning_rate": 4.303081879415416e-08, "loss": 0.4614, "step": 22888 }, { "epoch": 0.9594751789400878, "grad_norm": 2.203554153442383, "learning_rate": 4.294199376711217e-08, "loss": 0.497, "step": 22889 }, { "epoch": 0.9595170975550967, "grad_norm": 1.7997287511825562, "learning_rate": 4.285326011658153e-08, "loss": 0.4605, "step": 22890 }, { "epoch": 0.9595590161701057, "grad_norm": 1.9104030132293701, "learning_rate": 4.2764617844198166e-08, "loss": 0.4379, "step": 22891 }, { "epoch": 0.9596009347851147, "grad_norm": 2.060901165008545, "learning_rate": 4.267606695159632e-08, "loss": 0.5145, "step": 22892 }, { "epoch": 0.9596428534001237, "grad_norm": 1.672497034072876, "learning_rate": 4.258760744040746e-08, "loss": 0.464, "step": 22893 }, { "epoch": 0.9596847720151326, "grad_norm": 1.7937289476394653, "learning_rate": 4.2499239312262517e-08, "loss": 0.447, "step": 22894 }, { "epoch": 0.9597266906301416, "grad_norm": 1.629408836364746, "learning_rate": 4.241096256879074e-08, "loss": 0.5092, "step": 22895 }, { "epoch": 0.9597686092451505, "grad_norm": 2.0263516902923584, "learning_rate": 4.232277721161915e-08, "loss": 0.4894, "step": 22896 }, { "epoch": 0.9598105278601595, "grad_norm": 2.629850149154663, "learning_rate": 4.2234683242373676e-08, "loss": 0.5038, "step": 22897 }, { "epoch": 0.9598524464751684, "grad_norm": 2.6824705600738525, "learning_rate": 4.214668066267746e-08, "loss": 0.5123, "step": 22898 }, { "epoch": 0.9598943650901774, "grad_norm": 2.059047222137451, "learning_rate": 4.205876947415366e-08, "loss": 0.5118, "step": 22899 }, { "epoch": 0.9599362837051864, "grad_norm": 1.7833718061447144, "learning_rate": 4.197094967842152e-08, "loss": 0.5001, "step": 22900 }, { "epoch": 0.9599782023201954, "grad_norm": 1.9276658296585083, "learning_rate": 4.188322127710143e-08, "loss": 0.4419, "step": 22901 }, { "epoch": 0.9600201209352043, "grad_norm": 2.137800455093384, "learning_rate": 4.17955842718093e-08, "loss": 0.4543, "step": 22902 }, { "epoch": 0.9600620395502133, "grad_norm": 1.758776307106018, "learning_rate": 4.1708038664161066e-08, "loss": 0.4492, "step": 22903 }, { "epoch": 0.9601039581652222, "grad_norm": 1.988993525505066, "learning_rate": 4.1620584455771e-08, "loss": 0.4614, "step": 22904 }, { "epoch": 0.9601458767802312, "grad_norm": 2.084303140640259, "learning_rate": 4.153322164825058e-08, "loss": 0.4424, "step": 22905 }, { "epoch": 0.9601877953952401, "grad_norm": 3.07881498336792, "learning_rate": 4.144595024321019e-08, "loss": 0.5054, "step": 22906 }, { "epoch": 0.9602297140102491, "grad_norm": 1.8158944845199585, "learning_rate": 4.1358770242258536e-08, "loss": 0.4546, "step": 22907 }, { "epoch": 0.960271632625258, "grad_norm": 1.829936146736145, "learning_rate": 4.1271681647002684e-08, "loss": 0.4814, "step": 22908 }, { "epoch": 0.9603135512402671, "grad_norm": 1.8546582460403442, "learning_rate": 4.118468445904855e-08, "loss": 0.4813, "step": 22909 }, { "epoch": 0.960355469855276, "grad_norm": 2.1451282501220703, "learning_rate": 4.109777867999931e-08, "loss": 0.4431, "step": 22910 }, { "epoch": 0.960397388470285, "grad_norm": 1.968578815460205, "learning_rate": 4.1010964311457017e-08, "loss": 0.4829, "step": 22911 }, { "epoch": 0.9604393070852939, "grad_norm": 2.022573232650757, "learning_rate": 4.0924241355022045e-08, "loss": 0.5049, "step": 22912 }, { "epoch": 0.9604812257003028, "grad_norm": 2.427579402923584, "learning_rate": 4.0837609812293124e-08, "loss": 0.5131, "step": 22913 }, { "epoch": 0.9605231443153118, "grad_norm": 1.778289794921875, "learning_rate": 4.0751069684866753e-08, "loss": 0.444, "step": 22914 }, { "epoch": 0.9605650629303207, "grad_norm": 1.7838103771209717, "learning_rate": 4.066462097433832e-08, "loss": 0.5098, "step": 22915 }, { "epoch": 0.9606069815453298, "grad_norm": 6.029293537139893, "learning_rate": 4.057826368230211e-08, "loss": 0.4779, "step": 22916 }, { "epoch": 0.9606489001603387, "grad_norm": 1.7504761219024658, "learning_rate": 4.049199781034852e-08, "loss": 0.4678, "step": 22917 }, { "epoch": 0.9606908187753477, "grad_norm": 1.7874159812927246, "learning_rate": 4.040582336006904e-08, "loss": 0.491, "step": 22918 }, { "epoch": 0.9607327373903566, "grad_norm": 1.8089102506637573, "learning_rate": 4.031974033305242e-08, "loss": 0.4685, "step": 22919 }, { "epoch": 0.9607746560053656, "grad_norm": 1.7816245555877686, "learning_rate": 4.023374873088404e-08, "loss": 0.4213, "step": 22920 }, { "epoch": 0.9608165746203745, "grad_norm": 2.011305809020996, "learning_rate": 4.014784855515042e-08, "loss": 0.5176, "step": 22921 }, { "epoch": 0.9608584932353835, "grad_norm": 1.9434187412261963, "learning_rate": 4.0062039807433615e-08, "loss": 0.4844, "step": 22922 }, { "epoch": 0.9609004118503924, "grad_norm": 2.259025812149048, "learning_rate": 3.997632248931682e-08, "loss": 0.4658, "step": 22923 }, { "epoch": 0.9609423304654015, "grad_norm": 1.645658016204834, "learning_rate": 3.989069660237932e-08, "loss": 0.4658, "step": 22924 }, { "epoch": 0.9609842490804104, "grad_norm": 2.3223013877868652, "learning_rate": 3.980516214819985e-08, "loss": 0.394, "step": 22925 }, { "epoch": 0.9610261676954194, "grad_norm": 2.07932448387146, "learning_rate": 3.9719719128354927e-08, "loss": 0.4872, "step": 22926 }, { "epoch": 0.9610680863104283, "grad_norm": 1.5429177284240723, "learning_rate": 3.9634367544419406e-08, "loss": 0.4648, "step": 22927 }, { "epoch": 0.9611100049254373, "grad_norm": 2.2602384090423584, "learning_rate": 3.9549107397967025e-08, "loss": 0.5169, "step": 22928 }, { "epoch": 0.9611519235404462, "grad_norm": 2.1388802528381348, "learning_rate": 3.9463938690568745e-08, "loss": 0.5066, "step": 22929 }, { "epoch": 0.9611938421554552, "grad_norm": 2.017909288406372, "learning_rate": 3.937886142379555e-08, "loss": 0.537, "step": 22930 }, { "epoch": 0.9612357607704641, "grad_norm": 2.1377573013305664, "learning_rate": 3.9293875599215046e-08, "loss": 0.4587, "step": 22931 }, { "epoch": 0.9612776793854731, "grad_norm": 2.2418930530548096, "learning_rate": 3.920898121839378e-08, "loss": 0.4654, "step": 22932 }, { "epoch": 0.9613195980004821, "grad_norm": 2.0081284046173096, "learning_rate": 3.912417828289716e-08, "loss": 0.4772, "step": 22933 }, { "epoch": 0.9613615166154911, "grad_norm": 2.7765023708343506, "learning_rate": 3.9039466794288384e-08, "loss": 0.5415, "step": 22934 }, { "epoch": 0.9614034352305, "grad_norm": 1.869755744934082, "learning_rate": 3.895484675412786e-08, "loss": 0.4705, "step": 22935 }, { "epoch": 0.961445353845509, "grad_norm": 3.0117838382720947, "learning_rate": 3.887031816397713e-08, "loss": 0.4704, "step": 22936 }, { "epoch": 0.9614872724605179, "grad_norm": 2.456295967102051, "learning_rate": 3.878588102539271e-08, "loss": 0.46, "step": 22937 }, { "epoch": 0.9615291910755268, "grad_norm": 2.083216428756714, "learning_rate": 3.870153533993226e-08, "loss": 0.4835, "step": 22938 }, { "epoch": 0.9615711096905358, "grad_norm": 2.095858573913574, "learning_rate": 3.861728110915064e-08, "loss": 0.4529, "step": 22939 }, { "epoch": 0.9616130283055447, "grad_norm": 2.074545383453369, "learning_rate": 3.853311833459994e-08, "loss": 0.4615, "step": 22940 }, { "epoch": 0.9616549469205538, "grad_norm": 1.9659539461135864, "learning_rate": 3.8449047017832255e-08, "loss": 0.5186, "step": 22941 }, { "epoch": 0.9616968655355627, "grad_norm": 2.1346595287323, "learning_rate": 3.836506716039745e-08, "loss": 0.4719, "step": 22942 }, { "epoch": 0.9617387841505717, "grad_norm": 1.763791561126709, "learning_rate": 3.828117876384374e-08, "loss": 0.5125, "step": 22943 }, { "epoch": 0.9617807027655806, "grad_norm": 2.2132935523986816, "learning_rate": 3.819738182971655e-08, "loss": 0.4293, "step": 22944 }, { "epoch": 0.9618226213805896, "grad_norm": 1.902215600013733, "learning_rate": 3.811367635956131e-08, "loss": 0.4537, "step": 22945 }, { "epoch": 0.9618645399955985, "grad_norm": 1.6117526292800903, "learning_rate": 3.8030062354921235e-08, "loss": 0.4475, "step": 22946 }, { "epoch": 0.9619064586106075, "grad_norm": 1.718225121498108, "learning_rate": 3.794653981733676e-08, "loss": 0.4481, "step": 22947 }, { "epoch": 0.9619483772256164, "grad_norm": 3.0014305114746094, "learning_rate": 3.7863108748347753e-08, "loss": 0.4801, "step": 22948 }, { "epoch": 0.9619902958406255, "grad_norm": 3.158048629760742, "learning_rate": 3.7779769149493e-08, "loss": 0.4626, "step": 22949 }, { "epoch": 0.9620322144556344, "grad_norm": 1.7783336639404297, "learning_rate": 3.769652102230792e-08, "loss": 0.48, "step": 22950 }, { "epoch": 0.9620741330706434, "grad_norm": 1.8694114685058594, "learning_rate": 3.761336436832741e-08, "loss": 0.5111, "step": 22951 }, { "epoch": 0.9621160516856523, "grad_norm": 2.2758100032806396, "learning_rate": 3.753029918908413e-08, "loss": 0.5142, "step": 22952 }, { "epoch": 0.9621579703006613, "grad_norm": 1.677738070487976, "learning_rate": 3.744732548610908e-08, "loss": 0.4551, "step": 22953 }, { "epoch": 0.9621998889156702, "grad_norm": 2.0973641872406006, "learning_rate": 3.7364443260932134e-08, "loss": 0.523, "step": 22954 }, { "epoch": 0.9622418075306792, "grad_norm": 1.9760116338729858, "learning_rate": 3.7281652515081533e-08, "loss": 0.386, "step": 22955 }, { "epoch": 0.9622837261456881, "grad_norm": 1.942956805229187, "learning_rate": 3.719895325008216e-08, "loss": 0.5267, "step": 22956 }, { "epoch": 0.9623256447606972, "grad_norm": 2.0116000175476074, "learning_rate": 3.7116345467460014e-08, "loss": 0.4408, "step": 22957 }, { "epoch": 0.9623675633757061, "grad_norm": 1.8680223226547241, "learning_rate": 3.7033829168736654e-08, "loss": 0.4895, "step": 22958 }, { "epoch": 0.9624094819907151, "grad_norm": 1.64841628074646, "learning_rate": 3.69514043554331e-08, "loss": 0.4696, "step": 22959 }, { "epoch": 0.962451400605724, "grad_norm": 1.6627392768859863, "learning_rate": 3.686907102907033e-08, "loss": 0.4655, "step": 22960 }, { "epoch": 0.962493319220733, "grad_norm": 1.8465176820755005, "learning_rate": 3.6786829191163834e-08, "loss": 0.4901, "step": 22961 }, { "epoch": 0.9625352378357419, "grad_norm": 2.0019726753234863, "learning_rate": 3.670467884323126e-08, "loss": 0.5138, "step": 22962 }, { "epoch": 0.9625771564507508, "grad_norm": 1.7242423295974731, "learning_rate": 3.66226199867864e-08, "loss": 0.4814, "step": 22963 }, { "epoch": 0.9626190750657598, "grad_norm": 2.1766538619995117, "learning_rate": 3.6540652623341946e-08, "loss": 0.4462, "step": 22964 }, { "epoch": 0.9626609936807687, "grad_norm": 1.9408468008041382, "learning_rate": 3.645877675440834e-08, "loss": 0.4534, "step": 22965 }, { "epoch": 0.9627029122957778, "grad_norm": 1.7373948097229004, "learning_rate": 3.63769923814955e-08, "loss": 0.454, "step": 22966 }, { "epoch": 0.9627448309107867, "grad_norm": 1.9690840244293213, "learning_rate": 3.62952995061111e-08, "loss": 0.5007, "step": 22967 }, { "epoch": 0.9627867495257957, "grad_norm": 1.4527976512908936, "learning_rate": 3.621369812976005e-08, "loss": 0.4165, "step": 22968 }, { "epoch": 0.9628286681408046, "grad_norm": 1.8112767934799194, "learning_rate": 3.613218825394782e-08, "loss": 0.5135, "step": 22969 }, { "epoch": 0.9628705867558136, "grad_norm": 2.844036340713501, "learning_rate": 3.6050769880175976e-08, "loss": 0.5122, "step": 22970 }, { "epoch": 0.9629125053708225, "grad_norm": 2.7404961585998535, "learning_rate": 3.596944300994665e-08, "loss": 0.4517, "step": 22971 }, { "epoch": 0.9629544239858315, "grad_norm": 1.761605143547058, "learning_rate": 3.5888207644757535e-08, "loss": 0.4686, "step": 22972 }, { "epoch": 0.9629963426008404, "grad_norm": 1.7382937669754028, "learning_rate": 3.5807063786106325e-08, "loss": 0.453, "step": 22973 }, { "epoch": 0.9630382612158495, "grad_norm": 1.6933097839355469, "learning_rate": 3.572601143548904e-08, "loss": 0.4872, "step": 22974 }, { "epoch": 0.9630801798308584, "grad_norm": 2.023348093032837, "learning_rate": 3.5645050594400046e-08, "loss": 0.4665, "step": 22975 }, { "epoch": 0.9631220984458674, "grad_norm": 1.6262232065200806, "learning_rate": 3.5564181264331476e-08, "loss": 0.443, "step": 22976 }, { "epoch": 0.9631640170608763, "grad_norm": 1.5743892192840576, "learning_rate": 3.5483403446774364e-08, "loss": 0.4577, "step": 22977 }, { "epoch": 0.9632059356758853, "grad_norm": 1.635316252708435, "learning_rate": 3.540271714321697e-08, "loss": 0.4524, "step": 22978 }, { "epoch": 0.9632478542908942, "grad_norm": 2.4692862033843994, "learning_rate": 3.532212235514754e-08, "loss": 0.5041, "step": 22979 }, { "epoch": 0.9632897729059032, "grad_norm": 1.819183588027954, "learning_rate": 3.524161908405155e-08, "loss": 0.4637, "step": 22980 }, { "epoch": 0.9633316915209121, "grad_norm": 3.4928340911865234, "learning_rate": 3.5161207331412264e-08, "loss": 0.501, "step": 22981 }, { "epoch": 0.9633736101359212, "grad_norm": 2.4952785968780518, "learning_rate": 3.508088709871238e-08, "loss": 0.4691, "step": 22982 }, { "epoch": 0.9634155287509301, "grad_norm": 1.8744511604309082, "learning_rate": 3.5000658387432937e-08, "loss": 0.5057, "step": 22983 }, { "epoch": 0.9634574473659391, "grad_norm": 1.7367939949035645, "learning_rate": 3.49205211990522e-08, "loss": 0.458, "step": 22984 }, { "epoch": 0.963499365980948, "grad_norm": 2.0484254360198975, "learning_rate": 3.484047553504788e-08, "loss": 0.4377, "step": 22985 }, { "epoch": 0.963541284595957, "grad_norm": 1.6601533889770508, "learning_rate": 3.4760521396894895e-08, "loss": 0.4077, "step": 22986 }, { "epoch": 0.9635832032109659, "grad_norm": 2.0243687629699707, "learning_rate": 3.468065878606763e-08, "loss": 0.5404, "step": 22987 }, { "epoch": 0.9636251218259748, "grad_norm": 1.7993535995483398, "learning_rate": 3.460088770403824e-08, "loss": 0.4509, "step": 22988 }, { "epoch": 0.9636670404409838, "grad_norm": 1.8686370849609375, "learning_rate": 3.4521208152277216e-08, "loss": 0.5127, "step": 22989 }, { "epoch": 0.9637089590559927, "grad_norm": 1.8793977499008179, "learning_rate": 3.444162013225283e-08, "loss": 0.4973, "step": 22990 }, { "epoch": 0.9637508776710018, "grad_norm": 1.8748751878738403, "learning_rate": 3.4362123645432233e-08, "loss": 0.4724, "step": 22991 }, { "epoch": 0.9637927962860107, "grad_norm": 2.5864076614379883, "learning_rate": 3.428271869328148e-08, "loss": 0.5204, "step": 22992 }, { "epoch": 0.9638347149010197, "grad_norm": 2.754575490951538, "learning_rate": 3.4203405277263846e-08, "loss": 0.5079, "step": 22993 }, { "epoch": 0.9638766335160286, "grad_norm": 2.2094573974609375, "learning_rate": 3.412418339884094e-08, "loss": 0.4505, "step": 22994 }, { "epoch": 0.9639185521310376, "grad_norm": 1.88340425491333, "learning_rate": 3.4045053059474364e-08, "loss": 0.4824, "step": 22995 }, { "epoch": 0.9639604707460465, "grad_norm": 2.2662463188171387, "learning_rate": 3.3966014260621846e-08, "loss": 0.4959, "step": 22996 }, { "epoch": 0.9640023893610555, "grad_norm": 1.8810322284698486, "learning_rate": 3.3887067003740004e-08, "loss": 0.4943, "step": 22997 }, { "epoch": 0.9640443079760644, "grad_norm": 2.569605827331543, "learning_rate": 3.3808211290284886e-08, "loss": 0.4784, "step": 22998 }, { "epoch": 0.9640862265910735, "grad_norm": 2.1726136207580566, "learning_rate": 3.3729447121710334e-08, "loss": 0.4982, "step": 22999 }, { "epoch": 0.9641281452060824, "grad_norm": 3.242830514907837, "learning_rate": 3.365077449946741e-08, "loss": 0.4532, "step": 23000 }, { "epoch": 0.9641700638210914, "grad_norm": 1.9933689832687378, "learning_rate": 3.357219342500662e-08, "loss": 0.4812, "step": 23001 }, { "epoch": 0.9642119824361003, "grad_norm": 1.9730464220046997, "learning_rate": 3.349370389977624e-08, "loss": 0.5175, "step": 23002 }, { "epoch": 0.9642539010511093, "grad_norm": 4.960993766784668, "learning_rate": 3.341530592522402e-08, "loss": 0.5205, "step": 23003 }, { "epoch": 0.9642958196661182, "grad_norm": 1.8406296968460083, "learning_rate": 3.333699950279434e-08, "loss": 0.4169, "step": 23004 }, { "epoch": 0.9643377382811272, "grad_norm": 2.3806207180023193, "learning_rate": 3.3258784633930506e-08, "loss": 0.4538, "step": 23005 }, { "epoch": 0.9643796568961361, "grad_norm": 1.885832667350769, "learning_rate": 3.318066132007525e-08, "loss": 0.4557, "step": 23006 }, { "epoch": 0.9644215755111452, "grad_norm": 2.058844566345215, "learning_rate": 3.3102629562667966e-08, "loss": 0.4956, "step": 23007 }, { "epoch": 0.9644634941261541, "grad_norm": 2.446934461593628, "learning_rate": 3.3024689363146956e-08, "loss": 0.4983, "step": 23008 }, { "epoch": 0.9645054127411631, "grad_norm": 1.7552202939987183, "learning_rate": 3.29468407229494e-08, "loss": 0.4228, "step": 23009 }, { "epoch": 0.964547331356172, "grad_norm": 2.1142375469207764, "learning_rate": 3.2869083643509715e-08, "loss": 0.4485, "step": 23010 }, { "epoch": 0.964589249971181, "grad_norm": 2.063868761062622, "learning_rate": 3.279141812626174e-08, "loss": 0.5394, "step": 23011 }, { "epoch": 0.9646311685861899, "grad_norm": 2.025139570236206, "learning_rate": 3.271384417263712e-08, "loss": 0.4787, "step": 23012 }, { "epoch": 0.9646730872011988, "grad_norm": 1.9736573696136475, "learning_rate": 3.2636361784065816e-08, "loss": 0.4994, "step": 23013 }, { "epoch": 0.9647150058162078, "grad_norm": 2.2313153743743896, "learning_rate": 3.255897096197558e-08, "loss": 0.4832, "step": 23014 }, { "epoch": 0.9647569244312167, "grad_norm": 1.830579161643982, "learning_rate": 3.248167170779359e-08, "loss": 0.4645, "step": 23015 }, { "epoch": 0.9647988430462258, "grad_norm": 2.516139507293701, "learning_rate": 3.240446402294484e-08, "loss": 0.5572, "step": 23016 }, { "epoch": 0.9648407616612347, "grad_norm": 2.3760061264038086, "learning_rate": 3.232734790885261e-08, "loss": 0.4638, "step": 23017 }, { "epoch": 0.9648826802762437, "grad_norm": 2.587249517440796, "learning_rate": 3.2250323366937453e-08, "loss": 0.487, "step": 23018 }, { "epoch": 0.9649245988912526, "grad_norm": 1.922865867614746, "learning_rate": 3.217339039862044e-08, "loss": 0.4791, "step": 23019 }, { "epoch": 0.9649665175062616, "grad_norm": 2.2321505546569824, "learning_rate": 3.2096549005318776e-08, "loss": 0.4895, "step": 23020 }, { "epoch": 0.9650084361212705, "grad_norm": 1.7141474485397339, "learning_rate": 3.2019799188449105e-08, "loss": 0.429, "step": 23021 }, { "epoch": 0.9650503547362795, "grad_norm": 1.8077131509780884, "learning_rate": 3.1943140949426964e-08, "loss": 0.4376, "step": 23022 }, { "epoch": 0.9650922733512884, "grad_norm": 2.2145004272460938, "learning_rate": 3.1866574289664555e-08, "loss": 0.4623, "step": 23023 }, { "epoch": 0.9651341919662975, "grad_norm": 1.8837294578552246, "learning_rate": 3.179009921057352e-08, "loss": 0.4955, "step": 23024 }, { "epoch": 0.9651761105813064, "grad_norm": 2.3239896297454834, "learning_rate": 3.1713715713563855e-08, "loss": 0.4787, "step": 23025 }, { "epoch": 0.9652180291963154, "grad_norm": 1.6508920192718506, "learning_rate": 3.163742380004331e-08, "loss": 0.4494, "step": 23026 }, { "epoch": 0.9652599478113243, "grad_norm": 2.3090178966522217, "learning_rate": 3.1561223471418545e-08, "loss": 0.4661, "step": 23027 }, { "epoch": 0.9653018664263333, "grad_norm": 2.0950465202331543, "learning_rate": 3.148511472909399e-08, "loss": 0.4861, "step": 23028 }, { "epoch": 0.9653437850413422, "grad_norm": 1.815687656402588, "learning_rate": 3.140909757447241e-08, "loss": 0.4529, "step": 23029 }, { "epoch": 0.9653857036563512, "grad_norm": 2.004018545150757, "learning_rate": 3.1333172008954914e-08, "loss": 0.4768, "step": 23030 }, { "epoch": 0.9654276222713601, "grad_norm": 2.017518997192383, "learning_rate": 3.1257338033941485e-08, "loss": 0.4522, "step": 23031 }, { "epoch": 0.9654695408863692, "grad_norm": 1.9452760219573975, "learning_rate": 3.118159565083045e-08, "loss": 0.5088, "step": 23032 }, { "epoch": 0.9655114595013781, "grad_norm": 1.9809836149215698, "learning_rate": 3.1105944861016815e-08, "loss": 0.4918, "step": 23033 }, { "epoch": 0.9655533781163871, "grad_norm": 4.763397216796875, "learning_rate": 3.1030385665896666e-08, "loss": 0.5119, "step": 23034 }, { "epoch": 0.965595296731396, "grad_norm": 1.9793981313705444, "learning_rate": 3.095491806686113e-08, "loss": 0.4574, "step": 23035 }, { "epoch": 0.965637215346405, "grad_norm": 1.7351322174072266, "learning_rate": 3.087954206530297e-08, "loss": 0.4219, "step": 23036 }, { "epoch": 0.9656791339614139, "grad_norm": 4.216161251068115, "learning_rate": 3.080425766261052e-08, "loss": 0.4395, "step": 23037 }, { "epoch": 0.9657210525764228, "grad_norm": 1.6261026859283447, "learning_rate": 3.072906486017158e-08, "loss": 0.4539, "step": 23038 }, { "epoch": 0.9657629711914318, "grad_norm": 1.8787504434585571, "learning_rate": 3.065396365937279e-08, "loss": 0.5027, "step": 23039 }, { "epoch": 0.9658048898064407, "grad_norm": 1.7658069133758545, "learning_rate": 3.057895406159805e-08, "loss": 0.4793, "step": 23040 }, { "epoch": 0.9658468084214498, "grad_norm": 2.00093936920166, "learning_rate": 3.0504036068230156e-08, "loss": 0.5301, "step": 23041 }, { "epoch": 0.9658887270364587, "grad_norm": 2.179924964904785, "learning_rate": 3.042920968065077e-08, "loss": 0.4635, "step": 23042 }, { "epoch": 0.9659306456514677, "grad_norm": 2.026726722717285, "learning_rate": 3.035447490023824e-08, "loss": 0.5158, "step": 23043 }, { "epoch": 0.9659725642664766, "grad_norm": 1.9584441184997559, "learning_rate": 3.0279831728370903e-08, "loss": 0.4722, "step": 23044 }, { "epoch": 0.9660144828814856, "grad_norm": 2.015134334564209, "learning_rate": 3.020528016642432e-08, "loss": 0.5052, "step": 23045 }, { "epoch": 0.9660564014964945, "grad_norm": 2.0291335582733154, "learning_rate": 3.013082021577296e-08, "loss": 0.4799, "step": 23046 }, { "epoch": 0.9660983201115035, "grad_norm": 2.061417579650879, "learning_rate": 3.005645187778905e-08, "loss": 0.5415, "step": 23047 }, { "epoch": 0.9661402387265124, "grad_norm": 2.556077003479004, "learning_rate": 2.998217515384372e-08, "loss": 0.5013, "step": 23048 }, { "epoch": 0.9661821573415215, "grad_norm": 1.782890796661377, "learning_rate": 2.990799004530698e-08, "loss": 0.4784, "step": 23049 }, { "epoch": 0.9662240759565304, "grad_norm": 5.342665195465088, "learning_rate": 2.98338965535444e-08, "loss": 0.4717, "step": 23050 }, { "epoch": 0.9662659945715394, "grad_norm": 1.7231190204620361, "learning_rate": 2.975989467992324e-08, "loss": 0.4104, "step": 23051 }, { "epoch": 0.9663079131865483, "grad_norm": 1.9356478452682495, "learning_rate": 2.968598442580739e-08, "loss": 0.4536, "step": 23052 }, { "epoch": 0.9663498318015573, "grad_norm": 1.9832271337509155, "learning_rate": 2.96121657925591e-08, "loss": 0.4163, "step": 23053 }, { "epoch": 0.9663917504165662, "grad_norm": 2.930382490158081, "learning_rate": 2.9538438781539503e-08, "loss": 0.5286, "step": 23054 }, { "epoch": 0.9664336690315752, "grad_norm": 2.3637330532073975, "learning_rate": 2.9464803394106957e-08, "loss": 0.4775, "step": 23055 }, { "epoch": 0.9664755876465841, "grad_norm": 2.4542455673217773, "learning_rate": 2.9391259631619262e-08, "loss": 0.4837, "step": 23056 }, { "epoch": 0.9665175062615932, "grad_norm": 1.7314404249191284, "learning_rate": 2.9317807495432006e-08, "loss": 0.4863, "step": 23057 }, { "epoch": 0.9665594248766021, "grad_norm": 1.9385112524032593, "learning_rate": 2.92444469868991e-08, "loss": 0.4666, "step": 23058 }, { "epoch": 0.9666013434916111, "grad_norm": 2.089433431625366, "learning_rate": 2.91711781073728e-08, "loss": 0.4666, "step": 23059 }, { "epoch": 0.96664326210662, "grad_norm": 2.3240885734558105, "learning_rate": 2.9098000858204247e-08, "loss": 0.4386, "step": 23060 }, { "epoch": 0.966685180721629, "grad_norm": 2.1854565143585205, "learning_rate": 2.90249152407418e-08, "loss": 0.4359, "step": 23061 }, { "epoch": 0.9667270993366379, "grad_norm": 2.153710126876831, "learning_rate": 2.895192125633328e-08, "loss": 0.4597, "step": 23062 }, { "epoch": 0.9667690179516468, "grad_norm": 2.60485577583313, "learning_rate": 2.8879018906323165e-08, "loss": 0.4377, "step": 23063 }, { "epoch": 0.9668109365666558, "grad_norm": 1.958864688873291, "learning_rate": 2.8806208192056486e-08, "loss": 0.4981, "step": 23064 }, { "epoch": 0.9668528551816647, "grad_norm": 1.5902248620986938, "learning_rate": 2.8733489114874946e-08, "loss": 0.4576, "step": 23065 }, { "epoch": 0.9668947737966738, "grad_norm": 2.122877597808838, "learning_rate": 2.866086167611859e-08, "loss": 0.5374, "step": 23066 }, { "epoch": 0.9669366924116827, "grad_norm": 2.430896282196045, "learning_rate": 2.8588325877126898e-08, "loss": 0.4748, "step": 23067 }, { "epoch": 0.9669786110266917, "grad_norm": 5.283737659454346, "learning_rate": 2.8515881719236028e-08, "loss": 0.4976, "step": 23068 }, { "epoch": 0.9670205296417006, "grad_norm": 2.1355068683624268, "learning_rate": 2.8443529203782682e-08, "loss": 0.5276, "step": 23069 }, { "epoch": 0.9670624482567096, "grad_norm": 2.0175418853759766, "learning_rate": 2.8371268332099688e-08, "loss": 0.4518, "step": 23070 }, { "epoch": 0.9671043668717185, "grad_norm": 1.7497310638427734, "learning_rate": 2.829909910551931e-08, "loss": 0.4711, "step": 23071 }, { "epoch": 0.9671462854867275, "grad_norm": 1.9195936918258667, "learning_rate": 2.8227021525371602e-08, "loss": 0.4873, "step": 23072 }, { "epoch": 0.9671882041017364, "grad_norm": 1.9710568189620972, "learning_rate": 2.8155035592986045e-08, "loss": 0.4721, "step": 23073 }, { "epoch": 0.9672301227167455, "grad_norm": 7.3426384925842285, "learning_rate": 2.808314130968881e-08, "loss": 0.482, "step": 23074 }, { "epoch": 0.9672720413317544, "grad_norm": 1.8640109300613403, "learning_rate": 2.80113386768055e-08, "loss": 0.4742, "step": 23075 }, { "epoch": 0.9673139599467634, "grad_norm": 4.046444416046143, "learning_rate": 2.7939627695658946e-08, "loss": 0.5141, "step": 23076 }, { "epoch": 0.9673558785617723, "grad_norm": 1.951568365097046, "learning_rate": 2.7868008367571973e-08, "loss": 0.4352, "step": 23077 }, { "epoch": 0.9673977971767813, "grad_norm": 3.023271083831787, "learning_rate": 2.77964806938652e-08, "loss": 0.4924, "step": 23078 }, { "epoch": 0.9674397157917902, "grad_norm": 2.691901683807373, "learning_rate": 2.772504467585535e-08, "loss": 0.5312, "step": 23079 }, { "epoch": 0.9674816344067992, "grad_norm": 2.6558432579040527, "learning_rate": 2.765370031486081e-08, "loss": 0.4951, "step": 23080 }, { "epoch": 0.9675235530218081, "grad_norm": 2.3576278686523438, "learning_rate": 2.7582447612196638e-08, "loss": 0.4633, "step": 23081 }, { "epoch": 0.9675654716368172, "grad_norm": 1.9223417043685913, "learning_rate": 2.7511286569175676e-08, "loss": 0.4821, "step": 23082 }, { "epoch": 0.9676073902518261, "grad_norm": 2.4240145683288574, "learning_rate": 2.7440217187109653e-08, "loss": 0.5123, "step": 23083 }, { "epoch": 0.9676493088668351, "grad_norm": 2.028089761734009, "learning_rate": 2.7369239467309184e-08, "loss": 0.4059, "step": 23084 }, { "epoch": 0.967691227481844, "grad_norm": 2.0806188583374023, "learning_rate": 2.7298353411082113e-08, "loss": 0.5489, "step": 23085 }, { "epoch": 0.967733146096853, "grad_norm": 1.595145583152771, "learning_rate": 2.7227559019735172e-08, "loss": 0.5012, "step": 23086 }, { "epoch": 0.9677750647118619, "grad_norm": 2.33716082572937, "learning_rate": 2.7156856294573985e-08, "loss": 0.5193, "step": 23087 }, { "epoch": 0.9678169833268708, "grad_norm": 2.8010222911834717, "learning_rate": 2.7086245236900842e-08, "loss": 0.4501, "step": 23088 }, { "epoch": 0.9678589019418798, "grad_norm": 1.588104248046875, "learning_rate": 2.7015725848018037e-08, "loss": 0.4692, "step": 23089 }, { "epoch": 0.9679008205568888, "grad_norm": 1.553530216217041, "learning_rate": 2.694529812922564e-08, "loss": 0.4522, "step": 23090 }, { "epoch": 0.9679427391718978, "grad_norm": 5.0315165519714355, "learning_rate": 2.68749620818215e-08, "loss": 0.4562, "step": 23091 }, { "epoch": 0.9679846577869067, "grad_norm": 1.9699565172195435, "learning_rate": 2.6804717707102357e-08, "loss": 0.4574, "step": 23092 }, { "epoch": 0.9680265764019157, "grad_norm": 1.9561270475387573, "learning_rate": 2.673456500636329e-08, "loss": 0.4474, "step": 23093 }, { "epoch": 0.9680684950169246, "grad_norm": 1.5987391471862793, "learning_rate": 2.6664503980896593e-08, "loss": 0.4713, "step": 23094 }, { "epoch": 0.9681104136319336, "grad_norm": 2.3230535984039307, "learning_rate": 2.6594534631994573e-08, "loss": 0.4716, "step": 23095 }, { "epoch": 0.9681523322469425, "grad_norm": 1.7997729778289795, "learning_rate": 2.6524656960947305e-08, "loss": 0.5261, "step": 23096 }, { "epoch": 0.9681942508619515, "grad_norm": 2.007755756378174, "learning_rate": 2.6454870969041534e-08, "loss": 0.4568, "step": 23097 }, { "epoch": 0.9682361694769605, "grad_norm": 1.8394485712051392, "learning_rate": 2.6385176657564572e-08, "loss": 0.5065, "step": 23098 }, { "epoch": 0.9682780880919695, "grad_norm": 2.6583356857299805, "learning_rate": 2.6315574027801493e-08, "loss": 0.4524, "step": 23099 }, { "epoch": 0.9683200067069784, "grad_norm": 1.8351860046386719, "learning_rate": 2.6246063081035168e-08, "loss": 0.4872, "step": 23100 }, { "epoch": 0.9683619253219874, "grad_norm": 1.7716585397720337, "learning_rate": 2.6176643818546232e-08, "loss": 0.4316, "step": 23101 }, { "epoch": 0.9684038439369963, "grad_norm": 1.9204884767532349, "learning_rate": 2.6107316241614777e-08, "loss": 0.4339, "step": 23102 }, { "epoch": 0.9684457625520053, "grad_norm": 2.2292187213897705, "learning_rate": 2.603808035151867e-08, "loss": 0.4794, "step": 23103 }, { "epoch": 0.9684876811670142, "grad_norm": 2.1916866302490234, "learning_rate": 2.5968936149534107e-08, "loss": 0.5079, "step": 23104 }, { "epoch": 0.9685295997820232, "grad_norm": 1.7220447063446045, "learning_rate": 2.5899883636936185e-08, "loss": 0.5131, "step": 23105 }, { "epoch": 0.9685715183970321, "grad_norm": 1.73961341381073, "learning_rate": 2.5830922814997218e-08, "loss": 0.454, "step": 23106 }, { "epoch": 0.9686134370120412, "grad_norm": 1.8756276369094849, "learning_rate": 2.5762053684988407e-08, "loss": 0.4881, "step": 23107 }, { "epoch": 0.9686553556270501, "grad_norm": 1.7712688446044922, "learning_rate": 2.5693276248179855e-08, "loss": 0.4094, "step": 23108 }, { "epoch": 0.9686972742420591, "grad_norm": 3.6036102771759033, "learning_rate": 2.5624590505838876e-08, "loss": 0.4844, "step": 23109 }, { "epoch": 0.968739192857068, "grad_norm": 1.602253794670105, "learning_rate": 2.555599645923168e-08, "loss": 0.4585, "step": 23110 }, { "epoch": 0.968781111472077, "grad_norm": 1.9805907011032104, "learning_rate": 2.5487494109622813e-08, "loss": 0.4781, "step": 23111 }, { "epoch": 0.9688230300870859, "grad_norm": 1.9536340236663818, "learning_rate": 2.541908345827515e-08, "loss": 0.3941, "step": 23112 }, { "epoch": 0.9688649487020948, "grad_norm": 1.87322998046875, "learning_rate": 2.53507645064488e-08, "loss": 0.493, "step": 23113 }, { "epoch": 0.9689068673171038, "grad_norm": 1.7829267978668213, "learning_rate": 2.5282537255404415e-08, "loss": 0.4776, "step": 23114 }, { "epoch": 0.9689487859321128, "grad_norm": 2.100567579269409, "learning_rate": 2.5214401706398772e-08, "loss": 0.4942, "step": 23115 }, { "epoch": 0.9689907045471218, "grad_norm": 1.8049335479736328, "learning_rate": 2.514635786068864e-08, "loss": 0.4467, "step": 23116 }, { "epoch": 0.9690326231621307, "grad_norm": 1.95842444896698, "learning_rate": 2.5078405719527464e-08, "loss": 0.4723, "step": 23117 }, { "epoch": 0.9690745417771397, "grad_norm": 2.0600554943084717, "learning_rate": 2.5010545284168687e-08, "loss": 0.4682, "step": 23118 }, { "epoch": 0.9691164603921486, "grad_norm": 1.7191836833953857, "learning_rate": 2.4942776555862412e-08, "loss": 0.4563, "step": 23119 }, { "epoch": 0.9691583790071576, "grad_norm": 2.184587240219116, "learning_rate": 2.4875099535858206e-08, "loss": 0.4723, "step": 23120 }, { "epoch": 0.9692002976221665, "grad_norm": 2.4200079441070557, "learning_rate": 2.480751422540395e-08, "loss": 0.4505, "step": 23121 }, { "epoch": 0.9692422162371755, "grad_norm": 1.836758017539978, "learning_rate": 2.4740020625744764e-08, "loss": 0.4326, "step": 23122 }, { "epoch": 0.9692841348521845, "grad_norm": 2.2871220111846924, "learning_rate": 2.4672618738125765e-08, "loss": 0.4519, "step": 23123 }, { "epoch": 0.9693260534671935, "grad_norm": 2.1390221118927, "learning_rate": 2.4605308563788733e-08, "loss": 0.4565, "step": 23124 }, { "epoch": 0.9693679720822024, "grad_norm": 1.8969520330429077, "learning_rate": 2.4538090103973787e-08, "loss": 0.4387, "step": 23125 }, { "epoch": 0.9694098906972114, "grad_norm": 5.820730209350586, "learning_rate": 2.4470963359921607e-08, "loss": 0.4782, "step": 23126 }, { "epoch": 0.9694518093122203, "grad_norm": 5.808808326721191, "learning_rate": 2.4403928332868422e-08, "loss": 0.4726, "step": 23127 }, { "epoch": 0.9694937279272293, "grad_norm": 2.1591339111328125, "learning_rate": 2.4336985024050465e-08, "loss": 0.4925, "step": 23128 }, { "epoch": 0.9695356465422382, "grad_norm": 1.609742522239685, "learning_rate": 2.4270133434701747e-08, "loss": 0.4059, "step": 23129 }, { "epoch": 0.9695775651572472, "grad_norm": 1.5566532611846924, "learning_rate": 2.4203373566054068e-08, "loss": 0.5032, "step": 23130 }, { "epoch": 0.9696194837722562, "grad_norm": 3.3448925018310547, "learning_rate": 2.41367054193381e-08, "loss": 0.4622, "step": 23131 }, { "epoch": 0.9696614023872652, "grad_norm": 1.7308681011199951, "learning_rate": 2.4070128995783425e-08, "loss": 0.443, "step": 23132 }, { "epoch": 0.9697033210022741, "grad_norm": 1.8485187292099, "learning_rate": 2.4003644296616836e-08, "loss": 0.4873, "step": 23133 }, { "epoch": 0.9697452396172831, "grad_norm": 1.6837316751480103, "learning_rate": 2.393725132306346e-08, "loss": 0.472, "step": 23134 }, { "epoch": 0.969787158232292, "grad_norm": 1.9107931852340698, "learning_rate": 2.3870950076348433e-08, "loss": 0.5214, "step": 23135 }, { "epoch": 0.969829076847301, "grad_norm": 1.4752094745635986, "learning_rate": 2.3804740557692994e-08, "loss": 0.4232, "step": 23136 }, { "epoch": 0.9698709954623099, "grad_norm": 1.9513365030288696, "learning_rate": 2.373862276831729e-08, "loss": 0.4675, "step": 23137 }, { "epoch": 0.9699129140773188, "grad_norm": 2.024791717529297, "learning_rate": 2.3672596709440885e-08, "loss": 0.4887, "step": 23138 }, { "epoch": 0.9699548326923279, "grad_norm": 1.8559805154800415, "learning_rate": 2.3606662382281155e-08, "loss": 0.4942, "step": 23139 }, { "epoch": 0.9699967513073368, "grad_norm": 2.193192958831787, "learning_rate": 2.3540819788052115e-08, "loss": 0.4667, "step": 23140 }, { "epoch": 0.9700386699223458, "grad_norm": 2.1447622776031494, "learning_rate": 2.3475068927968913e-08, "loss": 0.5149, "step": 23141 }, { "epoch": 0.9700805885373547, "grad_norm": 1.8097865581512451, "learning_rate": 2.3409409803242243e-08, "loss": 0.4734, "step": 23142 }, { "epoch": 0.9701225071523637, "grad_norm": 1.7190333604812622, "learning_rate": 2.334384241508336e-08, "loss": 0.4758, "step": 23143 }, { "epoch": 0.9701644257673726, "grad_norm": 9.27406120300293, "learning_rate": 2.3278366764701298e-08, "loss": 0.4916, "step": 23144 }, { "epoch": 0.9702063443823816, "grad_norm": 1.5824908018112183, "learning_rate": 2.3212982853302313e-08, "loss": 0.4819, "step": 23145 }, { "epoch": 0.9702482629973905, "grad_norm": 2.1203324794769287, "learning_rate": 2.3147690682091552e-08, "loss": 0.4773, "step": 23146 }, { "epoch": 0.9702901816123995, "grad_norm": 1.8929325342178345, "learning_rate": 2.3082490252272493e-08, "loss": 0.4654, "step": 23147 }, { "epoch": 0.9703321002274085, "grad_norm": 2.0421862602233887, "learning_rate": 2.3017381565048068e-08, "loss": 0.452, "step": 23148 }, { "epoch": 0.9703740188424175, "grad_norm": 1.7315162420272827, "learning_rate": 2.295236462161732e-08, "loss": 0.4226, "step": 23149 }, { "epoch": 0.9704159374574264, "grad_norm": 2.4449872970581055, "learning_rate": 2.288743942317928e-08, "loss": 0.4282, "step": 23150 }, { "epoch": 0.9704578560724354, "grad_norm": 2.2211191654205322, "learning_rate": 2.2822605970930222e-08, "loss": 0.4465, "step": 23151 }, { "epoch": 0.9704997746874443, "grad_norm": 2.6812498569488525, "learning_rate": 2.275786426606641e-08, "loss": 0.4833, "step": 23152 }, { "epoch": 0.9705416933024533, "grad_norm": 2.1043028831481934, "learning_rate": 2.269321430978022e-08, "loss": 0.5267, "step": 23153 }, { "epoch": 0.9705836119174622, "grad_norm": 1.6936429738998413, "learning_rate": 2.262865610326348e-08, "loss": 0.4615, "step": 23154 }, { "epoch": 0.9706255305324712, "grad_norm": 2.131096363067627, "learning_rate": 2.25641896477069e-08, "loss": 0.471, "step": 23155 }, { "epoch": 0.9706674491474802, "grad_norm": 2.0197556018829346, "learning_rate": 2.249981494429787e-08, "loss": 0.4989, "step": 23156 }, { "epoch": 0.9707093677624892, "grad_norm": 2.0227785110473633, "learning_rate": 2.243553199422377e-08, "loss": 0.4849, "step": 23157 }, { "epoch": 0.9707512863774981, "grad_norm": 1.8679498434066772, "learning_rate": 2.2371340798669207e-08, "loss": 0.4658, "step": 23158 }, { "epoch": 0.9707932049925071, "grad_norm": 2.0134804248809814, "learning_rate": 2.230724135881823e-08, "loss": 0.5396, "step": 23159 }, { "epoch": 0.970835123607516, "grad_norm": 1.9215747117996216, "learning_rate": 2.2243233675851017e-08, "loss": 0.5013, "step": 23160 }, { "epoch": 0.970877042222525, "grad_norm": 2.3774447441101074, "learning_rate": 2.2179317750948837e-08, "loss": 0.5063, "step": 23161 }, { "epoch": 0.9709189608375339, "grad_norm": 1.7103056907653809, "learning_rate": 2.2115493585289084e-08, "loss": 0.5243, "step": 23162 }, { "epoch": 0.9709608794525428, "grad_norm": 2.8404550552368164, "learning_rate": 2.2051761180048593e-08, "loss": 0.4555, "step": 23163 }, { "epoch": 0.9710027980675519, "grad_norm": 3.7239577770233154, "learning_rate": 2.198812053640198e-08, "loss": 0.4406, "step": 23164 }, { "epoch": 0.9710447166825608, "grad_norm": 1.9441417455673218, "learning_rate": 2.1924571655522198e-08, "loss": 0.471, "step": 23165 }, { "epoch": 0.9710866352975698, "grad_norm": 1.9152017831802368, "learning_rate": 2.1861114538581085e-08, "loss": 0.4641, "step": 23166 }, { "epoch": 0.9711285539125787, "grad_norm": 2.8810300827026367, "learning_rate": 2.179774918674882e-08, "loss": 0.532, "step": 23167 }, { "epoch": 0.9711704725275877, "grad_norm": 2.7565999031066895, "learning_rate": 2.173447560119224e-08, "loss": 0.4695, "step": 23168 }, { "epoch": 0.9712123911425966, "grad_norm": 1.9440550804138184, "learning_rate": 2.167129378307875e-08, "loss": 0.5134, "step": 23169 }, { "epoch": 0.9712543097576056, "grad_norm": 1.9890681505203247, "learning_rate": 2.160820373357242e-08, "loss": 0.4768, "step": 23170 }, { "epoch": 0.9712962283726145, "grad_norm": 2.287813425064087, "learning_rate": 2.1545205453836204e-08, "loss": 0.4759, "step": 23171 }, { "epoch": 0.9713381469876236, "grad_norm": 1.8348301649093628, "learning_rate": 2.1482298945031955e-08, "loss": 0.4938, "step": 23172 }, { "epoch": 0.9713800656026325, "grad_norm": 2.227430820465088, "learning_rate": 2.141948420831874e-08, "loss": 0.4675, "step": 23173 }, { "epoch": 0.9714219842176415, "grad_norm": 1.5391062498092651, "learning_rate": 2.1356761244855084e-08, "loss": 0.4437, "step": 23174 }, { "epoch": 0.9714639028326504, "grad_norm": 2.09887433052063, "learning_rate": 2.1294130055796168e-08, "loss": 0.4791, "step": 23175 }, { "epoch": 0.9715058214476594, "grad_norm": 1.7943755388259888, "learning_rate": 2.1231590642297738e-08, "loss": 0.4184, "step": 23176 }, { "epoch": 0.9715477400626683, "grad_norm": 2.083240032196045, "learning_rate": 2.116914300551165e-08, "loss": 0.479, "step": 23177 }, { "epoch": 0.9715896586776773, "grad_norm": 2.190819025039673, "learning_rate": 2.1106787146589203e-08, "loss": 0.548, "step": 23178 }, { "epoch": 0.9716315772926862, "grad_norm": 1.892274022102356, "learning_rate": 2.1044523066680033e-08, "loss": 0.4879, "step": 23179 }, { "epoch": 0.9716734959076953, "grad_norm": 2.0123817920684814, "learning_rate": 2.0982350766932113e-08, "loss": 0.5214, "step": 23180 }, { "epoch": 0.9717154145227042, "grad_norm": 4.772147178649902, "learning_rate": 2.0920270248491192e-08, "loss": 0.5125, "step": 23181 }, { "epoch": 0.9717573331377132, "grad_norm": 1.923891305923462, "learning_rate": 2.085828151250191e-08, "loss": 0.4699, "step": 23182 }, { "epoch": 0.9717992517527221, "grad_norm": 1.9680289030075073, "learning_rate": 2.079638456010613e-08, "loss": 0.4613, "step": 23183 }, { "epoch": 0.9718411703677311, "grad_norm": 1.891627311706543, "learning_rate": 2.0734579392446274e-08, "loss": 0.462, "step": 23184 }, { "epoch": 0.97188308898274, "grad_norm": 2.276413917541504, "learning_rate": 2.067286601066032e-08, "loss": 0.5327, "step": 23185 }, { "epoch": 0.971925007597749, "grad_norm": 2.0894951820373535, "learning_rate": 2.06112444158868e-08, "loss": 0.456, "step": 23186 }, { "epoch": 0.9719669262127579, "grad_norm": 1.9476925134658813, "learning_rate": 2.054971460926092e-08, "loss": 0.4786, "step": 23187 }, { "epoch": 0.9720088448277668, "grad_norm": 2.2456159591674805, "learning_rate": 2.0488276591916768e-08, "loss": 0.5161, "step": 23188 }, { "epoch": 0.9720507634427759, "grad_norm": 2.1754488945007324, "learning_rate": 2.0426930364987886e-08, "loss": 0.436, "step": 23189 }, { "epoch": 0.9720926820577848, "grad_norm": 1.7574536800384521, "learning_rate": 2.0365675929603923e-08, "loss": 0.4396, "step": 23190 }, { "epoch": 0.9721346006727938, "grad_norm": 5.722115516662598, "learning_rate": 2.0304513286895088e-08, "loss": 0.5065, "step": 23191 }, { "epoch": 0.9721765192878027, "grad_norm": 1.7859364748001099, "learning_rate": 2.024344243798826e-08, "loss": 0.4609, "step": 23192 }, { "epoch": 0.9722184379028117, "grad_norm": 1.9349421262741089, "learning_rate": 2.0182463384008644e-08, "loss": 0.4862, "step": 23193 }, { "epoch": 0.9722603565178206, "grad_norm": 2.7837276458740234, "learning_rate": 2.012157612608201e-08, "loss": 0.4836, "step": 23194 }, { "epoch": 0.9723022751328296, "grad_norm": 2.0082008838653564, "learning_rate": 2.0060780665329126e-08, "loss": 0.4471, "step": 23195 }, { "epoch": 0.9723441937478385, "grad_norm": 2.2421693801879883, "learning_rate": 2.0000077002870765e-08, "loss": 0.4607, "step": 23196 }, { "epoch": 0.9723861123628476, "grad_norm": 1.7465649843215942, "learning_rate": 1.993946513982714e-08, "loss": 0.4342, "step": 23197 }, { "epoch": 0.9724280309778565, "grad_norm": 1.700892448425293, "learning_rate": 1.9878945077314025e-08, "loss": 0.4879, "step": 23198 }, { "epoch": 0.9724699495928655, "grad_norm": 2.366478204727173, "learning_rate": 1.9818516816447754e-08, "loss": 0.4749, "step": 23199 }, { "epoch": 0.9725118682078744, "grad_norm": 1.7006101608276367, "learning_rate": 1.975818035834298e-08, "loss": 0.5075, "step": 23200 }, { "epoch": 0.9725537868228834, "grad_norm": 2.1105666160583496, "learning_rate": 1.9697935704110492e-08, "loss": 0.4507, "step": 23201 }, { "epoch": 0.9725957054378923, "grad_norm": 2.1634371280670166, "learning_rate": 1.963778285486162e-08, "loss": 0.431, "step": 23202 }, { "epoch": 0.9726376240529013, "grad_norm": 1.9409042596817017, "learning_rate": 1.957772181170492e-08, "loss": 0.4561, "step": 23203 }, { "epoch": 0.9726795426679102, "grad_norm": 4.1139326095581055, "learning_rate": 1.951775257574784e-08, "loss": 0.4936, "step": 23204 }, { "epoch": 0.9727214612829193, "grad_norm": 1.7607005834579468, "learning_rate": 1.945787514809616e-08, "loss": 0.4839, "step": 23205 }, { "epoch": 0.9727633798979282, "grad_norm": 2.0943994522094727, "learning_rate": 1.9398089529852892e-08, "loss": 0.468, "step": 23206 }, { "epoch": 0.9728052985129372, "grad_norm": 1.7538689374923706, "learning_rate": 1.933839572211993e-08, "loss": 0.4365, "step": 23207 }, { "epoch": 0.9728472171279461, "grad_norm": 2.2241716384887695, "learning_rate": 1.9278793725998613e-08, "loss": 0.4492, "step": 23208 }, { "epoch": 0.9728891357429551, "grad_norm": 1.9807803630828857, "learning_rate": 1.921928354258695e-08, "loss": 0.4729, "step": 23209 }, { "epoch": 0.972931054357964, "grad_norm": 2.739584445953369, "learning_rate": 1.915986517298185e-08, "loss": 0.5329, "step": 23210 }, { "epoch": 0.972972972972973, "grad_norm": 1.8986157178878784, "learning_rate": 1.9100538618279095e-08, "loss": 0.4488, "step": 23211 }, { "epoch": 0.9730148915879819, "grad_norm": 2.0333571434020996, "learning_rate": 1.9041303879571704e-08, "loss": 0.5013, "step": 23212 }, { "epoch": 0.9730568102029908, "grad_norm": 1.8347163200378418, "learning_rate": 1.8982160957952132e-08, "loss": 0.5055, "step": 23213 }, { "epoch": 0.9730987288179999, "grad_norm": 1.6458377838134766, "learning_rate": 1.8923109854510625e-08, "loss": 0.4697, "step": 23214 }, { "epoch": 0.9731406474330088, "grad_norm": 3.0152571201324463, "learning_rate": 1.8864150570335193e-08, "loss": 0.4991, "step": 23215 }, { "epoch": 0.9731825660480178, "grad_norm": 1.8309303522109985, "learning_rate": 1.880528310651275e-08, "loss": 0.4753, "step": 23216 }, { "epoch": 0.9732244846630267, "grad_norm": 2.007481575012207, "learning_rate": 1.8746507464128538e-08, "loss": 0.4656, "step": 23217 }, { "epoch": 0.9732664032780357, "grad_norm": 1.8439831733703613, "learning_rate": 1.8687823644266688e-08, "loss": 0.4644, "step": 23218 }, { "epoch": 0.9733083218930446, "grad_norm": 1.9817907810211182, "learning_rate": 1.8629231648008006e-08, "loss": 0.467, "step": 23219 }, { "epoch": 0.9733502405080536, "grad_norm": 2.303412675857544, "learning_rate": 1.8570731476432734e-08, "loss": 0.4544, "step": 23220 }, { "epoch": 0.9733921591230625, "grad_norm": 2.0091583728790283, "learning_rate": 1.851232313061946e-08, "loss": 0.511, "step": 23221 }, { "epoch": 0.9734340777380716, "grad_norm": 2.1569604873657227, "learning_rate": 1.845400661164509e-08, "loss": 0.4588, "step": 23222 }, { "epoch": 0.9734759963530805, "grad_norm": 2.0204873085021973, "learning_rate": 1.8395781920584333e-08, "loss": 0.466, "step": 23223 }, { "epoch": 0.9735179149680895, "grad_norm": 1.6871132850646973, "learning_rate": 1.833764905851021e-08, "loss": 0.4662, "step": 23224 }, { "epoch": 0.9735598335830984, "grad_norm": 2.0313403606414795, "learning_rate": 1.8279608026494644e-08, "loss": 0.4771, "step": 23225 }, { "epoch": 0.9736017521981074, "grad_norm": 1.8353666067123413, "learning_rate": 1.822165882560789e-08, "loss": 0.4614, "step": 23226 }, { "epoch": 0.9736436708131163, "grad_norm": 2.0844743251800537, "learning_rate": 1.8163801456916874e-08, "loss": 0.5002, "step": 23227 }, { "epoch": 0.9736855894281253, "grad_norm": 1.999202013015747, "learning_rate": 1.810603592148963e-08, "loss": 0.4923, "step": 23228 }, { "epoch": 0.9737275080431342, "grad_norm": 1.6248198747634888, "learning_rate": 1.8048362220390303e-08, "loss": 0.4803, "step": 23229 }, { "epoch": 0.9737694266581433, "grad_norm": 1.6752002239227295, "learning_rate": 1.7990780354681936e-08, "loss": 0.4705, "step": 23230 }, { "epoch": 0.9738113452731522, "grad_norm": 2.275857925415039, "learning_rate": 1.79332903254259e-08, "loss": 0.4867, "step": 23231 }, { "epoch": 0.9738532638881612, "grad_norm": 1.8131211996078491, "learning_rate": 1.7875892133683014e-08, "loss": 0.4726, "step": 23232 }, { "epoch": 0.9738951825031701, "grad_norm": 1.9147043228149414, "learning_rate": 1.7818585780509655e-08, "loss": 0.5033, "step": 23233 }, { "epoch": 0.9739371011181791, "grad_norm": 2.0905511379241943, "learning_rate": 1.7761371266963313e-08, "loss": 0.506, "step": 23234 }, { "epoch": 0.973979019733188, "grad_norm": 1.7461379766464233, "learning_rate": 1.770424859409814e-08, "loss": 0.4615, "step": 23235 }, { "epoch": 0.974020938348197, "grad_norm": 1.779057264328003, "learning_rate": 1.764721776296774e-08, "loss": 0.471, "step": 23236 }, { "epoch": 0.9740628569632059, "grad_norm": 2.0824668407440186, "learning_rate": 1.7590278774622384e-08, "loss": 0.4784, "step": 23237 }, { "epoch": 0.9741047755782148, "grad_norm": 1.9382972717285156, "learning_rate": 1.7533431630112896e-08, "loss": 0.4855, "step": 23238 }, { "epoch": 0.9741466941932239, "grad_norm": 1.6266200542449951, "learning_rate": 1.747667633048622e-08, "loss": 0.4452, "step": 23239 }, { "epoch": 0.9741886128082328, "grad_norm": 2.6602706909179688, "learning_rate": 1.742001287678874e-08, "loss": 0.5106, "step": 23240 }, { "epoch": 0.9742305314232418, "grad_norm": 1.8987151384353638, "learning_rate": 1.7363441270065172e-08, "loss": 0.4914, "step": 23241 }, { "epoch": 0.9742724500382507, "grad_norm": 1.8271931409835815, "learning_rate": 1.7306961511358576e-08, "loss": 0.4773, "step": 23242 }, { "epoch": 0.9743143686532597, "grad_norm": 1.8488247394561768, "learning_rate": 1.7250573601709785e-08, "loss": 0.4702, "step": 23243 }, { "epoch": 0.9743562872682686, "grad_norm": 1.7618063688278198, "learning_rate": 1.7194277542157967e-08, "loss": 0.4513, "step": 23244 }, { "epoch": 0.9743982058832776, "grad_norm": 1.6477137804031372, "learning_rate": 1.713807333374118e-08, "loss": 0.4377, "step": 23245 }, { "epoch": 0.9744401244982865, "grad_norm": 2.5730528831481934, "learning_rate": 1.708196097749526e-08, "loss": 0.4074, "step": 23246 }, { "epoch": 0.9744820431132956, "grad_norm": 2.170966625213623, "learning_rate": 1.7025940474454938e-08, "loss": 0.5251, "step": 23247 }, { "epoch": 0.9745239617283045, "grad_norm": 1.950019359588623, "learning_rate": 1.697001182565272e-08, "loss": 0.4899, "step": 23248 }, { "epoch": 0.9745658803433135, "grad_norm": 1.5274758338928223, "learning_rate": 1.6914175032118895e-08, "loss": 0.4261, "step": 23249 }, { "epoch": 0.9746077989583224, "grad_norm": 2.3949339389801025, "learning_rate": 1.6858430094884302e-08, "loss": 0.4586, "step": 23250 }, { "epoch": 0.9746497175733314, "grad_norm": 1.687137246131897, "learning_rate": 1.680277701497479e-08, "loss": 0.448, "step": 23251 }, { "epoch": 0.9746916361883403, "grad_norm": 3.6267573833465576, "learning_rate": 1.6747215793417314e-08, "loss": 0.4967, "step": 23252 }, { "epoch": 0.9747335548033493, "grad_norm": 1.729910969734192, "learning_rate": 1.6691746431236054e-08, "loss": 0.4612, "step": 23253 }, { "epoch": 0.9747754734183582, "grad_norm": 2.158219575881958, "learning_rate": 1.6636368929452974e-08, "loss": 0.4742, "step": 23254 }, { "epoch": 0.9748173920333673, "grad_norm": 1.9979889392852783, "learning_rate": 1.658108328908892e-08, "loss": 0.5198, "step": 23255 }, { "epoch": 0.9748593106483762, "grad_norm": 2.3007586002349854, "learning_rate": 1.6525889511163075e-08, "loss": 0.4661, "step": 23256 }, { "epoch": 0.9749012292633852, "grad_norm": 4.643905162811279, "learning_rate": 1.6470787596693515e-08, "loss": 0.4542, "step": 23257 }, { "epoch": 0.9749431478783941, "grad_norm": 1.9863723516464233, "learning_rate": 1.6415777546694987e-08, "loss": 0.4631, "step": 23258 }, { "epoch": 0.9749850664934031, "grad_norm": 2.327122211456299, "learning_rate": 1.636085936218279e-08, "loss": 0.4668, "step": 23259 }, { "epoch": 0.975026985108412, "grad_norm": 1.9160916805267334, "learning_rate": 1.630603304416778e-08, "loss": 0.5173, "step": 23260 }, { "epoch": 0.975068903723421, "grad_norm": 1.7493345737457275, "learning_rate": 1.6251298593661368e-08, "loss": 0.4824, "step": 23261 }, { "epoch": 0.9751108223384299, "grad_norm": 1.6674269437789917, "learning_rate": 1.619665601167275e-08, "loss": 0.5018, "step": 23262 }, { "epoch": 0.9751527409534388, "grad_norm": 1.8955438137054443, "learning_rate": 1.61421052992089e-08, "loss": 0.4681, "step": 23263 }, { "epoch": 0.9751946595684479, "grad_norm": 1.7870121002197266, "learning_rate": 1.608764645727512e-08, "loss": 0.4545, "step": 23264 }, { "epoch": 0.9752365781834568, "grad_norm": 1.9017333984375, "learning_rate": 1.603327948687561e-08, "loss": 0.4228, "step": 23265 }, { "epoch": 0.9752784967984658, "grad_norm": 2.3644320964813232, "learning_rate": 1.5979004389012343e-08, "loss": 0.4965, "step": 23266 }, { "epoch": 0.9753204154134747, "grad_norm": 3.4489264488220215, "learning_rate": 1.5924821164685633e-08, "loss": 0.5356, "step": 23267 }, { "epoch": 0.9753623340284837, "grad_norm": 1.6231855154037476, "learning_rate": 1.5870729814895237e-08, "loss": 0.4513, "step": 23268 }, { "epoch": 0.9754042526434926, "grad_norm": 2.377136707305908, "learning_rate": 1.5816730340637577e-08, "loss": 0.4796, "step": 23269 }, { "epoch": 0.9754461712585016, "grad_norm": 2.007859945297241, "learning_rate": 1.5762822742907412e-08, "loss": 0.5103, "step": 23270 }, { "epoch": 0.9754880898735105, "grad_norm": 3.1532444953918457, "learning_rate": 1.5709007022700064e-08, "loss": 0.483, "step": 23271 }, { "epoch": 0.9755300084885196, "grad_norm": 1.9881954193115234, "learning_rate": 1.5655283181005843e-08, "loss": 0.4789, "step": 23272 }, { "epoch": 0.9755719271035285, "grad_norm": 2.260802984237671, "learning_rate": 1.5601651218815627e-08, "loss": 0.5125, "step": 23273 }, { "epoch": 0.9756138457185375, "grad_norm": 2.212503671646118, "learning_rate": 1.5548111137118627e-08, "loss": 0.465, "step": 23274 }, { "epoch": 0.9756557643335464, "grad_norm": 1.6524657011032104, "learning_rate": 1.5494662936901272e-08, "loss": 0.3888, "step": 23275 }, { "epoch": 0.9756976829485554, "grad_norm": 1.6587227582931519, "learning_rate": 1.5441306619149444e-08, "loss": 0.4244, "step": 23276 }, { "epoch": 0.9757396015635643, "grad_norm": 1.5821763277053833, "learning_rate": 1.5388042184845686e-08, "loss": 0.4605, "step": 23277 }, { "epoch": 0.9757815201785733, "grad_norm": 1.9994169473648071, "learning_rate": 1.533486963497255e-08, "loss": 0.4244, "step": 23278 }, { "epoch": 0.9758234387935822, "grad_norm": 2.283052921295166, "learning_rate": 1.5281788970509803e-08, "loss": 0.4912, "step": 23279 }, { "epoch": 0.9758653574085913, "grad_norm": 2.6111397743225098, "learning_rate": 1.522880019243611e-08, "loss": 0.4729, "step": 23280 }, { "epoch": 0.9759072760236002, "grad_norm": 2.6798861026763916, "learning_rate": 1.5175903301728467e-08, "loss": 0.4693, "step": 23281 }, { "epoch": 0.9759491946386092, "grad_norm": 1.8301769495010376, "learning_rate": 1.512309829936165e-08, "loss": 0.4744, "step": 23282 }, { "epoch": 0.9759911132536181, "grad_norm": 1.92431640625, "learning_rate": 1.5070385186309323e-08, "loss": 0.4765, "step": 23283 }, { "epoch": 0.9760330318686271, "grad_norm": 1.852213978767395, "learning_rate": 1.5017763963542375e-08, "loss": 0.4406, "step": 23284 }, { "epoch": 0.976074950483636, "grad_norm": 2.0046675205230713, "learning_rate": 1.49652346320317e-08, "loss": 0.4504, "step": 23285 }, { "epoch": 0.976116869098645, "grad_norm": 1.5454633235931396, "learning_rate": 1.4912797192745964e-08, "loss": 0.4195, "step": 23286 }, { "epoch": 0.9761587877136539, "grad_norm": 2.7791543006896973, "learning_rate": 1.4860451646650509e-08, "loss": 0.4911, "step": 23287 }, { "epoch": 0.9762007063286628, "grad_norm": 1.5501773357391357, "learning_rate": 1.4808197994710671e-08, "loss": 0.4249, "step": 23288 }, { "epoch": 0.9762426249436719, "grad_norm": 1.8645890951156616, "learning_rate": 1.4756036237890126e-08, "loss": 0.4676, "step": 23289 }, { "epoch": 0.9762845435586808, "grad_norm": 1.9467506408691406, "learning_rate": 1.4703966377150325e-08, "loss": 0.4838, "step": 23290 }, { "epoch": 0.9763264621736898, "grad_norm": 2.2667081356048584, "learning_rate": 1.4651988413451058e-08, "loss": 0.5032, "step": 23291 }, { "epoch": 0.9763683807886987, "grad_norm": 1.935326337814331, "learning_rate": 1.4600102347749889e-08, "loss": 0.4739, "step": 23292 }, { "epoch": 0.9764102994037077, "grad_norm": 1.846099615097046, "learning_rate": 1.4548308181003833e-08, "loss": 0.4739, "step": 23293 }, { "epoch": 0.9764522180187166, "grad_norm": 2.0384929180145264, "learning_rate": 1.4496605914168238e-08, "loss": 0.4591, "step": 23294 }, { "epoch": 0.9764941366337256, "grad_norm": 2.616171360015869, "learning_rate": 1.4444995548194563e-08, "loss": 0.489, "step": 23295 }, { "epoch": 0.9765360552487345, "grad_norm": 2.063931941986084, "learning_rate": 1.439347708403538e-08, "loss": 0.4759, "step": 23296 }, { "epoch": 0.9765779738637436, "grad_norm": 1.8043402433395386, "learning_rate": 1.4342050522639928e-08, "loss": 0.5551, "step": 23297 }, { "epoch": 0.9766198924787525, "grad_norm": 2.8514773845672607, "learning_rate": 1.4290715864956339e-08, "loss": 0.4625, "step": 23298 }, { "epoch": 0.9766618110937615, "grad_norm": 1.6499474048614502, "learning_rate": 1.4239473111931079e-08, "loss": 0.4126, "step": 23299 }, { "epoch": 0.9767037297087704, "grad_norm": 1.8632371425628662, "learning_rate": 1.4188322264508391e-08, "loss": 0.4936, "step": 23300 }, { "epoch": 0.9767456483237794, "grad_norm": 2.3224027156829834, "learning_rate": 1.4137263323631411e-08, "loss": 0.5184, "step": 23301 }, { "epoch": 0.9767875669387883, "grad_norm": 1.6123589277267456, "learning_rate": 1.4086296290241054e-08, "loss": 0.511, "step": 23302 }, { "epoch": 0.9768294855537973, "grad_norm": 2.0223655700683594, "learning_rate": 1.4035421165277118e-08, "loss": 0.4692, "step": 23303 }, { "epoch": 0.9768714041688062, "grad_norm": 1.774199366569519, "learning_rate": 1.3984637949677193e-08, "loss": 0.4307, "step": 23304 }, { "epoch": 0.9769133227838153, "grad_norm": 1.983830213546753, "learning_rate": 1.3933946644377749e-08, "loss": 0.4714, "step": 23305 }, { "epoch": 0.9769552413988242, "grad_norm": 1.9382747411727905, "learning_rate": 1.388334725031304e-08, "loss": 0.5255, "step": 23306 }, { "epoch": 0.9769971600138332, "grad_norm": 2.0381858348846436, "learning_rate": 1.3832839768415097e-08, "loss": 0.4772, "step": 23307 }, { "epoch": 0.9770390786288421, "grad_norm": 2.2529566287994385, "learning_rate": 1.3782424199616506e-08, "loss": 0.4912, "step": 23308 }, { "epoch": 0.9770809972438511, "grad_norm": 2.330075263977051, "learning_rate": 1.3732100544844861e-08, "loss": 0.4095, "step": 23309 }, { "epoch": 0.97712291585886, "grad_norm": 2.878650188446045, "learning_rate": 1.3681868805028864e-08, "loss": 0.4914, "step": 23310 }, { "epoch": 0.977164834473869, "grad_norm": 1.7799186706542969, "learning_rate": 1.3631728981094439e-08, "loss": 0.4998, "step": 23311 }, { "epoch": 0.9772067530888779, "grad_norm": 2.233416795730591, "learning_rate": 1.3581681073965847e-08, "loss": 0.4708, "step": 23312 }, { "epoch": 0.9772486717038869, "grad_norm": 2.6411185264587402, "learning_rate": 1.3531725084565127e-08, "loss": 0.4555, "step": 23313 }, { "epoch": 0.9772905903188959, "grad_norm": 2.771308422088623, "learning_rate": 1.3481861013813213e-08, "loss": 0.4771, "step": 23314 }, { "epoch": 0.9773325089339048, "grad_norm": 1.8889702558517456, "learning_rate": 1.3432088862629921e-08, "loss": 0.4677, "step": 23315 }, { "epoch": 0.9773744275489138, "grad_norm": 1.8111658096313477, "learning_rate": 1.3382408631932298e-08, "loss": 0.5037, "step": 23316 }, { "epoch": 0.9774163461639227, "grad_norm": 1.7829316854476929, "learning_rate": 1.333282032263572e-08, "loss": 0.4566, "step": 23317 }, { "epoch": 0.9774582647789317, "grad_norm": 2.2088327407836914, "learning_rate": 1.3283323935655567e-08, "loss": 0.4632, "step": 23318 }, { "epoch": 0.9775001833939406, "grad_norm": 2.5196573734283447, "learning_rate": 1.323391947190278e-08, "loss": 0.4638, "step": 23319 }, { "epoch": 0.9775421020089496, "grad_norm": 1.8087233304977417, "learning_rate": 1.3184606932288845e-08, "loss": 0.4419, "step": 23320 }, { "epoch": 0.9775840206239585, "grad_norm": 1.716131329536438, "learning_rate": 1.3135386317722487e-08, "loss": 0.4644, "step": 23321 }, { "epoch": 0.9776259392389676, "grad_norm": 1.8499200344085693, "learning_rate": 1.308625762911131e-08, "loss": 0.5053, "step": 23322 }, { "epoch": 0.9776678578539765, "grad_norm": 1.928640365600586, "learning_rate": 1.3037220867360701e-08, "loss": 0.4735, "step": 23323 }, { "epoch": 0.9777097764689855, "grad_norm": 1.7602266073226929, "learning_rate": 1.2988276033374935e-08, "loss": 0.4851, "step": 23324 }, { "epoch": 0.9777516950839944, "grad_norm": 1.7423019409179688, "learning_rate": 1.2939423128055516e-08, "loss": 0.4551, "step": 23325 }, { "epoch": 0.9777936136990034, "grad_norm": 2.9462594985961914, "learning_rate": 1.2890662152303946e-08, "loss": 0.4644, "step": 23326 }, { "epoch": 0.9778355323140123, "grad_norm": 1.7140744924545288, "learning_rate": 1.284199310701839e-08, "loss": 0.4807, "step": 23327 }, { "epoch": 0.9778774509290213, "grad_norm": 2.113856792449951, "learning_rate": 1.2793415993095915e-08, "loss": 0.5004, "step": 23328 }, { "epoch": 0.9779193695440302, "grad_norm": 2.4054982662200928, "learning_rate": 1.2744930811432465e-08, "loss": 0.4916, "step": 23329 }, { "epoch": 0.9779612881590393, "grad_norm": 3.189000368118286, "learning_rate": 1.2696537562921218e-08, "loss": 0.4856, "step": 23330 }, { "epoch": 0.9780032067740482, "grad_norm": 2.11984920501709, "learning_rate": 1.2648236248454792e-08, "loss": 0.4778, "step": 23331 }, { "epoch": 0.9780451253890572, "grad_norm": 1.9728256464004517, "learning_rate": 1.2600026868923032e-08, "loss": 0.4959, "step": 23332 }, { "epoch": 0.9780870440040661, "grad_norm": 1.9486638307571411, "learning_rate": 1.2551909425215226e-08, "loss": 0.493, "step": 23333 }, { "epoch": 0.9781289626190751, "grad_norm": 1.9337666034698486, "learning_rate": 1.2503883918217885e-08, "loss": 0.4755, "step": 23334 }, { "epoch": 0.978170881234084, "grad_norm": 1.9182732105255127, "learning_rate": 1.245595034881697e-08, "loss": 0.4184, "step": 23335 }, { "epoch": 0.978212799849093, "grad_norm": 2.2663731575012207, "learning_rate": 1.2408108717894551e-08, "loss": 0.5068, "step": 23336 }, { "epoch": 0.978254718464102, "grad_norm": 1.9940601587295532, "learning_rate": 1.2360359026334368e-08, "loss": 0.5144, "step": 23337 }, { "epoch": 0.9782966370791109, "grad_norm": 2.7537341117858887, "learning_rate": 1.231270127501516e-08, "loss": 0.4439, "step": 23338 }, { "epoch": 0.9783385556941199, "grad_norm": 1.9544068574905396, "learning_rate": 1.226513546481678e-08, "loss": 0.4735, "step": 23339 }, { "epoch": 0.9783804743091288, "grad_norm": 1.9147611856460571, "learning_rate": 1.2217661596614638e-08, "loss": 0.4427, "step": 23340 }, { "epoch": 0.9784223929241378, "grad_norm": 6.821877956390381, "learning_rate": 1.2170279671284702e-08, "loss": 0.4213, "step": 23341 }, { "epoch": 0.9784643115391467, "grad_norm": 2.3268418312072754, "learning_rate": 1.212298968970016e-08, "loss": 0.4393, "step": 23342 }, { "epoch": 0.9785062301541557, "grad_norm": 2.277528762817383, "learning_rate": 1.2075791652732539e-08, "loss": 0.4548, "step": 23343 }, { "epoch": 0.9785481487691646, "grad_norm": 2.2036454677581787, "learning_rate": 1.2028685561252252e-08, "loss": 0.4798, "step": 23344 }, { "epoch": 0.9785900673841736, "grad_norm": 1.7902772426605225, "learning_rate": 1.1981671416127495e-08, "loss": 0.4709, "step": 23345 }, { "epoch": 0.9786319859991826, "grad_norm": 2.7090930938720703, "learning_rate": 1.193474921822535e-08, "loss": 0.5251, "step": 23346 }, { "epoch": 0.9786739046141916, "grad_norm": 4.608430862426758, "learning_rate": 1.1887918968409573e-08, "loss": 0.4795, "step": 23347 }, { "epoch": 0.9787158232292005, "grad_norm": 1.8476877212524414, "learning_rate": 1.1841180667544471e-08, "loss": 0.4277, "step": 23348 }, { "epoch": 0.9787577418442095, "grad_norm": 1.8748034238815308, "learning_rate": 1.1794534316491579e-08, "loss": 0.4665, "step": 23349 }, { "epoch": 0.9787996604592184, "grad_norm": 1.8382668495178223, "learning_rate": 1.1747979916110208e-08, "loss": 0.4943, "step": 23350 }, { "epoch": 0.9788415790742274, "grad_norm": 2.2093594074249268, "learning_rate": 1.1701517467258561e-08, "loss": 0.4446, "step": 23351 }, { "epoch": 0.9788834976892363, "grad_norm": 1.6760382652282715, "learning_rate": 1.165514697079373e-08, "loss": 0.5127, "step": 23352 }, { "epoch": 0.9789254163042453, "grad_norm": 2.592193841934204, "learning_rate": 1.1608868427569476e-08, "loss": 0.4823, "step": 23353 }, { "epoch": 0.9789673349192543, "grad_norm": 1.6619973182678223, "learning_rate": 1.1562681838440116e-08, "loss": 0.4425, "step": 23354 }, { "epoch": 0.9790092535342633, "grad_norm": 1.7918891906738281, "learning_rate": 1.151658720425608e-08, "loss": 0.483, "step": 23355 }, { "epoch": 0.9790511721492722, "grad_norm": 2.255596876144409, "learning_rate": 1.1470584525867801e-08, "loss": 0.4985, "step": 23356 }, { "epoch": 0.9790930907642812, "grad_norm": 1.8017441034317017, "learning_rate": 1.1424673804122377e-08, "loss": 0.5381, "step": 23357 }, { "epoch": 0.9791350093792901, "grad_norm": 1.9318586587905884, "learning_rate": 1.137885503986691e-08, "loss": 0.4846, "step": 23358 }, { "epoch": 0.9791769279942991, "grad_norm": 1.8677846193313599, "learning_rate": 1.1333128233945167e-08, "loss": 0.4568, "step": 23359 }, { "epoch": 0.979218846609308, "grad_norm": 1.786884069442749, "learning_rate": 1.1287493387201477e-08, "loss": 0.4027, "step": 23360 }, { "epoch": 0.979260765224317, "grad_norm": 2.040708303451538, "learning_rate": 1.1241950500475162e-08, "loss": 0.4462, "step": 23361 }, { "epoch": 0.979302683839326, "grad_norm": 1.7436354160308838, "learning_rate": 1.1196499574607223e-08, "loss": 0.4114, "step": 23362 }, { "epoch": 0.9793446024543349, "grad_norm": 2.286771297454834, "learning_rate": 1.1151140610435319e-08, "loss": 0.4655, "step": 23363 }, { "epoch": 0.9793865210693439, "grad_norm": 2.0655622482299805, "learning_rate": 1.1105873608794893e-08, "loss": 0.4861, "step": 23364 }, { "epoch": 0.9794284396843528, "grad_norm": 10.967658996582031, "learning_rate": 1.1060698570520834e-08, "loss": 0.4811, "step": 23365 }, { "epoch": 0.9794703582993618, "grad_norm": 1.7393162250518799, "learning_rate": 1.1015615496445809e-08, "loss": 0.4644, "step": 23366 }, { "epoch": 0.9795122769143707, "grad_norm": 2.969665288925171, "learning_rate": 1.0970624387400819e-08, "loss": 0.5459, "step": 23367 }, { "epoch": 0.9795541955293797, "grad_norm": 2.4339635372161865, "learning_rate": 1.0925725244215202e-08, "loss": 0.4709, "step": 23368 }, { "epoch": 0.9795961141443886, "grad_norm": 1.9620815515518188, "learning_rate": 1.0880918067716628e-08, "loss": 0.4626, "step": 23369 }, { "epoch": 0.9796380327593976, "grad_norm": 2.710456132888794, "learning_rate": 1.0836202858731104e-08, "loss": 0.4196, "step": 23370 }, { "epoch": 0.9796799513744066, "grad_norm": 1.9290212392807007, "learning_rate": 1.079157961808297e-08, "loss": 0.4716, "step": 23371 }, { "epoch": 0.9797218699894156, "grad_norm": 1.7060036659240723, "learning_rate": 1.0747048346594347e-08, "loss": 0.4531, "step": 23372 }, { "epoch": 0.9797637886044245, "grad_norm": 1.8082844018936157, "learning_rate": 1.0702609045086799e-08, "loss": 0.4175, "step": 23373 }, { "epoch": 0.9798057072194335, "grad_norm": 1.810836911201477, "learning_rate": 1.0658261714379115e-08, "loss": 0.4438, "step": 23374 }, { "epoch": 0.9798476258344424, "grad_norm": 1.746842861175537, "learning_rate": 1.0614006355288974e-08, "loss": 0.4925, "step": 23375 }, { "epoch": 0.9798895444494514, "grad_norm": 2.5396645069122314, "learning_rate": 1.0569842968631838e-08, "loss": 0.4532, "step": 23376 }, { "epoch": 0.9799314630644603, "grad_norm": 1.6906238794326782, "learning_rate": 1.0525771555221498e-08, "loss": 0.4853, "step": 23377 }, { "epoch": 0.9799733816794693, "grad_norm": 2.081613302230835, "learning_rate": 1.0481792115871193e-08, "loss": 0.4519, "step": 23378 }, { "epoch": 0.9800153002944783, "grad_norm": 2.1298911571502686, "learning_rate": 1.0437904651391385e-08, "loss": 0.4435, "step": 23379 }, { "epoch": 0.9800572189094873, "grad_norm": 2.073361396789551, "learning_rate": 1.0394109162590871e-08, "loss": 0.493, "step": 23380 }, { "epoch": 0.9800991375244962, "grad_norm": 1.8274564743041992, "learning_rate": 1.035040565027734e-08, "loss": 0.5135, "step": 23381 }, { "epoch": 0.9801410561395052, "grad_norm": 2.1869165897369385, "learning_rate": 1.0306794115255703e-08, "loss": 0.4533, "step": 23382 }, { "epoch": 0.9801829747545141, "grad_norm": 1.9023211002349854, "learning_rate": 1.0263274558330316e-08, "loss": 0.4386, "step": 23383 }, { "epoch": 0.9802248933695231, "grad_norm": 1.6697850227355957, "learning_rate": 1.0219846980303317e-08, "loss": 0.4653, "step": 23384 }, { "epoch": 0.980266811984532, "grad_norm": 1.8741761445999146, "learning_rate": 1.0176511381975174e-08, "loss": 0.4614, "step": 23385 }, { "epoch": 0.980308730599541, "grad_norm": 1.7788392305374146, "learning_rate": 1.0133267764145249e-08, "loss": 0.4736, "step": 23386 }, { "epoch": 0.98035064921455, "grad_norm": 2.346895217895508, "learning_rate": 1.0090116127610128e-08, "loss": 0.4911, "step": 23387 }, { "epoch": 0.9803925678295589, "grad_norm": 2.7853713035583496, "learning_rate": 1.004705647316473e-08, "loss": 0.4951, "step": 23388 }, { "epoch": 0.9804344864445679, "grad_norm": 2.2179481983184814, "learning_rate": 1.0004088801604529e-08, "loss": 0.4806, "step": 23389 }, { "epoch": 0.9804764050595768, "grad_norm": 1.8852410316467285, "learning_rate": 9.961213113719448e-09, "loss": 0.4679, "step": 23390 }, { "epoch": 0.9805183236745858, "grad_norm": 8.655428886413574, "learning_rate": 9.918429410301634e-09, "loss": 0.4184, "step": 23391 }, { "epoch": 0.9805602422895947, "grad_norm": 1.843152642250061, "learning_rate": 9.875737692138787e-09, "loss": 0.4583, "step": 23392 }, { "epoch": 0.9806021609046037, "grad_norm": 1.9505195617675781, "learning_rate": 9.833137960018058e-09, "loss": 0.5176, "step": 23393 }, { "epoch": 0.9806440795196126, "grad_norm": 1.6275495290756226, "learning_rate": 9.790630214724372e-09, "loss": 0.4286, "step": 23394 }, { "epoch": 0.9806859981346217, "grad_norm": 2.3353686332702637, "learning_rate": 9.748214457042105e-09, "loss": 0.4236, "step": 23395 }, { "epoch": 0.9807279167496306, "grad_norm": 3.2293097972869873, "learning_rate": 9.705890687752295e-09, "loss": 0.4916, "step": 23396 }, { "epoch": 0.9807698353646396, "grad_norm": 2.1408846378326416, "learning_rate": 9.663658907635987e-09, "loss": 0.4975, "step": 23397 }, { "epoch": 0.9808117539796485, "grad_norm": 1.9386470317840576, "learning_rate": 9.62151911747089e-09, "loss": 0.4482, "step": 23398 }, { "epoch": 0.9808536725946575, "grad_norm": 1.8834028244018555, "learning_rate": 9.579471318034161e-09, "loss": 0.4816, "step": 23399 }, { "epoch": 0.9808955912096664, "grad_norm": 1.8868038654327393, "learning_rate": 9.537515510100737e-09, "loss": 0.4465, "step": 23400 }, { "epoch": 0.9809375098246754, "grad_norm": 2.0165257453918457, "learning_rate": 9.495651694444441e-09, "loss": 0.4778, "step": 23401 }, { "epoch": 0.9809794284396843, "grad_norm": 2.2092478275299072, "learning_rate": 9.453879871836324e-09, "loss": 0.4469, "step": 23402 }, { "epoch": 0.9810213470546933, "grad_norm": 1.6239203214645386, "learning_rate": 9.412200043046882e-09, "loss": 0.4673, "step": 23403 }, { "epoch": 0.9810632656697023, "grad_norm": 1.7195829153060913, "learning_rate": 9.370612208844388e-09, "loss": 0.4423, "step": 23404 }, { "epoch": 0.9811051842847113, "grad_norm": 1.5977461338043213, "learning_rate": 9.329116369994895e-09, "loss": 0.4457, "step": 23405 }, { "epoch": 0.9811471028997202, "grad_norm": 2.1585752964019775, "learning_rate": 9.287712527263904e-09, "loss": 0.4131, "step": 23406 }, { "epoch": 0.9811890215147292, "grad_norm": 1.8293474912643433, "learning_rate": 9.246400681414692e-09, "loss": 0.474, "step": 23407 }, { "epoch": 0.9812309401297381, "grad_norm": 1.7834036350250244, "learning_rate": 9.205180833208316e-09, "loss": 0.4554, "step": 23408 }, { "epoch": 0.9812728587447471, "grad_norm": 2.0092456340789795, "learning_rate": 9.164052983404726e-09, "loss": 0.4586, "step": 23409 }, { "epoch": 0.981314777359756, "grad_norm": 1.808905839920044, "learning_rate": 9.123017132762757e-09, "loss": 0.4405, "step": 23410 }, { "epoch": 0.981356695974765, "grad_norm": 2.105217933654785, "learning_rate": 9.082073282037917e-09, "loss": 0.5149, "step": 23411 }, { "epoch": 0.981398614589774, "grad_norm": 1.6743838787078857, "learning_rate": 9.04122143198516e-09, "loss": 0.497, "step": 23412 }, { "epoch": 0.9814405332047829, "grad_norm": 2.085693120956421, "learning_rate": 9.000461583357767e-09, "loss": 0.4998, "step": 23413 }, { "epoch": 0.9814824518197919, "grad_norm": 2.0661370754241943, "learning_rate": 8.959793736907362e-09, "loss": 0.4566, "step": 23414 }, { "epoch": 0.9815243704348008, "grad_norm": 2.1857402324676514, "learning_rate": 8.919217893382792e-09, "loss": 0.4459, "step": 23415 }, { "epoch": 0.9815662890498098, "grad_norm": 1.8056927919387817, "learning_rate": 8.8787340535329e-09, "loss": 0.4729, "step": 23416 }, { "epoch": 0.9816082076648187, "grad_norm": 2.204427480697632, "learning_rate": 8.838342218103201e-09, "loss": 0.4849, "step": 23417 }, { "epoch": 0.9816501262798277, "grad_norm": 2.3775808811187744, "learning_rate": 8.798042387838657e-09, "loss": 0.4689, "step": 23418 }, { "epoch": 0.9816920448948366, "grad_norm": 1.6974077224731445, "learning_rate": 8.757834563482558e-09, "loss": 0.4781, "step": 23419 }, { "epoch": 0.9817339635098457, "grad_norm": 1.6724388599395752, "learning_rate": 8.717718745774872e-09, "loss": 0.4481, "step": 23420 }, { "epoch": 0.9817758821248546, "grad_norm": 1.9697647094726562, "learning_rate": 8.677694935456116e-09, "loss": 0.5303, "step": 23421 }, { "epoch": 0.9818178007398636, "grad_norm": 4.084804534912109, "learning_rate": 8.637763133264033e-09, "loss": 0.4395, "step": 23422 }, { "epoch": 0.9818597193548725, "grad_norm": 2.0946152210235596, "learning_rate": 8.597923339934145e-09, "loss": 0.4238, "step": 23423 }, { "epoch": 0.9819016379698815, "grad_norm": 1.9370921850204468, "learning_rate": 8.558175556200865e-09, "loss": 0.4721, "step": 23424 }, { "epoch": 0.9819435565848904, "grad_norm": 1.7547458410263062, "learning_rate": 8.518519782797496e-09, "loss": 0.4909, "step": 23425 }, { "epoch": 0.9819854751998994, "grad_norm": 1.7808703184127808, "learning_rate": 8.478956020454565e-09, "loss": 0.4385, "step": 23426 }, { "epoch": 0.9820273938149083, "grad_norm": 1.9540425539016724, "learning_rate": 8.439484269901487e-09, "loss": 0.4413, "step": 23427 }, { "epoch": 0.9820693124299174, "grad_norm": 2.5605502128601074, "learning_rate": 8.400104531866016e-09, "loss": 0.5221, "step": 23428 }, { "epoch": 0.9821112310449263, "grad_norm": 1.6750702857971191, "learning_rate": 8.360816807074235e-09, "loss": 0.4084, "step": 23429 }, { "epoch": 0.9821531496599353, "grad_norm": 2.5703353881835938, "learning_rate": 8.321621096249456e-09, "loss": 0.4782, "step": 23430 }, { "epoch": 0.9821950682749442, "grad_norm": 1.7270808219909668, "learning_rate": 8.282517400115541e-09, "loss": 0.4497, "step": 23431 }, { "epoch": 0.9822369868899532, "grad_norm": 1.7824124097824097, "learning_rate": 8.24350571939192e-09, "loss": 0.4847, "step": 23432 }, { "epoch": 0.9822789055049621, "grad_norm": 2.0409040451049805, "learning_rate": 8.204586054798568e-09, "loss": 0.483, "step": 23433 }, { "epoch": 0.9823208241199711, "grad_norm": 2.065157175064087, "learning_rate": 8.165758407052693e-09, "loss": 0.4774, "step": 23434 }, { "epoch": 0.98236274273498, "grad_norm": 1.7787307500839233, "learning_rate": 8.127022776869831e-09, "loss": 0.5063, "step": 23435 }, { "epoch": 0.982404661349989, "grad_norm": 1.830104947090149, "learning_rate": 8.08837916496441e-09, "loss": 0.4618, "step": 23436 }, { "epoch": 0.982446579964998, "grad_norm": 1.9781986474990845, "learning_rate": 8.049827572049195e-09, "loss": 0.5078, "step": 23437 }, { "epoch": 0.9824884985800069, "grad_norm": 2.367867946624756, "learning_rate": 8.011367998833619e-09, "loss": 0.4544, "step": 23438 }, { "epoch": 0.9825304171950159, "grad_norm": 1.8003214597702026, "learning_rate": 7.973000446027113e-09, "loss": 0.4968, "step": 23439 }, { "epoch": 0.9825723358100248, "grad_norm": 1.8381518125534058, "learning_rate": 7.934724914337443e-09, "loss": 0.5149, "step": 23440 }, { "epoch": 0.9826142544250338, "grad_norm": 1.9368600845336914, "learning_rate": 7.896541404470159e-09, "loss": 0.4306, "step": 23441 }, { "epoch": 0.9826561730400427, "grad_norm": 2.149491786956787, "learning_rate": 7.85844991712803e-09, "loss": 0.4885, "step": 23442 }, { "epoch": 0.9826980916550517, "grad_norm": 1.9976379871368408, "learning_rate": 7.820450453014384e-09, "loss": 0.5046, "step": 23443 }, { "epoch": 0.9827400102700606, "grad_norm": 2.046588897705078, "learning_rate": 7.782543012828659e-09, "loss": 0.4496, "step": 23444 }, { "epoch": 0.9827819288850697, "grad_norm": 1.9031951427459717, "learning_rate": 7.744727597270852e-09, "loss": 0.4206, "step": 23445 }, { "epoch": 0.9828238475000786, "grad_norm": 2.1084601879119873, "learning_rate": 7.707004207037072e-09, "loss": 0.5236, "step": 23446 }, { "epoch": 0.9828657661150876, "grad_norm": 2.599782943725586, "learning_rate": 7.66937284282343e-09, "loss": 0.4677, "step": 23447 }, { "epoch": 0.9829076847300965, "grad_norm": 2.404745578765869, "learning_rate": 7.631833505322705e-09, "loss": 0.489, "step": 23448 }, { "epoch": 0.9829496033451055, "grad_norm": 1.9654688835144043, "learning_rate": 7.594386195227676e-09, "loss": 0.4839, "step": 23449 }, { "epoch": 0.9829915219601144, "grad_norm": 1.965448260307312, "learning_rate": 7.557030913228348e-09, "loss": 0.4808, "step": 23450 }, { "epoch": 0.9830334405751234, "grad_norm": 1.9923946857452393, "learning_rate": 7.51976766001361e-09, "loss": 0.4394, "step": 23451 }, { "epoch": 0.9830753591901323, "grad_norm": 2.4449684619903564, "learning_rate": 7.482596436269585e-09, "loss": 0.45, "step": 23452 }, { "epoch": 0.9831172778051414, "grad_norm": 2.086376905441284, "learning_rate": 7.445517242681832e-09, "loss": 0.4502, "step": 23453 }, { "epoch": 0.9831591964201503, "grad_norm": 2.0205612182617188, "learning_rate": 7.4085300799342505e-09, "loss": 0.5204, "step": 23454 }, { "epoch": 0.9832011150351593, "grad_norm": 1.8283377885818481, "learning_rate": 7.371634948708517e-09, "loss": 0.4718, "step": 23455 }, { "epoch": 0.9832430336501682, "grad_norm": 2.1055080890655518, "learning_rate": 7.3348318496840875e-09, "loss": 0.4889, "step": 23456 }, { "epoch": 0.9832849522651772, "grad_norm": 1.7315266132354736, "learning_rate": 7.298120783540419e-09, "loss": 0.4472, "step": 23457 }, { "epoch": 0.9833268708801861, "grad_norm": 1.9701884984970093, "learning_rate": 7.261501750953637e-09, "loss": 0.4888, "step": 23458 }, { "epoch": 0.9833687894951951, "grad_norm": 1.8377903699874878, "learning_rate": 7.224974752598202e-09, "loss": 0.4833, "step": 23459 }, { "epoch": 0.983410708110204, "grad_norm": 1.911102294921875, "learning_rate": 7.1885397891485745e-09, "loss": 0.4989, "step": 23460 }, { "epoch": 0.983452626725213, "grad_norm": 1.616533875465393, "learning_rate": 7.152196861275884e-09, "loss": 0.4805, "step": 23461 }, { "epoch": 0.983494545340222, "grad_norm": 2.08463191986084, "learning_rate": 7.115945969650151e-09, "loss": 0.4982, "step": 23462 }, { "epoch": 0.9835364639552309, "grad_norm": 2.1900105476379395, "learning_rate": 7.079787114939174e-09, "loss": 0.4796, "step": 23463 }, { "epoch": 0.9835783825702399, "grad_norm": 2.9957308769226074, "learning_rate": 7.043720297809642e-09, "loss": 0.4539, "step": 23464 }, { "epoch": 0.9836203011852488, "grad_norm": 2.3124990463256836, "learning_rate": 7.007745518927134e-09, "loss": 0.4972, "step": 23465 }, { "epoch": 0.9836622198002578, "grad_norm": 3.2902071475982666, "learning_rate": 6.971862778953898e-09, "loss": 0.4454, "step": 23466 }, { "epoch": 0.9837041384152667, "grad_norm": 2.1710705757141113, "learning_rate": 6.936072078551626e-09, "loss": 0.5185, "step": 23467 }, { "epoch": 0.9837460570302757, "grad_norm": 1.7238327264785767, "learning_rate": 6.900373418380346e-09, "loss": 0.4348, "step": 23468 }, { "epoch": 0.9837879756452846, "grad_norm": 1.9237149953842163, "learning_rate": 6.864766799097866e-09, "loss": 0.4866, "step": 23469 }, { "epoch": 0.9838298942602937, "grad_norm": 3.433534860610962, "learning_rate": 6.829252221360883e-09, "loss": 0.4104, "step": 23470 }, { "epoch": 0.9838718128753026, "grad_norm": 2.0496864318847656, "learning_rate": 6.793829685823317e-09, "loss": 0.5106, "step": 23471 }, { "epoch": 0.9839137314903116, "grad_norm": 1.8090786933898926, "learning_rate": 6.75849919313909e-09, "loss": 0.4169, "step": 23472 }, { "epoch": 0.9839556501053205, "grad_norm": 1.923202633857727, "learning_rate": 6.723260743958793e-09, "loss": 0.4799, "step": 23473 }, { "epoch": 0.9839975687203295, "grad_norm": 1.7691091299057007, "learning_rate": 6.688114338931906e-09, "loss": 0.4336, "step": 23474 }, { "epoch": 0.9840394873353384, "grad_norm": 1.8372737169265747, "learning_rate": 6.6530599787067996e-09, "loss": 0.4371, "step": 23475 }, { "epoch": 0.9840814059503474, "grad_norm": 2.0102429389953613, "learning_rate": 6.618097663929624e-09, "loss": 0.4907, "step": 23476 }, { "epoch": 0.9841233245653563, "grad_norm": 1.7203744649887085, "learning_rate": 6.583227395244307e-09, "loss": 0.4841, "step": 23477 }, { "epoch": 0.9841652431803654, "grad_norm": 1.856429100036621, "learning_rate": 6.548449173294224e-09, "loss": 0.5156, "step": 23478 }, { "epoch": 0.9842071617953743, "grad_norm": 1.8929365873336792, "learning_rate": 6.5137629987199746e-09, "loss": 0.4405, "step": 23479 }, { "epoch": 0.9842490804103833, "grad_norm": 2.054762601852417, "learning_rate": 6.479168872161601e-09, "loss": 0.4708, "step": 23480 }, { "epoch": 0.9842909990253922, "grad_norm": 1.981136679649353, "learning_rate": 6.444666794256371e-09, "loss": 0.5288, "step": 23481 }, { "epoch": 0.9843329176404012, "grad_norm": 2.091308832168579, "learning_rate": 6.4102567656398884e-09, "loss": 0.5207, "step": 23482 }, { "epoch": 0.9843748362554101, "grad_norm": 1.9473440647125244, "learning_rate": 6.3759387869472e-09, "loss": 0.5171, "step": 23483 }, { "epoch": 0.9844167548704191, "grad_norm": 1.760913372039795, "learning_rate": 6.341712858810578e-09, "loss": 0.5108, "step": 23484 }, { "epoch": 0.984458673485428, "grad_norm": 2.2287826538085938, "learning_rate": 6.307578981861184e-09, "loss": 0.4212, "step": 23485 }, { "epoch": 0.984500592100437, "grad_norm": 1.993794560432434, "learning_rate": 6.273537156727405e-09, "loss": 0.4463, "step": 23486 }, { "epoch": 0.984542510715446, "grad_norm": 2.3106720447540283, "learning_rate": 6.239587384038182e-09, "loss": 0.4792, "step": 23487 }, { "epoch": 0.9845844293304549, "grad_norm": 2.423276424407959, "learning_rate": 6.205729664418015e-09, "loss": 0.4839, "step": 23488 }, { "epoch": 0.9846263479454639, "grad_norm": 1.8100227117538452, "learning_rate": 6.171963998491404e-09, "loss": 0.494, "step": 23489 }, { "epoch": 0.9846682665604728, "grad_norm": 2.042412281036377, "learning_rate": 6.138290386881185e-09, "loss": 0.3906, "step": 23490 }, { "epoch": 0.9847101851754818, "grad_norm": 1.9157700538635254, "learning_rate": 6.104708830207417e-09, "loss": 0.5043, "step": 23491 }, { "epoch": 0.9847521037904907, "grad_norm": 1.570339322090149, "learning_rate": 6.071219329089606e-09, "loss": 0.4255, "step": 23492 }, { "epoch": 0.9847940224054997, "grad_norm": 2.1106488704681396, "learning_rate": 6.037821884145034e-09, "loss": 0.4773, "step": 23493 }, { "epoch": 0.9848359410205086, "grad_norm": 1.744557499885559, "learning_rate": 6.004516495989321e-09, "loss": 0.4549, "step": 23494 }, { "epoch": 0.9848778596355177, "grad_norm": 2.378934144973755, "learning_rate": 5.9713031652364196e-09, "loss": 0.463, "step": 23495 }, { "epoch": 0.9849197782505266, "grad_norm": 2.253786325454712, "learning_rate": 5.9381818924980625e-09, "loss": 0.4847, "step": 23496 }, { "epoch": 0.9849616968655356, "grad_norm": 1.998996615409851, "learning_rate": 5.905152678385984e-09, "loss": 0.5148, "step": 23497 }, { "epoch": 0.9850036154805445, "grad_norm": 1.7713366746902466, "learning_rate": 5.872215523507474e-09, "loss": 0.4452, "step": 23498 }, { "epoch": 0.9850455340955535, "grad_norm": 2.1360859870910645, "learning_rate": 5.839370428470936e-09, "loss": 0.4721, "step": 23499 }, { "epoch": 0.9850874527105624, "grad_norm": 1.882761001586914, "learning_rate": 5.806617393881442e-09, "loss": 0.5382, "step": 23500 }, { "epoch": 0.9851293713255714, "grad_norm": 2.2183616161346436, "learning_rate": 5.773956420342397e-09, "loss": 0.4888, "step": 23501 }, { "epoch": 0.9851712899405803, "grad_norm": 4.208333969116211, "learning_rate": 5.741387508456653e-09, "loss": 0.4436, "step": 23502 }, { "epoch": 0.9852132085555894, "grad_norm": 2.6890196800231934, "learning_rate": 5.70891065882373e-09, "loss": 0.5148, "step": 23503 }, { "epoch": 0.9852551271705983, "grad_norm": 2.1103458404541016, "learning_rate": 5.676525872042593e-09, "loss": 0.493, "step": 23504 }, { "epoch": 0.9852970457856073, "grad_norm": 2.4465086460113525, "learning_rate": 5.644233148710543e-09, "loss": 0.5246, "step": 23505 }, { "epoch": 0.9853389644006162, "grad_norm": 2.0065886974334717, "learning_rate": 5.612032489422103e-09, "loss": 0.5017, "step": 23506 }, { "epoch": 0.9853808830156252, "grad_norm": 1.7072607278823853, "learning_rate": 5.579923894771799e-09, "loss": 0.4552, "step": 23507 }, { "epoch": 0.9854228016306341, "grad_norm": 2.9399757385253906, "learning_rate": 5.547907365350824e-09, "loss": 0.5082, "step": 23508 }, { "epoch": 0.9854647202456431, "grad_norm": 1.9568047523498535, "learning_rate": 5.515982901749262e-09, "loss": 0.4682, "step": 23509 }, { "epoch": 0.985506638860652, "grad_norm": 2.1043150424957275, "learning_rate": 5.484150504556085e-09, "loss": 0.5064, "step": 23510 }, { "epoch": 0.9855485574756611, "grad_norm": 1.8104978799819946, "learning_rate": 5.452410174358047e-09, "loss": 0.445, "step": 23511 }, { "epoch": 0.98559047609067, "grad_norm": 2.1167869567871094, "learning_rate": 5.420761911740235e-09, "loss": 0.4972, "step": 23512 }, { "epoch": 0.9856323947056789, "grad_norm": 1.9024027585983276, "learning_rate": 5.389205717285517e-09, "loss": 0.4795, "step": 23513 }, { "epoch": 0.9856743133206879, "grad_norm": 2.3241653442382812, "learning_rate": 5.357741591576205e-09, "loss": 0.5234, "step": 23514 }, { "epoch": 0.9857162319356968, "grad_norm": 1.642275333404541, "learning_rate": 5.326369535191833e-09, "loss": 0.4163, "step": 23515 }, { "epoch": 0.9857581505507058, "grad_norm": 2.008094072341919, "learning_rate": 5.29508954871083e-09, "loss": 0.4746, "step": 23516 }, { "epoch": 0.9858000691657147, "grad_norm": 2.4623706340789795, "learning_rate": 5.263901632709956e-09, "loss": 0.4702, "step": 23517 }, { "epoch": 0.9858419877807237, "grad_norm": 2.199263095855713, "learning_rate": 5.232805787764306e-09, "loss": 0.4698, "step": 23518 }, { "epoch": 0.9858839063957326, "grad_norm": 1.8044427633285522, "learning_rate": 5.2018020144462e-09, "loss": 0.4319, "step": 23519 }, { "epoch": 0.9859258250107417, "grad_norm": 1.74221932888031, "learning_rate": 5.170890313328514e-09, "loss": 0.5037, "step": 23520 }, { "epoch": 0.9859677436257506, "grad_norm": 1.7100872993469238, "learning_rate": 5.140070684979681e-09, "loss": 0.462, "step": 23521 }, { "epoch": 0.9860096622407596, "grad_norm": 1.9644231796264648, "learning_rate": 5.109343129968691e-09, "loss": 0.4585, "step": 23522 }, { "epoch": 0.9860515808557685, "grad_norm": 6.753847599029541, "learning_rate": 5.078707648862313e-09, "loss": 0.4967, "step": 23523 }, { "epoch": 0.9860934994707775, "grad_norm": 2.3707656860351562, "learning_rate": 5.0481642422239854e-09, "loss": 0.4882, "step": 23524 }, { "epoch": 0.9861354180857864, "grad_norm": 3.1737060546875, "learning_rate": 5.0177129106177e-09, "loss": 0.4926, "step": 23525 }, { "epoch": 0.9861773367007954, "grad_norm": 1.7183079719543457, "learning_rate": 4.987353654604121e-09, "loss": 0.4532, "step": 23526 }, { "epoch": 0.9862192553158043, "grad_norm": 1.8822276592254639, "learning_rate": 4.957086474743911e-09, "loss": 0.438, "step": 23527 }, { "epoch": 0.9862611739308134, "grad_norm": 2.2736470699310303, "learning_rate": 4.926911371593845e-09, "loss": 0.4738, "step": 23528 }, { "epoch": 0.9863030925458223, "grad_norm": 2.205418586730957, "learning_rate": 4.896828345711258e-09, "loss": 0.4575, "step": 23529 }, { "epoch": 0.9863450111608313, "grad_norm": 1.8388020992279053, "learning_rate": 4.866837397649593e-09, "loss": 0.5099, "step": 23530 }, { "epoch": 0.9863869297758402, "grad_norm": 1.8781359195709229, "learning_rate": 4.836938527962298e-09, "loss": 0.4594, "step": 23531 }, { "epoch": 0.9864288483908492, "grad_norm": 1.7903004884719849, "learning_rate": 4.807131737200599e-09, "loss": 0.4493, "step": 23532 }, { "epoch": 0.9864707670058581, "grad_norm": 1.8866221904754639, "learning_rate": 4.7774170259135e-09, "loss": 0.4657, "step": 23533 }, { "epoch": 0.9865126856208671, "grad_norm": 2.2519919872283936, "learning_rate": 4.747794394649452e-09, "loss": 0.4619, "step": 23534 }, { "epoch": 0.986554604235876, "grad_norm": 1.894882082939148, "learning_rate": 4.71826384395413e-09, "loss": 0.4887, "step": 23535 }, { "epoch": 0.9865965228508851, "grad_norm": 2.0655441284179688, "learning_rate": 4.688825374371542e-09, "loss": 0.459, "step": 23536 }, { "epoch": 0.986638441465894, "grad_norm": 2.035959005355835, "learning_rate": 4.659478986444588e-09, "loss": 0.463, "step": 23537 }, { "epoch": 0.9866803600809029, "grad_norm": 1.942799687385559, "learning_rate": 4.630224680714501e-09, "loss": 0.4414, "step": 23538 }, { "epoch": 0.9867222786959119, "grad_norm": 1.784487247467041, "learning_rate": 4.60106245771974e-09, "loss": 0.4612, "step": 23539 }, { "epoch": 0.9867641973109208, "grad_norm": 1.7556582689285278, "learning_rate": 4.571992317999319e-09, "loss": 0.4667, "step": 23540 }, { "epoch": 0.9868061159259298, "grad_norm": 1.9048123359680176, "learning_rate": 4.543014262087808e-09, "loss": 0.4523, "step": 23541 }, { "epoch": 0.9868480345409387, "grad_norm": 2.172004222869873, "learning_rate": 4.514128290519782e-09, "loss": 0.4789, "step": 23542 }, { "epoch": 0.9868899531559477, "grad_norm": 1.8769395351409912, "learning_rate": 4.4853344038281456e-09, "loss": 0.4554, "step": 23543 }, { "epoch": 0.9869318717709566, "grad_norm": 2.3073058128356934, "learning_rate": 4.456632602543032e-09, "loss": 0.5098, "step": 23544 }, { "epoch": 0.9869737903859657, "grad_norm": 1.8506851196289062, "learning_rate": 4.428022887193461e-09, "loss": 0.4756, "step": 23545 }, { "epoch": 0.9870157090009746, "grad_norm": 2.2033743858337402, "learning_rate": 4.399505258307346e-09, "loss": 0.4241, "step": 23546 }, { "epoch": 0.9870576276159836, "grad_norm": 1.9652167558670044, "learning_rate": 4.3710797164103755e-09, "loss": 0.5091, "step": 23547 }, { "epoch": 0.9870995462309925, "grad_norm": 1.820227861404419, "learning_rate": 4.3427462620260205e-09, "loss": 0.4761, "step": 23548 }, { "epoch": 0.9871414648460015, "grad_norm": 1.9295499324798584, "learning_rate": 4.314504895676641e-09, "loss": 0.5488, "step": 23549 }, { "epoch": 0.9871833834610104, "grad_norm": 2.0693891048431396, "learning_rate": 4.286355617883486e-09, "loss": 0.5187, "step": 23550 }, { "epoch": 0.9872253020760194, "grad_norm": 1.9213789701461792, "learning_rate": 4.258298429165031e-09, "loss": 0.4388, "step": 23551 }, { "epoch": 0.9872672206910283, "grad_norm": 2.0801615715026855, "learning_rate": 4.230333330038083e-09, "loss": 0.4659, "step": 23552 }, { "epoch": 0.9873091393060374, "grad_norm": 1.919336199760437, "learning_rate": 4.202460321018897e-09, "loss": 0.4682, "step": 23553 }, { "epoch": 0.9873510579210463, "grad_norm": 1.729125738143921, "learning_rate": 4.174679402620396e-09, "loss": 0.4124, "step": 23554 }, { "epoch": 0.9873929765360553, "grad_norm": 1.8733490705490112, "learning_rate": 4.146990575355503e-09, "loss": 0.4948, "step": 23555 }, { "epoch": 0.9874348951510642, "grad_norm": 1.7527782917022705, "learning_rate": 4.1193938397338095e-09, "loss": 0.436, "step": 23556 }, { "epoch": 0.9874768137660732, "grad_norm": 2.1299548149108887, "learning_rate": 4.091889196264909e-09, "loss": 0.4713, "step": 23557 }, { "epoch": 0.9875187323810821, "grad_norm": 1.9276998043060303, "learning_rate": 4.0644766454556175e-09, "loss": 0.4819, "step": 23558 }, { "epoch": 0.9875606509960911, "grad_norm": 2.516388416290283, "learning_rate": 4.037156187810531e-09, "loss": 0.4649, "step": 23559 }, { "epoch": 0.9876025696111, "grad_norm": 2.122148036956787, "learning_rate": 4.0099278238342475e-09, "loss": 0.442, "step": 23560 }, { "epoch": 0.9876444882261091, "grad_norm": 1.7109276056289673, "learning_rate": 3.982791554028032e-09, "loss": 0.4902, "step": 23561 }, { "epoch": 0.987686406841118, "grad_norm": 1.8071483373641968, "learning_rate": 3.955747378892039e-09, "loss": 0.4597, "step": 23562 }, { "epoch": 0.9877283254561269, "grad_norm": 2.0339674949645996, "learning_rate": 3.928795298924759e-09, "loss": 0.4093, "step": 23563 }, { "epoch": 0.9877702440711359, "grad_norm": 1.8471814393997192, "learning_rate": 3.901935314624128e-09, "loss": 0.4354, "step": 23564 }, { "epoch": 0.9878121626861448, "grad_norm": 1.9646131992340088, "learning_rate": 3.875167426483639e-09, "loss": 0.498, "step": 23565 }, { "epoch": 0.9878540813011538, "grad_norm": 1.6791887283325195, "learning_rate": 3.848491634997897e-09, "loss": 0.4424, "step": 23566 }, { "epoch": 0.9878959999161627, "grad_norm": 2.1840529441833496, "learning_rate": 3.821907940658176e-09, "loss": 0.4723, "step": 23567 }, { "epoch": 0.9879379185311717, "grad_norm": 2.2657275199890137, "learning_rate": 3.795416343954639e-09, "loss": 0.4936, "step": 23568 }, { "epoch": 0.9879798371461807, "grad_norm": 1.969802737236023, "learning_rate": 3.769016845375228e-09, "loss": 0.4395, "step": 23569 }, { "epoch": 0.9880217557611897, "grad_norm": 2.118743896484375, "learning_rate": 3.742709445407333e-09, "loss": 0.4507, "step": 23570 }, { "epoch": 0.9880636743761986, "grad_norm": 3.5435311794281006, "learning_rate": 3.71649414453501e-09, "loss": 0.4576, "step": 23571 }, { "epoch": 0.9881055929912076, "grad_norm": 1.5749175548553467, "learning_rate": 3.6903709432423164e-09, "loss": 0.4697, "step": 23572 }, { "epoch": 0.9881475116062165, "grad_norm": 1.8493481874465942, "learning_rate": 3.6643398420105335e-09, "loss": 0.4926, "step": 23573 }, { "epoch": 0.9881894302212255, "grad_norm": 2.301983594894409, "learning_rate": 3.6384008413192784e-09, "loss": 0.4976, "step": 23574 }, { "epoch": 0.9882313488362344, "grad_norm": 1.840202808380127, "learning_rate": 3.6125539416465014e-09, "loss": 0.4703, "step": 23575 }, { "epoch": 0.9882732674512434, "grad_norm": 2.9614076614379883, "learning_rate": 3.586799143469599e-09, "loss": 0.4567, "step": 23576 }, { "epoch": 0.9883151860662523, "grad_norm": 1.845023512840271, "learning_rate": 3.5611364472626365e-09, "loss": 0.4888, "step": 23577 }, { "epoch": 0.9883571046812614, "grad_norm": 1.9855276346206665, "learning_rate": 3.535565853498568e-09, "loss": 0.5581, "step": 23578 }, { "epoch": 0.9883990232962703, "grad_norm": 1.7503771781921387, "learning_rate": 3.510087362648684e-09, "loss": 0.5402, "step": 23579 }, { "epoch": 0.9884409419112793, "grad_norm": 1.7151819467544556, "learning_rate": 3.484700975183164e-09, "loss": 0.3953, "step": 23580 }, { "epoch": 0.9884828605262882, "grad_norm": 1.83315110206604, "learning_rate": 3.4594066915694115e-09, "loss": 0.4688, "step": 23581 }, { "epoch": 0.9885247791412972, "grad_norm": 1.71964430809021, "learning_rate": 3.4342045122742752e-09, "loss": 0.5249, "step": 23582 }, { "epoch": 0.9885666977563061, "grad_norm": 1.7889938354492188, "learning_rate": 3.4090944377623836e-09, "loss": 0.428, "step": 23583 }, { "epoch": 0.9886086163713151, "grad_norm": 2.025127649307251, "learning_rate": 3.3840764684955896e-09, "loss": 0.48, "step": 23584 }, { "epoch": 0.988650534986324, "grad_norm": 1.9207803010940552, "learning_rate": 3.359150604936301e-09, "loss": 0.4372, "step": 23585 }, { "epoch": 0.9886924536013331, "grad_norm": 1.8501981496810913, "learning_rate": 3.33431684754304e-09, "loss": 0.5141, "step": 23586 }, { "epoch": 0.988734372216342, "grad_norm": 3.4328935146331787, "learning_rate": 3.3095751967737733e-09, "loss": 0.4941, "step": 23587 }, { "epoch": 0.9887762908313509, "grad_norm": 1.6211726665496826, "learning_rate": 3.284925653085358e-09, "loss": 0.4636, "step": 23588 }, { "epoch": 0.9888182094463599, "grad_norm": 2.1461076736450195, "learning_rate": 3.26036821693132e-09, "loss": 0.478, "step": 23589 }, { "epoch": 0.9888601280613688, "grad_norm": 1.7204234600067139, "learning_rate": 3.2359028887646304e-09, "loss": 0.435, "step": 23590 }, { "epoch": 0.9889020466763778, "grad_norm": 2.2823030948638916, "learning_rate": 3.2115296690360396e-09, "loss": 0.4905, "step": 23591 }, { "epoch": 0.9889439652913867, "grad_norm": 1.7383216619491577, "learning_rate": 3.1872485581951886e-09, "loss": 0.4467, "step": 23592 }, { "epoch": 0.9889858839063957, "grad_norm": 2.1993887424468994, "learning_rate": 3.1630595566894964e-09, "loss": 0.4949, "step": 23593 }, { "epoch": 0.9890278025214047, "grad_norm": 1.9084845781326294, "learning_rate": 3.1389626649652728e-09, "loss": 0.5154, "step": 23594 }, { "epoch": 0.9890697211364137, "grad_norm": 1.9326128959655762, "learning_rate": 3.114957883465497e-09, "loss": 0.4619, "step": 23595 }, { "epoch": 0.9891116397514226, "grad_norm": 2.0556046962738037, "learning_rate": 3.0910452126342584e-09, "loss": 0.497, "step": 23596 }, { "epoch": 0.9891535583664316, "grad_norm": 2.110933303833008, "learning_rate": 3.067224652911205e-09, "loss": 0.4521, "step": 23597 }, { "epoch": 0.9891954769814405, "grad_norm": 1.6930644512176514, "learning_rate": 3.043496204735985e-09, "loss": 0.4239, "step": 23598 }, { "epoch": 0.9892373955964495, "grad_norm": 1.85369074344635, "learning_rate": 3.0198598685460266e-09, "loss": 0.5087, "step": 23599 }, { "epoch": 0.9892793142114584, "grad_norm": 2.01542592048645, "learning_rate": 2.9963156447765374e-09, "loss": 0.4665, "step": 23600 }, { "epoch": 0.9893212328264674, "grad_norm": 2.0952396392822266, "learning_rate": 2.972863533861614e-09, "loss": 0.4888, "step": 23601 }, { "epoch": 0.9893631514414764, "grad_norm": 1.6486780643463135, "learning_rate": 2.949503536233689e-09, "loss": 0.4885, "step": 23602 }, { "epoch": 0.9894050700564854, "grad_norm": 1.9663444757461548, "learning_rate": 2.9262356523240833e-09, "loss": 0.4652, "step": 23603 }, { "epoch": 0.9894469886714943, "grad_norm": 1.8386116027832031, "learning_rate": 2.9030598825607882e-09, "loss": 0.5078, "step": 23604 }, { "epoch": 0.9894889072865033, "grad_norm": 2.0111169815063477, "learning_rate": 2.8799762273712396e-09, "loss": 0.5066, "step": 23605 }, { "epoch": 0.9895308259015122, "grad_norm": 2.3270957469940186, "learning_rate": 2.8569846871806527e-09, "loss": 0.479, "step": 23606 }, { "epoch": 0.9895727445165212, "grad_norm": 2.0722320079803467, "learning_rate": 2.8340852624136883e-09, "loss": 0.4424, "step": 23607 }, { "epoch": 0.9896146631315301, "grad_norm": 4.788729190826416, "learning_rate": 2.811277953492231e-09, "loss": 0.4608, "step": 23608 }, { "epoch": 0.9896565817465391, "grad_norm": 1.8623913526535034, "learning_rate": 2.7885627608359446e-09, "loss": 0.4747, "step": 23609 }, { "epoch": 0.989698500361548, "grad_norm": 1.9145594835281372, "learning_rate": 2.7659396848644936e-09, "loss": 0.4774, "step": 23610 }, { "epoch": 0.9897404189765571, "grad_norm": 2.0402021408081055, "learning_rate": 2.743408725994212e-09, "loss": 0.4459, "step": 23611 }, { "epoch": 0.989782337591566, "grad_norm": 2.6206839084625244, "learning_rate": 2.720969884640323e-09, "loss": 0.4736, "step": 23612 }, { "epoch": 0.9898242562065749, "grad_norm": 1.6976722478866577, "learning_rate": 2.6986231612174952e-09, "loss": 0.5139, "step": 23613 }, { "epoch": 0.9898661748215839, "grad_norm": 2.4245071411132812, "learning_rate": 2.6763685561365104e-09, "loss": 0.4816, "step": 23614 }, { "epoch": 0.9899080934365928, "grad_norm": 2.001251697540283, "learning_rate": 2.6542060698081517e-09, "loss": 0.5332, "step": 23615 }, { "epoch": 0.9899500120516018, "grad_norm": 2.679333448410034, "learning_rate": 2.632135702640981e-09, "loss": 0.4437, "step": 23616 }, { "epoch": 0.9899919306666107, "grad_norm": 2.1656720638275146, "learning_rate": 2.610157455041895e-09, "loss": 0.5197, "step": 23617 }, { "epoch": 0.9900338492816197, "grad_norm": 2.0646018981933594, "learning_rate": 2.5882713274155703e-09, "loss": 0.4646, "step": 23618 }, { "epoch": 0.9900757678966287, "grad_norm": 2.0399930477142334, "learning_rate": 2.5664773201661273e-09, "loss": 0.4813, "step": 23619 }, { "epoch": 0.9901176865116377, "grad_norm": 1.8084537982940674, "learning_rate": 2.5447754336949125e-09, "loss": 0.4713, "step": 23620 }, { "epoch": 0.9901596051266466, "grad_norm": 2.0068366527557373, "learning_rate": 2.523165668401606e-09, "loss": 0.4722, "step": 23621 }, { "epoch": 0.9902015237416556, "grad_norm": 2.6981022357940674, "learning_rate": 2.5016480246847773e-09, "loss": 0.4958, "step": 23622 }, { "epoch": 0.9902434423566645, "grad_norm": 2.618119955062866, "learning_rate": 2.4802225029418868e-09, "loss": 0.4427, "step": 23623 }, { "epoch": 0.9902853609716735, "grad_norm": 2.0525009632110596, "learning_rate": 2.458889103567064e-09, "loss": 0.4984, "step": 23624 }, { "epoch": 0.9903272795866824, "grad_norm": 10.718642234802246, "learning_rate": 2.4376478269533265e-09, "loss": 0.4254, "step": 23625 }, { "epoch": 0.9903691982016914, "grad_norm": 1.6784908771514893, "learning_rate": 2.4164986734931396e-09, "loss": 0.4429, "step": 23626 }, { "epoch": 0.9904111168167004, "grad_norm": 1.9170392751693726, "learning_rate": 2.3954416435756355e-09, "loss": 0.5189, "step": 23627 }, { "epoch": 0.9904530354317094, "grad_norm": 1.9381815195083618, "learning_rate": 2.374476737588838e-09, "loss": 0.4773, "step": 23628 }, { "epoch": 0.9904949540467183, "grad_norm": 1.8301277160644531, "learning_rate": 2.3536039559196587e-09, "loss": 0.462, "step": 23629 }, { "epoch": 0.9905368726617273, "grad_norm": 1.928971529006958, "learning_rate": 2.332823298953346e-09, "loss": 0.4896, "step": 23630 }, { "epoch": 0.9905787912767362, "grad_norm": 1.6844916343688965, "learning_rate": 2.3121347670718165e-09, "loss": 0.4382, "step": 23631 }, { "epoch": 0.9906207098917452, "grad_norm": 1.928808569908142, "learning_rate": 2.2915383606569864e-09, "loss": 0.4425, "step": 23632 }, { "epoch": 0.9906626285067541, "grad_norm": 2.359053134918213, "learning_rate": 2.2710340800885523e-09, "loss": 0.4495, "step": 23633 }, { "epoch": 0.9907045471217631, "grad_norm": 1.8267217874526978, "learning_rate": 2.250621925744545e-09, "loss": 0.4753, "step": 23634 }, { "epoch": 0.990746465736772, "grad_norm": 1.7032254934310913, "learning_rate": 2.23030189800133e-09, "loss": 0.5137, "step": 23635 }, { "epoch": 0.9907883843517811, "grad_norm": 2.2435667514801025, "learning_rate": 2.210073997233053e-09, "loss": 0.5096, "step": 23636 }, { "epoch": 0.99083030296679, "grad_norm": 1.793688178062439, "learning_rate": 2.1899382238127487e-09, "loss": 0.459, "step": 23637 }, { "epoch": 0.9908722215817989, "grad_norm": 2.3232035636901855, "learning_rate": 2.1698945781123416e-09, "loss": 0.4589, "step": 23638 }, { "epoch": 0.9909141401968079, "grad_norm": 2.0152647495269775, "learning_rate": 2.1499430604998707e-09, "loss": 0.4099, "step": 23639 }, { "epoch": 0.9909560588118168, "grad_norm": 1.996543049812317, "learning_rate": 2.13008367134393e-09, "loss": 0.4678, "step": 23640 }, { "epoch": 0.9909979774268258, "grad_norm": 1.81594979763031, "learning_rate": 2.110316411010893e-09, "loss": 0.4984, "step": 23641 }, { "epoch": 0.9910398960418347, "grad_norm": 1.8483705520629883, "learning_rate": 2.090641279864358e-09, "loss": 0.4694, "step": 23642 }, { "epoch": 0.9910818146568438, "grad_norm": 1.7909762859344482, "learning_rate": 2.0710582782679234e-09, "loss": 0.4716, "step": 23643 }, { "epoch": 0.9911237332718527, "grad_norm": 2.1911544799804688, "learning_rate": 2.0515674065818556e-09, "loss": 0.4727, "step": 23644 }, { "epoch": 0.9911656518868617, "grad_norm": 2.14510440826416, "learning_rate": 2.0321686651658677e-09, "loss": 0.5242, "step": 23645 }, { "epoch": 0.9912075705018706, "grad_norm": 2.1997122764587402, "learning_rate": 2.0128620543774512e-09, "loss": 0.453, "step": 23646 }, { "epoch": 0.9912494891168796, "grad_norm": 1.9546300172805786, "learning_rate": 1.9936475745718775e-09, "loss": 0.4766, "step": 23647 }, { "epoch": 0.9912914077318885, "grad_norm": 2.7090649604797363, "learning_rate": 1.974525226104418e-09, "loss": 0.4112, "step": 23648 }, { "epoch": 0.9913333263468975, "grad_norm": 2.1523430347442627, "learning_rate": 1.955495009326458e-09, "loss": 0.5153, "step": 23649 }, { "epoch": 0.9913752449619064, "grad_norm": 2.3983688354492188, "learning_rate": 1.9365569245899385e-09, "loss": 0.5098, "step": 23650 }, { "epoch": 0.9914171635769155, "grad_norm": 2.317270278930664, "learning_rate": 1.9177109722429143e-09, "loss": 0.4392, "step": 23651 }, { "epoch": 0.9914590821919244, "grad_norm": 3.9399383068084717, "learning_rate": 1.8989571526334405e-09, "loss": 0.4032, "step": 23652 }, { "epoch": 0.9915010008069334, "grad_norm": 2.151686906814575, "learning_rate": 1.8802954661073514e-09, "loss": 0.4809, "step": 23653 }, { "epoch": 0.9915429194219423, "grad_norm": 1.9337587356567383, "learning_rate": 1.861725913008261e-09, "loss": 0.4692, "step": 23654 }, { "epoch": 0.9915848380369513, "grad_norm": 1.796120285987854, "learning_rate": 1.843248493678118e-09, "loss": 0.4762, "step": 23655 }, { "epoch": 0.9916267566519602, "grad_norm": 2.716780424118042, "learning_rate": 1.8248632084583162e-09, "loss": 0.501, "step": 23656 }, { "epoch": 0.9916686752669692, "grad_norm": 1.9009050130844116, "learning_rate": 1.8065700576869182e-09, "loss": 0.4911, "step": 23657 }, { "epoch": 0.9917105938819781, "grad_norm": 1.8785823583602905, "learning_rate": 1.7883690417019871e-09, "loss": 0.4771, "step": 23658 }, { "epoch": 0.9917525124969871, "grad_norm": 2.0500805377960205, "learning_rate": 1.7702601608382553e-09, "loss": 0.4819, "step": 23659 }, { "epoch": 0.991794431111996, "grad_norm": 2.2167389392852783, "learning_rate": 1.7522434154304546e-09, "loss": 0.4858, "step": 23660 }, { "epoch": 0.9918363497270051, "grad_norm": 1.8443365097045898, "learning_rate": 1.7343188058099869e-09, "loss": 0.4541, "step": 23661 }, { "epoch": 0.991878268342014, "grad_norm": 2.0850484371185303, "learning_rate": 1.7164863323071435e-09, "loss": 0.4478, "step": 23662 }, { "epoch": 0.9919201869570229, "grad_norm": 1.8788456916809082, "learning_rate": 1.6987459952516606e-09, "loss": 0.4985, "step": 23663 }, { "epoch": 0.9919621055720319, "grad_norm": 1.8760076761245728, "learning_rate": 1.6810977949693884e-09, "loss": 0.4505, "step": 23664 }, { "epoch": 0.9920040241870408, "grad_norm": 1.845115303993225, "learning_rate": 1.663541731786178e-09, "loss": 0.4564, "step": 23665 }, { "epoch": 0.9920459428020498, "grad_norm": 1.9999680519104004, "learning_rate": 1.6460778060256589e-09, "loss": 0.5061, "step": 23666 }, { "epoch": 0.9920878614170587, "grad_norm": 2.0966503620147705, "learning_rate": 1.6287060180092407e-09, "loss": 0.4883, "step": 23667 }, { "epoch": 0.9921297800320678, "grad_norm": 2.0231006145477295, "learning_rate": 1.611426368057778e-09, "loss": 0.465, "step": 23668 }, { "epoch": 0.9921716986470767, "grad_norm": 1.8204779624938965, "learning_rate": 1.5942388564899047e-09, "loss": 0.4477, "step": 23669 }, { "epoch": 0.9922136172620857, "grad_norm": 1.7081513404846191, "learning_rate": 1.5771434836220345e-09, "loss": 0.3863, "step": 23670 }, { "epoch": 0.9922555358770946, "grad_norm": 1.8372013568878174, "learning_rate": 1.5601402497694706e-09, "loss": 0.5011, "step": 23671 }, { "epoch": 0.9922974544921036, "grad_norm": 1.638698697090149, "learning_rate": 1.543229155245296e-09, "loss": 0.3969, "step": 23672 }, { "epoch": 0.9923393731071125, "grad_norm": 1.8928313255310059, "learning_rate": 1.5264102003620385e-09, "loss": 0.4688, "step": 23673 }, { "epoch": 0.9923812917221215, "grad_norm": 1.7837557792663574, "learning_rate": 1.509683385428895e-09, "loss": 0.4566, "step": 23674 }, { "epoch": 0.9924232103371304, "grad_norm": 2.687187433242798, "learning_rate": 1.4930487107545078e-09, "loss": 0.4647, "step": 23675 }, { "epoch": 0.9924651289521395, "grad_norm": 1.9195270538330078, "learning_rate": 1.4765061766452981e-09, "loss": 0.4703, "step": 23676 }, { "epoch": 0.9925070475671484, "grad_norm": 2.3428213596343994, "learning_rate": 1.4600557834065775e-09, "loss": 0.4842, "step": 23677 }, { "epoch": 0.9925489661821574, "grad_norm": 1.8215209245681763, "learning_rate": 1.443697531341992e-09, "loss": 0.4397, "step": 23678 }, { "epoch": 0.9925908847971663, "grad_norm": 2.270639419555664, "learning_rate": 1.427431420751857e-09, "loss": 0.4683, "step": 23679 }, { "epoch": 0.9926328034121753, "grad_norm": 1.9470024108886719, "learning_rate": 1.4112574519370425e-09, "loss": 0.4564, "step": 23680 }, { "epoch": 0.9926747220271842, "grad_norm": 1.7936121225357056, "learning_rate": 1.395175625195644e-09, "loss": 0.4844, "step": 23681 }, { "epoch": 0.9927166406421932, "grad_norm": 2.53615665435791, "learning_rate": 1.3791859408235352e-09, "loss": 0.5237, "step": 23682 }, { "epoch": 0.9927585592572021, "grad_norm": 1.864071011543274, "learning_rate": 1.3632883991154812e-09, "loss": 0.4472, "step": 23683 }, { "epoch": 0.9928004778722112, "grad_norm": 2.742555856704712, "learning_rate": 1.3474830003651352e-09, "loss": 0.4776, "step": 23684 }, { "epoch": 0.9928423964872201, "grad_norm": 4.842590808868408, "learning_rate": 1.3317697448633759e-09, "loss": 0.4343, "step": 23685 }, { "epoch": 0.9928843151022291, "grad_norm": 1.9209201335906982, "learning_rate": 1.3161486328999717e-09, "loss": 0.459, "step": 23686 }, { "epoch": 0.992926233717238, "grad_norm": 2.459805965423584, "learning_rate": 1.3006196647630254e-09, "loss": 0.4723, "step": 23687 }, { "epoch": 0.9929681523322469, "grad_norm": 1.9115540981292725, "learning_rate": 1.2851828407384193e-09, "loss": 0.4624, "step": 23688 }, { "epoch": 0.9930100709472559, "grad_norm": 1.7101737260818481, "learning_rate": 1.2698381611109256e-09, "loss": 0.4442, "step": 23689 }, { "epoch": 0.9930519895622648, "grad_norm": 2.010380983352661, "learning_rate": 1.2545856261636514e-09, "loss": 0.4664, "step": 23690 }, { "epoch": 0.9930939081772738, "grad_norm": 1.8363873958587646, "learning_rate": 1.2394252361774828e-09, "loss": 0.4545, "step": 23691 }, { "epoch": 0.9931358267922827, "grad_norm": 1.7569793462753296, "learning_rate": 1.2243569914321962e-09, "loss": 0.4386, "step": 23692 }, { "epoch": 0.9931777454072918, "grad_norm": 1.8088656663894653, "learning_rate": 1.2093808922047923e-09, "loss": 0.4466, "step": 23693 }, { "epoch": 0.9932196640223007, "grad_norm": 2.012218713760376, "learning_rate": 1.1944969387722716e-09, "loss": 0.4672, "step": 23694 }, { "epoch": 0.9932615826373097, "grad_norm": 1.6378530263900757, "learning_rate": 1.179705131408304e-09, "loss": 0.4421, "step": 23695 }, { "epoch": 0.9933035012523186, "grad_norm": 2.037123918533325, "learning_rate": 1.1650054703854496e-09, "loss": 0.4762, "step": 23696 }, { "epoch": 0.9933454198673276, "grad_norm": 2.0920159816741943, "learning_rate": 1.1503979559757129e-09, "loss": 0.4651, "step": 23697 }, { "epoch": 0.9933873384823365, "grad_norm": 1.8313777446746826, "learning_rate": 1.1358825884472124e-09, "loss": 0.489, "step": 23698 }, { "epoch": 0.9934292570973455, "grad_norm": 1.9029844999313354, "learning_rate": 1.1214593680686225e-09, "loss": 0.4351, "step": 23699 }, { "epoch": 0.9934711757123544, "grad_norm": 2.059293508529663, "learning_rate": 1.1071282951047312e-09, "loss": 0.4417, "step": 23700 }, { "epoch": 0.9935130943273635, "grad_norm": 2.7092864513397217, "learning_rate": 1.0928893698203269e-09, "loss": 0.4725, "step": 23701 }, { "epoch": 0.9935550129423724, "grad_norm": 1.7453986406326294, "learning_rate": 1.078742592477422e-09, "loss": 0.4275, "step": 23702 }, { "epoch": 0.9935969315573814, "grad_norm": 2.1464121341705322, "learning_rate": 1.064687963336919e-09, "loss": 0.4573, "step": 23703 }, { "epoch": 0.9936388501723903, "grad_norm": 2.023138999938965, "learning_rate": 1.05072548265861e-09, "loss": 0.4463, "step": 23704 }, { "epoch": 0.9936807687873993, "grad_norm": 1.8003662824630737, "learning_rate": 1.0368551506989566e-09, "loss": 0.4936, "step": 23705 }, { "epoch": 0.9937226874024082, "grad_norm": 2.3212530612945557, "learning_rate": 1.0230769677144202e-09, "loss": 0.5028, "step": 23706 }, { "epoch": 0.9937646060174172, "grad_norm": 2.1605050563812256, "learning_rate": 1.0093909339586872e-09, "loss": 0.4832, "step": 23707 }, { "epoch": 0.9938065246324261, "grad_norm": 2.791364908218384, "learning_rate": 9.957970496832226e-10, "loss": 0.4665, "step": 23708 }, { "epoch": 0.9938484432474352, "grad_norm": 2.795786142349243, "learning_rate": 9.822953151400472e-10, "loss": 0.4743, "step": 23709 }, { "epoch": 0.9938903618624441, "grad_norm": 2.5745742321014404, "learning_rate": 9.688857305767407e-10, "loss": 0.4762, "step": 23710 }, { "epoch": 0.9939322804774531, "grad_norm": 3.750319242477417, "learning_rate": 9.55568296241438e-10, "loss": 0.5017, "step": 23711 }, { "epoch": 0.993974199092462, "grad_norm": 2.8001952171325684, "learning_rate": 9.423430123789434e-10, "loss": 0.4333, "step": 23712 }, { "epoch": 0.9940161177074709, "grad_norm": 1.716486930847168, "learning_rate": 9.29209879234061e-10, "loss": 0.4318, "step": 23713 }, { "epoch": 0.9940580363224799, "grad_norm": 1.7128244638442993, "learning_rate": 9.161688970471538e-10, "loss": 0.4594, "step": 23714 }, { "epoch": 0.9940999549374888, "grad_norm": 2.260326623916626, "learning_rate": 9.032200660602508e-10, "loss": 0.5016, "step": 23715 }, { "epoch": 0.9941418735524978, "grad_norm": 1.9803345203399658, "learning_rate": 8.903633865114947e-10, "loss": 0.4619, "step": 23716 }, { "epoch": 0.9941837921675067, "grad_norm": 1.8357024192810059, "learning_rate": 8.77598858637918e-10, "loss": 0.5052, "step": 23717 }, { "epoch": 0.9942257107825158, "grad_norm": 2.3143362998962402, "learning_rate": 8.649264826743331e-10, "loss": 0.4778, "step": 23718 }, { "epoch": 0.9942676293975247, "grad_norm": 2.207005739212036, "learning_rate": 8.523462588549969e-10, "loss": 0.4706, "step": 23719 }, { "epoch": 0.9943095480125337, "grad_norm": 2.2883460521698, "learning_rate": 8.398581874113909e-10, "loss": 0.5152, "step": 23720 }, { "epoch": 0.9943514666275426, "grad_norm": 1.7233483791351318, "learning_rate": 8.274622685738865e-10, "loss": 0.5158, "step": 23721 }, { "epoch": 0.9943933852425516, "grad_norm": 7.166957855224609, "learning_rate": 8.151585025711894e-10, "loss": 0.4647, "step": 23722 }, { "epoch": 0.9944353038575605, "grad_norm": 1.9296687841415405, "learning_rate": 8.029468896297854e-10, "loss": 0.4935, "step": 23723 }, { "epoch": 0.9944772224725695, "grad_norm": 1.733934760093689, "learning_rate": 7.908274299750496e-10, "loss": 0.4766, "step": 23724 }, { "epoch": 0.9945191410875784, "grad_norm": 1.6219578981399536, "learning_rate": 7.788001238295817e-10, "loss": 0.3955, "step": 23725 }, { "epoch": 0.9945610597025875, "grad_norm": 1.7435944080352783, "learning_rate": 7.668649714159815e-10, "loss": 0.4384, "step": 23726 }, { "epoch": 0.9946029783175964, "grad_norm": 2.0165960788726807, "learning_rate": 7.550219729546282e-10, "loss": 0.4576, "step": 23727 }, { "epoch": 0.9946448969326054, "grad_norm": 2.007343292236328, "learning_rate": 7.432711286625705e-10, "loss": 0.4645, "step": 23728 }, { "epoch": 0.9946868155476143, "grad_norm": 2.298855781555176, "learning_rate": 7.316124387568569e-10, "loss": 0.4891, "step": 23729 }, { "epoch": 0.9947287341626233, "grad_norm": 2.177635908126831, "learning_rate": 7.200459034528706e-10, "loss": 0.4976, "step": 23730 }, { "epoch": 0.9947706527776322, "grad_norm": 1.6677970886230469, "learning_rate": 7.085715229637746e-10, "loss": 0.4508, "step": 23731 }, { "epoch": 0.9948125713926412, "grad_norm": 2.0674726963043213, "learning_rate": 6.971892975005113e-10, "loss": 0.4726, "step": 23732 }, { "epoch": 0.9948544900076501, "grad_norm": 1.7925102710723877, "learning_rate": 6.858992272734677e-10, "loss": 0.4649, "step": 23733 }, { "epoch": 0.9948964086226592, "grad_norm": 1.909421682357788, "learning_rate": 6.747013124902558e-10, "loss": 0.489, "step": 23734 }, { "epoch": 0.9949383272376681, "grad_norm": 4.152475357055664, "learning_rate": 6.635955533579319e-10, "loss": 0.5054, "step": 23735 }, { "epoch": 0.9949802458526771, "grad_norm": 1.922108769416809, "learning_rate": 6.525819500807773e-10, "loss": 0.4451, "step": 23736 }, { "epoch": 0.995022164467686, "grad_norm": 1.8496965169906616, "learning_rate": 6.416605028614076e-10, "loss": 0.4751, "step": 23737 }, { "epoch": 0.9950640830826949, "grad_norm": 2.6156423091888428, "learning_rate": 6.308312119024385e-10, "loss": 0.4872, "step": 23738 }, { "epoch": 0.9951060016977039, "grad_norm": 2.1430470943450928, "learning_rate": 6.200940774020447e-10, "loss": 0.5087, "step": 23739 }, { "epoch": 0.9951479203127128, "grad_norm": 2.0174033641815186, "learning_rate": 6.094490995589564e-10, "loss": 0.5032, "step": 23740 }, { "epoch": 0.9951898389277218, "grad_norm": 1.9012205600738525, "learning_rate": 5.988962785691277e-10, "loss": 0.5118, "step": 23741 }, { "epoch": 0.9952317575427307, "grad_norm": 2.236349582672119, "learning_rate": 5.884356146274028e-10, "loss": 0.4596, "step": 23742 }, { "epoch": 0.9952736761577398, "grad_norm": 4.979459762573242, "learning_rate": 5.780671079258504e-10, "loss": 0.4693, "step": 23743 }, { "epoch": 0.9953155947727487, "grad_norm": 1.5975427627563477, "learning_rate": 5.67790758656539e-10, "loss": 0.4287, "step": 23744 }, { "epoch": 0.9953575133877577, "grad_norm": 1.880362629890442, "learning_rate": 5.576065670087616e-10, "loss": 0.5116, "step": 23745 }, { "epoch": 0.9953994320027666, "grad_norm": 2.0004734992980957, "learning_rate": 5.475145331695908e-10, "loss": 0.4277, "step": 23746 }, { "epoch": 0.9954413506177756, "grad_norm": 1.7182892560958862, "learning_rate": 5.37514657325544e-10, "loss": 0.4661, "step": 23747 }, { "epoch": 0.9954832692327845, "grad_norm": 2.195734977722168, "learning_rate": 5.276069396609185e-10, "loss": 0.4456, "step": 23748 }, { "epoch": 0.9955251878477935, "grad_norm": 2.869197368621826, "learning_rate": 5.177913803583457e-10, "loss": 0.5073, "step": 23749 }, { "epoch": 0.9955671064628024, "grad_norm": 2.0156731605529785, "learning_rate": 5.080679795987919e-10, "loss": 0.4873, "step": 23750 }, { "epoch": 0.9956090250778115, "grad_norm": 2.00136661529541, "learning_rate": 4.984367375610033e-10, "loss": 0.4836, "step": 23751 }, { "epoch": 0.9956509436928204, "grad_norm": 2.310770273208618, "learning_rate": 4.888976544231705e-10, "loss": 0.4962, "step": 23752 }, { "epoch": 0.9956928623078294, "grad_norm": 2.038902759552002, "learning_rate": 4.794507303612639e-10, "loss": 0.4604, "step": 23753 }, { "epoch": 0.9957347809228383, "grad_norm": 2.103130578994751, "learning_rate": 4.700959655484782e-10, "loss": 0.4035, "step": 23754 }, { "epoch": 0.9957766995378473, "grad_norm": 2.606362819671631, "learning_rate": 4.608333601580084e-10, "loss": 0.4702, "step": 23755 }, { "epoch": 0.9958186181528562, "grad_norm": 1.8731656074523926, "learning_rate": 4.516629143608287e-10, "loss": 0.4745, "step": 23756 }, { "epoch": 0.9958605367678652, "grad_norm": 1.9854729175567627, "learning_rate": 4.425846283251378e-10, "loss": 0.4625, "step": 23757 }, { "epoch": 0.9959024553828741, "grad_norm": 7.9337568283081055, "learning_rate": 4.335985022185796e-10, "loss": 0.4803, "step": 23758 }, { "epoch": 0.9959443739978832, "grad_norm": 1.7228931188583374, "learning_rate": 4.2470453620713225e-10, "loss": 0.4378, "step": 23759 }, { "epoch": 0.9959862926128921, "grad_norm": 2.6352782249450684, "learning_rate": 4.159027304545538e-10, "loss": 0.4648, "step": 23760 }, { "epoch": 0.9960282112279011, "grad_norm": 2.096132516860962, "learning_rate": 4.071930851229366e-10, "loss": 0.4866, "step": 23761 }, { "epoch": 0.99607012984291, "grad_norm": 1.8436613082885742, "learning_rate": 3.9857560037326326e-10, "loss": 0.4515, "step": 23762 }, { "epoch": 0.9961120484579189, "grad_norm": 1.8491243124008179, "learning_rate": 3.900502763642955e-10, "loss": 0.4681, "step": 23763 }, { "epoch": 0.9961539670729279, "grad_norm": 1.8991001844406128, "learning_rate": 3.816171132525748e-10, "loss": 0.4792, "step": 23764 }, { "epoch": 0.9961958856879368, "grad_norm": 1.8491225242614746, "learning_rate": 3.7327611119408745e-10, "loss": 0.4911, "step": 23765 }, { "epoch": 0.9962378043029458, "grad_norm": 2.04435133934021, "learning_rate": 3.6502727034259946e-10, "loss": 0.494, "step": 23766 }, { "epoch": 0.9962797229179547, "grad_norm": 1.8042258024215698, "learning_rate": 3.5687059084965617e-10, "loss": 0.479, "step": 23767 }, { "epoch": 0.9963216415329638, "grad_norm": 1.9033452272415161, "learning_rate": 3.48806072866803e-10, "loss": 0.4748, "step": 23768 }, { "epoch": 0.9963635601479727, "grad_norm": 2.704634666442871, "learning_rate": 3.408337165411446e-10, "loss": 0.4266, "step": 23769 }, { "epoch": 0.9964054787629817, "grad_norm": 2.2318875789642334, "learning_rate": 3.329535220203406e-10, "loss": 0.4457, "step": 23770 }, { "epoch": 0.9964473973779906, "grad_norm": 2.138383626937866, "learning_rate": 3.2516548945038526e-10, "loss": 0.5496, "step": 23771 }, { "epoch": 0.9964893159929996, "grad_norm": 2.1387386322021484, "learning_rate": 3.1746961897338723e-10, "loss": 0.4459, "step": 23772 }, { "epoch": 0.9965312346080085, "grad_norm": 1.7989202737808228, "learning_rate": 3.0986591073256523e-10, "loss": 0.4768, "step": 23773 }, { "epoch": 0.9965731532230175, "grad_norm": 1.9890536069869995, "learning_rate": 3.0235436486725225e-10, "loss": 0.4934, "step": 23774 }, { "epoch": 0.9966150718380264, "grad_norm": 2.0536046028137207, "learning_rate": 2.9493498151622614e-10, "loss": 0.4145, "step": 23775 }, { "epoch": 0.9966569904530355, "grad_norm": 10.925516128540039, "learning_rate": 2.876077608165995e-10, "loss": 0.485, "step": 23776 }, { "epoch": 0.9966989090680444, "grad_norm": 2.0720129013061523, "learning_rate": 2.8037270290215415e-10, "loss": 0.5072, "step": 23777 }, { "epoch": 0.9967408276830534, "grad_norm": 2.1425368785858154, "learning_rate": 2.7322980790778217e-10, "loss": 0.4997, "step": 23778 }, { "epoch": 0.9967827462980623, "grad_norm": 2.1985740661621094, "learning_rate": 2.6617907596449e-10, "loss": 0.484, "step": 23779 }, { "epoch": 0.9968246649130713, "grad_norm": 2.370600700378418, "learning_rate": 2.592205072021736e-10, "loss": 0.4332, "step": 23780 }, { "epoch": 0.9968665835280802, "grad_norm": 1.9632198810577393, "learning_rate": 2.523541017490638e-10, "loss": 0.5159, "step": 23781 }, { "epoch": 0.9969085021430892, "grad_norm": 1.8266907930374146, "learning_rate": 2.4557985973172603e-10, "loss": 0.4355, "step": 23782 }, { "epoch": 0.9969504207580981, "grad_norm": 1.8339698314666748, "learning_rate": 2.3889778127561544e-10, "loss": 0.4419, "step": 23783 }, { "epoch": 0.9969923393731072, "grad_norm": 1.8465133905410767, "learning_rate": 2.3230786650285663e-10, "loss": 0.4659, "step": 23784 }, { "epoch": 0.9970342579881161, "grad_norm": 1.645074725151062, "learning_rate": 2.258101155361292e-10, "loss": 0.4407, "step": 23785 }, { "epoch": 0.9970761766031251, "grad_norm": 1.7765754461288452, "learning_rate": 2.19404528494227e-10, "loss": 0.4707, "step": 23786 }, { "epoch": 0.997118095218134, "grad_norm": 2.8675644397735596, "learning_rate": 2.130911054959439e-10, "loss": 0.4918, "step": 23787 }, { "epoch": 0.9971600138331429, "grad_norm": 1.8513013124465942, "learning_rate": 2.0686984665674316e-10, "loss": 0.4716, "step": 23788 }, { "epoch": 0.9972019324481519, "grad_norm": 1.963639259338379, "learning_rate": 2.0074075209264298e-10, "loss": 0.4192, "step": 23789 }, { "epoch": 0.9972438510631608, "grad_norm": 1.9298373460769653, "learning_rate": 1.9470382191522087e-10, "loss": 0.4543, "step": 23790 }, { "epoch": 0.9972857696781698, "grad_norm": 2.100738286972046, "learning_rate": 1.8875905623660928e-10, "loss": 0.4744, "step": 23791 }, { "epoch": 0.9973276882931787, "grad_norm": 1.7614004611968994, "learning_rate": 1.8290645516616522e-10, "loss": 0.4984, "step": 23792 }, { "epoch": 0.9973696069081878, "grad_norm": 1.918920874595642, "learning_rate": 1.7714601881213546e-10, "loss": 0.4728, "step": 23793 }, { "epoch": 0.9974115255231967, "grad_norm": 3.3336246013641357, "learning_rate": 1.714777472799911e-10, "loss": 0.4423, "step": 23794 }, { "epoch": 0.9974534441382057, "grad_norm": 1.9732160568237305, "learning_rate": 1.6590164067520342e-10, "loss": 0.5039, "step": 23795 }, { "epoch": 0.9974953627532146, "grad_norm": 1.667852520942688, "learning_rate": 1.6041769909935778e-10, "loss": 0.4598, "step": 23796 }, { "epoch": 0.9975372813682236, "grad_norm": 2.0019752979278564, "learning_rate": 1.550259226540396e-10, "loss": 0.472, "step": 23797 }, { "epoch": 0.9975791999832325, "grad_norm": 1.903287649154663, "learning_rate": 1.4972631143916893e-10, "loss": 0.472, "step": 23798 }, { "epoch": 0.9976211185982415, "grad_norm": 1.8426625728607178, "learning_rate": 1.4451886555189032e-10, "loss": 0.4469, "step": 23799 }, { "epoch": 0.9976630372132504, "grad_norm": 2.033555507659912, "learning_rate": 1.3940358508823804e-10, "loss": 0.5187, "step": 23800 }, { "epoch": 0.9977049558282595, "grad_norm": 1.6380321979522705, "learning_rate": 1.3438047014258105e-10, "loss": 0.4413, "step": 23801 }, { "epoch": 0.9977468744432684, "grad_norm": 3.1238009929656982, "learning_rate": 1.2944952080762297e-10, "loss": 0.4919, "step": 23802 }, { "epoch": 0.9977887930582774, "grad_norm": 1.7445334196090698, "learning_rate": 1.2461073717384697e-10, "loss": 0.4349, "step": 23803 }, { "epoch": 0.9978307116732863, "grad_norm": 1.9066503047943115, "learning_rate": 1.1986411933118114e-10, "loss": 0.4794, "step": 23804 }, { "epoch": 0.9978726302882953, "grad_norm": 1.8979599475860596, "learning_rate": 1.1520966736622286e-10, "loss": 0.4832, "step": 23805 }, { "epoch": 0.9979145489033042, "grad_norm": 2.020509719848633, "learning_rate": 1.1064738136556952e-10, "loss": 0.4707, "step": 23806 }, { "epoch": 0.9979564675183132, "grad_norm": 2.601320743560791, "learning_rate": 1.0617726141304296e-10, "loss": 0.4097, "step": 23807 }, { "epoch": 0.9979983861333221, "grad_norm": 1.863446593284607, "learning_rate": 1.017993075907997e-10, "loss": 0.4844, "step": 23808 }, { "epoch": 0.9980403047483312, "grad_norm": 2.1538655757904053, "learning_rate": 9.7513519979886e-11, "loss": 0.4912, "step": 23809 }, { "epoch": 0.9980822233633401, "grad_norm": 1.84976065158844, "learning_rate": 9.331989865912772e-11, "loss": 0.4276, "step": 23810 }, { "epoch": 0.9981241419783491, "grad_norm": 2.4457850456237793, "learning_rate": 8.921844370568533e-11, "loss": 0.4626, "step": 23811 }, { "epoch": 0.998166060593358, "grad_norm": 2.206219434738159, "learning_rate": 8.520915519505401e-11, "loss": 0.4614, "step": 23812 }, { "epoch": 0.9982079792083669, "grad_norm": 1.870562195777893, "learning_rate": 8.129203320217383e-11, "loss": 0.4387, "step": 23813 }, { "epoch": 0.9982498978233759, "grad_norm": 1.7654213905334473, "learning_rate": 7.746707779809903e-11, "loss": 0.4489, "step": 23814 }, { "epoch": 0.9982918164383848, "grad_norm": 2.0775434970855713, "learning_rate": 7.373428905388392e-11, "loss": 0.4837, "step": 23815 }, { "epoch": 0.9983337350533938, "grad_norm": 1.8180088996887207, "learning_rate": 7.009366703780717e-11, "loss": 0.4323, "step": 23816 }, { "epoch": 0.9983756536684028, "grad_norm": 1.9508200883865356, "learning_rate": 6.654521181759244e-11, "loss": 0.4539, "step": 23817 }, { "epoch": 0.9984175722834118, "grad_norm": 2.2900068759918213, "learning_rate": 6.308892345818774e-11, "loss": 0.456, "step": 23818 }, { "epoch": 0.9984594908984207, "grad_norm": 1.9201868772506714, "learning_rate": 5.972480202398601e-11, "loss": 0.4521, "step": 23819 }, { "epoch": 0.9985014095134297, "grad_norm": 1.8219549655914307, "learning_rate": 5.645284757660463e-11, "loss": 0.5251, "step": 23820 }, { "epoch": 0.9985433281284386, "grad_norm": 2.0414934158325195, "learning_rate": 5.327306017599565e-11, "loss": 0.4295, "step": 23821 }, { "epoch": 0.9985852467434476, "grad_norm": 2.2175474166870117, "learning_rate": 5.018543988100089e-11, "loss": 0.4925, "step": 23822 }, { "epoch": 0.9986271653584565, "grad_norm": 1.8875004053115845, "learning_rate": 4.718998674879682e-11, "loss": 0.4688, "step": 23823 }, { "epoch": 0.9986690839734655, "grad_norm": 1.7163176536560059, "learning_rate": 4.42867008343395e-11, "loss": 0.4501, "step": 23824 }, { "epoch": 0.9987110025884745, "grad_norm": 1.9905873537063599, "learning_rate": 4.147558219091963e-11, "loss": 0.4389, "step": 23825 }, { "epoch": 0.9987529212034835, "grad_norm": 3.463869333267212, "learning_rate": 3.8756630871272794e-11, "loss": 0.5075, "step": 23826 }, { "epoch": 0.9987948398184924, "grad_norm": 2.390462636947632, "learning_rate": 3.612984692480392e-11, "loss": 0.4922, "step": 23827 }, { "epoch": 0.9988367584335014, "grad_norm": 1.8695424795150757, "learning_rate": 3.3595230399807724e-11, "loss": 0.496, "step": 23828 }, { "epoch": 0.9988786770485103, "grad_norm": 1.9180368185043335, "learning_rate": 3.115278134291355e-11, "loss": 0.4586, "step": 23829 }, { "epoch": 0.9989205956635193, "grad_norm": 1.8945233821868896, "learning_rate": 2.8802499800195672e-11, "loss": 0.5024, "step": 23830 }, { "epoch": 0.9989625142785282, "grad_norm": 2.4185824394226074, "learning_rate": 2.6544385813842555e-11, "loss": 0.482, "step": 23831 }, { "epoch": 0.9990044328935372, "grad_norm": 1.7257723808288574, "learning_rate": 2.4378439425487565e-11, "loss": 0.4212, "step": 23832 }, { "epoch": 0.9990463515085461, "grad_norm": 1.9241554737091064, "learning_rate": 2.2304660676208956e-11, "loss": 0.4821, "step": 23833 }, { "epoch": 0.9990882701235552, "grad_norm": 1.8026036024093628, "learning_rate": 2.0323049603199195e-11, "loss": 0.4438, "step": 23834 }, { "epoch": 0.9991301887385641, "grad_norm": 2.2761433124542236, "learning_rate": 1.8433606243095647e-11, "loss": 0.5014, "step": 23835 }, { "epoch": 0.9991721073535731, "grad_norm": 2.159959554672241, "learning_rate": 1.6636330631425446e-11, "loss": 0.4789, "step": 23836 }, { "epoch": 0.999214025968582, "grad_norm": 2.111297607421875, "learning_rate": 1.4931222800385058e-11, "loss": 0.481, "step": 23837 }, { "epoch": 0.9992559445835909, "grad_norm": 2.317007064819336, "learning_rate": 1.3318282781615843e-11, "loss": 0.4458, "step": 23838 }, { "epoch": 0.9992978631985999, "grad_norm": 1.9555898904800415, "learning_rate": 1.179751060509382e-11, "loss": 0.4305, "step": 23839 }, { "epoch": 0.9993397818136088, "grad_norm": 2.6479973793029785, "learning_rate": 1.0368906299129678e-11, "loss": 0.4682, "step": 23840 }, { "epoch": 0.9993817004286178, "grad_norm": 1.8162815570831299, "learning_rate": 9.032469889813655e-12, "loss": 0.4829, "step": 23841 }, { "epoch": 0.9994236190436268, "grad_norm": 3.2363388538360596, "learning_rate": 7.788201401570661e-12, "loss": 0.5252, "step": 23842 }, { "epoch": 0.9994655376586358, "grad_norm": 1.9247050285339355, "learning_rate": 6.636100857160266e-12, "loss": 0.486, "step": 23843 }, { "epoch": 0.9995074562736447, "grad_norm": 2.420443534851074, "learning_rate": 5.5761682782318196e-12, "loss": 0.4915, "step": 23844 }, { "epoch": 0.9995493748886537, "grad_norm": 3.2881531715393066, "learning_rate": 4.608403684769336e-12, "loss": 0.4449, "step": 23845 }, { "epoch": 0.9995912935036626, "grad_norm": 2.0606582164764404, "learning_rate": 3.73280709342616e-12, "loss": 0.4595, "step": 23846 }, { "epoch": 0.9996332121186716, "grad_norm": 1.927813172340393, "learning_rate": 2.9493785214107507e-12, "loss": 0.4681, "step": 23847 }, { "epoch": 0.9996751307336805, "grad_norm": 1.7477246522903442, "learning_rate": 2.2581179826008937e-12, "loss": 0.4691, "step": 23848 }, { "epoch": 0.9997170493486895, "grad_norm": 7.537197113037109, "learning_rate": 1.6590254897641544e-12, "loss": 0.4813, "step": 23849 }, { "epoch": 0.9997589679636985, "grad_norm": 2.48791241645813, "learning_rate": 1.1521010540027632e-12, "loss": 0.4458, "step": 23850 }, { "epoch": 0.9998008865787075, "grad_norm": 1.9044184684753418, "learning_rate": 7.373446847536159e-13, "loss": 0.4285, "step": 23851 }, { "epoch": 0.9998428051937164, "grad_norm": 3.337296485900879, "learning_rate": 4.1475638978827337e-13, "loss": 0.4888, "step": 23852 }, { "epoch": 0.9998847238087254, "grad_norm": 2.3598270416259766, "learning_rate": 1.8433617465785092e-13, "loss": 0.4935, "step": 23853 }, { "epoch": 0.9999266424237343, "grad_norm": 2.360267162322998, "learning_rate": 4.608404380324061e-14, "loss": 0.5229, "step": 23854 }, { "epoch": 0.9999685610387433, "grad_norm": 1.7850053310394287, "learning_rate": 0.0, "loss": 0.4649, "step": 23855 }, { "epoch": 0.9999685610387433, "step": 23855, "total_flos": 4.867355317124255e+19, "train_loss": 0.381923864584153, "train_runtime": 467796.4155, "train_samples_per_second": 13.055, "train_steps_per_second": 0.051 } ], "logging_steps": 1.0, "max_steps": 23855, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 200, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 4.867355317124255e+19, "train_batch_size": 1, "trial_name": null, "trial_params": null }