{ "best_metric": null, "best_model_checkpoint": null, "epoch": 5.69545389767707, "eval_steps": 500, "global_step": 5000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.005700441784238278, "grad_norm": 27.39983367919922, "learning_rate": 4.9999995544380894e-05, "loss": 16.8508, "num_input_tokens_seen": 97712, "step": 5 }, { "epoch": 0.011400883568476556, "grad_norm": 35.96725082397461, "learning_rate": 4.999998217752515e-05, "loss": 11.7177, "num_input_tokens_seen": 195504, "step": 10 }, { "epoch": 0.017101325352714837, "grad_norm": 13.93742847442627, "learning_rate": 4.999995989943754e-05, "loss": 6.3848, "num_input_tokens_seen": 293200, "step": 15 }, { "epoch": 0.022801767136953113, "grad_norm": 9.999053955078125, "learning_rate": 4.9999928710126e-05, "loss": 4.4249, "num_input_tokens_seen": 390960, "step": 20 }, { "epoch": 0.02850220892119139, "grad_norm": 11.072565078735352, "learning_rate": 4.999988860960165e-05, "loss": 3.7022, "num_input_tokens_seen": 488752, "step": 25 }, { "epoch": 0.034202650705429674, "grad_norm": 6.960489273071289, "learning_rate": 4.9999839597878784e-05, "loss": 2.5291, "num_input_tokens_seen": 586496, "step": 30 }, { "epoch": 0.039903092489667946, "grad_norm": 10.353778839111328, "learning_rate": 4.999978167497488e-05, "loss": 1.9498, "num_input_tokens_seen": 684240, "step": 35 }, { "epoch": 0.045603534273906225, "grad_norm": 12.493133544921875, "learning_rate": 4.999971484091057e-05, "loss": 1.551, "num_input_tokens_seen": 781936, "step": 40 }, { "epoch": 0.051303976058144504, "grad_norm": 16.28146743774414, "learning_rate": 4.999963909570968e-05, "loss": 1.4018, "num_input_tokens_seen": 879680, "step": 45 }, { "epoch": 0.05700441784238278, "grad_norm": 9.161210060119629, "learning_rate": 4.999955443939922e-05, "loss": 1.2093, "num_input_tokens_seen": 977440, "step": 50 }, { "epoch": 0.06270485962662106, "grad_norm": 13.297211647033691, "learning_rate": 4.9999460872009366e-05, "loss": 1.1716, "num_input_tokens_seen": 1075200, "step": 55 }, { "epoch": 0.06840530141085935, "grad_norm": 8.317058563232422, "learning_rate": 4.9999358393573445e-05, "loss": 1.1838, "num_input_tokens_seen": 1172880, "step": 60 }, { "epoch": 0.07410574319509762, "grad_norm": 7.10370397567749, "learning_rate": 4.9999247004128014e-05, "loss": 1.0844, "num_input_tokens_seen": 1270608, "step": 65 }, { "epoch": 0.07980618497933589, "grad_norm": 10.387308120727539, "learning_rate": 4.9999126703712775e-05, "loss": 1.0746, "num_input_tokens_seen": 1368368, "step": 70 }, { "epoch": 0.08550662676357418, "grad_norm": 10.262731552124023, "learning_rate": 4.999899749237059e-05, "loss": 1.0698, "num_input_tokens_seen": 1466016, "step": 75 }, { "epoch": 0.09120706854781245, "grad_norm": 9.300691604614258, "learning_rate": 4.9998859370147524e-05, "loss": 1.1167, "num_input_tokens_seen": 1563648, "step": 80 }, { "epoch": 0.09690751033205074, "grad_norm": 4.99461030960083, "learning_rate": 4.999871233709282e-05, "loss": 1.105, "num_input_tokens_seen": 1661456, "step": 85 }, { "epoch": 0.10260795211628901, "grad_norm": 8.910074234008789, "learning_rate": 4.9998556393258884e-05, "loss": 1.1096, "num_input_tokens_seen": 1759184, "step": 90 }, { "epoch": 0.1083083939005273, "grad_norm": 7.826836585998535, "learning_rate": 4.9998391538701293e-05, "loss": 1.1084, "num_input_tokens_seen": 1856848, "step": 95 }, { "epoch": 0.11400883568476557, "grad_norm": 8.221962928771973, "learning_rate": 4.999821777347883e-05, "loss": 1.1534, "num_input_tokens_seen": 1954544, "step": 100 }, { "epoch": 0.11970927746900385, "grad_norm": 8.834911346435547, "learning_rate": 4.9998035097653406e-05, "loss": 1.075, "num_input_tokens_seen": 2052224, "step": 105 }, { "epoch": 0.12540971925324212, "grad_norm": 7.220479488372803, "learning_rate": 4.9997843511290156e-05, "loss": 1.085, "num_input_tokens_seen": 2150000, "step": 110 }, { "epoch": 0.1311101610374804, "grad_norm": 7.887118816375732, "learning_rate": 4.999764301445736e-05, "loss": 1.0384, "num_input_tokens_seen": 2247808, "step": 115 }, { "epoch": 0.1368106028217187, "grad_norm": 9.087337493896484, "learning_rate": 4.9997433607226495e-05, "loss": 1.1907, "num_input_tokens_seen": 2345584, "step": 120 }, { "epoch": 0.14251104460595695, "grad_norm": 7.334683418273926, "learning_rate": 4.9997215289672194e-05, "loss": 1.0865, "num_input_tokens_seen": 2443360, "step": 125 }, { "epoch": 0.14821148639019524, "grad_norm": 6.93471097946167, "learning_rate": 4.9996988061872284e-05, "loss": 1.0469, "num_input_tokens_seen": 2541120, "step": 130 }, { "epoch": 0.15391192817443353, "grad_norm": 8.000816345214844, "learning_rate": 4.999675192390776e-05, "loss": 1.0966, "num_input_tokens_seen": 2638912, "step": 135 }, { "epoch": 0.15961236995867178, "grad_norm": 6.232958793640137, "learning_rate": 4.999650687586278e-05, "loss": 1.0418, "num_input_tokens_seen": 2736624, "step": 140 }, { "epoch": 0.16531281174291007, "grad_norm": 12.828268051147461, "learning_rate": 4.999625291782471e-05, "loss": 1.0684, "num_input_tokens_seen": 2834384, "step": 145 }, { "epoch": 0.17101325352714836, "grad_norm": 6.147856712341309, "learning_rate": 4.999599004988406e-05, "loss": 0.9802, "num_input_tokens_seen": 2932160, "step": 150 }, { "epoch": 0.17671369531138664, "grad_norm": 7.085546016693115, "learning_rate": 4.999571827213454e-05, "loss": 1.1506, "num_input_tokens_seen": 3029904, "step": 155 }, { "epoch": 0.1824141370956249, "grad_norm": 8.019725799560547, "learning_rate": 4.999543758467301e-05, "loss": 1.0185, "num_input_tokens_seen": 3127648, "step": 160 }, { "epoch": 0.1881145788798632, "grad_norm": 5.743609428405762, "learning_rate": 4.9995147987599536e-05, "loss": 1.0001, "num_input_tokens_seen": 3225360, "step": 165 }, { "epoch": 0.19381502066410147, "grad_norm": 6.749576568603516, "learning_rate": 4.999484948101734e-05, "loss": 1.0848, "num_input_tokens_seen": 3323152, "step": 170 }, { "epoch": 0.19951546244833976, "grad_norm": 7.745048522949219, "learning_rate": 4.9994542065032823e-05, "loss": 1.0074, "num_input_tokens_seen": 3420912, "step": 175 }, { "epoch": 0.20521590423257802, "grad_norm": 6.615988254547119, "learning_rate": 4.9994225739755565e-05, "loss": 1.0756, "num_input_tokens_seen": 3518752, "step": 180 }, { "epoch": 0.2109163460168163, "grad_norm": 6.993303298950195, "learning_rate": 4.999390050529831e-05, "loss": 1.0371, "num_input_tokens_seen": 3616560, "step": 185 }, { "epoch": 0.2166167878010546, "grad_norm": 6.100363731384277, "learning_rate": 4.9993566361777e-05, "loss": 0.9687, "num_input_tokens_seen": 3714320, "step": 190 }, { "epoch": 0.22231722958529285, "grad_norm": 5.574942111968994, "learning_rate": 4.999322330931074e-05, "loss": 1.0173, "num_input_tokens_seen": 3812144, "step": 195 }, { "epoch": 0.22801767136953113, "grad_norm": 7.2291975021362305, "learning_rate": 4.9992871348021804e-05, "loss": 1.0322, "num_input_tokens_seen": 3909824, "step": 200 }, { "epoch": 0.23371811315376942, "grad_norm": 6.874391078948975, "learning_rate": 4.999251047803565e-05, "loss": 1.0096, "num_input_tokens_seen": 4007600, "step": 205 }, { "epoch": 0.2394185549380077, "grad_norm": 9.4487886428833, "learning_rate": 4.9992140699480914e-05, "loss": 0.9313, "num_input_tokens_seen": 4105360, "step": 210 }, { "epoch": 0.24511899672224596, "grad_norm": 8.49326229095459, "learning_rate": 4.99917620124894e-05, "loss": 1.008, "num_input_tokens_seen": 4203072, "step": 215 }, { "epoch": 0.25081943850648425, "grad_norm": 8.347270965576172, "learning_rate": 4.999137441719609e-05, "loss": 0.9588, "num_input_tokens_seen": 4300784, "step": 220 }, { "epoch": 0.25651988029072254, "grad_norm": 7.798429489135742, "learning_rate": 4.999097791373915e-05, "loss": 1.0412, "num_input_tokens_seen": 4398448, "step": 225 }, { "epoch": 0.2622203220749608, "grad_norm": 7.584600448608398, "learning_rate": 4.99905725022599e-05, "loss": 0.9457, "num_input_tokens_seen": 4496256, "step": 230 }, { "epoch": 0.2679207638591991, "grad_norm": 5.460471153259277, "learning_rate": 4.9990158182902866e-05, "loss": 0.8931, "num_input_tokens_seen": 4594032, "step": 235 }, { "epoch": 0.2736212056434374, "grad_norm": 6.889909267425537, "learning_rate": 4.9989734955815715e-05, "loss": 0.846, "num_input_tokens_seen": 4691824, "step": 240 }, { "epoch": 0.2793216474276756, "grad_norm": 8.376734733581543, "learning_rate": 4.998930282114932e-05, "loss": 0.9712, "num_input_tokens_seen": 4789568, "step": 245 }, { "epoch": 0.2850220892119139, "grad_norm": 6.110357284545898, "learning_rate": 4.99888617790577e-05, "loss": 0.9427, "num_input_tokens_seen": 4887296, "step": 250 }, { "epoch": 0.2907225309961522, "grad_norm": 7.7024102210998535, "learning_rate": 4.998841182969808e-05, "loss": 0.8296, "num_input_tokens_seen": 4984976, "step": 255 }, { "epoch": 0.2964229727803905, "grad_norm": 6.920788288116455, "learning_rate": 4.998795297323083e-05, "loss": 1.0276, "num_input_tokens_seen": 5082688, "step": 260 }, { "epoch": 0.30212341456462877, "grad_norm": 7.553328514099121, "learning_rate": 4.9987485209819515e-05, "loss": 1.0488, "num_input_tokens_seen": 5180400, "step": 265 }, { "epoch": 0.30782385634886705, "grad_norm": 6.883415699005127, "learning_rate": 4.998700853963088e-05, "loss": 0.9426, "num_input_tokens_seen": 5278208, "step": 270 }, { "epoch": 0.31352429813310534, "grad_norm": 11.664554595947266, "learning_rate": 4.998652296283481e-05, "loss": 0.9294, "num_input_tokens_seen": 5375968, "step": 275 }, { "epoch": 0.31922473991734357, "grad_norm": 11.442777633666992, "learning_rate": 4.9986028479604416e-05, "loss": 1.0263, "num_input_tokens_seen": 5473760, "step": 280 }, { "epoch": 0.32492518170158186, "grad_norm": 7.503568649291992, "learning_rate": 4.9985525090115936e-05, "loss": 0.8616, "num_input_tokens_seen": 5571472, "step": 285 }, { "epoch": 0.33062562348582014, "grad_norm": 4.156918048858643, "learning_rate": 4.998501279454881e-05, "loss": 0.867, "num_input_tokens_seen": 5669136, "step": 290 }, { "epoch": 0.3363260652700584, "grad_norm": 8.972600936889648, "learning_rate": 4.998449159308565e-05, "loss": 0.9869, "num_input_tokens_seen": 5766816, "step": 295 }, { "epoch": 0.3420265070542967, "grad_norm": 6.943081855773926, "learning_rate": 4.9983961485912235e-05, "loss": 0.8677, "num_input_tokens_seen": 5864576, "step": 300 }, { "epoch": 0.347726948838535, "grad_norm": 5.271353244781494, "learning_rate": 4.9983422473217514e-05, "loss": 0.929, "num_input_tokens_seen": 5962384, "step": 305 }, { "epoch": 0.3534273906227733, "grad_norm": 6.874100685119629, "learning_rate": 4.998287455519363e-05, "loss": 0.8697, "num_input_tokens_seen": 6060160, "step": 310 }, { "epoch": 0.3591278324070115, "grad_norm": 6.316469192504883, "learning_rate": 4.998231773203587e-05, "loss": 0.8826, "num_input_tokens_seen": 6157920, "step": 315 }, { "epoch": 0.3648282741912498, "grad_norm": 6.3930816650390625, "learning_rate": 4.9981752003942734e-05, "loss": 0.9108, "num_input_tokens_seen": 6255600, "step": 320 }, { "epoch": 0.3705287159754881, "grad_norm": 7.396681785583496, "learning_rate": 4.998117737111587e-05, "loss": 0.9613, "num_input_tokens_seen": 6353424, "step": 325 }, { "epoch": 0.3762291577597264, "grad_norm": 9.793058395385742, "learning_rate": 4.998059383376009e-05, "loss": 0.8664, "num_input_tokens_seen": 6451184, "step": 330 }, { "epoch": 0.38192959954396466, "grad_norm": 4.0863423347473145, "learning_rate": 4.998000139208342e-05, "loss": 0.8693, "num_input_tokens_seen": 6549040, "step": 335 }, { "epoch": 0.38763004132820295, "grad_norm": 4.3018317222595215, "learning_rate": 4.997940004629702e-05, "loss": 0.9368, "num_input_tokens_seen": 6646752, "step": 340 }, { "epoch": 0.39333048311244123, "grad_norm": 16.874574661254883, "learning_rate": 4.9978789796615235e-05, "loss": 1.0444, "num_input_tokens_seen": 6744544, "step": 345 }, { "epoch": 0.3990309248966795, "grad_norm": 6.2149658203125, "learning_rate": 4.9978170643255604e-05, "loss": 0.9418, "num_input_tokens_seen": 6842256, "step": 350 }, { "epoch": 0.40473136668091775, "grad_norm": 6.908440113067627, "learning_rate": 4.997754258643882e-05, "loss": 0.8389, "num_input_tokens_seen": 6939984, "step": 355 }, { "epoch": 0.41043180846515603, "grad_norm": 9.332175254821777, "learning_rate": 4.997690562638874e-05, "loss": 0.9898, "num_input_tokens_seen": 7037776, "step": 360 }, { "epoch": 0.4161322502493943, "grad_norm": 7.081879138946533, "learning_rate": 4.9976259763332423e-05, "loss": 0.8761, "num_input_tokens_seen": 7135552, "step": 365 }, { "epoch": 0.4218326920336326, "grad_norm": 5.079131603240967, "learning_rate": 4.9975604997500084e-05, "loss": 0.8808, "num_input_tokens_seen": 7233248, "step": 370 }, { "epoch": 0.4275331338178709, "grad_norm": 7.381295680999756, "learning_rate": 4.99749413291251e-05, "loss": 0.9706, "num_input_tokens_seen": 7330976, "step": 375 }, { "epoch": 0.4332335756021092, "grad_norm": 4.044100284576416, "learning_rate": 4.9974268758444054e-05, "loss": 0.8972, "num_input_tokens_seen": 7428704, "step": 380 }, { "epoch": 0.43893401738634746, "grad_norm": 6.039126396179199, "learning_rate": 4.9973587285696674e-05, "loss": 0.7717, "num_input_tokens_seen": 7526480, "step": 385 }, { "epoch": 0.4446344591705857, "grad_norm": 5.874084949493408, "learning_rate": 4.997289691112588e-05, "loss": 0.9446, "num_input_tokens_seen": 7624320, "step": 390 }, { "epoch": 0.450334900954824, "grad_norm": 7.415895462036133, "learning_rate": 4.997219763497774e-05, "loss": 0.7123, "num_input_tokens_seen": 7722064, "step": 395 }, { "epoch": 0.45603534273906227, "grad_norm": 7.707664966583252, "learning_rate": 4.997148945750153e-05, "loss": 0.7859, "num_input_tokens_seen": 7819808, "step": 400 }, { "epoch": 0.46173578452330055, "grad_norm": 5.500309467315674, "learning_rate": 4.9970772378949655e-05, "loss": 0.826, "num_input_tokens_seen": 7917488, "step": 405 }, { "epoch": 0.46743622630753884, "grad_norm": 7.652528285980225, "learning_rate": 4.9970046399577734e-05, "loss": 0.8709, "num_input_tokens_seen": 8015264, "step": 410 }, { "epoch": 0.4731366680917771, "grad_norm": 6.417993545532227, "learning_rate": 4.996931151964455e-05, "loss": 0.9764, "num_input_tokens_seen": 8113024, "step": 415 }, { "epoch": 0.4788371098760154, "grad_norm": 5.648680210113525, "learning_rate": 4.996856773941202e-05, "loss": 0.8233, "num_input_tokens_seen": 8210784, "step": 420 }, { "epoch": 0.4845375516602537, "grad_norm": 8.321767807006836, "learning_rate": 4.9967815059145296e-05, "loss": 0.8556, "num_input_tokens_seen": 8308512, "step": 425 }, { "epoch": 0.4902379934444919, "grad_norm": 6.381886005401611, "learning_rate": 4.9967053479112656e-05, "loss": 0.7687, "num_input_tokens_seen": 8406208, "step": 430 }, { "epoch": 0.4959384352287302, "grad_norm": 7.855834007263184, "learning_rate": 4.996628299958557e-05, "loss": 0.7965, "num_input_tokens_seen": 8503952, "step": 435 }, { "epoch": 0.5016388770129685, "grad_norm": 8.358772277832031, "learning_rate": 4.996550362083866e-05, "loss": 0.7877, "num_input_tokens_seen": 8601616, "step": 440 }, { "epoch": 0.5073393187972067, "grad_norm": 8.553559303283691, "learning_rate": 4.996471534314976e-05, "loss": 0.76, "num_input_tokens_seen": 8699424, "step": 445 }, { "epoch": 0.5130397605814451, "grad_norm": 8.631624221801758, "learning_rate": 4.9963918166799836e-05, "loss": 0.8425, "num_input_tokens_seen": 8797088, "step": 450 }, { "epoch": 0.5187402023656833, "grad_norm": 11.236102104187012, "learning_rate": 4.9963112092073046e-05, "loss": 0.8332, "num_input_tokens_seen": 8894848, "step": 455 }, { "epoch": 0.5244406441499216, "grad_norm": 6.356544494628906, "learning_rate": 4.996229711925671e-05, "loss": 0.8231, "num_input_tokens_seen": 8992576, "step": 460 }, { "epoch": 0.5301410859341599, "grad_norm": 4.418157577514648, "learning_rate": 4.996147324864132e-05, "loss": 0.7168, "num_input_tokens_seen": 9090272, "step": 465 }, { "epoch": 0.5358415277183982, "grad_norm": 8.712305068969727, "learning_rate": 4.996064048052056e-05, "loss": 0.7672, "num_input_tokens_seen": 9188080, "step": 470 }, { "epoch": 0.5415419695026364, "grad_norm": 8.759718894958496, "learning_rate": 4.995979881519126e-05, "loss": 0.7601, "num_input_tokens_seen": 9285872, "step": 475 }, { "epoch": 0.5472424112868748, "grad_norm": 7.049539089202881, "learning_rate": 4.995894825295343e-05, "loss": 0.802, "num_input_tokens_seen": 9383584, "step": 480 }, { "epoch": 0.552942853071113, "grad_norm": 7.416094779968262, "learning_rate": 4.995808879411026e-05, "loss": 0.7645, "num_input_tokens_seen": 9481200, "step": 485 }, { "epoch": 0.5586432948553512, "grad_norm": 6.9029693603515625, "learning_rate": 4.995722043896809e-05, "loss": 0.6875, "num_input_tokens_seen": 9578944, "step": 490 }, { "epoch": 0.5643437366395896, "grad_norm": 7.398702621459961, "learning_rate": 4.995634318783646e-05, "loss": 0.7829, "num_input_tokens_seen": 9676688, "step": 495 }, { "epoch": 0.5700441784238278, "grad_norm": 7.631560802459717, "learning_rate": 4.9955457041028055e-05, "loss": 0.7324, "num_input_tokens_seen": 9774464, "step": 500 }, { "epoch": 0.5757446202080662, "grad_norm": 9.913789749145508, "learning_rate": 4.995456199885875e-05, "loss": 0.7578, "num_input_tokens_seen": 9872160, "step": 505 }, { "epoch": 0.5814450619923044, "grad_norm": 9.40986442565918, "learning_rate": 4.995365806164758e-05, "loss": 0.951, "num_input_tokens_seen": 9969904, "step": 510 }, { "epoch": 0.5871455037765427, "grad_norm": 5.404745578765869, "learning_rate": 4.995274522971675e-05, "loss": 0.7427, "num_input_tokens_seen": 10067648, "step": 515 }, { "epoch": 0.592845945560781, "grad_norm": 6.450439929962158, "learning_rate": 4.9951823503391634e-05, "loss": 0.75, "num_input_tokens_seen": 10165456, "step": 520 }, { "epoch": 0.5985463873450192, "grad_norm": 7.56156587600708, "learning_rate": 4.9950892883000786e-05, "loss": 0.7311, "num_input_tokens_seen": 10263152, "step": 525 }, { "epoch": 0.6042468291292575, "grad_norm": 5.007820129394531, "learning_rate": 4.994995336887593e-05, "loss": 0.7088, "num_input_tokens_seen": 10360848, "step": 530 }, { "epoch": 0.6099472709134958, "grad_norm": 6.651803016662598, "learning_rate": 4.994900496135195e-05, "loss": 0.7473, "num_input_tokens_seen": 10458496, "step": 535 }, { "epoch": 0.6156477126977341, "grad_norm": 4.845729351043701, "learning_rate": 4.9948047660766904e-05, "loss": 0.6939, "num_input_tokens_seen": 10556304, "step": 540 }, { "epoch": 0.6213481544819723, "grad_norm": 7.277071475982666, "learning_rate": 4.994708146746203e-05, "loss": 0.7219, "num_input_tokens_seen": 10654048, "step": 545 }, { "epoch": 0.6270485962662107, "grad_norm": 7.703381061553955, "learning_rate": 4.994610638178172e-05, "loss": 0.7795, "num_input_tokens_seen": 10751776, "step": 550 }, { "epoch": 0.6327490380504489, "grad_norm": 8.279520988464355, "learning_rate": 4.994512240407354e-05, "loss": 0.7027, "num_input_tokens_seen": 10849584, "step": 555 }, { "epoch": 0.6384494798346871, "grad_norm": 10.189576148986816, "learning_rate": 4.9944129534688234e-05, "loss": 0.6917, "num_input_tokens_seen": 10947264, "step": 560 }, { "epoch": 0.6441499216189255, "grad_norm": 6.311273574829102, "learning_rate": 4.994312777397972e-05, "loss": 0.7335, "num_input_tokens_seen": 11045120, "step": 565 }, { "epoch": 0.6498503634031637, "grad_norm": 9.937539100646973, "learning_rate": 4.994211712230504e-05, "loss": 0.6367, "num_input_tokens_seen": 11142864, "step": 570 }, { "epoch": 0.655550805187402, "grad_norm": 9.992775917053223, "learning_rate": 4.994109758002447e-05, "loss": 0.7662, "num_input_tokens_seen": 11240560, "step": 575 }, { "epoch": 0.6612512469716403, "grad_norm": 6.363308429718018, "learning_rate": 4.994006914750143e-05, "loss": 0.7291, "num_input_tokens_seen": 11338320, "step": 580 }, { "epoch": 0.6669516887558786, "grad_norm": 6.920602321624756, "learning_rate": 4.993903182510249e-05, "loss": 0.6525, "num_input_tokens_seen": 11436032, "step": 585 }, { "epoch": 0.6726521305401169, "grad_norm": 6.734442234039307, "learning_rate": 4.99379856131974e-05, "loss": 0.6581, "num_input_tokens_seen": 11533680, "step": 590 }, { "epoch": 0.6783525723243551, "grad_norm": 6.08076810836792, "learning_rate": 4.99369305121591e-05, "loss": 0.6868, "num_input_tokens_seen": 11631344, "step": 595 }, { "epoch": 0.6840530141085934, "grad_norm": 5.305174827575684, "learning_rate": 4.9935866522363665e-05, "loss": 0.7231, "num_input_tokens_seen": 11729104, "step": 600 }, { "epoch": 0.6897534558928317, "grad_norm": 5.337072849273682, "learning_rate": 4.9934793644190345e-05, "loss": 0.7082, "num_input_tokens_seen": 11826880, "step": 605 }, { "epoch": 0.69545389767707, "grad_norm": 6.563253879547119, "learning_rate": 4.993371187802159e-05, "loss": 0.7412, "num_input_tokens_seen": 11924592, "step": 610 }, { "epoch": 0.7011543394613082, "grad_norm": 6.92053747177124, "learning_rate": 4.993262122424298e-05, "loss": 0.6752, "num_input_tokens_seen": 12022256, "step": 615 }, { "epoch": 0.7068547812455466, "grad_norm": 10.413783073425293, "learning_rate": 4.9931521683243276e-05, "loss": 0.6955, "num_input_tokens_seen": 12120000, "step": 620 }, { "epoch": 0.7125552230297848, "grad_norm": 6.970921039581299, "learning_rate": 4.993041325541442e-05, "loss": 0.6883, "num_input_tokens_seen": 12217808, "step": 625 }, { "epoch": 0.718255664814023, "grad_norm": 5.135336875915527, "learning_rate": 4.992929594115151e-05, "loss": 0.6039, "num_input_tokens_seen": 12315616, "step": 630 }, { "epoch": 0.7239561065982614, "grad_norm": 7.350869655609131, "learning_rate": 4.99281697408528e-05, "loss": 0.7195, "num_input_tokens_seen": 12413376, "step": 635 }, { "epoch": 0.7296565483824996, "grad_norm": 6.427408218383789, "learning_rate": 4.992703465491974e-05, "loss": 0.5395, "num_input_tokens_seen": 12510960, "step": 640 }, { "epoch": 0.735356990166738, "grad_norm": 7.422171592712402, "learning_rate": 4.992589068375691e-05, "loss": 0.5605, "num_input_tokens_seen": 12608752, "step": 645 }, { "epoch": 0.7410574319509762, "grad_norm": 10.039104461669922, "learning_rate": 4.9924737827772104e-05, "loss": 0.6171, "num_input_tokens_seen": 12706448, "step": 650 }, { "epoch": 0.7467578737352145, "grad_norm": 6.769627094268799, "learning_rate": 4.992357608737623e-05, "loss": 0.6656, "num_input_tokens_seen": 12804144, "step": 655 }, { "epoch": 0.7524583155194527, "grad_norm": 6.161548614501953, "learning_rate": 4.992240546298341e-05, "loss": 0.6412, "num_input_tokens_seen": 12902000, "step": 660 }, { "epoch": 0.7581587573036911, "grad_norm": 9.02010440826416, "learning_rate": 4.9921225955010906e-05, "loss": 0.6899, "num_input_tokens_seen": 12999648, "step": 665 }, { "epoch": 0.7638591990879293, "grad_norm": 5.683040618896484, "learning_rate": 4.9920037563879155e-05, "loss": 0.5788, "num_input_tokens_seen": 13097424, "step": 670 }, { "epoch": 0.7695596408721675, "grad_norm": 5.576777935028076, "learning_rate": 4.9918840290011745e-05, "loss": 0.6354, "num_input_tokens_seen": 13195136, "step": 675 }, { "epoch": 0.7752600826564059, "grad_norm": 6.484269142150879, "learning_rate": 4.9917634133835466e-05, "loss": 0.6004, "num_input_tokens_seen": 13292912, "step": 680 }, { "epoch": 0.7809605244406441, "grad_norm": 5.845834732055664, "learning_rate": 4.991641909578023e-05, "loss": 0.6065, "num_input_tokens_seen": 13390560, "step": 685 }, { "epoch": 0.7866609662248825, "grad_norm": 7.066195011138916, "learning_rate": 4.9915195176279156e-05, "loss": 0.703, "num_input_tokens_seen": 13488304, "step": 690 }, { "epoch": 0.7923614080091207, "grad_norm": 7.687030792236328, "learning_rate": 4.9913962375768494e-05, "loss": 0.5684, "num_input_tokens_seen": 13586032, "step": 695 }, { "epoch": 0.798061849793359, "grad_norm": 5.923397064208984, "learning_rate": 4.9912720694687684e-05, "loss": 0.7124, "num_input_tokens_seen": 13683792, "step": 700 }, { "epoch": 0.8037622915775973, "grad_norm": 7.307689666748047, "learning_rate": 4.9911470133479324e-05, "loss": 0.585, "num_input_tokens_seen": 13781488, "step": 705 }, { "epoch": 0.8094627333618355, "grad_norm": 5.22707462310791, "learning_rate": 4.9910210692589164e-05, "loss": 0.6301, "num_input_tokens_seen": 13879264, "step": 710 }, { "epoch": 0.8151631751460738, "grad_norm": 6.6996870040893555, "learning_rate": 4.990894237246615e-05, "loss": 0.6073, "num_input_tokens_seen": 13976976, "step": 715 }, { "epoch": 0.8208636169303121, "grad_norm": 9.039154052734375, "learning_rate": 4.990766517356236e-05, "loss": 0.6611, "num_input_tokens_seen": 14074688, "step": 720 }, { "epoch": 0.8265640587145504, "grad_norm": 3.3328487873077393, "learning_rate": 4.9906379096333047e-05, "loss": 0.5829, "num_input_tokens_seen": 14172432, "step": 725 }, { "epoch": 0.8322645004987886, "grad_norm": 9.260608673095703, "learning_rate": 4.9905084141236646e-05, "loss": 0.7311, "num_input_tokens_seen": 14270112, "step": 730 }, { "epoch": 0.837964942283027, "grad_norm": 7.124883651733398, "learning_rate": 4.990378030873474e-05, "loss": 0.6354, "num_input_tokens_seen": 14367792, "step": 735 }, { "epoch": 0.8436653840672652, "grad_norm": 5.522550106048584, "learning_rate": 4.990246759929207e-05, "loss": 0.5578, "num_input_tokens_seen": 14465584, "step": 740 }, { "epoch": 0.8493658258515034, "grad_norm": 8.491950035095215, "learning_rate": 4.9901146013376556e-05, "loss": 0.6489, "num_input_tokens_seen": 14563344, "step": 745 }, { "epoch": 0.8550662676357418, "grad_norm": 6.821796417236328, "learning_rate": 4.989981555145928e-05, "loss": 0.451, "num_input_tokens_seen": 14661024, "step": 750 }, { "epoch": 0.86076670941998, "grad_norm": 8.655888557434082, "learning_rate": 4.9898476214014486e-05, "loss": 0.6291, "num_input_tokens_seen": 14758800, "step": 755 }, { "epoch": 0.8664671512042184, "grad_norm": 8.26075267791748, "learning_rate": 4.989712800151958e-05, "loss": 0.7259, "num_input_tokens_seen": 14856592, "step": 760 }, { "epoch": 0.8721675929884566, "grad_norm": 6.850794315338135, "learning_rate": 4.989577091445512e-05, "loss": 0.5639, "num_input_tokens_seen": 14954304, "step": 765 }, { "epoch": 0.8778680347726949, "grad_norm": 9.18870735168457, "learning_rate": 4.989440495330485e-05, "loss": 0.616, "num_input_tokens_seen": 15052016, "step": 770 }, { "epoch": 0.8835684765569332, "grad_norm": 9.08046817779541, "learning_rate": 4.989303011855567e-05, "loss": 0.5797, "num_input_tokens_seen": 15149664, "step": 775 }, { "epoch": 0.8892689183411714, "grad_norm": 5.607428073883057, "learning_rate": 4.989164641069763e-05, "loss": 0.5893, "num_input_tokens_seen": 15247360, "step": 780 }, { "epoch": 0.8949693601254097, "grad_norm": 6.935970783233643, "learning_rate": 4.9890253830223955e-05, "loss": 0.6095, "num_input_tokens_seen": 15345056, "step": 785 }, { "epoch": 0.900669801909648, "grad_norm": 6.799474239349365, "learning_rate": 4.988885237763102e-05, "loss": 0.5044, "num_input_tokens_seen": 15442752, "step": 790 }, { "epoch": 0.9063702436938863, "grad_norm": 6.294219017028809, "learning_rate": 4.98874420534184e-05, "loss": 0.5584, "num_input_tokens_seen": 15540464, "step": 795 }, { "epoch": 0.9120706854781245, "grad_norm": 5.488597869873047, "learning_rate": 4.988602285808877e-05, "loss": 0.4862, "num_input_tokens_seen": 15638128, "step": 800 }, { "epoch": 0.9177711272623629, "grad_norm": 8.307422637939453, "learning_rate": 4.988459479214802e-05, "loss": 0.5815, "num_input_tokens_seen": 15735872, "step": 805 }, { "epoch": 0.9234715690466011, "grad_norm": 10.344627380371094, "learning_rate": 4.988315785610519e-05, "loss": 0.5963, "num_input_tokens_seen": 15833680, "step": 810 }, { "epoch": 0.9291720108308394, "grad_norm": 10.354679107666016, "learning_rate": 4.9881712050472464e-05, "loss": 0.6225, "num_input_tokens_seen": 15931472, "step": 815 }, { "epoch": 0.9348724526150777, "grad_norm": 7.605050086975098, "learning_rate": 4.9880257375765194e-05, "loss": 0.645, "num_input_tokens_seen": 16029120, "step": 820 }, { "epoch": 0.9405728943993159, "grad_norm": 5.717419624328613, "learning_rate": 4.987879383250191e-05, "loss": 0.5142, "num_input_tokens_seen": 16126896, "step": 825 }, { "epoch": 0.9462733361835542, "grad_norm": 7.159694194793701, "learning_rate": 4.987732142120428e-05, "loss": 0.6613, "num_input_tokens_seen": 16224592, "step": 830 }, { "epoch": 0.9519737779677925, "grad_norm": 7.166426658630371, "learning_rate": 4.987584014239716e-05, "loss": 0.6094, "num_input_tokens_seen": 16322208, "step": 835 }, { "epoch": 0.9576742197520308, "grad_norm": 7.844811916351318, "learning_rate": 4.9874349996608536e-05, "loss": 0.5613, "num_input_tokens_seen": 16419904, "step": 840 }, { "epoch": 0.963374661536269, "grad_norm": 5.295498371124268, "learning_rate": 4.987285098436958e-05, "loss": 0.4958, "num_input_tokens_seen": 16517600, "step": 845 }, { "epoch": 0.9690751033205074, "grad_norm": 5.007256984710693, "learning_rate": 4.987134310621461e-05, "loss": 0.5119, "num_input_tokens_seen": 16615216, "step": 850 }, { "epoch": 0.9747755451047456, "grad_norm": 7.532383918762207, "learning_rate": 4.9869826362681096e-05, "loss": 0.4567, "num_input_tokens_seen": 16713040, "step": 855 }, { "epoch": 0.9804759868889839, "grad_norm": 6.256499767303467, "learning_rate": 4.9868300754309706e-05, "loss": 0.5088, "num_input_tokens_seen": 16810768, "step": 860 }, { "epoch": 0.9861764286732222, "grad_norm": 6.756839752197266, "learning_rate": 4.986676628164423e-05, "loss": 0.4097, "num_input_tokens_seen": 16908512, "step": 865 }, { "epoch": 0.9918768704574604, "grad_norm": 6.562160491943359, "learning_rate": 4.986522294523162e-05, "loss": 0.3819, "num_input_tokens_seen": 17006240, "step": 870 }, { "epoch": 0.9975773122416988, "grad_norm": 7.496212959289551, "learning_rate": 4.9863670745622015e-05, "loss": 0.4956, "num_input_tokens_seen": 17104000, "step": 875 }, { "epoch": 1.0022801767136953, "grad_norm": 9.028531074523926, "learning_rate": 4.986210968336868e-05, "loss": 0.5872, "num_input_tokens_seen": 17184592, "step": 880 }, { "epoch": 1.0079806184979336, "grad_norm": 6.04398775100708, "learning_rate": 4.986053975902807e-05, "loss": 0.48, "num_input_tokens_seen": 17282304, "step": 885 }, { "epoch": 1.013681060282172, "grad_norm": 9.602685928344727, "learning_rate": 4.985896097315977e-05, "loss": 0.5309, "num_input_tokens_seen": 17380080, "step": 890 }, { "epoch": 1.01938150206641, "grad_norm": 6.881324291229248, "learning_rate": 4.9857373326326545e-05, "loss": 0.5103, "num_input_tokens_seen": 17477760, "step": 895 }, { "epoch": 1.0250819438506484, "grad_norm": 8.762344360351562, "learning_rate": 4.985577681909431e-05, "loss": 0.5336, "num_input_tokens_seen": 17575456, "step": 900 }, { "epoch": 1.0307823856348868, "grad_norm": 7.028131484985352, "learning_rate": 4.985417145203214e-05, "loss": 0.4887, "num_input_tokens_seen": 17673184, "step": 905 }, { "epoch": 1.036482827419125, "grad_norm": 6.512467861175537, "learning_rate": 4.985255722571227e-05, "loss": 0.4787, "num_input_tokens_seen": 17770944, "step": 910 }, { "epoch": 1.0421832692033632, "grad_norm": 6.597855567932129, "learning_rate": 4.985093414071008e-05, "loss": 0.5185, "num_input_tokens_seen": 17868768, "step": 915 }, { "epoch": 1.0478837109876016, "grad_norm": 7.660828590393066, "learning_rate": 4.984930219760413e-05, "loss": 0.5056, "num_input_tokens_seen": 17966480, "step": 920 }, { "epoch": 1.05358415277184, "grad_norm": 6.880988121032715, "learning_rate": 4.984766139697611e-05, "loss": 0.5371, "num_input_tokens_seen": 18064336, "step": 925 }, { "epoch": 1.059284594556078, "grad_norm": 5.854323863983154, "learning_rate": 4.98460117394109e-05, "loss": 0.5041, "num_input_tokens_seen": 18162112, "step": 930 }, { "epoch": 1.0649850363403164, "grad_norm": 5.215938568115234, "learning_rate": 4.984435322549651e-05, "loss": 0.4857, "num_input_tokens_seen": 18259904, "step": 935 }, { "epoch": 1.0706854781245547, "grad_norm": 5.902091026306152, "learning_rate": 4.984268585582412e-05, "loss": 0.5047, "num_input_tokens_seen": 18357616, "step": 940 }, { "epoch": 1.0763859199087928, "grad_norm": 7.6616411209106445, "learning_rate": 4.9841009630988064e-05, "loss": 0.4147, "num_input_tokens_seen": 18455392, "step": 945 }, { "epoch": 1.0820863616930312, "grad_norm": 7.779905796051025, "learning_rate": 4.983932455158583e-05, "loss": 0.4762, "num_input_tokens_seen": 18553120, "step": 950 }, { "epoch": 1.0877868034772695, "grad_norm": 6.425886154174805, "learning_rate": 4.9837630618218056e-05, "loss": 0.4129, "num_input_tokens_seen": 18650864, "step": 955 }, { "epoch": 1.0934872452615079, "grad_norm": 8.044005393981934, "learning_rate": 4.983592783148856e-05, "loss": 0.4027, "num_input_tokens_seen": 18748624, "step": 960 }, { "epoch": 1.099187687045746, "grad_norm": 5.91091251373291, "learning_rate": 4.983421619200428e-05, "loss": 0.4064, "num_input_tokens_seen": 18846320, "step": 965 }, { "epoch": 1.1048881288299843, "grad_norm": 5.88447380065918, "learning_rate": 4.9832495700375346e-05, "loss": 0.4599, "num_input_tokens_seen": 18944144, "step": 970 }, { "epoch": 1.1105885706142227, "grad_norm": 7.686187744140625, "learning_rate": 4.983076635721502e-05, "loss": 0.4764, "num_input_tokens_seen": 19041904, "step": 975 }, { "epoch": 1.1162890123984608, "grad_norm": 6.358628273010254, "learning_rate": 4.982902816313972e-05, "loss": 0.4844, "num_input_tokens_seen": 19139664, "step": 980 }, { "epoch": 1.1219894541826991, "grad_norm": 6.775269508361816, "learning_rate": 4.982728111876903e-05, "loss": 0.4292, "num_input_tokens_seen": 19237488, "step": 985 }, { "epoch": 1.1276898959669375, "grad_norm": 7.491086483001709, "learning_rate": 4.982552522472569e-05, "loss": 0.4423, "num_input_tokens_seen": 19335152, "step": 990 }, { "epoch": 1.1333903377511758, "grad_norm": 7.664567947387695, "learning_rate": 4.982376048163557e-05, "loss": 0.4983, "num_input_tokens_seen": 19432976, "step": 995 }, { "epoch": 1.139090779535414, "grad_norm": 4.956116676330566, "learning_rate": 4.9821986890127734e-05, "loss": 0.4027, "num_input_tokens_seen": 19530704, "step": 1000 }, { "epoch": 1.1447912213196523, "grad_norm": 7.820266246795654, "learning_rate": 4.982020445083436e-05, "loss": 0.4131, "num_input_tokens_seen": 19628448, "step": 1005 }, { "epoch": 1.1504916631038906, "grad_norm": 5.402007102966309, "learning_rate": 4.981841316439081e-05, "loss": 0.4946, "num_input_tokens_seen": 19726176, "step": 1010 }, { "epoch": 1.1561921048881287, "grad_norm": 6.406668663024902, "learning_rate": 4.981661303143557e-05, "loss": 0.4701, "num_input_tokens_seen": 19823856, "step": 1015 }, { "epoch": 1.161892546672367, "grad_norm": 3.809847116470337, "learning_rate": 4.981480405261032e-05, "loss": 0.4063, "num_input_tokens_seen": 19921552, "step": 1020 }, { "epoch": 1.1675929884566054, "grad_norm": 8.321266174316406, "learning_rate": 4.981298622855984e-05, "loss": 0.38, "num_input_tokens_seen": 20019248, "step": 1025 }, { "epoch": 1.1732934302408438, "grad_norm": 4.611199855804443, "learning_rate": 4.981115955993213e-05, "loss": 0.3435, "num_input_tokens_seen": 20116992, "step": 1030 }, { "epoch": 1.1789938720250819, "grad_norm": 6.748137950897217, "learning_rate": 4.980932404737827e-05, "loss": 0.4443, "num_input_tokens_seen": 20214848, "step": 1035 }, { "epoch": 1.1846943138093202, "grad_norm": 7.327335834503174, "learning_rate": 4.980747969155255e-05, "loss": 0.5365, "num_input_tokens_seen": 20312608, "step": 1040 }, { "epoch": 1.1903947555935586, "grad_norm": 9.424795150756836, "learning_rate": 4.980562649311238e-05, "loss": 0.404, "num_input_tokens_seen": 20410288, "step": 1045 }, { "epoch": 1.196095197377797, "grad_norm": 6.2012152671813965, "learning_rate": 4.9803764452718335e-05, "loss": 0.4176, "num_input_tokens_seen": 20508080, "step": 1050 }, { "epoch": 1.201795639162035, "grad_norm": 6.23061990737915, "learning_rate": 4.980189357103414e-05, "loss": 0.3945, "num_input_tokens_seen": 20605856, "step": 1055 }, { "epoch": 1.2074960809462734, "grad_norm": 8.05282974243164, "learning_rate": 4.980001384872666e-05, "loss": 0.5353, "num_input_tokens_seen": 20703584, "step": 1060 }, { "epoch": 1.2131965227305117, "grad_norm": 7.0456156730651855, "learning_rate": 4.9798125286465935e-05, "loss": 0.4638, "num_input_tokens_seen": 20801376, "step": 1065 }, { "epoch": 1.2188969645147498, "grad_norm": 5.285283088684082, "learning_rate": 4.979622788492513e-05, "loss": 0.5492, "num_input_tokens_seen": 20899200, "step": 1070 }, { "epoch": 1.2245974062989882, "grad_norm": 7.358059883117676, "learning_rate": 4.9794321644780585e-05, "loss": 0.4979, "num_input_tokens_seen": 20996928, "step": 1075 }, { "epoch": 1.2302978480832265, "grad_norm": 6.339309215545654, "learning_rate": 4.979240656671177e-05, "loss": 0.3867, "num_input_tokens_seen": 21094752, "step": 1080 }, { "epoch": 1.2359982898674646, "grad_norm": 6.887006759643555, "learning_rate": 4.979048265140132e-05, "loss": 0.338, "num_input_tokens_seen": 21192480, "step": 1085 }, { "epoch": 1.241698731651703, "grad_norm": 7.377925395965576, "learning_rate": 4.9788549899535e-05, "loss": 0.3946, "num_input_tokens_seen": 21290144, "step": 1090 }, { "epoch": 1.2473991734359413, "grad_norm": 7.47123384475708, "learning_rate": 4.978660831180175e-05, "loss": 0.4831, "num_input_tokens_seen": 21387888, "step": 1095 }, { "epoch": 1.2530996152201794, "grad_norm": 9.348438262939453, "learning_rate": 4.978465788889365e-05, "loss": 0.4933, "num_input_tokens_seen": 21485536, "step": 1100 }, { "epoch": 1.2588000570044178, "grad_norm": 8.355619430541992, "learning_rate": 4.978269863150592e-05, "loss": 0.4139, "num_input_tokens_seen": 21583264, "step": 1105 }, { "epoch": 1.2645004987886561, "grad_norm": 5.923043727874756, "learning_rate": 4.978073054033694e-05, "loss": 0.3656, "num_input_tokens_seen": 21681040, "step": 1110 }, { "epoch": 1.2702009405728945, "grad_norm": 7.194669246673584, "learning_rate": 4.977875361608823e-05, "loss": 0.3487, "num_input_tokens_seen": 21778720, "step": 1115 }, { "epoch": 1.2759013823571328, "grad_norm": 7.351987838745117, "learning_rate": 4.9776767859464474e-05, "loss": 0.4004, "num_input_tokens_seen": 21876496, "step": 1120 }, { "epoch": 1.281601824141371, "grad_norm": 6.509387493133545, "learning_rate": 4.9774773271173494e-05, "loss": 0.3702, "num_input_tokens_seen": 21974256, "step": 1125 }, { "epoch": 1.2873022659256093, "grad_norm": 10.99763298034668, "learning_rate": 4.977276985192624e-05, "loss": 0.3921, "num_input_tokens_seen": 22071952, "step": 1130 }, { "epoch": 1.2930027077098476, "grad_norm": 5.566549777984619, "learning_rate": 4.977075760243686e-05, "loss": 0.4117, "num_input_tokens_seen": 22169696, "step": 1135 }, { "epoch": 1.2987031494940857, "grad_norm": 8.485737800598145, "learning_rate": 4.976873652342259e-05, "loss": 0.394, "num_input_tokens_seen": 22267456, "step": 1140 }, { "epoch": 1.304403591278324, "grad_norm": 7.3099284172058105, "learning_rate": 4.976670661560386e-05, "loss": 0.2883, "num_input_tokens_seen": 22365120, "step": 1145 }, { "epoch": 1.3101040330625624, "grad_norm": 6.294934272766113, "learning_rate": 4.976466787970423e-05, "loss": 0.3503, "num_input_tokens_seen": 22462880, "step": 1150 }, { "epoch": 1.3158044748468005, "grad_norm": 5.884027004241943, "learning_rate": 4.97626203164504e-05, "loss": 0.3098, "num_input_tokens_seen": 22560640, "step": 1155 }, { "epoch": 1.3215049166310389, "grad_norm": 7.804978847503662, "learning_rate": 4.9760563926572226e-05, "loss": 0.3423, "num_input_tokens_seen": 22658368, "step": 1160 }, { "epoch": 1.3272053584152772, "grad_norm": 7.155725002288818, "learning_rate": 4.97584987108027e-05, "loss": 0.3006, "num_input_tokens_seen": 22756176, "step": 1165 }, { "epoch": 1.3329058001995155, "grad_norm": 6.071112632751465, "learning_rate": 4.975642466987799e-05, "loss": 0.3357, "num_input_tokens_seen": 22853920, "step": 1170 }, { "epoch": 1.3386062419837537, "grad_norm": 5.568732738494873, "learning_rate": 4.9754341804537356e-05, "loss": 0.3445, "num_input_tokens_seen": 22951664, "step": 1175 }, { "epoch": 1.344306683767992, "grad_norm": 9.902073860168457, "learning_rate": 4.975225011552326e-05, "loss": 0.3621, "num_input_tokens_seen": 23049520, "step": 1180 }, { "epoch": 1.3500071255522303, "grad_norm": 5.503910064697266, "learning_rate": 4.975014960358126e-05, "loss": 0.3229, "num_input_tokens_seen": 23147280, "step": 1185 }, { "epoch": 1.3557075673364687, "grad_norm": 7.572802543640137, "learning_rate": 4.974804026946011e-05, "loss": 0.5356, "num_input_tokens_seen": 23245008, "step": 1190 }, { "epoch": 1.3614080091207068, "grad_norm": 7.335933208465576, "learning_rate": 4.9745922113911655e-05, "loss": 0.364, "num_input_tokens_seen": 23342768, "step": 1195 }, { "epoch": 1.3671084509049451, "grad_norm": 10.062085151672363, "learning_rate": 4.974379513769093e-05, "loss": 0.384, "num_input_tokens_seen": 23440480, "step": 1200 }, { "epoch": 1.3728088926891835, "grad_norm": 10.50871467590332, "learning_rate": 4.974165934155608e-05, "loss": 0.357, "num_input_tokens_seen": 23538192, "step": 1205 }, { "epoch": 1.3785093344734216, "grad_norm": 6.676635265350342, "learning_rate": 4.9739514726268416e-05, "loss": 0.316, "num_input_tokens_seen": 23635984, "step": 1210 }, { "epoch": 1.38420977625766, "grad_norm": 9.456487655639648, "learning_rate": 4.973736129259239e-05, "loss": 0.3407, "num_input_tokens_seen": 23733744, "step": 1215 }, { "epoch": 1.3899102180418983, "grad_norm": 7.790709495544434, "learning_rate": 4.9735199041295575e-05, "loss": 0.422, "num_input_tokens_seen": 23831440, "step": 1220 }, { "epoch": 1.3956106598261364, "grad_norm": 9.099756240844727, "learning_rate": 4.9733027973148727e-05, "loss": 0.4655, "num_input_tokens_seen": 23929184, "step": 1225 }, { "epoch": 1.4013111016103748, "grad_norm": 6.994203567504883, "learning_rate": 4.9730848088925706e-05, "loss": 0.388, "num_input_tokens_seen": 24026928, "step": 1230 }, { "epoch": 1.407011543394613, "grad_norm": 7.203540325164795, "learning_rate": 4.9728659389403535e-05, "loss": 0.4004, "num_input_tokens_seen": 24124688, "step": 1235 }, { "epoch": 1.4127119851788514, "grad_norm": 7.220198631286621, "learning_rate": 4.9726461875362377e-05, "loss": 0.3321, "num_input_tokens_seen": 24222416, "step": 1240 }, { "epoch": 1.4184124269630896, "grad_norm": 6.651162147521973, "learning_rate": 4.9724255547585534e-05, "loss": 0.2864, "num_input_tokens_seen": 24320096, "step": 1245 }, { "epoch": 1.424112868747328, "grad_norm": 7.986251354217529, "learning_rate": 4.9722040406859454e-05, "loss": 0.3401, "num_input_tokens_seen": 24417712, "step": 1250 }, { "epoch": 1.4298133105315662, "grad_norm": 6.927532196044922, "learning_rate": 4.971981645397371e-05, "loss": 0.344, "num_input_tokens_seen": 24515456, "step": 1255 }, { "epoch": 1.4355137523158046, "grad_norm": 8.963294982910156, "learning_rate": 4.9717583689721046e-05, "loss": 0.3394, "num_input_tokens_seen": 24613232, "step": 1260 }, { "epoch": 1.4412141941000427, "grad_norm": 9.106192588806152, "learning_rate": 4.9715342114897325e-05, "loss": 0.4323, "num_input_tokens_seen": 24710960, "step": 1265 }, { "epoch": 1.446914635884281, "grad_norm": 8.095370292663574, "learning_rate": 4.971309173030154e-05, "loss": 0.3961, "num_input_tokens_seen": 24808560, "step": 1270 }, { "epoch": 1.4526150776685194, "grad_norm": 7.318641662597656, "learning_rate": 4.9710832536735864e-05, "loss": 0.2917, "num_input_tokens_seen": 24906320, "step": 1275 }, { "epoch": 1.4583155194527575, "grad_norm": 7.140157699584961, "learning_rate": 4.970856453500557e-05, "loss": 0.3622, "num_input_tokens_seen": 25004016, "step": 1280 }, { "epoch": 1.4640159612369958, "grad_norm": 8.635784149169922, "learning_rate": 4.970628772591909e-05, "loss": 0.4472, "num_input_tokens_seen": 25101808, "step": 1285 }, { "epoch": 1.4697164030212342, "grad_norm": 9.65007495880127, "learning_rate": 4.970400211028798e-05, "loss": 0.3185, "num_input_tokens_seen": 25199568, "step": 1290 }, { "epoch": 1.4754168448054723, "grad_norm": 6.725943088531494, "learning_rate": 4.970170768892697e-05, "loss": 0.4232, "num_input_tokens_seen": 25297296, "step": 1295 }, { "epoch": 1.4811172865897106, "grad_norm": 7.719542980194092, "learning_rate": 4.9699404462653887e-05, "loss": 0.3133, "num_input_tokens_seen": 25395056, "step": 1300 }, { "epoch": 1.486817728373949, "grad_norm": 7.277915954589844, "learning_rate": 4.969709243228972e-05, "loss": 0.3103, "num_input_tokens_seen": 25492784, "step": 1305 }, { "epoch": 1.4925181701581873, "grad_norm": 6.372420310974121, "learning_rate": 4.96947715986586e-05, "loss": 0.3191, "num_input_tokens_seen": 25590528, "step": 1310 }, { "epoch": 1.4982186119424257, "grad_norm": 5.824290752410889, "learning_rate": 4.969244196258777e-05, "loss": 0.2663, "num_input_tokens_seen": 25688304, "step": 1315 }, { "epoch": 1.5039190537266638, "grad_norm": 5.841516971588135, "learning_rate": 4.969010352490764e-05, "loss": 0.3178, "num_input_tokens_seen": 25786096, "step": 1320 }, { "epoch": 1.5096194955109021, "grad_norm": 7.10455846786499, "learning_rate": 4.968775628645174e-05, "loss": 0.4365, "num_input_tokens_seen": 25883776, "step": 1325 }, { "epoch": 1.5153199372951405, "grad_norm": 6.748999118804932, "learning_rate": 4.9685400248056747e-05, "loss": 0.2147, "num_input_tokens_seen": 25981552, "step": 1330 }, { "epoch": 1.5210203790793786, "grad_norm": 4.732318878173828, "learning_rate": 4.968303541056246e-05, "loss": 0.3367, "num_input_tokens_seen": 26079312, "step": 1335 }, { "epoch": 1.526720820863617, "grad_norm": 4.386244297027588, "learning_rate": 4.9680661774811835e-05, "loss": 0.3207, "num_input_tokens_seen": 26177136, "step": 1340 }, { "epoch": 1.5324212626478553, "grad_norm": 9.144301414489746, "learning_rate": 4.967827934165095e-05, "loss": 0.2718, "num_input_tokens_seen": 26274944, "step": 1345 }, { "epoch": 1.5381217044320934, "grad_norm": 10.099781036376953, "learning_rate": 4.967588811192902e-05, "loss": 0.3752, "num_input_tokens_seen": 26372768, "step": 1350 }, { "epoch": 1.5438221462163317, "grad_norm": 7.383282661437988, "learning_rate": 4.96734880864984e-05, "loss": 0.2743, "num_input_tokens_seen": 26470608, "step": 1355 }, { "epoch": 1.54952258800057, "grad_norm": 8.287309646606445, "learning_rate": 4.967107926621457e-05, "loss": 0.2853, "num_input_tokens_seen": 26568368, "step": 1360 }, { "epoch": 1.5552230297848082, "grad_norm": 9.22373104095459, "learning_rate": 4.966866165193617e-05, "loss": 0.2913, "num_input_tokens_seen": 26666080, "step": 1365 }, { "epoch": 1.5609234715690468, "grad_norm": 6.172590255737305, "learning_rate": 4.966623524452494e-05, "loss": 0.2775, "num_input_tokens_seen": 26763792, "step": 1370 }, { "epoch": 1.5666239133532849, "grad_norm": 7.464315891265869, "learning_rate": 4.9663800044845784e-05, "loss": 0.3685, "num_input_tokens_seen": 26861488, "step": 1375 }, { "epoch": 1.572324355137523, "grad_norm": 7.123498916625977, "learning_rate": 4.9661356053766716e-05, "loss": 0.3636, "num_input_tokens_seen": 26959232, "step": 1380 }, { "epoch": 1.5780247969217616, "grad_norm": 7.7689619064331055, "learning_rate": 4.965890327215891e-05, "loss": 0.3052, "num_input_tokens_seen": 27057040, "step": 1385 }, { "epoch": 1.5837252387059997, "grad_norm": 5.988576412200928, "learning_rate": 4.965644170089665e-05, "loss": 0.3355, "num_input_tokens_seen": 27154768, "step": 1390 }, { "epoch": 1.589425680490238, "grad_norm": 9.67150592803955, "learning_rate": 4.965397134085735e-05, "loss": 0.3239, "num_input_tokens_seen": 27252480, "step": 1395 }, { "epoch": 1.5951261222744764, "grad_norm": 9.156477928161621, "learning_rate": 4.96514921929216e-05, "loss": 0.3421, "num_input_tokens_seen": 27350320, "step": 1400 }, { "epoch": 1.6008265640587145, "grad_norm": 8.114056587219238, "learning_rate": 4.964900425797306e-05, "loss": 0.405, "num_input_tokens_seen": 27448128, "step": 1405 }, { "epoch": 1.6065270058429528, "grad_norm": 8.292765617370605, "learning_rate": 4.9646507536898575e-05, "loss": 0.2936, "num_input_tokens_seen": 27545808, "step": 1410 }, { "epoch": 1.6122274476271912, "grad_norm": 6.832367420196533, "learning_rate": 4.964400203058809e-05, "loss": 0.2365, "num_input_tokens_seen": 27643456, "step": 1415 }, { "epoch": 1.6179278894114293, "grad_norm": 8.706232070922852, "learning_rate": 4.9641487739934684e-05, "loss": 0.3065, "num_input_tokens_seen": 27741168, "step": 1420 }, { "epoch": 1.6236283311956676, "grad_norm": 7.025177478790283, "learning_rate": 4.963896466583459e-05, "loss": 0.2376, "num_input_tokens_seen": 27838912, "step": 1425 }, { "epoch": 1.629328772979906, "grad_norm": 9.140198707580566, "learning_rate": 4.963643280918714e-05, "loss": 0.2518, "num_input_tokens_seen": 27936592, "step": 1430 }, { "epoch": 1.635029214764144, "grad_norm": 8.371499061584473, "learning_rate": 4.963389217089484e-05, "loss": 0.3488, "num_input_tokens_seen": 28034304, "step": 1435 }, { "epoch": 1.6407296565483827, "grad_norm": 4.390028476715088, "learning_rate": 4.963134275186327e-05, "loss": 0.2444, "num_input_tokens_seen": 28131984, "step": 1440 }, { "epoch": 1.6464300983326208, "grad_norm": 10.252416610717773, "learning_rate": 4.9628784553001185e-05, "loss": 0.3859, "num_input_tokens_seen": 28229680, "step": 1445 }, { "epoch": 1.652130540116859, "grad_norm": 8.285685539245605, "learning_rate": 4.962621757522044e-05, "loss": 0.3006, "num_input_tokens_seen": 28327440, "step": 1450 }, { "epoch": 1.6578309819010975, "grad_norm": 8.214794158935547, "learning_rate": 4.962364181943606e-05, "loss": 0.2718, "num_input_tokens_seen": 28425216, "step": 1455 }, { "epoch": 1.6635314236853356, "grad_norm": 12.243093490600586, "learning_rate": 4.9621057286566155e-05, "loss": 0.3569, "num_input_tokens_seen": 28522992, "step": 1460 }, { "epoch": 1.669231865469574, "grad_norm": 5.961976528167725, "learning_rate": 4.961846397753197e-05, "loss": 0.2414, "num_input_tokens_seen": 28620720, "step": 1465 }, { "epoch": 1.6749323072538123, "grad_norm": 7.293554782867432, "learning_rate": 4.961586189325791e-05, "loss": 0.2259, "num_input_tokens_seen": 28718464, "step": 1470 }, { "epoch": 1.6806327490380504, "grad_norm": 7.203713893890381, "learning_rate": 4.9613251034671465e-05, "loss": 0.2356, "num_input_tokens_seen": 28816368, "step": 1475 }, { "epoch": 1.6863331908222887, "grad_norm": 6.28717565536499, "learning_rate": 4.961063140270329e-05, "loss": 0.3129, "num_input_tokens_seen": 28914080, "step": 1480 }, { "epoch": 1.692033632606527, "grad_norm": 5.234409809112549, "learning_rate": 4.960800299828715e-05, "loss": 0.2614, "num_input_tokens_seen": 29011808, "step": 1485 }, { "epoch": 1.6977340743907652, "grad_norm": 7.488391399383545, "learning_rate": 4.960536582235993e-05, "loss": 0.2573, "num_input_tokens_seen": 29109488, "step": 1490 }, { "epoch": 1.7034345161750035, "grad_norm": 7.9980292320251465, "learning_rate": 4.960271987586166e-05, "loss": 0.2409, "num_input_tokens_seen": 29207232, "step": 1495 }, { "epoch": 1.7091349579592419, "grad_norm": 4.908660888671875, "learning_rate": 4.960006515973548e-05, "loss": 0.2969, "num_input_tokens_seen": 29304960, "step": 1500 }, { "epoch": 1.71483539974348, "grad_norm": 7.019242763519287, "learning_rate": 4.959740167492767e-05, "loss": 0.2576, "num_input_tokens_seen": 29402720, "step": 1505 }, { "epoch": 1.7205358415277185, "grad_norm": 5.726184844970703, "learning_rate": 4.959472942238762e-05, "loss": 0.2731, "num_input_tokens_seen": 29500480, "step": 1510 }, { "epoch": 1.7262362833119567, "grad_norm": 7.595127105712891, "learning_rate": 4.9592048403067845e-05, "loss": 0.3502, "num_input_tokens_seen": 29598240, "step": 1515 }, { "epoch": 1.731936725096195, "grad_norm": 5.473161697387695, "learning_rate": 4.958935861792402e-05, "loss": 0.3446, "num_input_tokens_seen": 29695952, "step": 1520 }, { "epoch": 1.7376371668804333, "grad_norm": 5.171034812927246, "learning_rate": 4.958666006791489e-05, "loss": 0.328, "num_input_tokens_seen": 29793696, "step": 1525 }, { "epoch": 1.7433376086646715, "grad_norm": 7.3467559814453125, "learning_rate": 4.958395275400237e-05, "loss": 0.2313, "num_input_tokens_seen": 29891456, "step": 1530 }, { "epoch": 1.7490380504489098, "grad_norm": 7.6113481521606445, "learning_rate": 4.958123667715147e-05, "loss": 0.3182, "num_input_tokens_seen": 29989280, "step": 1535 }, { "epoch": 1.7547384922331482, "grad_norm": 4.429864883422852, "learning_rate": 4.957851183833034e-05, "loss": 0.2573, "num_input_tokens_seen": 30087104, "step": 1540 }, { "epoch": 1.7604389340173863, "grad_norm": 7.2398200035095215, "learning_rate": 4.957577823851024e-05, "loss": 0.3694, "num_input_tokens_seen": 30184768, "step": 1545 }, { "epoch": 1.7661393758016246, "grad_norm": 5.151268005371094, "learning_rate": 4.957303587866557e-05, "loss": 0.1916, "num_input_tokens_seen": 30282496, "step": 1550 }, { "epoch": 1.771839817585863, "grad_norm": 3.622302293777466, "learning_rate": 4.957028475977384e-05, "loss": 0.2405, "num_input_tokens_seen": 30380288, "step": 1555 }, { "epoch": 1.777540259370101, "grad_norm": 6.740144729614258, "learning_rate": 4.9567524882815686e-05, "loss": 0.2632, "num_input_tokens_seen": 30478048, "step": 1560 }, { "epoch": 1.7832407011543394, "grad_norm": 8.049473762512207, "learning_rate": 4.956475624877486e-05, "loss": 0.4007, "num_input_tokens_seen": 30575728, "step": 1565 }, { "epoch": 1.7889411429385778, "grad_norm": 6.096712589263916, "learning_rate": 4.9561978858638245e-05, "loss": 0.3395, "num_input_tokens_seen": 30673488, "step": 1570 }, { "epoch": 1.7946415847228159, "grad_norm": 5.549604415893555, "learning_rate": 4.955919271339584e-05, "loss": 0.2917, "num_input_tokens_seen": 30771120, "step": 1575 }, { "epoch": 1.8003420265070544, "grad_norm": 6.270097732543945, "learning_rate": 4.9556397814040754e-05, "loss": 0.1805, "num_input_tokens_seen": 30868848, "step": 1580 }, { "epoch": 1.8060424682912926, "grad_norm": 5.557713985443115, "learning_rate": 4.955359416156925e-05, "loss": 0.2391, "num_input_tokens_seen": 30966576, "step": 1585 }, { "epoch": 1.811742910075531, "grad_norm": 5.296855926513672, "learning_rate": 4.955078175698067e-05, "loss": 0.3259, "num_input_tokens_seen": 31064320, "step": 1590 }, { "epoch": 1.8174433518597692, "grad_norm": 5.627151966094971, "learning_rate": 4.9547960601277496e-05, "loss": 0.2576, "num_input_tokens_seen": 31162048, "step": 1595 }, { "epoch": 1.8231437936440074, "grad_norm": 6.599062919616699, "learning_rate": 4.9545130695465336e-05, "loss": 0.2859, "num_input_tokens_seen": 31259840, "step": 1600 }, { "epoch": 1.8288442354282457, "grad_norm": 8.176803588867188, "learning_rate": 4.954229204055291e-05, "loss": 0.1917, "num_input_tokens_seen": 31357568, "step": 1605 }, { "epoch": 1.834544677212484, "grad_norm": 9.62484073638916, "learning_rate": 4.953944463755204e-05, "loss": 0.3755, "num_input_tokens_seen": 31455344, "step": 1610 }, { "epoch": 1.8402451189967222, "grad_norm": 6.320862293243408, "learning_rate": 4.9536588487477697e-05, "loss": 0.2781, "num_input_tokens_seen": 31553024, "step": 1615 }, { "epoch": 1.8459455607809605, "grad_norm": 4.023454189300537, "learning_rate": 4.953372359134795e-05, "loss": 0.2669, "num_input_tokens_seen": 31650848, "step": 1620 }, { "epoch": 1.8516460025651988, "grad_norm": 5.63836145401001, "learning_rate": 4.953084995018398e-05, "loss": 0.2577, "num_input_tokens_seen": 31748560, "step": 1625 }, { "epoch": 1.857346444349437, "grad_norm": 8.508479118347168, "learning_rate": 4.95279675650101e-05, "loss": 0.277, "num_input_tokens_seen": 31846224, "step": 1630 }, { "epoch": 1.8630468861336753, "grad_norm": 7.421855926513672, "learning_rate": 4.952507643685375e-05, "loss": 0.2915, "num_input_tokens_seen": 31944016, "step": 1635 }, { "epoch": 1.8687473279179136, "grad_norm": 8.737668991088867, "learning_rate": 4.952217656674546e-05, "loss": 0.2798, "num_input_tokens_seen": 32041680, "step": 1640 }, { "epoch": 1.8744477697021518, "grad_norm": 6.379103183746338, "learning_rate": 4.951926795571888e-05, "loss": 0.2403, "num_input_tokens_seen": 32139392, "step": 1645 }, { "epoch": 1.8801482114863903, "grad_norm": 3.9837377071380615, "learning_rate": 4.9516350604810793e-05, "loss": 0.1932, "num_input_tokens_seen": 32237184, "step": 1650 }, { "epoch": 1.8858486532706284, "grad_norm": 6.174622535705566, "learning_rate": 4.951342451506108e-05, "loss": 0.2904, "num_input_tokens_seen": 32334816, "step": 1655 }, { "epoch": 1.8915490950548668, "grad_norm": 5.5978899002075195, "learning_rate": 4.951048968751275e-05, "loss": 0.2017, "num_input_tokens_seen": 32432528, "step": 1660 }, { "epoch": 1.8972495368391051, "grad_norm": 6.84478759765625, "learning_rate": 4.9507546123211926e-05, "loss": 0.2464, "num_input_tokens_seen": 32530320, "step": 1665 }, { "epoch": 1.9029499786233433, "grad_norm": 4.2474799156188965, "learning_rate": 4.950459382320782e-05, "loss": 0.1859, "num_input_tokens_seen": 32628016, "step": 1670 }, { "epoch": 1.9086504204075816, "grad_norm": 7.542076587677002, "learning_rate": 4.9501632788552805e-05, "loss": 0.2051, "num_input_tokens_seen": 32725744, "step": 1675 }, { "epoch": 1.91435086219182, "grad_norm": 8.932976722717285, "learning_rate": 4.949866302030232e-05, "loss": 0.3001, "num_input_tokens_seen": 32823424, "step": 1680 }, { "epoch": 1.920051303976058, "grad_norm": 8.958706855773926, "learning_rate": 4.949568451951495e-05, "loss": 0.4515, "num_input_tokens_seen": 32921120, "step": 1685 }, { "epoch": 1.9257517457602964, "grad_norm": 8.6969633102417, "learning_rate": 4.9492697287252365e-05, "loss": 0.2328, "num_input_tokens_seen": 33018880, "step": 1690 }, { "epoch": 1.9314521875445347, "grad_norm": 5.649287223815918, "learning_rate": 4.948970132457938e-05, "loss": 0.2487, "num_input_tokens_seen": 33116656, "step": 1695 }, { "epoch": 1.9371526293287729, "grad_norm": 7.401125431060791, "learning_rate": 4.94866966325639e-05, "loss": 0.2839, "num_input_tokens_seen": 33214416, "step": 1700 }, { "epoch": 1.9428530711130114, "grad_norm": 8.189993858337402, "learning_rate": 4.9483683212276935e-05, "loss": 0.1811, "num_input_tokens_seen": 33312096, "step": 1705 }, { "epoch": 1.9485535128972495, "grad_norm": 8.221793174743652, "learning_rate": 4.948066106479262e-05, "loss": 0.2459, "num_input_tokens_seen": 33409792, "step": 1710 }, { "epoch": 1.9542539546814877, "grad_norm": 5.853153705596924, "learning_rate": 4.947763019118821e-05, "loss": 0.3363, "num_input_tokens_seen": 33507504, "step": 1715 }, { "epoch": 1.9599543964657262, "grad_norm": 7.756158351898193, "learning_rate": 4.947459059254405e-05, "loss": 0.2134, "num_input_tokens_seen": 33605136, "step": 1720 }, { "epoch": 1.9656548382499643, "grad_norm": 6.2845025062561035, "learning_rate": 4.9471542269943604e-05, "loss": 0.2498, "num_input_tokens_seen": 33702928, "step": 1725 }, { "epoch": 1.9713552800342027, "grad_norm": 2.1698145866394043, "learning_rate": 4.946848522447345e-05, "loss": 0.1366, "num_input_tokens_seen": 33800656, "step": 1730 }, { "epoch": 1.977055721818441, "grad_norm": 8.690340995788574, "learning_rate": 4.946541945722326e-05, "loss": 0.3645, "num_input_tokens_seen": 33898336, "step": 1735 }, { "epoch": 1.9827561636026791, "grad_norm": 7.308428764343262, "learning_rate": 4.946234496928583e-05, "loss": 0.1994, "num_input_tokens_seen": 33996096, "step": 1740 }, { "epoch": 1.9884566053869175, "grad_norm": 5.161618709564209, "learning_rate": 4.945926176175707e-05, "loss": 0.226, "num_input_tokens_seen": 34093792, "step": 1745 }, { "epoch": 1.9941570471711558, "grad_norm": 9.512948989868164, "learning_rate": 4.945616983573598e-05, "loss": 0.2135, "num_input_tokens_seen": 34191552, "step": 1750 }, { "epoch": 1.999857488955394, "grad_norm": 8.386415481567383, "learning_rate": 4.945306919232467e-05, "loss": 0.236, "num_input_tokens_seen": 34289248, "step": 1755 }, { "epoch": 2.0045603534273906, "grad_norm": 7.716324329376221, "learning_rate": 4.944995983262837e-05, "loss": 0.3453, "num_input_tokens_seen": 34369840, "step": 1760 }, { "epoch": 2.0102607952116287, "grad_norm": 3.110274076461792, "learning_rate": 4.9446841757755405e-05, "loss": 0.1964, "num_input_tokens_seen": 34467568, "step": 1765 }, { "epoch": 2.0159612369958673, "grad_norm": 6.07765007019043, "learning_rate": 4.944371496881721e-05, "loss": 0.2358, "num_input_tokens_seen": 34565248, "step": 1770 }, { "epoch": 2.0216616787801054, "grad_norm": 5.835047721862793, "learning_rate": 4.944057946692834e-05, "loss": 0.1317, "num_input_tokens_seen": 34662896, "step": 1775 }, { "epoch": 2.027362120564344, "grad_norm": 6.831048488616943, "learning_rate": 4.943743525320643e-05, "loss": 0.2355, "num_input_tokens_seen": 34760624, "step": 1780 }, { "epoch": 2.033062562348582, "grad_norm": 11.599326133728027, "learning_rate": 4.943428232877224e-05, "loss": 0.1869, "num_input_tokens_seen": 34858288, "step": 1785 }, { "epoch": 2.03876300413282, "grad_norm": 6.280679702758789, "learning_rate": 4.943112069474963e-05, "loss": 0.2707, "num_input_tokens_seen": 34955968, "step": 1790 }, { "epoch": 2.0444634459170588, "grad_norm": 7.326327800750732, "learning_rate": 4.942795035226555e-05, "loss": 0.2077, "num_input_tokens_seen": 35053744, "step": 1795 }, { "epoch": 2.050163887701297, "grad_norm": 4.219114303588867, "learning_rate": 4.9424771302450084e-05, "loss": 0.1575, "num_input_tokens_seen": 35151408, "step": 1800 }, { "epoch": 2.055864329485535, "grad_norm": 4.75279426574707, "learning_rate": 4.942158354643639e-05, "loss": 0.1663, "num_input_tokens_seen": 35249168, "step": 1805 }, { "epoch": 2.0615647712697736, "grad_norm": 7.397295951843262, "learning_rate": 4.9418387085360754e-05, "loss": 0.1872, "num_input_tokens_seen": 35346880, "step": 1810 }, { "epoch": 2.0672652130540117, "grad_norm": 7.649684906005859, "learning_rate": 4.941518192036254e-05, "loss": 0.2212, "num_input_tokens_seen": 35444688, "step": 1815 }, { "epoch": 2.07296565483825, "grad_norm": 3.966686725616455, "learning_rate": 4.941196805258423e-05, "loss": 0.1185, "num_input_tokens_seen": 35542416, "step": 1820 }, { "epoch": 2.0786660966224884, "grad_norm": 10.044607162475586, "learning_rate": 4.940874548317143e-05, "loss": 0.2099, "num_input_tokens_seen": 35640128, "step": 1825 }, { "epoch": 2.0843665384067265, "grad_norm": 5.567621231079102, "learning_rate": 4.9405514213272784e-05, "loss": 0.172, "num_input_tokens_seen": 35737872, "step": 1830 }, { "epoch": 2.0900669801909646, "grad_norm": 6.253763675689697, "learning_rate": 4.94022742440401e-05, "loss": 0.1485, "num_input_tokens_seen": 35835568, "step": 1835 }, { "epoch": 2.095767421975203, "grad_norm": 5.565789222717285, "learning_rate": 4.939902557662826e-05, "loss": 0.2586, "num_input_tokens_seen": 35933312, "step": 1840 }, { "epoch": 2.1014678637594413, "grad_norm": 4.763193607330322, "learning_rate": 4.939576821219525e-05, "loss": 0.2357, "num_input_tokens_seen": 36030944, "step": 1845 }, { "epoch": 2.10716830554368, "grad_norm": 11.802780151367188, "learning_rate": 4.9392502151902156e-05, "loss": 0.2471, "num_input_tokens_seen": 36128688, "step": 1850 }, { "epoch": 2.112868747327918, "grad_norm": 6.20152473449707, "learning_rate": 4.938922739691316e-05, "loss": 0.1398, "num_input_tokens_seen": 36226368, "step": 1855 }, { "epoch": 2.118569189112156, "grad_norm": 6.946401119232178, "learning_rate": 4.938594394839555e-05, "loss": 0.1601, "num_input_tokens_seen": 36324096, "step": 1860 }, { "epoch": 2.1242696308963946, "grad_norm": 4.598988056182861, "learning_rate": 4.938265180751971e-05, "loss": 0.1461, "num_input_tokens_seen": 36421840, "step": 1865 }, { "epoch": 2.1299700726806328, "grad_norm": 8.171868324279785, "learning_rate": 4.937935097545912e-05, "loss": 0.2531, "num_input_tokens_seen": 36519552, "step": 1870 }, { "epoch": 2.135670514464871, "grad_norm": 3.664236068725586, "learning_rate": 4.9376041453390365e-05, "loss": 0.1934, "num_input_tokens_seen": 36617280, "step": 1875 }, { "epoch": 2.1413709562491094, "grad_norm": 8.51440143585205, "learning_rate": 4.937272324249312e-05, "loss": 0.2024, "num_input_tokens_seen": 36714992, "step": 1880 }, { "epoch": 2.1470713980333476, "grad_norm": 6.9583001136779785, "learning_rate": 4.9369396343950154e-05, "loss": 0.2121, "num_input_tokens_seen": 36812784, "step": 1885 }, { "epoch": 2.1527718398175857, "grad_norm": 10.295254707336426, "learning_rate": 4.936606075894734e-05, "loss": 0.172, "num_input_tokens_seen": 36910688, "step": 1890 }, { "epoch": 2.1584722816018242, "grad_norm": 3.664754629135132, "learning_rate": 4.9362716488673654e-05, "loss": 0.163, "num_input_tokens_seen": 37008464, "step": 1895 }, { "epoch": 2.1641727233860624, "grad_norm": 9.738370895385742, "learning_rate": 4.9359363534321156e-05, "loss": 0.1591, "num_input_tokens_seen": 37106272, "step": 1900 }, { "epoch": 2.1698731651703005, "grad_norm": 6.579972743988037, "learning_rate": 4.9356001897085e-05, "loss": 0.1816, "num_input_tokens_seen": 37204048, "step": 1905 }, { "epoch": 2.175573606954539, "grad_norm": 6.322074890136719, "learning_rate": 4.935263157816345e-05, "loss": 0.183, "num_input_tokens_seen": 37301824, "step": 1910 }, { "epoch": 2.181274048738777, "grad_norm": 5.997386455535889, "learning_rate": 4.934925257875784e-05, "loss": 0.1722, "num_input_tokens_seen": 37399632, "step": 1915 }, { "epoch": 2.1869744905230157, "grad_norm": 7.246592998504639, "learning_rate": 4.9345864900072625e-05, "loss": 0.1017, "num_input_tokens_seen": 37497296, "step": 1920 }, { "epoch": 2.192674932307254, "grad_norm": 7.348723888397217, "learning_rate": 4.934246854331534e-05, "loss": 0.1756, "num_input_tokens_seen": 37595168, "step": 1925 }, { "epoch": 2.198375374091492, "grad_norm": 8.612208366394043, "learning_rate": 4.933906350969661e-05, "loss": 0.1674, "num_input_tokens_seen": 37692832, "step": 1930 }, { "epoch": 2.2040758158757305, "grad_norm": 16.743350982666016, "learning_rate": 4.933564980043015e-05, "loss": 0.2679, "num_input_tokens_seen": 37790512, "step": 1935 }, { "epoch": 2.2097762576599687, "grad_norm": 2.924848794937134, "learning_rate": 4.93322274167328e-05, "loss": 0.0981, "num_input_tokens_seen": 37888304, "step": 1940 }, { "epoch": 2.2154766994442068, "grad_norm": 8.15044116973877, "learning_rate": 4.9328796359824445e-05, "loss": 0.1621, "num_input_tokens_seen": 37986032, "step": 1945 }, { "epoch": 2.2211771412284453, "grad_norm": 9.112587928771973, "learning_rate": 4.932535663092809e-05, "loss": 0.2655, "num_input_tokens_seen": 38083776, "step": 1950 }, { "epoch": 2.2268775830126835, "grad_norm": 4.596200942993164, "learning_rate": 4.932190823126982e-05, "loss": 0.1608, "num_input_tokens_seen": 38181488, "step": 1955 }, { "epoch": 2.2325780247969216, "grad_norm": 5.918450355529785, "learning_rate": 4.9318451162078824e-05, "loss": 0.1119, "num_input_tokens_seen": 38279248, "step": 1960 }, { "epoch": 2.23827846658116, "grad_norm": 8.088936805725098, "learning_rate": 4.931498542458738e-05, "loss": 0.2202, "num_input_tokens_seen": 38377024, "step": 1965 }, { "epoch": 2.2439789083653983, "grad_norm": 6.410324573516846, "learning_rate": 4.931151102003082e-05, "loss": 0.1136, "num_input_tokens_seen": 38474768, "step": 1970 }, { "epoch": 2.249679350149637, "grad_norm": 8.893948554992676, "learning_rate": 4.930802794964763e-05, "loss": 0.1233, "num_input_tokens_seen": 38572432, "step": 1975 }, { "epoch": 2.255379791933875, "grad_norm": 2.6239705085754395, "learning_rate": 4.9304536214679315e-05, "loss": 0.1409, "num_input_tokens_seen": 38670112, "step": 1980 }, { "epoch": 2.261080233718113, "grad_norm": 8.951133728027344, "learning_rate": 4.930103581637052e-05, "loss": 0.159, "num_input_tokens_seen": 38767872, "step": 1985 }, { "epoch": 2.2667806755023516, "grad_norm": 8.211560249328613, "learning_rate": 4.929752675596896e-05, "loss": 0.1761, "num_input_tokens_seen": 38865584, "step": 1990 }, { "epoch": 2.2724811172865897, "grad_norm": 4.264492034912109, "learning_rate": 4.929400903472544e-05, "loss": 0.1206, "num_input_tokens_seen": 38963264, "step": 1995 }, { "epoch": 2.278181559070828, "grad_norm": 10.552608489990234, "learning_rate": 4.9290482653893846e-05, "loss": 0.1895, "num_input_tokens_seen": 39060944, "step": 2000 }, { "epoch": 2.2838820008550664, "grad_norm": 8.71176528930664, "learning_rate": 4.928694761473115e-05, "loss": 0.1604, "num_input_tokens_seen": 39158640, "step": 2005 }, { "epoch": 2.2895824426393045, "grad_norm": 6.852836608886719, "learning_rate": 4.928340391849742e-05, "loss": 0.2317, "num_input_tokens_seen": 39256352, "step": 2010 }, { "epoch": 2.2952828844235427, "grad_norm": 6.0051493644714355, "learning_rate": 4.9279851566455806e-05, "loss": 0.1945, "num_input_tokens_seen": 39354112, "step": 2015 }, { "epoch": 2.3009833262077812, "grad_norm": 8.092498779296875, "learning_rate": 4.927629055987254e-05, "loss": 0.1393, "num_input_tokens_seen": 39451824, "step": 2020 }, { "epoch": 2.3066837679920194, "grad_norm": 11.641897201538086, "learning_rate": 4.927272090001695e-05, "loss": 0.1692, "num_input_tokens_seen": 39549600, "step": 2025 }, { "epoch": 2.3123842097762575, "grad_norm": 6.343209266662598, "learning_rate": 4.9269142588161424e-05, "loss": 0.1058, "num_input_tokens_seen": 39647280, "step": 2030 }, { "epoch": 2.318084651560496, "grad_norm": 9.047554969787598, "learning_rate": 4.9265555625581464e-05, "loss": 0.1835, "num_input_tokens_seen": 39745040, "step": 2035 }, { "epoch": 2.323785093344734, "grad_norm": 7.915929317474365, "learning_rate": 4.9261960013555625e-05, "loss": 0.2291, "num_input_tokens_seen": 39842816, "step": 2040 }, { "epoch": 2.3294855351289723, "grad_norm": 5.318630695343018, "learning_rate": 4.925835575336557e-05, "loss": 0.1533, "num_input_tokens_seen": 39940576, "step": 2045 }, { "epoch": 2.335185976913211, "grad_norm": 7.653360366821289, "learning_rate": 4.9254742846296045e-05, "loss": 0.1978, "num_input_tokens_seen": 40038368, "step": 2050 }, { "epoch": 2.340886418697449, "grad_norm": 7.852567672729492, "learning_rate": 4.925112129363486e-05, "loss": 0.1531, "num_input_tokens_seen": 40136144, "step": 2055 }, { "epoch": 2.3465868604816875, "grad_norm": 6.86559534072876, "learning_rate": 4.92474910966729e-05, "loss": 0.0854, "num_input_tokens_seen": 40233952, "step": 2060 }, { "epoch": 2.3522873022659256, "grad_norm": 10.074113845825195, "learning_rate": 4.9243852256704183e-05, "loss": 0.1915, "num_input_tokens_seen": 40331696, "step": 2065 }, { "epoch": 2.3579877440501638, "grad_norm": 7.896622180938721, "learning_rate": 4.924020477502574e-05, "loss": 0.1495, "num_input_tokens_seen": 40429360, "step": 2070 }, { "epoch": 2.3636881858344023, "grad_norm": 8.397123336791992, "learning_rate": 4.923654865293773e-05, "loss": 0.1392, "num_input_tokens_seen": 40527136, "step": 2075 }, { "epoch": 2.3693886276186404, "grad_norm": 6.305740833282471, "learning_rate": 4.923288389174337e-05, "loss": 0.0875, "num_input_tokens_seen": 40624912, "step": 2080 }, { "epoch": 2.3750890694028786, "grad_norm": 8.387357711791992, "learning_rate": 4.9229210492748976e-05, "loss": 0.2358, "num_input_tokens_seen": 40722720, "step": 2085 }, { "epoch": 2.380789511187117, "grad_norm": 8.446219444274902, "learning_rate": 4.92255284572639e-05, "loss": 0.1482, "num_input_tokens_seen": 40820448, "step": 2090 }, { "epoch": 2.3864899529713552, "grad_norm": 6.30406379699707, "learning_rate": 4.9221837786600634e-05, "loss": 0.1603, "num_input_tokens_seen": 40918256, "step": 2095 }, { "epoch": 2.392190394755594, "grad_norm": 7.828269004821777, "learning_rate": 4.921813848207469e-05, "loss": 0.1764, "num_input_tokens_seen": 41015920, "step": 2100 }, { "epoch": 2.397890836539832, "grad_norm": 7.984679698944092, "learning_rate": 4.921443054500471e-05, "loss": 0.1809, "num_input_tokens_seen": 41113632, "step": 2105 }, { "epoch": 2.40359127832407, "grad_norm": 10.998138427734375, "learning_rate": 4.921071397671235e-05, "loss": 0.185, "num_input_tokens_seen": 41211344, "step": 2110 }, { "epoch": 2.4092917201083086, "grad_norm": 7.500617980957031, "learning_rate": 4.9206988778522414e-05, "loss": 0.116, "num_input_tokens_seen": 41308992, "step": 2115 }, { "epoch": 2.4149921618925467, "grad_norm": 7.988670349121094, "learning_rate": 4.9203254951762735e-05, "loss": 0.1457, "num_input_tokens_seen": 41406752, "step": 2120 }, { "epoch": 2.420692603676785, "grad_norm": 10.234942436218262, "learning_rate": 4.9199512497764226e-05, "loss": 0.2256, "num_input_tokens_seen": 41504464, "step": 2125 }, { "epoch": 2.4263930454610234, "grad_norm": 11.052645683288574, "learning_rate": 4.919576141786089e-05, "loss": 0.1721, "num_input_tokens_seen": 41602272, "step": 2130 }, { "epoch": 2.4320934872452615, "grad_norm": 3.908461332321167, "learning_rate": 4.91920017133898e-05, "loss": 0.1676, "num_input_tokens_seen": 41700000, "step": 2135 }, { "epoch": 2.4377939290294997, "grad_norm": 3.8486247062683105, "learning_rate": 4.9188233385691094e-05, "loss": 0.1458, "num_input_tokens_seen": 41797696, "step": 2140 }, { "epoch": 2.443494370813738, "grad_norm": 7.926904678344727, "learning_rate": 4.9184456436107994e-05, "loss": 0.202, "num_input_tokens_seen": 41895392, "step": 2145 }, { "epoch": 2.4491948125979763, "grad_norm": 8.783378601074219, "learning_rate": 4.91806708659868e-05, "loss": 0.1838, "num_input_tokens_seen": 41993120, "step": 2150 }, { "epoch": 2.4548952543822145, "grad_norm": 4.2609357833862305, "learning_rate": 4.917687667667686e-05, "loss": 0.1037, "num_input_tokens_seen": 42090880, "step": 2155 }, { "epoch": 2.460595696166453, "grad_norm": 10.791552543640137, "learning_rate": 4.917307386953062e-05, "loss": 0.1791, "num_input_tokens_seen": 42188576, "step": 2160 }, { "epoch": 2.466296137950691, "grad_norm": 8.570240020751953, "learning_rate": 4.9169262445903595e-05, "loss": 0.1608, "num_input_tokens_seen": 42286272, "step": 2165 }, { "epoch": 2.4719965797349293, "grad_norm": 5.907083034515381, "learning_rate": 4.9165442407154355e-05, "loss": 0.1657, "num_input_tokens_seen": 42384048, "step": 2170 }, { "epoch": 2.477697021519168, "grad_norm": 9.303824424743652, "learning_rate": 4.916161375464455e-05, "loss": 0.1839, "num_input_tokens_seen": 42481888, "step": 2175 }, { "epoch": 2.483397463303406, "grad_norm": 9.526128768920898, "learning_rate": 4.915777648973892e-05, "loss": 0.1084, "num_input_tokens_seen": 42579600, "step": 2180 }, { "epoch": 2.489097905087644, "grad_norm": 8.502360343933105, "learning_rate": 4.915393061380523e-05, "loss": 0.1205, "num_input_tokens_seen": 42677360, "step": 2185 }, { "epoch": 2.4947983468718826, "grad_norm": 8.056527137756348, "learning_rate": 4.9150076128214364e-05, "loss": 0.1244, "num_input_tokens_seen": 42775072, "step": 2190 }, { "epoch": 2.5004987886561207, "grad_norm": 3.5900754928588867, "learning_rate": 4.914621303434023e-05, "loss": 0.1198, "num_input_tokens_seen": 42872832, "step": 2195 }, { "epoch": 2.506199230440359, "grad_norm": 4.139026641845703, "learning_rate": 4.914234133355984e-05, "loss": 0.1016, "num_input_tokens_seen": 42970592, "step": 2200 }, { "epoch": 2.5118996722245974, "grad_norm": 8.35708999633789, "learning_rate": 4.9138461027253255e-05, "loss": 0.1066, "num_input_tokens_seen": 43068384, "step": 2205 }, { "epoch": 2.5176001140088355, "grad_norm": 2.149785041809082, "learning_rate": 4.913457211680361e-05, "loss": 0.0866, "num_input_tokens_seen": 43166240, "step": 2210 }, { "epoch": 2.523300555793074, "grad_norm": 3.6671578884124756, "learning_rate": 4.913067460359711e-05, "loss": 0.1831, "num_input_tokens_seen": 43264000, "step": 2215 }, { "epoch": 2.5290009975773122, "grad_norm": 7.476966857910156, "learning_rate": 4.912676848902301e-05, "loss": 0.1276, "num_input_tokens_seen": 43361712, "step": 2220 }, { "epoch": 2.534701439361551, "grad_norm": 14.653030395507812, "learning_rate": 4.912285377447366e-05, "loss": 0.1622, "num_input_tokens_seen": 43459392, "step": 2225 }, { "epoch": 2.540401881145789, "grad_norm": 7.570925712585449, "learning_rate": 4.9118930461344433e-05, "loss": 0.1279, "num_input_tokens_seen": 43557104, "step": 2230 }, { "epoch": 2.546102322930027, "grad_norm": 7.563347816467285, "learning_rate": 4.911499855103382e-05, "loss": 0.1015, "num_input_tokens_seen": 43654928, "step": 2235 }, { "epoch": 2.5518027647142656, "grad_norm": 5.281002521514893, "learning_rate": 4.9111058044943334e-05, "loss": 0.1255, "num_input_tokens_seen": 43752672, "step": 2240 }, { "epoch": 2.5575032064985037, "grad_norm": 5.972769737243652, "learning_rate": 4.910710894447757e-05, "loss": 0.0551, "num_input_tokens_seen": 43850512, "step": 2245 }, { "epoch": 2.563203648282742, "grad_norm": 1.96940279006958, "learning_rate": 4.9103151251044174e-05, "loss": 0.0708, "num_input_tokens_seen": 43948336, "step": 2250 }, { "epoch": 2.5689040900669804, "grad_norm": 10.9530668258667, "learning_rate": 4.909918496605387e-05, "loss": 0.1775, "num_input_tokens_seen": 44046080, "step": 2255 }, { "epoch": 2.5746045318512185, "grad_norm": 8.717411994934082, "learning_rate": 4.909521009092045e-05, "loss": 0.0874, "num_input_tokens_seen": 44143808, "step": 2260 }, { "epoch": 2.5803049736354566, "grad_norm": 11.679935455322266, "learning_rate": 4.909122662706074e-05, "loss": 0.2068, "num_input_tokens_seen": 44241600, "step": 2265 }, { "epoch": 2.586005415419695, "grad_norm": 5.130126953125, "learning_rate": 4.9087234575894644e-05, "loss": 0.0785, "num_input_tokens_seen": 44339312, "step": 2270 }, { "epoch": 2.5917058572039333, "grad_norm": 5.87738037109375, "learning_rate": 4.908323393884514e-05, "loss": 0.0893, "num_input_tokens_seen": 44437136, "step": 2275 }, { "epoch": 2.5974062989881714, "grad_norm": 9.397594451904297, "learning_rate": 4.9079224717338246e-05, "loss": 0.142, "num_input_tokens_seen": 44534912, "step": 2280 }, { "epoch": 2.60310674077241, "grad_norm": 3.5476927757263184, "learning_rate": 4.907520691280304e-05, "loss": 0.0855, "num_input_tokens_seen": 44632720, "step": 2285 }, { "epoch": 2.608807182556648, "grad_norm": 10.59860610961914, "learning_rate": 4.907118052667168e-05, "loss": 0.1536, "num_input_tokens_seen": 44730480, "step": 2290 }, { "epoch": 2.6145076243408862, "grad_norm": 7.346194744110107, "learning_rate": 4.906714556037936e-05, "loss": 0.1219, "num_input_tokens_seen": 44828112, "step": 2295 }, { "epoch": 2.620208066125125, "grad_norm": 4.199995517730713, "learning_rate": 4.9063102015364344e-05, "loss": 0.0867, "num_input_tokens_seen": 44925888, "step": 2300 }, { "epoch": 2.625908507909363, "grad_norm": 12.029003143310547, "learning_rate": 4.9059049893067954e-05, "loss": 0.1819, "num_input_tokens_seen": 45023728, "step": 2305 }, { "epoch": 2.631608949693601, "grad_norm": 9.11473560333252, "learning_rate": 4.9054989194934564e-05, "loss": 0.1298, "num_input_tokens_seen": 45121424, "step": 2310 }, { "epoch": 2.6373093914778396, "grad_norm": 17.91231918334961, "learning_rate": 4.905091992241161e-05, "loss": 0.1854, "num_input_tokens_seen": 45219200, "step": 2315 }, { "epoch": 2.6430098332620777, "grad_norm": 10.454273223876953, "learning_rate": 4.9046842076949576e-05, "loss": 0.2016, "num_input_tokens_seen": 45316944, "step": 2320 }, { "epoch": 2.648710275046316, "grad_norm": 11.895720481872559, "learning_rate": 4.904275566000202e-05, "loss": 0.2173, "num_input_tokens_seen": 45414688, "step": 2325 }, { "epoch": 2.6544107168305544, "grad_norm": 7.68233585357666, "learning_rate": 4.903866067302554e-05, "loss": 0.1429, "num_input_tokens_seen": 45512400, "step": 2330 }, { "epoch": 2.6601111586147925, "grad_norm": 6.063493251800537, "learning_rate": 4.9034557117479786e-05, "loss": 0.1397, "num_input_tokens_seen": 45610128, "step": 2335 }, { "epoch": 2.665811600399031, "grad_norm": 1.1613508462905884, "learning_rate": 4.903044499482747e-05, "loss": 0.0946, "num_input_tokens_seen": 45707920, "step": 2340 }, { "epoch": 2.671512042183269, "grad_norm": 10.7675142288208, "learning_rate": 4.902632430653435e-05, "loss": 0.1761, "num_input_tokens_seen": 45805744, "step": 2345 }, { "epoch": 2.6772124839675073, "grad_norm": 7.67887020111084, "learning_rate": 4.902219505406926e-05, "loss": 0.1615, "num_input_tokens_seen": 45903456, "step": 2350 }, { "epoch": 2.682912925751746, "grad_norm": 2.045400381088257, "learning_rate": 4.901805723890407e-05, "loss": 0.173, "num_input_tokens_seen": 46001264, "step": 2355 }, { "epoch": 2.688613367535984, "grad_norm": 8.454596519470215, "learning_rate": 4.9013910862513676e-05, "loss": 0.1894, "num_input_tokens_seen": 46098976, "step": 2360 }, { "epoch": 2.6943138093202226, "grad_norm": 7.203195095062256, "learning_rate": 4.9009755926376085e-05, "loss": 0.1496, "num_input_tokens_seen": 46196816, "step": 2365 }, { "epoch": 2.7000142511044607, "grad_norm": 10.533610343933105, "learning_rate": 4.9005592431972304e-05, "loss": 0.0768, "num_input_tokens_seen": 46294480, "step": 2370 }, { "epoch": 2.705714692888699, "grad_norm": 7.490808486938477, "learning_rate": 4.90014203807864e-05, "loss": 0.1114, "num_input_tokens_seen": 46392096, "step": 2375 }, { "epoch": 2.7114151346729374, "grad_norm": 7.0142083168029785, "learning_rate": 4.899723977430552e-05, "loss": 0.0883, "num_input_tokens_seen": 46489936, "step": 2380 }, { "epoch": 2.7171155764571755, "grad_norm": 8.30893611907959, "learning_rate": 4.899305061401983e-05, "loss": 0.1146, "num_input_tokens_seen": 46587648, "step": 2385 }, { "epoch": 2.7228160182414136, "grad_norm": 5.555058479309082, "learning_rate": 4.898885290142254e-05, "loss": 0.1212, "num_input_tokens_seen": 46685360, "step": 2390 }, { "epoch": 2.728516460025652, "grad_norm": 3.8162529468536377, "learning_rate": 4.898464663800995e-05, "loss": 0.1327, "num_input_tokens_seen": 46783072, "step": 2395 }, { "epoch": 2.7342169018098903, "grad_norm": 7.592154026031494, "learning_rate": 4.898043182528136e-05, "loss": 0.0871, "num_input_tokens_seen": 46880832, "step": 2400 }, { "epoch": 2.7399173435941284, "grad_norm": 4.684682846069336, "learning_rate": 4.897620846473915e-05, "loss": 0.0563, "num_input_tokens_seen": 46978576, "step": 2405 }, { "epoch": 2.745617785378367, "grad_norm": 1.159037709236145, "learning_rate": 4.897197655788872e-05, "loss": 0.1116, "num_input_tokens_seen": 47076304, "step": 2410 }, { "epoch": 2.751318227162605, "grad_norm": 9.983619689941406, "learning_rate": 4.8967736106238546e-05, "loss": 0.1072, "num_input_tokens_seen": 47174000, "step": 2415 }, { "epoch": 2.757018668946843, "grad_norm": 5.6760029792785645, "learning_rate": 4.8963487111300133e-05, "loss": 0.0847, "num_input_tokens_seen": 47271760, "step": 2420 }, { "epoch": 2.762719110731082, "grad_norm": 9.46414566040039, "learning_rate": 4.895922957458803e-05, "loss": 0.0821, "num_input_tokens_seen": 47369504, "step": 2425 }, { "epoch": 2.76841955251532, "grad_norm": 11.030294418334961, "learning_rate": 4.8954963497619836e-05, "loss": 0.1595, "num_input_tokens_seen": 47467312, "step": 2430 }, { "epoch": 2.774119994299558, "grad_norm": 4.053030967712402, "learning_rate": 4.895068888191618e-05, "loss": 0.0967, "num_input_tokens_seen": 47565024, "step": 2435 }, { "epoch": 2.7798204360837966, "grad_norm": 8.997540473937988, "learning_rate": 4.894640572900076e-05, "loss": 0.1222, "num_input_tokens_seen": 47662768, "step": 2440 }, { "epoch": 2.7855208778680347, "grad_norm": 9.141730308532715, "learning_rate": 4.89421140404003e-05, "loss": 0.1496, "num_input_tokens_seen": 47760640, "step": 2445 }, { "epoch": 2.791221319652273, "grad_norm": 6.096067905426025, "learning_rate": 4.8937813817644577e-05, "loss": 0.0965, "num_input_tokens_seen": 47858400, "step": 2450 }, { "epoch": 2.7969217614365114, "grad_norm": 7.94855260848999, "learning_rate": 4.89335050622664e-05, "loss": 0.1043, "num_input_tokens_seen": 47956112, "step": 2455 }, { "epoch": 2.8026222032207495, "grad_norm": 5.408116340637207, "learning_rate": 4.892918777580161e-05, "loss": 0.0953, "num_input_tokens_seen": 48053888, "step": 2460 }, { "epoch": 2.8083226450049876, "grad_norm": 10.836856842041016, "learning_rate": 4.8924861959789116e-05, "loss": 0.0829, "num_input_tokens_seen": 48151648, "step": 2465 }, { "epoch": 2.814023086789226, "grad_norm": 4.917642593383789, "learning_rate": 4.892052761577084e-05, "loss": 0.1339, "num_input_tokens_seen": 48249344, "step": 2470 }, { "epoch": 2.8197235285734643, "grad_norm": 4.280270099639893, "learning_rate": 4.891618474529178e-05, "loss": 0.0867, "num_input_tokens_seen": 48347088, "step": 2475 }, { "epoch": 2.825423970357703, "grad_norm": 10.1702241897583, "learning_rate": 4.8911833349899924e-05, "loss": 0.0944, "num_input_tokens_seen": 48444848, "step": 2480 }, { "epoch": 2.831124412141941, "grad_norm": 7.9395341873168945, "learning_rate": 4.890747343114634e-05, "loss": 0.1103, "num_input_tokens_seen": 48542528, "step": 2485 }, { "epoch": 2.836824853926179, "grad_norm": 7.740533828735352, "learning_rate": 4.8903104990585124e-05, "loss": 0.0763, "num_input_tokens_seen": 48640240, "step": 2490 }, { "epoch": 2.8425252957104177, "grad_norm": 2.646793842315674, "learning_rate": 4.8898728029773394e-05, "loss": 0.0821, "num_input_tokens_seen": 48737888, "step": 2495 }, { "epoch": 2.848225737494656, "grad_norm": 6.403926849365234, "learning_rate": 4.8894342550271314e-05, "loss": 0.0962, "num_input_tokens_seen": 48835600, "step": 2500 }, { "epoch": 2.8539261792788944, "grad_norm": 7.7822957038879395, "learning_rate": 4.888994855364209e-05, "loss": 0.0832, "num_input_tokens_seen": 48933312, "step": 2505 }, { "epoch": 2.8596266210631325, "grad_norm": 5.186079025268555, "learning_rate": 4.888554604145196e-05, "loss": 0.125, "num_input_tokens_seen": 49030960, "step": 2510 }, { "epoch": 2.8653270628473706, "grad_norm": 7.859062671661377, "learning_rate": 4.8881135015270206e-05, "loss": 0.0941, "num_input_tokens_seen": 49128672, "step": 2515 }, { "epoch": 2.871027504631609, "grad_norm": 7.483722686767578, "learning_rate": 4.887671547666912e-05, "loss": 0.1318, "num_input_tokens_seen": 49226416, "step": 2520 }, { "epoch": 2.8767279464158473, "grad_norm": 8.643847465515137, "learning_rate": 4.887228742722405e-05, "loss": 0.1856, "num_input_tokens_seen": 49324112, "step": 2525 }, { "epoch": 2.8824283882000854, "grad_norm": 8.750890731811523, "learning_rate": 4.8867850868513374e-05, "loss": 0.1006, "num_input_tokens_seen": 49421776, "step": 2530 }, { "epoch": 2.888128829984324, "grad_norm": 6.101781845092773, "learning_rate": 4.8863405802118514e-05, "loss": 0.1324, "num_input_tokens_seen": 49519568, "step": 2535 }, { "epoch": 2.893829271768562, "grad_norm": 7.980799198150635, "learning_rate": 4.8858952229623886e-05, "loss": 0.0907, "num_input_tokens_seen": 49617360, "step": 2540 }, { "epoch": 2.8995297135528, "grad_norm": 3.3241348266601562, "learning_rate": 4.8854490152616984e-05, "loss": 0.1104, "num_input_tokens_seen": 49715056, "step": 2545 }, { "epoch": 2.9052301553370388, "grad_norm": 10.823814392089844, "learning_rate": 4.88500195726883e-05, "loss": 0.1766, "num_input_tokens_seen": 49812848, "step": 2550 }, { "epoch": 2.910930597121277, "grad_norm": 6.91947078704834, "learning_rate": 4.884554049143139e-05, "loss": 0.1128, "num_input_tokens_seen": 49910496, "step": 2555 }, { "epoch": 2.916631038905515, "grad_norm": 4.440322399139404, "learning_rate": 4.884105291044279e-05, "loss": 0.0796, "num_input_tokens_seen": 50008224, "step": 2560 }, { "epoch": 2.9223314806897536, "grad_norm": 5.996119976043701, "learning_rate": 4.8836556831322125e-05, "loss": 0.1648, "num_input_tokens_seen": 50105952, "step": 2565 }, { "epoch": 2.9280319224739917, "grad_norm": 8.666937828063965, "learning_rate": 4.8832052255672e-05, "loss": 0.1488, "num_input_tokens_seen": 50203680, "step": 2570 }, { "epoch": 2.93373236425823, "grad_norm": 5.516872882843018, "learning_rate": 4.8827539185098085e-05, "loss": 0.1598, "num_input_tokens_seen": 50301504, "step": 2575 }, { "epoch": 2.9394328060424684, "grad_norm": 6.207433223724365, "learning_rate": 4.882301762120905e-05, "loss": 0.1003, "num_input_tokens_seen": 50399152, "step": 2580 }, { "epoch": 2.9451332478267065, "grad_norm": 5.3155598640441895, "learning_rate": 4.88184875656166e-05, "loss": 0.0675, "num_input_tokens_seen": 50496880, "step": 2585 }, { "epoch": 2.9508336896109446, "grad_norm": 8.118722915649414, "learning_rate": 4.881394901993549e-05, "loss": 0.0834, "num_input_tokens_seen": 50594656, "step": 2590 }, { "epoch": 2.956534131395183, "grad_norm": 8.280570983886719, "learning_rate": 4.880940198578347e-05, "loss": 0.1212, "num_input_tokens_seen": 50692496, "step": 2595 }, { "epoch": 2.9622345731794213, "grad_norm": 6.043283939361572, "learning_rate": 4.8804846464781334e-05, "loss": 0.1096, "num_input_tokens_seen": 50790272, "step": 2600 }, { "epoch": 2.9679350149636594, "grad_norm": 3.722360134124756, "learning_rate": 4.8800282458552885e-05, "loss": 0.155, "num_input_tokens_seen": 50888032, "step": 2605 }, { "epoch": 2.973635456747898, "grad_norm": 6.298059940338135, "learning_rate": 4.8795709968724974e-05, "loss": 0.072, "num_input_tokens_seen": 50985776, "step": 2610 }, { "epoch": 2.979335898532136, "grad_norm": 8.660189628601074, "learning_rate": 4.879112899692745e-05, "loss": 0.1247, "num_input_tokens_seen": 51083440, "step": 2615 }, { "epoch": 2.9850363403163747, "grad_norm": 14.756346702575684, "learning_rate": 4.8786539544793206e-05, "loss": 0.1067, "num_input_tokens_seen": 51181152, "step": 2620 }, { "epoch": 2.990736782100613, "grad_norm": 2.7973508834838867, "learning_rate": 4.878194161395816e-05, "loss": 0.0766, "num_input_tokens_seen": 51278912, "step": 2625 }, { "epoch": 2.9964372238848513, "grad_norm": 7.684298515319824, "learning_rate": 4.8777335206061216e-05, "loss": 0.0668, "num_input_tokens_seen": 51376640, "step": 2630 }, { "epoch": 3.0011400883568475, "grad_norm": 3.268608570098877, "learning_rate": 4.877272032274435e-05, "loss": 0.0698, "num_input_tokens_seen": 51457280, "step": 2635 }, { "epoch": 3.006840530141086, "grad_norm": 2.5024783611297607, "learning_rate": 4.876809696565252e-05, "loss": 0.0681, "num_input_tokens_seen": 51555088, "step": 2640 }, { "epoch": 3.012540971925324, "grad_norm": 2.1016647815704346, "learning_rate": 4.876346513643373e-05, "loss": 0.051, "num_input_tokens_seen": 51652864, "step": 2645 }, { "epoch": 3.0182414137095623, "grad_norm": 8.176024436950684, "learning_rate": 4.875882483673898e-05, "loss": 0.0712, "num_input_tokens_seen": 51750560, "step": 2650 }, { "epoch": 3.023941855493801, "grad_norm": 4.242624759674072, "learning_rate": 4.875417606822232e-05, "loss": 0.0761, "num_input_tokens_seen": 51848288, "step": 2655 }, { "epoch": 3.029642297278039, "grad_norm": 11.779088973999023, "learning_rate": 4.874951883254078e-05, "loss": 0.0485, "num_input_tokens_seen": 51946016, "step": 2660 }, { "epoch": 3.035342739062277, "grad_norm": 3.6110494136810303, "learning_rate": 4.874485313135446e-05, "loss": 0.0747, "num_input_tokens_seen": 52043776, "step": 2665 }, { "epoch": 3.0410431808465157, "grad_norm": 8.334362030029297, "learning_rate": 4.874017896632642e-05, "loss": 0.0614, "num_input_tokens_seen": 52141520, "step": 2670 }, { "epoch": 3.046743622630754, "grad_norm": 5.685539722442627, "learning_rate": 4.8735496339122776e-05, "loss": 0.0604, "num_input_tokens_seen": 52239200, "step": 2675 }, { "epoch": 3.052444064414992, "grad_norm": 4.587195873260498, "learning_rate": 4.8730805251412645e-05, "loss": 0.1134, "num_input_tokens_seen": 52336848, "step": 2680 }, { "epoch": 3.0581445061992305, "grad_norm": 2.6720361709594727, "learning_rate": 4.872610570486816e-05, "loss": 0.0946, "num_input_tokens_seen": 52434640, "step": 2685 }, { "epoch": 3.0638449479834686, "grad_norm": 5.53890323638916, "learning_rate": 4.872139770116447e-05, "loss": 0.0566, "num_input_tokens_seen": 52532400, "step": 2690 }, { "epoch": 3.069545389767707, "grad_norm": 0.7934585809707642, "learning_rate": 4.871668124197976e-05, "loss": 0.0163, "num_input_tokens_seen": 52630112, "step": 2695 }, { "epoch": 3.0752458315519453, "grad_norm": 9.233660697937012, "learning_rate": 4.871195632899518e-05, "loss": 0.0552, "num_input_tokens_seen": 52727840, "step": 2700 }, { "epoch": 3.0809462733361834, "grad_norm": 5.545113563537598, "learning_rate": 4.870722296389495e-05, "loss": 0.0711, "num_input_tokens_seen": 52825600, "step": 2705 }, { "epoch": 3.086646715120422, "grad_norm": 8.869247436523438, "learning_rate": 4.870248114836626e-05, "loss": 0.1192, "num_input_tokens_seen": 52923312, "step": 2710 }, { "epoch": 3.09234715690466, "grad_norm": 2.2247767448425293, "learning_rate": 4.8697730884099334e-05, "loss": 0.0258, "num_input_tokens_seen": 53020928, "step": 2715 }, { "epoch": 3.0980475986888982, "grad_norm": 0.8696161508560181, "learning_rate": 4.8692972172787396e-05, "loss": 0.0649, "num_input_tokens_seen": 53118720, "step": 2720 }, { "epoch": 3.103748040473137, "grad_norm": 6.550947189331055, "learning_rate": 4.86882050161267e-05, "loss": 0.0605, "num_input_tokens_seen": 53216512, "step": 2725 }, { "epoch": 3.109448482257375, "grad_norm": 5.182213306427002, "learning_rate": 4.8683429415816485e-05, "loss": 0.0933, "num_input_tokens_seen": 53314224, "step": 2730 }, { "epoch": 3.115148924041613, "grad_norm": 1.667688012123108, "learning_rate": 4.867864537355901e-05, "loss": 0.0777, "num_input_tokens_seen": 53411936, "step": 2735 }, { "epoch": 3.1208493658258516, "grad_norm": 10.697488784790039, "learning_rate": 4.867385289105955e-05, "loss": 0.1207, "num_input_tokens_seen": 53509664, "step": 2740 }, { "epoch": 3.1265498076100897, "grad_norm": 5.032103061676025, "learning_rate": 4.866905197002637e-05, "loss": 0.064, "num_input_tokens_seen": 53607408, "step": 2745 }, { "epoch": 3.1322502493943283, "grad_norm": 6.669914722442627, "learning_rate": 4.866424261217078e-05, "loss": 0.0425, "num_input_tokens_seen": 53705216, "step": 2750 }, { "epoch": 3.1379506911785664, "grad_norm": 3.654059886932373, "learning_rate": 4.865942481920706e-05, "loss": 0.0541, "num_input_tokens_seen": 53802960, "step": 2755 }, { "epoch": 3.1436511329628045, "grad_norm": 3.9376277923583984, "learning_rate": 4.865459859285251e-05, "loss": 0.0352, "num_input_tokens_seen": 53900720, "step": 2760 }, { "epoch": 3.149351574747043, "grad_norm": 3.594050168991089, "learning_rate": 4.864976393482743e-05, "loss": 0.0372, "num_input_tokens_seen": 53998384, "step": 2765 }, { "epoch": 3.155052016531281, "grad_norm": 5.50773811340332, "learning_rate": 4.864492084685514e-05, "loss": 0.0612, "num_input_tokens_seen": 54096144, "step": 2770 }, { "epoch": 3.1607524583155193, "grad_norm": 11.46947193145752, "learning_rate": 4.864006933066196e-05, "loss": 0.0896, "num_input_tokens_seen": 54193840, "step": 2775 }, { "epoch": 3.166452900099758, "grad_norm": 9.24869155883789, "learning_rate": 4.8635209387977197e-05, "loss": 0.0575, "num_input_tokens_seen": 54291568, "step": 2780 }, { "epoch": 3.172153341883996, "grad_norm": 5.757988929748535, "learning_rate": 4.8630341020533196e-05, "loss": 0.0832, "num_input_tokens_seen": 54389248, "step": 2785 }, { "epoch": 3.177853783668234, "grad_norm": 6.639283657073975, "learning_rate": 4.862546423006527e-05, "loss": 0.0882, "num_input_tokens_seen": 54486944, "step": 2790 }, { "epoch": 3.1835542254524727, "grad_norm": 5.969208240509033, "learning_rate": 4.8620579018311744e-05, "loss": 0.0486, "num_input_tokens_seen": 54584624, "step": 2795 }, { "epoch": 3.189254667236711, "grad_norm": 11.107736587524414, "learning_rate": 4.8615685387013956e-05, "loss": 0.0754, "num_input_tokens_seen": 54682384, "step": 2800 }, { "epoch": 3.194955109020949, "grad_norm": 9.802680969238281, "learning_rate": 4.861078333791624e-05, "loss": 0.0721, "num_input_tokens_seen": 54780160, "step": 2805 }, { "epoch": 3.2006555508051875, "grad_norm": 2.6495935916900635, "learning_rate": 4.860587287276592e-05, "loss": 0.0538, "num_input_tokens_seen": 54877872, "step": 2810 }, { "epoch": 3.2063559925894256, "grad_norm": 5.3208818435668945, "learning_rate": 4.8600953993313344e-05, "loss": 0.0571, "num_input_tokens_seen": 54975632, "step": 2815 }, { "epoch": 3.2120564343736637, "grad_norm": 5.696016311645508, "learning_rate": 4.859602670131185e-05, "loss": 0.0616, "num_input_tokens_seen": 55073408, "step": 2820 }, { "epoch": 3.2177568761579023, "grad_norm": 9.000017166137695, "learning_rate": 4.859109099851774e-05, "loss": 0.1114, "num_input_tokens_seen": 55171152, "step": 2825 }, { "epoch": 3.2234573179421404, "grad_norm": 6.167779922485352, "learning_rate": 4.8586146886690364e-05, "loss": 0.0335, "num_input_tokens_seen": 55268896, "step": 2830 }, { "epoch": 3.229157759726379, "grad_norm": 0.7552993893623352, "learning_rate": 4.8581194367592043e-05, "loss": 0.0157, "num_input_tokens_seen": 55366688, "step": 2835 }, { "epoch": 3.234858201510617, "grad_norm": 6.548010349273682, "learning_rate": 4.8576233442988095e-05, "loss": 0.0572, "num_input_tokens_seen": 55464368, "step": 2840 }, { "epoch": 3.240558643294855, "grad_norm": 0.6461604237556458, "learning_rate": 4.857126411464685e-05, "loss": 0.0241, "num_input_tokens_seen": 55562128, "step": 2845 }, { "epoch": 3.2462590850790938, "grad_norm": 7.866938591003418, "learning_rate": 4.856628638433962e-05, "loss": 0.0597, "num_input_tokens_seen": 55659792, "step": 2850 }, { "epoch": 3.251959526863332, "grad_norm": 5.226189136505127, "learning_rate": 4.85613002538407e-05, "loss": 0.0267, "num_input_tokens_seen": 55757504, "step": 2855 }, { "epoch": 3.25765996864757, "grad_norm": 6.863353252410889, "learning_rate": 4.855630572492742e-05, "loss": 0.0537, "num_input_tokens_seen": 55855344, "step": 2860 }, { "epoch": 3.2633604104318086, "grad_norm": 1.295962929725647, "learning_rate": 4.8551302799380055e-05, "loss": 0.0304, "num_input_tokens_seen": 55953072, "step": 2865 }, { "epoch": 3.2690608522160467, "grad_norm": 5.298805236816406, "learning_rate": 4.854629147898191e-05, "loss": 0.0321, "num_input_tokens_seen": 56050752, "step": 2870 }, { "epoch": 3.2747612940002853, "grad_norm": 12.122303009033203, "learning_rate": 4.854127176551925e-05, "loss": 0.1434, "num_input_tokens_seen": 56148560, "step": 2875 }, { "epoch": 3.2804617357845234, "grad_norm": 1.2280305624008179, "learning_rate": 4.8536243660781375e-05, "loss": 0.0707, "num_input_tokens_seen": 56246272, "step": 2880 }, { "epoch": 3.2861621775687615, "grad_norm": 5.140838623046875, "learning_rate": 4.8531207166560524e-05, "loss": 0.0457, "num_input_tokens_seen": 56343984, "step": 2885 }, { "epoch": 3.291862619353, "grad_norm": 2.1565487384796143, "learning_rate": 4.8526162284651974e-05, "loss": 0.0177, "num_input_tokens_seen": 56441792, "step": 2890 }, { "epoch": 3.297563061137238, "grad_norm": 2.832627534866333, "learning_rate": 4.852110901685396e-05, "loss": 0.0283, "num_input_tokens_seen": 56539600, "step": 2895 }, { "epoch": 3.3032635029214763, "grad_norm": 10.345179557800293, "learning_rate": 4.851604736496772e-05, "loss": 0.0475, "num_input_tokens_seen": 56637280, "step": 2900 }, { "epoch": 3.308963944705715, "grad_norm": 0.8519467711448669, "learning_rate": 4.8510977330797476e-05, "loss": 0.0266, "num_input_tokens_seen": 56735056, "step": 2905 }, { "epoch": 3.314664386489953, "grad_norm": 6.070542335510254, "learning_rate": 4.8505898916150436e-05, "loss": 0.0536, "num_input_tokens_seen": 56832864, "step": 2910 }, { "epoch": 3.320364828274191, "grad_norm": 3.4217629432678223, "learning_rate": 4.85008121228368e-05, "loss": 0.0251, "num_input_tokens_seen": 56930608, "step": 2915 }, { "epoch": 3.3260652700584297, "grad_norm": 5.251518726348877, "learning_rate": 4.849571695266977e-05, "loss": 0.0676, "num_input_tokens_seen": 57028336, "step": 2920 }, { "epoch": 3.331765711842668, "grad_norm": 9.286744117736816, "learning_rate": 4.849061340746549e-05, "loss": 0.1008, "num_input_tokens_seen": 57126128, "step": 2925 }, { "epoch": 3.337466153626906, "grad_norm": 5.496871471405029, "learning_rate": 4.848550148904314e-05, "loss": 0.1098, "num_input_tokens_seen": 57223840, "step": 2930 }, { "epoch": 3.3431665954111445, "grad_norm": 6.865820407867432, "learning_rate": 4.848038119922483e-05, "loss": 0.0545, "num_input_tokens_seen": 57321568, "step": 2935 }, { "epoch": 3.3488670371953826, "grad_norm": 4.949888229370117, "learning_rate": 4.847525253983572e-05, "loss": 0.1271, "num_input_tokens_seen": 57419328, "step": 2940 }, { "epoch": 3.3545674789796207, "grad_norm": 0.7395240068435669, "learning_rate": 4.847011551270391e-05, "loss": 0.0262, "num_input_tokens_seen": 57517008, "step": 2945 }, { "epoch": 3.3602679207638593, "grad_norm": 10.39633560180664, "learning_rate": 4.846497011966047e-05, "loss": 0.0333, "num_input_tokens_seen": 57614816, "step": 2950 }, { "epoch": 3.3659683625480974, "grad_norm": 5.424074649810791, "learning_rate": 4.845981636253949e-05, "loss": 0.066, "num_input_tokens_seen": 57712528, "step": 2955 }, { "epoch": 3.3716688043323355, "grad_norm": 5.526705265045166, "learning_rate": 4.845465424317802e-05, "loss": 0.0246, "num_input_tokens_seen": 57810208, "step": 2960 }, { "epoch": 3.377369246116574, "grad_norm": 3.3858978748321533, "learning_rate": 4.8449483763416095e-05, "loss": 0.0585, "num_input_tokens_seen": 57907968, "step": 2965 }, { "epoch": 3.383069687900812, "grad_norm": 4.47909688949585, "learning_rate": 4.844430492509674e-05, "loss": 0.0799, "num_input_tokens_seen": 58005744, "step": 2970 }, { "epoch": 3.3887701296850508, "grad_norm": 8.794025421142578, "learning_rate": 4.843911773006593e-05, "loss": 0.0286, "num_input_tokens_seen": 58103504, "step": 2975 }, { "epoch": 3.394470571469289, "grad_norm": 5.554230690002441, "learning_rate": 4.8433922180172653e-05, "loss": 0.0499, "num_input_tokens_seen": 58201232, "step": 2980 }, { "epoch": 3.400171013253527, "grad_norm": 4.487776279449463, "learning_rate": 4.842871827726886e-05, "loss": 0.0402, "num_input_tokens_seen": 58299024, "step": 2985 }, { "epoch": 3.4058714550377656, "grad_norm": 2.8140945434570312, "learning_rate": 4.8423506023209466e-05, "loss": 0.0566, "num_input_tokens_seen": 58396816, "step": 2990 }, { "epoch": 3.4115718968220037, "grad_norm": 3.601980686187744, "learning_rate": 4.8418285419852395e-05, "loss": 0.0412, "num_input_tokens_seen": 58494544, "step": 2995 }, { "epoch": 3.417272338606242, "grad_norm": 2.136195182800293, "learning_rate": 4.841305646905851e-05, "loss": 0.0304, "num_input_tokens_seen": 58592352, "step": 3000 }, { "epoch": 3.4229727803904804, "grad_norm": 5.654057502746582, "learning_rate": 4.8407819172691694e-05, "loss": 0.0304, "num_input_tokens_seen": 58690128, "step": 3005 }, { "epoch": 3.4286732221747185, "grad_norm": 2.4083597660064697, "learning_rate": 4.840257353261875e-05, "loss": 0.0383, "num_input_tokens_seen": 58787904, "step": 3010 }, { "epoch": 3.434373663958957, "grad_norm": 5.336053371429443, "learning_rate": 4.83973195507095e-05, "loss": 0.0915, "num_input_tokens_seen": 58885632, "step": 3015 }, { "epoch": 3.440074105743195, "grad_norm": 4.11752986907959, "learning_rate": 4.839205722883672e-05, "loss": 0.0503, "num_input_tokens_seen": 58983312, "step": 3020 }, { "epoch": 3.4457745475274333, "grad_norm": 13.777847290039062, "learning_rate": 4.838678656887616e-05, "loss": 0.1445, "num_input_tokens_seen": 59081072, "step": 3025 }, { "epoch": 3.451474989311672, "grad_norm": 9.075989723205566, "learning_rate": 4.838150757270655e-05, "loss": 0.0777, "num_input_tokens_seen": 59178896, "step": 3030 }, { "epoch": 3.45717543109591, "grad_norm": 7.0720014572143555, "learning_rate": 4.837622024220959e-05, "loss": 0.0592, "num_input_tokens_seen": 59276560, "step": 3035 }, { "epoch": 3.462875872880148, "grad_norm": 4.558810710906982, "learning_rate": 4.837092457926993e-05, "loss": 0.0274, "num_input_tokens_seen": 59374368, "step": 3040 }, { "epoch": 3.4685763146643867, "grad_norm": 14.200141906738281, "learning_rate": 4.8365620585775214e-05, "loss": 0.0558, "num_input_tokens_seen": 59472048, "step": 3045 }, { "epoch": 3.4742767564486248, "grad_norm": 5.859817028045654, "learning_rate": 4.836030826361605e-05, "loss": 0.0277, "num_input_tokens_seen": 59569840, "step": 3050 }, { "epoch": 3.479977198232863, "grad_norm": 8.385420799255371, "learning_rate": 4.835498761468601e-05, "loss": 0.0667, "num_input_tokens_seen": 59667584, "step": 3055 }, { "epoch": 3.4856776400171015, "grad_norm": 1.2888391017913818, "learning_rate": 4.834965864088164e-05, "loss": 0.0207, "num_input_tokens_seen": 59765392, "step": 3060 }, { "epoch": 3.4913780818013396, "grad_norm": 1.1023948192596436, "learning_rate": 4.834432134410245e-05, "loss": 0.0207, "num_input_tokens_seen": 59863152, "step": 3065 }, { "epoch": 3.4970785235855777, "grad_norm": 3.3756585121154785, "learning_rate": 4.8338975726250925e-05, "loss": 0.0416, "num_input_tokens_seen": 59960928, "step": 3070 }, { "epoch": 3.5027789653698163, "grad_norm": 0.9311105608940125, "learning_rate": 4.833362178923249e-05, "loss": 0.0316, "num_input_tokens_seen": 60058656, "step": 3075 }, { "epoch": 3.5084794071540544, "grad_norm": 10.324899673461914, "learning_rate": 4.8328259534955554e-05, "loss": 0.0793, "num_input_tokens_seen": 60156448, "step": 3080 }, { "epoch": 3.5141798489382925, "grad_norm": 4.7765703201293945, "learning_rate": 4.832288896533151e-05, "loss": 0.0476, "num_input_tokens_seen": 60254192, "step": 3085 }, { "epoch": 3.519880290722531, "grad_norm": 1.959538459777832, "learning_rate": 4.831751008227468e-05, "loss": 0.0346, "num_input_tokens_seen": 60351920, "step": 3090 }, { "epoch": 3.525580732506769, "grad_norm": 9.76518440246582, "learning_rate": 4.831212288770237e-05, "loss": 0.046, "num_input_tokens_seen": 60449696, "step": 3095 }, { "epoch": 3.5312811742910073, "grad_norm": 1.2289072275161743, "learning_rate": 4.8306727383534835e-05, "loss": 0.0225, "num_input_tokens_seen": 60547440, "step": 3100 }, { "epoch": 3.536981616075246, "grad_norm": 7.658115863800049, "learning_rate": 4.8301323571695314e-05, "loss": 0.0281, "num_input_tokens_seen": 60645200, "step": 3105 }, { "epoch": 3.542682057859484, "grad_norm": 4.308380126953125, "learning_rate": 4.829591145410997e-05, "loss": 0.0265, "num_input_tokens_seen": 60742880, "step": 3110 }, { "epoch": 3.5483824996437225, "grad_norm": 4.51566743850708, "learning_rate": 4.829049103270798e-05, "loss": 0.0473, "num_input_tokens_seen": 60840640, "step": 3115 }, { "epoch": 3.5540829414279607, "grad_norm": 4.3482255935668945, "learning_rate": 4.8285062309421426e-05, "loss": 0.0468, "num_input_tokens_seen": 60938400, "step": 3120 }, { "epoch": 3.559783383212199, "grad_norm": 7.6800994873046875, "learning_rate": 4.827962528618538e-05, "loss": 0.0282, "num_input_tokens_seen": 61036128, "step": 3125 }, { "epoch": 3.5654838249964373, "grad_norm": 8.757813453674316, "learning_rate": 4.8274179964937875e-05, "loss": 0.0225, "num_input_tokens_seen": 61133872, "step": 3130 }, { "epoch": 3.5711842667806755, "grad_norm": 1.4490429162979126, "learning_rate": 4.826872634761989e-05, "loss": 0.0375, "num_input_tokens_seen": 61231600, "step": 3135 }, { "epoch": 3.576884708564914, "grad_norm": 5.913198471069336, "learning_rate": 4.826326443617536e-05, "loss": 0.0422, "num_input_tokens_seen": 61329360, "step": 3140 }, { "epoch": 3.582585150349152, "grad_norm": 8.622357368469238, "learning_rate": 4.825779423255118e-05, "loss": 0.0399, "num_input_tokens_seen": 61427104, "step": 3145 }, { "epoch": 3.5882855921333903, "grad_norm": 6.383512496948242, "learning_rate": 4.825231573869721e-05, "loss": 0.0356, "num_input_tokens_seen": 61524848, "step": 3150 }, { "epoch": 3.593986033917629, "grad_norm": 15.792478561401367, "learning_rate": 4.824682895656624e-05, "loss": 0.0613, "num_input_tokens_seen": 61622512, "step": 3155 }, { "epoch": 3.599686475701867, "grad_norm": 1.2860291004180908, "learning_rate": 4.824133388811405e-05, "loss": 0.0439, "num_input_tokens_seen": 61720192, "step": 3160 }, { "epoch": 3.605386917486105, "grad_norm": 6.301830768585205, "learning_rate": 4.823583053529934e-05, "loss": 0.0353, "num_input_tokens_seen": 61817936, "step": 3165 }, { "epoch": 3.6110873592703436, "grad_norm": 4.263798236846924, "learning_rate": 4.823031890008379e-05, "loss": 0.0338, "num_input_tokens_seen": 61915664, "step": 3170 }, { "epoch": 3.6167878010545818, "grad_norm": 7.392456531524658, "learning_rate": 4.8224798984432005e-05, "loss": 0.0399, "num_input_tokens_seen": 62013456, "step": 3175 }, { "epoch": 3.62248824283882, "grad_norm": 2.850409746170044, "learning_rate": 4.8219270790311575e-05, "loss": 0.0422, "num_input_tokens_seen": 62111248, "step": 3180 }, { "epoch": 3.6281886846230584, "grad_norm": 3.5166022777557373, "learning_rate": 4.8213734319693004e-05, "loss": 0.0193, "num_input_tokens_seen": 62208960, "step": 3185 }, { "epoch": 3.6338891264072966, "grad_norm": 7.699153423309326, "learning_rate": 4.820818957454978e-05, "loss": 0.0698, "num_input_tokens_seen": 62306592, "step": 3190 }, { "epoch": 3.6395895681915347, "grad_norm": 0.7717591524124146, "learning_rate": 4.820263655685831e-05, "loss": 0.0257, "num_input_tokens_seen": 62404400, "step": 3195 }, { "epoch": 3.6452900099757732, "grad_norm": 6.028016567230225, "learning_rate": 4.819707526859797e-05, "loss": 0.0352, "num_input_tokens_seen": 62502160, "step": 3200 }, { "epoch": 3.6509904517600114, "grad_norm": 2.3986012935638428, "learning_rate": 4.819150571175108e-05, "loss": 0.043, "num_input_tokens_seen": 62599920, "step": 3205 }, { "epoch": 3.6566908935442495, "grad_norm": 3.4287400245666504, "learning_rate": 4.818592788830291e-05, "loss": 0.0289, "num_input_tokens_seen": 62697680, "step": 3210 }, { "epoch": 3.662391335328488, "grad_norm": 5.921146869659424, "learning_rate": 4.818034180024167e-05, "loss": 0.0331, "num_input_tokens_seen": 62795472, "step": 3215 }, { "epoch": 3.668091777112726, "grad_norm": 4.856356620788574, "learning_rate": 4.8174747449558515e-05, "loss": 0.0131, "num_input_tokens_seen": 62893136, "step": 3220 }, { "epoch": 3.6737922188969643, "grad_norm": 6.656949996948242, "learning_rate": 4.816914483824755e-05, "loss": 0.0426, "num_input_tokens_seen": 62990816, "step": 3225 }, { "epoch": 3.679492660681203, "grad_norm": 1.0884100198745728, "learning_rate": 4.816353396830583e-05, "loss": 0.032, "num_input_tokens_seen": 63088560, "step": 3230 }, { "epoch": 3.685193102465441, "grad_norm": 0.37009307742118835, "learning_rate": 4.815791484173333e-05, "loss": 0.0322, "num_input_tokens_seen": 63186272, "step": 3235 }, { "epoch": 3.690893544249679, "grad_norm": 2.093526840209961, "learning_rate": 4.815228746053301e-05, "loss": 0.0225, "num_input_tokens_seen": 63284016, "step": 3240 }, { "epoch": 3.6965939860339176, "grad_norm": 9.629427909851074, "learning_rate": 4.814665182671072e-05, "loss": 0.0321, "num_input_tokens_seen": 63381776, "step": 3245 }, { "epoch": 3.7022944278181558, "grad_norm": 7.924525260925293, "learning_rate": 4.8141007942275295e-05, "loss": 0.0641, "num_input_tokens_seen": 63479536, "step": 3250 }, { "epoch": 3.7079948696023943, "grad_norm": 3.5611679553985596, "learning_rate": 4.813535580923849e-05, "loss": 0.0731, "num_input_tokens_seen": 63577152, "step": 3255 }, { "epoch": 3.7136953113866324, "grad_norm": 0.575011670589447, "learning_rate": 4.812969542961502e-05, "loss": 0.0453, "num_input_tokens_seen": 63674928, "step": 3260 }, { "epoch": 3.719395753170871, "grad_norm": 5.894010066986084, "learning_rate": 4.8124026805422494e-05, "loss": 0.0257, "num_input_tokens_seen": 63772640, "step": 3265 }, { "epoch": 3.725096194955109, "grad_norm": 3.0350735187530518, "learning_rate": 4.811834993868152e-05, "loss": 0.0338, "num_input_tokens_seen": 63870336, "step": 3270 }, { "epoch": 3.7307966367393472, "grad_norm": 8.058395385742188, "learning_rate": 4.81126648314156e-05, "loss": 0.0421, "num_input_tokens_seen": 63968160, "step": 3275 }, { "epoch": 3.736497078523586, "grad_norm": 9.93237590789795, "learning_rate": 4.81069714856512e-05, "loss": 0.0448, "num_input_tokens_seen": 64065904, "step": 3280 }, { "epoch": 3.742197520307824, "grad_norm": 11.603642463684082, "learning_rate": 4.810126990341769e-05, "loss": 0.0901, "num_input_tokens_seen": 64163616, "step": 3285 }, { "epoch": 3.747897962092062, "grad_norm": 3.8158483505249023, "learning_rate": 4.809556008674741e-05, "loss": 0.0154, "num_input_tokens_seen": 64261376, "step": 3290 }, { "epoch": 3.7535984038763006, "grad_norm": 0.4274216890335083, "learning_rate": 4.8089842037675615e-05, "loss": 0.0094, "num_input_tokens_seen": 64359072, "step": 3295 }, { "epoch": 3.7592988456605387, "grad_norm": 4.152562618255615, "learning_rate": 4.808411575824051e-05, "loss": 0.0443, "num_input_tokens_seen": 64456816, "step": 3300 }, { "epoch": 3.764999287444777, "grad_norm": 4.328752040863037, "learning_rate": 4.807838125048322e-05, "loss": 0.0393, "num_input_tokens_seen": 64554464, "step": 3305 }, { "epoch": 3.7706997292290154, "grad_norm": 4.978052616119385, "learning_rate": 4.80726385164478e-05, "loss": 0.0324, "num_input_tokens_seen": 64652272, "step": 3310 }, { "epoch": 3.7764001710132535, "grad_norm": 6.3277082443237305, "learning_rate": 4.8066887558181265e-05, "loss": 0.0203, "num_input_tokens_seen": 64750016, "step": 3315 }, { "epoch": 3.7821006127974917, "grad_norm": 0.5800598859786987, "learning_rate": 4.806112837773351e-05, "loss": 0.015, "num_input_tokens_seen": 64847760, "step": 3320 }, { "epoch": 3.78780105458173, "grad_norm": 17.387359619140625, "learning_rate": 4.8055360977157426e-05, "loss": 0.0503, "num_input_tokens_seen": 64945504, "step": 3325 }, { "epoch": 3.7935014963659683, "grad_norm": 6.007382392883301, "learning_rate": 4.8049585358508776e-05, "loss": 0.0294, "num_input_tokens_seen": 65043232, "step": 3330 }, { "epoch": 3.7992019381502065, "grad_norm": 8.47810173034668, "learning_rate": 4.804380152384629e-05, "loss": 0.044, "num_input_tokens_seen": 65141024, "step": 3335 }, { "epoch": 3.804902379934445, "grad_norm": 9.82911491394043, "learning_rate": 4.8038009475231604e-05, "loss": 0.0369, "num_input_tokens_seen": 65238752, "step": 3340 }, { "epoch": 3.810602821718683, "grad_norm": 13.116619110107422, "learning_rate": 4.80322092147293e-05, "loss": 0.0289, "num_input_tokens_seen": 65336528, "step": 3345 }, { "epoch": 3.8163032635029213, "grad_norm": 1.19611656665802, "learning_rate": 4.802640074440686e-05, "loss": 0.0214, "num_input_tokens_seen": 65434272, "step": 3350 }, { "epoch": 3.82200370528716, "grad_norm": 0.3276759386062622, "learning_rate": 4.802058406633474e-05, "loss": 0.0193, "num_input_tokens_seen": 65532064, "step": 3355 }, { "epoch": 3.827704147071398, "grad_norm": 6.492347240447998, "learning_rate": 4.8014759182586274e-05, "loss": 0.0542, "num_input_tokens_seen": 65629792, "step": 3360 }, { "epoch": 3.833404588855636, "grad_norm": 3.1319868564605713, "learning_rate": 4.800892609523774e-05, "loss": 0.0361, "num_input_tokens_seen": 65727536, "step": 3365 }, { "epoch": 3.8391050306398746, "grad_norm": 0.28512752056121826, "learning_rate": 4.8003084806368336e-05, "loss": 0.0299, "num_input_tokens_seen": 65825200, "step": 3370 }, { "epoch": 3.8448054724241127, "grad_norm": 1.0629769563674927, "learning_rate": 4.7997235318060185e-05, "loss": 0.0643, "num_input_tokens_seen": 65922976, "step": 3375 }, { "epoch": 3.8505059142083513, "grad_norm": 9.550495147705078, "learning_rate": 4.799137763239835e-05, "loss": 0.024, "num_input_tokens_seen": 66020656, "step": 3380 }, { "epoch": 3.8562063559925894, "grad_norm": 5.962581157684326, "learning_rate": 4.798551175147079e-05, "loss": 0.0279, "num_input_tokens_seen": 66118384, "step": 3385 }, { "epoch": 3.8619067977768275, "grad_norm": 2.609731435775757, "learning_rate": 4.79796376773684e-05, "loss": 0.0399, "num_input_tokens_seen": 66216176, "step": 3390 }, { "epoch": 3.867607239561066, "grad_norm": 5.378483772277832, "learning_rate": 4.797375541218498e-05, "loss": 0.0118, "num_input_tokens_seen": 66313872, "step": 3395 }, { "epoch": 3.8733076813453042, "grad_norm": 7.734043598175049, "learning_rate": 4.796786495801727e-05, "loss": 0.0262, "num_input_tokens_seen": 66411664, "step": 3400 }, { "epoch": 3.879008123129543, "grad_norm": 7.185009479522705, "learning_rate": 4.796196631696491e-05, "loss": 0.0313, "num_input_tokens_seen": 66509440, "step": 3405 }, { "epoch": 3.884708564913781, "grad_norm": 4.7586164474487305, "learning_rate": 4.795605949113049e-05, "loss": 0.0137, "num_input_tokens_seen": 66607152, "step": 3410 }, { "epoch": 3.890409006698019, "grad_norm": 0.9074054956436157, "learning_rate": 4.795014448261947e-05, "loss": 0.0263, "num_input_tokens_seen": 66704880, "step": 3415 }, { "epoch": 3.8961094484822576, "grad_norm": 2.6224796772003174, "learning_rate": 4.794422129354026e-05, "loss": 0.0146, "num_input_tokens_seen": 66802656, "step": 3420 }, { "epoch": 3.9018098902664957, "grad_norm": 0.855692982673645, "learning_rate": 4.7938289926004185e-05, "loss": 0.0078, "num_input_tokens_seen": 66900480, "step": 3425 }, { "epoch": 3.907510332050734, "grad_norm": 1.3807679414749146, "learning_rate": 4.793235038212548e-05, "loss": 0.0188, "num_input_tokens_seen": 66998304, "step": 3430 }, { "epoch": 3.9132107738349724, "grad_norm": 0.8240529298782349, "learning_rate": 4.7926402664021275e-05, "loss": 0.0576, "num_input_tokens_seen": 67096000, "step": 3435 }, { "epoch": 3.9189112156192105, "grad_norm": 7.201174736022949, "learning_rate": 4.792044677381165e-05, "loss": 0.0205, "num_input_tokens_seen": 67193680, "step": 3440 }, { "epoch": 3.9246116574034486, "grad_norm": 10.291589736938477, "learning_rate": 4.791448271361957e-05, "loss": 0.0524, "num_input_tokens_seen": 67291472, "step": 3445 }, { "epoch": 3.930312099187687, "grad_norm": 3.4942891597747803, "learning_rate": 4.7908510485570925e-05, "loss": 0.0652, "num_input_tokens_seen": 67389216, "step": 3450 }, { "epoch": 3.9360125409719253, "grad_norm": 2.04127836227417, "learning_rate": 4.7902530091794505e-05, "loss": 0.0356, "num_input_tokens_seen": 67486912, "step": 3455 }, { "epoch": 3.9417129827561634, "grad_norm": 4.031794548034668, "learning_rate": 4.789654153442203e-05, "loss": 0.0419, "num_input_tokens_seen": 67584624, "step": 3460 }, { "epoch": 3.947413424540402, "grad_norm": 8.670853614807129, "learning_rate": 4.7890544815588115e-05, "loss": 0.0192, "num_input_tokens_seen": 67682320, "step": 3465 }, { "epoch": 3.95311386632464, "grad_norm": 7.351383686065674, "learning_rate": 4.788453993743028e-05, "loss": 0.0361, "num_input_tokens_seen": 67780064, "step": 3470 }, { "epoch": 3.9588143081088782, "grad_norm": 8.677574157714844, "learning_rate": 4.787852690208897e-05, "loss": 0.0235, "num_input_tokens_seen": 67877792, "step": 3475 }, { "epoch": 3.964514749893117, "grad_norm": 9.356419563293457, "learning_rate": 4.787250571170752e-05, "loss": 0.0572, "num_input_tokens_seen": 67975472, "step": 3480 }, { "epoch": 3.970215191677355, "grad_norm": 6.9926300048828125, "learning_rate": 4.786647636843219e-05, "loss": 0.0837, "num_input_tokens_seen": 68073200, "step": 3485 }, { "epoch": 3.975915633461593, "grad_norm": 4.828823089599609, "learning_rate": 4.786043887441213e-05, "loss": 0.0422, "num_input_tokens_seen": 68170976, "step": 3490 }, { "epoch": 3.9816160752458316, "grad_norm": 11.057583808898926, "learning_rate": 4.785439323179941e-05, "loss": 0.0326, "num_input_tokens_seen": 68268672, "step": 3495 }, { "epoch": 3.9873165170300697, "grad_norm": 0.949934184551239, "learning_rate": 4.784833944274899e-05, "loss": 0.0236, "num_input_tokens_seen": 68366432, "step": 3500 }, { "epoch": 3.993016958814308, "grad_norm": 6.844513416290283, "learning_rate": 4.784227750941873e-05, "loss": 0.0188, "num_input_tokens_seen": 68464128, "step": 3505 }, { "epoch": 3.9987174005985464, "grad_norm": 2.295914649963379, "learning_rate": 4.783620743396943e-05, "loss": 0.0186, "num_input_tokens_seen": 68561936, "step": 3510 }, { "epoch": 4.003420265070543, "grad_norm": 2.7240405082702637, "learning_rate": 4.783012921856474e-05, "loss": 0.0217, "num_input_tokens_seen": 68642496, "step": 3515 }, { "epoch": 4.009120706854781, "grad_norm": 9.799927711486816, "learning_rate": 4.782404286537124e-05, "loss": 0.0442, "num_input_tokens_seen": 68740256, "step": 3520 }, { "epoch": 4.01482114863902, "grad_norm": 4.823245525360107, "learning_rate": 4.781794837655843e-05, "loss": 0.0601, "num_input_tokens_seen": 68837968, "step": 3525 }, { "epoch": 4.020521590423257, "grad_norm": 6.3089213371276855, "learning_rate": 4.781184575429867e-05, "loss": 0.0181, "num_input_tokens_seen": 68935680, "step": 3530 }, { "epoch": 4.026222032207496, "grad_norm": 0.7296550869941711, "learning_rate": 4.780573500076723e-05, "loss": 0.0089, "num_input_tokens_seen": 69033408, "step": 3535 }, { "epoch": 4.0319224739917345, "grad_norm": 7.376620292663574, "learning_rate": 4.77996161181423e-05, "loss": 0.0136, "num_input_tokens_seen": 69131152, "step": 3540 }, { "epoch": 4.037622915775972, "grad_norm": 3.193028211593628, "learning_rate": 4.779348910860494e-05, "loss": 0.0251, "num_input_tokens_seen": 69228800, "step": 3545 }, { "epoch": 4.043323357560211, "grad_norm": 0.5682366490364075, "learning_rate": 4.7787353974339134e-05, "loss": 0.0037, "num_input_tokens_seen": 69326608, "step": 3550 }, { "epoch": 4.049023799344449, "grad_norm": 2.3201677799224854, "learning_rate": 4.778121071753174e-05, "loss": 0.0114, "num_input_tokens_seen": 69424368, "step": 3555 }, { "epoch": 4.054724241128688, "grad_norm": 2.5661370754241943, "learning_rate": 4.7775059340372516e-05, "loss": 0.0177, "num_input_tokens_seen": 69522032, "step": 3560 }, { "epoch": 4.060424682912926, "grad_norm": 0.4603801667690277, "learning_rate": 4.776889984505413e-05, "loss": 0.0249, "num_input_tokens_seen": 69619728, "step": 3565 }, { "epoch": 4.066125124697164, "grad_norm": 3.390105962753296, "learning_rate": 4.776273223377211e-05, "loss": 0.0172, "num_input_tokens_seen": 69717424, "step": 3570 }, { "epoch": 4.071825566481403, "grad_norm": 0.19697071611881256, "learning_rate": 4.7756556508724914e-05, "loss": 0.0153, "num_input_tokens_seen": 69815152, "step": 3575 }, { "epoch": 4.07752600826564, "grad_norm": 8.34150218963623, "learning_rate": 4.7750372672113874e-05, "loss": 0.0209, "num_input_tokens_seen": 69912960, "step": 3580 }, { "epoch": 4.083226450049879, "grad_norm": 0.32752522826194763, "learning_rate": 4.774418072614322e-05, "loss": 0.0138, "num_input_tokens_seen": 70010672, "step": 3585 }, { "epoch": 4.0889268918341175, "grad_norm": 3.4794821739196777, "learning_rate": 4.773798067302005e-05, "loss": 0.0562, "num_input_tokens_seen": 70108448, "step": 3590 }, { "epoch": 4.094627333618355, "grad_norm": 7.853202819824219, "learning_rate": 4.7731772514954384e-05, "loss": 0.0245, "num_input_tokens_seen": 70206144, "step": 3595 }, { "epoch": 4.100327775402594, "grad_norm": 0.33840203285217285, "learning_rate": 4.772555625415912e-05, "loss": 0.0092, "num_input_tokens_seen": 70303872, "step": 3600 }, { "epoch": 4.106028217186832, "grad_norm": 6.381319999694824, "learning_rate": 4.771933189285004e-05, "loss": 0.0101, "num_input_tokens_seen": 70401664, "step": 3605 }, { "epoch": 4.11172865897107, "grad_norm": 1.308600902557373, "learning_rate": 4.771309943324581e-05, "loss": 0.021, "num_input_tokens_seen": 70499408, "step": 3610 }, { "epoch": 4.1174291007553085, "grad_norm": 1.1248642206192017, "learning_rate": 4.7706858877567984e-05, "loss": 0.009, "num_input_tokens_seen": 70597200, "step": 3615 }, { "epoch": 4.123129542539547, "grad_norm": 0.797878623008728, "learning_rate": 4.770061022804102e-05, "loss": 0.0084, "num_input_tokens_seen": 70695008, "step": 3620 }, { "epoch": 4.128829984323785, "grad_norm": 5.1671905517578125, "learning_rate": 4.7694353486892224e-05, "loss": 0.0086, "num_input_tokens_seen": 70792784, "step": 3625 }, { "epoch": 4.134530426108023, "grad_norm": 1.4259310960769653, "learning_rate": 4.7688088656351827e-05, "loss": 0.0137, "num_input_tokens_seen": 70890576, "step": 3630 }, { "epoch": 4.140230867892262, "grad_norm": 0.6012780070304871, "learning_rate": 4.7681815738652916e-05, "loss": 0.0331, "num_input_tokens_seen": 70988352, "step": 3635 }, { "epoch": 4.1459313096765, "grad_norm": 0.923217236995697, "learning_rate": 4.767553473603147e-05, "loss": 0.0235, "num_input_tokens_seen": 71086128, "step": 3640 }, { "epoch": 4.151631751460738, "grad_norm": 0.6401808261871338, "learning_rate": 4.766924565072635e-05, "loss": 0.0056, "num_input_tokens_seen": 71183888, "step": 3645 }, { "epoch": 4.157332193244977, "grad_norm": 4.980163097381592, "learning_rate": 4.7662948484979304e-05, "loss": 0.0124, "num_input_tokens_seen": 71281648, "step": 3650 }, { "epoch": 4.163032635029214, "grad_norm": 0.19848279654979706, "learning_rate": 4.7656643241034946e-05, "loss": 0.0377, "num_input_tokens_seen": 71379440, "step": 3655 }, { "epoch": 4.168733076813453, "grad_norm": 0.8732094168663025, "learning_rate": 4.765032992114078e-05, "loss": 0.0071, "num_input_tokens_seen": 71477216, "step": 3660 }, { "epoch": 4.1744335185976915, "grad_norm": 1.789801001548767, "learning_rate": 4.7644008527547185e-05, "loss": 0.025, "num_input_tokens_seen": 71574992, "step": 3665 }, { "epoch": 4.180133960381929, "grad_norm": 2.9071710109710693, "learning_rate": 4.763767906250742e-05, "loss": 0.0172, "num_input_tokens_seen": 71672800, "step": 3670 }, { "epoch": 4.185834402166168, "grad_norm": 1.144612431526184, "learning_rate": 4.7631341528277615e-05, "loss": 0.0092, "num_input_tokens_seen": 71770512, "step": 3675 }, { "epoch": 4.191534843950406, "grad_norm": 2.124117374420166, "learning_rate": 4.7624995927116794e-05, "loss": 0.0214, "num_input_tokens_seen": 71868240, "step": 3680 }, { "epoch": 4.197235285734644, "grad_norm": 1.037842035293579, "learning_rate": 4.761864226128683e-05, "loss": 0.0173, "num_input_tokens_seen": 71965952, "step": 3685 }, { "epoch": 4.202935727518883, "grad_norm": 16.831375122070312, "learning_rate": 4.761228053305249e-05, "loss": 0.0419, "num_input_tokens_seen": 72063680, "step": 3690 }, { "epoch": 4.208636169303121, "grad_norm": 0.9212270975112915, "learning_rate": 4.76059107446814e-05, "loss": 0.0311, "num_input_tokens_seen": 72161472, "step": 3695 }, { "epoch": 4.21433661108736, "grad_norm": 7.812607288360596, "learning_rate": 4.759953289844409e-05, "loss": 0.0197, "num_input_tokens_seen": 72259120, "step": 3700 }, { "epoch": 4.220037052871597, "grad_norm": 2.24538516998291, "learning_rate": 4.759314699661392e-05, "loss": 0.0068, "num_input_tokens_seen": 72356848, "step": 3705 }, { "epoch": 4.225737494655836, "grad_norm": 1.110796570777893, "learning_rate": 4.758675304146715e-05, "loss": 0.0309, "num_input_tokens_seen": 72454608, "step": 3710 }, { "epoch": 4.2314379364400745, "grad_norm": 0.3017835021018982, "learning_rate": 4.75803510352829e-05, "loss": 0.0124, "num_input_tokens_seen": 72552288, "step": 3715 }, { "epoch": 4.237138378224312, "grad_norm": 3.7277181148529053, "learning_rate": 4.757394098034316e-05, "loss": 0.0754, "num_input_tokens_seen": 72650000, "step": 3720 }, { "epoch": 4.242838820008551, "grad_norm": 8.373753547668457, "learning_rate": 4.756752287893279e-05, "loss": 0.01, "num_input_tokens_seen": 72747856, "step": 3725 }, { "epoch": 4.248539261792789, "grad_norm": 3.710064172744751, "learning_rate": 4.7561096733339526e-05, "loss": 0.0109, "num_input_tokens_seen": 72845600, "step": 3730 }, { "epoch": 4.254239703577027, "grad_norm": 3.204511880874634, "learning_rate": 4.755466254585397e-05, "loss": 0.0271, "num_input_tokens_seen": 72943376, "step": 3735 }, { "epoch": 4.2599401453612655, "grad_norm": 2.9513449668884277, "learning_rate": 4.754822031876957e-05, "loss": 0.0119, "num_input_tokens_seen": 73041168, "step": 3740 }, { "epoch": 4.265640587145504, "grad_norm": 7.3270392417907715, "learning_rate": 4.754177005438266e-05, "loss": 0.0168, "num_input_tokens_seen": 73138832, "step": 3745 }, { "epoch": 4.271341028929742, "grad_norm": 11.671257972717285, "learning_rate": 4.753531175499243e-05, "loss": 0.0544, "num_input_tokens_seen": 73236592, "step": 3750 }, { "epoch": 4.27704147071398, "grad_norm": 2.340949773788452, "learning_rate": 4.7528845422900946e-05, "loss": 0.0058, "num_input_tokens_seen": 73334272, "step": 3755 }, { "epoch": 4.282741912498219, "grad_norm": 6.20223331451416, "learning_rate": 4.7522371060413126e-05, "loss": 0.0166, "num_input_tokens_seen": 73432016, "step": 3760 }, { "epoch": 4.288442354282457, "grad_norm": 2.861288547515869, "learning_rate": 4.751588866983676e-05, "loss": 0.0062, "num_input_tokens_seen": 73529760, "step": 3765 }, { "epoch": 4.294142796066695, "grad_norm": 0.3826698064804077, "learning_rate": 4.750939825348249e-05, "loss": 0.0276, "num_input_tokens_seen": 73627552, "step": 3770 }, { "epoch": 4.299843237850934, "grad_norm": 0.9756613373756409, "learning_rate": 4.7502899813663806e-05, "loss": 0.0052, "num_input_tokens_seen": 73725328, "step": 3775 }, { "epoch": 4.305543679635171, "grad_norm": 0.1972053200006485, "learning_rate": 4.749639335269709e-05, "loss": 0.0078, "num_input_tokens_seen": 73823024, "step": 3780 }, { "epoch": 4.31124412141941, "grad_norm": 3.610668897628784, "learning_rate": 4.748987887290156e-05, "loss": 0.0455, "num_input_tokens_seen": 73920736, "step": 3785 }, { "epoch": 4.3169445632036485, "grad_norm": 2.3730828762054443, "learning_rate": 4.7483356376599305e-05, "loss": 0.0169, "num_input_tokens_seen": 74018448, "step": 3790 }, { "epoch": 4.322645004987886, "grad_norm": 6.0831618309021, "learning_rate": 4.747682586611526e-05, "loss": 0.0107, "num_input_tokens_seen": 74116224, "step": 3795 }, { "epoch": 4.328345446772125, "grad_norm": 8.610361099243164, "learning_rate": 4.747028734377723e-05, "loss": 0.0209, "num_input_tokens_seen": 74214016, "step": 3800 }, { "epoch": 4.334045888556363, "grad_norm": 8.020880699157715, "learning_rate": 4.7463740811915856e-05, "loss": 0.0166, "num_input_tokens_seen": 74311712, "step": 3805 }, { "epoch": 4.339746330340601, "grad_norm": 0.3767450153827667, "learning_rate": 4.745718627286466e-05, "loss": 0.009, "num_input_tokens_seen": 74409504, "step": 3810 }, { "epoch": 4.3454467721248395, "grad_norm": 2.0386180877685547, "learning_rate": 4.7450623728959996e-05, "loss": 0.0143, "num_input_tokens_seen": 74507280, "step": 3815 }, { "epoch": 4.351147213909078, "grad_norm": 8.895707130432129, "learning_rate": 4.744405318254109e-05, "loss": 0.0129, "num_input_tokens_seen": 74604912, "step": 3820 }, { "epoch": 4.356847655693317, "grad_norm": 1.1221814155578613, "learning_rate": 4.743747463594999e-05, "loss": 0.0199, "num_input_tokens_seen": 74702720, "step": 3825 }, { "epoch": 4.362548097477554, "grad_norm": 0.9217889308929443, "learning_rate": 4.7430888091531635e-05, "loss": 0.0065, "num_input_tokens_seen": 74800448, "step": 3830 }, { "epoch": 4.368248539261793, "grad_norm": 1.3824939727783203, "learning_rate": 4.7424293551633785e-05, "loss": 0.0055, "num_input_tokens_seen": 74898160, "step": 3835 }, { "epoch": 4.3739489810460315, "grad_norm": 3.7707972526550293, "learning_rate": 4.741769101860707e-05, "loss": 0.0253, "num_input_tokens_seen": 74995824, "step": 3840 }, { "epoch": 4.379649422830269, "grad_norm": 0.15820211172103882, "learning_rate": 4.7411080494804944e-05, "loss": 0.0075, "num_input_tokens_seen": 75093584, "step": 3845 }, { "epoch": 4.385349864614508, "grad_norm": 13.08034610748291, "learning_rate": 4.7404461982583735e-05, "loss": 0.0158, "num_input_tokens_seen": 75191296, "step": 3850 }, { "epoch": 4.391050306398746, "grad_norm": 0.30530065298080444, "learning_rate": 4.739783548430262e-05, "loss": 0.0131, "num_input_tokens_seen": 75288960, "step": 3855 }, { "epoch": 4.396750748182984, "grad_norm": 0.16144217550754547, "learning_rate": 4.739120100232359e-05, "loss": 0.0319, "num_input_tokens_seen": 75386768, "step": 3860 }, { "epoch": 4.4024511899672225, "grad_norm": 8.027372360229492, "learning_rate": 4.7384558539011515e-05, "loss": 0.0352, "num_input_tokens_seen": 75484464, "step": 3865 }, { "epoch": 4.408151631751461, "grad_norm": 10.079002380371094, "learning_rate": 4.73779080967341e-05, "loss": 0.0151, "num_input_tokens_seen": 75582368, "step": 3870 }, { "epoch": 4.413852073535699, "grad_norm": 1.2833141088485718, "learning_rate": 4.7371249677861886e-05, "loss": 0.0081, "num_input_tokens_seen": 75680112, "step": 3875 }, { "epoch": 4.419552515319937, "grad_norm": 0.9999586343765259, "learning_rate": 4.736458328476826e-05, "loss": 0.0034, "num_input_tokens_seen": 75777840, "step": 3880 }, { "epoch": 4.425252957104176, "grad_norm": 7.5928215980529785, "learning_rate": 4.7357908919829464e-05, "loss": 0.012, "num_input_tokens_seen": 75875648, "step": 3885 }, { "epoch": 4.4309533988884136, "grad_norm": 0.11999927461147308, "learning_rate": 4.735122658542456e-05, "loss": 0.0093, "num_input_tokens_seen": 75973296, "step": 3890 }, { "epoch": 4.436653840672652, "grad_norm": 1.351083755493164, "learning_rate": 4.734453628393548e-05, "loss": 0.0051, "num_input_tokens_seen": 76071088, "step": 3895 }, { "epoch": 4.442354282456891, "grad_norm": 0.8617928624153137, "learning_rate": 4.733783801774696e-05, "loss": 0.0033, "num_input_tokens_seen": 76168848, "step": 3900 }, { "epoch": 4.448054724241128, "grad_norm": 0.1406688243150711, "learning_rate": 4.7331131789246614e-05, "loss": 0.0052, "num_input_tokens_seen": 76266512, "step": 3905 }, { "epoch": 4.453755166025367, "grad_norm": 0.1912948042154312, "learning_rate": 4.7324417600824854e-05, "loss": 0.0074, "num_input_tokens_seen": 76364288, "step": 3910 }, { "epoch": 4.4594556078096055, "grad_norm": 3.995418071746826, "learning_rate": 4.7317695454874964e-05, "loss": 0.0096, "num_input_tokens_seen": 76462016, "step": 3915 }, { "epoch": 4.465156049593843, "grad_norm": 0.7223560214042664, "learning_rate": 4.7310965353793044e-05, "loss": 0.003, "num_input_tokens_seen": 76559792, "step": 3920 }, { "epoch": 4.470856491378082, "grad_norm": 1.0632505416870117, "learning_rate": 4.730422729997804e-05, "loss": 0.035, "num_input_tokens_seen": 76657616, "step": 3925 }, { "epoch": 4.47655693316232, "grad_norm": 1.5412085056304932, "learning_rate": 4.729748129583171e-05, "loss": 0.0377, "num_input_tokens_seen": 76755312, "step": 3930 }, { "epoch": 4.482257374946558, "grad_norm": 1.5897432565689087, "learning_rate": 4.729072734375869e-05, "loss": 0.0166, "num_input_tokens_seen": 76853056, "step": 3935 }, { "epoch": 4.4879578167307965, "grad_norm": 0.3547525405883789, "learning_rate": 4.728396544616641e-05, "loss": 0.0201, "num_input_tokens_seen": 76950784, "step": 3940 }, { "epoch": 4.493658258515035, "grad_norm": 0.5280705094337463, "learning_rate": 4.727719560546514e-05, "loss": 0.0173, "num_input_tokens_seen": 77048592, "step": 3945 }, { "epoch": 4.499358700299274, "grad_norm": 5.78103494644165, "learning_rate": 4.7270417824068e-05, "loss": 0.0107, "num_input_tokens_seen": 77146336, "step": 3950 }, { "epoch": 4.505059142083511, "grad_norm": 5.893618583679199, "learning_rate": 4.726363210439092e-05, "loss": 0.0258, "num_input_tokens_seen": 77244000, "step": 3955 }, { "epoch": 4.51075958386775, "grad_norm": 5.622200965881348, "learning_rate": 4.725683844885266e-05, "loss": 0.0186, "num_input_tokens_seen": 77341856, "step": 3960 }, { "epoch": 4.516460025651988, "grad_norm": 3.430377244949341, "learning_rate": 4.725003685987482e-05, "loss": 0.0095, "num_input_tokens_seen": 77439648, "step": 3965 }, { "epoch": 4.522160467436226, "grad_norm": 16.007429122924805, "learning_rate": 4.724322733988183e-05, "loss": 0.0637, "num_input_tokens_seen": 77537440, "step": 3970 }, { "epoch": 4.527860909220465, "grad_norm": 0.6529415845870972, "learning_rate": 4.7236409891300934e-05, "loss": 0.0133, "num_input_tokens_seen": 77635136, "step": 3975 }, { "epoch": 4.533561351004703, "grad_norm": 0.014979444444179535, "learning_rate": 4.722958451656221e-05, "loss": 0.0353, "num_input_tokens_seen": 77732848, "step": 3980 }, { "epoch": 4.539261792788941, "grad_norm": 0.7652057409286499, "learning_rate": 4.722275121809856e-05, "loss": 0.0204, "num_input_tokens_seen": 77830576, "step": 3985 }, { "epoch": 4.5449622345731795, "grad_norm": 9.353001594543457, "learning_rate": 4.721590999834571e-05, "loss": 0.0329, "num_input_tokens_seen": 77928320, "step": 3990 }, { "epoch": 4.550662676357418, "grad_norm": 0.0680394098162651, "learning_rate": 4.720906085974221e-05, "loss": 0.0065, "num_input_tokens_seen": 78026032, "step": 3995 }, { "epoch": 4.556363118141656, "grad_norm": 2.323220729827881, "learning_rate": 4.720220380472942e-05, "loss": 0.0066, "num_input_tokens_seen": 78123696, "step": 4000 }, { "epoch": 4.562063559925894, "grad_norm": 0.5926280617713928, "learning_rate": 4.719533883575155e-05, "loss": 0.0043, "num_input_tokens_seen": 78221376, "step": 4005 }, { "epoch": 4.567764001710133, "grad_norm": 0.9745510220527649, "learning_rate": 4.7188465955255604e-05, "loss": 0.0147, "num_input_tokens_seen": 78319104, "step": 4010 }, { "epoch": 4.5734644434943705, "grad_norm": 10.33803653717041, "learning_rate": 4.7181585165691437e-05, "loss": 0.0112, "num_input_tokens_seen": 78416816, "step": 4015 }, { "epoch": 4.579164885278609, "grad_norm": 9.621374130249023, "learning_rate": 4.7174696469511674e-05, "loss": 0.0222, "num_input_tokens_seen": 78514656, "step": 4020 }, { "epoch": 4.584865327062848, "grad_norm": 4.345292568206787, "learning_rate": 4.716779986917182e-05, "loss": 0.0084, "num_input_tokens_seen": 78612400, "step": 4025 }, { "epoch": 4.590565768847085, "grad_norm": 1.6130069494247437, "learning_rate": 4.7160895367130125e-05, "loss": 0.0068, "num_input_tokens_seen": 78710256, "step": 4030 }, { "epoch": 4.596266210631324, "grad_norm": 1.0490821599960327, "learning_rate": 4.715398296584773e-05, "loss": 0.0086, "num_input_tokens_seen": 78807936, "step": 4035 }, { "epoch": 4.6019666524155625, "grad_norm": 10.216976165771484, "learning_rate": 4.714706266778854e-05, "loss": 0.0563, "num_input_tokens_seen": 78905744, "step": 4040 }, { "epoch": 4.6076670941998, "grad_norm": 0.1510315239429474, "learning_rate": 4.7140134475419304e-05, "loss": 0.0195, "num_input_tokens_seen": 79003584, "step": 4045 }, { "epoch": 4.613367535984039, "grad_norm": 0.32562801241874695, "learning_rate": 4.7133198391209566e-05, "loss": 0.0103, "num_input_tokens_seen": 79101408, "step": 4050 }, { "epoch": 4.619067977768277, "grad_norm": 1.025244116783142, "learning_rate": 4.7126254417631686e-05, "loss": 0.0022, "num_input_tokens_seen": 79199136, "step": 4055 }, { "epoch": 4.624768419552515, "grad_norm": 0.8182079195976257, "learning_rate": 4.7119302557160844e-05, "loss": 0.0032, "num_input_tokens_seen": 79296832, "step": 4060 }, { "epoch": 4.6304688613367535, "grad_norm": 0.6801844835281372, "learning_rate": 4.7112342812275026e-05, "loss": 0.012, "num_input_tokens_seen": 79394528, "step": 4065 }, { "epoch": 4.636169303120992, "grad_norm": 0.7363019585609436, "learning_rate": 4.7105375185455034e-05, "loss": 0.0055, "num_input_tokens_seen": 79492352, "step": 4070 }, { "epoch": 4.641869744905231, "grad_norm": 4.005695819854736, "learning_rate": 4.709839967918447e-05, "loss": 0.0195, "num_input_tokens_seen": 79590064, "step": 4075 }, { "epoch": 4.647570186689468, "grad_norm": 0.4576650857925415, "learning_rate": 4.709141629594975e-05, "loss": 0.0074, "num_input_tokens_seen": 79687856, "step": 4080 }, { "epoch": 4.653270628473707, "grad_norm": 0.05043329671025276, "learning_rate": 4.708442503824011e-05, "loss": 0.0175, "num_input_tokens_seen": 79785600, "step": 4085 }, { "epoch": 4.6589710702579445, "grad_norm": 0.19624769687652588, "learning_rate": 4.707742590854756e-05, "loss": 0.0029, "num_input_tokens_seen": 79883424, "step": 4090 }, { "epoch": 4.664671512042183, "grad_norm": 0.46751806139945984, "learning_rate": 4.7070418909366954e-05, "loss": 0.0192, "num_input_tokens_seen": 79981152, "step": 4095 }, { "epoch": 4.670371953826422, "grad_norm": 1.4042103290557861, "learning_rate": 4.706340404319593e-05, "loss": 0.002, "num_input_tokens_seen": 80078864, "step": 4100 }, { "epoch": 4.67607239561066, "grad_norm": 0.30372464656829834, "learning_rate": 4.705638131253492e-05, "loss": 0.0029, "num_input_tokens_seen": 80176672, "step": 4105 }, { "epoch": 4.681772837394898, "grad_norm": 12.20240306854248, "learning_rate": 4.704935071988718e-05, "loss": 0.0156, "num_input_tokens_seen": 80274272, "step": 4110 }, { "epoch": 4.6874732791791365, "grad_norm": 0.24599520862102509, "learning_rate": 4.704231226775877e-05, "loss": 0.0106, "num_input_tokens_seen": 80372080, "step": 4115 }, { "epoch": 4.693173720963375, "grad_norm": 5.936103820800781, "learning_rate": 4.7035265958658545e-05, "loss": 0.0063, "num_input_tokens_seen": 80469824, "step": 4120 }, { "epoch": 4.698874162747613, "grad_norm": 1.1616228818893433, "learning_rate": 4.702821179509814e-05, "loss": 0.0153, "num_input_tokens_seen": 80567536, "step": 4125 }, { "epoch": 4.704574604531851, "grad_norm": 11.066569328308105, "learning_rate": 4.702114977959203e-05, "loss": 0.0302, "num_input_tokens_seen": 80665344, "step": 4130 }, { "epoch": 4.71027504631609, "grad_norm": 0.9506548047065735, "learning_rate": 4.701407991465745e-05, "loss": 0.0058, "num_input_tokens_seen": 80763072, "step": 4135 }, { "epoch": 4.7159754881003275, "grad_norm": 0.13321082293987274, "learning_rate": 4.700700220281446e-05, "loss": 0.0023, "num_input_tokens_seen": 80860816, "step": 4140 }, { "epoch": 4.721675929884566, "grad_norm": 3.3196537494659424, "learning_rate": 4.699991664658591e-05, "loss": 0.0058, "num_input_tokens_seen": 80958480, "step": 4145 }, { "epoch": 4.727376371668805, "grad_norm": 0.7469080686569214, "learning_rate": 4.699282324849742e-05, "loss": 0.0398, "num_input_tokens_seen": 81056144, "step": 4150 }, { "epoch": 4.733076813453042, "grad_norm": 6.437920093536377, "learning_rate": 4.698572201107746e-05, "loss": 0.0205, "num_input_tokens_seen": 81153888, "step": 4155 }, { "epoch": 4.738777255237281, "grad_norm": 6.570982456207275, "learning_rate": 4.697861293685724e-05, "loss": 0.0083, "num_input_tokens_seen": 81251680, "step": 4160 }, { "epoch": 4.7444776970215194, "grad_norm": 2.7099902629852295, "learning_rate": 4.69714960283708e-05, "loss": 0.0038, "num_input_tokens_seen": 81349360, "step": 4165 }, { "epoch": 4.750178138805757, "grad_norm": 0.25943759083747864, "learning_rate": 4.696437128815494e-05, "loss": 0.0249, "num_input_tokens_seen": 81447104, "step": 4170 }, { "epoch": 4.755878580589996, "grad_norm": 1.0403450727462769, "learning_rate": 4.6957238718749295e-05, "loss": 0.0079, "num_input_tokens_seen": 81544896, "step": 4175 }, { "epoch": 4.761579022374234, "grad_norm": 6.538539886474609, "learning_rate": 4.6950098322696254e-05, "loss": 0.0292, "num_input_tokens_seen": 81642576, "step": 4180 }, { "epoch": 4.767279464158472, "grad_norm": 0.7306740283966064, "learning_rate": 4.6942950102541007e-05, "loss": 0.0153, "num_input_tokens_seen": 81740384, "step": 4185 }, { "epoch": 4.7729799059427105, "grad_norm": 1.1523919105529785, "learning_rate": 4.693579406083153e-05, "loss": 0.0137, "num_input_tokens_seen": 81838112, "step": 4190 }, { "epoch": 4.778680347726949, "grad_norm": 1.3222236633300781, "learning_rate": 4.69286302001186e-05, "loss": 0.0174, "num_input_tokens_seen": 81935856, "step": 4195 }, { "epoch": 4.784380789511188, "grad_norm": 4.96268367767334, "learning_rate": 4.692145852295576e-05, "loss": 0.0059, "num_input_tokens_seen": 82033616, "step": 4200 }, { "epoch": 4.790081231295425, "grad_norm": 1.5982474088668823, "learning_rate": 4.6914279031899364e-05, "loss": 0.017, "num_input_tokens_seen": 82131360, "step": 4205 }, { "epoch": 4.795781673079664, "grad_norm": 3.684070110321045, "learning_rate": 4.690709172950854e-05, "loss": 0.0113, "num_input_tokens_seen": 82229136, "step": 4210 }, { "epoch": 4.8014821148639015, "grad_norm": 12.136567115783691, "learning_rate": 4.689989661834518e-05, "loss": 0.0284, "num_input_tokens_seen": 82326864, "step": 4215 }, { "epoch": 4.80718255664814, "grad_norm": 0.33026665449142456, "learning_rate": 4.6892693700973994e-05, "loss": 0.0104, "num_input_tokens_seen": 82424672, "step": 4220 }, { "epoch": 4.812882998432379, "grad_norm": 8.315051078796387, "learning_rate": 4.688548297996245e-05, "loss": 0.017, "num_input_tokens_seen": 82522400, "step": 4225 }, { "epoch": 4.818583440216617, "grad_norm": 0.24934843182563782, "learning_rate": 4.687826445788081e-05, "loss": 0.0035, "num_input_tokens_seen": 82620208, "step": 4230 }, { "epoch": 4.824283882000855, "grad_norm": 0.1164386197924614, "learning_rate": 4.687103813730211e-05, "loss": 0.0092, "num_input_tokens_seen": 82717856, "step": 4235 }, { "epoch": 4.8299843237850935, "grad_norm": 11.615945816040039, "learning_rate": 4.686380402080218e-05, "loss": 0.0131, "num_input_tokens_seen": 82815632, "step": 4240 }, { "epoch": 4.835684765569331, "grad_norm": 0.6481454968452454, "learning_rate": 4.68565621109596e-05, "loss": 0.0011, "num_input_tokens_seen": 82913296, "step": 4245 }, { "epoch": 4.84138520735357, "grad_norm": 0.0655444860458374, "learning_rate": 4.6849312410355755e-05, "loss": 0.0198, "num_input_tokens_seen": 83011072, "step": 4250 }, { "epoch": 4.847085649137808, "grad_norm": 6.334054470062256, "learning_rate": 4.68420549215748e-05, "loss": 0.0048, "num_input_tokens_seen": 83108864, "step": 4255 }, { "epoch": 4.852786090922047, "grad_norm": 0.20101669430732727, "learning_rate": 4.6834789647203656e-05, "loss": 0.0048, "num_input_tokens_seen": 83206608, "step": 4260 }, { "epoch": 4.8584865327062845, "grad_norm": 14.852635383605957, "learning_rate": 4.6827516589832025e-05, "loss": 0.0461, "num_input_tokens_seen": 83304336, "step": 4265 }, { "epoch": 4.864186974490523, "grad_norm": 0.06325986981391907, "learning_rate": 4.68202357520524e-05, "loss": 0.0093, "num_input_tokens_seen": 83402064, "step": 4270 }, { "epoch": 4.869887416274762, "grad_norm": 1.859485149383545, "learning_rate": 4.681294713646002e-05, "loss": 0.0104, "num_input_tokens_seen": 83499824, "step": 4275 }, { "epoch": 4.875587858058999, "grad_norm": 0.47888869047164917, "learning_rate": 4.68056507456529e-05, "loss": 0.0106, "num_input_tokens_seen": 83597536, "step": 4280 }, { "epoch": 4.881288299843238, "grad_norm": 1.207236886024475, "learning_rate": 4.6798346582231855e-05, "loss": 0.0049, "num_input_tokens_seen": 83695296, "step": 4285 }, { "epoch": 4.886988741627476, "grad_norm": 0.27957767248153687, "learning_rate": 4.679103464880044e-05, "loss": 0.0017, "num_input_tokens_seen": 83793024, "step": 4290 }, { "epoch": 4.892689183411714, "grad_norm": 0.03679969534277916, "learning_rate": 4.678371494796499e-05, "loss": 0.0023, "num_input_tokens_seen": 83890752, "step": 4295 }, { "epoch": 4.898389625195953, "grad_norm": 1.9648526906967163, "learning_rate": 4.677638748233461e-05, "loss": 0.0168, "num_input_tokens_seen": 83988512, "step": 4300 }, { "epoch": 4.904090066980191, "grad_norm": 0.7433627247810364, "learning_rate": 4.676905225452117e-05, "loss": 0.0128, "num_input_tokens_seen": 84086352, "step": 4305 }, { "epoch": 4.909790508764429, "grad_norm": 1.4374768733978271, "learning_rate": 4.676170926713932e-05, "loss": 0.0019, "num_input_tokens_seen": 84184032, "step": 4310 }, { "epoch": 4.9154909505486675, "grad_norm": 0.46811923384666443, "learning_rate": 4.6754358522806454e-05, "loss": 0.0019, "num_input_tokens_seen": 84281776, "step": 4315 }, { "epoch": 4.921191392332906, "grad_norm": 2.098421573638916, "learning_rate": 4.6747000024142734e-05, "loss": 0.0169, "num_input_tokens_seen": 84379472, "step": 4320 }, { "epoch": 4.926891834117144, "grad_norm": 3.727424383163452, "learning_rate": 4.673963377377111e-05, "loss": 0.009, "num_input_tokens_seen": 84477232, "step": 4325 }, { "epoch": 4.932592275901382, "grad_norm": 9.418045043945312, "learning_rate": 4.6732259774317264e-05, "loss": 0.0283, "num_input_tokens_seen": 84574992, "step": 4330 }, { "epoch": 4.938292717685621, "grad_norm": 8.13887882232666, "learning_rate": 4.672487802840966e-05, "loss": 0.0163, "num_input_tokens_seen": 84672800, "step": 4335 }, { "epoch": 4.9439931594698585, "grad_norm": 0.15979628264904022, "learning_rate": 4.671748853867952e-05, "loss": 0.0126, "num_input_tokens_seen": 84770416, "step": 4340 }, { "epoch": 4.949693601254097, "grad_norm": 10.529417991638184, "learning_rate": 4.671009130776083e-05, "loss": 0.0189, "num_input_tokens_seen": 84868256, "step": 4345 }, { "epoch": 4.955394043038336, "grad_norm": 1.08811354637146, "learning_rate": 4.670268633829031e-05, "loss": 0.0016, "num_input_tokens_seen": 84965872, "step": 4350 }, { "epoch": 4.961094484822574, "grad_norm": 0.6671218872070312, "learning_rate": 4.6695273632907476e-05, "loss": 0.0025, "num_input_tokens_seen": 85063648, "step": 4355 }, { "epoch": 4.966794926606812, "grad_norm": 3.7817630767822266, "learning_rate": 4.668785319425458e-05, "loss": 0.0207, "num_input_tokens_seen": 85161424, "step": 4360 }, { "epoch": 4.97249536839105, "grad_norm": 3.2574493885040283, "learning_rate": 4.668042502497663e-05, "loss": 0.0183, "num_input_tokens_seen": 85259088, "step": 4365 }, { "epoch": 4.978195810175288, "grad_norm": 3.2037136554718018, "learning_rate": 4.66729891277214e-05, "loss": 0.0128, "num_input_tokens_seen": 85356816, "step": 4370 }, { "epoch": 4.983896251959527, "grad_norm": 3.986717462539673, "learning_rate": 4.66655455051394e-05, "loss": 0.0043, "num_input_tokens_seen": 85454656, "step": 4375 }, { "epoch": 4.989596693743765, "grad_norm": 1.5552836656570435, "learning_rate": 4.6658094159883916e-05, "loss": 0.0275, "num_input_tokens_seen": 85552432, "step": 4380 }, { "epoch": 4.995297135528004, "grad_norm": 0.39177027344703674, "learning_rate": 4.665063509461097e-05, "loss": 0.0053, "num_input_tokens_seen": 85650144, "step": 4385 }, { "epoch": 5.0, "grad_norm": 0.13293787837028503, "learning_rate": 4.6643168311979345e-05, "loss": 0.0034, "num_input_tokens_seen": 85730720, "step": 4390 }, { "epoch": 5.005700441784239, "grad_norm": 1.8142844438552856, "learning_rate": 4.663569381465058e-05, "loss": 0.0094, "num_input_tokens_seen": 85828432, "step": 4395 }, { "epoch": 5.011400883568476, "grad_norm": 0.5450536012649536, "learning_rate": 4.662821160528894e-05, "loss": 0.0019, "num_input_tokens_seen": 85926048, "step": 4400 }, { "epoch": 5.017101325352715, "grad_norm": 0.5705410242080688, "learning_rate": 4.662072168656146e-05, "loss": 0.0311, "num_input_tokens_seen": 86023760, "step": 4405 }, { "epoch": 5.022801767136953, "grad_norm": 0.47627347707748413, "learning_rate": 4.661322406113794e-05, "loss": 0.005, "num_input_tokens_seen": 86121552, "step": 4410 }, { "epoch": 5.028502208921191, "grad_norm": 5.517219066619873, "learning_rate": 4.6605718731690874e-05, "loss": 0.0048, "num_input_tokens_seen": 86219200, "step": 4415 }, { "epoch": 5.03420265070543, "grad_norm": 0.165016770362854, "learning_rate": 4.659820570089555e-05, "loss": 0.0025, "num_input_tokens_seen": 86316976, "step": 4420 }, { "epoch": 5.039903092489668, "grad_norm": 2.5400843620300293, "learning_rate": 4.659068497142998e-05, "loss": 0.0026, "num_input_tokens_seen": 86414736, "step": 4425 }, { "epoch": 5.045603534273906, "grad_norm": 1.7173391580581665, "learning_rate": 4.658315654597492e-05, "loss": 0.0037, "num_input_tokens_seen": 86512528, "step": 4430 }, { "epoch": 5.051303976058144, "grad_norm": 0.1867997944355011, "learning_rate": 4.657562042721388e-05, "loss": 0.001, "num_input_tokens_seen": 86610224, "step": 4435 }, { "epoch": 5.057004417842383, "grad_norm": 1.1393805742263794, "learning_rate": 4.65680766178331e-05, "loss": 0.0047, "num_input_tokens_seen": 86708000, "step": 4440 }, { "epoch": 5.062704859626621, "grad_norm": 4.353109836578369, "learning_rate": 4.656052512052158e-05, "loss": 0.0031, "num_input_tokens_seen": 86805696, "step": 4445 }, { "epoch": 5.068405301410859, "grad_norm": 0.10244199633598328, "learning_rate": 4.655296593797104e-05, "loss": 0.0167, "num_input_tokens_seen": 86903504, "step": 4450 }, { "epoch": 5.074105743195098, "grad_norm": 3.0064287185668945, "learning_rate": 4.654539907287594e-05, "loss": 0.0035, "num_input_tokens_seen": 87001264, "step": 4455 }, { "epoch": 5.079806184979336, "grad_norm": 2.2633399963378906, "learning_rate": 4.653782452793349e-05, "loss": 0.0022, "num_input_tokens_seen": 87099008, "step": 4460 }, { "epoch": 5.085506626763574, "grad_norm": 0.3934509754180908, "learning_rate": 4.653024230584364e-05, "loss": 0.0061, "num_input_tokens_seen": 87196672, "step": 4465 }, { "epoch": 5.091207068547813, "grad_norm": 0.034104038029909134, "learning_rate": 4.6522652409309064e-05, "loss": 0.0017, "num_input_tokens_seen": 87294416, "step": 4470 }, { "epoch": 5.096907510332051, "grad_norm": 2.047616720199585, "learning_rate": 4.651505484103518e-05, "loss": 0.0136, "num_input_tokens_seen": 87392128, "step": 4475 }, { "epoch": 5.102607952116289, "grad_norm": 4.767343044281006, "learning_rate": 4.6507449603730135e-05, "loss": 0.0118, "num_input_tokens_seen": 87489840, "step": 4480 }, { "epoch": 5.108308393900527, "grad_norm": 0.24816875159740448, "learning_rate": 4.6499836700104806e-05, "loss": 0.0083, "num_input_tokens_seen": 87587568, "step": 4485 }, { "epoch": 5.114008835684766, "grad_norm": 0.16580072045326233, "learning_rate": 4.6492216132872824e-05, "loss": 0.0053, "num_input_tokens_seen": 87685264, "step": 4490 }, { "epoch": 5.119709277469004, "grad_norm": 0.23322570323944092, "learning_rate": 4.648458790475052e-05, "loss": 0.0026, "num_input_tokens_seen": 87783088, "step": 4495 }, { "epoch": 5.125409719253242, "grad_norm": 0.2388758510351181, "learning_rate": 4.6476952018456974e-05, "loss": 0.0009, "num_input_tokens_seen": 87880832, "step": 4500 }, { "epoch": 5.131110161037481, "grad_norm": 2.167498826980591, "learning_rate": 4.646930847671401e-05, "loss": 0.009, "num_input_tokens_seen": 87978544, "step": 4505 }, { "epoch": 5.136810602821718, "grad_norm": 0.15172941982746124, "learning_rate": 4.646165728224616e-05, "loss": 0.0029, "num_input_tokens_seen": 88076304, "step": 4510 }, { "epoch": 5.142511044605957, "grad_norm": 1.221136450767517, "learning_rate": 4.645399843778068e-05, "loss": 0.0045, "num_input_tokens_seen": 88174016, "step": 4515 }, { "epoch": 5.1482114863901955, "grad_norm": 0.21661746501922607, "learning_rate": 4.644633194604756e-05, "loss": 0.013, "num_input_tokens_seen": 88271632, "step": 4520 }, { "epoch": 5.153911928174433, "grad_norm": 3.4261422157287598, "learning_rate": 4.6438657809779526e-05, "loss": 0.0069, "num_input_tokens_seen": 88369312, "step": 4525 }, { "epoch": 5.159612369958672, "grad_norm": 0.3439682126045227, "learning_rate": 4.6430976031712017e-05, "loss": 0.0014, "num_input_tokens_seen": 88467120, "step": 4530 }, { "epoch": 5.16531281174291, "grad_norm": 13.96764087677002, "learning_rate": 4.6423286614583195e-05, "loss": 0.0218, "num_input_tokens_seen": 88564848, "step": 4535 }, { "epoch": 5.171013253527148, "grad_norm": 0.09054847806692123, "learning_rate": 4.641558956113396e-05, "loss": 0.0054, "num_input_tokens_seen": 88662560, "step": 4540 }, { "epoch": 5.176713695311387, "grad_norm": 1.0485111474990845, "learning_rate": 4.640788487410791e-05, "loss": 0.0044, "num_input_tokens_seen": 88760400, "step": 4545 }, { "epoch": 5.182414137095625, "grad_norm": 0.10858794301748276, "learning_rate": 4.640017255625139e-05, "loss": 0.0009, "num_input_tokens_seen": 88858096, "step": 4550 }, { "epoch": 5.188114578879863, "grad_norm": 0.07652360200881958, "learning_rate": 4.639245261031344e-05, "loss": 0.0239, "num_input_tokens_seen": 88955856, "step": 4555 }, { "epoch": 5.193815020664101, "grad_norm": 0.6881747841835022, "learning_rate": 4.638472503904583e-05, "loss": 0.0009, "num_input_tokens_seen": 89053600, "step": 4560 }, { "epoch": 5.19951546244834, "grad_norm": 0.08055282384157181, "learning_rate": 4.637698984520307e-05, "loss": 0.0034, "num_input_tokens_seen": 89151296, "step": 4565 }, { "epoch": 5.205215904232578, "grad_norm": 0.08773194998502731, "learning_rate": 4.636924703154234e-05, "loss": 0.0121, "num_input_tokens_seen": 89249120, "step": 4570 }, { "epoch": 5.210916346016816, "grad_norm": 0.2949371039867401, "learning_rate": 4.636149660082358e-05, "loss": 0.0049, "num_input_tokens_seen": 89346832, "step": 4575 }, { "epoch": 5.216616787801055, "grad_norm": 7.335551738739014, "learning_rate": 4.635373855580942e-05, "loss": 0.0274, "num_input_tokens_seen": 89444576, "step": 4580 }, { "epoch": 5.222317229585292, "grad_norm": 2.2080814838409424, "learning_rate": 4.634597289926521e-05, "loss": 0.0128, "num_input_tokens_seen": 89542288, "step": 4585 }, { "epoch": 5.228017671369531, "grad_norm": 1.00960111618042, "learning_rate": 4.6338199633959025e-05, "loss": 0.0036, "num_input_tokens_seen": 89640096, "step": 4590 }, { "epoch": 5.23371811315377, "grad_norm": 0.1926228553056717, "learning_rate": 4.6330418762661624e-05, "loss": 0.0061, "num_input_tokens_seen": 89737872, "step": 4595 }, { "epoch": 5.239418554938008, "grad_norm": 0.0730406790971756, "learning_rate": 4.632263028814652e-05, "loss": 0.0383, "num_input_tokens_seen": 89835552, "step": 4600 }, { "epoch": 5.245118996722246, "grad_norm": 0.9148241281509399, "learning_rate": 4.6314834213189884e-05, "loss": 0.0167, "num_input_tokens_seen": 89933232, "step": 4605 }, { "epoch": 5.250819438506484, "grad_norm": 1.8269907236099243, "learning_rate": 4.630703054057063e-05, "loss": 0.006, "num_input_tokens_seen": 90030960, "step": 4610 }, { "epoch": 5.256519880290723, "grad_norm": 0.4568536877632141, "learning_rate": 4.6299219273070396e-05, "loss": 0.0105, "num_input_tokens_seen": 90128784, "step": 4615 }, { "epoch": 5.262220322074961, "grad_norm": 0.6757635474205017, "learning_rate": 4.629140041347347e-05, "loss": 0.0083, "num_input_tokens_seen": 90226576, "step": 4620 }, { "epoch": 5.267920763859199, "grad_norm": 8.738912582397461, "learning_rate": 4.628357396456692e-05, "loss": 0.0166, "num_input_tokens_seen": 90324304, "step": 4625 }, { "epoch": 5.273621205643438, "grad_norm": 1.5971795320510864, "learning_rate": 4.627573992914044e-05, "loss": 0.0029, "num_input_tokens_seen": 90421920, "step": 4630 }, { "epoch": 5.279321647427675, "grad_norm": 8.538966178894043, "learning_rate": 4.626789830998649e-05, "loss": 0.0098, "num_input_tokens_seen": 90519728, "step": 4635 }, { "epoch": 5.285022089211914, "grad_norm": 0.06448430567979813, "learning_rate": 4.626004910990021e-05, "loss": 0.0135, "num_input_tokens_seen": 90617440, "step": 4640 }, { "epoch": 5.2907225309961525, "grad_norm": 7.718270301818848, "learning_rate": 4.625219233167944e-05, "loss": 0.015, "num_input_tokens_seen": 90715248, "step": 4645 }, { "epoch": 5.29642297278039, "grad_norm": 0.2514442801475525, "learning_rate": 4.6244327978124734e-05, "loss": 0.0031, "num_input_tokens_seen": 90812960, "step": 4650 }, { "epoch": 5.302123414564629, "grad_norm": 0.28784915804862976, "learning_rate": 4.623645605203932e-05, "loss": 0.0063, "num_input_tokens_seen": 90910624, "step": 4655 }, { "epoch": 5.307823856348867, "grad_norm": 0.1487150639295578, "learning_rate": 4.6228576556229156e-05, "loss": 0.0035, "num_input_tokens_seen": 91008320, "step": 4660 }, { "epoch": 5.313524298133105, "grad_norm": 0.19565777480602264, "learning_rate": 4.622068949350289e-05, "loss": 0.0022, "num_input_tokens_seen": 91106128, "step": 4665 }, { "epoch": 5.319224739917344, "grad_norm": 0.2649058401584625, "learning_rate": 4.6212794866671836e-05, "loss": 0.0156, "num_input_tokens_seen": 91203968, "step": 4670 }, { "epoch": 5.324925181701582, "grad_norm": 1.254876732826233, "learning_rate": 4.620489267855006e-05, "loss": 0.0014, "num_input_tokens_seen": 91301696, "step": 4675 }, { "epoch": 5.33062562348582, "grad_norm": 0.03180227801203728, "learning_rate": 4.619698293195427e-05, "loss": 0.0046, "num_input_tokens_seen": 91399360, "step": 4680 }, { "epoch": 5.336326065270058, "grad_norm": 4.16030216217041, "learning_rate": 4.618906562970391e-05, "loss": 0.0031, "num_input_tokens_seen": 91497088, "step": 4685 }, { "epoch": 5.342026507054297, "grad_norm": 0.0919002890586853, "learning_rate": 4.6181140774621077e-05, "loss": 0.0021, "num_input_tokens_seen": 91594688, "step": 4690 }, { "epoch": 5.347726948838535, "grad_norm": 4.587754249572754, "learning_rate": 4.617320836953061e-05, "loss": 0.0129, "num_input_tokens_seen": 91692448, "step": 4695 }, { "epoch": 5.353427390622773, "grad_norm": 0.5592033863067627, "learning_rate": 4.6165268417259986e-05, "loss": 0.002, "num_input_tokens_seen": 91790160, "step": 4700 }, { "epoch": 5.359127832407012, "grad_norm": 9.988340377807617, "learning_rate": 4.6157320920639406e-05, "loss": 0.0083, "num_input_tokens_seen": 91887888, "step": 4705 }, { "epoch": 5.364828274191249, "grad_norm": 1.1014900207519531, "learning_rate": 4.6149365882501754e-05, "loss": 0.0049, "num_input_tokens_seen": 91985648, "step": 4710 }, { "epoch": 5.370528715975488, "grad_norm": 0.13604551553726196, "learning_rate": 4.614140330568261e-05, "loss": 0.0091, "num_input_tokens_seen": 92083408, "step": 4715 }, { "epoch": 5.3762291577597265, "grad_norm": 13.926383972167969, "learning_rate": 4.6133433193020206e-05, "loss": 0.0367, "num_input_tokens_seen": 92181072, "step": 4720 }, { "epoch": 5.381929599543964, "grad_norm": 0.2026086002588272, "learning_rate": 4.61254555473555e-05, "loss": 0.0112, "num_input_tokens_seen": 92278880, "step": 4725 }, { "epoch": 5.387630041328203, "grad_norm": 0.10835447162389755, "learning_rate": 4.6117470371532115e-05, "loss": 0.0094, "num_input_tokens_seen": 92376672, "step": 4730 }, { "epoch": 5.393330483112441, "grad_norm": 5.342576026916504, "learning_rate": 4.610947766839637e-05, "loss": 0.0153, "num_input_tokens_seen": 92474448, "step": 4735 }, { "epoch": 5.39903092489668, "grad_norm": 1.6363821029663086, "learning_rate": 4.610147744079725e-05, "loss": 0.0046, "num_input_tokens_seen": 92572160, "step": 4740 }, { "epoch": 5.404731366680918, "grad_norm": 0.9857316613197327, "learning_rate": 4.609346969158645e-05, "loss": 0.0092, "num_input_tokens_seen": 92669792, "step": 4745 }, { "epoch": 5.410431808465156, "grad_norm": 0.055682141333818436, "learning_rate": 4.60854544236183e-05, "loss": 0.003, "num_input_tokens_seen": 92767520, "step": 4750 }, { "epoch": 5.416132250249395, "grad_norm": 0.026181025430560112, "learning_rate": 4.607743163974987e-05, "loss": 0.0009, "num_input_tokens_seen": 92865344, "step": 4755 }, { "epoch": 5.421832692033632, "grad_norm": 0.0219236072152853, "learning_rate": 4.6069401342840854e-05, "loss": 0.003, "num_input_tokens_seen": 92963104, "step": 4760 }, { "epoch": 5.427533133817871, "grad_norm": 0.19581133127212524, "learning_rate": 4.606136353575366e-05, "loss": 0.0008, "num_input_tokens_seen": 93060912, "step": 4765 }, { "epoch": 5.4332335756021095, "grad_norm": 0.10174310952425003, "learning_rate": 4.6053318221353356e-05, "loss": 0.0006, "num_input_tokens_seen": 93158768, "step": 4770 }, { "epoch": 5.438934017386347, "grad_norm": 14.586435317993164, "learning_rate": 4.60452654025077e-05, "loss": 0.0157, "num_input_tokens_seen": 93256496, "step": 4775 }, { "epoch": 5.444634459170586, "grad_norm": 5.632846355438232, "learning_rate": 4.6037205082087095e-05, "loss": 0.0196, "num_input_tokens_seen": 93354208, "step": 4780 }, { "epoch": 5.450334900954824, "grad_norm": 0.09631127119064331, "learning_rate": 4.602913726296466e-05, "loss": 0.0012, "num_input_tokens_seen": 93451952, "step": 4785 }, { "epoch": 5.456035342739062, "grad_norm": 5.937401294708252, "learning_rate": 4.602106194801615e-05, "loss": 0.0037, "num_input_tokens_seen": 93549744, "step": 4790 }, { "epoch": 5.4617357845233006, "grad_norm": 0.3016761839389801, "learning_rate": 4.6012979140120016e-05, "loss": 0.0026, "num_input_tokens_seen": 93647520, "step": 4795 }, { "epoch": 5.467436226307539, "grad_norm": 0.0018981621833518147, "learning_rate": 4.600488884215737e-05, "loss": 0.0114, "num_input_tokens_seen": 93745280, "step": 4800 }, { "epoch": 5.473136668091777, "grad_norm": 0.2602160573005676, "learning_rate": 4.599679105701199e-05, "loss": 0.0043, "num_input_tokens_seen": 93842992, "step": 4805 }, { "epoch": 5.478837109876015, "grad_norm": 2.016597032546997, "learning_rate": 4.598868578757033e-05, "loss": 0.0043, "num_input_tokens_seen": 93940768, "step": 4810 }, { "epoch": 5.484537551660254, "grad_norm": 0.06573915481567383, "learning_rate": 4.5980573036721505e-05, "loss": 0.0025, "num_input_tokens_seen": 94038528, "step": 4815 }, { "epoch": 5.490237993444492, "grad_norm": 1.3520628213882446, "learning_rate": 4.597245280735731e-05, "loss": 0.0018, "num_input_tokens_seen": 94136224, "step": 4820 }, { "epoch": 5.49593843522873, "grad_norm": 14.249088287353516, "learning_rate": 4.59643251023722e-05, "loss": 0.0273, "num_input_tokens_seen": 94233888, "step": 4825 }, { "epoch": 5.501638877012969, "grad_norm": 2.345010280609131, "learning_rate": 4.595618992466328e-05, "loss": 0.0017, "num_input_tokens_seen": 94331568, "step": 4830 }, { "epoch": 5.507339318797206, "grad_norm": 0.5609773397445679, "learning_rate": 4.594804727713033e-05, "loss": 0.0045, "num_input_tokens_seen": 94429248, "step": 4835 }, { "epoch": 5.513039760581445, "grad_norm": 11.708130836486816, "learning_rate": 4.5939897162675804e-05, "loss": 0.0603, "num_input_tokens_seen": 94526912, "step": 4840 }, { "epoch": 5.5187402023656835, "grad_norm": 0.511098325252533, "learning_rate": 4.59317395842048e-05, "loss": 0.0015, "num_input_tokens_seen": 94624688, "step": 4845 }, { "epoch": 5.524440644149921, "grad_norm": 0.07705602049827576, "learning_rate": 4.592357454462508e-05, "loss": 0.0008, "num_input_tokens_seen": 94722496, "step": 4850 }, { "epoch": 5.53014108593416, "grad_norm": 0.49463194608688354, "learning_rate": 4.591540204684708e-05, "loss": 0.0226, "num_input_tokens_seen": 94820176, "step": 4855 }, { "epoch": 5.535841527718398, "grad_norm": 0.13507622480392456, "learning_rate": 4.590722209378387e-05, "loss": 0.0033, "num_input_tokens_seen": 94917984, "step": 4860 }, { "epoch": 5.541541969502637, "grad_norm": 0.11557400226593018, "learning_rate": 4.589903468835119e-05, "loss": 0.0048, "num_input_tokens_seen": 95015744, "step": 4865 }, { "epoch": 5.547242411286875, "grad_norm": 3.4503333568573, "learning_rate": 4.5890839833467455e-05, "loss": 0.0044, "num_input_tokens_seen": 95113504, "step": 4870 }, { "epoch": 5.552942853071113, "grad_norm": 0.1271464228630066, "learning_rate": 4.58826375320537e-05, "loss": 0.0021, "num_input_tokens_seen": 95211264, "step": 4875 }, { "epoch": 5.558643294855351, "grad_norm": 0.09808290749788284, "learning_rate": 4.587442778703362e-05, "loss": 0.0011, "num_input_tokens_seen": 95309040, "step": 4880 }, { "epoch": 5.564343736639589, "grad_norm": 2.261197328567505, "learning_rate": 4.586621060133362e-05, "loss": 0.0024, "num_input_tokens_seen": 95406768, "step": 4885 }, { "epoch": 5.570044178423828, "grad_norm": 0.1426500380039215, "learning_rate": 4.585798597788266e-05, "loss": 0.003, "num_input_tokens_seen": 95504512, "step": 4890 }, { "epoch": 5.5757446202080665, "grad_norm": 0.056966375559568405, "learning_rate": 4.584975391961242e-05, "loss": 0.0185, "num_input_tokens_seen": 95602240, "step": 4895 }, { "epoch": 5.581445061992304, "grad_norm": 13.30504322052002, "learning_rate": 4.584151442945725e-05, "loss": 0.0217, "num_input_tokens_seen": 95699968, "step": 4900 }, { "epoch": 5.587145503776543, "grad_norm": 1.4918162822723389, "learning_rate": 4.583326751035405e-05, "loss": 0.0303, "num_input_tokens_seen": 95797696, "step": 4905 }, { "epoch": 5.592845945560781, "grad_norm": 1.1220730543136597, "learning_rate": 4.582501316524247e-05, "loss": 0.0019, "num_input_tokens_seen": 95895424, "step": 4910 }, { "epoch": 5.598546387345019, "grad_norm": 1.1162631511688232, "learning_rate": 4.5816751397064764e-05, "loss": 0.0094, "num_input_tokens_seen": 95993056, "step": 4915 }, { "epoch": 5.6042468291292575, "grad_norm": 0.10550173372030258, "learning_rate": 4.5808482208765836e-05, "loss": 0.0277, "num_input_tokens_seen": 96090832, "step": 4920 }, { "epoch": 5.609947270913496, "grad_norm": 1.0097618103027344, "learning_rate": 4.580020560329322e-05, "loss": 0.0025, "num_input_tokens_seen": 96188544, "step": 4925 }, { "epoch": 5.615647712697734, "grad_norm": 0.717867374420166, "learning_rate": 4.579192158359712e-05, "loss": 0.0037, "num_input_tokens_seen": 96286368, "step": 4930 }, { "epoch": 5.621348154481972, "grad_norm": 0.8814383149147034, "learning_rate": 4.5783630152630365e-05, "loss": 0.024, "num_input_tokens_seen": 96384128, "step": 4935 }, { "epoch": 5.627048596266211, "grad_norm": 0.2208772897720337, "learning_rate": 4.577533131334844e-05, "loss": 0.0187, "num_input_tokens_seen": 96481888, "step": 4940 }, { "epoch": 5.632749038050449, "grad_norm": 9.015159606933594, "learning_rate": 4.5767025068709455e-05, "loss": 0.0203, "num_input_tokens_seen": 96579680, "step": 4945 }, { "epoch": 5.638449479834687, "grad_norm": 0.6479278206825256, "learning_rate": 4.5758711421674166e-05, "loss": 0.0253, "num_input_tokens_seen": 96677488, "step": 4950 }, { "epoch": 5.644149921618926, "grad_norm": 0.05944683775305748, "learning_rate": 4.575039037520598e-05, "loss": 0.001, "num_input_tokens_seen": 96775280, "step": 4955 }, { "epoch": 5.649850363403163, "grad_norm": 0.376200407743454, "learning_rate": 4.5742061932270906e-05, "loss": 0.0041, "num_input_tokens_seen": 96873072, "step": 4960 }, { "epoch": 5.655550805187402, "grad_norm": 0.14665372669696808, "learning_rate": 4.5733726095837634e-05, "loss": 0.0012, "num_input_tokens_seen": 96970912, "step": 4965 }, { "epoch": 5.6612512469716405, "grad_norm": 0.11682464182376862, "learning_rate": 4.572538286887748e-05, "loss": 0.029, "num_input_tokens_seen": 97068624, "step": 4970 }, { "epoch": 5.666951688755878, "grad_norm": 0.12137410789728165, "learning_rate": 4.571703225436435e-05, "loss": 0.0007, "num_input_tokens_seen": 97166384, "step": 4975 }, { "epoch": 5.672652130540117, "grad_norm": 0.09676264226436615, "learning_rate": 4.570867425527484e-05, "loss": 0.0009, "num_input_tokens_seen": 97264112, "step": 4980 }, { "epoch": 5.678352572324355, "grad_norm": 0.3440670371055603, "learning_rate": 4.570030887458815e-05, "loss": 0.0014, "num_input_tokens_seen": 97361872, "step": 4985 }, { "epoch": 5.684053014108594, "grad_norm": 0.712581992149353, "learning_rate": 4.569193611528612e-05, "loss": 0.0043, "num_input_tokens_seen": 97459616, "step": 4990 }, { "epoch": 5.6897534558928315, "grad_norm": 7.326711177825928, "learning_rate": 4.5683555980353197e-05, "loss": 0.009, "num_input_tokens_seen": 97557376, "step": 4995 }, { "epoch": 5.69545389767707, "grad_norm": 0.251907616853714, "learning_rate": 4.56751684727765e-05, "loss": 0.0112, "num_input_tokens_seen": 97655040, "step": 5000 } ], "logging_steps": 5, "max_steps": 26310, "num_input_tokens_seen": 97655040, "num_train_epochs": 30, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 9.099184256070451e+17, "train_batch_size": 2, "trial_name": null, "trial_params": null }