{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.0, "eval_steps": 500, "global_step": 476, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.004201680672268907, "grad_norm": 6.843970053848994, "learning_rate": 4.1666666666666667e-07, "loss": 1.4834, "step": 1 }, { "epoch": 0.02100840336134454, "grad_norm": 5.888026413386213, "learning_rate": 2.0833333333333334e-06, "loss": 1.4466, "step": 5 }, { "epoch": 0.04201680672268908, "grad_norm": 1.672754748671725, "learning_rate": 4.166666666666667e-06, "loss": 1.29, "step": 10 }, { "epoch": 0.06302521008403361, "grad_norm": 1.2549304333706048, "learning_rate": 6.25e-06, "loss": 1.265, "step": 15 }, { "epoch": 0.08403361344537816, "grad_norm": 1.0804048514467623, "learning_rate": 8.333333333333334e-06, "loss": 1.218, "step": 20 }, { "epoch": 0.10504201680672269, "grad_norm": 1.0012856727023651, "learning_rate": 1.0416666666666668e-05, "loss": 1.1874, "step": 25 }, { "epoch": 0.12605042016806722, "grad_norm": 1.0428770541540702, "learning_rate": 1.25e-05, "loss": 1.1738, "step": 30 }, { "epoch": 0.14705882352941177, "grad_norm": 0.8971861131120643, "learning_rate": 1.4583333333333333e-05, "loss": 1.1823, "step": 35 }, { "epoch": 0.16806722689075632, "grad_norm": 0.8761555093987056, "learning_rate": 1.6666666666666667e-05, "loss": 1.1509, "step": 40 }, { "epoch": 0.18907563025210083, "grad_norm": 0.9022691166860466, "learning_rate": 1.8750000000000002e-05, "loss": 1.1738, "step": 45 }, { "epoch": 0.21008403361344538, "grad_norm": 0.8595698667749491, "learning_rate": 1.9998922457512608e-05, "loss": 1.1827, "step": 50 }, { "epoch": 0.23109243697478993, "grad_norm": 0.8131261419160233, "learning_rate": 1.9986802771267902e-05, "loss": 1.1595, "step": 55 }, { "epoch": 0.25210084033613445, "grad_norm": 0.9352189654627542, "learning_rate": 1.996123284790336e-05, "loss": 1.1751, "step": 60 }, { "epoch": 0.27310924369747897, "grad_norm": 0.8183133797776339, "learning_rate": 1.9922247125020307e-05, "loss": 1.1368, "step": 65 }, { "epoch": 0.29411764705882354, "grad_norm": 0.7919702228536482, "learning_rate": 1.9869898108633834e-05, "loss": 1.1576, "step": 70 }, { "epoch": 0.31512605042016806, "grad_norm": 0.809220998201082, "learning_rate": 1.9804256302457653e-05, "loss": 1.1214, "step": 75 }, { "epoch": 0.33613445378151263, "grad_norm": 0.7934942224774806, "learning_rate": 1.972541011294959e-05, "loss": 1.1665, "step": 80 }, { "epoch": 0.35714285714285715, "grad_norm": 0.8016460068505575, "learning_rate": 1.963346573024568e-05, "loss": 1.14, "step": 85 }, { "epoch": 0.37815126050420167, "grad_norm": 0.7959323509300636, "learning_rate": 1.952854698514318e-05, "loss": 1.1427, "step": 90 }, { "epoch": 0.39915966386554624, "grad_norm": 0.834564941958375, "learning_rate": 1.9410795182325113e-05, "loss": 1.1582, "step": 95 }, { "epoch": 0.42016806722689076, "grad_norm": 0.8021887426628376, "learning_rate": 1.9280368910050943e-05, "loss": 1.1425, "step": 100 }, { "epoch": 0.4411764705882353, "grad_norm": 0.8535114934490843, "learning_rate": 1.9137443826569758e-05, "loss": 1.1507, "step": 105 }, { "epoch": 0.46218487394957986, "grad_norm": 0.8097650803240342, "learning_rate": 1.898221242354353e-05, "loss": 1.1376, "step": 110 }, { "epoch": 0.4831932773109244, "grad_norm": 0.854806698328117, "learning_rate": 1.881488376679912e-05, "loss": 1.1538, "step": 115 }, { "epoch": 0.5042016806722689, "grad_norm": 0.7411750487736666, "learning_rate": 1.8635683214758213e-05, "loss": 1.1548, "step": 120 }, { "epoch": 0.5252100840336135, "grad_norm": 0.7470899750611266, "learning_rate": 1.8444852114924325e-05, "loss": 1.1285, "step": 125 }, { "epoch": 0.5462184873949579, "grad_norm": 0.7855863245637623, "learning_rate": 1.8242647478835717e-05, "loss": 1.1428, "step": 130 }, { "epoch": 0.5672268907563025, "grad_norm": 0.74562538616598, "learning_rate": 1.8029341635921985e-05, "loss": 1.1303, "step": 135 }, { "epoch": 0.5882352941176471, "grad_norm": 0.751956755942676, "learning_rate": 1.780522186673046e-05, "loss": 1.157, "step": 140 }, { "epoch": 0.6092436974789915, "grad_norm": 0.7490360146750048, "learning_rate": 1.7570590016016455e-05, "loss": 1.1245, "step": 145 }, { "epoch": 0.6302521008403361, "grad_norm": 0.7962243135552697, "learning_rate": 1.7325762086218415e-05, "loss": 1.1577, "step": 150 }, { "epoch": 0.6512605042016807, "grad_norm": 0.7647261906536346, "learning_rate": 1.7071067811865477e-05, "loss": 1.1378, "step": 155 }, { "epoch": 0.6722689075630253, "grad_norm": 0.7569274861307836, "learning_rate": 1.680685021549063e-05, "loss": 1.1249, "step": 160 }, { "epoch": 0.6932773109243697, "grad_norm": 0.7606095560211281, "learning_rate": 1.6533465145647598e-05, "loss": 1.1405, "step": 165 }, { "epoch": 0.7142857142857143, "grad_norm": 0.7797117802105004, "learning_rate": 1.6251280797653606e-05, "loss": 1.1432, "step": 170 }, { "epoch": 0.7352941176470589, "grad_norm": 0.7411898433127342, "learning_rate": 1.5960677217703512e-05, "loss": 1.1341, "step": 175 }, { "epoch": 0.7563025210084033, "grad_norm": 0.7937630835311652, "learning_rate": 1.566204579102317e-05, "loss": 1.1771, "step": 180 }, { "epoch": 0.7773109243697479, "grad_norm": 0.7831403718494921, "learning_rate": 1.5355788714751378e-05, "loss": 1.1346, "step": 185 }, { "epoch": 0.7983193277310925, "grad_norm": 0.7670962454375009, "learning_rate": 1.5042318456260305e-05, "loss": 1.1411, "step": 190 }, { "epoch": 0.819327731092437, "grad_norm": 0.8187182133488576, "learning_rate": 1.4722057197643986e-05, "loss": 1.1287, "step": 195 }, { "epoch": 0.8403361344537815, "grad_norm": 0.7707658033677821, "learning_rate": 1.4395436267123017e-05, "loss": 1.1686, "step": 200 }, { "epoch": 0.8613445378151261, "grad_norm": 0.7929027157964723, "learning_rate": 1.40628955581312e-05, "loss": 1.1095, "step": 205 }, { "epoch": 0.8823529411764706, "grad_norm": 0.7671917924588447, "learning_rate": 1.3724882936866596e-05, "loss": 1.1532, "step": 210 }, { "epoch": 0.9033613445378151, "grad_norm": 0.743350743534685, "learning_rate": 1.3381853639104817e-05, "loss": 1.1177, "step": 215 }, { "epoch": 0.9243697478991597, "grad_norm": 0.7994542205340992, "learning_rate": 1.3034269657086993e-05, "loss": 1.0984, "step": 220 }, { "epoch": 0.9453781512605042, "grad_norm": 0.8028923468414233, "learning_rate": 1.2682599117308083e-05, "loss": 1.1225, "step": 225 }, { "epoch": 0.9663865546218487, "grad_norm": 0.7701315329198188, "learning_rate": 1.2327315650043605e-05, "loss": 1.1225, "step": 230 }, { "epoch": 0.9873949579831933, "grad_norm": 0.749625186589024, "learning_rate": 1.1968897751463841e-05, "loss": 1.1205, "step": 235 }, { "epoch": 1.0, "eval_loss": 1.122010350227356, "eval_runtime": 4.1604, "eval_samples_per_second": 35.574, "eval_steps_per_second": 0.721, "step": 238 }, { "epoch": 1.0084033613445378, "grad_norm": 0.8685596492224139, "learning_rate": 1.1607828139194683e-05, "loss": 1.0392, "step": 240 }, { "epoch": 1.0294117647058822, "grad_norm": 0.8195681790245358, "learning_rate": 1.1244593102192961e-05, "loss": 0.8804, "step": 245 }, { "epoch": 1.050420168067227, "grad_norm": 0.7862106935161252, "learning_rate": 1.0879681845811964e-05, "loss": 0.8759, "step": 250 }, { "epoch": 1.0714285714285714, "grad_norm": 0.7695442905422751, "learning_rate": 1.0513585832939104e-05, "loss": 0.8921, "step": 255 }, { "epoch": 1.092436974789916, "grad_norm": 0.7197984706671763, "learning_rate": 1.0146798122093167e-05, "loss": 0.8465, "step": 260 }, { "epoch": 1.1134453781512605, "grad_norm": 0.7436730962206078, "learning_rate": 9.779812703372538e-06, "loss": 0.8983, "step": 265 }, { "epoch": 1.134453781512605, "grad_norm": 0.7419938040761785, "learning_rate": 9.41312383314878e-06, "loss": 0.86, "step": 270 }, { "epoch": 1.1554621848739495, "grad_norm": 0.7727780995520752, "learning_rate": 9.047225368401622e-06, "loss": 0.8751, "step": 275 }, { "epoch": 1.1764705882352942, "grad_norm": 0.7184055924425042, "learning_rate": 8.682610101591813e-06, "loss": 0.8633, "step": 280 }, { "epoch": 1.1974789915966386, "grad_norm": 0.7424505911824795, "learning_rate": 8.319769096967681e-06, "loss": 0.881, "step": 285 }, { "epoch": 1.2184873949579833, "grad_norm": 0.7831013281249656, "learning_rate": 7.95919102919926e-06, "loss": 0.8758, "step": 290 }, { "epoch": 1.2394957983193278, "grad_norm": 0.7218415137209162, "learning_rate": 7.601361525230713e-06, "loss": 0.848, "step": 295 }, { "epoch": 1.2605042016806722, "grad_norm": 0.727646718952851, "learning_rate": 7.246762510237404e-06, "loss": 0.8928, "step": 300 }, { "epoch": 1.2815126050420167, "grad_norm": 0.7318861363154572, "learning_rate": 6.89587155856853e-06, "loss": 0.8861, "step": 305 }, { "epoch": 1.3025210084033614, "grad_norm": 0.7850969062028121, "learning_rate": 6.549161250549474e-06, "loss": 0.8785, "step": 310 }, { "epoch": 1.3235294117647058, "grad_norm": 0.7424275388998128, "learning_rate": 6.207098536010083e-06, "loss": 0.8732, "step": 315 }, { "epoch": 1.3445378151260505, "grad_norm": 0.7785063907714577, "learning_rate": 5.8701441053961185e-06, "loss": 0.8923, "step": 320 }, { "epoch": 1.365546218487395, "grad_norm": 0.7101355435294052, "learning_rate": 5.538751769310842e-06, "loss": 0.8756, "step": 325 }, { "epoch": 1.3865546218487395, "grad_norm": 0.7360750873465238, "learning_rate": 5.213367847322408e-06, "loss": 0.8657, "step": 330 }, { "epoch": 1.407563025210084, "grad_norm": 0.7493101168945231, "learning_rate": 4.894430566860144e-06, "loss": 0.8565, "step": 335 }, { "epoch": 1.4285714285714286, "grad_norm": 0.7246837540339062, "learning_rate": 4.58236947300939e-06, "loss": 0.867, "step": 340 }, { "epoch": 1.449579831932773, "grad_norm": 0.7052644676750524, "learning_rate": 4.277604849999666e-06, "loss": 0.8594, "step": 345 }, { "epoch": 1.4705882352941178, "grad_norm": 0.7349952399246719, "learning_rate": 3.980547155165429e-06, "loss": 0.8533, "step": 350 }, { "epoch": 1.4915966386554622, "grad_norm": 0.7284230275061265, "learning_rate": 3.691596466141666e-06, "loss": 0.9056, "step": 355 }, { "epoch": 1.5126050420168067, "grad_norm": 0.7193222649134673, "learning_rate": 3.4111419420388904e-06, "loss": 0.8662, "step": 360 }, { "epoch": 1.5336134453781511, "grad_norm": 0.731566479431413, "learning_rate": 3.139561299323206e-06, "loss": 0.8587, "step": 365 }, { "epoch": 1.5546218487394958, "grad_norm": 0.7391733181416821, "learning_rate": 2.877220303107373e-06, "loss": 0.8602, "step": 370 }, { "epoch": 1.5756302521008403, "grad_norm": 0.7390429200298151, "learning_rate": 2.624472274537925e-06, "loss": 0.8701, "step": 375 }, { "epoch": 1.596638655462185, "grad_norm": 0.741782430807988, "learning_rate": 2.381657614941858e-06, "loss": 0.8588, "step": 380 }, { "epoch": 1.6176470588235294, "grad_norm": 0.7082521628110617, "learning_rate": 2.149103347373753e-06, "loss": 0.8738, "step": 385 }, { "epoch": 1.638655462184874, "grad_norm": 0.7382222604301619, "learning_rate": 1.927122676180756e-06, "loss": 0.8501, "step": 390 }, { "epoch": 1.6596638655462184, "grad_norm": 0.7170057238227614, "learning_rate": 1.7160145651786098e-06, "loss": 0.8628, "step": 395 }, { "epoch": 1.680672268907563, "grad_norm": 0.7735455354416457, "learning_rate": 1.516063335006851e-06, "loss": 0.8799, "step": 400 }, { "epoch": 1.7016806722689075, "grad_norm": 0.7544008999017865, "learning_rate": 1.3275382802054704e-06, "loss": 0.8584, "step": 405 }, { "epoch": 1.7226890756302522, "grad_norm": 0.7234290509257019, "learning_rate": 1.1506933065287062e-06, "loss": 0.8878, "step": 410 }, { "epoch": 1.7436974789915967, "grad_norm": 0.7179219619602794, "learning_rate": 9.85766588984508e-07, "loss": 0.8701, "step": 415 }, { "epoch": 1.7647058823529411, "grad_norm": 0.710095564820816, "learning_rate": 8.329802510601559e-07, "loss": 0.8732, "step": 420 }, { "epoch": 1.7857142857142856, "grad_norm": 0.7465304276755028, "learning_rate": 6.925400655661118e-07, "loss": 0.897, "step": 425 }, { "epoch": 1.8067226890756303, "grad_norm": 0.7039905037617595, "learning_rate": 5.646351775009617e-07, "loss": 0.8484, "step": 430 }, { "epoch": 1.8277310924369747, "grad_norm": 0.7170317726401236, "learning_rate": 4.494378493107232e-07, "loss": 0.8518, "step": 435 }, { "epoch": 1.8487394957983194, "grad_norm": 0.7399232574218224, "learning_rate": 3.471032288855869e-07, "loss": 0.8629, "step": 440 }, { "epoch": 1.8697478991596639, "grad_norm": 0.7309518709786924, "learning_rate": 2.577691406065708e-07, "loss": 0.8723, "step": 445 }, { "epoch": 1.8907563025210083, "grad_norm": 0.7524923842002688, "learning_rate": 1.8155589972348453e-07, "loss": 0.8655, "step": 450 }, { "epoch": 1.9117647058823528, "grad_norm": 0.7470763175141618, "learning_rate": 1.1856615031422214e-07, "loss": 0.8656, "step": 455 }, { "epoch": 1.9327731092436975, "grad_norm": 0.7413561376156461, "learning_rate": 6.888472704359661e-08, "loss": 0.8555, "step": 460 }, { "epoch": 1.9537815126050422, "grad_norm": 0.725024144075873, "learning_rate": 3.2578540907926e-08, "loss": 0.8683, "step": 465 }, { "epoch": 1.9747899159663866, "grad_norm": 0.7137402433851012, "learning_rate": 9.696489119221942e-09, "loss": 0.8736, "step": 470 }, { "epoch": 1.995798319327731, "grad_norm": 0.7264722107888033, "learning_rate": 2.6938925037689467e-10, "loss": 0.8776, "step": 475 }, { "epoch": 2.0, "eval_loss": 1.1365480422973633, "eval_runtime": 5.9679, "eval_samples_per_second": 24.799, "eval_steps_per_second": 0.503, "step": 476 }, { "epoch": 2.0, "step": 476, "total_flos": 124772558045184.0, "train_loss": 1.015126781929441, "train_runtime": 3305.096, "train_samples_per_second": 9.214, "train_steps_per_second": 0.144 } ], "logging_steps": 5, "max_steps": 476, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 124772558045184.0, "train_batch_size": 8, "trial_name": null, "trial_params": null }