{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.11273957158962795, "eval_steps": 17, "global_step": 200, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0005636978579481398, "eval_loss": 6.74771785736084, "eval_runtime": 83.4995, "eval_samples_per_second": 35.785, "eval_steps_per_second": 4.479, "step": 1 }, { "epoch": 0.0016910935738444193, "grad_norm": 26.02037811279297, "learning_rate": 3e-05, "loss": 6.6508, "step": 3 }, { "epoch": 0.0033821871476888386, "grad_norm": 18.14887809753418, "learning_rate": 6e-05, "loss": 4.3309, "step": 6 }, { "epoch": 0.005073280721533258, "grad_norm": 3.2000560760498047, "learning_rate": 9e-05, "loss": 1.1379, "step": 9 }, { "epoch": 0.006764374295377677, "grad_norm": 7.714017391204834, "learning_rate": 9.997266286704631e-05, "loss": 0.787, "step": 12 }, { "epoch": 0.008455467869222097, "grad_norm": 1.9846014976501465, "learning_rate": 9.98292246503335e-05, "loss": 0.705, "step": 15 }, { "epoch": 0.009582863585118376, "eval_loss": 0.700009822845459, "eval_runtime": 83.9952, "eval_samples_per_second": 35.573, "eval_steps_per_second": 4.453, "step": 17 }, { "epoch": 0.010146561443066516, "grad_norm": 1.9030941724777222, "learning_rate": 9.956320346634876e-05, "loss": 0.7221, "step": 18 }, { "epoch": 0.011837655016910935, "grad_norm": 2.7882025241851807, "learning_rate": 9.917525374361912e-05, "loss": 0.7034, "step": 21 }, { "epoch": 0.013528748590755355, "grad_norm": 1.3202898502349854, "learning_rate": 9.86663298624003e-05, "loss": 0.7214, "step": 24 }, { "epoch": 0.015219842164599774, "grad_norm": 0.6887256503105164, "learning_rate": 9.803768380684242e-05, "loss": 0.6916, "step": 27 }, { "epoch": 0.016910935738444193, "grad_norm": 2.1117467880249023, "learning_rate": 9.729086208503174e-05, "loss": 0.7063, "step": 30 }, { "epoch": 0.018602029312288614, "grad_norm": 0.500767707824707, "learning_rate": 9.642770192448536e-05, "loss": 0.7011, "step": 33 }, { "epoch": 0.019165727170236752, "eval_loss": 0.6982219219207764, "eval_runtime": 84.042, "eval_samples_per_second": 35.554, "eval_steps_per_second": 4.45, "step": 34 }, { "epoch": 0.020293122886133032, "grad_norm": 0.8041173815727234, "learning_rate": 9.545032675245813e-05, "loss": 0.6928, "step": 36 }, { "epoch": 0.021984216459977453, "grad_norm": 0.7337102293968201, "learning_rate": 9.43611409721806e-05, "loss": 0.7076, "step": 39 }, { "epoch": 0.02367531003382187, "grad_norm": 0.6326396465301514, "learning_rate": 9.316282404787871e-05, "loss": 0.7006, "step": 42 }, { "epoch": 0.02536640360766629, "grad_norm": 0.4257495403289795, "learning_rate": 9.185832391312644e-05, "loss": 0.7013, "step": 45 }, { "epoch": 0.02705749718151071, "grad_norm": 0.2765394151210785, "learning_rate": 9.045084971874738e-05, "loss": 0.6911, "step": 48 }, { "epoch": 0.02874859075535513, "grad_norm": 0.4698239862918854, "learning_rate": 8.894386393810563e-05, "loss": 0.6933, "step": 51 }, { "epoch": 0.02874859075535513, "eval_loss": 0.6953115463256836, "eval_runtime": 84.0384, "eval_samples_per_second": 35.555, "eval_steps_per_second": 4.45, "step": 51 }, { "epoch": 0.030439684329199548, "grad_norm": 0.7247682213783264, "learning_rate": 8.73410738492077e-05, "loss": 0.7067, "step": 54 }, { "epoch": 0.032130777903043965, "grad_norm": 0.6278282999992371, "learning_rate": 8.564642241456986e-05, "loss": 0.697, "step": 57 }, { "epoch": 0.033821871476888386, "grad_norm": 0.45123350620269775, "learning_rate": 8.386407858128706e-05, "loss": 0.6992, "step": 60 }, { "epoch": 0.03551296505073281, "grad_norm": 0.5516684651374817, "learning_rate": 8.199842702516583e-05, "loss": 0.683, "step": 63 }, { "epoch": 0.03720405862457723, "grad_norm": 1.170713186264038, "learning_rate": 8.005405736415126e-05, "loss": 0.7026, "step": 66 }, { "epoch": 0.038331454340473504, "eval_loss": 0.696951150894165, "eval_runtime": 84.0534, "eval_samples_per_second": 35.549, "eval_steps_per_second": 4.45, "step": 68 }, { "epoch": 0.03889515219842165, "grad_norm": 1.1506222486495972, "learning_rate": 7.803575286758364e-05, "loss": 0.7467, "step": 69 }, { "epoch": 0.040586245772266064, "grad_norm": 0.87104731798172, "learning_rate": 7.594847868906076e-05, "loss": 0.6972, "step": 72 }, { "epoch": 0.042277339346110485, "grad_norm": 0.43808355927467346, "learning_rate": 7.379736965185368e-05, "loss": 0.6875, "step": 75 }, { "epoch": 0.043968432919954906, "grad_norm": 0.8352026343345642, "learning_rate": 7.158771761692464e-05, "loss": 0.6814, "step": 78 }, { "epoch": 0.04565952649379933, "grad_norm": 1.1601524353027344, "learning_rate": 6.932495846462261e-05, "loss": 0.7048, "step": 81 }, { "epoch": 0.04735062006764374, "grad_norm": 1.253566861152649, "learning_rate": 6.701465872208216e-05, "loss": 0.6852, "step": 84 }, { "epoch": 0.047914317925591886, "eval_loss": 0.6991522312164307, "eval_runtime": 84.042, "eval_samples_per_second": 35.554, "eval_steps_per_second": 4.45, "step": 85 }, { "epoch": 0.04904171364148816, "grad_norm": 1.2970808744430542, "learning_rate": 6.466250186922325e-05, "loss": 0.703, "step": 87 }, { "epoch": 0.05073280721533258, "grad_norm": 0.752733588218689, "learning_rate": 6.227427435703997e-05, "loss": 0.6913, "step": 90 }, { "epoch": 0.052423900789177004, "grad_norm": 0.5592336654663086, "learning_rate": 5.985585137257401e-05, "loss": 0.684, "step": 93 }, { "epoch": 0.05411499436302142, "grad_norm": 0.7148630023002625, "learning_rate": 5.74131823855921e-05, "loss": 0.7043, "step": 96 }, { "epoch": 0.05580608793686584, "grad_norm": 0.5819842219352722, "learning_rate": 5.495227651252315e-05, "loss": 0.6832, "step": 99 }, { "epoch": 0.05749718151071026, "grad_norm": 0.6519930958747864, "learning_rate": 5.247918773366112e-05, "loss": 0.6876, "step": 102 }, { "epoch": 0.05749718151071026, "eval_loss": 0.6744697690010071, "eval_runtime": 84.0459, "eval_samples_per_second": 35.552, "eval_steps_per_second": 4.45, "step": 102 }, { "epoch": 0.05918827508455468, "grad_norm": 1.0141414403915405, "learning_rate": 5e-05, "loss": 0.6754, "step": 105 }, { "epoch": 0.060879368658399095, "grad_norm": 2.331167697906494, "learning_rate": 4.7520812266338885e-05, "loss": 0.6436, "step": 108 }, { "epoch": 0.06257046223224352, "grad_norm": 3.2144436836242676, "learning_rate": 4.504772348747687e-05, "loss": 0.6441, "step": 111 }, { "epoch": 0.06426155580608793, "grad_norm": 3.9336791038513184, "learning_rate": 4.2586817614407895e-05, "loss": 0.5887, "step": 114 }, { "epoch": 0.06595264937993235, "grad_norm": 2.9044606685638428, "learning_rate": 4.0144148627425993e-05, "loss": 0.6684, "step": 117 }, { "epoch": 0.06708004509582864, "eval_loss": 0.6285394430160522, "eval_runtime": 84.0476, "eval_samples_per_second": 35.551, "eval_steps_per_second": 4.45, "step": 119 }, { "epoch": 0.06764374295377677, "grad_norm": 3.0893630981445312, "learning_rate": 3.772572564296005e-05, "loss": 0.6821, "step": 120 }, { "epoch": 0.0693348365276212, "grad_norm": 2.5429139137268066, "learning_rate": 3.533749813077677e-05, "loss": 0.5834, "step": 123 }, { "epoch": 0.07102593010146561, "grad_norm": 3.898667812347412, "learning_rate": 3.298534127791785e-05, "loss": 0.6205, "step": 126 }, { "epoch": 0.07271702367531004, "grad_norm": 3.492971897125244, "learning_rate": 3.0675041535377405e-05, "loss": 0.551, "step": 129 }, { "epoch": 0.07440811724915446, "grad_norm": 3.596611261367798, "learning_rate": 2.8412282383075363e-05, "loss": 0.6237, "step": 132 }, { "epoch": 0.07609921082299888, "grad_norm": 5.098214626312256, "learning_rate": 2.6202630348146324e-05, "loss": 0.6905, "step": 135 }, { "epoch": 0.07666290868094701, "eval_loss": 0.6011126637458801, "eval_runtime": 84.0372, "eval_samples_per_second": 35.556, "eval_steps_per_second": 4.45, "step": 136 }, { "epoch": 0.0777903043968433, "grad_norm": 3.057682514190674, "learning_rate": 2.405152131093926e-05, "loss": 0.614, "step": 138 }, { "epoch": 0.0794813979706877, "grad_norm": 2.849776268005371, "learning_rate": 2.196424713241637e-05, "loss": 0.55, "step": 141 }, { "epoch": 0.08117249154453213, "grad_norm": 4.259140491485596, "learning_rate": 1.9945942635848748e-05, "loss": 0.5797, "step": 144 }, { "epoch": 0.08286358511837655, "grad_norm": 3.7241506576538086, "learning_rate": 1.800157297483417e-05, "loss": 0.6275, "step": 147 }, { "epoch": 0.08455467869222097, "grad_norm": 2.4338207244873047, "learning_rate": 1.6135921418712956e-05, "loss": 0.577, "step": 150 }, { "epoch": 0.08624577226606539, "grad_norm": 2.8285350799560547, "learning_rate": 1.435357758543015e-05, "loss": 0.6133, "step": 153 }, { "epoch": 0.08624577226606539, "eval_loss": 0.5718837380409241, "eval_runtime": 84.0552, "eval_samples_per_second": 35.548, "eval_steps_per_second": 4.449, "step": 153 }, { "epoch": 0.08793686583990981, "grad_norm": 3.2385404109954834, "learning_rate": 1.2658926150792322e-05, "loss": 0.607, "step": 156 }, { "epoch": 0.08962795941375423, "grad_norm": 3.1674704551696777, "learning_rate": 1.1056136061894384e-05, "loss": 0.5506, "step": 159 }, { "epoch": 0.09131905298759865, "grad_norm": 3.280355930328369, "learning_rate": 9.549150281252633e-06, "loss": 0.5688, "step": 162 }, { "epoch": 0.09301014656144306, "grad_norm": 3.4832043647766113, "learning_rate": 8.141676086873572e-06, "loss": 0.6335, "step": 165 }, { "epoch": 0.09470124013528748, "grad_norm": 4.496728420257568, "learning_rate": 6.837175952121306e-06, "loss": 0.5274, "step": 168 }, { "epoch": 0.09582863585118377, "eval_loss": 0.5605720281600952, "eval_runtime": 84.0529, "eval_samples_per_second": 35.549, "eval_steps_per_second": 4.45, "step": 170 }, { "epoch": 0.0963923337091319, "grad_norm": 2.9714503288269043, "learning_rate": 5.6388590278194096e-06, "loss": 0.5568, "step": 171 }, { "epoch": 0.09808342728297632, "grad_norm": 2.8310670852661133, "learning_rate": 4.549673247541875e-06, "loss": 0.6169, "step": 174 }, { "epoch": 0.09977452085682074, "grad_norm": 3.3684370517730713, "learning_rate": 3.5722980755146517e-06, "loss": 0.4749, "step": 177 }, { "epoch": 0.10146561443066517, "grad_norm": 3.3378119468688965, "learning_rate": 2.7091379149682685e-06, "loss": 0.5579, "step": 180 }, { "epoch": 0.10315670800450959, "grad_norm": 3.996790885925293, "learning_rate": 1.962316193157593e-06, "loss": 0.5408, "step": 183 }, { "epoch": 0.10484780157835401, "grad_norm": 3.091064453125, "learning_rate": 1.333670137599713e-06, "loss": 0.5679, "step": 186 }, { "epoch": 0.10541149943630214, "eval_loss": 0.5559670925140381, "eval_runtime": 84.033, "eval_samples_per_second": 35.557, "eval_steps_per_second": 4.451, "step": 187 }, { "epoch": 0.10653889515219842, "grad_norm": 3.5897178649902344, "learning_rate": 8.247462563808817e-07, "loss": 0.5888, "step": 189 }, { "epoch": 0.10822998872604284, "grad_norm": 4.172910690307617, "learning_rate": 4.367965336512403e-07, "loss": 0.654, "step": 192 }, { "epoch": 0.10992108229988726, "grad_norm": 7.166314125061035, "learning_rate": 1.7077534966650766e-07, "loss": 0.6831, "step": 195 }, { "epoch": 0.11161217587373168, "grad_norm": 3.00948429107666, "learning_rate": 2.7337132953697554e-08, "loss": 0.5601, "step": 198 } ], "logging_steps": 3, "max_steps": 200, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 17, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 5.4428946137088e+16, "train_batch_size": 8, "trial_name": null, "trial_params": null }