{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 2.8421052631578947, "eval_steps": 500, "global_step": 42, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.07017543859649122, "grad_norm": 10.52865219116211, "learning_rate": 0.0, "loss": 5.0782, "step": 1 }, { "epoch": 0.14035087719298245, "grad_norm": 10.9165620803833, "learning_rate": 2.0000000000000003e-06, "loss": 5.1577, "step": 2 }, { "epoch": 0.21052631578947367, "grad_norm": 11.1675443649292, "learning_rate": 4.000000000000001e-06, "loss": 5.1071, "step": 3 }, { "epoch": 0.2807017543859649, "grad_norm": 10.448240280151367, "learning_rate": 6e-06, "loss": 5.1306, "step": 4 }, { "epoch": 0.3508771929824561, "grad_norm": 10.703569412231445, "learning_rate": 8.000000000000001e-06, "loss": 5.0578, "step": 5 }, { "epoch": 0.42105263157894735, "grad_norm": 10.621328353881836, "learning_rate": 1e-05, "loss": 4.9643, "step": 6 }, { "epoch": 0.49122807017543857, "grad_norm": 10.412773132324219, "learning_rate": 1.2e-05, "loss": 4.8105, "step": 7 }, { "epoch": 0.5614035087719298, "grad_norm": 10.143810272216797, "learning_rate": 1.4000000000000001e-05, "loss": 4.646, "step": 8 }, { "epoch": 0.631578947368421, "grad_norm": 8.884021759033203, "learning_rate": 1.6000000000000003e-05, "loss": 4.478, "step": 9 }, { "epoch": 0.7017543859649122, "grad_norm": 7.186990261077881, "learning_rate": 1.8e-05, "loss": 4.2099, "step": 10 }, { "epoch": 0.7719298245614035, "grad_norm": 5.728164196014404, "learning_rate": 2e-05, "loss": 4.0316, "step": 11 }, { "epoch": 0.8421052631578947, "grad_norm": 7.005011081695557, "learning_rate": 2.2000000000000003e-05, "loss": 3.9128, "step": 12 }, { "epoch": 0.9122807017543859, "grad_norm": 8.047316551208496, "learning_rate": 2.4e-05, "loss": 3.7518, "step": 13 }, { "epoch": 0.9824561403508771, "grad_norm": 6.397008895874023, "learning_rate": 2.6000000000000002e-05, "loss": 3.5923, "step": 14 }, { "epoch": 1.0, "grad_norm": 4.87739372253418, "learning_rate": 2.8000000000000003e-05, "loss": 3.3723, "step": 15 }, { "epoch": 1.0701754385964912, "grad_norm": 3.8167903423309326, "learning_rate": 3e-05, "loss": 3.3214, "step": 16 }, { "epoch": 1.1403508771929824, "grad_norm": 3.506105899810791, "learning_rate": 3.2000000000000005e-05, "loss": 3.2114, "step": 17 }, { "epoch": 1.2105263157894737, "grad_norm": 2.6847188472747803, "learning_rate": 3.4000000000000007e-05, "loss": 3.1914, "step": 18 }, { "epoch": 1.280701754385965, "grad_norm": 2.0849263668060303, "learning_rate": 3.6e-05, "loss": 3.1213, "step": 19 }, { "epoch": 1.3508771929824561, "grad_norm": 1.9912149906158447, "learning_rate": 3.8e-05, "loss": 3.0372, "step": 20 }, { "epoch": 1.4210526315789473, "grad_norm": 1.7372136116027832, "learning_rate": 4e-05, "loss": 2.948, "step": 21 }, { "epoch": 1.4912280701754386, "grad_norm": 1.7806267738342285, "learning_rate": 4.2e-05, "loss": 2.9269, "step": 22 }, { "epoch": 1.5614035087719298, "grad_norm": 1.5322002172470093, "learning_rate": 4.4000000000000006e-05, "loss": 2.8877, "step": 23 }, { "epoch": 1.631578947368421, "grad_norm": 1.6473444700241089, "learning_rate": 4.600000000000001e-05, "loss": 2.7834, "step": 24 }, { "epoch": 1.7017543859649122, "grad_norm": 1.4832264184951782, "learning_rate": 4.8e-05, "loss": 2.7808, "step": 25 }, { "epoch": 1.7719298245614035, "grad_norm": 1.4326348304748535, "learning_rate": 5e-05, "loss": 2.6874, "step": 26 }, { "epoch": 1.8421052631578947, "grad_norm": 2.185389280319214, "learning_rate": 5.2000000000000004e-05, "loss": 2.667, "step": 27 }, { "epoch": 1.912280701754386, "grad_norm": 4.019981384277344, "learning_rate": 5.4000000000000005e-05, "loss": 2.6309, "step": 28 }, { "epoch": 1.9824561403508771, "grad_norm": 1.368363618850708, "learning_rate": 5.6000000000000006e-05, "loss": 2.5584, "step": 29 }, { "epoch": 2.0, "grad_norm": 3.5658273696899414, "learning_rate": 5.8e-05, "loss": 2.4491, "step": 30 }, { "epoch": 2.0701754385964914, "grad_norm": 2.8816487789154053, "learning_rate": 6e-05, "loss": 2.4853, "step": 31 }, { "epoch": 2.1403508771929824, "grad_norm": 4.591163158416748, "learning_rate": 6.2e-05, "loss": 2.4324, "step": 32 }, { "epoch": 2.2105263157894735, "grad_norm": 4.660356044769287, "learning_rate": 6.400000000000001e-05, "loss": 2.3511, "step": 33 }, { "epoch": 2.280701754385965, "grad_norm": 2.6735446453094482, "learning_rate": 6.6e-05, "loss": 2.3241, "step": 34 }, { "epoch": 2.3508771929824563, "grad_norm": 6.208858966827393, "learning_rate": 6.800000000000001e-05, "loss": 2.291, "step": 35 }, { "epoch": 2.4210526315789473, "grad_norm": 3.5636632442474365, "learning_rate": 7e-05, "loss": 2.2498, "step": 36 }, { "epoch": 2.4912280701754383, "grad_norm": 3.1954169273376465, "learning_rate": 7.2e-05, "loss": 2.23, "step": 37 }, { "epoch": 2.56140350877193, "grad_norm": 3.8919479846954346, "learning_rate": 7.4e-05, "loss": 2.2094, "step": 38 }, { "epoch": 2.6315789473684212, "grad_norm": 2.745952606201172, "learning_rate": 7.6e-05, "loss": 2.1835, "step": 39 }, { "epoch": 2.7017543859649122, "grad_norm": 2.4738833904266357, "learning_rate": 7.800000000000001e-05, "loss": 2.1099, "step": 40 }, { "epoch": 2.7719298245614032, "grad_norm": 5.345929145812988, "learning_rate": 8e-05, "loss": 2.0947, "step": 41 }, { "epoch": 2.8421052631578947, "grad_norm": 2.14803147315979, "learning_rate": 8.2e-05, "loss": 2.0623, "step": 42 } ], "logging_steps": 1, "max_steps": 42, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 200, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 7.301010223949414e+17, "train_batch_size": 4, "trial_name": null, "trial_params": null }