{ "best_metric": 0.049842845648527145, "best_model_checkpoint": "./trained_models/Llama-3.1-8B-Instruct-CNADFTD-ADNI2NIFD-AN-fold-0-gathered-equally-represented-0234-v2/checkpoint-65", "epoch": 9.91139240506329, "eval_steps": 5, "global_step": 90, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.20253164556962025, "grad_norm": 20.267499923706055, "learning_rate": 2.9999999999999997e-05, "loss": 1.7395, "step": 2 }, { "epoch": 0.4050632911392405, "grad_norm": 19.267681121826172, "learning_rate": 5.9999999999999995e-05, "loss": 1.5413, "step": 4 }, { "epoch": 0.5063291139240507, "eval_loss": 0.1354750096797943, "eval_runtime": 10.1119, "eval_samples_per_second": 8.9, "eval_steps_per_second": 2.275, "step": 5 }, { "epoch": 0.6075949367088608, "grad_norm": 9.370927810668945, "learning_rate": 8.999999999999999e-05, "loss": 0.928, "step": 6 }, { "epoch": 0.810126582278481, "grad_norm": 14.292084693908691, "learning_rate": 0.00011999999999999999, "loss": 1.0332, "step": 8 }, { "epoch": 1.1012658227848102, "grad_norm": 16.358558654785156, "learning_rate": 0.00015, "loss": 0.7069, "step": 10 }, { "epoch": 1.1012658227848102, "eval_loss": 0.09530380368232727, "eval_runtime": 10.229, "eval_samples_per_second": 8.799, "eval_steps_per_second": 2.249, "step": 10 }, { "epoch": 1.3037974683544304, "grad_norm": 4.635617733001709, "learning_rate": 0.00017999999999999998, "loss": 0.7555, "step": 12 }, { "epoch": 1.5063291139240507, "grad_norm": 5.011940002441406, "learning_rate": 0.00020999999999999998, "loss": 0.4148, "step": 14 }, { "epoch": 1.6075949367088609, "eval_loss": 0.07413829118013382, "eval_runtime": 10.2852, "eval_samples_per_second": 8.75, "eval_steps_per_second": 2.236, "step": 15 }, { "epoch": 1.7088607594936709, "grad_norm": 3.0944323539733887, "learning_rate": 0.00023999999999999998, "loss": 0.4899, "step": 16 }, { "epoch": 1.9113924050632911, "grad_norm": 9.65760612487793, "learning_rate": 0.00027, "loss": 0.7349, "step": 18 }, { "epoch": 2.2025316455696204, "grad_norm": 6.6734418869018555, "learning_rate": 0.0003, "loss": 0.7657, "step": 20 }, { "epoch": 2.2025316455696204, "eval_loss": 0.06096666678786278, "eval_runtime": 10.2992, "eval_samples_per_second": 8.739, "eval_steps_per_second": 2.233, "step": 20 }, { "epoch": 2.4050632911392404, "grad_norm": 4.764143943786621, "learning_rate": 0.00029939614409928584, "loss": 0.5037, "step": 22 }, { "epoch": 2.607594936708861, "grad_norm": 2.2680883407592773, "learning_rate": 0.00029758943828979444, "loss": 0.4673, "step": 24 }, { "epoch": 2.708860759493671, "eval_loss": 0.05865984410047531, "eval_runtime": 10.3046, "eval_samples_per_second": 8.734, "eval_steps_per_second": 2.232, "step": 25 }, { "epoch": 2.810126582278481, "grad_norm": 4.016191005706787, "learning_rate": 0.00029459442910437797, "loss": 0.4189, "step": 26 }, { "epoch": 3.1012658227848102, "grad_norm": 6.397407531738281, "learning_rate": 0.00029043523059596053, "loss": 0.6435, "step": 28 }, { "epoch": 3.3037974683544302, "grad_norm": 5.727280139923096, "learning_rate": 0.0002851453301853628, "loss": 0.3936, "step": 30 }, { "epoch": 3.3037974683544302, "eval_loss": 0.06180719658732414, "eval_runtime": 10.3031, "eval_samples_per_second": 8.735, "eval_steps_per_second": 2.232, "step": 30 }, { "epoch": 3.5063291139240507, "grad_norm": 4.997829437255859, "learning_rate": 0.0002787673190402799, "loss": 0.3523, "step": 32 }, { "epoch": 3.708860759493671, "grad_norm": 4.903127670288086, "learning_rate": 0.0002713525491562421, "loss": 0.5887, "step": 34 }, { "epoch": 3.810126582278481, "eval_loss": 0.05977346748113632, "eval_runtime": 10.2946, "eval_samples_per_second": 8.742, "eval_steps_per_second": 2.234, "step": 35 }, { "epoch": 3.911392405063291, "grad_norm": 6.872932434082031, "learning_rate": 0.00026296071990054165, "loss": 0.4239, "step": 36 }, { "epoch": 4.2025316455696204, "grad_norm": 3.7034521102905273, "learning_rate": 0.0002536593973480297, "loss": 0.437, "step": 38 }, { "epoch": 4.405063291139241, "grad_norm": 4.1444244384765625, "learning_rate": 0.00024352347027881003, "loss": 0.2975, "step": 40 }, { "epoch": 4.405063291139241, "eval_loss": 0.06828329712152481, "eval_runtime": 10.2997, "eval_samples_per_second": 8.738, "eval_steps_per_second": 2.233, "step": 40 }, { "epoch": 4.6075949367088604, "grad_norm": 4.268646717071533, "learning_rate": 0.00023263454721781537, "loss": 0.2971, "step": 42 }, { "epoch": 4.810126582278481, "grad_norm": 9.77409839630127, "learning_rate": 0.0002210802993709498, "loss": 0.3884, "step": 44 }, { "epoch": 4.911392405063291, "eval_loss": 0.0632724016904831, "eval_runtime": 10.2951, "eval_samples_per_second": 8.742, "eval_steps_per_second": 2.234, "step": 45 }, { "epoch": 5.10126582278481, "grad_norm": 12.521153450012207, "learning_rate": 0.00020895375474808852, "loss": 0.634, "step": 46 }, { "epoch": 5.30379746835443, "grad_norm": 7.4212541580200195, "learning_rate": 0.0001963525491562421, "loss": 0.3246, "step": 48 }, { "epoch": 5.506329113924051, "grad_norm": 10.269664764404297, "learning_rate": 0.00018337814009344714, "loss": 0.4664, "step": 50 }, { "epoch": 5.506329113924051, "eval_loss": 0.09271334856748581, "eval_runtime": 10.2899, "eval_samples_per_second": 8.746, "eval_steps_per_second": 2.235, "step": 50 }, { "epoch": 5.708860759493671, "grad_norm": 6.81392765045166, "learning_rate": 0.00017013498987264832, "loss": 0.3999, "step": 52 }, { "epoch": 5.911392405063291, "grad_norm": 3.8285505771636963, "learning_rate": 0.00015672972455257723, "loss": 0.2666, "step": 54 }, { "epoch": 6.10126582278481, "eval_loss": 0.05789216235280037, "eval_runtime": 10.2896, "eval_samples_per_second": 8.747, "eval_steps_per_second": 2.235, "step": 55 }, { "epoch": 6.2025316455696204, "grad_norm": 3.7714805603027344, "learning_rate": 0.0001432702754474228, "loss": 0.3953, "step": 56 }, { "epoch": 6.405063291139241, "grad_norm": 2.9867966175079346, "learning_rate": 0.0001298650101273517, "loss": 0.227, "step": 58 }, { "epoch": 6.6075949367088604, "grad_norm": 4.1205010414123535, "learning_rate": 0.00011662185990655284, "loss": 0.2554, "step": 60 }, { "epoch": 6.6075949367088604, "eval_loss": 0.061419058591127396, "eval_runtime": 10.27, "eval_samples_per_second": 8.763, "eval_steps_per_second": 2.24, "step": 60 }, { "epoch": 6.810126582278481, "grad_norm": 2.6789913177490234, "learning_rate": 0.0001036474508437579, "loss": 0.2873, "step": 62 }, { "epoch": 7.10126582278481, "grad_norm": 4.4857683181762695, "learning_rate": 9.104624525191145e-05, "loss": 0.2828, "step": 64 }, { "epoch": 7.2025316455696204, "eval_loss": 0.049842845648527145, "eval_runtime": 10.2776, "eval_samples_per_second": 8.757, "eval_steps_per_second": 2.238, "step": 65 }, { "epoch": 7.30379746835443, "grad_norm": 1.8388667106628418, "learning_rate": 7.89197006290502e-05, "loss": 0.1712, "step": 66 }, { "epoch": 7.506329113924051, "grad_norm": 3.90205454826355, "learning_rate": 6.736545278218463e-05, "loss": 0.1491, "step": 68 }, { "epoch": 7.708860759493671, "grad_norm": 4.954835891723633, "learning_rate": 5.6476529721189974e-05, "loss": 0.233, "step": 70 }, { "epoch": 7.708860759493671, "eval_loss": 0.05690564960241318, "eval_runtime": 10.2862, "eval_samples_per_second": 8.75, "eval_steps_per_second": 2.236, "step": 70 }, { "epoch": 7.911392405063291, "grad_norm": 4.313951015472412, "learning_rate": 4.63406026519703e-05, "loss": 0.182, "step": 72 }, { "epoch": 8.20253164556962, "grad_norm": 3.9724860191345215, "learning_rate": 3.7039280099458366e-05, "loss": 0.1574, "step": 74 }, { "epoch": 8.30379746835443, "eval_loss": 0.06441806256771088, "eval_runtime": 10.2783, "eval_samples_per_second": 8.756, "eval_steps_per_second": 2.238, "step": 75 }, { "epoch": 8.405063291139241, "grad_norm": 7.101994037628174, "learning_rate": 2.8647450843757897e-05, "loss": 0.2221, "step": 76 }, { "epoch": 8.60759493670886, "grad_norm": 4.0686845779418945, "learning_rate": 2.1232680959720082e-05, "loss": 0.1784, "step": 78 }, { "epoch": 8.810126582278482, "grad_norm": 3.0517356395721436, "learning_rate": 1.4854669814637143e-05, "loss": 0.1949, "step": 80 }, { "epoch": 8.810126582278482, "eval_loss": 0.05705536901950836, "eval_runtime": 10.2777, "eval_samples_per_second": 8.757, "eval_steps_per_second": 2.238, "step": 80 }, { "epoch": 9.10126582278481, "grad_norm": 5.039515018463135, "learning_rate": 9.564769404039419e-06, "loss": 0.1881, "step": 82 }, { "epoch": 9.30379746835443, "grad_norm": 1.9921464920043945, "learning_rate": 5.405570895622013e-06, "loss": 0.0491, "step": 84 }, { "epoch": 9.405063291139241, "eval_loss": 0.057273320853710175, "eval_runtime": 10.276, "eval_samples_per_second": 8.758, "eval_steps_per_second": 2.238, "step": 85 }, { "epoch": 9.50632911392405, "grad_norm": 3.6120450496673584, "learning_rate": 2.4105617102055496e-06, "loss": 0.2364, "step": 86 }, { "epoch": 9.708860759493671, "grad_norm": 2.058825731277466, "learning_rate": 6.038559007141397e-07, "loss": 0.119, "step": 88 }, { "epoch": 9.91139240506329, "grad_norm": 3.1621336936950684, "learning_rate": 0.0, "loss": 0.1392, "step": 90 }, { "epoch": 9.91139240506329, "eval_loss": 0.05724988505244255, "eval_runtime": 10.2941, "eval_samples_per_second": 8.743, "eval_steps_per_second": 2.234, "step": 90 } ], "logging_steps": 2, "max_steps": 90, "num_input_tokens_seen": 0, "num_train_epochs": 10, "save_steps": 5, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 8.413422188878234e+16, "train_batch_size": 4, "trial_name": null, "trial_params": null }