|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.0, |
|
"eval_steps": 500, |
|
"global_step": 1682, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0011890606420927466, |
|
"grad_norm": 115.31934356689453, |
|
"learning_rate": 4.9970273483947685e-05, |
|
"loss": 9.3546, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.034482758620689655, |
|
"grad_norm": 18.599306106567383, |
|
"learning_rate": 4.913793103448276e-05, |
|
"loss": 2.4091, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.06896551724137931, |
|
"grad_norm": 12.393567085266113, |
|
"learning_rate": 4.827586206896552e-05, |
|
"loss": 1.1241, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.10344827586206896, |
|
"grad_norm": 50.094051361083984, |
|
"learning_rate": 4.741379310344828e-05, |
|
"loss": 1.0125, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.13793103448275862, |
|
"grad_norm": 39.95038986206055, |
|
"learning_rate": 4.655172413793104e-05, |
|
"loss": 0.9273, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.1724137931034483, |
|
"grad_norm": 13.281441688537598, |
|
"learning_rate": 4.5689655172413794e-05, |
|
"loss": 0.8734, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.20689655172413793, |
|
"grad_norm": 29.820255279541016, |
|
"learning_rate": 4.482758620689655e-05, |
|
"loss": 0.9182, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.2413793103448276, |
|
"grad_norm": 14.038665771484375, |
|
"learning_rate": 4.396551724137931e-05, |
|
"loss": 0.9129, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.27586206896551724, |
|
"grad_norm": 13.406801223754883, |
|
"learning_rate": 4.3103448275862066e-05, |
|
"loss": 0.7973, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.3103448275862069, |
|
"grad_norm": 15.321793556213379, |
|
"learning_rate": 4.224137931034483e-05, |
|
"loss": 0.8415, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.3448275862068966, |
|
"grad_norm": 27.121288299560547, |
|
"learning_rate": 4.1379310344827587e-05, |
|
"loss": 0.7995, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.3793103448275862, |
|
"grad_norm": 36.715858459472656, |
|
"learning_rate": 4.0517241379310344e-05, |
|
"loss": 0.6301, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 0.41379310344827586, |
|
"grad_norm": 16.650150299072266, |
|
"learning_rate": 3.965517241379311e-05, |
|
"loss": 0.5675, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 0.4482758620689655, |
|
"grad_norm": 230.8839569091797, |
|
"learning_rate": 3.8793103448275865e-05, |
|
"loss": 0.6154, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 0.4827586206896552, |
|
"grad_norm": 6.414034366607666, |
|
"learning_rate": 3.793103448275862e-05, |
|
"loss": 0.6439, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 0.5172413793103449, |
|
"grad_norm": 9.441433906555176, |
|
"learning_rate": 3.7068965517241385e-05, |
|
"loss": 0.5776, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.5517241379310345, |
|
"grad_norm": 23.88970947265625, |
|
"learning_rate": 3.620689655172414e-05, |
|
"loss": 0.6126, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 0.5862068965517241, |
|
"grad_norm": 10.72873592376709, |
|
"learning_rate": 3.53448275862069e-05, |
|
"loss": 0.529, |
|
"step": 493 |
|
}, |
|
{ |
|
"epoch": 0.6206896551724138, |
|
"grad_norm": 22.285120010375977, |
|
"learning_rate": 3.4482758620689657e-05, |
|
"loss": 0.5418, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 0.6551724137931034, |
|
"grad_norm": 28.53759002685547, |
|
"learning_rate": 3.3620689655172414e-05, |
|
"loss": 0.4864, |
|
"step": 551 |
|
}, |
|
{ |
|
"epoch": 0.6896551724137931, |
|
"grad_norm": 15.215431213378906, |
|
"learning_rate": 3.275862068965517e-05, |
|
"loss": 0.5474, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.7241379310344828, |
|
"grad_norm": 5.726695537567139, |
|
"learning_rate": 3.1896551724137935e-05, |
|
"loss": 0.4492, |
|
"step": 609 |
|
}, |
|
{ |
|
"epoch": 0.7586206896551724, |
|
"grad_norm": 20.181598663330078, |
|
"learning_rate": 3.103448275862069e-05, |
|
"loss": 0.4925, |
|
"step": 638 |
|
}, |
|
{ |
|
"epoch": 0.7931034482758621, |
|
"grad_norm": 16.79625701904297, |
|
"learning_rate": 3.017241379310345e-05, |
|
"loss": 0.4405, |
|
"step": 667 |
|
}, |
|
{ |
|
"epoch": 0.8275862068965517, |
|
"grad_norm": 5.420645236968994, |
|
"learning_rate": 2.9310344827586206e-05, |
|
"loss": 0.5411, |
|
"step": 696 |
|
}, |
|
{ |
|
"epoch": 0.8620689655172413, |
|
"grad_norm": 6.218174457550049, |
|
"learning_rate": 2.844827586206897e-05, |
|
"loss": 0.5516, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.896551724137931, |
|
"grad_norm": 27.71928596496582, |
|
"learning_rate": 2.7586206896551727e-05, |
|
"loss": 0.4341, |
|
"step": 754 |
|
}, |
|
{ |
|
"epoch": 0.9310344827586207, |
|
"grad_norm": 31.43971824645996, |
|
"learning_rate": 2.672413793103448e-05, |
|
"loss": 0.3883, |
|
"step": 783 |
|
}, |
|
{ |
|
"epoch": 0.9655172413793104, |
|
"grad_norm": 7.160225868225098, |
|
"learning_rate": 2.5862068965517244e-05, |
|
"loss": 0.4214, |
|
"step": 812 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 7.139071464538574, |
|
"learning_rate": 2.5e-05, |
|
"loss": 0.3728, |
|
"step": 841 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_cer": 0.019845431962462048, |
|
"eval_loss": 0.2980102300643921, |
|
"eval_runtime": 887.0299, |
|
"eval_samples_per_second": 1.896, |
|
"eval_steps_per_second": 0.238, |
|
"step": 841 |
|
}, |
|
{ |
|
"epoch": 1.0344827586206897, |
|
"grad_norm": 14.740863800048828, |
|
"learning_rate": 2.413793103448276e-05, |
|
"loss": 0.3343, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 1.0689655172413792, |
|
"grad_norm": 22.295618057250977, |
|
"learning_rate": 2.327586206896552e-05, |
|
"loss": 0.3419, |
|
"step": 899 |
|
}, |
|
{ |
|
"epoch": 1.103448275862069, |
|
"grad_norm": 2.4551053047180176, |
|
"learning_rate": 2.2413793103448276e-05, |
|
"loss": 0.3123, |
|
"step": 928 |
|
}, |
|
{ |
|
"epoch": 1.1379310344827587, |
|
"grad_norm": 7.861540794372559, |
|
"learning_rate": 2.1551724137931033e-05, |
|
"loss": 0.349, |
|
"step": 957 |
|
}, |
|
{ |
|
"epoch": 1.1724137931034484, |
|
"grad_norm": 7.037637233734131, |
|
"learning_rate": 2.0689655172413793e-05, |
|
"loss": 0.2869, |
|
"step": 986 |
|
}, |
|
{ |
|
"epoch": 1.206896551724138, |
|
"grad_norm": 11.390913009643555, |
|
"learning_rate": 1.9827586206896554e-05, |
|
"loss": 0.2826, |
|
"step": 1015 |
|
}, |
|
{ |
|
"epoch": 1.2413793103448276, |
|
"grad_norm": 2.8627371788024902, |
|
"learning_rate": 1.896551724137931e-05, |
|
"loss": 0.2573, |
|
"step": 1044 |
|
}, |
|
{ |
|
"epoch": 1.2758620689655173, |
|
"grad_norm": 16.846853256225586, |
|
"learning_rate": 1.810344827586207e-05, |
|
"loss": 0.2752, |
|
"step": 1073 |
|
}, |
|
{ |
|
"epoch": 1.3103448275862069, |
|
"grad_norm": 3.9017491340637207, |
|
"learning_rate": 1.7241379310344828e-05, |
|
"loss": 0.2555, |
|
"step": 1102 |
|
}, |
|
{ |
|
"epoch": 1.3448275862068966, |
|
"grad_norm": 5.074669361114502, |
|
"learning_rate": 1.6379310344827585e-05, |
|
"loss": 0.2452, |
|
"step": 1131 |
|
}, |
|
{ |
|
"epoch": 1.3793103448275863, |
|
"grad_norm": 18.69165802001953, |
|
"learning_rate": 1.5517241379310346e-05, |
|
"loss": 0.2666, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 1.4137931034482758, |
|
"grad_norm": 5.539306640625, |
|
"learning_rate": 1.4655172413793103e-05, |
|
"loss": 0.2741, |
|
"step": 1189 |
|
}, |
|
{ |
|
"epoch": 1.4482758620689655, |
|
"grad_norm": 5.758066654205322, |
|
"learning_rate": 1.3793103448275863e-05, |
|
"loss": 0.2365, |
|
"step": 1218 |
|
}, |
|
{ |
|
"epoch": 1.4827586206896552, |
|
"grad_norm": 4.248748779296875, |
|
"learning_rate": 1.2931034482758622e-05, |
|
"loss": 0.2521, |
|
"step": 1247 |
|
}, |
|
{ |
|
"epoch": 1.5172413793103448, |
|
"grad_norm": 11.129036903381348, |
|
"learning_rate": 1.206896551724138e-05, |
|
"loss": 0.2257, |
|
"step": 1276 |
|
}, |
|
{ |
|
"epoch": 1.5517241379310345, |
|
"grad_norm": 2.8656809329986572, |
|
"learning_rate": 1.1206896551724138e-05, |
|
"loss": 0.2106, |
|
"step": 1305 |
|
}, |
|
{ |
|
"epoch": 1.5862068965517242, |
|
"grad_norm": 2.041245460510254, |
|
"learning_rate": 1.0344827586206897e-05, |
|
"loss": 0.227, |
|
"step": 1334 |
|
}, |
|
{ |
|
"epoch": 1.6206896551724137, |
|
"grad_norm": 1.451377034187317, |
|
"learning_rate": 9.482758620689655e-06, |
|
"loss": 0.1961, |
|
"step": 1363 |
|
}, |
|
{ |
|
"epoch": 1.6551724137931034, |
|
"grad_norm": 4.332240104675293, |
|
"learning_rate": 8.620689655172414e-06, |
|
"loss": 0.195, |
|
"step": 1392 |
|
}, |
|
{ |
|
"epoch": 1.6896551724137931, |
|
"grad_norm": 5.428757667541504, |
|
"learning_rate": 7.758620689655173e-06, |
|
"loss": 0.1929, |
|
"step": 1421 |
|
}, |
|
{ |
|
"epoch": 1.7241379310344827, |
|
"grad_norm": 2.5995798110961914, |
|
"learning_rate": 6.896551724137932e-06, |
|
"loss": 0.1991, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 1.7586206896551724, |
|
"grad_norm": 2.2909860610961914, |
|
"learning_rate": 6.03448275862069e-06, |
|
"loss": 0.168, |
|
"step": 1479 |
|
}, |
|
{ |
|
"epoch": 1.793103448275862, |
|
"grad_norm": 2.049956798553467, |
|
"learning_rate": 5.172413793103448e-06, |
|
"loss": 0.18, |
|
"step": 1508 |
|
}, |
|
{ |
|
"epoch": 1.8275862068965516, |
|
"grad_norm": 2.291414737701416, |
|
"learning_rate": 4.310344827586207e-06, |
|
"loss": 0.1642, |
|
"step": 1537 |
|
}, |
|
{ |
|
"epoch": 1.8620689655172413, |
|
"grad_norm": 6.125259876251221, |
|
"learning_rate": 3.448275862068966e-06, |
|
"loss": 0.197, |
|
"step": 1566 |
|
}, |
|
{ |
|
"epoch": 1.896551724137931, |
|
"grad_norm": 2.359379291534424, |
|
"learning_rate": 2.586206896551724e-06, |
|
"loss": 0.1615, |
|
"step": 1595 |
|
}, |
|
{ |
|
"epoch": 1.9310344827586206, |
|
"grad_norm": 2.1470141410827637, |
|
"learning_rate": 1.724137931034483e-06, |
|
"loss": 0.1669, |
|
"step": 1624 |
|
}, |
|
{ |
|
"epoch": 1.9655172413793105, |
|
"grad_norm": 13.040664672851562, |
|
"learning_rate": 8.620689655172415e-07, |
|
"loss": 0.1641, |
|
"step": 1653 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 1.5605217218399048, |
|
"learning_rate": 0.0, |
|
"loss": 0.1551, |
|
"step": 1682 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_cer": 0.003201766491857577, |
|
"eval_loss": 0.17416498064994812, |
|
"eval_runtime": 882.6649, |
|
"eval_samples_per_second": 1.906, |
|
"eval_steps_per_second": 0.239, |
|
"step": 1682 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"step": 1682, |
|
"total_flos": 1.9906356553640313e+19, |
|
"train_loss": 0.47363961749354666, |
|
"train_runtime": 3260.3696, |
|
"train_samples_per_second": 4.125, |
|
"train_steps_per_second": 0.516 |
|
} |
|
], |
|
"logging_steps": 29, |
|
"max_steps": 1682, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 2, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.9906356553640313e+19, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|