|
{ |
|
"best_global_step": null, |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 20.0, |
|
"eval_steps": 500, |
|
"global_step": 820, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.24691358024691357, |
|
"grad_norm": 3.9386653900146484, |
|
"learning_rate": 4.945121951219512e-05, |
|
"loss": 2.6292, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.49382716049382713, |
|
"grad_norm": 2.194047451019287, |
|
"learning_rate": 4.884146341463415e-05, |
|
"loss": 1.7588, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.7407407407407407, |
|
"grad_norm": 3.3464252948760986, |
|
"learning_rate": 4.823170731707317e-05, |
|
"loss": 1.2321, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.9876543209876543, |
|
"grad_norm": 4.9754157066345215, |
|
"learning_rate": 4.76219512195122e-05, |
|
"loss": 1.1094, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 1.2222222222222223, |
|
"grad_norm": 1.6857168674468994, |
|
"learning_rate": 4.701219512195122e-05, |
|
"loss": 0.8016, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 1.4691358024691357, |
|
"grad_norm": 1.711855173110962, |
|
"learning_rate": 4.640243902439025e-05, |
|
"loss": 0.6045, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 1.7160493827160495, |
|
"grad_norm": 3.110994338989258, |
|
"learning_rate": 4.5792682926829275e-05, |
|
"loss": 0.611, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 1.9629629629629628, |
|
"grad_norm": 2.2832303047180176, |
|
"learning_rate": 4.5182926829268296e-05, |
|
"loss": 0.4029, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 2.197530864197531, |
|
"grad_norm": 2.3041841983795166, |
|
"learning_rate": 4.457317073170732e-05, |
|
"loss": 0.4292, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 2.4444444444444446, |
|
"grad_norm": 1.5648462772369385, |
|
"learning_rate": 4.3963414634146346e-05, |
|
"loss": 0.3602, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 2.691358024691358, |
|
"grad_norm": 0.8374887704849243, |
|
"learning_rate": 4.335365853658537e-05, |
|
"loss": 0.309, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 2.9382716049382713, |
|
"grad_norm": 0.9409658908843994, |
|
"learning_rate": 4.2743902439024395e-05, |
|
"loss": 0.2612, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 3.1728395061728394, |
|
"grad_norm": 3.718695878982544, |
|
"learning_rate": 4.2134146341463416e-05, |
|
"loss": 0.2936, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 3.419753086419753, |
|
"grad_norm": 3.2449347972869873, |
|
"learning_rate": 4.152439024390244e-05, |
|
"loss": 0.1758, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 3.6666666666666665, |
|
"grad_norm": 1.2654057741165161, |
|
"learning_rate": 4.0914634146341465e-05, |
|
"loss": 0.2181, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 3.9135802469135803, |
|
"grad_norm": 0.7008330225944519, |
|
"learning_rate": 4.030487804878049e-05, |
|
"loss": 0.1788, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 4.148148148148148, |
|
"grad_norm": 0.927452802658081, |
|
"learning_rate": 3.9695121951219514e-05, |
|
"loss": 0.2255, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 4.395061728395062, |
|
"grad_norm": 0.9944292902946472, |
|
"learning_rate": 3.908536585365854e-05, |
|
"loss": 0.2719, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 4.6419753086419755, |
|
"grad_norm": 1.1106724739074707, |
|
"learning_rate": 3.847560975609756e-05, |
|
"loss": 0.1458, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 4.888888888888889, |
|
"grad_norm": 3.588369369506836, |
|
"learning_rate": 3.786585365853659e-05, |
|
"loss": 0.2225, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 5.1234567901234565, |
|
"grad_norm": 14.498685836791992, |
|
"learning_rate": 3.725609756097561e-05, |
|
"loss": 0.1737, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 5.37037037037037, |
|
"grad_norm": 15.858694076538086, |
|
"learning_rate": 3.664634146341463e-05, |
|
"loss": 0.1728, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 5.617283950617284, |
|
"grad_norm": 3.0714364051818848, |
|
"learning_rate": 3.603658536585366e-05, |
|
"loss": 0.0774, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 5.864197530864198, |
|
"grad_norm": 0.775265634059906, |
|
"learning_rate": 3.542682926829268e-05, |
|
"loss": 0.2107, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 6.098765432098766, |
|
"grad_norm": 1.262907862663269, |
|
"learning_rate": 3.48170731707317e-05, |
|
"loss": 0.1312, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 6.345679012345679, |
|
"grad_norm": 0.7409253120422363, |
|
"learning_rate": 3.420731707317074e-05, |
|
"loss": 0.1074, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 6.592592592592593, |
|
"grad_norm": 2.9899144172668457, |
|
"learning_rate": 3.359756097560976e-05, |
|
"loss": 0.2688, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 6.839506172839506, |
|
"grad_norm": 0.6718606352806091, |
|
"learning_rate": 3.298780487804878e-05, |
|
"loss": 0.0848, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 7.074074074074074, |
|
"grad_norm": 0.3312334716320038, |
|
"learning_rate": 3.237804878048781e-05, |
|
"loss": 0.0607, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 7.320987654320987, |
|
"grad_norm": 2.2608187198638916, |
|
"learning_rate": 3.176829268292683e-05, |
|
"loss": 0.1274, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 7.567901234567901, |
|
"grad_norm": 0.7812129855155945, |
|
"learning_rate": 3.115853658536586e-05, |
|
"loss": 0.0824, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 7.814814814814815, |
|
"grad_norm": 0.8673564195632935, |
|
"learning_rate": 3.054878048780488e-05, |
|
"loss": 0.1774, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 8.049382716049383, |
|
"grad_norm": 1.7585481405258179, |
|
"learning_rate": 2.9939024390243903e-05, |
|
"loss": 0.1415, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 8.296296296296296, |
|
"grad_norm": 2.8376216888427734, |
|
"learning_rate": 2.9329268292682927e-05, |
|
"loss": 0.1173, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 8.54320987654321, |
|
"grad_norm": 1.6730660200119019, |
|
"learning_rate": 2.8719512195121952e-05, |
|
"loss": 0.1585, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 8.790123456790123, |
|
"grad_norm": 0.30232733488082886, |
|
"learning_rate": 2.810975609756098e-05, |
|
"loss": 0.0377, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 9.024691358024691, |
|
"grad_norm": 2.1726369857788086, |
|
"learning_rate": 2.7500000000000004e-05, |
|
"loss": 0.0568, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 9.271604938271604, |
|
"grad_norm": 0.7406997084617615, |
|
"learning_rate": 2.689024390243903e-05, |
|
"loss": 0.105, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 9.518518518518519, |
|
"grad_norm": 0.2634631097316742, |
|
"learning_rate": 2.628048780487805e-05, |
|
"loss": 0.0328, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 9.765432098765432, |
|
"grad_norm": 0.15690064430236816, |
|
"learning_rate": 2.5670731707317075e-05, |
|
"loss": 0.069, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"grad_norm": 0.09763717651367188, |
|
"learning_rate": 2.50609756097561e-05, |
|
"loss": 0.1503, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 10.246913580246913, |
|
"grad_norm": 0.9397181272506714, |
|
"learning_rate": 2.4451219512195124e-05, |
|
"loss": 0.1123, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 10.493827160493828, |
|
"grad_norm": 0.1911344975233078, |
|
"learning_rate": 2.3841463414634148e-05, |
|
"loss": 0.0439, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 10.74074074074074, |
|
"grad_norm": 2.133268117904663, |
|
"learning_rate": 2.3231707317073173e-05, |
|
"loss": 0.0466, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 10.987654320987655, |
|
"grad_norm": 3.681490898132324, |
|
"learning_rate": 2.2621951219512197e-05, |
|
"loss": 0.083, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 11.222222222222221, |
|
"grad_norm": 0.09520290791988373, |
|
"learning_rate": 2.2012195121951222e-05, |
|
"loss": 0.1275, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 11.469135802469136, |
|
"grad_norm": 0.18467915058135986, |
|
"learning_rate": 2.1402439024390243e-05, |
|
"loss": 0.0206, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 11.716049382716049, |
|
"grad_norm": 0.1692470908164978, |
|
"learning_rate": 2.0792682926829267e-05, |
|
"loss": 0.0694, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 11.962962962962964, |
|
"grad_norm": 0.21938467025756836, |
|
"learning_rate": 2.0182926829268295e-05, |
|
"loss": 0.1137, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 12.197530864197532, |
|
"grad_norm": 0.6872503161430359, |
|
"learning_rate": 1.957317073170732e-05, |
|
"loss": 0.0909, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 12.444444444444445, |
|
"grad_norm": 0.18933941423892975, |
|
"learning_rate": 1.896341463414634e-05, |
|
"loss": 0.0255, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 12.691358024691358, |
|
"grad_norm": 0.2120848298072815, |
|
"learning_rate": 1.8353658536585365e-05, |
|
"loss": 0.0383, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 12.938271604938272, |
|
"grad_norm": 0.12255409359931946, |
|
"learning_rate": 1.774390243902439e-05, |
|
"loss": 0.0941, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 13.17283950617284, |
|
"grad_norm": 0.10398901998996735, |
|
"learning_rate": 1.7134146341463418e-05, |
|
"loss": 0.0191, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 13.419753086419753, |
|
"grad_norm": 0.31278660893440247, |
|
"learning_rate": 1.652439024390244e-05, |
|
"loss": 0.0281, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 13.666666666666666, |
|
"grad_norm": 1.0805507898330688, |
|
"learning_rate": 1.5914634146341464e-05, |
|
"loss": 0.1726, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 13.91358024691358, |
|
"grad_norm": 0.4380980432033539, |
|
"learning_rate": 1.5304878048780488e-05, |
|
"loss": 0.0186, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 14.148148148148149, |
|
"grad_norm": 0.07801090180873871, |
|
"learning_rate": 1.4695121951219513e-05, |
|
"loss": 0.0473, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 14.395061728395062, |
|
"grad_norm": 0.10922332108020782, |
|
"learning_rate": 1.4085365853658535e-05, |
|
"loss": 0.023, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 14.641975308641975, |
|
"grad_norm": 0.2793682813644409, |
|
"learning_rate": 1.3475609756097562e-05, |
|
"loss": 0.0149, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 14.88888888888889, |
|
"grad_norm": 1.3744605779647827, |
|
"learning_rate": 1.2865853658536586e-05, |
|
"loss": 0.1422, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 15.123456790123457, |
|
"grad_norm": 0.1263941377401352, |
|
"learning_rate": 1.225609756097561e-05, |
|
"loss": 0.0124, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 15.37037037037037, |
|
"grad_norm": 0.13304823637008667, |
|
"learning_rate": 1.1646341463414635e-05, |
|
"loss": 0.0317, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 15.617283950617283, |
|
"grad_norm": 1.772745132446289, |
|
"learning_rate": 1.103658536585366e-05, |
|
"loss": 0.0443, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 15.864197530864198, |
|
"grad_norm": 0.06537698954343796, |
|
"learning_rate": 1.0426829268292683e-05, |
|
"loss": 0.0918, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 16.098765432098766, |
|
"grad_norm": 0.07521393895149231, |
|
"learning_rate": 9.817073170731709e-06, |
|
"loss": 0.0395, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 16.34567901234568, |
|
"grad_norm": 0.16562214493751526, |
|
"learning_rate": 9.207317073170732e-06, |
|
"loss": 0.1106, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 16.59259259259259, |
|
"grad_norm": 0.24475818872451782, |
|
"learning_rate": 8.597560975609756e-06, |
|
"loss": 0.019, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 16.839506172839506, |
|
"grad_norm": 0.32736074924468994, |
|
"learning_rate": 7.98780487804878e-06, |
|
"loss": 0.0326, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 17.074074074074073, |
|
"grad_norm": 0.05227570980787277, |
|
"learning_rate": 7.378048780487805e-06, |
|
"loss": 0.0103, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 17.320987654320987, |
|
"grad_norm": 0.09976907074451447, |
|
"learning_rate": 6.76829268292683e-06, |
|
"loss": 0.0864, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 17.567901234567902, |
|
"grad_norm": 0.04490213841199875, |
|
"learning_rate": 6.1585365853658535e-06, |
|
"loss": 0.0178, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 17.814814814814813, |
|
"grad_norm": 0.11479590833187103, |
|
"learning_rate": 5.548780487804878e-06, |
|
"loss": 0.0236, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 18.049382716049383, |
|
"grad_norm": 0.16496780514717102, |
|
"learning_rate": 4.9390243902439025e-06, |
|
"loss": 0.0481, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 18.296296296296298, |
|
"grad_norm": 6.131174564361572, |
|
"learning_rate": 4.329268292682927e-06, |
|
"loss": 0.0696, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 18.54320987654321, |
|
"grad_norm": 0.612872838973999, |
|
"learning_rate": 3.719512195121951e-06, |
|
"loss": 0.0194, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 18.790123456790123, |
|
"grad_norm": 1.3853808641433716, |
|
"learning_rate": 3.1097560975609757e-06, |
|
"loss": 0.037, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 19.02469135802469, |
|
"grad_norm": 0.06428851932287216, |
|
"learning_rate": 2.5e-06, |
|
"loss": 0.0269, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 19.271604938271604, |
|
"grad_norm": 0.15213467180728912, |
|
"learning_rate": 1.8902439024390245e-06, |
|
"loss": 0.017, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 19.51851851851852, |
|
"grad_norm": 2.0686116218566895, |
|
"learning_rate": 1.2804878048780488e-06, |
|
"loss": 0.088, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 19.765432098765434, |
|
"grad_norm": 0.05781777948141098, |
|
"learning_rate": 6.707317073170731e-07, |
|
"loss": 0.0367, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"grad_norm": 0.05907975882291794, |
|
"learning_rate": 6.097560975609757e-08, |
|
"loss": 0.0093, |
|
"step": 820 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 820, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 20, |
|
"save_steps": 100, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 430125247180800.0, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|