|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 4.985611510791367, |
|
"eval_steps": 500, |
|
"global_step": 1386, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.03597122302158273, |
|
"grad_norm": 3.388711929321289, |
|
"learning_rate": 2.857142857142857e-05, |
|
"loss": 1.2192, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.07194244604316546, |
|
"grad_norm": 7.04016637802124, |
|
"learning_rate": 5.714285714285714e-05, |
|
"loss": 0.4355, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.1079136690647482, |
|
"grad_norm": 1.9987560510635376, |
|
"learning_rate": 8.571428571428571e-05, |
|
"loss": 0.2583, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.14388489208633093, |
|
"grad_norm": 1.4927663803100586, |
|
"learning_rate": 0.00011428571428571428, |
|
"loss": 0.1959, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.17985611510791366, |
|
"grad_norm": 2.1035594940185547, |
|
"learning_rate": 0.00014285714285714287, |
|
"loss": 0.158, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.2158273381294964, |
|
"grad_norm": 1.8321762084960938, |
|
"learning_rate": 0.00017142857142857143, |
|
"loss": 0.1593, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.2517985611510791, |
|
"grad_norm": 0.9496012330055237, |
|
"learning_rate": 0.0002, |
|
"loss": 0.1172, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.28776978417266186, |
|
"grad_norm": 1.5062251091003418, |
|
"learning_rate": 0.0001999715070545774, |
|
"loss": 0.1394, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.3237410071942446, |
|
"grad_norm": 1.1900173425674438, |
|
"learning_rate": 0.00019988604445526827, |
|
"loss": 0.1144, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.3597122302158273, |
|
"grad_norm": 1.6312843561172485, |
|
"learning_rate": 0.0001997436609036963, |
|
"loss": 0.0954, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.39568345323741005, |
|
"grad_norm": 0.7715893387794495, |
|
"learning_rate": 0.00019954443753839667, |
|
"loss": 0.1043, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.4316546762589928, |
|
"grad_norm": 0.9805018901824951, |
|
"learning_rate": 0.00019928848788857887, |
|
"loss": 0.107, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.4676258992805755, |
|
"grad_norm": 0.79938805103302, |
|
"learning_rate": 0.00019897595780943102, |
|
"loss": 0.0939, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.5035971223021583, |
|
"grad_norm": 0.7441235184669495, |
|
"learning_rate": 0.00019860702539900287, |
|
"loss": 0.0876, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.539568345323741, |
|
"grad_norm": 1.4985952377319336, |
|
"learning_rate": 0.00019818190089671508, |
|
"loss": 0.0998, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.5755395683453237, |
|
"grad_norm": 0.49447914958000183, |
|
"learning_rate": 0.0001977008265635525, |
|
"loss": 0.084, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.6115107913669064, |
|
"grad_norm": 0.9330535531044006, |
|
"learning_rate": 0.00019716407654400952, |
|
"loss": 0.0711, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.6474820143884892, |
|
"grad_norm": 0.71977299451828, |
|
"learning_rate": 0.00019657195670986637, |
|
"loss": 0.0649, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.6834532374100719, |
|
"grad_norm": 0.6885129809379578, |
|
"learning_rate": 0.00019592480448588542, |
|
"loss": 0.0734, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.7194244604316546, |
|
"grad_norm": 0.8220723271369934, |
|
"learning_rate": 0.0001952229886575266, |
|
"loss": 0.0692, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.7553956834532374, |
|
"grad_norm": 0.4193010926246643, |
|
"learning_rate": 0.0001944669091607919, |
|
"loss": 0.0634, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.7913669064748201, |
|
"grad_norm": 0.5557839274406433, |
|
"learning_rate": 0.0001936569968543179, |
|
"loss": 0.0617, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.8273381294964028, |
|
"grad_norm": 0.6012473702430725, |
|
"learning_rate": 0.0001927937132738476, |
|
"loss": 0.0567, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.8633093525179856, |
|
"grad_norm": 0.5065937042236328, |
|
"learning_rate": 0.00019187755036921978, |
|
"loss": 0.0666, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.8992805755395683, |
|
"grad_norm": 0.8122988343238831, |
|
"learning_rate": 0.00019090903022402729, |
|
"loss": 0.0621, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.935251798561151, |
|
"grad_norm": 0.600527822971344, |
|
"learning_rate": 0.00018988870475810282, |
|
"loss": 0.0577, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.9712230215827338, |
|
"grad_norm": 0.41501474380493164, |
|
"learning_rate": 0.00018881715541300276, |
|
"loss": 0.0585, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 1.0071942446043165, |
|
"grad_norm": 0.7007383108139038, |
|
"learning_rate": 0.00018769499282066717, |
|
"loss": 0.0727, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 1.0431654676258992, |
|
"grad_norm": 0.4356389045715332, |
|
"learning_rate": 0.00018652285645544603, |
|
"loss": 0.0576, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 1.079136690647482, |
|
"grad_norm": 0.5948470234870911, |
|
"learning_rate": 0.00018530141426968902, |
|
"loss": 0.0641, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 1.1151079136690647, |
|
"grad_norm": 0.6867853403091431, |
|
"learning_rate": 0.00018403136231310684, |
|
"loss": 0.0535, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 1.1510791366906474, |
|
"grad_norm": 0.5092623829841614, |
|
"learning_rate": 0.00018271342433612113, |
|
"loss": 0.05, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 1.1870503597122302, |
|
"grad_norm": 0.6697306632995605, |
|
"learning_rate": 0.0001813483513774289, |
|
"loss": 0.0527, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 1.223021582733813, |
|
"grad_norm": 0.5235440731048584, |
|
"learning_rate": 0.0001799369213360163, |
|
"loss": 0.0477, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 1.2589928057553956, |
|
"grad_norm": 0.40158113837242126, |
|
"learning_rate": 0.0001784799385278661, |
|
"loss": 0.0587, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 1.2949640287769784, |
|
"grad_norm": 0.36186593770980835, |
|
"learning_rate": 0.000176978233227611, |
|
"loss": 0.0523, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 1.330935251798561, |
|
"grad_norm": 0.4480399489402771, |
|
"learning_rate": 0.00017543266119539422, |
|
"loss": 0.0413, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 1.3669064748201438, |
|
"grad_norm": 0.5076740980148315, |
|
"learning_rate": 0.00017384410318920697, |
|
"loss": 0.0465, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 1.4028776978417266, |
|
"grad_norm": 0.8270596265792847, |
|
"learning_rate": 0.0001722134644629807, |
|
"loss": 0.056, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 1.4388489208633093, |
|
"grad_norm": 0.520577609539032, |
|
"learning_rate": 0.00017054167425071995, |
|
"loss": 0.0401, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.474820143884892, |
|
"grad_norm": 0.7387260794639587, |
|
"learning_rate": 0.00016882968523697028, |
|
"loss": 0.0444, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 1.5107913669064748, |
|
"grad_norm": 0.34361258149147034, |
|
"learning_rate": 0.00016707847301392236, |
|
"loss": 0.0573, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 1.5467625899280577, |
|
"grad_norm": 0.49156174063682556, |
|
"learning_rate": 0.00016528903552546207, |
|
"loss": 0.0499, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 1.5827338129496402, |
|
"grad_norm": 0.37369754910469055, |
|
"learning_rate": 0.0001634623924984833, |
|
"loss": 0.047, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 1.6187050359712232, |
|
"grad_norm": 0.4199368357658386, |
|
"learning_rate": 0.0001615995848617876, |
|
"loss": 0.0422, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 1.6546762589928057, |
|
"grad_norm": 0.40282875299453735, |
|
"learning_rate": 0.0001597016741529014, |
|
"loss": 0.0388, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 1.6906474820143886, |
|
"grad_norm": 0.7056338787078857, |
|
"learning_rate": 0.0001577697419131497, |
|
"loss": 0.0453, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 1.7266187050359711, |
|
"grad_norm": 0.512008786201477, |
|
"learning_rate": 0.00015580488907132974, |
|
"loss": 0.0419, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 1.762589928057554, |
|
"grad_norm": 0.41075655817985535, |
|
"learning_rate": 0.00015380823531633729, |
|
"loss": 0.0465, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 1.7985611510791366, |
|
"grad_norm": 0.48174160718917847, |
|
"learning_rate": 0.0001517809184591017, |
|
"loss": 0.0386, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.8345323741007196, |
|
"grad_norm": 0.5647067427635193, |
|
"learning_rate": 0.0001497240937841944, |
|
"loss": 0.0425, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 1.870503597122302, |
|
"grad_norm": 0.4203431308269501, |
|
"learning_rate": 0.0001476389333914794, |
|
"loss": 0.0377, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 1.906474820143885, |
|
"grad_norm": 0.5694783329963684, |
|
"learning_rate": 0.0001455266255281821, |
|
"loss": 0.0374, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 1.9424460431654675, |
|
"grad_norm": 0.6108015179634094, |
|
"learning_rate": 0.00014338837391175582, |
|
"loss": 0.0417, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 1.9784172661870505, |
|
"grad_norm": 0.43549835681915283, |
|
"learning_rate": 0.00014122539704393265, |
|
"loss": 0.0339, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 2.014388489208633, |
|
"grad_norm": 0.4553022086620331, |
|
"learning_rate": 0.00013903892751634947, |
|
"loss": 0.0423, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 2.050359712230216, |
|
"grad_norm": 0.48167282342910767, |
|
"learning_rate": 0.0001368302113081447, |
|
"loss": 0.0411, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 2.0863309352517985, |
|
"grad_norm": 0.5101397037506104, |
|
"learning_rate": 0.0001346005070759258, |
|
"loss": 0.0495, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 2.1223021582733814, |
|
"grad_norm": 0.47834643721580505, |
|
"learning_rate": 0.00013235108543651272, |
|
"loss": 0.0352, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 2.158273381294964, |
|
"grad_norm": 0.4055221676826477, |
|
"learning_rate": 0.00013008322824286555, |
|
"loss": 0.0321, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 2.194244604316547, |
|
"grad_norm": 0.4231523275375366, |
|
"learning_rate": 0.00012779822785360912, |
|
"loss": 0.0467, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 2.2302158273381294, |
|
"grad_norm": 0.6262776851654053, |
|
"learning_rate": 0.00012549738639657115, |
|
"loss": 0.0363, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 2.2661870503597124, |
|
"grad_norm": 0.5341355800628662, |
|
"learning_rate": 0.00012318201502675285, |
|
"loss": 0.0328, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 2.302158273381295, |
|
"grad_norm": 0.2906142473220825, |
|
"learning_rate": 0.00012085343317915565, |
|
"loss": 0.0316, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 2.338129496402878, |
|
"grad_norm": 0.3438158929347992, |
|
"learning_rate": 0.00011851296781688952, |
|
"loss": 0.0315, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 2.3741007194244603, |
|
"grad_norm": 0.43991851806640625, |
|
"learning_rate": 0.00011616195267499102, |
|
"loss": 0.031, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 2.4100719424460433, |
|
"grad_norm": 0.3710590600967407, |
|
"learning_rate": 0.00011380172750038269, |
|
"loss": 0.0252, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 2.446043165467626, |
|
"grad_norm": 0.4207635223865509, |
|
"learning_rate": 0.00011143363728840625, |
|
"loss": 0.0409, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 2.4820143884892087, |
|
"grad_norm": 0.48165836930274963, |
|
"learning_rate": 0.00010905903151636501, |
|
"loss": 0.0282, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 2.5179856115107913, |
|
"grad_norm": 0.45046257972717285, |
|
"learning_rate": 0.00010667926337451217, |
|
"loss": 0.0257, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 2.553956834532374, |
|
"grad_norm": 0.4864603877067566, |
|
"learning_rate": 0.00010429568899492348, |
|
"loss": 0.0322, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 2.5899280575539567, |
|
"grad_norm": 0.3659379184246063, |
|
"learning_rate": 0.0001019096666786931, |
|
"loss": 0.0271, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 2.6258992805755397, |
|
"grad_norm": 0.2604960799217224, |
|
"learning_rate": 9.952255612189368e-05, |
|
"loss": 0.0326, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 2.661870503597122, |
|
"grad_norm": 0.4740797281265259, |
|
"learning_rate": 9.713571764074152e-05, |
|
"loss": 0.0306, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 2.697841726618705, |
|
"grad_norm": 0.2611696124076843, |
|
"learning_rate": 9.475051139640809e-05, |
|
"loss": 0.0271, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 2.7338129496402876, |
|
"grad_norm": 0.3587149679660797, |
|
"learning_rate": 9.236829661992023e-05, |
|
"loss": 0.0251, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 2.7697841726618706, |
|
"grad_norm": 0.4013247489929199, |
|
"learning_rate": 8.999043083759017e-05, |
|
"loss": 0.0256, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 2.805755395683453, |
|
"grad_norm": 0.40374016761779785, |
|
"learning_rate": 8.761826909741709e-05, |
|
"loss": 0.032, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 2.841726618705036, |
|
"grad_norm": 0.2673892080783844, |
|
"learning_rate": 8.525316319690092e-05, |
|
"loss": 0.0201, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 2.8776978417266186, |
|
"grad_norm": 0.44146105647087097, |
|
"learning_rate": 8.289646091270849e-05, |
|
"loss": 0.0305, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 2.9136690647482015, |
|
"grad_norm": 0.33300504088401794, |
|
"learning_rate": 8.054950523263096e-05, |
|
"loss": 0.0219, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 2.949640287769784, |
|
"grad_norm": 0.26247546076774597, |
|
"learning_rate": 7.821363359027048e-05, |
|
"loss": 0.0359, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 2.985611510791367, |
|
"grad_norm": 0.5542314052581787, |
|
"learning_rate": 7.589017710289139e-05, |
|
"loss": 0.0286, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 3.0215827338129495, |
|
"grad_norm": 0.4509626030921936, |
|
"learning_rate": 7.358045981287141e-05, |
|
"loss": 0.0234, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 3.0575539568345325, |
|
"grad_norm": 0.2706989049911499, |
|
"learning_rate": 7.128579793318428e-05, |
|
"loss": 0.0225, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 3.093525179856115, |
|
"grad_norm": 0.21301843225955963, |
|
"learning_rate": 6.900749909734406e-05, |
|
"loss": 0.0274, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 3.129496402877698, |
|
"grad_norm": 0.32899874448776245, |
|
"learning_rate": 6.674686161423843e-05, |
|
"loss": 0.0234, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 3.1654676258992804, |
|
"grad_norm": 0.4368799924850464, |
|
"learning_rate": 6.450517372827591e-05, |
|
"loss": 0.0199, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 3.2014388489208634, |
|
"grad_norm": 0.36518239974975586, |
|
"learning_rate": 6.22837128852683e-05, |
|
"loss": 0.0218, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 3.237410071942446, |
|
"grad_norm": 0.20987823605537415, |
|
"learning_rate": 6.008374500446676e-05, |
|
"loss": 0.0217, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 3.273381294964029, |
|
"grad_norm": 0.23875588178634644, |
|
"learning_rate": 5.790652375716652e-05, |
|
"loss": 0.0237, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 3.3093525179856114, |
|
"grad_norm": 0.320413738489151, |
|
"learning_rate": 5.575328985229098e-05, |
|
"loss": 0.0279, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 3.3453237410071943, |
|
"grad_norm": 0.20259268581867218, |
|
"learning_rate": 5.362527032936277e-05, |
|
"loss": 0.0265, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 3.381294964028777, |
|
"grad_norm": 0.22162136435508728, |
|
"learning_rate": 5.1523677859264516e-05, |
|
"loss": 0.0256, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 3.41726618705036, |
|
"grad_norm": 0.15575292706489563, |
|
"learning_rate": 4.944971005318716e-05, |
|
"loss": 0.0194, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 3.4532374100719423, |
|
"grad_norm": 0.29202401638031006, |
|
"learning_rate": 4.740454878016084e-05, |
|
"loss": 0.0276, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 3.4892086330935252, |
|
"grad_norm": 0.40271732211112976, |
|
"learning_rate": 4.538935949355623e-05, |
|
"loss": 0.0183, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 3.5251798561151078, |
|
"grad_norm": 0.22273583710193634, |
|
"learning_rate": 4.340529056694047e-05, |
|
"loss": 0.02, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 3.5611510791366907, |
|
"grad_norm": 0.1683393120765686, |
|
"learning_rate": 4.1453472639666457e-05, |
|
"loss": 0.0177, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 3.597122302158273, |
|
"grad_norm": 0.4574311375617981, |
|
"learning_rate": 3.9535017972567675e-05, |
|
"loss": 0.0205, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 3.633093525179856, |
|
"grad_norm": 0.31037506461143494, |
|
"learning_rate": 3.7651019814126654e-05, |
|
"loss": 0.0265, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 3.6690647482014387, |
|
"grad_norm": 0.349869966506958, |
|
"learning_rate": 3.5802551777477476e-05, |
|
"loss": 0.0153, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 3.7050359712230216, |
|
"grad_norm": 0.2690473794937134, |
|
"learning_rate": 3.399066722859782e-05, |
|
"loss": 0.0205, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 3.741007194244604, |
|
"grad_norm": 0.23748761415481567, |
|
"learning_rate": 3.2216398686038926e-05, |
|
"loss": 0.0145, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 3.776978417266187, |
|
"grad_norm": 0.24303750693798065, |
|
"learning_rate": 3.0480757232535772e-05, |
|
"loss": 0.0207, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 3.81294964028777, |
|
"grad_norm": 0.1993425041437149, |
|
"learning_rate": 2.8784731938832556e-05, |
|
"loss": 0.0237, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 3.8489208633093526, |
|
"grad_norm": 0.2566690444946289, |
|
"learning_rate": 2.7129289300051787e-05, |
|
"loss": 0.0194, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 3.884892086330935, |
|
"grad_norm": 0.2915847599506378, |
|
"learning_rate": 2.5515372684928683e-05, |
|
"loss": 0.0204, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 3.920863309352518, |
|
"grad_norm": 0.21342326700687408, |
|
"learning_rate": 2.394390179822382e-05, |
|
"loss": 0.0157, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 3.956834532374101, |
|
"grad_norm": 0.23817431926727295, |
|
"learning_rate": 2.2415772156621382e-05, |
|
"loss": 0.0224, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 3.9928057553956835, |
|
"grad_norm": 0.2008386105298996, |
|
"learning_rate": 2.0931854578410905e-05, |
|
"loss": 0.0128, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 4.028776978417266, |
|
"grad_norm": 0.16650381684303284, |
|
"learning_rate": 1.9492994687243714e-05, |
|
"loss": 0.0144, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 4.0647482014388485, |
|
"grad_norm": 0.20245954394340515, |
|
"learning_rate": 1.8100012430246837e-05, |
|
"loss": 0.0188, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 4.100719424460432, |
|
"grad_norm": 0.21814769506454468, |
|
"learning_rate": 1.6753701610768724e-05, |
|
"loss": 0.0128, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 4.136690647482014, |
|
"grad_norm": 0.10709039866924286, |
|
"learning_rate": 1.545482943602341e-05, |
|
"loss": 0.0191, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 4.172661870503597, |
|
"grad_norm": 0.14134275913238525, |
|
"learning_rate": 1.4204136079890584e-05, |
|
"loss": 0.0124, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 4.2086330935251794, |
|
"grad_norm": 0.33789917826652527, |
|
"learning_rate": 1.300233426112103e-05, |
|
"loss": 0.0152, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 4.244604316546763, |
|
"grad_norm": 0.17584295570850372, |
|
"learning_rate": 1.1850108837187335e-05, |
|
"loss": 0.0162, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 4.280575539568345, |
|
"grad_norm": 0.17790471017360687, |
|
"learning_rate": 1.0748116414011888e-05, |
|
"loss": 0.0189, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 4.316546762589928, |
|
"grad_norm": 0.2450784593820572, |
|
"learning_rate": 9.696984971794065e-06, |
|
"loss": 0.0151, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 4.35251798561151, |
|
"grad_norm": 0.4205743670463562, |
|
"learning_rate": 8.697313507150184e-06, |
|
"loss": 0.0228, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 4.388489208633094, |
|
"grad_norm": 0.15447357296943665, |
|
"learning_rate": 7.749671691769911e-06, |
|
"loss": 0.0174, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 4.424460431654676, |
|
"grad_norm": 0.1472749412059784, |
|
"learning_rate": 6.854599547783736e-06, |
|
"loss": 0.0191, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 4.460431654676259, |
|
"grad_norm": 0.16755138337612152, |
|
"learning_rate": 6.012607140026605e-06, |
|
"loss": 0.0153, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 4.496402877697841, |
|
"grad_norm": 0.16813568770885468, |
|
"learning_rate": 5.224174285372974e-06, |
|
"loss": 0.0181, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 4.532374100719425, |
|
"grad_norm": 0.10466030985116959, |
|
"learning_rate": 4.489750279308757e-06, |
|
"loss": 0.0105, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 4.568345323741007, |
|
"grad_norm": 0.1699487268924713, |
|
"learning_rate": 3.8097536398963963e-06, |
|
"loss": 0.0117, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 4.60431654676259, |
|
"grad_norm": 0.19925230741500854, |
|
"learning_rate": 3.184571869278574e-06, |
|
"loss": 0.0155, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 4.640287769784173, |
|
"grad_norm": 0.19352638721466064, |
|
"learning_rate": 2.6145612328566717e-06, |
|
"loss": 0.0118, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 4.676258992805756, |
|
"grad_norm": 0.3062892556190491, |
|
"learning_rate": 2.1000465562697856e-06, |
|
"loss": 0.0145, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 4.712230215827338, |
|
"grad_norm": 0.22001531720161438, |
|
"learning_rate": 1.6413210402898893e-06, |
|
"loss": 0.0134, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 4.748201438848921, |
|
"grad_norm": 0.2228756546974182, |
|
"learning_rate": 1.2386460937387822e-06, |
|
"loss": 0.0106, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 4.784172661870503, |
|
"grad_norm": 0.3254740834236145, |
|
"learning_rate": 8.922511845219971e-07, |
|
"loss": 0.0217, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 4.820143884892087, |
|
"grad_norm": 0.23890721797943115, |
|
"learning_rate": 6.023337088643665e-07, |
|
"loss": 0.0217, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 4.856115107913669, |
|
"grad_norm": 0.1692313402891159, |
|
"learning_rate": 3.6905887882213717e-07, |
|
"loss": 0.0171, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 4.892086330935252, |
|
"grad_norm": 0.1608191728591919, |
|
"learning_rate": 1.925596281353026e-07, |
|
"loss": 0.0124, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 4.928057553956835, |
|
"grad_norm": 0.1591498851776123, |
|
"learning_rate": 7.293653647421073e-08, |
|
"loss": 0.0223, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 4.9640287769784175, |
|
"grad_norm": 0.16027827560901642, |
|
"learning_rate": 1.0257772123312137e-08, |
|
"loss": 0.0149, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 4.985611510791367, |
|
"step": 1386, |
|
"total_flos": 4.876041943922688e+16, |
|
"train_loss": 0.05376770623662599, |
|
"train_runtime": 634.0011, |
|
"train_samples_per_second": 34.978, |
|
"train_steps_per_second": 2.186 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 1386, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 5, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 4.876041943922688e+16, |
|
"train_batch_size": 16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|