|
{ |
|
"best_global_step": 4870, |
|
"best_metric": 1.0858850479125977, |
|
"best_model_checkpoint": "/mnt/tank/scratch/rgurtsiev/workflow/STAGE1-DEMO/checkpoint-4870", |
|
"epoch": 10.0, |
|
"eval_steps": 500, |
|
"global_step": 4870, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.1026694045174538, |
|
"grad_norm": 1.1697688102722168, |
|
"learning_rate": 1.3415468856947298e-06, |
|
"loss": 3.6536, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.2053388090349076, |
|
"grad_norm": 1.3012523651123047, |
|
"learning_rate": 2.7104722792607808e-06, |
|
"loss": 3.6247, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.3080082135523614, |
|
"grad_norm": 1.3252615928649902, |
|
"learning_rate": 4.0793976728268315e-06, |
|
"loss": 3.5833, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.4106776180698152, |
|
"grad_norm": 1.2821773290634155, |
|
"learning_rate": 5.448323066392882e-06, |
|
"loss": 3.471, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.5133470225872689, |
|
"grad_norm": 1.5595322847366333, |
|
"learning_rate": 6.817248459958933e-06, |
|
"loss": 3.2428, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.6160164271047228, |
|
"grad_norm": 1.5284234285354614, |
|
"learning_rate": 8.186173853524983e-06, |
|
"loss": 2.9622, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.7186858316221766, |
|
"grad_norm": 1.7251927852630615, |
|
"learning_rate": 9.555099247091034e-06, |
|
"loss": 2.536, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.8213552361396304, |
|
"grad_norm": 1.3935151100158691, |
|
"learning_rate": 1.0924024640657084e-05, |
|
"loss": 2.0349, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.9240246406570842, |
|
"grad_norm": 0.7713799476623535, |
|
"learning_rate": 1.2292950034223137e-05, |
|
"loss": 1.7696, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 1.657446265220642, |
|
"eval_runtime": 396.4486, |
|
"eval_samples_per_second": 4.911, |
|
"eval_steps_per_second": 0.615, |
|
"step": 487 |
|
}, |
|
{ |
|
"epoch": 1.0266940451745379, |
|
"grad_norm": 0.8125890493392944, |
|
"learning_rate": 1.3661875427789186e-05, |
|
"loss": 1.6644, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.1293634496919918, |
|
"grad_norm": 0.7177293300628662, |
|
"learning_rate": 1.5030800821355238e-05, |
|
"loss": 1.5654, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 1.2320328542094456, |
|
"grad_norm": 0.9869002103805542, |
|
"learning_rate": 1.6399726214921288e-05, |
|
"loss": 1.543, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 1.3347022587268995, |
|
"grad_norm": 1.156468152999878, |
|
"learning_rate": 1.776865160848734e-05, |
|
"loss": 1.5008, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 1.4373716632443532, |
|
"grad_norm": 0.9758607149124146, |
|
"learning_rate": 1.913757700205339e-05, |
|
"loss": 1.4734, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 1.5400410677618068, |
|
"grad_norm": 1.0511637926101685, |
|
"learning_rate": 2.050650239561944e-05, |
|
"loss": 1.4343, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 1.6427104722792607, |
|
"grad_norm": 1.1518610715866089, |
|
"learning_rate": 2.187542778918549e-05, |
|
"loss": 1.4289, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 1.7453798767967146, |
|
"grad_norm": 1.1441558599472046, |
|
"learning_rate": 2.324435318275154e-05, |
|
"loss": 1.3902, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 1.8480492813141685, |
|
"grad_norm": 1.3006062507629395, |
|
"learning_rate": 2.4613278576317596e-05, |
|
"loss": 1.3834, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 1.9507186858316223, |
|
"grad_norm": 1.1244959831237793, |
|
"learning_rate": 2.5982203969883643e-05, |
|
"loss": 1.3556, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 1.3616195917129517, |
|
"eval_runtime": 396.64, |
|
"eval_samples_per_second": 4.909, |
|
"eval_steps_per_second": 0.615, |
|
"step": 974 |
|
}, |
|
{ |
|
"epoch": 2.0533880903490758, |
|
"grad_norm": 1.4575847387313843, |
|
"learning_rate": 2.7351129363449694e-05, |
|
"loss": 1.3388, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 2.1560574948665296, |
|
"grad_norm": 1.3578325510025024, |
|
"learning_rate": 2.8720054757015744e-05, |
|
"loss": 1.3067, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 2.2587268993839835, |
|
"grad_norm": 1.9210494756698608, |
|
"learning_rate": 3.0088980150581795e-05, |
|
"loss": 1.3183, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 2.3613963039014374, |
|
"grad_norm": 2.051852226257324, |
|
"learning_rate": 3.1457905544147846e-05, |
|
"loss": 1.274, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 2.4640657084188913, |
|
"grad_norm": 1.660125970840454, |
|
"learning_rate": 3.282683093771389e-05, |
|
"loss": 1.2661, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 2.566735112936345, |
|
"grad_norm": 4.411022663116455, |
|
"learning_rate": 3.419575633127995e-05, |
|
"loss": 1.2339, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 2.669404517453799, |
|
"grad_norm": 1.519783854484558, |
|
"learning_rate": 3.5564681724846e-05, |
|
"loss": 1.2286, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 2.7720739219712525, |
|
"grad_norm": 1.7709407806396484, |
|
"learning_rate": 3.693360711841205e-05, |
|
"loss": 1.219, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 2.8747433264887063, |
|
"grad_norm": 1.8909534215927124, |
|
"learning_rate": 3.8302532511978096e-05, |
|
"loss": 1.2072, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 2.97741273100616, |
|
"grad_norm": 1.4890103340148926, |
|
"learning_rate": 3.967145790554415e-05, |
|
"loss": 1.1951, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_loss": 1.2139294147491455, |
|
"eval_runtime": 396.2906, |
|
"eval_samples_per_second": 4.913, |
|
"eval_steps_per_second": 0.616, |
|
"step": 1461 |
|
}, |
|
{ |
|
"epoch": 3.080082135523614, |
|
"grad_norm": 2.104808807373047, |
|
"learning_rate": 3.998773779555414e-05, |
|
"loss": 1.2078, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 3.1827515400410675, |
|
"grad_norm": 1.3491119146347046, |
|
"learning_rate": 3.993426857404775e-05, |
|
"loss": 1.1717, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 3.2854209445585214, |
|
"grad_norm": 1.7482995986938477, |
|
"learning_rate": 3.983848290367661e-05, |
|
"loss": 1.1757, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 3.3880903490759753, |
|
"grad_norm": 1.8251243829727173, |
|
"learning_rate": 3.9700584118182686e-05, |
|
"loss": 1.1688, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 3.490759753593429, |
|
"grad_norm": 1.9155806303024292, |
|
"learning_rate": 3.9520864948993926e-05, |
|
"loss": 1.1742, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 3.593429158110883, |
|
"grad_norm": 1.8360294103622437, |
|
"learning_rate": 3.929970690381425e-05, |
|
"loss": 1.1474, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 3.696098562628337, |
|
"grad_norm": 1.893100380897522, |
|
"learning_rate": 3.903757945675931e-05, |
|
"loss": 1.1596, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 3.798767967145791, |
|
"grad_norm": 1.7409580945968628, |
|
"learning_rate": 3.8735039051757204e-05, |
|
"loss": 1.1496, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 3.9014373716632442, |
|
"grad_norm": 1.4829061031341553, |
|
"learning_rate": 3.83927279213298e-05, |
|
"loss": 1.1262, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_loss": 1.1556336879730225, |
|
"eval_runtime": 396.6647, |
|
"eval_samples_per_second": 4.908, |
|
"eval_steps_per_second": 0.615, |
|
"step": 1948 |
|
}, |
|
{ |
|
"epoch": 4.0041067761806985, |
|
"grad_norm": 1.6514705419540405, |
|
"learning_rate": 3.801137272326205e-05, |
|
"loss": 1.1419, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 4.1067761806981515, |
|
"grad_norm": 1.7151414155960083, |
|
"learning_rate": 3.759178299805344e-05, |
|
"loss": 1.1441, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 4.209445585215605, |
|
"grad_norm": 1.5632063150405884, |
|
"learning_rate": 3.713484945042608e-05, |
|
"loss": 1.1271, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 4.312114989733059, |
|
"grad_norm": 1.619882583618164, |
|
"learning_rate": 3.664154205853749e-05, |
|
"loss": 1.1016, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 4.414784394250513, |
|
"grad_norm": 1.6811403036117554, |
|
"learning_rate": 3.6112908014911744e-05, |
|
"loss": 1.1203, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 4.517453798767967, |
|
"grad_norm": 1.52841055393219, |
|
"learning_rate": 3.555006950346002e-05, |
|
"loss": 1.1152, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 4.620123203285421, |
|
"grad_norm": 1.482682704925537, |
|
"learning_rate": 3.4954221317309576e-05, |
|
"loss": 1.1063, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 4.722792607802875, |
|
"grad_norm": 1.508597731590271, |
|
"learning_rate": 3.4326628322497834e-05, |
|
"loss": 1.1095, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 4.825462012320329, |
|
"grad_norm": 1.5862213373184204, |
|
"learning_rate": 3.366862277291585e-05, |
|
"loss": 1.1062, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 4.9281314168377826, |
|
"grad_norm": 1.6203546524047852, |
|
"learning_rate": 3.298160148220079e-05, |
|
"loss": 1.1189, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_loss": 1.1254490613937378, |
|
"eval_runtime": 395.8276, |
|
"eval_samples_per_second": 4.919, |
|
"eval_steps_per_second": 0.616, |
|
"step": 2435 |
|
}, |
|
{ |
|
"epoch": 5.030800821355236, |
|
"grad_norm": 1.5772205591201782, |
|
"learning_rate": 3.2267022858581186e-05, |
|
"loss": 1.114, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 5.13347022587269, |
|
"grad_norm": 1.4345210790634155, |
|
"learning_rate": 3.152640380896904e-05, |
|
"loss": 1.0856, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 5.236139630390143, |
|
"grad_norm": 1.4504789113998413, |
|
"learning_rate": 3.0761316518871184e-05, |
|
"loss": 1.0866, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 5.338809034907597, |
|
"grad_norm": 1.841271996498108, |
|
"learning_rate": 2.997338511495513e-05, |
|
"loss": 1.0819, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 5.441478439425051, |
|
"grad_norm": 1.3801902532577515, |
|
"learning_rate": 2.916428221735433e-05, |
|
"loss": 1.1059, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 5.544147843942505, |
|
"grad_norm": 1.5078575611114502, |
|
"learning_rate": 2.8335725389031572e-05, |
|
"loss": 1.0974, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 5.646817248459959, |
|
"grad_norm": 1.6412862539291382, |
|
"learning_rate": 2.7489473489737663e-05, |
|
"loss": 1.1025, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 5.749486652977413, |
|
"grad_norm": 1.6640971899032593, |
|
"learning_rate": 2.6627322942305504e-05, |
|
"loss": 1.0769, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 5.852156057494867, |
|
"grad_norm": 1.8713879585266113, |
|
"learning_rate": 2.57511039192051e-05, |
|
"loss": 1.0739, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 5.95482546201232, |
|
"grad_norm": 1.477726936340332, |
|
"learning_rate": 2.4862676457455003e-05, |
|
"loss": 1.0914, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_loss": 1.1069694757461548, |
|
"eval_runtime": 396.2551, |
|
"eval_samples_per_second": 4.914, |
|
"eval_steps_per_second": 0.616, |
|
"step": 2922 |
|
}, |
|
{ |
|
"epoch": 6.057494866529774, |
|
"grad_norm": 1.5570802688598633, |
|
"learning_rate": 2.396392651013723e-05, |
|
"loss": 1.0646, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 6.160164271047228, |
|
"grad_norm": 1.5152437686920166, |
|
"learning_rate": 2.3056761942897655e-05, |
|
"loss": 1.0817, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 6.262833675564682, |
|
"grad_norm": 1.4843331575393677, |
|
"learning_rate": 2.2143108483930357e-05, |
|
"loss": 1.0762, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 6.365503080082135, |
|
"grad_norm": 1.504683256149292, |
|
"learning_rate": 2.122490563604342e-05, |
|
"loss": 1.0669, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 6.468172484599589, |
|
"grad_norm": 1.5838227272033691, |
|
"learning_rate": 2.0304102559483984e-05, |
|
"loss": 1.0597, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 6.570841889117043, |
|
"grad_norm": 1.61528480052948, |
|
"learning_rate": 1.9382653934262424e-05, |
|
"loss": 1.0839, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 6.673511293634497, |
|
"grad_norm": 1.4133933782577515, |
|
"learning_rate": 1.8462515810759347e-05, |
|
"loss": 1.0798, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 6.776180698151951, |
|
"grad_norm": 1.4809049367904663, |
|
"learning_rate": 1.7545641457423334e-05, |
|
"loss": 1.0666, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 6.878850102669404, |
|
"grad_norm": 1.6875197887420654, |
|
"learning_rate": 1.6633977214374384e-05, |
|
"loss": 1.0671, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 6.981519507186858, |
|
"grad_norm": 1.5088533163070679, |
|
"learning_rate": 1.5729458361714657e-05, |
|
"loss": 1.0756, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_loss": 1.0947245359420776, |
|
"eval_runtime": 396.3527, |
|
"eval_samples_per_second": 4.912, |
|
"eval_steps_per_second": 0.616, |
|
"step": 3409 |
|
}, |
|
{ |
|
"epoch": 7.084188911704312, |
|
"grad_norm": 1.9185519218444824, |
|
"learning_rate": 1.4834005011317545e-05, |
|
"loss": 1.055, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 7.186858316221766, |
|
"grad_norm": 1.6549229621887207, |
|
"learning_rate": 1.3949518030815667e-05, |
|
"loss": 1.0565, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 7.28952772073922, |
|
"grad_norm": 1.5964972972869873, |
|
"learning_rate": 1.3077875008440714e-05, |
|
"loss": 1.0696, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 7.392197125256674, |
|
"grad_norm": 1.5329166650772095, |
|
"learning_rate": 1.2220926267280677e-05, |
|
"loss": 1.0657, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 7.494866529774128, |
|
"grad_norm": 1.6698694229125977, |
|
"learning_rate": 1.1380490937415516e-05, |
|
"loss": 1.0596, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 7.597535934291582, |
|
"grad_norm": 1.5411231517791748, |
|
"learning_rate": 1.0558353094269314e-05, |
|
"loss": 1.059, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 7.700205338809035, |
|
"grad_norm": 1.7882490158081055, |
|
"learning_rate": 9.756257971376489e-06, |
|
"loss": 1.0731, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 7.8028747433264884, |
|
"grad_norm": 1.5504204034805298, |
|
"learning_rate": 8.975908255601452e-06, |
|
"loss": 1.0565, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 7.905544147843942, |
|
"grad_norm": 1.4786473512649536, |
|
"learning_rate": 8.21896047267632e-06, |
|
"loss": 1.0502, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_loss": 1.0888762474060059, |
|
"eval_runtime": 396.8423, |
|
"eval_samples_per_second": 4.906, |
|
"eval_steps_per_second": 0.615, |
|
"step": 3896 |
|
}, |
|
{ |
|
"epoch": 8.008213552361397, |
|
"grad_norm": 1.6214848756790161, |
|
"learning_rate": 7.487021470729523e-06, |
|
"loss": 1.0634, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 8.11088295687885, |
|
"grad_norm": 1.712120771408081, |
|
"learning_rate": 6.781645009269846e-06, |
|
"loss": 1.062, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 8.213552361396303, |
|
"grad_norm": 1.5750735998153687, |
|
"learning_rate": 6.1043284608671e-06, |
|
"loss": 1.0387, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 8.316221765913758, |
|
"grad_norm": 1.5793296098709106, |
|
"learning_rate": 5.45650963253082e-06, |
|
"loss": 1.0594, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 8.41889117043121, |
|
"grad_norm": 1.445952296257019, |
|
"learning_rate": 4.839563713534743e-06, |
|
"loss": 1.0614, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 8.521560574948666, |
|
"grad_norm": 1.731099009513855, |
|
"learning_rate": 4.254800356166155e-06, |
|
"loss": 1.0599, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 8.624229979466119, |
|
"grad_norm": 1.7094162702560425, |
|
"learning_rate": 3.7034608955971175e-06, |
|
"loss": 1.0437, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 8.726899383983573, |
|
"grad_norm": 1.6651116609573364, |
|
"learning_rate": 3.1867157147791585e-06, |
|
"loss": 1.0667, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 8.829568788501026, |
|
"grad_norm": 1.5388877391815186, |
|
"learning_rate": 2.705661759955318e-06, |
|
"loss": 1.0529, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 8.932238193018481, |
|
"grad_norm": 1.7173577547073364, |
|
"learning_rate": 2.261320212063589e-06, |
|
"loss": 1.0619, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_loss": 1.0863490104675293, |
|
"eval_runtime": 395.7285, |
|
"eval_samples_per_second": 4.92, |
|
"eval_steps_per_second": 0.617, |
|
"step": 4383 |
|
}, |
|
{ |
|
"epoch": 9.034907597535934, |
|
"grad_norm": 1.79407799243927, |
|
"learning_rate": 1.854634318974835e-06, |
|
"loss": 1.0494, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 9.137577002053389, |
|
"grad_norm": 1.6543513536453247, |
|
"learning_rate": 1.486467393167017e-06, |
|
"loss": 1.0497, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 9.240246406570842, |
|
"grad_norm": 1.5759410858154297, |
|
"learning_rate": 1.1576009790861598e-06, |
|
"loss": 1.0503, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 9.342915811088295, |
|
"grad_norm": 1.4627271890640259, |
|
"learning_rate": 8.687331940844545e-07, |
|
"loss": 1.0585, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 9.44558521560575, |
|
"grad_norm": 1.7757667303085327, |
|
"learning_rate": 6.204772464573161e-07, |
|
"loss": 1.0468, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 9.548254620123203, |
|
"grad_norm": 1.7230888605117798, |
|
"learning_rate": 4.1336013372533125e-07, |
|
"loss": 1.0485, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 9.650924024640657, |
|
"grad_norm": 1.7910829782485962, |
|
"learning_rate": 2.478215239243409e-07, |
|
"loss": 1.0498, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 9.75359342915811, |
|
"grad_norm": 1.7358592748641968, |
|
"learning_rate": 1.2421282227850662e-07, |
|
"loss": 1.0548, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 9.856262833675565, |
|
"grad_norm": 1.603044867515564, |
|
"learning_rate": 4.2796425237592306e-08, |
|
"loss": 1.0589, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 9.958932238193018, |
|
"grad_norm": 1.4900743961334229, |
|
"learning_rate": 3.745163461994494e-09, |
|
"loss": 1.056, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_loss": 1.0858850479125977, |
|
"eval_runtime": 396.092, |
|
"eval_samples_per_second": 4.916, |
|
"eval_steps_per_second": 0.616, |
|
"step": 4870 |
|
} |
|
], |
|
"logging_steps": 50, |
|
"max_steps": 4870, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 10, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 6.27791813148672e+17, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|