|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.0, |
|
"eval_steps": 500, |
|
"global_step": 2500, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.008, |
|
"grad_norm": 0.2994475388615314, |
|
"learning_rate": 1.6000000000000001e-06, |
|
"loss": 0.2223, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.016, |
|
"grad_norm": 0.273631506204062, |
|
"learning_rate": 3.2000000000000003e-06, |
|
"loss": 0.1771, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.024, |
|
"grad_norm": 0.2130209890157738, |
|
"learning_rate": 4.800000000000001e-06, |
|
"loss": 0.1598, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.032, |
|
"grad_norm": 0.18240027035932016, |
|
"learning_rate": 6.4000000000000006e-06, |
|
"loss": 0.1619, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 0.20027988136436473, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 0.1724, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.048, |
|
"grad_norm": 0.19349670254659063, |
|
"learning_rate": 9.600000000000001e-06, |
|
"loss": 0.1781, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.056, |
|
"grad_norm": 0.1522942785320451, |
|
"learning_rate": 1.1200000000000001e-05, |
|
"loss": 0.1741, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.064, |
|
"grad_norm": 0.20077878887262207, |
|
"learning_rate": 1.2800000000000001e-05, |
|
"loss": 0.1678, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.072, |
|
"grad_norm": 0.2006526811340805, |
|
"learning_rate": 1.4400000000000001e-05, |
|
"loss": 0.1754, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.18403850144652611, |
|
"learning_rate": 1.6000000000000003e-05, |
|
"loss": 0.1793, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.088, |
|
"grad_norm": 0.17392033537978838, |
|
"learning_rate": 1.76e-05, |
|
"loss": 0.1669, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.096, |
|
"grad_norm": 0.15129051337201097, |
|
"learning_rate": 1.9200000000000003e-05, |
|
"loss": 0.177, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.104, |
|
"grad_norm": 0.1686995282450159, |
|
"learning_rate": 1.9999781283802247e-05, |
|
"loss": 0.1838, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.112, |
|
"grad_norm": 0.18454731582348338, |
|
"learning_rate": 1.999803161162393e-05, |
|
"loss": 0.2058, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.1594563484769184, |
|
"learning_rate": 1.999453257340926e-05, |
|
"loss": 0.1953, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.128, |
|
"grad_norm": 0.16207625262493944, |
|
"learning_rate": 1.9989284781388617e-05, |
|
"loss": 0.1871, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.136, |
|
"grad_norm": 0.15859384233929794, |
|
"learning_rate": 1.9982289153773648e-05, |
|
"loss": 0.1948, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.144, |
|
"grad_norm": 0.17122386425268052, |
|
"learning_rate": 1.9973546914596622e-05, |
|
"loss": 0.1954, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.152, |
|
"grad_norm": 0.16538234415049946, |
|
"learning_rate": 1.996305959349627e-05, |
|
"loss": 0.1898, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.19838611436297024, |
|
"learning_rate": 1.9950829025450116e-05, |
|
"loss": 0.1903, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.168, |
|
"grad_norm": 0.18275997172338, |
|
"learning_rate": 1.993685735045343e-05, |
|
"loss": 0.189, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.176, |
|
"grad_norm": 0.19605227359376942, |
|
"learning_rate": 1.9921147013144782e-05, |
|
"loss": 0.202, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.184, |
|
"grad_norm": 0.1510786730603492, |
|
"learning_rate": 1.9903700762378303e-05, |
|
"loss": 0.1906, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.192, |
|
"grad_norm": 0.1731755052103561, |
|
"learning_rate": 1.9884521650742718e-05, |
|
"loss": 0.1818, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.1703147411704783, |
|
"learning_rate": 1.9863613034027224e-05, |
|
"loss": 0.1828, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.208, |
|
"grad_norm": 0.16465949216497772, |
|
"learning_rate": 1.9840978570634338e-05, |
|
"loss": 0.1998, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.216, |
|
"grad_norm": 0.13341671497634322, |
|
"learning_rate": 1.9816622220939762e-05, |
|
"loss": 0.1864, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.224, |
|
"grad_norm": 0.162293542213837, |
|
"learning_rate": 1.9790548246599447e-05, |
|
"loss": 0.1892, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.232, |
|
"grad_norm": 0.15597629673666613, |
|
"learning_rate": 1.976276120980393e-05, |
|
"loss": 0.1884, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.14793980628985084, |
|
"learning_rate": 1.973326597248006e-05, |
|
"loss": 0.1978, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.248, |
|
"grad_norm": 0.1447727611743318, |
|
"learning_rate": 1.9702067695440333e-05, |
|
"loss": 0.1883, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.256, |
|
"grad_norm": 0.16933785764552192, |
|
"learning_rate": 1.966917183747987e-05, |
|
"loss": 0.1898, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.264, |
|
"grad_norm": 0.1589759149940254, |
|
"learning_rate": 1.9634584154421316e-05, |
|
"loss": 0.1949, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.272, |
|
"grad_norm": 0.12763481414215103, |
|
"learning_rate": 1.95983106981077e-05, |
|
"loss": 0.1855, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.13722214819739667, |
|
"learning_rate": 1.9560357815343577e-05, |
|
"loss": 0.1887, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.288, |
|
"grad_norm": 0.17592856768775877, |
|
"learning_rate": 1.9520732146784493e-05, |
|
"loss": 0.1875, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.296, |
|
"grad_norm": 0.1540007177537391, |
|
"learning_rate": 1.947944062577507e-05, |
|
"loss": 0.1995, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.304, |
|
"grad_norm": 0.14462148850861567, |
|
"learning_rate": 1.9436490477135877e-05, |
|
"loss": 0.1905, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.312, |
|
"grad_norm": 0.14275084054493534, |
|
"learning_rate": 1.93918892158993e-05, |
|
"loss": 0.1946, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.13709865630962415, |
|
"learning_rate": 1.934564464599461e-05, |
|
"loss": 0.1848, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.328, |
|
"grad_norm": 0.12851355006510504, |
|
"learning_rate": 1.9297764858882516e-05, |
|
"loss": 0.2012, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.336, |
|
"grad_norm": 0.14422766911430207, |
|
"learning_rate": 1.924825823213939e-05, |
|
"loss": 0.1796, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.344, |
|
"grad_norm": 0.1456420238432456, |
|
"learning_rate": 1.9197133427991437e-05, |
|
"loss": 0.1894, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.352, |
|
"grad_norm": 0.11424505824104754, |
|
"learning_rate": 1.9144399391799043e-05, |
|
"loss": 0.1855, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.1505170978234406, |
|
"learning_rate": 1.909006535049163e-05, |
|
"loss": 0.1826, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.368, |
|
"grad_norm": 0.1325857585966927, |
|
"learning_rate": 1.903414081095315e-05, |
|
"loss": 0.1969, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.376, |
|
"grad_norm": 0.12049306945186798, |
|
"learning_rate": 1.897663555835872e-05, |
|
"loss": 0.1852, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.384, |
|
"grad_norm": 0.1321067212261413, |
|
"learning_rate": 1.8917559654462474e-05, |
|
"loss": 0.1902, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.392, |
|
"grad_norm": 0.12544038658282392, |
|
"learning_rate": 1.8856923435837024e-05, |
|
"loss": 0.1834, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.13754099065635406, |
|
"learning_rate": 1.879473751206489e-05, |
|
"loss": 0.1777, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.408, |
|
"grad_norm": 0.14299681112215643, |
|
"learning_rate": 1.8731012763882132e-05, |
|
"loss": 0.1891, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.416, |
|
"grad_norm": 0.12978945402957556, |
|
"learning_rate": 1.8665760341274505e-05, |
|
"loss": 0.1838, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.424, |
|
"grad_norm": 0.12828927955187144, |
|
"learning_rate": 1.859899166152657e-05, |
|
"loss": 0.1775, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.432, |
|
"grad_norm": 0.1399995330504017, |
|
"learning_rate": 1.8530718407223976e-05, |
|
"loss": 0.1949, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.1201971674290692, |
|
"learning_rate": 1.8460952524209355e-05, |
|
"loss": 0.1997, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.448, |
|
"grad_norm": 0.14220963488029062, |
|
"learning_rate": 1.8389706219492147e-05, |
|
"loss": 0.1864, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.456, |
|
"grad_norm": 0.12393318013758685, |
|
"learning_rate": 1.831699195911272e-05, |
|
"loss": 0.1932, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.464, |
|
"grad_norm": 0.13175301413953197, |
|
"learning_rate": 1.8242822465961177e-05, |
|
"loss": 0.1868, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.472, |
|
"grad_norm": 0.1439474730329765, |
|
"learning_rate": 1.8167210717551224e-05, |
|
"loss": 0.1825, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.1189834140430123, |
|
"learning_rate": 1.8090169943749477e-05, |
|
"loss": 0.1793, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.488, |
|
"grad_norm": 0.13694513887911383, |
|
"learning_rate": 1.8011713624460608e-05, |
|
"loss": 0.1838, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.496, |
|
"grad_norm": 0.1342792821222079, |
|
"learning_rate": 1.793185548726878e-05, |
|
"loss": 0.1981, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.504, |
|
"grad_norm": 0.12267440816211221, |
|
"learning_rate": 1.785060950503568e-05, |
|
"loss": 0.1945, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.512, |
|
"grad_norm": 0.14191462759384785, |
|
"learning_rate": 1.7767989893455696e-05, |
|
"loss": 0.1939, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 0.15143689437873276, |
|
"learning_rate": 1.7684011108568593e-05, |
|
"loss": 0.1832, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.528, |
|
"grad_norm": 0.10997527630360879, |
|
"learning_rate": 1.759868784423009e-05, |
|
"loss": 0.1765, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.536, |
|
"grad_norm": 0.1262528956359292, |
|
"learning_rate": 1.7512035029540887e-05, |
|
"loss": 0.1791, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.544, |
|
"grad_norm": 0.15602459318561945, |
|
"learning_rate": 1.74240678262345e-05, |
|
"loss": 0.1743, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.552, |
|
"grad_norm": 0.13025666149189777, |
|
"learning_rate": 1.73348016260244e-05, |
|
"loss": 0.1834, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.1387483634701272, |
|
"learning_rate": 1.7244252047910893e-05, |
|
"loss": 0.1844, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.568, |
|
"grad_norm": 0.14296180482072157, |
|
"learning_rate": 1.7152434935448257e-05, |
|
"loss": 0.1728, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.576, |
|
"grad_norm": 0.13975560544538343, |
|
"learning_rate": 1.705936635397259e-05, |
|
"loss": 0.1861, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.584, |
|
"grad_norm": 0.11624329808561513, |
|
"learning_rate": 1.6965062587790823e-05, |
|
"loss": 0.1851, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.592, |
|
"grad_norm": 0.12204929575785449, |
|
"learning_rate": 1.6869540137331445e-05, |
|
"loss": 0.1851, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.11843883632568544, |
|
"learning_rate": 1.6772815716257414e-05, |
|
"loss": 0.1926, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.608, |
|
"grad_norm": 0.13379905247254648, |
|
"learning_rate": 1.667490624854173e-05, |
|
"loss": 0.1825, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.616, |
|
"grad_norm": 0.14954402393242697, |
|
"learning_rate": 1.6575828865506246e-05, |
|
"loss": 0.184, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.624, |
|
"grad_norm": 0.11855929270795366, |
|
"learning_rate": 1.647560090282419e-05, |
|
"loss": 0.1741, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.632, |
|
"grad_norm": 0.15098593909944216, |
|
"learning_rate": 1.63742398974869e-05, |
|
"loss": 0.18, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.12622455885460282, |
|
"learning_rate": 1.6271763584735373e-05, |
|
"loss": 0.1795, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.648, |
|
"grad_norm": 0.1214479548976728, |
|
"learning_rate": 1.616818989495711e-05, |
|
"loss": 0.1754, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.656, |
|
"grad_norm": 0.11633961346066299, |
|
"learning_rate": 1.6063536950548825e-05, |
|
"loss": 0.1791, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.664, |
|
"grad_norm": 0.1349935261273476, |
|
"learning_rate": 1.595782306274553e-05, |
|
"loss": 0.1901, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.672, |
|
"grad_norm": 0.12313061321761182, |
|
"learning_rate": 1.5851066728416617e-05, |
|
"loss": 0.1813, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.16151825468675773, |
|
"learning_rate": 1.5743286626829437e-05, |
|
"loss": 0.1754, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.688, |
|
"grad_norm": 0.14085102248573694, |
|
"learning_rate": 1.5634501616380967e-05, |
|
"loss": 0.1885, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.696, |
|
"grad_norm": 0.14051153835159377, |
|
"learning_rate": 1.5524730731298136e-05, |
|
"loss": 0.1883, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.704, |
|
"grad_norm": 0.10696859105735423, |
|
"learning_rate": 1.541399317830738e-05, |
|
"loss": 0.1785, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.712, |
|
"grad_norm": 0.12489614347746164, |
|
"learning_rate": 1.530230833327405e-05, |
|
"loss": 0.1868, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 0.13020128972045483, |
|
"learning_rate": 1.5189695737812153e-05, |
|
"loss": 0.1803, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.728, |
|
"grad_norm": 0.12378887718286205, |
|
"learning_rate": 1.5076175095865171e-05, |
|
"loss": 0.1885, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.736, |
|
"grad_norm": 0.12669683188981912, |
|
"learning_rate": 1.4961766270258422e-05, |
|
"loss": 0.1735, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.744, |
|
"grad_norm": 0.12139646601130864, |
|
"learning_rate": 1.4846489279223653e-05, |
|
"loss": 0.1842, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.752, |
|
"grad_norm": 0.12564036718714786, |
|
"learning_rate": 1.473036429289641e-05, |
|
"loss": 0.1862, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 0.12879466675814272, |
|
"learning_rate": 1.461341162978688e-05, |
|
"loss": 0.1794, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.768, |
|
"grad_norm": 0.13118866753606923, |
|
"learning_rate": 1.4495651753224706e-05, |
|
"loss": 0.1752, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.776, |
|
"grad_norm": 0.1397568247577318, |
|
"learning_rate": 1.437710526777852e-05, |
|
"loss": 0.1837, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.784, |
|
"grad_norm": 0.1199748082038019, |
|
"learning_rate": 1.4257792915650728e-05, |
|
"loss": 0.1751, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.792, |
|
"grad_norm": 0.1275568062184644, |
|
"learning_rate": 1.4137735573048232e-05, |
|
"loss": 0.1681, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.11731254237226522, |
|
"learning_rate": 1.4016954246529697e-05, |
|
"loss": 0.1672, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.808, |
|
"grad_norm": 0.12803490644945145, |
|
"learning_rate": 1.3895470069330003e-05, |
|
"loss": 0.1775, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.816, |
|
"grad_norm": 0.13148610169928593, |
|
"learning_rate": 1.377330429766256e-05, |
|
"loss": 0.179, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.824, |
|
"grad_norm": 0.12498379605499213, |
|
"learning_rate": 1.3650478307000059e-05, |
|
"loss": 0.1743, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.832, |
|
"grad_norm": 0.10768287920272084, |
|
"learning_rate": 1.3527013588334415e-05, |
|
"loss": 0.1798, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.11876441348756149, |
|
"learning_rate": 1.3402931744416432e-05, |
|
"loss": 0.1669, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.848, |
|
"grad_norm": 0.11670565624337659, |
|
"learning_rate": 1.3278254485975977e-05, |
|
"loss": 0.176, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.856, |
|
"grad_norm": 0.12614319023396608, |
|
"learning_rate": 1.3153003627923217e-05, |
|
"loss": 0.1717, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.864, |
|
"grad_norm": 0.15423610893363637, |
|
"learning_rate": 1.3027201085531633e-05, |
|
"loss": 0.1787, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.872, |
|
"grad_norm": 0.1437966604387281, |
|
"learning_rate": 1.2900868870603502e-05, |
|
"loss": 0.1713, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.13194227493290273, |
|
"learning_rate": 1.2774029087618448e-05, |
|
"loss": 0.1859, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.888, |
|
"grad_norm": 0.11652085362332747, |
|
"learning_rate": 1.2646703929865817e-05, |
|
"loss": 0.1798, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.896, |
|
"grad_norm": 0.11917839406476669, |
|
"learning_rate": 1.2518915675561482e-05, |
|
"loss": 0.1868, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.904, |
|
"grad_norm": 0.11813364230838641, |
|
"learning_rate": 1.2390686683949799e-05, |
|
"loss": 0.1575, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.912, |
|
"grad_norm": 0.1196110875735261, |
|
"learning_rate": 1.2262039391391405e-05, |
|
"loss": 0.17, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.12300092078788795, |
|
"learning_rate": 1.213299630743747e-05, |
|
"loss": 0.1682, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.928, |
|
"grad_norm": 0.1208425846118018, |
|
"learning_rate": 1.2003580010891214e-05, |
|
"loss": 0.1703, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.936, |
|
"grad_norm": 0.12083725425562747, |
|
"learning_rate": 1.187381314585725e-05, |
|
"loss": 0.1797, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.944, |
|
"grad_norm": 0.1178408143433832, |
|
"learning_rate": 1.1743718417779518e-05, |
|
"loss": 0.1792, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.952, |
|
"grad_norm": 0.11740271053188309, |
|
"learning_rate": 1.1613318589468512e-05, |
|
"loss": 0.1726, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.1090840603269972, |
|
"learning_rate": 1.148263647711842e-05, |
|
"loss": 0.1756, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.968, |
|
"grad_norm": 0.124940945129332, |
|
"learning_rate": 1.135169494631497e-05, |
|
"loss": 0.1837, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 0.976, |
|
"grad_norm": 0.1257424082327308, |
|
"learning_rate": 1.1220516908034602e-05, |
|
"loss": 0.1785, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.984, |
|
"grad_norm": 0.11430280573700291, |
|
"learning_rate": 1.1089125314635727e-05, |
|
"loss": 0.1731, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.992, |
|
"grad_norm": 0.13276298761961705, |
|
"learning_rate": 1.0957543155842703e-05, |
|
"loss": 0.1718, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.1161581920629144, |
|
"learning_rate": 1.0825793454723325e-05, |
|
"loss": 0.1696, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 1.008, |
|
"grad_norm": 0.09621449723576145, |
|
"learning_rate": 1.0693899263660442e-05, |
|
"loss": 0.1113, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 1.016, |
|
"grad_norm": 0.10015368517162866, |
|
"learning_rate": 1.0561883660318456e-05, |
|
"loss": 0.1003, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 1.024, |
|
"grad_norm": 0.1076609073322909, |
|
"learning_rate": 1.0429769743605406e-05, |
|
"loss": 0.1092, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 1.032, |
|
"grad_norm": 0.1209467723696444, |
|
"learning_rate": 1.0297580629631324e-05, |
|
"loss": 0.1047, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"grad_norm": 0.11347187764817329, |
|
"learning_rate": 1.0165339447663586e-05, |
|
"loss": 0.1012, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 1.048, |
|
"grad_norm": 0.13787078556126084, |
|
"learning_rate": 1.0033069336079952e-05, |
|
"loss": 0.0996, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 1.056, |
|
"grad_norm": 0.11234445761651994, |
|
"learning_rate": 9.900793438320037e-06, |
|
"loss": 0.0986, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 1.064, |
|
"grad_norm": 0.11599014325689881, |
|
"learning_rate": 9.768534898835864e-06, |
|
"loss": 0.1063, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 1.072, |
|
"grad_norm": 0.11347494815929351, |
|
"learning_rate": 9.636316859042258e-06, |
|
"loss": 0.1039, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"grad_norm": 0.12183008770849793, |
|
"learning_rate": 9.504162453267776e-06, |
|
"loss": 0.1032, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 1.088, |
|
"grad_norm": 0.11562604188712092, |
|
"learning_rate": 9.372094804706867e-06, |
|
"loss": 0.1011, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 1.096, |
|
"grad_norm": 0.10067772672412269, |
|
"learning_rate": 9.24013702137397e-06, |
|
"loss": 0.0972, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 1.104, |
|
"grad_norm": 0.11541149806656344, |
|
"learning_rate": 9.108312192060298e-06, |
|
"loss": 0.0958, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 1.112, |
|
"grad_norm": 0.129106816260335, |
|
"learning_rate": 8.97664338229395e-06, |
|
"loss": 0.1, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"grad_norm": 0.11074093024717463, |
|
"learning_rate": 8.84515363030414e-06, |
|
"loss": 0.0976, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 1.1280000000000001, |
|
"grad_norm": 0.10927701173789658, |
|
"learning_rate": 8.713865942990143e-06, |
|
"loss": 0.1004, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 1.1360000000000001, |
|
"grad_norm": 0.129800724137463, |
|
"learning_rate": 8.582803291895758e-06, |
|
"loss": 0.1097, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 1.144, |
|
"grad_norm": 0.1109925526353623, |
|
"learning_rate": 8.451988609189987e-06, |
|
"loss": 0.1047, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 1.152, |
|
"grad_norm": 0.12402578949666784, |
|
"learning_rate": 8.321444783654524e-06, |
|
"loss": 0.1079, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 0.1112595703457445, |
|
"learning_rate": 8.191194656678905e-06, |
|
"loss": 0.0976, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 1.168, |
|
"grad_norm": 0.12036457116059705, |
|
"learning_rate": 8.06126101826392e-06, |
|
"loss": 0.0973, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 1.176, |
|
"grad_norm": 0.12090306774036914, |
|
"learning_rate": 7.931666603034034e-06, |
|
"loss": 0.1008, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 1.184, |
|
"grad_norm": 0.12317909704896644, |
|
"learning_rate": 7.80243408625947e-06, |
|
"loss": 0.0948, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 1.192, |
|
"grad_norm": 0.10705938042955583, |
|
"learning_rate": 7.673586079888699e-06, |
|
"loss": 0.1067, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 0.14259766047955233, |
|
"learning_rate": 7.545145128592009e-06, |
|
"loss": 0.1007, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.208, |
|
"grad_norm": 0.11378740748534237, |
|
"learning_rate": 7.4171337058168365e-06, |
|
"loss": 0.103, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 1.216, |
|
"grad_norm": 0.11908875076001751, |
|
"learning_rate": 7.28957420985556e-06, |
|
"loss": 0.099, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 1.224, |
|
"grad_norm": 0.1177996071960852, |
|
"learning_rate": 7.16248895992645e-06, |
|
"loss": 0.0974, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 1.232, |
|
"grad_norm": 0.12558719327814405, |
|
"learning_rate": 7.035900192268464e-06, |
|
"loss": 0.1114, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"grad_norm": 0.12231602378332272, |
|
"learning_rate": 6.909830056250527e-06, |
|
"loss": 0.1017, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 1.248, |
|
"grad_norm": 0.1411428231676542, |
|
"learning_rate": 6.784300610496049e-06, |
|
"loss": 0.1037, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 1.256, |
|
"grad_norm": 0.11433442906651817, |
|
"learning_rate": 6.659333819023291e-06, |
|
"loss": 0.1017, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 1.264, |
|
"grad_norm": 0.1587733997288609, |
|
"learning_rate": 6.534951547402322e-06, |
|
"loss": 0.1022, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 1.272, |
|
"grad_norm": 0.11284632840723509, |
|
"learning_rate": 6.411175558929152e-06, |
|
"loss": 0.1016, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 0.10824650118751981, |
|
"learning_rate": 6.2880275108177915e-06, |
|
"loss": 0.1027, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 1.288, |
|
"grad_norm": 0.13145070635475115, |
|
"learning_rate": 6.165528950410884e-06, |
|
"loss": 0.1066, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 1.296, |
|
"grad_norm": 0.13085006869996493, |
|
"learning_rate": 6.04370131140952e-06, |
|
"loss": 0.0992, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 1.304, |
|
"grad_norm": 0.12602863363341382, |
|
"learning_rate": 5.922565910122967e-06, |
|
"loss": 0.0953, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 1.312, |
|
"grad_norm": 0.1225471587130614, |
|
"learning_rate": 5.802143941738945e-06, |
|
"loss": 0.106, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"grad_norm": 0.11766655619511902, |
|
"learning_rate": 5.6824564766150724e-06, |
|
"loss": 0.1063, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 1.328, |
|
"grad_norm": 0.12243121066150191, |
|
"learning_rate": 5.563524456592163e-06, |
|
"loss": 0.1005, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 1.336, |
|
"grad_norm": 0.1137092367870251, |
|
"learning_rate": 5.445368691330008e-06, |
|
"loss": 0.1014, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 1.3439999999999999, |
|
"grad_norm": 0.11563617059718279, |
|
"learning_rate": 5.328009854666303e-06, |
|
"loss": 0.0963, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 1.3519999999999999, |
|
"grad_norm": 0.10812153213525164, |
|
"learning_rate": 5.211468480999304e-06, |
|
"loss": 0.1039, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 1.3599999999999999, |
|
"grad_norm": 0.10528899506093697, |
|
"learning_rate": 5.095764961694923e-06, |
|
"loss": 0.1003, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 1.3679999999999999, |
|
"grad_norm": 0.11815436297994328, |
|
"learning_rate": 4.980919541518796e-06, |
|
"loss": 0.0982, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 1.376, |
|
"grad_norm": 0.12086679854967129, |
|
"learning_rate": 4.866952315094088e-06, |
|
"loss": 0.0971, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 1.384, |
|
"grad_norm": 0.1216775616601072, |
|
"learning_rate": 4.753883223385467e-06, |
|
"loss": 0.0945, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 1.392, |
|
"grad_norm": 0.10648914490397961, |
|
"learning_rate": 4.641732050210032e-06, |
|
"loss": 0.1039, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 0.11378893301113804, |
|
"learning_rate": 4.530518418775734e-06, |
|
"loss": 0.0935, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 1.408, |
|
"grad_norm": 0.13321885782018256, |
|
"learning_rate": 4.420261788247841e-06, |
|
"loss": 0.1007, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 1.416, |
|
"grad_norm": 0.1166412836830051, |
|
"learning_rate": 4.3109814503441894e-06, |
|
"loss": 0.1, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 1.424, |
|
"grad_norm": 0.12573437053218914, |
|
"learning_rate": 4.202696525959667e-06, |
|
"loss": 0.1018, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 1.432, |
|
"grad_norm": 0.10766865319844147, |
|
"learning_rate": 4.0954259618206295e-06, |
|
"loss": 0.1055, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"grad_norm": 0.11916472873837182, |
|
"learning_rate": 3.989188527169749e-06, |
|
"loss": 0.1006, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 1.448, |
|
"grad_norm": 0.12141420740790235, |
|
"learning_rate": 3.884002810481959e-06, |
|
"loss": 0.1053, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 1.456, |
|
"grad_norm": 0.10131664100763009, |
|
"learning_rate": 3.7798872162119948e-06, |
|
"loss": 0.095, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 1.464, |
|
"grad_norm": 0.10718874752624193, |
|
"learning_rate": 3.676859961574162e-06, |
|
"loss": 0.0966, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 1.472, |
|
"grad_norm": 0.11477403835967784, |
|
"learning_rate": 3.5749390733548382e-06, |
|
"loss": 0.0973, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"grad_norm": 0.10635089908850691, |
|
"learning_rate": 3.4741423847583134e-06, |
|
"loss": 0.0976, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 1.488, |
|
"grad_norm": 0.10600348235646671, |
|
"learning_rate": 3.3744875322865035e-06, |
|
"loss": 0.0974, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 1.496, |
|
"grad_norm": 0.1097923236145172, |
|
"learning_rate": 3.2759919526530536e-06, |
|
"loss": 0.1018, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 1.504, |
|
"grad_norm": 0.12057262052779812, |
|
"learning_rate": 3.178672879732435e-06, |
|
"loss": 0.0921, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 1.512, |
|
"grad_norm": 0.10322870417117935, |
|
"learning_rate": 3.0825473415445073e-06, |
|
"loss": 0.0969, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"grad_norm": 0.11162000352158546, |
|
"learning_rate": 2.9876321572751143e-06, |
|
"loss": 0.1004, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 1.528, |
|
"grad_norm": 0.11080139256806071, |
|
"learning_rate": 2.8939439343332086e-06, |
|
"loss": 0.0925, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 1.536, |
|
"grad_norm": 0.11129458730658273, |
|
"learning_rate": 2.8014990654450325e-06, |
|
"loss": 0.0965, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 1.544, |
|
"grad_norm": 0.11117549882577657, |
|
"learning_rate": 2.7103137257858867e-06, |
|
"loss": 0.0913, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 1.552, |
|
"grad_norm": 0.13048983161003938, |
|
"learning_rate": 2.6204038701499056e-06, |
|
"loss": 0.0992, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"grad_norm": 0.11734607541924097, |
|
"learning_rate": 2.5317852301584642e-06, |
|
"loss": 0.1012, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 1.568, |
|
"grad_norm": 0.1083555940829194, |
|
"learning_rate": 2.4444733115075823e-06, |
|
"loss": 0.0971, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 1.576, |
|
"grad_norm": 0.10908207209378103, |
|
"learning_rate": 2.3584833912548887e-06, |
|
"loss": 0.0979, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 1.584, |
|
"grad_norm": 0.11841797427563601, |
|
"learning_rate": 2.2738305151465646e-06, |
|
"loss": 0.0942, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 1.592, |
|
"grad_norm": 0.11315769991653934, |
|
"learning_rate": 2.190529494984782e-06, |
|
"loss": 0.0888, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"grad_norm": 0.11433802488811017, |
|
"learning_rate": 2.1085949060360654e-06, |
|
"loss": 0.0875, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.608, |
|
"grad_norm": 0.12338526437657736, |
|
"learning_rate": 2.0280410844810426e-06, |
|
"loss": 0.1007, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 1.616, |
|
"grad_norm": 0.1140721727012241, |
|
"learning_rate": 1.9488821249060297e-06, |
|
"loss": 0.0961, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 1.624, |
|
"grad_norm": 0.10596121333045593, |
|
"learning_rate": 1.8711318778368792e-06, |
|
"loss": 0.0955, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 1.6320000000000001, |
|
"grad_norm": 0.1216702589106228, |
|
"learning_rate": 1.7948039473155553e-06, |
|
"loss": 0.0992, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 1.6400000000000001, |
|
"grad_norm": 0.12842132114696375, |
|
"learning_rate": 1.7199116885197996e-06, |
|
"loss": 0.0961, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 1.6480000000000001, |
|
"grad_norm": 0.12054578218454282, |
|
"learning_rate": 1.646468205426377e-06, |
|
"loss": 0.0915, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 1.6560000000000001, |
|
"grad_norm": 0.13571464476020081, |
|
"learning_rate": 1.5744863485182537e-06, |
|
"loss": 0.1007, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 1.6640000000000001, |
|
"grad_norm": 0.09706739858443193, |
|
"learning_rate": 1.5039787125361327e-06, |
|
"loss": 0.094, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 1.6720000000000002, |
|
"grad_norm": 0.10675797230018248, |
|
"learning_rate": 1.4349576342747462e-06, |
|
"loss": 0.0938, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 1.6800000000000002, |
|
"grad_norm": 0.11223016492676532, |
|
"learning_rate": 1.367435190424261e-06, |
|
"loss": 0.0938, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 1.688, |
|
"grad_norm": 0.12732260810929744, |
|
"learning_rate": 1.3014231954572287e-06, |
|
"loss": 0.098, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 1.696, |
|
"grad_norm": 0.12300876323192528, |
|
"learning_rate": 1.2369331995613664e-06, |
|
"loss": 0.0964, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 1.704, |
|
"grad_norm": 0.12534390954335137, |
|
"learning_rate": 1.1739764866186309e-06, |
|
"loss": 0.0896, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 1.712, |
|
"grad_norm": 0.11609921777362982, |
|
"learning_rate": 1.112564072230863e-06, |
|
"loss": 0.0959, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"grad_norm": 0.10391980074321304, |
|
"learning_rate": 1.0527067017923654e-06, |
|
"loss": 0.0961, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 1.728, |
|
"grad_norm": 0.1278354882865009, |
|
"learning_rate": 9.944148486097793e-07, |
|
"loss": 0.0976, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 1.736, |
|
"grad_norm": 0.10681775464138107, |
|
"learning_rate": 9.376987120695547e-07, |
|
"loss": 0.0967, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 1.744, |
|
"grad_norm": 0.12461590835029315, |
|
"learning_rate": 8.825682158533555e-07, |
|
"loss": 0.1019, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 1.752, |
|
"grad_norm": 0.12925806225583442, |
|
"learning_rate": 8.290330062017015e-07, |
|
"loss": 0.0991, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"grad_norm": 0.12035123190303527, |
|
"learning_rate": 7.771024502261526e-07, |
|
"loss": 0.095, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 1.768, |
|
"grad_norm": 0.11830596930835367, |
|
"learning_rate": 7.267856342703461e-07, |
|
"loss": 0.1003, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 1.776, |
|
"grad_norm": 0.10309290288715313, |
|
"learning_rate": 6.780913623201346e-07, |
|
"loss": 0.0934, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 1.784, |
|
"grad_norm": 0.11823660601228905, |
|
"learning_rate": 6.310281544631547e-07, |
|
"loss": 0.091, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 1.792, |
|
"grad_norm": 0.1240524461220155, |
|
"learning_rate": 5.856042453980526e-07, |
|
"loss": 0.0933, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"grad_norm": 0.10954545218644868, |
|
"learning_rate": 5.418275829936537e-07, |
|
"loss": 0.095, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 1.808, |
|
"grad_norm": 0.09984478082295495, |
|
"learning_rate": 4.997058268983135e-07, |
|
"loss": 0.0884, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 1.8159999999999998, |
|
"grad_norm": 0.09980866667906611, |
|
"learning_rate": 4.592463471997022e-07, |
|
"loss": 0.091, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 1.8239999999999998, |
|
"grad_norm": 0.11065649275917133, |
|
"learning_rate": 4.204562231352516e-07, |
|
"loss": 0.0996, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 1.8319999999999999, |
|
"grad_norm": 0.10945247265836085, |
|
"learning_rate": 3.83342241853496e-07, |
|
"loss": 0.093, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 1.8399999999999999, |
|
"grad_norm": 0.11965279820742956, |
|
"learning_rate": 3.4791089722651437e-07, |
|
"loss": 0.1022, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 1.8479999999999999, |
|
"grad_norm": 0.1181705374127321, |
|
"learning_rate": 3.1416838871368925e-07, |
|
"loss": 0.1009, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 1.8559999999999999, |
|
"grad_norm": 0.11160263850458792, |
|
"learning_rate": 2.8212062027698995e-07, |
|
"loss": 0.0963, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 1.8639999999999999, |
|
"grad_norm": 0.13030757683951363, |
|
"learning_rate": 2.5177319934793995e-07, |
|
"loss": 0.0983, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 1.8719999999999999, |
|
"grad_norm": 0.11874772797683317, |
|
"learning_rate": 2.2313143584648423e-07, |
|
"loss": 0.092, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"grad_norm": 0.11341961857438242, |
|
"learning_rate": 1.9620034125190645e-07, |
|
"loss": 0.092, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 1.888, |
|
"grad_norm": 0.12018288004198623, |
|
"learning_rate": 1.7098462772596302e-07, |
|
"loss": 0.1013, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 1.896, |
|
"grad_norm": 0.1076575691013852, |
|
"learning_rate": 1.474887072883935e-07, |
|
"loss": 0.0911, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 1.904, |
|
"grad_norm": 0.11337036877085487, |
|
"learning_rate": 1.2571669104494254e-07, |
|
"loss": 0.0881, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 1.912, |
|
"grad_norm": 0.1610883586396456, |
|
"learning_rate": 1.0567238846803995e-07, |
|
"loss": 0.0977, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"grad_norm": 0.11183827260464378, |
|
"learning_rate": 8.735930673024806e-08, |
|
"loss": 0.0869, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 1.928, |
|
"grad_norm": 0.14085464266938202, |
|
"learning_rate": 7.078065009060941e-08, |
|
"loss": 0.0979, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 1.936, |
|
"grad_norm": 0.12681513850389028, |
|
"learning_rate": 5.5939319333998546e-08, |
|
"loss": 0.0955, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 1.944, |
|
"grad_norm": 0.12408511876527423, |
|
"learning_rate": 4.2837911263562406e-08, |
|
"loss": 0.0983, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 1.952, |
|
"grad_norm": 0.12057693933072995, |
|
"learning_rate": 3.147871824635717e-08, |
|
"loss": 0.0978, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"grad_norm": 0.12061596653924865, |
|
"learning_rate": 2.1863727812254653e-08, |
|
"loss": 0.0941, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 1.968, |
|
"grad_norm": 0.11057835512679115, |
|
"learning_rate": 1.3994622306173766e-08, |
|
"loss": 0.0977, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 1.976, |
|
"grad_norm": 0.12585430211339457, |
|
"learning_rate": 7.872778593728258e-09, |
|
"loss": 0.0958, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 1.984, |
|
"grad_norm": 0.11787146911875013, |
|
"learning_rate": 3.499267820307184e-09, |
|
"loss": 0.0919, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 1.992, |
|
"grad_norm": 0.1087435338716272, |
|
"learning_rate": 8.748552236603758e-10, |
|
"loss": 0.0936, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 0.11470541946837681, |
|
"learning_rate": 0.0, |
|
"loss": 0.0898, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"step": 2500, |
|
"total_flos": 22262632857600.0, |
|
"train_loss": 0.14060430166721344, |
|
"train_runtime": 9775.4364, |
|
"train_samples_per_second": 8.184, |
|
"train_steps_per_second": 0.256 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 2500, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 2, |
|
"save_steps": 10000, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 22262632857600.0, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|