{ "best_metric": 0.9340733266061217, "best_model_checkpoint": "/mnt/data4_HDD_14TB/yang/voxceleb-checkpoints/etdnn/voxceleb1/pretrain/ce-len3-bs256-lr5e-4/checkpoint-5230", "epoch": 10.0, "eval_steps": 500, "global_step": 5230, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.03824091778202677, "grad_norm": 10.999133110046387, "learning_rate": 1.9120458891013384e-05, "loss": 7.1473, "step": 20 }, { "epoch": 0.07648183556405354, "grad_norm": 10.200960159301758, "learning_rate": 3.824091778202677e-05, "loss": 7.1283, "step": 40 }, { "epoch": 0.1147227533460803, "grad_norm": 9.468878746032715, "learning_rate": 5.736137667304015e-05, "loss": 7.0847, "step": 60 }, { "epoch": 0.15296367112810708, "grad_norm": 8.736529350280762, "learning_rate": 7.648183556405354e-05, "loss": 7.0321, "step": 80 }, { "epoch": 0.19120458891013384, "grad_norm": 6.856561183929443, "learning_rate": 9.560229445506692e-05, "loss": 6.955, "step": 100 }, { "epoch": 0.2294455066921606, "grad_norm": 5.473419666290283, "learning_rate": 0.0001147227533460803, "loss": 6.8474, "step": 120 }, { "epoch": 0.2676864244741874, "grad_norm": 3.7059528827667236, "learning_rate": 0.0001338432122370937, "loss": 6.7002, "step": 140 }, { "epoch": 0.30592734225621415, "grad_norm": 2.9125759601593018, "learning_rate": 0.00015296367112810707, "loss": 6.5368, "step": 160 }, { "epoch": 0.3441682600382409, "grad_norm": 2.5546936988830566, "learning_rate": 0.00017208413001912047, "loss": 6.3399, "step": 180 }, { "epoch": 0.3824091778202677, "grad_norm": 2.418347120285034, "learning_rate": 0.00019120458891013384, "loss": 6.1511, "step": 200 }, { "epoch": 0.42065009560229444, "grad_norm": 2.2986104488372803, "learning_rate": 0.0002103250478011472, "loss": 6.015, "step": 220 }, { "epoch": 0.4588910133843212, "grad_norm": 2.227661609649658, "learning_rate": 0.0002294455066921606, "loss": 5.8224, "step": 240 }, { "epoch": 0.497131931166348, "grad_norm": 2.2191731929779053, "learning_rate": 0.00024856596558317403, "loss": 5.6838, "step": 260 }, { "epoch": 0.5353728489483748, "grad_norm": 2.2345361709594727, "learning_rate": 0.0002676864244741874, "loss": 5.5764, "step": 280 }, { "epoch": 0.5736137667304015, "grad_norm": 2.2029287815093994, "learning_rate": 0.0002868068833652008, "loss": 5.4715, "step": 300 }, { "epoch": 0.6118546845124283, "grad_norm": 2.1917331218719482, "learning_rate": 0.00030592734225621415, "loss": 5.321, "step": 320 }, { "epoch": 0.6500956022944551, "grad_norm": 2.2598936557769775, "learning_rate": 0.0003250478011472275, "loss": 5.2218, "step": 340 }, { "epoch": 0.6883365200764818, "grad_norm": 2.3335328102111816, "learning_rate": 0.00034416826003824094, "loss": 5.1449, "step": 360 }, { "epoch": 0.7265774378585086, "grad_norm": 2.3767151832580566, "learning_rate": 0.0003632887189292543, "loss": 5.0062, "step": 380 }, { "epoch": 0.7648183556405354, "grad_norm": 2.357161521911621, "learning_rate": 0.0003824091778202677, "loss": 4.9352, "step": 400 }, { "epoch": 0.8030592734225621, "grad_norm": 2.271972179412842, "learning_rate": 0.00040152963671128105, "loss": 4.8959, "step": 420 }, { "epoch": 0.8413001912045889, "grad_norm": 2.4624767303466797, "learning_rate": 0.0004206500956022944, "loss": 4.7951, "step": 440 }, { "epoch": 0.8795411089866156, "grad_norm": 2.1839234828948975, "learning_rate": 0.00043977055449330785, "loss": 4.6734, "step": 460 }, { "epoch": 0.9177820267686424, "grad_norm": 2.3169806003570557, "learning_rate": 0.0004588910133843212, "loss": 4.6458, "step": 480 }, { "epoch": 0.9560229445506692, "grad_norm": 2.3231122493743896, "learning_rate": 0.0004780114722753346, "loss": 4.5657, "step": 500 }, { "epoch": 0.994263862332696, "grad_norm": 2.34975266456604, "learning_rate": 0.0004971319311663481, "loss": 4.489, "step": 520 }, { "epoch": 1.0, "eval_accuracy": 0.1722166162125799, "eval_loss": 4.208949089050293, "eval_runtime": 1092.4973, "eval_samples_per_second": 13.606, "eval_steps_per_second": 13.606, "step": 523 }, { "epoch": 1.0325047801147227, "grad_norm": 2.3869619369506836, "learning_rate": 0.0004981941788825154, "loss": 4.3496, "step": 540 }, { "epoch": 1.0707456978967496, "grad_norm": 2.2834343910217285, "learning_rate": 0.0004960696834501806, "loss": 4.3108, "step": 560 }, { "epoch": 1.1089866156787762, "grad_norm": 2.192180633544922, "learning_rate": 0.0004939451880178457, "loss": 4.229, "step": 580 }, { "epoch": 1.147227533460803, "grad_norm": 2.161285161972046, "learning_rate": 0.0004918206925855109, "loss": 4.1804, "step": 600 }, { "epoch": 1.1854684512428297, "grad_norm": 2.2634170055389404, "learning_rate": 0.0004896961971531761, "loss": 4.118, "step": 620 }, { "epoch": 1.2237093690248566, "grad_norm": 2.1758434772491455, "learning_rate": 0.00048757170172084133, "loss": 4.0295, "step": 640 }, { "epoch": 1.2619502868068833, "grad_norm": 2.25357985496521, "learning_rate": 0.0004854472062885065, "loss": 3.9559, "step": 660 }, { "epoch": 1.3001912045889101, "grad_norm": 2.2077853679656982, "learning_rate": 0.0004833227108561717, "loss": 3.9295, "step": 680 }, { "epoch": 1.338432122370937, "grad_norm": 2.28712797164917, "learning_rate": 0.0004811982154238369, "loss": 3.8221, "step": 700 }, { "epoch": 1.3766730401529637, "grad_norm": 2.2988157272338867, "learning_rate": 0.000479073719991502, "loss": 3.7764, "step": 720 }, { "epoch": 1.4149139579349903, "grad_norm": 2.391023874282837, "learning_rate": 0.0004769492245591672, "loss": 3.7518, "step": 740 }, { "epoch": 1.4531548757170172, "grad_norm": 2.065288782119751, "learning_rate": 0.00047482472912683236, "loss": 3.7074, "step": 760 }, { "epoch": 1.491395793499044, "grad_norm": 2.288686990737915, "learning_rate": 0.00047270023369449756, "loss": 3.6558, "step": 780 }, { "epoch": 1.5296367112810707, "grad_norm": 2.228173017501831, "learning_rate": 0.00047057573826216276, "loss": 3.6038, "step": 800 }, { "epoch": 1.5678776290630974, "grad_norm": 2.1894149780273438, "learning_rate": 0.0004684512428298279, "loss": 3.5405, "step": 820 }, { "epoch": 1.6061185468451242, "grad_norm": 2.234192371368408, "learning_rate": 0.0004663267473974931, "loss": 3.4855, "step": 840 }, { "epoch": 1.644359464627151, "grad_norm": 2.309387445449829, "learning_rate": 0.0004642022519651583, "loss": 3.4771, "step": 860 }, { "epoch": 1.682600382409178, "grad_norm": 2.183579206466675, "learning_rate": 0.00046207775653282345, "loss": 3.3984, "step": 880 }, { "epoch": 1.7208413001912046, "grad_norm": 2.286858558654785, "learning_rate": 0.00045995326110048865, "loss": 3.3318, "step": 900 }, { "epoch": 1.7590822179732313, "grad_norm": 2.2087490558624268, "learning_rate": 0.00045782876566815385, "loss": 3.3256, "step": 920 }, { "epoch": 1.7973231357552581, "grad_norm": 2.2571709156036377, "learning_rate": 0.000455704270235819, "loss": 3.2644, "step": 940 }, { "epoch": 1.835564053537285, "grad_norm": 2.346696138381958, "learning_rate": 0.0004535797748034842, "loss": 3.2687, "step": 960 }, { "epoch": 1.8738049713193117, "grad_norm": 2.309258460998535, "learning_rate": 0.0004514552793711494, "loss": 3.2346, "step": 980 }, { "epoch": 1.9120458891013383, "grad_norm": 2.200920820236206, "learning_rate": 0.00044933078393881453, "loss": 3.1507, "step": 1000 }, { "epoch": 1.9502868068833652, "grad_norm": 2.1627931594848633, "learning_rate": 0.00044720628850647973, "loss": 3.0884, "step": 1020 }, { "epoch": 1.988527724665392, "grad_norm": 2.350935220718384, "learning_rate": 0.00044508179307414493, "loss": 3.0685, "step": 1040 }, { "epoch": 2.0, "eval_accuracy": 0.41096535486041036, "eval_loss": 2.762111186981201, "eval_runtime": 1228.094, "eval_samples_per_second": 12.104, "eval_steps_per_second": 12.104, "step": 1046 }, { "epoch": 2.026768642447419, "grad_norm": 2.137504816055298, "learning_rate": 0.0004429572976418101, "loss": 3.0334, "step": 1060 }, { "epoch": 2.0650095602294454, "grad_norm": 2.2636399269104004, "learning_rate": 0.0004408328022094752, "loss": 2.8843, "step": 1080 }, { "epoch": 2.1032504780114722, "grad_norm": 2.141150712966919, "learning_rate": 0.0004387083067771405, "loss": 2.909, "step": 1100 }, { "epoch": 2.141491395793499, "grad_norm": 2.243589401245117, "learning_rate": 0.0004365838113448056, "loss": 2.8407, "step": 1120 }, { "epoch": 2.179732313575526, "grad_norm": 2.3046114444732666, "learning_rate": 0.00043445931591247077, "loss": 2.8377, "step": 1140 }, { "epoch": 2.2179732313575524, "grad_norm": 2.375234603881836, "learning_rate": 0.000432334820480136, "loss": 2.7642, "step": 1160 }, { "epoch": 2.2562141491395793, "grad_norm": 2.36006236076355, "learning_rate": 0.00043021032504780116, "loss": 2.8011, "step": 1180 }, { "epoch": 2.294455066921606, "grad_norm": 2.246645450592041, "learning_rate": 0.0004280858296154663, "loss": 2.7502, "step": 1200 }, { "epoch": 2.332695984703633, "grad_norm": 2.420032501220703, "learning_rate": 0.0004259613341831315, "loss": 2.6646, "step": 1220 }, { "epoch": 2.3709369024856595, "grad_norm": 2.3049144744873047, "learning_rate": 0.0004238368387507967, "loss": 2.7294, "step": 1240 }, { "epoch": 2.4091778202676863, "grad_norm": 2.3154027462005615, "learning_rate": 0.00042171234331846185, "loss": 2.7323, "step": 1260 }, { "epoch": 2.447418738049713, "grad_norm": 2.3349056243896484, "learning_rate": 0.00041958784788612705, "loss": 2.6435, "step": 1280 }, { "epoch": 2.48565965583174, "grad_norm": 2.3369665145874023, "learning_rate": 0.00041746335245379225, "loss": 2.5938, "step": 1300 }, { "epoch": 2.5239005736137665, "grad_norm": 2.115408182144165, "learning_rate": 0.0004153388570214574, "loss": 2.5557, "step": 1320 }, { "epoch": 2.5621414913957934, "grad_norm": 2.2280750274658203, "learning_rate": 0.00041321436158912254, "loss": 2.5262, "step": 1340 }, { "epoch": 2.6003824091778203, "grad_norm": 2.3640174865722656, "learning_rate": 0.0004110898661567878, "loss": 2.5302, "step": 1360 }, { "epoch": 2.638623326959847, "grad_norm": 2.220531940460205, "learning_rate": 0.00040896537072445294, "loss": 2.4982, "step": 1380 }, { "epoch": 2.676864244741874, "grad_norm": 2.19985294342041, "learning_rate": 0.0004068408752921181, "loss": 2.4884, "step": 1400 }, { "epoch": 2.7151051625239004, "grad_norm": 2.083179235458374, "learning_rate": 0.00040471637985978334, "loss": 2.4422, "step": 1420 }, { "epoch": 2.7533460803059273, "grad_norm": 2.171346664428711, "learning_rate": 0.0004025918844274485, "loss": 2.4187, "step": 1440 }, { "epoch": 2.791586998087954, "grad_norm": 2.330702304840088, "learning_rate": 0.0004004673889951136, "loss": 2.3988, "step": 1460 }, { "epoch": 2.8298279158699806, "grad_norm": 2.3132684230804443, "learning_rate": 0.0003983428935627789, "loss": 2.3606, "step": 1480 }, { "epoch": 2.8680688336520075, "grad_norm": 2.188427686691284, "learning_rate": 0.000396218398130444, "loss": 2.3411, "step": 1500 }, { "epoch": 2.9063097514340344, "grad_norm": 2.1979410648345947, "learning_rate": 0.00039409390269810917, "loss": 2.3314, "step": 1520 }, { "epoch": 2.9445506692160612, "grad_norm": 2.246650218963623, "learning_rate": 0.0003919694072657744, "loss": 2.3221, "step": 1540 }, { "epoch": 2.982791586998088, "grad_norm": 2.2953240871429443, "learning_rate": 0.00038984491183343957, "loss": 2.2892, "step": 1560 }, { "epoch": 3.0, "eval_accuracy": 0.6542885973763874, "eval_loss": 1.6626524925231934, "eval_runtime": 1164.8432, "eval_samples_per_second": 12.761, "eval_steps_per_second": 12.761, "step": 1569 }, { "epoch": 3.0210325047801145, "grad_norm": 2.2067298889160156, "learning_rate": 0.0003877204164011047, "loss": 2.2228, "step": 1580 }, { "epoch": 3.0592734225621414, "grad_norm": 2.2736258506774902, "learning_rate": 0.00038559592096876997, "loss": 2.1788, "step": 1600 }, { "epoch": 3.0975143403441683, "grad_norm": 2.324817419052124, "learning_rate": 0.0003834714255364351, "loss": 2.1743, "step": 1620 }, { "epoch": 3.135755258126195, "grad_norm": 2.188101291656494, "learning_rate": 0.00038134693010410026, "loss": 2.1267, "step": 1640 }, { "epoch": 3.173996175908222, "grad_norm": 2.2637600898742676, "learning_rate": 0.00037922243467176546, "loss": 2.0835, "step": 1660 }, { "epoch": 3.2122370936902485, "grad_norm": 2.2315452098846436, "learning_rate": 0.00037709793923943065, "loss": 2.0944, "step": 1680 }, { "epoch": 3.2504780114722753, "grad_norm": 2.236682176589966, "learning_rate": 0.0003749734438070958, "loss": 2.0867, "step": 1700 }, { "epoch": 3.288718929254302, "grad_norm": 2.1970863342285156, "learning_rate": 0.000372848948374761, "loss": 2.0138, "step": 1720 }, { "epoch": 3.3269598470363286, "grad_norm": 2.2857913970947266, "learning_rate": 0.0003707244529424262, "loss": 2.0081, "step": 1740 }, { "epoch": 3.3652007648183555, "grad_norm": 2.256211280822754, "learning_rate": 0.00036859995751009134, "loss": 1.9887, "step": 1760 }, { "epoch": 3.4034416826003824, "grad_norm": 2.3634960651397705, "learning_rate": 0.00036647546207775654, "loss": 1.9965, "step": 1780 }, { "epoch": 3.4416826003824093, "grad_norm": 2.2254064083099365, "learning_rate": 0.00036435096664542174, "loss": 2.0024, "step": 1800 }, { "epoch": 3.479923518164436, "grad_norm": 2.271703004837036, "learning_rate": 0.0003622264712130869, "loss": 2.0024, "step": 1820 }, { "epoch": 3.5181644359464626, "grad_norm": 2.220524549484253, "learning_rate": 0.0003601019757807521, "loss": 1.9561, "step": 1840 }, { "epoch": 3.5564053537284894, "grad_norm": 2.3010692596435547, "learning_rate": 0.0003579774803484173, "loss": 1.9641, "step": 1860 }, { "epoch": 3.5946462715105163, "grad_norm": 2.2860300540924072, "learning_rate": 0.00035585298491608243, "loss": 1.9032, "step": 1880 }, { "epoch": 3.632887189292543, "grad_norm": 2.2572293281555176, "learning_rate": 0.00035372848948374763, "loss": 1.9115, "step": 1900 }, { "epoch": 3.67112810707457, "grad_norm": 2.441771984100342, "learning_rate": 0.0003516039940514128, "loss": 1.8957, "step": 1920 }, { "epoch": 3.7093690248565965, "grad_norm": 2.3750736713409424, "learning_rate": 0.00034947949861907797, "loss": 1.8638, "step": 1940 }, { "epoch": 3.7476099426386233, "grad_norm": 2.114288806915283, "learning_rate": 0.00034735500318674317, "loss": 1.8105, "step": 1960 }, { "epoch": 3.78585086042065, "grad_norm": 2.3129258155822754, "learning_rate": 0.0003452305077544083, "loss": 1.8908, "step": 1980 }, { "epoch": 3.8240917782026767, "grad_norm": 2.217604398727417, "learning_rate": 0.0003431060123220735, "loss": 1.8116, "step": 2000 }, { "epoch": 3.8623326959847035, "grad_norm": 2.2592735290527344, "learning_rate": 0.0003409815168897387, "loss": 1.8325, "step": 2020 }, { "epoch": 3.9005736137667304, "grad_norm": 2.321389675140381, "learning_rate": 0.00033885702145740386, "loss": 1.7757, "step": 2040 }, { "epoch": 3.9388145315487573, "grad_norm": 2.0709521770477295, "learning_rate": 0.00033673252602506906, "loss": 1.7801, "step": 2060 }, { "epoch": 3.977055449330784, "grad_norm": 2.3316352367401123, "learning_rate": 0.00033460803059273426, "loss": 1.7576, "step": 2080 }, { "epoch": 4.0, "eval_accuracy": 0.758627648839556, "eval_loss": 1.1761366128921509, "eval_runtime": 910.1664, "eval_samples_per_second": 16.332, "eval_steps_per_second": 16.332, "step": 2092 }, { "epoch": 4.015296367112811, "grad_norm": 2.2446036338806152, "learning_rate": 0.0003324835351603994, "loss": 1.7157, "step": 2100 }, { "epoch": 4.053537284894838, "grad_norm": 2.1550064086914062, "learning_rate": 0.0003303590397280646, "loss": 1.6126, "step": 2120 }, { "epoch": 4.091778202676864, "grad_norm": 2.3542799949645996, "learning_rate": 0.0003282345442957298, "loss": 1.6417, "step": 2140 }, { "epoch": 4.130019120458891, "grad_norm": 2.3036603927612305, "learning_rate": 0.00032611004886339494, "loss": 1.6108, "step": 2160 }, { "epoch": 4.168260038240918, "grad_norm": 2.168532133102417, "learning_rate": 0.00032398555343106014, "loss": 1.5939, "step": 2180 }, { "epoch": 4.2065009560229445, "grad_norm": 2.257270097732544, "learning_rate": 0.0003218610579987253, "loss": 1.6157, "step": 2200 }, { "epoch": 4.244741873804971, "grad_norm": 2.2476625442504883, "learning_rate": 0.0003197365625663905, "loss": 1.6477, "step": 2220 }, { "epoch": 4.282982791586998, "grad_norm": 2.162182092666626, "learning_rate": 0.00031761206713405563, "loss": 1.5589, "step": 2240 }, { "epoch": 4.321223709369025, "grad_norm": 2.084498167037964, "learning_rate": 0.00031548757170172083, "loss": 1.5683, "step": 2260 }, { "epoch": 4.359464627151052, "grad_norm": 2.1769893169403076, "learning_rate": 0.00031336307626938603, "loss": 1.5324, "step": 2280 }, { "epoch": 4.397705544933078, "grad_norm": 2.1961722373962402, "learning_rate": 0.0003112385808370512, "loss": 1.5551, "step": 2300 }, { "epoch": 4.435946462715105, "grad_norm": 2.2195024490356445, "learning_rate": 0.0003091140854047164, "loss": 1.5724, "step": 2320 }, { "epoch": 4.474187380497132, "grad_norm": 2.18676495552063, "learning_rate": 0.0003069895899723816, "loss": 1.5322, "step": 2340 }, { "epoch": 4.512428298279159, "grad_norm": 2.2305195331573486, "learning_rate": 0.0003048650945400467, "loss": 1.5207, "step": 2360 }, { "epoch": 4.550669216061186, "grad_norm": 2.297010898590088, "learning_rate": 0.0003027405991077119, "loss": 1.5712, "step": 2380 }, { "epoch": 4.588910133843212, "grad_norm": 2.261610746383667, "learning_rate": 0.0003006161036753771, "loss": 1.513, "step": 2400 }, { "epoch": 4.627151051625239, "grad_norm": 2.1144425868988037, "learning_rate": 0.00029849160824304226, "loss": 1.4854, "step": 2420 }, { "epoch": 4.665391969407266, "grad_norm": 2.344909191131592, "learning_rate": 0.00029636711281070746, "loss": 1.5025, "step": 2440 }, { "epoch": 4.7036328871892925, "grad_norm": 2.1943275928497314, "learning_rate": 0.00029424261737837266, "loss": 1.4917, "step": 2460 }, { "epoch": 4.741873804971319, "grad_norm": 2.191070079803467, "learning_rate": 0.0002921181219460378, "loss": 1.4539, "step": 2480 }, { "epoch": 4.780114722753346, "grad_norm": 2.142045259475708, "learning_rate": 0.000289993626513703, "loss": 1.4688, "step": 2500 }, { "epoch": 4.818355640535373, "grad_norm": 2.203152656555176, "learning_rate": 0.0002878691310813682, "loss": 1.4236, "step": 2520 }, { "epoch": 4.8565965583174, "grad_norm": 2.264139175415039, "learning_rate": 0.00028574463564903335, "loss": 1.4211, "step": 2540 }, { "epoch": 4.894837476099426, "grad_norm": 2.173497438430786, "learning_rate": 0.0002836201402166985, "loss": 1.4236, "step": 2560 }, { "epoch": 4.933078393881453, "grad_norm": 2.069350481033325, "learning_rate": 0.00028149564478436375, "loss": 1.3804, "step": 2580 }, { "epoch": 4.97131931166348, "grad_norm": 2.1907660961151123, "learning_rate": 0.0002793711493520289, "loss": 1.3706, "step": 2600 }, { "epoch": 5.0, "eval_accuracy": 0.8203834510595358, "eval_loss": 0.8903235793113708, "eval_runtime": 813.2495, "eval_samples_per_second": 18.279, "eval_steps_per_second": 18.279, "step": 2615 }, { "epoch": 5.009560229445507, "grad_norm": 2.2395236492156982, "learning_rate": 0.00027724665391969404, "loss": 1.3611, "step": 2620 }, { "epoch": 5.047801147227533, "grad_norm": 2.164022922515869, "learning_rate": 0.0002751221584873593, "loss": 1.2885, "step": 2640 }, { "epoch": 5.08604206500956, "grad_norm": 2.335524559020996, "learning_rate": 0.00027299766305502443, "loss": 1.3024, "step": 2660 }, { "epoch": 5.124282982791587, "grad_norm": 2.3510003089904785, "learning_rate": 0.0002708731676226896, "loss": 1.2819, "step": 2680 }, { "epoch": 5.162523900573614, "grad_norm": 2.2182960510253906, "learning_rate": 0.00026874867219035483, "loss": 1.2927, "step": 2700 }, { "epoch": 5.2007648183556405, "grad_norm": 2.164422035217285, "learning_rate": 0.00026662417675802, "loss": 1.2905, "step": 2720 }, { "epoch": 5.239005736137667, "grad_norm": 2.2719788551330566, "learning_rate": 0.0002644996813256851, "loss": 1.3067, "step": 2740 }, { "epoch": 5.277246653919694, "grad_norm": 2.1733806133270264, "learning_rate": 0.0002623751858933504, "loss": 1.2778, "step": 2760 }, { "epoch": 5.315487571701721, "grad_norm": 2.0940115451812744, "learning_rate": 0.0002602506904610155, "loss": 1.2553, "step": 2780 }, { "epoch": 5.353728489483748, "grad_norm": 2.049928665161133, "learning_rate": 0.00025812619502868067, "loss": 1.2572, "step": 2800 }, { "epoch": 5.3919694072657744, "grad_norm": 2.039212703704834, "learning_rate": 0.0002560016995963459, "loss": 1.2139, "step": 2820 }, { "epoch": 5.430210325047801, "grad_norm": 2.2491774559020996, "learning_rate": 0.00025387720416401106, "loss": 1.2591, "step": 2840 }, { "epoch": 5.468451242829828, "grad_norm": 2.1145877838134766, "learning_rate": 0.0002517527087316762, "loss": 1.2458, "step": 2860 }, { "epoch": 5.506692160611855, "grad_norm": 2.1274683475494385, "learning_rate": 0.0002496282132993414, "loss": 1.227, "step": 2880 }, { "epoch": 5.544933078393882, "grad_norm": 2.0846710205078125, "learning_rate": 0.0002475037178670066, "loss": 1.1897, "step": 2900 }, { "epoch": 5.583173996175908, "grad_norm": 2.143688678741455, "learning_rate": 0.0002453792224346718, "loss": 1.2133, "step": 2920 }, { "epoch": 5.621414913957935, "grad_norm": 2.2568676471710205, "learning_rate": 0.00024325472700233695, "loss": 1.2194, "step": 2940 }, { "epoch": 5.659655831739962, "grad_norm": 2.266052484512329, "learning_rate": 0.00024113023157000212, "loss": 1.1895, "step": 2960 }, { "epoch": 5.6978967495219885, "grad_norm": 2.352609395980835, "learning_rate": 0.0002390057361376673, "loss": 1.1849, "step": 2980 }, { "epoch": 5.736137667304015, "grad_norm": 2.1269915103912354, "learning_rate": 0.0002368812407053325, "loss": 1.205, "step": 3000 }, { "epoch": 5.774378585086042, "grad_norm": 2.20017147064209, "learning_rate": 0.00023475674527299767, "loss": 1.1366, "step": 3020 }, { "epoch": 5.812619502868069, "grad_norm": 2.0107390880584717, "learning_rate": 0.00023263224984066284, "loss": 1.1268, "step": 3040 }, { "epoch": 5.850860420650095, "grad_norm": 2.091771364212036, "learning_rate": 0.00023050775440832804, "loss": 1.1619, "step": 3060 }, { "epoch": 5.8891013384321225, "grad_norm": 2.261592388153076, "learning_rate": 0.0002283832589759932, "loss": 1.1789, "step": 3080 }, { "epoch": 5.927342256214149, "grad_norm": 2.126786947250366, "learning_rate": 0.00022625876354365838, "loss": 1.1492, "step": 3100 }, { "epoch": 5.965583173996176, "grad_norm": 2.123417615890503, "learning_rate": 0.00022413426811132355, "loss": 1.1258, "step": 3120 }, { "epoch": 6.0, "eval_accuracy": 0.8432559704002691, "eval_loss": 0.7555378079414368, "eval_runtime": 1118.6028, "eval_samples_per_second": 13.289, "eval_steps_per_second": 13.289, "step": 3138 }, { "epoch": 6.003824091778203, "grad_norm": 2.0020689964294434, "learning_rate": 0.00022200977267898875, "loss": 1.1086, "step": 3140 }, { "epoch": 6.042065009560229, "grad_norm": 2.0867371559143066, "learning_rate": 0.00021988527724665392, "loss": 1.0815, "step": 3160 }, { "epoch": 6.080305927342256, "grad_norm": 1.9533103704452515, "learning_rate": 0.0002177607818143191, "loss": 1.0412, "step": 3180 }, { "epoch": 6.118546845124283, "grad_norm": 2.2522337436676025, "learning_rate": 0.0002156362863819843, "loss": 1.0312, "step": 3200 }, { "epoch": 6.15678776290631, "grad_norm": 2.058983087539673, "learning_rate": 0.00021351179094964944, "loss": 1.0449, "step": 3220 }, { "epoch": 6.195028680688337, "grad_norm": 2.1774373054504395, "learning_rate": 0.00021138729551731464, "loss": 1.0532, "step": 3240 }, { "epoch": 6.233269598470363, "grad_norm": 2.0248055458068848, "learning_rate": 0.00020926280008497984, "loss": 1.0461, "step": 3260 }, { "epoch": 6.27151051625239, "grad_norm": 2.066737651824951, "learning_rate": 0.00020713830465264498, "loss": 1.0243, "step": 3280 }, { "epoch": 6.309751434034417, "grad_norm": 2.100133180618286, "learning_rate": 0.00020501380922031018, "loss": 1.0309, "step": 3300 }, { "epoch": 6.347992351816444, "grad_norm": 2.192960739135742, "learning_rate": 0.00020288931378797538, "loss": 1.0147, "step": 3320 }, { "epoch": 6.3862332695984705, "grad_norm": 2.123382329940796, "learning_rate": 0.00020076481835564053, "loss": 1.0229, "step": 3340 }, { "epoch": 6.424474187380497, "grad_norm": 2.162189245223999, "learning_rate": 0.00019864032292330573, "loss": 1.011, "step": 3360 }, { "epoch": 6.462715105162524, "grad_norm": 2.254284143447876, "learning_rate": 0.0001965158274909709, "loss": 0.9929, "step": 3380 }, { "epoch": 6.500956022944551, "grad_norm": 2.2183146476745605, "learning_rate": 0.00019439133205863607, "loss": 0.9852, "step": 3400 }, { "epoch": 6.539196940726577, "grad_norm": 2.1704065799713135, "learning_rate": 0.00019226683662630127, "loss": 1.0129, "step": 3420 }, { "epoch": 6.577437858508604, "grad_norm": 2.0828094482421875, "learning_rate": 0.00019014234119396644, "loss": 0.9667, "step": 3440 }, { "epoch": 6.615678776290631, "grad_norm": 2.3150341510772705, "learning_rate": 0.0001880178457616316, "loss": 0.9769, "step": 3460 }, { "epoch": 6.653919694072657, "grad_norm": 2.0636775493621826, "learning_rate": 0.0001858933503292968, "loss": 0.979, "step": 3480 }, { "epoch": 6.692160611854685, "grad_norm": 2.171602487564087, "learning_rate": 0.00018376885489696196, "loss": 0.9667, "step": 3500 }, { "epoch": 6.730401529636711, "grad_norm": 2.1150522232055664, "learning_rate": 0.00018164435946462716, "loss": 0.9593, "step": 3520 }, { "epoch": 6.768642447418738, "grad_norm": 2.096452236175537, "learning_rate": 0.00017951986403229233, "loss": 0.9828, "step": 3540 }, { "epoch": 6.806883365200765, "grad_norm": 2.2779390811920166, "learning_rate": 0.0001773953685999575, "loss": 0.9648, "step": 3560 }, { "epoch": 6.845124282982791, "grad_norm": 2.2793240547180176, "learning_rate": 0.0001752708731676227, "loss": 0.9639, "step": 3580 }, { "epoch": 6.8833652007648185, "grad_norm": 2.0822837352752686, "learning_rate": 0.00017314637773528787, "loss": 0.9421, "step": 3600 }, { "epoch": 6.921606118546845, "grad_norm": 2.1541502475738525, "learning_rate": 0.00017102188230295304, "loss": 0.9277, "step": 3620 }, { "epoch": 6.959847036328872, "grad_norm": 2.1241962909698486, "learning_rate": 0.00016889738687061824, "loss": 0.9335, "step": 3640 }, { "epoch": 6.998087954110899, "grad_norm": 2.2354745864868164, "learning_rate": 0.00016677289143828341, "loss": 0.9379, "step": 3660 }, { "epoch": 7.0, "eval_accuracy": 0.889673730238816, "eval_loss": 0.5586521029472351, "eval_runtime": 904.3879, "eval_samples_per_second": 16.437, "eval_steps_per_second": 16.437, "step": 3661 }, { "epoch": 7.036328871892925, "grad_norm": 2.109501600265503, "learning_rate": 0.0001646483960059486, "loss": 0.8485, "step": 3680 }, { "epoch": 7.074569789674952, "grad_norm": 2.2034125328063965, "learning_rate": 0.00016252390057361376, "loss": 0.897, "step": 3700 }, { "epoch": 7.112810707456979, "grad_norm": 2.053767204284668, "learning_rate": 0.00016039940514127896, "loss": 0.859, "step": 3720 }, { "epoch": 7.151051625239006, "grad_norm": 1.977229118347168, "learning_rate": 0.00015827490970894413, "loss": 0.8643, "step": 3740 }, { "epoch": 7.189292543021033, "grad_norm": 2.0782172679901123, "learning_rate": 0.0001561504142766093, "loss": 0.8665, "step": 3760 }, { "epoch": 7.227533460803059, "grad_norm": 1.9473881721496582, "learning_rate": 0.0001540259188442745, "loss": 0.8748, "step": 3780 }, { "epoch": 7.265774378585086, "grad_norm": 2.0109825134277344, "learning_rate": 0.00015190142341193967, "loss": 0.8589, "step": 3800 }, { "epoch": 7.304015296367113, "grad_norm": 2.1534509658813477, "learning_rate": 0.00014977692797960484, "loss": 0.8373, "step": 3820 }, { "epoch": 7.342256214149139, "grad_norm": 2.1457502841949463, "learning_rate": 0.00014765243254727004, "loss": 0.8487, "step": 3840 }, { "epoch": 7.3804971319311665, "grad_norm": 2.200446844100952, "learning_rate": 0.0001455279371149352, "loss": 0.841, "step": 3860 }, { "epoch": 7.418738049713193, "grad_norm": 2.0335378646850586, "learning_rate": 0.0001434034416826004, "loss": 0.846, "step": 3880 }, { "epoch": 7.45697896749522, "grad_norm": 2.1471550464630127, "learning_rate": 0.0001412789462502656, "loss": 0.8442, "step": 3900 }, { "epoch": 7.495219885277247, "grad_norm": 2.2554261684417725, "learning_rate": 0.00013915445081793073, "loss": 0.8651, "step": 3920 }, { "epoch": 7.533460803059273, "grad_norm": 2.1498897075653076, "learning_rate": 0.00013702995538559593, "loss": 0.839, "step": 3940 }, { "epoch": 7.5717017208413, "grad_norm": 2.1460678577423096, "learning_rate": 0.00013490545995326113, "loss": 0.8449, "step": 3960 }, { "epoch": 7.609942638623327, "grad_norm": 2.194730281829834, "learning_rate": 0.00013278096452092628, "loss": 0.8075, "step": 3980 }, { "epoch": 7.648183556405353, "grad_norm": 2.291327714920044, "learning_rate": 0.00013065646908859147, "loss": 0.8163, "step": 4000 }, { "epoch": 7.686424474187381, "grad_norm": 2.049647808074951, "learning_rate": 0.00012853197365625662, "loss": 0.7999, "step": 4020 }, { "epoch": 7.724665391969407, "grad_norm": 2.026001214981079, "learning_rate": 0.00012640747822392182, "loss": 0.7721, "step": 4040 }, { "epoch": 7.762906309751434, "grad_norm": 2.1242027282714844, "learning_rate": 0.00012428298279158702, "loss": 0.7957, "step": 4060 }, { "epoch": 7.801147227533461, "grad_norm": 2.078123092651367, "learning_rate": 0.0001221584873592522, "loss": 0.8244, "step": 4080 }, { "epoch": 7.839388145315487, "grad_norm": 2.028822183609009, "learning_rate": 0.00012003399192691736, "loss": 0.8066, "step": 4100 }, { "epoch": 7.8776290630975145, "grad_norm": 2.30352520942688, "learning_rate": 0.00011790949649458253, "loss": 0.7892, "step": 4120 }, { "epoch": 7.915869980879541, "grad_norm": 2.100221872329712, "learning_rate": 0.00011578500106224772, "loss": 0.7703, "step": 4140 }, { "epoch": 7.954110898661568, "grad_norm": 2.1163010597229004, "learning_rate": 0.00011366050562991289, "loss": 0.7958, "step": 4160 }, { "epoch": 7.992351816443595, "grad_norm": 2.194089412689209, "learning_rate": 0.00011153601019757808, "loss": 0.7925, "step": 4180 }, { "epoch": 8.0, "eval_accuracy": 0.9117389841910528, "eval_loss": 0.4518139362335205, "eval_runtime": 1278.572, "eval_samples_per_second": 11.626, "eval_steps_per_second": 11.626, "step": 4184 }, { "epoch": 8.030592734225621, "grad_norm": 2.024029016494751, "learning_rate": 0.00010941151476524326, "loss": 0.7768, "step": 4200 }, { "epoch": 8.068833652007648, "grad_norm": 2.101900577545166, "learning_rate": 0.00010728701933290843, "loss": 0.7174, "step": 4220 }, { "epoch": 8.107074569789676, "grad_norm": 1.8774019479751587, "learning_rate": 0.0001051625239005736, "loss": 0.7331, "step": 4240 }, { "epoch": 8.145315487571702, "grad_norm": 2.0548768043518066, "learning_rate": 0.0001030380284682388, "loss": 0.6988, "step": 4260 }, { "epoch": 8.183556405353729, "grad_norm": 2.09981107711792, "learning_rate": 0.00010091353303590398, "loss": 0.7329, "step": 4280 }, { "epoch": 8.221797323135755, "grad_norm": 1.9958146810531616, "learning_rate": 9.878903760356915e-05, "loss": 0.7085, "step": 4300 }, { "epoch": 8.260038240917781, "grad_norm": 1.9749237298965454, "learning_rate": 9.666454217123433e-05, "loss": 0.715, "step": 4320 }, { "epoch": 8.29827915869981, "grad_norm": 1.9691994190216064, "learning_rate": 9.454004673889952e-05, "loss": 0.7151, "step": 4340 }, { "epoch": 8.336520076481836, "grad_norm": 2.1549344062805176, "learning_rate": 9.241555130656469e-05, "loss": 0.7249, "step": 4360 }, { "epoch": 8.374760994263863, "grad_norm": 2.2257604598999023, "learning_rate": 9.029105587422986e-05, "loss": 0.7207, "step": 4380 }, { "epoch": 8.413001912045889, "grad_norm": 2.0424046516418457, "learning_rate": 8.816656044189505e-05, "loss": 0.6981, "step": 4400 }, { "epoch": 8.451242829827915, "grad_norm": 1.8724883794784546, "learning_rate": 8.604206500956024e-05, "loss": 0.7153, "step": 4420 }, { "epoch": 8.489483747609942, "grad_norm": 2.0893712043762207, "learning_rate": 8.391756957722541e-05, "loss": 0.7254, "step": 4440 }, { "epoch": 8.52772466539197, "grad_norm": 2.0398759841918945, "learning_rate": 8.179307414489059e-05, "loss": 0.7184, "step": 4460 }, { "epoch": 8.565965583173996, "grad_norm": 2.1322920322418213, "learning_rate": 7.966857871255577e-05, "loss": 0.7189, "step": 4480 }, { "epoch": 8.604206500956023, "grad_norm": 2.0477068424224854, "learning_rate": 7.754408328022095e-05, "loss": 0.705, "step": 4500 }, { "epoch": 8.64244741873805, "grad_norm": 2.0737295150756836, "learning_rate": 7.541958784788614e-05, "loss": 0.7285, "step": 4520 }, { "epoch": 8.680688336520076, "grad_norm": 1.9542038440704346, "learning_rate": 7.329509241555131e-05, "loss": 0.7028, "step": 4540 }, { "epoch": 8.718929254302104, "grad_norm": 1.9409737586975098, "learning_rate": 7.117059698321648e-05, "loss": 0.7031, "step": 4560 }, { "epoch": 8.75717017208413, "grad_norm": 1.982315182685852, "learning_rate": 6.904610155088168e-05, "loss": 0.6909, "step": 4580 }, { "epoch": 8.795411089866157, "grad_norm": 2.2722830772399902, "learning_rate": 6.692160611854685e-05, "loss": 0.7083, "step": 4600 }, { "epoch": 8.833652007648183, "grad_norm": 1.9158116579055786, "learning_rate": 6.479711068621202e-05, "loss": 0.6885, "step": 4620 }, { "epoch": 8.87189292543021, "grad_norm": 2.058122396469116, "learning_rate": 6.26726152538772e-05, "loss": 0.6912, "step": 4640 }, { "epoch": 8.910133843212238, "grad_norm": 2.013585090637207, "learning_rate": 6.054811982154239e-05, "loss": 0.6937, "step": 4660 }, { "epoch": 8.948374760994264, "grad_norm": 2.1662707328796387, "learning_rate": 5.8423624389207567e-05, "loss": 0.7003, "step": 4680 }, { "epoch": 8.98661567877629, "grad_norm": 2.06309175491333, "learning_rate": 5.6299128956872745e-05, "loss": 0.6733, "step": 4700 }, { "epoch": 9.0, "eval_accuracy": 0.929297006390851, "eval_loss": 0.38889044523239136, "eval_runtime": 1336.4014, "eval_samples_per_second": 11.123, "eval_steps_per_second": 11.123, "step": 4707 }, { "epoch": 9.024856596558317, "grad_norm": 1.8986388444900513, "learning_rate": 5.4174633524537924e-05, "loss": 0.6593, "step": 4720 }, { "epoch": 9.063097514340344, "grad_norm": 1.9107776880264282, "learning_rate": 5.20501380922031e-05, "loss": 0.6354, "step": 4740 }, { "epoch": 9.101338432122372, "grad_norm": 2.0714404582977295, "learning_rate": 4.992564265986829e-05, "loss": 0.6531, "step": 4760 }, { "epoch": 9.139579349904398, "grad_norm": 1.9571270942687988, "learning_rate": 4.780114722753346e-05, "loss": 0.6429, "step": 4780 }, { "epoch": 9.177820267686425, "grad_norm": 1.9968469142913818, "learning_rate": 4.5676651795198646e-05, "loss": 0.6498, "step": 4800 }, { "epoch": 9.216061185468451, "grad_norm": 1.982832670211792, "learning_rate": 4.355215636286382e-05, "loss": 0.6141, "step": 4820 }, { "epoch": 9.254302103250478, "grad_norm": 2.035123825073242, "learning_rate": 4.1427660930529e-05, "loss": 0.6269, "step": 4840 }, { "epoch": 9.292543021032504, "grad_norm": 1.8670984506607056, "learning_rate": 3.930316549819418e-05, "loss": 0.6143, "step": 4860 }, { "epoch": 9.330783938814532, "grad_norm": 1.9521456956863403, "learning_rate": 3.7178670065859354e-05, "loss": 0.6352, "step": 4880 }, { "epoch": 9.369024856596559, "grad_norm": 1.9804260730743408, "learning_rate": 3.505417463352454e-05, "loss": 0.6214, "step": 4900 }, { "epoch": 9.407265774378585, "grad_norm": 2.149338483810425, "learning_rate": 3.292967920118971e-05, "loss": 0.655, "step": 4920 }, { "epoch": 9.445506692160611, "grad_norm": 1.8145477771759033, "learning_rate": 3.08051837688549e-05, "loss": 0.6095, "step": 4940 }, { "epoch": 9.483747609942638, "grad_norm": 2.009587287902832, "learning_rate": 2.8680688336520076e-05, "loss": 0.6314, "step": 4960 }, { "epoch": 9.521988527724666, "grad_norm": 1.946964979171753, "learning_rate": 2.655619290418526e-05, "loss": 0.6272, "step": 4980 }, { "epoch": 9.560229445506693, "grad_norm": 2.2486774921417236, "learning_rate": 2.4431697471850437e-05, "loss": 0.6133, "step": 5000 }, { "epoch": 9.598470363288719, "grad_norm": 1.9570778608322144, "learning_rate": 2.2307202039515616e-05, "loss": 0.6185, "step": 5020 }, { "epoch": 9.636711281070745, "grad_norm": 1.8994139432907104, "learning_rate": 2.0182706607180795e-05, "loss": 0.6366, "step": 5040 }, { "epoch": 9.674952198852772, "grad_norm": 2.102374792098999, "learning_rate": 1.8058211174845977e-05, "loss": 0.6414, "step": 5060 }, { "epoch": 9.7131931166348, "grad_norm": 2.073791027069092, "learning_rate": 1.5933715742511156e-05, "loss": 0.6053, "step": 5080 }, { "epoch": 9.751434034416826, "grad_norm": 1.9205766916275024, "learning_rate": 1.3809220310176333e-05, "loss": 0.6032, "step": 5100 }, { "epoch": 9.789674952198853, "grad_norm": 2.04294490814209, "learning_rate": 1.1684724877841512e-05, "loss": 0.6202, "step": 5120 }, { "epoch": 9.82791586998088, "grad_norm": 1.9936119318008423, "learning_rate": 9.560229445506692e-06, "loss": 0.5882, "step": 5140 }, { "epoch": 9.866156787762906, "grad_norm": 1.8728615045547485, "learning_rate": 7.435734013171872e-06, "loss": 0.5955, "step": 5160 }, { "epoch": 9.904397705544934, "grad_norm": 2.069910764694214, "learning_rate": 5.311238580837051e-06, "loss": 0.6049, "step": 5180 }, { "epoch": 9.94263862332696, "grad_norm": 1.9873836040496826, "learning_rate": 3.1867431485022306e-06, "loss": 0.6251, "step": 5200 }, { "epoch": 9.980879541108987, "grad_norm": 1.932947039604187, "learning_rate": 1.0622477161674102e-06, "loss": 0.6187, "step": 5220 }, { "epoch": 10.0, "eval_accuracy": 0.9340733266061217, "eval_loss": 0.35941872000694275, "eval_runtime": 1417.2186, "eval_samples_per_second": 10.489, "eval_steps_per_second": 10.489, "step": 5230 }, { "epoch": 10.0, "step": 5230, "total_flos": 2.7283641066057605e+18, "train_loss": 1.9864544938672797, "train_runtime": 94798.7749, "train_samples_per_second": 14.112, "train_steps_per_second": 0.055 } ], "logging_steps": 20, "max_steps": 5230, "num_input_tokens_seen": 0, "num_train_epochs": 10, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.7283641066057605e+18, "train_batch_size": 256, "trial_name": null, "trial_params": null }