{ "best_metric": null, "best_model_checkpoint": null, "epoch": 6.0, "eval_steps": 500, "global_step": 7692, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0078003120124804995, "grad_norm": 2.144841432571411, "learning_rate": 0.0009989971027412525, "loss": 1.1174, "step": 10 }, { "epoch": 0.015600624024960999, "grad_norm": 1.0625462532043457, "learning_rate": 0.0009978827724537553, "loss": 0.5035, "step": 20 }, { "epoch": 0.0234009360374415, "grad_norm": 1.619573950767517, "learning_rate": 0.0009967684421662581, "loss": 0.3841, "step": 30 }, { "epoch": 0.031201248049921998, "grad_norm": 3.2681403160095215, "learning_rate": 0.000995654111878761, "loss": 0.4974, "step": 40 }, { "epoch": 0.0390015600624025, "grad_norm": 1.5944786071777344, "learning_rate": 0.0009945397815912637, "loss": 0.4797, "step": 50 }, { "epoch": 0.046801872074883, "grad_norm": 1.8234444856643677, "learning_rate": 0.0009934254513037665, "loss": 0.637, "step": 60 }, { "epoch": 0.054602184087363496, "grad_norm": 3.790844440460205, "learning_rate": 0.0009923111210162691, "loss": 0.4479, "step": 70 }, { "epoch": 0.062402496099843996, "grad_norm": 2.9351046085357666, "learning_rate": 0.000991196790728772, "loss": 0.4332, "step": 80 }, { "epoch": 0.07020280811232449, "grad_norm": 2.2770025730133057, "learning_rate": 0.0009900824604412747, "loss": 0.4399, "step": 90 }, { "epoch": 0.078003120124805, "grad_norm": 1.86591637134552, "learning_rate": 0.0009889681301537775, "loss": 0.407, "step": 100 }, { "epoch": 0.08580343213728549, "grad_norm": 1.6851640939712524, "learning_rate": 0.0009878537998662805, "loss": 0.3862, "step": 110 }, { "epoch": 0.093603744149766, "grad_norm": 2.5469818115234375, "learning_rate": 0.0009867394695787831, "loss": 0.4166, "step": 120 }, { "epoch": 0.10140405616224649, "grad_norm": 1.8210259675979614, "learning_rate": 0.000985625139291286, "loss": 0.3785, "step": 130 }, { "epoch": 0.10920436817472699, "grad_norm": 2.031057119369507, "learning_rate": 0.0009845108090037887, "loss": 0.4177, "step": 140 }, { "epoch": 0.11700468018720749, "grad_norm": 1.6646612882614136, "learning_rate": 0.0009833964787162915, "loss": 0.402, "step": 150 }, { "epoch": 0.12480499219968799, "grad_norm": 1.8680285215377808, "learning_rate": 0.0009822821484287943, "loss": 0.3282, "step": 160 }, { "epoch": 0.13260530421216848, "grad_norm": 1.8039604425430298, "learning_rate": 0.0009811678181412971, "loss": 0.3536, "step": 170 }, { "epoch": 0.14040561622464898, "grad_norm": 3.3018901348114014, "learning_rate": 0.0009800534878538, "loss": 0.4595, "step": 180 }, { "epoch": 0.1482059282371295, "grad_norm": 3.684013843536377, "learning_rate": 0.0009789391575663027, "loss": 0.4288, "step": 190 }, { "epoch": 0.15600624024961, "grad_norm": 1.4512592554092407, "learning_rate": 0.0009778248272788055, "loss": 0.5086, "step": 200 }, { "epoch": 0.16380655226209048, "grad_norm": 2.3981761932373047, "learning_rate": 0.0009767104969913081, "loss": 0.4084, "step": 210 }, { "epoch": 0.17160686427457097, "grad_norm": 3.7943010330200195, "learning_rate": 0.000975596166703811, "loss": 0.4524, "step": 220 }, { "epoch": 0.1794071762870515, "grad_norm": 2.657606840133667, "learning_rate": 0.0009744818364163138, "loss": 0.3592, "step": 230 }, { "epoch": 0.187207488299532, "grad_norm": 2.7629363536834717, "learning_rate": 0.0009733675061288166, "loss": 0.4263, "step": 240 }, { "epoch": 0.19500780031201248, "grad_norm": 1.3749983310699463, "learning_rate": 0.0009722531758413193, "loss": 0.48, "step": 250 }, { "epoch": 0.20280811232449297, "grad_norm": 2.648716449737549, "learning_rate": 0.0009711388455538221, "loss": 0.416, "step": 260 }, { "epoch": 0.21060842433697347, "grad_norm": 1.5672308206558228, "learning_rate": 0.0009700245152663249, "loss": 0.4223, "step": 270 }, { "epoch": 0.21840873634945399, "grad_norm": 2.618163585662842, "learning_rate": 0.0009689101849788277, "loss": 0.4172, "step": 280 }, { "epoch": 0.22620904836193448, "grad_norm": 3.6365268230438232, "learning_rate": 0.0009677958546913305, "loss": 0.5501, "step": 290 }, { "epoch": 0.23400936037441497, "grad_norm": 2.740039825439453, "learning_rate": 0.0009666815244038332, "loss": 0.3553, "step": 300 }, { "epoch": 0.24180967238689546, "grad_norm": 3.406210422515869, "learning_rate": 0.000965567194116336, "loss": 0.3518, "step": 310 }, { "epoch": 0.24960998439937598, "grad_norm": 1.4707075357437134, "learning_rate": 0.000964452863828839, "loss": 0.3452, "step": 320 }, { "epoch": 0.2574102964118565, "grad_norm": 1.608324408531189, "learning_rate": 0.0009633385335413417, "loss": 0.4908, "step": 330 }, { "epoch": 0.26521060842433697, "grad_norm": 4.090480327606201, "learning_rate": 0.0009622242032538444, "loss": 0.4597, "step": 340 }, { "epoch": 0.27301092043681746, "grad_norm": 2.2214395999908447, "learning_rate": 0.0009611098729663472, "loss": 0.4552, "step": 350 }, { "epoch": 0.28081123244929795, "grad_norm": 1.9134166240692139, "learning_rate": 0.00095999554267885, "loss": 0.3571, "step": 360 }, { "epoch": 0.28861154446177845, "grad_norm": 1.8127851486206055, "learning_rate": 0.0009588812123913528, "loss": 0.3808, "step": 370 }, { "epoch": 0.296411856474259, "grad_norm": 2.2262885570526123, "learning_rate": 0.0009577668821038556, "loss": 0.4099, "step": 380 }, { "epoch": 0.3042121684867395, "grad_norm": 2.8041303157806396, "learning_rate": 0.0009566525518163583, "loss": 0.3988, "step": 390 }, { "epoch": 0.31201248049922, "grad_norm": 6.797432899475098, "learning_rate": 0.0009555382215288611, "loss": 0.4728, "step": 400 }, { "epoch": 0.31981279251170047, "grad_norm": 3.1861369609832764, "learning_rate": 0.000954423891241364, "loss": 0.3502, "step": 410 }, { "epoch": 0.32761310452418096, "grad_norm": 2.9223642349243164, "learning_rate": 0.0009533095609538667, "loss": 0.4215, "step": 420 }, { "epoch": 0.33541341653666146, "grad_norm": 6.848895072937012, "learning_rate": 0.0009521952306663694, "loss": 0.415, "step": 430 }, { "epoch": 0.34321372854914195, "grad_norm": 9.054282188415527, "learning_rate": 0.0009510809003788722, "loss": 0.4667, "step": 440 }, { "epoch": 0.35101404056162244, "grad_norm": 2.3005900382995605, "learning_rate": 0.0009499665700913752, "loss": 0.6565, "step": 450 }, { "epoch": 0.358814352574103, "grad_norm": 2.9467573165893555, "learning_rate": 0.000948852239803878, "loss": 0.3837, "step": 460 }, { "epoch": 0.3666146645865835, "grad_norm": 1.5977652072906494, "learning_rate": 0.0009477379095163808, "loss": 0.407, "step": 470 }, { "epoch": 0.374414976599064, "grad_norm": 2.8274600505828857, "learning_rate": 0.0009466235792288835, "loss": 0.4501, "step": 480 }, { "epoch": 0.38221528861154447, "grad_norm": 1.8566502332687378, "learning_rate": 0.0009455092489413863, "loss": 0.3146, "step": 490 }, { "epoch": 0.39001560062402496, "grad_norm": 2.5871951580047607, "learning_rate": 0.0009443949186538891, "loss": 0.3335, "step": 500 }, { "epoch": 0.39781591263650545, "grad_norm": 2.0552711486816406, "learning_rate": 0.0009432805883663919, "loss": 0.3698, "step": 510 }, { "epoch": 0.40561622464898595, "grad_norm": 1.5244548320770264, "learning_rate": 0.0009421662580788946, "loss": 0.581, "step": 520 }, { "epoch": 0.41341653666146644, "grad_norm": 1.5146633386611938, "learning_rate": 0.0009410519277913974, "loss": 0.3634, "step": 530 }, { "epoch": 0.42121684867394693, "grad_norm": 2.394819736480713, "learning_rate": 0.0009399375975039002, "loss": 0.338, "step": 540 }, { "epoch": 0.4290171606864275, "grad_norm": 2.3500325679779053, "learning_rate": 0.000938823267216403, "loss": 0.4025, "step": 550 }, { "epoch": 0.43681747269890797, "grad_norm": 2.4186370372772217, "learning_rate": 0.0009377089369289058, "loss": 0.4095, "step": 560 }, { "epoch": 0.44461778471138846, "grad_norm": 1.9770065546035767, "learning_rate": 0.0009365946066414085, "loss": 0.4275, "step": 570 }, { "epoch": 0.45241809672386896, "grad_norm": 1.7679632902145386, "learning_rate": 0.0009354802763539114, "loss": 0.3949, "step": 580 }, { "epoch": 0.46021840873634945, "grad_norm": 2.0794620513916016, "learning_rate": 0.0009343659460664142, "loss": 0.4208, "step": 590 }, { "epoch": 0.46801872074882994, "grad_norm": 2.677424192428589, "learning_rate": 0.000933251615778917, "loss": 0.4089, "step": 600 }, { "epoch": 0.47581903276131043, "grad_norm": 1.526112675666809, "learning_rate": 0.0009321372854914197, "loss": 0.4196, "step": 610 }, { "epoch": 0.4836193447737909, "grad_norm": 1.8656370639801025, "learning_rate": 0.0009310229552039225, "loss": 0.3732, "step": 620 }, { "epoch": 0.4914196567862715, "grad_norm": 3.3338847160339355, "learning_rate": 0.0009299086249164253, "loss": 0.4231, "step": 630 }, { "epoch": 0.49921996879875197, "grad_norm": 2.1057350635528564, "learning_rate": 0.0009287942946289281, "loss": 0.3921, "step": 640 }, { "epoch": 0.5070202808112324, "grad_norm": 1.544977068901062, "learning_rate": 0.0009276799643414309, "loss": 0.3748, "step": 650 }, { "epoch": 0.514820592823713, "grad_norm": 3.4070258140563965, "learning_rate": 0.0009265656340539336, "loss": 0.4027, "step": 660 }, { "epoch": 0.5226209048361935, "grad_norm": 5.5486931800842285, "learning_rate": 0.0009254513037664364, "loss": 0.4326, "step": 670 }, { "epoch": 0.5304212168486739, "grad_norm": 3.6824769973754883, "learning_rate": 0.0009243369734789392, "loss": 0.4809, "step": 680 }, { "epoch": 0.5382215288611545, "grad_norm": 6.3154778480529785, "learning_rate": 0.000923222643191442, "loss": 0.3588, "step": 690 }, { "epoch": 0.5460218408736349, "grad_norm": 3.133465528488159, "learning_rate": 0.0009221083129039447, "loss": 0.4712, "step": 700 }, { "epoch": 0.5538221528861155, "grad_norm": 4.222598552703857, "learning_rate": 0.0009209939826164475, "loss": 0.4135, "step": 710 }, { "epoch": 0.5616224648985959, "grad_norm": 4.6125078201293945, "learning_rate": 0.0009198796523289504, "loss": 0.393, "step": 720 }, { "epoch": 0.5694227769110765, "grad_norm": 6.543318748474121, "learning_rate": 0.0009187653220414532, "loss": 0.6138, "step": 730 }, { "epoch": 0.5772230889235569, "grad_norm": 2.596463680267334, "learning_rate": 0.000917650991753956, "loss": 0.3694, "step": 740 }, { "epoch": 0.5850234009360374, "grad_norm": 2.428490161895752, "learning_rate": 0.0009165366614664587, "loss": 0.4535, "step": 750 }, { "epoch": 0.592823712948518, "grad_norm": 1.8790688514709473, "learning_rate": 0.0009154223311789615, "loss": 0.3986, "step": 760 }, { "epoch": 0.6006240249609984, "grad_norm": 3.141587734222412, "learning_rate": 0.0009143080008914643, "loss": 0.3881, "step": 770 }, { "epoch": 0.608424336973479, "grad_norm": 2.125810146331787, "learning_rate": 0.0009131936706039671, "loss": 0.4225, "step": 780 }, { "epoch": 0.6162246489859594, "grad_norm": 2.532404661178589, "learning_rate": 0.0009120793403164698, "loss": 0.3423, "step": 790 }, { "epoch": 0.62402496099844, "grad_norm": 3.6324350833892822, "learning_rate": 0.0009109650100289726, "loss": 0.4205, "step": 800 }, { "epoch": 0.6318252730109204, "grad_norm": 1.4804415702819824, "learning_rate": 0.0009098506797414754, "loss": 0.4042, "step": 810 }, { "epoch": 0.6396255850234009, "grad_norm": 1.5140562057495117, "learning_rate": 0.0009087363494539782, "loss": 0.3056, "step": 820 }, { "epoch": 0.6474258970358814, "grad_norm": 2.470576047897339, "learning_rate": 0.000907622019166481, "loss": 0.391, "step": 830 }, { "epoch": 0.6552262090483619, "grad_norm": 3.4496209621429443, "learning_rate": 0.0009065076888789837, "loss": 0.4163, "step": 840 }, { "epoch": 0.6630265210608425, "grad_norm": 1.8823250532150269, "learning_rate": 0.0009053933585914866, "loss": 0.3877, "step": 850 }, { "epoch": 0.6708268330733229, "grad_norm": 2.492297410964966, "learning_rate": 0.0009042790283039894, "loss": 0.3542, "step": 860 }, { "epoch": 0.6786271450858035, "grad_norm": 3.977569341659546, "learning_rate": 0.0009031646980164922, "loss": 0.4168, "step": 870 }, { "epoch": 0.6864274570982839, "grad_norm": 3.938462495803833, "learning_rate": 0.0009020503677289949, "loss": 0.4894, "step": 880 }, { "epoch": 0.6942277691107644, "grad_norm": 1.3457701206207275, "learning_rate": 0.0009009360374414977, "loss": 0.4903, "step": 890 }, { "epoch": 0.7020280811232449, "grad_norm": 6.3473124504089355, "learning_rate": 0.0008998217071540005, "loss": 0.4766, "step": 900 }, { "epoch": 0.7098283931357254, "grad_norm": 3.145792245864868, "learning_rate": 0.0008987073768665033, "loss": 0.4119, "step": 910 }, { "epoch": 0.717628705148206, "grad_norm": 1.809446930885315, "learning_rate": 0.0008975930465790061, "loss": 0.3991, "step": 920 }, { "epoch": 0.7254290171606864, "grad_norm": 1.7960044145584106, "learning_rate": 0.0008964787162915088, "loss": 0.3095, "step": 930 }, { "epoch": 0.733229329173167, "grad_norm": 2.9710285663604736, "learning_rate": 0.0008953643860040116, "loss": 0.5104, "step": 940 }, { "epoch": 0.7410296411856474, "grad_norm": 2.460524797439575, "learning_rate": 0.0008942500557165144, "loss": 0.4332, "step": 950 }, { "epoch": 0.748829953198128, "grad_norm": 1.6166704893112183, "learning_rate": 0.0008931357254290172, "loss": 0.3856, "step": 960 }, { "epoch": 0.7566302652106084, "grad_norm": 1.747750163078308, "learning_rate": 0.0008920213951415199, "loss": 0.3973, "step": 970 }, { "epoch": 0.7644305772230889, "grad_norm": 1.4469414949417114, "learning_rate": 0.0008909070648540227, "loss": 0.3344, "step": 980 }, { "epoch": 0.7722308892355694, "grad_norm": 4.661273956298828, "learning_rate": 0.0008897927345665256, "loss": 0.6003, "step": 990 }, { "epoch": 0.7800312012480499, "grad_norm": 3.6588950157165527, "learning_rate": 0.0008886784042790284, "loss": 0.473, "step": 1000 }, { "epoch": 0.7878315132605305, "grad_norm": 2.208383798599243, "learning_rate": 0.0008876755070202809, "loss": 0.3899, "step": 1010 }, { "epoch": 0.7956318252730109, "grad_norm": 2.3569576740264893, "learning_rate": 0.0008865611767327836, "loss": 0.3871, "step": 1020 }, { "epoch": 0.8034321372854915, "grad_norm": 2.7071454524993896, "learning_rate": 0.0008854468464452864, "loss": 0.3149, "step": 1030 }, { "epoch": 0.8112324492979719, "grad_norm": 2.8024532794952393, "learning_rate": 0.0008843325161577892, "loss": 0.3328, "step": 1040 }, { "epoch": 0.8190327613104524, "grad_norm": 3.0969290733337402, "learning_rate": 0.000883218185870292, "loss": 0.3948, "step": 1050 }, { "epoch": 0.8268330733229329, "grad_norm": 2.982484817504883, "learning_rate": 0.0008821038555827947, "loss": 0.4323, "step": 1060 }, { "epoch": 0.8346333853354134, "grad_norm": 3.133814573287964, "learning_rate": 0.0008809895252952975, "loss": 0.4393, "step": 1070 }, { "epoch": 0.8424336973478939, "grad_norm": 3.3123364448547363, "learning_rate": 0.0008798751950078003, "loss": 0.3244, "step": 1080 }, { "epoch": 0.8502340093603744, "grad_norm": 2.308555841445923, "learning_rate": 0.0008787608647203032, "loss": 0.424, "step": 1090 }, { "epoch": 0.858034321372855, "grad_norm": 3.654137134552002, "learning_rate": 0.000877646534432806, "loss": 0.3445, "step": 1100 }, { "epoch": 0.8658346333853354, "grad_norm": 2.149843692779541, "learning_rate": 0.0008765322041453087, "loss": 0.3398, "step": 1110 }, { "epoch": 0.8736349453978159, "grad_norm": 3.1334431171417236, "learning_rate": 0.0008754178738578115, "loss": 0.3333, "step": 1120 }, { "epoch": 0.8814352574102964, "grad_norm": 2.2942090034484863, "learning_rate": 0.0008743035435703143, "loss": 0.4188, "step": 1130 }, { "epoch": 0.8892355694227769, "grad_norm": 2.0195343494415283, "learning_rate": 0.0008731892132828171, "loss": 0.4047, "step": 1140 }, { "epoch": 0.8970358814352574, "grad_norm": 2.3850839138031006, "learning_rate": 0.0008720748829953198, "loss": 0.3931, "step": 1150 }, { "epoch": 0.9048361934477379, "grad_norm": 1.6200228929519653, "learning_rate": 0.0008709605527078226, "loss": 0.409, "step": 1160 }, { "epoch": 0.9126365054602185, "grad_norm": 2.9001989364624023, "learning_rate": 0.0008698462224203254, "loss": 0.4289, "step": 1170 }, { "epoch": 0.9204368174726989, "grad_norm": 1.52889883518219, "learning_rate": 0.0008687318921328282, "loss": 0.3575, "step": 1180 }, { "epoch": 0.9282371294851794, "grad_norm": 2.359733819961548, "learning_rate": 0.000867617561845331, "loss": 0.3837, "step": 1190 }, { "epoch": 0.9360374414976599, "grad_norm": 2.3807597160339355, "learning_rate": 0.0008665032315578337, "loss": 0.4206, "step": 1200 }, { "epoch": 0.9438377535101404, "grad_norm": 1.8366179466247559, "learning_rate": 0.0008653889012703365, "loss": 0.3101, "step": 1210 }, { "epoch": 0.9516380655226209, "grad_norm": 3.1048014163970947, "learning_rate": 0.0008642745709828393, "loss": 0.4104, "step": 1220 }, { "epoch": 0.9594383775351014, "grad_norm": 1.5314342975616455, "learning_rate": 0.0008631602406953422, "loss": 0.3539, "step": 1230 }, { "epoch": 0.9672386895475819, "grad_norm": 2.8501791954040527, "learning_rate": 0.0008620459104078449, "loss": 0.4104, "step": 1240 }, { "epoch": 0.9750390015600624, "grad_norm": 5.708191394805908, "learning_rate": 0.0008609315801203477, "loss": 0.4038, "step": 1250 }, { "epoch": 0.982839313572543, "grad_norm": 2.173867702484131, "learning_rate": 0.0008598172498328505, "loss": 0.338, "step": 1260 }, { "epoch": 0.9906396255850234, "grad_norm": 2.7057418823242188, "learning_rate": 0.0008587029195453533, "loss": 0.3941, "step": 1270 }, { "epoch": 0.9984399375975039, "grad_norm": 1.3492989540100098, "learning_rate": 0.0008575885892578561, "loss": 0.3696, "step": 1280 }, { "epoch": 1.0, "eval_loss": 0.4044143855571747, "eval_runtime": 635.8286, "eval_samples_per_second": 0.865, "eval_steps_per_second": 0.865, "eval_wer": 26.665379416875844, "step": 1282 }, { "epoch": 1.0062402496099845, "grad_norm": 2.5727291107177734, "learning_rate": 0.0008564742589703588, "loss": 0.3369, "step": 1290 }, { "epoch": 1.0140405616224648, "grad_norm": 2.681490182876587, "learning_rate": 0.0008553599286828616, "loss": 0.2434, "step": 1300 }, { "epoch": 1.0218408736349454, "grad_norm": 2.0296504497528076, "learning_rate": 0.0008542455983953644, "loss": 0.2122, "step": 1310 }, { "epoch": 1.029641185647426, "grad_norm": 2.277512311935425, "learning_rate": 0.0008531312681078672, "loss": 0.3932, "step": 1320 }, { "epoch": 1.0374414976599065, "grad_norm": 4.77215576171875, "learning_rate": 0.0008520169378203699, "loss": 0.2658, "step": 1330 }, { "epoch": 1.045241809672387, "grad_norm": 1.4027091264724731, "learning_rate": 0.0008509026075328727, "loss": 0.3524, "step": 1340 }, { "epoch": 1.0530421216848673, "grad_norm": 2.2849514484405518, "learning_rate": 0.0008497882772453755, "loss": 0.2705, "step": 1350 }, { "epoch": 1.0608424336973479, "grad_norm": 1.2896777391433716, "learning_rate": 0.0008486739469578784, "loss": 0.264, "step": 1360 }, { "epoch": 1.0686427457098284, "grad_norm": 2.4552128314971924, "learning_rate": 0.0008475596166703812, "loss": 0.2833, "step": 1370 }, { "epoch": 1.076443057722309, "grad_norm": 2.044693946838379, "learning_rate": 0.0008464452863828839, "loss": 0.2116, "step": 1380 }, { "epoch": 1.0842433697347893, "grad_norm": 1.3727463483810425, "learning_rate": 0.0008453309560953867, "loss": 0.2533, "step": 1390 }, { "epoch": 1.0920436817472698, "grad_norm": 1.6917822360992432, "learning_rate": 0.0008442166258078895, "loss": 0.4259, "step": 1400 }, { "epoch": 1.0998439937597504, "grad_norm": 2.198549747467041, "learning_rate": 0.0008431022955203923, "loss": 0.3161, "step": 1410 }, { "epoch": 1.107644305772231, "grad_norm": 1.7467869520187378, "learning_rate": 0.000841987965232895, "loss": 0.2377, "step": 1420 }, { "epoch": 1.1154446177847115, "grad_norm": 2.5347695350646973, "learning_rate": 0.0008408736349453978, "loss": 0.3258, "step": 1430 }, { "epoch": 1.1232449297971918, "grad_norm": 1.9081774950027466, "learning_rate": 0.0008397593046579006, "loss": 0.2462, "step": 1440 }, { "epoch": 1.1310452418096724, "grad_norm": 1.5889848470687866, "learning_rate": 0.0008386449743704034, "loss": 0.2404, "step": 1450 }, { "epoch": 1.138845553822153, "grad_norm": 1.8944768905639648, "learning_rate": 0.0008375306440829062, "loss": 0.2707, "step": 1460 }, { "epoch": 1.1466458658346335, "grad_norm": 2.5448453426361084, "learning_rate": 0.0008364163137954089, "loss": 0.3342, "step": 1470 }, { "epoch": 1.154446177847114, "grad_norm": 2.0936005115509033, "learning_rate": 0.0008353019835079117, "loss": 0.385, "step": 1480 }, { "epoch": 1.1622464898595943, "grad_norm": 2.614129066467285, "learning_rate": 0.0008341876532204145, "loss": 0.2817, "step": 1490 }, { "epoch": 1.1700468018720749, "grad_norm": 1.6156001091003418, "learning_rate": 0.0008330733229329174, "loss": 0.3527, "step": 1500 }, { "epoch": 1.1778471138845554, "grad_norm": 1.4294220209121704, "learning_rate": 0.0008319589926454201, "loss": 0.2469, "step": 1510 }, { "epoch": 1.185647425897036, "grad_norm": 3.197176456451416, "learning_rate": 0.0008308446623579229, "loss": 0.2753, "step": 1520 }, { "epoch": 1.1934477379095163, "grad_norm": 2.1629223823547363, "learning_rate": 0.0008297303320704257, "loss": 0.3951, "step": 1530 }, { "epoch": 1.2012480499219969, "grad_norm": 2.9824419021606445, "learning_rate": 0.0008286160017829285, "loss": 0.3278, "step": 1540 }, { "epoch": 1.2090483619344774, "grad_norm": 2.866138219833374, "learning_rate": 0.0008275016714954313, "loss": 0.4267, "step": 1550 }, { "epoch": 1.216848673946958, "grad_norm": 2.36781644821167, "learning_rate": 0.000826387341207934, "loss": 0.3414, "step": 1560 }, { "epoch": 1.2246489859594383, "grad_norm": 1.8305447101593018, "learning_rate": 0.0008252730109204368, "loss": 0.2925, "step": 1570 }, { "epoch": 1.2324492979719188, "grad_norm": 1.9879776239395142, "learning_rate": 0.0008241586806329396, "loss": 0.357, "step": 1580 }, { "epoch": 1.2402496099843994, "grad_norm": 2.183350086212158, "learning_rate": 0.0008230443503454424, "loss": 0.3409, "step": 1590 }, { "epoch": 1.24804992199688, "grad_norm": 2.197072744369507, "learning_rate": 0.0008219300200579451, "loss": 0.33, "step": 1600 }, { "epoch": 1.2558502340093605, "grad_norm": 3.2065696716308594, "learning_rate": 0.0008208156897704479, "loss": 0.2853, "step": 1610 }, { "epoch": 1.2636505460218408, "grad_norm": 2.0581350326538086, "learning_rate": 0.0008197013594829507, "loss": 0.3647, "step": 1620 }, { "epoch": 1.2714508580343213, "grad_norm": 3.149153232574463, "learning_rate": 0.0008185870291954536, "loss": 0.3921, "step": 1630 }, { "epoch": 1.2792511700468019, "grad_norm": 2.5097105503082275, "learning_rate": 0.0008174726989079563, "loss": 0.3302, "step": 1640 }, { "epoch": 1.2870514820592824, "grad_norm": 2.7537474632263184, "learning_rate": 0.0008163583686204591, "loss": 0.3286, "step": 1650 }, { "epoch": 1.294851794071763, "grad_norm": 1.966965675354004, "learning_rate": 0.0008152440383329619, "loss": 0.2785, "step": 1660 }, { "epoch": 1.3026521060842433, "grad_norm": 1.9159988164901733, "learning_rate": 0.0008141297080454647, "loss": 0.3299, "step": 1670 }, { "epoch": 1.3104524180967239, "grad_norm": 2.2212252616882324, "learning_rate": 0.0008130153777579675, "loss": 0.2807, "step": 1680 }, { "epoch": 1.3182527301092044, "grad_norm": 4.194318771362305, "learning_rate": 0.0008119010474704702, "loss": 0.3161, "step": 1690 }, { "epoch": 1.3260530421216847, "grad_norm": 1.7189604043960571, "learning_rate": 0.000810786717182973, "loss": 0.3356, "step": 1700 }, { "epoch": 1.3338533541341655, "grad_norm": 1.5196418762207031, "learning_rate": 0.0008096723868954758, "loss": 0.2616, "step": 1710 }, { "epoch": 1.3416536661466458, "grad_norm": 1.497450351715088, "learning_rate": 0.0008085580566079786, "loss": 0.2946, "step": 1720 }, { "epoch": 1.3494539781591264, "grad_norm": 1.74885892868042, "learning_rate": 0.0008074437263204813, "loss": 0.2801, "step": 1730 }, { "epoch": 1.357254290171607, "grad_norm": 2.040701389312744, "learning_rate": 0.0008063293960329841, "loss": 0.3203, "step": 1740 }, { "epoch": 1.3650546021840873, "grad_norm": 3.760457754135132, "learning_rate": 0.0008052150657454869, "loss": 0.4569, "step": 1750 }, { "epoch": 1.3728549141965678, "grad_norm": 2.92971134185791, "learning_rate": 0.0008041007354579897, "loss": 0.3321, "step": 1760 }, { "epoch": 1.3806552262090483, "grad_norm": 1.9461047649383545, "learning_rate": 0.0008029864051704926, "loss": 0.2696, "step": 1770 }, { "epoch": 1.388455538221529, "grad_norm": 3.2626147270202637, "learning_rate": 0.0008018720748829953, "loss": 0.3322, "step": 1780 }, { "epoch": 1.3962558502340094, "grad_norm": 2.1270642280578613, "learning_rate": 0.0008007577445954981, "loss": 0.2965, "step": 1790 }, { "epoch": 1.4040561622464898, "grad_norm": 2.3174221515655518, "learning_rate": 0.0007996434143080009, "loss": 0.3425, "step": 1800 }, { "epoch": 1.4118564742589703, "grad_norm": 2.5749576091766357, "learning_rate": 0.0007985290840205037, "loss": 0.3622, "step": 1810 }, { "epoch": 1.4196567862714509, "grad_norm": 1.873813509941101, "learning_rate": 0.0007974147537330064, "loss": 0.2497, "step": 1820 }, { "epoch": 1.4274570982839314, "grad_norm": 3.633928060531616, "learning_rate": 0.0007963004234455092, "loss": 0.4058, "step": 1830 }, { "epoch": 1.435257410296412, "grad_norm": 2.356269598007202, "learning_rate": 0.000795186093158012, "loss": 0.2635, "step": 1840 }, { "epoch": 1.4430577223088923, "grad_norm": 1.9108752012252808, "learning_rate": 0.0007940717628705148, "loss": 0.336, "step": 1850 }, { "epoch": 1.4508580343213728, "grad_norm": 1.5505330562591553, "learning_rate": 0.0007929574325830176, "loss": 0.3493, "step": 1860 }, { "epoch": 1.4586583463338534, "grad_norm": 1.9970422983169556, "learning_rate": 0.0007918431022955203, "loss": 0.273, "step": 1870 }, { "epoch": 1.466458658346334, "grad_norm": 2.753758192062378, "learning_rate": 0.0007907287720080231, "loss": 0.2845, "step": 1880 }, { "epoch": 1.4742589703588145, "grad_norm": NaN, "learning_rate": 0.0007897258747492757, "loss": 0.2544, "step": 1890 }, { "epoch": 1.4820592823712948, "grad_norm": 3.0995099544525146, "learning_rate": 0.0007886115444617785, "loss": 0.2884, "step": 1900 }, { "epoch": 1.4898595943837754, "grad_norm": 5.728559970855713, "learning_rate": 0.0007874972141742812, "loss": 0.2681, "step": 1910 }, { "epoch": 1.497659906396256, "grad_norm": 1.492622971534729, "learning_rate": 0.000786382883886784, "loss": 0.2891, "step": 1920 }, { "epoch": 1.5054602184087362, "grad_norm": 1.7419252395629883, "learning_rate": 0.0007852685535992868, "loss": 0.4089, "step": 1930 }, { "epoch": 1.513260530421217, "grad_norm": 6.814690589904785, "learning_rate": 0.0007841542233117896, "loss": 0.3372, "step": 1940 }, { "epoch": 1.5210608424336973, "grad_norm": 2.380725860595703, "learning_rate": 0.0007830398930242924, "loss": 0.3189, "step": 1950 }, { "epoch": 1.5288611544461779, "grad_norm": 5.004116058349609, "learning_rate": 0.0007819255627367951, "loss": 0.3018, "step": 1960 }, { "epoch": 1.5366614664586584, "grad_norm": 2.604365825653076, "learning_rate": 0.0007808112324492979, "loss": 0.3054, "step": 1970 }, { "epoch": 1.5444617784711387, "grad_norm": 1.585584044456482, "learning_rate": 0.0007796969021618007, "loss": 0.3477, "step": 1980 }, { "epoch": 1.5522620904836193, "grad_norm": 1.8678693771362305, "learning_rate": 0.0007785825718743035, "loss": 0.3577, "step": 1990 }, { "epoch": 1.5600624024960998, "grad_norm": 1.654689073562622, "learning_rate": 0.0007774682415868062, "loss": 0.2625, "step": 2000 }, { "epoch": 1.5678627145085804, "grad_norm": 2.108919858932495, "learning_rate": 0.0007763539112993092, "loss": 0.2497, "step": 2010 }, { "epoch": 1.575663026521061, "grad_norm": 7.198604106903076, "learning_rate": 0.000775239581011812, "loss": 0.3382, "step": 2020 }, { "epoch": 1.5834633385335413, "grad_norm": 2.2285892963409424, "learning_rate": 0.0007741252507243148, "loss": 0.2598, "step": 2030 }, { "epoch": 1.5912636505460218, "grad_norm": 1.7743014097213745, "learning_rate": 0.0007730109204368176, "loss": 0.2757, "step": 2040 }, { "epoch": 1.5990639625585024, "grad_norm": 1.7763789892196655, "learning_rate": 0.0007718965901493203, "loss": 0.2703, "step": 2050 }, { "epoch": 1.6068642745709827, "grad_norm": 2.159956693649292, "learning_rate": 0.000770782259861823, "loss": 0.2824, "step": 2060 }, { "epoch": 1.6146645865834635, "grad_norm": 1.4845560789108276, "learning_rate": 0.0007696679295743259, "loss": 0.2528, "step": 2070 }, { "epoch": 1.6224648985959438, "grad_norm": 3.627887010574341, "learning_rate": 0.0007685535992868287, "loss": 0.3197, "step": 2080 }, { "epoch": 1.6302652106084243, "grad_norm": 2.2174973487854004, "learning_rate": 0.0007674392689993314, "loss": 0.2994, "step": 2090 }, { "epoch": 1.6380655226209049, "grad_norm": 2.5977325439453125, "learning_rate": 0.0007663249387118342, "loss": 0.2991, "step": 2100 }, { "epoch": 1.6458658346333852, "grad_norm": 1.9066824913024902, "learning_rate": 0.000765210608424337, "loss": 0.2166, "step": 2110 }, { "epoch": 1.653666146645866, "grad_norm": 1.7197297811508179, "learning_rate": 0.0007640962781368398, "loss": 0.2948, "step": 2120 }, { "epoch": 1.6614664586583463, "grad_norm": 2.054304361343384, "learning_rate": 0.0007629819478493426, "loss": 0.3073, "step": 2130 }, { "epoch": 1.6692667706708268, "grad_norm": 1.7934963703155518, "learning_rate": 0.0007618676175618453, "loss": 0.2667, "step": 2140 }, { "epoch": 1.6770670826833074, "grad_norm": 2.5259838104248047, "learning_rate": 0.0007607532872743482, "loss": 0.3322, "step": 2150 }, { "epoch": 1.6848673946957877, "grad_norm": 3.6354122161865234, "learning_rate": 0.000759638956986851, "loss": 0.3909, "step": 2160 }, { "epoch": 1.6926677067082685, "grad_norm": 1.6722809076309204, "learning_rate": 0.0007585246266993538, "loss": 0.2987, "step": 2170 }, { "epoch": 1.7004680187207488, "grad_norm": 4.3235015869140625, "learning_rate": 0.0007574102964118565, "loss": 0.3112, "step": 2180 }, { "epoch": 1.7082683307332294, "grad_norm": 2.236316442489624, "learning_rate": 0.0007562959661243593, "loss": 0.3174, "step": 2190 }, { "epoch": 1.71606864274571, "grad_norm": 21.32891273498535, "learning_rate": 0.0007551816358368621, "loss": 0.3696, "step": 2200 }, { "epoch": 1.7238689547581902, "grad_norm": 2.4251410961151123, "learning_rate": 0.0007540673055493649, "loss": 0.3171, "step": 2210 }, { "epoch": 1.7316692667706708, "grad_norm": 2.4152424335479736, "learning_rate": 0.0007529529752618677, "loss": 0.328, "step": 2220 }, { "epoch": 1.7394695787831513, "grad_norm": 2.0988574028015137, "learning_rate": 0.0007518386449743704, "loss": 0.3061, "step": 2230 }, { "epoch": 1.7472698907956317, "grad_norm": 4.469291687011719, "learning_rate": 0.0007507243146868732, "loss": 0.3315, "step": 2240 }, { "epoch": 1.7550702028081124, "grad_norm": 2.4917778968811035, "learning_rate": 0.000749609984399376, "loss": 0.4296, "step": 2250 }, { "epoch": 1.7628705148205928, "grad_norm": 3.073840379714966, "learning_rate": 0.0007484956541118788, "loss": 0.3685, "step": 2260 }, { "epoch": 1.7706708268330733, "grad_norm": 2.205733299255371, "learning_rate": 0.0007473813238243815, "loss": 0.2691, "step": 2270 }, { "epoch": 1.7784711388455539, "grad_norm": 2.3948941230773926, "learning_rate": 0.0007462669935368844, "loss": 0.3065, "step": 2280 }, { "epoch": 1.7862714508580342, "grad_norm": 2.6060824394226074, "learning_rate": 0.0007451526632493872, "loss": 0.3354, "step": 2290 }, { "epoch": 1.794071762870515, "grad_norm": 3.2586774826049805, "learning_rate": 0.00074403833296189, "loss": 0.261, "step": 2300 }, { "epoch": 1.8018720748829953, "grad_norm": 1.6417285203933716, "learning_rate": 0.0007429240026743928, "loss": 0.3828, "step": 2310 }, { "epoch": 1.8096723868954758, "grad_norm": 4.006927967071533, "learning_rate": 0.0007418096723868955, "loss": 0.4228, "step": 2320 }, { "epoch": 1.8174726989079564, "grad_norm": 2.5880374908447266, "learning_rate": 0.0007406953420993983, "loss": 0.2766, "step": 2330 }, { "epoch": 1.8252730109204367, "grad_norm": 1.602337121963501, "learning_rate": 0.0007395810118119011, "loss": 0.2794, "step": 2340 }, { "epoch": 1.8330733229329175, "grad_norm": 5.932153224945068, "learning_rate": 0.0007384666815244039, "loss": 0.3471, "step": 2350 }, { "epoch": 1.8408736349453978, "grad_norm": 4.076808452606201, "learning_rate": 0.0007373523512369066, "loss": 0.2649, "step": 2360 }, { "epoch": 1.8486739469578783, "grad_norm": 4.666397571563721, "learning_rate": 0.0007362380209494094, "loss": 0.2751, "step": 2370 }, { "epoch": 1.856474258970359, "grad_norm": 3.792745590209961, "learning_rate": 0.0007351236906619122, "loss": 0.2606, "step": 2380 }, { "epoch": 1.8642745709828392, "grad_norm": 2.5275423526763916, "learning_rate": 0.000734009360374415, "loss": 0.3469, "step": 2390 }, { "epoch": 1.8720748829953198, "grad_norm": 1.59649658203125, "learning_rate": 0.0007328950300869178, "loss": 0.2724, "step": 2400 }, { "epoch": 1.8798751950078003, "grad_norm": 3.7428267002105713, "learning_rate": 0.0007317806997994206, "loss": 0.4426, "step": 2410 }, { "epoch": 1.8876755070202809, "grad_norm": 3.7439956665039062, "learning_rate": 0.0007306663695119234, "loss": 0.2901, "step": 2420 }, { "epoch": 1.8954758190327614, "grad_norm": 2.3777198791503906, "learning_rate": 0.0007295520392244262, "loss": 0.2837, "step": 2430 }, { "epoch": 1.9032761310452417, "grad_norm": 2.7654988765716553, "learning_rate": 0.000728437708936929, "loss": 0.277, "step": 2440 }, { "epoch": 1.9110764430577223, "grad_norm": 1.8758680820465088, "learning_rate": 0.0007273233786494317, "loss": 0.2706, "step": 2450 }, { "epoch": 1.9188767550702028, "grad_norm": 2.8725340366363525, "learning_rate": 0.0007262090483619345, "loss": 0.2566, "step": 2460 }, { "epoch": 1.9266770670826832, "grad_norm": 2.4021458625793457, "learning_rate": 0.0007250947180744373, "loss": 0.2645, "step": 2470 }, { "epoch": 1.934477379095164, "grad_norm": 2.8407838344573975, "learning_rate": 0.0007239803877869401, "loss": 0.3152, "step": 2480 }, { "epoch": 1.9422776911076443, "grad_norm": 3.606403112411499, "learning_rate": 0.0007228660574994429, "loss": 0.373, "step": 2490 }, { "epoch": 1.9500780031201248, "grad_norm": 2.362473487854004, "learning_rate": 0.0007217517272119456, "loss": 0.4332, "step": 2500 }, { "epoch": 1.9578783151326054, "grad_norm": 1.9711815118789673, "learning_rate": 0.0007206373969244484, "loss": 0.2606, "step": 2510 }, { "epoch": 1.9656786271450857, "grad_norm": 2.683908224105835, "learning_rate": 0.0007195230666369512, "loss": 0.3582, "step": 2520 }, { "epoch": 1.9734789391575664, "grad_norm": 2.5902493000030518, "learning_rate": 0.000718408736349454, "loss": 0.2694, "step": 2530 }, { "epoch": 1.9812792511700468, "grad_norm": 3.92708420753479, "learning_rate": 0.0007172944060619567, "loss": 0.3083, "step": 2540 }, { "epoch": 1.9890795631825273, "grad_norm": 2.6788370609283447, "learning_rate": 0.0007161800757744596, "loss": 0.2587, "step": 2550 }, { "epoch": 1.9968798751950079, "grad_norm": 2.413313627243042, "learning_rate": 0.0007150657454869624, "loss": 0.2732, "step": 2560 }, { "epoch": 2.0, "eval_loss": 0.3617618680000305, "eval_runtime": 36952.264, "eval_samples_per_second": 0.015, "eval_steps_per_second": 0.015, "eval_wer": 22.793975670978952, "step": 2564 }, { "epoch": 2.004680187207488, "grad_norm": 1.572451114654541, "learning_rate": 0.0007139514151994652, "loss": 0.2147, "step": 2570 }, { "epoch": 2.012480499219969, "grad_norm": 2.4759023189544678, "learning_rate": 0.000712837084911968, "loss": 0.2015, "step": 2580 }, { "epoch": 2.0202808112324493, "grad_norm": 1.6903563737869263, "learning_rate": 0.0007117227546244707, "loss": 0.1703, "step": 2590 }, { "epoch": 2.0280811232449296, "grad_norm": 3.494985580444336, "learning_rate": 0.0007106084243369735, "loss": 0.2063, "step": 2600 }, { "epoch": 2.0358814352574104, "grad_norm": 1.5024439096450806, "learning_rate": 0.0007094940940494763, "loss": 0.1807, "step": 2610 }, { "epoch": 2.0436817472698907, "grad_norm": 1.9423105716705322, "learning_rate": 0.0007083797637619791, "loss": 0.1953, "step": 2620 }, { "epoch": 2.0514820592823715, "grad_norm": 0.8572360277175903, "learning_rate": 0.0007072654334744818, "loss": 0.3423, "step": 2630 }, { "epoch": 2.059282371294852, "grad_norm": 2.574855327606201, "learning_rate": 0.0007061511031869846, "loss": 0.1916, "step": 2640 }, { "epoch": 2.067082683307332, "grad_norm": 2.2941219806671143, "learning_rate": 0.0007050367728994874, "loss": 0.2168, "step": 2650 }, { "epoch": 2.074882995319813, "grad_norm": 2.0448150634765625, "learning_rate": 0.0007039224426119902, "loss": 0.2412, "step": 2660 }, { "epoch": 2.0826833073322932, "grad_norm": 1.8044381141662598, "learning_rate": 0.000702808112324493, "loss": 0.3261, "step": 2670 }, { "epoch": 2.090483619344774, "grad_norm": 2.752220630645752, "learning_rate": 0.0007016937820369958, "loss": 0.2007, "step": 2680 }, { "epoch": 2.0982839313572543, "grad_norm": 1.4598050117492676, "learning_rate": 0.0007005794517494986, "loss": 0.1856, "step": 2690 }, { "epoch": 2.1060842433697347, "grad_norm": 3.578192710876465, "learning_rate": 0.0006994651214620014, "loss": 0.1991, "step": 2700 }, { "epoch": 2.1138845553822154, "grad_norm": 2.6971054077148438, "learning_rate": 0.0006983507911745042, "loss": 0.2134, "step": 2710 }, { "epoch": 2.1216848673946958, "grad_norm": 2.437596559524536, "learning_rate": 0.0006972364608870069, "loss": 0.228, "step": 2720 }, { "epoch": 2.129485179407176, "grad_norm": 2.6254658699035645, "learning_rate": 0.0006961221305995097, "loss": 0.1671, "step": 2730 }, { "epoch": 2.137285491419657, "grad_norm": 1.3765720129013062, "learning_rate": 0.0006950078003120125, "loss": 0.2391, "step": 2740 }, { "epoch": 2.145085803432137, "grad_norm": 2.192396879196167, "learning_rate": 0.0006938934700245153, "loss": 0.2331, "step": 2750 }, { "epoch": 2.152886115444618, "grad_norm": 1.4418809413909912, "learning_rate": 0.0006927791397370181, "loss": 0.2041, "step": 2760 }, { "epoch": 2.1606864274570983, "grad_norm": 2.553459882736206, "learning_rate": 0.0006916648094495208, "loss": 0.2342, "step": 2770 }, { "epoch": 2.1684867394695786, "grad_norm": 1.7199114561080933, "learning_rate": 0.0006905504791620236, "loss": 0.1841, "step": 2780 }, { "epoch": 2.1762870514820594, "grad_norm": 2.2145979404449463, "learning_rate": 0.0006894361488745264, "loss": 0.1821, "step": 2790 }, { "epoch": 2.1840873634945397, "grad_norm": 2.434779405593872, "learning_rate": 0.0006883218185870292, "loss": 0.2115, "step": 2800 }, { "epoch": 2.1918876755070205, "grad_norm": 1.0518474578857422, "learning_rate": 0.0006872074882995319, "loss": 0.2186, "step": 2810 }, { "epoch": 2.199687987519501, "grad_norm": 1.9856785535812378, "learning_rate": 0.0006860931580120348, "loss": 0.2448, "step": 2820 }, { "epoch": 2.207488299531981, "grad_norm": 17.3148136138916, "learning_rate": 0.0006849788277245376, "loss": 0.2372, "step": 2830 }, { "epoch": 2.215288611544462, "grad_norm": 3.4590959548950195, "learning_rate": 0.0006838644974370404, "loss": 0.3179, "step": 2840 }, { "epoch": 2.223088923556942, "grad_norm": 2.127650260925293, "learning_rate": 0.0006827501671495432, "loss": 0.2445, "step": 2850 }, { "epoch": 2.230889235569423, "grad_norm": 1.895729660987854, "learning_rate": 0.0006816358368620459, "loss": 0.1966, "step": 2860 }, { "epoch": 2.2386895475819033, "grad_norm": 1.4693467617034912, "learning_rate": 0.0006805215065745487, "loss": 0.2351, "step": 2870 }, { "epoch": 2.2464898595943836, "grad_norm": 1.500453233718872, "learning_rate": 0.0006794071762870515, "loss": 0.1946, "step": 2880 }, { "epoch": 2.2542901716068644, "grad_norm": 2.8631374835968018, "learning_rate": 0.0006782928459995543, "loss": 0.2367, "step": 2890 }, { "epoch": 2.2620904836193447, "grad_norm": 1.5281766653060913, "learning_rate": 0.000677178515712057, "loss": 0.1855, "step": 2900 }, { "epoch": 2.2698907956318255, "grad_norm": 2.3027195930480957, "learning_rate": 0.0006760641854245598, "loss": 0.2117, "step": 2910 }, { "epoch": 2.277691107644306, "grad_norm": 2.9962668418884277, "learning_rate": 0.0006749498551370626, "loss": 0.2441, "step": 2920 }, { "epoch": 2.285491419656786, "grad_norm": 2.489192008972168, "learning_rate": 0.0006738355248495654, "loss": 0.2212, "step": 2930 }, { "epoch": 2.293291731669267, "grad_norm": 1.366945743560791, "learning_rate": 0.0006727211945620682, "loss": 0.2003, "step": 2940 }, { "epoch": 2.3010920436817472, "grad_norm": 1.4768056869506836, "learning_rate": 0.000671606864274571, "loss": 0.3107, "step": 2950 }, { "epoch": 2.308892355694228, "grad_norm": 1.797979474067688, "learning_rate": 0.0006704925339870738, "loss": 0.1978, "step": 2960 }, { "epoch": 2.3166926677067083, "grad_norm": 1.9439573287963867, "learning_rate": 0.0006693782036995766, "loss": 0.2306, "step": 2970 }, { "epoch": 2.3244929797191887, "grad_norm": 1.582478642463684, "learning_rate": 0.0006682638734120794, "loss": 0.2079, "step": 2980 }, { "epoch": 2.3322932917316694, "grad_norm": 2.1245157718658447, "learning_rate": 0.0006671495431245821, "loss": 0.2258, "step": 2990 }, { "epoch": 2.3400936037441498, "grad_norm": 2.427675724029541, "learning_rate": 0.0006660352128370849, "loss": 0.2846, "step": 3000 }, { "epoch": 2.34789391575663, "grad_norm": 0.9095527529716492, "learning_rate": 0.0006649208825495877, "loss": 0.1746, "step": 3010 }, { "epoch": 2.355694227769111, "grad_norm": 2.0468294620513916, "learning_rate": 0.0006638065522620905, "loss": 0.2645, "step": 3020 }, { "epoch": 2.363494539781591, "grad_norm": 1.5703836679458618, "learning_rate": 0.0006626922219745933, "loss": 0.1896, "step": 3030 }, { "epoch": 2.371294851794072, "grad_norm": 2.37263822555542, "learning_rate": 0.000661577891687096, "loss": 0.232, "step": 3040 }, { "epoch": 2.3790951638065523, "grad_norm": 1.6041433811187744, "learning_rate": 0.0006604635613995988, "loss": 0.2135, "step": 3050 }, { "epoch": 2.3868954758190326, "grad_norm": 1.87883722782135, "learning_rate": 0.0006593492311121016, "loss": 0.1993, "step": 3060 }, { "epoch": 2.3946957878315134, "grad_norm": 2.5502099990844727, "learning_rate": 0.0006582349008246044, "loss": 0.2468, "step": 3070 }, { "epoch": 2.4024960998439937, "grad_norm": 2.681384801864624, "learning_rate": 0.0006571205705371071, "loss": 0.2364, "step": 3080 }, { "epoch": 2.410296411856474, "grad_norm": 1.2032707929611206, "learning_rate": 0.00065600624024961, "loss": 0.2732, "step": 3090 }, { "epoch": 2.418096723868955, "grad_norm": 1.553661584854126, "learning_rate": 0.0006548919099621128, "loss": 0.2214, "step": 3100 }, { "epoch": 2.425897035881435, "grad_norm": 0.9736389517784119, "learning_rate": 0.0006537775796746156, "loss": 0.1778, "step": 3110 }, { "epoch": 2.433697347893916, "grad_norm": 1.911352276802063, "learning_rate": 0.0006526632493871184, "loss": 0.2062, "step": 3120 }, { "epoch": 2.4414976599063962, "grad_norm": 1.4338595867156982, "learning_rate": 0.0006515489190996211, "loss": 0.193, "step": 3130 }, { "epoch": 2.4492979719188765, "grad_norm": 1.3027153015136719, "learning_rate": 0.0006504345888121239, "loss": 0.3175, "step": 3140 }, { "epoch": 2.4570982839313573, "grad_norm": 2.262709140777588, "learning_rate": 0.0006493202585246267, "loss": 0.2874, "step": 3150 }, { "epoch": 2.4648985959438376, "grad_norm": 1.36016845703125, "learning_rate": 0.0006482059282371295, "loss": 0.1603, "step": 3160 }, { "epoch": 2.4726989079563184, "grad_norm": 3.441779613494873, "learning_rate": 0.0006470915979496322, "loss": 0.4564, "step": 3170 }, { "epoch": 2.4804992199687987, "grad_norm": 1.8022549152374268, "learning_rate": 0.000645977267662135, "loss": 0.1928, "step": 3180 }, { "epoch": 2.488299531981279, "grad_norm": 2.104497194290161, "learning_rate": 0.0006448629373746378, "loss": 0.2047, "step": 3190 }, { "epoch": 2.49609984399376, "grad_norm": 2.5888454914093018, "learning_rate": 0.0006437486070871406, "loss": 0.2591, "step": 3200 }, { "epoch": 2.50390015600624, "grad_norm": 1.3476183414459229, "learning_rate": 0.0006426342767996434, "loss": 0.2636, "step": 3210 }, { "epoch": 2.511700468018721, "grad_norm": 2.888965606689453, "learning_rate": 0.0006415199465121462, "loss": 0.216, "step": 3220 }, { "epoch": 2.5195007800312013, "grad_norm": 1.8190357685089111, "learning_rate": 0.000640405616224649, "loss": 0.3498, "step": 3230 }, { "epoch": 2.5273010920436816, "grad_norm": 1.5132287740707397, "learning_rate": 0.0006392912859371518, "loss": 0.227, "step": 3240 }, { "epoch": 2.5351014040561624, "grad_norm": 2.8061325550079346, "learning_rate": 0.0006381769556496546, "loss": 0.2089, "step": 3250 }, { "epoch": 2.5429017160686427, "grad_norm": 2.2521767616271973, "learning_rate": 0.0006370626253621573, "loss": 0.2362, "step": 3260 }, { "epoch": 2.5507020280811235, "grad_norm": 1.4635623693466187, "learning_rate": 0.0006359482950746601, "loss": 0.1901, "step": 3270 }, { "epoch": 2.5585023400936038, "grad_norm": 1.9882755279541016, "learning_rate": 0.0006348339647871629, "loss": 0.2387, "step": 3280 }, { "epoch": 2.566302652106084, "grad_norm": 6.737502098083496, "learning_rate": 0.0006337196344996657, "loss": 0.2278, "step": 3290 }, { "epoch": 2.574102964118565, "grad_norm": 1.686926245689392, "learning_rate": 0.0006326053042121685, "loss": 0.3118, "step": 3300 }, { "epoch": 2.581903276131045, "grad_norm": 1.436584234237671, "learning_rate": 0.0006314909739246712, "loss": 0.1787, "step": 3310 }, { "epoch": 2.589703588143526, "grad_norm": 3.646476984024048, "learning_rate": 0.000630376643637174, "loss": 0.2165, "step": 3320 }, { "epoch": 2.5975039001560063, "grad_norm": 1.8787113428115845, "learning_rate": 0.0006292623133496768, "loss": 0.1993, "step": 3330 }, { "epoch": 2.6053042121684866, "grad_norm": 2.4485440254211426, "learning_rate": 0.0006281479830621796, "loss": 0.24, "step": 3340 }, { "epoch": 2.6131045241809674, "grad_norm": 2.73197865486145, "learning_rate": 0.0006270336527746824, "loss": 0.228, "step": 3350 }, { "epoch": 2.6209048361934477, "grad_norm": 3.2943315505981445, "learning_rate": 0.0006259193224871852, "loss": 0.2289, "step": 3360 }, { "epoch": 2.6287051482059285, "grad_norm": 2.165308952331543, "learning_rate": 0.000624804992199688, "loss": 0.2435, "step": 3370 }, { "epoch": 2.636505460218409, "grad_norm": 2.3766629695892334, "learning_rate": 0.0006236906619121908, "loss": 0.26, "step": 3380 }, { "epoch": 2.644305772230889, "grad_norm": 1.1527057886123657, "learning_rate": 0.0006225763316246936, "loss": 0.2486, "step": 3390 }, { "epoch": 2.6521060842433695, "grad_norm": 2.6304874420166016, "learning_rate": 0.0006214620013371963, "loss": 0.3827, "step": 3400 }, { "epoch": 2.6599063962558502, "grad_norm": 1.5219537019729614, "learning_rate": 0.0006203476710496991, "loss": 0.2231, "step": 3410 }, { "epoch": 2.667706708268331, "grad_norm": 1.9267528057098389, "learning_rate": 0.0006192333407622019, "loss": 0.2468, "step": 3420 }, { "epoch": 2.6755070202808113, "grad_norm": 2.247861385345459, "learning_rate": 0.0006181190104747047, "loss": 0.2463, "step": 3430 }, { "epoch": 2.6833073322932917, "grad_norm": 2.6133053302764893, "learning_rate": 0.0006170046801872074, "loss": 0.2636, "step": 3440 }, { "epoch": 2.691107644305772, "grad_norm": 1.4675954580307007, "learning_rate": 0.0006158903498997102, "loss": 0.2141, "step": 3450 }, { "epoch": 2.6989079563182528, "grad_norm": 3.3972527980804443, "learning_rate": 0.000614776019612213, "loss": 0.2102, "step": 3460 }, { "epoch": 2.706708268330733, "grad_norm": 2.0020527839660645, "learning_rate": 0.0006136616893247158, "loss": 0.2618, "step": 3470 }, { "epoch": 2.714508580343214, "grad_norm": 1.4799424409866333, "learning_rate": 0.0006125473590372186, "loss": 0.2644, "step": 3480 }, { "epoch": 2.722308892355694, "grad_norm": 2.207921028137207, "learning_rate": 0.0006114330287497214, "loss": 0.2496, "step": 3490 }, { "epoch": 2.7301092043681745, "grad_norm": 2.7361607551574707, "learning_rate": 0.0006103186984622242, "loss": 0.1917, "step": 3500 }, { "epoch": 2.7379095163806553, "grad_norm": 2.6031532287597656, "learning_rate": 0.000609204368174727, "loss": 0.2358, "step": 3510 }, { "epoch": 2.7457098283931356, "grad_norm": 1.442651629447937, "learning_rate": 0.0006080900378872299, "loss": 0.2929, "step": 3520 }, { "epoch": 2.7535101404056164, "grad_norm": 2.853076457977295, "learning_rate": 0.0006069757075997325, "loss": 0.2287, "step": 3530 }, { "epoch": 2.7613104524180967, "grad_norm": 1.827863335609436, "learning_rate": 0.0006058613773122353, "loss": 0.284, "step": 3540 }, { "epoch": 2.769110764430577, "grad_norm": 1.29642915725708, "learning_rate": 0.0006047470470247382, "loss": 0.2156, "step": 3550 }, { "epoch": 2.776911076443058, "grad_norm": 2.0758543014526367, "learning_rate": 0.000603632716737241, "loss": 0.4312, "step": 3560 }, { "epoch": 2.784711388455538, "grad_norm": 2.1832942962646484, "learning_rate": 0.0006025183864497438, "loss": 0.22, "step": 3570 }, { "epoch": 2.792511700468019, "grad_norm": 1.541040301322937, "learning_rate": 0.0006014040561622464, "loss": 0.2452, "step": 3580 }, { "epoch": 2.800312012480499, "grad_norm": 1.5947539806365967, "learning_rate": 0.0006002897258747492, "loss": 0.2708, "step": 3590 }, { "epoch": 2.8081123244929795, "grad_norm": 2.130390167236328, "learning_rate": 0.000599175395587252, "loss": 0.2194, "step": 3600 }, { "epoch": 2.8159126365054603, "grad_norm": 2.382166862487793, "learning_rate": 0.0005980610652997549, "loss": 0.1931, "step": 3610 }, { "epoch": 2.8237129485179406, "grad_norm": 4.421852111816406, "learning_rate": 0.0005969467350122577, "loss": 0.3081, "step": 3620 }, { "epoch": 2.8315132605304214, "grad_norm": 1.1889768838882446, "learning_rate": 0.0005958324047247605, "loss": 0.2134, "step": 3630 }, { "epoch": 2.8393135725429017, "grad_norm": 2.6204874515533447, "learning_rate": 0.0005947180744372633, "loss": 0.2128, "step": 3640 }, { "epoch": 2.847113884555382, "grad_norm": 1.4705913066864014, "learning_rate": 0.0005936037441497661, "loss": 0.1894, "step": 3650 }, { "epoch": 2.854914196567863, "grad_norm": 3.110135555267334, "learning_rate": 0.0005924894138622688, "loss": 0.2072, "step": 3660 }, { "epoch": 2.862714508580343, "grad_norm": 1.3255491256713867, "learning_rate": 0.0005913750835747716, "loss": 0.2325, "step": 3670 }, { "epoch": 2.870514820592824, "grad_norm": 2.2520713806152344, "learning_rate": 0.0005902607532872744, "loss": 0.2121, "step": 3680 }, { "epoch": 2.8783151326053042, "grad_norm": 1.4630913734436035, "learning_rate": 0.0005891464229997772, "loss": 0.2281, "step": 3690 }, { "epoch": 2.8861154446177846, "grad_norm": 2.0491292476654053, "learning_rate": 0.00058803209271228, "loss": 0.2227, "step": 3700 }, { "epoch": 2.8939157566302653, "grad_norm": 6.133053302764893, "learning_rate": 0.0005869177624247827, "loss": 0.3106, "step": 3710 }, { "epoch": 2.9017160686427457, "grad_norm": 2.3226382732391357, "learning_rate": 0.0005858034321372855, "loss": 0.2193, "step": 3720 }, { "epoch": 2.9095163806552264, "grad_norm": 1.351330041885376, "learning_rate": 0.0005846891018497883, "loss": 0.1928, "step": 3730 }, { "epoch": 2.9173166926677068, "grad_norm": 0.9889002442359924, "learning_rate": 0.0005835747715622911, "loss": 0.2395, "step": 3740 }, { "epoch": 2.925117004680187, "grad_norm": 3.9231808185577393, "learning_rate": 0.0005824604412747938, "loss": 0.2288, "step": 3750 }, { "epoch": 2.932917316692668, "grad_norm": 1.1299773454666138, "learning_rate": 0.0005813461109872967, "loss": 0.1996, "step": 3760 }, { "epoch": 2.940717628705148, "grad_norm": 1.894411563873291, "learning_rate": 0.0005802317806997995, "loss": 0.2444, "step": 3770 }, { "epoch": 2.948517940717629, "grad_norm": 3.100918769836426, "learning_rate": 0.0005791174504123023, "loss": 0.2723, "step": 3780 }, { "epoch": 2.9563182527301093, "grad_norm": 2.921398639678955, "learning_rate": 0.0005780031201248051, "loss": 0.1963, "step": 3790 }, { "epoch": 2.9641185647425896, "grad_norm": 2.898193120956421, "learning_rate": 0.0005768887898373078, "loss": 0.2262, "step": 3800 }, { "epoch": 2.97191887675507, "grad_norm": 1.8510135412216187, "learning_rate": 0.0005757744595498106, "loss": 0.2253, "step": 3810 }, { "epoch": 2.9797191887675507, "grad_norm": 2.716972827911377, "learning_rate": 0.0005746601292623134, "loss": 0.222, "step": 3820 }, { "epoch": 2.9875195007800315, "grad_norm": 2.466111660003662, "learning_rate": 0.0005735457989748162, "loss": 0.2903, "step": 3830 }, { "epoch": 2.995319812792512, "grad_norm": 2.409747838973999, "learning_rate": 0.0005724314686873189, "loss": 0.3088, "step": 3840 }, { "epoch": 3.0, "eval_loss": 0.34428906440734863, "eval_runtime": 557.5413, "eval_samples_per_second": 0.986, "eval_steps_per_second": 0.986, "eval_wer": 24.454527901139215, "step": 3846 }, { "epoch": 3.003120124804992, "grad_norm": 1.5137571096420288, "learning_rate": 0.0005713171383998217, "loss": 0.197, "step": 3850 }, { "epoch": 3.010920436817473, "grad_norm": 0.7653738856315613, "learning_rate": 0.0005702028081123245, "loss": 0.1295, "step": 3860 }, { "epoch": 3.0187207488299532, "grad_norm": 1.05947744846344, "learning_rate": 0.0005690884778248273, "loss": 0.1684, "step": 3870 }, { "epoch": 3.0265210608424336, "grad_norm": 1.4757158756256104, "learning_rate": 0.0005679741475373301, "loss": 0.1379, "step": 3880 }, { "epoch": 3.0343213728549143, "grad_norm": 1.789844274520874, "learning_rate": 0.0005668598172498329, "loss": 0.1711, "step": 3890 }, { "epoch": 3.0421216848673946, "grad_norm": 1.6123414039611816, "learning_rate": 0.0005657454869623357, "loss": 0.1375, "step": 3900 }, { "epoch": 3.049921996879875, "grad_norm": 1.1085964441299438, "learning_rate": 0.0005646311566748385, "loss": 0.1552, "step": 3910 }, { "epoch": 3.0577223088923557, "grad_norm": 0.5497006773948669, "learning_rate": 0.0005635168263873413, "loss": 0.2185, "step": 3920 }, { "epoch": 3.065522620904836, "grad_norm": 1.584539532661438, "learning_rate": 0.000562402496099844, "loss": 0.1398, "step": 3930 }, { "epoch": 3.073322932917317, "grad_norm": 1.2666114568710327, "learning_rate": 0.0005612881658123468, "loss": 0.1372, "step": 3940 }, { "epoch": 3.081123244929797, "grad_norm": 2.3011374473571777, "learning_rate": 0.0005601738355248496, "loss": 0.1354, "step": 3950 }, { "epoch": 3.0889235569422775, "grad_norm": 1.05606210231781, "learning_rate": 0.0005590595052373524, "loss": 0.1158, "step": 3960 }, { "epoch": 3.0967238689547583, "grad_norm": 1.0332001447677612, "learning_rate": 0.0005579451749498552, "loss": 0.1333, "step": 3970 }, { "epoch": 3.1045241809672386, "grad_norm": 1.020912766456604, "learning_rate": 0.0005568308446623579, "loss": 0.1259, "step": 3980 }, { "epoch": 3.1123244929797194, "grad_norm": 1.2610992193222046, "learning_rate": 0.0005557165143748607, "loss": 0.1391, "step": 3990 }, { "epoch": 3.1201248049921997, "grad_norm": 1.1720095872879028, "learning_rate": 0.0005546021840873635, "loss": 0.188, "step": 4000 }, { "epoch": 3.12792511700468, "grad_norm": 3.7483229637145996, "learning_rate": 0.0005534878537998663, "loss": 0.1198, "step": 4010 }, { "epoch": 3.135725429017161, "grad_norm": 0.9843854308128357, "learning_rate": 0.000552373523512369, "loss": 0.1405, "step": 4020 }, { "epoch": 3.143525741029641, "grad_norm": 2.162191152572632, "learning_rate": 0.0005512591932248719, "loss": 0.1268, "step": 4030 }, { "epoch": 3.151326053042122, "grad_norm": 1.1201051473617554, "learning_rate": 0.0005501448629373747, "loss": 0.1354, "step": 4040 }, { "epoch": 3.159126365054602, "grad_norm": 8.070413589477539, "learning_rate": 0.0005490305326498775, "loss": 0.1438, "step": 4050 }, { "epoch": 3.1669266770670825, "grad_norm": 0.971108078956604, "learning_rate": 0.0005479162023623803, "loss": 0.1521, "step": 4060 }, { "epoch": 3.1747269890795633, "grad_norm": 0.9375523924827576, "learning_rate": 0.000546801872074883, "loss": 0.1313, "step": 4070 }, { "epoch": 3.1825273010920436, "grad_norm": 1.3351547718048096, "learning_rate": 0.0005456875417873858, "loss": 0.1626, "step": 4080 }, { "epoch": 3.1903276131045244, "grad_norm": 2.073784589767456, "learning_rate": 0.0005445732114998886, "loss": 0.1435, "step": 4090 }, { "epoch": 3.1981279251170047, "grad_norm": 1.691977620124817, "learning_rate": 0.0005434588812123914, "loss": 0.1742, "step": 4100 }, { "epoch": 3.205928237129485, "grad_norm": 22.18892478942871, "learning_rate": 0.0005423445509248941, "loss": 0.1618, "step": 4110 }, { "epoch": 3.213728549141966, "grad_norm": 1.7861751317977905, "learning_rate": 0.0005412302206373969, "loss": 0.1351, "step": 4120 }, { "epoch": 3.221528861154446, "grad_norm": 1.1942836046218872, "learning_rate": 0.0005401158903498997, "loss": 0.1403, "step": 4130 }, { "epoch": 3.2293291731669265, "grad_norm": 4.223365306854248, "learning_rate": 0.0005390015600624025, "loss": 0.1568, "step": 4140 }, { "epoch": 3.2371294851794072, "grad_norm": 1.828805685043335, "learning_rate": 0.0005378872297749053, "loss": 0.1732, "step": 4150 }, { "epoch": 3.2449297971918876, "grad_norm": 1.3445240259170532, "learning_rate": 0.0005367728994874081, "loss": 0.1609, "step": 4160 }, { "epoch": 3.2527301092043683, "grad_norm": 1.8210989236831665, "learning_rate": 0.0005356585691999109, "loss": 0.1529, "step": 4170 }, { "epoch": 3.2605304212168487, "grad_norm": 1.5427830219268799, "learning_rate": 0.0005345442389124137, "loss": 0.1245, "step": 4180 }, { "epoch": 3.268330733229329, "grad_norm": 2.216970920562744, "learning_rate": 0.0005334299086249165, "loss": 0.1321, "step": 4190 }, { "epoch": 3.2761310452418098, "grad_norm": 2.470158338546753, "learning_rate": 0.0005323155783374192, "loss": 0.1466, "step": 4200 }, { "epoch": 3.28393135725429, "grad_norm": 1.8444827795028687, "learning_rate": 0.000531201248049922, "loss": 0.141, "step": 4210 }, { "epoch": 3.291731669266771, "grad_norm": 1.771625280380249, "learning_rate": 0.0005300869177624248, "loss": 0.1443, "step": 4220 }, { "epoch": 3.299531981279251, "grad_norm": 2.486222267150879, "learning_rate": 0.0005289725874749276, "loss": 0.1665, "step": 4230 }, { "epoch": 3.3073322932917315, "grad_norm": 2.0185546875, "learning_rate": 0.0005278582571874304, "loss": 0.2224, "step": 4240 }, { "epoch": 3.3151326053042123, "grad_norm": 1.810857892036438, "learning_rate": 0.0005267439268999331, "loss": 0.1384, "step": 4250 }, { "epoch": 3.3229329173166926, "grad_norm": 0.9884628057479858, "learning_rate": 0.0005256295966124359, "loss": 0.1028, "step": 4260 }, { "epoch": 3.330733229329173, "grad_norm": 0.9811989068984985, "learning_rate": 0.0005245152663249387, "loss": 0.1327, "step": 4270 }, { "epoch": 3.3385335413416537, "grad_norm": 0.9268475770950317, "learning_rate": 0.0005234009360374415, "loss": 0.2011, "step": 4280 }, { "epoch": 3.346333853354134, "grad_norm": 2.9813990592956543, "learning_rate": 0.0005222866057499443, "loss": 0.1729, "step": 4290 }, { "epoch": 3.354134165366615, "grad_norm": 1.216805338859558, "learning_rate": 0.0005211722754624471, "loss": 0.148, "step": 4300 }, { "epoch": 3.361934477379095, "grad_norm": 4.585384845733643, "learning_rate": 0.0005200579451749499, "loss": 0.28, "step": 4310 }, { "epoch": 3.3697347893915754, "grad_norm": 1.331865906715393, "learning_rate": 0.0005189436148874527, "loss": 0.1757, "step": 4320 }, { "epoch": 3.377535101404056, "grad_norm": 3.4767305850982666, "learning_rate": 0.0005178292845999555, "loss": 0.1955, "step": 4330 }, { "epoch": 3.3853354134165365, "grad_norm": 1.2598906755447388, "learning_rate": 0.0005167149543124582, "loss": 0.1633, "step": 4340 }, { "epoch": 3.3931357254290173, "grad_norm": 1.9768121242523193, "learning_rate": 0.000515600624024961, "loss": 0.1314, "step": 4350 }, { "epoch": 3.4009360374414976, "grad_norm": 1.916244387626648, "learning_rate": 0.0005144862937374638, "loss": 0.1826, "step": 4360 }, { "epoch": 3.408736349453978, "grad_norm": 1.159989595413208, "learning_rate": 0.0005133719634499666, "loss": 0.1655, "step": 4370 }, { "epoch": 3.4165366614664587, "grad_norm": 2.1428520679473877, "learning_rate": 0.0005122576331624693, "loss": 0.1868, "step": 4380 }, { "epoch": 3.424336973478939, "grad_norm": 1.3343071937561035, "learning_rate": 0.0005111433028749721, "loss": 0.1539, "step": 4390 }, { "epoch": 3.43213728549142, "grad_norm": 2.33966326713562, "learning_rate": 0.0005100289725874749, "loss": 0.1613, "step": 4400 }, { "epoch": 3.4399375975039, "grad_norm": 1.662204384803772, "learning_rate": 0.0005089146422999777, "loss": 0.1554, "step": 4410 }, { "epoch": 3.4477379095163805, "grad_norm": 0.7760083675384521, "learning_rate": 0.0005078003120124805, "loss": 0.0962, "step": 4420 }, { "epoch": 3.4555382215288613, "grad_norm": 1.860338807106018, "learning_rate": 0.0005066859817249833, "loss": 0.1757, "step": 4430 }, { "epoch": 3.4633385335413416, "grad_norm": 0.6259675025939941, "learning_rate": 0.0005055716514374861, "loss": 0.176, "step": 4440 }, { "epoch": 3.4711388455538223, "grad_norm": 2.819291591644287, "learning_rate": 0.0005044573211499889, "loss": 0.2447, "step": 4450 }, { "epoch": 3.4789391575663027, "grad_norm": 1.125827431678772, "learning_rate": 0.0005033429908624917, "loss": 0.1265, "step": 4460 }, { "epoch": 3.486739469578783, "grad_norm": 1.1929935216903687, "learning_rate": 0.0005022286605749944, "loss": 0.1428, "step": 4470 }, { "epoch": 3.4945397815912638, "grad_norm": 1.2320295572280884, "learning_rate": 0.0005011143302874972, "loss": 0.1372, "step": 4480 }, { "epoch": 3.502340093603744, "grad_norm": 1.1627320051193237, "learning_rate": 0.0005, "loss": 0.1143, "step": 4490 }, { "epoch": 3.510140405616225, "grad_norm": 2.7850723266601562, "learning_rate": 0.0004988856697125028, "loss": 0.1431, "step": 4500 }, { "epoch": 3.517940717628705, "grad_norm": 1.2920804023742676, "learning_rate": 0.0004977713394250056, "loss": 0.17, "step": 4510 }, { "epoch": 3.5257410296411855, "grad_norm": 1.2807234525680542, "learning_rate": 0.0004966570091375084, "loss": 0.1379, "step": 4520 }, { "epoch": 3.5335413416536663, "grad_norm": 1.2454224824905396, "learning_rate": 0.0004955426788500111, "loss": 0.1439, "step": 4530 }, { "epoch": 3.5413416536661466, "grad_norm": 3.7264480590820312, "learning_rate": 0.0004944283485625139, "loss": 0.1691, "step": 4540 }, { "epoch": 3.5491419656786274, "grad_norm": 1.653113842010498, "learning_rate": 0.0004933140182750167, "loss": 0.157, "step": 4550 }, { "epoch": 3.5569422776911077, "grad_norm": 2.0576372146606445, "learning_rate": 0.0004921996879875195, "loss": 0.171, "step": 4560 }, { "epoch": 3.564742589703588, "grad_norm": 1.499207615852356, "learning_rate": 0.0004910853577000223, "loss": 0.1297, "step": 4570 }, { "epoch": 3.572542901716069, "grad_norm": 1.2240318059921265, "learning_rate": 0.0004899710274125251, "loss": 0.2274, "step": 4580 }, { "epoch": 3.580343213728549, "grad_norm": 1.4562804698944092, "learning_rate": 0.0004888566971250279, "loss": 0.1884, "step": 4590 }, { "epoch": 3.58814352574103, "grad_norm": 2.0862207412719727, "learning_rate": 0.00048774236683753067, "loss": 0.1323, "step": 4600 }, { "epoch": 3.5959438377535102, "grad_norm": 1.8791660070419312, "learning_rate": 0.00048662803655003347, "loss": 0.1446, "step": 4610 }, { "epoch": 3.6037441497659906, "grad_norm": 2.589106798171997, "learning_rate": 0.0004855137062625362, "loss": 0.1722, "step": 4620 }, { "epoch": 3.611544461778471, "grad_norm": 1.3053058385849, "learning_rate": 0.000484399375975039, "loss": 0.1458, "step": 4630 }, { "epoch": 3.6193447737909517, "grad_norm": 1.4898595809936523, "learning_rate": 0.00048328504568754177, "loss": 0.155, "step": 4640 }, { "epoch": 3.627145085803432, "grad_norm": 1.9109569787979126, "learning_rate": 0.00048217071540004457, "loss": 0.1526, "step": 4650 }, { "epoch": 3.6349453978159127, "grad_norm": 2.045905351638794, "learning_rate": 0.00048105638511254737, "loss": 0.1281, "step": 4660 }, { "epoch": 3.642745709828393, "grad_norm": 2.0706069469451904, "learning_rate": 0.0004799420548250502, "loss": 0.2016, "step": 4670 }, { "epoch": 3.6505460218408734, "grad_norm": 1.3242971897125244, "learning_rate": 0.000478827724537553, "loss": 0.1715, "step": 4680 }, { "epoch": 3.658346333853354, "grad_norm": 1.2207064628601074, "learning_rate": 0.0004777133942500557, "loss": 0.1262, "step": 4690 }, { "epoch": 3.6661466458658345, "grad_norm": 2.0521817207336426, "learning_rate": 0.0004765990639625585, "loss": 0.1929, "step": 4700 }, { "epoch": 3.6739469578783153, "grad_norm": 1.009435772895813, "learning_rate": 0.00047548473367506127, "loss": 0.2248, "step": 4710 }, { "epoch": 3.6817472698907956, "grad_norm": 1.105904221534729, "learning_rate": 0.0004743704033875641, "loss": 0.1554, "step": 4720 }, { "epoch": 3.689547581903276, "grad_norm": 1.9131145477294922, "learning_rate": 0.0004732560731000669, "loss": 0.1435, "step": 4730 }, { "epoch": 3.6973478939157567, "grad_norm": 4.082576751708984, "learning_rate": 0.0004721417428125697, "loss": 0.1539, "step": 4740 }, { "epoch": 3.705148205928237, "grad_norm": 0.8368641138076782, "learning_rate": 0.0004710274125250724, "loss": 0.1296, "step": 4750 }, { "epoch": 3.712948517940718, "grad_norm": 2.7137794494628906, "learning_rate": 0.00046991308223757523, "loss": 0.1557, "step": 4760 }, { "epoch": 3.720748829953198, "grad_norm": 1.2786542177200317, "learning_rate": 0.00046879875195007803, "loss": 0.1529, "step": 4770 }, { "epoch": 3.7285491419656784, "grad_norm": 2.214386463165283, "learning_rate": 0.0004676844216625808, "loss": 0.143, "step": 4780 }, { "epoch": 3.736349453978159, "grad_norm": 1.2340818643569946, "learning_rate": 0.0004665700913750836, "loss": 0.1481, "step": 4790 }, { "epoch": 3.7441497659906395, "grad_norm": 1.804880976676941, "learning_rate": 0.0004654557610875864, "loss": 0.1385, "step": 4800 }, { "epoch": 3.7519500780031203, "grad_norm": 1.0058701038360596, "learning_rate": 0.0004643414308000892, "loss": 0.2446, "step": 4810 }, { "epoch": 3.7597503900156006, "grad_norm": 2.6180830001831055, "learning_rate": 0.00046322710051259193, "loss": 0.18, "step": 4820 }, { "epoch": 3.767550702028081, "grad_norm": 0.8924421072006226, "learning_rate": 0.00046211277022509473, "loss": 0.1533, "step": 4830 }, { "epoch": 3.7753510140405617, "grad_norm": 1.5280059576034546, "learning_rate": 0.0004609984399375975, "loss": 0.1567, "step": 4840 }, { "epoch": 3.783151326053042, "grad_norm": 2.1860077381134033, "learning_rate": 0.0004598841096501003, "loss": 0.1502, "step": 4850 }, { "epoch": 3.790951638065523, "grad_norm": 1.2397303581237793, "learning_rate": 0.00045876977936260314, "loss": 0.163, "step": 4860 }, { "epoch": 3.798751950078003, "grad_norm": 1.8644335269927979, "learning_rate": 0.0004576554490751059, "loss": 0.1843, "step": 4870 }, { "epoch": 3.8065522620904835, "grad_norm": 4.985640525817871, "learning_rate": 0.0004565411187876087, "loss": 0.2699, "step": 4880 }, { "epoch": 3.8143525741029642, "grad_norm": 1.4370672702789307, "learning_rate": 0.00045542678850011144, "loss": 0.1285, "step": 4890 }, { "epoch": 3.8221528861154446, "grad_norm": 1.725677728652954, "learning_rate": 0.00045431245821261424, "loss": 0.1609, "step": 4900 }, { "epoch": 3.8299531981279253, "grad_norm": 0.9739437699317932, "learning_rate": 0.000453198127925117, "loss": 0.1614, "step": 4910 }, { "epoch": 3.8377535101404057, "grad_norm": 3.672434091567993, "learning_rate": 0.0004520837976376198, "loss": 0.1747, "step": 4920 }, { "epoch": 3.845553822152886, "grad_norm": 1.3917264938354492, "learning_rate": 0.0004509694673501226, "loss": 0.159, "step": 4930 }, { "epoch": 3.8533541341653668, "grad_norm": 2.578519105911255, "learning_rate": 0.0004498551370626254, "loss": 0.1394, "step": 4940 }, { "epoch": 3.861154446177847, "grad_norm": 1.3072192668914795, "learning_rate": 0.00044874080677512814, "loss": 0.1167, "step": 4950 }, { "epoch": 3.868954758190328, "grad_norm": 1.1134214401245117, "learning_rate": 0.00044762647648763094, "loss": 0.1445, "step": 4960 }, { "epoch": 3.876755070202808, "grad_norm": 2.7288825511932373, "learning_rate": 0.00044651214620013374, "loss": 0.1693, "step": 4970 }, { "epoch": 3.8845553822152885, "grad_norm": 1.1238129138946533, "learning_rate": 0.0004453978159126365, "loss": 0.1455, "step": 4980 }, { "epoch": 3.892355694227769, "grad_norm": 1.7059211730957031, "learning_rate": 0.0004442834856251393, "loss": 0.1069, "step": 4990 }, { "epoch": 3.9001560062402496, "grad_norm": 1.0301696062088013, "learning_rate": 0.0004431691553376421, "loss": 0.3035, "step": 5000 }, { "epoch": 3.9079563182527304, "grad_norm": 1.744986891746521, "learning_rate": 0.0004420548250501449, "loss": 0.1249, "step": 5010 }, { "epoch": 3.9157566302652107, "grad_norm": 0.8731406331062317, "learning_rate": 0.00044094049476264764, "loss": 0.1714, "step": 5020 }, { "epoch": 3.923556942277691, "grad_norm": 1.1749571561813354, "learning_rate": 0.00043982616447515045, "loss": 0.264, "step": 5030 }, { "epoch": 3.9313572542901714, "grad_norm": 2.552839756011963, "learning_rate": 0.0004387118341876532, "loss": 0.2517, "step": 5040 }, { "epoch": 3.939157566302652, "grad_norm": 1.9722734689712524, "learning_rate": 0.000437597503900156, "loss": 0.1854, "step": 5050 }, { "epoch": 3.9469578783151325, "grad_norm": 2.063805341720581, "learning_rate": 0.0004364831736126588, "loss": 0.1574, "step": 5060 }, { "epoch": 3.954758190327613, "grad_norm": 4.838432312011719, "learning_rate": 0.0004353688433251616, "loss": 0.1939, "step": 5070 }, { "epoch": 3.9625585023400935, "grad_norm": 1.9519829750061035, "learning_rate": 0.0004342545130376644, "loss": 0.1334, "step": 5080 }, { "epoch": 3.970358814352574, "grad_norm": 1.3953014612197876, "learning_rate": 0.00043314018275016715, "loss": 0.1504, "step": 5090 }, { "epoch": 3.9781591263650546, "grad_norm": 1.6272999048233032, "learning_rate": 0.00043202585246266995, "loss": 0.153, "step": 5100 }, { "epoch": 3.985959438377535, "grad_norm": 1.610538125038147, "learning_rate": 0.0004309115221751727, "loss": 0.1727, "step": 5110 }, { "epoch": 3.9937597503900157, "grad_norm": 1.6635684967041016, "learning_rate": 0.0004297971918876755, "loss": 0.1316, "step": 5120 }, { "epoch": 4.0, "eval_loss": 0.31802308559417725, "eval_runtime": 526.9062, "eval_samples_per_second": 1.044, "eval_steps_per_second": 1.044, "eval_wer": 21.84784707472485, "step": 5128 }, { "epoch": 4.001560062402496, "grad_norm": 0.6206927299499512, "learning_rate": 0.0004286828616001783, "loss": 0.1324, "step": 5130 }, { "epoch": 4.009360374414976, "grad_norm": 1.6887155771255493, "learning_rate": 0.0004275685313126811, "loss": 0.1079, "step": 5140 }, { "epoch": 4.017160686427457, "grad_norm": 0.9051909446716309, "learning_rate": 0.0004264542010251839, "loss": 0.0854, "step": 5150 }, { "epoch": 4.024960998439938, "grad_norm": 0.5828364491462708, "learning_rate": 0.00042533987073768665, "loss": 0.0619, "step": 5160 }, { "epoch": 4.032761310452418, "grad_norm": 1.66502046585083, "learning_rate": 0.00042422554045018946, "loss": 0.1545, "step": 5170 }, { "epoch": 4.040561622464899, "grad_norm": 0.7514560222625732, "learning_rate": 0.0004231112101626922, "loss": 0.0779, "step": 5180 }, { "epoch": 4.048361934477379, "grad_norm": 0.8129399418830872, "learning_rate": 0.000421996879875195, "loss": 0.0423, "step": 5190 }, { "epoch": 4.056162246489859, "grad_norm": 1.5373820066452026, "learning_rate": 0.0004208825495876978, "loss": 0.0664, "step": 5200 }, { "epoch": 4.0639625585023404, "grad_norm": 0.8093292117118835, "learning_rate": 0.0004197682193002006, "loss": 0.066, "step": 5210 }, { "epoch": 4.071762870514821, "grad_norm": 1.3024191856384277, "learning_rate": 0.00041865388901270336, "loss": 0.1046, "step": 5220 }, { "epoch": 4.079563182527301, "grad_norm": 0.7004597187042236, "learning_rate": 0.00041753955872520616, "loss": 0.07, "step": 5230 }, { "epoch": 4.087363494539781, "grad_norm": 1.7296199798583984, "learning_rate": 0.00041642522843770896, "loss": 0.0882, "step": 5240 }, { "epoch": 4.095163806552262, "grad_norm": 0.8828613758087158, "learning_rate": 0.0004153108981502117, "loss": 0.0802, "step": 5250 }, { "epoch": 4.102964118564743, "grad_norm": 1.9043530225753784, "learning_rate": 0.0004141965678627145, "loss": 0.0688, "step": 5260 }, { "epoch": 4.110764430577223, "grad_norm": 0.7490746974945068, "learning_rate": 0.0004130822375752173, "loss": 0.0736, "step": 5270 }, { "epoch": 4.118564742589704, "grad_norm": 0.5277013182640076, "learning_rate": 0.0004119679072877201, "loss": 0.0669, "step": 5280 }, { "epoch": 4.126365054602184, "grad_norm": 2.074106216430664, "learning_rate": 0.00041085357700022286, "loss": 0.1113, "step": 5290 }, { "epoch": 4.134165366614664, "grad_norm": 1.2340360879898071, "learning_rate": 0.00040973924671272566, "loss": 0.0687, "step": 5300 }, { "epoch": 4.1419656786271455, "grad_norm": 1.7778706550598145, "learning_rate": 0.0004086249164252284, "loss": 0.0809, "step": 5310 }, { "epoch": 4.149765990639626, "grad_norm": 1.469011664390564, "learning_rate": 0.0004075105861377312, "loss": 0.0774, "step": 5320 }, { "epoch": 4.157566302652106, "grad_norm": 1.4440765380859375, "learning_rate": 0.00040639625585023407, "loss": 0.0912, "step": 5330 }, { "epoch": 4.1653666146645865, "grad_norm": 2.316056251525879, "learning_rate": 0.0004052819255627368, "loss": 0.089, "step": 5340 }, { "epoch": 4.173166926677067, "grad_norm": 1.2539482116699219, "learning_rate": 0.0004041675952752396, "loss": 0.0703, "step": 5350 }, { "epoch": 4.180967238689548, "grad_norm": 1.6124414205551147, "learning_rate": 0.00040305326498774237, "loss": 0.1678, "step": 5360 }, { "epoch": 4.188767550702028, "grad_norm": 2.8829872608184814, "learning_rate": 0.00040193893470024517, "loss": 0.0925, "step": 5370 }, { "epoch": 4.196567862714509, "grad_norm": 1.8418601751327515, "learning_rate": 0.0004008246044127479, "loss": 0.0982, "step": 5380 }, { "epoch": 4.204368174726989, "grad_norm": 3.860861301422119, "learning_rate": 0.0003997102741252507, "loss": 0.1058, "step": 5390 }, { "epoch": 4.212168486739469, "grad_norm": 1.5564589500427246, "learning_rate": 0.0003985959438377535, "loss": 0.0972, "step": 5400 }, { "epoch": 4.21996879875195, "grad_norm": 1.1446703672409058, "learning_rate": 0.0003974816135502563, "loss": 0.0947, "step": 5410 }, { "epoch": 4.227769110764431, "grad_norm": 1.7438228130340576, "learning_rate": 0.0003963672832627591, "loss": 0.1094, "step": 5420 }, { "epoch": 4.235569422776911, "grad_norm": 1.3870216608047485, "learning_rate": 0.00039525295297526187, "loss": 0.0855, "step": 5430 }, { "epoch": 4.2433697347893915, "grad_norm": 1.2293809652328491, "learning_rate": 0.00039413862268776467, "loss": 0.0656, "step": 5440 }, { "epoch": 4.251170046801872, "grad_norm": 1.1131449937820435, "learning_rate": 0.0003930242924002674, "loss": 0.0854, "step": 5450 }, { "epoch": 4.258970358814352, "grad_norm": 1.5234099626541138, "learning_rate": 0.0003919099621127702, "loss": 0.1031, "step": 5460 }, { "epoch": 4.266770670826833, "grad_norm": 1.6739901304244995, "learning_rate": 0.000390795631825273, "loss": 0.0738, "step": 5470 }, { "epoch": 4.274570982839314, "grad_norm": 1.2000303268432617, "learning_rate": 0.0003896813015377758, "loss": 0.0685, "step": 5480 }, { "epoch": 4.282371294851794, "grad_norm": 2.348032236099243, "learning_rate": 0.0003885669712502786, "loss": 0.1005, "step": 5490 }, { "epoch": 4.290171606864274, "grad_norm": 2.87178373336792, "learning_rate": 0.0003874526409627814, "loss": 0.1776, "step": 5500 }, { "epoch": 4.297971918876755, "grad_norm": 1.1684914827346802, "learning_rate": 0.0003863383106752842, "loss": 0.0726, "step": 5510 }, { "epoch": 4.305772230889236, "grad_norm": 1.1053632497787476, "learning_rate": 0.0003852239803877869, "loss": 0.0917, "step": 5520 }, { "epoch": 4.313572542901716, "grad_norm": 1.0637143850326538, "learning_rate": 0.0003841096501002897, "loss": 0.146, "step": 5530 }, { "epoch": 4.3213728549141965, "grad_norm": 1.628104567527771, "learning_rate": 0.00038299531981279253, "loss": 0.0889, "step": 5540 }, { "epoch": 4.329173166926677, "grad_norm": 2.666768789291382, "learning_rate": 0.00038188098952529533, "loss": 0.1652, "step": 5550 }, { "epoch": 4.336973478939157, "grad_norm": 0.9281901717185974, "learning_rate": 0.0003807666592377981, "loss": 0.1052, "step": 5560 }, { "epoch": 4.344773790951638, "grad_norm": 1.086384892463684, "learning_rate": 0.0003796523289503009, "loss": 0.0889, "step": 5570 }, { "epoch": 4.352574102964119, "grad_norm": 0.9888959527015686, "learning_rate": 0.00037853799866280363, "loss": 0.0876, "step": 5580 }, { "epoch": 4.360374414976599, "grad_norm": 1.622720718383789, "learning_rate": 0.00037742366837530643, "loss": 0.0892, "step": 5590 }, { "epoch": 4.368174726989079, "grad_norm": 0.9988596439361572, "learning_rate": 0.0003763093380878093, "loss": 0.0997, "step": 5600 }, { "epoch": 4.37597503900156, "grad_norm": 1.840929388999939, "learning_rate": 0.00037519500780031203, "loss": 0.0719, "step": 5610 }, { "epoch": 4.383775351014041, "grad_norm": 1.2531293630599976, "learning_rate": 0.00037408067751281484, "loss": 0.0772, "step": 5620 }, { "epoch": 4.391575663026521, "grad_norm": 0.6823952794075012, "learning_rate": 0.0003729663472253176, "loss": 0.1192, "step": 5630 }, { "epoch": 4.399375975039002, "grad_norm": 0.9205799102783203, "learning_rate": 0.0003718520169378204, "loss": 0.07, "step": 5640 }, { "epoch": 4.407176287051482, "grad_norm": 1.4101178646087646, "learning_rate": 0.00037073768665032313, "loss": 0.1165, "step": 5650 }, { "epoch": 4.414976599063962, "grad_norm": 0.9926198124885559, "learning_rate": 0.00036962335636282593, "loss": 0.0749, "step": 5660 }, { "epoch": 4.422776911076443, "grad_norm": 1.9218907356262207, "learning_rate": 0.00036850902607532874, "loss": 0.0739, "step": 5670 }, { "epoch": 4.430577223088924, "grad_norm": 2.6744730472564697, "learning_rate": 0.00036739469578783154, "loss": 0.1022, "step": 5680 }, { "epoch": 4.438377535101404, "grad_norm": 1.0124350786209106, "learning_rate": 0.00036628036550033434, "loss": 0.0892, "step": 5690 }, { "epoch": 4.446177847113884, "grad_norm": 1.43318510055542, "learning_rate": 0.0003651660352128371, "loss": 0.082, "step": 5700 }, { "epoch": 4.453978159126365, "grad_norm": 1.302046537399292, "learning_rate": 0.0003640517049253399, "loss": 0.0879, "step": 5710 }, { "epoch": 4.461778471138846, "grad_norm": 0.7893558740615845, "learning_rate": 0.00036293737463784264, "loss": 0.0849, "step": 5720 }, { "epoch": 4.469578783151326, "grad_norm": 1.573437213897705, "learning_rate": 0.00036182304435034544, "loss": 0.1125, "step": 5730 }, { "epoch": 4.477379095163807, "grad_norm": 3.517354965209961, "learning_rate": 0.00036070871406284824, "loss": 0.1041, "step": 5740 }, { "epoch": 4.485179407176287, "grad_norm": 1.05705726146698, "learning_rate": 0.00035959438377535104, "loss": 0.0815, "step": 5750 }, { "epoch": 4.492979719188767, "grad_norm": 0.7418360710144043, "learning_rate": 0.0003584800534878538, "loss": 0.0698, "step": 5760 }, { "epoch": 4.5007800312012485, "grad_norm": 2.2736730575561523, "learning_rate": 0.0003573657232003566, "loss": 0.0742, "step": 5770 }, { "epoch": 4.508580343213729, "grad_norm": 0.7430834174156189, "learning_rate": 0.0003562513929128594, "loss": 0.1088, "step": 5780 }, { "epoch": 4.516380655226209, "grad_norm": 3.169813632965088, "learning_rate": 0.00035513706262536214, "loss": 0.0999, "step": 5790 }, { "epoch": 4.5241809672386895, "grad_norm": 1.5269415378570557, "learning_rate": 0.000354022732337865, "loss": 0.1035, "step": 5800 }, { "epoch": 4.53198127925117, "grad_norm": 1.1650433540344238, "learning_rate": 0.00035290840205036775, "loss": 0.0838, "step": 5810 }, { "epoch": 4.539781591263651, "grad_norm": 1.535849690437317, "learning_rate": 0.00035179407176287055, "loss": 0.1141, "step": 5820 }, { "epoch": 4.547581903276131, "grad_norm": 0.7315622568130493, "learning_rate": 0.0003506797414753733, "loss": 0.1007, "step": 5830 }, { "epoch": 4.555382215288612, "grad_norm": 0.8726515769958496, "learning_rate": 0.0003495654111878761, "loss": 0.0897, "step": 5840 }, { "epoch": 4.563182527301092, "grad_norm": 3.0370798110961914, "learning_rate": 0.00034845108090037885, "loss": 0.0912, "step": 5850 }, { "epoch": 4.570982839313572, "grad_norm": 2.573432207107544, "learning_rate": 0.00034733675061288165, "loss": 0.1582, "step": 5860 }, { "epoch": 4.5787831513260535, "grad_norm": 1.1244597434997559, "learning_rate": 0.00034622242032538445, "loss": 0.1557, "step": 5870 }, { "epoch": 4.586583463338534, "grad_norm": 0.8730902671813965, "learning_rate": 0.00034510809003788725, "loss": 0.0922, "step": 5880 }, { "epoch": 4.594383775351014, "grad_norm": 2.5110976696014404, "learning_rate": 0.00034399375975039005, "loss": 0.1827, "step": 5890 }, { "epoch": 4.6021840873634945, "grad_norm": 1.3127886056900024, "learning_rate": 0.0003428794294628928, "loss": 0.0902, "step": 5900 }, { "epoch": 4.609984399375975, "grad_norm": 2.67976450920105, "learning_rate": 0.0003417650991753956, "loss": 0.1136, "step": 5910 }, { "epoch": 4.617784711388456, "grad_norm": 0.5845116376876831, "learning_rate": 0.00034065076888789835, "loss": 0.0835, "step": 5920 }, { "epoch": 4.625585023400936, "grad_norm": 0.5674394965171814, "learning_rate": 0.00033953643860040115, "loss": 0.0862, "step": 5930 }, { "epoch": 4.633385335413417, "grad_norm": 1.654697299003601, "learning_rate": 0.00033842210831290395, "loss": 0.0904, "step": 5940 }, { "epoch": 4.641185647425897, "grad_norm": 1.1088430881500244, "learning_rate": 0.00033730777802540676, "loss": 0.1512, "step": 5950 }, { "epoch": 4.648985959438377, "grad_norm": 0.6826479434967041, "learning_rate": 0.0003361934477379095, "loss": 0.0758, "step": 5960 }, { "epoch": 4.656786271450858, "grad_norm": 2.3032681941986084, "learning_rate": 0.0003350791174504123, "loss": 0.1093, "step": 5970 }, { "epoch": 4.664586583463339, "grad_norm": 0.9655252695083618, "learning_rate": 0.0003339647871629151, "loss": 0.0698, "step": 5980 }, { "epoch": 4.672386895475819, "grad_norm": 2.330497980117798, "learning_rate": 0.00033285045687541786, "loss": 0.1139, "step": 5990 }, { "epoch": 4.6801872074882995, "grad_norm": 0.5093659162521362, "learning_rate": 0.00033173612658792066, "loss": 0.0919, "step": 6000 }, { "epoch": 4.68798751950078, "grad_norm": 0.9975968599319458, "learning_rate": 0.00033062179630042346, "loss": 0.1115, "step": 6010 }, { "epoch": 4.69578783151326, "grad_norm": 0.7988600730895996, "learning_rate": 0.00032950746601292626, "loss": 0.0684, "step": 6020 }, { "epoch": 4.703588143525741, "grad_norm": 3.174332857131958, "learning_rate": 0.000328393135725429, "loss": 0.1761, "step": 6030 }, { "epoch": 4.711388455538222, "grad_norm": 0.6243475675582886, "learning_rate": 0.0003272788054379318, "loss": 0.087, "step": 6040 }, { "epoch": 4.719188767550702, "grad_norm": 0.5919020175933838, "learning_rate": 0.00032616447515043456, "loss": 0.1119, "step": 6050 }, { "epoch": 4.726989079563182, "grad_norm": 1.2413996458053589, "learning_rate": 0.00032505014486293736, "loss": 0.1033, "step": 6060 }, { "epoch": 4.734789391575663, "grad_norm": 0.5087947249412537, "learning_rate": 0.0003239358145754402, "loss": 0.1037, "step": 6070 }, { "epoch": 4.742589703588144, "grad_norm": 0.8413617014884949, "learning_rate": 0.00032282148428794296, "loss": 0.0719, "step": 6080 }, { "epoch": 4.750390015600624, "grad_norm": 0.6793035864830017, "learning_rate": 0.00032170715400044577, "loss": 0.0719, "step": 6090 }, { "epoch": 4.758190327613105, "grad_norm": 1.9824508428573608, "learning_rate": 0.0003207042567416982, "loss": 0.0926, "step": 6100 }, { "epoch": 4.765990639625585, "grad_norm": 0.9389640688896179, "learning_rate": 0.00031958992645420107, "loss": 0.072, "step": 6110 }, { "epoch": 4.773790951638065, "grad_norm": 2.0503666400909424, "learning_rate": 0.0003184755961667038, "loss": 0.133, "step": 6120 }, { "epoch": 4.7815912636505455, "grad_norm": 0.9717590808868408, "learning_rate": 0.0003173612658792066, "loss": 0.1067, "step": 6130 }, { "epoch": 4.789391575663027, "grad_norm": 0.9370633363723755, "learning_rate": 0.00031624693559170936, "loss": 0.0811, "step": 6140 }, { "epoch": 4.797191887675507, "grad_norm": 1.7552763223648071, "learning_rate": 0.00031513260530421217, "loss": 0.1013, "step": 6150 }, { "epoch": 4.804992199687987, "grad_norm": 1.8624671697616577, "learning_rate": 0.00031401827501671497, "loss": 0.1082, "step": 6160 }, { "epoch": 4.812792511700468, "grad_norm": 1.0293713808059692, "learning_rate": 0.0003129039447292177, "loss": 0.1455, "step": 6170 }, { "epoch": 4.820592823712948, "grad_norm": 1.5493659973144531, "learning_rate": 0.00031178961444172057, "loss": 0.0757, "step": 6180 }, { "epoch": 4.828393135725429, "grad_norm": 0.4452805817127228, "learning_rate": 0.0003106752841542233, "loss": 0.0941, "step": 6190 }, { "epoch": 4.83619344773791, "grad_norm": 1.9039082527160645, "learning_rate": 0.0003095609538667261, "loss": 0.0941, "step": 6200 }, { "epoch": 4.84399375975039, "grad_norm": 1.5562025308609009, "learning_rate": 0.00030844662357922887, "loss": 0.0903, "step": 6210 }, { "epoch": 4.85179407176287, "grad_norm": 1.0633869171142578, "learning_rate": 0.00030733229329173167, "loss": 0.0689, "step": 6220 }, { "epoch": 4.859594383775351, "grad_norm": 0.3800029754638672, "learning_rate": 0.0003062179630042344, "loss": 0.0635, "step": 6230 }, { "epoch": 4.867394695787832, "grad_norm": 0.5013501644134521, "learning_rate": 0.0003051036327167372, "loss": 0.0706, "step": 6240 }, { "epoch": 4.875195007800312, "grad_norm": 0.3566785454750061, "learning_rate": 0.0003039893024292401, "loss": 0.0896, "step": 6250 }, { "epoch": 4.8829953198127924, "grad_norm": 1.4850417375564575, "learning_rate": 0.0003028749721417428, "loss": 0.0675, "step": 6260 }, { "epoch": 4.890795631825273, "grad_norm": 0.6070088148117065, "learning_rate": 0.0003017606418542456, "loss": 0.0996, "step": 6270 }, { "epoch": 4.898595943837753, "grad_norm": 1.4902043342590332, "learning_rate": 0.0003006463115667484, "loss": 0.07, "step": 6280 }, { "epoch": 4.906396255850234, "grad_norm": 1.594970941543579, "learning_rate": 0.0002995319812792512, "loss": 0.0948, "step": 6290 }, { "epoch": 4.914196567862715, "grad_norm": 1.2763868570327759, "learning_rate": 0.0002984176509917539, "loss": 0.0839, "step": 6300 }, { "epoch": 4.921996879875195, "grad_norm": 1.6840524673461914, "learning_rate": 0.0002973033207042567, "loss": 0.0996, "step": 6310 }, { "epoch": 4.929797191887675, "grad_norm": 1.1521170139312744, "learning_rate": 0.00029618899041675953, "loss": 0.0839, "step": 6320 }, { "epoch": 4.937597503900156, "grad_norm": 0.8535686731338501, "learning_rate": 0.00029507466012926233, "loss": 0.0888, "step": 6330 }, { "epoch": 4.945397815912637, "grad_norm": 1.0015994310379028, "learning_rate": 0.00029396032984176513, "loss": 0.2152, "step": 6340 }, { "epoch": 4.953198127925117, "grad_norm": 1.499544620513916, "learning_rate": 0.0002928459995542679, "loss": 0.0897, "step": 6350 }, { "epoch": 4.9609984399375975, "grad_norm": 2.329367160797119, "learning_rate": 0.0002917316692667707, "loss": 0.1614, "step": 6360 }, { "epoch": 4.968798751950078, "grad_norm": 1.5685973167419434, "learning_rate": 0.00029061733897927343, "loss": 0.0717, "step": 6370 }, { "epoch": 4.976599063962558, "grad_norm": 1.2372586727142334, "learning_rate": 0.0002895030086917763, "loss": 0.1291, "step": 6380 }, { "epoch": 4.984399375975039, "grad_norm": 1.1887118816375732, "learning_rate": 0.00028838867840427903, "loss": 0.0816, "step": 6390 }, { "epoch": 4.99219968798752, "grad_norm": 3.547868251800537, "learning_rate": 0.00028727434811678183, "loss": 0.0955, "step": 6400 }, { "epoch": 5.0, "grad_norm": 1.3139488697052002, "learning_rate": 0.0002861600178292846, "loss": 0.1174, "step": 6410 }, { "epoch": 5.0, "eval_loss": 0.3110225200653076, "eval_runtime": 535.2027, "eval_samples_per_second": 1.028, "eval_steps_per_second": 1.028, "eval_wer": 19.134968140567675, "step": 6410 }, { "epoch": 5.00780031201248, "grad_norm": 0.7019069194793701, "learning_rate": 0.0002850456875417874, "loss": 0.0452, "step": 6420 }, { "epoch": 5.015600624024961, "grad_norm": 0.933160662651062, "learning_rate": 0.0002839313572542902, "loss": 0.049, "step": 6430 }, { "epoch": 5.023400936037442, "grad_norm": 1.0783261060714722, "learning_rate": 0.00028281702696679293, "loss": 0.0417, "step": 6440 }, { "epoch": 5.031201248049922, "grad_norm": 0.5213541388511658, "learning_rate": 0.0002817026966792958, "loss": 0.0333, "step": 6450 }, { "epoch": 5.0390015600624025, "grad_norm": 2.4255504608154297, "learning_rate": 0.00028058836639179854, "loss": 0.0515, "step": 6460 }, { "epoch": 5.046801872074883, "grad_norm": 0.4260399639606476, "learning_rate": 0.00027947403610430134, "loss": 0.039, "step": 6470 }, { "epoch": 5.054602184087363, "grad_norm": 0.2641662657260895, "learning_rate": 0.0002783597058168041, "loss": 0.03, "step": 6480 }, { "epoch": 5.062402496099844, "grad_norm": 0.9397852420806885, "learning_rate": 0.0002772453755293069, "loss": 0.0489, "step": 6490 }, { "epoch": 5.070202808112325, "grad_norm": 1.324392557144165, "learning_rate": 0.00027613104524180964, "loss": 0.0379, "step": 6500 }, { "epoch": 5.078003120124805, "grad_norm": 0.6886770129203796, "learning_rate": 0.00027501671495431244, "loss": 0.0332, "step": 6510 }, { "epoch": 5.085803432137285, "grad_norm": 0.9160327911376953, "learning_rate": 0.0002739023846668153, "loss": 0.0386, "step": 6520 }, { "epoch": 5.093603744149766, "grad_norm": 1.2860448360443115, "learning_rate": 0.00027278805437931804, "loss": 0.0409, "step": 6530 }, { "epoch": 5.101404056162247, "grad_norm": 1.0299162864685059, "learning_rate": 0.00027167372409182084, "loss": 0.0446, "step": 6540 }, { "epoch": 5.109204368174727, "grad_norm": 0.8494958281517029, "learning_rate": 0.0002705593938043236, "loss": 0.0424, "step": 6550 }, { "epoch": 5.1170046801872076, "grad_norm": 0.21834442019462585, "learning_rate": 0.0002694450635168264, "loss": 0.0541, "step": 6560 }, { "epoch": 5.124804992199688, "grad_norm": 0.4797457754611969, "learning_rate": 0.00026833073322932914, "loss": 0.0385, "step": 6570 }, { "epoch": 5.132605304212168, "grad_norm": 1.514221429824829, "learning_rate": 0.000267216402941832, "loss": 0.077, "step": 6580 }, { "epoch": 5.140405616224649, "grad_norm": 1.4447021484375, "learning_rate": 0.00026610207265433475, "loss": 0.032, "step": 6590 }, { "epoch": 5.14820592823713, "grad_norm": 0.44626960158348083, "learning_rate": 0.00026498774236683755, "loss": 0.0287, "step": 6600 }, { "epoch": 5.15600624024961, "grad_norm": 0.3203662037849426, "learning_rate": 0.00026387341207934035, "loss": 0.032, "step": 6610 }, { "epoch": 5.16380655226209, "grad_norm": 0.5742075443267822, "learning_rate": 0.0002627590817918431, "loss": 0.0331, "step": 6620 }, { "epoch": 5.171606864274571, "grad_norm": 0.7284305095672607, "learning_rate": 0.0002616447515043459, "loss": 0.0416, "step": 6630 }, { "epoch": 5.179407176287052, "grad_norm": 1.0046939849853516, "learning_rate": 0.00026053042121684865, "loss": 0.0819, "step": 6640 }, { "epoch": 5.187207488299532, "grad_norm": 0.6021052002906799, "learning_rate": 0.0002594160909293515, "loss": 0.0358, "step": 6650 }, { "epoch": 5.195007800312013, "grad_norm": 1.0789178609848022, "learning_rate": 0.00025830176064185425, "loss": 0.0445, "step": 6660 }, { "epoch": 5.202808112324493, "grad_norm": 0.5201112627983093, "learning_rate": 0.00025718743035435705, "loss": 0.054, "step": 6670 }, { "epoch": 5.210608424336973, "grad_norm": 0.5282875299453735, "learning_rate": 0.0002560731000668598, "loss": 0.051, "step": 6680 }, { "epoch": 5.218408736349454, "grad_norm": 0.6322522163391113, "learning_rate": 0.0002549587697793626, "loss": 0.029, "step": 6690 }, { "epoch": 5.226209048361935, "grad_norm": 0.8109174370765686, "learning_rate": 0.0002538444394918654, "loss": 0.0529, "step": 6700 }, { "epoch": 5.234009360374415, "grad_norm": 0.8033832907676697, "learning_rate": 0.00025273010920436815, "loss": 0.0515, "step": 6710 }, { "epoch": 5.241809672386895, "grad_norm": 0.7336683869361877, "learning_rate": 0.000251615778916871, "loss": 0.0479, "step": 6720 }, { "epoch": 5.249609984399376, "grad_norm": 0.5804153680801392, "learning_rate": 0.00025050144862937376, "loss": 0.0497, "step": 6730 }, { "epoch": 5.257410296411856, "grad_norm": 0.6198682188987732, "learning_rate": 0.00024938711834187656, "loss": 0.0482, "step": 6740 }, { "epoch": 5.265210608424337, "grad_norm": 0.6303295493125916, "learning_rate": 0.00024827278805437936, "loss": 0.0375, "step": 6750 }, { "epoch": 5.273010920436818, "grad_norm": 0.33019986748695374, "learning_rate": 0.0002471584577668821, "loss": 0.0425, "step": 6760 }, { "epoch": 5.280811232449298, "grad_norm": 0.8694418668746948, "learning_rate": 0.0002460441274793849, "loss": 0.0556, "step": 6770 }, { "epoch": 5.288611544461778, "grad_norm": 0.43182373046875, "learning_rate": 0.00024492979719188766, "loss": 0.0281, "step": 6780 }, { "epoch": 5.296411856474259, "grad_norm": 1.0769046545028687, "learning_rate": 0.00024381546690439046, "loss": 0.043, "step": 6790 }, { "epoch": 5.30421216848674, "grad_norm": 2.0707390308380127, "learning_rate": 0.00024270113661689326, "loss": 0.0923, "step": 6800 }, { "epoch": 5.31201248049922, "grad_norm": 1.116832971572876, "learning_rate": 0.00024158680632939603, "loss": 0.0402, "step": 6810 }, { "epoch": 5.3198127925117005, "grad_norm": 2.6617612838745117, "learning_rate": 0.0002404724760418988, "loss": 0.0474, "step": 6820 }, { "epoch": 5.327613104524181, "grad_norm": 0.6436609625816345, "learning_rate": 0.00023935814575440158, "loss": 0.0437, "step": 6830 }, { "epoch": 5.335413416536661, "grad_norm": 0.6442177891731262, "learning_rate": 0.0002382438154669044, "loss": 0.0556, "step": 6840 }, { "epoch": 5.343213728549142, "grad_norm": 0.9249531030654907, "learning_rate": 0.0002371294851794072, "loss": 0.0409, "step": 6850 }, { "epoch": 5.351014040561623, "grad_norm": 1.3276004791259766, "learning_rate": 0.00023601515489190996, "loss": 0.0951, "step": 6860 }, { "epoch": 5.358814352574103, "grad_norm": 0.31066539883613586, "learning_rate": 0.00023490082460441276, "loss": 0.0273, "step": 6870 }, { "epoch": 5.366614664586583, "grad_norm": 0.582923948764801, "learning_rate": 0.00023378649431691554, "loss": 0.0417, "step": 6880 }, { "epoch": 5.374414976599064, "grad_norm": 1.0889739990234375, "learning_rate": 0.00023267216402941831, "loss": 0.0488, "step": 6890 }, { "epoch": 5.382215288611545, "grad_norm": 3.5691936016082764, "learning_rate": 0.00023155783374192112, "loss": 0.0528, "step": 6900 }, { "epoch": 5.390015600624025, "grad_norm": 0.6329577565193176, "learning_rate": 0.0002304435034544239, "loss": 0.0854, "step": 6910 }, { "epoch": 5.3978159126365055, "grad_norm": 0.3939153254032135, "learning_rate": 0.00022932917316692667, "loss": 0.0324, "step": 6920 }, { "epoch": 5.405616224648986, "grad_norm": 0.47252392768859863, "learning_rate": 0.00022821484287942947, "loss": 0.0508, "step": 6930 }, { "epoch": 5.413416536661466, "grad_norm": 1.2640843391418457, "learning_rate": 0.00022710051259193227, "loss": 0.1469, "step": 6940 }, { "epoch": 5.4212168486739465, "grad_norm": 0.6980273723602295, "learning_rate": 0.00022598618230443504, "loss": 0.0293, "step": 6950 }, { "epoch": 5.429017160686428, "grad_norm": 0.6737593412399292, "learning_rate": 0.00022487185201693782, "loss": 0.0307, "step": 6960 }, { "epoch": 5.436817472698908, "grad_norm": 0.4558241069316864, "learning_rate": 0.00022375752172944062, "loss": 0.0376, "step": 6970 }, { "epoch": 5.444617784711388, "grad_norm": 0.269889771938324, "learning_rate": 0.0002226431914419434, "loss": 0.0367, "step": 6980 }, { "epoch": 5.452418096723869, "grad_norm": 1.621789574623108, "learning_rate": 0.00022152886115444617, "loss": 0.0366, "step": 6990 }, { "epoch": 5.460218408736349, "grad_norm": 1.0347936153411865, "learning_rate": 0.00022041453086694895, "loss": 0.0335, "step": 7000 }, { "epoch": 5.46801872074883, "grad_norm": 0.5214996337890625, "learning_rate": 0.00021930020057945175, "loss": 0.0514, "step": 7010 }, { "epoch": 5.4758190327613105, "grad_norm": 4.288679122924805, "learning_rate": 0.00021818587029195455, "loss": 0.0516, "step": 7020 }, { "epoch": 5.483619344773791, "grad_norm": 1.2256821393966675, "learning_rate": 0.00021707154000445732, "loss": 0.0315, "step": 7030 }, { "epoch": 5.491419656786271, "grad_norm": 0.5737278461456299, "learning_rate": 0.00021595720971696013, "loss": 0.0282, "step": 7040 }, { "epoch": 5.4992199687987515, "grad_norm": 0.371231347322464, "learning_rate": 0.0002148428794294629, "loss": 0.0304, "step": 7050 }, { "epoch": 5.507020280811233, "grad_norm": 0.34915468096733093, "learning_rate": 0.00021372854914196568, "loss": 0.0434, "step": 7060 }, { "epoch": 5.514820592823713, "grad_norm": 0.6915215253829956, "learning_rate": 0.00021261421885446848, "loss": 0.0335, "step": 7070 }, { "epoch": 5.522620904836193, "grad_norm": 1.4587883949279785, "learning_rate": 0.00021149988856697125, "loss": 0.0327, "step": 7080 }, { "epoch": 5.530421216848674, "grad_norm": 1.0360335111618042, "learning_rate": 0.00021038555827947403, "loss": 0.0434, "step": 7090 }, { "epoch": 5.538221528861154, "grad_norm": 0.5000662803649902, "learning_rate": 0.0002092712279919768, "loss": 0.0442, "step": 7100 }, { "epoch": 5.546021840873635, "grad_norm": 0.7996264696121216, "learning_rate": 0.00020815689770447963, "loss": 0.0433, "step": 7110 }, { "epoch": 5.553822152886116, "grad_norm": 0.2298089563846588, "learning_rate": 0.0002070425674169824, "loss": 0.05, "step": 7120 }, { "epoch": 5.561622464898596, "grad_norm": 0.8946752548217773, "learning_rate": 0.00020592823712948518, "loss": 0.046, "step": 7130 }, { "epoch": 5.569422776911076, "grad_norm": 1.9877400398254395, "learning_rate": 0.00020481390684198798, "loss": 0.0586, "step": 7140 }, { "epoch": 5.577223088923557, "grad_norm": 0.7657136917114258, "learning_rate": 0.00020369957655449076, "loss": 0.038, "step": 7150 }, { "epoch": 5.585023400936038, "grad_norm": 0.5632153153419495, "learning_rate": 0.00020258524626699353, "loss": 0.0442, "step": 7160 }, { "epoch": 5.592823712948518, "grad_norm": 0.7438672780990601, "learning_rate": 0.00020147091597949633, "loss": 0.116, "step": 7170 }, { "epoch": 5.600624024960998, "grad_norm": 3.3054873943328857, "learning_rate": 0.0002003565856919991, "loss": 0.0454, "step": 7180 }, { "epoch": 5.608424336973479, "grad_norm": 1.7627991437911987, "learning_rate": 0.00019924225540450188, "loss": 0.1495, "step": 7190 }, { "epoch": 5.616224648985959, "grad_norm": 0.8701293468475342, "learning_rate": 0.00019812792511700469, "loss": 0.0439, "step": 7200 }, { "epoch": 5.62402496099844, "grad_norm": 0.5573844909667969, "learning_rate": 0.0001970135948295075, "loss": 0.0389, "step": 7210 }, { "epoch": 5.631825273010921, "grad_norm": 0.8145976662635803, "learning_rate": 0.00019589926454201026, "loss": 0.0489, "step": 7220 }, { "epoch": 5.639625585023401, "grad_norm": 0.5111231207847595, "learning_rate": 0.00019478493425451304, "loss": 0.0333, "step": 7230 }, { "epoch": 5.647425897035881, "grad_norm": 0.35958966612815857, "learning_rate": 0.00019367060396701584, "loss": 0.0438, "step": 7240 }, { "epoch": 5.655226209048362, "grad_norm": 0.3431423604488373, "learning_rate": 0.0001925562736795186, "loss": 0.0355, "step": 7250 }, { "epoch": 5.663026521060843, "grad_norm": 0.5120709538459778, "learning_rate": 0.0001914419433920214, "loss": 0.0305, "step": 7260 }, { "epoch": 5.670826833073323, "grad_norm": 0.9058636426925659, "learning_rate": 0.0001903276131045242, "loss": 0.1478, "step": 7270 }, { "epoch": 5.6786271450858035, "grad_norm": 1.0260646343231201, "learning_rate": 0.00018921328281702696, "loss": 0.0438, "step": 7280 }, { "epoch": 5.686427457098284, "grad_norm": 1.2316080331802368, "learning_rate": 0.00018809895252952974, "loss": 0.0305, "step": 7290 }, { "epoch": 5.694227769110764, "grad_norm": 1.0001583099365234, "learning_rate": 0.00018698462224203254, "loss": 0.0475, "step": 7300 }, { "epoch": 5.702028081123245, "grad_norm": 0.3323913812637329, "learning_rate": 0.00018587029195453534, "loss": 0.0289, "step": 7310 }, { "epoch": 5.709828393135726, "grad_norm": 0.7816438674926758, "learning_rate": 0.00018475596166703812, "loss": 0.0375, "step": 7320 }, { "epoch": 5.717628705148206, "grad_norm": 0.9658516645431519, "learning_rate": 0.0001836416313795409, "loss": 0.0345, "step": 7330 }, { "epoch": 5.725429017160686, "grad_norm": 0.5915180444717407, "learning_rate": 0.0001825273010920437, "loss": 0.0494, "step": 7340 }, { "epoch": 5.733229329173167, "grad_norm": 0.6771920919418335, "learning_rate": 0.00018141297080454647, "loss": 0.049, "step": 7350 }, { "epoch": 5.741029641185648, "grad_norm": 0.3479750454425812, "learning_rate": 0.00018029864051704924, "loss": 0.0706, "step": 7360 }, { "epoch": 5.748829953198128, "grad_norm": 0.5420498251914978, "learning_rate": 0.00017918431022955205, "loss": 0.0448, "step": 7370 }, { "epoch": 5.7566302652106085, "grad_norm": 1.378409504890442, "learning_rate": 0.00017806997994205482, "loss": 0.031, "step": 7380 }, { "epoch": 5.764430577223089, "grad_norm": 1.59050714969635, "learning_rate": 0.00017695564965455762, "loss": 0.0392, "step": 7390 }, { "epoch": 5.772230889235569, "grad_norm": 0.46850696206092834, "learning_rate": 0.0001758413193670604, "loss": 0.0367, "step": 7400 }, { "epoch": 5.78003120124805, "grad_norm": 0.560463011264801, "learning_rate": 0.0001747269890795632, "loss": 0.028, "step": 7410 }, { "epoch": 5.787831513260531, "grad_norm": 0.3987765610218048, "learning_rate": 0.00017361265879206597, "loss": 0.0276, "step": 7420 }, { "epoch": 5.795631825273011, "grad_norm": 1.1987881660461426, "learning_rate": 0.00017249832850456875, "loss": 0.0309, "step": 7430 }, { "epoch": 5.803432137285491, "grad_norm": 1.4303110837936401, "learning_rate": 0.00017138399821707155, "loss": 0.0723, "step": 7440 }, { "epoch": 5.811232449297972, "grad_norm": 0.39652517437934875, "learning_rate": 0.00017026966792957433, "loss": 0.0368, "step": 7450 }, { "epoch": 5.819032761310453, "grad_norm": 2.1430587768554688, "learning_rate": 0.0001691553376420771, "loss": 0.0466, "step": 7460 }, { "epoch": 5.826833073322933, "grad_norm": 0.5818430185317993, "learning_rate": 0.00016804100735457988, "loss": 0.0402, "step": 7470 }, { "epoch": 5.8346333853354135, "grad_norm": 0.9637795090675354, "learning_rate": 0.0001669266770670827, "loss": 0.0419, "step": 7480 }, { "epoch": 5.842433697347894, "grad_norm": 0.8920093178749084, "learning_rate": 0.00016581234677958548, "loss": 0.0428, "step": 7490 }, { "epoch": 5.850234009360374, "grad_norm": 0.36178743839263916, "learning_rate": 0.00016469801649208825, "loss": 0.1173, "step": 7500 }, { "epoch": 5.858034321372855, "grad_norm": 1.9419833421707153, "learning_rate": 0.00016358368620459106, "loss": 0.079, "step": 7510 }, { "epoch": 5.865834633385336, "grad_norm": 0.8884839415550232, "learning_rate": 0.00016246935591709383, "loss": 0.0332, "step": 7520 }, { "epoch": 5.873634945397816, "grad_norm": 0.42523589730262756, "learning_rate": 0.0001613550256295966, "loss": 0.0328, "step": 7530 }, { "epoch": 5.881435257410296, "grad_norm": 0.5401551127433777, "learning_rate": 0.0001602406953420994, "loss": 0.0281, "step": 7540 }, { "epoch": 5.889235569422777, "grad_norm": 0.36174464225769043, "learning_rate": 0.00015912636505460218, "loss": 0.0714, "step": 7550 }, { "epoch": 5.897035881435257, "grad_norm": 0.7682410478591919, "learning_rate": 0.00015801203476710496, "loss": 0.0404, "step": 7560 }, { "epoch": 5.904836193447738, "grad_norm": 0.7188004851341248, "learning_rate": 0.00015689770447960776, "loss": 0.0436, "step": 7570 }, { "epoch": 5.912636505460219, "grad_norm": 1.0970255136489868, "learning_rate": 0.00015578337419211056, "loss": 0.0388, "step": 7580 }, { "epoch": 5.920436817472699, "grad_norm": 0.6196308732032776, "learning_rate": 0.00015466904390461334, "loss": 0.0419, "step": 7590 }, { "epoch": 5.928237129485179, "grad_norm": 0.7572742104530334, "learning_rate": 0.0001535547136171161, "loss": 0.0342, "step": 7600 }, { "epoch": 5.9360374414976596, "grad_norm": 0.1703101396560669, "learning_rate": 0.0001524403833296189, "loss": 0.0336, "step": 7610 }, { "epoch": 5.943837753510141, "grad_norm": 2.856678009033203, "learning_rate": 0.0001513260530421217, "loss": 0.0442, "step": 7620 }, { "epoch": 5.951638065522621, "grad_norm": 0.9317063093185425, "learning_rate": 0.00015021172275462446, "loss": 0.0444, "step": 7630 }, { "epoch": 5.959438377535101, "grad_norm": 0.6945188641548157, "learning_rate": 0.00014909739246712726, "loss": 0.0413, "step": 7640 }, { "epoch": 5.967238689547582, "grad_norm": 0.31758037209510803, "learning_rate": 0.00014798306217963004, "loss": 0.045, "step": 7650 }, { "epoch": 5.975039001560062, "grad_norm": 2.0739245414733887, "learning_rate": 0.00014686873189213281, "loss": 0.0374, "step": 7660 }, { "epoch": 5.982839313572543, "grad_norm": 0.7501590251922607, "learning_rate": 0.00014575440160463562, "loss": 0.0351, "step": 7670 }, { "epoch": 5.990639625585024, "grad_norm": 0.6548338532447815, "learning_rate": 0.00014464007131713842, "loss": 0.0251, "step": 7680 }, { "epoch": 5.998439937597504, "grad_norm": 0.6538023352622986, "learning_rate": 0.0001435257410296412, "loss": 0.0357, "step": 7690 }, { "epoch": 6.0, "eval_loss": 0.3014097511768341, "eval_runtime": 536.393, "eval_samples_per_second": 1.025, "eval_steps_per_second": 1.025, "eval_wer": 18.922571925082064, "step": 7692 } ], "logging_steps": 10, "max_steps": 8974, "num_input_tokens_seen": 0, "num_train_epochs": 7, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 6.77699382509568e+18, "train_batch_size": 3, "trial_name": null, "trial_params": null }