{ "best_metric": null, "best_model_checkpoint": null, "epoch": 5.0, "eval_steps": 3000, "global_step": 566190, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00044154789028418023, "grad_norm": 4.407685279846191, "learning_rate": 4.800000000000001e-06, "loss": 4.2565, "step": 50 }, { "epoch": 0.0008830957805683605, "grad_norm": 2.3940842151641846, "learning_rate": 9.800000000000001e-06, "loss": 4.3171, "step": 100 }, { "epoch": 0.0013246436708525408, "grad_norm": 2.555833339691162, "learning_rate": 1.48e-05, "loss": 3.8795, "step": 150 }, { "epoch": 0.001766191561136721, "grad_norm": 0.9911451935768127, "learning_rate": 1.97e-05, "loss": 3.9588, "step": 200 }, { "epoch": 0.0022077394514209013, "grad_norm": 1.1606439352035522, "learning_rate": 2.47e-05, "loss": 3.4062, "step": 250 }, { "epoch": 0.0026492873417050815, "grad_norm": 2.9307570457458496, "learning_rate": 2.97e-05, "loss": 3.5316, "step": 300 }, { "epoch": 0.0030908352319892617, "grad_norm": 2.6516165733337402, "learning_rate": 3.4699999999999996e-05, "loss": 3.5324, "step": 350 }, { "epoch": 0.003532383122273442, "grad_norm": 1.69759202003479, "learning_rate": 3.97e-05, "loss": 3.6864, "step": 400 }, { "epoch": 0.003973931012557622, "grad_norm": 3.0430002212524414, "learning_rate": 4.47e-05, "loss": 3.6768, "step": 450 }, { "epoch": 0.004415478902841803, "grad_norm": 4.26913595199585, "learning_rate": 4.97e-05, "loss": 3.9357, "step": 500 }, { "epoch": 0.004857026793125983, "grad_norm": 1.9752435684204102, "learning_rate": 5.470000000000001e-05, "loss": 3.4989, "step": 550 }, { "epoch": 0.005298574683410163, "grad_norm": 2.631711006164551, "learning_rate": 5.96e-05, "loss": 3.4577, "step": 600 }, { "epoch": 0.005740122573694343, "grad_norm": 4.840415000915527, "learning_rate": 6.460000000000001e-05, "loss": 3.6496, "step": 650 }, { "epoch": 0.006181670463978523, "grad_norm": 0.7839917540550232, "learning_rate": 6.96e-05, "loss": 3.4212, "step": 700 }, { "epoch": 0.0066232183542627035, "grad_norm": 2.6035537719726562, "learning_rate": 7.46e-05, "loss": 3.5461, "step": 750 }, { "epoch": 0.007064766244546884, "grad_norm": 2.7785916328430176, "learning_rate": 7.960000000000001e-05, "loss": 3.5111, "step": 800 }, { "epoch": 0.007506314134831064, "grad_norm": 3.2137913703918457, "learning_rate": 8.46e-05, "loss": 3.4488, "step": 850 }, { "epoch": 0.007947862025115243, "grad_norm": 2.4404215812683105, "learning_rate": 8.960000000000001e-05, "loss": 3.6674, "step": 900 }, { "epoch": 0.008389409915399423, "grad_norm": 2.055772542953491, "learning_rate": 9.46e-05, "loss": 3.5125, "step": 950 }, { "epoch": 0.008830957805683605, "grad_norm": 6.80244255065918, "learning_rate": 9.960000000000001e-05, "loss": 3.7395, "step": 1000 }, { "epoch": 0.009272505695967785, "grad_norm": 1.6064447164535522, "learning_rate": 9.999999836556943e-05, "loss": 3.2688, "step": 1050 }, { "epoch": 0.009714053586251966, "grad_norm": 1.4385908842086792, "learning_rate": 9.999999288142163e-05, "loss": 3.7015, "step": 1100 }, { "epoch": 0.010155601476536146, "grad_norm": 2.0870840549468994, "learning_rate": 9.999998353519836e-05, "loss": 3.6805, "step": 1150 }, { "epoch": 0.010597149366820326, "grad_norm": 1.5181084871292114, "learning_rate": 9.999997032690033e-05, "loss": 3.9943, "step": 1200 }, { "epoch": 0.011038697257104506, "grad_norm": 2.0641415119171143, "learning_rate": 9.99999532565286e-05, "loss": 3.4652, "step": 1250 }, { "epoch": 0.011480245147388686, "grad_norm": 4.051513671875, "learning_rate": 9.999993232408446e-05, "loss": 3.784, "step": 1300 }, { "epoch": 0.011921793037672867, "grad_norm": 2.475578546524048, "learning_rate": 9.999990752956954e-05, "loss": 3.6213, "step": 1350 }, { "epoch": 0.012363340927957047, "grad_norm": 2.650278091430664, "learning_rate": 9.999987887298574e-05, "loss": 3.2759, "step": 1400 }, { "epoch": 0.012804888818241227, "grad_norm": 1.043623924255371, "learning_rate": 9.999984635433529e-05, "loss": 3.4098, "step": 1450 }, { "epoch": 0.013246436708525407, "grad_norm": 2.3483691215515137, "learning_rate": 9.999980997362069e-05, "loss": 3.5984, "step": 1500 }, { "epoch": 0.013687984598809587, "grad_norm": 1.8058457374572754, "learning_rate": 9.999976973084477e-05, "loss": 3.5686, "step": 1550 }, { "epoch": 0.014129532489093767, "grad_norm": 1.9905356168746948, "learning_rate": 9.999972562601062e-05, "loss": 3.5255, "step": 1600 }, { "epoch": 0.014571080379377948, "grad_norm": 8.096346855163574, "learning_rate": 9.999967765912164e-05, "loss": 3.8234, "step": 1650 }, { "epoch": 0.015012628269662128, "grad_norm": 3.1411712169647217, "learning_rate": 9.999962583018154e-05, "loss": 3.3328, "step": 1700 }, { "epoch": 0.015454176159946308, "grad_norm": 4.46273946762085, "learning_rate": 9.999957013919435e-05, "loss": 3.2951, "step": 1750 }, { "epoch": 0.015895724050230486, "grad_norm": 1.3609850406646729, "learning_rate": 9.999951058616435e-05, "loss": 3.3173, "step": 1800 }, { "epoch": 0.016337271940514667, "grad_norm": 0.8974264860153198, "learning_rate": 9.999944717109613e-05, "loss": 3.5905, "step": 1850 }, { "epoch": 0.016778819830798847, "grad_norm": 4.310591220855713, "learning_rate": 9.999937989399459e-05, "loss": 3.2723, "step": 1900 }, { "epoch": 0.017220367721083027, "grad_norm": 10.745691299438477, "learning_rate": 9.999930875486495e-05, "loss": 3.3104, "step": 1950 }, { "epoch": 0.01766191561136721, "grad_norm": 1.678096890449524, "learning_rate": 9.999923375371269e-05, "loss": 3.4128, "step": 2000 }, { "epoch": 0.01810346350165139, "grad_norm": 1.2710167169570923, "learning_rate": 9.99991548905436e-05, "loss": 3.8629, "step": 2050 }, { "epoch": 0.01854501139193557, "grad_norm": 1.6546989679336548, "learning_rate": 9.999907216536378e-05, "loss": 3.9797, "step": 2100 }, { "epoch": 0.01898655928221975, "grad_norm": 2.4010069370269775, "learning_rate": 9.999898557817962e-05, "loss": 3.4582, "step": 2150 }, { "epoch": 0.01942810717250393, "grad_norm": 11.584942817687988, "learning_rate": 9.999889512899778e-05, "loss": 3.4392, "step": 2200 }, { "epoch": 0.01986965506278811, "grad_norm": 2.588017702102661, "learning_rate": 9.999880081782529e-05, "loss": 3.2102, "step": 2250 }, { "epoch": 0.02031120295307229, "grad_norm": 4.959451198577881, "learning_rate": 9.99987026446694e-05, "loss": 3.4403, "step": 2300 }, { "epoch": 0.020752750843356472, "grad_norm": 2.886789560317993, "learning_rate": 9.999860060953772e-05, "loss": 3.3001, "step": 2350 }, { "epoch": 0.021194298733640652, "grad_norm": 7.682253837585449, "learning_rate": 9.999849471243812e-05, "loss": 3.8255, "step": 2400 }, { "epoch": 0.021635846623924832, "grad_norm": 9.694388389587402, "learning_rate": 9.999838495337877e-05, "loss": 3.8851, "step": 2450 }, { "epoch": 0.022077394514209012, "grad_norm": 2.9934961795806885, "learning_rate": 9.999827133236815e-05, "loss": 3.6424, "step": 2500 }, { "epoch": 0.022518942404493193, "grad_norm": 1.4004652500152588, "learning_rate": 9.999815384941506e-05, "loss": 3.5799, "step": 2550 }, { "epoch": 0.022960490294777373, "grad_norm": 2.06247615814209, "learning_rate": 9.999803250452856e-05, "loss": 3.4358, "step": 2600 }, { "epoch": 0.023402038185061553, "grad_norm": 2.051114082336426, "learning_rate": 9.999790729771798e-05, "loss": 3.6721, "step": 2650 }, { "epoch": 0.023843586075345733, "grad_norm": 2.4515838623046875, "learning_rate": 9.999777822899307e-05, "loss": 3.6419, "step": 2700 }, { "epoch": 0.024285133965629913, "grad_norm": 2.6057190895080566, "learning_rate": 9.999764529836375e-05, "loss": 3.4357, "step": 2750 }, { "epoch": 0.024726681855914093, "grad_norm": 2.1552577018737793, "learning_rate": 9.99975085058403e-05, "loss": 3.6189, "step": 2800 }, { "epoch": 0.025168229746198274, "grad_norm": 4.0472846031188965, "learning_rate": 9.999736785143327e-05, "loss": 3.5135, "step": 2850 }, { "epoch": 0.025609777636482454, "grad_norm": 1.694504976272583, "learning_rate": 9.999722333515355e-05, "loss": 3.5287, "step": 2900 }, { "epoch": 0.026051325526766634, "grad_norm": 8.814433097839355, "learning_rate": 9.99970749570123e-05, "loss": 3.4025, "step": 2950 }, { "epoch": 0.026492873417050814, "grad_norm": 2.079963207244873, "learning_rate": 9.999692579966685e-05, "loss": 3.8211, "step": 3000 }, { "epoch": 0.026492873417050814, "eval_asr_loss": 0.8104195239432453, "eval_loss": 3.4588043689727783, "eval_runtime": 20.2586, "eval_samples_per_second": 37.91, "eval_steps_per_second": 9.477, "eval_tts_loss": 6.061187094347814, "step": 3000 }, { "epoch": 0.026934421307334994, "grad_norm": 2.6620876789093018, "learning_rate": 9.999676977507384e-05, "loss": 3.4284, "step": 3050 }, { "epoch": 0.027375969197619174, "grad_norm": 10.126931190490723, "learning_rate": 9.999660988865433e-05, "loss": 3.756, "step": 3100 }, { "epoch": 0.027817517087903355, "grad_norm": 1.450592279434204, "learning_rate": 9.999644614042068e-05, "loss": 3.2147, "step": 3150 }, { "epoch": 0.028259064978187535, "grad_norm": 1.703476071357727, "learning_rate": 9.99962785303855e-05, "loss": 3.5119, "step": 3200 }, { "epoch": 0.028700612868471715, "grad_norm": 2.767446756362915, "learning_rate": 9.999610705856177e-05, "loss": 3.2736, "step": 3250 }, { "epoch": 0.029142160758755895, "grad_norm": 1.1504226922988892, "learning_rate": 9.999593172496273e-05, "loss": 3.4068, "step": 3300 }, { "epoch": 0.029583708649040075, "grad_norm": 1.6964422464370728, "learning_rate": 9.99957525296019e-05, "loss": 3.4917, "step": 3350 }, { "epoch": 0.030025256539324256, "grad_norm": 1.8230127096176147, "learning_rate": 9.999556947249316e-05, "loss": 3.1574, "step": 3400 }, { "epoch": 0.030466804429608436, "grad_norm": 3.8012404441833496, "learning_rate": 9.999538255365062e-05, "loss": 3.6179, "step": 3450 }, { "epoch": 0.030908352319892616, "grad_norm": 3.4073257446289062, "learning_rate": 9.999519177308871e-05, "loss": 3.6502, "step": 3500 }, { "epoch": 0.03134990021017679, "grad_norm": 1.7333426475524902, "learning_rate": 9.99949971308222e-05, "loss": 3.5175, "step": 3550 }, { "epoch": 0.03179144810046097, "grad_norm": 1.5845162868499756, "learning_rate": 9.99947986268661e-05, "loss": 3.4373, "step": 3600 }, { "epoch": 0.03223299599074515, "grad_norm": 4.4948225021362305, "learning_rate": 9.999459626123576e-05, "loss": 3.6571, "step": 3650 }, { "epoch": 0.03267454388102933, "grad_norm": 3.1738150119781494, "learning_rate": 9.99943900339468e-05, "loss": 3.6251, "step": 3700 }, { "epoch": 0.03311609177131351, "grad_norm": 2.0190846920013428, "learning_rate": 9.999417994501514e-05, "loss": 3.5417, "step": 3750 }, { "epoch": 0.033557639661597694, "grad_norm": 3.240422248840332, "learning_rate": 9.999396599445701e-05, "loss": 3.5891, "step": 3800 }, { "epoch": 0.033999187551881874, "grad_norm": 5.438379287719727, "learning_rate": 9.999374818228895e-05, "loss": 3.6557, "step": 3850 }, { "epoch": 0.034440735442166054, "grad_norm": 1.6964021921157837, "learning_rate": 9.999352650852778e-05, "loss": 3.4464, "step": 3900 }, { "epoch": 0.03488228333245024, "grad_norm": 1.759133219718933, "learning_rate": 9.999330097319063e-05, "loss": 3.5975, "step": 3950 }, { "epoch": 0.03532383122273442, "grad_norm": 1.7858753204345703, "learning_rate": 9.999307157629489e-05, "loss": 3.8846, "step": 4000 }, { "epoch": 0.0357653791130186, "grad_norm": 1.641852855682373, "learning_rate": 9.999283831785833e-05, "loss": 3.7919, "step": 4050 }, { "epoch": 0.03620692700330278, "grad_norm": 1.8461246490478516, "learning_rate": 9.99926011978989e-05, "loss": 3.2686, "step": 4100 }, { "epoch": 0.03664847489358696, "grad_norm": 2.8383119106292725, "learning_rate": 9.999236021643498e-05, "loss": 3.1711, "step": 4150 }, { "epoch": 0.03709002278387114, "grad_norm": 1.094404935836792, "learning_rate": 9.999211537348514e-05, "loss": 3.4461, "step": 4200 }, { "epoch": 0.03753157067415532, "grad_norm": 1.1627689599990845, "learning_rate": 9.999186666906832e-05, "loss": 3.2783, "step": 4250 }, { "epoch": 0.0379731185644395, "grad_norm": 2.673081398010254, "learning_rate": 9.999161410320373e-05, "loss": 3.3066, "step": 4300 }, { "epoch": 0.03841466645472368, "grad_norm": 2.3519914150238037, "learning_rate": 9.999135767591083e-05, "loss": 3.5882, "step": 4350 }, { "epoch": 0.03885621434500786, "grad_norm": 2.861649990081787, "learning_rate": 9.999109738720949e-05, "loss": 3.3404, "step": 4400 }, { "epoch": 0.03929776223529204, "grad_norm": 2.392594575881958, "learning_rate": 9.999083323711979e-05, "loss": 3.4555, "step": 4450 }, { "epoch": 0.03973931012557622, "grad_norm": 2.1783764362335205, "learning_rate": 9.999056522566213e-05, "loss": 3.3032, "step": 4500 }, { "epoch": 0.0401808580158604, "grad_norm": 3.0805599689483643, "learning_rate": 9.99902933528572e-05, "loss": 3.7245, "step": 4550 }, { "epoch": 0.04062240590614458, "grad_norm": 0.8756254315376282, "learning_rate": 9.999001761872601e-05, "loss": 3.0947, "step": 4600 }, { "epoch": 0.041063953796428763, "grad_norm": 8.237518310546875, "learning_rate": 9.998973802328988e-05, "loss": 3.369, "step": 4650 }, { "epoch": 0.041505501686712944, "grad_norm": 1.5842547416687012, "learning_rate": 9.998945456657039e-05, "loss": 3.6075, "step": 4700 }, { "epoch": 0.041947049576997124, "grad_norm": 2.602689504623413, "learning_rate": 9.998916724858942e-05, "loss": 3.7514, "step": 4750 }, { "epoch": 0.042388597467281304, "grad_norm": 3.264051914215088, "learning_rate": 9.998887606936918e-05, "loss": 3.435, "step": 4800 }, { "epoch": 0.042830145357565484, "grad_norm": 2.2757341861724854, "learning_rate": 9.998858102893215e-05, "loss": 3.6671, "step": 4850 }, { "epoch": 0.043271693247849664, "grad_norm": 1.5097368955612183, "learning_rate": 9.998828212730115e-05, "loss": 3.4472, "step": 4900 }, { "epoch": 0.043713241138133845, "grad_norm": 2.7084147930145264, "learning_rate": 9.998797936449922e-05, "loss": 3.8243, "step": 4950 }, { "epoch": 0.044154789028418025, "grad_norm": 2.9833407402038574, "learning_rate": 9.998767274054978e-05, "loss": 3.7378, "step": 5000 }, { "epoch": 0.044596336918702205, "grad_norm": 9.48874568939209, "learning_rate": 9.998736225547651e-05, "loss": 3.4267, "step": 5050 }, { "epoch": 0.045037884808986385, "grad_norm": 1.7265108823776245, "learning_rate": 9.998704790930337e-05, "loss": 3.4905, "step": 5100 }, { "epoch": 0.045479432699270565, "grad_norm": 1.690320372581482, "learning_rate": 9.998672970205467e-05, "loss": 3.4667, "step": 5150 }, { "epoch": 0.045920980589554745, "grad_norm": 3.318742275238037, "learning_rate": 9.998640763375497e-05, "loss": 3.4528, "step": 5200 }, { "epoch": 0.046362528479838926, "grad_norm": 3.2868552207946777, "learning_rate": 9.998608170442915e-05, "loss": 3.5877, "step": 5250 }, { "epoch": 0.046804076370123106, "grad_norm": 3.8400142192840576, "learning_rate": 9.99857519141024e-05, "loss": 3.5318, "step": 5300 }, { "epoch": 0.047245624260407286, "grad_norm": 1.9866262674331665, "learning_rate": 9.998541826280018e-05, "loss": 3.7204, "step": 5350 }, { "epoch": 0.047687172150691466, "grad_norm": 2.3364500999450684, "learning_rate": 9.998508075054826e-05, "loss": 3.3647, "step": 5400 }, { "epoch": 0.048128720040975646, "grad_norm": 3.461249589920044, "learning_rate": 9.99847393773727e-05, "loss": 3.4072, "step": 5450 }, { "epoch": 0.048570267931259826, "grad_norm": 1.401358962059021, "learning_rate": 9.99843941432999e-05, "loss": 3.506, "step": 5500 }, { "epoch": 0.04901181582154401, "grad_norm": 2.5606305599212646, "learning_rate": 9.99840450483565e-05, "loss": 3.3622, "step": 5550 }, { "epoch": 0.04945336371182819, "grad_norm": 2.909435510635376, "learning_rate": 9.998369209256947e-05, "loss": 3.1674, "step": 5600 }, { "epoch": 0.04989491160211237, "grad_norm": 1.5808318853378296, "learning_rate": 9.998333527596607e-05, "loss": 3.5161, "step": 5650 }, { "epoch": 0.05033645949239655, "grad_norm": 1.6047215461730957, "learning_rate": 9.998297459857387e-05, "loss": 3.528, "step": 5700 }, { "epoch": 0.05077800738268073, "grad_norm": 2.816723346710205, "learning_rate": 9.998261006042074e-05, "loss": 3.1358, "step": 5750 }, { "epoch": 0.05121955527296491, "grad_norm": 1.9394768476486206, "learning_rate": 9.99822416615348e-05, "loss": 2.8468, "step": 5800 }, { "epoch": 0.05166110316324909, "grad_norm": 7.701488971710205, "learning_rate": 9.998186940194454e-05, "loss": 3.5079, "step": 5850 }, { "epoch": 0.05210265105353327, "grad_norm": 1.9361647367477417, "learning_rate": 9.998150084191847e-05, "loss": 3.1601, "step": 5900 }, { "epoch": 0.05254419894381745, "grad_norm": 1.7290771007537842, "learning_rate": 9.998112093821874e-05, "loss": 3.5756, "step": 5950 }, { "epoch": 0.05298574683410163, "grad_norm": 7.283807754516602, "learning_rate": 9.998073717390126e-05, "loss": 3.3956, "step": 6000 }, { "epoch": 0.05298574683410163, "eval_asr_loss": 0.8455491743058053, "eval_loss": 3.3733489513397217, "eval_runtime": 20.0292, "eval_samples_per_second": 38.344, "eval_steps_per_second": 9.586, "eval_tts_loss": 5.977993885297463, "step": 6000 }, { "epoch": 0.05342729472438581, "grad_norm": 2.8105592727661133, "learning_rate": 9.998034954899564e-05, "loss": 3.5391, "step": 6050 }, { "epoch": 0.05386884261466999, "grad_norm": 2.575618267059326, "learning_rate": 9.997995806353187e-05, "loss": 3.4848, "step": 6100 }, { "epoch": 0.05431039050495417, "grad_norm": 1.944770336151123, "learning_rate": 9.997956271754013e-05, "loss": 3.4319, "step": 6150 }, { "epoch": 0.05475193839523835, "grad_norm": 2.9254565238952637, "learning_rate": 9.997916351105098e-05, "loss": 3.5138, "step": 6200 }, { "epoch": 0.05519348628552253, "grad_norm": 2.876880645751953, "learning_rate": 9.997876044409529e-05, "loss": 3.2436, "step": 6250 }, { "epoch": 0.05563503417580671, "grad_norm": 1.5080561637878418, "learning_rate": 9.997835351670415e-05, "loss": 3.3394, "step": 6300 }, { "epoch": 0.05607658206609089, "grad_norm": 2.99318790435791, "learning_rate": 9.9977942728909e-05, "loss": 3.2061, "step": 6350 }, { "epoch": 0.05651812995637507, "grad_norm": 2.1895508766174316, "learning_rate": 9.997752808074157e-05, "loss": 3.586, "step": 6400 }, { "epoch": 0.05695967784665925, "grad_norm": 2.089017868041992, "learning_rate": 9.997710957223389e-05, "loss": 3.6418, "step": 6450 }, { "epoch": 0.05740122573694343, "grad_norm": 1.3540375232696533, "learning_rate": 9.997668720341831e-05, "loss": 3.5735, "step": 6500 }, { "epoch": 0.05784277362722761, "grad_norm": 2.3487038612365723, "learning_rate": 9.997626097432741e-05, "loss": 3.4285, "step": 6550 }, { "epoch": 0.05828432151751179, "grad_norm": 2.5517020225524902, "learning_rate": 9.997583088499415e-05, "loss": 3.5823, "step": 6600 }, { "epoch": 0.05872586940779597, "grad_norm": 1.7537412643432617, "learning_rate": 9.997539693545174e-05, "loss": 3.1481, "step": 6650 }, { "epoch": 0.05916741729808015, "grad_norm": 2.911041736602783, "learning_rate": 9.99749591257337e-05, "loss": 3.7261, "step": 6700 }, { "epoch": 0.05960896518836433, "grad_norm": 7.761764049530029, "learning_rate": 9.997451745587382e-05, "loss": 3.8374, "step": 6750 }, { "epoch": 0.06005051307864851, "grad_norm": 6.103822708129883, "learning_rate": 9.997407192590625e-05, "loss": 3.1474, "step": 6800 }, { "epoch": 0.06049206096893269, "grad_norm": 8.443099021911621, "learning_rate": 9.99736225358654e-05, "loss": 3.4926, "step": 6850 }, { "epoch": 0.06093360885921687, "grad_norm": 6.040070056915283, "learning_rate": 9.997316928578595e-05, "loss": 3.1745, "step": 6900 }, { "epoch": 0.06137515674950105, "grad_norm": 1.7983872890472412, "learning_rate": 9.997271217570296e-05, "loss": 3.5779, "step": 6950 }, { "epoch": 0.06181670463978523, "grad_norm": 2.881488800048828, "learning_rate": 9.99722512056517e-05, "loss": 3.6113, "step": 7000 }, { "epoch": 0.06225825253006941, "grad_norm": 2.0535004138946533, "learning_rate": 9.997178637566777e-05, "loss": 3.5762, "step": 7050 }, { "epoch": 0.06269980042035359, "grad_norm": 2.7159841060638428, "learning_rate": 9.997131768578711e-05, "loss": 3.4012, "step": 7100 }, { "epoch": 0.06314134831063777, "grad_norm": 3.324490785598755, "learning_rate": 9.997084513604591e-05, "loss": 3.8183, "step": 7150 }, { "epoch": 0.06358289620092195, "grad_norm": 1.6304800510406494, "learning_rate": 9.997036872648064e-05, "loss": 3.4161, "step": 7200 }, { "epoch": 0.06402444409120613, "grad_norm": 2.0595145225524902, "learning_rate": 9.996988845712813e-05, "loss": 3.4944, "step": 7250 }, { "epoch": 0.0644659919814903, "grad_norm": 2.9137229919433594, "learning_rate": 9.996940432802548e-05, "loss": 3.6446, "step": 7300 }, { "epoch": 0.06490753987177449, "grad_norm": 2.6019785404205322, "learning_rate": 9.996891633921007e-05, "loss": 3.5755, "step": 7350 }, { "epoch": 0.06534908776205867, "grad_norm": 1.3080384731292725, "learning_rate": 9.99684244907196e-05, "loss": 3.432, "step": 7400 }, { "epoch": 0.06579063565234285, "grad_norm": 1.085084080696106, "learning_rate": 9.996792878259204e-05, "loss": 3.5939, "step": 7450 }, { "epoch": 0.06623218354262703, "grad_norm": 2.416012763977051, "learning_rate": 9.996742921486573e-05, "loss": 3.3125, "step": 7500 }, { "epoch": 0.0666737314329112, "grad_norm": 2.329331159591675, "learning_rate": 9.99669257875792e-05, "loss": 3.4727, "step": 7550 }, { "epoch": 0.06711527932319539, "grad_norm": 1.2133623361587524, "learning_rate": 9.996641850077135e-05, "loss": 3.6817, "step": 7600 }, { "epoch": 0.06755682721347957, "grad_norm": 3.447230577468872, "learning_rate": 9.99659073544814e-05, "loss": 3.3487, "step": 7650 }, { "epoch": 0.06799837510376375, "grad_norm": 1.8407869338989258, "learning_rate": 9.99653923487488e-05, "loss": 3.4952, "step": 7700 }, { "epoch": 0.06843992299404793, "grad_norm": 3.4017183780670166, "learning_rate": 9.996487348361331e-05, "loss": 3.5987, "step": 7750 }, { "epoch": 0.06888147088433211, "grad_norm": 2.014235734939575, "learning_rate": 9.996435075911507e-05, "loss": 3.4913, "step": 7800 }, { "epoch": 0.0693230187746163, "grad_norm": 8.113083839416504, "learning_rate": 9.99638241752944e-05, "loss": 3.2701, "step": 7850 }, { "epoch": 0.06976456666490048, "grad_norm": 3.6530191898345947, "learning_rate": 9.996329373219199e-05, "loss": 3.7682, "step": 7900 }, { "epoch": 0.07020611455518466, "grad_norm": 3.110872507095337, "learning_rate": 9.996275942984882e-05, "loss": 2.9624, "step": 7950 }, { "epoch": 0.07064766244546884, "grad_norm": 5.682243347167969, "learning_rate": 9.996222126830616e-05, "loss": 3.1075, "step": 8000 }, { "epoch": 0.07108921033575302, "grad_norm": 8.339673042297363, "learning_rate": 9.996167924760556e-05, "loss": 3.4832, "step": 8050 }, { "epoch": 0.0715307582260372, "grad_norm": 2.9629604816436768, "learning_rate": 9.996113336778891e-05, "loss": 3.5399, "step": 8100 }, { "epoch": 0.07197230611632138, "grad_norm": 4.072164058685303, "learning_rate": 9.996058362889837e-05, "loss": 3.3337, "step": 8150 }, { "epoch": 0.07241385400660556, "grad_norm": 2.1256349086761475, "learning_rate": 9.99600300309764e-05, "loss": 3.4296, "step": 8200 }, { "epoch": 0.07285540189688974, "grad_norm": 0.9664581418037415, "learning_rate": 9.995947257406575e-05, "loss": 3.3867, "step": 8250 }, { "epoch": 0.07329694978717392, "grad_norm": 2.7981796264648438, "learning_rate": 9.995891125820948e-05, "loss": 3.5777, "step": 8300 }, { "epoch": 0.0737384976774581, "grad_norm": 2.2892305850982666, "learning_rate": 9.995834608345098e-05, "loss": 3.448, "step": 8350 }, { "epoch": 0.07418004556774228, "grad_norm": 3.462890386581421, "learning_rate": 9.995777704983387e-05, "loss": 3.6835, "step": 8400 }, { "epoch": 0.07462159345802646, "grad_norm": 2.887977123260498, "learning_rate": 9.995720415740209e-05, "loss": 3.8102, "step": 8450 }, { "epoch": 0.07506314134831064, "grad_norm": 1.103590726852417, "learning_rate": 9.995662740619993e-05, "loss": 3.3465, "step": 8500 }, { "epoch": 0.07550468923859482, "grad_norm": 1.5264739990234375, "learning_rate": 9.995604679627193e-05, "loss": 3.3524, "step": 8550 }, { "epoch": 0.075946237128879, "grad_norm": 10.508204460144043, "learning_rate": 9.995546232766293e-05, "loss": 2.9848, "step": 8600 }, { "epoch": 0.07638778501916318, "grad_norm": 11.036320686340332, "learning_rate": 9.995487400041806e-05, "loss": 3.6104, "step": 8650 }, { "epoch": 0.07682933290944736, "grad_norm": 0.900551974773407, "learning_rate": 9.995428181458279e-05, "loss": 3.2115, "step": 8700 }, { "epoch": 0.07727088079973154, "grad_norm": 1.0776323080062866, "learning_rate": 9.995368577020285e-05, "loss": 3.378, "step": 8750 }, { "epoch": 0.07771242869001573, "grad_norm": 1.9495457410812378, "learning_rate": 9.995308586732427e-05, "loss": 3.134, "step": 8800 }, { "epoch": 0.0781539765802999, "grad_norm": 2.638998508453369, "learning_rate": 9.99524821059934e-05, "loss": 3.158, "step": 8850 }, { "epoch": 0.07859552447058409, "grad_norm": 1.9697933197021484, "learning_rate": 9.995187448625688e-05, "loss": 3.4271, "step": 8900 }, { "epoch": 0.07903707236086827, "grad_norm": 1.1985995769500732, "learning_rate": 9.995126300816163e-05, "loss": 3.3212, "step": 8950 }, { "epoch": 0.07947862025115245, "grad_norm": 5.711977005004883, "learning_rate": 9.995064767175488e-05, "loss": 3.256, "step": 9000 }, { "epoch": 0.07947862025115245, "eval_asr_loss": 0.8664082059474173, "eval_loss": 3.2935588359832764, "eval_runtime": 19.9624, "eval_samples_per_second": 38.472, "eval_steps_per_second": 9.618, "eval_tts_loss": 5.950566485041761, "step": 9000 }, { "epoch": 0.07992016814143663, "grad_norm": 2.0665061473846436, "learning_rate": 9.995002847708418e-05, "loss": 3.5921, "step": 9050 }, { "epoch": 0.0803617160317208, "grad_norm": 2.0567665100097656, "learning_rate": 9.994940542419733e-05, "loss": 3.4312, "step": 9100 }, { "epoch": 0.08080326392200499, "grad_norm": 5.3886847496032715, "learning_rate": 9.994879108917332e-05, "loss": 3.2038, "step": 9150 }, { "epoch": 0.08124481181228917, "grad_norm": 5.97287130355835, "learning_rate": 9.994816039716079e-05, "loss": 3.4363, "step": 9200 }, { "epoch": 0.08168635970257335, "grad_norm": 3.153740406036377, "learning_rate": 9.994752584707642e-05, "loss": 3.3041, "step": 9250 }, { "epoch": 0.08212790759285753, "grad_norm": 2.116394519805908, "learning_rate": 9.994688743896923e-05, "loss": 3.906, "step": 9300 }, { "epoch": 0.08256945548314171, "grad_norm": 2.2864694595336914, "learning_rate": 9.994624517288851e-05, "loss": 3.5515, "step": 9350 }, { "epoch": 0.08301100337342589, "grad_norm": 6.8061723709106445, "learning_rate": 9.994559904888388e-05, "loss": 3.5991, "step": 9400 }, { "epoch": 0.08345255126371007, "grad_norm": 5.035069465637207, "learning_rate": 9.994494906700527e-05, "loss": 3.7135, "step": 9450 }, { "epoch": 0.08389409915399425, "grad_norm": 2.2650699615478516, "learning_rate": 9.994429522730284e-05, "loss": 3.3555, "step": 9500 }, { "epoch": 0.08433564704427843, "grad_norm": 3.1491940021514893, "learning_rate": 9.994363752982714e-05, "loss": 3.0579, "step": 9550 }, { "epoch": 0.08477719493456261, "grad_norm": 3.5358498096466064, "learning_rate": 9.994297597462894e-05, "loss": 3.6277, "step": 9600 }, { "epoch": 0.08521874282484679, "grad_norm": 3.785895824432373, "learning_rate": 9.994231056175936e-05, "loss": 3.2452, "step": 9650 }, { "epoch": 0.08566029071513097, "grad_norm": 1.632702350616455, "learning_rate": 9.994164129126977e-05, "loss": 3.1927, "step": 9700 }, { "epoch": 0.08610183860541515, "grad_norm": 3.5743367671966553, "learning_rate": 9.99409681632119e-05, "loss": 3.6792, "step": 9750 }, { "epoch": 0.08654338649569933, "grad_norm": 9.39389419555664, "learning_rate": 9.994029117763773e-05, "loss": 3.1351, "step": 9800 }, { "epoch": 0.08698493438598351, "grad_norm": 2.9739062786102295, "learning_rate": 9.993961033459953e-05, "loss": 3.1706, "step": 9850 }, { "epoch": 0.08742648227626769, "grad_norm": 3.320171356201172, "learning_rate": 9.99389256341499e-05, "loss": 3.6547, "step": 9900 }, { "epoch": 0.08786803016655187, "grad_norm": 8.49543285369873, "learning_rate": 9.993823707634176e-05, "loss": 3.701, "step": 9950 }, { "epoch": 0.08830957805683605, "grad_norm": 1.3553107976913452, "learning_rate": 9.993754466122827e-05, "loss": 3.4541, "step": 10000 }, { "epoch": 0.08875112594712023, "grad_norm": 1.0776971578598022, "learning_rate": 9.993684838886289e-05, "loss": 3.7029, "step": 10050 }, { "epoch": 0.08919267383740441, "grad_norm": 1.6787606477737427, "learning_rate": 9.993614825929945e-05, "loss": 3.6013, "step": 10100 }, { "epoch": 0.08963422172768859, "grad_norm": 1.4465861320495605, "learning_rate": 9.993544427259198e-05, "loss": 3.2941, "step": 10150 }, { "epoch": 0.09007576961797277, "grad_norm": 3.221482276916504, "learning_rate": 9.99347364287949e-05, "loss": 3.5483, "step": 10200 }, { "epoch": 0.09051731750825695, "grad_norm": 1.0219640731811523, "learning_rate": 9.993402472796284e-05, "loss": 3.2682, "step": 10250 }, { "epoch": 0.09095886539854113, "grad_norm": 4.225593090057373, "learning_rate": 9.993330917015082e-05, "loss": 3.4346, "step": 10300 }, { "epoch": 0.09140041328882531, "grad_norm": 1.145766019821167, "learning_rate": 9.993258975541408e-05, "loss": 3.5205, "step": 10350 }, { "epoch": 0.09184196117910949, "grad_norm": 4.837847709655762, "learning_rate": 9.99318664838082e-05, "loss": 3.3985, "step": 10400 }, { "epoch": 0.09228350906939367, "grad_norm": 3.118101119995117, "learning_rate": 9.993113935538903e-05, "loss": 3.6283, "step": 10450 }, { "epoch": 0.09272505695967785, "grad_norm": 3.6168124675750732, "learning_rate": 9.993040837021277e-05, "loss": 3.1753, "step": 10500 }, { "epoch": 0.09316660484996203, "grad_norm": 3.866116523742676, "learning_rate": 9.992967352833584e-05, "loss": 3.1334, "step": 10550 }, { "epoch": 0.09360815274024621, "grad_norm": 2.1199982166290283, "learning_rate": 9.992893482981505e-05, "loss": 3.3849, "step": 10600 }, { "epoch": 0.09404970063053039, "grad_norm": 1.948744297027588, "learning_rate": 9.99281922747074e-05, "loss": 3.3414, "step": 10650 }, { "epoch": 0.09449124852081457, "grad_norm": 5.387505054473877, "learning_rate": 9.99274458630703e-05, "loss": 3.6585, "step": 10700 }, { "epoch": 0.09493279641109875, "grad_norm": 2.675678253173828, "learning_rate": 9.992669559496136e-05, "loss": 3.1715, "step": 10750 }, { "epoch": 0.09537434430138293, "grad_norm": 1.5642513036727905, "learning_rate": 9.992594147043856e-05, "loss": 3.2574, "step": 10800 }, { "epoch": 0.09581589219166711, "grad_norm": 2.7304491996765137, "learning_rate": 9.992518348956014e-05, "loss": 3.547, "step": 10850 }, { "epoch": 0.09625744008195129, "grad_norm": 3.1260108947753906, "learning_rate": 9.992442165238465e-05, "loss": 3.6459, "step": 10900 }, { "epoch": 0.09669898797223547, "grad_norm": 1.5745757818222046, "learning_rate": 9.992365595897092e-05, "loss": 3.0718, "step": 10950 }, { "epoch": 0.09714053586251965, "grad_norm": 4.948554515838623, "learning_rate": 9.992288640937812e-05, "loss": 2.9318, "step": 11000 }, { "epoch": 0.09758208375280383, "grad_norm": 3.8740906715393066, "learning_rate": 9.992211300366568e-05, "loss": 3.5334, "step": 11050 }, { "epoch": 0.09802363164308801, "grad_norm": 4.66175651550293, "learning_rate": 9.992133574189335e-05, "loss": 3.4052, "step": 11100 }, { "epoch": 0.0984651795333722, "grad_norm": 3.1660337448120117, "learning_rate": 9.992055462412113e-05, "loss": 3.4354, "step": 11150 }, { "epoch": 0.09890672742365637, "grad_norm": 3.1641385555267334, "learning_rate": 9.99197696504094e-05, "loss": 3.4509, "step": 11200 }, { "epoch": 0.09934827531394055, "grad_norm": 2.3828437328338623, "learning_rate": 9.991898082081874e-05, "loss": 3.1644, "step": 11250 }, { "epoch": 0.09978982320422473, "grad_norm": 3.6782753467559814, "learning_rate": 9.991818813541014e-05, "loss": 3.6032, "step": 11300 }, { "epoch": 0.10023137109450891, "grad_norm": 1.4960347414016724, "learning_rate": 9.991739159424481e-05, "loss": 3.3263, "step": 11350 }, { "epoch": 0.1006729189847931, "grad_norm": 3.1267852783203125, "learning_rate": 9.991659119738423e-05, "loss": 3.4492, "step": 11400 }, { "epoch": 0.10111446687507727, "grad_norm": 1.307042121887207, "learning_rate": 9.991578694489028e-05, "loss": 3.2172, "step": 11450 }, { "epoch": 0.10155601476536145, "grad_norm": 2.0429134368896484, "learning_rate": 9.991497883682506e-05, "loss": 3.171, "step": 11500 }, { "epoch": 0.10199756265564563, "grad_norm": 2.784212350845337, "learning_rate": 9.991416687325101e-05, "loss": 3.4447, "step": 11550 }, { "epoch": 0.10243911054592982, "grad_norm": 6.423585414886475, "learning_rate": 9.991335105423081e-05, "loss": 3.2792, "step": 11600 }, { "epoch": 0.102880658436214, "grad_norm": 2.206615924835205, "learning_rate": 9.99125313798275e-05, "loss": 3.2817, "step": 11650 }, { "epoch": 0.10332220632649818, "grad_norm": 4.011673450469971, "learning_rate": 9.991170785010438e-05, "loss": 3.692, "step": 11700 }, { "epoch": 0.10376375421678236, "grad_norm": 1.4347039461135864, "learning_rate": 9.991088046512507e-05, "loss": 3.3256, "step": 11750 }, { "epoch": 0.10420530210706654, "grad_norm": 3.251509428024292, "learning_rate": 9.991004922495348e-05, "loss": 3.3607, "step": 11800 }, { "epoch": 0.10464684999735072, "grad_norm": 5.043297290802002, "learning_rate": 9.990921412965381e-05, "loss": 3.0377, "step": 11850 }, { "epoch": 0.1050883978876349, "grad_norm": 1.7166991233825684, "learning_rate": 9.990837517929057e-05, "loss": 3.2673, "step": 11900 }, { "epoch": 0.10552994577791908, "grad_norm": 2.2065978050231934, "learning_rate": 9.990753237392854e-05, "loss": 3.5916, "step": 11950 }, { "epoch": 0.10597149366820326, "grad_norm": 3.4584174156188965, "learning_rate": 9.990668571363286e-05, "loss": 3.1674, "step": 12000 }, { "epoch": 0.10597149366820326, "eval_asr_loss": 0.8610385786988017, "eval_loss": 3.250169515609741, "eval_runtime": 20.2526, "eval_samples_per_second": 37.921, "eval_steps_per_second": 9.48, "eval_tts_loss": 6.012793056516366, "step": 12000 }, { "epoch": 0.10641304155848744, "grad_norm": 3.1279189586639404, "learning_rate": 9.99058351984689e-05, "loss": 3.5257, "step": 12050 }, { "epoch": 0.10685458944877162, "grad_norm": 5.271905899047852, "learning_rate": 9.990498082850234e-05, "loss": 3.2174, "step": 12100 }, { "epoch": 0.1072961373390558, "grad_norm": 2.162796974182129, "learning_rate": 9.990412260379922e-05, "loss": 3.4223, "step": 12150 }, { "epoch": 0.10773768522933998, "grad_norm": 2.182065725326538, "learning_rate": 9.990326052442579e-05, "loss": 3.3126, "step": 12200 }, { "epoch": 0.10817923311962416, "grad_norm": 1.2493520975112915, "learning_rate": 9.990239459044866e-05, "loss": 3.5304, "step": 12250 }, { "epoch": 0.10862078100990834, "grad_norm": 2.064229726791382, "learning_rate": 9.99015248019347e-05, "loss": 3.4558, "step": 12300 }, { "epoch": 0.10906232890019252, "grad_norm": 2.22754168510437, "learning_rate": 9.99006511589511e-05, "loss": 3.3677, "step": 12350 }, { "epoch": 0.1095038767904767, "grad_norm": 4.160098075866699, "learning_rate": 9.989977366156535e-05, "loss": 3.5189, "step": 12400 }, { "epoch": 0.10994542468076088, "grad_norm": 0.9621350169181824, "learning_rate": 9.989889230984522e-05, "loss": 3.7358, "step": 12450 }, { "epoch": 0.11038697257104506, "grad_norm": 1.2985560894012451, "learning_rate": 9.989800710385879e-05, "loss": 3.6572, "step": 12500 }, { "epoch": 0.11082852046132924, "grad_norm": 0.9708568453788757, "learning_rate": 9.989711804367443e-05, "loss": 3.2729, "step": 12550 }, { "epoch": 0.11127006835161342, "grad_norm": 1.924854040145874, "learning_rate": 9.989622512936083e-05, "loss": 3.4972, "step": 12600 }, { "epoch": 0.1117116162418976, "grad_norm": 4.869857311248779, "learning_rate": 9.989532836098691e-05, "loss": 3.5102, "step": 12650 }, { "epoch": 0.11215316413218178, "grad_norm": 5.745826244354248, "learning_rate": 9.9894427738622e-05, "loss": 3.3836, "step": 12700 }, { "epoch": 0.11259471202246596, "grad_norm": 5.497977256774902, "learning_rate": 9.989352326233566e-05, "loss": 3.3298, "step": 12750 }, { "epoch": 0.11303625991275014, "grad_norm": 9.22269344329834, "learning_rate": 9.98926149321977e-05, "loss": 3.3542, "step": 12800 }, { "epoch": 0.11347780780303432, "grad_norm": 2.255528211593628, "learning_rate": 9.989172102972332e-05, "loss": 3.8121, "step": 12850 }, { "epoch": 0.1139193556933185, "grad_norm": 2.629471778869629, "learning_rate": 9.98908050691665e-05, "loss": 3.456, "step": 12900 }, { "epoch": 0.11436090358360268, "grad_norm": 6.432432174682617, "learning_rate": 9.988988525496805e-05, "loss": 3.1076, "step": 12950 }, { "epoch": 0.11480245147388686, "grad_norm": 1.9251166582107544, "learning_rate": 9.988896158719903e-05, "loss": 3.5083, "step": 13000 }, { "epoch": 0.11524399936417104, "grad_norm": 1.9608491659164429, "learning_rate": 9.988803406593077e-05, "loss": 3.4977, "step": 13050 }, { "epoch": 0.11568554725445522, "grad_norm": 2.201385259628296, "learning_rate": 9.988710269123491e-05, "loss": 3.3394, "step": 13100 }, { "epoch": 0.1161270951447394, "grad_norm": 0.8397690653800964, "learning_rate": 9.98861674631834e-05, "loss": 3.0705, "step": 13150 }, { "epoch": 0.11656864303502358, "grad_norm": 3.4315147399902344, "learning_rate": 9.988522838184848e-05, "loss": 2.9964, "step": 13200 }, { "epoch": 0.11701019092530776, "grad_norm": 1.0664474964141846, "learning_rate": 9.988428544730267e-05, "loss": 3.5425, "step": 13250 }, { "epoch": 0.11745173881559194, "grad_norm": 1.7559316158294678, "learning_rate": 9.988333865961883e-05, "loss": 3.4283, "step": 13300 }, { "epoch": 0.11789328670587612, "grad_norm": 1.3742626905441284, "learning_rate": 9.988238801887006e-05, "loss": 3.0457, "step": 13350 }, { "epoch": 0.1183348345961603, "grad_norm": 6.306196212768555, "learning_rate": 9.988143352512982e-05, "loss": 3.4985, "step": 13400 }, { "epoch": 0.11877638248644448, "grad_norm": 0.7708596587181091, "learning_rate": 9.98804751784718e-05, "loss": 3.3379, "step": 13450 }, { "epoch": 0.11921793037672866, "grad_norm": 7.315723419189453, "learning_rate": 9.987951297897008e-05, "loss": 3.4271, "step": 13500 }, { "epoch": 0.11965947826701284, "grad_norm": 1.3316899538040161, "learning_rate": 9.987854692669894e-05, "loss": 3.362, "step": 13550 }, { "epoch": 0.12010102615729702, "grad_norm": 2.5768868923187256, "learning_rate": 9.9877577021733e-05, "loss": 3.7057, "step": 13600 }, { "epoch": 0.1205425740475812, "grad_norm": 1.0544745922088623, "learning_rate": 9.987660326414718e-05, "loss": 3.2666, "step": 13650 }, { "epoch": 0.12098412193786538, "grad_norm": 1.6136611700057983, "learning_rate": 9.98756256540167e-05, "loss": 3.019, "step": 13700 }, { "epoch": 0.12142566982814956, "grad_norm": 3.0674450397491455, "learning_rate": 9.987464419141707e-05, "loss": 3.3559, "step": 13750 }, { "epoch": 0.12186721771843374, "grad_norm": 5.650421142578125, "learning_rate": 9.987365887642412e-05, "loss": 3.1113, "step": 13800 }, { "epoch": 0.12230876560871792, "grad_norm": 6.18209981918335, "learning_rate": 9.987266970911393e-05, "loss": 3.6087, "step": 13850 }, { "epoch": 0.1227503134990021, "grad_norm": 3.334939956665039, "learning_rate": 9.98716766895629e-05, "loss": 3.2844, "step": 13900 }, { "epoch": 0.12319186138928628, "grad_norm": 1.9125257730484009, "learning_rate": 9.987067981784774e-05, "loss": 3.5561, "step": 13950 }, { "epoch": 0.12363340927957046, "grad_norm": 2.013408899307251, "learning_rate": 9.986967909404547e-05, "loss": 3.7224, "step": 14000 }, { "epoch": 0.12407495716985464, "grad_norm": 1.1568374633789062, "learning_rate": 9.986867451823337e-05, "loss": 3.348, "step": 14050 }, { "epoch": 0.12451650506013882, "grad_norm": 2.9201159477233887, "learning_rate": 9.986766609048904e-05, "loss": 3.358, "step": 14100 }, { "epoch": 0.124958052950423, "grad_norm": 6.0850982666015625, "learning_rate": 9.986665381089038e-05, "loss": 3.3947, "step": 14150 }, { "epoch": 0.12539960084070717, "grad_norm": 2.2208523750305176, "learning_rate": 9.986563767951555e-05, "loss": 3.3521, "step": 14200 }, { "epoch": 0.12584114873099136, "grad_norm": 3.7602078914642334, "learning_rate": 9.986461769644306e-05, "loss": 3.3009, "step": 14250 }, { "epoch": 0.12628269662127553, "grad_norm": 5.159857749938965, "learning_rate": 9.98635938617517e-05, "loss": 3.0826, "step": 14300 }, { "epoch": 0.12672424451155972, "grad_norm": 1.6318804025650024, "learning_rate": 9.986256617552054e-05, "loss": 3.1461, "step": 14350 }, { "epoch": 0.1271657924018439, "grad_norm": 1.6564624309539795, "learning_rate": 9.986153463782897e-05, "loss": 3.6095, "step": 14400 }, { "epoch": 0.12760734029212809, "grad_norm": 6.902859210968018, "learning_rate": 9.986049924875666e-05, "loss": 3.4435, "step": 14450 }, { "epoch": 0.12804888818241225, "grad_norm": 2.513582229614258, "learning_rate": 9.985946000838359e-05, "loss": 3.6592, "step": 14500 }, { "epoch": 0.12849043607269645, "grad_norm": 1.5714596509933472, "learning_rate": 9.985841691679004e-05, "loss": 3.3225, "step": 14550 }, { "epoch": 0.1289319839629806, "grad_norm": 2.4705002307891846, "learning_rate": 9.985736997405655e-05, "loss": 3.3319, "step": 14600 }, { "epoch": 0.1293735318532648, "grad_norm": 0.8616177439689636, "learning_rate": 9.985631918026401e-05, "loss": 3.1653, "step": 14650 }, { "epoch": 0.12981507974354897, "grad_norm": 1.7353193759918213, "learning_rate": 9.985526453549359e-05, "loss": 3.4691, "step": 14700 }, { "epoch": 0.13025662763383317, "grad_norm": 5.408633708953857, "learning_rate": 9.985420603982673e-05, "loss": 3.5129, "step": 14750 }, { "epoch": 0.13069817552411733, "grad_norm": 1.2819982767105103, "learning_rate": 9.985314369334523e-05, "loss": 3.2823, "step": 14800 }, { "epoch": 0.13113972341440153, "grad_norm": 8.26905632019043, "learning_rate": 9.98520774961311e-05, "loss": 3.2017, "step": 14850 }, { "epoch": 0.1315812713046857, "grad_norm": 2.404160737991333, "learning_rate": 9.985100744826674e-05, "loss": 3.2397, "step": 14900 }, { "epoch": 0.1320228191949699, "grad_norm": 1.9694509506225586, "learning_rate": 9.984993354983477e-05, "loss": 3.7057, "step": 14950 }, { "epoch": 0.13246436708525405, "grad_norm": 7.179697036743164, "learning_rate": 9.984885580091814e-05, "loss": 2.8689, "step": 15000 }, { "epoch": 0.13246436708525405, "eval_asr_loss": 0.8746035999883381, "eval_loss": 3.1943347454071045, "eval_runtime": 20.0975, "eval_samples_per_second": 38.214, "eval_steps_per_second": 9.553, "eval_tts_loss": 5.963482838455088, "step": 15000 }, { "epoch": 0.13290591497553825, "grad_norm": 1.963370680809021, "learning_rate": 9.98477742016001e-05, "loss": 3.0848, "step": 15050 }, { "epoch": 0.1333474628658224, "grad_norm": 5.354642868041992, "learning_rate": 9.984668875196421e-05, "loss": 3.4422, "step": 15100 }, { "epoch": 0.1337890107561066, "grad_norm": 0.9628197550773621, "learning_rate": 9.98455994520943e-05, "loss": 3.6956, "step": 15150 }, { "epoch": 0.13423055864639077, "grad_norm": 1.9591059684753418, "learning_rate": 9.984450630207451e-05, "loss": 3.8912, "step": 15200 }, { "epoch": 0.13467210653667497, "grad_norm": 2.4057066440582275, "learning_rate": 9.984340930198927e-05, "loss": 3.4458, "step": 15250 }, { "epoch": 0.13511365442695913, "grad_norm": 1.3141722679138184, "learning_rate": 9.984230845192336e-05, "loss": 3.3671, "step": 15300 }, { "epoch": 0.13555520231724333, "grad_norm": 3.438781261444092, "learning_rate": 9.984120375196174e-05, "loss": 3.2085, "step": 15350 }, { "epoch": 0.1359967502075275, "grad_norm": 1.443537712097168, "learning_rate": 9.984011741091279e-05, "loss": 3.2912, "step": 15400 }, { "epoch": 0.1364382980978117, "grad_norm": 2.5141725540161133, "learning_rate": 9.983900508840976e-05, "loss": 3.4251, "step": 15450 }, { "epoch": 0.13687984598809586, "grad_norm": 3.5416128635406494, "learning_rate": 9.98378889162662e-05, "loss": 3.4953, "step": 15500 }, { "epoch": 0.13732139387838005, "grad_norm": 3.750843048095703, "learning_rate": 9.983676889456833e-05, "loss": 3.6289, "step": 15550 }, { "epoch": 0.13776294176866422, "grad_norm": 8.563606262207031, "learning_rate": 9.983564502340267e-05, "loss": 3.3268, "step": 15600 }, { "epoch": 0.1382044896589484, "grad_norm": 3.2222371101379395, "learning_rate": 9.983451730285603e-05, "loss": 3.329, "step": 15650 }, { "epoch": 0.1386460375492326, "grad_norm": 2.218827724456787, "learning_rate": 9.983338573301552e-05, "loss": 3.8973, "step": 15700 }, { "epoch": 0.13908758543951677, "grad_norm": 1.1404446363449097, "learning_rate": 9.983225031396852e-05, "loss": 3.5931, "step": 15750 }, { "epoch": 0.13952913332980096, "grad_norm": 0.8973853588104248, "learning_rate": 9.983111104580276e-05, "loss": 3.2535, "step": 15800 }, { "epoch": 0.13997068122008513, "grad_norm": 8.366558074951172, "learning_rate": 9.982996792860623e-05, "loss": 3.0558, "step": 15850 }, { "epoch": 0.14041222911036932, "grad_norm": 1.4844486713409424, "learning_rate": 9.982882096246722e-05, "loss": 3.4828, "step": 15900 }, { "epoch": 0.1408537770006535, "grad_norm": 2.583129644393921, "learning_rate": 9.982767014747432e-05, "loss": 3.6149, "step": 15950 }, { "epoch": 0.14129532489093768, "grad_norm": 3.2486109733581543, "learning_rate": 9.982651548371644e-05, "loss": 3.1331, "step": 16000 }, { "epoch": 0.14173687278122185, "grad_norm": 2.9078361988067627, "learning_rate": 9.982535697128275e-05, "loss": 2.9418, "step": 16050 }, { "epoch": 0.14217842067150605, "grad_norm": 1.986458659172058, "learning_rate": 9.982419461026273e-05, "loss": 3.4626, "step": 16100 }, { "epoch": 0.1426199685617902, "grad_norm": 5.311156272888184, "learning_rate": 9.982302840074617e-05, "loss": 3.1937, "step": 16150 }, { "epoch": 0.1430615164520744, "grad_norm": 1.4057707786560059, "learning_rate": 9.982185834282319e-05, "loss": 4.0044, "step": 16200 }, { "epoch": 0.14350306434235857, "grad_norm": 1.7638394832611084, "learning_rate": 9.98206844365841e-05, "loss": 3.657, "step": 16250 }, { "epoch": 0.14394461223264277, "grad_norm": 2.069039821624756, "learning_rate": 9.98195066821196e-05, "loss": 3.5534, "step": 16300 }, { "epoch": 0.14438616012292693, "grad_norm": 7.656248092651367, "learning_rate": 9.981832507952067e-05, "loss": 3.1842, "step": 16350 }, { "epoch": 0.14482770801321113, "grad_norm": 3.7678370475769043, "learning_rate": 9.981713962887859e-05, "loss": 3.515, "step": 16400 }, { "epoch": 0.1452692559034953, "grad_norm": 2.7892682552337646, "learning_rate": 9.98159503302849e-05, "loss": 3.6017, "step": 16450 }, { "epoch": 0.1457108037937795, "grad_norm": 1.75071120262146, "learning_rate": 9.981475718383147e-05, "loss": 3.6968, "step": 16500 }, { "epoch": 0.14615235168406365, "grad_norm": 0.988599419593811, "learning_rate": 9.981356018961047e-05, "loss": 3.638, "step": 16550 }, { "epoch": 0.14659389957434785, "grad_norm": 3.183884382247925, "learning_rate": 9.981235934771436e-05, "loss": 3.4087, "step": 16600 }, { "epoch": 0.147035447464632, "grad_norm": 2.3756110668182373, "learning_rate": 9.981115465823587e-05, "loss": 3.493, "step": 16650 }, { "epoch": 0.1474769953549162, "grad_norm": 1.6157623529434204, "learning_rate": 9.980994612126807e-05, "loss": 3.4459, "step": 16700 }, { "epoch": 0.14791854324520037, "grad_norm": 3.4954352378845215, "learning_rate": 9.980873373690431e-05, "loss": 3.6551, "step": 16750 }, { "epoch": 0.14836009113548457, "grad_norm": 2.129396915435791, "learning_rate": 9.980751750523825e-05, "loss": 3.5101, "step": 16800 }, { "epoch": 0.14880163902576873, "grad_norm": 1.6230214834213257, "learning_rate": 9.98062974263638e-05, "loss": 3.3445, "step": 16850 }, { "epoch": 0.14924318691605293, "grad_norm": 2.7362122535705566, "learning_rate": 9.980507350037522e-05, "loss": 3.4352, "step": 16900 }, { "epoch": 0.1496847348063371, "grad_norm": 4.6118364334106445, "learning_rate": 9.980384572736706e-05, "loss": 2.9692, "step": 16950 }, { "epoch": 0.1501262826966213, "grad_norm": 3.5687484741210938, "learning_rate": 9.980261410743414e-05, "loss": 3.669, "step": 17000 }, { "epoch": 0.15056783058690545, "grad_norm": 1.4324193000793457, "learning_rate": 9.980137864067158e-05, "loss": 3.0393, "step": 17050 }, { "epoch": 0.15100937847718965, "grad_norm": 2.652137517929077, "learning_rate": 9.980013932717484e-05, "loss": 3.6648, "step": 17100 }, { "epoch": 0.15145092636747381, "grad_norm": 7.182887554168701, "learning_rate": 9.97988961670396e-05, "loss": 3.3025, "step": 17150 }, { "epoch": 0.151892474257758, "grad_norm": 2.4190828800201416, "learning_rate": 9.979764916036195e-05, "loss": 3.5013, "step": 17200 }, { "epoch": 0.15233402214804218, "grad_norm": 4.525496006011963, "learning_rate": 9.979639830723816e-05, "loss": 3.4685, "step": 17250 }, { "epoch": 0.15277557003832637, "grad_norm": 1.242196798324585, "learning_rate": 9.979514360776486e-05, "loss": 3.3286, "step": 17300 }, { "epoch": 0.15321711792861054, "grad_norm": 4.049784183502197, "learning_rate": 9.979388506203897e-05, "loss": 3.1067, "step": 17350 }, { "epoch": 0.15365866581889473, "grad_norm": 2.2166428565979004, "learning_rate": 9.97926226701577e-05, "loss": 3.1167, "step": 17400 }, { "epoch": 0.1541002137091789, "grad_norm": 1.4584290981292725, "learning_rate": 9.979135643221856e-05, "loss": 3.4869, "step": 17450 }, { "epoch": 0.1545417615994631, "grad_norm": 2.357682228088379, "learning_rate": 9.979008634831934e-05, "loss": 3.3837, "step": 17500 }, { "epoch": 0.15498330948974726, "grad_norm": 3.9010651111602783, "learning_rate": 9.978881241855817e-05, "loss": 3.0948, "step": 17550 }, { "epoch": 0.15542485738003145, "grad_norm": 2.2975761890411377, "learning_rate": 9.978753464303343e-05, "loss": 3.487, "step": 17600 }, { "epoch": 0.15586640527031562, "grad_norm": 1.2210371494293213, "learning_rate": 9.978625302184383e-05, "loss": 3.7763, "step": 17650 }, { "epoch": 0.1563079531605998, "grad_norm": 1.8714485168457031, "learning_rate": 9.978496755508836e-05, "loss": 2.9216, "step": 17700 }, { "epoch": 0.15674950105088398, "grad_norm": 3.4243338108062744, "learning_rate": 9.978367824286629e-05, "loss": 3.5642, "step": 17750 }, { "epoch": 0.15719104894116817, "grad_norm": 1.6219075918197632, "learning_rate": 9.978238508527727e-05, "loss": 3.3517, "step": 17800 }, { "epoch": 0.15763259683145234, "grad_norm": 0.7262994647026062, "learning_rate": 9.97810880824211e-05, "loss": 2.9789, "step": 17850 }, { "epoch": 0.15807414472173653, "grad_norm": 2.3567357063293457, "learning_rate": 9.977981328904049e-05, "loss": 3.4182, "step": 17900 }, { "epoch": 0.1585156926120207, "grad_norm": 2.0587501525878906, "learning_rate": 9.977850867285131e-05, "loss": 3.4566, "step": 17950 }, { "epoch": 0.1589572405023049, "grad_norm": 1.3674755096435547, "learning_rate": 9.977720021169445e-05, "loss": 3.4426, "step": 18000 }, { "epoch": 0.1589572405023049, "eval_asr_loss": 0.8684477998103262, "eval_loss": 3.168827772140503, "eval_runtime": 20.2426, "eval_samples_per_second": 37.94, "eval_steps_per_second": 9.485, "eval_tts_loss": 5.975099919698394, "step": 18000 }, { "epoch": 0.15939878839258906, "grad_norm": 1.7318178415298462, "learning_rate": 9.977588790567097e-05, "loss": 3.1039, "step": 18050 }, { "epoch": 0.15984033628287325, "grad_norm": 4.160914897918701, "learning_rate": 9.977457175488225e-05, "loss": 3.4418, "step": 18100 }, { "epoch": 0.16028188417315742, "grad_norm": 1.6405212879180908, "learning_rate": 9.977325175942992e-05, "loss": 3.537, "step": 18150 }, { "epoch": 0.1607234320634416, "grad_norm": 3.2284727096557617, "learning_rate": 9.977192791941596e-05, "loss": 3.5699, "step": 18200 }, { "epoch": 0.16116497995372578, "grad_norm": 6.879377841949463, "learning_rate": 9.977060023494263e-05, "loss": 3.3869, "step": 18250 }, { "epoch": 0.16160652784400997, "grad_norm": 5.974034309387207, "learning_rate": 9.976926870611247e-05, "loss": 3.3198, "step": 18300 }, { "epoch": 0.16204807573429414, "grad_norm": 1.0853605270385742, "learning_rate": 9.976793333302834e-05, "loss": 3.2399, "step": 18350 }, { "epoch": 0.16248962362457833, "grad_norm": 2.7796192169189453, "learning_rate": 9.976659411579337e-05, "loss": 3.3726, "step": 18400 }, { "epoch": 0.1629311715148625, "grad_norm": 2.083576202392578, "learning_rate": 9.976525105451102e-05, "loss": 3.3063, "step": 18450 }, { "epoch": 0.1633727194051467, "grad_norm": 1.6082487106323242, "learning_rate": 9.976390414928501e-05, "loss": 3.2636, "step": 18500 }, { "epoch": 0.16381426729543086, "grad_norm": 2.3426663875579834, "learning_rate": 9.97625534002194e-05, "loss": 3.4482, "step": 18550 }, { "epoch": 0.16425581518571505, "grad_norm": 4.199110507965088, "learning_rate": 9.976119880741852e-05, "loss": 3.5328, "step": 18600 }, { "epoch": 0.16469736307599922, "grad_norm": 2.411940097808838, "learning_rate": 9.975984037098698e-05, "loss": 3.1049, "step": 18650 }, { "epoch": 0.16513891096628341, "grad_norm": 8.064169883728027, "learning_rate": 9.975847809102974e-05, "loss": 3.2858, "step": 18700 }, { "epoch": 0.16558045885656758, "grad_norm": 2.0680835247039795, "learning_rate": 9.9757111967652e-05, "loss": 3.1414, "step": 18750 }, { "epoch": 0.16602200674685177, "grad_norm": 2.7678062915802, "learning_rate": 9.975574200095927e-05, "loss": 3.1492, "step": 18800 }, { "epoch": 0.16646355463713594, "grad_norm": 3.402519941329956, "learning_rate": 9.975436819105742e-05, "loss": 3.1184, "step": 18850 }, { "epoch": 0.16690510252742013, "grad_norm": 1.3270809650421143, "learning_rate": 9.97529905380525e-05, "loss": 3.4959, "step": 18900 }, { "epoch": 0.1673466504177043, "grad_norm": 7.308635711669922, "learning_rate": 9.975160904205098e-05, "loss": 3.4175, "step": 18950 }, { "epoch": 0.1677881983079885, "grad_norm": 1.5945827960968018, "learning_rate": 9.975022370315952e-05, "loss": 3.2295, "step": 19000 }, { "epoch": 0.16822974619827266, "grad_norm": 4.1863861083984375, "learning_rate": 9.974883452148517e-05, "loss": 3.1561, "step": 19050 }, { "epoch": 0.16867129408855686, "grad_norm": 1.7289392948150635, "learning_rate": 9.97474414971352e-05, "loss": 3.5214, "step": 19100 }, { "epoch": 0.16911284197884102, "grad_norm": 1.1129475831985474, "learning_rate": 9.974604463021722e-05, "loss": 2.9315, "step": 19150 }, { "epoch": 0.16955438986912522, "grad_norm": 1.6852185726165771, "learning_rate": 9.974464392083913e-05, "loss": 3.4969, "step": 19200 }, { "epoch": 0.16999593775940938, "grad_norm": 6.104635238647461, "learning_rate": 9.974323936910913e-05, "loss": 3.6542, "step": 19250 }, { "epoch": 0.17043748564969358, "grad_norm": 2.2778172492980957, "learning_rate": 9.974183097513567e-05, "loss": 3.1306, "step": 19300 }, { "epoch": 0.17087903353997774, "grad_norm": 1.4735571146011353, "learning_rate": 9.974041873902757e-05, "loss": 3.654, "step": 19350 }, { "epoch": 0.17132058143026194, "grad_norm": 3.6428987979888916, "learning_rate": 9.973900266089393e-05, "loss": 3.7674, "step": 19400 }, { "epoch": 0.1717621293205461, "grad_norm": 3.5552337169647217, "learning_rate": 9.97375827408441e-05, "loss": 3.5145, "step": 19450 }, { "epoch": 0.1722036772108303, "grad_norm": 3.519054651260376, "learning_rate": 9.973615897898777e-05, "loss": 3.1466, "step": 19500 }, { "epoch": 0.17264522510111446, "grad_norm": 5.785290241241455, "learning_rate": 9.97347313754349e-05, "loss": 3.207, "step": 19550 }, { "epoch": 0.17308677299139866, "grad_norm": 1.1869070529937744, "learning_rate": 9.973329993029577e-05, "loss": 3.4333, "step": 19600 }, { "epoch": 0.17352832088168282, "grad_norm": 1.620424747467041, "learning_rate": 9.973186464368096e-05, "loss": 3.4715, "step": 19650 }, { "epoch": 0.17396986877196702, "grad_norm": 3.2140984535217285, "learning_rate": 9.973042551570131e-05, "loss": 3.5503, "step": 19700 }, { "epoch": 0.17441141666225118, "grad_norm": 8.555562973022461, "learning_rate": 9.972898254646799e-05, "loss": 3.9534, "step": 19750 }, { "epoch": 0.17485296455253538, "grad_norm": 2.755021095275879, "learning_rate": 9.972753573609245e-05, "loss": 3.2534, "step": 19800 }, { "epoch": 0.17529451244281954, "grad_norm": 2.5277762413024902, "learning_rate": 9.972608508468646e-05, "loss": 3.9156, "step": 19850 }, { "epoch": 0.17573606033310374, "grad_norm": 6.709779739379883, "learning_rate": 9.972463059236207e-05, "loss": 3.3525, "step": 19900 }, { "epoch": 0.1761776082233879, "grad_norm": 2.41070294380188, "learning_rate": 9.972317225923161e-05, "loss": 3.3441, "step": 19950 }, { "epoch": 0.1766191561136721, "grad_norm": 1.5217126607894897, "learning_rate": 9.972171008540774e-05, "loss": 3.2966, "step": 20000 }, { "epoch": 0.17706070400395627, "grad_norm": 5.1036763191223145, "learning_rate": 9.972024407100337e-05, "loss": 3.1964, "step": 20050 }, { "epoch": 0.17750225189424046, "grad_norm": 1.291801929473877, "learning_rate": 9.97187742161318e-05, "loss": 3.6654, "step": 20100 }, { "epoch": 0.17794379978452463, "grad_norm": 1.6413884162902832, "learning_rate": 9.97173005209065e-05, "loss": 3.5594, "step": 20150 }, { "epoch": 0.17838534767480882, "grad_norm": 2.1488759517669678, "learning_rate": 9.971582298544133e-05, "loss": 3.1799, "step": 20200 }, { "epoch": 0.17882689556509299, "grad_norm": 4.286724090576172, "learning_rate": 9.971434160985041e-05, "loss": 3.1923, "step": 20250 }, { "epoch": 0.17926844345537718, "grad_norm": 2.672957420349121, "learning_rate": 9.971285639424817e-05, "loss": 3.216, "step": 20300 }, { "epoch": 0.17970999134566135, "grad_norm": 1.8457285165786743, "learning_rate": 9.971136733874931e-05, "loss": 3.0968, "step": 20350 }, { "epoch": 0.18015153923594554, "grad_norm": 2.780160665512085, "learning_rate": 9.970987444346889e-05, "loss": 3.7793, "step": 20400 }, { "epoch": 0.1805930871262297, "grad_norm": 4.923703670501709, "learning_rate": 9.970837770852218e-05, "loss": 3.0386, "step": 20450 }, { "epoch": 0.1810346350165139, "grad_norm": 6.101312637329102, "learning_rate": 9.97068771340248e-05, "loss": 3.6275, "step": 20500 }, { "epoch": 0.18147618290679807, "grad_norm": 1.9402213096618652, "learning_rate": 9.970537272009269e-05, "loss": 3.0918, "step": 20550 }, { "epoch": 0.18191773079708226, "grad_norm": 0.9089447259902954, "learning_rate": 9.970386446684201e-05, "loss": 3.5084, "step": 20600 }, { "epoch": 0.18235927868736643, "grad_norm": 0.5540192723274231, "learning_rate": 9.970235237438927e-05, "loss": 3.3147, "step": 20650 }, { "epoch": 0.18280082657765062, "grad_norm": 3.3821587562561035, "learning_rate": 9.970083644285129e-05, "loss": 3.7705, "step": 20700 }, { "epoch": 0.1832423744679348, "grad_norm": 2.602141857147217, "learning_rate": 9.969931667234512e-05, "loss": 3.231, "step": 20750 }, { "epoch": 0.18368392235821898, "grad_norm": 2.034205198287964, "learning_rate": 9.969779306298818e-05, "loss": 3.5487, "step": 20800 }, { "epoch": 0.18412547024850315, "grad_norm": 2.844647169113159, "learning_rate": 9.969626561489817e-05, "loss": 3.5581, "step": 20850 }, { "epoch": 0.18456701813878734, "grad_norm": 1.6377545595169067, "learning_rate": 9.969473432819305e-05, "loss": 3.8547, "step": 20900 }, { "epoch": 0.1850085660290715, "grad_norm": 7.625864505767822, "learning_rate": 9.96931992029911e-05, "loss": 3.1848, "step": 20950 }, { "epoch": 0.1854501139193557, "grad_norm": 1.9800269603729248, "learning_rate": 9.969166023941089e-05, "loss": 3.1796, "step": 21000 }, { "epoch": 0.1854501139193557, "eval_asr_loss": 0.8812751858267902, "eval_loss": 3.14119029045105, "eval_runtime": 19.8885, "eval_samples_per_second": 38.615, "eval_steps_per_second": 9.654, "eval_tts_loss": 5.95043236252646, "step": 21000 }, { "epoch": 0.18589166180963987, "grad_norm": 0.8084915280342102, "learning_rate": 9.969011743757131e-05, "loss": 3.1867, "step": 21050 }, { "epoch": 0.18633320969992406, "grad_norm": 4.80615234375, "learning_rate": 9.968857079759151e-05, "loss": 3.3244, "step": 21100 }, { "epoch": 0.18677475759020823, "grad_norm": 1.4828674793243408, "learning_rate": 9.968702031959096e-05, "loss": 2.8393, "step": 21150 }, { "epoch": 0.18721630548049242, "grad_norm": 3.2170047760009766, "learning_rate": 9.968546600368942e-05, "loss": 3.3566, "step": 21200 }, { "epoch": 0.1876578533707766, "grad_norm": 2.004711151123047, "learning_rate": 9.968390785000696e-05, "loss": 3.5387, "step": 21250 }, { "epoch": 0.18809940126106078, "grad_norm": 2.858024835586548, "learning_rate": 9.968234585866393e-05, "loss": 3.4617, "step": 21300 }, { "epoch": 0.18854094915134495, "grad_norm": 2.3722898960113525, "learning_rate": 9.968078002978097e-05, "loss": 3.3284, "step": 21350 }, { "epoch": 0.18898249704162914, "grad_norm": 3.7834815979003906, "learning_rate": 9.967921036347904e-05, "loss": 3.2961, "step": 21400 }, { "epoch": 0.1894240449319133, "grad_norm": 1.9320305585861206, "learning_rate": 9.967763685987937e-05, "loss": 3.0897, "step": 21450 }, { "epoch": 0.1898655928221975, "grad_norm": 2.595550537109375, "learning_rate": 9.96760595191035e-05, "loss": 3.1881, "step": 21500 }, { "epoch": 0.19030714071248167, "grad_norm": 2.7209346294403076, "learning_rate": 9.967447834127328e-05, "loss": 3.9049, "step": 21550 }, { "epoch": 0.19074868860276586, "grad_norm": 2.9117844104766846, "learning_rate": 9.967289332651085e-05, "loss": 3.7096, "step": 21600 }, { "epoch": 0.19119023649305003, "grad_norm": 3.910952091217041, "learning_rate": 9.96713044749386e-05, "loss": 3.2424, "step": 21650 }, { "epoch": 0.19163178438333422, "grad_norm": 0.9400450587272644, "learning_rate": 9.96697117866793e-05, "loss": 2.9279, "step": 21700 }, { "epoch": 0.1920733322736184, "grad_norm": 3.050595760345459, "learning_rate": 9.966811526185594e-05, "loss": 3.4662, "step": 21750 }, { "epoch": 0.19251488016390259, "grad_norm": 4.441404819488525, "learning_rate": 9.966651490059184e-05, "loss": 3.0896, "step": 21800 }, { "epoch": 0.19295642805418675, "grad_norm": 3.2662458419799805, "learning_rate": 9.966491070301064e-05, "loss": 3.1643, "step": 21850 }, { "epoch": 0.19339797594447095, "grad_norm": 1.5034865140914917, "learning_rate": 9.966330266923623e-05, "loss": 3.1728, "step": 21900 }, { "epoch": 0.1938395238347551, "grad_norm": 1.3567650318145752, "learning_rate": 9.966169079939282e-05, "loss": 3.433, "step": 21950 }, { "epoch": 0.1942810717250393, "grad_norm": 2.380711793899536, "learning_rate": 9.966007509360492e-05, "loss": 3.3513, "step": 22000 }, { "epoch": 0.19472261961532347, "grad_norm": 1.819464921951294, "learning_rate": 9.965845555199732e-05, "loss": 3.0924, "step": 22050 }, { "epoch": 0.19516416750560767, "grad_norm": 1.84955894947052, "learning_rate": 9.965683217469513e-05, "loss": 3.3711, "step": 22100 }, { "epoch": 0.19560571539589183, "grad_norm": 2.3455941677093506, "learning_rate": 9.965520496182371e-05, "loss": 3.5974, "step": 22150 }, { "epoch": 0.19604726328617603, "grad_norm": 1.5657644271850586, "learning_rate": 9.965357391350877e-05, "loss": 3.5424, "step": 22200 }, { "epoch": 0.1964888111764602, "grad_norm": 2.460850477218628, "learning_rate": 9.965193902987632e-05, "loss": 3.4851, "step": 22250 }, { "epoch": 0.1969303590667444, "grad_norm": 4.1430768966674805, "learning_rate": 9.96503003110526e-05, "loss": 3.6162, "step": 22300 }, { "epoch": 0.19737190695702855, "grad_norm": 1.511238932609558, "learning_rate": 9.964865775716421e-05, "loss": 3.4041, "step": 22350 }, { "epoch": 0.19781345484731275, "grad_norm": 5.789384841918945, "learning_rate": 9.964701136833801e-05, "loss": 3.6231, "step": 22400 }, { "epoch": 0.1982550027375969, "grad_norm": 1.697780728340149, "learning_rate": 9.964536114470118e-05, "loss": 3.5089, "step": 22450 }, { "epoch": 0.1986965506278811, "grad_norm": 1.9673670530319214, "learning_rate": 9.964370708638118e-05, "loss": 3.1656, "step": 22500 }, { "epoch": 0.19913809851816527, "grad_norm": 1.157395601272583, "learning_rate": 9.964204919350577e-05, "loss": 3.5386, "step": 22550 }, { "epoch": 0.19957964640844947, "grad_norm": 1.872974157333374, "learning_rate": 9.964038746620303e-05, "loss": 3.8252, "step": 22600 }, { "epoch": 0.20002119429873363, "grad_norm": 2.0033257007598877, "learning_rate": 9.963872190460128e-05, "loss": 3.0679, "step": 22650 }, { "epoch": 0.20046274218901783, "grad_norm": 1.9374537467956543, "learning_rate": 9.963705250882919e-05, "loss": 3.4905, "step": 22700 }, { "epoch": 0.200904290079302, "grad_norm": 1.368355393409729, "learning_rate": 9.963537927901571e-05, "loss": 3.1261, "step": 22750 }, { "epoch": 0.2013458379695862, "grad_norm": 2.8965837955474854, "learning_rate": 9.963370221529007e-05, "loss": 3.434, "step": 22800 }, { "epoch": 0.20178738585987036, "grad_norm": 4.269700527191162, "learning_rate": 9.96320213177818e-05, "loss": 3.3089, "step": 22850 }, { "epoch": 0.20222893375015455, "grad_norm": 2.0559756755828857, "learning_rate": 9.963033658662078e-05, "loss": 3.0824, "step": 22900 }, { "epoch": 0.20267048164043872, "grad_norm": 2.692986011505127, "learning_rate": 9.962868183079845e-05, "loss": 2.9178, "step": 22950 }, { "epoch": 0.2031120295307229, "grad_norm": 2.5358245372772217, "learning_rate": 9.962698950938911e-05, "loss": 3.8775, "step": 23000 }, { "epoch": 0.20355357742100708, "grad_norm": 2.653918504714966, "learning_rate": 9.962529335471565e-05, "loss": 3.9464, "step": 23050 }, { "epoch": 0.20399512531129127, "grad_norm": 2.1768572330474854, "learning_rate": 9.962362740422908e-05, "loss": 3.3436, "step": 23100 }, { "epoch": 0.20443667320157544, "grad_norm": 1.4818288087844849, "learning_rate": 9.962192366007948e-05, "loss": 3.3089, "step": 23150 }, { "epoch": 0.20487822109185963, "grad_norm": 8.97103500366211, "learning_rate": 9.962021608305707e-05, "loss": 3.2528, "step": 23200 }, { "epoch": 0.2053197689821438, "grad_norm": 2.6199610233306885, "learning_rate": 9.961850467329372e-05, "loss": 3.3929, "step": 23250 }, { "epoch": 0.205761316872428, "grad_norm": 1.7377407550811768, "learning_rate": 9.961678943092164e-05, "loss": 3.4344, "step": 23300 }, { "epoch": 0.20620286476271216, "grad_norm": 1.5331953763961792, "learning_rate": 9.961507035607332e-05, "loss": 3.6088, "step": 23350 }, { "epoch": 0.20664441265299635, "grad_norm": 5.9846014976501465, "learning_rate": 9.961334744888154e-05, "loss": 3.5139, "step": 23400 }, { "epoch": 0.20708596054328052, "grad_norm": 1.6636093854904175, "learning_rate": 9.961162070947936e-05, "loss": 3.783, "step": 23450 }, { "epoch": 0.2075275084335647, "grad_norm": 4.730373382568359, "learning_rate": 9.960989013800019e-05, "loss": 3.5666, "step": 23500 }, { "epoch": 0.20796905632384888, "grad_norm": 1.7042059898376465, "learning_rate": 9.960815573457769e-05, "loss": 3.4093, "step": 23550 }, { "epoch": 0.20841060421413307, "grad_norm": 1.6063077449798584, "learning_rate": 9.96064174993458e-05, "loss": 2.9415, "step": 23600 }, { "epoch": 0.20885215210441724, "grad_norm": 2.5845298767089844, "learning_rate": 9.960467543243884e-05, "loss": 3.421, "step": 23650 }, { "epoch": 0.20929369999470143, "grad_norm": 2.843557119369507, "learning_rate": 9.960292953399131e-05, "loss": 3.7182, "step": 23700 }, { "epoch": 0.2097352478849856, "grad_norm": 2.7299513816833496, "learning_rate": 9.960117980413811e-05, "loss": 3.4932, "step": 23750 }, { "epoch": 0.2101767957752698, "grad_norm": 3.5709776878356934, "learning_rate": 9.959942624301437e-05, "loss": 3.4617, "step": 23800 }, { "epoch": 0.21061834366555396, "grad_norm": 1.636993408203125, "learning_rate": 9.959766885075556e-05, "loss": 3.2784, "step": 23850 }, { "epoch": 0.21105989155583815, "grad_norm": 1.2677909135818481, "learning_rate": 9.959590762749738e-05, "loss": 3.5762, "step": 23900 }, { "epoch": 0.21150143944612232, "grad_norm": 4.043791770935059, "learning_rate": 9.959414257337591e-05, "loss": 3.1917, "step": 23950 }, { "epoch": 0.2119429873364065, "grad_norm": 3.0787148475646973, "learning_rate": 9.959237368852747e-05, "loss": 3.307, "step": 24000 }, { "epoch": 0.2119429873364065, "eval_asr_loss": 0.8739951278052379, "eval_loss": 3.1280133724212646, "eval_runtime": 20.4078, "eval_samples_per_second": 37.633, "eval_steps_per_second": 9.408, "eval_tts_loss": 5.9267985776675065, "step": 24000 }, { "epoch": 0.21238453522669068, "grad_norm": 2.54976487159729, "learning_rate": 9.95906009730887e-05, "loss": 3.178, "step": 24050 }, { "epoch": 0.21282608311697487, "grad_norm": 1.0521984100341797, "learning_rate": 9.958882442719652e-05, "loss": 3.7995, "step": 24100 }, { "epoch": 0.21326763100725904, "grad_norm": 1.708702564239502, "learning_rate": 9.958704405098815e-05, "loss": 3.5033, "step": 24150 }, { "epoch": 0.21370917889754323, "grad_norm": 2.1780407428741455, "learning_rate": 9.958525984460112e-05, "loss": 3.4977, "step": 24200 }, { "epoch": 0.2141507267878274, "grad_norm": 2.227015972137451, "learning_rate": 9.958347180817323e-05, "loss": 2.9502, "step": 24250 }, { "epoch": 0.2145922746781116, "grad_norm": 1.5458627939224243, "learning_rate": 9.95816799418426e-05, "loss": 3.3266, "step": 24300 }, { "epoch": 0.21503382256839576, "grad_norm": 1.8348188400268555, "learning_rate": 9.957988424574764e-05, "loss": 3.3448, "step": 24350 }, { "epoch": 0.21547537045867995, "grad_norm": 2.4458820819854736, "learning_rate": 9.957808472002704e-05, "loss": 3.9597, "step": 24400 }, { "epoch": 0.21591691834896412, "grad_norm": 1.2153455018997192, "learning_rate": 9.957628136481981e-05, "loss": 3.0095, "step": 24450 }, { "epoch": 0.21635846623924831, "grad_norm": 1.6723811626434326, "learning_rate": 9.957447418026525e-05, "loss": 3.3562, "step": 24500 }, { "epoch": 0.21680001412953248, "grad_norm": 4.406776428222656, "learning_rate": 9.957266316650291e-05, "loss": 3.4466, "step": 24550 }, { "epoch": 0.21724156201981668, "grad_norm": 5.27102518081665, "learning_rate": 9.957084832367273e-05, "loss": 3.3489, "step": 24600 }, { "epoch": 0.21768310991010084, "grad_norm": 2.6601903438568115, "learning_rate": 9.956902965191486e-05, "loss": 3.3423, "step": 24650 }, { "epoch": 0.21812465780038504, "grad_norm": 1.5734961032867432, "learning_rate": 9.956720715136977e-05, "loss": 3.3207, "step": 24700 }, { "epoch": 0.2185662056906692, "grad_norm": 2.8826043605804443, "learning_rate": 9.956538082217826e-05, "loss": 3.1197, "step": 24750 }, { "epoch": 0.2190077535809534, "grad_norm": 1.6438381671905518, "learning_rate": 9.956355066448138e-05, "loss": 3.1748, "step": 24800 }, { "epoch": 0.21944930147123756, "grad_norm": 1.6108850240707397, "learning_rate": 9.956171667842048e-05, "loss": 3.7212, "step": 24850 }, { "epoch": 0.21989084936152176, "grad_norm": 1.1622767448425293, "learning_rate": 9.955987886413725e-05, "loss": 3.553, "step": 24900 }, { "epoch": 0.22033239725180592, "grad_norm": 1.478102207183838, "learning_rate": 9.955803722177363e-05, "loss": 3.2171, "step": 24950 }, { "epoch": 0.22077394514209012, "grad_norm": 2.2040581703186035, "learning_rate": 9.95561917514719e-05, "loss": 3.4424, "step": 25000 }, { "epoch": 0.22121549303237428, "grad_norm": 1.8141365051269531, "learning_rate": 9.955434245337457e-05, "loss": 3.6843, "step": 25050 }, { "epoch": 0.22165704092265848, "grad_norm": 1.6195813417434692, "learning_rate": 9.955248932762447e-05, "loss": 3.7377, "step": 25100 }, { "epoch": 0.22209858881294264, "grad_norm": 2.677534580230713, "learning_rate": 9.955063237436477e-05, "loss": 3.5157, "step": 25150 }, { "epoch": 0.22254013670322684, "grad_norm": 1.7741618156433105, "learning_rate": 9.954880884685869e-05, "loss": 3.4256, "step": 25200 }, { "epoch": 0.222981684593511, "grad_norm": 0.9014108777046204, "learning_rate": 9.954694431555343e-05, "loss": 3.5028, "step": 25250 }, { "epoch": 0.2234232324837952, "grad_norm": 2.5479094982147217, "learning_rate": 9.954507595716685e-05, "loss": 3.7392, "step": 25300 }, { "epoch": 0.22386478037407936, "grad_norm": 1.6607797145843506, "learning_rate": 9.954320377184328e-05, "loss": 3.7739, "step": 25350 }, { "epoch": 0.22430632826436356, "grad_norm": 2.8050482273101807, "learning_rate": 9.954132775972735e-05, "loss": 3.3908, "step": 25400 }, { "epoch": 0.22474787615464772, "grad_norm": 0.8809208869934082, "learning_rate": 9.953944792096392e-05, "loss": 3.4604, "step": 25450 }, { "epoch": 0.22518942404493192, "grad_norm": 4.810263156890869, "learning_rate": 9.953756425569824e-05, "loss": 3.3016, "step": 25500 }, { "epoch": 0.22563097193521608, "grad_norm": 0.7430752515792847, "learning_rate": 9.95356767640758e-05, "loss": 3.3905, "step": 25550 }, { "epoch": 0.22607251982550028, "grad_norm": 1.2698413133621216, "learning_rate": 9.953378544624237e-05, "loss": 3.4881, "step": 25600 }, { "epoch": 0.22651406771578445, "grad_norm": 2.968998908996582, "learning_rate": 9.953189030234405e-05, "loss": 3.2608, "step": 25650 }, { "epoch": 0.22695561560606864, "grad_norm": 8.66765308380127, "learning_rate": 9.95299913325272e-05, "loss": 3.1594, "step": 25700 }, { "epoch": 0.2273971634963528, "grad_norm": 3.061037302017212, "learning_rate": 9.952808853693856e-05, "loss": 3.5046, "step": 25750 }, { "epoch": 0.227838711386637, "grad_norm": 1.6408863067626953, "learning_rate": 9.952618191572503e-05, "loss": 2.9974, "step": 25800 }, { "epoch": 0.22828025927692117, "grad_norm": 1.7766352891921997, "learning_rate": 9.952427146903393e-05, "loss": 3.4394, "step": 25850 }, { "epoch": 0.22872180716720536, "grad_norm": 1.8550622463226318, "learning_rate": 9.952235719701283e-05, "loss": 3.2499, "step": 25900 }, { "epoch": 0.22916335505748953, "grad_norm": 4.626537799835205, "learning_rate": 9.952043909980955e-05, "loss": 3.3191, "step": 25950 }, { "epoch": 0.22960490294777372, "grad_norm": 1.894518256187439, "learning_rate": 9.951851717757228e-05, "loss": 3.2887, "step": 26000 }, { "epoch": 0.2300464508380579, "grad_norm": 2.2566466331481934, "learning_rate": 9.951659143044947e-05, "loss": 3.2224, "step": 26050 }, { "epoch": 0.23048799872834208, "grad_norm": 1.502895474433899, "learning_rate": 9.951466185858985e-05, "loss": 3.4682, "step": 26100 }, { "epoch": 0.23092954661862625, "grad_norm": 3.664829730987549, "learning_rate": 9.951272846214247e-05, "loss": 3.0882, "step": 26150 }, { "epoch": 0.23137109450891044, "grad_norm": 2.9891955852508545, "learning_rate": 9.951079124125668e-05, "loss": 3.1857, "step": 26200 }, { "epoch": 0.2318126423991946, "grad_norm": 1.8004250526428223, "learning_rate": 9.950885019608211e-05, "loss": 3.3859, "step": 26250 }, { "epoch": 0.2322541902894788, "grad_norm": 2.643892526626587, "learning_rate": 9.950690532676869e-05, "loss": 3.1025, "step": 26300 }, { "epoch": 0.23269573817976297, "grad_norm": 8.552020072937012, "learning_rate": 9.950495663346662e-05, "loss": 2.8521, "step": 26350 }, { "epoch": 0.23313728607004716, "grad_norm": 2.302180528640747, "learning_rate": 9.950300411632645e-05, "loss": 3.6808, "step": 26400 }, { "epoch": 0.23357883396033133, "grad_norm": 2.410003185272217, "learning_rate": 9.950104777549898e-05, "loss": 3.2536, "step": 26450 }, { "epoch": 0.23402038185061552, "grad_norm": 3.413940906524658, "learning_rate": 9.949908761113534e-05, "loss": 3.6206, "step": 26500 }, { "epoch": 0.2344619297408997, "grad_norm": 1.564780354499817, "learning_rate": 9.949712362338692e-05, "loss": 3.4719, "step": 26550 }, { "epoch": 0.23490347763118388, "grad_norm": 1.3204708099365234, "learning_rate": 9.94951558124054e-05, "loss": 3.95, "step": 26600 }, { "epoch": 0.23534502552146805, "grad_norm": 2.4696288108825684, "learning_rate": 9.949318417834283e-05, "loss": 3.8591, "step": 26650 }, { "epoch": 0.23578657341175224, "grad_norm": 1.5461256504058838, "learning_rate": 9.949120872135144e-05, "loss": 3.6145, "step": 26700 }, { "epoch": 0.2362281213020364, "grad_norm": 1.8961230516433716, "learning_rate": 9.948922944158388e-05, "loss": 3.4088, "step": 26750 }, { "epoch": 0.2366696691923206, "grad_norm": 3.098393201828003, "learning_rate": 9.9487246339193e-05, "loss": 3.1048, "step": 26800 }, { "epoch": 0.23711121708260477, "grad_norm": 0.8614824414253235, "learning_rate": 9.948525941433195e-05, "loss": 3.3338, "step": 26850 }, { "epoch": 0.23755276497288896, "grad_norm": 2.924774646759033, "learning_rate": 9.948326866715426e-05, "loss": 3.2607, "step": 26900 }, { "epoch": 0.23799431286317313, "grad_norm": 5.287550449371338, "learning_rate": 9.948127409781367e-05, "loss": 3.4998, "step": 26950 }, { "epoch": 0.23843586075345732, "grad_norm": 1.865439534187317, "learning_rate": 9.947927570646424e-05, "loss": 3.1788, "step": 27000 }, { "epoch": 0.23843586075345732, "eval_asr_loss": 0.8890385764932143, "eval_loss": 3.1024086475372314, "eval_runtime": 19.895, "eval_samples_per_second": 38.603, "eval_steps_per_second": 9.651, "eval_tts_loss": 5.952762275795868, "step": 27000 }, { "epoch": 0.2388774086437415, "grad_norm": 4.502477169036865, "learning_rate": 9.947727349326033e-05, "loss": 2.8551, "step": 27050 }, { "epoch": 0.23931895653402568, "grad_norm": 2.6308412551879883, "learning_rate": 9.94752674583566e-05, "loss": 3.4248, "step": 27100 }, { "epoch": 0.23976050442430985, "grad_norm": 1.3937596082687378, "learning_rate": 9.9473257601908e-05, "loss": 3.0961, "step": 27150 }, { "epoch": 0.24020205231459404, "grad_norm": 2.9952657222747803, "learning_rate": 9.947124392406977e-05, "loss": 3.7838, "step": 27200 }, { "epoch": 0.2406436002048782, "grad_norm": 3.2740750312805176, "learning_rate": 9.946922642499744e-05, "loss": 3.5869, "step": 27250 }, { "epoch": 0.2410851480951624, "grad_norm": 1.290229082107544, "learning_rate": 9.946720510484686e-05, "loss": 3.521, "step": 27300 }, { "epoch": 0.24152669598544657, "grad_norm": 0.997048020362854, "learning_rate": 9.946517996377417e-05, "loss": 3.1288, "step": 27350 }, { "epoch": 0.24196824387573077, "grad_norm": 1.3124210834503174, "learning_rate": 9.946315100193577e-05, "loss": 2.9812, "step": 27400 }, { "epoch": 0.24240979176601493, "grad_norm": 1.2291759252548218, "learning_rate": 9.946111821948839e-05, "loss": 3.3039, "step": 27450 }, { "epoch": 0.24285133965629913, "grad_norm": 1.4661563634872437, "learning_rate": 9.945908161658904e-05, "loss": 3.271, "step": 27500 }, { "epoch": 0.2432928875465833, "grad_norm": 2.156710624694824, "learning_rate": 9.945704119339506e-05, "loss": 3.7045, "step": 27550 }, { "epoch": 0.24373443543686749, "grad_norm": 3.8776209354400635, "learning_rate": 9.9454996950064e-05, "loss": 3.6893, "step": 27600 }, { "epoch": 0.24417598332715165, "grad_norm": 1.0397090911865234, "learning_rate": 9.945294888675381e-05, "loss": 3.5694, "step": 27650 }, { "epoch": 0.24461753121743585, "grad_norm": 6.586629867553711, "learning_rate": 9.945089700362266e-05, "loss": 3.4295, "step": 27700 }, { "epoch": 0.24505907910772, "grad_norm": 1.8651010990142822, "learning_rate": 9.944884130082905e-05, "loss": 3.4199, "step": 27750 }, { "epoch": 0.2455006269980042, "grad_norm": 2.351825475692749, "learning_rate": 9.944678177853176e-05, "loss": 3.88, "step": 27800 }, { "epoch": 0.24594217488828837, "grad_norm": 3.991497755050659, "learning_rate": 9.944471843688987e-05, "loss": 3.5451, "step": 27850 }, { "epoch": 0.24638372277857257, "grad_norm": 4.23520040512085, "learning_rate": 9.944265127606278e-05, "loss": 3.2976, "step": 27900 }, { "epoch": 0.24682527066885673, "grad_norm": 3.55438232421875, "learning_rate": 9.944058029621013e-05, "loss": 2.9787, "step": 27950 }, { "epoch": 0.24726681855914093, "grad_norm": 6.835021495819092, "learning_rate": 9.94385054974919e-05, "loss": 3.151, "step": 28000 }, { "epoch": 0.2477083664494251, "grad_norm": 2.232171058654785, "learning_rate": 9.943642688006834e-05, "loss": 3.353, "step": 28050 }, { "epoch": 0.2481499143397093, "grad_norm": 2.0550529956817627, "learning_rate": 9.94343444441e-05, "loss": 3.3, "step": 28100 }, { "epoch": 0.24859146222999345, "grad_norm": 0.9220274686813354, "learning_rate": 9.943225818974775e-05, "loss": 3.3238, "step": 28150 }, { "epoch": 0.24903301012027765, "grad_norm": 3.3017609119415283, "learning_rate": 9.943016811717275e-05, "loss": 3.652, "step": 28200 }, { "epoch": 0.24947455801056181, "grad_norm": 2.590731620788574, "learning_rate": 9.94280742265364e-05, "loss": 3.362, "step": 28250 }, { "epoch": 0.249916105900846, "grad_norm": 1.249603271484375, "learning_rate": 9.942597651800046e-05, "loss": 3.2159, "step": 28300 }, { "epoch": 0.2503576537911302, "grad_norm": 3.9774234294891357, "learning_rate": 9.942387499172694e-05, "loss": 3.2234, "step": 28350 }, { "epoch": 0.25079920168141434, "grad_norm": 1.4465439319610596, "learning_rate": 9.94217696478782e-05, "loss": 3.3785, "step": 28400 }, { "epoch": 0.25124074957169856, "grad_norm": 0.9342706799507141, "learning_rate": 9.941966048661682e-05, "loss": 3.1059, "step": 28450 }, { "epoch": 0.25168229746198273, "grad_norm": 2.685728073120117, "learning_rate": 9.941754750810573e-05, "loss": 3.5622, "step": 28500 }, { "epoch": 0.2521238453522669, "grad_norm": 6.555978775024414, "learning_rate": 9.941543071250816e-05, "loss": 3.3656, "step": 28550 }, { "epoch": 0.25256539324255106, "grad_norm": 0.8330821394920349, "learning_rate": 9.94133100999876e-05, "loss": 3.5714, "step": 28600 }, { "epoch": 0.2530069411328353, "grad_norm": 22.39084243774414, "learning_rate": 9.941118567070784e-05, "loss": 3.1974, "step": 28650 }, { "epoch": 0.25344848902311945, "grad_norm": 1.9922549724578857, "learning_rate": 9.940905742483298e-05, "loss": 3.5825, "step": 28700 }, { "epoch": 0.2538900369134036, "grad_norm": 1.42471444606781, "learning_rate": 9.94069253625274e-05, "loss": 3.3538, "step": 28750 }, { "epoch": 0.2543315848036878, "grad_norm": 3.5649797916412354, "learning_rate": 9.940478948395582e-05, "loss": 3.1705, "step": 28800 }, { "epoch": 0.254773132693972, "grad_norm": 2.5138847827911377, "learning_rate": 9.940264978928318e-05, "loss": 3.7867, "step": 28850 }, { "epoch": 0.25521468058425617, "grad_norm": 1.892633080482483, "learning_rate": 9.940050627867476e-05, "loss": 3.1689, "step": 28900 }, { "epoch": 0.25565622847454034, "grad_norm": 1.2871747016906738, "learning_rate": 9.939835895229615e-05, "loss": 3.651, "step": 28950 }, { "epoch": 0.2560977763648245, "grad_norm": 4.025623798370361, "learning_rate": 9.939620781031318e-05, "loss": 3.472, "step": 29000 }, { "epoch": 0.2565393242551087, "grad_norm": 2.834179162979126, "learning_rate": 9.939405285289203e-05, "loss": 3.3033, "step": 29050 }, { "epoch": 0.2569808721453929, "grad_norm": 2.194532871246338, "learning_rate": 9.939189408019916e-05, "loss": 3.3813, "step": 29100 }, { "epoch": 0.25742242003567706, "grad_norm": 2.8828024864196777, "learning_rate": 9.938973149240127e-05, "loss": 3.283, "step": 29150 }, { "epoch": 0.2578639679259612, "grad_norm": 3.8372597694396973, "learning_rate": 9.938756508966547e-05, "loss": 3.324, "step": 29200 }, { "epoch": 0.25830551581624545, "grad_norm": 1.0322990417480469, "learning_rate": 9.938539487215903e-05, "loss": 3.1929, "step": 29250 }, { "epoch": 0.2587470637065296, "grad_norm": 4.666669845581055, "learning_rate": 9.938322084004964e-05, "loss": 3.2645, "step": 29300 }, { "epoch": 0.2591886115968138, "grad_norm": 2.352522850036621, "learning_rate": 9.938104299350519e-05, "loss": 2.8695, "step": 29350 }, { "epoch": 0.25963015948709794, "grad_norm": 1.9813988208770752, "learning_rate": 9.93788613326939e-05, "loss": 3.0595, "step": 29400 }, { "epoch": 0.26007170737738217, "grad_norm": 4.507404804229736, "learning_rate": 9.93766758577843e-05, "loss": 3.3092, "step": 29450 }, { "epoch": 0.26051325526766633, "grad_norm": 0.7371171712875366, "learning_rate": 9.93744865689452e-05, "loss": 3.3464, "step": 29500 }, { "epoch": 0.2609548031579505, "grad_norm": 1.4071518182754517, "learning_rate": 9.937229346634566e-05, "loss": 2.8855, "step": 29550 }, { "epoch": 0.26139635104823467, "grad_norm": 2.9686851501464844, "learning_rate": 9.937009655015515e-05, "loss": 3.1077, "step": 29600 }, { "epoch": 0.2618378989385189, "grad_norm": 2.3262200355529785, "learning_rate": 9.936789582054331e-05, "loss": 3.1181, "step": 29650 }, { "epoch": 0.26227944682880305, "grad_norm": 1.0248990058898926, "learning_rate": 9.936569127768017e-05, "loss": 3.4325, "step": 29700 }, { "epoch": 0.2627209947190872, "grad_norm": 3.9115331172943115, "learning_rate": 9.936348292173596e-05, "loss": 3.132, "step": 29750 }, { "epoch": 0.2631625426093714, "grad_norm": 4.591248512268066, "learning_rate": 9.936127075288128e-05, "loss": 3.151, "step": 29800 }, { "epoch": 0.2636040904996556, "grad_norm": 2.473047971725464, "learning_rate": 9.935905477128703e-05, "loss": 3.4772, "step": 29850 }, { "epoch": 0.2640456383899398, "grad_norm": 1.3905609846115112, "learning_rate": 9.935683497712434e-05, "loss": 2.9894, "step": 29900 }, { "epoch": 0.26448718628022394, "grad_norm": 2.24664568901062, "learning_rate": 9.935461137056469e-05, "loss": 3.689, "step": 29950 }, { "epoch": 0.2649287341705081, "grad_norm": 1.1567890644073486, "learning_rate": 9.935238395177981e-05, "loss": 3.2017, "step": 30000 }, { "epoch": 0.2649287341705081, "eval_asr_loss": 0.8843846080425863, "eval_loss": 3.092066526412964, "eval_runtime": 19.8662, "eval_samples_per_second": 38.659, "eval_steps_per_second": 9.665, "eval_tts_loss": 5.902505457880168, "step": 30000 }, { "epoch": 0.26537028206079233, "grad_norm": 5.792656421661377, "learning_rate": 9.935015272094178e-05, "loss": 3.3117, "step": 30050 }, { "epoch": 0.2658118299510765, "grad_norm": 3.329699993133545, "learning_rate": 9.934791767822292e-05, "loss": 3.4779, "step": 30100 }, { "epoch": 0.26625337784136066, "grad_norm": 1.1682640314102173, "learning_rate": 9.934567882379589e-05, "loss": 3.0166, "step": 30150 }, { "epoch": 0.2666949257316448, "grad_norm": 2.778029441833496, "learning_rate": 9.93434361578336e-05, "loss": 3.209, "step": 30200 }, { "epoch": 0.26713647362192905, "grad_norm": 2.259307622909546, "learning_rate": 9.934118968050929e-05, "loss": 3.0044, "step": 30250 }, { "epoch": 0.2675780215122132, "grad_norm": 2.9219155311584473, "learning_rate": 9.933898443511526e-05, "loss": 3.7484, "step": 30300 }, { "epoch": 0.2680195694024974, "grad_norm": 1.4216759204864502, "learning_rate": 9.933673041180634e-05, "loss": 3.2729, "step": 30350 }, { "epoch": 0.26846111729278155, "grad_norm": 0.8821609020233154, "learning_rate": 9.933447257765338e-05, "loss": 2.7456, "step": 30400 }, { "epoch": 0.26890266518306577, "grad_norm": 4.904880046844482, "learning_rate": 9.933221093283074e-05, "loss": 3.2778, "step": 30450 }, { "epoch": 0.26934421307334994, "grad_norm": 1.9883630275726318, "learning_rate": 9.932994547751313e-05, "loss": 3.4769, "step": 30500 }, { "epoch": 0.2697857609636341, "grad_norm": 3.2121076583862305, "learning_rate": 9.932767621187555e-05, "loss": 3.2237, "step": 30550 }, { "epoch": 0.27022730885391827, "grad_norm": 0.9768474102020264, "learning_rate": 9.932540313609326e-05, "loss": 3.2586, "step": 30600 }, { "epoch": 0.2706688567442025, "grad_norm": 1.543562889099121, "learning_rate": 9.932312625034186e-05, "loss": 3.5258, "step": 30650 }, { "epoch": 0.27111040463448666, "grad_norm": 1.3274840116500854, "learning_rate": 9.932084555479719e-05, "loss": 3.2218, "step": 30700 }, { "epoch": 0.2715519525247708, "grad_norm": 3.6982388496398926, "learning_rate": 9.931856104963543e-05, "loss": 3.4994, "step": 30750 }, { "epoch": 0.271993500415055, "grad_norm": 1.6436798572540283, "learning_rate": 9.931627273503304e-05, "loss": 3.4035, "step": 30800 }, { "epoch": 0.2724350483053392, "grad_norm": 4.315684795379639, "learning_rate": 9.931398061116677e-05, "loss": 3.4837, "step": 30850 }, { "epoch": 0.2728765961956234, "grad_norm": 1.2069634199142456, "learning_rate": 9.931168467821367e-05, "loss": 3.2571, "step": 30900 }, { "epoch": 0.27331814408590754, "grad_norm": 3.137390375137329, "learning_rate": 9.930938493635109e-05, "loss": 3.2457, "step": 30950 }, { "epoch": 0.2737596919761917, "grad_norm": 1.6049838066101074, "learning_rate": 9.930708138575664e-05, "loss": 2.9232, "step": 31000 }, { "epoch": 0.27420123986647593, "grad_norm": 2.4024581909179688, "learning_rate": 9.930477402660827e-05, "loss": 3.1933, "step": 31050 }, { "epoch": 0.2746427877567601, "grad_norm": 2.070897340774536, "learning_rate": 9.93024628590842e-05, "loss": 3.1729, "step": 31100 }, { "epoch": 0.27508433564704426, "grad_norm": 2.37459397315979, "learning_rate": 9.930014788336293e-05, "loss": 3.0385, "step": 31150 }, { "epoch": 0.27552588353732843, "grad_norm": 3.5087156295776367, "learning_rate": 9.92978290996233e-05, "loss": 3.1936, "step": 31200 }, { "epoch": 0.27596743142761265, "grad_norm": 6.348201751708984, "learning_rate": 9.929550650804443e-05, "loss": 3.2907, "step": 31250 }, { "epoch": 0.2764089793178968, "grad_norm": 1.3216320276260376, "learning_rate": 9.929318010880566e-05, "loss": 3.1206, "step": 31300 }, { "epoch": 0.276850527208181, "grad_norm": 3.7011451721191406, "learning_rate": 9.929084990208674e-05, "loss": 3.6797, "step": 31350 }, { "epoch": 0.2772920750984652, "grad_norm": 1.47987961769104, "learning_rate": 9.928851588806763e-05, "loss": 3.2982, "step": 31400 }, { "epoch": 0.2777336229887494, "grad_norm": 2.7035892009735107, "learning_rate": 9.928617806692863e-05, "loss": 3.5924, "step": 31450 }, { "epoch": 0.27817517087903354, "grad_norm": 1.140507698059082, "learning_rate": 9.92838364388503e-05, "loss": 3.4756, "step": 31500 }, { "epoch": 0.2786167187693177, "grad_norm": 1.1196190118789673, "learning_rate": 9.928149100401354e-05, "loss": 3.5627, "step": 31550 }, { "epoch": 0.27905826665960193, "grad_norm": 3.444877862930298, "learning_rate": 9.927914176259947e-05, "loss": 3.3291, "step": 31600 }, { "epoch": 0.2794998145498861, "grad_norm": 4.410919189453125, "learning_rate": 9.927678871478958e-05, "loss": 3.1749, "step": 31650 }, { "epoch": 0.27994136244017026, "grad_norm": 1.771317481994629, "learning_rate": 9.927443186076563e-05, "loss": 2.6498, "step": 31700 }, { "epoch": 0.2803829103304544, "grad_norm": 2.1711716651916504, "learning_rate": 9.927207120070964e-05, "loss": 3.6787, "step": 31750 }, { "epoch": 0.28082445822073865, "grad_norm": 4.0384202003479, "learning_rate": 9.926970673480398e-05, "loss": 3.6298, "step": 31800 }, { "epoch": 0.2812660061110228, "grad_norm": 2.198716640472412, "learning_rate": 9.926733846323125e-05, "loss": 2.9091, "step": 31850 }, { "epoch": 0.281707554001307, "grad_norm": 1.819726586341858, "learning_rate": 9.92649663861744e-05, "loss": 3.5799, "step": 31900 }, { "epoch": 0.28214910189159115, "grad_norm": 1.7500370740890503, "learning_rate": 9.926259050381667e-05, "loss": 3.4091, "step": 31950 }, { "epoch": 0.28259064978187537, "grad_norm": 4.203965663909912, "learning_rate": 9.926021081634155e-05, "loss": 3.2101, "step": 32000 }, { "epoch": 0.28303219767215954, "grad_norm": 2.5612051486968994, "learning_rate": 9.925782732393285e-05, "loss": 3.0251, "step": 32050 }, { "epoch": 0.2834737455624437, "grad_norm": 1.427227258682251, "learning_rate": 9.92554400267747e-05, "loss": 3.2904, "step": 32100 }, { "epoch": 0.28391529345272787, "grad_norm": 1.6692140102386475, "learning_rate": 9.925304892505146e-05, "loss": 3.74, "step": 32150 }, { "epoch": 0.2843568413430121, "grad_norm": 1.287840485572815, "learning_rate": 9.925065401894786e-05, "loss": 3.5495, "step": 32200 }, { "epoch": 0.28479838923329626, "grad_norm": 1.882232427597046, "learning_rate": 9.924825530864887e-05, "loss": 3.2056, "step": 32250 }, { "epoch": 0.2852399371235804, "grad_norm": 1.2615700960159302, "learning_rate": 9.924585279433976e-05, "loss": 3.1918, "step": 32300 }, { "epoch": 0.2856814850138646, "grad_norm": 1.7634786367416382, "learning_rate": 9.924344647620612e-05, "loss": 3.5558, "step": 32350 }, { "epoch": 0.2861230329041488, "grad_norm": 1.2818759679794312, "learning_rate": 9.924103635443381e-05, "loss": 3.0904, "step": 32400 }, { "epoch": 0.286564580794433, "grad_norm": 3.8670778274536133, "learning_rate": 9.9238622429209e-05, "loss": 3.8008, "step": 32450 }, { "epoch": 0.28700612868471714, "grad_norm": 2.2975099086761475, "learning_rate": 9.923625309255875e-05, "loss": 2.9886, "step": 32500 }, { "epoch": 0.2874476765750013, "grad_norm": 3.461240530014038, "learning_rate": 9.923383163704833e-05, "loss": 3.3378, "step": 32550 }, { "epoch": 0.28788922446528553, "grad_norm": 1.2354512214660645, "learning_rate": 9.923140637864192e-05, "loss": 3.4829, "step": 32600 }, { "epoch": 0.2883307723555697, "grad_norm": 2.5937671661376953, "learning_rate": 9.922897731752683e-05, "loss": 3.3737, "step": 32650 }, { "epoch": 0.28877232024585386, "grad_norm": 2.7086663246154785, "learning_rate": 9.922654445389069e-05, "loss": 3.3247, "step": 32700 }, { "epoch": 0.28921386813613803, "grad_norm": 1.1576141119003296, "learning_rate": 9.922410778792142e-05, "loss": 3.3032, "step": 32750 }, { "epoch": 0.28965541602642225, "grad_norm": 4.800418853759766, "learning_rate": 9.922166731980725e-05, "loss": 3.7363, "step": 32800 }, { "epoch": 0.2900969639167064, "grad_norm": 1.6237969398498535, "learning_rate": 9.921922304973663e-05, "loss": 3.6136, "step": 32850 }, { "epoch": 0.2905385118069906, "grad_norm": 3.0943002700805664, "learning_rate": 9.921677497789843e-05, "loss": 3.18, "step": 32900 }, { "epoch": 0.29098005969727475, "grad_norm": 2.4920825958251953, "learning_rate": 9.921432310448169e-05, "loss": 3.2806, "step": 32950 }, { "epoch": 0.291421607587559, "grad_norm": 4.264760971069336, "learning_rate": 9.921186742967581e-05, "loss": 2.9155, "step": 33000 }, { "epoch": 0.291421607587559, "eval_asr_loss": 0.895976054242954, "eval_loss": 3.079402208328247, "eval_runtime": 20.0371, "eval_samples_per_second": 38.329, "eval_steps_per_second": 9.582, "eval_tts_loss": 5.968250579680352, "step": 33000 }, { "epoch": 0.29186315547784314, "grad_norm": 3.386139154434204, "learning_rate": 9.92094079536705e-05, "loss": 3.9219, "step": 33050 }, { "epoch": 0.2923047033681273, "grad_norm": 1.537395715713501, "learning_rate": 9.92069446766557e-05, "loss": 3.1577, "step": 33100 }, { "epoch": 0.29274625125841147, "grad_norm": 0.6773030757904053, "learning_rate": 9.920447759882168e-05, "loss": 2.9516, "step": 33150 }, { "epoch": 0.2931877991486957, "grad_norm": 2.179901599884033, "learning_rate": 9.9202006720359e-05, "loss": 3.7766, "step": 33200 }, { "epoch": 0.29362934703897986, "grad_norm": 4.384719371795654, "learning_rate": 9.919953204145853e-05, "loss": 3.4504, "step": 33250 }, { "epoch": 0.294070894929264, "grad_norm": 5.59550666809082, "learning_rate": 9.919705356231141e-05, "loss": 3.2456, "step": 33300 }, { "epoch": 0.2945124428195482, "grad_norm": 1.8736000061035156, "learning_rate": 9.919457128310909e-05, "loss": 3.5584, "step": 33350 }, { "epoch": 0.2949539907098324, "grad_norm": 2.3086678981781006, "learning_rate": 9.919208520404328e-05, "loss": 3.3485, "step": 33400 }, { "epoch": 0.2953955386001166, "grad_norm": 1.736769676208496, "learning_rate": 9.918959532530604e-05, "loss": 3.426, "step": 33450 }, { "epoch": 0.29583708649040075, "grad_norm": 1.8966106176376343, "learning_rate": 9.918710164708967e-05, "loss": 3.0041, "step": 33500 }, { "epoch": 0.2962786343806849, "grad_norm": 3.222245216369629, "learning_rate": 9.918460416958679e-05, "loss": 3.0005, "step": 33550 }, { "epoch": 0.29672018227096914, "grad_norm": 0.8299497365951538, "learning_rate": 9.918210289299031e-05, "loss": 3.3702, "step": 33600 }, { "epoch": 0.2971617301612533, "grad_norm": 3.876796245574951, "learning_rate": 9.917959781749345e-05, "loss": 3.243, "step": 33650 }, { "epoch": 0.29760327805153747, "grad_norm": 4.3070807456970215, "learning_rate": 9.917708894328968e-05, "loss": 3.4787, "step": 33700 }, { "epoch": 0.29804482594182163, "grad_norm": 2.306934356689453, "learning_rate": 9.917457627057279e-05, "loss": 3.2177, "step": 33750 }, { "epoch": 0.29848637383210586, "grad_norm": 3.8237075805664062, "learning_rate": 9.917205979953687e-05, "loss": 2.9533, "step": 33800 }, { "epoch": 0.29892792172239, "grad_norm": 1.627469778060913, "learning_rate": 9.91695395303763e-05, "loss": 3.3876, "step": 33850 }, { "epoch": 0.2993694696126742, "grad_norm": 1.0928808450698853, "learning_rate": 9.916701546328576e-05, "loss": 2.8597, "step": 33900 }, { "epoch": 0.29981101750295835, "grad_norm": 2.0172150135040283, "learning_rate": 9.916448759846017e-05, "loss": 3.6425, "step": 33950 }, { "epoch": 0.3002525653932426, "grad_norm": 2.5472891330718994, "learning_rate": 9.916195593609486e-05, "loss": 3.4216, "step": 34000 }, { "epoch": 0.30069411328352674, "grad_norm": 2.0118589401245117, "learning_rate": 9.91594712227922e-05, "loss": 3.1182, "step": 34050 }, { "epoch": 0.3011356611738109, "grad_norm": 1.4318723678588867, "learning_rate": 9.915693204187532e-05, "loss": 3.2758, "step": 34100 }, { "epoch": 0.3015772090640951, "grad_norm": 6.2676801681518555, "learning_rate": 9.915438906400228e-05, "loss": 3.2553, "step": 34150 }, { "epoch": 0.3020187569543793, "grad_norm": 1.1752573251724243, "learning_rate": 9.915184228936952e-05, "loss": 3.4345, "step": 34200 }, { "epoch": 0.30246030484466346, "grad_norm": 1.754859447479248, "learning_rate": 9.914929171817371e-05, "loss": 3.5339, "step": 34250 }, { "epoch": 0.30290185273494763, "grad_norm": 1.1092463731765747, "learning_rate": 9.914673735061191e-05, "loss": 3.7403, "step": 34300 }, { "epoch": 0.3033434006252318, "grad_norm": 2.3356130123138428, "learning_rate": 9.91441791868814e-05, "loss": 3.7242, "step": 34350 }, { "epoch": 0.303784948515516, "grad_norm": 1.9679780006408691, "learning_rate": 9.914161722717976e-05, "loss": 3.0423, "step": 34400 }, { "epoch": 0.3042264964058002, "grad_norm": 2.5788609981536865, "learning_rate": 9.91390514717049e-05, "loss": 3.3403, "step": 34450 }, { "epoch": 0.30466804429608435, "grad_norm": 5.141701698303223, "learning_rate": 9.913648192065503e-05, "loss": 3.0419, "step": 34500 }, { "epoch": 0.3051095921863685, "grad_norm": 0.9429007172584534, "learning_rate": 9.913390857422858e-05, "loss": 3.4227, "step": 34550 }, { "epoch": 0.30555114007665274, "grad_norm": 5.526501655578613, "learning_rate": 9.913133143262432e-05, "loss": 3.4853, "step": 34600 }, { "epoch": 0.3059926879669369, "grad_norm": 1.2435593605041504, "learning_rate": 9.912875049604135e-05, "loss": 3.1218, "step": 34650 }, { "epoch": 0.30643423585722107, "grad_norm": 3.752776861190796, "learning_rate": 9.912616576467899e-05, "loss": 3.29, "step": 34700 }, { "epoch": 0.30687578374750524, "grad_norm": 1.3690961599349976, "learning_rate": 9.91235772387369e-05, "loss": 3.5259, "step": 34750 }, { "epoch": 0.30731733163778946, "grad_norm": 5.097490310668945, "learning_rate": 9.912098491841504e-05, "loss": 3.4285, "step": 34800 }, { "epoch": 0.3077588795280736, "grad_norm": 2.029536485671997, "learning_rate": 9.911838880391362e-05, "loss": 3.5238, "step": 34850 }, { "epoch": 0.3082004274183578, "grad_norm": 1.1511303186416626, "learning_rate": 9.911578889543318e-05, "loss": 3.5973, "step": 34900 }, { "epoch": 0.30864197530864196, "grad_norm": 0.7825906872749329, "learning_rate": 9.911318519317454e-05, "loss": 3.5027, "step": 34950 }, { "epoch": 0.3090835231989262, "grad_norm": 1.2858946323394775, "learning_rate": 9.91105776973388e-05, "loss": 3.5128, "step": 35000 }, { "epoch": 0.30952507108921035, "grad_norm": 3.028378486633301, "learning_rate": 9.910796640812739e-05, "loss": 3.2618, "step": 35050 }, { "epoch": 0.3099666189794945, "grad_norm": 2.016479253768921, "learning_rate": 9.910535132574198e-05, "loss": 3.4658, "step": 35100 }, { "epoch": 0.3104081668697787, "grad_norm": 5.971153736114502, "learning_rate": 9.91027324503846e-05, "loss": 3.5681, "step": 35150 }, { "epoch": 0.3108497147600629, "grad_norm": 4.901987552642822, "learning_rate": 9.910010978225753e-05, "loss": 3.2022, "step": 35200 }, { "epoch": 0.31129126265034707, "grad_norm": 1.5224626064300537, "learning_rate": 9.90974833215633e-05, "loss": 3.4693, "step": 35250 }, { "epoch": 0.31173281054063123, "grad_norm": 1.7592462301254272, "learning_rate": 9.909485306850483e-05, "loss": 3.2146, "step": 35300 }, { "epoch": 0.3121743584309154, "grad_norm": 2.2695250511169434, "learning_rate": 9.909221902328529e-05, "loss": 3.3325, "step": 35350 }, { "epoch": 0.3126159063211996, "grad_norm": 1.2258449792861938, "learning_rate": 9.908958118610811e-05, "loss": 3.4704, "step": 35400 }, { "epoch": 0.3130574542114838, "grad_norm": 5.258679389953613, "learning_rate": 9.908693955717705e-05, "loss": 3.579, "step": 35450 }, { "epoch": 0.31349900210176795, "grad_norm": 1.850501298904419, "learning_rate": 9.908429413669615e-05, "loss": 3.2378, "step": 35500 }, { "epoch": 0.3139405499920521, "grad_norm": 1.2289330959320068, "learning_rate": 9.908164492486976e-05, "loss": 3.4111, "step": 35550 }, { "epoch": 0.31438209788233634, "grad_norm": 2.0499982833862305, "learning_rate": 9.907899192190249e-05, "loss": 3.6974, "step": 35600 }, { "epoch": 0.3148236457726205, "grad_norm": 1.0805225372314453, "learning_rate": 9.907633512799927e-05, "loss": 3.1798, "step": 35650 }, { "epoch": 0.3152651936629047, "grad_norm": 1.2690644264221191, "learning_rate": 9.907367454336533e-05, "loss": 2.9274, "step": 35700 }, { "epoch": 0.31570674155318884, "grad_norm": 5.509250164031982, "learning_rate": 9.907101016820616e-05, "loss": 3.1606, "step": 35750 }, { "epoch": 0.31614828944347306, "grad_norm": 3.2590670585632324, "learning_rate": 9.906834200272756e-05, "loss": 3.2878, "step": 35800 }, { "epoch": 0.31658983733375723, "grad_norm": 1.7757644653320312, "learning_rate": 9.906567004713563e-05, "loss": 3.3347, "step": 35850 }, { "epoch": 0.3170313852240414, "grad_norm": 6.056863784790039, "learning_rate": 9.906299430163675e-05, "loss": 3.3504, "step": 35900 }, { "epoch": 0.31747293311432556, "grad_norm": 1.0027168989181519, "learning_rate": 9.90603147664376e-05, "loss": 3.0737, "step": 35950 }, { "epoch": 0.3179144810046098, "grad_norm": 1.916014313697815, "learning_rate": 9.905763144174517e-05, "loss": 3.1648, "step": 36000 }, { "epoch": 0.3179144810046098, "eval_asr_loss": 0.8892677519164152, "eval_loss": 3.0999526977539062, "eval_runtime": 20.2917, "eval_samples_per_second": 37.848, "eval_steps_per_second": 9.462, "eval_tts_loss": 5.944016528967177, "step": 36000 }, { "epoch": 0.31835602889489395, "grad_norm": 2.1346523761749268, "learning_rate": 9.90549443277667e-05, "loss": 3.5172, "step": 36050 }, { "epoch": 0.3187975767851781, "grad_norm": 1.5300114154815674, "learning_rate": 9.905225342470974e-05, "loss": 3.2909, "step": 36100 }, { "epoch": 0.3192391246754623, "grad_norm": 1.7766307592391968, "learning_rate": 9.904955873278216e-05, "loss": 3.5821, "step": 36150 }, { "epoch": 0.3196806725657465, "grad_norm": 1.8257805109024048, "learning_rate": 9.904686025219211e-05, "loss": 3.3543, "step": 36200 }, { "epoch": 0.32012222045603067, "grad_norm": 2.9343011379241943, "learning_rate": 9.9044157983148e-05, "loss": 3.5759, "step": 36250 }, { "epoch": 0.32056376834631484, "grad_norm": 2.4320614337921143, "learning_rate": 9.904145192585857e-05, "loss": 3.2795, "step": 36300 }, { "epoch": 0.321005316236599, "grad_norm": 2.0806117057800293, "learning_rate": 9.903874208053283e-05, "loss": 2.9637, "step": 36350 }, { "epoch": 0.3214468641268832, "grad_norm": 10.894238471984863, "learning_rate": 9.903602844738012e-05, "loss": 3.5964, "step": 36400 }, { "epoch": 0.3218884120171674, "grad_norm": 1.5613411664962769, "learning_rate": 9.903331102661001e-05, "loss": 3.2076, "step": 36450 }, { "epoch": 0.32232995990745156, "grad_norm": 3.5824601650238037, "learning_rate": 9.903058981843242e-05, "loss": 3.3151, "step": 36500 }, { "epoch": 0.3227715077977357, "grad_norm": 4.121826171875, "learning_rate": 9.902786482305753e-05, "loss": 3.1602, "step": 36550 }, { "epoch": 0.32321305568801995, "grad_norm": 2.1794791221618652, "learning_rate": 9.902513604069583e-05, "loss": 3.3487, "step": 36600 }, { "epoch": 0.3236546035783041, "grad_norm": 3.4854958057403564, "learning_rate": 9.902240347155809e-05, "loss": 3.5573, "step": 36650 }, { "epoch": 0.3240961514685883, "grad_norm": 2.346065044403076, "learning_rate": 9.901966711585538e-05, "loss": 2.7336, "step": 36700 }, { "epoch": 0.32453769935887244, "grad_norm": 1.1362333297729492, "learning_rate": 9.901692697379905e-05, "loss": 3.5597, "step": 36750 }, { "epoch": 0.32497924724915667, "grad_norm": 1.6578530073165894, "learning_rate": 9.901418304560077e-05, "loss": 3.4961, "step": 36800 }, { "epoch": 0.32542079513944083, "grad_norm": 1.7271004915237427, "learning_rate": 9.901143533147248e-05, "loss": 3.1233, "step": 36850 }, { "epoch": 0.325862343029725, "grad_norm": 5.3198652267456055, "learning_rate": 9.900868383162643e-05, "loss": 3.6441, "step": 36900 }, { "epoch": 0.32630389092000917, "grad_norm": 1.4437021017074585, "learning_rate": 9.900592854627511e-05, "loss": 3.239, "step": 36950 }, { "epoch": 0.3267454388102934, "grad_norm": 3.09059739112854, "learning_rate": 9.900316947563138e-05, "loss": 3.3918, "step": 37000 }, { "epoch": 0.32718698670057755, "grad_norm": 1.7625781297683716, "learning_rate": 9.900040661990835e-05, "loss": 3.5584, "step": 37050 }, { "epoch": 0.3276285345908617, "grad_norm": 1.064771294593811, "learning_rate": 9.899763997931942e-05, "loss": 3.0205, "step": 37100 }, { "epoch": 0.3280700824811459, "grad_norm": 2.178464889526367, "learning_rate": 9.899486955407826e-05, "loss": 3.2899, "step": 37150 }, { "epoch": 0.3285116303714301, "grad_norm": 1.889468789100647, "learning_rate": 9.899209534439891e-05, "loss": 3.3759, "step": 37200 }, { "epoch": 0.3289531782617143, "grad_norm": 3.320777654647827, "learning_rate": 9.898931735049565e-05, "loss": 3.6612, "step": 37250 }, { "epoch": 0.32939472615199844, "grad_norm": 4.364492416381836, "learning_rate": 9.898653557258303e-05, "loss": 3.5972, "step": 37300 }, { "epoch": 0.3298362740422826, "grad_norm": 4.426956653594971, "learning_rate": 9.898375001087593e-05, "loss": 3.5658, "step": 37350 }, { "epoch": 0.33027782193256683, "grad_norm": 1.2732926607131958, "learning_rate": 9.898096066558951e-05, "loss": 3.1916, "step": 37400 }, { "epoch": 0.330719369822851, "grad_norm": 1.8885505199432373, "learning_rate": 9.897816753693922e-05, "loss": 3.6108, "step": 37450 }, { "epoch": 0.33116091771313516, "grad_norm": 5.467677116394043, "learning_rate": 9.897537062514081e-05, "loss": 3.079, "step": 37500 }, { "epoch": 0.3316024656034193, "grad_norm": 3.006726026535034, "learning_rate": 9.897256993041033e-05, "loss": 3.3902, "step": 37550 }, { "epoch": 0.33204401349370355, "grad_norm": 3.979788064956665, "learning_rate": 9.896976545296411e-05, "loss": 2.8985, "step": 37600 }, { "epoch": 0.3324855613839877, "grad_norm": 0.933592677116394, "learning_rate": 9.896695719301874e-05, "loss": 3.3717, "step": 37650 }, { "epoch": 0.3329271092742719, "grad_norm": 2.022643804550171, "learning_rate": 9.896414515079115e-05, "loss": 3.2212, "step": 37700 }, { "epoch": 0.33336865716455605, "grad_norm": 1.9231008291244507, "learning_rate": 9.896132932649857e-05, "loss": 3.0829, "step": 37750 }, { "epoch": 0.33381020505484027, "grad_norm": 6.672185897827148, "learning_rate": 9.895850972035847e-05, "loss": 3.0794, "step": 37800 }, { "epoch": 0.33425175294512444, "grad_norm": 2.566579580307007, "learning_rate": 9.895568633258865e-05, "loss": 3.2348, "step": 37850 }, { "epoch": 0.3346933008354086, "grad_norm": 3.8691163063049316, "learning_rate": 9.895285916340719e-05, "loss": 3.1823, "step": 37900 }, { "epoch": 0.33513484872569277, "grad_norm": 1.986522912979126, "learning_rate": 9.895002821303246e-05, "loss": 3.4586, "step": 37950 }, { "epoch": 0.335576396615977, "grad_norm": 1.734668254852295, "learning_rate": 9.894719348168315e-05, "loss": 3.7534, "step": 38000 }, { "epoch": 0.33601794450626116, "grad_norm": 1.9242331981658936, "learning_rate": 9.894435496957819e-05, "loss": 3.3395, "step": 38050 }, { "epoch": 0.3364594923965453, "grad_norm": 1.77113676071167, "learning_rate": 9.894151267693686e-05, "loss": 3.352, "step": 38100 }, { "epoch": 0.3369010402868295, "grad_norm": 0.799421489238739, "learning_rate": 9.893866660397867e-05, "loss": 3.7193, "step": 38150 }, { "epoch": 0.3373425881771137, "grad_norm": 1.236433982849121, "learning_rate": 9.89358167509235e-05, "loss": 3.5647, "step": 38200 }, { "epoch": 0.3377841360673979, "grad_norm": 3.1852049827575684, "learning_rate": 9.893296311799143e-05, "loss": 3.477, "step": 38250 }, { "epoch": 0.33822568395768204, "grad_norm": 2.2389414310455322, "learning_rate": 9.89301057054029e-05, "loss": 2.8368, "step": 38300 }, { "epoch": 0.3386672318479662, "grad_norm": 6.785423278808594, "learning_rate": 9.892724451337863e-05, "loss": 3.3759, "step": 38350 }, { "epoch": 0.33910877973825043, "grad_norm": 4.653458595275879, "learning_rate": 9.892437954213958e-05, "loss": 3.353, "step": 38400 }, { "epoch": 0.3395503276285346, "grad_norm": 3.1158320903778076, "learning_rate": 9.892151079190711e-05, "loss": 3.5522, "step": 38450 }, { "epoch": 0.33999187551881876, "grad_norm": 2.8249120712280273, "learning_rate": 9.891863826290277e-05, "loss": 3.3592, "step": 38500 }, { "epoch": 0.34043342340910293, "grad_norm": 2.011695384979248, "learning_rate": 9.891576195534842e-05, "loss": 3.3599, "step": 38550 }, { "epoch": 0.34087497129938715, "grad_norm": 1.8144574165344238, "learning_rate": 9.891288186946628e-05, "loss": 3.3011, "step": 38600 }, { "epoch": 0.3413165191896713, "grad_norm": 2.060272455215454, "learning_rate": 9.890999800547878e-05, "loss": 3.4336, "step": 38650 }, { "epoch": 0.3417580670799555, "grad_norm": 1.401672124862671, "learning_rate": 9.890711036360867e-05, "loss": 3.1762, "step": 38700 }, { "epoch": 0.34219961497023965, "grad_norm": 1.6355420351028442, "learning_rate": 9.8904218944079e-05, "loss": 3.6908, "step": 38750 }, { "epoch": 0.3426411628605239, "grad_norm": 2.591089963912964, "learning_rate": 9.890132374711312e-05, "loss": 3.3153, "step": 38800 }, { "epoch": 0.34308271075080804, "grad_norm": 2.8402106761932373, "learning_rate": 9.889842477293465e-05, "loss": 3.5123, "step": 38850 }, { "epoch": 0.3435242586410922, "grad_norm": 0.9350060820579529, "learning_rate": 9.889552202176751e-05, "loss": 3.3598, "step": 38900 }, { "epoch": 0.3439658065313764, "grad_norm": 2.9623401165008545, "learning_rate": 9.889261549383591e-05, "loss": 3.1874, "step": 38950 }, { "epoch": 0.3444073544216606, "grad_norm": 1.7671315670013428, "learning_rate": 9.888970518936438e-05, "loss": 3.5201, "step": 39000 }, { "epoch": 0.3444073544216606, "eval_asr_loss": 0.8851676637238344, "eval_loss": 3.074697494506836, "eval_runtime": 20.3913, "eval_samples_per_second": 37.663, "eval_steps_per_second": 9.416, "eval_tts_loss": 5.894229360095819, "step": 39000 }, { "epoch": 0.34484890231194476, "grad_norm": 1.1440892219543457, "learning_rate": 9.888679110857768e-05, "loss": 3.4325, "step": 39050 }, { "epoch": 0.3452904502022289, "grad_norm": 2.065777540206909, "learning_rate": 9.888387325170093e-05, "loss": 3.3255, "step": 39100 }, { "epoch": 0.3457319980925131, "grad_norm": 5.079590320587158, "learning_rate": 9.888095161895947e-05, "loss": 3.1736, "step": 39150 }, { "epoch": 0.3461735459827973, "grad_norm": 2.521078586578369, "learning_rate": 9.887802621057902e-05, "loss": 3.6232, "step": 39200 }, { "epoch": 0.3466150938730815, "grad_norm": 3.053846836090088, "learning_rate": 9.88750970267855e-05, "loss": 3.2633, "step": 39250 }, { "epoch": 0.34705664176336565, "grad_norm": 1.3491613864898682, "learning_rate": 9.88721640678052e-05, "loss": 3.7012, "step": 39300 }, { "epoch": 0.3474981896536498, "grad_norm": 1.5826070308685303, "learning_rate": 9.886922733386463e-05, "loss": 3.3255, "step": 39350 }, { "epoch": 0.34793973754393404, "grad_norm": 5.572771072387695, "learning_rate": 9.886628682519069e-05, "loss": 3.2976, "step": 39400 }, { "epoch": 0.3483812854342182, "grad_norm": 1.5839028358459473, "learning_rate": 9.886334254201041e-05, "loss": 3.3149, "step": 39450 }, { "epoch": 0.34882283332450237, "grad_norm": 3.295896053314209, "learning_rate": 9.88603944845513e-05, "loss": 2.9223, "step": 39500 }, { "epoch": 0.34926438121478653, "grad_norm": 2.2568423748016357, "learning_rate": 9.885744265304104e-05, "loss": 3.527, "step": 39550 }, { "epoch": 0.34970592910507076, "grad_norm": 0.9765694737434387, "learning_rate": 9.885448704770762e-05, "loss": 3.4622, "step": 39600 }, { "epoch": 0.3501474769953549, "grad_norm": 1.8776651620864868, "learning_rate": 9.885152766877936e-05, "loss": 3.327, "step": 39650 }, { "epoch": 0.3505890248856391, "grad_norm": 3.4537465572357178, "learning_rate": 9.884856451648483e-05, "loss": 3.5304, "step": 39700 }, { "epoch": 0.35103057277592326, "grad_norm": 2.3317794799804688, "learning_rate": 9.884559759105292e-05, "loss": 3.485, "step": 39750 }, { "epoch": 0.3514721206662075, "grad_norm": 2.5425727367401123, "learning_rate": 9.88426268927128e-05, "loss": 3.5912, "step": 39800 }, { "epoch": 0.35191366855649164, "grad_norm": 0.6102803945541382, "learning_rate": 9.883965242169392e-05, "loss": 3.0589, "step": 39850 }, { "epoch": 0.3523552164467758, "grad_norm": 2.142951726913452, "learning_rate": 9.883667417822604e-05, "loss": 3.4354, "step": 39900 }, { "epoch": 0.35279676433706, "grad_norm": 0.8856062293052673, "learning_rate": 9.883369216253919e-05, "loss": 3.3475, "step": 39950 }, { "epoch": 0.3532383122273442, "grad_norm": 1.8055260181427002, "learning_rate": 9.883070637486374e-05, "loss": 3.2718, "step": 40000 }, { "epoch": 0.35367986011762836, "grad_norm": 1.1027382612228394, "learning_rate": 9.882777664358069e-05, "loss": 3.1336, "step": 40050 }, { "epoch": 0.35412140800791253, "grad_norm": 5.140567779541016, "learning_rate": 9.882478338804843e-05, "loss": 3.3178, "step": 40100 }, { "epoch": 0.3545629558981967, "grad_norm": 2.794602155685425, "learning_rate": 9.882178636121568e-05, "loss": 3.8837, "step": 40150 }, { "epoch": 0.3550045037884809, "grad_norm": 2.1202375888824463, "learning_rate": 9.881878556331392e-05, "loss": 3.5493, "step": 40200 }, { "epoch": 0.3554460516787651, "grad_norm": 0.8689738512039185, "learning_rate": 9.881578099457496e-05, "loss": 3.29, "step": 40250 }, { "epoch": 0.35588759956904925, "grad_norm": 2.1677334308624268, "learning_rate": 9.881277265523085e-05, "loss": 3.2762, "step": 40300 }, { "epoch": 0.3563291474593334, "grad_norm": 1.659857988357544, "learning_rate": 9.880976054551399e-05, "loss": 3.8149, "step": 40350 }, { "epoch": 0.35677069534961764, "grad_norm": 2.109435558319092, "learning_rate": 9.880674466565701e-05, "loss": 3.3179, "step": 40400 }, { "epoch": 0.3572122432399018, "grad_norm": 2.959153175354004, "learning_rate": 9.880372501589289e-05, "loss": 3.5068, "step": 40450 }, { "epoch": 0.35765379113018597, "grad_norm": 0.8862214088439941, "learning_rate": 9.880070159645483e-05, "loss": 3.593, "step": 40500 }, { "epoch": 0.35809533902047014, "grad_norm": 4.1659464836120605, "learning_rate": 9.879767440757641e-05, "loss": 3.1608, "step": 40550 }, { "epoch": 0.35853688691075436, "grad_norm": 4.7671799659729, "learning_rate": 9.879464344949143e-05, "loss": 3.3211, "step": 40600 }, { "epoch": 0.3589784348010385, "grad_norm": 5.715523719787598, "learning_rate": 9.8791608722434e-05, "loss": 3.0981, "step": 40650 }, { "epoch": 0.3594199826913227, "grad_norm": 1.8158924579620361, "learning_rate": 9.878857022663855e-05, "loss": 2.8897, "step": 40700 }, { "epoch": 0.35986153058160686, "grad_norm": 2.091034412384033, "learning_rate": 9.878552796233976e-05, "loss": 3.6337, "step": 40750 }, { "epoch": 0.3603030784718911, "grad_norm": 3.8585503101348877, "learning_rate": 9.878248192977262e-05, "loss": 3.5439, "step": 40800 }, { "epoch": 0.36074462636217525, "grad_norm": 2.2025511264801025, "learning_rate": 9.877943212917242e-05, "loss": 3.3184, "step": 40850 }, { "epoch": 0.3611861742524594, "grad_norm": 3.991384744644165, "learning_rate": 9.877637856077473e-05, "loss": 3.4019, "step": 40900 }, { "epoch": 0.3616277221427436, "grad_norm": 4.2655534744262695, "learning_rate": 9.87733212248154e-05, "loss": 2.82, "step": 40950 }, { "epoch": 0.3620692700330278, "grad_norm": 1.683498740196228, "learning_rate": 9.877026012153059e-05, "loss": 3.3487, "step": 41000 }, { "epoch": 0.36251081792331197, "grad_norm": 2.264552593231201, "learning_rate": 9.876719525115675e-05, "loss": 2.975, "step": 41050 }, { "epoch": 0.36295236581359613, "grad_norm": 2.6153547763824463, "learning_rate": 9.876412661393061e-05, "loss": 3.3764, "step": 41100 }, { "epoch": 0.3633939137038803, "grad_norm": 1.591480016708374, "learning_rate": 9.876105421008919e-05, "loss": 2.965, "step": 41150 }, { "epoch": 0.3638354615941645, "grad_norm": 2.9769952297210693, "learning_rate": 9.875797803986983e-05, "loss": 3.2887, "step": 41200 }, { "epoch": 0.3642770094844487, "grad_norm": 5.074538230895996, "learning_rate": 9.875489810351011e-05, "loss": 3.5681, "step": 41250 }, { "epoch": 0.36471855737473285, "grad_norm": 2.3003244400024414, "learning_rate": 9.875181440124794e-05, "loss": 3.443, "step": 41300 }, { "epoch": 0.365160105265017, "grad_norm": 2.3685524463653564, "learning_rate": 9.874872693332152e-05, "loss": 3.1724, "step": 41350 }, { "epoch": 0.36560165315530124, "grad_norm": 0.6377138495445251, "learning_rate": 9.874563569996931e-05, "loss": 3.4932, "step": 41400 }, { "epoch": 0.3660432010455854, "grad_norm": 1.721374750137329, "learning_rate": 9.874254070143009e-05, "loss": 3.2422, "step": 41450 }, { "epoch": 0.3664847489358696, "grad_norm": 2.2348365783691406, "learning_rate": 9.873944193794294e-05, "loss": 3.0912, "step": 41500 }, { "epoch": 0.36692629682615374, "grad_norm": 1.8013442754745483, "learning_rate": 9.873633940974719e-05, "loss": 3.4616, "step": 41550 }, { "epoch": 0.36736784471643796, "grad_norm": 1.4249016046524048, "learning_rate": 9.873323311708249e-05, "loss": 3.2594, "step": 41600 }, { "epoch": 0.36780939260672213, "grad_norm": 1.336929440498352, "learning_rate": 9.873012306018879e-05, "loss": 3.4553, "step": 41650 }, { "epoch": 0.3682509404970063, "grad_norm": 1.3156746625900269, "learning_rate": 9.872700923930629e-05, "loss": 3.5638, "step": 41700 }, { "epoch": 0.36869248838729046, "grad_norm": 1.1224370002746582, "learning_rate": 9.872389165467553e-05, "loss": 3.131, "step": 41750 }, { "epoch": 0.3691340362775747, "grad_norm": 1.9698013067245483, "learning_rate": 9.87207703065373e-05, "loss": 3.2832, "step": 41800 }, { "epoch": 0.36957558416785885, "grad_norm": 1.916576623916626, "learning_rate": 9.87176451951327e-05, "loss": 3.8032, "step": 41850 }, { "epoch": 0.370017132058143, "grad_norm": 1.1512713432312012, "learning_rate": 9.871451632070312e-05, "loss": 3.4968, "step": 41900 }, { "epoch": 0.3704586799484272, "grad_norm": 1.165654182434082, "learning_rate": 9.871138368349025e-05, "loss": 3.7091, "step": 41950 }, { "epoch": 0.3709002278387114, "grad_norm": 0.8057221174240112, "learning_rate": 9.870824728373604e-05, "loss": 2.845, "step": 42000 }, { "epoch": 0.3709002278387114, "eval_asr_loss": 0.8905537465911811, "eval_loss": 3.063842535018921, "eval_runtime": 20.3425, "eval_samples_per_second": 37.754, "eval_steps_per_second": 9.438, "eval_tts_loss": 5.883050221472156, "step": 42000 }, { "epoch": 0.37134177572899557, "grad_norm": 9.080826759338379, "learning_rate": 9.870510712168278e-05, "loss": 3.178, "step": 42050 }, { "epoch": 0.37178332361927974, "grad_norm": 1.3995147943496704, "learning_rate": 9.870196319757299e-05, "loss": 3.316, "step": 42100 }, { "epoch": 0.3722248715095639, "grad_norm": 2.64806866645813, "learning_rate": 9.869887850223219e-05, "loss": 3.0827, "step": 42150 }, { "epoch": 0.3726664193998481, "grad_norm": 0.9530137181282043, "learning_rate": 9.869572712996722e-05, "loss": 3.2979, "step": 42200 }, { "epoch": 0.3731079672901323, "grad_norm": 2.155907392501831, "learning_rate": 9.869257199637023e-05, "loss": 3.2503, "step": 42250 }, { "epoch": 0.37354951518041646, "grad_norm": 1.780704379081726, "learning_rate": 9.868941310168498e-05, "loss": 3.4232, "step": 42300 }, { "epoch": 0.3739910630707006, "grad_norm": 0.6130134463310242, "learning_rate": 9.868625044615543e-05, "loss": 3.2393, "step": 42350 }, { "epoch": 0.37443261096098485, "grad_norm": 1.8535226583480835, "learning_rate": 9.868308403002589e-05, "loss": 3.2783, "step": 42400 }, { "epoch": 0.374874158851269, "grad_norm": 3.596348285675049, "learning_rate": 9.867991385354092e-05, "loss": 3.157, "step": 42450 }, { "epoch": 0.3753157067415532, "grad_norm": 8.728959083557129, "learning_rate": 9.86767399169454e-05, "loss": 3.542, "step": 42500 }, { "epoch": 0.37575725463183735, "grad_norm": 1.9934968948364258, "learning_rate": 9.867356222048448e-05, "loss": 3.4873, "step": 42550 }, { "epoch": 0.37619880252212157, "grad_norm": 2.618511915206909, "learning_rate": 9.867038076440362e-05, "loss": 3.1409, "step": 42600 }, { "epoch": 0.37664035041240573, "grad_norm": 1.344989538192749, "learning_rate": 9.866719554894857e-05, "loss": 3.4496, "step": 42650 }, { "epoch": 0.3770818983026899, "grad_norm": 1.429974913597107, "learning_rate": 9.866400657436534e-05, "loss": 3.1867, "step": 42700 }, { "epoch": 0.37752344619297407, "grad_norm": 1.021186351776123, "learning_rate": 9.866081384090026e-05, "loss": 2.994, "step": 42750 }, { "epoch": 0.3779649940832583, "grad_norm": 2.500567674636841, "learning_rate": 9.865761734879994e-05, "loss": 2.9304, "step": 42800 }, { "epoch": 0.37840654197354245, "grad_norm": 1.401653528213501, "learning_rate": 9.865441709831129e-05, "loss": 3.5495, "step": 42850 }, { "epoch": 0.3788480898638266, "grad_norm": 2.189068555831909, "learning_rate": 9.865121308968151e-05, "loss": 3.5045, "step": 42900 }, { "epoch": 0.3792896377541108, "grad_norm": 1.583028793334961, "learning_rate": 9.864800532315806e-05, "loss": 3.4075, "step": 42950 }, { "epoch": 0.379731185644395, "grad_norm": 1.778917908668518, "learning_rate": 9.864479379898871e-05, "loss": 3.3313, "step": 43000 }, { "epoch": 0.3801727335346792, "grad_norm": 2.908151626586914, "learning_rate": 9.864157851742154e-05, "loss": 3.1747, "step": 43050 }, { "epoch": 0.38061428142496334, "grad_norm": 3.3366377353668213, "learning_rate": 9.86383594787049e-05, "loss": 3.5529, "step": 43100 }, { "epoch": 0.3810558293152475, "grad_norm": 1.980222225189209, "learning_rate": 9.863513668308744e-05, "loss": 3.559, "step": 43150 }, { "epoch": 0.38149737720553173, "grad_norm": 2.1785378456115723, "learning_rate": 9.863191013081807e-05, "loss": 3.3076, "step": 43200 }, { "epoch": 0.3819389250958159, "grad_norm": 3.69520902633667, "learning_rate": 9.862867982214604e-05, "loss": 3.4053, "step": 43250 }, { "epoch": 0.38238047298610006, "grad_norm": 5.211033821105957, "learning_rate": 9.862544575732085e-05, "loss": 3.2504, "step": 43300 }, { "epoch": 0.38282202087638423, "grad_norm": 2.920567750930786, "learning_rate": 9.862220793659232e-05, "loss": 3.1881, "step": 43350 }, { "epoch": 0.38326356876666845, "grad_norm": 2.8354828357696533, "learning_rate": 9.86189663602105e-05, "loss": 3.6872, "step": 43400 }, { "epoch": 0.3837051166569526, "grad_norm": 3.6911203861236572, "learning_rate": 9.861572102842583e-05, "loss": 3.3914, "step": 43450 }, { "epoch": 0.3841466645472368, "grad_norm": 3.9869236946105957, "learning_rate": 9.861247194148895e-05, "loss": 3.3991, "step": 43500 }, { "epoch": 0.38458821243752095, "grad_norm": 2.9851901531219482, "learning_rate": 9.860928419328401e-05, "loss": 3.4688, "step": 43550 }, { "epoch": 0.38502976032780517, "grad_norm": 5.538595199584961, "learning_rate": 9.860602767188645e-05, "loss": 3.3237, "step": 43600 }, { "epoch": 0.38547130821808934, "grad_norm": 0.9150497317314148, "learning_rate": 9.860276739608543e-05, "loss": 3.2588, "step": 43650 }, { "epoch": 0.3859128561083735, "grad_norm": 3.647045612335205, "learning_rate": 9.859950336613275e-05, "loss": 3.1377, "step": 43700 }, { "epoch": 0.38635440399865767, "grad_norm": 0.6097536087036133, "learning_rate": 9.859623558228055e-05, "loss": 3.1102, "step": 43750 }, { "epoch": 0.3867959518889419, "grad_norm": 1.7617607116699219, "learning_rate": 9.859296404478124e-05, "loss": 2.9313, "step": 43800 }, { "epoch": 0.38723749977922606, "grad_norm": 2.4368743896484375, "learning_rate": 9.858968875388752e-05, "loss": 3.2671, "step": 43850 }, { "epoch": 0.3876790476695102, "grad_norm": 1.931584358215332, "learning_rate": 9.858640970985237e-05, "loss": 3.4481, "step": 43900 }, { "epoch": 0.3881205955597944, "grad_norm": 1.9381108283996582, "learning_rate": 9.858312691292906e-05, "loss": 3.2431, "step": 43950 }, { "epoch": 0.3885621434500786, "grad_norm": 2.4838809967041016, "learning_rate": 9.857984036337118e-05, "loss": 3.4532, "step": 44000 }, { "epoch": 0.3890036913403628, "grad_norm": 2.848560094833374, "learning_rate": 9.857655006143259e-05, "loss": 3.3183, "step": 44050 }, { "epoch": 0.38944523923064694, "grad_norm": 2.6281092166900635, "learning_rate": 9.85732560073674e-05, "loss": 3.348, "step": 44100 }, { "epoch": 0.3898867871209311, "grad_norm": 1.351362943649292, "learning_rate": 9.856995820143006e-05, "loss": 3.2875, "step": 44150 }, { "epoch": 0.39032833501121533, "grad_norm": 11.539451599121094, "learning_rate": 9.856665664387535e-05, "loss": 3.8226, "step": 44200 }, { "epoch": 0.3907698829014995, "grad_norm": 2.561694383621216, "learning_rate": 9.856335133495822e-05, "loss": 3.7136, "step": 44250 }, { "epoch": 0.39121143079178367, "grad_norm": 1.2988532781600952, "learning_rate": 9.856004227493402e-05, "loss": 3.5633, "step": 44300 }, { "epoch": 0.39165297868206783, "grad_norm": 2.9327540397644043, "learning_rate": 9.855672946405831e-05, "loss": 3.2624, "step": 44350 }, { "epoch": 0.39209452657235205, "grad_norm": 2.4923012256622314, "learning_rate": 9.855341290258703e-05, "loss": 3.0418, "step": 44400 }, { "epoch": 0.3925360744626362, "grad_norm": 0.7568151354789734, "learning_rate": 9.85500925907763e-05, "loss": 3.5547, "step": 44450 }, { "epoch": 0.3929776223529204, "grad_norm": 1.1965677738189697, "learning_rate": 9.854676852888262e-05, "loss": 3.4921, "step": 44500 }, { "epoch": 0.39341917024320455, "grad_norm": 0.5616858601570129, "learning_rate": 9.854344071716272e-05, "loss": 3.2839, "step": 44550 }, { "epoch": 0.3938607181334888, "grad_norm": 3.225074052810669, "learning_rate": 9.854010915587369e-05, "loss": 3.4744, "step": 44600 }, { "epoch": 0.39430226602377294, "grad_norm": 3.9845869541168213, "learning_rate": 9.853677384527282e-05, "loss": 3.4246, "step": 44650 }, { "epoch": 0.3947438139140571, "grad_norm": 3.029383897781372, "learning_rate": 9.853343478561773e-05, "loss": 3.408, "step": 44700 }, { "epoch": 0.3951853618043413, "grad_norm": 1.8023637533187866, "learning_rate": 9.853009197716638e-05, "loss": 3.1036, "step": 44750 }, { "epoch": 0.3956269096946255, "grad_norm": 0.8640767931938171, "learning_rate": 9.852674542017694e-05, "loss": 3.3368, "step": 44800 }, { "epoch": 0.39606845758490966, "grad_norm": 0.6954994797706604, "learning_rate": 9.852339511490792e-05, "loss": 2.9964, "step": 44850 }, { "epoch": 0.3965100054751938, "grad_norm": 8.39139461517334, "learning_rate": 9.85200410616181e-05, "loss": 2.9249, "step": 44900 }, { "epoch": 0.396951553365478, "grad_norm": 4.7654900550842285, "learning_rate": 9.851668326056652e-05, "loss": 3.0992, "step": 44950 }, { "epoch": 0.3973931012557622, "grad_norm": 1.2103265523910522, "learning_rate": 9.851332171201258e-05, "loss": 3.3672, "step": 45000 }, { "epoch": 0.3973931012557622, "eval_asr_loss": 0.8913320451830068, "eval_loss": 3.0676581859588623, "eval_runtime": 20.2524, "eval_samples_per_second": 37.921, "eval_steps_per_second": 9.48, "eval_tts_loss": 5.97925205245277, "step": 45000 }, { "epoch": 0.3978346491460464, "grad_norm": 2.4184420108795166, "learning_rate": 9.850995641621592e-05, "loss": 3.6283, "step": 45050 }, { "epoch": 0.39827619703633055, "grad_norm": 1.4427285194396973, "learning_rate": 9.850658737343648e-05, "loss": 3.5224, "step": 45100 }, { "epoch": 0.3987177449266147, "grad_norm": 6.392005920410156, "learning_rate": 9.850321458393447e-05, "loss": 3.0993, "step": 45150 }, { "epoch": 0.39915929281689894, "grad_norm": 1.4388530254364014, "learning_rate": 9.849983804797046e-05, "loss": 3.0099, "step": 45200 }, { "epoch": 0.3996008407071831, "grad_norm": 2.679241180419922, "learning_rate": 9.849645776580522e-05, "loss": 3.3363, "step": 45250 }, { "epoch": 0.40004238859746727, "grad_norm": 1.61045503616333, "learning_rate": 9.849307373769985e-05, "loss": 3.1612, "step": 45300 }, { "epoch": 0.40048393648775144, "grad_norm": 1.9430288076400757, "learning_rate": 9.848968596391574e-05, "loss": 3.4328, "step": 45350 }, { "epoch": 0.40092548437803566, "grad_norm": 1.6506637334823608, "learning_rate": 9.848629444471456e-05, "loss": 3.6278, "step": 45400 }, { "epoch": 0.4013670322683198, "grad_norm": 2.5751986503601074, "learning_rate": 9.848289918035831e-05, "loss": 3.3468, "step": 45450 }, { "epoch": 0.401808580158604, "grad_norm": 2.1925528049468994, "learning_rate": 9.84795001711092e-05, "loss": 3.3276, "step": 45500 }, { "epoch": 0.40225012804888816, "grad_norm": 3.269740581512451, "learning_rate": 9.847609741722981e-05, "loss": 2.7999, "step": 45550 }, { "epoch": 0.4026916759391724, "grad_norm": 4.853845596313477, "learning_rate": 9.847269091898294e-05, "loss": 3.1533, "step": 45600 }, { "epoch": 0.40313322382945654, "grad_norm": 2.3576786518096924, "learning_rate": 9.846928067663175e-05, "loss": 3.4268, "step": 45650 }, { "epoch": 0.4035747717197407, "grad_norm": 1.3526619672775269, "learning_rate": 9.846586669043964e-05, "loss": 3.156, "step": 45700 }, { "epoch": 0.4040163196100249, "grad_norm": 1.5202453136444092, "learning_rate": 9.846244896067031e-05, "loss": 3.2914, "step": 45750 }, { "epoch": 0.4044578675003091, "grad_norm": 1.80320143699646, "learning_rate": 9.845902748758773e-05, "loss": 3.2984, "step": 45800 }, { "epoch": 0.40489941539059326, "grad_norm": 1.7634551525115967, "learning_rate": 9.845560227145621e-05, "loss": 3.1742, "step": 45850 }, { "epoch": 0.40534096328087743, "grad_norm": 2.447131633758545, "learning_rate": 9.84521733125403e-05, "loss": 3.0532, "step": 45900 }, { "epoch": 0.4057825111711616, "grad_norm": 2.7054214477539062, "learning_rate": 9.844874061110485e-05, "loss": 3.5135, "step": 45950 }, { "epoch": 0.4062240590614458, "grad_norm": 0.9877326488494873, "learning_rate": 9.844530416741505e-05, "loss": 3.4712, "step": 46000 }, { "epoch": 0.40666560695173, "grad_norm": 1.2479236125946045, "learning_rate": 9.84418639817363e-05, "loss": 3.5099, "step": 46050 }, { "epoch": 0.40710715484201415, "grad_norm": 3.176802635192871, "learning_rate": 9.843842005433431e-05, "loss": 3.5577, "step": 46100 }, { "epoch": 0.4075487027322983, "grad_norm": 1.3120348453521729, "learning_rate": 9.843497238547513e-05, "loss": 2.9955, "step": 46150 }, { "epoch": 0.40799025062258254, "grad_norm": 1.1844558715820312, "learning_rate": 9.843152097542505e-05, "loss": 3.6275, "step": 46200 }, { "epoch": 0.4084317985128667, "grad_norm": 2.701456308364868, "learning_rate": 9.842806582445067e-05, "loss": 3.8021, "step": 46250 }, { "epoch": 0.4088733464031509, "grad_norm": 2.6151137351989746, "learning_rate": 9.842460693281885e-05, "loss": 3.6685, "step": 46300 }, { "epoch": 0.40931489429343504, "grad_norm": 2.728315591812134, "learning_rate": 9.842114430079679e-05, "loss": 3.4043, "step": 46350 }, { "epoch": 0.40975644218371926, "grad_norm": 2.2898852825164795, "learning_rate": 9.841767792865191e-05, "loss": 3.6381, "step": 46400 }, { "epoch": 0.4101979900740034, "grad_norm": 1.9836546182632446, "learning_rate": 9.8414207816652e-05, "loss": 3.0714, "step": 46450 }, { "epoch": 0.4106395379642876, "grad_norm": 2.8951032161712646, "learning_rate": 9.841073396506506e-05, "loss": 3.4733, "step": 46500 }, { "epoch": 0.41108108585457176, "grad_norm": 1.4405927658081055, "learning_rate": 9.840725637415945e-05, "loss": 3.2872, "step": 46550 }, { "epoch": 0.411522633744856, "grad_norm": 3.1269359588623047, "learning_rate": 9.840377504420376e-05, "loss": 3.4911, "step": 46600 }, { "epoch": 0.41196418163514015, "grad_norm": 3.698716163635254, "learning_rate": 9.840028997546689e-05, "loss": 3.1268, "step": 46650 }, { "epoch": 0.4124057295254243, "grad_norm": 2.955763578414917, "learning_rate": 9.839680116821804e-05, "loss": 3.5459, "step": 46700 }, { "epoch": 0.4128472774157085, "grad_norm": 0.848429799079895, "learning_rate": 9.83933086227267e-05, "loss": 3.2194, "step": 46750 }, { "epoch": 0.4132888253059927, "grad_norm": 8.197107315063477, "learning_rate": 9.838981233926264e-05, "loss": 3.2381, "step": 46800 }, { "epoch": 0.41373037319627687, "grad_norm": 1.4284354448318481, "learning_rate": 9.83863123180959e-05, "loss": 3.4783, "step": 46850 }, { "epoch": 0.41417192108656103, "grad_norm": 1.3862097263336182, "learning_rate": 9.838280855949684e-05, "loss": 3.3749, "step": 46900 }, { "epoch": 0.41461346897684526, "grad_norm": 1.5118871927261353, "learning_rate": 9.837930106373609e-05, "loss": 3.2749, "step": 46950 }, { "epoch": 0.4150550168671294, "grad_norm": 4.723493576049805, "learning_rate": 9.83757898310846e-05, "loss": 2.9933, "step": 47000 }, { "epoch": 0.4154965647574136, "grad_norm": 6.659562110900879, "learning_rate": 9.837227486181353e-05, "loss": 3.4323, "step": 47050 }, { "epoch": 0.41593811264769776, "grad_norm": 2.5123307704925537, "learning_rate": 9.836875615619443e-05, "loss": 2.9715, "step": 47100 }, { "epoch": 0.416379660537982, "grad_norm": 2.6539173126220703, "learning_rate": 9.836523371449906e-05, "loss": 3.1207, "step": 47150 }, { "epoch": 0.41682120842826614, "grad_norm": 3.233375072479248, "learning_rate": 9.836170753699952e-05, "loss": 3.1869, "step": 47200 }, { "epoch": 0.4172627563185503, "grad_norm": 1.2898887395858765, "learning_rate": 9.835817762396816e-05, "loss": 3.2155, "step": 47250 }, { "epoch": 0.4177043042088345, "grad_norm": 0.9073876738548279, "learning_rate": 9.835464397567765e-05, "loss": 3.363, "step": 47300 }, { "epoch": 0.4181458520991187, "grad_norm": 2.146453380584717, "learning_rate": 9.835110659240092e-05, "loss": 3.6663, "step": 47350 }, { "epoch": 0.41858739998940286, "grad_norm": 4.763782501220703, "learning_rate": 9.834756547441122e-05, "loss": 3.5419, "step": 47400 }, { "epoch": 0.41902894787968703, "grad_norm": 2.6837880611419678, "learning_rate": 9.834402062198207e-05, "loss": 3.134, "step": 47450 }, { "epoch": 0.4194704957699712, "grad_norm": 2.5617268085479736, "learning_rate": 9.834047203538726e-05, "loss": 3.4525, "step": 47500 }, { "epoch": 0.4199120436602554, "grad_norm": 2.1954944133758545, "learning_rate": 9.83369197149009e-05, "loss": 3.7715, "step": 47550 }, { "epoch": 0.4203535915505396, "grad_norm": 1.5438072681427002, "learning_rate": 9.833336366079737e-05, "loss": 3.6781, "step": 47600 }, { "epoch": 0.42079513944082375, "grad_norm": 1.757779598236084, "learning_rate": 9.832980387335138e-05, "loss": 3.4219, "step": 47650 }, { "epoch": 0.4212366873311079, "grad_norm": 2.8832337856292725, "learning_rate": 9.832624035283785e-05, "loss": 2.9347, "step": 47700 }, { "epoch": 0.42167823522139214, "grad_norm": 2.0708539485931396, "learning_rate": 9.832274448117773e-05, "loss": 3.1908, "step": 47750 }, { "epoch": 0.4221197831116763, "grad_norm": 1.883726716041565, "learning_rate": 9.831917357000283e-05, "loss": 3.5663, "step": 47800 }, { "epoch": 0.42256133100196047, "grad_norm": 1.2137542963027954, "learning_rate": 9.83155989265815e-05, "loss": 3.3599, "step": 47850 }, { "epoch": 0.42300287889224464, "grad_norm": 1.1021403074264526, "learning_rate": 9.831202055118986e-05, "loss": 3.3176, "step": 47900 }, { "epoch": 0.42344442678252886, "grad_norm": 3.1852636337280273, "learning_rate": 9.830843844410431e-05, "loss": 3.2934, "step": 47950 }, { "epoch": 0.423885974672813, "grad_norm": 2.2362117767333984, "learning_rate": 9.830492435893767e-05, "loss": 3.5098, "step": 48000 }, { "epoch": 0.423885974672813, "eval_asr_loss": 0.8946412356905893, "eval_loss": 3.0450656414031982, "eval_runtime": 20.5825, "eval_samples_per_second": 37.313, "eval_steps_per_second": 9.328, "eval_tts_loss": 5.9677758015827616, "step": 48000 }, { "epoch": 0.4243275225630972, "grad_norm": 0.6300576329231262, "learning_rate": 9.830133486391473e-05, "loss": 3.5818, "step": 48050 }, { "epoch": 0.42476907045338136, "grad_norm": 1.3167023658752441, "learning_rate": 9.829774163802325e-05, "loss": 3.4292, "step": 48100 }, { "epoch": 0.4252106183436656, "grad_norm": 2.811063051223755, "learning_rate": 9.829414468154079e-05, "loss": 3.7977, "step": 48150 }, { "epoch": 0.42565216623394975, "grad_norm": 2.503530502319336, "learning_rate": 9.829054399474516e-05, "loss": 3.6608, "step": 48200 }, { "epoch": 0.4260937141242339, "grad_norm": 2.051328420639038, "learning_rate": 9.828693957791452e-05, "loss": 3.3447, "step": 48250 }, { "epoch": 0.4265352620145181, "grad_norm": 0.9550831913948059, "learning_rate": 9.828333143132724e-05, "loss": 3.2968, "step": 48300 }, { "epoch": 0.4269768099048023, "grad_norm": 3.1451058387756348, "learning_rate": 9.827971955526205e-05, "loss": 3.3196, "step": 48350 }, { "epoch": 0.42741835779508647, "grad_norm": 1.7226669788360596, "learning_rate": 9.827610394999791e-05, "loss": 3.1922, "step": 48400 }, { "epoch": 0.42785990568537063, "grad_norm": 2.289461851119995, "learning_rate": 9.827248461581412e-05, "loss": 3.327, "step": 48450 }, { "epoch": 0.4283014535756548, "grad_norm": 3.5129716396331787, "learning_rate": 9.826886155299023e-05, "loss": 3.4937, "step": 48500 }, { "epoch": 0.428743001465939, "grad_norm": 2.9753642082214355, "learning_rate": 9.826523476180609e-05, "loss": 3.3748, "step": 48550 }, { "epoch": 0.4291845493562232, "grad_norm": 1.0942645072937012, "learning_rate": 9.826160424254185e-05, "loss": 3.391, "step": 48600 }, { "epoch": 0.42962609724650735, "grad_norm": 2.1059162616729736, "learning_rate": 9.825796999547792e-05, "loss": 3.7639, "step": 48650 }, { "epoch": 0.4300676451367915, "grad_norm": 1.9335929155349731, "learning_rate": 9.825433202089503e-05, "loss": 2.9592, "step": 48700 }, { "epoch": 0.43050919302707574, "grad_norm": 0.7627319693565369, "learning_rate": 9.825069031907416e-05, "loss": 3.3292, "step": 48750 }, { "epoch": 0.4309507409173599, "grad_norm": 3.433350086212158, "learning_rate": 9.824704489029664e-05, "loss": 3.3546, "step": 48800 }, { "epoch": 0.4313922888076441, "grad_norm": 2.693155527114868, "learning_rate": 9.824339573484402e-05, "loss": 3.4344, "step": 48850 }, { "epoch": 0.43183383669792824, "grad_norm": 1.1923327445983887, "learning_rate": 9.823974285299817e-05, "loss": 3.3486, "step": 48900 }, { "epoch": 0.43227538458821246, "grad_norm": 1.9175432920455933, "learning_rate": 9.823608624504123e-05, "loss": 3.4314, "step": 48950 }, { "epoch": 0.43271693247849663, "grad_norm": 0.7622149586677551, "learning_rate": 9.823242591125568e-05, "loss": 3.1681, "step": 49000 }, { "epoch": 0.4331584803687808, "grad_norm": 2.3033721446990967, "learning_rate": 9.822876185192421e-05, "loss": 3.4592, "step": 49050 }, { "epoch": 0.43360002825906496, "grad_norm": 0.886083722114563, "learning_rate": 9.822509406732987e-05, "loss": 3.6985, "step": 49100 }, { "epoch": 0.4340415761493492, "grad_norm": 6.617265701293945, "learning_rate": 9.822142255775594e-05, "loss": 3.1759, "step": 49150 }, { "epoch": 0.43448312403963335, "grad_norm": 4.3931121826171875, "learning_rate": 9.821774732348602e-05, "loss": 3.4327, "step": 49200 }, { "epoch": 0.4349246719299175, "grad_norm": 1.2443506717681885, "learning_rate": 9.8214068364804e-05, "loss": 3.5825, "step": 49250 }, { "epoch": 0.4353662198202017, "grad_norm": 2.791949510574341, "learning_rate": 9.821038568199403e-05, "loss": 3.1509, "step": 49300 }, { "epoch": 0.4358077677104859, "grad_norm": 1.996881127357483, "learning_rate": 9.820669927534061e-05, "loss": 2.9644, "step": 49350 }, { "epoch": 0.43624931560077007, "grad_norm": 1.907235026359558, "learning_rate": 9.820300914512842e-05, "loss": 3.3867, "step": 49400 }, { "epoch": 0.43669086349105424, "grad_norm": 3.5670182704925537, "learning_rate": 9.819931529164254e-05, "loss": 3.4446, "step": 49450 }, { "epoch": 0.4371324113813384, "grad_norm": 1.6872813701629639, "learning_rate": 9.819561771516826e-05, "loss": 2.9724, "step": 49500 }, { "epoch": 0.4375739592716226, "grad_norm": 3.4050118923187256, "learning_rate": 9.819191641599121e-05, "loss": 3.2879, "step": 49550 }, { "epoch": 0.4380155071619068, "grad_norm": 2.1053950786590576, "learning_rate": 9.818821139439727e-05, "loss": 3.4533, "step": 49600 }, { "epoch": 0.43845705505219096, "grad_norm": 1.7744102478027344, "learning_rate": 9.818450265067261e-05, "loss": 3.3481, "step": 49650 }, { "epoch": 0.4388986029424751, "grad_norm": 2.2178120613098145, "learning_rate": 9.818079018510374e-05, "loss": 3.3662, "step": 49700 }, { "epoch": 0.43934015083275935, "grad_norm": 2.2515435218811035, "learning_rate": 9.817707399797736e-05, "loss": 3.336, "step": 49750 }, { "epoch": 0.4397816987230435, "grad_norm": 3.4062695503234863, "learning_rate": 9.817335408958056e-05, "loss": 3.2625, "step": 49800 }, { "epoch": 0.4402232466133277, "grad_norm": 1.6740872859954834, "learning_rate": 9.816963046020065e-05, "loss": 3.7033, "step": 49850 }, { "epoch": 0.44066479450361185, "grad_norm": 0.949266254901886, "learning_rate": 9.816590311012525e-05, "loss": 3.2875, "step": 49900 }, { "epoch": 0.44110634239389607, "grad_norm": 2.8111281394958496, "learning_rate": 9.816217203964228e-05, "loss": 3.2595, "step": 49950 }, { "epoch": 0.44154789028418023, "grad_norm": 1.117742657661438, "learning_rate": 9.815843724903993e-05, "loss": 3.2435, "step": 50000 }, { "epoch": 0.4419894381744644, "grad_norm": 2.1841189861297607, "learning_rate": 9.815469873860666e-05, "loss": 3.5207, "step": 50050 }, { "epoch": 0.44243098606474857, "grad_norm": 4.919894695281982, "learning_rate": 9.815095650863124e-05, "loss": 3.3703, "step": 50100 }, { "epoch": 0.4428725339550328, "grad_norm": 3.6214163303375244, "learning_rate": 9.814721055940275e-05, "loss": 3.5246, "step": 50150 }, { "epoch": 0.44331408184531695, "grad_norm": 3.0915398597717285, "learning_rate": 9.814346089121053e-05, "loss": 3.7959, "step": 50200 }, { "epoch": 0.4437556297356011, "grad_norm": 1.7853915691375732, "learning_rate": 9.813970750434419e-05, "loss": 3.4, "step": 50250 }, { "epoch": 0.4441971776258853, "grad_norm": 0.6693574786186218, "learning_rate": 9.813595039909367e-05, "loss": 3.2674, "step": 50300 }, { "epoch": 0.4446387255161695, "grad_norm": 7.433414936065674, "learning_rate": 9.813218957574914e-05, "loss": 3.1686, "step": 50350 }, { "epoch": 0.4450802734064537, "grad_norm": 2.307650566101074, "learning_rate": 9.812842503460114e-05, "loss": 3.2572, "step": 50400 }, { "epoch": 0.44552182129673784, "grad_norm": 4.549615383148193, "learning_rate": 9.812465677594041e-05, "loss": 3.5429, "step": 50450 }, { "epoch": 0.445963369187022, "grad_norm": 1.0888493061065674, "learning_rate": 9.812088480005804e-05, "loss": 3.3837, "step": 50500 }, { "epoch": 0.44640491707730623, "grad_norm": 2.1504311561584473, "learning_rate": 9.811718465752564e-05, "loss": 3.7276, "step": 50550 }, { "epoch": 0.4468464649675904, "grad_norm": 3.0695831775665283, "learning_rate": 9.811340532240422e-05, "loss": 3.4433, "step": 50600 }, { "epoch": 0.44728801285787456, "grad_norm": 1.7197152376174927, "learning_rate": 9.810962227093022e-05, "loss": 3.5762, "step": 50650 }, { "epoch": 0.44772956074815873, "grad_norm": 1.3767517805099487, "learning_rate": 9.810583550339588e-05, "loss": 3.0935, "step": 50700 }, { "epoch": 0.44817110863844295, "grad_norm": 1.5721153020858765, "learning_rate": 9.810204502009364e-05, "loss": 3.4474, "step": 50750 }, { "epoch": 0.4486126565287271, "grad_norm": 4.190256118774414, "learning_rate": 9.809825082131633e-05, "loss": 3.1628, "step": 50800 }, { "epoch": 0.4490542044190113, "grad_norm": 2.406953811645508, "learning_rate": 9.809445290735702e-05, "loss": 3.2441, "step": 50850 }, { "epoch": 0.44949575230929545, "grad_norm": 4.159549236297607, "learning_rate": 9.809065127850903e-05, "loss": 3.0559, "step": 50900 }, { "epoch": 0.44993730019957967, "grad_norm": 1.8754595518112183, "learning_rate": 9.808684593506605e-05, "loss": 3.7703, "step": 50950 }, { "epoch": 0.45037884808986384, "grad_norm": 1.29695725440979, "learning_rate": 9.808303687732196e-05, "loss": 3.1498, "step": 51000 }, { "epoch": 0.45037884808986384, "eval_asr_loss": 0.9004947789321158, "eval_loss": 3.0421018600463867, "eval_runtime": 20.6473, "eval_samples_per_second": 37.196, "eval_steps_per_second": 9.299, "eval_tts_loss": 5.865643483398774, "step": 51000 }, { "epoch": 0.450820395980148, "grad_norm": 2.4158124923706055, "learning_rate": 9.807922410557102e-05, "loss": 3.3356, "step": 51050 }, { "epoch": 0.45126194387043217, "grad_norm": 0.9025440812110901, "learning_rate": 9.807540762010772e-05, "loss": 3.336, "step": 51100 }, { "epoch": 0.4517034917607164, "grad_norm": 5.0518364906311035, "learning_rate": 9.807158742122684e-05, "loss": 3.2288, "step": 51150 }, { "epoch": 0.45214503965100056, "grad_norm": 1.4660353660583496, "learning_rate": 9.806776350922346e-05, "loss": 3.0611, "step": 51200 }, { "epoch": 0.4525865875412847, "grad_norm": 1.7514241933822632, "learning_rate": 9.806393588439297e-05, "loss": 3.304, "step": 51250 }, { "epoch": 0.4530281354315689, "grad_norm": 1.4853622913360596, "learning_rate": 9.806010454703099e-05, "loss": 3.436, "step": 51300 }, { "epoch": 0.4534696833218531, "grad_norm": 2.174837112426758, "learning_rate": 9.805626949743347e-05, "loss": 3.6087, "step": 51350 }, { "epoch": 0.4539112312121373, "grad_norm": 1.5965477228164673, "learning_rate": 9.805243073589665e-05, "loss": 3.4667, "step": 51400 }, { "epoch": 0.45435277910242144, "grad_norm": 2.395080804824829, "learning_rate": 9.804858826271703e-05, "loss": 3.5994, "step": 51450 }, { "epoch": 0.4547943269927056, "grad_norm": 4.734639644622803, "learning_rate": 9.804474207819139e-05, "loss": 3.2731, "step": 51500 }, { "epoch": 0.45523587488298983, "grad_norm": 1.91262948513031, "learning_rate": 9.804089218261684e-05, "loss": 3.4359, "step": 51550 }, { "epoch": 0.455677422773274, "grad_norm": 4.3235321044921875, "learning_rate": 9.803703857629075e-05, "loss": 3.6903, "step": 51600 }, { "epoch": 0.45611897066355817, "grad_norm": 2.399240016937256, "learning_rate": 9.803318125951075e-05, "loss": 3.1958, "step": 51650 }, { "epoch": 0.45656051855384233, "grad_norm": 3.3152384757995605, "learning_rate": 9.802932023257483e-05, "loss": 3.4387, "step": 51700 }, { "epoch": 0.45700206644412655, "grad_norm": 3.167579412460327, "learning_rate": 9.802545549578119e-05, "loss": 3.6247, "step": 51750 }, { "epoch": 0.4574436143344107, "grad_norm": 1.6194385290145874, "learning_rate": 9.802158704942837e-05, "loss": 3.3378, "step": 51800 }, { "epoch": 0.4578851622246949, "grad_norm": 1.213178277015686, "learning_rate": 9.801771489381515e-05, "loss": 3.0608, "step": 51850 }, { "epoch": 0.45832671011497905, "grad_norm": 4.134045600891113, "learning_rate": 9.8013916582878e-05, "loss": 3.1588, "step": 51900 }, { "epoch": 0.4587682580052633, "grad_norm": 3.060990571975708, "learning_rate": 9.801003708381187e-05, "loss": 3.3848, "step": 51950 }, { "epoch": 0.45920980589554744, "grad_norm": 2.216519594192505, "learning_rate": 9.800615387637748e-05, "loss": 3.2701, "step": 52000 }, { "epoch": 0.4596513537858316, "grad_norm": 2.748791456222534, "learning_rate": 9.80022669608748e-05, "loss": 3.6429, "step": 52050 }, { "epoch": 0.4600929016761158, "grad_norm": 2.1832919120788574, "learning_rate": 9.799837633760403e-05, "loss": 3.4686, "step": 52100 }, { "epoch": 0.4605344495664, "grad_norm": 10.31329345703125, "learning_rate": 9.799448200686569e-05, "loss": 3.4104, "step": 52150 }, { "epoch": 0.46097599745668416, "grad_norm": 3.921966791152954, "learning_rate": 9.799058396896062e-05, "loss": 3.1034, "step": 52200 }, { "epoch": 0.4614175453469683, "grad_norm": 1.399453043937683, "learning_rate": 9.798668222418988e-05, "loss": 2.6928, "step": 52250 }, { "epoch": 0.4618590932372525, "grad_norm": 2.0908732414245605, "learning_rate": 9.798277677285483e-05, "loss": 3.2176, "step": 52300 }, { "epoch": 0.4623006411275367, "grad_norm": 1.981628656387329, "learning_rate": 9.797886761525719e-05, "loss": 3.1046, "step": 52350 }, { "epoch": 0.4627421890178209, "grad_norm": 2.1923952102661133, "learning_rate": 9.797495475169886e-05, "loss": 2.9833, "step": 52400 }, { "epoch": 0.46318373690810505, "grad_norm": 4.177052021026611, "learning_rate": 9.79710381824821e-05, "loss": 2.9391, "step": 52450 }, { "epoch": 0.4636252847983892, "grad_norm": 1.3210045099258423, "learning_rate": 9.796711790790941e-05, "loss": 3.5919, "step": 52500 }, { "epoch": 0.46406683268867344, "grad_norm": 1.5212137699127197, "learning_rate": 9.796319392828361e-05, "loss": 3.4421, "step": 52550 }, { "epoch": 0.4645083805789576, "grad_norm": 9.704986572265625, "learning_rate": 9.795926624390781e-05, "loss": 3.2297, "step": 52600 }, { "epoch": 0.46494992846924177, "grad_norm": 0.9471487402915955, "learning_rate": 9.795533485508537e-05, "loss": 3.6265, "step": 52650 }, { "epoch": 0.46539147635952594, "grad_norm": 2.34822154045105, "learning_rate": 9.795139976211996e-05, "loss": 3.2869, "step": 52700 }, { "epoch": 0.46583302424981016, "grad_norm": 1.4003323316574097, "learning_rate": 9.794746096531554e-05, "loss": 3.215, "step": 52750 }, { "epoch": 0.4662745721400943, "grad_norm": 0.5006351470947266, "learning_rate": 9.794351846497634e-05, "loss": 3.3245, "step": 52800 }, { "epoch": 0.4667161200303785, "grad_norm": 1.4579681158065796, "learning_rate": 9.793957226140688e-05, "loss": 3.0416, "step": 52850 }, { "epoch": 0.46715766792066266, "grad_norm": 2.5864365100860596, "learning_rate": 9.793562235491198e-05, "loss": 3.2932, "step": 52900 }, { "epoch": 0.4675992158109469, "grad_norm": 1.7026253938674927, "learning_rate": 9.793166874579673e-05, "loss": 3.1959, "step": 52950 }, { "epoch": 0.46804076370123104, "grad_norm": 5.3310394287109375, "learning_rate": 9.792771143436654e-05, "loss": 3.2207, "step": 53000 }, { "epoch": 0.4684823115915152, "grad_norm": 1.704520583152771, "learning_rate": 9.792375042092704e-05, "loss": 3.1596, "step": 53050 }, { "epoch": 0.4689238594817994, "grad_norm": 2.3095810413360596, "learning_rate": 9.79197857057842e-05, "loss": 3.48, "step": 53100 }, { "epoch": 0.4693654073720836, "grad_norm": 3.7878098487854004, "learning_rate": 9.791581728924427e-05, "loss": 3.4366, "step": 53150 }, { "epoch": 0.46980695526236776, "grad_norm": 1.6754798889160156, "learning_rate": 9.791184517161377e-05, "loss": 3.1798, "step": 53200 }, { "epoch": 0.47024850315265193, "grad_norm": 2.6600844860076904, "learning_rate": 9.79078693531995e-05, "loss": 3.378, "step": 53250 }, { "epoch": 0.4706900510429361, "grad_norm": 1.902320146560669, "learning_rate": 9.790388983430859e-05, "loss": 2.8402, "step": 53300 }, { "epoch": 0.4711315989332203, "grad_norm": 1.9587918519973755, "learning_rate": 9.789990661524839e-05, "loss": 3.6821, "step": 53350 }, { "epoch": 0.4715731468235045, "grad_norm": 0.7448610067367554, "learning_rate": 9.789591969632658e-05, "loss": 3.3176, "step": 53400 }, { "epoch": 0.47201469471378865, "grad_norm": 4.580758094787598, "learning_rate": 9.789192907785111e-05, "loss": 3.2808, "step": 53450 }, { "epoch": 0.4724562426040728, "grad_norm": 2.524980068206787, "learning_rate": 9.788793476013025e-05, "loss": 3.3751, "step": 53500 }, { "epoch": 0.47289779049435704, "grad_norm": 3.0498645305633545, "learning_rate": 9.788393674347249e-05, "loss": 3.1801, "step": 53550 }, { "epoch": 0.4733393383846412, "grad_norm": 2.230107545852661, "learning_rate": 9.787993502818669e-05, "loss": 3.7616, "step": 53600 }, { "epoch": 0.4737808862749254, "grad_norm": 0.8255630731582642, "learning_rate": 9.78759296145819e-05, "loss": 3.0498, "step": 53650 }, { "epoch": 0.47422243416520954, "grad_norm": 4.1301727294921875, "learning_rate": 9.787192050296752e-05, "loss": 3.281, "step": 53700 }, { "epoch": 0.47466398205549376, "grad_norm": 3.9466300010681152, "learning_rate": 9.786790769365322e-05, "loss": 3.1479, "step": 53750 }, { "epoch": 0.4751055299457779, "grad_norm": 1.8435554504394531, "learning_rate": 9.786389118694897e-05, "loss": 3.4114, "step": 53800 }, { "epoch": 0.4755470778360621, "grad_norm": 0.7896223664283752, "learning_rate": 9.785987098316499e-05, "loss": 3.4087, "step": 53850 }, { "epoch": 0.47598862572634626, "grad_norm": 9.002167701721191, "learning_rate": 9.785584708261182e-05, "loss": 3.2257, "step": 53900 }, { "epoch": 0.4764301736166305, "grad_norm": 0.6309633851051331, "learning_rate": 9.785181948560028e-05, "loss": 3.1985, "step": 53950 }, { "epoch": 0.47687172150691465, "grad_norm": 3.459434747695923, "learning_rate": 9.784778819244144e-05, "loss": 3.3152, "step": 54000 }, { "epoch": 0.47687172150691465, "eval_asr_loss": 0.8963724033623304, "eval_loss": 3.0343711376190186, "eval_runtime": 20.7535, "eval_samples_per_second": 37.006, "eval_steps_per_second": 9.251, "eval_tts_loss": 5.906121494862091, "step": 54000 }, { "epoch": 0.4773132693971988, "grad_norm": 2.1301255226135254, "learning_rate": 9.78437532034467e-05, "loss": 3.7664, "step": 54050 }, { "epoch": 0.477754817287483, "grad_norm": 2.864131450653076, "learning_rate": 9.783971451892773e-05, "loss": 3.0815, "step": 54100 }, { "epoch": 0.4781963651777672, "grad_norm": 2.723444938659668, "learning_rate": 9.783567213919649e-05, "loss": 3.2961, "step": 54150 }, { "epoch": 0.47863791306805137, "grad_norm": 2.7087883949279785, "learning_rate": 9.783162606456521e-05, "loss": 3.4694, "step": 54200 }, { "epoch": 0.47907946095833553, "grad_norm": 2.582592248916626, "learning_rate": 9.782757629534642e-05, "loss": 3.4529, "step": 54250 }, { "epoch": 0.4795210088486197, "grad_norm": 1.1580854654312134, "learning_rate": 9.782352283185293e-05, "loss": 3.357, "step": 54300 }, { "epoch": 0.4799625567389039, "grad_norm": 3.442847490310669, "learning_rate": 9.781946567439781e-05, "loss": 3.2244, "step": 54350 }, { "epoch": 0.4804041046291881, "grad_norm": 4.473285675048828, "learning_rate": 9.781540482329447e-05, "loss": 3.2563, "step": 54400 }, { "epoch": 0.48084565251947226, "grad_norm": 2.0757851600646973, "learning_rate": 9.78113402788566e-05, "loss": 3.3494, "step": 54450 }, { "epoch": 0.4812872004097564, "grad_norm": 1.0765964984893799, "learning_rate": 9.78072720413981e-05, "loss": 2.973, "step": 54500 }, { "epoch": 0.48172874830004064, "grad_norm": 2.187049150466919, "learning_rate": 9.780320011123322e-05, "loss": 3.5087, "step": 54550 }, { "epoch": 0.4821702961903248, "grad_norm": 2.3897063732147217, "learning_rate": 9.779912448867649e-05, "loss": 3.3432, "step": 54600 }, { "epoch": 0.482611844080609, "grad_norm": 1.4350836277008057, "learning_rate": 9.779504517404274e-05, "loss": 3.547, "step": 54650 }, { "epoch": 0.48305339197089314, "grad_norm": 1.5999610424041748, "learning_rate": 9.779096216764703e-05, "loss": 3.1709, "step": 54700 }, { "epoch": 0.48349493986117736, "grad_norm": 4.64166784286499, "learning_rate": 9.778687546980474e-05, "loss": 3.6566, "step": 54750 }, { "epoch": 0.48393648775146153, "grad_norm": 2.923461675643921, "learning_rate": 9.778278508083154e-05, "loss": 3.4523, "step": 54800 }, { "epoch": 0.4843780356417457, "grad_norm": 1.1621273756027222, "learning_rate": 9.77786910010434e-05, "loss": 3.4142, "step": 54850 }, { "epoch": 0.48481958353202986, "grad_norm": 2.0562987327575684, "learning_rate": 9.77745932307565e-05, "loss": 3.2347, "step": 54900 }, { "epoch": 0.4852611314223141, "grad_norm": 2.8436999320983887, "learning_rate": 9.777049177028742e-05, "loss": 3.4778, "step": 54950 }, { "epoch": 0.48570267931259825, "grad_norm": 3.21048641204834, "learning_rate": 9.776638661995292e-05, "loss": 3.4173, "step": 55000 }, { "epoch": 0.4861442272028824, "grad_norm": 2.347586154937744, "learning_rate": 9.77622777800701e-05, "loss": 3.6781, "step": 55050 }, { "epoch": 0.4865857750931666, "grad_norm": 1.2790504693984985, "learning_rate": 9.775816525095633e-05, "loss": 3.5449, "step": 55100 }, { "epoch": 0.4870273229834508, "grad_norm": 2.234579563140869, "learning_rate": 9.775404903292929e-05, "loss": 3.403, "step": 55150 }, { "epoch": 0.48746887087373497, "grad_norm": 3.1236841678619385, "learning_rate": 9.774992912630689e-05, "loss": 3.5316, "step": 55200 }, { "epoch": 0.48791041876401914, "grad_norm": 2.067030906677246, "learning_rate": 9.774580553140736e-05, "loss": 3.3101, "step": 55250 }, { "epoch": 0.4883519666543033, "grad_norm": 2.3011932373046875, "learning_rate": 9.774167824854925e-05, "loss": 3.9037, "step": 55300 }, { "epoch": 0.4887935145445875, "grad_norm": 0.8518275618553162, "learning_rate": 9.773754727805132e-05, "loss": 3.7854, "step": 55350 }, { "epoch": 0.4892350624348717, "grad_norm": 2.5740087032318115, "learning_rate": 9.773341262023265e-05, "loss": 3.6587, "step": 55400 }, { "epoch": 0.48967661032515586, "grad_norm": 2.541947364807129, "learning_rate": 9.772927427541266e-05, "loss": 3.3046, "step": 55450 }, { "epoch": 0.49011815821544, "grad_norm": 10.81189250946045, "learning_rate": 9.772513224391093e-05, "loss": 3.5087, "step": 55500 }, { "epoch": 0.49055970610572425, "grad_norm": 2.659097909927368, "learning_rate": 9.772098652604745e-05, "loss": 3.7244, "step": 55550 }, { "epoch": 0.4910012539960084, "grad_norm": 1.6622788906097412, "learning_rate": 9.771683712214241e-05, "loss": 3.9154, "step": 55600 }, { "epoch": 0.4914428018862926, "grad_norm": 4.244743347167969, "learning_rate": 9.771268403251634e-05, "loss": 3.1095, "step": 55650 }, { "epoch": 0.49188434977657675, "grad_norm": 2.233381986618042, "learning_rate": 9.770852725749002e-05, "loss": 2.7214, "step": 55700 }, { "epoch": 0.49232589766686097, "grad_norm": 3.992842674255371, "learning_rate": 9.770436679738452e-05, "loss": 2.8508, "step": 55750 }, { "epoch": 0.49276744555714513, "grad_norm": 2.8735084533691406, "learning_rate": 9.770020265252122e-05, "loss": 3.3086, "step": 55800 }, { "epoch": 0.4932089934474293, "grad_norm": 0.7846609950065613, "learning_rate": 9.769603482322173e-05, "loss": 3.0594, "step": 55850 }, { "epoch": 0.49365054133771347, "grad_norm": 3.95578670501709, "learning_rate": 9.769186330980802e-05, "loss": 3.2712, "step": 55900 }, { "epoch": 0.4940920892279977, "grad_norm": 1.6080371141433716, "learning_rate": 9.768768811260229e-05, "loss": 3.3038, "step": 55950 }, { "epoch": 0.49453363711828185, "grad_norm": 1.84148108959198, "learning_rate": 9.7683509231927e-05, "loss": 3.2961, "step": 56000 }, { "epoch": 0.494975185008566, "grad_norm": 3.487783670425415, "learning_rate": 9.7679326668105e-05, "loss": 3.0265, "step": 56050 }, { "epoch": 0.4954167328988502, "grad_norm": 1.8693515062332153, "learning_rate": 9.767514042145931e-05, "loss": 3.3785, "step": 56100 }, { "epoch": 0.4958582807891344, "grad_norm": 0.690427839756012, "learning_rate": 9.767095049231329e-05, "loss": 2.9811, "step": 56150 }, { "epoch": 0.4962998286794186, "grad_norm": 3.2560925483703613, "learning_rate": 9.766675688099059e-05, "loss": 3.3119, "step": 56200 }, { "epoch": 0.49674137656970274, "grad_norm": 2.757688522338867, "learning_rate": 9.766255958781512e-05, "loss": 3.3075, "step": 56250 }, { "epoch": 0.4971829244599869, "grad_norm": 1.4115214347839355, "learning_rate": 9.765835861311108e-05, "loss": 3.3612, "step": 56300 }, { "epoch": 0.49762447235027113, "grad_norm": 3.8688366413116455, "learning_rate": 9.765415395720298e-05, "loss": 3.1324, "step": 56350 }, { "epoch": 0.4980660202405553, "grad_norm": 3.989609479904175, "learning_rate": 9.764994562041559e-05, "loss": 3.1485, "step": 56400 }, { "epoch": 0.49850756813083946, "grad_norm": 2.1451547145843506, "learning_rate": 9.764573360307394e-05, "loss": 3.2281, "step": 56450 }, { "epoch": 0.49894911602112363, "grad_norm": 1.555163025856018, "learning_rate": 9.76415179055034e-05, "loss": 3.0442, "step": 56500 }, { "epoch": 0.49939066391140785, "grad_norm": 2.0306787490844727, "learning_rate": 9.76372985280296e-05, "loss": 3.3319, "step": 56550 }, { "epoch": 0.499832211801692, "grad_norm": 3.7980337142944336, "learning_rate": 9.763307547097844e-05, "loss": 3.6434, "step": 56600 }, { "epoch": 0.5002737596919762, "grad_norm": 4.4818525314331055, "learning_rate": 9.762884873467611e-05, "loss": 3.2294, "step": 56650 }, { "epoch": 0.5007153075822603, "grad_norm": 1.937999963760376, "learning_rate": 9.76246183194491e-05, "loss": 3.5661, "step": 56700 }, { "epoch": 0.5011568554725445, "grad_norm": 1.306862711906433, "learning_rate": 9.762038422562417e-05, "loss": 2.9257, "step": 56750 }, { "epoch": 0.5015984033628287, "grad_norm": 1.1766592264175415, "learning_rate": 9.761614645352835e-05, "loss": 3.3017, "step": 56800 }, { "epoch": 0.502039951253113, "grad_norm": 3.1086044311523438, "learning_rate": 9.7611905003489e-05, "loss": 3.5308, "step": 56850 }, { "epoch": 0.5024814991433971, "grad_norm": 2.8829550743103027, "learning_rate": 9.760765987583373e-05, "loss": 3.1106, "step": 56900 }, { "epoch": 0.5029230470336813, "grad_norm": 2.7359766960144043, "learning_rate": 9.760341107089044e-05, "loss": 3.509, "step": 56950 }, { "epoch": 0.5033645949239655, "grad_norm": 7.022033214569092, "learning_rate": 9.759915858898732e-05, "loss": 3.1945, "step": 57000 }, { "epoch": 0.5033645949239655, "eval_asr_loss": 0.9032422534916057, "eval_loss": 3.0304603576660156, "eval_runtime": 51.0735, "eval_samples_per_second": 15.037, "eval_steps_per_second": 3.759, "eval_tts_loss": 5.9164860032310695, "step": 57000 }, { "epoch": 0.5038061428142496, "grad_norm": 1.487955093383789, "learning_rate": 9.75949024304528e-05, "loss": 3.1603, "step": 57050 }, { "epoch": 0.5042476907045338, "grad_norm": 2.423590660095215, "learning_rate": 9.759064259561568e-05, "loss": 3.6678, "step": 57100 }, { "epoch": 0.504689238594818, "grad_norm": 1.796743631362915, "learning_rate": 9.758637908480497e-05, "loss": 3.4576, "step": 57150 }, { "epoch": 0.5051307864851021, "grad_norm": 6.50948429107666, "learning_rate": 9.758211189834999e-05, "loss": 3.2279, "step": 57200 }, { "epoch": 0.5055723343753864, "grad_norm": 1.0165084600448608, "learning_rate": 9.757784103658036e-05, "loss": 3.3161, "step": 57250 }, { "epoch": 0.5060138822656706, "grad_norm": 3.4719436168670654, "learning_rate": 9.757356649982595e-05, "loss": 3.6583, "step": 57300 }, { "epoch": 0.5064554301559547, "grad_norm": 2.7984747886657715, "learning_rate": 9.756928828841694e-05, "loss": 3.507, "step": 57350 }, { "epoch": 0.5068969780462389, "grad_norm": 4.911170959472656, "learning_rate": 9.75650064026838e-05, "loss": 3.0097, "step": 57400 }, { "epoch": 0.5073385259365231, "grad_norm": 3.57440185546875, "learning_rate": 9.756072084295725e-05, "loss": 2.9845, "step": 57450 }, { "epoch": 0.5077800738268072, "grad_norm": 1.4227149486541748, "learning_rate": 9.755651743023584e-05, "loss": 3.6389, "step": 57500 }, { "epoch": 0.5082216217170914, "grad_norm": 2.1636316776275635, "learning_rate": 9.755222459697919e-05, "loss": 2.9495, "step": 57550 }, { "epoch": 0.5086631696073756, "grad_norm": 5.281383991241455, "learning_rate": 9.754792809071644e-05, "loss": 3.388, "step": 57600 }, { "epoch": 0.5091047174976598, "grad_norm": 1.8442537784576416, "learning_rate": 9.754362791177944e-05, "loss": 3.7662, "step": 57650 }, { "epoch": 0.509546265387944, "grad_norm": 2.220319986343384, "learning_rate": 9.753932406050032e-05, "loss": 3.1771, "step": 57700 }, { "epoch": 0.5099878132782282, "grad_norm": 2.379897356033325, "learning_rate": 9.753501653721155e-05, "loss": 3.2315, "step": 57750 }, { "epoch": 0.5104293611685123, "grad_norm": 1.3445017337799072, "learning_rate": 9.753070534224584e-05, "loss": 3.4884, "step": 57800 }, { "epoch": 0.5108709090587965, "grad_norm": 1.7846251726150513, "learning_rate": 9.752639047593619e-05, "loss": 3.5118, "step": 57850 }, { "epoch": 0.5113124569490807, "grad_norm": 0.7859761714935303, "learning_rate": 9.752207193861589e-05, "loss": 3.4603, "step": 57900 }, { "epoch": 0.5117540048393648, "grad_norm": 1.91860830783844, "learning_rate": 9.751774973061851e-05, "loss": 3.354, "step": 57950 }, { "epoch": 0.512195552729649, "grad_norm": 2.4331154823303223, "learning_rate": 9.751342385227791e-05, "loss": 3.1493, "step": 58000 }, { "epoch": 0.5126371006199333, "grad_norm": 2.020382881164551, "learning_rate": 9.750909430392821e-05, "loss": 3.4823, "step": 58050 }, { "epoch": 0.5130786485102174, "grad_norm": 1.6145989894866943, "learning_rate": 9.750476108590383e-05, "loss": 3.2045, "step": 58100 }, { "epoch": 0.5135201964005016, "grad_norm": 4.1392364501953125, "learning_rate": 9.750042419853949e-05, "loss": 3.2311, "step": 58150 }, { "epoch": 0.5139617442907858, "grad_norm": 2.6319990158081055, "learning_rate": 9.749608364217018e-05, "loss": 3.4167, "step": 58200 }, { "epoch": 0.51440329218107, "grad_norm": 0.8864327073097229, "learning_rate": 9.749173941713114e-05, "loss": 3.4851, "step": 58250 }, { "epoch": 0.5148448400713541, "grad_norm": 1.8275792598724365, "learning_rate": 9.748739152375799e-05, "loss": 3.429, "step": 58300 }, { "epoch": 0.5152863879616383, "grad_norm": 1.4619648456573486, "learning_rate": 9.74830399623865e-05, "loss": 3.1952, "step": 58350 }, { "epoch": 0.5157279358519224, "grad_norm": 2.958252429962158, "learning_rate": 9.747868473335283e-05, "loss": 3.3281, "step": 58400 }, { "epoch": 0.5161694837422067, "grad_norm": 3.7064812183380127, "learning_rate": 9.747432583699334e-05, "loss": 3.1441, "step": 58450 }, { "epoch": 0.5166110316324909, "grad_norm": 2.135683298110962, "learning_rate": 9.746996327364478e-05, "loss": 3.3074, "step": 58500 }, { "epoch": 0.5170525795227751, "grad_norm": 4.96755838394165, "learning_rate": 9.746559704364409e-05, "loss": 3.1264, "step": 58550 }, { "epoch": 0.5174941274130592, "grad_norm": 1.5454546213150024, "learning_rate": 9.74612271473285e-05, "loss": 3.3129, "step": 58600 }, { "epoch": 0.5179356753033434, "grad_norm": 2.1017255783081055, "learning_rate": 9.74568535850356e-05, "loss": 3.2598, "step": 58650 }, { "epoch": 0.5183772231936276, "grad_norm": 1.2823448181152344, "learning_rate": 9.745247635710318e-05, "loss": 3.6969, "step": 58700 }, { "epoch": 0.5188187710839117, "grad_norm": 2.371886968612671, "learning_rate": 9.744809546386933e-05, "loss": 3.2902, "step": 58750 }, { "epoch": 0.5192603189741959, "grad_norm": 1.7832865715026855, "learning_rate": 9.744371090567246e-05, "loss": 2.8521, "step": 58800 }, { "epoch": 0.5197018668644802, "grad_norm": 1.12856125831604, "learning_rate": 9.743932268285124e-05, "loss": 3.7715, "step": 58850 }, { "epoch": 0.5201434147547643, "grad_norm": 1.2334407567977905, "learning_rate": 9.743493079574461e-05, "loss": 3.2311, "step": 58900 }, { "epoch": 0.5205849626450485, "grad_norm": 2.203479528427124, "learning_rate": 9.743053524469182e-05, "loss": 3.438, "step": 58950 }, { "epoch": 0.5210265105353327, "grad_norm": 0.8005960583686829, "learning_rate": 9.742613603003238e-05, "loss": 3.6582, "step": 59000 }, { "epoch": 0.5214680584256168, "grad_norm": 2.671816110610962, "learning_rate": 9.742173315210608e-05, "loss": 2.9465, "step": 59050 }, { "epoch": 0.521909606315901, "grad_norm": 0.9956416487693787, "learning_rate": 9.741732661125304e-05, "loss": 3.3792, "step": 59100 }, { "epoch": 0.5223511542061852, "grad_norm": 1.9408859014511108, "learning_rate": 9.741291640781359e-05, "loss": 3.2726, "step": 59150 }, { "epoch": 0.5227927020964693, "grad_norm": 1.1133266687393188, "learning_rate": 9.740850254212841e-05, "loss": 3.3387, "step": 59200 }, { "epoch": 0.5232342499867536, "grad_norm": 0.527526319026947, "learning_rate": 9.740408501453841e-05, "loss": 3.0193, "step": 59250 }, { "epoch": 0.5236757978770378, "grad_norm": 1.995932936668396, "learning_rate": 9.7399752285049e-05, "loss": 3.5348, "step": 59300 }, { "epoch": 0.5241173457673219, "grad_norm": 1.4613289833068848, "learning_rate": 9.739532750789442e-05, "loss": 3.0866, "step": 59350 }, { "epoch": 0.5245588936576061, "grad_norm": 2.9688708782196045, "learning_rate": 9.739089906985268e-05, "loss": 3.5847, "step": 59400 }, { "epoch": 0.5250004415478903, "grad_norm": 1.3395789861679077, "learning_rate": 9.738646697126586e-05, "loss": 3.353, "step": 59450 }, { "epoch": 0.5254419894381744, "grad_norm": 4.93052864074707, "learning_rate": 9.738203121247627e-05, "loss": 3.626, "step": 59500 }, { "epoch": 0.5258835373284586, "grad_norm": 3.3152246475219727, "learning_rate": 9.737759179382658e-05, "loss": 2.8786, "step": 59550 }, { "epoch": 0.5263250852187428, "grad_norm": 1.5898215770721436, "learning_rate": 9.737314871565965e-05, "loss": 3.7041, "step": 59600 }, { "epoch": 0.526766633109027, "grad_norm": 2.1526150703430176, "learning_rate": 9.736870197831871e-05, "loss": 3.5867, "step": 59650 }, { "epoch": 0.5272081809993112, "grad_norm": 2.351823568344116, "learning_rate": 9.736425158214723e-05, "loss": 3.4919, "step": 59700 }, { "epoch": 0.5276497288895954, "grad_norm": 2.4668421745300293, "learning_rate": 9.735979752748894e-05, "loss": 3.3577, "step": 59750 }, { "epoch": 0.5280912767798795, "grad_norm": 2.4981229305267334, "learning_rate": 9.73553398146879e-05, "loss": 3.368, "step": 59800 }, { "epoch": 0.5285328246701637, "grad_norm": 1.9264018535614014, "learning_rate": 9.735087844408839e-05, "loss": 3.7379, "step": 59850 }, { "epoch": 0.5289743725604479, "grad_norm": 4.922735691070557, "learning_rate": 9.734641341603507e-05, "loss": 3.1943, "step": 59900 }, { "epoch": 0.529415920450732, "grad_norm": 5.6230645179748535, "learning_rate": 9.73419447308728e-05, "loss": 3.3315, "step": 59950 }, { "epoch": 0.5298574683410162, "grad_norm": 1.985254168510437, "learning_rate": 9.733747238894673e-05, "loss": 3.8071, "step": 60000 }, { "epoch": 0.5298574683410162, "eval_asr_loss": 0.8992706317769239, "eval_loss": 3.0059359073638916, "eval_runtime": 20.6646, "eval_samples_per_second": 37.165, "eval_steps_per_second": 9.291, "eval_tts_loss": 5.9863666861098865, "step": 60000 }, { "epoch": 0.5302990162313005, "grad_norm": 1.029417872428894, "learning_rate": 9.733299639060233e-05, "loss": 3.503, "step": 60050 }, { "epoch": 0.5307405641215847, "grad_norm": 6.666355609893799, "learning_rate": 9.732851673618535e-05, "loss": 3.2843, "step": 60100 }, { "epoch": 0.5311821120118688, "grad_norm": 1.5539653301239014, "learning_rate": 9.732403342604177e-05, "loss": 3.374, "step": 60150 }, { "epoch": 0.531623659902153, "grad_norm": 1.662533164024353, "learning_rate": 9.731954646051792e-05, "loss": 3.1763, "step": 60200 }, { "epoch": 0.5320652077924372, "grad_norm": 1.5471640825271606, "learning_rate": 9.731505583996035e-05, "loss": 3.3524, "step": 60250 }, { "epoch": 0.5325067556827213, "grad_norm": 3.6004583835601807, "learning_rate": 9.731056156471594e-05, "loss": 3.4967, "step": 60300 }, { "epoch": 0.5329483035730055, "grad_norm": 5.77100944519043, "learning_rate": 9.730606363513184e-05, "loss": 3.3075, "step": 60350 }, { "epoch": 0.5333898514632897, "grad_norm": 2.9445252418518066, "learning_rate": 9.730156205155545e-05, "loss": 3.681, "step": 60400 }, { "epoch": 0.5338313993535739, "grad_norm": 3.9473483562469482, "learning_rate": 9.729705681433451e-05, "loss": 2.9724, "step": 60450 }, { "epoch": 0.5342729472438581, "grad_norm": 2.4145290851593018, "learning_rate": 9.729254792381698e-05, "loss": 3.1867, "step": 60500 }, { "epoch": 0.5347144951341423, "grad_norm": 2.9569525718688965, "learning_rate": 9.728803538035119e-05, "loss": 3.6975, "step": 60550 }, { "epoch": 0.5351560430244264, "grad_norm": 3.94183349609375, "learning_rate": 9.728351918428562e-05, "loss": 3.6706, "step": 60600 }, { "epoch": 0.5355975909147106, "grad_norm": 1.8953005075454712, "learning_rate": 9.727899933596917e-05, "loss": 3.3651, "step": 60650 }, { "epoch": 0.5360391388049948, "grad_norm": 1.777334213256836, "learning_rate": 9.727447583575092e-05, "loss": 3.7644, "step": 60700 }, { "epoch": 0.5364806866952789, "grad_norm": 3.474303960800171, "learning_rate": 9.72699486839803e-05, "loss": 3.2085, "step": 60750 }, { "epoch": 0.5369222345855631, "grad_norm": 2.88875412940979, "learning_rate": 9.726541788100696e-05, "loss": 3.8591, "step": 60800 }, { "epoch": 0.5373637824758474, "grad_norm": 3.021052598953247, "learning_rate": 9.726088342718091e-05, "loss": 3.3407, "step": 60850 }, { "epoch": 0.5378053303661315, "grad_norm": 1.9368441104888916, "learning_rate": 9.725634532285238e-05, "loss": 3.4342, "step": 60900 }, { "epoch": 0.5382468782564157, "grad_norm": 1.3616617918014526, "learning_rate": 9.725180356837187e-05, "loss": 3.1015, "step": 60950 }, { "epoch": 0.5386884261466999, "grad_norm": 1.2605106830596924, "learning_rate": 9.724725816409024e-05, "loss": 3.4434, "step": 61000 }, { "epoch": 0.539129974036984, "grad_norm": 2.5226216316223145, "learning_rate": 9.724270911035856e-05, "loss": 3.2771, "step": 61050 }, { "epoch": 0.5395715219272682, "grad_norm": 1.5489667654037476, "learning_rate": 9.723815640752818e-05, "loss": 2.9834, "step": 61100 }, { "epoch": 0.5400130698175524, "grad_norm": 0.3823397159576416, "learning_rate": 9.723360005595082e-05, "loss": 3.4476, "step": 61150 }, { "epoch": 0.5404546177078365, "grad_norm": 1.4170905351638794, "learning_rate": 9.722904005597838e-05, "loss": 2.9736, "step": 61200 }, { "epoch": 0.5408961655981208, "grad_norm": 3.013650417327881, "learning_rate": 9.722447640796306e-05, "loss": 3.4688, "step": 61250 }, { "epoch": 0.541337713488405, "grad_norm": 2.0270910263061523, "learning_rate": 9.721990911225742e-05, "loss": 3.4488, "step": 61300 }, { "epoch": 0.5417792613786891, "grad_norm": 2.3157079219818115, "learning_rate": 9.72153381692142e-05, "loss": 3.4686, "step": 61350 }, { "epoch": 0.5422208092689733, "grad_norm": 2.65665864944458, "learning_rate": 9.721076357918648e-05, "loss": 3.0758, "step": 61400 }, { "epoch": 0.5426623571592575, "grad_norm": 4.086440086364746, "learning_rate": 9.720618534252761e-05, "loss": 3.0561, "step": 61450 }, { "epoch": 0.5431039050495416, "grad_norm": 1.6047959327697754, "learning_rate": 9.720160345959122e-05, "loss": 3.2575, "step": 61500 }, { "epoch": 0.5435454529398258, "grad_norm": 1.371722936630249, "learning_rate": 9.719701793073121e-05, "loss": 3.197, "step": 61550 }, { "epoch": 0.54398700083011, "grad_norm": 1.1012533903121948, "learning_rate": 9.71924287563018e-05, "loss": 3.1246, "step": 61600 }, { "epoch": 0.5444285487203943, "grad_norm": 2.610647201538086, "learning_rate": 9.718783593665745e-05, "loss": 3.4834, "step": 61650 }, { "epoch": 0.5448700966106784, "grad_norm": 3.1460602283477783, "learning_rate": 9.71832394721529e-05, "loss": 3.5381, "step": 61700 }, { "epoch": 0.5453116445009626, "grad_norm": 2.426410675048828, "learning_rate": 9.717863936314322e-05, "loss": 3.0234, "step": 61750 }, { "epoch": 0.5457531923912468, "grad_norm": 1.7792102098464966, "learning_rate": 9.71740356099837e-05, "loss": 3.3962, "step": 61800 }, { "epoch": 0.5461947402815309, "grad_norm": 0.535715639591217, "learning_rate": 9.716942821302995e-05, "loss": 3.2452, "step": 61850 }, { "epoch": 0.5466362881718151, "grad_norm": 3.1257872581481934, "learning_rate": 9.716481717263787e-05, "loss": 3.2172, "step": 61900 }, { "epoch": 0.5470778360620993, "grad_norm": 2.848803997039795, "learning_rate": 9.71602024891636e-05, "loss": 3.5478, "step": 61950 }, { "epoch": 0.5475193839523834, "grad_norm": 2.5586445331573486, "learning_rate": 9.71555841629636e-05, "loss": 3.5599, "step": 62000 }, { "epoch": 0.5479609318426677, "grad_norm": 5.202526569366455, "learning_rate": 9.715096219439458e-05, "loss": 3.3383, "step": 62050 }, { "epoch": 0.5484024797329519, "grad_norm": 7.311855316162109, "learning_rate": 9.714633658381358e-05, "loss": 3.5755, "step": 62100 }, { "epoch": 0.548844027623236, "grad_norm": 2.228634834289551, "learning_rate": 9.714170733157784e-05, "loss": 3.1598, "step": 62150 }, { "epoch": 0.5492855755135202, "grad_norm": 0.7159774899482727, "learning_rate": 9.713707443804499e-05, "loss": 3.3646, "step": 62200 }, { "epoch": 0.5497271234038044, "grad_norm": 1.7473576068878174, "learning_rate": 9.713243790357282e-05, "loss": 3.153, "step": 62250 }, { "epoch": 0.5501686712940885, "grad_norm": 2.541585922241211, "learning_rate": 9.712779772851952e-05, "loss": 3.4916, "step": 62300 }, { "epoch": 0.5506102191843727, "grad_norm": 5.356276512145996, "learning_rate": 9.712315391324346e-05, "loss": 3.2563, "step": 62350 }, { "epoch": 0.5510517670746569, "grad_norm": 2.213956356048584, "learning_rate": 9.711850645810336e-05, "loss": 3.1959, "step": 62400 }, { "epoch": 0.5514933149649411, "grad_norm": 1.0666532516479492, "learning_rate": 9.711385536345818e-05, "loss": 3.4801, "step": 62450 }, { "epoch": 0.5519348628552253, "grad_norm": 5.453732967376709, "learning_rate": 9.71092006296672e-05, "loss": 3.1979, "step": 62500 }, { "epoch": 0.5523764107455095, "grad_norm": 6.855234622955322, "learning_rate": 9.710454225708994e-05, "loss": 3.6221, "step": 62550 }, { "epoch": 0.5528179586357936, "grad_norm": 1.6029174327850342, "learning_rate": 9.709988024608623e-05, "loss": 3.4752, "step": 62600 }, { "epoch": 0.5532595065260778, "grad_norm": 1.8403346538543701, "learning_rate": 9.709521459701616e-05, "loss": 3.4726, "step": 62650 }, { "epoch": 0.553701054416362, "grad_norm": 4.461715221405029, "learning_rate": 9.709054531024011e-05, "loss": 2.9593, "step": 62700 }, { "epoch": 0.5541426023066461, "grad_norm": 0.8452909588813782, "learning_rate": 9.708587238611877e-05, "loss": 3.2217, "step": 62750 }, { "epoch": 0.5545841501969304, "grad_norm": 1.3129770755767822, "learning_rate": 9.708119582501305e-05, "loss": 3.3426, "step": 62800 }, { "epoch": 0.5550256980872146, "grad_norm": 2.4353716373443604, "learning_rate": 9.707651562728419e-05, "loss": 3.4732, "step": 62850 }, { "epoch": 0.5554672459774987, "grad_norm": 1.2511876821517944, "learning_rate": 9.707183179329371e-05, "loss": 3.3568, "step": 62900 }, { "epoch": 0.5559087938677829, "grad_norm": 4.347702980041504, "learning_rate": 9.706714432340336e-05, "loss": 3.1255, "step": 62950 }, { "epoch": 0.5563503417580671, "grad_norm": 1.2785756587982178, "learning_rate": 9.706245321797525e-05, "loss": 3.5741, "step": 63000 }, { "epoch": 0.5563503417580671, "eval_asr_loss": 0.9045116492456787, "eval_loss": 3.014928102493286, "eval_runtime": 20.6473, "eval_samples_per_second": 37.196, "eval_steps_per_second": 9.299, "eval_tts_loss": 5.952768747740511, "step": 63000 }, { "epoch": 0.5567918896483512, "grad_norm": 1.4415526390075684, "learning_rate": 9.705775847737169e-05, "loss": 3.5287, "step": 63050 }, { "epoch": 0.5572334375386354, "grad_norm": 1.9535558223724365, "learning_rate": 9.705306010195533e-05, "loss": 3.3174, "step": 63100 }, { "epoch": 0.5576749854289196, "grad_norm": 1.3182282447814941, "learning_rate": 9.704835809208907e-05, "loss": 3.4299, "step": 63150 }, { "epoch": 0.5581165333192039, "grad_norm": 1.156562328338623, "learning_rate": 9.704365244813613e-05, "loss": 3.2571, "step": 63200 }, { "epoch": 0.558558081209488, "grad_norm": 1.929991364479065, "learning_rate": 9.703894317045993e-05, "loss": 3.0214, "step": 63250 }, { "epoch": 0.5589996290997722, "grad_norm": 1.8567146062850952, "learning_rate": 9.703423025942426e-05, "loss": 3.2539, "step": 63300 }, { "epoch": 0.5594411769900564, "grad_norm": 0.45442625880241394, "learning_rate": 9.702951371539315e-05, "loss": 3.5004, "step": 63350 }, { "epoch": 0.5598827248803405, "grad_norm": 2.3628077507019043, "learning_rate": 9.702479353873089e-05, "loss": 3.1837, "step": 63400 }, { "epoch": 0.5603242727706247, "grad_norm": 3.7257354259490967, "learning_rate": 9.702006972980208e-05, "loss": 3.503, "step": 63450 }, { "epoch": 0.5607658206609089, "grad_norm": 3.1037049293518066, "learning_rate": 9.701534228897163e-05, "loss": 3.2654, "step": 63500 }, { "epoch": 0.561207368551193, "grad_norm": 2.567664861679077, "learning_rate": 9.701061121660464e-05, "loss": 3.3444, "step": 63550 }, { "epoch": 0.5616489164414773, "grad_norm": 3.259284019470215, "learning_rate": 9.700587651306658e-05, "loss": 3.0933, "step": 63600 }, { "epoch": 0.5620904643317615, "grad_norm": 3.6262829303741455, "learning_rate": 9.700113817872317e-05, "loss": 3.4273, "step": 63650 }, { "epoch": 0.5625320122220456, "grad_norm": 2.7692084312438965, "learning_rate": 9.699639621394039e-05, "loss": 3.5773, "step": 63700 }, { "epoch": 0.5629735601123298, "grad_norm": 2.4339802265167236, "learning_rate": 9.699165061908451e-05, "loss": 2.8369, "step": 63750 }, { "epoch": 0.563415108002614, "grad_norm": 1.9642480611801147, "learning_rate": 9.69869013945221e-05, "loss": 3.533, "step": 63800 }, { "epoch": 0.5638566558928981, "grad_norm": 4.266359329223633, "learning_rate": 9.698214854062e-05, "loss": 3.3043, "step": 63850 }, { "epoch": 0.5642982037831823, "grad_norm": 1.5324000120162964, "learning_rate": 9.697739205774532e-05, "loss": 3.6485, "step": 63900 }, { "epoch": 0.5647397516734665, "grad_norm": 3.5735909938812256, "learning_rate": 9.6972727184053e-05, "loss": 3.3017, "step": 63950 }, { "epoch": 0.5651812995637507, "grad_norm": 1.4858098030090332, "learning_rate": 9.696796351689678e-05, "loss": 3.4435, "step": 64000 }, { "epoch": 0.5656228474540349, "grad_norm": 3.1134603023529053, "learning_rate": 9.696319622186367e-05, "loss": 2.8183, "step": 64050 }, { "epoch": 0.5660643953443191, "grad_norm": 0.87113356590271, "learning_rate": 9.695842529932186e-05, "loss": 3.5863, "step": 64100 }, { "epoch": 0.5665059432346032, "grad_norm": 2.494359016418457, "learning_rate": 9.695365074963992e-05, "loss": 3.2139, "step": 64150 }, { "epoch": 0.5669474911248874, "grad_norm": 4.142000198364258, "learning_rate": 9.694887257318659e-05, "loss": 3.4118, "step": 64200 }, { "epoch": 0.5673890390151716, "grad_norm": 7.738039493560791, "learning_rate": 9.694409077033097e-05, "loss": 3.3375, "step": 64250 }, { "epoch": 0.5678305869054557, "grad_norm": 2.2657651901245117, "learning_rate": 9.693930534144243e-05, "loss": 3.5303, "step": 64300 }, { "epoch": 0.5682721347957399, "grad_norm": 1.5491681098937988, "learning_rate": 9.693451628689059e-05, "loss": 3.1383, "step": 64350 }, { "epoch": 0.5687136826860242, "grad_norm": 5.7371745109558105, "learning_rate": 9.692972360704534e-05, "loss": 3.3374, "step": 64400 }, { "epoch": 0.5691552305763083, "grad_norm": 1.6547493934631348, "learning_rate": 9.69249273022769e-05, "loss": 3.5325, "step": 64450 }, { "epoch": 0.5695967784665925, "grad_norm": 4.246345043182373, "learning_rate": 9.692012737295574e-05, "loss": 3.487, "step": 64500 }, { "epoch": 0.5700383263568767, "grad_norm": 4.030117988586426, "learning_rate": 9.69153238194526e-05, "loss": 3.0569, "step": 64550 }, { "epoch": 0.5704798742471608, "grad_norm": 1.0426106452941895, "learning_rate": 9.691051664213855e-05, "loss": 3.3855, "step": 64600 }, { "epoch": 0.570921422137445, "grad_norm": 1.773262619972229, "learning_rate": 9.690570584138486e-05, "loss": 3.0886, "step": 64650 }, { "epoch": 0.5713629700277292, "grad_norm": 2.984727144241333, "learning_rate": 9.690089141756316e-05, "loss": 3.3807, "step": 64700 }, { "epoch": 0.5718045179180133, "grad_norm": 2.3156557083129883, "learning_rate": 9.689607337104528e-05, "loss": 3.3806, "step": 64750 }, { "epoch": 0.5722460658082976, "grad_norm": 3.1280009746551514, "learning_rate": 9.689125170220341e-05, "loss": 3.223, "step": 64800 }, { "epoch": 0.5726876136985818, "grad_norm": 1.362710952758789, "learning_rate": 9.688642641140999e-05, "loss": 2.7777, "step": 64850 }, { "epoch": 0.573129161588866, "grad_norm": 2.129289150238037, "learning_rate": 9.68815974990377e-05, "loss": 3.2796, "step": 64900 }, { "epoch": 0.5735707094791501, "grad_norm": 3.2772629261016846, "learning_rate": 9.687676496545955e-05, "loss": 3.4964, "step": 64950 }, { "epoch": 0.5740122573694343, "grad_norm": 2.643249034881592, "learning_rate": 9.68719288110488e-05, "loss": 3.6755, "step": 65000 }, { "epoch": 0.5744538052597185, "grad_norm": 1.4029085636138916, "learning_rate": 9.686708903617902e-05, "loss": 3.1452, "step": 65050 }, { "epoch": 0.5748953531500026, "grad_norm": 0.7895709276199341, "learning_rate": 9.686224564122403e-05, "loss": 3.233, "step": 65100 }, { "epoch": 0.5753369010402868, "grad_norm": 1.5120134353637695, "learning_rate": 9.685739862655793e-05, "loss": 3.2801, "step": 65150 }, { "epoch": 0.5757784489305711, "grad_norm": 1.919240117073059, "learning_rate": 9.685254799255517e-05, "loss": 3.3301, "step": 65200 }, { "epoch": 0.5762199968208552, "grad_norm": 2.2543394565582275, "learning_rate": 9.684769373959033e-05, "loss": 2.9593, "step": 65250 }, { "epoch": 0.5766615447111394, "grad_norm": 1.0832370519638062, "learning_rate": 9.684283586803843e-05, "loss": 3.6906, "step": 65300 }, { "epoch": 0.5771030926014236, "grad_norm": 3.100877285003662, "learning_rate": 9.683797437827466e-05, "loss": 3.347, "step": 65350 }, { "epoch": 0.5775446404917077, "grad_norm": 3.069957733154297, "learning_rate": 9.683310927067455e-05, "loss": 3.0518, "step": 65400 }, { "epoch": 0.5779861883819919, "grad_norm": 1.4712010622024536, "learning_rate": 9.682824054561389e-05, "loss": 3.2539, "step": 65450 }, { "epoch": 0.5784277362722761, "grad_norm": 1.5608115196228027, "learning_rate": 9.682336820346874e-05, "loss": 3.4519, "step": 65500 }, { "epoch": 0.5788692841625602, "grad_norm": 2.5109148025512695, "learning_rate": 9.681849224461544e-05, "loss": 3.3817, "step": 65550 }, { "epoch": 0.5793108320528445, "grad_norm": 3.372307300567627, "learning_rate": 9.68136126694306e-05, "loss": 3.4769, "step": 65600 }, { "epoch": 0.5797523799431287, "grad_norm": 2.304180860519409, "learning_rate": 9.680872947829118e-05, "loss": 3.8492, "step": 65650 }, { "epoch": 0.5801939278334128, "grad_norm": 0.9315201044082642, "learning_rate": 9.680384267157434e-05, "loss": 3.7529, "step": 65700 }, { "epoch": 0.580635475723697, "grad_norm": 1.882952332496643, "learning_rate": 9.679895224965752e-05, "loss": 3.0259, "step": 65750 }, { "epoch": 0.5810770236139812, "grad_norm": 3.846442222595215, "learning_rate": 9.679405821291849e-05, "loss": 3.0785, "step": 65800 }, { "epoch": 0.5815185715042653, "grad_norm": 1.7225450277328491, "learning_rate": 9.678916056173526e-05, "loss": 3.5125, "step": 65850 }, { "epoch": 0.5819601193945495, "grad_norm": 0.8811081051826477, "learning_rate": 9.678425929648614e-05, "loss": 3.0916, "step": 65900 }, { "epoch": 0.5824016672848337, "grad_norm": 3.4488401412963867, "learning_rate": 9.67793544175497e-05, "loss": 3.0175, "step": 65950 }, { "epoch": 0.582843215175118, "grad_norm": 2.16182804107666, "learning_rate": 9.677444592530483e-05, "loss": 3.6373, "step": 66000 }, { "epoch": 0.582843215175118, "eval_asr_loss": 0.8874591935323445, "eval_loss": 3.00752854347229, "eval_runtime": 21.0653, "eval_samples_per_second": 36.458, "eval_steps_per_second": 9.115, "eval_tts_loss": 5.963853557317323, "step": 66000 }, { "epoch": 0.5832847630654021, "grad_norm": 2.242957353591919, "learning_rate": 9.676953382013063e-05, "loss": 3.0162, "step": 66050 }, { "epoch": 0.5837263109556863, "grad_norm": 2.5106394290924072, "learning_rate": 9.676461810240654e-05, "loss": 3.4729, "step": 66100 }, { "epoch": 0.5841678588459704, "grad_norm": 3.770650863647461, "learning_rate": 9.675969877251225e-05, "loss": 3.4405, "step": 66150 }, { "epoch": 0.5846094067362546, "grad_norm": 2.2024853229522705, "learning_rate": 9.675477583082775e-05, "loss": 3.5091, "step": 66200 }, { "epoch": 0.5850509546265388, "grad_norm": 5.273355007171631, "learning_rate": 9.674984927773328e-05, "loss": 3.5578, "step": 66250 }, { "epoch": 0.5854925025168229, "grad_norm": 2.3930509090423584, "learning_rate": 9.674491911360939e-05, "loss": 3.269, "step": 66300 }, { "epoch": 0.5859340504071071, "grad_norm": 0.5656553506851196, "learning_rate": 9.673998533883687e-05, "loss": 3.4693, "step": 66350 }, { "epoch": 0.5863755982973914, "grad_norm": 1.1394141912460327, "learning_rate": 9.673504795379683e-05, "loss": 3.4782, "step": 66400 }, { "epoch": 0.5868171461876756, "grad_norm": 2.660053253173828, "learning_rate": 9.673010695887064e-05, "loss": 3.5723, "step": 66450 }, { "epoch": 0.5872586940779597, "grad_norm": 2.931849718093872, "learning_rate": 9.672516235443994e-05, "loss": 3.0214, "step": 66500 }, { "epoch": 0.5877002419682439, "grad_norm": 3.5768563747406006, "learning_rate": 9.672021414088667e-05, "loss": 3.0214, "step": 66550 }, { "epoch": 0.588141789858528, "grad_norm": 6.038229465484619, "learning_rate": 9.671526231859305e-05, "loss": 3.4275, "step": 66600 }, { "epoch": 0.5885833377488122, "grad_norm": 3.714881181716919, "learning_rate": 9.671030688794153e-05, "loss": 2.9659, "step": 66650 }, { "epoch": 0.5890248856390964, "grad_norm": 1.5607867240905762, "learning_rate": 9.670544706544311e-05, "loss": 3.4904, "step": 66700 }, { "epoch": 0.5894664335293806, "grad_norm": 1.8332650661468506, "learning_rate": 9.67004844913725e-05, "loss": 3.0832, "step": 66750 }, { "epoch": 0.5899079814196648, "grad_norm": 2.306621789932251, "learning_rate": 9.669551831008545e-05, "loss": 3.1016, "step": 66800 }, { "epoch": 0.590349529309949, "grad_norm": 1.265374779701233, "learning_rate": 9.66905485219656e-05, "loss": 3.3778, "step": 66850 }, { "epoch": 0.5907910772002332, "grad_norm": 2.479994773864746, "learning_rate": 9.66855751273968e-05, "loss": 3.3998, "step": 66900 }, { "epoch": 0.5912326250905173, "grad_norm": 2.6619317531585693, "learning_rate": 9.66805981267632e-05, "loss": 3.0006, "step": 66950 }, { "epoch": 0.5916741729808015, "grad_norm": 2.5897438526153564, "learning_rate": 9.667561752044922e-05, "loss": 3.6474, "step": 67000 }, { "epoch": 0.5921157208710857, "grad_norm": 5.512159824371338, "learning_rate": 9.667063330883961e-05, "loss": 3.457, "step": 67050 }, { "epoch": 0.5925572687613698, "grad_norm": 2.0321197509765625, "learning_rate": 9.666564549231931e-05, "loss": 3.653, "step": 67100 }, { "epoch": 0.592998816651654, "grad_norm": 2.648627519607544, "learning_rate": 9.666065407127361e-05, "loss": 2.8554, "step": 67150 }, { "epoch": 0.5934403645419383, "grad_norm": 2.1089653968811035, "learning_rate": 9.665565904608806e-05, "loss": 3.0816, "step": 67200 }, { "epoch": 0.5938819124322224, "grad_norm": 0.6567474007606506, "learning_rate": 9.665066041714849e-05, "loss": 3.4614, "step": 67250 }, { "epoch": 0.5943234603225066, "grad_norm": 4.268828392028809, "learning_rate": 9.664565818484097e-05, "loss": 3.1899, "step": 67300 }, { "epoch": 0.5947650082127908, "grad_norm": 4.196305751800537, "learning_rate": 9.664065234955191e-05, "loss": 3.3985, "step": 67350 }, { "epoch": 0.5952065561030749, "grad_norm": 4.905257225036621, "learning_rate": 9.663564291166795e-05, "loss": 3.3829, "step": 67400 }, { "epoch": 0.5956481039933591, "grad_norm": 3.6008496284484863, "learning_rate": 9.663062987157604e-05, "loss": 2.946, "step": 67450 }, { "epoch": 0.5960896518836433, "grad_norm": 1.2707490921020508, "learning_rate": 9.66256132296634e-05, "loss": 3.4213, "step": 67500 }, { "epoch": 0.5965311997739274, "grad_norm": 3.5220985412597656, "learning_rate": 9.662059298631749e-05, "loss": 3.0964, "step": 67550 }, { "epoch": 0.5969727476642117, "grad_norm": 1.3247735500335693, "learning_rate": 9.66155691419261e-05, "loss": 3.4796, "step": 67600 }, { "epoch": 0.5974142955544959, "grad_norm": 1.086634874343872, "learning_rate": 9.66105416968773e-05, "loss": 3.4287, "step": 67650 }, { "epoch": 0.59785584344478, "grad_norm": 3.1261913776397705, "learning_rate": 9.660551065155938e-05, "loss": 3.239, "step": 67700 }, { "epoch": 0.5982973913350642, "grad_norm": 5.029034614562988, "learning_rate": 9.660047600636099e-05, "loss": 2.8555, "step": 67750 }, { "epoch": 0.5987389392253484, "grad_norm": 2.428929328918457, "learning_rate": 9.659543776167097e-05, "loss": 3.331, "step": 67800 }, { "epoch": 0.5991804871156325, "grad_norm": 2.1692111492156982, "learning_rate": 9.65903959178785e-05, "loss": 3.2252, "step": 67850 }, { "epoch": 0.5996220350059167, "grad_norm": 5.579278945922852, "learning_rate": 9.658535047537303e-05, "loss": 3.2612, "step": 67900 }, { "epoch": 0.6000635828962009, "grad_norm": 1.7974921464920044, "learning_rate": 9.658030143454426e-05, "loss": 3.3738, "step": 67950 }, { "epoch": 0.6005051307864852, "grad_norm": 2.329169750213623, "learning_rate": 9.657524879578221e-05, "loss": 3.358, "step": 68000 }, { "epoch": 0.6009466786767693, "grad_norm": 2.0072598457336426, "learning_rate": 9.657019255947712e-05, "loss": 3.1542, "step": 68050 }, { "epoch": 0.6013882265670535, "grad_norm": 1.212756872177124, "learning_rate": 9.656513272601957e-05, "loss": 2.8526, "step": 68100 }, { "epoch": 0.6018297744573377, "grad_norm": 2.1172633171081543, "learning_rate": 9.656006929580036e-05, "loss": 3.3746, "step": 68150 }, { "epoch": 0.6022713223476218, "grad_norm": 3.0865988731384277, "learning_rate": 9.655500226921064e-05, "loss": 3.4431, "step": 68200 }, { "epoch": 0.602712870237906, "grad_norm": 2.7705607414245605, "learning_rate": 9.654993164664175e-05, "loss": 3.3185, "step": 68250 }, { "epoch": 0.6031544181281902, "grad_norm": 2.9133379459381104, "learning_rate": 9.654485742848538e-05, "loss": 3.2805, "step": 68300 }, { "epoch": 0.6035959660184743, "grad_norm": 1.945061445236206, "learning_rate": 9.653977961513347e-05, "loss": 3.5587, "step": 68350 }, { "epoch": 0.6040375139087586, "grad_norm": 0.5719209313392639, "learning_rate": 9.653469820697822e-05, "loss": 3.039, "step": 68400 }, { "epoch": 0.6044790617990428, "grad_norm": 2.7784836292266846, "learning_rate": 9.652961320441214e-05, "loss": 3.6163, "step": 68450 }, { "epoch": 0.6049206096893269, "grad_norm": 1.212082028388977, "learning_rate": 9.652452460782799e-05, "loss": 2.763, "step": 68500 }, { "epoch": 0.6053621575796111, "grad_norm": 6.158472061157227, "learning_rate": 9.651943241761884e-05, "loss": 3.4118, "step": 68550 }, { "epoch": 0.6058037054698953, "grad_norm": 1.3418790102005005, "learning_rate": 9.6514336634178e-05, "loss": 3.3112, "step": 68600 }, { "epoch": 0.6062452533601794, "grad_norm": 2.3166797161102295, "learning_rate": 9.65092372578991e-05, "loss": 3.3965, "step": 68650 }, { "epoch": 0.6066868012504636, "grad_norm": 2.7341291904449463, "learning_rate": 9.6504134289176e-05, "loss": 3.4362, "step": 68700 }, { "epoch": 0.6071283491407478, "grad_norm": 5.785979270935059, "learning_rate": 9.649902772840287e-05, "loss": 3.5519, "step": 68750 }, { "epoch": 0.607569897031032, "grad_norm": 1.710016131401062, "learning_rate": 9.649391757597414e-05, "loss": 3.8108, "step": 68800 }, { "epoch": 0.6080114449213162, "grad_norm": 1.7782738208770752, "learning_rate": 9.648880383228455e-05, "loss": 3.0538, "step": 68850 }, { "epoch": 0.6084529928116004, "grad_norm": 2.464695930480957, "learning_rate": 9.648368649772907e-05, "loss": 3.4832, "step": 68900 }, { "epoch": 0.6088945407018845, "grad_norm": 2.7873873710632324, "learning_rate": 9.647856557270299e-05, "loss": 3.5608, "step": 68950 }, { "epoch": 0.6093360885921687, "grad_norm": 2.8136725425720215, "learning_rate": 9.647344105760183e-05, "loss": 2.8681, "step": 69000 }, { "epoch": 0.6093360885921687, "eval_asr_loss": 0.9085685970427385, "eval_loss": 3.00970458984375, "eval_runtime": 20.9691, "eval_samples_per_second": 36.625, "eval_steps_per_second": 9.156, "eval_tts_loss": 5.906105015571997, "step": 69000 }, { "epoch": 0.6097776364824529, "grad_norm": 1.4982023239135742, "learning_rate": 9.646831295282143e-05, "loss": 3.7657, "step": 69050 }, { "epoch": 0.610219184372737, "grad_norm": 1.8872673511505127, "learning_rate": 9.646318125875791e-05, "loss": 3.1832, "step": 69100 }, { "epoch": 0.6106607322630212, "grad_norm": 3.1537160873413086, "learning_rate": 9.645804597580761e-05, "loss": 3.028, "step": 69150 }, { "epoch": 0.6111022801533055, "grad_norm": 2.5125792026519775, "learning_rate": 9.645290710436722e-05, "loss": 3.2232, "step": 69200 }, { "epoch": 0.6115438280435896, "grad_norm": 2.1751527786254883, "learning_rate": 9.644776464483368e-05, "loss": 3.1004, "step": 69250 }, { "epoch": 0.6119853759338738, "grad_norm": 1.4385361671447754, "learning_rate": 9.644261859760417e-05, "loss": 2.8845, "step": 69300 }, { "epoch": 0.612426923824158, "grad_norm": 1.0426270961761475, "learning_rate": 9.643746896307622e-05, "loss": 3.3967, "step": 69350 }, { "epoch": 0.6128684717144421, "grad_norm": 5.886585712432861, "learning_rate": 9.643231574164755e-05, "loss": 3.2875, "step": 69400 }, { "epoch": 0.6133100196047263, "grad_norm": 3.5066511631011963, "learning_rate": 9.642715893371622e-05, "loss": 2.8433, "step": 69450 }, { "epoch": 0.6137515674950105, "grad_norm": 2.092880964279175, "learning_rate": 9.642199853968055e-05, "loss": 3.2078, "step": 69500 }, { "epoch": 0.6141931153852946, "grad_norm": 1.670851707458496, "learning_rate": 9.641683455993916e-05, "loss": 3.0111, "step": 69550 }, { "epoch": 0.6146346632755789, "grad_norm": 3.2171874046325684, "learning_rate": 9.641177038132528e-05, "loss": 2.9596, "step": 69600 }, { "epoch": 0.6150762111658631, "grad_norm": 1.39114248752594, "learning_rate": 9.640659930306355e-05, "loss": 3.2125, "step": 69650 }, { "epoch": 0.6155177590561473, "grad_norm": 0.8414401412010193, "learning_rate": 9.640142464028551e-05, "loss": 3.3546, "step": 69700 }, { "epoch": 0.6159593069464314, "grad_norm": 2.521662473678589, "learning_rate": 9.639624639339092e-05, "loss": 3.2137, "step": 69750 }, { "epoch": 0.6164008548367156, "grad_norm": 2.6045455932617188, "learning_rate": 9.63910645627797e-05, "loss": 3.4016, "step": 69800 }, { "epoch": 0.6168424027269998, "grad_norm": 2.437931537628174, "learning_rate": 9.638587914885215e-05, "loss": 3.0942, "step": 69850 }, { "epoch": 0.6172839506172839, "grad_norm": 2.744288682937622, "learning_rate": 9.638069015200877e-05, "loss": 3.4627, "step": 69900 }, { "epoch": 0.6177254985075681, "grad_norm": 8.556507110595703, "learning_rate": 9.637549757265037e-05, "loss": 2.9197, "step": 69950 }, { "epoch": 0.6181670463978524, "grad_norm": 2.708615303039551, "learning_rate": 9.637030141117803e-05, "loss": 3.4477, "step": 70000 }, { "epoch": 0.6186085942881365, "grad_norm": 2.7415497303009033, "learning_rate": 9.636510166799313e-05, "loss": 3.3546, "step": 70050 }, { "epoch": 0.6190501421784207, "grad_norm": 2.8086228370666504, "learning_rate": 9.635989834349728e-05, "loss": 3.3978, "step": 70100 }, { "epoch": 0.6194916900687049, "grad_norm": 2.983367681503296, "learning_rate": 9.635469143809239e-05, "loss": 3.243, "step": 70150 }, { "epoch": 0.619933237958989, "grad_norm": 0.500150740146637, "learning_rate": 9.634948095218068e-05, "loss": 3.3688, "step": 70200 }, { "epoch": 0.6203747858492732, "grad_norm": 2.545722484588623, "learning_rate": 9.63442668861646e-05, "loss": 3.6102, "step": 70250 }, { "epoch": 0.6208163337395574, "grad_norm": 0.7100191712379456, "learning_rate": 9.633904924044687e-05, "loss": 3.7465, "step": 70300 }, { "epoch": 0.6212578816298415, "grad_norm": 1.9451789855957031, "learning_rate": 9.633382801543055e-05, "loss": 3.3812, "step": 70350 }, { "epoch": 0.6216994295201258, "grad_norm": 2.1257078647613525, "learning_rate": 9.632860321151892e-05, "loss": 3.0553, "step": 70400 }, { "epoch": 0.62214097741041, "grad_norm": 0.8686696290969849, "learning_rate": 9.632337482911553e-05, "loss": 2.8603, "step": 70450 }, { "epoch": 0.6225825253006941, "grad_norm": 6.291978359222412, "learning_rate": 9.631814286862426e-05, "loss": 3.1015, "step": 70500 }, { "epoch": 0.6230240731909783, "grad_norm": 3.416860818862915, "learning_rate": 9.631290733044921e-05, "loss": 3.5239, "step": 70550 }, { "epoch": 0.6234656210812625, "grad_norm": 1.6016830205917358, "learning_rate": 9.63076682149948e-05, "loss": 3.2555, "step": 70600 }, { "epoch": 0.6239071689715466, "grad_norm": 2.5389463901519775, "learning_rate": 9.630242552266569e-05, "loss": 3.5458, "step": 70650 }, { "epoch": 0.6243487168618308, "grad_norm": 2.5725655555725098, "learning_rate": 9.629717925386683e-05, "loss": 3.0217, "step": 70700 }, { "epoch": 0.624790264752115, "grad_norm": 2.941615581512451, "learning_rate": 9.629192940900348e-05, "loss": 3.2919, "step": 70750 }, { "epoch": 0.6252318126423992, "grad_norm": 2.9366567134857178, "learning_rate": 9.628667598848113e-05, "loss": 3.0738, "step": 70800 }, { "epoch": 0.6256733605326834, "grad_norm": 4.780130863189697, "learning_rate": 9.628141899270554e-05, "loss": 3.4164, "step": 70850 }, { "epoch": 0.6261149084229676, "grad_norm": 2.4676146507263184, "learning_rate": 9.62761584220828e-05, "loss": 2.6195, "step": 70900 }, { "epoch": 0.6265564563132517, "grad_norm": 2.815596103668213, "learning_rate": 9.627089427701923e-05, "loss": 3.1961, "step": 70950 }, { "epoch": 0.6269980042035359, "grad_norm": 1.198612093925476, "learning_rate": 9.626562655792145e-05, "loss": 3.446, "step": 71000 }, { "epoch": 0.6274395520938201, "grad_norm": 0.4727330803871155, "learning_rate": 9.626035526519632e-05, "loss": 3.252, "step": 71050 }, { "epoch": 0.6278810999841042, "grad_norm": 2.9533727169036865, "learning_rate": 9.625508039925104e-05, "loss": 3.3784, "step": 71100 }, { "epoch": 0.6283226478743884, "grad_norm": 1.3359452486038208, "learning_rate": 9.624980196049303e-05, "loss": 3.5926, "step": 71150 }, { "epoch": 0.6287641957646727, "grad_norm": 1.3383703231811523, "learning_rate": 9.624451994932999e-05, "loss": 3.4367, "step": 71200 }, { "epoch": 0.6292057436549569, "grad_norm": 3.669863224029541, "learning_rate": 9.623923436616996e-05, "loss": 3.4483, "step": 71250 }, { "epoch": 0.629647291545241, "grad_norm": 3.541435718536377, "learning_rate": 9.623394521142113e-05, "loss": 3.5803, "step": 71300 }, { "epoch": 0.6300888394355252, "grad_norm": 1.2664039134979248, "learning_rate": 9.622865248549211e-05, "loss": 3.02, "step": 71350 }, { "epoch": 0.6305303873258093, "grad_norm": 3.8231582641601562, "learning_rate": 9.622335618879168e-05, "loss": 3.2505, "step": 71400 }, { "epoch": 0.6309719352160935, "grad_norm": 1.3903136253356934, "learning_rate": 9.621805632172896e-05, "loss": 2.9738, "step": 71450 }, { "epoch": 0.6314134831063777, "grad_norm": 1.4685653448104858, "learning_rate": 9.621275288471329e-05, "loss": 3.4252, "step": 71500 }, { "epoch": 0.6318550309966618, "grad_norm": 4.989869594573975, "learning_rate": 9.620744587815435e-05, "loss": 3.2167, "step": 71550 }, { "epoch": 0.6322965788869461, "grad_norm": 1.4381548166275024, "learning_rate": 9.620213530246205e-05, "loss": 3.5436, "step": 71600 }, { "epoch": 0.6327381267772303, "grad_norm": 1.2252341508865356, "learning_rate": 9.619682115804656e-05, "loss": 3.0586, "step": 71650 }, { "epoch": 0.6331796746675145, "grad_norm": 4.090355396270752, "learning_rate": 9.619150344531838e-05, "loss": 3.4366, "step": 71700 }, { "epoch": 0.6336212225577986, "grad_norm": 1.2702091932296753, "learning_rate": 9.618618216468824e-05, "loss": 2.9698, "step": 71750 }, { "epoch": 0.6340627704480828, "grad_norm": 3.03657865524292, "learning_rate": 9.61808573165672e-05, "loss": 3.0061, "step": 71800 }, { "epoch": 0.634504318338367, "grad_norm": 2.344271421432495, "learning_rate": 9.617552890136652e-05, "loss": 2.6665, "step": 71850 }, { "epoch": 0.6349458662286511, "grad_norm": 1.6858032941818237, "learning_rate": 9.61701969194978e-05, "loss": 3.1685, "step": 71900 }, { "epoch": 0.6353874141189353, "grad_norm": 3.873769760131836, "learning_rate": 9.616486137137287e-05, "loss": 3.5466, "step": 71950 }, { "epoch": 0.6358289620092196, "grad_norm": 1.7075845003128052, "learning_rate": 9.615952225740385e-05, "loss": 3.3663, "step": 72000 }, { "epoch": 0.6358289620092196, "eval_asr_loss": 0.9085645562718005, "eval_loss": 3.0017716884613037, "eval_runtime": 20.9003, "eval_samples_per_second": 36.746, "eval_steps_per_second": 9.186, "eval_tts_loss": 5.94018095728795, "step": 72000 }, { "epoch": 0.6362705098995037, "grad_norm": 2.023348331451416, "learning_rate": 9.615417957800319e-05, "loss": 3.2043, "step": 72050 }, { "epoch": 0.6367120577897879, "grad_norm": 1.3778334856033325, "learning_rate": 9.614883333358351e-05, "loss": 3.2908, "step": 72100 }, { "epoch": 0.6371536056800721, "grad_norm": 2.4582135677337646, "learning_rate": 9.61434835245578e-05, "loss": 3.4521, "step": 72150 }, { "epoch": 0.6375951535703562, "grad_norm": 4.403940200805664, "learning_rate": 9.613813015133926e-05, "loss": 3.1724, "step": 72200 }, { "epoch": 0.6380367014606404, "grad_norm": 1.1220415830612183, "learning_rate": 9.613277321434141e-05, "loss": 3.2921, "step": 72250 }, { "epoch": 0.6384782493509246, "grad_norm": 5.00771427154541, "learning_rate": 9.612741271397802e-05, "loss": 3.4677, "step": 72300 }, { "epoch": 0.6389197972412087, "grad_norm": 2.081688404083252, "learning_rate": 9.612204865066317e-05, "loss": 2.9708, "step": 72350 }, { "epoch": 0.639361345131493, "grad_norm": 1.0858304500579834, "learning_rate": 9.611668102481114e-05, "loss": 3.3104, "step": 72400 }, { "epoch": 0.6398028930217772, "grad_norm": 4.691798686981201, "learning_rate": 9.611141729550216e-05, "loss": 3.4249, "step": 72450 }, { "epoch": 0.6402444409120613, "grad_norm": 1.6770505905151367, "learning_rate": 9.610604261705e-05, "loss": 3.3248, "step": 72500 }, { "epoch": 0.6406859888023455, "grad_norm": 0.5892982482910156, "learning_rate": 9.610066437729704e-05, "loss": 3.2216, "step": 72550 }, { "epoch": 0.6411275366926297, "grad_norm": 2.631664514541626, "learning_rate": 9.609528257665864e-05, "loss": 3.4193, "step": 72600 }, { "epoch": 0.6415690845829138, "grad_norm": 1.2707042694091797, "learning_rate": 9.608989721555055e-05, "loss": 3.3635, "step": 72650 }, { "epoch": 0.642010632473198, "grad_norm": 3.7268316745758057, "learning_rate": 9.608450829438873e-05, "loss": 3.5901, "step": 72700 }, { "epoch": 0.6424521803634822, "grad_norm": 3.723600387573242, "learning_rate": 9.607911581358943e-05, "loss": 3.157, "step": 72750 }, { "epoch": 0.6428937282537664, "grad_norm": 1.398940086364746, "learning_rate": 9.607371977356918e-05, "loss": 3.2389, "step": 72800 }, { "epoch": 0.6433352761440506, "grad_norm": 3.619192600250244, "learning_rate": 9.606832017474477e-05, "loss": 3.5136, "step": 72850 }, { "epoch": 0.6437768240343348, "grad_norm": 1.5558445453643799, "learning_rate": 9.606291701753327e-05, "loss": 3.4748, "step": 72900 }, { "epoch": 0.644218371924619, "grad_norm": 3.18674635887146, "learning_rate": 9.605751030235204e-05, "loss": 2.6554, "step": 72950 }, { "epoch": 0.6446599198149031, "grad_norm": 4.52305793762207, "learning_rate": 9.605210002961868e-05, "loss": 3.6155, "step": 73000 }, { "epoch": 0.6451014677051873, "grad_norm": 5.28046989440918, "learning_rate": 9.604668619975112e-05, "loss": 3.404, "step": 73050 }, { "epoch": 0.6455430155954714, "grad_norm": 2.169076442718506, "learning_rate": 9.604126881316751e-05, "loss": 3.3904, "step": 73100 }, { "epoch": 0.6459845634857556, "grad_norm": 1.4948843717575073, "learning_rate": 9.60358478702863e-05, "loss": 3.1796, "step": 73150 }, { "epoch": 0.6464261113760399, "grad_norm": 3.89113450050354, "learning_rate": 9.603042337152622e-05, "loss": 3.6965, "step": 73200 }, { "epoch": 0.6468676592663241, "grad_norm": 0.8911652565002441, "learning_rate": 9.602499531730625e-05, "loss": 3.4166, "step": 73250 }, { "epoch": 0.6473092071566082, "grad_norm": 1.3251335620880127, "learning_rate": 9.601956370804567e-05, "loss": 3.0175, "step": 73300 }, { "epoch": 0.6477507550468924, "grad_norm": 3.875401735305786, "learning_rate": 9.601412854416403e-05, "loss": 3.5598, "step": 73350 }, { "epoch": 0.6481923029371766, "grad_norm": 2.4360857009887695, "learning_rate": 9.600868982608116e-05, "loss": 3.3592, "step": 73400 }, { "epoch": 0.6486338508274607, "grad_norm": 2.4810409545898438, "learning_rate": 9.600324755421715e-05, "loss": 3.4158, "step": 73450 }, { "epoch": 0.6490753987177449, "grad_norm": 1.3619318008422852, "learning_rate": 9.599780172899234e-05, "loss": 2.9944, "step": 73500 }, { "epoch": 0.649516946608029, "grad_norm": 2.7849841117858887, "learning_rate": 9.599235235082741e-05, "loss": 3.6618, "step": 73550 }, { "epoch": 0.6499584944983133, "grad_norm": 5.122074127197266, "learning_rate": 9.598689942014326e-05, "loss": 3.0184, "step": 73600 }, { "epoch": 0.6504000423885975, "grad_norm": 2.697990655899048, "learning_rate": 9.59814429373611e-05, "loss": 3.0678, "step": 73650 }, { "epoch": 0.6508415902788817, "grad_norm": 1.3273183107376099, "learning_rate": 9.597598290290237e-05, "loss": 2.9799, "step": 73700 }, { "epoch": 0.6512831381691658, "grad_norm": 1.5726739168167114, "learning_rate": 9.597051931718882e-05, "loss": 3.1769, "step": 73750 }, { "epoch": 0.65172468605945, "grad_norm": 2.7319064140319824, "learning_rate": 9.596505218064249e-05, "loss": 3.5969, "step": 73800 }, { "epoch": 0.6521662339497342, "grad_norm": 3.6344516277313232, "learning_rate": 9.595958149368563e-05, "loss": 3.2977, "step": 73850 }, { "epoch": 0.6526077818400183, "grad_norm": 2.6774423122406006, "learning_rate": 9.595410725674084e-05, "loss": 3.1628, "step": 73900 }, { "epoch": 0.6530493297303025, "grad_norm": 4.738192081451416, "learning_rate": 9.594862947023094e-05, "loss": 3.6289, "step": 73950 }, { "epoch": 0.6534908776205868, "grad_norm": 1.0268847942352295, "learning_rate": 9.594314813457905e-05, "loss": 3.6385, "step": 74000 }, { "epoch": 0.6539324255108709, "grad_norm": 2.7158186435699463, "learning_rate": 9.593766325020855e-05, "loss": 3.4732, "step": 74050 }, { "epoch": 0.6543739734011551, "grad_norm": 2.4069275856018066, "learning_rate": 9.593217481754311e-05, "loss": 3.493, "step": 74100 }, { "epoch": 0.6548155212914393, "grad_norm": 1.4273890256881714, "learning_rate": 9.592668283700665e-05, "loss": 3.4282, "step": 74150 }, { "epoch": 0.6552570691817234, "grad_norm": 1.4611910581588745, "learning_rate": 9.59211873090234e-05, "loss": 3.5503, "step": 74200 }, { "epoch": 0.6556986170720076, "grad_norm": 1.2271146774291992, "learning_rate": 9.591568823401782e-05, "loss": 3.081, "step": 74250 }, { "epoch": 0.6561401649622918, "grad_norm": 2.7792177200317383, "learning_rate": 9.591018561241467e-05, "loss": 3.4418, "step": 74300 }, { "epoch": 0.6565817128525759, "grad_norm": 5.776242733001709, "learning_rate": 9.5904679444639e-05, "loss": 3.5402, "step": 74350 }, { "epoch": 0.6570232607428602, "grad_norm": 2.6737334728240967, "learning_rate": 9.58991697311161e-05, "loss": 3.2469, "step": 74400 }, { "epoch": 0.6574648086331444, "grad_norm": 1.3555004596710205, "learning_rate": 9.589365647227155e-05, "loss": 3.3199, "step": 74450 }, { "epoch": 0.6579063565234285, "grad_norm": 3.3269264698028564, "learning_rate": 9.58881396685312e-05, "loss": 3.462, "step": 74500 }, { "epoch": 0.6583479044137127, "grad_norm": 2.6994597911834717, "learning_rate": 9.58826193203212e-05, "loss": 3.3227, "step": 74550 }, { "epoch": 0.6587894523039969, "grad_norm": 3.3708934783935547, "learning_rate": 9.587709542806792e-05, "loss": 3.1164, "step": 74600 }, { "epoch": 0.659231000194281, "grad_norm": 1.3764150142669678, "learning_rate": 9.587156799219804e-05, "loss": 3.3193, "step": 74650 }, { "epoch": 0.6596725480845652, "grad_norm": 4.554634094238281, "learning_rate": 9.586603701313852e-05, "loss": 3.5422, "step": 74700 }, { "epoch": 0.6601140959748494, "grad_norm": 4.359382152557373, "learning_rate": 9.586050249131658e-05, "loss": 3.2248, "step": 74750 }, { "epoch": 0.6605556438651337, "grad_norm": 4.0737714767456055, "learning_rate": 9.585496442715968e-05, "loss": 3.3829, "step": 74800 }, { "epoch": 0.6609971917554178, "grad_norm": 2.545942544937134, "learning_rate": 9.584942282109564e-05, "loss": 3.2756, "step": 74850 }, { "epoch": 0.661438739645702, "grad_norm": 3.28908109664917, "learning_rate": 9.584387767355247e-05, "loss": 3.4446, "step": 74900 }, { "epoch": 0.6618802875359862, "grad_norm": 5.4131340980529785, "learning_rate": 9.58383289849585e-05, "loss": 2.9662, "step": 74950 }, { "epoch": 0.6623218354262703, "grad_norm": 1.5948145389556885, "learning_rate": 9.58327767557423e-05, "loss": 3.3822, "step": 75000 }, { "epoch": 0.6623218354262703, "eval_asr_loss": 0.8949383808222489, "eval_loss": 3.007831573486328, "eval_runtime": 20.8134, "eval_samples_per_second": 36.899, "eval_steps_per_second": 9.225, "eval_tts_loss": 5.953438474404635, "step": 75000 }, { "epoch": 0.6627633833165545, "grad_norm": 3.2304179668426514, "learning_rate": 9.582722098633276e-05, "loss": 3.3734, "step": 75050 }, { "epoch": 0.6632049312068387, "grad_norm": 3.123548746109009, "learning_rate": 9.5821661677159e-05, "loss": 3.4995, "step": 75100 }, { "epoch": 0.6636464790971228, "grad_norm": 2.3339643478393555, "learning_rate": 9.581609882865044e-05, "loss": 3.3416, "step": 75150 }, { "epoch": 0.6640880269874071, "grad_norm": 1.0405722856521606, "learning_rate": 9.58106438036635e-05, "loss": 3.3155, "step": 75200 }, { "epoch": 0.6645295748776913, "grad_norm": 3.653280258178711, "learning_rate": 9.580507394853993e-05, "loss": 2.9206, "step": 75250 }, { "epoch": 0.6649711227679754, "grad_norm": 2.6631522178649902, "learning_rate": 9.579950055536279e-05, "loss": 3.1143, "step": 75300 }, { "epoch": 0.6654126706582596, "grad_norm": 2.752356767654419, "learning_rate": 9.579392362456264e-05, "loss": 3.4541, "step": 75350 }, { "epoch": 0.6658542185485438, "grad_norm": 1.8876862525939941, "learning_rate": 9.578834315657019e-05, "loss": 2.942, "step": 75400 }, { "epoch": 0.6662957664388279, "grad_norm": 3.700727939605713, "learning_rate": 9.57827591518165e-05, "loss": 3.4877, "step": 75450 }, { "epoch": 0.6667373143291121, "grad_norm": 1.4538304805755615, "learning_rate": 9.57771716107329e-05, "loss": 3.5755, "step": 75500 }, { "epoch": 0.6671788622193963, "grad_norm": 4.2887349128723145, "learning_rate": 9.577158053375096e-05, "loss": 3.3367, "step": 75550 }, { "epoch": 0.6676204101096805, "grad_norm": 4.3281025886535645, "learning_rate": 9.576598592130257e-05, "loss": 3.5268, "step": 75600 }, { "epoch": 0.6680619579999647, "grad_norm": 2.3635294437408447, "learning_rate": 9.576038777381984e-05, "loss": 3.6769, "step": 75650 }, { "epoch": 0.6685035058902489, "grad_norm": 4.012321949005127, "learning_rate": 9.57547860917352e-05, "loss": 3.1022, "step": 75700 }, { "epoch": 0.668945053780533, "grad_norm": 1.8018879890441895, "learning_rate": 9.574918087548132e-05, "loss": 3.4172, "step": 75750 }, { "epoch": 0.6693866016708172, "grad_norm": 2.358532190322876, "learning_rate": 9.574357212549116e-05, "loss": 3.1392, "step": 75800 }, { "epoch": 0.6698281495611014, "grad_norm": 2.4655981063842773, "learning_rate": 9.573795984219793e-05, "loss": 3.3193, "step": 75850 }, { "epoch": 0.6702696974513855, "grad_norm": 2.9911704063415527, "learning_rate": 9.573234402603517e-05, "loss": 3.563, "step": 75900 }, { "epoch": 0.6707112453416697, "grad_norm": 5.0430707931518555, "learning_rate": 9.572672467743664e-05, "loss": 3.2619, "step": 75950 }, { "epoch": 0.671152793231954, "grad_norm": 2.1028594970703125, "learning_rate": 9.572110179683637e-05, "loss": 3.6662, "step": 76000 }, { "epoch": 0.6715943411222381, "grad_norm": 2.4186270236968994, "learning_rate": 9.571547538466868e-05, "loss": 3.0861, "step": 76050 }, { "epoch": 0.6720358890125223, "grad_norm": 1.6107758283615112, "learning_rate": 9.570984544136819e-05, "loss": 3.9137, "step": 76100 }, { "epoch": 0.6724774369028065, "grad_norm": 3.140552043914795, "learning_rate": 9.570432467144772e-05, "loss": 3.643, "step": 76150 }, { "epoch": 0.6729189847930906, "grad_norm": 3.5983471870422363, "learning_rate": 9.569868773778745e-05, "loss": 3.6072, "step": 76200 }, { "epoch": 0.6733605326833748, "grad_norm": 3.1312403678894043, "learning_rate": 9.569304727429107e-05, "loss": 3.3982, "step": 76250 }, { "epoch": 0.673802080573659, "grad_norm": 1.2464442253112793, "learning_rate": 9.568740328139423e-05, "loss": 3.36, "step": 76300 }, { "epoch": 0.6742436284639431, "grad_norm": 1.5809202194213867, "learning_rate": 9.568175575953293e-05, "loss": 3.404, "step": 76350 }, { "epoch": 0.6746851763542274, "grad_norm": 2.486264944076538, "learning_rate": 9.567610470914334e-05, "loss": 3.4402, "step": 76400 }, { "epoch": 0.6751267242445116, "grad_norm": 1.804176926612854, "learning_rate": 9.5670450130662e-05, "loss": 3.2879, "step": 76450 }, { "epoch": 0.6755682721347958, "grad_norm": 1.7221146821975708, "learning_rate": 9.566479202452567e-05, "loss": 3.2473, "step": 76500 }, { "epoch": 0.6760098200250799, "grad_norm": 2.4581539630889893, "learning_rate": 9.565913039117138e-05, "loss": 3.7328, "step": 76550 }, { "epoch": 0.6764513679153641, "grad_norm": 1.7588177919387817, "learning_rate": 9.565346523103643e-05, "loss": 3.1056, "step": 76600 }, { "epoch": 0.6768929158056483, "grad_norm": 2.841600179672241, "learning_rate": 9.564779654455844e-05, "loss": 3.3631, "step": 76650 }, { "epoch": 0.6773344636959324, "grad_norm": 2.2738077640533447, "learning_rate": 9.564212433217523e-05, "loss": 3.4339, "step": 76700 }, { "epoch": 0.6777760115862166, "grad_norm": 1.1713546514511108, "learning_rate": 9.563644859432497e-05, "loss": 3.2702, "step": 76750 }, { "epoch": 0.6782175594765009, "grad_norm": 2.2375364303588867, "learning_rate": 9.563076933144602e-05, "loss": 3.6829, "step": 76800 }, { "epoch": 0.678659107366785, "grad_norm": 1.015380620956421, "learning_rate": 9.562508654397708e-05, "loss": 3.0635, "step": 76850 }, { "epoch": 0.6791006552570692, "grad_norm": 0.6230117082595825, "learning_rate": 9.56194002323571e-05, "loss": 3.5006, "step": 76900 }, { "epoch": 0.6795422031473534, "grad_norm": 1.3702956438064575, "learning_rate": 9.56137103970253e-05, "loss": 3.107, "step": 76950 }, { "epoch": 0.6799837510376375, "grad_norm": 1.878818392753601, "learning_rate": 9.560801703842117e-05, "loss": 3.6604, "step": 77000 }, { "epoch": 0.6804252989279217, "grad_norm": 1.3559465408325195, "learning_rate": 9.560232015698444e-05, "loss": 3.634, "step": 77050 }, { "epoch": 0.6808668468182059, "grad_norm": 15.994784355163574, "learning_rate": 9.559661975315519e-05, "loss": 3.4924, "step": 77100 }, { "epoch": 0.68130839470849, "grad_norm": 0.9013010859489441, "learning_rate": 9.559091582737372e-05, "loss": 3.474, "step": 77150 }, { "epoch": 0.6817499425987743, "grad_norm": 1.3329524993896484, "learning_rate": 9.558520838008059e-05, "loss": 3.4002, "step": 77200 }, { "epoch": 0.6821914904890585, "grad_norm": 2.5802180767059326, "learning_rate": 9.557949741171667e-05, "loss": 3.1695, "step": 77250 }, { "epoch": 0.6826330383793426, "grad_norm": 2.464118242263794, "learning_rate": 9.557378292272307e-05, "loss": 3.337, "step": 77300 }, { "epoch": 0.6830745862696268, "grad_norm": 1.7328948974609375, "learning_rate": 9.556806491354121e-05, "loss": 3.085, "step": 77350 }, { "epoch": 0.683516134159911, "grad_norm": 1.6609792709350586, "learning_rate": 9.556234338461274e-05, "loss": 3.7844, "step": 77400 }, { "epoch": 0.6839576820501951, "grad_norm": 3.412128210067749, "learning_rate": 9.55566183363796e-05, "loss": 3.3013, "step": 77450 }, { "epoch": 0.6843992299404793, "grad_norm": 1.4039990901947021, "learning_rate": 9.555088976928399e-05, "loss": 3.2333, "step": 77500 }, { "epoch": 0.6848407778307635, "grad_norm": 1.7826491594314575, "learning_rate": 9.55451576837684e-05, "loss": 3.4012, "step": 77550 }, { "epoch": 0.6852823257210477, "grad_norm": 3.7740890979766846, "learning_rate": 9.553942208027559e-05, "loss": 3.2525, "step": 77600 }, { "epoch": 0.6857238736113319, "grad_norm": 3.7978830337524414, "learning_rate": 9.553368295924861e-05, "loss": 3.3832, "step": 77650 }, { "epoch": 0.6861654215016161, "grad_norm": 1.697966456413269, "learning_rate": 9.552794032113073e-05, "loss": 3.2114, "step": 77700 }, { "epoch": 0.6866069693919002, "grad_norm": 1.4913134574890137, "learning_rate": 9.552219416636552e-05, "loss": 3.1685, "step": 77750 }, { "epoch": 0.6870485172821844, "grad_norm": 2.034682512283325, "learning_rate": 9.551644449539683e-05, "loss": 2.8857, "step": 77800 }, { "epoch": 0.6874900651724686, "grad_norm": 1.9270625114440918, "learning_rate": 9.551069130866877e-05, "loss": 3.5445, "step": 77850 }, { "epoch": 0.6879316130627527, "grad_norm": 5.204434871673584, "learning_rate": 9.550493460662572e-05, "loss": 3.3383, "step": 77900 }, { "epoch": 0.6883731609530369, "grad_norm": 3.010862350463867, "learning_rate": 9.549917438971235e-05, "loss": 3.3101, "step": 77950 }, { "epoch": 0.6888147088433212, "grad_norm": 0.8644163608551025, "learning_rate": 9.549341065837357e-05, "loss": 3.3201, "step": 78000 }, { "epoch": 0.6888147088433212, "eval_asr_loss": 0.8925247313395483, "eval_loss": 2.992865800857544, "eval_runtime": 20.774, "eval_samples_per_second": 36.969, "eval_steps_per_second": 9.242, "eval_tts_loss": 5.960387531532541, "step": 78000 }, { "epoch": 0.6892562567336054, "grad_norm": 2.4011969566345215, "learning_rate": 9.54876434130546e-05, "loss": 3.9292, "step": 78050 }, { "epoch": 0.6896978046238895, "grad_norm": 3.125196695327759, "learning_rate": 9.54818726542009e-05, "loss": 3.247, "step": 78100 }, { "epoch": 0.6901393525141737, "grad_norm": 1.368432879447937, "learning_rate": 9.547609838225821e-05, "loss": 2.951, "step": 78150 }, { "epoch": 0.6905809004044579, "grad_norm": 2.4032599925994873, "learning_rate": 9.547032059767253e-05, "loss": 3.5267, "step": 78200 }, { "epoch": 0.691022448294742, "grad_norm": 2.1227686405181885, "learning_rate": 9.546453930089019e-05, "loss": 3.2206, "step": 78250 }, { "epoch": 0.6914639961850262, "grad_norm": 1.2852637767791748, "learning_rate": 9.54587544923577e-05, "loss": 3.4086, "step": 78300 }, { "epoch": 0.6919055440753105, "grad_norm": 1.241897463798523, "learning_rate": 9.54529661725219e-05, "loss": 3.1159, "step": 78350 }, { "epoch": 0.6923470919655946, "grad_norm": 1.472129225730896, "learning_rate": 9.544717434182991e-05, "loss": 3.2442, "step": 78400 }, { "epoch": 0.6927886398558788, "grad_norm": 2.743114471435547, "learning_rate": 9.544137900072907e-05, "loss": 3.2978, "step": 78450 }, { "epoch": 0.693230187746163, "grad_norm": 2.720283031463623, "learning_rate": 9.543558014966703e-05, "loss": 3.5339, "step": 78500 }, { "epoch": 0.6936717356364471, "grad_norm": 2.2413861751556396, "learning_rate": 9.542977778909173e-05, "loss": 3.6654, "step": 78550 }, { "epoch": 0.6941132835267313, "grad_norm": 0.8887834548950195, "learning_rate": 9.542397191945129e-05, "loss": 2.9958, "step": 78600 }, { "epoch": 0.6945548314170155, "grad_norm": 1.2447916269302368, "learning_rate": 9.541816254119425e-05, "loss": 3.0432, "step": 78650 }, { "epoch": 0.6949963793072996, "grad_norm": 2.320293664932251, "learning_rate": 9.541234965476925e-05, "loss": 3.4944, "step": 78700 }, { "epoch": 0.6954379271975839, "grad_norm": 2.7362334728240967, "learning_rate": 9.540653326062534e-05, "loss": 3.4683, "step": 78750 }, { "epoch": 0.6958794750878681, "grad_norm": 1.2431176900863647, "learning_rate": 9.540071335921176e-05, "loss": 3.2625, "step": 78800 }, { "epoch": 0.6963210229781522, "grad_norm": 2.1052627563476562, "learning_rate": 9.539488995097807e-05, "loss": 3.3472, "step": 78850 }, { "epoch": 0.6967625708684364, "grad_norm": 2.4554333686828613, "learning_rate": 9.538906303637407e-05, "loss": 3.327, "step": 78900 }, { "epoch": 0.6972041187587206, "grad_norm": 2.1302428245544434, "learning_rate": 9.538323261584985e-05, "loss": 3.4227, "step": 78950 }, { "epoch": 0.6976456666490047, "grad_norm": 1.9906952381134033, "learning_rate": 9.537739868985574e-05, "loss": 3.8786, "step": 79000 }, { "epoch": 0.6980872145392889, "grad_norm": 4.157954216003418, "learning_rate": 9.537156125884236e-05, "loss": 3.0816, "step": 79050 }, { "epoch": 0.6985287624295731, "grad_norm": 4.053858757019043, "learning_rate": 9.536572032326061e-05, "loss": 3.2263, "step": 79100 }, { "epoch": 0.6989703103198573, "grad_norm": 1.9072597026824951, "learning_rate": 9.535987588356168e-05, "loss": 3.4913, "step": 79150 }, { "epoch": 0.6994118582101415, "grad_norm": 4.11290168762207, "learning_rate": 9.535402794019696e-05, "loss": 3.5914, "step": 79200 }, { "epoch": 0.6998534061004257, "grad_norm": 3.6236138343811035, "learning_rate": 9.534817649361816e-05, "loss": 3.0452, "step": 79250 }, { "epoch": 0.7002949539907098, "grad_norm": 4.064155578613281, "learning_rate": 9.534232154427728e-05, "loss": 3.5526, "step": 79300 }, { "epoch": 0.700736501880994, "grad_norm": 2.754270553588867, "learning_rate": 9.533646309262657e-05, "loss": 3.3366, "step": 79350 }, { "epoch": 0.7011780497712782, "grad_norm": 2.6481540203094482, "learning_rate": 9.533060113911852e-05, "loss": 3.6007, "step": 79400 }, { "epoch": 0.7016195976615623, "grad_norm": 3.322692632675171, "learning_rate": 9.532473568420591e-05, "loss": 3.1894, "step": 79450 }, { "epoch": 0.7020611455518465, "grad_norm": 2.9386980533599854, "learning_rate": 9.531886672834182e-05, "loss": 3.1106, "step": 79500 }, { "epoch": 0.7025026934421308, "grad_norm": 1.7416975498199463, "learning_rate": 9.531299427197957e-05, "loss": 3.5195, "step": 79550 }, { "epoch": 0.702944241332415, "grad_norm": 2.034696578979492, "learning_rate": 9.530711831557274e-05, "loss": 3.3598, "step": 79600 }, { "epoch": 0.7033857892226991, "grad_norm": 2.048043966293335, "learning_rate": 9.530123885957523e-05, "loss": 3.3274, "step": 79650 }, { "epoch": 0.7038273371129833, "grad_norm": 0.8644744157791138, "learning_rate": 9.529535590444115e-05, "loss": 3.3067, "step": 79700 }, { "epoch": 0.7042688850032675, "grad_norm": 3.338059186935425, "learning_rate": 9.528946945062493e-05, "loss": 3.2878, "step": 79750 }, { "epoch": 0.7047104328935516, "grad_norm": 1.424256682395935, "learning_rate": 9.528357949858123e-05, "loss": 3.5114, "step": 79800 }, { "epoch": 0.7051519807838358, "grad_norm": 3.7967846393585205, "learning_rate": 9.527768604876501e-05, "loss": 3.4413, "step": 79850 }, { "epoch": 0.70559352867412, "grad_norm": 3.256422996520996, "learning_rate": 9.52717891016315e-05, "loss": 3.0054, "step": 79900 }, { "epoch": 0.7060350765644042, "grad_norm": 3.2881686687469482, "learning_rate": 9.526588865763614e-05, "loss": 3.3926, "step": 79950 }, { "epoch": 0.7064766244546884, "grad_norm": 2.3848628997802734, "learning_rate": 9.525998471723476e-05, "loss": 3.3552, "step": 80000 }, { "epoch": 0.7069181723449726, "grad_norm": 3.0554099082946777, "learning_rate": 9.525407728088333e-05, "loss": 3.4239, "step": 80050 }, { "epoch": 0.7073597202352567, "grad_norm": 2.6418068408966064, "learning_rate": 9.524816634903819e-05, "loss": 3.2979, "step": 80100 }, { "epoch": 0.7078012681255409, "grad_norm": 1.501378059387207, "learning_rate": 9.52422519221559e-05, "loss": 3.7132, "step": 80150 }, { "epoch": 0.7082428160158251, "grad_norm": 3.162963390350342, "learning_rate": 9.523633400069327e-05, "loss": 3.7012, "step": 80200 }, { "epoch": 0.7086843639061092, "grad_norm": 0.6245143413543701, "learning_rate": 9.523041258510746e-05, "loss": 3.423, "step": 80250 }, { "epoch": 0.7091259117963934, "grad_norm": 0.8148535490036011, "learning_rate": 9.522448767585579e-05, "loss": 3.3979, "step": 80300 }, { "epoch": 0.7095674596866777, "grad_norm": 2.1384999752044678, "learning_rate": 9.521855927339595e-05, "loss": 2.7129, "step": 80350 }, { "epoch": 0.7100090075769618, "grad_norm": 3.697734832763672, "learning_rate": 9.521262737818586e-05, "loss": 3.0784, "step": 80400 }, { "epoch": 0.710450555467246, "grad_norm": 6.074985027313232, "learning_rate": 9.520669199068368e-05, "loss": 3.4716, "step": 80450 }, { "epoch": 0.7108921033575302, "grad_norm": 3.228442907333374, "learning_rate": 9.520075311134788e-05, "loss": 3.3426, "step": 80500 }, { "epoch": 0.7113336512478143, "grad_norm": 1.2277064323425293, "learning_rate": 9.51948107406372e-05, "loss": 3.2555, "step": 80550 }, { "epoch": 0.7117751991380985, "grad_norm": 0.9551520943641663, "learning_rate": 9.518886487901065e-05, "loss": 3.5223, "step": 80600 }, { "epoch": 0.7122167470283827, "grad_norm": 2.2914512157440186, "learning_rate": 9.518291552692746e-05, "loss": 2.9172, "step": 80650 }, { "epoch": 0.7126582949186668, "grad_norm": 3.3304460048675537, "learning_rate": 9.517696268484719e-05, "loss": 2.6855, "step": 80700 }, { "epoch": 0.7130998428089511, "grad_norm": 1.3867692947387695, "learning_rate": 9.517100635322966e-05, "loss": 2.9664, "step": 80750 }, { "epoch": 0.7135413906992353, "grad_norm": 1.0515697002410889, "learning_rate": 9.51650465325349e-05, "loss": 3.2619, "step": 80800 }, { "epoch": 0.7139829385895194, "grad_norm": 1.7852091789245605, "learning_rate": 9.515908322322329e-05, "loss": 3.0423, "step": 80850 }, { "epoch": 0.7144244864798036, "grad_norm": 1.8944436311721802, "learning_rate": 9.515311642575545e-05, "loss": 3.1392, "step": 80900 }, { "epoch": 0.7148660343700878, "grad_norm": 0.9419056177139282, "learning_rate": 9.514714614059225e-05, "loss": 3.1358, "step": 80950 }, { "epoch": 0.7153075822603719, "grad_norm": 3.2333171367645264, "learning_rate": 9.514117236819485e-05, "loss": 3.8758, "step": 81000 }, { "epoch": 0.7153075822603719, "eval_asr_loss": 0.9185007565241553, "eval_loss": 2.989028215408325, "eval_runtime": 20.4458, "eval_samples_per_second": 37.563, "eval_steps_per_second": 9.391, "eval_tts_loss": 5.958070210757005, "step": 81000 }, { "epoch": 0.7157491301506561, "grad_norm": 2.493988037109375, "learning_rate": 9.513519510902466e-05, "loss": 3.6492, "step": 81050 }, { "epoch": 0.7161906780409403, "grad_norm": 2.0979015827178955, "learning_rate": 9.512921436354339e-05, "loss": 3.7933, "step": 81100 }, { "epoch": 0.7166322259312246, "grad_norm": 2.046807050704956, "learning_rate": 9.512323013221299e-05, "loss": 3.291, "step": 81150 }, { "epoch": 0.7170737738215087, "grad_norm": 2.943680763244629, "learning_rate": 9.51172424154957e-05, "loss": 3.0745, "step": 81200 }, { "epoch": 0.7175153217117929, "grad_norm": 2.4897234439849854, "learning_rate": 9.5111251213854e-05, "loss": 3.1611, "step": 81250 }, { "epoch": 0.717956869602077, "grad_norm": 4.505876541137695, "learning_rate": 9.51052565277507e-05, "loss": 3.5499, "step": 81300 }, { "epoch": 0.7183984174923612, "grad_norm": 1.430411458015442, "learning_rate": 9.50992583576488e-05, "loss": 3.1281, "step": 81350 }, { "epoch": 0.7188399653826454, "grad_norm": 3.1512632369995117, "learning_rate": 9.509325670401164e-05, "loss": 3.4918, "step": 81400 }, { "epoch": 0.7192815132729296, "grad_norm": 1.2546844482421875, "learning_rate": 9.508725156730276e-05, "loss": 3.6289, "step": 81450 }, { "epoch": 0.7197230611632137, "grad_norm": 3.754129409790039, "learning_rate": 9.508124294798604e-05, "loss": 2.9913, "step": 81500 }, { "epoch": 0.720164609053498, "grad_norm": 3.811619997024536, "learning_rate": 9.507523084652556e-05, "loss": 3.233, "step": 81550 }, { "epoch": 0.7206061569437822, "grad_norm": 1.469282627105713, "learning_rate": 9.5069335609166e-05, "loss": 3.6196, "step": 81600 }, { "epoch": 0.7210477048340663, "grad_norm": 2.8807554244995117, "learning_rate": 9.50633166144312e-05, "loss": 2.9091, "step": 81650 }, { "epoch": 0.7214892527243505, "grad_norm": 2.3154757022857666, "learning_rate": 9.505729413893734e-05, "loss": 3.2319, "step": 81700 }, { "epoch": 0.7219308006146347, "grad_norm": 1.580367922782898, "learning_rate": 9.505126818314955e-05, "loss": 3.5418, "step": 81750 }, { "epoch": 0.7223723485049188, "grad_norm": 1.0496435165405273, "learning_rate": 9.504523874753335e-05, "loss": 3.6296, "step": 81800 }, { "epoch": 0.722813896395203, "grad_norm": 0.9904226660728455, "learning_rate": 9.50392058325544e-05, "loss": 3.2764, "step": 81850 }, { "epoch": 0.7232554442854872, "grad_norm": 1.5296332836151123, "learning_rate": 9.503316943867876e-05, "loss": 3.6219, "step": 81900 }, { "epoch": 0.7236969921757714, "grad_norm": 1.79270601272583, "learning_rate": 9.502712956637263e-05, "loss": 3.467, "step": 81950 }, { "epoch": 0.7241385400660556, "grad_norm": 3.750113010406494, "learning_rate": 9.502108621610255e-05, "loss": 2.9975, "step": 82000 }, { "epoch": 0.7245800879563398, "grad_norm": 1.8261284828186035, "learning_rate": 9.501503938833534e-05, "loss": 3.2391, "step": 82050 }, { "epoch": 0.7250216358466239, "grad_norm": 1.2408541440963745, "learning_rate": 9.500911012370587e-05, "loss": 3.4596, "step": 82100 }, { "epoch": 0.7254631837369081, "grad_norm": 1.0297032594680786, "learning_rate": 9.50030564118725e-05, "loss": 3.1892, "step": 82150 }, { "epoch": 0.7259047316271923, "grad_norm": 2.6377224922180176, "learning_rate": 9.499699922393466e-05, "loss": 3.624, "step": 82200 }, { "epoch": 0.7263462795174764, "grad_norm": 3.4818551540374756, "learning_rate": 9.499093856036019e-05, "loss": 3.1586, "step": 82250 }, { "epoch": 0.7267878274077606, "grad_norm": 2.5095982551574707, "learning_rate": 9.498487442161721e-05, "loss": 3.2636, "step": 82300 }, { "epoch": 0.7272293752980449, "grad_norm": 1.323671579360962, "learning_rate": 9.497880680817414e-05, "loss": 3.3924, "step": 82350 }, { "epoch": 0.727670923188329, "grad_norm": 3.4375061988830566, "learning_rate": 9.497273572049967e-05, "loss": 3.2829, "step": 82400 }, { "epoch": 0.7281124710786132, "grad_norm": 1.7869148254394531, "learning_rate": 9.496666115906272e-05, "loss": 3.5715, "step": 82450 }, { "epoch": 0.7285540189688974, "grad_norm": 2.402039051055908, "learning_rate": 9.49605831243325e-05, "loss": 3.3708, "step": 82500 }, { "epoch": 0.7289955668591815, "grad_norm": 2.0941081047058105, "learning_rate": 9.49545016167785e-05, "loss": 3.3561, "step": 82550 }, { "epoch": 0.7294371147494657, "grad_norm": 3.094398021697998, "learning_rate": 9.494841663687043e-05, "loss": 3.5185, "step": 82600 }, { "epoch": 0.7298786626397499, "grad_norm": 1.317842960357666, "learning_rate": 9.494232818507833e-05, "loss": 3.3928, "step": 82650 }, { "epoch": 0.730320210530034, "grad_norm": 2.3808138370513916, "learning_rate": 9.493623626187249e-05, "loss": 3.375, "step": 82700 }, { "epoch": 0.7307617584203183, "grad_norm": 1.0954011678695679, "learning_rate": 9.493014086772345e-05, "loss": 3.2433, "step": 82750 }, { "epoch": 0.7312033063106025, "grad_norm": 2.7866711616516113, "learning_rate": 9.492404200310201e-05, "loss": 2.8652, "step": 82800 }, { "epoch": 0.7316448542008867, "grad_norm": 0.6492549777030945, "learning_rate": 9.491793966847928e-05, "loss": 3.688, "step": 82850 }, { "epoch": 0.7320864020911708, "grad_norm": 2.792008876800537, "learning_rate": 9.491183386432659e-05, "loss": 3.7061, "step": 82900 }, { "epoch": 0.732527949981455, "grad_norm": 3.021859884262085, "learning_rate": 9.49057245911156e-05, "loss": 3.5694, "step": 82950 }, { "epoch": 0.7329694978717392, "grad_norm": 2.1141061782836914, "learning_rate": 9.489961184931815e-05, "loss": 3.1073, "step": 83000 }, { "epoch": 0.7334110457620233, "grad_norm": 1.2735497951507568, "learning_rate": 9.489349563940644e-05, "loss": 3.3421, "step": 83050 }, { "epoch": 0.7338525936523075, "grad_norm": 1.5497995615005493, "learning_rate": 9.488737596185286e-05, "loss": 3.6448, "step": 83100 }, { "epoch": 0.7342941415425918, "grad_norm": 2.1992032527923584, "learning_rate": 9.488125281713013e-05, "loss": 3.4424, "step": 83150 }, { "epoch": 0.7347356894328759, "grad_norm": 4.297890663146973, "learning_rate": 9.487512620571121e-05, "loss": 3.191, "step": 83200 }, { "epoch": 0.7351772373231601, "grad_norm": 2.3956549167633057, "learning_rate": 9.486899612806931e-05, "loss": 3.0592, "step": 83250 }, { "epoch": 0.7356187852134443, "grad_norm": 1.5929914712905884, "learning_rate": 9.486286258467793e-05, "loss": 3.214, "step": 83300 }, { "epoch": 0.7360603331037284, "grad_norm": 1.5399370193481445, "learning_rate": 9.485672557601087e-05, "loss": 3.3211, "step": 83350 }, { "epoch": 0.7365018809940126, "grad_norm": 3.4927520751953125, "learning_rate": 9.48505851025421e-05, "loss": 3.287, "step": 83400 }, { "epoch": 0.7369434288842968, "grad_norm": 1.6479640007019043, "learning_rate": 9.484444116474597e-05, "loss": 3.2647, "step": 83450 }, { "epoch": 0.7373849767745809, "grad_norm": 4.172652721405029, "learning_rate": 9.483829376309703e-05, "loss": 3.0049, "step": 83500 }, { "epoch": 0.7378265246648652, "grad_norm": 2.3201801776885986, "learning_rate": 9.48321428980701e-05, "loss": 2.9019, "step": 83550 }, { "epoch": 0.7382680725551494, "grad_norm": 3.2628872394561768, "learning_rate": 9.482598857014031e-05, "loss": 3.2633, "step": 83600 }, { "epoch": 0.7387096204454335, "grad_norm": 1.8258609771728516, "learning_rate": 9.481983077978302e-05, "loss": 3.2017, "step": 83650 }, { "epoch": 0.7391511683357177, "grad_norm": 2.0017619132995605, "learning_rate": 9.481366952747386e-05, "loss": 3.2454, "step": 83700 }, { "epoch": 0.7395927162260019, "grad_norm": 2.6693084239959717, "learning_rate": 9.480750481368872e-05, "loss": 3.6843, "step": 83750 }, { "epoch": 0.740034264116286, "grad_norm": 1.9272335767745972, "learning_rate": 9.480133663890382e-05, "loss": 3.0776, "step": 83800 }, { "epoch": 0.7404758120065702, "grad_norm": 0.973445475101471, "learning_rate": 9.479516500359555e-05, "loss": 3.5528, "step": 83850 }, { "epoch": 0.7409173598968544, "grad_norm": 1.443723201751709, "learning_rate": 9.478898990824064e-05, "loss": 3.3918, "step": 83900 }, { "epoch": 0.7413589077871386, "grad_norm": 3.7001953125, "learning_rate": 9.478281135331606e-05, "loss": 3.3402, "step": 83950 }, { "epoch": 0.7418004556774228, "grad_norm": 2.7895514965057373, "learning_rate": 9.477662933929905e-05, "loss": 3.2964, "step": 84000 }, { "epoch": 0.7418004556774228, "eval_asr_loss": 0.907389885219024, "eval_loss": 2.9814484119415283, "eval_runtime": 21.1732, "eval_samples_per_second": 36.272, "eval_steps_per_second": 9.068, "eval_tts_loss": 5.974965912665154, "step": 84000 }, { "epoch": 0.742242003567707, "grad_norm": 2.343669891357422, "learning_rate": 9.477044386666711e-05, "loss": 3.0479, "step": 84050 }, { "epoch": 0.7426835514579911, "grad_norm": 2.6981098651885986, "learning_rate": 9.476425493589804e-05, "loss": 3.538, "step": 84100 }, { "epoch": 0.7431250993482753, "grad_norm": 2.0038726329803467, "learning_rate": 9.475806254746984e-05, "loss": 3.3287, "step": 84150 }, { "epoch": 0.7435666472385595, "grad_norm": 2.3365066051483154, "learning_rate": 9.475186670186088e-05, "loss": 3.4029, "step": 84200 }, { "epoch": 0.7440081951288436, "grad_norm": 3.0482475757598877, "learning_rate": 9.474566739954966e-05, "loss": 3.1918, "step": 84250 }, { "epoch": 0.7444497430191278, "grad_norm": 2.1410951614379883, "learning_rate": 9.47394646410151e-05, "loss": 3.1646, "step": 84300 }, { "epoch": 0.7448912909094121, "grad_norm": 2.7029786109924316, "learning_rate": 9.473325842673626e-05, "loss": 3.3533, "step": 84350 }, { "epoch": 0.7453328387996963, "grad_norm": 1.4176026582717896, "learning_rate": 9.472704875719254e-05, "loss": 3.5932, "step": 84400 }, { "epoch": 0.7457743866899804, "grad_norm": 3.4000325202941895, "learning_rate": 9.472083563286358e-05, "loss": 2.8913, "step": 84450 }, { "epoch": 0.7462159345802646, "grad_norm": 3.8159403800964355, "learning_rate": 9.47146190542293e-05, "loss": 3.0607, "step": 84500 }, { "epoch": 0.7466574824705488, "grad_norm": 2.903031349182129, "learning_rate": 9.470839902176985e-05, "loss": 3.1426, "step": 84550 }, { "epoch": 0.7470990303608329, "grad_norm": 1.7477384805679321, "learning_rate": 9.470217553596568e-05, "loss": 3.1242, "step": 84600 }, { "epoch": 0.7475405782511171, "grad_norm": 1.4132603406906128, "learning_rate": 9.469594859729753e-05, "loss": 3.1462, "step": 84650 }, { "epoch": 0.7479821261414012, "grad_norm": 1.7881561517715454, "learning_rate": 9.468971820624636e-05, "loss": 3.1489, "step": 84700 }, { "epoch": 0.7484236740316855, "grad_norm": 0.7184849381446838, "learning_rate": 9.468348436329341e-05, "loss": 3.5103, "step": 84750 }, { "epoch": 0.7488652219219697, "grad_norm": 5.096653938293457, "learning_rate": 9.46772470689202e-05, "loss": 3.6296, "step": 84800 }, { "epoch": 0.7493067698122539, "grad_norm": 1.7660068273544312, "learning_rate": 9.467100632360851e-05, "loss": 3.2378, "step": 84850 }, { "epoch": 0.749748317702538, "grad_norm": 3.2901999950408936, "learning_rate": 9.466476212784038e-05, "loss": 2.8817, "step": 84900 }, { "epoch": 0.7501898655928222, "grad_norm": 0.6989944577217102, "learning_rate": 9.465851448209813e-05, "loss": 3.2027, "step": 84950 }, { "epoch": 0.7506314134831064, "grad_norm": 4.7878875732421875, "learning_rate": 9.465226338686431e-05, "loss": 3.0253, "step": 85000 }, { "epoch": 0.7510729613733905, "grad_norm": 2.0710065364837646, "learning_rate": 9.46460088426218e-05, "loss": 3.5431, "step": 85050 }, { "epoch": 0.7515145092636747, "grad_norm": 1.119075059890747, "learning_rate": 9.463975084985369e-05, "loss": 3.4444, "step": 85100 }, { "epoch": 0.751956057153959, "grad_norm": 2.28783917427063, "learning_rate": 9.463348940904335e-05, "loss": 3.6863, "step": 85150 }, { "epoch": 0.7523976050442431, "grad_norm": 2.2328102588653564, "learning_rate": 9.462722452067445e-05, "loss": 3.2708, "step": 85200 }, { "epoch": 0.7528391529345273, "grad_norm": 1.885800838470459, "learning_rate": 9.462095618523089e-05, "loss": 3.7459, "step": 85250 }, { "epoch": 0.7532807008248115, "grad_norm": 1.719815731048584, "learning_rate": 9.461468440319681e-05, "loss": 3.7436, "step": 85300 }, { "epoch": 0.7537222487150956, "grad_norm": 1.0964494943618774, "learning_rate": 9.460840917505668e-05, "loss": 2.7071, "step": 85350 }, { "epoch": 0.7541637966053798, "grad_norm": 2.1939446926116943, "learning_rate": 9.460213050129525e-05, "loss": 3.4522, "step": 85400 }, { "epoch": 0.754605344495664, "grad_norm": 3.344871759414673, "learning_rate": 9.459584838239743e-05, "loss": 3.3535, "step": 85450 }, { "epoch": 0.7550468923859481, "grad_norm": 3.5036964416503906, "learning_rate": 9.458956281884848e-05, "loss": 3.1035, "step": 85500 }, { "epoch": 0.7554884402762324, "grad_norm": 1.8744654655456543, "learning_rate": 9.458327381113392e-05, "loss": 3.4614, "step": 85550 }, { "epoch": 0.7559299881665166, "grad_norm": 3.193300485610962, "learning_rate": 9.457698135973951e-05, "loss": 3.2385, "step": 85600 }, { "epoch": 0.7563715360568007, "grad_norm": 2.4154229164123535, "learning_rate": 9.457068546515129e-05, "loss": 3.1358, "step": 85650 }, { "epoch": 0.7568130839470849, "grad_norm": 1.8072885274887085, "learning_rate": 9.456438612785556e-05, "loss": 3.0887, "step": 85700 }, { "epoch": 0.7572546318373691, "grad_norm": 4.455030918121338, "learning_rate": 9.455808334833892e-05, "loss": 2.7886, "step": 85750 }, { "epoch": 0.7576961797276532, "grad_norm": 3.4446539878845215, "learning_rate": 9.455177712708815e-05, "loss": 3.045, "step": 85800 }, { "epoch": 0.7581377276179374, "grad_norm": 1.5495901107788086, "learning_rate": 9.45454674645904e-05, "loss": 3.3734, "step": 85850 }, { "epoch": 0.7585792755082216, "grad_norm": 1.636919379234314, "learning_rate": 9.453915436133303e-05, "loss": 3.1687, "step": 85900 }, { "epoch": 0.7590208233985059, "grad_norm": 1.9204362630844116, "learning_rate": 9.453283781780365e-05, "loss": 3.5536, "step": 85950 }, { "epoch": 0.75946237128879, "grad_norm": 2.4546802043914795, "learning_rate": 9.45265178344902e-05, "loss": 3.0949, "step": 86000 }, { "epoch": 0.7599039191790742, "grad_norm": 5.441671848297119, "learning_rate": 9.45201944118808e-05, "loss": 3.2127, "step": 86050 }, { "epoch": 0.7603454670693583, "grad_norm": 3.458197593688965, "learning_rate": 9.45138675504639e-05, "loss": 3.4892, "step": 86100 }, { "epoch": 0.7607870149596425, "grad_norm": 1.5939244031906128, "learning_rate": 9.450753725072822e-05, "loss": 3.2989, "step": 86150 }, { "epoch": 0.7612285628499267, "grad_norm": 2.7477095127105713, "learning_rate": 9.450120351316268e-05, "loss": 3.1591, "step": 86200 }, { "epoch": 0.7616701107402108, "grad_norm": 2.2054810523986816, "learning_rate": 9.449486633825654e-05, "loss": 3.2117, "step": 86250 }, { "epoch": 0.762111658630495, "grad_norm": 2.03580379486084, "learning_rate": 9.448852572649926e-05, "loss": 2.9641, "step": 86300 }, { "epoch": 0.7625532065207793, "grad_norm": 1.0966379642486572, "learning_rate": 9.448218167838065e-05, "loss": 3.3211, "step": 86350 }, { "epoch": 0.7629947544110635, "grad_norm": 2.210423707962036, "learning_rate": 9.44758341943907e-05, "loss": 3.2672, "step": 86400 }, { "epoch": 0.7634363023013476, "grad_norm": 3.2034225463867188, "learning_rate": 9.446948327501967e-05, "loss": 3.6083, "step": 86450 }, { "epoch": 0.7638778501916318, "grad_norm": 2.616360664367676, "learning_rate": 9.446312892075818e-05, "loss": 3.1633, "step": 86500 }, { "epoch": 0.764319398081916, "grad_norm": 2.7256109714508057, "learning_rate": 9.445689832152417e-05, "loss": 3.1589, "step": 86550 }, { "epoch": 0.7647609459722001, "grad_norm": 3.060117483139038, "learning_rate": 9.445053716762777e-05, "loss": 3.142, "step": 86600 }, { "epoch": 0.7652024938624843, "grad_norm": 1.2337734699249268, "learning_rate": 9.444417258030429e-05, "loss": 3.1518, "step": 86650 }, { "epoch": 0.7656440417527685, "grad_norm": 2.3434503078460693, "learning_rate": 9.443780456004537e-05, "loss": 3.2844, "step": 86700 }, { "epoch": 0.7660855896430527, "grad_norm": 2.32765793800354, "learning_rate": 9.443143310734285e-05, "loss": 3.131, "step": 86750 }, { "epoch": 0.7665271375333369, "grad_norm": 1.1229861974716187, "learning_rate": 9.442505822268891e-05, "loss": 3.4084, "step": 86800 }, { "epoch": 0.7669686854236211, "grad_norm": 2.0934159755706787, "learning_rate": 9.441867990657594e-05, "loss": 3.3533, "step": 86850 }, { "epoch": 0.7674102333139052, "grad_norm": 5.029802322387695, "learning_rate": 9.441229815949658e-05, "loss": 3.2617, "step": 86900 }, { "epoch": 0.7678517812041894, "grad_norm": 3.596031665802002, "learning_rate": 9.440591298194382e-05, "loss": 3.3018, "step": 86950 }, { "epoch": 0.7682933290944736, "grad_norm": 0.9849772453308105, "learning_rate": 9.439952437441082e-05, "loss": 2.6819, "step": 87000 }, { "epoch": 0.7682933290944736, "eval_asr_loss": 0.9071832583308085, "eval_loss": 2.9738311767578125, "eval_runtime": 20.831, "eval_samples_per_second": 36.868, "eval_steps_per_second": 9.217, "eval_tts_loss": 5.941716546998216, "step": 87000 }, { "epoch": 0.7687348769847577, "grad_norm": 1.9874849319458008, "learning_rate": 9.439313233739105e-05, "loss": 3.5471, "step": 87050 }, { "epoch": 0.7691764248750419, "grad_norm": 2.417670249938965, "learning_rate": 9.438673687137827e-05, "loss": 3.2482, "step": 87100 }, { "epoch": 0.7696179727653262, "grad_norm": 1.2283480167388916, "learning_rate": 9.438033797686642e-05, "loss": 3.6392, "step": 87150 }, { "epoch": 0.7700595206556103, "grad_norm": 4.871400833129883, "learning_rate": 9.437393565434984e-05, "loss": 3.7572, "step": 87200 }, { "epoch": 0.7705010685458945, "grad_norm": 2.2227866649627686, "learning_rate": 9.436752990432298e-05, "loss": 3.0988, "step": 87250 }, { "epoch": 0.7709426164361787, "grad_norm": 1.317663550376892, "learning_rate": 9.436112072728067e-05, "loss": 2.342, "step": 87300 }, { "epoch": 0.7713841643264628, "grad_norm": 1.2101949453353882, "learning_rate": 9.435470812371796e-05, "loss": 3.0555, "step": 87350 }, { "epoch": 0.771825712216747, "grad_norm": 2.3560891151428223, "learning_rate": 9.434829209413016e-05, "loss": 3.287, "step": 87400 }, { "epoch": 0.7722672601070312, "grad_norm": 2.26143217086792, "learning_rate": 9.434187263901286e-05, "loss": 3.1895, "step": 87450 }, { "epoch": 0.7727088079973153, "grad_norm": 2.7307794094085693, "learning_rate": 9.43354497588619e-05, "loss": 3.0531, "step": 87500 }, { "epoch": 0.7731503558875996, "grad_norm": 2.08046293258667, "learning_rate": 9.43290234541734e-05, "loss": 3.5845, "step": 87550 }, { "epoch": 0.7735919037778838, "grad_norm": 1.427642822265625, "learning_rate": 9.432259372544375e-05, "loss": 3.372, "step": 87600 }, { "epoch": 0.774033451668168, "grad_norm": 1.8793132305145264, "learning_rate": 9.431616057316957e-05, "loss": 3.44, "step": 87650 }, { "epoch": 0.7744749995584521, "grad_norm": 5.893092632293701, "learning_rate": 9.430972399784779e-05, "loss": 3.3203, "step": 87700 }, { "epoch": 0.7749165474487363, "grad_norm": 1.8131071329116821, "learning_rate": 9.430328399997555e-05, "loss": 3.1405, "step": 87750 }, { "epoch": 0.7753580953390204, "grad_norm": 1.3549493551254272, "learning_rate": 9.429696948198171e-05, "loss": 3.4958, "step": 87800 }, { "epoch": 0.7757996432293046, "grad_norm": 2.8641469478607178, "learning_rate": 9.429052270892739e-05, "loss": 3.0084, "step": 87850 }, { "epoch": 0.7762411911195888, "grad_norm": 1.4089255332946777, "learning_rate": 9.428407251480576e-05, "loss": 3.5739, "step": 87900 }, { "epoch": 0.7766827390098731, "grad_norm": 2.792038679122925, "learning_rate": 9.427761890011504e-05, "loss": 3.0026, "step": 87950 }, { "epoch": 0.7771242869001572, "grad_norm": 2.8735337257385254, "learning_rate": 9.427116186535371e-05, "loss": 2.9788, "step": 88000 }, { "epoch": 0.7775658347904414, "grad_norm": 2.6922061443328857, "learning_rate": 9.426470141102058e-05, "loss": 3.5371, "step": 88050 }, { "epoch": 0.7780073826807256, "grad_norm": 2.8780505657196045, "learning_rate": 9.425823753761458e-05, "loss": 3.4691, "step": 88100 }, { "epoch": 0.7784489305710097, "grad_norm": 3.15449595451355, "learning_rate": 9.425177024563504e-05, "loss": 3.0253, "step": 88150 }, { "epoch": 0.7788904784612939, "grad_norm": 2.4547994136810303, "learning_rate": 9.42452995355815e-05, "loss": 3.22, "step": 88200 }, { "epoch": 0.779332026351578, "grad_norm": 1.3049603700637817, "learning_rate": 9.423882540795375e-05, "loss": 3.3362, "step": 88250 }, { "epoch": 0.7797735742418622, "grad_norm": 2.944639205932617, "learning_rate": 9.423234786325188e-05, "loss": 2.6225, "step": 88300 }, { "epoch": 0.7802151221321465, "grad_norm": 4.7420783042907715, "learning_rate": 9.42258669019762e-05, "loss": 2.852, "step": 88350 }, { "epoch": 0.7806566700224307, "grad_norm": 2.200601100921631, "learning_rate": 9.421938252462734e-05, "loss": 3.4574, "step": 88400 }, { "epoch": 0.7810982179127148, "grad_norm": 0.8284490704536438, "learning_rate": 9.421289473170615e-05, "loss": 3.3901, "step": 88450 }, { "epoch": 0.781539765802999, "grad_norm": 2.629199266433716, "learning_rate": 9.420640352371375e-05, "loss": 3.0143, "step": 88500 }, { "epoch": 0.7819813136932832, "grad_norm": 2.411001205444336, "learning_rate": 9.419990890115155e-05, "loss": 3.4572, "step": 88550 }, { "epoch": 0.7824228615835673, "grad_norm": 3.345630168914795, "learning_rate": 9.419341086452118e-05, "loss": 2.9808, "step": 88600 }, { "epoch": 0.7828644094738515, "grad_norm": 1.7286512851715088, "learning_rate": 9.418690941432458e-05, "loss": 3.3402, "step": 88650 }, { "epoch": 0.7833059573641357, "grad_norm": 2.867637872695923, "learning_rate": 9.418040455106391e-05, "loss": 3.3708, "step": 88700 }, { "epoch": 0.7837475052544199, "grad_norm": 1.4467108249664307, "learning_rate": 9.417389627524163e-05, "loss": 3.4814, "step": 88750 }, { "epoch": 0.7841890531447041, "grad_norm": 1.6191805601119995, "learning_rate": 9.416738458736045e-05, "loss": 3.4584, "step": 88800 }, { "epoch": 0.7846306010349883, "grad_norm": 2.4342501163482666, "learning_rate": 9.416086948792333e-05, "loss": 3.4172, "step": 88850 }, { "epoch": 0.7850721489252724, "grad_norm": 1.3565874099731445, "learning_rate": 9.415435097743352e-05, "loss": 3.6409, "step": 88900 }, { "epoch": 0.7855136968155566, "grad_norm": 0.9413079619407654, "learning_rate": 9.414782905639452e-05, "loss": 3.1165, "step": 88950 }, { "epoch": 0.7859552447058408, "grad_norm": 2.325162887573242, "learning_rate": 9.414130372531009e-05, "loss": 3.3463, "step": 89000 }, { "epoch": 0.7863967925961249, "grad_norm": 2.1379945278167725, "learning_rate": 9.413477498468423e-05, "loss": 3.416, "step": 89050 }, { "epoch": 0.7868383404864091, "grad_norm": 1.78729248046875, "learning_rate": 9.412824283502129e-05, "loss": 3.6121, "step": 89100 }, { "epoch": 0.7872798883766934, "grad_norm": 1.3895008563995361, "learning_rate": 9.412170727682576e-05, "loss": 3.1819, "step": 89150 }, { "epoch": 0.7877214362669775, "grad_norm": 3.693272352218628, "learning_rate": 9.41151683106025e-05, "loss": 3.3451, "step": 89200 }, { "epoch": 0.7881629841572617, "grad_norm": 1.3732582330703735, "learning_rate": 9.410862593685657e-05, "loss": 3.0712, "step": 89250 }, { "epoch": 0.7886045320475459, "grad_norm": 1.9743282794952393, "learning_rate": 9.410208015609333e-05, "loss": 3.7555, "step": 89300 }, { "epoch": 0.78904607993783, "grad_norm": 2.194549083709717, "learning_rate": 9.409553096881835e-05, "loss": 2.9958, "step": 89350 }, { "epoch": 0.7894876278281142, "grad_norm": 1.8362387418746948, "learning_rate": 9.408897837553754e-05, "loss": 3.4235, "step": 89400 }, { "epoch": 0.7899291757183984, "grad_norm": 1.8394923210144043, "learning_rate": 9.4082422376757e-05, "loss": 3.3358, "step": 89450 }, { "epoch": 0.7903707236086825, "grad_norm": 2.9508590698242188, "learning_rate": 9.407586297298314e-05, "loss": 3.5187, "step": 89500 }, { "epoch": 0.7908122714989668, "grad_norm": 1.0827395915985107, "learning_rate": 9.406930016472262e-05, "loss": 2.985, "step": 89550 }, { "epoch": 0.791253819389251, "grad_norm": 6.141767978668213, "learning_rate": 9.406273395248236e-05, "loss": 3.3644, "step": 89600 }, { "epoch": 0.7916953672795352, "grad_norm": 2.3642897605895996, "learning_rate": 9.405616433676954e-05, "loss": 2.9457, "step": 89650 }, { "epoch": 0.7921369151698193, "grad_norm": 1.307497501373291, "learning_rate": 9.404959131809163e-05, "loss": 3.5927, "step": 89700 }, { "epoch": 0.7925784630601035, "grad_norm": 0.8792401552200317, "learning_rate": 9.404301489695629e-05, "loss": 3.4736, "step": 89750 }, { "epoch": 0.7930200109503877, "grad_norm": 1.8458104133605957, "learning_rate": 9.403643507387155e-05, "loss": 3.1718, "step": 89800 }, { "epoch": 0.7934615588406718, "grad_norm": 0.40211722254753113, "learning_rate": 9.402985184934561e-05, "loss": 3.1028, "step": 89850 }, { "epoch": 0.793903106730956, "grad_norm": 1.4072984457015991, "learning_rate": 9.402326522388695e-05, "loss": 3.2689, "step": 89900 }, { "epoch": 0.7943446546212403, "grad_norm": 3.634361505508423, "learning_rate": 9.401667519800439e-05, "loss": 3.2542, "step": 89950 }, { "epoch": 0.7947862025115244, "grad_norm": 2.3660759925842285, "learning_rate": 9.401008177220691e-05, "loss": 3.3141, "step": 90000 }, { "epoch": 0.7947862025115244, "eval_asr_loss": 0.913769633646809, "eval_loss": 2.972637891769409, "eval_runtime": 20.6823, "eval_samples_per_second": 37.133, "eval_steps_per_second": 9.283, "eval_tts_loss": 5.941250072060806, "step": 90000 }, { "epoch": 0.7952277504018086, "grad_norm": 1.8747526407241821, "learning_rate": 9.400348494700382e-05, "loss": 3.2776, "step": 90050 }, { "epoch": 0.7956692982920928, "grad_norm": 2.4819157123565674, "learning_rate": 9.399688472290464e-05, "loss": 3.3672, "step": 90100 }, { "epoch": 0.7961108461823769, "grad_norm": 3.1009058952331543, "learning_rate": 9.399028110041921e-05, "loss": 3.7072, "step": 90150 }, { "epoch": 0.7965523940726611, "grad_norm": 2.1817233562469482, "learning_rate": 9.39836740800576e-05, "loss": 3.1293, "step": 90200 }, { "epoch": 0.7969939419629453, "grad_norm": 2.2436323165893555, "learning_rate": 9.397706366233011e-05, "loss": 3.7612, "step": 90250 }, { "epoch": 0.7974354898532294, "grad_norm": 2.8587071895599365, "learning_rate": 9.397058215732492e-05, "loss": 3.574, "step": 90300 }, { "epoch": 0.7978770377435137, "grad_norm": 1.305193543434143, "learning_rate": 9.396396501431968e-05, "loss": 3.2321, "step": 90350 }, { "epoch": 0.7983185856337979, "grad_norm": 4.059189319610596, "learning_rate": 9.395734447547093e-05, "loss": 3.2663, "step": 90400 }, { "epoch": 0.798760133524082, "grad_norm": 1.4845237731933594, "learning_rate": 9.395072054129008e-05, "loss": 3.2744, "step": 90450 }, { "epoch": 0.7992016814143662, "grad_norm": 3.5800607204437256, "learning_rate": 9.394409321228877e-05, "loss": 3.3557, "step": 90500 }, { "epoch": 0.7996432293046504, "grad_norm": 1.589889645576477, "learning_rate": 9.393746248897888e-05, "loss": 3.1891, "step": 90550 }, { "epoch": 0.8000847771949345, "grad_norm": 2.2087574005126953, "learning_rate": 9.39308283718726e-05, "loss": 3.4591, "step": 90600 }, { "epoch": 0.8005263250852187, "grad_norm": 2.9878342151641846, "learning_rate": 9.392419086148234e-05, "loss": 3.1537, "step": 90650 }, { "epoch": 0.8009678729755029, "grad_norm": 4.756255626678467, "learning_rate": 9.391754995832083e-05, "loss": 3.1161, "step": 90700 }, { "epoch": 0.8014094208657871, "grad_norm": 3.2718210220336914, "learning_rate": 9.391090566290097e-05, "loss": 3.6481, "step": 90750 }, { "epoch": 0.8018509687560713, "grad_norm": 1.9818165302276611, "learning_rate": 9.3904257975736e-05, "loss": 3.0561, "step": 90800 }, { "epoch": 0.8022925166463555, "grad_norm": 0.7319821715354919, "learning_rate": 9.38976068973394e-05, "loss": 3.6211, "step": 90850 }, { "epoch": 0.8027340645366396, "grad_norm": 1.0141242742538452, "learning_rate": 9.389095242822492e-05, "loss": 3.3727, "step": 90900 }, { "epoch": 0.8031756124269238, "grad_norm": 2.862752676010132, "learning_rate": 9.388429456890654e-05, "loss": 3.2647, "step": 90950 }, { "epoch": 0.803617160317208, "grad_norm": 2.0796523094177246, "learning_rate": 9.387763331989854e-05, "loss": 2.9279, "step": 91000 }, { "epoch": 0.8040587082074921, "grad_norm": 3.659575939178467, "learning_rate": 9.387096868171543e-05, "loss": 3.3732, "step": 91050 }, { "epoch": 0.8045002560977763, "grad_norm": 1.8139934539794922, "learning_rate": 9.3864300654872e-05, "loss": 3.4054, "step": 91100 }, { "epoch": 0.8049418039880606, "grad_norm": 1.9845603704452515, "learning_rate": 9.385762923988332e-05, "loss": 3.2696, "step": 91150 }, { "epoch": 0.8053833518783448, "grad_norm": 1.5647777318954468, "learning_rate": 9.385095443726466e-05, "loss": 2.9517, "step": 91200 }, { "epoch": 0.8058248997686289, "grad_norm": 2.7620863914489746, "learning_rate": 9.384427624753164e-05, "loss": 3.3531, "step": 91250 }, { "epoch": 0.8062664476589131, "grad_norm": 4.5521039962768555, "learning_rate": 9.383759467120006e-05, "loss": 2.9471, "step": 91300 }, { "epoch": 0.8067079955491973, "grad_norm": 6.784761428833008, "learning_rate": 9.3830909708786e-05, "loss": 3.0502, "step": 91350 }, { "epoch": 0.8071495434394814, "grad_norm": 2.283494472503662, "learning_rate": 9.382422136080587e-05, "loss": 3.2579, "step": 91400 }, { "epoch": 0.8075910913297656, "grad_norm": 2.49485182762146, "learning_rate": 9.381752962777626e-05, "loss": 3.2323, "step": 91450 }, { "epoch": 0.8080326392200498, "grad_norm": 1.1890743970870972, "learning_rate": 9.381083451021402e-05, "loss": 3.3812, "step": 91500 }, { "epoch": 0.808474187110334, "grad_norm": 1.023622989654541, "learning_rate": 9.380413600863633e-05, "loss": 3.1125, "step": 91550 }, { "epoch": 0.8089157350006182, "grad_norm": 3.7999751567840576, "learning_rate": 9.37974341235606e-05, "loss": 3.0946, "step": 91600 }, { "epoch": 0.8093572828909024, "grad_norm": 2.7334210872650146, "learning_rate": 9.379072885550446e-05, "loss": 3.2808, "step": 91650 }, { "epoch": 0.8097988307811865, "grad_norm": 1.0871787071228027, "learning_rate": 9.378402020498585e-05, "loss": 2.9201, "step": 91700 }, { "epoch": 0.8102403786714707, "grad_norm": 1.837319254875183, "learning_rate": 9.377730817252296e-05, "loss": 3.3392, "step": 91750 }, { "epoch": 0.8106819265617549, "grad_norm": 1.857426643371582, "learning_rate": 9.377059275863423e-05, "loss": 3.4567, "step": 91800 }, { "epoch": 0.811123474452039, "grad_norm": 2.4178388118743896, "learning_rate": 9.37638739638384e-05, "loss": 3.1172, "step": 91850 }, { "epoch": 0.8115650223423232, "grad_norm": 1.7428170442581177, "learning_rate": 9.375715178865438e-05, "loss": 3.2541, "step": 91900 }, { "epoch": 0.8120065702326075, "grad_norm": 2.188565731048584, "learning_rate": 9.375042623360143e-05, "loss": 3.0557, "step": 91950 }, { "epoch": 0.8124481181228916, "grad_norm": 2.7001125812530518, "learning_rate": 9.374369729919907e-05, "loss": 3.4467, "step": 92000 }, { "epoch": 0.8128896660131758, "grad_norm": 0.5670623183250427, "learning_rate": 9.373696498596703e-05, "loss": 3.4051, "step": 92050 }, { "epoch": 0.81333121390346, "grad_norm": 0.8684957027435303, "learning_rate": 9.37302292944253e-05, "loss": 3.4903, "step": 92100 }, { "epoch": 0.8137727617937441, "grad_norm": 1.7560847997665405, "learning_rate": 9.372349022509419e-05, "loss": 2.8986, "step": 92150 }, { "epoch": 0.8142143096840283, "grad_norm": 1.301934838294983, "learning_rate": 9.371674777849423e-05, "loss": 3.2969, "step": 92200 }, { "epoch": 0.8146558575743125, "grad_norm": 5.415755748748779, "learning_rate": 9.37100019551462e-05, "loss": 3.5175, "step": 92250 }, { "epoch": 0.8150974054645966, "grad_norm": 1.403662085533142, "learning_rate": 9.370325275557119e-05, "loss": 3.3363, "step": 92300 }, { "epoch": 0.8155389533548809, "grad_norm": 0.34755173325538635, "learning_rate": 9.369650018029047e-05, "loss": 2.8427, "step": 92350 }, { "epoch": 0.8159805012451651, "grad_norm": 4.63238000869751, "learning_rate": 9.368974422982567e-05, "loss": 3.401, "step": 92400 }, { "epoch": 0.8164220491354492, "grad_norm": 5.1991286277771, "learning_rate": 9.36829849046986e-05, "loss": 3.0604, "step": 92450 }, { "epoch": 0.8168635970257334, "grad_norm": 4.037189960479736, "learning_rate": 9.367622220543136e-05, "loss": 3.4574, "step": 92500 }, { "epoch": 0.8173051449160176, "grad_norm": 3.1938400268554688, "learning_rate": 9.366945613254631e-05, "loss": 3.4916, "step": 92550 }, { "epoch": 0.8177466928063017, "grad_norm": 1.4730736017227173, "learning_rate": 9.36626866865661e-05, "loss": 3.5001, "step": 92600 }, { "epoch": 0.8181882406965859, "grad_norm": 2.499075174331665, "learning_rate": 9.36559138680136e-05, "loss": 3.3653, "step": 92650 }, { "epoch": 0.8186297885868701, "grad_norm": 3.067751884460449, "learning_rate": 9.364913767741192e-05, "loss": 2.9999, "step": 92700 }, { "epoch": 0.8190713364771544, "grad_norm": 3.1233701705932617, "learning_rate": 9.36423581152845e-05, "loss": 3.377, "step": 92750 }, { "epoch": 0.8195128843674385, "grad_norm": 1.1378720998764038, "learning_rate": 9.363557518215498e-05, "loss": 2.8108, "step": 92800 }, { "epoch": 0.8199544322577227, "grad_norm": 1.5843881368637085, "learning_rate": 9.362878887854732e-05, "loss": 3.4633, "step": 92850 }, { "epoch": 0.8203959801480069, "grad_norm": 3.084805965423584, "learning_rate": 9.362199920498566e-05, "loss": 3.3343, "step": 92900 }, { "epoch": 0.820837528038291, "grad_norm": 1.3744654655456543, "learning_rate": 9.361520616199446e-05, "loss": 2.9136, "step": 92950 }, { "epoch": 0.8212790759285752, "grad_norm": 3.1736974716186523, "learning_rate": 9.360840975009845e-05, "loss": 3.1212, "step": 93000 }, { "epoch": 0.8212790759285752, "eval_asr_loss": 0.9125425967281584, "eval_loss": 2.9801833629608154, "eval_runtime": 20.5623, "eval_samples_per_second": 37.35, "eval_steps_per_second": 9.337, "eval_tts_loss": 5.957149096143686, "step": 93000 }, { "epoch": 0.8217206238188594, "grad_norm": 2.6484904289245605, "learning_rate": 9.360160996982256e-05, "loss": 3.3256, "step": 93050 }, { "epoch": 0.8221621717091435, "grad_norm": 1.6820108890533447, "learning_rate": 9.359480682169204e-05, "loss": 3.3486, "step": 93100 }, { "epoch": 0.8226037195994278, "grad_norm": 1.407516360282898, "learning_rate": 9.358800030623236e-05, "loss": 3.4169, "step": 93150 }, { "epoch": 0.823045267489712, "grad_norm": 1.139156699180603, "learning_rate": 9.358119042396929e-05, "loss": 2.8844, "step": 93200 }, { "epoch": 0.8234868153799961, "grad_norm": 1.5392656326293945, "learning_rate": 9.35743771754288e-05, "loss": 3.5062, "step": 93250 }, { "epoch": 0.8239283632702803, "grad_norm": 1.2801883220672607, "learning_rate": 9.356756056113717e-05, "loss": 2.9979, "step": 93300 }, { "epoch": 0.8243699111605645, "grad_norm": 2.373271942138672, "learning_rate": 9.356074058162094e-05, "loss": 2.8897, "step": 93350 }, { "epoch": 0.8248114590508486, "grad_norm": 0.7981138229370117, "learning_rate": 9.355391723740687e-05, "loss": 3.7609, "step": 93400 }, { "epoch": 0.8252530069411328, "grad_norm": 1.8183437585830688, "learning_rate": 9.354709052902204e-05, "loss": 3.2145, "step": 93450 }, { "epoch": 0.825694554831417, "grad_norm": 1.9743393659591675, "learning_rate": 9.354026045699372e-05, "loss": 3.5426, "step": 93500 }, { "epoch": 0.8261361027217012, "grad_norm": 2.351931571960449, "learning_rate": 9.35334270218495e-05, "loss": 3.347, "step": 93550 }, { "epoch": 0.8265776506119854, "grad_norm": 1.7274210453033447, "learning_rate": 9.352659022411718e-05, "loss": 2.9723, "step": 93600 }, { "epoch": 0.8270191985022696, "grad_norm": 3.6531035900115967, "learning_rate": 9.351975006432487e-05, "loss": 3.311, "step": 93650 }, { "epoch": 0.8274607463925537, "grad_norm": 5.743092060089111, "learning_rate": 9.351290654300091e-05, "loss": 3.1698, "step": 93700 }, { "epoch": 0.8279022942828379, "grad_norm": 1.3943450450897217, "learning_rate": 9.35060596606739e-05, "loss": 3.0665, "step": 93750 }, { "epoch": 0.8283438421731221, "grad_norm": 3.4705088138580322, "learning_rate": 9.349920941787269e-05, "loss": 3.5611, "step": 93800 }, { "epoch": 0.8287853900634062, "grad_norm": 1.309612512588501, "learning_rate": 9.34923558151264e-05, "loss": 3.4221, "step": 93850 }, { "epoch": 0.8292269379536905, "grad_norm": 1.8064721822738647, "learning_rate": 9.348549885296446e-05, "loss": 3.564, "step": 93900 }, { "epoch": 0.8296684858439747, "grad_norm": 1.7501392364501953, "learning_rate": 9.347863853191646e-05, "loss": 3.0225, "step": 93950 }, { "epoch": 0.8301100337342588, "grad_norm": 3.4473793506622314, "learning_rate": 9.347177485251233e-05, "loss": 3.0757, "step": 94000 }, { "epoch": 0.830551581624543, "grad_norm": 1.1141574382781982, "learning_rate": 9.346490781528221e-05, "loss": 3.0982, "step": 94050 }, { "epoch": 0.8309931295148272, "grad_norm": 3.239313840866089, "learning_rate": 9.345803742075656e-05, "loss": 3.463, "step": 94100 }, { "epoch": 0.8314346774051113, "grad_norm": 2.90984845161438, "learning_rate": 9.345116366946601e-05, "loss": 3.131, "step": 94150 }, { "epoch": 0.8318762252953955, "grad_norm": 3.623865842819214, "learning_rate": 9.344428656194153e-05, "loss": 3.4718, "step": 94200 }, { "epoch": 0.8323177731856797, "grad_norm": 0.8482750058174133, "learning_rate": 9.343740609871431e-05, "loss": 3.3195, "step": 94250 }, { "epoch": 0.832759321075964, "grad_norm": 1.9041322469711304, "learning_rate": 9.343052228031582e-05, "loss": 3.3709, "step": 94300 }, { "epoch": 0.8332008689662481, "grad_norm": 3.6412575244903564, "learning_rate": 9.342363510727773e-05, "loss": 3.3433, "step": 94350 }, { "epoch": 0.8336424168565323, "grad_norm": 1.070917010307312, "learning_rate": 9.341674458013208e-05, "loss": 3.5517, "step": 94400 }, { "epoch": 0.8340839647468165, "grad_norm": 4.740517616271973, "learning_rate": 9.340985069941105e-05, "loss": 3.3758, "step": 94450 }, { "epoch": 0.8345255126371006, "grad_norm": 0.6648314595222473, "learning_rate": 9.340295346564719e-05, "loss": 3.5991, "step": 94500 }, { "epoch": 0.8349670605273848, "grad_norm": 3.5965664386749268, "learning_rate": 9.339605287937319e-05, "loss": 3.4054, "step": 94550 }, { "epoch": 0.835408608417669, "grad_norm": 2.838550329208374, "learning_rate": 9.33891489411221e-05, "loss": 3.0964, "step": 94600 }, { "epoch": 0.8358501563079531, "grad_norm": 1.3823540210723877, "learning_rate": 9.338224165142719e-05, "loss": 3.5835, "step": 94650 }, { "epoch": 0.8362917041982374, "grad_norm": 2.0661122798919678, "learning_rate": 9.337533101082199e-05, "loss": 3.2928, "step": 94700 }, { "epoch": 0.8367332520885216, "grad_norm": 1.8618370294570923, "learning_rate": 9.336841701984027e-05, "loss": 3.0778, "step": 94750 }, { "epoch": 0.8371747999788057, "grad_norm": 4.694985389709473, "learning_rate": 9.336149967901609e-05, "loss": 3.0738, "step": 94800 }, { "epoch": 0.8376163478690899, "grad_norm": 2.5283780097961426, "learning_rate": 9.335457898888376e-05, "loss": 3.4458, "step": 94850 }, { "epoch": 0.8380578957593741, "grad_norm": 1.0920939445495605, "learning_rate": 9.334765494997784e-05, "loss": 3.2661, "step": 94900 }, { "epoch": 0.8384994436496582, "grad_norm": 2.00156307220459, "learning_rate": 9.334072756283314e-05, "loss": 3.2389, "step": 94950 }, { "epoch": 0.8389409915399424, "grad_norm": 2.5333197116851807, "learning_rate": 9.333379682798475e-05, "loss": 3.5175, "step": 95000 }, { "epoch": 0.8393825394302266, "grad_norm": 2.970446825027466, "learning_rate": 9.332686274596803e-05, "loss": 3.0734, "step": 95050 }, { "epoch": 0.8398240873205108, "grad_norm": 1.0278738737106323, "learning_rate": 9.331992531731857e-05, "loss": 3.5306, "step": 95100 }, { "epoch": 0.840265635210795, "grad_norm": 2.5087051391601562, "learning_rate": 9.331298454257221e-05, "loss": 3.3883, "step": 95150 }, { "epoch": 0.8407071831010792, "grad_norm": 1.317252278327942, "learning_rate": 9.330604042226507e-05, "loss": 3.2772, "step": 95200 }, { "epoch": 0.8411487309913633, "grad_norm": 2.6759376525878906, "learning_rate": 9.329909295693355e-05, "loss": 2.9157, "step": 95250 }, { "epoch": 0.8415902788816475, "grad_norm": 1.9456520080566406, "learning_rate": 9.329214214711426e-05, "loss": 3.2917, "step": 95300 }, { "epoch": 0.8420318267719317, "grad_norm": 2.113647937774658, "learning_rate": 9.328532710918674e-05, "loss": 3.3203, "step": 95350 }, { "epoch": 0.8424733746622158, "grad_norm": 1.5438154935836792, "learning_rate": 9.327836967886584e-05, "loss": 3.6364, "step": 95400 }, { "epoch": 0.8429149225525, "grad_norm": 1.4090991020202637, "learning_rate": 9.32714089056579e-05, "loss": 3.2839, "step": 95450 }, { "epoch": 0.8433564704427843, "grad_norm": 3.7097344398498535, "learning_rate": 9.326444479010055e-05, "loss": 3.1992, "step": 95500 }, { "epoch": 0.8437980183330684, "grad_norm": 1.130540370941162, "learning_rate": 9.32574773327317e-05, "loss": 3.123, "step": 95550 }, { "epoch": 0.8442395662233526, "grad_norm": 3.1675918102264404, "learning_rate": 9.325050653408954e-05, "loss": 3.4446, "step": 95600 }, { "epoch": 0.8446811141136368, "grad_norm": 7.838589191436768, "learning_rate": 9.32435323947125e-05, "loss": 3.1731, "step": 95650 }, { "epoch": 0.8451226620039209, "grad_norm": 2.1061744689941406, "learning_rate": 9.32365549151393e-05, "loss": 3.1445, "step": 95700 }, { "epoch": 0.8455642098942051, "grad_norm": 1.585892677307129, "learning_rate": 9.322957409590884e-05, "loss": 3.248, "step": 95750 }, { "epoch": 0.8460057577844893, "grad_norm": 3.2491941452026367, "learning_rate": 9.322258993756037e-05, "loss": 3.1858, "step": 95800 }, { "epoch": 0.8464473056747734, "grad_norm": 4.585638999938965, "learning_rate": 9.321560244063334e-05, "loss": 3.3065, "step": 95850 }, { "epoch": 0.8468888535650577, "grad_norm": 1.343159556388855, "learning_rate": 9.320861160566748e-05, "loss": 3.3079, "step": 95900 }, { "epoch": 0.8473304014553419, "grad_norm": 4.3721699714660645, "learning_rate": 9.320161743320277e-05, "loss": 3.1931, "step": 95950 }, { "epoch": 0.847771949345626, "grad_norm": 1.243537425994873, "learning_rate": 9.319461992377945e-05, "loss": 3.2813, "step": 96000 }, { "epoch": 0.847771949345626, "eval_asr_loss": 0.9113168214024796, "eval_loss": 2.9808547496795654, "eval_runtime": 20.2448, "eval_samples_per_second": 37.936, "eval_steps_per_second": 9.484, "eval_tts_loss": 5.98077398558446, "step": 96000 }, { "epoch": 0.8482134972359102, "grad_norm": 3.9162681102752686, "learning_rate": 9.318761907793804e-05, "loss": 3.2729, "step": 96050 }, { "epoch": 0.8486550451261944, "grad_norm": 0.9850695729255676, "learning_rate": 9.318061489621925e-05, "loss": 3.4972, "step": 96100 }, { "epoch": 0.8490965930164786, "grad_norm": 2.3384222984313965, "learning_rate": 9.317360737916415e-05, "loss": 3.5748, "step": 96150 }, { "epoch": 0.8495381409067627, "grad_norm": 1.6895709037780762, "learning_rate": 9.316659652731395e-05, "loss": 3.5112, "step": 96200 }, { "epoch": 0.8499796887970469, "grad_norm": 4.779604434967041, "learning_rate": 9.315958234121023e-05, "loss": 3.4245, "step": 96250 }, { "epoch": 0.8504212366873312, "grad_norm": 3.7872555255889893, "learning_rate": 9.315256482139475e-05, "loss": 3.4721, "step": 96300 }, { "epoch": 0.8508627845776153, "grad_norm": 1.2569001913070679, "learning_rate": 9.314554396840956e-05, "loss": 3.2695, "step": 96350 }, { "epoch": 0.8513043324678995, "grad_norm": 2.095500946044922, "learning_rate": 9.313851978279697e-05, "loss": 3.1881, "step": 96400 }, { "epoch": 0.8517458803581837, "grad_norm": 0.9828936457633972, "learning_rate": 9.313149226509952e-05, "loss": 3.0915, "step": 96450 }, { "epoch": 0.8521874282484678, "grad_norm": 2.886770725250244, "learning_rate": 9.312446141586004e-05, "loss": 3.4099, "step": 96500 }, { "epoch": 0.852628976138752, "grad_norm": 4.082062721252441, "learning_rate": 9.31174272356216e-05, "loss": 3.0428, "step": 96550 }, { "epoch": 0.8530705240290362, "grad_norm": 1.261383295059204, "learning_rate": 9.311038972492754e-05, "loss": 3.2126, "step": 96600 }, { "epoch": 0.8535120719193203, "grad_norm": 1.2416714429855347, "learning_rate": 9.310334888432142e-05, "loss": 3.2021, "step": 96650 }, { "epoch": 0.8539536198096046, "grad_norm": 3.1885976791381836, "learning_rate": 9.309630471434712e-05, "loss": 2.7708, "step": 96700 }, { "epoch": 0.8543951676998888, "grad_norm": 0.710462749004364, "learning_rate": 9.308925721554874e-05, "loss": 3.199, "step": 96750 }, { "epoch": 0.8548367155901729, "grad_norm": 5.235406875610352, "learning_rate": 9.308220638847059e-05, "loss": 3.2963, "step": 96800 }, { "epoch": 0.8552782634804571, "grad_norm": 1.9540880918502808, "learning_rate": 9.307515223365735e-05, "loss": 3.3895, "step": 96850 }, { "epoch": 0.8557198113707413, "grad_norm": 2.796398162841797, "learning_rate": 9.306809475165385e-05, "loss": 3.4795, "step": 96900 }, { "epoch": 0.8561613592610254, "grad_norm": 1.8404650688171387, "learning_rate": 9.306103394300525e-05, "loss": 3.2471, "step": 96950 }, { "epoch": 0.8566029071513096, "grad_norm": 1.5861579179763794, "learning_rate": 9.305396980825693e-05, "loss": 2.8261, "step": 97000 }, { "epoch": 0.8570444550415938, "grad_norm": 5.508571624755859, "learning_rate": 9.304690234795453e-05, "loss": 3.0444, "step": 97050 }, { "epoch": 0.857486002931878, "grad_norm": 3.7076914310455322, "learning_rate": 9.303983156264393e-05, "loss": 3.3217, "step": 97100 }, { "epoch": 0.8579275508221622, "grad_norm": 4.032550811767578, "learning_rate": 9.303275745287133e-05, "loss": 3.3205, "step": 97150 }, { "epoch": 0.8583690987124464, "grad_norm": 0.970733106136322, "learning_rate": 9.30256800191831e-05, "loss": 3.1487, "step": 97200 }, { "epoch": 0.8588106466027305, "grad_norm": 2.3852109909057617, "learning_rate": 9.301859926212595e-05, "loss": 3.1561, "step": 97250 }, { "epoch": 0.8592521944930147, "grad_norm": 1.1150777339935303, "learning_rate": 9.301151518224679e-05, "loss": 2.9001, "step": 97300 }, { "epoch": 0.8596937423832989, "grad_norm": 3.74100923538208, "learning_rate": 9.300442778009282e-05, "loss": 3.2136, "step": 97350 }, { "epoch": 0.860135290273583, "grad_norm": 2.762349843978882, "learning_rate": 9.299733705621145e-05, "loss": 3.4975, "step": 97400 }, { "epoch": 0.8605768381638672, "grad_norm": 0.8698297142982483, "learning_rate": 9.299024301115043e-05, "loss": 3.3572, "step": 97450 }, { "epoch": 0.8610183860541515, "grad_norm": 3.2828052043914795, "learning_rate": 9.298314564545765e-05, "loss": 3.3767, "step": 97500 }, { "epoch": 0.8614599339444357, "grad_norm": 2.659813642501831, "learning_rate": 9.297604495968138e-05, "loss": 3.5504, "step": 97550 }, { "epoch": 0.8619014818347198, "grad_norm": 1.7123510837554932, "learning_rate": 9.296894095437005e-05, "loss": 3.4038, "step": 97600 }, { "epoch": 0.862343029725004, "grad_norm": 2.9085240364074707, "learning_rate": 9.296183363007241e-05, "loss": 2.8643, "step": 97650 }, { "epoch": 0.8627845776152882, "grad_norm": 1.4961036443710327, "learning_rate": 9.295472298733742e-05, "loss": 3.1688, "step": 97700 }, { "epoch": 0.8632261255055723, "grad_norm": 2.028132677078247, "learning_rate": 9.294760902671432e-05, "loss": 3.3077, "step": 97750 }, { "epoch": 0.8636676733958565, "grad_norm": 1.637831449508667, "learning_rate": 9.294049174875261e-05, "loss": 2.9796, "step": 97800 }, { "epoch": 0.8641092212861406, "grad_norm": 1.5504677295684814, "learning_rate": 9.293337115400205e-05, "loss": 2.8288, "step": 97850 }, { "epoch": 0.8645507691764249, "grad_norm": 3.117736577987671, "learning_rate": 9.292638975372799e-05, "loss": 2.9301, "step": 97900 }, { "epoch": 0.8649923170667091, "grad_norm": 2.2668349742889404, "learning_rate": 9.291926259335835e-05, "loss": 3.8076, "step": 97950 }, { "epoch": 0.8654338649569933, "grad_norm": 3.2014098167419434, "learning_rate": 9.29121321178396e-05, "loss": 3.1708, "step": 98000 }, { "epoch": 0.8658754128472774, "grad_norm": 3.2920122146606445, "learning_rate": 9.290499832772255e-05, "loss": 3.3037, "step": 98050 }, { "epoch": 0.8663169607375616, "grad_norm": 1.8121235370635986, "learning_rate": 9.28978612235582e-05, "loss": 3.2196, "step": 98100 }, { "epoch": 0.8667585086278458, "grad_norm": 1.6291028261184692, "learning_rate": 9.289072080589783e-05, "loss": 3.2285, "step": 98150 }, { "epoch": 0.8672000565181299, "grad_norm": 1.6624693870544434, "learning_rate": 9.288357707529297e-05, "loss": 3.1948, "step": 98200 }, { "epoch": 0.8676416044084141, "grad_norm": 2.4642486572265625, "learning_rate": 9.287643003229543e-05, "loss": 3.0989, "step": 98250 }, { "epoch": 0.8680831522986984, "grad_norm": 2.3431122303009033, "learning_rate": 9.286927967745725e-05, "loss": 3.0297, "step": 98300 }, { "epoch": 0.8685247001889825, "grad_norm": 1.7271332740783691, "learning_rate": 9.286212601133076e-05, "loss": 3.7604, "step": 98350 }, { "epoch": 0.8689662480792667, "grad_norm": 2.332041025161743, "learning_rate": 9.285496903446846e-05, "loss": 3.4561, "step": 98400 }, { "epoch": 0.8694077959695509, "grad_norm": 1.6665946245193481, "learning_rate": 9.284780874742321e-05, "loss": 3.2693, "step": 98450 }, { "epoch": 0.869849343859835, "grad_norm": 1.6122417449951172, "learning_rate": 9.284064515074809e-05, "loss": 3.2516, "step": 98500 }, { "epoch": 0.8702908917501192, "grad_norm": 0.9736064672470093, "learning_rate": 9.283347824499639e-05, "loss": 3.3643, "step": 98550 }, { "epoch": 0.8707324396404034, "grad_norm": 2.6523666381835938, "learning_rate": 9.282630803072173e-05, "loss": 3.4379, "step": 98600 }, { "epoch": 0.8711739875306875, "grad_norm": 3.84800386428833, "learning_rate": 9.281913450847792e-05, "loss": 3.6332, "step": 98650 }, { "epoch": 0.8716155354209718, "grad_norm": 1.0476601123809814, "learning_rate": 9.281195767881908e-05, "loss": 3.1238, "step": 98700 }, { "epoch": 0.872057083311256, "grad_norm": 2.119279623031616, "learning_rate": 9.280477754229952e-05, "loss": 3.4006, "step": 98750 }, { "epoch": 0.8724986312015401, "grad_norm": 1.7416213750839233, "learning_rate": 9.279759409947388e-05, "loss": 3.5637, "step": 98800 }, { "epoch": 0.8729401790918243, "grad_norm": 4.38261079788208, "learning_rate": 9.279040735089702e-05, "loss": 3.233, "step": 98850 }, { "epoch": 0.8733817269821085, "grad_norm": 4.000781059265137, "learning_rate": 9.278321729712403e-05, "loss": 3.5001, "step": 98900 }, { "epoch": 0.8738232748723926, "grad_norm": 2.074897050857544, "learning_rate": 9.27760239387103e-05, "loss": 3.0254, "step": 98950 }, { "epoch": 0.8742648227626768, "grad_norm": 3.6822054386138916, "learning_rate": 9.276882727621146e-05, "loss": 3.2075, "step": 99000 }, { "epoch": 0.8742648227626768, "eval_asr_loss": 0.9122765446783635, "eval_loss": 2.9597482681274414, "eval_runtime": 21.0264, "eval_samples_per_second": 36.525, "eval_steps_per_second": 9.131, "eval_tts_loss": 5.931117558903506, "step": 99000 }, { "epoch": 0.874706370652961, "grad_norm": 2.583610773086548, "learning_rate": 9.276162731018336e-05, "loss": 2.9525, "step": 99050 }, { "epoch": 0.8751479185432453, "grad_norm": 3.8291828632354736, "learning_rate": 9.275442404118217e-05, "loss": 3.5362, "step": 99100 }, { "epoch": 0.8755894664335294, "grad_norm": 1.9839342832565308, "learning_rate": 9.274721746976427e-05, "loss": 3.168, "step": 99150 }, { "epoch": 0.8760310143238136, "grad_norm": 2.494183301925659, "learning_rate": 9.274000759648632e-05, "loss": 3.0813, "step": 99200 }, { "epoch": 0.8764725622140978, "grad_norm": 1.9744882583618164, "learning_rate": 9.273279442190519e-05, "loss": 3.3301, "step": 99250 }, { "epoch": 0.8769141101043819, "grad_norm": 1.0826927423477173, "learning_rate": 9.272557794657805e-05, "loss": 3.2976, "step": 99300 }, { "epoch": 0.8773556579946661, "grad_norm": 3.1494545936584473, "learning_rate": 9.271835817106234e-05, "loss": 3.2733, "step": 99350 }, { "epoch": 0.8777972058849502, "grad_norm": 2.1785781383514404, "learning_rate": 9.271113509591568e-05, "loss": 3.1884, "step": 99400 }, { "epoch": 0.8782387537752344, "grad_norm": 1.9323933124542236, "learning_rate": 9.270390872169602e-05, "loss": 3.4076, "step": 99450 }, { "epoch": 0.8786803016655187, "grad_norm": 2.127812385559082, "learning_rate": 9.269667904896153e-05, "loss": 3.3259, "step": 99500 }, { "epoch": 0.8791218495558029, "grad_norm": 2.901846170425415, "learning_rate": 9.268944607827064e-05, "loss": 3.878, "step": 99550 }, { "epoch": 0.879563397446087, "grad_norm": 2.7331161499023438, "learning_rate": 9.268220981018204e-05, "loss": 3.0213, "step": 99600 }, { "epoch": 0.8800049453363712, "grad_norm": 1.2581075429916382, "learning_rate": 9.267497024525466e-05, "loss": 3.0469, "step": 99650 }, { "epoch": 0.8804464932266554, "grad_norm": 1.963077187538147, "learning_rate": 9.266772738404771e-05, "loss": 3.0544, "step": 99700 }, { "epoch": 0.8808880411169395, "grad_norm": 2.728001594543457, "learning_rate": 9.266048122712064e-05, "loss": 3.0793, "step": 99750 }, { "epoch": 0.8813295890072237, "grad_norm": 2.3829917907714844, "learning_rate": 9.265323177503312e-05, "loss": 3.2839, "step": 99800 }, { "epoch": 0.8817711368975079, "grad_norm": 0.8417502045631409, "learning_rate": 9.264597902834515e-05, "loss": 3.1644, "step": 99850 }, { "epoch": 0.8822126847877921, "grad_norm": 3.0022170543670654, "learning_rate": 9.263872298761693e-05, "loss": 3.4081, "step": 99900 }, { "epoch": 0.8826542326780763, "grad_norm": 5.986268043518066, "learning_rate": 9.263146365340891e-05, "loss": 3.1996, "step": 99950 }, { "epoch": 0.8830957805683605, "grad_norm": 2.8962159156799316, "learning_rate": 9.262420102628184e-05, "loss": 3.1852, "step": 100000 }, { "epoch": 0.8835373284586446, "grad_norm": 2.588916540145874, "learning_rate": 9.261693510679668e-05, "loss": 3.1378, "step": 100050 }, { "epoch": 0.8839788763489288, "grad_norm": 2.7760961055755615, "learning_rate": 9.260966589551468e-05, "loss": 3.2112, "step": 100100 }, { "epoch": 0.884420424239213, "grad_norm": 1.5340514183044434, "learning_rate": 9.260239339299727e-05, "loss": 3.5164, "step": 100150 }, { "epoch": 0.8848619721294971, "grad_norm": 9.54226303100586, "learning_rate": 9.259511759980625e-05, "loss": 3.5849, "step": 100200 }, { "epoch": 0.8853035200197813, "grad_norm": 1.6671772003173828, "learning_rate": 9.258783851650361e-05, "loss": 3.4115, "step": 100250 }, { "epoch": 0.8857450679100656, "grad_norm": 1.8623788356781006, "learning_rate": 9.258055614365156e-05, "loss": 3.3499, "step": 100300 }, { "epoch": 0.8861866158003497, "grad_norm": 1.5677406787872314, "learning_rate": 9.257327048181262e-05, "loss": 3.2248, "step": 100350 }, { "epoch": 0.8866281636906339, "grad_norm": 1.939414381980896, "learning_rate": 9.256598153154954e-05, "loss": 3.419, "step": 100400 }, { "epoch": 0.8870697115809181, "grad_norm": 2.109490394592285, "learning_rate": 9.255883517040522e-05, "loss": 3.5276, "step": 100450 }, { "epoch": 0.8875112594712022, "grad_norm": 3.756086826324463, "learning_rate": 9.255153971072361e-05, "loss": 3.6449, "step": 100500 }, { "epoch": 0.8879528073614864, "grad_norm": 3.2874066829681396, "learning_rate": 9.254424096429637e-05, "loss": 3.452, "step": 100550 }, { "epoch": 0.8883943552517706, "grad_norm": 1.6661789417266846, "learning_rate": 9.253693893168729e-05, "loss": 3.2744, "step": 100600 }, { "epoch": 0.8888359031420547, "grad_norm": 3.6185531616210938, "learning_rate": 9.252963361346036e-05, "loss": 2.9505, "step": 100650 }, { "epoch": 0.889277451032339, "grad_norm": 5.086017608642578, "learning_rate": 9.25223250101799e-05, "loss": 2.9579, "step": 100700 }, { "epoch": 0.8897189989226232, "grad_norm": 3.7332069873809814, "learning_rate": 9.25150131224104e-05, "loss": 3.1295, "step": 100750 }, { "epoch": 0.8901605468129073, "grad_norm": 3.6353108882904053, "learning_rate": 9.250769795071666e-05, "loss": 3.6284, "step": 100800 }, { "epoch": 0.8906020947031915, "grad_norm": 6.214040279388428, "learning_rate": 9.250037949566368e-05, "loss": 2.6618, "step": 100850 }, { "epoch": 0.8910436425934757, "grad_norm": 3.569542169570923, "learning_rate": 9.249305775781681e-05, "loss": 3.5992, "step": 100900 }, { "epoch": 0.8914851904837598, "grad_norm": 2.4265153408050537, "learning_rate": 9.248573273774154e-05, "loss": 3.8598, "step": 100950 }, { "epoch": 0.891926738374044, "grad_norm": 2.2118725776672363, "learning_rate": 9.247840443600368e-05, "loss": 2.7226, "step": 101000 }, { "epoch": 0.8923682862643282, "grad_norm": 1.527198076248169, "learning_rate": 9.24710728531693e-05, "loss": 3.3547, "step": 101050 }, { "epoch": 0.8928098341546125, "grad_norm": 2.138624668121338, "learning_rate": 9.246373798980468e-05, "loss": 3.5805, "step": 101100 }, { "epoch": 0.8932513820448966, "grad_norm": 3.642125129699707, "learning_rate": 9.245639984647639e-05, "loss": 3.2457, "step": 101150 }, { "epoch": 0.8936929299351808, "grad_norm": 1.057465672492981, "learning_rate": 9.244905842375122e-05, "loss": 3.2628, "step": 101200 }, { "epoch": 0.894134477825465, "grad_norm": 1.685719609260559, "learning_rate": 9.244171372219626e-05, "loss": 3.8842, "step": 101250 }, { "epoch": 0.8945760257157491, "grad_norm": 1.3749964237213135, "learning_rate": 9.243436574237878e-05, "loss": 3.0899, "step": 101300 }, { "epoch": 0.8950175736060333, "grad_norm": 2.584540367126465, "learning_rate": 9.24270144848664e-05, "loss": 3.6703, "step": 101350 }, { "epoch": 0.8954591214963175, "grad_norm": 3.092801570892334, "learning_rate": 9.241965995022692e-05, "loss": 3.1755, "step": 101400 }, { "epoch": 0.8959006693866016, "grad_norm": 1.243997573852539, "learning_rate": 9.241230213902842e-05, "loss": 2.8785, "step": 101450 }, { "epoch": 0.8963422172768859, "grad_norm": 4.314537048339844, "learning_rate": 9.240494105183921e-05, "loss": 3.3101, "step": 101500 }, { "epoch": 0.8967837651671701, "grad_norm": 2.9184675216674805, "learning_rate": 9.239757668922791e-05, "loss": 3.318, "step": 101550 }, { "epoch": 0.8972253130574542, "grad_norm": 2.036470890045166, "learning_rate": 9.239020905176331e-05, "loss": 3.1143, "step": 101600 }, { "epoch": 0.8976668609477384, "grad_norm": 4.395949363708496, "learning_rate": 9.238283814001454e-05, "loss": 3.2825, "step": 101650 }, { "epoch": 0.8981084088380226, "grad_norm": 2.0808677673339844, "learning_rate": 9.23754639545509e-05, "loss": 3.7597, "step": 101700 }, { "epoch": 0.8985499567283067, "grad_norm": 1.0310138463974, "learning_rate": 9.236808649594202e-05, "loss": 3.3764, "step": 101750 }, { "epoch": 0.8989915046185909, "grad_norm": 1.490734338760376, "learning_rate": 9.236070576475773e-05, "loss": 3.3401, "step": 101800 }, { "epoch": 0.8994330525088751, "grad_norm": 1.4883259534835815, "learning_rate": 9.235332176156812e-05, "loss": 3.3028, "step": 101850 }, { "epoch": 0.8998746003991593, "grad_norm": 1.0461546182632446, "learning_rate": 9.234593448694354e-05, "loss": 3.007, "step": 101900 }, { "epoch": 0.9003161482894435, "grad_norm": 1.0872769355773926, "learning_rate": 9.233854394145462e-05, "loss": 3.1332, "step": 101950 }, { "epoch": 0.9007576961797277, "grad_norm": 4.982665061950684, "learning_rate": 9.23311501256722e-05, "loss": 3.3656, "step": 102000 }, { "epoch": 0.9007576961797277, "eval_asr_loss": 0.903455578837743, "eval_loss": 2.968921661376953, "eval_runtime": 20.4043, "eval_samples_per_second": 37.639, "eval_steps_per_second": 9.41, "eval_tts_loss": 5.955756151524818, "step": 102000 }, { "epoch": 0.9011992440700118, "grad_norm": 1.7485255002975464, "learning_rate": 9.232375304016738e-05, "loss": 3.2068, "step": 102050 }, { "epoch": 0.901640791960296, "grad_norm": 2.655881643295288, "learning_rate": 9.231635268551153e-05, "loss": 3.2794, "step": 102100 }, { "epoch": 0.9020823398505802, "grad_norm": 1.1869397163391113, "learning_rate": 9.230894906227628e-05, "loss": 3.6545, "step": 102150 }, { "epoch": 0.9025238877408643, "grad_norm": 1.75388765335083, "learning_rate": 9.23016903408811e-05, "loss": 3.4393, "step": 102200 }, { "epoch": 0.9029654356311485, "grad_norm": 2.8003134727478027, "learning_rate": 9.229428024754599e-05, "loss": 3.6308, "step": 102250 }, { "epoch": 0.9034069835214328, "grad_norm": 2.303340196609497, "learning_rate": 9.228686688733635e-05, "loss": 3.4285, "step": 102300 }, { "epoch": 0.903848531411717, "grad_norm": 1.432195782661438, "learning_rate": 9.227945026082483e-05, "loss": 3.0429, "step": 102350 }, { "epoch": 0.9042900793020011, "grad_norm": 3.9534683227539062, "learning_rate": 9.227203036858429e-05, "loss": 3.385, "step": 102400 }, { "epoch": 0.9047316271922853, "grad_norm": 1.3392046689987183, "learning_rate": 9.226460721118788e-05, "loss": 2.9859, "step": 102450 }, { "epoch": 0.9051731750825694, "grad_norm": 1.7585742473602295, "learning_rate": 9.225718078920896e-05, "loss": 2.8935, "step": 102500 }, { "epoch": 0.9056147229728536, "grad_norm": 1.3776382207870483, "learning_rate": 9.224975110322113e-05, "loss": 3.2253, "step": 102550 }, { "epoch": 0.9060562708631378, "grad_norm": 0.8360838294029236, "learning_rate": 9.22423181537983e-05, "loss": 2.9125, "step": 102600 }, { "epoch": 0.906497818753422, "grad_norm": 2.6101293563842773, "learning_rate": 9.223488194151461e-05, "loss": 3.0466, "step": 102650 }, { "epoch": 0.9069393666437062, "grad_norm": 2.8164517879486084, "learning_rate": 9.222744246694441e-05, "loss": 3.3639, "step": 102700 }, { "epoch": 0.9073809145339904, "grad_norm": 2.205768585205078, "learning_rate": 9.221999973066238e-05, "loss": 3.3834, "step": 102750 }, { "epoch": 0.9078224624242746, "grad_norm": 3.8350329399108887, "learning_rate": 9.221255373324338e-05, "loss": 2.7993, "step": 102800 }, { "epoch": 0.9082640103145587, "grad_norm": 3.8770978450775146, "learning_rate": 9.220510447526256e-05, "loss": 2.9693, "step": 102850 }, { "epoch": 0.9087055582048429, "grad_norm": 2.760380983352661, "learning_rate": 9.219765195729529e-05, "loss": 3.5722, "step": 102900 }, { "epoch": 0.9091471060951271, "grad_norm": 3.3382771015167236, "learning_rate": 9.219019617991724e-05, "loss": 3.5588, "step": 102950 }, { "epoch": 0.9095886539854112, "grad_norm": 2.2112584114074707, "learning_rate": 9.218273714370432e-05, "loss": 3.3712, "step": 103000 }, { "epoch": 0.9100302018756954, "grad_norm": 2.458007335662842, "learning_rate": 9.217527484923262e-05, "loss": 2.5991, "step": 103050 }, { "epoch": 0.9104717497659797, "grad_norm": 2.0607690811157227, "learning_rate": 9.21678092970786e-05, "loss": 3.4419, "step": 103100 }, { "epoch": 0.9109132976562638, "grad_norm": 3.3169896602630615, "learning_rate": 9.216034048781887e-05, "loss": 2.8659, "step": 103150 }, { "epoch": 0.911354845546548, "grad_norm": 2.7883317470550537, "learning_rate": 9.215286842203035e-05, "loss": 3.6627, "step": 103200 }, { "epoch": 0.9117963934368322, "grad_norm": 1.9193239212036133, "learning_rate": 9.21453931002902e-05, "loss": 3.2209, "step": 103250 }, { "epoch": 0.9122379413271163, "grad_norm": 1.5039438009262085, "learning_rate": 9.213791452317579e-05, "loss": 3.2848, "step": 103300 }, { "epoch": 0.9126794892174005, "grad_norm": 1.5526306629180908, "learning_rate": 9.213043269126483e-05, "loss": 3.0704, "step": 103350 }, { "epoch": 0.9131210371076847, "grad_norm": 1.0206029415130615, "learning_rate": 9.212294760513518e-05, "loss": 3.675, "step": 103400 }, { "epoch": 0.9135625849979688, "grad_norm": 4.434154510498047, "learning_rate": 9.211545926536504e-05, "loss": 3.4542, "step": 103450 }, { "epoch": 0.9140041328882531, "grad_norm": 2.1419999599456787, "learning_rate": 9.210796767253278e-05, "loss": 3.8326, "step": 103500 }, { "epoch": 0.9144456807785373, "grad_norm": 0.9824464321136475, "learning_rate": 9.210047282721709e-05, "loss": 3.2751, "step": 103550 }, { "epoch": 0.9148872286688214, "grad_norm": 3.2531392574310303, "learning_rate": 9.20929747299969e-05, "loss": 3.1043, "step": 103600 }, { "epoch": 0.9153287765591056, "grad_norm": 7.856770992279053, "learning_rate": 9.208547338145132e-05, "loss": 3.0719, "step": 103650 }, { "epoch": 0.9157703244493898, "grad_norm": 0.7467933893203735, "learning_rate": 9.207796878215979e-05, "loss": 3.6067, "step": 103700 }, { "epoch": 0.9162118723396739, "grad_norm": 11.154623985290527, "learning_rate": 9.2070460932702e-05, "loss": 3.2332, "step": 103750 }, { "epoch": 0.9166534202299581, "grad_norm": 4.158257484436035, "learning_rate": 9.206294983365785e-05, "loss": 3.4292, "step": 103800 }, { "epoch": 0.9170949681202423, "grad_norm": 1.339894413948059, "learning_rate": 9.205543548560751e-05, "loss": 3.1509, "step": 103850 }, { "epoch": 0.9175365160105265, "grad_norm": 1.5442434549331665, "learning_rate": 9.204791788913138e-05, "loss": 3.0922, "step": 103900 }, { "epoch": 0.9179780639008107, "grad_norm": 1.0537227392196655, "learning_rate": 9.204039704481018e-05, "loss": 3.2525, "step": 103950 }, { "epoch": 0.9184196117910949, "grad_norm": 2.2441678047180176, "learning_rate": 9.203287295322478e-05, "loss": 2.9392, "step": 104000 }, { "epoch": 0.918861159681379, "grad_norm": 3.044712543487549, "learning_rate": 9.202534561495637e-05, "loss": 3.0516, "step": 104050 }, { "epoch": 0.9193027075716632, "grad_norm": 1.7708096504211426, "learning_rate": 9.201781503058638e-05, "loss": 3.3234, "step": 104100 }, { "epoch": 0.9197442554619474, "grad_norm": 2.0960426330566406, "learning_rate": 9.201028120069648e-05, "loss": 3.808, "step": 104150 }, { "epoch": 0.9201858033522315, "grad_norm": 6.793859958648682, "learning_rate": 9.20027441258686e-05, "loss": 3.6451, "step": 104200 }, { "epoch": 0.9206273512425157, "grad_norm": 2.1247751712799072, "learning_rate": 9.19952038066849e-05, "loss": 3.6018, "step": 104250 }, { "epoch": 0.9210688991328, "grad_norm": 1.15399169921875, "learning_rate": 9.198766024372783e-05, "loss": 3.3271, "step": 104300 }, { "epoch": 0.9215104470230842, "grad_norm": 1.7735700607299805, "learning_rate": 9.198011343758002e-05, "loss": 3.0006, "step": 104350 }, { "epoch": 0.9219519949133683, "grad_norm": 0.8139469027519226, "learning_rate": 9.197256338882445e-05, "loss": 3.0361, "step": 104400 }, { "epoch": 0.9223935428036525, "grad_norm": 3.3219869136810303, "learning_rate": 9.196501009804429e-05, "loss": 3.1615, "step": 104450 }, { "epoch": 0.9228350906939367, "grad_norm": 2.1719770431518555, "learning_rate": 9.195745356582292e-05, "loss": 3.8578, "step": 104500 }, { "epoch": 0.9232766385842208, "grad_norm": 1.2096954584121704, "learning_rate": 9.194989379274406e-05, "loss": 3.1797, "step": 104550 }, { "epoch": 0.923718186474505, "grad_norm": 1.369635820388794, "learning_rate": 9.194233077939165e-05, "loss": 3.1194, "step": 104600 }, { "epoch": 0.9241597343647892, "grad_norm": 1.7169206142425537, "learning_rate": 9.193476452634983e-05, "loss": 3.3546, "step": 104650 }, { "epoch": 0.9246012822550734, "grad_norm": 5.232157230377197, "learning_rate": 9.192719503420304e-05, "loss": 3.2173, "step": 104700 }, { "epoch": 0.9250428301453576, "grad_norm": 1.7149889469146729, "learning_rate": 9.191962230353598e-05, "loss": 3.0774, "step": 104750 }, { "epoch": 0.9254843780356418, "grad_norm": 1.5316158533096313, "learning_rate": 9.191204633493355e-05, "loss": 3.6406, "step": 104800 }, { "epoch": 0.9259259259259259, "grad_norm": 2.05110239982605, "learning_rate": 9.190446712898097e-05, "loss": 3.3836, "step": 104850 }, { "epoch": 0.9263674738162101, "grad_norm": 3.319016695022583, "learning_rate": 9.189688468626363e-05, "loss": 3.1945, "step": 104900 }, { "epoch": 0.9268090217064943, "grad_norm": 2.672206163406372, "learning_rate": 9.188929900736722e-05, "loss": 3.5687, "step": 104950 }, { "epoch": 0.9272505695967784, "grad_norm": 1.4646509885787964, "learning_rate": 9.188171009287767e-05, "loss": 3.329, "step": 105000 }, { "epoch": 0.9272505695967784, "eval_asr_loss": 0.9064391918799752, "eval_loss": 2.951792001724243, "eval_runtime": 20.8436, "eval_samples_per_second": 36.846, "eval_steps_per_second": 9.211, "eval_tts_loss": 5.888024541957233, "step": 105000 }, { "epoch": 0.9276921174870626, "grad_norm": 8.195231437683105, "learning_rate": 9.187411794338117e-05, "loss": 3.1485, "step": 105050 }, { "epoch": 0.9281336653773469, "grad_norm": 2.3274195194244385, "learning_rate": 9.186652255946413e-05, "loss": 3.4639, "step": 105100 }, { "epoch": 0.928575213267631, "grad_norm": 2.3561298847198486, "learning_rate": 9.185892394171325e-05, "loss": 2.9877, "step": 105150 }, { "epoch": 0.9290167611579152, "grad_norm": 2.7232580184936523, "learning_rate": 9.185132209071545e-05, "loss": 3.1942, "step": 105200 }, { "epoch": 0.9294583090481994, "grad_norm": 2.004868268966675, "learning_rate": 9.18437170070579e-05, "loss": 3.271, "step": 105250 }, { "epoch": 0.9298998569384835, "grad_norm": 1.0723556280136108, "learning_rate": 9.183610869132804e-05, "loss": 3.7162, "step": 105300 }, { "epoch": 0.9303414048287677, "grad_norm": 2.3168883323669434, "learning_rate": 9.182849714411354e-05, "loss": 3.3603, "step": 105350 }, { "epoch": 0.9307829527190519, "grad_norm": 3.932926654815674, "learning_rate": 9.182088236600235e-05, "loss": 3.2771, "step": 105400 }, { "epoch": 0.931224500609336, "grad_norm": 2.3578920364379883, "learning_rate": 9.18132643575826e-05, "loss": 3.2315, "step": 105450 }, { "epoch": 0.9316660484996203, "grad_norm": 2.431108236312866, "learning_rate": 9.180564311944276e-05, "loss": 3.5825, "step": 105500 }, { "epoch": 0.9321075963899045, "grad_norm": 1.2201814651489258, "learning_rate": 9.179801865217149e-05, "loss": 3.2987, "step": 105550 }, { "epoch": 0.9325491442801886, "grad_norm": 3.136932134628296, "learning_rate": 9.17903909563577e-05, "loss": 3.5143, "step": 105600 }, { "epoch": 0.9329906921704728, "grad_norm": 1.9136083126068115, "learning_rate": 9.178276003259061e-05, "loss": 2.8707, "step": 105650 }, { "epoch": 0.933432240060757, "grad_norm": 3.5375232696533203, "learning_rate": 9.177512588145959e-05, "loss": 3.5015, "step": 105700 }, { "epoch": 0.9338737879510411, "grad_norm": 3.3134589195251465, "learning_rate": 9.176748850355434e-05, "loss": 3.313, "step": 105750 }, { "epoch": 0.9343153358413253, "grad_norm": 4.367105484008789, "learning_rate": 9.175984789946479e-05, "loss": 3.431, "step": 105800 }, { "epoch": 0.9347568837316095, "grad_norm": 0.8827017545700073, "learning_rate": 9.17522040697811e-05, "loss": 3.2357, "step": 105850 }, { "epoch": 0.9351984316218938, "grad_norm": 2.8383736610412598, "learning_rate": 9.17445570150937e-05, "loss": 3.4118, "step": 105900 }, { "epoch": 0.9356399795121779, "grad_norm": 2.147942543029785, "learning_rate": 9.173690673599325e-05, "loss": 3.4423, "step": 105950 }, { "epoch": 0.9360815274024621, "grad_norm": 0.8061859607696533, "learning_rate": 9.172925323307066e-05, "loss": 2.889, "step": 106000 }, { "epoch": 0.9365230752927463, "grad_norm": 2.103886127471924, "learning_rate": 9.172159650691714e-05, "loss": 3.1968, "step": 106050 }, { "epoch": 0.9369646231830304, "grad_norm": 1.432693362236023, "learning_rate": 9.171393655812406e-05, "loss": 3.4562, "step": 106100 }, { "epoch": 0.9374061710733146, "grad_norm": 1.1502761840820312, "learning_rate": 9.17062733872831e-05, "loss": 3.4893, "step": 106150 }, { "epoch": 0.9378477189635988, "grad_norm": 2.3663687705993652, "learning_rate": 9.169860699498619e-05, "loss": 3.1563, "step": 106200 }, { "epoch": 0.9382892668538829, "grad_norm": 1.9978609085083008, "learning_rate": 9.169093738182548e-05, "loss": 3.0778, "step": 106250 }, { "epoch": 0.9387308147441672, "grad_norm": 1.2889201641082764, "learning_rate": 9.168326454839339e-05, "loss": 3.3576, "step": 106300 }, { "epoch": 0.9391723626344514, "grad_norm": 1.1832932233810425, "learning_rate": 9.167558849528257e-05, "loss": 2.9429, "step": 106350 }, { "epoch": 0.9396139105247355, "grad_norm": 1.980592966079712, "learning_rate": 9.166790922308594e-05, "loss": 3.0273, "step": 106400 }, { "epoch": 0.9400554584150197, "grad_norm": 5.848330020904541, "learning_rate": 9.166038041374784e-05, "loss": 3.5707, "step": 106450 }, { "epoch": 0.9404970063053039, "grad_norm": 1.3137308359146118, "learning_rate": 9.165269476951148e-05, "loss": 3.1561, "step": 106500 }, { "epoch": 0.940938554195588, "grad_norm": 2.650174617767334, "learning_rate": 9.164500590795764e-05, "loss": 3.3189, "step": 106550 }, { "epoch": 0.9413801020858722, "grad_norm": 7.973121643066406, "learning_rate": 9.163731382968025e-05, "loss": 2.9166, "step": 106600 }, { "epoch": 0.9418216499761564, "grad_norm": 3.19098162651062, "learning_rate": 9.162961853527343e-05, "loss": 3.4358, "step": 106650 }, { "epoch": 0.9422631978664406, "grad_norm": 2.6148929595947266, "learning_rate": 9.16219200253316e-05, "loss": 3.0046, "step": 106700 }, { "epoch": 0.9427047457567248, "grad_norm": 2.952760696411133, "learning_rate": 9.161421830044938e-05, "loss": 3.4446, "step": 106750 }, { "epoch": 0.943146293647009, "grad_norm": 1.8998844623565674, "learning_rate": 9.160651336122169e-05, "loss": 3.002, "step": 106800 }, { "epoch": 0.9435878415372931, "grad_norm": 3.6690402030944824, "learning_rate": 9.159880520824364e-05, "loss": 3.2972, "step": 106850 }, { "epoch": 0.9440293894275773, "grad_norm": 1.2669631242752075, "learning_rate": 9.159109384211065e-05, "loss": 3.7589, "step": 106900 }, { "epoch": 0.9444709373178615, "grad_norm": 2.6425857543945312, "learning_rate": 9.158337926341833e-05, "loss": 3.1515, "step": 106950 }, { "epoch": 0.9449124852081456, "grad_norm": 4.218154430389404, "learning_rate": 9.157566147276258e-05, "loss": 2.7466, "step": 107000 }, { "epoch": 0.9453540330984298, "grad_norm": 1.4164811372756958, "learning_rate": 9.156794047073953e-05, "loss": 3.019, "step": 107050 }, { "epoch": 0.9457955809887141, "grad_norm": 2.938023090362549, "learning_rate": 9.156021625794556e-05, "loss": 3.3921, "step": 107100 }, { "epoch": 0.9462371288789982, "grad_norm": 2.360468864440918, "learning_rate": 9.155248883497732e-05, "loss": 2.9222, "step": 107150 }, { "epoch": 0.9466786767692824, "grad_norm": 2.1282787322998047, "learning_rate": 9.154475820243167e-05, "loss": 3.0558, "step": 107200 }, { "epoch": 0.9471202246595666, "grad_norm": 3.2383251190185547, "learning_rate": 9.153702436090572e-05, "loss": 3.1759, "step": 107250 }, { "epoch": 0.9475617725498507, "grad_norm": 1.6410382986068726, "learning_rate": 9.152928731099688e-05, "loss": 3.0554, "step": 107300 }, { "epoch": 0.9480033204401349, "grad_norm": 1.9488804340362549, "learning_rate": 9.152154705330274e-05, "loss": 2.8708, "step": 107350 }, { "epoch": 0.9484448683304191, "grad_norm": 0.915654182434082, "learning_rate": 9.15138035884212e-05, "loss": 2.926, "step": 107400 }, { "epoch": 0.9488864162207032, "grad_norm": 5.75622034072876, "learning_rate": 9.150605691695036e-05, "loss": 3.4072, "step": 107450 }, { "epoch": 0.9493279641109875, "grad_norm": 1.4961851835250854, "learning_rate": 9.149830703948856e-05, "loss": 3.4192, "step": 107500 }, { "epoch": 0.9497695120012717, "grad_norm": 2.9729695320129395, "learning_rate": 9.149055395663446e-05, "loss": 3.0369, "step": 107550 }, { "epoch": 0.9502110598915559, "grad_norm": 3.2683751583099365, "learning_rate": 9.148279766898688e-05, "loss": 3.4268, "step": 107600 }, { "epoch": 0.95065260778184, "grad_norm": 3.0689873695373535, "learning_rate": 9.147503817714496e-05, "loss": 3.1971, "step": 107650 }, { "epoch": 0.9510941556721242, "grad_norm": 7.556708335876465, "learning_rate": 9.146727548170803e-05, "loss": 3.5455, "step": 107700 }, { "epoch": 0.9515357035624084, "grad_norm": 2.019350528717041, "learning_rate": 9.145950958327569e-05, "loss": 3.5969, "step": 107750 }, { "epoch": 0.9519772514526925, "grad_norm": 1.1041678190231323, "learning_rate": 9.145174048244782e-05, "loss": 3.1785, "step": 107800 }, { "epoch": 0.9524187993429767, "grad_norm": 1.0742287635803223, "learning_rate": 9.144396817982449e-05, "loss": 3.0938, "step": 107850 }, { "epoch": 0.952860347233261, "grad_norm": 3.854574203491211, "learning_rate": 9.143619267600605e-05, "loss": 2.8083, "step": 107900 }, { "epoch": 0.9533018951235451, "grad_norm": 2.129457473754883, "learning_rate": 9.14284139715931e-05, "loss": 3.2925, "step": 107950 }, { "epoch": 0.9537434430138293, "grad_norm": 4.399311542510986, "learning_rate": 9.142063206718647e-05, "loss": 3.5332, "step": 108000 }, { "epoch": 0.9537434430138293, "eval_asr_loss": 0.9242537100154695, "eval_loss": 2.951432466506958, "eval_runtime": 20.2039, "eval_samples_per_second": 38.012, "eval_steps_per_second": 9.503, "eval_tts_loss": 5.889301824693111, "step": 108000 }, { "epoch": 0.9541849909041135, "grad_norm": 2.395542860031128, "learning_rate": 9.141284696338725e-05, "loss": 3.7637, "step": 108050 }, { "epoch": 0.9546265387943976, "grad_norm": 5.644167900085449, "learning_rate": 9.140505866079678e-05, "loss": 3.4879, "step": 108100 }, { "epoch": 0.9550680866846818, "grad_norm": 2.9684510231018066, "learning_rate": 9.139726716001662e-05, "loss": 3.2029, "step": 108150 }, { "epoch": 0.955509634574966, "grad_norm": 2.128922939300537, "learning_rate": 9.138947246164862e-05, "loss": 3.6179, "step": 108200 }, { "epoch": 0.9559511824652501, "grad_norm": 1.3739852905273438, "learning_rate": 9.138167456629486e-05, "loss": 3.5268, "step": 108250 }, { "epoch": 0.9563927303555344, "grad_norm": 1.0485423803329468, "learning_rate": 9.137387347455761e-05, "loss": 3.3022, "step": 108300 }, { "epoch": 0.9568342782458186, "grad_norm": 2.3977785110473633, "learning_rate": 9.13660691870395e-05, "loss": 3.52, "step": 108350 }, { "epoch": 0.9572758261361027, "grad_norm": 3.1857287883758545, "learning_rate": 9.135826170434331e-05, "loss": 3.6409, "step": 108400 }, { "epoch": 0.9577173740263869, "grad_norm": 1.380523920059204, "learning_rate": 9.135045102707212e-05, "loss": 3.4588, "step": 108450 }, { "epoch": 0.9581589219166711, "grad_norm": 2.0876681804656982, "learning_rate": 9.134263715582922e-05, "loss": 3.5122, "step": 108500 }, { "epoch": 0.9586004698069552, "grad_norm": 1.6372121572494507, "learning_rate": 9.133482009121816e-05, "loss": 2.4488, "step": 108550 }, { "epoch": 0.9590420176972394, "grad_norm": 1.7906489372253418, "learning_rate": 9.132699983384278e-05, "loss": 3.528, "step": 108600 }, { "epoch": 0.9594835655875236, "grad_norm": 3.648322582244873, "learning_rate": 9.13191763843071e-05, "loss": 3.2061, "step": 108650 }, { "epoch": 0.9599251134778078, "grad_norm": 2.764986991882324, "learning_rate": 9.131134974321542e-05, "loss": 3.2878, "step": 108700 }, { "epoch": 0.960366661368092, "grad_norm": 2.228464365005493, "learning_rate": 9.130351991117229e-05, "loss": 3.1307, "step": 108750 }, { "epoch": 0.9608082092583762, "grad_norm": 1.0917402505874634, "learning_rate": 9.129568688878248e-05, "loss": 2.9365, "step": 108800 }, { "epoch": 0.9612497571486603, "grad_norm": 0.8567925095558167, "learning_rate": 9.128785067665104e-05, "loss": 3.3896, "step": 108850 }, { "epoch": 0.9616913050389445, "grad_norm": 0.9180238246917725, "learning_rate": 9.128001127538325e-05, "loss": 3.7838, "step": 108900 }, { "epoch": 0.9621328529292287, "grad_norm": 6.334617614746094, "learning_rate": 9.127216868558463e-05, "loss": 3.0286, "step": 108950 }, { "epoch": 0.9625744008195128, "grad_norm": 1.8890008926391602, "learning_rate": 9.126432290786096e-05, "loss": 3.2409, "step": 109000 }, { "epoch": 0.963015948709797, "grad_norm": 1.5762968063354492, "learning_rate": 9.125647394281824e-05, "loss": 3.1578, "step": 109050 }, { "epoch": 0.9634574966000813, "grad_norm": 1.3965164422988892, "learning_rate": 9.124862179106278e-05, "loss": 3.3663, "step": 109100 }, { "epoch": 0.9638990444903655, "grad_norm": 2.264342784881592, "learning_rate": 9.124076645320104e-05, "loss": 3.2378, "step": 109150 }, { "epoch": 0.9643405923806496, "grad_norm": 2.7489473819732666, "learning_rate": 9.12329079298398e-05, "loss": 3.5606, "step": 109200 }, { "epoch": 0.9647821402709338, "grad_norm": 1.9388216733932495, "learning_rate": 9.122504622158608e-05, "loss": 3.2344, "step": 109250 }, { "epoch": 0.965223688161218, "grad_norm": 1.1815500259399414, "learning_rate": 9.121733865809997e-05, "loss": 2.8723, "step": 109300 }, { "epoch": 0.9656652360515021, "grad_norm": 0.5941036343574524, "learning_rate": 9.120947064555085e-05, "loss": 3.0666, "step": 109350 }, { "epoch": 0.9661067839417863, "grad_norm": 2.4697368144989014, "learning_rate": 9.120159944991959e-05, "loss": 3.2941, "step": 109400 }, { "epoch": 0.9665483318320705, "grad_norm": 2.9544789791107178, "learning_rate": 9.119372507181415e-05, "loss": 3.0451, "step": 109450 }, { "epoch": 0.9669898797223547, "grad_norm": 4.041789531707764, "learning_rate": 9.118584751184273e-05, "loss": 3.1865, "step": 109500 }, { "epoch": 0.9674314276126389, "grad_norm": 0.4819214344024658, "learning_rate": 9.117796677061387e-05, "loss": 3.2453, "step": 109550 }, { "epoch": 0.9678729755029231, "grad_norm": 3.580658197402954, "learning_rate": 9.117008284873625e-05, "loss": 2.897, "step": 109600 }, { "epoch": 0.9683145233932072, "grad_norm": 1.2903705835342407, "learning_rate": 9.116219574681882e-05, "loss": 3.5748, "step": 109650 }, { "epoch": 0.9687560712834914, "grad_norm": 2.910269260406494, "learning_rate": 9.115430546547083e-05, "loss": 3.1729, "step": 109700 }, { "epoch": 0.9691976191737756, "grad_norm": 2.1742489337921143, "learning_rate": 9.114641200530171e-05, "loss": 3.8506, "step": 109750 }, { "epoch": 0.9696391670640597, "grad_norm": 1.4163600206375122, "learning_rate": 9.11385153669212e-05, "loss": 3.486, "step": 109800 }, { "epoch": 0.970080714954344, "grad_norm": 2.7012150287628174, "learning_rate": 9.11306155509392e-05, "loss": 3.0437, "step": 109850 }, { "epoch": 0.9705222628446282, "grad_norm": 3.0786678791046143, "learning_rate": 9.112271255796593e-05, "loss": 2.9614, "step": 109900 }, { "epoch": 0.9709638107349123, "grad_norm": 2.5239641666412354, "learning_rate": 9.111480638861183e-05, "loss": 2.8935, "step": 109950 }, { "epoch": 0.9714053586251965, "grad_norm": 1.9971002340316772, "learning_rate": 9.110689704348758e-05, "loss": 3.0526, "step": 110000 }, { "epoch": 0.9718469065154807, "grad_norm": 2.921504259109497, "learning_rate": 9.10989845232041e-05, "loss": 2.9988, "step": 110050 }, { "epoch": 0.9722884544057648, "grad_norm": 3.520411491394043, "learning_rate": 9.109106882837259e-05, "loss": 2.7144, "step": 110100 }, { "epoch": 0.972730002296049, "grad_norm": 11.24007511138916, "learning_rate": 9.108314995960445e-05, "loss": 3.398, "step": 110150 }, { "epoch": 0.9731715501863332, "grad_norm": 1.7013554573059082, "learning_rate": 9.107522791751135e-05, "loss": 3.3744, "step": 110200 }, { "epoch": 0.9736130980766174, "grad_norm": 5.85800313949585, "learning_rate": 9.106746123808995e-05, "loss": 2.8389, "step": 110250 }, { "epoch": 0.9740546459669016, "grad_norm": 0.44112929701805115, "learning_rate": 9.105953291461892e-05, "loss": 3.5296, "step": 110300 }, { "epoch": 0.9744961938571858, "grad_norm": 1.3203797340393066, "learning_rate": 9.105160141964715e-05, "loss": 3.5493, "step": 110350 }, { "epoch": 0.9749377417474699, "grad_norm": 2.4732491970062256, "learning_rate": 9.104366675378729e-05, "loss": 3.4198, "step": 110400 }, { "epoch": 0.9753792896377541, "grad_norm": 3.4213356971740723, "learning_rate": 9.10357289176522e-05, "loss": 3.2578, "step": 110450 }, { "epoch": 0.9758208375280383, "grad_norm": 2.482844352722168, "learning_rate": 9.102778791185502e-05, "loss": 3.2519, "step": 110500 }, { "epoch": 0.9762623854183224, "grad_norm": 2.138728380203247, "learning_rate": 9.101984373700912e-05, "loss": 3.0722, "step": 110550 }, { "epoch": 0.9767039333086066, "grad_norm": 1.6257545948028564, "learning_rate": 9.101189639372815e-05, "loss": 2.7877, "step": 110600 }, { "epoch": 0.9771454811988909, "grad_norm": 1.8852672576904297, "learning_rate": 9.100394588262595e-05, "loss": 3.0108, "step": 110650 }, { "epoch": 0.977587029089175, "grad_norm": 4.819422245025635, "learning_rate": 9.099599220431662e-05, "loss": 3.0413, "step": 110700 }, { "epoch": 0.9780285769794592, "grad_norm": 1.705790400505066, "learning_rate": 9.098803535941456e-05, "loss": 3.0909, "step": 110750 }, { "epoch": 0.9784701248697434, "grad_norm": 2.922133445739746, "learning_rate": 9.098007534853431e-05, "loss": 3.318, "step": 110800 }, { "epoch": 0.9789116727600276, "grad_norm": 0.9488845467567444, "learning_rate": 9.097211217229074e-05, "loss": 3.0227, "step": 110850 }, { "epoch": 0.9793532206503117, "grad_norm": 1.900322675704956, "learning_rate": 9.096414583129895e-05, "loss": 2.8965, "step": 110900 }, { "epoch": 0.9797947685405959, "grad_norm": 4.18387508392334, "learning_rate": 9.095617632617426e-05, "loss": 2.5843, "step": 110950 }, { "epoch": 0.98023631643088, "grad_norm": 4.1672444343566895, "learning_rate": 9.094820365753224e-05, "loss": 3.7352, "step": 111000 }, { "epoch": 0.98023631643088, "eval_asr_loss": 0.9174097384220402, "eval_loss": 2.9519290924072266, "eval_runtime": 20.6953, "eval_samples_per_second": 37.11, "eval_steps_per_second": 9.277, "eval_tts_loss": 5.933492346852046, "step": 111000 }, { "epoch": 0.9806778643211643, "grad_norm": 1.7948575019836426, "learning_rate": 9.094022782598873e-05, "loss": 3.3024, "step": 111050 }, { "epoch": 0.9811194122114485, "grad_norm": 1.7361973524093628, "learning_rate": 9.093224883215975e-05, "loss": 3.2393, "step": 111100 }, { "epoch": 0.9815609601017327, "grad_norm": 3.170607805252075, "learning_rate": 9.092426667666167e-05, "loss": 2.9186, "step": 111150 }, { "epoch": 0.9820025079920168, "grad_norm": 2.6592564582824707, "learning_rate": 9.091628136011103e-05, "loss": 3.1093, "step": 111200 }, { "epoch": 0.982444055882301, "grad_norm": 3.301297187805176, "learning_rate": 9.090829288312459e-05, "loss": 3.4304, "step": 111250 }, { "epoch": 0.9828856037725852, "grad_norm": 1.6504911184310913, "learning_rate": 9.090030124631943e-05, "loss": 2.9901, "step": 111300 }, { "epoch": 0.9833271516628693, "grad_norm": 1.7255781888961792, "learning_rate": 9.089230645031281e-05, "loss": 3.1932, "step": 111350 }, { "epoch": 0.9837686995531535, "grad_norm": 1.882737159729004, "learning_rate": 9.088446848576421e-05, "loss": 3.1261, "step": 111400 }, { "epoch": 0.9842102474434378, "grad_norm": 2.1880311965942383, "learning_rate": 9.087646743636082e-05, "loss": 2.9171, "step": 111450 }, { "epoch": 0.9846517953337219, "grad_norm": 2.5656466484069824, "learning_rate": 9.086846322959693e-05, "loss": 3.2527, "step": 111500 }, { "epoch": 0.9850933432240061, "grad_norm": 1.9035078287124634, "learning_rate": 9.086045586609081e-05, "loss": 2.9598, "step": 111550 }, { "epoch": 0.9855348911142903, "grad_norm": 2.160431385040283, "learning_rate": 9.085244534646095e-05, "loss": 2.8841, "step": 111600 }, { "epoch": 0.9859764390045744, "grad_norm": 1.5755068063735962, "learning_rate": 9.084443167132613e-05, "loss": 3.3504, "step": 111650 }, { "epoch": 0.9864179868948586, "grad_norm": 1.7207096815109253, "learning_rate": 9.083641484130529e-05, "loss": 3.3112, "step": 111700 }, { "epoch": 0.9868595347851428, "grad_norm": 2.6062755584716797, "learning_rate": 9.082839485701771e-05, "loss": 3.3106, "step": 111750 }, { "epoch": 0.9873010826754269, "grad_norm": 2.2912654876708984, "learning_rate": 9.082037171908282e-05, "loss": 3.0499, "step": 111800 }, { "epoch": 0.9877426305657112, "grad_norm": 1.4460901021957397, "learning_rate": 9.081234542812038e-05, "loss": 3.7295, "step": 111850 }, { "epoch": 0.9881841784559954, "grad_norm": 1.997914433479309, "learning_rate": 9.080431598475032e-05, "loss": 3.2918, "step": 111900 }, { "epoch": 0.9886257263462795, "grad_norm": 2.2903008460998535, "learning_rate": 9.079628338959286e-05, "loss": 3.5094, "step": 111950 }, { "epoch": 0.9890672742365637, "grad_norm": 2.559363603591919, "learning_rate": 9.078824764326845e-05, "loss": 3.3677, "step": 112000 }, { "epoch": 0.9895088221268479, "grad_norm": 3.8502397537231445, "learning_rate": 9.07802087463978e-05, "loss": 3.0312, "step": 112050 }, { "epoch": 0.989950370017132, "grad_norm": 3.3244760036468506, "learning_rate": 9.077216669960181e-05, "loss": 3.4518, "step": 112100 }, { "epoch": 0.9903919179074162, "grad_norm": 3.6893279552459717, "learning_rate": 9.076412150350169e-05, "loss": 3.1172, "step": 112150 }, { "epoch": 0.9908334657977004, "grad_norm": 1.7387036085128784, "learning_rate": 9.075607315871885e-05, "loss": 3.1974, "step": 112200 }, { "epoch": 0.9912750136879847, "grad_norm": 3.1072323322296143, "learning_rate": 9.074802166587496e-05, "loss": 2.914, "step": 112250 }, { "epoch": 0.9917165615782688, "grad_norm": 2.5612170696258545, "learning_rate": 9.073996702559193e-05, "loss": 3.5433, "step": 112300 }, { "epoch": 0.992158109468553, "grad_norm": 2.5611658096313477, "learning_rate": 9.073190923849193e-05, "loss": 3.3504, "step": 112350 }, { "epoch": 0.9925996573588372, "grad_norm": 2.5978565216064453, "learning_rate": 9.07238483051973e-05, "loss": 3.011, "step": 112400 }, { "epoch": 0.9930412052491213, "grad_norm": 4.176834583282471, "learning_rate": 9.071578422633075e-05, "loss": 3.2228, "step": 112450 }, { "epoch": 0.9934827531394055, "grad_norm": 1.0866472721099854, "learning_rate": 9.070771700251512e-05, "loss": 3.4185, "step": 112500 }, { "epoch": 0.9939243010296896, "grad_norm": 2.3132870197296143, "learning_rate": 9.069964663437354e-05, "loss": 3.3366, "step": 112550 }, { "epoch": 0.9943658489199738, "grad_norm": 1.7939915657043457, "learning_rate": 9.069157312252938e-05, "loss": 3.7442, "step": 112600 }, { "epoch": 0.9948073968102581, "grad_norm": 1.947704553604126, "learning_rate": 9.068349646760626e-05, "loss": 3.3506, "step": 112650 }, { "epoch": 0.9952489447005423, "grad_norm": 1.6947288513183594, "learning_rate": 9.0675416670228e-05, "loss": 3.4434, "step": 112700 }, { "epoch": 0.9956904925908264, "grad_norm": 5.008423805236816, "learning_rate": 9.066733373101875e-05, "loss": 3.3881, "step": 112750 }, { "epoch": 0.9961320404811106, "grad_norm": 0.6654936671257019, "learning_rate": 9.06592476506028e-05, "loss": 3.3806, "step": 112800 }, { "epoch": 0.9965735883713948, "grad_norm": 4.415541172027588, "learning_rate": 9.065115842960475e-05, "loss": 2.8368, "step": 112850 }, { "epoch": 0.9970151362616789, "grad_norm": 2.217128276824951, "learning_rate": 9.064306606864943e-05, "loss": 3.4863, "step": 112900 }, { "epoch": 0.9974566841519631, "grad_norm": 2.124258518218994, "learning_rate": 9.063497056836189e-05, "loss": 3.4668, "step": 112950 }, { "epoch": 0.9978982320422473, "grad_norm": 2.5500919818878174, "learning_rate": 9.062687192936745e-05, "loss": 3.2656, "step": 113000 }, { "epoch": 0.9983397799325315, "grad_norm": 1.8848084211349487, "learning_rate": 9.061877015229166e-05, "loss": 3.3762, "step": 113050 }, { "epoch": 0.9987813278228157, "grad_norm": 1.6677478551864624, "learning_rate": 9.061066523776032e-05, "loss": 3.2986, "step": 113100 }, { "epoch": 0.9992228757130999, "grad_norm": 2.6204864978790283, "learning_rate": 9.060255718639945e-05, "loss": 3.4573, "step": 113150 }, { "epoch": 0.999664423603384, "grad_norm": 2.433886766433716, "learning_rate": 9.059444599883534e-05, "loss": 3.19, "step": 113200 }, { "epoch": 1.0001059714936682, "grad_norm": 3.542097568511963, "learning_rate": 9.058633167569451e-05, "loss": 3.1019, "step": 113250 }, { "epoch": 1.0005475193839524, "grad_norm": 3.6584270000457764, "learning_rate": 9.057821421760372e-05, "loss": 3.1219, "step": 113300 }, { "epoch": 1.0009890672742365, "grad_norm": 2.3734920024871826, "learning_rate": 9.057009362518998e-05, "loss": 3.0769, "step": 113350 }, { "epoch": 1.0014306151645207, "grad_norm": 1.0202910900115967, "learning_rate": 9.056196989908053e-05, "loss": 3.166, "step": 113400 }, { "epoch": 1.0018721630548049, "grad_norm": 5.295326232910156, "learning_rate": 9.055384303990285e-05, "loss": 3.2082, "step": 113450 }, { "epoch": 1.002313710945089, "grad_norm": 3.40751314163208, "learning_rate": 9.054571304828468e-05, "loss": 3.1396, "step": 113500 }, { "epoch": 1.0027552588353732, "grad_norm": 1.1059200763702393, "learning_rate": 9.0537579924854e-05, "loss": 3.0367, "step": 113550 }, { "epoch": 1.0031968067256574, "grad_norm": 1.1731442213058472, "learning_rate": 9.052944367023902e-05, "loss": 3.0164, "step": 113600 }, { "epoch": 1.0036383546159418, "grad_norm": 2.744382381439209, "learning_rate": 9.05213042850682e-05, "loss": 2.7251, "step": 113650 }, { "epoch": 1.004079902506226, "grad_norm": 3.96297550201416, "learning_rate": 9.051316176997023e-05, "loss": 3.0946, "step": 113700 }, { "epoch": 1.00452145039651, "grad_norm": 1.3721880912780762, "learning_rate": 9.050501612557406e-05, "loss": 3.1733, "step": 113750 }, { "epoch": 1.0049629982867943, "grad_norm": 4.159188270568848, "learning_rate": 9.049686735250886e-05, "loss": 3.3159, "step": 113800 }, { "epoch": 1.0054045461770784, "grad_norm": 2.8054113388061523, "learning_rate": 9.048871545140407e-05, "loss": 3.2084, "step": 113850 }, { "epoch": 1.0058460940673626, "grad_norm": 5.3158392906188965, "learning_rate": 9.048056042288933e-05, "loss": 3.0027, "step": 113900 }, { "epoch": 1.0062876419576468, "grad_norm": 0.9245600700378418, "learning_rate": 9.047240226759457e-05, "loss": 3.0888, "step": 113950 }, { "epoch": 1.006729189847931, "grad_norm": 2.9411399364471436, "learning_rate": 9.046424098614993e-05, "loss": 3.3146, "step": 114000 }, { "epoch": 1.006729189847931, "eval_asr_loss": 0.9158940412254022, "eval_loss": 2.9361159801483154, "eval_runtime": 20.4244, "eval_samples_per_second": 37.602, "eval_steps_per_second": 9.401, "eval_tts_loss": 6.045867164957902, "step": 114000 }, { "epoch": 1.007170737738215, "grad_norm": 1.7878565788269043, "learning_rate": 9.045607657918583e-05, "loss": 3.6228, "step": 114050 }, { "epoch": 1.0076122856284992, "grad_norm": 1.3481472730636597, "learning_rate": 9.044790904733282e-05, "loss": 3.5248, "step": 114100 }, { "epoch": 1.0080538335187834, "grad_norm": 1.8237581253051758, "learning_rate": 9.043973839122186e-05, "loss": 3.0638, "step": 114150 }, { "epoch": 1.0084953814090676, "grad_norm": 3.578498363494873, "learning_rate": 9.0431564611484e-05, "loss": 3.2789, "step": 114200 }, { "epoch": 1.0089369292993517, "grad_norm": 1.5951688289642334, "learning_rate": 9.042338770875067e-05, "loss": 3.259, "step": 114250 }, { "epoch": 1.009378477189636, "grad_norm": 2.594109058380127, "learning_rate": 9.041520768365338e-05, "loss": 2.8875, "step": 114300 }, { "epoch": 1.00982002507992, "grad_norm": 2.1558804512023926, "learning_rate": 9.040702453682402e-05, "loss": 2.6654, "step": 114350 }, { "epoch": 1.0102615729702042, "grad_norm": 2.0555760860443115, "learning_rate": 9.039883826889466e-05, "loss": 3.57, "step": 114400 }, { "epoch": 1.0107031208604886, "grad_norm": 3.0848312377929688, "learning_rate": 9.03906488804976e-05, "loss": 2.8284, "step": 114450 }, { "epoch": 1.0111446687507728, "grad_norm": 0.27899661660194397, "learning_rate": 9.038245637226543e-05, "loss": 3.164, "step": 114500 }, { "epoch": 1.011586216641057, "grad_norm": 1.8623350858688354, "learning_rate": 9.037426074483093e-05, "loss": 2.9045, "step": 114550 }, { "epoch": 1.0120277645313411, "grad_norm": 1.1736133098602295, "learning_rate": 9.036606199882716e-05, "loss": 3.3847, "step": 114600 }, { "epoch": 1.0124693124216253, "grad_norm": 6.291317462921143, "learning_rate": 9.035786013488739e-05, "loss": 3.3651, "step": 114650 }, { "epoch": 1.0129108603119095, "grad_norm": 3.4732656478881836, "learning_rate": 9.034965515364515e-05, "loss": 3.8414, "step": 114700 }, { "epoch": 1.0133524082021936, "grad_norm": 3.2274088859558105, "learning_rate": 9.034144705573422e-05, "loss": 2.9208, "step": 114750 }, { "epoch": 1.0137939560924778, "grad_norm": 1.771211862564087, "learning_rate": 9.033323584178857e-05, "loss": 3.3398, "step": 114800 }, { "epoch": 1.014235503982762, "grad_norm": 2.4299356937408447, "learning_rate": 9.032502151244246e-05, "loss": 3.5498, "step": 114850 }, { "epoch": 1.0146770518730461, "grad_norm": 3.958322286605835, "learning_rate": 9.031680406833041e-05, "loss": 3.13, "step": 114900 }, { "epoch": 1.0151185997633303, "grad_norm": 1.6944149732589722, "learning_rate": 9.030858351008711e-05, "loss": 3.3474, "step": 114950 }, { "epoch": 1.0155601476536145, "grad_norm": 1.9284944534301758, "learning_rate": 9.030035983834754e-05, "loss": 3.3015, "step": 115000 }, { "epoch": 1.0160016955438986, "grad_norm": 9.017916679382324, "learning_rate": 9.02921330537469e-05, "loss": 3.0109, "step": 115050 }, { "epoch": 1.0164432434341828, "grad_norm": 2.3894286155700684, "learning_rate": 9.028390315692066e-05, "loss": 3.0692, "step": 115100 }, { "epoch": 1.016884791324467, "grad_norm": 3.3184964656829834, "learning_rate": 9.02756701485045e-05, "loss": 2.9476, "step": 115150 }, { "epoch": 1.0173263392147511, "grad_norm": 1.9609465599060059, "learning_rate": 9.026743402913435e-05, "loss": 2.94, "step": 115200 }, { "epoch": 1.0177678871050355, "grad_norm": 1.2730053663253784, "learning_rate": 9.025919479944638e-05, "loss": 3.3829, "step": 115250 }, { "epoch": 1.0182094349953197, "grad_norm": 2.1093947887420654, "learning_rate": 9.025095246007699e-05, "loss": 3.3117, "step": 115300 }, { "epoch": 1.0186509828856039, "grad_norm": 4.5406670570373535, "learning_rate": 9.024270701166285e-05, "loss": 3.3581, "step": 115350 }, { "epoch": 1.019092530775888, "grad_norm": 2.3314759731292725, "learning_rate": 9.023445845484083e-05, "loss": 3.2504, "step": 115400 }, { "epoch": 1.0195340786661722, "grad_norm": 1.8314716815948486, "learning_rate": 9.022620679024807e-05, "loss": 3.0811, "step": 115450 }, { "epoch": 1.0199756265564563, "grad_norm": 3.113945960998535, "learning_rate": 9.021795201852197e-05, "loss": 2.9139, "step": 115500 }, { "epoch": 1.0204171744467405, "grad_norm": 7.921067714691162, "learning_rate": 9.020969414030009e-05, "loss": 3.1436, "step": 115550 }, { "epoch": 1.0208587223370247, "grad_norm": 2.9270060062408447, "learning_rate": 9.020143315622032e-05, "loss": 3.2475, "step": 115600 }, { "epoch": 1.0213002702273088, "grad_norm": 2.903693675994873, "learning_rate": 9.019316906692072e-05, "loss": 3.5003, "step": 115650 }, { "epoch": 1.021741818117593, "grad_norm": 2.832529306411743, "learning_rate": 9.018490187303966e-05, "loss": 3.0619, "step": 115700 }, { "epoch": 1.0221833660078772, "grad_norm": 2.0733742713928223, "learning_rate": 9.017663157521567e-05, "loss": 3.1382, "step": 115750 }, { "epoch": 1.0226249138981613, "grad_norm": 2.4163548946380615, "learning_rate": 9.016835817408759e-05, "loss": 3.6552, "step": 115800 }, { "epoch": 1.0230664617884455, "grad_norm": 4.126326084136963, "learning_rate": 9.016008167029445e-05, "loss": 3.3899, "step": 115850 }, { "epoch": 1.0235080096787297, "grad_norm": 2.870994806289673, "learning_rate": 9.015180206447556e-05, "loss": 3.1306, "step": 115900 }, { "epoch": 1.0239495575690138, "grad_norm": 2.4682250022888184, "learning_rate": 9.014351935727041e-05, "loss": 3.3603, "step": 115950 }, { "epoch": 1.024391105459298, "grad_norm": 1.5408371686935425, "learning_rate": 9.013523354931883e-05, "loss": 3.3107, "step": 116000 }, { "epoch": 1.0248326533495824, "grad_norm": 3.0599887371063232, "learning_rate": 9.012694464126077e-05, "loss": 3.2021, "step": 116050 }, { "epoch": 1.0252742012398666, "grad_norm": 2.876763105392456, "learning_rate": 9.011865263373653e-05, "loss": 3.4132, "step": 116100 }, { "epoch": 1.0257157491301507, "grad_norm": 1.7822761535644531, "learning_rate": 9.011035752738655e-05, "loss": 3.1308, "step": 116150 }, { "epoch": 1.026157297020435, "grad_norm": 2.372128963470459, "learning_rate": 9.010205932285159e-05, "loss": 2.9991, "step": 116200 }, { "epoch": 1.026598844910719, "grad_norm": 3.578298568725586, "learning_rate": 9.00937580207726e-05, "loss": 3.0682, "step": 116250 }, { "epoch": 1.0270403928010032, "grad_norm": 2.4689536094665527, "learning_rate": 9.008545362179077e-05, "loss": 3.5783, "step": 116300 }, { "epoch": 1.0274819406912874, "grad_norm": 1.7478725910186768, "learning_rate": 9.007714612654759e-05, "loss": 3.2391, "step": 116350 }, { "epoch": 1.0279234885815716, "grad_norm": 1.0738362073898315, "learning_rate": 9.006883553568469e-05, "loss": 2.953, "step": 116400 }, { "epoch": 1.0283650364718557, "grad_norm": 2.858876943588257, "learning_rate": 9.006052184984403e-05, "loss": 3.3315, "step": 116450 }, { "epoch": 1.02880658436214, "grad_norm": 3.038067102432251, "learning_rate": 9.005220506966776e-05, "loss": 3.0056, "step": 116500 }, { "epoch": 1.029248132252424, "grad_norm": 2.7625601291656494, "learning_rate": 9.004388519579829e-05, "loss": 3.4374, "step": 116550 }, { "epoch": 1.0296896801427082, "grad_norm": 1.6374485492706299, "learning_rate": 9.003556222887823e-05, "loss": 3.2502, "step": 116600 }, { "epoch": 1.0301312280329924, "grad_norm": 4.125757694244385, "learning_rate": 9.002723616955049e-05, "loss": 2.8794, "step": 116650 }, { "epoch": 1.0305727759232766, "grad_norm": 1.5545926094055176, "learning_rate": 9.001890701845819e-05, "loss": 3.3588, "step": 116700 }, { "epoch": 1.0310143238135607, "grad_norm": 6.130286693572998, "learning_rate": 9.001057477624467e-05, "loss": 3.479, "step": 116750 }, { "epoch": 1.031455871703845, "grad_norm": 1.1050196886062622, "learning_rate": 9.000223944355351e-05, "loss": 3.1521, "step": 116800 }, { "epoch": 1.0318974195941293, "grad_norm": 4.7166056632995605, "learning_rate": 8.999390102102858e-05, "loss": 3.0597, "step": 116850 }, { "epoch": 1.0323389674844135, "grad_norm": 5.70416259765625, "learning_rate": 8.998555950931392e-05, "loss": 3.5213, "step": 116900 }, { "epoch": 1.0327805153746976, "grad_norm": 2.4848475456237793, "learning_rate": 8.997721490905386e-05, "loss": 3.5041, "step": 116950 }, { "epoch": 1.0332220632649818, "grad_norm": 1.464582920074463, "learning_rate": 8.996886722089295e-05, "loss": 3.0542, "step": 117000 }, { "epoch": 1.0332220632649818, "eval_asr_loss": 0.9235363394382403, "eval_loss": 2.937559127807617, "eval_runtime": 22.1307, "eval_samples_per_second": 34.703, "eval_steps_per_second": 8.676, "eval_tts_loss": 5.999984977433617, "step": 117000 }, { "epoch": 1.033663611155266, "grad_norm": 0.5302391648292542, "learning_rate": 8.996051644547598e-05, "loss": 3.2493, "step": 117050 }, { "epoch": 1.0341051590455501, "grad_norm": 1.5811774730682373, "learning_rate": 8.995216258344796e-05, "loss": 3.317, "step": 117100 }, { "epoch": 1.0345467069358343, "grad_norm": 1.782463550567627, "learning_rate": 8.994380563545417e-05, "loss": 2.9838, "step": 117150 }, { "epoch": 1.0349882548261184, "grad_norm": 4.319481372833252, "learning_rate": 8.99354456021401e-05, "loss": 3.1259, "step": 117200 }, { "epoch": 1.0354298027164026, "grad_norm": 2.537027597427368, "learning_rate": 8.99270824841515e-05, "loss": 3.2959, "step": 117250 }, { "epoch": 1.0358713506066868, "grad_norm": 0.6770037412643433, "learning_rate": 8.991871628213436e-05, "loss": 3.3386, "step": 117300 }, { "epoch": 1.036312898496971, "grad_norm": 2.032209634780884, "learning_rate": 8.991034699673488e-05, "loss": 3.0363, "step": 117350 }, { "epoch": 1.0367544463872551, "grad_norm": 1.5942469835281372, "learning_rate": 8.990197462859952e-05, "loss": 3.5947, "step": 117400 }, { "epoch": 1.0371959942775393, "grad_norm": 3.661116361618042, "learning_rate": 8.989359917837498e-05, "loss": 3.1945, "step": 117450 }, { "epoch": 1.0376375421678234, "grad_norm": 1.773616075515747, "learning_rate": 8.988522064670822e-05, "loss": 2.7827, "step": 117500 }, { "epoch": 1.0380790900581076, "grad_norm": 7.883002758026123, "learning_rate": 8.987683903424636e-05, "loss": 3.0637, "step": 117550 }, { "epoch": 1.0385206379483918, "grad_norm": 1.7299555540084839, "learning_rate": 8.986845434163682e-05, "loss": 3.4542, "step": 117600 }, { "epoch": 1.0389621858386762, "grad_norm": 1.8580836057662964, "learning_rate": 8.986006656952727e-05, "loss": 3.1663, "step": 117650 }, { "epoch": 1.0394037337289603, "grad_norm": 1.1546132564544678, "learning_rate": 8.985167571856556e-05, "loss": 3.1597, "step": 117700 }, { "epoch": 1.0398452816192445, "grad_norm": 4.701486110687256, "learning_rate": 8.984328178939985e-05, "loss": 3.4678, "step": 117750 }, { "epoch": 1.0402868295095287, "grad_norm": 1.7120305299758911, "learning_rate": 8.98348847826785e-05, "loss": 3.0328, "step": 117800 }, { "epoch": 1.0407283773998128, "grad_norm": 2.6418991088867188, "learning_rate": 8.982648469905006e-05, "loss": 3.137, "step": 117850 }, { "epoch": 1.041169925290097, "grad_norm": 1.3650034666061401, "learning_rate": 8.981808153916341e-05, "loss": 2.8192, "step": 117900 }, { "epoch": 1.0416114731803812, "grad_norm": 0.9207521677017212, "learning_rate": 8.980967530366762e-05, "loss": 3.3502, "step": 117950 }, { "epoch": 1.0420530210706653, "grad_norm": 4.987005233764648, "learning_rate": 8.980126599321197e-05, "loss": 3.1026, "step": 118000 }, { "epoch": 1.0424945689609495, "grad_norm": 5.551382064819336, "learning_rate": 8.979285360844602e-05, "loss": 3.0126, "step": 118050 }, { "epoch": 1.0429361168512337, "grad_norm": 3.7902963161468506, "learning_rate": 8.97844381500196e-05, "loss": 3.6057, "step": 118100 }, { "epoch": 1.0433776647415178, "grad_norm": 2.5070292949676514, "learning_rate": 8.977601961858267e-05, "loss": 3.2931, "step": 118150 }, { "epoch": 1.043819212631802, "grad_norm": 1.3204067945480347, "learning_rate": 8.97675980147855e-05, "loss": 3.6482, "step": 118200 }, { "epoch": 1.0442607605220862, "grad_norm": 1.2875250577926636, "learning_rate": 8.975917333927862e-05, "loss": 2.8817, "step": 118250 }, { "epoch": 1.0447023084123703, "grad_norm": 3.1866626739501953, "learning_rate": 8.975074559271275e-05, "loss": 3.4228, "step": 118300 }, { "epoch": 1.0451438563026545, "grad_norm": 2.342751979827881, "learning_rate": 8.974231477573885e-05, "loss": 3.4827, "step": 118350 }, { "epoch": 1.0455854041929387, "grad_norm": 4.2284979820251465, "learning_rate": 8.973388088900814e-05, "loss": 2.6648, "step": 118400 }, { "epoch": 1.046026952083223, "grad_norm": 1.4045745134353638, "learning_rate": 8.972544393317208e-05, "loss": 2.9577, "step": 118450 }, { "epoch": 1.0464684999735072, "grad_norm": 1.3765054941177368, "learning_rate": 8.971700390888233e-05, "loss": 3.0889, "step": 118500 }, { "epoch": 1.0469100478637914, "grad_norm": 1.7757630348205566, "learning_rate": 8.970856081679081e-05, "loss": 3.2674, "step": 118550 }, { "epoch": 1.0473515957540755, "grad_norm": 2.229665517807007, "learning_rate": 8.970011465754969e-05, "loss": 3.1471, "step": 118600 }, { "epoch": 1.0477931436443597, "grad_norm": 2.3084306716918945, "learning_rate": 8.969166543181136e-05, "loss": 3.3577, "step": 118650 }, { "epoch": 1.0482346915346439, "grad_norm": 1.667510747909546, "learning_rate": 8.968321314022845e-05, "loss": 2.8539, "step": 118700 }, { "epoch": 1.048676239424928, "grad_norm": 2.1955771446228027, "learning_rate": 8.967475778345385e-05, "loss": 3.0149, "step": 118750 }, { "epoch": 1.0491177873152122, "grad_norm": 3.0574069023132324, "learning_rate": 8.966629936214062e-05, "loss": 3.5799, "step": 118800 }, { "epoch": 1.0495593352054964, "grad_norm": 3.2477333545684814, "learning_rate": 8.965783787694212e-05, "loss": 3.618, "step": 118850 }, { "epoch": 1.0500008830957805, "grad_norm": 1.1270619630813599, "learning_rate": 8.964937332851193e-05, "loss": 3.346, "step": 118900 }, { "epoch": 1.0504424309860647, "grad_norm": 3.402944803237915, "learning_rate": 8.964090571750389e-05, "loss": 3.252, "step": 118950 }, { "epoch": 1.0508839788763489, "grad_norm": 3.217515230178833, "learning_rate": 8.963243504457202e-05, "loss": 2.9268, "step": 119000 }, { "epoch": 1.051325526766633, "grad_norm": 1.877545714378357, "learning_rate": 8.962396131037061e-05, "loss": 3.0917, "step": 119050 }, { "epoch": 1.0517670746569172, "grad_norm": 0.7133153676986694, "learning_rate": 8.96154845155542e-05, "loss": 3.386, "step": 119100 }, { "epoch": 1.0522086225472014, "grad_norm": 1.9354885816574097, "learning_rate": 8.960700466077751e-05, "loss": 2.8361, "step": 119150 }, { "epoch": 1.0526501704374855, "grad_norm": 1.1146783828735352, "learning_rate": 8.95985217466956e-05, "loss": 3.2829, "step": 119200 }, { "epoch": 1.05309171832777, "grad_norm": 1.3031646013259888, "learning_rate": 8.959003577396367e-05, "loss": 3.1304, "step": 119250 }, { "epoch": 1.053533266218054, "grad_norm": 2.9102251529693604, "learning_rate": 8.958154674323717e-05, "loss": 3.3583, "step": 119300 }, { "epoch": 1.0539748141083383, "grad_norm": 1.9654988050460815, "learning_rate": 8.957305465517185e-05, "loss": 3.1736, "step": 119350 }, { "epoch": 1.0544163619986224, "grad_norm": 4.362683296203613, "learning_rate": 8.956455951042361e-05, "loss": 2.8865, "step": 119400 }, { "epoch": 1.0548579098889066, "grad_norm": 2.9880433082580566, "learning_rate": 8.955606130964865e-05, "loss": 3.3161, "step": 119450 }, { "epoch": 1.0552994577791908, "grad_norm": 2.5000228881835938, "learning_rate": 8.954756005350339e-05, "loss": 2.7501, "step": 119500 }, { "epoch": 1.055741005669475, "grad_norm": 3.525005578994751, "learning_rate": 8.953905574264445e-05, "loss": 3.082, "step": 119550 }, { "epoch": 1.056182553559759, "grad_norm": 3.375227928161621, "learning_rate": 8.953054837772874e-05, "loss": 3.4092, "step": 119600 }, { "epoch": 1.0566241014500433, "grad_norm": 7.511069297790527, "learning_rate": 8.952203795941339e-05, "loss": 3.2211, "step": 119650 }, { "epoch": 1.0570656493403274, "grad_norm": 1.7979106903076172, "learning_rate": 8.951369478768952e-05, "loss": 2.9699, "step": 119700 }, { "epoch": 1.0575071972306116, "grad_norm": 1.3295596837997437, "learning_rate": 8.95051783255824e-05, "loss": 2.8737, "step": 119750 }, { "epoch": 1.0579487451208958, "grad_norm": 1.0681979656219482, "learning_rate": 8.949682923220605e-05, "loss": 3.1526, "step": 119800 }, { "epoch": 1.05839029301118, "grad_norm": 8.75452709197998, "learning_rate": 8.948830672888615e-05, "loss": 3.0168, "step": 119850 }, { "epoch": 1.058831840901464, "grad_norm": 2.01828670501709, "learning_rate": 8.947978117542938e-05, "loss": 2.943, "step": 119900 }, { "epoch": 1.0592733887917483, "grad_norm": 2.2010951042175293, "learning_rate": 8.947125257249429e-05, "loss": 3.1595, "step": 119950 }, { "epoch": 1.0597149366820324, "grad_norm": 1.8590667247772217, "learning_rate": 8.946272092073967e-05, "loss": 3.2562, "step": 120000 }, { "epoch": 1.0597149366820324, "eval_asr_loss": 0.9155421388644238, "eval_loss": 2.9473683834075928, "eval_runtime": 20.7175, "eval_samples_per_second": 37.07, "eval_steps_per_second": 9.268, "eval_tts_loss": 5.970084495848347, "step": 120000 }, { "epoch": 1.0601564845723168, "grad_norm": 1.69383704662323, "learning_rate": 8.945418622082446e-05, "loss": 3.316, "step": 120050 }, { "epoch": 1.060598032462601, "grad_norm": 1.9276316165924072, "learning_rate": 8.944564847340796e-05, "loss": 3.0642, "step": 120100 }, { "epoch": 1.0610395803528851, "grad_norm": 4.82901668548584, "learning_rate": 8.943710767914958e-05, "loss": 2.7604, "step": 120150 }, { "epoch": 1.0614811282431693, "grad_norm": 0.9138553142547607, "learning_rate": 8.942856383870905e-05, "loss": 3.1058, "step": 120200 }, { "epoch": 1.0619226761334535, "grad_norm": 1.6439554691314697, "learning_rate": 8.942001695274632e-05, "loss": 3.7622, "step": 120250 }, { "epoch": 1.0623642240237376, "grad_norm": 3.0835959911346436, "learning_rate": 8.941146702192154e-05, "loss": 3.4381, "step": 120300 }, { "epoch": 1.0628057719140218, "grad_norm": 1.632073163986206, "learning_rate": 8.940291404689513e-05, "loss": 2.7266, "step": 120350 }, { "epoch": 1.063247319804306, "grad_norm": 1.424759864807129, "learning_rate": 8.939435802832776e-05, "loss": 3.1694, "step": 120400 }, { "epoch": 1.0636888676945901, "grad_norm": 1.0503005981445312, "learning_rate": 8.938579896688027e-05, "loss": 3.2015, "step": 120450 }, { "epoch": 1.0641304155848743, "grad_norm": 3.014944314956665, "learning_rate": 8.93772368632138e-05, "loss": 3.1513, "step": 120500 }, { "epoch": 1.0645719634751585, "grad_norm": 4.032464981079102, "learning_rate": 8.936867171798968e-05, "loss": 3.2884, "step": 120550 }, { "epoch": 1.0650135113654426, "grad_norm": 1.6288483142852783, "learning_rate": 8.93601035318695e-05, "loss": 3.1507, "step": 120600 }, { "epoch": 1.0654550592557268, "grad_norm": 2.872183322906494, "learning_rate": 8.935153230551512e-05, "loss": 3.4061, "step": 120650 }, { "epoch": 1.065896607146011, "grad_norm": 0.9016225337982178, "learning_rate": 8.934295803958854e-05, "loss": 3.262, "step": 120700 }, { "epoch": 1.0663381550362951, "grad_norm": 3.425549030303955, "learning_rate": 8.933438073475206e-05, "loss": 3.2199, "step": 120750 }, { "epoch": 1.0667797029265795, "grad_norm": 1.639690637588501, "learning_rate": 8.932580039166823e-05, "loss": 3.2348, "step": 120800 }, { "epoch": 1.0672212508168637, "grad_norm": 1.3123271465301514, "learning_rate": 8.931721701099977e-05, "loss": 2.9831, "step": 120850 }, { "epoch": 1.0676627987071479, "grad_norm": 2.0268373489379883, "learning_rate": 8.930863059340973e-05, "loss": 3.5563, "step": 120900 }, { "epoch": 1.068104346597432, "grad_norm": 2.0526180267333984, "learning_rate": 8.930004113956127e-05, "loss": 3.1924, "step": 120950 }, { "epoch": 1.0685458944877162, "grad_norm": 1.7065848112106323, "learning_rate": 8.929144865011791e-05, "loss": 3.0529, "step": 121000 }, { "epoch": 1.0689874423780004, "grad_norm": 2.654390335083008, "learning_rate": 8.92828531257433e-05, "loss": 3.4986, "step": 121050 }, { "epoch": 1.0694289902682845, "grad_norm": 2.723402738571167, "learning_rate": 8.927425456710141e-05, "loss": 3.3786, "step": 121100 }, { "epoch": 1.0698705381585687, "grad_norm": 1.8961224555969238, "learning_rate": 8.92656529748564e-05, "loss": 3.4482, "step": 121150 }, { "epoch": 1.0703120860488529, "grad_norm": 2.6403260231018066, "learning_rate": 8.925704834967266e-05, "loss": 3.0291, "step": 121200 }, { "epoch": 1.070753633939137, "grad_norm": 2.748976469039917, "learning_rate": 8.924844069221481e-05, "loss": 2.7512, "step": 121250 }, { "epoch": 1.0711951818294212, "grad_norm": 3.6677534580230713, "learning_rate": 8.923983000314774e-05, "loss": 3.4528, "step": 121300 }, { "epoch": 1.0716367297197054, "grad_norm": 1.444061517715454, "learning_rate": 8.923121628313654e-05, "loss": 3.1281, "step": 121350 }, { "epoch": 1.0720782776099895, "grad_norm": 3.1170246601104736, "learning_rate": 8.922259953284658e-05, "loss": 3.3235, "step": 121400 }, { "epoch": 1.0725198255002737, "grad_norm": 2.0666680335998535, "learning_rate": 8.921397975294337e-05, "loss": 3.0791, "step": 121450 }, { "epoch": 1.0729613733905579, "grad_norm": 0.9159076809883118, "learning_rate": 8.920535694409276e-05, "loss": 3.1767, "step": 121500 }, { "epoch": 1.073402921280842, "grad_norm": 1.4360718727111816, "learning_rate": 8.919673110696078e-05, "loss": 3.1677, "step": 121550 }, { "epoch": 1.0738444691711262, "grad_norm": 6.167787551879883, "learning_rate": 8.91881022422137e-05, "loss": 2.988, "step": 121600 }, { "epoch": 1.0742860170614106, "grad_norm": 1.9026552438735962, "learning_rate": 8.917947035051804e-05, "loss": 3.4476, "step": 121650 }, { "epoch": 1.0747275649516947, "grad_norm": 2.0857155323028564, "learning_rate": 8.91708354325405e-05, "loss": 3.2904, "step": 121700 }, { "epoch": 1.075169112841979, "grad_norm": 2.137205123901367, "learning_rate": 8.916219748894811e-05, "loss": 3.1856, "step": 121750 }, { "epoch": 1.075610660732263, "grad_norm": 1.6799383163452148, "learning_rate": 8.915355652040804e-05, "loss": 3.3108, "step": 121800 }, { "epoch": 1.0760522086225472, "grad_norm": 2.878725290298462, "learning_rate": 8.914491252758773e-05, "loss": 3.303, "step": 121850 }, { "epoch": 1.0764937565128314, "grad_norm": 1.8924479484558105, "learning_rate": 8.913626551115489e-05, "loss": 2.9533, "step": 121900 }, { "epoch": 1.0769353044031156, "grad_norm": 3.5954272747039795, "learning_rate": 8.912761547177737e-05, "loss": 3.0536, "step": 121950 }, { "epoch": 1.0773768522933997, "grad_norm": 3.484809160232544, "learning_rate": 8.911896241012337e-05, "loss": 2.6292, "step": 122000 }, { "epoch": 1.077818400183684, "grad_norm": 2.250790596008301, "learning_rate": 8.91104794781339e-05, "loss": 3.3854, "step": 122050 }, { "epoch": 1.078259948073968, "grad_norm": 1.3487905263900757, "learning_rate": 8.910182043434451e-05, "loss": 3.3714, "step": 122100 }, { "epoch": 1.0787014959642522, "grad_norm": 1.1757347583770752, "learning_rate": 8.909315837027104e-05, "loss": 3.3735, "step": 122150 }, { "epoch": 1.0791430438545364, "grad_norm": 1.4942172765731812, "learning_rate": 8.908449328658259e-05, "loss": 3.4967, "step": 122200 }, { "epoch": 1.0795845917448206, "grad_norm": 1.2587921619415283, "learning_rate": 8.907582518394846e-05, "loss": 3.3902, "step": 122250 }, { "epoch": 1.0800261396351047, "grad_norm": 5.078591823577881, "learning_rate": 8.90671540630382e-05, "loss": 2.8449, "step": 122300 }, { "epoch": 1.080467687525389, "grad_norm": 1.2880104780197144, "learning_rate": 8.905847992452154e-05, "loss": 3.0292, "step": 122350 }, { "epoch": 1.0809092354156733, "grad_norm": 2.9822380542755127, "learning_rate": 8.904980276906854e-05, "loss": 3.5007, "step": 122400 }, { "epoch": 1.0813507833059575, "grad_norm": 1.5988084077835083, "learning_rate": 8.90411225973494e-05, "loss": 3.1964, "step": 122450 }, { "epoch": 1.0817923311962416, "grad_norm": 2.4721879959106445, "learning_rate": 8.90324394100346e-05, "loss": 3.3969, "step": 122500 }, { "epoch": 1.0822338790865258, "grad_norm": 2.2065553665161133, "learning_rate": 8.902375320779483e-05, "loss": 3.1513, "step": 122550 }, { "epoch": 1.08267542697681, "grad_norm": 4.793309688568115, "learning_rate": 8.901506399130104e-05, "loss": 3.1388, "step": 122600 }, { "epoch": 1.0831169748670941, "grad_norm": 2.873149871826172, "learning_rate": 8.900637176122439e-05, "loss": 3.1295, "step": 122650 }, { "epoch": 1.0835585227573783, "grad_norm": 1.5303065776824951, "learning_rate": 8.89976765182363e-05, "loss": 3.1349, "step": 122700 }, { "epoch": 1.0840000706476625, "grad_norm": 2.5381152629852295, "learning_rate": 8.89889782630084e-05, "loss": 2.9839, "step": 122750 }, { "epoch": 1.0844416185379466, "grad_norm": 1.606925368309021, "learning_rate": 8.898027699621253e-05, "loss": 2.9266, "step": 122800 }, { "epoch": 1.0848831664282308, "grad_norm": 2.0156993865966797, "learning_rate": 8.897157271852079e-05, "loss": 3.6531, "step": 122850 }, { "epoch": 1.085324714318515, "grad_norm": 2.879831552505493, "learning_rate": 8.896286543060555e-05, "loss": 3.1662, "step": 122900 }, { "epoch": 1.0857662622087991, "grad_norm": 1.4810205698013306, "learning_rate": 8.895415513313934e-05, "loss": 3.2344, "step": 122950 }, { "epoch": 1.0862078100990833, "grad_norm": 3.132847547531128, "learning_rate": 8.894544182679497e-05, "loss": 3.4795, "step": 123000 }, { "epoch": 1.0862078100990833, "eval_asr_loss": 0.9171379533371528, "eval_loss": 2.9432666301727295, "eval_runtime": 20.3376, "eval_samples_per_second": 37.763, "eval_steps_per_second": 9.441, "eval_tts_loss": 6.0161064579099435, "step": 123000 }, { "epoch": 1.0866493579893675, "grad_norm": 3.2920241355895996, "learning_rate": 8.893672551224547e-05, "loss": 3.0528, "step": 123050 }, { "epoch": 1.0870909058796516, "grad_norm": 3.9480741024017334, "learning_rate": 8.892800619016409e-05, "loss": 2.9215, "step": 123100 }, { "epoch": 1.0875324537699358, "grad_norm": 1.113074541091919, "learning_rate": 8.891928386122434e-05, "loss": 2.6988, "step": 123150 }, { "epoch": 1.08797400166022, "grad_norm": 3.499051094055176, "learning_rate": 8.891055852609992e-05, "loss": 3.4827, "step": 123200 }, { "epoch": 1.0884155495505043, "grad_norm": 3.23872971534729, "learning_rate": 8.890183018546483e-05, "loss": 3.1988, "step": 123250 }, { "epoch": 1.0888570974407885, "grad_norm": 2.035611391067505, "learning_rate": 8.889309883999321e-05, "loss": 3.414, "step": 123300 }, { "epoch": 1.0892986453310727, "grad_norm": 2.9164116382598877, "learning_rate": 8.888436449035955e-05, "loss": 2.8317, "step": 123350 }, { "epoch": 1.0897401932213568, "grad_norm": 2.3770909309387207, "learning_rate": 8.887562713723843e-05, "loss": 2.7964, "step": 123400 }, { "epoch": 1.090181741111641, "grad_norm": 1.5850526094436646, "learning_rate": 8.886688678130477e-05, "loss": 2.9763, "step": 123450 }, { "epoch": 1.0906232890019252, "grad_norm": 3.0644419193267822, "learning_rate": 8.885814342323369e-05, "loss": 3.0505, "step": 123500 }, { "epoch": 1.0910648368922093, "grad_norm": 4.172904968261719, "learning_rate": 8.884939706370055e-05, "loss": 3.088, "step": 123550 }, { "epoch": 1.0915063847824935, "grad_norm": 2.7594523429870605, "learning_rate": 8.884064770338092e-05, "loss": 3.414, "step": 123600 }, { "epoch": 1.0919479326727777, "grad_norm": 2.1696181297302246, "learning_rate": 8.88318953429506e-05, "loss": 3.0237, "step": 123650 }, { "epoch": 1.0923894805630618, "grad_norm": 2.1066668033599854, "learning_rate": 8.882313998308564e-05, "loss": 3.3296, "step": 123700 }, { "epoch": 1.092831028453346, "grad_norm": 2.821223020553589, "learning_rate": 8.881438162446236e-05, "loss": 2.9404, "step": 123750 }, { "epoch": 1.0932725763436302, "grad_norm": 1.817001223564148, "learning_rate": 8.880562026775721e-05, "loss": 2.9152, "step": 123800 }, { "epoch": 1.0937141242339143, "grad_norm": 1.3523433208465576, "learning_rate": 8.879685591364697e-05, "loss": 3.2203, "step": 123850 }, { "epoch": 1.0941556721241985, "grad_norm": 3.8314602375030518, "learning_rate": 8.87880885628086e-05, "loss": 2.6743, "step": 123900 }, { "epoch": 1.0945972200144827, "grad_norm": 5.9612321853637695, "learning_rate": 8.87793182159193e-05, "loss": 3.0251, "step": 123950 }, { "epoch": 1.095038767904767, "grad_norm": 9.792705535888672, "learning_rate": 8.87705448736565e-05, "loss": 3.019, "step": 124000 }, { "epoch": 1.0954803157950512, "grad_norm": 2.604074001312256, "learning_rate": 8.876176853669786e-05, "loss": 3.308, "step": 124050 }, { "epoch": 1.0959218636853354, "grad_norm": 5.136208534240723, "learning_rate": 8.87529892057213e-05, "loss": 3.1035, "step": 124100 }, { "epoch": 1.0963634115756196, "grad_norm": 3.2921555042266846, "learning_rate": 8.874438255722164e-05, "loss": 3.6894, "step": 124150 }, { "epoch": 1.0968049594659037, "grad_norm": 2.1269774436950684, "learning_rate": 8.87355973000904e-05, "loss": 3.0221, "step": 124200 }, { "epoch": 1.097246507356188, "grad_norm": 2.0114710330963135, "learning_rate": 8.872680905096278e-05, "loss": 3.4012, "step": 124250 }, { "epoch": 1.097688055246472, "grad_norm": 4.107577323913574, "learning_rate": 8.871801781051753e-05, "loss": 3.2551, "step": 124300 }, { "epoch": 1.0981296031367562, "grad_norm": 2.8425934314727783, "learning_rate": 8.870922357943373e-05, "loss": 3.2815, "step": 124350 }, { "epoch": 1.0985711510270404, "grad_norm": 2.750260829925537, "learning_rate": 8.870042635839065e-05, "loss": 2.9947, "step": 124400 }, { "epoch": 1.0990126989173246, "grad_norm": 2.1644296646118164, "learning_rate": 8.869162614806781e-05, "loss": 3.4007, "step": 124450 }, { "epoch": 1.0994542468076087, "grad_norm": 4.535878658294678, "learning_rate": 8.868282294914493e-05, "loss": 2.8193, "step": 124500 }, { "epoch": 1.099895794697893, "grad_norm": 0.610397458076477, "learning_rate": 8.867401676230202e-05, "loss": 3.3091, "step": 124550 }, { "epoch": 1.100337342588177, "grad_norm": 1.3316540718078613, "learning_rate": 8.866520758821926e-05, "loss": 2.9345, "step": 124600 }, { "epoch": 1.1007788904784612, "grad_norm": 2.5452449321746826, "learning_rate": 8.865639542757706e-05, "loss": 3.0348, "step": 124650 }, { "epoch": 1.1012204383687454, "grad_norm": 1.1717711687088013, "learning_rate": 8.864758028105614e-05, "loss": 3.6851, "step": 124700 }, { "epoch": 1.1016619862590296, "grad_norm": 1.1457619667053223, "learning_rate": 8.863876214933735e-05, "loss": 2.7013, "step": 124750 }, { "epoch": 1.1021035341493137, "grad_norm": 3.3232429027557373, "learning_rate": 8.862994103310183e-05, "loss": 3.4489, "step": 124800 }, { "epoch": 1.1025450820395981, "grad_norm": 2.6263210773468018, "learning_rate": 8.862111693303094e-05, "loss": 3.5465, "step": 124850 }, { "epoch": 1.1029866299298823, "grad_norm": 2.42629337310791, "learning_rate": 8.861228984980626e-05, "loss": 3.1175, "step": 124900 }, { "epoch": 1.1034281778201664, "grad_norm": 2.722221612930298, "learning_rate": 8.860345978410962e-05, "loss": 3.1893, "step": 124950 }, { "epoch": 1.1038697257104506, "grad_norm": 2.3193819522857666, "learning_rate": 8.859462673662305e-05, "loss": 3.6755, "step": 125000 }, { "epoch": 1.1043112736007348, "grad_norm": 2.0361616611480713, "learning_rate": 8.858579070802883e-05, "loss": 3.7692, "step": 125050 }, { "epoch": 1.104752821491019, "grad_norm": 2.403743028640747, "learning_rate": 8.85769516990095e-05, "loss": 3.03, "step": 125100 }, { "epoch": 1.105194369381303, "grad_norm": 6.095032691955566, "learning_rate": 8.856810971024776e-05, "loss": 3.4562, "step": 125150 }, { "epoch": 1.1056359172715873, "grad_norm": 3.28342866897583, "learning_rate": 8.85592647424266e-05, "loss": 3.3188, "step": 125200 }, { "epoch": 1.1060774651618714, "grad_norm": 2.5184452533721924, "learning_rate": 8.85504167962292e-05, "loss": 3.3513, "step": 125250 }, { "epoch": 1.1065190130521556, "grad_norm": 1.7279000282287598, "learning_rate": 8.854156587233902e-05, "loss": 3.5462, "step": 125300 }, { "epoch": 1.1069605609424398, "grad_norm": 2.6725499629974365, "learning_rate": 8.853271197143969e-05, "loss": 2.5798, "step": 125350 }, { "epoch": 1.107402108832724, "grad_norm": 1.1816620826721191, "learning_rate": 8.85238550942151e-05, "loss": 3.5221, "step": 125400 }, { "epoch": 1.107843656723008, "grad_norm": 1.639984369277954, "learning_rate": 8.851499524134939e-05, "loss": 2.9796, "step": 125450 }, { "epoch": 1.1082852046132923, "grad_norm": 5.580627918243408, "learning_rate": 8.850613241352688e-05, "loss": 3.2337, "step": 125500 }, { "epoch": 1.1087267525035764, "grad_norm": 4.015604019165039, "learning_rate": 8.849726661143217e-05, "loss": 3.013, "step": 125550 }, { "epoch": 1.1091683003938608, "grad_norm": 5.388280391693115, "learning_rate": 8.848839783575007e-05, "loss": 3.2899, "step": 125600 }, { "epoch": 1.109609848284145, "grad_norm": 1.1181635856628418, "learning_rate": 8.847952608716559e-05, "loss": 3.2404, "step": 125650 }, { "epoch": 1.1100513961744292, "grad_norm": 1.2967584133148193, "learning_rate": 8.847065136636403e-05, "loss": 3.1253, "step": 125700 }, { "epoch": 1.1104929440647133, "grad_norm": 2.1886446475982666, "learning_rate": 8.846177367403088e-05, "loss": 3.3288, "step": 125750 }, { "epoch": 1.1109344919549975, "grad_norm": 2.3884706497192383, "learning_rate": 8.845289301085184e-05, "loss": 3.7135, "step": 125800 }, { "epoch": 1.1113760398452817, "grad_norm": 1.3103140592575073, "learning_rate": 8.844400937751291e-05, "loss": 2.7412, "step": 125850 }, { "epoch": 1.1118175877355658, "grad_norm": 2.5914974212646484, "learning_rate": 8.843512277470023e-05, "loss": 2.9458, "step": 125900 }, { "epoch": 1.11225913562585, "grad_norm": 1.5716588497161865, "learning_rate": 8.842623320310023e-05, "loss": 3.558, "step": 125950 }, { "epoch": 1.1127006835161342, "grad_norm": 2.2623939514160156, "learning_rate": 8.841734066339959e-05, "loss": 3.2775, "step": 126000 }, { "epoch": 1.1127006835161342, "eval_asr_loss": 0.9239404375777175, "eval_loss": 2.9568049907684326, "eval_runtime": 20.3306, "eval_samples_per_second": 37.776, "eval_steps_per_second": 9.444, "eval_tts_loss": 6.011922956928332, "step": 126000 }, { "epoch": 1.1131422314064183, "grad_norm": 1.213431477546692, "learning_rate": 8.840844515628515e-05, "loss": 3.3045, "step": 126050 }, { "epoch": 1.1135837792967025, "grad_norm": 1.4413515329360962, "learning_rate": 8.839954668244399e-05, "loss": 3.3209, "step": 126100 }, { "epoch": 1.1140253271869867, "grad_norm": 0.6168299913406372, "learning_rate": 8.839064524256348e-05, "loss": 3.7805, "step": 126150 }, { "epoch": 1.1144668750772708, "grad_norm": 1.9081262350082397, "learning_rate": 8.838174083733117e-05, "loss": 3.1688, "step": 126200 }, { "epoch": 1.114908422967555, "grad_norm": 1.6664979457855225, "learning_rate": 8.837283346743485e-05, "loss": 3.2032, "step": 126250 }, { "epoch": 1.1153499708578392, "grad_norm": 5.024023532867432, "learning_rate": 8.836392313356255e-05, "loss": 3.2986, "step": 126300 }, { "epoch": 1.1157915187481233, "grad_norm": 1.7272241115570068, "learning_rate": 8.835500983640248e-05, "loss": 2.9631, "step": 126350 }, { "epoch": 1.1162330666384075, "grad_norm": 4.159533977508545, "learning_rate": 8.834609357664314e-05, "loss": 3.1184, "step": 126400 }, { "epoch": 1.1166746145286919, "grad_norm": 1.9124999046325684, "learning_rate": 8.833717435497325e-05, "loss": 3.4903, "step": 126450 }, { "epoch": 1.117116162418976, "grad_norm": 3.8171277046203613, "learning_rate": 8.832825217208174e-05, "loss": 2.9298, "step": 126500 }, { "epoch": 1.1175577103092602, "grad_norm": 3.396512508392334, "learning_rate": 8.831932702865774e-05, "loss": 3.398, "step": 126550 }, { "epoch": 1.1179992581995444, "grad_norm": 0.947153627872467, "learning_rate": 8.831039892539069e-05, "loss": 3.2457, "step": 126600 }, { "epoch": 1.1184408060898285, "grad_norm": 5.9848456382751465, "learning_rate": 8.830146786297016e-05, "loss": 2.9657, "step": 126650 }, { "epoch": 1.1188823539801127, "grad_norm": 1.9373352527618408, "learning_rate": 8.829253384208604e-05, "loss": 3.4747, "step": 126700 }, { "epoch": 1.1193239018703969, "grad_norm": 2.3222639560699463, "learning_rate": 8.828359686342839e-05, "loss": 3.1783, "step": 126750 }, { "epoch": 1.119765449760681, "grad_norm": 1.1654438972473145, "learning_rate": 8.82746569276875e-05, "loss": 3.1918, "step": 126800 }, { "epoch": 1.1202069976509652, "grad_norm": 0.7595392465591431, "learning_rate": 8.826571403555395e-05, "loss": 3.2987, "step": 126850 }, { "epoch": 1.1206485455412494, "grad_norm": 3.6976282596588135, "learning_rate": 8.825676818771846e-05, "loss": 3.1541, "step": 126900 }, { "epoch": 1.1210900934315335, "grad_norm": 4.80150842666626, "learning_rate": 8.824799838988361e-05, "loss": 2.8695, "step": 126950 }, { "epoch": 1.1215316413218177, "grad_norm": 2.4069314002990723, "learning_rate": 8.823904669179708e-05, "loss": 3.416, "step": 127000 }, { "epoch": 1.1219731892121019, "grad_norm": 3.3130557537078857, "learning_rate": 8.823009204006847e-05, "loss": 3.0622, "step": 127050 }, { "epoch": 1.122414737102386, "grad_norm": 4.023521900177002, "learning_rate": 8.822113443538942e-05, "loss": 3.5856, "step": 127100 }, { "epoch": 1.1228562849926702, "grad_norm": 2.4293367862701416, "learning_rate": 8.821217387845184e-05, "loss": 3.1485, "step": 127150 }, { "epoch": 1.1232978328829546, "grad_norm": 1.6036757230758667, "learning_rate": 8.820321036994787e-05, "loss": 3.402, "step": 127200 }, { "epoch": 1.1237393807732388, "grad_norm": 5.501082420349121, "learning_rate": 8.819424391056985e-05, "loss": 3.4117, "step": 127250 }, { "epoch": 1.124180928663523, "grad_norm": 0.864111065864563, "learning_rate": 8.818527450101035e-05, "loss": 2.8882, "step": 127300 }, { "epoch": 1.124622476553807, "grad_norm": 1.7957173585891724, "learning_rate": 8.817630214196222e-05, "loss": 3.1329, "step": 127350 }, { "epoch": 1.1250640244440913, "grad_norm": 2.793177604675293, "learning_rate": 8.816732683411846e-05, "loss": 3.602, "step": 127400 }, { "epoch": 1.1255055723343754, "grad_norm": 3.6395950317382812, "learning_rate": 8.815834857817234e-05, "loss": 3.6478, "step": 127450 }, { "epoch": 1.1259471202246596, "grad_norm": 3.0564305782318115, "learning_rate": 8.814936737481739e-05, "loss": 3.1558, "step": 127500 }, { "epoch": 1.1263886681149438, "grad_norm": 2.4412167072296143, "learning_rate": 8.81403832247473e-05, "loss": 3.0928, "step": 127550 }, { "epoch": 1.126830216005228, "grad_norm": 0.747972309589386, "learning_rate": 8.813139612865601e-05, "loss": 3.1837, "step": 127600 }, { "epoch": 1.127271763895512, "grad_norm": 1.7288129329681396, "learning_rate": 8.812240608723774e-05, "loss": 3.0548, "step": 127650 }, { "epoch": 1.1277133117857963, "grad_norm": 2.70778226852417, "learning_rate": 8.811359298976078e-05, "loss": 3.1048, "step": 127700 }, { "epoch": 1.1281548596760804, "grad_norm": 2.5320427417755127, "learning_rate": 8.810459711864387e-05, "loss": 3.8118, "step": 127750 }, { "epoch": 1.1285964075663646, "grad_norm": 1.2388442754745483, "learning_rate": 8.809559830426997e-05, "loss": 3.3423, "step": 127800 }, { "epoch": 1.1290379554566488, "grad_norm": 1.5890963077545166, "learning_rate": 8.808659654733411e-05, "loss": 2.7988, "step": 127850 }, { "epoch": 1.129479503346933, "grad_norm": 3.5289838314056396, "learning_rate": 8.807759184853165e-05, "loss": 3.3222, "step": 127900 }, { "epoch": 1.129921051237217, "grad_norm": 2.6470303535461426, "learning_rate": 8.806858420855813e-05, "loss": 2.5924, "step": 127950 }, { "epoch": 1.1303625991275013, "grad_norm": 1.1539044380187988, "learning_rate": 8.805957362810926e-05, "loss": 3.5027, "step": 128000 }, { "epoch": 1.1308041470177856, "grad_norm": 2.7577285766601562, "learning_rate": 8.805056010788108e-05, "loss": 2.8623, "step": 128050 }, { "epoch": 1.1312456949080698, "grad_norm": 3.473392963409424, "learning_rate": 8.804154364856979e-05, "loss": 3.6265, "step": 128100 }, { "epoch": 1.131687242798354, "grad_norm": 1.1800062656402588, "learning_rate": 8.803252425087183e-05, "loss": 3.161, "step": 128150 }, { "epoch": 1.1321287906886381, "grad_norm": 3.8547286987304688, "learning_rate": 8.802350191548387e-05, "loss": 3.247, "step": 128200 }, { "epoch": 1.1325703385789223, "grad_norm": 2.5325534343719482, "learning_rate": 8.801447664310282e-05, "loss": 3.2594, "step": 128250 }, { "epoch": 1.1330118864692065, "grad_norm": 2.247164726257324, "learning_rate": 8.800544843442582e-05, "loss": 3.6423, "step": 128300 }, { "epoch": 1.1334534343594906, "grad_norm": 4.375690460205078, "learning_rate": 8.799641729015019e-05, "loss": 3.6558, "step": 128350 }, { "epoch": 1.1338949822497748, "grad_norm": 3.372756242752075, "learning_rate": 8.798738321097353e-05, "loss": 3.1197, "step": 128400 }, { "epoch": 1.134336530140059, "grad_norm": 3.348130464553833, "learning_rate": 8.797834619759363e-05, "loss": 2.9325, "step": 128450 }, { "epoch": 1.1347780780303431, "grad_norm": 0.7869385480880737, "learning_rate": 8.796930625070853e-05, "loss": 3.1624, "step": 128500 }, { "epoch": 1.1352196259206273, "grad_norm": 2.3363020420074463, "learning_rate": 8.796026337101653e-05, "loss": 3.1033, "step": 128550 }, { "epoch": 1.1356611738109115, "grad_norm": 1.4216840267181396, "learning_rate": 8.795121755921604e-05, "loss": 3.6407, "step": 128600 }, { "epoch": 1.1361027217011956, "grad_norm": 1.4757707118988037, "learning_rate": 8.794216881600583e-05, "loss": 3.5712, "step": 128650 }, { "epoch": 1.1365442695914798, "grad_norm": 1.487884283065796, "learning_rate": 8.793311714208481e-05, "loss": 3.1033, "step": 128700 }, { "epoch": 1.1369858174817642, "grad_norm": 3.3556032180786133, "learning_rate": 8.792406253815215e-05, "loss": 2.8425, "step": 128750 }, { "epoch": 1.1374273653720484, "grad_norm": 1.508814811706543, "learning_rate": 8.791500500490726e-05, "loss": 3.3935, "step": 128800 }, { "epoch": 1.1378689132623325, "grad_norm": 1.8587937355041504, "learning_rate": 8.790594454304974e-05, "loss": 3.4744, "step": 128850 }, { "epoch": 1.1383104611526167, "grad_norm": 0.7840842604637146, "learning_rate": 8.789688115327944e-05, "loss": 3.2839, "step": 128900 }, { "epoch": 1.1387520090429009, "grad_norm": 3.428917646408081, "learning_rate": 8.78878148362964e-05, "loss": 3.228, "step": 128950 }, { "epoch": 1.139193556933185, "grad_norm": 1.454732894897461, "learning_rate": 8.787874559280096e-05, "loss": 2.5992, "step": 129000 }, { "epoch": 1.139193556933185, "eval_asr_loss": 0.9056011822879443, "eval_loss": 2.9360103607177734, "eval_runtime": 20.5894, "eval_samples_per_second": 37.301, "eval_steps_per_second": 9.325, "eval_tts_loss": 6.019627850743963, "step": 129000 }, { "epoch": 1.1396351048234692, "grad_norm": 3.9071409702301025, "learning_rate": 8.786967342349364e-05, "loss": 2.7655, "step": 129050 }, { "epoch": 1.1400766527137534, "grad_norm": 1.2048531770706177, "learning_rate": 8.786059832907516e-05, "loss": 2.8011, "step": 129100 }, { "epoch": 1.1405182006040375, "grad_norm": 1.1813719272613525, "learning_rate": 8.78515203102465e-05, "loss": 3.2859, "step": 129150 }, { "epoch": 1.1409597484943217, "grad_norm": 5.590451717376709, "learning_rate": 8.784243936770889e-05, "loss": 3.3888, "step": 129200 }, { "epoch": 1.1414012963846059, "grad_norm": 1.4032896757125854, "learning_rate": 8.783335550216372e-05, "loss": 3.2597, "step": 129250 }, { "epoch": 1.14184284427489, "grad_norm": 1.308907151222229, "learning_rate": 8.782426871431265e-05, "loss": 3.2236, "step": 129300 }, { "epoch": 1.1422843921651742, "grad_norm": 4.8072028160095215, "learning_rate": 8.781517900485757e-05, "loss": 2.996, "step": 129350 }, { "epoch": 1.1427259400554584, "grad_norm": 1.4444926977157593, "learning_rate": 8.780608637450056e-05, "loss": 2.6338, "step": 129400 }, { "epoch": 1.1431674879457425, "grad_norm": 4.736634731292725, "learning_rate": 8.7796990823944e-05, "loss": 3.0496, "step": 129450 }, { "epoch": 1.1436090358360267, "grad_norm": 1.504779577255249, "learning_rate": 8.778789235389037e-05, "loss": 3.0389, "step": 129500 }, { "epoch": 1.1440505837263109, "grad_norm": 3.4063234329223633, "learning_rate": 8.77787909650425e-05, "loss": 3.2908, "step": 129550 }, { "epoch": 1.144492131616595, "grad_norm": 2.3659403324127197, "learning_rate": 8.776968665810339e-05, "loss": 3.567, "step": 129600 }, { "epoch": 1.1449336795068794, "grad_norm": 2.675527334213257, "learning_rate": 8.776057943377624e-05, "loss": 3.2116, "step": 129650 }, { "epoch": 1.1453752273971636, "grad_norm": 2.0578064918518066, "learning_rate": 8.775146929276457e-05, "loss": 3.227, "step": 129700 }, { "epoch": 1.1458167752874477, "grad_norm": 3.0254499912261963, "learning_rate": 8.774235623577199e-05, "loss": 2.9167, "step": 129750 }, { "epoch": 1.146258323177732, "grad_norm": 1.1867756843566895, "learning_rate": 8.773324026350245e-05, "loss": 3.1534, "step": 129800 }, { "epoch": 1.146699871068016, "grad_norm": 0.8575452566146851, "learning_rate": 8.772412137666005e-05, "loss": 3.04, "step": 129850 }, { "epoch": 1.1471414189583002, "grad_norm": 1.0062538385391235, "learning_rate": 8.771499957594917e-05, "loss": 3.2266, "step": 129900 }, { "epoch": 1.1475829668485844, "grad_norm": 2.8599894046783447, "learning_rate": 8.77058748620744e-05, "loss": 3.3753, "step": 129950 }, { "epoch": 1.1480245147388686, "grad_norm": 2.6259355545043945, "learning_rate": 8.769674723574052e-05, "loss": 3.4298, "step": 130000 }, { "epoch": 1.1484660626291527, "grad_norm": 0.4966769516468048, "learning_rate": 8.76876166976526e-05, "loss": 3.2482, "step": 130050 }, { "epoch": 1.148907610519437, "grad_norm": 4.237811088562012, "learning_rate": 8.767848324851584e-05, "loss": 3.8295, "step": 130100 }, { "epoch": 1.149349158409721, "grad_norm": 1.2462416887283325, "learning_rate": 8.766934688903577e-05, "loss": 2.9447, "step": 130150 }, { "epoch": 1.1497907063000052, "grad_norm": 4.463748455047607, "learning_rate": 8.766020761991808e-05, "loss": 3.1586, "step": 130200 }, { "epoch": 1.1502322541902894, "grad_norm": 1.7702147960662842, "learning_rate": 8.76510654418687e-05, "loss": 3.2017, "step": 130250 }, { "epoch": 1.1506738020805736, "grad_norm": 1.8763153553009033, "learning_rate": 8.764192035559378e-05, "loss": 3.2077, "step": 130300 }, { "epoch": 1.151115349970858, "grad_norm": 1.445235013961792, "learning_rate": 8.763277236179971e-05, "loss": 2.9795, "step": 130350 }, { "epoch": 1.1515568978611421, "grad_norm": 3.101717233657837, "learning_rate": 8.762362146119309e-05, "loss": 3.5737, "step": 130400 }, { "epoch": 1.1519984457514263, "grad_norm": 2.088386297225952, "learning_rate": 8.761446765448076e-05, "loss": 3.071, "step": 130450 }, { "epoch": 1.1524399936417105, "grad_norm": 1.0259060859680176, "learning_rate": 8.760531094236975e-05, "loss": 3.0661, "step": 130500 }, { "epoch": 1.1528815415319946, "grad_norm": 1.4046655893325806, "learning_rate": 8.759615132556736e-05, "loss": 2.9523, "step": 130550 }, { "epoch": 1.1533230894222788, "grad_norm": 2.649925947189331, "learning_rate": 8.75869888047811e-05, "loss": 3.2149, "step": 130600 }, { "epoch": 1.153764637312563, "grad_norm": 3.931152582168579, "learning_rate": 8.757782338071866e-05, "loss": 3.5996, "step": 130650 }, { "epoch": 1.1542061852028471, "grad_norm": 0.7651322484016418, "learning_rate": 8.756865505408802e-05, "loss": 2.8727, "step": 130700 }, { "epoch": 1.1546477330931313, "grad_norm": 3.1403989791870117, "learning_rate": 8.755948382559735e-05, "loss": 3.5253, "step": 130750 }, { "epoch": 1.1550892809834155, "grad_norm": 0.9329702854156494, "learning_rate": 8.755030969595505e-05, "loss": 2.9106, "step": 130800 }, { "epoch": 1.1555308288736996, "grad_norm": 1.3300598859786987, "learning_rate": 8.754113266586977e-05, "loss": 3.4129, "step": 130850 }, { "epoch": 1.1559723767639838, "grad_norm": 4.641824722290039, "learning_rate": 8.75319527360503e-05, "loss": 3.3374, "step": 130900 }, { "epoch": 1.156413924654268, "grad_norm": 2.1139044761657715, "learning_rate": 8.752276990720576e-05, "loss": 3.4172, "step": 130950 }, { "epoch": 1.1568554725445521, "grad_norm": 1.589998483657837, "learning_rate": 8.751358418004542e-05, "loss": 2.8262, "step": 131000 }, { "epoch": 1.1572970204348363, "grad_norm": 1.7650665044784546, "learning_rate": 8.75043955552788e-05, "loss": 3.6119, "step": 131050 }, { "epoch": 1.1577385683251205, "grad_norm": 4.2411603927612305, "learning_rate": 8.749520403361566e-05, "loss": 3.1808, "step": 131100 }, { "epoch": 1.1581801162154046, "grad_norm": 1.8814787864685059, "learning_rate": 8.748600961576596e-05, "loss": 2.9927, "step": 131150 }, { "epoch": 1.1586216641056888, "grad_norm": 2.2863402366638184, "learning_rate": 8.74768123024399e-05, "loss": 3.0464, "step": 131200 }, { "epoch": 1.1590632119959732, "grad_norm": 1.000219464302063, "learning_rate": 8.746761209434786e-05, "loss": 3.3005, "step": 131250 }, { "epoch": 1.1595047598862573, "grad_norm": 3.1498961448669434, "learning_rate": 8.745840899220051e-05, "loss": 2.8365, "step": 131300 }, { "epoch": 1.1599463077765415, "grad_norm": 6.975586891174316, "learning_rate": 8.744920299670871e-05, "loss": 3.2125, "step": 131350 }, { "epoch": 1.1603878556668257, "grad_norm": 1.3487223386764526, "learning_rate": 8.743999410858354e-05, "loss": 3.1216, "step": 131400 }, { "epoch": 1.1608294035571098, "grad_norm": 2.6848034858703613, "learning_rate": 8.74307823285363e-05, "loss": 2.8299, "step": 131450 }, { "epoch": 1.161270951447394, "grad_norm": 3.3084640502929688, "learning_rate": 8.742156765727853e-05, "loss": 3.5291, "step": 131500 }, { "epoch": 1.1617124993376782, "grad_norm": 3.785712480545044, "learning_rate": 8.741235009552197e-05, "loss": 3.1219, "step": 131550 }, { "epoch": 1.1621540472279623, "grad_norm": 2.4641273021698, "learning_rate": 8.740312964397863e-05, "loss": 3.6626, "step": 131600 }, { "epoch": 1.1625955951182465, "grad_norm": 2.6847565174102783, "learning_rate": 8.73939063033607e-05, "loss": 3.425, "step": 131650 }, { "epoch": 1.1630371430085307, "grad_norm": 3.19730281829834, "learning_rate": 8.738468007438059e-05, "loss": 3.5178, "step": 131700 }, { "epoch": 1.1634786908988148, "grad_norm": 1.92465341091156, "learning_rate": 8.737545095775094e-05, "loss": 3.1112, "step": 131750 }, { "epoch": 1.163920238789099, "grad_norm": 3.2829654216766357, "learning_rate": 8.736621895418467e-05, "loss": 3.4692, "step": 131800 }, { "epoch": 1.1643617866793832, "grad_norm": 3.4545881748199463, "learning_rate": 8.7357168790471e-05, "loss": 2.9955, "step": 131850 }, { "epoch": 1.1648033345696673, "grad_norm": 1.9514338970184326, "learning_rate": 8.734793107287413e-05, "loss": 2.8255, "step": 131900 }, { "epoch": 1.1652448824599517, "grad_norm": 2.7962305545806885, "learning_rate": 8.73386904704663e-05, "loss": 3.2173, "step": 131950 }, { "epoch": 1.165686430350236, "grad_norm": 4.475177764892578, "learning_rate": 8.732963188195087e-05, "loss": 3.1302, "step": 132000 }, { "epoch": 1.165686430350236, "eval_asr_loss": 0.9138820392081759, "eval_loss": 2.9405391216278076, "eval_runtime": 20.3253, "eval_samples_per_second": 37.785, "eval_steps_per_second": 9.446, "eval_tts_loss": 5.998060327010354, "step": 132000 }, { "epoch": 1.16612797824052, "grad_norm": 6.156948089599609, "learning_rate": 8.732038556972324e-05, "loss": 2.9282, "step": 132050 }, { "epoch": 1.1665695261308042, "grad_norm": 3.0831196308135986, "learning_rate": 8.731113637481229e-05, "loss": 3.2877, "step": 132100 }, { "epoch": 1.1670110740210884, "grad_norm": 2.766617774963379, "learning_rate": 8.730188429793244e-05, "loss": 3.6187, "step": 132150 }, { "epoch": 1.1674526219113726, "grad_norm": 2.436147451400757, "learning_rate": 8.729262933979835e-05, "loss": 3.2395, "step": 132200 }, { "epoch": 1.1678941698016567, "grad_norm": 7.238566875457764, "learning_rate": 8.728337150112486e-05, "loss": 3.2521, "step": 132250 }, { "epoch": 1.168335717691941, "grad_norm": 1.9058992862701416, "learning_rate": 8.727411078262711e-05, "loss": 3.2199, "step": 132300 }, { "epoch": 1.168777265582225, "grad_norm": 1.8208065032958984, "learning_rate": 8.726484718502035e-05, "loss": 3.5917, "step": 132350 }, { "epoch": 1.1692188134725092, "grad_norm": 2.439476490020752, "learning_rate": 8.725558070902014e-05, "loss": 3.2484, "step": 132400 }, { "epoch": 1.1696603613627934, "grad_norm": 2.0772078037261963, "learning_rate": 8.724631135534225e-05, "loss": 3.0087, "step": 132450 }, { "epoch": 1.1701019092530776, "grad_norm": 1.8260692358016968, "learning_rate": 8.723703912470264e-05, "loss": 3.1512, "step": 132500 }, { "epoch": 1.1705434571433617, "grad_norm": 2.0503768920898438, "learning_rate": 8.722776401781751e-05, "loss": 3.2767, "step": 132550 }, { "epoch": 1.1709850050336459, "grad_norm": 1.1883853673934937, "learning_rate": 8.721848603540331e-05, "loss": 3.66, "step": 132600 }, { "epoch": 1.17142655292393, "grad_norm": 3.468325138092041, "learning_rate": 8.720920517817665e-05, "loss": 2.6868, "step": 132650 }, { "epoch": 1.1718681008142142, "grad_norm": 2.5057759284973145, "learning_rate": 8.719992144685442e-05, "loss": 3.3156, "step": 132700 }, { "epoch": 1.1723096487044984, "grad_norm": 2.8704116344451904, "learning_rate": 8.719063484215372e-05, "loss": 3.309, "step": 132750 }, { "epoch": 1.1727511965947826, "grad_norm": 2.1519346237182617, "learning_rate": 8.718134536479184e-05, "loss": 3.0556, "step": 132800 }, { "epoch": 1.173192744485067, "grad_norm": 1.5782400369644165, "learning_rate": 8.717205301548631e-05, "loss": 3.1093, "step": 132850 }, { "epoch": 1.173634292375351, "grad_norm": 3.0821375846862793, "learning_rate": 8.71627577949549e-05, "loss": 3.0993, "step": 132900 }, { "epoch": 1.1740758402656353, "grad_norm": 1.1189353466033936, "learning_rate": 8.715345970391557e-05, "loss": 2.7387, "step": 132950 }, { "epoch": 1.1745173881559194, "grad_norm": 2.6682019233703613, "learning_rate": 8.714415874308655e-05, "loss": 2.7153, "step": 133000 }, { "epoch": 1.1749589360462036, "grad_norm": 1.7515902519226074, "learning_rate": 8.713485491318622e-05, "loss": 3.4875, "step": 133050 }, { "epoch": 1.1754004839364878, "grad_norm": 1.7068076133728027, "learning_rate": 8.712554821493326e-05, "loss": 3.2121, "step": 133100 }, { "epoch": 1.175842031826772, "grad_norm": 3.6428372859954834, "learning_rate": 8.71162386490465e-05, "loss": 3.0579, "step": 133150 }, { "epoch": 1.176283579717056, "grad_norm": 0.455516517162323, "learning_rate": 8.710692621624506e-05, "loss": 2.8844, "step": 133200 }, { "epoch": 1.1767251276073403, "grad_norm": 1.2548800706863403, "learning_rate": 8.709761091724821e-05, "loss": 2.8791, "step": 133250 }, { "epoch": 1.1771666754976244, "grad_norm": 1.0584521293640137, "learning_rate": 8.70882927527755e-05, "loss": 3.4511, "step": 133300 }, { "epoch": 1.1776082233879086, "grad_norm": 0.7974949479103088, "learning_rate": 8.707897172354666e-05, "loss": 3.153, "step": 133350 }, { "epoch": 1.1780497712781928, "grad_norm": 2.0270535945892334, "learning_rate": 8.70696478302817e-05, "loss": 3.2497, "step": 133400 }, { "epoch": 1.178491319168477, "grad_norm": 1.8549659252166748, "learning_rate": 8.706032107370079e-05, "loss": 3.1353, "step": 133450 }, { "epoch": 1.178932867058761, "grad_norm": 3.472029685974121, "learning_rate": 8.705099145452432e-05, "loss": 3.4737, "step": 133500 }, { "epoch": 1.1793744149490455, "grad_norm": 2.2693088054656982, "learning_rate": 8.704165897347294e-05, "loss": 2.7868, "step": 133550 }, { "epoch": 1.1798159628393297, "grad_norm": 1.4861036539077759, "learning_rate": 8.703232363126753e-05, "loss": 3.4654, "step": 133600 }, { "epoch": 1.1802575107296138, "grad_norm": 1.4503135681152344, "learning_rate": 8.702298542862913e-05, "loss": 3.1721, "step": 133650 }, { "epoch": 1.180699058619898, "grad_norm": 1.2555351257324219, "learning_rate": 8.701364436627906e-05, "loss": 2.4155, "step": 133700 }, { "epoch": 1.1811406065101822, "grad_norm": 2.5926589965820312, "learning_rate": 8.700430044493881e-05, "loss": 3.0593, "step": 133750 }, { "epoch": 1.1815821544004663, "grad_norm": 1.0408382415771484, "learning_rate": 8.699495366533015e-05, "loss": 2.9217, "step": 133800 }, { "epoch": 1.1820237022907505, "grad_norm": 3.3952131271362305, "learning_rate": 8.698560402817503e-05, "loss": 3.1385, "step": 133850 }, { "epoch": 1.1824652501810347, "grad_norm": 2.9611570835113525, "learning_rate": 8.697625153419563e-05, "loss": 3.0722, "step": 133900 }, { "epoch": 1.1829067980713188, "grad_norm": 3.736604690551758, "learning_rate": 8.696689618411434e-05, "loss": 3.0413, "step": 133950 }, { "epoch": 1.183348345961603, "grad_norm": 3.9566078186035156, "learning_rate": 8.69575379786538e-05, "loss": 3.11, "step": 134000 }, { "epoch": 1.1837898938518872, "grad_norm": 3.8381118774414062, "learning_rate": 8.694817691853682e-05, "loss": 2.9578, "step": 134050 }, { "epoch": 1.1842314417421713, "grad_norm": 1.4392008781433105, "learning_rate": 8.693881300448651e-05, "loss": 3.1636, "step": 134100 }, { "epoch": 1.1846729896324555, "grad_norm": 2.4134435653686523, "learning_rate": 8.692944623722612e-05, "loss": 3.2271, "step": 134150 }, { "epoch": 1.1851145375227397, "grad_norm": 3.1550133228302, "learning_rate": 8.692007661747917e-05, "loss": 3.7818, "step": 134200 }, { "epoch": 1.1855560854130238, "grad_norm": 1.9582149982452393, "learning_rate": 8.691070414596936e-05, "loss": 3.145, "step": 134250 }, { "epoch": 1.185997633303308, "grad_norm": 1.6500295400619507, "learning_rate": 8.690132882342064e-05, "loss": 3.1886, "step": 134300 }, { "epoch": 1.1864391811935922, "grad_norm": 3.904228687286377, "learning_rate": 8.68919506505572e-05, "loss": 3.3516, "step": 134350 }, { "epoch": 1.1868807290838763, "grad_norm": 3.5070960521698, "learning_rate": 8.688256962810339e-05, "loss": 3.1526, "step": 134400 }, { "epoch": 1.1873222769741607, "grad_norm": 3.185276508331299, "learning_rate": 8.687318575678385e-05, "loss": 3.4449, "step": 134450 }, { "epoch": 1.1877638248644449, "grad_norm": 3.7565488815307617, "learning_rate": 8.686379903732338e-05, "loss": 3.4953, "step": 134500 }, { "epoch": 1.188205372754729, "grad_norm": 0.8475595116615295, "learning_rate": 8.685440947044703e-05, "loss": 3.0571, "step": 134550 }, { "epoch": 1.1886469206450132, "grad_norm": 4.451568603515625, "learning_rate": 8.684501705688006e-05, "loss": 3.0146, "step": 134600 }, { "epoch": 1.1890884685352974, "grad_norm": 1.2072436809539795, "learning_rate": 8.683562179734796e-05, "loss": 2.8346, "step": 134650 }, { "epoch": 1.1895300164255815, "grad_norm": 1.3233250379562378, "learning_rate": 8.682622369257644e-05, "loss": 2.8381, "step": 134700 }, { "epoch": 1.1899715643158657, "grad_norm": 1.7606003284454346, "learning_rate": 8.681682274329141e-05, "loss": 3.3425, "step": 134750 }, { "epoch": 1.1904131122061499, "grad_norm": 1.8309063911437988, "learning_rate": 8.680741895021902e-05, "loss": 3.1333, "step": 134800 }, { "epoch": 1.190854660096434, "grad_norm": 5.641381740570068, "learning_rate": 8.679801231408564e-05, "loss": 3.0138, "step": 134850 }, { "epoch": 1.1912962079867182, "grad_norm": 5.100592136383057, "learning_rate": 8.678860283561783e-05, "loss": 3.3786, "step": 134900 }, { "epoch": 1.1917377558770024, "grad_norm": 1.4024208784103394, "learning_rate": 8.677919051554245e-05, "loss": 3.2306, "step": 134950 }, { "epoch": 1.1921793037672865, "grad_norm": 3.9913697242736816, "learning_rate": 8.676977535458644e-05, "loss": 2.5345, "step": 135000 }, { "epoch": 1.1921793037672865, "eval_asr_loss": 0.9235845024814252, "eval_loss": 2.937715768814087, "eval_runtime": 20.5852, "eval_samples_per_second": 37.308, "eval_steps_per_second": 9.327, "eval_tts_loss": 5.992201569014508, "step": 135000 }, { "epoch": 1.1926208516575707, "grad_norm": 3.502964735031128, "learning_rate": 8.67603573534771e-05, "loss": 3.4259, "step": 135050 }, { "epoch": 1.1930623995478549, "grad_norm": 2.7115063667297363, "learning_rate": 8.675093651294186e-05, "loss": 3.2835, "step": 135100 }, { "epoch": 1.1935039474381393, "grad_norm": 2.064483642578125, "learning_rate": 8.674151283370842e-05, "loss": 3.294, "step": 135150 }, { "epoch": 1.1939454953284234, "grad_norm": 1.643545389175415, "learning_rate": 8.673208631650467e-05, "loss": 3.5124, "step": 135200 }, { "epoch": 1.1943870432187076, "grad_norm": 3.38920521736145, "learning_rate": 8.672265696205874e-05, "loss": 3.3957, "step": 135250 }, { "epoch": 1.1948285911089918, "grad_norm": 2.7786691188812256, "learning_rate": 8.671322477109896e-05, "loss": 2.9475, "step": 135300 }, { "epoch": 1.195270138999276, "grad_norm": 1.6041375398635864, "learning_rate": 8.670378974435388e-05, "loss": 3.398, "step": 135350 }, { "epoch": 1.19571168688956, "grad_norm": 1.7173302173614502, "learning_rate": 8.669435188255228e-05, "loss": 3.0656, "step": 135400 }, { "epoch": 1.1961532347798443, "grad_norm": 15.721769332885742, "learning_rate": 8.668491118642316e-05, "loss": 3.0591, "step": 135450 }, { "epoch": 1.1965947826701284, "grad_norm": 2.442812204360962, "learning_rate": 8.667546765669572e-05, "loss": 3.1348, "step": 135500 }, { "epoch": 1.1970363305604126, "grad_norm": 1.5753581523895264, "learning_rate": 8.666621024910873e-05, "loss": 3.1149, "step": 135550 }, { "epoch": 1.1974778784506968, "grad_norm": 2.9639949798583984, "learning_rate": 8.665676111100882e-05, "loss": 3.1263, "step": 135600 }, { "epoch": 1.197919426340981, "grad_norm": 2.138810634613037, "learning_rate": 8.664730914148498e-05, "loss": 3.5442, "step": 135650 }, { "epoch": 1.198360974231265, "grad_norm": 2.3526198863983154, "learning_rate": 8.663785434126725e-05, "loss": 3.5807, "step": 135700 }, { "epoch": 1.1988025221215493, "grad_norm": 1.0211713314056396, "learning_rate": 8.662839671108594e-05, "loss": 3.157, "step": 135750 }, { "epoch": 1.1992440700118334, "grad_norm": 3.1189095973968506, "learning_rate": 8.661893625167158e-05, "loss": 3.0157, "step": 135800 }, { "epoch": 1.1996856179021176, "grad_norm": 1.5837284326553345, "learning_rate": 8.660947296375495e-05, "loss": 2.9758, "step": 135850 }, { "epoch": 1.2001271657924018, "grad_norm": 2.6841766834259033, "learning_rate": 8.660000684806694e-05, "loss": 3.0694, "step": 135900 }, { "epoch": 1.200568713682686, "grad_norm": 3.424384832382202, "learning_rate": 8.659053790533877e-05, "loss": 2.7653, "step": 135950 }, { "epoch": 1.20101026157297, "grad_norm": 2.5790984630584717, "learning_rate": 8.658106613630184e-05, "loss": 3.3548, "step": 136000 }, { "epoch": 1.2014518094632545, "grad_norm": 0.8066131472587585, "learning_rate": 8.657159154168774e-05, "loss": 2.7384, "step": 136050 }, { "epoch": 1.2018933573535386, "grad_norm": 1.3459528684616089, "learning_rate": 8.65621141222283e-05, "loss": 3.3098, "step": 136100 }, { "epoch": 1.2023349052438228, "grad_norm": 1.5684973001480103, "learning_rate": 8.65526338786556e-05, "loss": 3.2824, "step": 136150 }, { "epoch": 1.202776453134107, "grad_norm": 1.0687460899353027, "learning_rate": 8.654315081170187e-05, "loss": 2.8486, "step": 136200 }, { "epoch": 1.2032180010243911, "grad_norm": 1.0528970956802368, "learning_rate": 8.653366492209962e-05, "loss": 3.1545, "step": 136250 }, { "epoch": 1.2036595489146753, "grad_norm": 4.352191925048828, "learning_rate": 8.652417621058158e-05, "loss": 3.1152, "step": 136300 }, { "epoch": 1.2041010968049595, "grad_norm": 1.8942192792892456, "learning_rate": 8.651468467788061e-05, "loss": 3.6867, "step": 136350 }, { "epoch": 1.2045426446952436, "grad_norm": 2.341681480407715, "learning_rate": 8.65051903247299e-05, "loss": 3.1624, "step": 136400 }, { "epoch": 1.2049841925855278, "grad_norm": 1.9934921264648438, "learning_rate": 8.649569315186279e-05, "loss": 3.2298, "step": 136450 }, { "epoch": 1.205425740475812, "grad_norm": 1.4484180212020874, "learning_rate": 8.648619316001286e-05, "loss": 3.3831, "step": 136500 }, { "epoch": 1.2058672883660961, "grad_norm": 1.979071855545044, "learning_rate": 8.647669034991389e-05, "loss": 2.9035, "step": 136550 }, { "epoch": 1.2063088362563803, "grad_norm": 2.6589207649230957, "learning_rate": 8.646718472229991e-05, "loss": 3.3321, "step": 136600 }, { "epoch": 1.2067503841466645, "grad_norm": 1.3302929401397705, "learning_rate": 8.645767627790515e-05, "loss": 3.5172, "step": 136650 }, { "epoch": 1.2071919320369486, "grad_norm": 4.300166606903076, "learning_rate": 8.644816501746405e-05, "loss": 3.1273, "step": 136700 }, { "epoch": 1.207633479927233, "grad_norm": 2.7277801036834717, "learning_rate": 8.643865094171126e-05, "loss": 3.6246, "step": 136750 }, { "epoch": 1.2080750278175172, "grad_norm": 1.1152167320251465, "learning_rate": 8.642913405138168e-05, "loss": 2.9205, "step": 136800 }, { "epoch": 1.2085165757078014, "grad_norm": 3.885781764984131, "learning_rate": 8.641961434721042e-05, "loss": 3.4886, "step": 136850 }, { "epoch": 1.2089581235980855, "grad_norm": 1.0696148872375488, "learning_rate": 8.641009182993276e-05, "loss": 2.7299, "step": 136900 }, { "epoch": 1.2093996714883697, "grad_norm": 2.8801472187042236, "learning_rate": 8.640056650028428e-05, "loss": 3.257, "step": 136950 }, { "epoch": 1.2098412193786539, "grad_norm": 3.712364435195923, "learning_rate": 8.639103835900069e-05, "loss": 3.5346, "step": 137000 }, { "epoch": 1.210282767268938, "grad_norm": 1.6460645198822021, "learning_rate": 8.638150740681796e-05, "loss": 3.5413, "step": 137050 }, { "epoch": 1.2107243151592222, "grad_norm": 1.8404775857925415, "learning_rate": 8.637197364447231e-05, "loss": 3.5827, "step": 137100 }, { "epoch": 1.2111658630495064, "grad_norm": 4.571069240570068, "learning_rate": 8.63624370727001e-05, "loss": 3.0262, "step": 137150 }, { "epoch": 1.2116074109397905, "grad_norm": 3.377689838409424, "learning_rate": 8.6352897692238e-05, "loss": 3.0937, "step": 137200 }, { "epoch": 1.2120489588300747, "grad_norm": 3.101484775543213, "learning_rate": 8.63433555038228e-05, "loss": 3.4086, "step": 137250 }, { "epoch": 1.2124905067203589, "grad_norm": 3.2903013229370117, "learning_rate": 8.633381050819157e-05, "loss": 3.3043, "step": 137300 }, { "epoch": 1.212932054610643, "grad_norm": 9.131916046142578, "learning_rate": 8.632426270608159e-05, "loss": 3.4324, "step": 137350 }, { "epoch": 1.2133736025009272, "grad_norm": 1.1834709644317627, "learning_rate": 8.631471209823032e-05, "loss": 3.271, "step": 137400 }, { "epoch": 1.2138151503912114, "grad_norm": 1.6877617835998535, "learning_rate": 8.63051586853755e-05, "loss": 2.7578, "step": 137450 }, { "epoch": 1.2142566982814955, "grad_norm": 3.62005352973938, "learning_rate": 8.629560246825501e-05, "loss": 2.5288, "step": 137500 }, { "epoch": 1.2146982461717797, "grad_norm": 2.522944450378418, "learning_rate": 8.628604344760701e-05, "loss": 2.7992, "step": 137550 }, { "epoch": 1.2151397940620638, "grad_norm": 2.285325050354004, "learning_rate": 8.627648162416985e-05, "loss": 3.0256, "step": 137600 }, { "epoch": 1.2155813419523482, "grad_norm": 2.962488889694214, "learning_rate": 8.626691699868211e-05, "loss": 3.6971, "step": 137650 }, { "epoch": 1.2160228898426324, "grad_norm": 9.112953186035156, "learning_rate": 8.625734957188257e-05, "loss": 3.3418, "step": 137700 }, { "epoch": 1.2164644377329166, "grad_norm": 1.3443338871002197, "learning_rate": 8.624777934451022e-05, "loss": 3.3575, "step": 137750 }, { "epoch": 1.2169059856232007, "grad_norm": 6.0494256019592285, "learning_rate": 8.623820631730429e-05, "loss": 3.2255, "step": 137800 }, { "epoch": 1.217347533513485, "grad_norm": 2.178502321243286, "learning_rate": 8.62286304910042e-05, "loss": 3.2216, "step": 137850 }, { "epoch": 1.217789081403769, "grad_norm": 3.197324275970459, "learning_rate": 8.621905186634964e-05, "loss": 3.1341, "step": 137900 }, { "epoch": 1.2182306292940532, "grad_norm": 1.0135949850082397, "learning_rate": 8.620947044408043e-05, "loss": 3.2025, "step": 137950 }, { "epoch": 1.2186721771843374, "grad_norm": 2.018465995788574, "learning_rate": 8.619988622493669e-05, "loss": 3.1432, "step": 138000 }, { "epoch": 1.2186721771843374, "eval_asr_loss": 0.9245979018125478, "eval_loss": 2.9356586933135986, "eval_runtime": 20.4935, "eval_samples_per_second": 37.475, "eval_steps_per_second": 9.369, "eval_tts_loss": 5.940896828933876, "step": 138000 }, { "epoch": 1.2191137250746216, "grad_norm": 3.4536030292510986, "learning_rate": 8.61902992096587e-05, "loss": 3.4214, "step": 138050 }, { "epoch": 1.2195552729649057, "grad_norm": 2.176469087600708, "learning_rate": 8.618070939898699e-05, "loss": 3.1659, "step": 138100 }, { "epoch": 1.21999682085519, "grad_norm": 4.467554092407227, "learning_rate": 8.617111679366226e-05, "loss": 2.888, "step": 138150 }, { "epoch": 1.220438368745474, "grad_norm": 1.241233468055725, "learning_rate": 8.616152139442551e-05, "loss": 2.8083, "step": 138200 }, { "epoch": 1.2208799166357582, "grad_norm": 2.9702117443084717, "learning_rate": 8.615192320201784e-05, "loss": 3.3338, "step": 138250 }, { "epoch": 1.2213214645260424, "grad_norm": 1.1560981273651123, "learning_rate": 8.614232221718069e-05, "loss": 3.3368, "step": 138300 }, { "epoch": 1.2217630124163268, "grad_norm": 0.8796913623809814, "learning_rate": 8.613271844065561e-05, "loss": 3.2698, "step": 138350 }, { "epoch": 1.222204560306611, "grad_norm": 3.3393924236297607, "learning_rate": 8.612311187318442e-05, "loss": 3.3047, "step": 138400 }, { "epoch": 1.2226461081968951, "grad_norm": 3.1728880405426025, "learning_rate": 8.611350251550916e-05, "loss": 3.1528, "step": 138450 }, { "epoch": 1.2230876560871793, "grad_norm": 0.9648553133010864, "learning_rate": 8.610389036837207e-05, "loss": 3.1583, "step": 138500 }, { "epoch": 1.2235292039774635, "grad_norm": 2.8733668327331543, "learning_rate": 8.60942754325156e-05, "loss": 3.0704, "step": 138550 }, { "epoch": 1.2239707518677476, "grad_norm": 8.728034973144531, "learning_rate": 8.608465770868244e-05, "loss": 3.431, "step": 138600 }, { "epoch": 1.2244122997580318, "grad_norm": 4.511626243591309, "learning_rate": 8.607503719761542e-05, "loss": 2.9597, "step": 138650 }, { "epoch": 1.224853847648316, "grad_norm": 4.3571014404296875, "learning_rate": 8.606541390005772e-05, "loss": 2.9135, "step": 138700 }, { "epoch": 1.2252953955386001, "grad_norm": 2.458829164505005, "learning_rate": 8.60557878167526e-05, "loss": 3.7083, "step": 138750 }, { "epoch": 1.2257369434288843, "grad_norm": 3.1015942096710205, "learning_rate": 8.604615894844364e-05, "loss": 3.2546, "step": 138800 }, { "epoch": 1.2261784913191685, "grad_norm": 4.6130571365356445, "learning_rate": 8.603652729587455e-05, "loss": 3.2822, "step": 138850 }, { "epoch": 1.2266200392094526, "grad_norm": 0.9187746047973633, "learning_rate": 8.60268928597893e-05, "loss": 2.8447, "step": 138900 }, { "epoch": 1.2270615870997368, "grad_norm": 1.320948600769043, "learning_rate": 8.601725564093209e-05, "loss": 3.2058, "step": 138950 }, { "epoch": 1.227503134990021, "grad_norm": 1.928350567817688, "learning_rate": 8.600761564004727e-05, "loss": 3.5719, "step": 139000 }, { "epoch": 1.2279446828803051, "grad_norm": 1.5674891471862793, "learning_rate": 8.599797285787951e-05, "loss": 2.9566, "step": 139050 }, { "epoch": 1.2283862307705893, "grad_norm": 0.9713530540466309, "learning_rate": 8.598832729517359e-05, "loss": 3.3049, "step": 139100 }, { "epoch": 1.2288277786608734, "grad_norm": 2.7060794830322266, "learning_rate": 8.597867895267455e-05, "loss": 3.2953, "step": 139150 }, { "epoch": 1.2292693265511576, "grad_norm": 1.6016948223114014, "learning_rate": 8.596902783112765e-05, "loss": 3.6453, "step": 139200 }, { "epoch": 1.229710874441442, "grad_norm": 3.066992998123169, "learning_rate": 8.595937393127838e-05, "loss": 3.7555, "step": 139250 }, { "epoch": 1.2301524223317262, "grad_norm": 1.1654362678527832, "learning_rate": 8.59497172538724e-05, "loss": 3.4442, "step": 139300 }, { "epoch": 1.2305939702220103, "grad_norm": 1.2640044689178467, "learning_rate": 8.594005779965559e-05, "loss": 3.2677, "step": 139350 }, { "epoch": 1.2310355181122945, "grad_norm": 1.9451007843017578, "learning_rate": 8.593039556937408e-05, "loss": 3.0464, "step": 139400 }, { "epoch": 1.2314770660025787, "grad_norm": 2.010702610015869, "learning_rate": 8.59207305637742e-05, "loss": 3.5588, "step": 139450 }, { "epoch": 1.2319186138928628, "grad_norm": 1.7903605699539185, "learning_rate": 8.591106278360247e-05, "loss": 2.9184, "step": 139500 }, { "epoch": 1.232360161783147, "grad_norm": 4.142259120941162, "learning_rate": 8.590139222960568e-05, "loss": 3.2671, "step": 139550 }, { "epoch": 1.2328017096734312, "grad_norm": 6.523422718048096, "learning_rate": 8.589171890253073e-05, "loss": 3.1727, "step": 139600 }, { "epoch": 1.2332432575637153, "grad_norm": 2.88853120803833, "learning_rate": 8.588204280312488e-05, "loss": 3.1479, "step": 139650 }, { "epoch": 1.2336848054539995, "grad_norm": 3.082754135131836, "learning_rate": 8.587236393213549e-05, "loss": 2.7581, "step": 139700 }, { "epoch": 1.2341263533442837, "grad_norm": 2.5806145668029785, "learning_rate": 8.586268229031019e-05, "loss": 3.3388, "step": 139750 }, { "epoch": 1.2345679012345678, "grad_norm": 6.057657718658447, "learning_rate": 8.585299787839678e-05, "loss": 3.3327, "step": 139800 }, { "epoch": 1.235009449124852, "grad_norm": 5.6322021484375, "learning_rate": 8.58433106971433e-05, "loss": 3.6759, "step": 139850 }, { "epoch": 1.2354509970151362, "grad_norm": 0.7276424169540405, "learning_rate": 8.583362074729802e-05, "loss": 3.641, "step": 139900 }, { "epoch": 1.2358925449054206, "grad_norm": 0.8799581527709961, "learning_rate": 8.582392802960939e-05, "loss": 3.5811, "step": 139950 }, { "epoch": 1.2363340927957047, "grad_norm": 3.2745556831359863, "learning_rate": 8.58142325448261e-05, "loss": 3.1727, "step": 140000 }, { "epoch": 1.2367756406859889, "grad_norm": 1.2958625555038452, "learning_rate": 8.580472828582498e-05, "loss": 3.2812, "step": 140050 }, { "epoch": 1.237217188576273, "grad_norm": 0.9326485395431519, "learning_rate": 8.579502732440385e-05, "loss": 3.3295, "step": 140100 }, { "epoch": 1.2376587364665572, "grad_norm": 1.5448970794677734, "learning_rate": 8.578532359812039e-05, "loss": 2.8871, "step": 140150 }, { "epoch": 1.2381002843568414, "grad_norm": 2.073150873184204, "learning_rate": 8.577561710772413e-05, "loss": 3.5859, "step": 140200 }, { "epoch": 1.2385418322471256, "grad_norm": 1.6688034534454346, "learning_rate": 8.576590785396483e-05, "loss": 3.0485, "step": 140250 }, { "epoch": 1.2389833801374097, "grad_norm": 2.4727234840393066, "learning_rate": 8.575619583759242e-05, "loss": 3.1929, "step": 140300 }, { "epoch": 1.2394249280276939, "grad_norm": 3.2702481746673584, "learning_rate": 8.574648105935709e-05, "loss": 3.1036, "step": 140350 }, { "epoch": 1.239866475917978, "grad_norm": 2.7264039516448975, "learning_rate": 8.573676352000923e-05, "loss": 3.4255, "step": 140400 }, { "epoch": 1.2403080238082622, "grad_norm": 3.251349925994873, "learning_rate": 8.572704322029942e-05, "loss": 3.3815, "step": 140450 }, { "epoch": 1.2407495716985464, "grad_norm": 3.0462658405303955, "learning_rate": 8.571732016097849e-05, "loss": 3.0341, "step": 140500 }, { "epoch": 1.2411911195888305, "grad_norm": 1.1782664060592651, "learning_rate": 8.570759434279744e-05, "loss": 3.2994, "step": 140550 }, { "epoch": 1.2416326674791147, "grad_norm": 1.4393571615219116, "learning_rate": 8.56978657665075e-05, "loss": 3.3121, "step": 140600 }, { "epoch": 1.2420742153693989, "grad_norm": 1.4110065698623657, "learning_rate": 8.568813443286017e-05, "loss": 3.038, "step": 140650 }, { "epoch": 1.242515763259683, "grad_norm": 1.5697439908981323, "learning_rate": 8.567840034260706e-05, "loss": 2.8957, "step": 140700 }, { "epoch": 1.2429573111499672, "grad_norm": 2.0723979473114014, "learning_rate": 8.566866349650009e-05, "loss": 3.1198, "step": 140750 }, { "epoch": 1.2433988590402516, "grad_norm": 3.4420316219329834, "learning_rate": 8.565892389529131e-05, "loss": 3.6046, "step": 140800 }, { "epoch": 1.2438404069305358, "grad_norm": 0.7216135263442993, "learning_rate": 8.564918153973305e-05, "loss": 3.0498, "step": 140850 }, { "epoch": 1.24428195482082, "grad_norm": 2.0984039306640625, "learning_rate": 8.563943643057779e-05, "loss": 3.2841, "step": 140900 }, { "epoch": 1.244723502711104, "grad_norm": 2.13993501663208, "learning_rate": 8.56296885685783e-05, "loss": 3.3196, "step": 140950 }, { "epoch": 1.2451650506013883, "grad_norm": 3.896669387817383, "learning_rate": 8.561993795448748e-05, "loss": 2.9746, "step": 141000 }, { "epoch": 1.2451650506013883, "eval_asr_loss": 0.9286146542702917, "eval_loss": 2.938673734664917, "eval_runtime": 20.5886, "eval_samples_per_second": 37.302, "eval_steps_per_second": 9.326, "eval_tts_loss": 5.989402265710523, "step": 141000 }, { "epoch": 1.2456065984916724, "grad_norm": 2.933366298675537, "learning_rate": 8.561018458905852e-05, "loss": 2.8598, "step": 141050 }, { "epoch": 1.2460481463819566, "grad_norm": 1.264729380607605, "learning_rate": 8.560042847304476e-05, "loss": 3.3088, "step": 141100 }, { "epoch": 1.2464896942722408, "grad_norm": 3.7516424655914307, "learning_rate": 8.559066960719978e-05, "loss": 3.1826, "step": 141150 }, { "epoch": 1.246931242162525, "grad_norm": 1.529587745666504, "learning_rate": 8.558110325151189e-05, "loss": 3.0593, "step": 141200 }, { "epoch": 1.247372790052809, "grad_norm": 2.4243791103363037, "learning_rate": 8.557133894322516e-05, "loss": 3.3451, "step": 141250 }, { "epoch": 1.2478143379430933, "grad_norm": 1.9787139892578125, "learning_rate": 8.55615718873541e-05, "loss": 2.6297, "step": 141300 }, { "epoch": 1.2482558858333774, "grad_norm": 1.8320086002349854, "learning_rate": 8.555180208465318e-05, "loss": 3.2509, "step": 141350 }, { "epoch": 1.2486974337236616, "grad_norm": 5.408695220947266, "learning_rate": 8.554202953587701e-05, "loss": 3.4093, "step": 141400 }, { "epoch": 1.2491389816139458, "grad_norm": 3.7261509895324707, "learning_rate": 8.553225424178045e-05, "loss": 3.1478, "step": 141450 }, { "epoch": 1.2495805295042302, "grad_norm": 1.5579651594161987, "learning_rate": 8.552247620311855e-05, "loss": 3.2053, "step": 141500 }, { "epoch": 1.2500220773945143, "grad_norm": 2.1170806884765625, "learning_rate": 8.55126954206466e-05, "loss": 3.3457, "step": 141550 }, { "epoch": 1.2504636252847985, "grad_norm": 4.129932403564453, "learning_rate": 8.550291189512005e-05, "loss": 2.4842, "step": 141600 }, { "epoch": 1.2509051731750827, "grad_norm": 0.851900041103363, "learning_rate": 8.549312562729463e-05, "loss": 3.0324, "step": 141650 }, { "epoch": 1.2513467210653668, "grad_norm": 2.4202487468719482, "learning_rate": 8.548333661792621e-05, "loss": 3.6167, "step": 141700 }, { "epoch": 1.251788268955651, "grad_norm": 3.154711961746216, "learning_rate": 8.547354486777094e-05, "loss": 2.9288, "step": 141750 }, { "epoch": 1.2522298168459352, "grad_norm": 2.6830191612243652, "learning_rate": 8.546375037758512e-05, "loss": 3.7127, "step": 141800 }, { "epoch": 1.2526713647362193, "grad_norm": 1.3863435983657837, "learning_rate": 8.545395314812532e-05, "loss": 3.1614, "step": 141850 }, { "epoch": 1.2531129126265035, "grad_norm": 4.197157382965088, "learning_rate": 8.544415318014828e-05, "loss": 3.1192, "step": 141900 }, { "epoch": 1.2535544605167876, "grad_norm": 2.338468074798584, "learning_rate": 8.543435047441096e-05, "loss": 3.8493, "step": 141950 }, { "epoch": 1.2539960084070718, "grad_norm": 3.277184247970581, "learning_rate": 8.542454503167054e-05, "loss": 3.4279, "step": 142000 }, { "epoch": 1.254437556297356, "grad_norm": 1.9274518489837646, "learning_rate": 8.541473685268442e-05, "loss": 3.132, "step": 142050 }, { "epoch": 1.2548791041876401, "grad_norm": 2.079745054244995, "learning_rate": 8.540492593821016e-05, "loss": 3.3633, "step": 142100 }, { "epoch": 1.2553206520779243, "grad_norm": 1.85688316822052, "learning_rate": 8.539511228900563e-05, "loss": 2.9476, "step": 142150 }, { "epoch": 1.2557621999682085, "grad_norm": 1.192519187927246, "learning_rate": 8.53852959058288e-05, "loss": 3.0553, "step": 142200 }, { "epoch": 1.2562037478584926, "grad_norm": 1.694144606590271, "learning_rate": 8.537547678943792e-05, "loss": 3.1106, "step": 142250 }, { "epoch": 1.2566452957487768, "grad_norm": 1.598209023475647, "learning_rate": 8.536565494059143e-05, "loss": 3.225, "step": 142300 }, { "epoch": 1.257086843639061, "grad_norm": 1.6346831321716309, "learning_rate": 8.5355830360048e-05, "loss": 3.1595, "step": 142350 }, { "epoch": 1.2575283915293451, "grad_norm": 1.3039432764053345, "learning_rate": 8.534600304856646e-05, "loss": 3.2784, "step": 142400 }, { "epoch": 1.2579699394196295, "grad_norm": 3.690321207046509, "learning_rate": 8.533617300690593e-05, "loss": 2.9288, "step": 142450 }, { "epoch": 1.2584114873099137, "grad_norm": 2.3511857986450195, "learning_rate": 8.532634023582567e-05, "loss": 3.2557, "step": 142500 }, { "epoch": 1.2588530352001979, "grad_norm": 2.7391929626464844, "learning_rate": 8.531650473608518e-05, "loss": 3.4044, "step": 142550 }, { "epoch": 1.259294583090482, "grad_norm": 4.795861721038818, "learning_rate": 8.530666650844419e-05, "loss": 2.7433, "step": 142600 }, { "epoch": 1.2597361309807662, "grad_norm": 2.0584652423858643, "learning_rate": 8.529682555366259e-05, "loss": 3.2387, "step": 142650 }, { "epoch": 1.2601776788710504, "grad_norm": 3.6735129356384277, "learning_rate": 8.528698187250052e-05, "loss": 3.4929, "step": 142700 }, { "epoch": 1.2606192267613345, "grad_norm": 2.488990545272827, "learning_rate": 8.527713546571834e-05, "loss": 3.4303, "step": 142750 }, { "epoch": 1.2610607746516187, "grad_norm": 3.3868706226348877, "learning_rate": 8.526728633407659e-05, "loss": 3.2953, "step": 142800 }, { "epoch": 1.2615023225419029, "grad_norm": 2.3298208713531494, "learning_rate": 8.525743447833601e-05, "loss": 3.8784, "step": 142850 }, { "epoch": 1.261943870432187, "grad_norm": 7.178310394287109, "learning_rate": 8.52475798992576e-05, "loss": 2.9547, "step": 142900 }, { "epoch": 1.2623854183224712, "grad_norm": 1.7926055192947388, "learning_rate": 8.523772259760255e-05, "loss": 3.4618, "step": 142950 }, { "epoch": 1.2628269662127554, "grad_norm": 2.2966785430908203, "learning_rate": 8.522786257413221e-05, "loss": 2.9838, "step": 143000 }, { "epoch": 1.2632685141030398, "grad_norm": 4.8272600173950195, "learning_rate": 8.52179998296082e-05, "loss": 3.4569, "step": 143050 }, { "epoch": 1.263710061993324, "grad_norm": 1.6934598684310913, "learning_rate": 8.520813436479238e-05, "loss": 3.3623, "step": 143100 }, { "epoch": 1.264151609883608, "grad_norm": 2.3045802116394043, "learning_rate": 8.519826618044672e-05, "loss": 2.8705, "step": 143150 }, { "epoch": 1.2645931577738923, "grad_norm": 1.5127077102661133, "learning_rate": 8.518839527733346e-05, "loss": 3.7103, "step": 143200 }, { "epoch": 1.2650347056641764, "grad_norm": 1.3205996751785278, "learning_rate": 8.517852165621507e-05, "loss": 3.3271, "step": 143250 }, { "epoch": 1.2654762535544606, "grad_norm": 4.060527801513672, "learning_rate": 8.516864531785417e-05, "loss": 2.9748, "step": 143300 }, { "epoch": 1.2659178014447448, "grad_norm": 2.6280086040496826, "learning_rate": 8.515876626301367e-05, "loss": 3.334, "step": 143350 }, { "epoch": 1.266359349335029, "grad_norm": 1.7252777814865112, "learning_rate": 8.514888449245659e-05, "loss": 2.9757, "step": 143400 }, { "epoch": 1.266800897225313, "grad_norm": 2.3652231693267822, "learning_rate": 8.513900000694624e-05, "loss": 3.2599, "step": 143450 }, { "epoch": 1.2672424451155972, "grad_norm": 3.4115631580352783, "learning_rate": 8.512911280724612e-05, "loss": 3.4985, "step": 143500 }, { "epoch": 1.2676839930058814, "grad_norm": 1.9971885681152344, "learning_rate": 8.51192228941199e-05, "loss": 3.3195, "step": 143550 }, { "epoch": 1.2681255408961656, "grad_norm": 1.7323343753814697, "learning_rate": 8.510933026833154e-05, "loss": 2.918, "step": 143600 }, { "epoch": 1.2685670887864497, "grad_norm": 1.7640634775161743, "learning_rate": 8.509943493064512e-05, "loss": 2.8826, "step": 143650 }, { "epoch": 1.269008636676734, "grad_norm": 1.5099214315414429, "learning_rate": 8.508973486936555e-05, "loss": 3.094, "step": 143700 }, { "epoch": 1.269450184567018, "grad_norm": 1.7164794206619263, "learning_rate": 8.507983416437615e-05, "loss": 3.541, "step": 143750 }, { "epoch": 1.2698917324573022, "grad_norm": 1.553723692893982, "learning_rate": 8.506993074976701e-05, "loss": 3.3889, "step": 143800 }, { "epoch": 1.2703332803475864, "grad_norm": 1.3120352029800415, "learning_rate": 8.506002462630309e-05, "loss": 2.9932, "step": 143850 }, { "epoch": 1.2707748282378706, "grad_norm": 4.442413806915283, "learning_rate": 8.50501157947496e-05, "loss": 3.1092, "step": 143900 }, { "epoch": 1.2712163761281547, "grad_norm": 1.6406073570251465, "learning_rate": 8.504020425587187e-05, "loss": 3.1981, "step": 143950 }, { "epoch": 1.271657924018439, "grad_norm": 2.6018967628479004, "learning_rate": 8.503029001043548e-05, "loss": 3.1659, "step": 144000 }, { "epoch": 1.271657924018439, "eval_asr_loss": 0.9322961166804568, "eval_loss": 2.9280214309692383, "eval_runtime": 20.2032, "eval_samples_per_second": 38.014, "eval_steps_per_second": 9.503, "eval_tts_loss": 5.993786748371479, "step": 144000 }, { "epoch": 1.2720994719087233, "grad_norm": 1.6098721027374268, "learning_rate": 8.502037305920624e-05, "loss": 3.3989, "step": 144050 }, { "epoch": 1.2725410197990075, "grad_norm": 1.1011525392532349, "learning_rate": 8.501045340295015e-05, "loss": 2.9338, "step": 144100 }, { "epoch": 1.2729825676892916, "grad_norm": 1.1838862895965576, "learning_rate": 8.500053104243342e-05, "loss": 3.4187, "step": 144150 }, { "epoch": 1.2734241155795758, "grad_norm": 3.3150293827056885, "learning_rate": 8.499060597842247e-05, "loss": 3.6549, "step": 144200 }, { "epoch": 1.27386566346986, "grad_norm": 6.420579433441162, "learning_rate": 8.498067821168389e-05, "loss": 3.0136, "step": 144250 }, { "epoch": 1.2743072113601441, "grad_norm": 1.6961369514465332, "learning_rate": 8.497074774298456e-05, "loss": 2.9492, "step": 144300 }, { "epoch": 1.2747487592504283, "grad_norm": 1.9354748725891113, "learning_rate": 8.496081457309153e-05, "loss": 3.2681, "step": 144350 }, { "epoch": 1.2751903071407125, "grad_norm": 1.1011364459991455, "learning_rate": 8.495087870277201e-05, "loss": 3.1899, "step": 144400 }, { "epoch": 1.2756318550309966, "grad_norm": 2.966367721557617, "learning_rate": 8.49409401327935e-05, "loss": 2.9833, "step": 144450 }, { "epoch": 1.2760734029212808, "grad_norm": 5.4192938804626465, "learning_rate": 8.493099886392364e-05, "loss": 3.3008, "step": 144500 }, { "epoch": 1.276514950811565, "grad_norm": 1.545189619064331, "learning_rate": 8.492105489693034e-05, "loss": 3.3218, "step": 144550 }, { "epoch": 1.2769564987018491, "grad_norm": 1.899052619934082, "learning_rate": 8.491110823258166e-05, "loss": 3.1562, "step": 144600 }, { "epoch": 1.2773980465921335, "grad_norm": 2.4317712783813477, "learning_rate": 8.49011588716459e-05, "loss": 3.5698, "step": 144650 }, { "epoch": 1.2778395944824177, "grad_norm": 0.8498044610023499, "learning_rate": 8.489120681489157e-05, "loss": 3.1922, "step": 144700 }, { "epoch": 1.2782811423727019, "grad_norm": 3.520315170288086, "learning_rate": 8.48812520630874e-05, "loss": 2.8446, "step": 144750 }, { "epoch": 1.278722690262986, "grad_norm": 2.8437373638153076, "learning_rate": 8.48712946170023e-05, "loss": 2.854, "step": 144800 }, { "epoch": 1.2791642381532702, "grad_norm": 2.23712420463562, "learning_rate": 8.486133447740537e-05, "loss": 3.4061, "step": 144850 }, { "epoch": 1.2796057860435543, "grad_norm": 3.989715337753296, "learning_rate": 8.485137164506596e-05, "loss": 3.206, "step": 144900 }, { "epoch": 1.2800473339338385, "grad_norm": 0.8477696776390076, "learning_rate": 8.484140612075364e-05, "loss": 3.3947, "step": 144950 }, { "epoch": 1.2804888818241227, "grad_norm": 1.0561884641647339, "learning_rate": 8.483143790523813e-05, "loss": 3.4777, "step": 145000 }, { "epoch": 1.2809304297144068, "grad_norm": 3.1107983589172363, "learning_rate": 8.482146699928941e-05, "loss": 3.2744, "step": 145050 }, { "epoch": 1.281371977604691, "grad_norm": 2.971332311630249, "learning_rate": 8.481149340367764e-05, "loss": 3.6498, "step": 145100 }, { "epoch": 1.2818135254949752, "grad_norm": 2.0575344562530518, "learning_rate": 8.480151711917321e-05, "loss": 2.7598, "step": 145150 }, { "epoch": 1.2822550733852593, "grad_norm": 1.7395824193954468, "learning_rate": 8.479153814654667e-05, "loss": 3.0415, "step": 145200 }, { "epoch": 1.2826966212755435, "grad_norm": 1.918105125427246, "learning_rate": 8.478155648656885e-05, "loss": 3.4108, "step": 145250 }, { "epoch": 1.2831381691658277, "grad_norm": 1.5583088397979736, "learning_rate": 8.477157214001072e-05, "loss": 3.1427, "step": 145300 }, { "epoch": 1.2835797170561118, "grad_norm": 2.076531410217285, "learning_rate": 8.47615851076435e-05, "loss": 3.1514, "step": 145350 }, { "epoch": 1.284021264946396, "grad_norm": 1.9906885623931885, "learning_rate": 8.47515953902386e-05, "loss": 3.2949, "step": 145400 }, { "epoch": 1.2844628128366802, "grad_norm": 1.319674015045166, "learning_rate": 8.474160298856764e-05, "loss": 3.5379, "step": 145450 }, { "epoch": 1.2849043607269643, "grad_norm": 2.557248830795288, "learning_rate": 8.473160790340245e-05, "loss": 3.4775, "step": 145500 }, { "epoch": 1.2853459086172485, "grad_norm": 1.7793129682540894, "learning_rate": 8.472161013551506e-05, "loss": 2.6998, "step": 145550 }, { "epoch": 1.2857874565075327, "grad_norm": 2.644193410873413, "learning_rate": 8.471160968567773e-05, "loss": 3.0271, "step": 145600 }, { "epoch": 1.286229004397817, "grad_norm": 2.9333336353302, "learning_rate": 8.470160655466289e-05, "loss": 3.0707, "step": 145650 }, { "epoch": 1.2866705522881012, "grad_norm": 1.7270702123641968, "learning_rate": 8.46916007432432e-05, "loss": 3.2489, "step": 145700 }, { "epoch": 1.2871121001783854, "grad_norm": 2.789318084716797, "learning_rate": 8.468159225219153e-05, "loss": 3.3415, "step": 145750 }, { "epoch": 1.2875536480686696, "grad_norm": 2.3298559188842773, "learning_rate": 8.467158108228094e-05, "loss": 3.4328, "step": 145800 }, { "epoch": 1.2879951959589537, "grad_norm": 2.0260627269744873, "learning_rate": 8.466156723428475e-05, "loss": 2.9463, "step": 145850 }, { "epoch": 1.288436743849238, "grad_norm": 1.130997896194458, "learning_rate": 8.465155070897639e-05, "loss": 3.4273, "step": 145900 }, { "epoch": 1.288878291739522, "grad_norm": 5.2059221267700195, "learning_rate": 8.464153150712957e-05, "loss": 3.4943, "step": 145950 }, { "epoch": 1.2893198396298062, "grad_norm": 1.7285943031311035, "learning_rate": 8.46315096295182e-05, "loss": 3.4631, "step": 146000 }, { "epoch": 1.2897613875200904, "grad_norm": 3.8934295177459717, "learning_rate": 8.462148507691638e-05, "loss": 3.4679, "step": 146050 }, { "epoch": 1.2902029354103746, "grad_norm": 4.572065353393555, "learning_rate": 8.461145785009843e-05, "loss": 3.2549, "step": 146100 }, { "epoch": 1.2906444833006587, "grad_norm": 2.979644775390625, "learning_rate": 8.460142794983883e-05, "loss": 3.023, "step": 146150 }, { "epoch": 1.291086031190943, "grad_norm": 2.907404661178589, "learning_rate": 8.459139537691235e-05, "loss": 3.2693, "step": 146200 }, { "epoch": 1.2915275790812273, "grad_norm": 3.809751510620117, "learning_rate": 8.458136013209391e-05, "loss": 3.6373, "step": 146250 }, { "epoch": 1.2919691269715115, "grad_norm": 2.584719181060791, "learning_rate": 8.457132221615862e-05, "loss": 3.107, "step": 146300 }, { "epoch": 1.2924106748617956, "grad_norm": 1.6046016216278076, "learning_rate": 8.456128162988186e-05, "loss": 3.3261, "step": 146350 }, { "epoch": 1.2928522227520798, "grad_norm": 1.9032001495361328, "learning_rate": 8.455123837403916e-05, "loss": 3.2709, "step": 146400 }, { "epoch": 1.293293770642364, "grad_norm": 0.8189128041267395, "learning_rate": 8.45411924494063e-05, "loss": 2.8643, "step": 146450 }, { "epoch": 1.2937353185326481, "grad_norm": 2.8768742084503174, "learning_rate": 8.45311438567592e-05, "loss": 3.0366, "step": 146500 }, { "epoch": 1.2941768664229323, "grad_norm": 1.3201286792755127, "learning_rate": 8.452109259687407e-05, "loss": 3.1472, "step": 146550 }, { "epoch": 1.2946184143132164, "grad_norm": 2.3641459941864014, "learning_rate": 8.451103867052725e-05, "loss": 3.7229, "step": 146600 }, { "epoch": 1.2950599622035006, "grad_norm": 7.314673900604248, "learning_rate": 8.450098207849534e-05, "loss": 3.1165, "step": 146650 }, { "epoch": 1.2955015100937848, "grad_norm": 2.273407220840454, "learning_rate": 8.449092282155513e-05, "loss": 3.0138, "step": 146700 }, { "epoch": 1.295943057984069, "grad_norm": 2.268495798110962, "learning_rate": 8.448086090048361e-05, "loss": 2.4604, "step": 146750 }, { "epoch": 1.2963846058743531, "grad_norm": 1.9816882610321045, "learning_rate": 8.447079631605797e-05, "loss": 3.2556, "step": 146800 }, { "epoch": 1.2968261537646373, "grad_norm": 1.5324925184249878, "learning_rate": 8.446072906905562e-05, "loss": 3.2377, "step": 146850 }, { "epoch": 1.2972677016549214, "grad_norm": 4.843852996826172, "learning_rate": 8.445065916025416e-05, "loss": 3.302, "step": 146900 }, { "epoch": 1.2977092495452056, "grad_norm": 4.478860855102539, "learning_rate": 8.444058659043143e-05, "loss": 3.0785, "step": 146950 }, { "epoch": 1.2981507974354898, "grad_norm": 4.566507339477539, "learning_rate": 8.443051136036542e-05, "loss": 3.5216, "step": 147000 }, { "epoch": 1.2981507974354898, "eval_asr_loss": 0.9203305681771484, "eval_loss": 2.9263274669647217, "eval_runtime": 20.2234, "eval_samples_per_second": 37.976, "eval_steps_per_second": 9.494, "eval_tts_loss": 5.990303801165284, "step": 147000 }, { "epoch": 1.298592345325774, "grad_norm": 1.1812950372695923, "learning_rate": 8.442043347083438e-05, "loss": 3.3527, "step": 147050 }, { "epoch": 1.299033893216058, "grad_norm": 0.6211222410202026, "learning_rate": 8.441035292261672e-05, "loss": 3.4486, "step": 147100 }, { "epoch": 1.2994754411063423, "grad_norm": 3.0547852516174316, "learning_rate": 8.440047140665608e-05, "loss": 2.8583, "step": 147150 }, { "epoch": 1.2999169889966264, "grad_norm": 1.5730994939804077, "learning_rate": 8.439038559653625e-05, "loss": 3.1623, "step": 147200 }, { "epoch": 1.3003585368869108, "grad_norm": 4.245264053344727, "learning_rate": 8.438029713005078e-05, "loss": 3.1082, "step": 147250 }, { "epoch": 1.300800084777195, "grad_norm": 2.3482489585876465, "learning_rate": 8.437020600797888e-05, "loss": 2.945, "step": 147300 }, { "epoch": 1.3012416326674792, "grad_norm": 0.9055774807929993, "learning_rate": 8.436011223110004e-05, "loss": 3.1753, "step": 147350 }, { "epoch": 1.3016831805577633, "grad_norm": 2.761030912399292, "learning_rate": 8.435001580019388e-05, "loss": 3.1069, "step": 147400 }, { "epoch": 1.3021247284480475, "grad_norm": 3.1822400093078613, "learning_rate": 8.433991671604028e-05, "loss": 3.3119, "step": 147450 }, { "epoch": 1.3025662763383317, "grad_norm": 2.4787344932556152, "learning_rate": 8.43298149794193e-05, "loss": 3.4397, "step": 147500 }, { "epoch": 1.3030078242286158, "grad_norm": 3.7208900451660156, "learning_rate": 8.431971059111124e-05, "loss": 3.4477, "step": 147550 }, { "epoch": 1.3034493721189, "grad_norm": 3.309570550918579, "learning_rate": 8.430960355189654e-05, "loss": 3.03, "step": 147600 }, { "epoch": 1.3038909200091842, "grad_norm": 0.6187686920166016, "learning_rate": 8.429949386255591e-05, "loss": 3.2692, "step": 147650 }, { "epoch": 1.3043324678994683, "grad_norm": 4.8789381980896, "learning_rate": 8.428938152387024e-05, "loss": 3.1269, "step": 147700 }, { "epoch": 1.3047740157897525, "grad_norm": 1.8257217407226562, "learning_rate": 8.42792665366206e-05, "loss": 3.6406, "step": 147750 }, { "epoch": 1.3052155636800367, "grad_norm": 0.7555809020996094, "learning_rate": 8.42691489015883e-05, "loss": 3.3547, "step": 147800 }, { "epoch": 1.305657111570321, "grad_norm": 2.5792572498321533, "learning_rate": 8.425902861955485e-05, "loss": 2.7654, "step": 147850 }, { "epoch": 1.3060986594606052, "grad_norm": 2.951904296875, "learning_rate": 8.424890569130195e-05, "loss": 3.4109, "step": 147900 }, { "epoch": 1.3065402073508894, "grad_norm": 2.7723746299743652, "learning_rate": 8.423878011761149e-05, "loss": 3.6622, "step": 147950 }, { "epoch": 1.3069817552411735, "grad_norm": 3.5301687717437744, "learning_rate": 8.422865189926561e-05, "loss": 3.3498, "step": 148000 }, { "epoch": 1.3074233031314577, "grad_norm": 2.9211134910583496, "learning_rate": 8.421852103704664e-05, "loss": 2.8255, "step": 148050 }, { "epoch": 1.3078648510217419, "grad_norm": 2.1544718742370605, "learning_rate": 8.420838753173705e-05, "loss": 3.3958, "step": 148100 }, { "epoch": 1.308306398912026, "grad_norm": 1.1590626239776611, "learning_rate": 8.419825138411963e-05, "loss": 3.156, "step": 148150 }, { "epoch": 1.3087479468023102, "grad_norm": 2.3119921684265137, "learning_rate": 8.418811259497727e-05, "loss": 3.5458, "step": 148200 }, { "epoch": 1.3091894946925944, "grad_norm": 5.003969192504883, "learning_rate": 8.417797116509312e-05, "loss": 3.0429, "step": 148250 }, { "epoch": 1.3096310425828785, "grad_norm": 3.385624408721924, "learning_rate": 8.416782709525053e-05, "loss": 3.5067, "step": 148300 }, { "epoch": 1.3100725904731627, "grad_norm": 2.701323986053467, "learning_rate": 8.415768038623302e-05, "loss": 3.4576, "step": 148350 }, { "epoch": 1.3105141383634469, "grad_norm": 1.0806344747543335, "learning_rate": 8.414753103882434e-05, "loss": 3.084, "step": 148400 }, { "epoch": 1.310955686253731, "grad_norm": 2.5085790157318115, "learning_rate": 8.413737905380848e-05, "loss": 3.1328, "step": 148450 }, { "epoch": 1.3113972341440152, "grad_norm": 2.03137469291687, "learning_rate": 8.412722443196953e-05, "loss": 3.4205, "step": 148500 }, { "epoch": 1.3118387820342994, "grad_norm": 2.739572286605835, "learning_rate": 8.41170671740919e-05, "loss": 3.5465, "step": 148550 }, { "epoch": 1.3122803299245835, "grad_norm": 1.639987826347351, "learning_rate": 8.410690728096013e-05, "loss": 3.1524, "step": 148600 }, { "epoch": 1.3127218778148677, "grad_norm": 3.545746088027954, "learning_rate": 8.4096744753359e-05, "loss": 2.8028, "step": 148650 }, { "epoch": 1.3131634257051519, "grad_norm": 5.186053276062012, "learning_rate": 8.408657959207347e-05, "loss": 3.3191, "step": 148700 }, { "epoch": 1.313604973595436, "grad_norm": 0.6773577928543091, "learning_rate": 8.40764117978887e-05, "loss": 3.2917, "step": 148750 }, { "epoch": 1.3140465214857202, "grad_norm": 1.8338830471038818, "learning_rate": 8.406624137159008e-05, "loss": 3.3637, "step": 148800 }, { "epoch": 1.3144880693760046, "grad_norm": 1.1749883890151978, "learning_rate": 8.405606831396318e-05, "loss": 3.0744, "step": 148850 }, { "epoch": 1.3149296172662888, "grad_norm": 2.124967575073242, "learning_rate": 8.404589262579381e-05, "loss": 3.4487, "step": 148900 }, { "epoch": 1.315371165156573, "grad_norm": 2.8038904666900635, "learning_rate": 8.403571430786793e-05, "loss": 3.1075, "step": 148950 }, { "epoch": 1.315812713046857, "grad_norm": 4.513190746307373, "learning_rate": 8.402553336097172e-05, "loss": 2.9033, "step": 149000 }, { "epoch": 1.3162542609371413, "grad_norm": 3.4421756267547607, "learning_rate": 8.40153497858916e-05, "loss": 3.0916, "step": 149050 }, { "epoch": 1.3166958088274254, "grad_norm": 1.1779747009277344, "learning_rate": 8.400516358341415e-05, "loss": 2.8582, "step": 149100 }, { "epoch": 1.3171373567177096, "grad_norm": 6.414771556854248, "learning_rate": 8.399497475432616e-05, "loss": 3.1022, "step": 149150 }, { "epoch": 1.3175789046079938, "grad_norm": 1.1380168199539185, "learning_rate": 8.398478329941466e-05, "loss": 3.1022, "step": 149200 }, { "epoch": 1.318020452498278, "grad_norm": 2.7382118701934814, "learning_rate": 8.397458921946682e-05, "loss": 3.0475, "step": 149250 }, { "epoch": 1.318462000388562, "grad_norm": 3.204192876815796, "learning_rate": 8.396439251527004e-05, "loss": 3.0297, "step": 149300 }, { "epoch": 1.3189035482788463, "grad_norm": 1.7281981706619263, "learning_rate": 8.395419318761197e-05, "loss": 3.0445, "step": 149350 }, { "epoch": 1.3193450961691304, "grad_norm": 2.468430280685425, "learning_rate": 8.39439912372804e-05, "loss": 2.9012, "step": 149400 }, { "epoch": 1.3197866440594148, "grad_norm": 3.0778887271881104, "learning_rate": 8.393378666506333e-05, "loss": 3.4102, "step": 149450 }, { "epoch": 1.320228191949699, "grad_norm": 1.3242723941802979, "learning_rate": 8.3923579471749e-05, "loss": 2.794, "step": 149500 }, { "epoch": 1.3206697398399831, "grad_norm": 5.034628391265869, "learning_rate": 8.391336965812581e-05, "loss": 3.0745, "step": 149550 }, { "epoch": 1.3211112877302673, "grad_norm": 3.1892762184143066, "learning_rate": 8.390315722498241e-05, "loss": 3.0208, "step": 149600 }, { "epoch": 1.3215528356205515, "grad_norm": 1.5829938650131226, "learning_rate": 8.389294217310759e-05, "loss": 3.2962, "step": 149650 }, { "epoch": 1.3219943835108356, "grad_norm": 3.269179105758667, "learning_rate": 8.38827245032904e-05, "loss": 3.3731, "step": 149700 }, { "epoch": 1.3224359314011198, "grad_norm": 2.511181116104126, "learning_rate": 8.387250421632005e-05, "loss": 3.2497, "step": 149750 }, { "epoch": 1.322877479291404, "grad_norm": 1.1484040021896362, "learning_rate": 8.386228131298597e-05, "loss": 3.205, "step": 149800 }, { "epoch": 1.3233190271816881, "grad_norm": 0.9007138609886169, "learning_rate": 8.385205579407782e-05, "loss": 2.6462, "step": 149850 }, { "epoch": 1.3237605750719723, "grad_norm": 1.4533042907714844, "learning_rate": 8.38418276603854e-05, "loss": 3.2513, "step": 149900 }, { "epoch": 1.3242021229622565, "grad_norm": 3.494771957397461, "learning_rate": 8.38315969126988e-05, "loss": 3.298, "step": 149950 }, { "epoch": 1.3246436708525406, "grad_norm": 3.4420292377471924, "learning_rate": 8.382136355180819e-05, "loss": 3.4275, "step": 150000 }, { "epoch": 1.3246436708525406, "eval_asr_loss": 0.9286041771558345, "eval_loss": 2.9267418384552, "eval_runtime": 20.4668, "eval_samples_per_second": 37.524, "eval_steps_per_second": 9.381, "eval_tts_loss": 5.983196550111349, "step": 150000 }, { "epoch": 1.3250852187428248, "grad_norm": 4.632486820220947, "learning_rate": 8.381112757850405e-05, "loss": 2.9501, "step": 150050 }, { "epoch": 1.325526766633109, "grad_norm": 1.6033202409744263, "learning_rate": 8.380088899357701e-05, "loss": 2.4565, "step": 150100 }, { "epoch": 1.3259683145233931, "grad_norm": 2.7419354915618896, "learning_rate": 8.37906477978179e-05, "loss": 3.2429, "step": 150150 }, { "epoch": 1.3264098624136773, "grad_norm": 1.273281216621399, "learning_rate": 8.378040399201783e-05, "loss": 3.1723, "step": 150200 }, { "epoch": 1.3268514103039615, "grad_norm": 4.021177291870117, "learning_rate": 8.377015757696797e-05, "loss": 2.9459, "step": 150250 }, { "epoch": 1.3272929581942456, "grad_norm": 2.726773500442505, "learning_rate": 8.375990855345981e-05, "loss": 3.2983, "step": 150300 }, { "epoch": 1.3277345060845298, "grad_norm": 1.6660208702087402, "learning_rate": 8.374965692228499e-05, "loss": 2.8783, "step": 150350 }, { "epoch": 1.328176053974814, "grad_norm": 12.518718719482422, "learning_rate": 8.373940268423535e-05, "loss": 3.518, "step": 150400 }, { "epoch": 1.3286176018650984, "grad_norm": 3.1526901721954346, "learning_rate": 8.372914584010297e-05, "loss": 3.4692, "step": 150450 }, { "epoch": 1.3290591497553825, "grad_norm": 0.5204312801361084, "learning_rate": 8.371888639068008e-05, "loss": 3.257, "step": 150500 }, { "epoch": 1.3295006976456667, "grad_norm": 3.0339365005493164, "learning_rate": 8.370862433675915e-05, "loss": 3.2844, "step": 150550 }, { "epoch": 1.3299422455359509, "grad_norm": 1.2331526279449463, "learning_rate": 8.369835967913282e-05, "loss": 3.2728, "step": 150600 }, { "epoch": 1.330383793426235, "grad_norm": 4.784111022949219, "learning_rate": 8.368809241859397e-05, "loss": 3.0626, "step": 150650 }, { "epoch": 1.3308253413165192, "grad_norm": 0.7437164783477783, "learning_rate": 8.367782255593564e-05, "loss": 3.2591, "step": 150700 }, { "epoch": 1.3312668892068034, "grad_norm": 1.0057419538497925, "learning_rate": 8.366755009195111e-05, "loss": 3.4121, "step": 150750 }, { "epoch": 1.3317084370970875, "grad_norm": 1.6400574445724487, "learning_rate": 8.365727502743381e-05, "loss": 3.0203, "step": 150800 }, { "epoch": 1.3321499849873717, "grad_norm": 2.0449421405792236, "learning_rate": 8.364699736317743e-05, "loss": 3.4384, "step": 150850 }, { "epoch": 1.3325915328776559, "grad_norm": 1.9299911260604858, "learning_rate": 8.363671709997582e-05, "loss": 3.2457, "step": 150900 }, { "epoch": 1.33303308076794, "grad_norm": 2.1701300144195557, "learning_rate": 8.362643423862305e-05, "loss": 3.1797, "step": 150950 }, { "epoch": 1.3334746286582242, "grad_norm": 1.6506251096725464, "learning_rate": 8.361614877991338e-05, "loss": 2.9922, "step": 151000 }, { "epoch": 1.3339161765485086, "grad_norm": 1.9627467393875122, "learning_rate": 8.360586072464126e-05, "loss": 3.407, "step": 151050 }, { "epoch": 1.3343577244387927, "grad_norm": 2.9375967979431152, "learning_rate": 8.359557007360136e-05, "loss": 3.3929, "step": 151100 }, { "epoch": 1.334799272329077, "grad_norm": 2.7383005619049072, "learning_rate": 8.358527682758858e-05, "loss": 3.443, "step": 151150 }, { "epoch": 1.335240820219361, "grad_norm": 1.5693986415863037, "learning_rate": 8.357498098739793e-05, "loss": 3.4043, "step": 151200 }, { "epoch": 1.3356823681096452, "grad_norm": 4.640061855316162, "learning_rate": 8.356468255382474e-05, "loss": 3.5593, "step": 151250 }, { "epoch": 1.3361239159999294, "grad_norm": 2.2127010822296143, "learning_rate": 8.355438152766442e-05, "loss": 3.7348, "step": 151300 }, { "epoch": 1.3365654638902136, "grad_norm": 2.803560972213745, "learning_rate": 8.354407790971268e-05, "loss": 3.3592, "step": 151350 }, { "epoch": 1.3370070117804977, "grad_norm": 1.9920105934143066, "learning_rate": 8.353377170076536e-05, "loss": 3.2962, "step": 151400 }, { "epoch": 1.337448559670782, "grad_norm": 1.280686378479004, "learning_rate": 8.352346290161852e-05, "loss": 3.347, "step": 151450 }, { "epoch": 1.337890107561066, "grad_norm": 2.0447115898132324, "learning_rate": 8.351315151306845e-05, "loss": 3.2339, "step": 151500 }, { "epoch": 1.3383316554513502, "grad_norm": 2.681647777557373, "learning_rate": 8.350283753591161e-05, "loss": 3.2523, "step": 151550 }, { "epoch": 1.3387732033416344, "grad_norm": 4.001578330993652, "learning_rate": 8.349252097094467e-05, "loss": 3.0276, "step": 151600 }, { "epoch": 1.3392147512319186, "grad_norm": 1.5295051336288452, "learning_rate": 8.348220181896449e-05, "loss": 3.6369, "step": 151650 }, { "epoch": 1.3396562991222027, "grad_norm": 3.501155138015747, "learning_rate": 8.347208654087182e-05, "loss": 3.1035, "step": 151700 }, { "epoch": 1.340097847012487, "grad_norm": 3.054642915725708, "learning_rate": 8.346176226895712e-05, "loss": 3.1906, "step": 151750 }, { "epoch": 1.340539394902771, "grad_norm": 2.961508274078369, "learning_rate": 8.345143541240505e-05, "loss": 2.7505, "step": 151800 }, { "epoch": 1.3409809427930552, "grad_norm": 2.213796377182007, "learning_rate": 8.344110597201324e-05, "loss": 3.4039, "step": 151850 }, { "epoch": 1.3414224906833394, "grad_norm": 3.433262825012207, "learning_rate": 8.343077394857957e-05, "loss": 3.3918, "step": 151900 }, { "epoch": 1.3418640385736236, "grad_norm": 2.9317116737365723, "learning_rate": 8.34204393429021e-05, "loss": 3.7814, "step": 151950 }, { "epoch": 1.3423055864639077, "grad_norm": 1.5801756381988525, "learning_rate": 8.341010215577909e-05, "loss": 3.4844, "step": 152000 }, { "epoch": 1.3427471343541921, "grad_norm": 4.749630451202393, "learning_rate": 8.3399762388009e-05, "loss": 3.5081, "step": 152050 }, { "epoch": 1.3431886822444763, "grad_norm": 1.1885944604873657, "learning_rate": 8.338942004039047e-05, "loss": 3.2524, "step": 152100 }, { "epoch": 1.3436302301347605, "grad_norm": 2.543466567993164, "learning_rate": 8.33790751137224e-05, "loss": 3.0227, "step": 152150 }, { "epoch": 1.3440717780250446, "grad_norm": 1.6229182481765747, "learning_rate": 8.33687276088038e-05, "loss": 3.2959, "step": 152200 }, { "epoch": 1.3445133259153288, "grad_norm": 1.94907808303833, "learning_rate": 8.335837752643397e-05, "loss": 2.9457, "step": 152250 }, { "epoch": 1.344954873805613, "grad_norm": 1.4191594123840332, "learning_rate": 8.334802486741233e-05, "loss": 3.6033, "step": 152300 }, { "epoch": 1.3453964216958971, "grad_norm": 2.68353271484375, "learning_rate": 8.333766963253858e-05, "loss": 2.968, "step": 152350 }, { "epoch": 1.3458379695861813, "grad_norm": 3.616288185119629, "learning_rate": 8.332731182261251e-05, "loss": 2.9924, "step": 152400 }, { "epoch": 1.3462795174764655, "grad_norm": 1.501586675643921, "learning_rate": 8.331695143843424e-05, "loss": 3.0411, "step": 152450 }, { "epoch": 1.3467210653667496, "grad_norm": 2.923124074935913, "learning_rate": 8.330658848080397e-05, "loss": 3.5488, "step": 152500 }, { "epoch": 1.3471626132570338, "grad_norm": 2.4280457496643066, "learning_rate": 8.329622295052218e-05, "loss": 3.202, "step": 152550 }, { "epoch": 1.347604161147318, "grad_norm": 1.1072304248809814, "learning_rate": 8.328585484838952e-05, "loss": 2.9447, "step": 152600 }, { "epoch": 1.3480457090376023, "grad_norm": 2.937720775604248, "learning_rate": 8.327548417520681e-05, "loss": 3.2629, "step": 152650 }, { "epoch": 1.3484872569278865, "grad_norm": 1.4031856060028076, "learning_rate": 8.326511093177511e-05, "loss": 3.7229, "step": 152700 }, { "epoch": 1.3489288048181707, "grad_norm": 1.5467109680175781, "learning_rate": 8.325473511889569e-05, "loss": 3.2737, "step": 152750 }, { "epoch": 1.3493703527084548, "grad_norm": 1.7189159393310547, "learning_rate": 8.324435673736997e-05, "loss": 3.3052, "step": 152800 }, { "epoch": 1.349811900598739, "grad_norm": 1.80332612991333, "learning_rate": 8.323397578799959e-05, "loss": 3.8028, "step": 152850 }, { "epoch": 1.3502534484890232, "grad_norm": 3.8326416015625, "learning_rate": 8.32235922715864e-05, "loss": 2.936, "step": 152900 }, { "epoch": 1.3506949963793073, "grad_norm": 2.770670175552368, "learning_rate": 8.321320618893244e-05, "loss": 2.9054, "step": 152950 }, { "epoch": 1.3511365442695915, "grad_norm": 2.175163507461548, "learning_rate": 8.320281754083994e-05, "loss": 3.6902, "step": 153000 }, { "epoch": 1.3511365442695915, "eval_asr_loss": 0.9108942616081308, "eval_loss": 2.9208271503448486, "eval_runtime": 20.3468, "eval_samples_per_second": 37.745, "eval_steps_per_second": 9.436, "eval_tts_loss": 5.98267045013614, "step": 153000 }, { "epoch": 1.3515780921598757, "grad_norm": 2.5464835166931152, "learning_rate": 8.319242632811132e-05, "loss": 3.3609, "step": 153050 }, { "epoch": 1.3520196400501598, "grad_norm": 2.2353081703186035, "learning_rate": 8.318203255154926e-05, "loss": 3.2348, "step": 153100 }, { "epoch": 1.352461187940444, "grad_norm": 1.6320254802703857, "learning_rate": 8.317163621195654e-05, "loss": 3.2052, "step": 153150 }, { "epoch": 1.3529027358307282, "grad_norm": 2.5109684467315674, "learning_rate": 8.316123731013621e-05, "loss": 3.5582, "step": 153200 }, { "epoch": 1.3533442837210123, "grad_norm": 1.6748558282852173, "learning_rate": 8.31508358468915e-05, "loss": 2.6611, "step": 153250 }, { "epoch": 1.3537858316112965, "grad_norm": 0.912735641002655, "learning_rate": 8.314043182302583e-05, "loss": 3.4849, "step": 153300 }, { "epoch": 1.3542273795015807, "grad_norm": 1.289677381515503, "learning_rate": 8.313002523934284e-05, "loss": 2.6686, "step": 153350 }, { "epoch": 1.3546689273918648, "grad_norm": 2.805980920791626, "learning_rate": 8.311961609664632e-05, "loss": 3.393, "step": 153400 }, { "epoch": 1.355110475282149, "grad_norm": 2.8691484928131104, "learning_rate": 8.310920439574029e-05, "loss": 2.8758, "step": 153450 }, { "epoch": 1.3555520231724332, "grad_norm": 2.143998861312866, "learning_rate": 8.309879013742898e-05, "loss": 3.4699, "step": 153500 }, { "epoch": 1.3559935710627173, "grad_norm": 1.3978588581085205, "learning_rate": 8.308837332251682e-05, "loss": 3.5442, "step": 153550 }, { "epoch": 1.3564351189530015, "grad_norm": 2.2007105350494385, "learning_rate": 8.30779539518084e-05, "loss": 2.9063, "step": 153600 }, { "epoch": 1.356876666843286, "grad_norm": 2.752504348754883, "learning_rate": 8.306753202610853e-05, "loss": 3.4736, "step": 153650 }, { "epoch": 1.35731821473357, "grad_norm": 3.7296342849731445, "learning_rate": 8.305710754622221e-05, "loss": 3.1397, "step": 153700 }, { "epoch": 1.3577597626238542, "grad_norm": 2.1975762844085693, "learning_rate": 8.304668051295464e-05, "loss": 3.0458, "step": 153750 }, { "epoch": 1.3582013105141384, "grad_norm": 1.018903136253357, "learning_rate": 8.303625092711125e-05, "loss": 3.3913, "step": 153800 }, { "epoch": 1.3586428584044226, "grad_norm": 1.9552948474884033, "learning_rate": 8.30258187894976e-05, "loss": 3.2726, "step": 153850 }, { "epoch": 1.3590844062947067, "grad_norm": 6.85836935043335, "learning_rate": 8.301538410091949e-05, "loss": 3.5054, "step": 153900 }, { "epoch": 1.359525954184991, "grad_norm": 2.9646308422088623, "learning_rate": 8.300494686218295e-05, "loss": 3.0811, "step": 153950 }, { "epoch": 1.359967502075275, "grad_norm": 3.7286722660064697, "learning_rate": 8.299450707409411e-05, "loss": 3.1636, "step": 154000 }, { "epoch": 1.3604090499655592, "grad_norm": 1.8345937728881836, "learning_rate": 8.298406473745939e-05, "loss": 3.6935, "step": 154050 }, { "epoch": 1.3608505978558434, "grad_norm": 2.4694504737854004, "learning_rate": 8.297361985308538e-05, "loss": 3.09, "step": 154100 }, { "epoch": 1.3612921457461276, "grad_norm": 1.6211429834365845, "learning_rate": 8.296317242177883e-05, "loss": 3.5075, "step": 154150 }, { "epoch": 1.3617336936364117, "grad_norm": 3.931645154953003, "learning_rate": 8.295272244434672e-05, "loss": 2.8835, "step": 154200 }, { "epoch": 1.3621752415266961, "grad_norm": 2.9920766353607178, "learning_rate": 8.294226992159626e-05, "loss": 3.3678, "step": 154250 }, { "epoch": 1.3626167894169803, "grad_norm": 0.9993804097175598, "learning_rate": 8.293181485433477e-05, "loss": 3.493, "step": 154300 }, { "epoch": 1.3630583373072644, "grad_norm": 1.9843803644180298, "learning_rate": 8.292135724336984e-05, "loss": 3.2889, "step": 154350 }, { "epoch": 1.3634998851975486, "grad_norm": 2.757063865661621, "learning_rate": 8.291089708950925e-05, "loss": 3.6329, "step": 154400 }, { "epoch": 1.3639414330878328, "grad_norm": 2.337742805480957, "learning_rate": 8.290043439356091e-05, "loss": 2.9563, "step": 154450 }, { "epoch": 1.364382980978117, "grad_norm": 3.824697971343994, "learning_rate": 8.288996915633301e-05, "loss": 3.0373, "step": 154500 }, { "epoch": 1.364824528868401, "grad_norm": 3.1068763732910156, "learning_rate": 8.28795013786339e-05, "loss": 3.5153, "step": 154550 }, { "epoch": 1.3652660767586853, "grad_norm": 1.7246665954589844, "learning_rate": 8.286903106127211e-05, "loss": 3.1697, "step": 154600 }, { "epoch": 1.3657076246489694, "grad_norm": 1.9760079383850098, "learning_rate": 8.285855820505638e-05, "loss": 3.2777, "step": 154650 }, { "epoch": 1.3661491725392536, "grad_norm": 1.5915566682815552, "learning_rate": 8.284808281079567e-05, "loss": 3.2404, "step": 154700 }, { "epoch": 1.3665907204295378, "grad_norm": 3.363936424255371, "learning_rate": 8.283760487929911e-05, "loss": 3.1442, "step": 154750 }, { "epoch": 1.367032268319822, "grad_norm": 1.9960918426513672, "learning_rate": 8.282712441137603e-05, "loss": 3.5091, "step": 154800 }, { "epoch": 1.367473816210106, "grad_norm": 1.3129304647445679, "learning_rate": 8.281664140783594e-05, "loss": 2.9314, "step": 154850 }, { "epoch": 1.3679153641003903, "grad_norm": 2.3583755493164062, "learning_rate": 8.280615586948859e-05, "loss": 3.226, "step": 154900 }, { "epoch": 1.3683569119906744, "grad_norm": 3.2506699562072754, "learning_rate": 8.27956677971439e-05, "loss": 3.0299, "step": 154950 }, { "epoch": 1.3687984598809586, "grad_norm": 1.7959414720535278, "learning_rate": 8.278517719161196e-05, "loss": 3.1626, "step": 155000 }, { "epoch": 1.3692400077712428, "grad_norm": 1.9496901035308838, "learning_rate": 8.277468405370308e-05, "loss": 3.2722, "step": 155050 }, { "epoch": 1.369681555661527, "grad_norm": 2.7062554359436035, "learning_rate": 8.27643983224214e-05, "loss": 3.1118, "step": 155100 }, { "epoch": 1.370123103551811, "grad_norm": 5.150788307189941, "learning_rate": 8.275390017279756e-05, "loss": 3.0469, "step": 155150 }, { "epoch": 1.3705646514420953, "grad_norm": 4.183962345123291, "learning_rate": 8.274339949321266e-05, "loss": 2.9026, "step": 155200 }, { "epoch": 1.3710061993323797, "grad_norm": 3.3120381832122803, "learning_rate": 8.273289628447782e-05, "loss": 2.9835, "step": 155250 }, { "epoch": 1.3714477472226638, "grad_norm": 5.01318883895874, "learning_rate": 8.272239054740431e-05, "loss": 3.1624, "step": 155300 }, { "epoch": 1.371889295112948, "grad_norm": 2.8328940868377686, "learning_rate": 8.27118822828036e-05, "loss": 3.1617, "step": 155350 }, { "epoch": 1.3723308430032322, "grad_norm": 1.2944576740264893, "learning_rate": 8.270137149148737e-05, "loss": 2.9411, "step": 155400 }, { "epoch": 1.3727723908935163, "grad_norm": 0.6006796956062317, "learning_rate": 8.269085817426751e-05, "loss": 3.0553, "step": 155450 }, { "epoch": 1.3732139387838005, "grad_norm": 2.5675642490386963, "learning_rate": 8.268034233195605e-05, "loss": 3.7048, "step": 155500 }, { "epoch": 1.3736554866740847, "grad_norm": 2.5069117546081543, "learning_rate": 8.266982396536527e-05, "loss": 2.8343, "step": 155550 }, { "epoch": 1.3740970345643688, "grad_norm": 2.1562747955322266, "learning_rate": 8.265930307530763e-05, "loss": 3.1942, "step": 155600 }, { "epoch": 1.374538582454653, "grad_norm": 2.492576837539673, "learning_rate": 8.264877966259577e-05, "loss": 3.0843, "step": 155650 }, { "epoch": 1.3749801303449372, "grad_norm": 3.6967947483062744, "learning_rate": 8.26382537280425e-05, "loss": 3.0236, "step": 155700 }, { "epoch": 1.3754216782352213, "grad_norm": 2.7098875045776367, "learning_rate": 8.262772527246094e-05, "loss": 3.6515, "step": 155750 }, { "epoch": 1.3758632261255055, "grad_norm": 2.622032642364502, "learning_rate": 8.261719429666426e-05, "loss": 3.2831, "step": 155800 }, { "epoch": 1.3763047740157899, "grad_norm": 4.229182243347168, "learning_rate": 8.26066608014659e-05, "loss": 3.383, "step": 155850 }, { "epoch": 1.376746321906074, "grad_norm": 2.306593179702759, "learning_rate": 8.259612478767949e-05, "loss": 2.8599, "step": 155900 }, { "epoch": 1.3771878697963582, "grad_norm": 4.549513339996338, "learning_rate": 8.258579705141899e-05, "loss": 2.933, "step": 155950 }, { "epoch": 1.3776294176866424, "grad_norm": 5.794224262237549, "learning_rate": 8.257525605322936e-05, "loss": 3.1477, "step": 156000 }, { "epoch": 1.3776294176866424, "eval_asr_loss": 0.9160445507592706, "eval_loss": 2.932405471801758, "eval_runtime": 20.6052, "eval_samples_per_second": 37.272, "eval_steps_per_second": 9.318, "eval_tts_loss": 6.016129709398038, "step": 156000 }, { "epoch": 1.3780709655769265, "grad_norm": 1.7329176664352417, "learning_rate": 8.256471253887742e-05, "loss": 3.2825, "step": 156050 }, { "epoch": 1.3785125134672107, "grad_norm": 4.851804256439209, "learning_rate": 8.255416650917758e-05, "loss": 3.6591, "step": 156100 }, { "epoch": 1.3789540613574949, "grad_norm": 0.9085479378700256, "learning_rate": 8.254361796494443e-05, "loss": 2.9981, "step": 156150 }, { "epoch": 1.379395609247779, "grad_norm": 1.4775174856185913, "learning_rate": 8.253306690699277e-05, "loss": 3.4719, "step": 156200 }, { "epoch": 1.3798371571380632, "grad_norm": 2.0856218338012695, "learning_rate": 8.252251333613755e-05, "loss": 3.1563, "step": 156250 }, { "epoch": 1.3802787050283474, "grad_norm": 1.0467510223388672, "learning_rate": 8.251195725319394e-05, "loss": 3.5563, "step": 156300 }, { "epoch": 1.3807202529186315, "grad_norm": 1.6420974731445312, "learning_rate": 8.250139865897735e-05, "loss": 3.1661, "step": 156350 }, { "epoch": 1.3811618008089157, "grad_norm": 1.7073570489883423, "learning_rate": 8.249083755430332e-05, "loss": 3.0233, "step": 156400 }, { "epoch": 1.3816033486991999, "grad_norm": 3.6677961349487305, "learning_rate": 8.248027393998758e-05, "loss": 3.0304, "step": 156450 }, { "epoch": 1.382044896589484, "grad_norm": 3.6211111545562744, "learning_rate": 8.246970781684611e-05, "loss": 3.0123, "step": 156500 }, { "epoch": 1.3824864444797682, "grad_norm": 2.3987412452697754, "learning_rate": 8.245913918569506e-05, "loss": 3.2113, "step": 156550 }, { "epoch": 1.3829279923700524, "grad_norm": 3.1175930500030518, "learning_rate": 8.244856804735076e-05, "loss": 3.2382, "step": 156600 }, { "epoch": 1.3833695402603365, "grad_norm": 1.2450615167617798, "learning_rate": 8.243799440262972e-05, "loss": 3.434, "step": 156650 }, { "epoch": 1.3838110881506207, "grad_norm": 2.2003071308135986, "learning_rate": 8.242741825234866e-05, "loss": 3.3624, "step": 156700 }, { "epoch": 1.3842526360409049, "grad_norm": 3.6320629119873047, "learning_rate": 8.241683959732453e-05, "loss": 3.0379, "step": 156750 }, { "epoch": 1.3846941839311893, "grad_norm": 2.458106756210327, "learning_rate": 8.240625843837444e-05, "loss": 3.3652, "step": 156800 }, { "epoch": 1.3851357318214734, "grad_norm": 2.492220878601074, "learning_rate": 8.239567477631569e-05, "loss": 3.3358, "step": 156850 }, { "epoch": 1.3855772797117576, "grad_norm": 1.127402424812317, "learning_rate": 8.238508861196574e-05, "loss": 3.317, "step": 156900 }, { "epoch": 1.3860188276020418, "grad_norm": 2.1057465076446533, "learning_rate": 8.237449994614234e-05, "loss": 3.3922, "step": 156950 }, { "epoch": 1.386460375492326, "grad_norm": 0.6556951999664307, "learning_rate": 8.236390877966332e-05, "loss": 3.4174, "step": 157000 }, { "epoch": 1.38690192338261, "grad_norm": 2.198021173477173, "learning_rate": 8.235331511334681e-05, "loss": 3.4776, "step": 157050 }, { "epoch": 1.3873434712728943, "grad_norm": 2.8309218883514404, "learning_rate": 8.234271894801105e-05, "loss": 2.996, "step": 157100 }, { "epoch": 1.3877850191631784, "grad_norm": 3.436307430267334, "learning_rate": 8.23321202844745e-05, "loss": 3.3207, "step": 157150 }, { "epoch": 1.3882265670534626, "grad_norm": 5.650524139404297, "learning_rate": 8.232151912355583e-05, "loss": 3.2116, "step": 157200 }, { "epoch": 1.3886681149437468, "grad_norm": 3.0345571041107178, "learning_rate": 8.231091546607388e-05, "loss": 3.2176, "step": 157250 }, { "epoch": 1.389109662834031, "grad_norm": 1.683851718902588, "learning_rate": 8.230030931284771e-05, "loss": 3.0684, "step": 157300 }, { "epoch": 1.389551210724315, "grad_norm": 3.6749675273895264, "learning_rate": 8.228970066469655e-05, "loss": 2.7936, "step": 157350 }, { "epoch": 1.3899927586145995, "grad_norm": 2.033708333969116, "learning_rate": 8.227908952243981e-05, "loss": 2.6264, "step": 157400 }, { "epoch": 1.3904343065048836, "grad_norm": 3.007284641265869, "learning_rate": 8.226847588689713e-05, "loss": 3.3704, "step": 157450 }, { "epoch": 1.3908758543951678, "grad_norm": 3.3445730209350586, "learning_rate": 8.22578597588883e-05, "loss": 2.9904, "step": 157500 }, { "epoch": 1.391317402285452, "grad_norm": 5.233058929443359, "learning_rate": 8.224724113923336e-05, "loss": 3.0608, "step": 157550 }, { "epoch": 1.3917589501757361, "grad_norm": 2.4576175212860107, "learning_rate": 8.223662002875249e-05, "loss": 3.3851, "step": 157600 }, { "epoch": 1.3922004980660203, "grad_norm": 3.7473318576812744, "learning_rate": 8.222599642826608e-05, "loss": 3.0497, "step": 157650 }, { "epoch": 1.3926420459563045, "grad_norm": 4.362736225128174, "learning_rate": 8.22153703385947e-05, "loss": 3.0953, "step": 157700 }, { "epoch": 1.3930835938465886, "grad_norm": 2.2473487854003906, "learning_rate": 8.220474176055917e-05, "loss": 3.354, "step": 157750 }, { "epoch": 1.3935251417368728, "grad_norm": 5.24487829208374, "learning_rate": 8.21941106949804e-05, "loss": 3.2924, "step": 157800 }, { "epoch": 1.393966689627157, "grad_norm": 5.974734783172607, "learning_rate": 8.218347714267958e-05, "loss": 2.666, "step": 157850 }, { "epoch": 1.3944082375174411, "grad_norm": 1.5689363479614258, "learning_rate": 8.217284110447807e-05, "loss": 3.3073, "step": 157900 }, { "epoch": 1.3948497854077253, "grad_norm": 2.4762063026428223, "learning_rate": 8.216220258119739e-05, "loss": 2.9431, "step": 157950 }, { "epoch": 1.3952913332980095, "grad_norm": 1.6257416009902954, "learning_rate": 8.21515615736593e-05, "loss": 3.1226, "step": 158000 }, { "epoch": 1.3957328811882936, "grad_norm": 1.930908203125, "learning_rate": 8.21409180826857e-05, "loss": 3.1131, "step": 158050 }, { "epoch": 1.3961744290785778, "grad_norm": 2.8949382305145264, "learning_rate": 8.213027210909874e-05, "loss": 3.1718, "step": 158100 }, { "epoch": 1.396615976968862, "grad_norm": 3.7047338485717773, "learning_rate": 8.21196236537207e-05, "loss": 3.3025, "step": 158150 }, { "epoch": 1.3970575248591461, "grad_norm": 3.0060231685638428, "learning_rate": 8.210897271737412e-05, "loss": 3.3839, "step": 158200 }, { "epoch": 1.3974990727494303, "grad_norm": 2.41214919090271, "learning_rate": 8.209831930088167e-05, "loss": 3.3249, "step": 158250 }, { "epoch": 1.3979406206397145, "grad_norm": 1.3485262393951416, "learning_rate": 8.208766340506623e-05, "loss": 3.095, "step": 158300 }, { "epoch": 1.3983821685299986, "grad_norm": 3.0010628700256348, "learning_rate": 8.207700503075089e-05, "loss": 3.0024, "step": 158350 }, { "epoch": 1.398823716420283, "grad_norm": 1.9099316596984863, "learning_rate": 8.206634417875891e-05, "loss": 3.6751, "step": 158400 }, { "epoch": 1.3992652643105672, "grad_norm": 1.0043919086456299, "learning_rate": 8.205568084991377e-05, "loss": 3.402, "step": 158450 }, { "epoch": 1.3997068122008514, "grad_norm": 0.9302772283554077, "learning_rate": 8.204501504503913e-05, "loss": 3.4862, "step": 158500 }, { "epoch": 1.4001483600911355, "grad_norm": 3.9256772994995117, "learning_rate": 8.203434676495876e-05, "loss": 3.1125, "step": 158550 }, { "epoch": 1.4005899079814197, "grad_norm": 1.6073594093322754, "learning_rate": 8.202367601049678e-05, "loss": 3.2281, "step": 158600 }, { "epoch": 1.4010314558717039, "grad_norm": 1.277282953262329, "learning_rate": 8.201300278247736e-05, "loss": 3.1331, "step": 158650 }, { "epoch": 1.401473003761988, "grad_norm": 2.780365467071533, "learning_rate": 8.200232708172495e-05, "loss": 3.6974, "step": 158700 }, { "epoch": 1.4019145516522722, "grad_norm": 3.2038660049438477, "learning_rate": 8.199164890906415e-05, "loss": 3.2937, "step": 158750 }, { "epoch": 1.4023560995425564, "grad_norm": 5.772561073303223, "learning_rate": 8.198096826531974e-05, "loss": 3.1028, "step": 158800 }, { "epoch": 1.4027976474328405, "grad_norm": 2.543661117553711, "learning_rate": 8.197028515131673e-05, "loss": 3.1588, "step": 158850 }, { "epoch": 1.4032391953231247, "grad_norm": 2.7590420246124268, "learning_rate": 8.195959956788031e-05, "loss": 3.1061, "step": 158900 }, { "epoch": 1.4036807432134089, "grad_norm": 2.1173219680786133, "learning_rate": 8.194891151583579e-05, "loss": 3.132, "step": 158950 }, { "epoch": 1.4041222911036932, "grad_norm": 1.821299433708191, "learning_rate": 8.19382209960088e-05, "loss": 3.4133, "step": 159000 }, { "epoch": 1.4041222911036932, "eval_asr_loss": 0.9242424718126228, "eval_loss": 2.9176743030548096, "eval_runtime": 20.4126, "eval_samples_per_second": 37.624, "eval_steps_per_second": 9.406, "eval_tts_loss": 5.999804931002174, "step": 159000 }, { "epoch": 1.4045638389939774, "grad_norm": 2.2017438411712646, "learning_rate": 8.192752800922508e-05, "loss": 3.2473, "step": 159050 }, { "epoch": 1.4050053868842616, "grad_norm": 5.148829460144043, "learning_rate": 8.191683255631053e-05, "loss": 3.1951, "step": 159100 }, { "epoch": 1.4054469347745457, "grad_norm": 2.3548662662506104, "learning_rate": 8.190613463809134e-05, "loss": 3.5011, "step": 159150 }, { "epoch": 1.40588848266483, "grad_norm": 3.541984796524048, "learning_rate": 8.189543425539378e-05, "loss": 2.8037, "step": 159200 }, { "epoch": 1.406330030555114, "grad_norm": 2.6860177516937256, "learning_rate": 8.18847314090444e-05, "loss": 3.3826, "step": 159250 }, { "epoch": 1.4067715784453982, "grad_norm": 2.60087251663208, "learning_rate": 8.187402609986989e-05, "loss": 3.7379, "step": 159300 }, { "epoch": 1.4072131263356824, "grad_norm": 4.26237154006958, "learning_rate": 8.186331832869715e-05, "loss": 2.9288, "step": 159350 }, { "epoch": 1.4076546742259666, "grad_norm": 2.6659810543060303, "learning_rate": 8.185260809635326e-05, "loss": 3.4246, "step": 159400 }, { "epoch": 1.4080962221162507, "grad_norm": 3.74465012550354, "learning_rate": 8.184189540366551e-05, "loss": 2.8438, "step": 159450 }, { "epoch": 1.408537770006535, "grad_norm": 4.397085189819336, "learning_rate": 8.183118025146134e-05, "loss": 3.489, "step": 159500 }, { "epoch": 1.408979317896819, "grad_norm": 4.9571452140808105, "learning_rate": 8.182046264056842e-05, "loss": 3.2969, "step": 159550 }, { "epoch": 1.4094208657871032, "grad_norm": 1.2717839479446411, "learning_rate": 8.180974257181457e-05, "loss": 3.4956, "step": 159600 }, { "epoch": 1.4098624136773874, "grad_norm": 2.1157987117767334, "learning_rate": 8.179902004602787e-05, "loss": 3.4988, "step": 159650 }, { "epoch": 1.4103039615676716, "grad_norm": 2.7032673358917236, "learning_rate": 8.178829506403651e-05, "loss": 3.6502, "step": 159700 }, { "epoch": 1.4107455094579557, "grad_norm": 1.5434041023254395, "learning_rate": 8.177756762666892e-05, "loss": 3.6715, "step": 159750 }, { "epoch": 1.41118705734824, "grad_norm": 2.1543235778808594, "learning_rate": 8.17668377347537e-05, "loss": 3.4379, "step": 159800 }, { "epoch": 1.411628605238524, "grad_norm": 2.2687978744506836, "learning_rate": 8.175610538911961e-05, "loss": 3.2811, "step": 159850 }, { "epoch": 1.4120701531288082, "grad_norm": 3.682598829269409, "learning_rate": 8.174537059059571e-05, "loss": 3.3635, "step": 159900 }, { "epoch": 1.4125117010190924, "grad_norm": 3.4800243377685547, "learning_rate": 8.173463334001109e-05, "loss": 3.2212, "step": 159950 }, { "epoch": 1.4129532489093768, "grad_norm": 2.2715086936950684, "learning_rate": 8.172389363819514e-05, "loss": 3.5366, "step": 160000 }, { "epoch": 1.413394796799661, "grad_norm": 4.865034103393555, "learning_rate": 8.171315148597744e-05, "loss": 3.4287, "step": 160050 }, { "epoch": 1.4138363446899451, "grad_norm": 1.0195260047912598, "learning_rate": 8.17024068841877e-05, "loss": 3.4387, "step": 160100 }, { "epoch": 1.4142778925802293, "grad_norm": 0.7502059936523438, "learning_rate": 8.169165983365588e-05, "loss": 2.8933, "step": 160150 }, { "epoch": 1.4147194404705135, "grad_norm": 1.9067836999893188, "learning_rate": 8.168091033521204e-05, "loss": 3.5911, "step": 160200 }, { "epoch": 1.4151609883607976, "grad_norm": 1.9587763547897339, "learning_rate": 8.167015838968653e-05, "loss": 3.1981, "step": 160250 }, { "epoch": 1.4156025362510818, "grad_norm": 2.4803414344787598, "learning_rate": 8.165940399790986e-05, "loss": 2.9892, "step": 160300 }, { "epoch": 1.416044084141366, "grad_norm": 1.6044929027557373, "learning_rate": 8.164864716071266e-05, "loss": 3.2477, "step": 160350 }, { "epoch": 1.4164856320316501, "grad_norm": 2.8173727989196777, "learning_rate": 8.163788787892586e-05, "loss": 2.7043, "step": 160400 }, { "epoch": 1.4169271799219343, "grad_norm": 2.8519928455352783, "learning_rate": 8.16271261533805e-05, "loss": 3.2307, "step": 160450 }, { "epoch": 1.4173687278122185, "grad_norm": 2.7776854038238525, "learning_rate": 8.161636198490784e-05, "loss": 3.343, "step": 160500 }, { "epoch": 1.4178102757025026, "grad_norm": 2.121873378753662, "learning_rate": 8.160559537433932e-05, "loss": 3.2525, "step": 160550 }, { "epoch": 1.418251823592787, "grad_norm": 2.0176658630371094, "learning_rate": 8.159482632250655e-05, "loss": 3.5169, "step": 160600 }, { "epoch": 1.4186933714830712, "grad_norm": 2.57165265083313, "learning_rate": 8.158405483024138e-05, "loss": 3.3809, "step": 160650 }, { "epoch": 1.4191349193733553, "grad_norm": 0.6696668267250061, "learning_rate": 8.15732808983758e-05, "loss": 2.9887, "step": 160700 }, { "epoch": 1.4195764672636395, "grad_norm": 2.2909727096557617, "learning_rate": 8.156250452774198e-05, "loss": 3.1004, "step": 160750 }, { "epoch": 1.4200180151539237, "grad_norm": 8.691178321838379, "learning_rate": 8.155172571917233e-05, "loss": 2.8258, "step": 160800 }, { "epoch": 1.4204595630442078, "grad_norm": 3.109480381011963, "learning_rate": 8.154094447349943e-05, "loss": 3.3217, "step": 160850 }, { "epoch": 1.420901110934492, "grad_norm": 4.277402400970459, "learning_rate": 8.153016079155604e-05, "loss": 3.426, "step": 160900 }, { "epoch": 1.4213426588247762, "grad_norm": 1.419296145439148, "learning_rate": 8.151937467417508e-05, "loss": 3.1784, "step": 160950 }, { "epoch": 1.4217842067150603, "grad_norm": 1.5565255880355835, "learning_rate": 8.150858612218972e-05, "loss": 3.0897, "step": 161000 }, { "epoch": 1.4222257546053445, "grad_norm": 3.24021315574646, "learning_rate": 8.149779513643325e-05, "loss": 3.2465, "step": 161050 }, { "epoch": 1.4226673024956287, "grad_norm": 5.220244407653809, "learning_rate": 8.148721760995049e-05, "loss": 3.3344, "step": 161100 }, { "epoch": 1.4231088503859128, "grad_norm": 1.918340802192688, "learning_rate": 8.147642180778646e-05, "loss": 3.5066, "step": 161150 }, { "epoch": 1.423550398276197, "grad_norm": 3.6554203033447266, "learning_rate": 8.146562357433577e-05, "loss": 3.2737, "step": 161200 }, { "epoch": 1.4239919461664812, "grad_norm": 1.7127411365509033, "learning_rate": 8.145482291043247e-05, "loss": 3.3952, "step": 161250 }, { "epoch": 1.4244334940567653, "grad_norm": 1.1752212047576904, "learning_rate": 8.144401981691085e-05, "loss": 2.9625, "step": 161300 }, { "epoch": 1.4248750419470495, "grad_norm": 1.228081464767456, "learning_rate": 8.143321429460533e-05, "loss": 3.2811, "step": 161350 }, { "epoch": 1.4253165898373337, "grad_norm": 3.1649067401885986, "learning_rate": 8.142240634435052e-05, "loss": 3.2813, "step": 161400 }, { "epoch": 1.4257581377276178, "grad_norm": 2.143840789794922, "learning_rate": 8.141159596698131e-05, "loss": 3.4526, "step": 161450 }, { "epoch": 1.426199685617902, "grad_norm": 4.533790588378906, "learning_rate": 8.140078316333266e-05, "loss": 3.4879, "step": 161500 }, { "epoch": 1.4266412335081862, "grad_norm": 3.0826544761657715, "learning_rate": 8.138996793423979e-05, "loss": 3.1003, "step": 161550 }, { "epoch": 1.4270827813984706, "grad_norm": 8.362160682678223, "learning_rate": 8.137915028053806e-05, "loss": 2.9193, "step": 161600 }, { "epoch": 1.4275243292887547, "grad_norm": 3.49751353263855, "learning_rate": 8.136854662836015e-05, "loss": 3.1116, "step": 161650 }, { "epoch": 1.427965877179039, "grad_norm": 1.2915788888931274, "learning_rate": 8.135772417639821e-05, "loss": 3.0923, "step": 161700 }, { "epoch": 1.428407425069323, "grad_norm": 2.0796115398406982, "learning_rate": 8.134689930231797e-05, "loss": 3.4672, "step": 161750 }, { "epoch": 1.4288489729596072, "grad_norm": 1.242505431175232, "learning_rate": 8.133607200695558e-05, "loss": 3.2748, "step": 161800 }, { "epoch": 1.4292905208498914, "grad_norm": 5.859312057495117, "learning_rate": 8.132524229114736e-05, "loss": 2.9677, "step": 161850 }, { "epoch": 1.4297320687401756, "grad_norm": 2.893573522567749, "learning_rate": 8.131441015572977e-05, "loss": 3.486, "step": 161900 }, { "epoch": 1.4301736166304597, "grad_norm": 2.6132123470306396, "learning_rate": 8.130357560153958e-05, "loss": 3.5871, "step": 161950 }, { "epoch": 1.4306151645207439, "grad_norm": 2.7765181064605713, "learning_rate": 8.129273862941361e-05, "loss": 2.9948, "step": 162000 }, { "epoch": 1.4306151645207439, "eval_asr_loss": 0.9175770946416757, "eval_loss": 2.9294345378875732, "eval_runtime": 20.6294, "eval_samples_per_second": 37.228, "eval_steps_per_second": 9.307, "eval_tts_loss": 6.006233162385569, "step": 162000 }, { "epoch": 1.431056712411028, "grad_norm": 2.137802839279175, "learning_rate": 8.128189924018891e-05, "loss": 3.1945, "step": 162050 }, { "epoch": 1.4314982603013122, "grad_norm": 3.257875442504883, "learning_rate": 8.127105743470277e-05, "loss": 3.7324, "step": 162100 }, { "epoch": 1.4319398081915964, "grad_norm": 4.234621047973633, "learning_rate": 8.126021321379263e-05, "loss": 3.1456, "step": 162150 }, { "epoch": 1.4323813560818808, "grad_norm": 2.000067710876465, "learning_rate": 8.124936657829607e-05, "loss": 3.0195, "step": 162200 }, { "epoch": 1.432822903972165, "grad_norm": 0.7076943516731262, "learning_rate": 8.123851752905096e-05, "loss": 3.2913, "step": 162250 }, { "epoch": 1.433264451862449, "grad_norm": 1.7147762775421143, "learning_rate": 8.122766606689523e-05, "loss": 3.3676, "step": 162300 }, { "epoch": 1.4337059997527333, "grad_norm": 1.8127288818359375, "learning_rate": 8.121681219266713e-05, "loss": 3.4248, "step": 162350 }, { "epoch": 1.4341475476430174, "grad_norm": 2.52752685546875, "learning_rate": 8.120595590720499e-05, "loss": 3.3246, "step": 162400 }, { "epoch": 1.4345890955333016, "grad_norm": 2.496375322341919, "learning_rate": 8.119509721134735e-05, "loss": 3.0343, "step": 162450 }, { "epoch": 1.4350306434235858, "grad_norm": 0.9835373163223267, "learning_rate": 8.118423610593301e-05, "loss": 3.2281, "step": 162500 }, { "epoch": 1.43547219131387, "grad_norm": 3.728062629699707, "learning_rate": 8.117337259180084e-05, "loss": 3.1614, "step": 162550 }, { "epoch": 1.435913739204154, "grad_norm": 4.0812907218933105, "learning_rate": 8.116250666979e-05, "loss": 3.3182, "step": 162600 }, { "epoch": 1.4363552870944383, "grad_norm": 1.9645757675170898, "learning_rate": 8.115163834073977e-05, "loss": 3.0928, "step": 162650 }, { "epoch": 1.4367968349847224, "grad_norm": 2.8638205528259277, "learning_rate": 8.114076760548963e-05, "loss": 3.1087, "step": 162700 }, { "epoch": 1.4372383828750066, "grad_norm": 4.352683067321777, "learning_rate": 8.112989446487927e-05, "loss": 3.3871, "step": 162750 }, { "epoch": 1.4376799307652908, "grad_norm": 3.6510112285614014, "learning_rate": 8.11190189197485e-05, "loss": 2.9083, "step": 162800 }, { "epoch": 1.438121478655575, "grad_norm": 1.543848991394043, "learning_rate": 8.110814097093743e-05, "loss": 3.3145, "step": 162850 }, { "epoch": 1.438563026545859, "grad_norm": 1.975798487663269, "learning_rate": 8.109726061928627e-05, "loss": 3.253, "step": 162900 }, { "epoch": 1.4390045744361433, "grad_norm": 3.0668487548828125, "learning_rate": 8.10863778656354e-05, "loss": 3.3423, "step": 162950 }, { "epoch": 1.4394461223264274, "grad_norm": 0.9884811639785767, "learning_rate": 8.107549271082545e-05, "loss": 2.8906, "step": 163000 }, { "epoch": 1.4398876702167116, "grad_norm": 2.0988800525665283, "learning_rate": 8.106460515569721e-05, "loss": 3.0113, "step": 163050 }, { "epoch": 1.4403292181069958, "grad_norm": 2.1560781002044678, "learning_rate": 8.105371520109163e-05, "loss": 3.366, "step": 163100 }, { "epoch": 1.44077076599728, "grad_norm": 3.452460765838623, "learning_rate": 8.104282284784989e-05, "loss": 3.4599, "step": 163150 }, { "epoch": 1.4412123138875643, "grad_norm": 2.827583074569702, "learning_rate": 8.10319280968133e-05, "loss": 3.0957, "step": 163200 }, { "epoch": 1.4416538617778485, "grad_norm": 5.022325038909912, "learning_rate": 8.102103094882343e-05, "loss": 2.9519, "step": 163250 }, { "epoch": 1.4420954096681327, "grad_norm": 2.3892874717712402, "learning_rate": 8.101013140472195e-05, "loss": 3.0009, "step": 163300 }, { "epoch": 1.4425369575584168, "grad_norm": 2.8516552448272705, "learning_rate": 8.09992294653508e-05, "loss": 3.0863, "step": 163350 }, { "epoch": 1.442978505448701, "grad_norm": 1.9601846933364868, "learning_rate": 8.098832513155203e-05, "loss": 3.2776, "step": 163400 }, { "epoch": 1.4434200533389852, "grad_norm": 2.186326503753662, "learning_rate": 8.097741840416791e-05, "loss": 3.3627, "step": 163450 }, { "epoch": 1.4438616012292693, "grad_norm": 4.046080589294434, "learning_rate": 8.096650928404092e-05, "loss": 2.9213, "step": 163500 }, { "epoch": 1.4443031491195535, "grad_norm": 2.9167070388793945, "learning_rate": 8.095559777201365e-05, "loss": 3.2593, "step": 163550 }, { "epoch": 1.4447446970098377, "grad_norm": 1.3748699426651, "learning_rate": 8.094468386892896e-05, "loss": 3.3712, "step": 163600 }, { "epoch": 1.4451862449001218, "grad_norm": 2.320512056350708, "learning_rate": 8.093376757562985e-05, "loss": 3.0272, "step": 163650 }, { "epoch": 1.445627792790406, "grad_norm": 7.251769542694092, "learning_rate": 8.092284889295949e-05, "loss": 3.4075, "step": 163700 }, { "epoch": 1.4460693406806902, "grad_norm": 2.5889413356781006, "learning_rate": 8.091192782176129e-05, "loss": 2.8281, "step": 163750 }, { "epoch": 1.4465108885709745, "grad_norm": 2.878997325897217, "learning_rate": 8.090100436287877e-05, "loss": 3.5972, "step": 163800 }, { "epoch": 1.4469524364612587, "grad_norm": 1.8390371799468994, "learning_rate": 8.089007851715571e-05, "loss": 3.3453, "step": 163850 }, { "epoch": 1.4473939843515429, "grad_norm": 2.4774179458618164, "learning_rate": 8.087915028543603e-05, "loss": 2.9448, "step": 163900 }, { "epoch": 1.447835532241827, "grad_norm": 2.5568761825561523, "learning_rate": 8.086821966856381e-05, "loss": 2.6824, "step": 163950 }, { "epoch": 1.4482770801321112, "grad_norm": 2.712879180908203, "learning_rate": 8.085728666738341e-05, "loss": 3.2722, "step": 164000 }, { "epoch": 1.4487186280223954, "grad_norm": 2.8086280822753906, "learning_rate": 8.084635128273926e-05, "loss": 3.4506, "step": 164050 }, { "epoch": 1.4491601759126795, "grad_norm": 2.9732656478881836, "learning_rate": 8.083541351547603e-05, "loss": 3.462, "step": 164100 }, { "epoch": 1.4496017238029637, "grad_norm": 3.0905792713165283, "learning_rate": 8.082447336643861e-05, "loss": 2.8562, "step": 164150 }, { "epoch": 1.4500432716932479, "grad_norm": 2.6996688842773438, "learning_rate": 8.0813530836472e-05, "loss": 3.3813, "step": 164200 }, { "epoch": 1.450484819583532, "grad_norm": 1.9003500938415527, "learning_rate": 8.08025859264214e-05, "loss": 3.3731, "step": 164250 }, { "epoch": 1.4509263674738162, "grad_norm": 1.5996307134628296, "learning_rate": 8.079163863713227e-05, "loss": 3.1851, "step": 164300 }, { "epoch": 1.4513679153641004, "grad_norm": 1.505954384803772, "learning_rate": 8.078068896945013e-05, "loss": 3.2188, "step": 164350 }, { "epoch": 1.4518094632543845, "grad_norm": 3.6162900924682617, "learning_rate": 8.076973692422081e-05, "loss": 3.2365, "step": 164400 }, { "epoch": 1.4522510111446687, "grad_norm": 2.8672714233398438, "learning_rate": 8.07587825022902e-05, "loss": 3.571, "step": 164450 }, { "epoch": 1.4526925590349529, "grad_norm": 1.8721016645431519, "learning_rate": 8.07478257045045e-05, "loss": 2.9197, "step": 164500 }, { "epoch": 1.453134106925237, "grad_norm": 3.9616005420684814, "learning_rate": 8.073686653170997e-05, "loss": 3.3142, "step": 164550 }, { "epoch": 1.4535756548155212, "grad_norm": 1.1517298221588135, "learning_rate": 8.072590498475316e-05, "loss": 3.1608, "step": 164600 }, { "epoch": 1.4540172027058054, "grad_norm": 3.2245967388153076, "learning_rate": 8.071494106448074e-05, "loss": 3.3427, "step": 164650 }, { "epoch": 1.4544587505960895, "grad_norm": 1.755993366241455, "learning_rate": 8.070397477173957e-05, "loss": 3.0496, "step": 164700 }, { "epoch": 1.4549002984863737, "grad_norm": 3.7342405319213867, "learning_rate": 8.06930061073767e-05, "loss": 3.2357, "step": 164750 }, { "epoch": 1.455341846376658, "grad_norm": 2.7675912380218506, "learning_rate": 8.06820350722394e-05, "loss": 3.2516, "step": 164800 }, { "epoch": 1.4557833942669423, "grad_norm": 2.1434454917907715, "learning_rate": 8.067106166717504e-05, "loss": 2.7675, "step": 164850 }, { "epoch": 1.4562249421572264, "grad_norm": 1.979413390159607, "learning_rate": 8.066008589303127e-05, "loss": 3.609, "step": 164900 }, { "epoch": 1.4566664900475106, "grad_norm": 4.171022891998291, "learning_rate": 8.064910775065584e-05, "loss": 3.2429, "step": 164950 }, { "epoch": 1.4571080379377948, "grad_norm": 11.221723556518555, "learning_rate": 8.063812724089674e-05, "loss": 2.7889, "step": 165000 }, { "epoch": 1.4571080379377948, "eval_asr_loss": 0.9156917675408397, "eval_loss": 2.9276485443115234, "eval_runtime": 20.3059, "eval_samples_per_second": 37.821, "eval_steps_per_second": 9.455, "eval_tts_loss": 5.998138508365599, "step": 165000 }, { "epoch": 1.457549585828079, "grad_norm": 5.32289981842041, "learning_rate": 8.062714436460213e-05, "loss": 3.4838, "step": 165050 }, { "epoch": 1.457991133718363, "grad_norm": 3.618535280227661, "learning_rate": 8.06161591226203e-05, "loss": 2.8096, "step": 165100 }, { "epoch": 1.4584326816086473, "grad_norm": 1.2320194244384766, "learning_rate": 8.060517151579983e-05, "loss": 3.3336, "step": 165150 }, { "epoch": 1.4588742294989314, "grad_norm": 1.9119493961334229, "learning_rate": 8.059418154498936e-05, "loss": 3.0939, "step": 165200 }, { "epoch": 1.4593157773892156, "grad_norm": 2.608098268508911, "learning_rate": 8.05831892110378e-05, "loss": 2.8796, "step": 165250 }, { "epoch": 1.4597573252794998, "grad_norm": 0.9068975448608398, "learning_rate": 8.05721945147942e-05, "loss": 3.6585, "step": 165300 }, { "epoch": 1.460198873169784, "grad_norm": 2.3178799152374268, "learning_rate": 8.056119745710784e-05, "loss": 3.5646, "step": 165350 }, { "epoch": 1.4606404210600683, "grad_norm": 1.2207263708114624, "learning_rate": 8.055019803882813e-05, "loss": 2.7127, "step": 165400 }, { "epoch": 1.4610819689503525, "grad_norm": 2.106874465942383, "learning_rate": 8.053919626080466e-05, "loss": 2.8729, "step": 165450 }, { "epoch": 1.4615235168406366, "grad_norm": 1.8500902652740479, "learning_rate": 8.052819212388725e-05, "loss": 3.3694, "step": 165500 }, { "epoch": 1.4619650647309208, "grad_norm": 3.1178908348083496, "learning_rate": 8.051718562892587e-05, "loss": 3.0963, "step": 165550 }, { "epoch": 1.462406612621205, "grad_norm": 3.855515718460083, "learning_rate": 8.050617677677069e-05, "loss": 3.8821, "step": 165600 }, { "epoch": 1.4628481605114891, "grad_norm": 1.1022988557815552, "learning_rate": 8.049516556827203e-05, "loss": 3.3908, "step": 165650 }, { "epoch": 1.4632897084017733, "grad_norm": 8.135658264160156, "learning_rate": 8.048415200428042e-05, "loss": 3.2926, "step": 165700 }, { "epoch": 1.4637312562920575, "grad_norm": 2.4983675479888916, "learning_rate": 8.047313608564658e-05, "loss": 3.1689, "step": 165750 }, { "epoch": 1.4641728041823416, "grad_norm": 3.4673867225646973, "learning_rate": 8.046211781322135e-05, "loss": 3.1294, "step": 165800 }, { "epoch": 1.4646143520726258, "grad_norm": 1.9203253984451294, "learning_rate": 8.045109718785586e-05, "loss": 3.5738, "step": 165850 }, { "epoch": 1.46505589996291, "grad_norm": 2.1311495304107666, "learning_rate": 8.044007421040132e-05, "loss": 3.3138, "step": 165900 }, { "epoch": 1.4654974478531941, "grad_norm": 3.7336440086364746, "learning_rate": 8.042904888170917e-05, "loss": 3.198, "step": 165950 }, { "epoch": 1.4659389957434783, "grad_norm": 3.2186074256896973, "learning_rate": 8.041802120263102e-05, "loss": 3.0895, "step": 166000 }, { "epoch": 1.4663805436337625, "grad_norm": 2.4467952251434326, "learning_rate": 8.040699117401868e-05, "loss": 3.1398, "step": 166050 }, { "epoch": 1.4668220915240466, "grad_norm": 1.0022852420806885, "learning_rate": 8.039595879672411e-05, "loss": 3.3921, "step": 166100 }, { "epoch": 1.4672636394143308, "grad_norm": 2.3139004707336426, "learning_rate": 8.038492407159947e-05, "loss": 3.372, "step": 166150 }, { "epoch": 1.467705187304615, "grad_norm": 3.8514606952667236, "learning_rate": 8.037388699949711e-05, "loss": 3.589, "step": 166200 }, { "epoch": 1.4681467351948991, "grad_norm": 2.21282958984375, "learning_rate": 8.036284758126953e-05, "loss": 3.3278, "step": 166250 }, { "epoch": 1.4685882830851833, "grad_norm": 0.6544881463050842, "learning_rate": 8.035180581776945e-05, "loss": 2.9854, "step": 166300 }, { "epoch": 1.4690298309754675, "grad_norm": 1.3662303686141968, "learning_rate": 8.034076170984974e-05, "loss": 2.9016, "step": 166350 }, { "epoch": 1.4694713788657519, "grad_norm": 2.929163932800293, "learning_rate": 8.032971525836348e-05, "loss": 4.1682, "step": 166400 }, { "epoch": 1.469912926756036, "grad_norm": 2.474565029144287, "learning_rate": 8.031866646416391e-05, "loss": 2.9599, "step": 166450 }, { "epoch": 1.4703544746463202, "grad_norm": 3.5583083629608154, "learning_rate": 8.030761532810444e-05, "loss": 3.309, "step": 166500 }, { "epoch": 1.4707960225366044, "grad_norm": 0.5456688404083252, "learning_rate": 8.029656185103868e-05, "loss": 3.472, "step": 166550 }, { "epoch": 1.4712375704268885, "grad_norm": 1.5131019353866577, "learning_rate": 8.028550603382045e-05, "loss": 3.2333, "step": 166600 }, { "epoch": 1.4716791183171727, "grad_norm": 2.514101028442383, "learning_rate": 8.027444787730367e-05, "loss": 2.9003, "step": 166650 }, { "epoch": 1.4721206662074569, "grad_norm": 2.0266590118408203, "learning_rate": 8.026338738234251e-05, "loss": 2.8004, "step": 166700 }, { "epoch": 1.472562214097741, "grad_norm": 2.531409740447998, "learning_rate": 8.025232454979133e-05, "loss": 3.714, "step": 166750 }, { "epoch": 1.4730037619880252, "grad_norm": 2.3518412113189697, "learning_rate": 8.024125938050458e-05, "loss": 3.3732, "step": 166800 }, { "epoch": 1.4734453098783094, "grad_norm": 5.866429328918457, "learning_rate": 8.023019187533699e-05, "loss": 3.3941, "step": 166850 }, { "epoch": 1.4738868577685935, "grad_norm": 1.075089454650879, "learning_rate": 8.021912203514341e-05, "loss": 2.9005, "step": 166900 }, { "epoch": 1.4743284056588777, "grad_norm": 1.00244140625, "learning_rate": 8.020804986077892e-05, "loss": 3.8404, "step": 166950 }, { "epoch": 1.474769953549162, "grad_norm": 3.055591106414795, "learning_rate": 8.019697535309873e-05, "loss": 3.1253, "step": 167000 }, { "epoch": 1.4752115014394462, "grad_norm": 3.31400203704834, "learning_rate": 8.018589851295826e-05, "loss": 3.6073, "step": 167050 }, { "epoch": 1.4756530493297304, "grad_norm": 2.1024818420410156, "learning_rate": 8.01748193412131e-05, "loss": 3.0527, "step": 167100 }, { "epoch": 1.4760945972200146, "grad_norm": 2.65166974067688, "learning_rate": 8.016373783871902e-05, "loss": 2.9383, "step": 167150 }, { "epoch": 1.4765361451102987, "grad_norm": 2.0548911094665527, "learning_rate": 8.015265400633197e-05, "loss": 2.886, "step": 167200 }, { "epoch": 1.476977693000583, "grad_norm": 2.2426414489746094, "learning_rate": 8.014156784490809e-05, "loss": 3.4212, "step": 167250 }, { "epoch": 1.477419240890867, "grad_norm": 1.2684077024459839, "learning_rate": 8.013047935530368e-05, "loss": 2.9928, "step": 167300 }, { "epoch": 1.4778607887811512, "grad_norm": 3.0890889167785645, "learning_rate": 8.011938853837524e-05, "loss": 3.1451, "step": 167350 }, { "epoch": 1.4783023366714354, "grad_norm": 6.254854202270508, "learning_rate": 8.010829539497943e-05, "loss": 2.9005, "step": 167400 }, { "epoch": 1.4787438845617196, "grad_norm": 2.656203508377075, "learning_rate": 8.009719992597313e-05, "loss": 3.1927, "step": 167450 }, { "epoch": 1.4791854324520037, "grad_norm": 2.7394678592681885, "learning_rate": 8.008610213221335e-05, "loss": 3.6834, "step": 167500 }, { "epoch": 1.479626980342288, "grad_norm": 4.921375274658203, "learning_rate": 8.007522403967907e-05, "loss": 3.184, "step": 167550 }, { "epoch": 1.480068528232572, "grad_norm": 5.552090167999268, "learning_rate": 8.00641216454365e-05, "loss": 3.1165, "step": 167600 }, { "epoch": 1.4805100761228562, "grad_norm": 1.2124018669128418, "learning_rate": 8.005301692899549e-05, "loss": 3.0754, "step": 167650 }, { "epoch": 1.4809516240131404, "grad_norm": 1.2512977123260498, "learning_rate": 8.004190989121377e-05, "loss": 2.6987, "step": 167700 }, { "epoch": 1.4813931719034246, "grad_norm": 4.063098907470703, "learning_rate": 8.003080053294925e-05, "loss": 3.5446, "step": 167750 }, { "epoch": 1.4818347197937087, "grad_norm": 3.4058032035827637, "learning_rate": 8.001968885506004e-05, "loss": 3.1717, "step": 167800 }, { "epoch": 1.482276267683993, "grad_norm": 1.3887379169464111, "learning_rate": 8.000857485840442e-05, "loss": 3.2566, "step": 167850 }, { "epoch": 1.482717815574277, "grad_norm": 3.54004168510437, "learning_rate": 7.999745854384086e-05, "loss": 2.7738, "step": 167900 }, { "epoch": 1.4831593634645612, "grad_norm": 0.8720212578773499, "learning_rate": 7.9986339912228e-05, "loss": 2.6591, "step": 167950 }, { "epoch": 1.4836009113548456, "grad_norm": 2.6884117126464844, "learning_rate": 7.997521896442467e-05, "loss": 2.9426, "step": 168000 }, { "epoch": 1.4836009113548456, "eval_asr_loss": 0.9104613695318818, "eval_loss": 2.9150123596191406, "eval_runtime": 20.6114, "eval_samples_per_second": 37.261, "eval_steps_per_second": 9.315, "eval_tts_loss": 5.9395940312615085, "step": 168000 }, { "epoch": 1.4840424592451298, "grad_norm": 1.3045971393585205, "learning_rate": 7.996409570128984e-05, "loss": 3.0543, "step": 168050 }, { "epoch": 1.484484007135414, "grad_norm": 1.5941531658172607, "learning_rate": 7.995297012368273e-05, "loss": 2.7356, "step": 168100 }, { "epoch": 1.4849255550256981, "grad_norm": 6.986129283905029, "learning_rate": 7.994184223246263e-05, "loss": 3.7545, "step": 168150 }, { "epoch": 1.4853671029159823, "grad_norm": 1.7898863554000854, "learning_rate": 7.993071202848916e-05, "loss": 3.2304, "step": 168200 }, { "epoch": 1.4858086508062665, "grad_norm": 1.5254120826721191, "learning_rate": 7.991957951262196e-05, "loss": 2.9966, "step": 168250 }, { "epoch": 1.4862501986965506, "grad_norm": 5.384498596191406, "learning_rate": 7.990844468572098e-05, "loss": 3.0987, "step": 168300 }, { "epoch": 1.4866917465868348, "grad_norm": 2.3323962688446045, "learning_rate": 7.989730754864624e-05, "loss": 2.9712, "step": 168350 }, { "epoch": 1.487133294477119, "grad_norm": 3.0033323764801025, "learning_rate": 7.988616810225804e-05, "loss": 3.1348, "step": 168400 }, { "epoch": 1.4875748423674031, "grad_norm": 3.1256043910980225, "learning_rate": 7.987502634741676e-05, "loss": 2.8878, "step": 168450 }, { "epoch": 1.4880163902576873, "grad_norm": 1.6998767852783203, "learning_rate": 7.986388228498303e-05, "loss": 3.1979, "step": 168500 }, { "epoch": 1.4884579381479714, "grad_norm": 1.2934191226959229, "learning_rate": 7.985273591581763e-05, "loss": 3.2321, "step": 168550 }, { "epoch": 1.4888994860382558, "grad_norm": 1.3229929208755493, "learning_rate": 7.984158724078154e-05, "loss": 3.4828, "step": 168600 }, { "epoch": 1.48934103392854, "grad_norm": 1.3975541591644287, "learning_rate": 7.983043626073587e-05, "loss": 3.2538, "step": 168650 }, { "epoch": 1.4897825818188242, "grad_norm": 0.7398980259895325, "learning_rate": 7.981928297654197e-05, "loss": 3.1933, "step": 168700 }, { "epoch": 1.4902241297091083, "grad_norm": 2.5116653442382812, "learning_rate": 7.980812738906129e-05, "loss": 3.2666, "step": 168750 }, { "epoch": 1.4906656775993925, "grad_norm": 4.653420925140381, "learning_rate": 7.979696949915556e-05, "loss": 2.9902, "step": 168800 }, { "epoch": 1.4911072254896767, "grad_norm": 2.0215892791748047, "learning_rate": 7.97858093076866e-05, "loss": 3.4902, "step": 168850 }, { "epoch": 1.4915487733799608, "grad_norm": 2.4749724864959717, "learning_rate": 7.977464681551646e-05, "loss": 3.2653, "step": 168900 }, { "epoch": 1.491990321270245, "grad_norm": 0.8826634883880615, "learning_rate": 7.97634820235073e-05, "loss": 2.8773, "step": 168950 }, { "epoch": 1.4924318691605292, "grad_norm": 2.755411386489868, "learning_rate": 7.975231493252157e-05, "loss": 3.1709, "step": 169000 }, { "epoch": 1.4928734170508133, "grad_norm": 1.458707332611084, "learning_rate": 7.97411455434218e-05, "loss": 3.195, "step": 169050 }, { "epoch": 1.4933149649410975, "grad_norm": 2.2169852256774902, "learning_rate": 7.972997385707073e-05, "loss": 3.069, "step": 169100 }, { "epoch": 1.4937565128313817, "grad_norm": 4.073314189910889, "learning_rate": 7.971879987433128e-05, "loss": 3.0165, "step": 169150 }, { "epoch": 1.4941980607216658, "grad_norm": 1.273754596710205, "learning_rate": 7.970762359606656e-05, "loss": 3.1862, "step": 169200 }, { "epoch": 1.49463960861195, "grad_norm": 2.835773468017578, "learning_rate": 7.96964450231398e-05, "loss": 3.2468, "step": 169250 }, { "epoch": 1.4950811565022342, "grad_norm": 3.9089925289154053, "learning_rate": 7.968526415641452e-05, "loss": 2.8939, "step": 169300 }, { "epoch": 1.4955227043925183, "grad_norm": 3.043555498123169, "learning_rate": 7.967408099675429e-05, "loss": 3.6233, "step": 169350 }, { "epoch": 1.4959642522828025, "grad_norm": 1.930454969406128, "learning_rate": 7.966289554502293e-05, "loss": 2.4929, "step": 169400 }, { "epoch": 1.4964058001730867, "grad_norm": 5.453582286834717, "learning_rate": 7.965170780208441e-05, "loss": 3.0559, "step": 169450 }, { "epoch": 1.4968473480633708, "grad_norm": 1.6898558139801025, "learning_rate": 7.964051776880292e-05, "loss": 3.0782, "step": 169500 }, { "epoch": 1.497288895953655, "grad_norm": 1.4121713638305664, "learning_rate": 7.962932544604277e-05, "loss": 3.6197, "step": 169550 }, { "epoch": 1.4977304438439394, "grad_norm": 6.987101078033447, "learning_rate": 7.96181308346685e-05, "loss": 3.3473, "step": 169600 }, { "epoch": 1.4981719917342236, "grad_norm": 3.6944234371185303, "learning_rate": 7.960693393554474e-05, "loss": 3.2968, "step": 169650 }, { "epoch": 1.4986135396245077, "grad_norm": 1.7384791374206543, "learning_rate": 7.959573474953643e-05, "loss": 3.2508, "step": 169700 }, { "epoch": 1.4990550875147919, "grad_norm": 3.552605152130127, "learning_rate": 7.958453327750854e-05, "loss": 3.7686, "step": 169750 }, { "epoch": 1.499496635405076, "grad_norm": 1.9855374097824097, "learning_rate": 7.957332952032634e-05, "loss": 3.5705, "step": 169800 }, { "epoch": 1.4999381832953602, "grad_norm": 2.5507972240448, "learning_rate": 7.956212347885521e-05, "loss": 3.1468, "step": 169850 }, { "epoch": 1.5003797311856444, "grad_norm": 2.040591239929199, "learning_rate": 7.955091515396073e-05, "loss": 3.2078, "step": 169900 }, { "epoch": 1.5008212790759285, "grad_norm": 2.0551323890686035, "learning_rate": 7.953970454650863e-05, "loss": 3.1355, "step": 169950 }, { "epoch": 1.5012628269662127, "grad_norm": 1.476214051246643, "learning_rate": 7.952849165736483e-05, "loss": 3.1195, "step": 170000 }, { "epoch": 1.5017043748564969, "grad_norm": 4.278891086578369, "learning_rate": 7.951727648739547e-05, "loss": 2.9388, "step": 170050 }, { "epoch": 1.5021459227467813, "grad_norm": 3.7804348468780518, "learning_rate": 7.950605903746679e-05, "loss": 3.2333, "step": 170100 }, { "epoch": 1.5025874706370654, "grad_norm": 2.9571738243103027, "learning_rate": 7.949483930844524e-05, "loss": 3.5166, "step": 170150 }, { "epoch": 1.5030290185273496, "grad_norm": 1.926645040512085, "learning_rate": 7.948361730119747e-05, "loss": 2.9482, "step": 170200 }, { "epoch": 1.5034705664176338, "grad_norm": 2.891768217086792, "learning_rate": 7.947261752459494e-05, "loss": 3.2691, "step": 170250 }, { "epoch": 1.503912114307918, "grad_norm": 2.981902599334717, "learning_rate": 7.946139100901666e-05, "loss": 3.146, "step": 170300 }, { "epoch": 1.504353662198202, "grad_norm": 5.165239334106445, "learning_rate": 7.945016221779575e-05, "loss": 3.2342, "step": 170350 }, { "epoch": 1.5047952100884863, "grad_norm": 2.168894052505493, "learning_rate": 7.943893115179954e-05, "loss": 3.3417, "step": 170400 }, { "epoch": 1.5052367579787704, "grad_norm": 1.8027621507644653, "learning_rate": 7.942769781189552e-05, "loss": 3.2269, "step": 170450 }, { "epoch": 1.5056783058690546, "grad_norm": 2.0628678798675537, "learning_rate": 7.941646219895141e-05, "loss": 3.0221, "step": 170500 }, { "epoch": 1.5061198537593388, "grad_norm": 2.1540791988372803, "learning_rate": 7.940522431383502e-05, "loss": 3.2091, "step": 170550 }, { "epoch": 1.506561401649623, "grad_norm": 2.1455180644989014, "learning_rate": 7.939398415741441e-05, "loss": 3.0678, "step": 170600 }, { "epoch": 1.507002949539907, "grad_norm": 1.412413477897644, "learning_rate": 7.938274173055777e-05, "loss": 3.4302, "step": 170650 }, { "epoch": 1.5074444974301913, "grad_norm": 1.9970897436141968, "learning_rate": 7.937149703413349e-05, "loss": 3.1186, "step": 170700 }, { "epoch": 1.5078860453204754, "grad_norm": 4.280227184295654, "learning_rate": 7.936025006901014e-05, "loss": 3.2719, "step": 170750 }, { "epoch": 1.5083275932107596, "grad_norm": 3.5186574459075928, "learning_rate": 7.93490008360564e-05, "loss": 3.2983, "step": 170800 }, { "epoch": 1.5087691411010438, "grad_norm": 1.2954347133636475, "learning_rate": 7.933774933614126e-05, "loss": 3.1792, "step": 170850 }, { "epoch": 1.509210688991328, "grad_norm": 2.387403726577759, "learning_rate": 7.932672066765598e-05, "loss": 2.9762, "step": 170900 }, { "epoch": 1.509652236881612, "grad_norm": 5.01585578918457, "learning_rate": 7.93154646817213e-05, "loss": 3.2197, "step": 170950 }, { "epoch": 1.5100937847718963, "grad_norm": 3.7455615997314453, "learning_rate": 7.930420643141558e-05, "loss": 3.4165, "step": 171000 }, { "epoch": 1.5100937847718963, "eval_asr_loss": 0.9273673528471224, "eval_loss": 2.9135005474090576, "eval_runtime": 20.4893, "eval_samples_per_second": 37.483, "eval_steps_per_second": 9.371, "eval_tts_loss": 5.9611810830116205, "step": 171000 }, { "epoch": 1.5105353326621804, "grad_norm": 1.949747920036316, "learning_rate": 7.92929459176084e-05, "loss": 3.3078, "step": 171050 }, { "epoch": 1.5109768805524646, "grad_norm": 3.1064343452453613, "learning_rate": 7.928168314116953e-05, "loss": 3.5604, "step": 171100 }, { "epoch": 1.5114184284427488, "grad_norm": 1.3786486387252808, "learning_rate": 7.927041810296896e-05, "loss": 3.3521, "step": 171150 }, { "epoch": 1.511859976333033, "grad_norm": 3.0906989574432373, "learning_rate": 7.925915080387679e-05, "loss": 3.719, "step": 171200 }, { "epoch": 1.512301524223317, "grad_norm": 1.4307838678359985, "learning_rate": 7.924788124476331e-05, "loss": 3.4073, "step": 171250 }, { "epoch": 1.5127430721136015, "grad_norm": 1.856938362121582, "learning_rate": 7.923660942649904e-05, "loss": 3.062, "step": 171300 }, { "epoch": 1.5131846200038856, "grad_norm": 2.1577956676483154, "learning_rate": 7.92253353499546e-05, "loss": 3.0501, "step": 171350 }, { "epoch": 1.5136261678941698, "grad_norm": 1.228046178817749, "learning_rate": 7.921405901600084e-05, "loss": 3.1004, "step": 171400 }, { "epoch": 1.514067715784454, "grad_norm": 2.502464771270752, "learning_rate": 7.920278042550875e-05, "loss": 2.6685, "step": 171450 }, { "epoch": 1.5145092636747381, "grad_norm": 3.5946199893951416, "learning_rate": 7.919149957934948e-05, "loss": 3.0592, "step": 171500 }, { "epoch": 1.5149508115650223, "grad_norm": 2.893009901046753, "learning_rate": 7.918021647839443e-05, "loss": 3.1573, "step": 171550 }, { "epoch": 1.5153923594553065, "grad_norm": 1.8947813510894775, "learning_rate": 7.91689311235151e-05, "loss": 3.3152, "step": 171600 }, { "epoch": 1.5158339073455906, "grad_norm": 3.0196502208709717, "learning_rate": 7.915764351558317e-05, "loss": 3.2163, "step": 171650 }, { "epoch": 1.516275455235875, "grad_norm": 3.1389248371124268, "learning_rate": 7.914635365547053e-05, "loss": 3.2143, "step": 171700 }, { "epoch": 1.5167170031261592, "grad_norm": 3.7226805686950684, "learning_rate": 7.913506154404924e-05, "loss": 3.1146, "step": 171750 }, { "epoch": 1.5171585510164434, "grad_norm": 1.4495065212249756, "learning_rate": 7.912376718219148e-05, "loss": 3.4301, "step": 171800 }, { "epoch": 1.5176000989067275, "grad_norm": 4.089288711547852, "learning_rate": 7.911247057076969e-05, "loss": 2.8631, "step": 171850 }, { "epoch": 1.5180416467970117, "grad_norm": 1.8996052742004395, "learning_rate": 7.91011717106564e-05, "loss": 3.3218, "step": 171900 }, { "epoch": 1.5184831946872959, "grad_norm": 4.082929611206055, "learning_rate": 7.908987060272437e-05, "loss": 3.0405, "step": 171950 }, { "epoch": 1.51892474257758, "grad_norm": 7.022767543792725, "learning_rate": 7.90785672478465e-05, "loss": 3.5964, "step": 172000 }, { "epoch": 1.5193662904678642, "grad_norm": 6.730228900909424, "learning_rate": 7.90672616468959e-05, "loss": 3.3314, "step": 172050 }, { "epoch": 1.5198078383581484, "grad_norm": 2.4673285484313965, "learning_rate": 7.905595380074582e-05, "loss": 3.2224, "step": 172100 }, { "epoch": 1.5202493862484325, "grad_norm": 2.011349678039551, "learning_rate": 7.904464371026971e-05, "loss": 3.4183, "step": 172150 }, { "epoch": 1.5206909341387167, "grad_norm": 1.7207785844802856, "learning_rate": 7.903333137634115e-05, "loss": 3.1136, "step": 172200 }, { "epoch": 1.5211324820290009, "grad_norm": 4.2923150062561035, "learning_rate": 7.902201679983393e-05, "loss": 3.1425, "step": 172250 }, { "epoch": 1.521574029919285, "grad_norm": 3.7688066959381104, "learning_rate": 7.901069998162203e-05, "loss": 3.4813, "step": 172300 }, { "epoch": 1.5220155778095692, "grad_norm": 1.0002552270889282, "learning_rate": 7.899938092257955e-05, "loss": 3.5722, "step": 172350 }, { "epoch": 1.5224571256998534, "grad_norm": 1.6142280101776123, "learning_rate": 7.89880596235808e-05, "loss": 2.8011, "step": 172400 }, { "epoch": 1.5228986735901375, "grad_norm": 5.061013698577881, "learning_rate": 7.897673608550026e-05, "loss": 2.8856, "step": 172450 }, { "epoch": 1.5233402214804217, "grad_norm": 4.347459316253662, "learning_rate": 7.896541030921256e-05, "loss": 3.1504, "step": 172500 }, { "epoch": 1.5237817693707059, "grad_norm": 1.6337541341781616, "learning_rate": 7.895408229559256e-05, "loss": 3.1389, "step": 172550 }, { "epoch": 1.52422331726099, "grad_norm": 3.7045750617980957, "learning_rate": 7.89427520455152e-05, "loss": 3.4723, "step": 172600 }, { "epoch": 1.5246648651512742, "grad_norm": 2.3034849166870117, "learning_rate": 7.893141955985568e-05, "loss": 3.1481, "step": 172650 }, { "epoch": 1.5251064130415584, "grad_norm": 1.3366230726242065, "learning_rate": 7.892008483948933e-05, "loss": 2.9249, "step": 172700 }, { "epoch": 1.5255479609318425, "grad_norm": 1.471297025680542, "learning_rate": 7.890874788529166e-05, "loss": 3.4342, "step": 172750 }, { "epoch": 1.5259895088221267, "grad_norm": 4.741721153259277, "learning_rate": 7.889740869813835e-05, "loss": 3.3231, "step": 172800 }, { "epoch": 1.526431056712411, "grad_norm": 1.3582066297531128, "learning_rate": 7.888606727890528e-05, "loss": 3.2317, "step": 172850 }, { "epoch": 1.5268726046026952, "grad_norm": 2.5067501068115234, "learning_rate": 7.887472362846844e-05, "loss": 3.1801, "step": 172900 }, { "epoch": 1.5273141524929794, "grad_norm": 2.211074113845825, "learning_rate": 7.886337774770407e-05, "loss": 3.4487, "step": 172950 }, { "epoch": 1.5277557003832636, "grad_norm": 2.118274450302124, "learning_rate": 7.88520296374885e-05, "loss": 3.4356, "step": 173000 }, { "epoch": 1.5281972482735477, "grad_norm": 2.6437795162200928, "learning_rate": 7.884067929869832e-05, "loss": 3.2181, "step": 173050 }, { "epoch": 1.528638796163832, "grad_norm": 7.714078426361084, "learning_rate": 7.882932673221022e-05, "loss": 3.0523, "step": 173100 }, { "epoch": 1.529080344054116, "grad_norm": 3.1765506267547607, "learning_rate": 7.881797193890108e-05, "loss": 2.9752, "step": 173150 }, { "epoch": 1.5295218919444002, "grad_norm": 1.7637883424758911, "learning_rate": 7.880661491964799e-05, "loss": 2.7156, "step": 173200 }, { "epoch": 1.5299634398346846, "grad_norm": 2.4958198070526123, "learning_rate": 7.879525567532818e-05, "loss": 3.2052, "step": 173250 }, { "epoch": 1.5304049877249688, "grad_norm": 1.5706040859222412, "learning_rate": 7.878389420681902e-05, "loss": 2.9912, "step": 173300 }, { "epoch": 1.530846535615253, "grad_norm": 5.228756904602051, "learning_rate": 7.877253051499813e-05, "loss": 3.2192, "step": 173350 }, { "epoch": 1.5312880835055371, "grad_norm": 4.856530666351318, "learning_rate": 7.876116460074325e-05, "loss": 2.5722, "step": 173400 }, { "epoch": 1.5317296313958213, "grad_norm": 1.8919997215270996, "learning_rate": 7.874979646493228e-05, "loss": 3.1431, "step": 173450 }, { "epoch": 1.5321711792861055, "grad_norm": 3.6233277320861816, "learning_rate": 7.873842610844332e-05, "loss": 3.6041, "step": 173500 }, { "epoch": 1.5326127271763896, "grad_norm": 0.6207433938980103, "learning_rate": 7.872705353215464e-05, "loss": 2.8251, "step": 173550 }, { "epoch": 1.5330542750666738, "grad_norm": 3.8003902435302734, "learning_rate": 7.871567873694468e-05, "loss": 3.3234, "step": 173600 }, { "epoch": 1.533495822956958, "grad_norm": 2.3610243797302246, "learning_rate": 7.870430172369204e-05, "loss": 2.9732, "step": 173650 }, { "epoch": 1.5339373708472421, "grad_norm": 3.0918922424316406, "learning_rate": 7.869292249327549e-05, "loss": 3.3379, "step": 173700 }, { "epoch": 1.5343789187375263, "grad_norm": 3.1839041709899902, "learning_rate": 7.868154104657399e-05, "loss": 3.1345, "step": 173750 }, { "epoch": 1.5348204666278105, "grad_norm": 2.7104201316833496, "learning_rate": 7.867015738446666e-05, "loss": 3.0175, "step": 173800 }, { "epoch": 1.5352620145180946, "grad_norm": 6.00855016708374, "learning_rate": 7.865877150783279e-05, "loss": 3.308, "step": 173850 }, { "epoch": 1.5357035624083788, "grad_norm": 1.8182517290115356, "learning_rate": 7.864738341755183e-05, "loss": 3.6535, "step": 173900 }, { "epoch": 1.536145110298663, "grad_norm": 1.9349044561386108, "learning_rate": 7.863599311450343e-05, "loss": 3.2833, "step": 173950 }, { "epoch": 1.5365866581889471, "grad_norm": 3.1747500896453857, "learning_rate": 7.862460059956738e-05, "loss": 3.1619, "step": 174000 }, { "epoch": 1.5365866581889471, "eval_asr_loss": 0.9183246925579431, "eval_loss": 2.908951997756958, "eval_runtime": 29.1305, "eval_samples_per_second": 26.364, "eval_steps_per_second": 6.591, "eval_tts_loss": 5.985057644157511, "step": 174000 }, { "epoch": 1.5370282060792313, "grad_norm": 3.5058462619781494, "learning_rate": 7.861320587362366e-05, "loss": 2.9589, "step": 174050 }, { "epoch": 1.5374697539695155, "grad_norm": 1.7006515264511108, "learning_rate": 7.860180893755243e-05, "loss": 3.1892, "step": 174100 }, { "epoch": 1.5379113018597996, "grad_norm": 6.1886444091796875, "learning_rate": 7.8590409792234e-05, "loss": 2.8409, "step": 174150 }, { "epoch": 1.5383528497500838, "grad_norm": 5.43979549407959, "learning_rate": 7.857900843854884e-05, "loss": 3.1715, "step": 174200 }, { "epoch": 1.538794397640368, "grad_norm": 1.8094000816345215, "learning_rate": 7.856760487737763e-05, "loss": 3.4528, "step": 174250 }, { "epoch": 1.5392359455306521, "grad_norm": 3.401740312576294, "learning_rate": 7.855619910960119e-05, "loss": 3.4312, "step": 174300 }, { "epoch": 1.5396774934209363, "grad_norm": 3.1682894229888916, "learning_rate": 7.854479113610052e-05, "loss": 3.1221, "step": 174350 }, { "epoch": 1.5401190413112205, "grad_norm": 5.100546360015869, "learning_rate": 7.853338095775677e-05, "loss": 3.1851, "step": 174400 }, { "epoch": 1.5405605892015048, "grad_norm": 0.7844409346580505, "learning_rate": 7.85219685754513e-05, "loss": 3.2288, "step": 174450 }, { "epoch": 1.541002137091789, "grad_norm": 2.428945302963257, "learning_rate": 7.851055399006565e-05, "loss": 2.9583, "step": 174500 }, { "epoch": 1.5414436849820732, "grad_norm": 3.063448905944824, "learning_rate": 7.849913720248143e-05, "loss": 2.7642, "step": 174550 }, { "epoch": 1.5418852328723573, "grad_norm": 3.730762243270874, "learning_rate": 7.848771821358055e-05, "loss": 3.4903, "step": 174600 }, { "epoch": 1.5423267807626415, "grad_norm": 1.0115082263946533, "learning_rate": 7.847629702424499e-05, "loss": 3.4874, "step": 174650 }, { "epoch": 1.5427683286529257, "grad_norm": 4.5162835121154785, "learning_rate": 7.846487363535697e-05, "loss": 3.3289, "step": 174700 }, { "epoch": 1.5432098765432098, "grad_norm": 2.281966209411621, "learning_rate": 7.845344804779884e-05, "loss": 3.6502, "step": 174750 }, { "epoch": 1.543651424433494, "grad_norm": 3.5982887744903564, "learning_rate": 7.844202026245311e-05, "loss": 3.0063, "step": 174800 }, { "epoch": 1.5440929723237784, "grad_norm": 4.303706645965576, "learning_rate": 7.84305902802025e-05, "loss": 3.3499, "step": 174850 }, { "epoch": 1.5445345202140626, "grad_norm": 0.5809865593910217, "learning_rate": 7.841915810192987e-05, "loss": 3.3385, "step": 174900 }, { "epoch": 1.5449760681043467, "grad_norm": 1.4233925342559814, "learning_rate": 7.840772372851827e-05, "loss": 3.3227, "step": 174950 }, { "epoch": 1.545417615994631, "grad_norm": 1.0716662406921387, "learning_rate": 7.839628716085092e-05, "loss": 2.7928, "step": 175000 }, { "epoch": 1.545859163884915, "grad_norm": 1.625321388244629, "learning_rate": 7.838484839981117e-05, "loss": 2.9877, "step": 175050 }, { "epoch": 1.5463007117751992, "grad_norm": 1.8648103475570679, "learning_rate": 7.837340744628257e-05, "loss": 3.0413, "step": 175100 }, { "epoch": 1.5467422596654834, "grad_norm": 4.233791828155518, "learning_rate": 7.836196430114884e-05, "loss": 3.1447, "step": 175150 }, { "epoch": 1.5471838075557676, "grad_norm": 2.086745023727417, "learning_rate": 7.835051896529388e-05, "loss": 3.4688, "step": 175200 }, { "epoch": 1.5476253554460517, "grad_norm": 1.6742719411849976, "learning_rate": 7.833930041157026e-05, "loss": 3.1915, "step": 175250 }, { "epoch": 1.548066903336336, "grad_norm": 2.0787861347198486, "learning_rate": 7.832785074069555e-05, "loss": 3.1022, "step": 175300 }, { "epoch": 1.54850845122662, "grad_norm": 3.8537216186523438, "learning_rate": 7.831639888173457e-05, "loss": 2.9691, "step": 175350 }, { "epoch": 1.5489499991169042, "grad_norm": 1.3162223100662231, "learning_rate": 7.830494483557191e-05, "loss": 3.1025, "step": 175400 }, { "epoch": 1.5493915470071884, "grad_norm": 3.5656752586364746, "learning_rate": 7.829348860309228e-05, "loss": 2.9325, "step": 175450 }, { "epoch": 1.5498330948974726, "grad_norm": 1.3675990104675293, "learning_rate": 7.828203018518056e-05, "loss": 2.8666, "step": 175500 }, { "epoch": 1.5502746427877567, "grad_norm": 7.35089111328125, "learning_rate": 7.827056958272183e-05, "loss": 3.3025, "step": 175550 }, { "epoch": 1.550716190678041, "grad_norm": 1.7935676574707031, "learning_rate": 7.825910679660134e-05, "loss": 3.3327, "step": 175600 }, { "epoch": 1.551157738568325, "grad_norm": 3.305955171585083, "learning_rate": 7.824764182770448e-05, "loss": 3.1936, "step": 175650 }, { "epoch": 1.5515992864586092, "grad_norm": 2.892768621444702, "learning_rate": 7.823617467691681e-05, "loss": 3.1448, "step": 175700 }, { "epoch": 1.5520408343488934, "grad_norm": 2.3808202743530273, "learning_rate": 7.82247053451241e-05, "loss": 3.2515, "step": 175750 }, { "epoch": 1.5524823822391776, "grad_norm": 3.7346816062927246, "learning_rate": 7.821323383321224e-05, "loss": 3.2934, "step": 175800 }, { "epoch": 1.5529239301294617, "grad_norm": 1.123934030532837, "learning_rate": 7.82017601420673e-05, "loss": 3.5983, "step": 175850 }, { "epoch": 1.553365478019746, "grad_norm": 2.3236446380615234, "learning_rate": 7.819028427257553e-05, "loss": 3.1701, "step": 175900 }, { "epoch": 1.55380702591003, "grad_norm": 1.2918416261672974, "learning_rate": 7.817880622562334e-05, "loss": 3.1689, "step": 175950 }, { "epoch": 1.5542485738003142, "grad_norm": 3.1949520111083984, "learning_rate": 7.816732600209735e-05, "loss": 3.2805, "step": 176000 }, { "epoch": 1.5546901216905986, "grad_norm": 1.407420039176941, "learning_rate": 7.815584360288424e-05, "loss": 3.6234, "step": 176050 }, { "epoch": 1.5551316695808828, "grad_norm": 2.7942097187042236, "learning_rate": 7.814435902887098e-05, "loss": 2.9961, "step": 176100 }, { "epoch": 1.555573217471167, "grad_norm": 5.137963771820068, "learning_rate": 7.813287228094466e-05, "loss": 2.9095, "step": 176150 }, { "epoch": 1.5560147653614511, "grad_norm": 1.7814692258834839, "learning_rate": 7.81213833599925e-05, "loss": 3.0931, "step": 176200 }, { "epoch": 1.5564563132517353, "grad_norm": 0.8939611911773682, "learning_rate": 7.810989226690193e-05, "loss": 3.306, "step": 176250 }, { "epoch": 1.5568978611420194, "grad_norm": 0.7645553946495056, "learning_rate": 7.809839900256056e-05, "loss": 3.0424, "step": 176300 }, { "epoch": 1.5573394090323036, "grad_norm": 2.913379192352295, "learning_rate": 7.808690356785613e-05, "loss": 3.2549, "step": 176350 }, { "epoch": 1.5577809569225878, "grad_norm": 2.6756792068481445, "learning_rate": 7.807540596367658e-05, "loss": 3.6026, "step": 176400 }, { "epoch": 1.5582225048128722, "grad_norm": 1.4105322360992432, "learning_rate": 7.806390619090998e-05, "loss": 3.085, "step": 176450 }, { "epoch": 1.5586640527031563, "grad_norm": 5.497186660766602, "learning_rate": 7.80524042504446e-05, "loss": 2.7705, "step": 176500 }, { "epoch": 1.5591056005934405, "grad_norm": 6.118846416473389, "learning_rate": 7.804090014316887e-05, "loss": 3.3862, "step": 176550 }, { "epoch": 1.5595471484837247, "grad_norm": 1.7508882284164429, "learning_rate": 7.80293938699714e-05, "loss": 2.9305, "step": 176600 }, { "epoch": 1.5599886963740088, "grad_norm": 3.990896701812744, "learning_rate": 7.801788543174092e-05, "loss": 3.2443, "step": 176650 }, { "epoch": 1.560430244264293, "grad_norm": 4.021361351013184, "learning_rate": 7.800637482936637e-05, "loss": 2.9178, "step": 176700 }, { "epoch": 1.5608717921545772, "grad_norm": 1.7801082134246826, "learning_rate": 7.799486206373685e-05, "loss": 3.4446, "step": 176750 }, { "epoch": 1.5613133400448613, "grad_norm": 2.1543657779693604, "learning_rate": 7.798334713574165e-05, "loss": 3.2613, "step": 176800 }, { "epoch": 1.5617548879351455, "grad_norm": 2.6019632816314697, "learning_rate": 7.797183004627015e-05, "loss": 3.6306, "step": 176850 }, { "epoch": 1.5621964358254297, "grad_norm": 5.682262420654297, "learning_rate": 7.796031079621197e-05, "loss": 3.2412, "step": 176900 }, { "epoch": 1.5626379837157138, "grad_norm": 2.869284152984619, "learning_rate": 7.794878938645691e-05, "loss": 3.2968, "step": 176950 }, { "epoch": 1.563079531605998, "grad_norm": 2.434187173843384, "learning_rate": 7.793726581789485e-05, "loss": 3.1968, "step": 177000 }, { "epoch": 1.563079531605998, "eval_asr_loss": 0.9133994808530763, "eval_loss": 2.902208089828491, "eval_runtime": 21.1294, "eval_samples_per_second": 36.347, "eval_steps_per_second": 9.087, "eval_tts_loss": 5.981479948869558, "step": 177000 }, { "epoch": 1.5635210794962822, "grad_norm": 3.677980661392212, "learning_rate": 7.792574009141591e-05, "loss": 3.076, "step": 177050 }, { "epoch": 1.5639626273865663, "grad_norm": 1.4063122272491455, "learning_rate": 7.791421220791034e-05, "loss": 2.9536, "step": 177100 }, { "epoch": 1.5644041752768505, "grad_norm": 2.6018192768096924, "learning_rate": 7.79026821682686e-05, "loss": 3.0829, "step": 177150 }, { "epoch": 1.5648457231671347, "grad_norm": 2.5115301609039307, "learning_rate": 7.789114997338125e-05, "loss": 3.1509, "step": 177200 }, { "epoch": 1.5652872710574188, "grad_norm": 1.8205089569091797, "learning_rate": 7.787984633223084e-05, "loss": 2.9936, "step": 177250 }, { "epoch": 1.565728818947703, "grad_norm": 3.0705020427703857, "learning_rate": 7.786830987258534e-05, "loss": 3.2129, "step": 177300 }, { "epoch": 1.5661703668379872, "grad_norm": 1.5330673456192017, "learning_rate": 7.785677126034919e-05, "loss": 2.8801, "step": 177350 }, { "epoch": 1.5666119147282713, "grad_norm": 3.940424680709839, "learning_rate": 7.784523049641369e-05, "loss": 3.1278, "step": 177400 }, { "epoch": 1.5670534626185555, "grad_norm": 2.2547924518585205, "learning_rate": 7.783368758167027e-05, "loss": 2.8893, "step": 177450 }, { "epoch": 1.5674950105088397, "grad_norm": 2.2895209789276123, "learning_rate": 7.78221425170105e-05, "loss": 3.0462, "step": 177500 }, { "epoch": 1.5679365583991238, "grad_norm": 1.0775874853134155, "learning_rate": 7.781059530332613e-05, "loss": 3.2746, "step": 177550 }, { "epoch": 1.568378106289408, "grad_norm": 3.228985071182251, "learning_rate": 7.77990459415091e-05, "loss": 3.1635, "step": 177600 }, { "epoch": 1.5688196541796924, "grad_norm": 2.3288896083831787, "learning_rate": 7.778749443245153e-05, "loss": 2.9833, "step": 177650 }, { "epoch": 1.5692612020699765, "grad_norm": 1.9562253952026367, "learning_rate": 7.777594077704561e-05, "loss": 2.9966, "step": 177700 }, { "epoch": 1.5697027499602607, "grad_norm": 3.997286558151245, "learning_rate": 7.776438497618382e-05, "loss": 3.0108, "step": 177750 }, { "epoch": 1.5701442978505449, "grad_norm": 4.3550801277160645, "learning_rate": 7.775282703075872e-05, "loss": 3.4799, "step": 177800 }, { "epoch": 1.570585845740829, "grad_norm": 2.5383787155151367, "learning_rate": 7.774126694166307e-05, "loss": 3.0563, "step": 177850 }, { "epoch": 1.5710273936311132, "grad_norm": 2.1461341381073, "learning_rate": 7.772970470978978e-05, "loss": 3.5388, "step": 177900 }, { "epoch": 1.5714689415213974, "grad_norm": 1.3384110927581787, "learning_rate": 7.771814033603196e-05, "loss": 3.2212, "step": 177950 }, { "epoch": 1.5719104894116815, "grad_norm": 2.8407857418060303, "learning_rate": 7.770657382128284e-05, "loss": 3.0422, "step": 178000 }, { "epoch": 1.572352037301966, "grad_norm": 2.219881772994995, "learning_rate": 7.769500516643582e-05, "loss": 3.061, "step": 178050 }, { "epoch": 1.57279358519225, "grad_norm": 2.5825889110565186, "learning_rate": 7.76834343723845e-05, "loss": 3.6231, "step": 178100 }, { "epoch": 1.5732351330825343, "grad_norm": 2.435443162918091, "learning_rate": 7.767186144002264e-05, "loss": 3.1914, "step": 178150 }, { "epoch": 1.5736766809728184, "grad_norm": 3.284209728240967, "learning_rate": 7.766028637024411e-05, "loss": 2.9997, "step": 178200 }, { "epoch": 1.5741182288631026, "grad_norm": 3.507371187210083, "learning_rate": 7.764870916394302e-05, "loss": 3.117, "step": 178250 }, { "epoch": 1.5745597767533868, "grad_norm": 4.541454315185547, "learning_rate": 7.763712982201362e-05, "loss": 3.203, "step": 178300 }, { "epoch": 1.575001324643671, "grad_norm": 1.8233087062835693, "learning_rate": 7.762554834535028e-05, "loss": 3.2556, "step": 178350 }, { "epoch": 1.575442872533955, "grad_norm": 2.1894471645355225, "learning_rate": 7.76139647348476e-05, "loss": 3.4755, "step": 178400 }, { "epoch": 1.5758844204242393, "grad_norm": 2.089081287384033, "learning_rate": 7.760237899140028e-05, "loss": 3.3049, "step": 178450 }, { "epoch": 1.5763259683145234, "grad_norm": 3.1048781871795654, "learning_rate": 7.759079111590326e-05, "loss": 3.1276, "step": 178500 }, { "epoch": 1.5767675162048076, "grad_norm": 5.147495269775391, "learning_rate": 7.757920110925159e-05, "loss": 3.19, "step": 178550 }, { "epoch": 1.5772090640950918, "grad_norm": 5.037656784057617, "learning_rate": 7.756784083594946e-05, "loss": 3.3464, "step": 178600 }, { "epoch": 1.577650611985376, "grad_norm": 1.2267601490020752, "learning_rate": 7.755624661225285e-05, "loss": 3.5405, "step": 178650 }, { "epoch": 1.57809215987566, "grad_norm": 0.7518936991691589, "learning_rate": 7.754465026006984e-05, "loss": 3.4172, "step": 178700 }, { "epoch": 1.5785337077659443, "grad_norm": 2.8501391410827637, "learning_rate": 7.75330517802962e-05, "loss": 3.3903, "step": 178750 }, { "epoch": 1.5789752556562284, "grad_norm": 1.8526347875595093, "learning_rate": 7.752145117382777e-05, "loss": 3.047, "step": 178800 }, { "epoch": 1.5794168035465126, "grad_norm": 1.4192283153533936, "learning_rate": 7.750984844156061e-05, "loss": 3.1864, "step": 178850 }, { "epoch": 1.5798583514367968, "grad_norm": 0.8733697533607483, "learning_rate": 7.749824358439093e-05, "loss": 3.2894, "step": 178900 }, { "epoch": 1.580299899327081, "grad_norm": 2.241514205932617, "learning_rate": 7.748663660321511e-05, "loss": 3.0506, "step": 178950 }, { "epoch": 1.580741447217365, "grad_norm": 4.648041725158691, "learning_rate": 7.74750274989297e-05, "loss": 3.3081, "step": 179000 }, { "epoch": 1.5811829951076493, "grad_norm": 2.8296263217926025, "learning_rate": 7.74634162724314e-05, "loss": 3.2342, "step": 179050 }, { "epoch": 1.5816245429979334, "grad_norm": 2.477835178375244, "learning_rate": 7.745180292461707e-05, "loss": 2.7343, "step": 179100 }, { "epoch": 1.5820660908882176, "grad_norm": 1.494752287864685, "learning_rate": 7.744018745638375e-05, "loss": 3.4856, "step": 179150 }, { "epoch": 1.5825076387785018, "grad_norm": 3.970940589904785, "learning_rate": 7.742856986862864e-05, "loss": 3.1116, "step": 179200 }, { "epoch": 1.5829491866687861, "grad_norm": 2.4596049785614014, "learning_rate": 7.741695016224908e-05, "loss": 2.8419, "step": 179250 }, { "epoch": 1.5833907345590703, "grad_norm": 1.8739557266235352, "learning_rate": 7.740532833814263e-05, "loss": 3.136, "step": 179300 }, { "epoch": 1.5838322824493545, "grad_norm": 1.3350729942321777, "learning_rate": 7.739370439720693e-05, "loss": 3.5008, "step": 179350 }, { "epoch": 1.5842738303396386, "grad_norm": 5.471927165985107, "learning_rate": 7.738207834033987e-05, "loss": 3.4203, "step": 179400 }, { "epoch": 1.5847153782299228, "grad_norm": 2.8407082557678223, "learning_rate": 7.737045016843944e-05, "loss": 3.3583, "step": 179450 }, { "epoch": 1.585156926120207, "grad_norm": 2.555143356323242, "learning_rate": 7.735881988240386e-05, "loss": 3.4862, "step": 179500 }, { "epoch": 1.5855984740104911, "grad_norm": 3.2967026233673096, "learning_rate": 7.734718748313143e-05, "loss": 3.1858, "step": 179550 }, { "epoch": 1.5860400219007753, "grad_norm": 0.6562075614929199, "learning_rate": 7.733555297152064e-05, "loss": 3.2385, "step": 179600 }, { "epoch": 1.5864815697910597, "grad_norm": 1.7235629558563232, "learning_rate": 7.73239163484702e-05, "loss": 3.3023, "step": 179650 }, { "epoch": 1.5869231176813439, "grad_norm": 3.370729446411133, "learning_rate": 7.731227761487893e-05, "loss": 3.463, "step": 179700 }, { "epoch": 1.587364665571628, "grad_norm": 4.0230207443237305, "learning_rate": 7.73006367716458e-05, "loss": 3.0753, "step": 179750 }, { "epoch": 1.5878062134619122, "grad_norm": 3.8130829334259033, "learning_rate": 7.728899381966999e-05, "loss": 2.8752, "step": 179800 }, { "epoch": 1.5882477613521964, "grad_norm": 4.115664482116699, "learning_rate": 7.72773487598508e-05, "loss": 3.072, "step": 179850 }, { "epoch": 1.5886893092424805, "grad_norm": 1.7616603374481201, "learning_rate": 7.726570159308774e-05, "loss": 3.1123, "step": 179900 }, { "epoch": 1.5891308571327647, "grad_norm": 3.976717472076416, "learning_rate": 7.725405232028044e-05, "loss": 3.1311, "step": 179950 }, { "epoch": 1.5895724050230489, "grad_norm": 1.6560734510421753, "learning_rate": 7.724240094232868e-05, "loss": 3.2467, "step": 180000 }, { "epoch": 1.5895724050230489, "eval_asr_loss": 0.905089343649495, "eval_loss": 2.9030532836914062, "eval_runtime": 20.6595, "eval_samples_per_second": 37.174, "eval_steps_per_second": 9.294, "eval_tts_loss": 5.932771591964687, "step": 180000 }, { "epoch": 1.590013952913333, "grad_norm": 3.773517608642578, "learning_rate": 7.723074746013248e-05, "loss": 3.554, "step": 180050 }, { "epoch": 1.5904555008036172, "grad_norm": 3.9356374740600586, "learning_rate": 7.721909187459193e-05, "loss": 3.2263, "step": 180100 }, { "epoch": 1.5908970486939014, "grad_norm": 4.545897483825684, "learning_rate": 7.720743418660736e-05, "loss": 3.3664, "step": 180150 }, { "epoch": 1.5913385965841855, "grad_norm": 3.460737466812134, "learning_rate": 7.719577439707921e-05, "loss": 3.0415, "step": 180200 }, { "epoch": 1.5917801444744697, "grad_norm": 1.8591420650482178, "learning_rate": 7.718411250690808e-05, "loss": 3.3958, "step": 180250 }, { "epoch": 1.5922216923647539, "grad_norm": 3.867363929748535, "learning_rate": 7.717244851699481e-05, "loss": 3.0154, "step": 180300 }, { "epoch": 1.592663240255038, "grad_norm": 1.8669315576553345, "learning_rate": 7.716078242824028e-05, "loss": 3.0492, "step": 180350 }, { "epoch": 1.5931047881453222, "grad_norm": 2.609985589981079, "learning_rate": 7.714911424154566e-05, "loss": 3.1507, "step": 180400 }, { "epoch": 1.5935463360356064, "grad_norm": 3.6491219997406006, "learning_rate": 7.713744395781214e-05, "loss": 3.1012, "step": 180450 }, { "epoch": 1.5939878839258905, "grad_norm": 1.0326428413391113, "learning_rate": 7.71257715779412e-05, "loss": 3.4021, "step": 180500 }, { "epoch": 1.5944294318161747, "grad_norm": 3.4464635848999023, "learning_rate": 7.711409710283443e-05, "loss": 3.2061, "step": 180550 }, { "epoch": 1.5948709797064589, "grad_norm": 3.594515085220337, "learning_rate": 7.71024205333936e-05, "loss": 3.0179, "step": 180600 }, { "epoch": 1.595312527596743, "grad_norm": 3.2505548000335693, "learning_rate": 7.709074187052059e-05, "loss": 2.9457, "step": 180650 }, { "epoch": 1.5957540754870272, "grad_norm": 3.6564249992370605, "learning_rate": 7.70790611151175e-05, "loss": 2.9484, "step": 180700 }, { "epoch": 1.5961956233773114, "grad_norm": 2.443648099899292, "learning_rate": 7.706737826808654e-05, "loss": 3.2759, "step": 180750 }, { "epoch": 1.5966371712675955, "grad_norm": 3.290959119796753, "learning_rate": 7.705569333033016e-05, "loss": 3.3899, "step": 180800 }, { "epoch": 1.59707871915788, "grad_norm": 4.142195224761963, "learning_rate": 7.704400630275089e-05, "loss": 3.1456, "step": 180850 }, { "epoch": 1.597520267048164, "grad_norm": 3.0236165523529053, "learning_rate": 7.703231718625143e-05, "loss": 3.1064, "step": 180900 }, { "epoch": 1.5979618149384482, "grad_norm": 1.8898965120315552, "learning_rate": 7.702062598173473e-05, "loss": 3.1881, "step": 180950 }, { "epoch": 1.5984033628287324, "grad_norm": 4.283849716186523, "learning_rate": 7.700893269010378e-05, "loss": 3.1001, "step": 181000 }, { "epoch": 1.5988449107190166, "grad_norm": 1.4696093797683716, "learning_rate": 7.699723731226182e-05, "loss": 3.0011, "step": 181050 }, { "epoch": 1.5992864586093007, "grad_norm": 2.0414416790008545, "learning_rate": 7.698553984911223e-05, "loss": 3.6156, "step": 181100 }, { "epoch": 1.599728006499585, "grad_norm": 4.270824432373047, "learning_rate": 7.697384030155848e-05, "loss": 3.7908, "step": 181150 }, { "epoch": 1.600169554389869, "grad_norm": 6.27476167678833, "learning_rate": 7.69621386705043e-05, "loss": 3.1732, "step": 181200 }, { "epoch": 1.6006111022801535, "grad_norm": 2.570694923400879, "learning_rate": 7.695043495685356e-05, "loss": 3.2865, "step": 181250 }, { "epoch": 1.6010526501704376, "grad_norm": 3.351912260055542, "learning_rate": 7.693872916151025e-05, "loss": 3.4516, "step": 181300 }, { "epoch": 1.6014941980607218, "grad_norm": 2.8289058208465576, "learning_rate": 7.692702128537855e-05, "loss": 3.7611, "step": 181350 }, { "epoch": 1.601935745951006, "grad_norm": 5.322634696960449, "learning_rate": 7.691531132936277e-05, "loss": 3.226, "step": 181400 }, { "epoch": 1.6023772938412901, "grad_norm": 1.7941278219223022, "learning_rate": 7.690359929436743e-05, "loss": 3.2578, "step": 181450 }, { "epoch": 1.6028188417315743, "grad_norm": 0.7209708094596863, "learning_rate": 7.689188518129719e-05, "loss": 3.514, "step": 181500 }, { "epoch": 1.6032603896218585, "grad_norm": 6.211198806762695, "learning_rate": 7.688016899105685e-05, "loss": 2.9773, "step": 181550 }, { "epoch": 1.6037019375121426, "grad_norm": 0.9564201831817627, "learning_rate": 7.68684507245514e-05, "loss": 3.0722, "step": 181600 }, { "epoch": 1.6041434854024268, "grad_norm": 1.6184066534042358, "learning_rate": 7.685673038268596e-05, "loss": 2.9939, "step": 181650 }, { "epoch": 1.604585033292711, "grad_norm": 1.1458969116210938, "learning_rate": 7.684500796636585e-05, "loss": 3.48, "step": 181700 }, { "epoch": 1.6050265811829951, "grad_norm": 1.6181378364562988, "learning_rate": 7.68332834764965e-05, "loss": 2.8879, "step": 181750 }, { "epoch": 1.6054681290732793, "grad_norm": 6.061189651489258, "learning_rate": 7.682155691398355e-05, "loss": 3.367, "step": 181800 }, { "epoch": 1.6059096769635635, "grad_norm": 3.899233818054199, "learning_rate": 7.680982827973275e-05, "loss": 3.1318, "step": 181850 }, { "epoch": 1.6063512248538476, "grad_norm": 3.1183807849884033, "learning_rate": 7.679809757465007e-05, "loss": 3.4564, "step": 181900 }, { "epoch": 1.6067927727441318, "grad_norm": 0.9795408844947815, "learning_rate": 7.678636479964159e-05, "loss": 3.5955, "step": 181950 }, { "epoch": 1.607234320634416, "grad_norm": 1.9428582191467285, "learning_rate": 7.677462995561357e-05, "loss": 3.2042, "step": 182000 }, { "epoch": 1.6076758685247001, "grad_norm": 3.8682286739349365, "learning_rate": 7.676289304347245e-05, "loss": 3.096, "step": 182050 }, { "epoch": 1.6081174164149843, "grad_norm": 1.1212639808654785, "learning_rate": 7.675115406412476e-05, "loss": 3.3886, "step": 182100 }, { "epoch": 1.6085589643052685, "grad_norm": 2.1383824348449707, "learning_rate": 7.673941301847727e-05, "loss": 3.1706, "step": 182150 }, { "epoch": 1.6090005121955526, "grad_norm": 1.5465537309646606, "learning_rate": 7.672766990743687e-05, "loss": 3.1482, "step": 182200 }, { "epoch": 1.6094420600858368, "grad_norm": 1.8144700527191162, "learning_rate": 7.67159247319106e-05, "loss": 3.2361, "step": 182250 }, { "epoch": 1.609883607976121, "grad_norm": 1.2655187845230103, "learning_rate": 7.67041774928057e-05, "loss": 2.5676, "step": 182300 }, { "epoch": 1.6103251558664051, "grad_norm": 2.401733636856079, "learning_rate": 7.669242819102953e-05, "loss": 3.006, "step": 182350 }, { "epoch": 1.6107667037566893, "grad_norm": 0.8709122538566589, "learning_rate": 7.668067682748963e-05, "loss": 3.1246, "step": 182400 }, { "epoch": 1.6112082516469737, "grad_norm": 0.2746553421020508, "learning_rate": 7.666892340309371e-05, "loss": 2.8896, "step": 182450 }, { "epoch": 1.6116497995372578, "grad_norm": 2.582265615463257, "learning_rate": 7.665716791874958e-05, "loss": 3.1229, "step": 182500 }, { "epoch": 1.612091347427542, "grad_norm": 2.9576897621154785, "learning_rate": 7.664541037536526e-05, "loss": 2.6456, "step": 182550 }, { "epoch": 1.6125328953178262, "grad_norm": 3.0392637252807617, "learning_rate": 7.663365077384898e-05, "loss": 3.5719, "step": 182600 }, { "epoch": 1.6129744432081103, "grad_norm": 1.836568832397461, "learning_rate": 7.6621889115109e-05, "loss": 3.2949, "step": 182650 }, { "epoch": 1.6134159910983945, "grad_norm": 1.3547306060791016, "learning_rate": 7.661012540005381e-05, "loss": 3.3838, "step": 182700 }, { "epoch": 1.6138575389886787, "grad_norm": 3.1797091960906982, "learning_rate": 7.659835962959211e-05, "loss": 3.1803, "step": 182750 }, { "epoch": 1.6142990868789628, "grad_norm": 1.4783034324645996, "learning_rate": 7.658659180463267e-05, "loss": 3.308, "step": 182800 }, { "epoch": 1.6147406347692472, "grad_norm": 3.789548873901367, "learning_rate": 7.657482192608445e-05, "loss": 3.2409, "step": 182850 }, { "epoch": 1.6151821826595314, "grad_norm": 5.076577186584473, "learning_rate": 7.65630499948566e-05, "loss": 2.4807, "step": 182900 }, { "epoch": 1.6156237305498156, "grad_norm": 4.224163055419922, "learning_rate": 7.655127601185837e-05, "loss": 2.828, "step": 182950 }, { "epoch": 1.6160652784400997, "grad_norm": 3.1117782592773438, "learning_rate": 7.653949997799924e-05, "loss": 2.9108, "step": 183000 }, { "epoch": 1.6160652784400997, "eval_asr_loss": 0.9132444849537853, "eval_loss": 2.907047986984253, "eval_runtime": 21.3259, "eval_samples_per_second": 36.013, "eval_steps_per_second": 9.003, "eval_tts_loss": 5.958474148858937, "step": 183000 }, { "epoch": 1.616506826330384, "grad_norm": 1.899177074432373, "learning_rate": 7.652772189418875e-05, "loss": 3.2806, "step": 183050 }, { "epoch": 1.616948374220668, "grad_norm": 4.129954814910889, "learning_rate": 7.651594176133672e-05, "loss": 2.7331, "step": 183100 }, { "epoch": 1.6173899221109522, "grad_norm": 1.1719375848770142, "learning_rate": 7.650415958035303e-05, "loss": 3.4316, "step": 183150 }, { "epoch": 1.6178314700012364, "grad_norm": 1.4842125177383423, "learning_rate": 7.649237535214776e-05, "loss": 3.2723, "step": 183200 }, { "epoch": 1.6182730178915206, "grad_norm": 2.0932915210723877, "learning_rate": 7.648058907763114e-05, "loss": 3.2312, "step": 183250 }, { "epoch": 1.6187145657818047, "grad_norm": 2.771362066268921, "learning_rate": 7.646880075771357e-05, "loss": 3.5672, "step": 183300 }, { "epoch": 1.619156113672089, "grad_norm": 1.0976365804672241, "learning_rate": 7.645724622062386e-05, "loss": 3.2644, "step": 183350 }, { "epoch": 1.619597661562373, "grad_norm": 2.213742971420288, "learning_rate": 7.644545385349884e-05, "loss": 3.3306, "step": 183400 }, { "epoch": 1.6200392094526572, "grad_norm": 4.350307464599609, "learning_rate": 7.643365944368675e-05, "loss": 3.603, "step": 183450 }, { "epoch": 1.6204807573429414, "grad_norm": 2.864485263824463, "learning_rate": 7.642186299209865e-05, "loss": 3.2446, "step": 183500 }, { "epoch": 1.6209223052332256, "grad_norm": 3.199735403060913, "learning_rate": 7.641006449964566e-05, "loss": 3.2124, "step": 183550 }, { "epoch": 1.6213638531235097, "grad_norm": 4.11868143081665, "learning_rate": 7.639826396723914e-05, "loss": 3.0056, "step": 183600 }, { "epoch": 1.621805401013794, "grad_norm": 1.4297511577606201, "learning_rate": 7.638646139579058e-05, "loss": 3.5371, "step": 183650 }, { "epoch": 1.622246948904078, "grad_norm": 2.8237781524658203, "learning_rate": 7.637465678621162e-05, "loss": 3.2139, "step": 183700 }, { "epoch": 1.6226884967943622, "grad_norm": 4.127041339874268, "learning_rate": 7.636308629230888e-05, "loss": 3.174, "step": 183750 }, { "epoch": 1.6231300446846464, "grad_norm": 3.040043592453003, "learning_rate": 7.635127764992194e-05, "loss": 3.4317, "step": 183800 }, { "epoch": 1.6235715925749306, "grad_norm": 2.9634487628936768, "learning_rate": 7.633946697212221e-05, "loss": 3.2735, "step": 183850 }, { "epoch": 1.6240131404652147, "grad_norm": 4.035878658294678, "learning_rate": 7.632765425982202e-05, "loss": 3.3785, "step": 183900 }, { "epoch": 1.6244546883554989, "grad_norm": 3.683920383453369, "learning_rate": 7.631583951393382e-05, "loss": 3.1896, "step": 183950 }, { "epoch": 1.624896236245783, "grad_norm": 2.083125352859497, "learning_rate": 7.630402273537014e-05, "loss": 2.9794, "step": 184000 }, { "epoch": 1.6253377841360674, "grad_norm": 2.1921465396881104, "learning_rate": 7.629220392504377e-05, "loss": 3.3312, "step": 184050 }, { "epoch": 1.6257793320263516, "grad_norm": 1.8298650979995728, "learning_rate": 7.628038308386757e-05, "loss": 3.3576, "step": 184100 }, { "epoch": 1.6262208799166358, "grad_norm": 1.597791314125061, "learning_rate": 7.626856021275465e-05, "loss": 3.5873, "step": 184150 }, { "epoch": 1.62666242780692, "grad_norm": 3.174025774002075, "learning_rate": 7.62567353126182e-05, "loss": 3.1898, "step": 184200 }, { "epoch": 1.627103975697204, "grad_norm": 2.265221357345581, "learning_rate": 7.624490838437161e-05, "loss": 3.1095, "step": 184250 }, { "epoch": 1.6275455235874883, "grad_norm": 2.2410521507263184, "learning_rate": 7.623307942892839e-05, "loss": 2.921, "step": 184300 }, { "epoch": 1.6279870714777724, "grad_norm": 2.9074504375457764, "learning_rate": 7.622124844720224e-05, "loss": 3.3522, "step": 184350 }, { "epoch": 1.6284286193680566, "grad_norm": 2.2740726470947266, "learning_rate": 7.620941544010697e-05, "loss": 3.4447, "step": 184400 }, { "epoch": 1.628870167258341, "grad_norm": 3.019122838973999, "learning_rate": 7.619758040855664e-05, "loss": 3.183, "step": 184450 }, { "epoch": 1.6293117151486252, "grad_norm": 2.1911163330078125, "learning_rate": 7.618574335346536e-05, "loss": 3.0883, "step": 184500 }, { "epoch": 1.6297532630389093, "grad_norm": 1.324202299118042, "learning_rate": 7.617390427574744e-05, "loss": 3.1827, "step": 184550 }, { "epoch": 1.6301948109291935, "grad_norm": 2.130458116531372, "learning_rate": 7.61620631763174e-05, "loss": 3.4636, "step": 184600 }, { "epoch": 1.6306363588194777, "grad_norm": 3.302278757095337, "learning_rate": 7.61502200560898e-05, "loss": 3.1299, "step": 184650 }, { "epoch": 1.6310779067097618, "grad_norm": 0.9419139623641968, "learning_rate": 7.613837491597944e-05, "loss": 3.2849, "step": 184700 }, { "epoch": 1.631519454600046, "grad_norm": 4.2170186042785645, "learning_rate": 7.612652775690127e-05, "loss": 2.7851, "step": 184750 }, { "epoch": 1.6319610024903302, "grad_norm": 3.18815541267395, "learning_rate": 7.611467857977039e-05, "loss": 3.372, "step": 184800 }, { "epoch": 1.6324025503806143, "grad_norm": 4.340141773223877, "learning_rate": 7.610282738550204e-05, "loss": 3.2953, "step": 184850 }, { "epoch": 1.6328440982708985, "grad_norm": 4.1585869789123535, "learning_rate": 7.60909741750116e-05, "loss": 2.9285, "step": 184900 }, { "epoch": 1.6332856461611827, "grad_norm": 2.83577299118042, "learning_rate": 7.607911894921467e-05, "loss": 3.2021, "step": 184950 }, { "epoch": 1.6337271940514668, "grad_norm": 1.7795286178588867, "learning_rate": 7.606726170902693e-05, "loss": 3.1474, "step": 185000 }, { "epoch": 1.634168741941751, "grad_norm": 2.461143732070923, "learning_rate": 7.605540245536427e-05, "loss": 2.8676, "step": 185050 }, { "epoch": 1.6346102898320352, "grad_norm": 1.4866169691085815, "learning_rate": 7.604354118914272e-05, "loss": 3.1229, "step": 185100 }, { "epoch": 1.6350518377223193, "grad_norm": 3.6347692012786865, "learning_rate": 7.603167791127845e-05, "loss": 3.7924, "step": 185150 }, { "epoch": 1.6354933856126035, "grad_norm": 2.398576259613037, "learning_rate": 7.60198126226878e-05, "loss": 3.2153, "step": 185200 }, { "epoch": 1.6359349335028877, "grad_norm": 2.5132200717926025, "learning_rate": 7.600794532428728e-05, "loss": 3.4534, "step": 185250 }, { "epoch": 1.6363764813931718, "grad_norm": 2.564957857131958, "learning_rate": 7.59960760169935e-05, "loss": 3.1164, "step": 185300 }, { "epoch": 1.636818029283456, "grad_norm": 4.661256313323975, "learning_rate": 7.598420470172331e-05, "loss": 3.5114, "step": 185350 }, { "epoch": 1.6372595771737402, "grad_norm": 1.975266695022583, "learning_rate": 7.597233137939364e-05, "loss": 3.399, "step": 185400 }, { "epoch": 1.6377011250640243, "grad_norm": 3.3702497482299805, "learning_rate": 7.59604560509216e-05, "loss": 3.1902, "step": 185450 }, { "epoch": 1.6381426729543085, "grad_norm": 3.0457959175109863, "learning_rate": 7.594857871722449e-05, "loss": 2.8022, "step": 185500 }, { "epoch": 1.6385842208445927, "grad_norm": 3.148191213607788, "learning_rate": 7.59366993792197e-05, "loss": 3.1881, "step": 185550 }, { "epoch": 1.6390257687348768, "grad_norm": 2.3851609230041504, "learning_rate": 7.592481803782483e-05, "loss": 2.9736, "step": 185600 }, { "epoch": 1.6394673166251612, "grad_norm": 0.8741468191146851, "learning_rate": 7.591293469395758e-05, "loss": 3.0127, "step": 185650 }, { "epoch": 1.6399088645154454, "grad_norm": 4.1316046714782715, "learning_rate": 7.590104934853588e-05, "loss": 3.2769, "step": 185700 }, { "epoch": 1.6403504124057295, "grad_norm": 1.8304107189178467, "learning_rate": 7.588916200247775e-05, "loss": 3.5342, "step": 185750 }, { "epoch": 1.6407919602960137, "grad_norm": 2.1133339405059814, "learning_rate": 7.587727265670137e-05, "loss": 3.3722, "step": 185800 }, { "epoch": 1.6412335081862979, "grad_norm": 2.7472083568573, "learning_rate": 7.586538131212512e-05, "loss": 3.2366, "step": 185850 }, { "epoch": 1.641675056076582, "grad_norm": 2.5508406162261963, "learning_rate": 7.58534879696675e-05, "loss": 3.3287, "step": 185900 }, { "epoch": 1.6421166039668662, "grad_norm": 5.497042179107666, "learning_rate": 7.584159263024718e-05, "loss": 3.3166, "step": 185950 }, { "epoch": 1.6425581518571504, "grad_norm": 4.364591598510742, "learning_rate": 7.582969529478294e-05, "loss": 3.5261, "step": 186000 }, { "epoch": 1.6425581518571504, "eval_asr_loss": 0.9146858002053176, "eval_loss": 2.897829055786133, "eval_runtime": 21.1777, "eval_samples_per_second": 36.265, "eval_steps_per_second": 9.066, "eval_tts_loss": 5.956395547650511, "step": 186000 }, { "epoch": 1.6429996997474348, "grad_norm": 3.4542715549468994, "learning_rate": 7.581779596419378e-05, "loss": 2.9081, "step": 186050 }, { "epoch": 1.643441247637719, "grad_norm": 4.252161502838135, "learning_rate": 7.58058946393988e-05, "loss": 3.3154, "step": 186100 }, { "epoch": 1.643882795528003, "grad_norm": 3.3466477394104004, "learning_rate": 7.579399132131731e-05, "loss": 3.35, "step": 186150 }, { "epoch": 1.6443243434182873, "grad_norm": 3.145517349243164, "learning_rate": 7.578208601086868e-05, "loss": 3.3923, "step": 186200 }, { "epoch": 1.6447658913085714, "grad_norm": 3.68949556350708, "learning_rate": 7.577017870897253e-05, "loss": 3.0109, "step": 186250 }, { "epoch": 1.6452074391988556, "grad_norm": 1.4948679208755493, "learning_rate": 7.575826941654862e-05, "loss": 3.4221, "step": 186300 }, { "epoch": 1.6456489870891398, "grad_norm": 6.5033488273620605, "learning_rate": 7.574635813451684e-05, "loss": 2.666, "step": 186350 }, { "epoch": 1.646090534979424, "grad_norm": 2.8202450275421143, "learning_rate": 7.57344448637972e-05, "loss": 3.3709, "step": 186400 }, { "epoch": 1.646532082869708, "grad_norm": 1.8358458280563354, "learning_rate": 7.57225296053099e-05, "loss": 3.1007, "step": 186450 }, { "epoch": 1.6469736307599923, "grad_norm": 1.9176243543624878, "learning_rate": 7.571061235997531e-05, "loss": 3.1518, "step": 186500 }, { "epoch": 1.6474151786502764, "grad_norm": 0.955685555934906, "learning_rate": 7.569869312871396e-05, "loss": 3.6704, "step": 186550 }, { "epoch": 1.6478567265405606, "grad_norm": 4.65590763092041, "learning_rate": 7.568677191244646e-05, "loss": 2.8114, "step": 186600 }, { "epoch": 1.6482982744308448, "grad_norm": 1.1020005941390991, "learning_rate": 7.567484871209367e-05, "loss": 3.2136, "step": 186650 }, { "epoch": 1.648739822321129, "grad_norm": 2.0391685962677, "learning_rate": 7.566292352857651e-05, "loss": 3.4038, "step": 186700 }, { "epoch": 1.649181370211413, "grad_norm": 2.0860207080841064, "learning_rate": 7.565099636281614e-05, "loss": 3.4063, "step": 186750 }, { "epoch": 1.6496229181016973, "grad_norm": 2.8292300701141357, "learning_rate": 7.563906721573379e-05, "loss": 3.2938, "step": 186800 }, { "epoch": 1.6500644659919814, "grad_norm": 3.6155056953430176, "learning_rate": 7.562713608825094e-05, "loss": 3.0306, "step": 186850 }, { "epoch": 1.6505060138822656, "grad_norm": 1.792187213897705, "learning_rate": 7.561520298128913e-05, "loss": 2.9088, "step": 186900 }, { "epoch": 1.6509475617725498, "grad_norm": 3.0184032917022705, "learning_rate": 7.56032678957701e-05, "loss": 3.7706, "step": 186950 }, { "epoch": 1.651389109662834, "grad_norm": 1.4571894407272339, "learning_rate": 7.559133083261574e-05, "loss": 2.7708, "step": 187000 }, { "epoch": 1.651830657553118, "grad_norm": 1.022334337234497, "learning_rate": 7.557939179274807e-05, "loss": 3.2407, "step": 187050 }, { "epoch": 1.6522722054434023, "grad_norm": 2.326892137527466, "learning_rate": 7.556745077708928e-05, "loss": 3.0828, "step": 187100 }, { "epoch": 1.6527137533336864, "grad_norm": 7.504704475402832, "learning_rate": 7.555550778656176e-05, "loss": 2.9196, "step": 187150 }, { "epoch": 1.6531553012239706, "grad_norm": 1.7373300790786743, "learning_rate": 7.554356282208792e-05, "loss": 3.4106, "step": 187200 }, { "epoch": 1.653596849114255, "grad_norm": 7.9670538902282715, "learning_rate": 7.553185484267011e-05, "loss": 2.9811, "step": 187250 }, { "epoch": 1.6540383970045391, "grad_norm": 4.288424015045166, "learning_rate": 7.551990597250481e-05, "loss": 2.9887, "step": 187300 }, { "epoch": 1.6544799448948233, "grad_norm": 3.482987642288208, "learning_rate": 7.550795513114318e-05, "loss": 3.6097, "step": 187350 }, { "epoch": 1.6549214927851075, "grad_norm": 2.8395724296569824, "learning_rate": 7.549600231950832e-05, "loss": 3.4265, "step": 187400 }, { "epoch": 1.6553630406753916, "grad_norm": 2.013218641281128, "learning_rate": 7.548404753852349e-05, "loss": 3.6109, "step": 187450 }, { "epoch": 1.6558045885656758, "grad_norm": 1.625708818435669, "learning_rate": 7.547209078911207e-05, "loss": 3.0623, "step": 187500 }, { "epoch": 1.65624613645596, "grad_norm": 1.8119765520095825, "learning_rate": 7.546013207219765e-05, "loss": 3.1547, "step": 187550 }, { "epoch": 1.6566876843462441, "grad_norm": 5.0249481201171875, "learning_rate": 7.54481713887039e-05, "loss": 2.7106, "step": 187600 }, { "epoch": 1.6571292322365285, "grad_norm": 1.6827908754348755, "learning_rate": 7.543620873955472e-05, "loss": 3.3438, "step": 187650 }, { "epoch": 1.6575707801268127, "grad_norm": 1.1493171453475952, "learning_rate": 7.542424412567412e-05, "loss": 2.9085, "step": 187700 }, { "epoch": 1.6580123280170969, "grad_norm": 1.2166669368743896, "learning_rate": 7.541227754798624e-05, "loss": 2.9538, "step": 187750 }, { "epoch": 1.658453875907381, "grad_norm": 3.9026665687561035, "learning_rate": 7.540030900741539e-05, "loss": 3.4233, "step": 187800 }, { "epoch": 1.6588954237976652, "grad_norm": 2.1174581050872803, "learning_rate": 7.538833850488609e-05, "loss": 3.0982, "step": 187850 }, { "epoch": 1.6593369716879494, "grad_norm": 1.8407869338989258, "learning_rate": 7.53763660413229e-05, "loss": 3.239, "step": 187900 }, { "epoch": 1.6597785195782335, "grad_norm": 4.55293607711792, "learning_rate": 7.536439161765063e-05, "loss": 3.4121, "step": 187950 }, { "epoch": 1.6602200674685177, "grad_norm": 2.4583520889282227, "learning_rate": 7.53524152347942e-05, "loss": 3.2692, "step": 188000 }, { "epoch": 1.6606616153588019, "grad_norm": 3.5421550273895264, "learning_rate": 7.534043689367867e-05, "loss": 3.1821, "step": 188050 }, { "epoch": 1.661103163249086, "grad_norm": 3.7831602096557617, "learning_rate": 7.532845659522927e-05, "loss": 3.74, "step": 188100 }, { "epoch": 1.6615447111393702, "grad_norm": 1.2975730895996094, "learning_rate": 7.531647434037137e-05, "loss": 2.9245, "step": 188150 }, { "epoch": 1.6619862590296544, "grad_norm": 3.5446507930755615, "learning_rate": 7.530449013003051e-05, "loss": 3.12, "step": 188200 }, { "epoch": 1.6624278069199385, "grad_norm": 1.0230252742767334, "learning_rate": 7.529250396513237e-05, "loss": 3.3197, "step": 188250 }, { "epoch": 1.6628693548102227, "grad_norm": 2.7343640327453613, "learning_rate": 7.528051584660276e-05, "loss": 3.3, "step": 188300 }, { "epoch": 1.6633109027005069, "grad_norm": 0.8870599269866943, "learning_rate": 7.526852577536769e-05, "loss": 3.054, "step": 188350 }, { "epoch": 1.663752450590791, "grad_norm": 3.2465200424194336, "learning_rate": 7.525653375235327e-05, "loss": 2.9143, "step": 188400 }, { "epoch": 1.6641939984810752, "grad_norm": 3.82564377784729, "learning_rate": 7.524453977848578e-05, "loss": 3.2014, "step": 188450 }, { "epoch": 1.6646355463713594, "grad_norm": 1.1681989431381226, "learning_rate": 7.523254385469166e-05, "loss": 3.2335, "step": 188500 }, { "epoch": 1.6650770942616435, "grad_norm": 1.4739675521850586, "learning_rate": 7.522054598189747e-05, "loss": 2.8809, "step": 188550 }, { "epoch": 1.6655186421519277, "grad_norm": 2.9741079807281494, "learning_rate": 7.520854616103002e-05, "loss": 3.5088, "step": 188600 }, { "epoch": 1.6659601900422119, "grad_norm": 2.840975284576416, "learning_rate": 7.519654439301609e-05, "loss": 3.3214, "step": 188650 }, { "epoch": 1.666401737932496, "grad_norm": 3.7124547958374023, "learning_rate": 7.51845406787828e-05, "loss": 3.6984, "step": 188700 }, { "epoch": 1.6668432858227802, "grad_norm": 6.617938995361328, "learning_rate": 7.517253501925727e-05, "loss": 3.3593, "step": 188750 }, { "epoch": 1.6672848337130646, "grad_norm": 1.4384989738464355, "learning_rate": 7.51605274153669e-05, "loss": 3.458, "step": 188800 }, { "epoch": 1.6677263816033487, "grad_norm": 4.336093902587891, "learning_rate": 7.514851786803913e-05, "loss": 3.294, "step": 188850 }, { "epoch": 1.668167929493633, "grad_norm": 1.1996614933013916, "learning_rate": 7.51365063782016e-05, "loss": 3.3519, "step": 188900 }, { "epoch": 1.668609477383917, "grad_norm": 1.576010823249817, "learning_rate": 7.512449294678212e-05, "loss": 3.2801, "step": 188950 }, { "epoch": 1.6690510252742012, "grad_norm": 0.6536944508552551, "learning_rate": 7.511247757470859e-05, "loss": 3.2111, "step": 189000 }, { "epoch": 1.6690510252742012, "eval_asr_loss": 0.9149196284461532, "eval_loss": 2.8986892700195312, "eval_runtime": 20.5204, "eval_samples_per_second": 37.426, "eval_steps_per_second": 9.357, "eval_tts_loss": 5.990396785670996, "step": 189000 }, { "epoch": 1.6694925731644854, "grad_norm": 3.608792543411255, "learning_rate": 7.510046026290912e-05, "loss": 3.178, "step": 189050 }, { "epoch": 1.6699341210547696, "grad_norm": 4.6743245124816895, "learning_rate": 7.508844101231195e-05, "loss": 2.9165, "step": 189100 }, { "epoch": 1.6703756689450537, "grad_norm": 1.1375796794891357, "learning_rate": 7.507641982384543e-05, "loss": 3.3253, "step": 189150 }, { "epoch": 1.6708172168353381, "grad_norm": 0.7114461660385132, "learning_rate": 7.506439669843816e-05, "loss": 3.0497, "step": 189200 }, { "epoch": 1.6712587647256223, "grad_norm": 4.1343889236450195, "learning_rate": 7.505237163701877e-05, "loss": 3.3836, "step": 189250 }, { "epoch": 1.6717003126159065, "grad_norm": 2.8716280460357666, "learning_rate": 7.504034464051609e-05, "loss": 3.3779, "step": 189300 }, { "epoch": 1.6721418605061906, "grad_norm": 1.575553059577942, "learning_rate": 7.502831570985915e-05, "loss": 3.185, "step": 189350 }, { "epoch": 1.6725834083964748, "grad_norm": 2.8956246376037598, "learning_rate": 7.501628484597702e-05, "loss": 3.5845, "step": 189400 }, { "epoch": 1.673024956286759, "grad_norm": 4.885554790496826, "learning_rate": 7.500425204979906e-05, "loss": 3.3725, "step": 189450 }, { "epoch": 1.6734665041770431, "grad_norm": 6.7091264724731445, "learning_rate": 7.499221732225464e-05, "loss": 3.3188, "step": 189500 }, { "epoch": 1.6739080520673273, "grad_norm": 1.5392342805862427, "learning_rate": 7.498018066427335e-05, "loss": 3.2148, "step": 189550 }, { "epoch": 1.6743495999576115, "grad_norm": 1.949496865272522, "learning_rate": 7.496814207678494e-05, "loss": 3.519, "step": 189600 }, { "epoch": 1.6747911478478956, "grad_norm": 2.188905715942383, "learning_rate": 7.495610156071927e-05, "loss": 2.8393, "step": 189650 }, { "epoch": 1.6752326957381798, "grad_norm": 2.987746477127075, "learning_rate": 7.494405911700638e-05, "loss": 3.0112, "step": 189700 }, { "epoch": 1.675674243628464, "grad_norm": 1.8022232055664062, "learning_rate": 7.493201474657646e-05, "loss": 3.2487, "step": 189750 }, { "epoch": 1.6761157915187481, "grad_norm": 0.8079404830932617, "learning_rate": 7.491996845035979e-05, "loss": 3.4419, "step": 189800 }, { "epoch": 1.6765573394090323, "grad_norm": 2.956101655960083, "learning_rate": 7.490792022928689e-05, "loss": 3.4542, "step": 189850 }, { "epoch": 1.6769988872993165, "grad_norm": 1.22069251537323, "learning_rate": 7.489587008428834e-05, "loss": 3.2008, "step": 189900 }, { "epoch": 1.6774404351896006, "grad_norm": 1.820133924484253, "learning_rate": 7.488381801629498e-05, "loss": 3.1799, "step": 189950 }, { "epoch": 1.6778819830798848, "grad_norm": 1.816726565361023, "learning_rate": 7.487176402623766e-05, "loss": 3.0303, "step": 190000 }, { "epoch": 1.678323530970169, "grad_norm": 2.491438150405884, "learning_rate": 7.485970811504748e-05, "loss": 3.1211, "step": 190050 }, { "epoch": 1.6787650788604531, "grad_norm": 1.7439849376678467, "learning_rate": 7.484765028365564e-05, "loss": 2.8709, "step": 190100 }, { "epoch": 1.6792066267507373, "grad_norm": 3.090697765350342, "learning_rate": 7.483559053299354e-05, "loss": 3.8603, "step": 190150 }, { "epoch": 1.6796481746410215, "grad_norm": 4.9195122718811035, "learning_rate": 7.482352886399265e-05, "loss": 3.1793, "step": 190200 }, { "epoch": 1.6800897225313056, "grad_norm": 6.847498416900635, "learning_rate": 7.481146527758468e-05, "loss": 3.0936, "step": 190250 }, { "epoch": 1.6805312704215898, "grad_norm": 2.2523751258850098, "learning_rate": 7.479939977470137e-05, "loss": 2.8686, "step": 190300 }, { "epoch": 1.680972818311874, "grad_norm": 3.1140174865722656, "learning_rate": 7.478733235627476e-05, "loss": 3.4243, "step": 190350 }, { "epoch": 1.6814143662021583, "grad_norm": 6.4473066329956055, "learning_rate": 7.477526302323691e-05, "loss": 2.9147, "step": 190400 }, { "epoch": 1.6818559140924425, "grad_norm": 5.672176837921143, "learning_rate": 7.476319177652005e-05, "loss": 2.9304, "step": 190450 }, { "epoch": 1.6822974619827267, "grad_norm": 2.4453301429748535, "learning_rate": 7.475111861705664e-05, "loss": 3.141, "step": 190500 }, { "epoch": 1.6827390098730108, "grad_norm": 4.314248561859131, "learning_rate": 7.473904354577918e-05, "loss": 2.9855, "step": 190550 }, { "epoch": 1.683180557763295, "grad_norm": 4.665952205657959, "learning_rate": 7.472696656362042e-05, "loss": 3.5754, "step": 190600 }, { "epoch": 1.6836221056535792, "grad_norm": 3.7188048362731934, "learning_rate": 7.471488767151315e-05, "loss": 3.052, "step": 190650 }, { "epoch": 1.6840636535438633, "grad_norm": 2.8768093585968018, "learning_rate": 7.470280687039037e-05, "loss": 3.331, "step": 190700 }, { "epoch": 1.6845052014341475, "grad_norm": 2.511436700820923, "learning_rate": 7.469072416118526e-05, "loss": 3.256, "step": 190750 }, { "epoch": 1.684946749324432, "grad_norm": 2.553476572036743, "learning_rate": 7.467863954483105e-05, "loss": 2.9051, "step": 190800 }, { "epoch": 1.685388297214716, "grad_norm": 2.4301607608795166, "learning_rate": 7.466655302226122e-05, "loss": 3.1829, "step": 190850 }, { "epoch": 1.6858298451050002, "grad_norm": 2.7761809825897217, "learning_rate": 7.465446459440935e-05, "loss": 3.1125, "step": 190900 }, { "epoch": 1.6862713929952844, "grad_norm": 4.545463562011719, "learning_rate": 7.464237426220914e-05, "loss": 3.0704, "step": 190950 }, { "epoch": 1.6867129408855686, "grad_norm": 2.1818442344665527, "learning_rate": 7.463028202659448e-05, "loss": 3.6544, "step": 191000 }, { "epoch": 1.6871544887758527, "grad_norm": 3.3841798305511475, "learning_rate": 7.461818788849938e-05, "loss": 3.333, "step": 191050 }, { "epoch": 1.687596036666137, "grad_norm": 1.6875982284545898, "learning_rate": 7.460609184885802e-05, "loss": 3.605, "step": 191100 }, { "epoch": 1.688037584556421, "grad_norm": 2.939258337020874, "learning_rate": 7.459399390860476e-05, "loss": 2.9491, "step": 191150 }, { "epoch": 1.6884791324467052, "grad_norm": 4.111961364746094, "learning_rate": 7.458189406867398e-05, "loss": 3.0345, "step": 191200 }, { "epoch": 1.6889206803369894, "grad_norm": 4.00358772277832, "learning_rate": 7.456979233000034e-05, "loss": 2.8185, "step": 191250 }, { "epoch": 1.6893622282272736, "grad_norm": 2.474588394165039, "learning_rate": 7.45576886935186e-05, "loss": 3.2167, "step": 191300 }, { "epoch": 1.6898037761175577, "grad_norm": 2.620952606201172, "learning_rate": 7.454558316016364e-05, "loss": 3.115, "step": 191350 }, { "epoch": 1.6902453240078419, "grad_norm": 5.4351701736450195, "learning_rate": 7.453347573087052e-05, "loss": 3.096, "step": 191400 }, { "epoch": 1.690686871898126, "grad_norm": 0.8584945797920227, "learning_rate": 7.452136640657445e-05, "loss": 3.0888, "step": 191450 }, { "epoch": 1.6911284197884102, "grad_norm": 0.5445427894592285, "learning_rate": 7.450925518821075e-05, "loss": 3.0229, "step": 191500 }, { "epoch": 1.6915699676786944, "grad_norm": 1.948606014251709, "learning_rate": 7.449714207671494e-05, "loss": 3.092, "step": 191550 }, { "epoch": 1.6920115155689786, "grad_norm": 3.2247259616851807, "learning_rate": 7.448502707302259e-05, "loss": 2.8165, "step": 191600 }, { "epoch": 1.6924530634592627, "grad_norm": 4.422319412231445, "learning_rate": 7.447291017806957e-05, "loss": 3.4485, "step": 191650 }, { "epoch": 1.6928946113495469, "grad_norm": 1.4744811058044434, "learning_rate": 7.446079139279174e-05, "loss": 3.064, "step": 191700 }, { "epoch": 1.693336159239831, "grad_norm": 2.2750120162963867, "learning_rate": 7.444867071812521e-05, "loss": 3.4599, "step": 191750 }, { "epoch": 1.6937777071301152, "grad_norm": 1.9034446477890015, "learning_rate": 7.443654815500617e-05, "loss": 2.9978, "step": 191800 }, { "epoch": 1.6942192550203994, "grad_norm": 2.9149630069732666, "learning_rate": 7.442442370437099e-05, "loss": 3.5718, "step": 191850 }, { "epoch": 1.6946608029106836, "grad_norm": 2.397642135620117, "learning_rate": 7.441229736715623e-05, "loss": 3.0129, "step": 191900 }, { "epoch": 1.6951023508009677, "grad_norm": 4.1654486656188965, "learning_rate": 7.440016914429848e-05, "loss": 2.7316, "step": 191950 }, { "epoch": 1.695543898691252, "grad_norm": 3.3984122276306152, "learning_rate": 7.438803903673459e-05, "loss": 3.4284, "step": 192000 }, { "epoch": 1.695543898691252, "eval_asr_loss": 0.9019791327945982, "eval_loss": 2.891016721725464, "eval_runtime": 20.3988, "eval_samples_per_second": 37.649, "eval_steps_per_second": 9.412, "eval_tts_loss": 5.950324852027226, "step": 192000 }, { "epoch": 1.6959854465815363, "grad_norm": 3.025519609451294, "learning_rate": 7.437590704540147e-05, "loss": 3.2471, "step": 192050 }, { "epoch": 1.6964269944718204, "grad_norm": 3.1683413982391357, "learning_rate": 7.436377317123624e-05, "loss": 3.032, "step": 192100 }, { "epoch": 1.6968685423621046, "grad_norm": 1.259204387664795, "learning_rate": 7.435163741517614e-05, "loss": 3.3301, "step": 192150 }, { "epoch": 1.6973100902523888, "grad_norm": 2.364607572555542, "learning_rate": 7.433949977815852e-05, "loss": 3.5812, "step": 192200 }, { "epoch": 1.697751638142673, "grad_norm": 2.005537748336792, "learning_rate": 7.432736026112097e-05, "loss": 3.3359, "step": 192250 }, { "epoch": 1.698193186032957, "grad_norm": 3.5940768718719482, "learning_rate": 7.431546171133245e-05, "loss": 3.0394, "step": 192300 }, { "epoch": 1.6986347339232413, "grad_norm": 1.2887500524520874, "learning_rate": 7.43033184746218e-05, "loss": 3.191, "step": 192350 }, { "epoch": 1.6990762818135257, "grad_norm": 1.4926159381866455, "learning_rate": 7.429117336068591e-05, "loss": 3.4094, "step": 192400 }, { "epoch": 1.6995178297038098, "grad_norm": 2.109112501144409, "learning_rate": 7.427902637046287e-05, "loss": 3.4759, "step": 192450 }, { "epoch": 1.699959377594094, "grad_norm": 1.6524863243103027, "learning_rate": 7.426687750489089e-05, "loss": 3.1766, "step": 192500 }, { "epoch": 1.7004009254843782, "grad_norm": 3.5204215049743652, "learning_rate": 7.425472676490842e-05, "loss": 3.17, "step": 192550 }, { "epoch": 1.7008424733746623, "grad_norm": 1.593204140663147, "learning_rate": 7.424257415145399e-05, "loss": 3.2035, "step": 192600 }, { "epoch": 1.7012840212649465, "grad_norm": 1.7033207416534424, "learning_rate": 7.423041966546626e-05, "loss": 3.3376, "step": 192650 }, { "epoch": 1.7017255691552307, "grad_norm": 4.6197733879089355, "learning_rate": 7.42182633078841e-05, "loss": 3.334, "step": 192700 }, { "epoch": 1.7021671170455148, "grad_norm": 2.908355236053467, "learning_rate": 7.420610507964644e-05, "loss": 3.0815, "step": 192750 }, { "epoch": 1.702608664935799, "grad_norm": 1.182983160018921, "learning_rate": 7.419394498169244e-05, "loss": 2.8714, "step": 192800 }, { "epoch": 1.7030502128260832, "grad_norm": 4.0162224769592285, "learning_rate": 7.418178301496134e-05, "loss": 2.9672, "step": 192850 }, { "epoch": 1.7034917607163673, "grad_norm": 3.590345859527588, "learning_rate": 7.416961918039256e-05, "loss": 3.1313, "step": 192900 }, { "epoch": 1.7039333086066515, "grad_norm": 0.6673361659049988, "learning_rate": 7.415745347892565e-05, "loss": 3.1233, "step": 192950 }, { "epoch": 1.7043748564969357, "grad_norm": 2.611398935317993, "learning_rate": 7.414528591150032e-05, "loss": 3.2982, "step": 193000 }, { "epoch": 1.7048164043872198, "grad_norm": 3.113255739212036, "learning_rate": 7.413311647905638e-05, "loss": 3.0646, "step": 193050 }, { "epoch": 1.705257952277504, "grad_norm": 3.388859987258911, "learning_rate": 7.412094518253385e-05, "loss": 3.4217, "step": 193100 }, { "epoch": 1.7056995001677882, "grad_norm": 3.73885440826416, "learning_rate": 7.410877202287282e-05, "loss": 3.0224, "step": 193150 }, { "epoch": 1.7061410480580723, "grad_norm": 2.581843137741089, "learning_rate": 7.409659700101362e-05, "loss": 3.3819, "step": 193200 }, { "epoch": 1.7065825959483565, "grad_norm": 3.6114909648895264, "learning_rate": 7.408442011789661e-05, "loss": 2.8756, "step": 193250 }, { "epoch": 1.7070241438386407, "grad_norm": 1.801611304283142, "learning_rate": 7.407224137446239e-05, "loss": 3.2742, "step": 193300 }, { "epoch": 1.7074656917289248, "grad_norm": 2.8728644847869873, "learning_rate": 7.406006077165165e-05, "loss": 3.0712, "step": 193350 }, { "epoch": 1.707907239619209, "grad_norm": 3.180121660232544, "learning_rate": 7.404787831040525e-05, "loss": 3.21, "step": 193400 }, { "epoch": 1.7083487875094931, "grad_norm": 3.0456016063690186, "learning_rate": 7.403569399166415e-05, "loss": 3.0706, "step": 193450 }, { "epoch": 1.7087903353997773, "grad_norm": 1.180357813835144, "learning_rate": 7.402350781636951e-05, "loss": 2.9831, "step": 193500 }, { "epoch": 1.7092318832900615, "grad_norm": 2.2708914279937744, "learning_rate": 7.40113197854626e-05, "loss": 3.3696, "step": 193550 }, { "epoch": 1.7096734311803459, "grad_norm": 3.7013955116271973, "learning_rate": 7.399912989988487e-05, "loss": 2.7012, "step": 193600 }, { "epoch": 1.71011497907063, "grad_norm": 1.7447788715362549, "learning_rate": 7.398693816057783e-05, "loss": 3.5764, "step": 193650 }, { "epoch": 1.7105565269609142, "grad_norm": 1.3203649520874023, "learning_rate": 7.397474456848324e-05, "loss": 3.1377, "step": 193700 }, { "epoch": 1.7109980748511984, "grad_norm": 3.331073045730591, "learning_rate": 7.396254912454292e-05, "loss": 3.1011, "step": 193750 }, { "epoch": 1.7114396227414825, "grad_norm": 2.029482126235962, "learning_rate": 7.395035182969889e-05, "loss": 3.3574, "step": 193800 }, { "epoch": 1.7118811706317667, "grad_norm": 1.7867449522018433, "learning_rate": 7.393815268489328e-05, "loss": 3.1734, "step": 193850 }, { "epoch": 1.7123227185220509, "grad_norm": 2.170999526977539, "learning_rate": 7.392595169106834e-05, "loss": 3.3209, "step": 193900 }, { "epoch": 1.712764266412335, "grad_norm": 3.1686041355133057, "learning_rate": 7.391374884916654e-05, "loss": 3.1962, "step": 193950 }, { "epoch": 1.7132058143026194, "grad_norm": 3.2382314205169678, "learning_rate": 7.390154416013042e-05, "loss": 3.3819, "step": 194000 }, { "epoch": 1.7136473621929036, "grad_norm": 2.528313398361206, "learning_rate": 7.388933762490268e-05, "loss": 3.1662, "step": 194050 }, { "epoch": 1.7140889100831878, "grad_norm": 2.6687514781951904, "learning_rate": 7.38771292444262e-05, "loss": 2.9224, "step": 194100 }, { "epoch": 1.714530457973472, "grad_norm": 2.152829647064209, "learning_rate": 7.386491901964394e-05, "loss": 3.3078, "step": 194150 }, { "epoch": 1.714972005863756, "grad_norm": 3.414506673812866, "learning_rate": 7.385270695149906e-05, "loss": 3.1442, "step": 194200 }, { "epoch": 1.7154135537540403, "grad_norm": 1.722166895866394, "learning_rate": 7.384049304093485e-05, "loss": 3.3816, "step": 194250 }, { "epoch": 1.7158551016443244, "grad_norm": 3.0730419158935547, "learning_rate": 7.382827728889468e-05, "loss": 3.1881, "step": 194300 }, { "epoch": 1.7162966495346086, "grad_norm": 0.8389607667922974, "learning_rate": 7.381605969632218e-05, "loss": 3.1474, "step": 194350 }, { "epoch": 1.7167381974248928, "grad_norm": 1.5451440811157227, "learning_rate": 7.380384026416102e-05, "loss": 2.7997, "step": 194400 }, { "epoch": 1.717179745315177, "grad_norm": 1.2447178363800049, "learning_rate": 7.379161899335504e-05, "loss": 3.1987, "step": 194450 }, { "epoch": 1.717621293205461, "grad_norm": 1.0391019582748413, "learning_rate": 7.377939588484823e-05, "loss": 3.5185, "step": 194500 }, { "epoch": 1.7180628410957453, "grad_norm": 1.7594729661941528, "learning_rate": 7.376717093958477e-05, "loss": 3.1056, "step": 194550 }, { "epoch": 1.7185043889860294, "grad_norm": 4.89455509185791, "learning_rate": 7.375494415850885e-05, "loss": 3.3591, "step": 194600 }, { "epoch": 1.7189459368763136, "grad_norm": 1.501932144165039, "learning_rate": 7.374271554256495e-05, "loss": 3.4794, "step": 194650 }, { "epoch": 1.7193874847665978, "grad_norm": 3.946047067642212, "learning_rate": 7.373048509269759e-05, "loss": 2.8626, "step": 194700 }, { "epoch": 1.719829032656882, "grad_norm": 3.7815029621124268, "learning_rate": 7.371849747346553e-05, "loss": 3.3103, "step": 194750 }, { "epoch": 1.720270580547166, "grad_norm": 1.0685265064239502, "learning_rate": 7.370626339521693e-05, "loss": 3.1576, "step": 194800 }, { "epoch": 1.7207121284374503, "grad_norm": 2.738694190979004, "learning_rate": 7.369402748586054e-05, "loss": 3.3068, "step": 194850 }, { "epoch": 1.7211536763277344, "grad_norm": 2.3966355323791504, "learning_rate": 7.368178974634143e-05, "loss": 3.4519, "step": 194900 }, { "epoch": 1.7215952242180186, "grad_norm": 2.5047898292541504, "learning_rate": 7.366955017760488e-05, "loss": 3.3452, "step": 194950 }, { "epoch": 1.7220367721083027, "grad_norm": 1.0891519784927368, "learning_rate": 7.36573087805963e-05, "loss": 3.4548, "step": 195000 }, { "epoch": 1.7220367721083027, "eval_asr_loss": 0.9187432370274967, "eval_loss": 2.88369083404541, "eval_runtime": 20.6655, "eval_samples_per_second": 37.163, "eval_steps_per_second": 9.291, "eval_tts_loss": 5.93669020019495, "step": 195000 }, { "epoch": 1.722478319998587, "grad_norm": 2.0451600551605225, "learning_rate": 7.364506555626122e-05, "loss": 2.8664, "step": 195050 }, { "epoch": 1.722919867888871, "grad_norm": 2.9029061794281006, "learning_rate": 7.363282050554532e-05, "loss": 3.5179, "step": 195100 }, { "epoch": 1.7233614157791552, "grad_norm": 2.5034167766571045, "learning_rate": 7.362057362939445e-05, "loss": 3.0646, "step": 195150 }, { "epoch": 1.7238029636694396, "grad_norm": 3.19059681892395, "learning_rate": 7.360832492875455e-05, "loss": 2.7585, "step": 195200 }, { "epoch": 1.7242445115597238, "grad_norm": 0.7491962313652039, "learning_rate": 7.359607440457176e-05, "loss": 3.0848, "step": 195250 }, { "epoch": 1.724686059450008, "grad_norm": 1.9276366233825684, "learning_rate": 7.358382205779231e-05, "loss": 3.3159, "step": 195300 }, { "epoch": 1.7251276073402921, "grad_norm": 4.427799224853516, "learning_rate": 7.357156788936257e-05, "loss": 2.8902, "step": 195350 }, { "epoch": 1.7255691552305763, "grad_norm": 6.118712902069092, "learning_rate": 7.35593119002291e-05, "loss": 3.2164, "step": 195400 }, { "epoch": 1.7260107031208605, "grad_norm": 2.9821903705596924, "learning_rate": 7.354705409133858e-05, "loss": 3.344, "step": 195450 }, { "epoch": 1.7264522510111446, "grad_norm": 3.2221412658691406, "learning_rate": 7.353479446363778e-05, "loss": 3.3174, "step": 195500 }, { "epoch": 1.7268937989014288, "grad_norm": 2.3944759368896484, "learning_rate": 7.35225330180737e-05, "loss": 3.2564, "step": 195550 }, { "epoch": 1.7273353467917132, "grad_norm": 1.8751672506332397, "learning_rate": 7.351026975559338e-05, "loss": 3.0801, "step": 195600 }, { "epoch": 1.7277768946819974, "grad_norm": 1.0560152530670166, "learning_rate": 7.349800467714412e-05, "loss": 3.3012, "step": 195650 }, { "epoch": 1.7282184425722815, "grad_norm": 2.2292487621307373, "learning_rate": 7.348573778367323e-05, "loss": 3.2392, "step": 195700 }, { "epoch": 1.7286599904625657, "grad_norm": 2.8452000617980957, "learning_rate": 7.347346907612825e-05, "loss": 3.59, "step": 195750 }, { "epoch": 1.7291015383528499, "grad_norm": 4.416501998901367, "learning_rate": 7.346119855545685e-05, "loss": 3.4565, "step": 195800 }, { "epoch": 1.729543086243134, "grad_norm": 4.022757053375244, "learning_rate": 7.344892622260678e-05, "loss": 3.414, "step": 195850 }, { "epoch": 1.7299846341334182, "grad_norm": 1.0071609020233154, "learning_rate": 7.343665207852603e-05, "loss": 2.9478, "step": 195900 }, { "epoch": 1.7304261820237024, "grad_norm": 0.6758058667182922, "learning_rate": 7.342437612416263e-05, "loss": 3.2096, "step": 195950 }, { "epoch": 1.7308677299139865, "grad_norm": 2.8905625343322754, "learning_rate": 7.34120983604648e-05, "loss": 3.07, "step": 196000 }, { "epoch": 1.7313092778042707, "grad_norm": 2.475128173828125, "learning_rate": 7.339981878838092e-05, "loss": 3.4404, "step": 196050 }, { "epoch": 1.7317508256945549, "grad_norm": 0.9189386963844299, "learning_rate": 7.338753740885945e-05, "loss": 2.84, "step": 196100 }, { "epoch": 1.732192373584839, "grad_norm": 1.5804566144943237, "learning_rate": 7.337525422284904e-05, "loss": 3.3711, "step": 196150 }, { "epoch": 1.7326339214751232, "grad_norm": 1.715122938156128, "learning_rate": 7.336296923129844e-05, "loss": 3.3647, "step": 196200 }, { "epoch": 1.7330754693654074, "grad_norm": 1.8856436014175415, "learning_rate": 7.33506824351566e-05, "loss": 3.5442, "step": 196250 }, { "epoch": 1.7335170172556915, "grad_norm": 2.374765396118164, "learning_rate": 7.333839383537254e-05, "loss": 3.2726, "step": 196300 }, { "epoch": 1.7339585651459757, "grad_norm": 2.1815669536590576, "learning_rate": 7.332610343289545e-05, "loss": 3.4051, "step": 196350 }, { "epoch": 1.7344001130362598, "grad_norm": 2.3607351779937744, "learning_rate": 7.331381122867468e-05, "loss": 3.063, "step": 196400 }, { "epoch": 1.734841660926544, "grad_norm": 0.8316501379013062, "learning_rate": 7.330151722365972e-05, "loss": 2.9158, "step": 196450 }, { "epoch": 1.7352832088168282, "grad_norm": 2.3766934871673584, "learning_rate": 7.32892214188001e-05, "loss": 2.7718, "step": 196500 }, { "epoch": 1.7357247567071123, "grad_norm": 2.467963695526123, "learning_rate": 7.327692381504563e-05, "loss": 3.1836, "step": 196550 }, { "epoch": 1.7361663045973965, "grad_norm": 3.188460111618042, "learning_rate": 7.326462441334618e-05, "loss": 3.3165, "step": 196600 }, { "epoch": 1.7366078524876807, "grad_norm": 1.6318292617797852, "learning_rate": 7.325256925623005e-05, "loss": 3.717, "step": 196650 }, { "epoch": 1.7370494003779648, "grad_norm": 3.8458199501037598, "learning_rate": 7.324026629740243e-05, "loss": 3.4363, "step": 196700 }, { "epoch": 1.737490948268249, "grad_norm": 4.6386895179748535, "learning_rate": 7.322796154346132e-05, "loss": 3.296, "step": 196750 }, { "epoch": 1.7379324961585334, "grad_norm": 2.404712438583374, "learning_rate": 7.321565499535713e-05, "loss": 3.5477, "step": 196800 }, { "epoch": 1.7383740440488176, "grad_norm": 5.802834510803223, "learning_rate": 7.320334665404046e-05, "loss": 2.9515, "step": 196850 }, { "epoch": 1.7388155919391017, "grad_norm": 1.9599117040634155, "learning_rate": 7.319103652046201e-05, "loss": 3.067, "step": 196900 }, { "epoch": 1.739257139829386, "grad_norm": 3.9484918117523193, "learning_rate": 7.317872459557267e-05, "loss": 2.9862, "step": 196950 }, { "epoch": 1.73969868771967, "grad_norm": 4.071539878845215, "learning_rate": 7.316641088032339e-05, "loss": 2.9783, "step": 197000 }, { "epoch": 1.7401402356099542, "grad_norm": 2.045924663543701, "learning_rate": 7.315409537566532e-05, "loss": 3.0458, "step": 197050 }, { "epoch": 1.7405817835002384, "grad_norm": 3.734661817550659, "learning_rate": 7.314177808254971e-05, "loss": 3.3027, "step": 197100 }, { "epoch": 1.7410233313905226, "grad_norm": 2.9852235317230225, "learning_rate": 7.312945900192798e-05, "loss": 3.608, "step": 197150 }, { "epoch": 1.741464879280807, "grad_norm": 2.508160352706909, "learning_rate": 7.311713813475168e-05, "loss": 3.5748, "step": 197200 }, { "epoch": 1.7419064271710911, "grad_norm": 5.302500247955322, "learning_rate": 7.310481548197249e-05, "loss": 3.5315, "step": 197250 }, { "epoch": 1.7423479750613753, "grad_norm": 1.832080602645874, "learning_rate": 7.309249104454221e-05, "loss": 3.175, "step": 197300 }, { "epoch": 1.7427895229516595, "grad_norm": 4.390317916870117, "learning_rate": 7.308016482341285e-05, "loss": 2.9957, "step": 197350 }, { "epoch": 1.7432310708419436, "grad_norm": 1.2284395694732666, "learning_rate": 7.306783681953645e-05, "loss": 3.1387, "step": 197400 }, { "epoch": 1.7436726187322278, "grad_norm": 4.071477890014648, "learning_rate": 7.305550703386525e-05, "loss": 3.0247, "step": 197450 }, { "epoch": 1.744114166622512, "grad_norm": 4.2401227951049805, "learning_rate": 7.304317546735167e-05, "loss": 3.0524, "step": 197500 }, { "epoch": 1.7445557145127961, "grad_norm": 1.551368236541748, "learning_rate": 7.303084212094817e-05, "loss": 3.0328, "step": 197550 }, { "epoch": 1.7449972624030803, "grad_norm": 3.074233055114746, "learning_rate": 7.30185069956074e-05, "loss": 3.1914, "step": 197600 }, { "epoch": 1.7454388102933645, "grad_norm": 1.8998547792434692, "learning_rate": 7.300617009228215e-05, "loss": 3.3625, "step": 197650 }, { "epoch": 1.7458803581836486, "grad_norm": 1.0645414590835571, "learning_rate": 7.299383141192539e-05, "loss": 3.0895, "step": 197700 }, { "epoch": 1.7463219060739328, "grad_norm": 1.3525128364562988, "learning_rate": 7.29814909554901e-05, "loss": 3.1057, "step": 197750 }, { "epoch": 1.746763453964217, "grad_norm": 1.9818834066390991, "learning_rate": 7.296914872392952e-05, "loss": 3.3467, "step": 197800 }, { "epoch": 1.7472050018545011, "grad_norm": 2.3736205101013184, "learning_rate": 7.295680471819697e-05, "loss": 3.0871, "step": 197850 }, { "epoch": 1.7476465497447853, "grad_norm": 1.744696021080017, "learning_rate": 7.294445893924593e-05, "loss": 3.3649, "step": 197900 }, { "epoch": 1.7480880976350694, "grad_norm": 2.553920269012451, "learning_rate": 7.293211138803e-05, "loss": 2.8486, "step": 197950 }, { "epoch": 1.7485296455253536, "grad_norm": 2.8582518100738525, "learning_rate": 7.291976206550292e-05, "loss": 2.8678, "step": 198000 }, { "epoch": 1.7485296455253536, "eval_asr_loss": 0.9277813797653791, "eval_loss": 2.883857488632202, "eval_runtime": 20.4712, "eval_samples_per_second": 37.516, "eval_steps_per_second": 9.379, "eval_tts_loss": 5.936648276907347, "step": 198000 }, { "epoch": 1.7489711934156378, "grad_norm": 2.297837495803833, "learning_rate": 7.290741097261856e-05, "loss": 3.0111, "step": 198050 }, { "epoch": 1.749412741305922, "grad_norm": 3.550356388092041, "learning_rate": 7.289505811033099e-05, "loss": 3.5257, "step": 198100 }, { "epoch": 1.7498542891962061, "grad_norm": 4.377917289733887, "learning_rate": 7.288270347959429e-05, "loss": 3.2884, "step": 198150 }, { "epoch": 1.7502958370864903, "grad_norm": 1.634263038635254, "learning_rate": 7.28703470813628e-05, "loss": 2.6872, "step": 198200 }, { "epoch": 1.7507373849767744, "grad_norm": 4.123717308044434, "learning_rate": 7.285798891659092e-05, "loss": 3.2004, "step": 198250 }, { "epoch": 1.7511789328670586, "grad_norm": 1.9348206520080566, "learning_rate": 7.284562898623325e-05, "loss": 3.2221, "step": 198300 }, { "epoch": 1.7516204807573428, "grad_norm": 3.293262004852295, "learning_rate": 7.283326729124446e-05, "loss": 3.2032, "step": 198350 }, { "epoch": 1.7520620286476272, "grad_norm": 2.118044376373291, "learning_rate": 7.282090383257937e-05, "loss": 3.2974, "step": 198400 }, { "epoch": 1.7525035765379113, "grad_norm": 1.9033870697021484, "learning_rate": 7.280853861119298e-05, "loss": 3.4477, "step": 198450 }, { "epoch": 1.7529451244281955, "grad_norm": 3.468416452407837, "learning_rate": 7.279617162804042e-05, "loss": 2.9804, "step": 198500 }, { "epoch": 1.7533866723184797, "grad_norm": 3.426452159881592, "learning_rate": 7.278380288407689e-05, "loss": 3.0201, "step": 198550 }, { "epoch": 1.7538282202087638, "grad_norm": 4.912716865539551, "learning_rate": 7.277143238025779e-05, "loss": 3.0869, "step": 198600 }, { "epoch": 1.754269768099048, "grad_norm": 1.0470197200775146, "learning_rate": 7.275906011753863e-05, "loss": 3.5107, "step": 198650 }, { "epoch": 1.7547113159893322, "grad_norm": 2.836777687072754, "learning_rate": 7.274668609687507e-05, "loss": 3.0968, "step": 198700 }, { "epoch": 1.7551528638796163, "grad_norm": 3.8205599784851074, "learning_rate": 7.273431031922291e-05, "loss": 3.0864, "step": 198750 }, { "epoch": 1.7555944117699007, "grad_norm": 5.9745893478393555, "learning_rate": 7.272193278553803e-05, "loss": 3.2796, "step": 198800 }, { "epoch": 1.7560359596601849, "grad_norm": 1.6742478609085083, "learning_rate": 7.270955349677654e-05, "loss": 3.5679, "step": 198850 }, { "epoch": 1.756477507550469, "grad_norm": 1.8220934867858887, "learning_rate": 7.269717245389461e-05, "loss": 2.9193, "step": 198900 }, { "epoch": 1.7569190554407532, "grad_norm": 0.8500947952270508, "learning_rate": 7.268478965784857e-05, "loss": 2.9982, "step": 198950 }, { "epoch": 1.7573606033310374, "grad_norm": 3.3522162437438965, "learning_rate": 7.267240510959489e-05, "loss": 3.1133, "step": 199000 }, { "epoch": 1.7578021512213216, "grad_norm": 1.2531402111053467, "learning_rate": 7.266001881009018e-05, "loss": 2.583, "step": 199050 }, { "epoch": 1.7582436991116057, "grad_norm": 3.770021677017212, "learning_rate": 7.264763076029116e-05, "loss": 3.4063, "step": 199100 }, { "epoch": 1.7586852470018899, "grad_norm": 2.973428249359131, "learning_rate": 7.263524096115471e-05, "loss": 3.1737, "step": 199150 }, { "epoch": 1.759126794892174, "grad_norm": 1.8131953477859497, "learning_rate": 7.262284941363784e-05, "loss": 3.3857, "step": 199200 }, { "epoch": 1.7595683427824582, "grad_norm": 2.9858899116516113, "learning_rate": 7.261045611869767e-05, "loss": 3.3893, "step": 199250 }, { "epoch": 1.7600098906727424, "grad_norm": 3.661966323852539, "learning_rate": 7.259806107729155e-05, "loss": 3.2071, "step": 199300 }, { "epoch": 1.7604514385630265, "grad_norm": 3.4197046756744385, "learning_rate": 7.258566429037679e-05, "loss": 3.0726, "step": 199350 }, { "epoch": 1.7608929864533107, "grad_norm": 3.3999383449554443, "learning_rate": 7.2573265758911e-05, "loss": 3.6426, "step": 199400 }, { "epoch": 1.7613345343435949, "grad_norm": 1.2242519855499268, "learning_rate": 7.256086548385183e-05, "loss": 2.9793, "step": 199450 }, { "epoch": 1.761776082233879, "grad_norm": 5.0131964683532715, "learning_rate": 7.254846346615713e-05, "loss": 3.4929, "step": 199500 }, { "epoch": 1.7622176301241632, "grad_norm": 3.7067580223083496, "learning_rate": 7.253605970678483e-05, "loss": 3.1712, "step": 199550 }, { "epoch": 1.7626591780144474, "grad_norm": 1.0138520002365112, "learning_rate": 7.2523654206693e-05, "loss": 3.5418, "step": 199600 }, { "epoch": 1.7631007259047315, "grad_norm": 0.7684496641159058, "learning_rate": 7.251124696683989e-05, "loss": 2.7712, "step": 199650 }, { "epoch": 1.7635422737950157, "grad_norm": 3.878993272781372, "learning_rate": 7.249883798818386e-05, "loss": 3.6168, "step": 199700 }, { "epoch": 1.7639838216852999, "grad_norm": 3.2014262676239014, "learning_rate": 7.248642727168335e-05, "loss": 3.2813, "step": 199750 }, { "epoch": 1.764425369575584, "grad_norm": 3.0954771041870117, "learning_rate": 7.247401481829703e-05, "loss": 3.6834, "step": 199800 }, { "epoch": 1.7648669174658682, "grad_norm": 1.132131814956665, "learning_rate": 7.246160062898364e-05, "loss": 2.5174, "step": 199850 }, { "epoch": 1.7653084653561524, "grad_norm": 2.3023476600646973, "learning_rate": 7.244918470470206e-05, "loss": 3.2556, "step": 199900 }, { "epoch": 1.7657500132464365, "grad_norm": 4.4456586837768555, "learning_rate": 7.243676704641134e-05, "loss": 3.2855, "step": 199950 }, { "epoch": 1.766191561136721, "grad_norm": 3.184326171875, "learning_rate": 7.242434765507061e-05, "loss": 3.531, "step": 200000 }, { "epoch": 1.766633109027005, "grad_norm": 1.8845802545547485, "learning_rate": 7.241192653163918e-05, "loss": 2.9489, "step": 200050 }, { "epoch": 1.7670746569172893, "grad_norm": 2.8838372230529785, "learning_rate": 7.239950367707647e-05, "loss": 2.9885, "step": 200100 }, { "epoch": 1.7675162048075734, "grad_norm": 0.9097310304641724, "learning_rate": 7.238707909234206e-05, "loss": 3.262, "step": 200150 }, { "epoch": 1.7679577526978576, "grad_norm": 2.829453229904175, "learning_rate": 7.237465277839561e-05, "loss": 3.2191, "step": 200200 }, { "epoch": 1.7683993005881418, "grad_norm": 1.0789093971252441, "learning_rate": 7.236222473619697e-05, "loss": 3.1551, "step": 200250 }, { "epoch": 1.768840848478426, "grad_norm": 1.0863782167434692, "learning_rate": 7.234979496670609e-05, "loss": 2.9866, "step": 200300 }, { "epoch": 1.76928239636871, "grad_norm": 1.4690102338790894, "learning_rate": 7.233736347088307e-05, "loss": 2.9813, "step": 200350 }, { "epoch": 1.7697239442589945, "grad_norm": 2.871446132659912, "learning_rate": 7.232493024968815e-05, "loss": 3.1252, "step": 200400 }, { "epoch": 1.7701654921492787, "grad_norm": 1.6246694326400757, "learning_rate": 7.231249530408168e-05, "loss": 3.0882, "step": 200450 }, { "epoch": 1.7706070400395628, "grad_norm": 6.12792444229126, "learning_rate": 7.230005863502413e-05, "loss": 2.9357, "step": 200500 }, { "epoch": 1.771048587929847, "grad_norm": 2.492288589477539, "learning_rate": 7.228762024347616e-05, "loss": 3.2043, "step": 200550 }, { "epoch": 1.7714901358201312, "grad_norm": 1.2008299827575684, "learning_rate": 7.227518013039851e-05, "loss": 2.7181, "step": 200600 }, { "epoch": 1.7719316837104153, "grad_norm": 3.010150909423828, "learning_rate": 7.226273829675211e-05, "loss": 3.1411, "step": 200650 }, { "epoch": 1.7723732316006995, "grad_norm": 2.3677845001220703, "learning_rate": 7.225029474349794e-05, "loss": 3.4423, "step": 200700 }, { "epoch": 1.7728147794909836, "grad_norm": 4.701264381408691, "learning_rate": 7.223784947159717e-05, "loss": 3.2663, "step": 200750 }, { "epoch": 1.7732563273812678, "grad_norm": 0.874535322189331, "learning_rate": 7.222540248201112e-05, "loss": 3.0042, "step": 200800 }, { "epoch": 1.773697875271552, "grad_norm": 2.3065900802612305, "learning_rate": 7.221295377570119e-05, "loss": 3.1783, "step": 200850 }, { "epoch": 1.7741394231618361, "grad_norm": 3.3488705158233643, "learning_rate": 7.220050335362894e-05, "loss": 3.3362, "step": 200900 }, { "epoch": 1.7745809710521203, "grad_norm": 1.8239176273345947, "learning_rate": 7.218805121675605e-05, "loss": 3.3862, "step": 200950 }, { "epoch": 1.7750225189424045, "grad_norm": 2.1905157566070557, "learning_rate": 7.217559736604435e-05, "loss": 3.5234, "step": 201000 }, { "epoch": 1.7750225189424045, "eval_asr_loss": 0.910745237708085, "eval_loss": 2.8749611377716064, "eval_runtime": 20.3917, "eval_samples_per_second": 37.662, "eval_steps_per_second": 9.416, "eval_tts_loss": 5.940568454043894, "step": 201000 }, { "epoch": 1.7754640668326886, "grad_norm": 1.463396430015564, "learning_rate": 7.216314180245581e-05, "loss": 3.2108, "step": 201050 }, { "epoch": 1.7759056147229728, "grad_norm": 2.2169008255004883, "learning_rate": 7.21506845269525e-05, "loss": 3.3798, "step": 201100 }, { "epoch": 1.776347162613257, "grad_norm": 2.9501075744628906, "learning_rate": 7.213822554049664e-05, "loss": 3.1308, "step": 201150 }, { "epoch": 1.7767887105035411, "grad_norm": 2.21448016166687, "learning_rate": 7.212576484405058e-05, "loss": 3.1955, "step": 201200 }, { "epoch": 1.7772302583938253, "grad_norm": 2.7748055458068848, "learning_rate": 7.211330243857682e-05, "loss": 3.0091, "step": 201250 }, { "epoch": 1.7776718062841095, "grad_norm": 2.3246281147003174, "learning_rate": 7.210083832503796e-05, "loss": 3.2605, "step": 201300 }, { "epoch": 1.7781133541743936, "grad_norm": 2.713412046432495, "learning_rate": 7.208837250439675e-05, "loss": 3.0552, "step": 201350 }, { "epoch": 1.7785549020646778, "grad_norm": 1.687602162361145, "learning_rate": 7.207590497761604e-05, "loss": 3.2309, "step": 201400 }, { "epoch": 1.778996449954962, "grad_norm": 6.353885650634766, "learning_rate": 7.20634357456589e-05, "loss": 3.3475, "step": 201450 }, { "epoch": 1.7794379978452461, "grad_norm": 2.1135149002075195, "learning_rate": 7.205096480948844e-05, "loss": 2.8696, "step": 201500 }, { "epoch": 1.7798795457355303, "grad_norm": 2.8203511238098145, "learning_rate": 7.203849217006792e-05, "loss": 2.8179, "step": 201550 }, { "epoch": 1.7803210936258147, "grad_norm": 3.2417795658111572, "learning_rate": 7.202601782836075e-05, "loss": 3.2816, "step": 201600 }, { "epoch": 1.7807626415160989, "grad_norm": 3.100700855255127, "learning_rate": 7.201354178533051e-05, "loss": 3.1216, "step": 201650 }, { "epoch": 1.781204189406383, "grad_norm": 3.2376022338867188, "learning_rate": 7.200106404194082e-05, "loss": 3.1, "step": 201700 }, { "epoch": 1.7816457372966672, "grad_norm": 2.8501505851745605, "learning_rate": 7.198858459915549e-05, "loss": 2.8905, "step": 201750 }, { "epoch": 1.7820872851869514, "grad_norm": 3.309192180633545, "learning_rate": 7.197610345793848e-05, "loss": 2.9565, "step": 201800 }, { "epoch": 1.7825288330772355, "grad_norm": 2.582115888595581, "learning_rate": 7.196362061925381e-05, "loss": 3.7028, "step": 201850 }, { "epoch": 1.7829703809675197, "grad_norm": 2.774749517440796, "learning_rate": 7.195113608406569e-05, "loss": 3.2302, "step": 201900 }, { "epoch": 1.7834119288578039, "grad_norm": 1.1029777526855469, "learning_rate": 7.193864985333847e-05, "loss": 3.3933, "step": 201950 }, { "epoch": 1.7838534767480883, "grad_norm": 3.4153521060943604, "learning_rate": 7.192616192803656e-05, "loss": 3.4729, "step": 202000 }, { "epoch": 1.7842950246383724, "grad_norm": 4.102583885192871, "learning_rate": 7.191367230912459e-05, "loss": 2.8795, "step": 202050 }, { "epoch": 1.7847365725286566, "grad_norm": 0.49508291482925415, "learning_rate": 7.190118099756724e-05, "loss": 3.681, "step": 202100 }, { "epoch": 1.7851781204189408, "grad_norm": 1.6070566177368164, "learning_rate": 7.188893787096637e-05, "loss": 3.4495, "step": 202150 }, { "epoch": 1.785619668309225, "grad_norm": 1.8097330331802368, "learning_rate": 7.187644321081782e-05, "loss": 3.138, "step": 202200 }, { "epoch": 1.786061216199509, "grad_norm": 1.518096923828125, "learning_rate": 7.186394686089954e-05, "loss": 2.7297, "step": 202250 }, { "epoch": 1.7865027640897932, "grad_norm": 3.6704065799713135, "learning_rate": 7.185144882217678e-05, "loss": 2.875, "step": 202300 }, { "epoch": 1.7869443119800774, "grad_norm": 1.533138632774353, "learning_rate": 7.18389490956149e-05, "loss": 2.9326, "step": 202350 }, { "epoch": 1.7873858598703616, "grad_norm": 2.0090863704681396, "learning_rate": 7.182644768217936e-05, "loss": 3.4076, "step": 202400 }, { "epoch": 1.7878274077606457, "grad_norm": 3.815380573272705, "learning_rate": 7.181394458283583e-05, "loss": 3.3829, "step": 202450 }, { "epoch": 1.78826895565093, "grad_norm": 2.4792747497558594, "learning_rate": 7.180143979855008e-05, "loss": 3.314, "step": 202500 }, { "epoch": 1.788710503541214, "grad_norm": 1.389562964439392, "learning_rate": 7.178893333028794e-05, "loss": 3.0, "step": 202550 }, { "epoch": 1.7891520514314982, "grad_norm": 1.2650171518325806, "learning_rate": 7.177642517901549e-05, "loss": 3.5354, "step": 202600 }, { "epoch": 1.7895935993217824, "grad_norm": 3.26472806930542, "learning_rate": 7.176391534569885e-05, "loss": 3.3161, "step": 202650 }, { "epoch": 1.7900351472120666, "grad_norm": 2.080465793609619, "learning_rate": 7.17514038313043e-05, "loss": 2.8847, "step": 202700 }, { "epoch": 1.7904766951023507, "grad_norm": 2.661543369293213, "learning_rate": 7.173889063679826e-05, "loss": 3.2848, "step": 202750 }, { "epoch": 1.790918242992635, "grad_norm": 1.5421067476272583, "learning_rate": 7.172637576314724e-05, "loss": 3.3914, "step": 202800 }, { "epoch": 1.791359790882919, "grad_norm": 4.784469127655029, "learning_rate": 7.171385921131793e-05, "loss": 3.0531, "step": 202850 }, { "epoch": 1.7918013387732032, "grad_norm": 5.333433151245117, "learning_rate": 7.170134098227713e-05, "loss": 3.3758, "step": 202900 }, { "epoch": 1.7922428866634874, "grad_norm": 1.1900759935379028, "learning_rate": 7.168882107699178e-05, "loss": 3.2778, "step": 202950 }, { "epoch": 1.7926844345537716, "grad_norm": 4.2666096687316895, "learning_rate": 7.167629949642889e-05, "loss": 3.0026, "step": 203000 }, { "epoch": 1.7931259824440557, "grad_norm": 2.9323208332061768, "learning_rate": 7.166377624155567e-05, "loss": 3.0496, "step": 203050 }, { "epoch": 1.79356753033434, "grad_norm": 4.785465717315674, "learning_rate": 7.165125131333946e-05, "loss": 3.0047, "step": 203100 }, { "epoch": 1.794009078224624, "grad_norm": 2.6500496864318848, "learning_rate": 7.163872471274768e-05, "loss": 2.8693, "step": 203150 }, { "epoch": 1.7944506261149085, "grad_norm": 4.793989181518555, "learning_rate": 7.16261964407479e-05, "loss": 3.3363, "step": 203200 }, { "epoch": 1.7948921740051926, "grad_norm": 4.713131904602051, "learning_rate": 7.161366649830783e-05, "loss": 3.1709, "step": 203250 }, { "epoch": 1.7953337218954768, "grad_norm": 1.1408799886703491, "learning_rate": 7.16011348863953e-05, "loss": 3.091, "step": 203300 }, { "epoch": 1.795775269785761, "grad_norm": 4.1913676261901855, "learning_rate": 7.158860160597828e-05, "loss": 2.9463, "step": 203350 }, { "epoch": 1.7962168176760451, "grad_norm": 3.7222936153411865, "learning_rate": 7.157606665802484e-05, "loss": 3.5081, "step": 203400 }, { "epoch": 1.7966583655663293, "grad_norm": 2.148895740509033, "learning_rate": 7.156353004350321e-05, "loss": 2.9903, "step": 203450 }, { "epoch": 1.7970999134566135, "grad_norm": 1.5505282878875732, "learning_rate": 7.155099176338176e-05, "loss": 3.1496, "step": 203500 }, { "epoch": 1.7975414613468979, "grad_norm": 3.622462511062622, "learning_rate": 7.153845181862893e-05, "loss": 2.7559, "step": 203550 }, { "epoch": 1.797983009237182, "grad_norm": 4.700122356414795, "learning_rate": 7.152591021021332e-05, "loss": 3.12, "step": 203600 }, { "epoch": 1.7984245571274662, "grad_norm": 1.3534070253372192, "learning_rate": 7.15133669391037e-05, "loss": 2.584, "step": 203650 }, { "epoch": 1.7988661050177503, "grad_norm": 2.837641954421997, "learning_rate": 7.150082200626891e-05, "loss": 3.6518, "step": 203700 }, { "epoch": 1.7993076529080345, "grad_norm": 3.3624727725982666, "learning_rate": 7.148827541267794e-05, "loss": 3.2949, "step": 203750 }, { "epoch": 1.7997492007983187, "grad_norm": 2.1359448432922363, "learning_rate": 7.147572715929991e-05, "loss": 3.3417, "step": 203800 }, { "epoch": 1.8001907486886028, "grad_norm": 2.3549392223358154, "learning_rate": 7.146317724710407e-05, "loss": 2.98, "step": 203850 }, { "epoch": 1.800632296578887, "grad_norm": 2.2345235347747803, "learning_rate": 7.145062567705979e-05, "loss": 3.5184, "step": 203900 }, { "epoch": 1.8010738444691712, "grad_norm": 2.326660394668579, "learning_rate": 7.143807245013656e-05, "loss": 2.8502, "step": 203950 }, { "epoch": 1.8015153923594553, "grad_norm": 2.0236425399780273, "learning_rate": 7.142551756730403e-05, "loss": 3.163, "step": 204000 }, { "epoch": 1.8015153923594553, "eval_asr_loss": 0.9254957463829835, "eval_loss": 2.87661075592041, "eval_runtime": 20.4661, "eval_samples_per_second": 37.525, "eval_steps_per_second": 9.381, "eval_tts_loss": 5.962671826088398, "step": 204000 }, { "epoch": 1.8019569402497395, "grad_norm": 1.7293661832809448, "learning_rate": 7.141296102953195e-05, "loss": 2.9224, "step": 204050 }, { "epoch": 1.8023984881400237, "grad_norm": 1.66663658618927, "learning_rate": 7.140040283779021e-05, "loss": 3.1893, "step": 204100 }, { "epoch": 1.8028400360303078, "grad_norm": 2.202199935913086, "learning_rate": 7.138784299304882e-05, "loss": 3.2132, "step": 204150 }, { "epoch": 1.803281583920592, "grad_norm": 2.037201404571533, "learning_rate": 7.137528149627792e-05, "loss": 3.3609, "step": 204200 }, { "epoch": 1.8037231318108762, "grad_norm": 2.08586049079895, "learning_rate": 7.136271834844778e-05, "loss": 3.123, "step": 204250 }, { "epoch": 1.8041646797011603, "grad_norm": 2.966116189956665, "learning_rate": 7.135040486265179e-05, "loss": 2.9609, "step": 204300 }, { "epoch": 1.8046062275914445, "grad_norm": 1.4813175201416016, "learning_rate": 7.133783844858734e-05, "loss": 3.1627, "step": 204350 }, { "epoch": 1.8050477754817287, "grad_norm": 1.58802330493927, "learning_rate": 7.132527038635583e-05, "loss": 3.3023, "step": 204400 }, { "epoch": 1.8054893233720128, "grad_norm": 1.9538745880126953, "learning_rate": 7.1312700676928e-05, "loss": 3.1937, "step": 204450 }, { "epoch": 1.805930871262297, "grad_norm": 2.2432916164398193, "learning_rate": 7.130012932127477e-05, "loss": 3.3247, "step": 204500 }, { "epoch": 1.8063724191525812, "grad_norm": 3.8045809268951416, "learning_rate": 7.128755632036717e-05, "loss": 3.459, "step": 204550 }, { "epoch": 1.8068139670428653, "grad_norm": 4.374520301818848, "learning_rate": 7.127498167517636e-05, "loss": 3.141, "step": 204600 }, { "epoch": 1.8072555149331495, "grad_norm": 2.797487735748291, "learning_rate": 7.126240538667361e-05, "loss": 3.7096, "step": 204650 }, { "epoch": 1.8076970628234337, "grad_norm": 1.1959338188171387, "learning_rate": 7.124982745583037e-05, "loss": 3.4067, "step": 204700 }, { "epoch": 1.808138610713718, "grad_norm": 4.910750389099121, "learning_rate": 7.123724788361814e-05, "loss": 3.1069, "step": 204750 }, { "epoch": 1.8085801586040022, "grad_norm": 1.9608979225158691, "learning_rate": 7.12246666710086e-05, "loss": 3.6457, "step": 204800 }, { "epoch": 1.8090217064942864, "grad_norm": 4.217527866363525, "learning_rate": 7.121208381897353e-05, "loss": 3.1097, "step": 204850 }, { "epoch": 1.8094632543845706, "grad_norm": 1.2531399726867676, "learning_rate": 7.119949932848486e-05, "loss": 3.1505, "step": 204900 }, { "epoch": 1.8099048022748547, "grad_norm": 2.24678897857666, "learning_rate": 7.118691320051464e-05, "loss": 3.1959, "step": 204950 }, { "epoch": 1.810346350165139, "grad_norm": 1.1606642007827759, "learning_rate": 7.117432543603503e-05, "loss": 3.3106, "step": 205000 }, { "epoch": 1.810787898055423, "grad_norm": 6.087655067443848, "learning_rate": 7.116173603601834e-05, "loss": 3.097, "step": 205050 }, { "epoch": 1.8112294459457072, "grad_norm": 2.541350841522217, "learning_rate": 7.114914500143698e-05, "loss": 3.3576, "step": 205100 }, { "epoch": 1.8116709938359916, "grad_norm": 3.244662284851074, "learning_rate": 7.113655233326351e-05, "loss": 3.3987, "step": 205150 }, { "epoch": 1.8121125417262758, "grad_norm": 1.9999686479568481, "learning_rate": 7.11239580324706e-05, "loss": 3.0979, "step": 205200 }, { "epoch": 1.81255408961656, "grad_norm": 2.8434536457061768, "learning_rate": 7.111136210003106e-05, "loss": 3.2806, "step": 205250 }, { "epoch": 1.8129956375068441, "grad_norm": 1.383022665977478, "learning_rate": 7.109876453691781e-05, "loss": 2.9514, "step": 205300 }, { "epoch": 1.8134371853971283, "grad_norm": 4.14407205581665, "learning_rate": 7.108616534410394e-05, "loss": 3.2434, "step": 205350 }, { "epoch": 1.8138787332874124, "grad_norm": 4.073207855224609, "learning_rate": 7.107356452256258e-05, "loss": 3.2738, "step": 205400 }, { "epoch": 1.8143202811776966, "grad_norm": 2.9110472202301025, "learning_rate": 7.106096207326707e-05, "loss": 3.0951, "step": 205450 }, { "epoch": 1.8147618290679808, "grad_norm": 4.486078262329102, "learning_rate": 7.104835799719083e-05, "loss": 3.1462, "step": 205500 }, { "epoch": 1.815203376958265, "grad_norm": 2.024273157119751, "learning_rate": 7.103575229530742e-05, "loss": 3.2425, "step": 205550 }, { "epoch": 1.8156449248485491, "grad_norm": 1.897672414779663, "learning_rate": 7.102314496859052e-05, "loss": 3.2251, "step": 205600 }, { "epoch": 1.8160864727388333, "grad_norm": 3.416689395904541, "learning_rate": 7.101053601801392e-05, "loss": 2.9973, "step": 205650 }, { "epoch": 1.8165280206291174, "grad_norm": 1.4610952138900757, "learning_rate": 7.099792544455161e-05, "loss": 3.0895, "step": 205700 }, { "epoch": 1.8169695685194016, "grad_norm": 3.8125312328338623, "learning_rate": 7.09853132491776e-05, "loss": 2.9728, "step": 205750 }, { "epoch": 1.8174111164096858, "grad_norm": 2.8992254734039307, "learning_rate": 7.09726994328661e-05, "loss": 3.2576, "step": 205800 }, { "epoch": 1.81785266429997, "grad_norm": 3.9919283390045166, "learning_rate": 7.09600839965914e-05, "loss": 3.0434, "step": 205850 }, { "epoch": 1.8182942121902541, "grad_norm": 1.9868943691253662, "learning_rate": 7.094746694132795e-05, "loss": 3.1502, "step": 205900 }, { "epoch": 1.8187357600805383, "grad_norm": 3.0747931003570557, "learning_rate": 7.093484826805032e-05, "loss": 2.5582, "step": 205950 }, { "epoch": 1.8191773079708224, "grad_norm": 3.5087995529174805, "learning_rate": 7.092222797773318e-05, "loss": 3.1024, "step": 206000 }, { "epoch": 1.8196188558611066, "grad_norm": 2.8471686840057373, "learning_rate": 7.090960607135134e-05, "loss": 3.008, "step": 206050 }, { "epoch": 1.8200604037513908, "grad_norm": 3.2857754230499268, "learning_rate": 7.089698254987975e-05, "loss": 3.1496, "step": 206100 }, { "epoch": 1.820501951641675, "grad_norm": 3.396155834197998, "learning_rate": 7.088435741429344e-05, "loss": 2.8294, "step": 206150 }, { "epoch": 1.820943499531959, "grad_norm": 3.1792287826538086, "learning_rate": 7.087173066556762e-05, "loss": 3.3667, "step": 206200 }, { "epoch": 1.8213850474222433, "grad_norm": 2.4418413639068604, "learning_rate": 7.085910230467761e-05, "loss": 2.8149, "step": 206250 }, { "epoch": 1.8218265953125274, "grad_norm": 1.8641915321350098, "learning_rate": 7.08464723325988e-05, "loss": 2.9135, "step": 206300 }, { "epoch": 1.8222681432028118, "grad_norm": 4.3222880363464355, "learning_rate": 7.083384075030681e-05, "loss": 3.2363, "step": 206350 }, { "epoch": 1.822709691093096, "grad_norm": 2.6220529079437256, "learning_rate": 7.082120755877726e-05, "loss": 3.1948, "step": 206400 }, { "epoch": 1.8231512389833802, "grad_norm": 3.988168954849243, "learning_rate": 7.080857275898598e-05, "loss": 3.438, "step": 206450 }, { "epoch": 1.8235927868736643, "grad_norm": 2.376438856124878, "learning_rate": 7.079593635190892e-05, "loss": 3.5711, "step": 206500 }, { "epoch": 1.8240343347639485, "grad_norm": 5.018946170806885, "learning_rate": 7.078329833852212e-05, "loss": 2.6372, "step": 206550 }, { "epoch": 1.8244758826542327, "grad_norm": 1.6604958772659302, "learning_rate": 7.077065871980177e-05, "loss": 3.1432, "step": 206600 }, { "epoch": 1.8249174305445168, "grad_norm": 1.9766267538070679, "learning_rate": 7.075801749672415e-05, "loss": 3.1183, "step": 206650 }, { "epoch": 1.825358978434801, "grad_norm": 2.3561177253723145, "learning_rate": 7.074537467026569e-05, "loss": 3.378, "step": 206700 }, { "epoch": 1.8258005263250854, "grad_norm": 0.6396112442016602, "learning_rate": 7.073273024140298e-05, "loss": 3.4214, "step": 206750 }, { "epoch": 1.8262420742153695, "grad_norm": 3.366499662399292, "learning_rate": 7.072008421111265e-05, "loss": 2.9011, "step": 206800 }, { "epoch": 1.8266836221056537, "grad_norm": 2.9011270999908447, "learning_rate": 7.070743658037152e-05, "loss": 3.3191, "step": 206850 }, { "epoch": 1.8271251699959379, "grad_norm": 2.483954906463623, "learning_rate": 7.069504035042934e-05, "loss": 2.9523, "step": 206900 }, { "epoch": 1.827566717886222, "grad_norm": 3.9285833835601807, "learning_rate": 7.068238955367785e-05, "loss": 2.9662, "step": 206950 }, { "epoch": 1.8280082657765062, "grad_norm": 2.934602975845337, "learning_rate": 7.066973715938715e-05, "loss": 3.7216, "step": 207000 }, { "epoch": 1.8280082657765062, "eval_asr_loss": 0.9368323913638121, "eval_loss": 2.8737800121307373, "eval_runtime": 20.803, "eval_samples_per_second": 36.918, "eval_steps_per_second": 9.229, "eval_tts_loss": 5.962727203036098, "step": 207000 }, { "epoch": 1.8284498136667904, "grad_norm": 3.736560821533203, "learning_rate": 7.065708316853454e-05, "loss": 2.9009, "step": 207050 }, { "epoch": 1.8288913615570745, "grad_norm": 1.8733489513397217, "learning_rate": 7.06444275820974e-05, "loss": 3.076, "step": 207100 }, { "epoch": 1.8293329094473587, "grad_norm": 4.141226768493652, "learning_rate": 7.06317704010533e-05, "loss": 3.1736, "step": 207150 }, { "epoch": 1.8297744573376429, "grad_norm": 2.7284300327301025, "learning_rate": 7.06191116263799e-05, "loss": 2.8449, "step": 207200 }, { "epoch": 1.830216005227927, "grad_norm": 1.1890994310379028, "learning_rate": 7.060645125905497e-05, "loss": 3.0688, "step": 207250 }, { "epoch": 1.8306575531182112, "grad_norm": 2.6329152584075928, "learning_rate": 7.05937893000564e-05, "loss": 3.0454, "step": 207300 }, { "epoch": 1.8310991010084954, "grad_norm": 4.558962821960449, "learning_rate": 7.058112575036224e-05, "loss": 3.0297, "step": 207350 }, { "epoch": 1.8315406488987795, "grad_norm": 2.9594004154205322, "learning_rate": 7.056846061095063e-05, "loss": 3.09, "step": 207400 }, { "epoch": 1.8319821967890637, "grad_norm": 2.892536163330078, "learning_rate": 7.055579388279988e-05, "loss": 2.9773, "step": 207450 }, { "epoch": 1.8324237446793479, "grad_norm": 2.9856982231140137, "learning_rate": 7.054312556688835e-05, "loss": 3.4105, "step": 207500 }, { "epoch": 1.832865292569632, "grad_norm": 3.486680746078491, "learning_rate": 7.053045566419457e-05, "loss": 3.513, "step": 207550 }, { "epoch": 1.8333068404599162, "grad_norm": 1.4230419397354126, "learning_rate": 7.051778417569719e-05, "loss": 3.2426, "step": 207600 }, { "epoch": 1.8337483883502004, "grad_norm": 1.3702068328857422, "learning_rate": 7.050511110237497e-05, "loss": 3.2997, "step": 207650 }, { "epoch": 1.8341899362404845, "grad_norm": 3.2727692127227783, "learning_rate": 7.049243644520677e-05, "loss": 3.0542, "step": 207700 }, { "epoch": 1.8346314841307687, "grad_norm": 2.400622606277466, "learning_rate": 7.047976020517164e-05, "loss": 2.4248, "step": 207750 }, { "epoch": 1.8350730320210529, "grad_norm": 2.2951934337615967, "learning_rate": 7.04670823832487e-05, "loss": 3.0014, "step": 207800 }, { "epoch": 1.835514579911337, "grad_norm": 4.202934741973877, "learning_rate": 7.045440298041721e-05, "loss": 3.0171, "step": 207850 }, { "epoch": 1.8359561278016212, "grad_norm": 5.109984874725342, "learning_rate": 7.044172199765652e-05, "loss": 3.0452, "step": 207900 }, { "epoch": 1.8363976756919056, "grad_norm": 0.8654121160507202, "learning_rate": 7.042903943594617e-05, "loss": 3.245, "step": 207950 }, { "epoch": 1.8368392235821898, "grad_norm": 2.106384754180908, "learning_rate": 7.041635529626574e-05, "loss": 3.1378, "step": 208000 }, { "epoch": 1.837280771472474, "grad_norm": 2.5134472846984863, "learning_rate": 7.040366957959499e-05, "loss": 3.4982, "step": 208050 }, { "epoch": 1.837722319362758, "grad_norm": 3.753645181655884, "learning_rate": 7.039123604820596e-05, "loss": 3.2786, "step": 208100 }, { "epoch": 1.8381638672530423, "grad_norm": 2.0565855503082275, "learning_rate": 7.037854721198529e-05, "loss": 2.9289, "step": 208150 }, { "epoch": 1.8386054151433264, "grad_norm": 2.073881149291992, "learning_rate": 7.036585680169465e-05, "loss": 3.1887, "step": 208200 }, { "epoch": 1.8390469630336106, "grad_norm": 1.2692102193832397, "learning_rate": 7.035316481831427e-05, "loss": 2.9537, "step": 208250 }, { "epoch": 1.8394885109238948, "grad_norm": 1.5824757814407349, "learning_rate": 7.034047126282451e-05, "loss": 2.8978, "step": 208300 }, { "epoch": 1.8399300588141791, "grad_norm": 4.089334964752197, "learning_rate": 7.032777613620582e-05, "loss": 3.3272, "step": 208350 }, { "epoch": 1.8403716067044633, "grad_norm": 2.8007025718688965, "learning_rate": 7.031507943943879e-05, "loss": 3.1612, "step": 208400 }, { "epoch": 1.8408131545947475, "grad_norm": 1.6686805486679077, "learning_rate": 7.030238117350414e-05, "loss": 2.7393, "step": 208450 }, { "epoch": 1.8412547024850316, "grad_norm": 3.463226079940796, "learning_rate": 7.028968133938269e-05, "loss": 2.7314, "step": 208500 }, { "epoch": 1.8416962503753158, "grad_norm": 4.528909683227539, "learning_rate": 7.027697993805544e-05, "loss": 3.1757, "step": 208550 }, { "epoch": 1.8421377982656, "grad_norm": 2.405137538909912, "learning_rate": 7.026427697050342e-05, "loss": 3.126, "step": 208600 }, { "epoch": 1.8425793461558841, "grad_norm": 2.5665478706359863, "learning_rate": 7.025157243770782e-05, "loss": 3.4373, "step": 208650 }, { "epoch": 1.8430208940461683, "grad_norm": 1.1757197380065918, "learning_rate": 7.023886634065e-05, "loss": 3.4072, "step": 208700 }, { "epoch": 1.8434624419364525, "grad_norm": 1.8347642421722412, "learning_rate": 7.022615868031138e-05, "loss": 2.8205, "step": 208750 }, { "epoch": 1.8439039898267366, "grad_norm": 3.2277307510375977, "learning_rate": 7.02134494576735e-05, "loss": 2.8286, "step": 208800 }, { "epoch": 1.8443455377170208, "grad_norm": 5.538561820983887, "learning_rate": 7.020073867371806e-05, "loss": 3.152, "step": 208850 }, { "epoch": 1.844787085607305, "grad_norm": 2.4161622524261475, "learning_rate": 7.018802632942687e-05, "loss": 3.0617, "step": 208900 }, { "epoch": 1.8452286334975891, "grad_norm": 3.8103973865509033, "learning_rate": 7.017531242578182e-05, "loss": 3.1433, "step": 208950 }, { "epoch": 1.8456701813878733, "grad_norm": 1.9754663705825806, "learning_rate": 7.016259696376496e-05, "loss": 3.0137, "step": 209000 }, { "epoch": 1.8461117292781575, "grad_norm": 3.691028594970703, "learning_rate": 7.014987994435846e-05, "loss": 3.0269, "step": 209050 }, { "epoch": 1.8465532771684416, "grad_norm": 4.685206413269043, "learning_rate": 7.013716136854462e-05, "loss": 3.7514, "step": 209100 }, { "epoch": 1.8469948250587258, "grad_norm": 4.822083473205566, "learning_rate": 7.012444123730579e-05, "loss": 3.11, "step": 209150 }, { "epoch": 1.84743637294901, "grad_norm": 2.390188694000244, "learning_rate": 7.011171955162455e-05, "loss": 3.4009, "step": 209200 }, { "epoch": 1.8478779208392941, "grad_norm": 2.3603127002716064, "learning_rate": 7.009899631248352e-05, "loss": 3.5394, "step": 209250 }, { "epoch": 1.8483194687295783, "grad_norm": 1.7594237327575684, "learning_rate": 7.008627152086545e-05, "loss": 3.0626, "step": 209300 }, { "epoch": 1.8487610166198625, "grad_norm": 4.551392555236816, "learning_rate": 7.007354517775323e-05, "loss": 3.2904, "step": 209350 }, { "epoch": 1.8492025645101466, "grad_norm": 3.563063144683838, "learning_rate": 7.006081728412985e-05, "loss": 3.3053, "step": 209400 }, { "epoch": 1.8496441124004308, "grad_norm": 2.135427236557007, "learning_rate": 7.004808784097847e-05, "loss": 3.666, "step": 209450 }, { "epoch": 1.850085660290715, "grad_norm": 2.4800546169281006, "learning_rate": 7.00353568492823e-05, "loss": 3.5286, "step": 209500 }, { "epoch": 1.8505272081809994, "grad_norm": 4.204072952270508, "learning_rate": 7.002262431002471e-05, "loss": 2.979, "step": 209550 }, { "epoch": 1.8509687560712835, "grad_norm": 2.9853248596191406, "learning_rate": 7.000989022418916e-05, "loss": 3.5786, "step": 209600 }, { "epoch": 1.8514103039615677, "grad_norm": 0.9208089709281921, "learning_rate": 6.999715459275929e-05, "loss": 3.2202, "step": 209650 }, { "epoch": 1.8518518518518519, "grad_norm": 1.4217525720596313, "learning_rate": 6.99844174167188e-05, "loss": 3.2799, "step": 209700 }, { "epoch": 1.852293399742136, "grad_norm": 0.9643871784210205, "learning_rate": 6.997167869705152e-05, "loss": 3.3573, "step": 209750 }, { "epoch": 1.8527349476324202, "grad_norm": 1.374295949935913, "learning_rate": 6.995893843474141e-05, "loss": 2.9871, "step": 209800 }, { "epoch": 1.8531764955227044, "grad_norm": 3.098726511001587, "learning_rate": 6.994619663077255e-05, "loss": 3.1085, "step": 209850 }, { "epoch": 1.8536180434129885, "grad_norm": 7.691214084625244, "learning_rate": 6.993345328612917e-05, "loss": 2.9589, "step": 209900 }, { "epoch": 1.854059591303273, "grad_norm": 2.3036842346191406, "learning_rate": 6.992070840179552e-05, "loss": 3.3949, "step": 209950 }, { "epoch": 1.854501139193557, "grad_norm": 2.4765408039093018, "learning_rate": 6.990796197875609e-05, "loss": 3.071, "step": 210000 }, { "epoch": 1.854501139193557, "eval_asr_loss": 0.9128160601105207, "eval_loss": 2.8701679706573486, "eval_runtime": 20.6971, "eval_samples_per_second": 37.107, "eval_steps_per_second": 9.277, "eval_tts_loss": 5.965158913742212, "step": 210000 }, { "epoch": 1.8549426870838412, "grad_norm": 5.095572471618652, "learning_rate": 6.989521401799539e-05, "loss": 2.8314, "step": 210050 }, { "epoch": 1.8553842349741254, "grad_norm": 1.9772043228149414, "learning_rate": 6.988246452049814e-05, "loss": 2.9511, "step": 210100 }, { "epoch": 1.8558257828644096, "grad_norm": 3.964059352874756, "learning_rate": 6.986971348724909e-05, "loss": 3.0983, "step": 210150 }, { "epoch": 1.8562673307546937, "grad_norm": 2.66196608543396, "learning_rate": 6.985696091923314e-05, "loss": 3.0646, "step": 210200 }, { "epoch": 1.856708878644978, "grad_norm": 3.449439764022827, "learning_rate": 6.984420681743537e-05, "loss": 3.2398, "step": 210250 }, { "epoch": 1.857150426535262, "grad_norm": 10.892763137817383, "learning_rate": 6.98314511828409e-05, "loss": 3.1401, "step": 210300 }, { "epoch": 1.8575919744255462, "grad_norm": 1.941422939300537, "learning_rate": 6.981869401643498e-05, "loss": 3.3163, "step": 210350 }, { "epoch": 1.8580335223158304, "grad_norm": 3.753326892852783, "learning_rate": 6.980593531920302e-05, "loss": 2.955, "step": 210400 }, { "epoch": 1.8584750702061146, "grad_norm": 3.6323447227478027, "learning_rate": 6.979317509213049e-05, "loss": 2.9434, "step": 210450 }, { "epoch": 1.8589166180963987, "grad_norm": 1.4818679094314575, "learning_rate": 6.978041333620303e-05, "loss": 2.8079, "step": 210500 }, { "epoch": 1.859358165986683, "grad_norm": 2.958521842956543, "learning_rate": 6.976765005240636e-05, "loss": 3.0063, "step": 210550 }, { "epoch": 1.859799713876967, "grad_norm": 5.629188537597656, "learning_rate": 6.975488524172637e-05, "loss": 3.156, "step": 210600 }, { "epoch": 1.8602412617672512, "grad_norm": 5.116648197174072, "learning_rate": 6.9742118905149e-05, "loss": 2.5961, "step": 210650 }, { "epoch": 1.8606828096575354, "grad_norm": 3.616987466812134, "learning_rate": 6.972935104366036e-05, "loss": 3.3869, "step": 210700 }, { "epoch": 1.8611243575478196, "grad_norm": 1.0495541095733643, "learning_rate": 6.971658165824665e-05, "loss": 3.5673, "step": 210750 }, { "epoch": 1.8615659054381037, "grad_norm": 4.416067600250244, "learning_rate": 6.970381074989419e-05, "loss": 2.5567, "step": 210800 }, { "epoch": 1.862007453328388, "grad_norm": 4.130253314971924, "learning_rate": 6.969103831958944e-05, "loss": 3.2104, "step": 210850 }, { "epoch": 1.862449001218672, "grad_norm": 2.2265453338623047, "learning_rate": 6.967826436831896e-05, "loss": 3.3314, "step": 210900 }, { "epoch": 1.8628905491089562, "grad_norm": 3.0271718502044678, "learning_rate": 6.966548889706941e-05, "loss": 3.1078, "step": 210950 }, { "epoch": 1.8633320969992404, "grad_norm": 5.183620452880859, "learning_rate": 6.96527119068276e-05, "loss": 2.9481, "step": 211000 }, { "epoch": 1.8637736448895246, "grad_norm": 1.8804371356964111, "learning_rate": 6.963993339858047e-05, "loss": 2.826, "step": 211050 }, { "epoch": 1.8642151927798087, "grad_norm": 2.7824251651763916, "learning_rate": 6.962715337331501e-05, "loss": 3.0962, "step": 211100 }, { "epoch": 1.8646567406700931, "grad_norm": 2.67588472366333, "learning_rate": 6.961437183201838e-05, "loss": 2.7954, "step": 211150 }, { "epoch": 1.8650982885603773, "grad_norm": 2.0343751907348633, "learning_rate": 6.960158877567786e-05, "loss": 3.0353, "step": 211200 }, { "epoch": 1.8655398364506615, "grad_norm": 2.9644811153411865, "learning_rate": 6.958880420528083e-05, "loss": 2.8633, "step": 211250 }, { "epoch": 1.8659813843409456, "grad_norm": 2.8709774017333984, "learning_rate": 6.957601812181478e-05, "loss": 3.2773, "step": 211300 }, { "epoch": 1.8664229322312298, "grad_norm": 2.6100618839263916, "learning_rate": 6.95632305262673e-05, "loss": 2.8437, "step": 211350 }, { "epoch": 1.866864480121514, "grad_norm": 2.5091559886932373, "learning_rate": 6.955044141962619e-05, "loss": 3.3564, "step": 211400 }, { "epoch": 1.8673060280117981, "grad_norm": 2.1142711639404297, "learning_rate": 6.953765080287925e-05, "loss": 3.5474, "step": 211450 }, { "epoch": 1.8677475759020823, "grad_norm": 1.718662977218628, "learning_rate": 6.952485867701446e-05, "loss": 3.3915, "step": 211500 }, { "epoch": 1.8681891237923667, "grad_norm": 1.8452847003936768, "learning_rate": 6.951206504301991e-05, "loss": 2.9542, "step": 211550 }, { "epoch": 1.8686306716826508, "grad_norm": 0.9762416481971741, "learning_rate": 6.949926990188378e-05, "loss": 3.5469, "step": 211600 }, { "epoch": 1.869072219572935, "grad_norm": 2.9833409786224365, "learning_rate": 6.948647325459441e-05, "loss": 3.4791, "step": 211650 }, { "epoch": 1.8695137674632192, "grad_norm": 3.5274078845977783, "learning_rate": 6.947367510214022e-05, "loss": 2.7129, "step": 211700 }, { "epoch": 1.8699553153535033, "grad_norm": 1.9630882740020752, "learning_rate": 6.946087544550975e-05, "loss": 3.2424, "step": 211750 }, { "epoch": 1.8703968632437875, "grad_norm": 1.6840051412582397, "learning_rate": 6.944807428569168e-05, "loss": 3.1344, "step": 211800 }, { "epoch": 1.8708384111340717, "grad_norm": 2.4087986946105957, "learning_rate": 6.943527162367481e-05, "loss": 3.6155, "step": 211850 }, { "epoch": 1.8712799590243558, "grad_norm": 2.1375205516815186, "learning_rate": 6.942246746044798e-05, "loss": 3.2264, "step": 211900 }, { "epoch": 1.87172150691464, "grad_norm": 1.4749163389205933, "learning_rate": 6.940966179700026e-05, "loss": 3.4048, "step": 211950 }, { "epoch": 1.8721630548049242, "grad_norm": 2.870462656021118, "learning_rate": 6.939685463432073e-05, "loss": 3.3072, "step": 212000 }, { "epoch": 1.8726046026952083, "grad_norm": 2.298576593399048, "learning_rate": 6.93840459733987e-05, "loss": 3.2295, "step": 212050 }, { "epoch": 1.8730461505854925, "grad_norm": 2.6228911876678467, "learning_rate": 6.937123581522345e-05, "loss": 3.0334, "step": 212100 }, { "epoch": 1.8734876984757767, "grad_norm": 3.790313243865967, "learning_rate": 6.935842416078451e-05, "loss": 3.6681, "step": 212150 }, { "epoch": 1.8739292463660608, "grad_norm": 1.6933039426803589, "learning_rate": 6.934561101107147e-05, "loss": 3.3708, "step": 212200 }, { "epoch": 1.874370794256345, "grad_norm": 1.3448114395141602, "learning_rate": 6.9332796367074e-05, "loss": 3.5045, "step": 212250 }, { "epoch": 1.8748123421466292, "grad_norm": 1.4509389400482178, "learning_rate": 6.931998022978197e-05, "loss": 3.2581, "step": 212300 }, { "epoch": 1.8752538900369133, "grad_norm": 2.475632667541504, "learning_rate": 6.930716260018529e-05, "loss": 3.5627, "step": 212350 }, { "epoch": 1.8756954379271975, "grad_norm": 1.6319446563720703, "learning_rate": 6.929434347927401e-05, "loss": 3.4894, "step": 212400 }, { "epoch": 1.8761369858174817, "grad_norm": 3.2146284580230713, "learning_rate": 6.928152286803833e-05, "loss": 3.3476, "step": 212450 }, { "epoch": 1.8765785337077658, "grad_norm": 2.51674747467041, "learning_rate": 6.926870076746848e-05, "loss": 3.1805, "step": 212500 }, { "epoch": 1.87702008159805, "grad_norm": 3.0522384643554688, "learning_rate": 6.92558771785549e-05, "loss": 3.1066, "step": 212550 }, { "epoch": 1.8774616294883342, "grad_norm": 4.367920875549316, "learning_rate": 6.924305210228812e-05, "loss": 3.4727, "step": 212600 }, { "epoch": 1.8779031773786183, "grad_norm": 1.442237138748169, "learning_rate": 6.923022553965873e-05, "loss": 2.9928, "step": 212650 }, { "epoch": 1.8783447252689025, "grad_norm": 5.329409599304199, "learning_rate": 6.921739749165747e-05, "loss": 2.8959, "step": 212700 }, { "epoch": 1.878786273159187, "grad_norm": 1.250346302986145, "learning_rate": 6.92045679592752e-05, "loss": 3.0655, "step": 212750 }, { "epoch": 1.879227821049471, "grad_norm": 3.5236244201660156, "learning_rate": 6.919173694350293e-05, "loss": 3.4609, "step": 212800 }, { "epoch": 1.8796693689397552, "grad_norm": 1.863847017288208, "learning_rate": 6.917890444533173e-05, "loss": 3.4643, "step": 212850 }, { "epoch": 1.8801109168300394, "grad_norm": 1.5934778451919556, "learning_rate": 6.916607046575276e-05, "loss": 2.9508, "step": 212900 }, { "epoch": 1.8805524647203236, "grad_norm": 1.6757519245147705, "learning_rate": 6.915323500575739e-05, "loss": 2.9168, "step": 212950 }, { "epoch": 1.8809940126106077, "grad_norm": 3.9492690563201904, "learning_rate": 6.914065481961739e-05, "loss": 3.1243, "step": 213000 }, { "epoch": 1.8809940126106077, "eval_asr_loss": 0.9141171246415224, "eval_loss": 2.869227647781372, "eval_runtime": 20.5884, "eval_samples_per_second": 37.303, "eval_steps_per_second": 9.326, "eval_tts_loss": 5.979077017942771, "step": 213000 }, { "epoch": 1.881435560500892, "grad_norm": 1.5101724863052368, "learning_rate": 6.912781643132254e-05, "loss": 3.2316, "step": 213050 }, { "epoch": 1.881877108391176, "grad_norm": 1.5922937393188477, "learning_rate": 6.911497656556606e-05, "loss": 3.423, "step": 213100 }, { "epoch": 1.8823186562814604, "grad_norm": 2.6681389808654785, "learning_rate": 6.910213522333973e-05, "loss": 3.3049, "step": 213150 }, { "epoch": 1.8827602041717446, "grad_norm": 3.4186484813690186, "learning_rate": 6.908929240563545e-05, "loss": 2.7473, "step": 213200 }, { "epoch": 1.8832017520620288, "grad_norm": 2.218273639678955, "learning_rate": 6.90764481134452e-05, "loss": 3.5811, "step": 213250 }, { "epoch": 1.883643299952313, "grad_norm": 3.6075308322906494, "learning_rate": 6.906360234776109e-05, "loss": 3.3159, "step": 213300 }, { "epoch": 1.884084847842597, "grad_norm": 3.542436122894287, "learning_rate": 6.905075510957536e-05, "loss": 3.2598, "step": 213350 }, { "epoch": 1.8845263957328813, "grad_norm": 3.3695287704467773, "learning_rate": 6.903790639988032e-05, "loss": 3.4699, "step": 213400 }, { "epoch": 1.8849679436231654, "grad_norm": 4.290796279907227, "learning_rate": 6.902505621966848e-05, "loss": 3.405, "step": 213450 }, { "epoch": 1.8854094915134496, "grad_norm": 2.021557331085205, "learning_rate": 6.901220456993236e-05, "loss": 3.3985, "step": 213500 }, { "epoch": 1.8858510394037338, "grad_norm": 2.5229990482330322, "learning_rate": 6.899935145166467e-05, "loss": 3.265, "step": 213550 }, { "epoch": 1.886292587294018, "grad_norm": 1.655785322189331, "learning_rate": 6.898649686585817e-05, "loss": 2.9395, "step": 213600 }, { "epoch": 1.886734135184302, "grad_norm": 1.8800694942474365, "learning_rate": 6.89736408135058e-05, "loss": 3.3818, "step": 213650 }, { "epoch": 1.8871756830745863, "grad_norm": 1.9261225461959839, "learning_rate": 6.896078329560056e-05, "loss": 3.5134, "step": 213700 }, { "epoch": 1.8876172309648704, "grad_norm": 2.3512516021728516, "learning_rate": 6.894792431313561e-05, "loss": 2.9406, "step": 213750 }, { "epoch": 1.8880587788551546, "grad_norm": 1.5226099491119385, "learning_rate": 6.893506386710415e-05, "loss": 2.9769, "step": 213800 }, { "epoch": 1.8885003267454388, "grad_norm": 2.3892409801483154, "learning_rate": 6.89222019584996e-05, "loss": 3.4092, "step": 213850 }, { "epoch": 1.888941874635723, "grad_norm": 1.6916769742965698, "learning_rate": 6.890933858831539e-05, "loss": 3.077, "step": 213900 }, { "epoch": 1.889383422526007, "grad_norm": 0.7223499417304993, "learning_rate": 6.889647375754512e-05, "loss": 3.2138, "step": 213950 }, { "epoch": 1.8898249704162913, "grad_norm": 2.6566531658172607, "learning_rate": 6.88836074671825e-05, "loss": 3.0725, "step": 214000 }, { "epoch": 1.8902665183065754, "grad_norm": 2.1201887130737305, "learning_rate": 6.887073971822131e-05, "loss": 3.2355, "step": 214050 }, { "epoch": 1.8907080661968596, "grad_norm": 2.0156757831573486, "learning_rate": 6.885787051165553e-05, "loss": 3.1942, "step": 214100 }, { "epoch": 1.8911496140871438, "grad_norm": 3.2175350189208984, "learning_rate": 6.884499984847914e-05, "loss": 3.2014, "step": 214150 }, { "epoch": 1.891591161977428, "grad_norm": 3.1125988960266113, "learning_rate": 6.88321277296863e-05, "loss": 2.7858, "step": 214200 }, { "epoch": 1.892032709867712, "grad_norm": 2.6365787982940674, "learning_rate": 6.881925415627131e-05, "loss": 2.9779, "step": 214250 }, { "epoch": 1.8924742577579963, "grad_norm": 4.814088821411133, "learning_rate": 6.880637912922852e-05, "loss": 2.564, "step": 214300 }, { "epoch": 1.8929158056482807, "grad_norm": 2.180482864379883, "learning_rate": 6.87935026495524e-05, "loss": 3.4417, "step": 214350 }, { "epoch": 1.8933573535385648, "grad_norm": 3.2703518867492676, "learning_rate": 6.878062471823756e-05, "loss": 3.5152, "step": 214400 }, { "epoch": 1.893798901428849, "grad_norm": 2.8934197425842285, "learning_rate": 6.876774533627873e-05, "loss": 3.3839, "step": 214450 }, { "epoch": 1.8942404493191332, "grad_norm": 2.586165189743042, "learning_rate": 6.875486450467073e-05, "loss": 3.1851, "step": 214500 }, { "epoch": 1.8946819972094173, "grad_norm": 4.4620232582092285, "learning_rate": 6.874198222440845e-05, "loss": 3.3091, "step": 214550 }, { "epoch": 1.8951235450997015, "grad_norm": 2.581688642501831, "learning_rate": 6.8729098496487e-05, "loss": 2.9032, "step": 214600 }, { "epoch": 1.8955650929899857, "grad_norm": 4.92921781539917, "learning_rate": 6.87162133219015e-05, "loss": 2.9231, "step": 214650 }, { "epoch": 1.8960066408802698, "grad_norm": 2.6516809463500977, "learning_rate": 6.870332670164722e-05, "loss": 3.3176, "step": 214700 }, { "epoch": 1.8964481887705542, "grad_norm": 3.65873122215271, "learning_rate": 6.869043863671955e-05, "loss": 3.1889, "step": 214750 }, { "epoch": 1.8968897366608384, "grad_norm": 5.168747901916504, "learning_rate": 6.8677549128114e-05, "loss": 3.8415, "step": 214800 }, { "epoch": 1.8973312845511225, "grad_norm": 20.138532638549805, "learning_rate": 6.866465817682616e-05, "loss": 3.2379, "step": 214850 }, { "epoch": 1.8977728324414067, "grad_norm": 1.795832633972168, "learning_rate": 6.865176578385175e-05, "loss": 2.9264, "step": 214900 }, { "epoch": 1.8982143803316909, "grad_norm": 4.119626998901367, "learning_rate": 6.863887195018658e-05, "loss": 3.3952, "step": 214950 }, { "epoch": 1.898655928221975, "grad_norm": 0.8645926117897034, "learning_rate": 6.862597667682661e-05, "loss": 3.247, "step": 215000 }, { "epoch": 1.8990974761122592, "grad_norm": 1.5755289793014526, "learning_rate": 6.86130799647679e-05, "loss": 3.2569, "step": 215050 }, { "epoch": 1.8995390240025434, "grad_norm": 2.1388795375823975, "learning_rate": 6.860018181500659e-05, "loss": 3.1373, "step": 215100 }, { "epoch": 1.8999805718928275, "grad_norm": 2.395258665084839, "learning_rate": 6.858728222853895e-05, "loss": 3.3231, "step": 215150 }, { "epoch": 1.9004221197831117, "grad_norm": 2.154432773590088, "learning_rate": 6.857438120636137e-05, "loss": 3.3407, "step": 215200 }, { "epoch": 1.9008636676733959, "grad_norm": 4.255756378173828, "learning_rate": 6.856147874947036e-05, "loss": 3.256, "step": 215250 }, { "epoch": 1.90130521556368, "grad_norm": 4.8996357917785645, "learning_rate": 6.854857485886252e-05, "loss": 3.2052, "step": 215300 }, { "epoch": 1.9017467634539642, "grad_norm": 3.192509412765503, "learning_rate": 6.853566953553454e-05, "loss": 3.4194, "step": 215350 }, { "epoch": 1.9021883113442484, "grad_norm": 4.064058303833008, "learning_rate": 6.852276278048329e-05, "loss": 3.2169, "step": 215400 }, { "epoch": 1.9026298592345325, "grad_norm": 2.9588820934295654, "learning_rate": 6.850985459470567e-05, "loss": 3.0224, "step": 215450 }, { "epoch": 1.9030714071248167, "grad_norm": 4.584090709686279, "learning_rate": 6.849694497919874e-05, "loss": 3.308, "step": 215500 }, { "epoch": 1.9035129550151009, "grad_norm": 2.459688186645508, "learning_rate": 6.848403393495966e-05, "loss": 2.8452, "step": 215550 }, { "epoch": 1.903954502905385, "grad_norm": 3.800164222717285, "learning_rate": 6.847112146298568e-05, "loss": 2.9931, "step": 215600 }, { "epoch": 1.9043960507956692, "grad_norm": 3.536306858062744, "learning_rate": 6.845820756427424e-05, "loss": 3.2442, "step": 215650 }, { "epoch": 1.9048375986859534, "grad_norm": 3.559290647506714, "learning_rate": 6.844529223982275e-05, "loss": 2.9447, "step": 215700 }, { "epoch": 1.9052791465762375, "grad_norm": 4.145590305328369, "learning_rate": 6.843237549062886e-05, "loss": 3.1105, "step": 215750 }, { "epoch": 1.9057206944665217, "grad_norm": 0.8923916220664978, "learning_rate": 6.841945731769028e-05, "loss": 2.9162, "step": 215800 }, { "epoch": 1.9061622423568059, "grad_norm": 3.795706272125244, "learning_rate": 6.840653772200479e-05, "loss": 3.0548, "step": 215850 }, { "epoch": 1.90660379024709, "grad_norm": 2.7242794036865234, "learning_rate": 6.839361670457037e-05, "loss": 2.8588, "step": 215900 }, { "epoch": 1.9070453381373744, "grad_norm": 2.8802475929260254, "learning_rate": 6.8380694266385e-05, "loss": 2.8865, "step": 215950 }, { "epoch": 1.9074868860276586, "grad_norm": 2.9323482513427734, "learning_rate": 6.836777040844688e-05, "loss": 3.0616, "step": 216000 }, { "epoch": 1.9074868860276586, "eval_asr_loss": 0.9128535109015836, "eval_loss": 2.869661331176758, "eval_runtime": 20.623, "eval_samples_per_second": 37.24, "eval_steps_per_second": 9.31, "eval_tts_loss": 5.943447776517638, "step": 216000 }, { "epoch": 1.9079284339179428, "grad_norm": 1.7712619304656982, "learning_rate": 6.835484513175427e-05, "loss": 3.6614, "step": 216050 }, { "epoch": 1.908369981808227, "grad_norm": 4.220677852630615, "learning_rate": 6.834191843730549e-05, "loss": 3.0352, "step": 216100 }, { "epoch": 1.908811529698511, "grad_norm": 3.0093469619750977, "learning_rate": 6.832899032609904e-05, "loss": 3.3183, "step": 216150 }, { "epoch": 1.9092530775887953, "grad_norm": 1.6805152893066406, "learning_rate": 6.831606079913352e-05, "loss": 3.3335, "step": 216200 }, { "epoch": 1.9096946254790794, "grad_norm": 5.431236267089844, "learning_rate": 6.830312985740764e-05, "loss": 3.4567, "step": 216250 }, { "epoch": 1.9101361733693636, "grad_norm": 1.6891889572143555, "learning_rate": 6.829019750192016e-05, "loss": 2.7522, "step": 216300 }, { "epoch": 1.910577721259648, "grad_norm": 3.535320997238159, "learning_rate": 6.827726373367003e-05, "loss": 3.6283, "step": 216350 }, { "epoch": 1.9110192691499321, "grad_norm": 5.188021183013916, "learning_rate": 6.826432855365626e-05, "loss": 3.189, "step": 216400 }, { "epoch": 1.9114608170402163, "grad_norm": 3.6246254444122314, "learning_rate": 6.825165070851258e-05, "loss": 3.0975, "step": 216450 }, { "epoch": 1.9119023649305005, "grad_norm": 4.558775901794434, "learning_rate": 6.823871273615455e-05, "loss": 3.5297, "step": 216500 }, { "epoch": 1.9123439128207846, "grad_norm": 3.1083285808563232, "learning_rate": 6.822577335501065e-05, "loss": 3.217, "step": 216550 }, { "epoch": 1.9127854607110688, "grad_norm": 2.8470730781555176, "learning_rate": 6.821283256608029e-05, "loss": 3.5633, "step": 216600 }, { "epoch": 1.913227008601353, "grad_norm": 2.6485366821289062, "learning_rate": 6.819989037036303e-05, "loss": 3.0846, "step": 216650 }, { "epoch": 1.9136685564916371, "grad_norm": 4.975265026092529, "learning_rate": 6.81869467688586e-05, "loss": 2.9518, "step": 216700 }, { "epoch": 1.9141101043819213, "grad_norm": 3.9095828533172607, "learning_rate": 6.817400176256674e-05, "loss": 3.0829, "step": 216750 }, { "epoch": 1.9145516522722055, "grad_norm": 2.3637115955352783, "learning_rate": 6.816105535248735e-05, "loss": 3.1916, "step": 216800 }, { "epoch": 1.9149932001624896, "grad_norm": 1.6925334930419922, "learning_rate": 6.814810753962045e-05, "loss": 3.2711, "step": 216850 }, { "epoch": 1.9154347480527738, "grad_norm": 1.323299527168274, "learning_rate": 6.813515832496613e-05, "loss": 3.3585, "step": 216900 }, { "epoch": 1.915876295943058, "grad_norm": 4.358603000640869, "learning_rate": 6.812220770952462e-05, "loss": 3.396, "step": 216950 }, { "epoch": 1.9163178438333421, "grad_norm": 2.25028920173645, "learning_rate": 6.810925569429622e-05, "loss": 3.6563, "step": 217000 }, { "epoch": 1.9167593917236263, "grad_norm": 1.2531156539916992, "learning_rate": 6.809630228028138e-05, "loss": 3.1924, "step": 217050 }, { "epoch": 1.9172009396139105, "grad_norm": 2.4231159687042236, "learning_rate": 6.808334746848064e-05, "loss": 3.3104, "step": 217100 }, { "epoch": 1.9176424875041946, "grad_norm": 4.368786334991455, "learning_rate": 6.807039125989466e-05, "loss": 3.348, "step": 217150 }, { "epoch": 1.9180840353944788, "grad_norm": 2.5258097648620605, "learning_rate": 6.80574336555242e-05, "loss": 2.6929, "step": 217200 }, { "epoch": 1.918525583284763, "grad_norm": 3.867431163787842, "learning_rate": 6.804447465637008e-05, "loss": 3.1379, "step": 217250 }, { "epoch": 1.9189671311750471, "grad_norm": 1.6516950130462646, "learning_rate": 6.803151426343331e-05, "loss": 3.2439, "step": 217300 }, { "epoch": 1.9194086790653313, "grad_norm": 1.0718307495117188, "learning_rate": 6.8018552477715e-05, "loss": 3.08, "step": 217350 }, { "epoch": 1.9198502269556155, "grad_norm": 1.568926453590393, "learning_rate": 6.800558930021627e-05, "loss": 3.207, "step": 217400 }, { "epoch": 1.9202917748458996, "grad_norm": 2.034952402114868, "learning_rate": 6.799262473193846e-05, "loss": 2.7577, "step": 217450 }, { "epoch": 1.9207333227361838, "grad_norm": 1.5532232522964478, "learning_rate": 6.797965877388297e-05, "loss": 3.395, "step": 217500 }, { "epoch": 1.9211748706264682, "grad_norm": 2.738466739654541, "learning_rate": 6.79666914270513e-05, "loss": 3.2612, "step": 217550 }, { "epoch": 1.9216164185167524, "grad_norm": 1.9071691036224365, "learning_rate": 6.795372269244506e-05, "loss": 3.145, "step": 217600 }, { "epoch": 1.9220579664070365, "grad_norm": 4.6840596199035645, "learning_rate": 6.794075257106599e-05, "loss": 3.3608, "step": 217650 }, { "epoch": 1.9224995142973207, "grad_norm": 1.42924165725708, "learning_rate": 6.792778106391591e-05, "loss": 2.9166, "step": 217700 }, { "epoch": 1.9229410621876049, "grad_norm": 3.003345012664795, "learning_rate": 6.79148081719968e-05, "loss": 3.1679, "step": 217750 }, { "epoch": 1.923382610077889, "grad_norm": 2.117689609527588, "learning_rate": 6.790183389631065e-05, "loss": 3.0708, "step": 217800 }, { "epoch": 1.9238241579681732, "grad_norm": 1.345949411392212, "learning_rate": 6.788885823785962e-05, "loss": 3.379, "step": 217850 }, { "epoch": 1.9242657058584574, "grad_norm": 2.085387945175171, "learning_rate": 6.787588119764599e-05, "loss": 3.3116, "step": 217900 }, { "epoch": 1.9247072537487417, "grad_norm": 1.578888177871704, "learning_rate": 6.786290277667214e-05, "loss": 3.804, "step": 217950 }, { "epoch": 1.925148801639026, "grad_norm": 2.3243682384490967, "learning_rate": 6.78499229759405e-05, "loss": 3.2998, "step": 218000 }, { "epoch": 1.92559034952931, "grad_norm": 1.385135293006897, "learning_rate": 6.783694179645368e-05, "loss": 3.2463, "step": 218050 }, { "epoch": 1.9260318974195942, "grad_norm": 4.232430934906006, "learning_rate": 6.782395923921436e-05, "loss": 3.0034, "step": 218100 }, { "epoch": 1.9264734453098784, "grad_norm": 0.8287903070449829, "learning_rate": 6.781097530522534e-05, "loss": 3.0441, "step": 218150 }, { "epoch": 1.9269149932001626, "grad_norm": 1.6629900932312012, "learning_rate": 6.779798999548949e-05, "loss": 3.075, "step": 218200 }, { "epoch": 1.9273565410904467, "grad_norm": 6.08732795715332, "learning_rate": 6.778500331100985e-05, "loss": 3.3407, "step": 218250 }, { "epoch": 1.927798088980731, "grad_norm": 2.5804171562194824, "learning_rate": 6.777201525278954e-05, "loss": 3.2754, "step": 218300 }, { "epoch": 1.928239636871015, "grad_norm": 2.3154220581054688, "learning_rate": 6.77590258218317e-05, "loss": 3.1724, "step": 218350 }, { "epoch": 1.9286811847612992, "grad_norm": 2.769705057144165, "learning_rate": 6.774603501913975e-05, "loss": 3.3089, "step": 218400 }, { "epoch": 1.9291227326515834, "grad_norm": 1.773826241493225, "learning_rate": 6.773304284571707e-05, "loss": 3.05, "step": 218450 }, { "epoch": 1.9295642805418676, "grad_norm": 1.2457982301712036, "learning_rate": 6.77200493025672e-05, "loss": 2.9031, "step": 218500 }, { "epoch": 1.9300058284321517, "grad_norm": 4.650148391723633, "learning_rate": 6.770705439069377e-05, "loss": 3.1762, "step": 218550 }, { "epoch": 1.930447376322436, "grad_norm": 4.885578632354736, "learning_rate": 6.769405811110055e-05, "loss": 2.9341, "step": 218600 }, { "epoch": 1.93088892421272, "grad_norm": 5.776822090148926, "learning_rate": 6.768106046479138e-05, "loss": 3.1056, "step": 218650 }, { "epoch": 1.9313304721030042, "grad_norm": 3.6868488788604736, "learning_rate": 6.766806145277023e-05, "loss": 3.2905, "step": 218700 }, { "epoch": 1.9317720199932884, "grad_norm": 2.8849265575408936, "learning_rate": 6.765506107604113e-05, "loss": 3.3836, "step": 218750 }, { "epoch": 1.9322135678835726, "grad_norm": 2.124006986618042, "learning_rate": 6.764205933560829e-05, "loss": 3.4177, "step": 218800 }, { "epoch": 1.9326551157738567, "grad_norm": 7.862915992736816, "learning_rate": 6.762905623247596e-05, "loss": 2.9393, "step": 218850 }, { "epoch": 1.933096663664141, "grad_norm": 3.948845624923706, "learning_rate": 6.761605176764854e-05, "loss": 3.3315, "step": 218900 }, { "epoch": 1.933538211554425, "grad_norm": 0.9531094431877136, "learning_rate": 6.760304594213048e-05, "loss": 3.3055, "step": 218950 }, { "epoch": 1.9339797594447092, "grad_norm": 1.7013096809387207, "learning_rate": 6.75900387569264e-05, "loss": 3.2939, "step": 219000 }, { "epoch": 1.9339797594447092, "eval_asr_loss": 0.922124789094716, "eval_loss": 2.875542402267456, "eval_runtime": 20.7112, "eval_samples_per_second": 37.081, "eval_steps_per_second": 9.27, "eval_tts_loss": 5.956120436386516, "step": 219000 }, { "epoch": 1.9344213073349934, "grad_norm": 5.030139446258545, "learning_rate": 6.757703021304099e-05, "loss": 3.032, "step": 219050 }, { "epoch": 1.9348628552252778, "grad_norm": 3.529791831970215, "learning_rate": 6.756402031147905e-05, "loss": 2.8338, "step": 219100 }, { "epoch": 1.935304403115562, "grad_norm": 2.3809285163879395, "learning_rate": 6.755100905324548e-05, "loss": 3.3668, "step": 219150 }, { "epoch": 1.9357459510058461, "grad_norm": 4.645364284515381, "learning_rate": 6.753799643934526e-05, "loss": 3.2265, "step": 219200 }, { "epoch": 1.9361874988961303, "grad_norm": 3.215935468673706, "learning_rate": 6.752498247078355e-05, "loss": 3.178, "step": 219250 }, { "epoch": 1.9366290467864145, "grad_norm": 4.604817867279053, "learning_rate": 6.751196714856556e-05, "loss": 3.4193, "step": 219300 }, { "epoch": 1.9370705946766986, "grad_norm": 5.685570240020752, "learning_rate": 6.74989504736966e-05, "loss": 3.0059, "step": 219350 }, { "epoch": 1.9375121425669828, "grad_norm": 2.7423245906829834, "learning_rate": 6.74859324471821e-05, "loss": 2.7509, "step": 219400 }, { "epoch": 1.937953690457267, "grad_norm": 2.651466131210327, "learning_rate": 6.747291307002759e-05, "loss": 3.3804, "step": 219450 }, { "epoch": 1.9383952383475513, "grad_norm": 0.9820277690887451, "learning_rate": 6.745989234323872e-05, "loss": 3.3264, "step": 219500 }, { "epoch": 1.9388367862378355, "grad_norm": 1.3179621696472168, "learning_rate": 6.744687026782123e-05, "loss": 3.0364, "step": 219550 }, { "epoch": 1.9392783341281197, "grad_norm": 3.5786843299865723, "learning_rate": 6.743384684478093e-05, "loss": 2.9091, "step": 219600 }, { "epoch": 1.9397198820184038, "grad_norm": 4.74429178237915, "learning_rate": 6.742082207512381e-05, "loss": 3.3593, "step": 219650 }, { "epoch": 1.940161429908688, "grad_norm": 2.6699705123901367, "learning_rate": 6.740779595985591e-05, "loss": 3.6889, "step": 219700 }, { "epoch": 1.9406029777989722, "grad_norm": 2.328733205795288, "learning_rate": 6.73947684999834e-05, "loss": 3.5794, "step": 219750 }, { "epoch": 1.9410445256892563, "grad_norm": 2.218416213989258, "learning_rate": 6.738173969651251e-05, "loss": 3.1906, "step": 219800 }, { "epoch": 1.9414860735795405, "grad_norm": 1.3207190036773682, "learning_rate": 6.736870955044963e-05, "loss": 3.1299, "step": 219850 }, { "epoch": 1.9419276214698247, "grad_norm": 2.481778144836426, "learning_rate": 6.735567806280121e-05, "loss": 2.9451, "step": 219900 }, { "epoch": 1.9423691693601088, "grad_norm": 2.82830548286438, "learning_rate": 6.734264523457383e-05, "loss": 3.3885, "step": 219950 }, { "epoch": 1.942810717250393, "grad_norm": 1.681307315826416, "learning_rate": 6.732987176325148e-05, "loss": 3.4272, "step": 220000 }, { "epoch": 1.9432522651406772, "grad_norm": 4.030539035797119, "learning_rate": 6.731683628364775e-05, "loss": 3.3767, "step": 220050 }, { "epoch": 1.9436938130309613, "grad_norm": 2.7558698654174805, "learning_rate": 6.730379946646527e-05, "loss": 3.6885, "step": 220100 }, { "epoch": 1.9441353609212455, "grad_norm": 2.890310525894165, "learning_rate": 6.729076131271102e-05, "loss": 3.3828, "step": 220150 }, { "epoch": 1.9445769088115297, "grad_norm": 1.815947413444519, "learning_rate": 6.727772182339209e-05, "loss": 3.3614, "step": 220200 }, { "epoch": 1.9450184567018138, "grad_norm": 1.3657796382904053, "learning_rate": 6.726468099951563e-05, "loss": 3.1284, "step": 220250 }, { "epoch": 1.945460004592098, "grad_norm": 2.1993958950042725, "learning_rate": 6.725163884208898e-05, "loss": 3.0117, "step": 220300 }, { "epoch": 1.9459015524823822, "grad_norm": 3.3096845149993896, "learning_rate": 6.723859535211951e-05, "loss": 3.5753, "step": 220350 }, { "epoch": 1.9463431003726663, "grad_norm": 2.9292962551116943, "learning_rate": 6.722555053061474e-05, "loss": 3.0053, "step": 220400 }, { "epoch": 1.9467846482629505, "grad_norm": 1.084033727645874, "learning_rate": 6.721250437858225e-05, "loss": 2.8641, "step": 220450 }, { "epoch": 1.9472261961532347, "grad_norm": 1.6132060289382935, "learning_rate": 6.719945689702975e-05, "loss": 3.2127, "step": 220500 }, { "epoch": 1.9476677440435188, "grad_norm": 3.2746384143829346, "learning_rate": 6.718640808696508e-05, "loss": 2.9314, "step": 220550 }, { "epoch": 1.948109291933803, "grad_norm": 2.0202183723449707, "learning_rate": 6.71733579493961e-05, "loss": 3.4692, "step": 220600 }, { "epoch": 1.9485508398240872, "grad_norm": 1.8772021532058716, "learning_rate": 6.716030648533085e-05, "loss": 2.9944, "step": 220650 }, { "epoch": 1.9489923877143716, "grad_norm": 2.3999502658843994, "learning_rate": 6.714725369577744e-05, "loss": 2.832, "step": 220700 }, { "epoch": 1.9494339356046557, "grad_norm": 4.706568241119385, "learning_rate": 6.713419958174407e-05, "loss": 3.0989, "step": 220750 }, { "epoch": 1.9498754834949399, "grad_norm": 4.1655097007751465, "learning_rate": 6.712114414423909e-05, "loss": 2.9818, "step": 220800 }, { "epoch": 1.950317031385224, "grad_norm": 3.7588179111480713, "learning_rate": 6.710808738427092e-05, "loss": 3.2759, "step": 220850 }, { "epoch": 1.9507585792755082, "grad_norm": 1.6095143556594849, "learning_rate": 6.709502930284805e-05, "loss": 3.4051, "step": 220900 }, { "epoch": 1.9512001271657924, "grad_norm": 2.343217372894287, "learning_rate": 6.708196990097914e-05, "loss": 3.5032, "step": 220950 }, { "epoch": 1.9516416750560766, "grad_norm": 3.6511197090148926, "learning_rate": 6.70689091796729e-05, "loss": 3.2668, "step": 221000 }, { "epoch": 1.9520832229463607, "grad_norm": 4.484455585479736, "learning_rate": 6.705584713993817e-05, "loss": 3.2429, "step": 221050 }, { "epoch": 1.952524770836645, "grad_norm": 1.386128306388855, "learning_rate": 6.704278378278387e-05, "loss": 3.1461, "step": 221100 }, { "epoch": 1.9529663187269293, "grad_norm": 1.4120815992355347, "learning_rate": 6.702971910921904e-05, "loss": 3.2366, "step": 221150 }, { "epoch": 1.9534078666172134, "grad_norm": 1.724896788597107, "learning_rate": 6.701665312025284e-05, "loss": 3.384, "step": 221200 }, { "epoch": 1.9538494145074976, "grad_norm": 0.8912140727043152, "learning_rate": 6.700358581689446e-05, "loss": 3.071, "step": 221250 }, { "epoch": 1.9542909623977818, "grad_norm": 5.190890312194824, "learning_rate": 6.699051720015326e-05, "loss": 3.1861, "step": 221300 }, { "epoch": 1.954732510288066, "grad_norm": 3.6501457691192627, "learning_rate": 6.697744727103871e-05, "loss": 2.9187, "step": 221350 }, { "epoch": 1.95517405817835, "grad_norm": 2.820183277130127, "learning_rate": 6.696437603056029e-05, "loss": 3.0877, "step": 221400 }, { "epoch": 1.9556156060686343, "grad_norm": 5.598944664001465, "learning_rate": 6.695130347972769e-05, "loss": 3.1602, "step": 221450 }, { "epoch": 1.9560571539589184, "grad_norm": 2.189221143722534, "learning_rate": 6.693822961955063e-05, "loss": 2.9852, "step": 221500 }, { "epoch": 1.9564987018492026, "grad_norm": 2.1708157062530518, "learning_rate": 6.6925154451039e-05, "loss": 3.4098, "step": 221550 }, { "epoch": 1.9569402497394868, "grad_norm": 1.5089625120162964, "learning_rate": 6.691207797520267e-05, "loss": 3.1213, "step": 221600 }, { "epoch": 1.957381797629771, "grad_norm": 2.335033416748047, "learning_rate": 6.689900019305174e-05, "loss": 3.354, "step": 221650 }, { "epoch": 1.957823345520055, "grad_norm": 3.1629550457000732, "learning_rate": 6.688592110559635e-05, "loss": 3.287, "step": 221700 }, { "epoch": 1.9582648934103393, "grad_norm": 6.7626142501831055, "learning_rate": 6.687284071384674e-05, "loss": 3.0697, "step": 221750 }, { "epoch": 1.9587064413006234, "grad_norm": 2.335103988647461, "learning_rate": 6.685975901881325e-05, "loss": 3.7996, "step": 221800 }, { "epoch": 1.9591479891909076, "grad_norm": 4.960870265960693, "learning_rate": 6.684667602150636e-05, "loss": 3.2685, "step": 221850 }, { "epoch": 1.9595895370811918, "grad_norm": 2.5751452445983887, "learning_rate": 6.683359172293659e-05, "loss": 3.0824, "step": 221900 }, { "epoch": 1.960031084971476, "grad_norm": 1.5638737678527832, "learning_rate": 6.68205061241146e-05, "loss": 3.1128, "step": 221950 }, { "epoch": 1.96047263286176, "grad_norm": 2.9490702152252197, "learning_rate": 6.680741922605116e-05, "loss": 3.1104, "step": 222000 }, { "epoch": 1.96047263286176, "eval_asr_loss": 0.9345355808110474, "eval_loss": 2.8736743927001953, "eval_runtime": 20.8897, "eval_samples_per_second": 36.765, "eval_steps_per_second": 9.191, "eval_tts_loss": 5.937240915996316, "step": 222000 }, { "epoch": 1.9609141807520443, "grad_norm": 1.6358108520507812, "learning_rate": 6.67943310297571e-05, "loss": 3.041, "step": 222050 }, { "epoch": 1.9613557286423284, "grad_norm": 3.1550421714782715, "learning_rate": 6.678124153624338e-05, "loss": 2.9968, "step": 222100 }, { "epoch": 1.9617972765326126, "grad_norm": 4.3898091316223145, "learning_rate": 6.676815074652106e-05, "loss": 2.8248, "step": 222150 }, { "epoch": 1.9622388244228968, "grad_norm": 2.0046401023864746, "learning_rate": 6.675505866160127e-05, "loss": 3.1247, "step": 222200 }, { "epoch": 1.962680372313181, "grad_norm": 4.28106689453125, "learning_rate": 6.674196528249528e-05, "loss": 3.3193, "step": 222250 }, { "epoch": 1.9631219202034653, "grad_norm": 4.919691562652588, "learning_rate": 6.672887061021444e-05, "loss": 3.5789, "step": 222300 }, { "epoch": 1.9635634680937495, "grad_norm": 3.231856346130371, "learning_rate": 6.671577464577022e-05, "loss": 2.8704, "step": 222350 }, { "epoch": 1.9640050159840337, "grad_norm": 3.4530904293060303, "learning_rate": 6.670267739017413e-05, "loss": 2.9751, "step": 222400 }, { "epoch": 1.9644465638743178, "grad_norm": 1.820241093635559, "learning_rate": 6.668957884443786e-05, "loss": 3.6656, "step": 222450 }, { "epoch": 1.964888111764602, "grad_norm": 4.895803928375244, "learning_rate": 6.667674101889736e-05, "loss": 2.7195, "step": 222500 }, { "epoch": 1.9653296596548862, "grad_norm": 1.803472876548767, "learning_rate": 6.666363992166847e-05, "loss": 3.4702, "step": 222550 }, { "epoch": 1.9657712075451703, "grad_norm": 4.227266788482666, "learning_rate": 6.665053753731469e-05, "loss": 3.2673, "step": 222600 }, { "epoch": 1.9662127554354545, "grad_norm": 2.4153201580047607, "learning_rate": 6.66374338668481e-05, "loss": 3.3712, "step": 222650 }, { "epoch": 1.9666543033257389, "grad_norm": 4.17147159576416, "learning_rate": 6.66243289112808e-05, "loss": 3.1089, "step": 222700 }, { "epoch": 1.967095851216023, "grad_norm": 1.907220721244812, "learning_rate": 6.661122267162508e-05, "loss": 3.0767, "step": 222750 }, { "epoch": 1.9675373991063072, "grad_norm": 3.343963861465454, "learning_rate": 6.659811514889324e-05, "loss": 3.0645, "step": 222800 }, { "epoch": 1.9679789469965914, "grad_norm": 2.7508366107940674, "learning_rate": 6.658500634409779e-05, "loss": 3.276, "step": 222850 }, { "epoch": 1.9684204948868755, "grad_norm": 6.333864212036133, "learning_rate": 6.657189625825122e-05, "loss": 2.9536, "step": 222900 }, { "epoch": 1.9688620427771597, "grad_norm": 4.854486465454102, "learning_rate": 6.655878489236617e-05, "loss": 3.2436, "step": 222950 }, { "epoch": 1.9693035906674439, "grad_norm": 2.8148059844970703, "learning_rate": 6.654567224745542e-05, "loss": 3.1323, "step": 223000 }, { "epoch": 1.969745138557728, "grad_norm": 3.6999666690826416, "learning_rate": 6.653255832453181e-05, "loss": 3.123, "step": 223050 }, { "epoch": 1.9701866864480122, "grad_norm": 4.362318515777588, "learning_rate": 6.651944312460823e-05, "loss": 3.1069, "step": 223100 }, { "epoch": 1.9706282343382964, "grad_norm": 4.602596759796143, "learning_rate": 6.650632664869779e-05, "loss": 2.7696, "step": 223150 }, { "epoch": 1.9710697822285805, "grad_norm": 4.410917282104492, "learning_rate": 6.649320889781356e-05, "loss": 2.7429, "step": 223200 }, { "epoch": 1.9715113301188647, "grad_norm": 2.3495771884918213, "learning_rate": 6.648008987296878e-05, "loss": 3.1646, "step": 223250 }, { "epoch": 1.9719528780091489, "grad_norm": 4.457131385803223, "learning_rate": 6.646696957517684e-05, "loss": 3.0327, "step": 223300 }, { "epoch": 1.972394425899433, "grad_norm": 3.8774821758270264, "learning_rate": 6.645384800545112e-05, "loss": 3.4975, "step": 223350 }, { "epoch": 1.9728359737897172, "grad_norm": 1.317435622215271, "learning_rate": 6.644072516480517e-05, "loss": 3.2134, "step": 223400 }, { "epoch": 1.9732775216800014, "grad_norm": 0.6410670280456543, "learning_rate": 6.642760105425261e-05, "loss": 2.9508, "step": 223450 }, { "epoch": 1.9737190695702855, "grad_norm": 1.857003927230835, "learning_rate": 6.641447567480717e-05, "loss": 2.7722, "step": 223500 }, { "epoch": 1.9741606174605697, "grad_norm": 3.14223575592041, "learning_rate": 6.640134902748267e-05, "loss": 3.4729, "step": 223550 }, { "epoch": 1.9746021653508539, "grad_norm": 2.031442165374756, "learning_rate": 6.638822111329304e-05, "loss": 3.0931, "step": 223600 }, { "epoch": 1.975043713241138, "grad_norm": 1.728389024734497, "learning_rate": 6.637509193325231e-05, "loss": 3.4318, "step": 223650 }, { "epoch": 1.9754852611314222, "grad_norm": 3.0955426692962646, "learning_rate": 6.636196148837456e-05, "loss": 3.1537, "step": 223700 }, { "epoch": 1.9759268090217064, "grad_norm": 3.6021342277526855, "learning_rate": 6.634882977967406e-05, "loss": 3.1403, "step": 223750 }, { "epoch": 1.9763683569119905, "grad_norm": 4.108619213104248, "learning_rate": 6.633569680816509e-05, "loss": 3.1472, "step": 223800 }, { "epoch": 1.9768099048022747, "grad_norm": 3.5418221950531006, "learning_rate": 6.632256257486204e-05, "loss": 3.197, "step": 223850 }, { "epoch": 1.977251452692559, "grad_norm": 2.8276524543762207, "learning_rate": 6.630942708077947e-05, "loss": 3.3344, "step": 223900 }, { "epoch": 1.9776930005828433, "grad_norm": 3.096320867538452, "learning_rate": 6.629629032693195e-05, "loss": 2.9784, "step": 223950 }, { "epoch": 1.9781345484731274, "grad_norm": 1.5722672939300537, "learning_rate": 6.628315231433418e-05, "loss": 3.1677, "step": 224000 }, { "epoch": 1.9785760963634116, "grad_norm": 2.9355618953704834, "learning_rate": 6.627001304400098e-05, "loss": 2.979, "step": 224050 }, { "epoch": 1.9790176442536958, "grad_norm": 3.031902313232422, "learning_rate": 6.625687251694721e-05, "loss": 2.9146, "step": 224100 }, { "epoch": 1.97945919214398, "grad_norm": 2.6479899883270264, "learning_rate": 6.624373073418793e-05, "loss": 2.7404, "step": 224150 }, { "epoch": 1.979900740034264, "grad_norm": 2.275542974472046, "learning_rate": 6.623058769673819e-05, "loss": 3.2156, "step": 224200 }, { "epoch": 1.9803422879245483, "grad_norm": 1.3600300550460815, "learning_rate": 6.621744340561315e-05, "loss": 3.1979, "step": 224250 }, { "epoch": 1.9807838358148326, "grad_norm": 1.7294834852218628, "learning_rate": 6.620429786182816e-05, "loss": 3.1528, "step": 224300 }, { "epoch": 1.9812253837051168, "grad_norm": 1.5742638111114502, "learning_rate": 6.619115106639854e-05, "loss": 3.531, "step": 224350 }, { "epoch": 1.981666931595401, "grad_norm": 2.426547050476074, "learning_rate": 6.617800302033983e-05, "loss": 3.0334, "step": 224400 }, { "epoch": 1.9821084794856851, "grad_norm": 1.9069656133651733, "learning_rate": 6.616485372466756e-05, "loss": 3.4618, "step": 224450 }, { "epoch": 1.9825500273759693, "grad_norm": 4.325615882873535, "learning_rate": 6.615170318039741e-05, "loss": 3.1056, "step": 224500 }, { "epoch": 1.9829915752662535, "grad_norm": 2.732570171356201, "learning_rate": 6.613855138854514e-05, "loss": 3.4351, "step": 224550 }, { "epoch": 1.9834331231565376, "grad_norm": 3.881892681121826, "learning_rate": 6.612539835012664e-05, "loss": 3.6275, "step": 224600 }, { "epoch": 1.9838746710468218, "grad_norm": 4.55607271194458, "learning_rate": 6.611224406615786e-05, "loss": 3.1314, "step": 224650 }, { "epoch": 1.984316218937106, "grad_norm": 2.960299015045166, "learning_rate": 6.609908853765485e-05, "loss": 3.1012, "step": 224700 }, { "epoch": 1.9847577668273901, "grad_norm": 4.393443584442139, "learning_rate": 6.608593176563377e-05, "loss": 3.4316, "step": 224750 }, { "epoch": 1.9851993147176743, "grad_norm": 3.698000907897949, "learning_rate": 6.607277375111086e-05, "loss": 3.4583, "step": 224800 }, { "epoch": 1.9856408626079585, "grad_norm": 2.1111578941345215, "learning_rate": 6.605961449510248e-05, "loss": 3.3533, "step": 224850 }, { "epoch": 1.9860824104982426, "grad_norm": 2.465977668762207, "learning_rate": 6.604645399862504e-05, "loss": 3.1043, "step": 224900 }, { "epoch": 1.9865239583885268, "grad_norm": 2.8451077938079834, "learning_rate": 6.603329226269511e-05, "loss": 3.1937, "step": 224950 }, { "epoch": 1.986965506278811, "grad_norm": 0.7063005566596985, "learning_rate": 6.602012928832932e-05, "loss": 2.9482, "step": 225000 }, { "epoch": 1.986965506278811, "eval_asr_loss": 0.9346238268861363, "eval_loss": 2.8609893321990967, "eval_runtime": 20.7029, "eval_samples_per_second": 37.096, "eval_steps_per_second": 9.274, "eval_tts_loss": 5.910735508017596, "step": 225000 }, { "epoch": 1.9874070541690951, "grad_norm": 5.629608631134033, "learning_rate": 6.600696507654439e-05, "loss": 3.2127, "step": 225050 }, { "epoch": 1.9878486020593793, "grad_norm": 2.0940353870391846, "learning_rate": 6.599379962835713e-05, "loss": 3.0928, "step": 225100 }, { "epoch": 1.9882901499496635, "grad_norm": 0.8996180891990662, "learning_rate": 6.598063294478448e-05, "loss": 3.1409, "step": 225150 }, { "epoch": 1.9887316978399476, "grad_norm": 5.146057605743408, "learning_rate": 6.596746502684345e-05, "loss": 2.87, "step": 225200 }, { "epoch": 1.9891732457302318, "grad_norm": 6.451409339904785, "learning_rate": 6.595429587555114e-05, "loss": 3.2195, "step": 225250 }, { "epoch": 1.989614793620516, "grad_norm": 1.7646028995513916, "learning_rate": 6.594112549192475e-05, "loss": 3.5659, "step": 225300 }, { "epoch": 1.9900563415108001, "grad_norm": 1.4490691423416138, "learning_rate": 6.59279538769816e-05, "loss": 3.3229, "step": 225350 }, { "epoch": 1.9904978894010843, "grad_norm": 2.59922194480896, "learning_rate": 6.591478103173909e-05, "loss": 2.8994, "step": 225400 }, { "epoch": 1.9909394372913685, "grad_norm": 1.5313228368759155, "learning_rate": 6.59016069572147e-05, "loss": 3.2554, "step": 225450 }, { "epoch": 1.9913809851816529, "grad_norm": 5.3769917488098145, "learning_rate": 6.588843165442599e-05, "loss": 3.1764, "step": 225500 }, { "epoch": 1.991822533071937, "grad_norm": 3.157618522644043, "learning_rate": 6.58752551243907e-05, "loss": 3.2877, "step": 225550 }, { "epoch": 1.9922640809622212, "grad_norm": 1.7196130752563477, "learning_rate": 6.586207736812655e-05, "loss": 2.787, "step": 225600 }, { "epoch": 1.9927056288525054, "grad_norm": 0.6843865513801575, "learning_rate": 6.584889838665143e-05, "loss": 3.2724, "step": 225650 }, { "epoch": 1.9931471767427895, "grad_norm": 2.9632790088653564, "learning_rate": 6.583571818098331e-05, "loss": 3.3264, "step": 225700 }, { "epoch": 1.9935887246330737, "grad_norm": 2.421201705932617, "learning_rate": 6.582253675214023e-05, "loss": 2.9584, "step": 225750 }, { "epoch": 1.9940302725233578, "grad_norm": 5.2676496505737305, "learning_rate": 6.580935410114037e-05, "loss": 3.5463, "step": 225800 }, { "epoch": 1.994471820413642, "grad_norm": 1.7726913690567017, "learning_rate": 6.579617022900196e-05, "loss": 3.0083, "step": 225850 }, { "epoch": 1.9949133683039264, "grad_norm": 2.7158071994781494, "learning_rate": 6.578298513674334e-05, "loss": 3.5157, "step": 225900 }, { "epoch": 1.9953549161942106, "grad_norm": 2.830840826034546, "learning_rate": 6.576979882538297e-05, "loss": 3.3464, "step": 225950 }, { "epoch": 1.9957964640844947, "grad_norm": 2.3365275859832764, "learning_rate": 6.575687505845887e-05, "loss": 3.4291, "step": 226000 }, { "epoch": 1.996238011974779, "grad_norm": 1.796595573425293, "learning_rate": 6.574368633628195e-05, "loss": 2.8841, "step": 226050 }, { "epoch": 1.996679559865063, "grad_norm": 4.268457889556885, "learning_rate": 6.573049639803877e-05, "loss": 3.2318, "step": 226100 }, { "epoch": 1.9971211077553472, "grad_norm": 2.008795738220215, "learning_rate": 6.571730524474815e-05, "loss": 2.9898, "step": 226150 }, { "epoch": 1.9975626556456314, "grad_norm": 1.9656585454940796, "learning_rate": 6.570411287742894e-05, "loss": 3.2079, "step": 226200 }, { "epoch": 1.9980042035359156, "grad_norm": 1.4856351613998413, "learning_rate": 6.569091929710021e-05, "loss": 2.7014, "step": 226250 }, { "epoch": 1.9984457514261997, "grad_norm": 0.6530186533927917, "learning_rate": 6.5677724504781e-05, "loss": 3.0105, "step": 226300 }, { "epoch": 1.998887299316484, "grad_norm": 5.405598163604736, "learning_rate": 6.566452850149052e-05, "loss": 3.3306, "step": 226350 }, { "epoch": 1.999328847206768, "grad_norm": 2.492218017578125, "learning_rate": 6.565133128824805e-05, "loss": 3.5534, "step": 226400 }, { "epoch": 1.9997703950970522, "grad_norm": 3.143587350845337, "learning_rate": 6.563813286607296e-05, "loss": 2.6712, "step": 226450 }, { "epoch": 2.0002119429873364, "grad_norm": 4.805309772491455, "learning_rate": 6.562493323598471e-05, "loss": 3.1063, "step": 226500 }, { "epoch": 2.0006534908776206, "grad_norm": 3.23760986328125, "learning_rate": 6.561173239900286e-05, "loss": 3.1377, "step": 226550 }, { "epoch": 2.0010950387679047, "grad_norm": 3.0754528045654297, "learning_rate": 6.559853035614707e-05, "loss": 3.1366, "step": 226600 }, { "epoch": 2.001536586658189, "grad_norm": 3.882094144821167, "learning_rate": 6.558532710843709e-05, "loss": 3.3212, "step": 226650 }, { "epoch": 2.001978134548473, "grad_norm": 1.7975034713745117, "learning_rate": 6.557212265689273e-05, "loss": 2.9501, "step": 226700 }, { "epoch": 2.0024196824387572, "grad_norm": 4.0580549240112305, "learning_rate": 6.555891700253396e-05, "loss": 3.0891, "step": 226750 }, { "epoch": 2.0028612303290414, "grad_norm": 2.229588747024536, "learning_rate": 6.554571014638077e-05, "loss": 2.6984, "step": 226800 }, { "epoch": 2.0033027782193256, "grad_norm": 1.334570050239563, "learning_rate": 6.553250208945332e-05, "loss": 2.8494, "step": 226850 }, { "epoch": 2.0037443261096097, "grad_norm": 4.698803901672363, "learning_rate": 6.551929283277178e-05, "loss": 2.8055, "step": 226900 }, { "epoch": 2.004185873999894, "grad_norm": 2.259545087814331, "learning_rate": 6.550608237735645e-05, "loss": 3.5065, "step": 226950 }, { "epoch": 2.004627421890178, "grad_norm": 2.018728256225586, "learning_rate": 6.549287072422778e-05, "loss": 3.5027, "step": 227000 }, { "epoch": 2.0050689697804622, "grad_norm": 3.1825854778289795, "learning_rate": 6.547965787440621e-05, "loss": 3.5209, "step": 227050 }, { "epoch": 2.0055105176707464, "grad_norm": 2.5651297569274902, "learning_rate": 6.546644382891232e-05, "loss": 3.0002, "step": 227100 }, { "epoch": 2.0059520655610306, "grad_norm": 2.613448143005371, "learning_rate": 6.54532285887668e-05, "loss": 3.1573, "step": 227150 }, { "epoch": 2.0063936134513147, "grad_norm": 2.843972682952881, "learning_rate": 6.54400121549904e-05, "loss": 3.059, "step": 227200 }, { "epoch": 2.006835161341599, "grad_norm": 2.9624767303466797, "learning_rate": 6.542679452860404e-05, "loss": 2.7594, "step": 227250 }, { "epoch": 2.0072767092318835, "grad_norm": 2.1793372631073, "learning_rate": 6.541357571062856e-05, "loss": 3.4074, "step": 227300 }, { "epoch": 2.0077182571221677, "grad_norm": 1.0161021947860718, "learning_rate": 6.540035570208509e-05, "loss": 2.8815, "step": 227350 }, { "epoch": 2.008159805012452, "grad_norm": 4.125101089477539, "learning_rate": 6.538713450399472e-05, "loss": 2.9716, "step": 227400 }, { "epoch": 2.008601352902736, "grad_norm": 3.0919394493103027, "learning_rate": 6.537391211737868e-05, "loss": 3.2971, "step": 227450 }, { "epoch": 2.00904290079302, "grad_norm": 3.4798877239227295, "learning_rate": 6.536068854325832e-05, "loss": 3.0346, "step": 227500 }, { "epoch": 2.0094844486833043, "grad_norm": 2.032561779022217, "learning_rate": 6.534746378265502e-05, "loss": 2.8012, "step": 227550 }, { "epoch": 2.0099259965735885, "grad_norm": 4.441436290740967, "learning_rate": 6.533423783659027e-05, "loss": 3.1085, "step": 227600 }, { "epoch": 2.0103675444638727, "grad_norm": 2.3486709594726562, "learning_rate": 6.53210107060857e-05, "loss": 3.292, "step": 227650 }, { "epoch": 2.010809092354157, "grad_norm": 3.3359105587005615, "learning_rate": 6.530778239216294e-05, "loss": 3.2894, "step": 227700 }, { "epoch": 2.011250640244441, "grad_norm": 1.950713038444519, "learning_rate": 6.529455289584382e-05, "loss": 3.3267, "step": 227750 }, { "epoch": 2.011692188134725, "grad_norm": 4.134000778198242, "learning_rate": 6.528132221815018e-05, "loss": 3.3464, "step": 227800 }, { "epoch": 2.0121337360250093, "grad_norm": 5.741815567016602, "learning_rate": 6.526809036010397e-05, "loss": 2.8544, "step": 227850 }, { "epoch": 2.0125752839152935, "grad_norm": 4.831362247467041, "learning_rate": 6.525512199502562e-05, "loss": 2.9295, "step": 227900 }, { "epoch": 2.0130168318055777, "grad_norm": 3.7244808673858643, "learning_rate": 6.52418878028967e-05, "loss": 3.2785, "step": 227950 }, { "epoch": 2.013458379695862, "grad_norm": 1.5436469316482544, "learning_rate": 6.522865243346119e-05, "loss": 3.2067, "step": 228000 }, { "epoch": 2.013458379695862, "eval_asr_loss": 0.9319613288593795, "eval_loss": 2.857133626937866, "eval_runtime": 20.3432, "eval_samples_per_second": 37.752, "eval_steps_per_second": 9.438, "eval_tts_loss": 6.035552883036707, "step": 228000 }, { "epoch": 2.013899927586146, "grad_norm": 3.8790123462677, "learning_rate": 6.521541588774139e-05, "loss": 2.8997, "step": 228050 }, { "epoch": 2.01434147547643, "grad_norm": 5.026811599731445, "learning_rate": 6.520217816675975e-05, "loss": 2.8402, "step": 228100 }, { "epoch": 2.0147830233667143, "grad_norm": 2.0185930728912354, "learning_rate": 6.518893927153876e-05, "loss": 2.6189, "step": 228150 }, { "epoch": 2.0152245712569985, "grad_norm": 1.0025949478149414, "learning_rate": 6.517569920310102e-05, "loss": 2.6175, "step": 228200 }, { "epoch": 2.0156661191472827, "grad_norm": 1.5849015712738037, "learning_rate": 6.516245796246919e-05, "loss": 3.1103, "step": 228250 }, { "epoch": 2.016107667037567, "grad_norm": 2.3523590564727783, "learning_rate": 6.514921555066604e-05, "loss": 3.604, "step": 228300 }, { "epoch": 2.016549214927851, "grad_norm": 3.9185702800750732, "learning_rate": 6.513597196871448e-05, "loss": 2.7966, "step": 228350 }, { "epoch": 2.016990762818135, "grad_norm": 1.396605372428894, "learning_rate": 6.512272721763743e-05, "loss": 3.2534, "step": 228400 }, { "epoch": 2.0174323107084193, "grad_norm": 1.2401235103607178, "learning_rate": 6.510948129845793e-05, "loss": 2.8992, "step": 228450 }, { "epoch": 2.0178738585987035, "grad_norm": 2.9374325275421143, "learning_rate": 6.509623421219913e-05, "loss": 2.8456, "step": 228500 }, { "epoch": 2.0183154064889877, "grad_norm": 0.8914576768875122, "learning_rate": 6.508298595988425e-05, "loss": 2.9675, "step": 228550 }, { "epoch": 2.018756954379272, "grad_norm": 1.5249571800231934, "learning_rate": 6.506973654253658e-05, "loss": 3.0548, "step": 228600 }, { "epoch": 2.019198502269556, "grad_norm": 1.9931347370147705, "learning_rate": 6.505648596117956e-05, "loss": 3.2147, "step": 228650 }, { "epoch": 2.01964005015984, "grad_norm": 2.841373920440674, "learning_rate": 6.50432342168367e-05, "loss": 3.4644, "step": 228700 }, { "epoch": 2.0200815980501243, "grad_norm": 2.344156503677368, "learning_rate": 6.502998131053155e-05, "loss": 3.1103, "step": 228750 }, { "epoch": 2.0205231459404085, "grad_norm": 3.8044936656951904, "learning_rate": 6.50167272432878e-05, "loss": 3.1707, "step": 228800 }, { "epoch": 2.0209646938306927, "grad_norm": 2.157287120819092, "learning_rate": 6.500347201612919e-05, "loss": 3.36, "step": 228850 }, { "epoch": 2.0214062417209773, "grad_norm": 1.3241900205612183, "learning_rate": 6.49902156300796e-05, "loss": 2.7496, "step": 228900 }, { "epoch": 2.0218477896112614, "grad_norm": 4.144221305847168, "learning_rate": 6.497695808616296e-05, "loss": 2.9482, "step": 228950 }, { "epoch": 2.0222893375015456, "grad_norm": 2.7052125930786133, "learning_rate": 6.49636993854033e-05, "loss": 2.9198, "step": 229000 }, { "epoch": 2.0227308853918298, "grad_norm": 2.5557072162628174, "learning_rate": 6.495043952882477e-05, "loss": 3.2878, "step": 229050 }, { "epoch": 2.023172433282114, "grad_norm": 2.4937329292297363, "learning_rate": 6.493717851745157e-05, "loss": 3.0772, "step": 229100 }, { "epoch": 2.023613981172398, "grad_norm": 1.7631187438964844, "learning_rate": 6.492391635230797e-05, "loss": 2.9628, "step": 229150 }, { "epoch": 2.0240555290626823, "grad_norm": 2.703622341156006, "learning_rate": 6.49106530344184e-05, "loss": 3.2114, "step": 229200 }, { "epoch": 2.0244970769529664, "grad_norm": 1.6925734281539917, "learning_rate": 6.489738856480731e-05, "loss": 3.0406, "step": 229250 }, { "epoch": 2.0249386248432506, "grad_norm": 2.586226224899292, "learning_rate": 6.488412294449929e-05, "loss": 2.7732, "step": 229300 }, { "epoch": 2.0253801727335348, "grad_norm": 2.5252676010131836, "learning_rate": 6.487085617451898e-05, "loss": 2.9559, "step": 229350 }, { "epoch": 2.025821720623819, "grad_norm": 4.164633274078369, "learning_rate": 6.485758825589113e-05, "loss": 2.6783, "step": 229400 }, { "epoch": 2.026263268514103, "grad_norm": 3.092494010925293, "learning_rate": 6.484431918964058e-05, "loss": 3.1991, "step": 229450 }, { "epoch": 2.0267048164043873, "grad_norm": 2.40688419342041, "learning_rate": 6.483104897679227e-05, "loss": 3.2162, "step": 229500 }, { "epoch": 2.0271463642946714, "grad_norm": 5.002386093139648, "learning_rate": 6.481777761837116e-05, "loss": 2.9303, "step": 229550 }, { "epoch": 2.0275879121849556, "grad_norm": 1.425470232963562, "learning_rate": 6.480450511540238e-05, "loss": 2.9637, "step": 229600 }, { "epoch": 2.0280294600752398, "grad_norm": 2.9680776596069336, "learning_rate": 6.479123146891112e-05, "loss": 3.3259, "step": 229650 }, { "epoch": 2.028471007965524, "grad_norm": 5.0673041343688965, "learning_rate": 6.477795667992267e-05, "loss": 3.1311, "step": 229700 }, { "epoch": 2.028912555855808, "grad_norm": 1.1483591794967651, "learning_rate": 6.476468074946238e-05, "loss": 3.1779, "step": 229750 }, { "epoch": 2.0293541037460923, "grad_norm": 3.565528631210327, "learning_rate": 6.475140367855567e-05, "loss": 3.0531, "step": 229800 }, { "epoch": 2.0297956516363764, "grad_norm": 1.5797202587127686, "learning_rate": 6.473812546822815e-05, "loss": 3.2312, "step": 229850 }, { "epoch": 2.0302371995266606, "grad_norm": 3.3140475749969482, "learning_rate": 6.47248461195054e-05, "loss": 3.2321, "step": 229900 }, { "epoch": 2.0306787474169448, "grad_norm": 4.104597568511963, "learning_rate": 6.471156563341317e-05, "loss": 3.2755, "step": 229950 }, { "epoch": 2.031120295307229, "grad_norm": 2.909299850463867, "learning_rate": 6.469828401097722e-05, "loss": 2.8798, "step": 230000 }, { "epoch": 2.031561843197513, "grad_norm": 1.7076945304870605, "learning_rate": 6.468500125322346e-05, "loss": 3.3038, "step": 230050 }, { "epoch": 2.0320033910877973, "grad_norm": 4.370232582092285, "learning_rate": 6.467171736117791e-05, "loss": 2.9408, "step": 230100 }, { "epoch": 2.0324449389780814, "grad_norm": 2.5762276649475098, "learning_rate": 6.465843233586657e-05, "loss": 2.9553, "step": 230150 }, { "epoch": 2.0328864868683656, "grad_norm": 3.14599347114563, "learning_rate": 6.464514617831567e-05, "loss": 3.0772, "step": 230200 }, { "epoch": 2.0333280347586498, "grad_norm": 2.719356060028076, "learning_rate": 6.463185888955138e-05, "loss": 2.985, "step": 230250 }, { "epoch": 2.033769582648934, "grad_norm": 3.8772943019866943, "learning_rate": 6.461857047060009e-05, "loss": 3.143, "step": 230300 }, { "epoch": 2.034211130539218, "grad_norm": 2.380610704421997, "learning_rate": 6.460528092248819e-05, "loss": 2.863, "step": 230350 }, { "epoch": 2.0346526784295023, "grad_norm": 4.021361827850342, "learning_rate": 6.459199024624219e-05, "loss": 2.7805, "step": 230400 }, { "epoch": 2.0350942263197864, "grad_norm": 3.375150442123413, "learning_rate": 6.457869844288867e-05, "loss": 2.8781, "step": 230450 }, { "epoch": 2.035535774210071, "grad_norm": 6.324542045593262, "learning_rate": 6.456540551345431e-05, "loss": 2.6493, "step": 230500 }, { "epoch": 2.035977322100355, "grad_norm": 2.199507474899292, "learning_rate": 6.455211145896589e-05, "loss": 3.4413, "step": 230550 }, { "epoch": 2.0364188699906394, "grad_norm": 1.475187063217163, "learning_rate": 6.453881628045028e-05, "loss": 3.3786, "step": 230600 }, { "epoch": 2.0368604178809235, "grad_norm": 1.968745470046997, "learning_rate": 6.452551997893438e-05, "loss": 3.3907, "step": 230650 }, { "epoch": 2.0373019657712077, "grad_norm": 3.9331090450286865, "learning_rate": 6.451222255544523e-05, "loss": 3.0484, "step": 230700 }, { "epoch": 2.037743513661492, "grad_norm": 4.076101779937744, "learning_rate": 6.449892401100996e-05, "loss": 3.0525, "step": 230750 }, { "epoch": 2.038185061551776, "grad_norm": 1.4716242551803589, "learning_rate": 6.448562434665575e-05, "loss": 3.0944, "step": 230800 }, { "epoch": 2.03862660944206, "grad_norm": 3.958530902862549, "learning_rate": 6.447232356340991e-05, "loss": 3.3774, "step": 230850 }, { "epoch": 2.0390681573323444, "grad_norm": 3.1192853450775146, "learning_rate": 6.445902166229978e-05, "loss": 3.4414, "step": 230900 }, { "epoch": 2.0395097052226285, "grad_norm": 1.0482338666915894, "learning_rate": 6.444571864435283e-05, "loss": 3.15, "step": 230950 }, { "epoch": 2.0399512531129127, "grad_norm": 1.4265995025634766, "learning_rate": 6.443241451059662e-05, "loss": 2.9373, "step": 231000 }, { "epoch": 2.0399512531129127, "eval_asr_loss": 0.9325178723528175, "eval_loss": 2.8670003414154053, "eval_runtime": 20.813, "eval_samples_per_second": 36.9, "eval_steps_per_second": 9.225, "eval_tts_loss": 6.1085950902518045, "step": 231000 }, { "epoch": 2.040392801003197, "grad_norm": 1.4011311531066895, "learning_rate": 6.441910926205877e-05, "loss": 2.8658, "step": 231050 }, { "epoch": 2.040834348893481, "grad_norm": 1.7842819690704346, "learning_rate": 6.440580289976701e-05, "loss": 2.9198, "step": 231100 }, { "epoch": 2.041275896783765, "grad_norm": 1.5047529935836792, "learning_rate": 6.439249542474912e-05, "loss": 2.9665, "step": 231150 }, { "epoch": 2.0417174446740494, "grad_norm": 1.1403131484985352, "learning_rate": 6.4379186838033e-05, "loss": 2.9217, "step": 231200 }, { "epoch": 2.0421589925643335, "grad_norm": 2.5434775352478027, "learning_rate": 6.436587714064665e-05, "loss": 3.1017, "step": 231250 }, { "epoch": 2.0426005404546177, "grad_norm": 4.1038312911987305, "learning_rate": 6.43525663336181e-05, "loss": 3.3316, "step": 231300 }, { "epoch": 2.043042088344902, "grad_norm": 3.0226967334747314, "learning_rate": 6.433925441797549e-05, "loss": 2.7633, "step": 231350 }, { "epoch": 2.043483636235186, "grad_norm": 2.997720718383789, "learning_rate": 6.432594139474709e-05, "loss": 2.9086, "step": 231400 }, { "epoch": 2.04392518412547, "grad_norm": 1.9566609859466553, "learning_rate": 6.43126272649612e-05, "loss": 2.9525, "step": 231450 }, { "epoch": 2.0443667320157544, "grad_norm": 2.1925888061523438, "learning_rate": 6.429931202964621e-05, "loss": 3.3177, "step": 231500 }, { "epoch": 2.0448082799060385, "grad_norm": 3.2015256881713867, "learning_rate": 6.428599568983062e-05, "loss": 2.914, "step": 231550 }, { "epoch": 2.0452498277963227, "grad_norm": 3.4477548599243164, "learning_rate": 6.427267824654299e-05, "loss": 3.3198, "step": 231600 }, { "epoch": 2.045691375686607, "grad_norm": 2.724602222442627, "learning_rate": 6.425935970081203e-05, "loss": 2.8969, "step": 231650 }, { "epoch": 2.046132923576891, "grad_norm": 3.4831771850585938, "learning_rate": 6.424604005366642e-05, "loss": 3.0436, "step": 231700 }, { "epoch": 2.046574471467175, "grad_norm": 2.3419885635375977, "learning_rate": 6.423271930613503e-05, "loss": 3.4779, "step": 231750 }, { "epoch": 2.0470160193574594, "grad_norm": 2.4602243900299072, "learning_rate": 6.421939745924677e-05, "loss": 2.6744, "step": 231800 }, { "epoch": 2.0474575672477435, "grad_norm": 6.231406211853027, "learning_rate": 6.420607451403062e-05, "loss": 2.7832, "step": 231850 }, { "epoch": 2.0478991151380277, "grad_norm": 1.2035382986068726, "learning_rate": 6.419275047151569e-05, "loss": 3.0244, "step": 231900 }, { "epoch": 2.048340663028312, "grad_norm": 4.7911481857299805, "learning_rate": 6.417942533273111e-05, "loss": 2.9757, "step": 231950 }, { "epoch": 2.048782210918596, "grad_norm": 3.543363332748413, "learning_rate": 6.416609909870619e-05, "loss": 2.6808, "step": 232000 }, { "epoch": 2.04922375880888, "grad_norm": 2.8009135723114014, "learning_rate": 6.415277177047025e-05, "loss": 2.9496, "step": 232050 }, { "epoch": 2.049665306699165, "grad_norm": 4.706131458282471, "learning_rate": 6.413944334905268e-05, "loss": 2.9122, "step": 232100 }, { "epoch": 2.050106854589449, "grad_norm": 3.3189642429351807, "learning_rate": 6.412611383548301e-05, "loss": 3.0313, "step": 232150 }, { "epoch": 2.050548402479733, "grad_norm": 3.766918182373047, "learning_rate": 6.411304985357103e-05, "loss": 3.2675, "step": 232200 }, { "epoch": 2.0509899503700173, "grad_norm": 1.4554075002670288, "learning_rate": 6.409971818057778e-05, "loss": 3.2086, "step": 232250 }, { "epoch": 2.0514314982603015, "grad_norm": 5.887540340423584, "learning_rate": 6.408638541850088e-05, "loss": 2.8141, "step": 232300 }, { "epoch": 2.0518730461505856, "grad_norm": 4.158108711242676, "learning_rate": 6.407305156837013e-05, "loss": 3.099, "step": 232350 }, { "epoch": 2.05231459404087, "grad_norm": 2.646940231323242, "learning_rate": 6.40597166312155e-05, "loss": 3.4103, "step": 232400 }, { "epoch": 2.052756141931154, "grad_norm": 1.4084917306900024, "learning_rate": 6.4046380608067e-05, "loss": 3.0681, "step": 232450 }, { "epoch": 2.053197689821438, "grad_norm": 3.3093457221984863, "learning_rate": 6.403304349995468e-05, "loss": 3.1133, "step": 232500 }, { "epoch": 2.0536392377117223, "grad_norm": 7.799812316894531, "learning_rate": 6.401970530790877e-05, "loss": 3.3527, "step": 232550 }, { "epoch": 2.0540807856020065, "grad_norm": 4.245431900024414, "learning_rate": 6.40063660329595e-05, "loss": 2.9631, "step": 232600 }, { "epoch": 2.0545223334922906, "grad_norm": 2.25030517578125, "learning_rate": 6.399302567613721e-05, "loss": 3.3969, "step": 232650 }, { "epoch": 2.054963881382575, "grad_norm": 1.1634451150894165, "learning_rate": 6.397968423847236e-05, "loss": 3.495, "step": 232700 }, { "epoch": 2.055405429272859, "grad_norm": 2.125120162963867, "learning_rate": 6.396634172099544e-05, "loss": 2.9466, "step": 232750 }, { "epoch": 2.055846977163143, "grad_norm": 2.8387858867645264, "learning_rate": 6.395299812473705e-05, "loss": 2.7765, "step": 232800 }, { "epoch": 2.0562885250534273, "grad_norm": 3.0474061965942383, "learning_rate": 6.39396534507279e-05, "loss": 3.199, "step": 232850 }, { "epoch": 2.0567300729437115, "grad_norm": 5.258384704589844, "learning_rate": 6.392630769999868e-05, "loss": 3.1556, "step": 232900 }, { "epoch": 2.0571716208339956, "grad_norm": 2.827653169631958, "learning_rate": 6.391296087358032e-05, "loss": 2.6953, "step": 232950 }, { "epoch": 2.05761316872428, "grad_norm": 2.8440933227539062, "learning_rate": 6.389961297250368e-05, "loss": 3.2708, "step": 233000 }, { "epoch": 2.058054716614564, "grad_norm": 3.5404655933380127, "learning_rate": 6.388626399779982e-05, "loss": 3.0693, "step": 233050 }, { "epoch": 2.058496264504848, "grad_norm": 1.7443832159042358, "learning_rate": 6.387291395049982e-05, "loss": 3.4108, "step": 233100 }, { "epoch": 2.0589378123951323, "grad_norm": 5.929326057434082, "learning_rate": 6.385956283163485e-05, "loss": 3.2485, "step": 233150 }, { "epoch": 2.0593793602854165, "grad_norm": 1.6079530715942383, "learning_rate": 6.38462106422362e-05, "loss": 3.4657, "step": 233200 }, { "epoch": 2.0598209081757006, "grad_norm": 3.8741631507873535, "learning_rate": 6.383285738333515e-05, "loss": 2.5408, "step": 233250 }, { "epoch": 2.060262456065985, "grad_norm": 3.9816033840179443, "learning_rate": 6.38195030559632e-05, "loss": 3.1404, "step": 233300 }, { "epoch": 2.060704003956269, "grad_norm": 1.1695952415466309, "learning_rate": 6.380614766115181e-05, "loss": 3.1305, "step": 233350 }, { "epoch": 2.061145551846553, "grad_norm": 1.1502101421356201, "learning_rate": 6.379279119993259e-05, "loss": 2.7873, "step": 233400 }, { "epoch": 2.0615870997368373, "grad_norm": 2.502312660217285, "learning_rate": 6.37794336733372e-05, "loss": 2.8025, "step": 233450 }, { "epoch": 2.0620286476271215, "grad_norm": 2.5723695755004883, "learning_rate": 6.376607508239742e-05, "loss": 3.1278, "step": 233500 }, { "epoch": 2.0624701955174056, "grad_norm": 1.911758542060852, "learning_rate": 6.375271542814507e-05, "loss": 3.3527, "step": 233550 }, { "epoch": 2.06291174340769, "grad_norm": 3.6591827869415283, "learning_rate": 6.373935471161206e-05, "loss": 2.9983, "step": 233600 }, { "epoch": 2.0633532912979744, "grad_norm": 1.4186557531356812, "learning_rate": 6.372599293383042e-05, "loss": 3.1304, "step": 233650 }, { "epoch": 2.0637948391882586, "grad_norm": 3.396662950515747, "learning_rate": 6.371263009583222e-05, "loss": 3.1124, "step": 233700 }, { "epoch": 2.0642363870785427, "grad_norm": 3.51399302482605, "learning_rate": 6.36992661986496e-05, "loss": 3.1643, "step": 233750 }, { "epoch": 2.064677934968827, "grad_norm": 1.7544376850128174, "learning_rate": 6.368590124331486e-05, "loss": 3.1829, "step": 233800 }, { "epoch": 2.065119482859111, "grad_norm": 3.821058750152588, "learning_rate": 6.367253523086028e-05, "loss": 2.7627, "step": 233850 }, { "epoch": 2.0655610307493952, "grad_norm": 3.183556079864502, "learning_rate": 6.365916816231832e-05, "loss": 2.8705, "step": 233900 }, { "epoch": 2.0660025786396794, "grad_norm": 3.5835275650024414, "learning_rate": 6.364580003872142e-05, "loss": 2.798, "step": 233950 }, { "epoch": 2.0664441265299636, "grad_norm": 2.4540653228759766, "learning_rate": 6.36324308611022e-05, "loss": 2.8802, "step": 234000 }, { "epoch": 2.0664441265299636, "eval_asr_loss": 0.9346235701288164, "eval_loss": 2.8734447956085205, "eval_runtime": 20.7109, "eval_samples_per_second": 37.082, "eval_steps_per_second": 9.27, "eval_tts_loss": 6.045795924074517, "step": 234000 }, { "epoch": 2.0668856744202477, "grad_norm": 1.7151188850402832, "learning_rate": 6.361906063049328e-05, "loss": 3.08, "step": 234050 }, { "epoch": 2.067327222310532, "grad_norm": 3.399620771408081, "learning_rate": 6.360568934792743e-05, "loss": 3.4007, "step": 234100 }, { "epoch": 2.067768770200816, "grad_norm": 2.422654151916504, "learning_rate": 6.359231701443742e-05, "loss": 3.1204, "step": 234150 }, { "epoch": 2.0682103180911002, "grad_norm": 4.625797748565674, "learning_rate": 6.357894363105621e-05, "loss": 2.9898, "step": 234200 }, { "epoch": 2.0686518659813844, "grad_norm": 1.8986616134643555, "learning_rate": 6.356556919881673e-05, "loss": 2.9439, "step": 234250 }, { "epoch": 2.0690934138716686, "grad_norm": 1.4936383962631226, "learning_rate": 6.355219371875208e-05, "loss": 2.9849, "step": 234300 }, { "epoch": 2.0695349617619527, "grad_norm": 3.0682079792022705, "learning_rate": 6.353881719189538e-05, "loss": 3.2306, "step": 234350 }, { "epoch": 2.069976509652237, "grad_norm": 1.3556572198867798, "learning_rate": 6.352570718097391e-05, "loss": 3.268, "step": 234400 }, { "epoch": 2.070418057542521, "grad_norm": 1.753355860710144, "learning_rate": 6.351232858451726e-05, "loss": 3.4784, "step": 234450 }, { "epoch": 2.0708596054328052, "grad_norm": 5.882805347442627, "learning_rate": 6.349894894434779e-05, "loss": 2.8673, "step": 234500 }, { "epoch": 2.0713011533230894, "grad_norm": 3.6827399730682373, "learning_rate": 6.348556826149898e-05, "loss": 3.0578, "step": 234550 }, { "epoch": 2.0717427012133736, "grad_norm": 6.412780284881592, "learning_rate": 6.34721865370044e-05, "loss": 3.0419, "step": 234600 }, { "epoch": 2.0721842491036577, "grad_norm": 1.9119713306427002, "learning_rate": 6.345880377189763e-05, "loss": 3.4086, "step": 234650 }, { "epoch": 2.072625796993942, "grad_norm": 2.772346258163452, "learning_rate": 6.34454199672124e-05, "loss": 3.0575, "step": 234700 }, { "epoch": 2.073067344884226, "grad_norm": 4.332114219665527, "learning_rate": 6.34320351239825e-05, "loss": 3.135, "step": 234750 }, { "epoch": 2.0735088927745102, "grad_norm": 2.814565658569336, "learning_rate": 6.341864924324178e-05, "loss": 3.1281, "step": 234800 }, { "epoch": 2.0739504406647944, "grad_norm": 6.352789402008057, "learning_rate": 6.340526232602419e-05, "loss": 3.1765, "step": 234850 }, { "epoch": 2.0743919885550786, "grad_norm": 4.205619812011719, "learning_rate": 6.339187437336376e-05, "loss": 2.7536, "step": 234900 }, { "epoch": 2.0748335364453627, "grad_norm": 2.4051403999328613, "learning_rate": 6.337848538629458e-05, "loss": 2.8179, "step": 234950 }, { "epoch": 2.075275084335647, "grad_norm": 2.8886311054229736, "learning_rate": 6.336509536585089e-05, "loss": 2.9021, "step": 235000 }, { "epoch": 2.075716632225931, "grad_norm": 3.565911054611206, "learning_rate": 6.335170431306689e-05, "loss": 2.9569, "step": 235050 }, { "epoch": 2.0761581801162152, "grad_norm": 3.9193966388702393, "learning_rate": 6.333831222897695e-05, "loss": 3.364, "step": 235100 }, { "epoch": 2.0765997280064994, "grad_norm": 3.124962091445923, "learning_rate": 6.33249191146155e-05, "loss": 3.3793, "step": 235150 }, { "epoch": 2.0770412758967836, "grad_norm": 2.5896620750427246, "learning_rate": 6.331152497101705e-05, "loss": 3.0124, "step": 235200 }, { "epoch": 2.0774828237870677, "grad_norm": 1.9518678188323975, "learning_rate": 6.329812979921615e-05, "loss": 3.1021, "step": 235250 }, { "epoch": 2.0779243716773523, "grad_norm": 1.719233512878418, "learning_rate": 6.32847336002475e-05, "loss": 3.1602, "step": 235300 }, { "epoch": 2.0783659195676365, "grad_norm": 2.6694281101226807, "learning_rate": 6.327133637514583e-05, "loss": 2.9971, "step": 235350 }, { "epoch": 2.0788074674579207, "grad_norm": 3.187509775161743, "learning_rate": 6.325793812494598e-05, "loss": 2.6088, "step": 235400 }, { "epoch": 2.079249015348205, "grad_norm": 4.0148606300354, "learning_rate": 6.32445388506828e-05, "loss": 3.0347, "step": 235450 }, { "epoch": 2.079690563238489, "grad_norm": 2.926313877105713, "learning_rate": 6.323113855339134e-05, "loss": 3.0781, "step": 235500 }, { "epoch": 2.080132111128773, "grad_norm": 1.753333568572998, "learning_rate": 6.32177372341066e-05, "loss": 2.9386, "step": 235550 }, { "epoch": 2.0805736590190573, "grad_norm": 2.3865480422973633, "learning_rate": 6.320433489386375e-05, "loss": 3.2847, "step": 235600 }, { "epoch": 2.0810152069093415, "grad_norm": 3.722965717315674, "learning_rate": 6.3190931533698e-05, "loss": 2.9328, "step": 235650 }, { "epoch": 2.0814567547996257, "grad_norm": 2.7150113582611084, "learning_rate": 6.317752715464464e-05, "loss": 3.2807, "step": 235700 }, { "epoch": 2.08189830268991, "grad_norm": 2.8550362586975098, "learning_rate": 6.316412175773904e-05, "loss": 3.0331, "step": 235750 }, { "epoch": 2.082339850580194, "grad_norm": 3.4029903411865234, "learning_rate": 6.315071534401669e-05, "loss": 3.0143, "step": 235800 }, { "epoch": 2.082781398470478, "grad_norm": 2.1413564682006836, "learning_rate": 6.313730791451306e-05, "loss": 3.216, "step": 235850 }, { "epoch": 2.0832229463607623, "grad_norm": 3.0733673572540283, "learning_rate": 6.312389947026383e-05, "loss": 3.28, "step": 235900 }, { "epoch": 2.0836644942510465, "grad_norm": 4.860799789428711, "learning_rate": 6.311049001230463e-05, "loss": 3.0566, "step": 235950 }, { "epoch": 2.0841060421413307, "grad_norm": 1.6133939027786255, "learning_rate": 6.309707954167126e-05, "loss": 2.8544, "step": 236000 }, { "epoch": 2.084547590031615, "grad_norm": 3.4890334606170654, "learning_rate": 6.308366805939955e-05, "loss": 3.4803, "step": 236050 }, { "epoch": 2.084989137921899, "grad_norm": 5.022340297698975, "learning_rate": 6.307025556652543e-05, "loss": 2.8892, "step": 236100 }, { "epoch": 2.085430685812183, "grad_norm": 2.1764464378356934, "learning_rate": 6.305684206408491e-05, "loss": 2.8681, "step": 236150 }, { "epoch": 2.0858722337024673, "grad_norm": 4.524923801422119, "learning_rate": 6.304342755311406e-05, "loss": 2.9398, "step": 236200 }, { "epoch": 2.0863137815927515, "grad_norm": 2.636495590209961, "learning_rate": 6.3030012034649e-05, "loss": 2.9677, "step": 236250 }, { "epoch": 2.0867553294830357, "grad_norm": 2.2745394706726074, "learning_rate": 6.301659550972604e-05, "loss": 3.0237, "step": 236300 }, { "epoch": 2.08719687737332, "grad_norm": 6.002178192138672, "learning_rate": 6.300317797938144e-05, "loss": 2.7733, "step": 236350 }, { "epoch": 2.087638425263604, "grad_norm": NaN, "learning_rate": 6.299002782518247e-05, "loss": 3.1529, "step": 236400 }, { "epoch": 2.088079973153888, "grad_norm": 5.430758953094482, "learning_rate": 6.29766083071607e-05, "loss": 2.8299, "step": 236450 }, { "epoch": 2.0885215210441723, "grad_norm": 2.8759422302246094, "learning_rate": 6.296318778680596e-05, "loss": 2.7195, "step": 236500 }, { "epoch": 2.0889630689344565, "grad_norm": 1.781813144683838, "learning_rate": 6.29497662651549e-05, "loss": 3.2773, "step": 236550 }, { "epoch": 2.0894046168247407, "grad_norm": 6.49731969833374, "learning_rate": 6.293634374324419e-05, "loss": 3.0205, "step": 236600 }, { "epoch": 2.089846164715025, "grad_norm": 2.8632984161376953, "learning_rate": 6.292292022211063e-05, "loss": 3.0186, "step": 236650 }, { "epoch": 2.090287712605309, "grad_norm": 2.556323289871216, "learning_rate": 6.290949570279107e-05, "loss": 2.7086, "step": 236700 }, { "epoch": 2.090729260495593, "grad_norm": 3.0288796424865723, "learning_rate": 6.289607018632244e-05, "loss": 3.0225, "step": 236750 }, { "epoch": 2.0911708083858773, "grad_norm": 3.256965160369873, "learning_rate": 6.288264367374173e-05, "loss": 3.4244, "step": 236800 }, { "epoch": 2.091612356276162, "grad_norm": 8.172772407531738, "learning_rate": 6.286921616608606e-05, "loss": 3.2433, "step": 236850 }, { "epoch": 2.092053904166446, "grad_norm": 3.213986873626709, "learning_rate": 6.285578766439255e-05, "loss": 3.3647, "step": 236900 }, { "epoch": 2.0924954520567303, "grad_norm": 6.556624412536621, "learning_rate": 6.284235816969847e-05, "loss": 3.0929, "step": 236950 }, { "epoch": 2.0929369999470144, "grad_norm": 4.544995307922363, "learning_rate": 6.282892768304112e-05, "loss": 2.7749, "step": 237000 }, { "epoch": 2.0929369999470144, "eval_asr_loss": 0.9273501776573921, "eval_loss": 2.862983465194702, "eval_runtime": 20.8111, "eval_samples_per_second": 36.903, "eval_steps_per_second": 9.226, "eval_tts_loss": 6.039148560531288, "step": 237000 }, { "epoch": 2.0933785478372986, "grad_norm": 2.387843608856201, "learning_rate": 6.281549620545788e-05, "loss": 3.3171, "step": 237050 }, { "epoch": 2.0938200957275828, "grad_norm": 1.6893880367279053, "learning_rate": 6.280206373798624e-05, "loss": 3.3243, "step": 237100 }, { "epoch": 2.094261643617867, "grad_norm": 2.2369627952575684, "learning_rate": 6.27888989604742e-05, "loss": 2.6777, "step": 237150 }, { "epoch": 2.094703191508151, "grad_norm": 3.7619330883026123, "learning_rate": 6.277546453608453e-05, "loss": 2.8815, "step": 237200 }, { "epoch": 2.0951447393984353, "grad_norm": 2.0962555408477783, "learning_rate": 6.276202912489859e-05, "loss": 3.2917, "step": 237250 }, { "epoch": 2.0955862872887194, "grad_norm": 1.178427815437317, "learning_rate": 6.274859272795408e-05, "loss": 2.746, "step": 237300 }, { "epoch": 2.0960278351790036, "grad_norm": 1.9152957201004028, "learning_rate": 6.273515534628891e-05, "loss": 3.2423, "step": 237350 }, { "epoch": 2.0964693830692878, "grad_norm": 2.9031455516815186, "learning_rate": 6.272171698094098e-05, "loss": 3.0746, "step": 237400 }, { "epoch": 2.096910930959572, "grad_norm": 2.8071396350860596, "learning_rate": 6.270827763294828e-05, "loss": 3.2411, "step": 237450 }, { "epoch": 2.097352478849856, "grad_norm": 4.363584041595459, "learning_rate": 6.269483730334891e-05, "loss": 3.1862, "step": 237500 }, { "epoch": 2.0977940267401403, "grad_norm": 1.6039600372314453, "learning_rate": 6.268139599318099e-05, "loss": 3.4092, "step": 237550 }, { "epoch": 2.0982355746304244, "grad_norm": 5.065496921539307, "learning_rate": 6.266795370348278e-05, "loss": 2.8442, "step": 237600 }, { "epoch": 2.0986771225207086, "grad_norm": 3.6026055812835693, "learning_rate": 6.265451043529256e-05, "loss": 3.2712, "step": 237650 }, { "epoch": 2.0991186704109928, "grad_norm": 3.2399210929870605, "learning_rate": 6.264106618964872e-05, "loss": 3.0832, "step": 237700 }, { "epoch": 2.099560218301277, "grad_norm": 3.737478494644165, "learning_rate": 6.26276209675897e-05, "loss": 3.2228, "step": 237750 }, { "epoch": 2.100001766191561, "grad_norm": 1.1522648334503174, "learning_rate": 6.261417477015404e-05, "loss": 3.1848, "step": 237800 }, { "epoch": 2.1004433140818453, "grad_norm": 3.852766513824463, "learning_rate": 6.260072759838035e-05, "loss": 2.6, "step": 237850 }, { "epoch": 2.1008848619721294, "grad_norm": 3.893470287322998, "learning_rate": 6.258727945330731e-05, "loss": 2.656, "step": 237900 }, { "epoch": 2.1013264098624136, "grad_norm": 2.8423218727111816, "learning_rate": 6.257383033597367e-05, "loss": 2.7779, "step": 237950 }, { "epoch": 2.1017679577526978, "grad_norm": 7.985106468200684, "learning_rate": 6.256038024741827e-05, "loss": 2.403, "step": 238000 }, { "epoch": 2.102209505642982, "grad_norm": 1.0427534580230713, "learning_rate": 6.254692918867997e-05, "loss": 3.2697, "step": 238050 }, { "epoch": 2.102651053533266, "grad_norm": 3.9906439781188965, "learning_rate": 6.25334771607978e-05, "loss": 3.2458, "step": 238100 }, { "epoch": 2.1030926014235503, "grad_norm": 3.478618860244751, "learning_rate": 6.252002416481082e-05, "loss": 3.1903, "step": 238150 }, { "epoch": 2.1035341493138344, "grad_norm": 4.565757751464844, "learning_rate": 6.250657020175811e-05, "loss": 3.3304, "step": 238200 }, { "epoch": 2.1039756972041186, "grad_norm": 4.789504051208496, "learning_rate": 6.249311527267892e-05, "loss": 3.4375, "step": 238250 }, { "epoch": 2.1044172450944028, "grad_norm": 2.925300121307373, "learning_rate": 6.247965937861251e-05, "loss": 3.4789, "step": 238300 }, { "epoch": 2.104858792984687, "grad_norm": 2.9117133617401123, "learning_rate": 6.246620252059824e-05, "loss": 3.0623, "step": 238350 }, { "epoch": 2.105300340874971, "grad_norm": 0.9601924419403076, "learning_rate": 6.245274469967553e-05, "loss": 3.3377, "step": 238400 }, { "epoch": 2.1057418887652553, "grad_norm": 3.297055959701538, "learning_rate": 6.24392859168839e-05, "loss": 2.9941, "step": 238450 }, { "epoch": 2.10618343665554, "grad_norm": 2.3726651668548584, "learning_rate": 6.24258261732629e-05, "loss": 3.4003, "step": 238500 }, { "epoch": 2.106624984545824, "grad_norm": 2.375788927078247, "learning_rate": 6.241236546985221e-05, "loss": 3.079, "step": 238550 }, { "epoch": 2.107066532436108, "grad_norm": 0.5383282899856567, "learning_rate": 6.239890380769154e-05, "loss": 2.3831, "step": 238600 }, { "epoch": 2.1075080803263924, "grad_norm": 3.9929215908050537, "learning_rate": 6.238544118782069e-05, "loss": 3.3199, "step": 238650 }, { "epoch": 2.1079496282166765, "grad_norm": 2.6869723796844482, "learning_rate": 6.23719776112795e-05, "loss": 3.6369, "step": 238700 }, { "epoch": 2.1083911761069607, "grad_norm": 3.927208662033081, "learning_rate": 6.2358513079108e-05, "loss": 3.2033, "step": 238750 }, { "epoch": 2.108832723997245, "grad_norm": 2.8071539402008057, "learning_rate": 6.234504759234613e-05, "loss": 3.4232, "step": 238800 }, { "epoch": 2.109274271887529, "grad_norm": 1.4641625881195068, "learning_rate": 6.233158115203403e-05, "loss": 2.7924, "step": 238850 }, { "epoch": 2.109715819777813, "grad_norm": 1.6354135274887085, "learning_rate": 6.231811375921184e-05, "loss": 2.7455, "step": 238900 }, { "epoch": 2.1101573676680974, "grad_norm": 2.952073574066162, "learning_rate": 6.230464541491985e-05, "loss": 3.3115, "step": 238950 }, { "epoch": 2.1105989155583815, "grad_norm": 1.168439507484436, "learning_rate": 6.229117612019832e-05, "loss": 2.6336, "step": 239000 }, { "epoch": 2.1110404634486657, "grad_norm": 2.487980842590332, "learning_rate": 6.227770587608767e-05, "loss": 2.9631, "step": 239050 }, { "epoch": 2.11148201133895, "grad_norm": 3.3226206302642822, "learning_rate": 6.226423468362834e-05, "loss": 3.1342, "step": 239100 }, { "epoch": 2.111923559229234, "grad_norm": 1.275766372680664, "learning_rate": 6.225076254386088e-05, "loss": 2.7905, "step": 239150 }, { "epoch": 2.112365107119518, "grad_norm": 3.030874252319336, "learning_rate": 6.223728945782591e-05, "loss": 2.9667, "step": 239200 }, { "epoch": 2.1128066550098024, "grad_norm": 4.319136142730713, "learning_rate": 6.222381542656408e-05, "loss": 3.2937, "step": 239250 }, { "epoch": 2.1132482029000865, "grad_norm": 2.6660947799682617, "learning_rate": 6.221034045111617e-05, "loss": 3.058, "step": 239300 }, { "epoch": 2.1136897507903707, "grad_norm": 3.479022741317749, "learning_rate": 6.2196864532523e-05, "loss": 3.3027, "step": 239350 }, { "epoch": 2.114131298680655, "grad_norm": 1.8305370807647705, "learning_rate": 6.218338767182548e-05, "loss": 2.9988, "step": 239400 }, { "epoch": 2.114572846570939, "grad_norm": 2.012442111968994, "learning_rate": 6.216990987006457e-05, "loss": 3.2906, "step": 239450 }, { "epoch": 2.115014394461223, "grad_norm": 1.7256959676742554, "learning_rate": 6.215643112828133e-05, "loss": 2.9947, "step": 239500 }, { "epoch": 2.1154559423515074, "grad_norm": 2.1054117679595947, "learning_rate": 6.214295144751684e-05, "loss": 2.5548, "step": 239550 }, { "epoch": 2.1158974902417915, "grad_norm": 3.857567071914673, "learning_rate": 6.212947082881234e-05, "loss": 2.6859, "step": 239600 }, { "epoch": 2.1163390381320757, "grad_norm": 3.1762421131134033, "learning_rate": 6.211598927320907e-05, "loss": 2.8899, "step": 239650 }, { "epoch": 2.11678058602236, "grad_norm": 4.7173237800598145, "learning_rate": 6.210250678174837e-05, "loss": 3.0474, "step": 239700 }, { "epoch": 2.117222133912644, "grad_norm": 1.7269694805145264, "learning_rate": 6.208902335547166e-05, "loss": 3.3108, "step": 239750 }, { "epoch": 2.117663681802928, "grad_norm": 4.368720054626465, "learning_rate": 6.207553899542039e-05, "loss": 3.0016, "step": 239800 }, { "epoch": 2.1181052296932124, "grad_norm": 2.5750601291656494, "learning_rate": 6.206205370263612e-05, "loss": 2.912, "step": 239850 }, { "epoch": 2.1185467775834965, "grad_norm": 4.731289386749268, "learning_rate": 6.204856747816052e-05, "loss": 2.9171, "step": 239900 }, { "epoch": 2.1189883254737807, "grad_norm": 1.6712313890457153, "learning_rate": 6.203508032303524e-05, "loss": 2.8296, "step": 239950 }, { "epoch": 2.119429873364065, "grad_norm": 0.9750666618347168, "learning_rate": 6.202159223830204e-05, "loss": 3.4434, "step": 240000 }, { "epoch": 2.119429873364065, "eval_asr_loss": 0.9326350055396521, "eval_loss": 2.8607914447784424, "eval_runtime": 20.6006, "eval_samples_per_second": 37.28, "eval_steps_per_second": 9.32, "eval_tts_loss": 6.059072890020006, "step": 240000 }, { "epoch": 2.1198714212543495, "grad_norm": 5.349383354187012, "learning_rate": 6.200810322500278e-05, "loss": 3.0129, "step": 240050 }, { "epoch": 2.1203129691446336, "grad_norm": 3.87054443359375, "learning_rate": 6.19946132841794e-05, "loss": 3.0981, "step": 240100 }, { "epoch": 2.120754517034918, "grad_norm": 2.3254106044769287, "learning_rate": 6.198112241687383e-05, "loss": 3.0004, "step": 240150 }, { "epoch": 2.121196064925202, "grad_norm": 4.357524871826172, "learning_rate": 6.196763062412816e-05, "loss": 2.987, "step": 240200 }, { "epoch": 2.121637612815486, "grad_norm": 1.6933478116989136, "learning_rate": 6.19541379069845e-05, "loss": 2.9078, "step": 240250 }, { "epoch": 2.1220791607057703, "grad_norm": 1.5506831407546997, "learning_rate": 6.194064426648507e-05, "loss": 3.2328, "step": 240300 }, { "epoch": 2.1225207085960545, "grad_norm": 4.903199672698975, "learning_rate": 6.192714970367211e-05, "loss": 3.0385, "step": 240350 }, { "epoch": 2.1229622564863386, "grad_norm": 1.618307113647461, "learning_rate": 6.191365421958797e-05, "loss": 3.3363, "step": 240400 }, { "epoch": 2.123403804376623, "grad_norm": 3.6995747089385986, "learning_rate": 6.190015781527508e-05, "loss": 2.7186, "step": 240450 }, { "epoch": 2.123845352266907, "grad_norm": 4.0562944412231445, "learning_rate": 6.18866604917759e-05, "loss": 3.1056, "step": 240500 }, { "epoch": 2.124286900157191, "grad_norm": 2.19975209236145, "learning_rate": 6.1873162250133e-05, "loss": 3.6208, "step": 240550 }, { "epoch": 2.1247284480474753, "grad_norm": 2.4086647033691406, "learning_rate": 6.1859663091389e-05, "loss": 3.0413, "step": 240600 }, { "epoch": 2.1251699959377595, "grad_norm": 1.795384407043457, "learning_rate": 6.184616301658658e-05, "loss": 2.8259, "step": 240650 }, { "epoch": 2.1256115438280436, "grad_norm": 2.943624973297119, "learning_rate": 6.183266202676854e-05, "loss": 2.8549, "step": 240700 }, { "epoch": 2.126053091718328, "grad_norm": 5.150783538818359, "learning_rate": 6.181916012297767e-05, "loss": 2.8026, "step": 240750 }, { "epoch": 2.126494639608612, "grad_norm": 2.3909823894500732, "learning_rate": 6.180565730625692e-05, "loss": 3.0313, "step": 240800 }, { "epoch": 2.126936187498896, "grad_norm": 3.5351691246032715, "learning_rate": 6.179215357764925e-05, "loss": 2.6941, "step": 240850 }, { "epoch": 2.1273777353891803, "grad_norm": 4.411786079406738, "learning_rate": 6.177864893819771e-05, "loss": 3.0214, "step": 240900 }, { "epoch": 2.1278192832794645, "grad_norm": 3.681881904602051, "learning_rate": 6.176514338894543e-05, "loss": 3.1937, "step": 240950 }, { "epoch": 2.1282608311697486, "grad_norm": 3.171037435531616, "learning_rate": 6.175163693093556e-05, "loss": 2.8394, "step": 241000 }, { "epoch": 2.128702379060033, "grad_norm": 3.15596866607666, "learning_rate": 6.173812956521141e-05, "loss": 3.3151, "step": 241050 }, { "epoch": 2.129143926950317, "grad_norm": 0.9684367179870605, "learning_rate": 6.17246212928163e-05, "loss": 3.0942, "step": 241100 }, { "epoch": 2.129585474840601, "grad_norm": 1.9763288497924805, "learning_rate": 6.171111211479358e-05, "loss": 2.7623, "step": 241150 }, { "epoch": 2.1300270227308853, "grad_norm": 4.591089248657227, "learning_rate": 6.16976020321868e-05, "loss": 3.0964, "step": 241200 }, { "epoch": 2.1304685706211695, "grad_norm": 2.27897047996521, "learning_rate": 6.168409104603943e-05, "loss": 3.069, "step": 241250 }, { "epoch": 2.1309101185114536, "grad_norm": 2.3860621452331543, "learning_rate": 6.167057915739511e-05, "loss": 3.6816, "step": 241300 }, { "epoch": 2.131351666401738, "grad_norm": 3.626079797744751, "learning_rate": 6.165706636729752e-05, "loss": 3.0158, "step": 241350 }, { "epoch": 2.131793214292022, "grad_norm": 4.617300987243652, "learning_rate": 6.164355267679037e-05, "loss": 3.0871, "step": 241400 }, { "epoch": 2.132234762182306, "grad_norm": 4.720652103424072, "learning_rate": 6.163003808691754e-05, "loss": 3.0995, "step": 241450 }, { "epoch": 2.1326763100725903, "grad_norm": 4.754398345947266, "learning_rate": 6.16165225987229e-05, "loss": 3.0797, "step": 241500 }, { "epoch": 2.1331178579628745, "grad_norm": 1.57192862033844, "learning_rate": 6.160300621325037e-05, "loss": 3.4888, "step": 241550 }, { "epoch": 2.133559405853159, "grad_norm": 3.342073440551758, "learning_rate": 6.158948893154401e-05, "loss": 3.5425, "step": 241600 }, { "epoch": 2.134000953743443, "grad_norm": 1.9215145111083984, "learning_rate": 6.15759707546479e-05, "loss": 2.7573, "step": 241650 }, { "epoch": 2.1344425016337274, "grad_norm": 2.9148123264312744, "learning_rate": 6.156245168360621e-05, "loss": 3.0206, "step": 241700 }, { "epoch": 2.1348840495240116, "grad_norm": 2.3337414264678955, "learning_rate": 6.154893171946319e-05, "loss": 3.3977, "step": 241750 }, { "epoch": 2.1353255974142957, "grad_norm": 3.570500612258911, "learning_rate": 6.15354108632631e-05, "loss": 3.1962, "step": 241800 }, { "epoch": 2.13576714530458, "grad_norm": 0.8573853969573975, "learning_rate": 6.152188911605037e-05, "loss": 2.7685, "step": 241850 }, { "epoch": 2.136208693194864, "grad_norm": 6.819091796875, "learning_rate": 6.150836647886937e-05, "loss": 3.1988, "step": 241900 }, { "epoch": 2.1366502410851482, "grad_norm": 3.854794979095459, "learning_rate": 6.149484295276467e-05, "loss": 3.0705, "step": 241950 }, { "epoch": 2.1370917889754324, "grad_norm": 4.04559850692749, "learning_rate": 6.148131853878083e-05, "loss": 3.1822, "step": 242000 }, { "epoch": 2.1375333368657166, "grad_norm": 3.1430141925811768, "learning_rate": 6.146779323796248e-05, "loss": 2.98, "step": 242050 }, { "epoch": 2.1379748847560007, "grad_norm": 2.6168956756591797, "learning_rate": 6.145426705135436e-05, "loss": 2.991, "step": 242100 }, { "epoch": 2.138416432646285, "grad_norm": 2.851386308670044, "learning_rate": 6.144073998000123e-05, "loss": 2.8305, "step": 242150 }, { "epoch": 2.138857980536569, "grad_norm": 2.7866199016571045, "learning_rate": 6.142721202494795e-05, "loss": 3.1857, "step": 242200 }, { "epoch": 2.1392995284268532, "grad_norm": 1.859724760055542, "learning_rate": 6.141368318723946e-05, "loss": 2.7095, "step": 242250 }, { "epoch": 2.1397410763171374, "grad_norm": 2.3074333667755127, "learning_rate": 6.14001534679207e-05, "loss": 3.1211, "step": 242300 }, { "epoch": 2.1401826242074216, "grad_norm": 3.469409942626953, "learning_rate": 6.138662286803677e-05, "loss": 2.8818, "step": 242350 }, { "epoch": 2.1406241720977057, "grad_norm": 4.329566478729248, "learning_rate": 6.137309138863278e-05, "loss": 3.3215, "step": 242400 }, { "epoch": 2.14106571998799, "grad_norm": 4.724846363067627, "learning_rate": 6.135955903075392e-05, "loss": 3.3933, "step": 242450 }, { "epoch": 2.141507267878274, "grad_norm": 1.137832522392273, "learning_rate": 6.134602579544546e-05, "loss": 2.7406, "step": 242500 }, { "epoch": 2.1419488157685582, "grad_norm": 2.7646424770355225, "learning_rate": 6.133249168375269e-05, "loss": 3.569, "step": 242550 }, { "epoch": 2.1423903636588424, "grad_norm": 6.01995325088501, "learning_rate": 6.131895669672106e-05, "loss": 3.3849, "step": 242600 }, { "epoch": 2.1428319115491266, "grad_norm": 1.6857974529266357, "learning_rate": 6.130542083539601e-05, "loss": 3.0329, "step": 242650 }, { "epoch": 2.1432734594394107, "grad_norm": 5.805901050567627, "learning_rate": 6.129188410082305e-05, "loss": 2.8821, "step": 242700 }, { "epoch": 2.143715007329695, "grad_norm": 2.3712074756622314, "learning_rate": 6.12783464940478e-05, "loss": 3.1749, "step": 242750 }, { "epoch": 2.144156555219979, "grad_norm": 1.9185987710952759, "learning_rate": 6.126480801611593e-05, "loss": 3.5455, "step": 242800 }, { "epoch": 2.144598103110263, "grad_norm": 2.8377609252929688, "learning_rate": 6.125126866807315e-05, "loss": 3.174, "step": 242850 }, { "epoch": 2.1450396510005474, "grad_norm": 1.1797131299972534, "learning_rate": 6.123772845096529e-05, "loss": 3.1248, "step": 242900 }, { "epoch": 2.1454811988908316, "grad_norm": 2.4720191955566406, "learning_rate": 6.122418736583819e-05, "loss": 3.3231, "step": 242950 }, { "epoch": 2.1459227467811157, "grad_norm": 2.443220376968384, "learning_rate": 6.121064541373779e-05, "loss": 3.1198, "step": 243000 }, { "epoch": 2.1459227467811157, "eval_asr_loss": 0.9242461907759509, "eval_loss": 2.86102032661438, "eval_runtime": 20.6095, "eval_samples_per_second": 37.264, "eval_steps_per_second": 9.316, "eval_tts_loss": 6.029543186687746, "step": 243000 }, { "epoch": 2.1463642946714, "grad_norm": 3.9056832790374756, "learning_rate": 6.11971025957101e-05, "loss": 3.4246, "step": 243050 }, { "epoch": 2.146805842561684, "grad_norm": 5.718588352203369, "learning_rate": 6.118355891280119e-05, "loss": 3.04, "step": 243100 }, { "epoch": 2.147247390451968, "grad_norm": 1.7407891750335693, "learning_rate": 6.117001436605718e-05, "loss": 3.231, "step": 243150 }, { "epoch": 2.1476889383422524, "grad_norm": 2.6677162647247314, "learning_rate": 6.11567398731635e-05, "loss": 3.4042, "step": 243200 }, { "epoch": 2.148130486232537, "grad_norm": 2.967630386352539, "learning_rate": 6.114319361911259e-05, "loss": 3.191, "step": 243250 }, { "epoch": 2.148572034122821, "grad_norm": 4.103931903839111, "learning_rate": 6.112964650434443e-05, "loss": 3.0904, "step": 243300 }, { "epoch": 2.1490135820131053, "grad_norm": 4.89622163772583, "learning_rate": 6.111609852990548e-05, "loss": 2.9498, "step": 243350 }, { "epoch": 2.1494551299033895, "grad_norm": 2.395061492919922, "learning_rate": 6.110254969684219e-05, "loss": 3.1946, "step": 243400 }, { "epoch": 2.1498966777936737, "grad_norm": 3.4221737384796143, "learning_rate": 6.108900000620106e-05, "loss": 3.2357, "step": 243450 }, { "epoch": 2.150338225683958, "grad_norm": 4.201228141784668, "learning_rate": 6.107544945902872e-05, "loss": 2.9938, "step": 243500 }, { "epoch": 2.150779773574242, "grad_norm": 4.521472930908203, "learning_rate": 6.106189805637184e-05, "loss": 3.3467, "step": 243550 }, { "epoch": 2.151221321464526, "grad_norm": 4.29344367980957, "learning_rate": 6.104834579927711e-05, "loss": 3.0526, "step": 243600 }, { "epoch": 2.1516628693548103, "grad_norm": 2.7563273906707764, "learning_rate": 6.103506375935755e-05, "loss": 3.5406, "step": 243650 }, { "epoch": 2.1521044172450945, "grad_norm": 5.406722068786621, "learning_rate": 6.102150981356427e-05, "loss": 2.2932, "step": 243700 }, { "epoch": 2.1525459651353787, "grad_norm": 2.6092753410339355, "learning_rate": 6.100795501645281e-05, "loss": 3.2518, "step": 243750 }, { "epoch": 2.152987513025663, "grad_norm": 3.871994972229004, "learning_rate": 6.0994399369070176e-05, "loss": 2.6694, "step": 243800 }, { "epoch": 2.153429060915947, "grad_norm": 5.2820892333984375, "learning_rate": 6.098084287246343e-05, "loss": 3.4533, "step": 243850 }, { "epoch": 2.153870608806231, "grad_norm": 2.723698139190674, "learning_rate": 6.096728552767967e-05, "loss": 3.1388, "step": 243900 }, { "epoch": 2.1543121566965153, "grad_norm": 3.1074745655059814, "learning_rate": 6.095372733576611e-05, "loss": 3.4028, "step": 243950 }, { "epoch": 2.1547537045867995, "grad_norm": 3.3812079429626465, "learning_rate": 6.094016829776998e-05, "loss": 2.7625, "step": 244000 }, { "epoch": 2.1551952524770837, "grad_norm": 3.0963339805603027, "learning_rate": 6.092660841473865e-05, "loss": 2.8601, "step": 244050 }, { "epoch": 2.155636800367368, "grad_norm": 4.549088478088379, "learning_rate": 6.091304768771947e-05, "loss": 3.0102, "step": 244100 }, { "epoch": 2.156078348257652, "grad_norm": 2.3300344944000244, "learning_rate": 6.089948611775988e-05, "loss": 3.0406, "step": 244150 }, { "epoch": 2.156519896147936, "grad_norm": 2.152743339538574, "learning_rate": 6.088592370590742e-05, "loss": 2.8996, "step": 244200 }, { "epoch": 2.1569614440382203, "grad_norm": 1.4900521039962769, "learning_rate": 6.087236045320966e-05, "loss": 3.0599, "step": 244250 }, { "epoch": 2.1574029919285045, "grad_norm": 2.646312713623047, "learning_rate": 6.0858796360714256e-05, "loss": 2.868, "step": 244300 }, { "epoch": 2.1578445398187887, "grad_norm": 3.2758681774139404, "learning_rate": 6.0845231429468916e-05, "loss": 3.0807, "step": 244350 }, { "epoch": 2.158286087709073, "grad_norm": 2.091062545776367, "learning_rate": 6.0831665660521395e-05, "loss": 3.0439, "step": 244400 }, { "epoch": 2.158727635599357, "grad_norm": 1.417259693145752, "learning_rate": 6.0818099054919574e-05, "loss": 3.0763, "step": 244450 }, { "epoch": 2.159169183489641, "grad_norm": 0.9007434844970703, "learning_rate": 6.080453161371133e-05, "loss": 3.0514, "step": 244500 }, { "epoch": 2.1596107313799253, "grad_norm": 2.9796414375305176, "learning_rate": 6.079096333794463e-05, "loss": 2.9235, "step": 244550 }, { "epoch": 2.1600522792702095, "grad_norm": 2.401273250579834, "learning_rate": 6.077739422866752e-05, "loss": 2.8255, "step": 244600 }, { "epoch": 2.1604938271604937, "grad_norm": 2.502058982849121, "learning_rate": 6.0763824286928096e-05, "loss": 3.7429, "step": 244650 }, { "epoch": 2.160935375050778, "grad_norm": 4.2916178703308105, "learning_rate": 6.075025351377453e-05, "loss": 2.7973, "step": 244700 }, { "epoch": 2.161376922941062, "grad_norm": 2.6151785850524902, "learning_rate": 6.0736681910255024e-05, "loss": 3.1876, "step": 244750 }, { "epoch": 2.1618184708313466, "grad_norm": 5.386536598205566, "learning_rate": 6.0723109477417896e-05, "loss": 3.115, "step": 244800 }, { "epoch": 2.1622600187216303, "grad_norm": 3.8035407066345215, "learning_rate": 6.070953621631148e-05, "loss": 2.852, "step": 244850 }, { "epoch": 2.162701566611915, "grad_norm": 0.988106906414032, "learning_rate": 6.0695962127984204e-05, "loss": 3.1852, "step": 244900 }, { "epoch": 2.163143114502199, "grad_norm": 2.1697728633880615, "learning_rate": 6.068238721348456e-05, "loss": 3.2436, "step": 244950 }, { "epoch": 2.1635846623924833, "grad_norm": 1.6820136308670044, "learning_rate": 6.06688114738611e-05, "loss": 2.865, "step": 245000 }, { "epoch": 2.1640262102827674, "grad_norm": 4.40377950668335, "learning_rate": 6.065523491016241e-05, "loss": 2.9429, "step": 245050 }, { "epoch": 2.1644677581730516, "grad_norm": 2.3662831783294678, "learning_rate": 6.064165752343718e-05, "loss": 3.0624, "step": 245100 }, { "epoch": 2.1649093060633358, "grad_norm": 3.140277147293091, "learning_rate": 6.062807931473413e-05, "loss": 2.7547, "step": 245150 }, { "epoch": 2.16535085395362, "grad_norm": 5.113159656524658, "learning_rate": 6.0614771873733054e-05, "loss": 3.41, "step": 245200 }, { "epoch": 2.165792401843904, "grad_norm": 1.3435204029083252, "learning_rate": 6.06011920406082e-05, "loss": 3.3534, "step": 245250 }, { "epoch": 2.1662339497341883, "grad_norm": 1.744178056716919, "learning_rate": 6.058761138863115e-05, "loss": 2.7636, "step": 245300 }, { "epoch": 2.1666754976244724, "grad_norm": 6.074828624725342, "learning_rate": 6.0574029918850905e-05, "loss": 3.0243, "step": 245350 }, { "epoch": 2.1671170455147566, "grad_norm": 3.06844425201416, "learning_rate": 6.056044763231652e-05, "loss": 3.4827, "step": 245400 }, { "epoch": 2.1675585934050408, "grad_norm": 4.505370616912842, "learning_rate": 6.0546864530077084e-05, "loss": 2.7537, "step": 245450 }, { "epoch": 2.168000141295325, "grad_norm": 3.3782474994659424, "learning_rate": 6.053328061318182e-05, "loss": 2.8607, "step": 245500 }, { "epoch": 2.168441689185609, "grad_norm": 2.681347608566284, "learning_rate": 6.051969588267994e-05, "loss": 2.8934, "step": 245550 }, { "epoch": 2.1688832370758933, "grad_norm": 1.9563655853271484, "learning_rate": 6.050611033962076e-05, "loss": 3.3919, "step": 245600 }, { "epoch": 2.1693247849661774, "grad_norm": 1.3241103887557983, "learning_rate": 6.049252398505364e-05, "loss": 3.2921, "step": 245650 }, { "epoch": 2.1697663328564616, "grad_norm": 3.78419828414917, "learning_rate": 6.047893682002802e-05, "loss": 3.1427, "step": 245700 }, { "epoch": 2.1702078807467458, "grad_norm": 3.2498936653137207, "learning_rate": 6.0465348845593394e-05, "loss": 3.0815, "step": 245750 }, { "epoch": 2.17064942863703, "grad_norm": 2.077392578125, "learning_rate": 6.04517600627993e-05, "loss": 2.8622, "step": 245800 }, { "epoch": 2.171090976527314, "grad_norm": 2.2121798992156982, "learning_rate": 6.0438170472695374e-05, "loss": 3.1503, "step": 245850 }, { "epoch": 2.1715325244175983, "grad_norm": 1.9977270364761353, "learning_rate": 6.0424580076331305e-05, "loss": 2.6607, "step": 245900 }, { "epoch": 2.1719740723078824, "grad_norm": 4.047658443450928, "learning_rate": 6.041098887475681e-05, "loss": 3.0044, "step": 245950 }, { "epoch": 2.1724156201981666, "grad_norm": 1.35835862159729, "learning_rate": 6.0397396869021714e-05, "loss": 3.186, "step": 246000 }, { "epoch": 2.1724156201981666, "eval_asr_loss": 0.9392569830705412, "eval_loss": 2.8575477600097656, "eval_runtime": 20.4511, "eval_samples_per_second": 37.553, "eval_steps_per_second": 9.388, "eval_tts_loss": 6.0642424824213865, "step": 246000 }, { "epoch": 2.1728571680884508, "grad_norm": 2.7876996994018555, "learning_rate": 6.038380406017588e-05, "loss": 3.0091, "step": 246050 }, { "epoch": 2.173298715978735, "grad_norm": 23.924245834350586, "learning_rate": 6.0370210449269224e-05, "loss": 3.028, "step": 246100 }, { "epoch": 2.173740263869019, "grad_norm": 2.4650650024414062, "learning_rate": 6.0356616037351755e-05, "loss": 2.8779, "step": 246150 }, { "epoch": 2.1741818117593033, "grad_norm": 1.138063669204712, "learning_rate": 6.034302082547351e-05, "loss": 2.9028, "step": 246200 }, { "epoch": 2.1746233596495874, "grad_norm": 2.4580750465393066, "learning_rate": 6.032942481468462e-05, "loss": 3.2762, "step": 246250 }, { "epoch": 2.1750649075398716, "grad_norm": 4.300107002258301, "learning_rate": 6.0315828006035245e-05, "loss": 3.3666, "step": 246300 }, { "epoch": 2.1755064554301558, "grad_norm": 2.1569507122039795, "learning_rate": 6.0302230400575646e-05, "loss": 3.1456, "step": 246350 }, { "epoch": 2.17594800332044, "grad_norm": 4.602993488311768, "learning_rate": 6.0288631999356104e-05, "loss": 3.2562, "step": 246400 }, { "epoch": 2.1763895512107245, "grad_norm": 4.042548179626465, "learning_rate": 6.0275032803426975e-05, "loss": 2.946, "step": 246450 }, { "epoch": 2.1768310991010087, "grad_norm": 3.4596941471099854, "learning_rate": 6.0261432813838715e-05, "loss": 2.9021, "step": 246500 }, { "epoch": 2.177272646991293, "grad_norm": 6.044607639312744, "learning_rate": 6.024783203164177e-05, "loss": 2.8478, "step": 246550 }, { "epoch": 2.177714194881577, "grad_norm": 2.1571695804595947, "learning_rate": 6.023423045788669e-05, "loss": 3.2925, "step": 246600 }, { "epoch": 2.178155742771861, "grad_norm": 4.901209354400635, "learning_rate": 6.02206280936241e-05, "loss": 3.12, "step": 246650 }, { "epoch": 2.1785972906621454, "grad_norm": 2.6473047733306885, "learning_rate": 6.0207024939904655e-05, "loss": 3.1452, "step": 246700 }, { "epoch": 2.1790388385524295, "grad_norm": 2.9098074436187744, "learning_rate": 6.0193420997779096e-05, "loss": 3.2063, "step": 246750 }, { "epoch": 2.1794803864427137, "grad_norm": 7.190146446228027, "learning_rate": 6.0179816268298194e-05, "loss": 3.2197, "step": 246800 }, { "epoch": 2.179921934332998, "grad_norm": 2.408454418182373, "learning_rate": 6.0166210752512806e-05, "loss": 2.8101, "step": 246850 }, { "epoch": 2.180363482223282, "grad_norm": 5.173770904541016, "learning_rate": 6.015260445147386e-05, "loss": 3.2971, "step": 246900 }, { "epoch": 2.180805030113566, "grad_norm": 4.6776957511901855, "learning_rate": 6.01389973662323e-05, "loss": 3.2507, "step": 246950 }, { "epoch": 2.1812465780038504, "grad_norm": 2.6588592529296875, "learning_rate": 6.012538949783916e-05, "loss": 3.041, "step": 247000 }, { "epoch": 2.1816881258941345, "grad_norm": 2.170778274536133, "learning_rate": 6.011178084734556e-05, "loss": 3.2546, "step": 247050 }, { "epoch": 2.1821296737844187, "grad_norm": 3.928318977355957, "learning_rate": 6.009817141580262e-05, "loss": 3.4263, "step": 247100 }, { "epoch": 2.182571221674703, "grad_norm": 2.583965301513672, "learning_rate": 6.008456120426158e-05, "loss": 2.922, "step": 247150 }, { "epoch": 2.183012769564987, "grad_norm": 2.960022211074829, "learning_rate": 6.007095021377369e-05, "loss": 3.0257, "step": 247200 }, { "epoch": 2.183454317455271, "grad_norm": 1.4887888431549072, "learning_rate": 6.0057338445390296e-05, "loss": 3.0247, "step": 247250 }, { "epoch": 2.1838958653455554, "grad_norm": 2.6487021446228027, "learning_rate": 6.004372590016281e-05, "loss": 3.2311, "step": 247300 }, { "epoch": 2.1843374132358395, "grad_norm": 3.543104648590088, "learning_rate": 6.0030112579142636e-05, "loss": 3.1826, "step": 247350 }, { "epoch": 2.1847789611261237, "grad_norm": 4.436171531677246, "learning_rate": 6.0016498483381345e-05, "loss": 3.0521, "step": 247400 }, { "epoch": 2.185220509016408, "grad_norm": 1.3873144388198853, "learning_rate": 6.000288361393047e-05, "loss": 3.2691, "step": 247450 }, { "epoch": 2.185662056906692, "grad_norm": 2.788475513458252, "learning_rate": 5.998926797184167e-05, "loss": 3.2972, "step": 247500 }, { "epoch": 2.186103604796976, "grad_norm": 5.260395050048828, "learning_rate": 5.997565155816661e-05, "loss": 3.1924, "step": 247550 }, { "epoch": 2.1865451526872604, "grad_norm": 3.8101744651794434, "learning_rate": 5.996203437395706e-05, "loss": 3.6127, "step": 247600 }, { "epoch": 2.1869867005775445, "grad_norm": 6.26643705368042, "learning_rate": 5.994841642026485e-05, "loss": 2.5554, "step": 247650 }, { "epoch": 2.1874282484678287, "grad_norm": 4.825605869293213, "learning_rate": 5.993479769814182e-05, "loss": 3.6132, "step": 247700 }, { "epoch": 2.187869796358113, "grad_norm": 4.736049652099609, "learning_rate": 5.992117820863989e-05, "loss": 2.9763, "step": 247750 }, { "epoch": 2.188311344248397, "grad_norm": 1.331676959991455, "learning_rate": 5.99075579528111e-05, "loss": 3.1284, "step": 247800 }, { "epoch": 2.188752892138681, "grad_norm": 2.1483421325683594, "learning_rate": 5.989393693170746e-05, "loss": 2.9496, "step": 247850 }, { "epoch": 2.1891944400289653, "grad_norm": 3.0510306358337402, "learning_rate": 5.988031514638111e-05, "loss": 3.2423, "step": 247900 }, { "epoch": 2.1896359879192495, "grad_norm": 2.3355166912078857, "learning_rate": 5.9866692597884174e-05, "loss": 3.1941, "step": 247950 }, { "epoch": 2.190077535809534, "grad_norm": 3.830967664718628, "learning_rate": 5.9853069287268906e-05, "loss": 2.9566, "step": 248000 }, { "epoch": 2.190519083699818, "grad_norm": 4.699981212615967, "learning_rate": 5.9839445215587596e-05, "loss": 2.9206, "step": 248050 }, { "epoch": 2.1909606315901025, "grad_norm": 6.072689533233643, "learning_rate": 5.982582038389257e-05, "loss": 3.256, "step": 248100 }, { "epoch": 2.1914021794803866, "grad_norm": 4.4500203132629395, "learning_rate": 5.9812194793236245e-05, "loss": 3.0591, "step": 248150 }, { "epoch": 2.191843727370671, "grad_norm": 3.2074832916259766, "learning_rate": 5.979856844467108e-05, "loss": 2.9361, "step": 248200 }, { "epoch": 2.192285275260955, "grad_norm": 1.263113260269165, "learning_rate": 5.978494133924959e-05, "loss": 3.0833, "step": 248250 }, { "epoch": 2.192726823151239, "grad_norm": 1.6638262271881104, "learning_rate": 5.977131347802435e-05, "loss": 3.2541, "step": 248300 }, { "epoch": 2.1931683710415233, "grad_norm": 2.9127674102783203, "learning_rate": 5.9757684862048004e-05, "loss": 3.2699, "step": 248350 }, { "epoch": 2.1936099189318075, "grad_norm": 1.5334758758544922, "learning_rate": 5.974405549237324e-05, "loss": 3.212, "step": 248400 }, { "epoch": 2.1940514668220916, "grad_norm": 2.567260980606079, "learning_rate": 5.973042537005283e-05, "loss": 3.1895, "step": 248450 }, { "epoch": 2.194493014712376, "grad_norm": 2.907191753387451, "learning_rate": 5.971679449613956e-05, "loss": 3.2427, "step": 248500 }, { "epoch": 2.19493456260266, "grad_norm": 1.528548240661621, "learning_rate": 5.970316287168631e-05, "loss": 3.1348, "step": 248550 }, { "epoch": 2.195376110492944, "grad_norm": 2.61979079246521, "learning_rate": 5.9689530497746023e-05, "loss": 3.1506, "step": 248600 }, { "epoch": 2.1958176583832283, "grad_norm": 1.5362796783447266, "learning_rate": 5.967589737537166e-05, "loss": 2.6354, "step": 248650 }, { "epoch": 2.1962592062735125, "grad_norm": 4.144810676574707, "learning_rate": 5.966226350561628e-05, "loss": 3.2019, "step": 248700 }, { "epoch": 2.1967007541637966, "grad_norm": 2.271742820739746, "learning_rate": 5.964862888953297e-05, "loss": 3.2604, "step": 248750 }, { "epoch": 2.197142302054081, "grad_norm": 2.973482131958008, "learning_rate": 5.96349935281749e-05, "loss": 2.8219, "step": 248800 }, { "epoch": 2.197583849944365, "grad_norm": 4.361788749694824, "learning_rate": 5.9621357422595295e-05, "loss": 3.1846, "step": 248850 }, { "epoch": 2.198025397834649, "grad_norm": 0.705970048904419, "learning_rate": 5.960772057384739e-05, "loss": 3.398, "step": 248900 }, { "epoch": 2.1984669457249333, "grad_norm": 1.652596354484558, "learning_rate": 5.959408298298456e-05, "loss": 3.2226, "step": 248950 }, { "epoch": 2.1989084936152175, "grad_norm": 1.5994527339935303, "learning_rate": 5.958044465106016e-05, "loss": 2.5712, "step": 249000 }, { "epoch": 2.1989084936152175, "eval_asr_loss": 0.9304255538908192, "eval_loss": 2.846482992172241, "eval_runtime": 20.4139, "eval_samples_per_second": 37.621, "eval_steps_per_second": 9.405, "eval_tts_loss": 6.0102194777481515, "step": 249000 }, { "epoch": 2.1993500415055016, "grad_norm": 2.3923439979553223, "learning_rate": 5.956680557912766e-05, "loss": 3.1958, "step": 249050 }, { "epoch": 2.199791589395786, "grad_norm": 3.9689152240753174, "learning_rate": 5.955316576824056e-05, "loss": 3.1784, "step": 249100 }, { "epoch": 2.20023313728607, "grad_norm": 2.680379629135132, "learning_rate": 5.9539525219452396e-05, "loss": 3.2746, "step": 249150 }, { "epoch": 2.200674685176354, "grad_norm": 2.977792978286743, "learning_rate": 5.952588393381682e-05, "loss": 2.9965, "step": 249200 }, { "epoch": 2.2011162330666383, "grad_norm": 7.2773261070251465, "learning_rate": 5.951224191238749e-05, "loss": 3.0737, "step": 249250 }, { "epoch": 2.2015577809569224, "grad_norm": 3.3566043376922607, "learning_rate": 5.9498872018535125e-05, "loss": 3.2456, "step": 249300 }, { "epoch": 2.2019993288472066, "grad_norm": 4.931230068206787, "learning_rate": 5.948522854334293e-05, "loss": 2.9961, "step": 249350 }, { "epoch": 2.202440876737491, "grad_norm": 2.0846118927001953, "learning_rate": 5.9471584335497246e-05, "loss": 3.0981, "step": 249400 }, { "epoch": 2.202882424627775, "grad_norm": 5.715150833129883, "learning_rate": 5.9457939396051985e-05, "loss": 2.7673, "step": 249450 }, { "epoch": 2.203323972518059, "grad_norm": 1.7555757761001587, "learning_rate": 5.944429372606111e-05, "loss": 3.1935, "step": 249500 }, { "epoch": 2.2037655204083433, "grad_norm": 3.721235752105713, "learning_rate": 5.943064732657864e-05, "loss": 3.0966, "step": 249550 }, { "epoch": 2.2042070682986274, "grad_norm": 3.5476677417755127, "learning_rate": 5.9417000198658625e-05, "loss": 2.8752, "step": 249600 }, { "epoch": 2.204648616188912, "grad_norm": 4.5587286949157715, "learning_rate": 5.94033523433552e-05, "loss": 3.0805, "step": 249650 }, { "epoch": 2.2050901640791962, "grad_norm": 4.7551703453063965, "learning_rate": 5.938970376172254e-05, "loss": 2.7924, "step": 249700 }, { "epoch": 2.2055317119694804, "grad_norm": 3.6469037532806396, "learning_rate": 5.93760544548149e-05, "loss": 3.1216, "step": 249750 }, { "epoch": 2.2059732598597646, "grad_norm": 3.2515978813171387, "learning_rate": 5.9362404423686534e-05, "loss": 3.1162, "step": 249800 }, { "epoch": 2.2064148077500487, "grad_norm": 4.0924787521362305, "learning_rate": 5.9348753669391844e-05, "loss": 3.6559, "step": 249850 }, { "epoch": 2.206856355640333, "grad_norm": 2.8159549236297607, "learning_rate": 5.9335102192985194e-05, "loss": 2.7405, "step": 249900 }, { "epoch": 2.207297903530617, "grad_norm": 1.7060962915420532, "learning_rate": 5.9321449995521074e-05, "loss": 3.0616, "step": 249950 }, { "epoch": 2.2077394514209012, "grad_norm": 3.1570303440093994, "learning_rate": 5.930779707805397e-05, "loss": 3.4593, "step": 250000 }, { "epoch": 2.2081809993111854, "grad_norm": 2.408595561981201, "learning_rate": 5.929414344163846e-05, "loss": 3.0959, "step": 250050 }, { "epoch": 2.2086225472014696, "grad_norm": 3.342355966567993, "learning_rate": 5.9280489087329205e-05, "loss": 3.0539, "step": 250100 }, { "epoch": 2.2090640950917537, "grad_norm": 5.067995548248291, "learning_rate": 5.926683401618086e-05, "loss": 3.0701, "step": 250150 }, { "epoch": 2.209505642982038, "grad_norm": 1.7544509172439575, "learning_rate": 5.925317822924815e-05, "loss": 3.0253, "step": 250200 }, { "epoch": 2.209947190872322, "grad_norm": 2.652850866317749, "learning_rate": 5.923952172758591e-05, "loss": 3.3815, "step": 250250 }, { "epoch": 2.210388738762606, "grad_norm": 4.625723838806152, "learning_rate": 5.9225864512248955e-05, "loss": 2.9297, "step": 250300 }, { "epoch": 2.2108302866528904, "grad_norm": 2.095754384994507, "learning_rate": 5.9212206584292196e-05, "loss": 2.9707, "step": 250350 }, { "epoch": 2.2112718345431746, "grad_norm": 5.024804592132568, "learning_rate": 5.9198547944770596e-05, "loss": 3.1168, "step": 250400 }, { "epoch": 2.2117133824334587, "grad_norm": 9.1120023727417, "learning_rate": 5.918488859473916e-05, "loss": 2.9197, "step": 250450 }, { "epoch": 2.212154930323743, "grad_norm": 1.8666355609893799, "learning_rate": 5.9171228535253e-05, "loss": 3.035, "step": 250500 }, { "epoch": 2.212596478214027, "grad_norm": 1.0906444787979126, "learning_rate": 5.9157567767367175e-05, "loss": 3.2451, "step": 250550 }, { "epoch": 2.213038026104311, "grad_norm": 2.957937479019165, "learning_rate": 5.91439062921369e-05, "loss": 2.8599, "step": 250600 }, { "epoch": 2.2134795739945954, "grad_norm": 2.6834259033203125, "learning_rate": 5.913024411061739e-05, "loss": 3.3457, "step": 250650 }, { "epoch": 2.2139211218848796, "grad_norm": 2.9120137691497803, "learning_rate": 5.911658122386397e-05, "loss": 3.1717, "step": 250700 }, { "epoch": 2.2143626697751637, "grad_norm": 1.45148503780365, "learning_rate": 5.9102917632931945e-05, "loss": 3.0067, "step": 250750 }, { "epoch": 2.214804217665448, "grad_norm": 2.60270357131958, "learning_rate": 5.9089253338876715e-05, "loss": 3.1368, "step": 250800 }, { "epoch": 2.215245765555732, "grad_norm": 1.3488702774047852, "learning_rate": 5.907558834275374e-05, "loss": 2.949, "step": 250850 }, { "epoch": 2.215687313446016, "grad_norm": 3.042109727859497, "learning_rate": 5.906192264561855e-05, "loss": 3.3949, "step": 250900 }, { "epoch": 2.2161288613363004, "grad_norm": 1.2870628833770752, "learning_rate": 5.9048256248526644e-05, "loss": 2.9223, "step": 250950 }, { "epoch": 2.2165704092265845, "grad_norm": 4.794906139373779, "learning_rate": 5.9034589152533695e-05, "loss": 3.2492, "step": 251000 }, { "epoch": 2.2170119571168687, "grad_norm": 3.8245151042938232, "learning_rate": 5.902092135869533e-05, "loss": 3.2285, "step": 251050 }, { "epoch": 2.217453505007153, "grad_norm": 0.950536847114563, "learning_rate": 5.90072528680673e-05, "loss": 3.144, "step": 251100 }, { "epoch": 2.217895052897437, "grad_norm": 1.9943549633026123, "learning_rate": 5.8993583681705354e-05, "loss": 2.9898, "step": 251150 }, { "epoch": 2.2183366007877217, "grad_norm": 2.278146982192993, "learning_rate": 5.897991380066533e-05, "loss": 3.2155, "step": 251200 }, { "epoch": 2.218778148678006, "grad_norm": 6.75840425491333, "learning_rate": 5.8966243226003104e-05, "loss": 2.9816, "step": 251250 }, { "epoch": 2.21921969656829, "grad_norm": 2.8531532287597656, "learning_rate": 5.895257195877465e-05, "loss": 3.243, "step": 251300 }, { "epoch": 2.219661244458574, "grad_norm": 3.177061080932617, "learning_rate": 5.8939173445980634e-05, "loss": 3.2117, "step": 251350 }, { "epoch": 2.2201027923488583, "grad_norm": 1.6812225580215454, "learning_rate": 5.892550081058642e-05, "loss": 3.1777, "step": 251400 }, { "epoch": 2.2205443402391425, "grad_norm": 0.9270955920219421, "learning_rate": 5.891182748577294e-05, "loss": 3.4789, "step": 251450 }, { "epoch": 2.2209858881294267, "grad_norm": 1.8096257448196411, "learning_rate": 5.889815347259635e-05, "loss": 2.7918, "step": 251500 }, { "epoch": 2.221427436019711, "grad_norm": 4.176187992095947, "learning_rate": 5.8884478772112864e-05, "loss": 3.036, "step": 251550 }, { "epoch": 2.221868983909995, "grad_norm": 0.8330768346786499, "learning_rate": 5.887080338537873e-05, "loss": 3.2742, "step": 251600 }, { "epoch": 2.222310531800279, "grad_norm": 2.405538558959961, "learning_rate": 5.885712731345026e-05, "loss": 2.8618, "step": 251650 }, { "epoch": 2.2227520796905633, "grad_norm": 5.042022705078125, "learning_rate": 5.884345055738381e-05, "loss": 2.7369, "step": 251700 }, { "epoch": 2.2231936275808475, "grad_norm": 3.783979654312134, "learning_rate": 5.8829773118235774e-05, "loss": 3.2306, "step": 251750 }, { "epoch": 2.2236351754711317, "grad_norm": 2.477170705795288, "learning_rate": 5.881609499706267e-05, "loss": 2.8828, "step": 251800 }, { "epoch": 2.224076723361416, "grad_norm": 1.5696929693222046, "learning_rate": 5.8802416194920985e-05, "loss": 3.2756, "step": 251850 }, { "epoch": 2.2245182712517, "grad_norm": 2.808018684387207, "learning_rate": 5.8788736712867265e-05, "loss": 3.8227, "step": 251900 }, { "epoch": 2.224959819141984, "grad_norm": 4.302034854888916, "learning_rate": 5.877505655195818e-05, "loss": 3.4848, "step": 251950 }, { "epoch": 2.2254013670322683, "grad_norm": 1.6220154762268066, "learning_rate": 5.8761375713250376e-05, "loss": 3.2193, "step": 252000 }, { "epoch": 2.2254013670322683, "eval_asr_loss": 0.9349989803252292, "eval_loss": 2.84177303314209, "eval_runtime": 20.8268, "eval_samples_per_second": 36.876, "eval_steps_per_second": 9.219, "eval_tts_loss": 6.0021888255360905, "step": 252000 }, { "epoch": 2.2258429149225525, "grad_norm": 2.4672141075134277, "learning_rate": 5.874769419780061e-05, "loss": 3.1268, "step": 252050 }, { "epoch": 2.2262844628128367, "grad_norm": 0.6886611580848694, "learning_rate": 5.873401200666563e-05, "loss": 2.5239, "step": 252100 }, { "epoch": 2.226726010703121, "grad_norm": 2.589230537414551, "learning_rate": 5.8720329140902284e-05, "loss": 3.2045, "step": 252150 }, { "epoch": 2.227167558593405, "grad_norm": 1.5318831205368042, "learning_rate": 5.870664560156747e-05, "loss": 3.0298, "step": 252200 }, { "epoch": 2.227609106483689, "grad_norm": 2.816803216934204, "learning_rate": 5.869296138971809e-05, "loss": 3.2201, "step": 252250 }, { "epoch": 2.2280506543739733, "grad_norm": 2.9629454612731934, "learning_rate": 5.8679276506411184e-05, "loss": 3.2301, "step": 252300 }, { "epoch": 2.2284922022642575, "grad_norm": 3.7039413452148438, "learning_rate": 5.866559095270375e-05, "loss": 3.3573, "step": 252350 }, { "epoch": 2.2289337501545416, "grad_norm": 4.958855152130127, "learning_rate": 5.865190472965291e-05, "loss": 2.9811, "step": 252400 }, { "epoch": 2.229375298044826, "grad_norm": 4.706749439239502, "learning_rate": 5.8638491582684894e-05, "loss": 2.9408, "step": 252450 }, { "epoch": 2.22981684593511, "grad_norm": 2.872765064239502, "learning_rate": 5.8624804037452916e-05, "loss": 3.0163, "step": 252500 }, { "epoch": 2.230258393825394, "grad_norm": 2.8682596683502197, "learning_rate": 5.861111582602796e-05, "loss": 3.4345, "step": 252550 }, { "epoch": 2.2306999417156783, "grad_norm": 2.9680938720703125, "learning_rate": 5.8597426949467325e-05, "loss": 3.4903, "step": 252600 }, { "epoch": 2.2311414896059625, "grad_norm": 1.2532522678375244, "learning_rate": 5.858373740882837e-05, "loss": 2.7181, "step": 252650 }, { "epoch": 2.2315830374962466, "grad_norm": 1.0817298889160156, "learning_rate": 5.8570047205168475e-05, "loss": 2.8678, "step": 252700 }, { "epoch": 2.2320245853865313, "grad_norm": 4.356616497039795, "learning_rate": 5.8556356339545124e-05, "loss": 2.9539, "step": 252750 }, { "epoch": 2.232466133276815, "grad_norm": 1.9056625366210938, "learning_rate": 5.854266481301578e-05, "loss": 3.1982, "step": 252800 }, { "epoch": 2.2329076811670996, "grad_norm": 3.5170702934265137, "learning_rate": 5.8528972626638035e-05, "loss": 2.9899, "step": 252850 }, { "epoch": 2.2333492290573838, "grad_norm": 2.9625301361083984, "learning_rate": 5.851527978146948e-05, "loss": 3.3675, "step": 252900 }, { "epoch": 2.233790776947668, "grad_norm": 2.527756929397583, "learning_rate": 5.8501586278567755e-05, "loss": 2.8116, "step": 252950 }, { "epoch": 2.234232324837952, "grad_norm": 3.2555031776428223, "learning_rate": 5.848789211899058e-05, "loss": 2.7627, "step": 253000 }, { "epoch": 2.2346738727282363, "grad_norm": 1.6485633850097656, "learning_rate": 5.847419730379572e-05, "loss": 3.2575, "step": 253050 }, { "epoch": 2.2351154206185204, "grad_norm": 2.901740074157715, "learning_rate": 5.846050183404099e-05, "loss": 3.4547, "step": 253100 }, { "epoch": 2.2355569685088046, "grad_norm": 1.9311729669570923, "learning_rate": 5.8446805710784226e-05, "loss": 3.0989, "step": 253150 }, { "epoch": 2.2359985163990888, "grad_norm": 5.239856243133545, "learning_rate": 5.843310893508333e-05, "loss": 3.2647, "step": 253200 }, { "epoch": 2.236440064289373, "grad_norm": 3.5403478145599365, "learning_rate": 5.84194115079963e-05, "loss": 3.1479, "step": 253250 }, { "epoch": 2.236881612179657, "grad_norm": 2.1544852256774902, "learning_rate": 5.840571343058113e-05, "loss": 3.427, "step": 253300 }, { "epoch": 2.2373231600699413, "grad_norm": 2.9125540256500244, "learning_rate": 5.839201470389587e-05, "loss": 3.4527, "step": 253350 }, { "epoch": 2.2377647079602254, "grad_norm": 3.3283560276031494, "learning_rate": 5.837831532899863e-05, "loss": 3.0141, "step": 253400 }, { "epoch": 2.2382062558505096, "grad_norm": 4.432610511779785, "learning_rate": 5.836461530694759e-05, "loss": 2.6997, "step": 253450 }, { "epoch": 2.2386478037407938, "grad_norm": 3.704026699066162, "learning_rate": 5.835091463880094e-05, "loss": 3.5248, "step": 253500 }, { "epoch": 2.239089351631078, "grad_norm": 6.4238810539245605, "learning_rate": 5.833721332561694e-05, "loss": 3.3761, "step": 253550 }, { "epoch": 2.239530899521362, "grad_norm": 3.5807857513427734, "learning_rate": 5.8323511368453906e-05, "loss": 3.2106, "step": 253600 }, { "epoch": 2.2399724474116463, "grad_norm": 4.876890659332275, "learning_rate": 5.8309808768370214e-05, "loss": 3.2801, "step": 253650 }, { "epoch": 2.2404139953019304, "grad_norm": 4.392118453979492, "learning_rate": 5.829610552642426e-05, "loss": 3.2126, "step": 253700 }, { "epoch": 2.2408555431922146, "grad_norm": 3.744412422180176, "learning_rate": 5.82824016436745e-05, "loss": 3.1052, "step": 253750 }, { "epoch": 2.2412970910824987, "grad_norm": 3.0900723934173584, "learning_rate": 5.826869712117944e-05, "loss": 3.1165, "step": 253800 }, { "epoch": 2.241738638972783, "grad_norm": 2.942573308944702, "learning_rate": 5.825499195999765e-05, "loss": 3.0981, "step": 253850 }, { "epoch": 2.242180186863067, "grad_norm": 2.052508592605591, "learning_rate": 5.8241286161187737e-05, "loss": 3.4993, "step": 253900 }, { "epoch": 2.2426217347533512, "grad_norm": 2.6822502613067627, "learning_rate": 5.822757972580832e-05, "loss": 3.0703, "step": 253950 }, { "epoch": 2.2430632826436354, "grad_norm": 2.0350875854492188, "learning_rate": 5.821387265491817e-05, "loss": 3.3772, "step": 254000 }, { "epoch": 2.2435048305339196, "grad_norm": 4.792269229888916, "learning_rate": 5.8200164949575996e-05, "loss": 2.8125, "step": 254050 }, { "epoch": 2.2439463784242037, "grad_norm": 5.7412285804748535, "learning_rate": 5.8186456610840624e-05, "loss": 2.6822, "step": 254100 }, { "epoch": 2.244387926314488, "grad_norm": 7.473790168762207, "learning_rate": 5.81727476397709e-05, "loss": 3.1313, "step": 254150 }, { "epoch": 2.244829474204772, "grad_norm": 3.2918715476989746, "learning_rate": 5.8159038037425704e-05, "loss": 3.1051, "step": 254200 }, { "epoch": 2.2452710220950562, "grad_norm": 1.7990907430648804, "learning_rate": 5.814532780486402e-05, "loss": 3.1019, "step": 254250 }, { "epoch": 2.2457125699853404, "grad_norm": 3.1198318004608154, "learning_rate": 5.813161694314484e-05, "loss": 3.4328, "step": 254300 }, { "epoch": 2.2461541178756246, "grad_norm": 2.5575218200683594, "learning_rate": 5.811790545332719e-05, "loss": 3.2553, "step": 254350 }, { "epoch": 2.246595665765909, "grad_norm": 1.4484281539916992, "learning_rate": 5.810419333647019e-05, "loss": 3.2154, "step": 254400 }, { "epoch": 2.2470372136561934, "grad_norm": 4.266369342803955, "learning_rate": 5.809048059363298e-05, "loss": 3.2351, "step": 254450 }, { "epoch": 2.2474787615464775, "grad_norm": 2.723226308822632, "learning_rate": 5.807676722587474e-05, "loss": 3.0583, "step": 254500 }, { "epoch": 2.2479203094367617, "grad_norm": 2.533464193344116, "learning_rate": 5.806305323425473e-05, "loss": 2.9562, "step": 254550 }, { "epoch": 2.248361857327046, "grad_norm": 3.364504814147949, "learning_rate": 5.804933861983222e-05, "loss": 3.337, "step": 254600 }, { "epoch": 2.24880340521733, "grad_norm": 2.2846782207489014, "learning_rate": 5.803562338366657e-05, "loss": 2.9944, "step": 254650 }, { "epoch": 2.249244953107614, "grad_norm": 1.4509943723678589, "learning_rate": 5.8021907526817156e-05, "loss": 3.4156, "step": 254700 }, { "epoch": 2.2496865009978984, "grad_norm": 1.3979560136795044, "learning_rate": 5.800819105034338e-05, "loss": 3.1632, "step": 254750 }, { "epoch": 2.2501280488881825, "grad_norm": 6.0780110359191895, "learning_rate": 5.7994473955304786e-05, "loss": 2.7354, "step": 254800 }, { "epoch": 2.2505695967784667, "grad_norm": 2.125615119934082, "learning_rate": 5.798075624276085e-05, "loss": 2.7818, "step": 254850 }, { "epoch": 2.251011144668751, "grad_norm": 6.371503829956055, "learning_rate": 5.7967037913771185e-05, "loss": 3.0069, "step": 254900 }, { "epoch": 2.251452692559035, "grad_norm": 6.355823993682861, "learning_rate": 5.795331896939539e-05, "loss": 3.3613, "step": 254950 }, { "epoch": 2.251894240449319, "grad_norm": 3.9638240337371826, "learning_rate": 5.7939599410693134e-05, "loss": 3.2695, "step": 255000 }, { "epoch": 2.251894240449319, "eval_asr_loss": 0.9167550158655973, "eval_loss": 2.836369752883911, "eval_runtime": 20.729, "eval_samples_per_second": 37.05, "eval_steps_per_second": 9.262, "eval_tts_loss": 6.040006003285445, "step": 255000 }, { "epoch": 2.2523357883396034, "grad_norm": 5.163956642150879, "learning_rate": 5.792587923872418e-05, "loss": 2.7723, "step": 255050 }, { "epoch": 2.2527773362298875, "grad_norm": 2.9461264610290527, "learning_rate": 5.791215845454824e-05, "loss": 3.3687, "step": 255100 }, { "epoch": 2.2532188841201717, "grad_norm": 5.078152179718018, "learning_rate": 5.789843705922516e-05, "loss": 3.267, "step": 255150 }, { "epoch": 2.253660432010456, "grad_norm": 3.985874652862549, "learning_rate": 5.7884715053814794e-05, "loss": 3.0642, "step": 255200 }, { "epoch": 2.25410197990074, "grad_norm": 3.860677480697632, "learning_rate": 5.787099243937706e-05, "loss": 3.2701, "step": 255250 }, { "epoch": 2.254543527791024, "grad_norm": 2.267491102218628, "learning_rate": 5.785726921697189e-05, "loss": 2.7865, "step": 255300 }, { "epoch": 2.2549850756813083, "grad_norm": 1.3620892763137817, "learning_rate": 5.7843545387659305e-05, "loss": 2.7667, "step": 255350 }, { "epoch": 2.2554266235715925, "grad_norm": 4.178198337554932, "learning_rate": 5.7829820952499345e-05, "loss": 2.8948, "step": 255400 }, { "epoch": 2.2558681714618767, "grad_norm": 4.220424652099609, "learning_rate": 5.781609591255213e-05, "loss": 3.1193, "step": 255450 }, { "epoch": 2.256309719352161, "grad_norm": 4.8280510902404785, "learning_rate": 5.780237026887776e-05, "loss": 2.9872, "step": 255500 }, { "epoch": 2.256751267242445, "grad_norm": 3.40598464012146, "learning_rate": 5.778864402253646e-05, "loss": 3.1323, "step": 255550 }, { "epoch": 2.257192815132729, "grad_norm": 1.5590683221817017, "learning_rate": 5.777491717458846e-05, "loss": 3.5949, "step": 255600 }, { "epoch": 2.2576343630230133, "grad_norm": 4.328376770019531, "learning_rate": 5.776118972609403e-05, "loss": 3.1563, "step": 255650 }, { "epoch": 2.2580759109132975, "grad_norm": 1.7125028371810913, "learning_rate": 5.774746167811351e-05, "loss": 3.0802, "step": 255700 }, { "epoch": 2.2585174588035817, "grad_norm": 2.554837703704834, "learning_rate": 5.773373303170727e-05, "loss": 2.9005, "step": 255750 }, { "epoch": 2.258959006693866, "grad_norm": 3.7836754322052, "learning_rate": 5.772000378793574e-05, "loss": 3.1204, "step": 255800 }, { "epoch": 2.25940055458415, "grad_norm": 4.424196720123291, "learning_rate": 5.770627394785938e-05, "loss": 2.9913, "step": 255850 }, { "epoch": 2.259842102474434, "grad_norm": 5.351930141448975, "learning_rate": 5.76925435125387e-05, "loss": 3.1906, "step": 255900 }, { "epoch": 2.260283650364719, "grad_norm": 4.81837797164917, "learning_rate": 5.767881248303426e-05, "loss": 3.2567, "step": 255950 }, { "epoch": 2.2607251982550025, "grad_norm": 0.9730196595191956, "learning_rate": 5.766508086040667e-05, "loss": 2.691, "step": 256000 }, { "epoch": 2.261166746145287, "grad_norm": 2.4241783618927, "learning_rate": 5.765134864571659e-05, "loss": 3.2201, "step": 256050 }, { "epoch": 2.2616082940355713, "grad_norm": 1.6978586912155151, "learning_rate": 5.763761584002472e-05, "loss": 3.03, "step": 256100 }, { "epoch": 2.2620498419258555, "grad_norm": 2.6182703971862793, "learning_rate": 5.762388244439176e-05, "loss": 2.9427, "step": 256150 }, { "epoch": 2.2624913898161396, "grad_norm": 3.28235125541687, "learning_rate": 5.7610148459878565e-05, "loss": 2.7524, "step": 256200 }, { "epoch": 2.262932937706424, "grad_norm": 3.038163423538208, "learning_rate": 5.759641388754591e-05, "loss": 2.885, "step": 256250 }, { "epoch": 2.263374485596708, "grad_norm": 6.547210693359375, "learning_rate": 5.75826787284547e-05, "loss": 2.6655, "step": 256300 }, { "epoch": 2.263816033486992, "grad_norm": 2.0644569396972656, "learning_rate": 5.7568942983665865e-05, "loss": 2.9917, "step": 256350 }, { "epoch": 2.2642575813772763, "grad_norm": 2.1220710277557373, "learning_rate": 5.755520665424036e-05, "loss": 3.3596, "step": 256400 }, { "epoch": 2.2646991292675605, "grad_norm": 2.3702235221862793, "learning_rate": 5.7541469741239216e-05, "loss": 3.0393, "step": 256450 }, { "epoch": 2.2651406771578446, "grad_norm": 2.5091640949249268, "learning_rate": 5.752773224572347e-05, "loss": 2.9858, "step": 256500 }, { "epoch": 2.265582225048129, "grad_norm": 2.7721481323242188, "learning_rate": 5.751399416875425e-05, "loss": 2.7142, "step": 256550 }, { "epoch": 2.266023772938413, "grad_norm": 4.742951393127441, "learning_rate": 5.7500255511392706e-05, "loss": 3.4102, "step": 256600 }, { "epoch": 2.266465320828697, "grad_norm": 3.379053831100464, "learning_rate": 5.748651627470001e-05, "loss": 3.2227, "step": 256650 }, { "epoch": 2.2669068687189813, "grad_norm": 4.015255451202393, "learning_rate": 5.7473051261696856e-05, "loss": 2.7462, "step": 256700 }, { "epoch": 2.2673484166092654, "grad_norm": 1.5143048763275146, "learning_rate": 5.745931088105943e-05, "loss": 3.0989, "step": 256750 }, { "epoch": 2.2677899644995496, "grad_norm": 3.181206226348877, "learning_rate": 5.744556992425347e-05, "loss": 3.2307, "step": 256800 }, { "epoch": 2.268231512389834, "grad_norm": 2.5234804153442383, "learning_rate": 5.7431828392340394e-05, "loss": 2.9222, "step": 256850 }, { "epoch": 2.268673060280118, "grad_norm": 3.6337316036224365, "learning_rate": 5.741808628638159e-05, "loss": 2.9027, "step": 256900 }, { "epoch": 2.269114608170402, "grad_norm": 3.142951488494873, "learning_rate": 5.740434360743852e-05, "loss": 3.5821, "step": 256950 }, { "epoch": 2.2695561560606863, "grad_norm": 0.9363364577293396, "learning_rate": 5.739060035657269e-05, "loss": 3.2935, "step": 257000 }, { "epoch": 2.2699977039509704, "grad_norm": 2.494088649749756, "learning_rate": 5.737685653484565e-05, "loss": 2.9355, "step": 257050 }, { "epoch": 2.2704392518412546, "grad_norm": 1.426322102546692, "learning_rate": 5.736311214331902e-05, "loss": 2.9509, "step": 257100 }, { "epoch": 2.270880799731539, "grad_norm": 1.4795937538146973, "learning_rate": 5.734936718305439e-05, "loss": 2.9341, "step": 257150 }, { "epoch": 2.271322347621823, "grad_norm": 4.622045993804932, "learning_rate": 5.733562165511347e-05, "loss": 2.8467, "step": 257200 }, { "epoch": 2.271763895512107, "grad_norm": 1.1926823854446411, "learning_rate": 5.732187556055798e-05, "loss": 2.7719, "step": 257250 }, { "epoch": 2.2722054434023913, "grad_norm": 3.47904634475708, "learning_rate": 5.7308128900449695e-05, "loss": 3.3598, "step": 257300 }, { "epoch": 2.2726469912926754, "grad_norm": 4.154747009277344, "learning_rate": 5.729438167585043e-05, "loss": 3.5488, "step": 257350 }, { "epoch": 2.2730885391829596, "grad_norm": 2.0710361003875732, "learning_rate": 5.7280633887822034e-05, "loss": 2.9576, "step": 257400 }, { "epoch": 2.2735300870732438, "grad_norm": 1.667640209197998, "learning_rate": 5.72668855374264e-05, "loss": 3.0993, "step": 257450 }, { "epoch": 2.2739716349635284, "grad_norm": 2.4730777740478516, "learning_rate": 5.725313662572549e-05, "loss": 3.0302, "step": 257500 }, { "epoch": 2.274413182853812, "grad_norm": 4.713761806488037, "learning_rate": 5.7239387153781255e-05, "loss": 3.0969, "step": 257550 }, { "epoch": 2.2748547307440967, "grad_norm": 3.9501495361328125, "learning_rate": 5.7225637122655776e-05, "loss": 3.1804, "step": 257600 }, { "epoch": 2.2752962786343804, "grad_norm": 7.691797256469727, "learning_rate": 5.72118865334111e-05, "loss": 3.1689, "step": 257650 }, { "epoch": 2.275737826524665, "grad_norm": 2.7920868396759033, "learning_rate": 5.719813538710933e-05, "loss": 2.7669, "step": 257700 }, { "epoch": 2.276179374414949, "grad_norm": 3.043945550918579, "learning_rate": 5.718438368481265e-05, "loss": 3.2661, "step": 257750 }, { "epoch": 2.2766209223052334, "grad_norm": 2.6614737510681152, "learning_rate": 5.7170631427583244e-05, "loss": 3.4721, "step": 257800 }, { "epoch": 2.2770624701955176, "grad_norm": 1.1885290145874023, "learning_rate": 5.715687861648338e-05, "loss": 2.8085, "step": 257850 }, { "epoch": 2.2775040180858017, "grad_norm": 3.0696117877960205, "learning_rate": 5.7143125252575325e-05, "loss": 3.1958, "step": 257900 }, { "epoch": 2.277945565976086, "grad_norm": 3.802065134048462, "learning_rate": 5.712937133692141e-05, "loss": 3.0814, "step": 257950 }, { "epoch": 2.27838711386637, "grad_norm": 3.4466001987457275, "learning_rate": 5.7115616870584024e-05, "loss": 3.0659, "step": 258000 }, { "epoch": 2.27838711386637, "eval_asr_loss": 0.918105379332637, "eval_loss": 2.826794385910034, "eval_runtime": 20.7768, "eval_samples_per_second": 36.964, "eval_steps_per_second": 9.241, "eval_tts_loss": 6.010721179663374, "step": 258000 }, { "epoch": 2.278828661756654, "grad_norm": 3.1393704414367676, "learning_rate": 5.710186185462556e-05, "loss": 3.3603, "step": 258050 }, { "epoch": 2.2792702096469384, "grad_norm": 3.025592088699341, "learning_rate": 5.70881062901085e-05, "loss": 3.3545, "step": 258100 }, { "epoch": 2.2797117575372225, "grad_norm": 2.151677131652832, "learning_rate": 5.707435017809534e-05, "loss": 3.6102, "step": 258150 }, { "epoch": 2.2801533054275067, "grad_norm": 2.546576738357544, "learning_rate": 5.7060593519648596e-05, "loss": 2.8791, "step": 258200 }, { "epoch": 2.280594853317791, "grad_norm": 3.4382550716400146, "learning_rate": 5.704683631583089e-05, "loss": 3.2175, "step": 258250 }, { "epoch": 2.281036401208075, "grad_norm": 7.082617282867432, "learning_rate": 5.703307856770485e-05, "loss": 2.7008, "step": 258300 }, { "epoch": 2.281477949098359, "grad_norm": 1.0854841470718384, "learning_rate": 5.70193202763331e-05, "loss": 3.1032, "step": 258350 }, { "epoch": 2.2819194969886434, "grad_norm": 3.358534336090088, "learning_rate": 5.700556144277839e-05, "loss": 2.8708, "step": 258400 }, { "epoch": 2.2823610448789275, "grad_norm": 2.6743767261505127, "learning_rate": 5.699180206810347e-05, "loss": 2.8837, "step": 258450 }, { "epoch": 2.2828025927692117, "grad_norm": 4.866109371185303, "learning_rate": 5.697804215337113e-05, "loss": 3.2468, "step": 258500 }, { "epoch": 2.283244140659496, "grad_norm": 4.143941879272461, "learning_rate": 5.6964281699644194e-05, "loss": 3.0353, "step": 258550 }, { "epoch": 2.28368568854978, "grad_norm": 1.2687678337097168, "learning_rate": 5.6950520707985554e-05, "loss": 3.2586, "step": 258600 }, { "epoch": 2.284127236440064, "grad_norm": 3.731055974960327, "learning_rate": 5.6936759179458134e-05, "loss": 3.171, "step": 258650 }, { "epoch": 2.2845687843303484, "grad_norm": 5.414038181304932, "learning_rate": 5.69229971151249e-05, "loss": 2.9279, "step": 258700 }, { "epoch": 2.2850103322206325, "grad_norm": 1.276946783065796, "learning_rate": 5.6909234516048826e-05, "loss": 3.5517, "step": 258750 }, { "epoch": 2.2854518801109167, "grad_norm": 5.130809307098389, "learning_rate": 5.6895471383292985e-05, "loss": 2.6182, "step": 258800 }, { "epoch": 2.285893428001201, "grad_norm": 3.0289876461029053, "learning_rate": 5.6881707717920454e-05, "loss": 3.1079, "step": 258850 }, { "epoch": 2.286334975891485, "grad_norm": 1.6382182836532593, "learning_rate": 5.686821881013521e-05, "loss": 3.1695, "step": 258900 }, { "epoch": 2.286776523781769, "grad_norm": 3.9014601707458496, "learning_rate": 5.68544540933181e-05, "loss": 3.1019, "step": 258950 }, { "epoch": 2.2872180716720534, "grad_norm": 2.067531108856201, "learning_rate": 5.684068884705254e-05, "loss": 2.6226, "step": 259000 }, { "epoch": 2.2876596195623375, "grad_norm": 2.9713211059570312, "learning_rate": 5.682692307240176e-05, "loss": 2.7708, "step": 259050 }, { "epoch": 2.2881011674526217, "grad_norm": 2.8738832473754883, "learning_rate": 5.681315677042906e-05, "loss": 3.0192, "step": 259100 }, { "epoch": 2.2885427153429063, "grad_norm": 1.064266562461853, "learning_rate": 5.67993899421978e-05, "loss": 3.2214, "step": 259150 }, { "epoch": 2.28898426323319, "grad_norm": 2.092491865158081, "learning_rate": 5.67856225887713e-05, "loss": 3.165, "step": 259200 }, { "epoch": 2.2894258111234747, "grad_norm": 4.751421928405762, "learning_rate": 5.677185471121299e-05, "loss": 2.9537, "step": 259250 }, { "epoch": 2.289867359013759, "grad_norm": 4.3732075691223145, "learning_rate": 5.6758086310586344e-05, "loss": 2.8884, "step": 259300 }, { "epoch": 2.290308906904043, "grad_norm": 3.341885805130005, "learning_rate": 5.674431738795484e-05, "loss": 3.1115, "step": 259350 }, { "epoch": 2.290750454794327, "grad_norm": 2.119860887527466, "learning_rate": 5.6730547944381994e-05, "loss": 2.6533, "step": 259400 }, { "epoch": 2.2911920026846113, "grad_norm": 2.396080255508423, "learning_rate": 5.67167779809314e-05, "loss": 2.6871, "step": 259450 }, { "epoch": 2.2916335505748955, "grad_norm": 2.1403701305389404, "learning_rate": 5.670300749866666e-05, "loss": 3.2967, "step": 259500 }, { "epoch": 2.2920750984651797, "grad_norm": 2.3258633613586426, "learning_rate": 5.6689236498651445e-05, "loss": 3.2124, "step": 259550 }, { "epoch": 2.292516646355464, "grad_norm": 4.450578212738037, "learning_rate": 5.6675464981949424e-05, "loss": 3.0109, "step": 259600 }, { "epoch": 2.292958194245748, "grad_norm": 3.3391926288604736, "learning_rate": 5.666169294962432e-05, "loss": 3.0317, "step": 259650 }, { "epoch": 2.293399742136032, "grad_norm": 3.5569305419921875, "learning_rate": 5.664792040273995e-05, "loss": 3.3952, "step": 259700 }, { "epoch": 2.2938412900263163, "grad_norm": 2.806424379348755, "learning_rate": 5.66341473423601e-05, "loss": 3.1943, "step": 259750 }, { "epoch": 2.2942828379166005, "grad_norm": 1.9792596101760864, "learning_rate": 5.662037376954862e-05, "loss": 2.7434, "step": 259800 }, { "epoch": 2.2947243858068846, "grad_norm": 2.8331778049468994, "learning_rate": 5.6606599685369396e-05, "loss": 3.6059, "step": 259850 }, { "epoch": 2.295165933697169, "grad_norm": 2.8929219245910645, "learning_rate": 5.659282509088637e-05, "loss": 2.5122, "step": 259900 }, { "epoch": 2.295607481587453, "grad_norm": 6.325002670288086, "learning_rate": 5.657904998716353e-05, "loss": 3.2667, "step": 259950 }, { "epoch": 2.296049029477737, "grad_norm": 3.9959585666656494, "learning_rate": 5.656554989247607e-05, "loss": 3.2405, "step": 260000 }, { "epoch": 2.2964905773680213, "grad_norm": 2.478994369506836, "learning_rate": 5.6551773783597426e-05, "loss": 3.2724, "step": 260050 }, { "epoch": 2.2969321252583055, "grad_norm": 7.020759582519531, "learning_rate": 5.6537997168649815e-05, "loss": 3.1246, "step": 260100 }, { "epoch": 2.2973736731485896, "grad_norm": 4.527580738067627, "learning_rate": 5.6524220048697365e-05, "loss": 2.9518, "step": 260150 }, { "epoch": 2.297815221038874, "grad_norm": 2.7438511848449707, "learning_rate": 5.651044242480423e-05, "loss": 2.6793, "step": 260200 }, { "epoch": 2.298256768929158, "grad_norm": 4.815626621246338, "learning_rate": 5.649666429803464e-05, "loss": 2.8741, "step": 260250 }, { "epoch": 2.298698316819442, "grad_norm": 3.471956491470337, "learning_rate": 5.648288566945282e-05, "loss": 2.8175, "step": 260300 }, { "epoch": 2.2991398647097263, "grad_norm": 8.08088493347168, "learning_rate": 5.646910654012306e-05, "loss": 3.0679, "step": 260350 }, { "epoch": 2.2995814126000105, "grad_norm": 2.5349581241607666, "learning_rate": 5.6455326911109674e-05, "loss": 2.8856, "step": 260400 }, { "epoch": 2.3000229604902946, "grad_norm": 3.89231538772583, "learning_rate": 5.644154678347702e-05, "loss": 3.1039, "step": 260450 }, { "epoch": 2.300464508380579, "grad_norm": 2.5363729000091553, "learning_rate": 5.6427766158289494e-05, "loss": 2.7886, "step": 260500 }, { "epoch": 2.300906056270863, "grad_norm": 2.0707285404205322, "learning_rate": 5.641398503661154e-05, "loss": 3.0274, "step": 260550 }, { "epoch": 2.301347604161147, "grad_norm": 2.4267489910125732, "learning_rate": 5.6400203419507633e-05, "loss": 3.3627, "step": 260600 }, { "epoch": 2.3017891520514313, "grad_norm": 2.2621796131134033, "learning_rate": 5.638642130804228e-05, "loss": 3.1289, "step": 260650 }, { "epoch": 2.302230699941716, "grad_norm": 3.468550205230713, "learning_rate": 5.6372638703280045e-05, "loss": 3.4757, "step": 260700 }, { "epoch": 2.3026722478319996, "grad_norm": 1.555128812789917, "learning_rate": 5.635885560628551e-05, "loss": 3.2301, "step": 260750 }, { "epoch": 2.3031137957222843, "grad_norm": 4.752922058105469, "learning_rate": 5.6345072018123294e-05, "loss": 3.7805, "step": 260800 }, { "epoch": 2.303555343612568, "grad_norm": 4.242067337036133, "learning_rate": 5.633128793985808e-05, "loss": 3.3764, "step": 260850 }, { "epoch": 2.3039968915028526, "grad_norm": 4.379598140716553, "learning_rate": 5.6317503372554546e-05, "loss": 2.776, "step": 260900 }, { "epoch": 2.3044384393931368, "grad_norm": 2.2684667110443115, "learning_rate": 5.6303718317277444e-05, "loss": 2.8038, "step": 260950 }, { "epoch": 2.304879987283421, "grad_norm": 1.9367856979370117, "learning_rate": 5.628993277509157e-05, "loss": 3.3306, "step": 261000 }, { "epoch": 2.304879987283421, "eval_asr_loss": 0.9354507490389332, "eval_loss": 2.8315999507904053, "eval_runtime": 20.34, "eval_samples_per_second": 37.758, "eval_steps_per_second": 9.44, "eval_tts_loss": 5.994351006919104, "step": 261000 }, { "epoch": 2.305321535173705, "grad_norm": 3.824610710144043, "learning_rate": 5.627614674706172e-05, "loss": 2.7628, "step": 261050 }, { "epoch": 2.3057630830639892, "grad_norm": 4.425666332244873, "learning_rate": 5.626236023425275e-05, "loss": 3.1435, "step": 261100 }, { "epoch": 2.3062046309542734, "grad_norm": 3.835205554962158, "learning_rate": 5.624857323772956e-05, "loss": 3.0374, "step": 261150 }, { "epoch": 2.3066461788445576, "grad_norm": 1.3904534578323364, "learning_rate": 5.6234785758557074e-05, "loss": 3.2103, "step": 261200 }, { "epoch": 2.3070877267348417, "grad_norm": 4.534204483032227, "learning_rate": 5.622099779780027e-05, "loss": 3.1799, "step": 261250 }, { "epoch": 2.307529274625126, "grad_norm": 4.1542253494262695, "learning_rate": 5.6207209356524113e-05, "loss": 3.4636, "step": 261300 }, { "epoch": 2.30797082251541, "grad_norm": 2.777684450149536, "learning_rate": 5.6193420435793685e-05, "loss": 3.0215, "step": 261350 }, { "epoch": 2.3084123704056942, "grad_norm": 1.5122051239013672, "learning_rate": 5.617963103667404e-05, "loss": 3.2348, "step": 261400 }, { "epoch": 2.3088539182959784, "grad_norm": 2.424396514892578, "learning_rate": 5.616584116023032e-05, "loss": 2.7274, "step": 261450 }, { "epoch": 2.3092954661862626, "grad_norm": 1.3294912576675415, "learning_rate": 5.615205080752763e-05, "loss": 3.0335, "step": 261500 }, { "epoch": 2.3097370140765467, "grad_norm": 2.8590776920318604, "learning_rate": 5.613825997963119e-05, "loss": 2.6329, "step": 261550 }, { "epoch": 2.310178561966831, "grad_norm": 3.0827012062072754, "learning_rate": 5.612446867760621e-05, "loss": 3.2388, "step": 261600 }, { "epoch": 2.310620109857115, "grad_norm": 2.782808780670166, "learning_rate": 5.6110676902517975e-05, "loss": 2.6786, "step": 261650 }, { "epoch": 2.3110616577473992, "grad_norm": 4.4579925537109375, "learning_rate": 5.609688465543176e-05, "loss": 3.0859, "step": 261700 }, { "epoch": 2.3115032056376834, "grad_norm": 0.8005117774009705, "learning_rate": 5.60830919374129e-05, "loss": 3.1809, "step": 261750 }, { "epoch": 2.3119447535279676, "grad_norm": 5.154500961303711, "learning_rate": 5.606929874952678e-05, "loss": 2.8245, "step": 261800 }, { "epoch": 2.3123863014182517, "grad_norm": 2.2137515544891357, "learning_rate": 5.605550509283879e-05, "loss": 3.1944, "step": 261850 }, { "epoch": 2.312827849308536, "grad_norm": 3.4326372146606445, "learning_rate": 5.6041710968414385e-05, "loss": 3.2534, "step": 261900 }, { "epoch": 2.31326939719882, "grad_norm": 4.479403495788574, "learning_rate": 5.6027916377319036e-05, "loss": 2.7042, "step": 261950 }, { "epoch": 2.3137109450891042, "grad_norm": 11.083107948303223, "learning_rate": 5.6014121320618275e-05, "loss": 2.9227, "step": 262000 }, { "epoch": 2.3141524929793884, "grad_norm": 1.5570802688598633, "learning_rate": 5.600032579937763e-05, "loss": 2.7361, "step": 262050 }, { "epoch": 2.3145940408696726, "grad_norm": 2.347303867340088, "learning_rate": 5.598652981466268e-05, "loss": 3.0655, "step": 262100 }, { "epoch": 2.3150355887599567, "grad_norm": 2.3728644847869873, "learning_rate": 5.5972733367539086e-05, "loss": 3.3828, "step": 262150 }, { "epoch": 2.315477136650241, "grad_norm": 4.1391496658325195, "learning_rate": 5.5958936459072486e-05, "loss": 3.397, "step": 262200 }, { "epoch": 2.315918684540525, "grad_norm": 2.413208484649658, "learning_rate": 5.594513909032858e-05, "loss": 2.9365, "step": 262250 }, { "epoch": 2.3163602324308092, "grad_norm": 2.8490779399871826, "learning_rate": 5.5931341262373094e-05, "loss": 2.8716, "step": 262300 }, { "epoch": 2.316801780321094, "grad_norm": 4.042685508728027, "learning_rate": 5.591754297627177e-05, "loss": 2.5153, "step": 262350 }, { "epoch": 2.3172433282113776, "grad_norm": 1.9091929197311401, "learning_rate": 5.590374423309046e-05, "loss": 3.1697, "step": 262400 }, { "epoch": 2.317684876101662, "grad_norm": 2.742373466491699, "learning_rate": 5.588994503389495e-05, "loss": 3.2971, "step": 262450 }, { "epoch": 2.3181264239919463, "grad_norm": 3.162210464477539, "learning_rate": 5.587614537975113e-05, "loss": 2.7407, "step": 262500 }, { "epoch": 2.3185679718822305, "grad_norm": 5.701384544372559, "learning_rate": 5.5862345271724915e-05, "loss": 2.8726, "step": 262550 }, { "epoch": 2.3190095197725147, "grad_norm": 3.7417452335357666, "learning_rate": 5.584854471088223e-05, "loss": 3.0792, "step": 262600 }, { "epoch": 2.319451067662799, "grad_norm": 2.8303890228271484, "learning_rate": 5.583474369828907e-05, "loss": 2.909, "step": 262650 }, { "epoch": 2.319892615553083, "grad_norm": 2.124849796295166, "learning_rate": 5.582094223501144e-05, "loss": 3.4017, "step": 262700 }, { "epoch": 2.320334163443367, "grad_norm": 2.6295056343078613, "learning_rate": 5.580714032211537e-05, "loss": 3.0611, "step": 262750 }, { "epoch": 2.3207757113336513, "grad_norm": 4.357548713684082, "learning_rate": 5.579333796066697e-05, "loss": 2.8004, "step": 262800 }, { "epoch": 2.3212172592239355, "grad_norm": 2.4115006923675537, "learning_rate": 5.577953515173232e-05, "loss": 3.1359, "step": 262850 }, { "epoch": 2.3216588071142197, "grad_norm": 4.313747882843018, "learning_rate": 5.57657318963776e-05, "loss": 2.901, "step": 262900 }, { "epoch": 2.322100355004504, "grad_norm": 3.3670005798339844, "learning_rate": 5.575192819566898e-05, "loss": 2.9926, "step": 262950 }, { "epoch": 2.322541902894788, "grad_norm": 4.261048793792725, "learning_rate": 5.573812405067269e-05, "loss": 3.1137, "step": 263000 }, { "epoch": 2.322983450785072, "grad_norm": 2.3910257816314697, "learning_rate": 5.572431946245497e-05, "loss": 3.0283, "step": 263050 }, { "epoch": 2.3234249986753563, "grad_norm": 5.147556304931641, "learning_rate": 5.57105144320821e-05, "loss": 3.0282, "step": 263100 }, { "epoch": 2.3238665465656405, "grad_norm": 3.097411632537842, "learning_rate": 5.5696708960620425e-05, "loss": 3.1169, "step": 263150 }, { "epoch": 2.3243080944559247, "grad_norm": 2.724949359893799, "learning_rate": 5.568290304913629e-05, "loss": 3.2258, "step": 263200 }, { "epoch": 2.324749642346209, "grad_norm": 1.34285569190979, "learning_rate": 5.566909669869608e-05, "loss": 2.9444, "step": 263250 }, { "epoch": 2.325191190236493, "grad_norm": 3.5121731758117676, "learning_rate": 5.565528991036622e-05, "loss": 3.3111, "step": 263300 }, { "epoch": 2.325632738126777, "grad_norm": 4.275054931640625, "learning_rate": 5.5641482685213174e-05, "loss": 2.8908, "step": 263350 }, { "epoch": 2.3260742860170613, "grad_norm": 1.9796489477157593, "learning_rate": 5.5627675024303436e-05, "loss": 2.8192, "step": 263400 }, { "epoch": 2.3265158339073455, "grad_norm": 3.5042240619659424, "learning_rate": 5.561386692870352e-05, "loss": 2.9817, "step": 263450 }, { "epoch": 2.3269573817976297, "grad_norm": 5.757020950317383, "learning_rate": 5.560005839947997e-05, "loss": 3.1829, "step": 263500 }, { "epoch": 2.327398929687914, "grad_norm": 4.141802787780762, "learning_rate": 5.558624943769941e-05, "loss": 3.3263, "step": 263550 }, { "epoch": 2.327840477578198, "grad_norm": 2.35674786567688, "learning_rate": 5.557244004442846e-05, "loss": 3.4281, "step": 263600 }, { "epoch": 2.328282025468482, "grad_norm": 1.373637080192566, "learning_rate": 5.5558630220733766e-05, "loss": 2.8377, "step": 263650 }, { "epoch": 2.3287235733587663, "grad_norm": 1.4962685108184814, "learning_rate": 5.554481996768202e-05, "loss": 3.1553, "step": 263700 }, { "epoch": 2.3291651212490505, "grad_norm": 4.954227447509766, "learning_rate": 5.553100928633994e-05, "loss": 2.9659, "step": 263750 }, { "epoch": 2.3296066691393347, "grad_norm": 1.6870756149291992, "learning_rate": 5.551719817777431e-05, "loss": 2.9963, "step": 263800 }, { "epoch": 2.330048217029619, "grad_norm": 2.317361831665039, "learning_rate": 5.5503386643051906e-05, "loss": 2.9823, "step": 263850 }, { "epoch": 2.3304897649199035, "grad_norm": 1.8838040828704834, "learning_rate": 5.5489574683239524e-05, "loss": 2.7254, "step": 263900 }, { "epoch": 2.330931312810187, "grad_norm": 3.2595179080963135, "learning_rate": 5.547576229940407e-05, "loss": 3.31, "step": 263950 }, { "epoch": 2.331372860700472, "grad_norm": 3.8841443061828613, "learning_rate": 5.5461949492612406e-05, "loss": 3.3577, "step": 264000 }, { "epoch": 2.331372860700472, "eval_asr_loss": 0.9167385954505524, "eval_loss": 2.828986406326294, "eval_runtime": 20.6332, "eval_samples_per_second": 37.222, "eval_steps_per_second": 9.305, "eval_tts_loss": 6.048375752071973, "step": 264000 }, { "epoch": 2.3318144085907555, "grad_norm": 2.737771987915039, "learning_rate": 5.544813626393147e-05, "loss": 2.7298, "step": 264050 }, { "epoch": 2.33225595648104, "grad_norm": 2.2012710571289062, "learning_rate": 5.543459889153543e-05, "loss": 2.9061, "step": 264100 }, { "epoch": 2.3326975043713243, "grad_norm": 5.5074310302734375, "learning_rate": 5.542078483066145e-05, "loss": 2.8751, "step": 264150 }, { "epoch": 2.3331390522616084, "grad_norm": 4.178867816925049, "learning_rate": 5.5406970351077834e-05, "loss": 3.0543, "step": 264200 }, { "epoch": 2.3335806001518926, "grad_norm": 4.458735942840576, "learning_rate": 5.53931554538516e-05, "loss": 3.3491, "step": 264250 }, { "epoch": 2.334022148042177, "grad_norm": 2.34615421295166, "learning_rate": 5.537934014004984e-05, "loss": 3.072, "step": 264300 }, { "epoch": 2.334463695932461, "grad_norm": 2.2701241970062256, "learning_rate": 5.5365524410739676e-05, "loss": 3.0502, "step": 264350 }, { "epoch": 2.334905243822745, "grad_norm": 2.0921719074249268, "learning_rate": 5.535170826698824e-05, "loss": 3.3448, "step": 264400 }, { "epoch": 2.3353467917130293, "grad_norm": 2.956352710723877, "learning_rate": 5.5337891709862724e-05, "loss": 3.7862, "step": 264450 }, { "epoch": 2.3357883396033134, "grad_norm": 4.650622367858887, "learning_rate": 5.532407474043035e-05, "loss": 3.5074, "step": 264500 }, { "epoch": 2.3362298874935976, "grad_norm": 3.9611144065856934, "learning_rate": 5.531025735975833e-05, "loss": 3.0186, "step": 264550 }, { "epoch": 2.336671435383882, "grad_norm": 2.757202625274658, "learning_rate": 5.529643956891397e-05, "loss": 3.0246, "step": 264600 }, { "epoch": 2.337112983274166, "grad_norm": 1.2094484567642212, "learning_rate": 5.528262136896456e-05, "loss": 3.1786, "step": 264650 }, { "epoch": 2.33755453116445, "grad_norm": 1.3035411834716797, "learning_rate": 5.5268802760977457e-05, "loss": 2.6192, "step": 264700 }, { "epoch": 2.3379960790547343, "grad_norm": 3.1371283531188965, "learning_rate": 5.5254983746020005e-05, "loss": 3.1407, "step": 264750 }, { "epoch": 2.3384376269450184, "grad_norm": 2.692570686340332, "learning_rate": 5.524116432515962e-05, "loss": 2.8352, "step": 264800 }, { "epoch": 2.3388791748353026, "grad_norm": 2.4653825759887695, "learning_rate": 5.522734449946375e-05, "loss": 3.1506, "step": 264850 }, { "epoch": 2.3393207227255868, "grad_norm": 2.7341175079345703, "learning_rate": 5.521352426999983e-05, "loss": 3.2424, "step": 264900 }, { "epoch": 2.339762270615871, "grad_norm": 5.063891410827637, "learning_rate": 5.5199703637835367e-05, "loss": 2.9466, "step": 264950 }, { "epoch": 2.340203818506155, "grad_norm": 2.0876517295837402, "learning_rate": 5.51858826040379e-05, "loss": 3.0843, "step": 265000 }, { "epoch": 2.3406453663964393, "grad_norm": 4.019355297088623, "learning_rate": 5.517206116967497e-05, "loss": 2.7642, "step": 265050 }, { "epoch": 2.3410869142867234, "grad_norm": 2.348693609237671, "learning_rate": 5.515823933581418e-05, "loss": 2.9398, "step": 265100 }, { "epoch": 2.3415284621770076, "grad_norm": 3.2666256427764893, "learning_rate": 5.514441710352314e-05, "loss": 3.1423, "step": 265150 }, { "epoch": 2.3419700100672918, "grad_norm": 5.016966819763184, "learning_rate": 5.513059447386948e-05, "loss": 3.0213, "step": 265200 }, { "epoch": 2.342411557957576, "grad_norm": 2.02489972114563, "learning_rate": 5.511677144792094e-05, "loss": 3.0181, "step": 265250 }, { "epoch": 2.34285310584786, "grad_norm": 4.673335075378418, "learning_rate": 5.510294802674516e-05, "loss": 2.7512, "step": 265300 }, { "epoch": 2.3432946537381443, "grad_norm": 1.0106866359710693, "learning_rate": 5.508912421140994e-05, "loss": 3.1089, "step": 265350 }, { "epoch": 2.3437362016284284, "grad_norm": 4.392617702484131, "learning_rate": 5.5075300002983024e-05, "loss": 2.6871, "step": 265400 }, { "epoch": 2.3441777495187126, "grad_norm": 3.6103897094726562, "learning_rate": 5.5061475402532214e-05, "loss": 2.8084, "step": 265450 }, { "epoch": 2.3446192974089968, "grad_norm": 3.152289867401123, "learning_rate": 5.504765041112536e-05, "loss": 3.015, "step": 265500 }, { "epoch": 2.3450608452992814, "grad_norm": 3.9061062335968018, "learning_rate": 5.503382502983031e-05, "loss": 3.2465, "step": 265550 }, { "epoch": 2.345502393189565, "grad_norm": 1.829247236251831, "learning_rate": 5.501999925971495e-05, "loss": 2.9146, "step": 265600 }, { "epoch": 2.3459439410798497, "grad_norm": 2.7599451541900635, "learning_rate": 5.5006173101847236e-05, "loss": 3.2354, "step": 265650 }, { "epoch": 2.346385488970134, "grad_norm": 4.144500732421875, "learning_rate": 5.499234655729508e-05, "loss": 2.7549, "step": 265700 }, { "epoch": 2.346827036860418, "grad_norm": 2.0739288330078125, "learning_rate": 5.49785196271265e-05, "loss": 3.2811, "step": 265750 }, { "epoch": 2.347268584750702, "grad_norm": 2.2752270698547363, "learning_rate": 5.496469231240949e-05, "loss": 3.4223, "step": 265800 }, { "epoch": 2.3477101326409864, "grad_norm": 4.951728820800781, "learning_rate": 5.49508646142121e-05, "loss": 3.0825, "step": 265850 }, { "epoch": 2.3481516805312705, "grad_norm": 2.364725112915039, "learning_rate": 5.49370365336024e-05, "loss": 3.3432, "step": 265900 }, { "epoch": 2.3485932284215547, "grad_norm": 1.3857911825180054, "learning_rate": 5.492320807164849e-05, "loss": 2.8634, "step": 265950 }, { "epoch": 2.349034776311839, "grad_norm": 2.826765298843384, "learning_rate": 5.490937922941851e-05, "loss": 3.213, "step": 266000 }, { "epoch": 2.349476324202123, "grad_norm": 3.6707797050476074, "learning_rate": 5.489555000798062e-05, "loss": 3.4422, "step": 266050 }, { "epoch": 2.349917872092407, "grad_norm": 1.660170078277588, "learning_rate": 5.4881720408402995e-05, "loss": 3.082, "step": 266100 }, { "epoch": 2.3503594199826914, "grad_norm": 1.5330767631530762, "learning_rate": 5.4867890431753865e-05, "loss": 3.1163, "step": 266150 }, { "epoch": 2.3508009678729755, "grad_norm": 2.553494691848755, "learning_rate": 5.485406007910149e-05, "loss": 2.994, "step": 266200 }, { "epoch": 2.3512425157632597, "grad_norm": 3.286794900894165, "learning_rate": 5.4840229351514116e-05, "loss": 3.6846, "step": 266250 }, { "epoch": 2.351684063653544, "grad_norm": 3.5662360191345215, "learning_rate": 5.4826398250060086e-05, "loss": 3.0479, "step": 266300 }, { "epoch": 2.352125611543828, "grad_norm": 3.3685927391052246, "learning_rate": 5.481256677580769e-05, "loss": 2.966, "step": 266350 }, { "epoch": 2.352567159434112, "grad_norm": 3.6049554347991943, "learning_rate": 5.479873492982534e-05, "loss": 3.139, "step": 266400 }, { "epoch": 2.3530087073243964, "grad_norm": 4.308905124664307, "learning_rate": 5.4784902713181416e-05, "loss": 2.8155, "step": 266450 }, { "epoch": 2.3534502552146805, "grad_norm": 5.833682537078857, "learning_rate": 5.4771070126944326e-05, "loss": 3.2216, "step": 266500 }, { "epoch": 2.3538918031049647, "grad_norm": 3.323594570159912, "learning_rate": 5.475723717218253e-05, "loss": 3.1389, "step": 266550 }, { "epoch": 2.354333350995249, "grad_norm": 1.5683462619781494, "learning_rate": 5.474340384996449e-05, "loss": 3.071, "step": 266600 }, { "epoch": 2.354774898885533, "grad_norm": 4.281044006347656, "learning_rate": 5.4729570161358736e-05, "loss": 3.1891, "step": 266650 }, { "epoch": 2.355216446775817, "grad_norm": 1.4462392330169678, "learning_rate": 5.471573610743379e-05, "loss": 3.3321, "step": 266700 }, { "epoch": 2.3556579946661014, "grad_norm": 1.8511630296707153, "learning_rate": 5.470190168925821e-05, "loss": 2.6668, "step": 266750 }, { "epoch": 2.3560995425563855, "grad_norm": 3.6572837829589844, "learning_rate": 5.46880669079006e-05, "loss": 3.1471, "step": 266800 }, { "epoch": 2.3565410904466697, "grad_norm": 2.89141583442688, "learning_rate": 5.4674508470840816e-05, "loss": 3.191, "step": 266850 }, { "epoch": 2.356982638336954, "grad_norm": 11.647260665893555, "learning_rate": 5.4660672973535434e-05, "loss": 3.075, "step": 266900 }, { "epoch": 2.357424186227238, "grad_norm": 2.813277244567871, "learning_rate": 5.464683711623261e-05, "loss": 2.9424, "step": 266950 }, { "epoch": 2.357865734117522, "grad_norm": 2.9752204418182373, "learning_rate": 5.463300090000101e-05, "loss": 3.0976, "step": 267000 }, { "epoch": 2.357865734117522, "eval_asr_loss": 0.9327639954810589, "eval_loss": 2.834068536758423, "eval_runtime": 20.3783, "eval_samples_per_second": 37.687, "eval_steps_per_second": 9.422, "eval_tts_loss": 6.044620468470732, "step": 267000 }, { "epoch": 2.3583072820078064, "grad_norm": 5.590354919433594, "learning_rate": 5.461916432590938e-05, "loss": 3.0794, "step": 267050 }, { "epoch": 2.358748829898091, "grad_norm": 6.0784149169921875, "learning_rate": 5.460532739502647e-05, "loss": 3.1299, "step": 267100 }, { "epoch": 2.3591903777883747, "grad_norm": 3.448201894760132, "learning_rate": 5.4591490108421085e-05, "loss": 3.0882, "step": 267150 }, { "epoch": 2.3596319256786593, "grad_norm": 1.0501567125320435, "learning_rate": 5.457765246716201e-05, "loss": 3.3177, "step": 267200 }, { "epoch": 2.3600734735689435, "grad_norm": 2.9930355548858643, "learning_rate": 5.45638144723181e-05, "loss": 3.4531, "step": 267250 }, { "epoch": 2.3605150214592276, "grad_norm": 1.5537875890731812, "learning_rate": 5.454997612495822e-05, "loss": 3.6521, "step": 267300 }, { "epoch": 2.360956569349512, "grad_norm": 2.3104841709136963, "learning_rate": 5.453613742615129e-05, "loss": 2.8759, "step": 267350 }, { "epoch": 2.361398117239796, "grad_norm": 4.198497295379639, "learning_rate": 5.4522298376966185e-05, "loss": 2.7719, "step": 267400 }, { "epoch": 2.36183966513008, "grad_norm": 4.149351596832275, "learning_rate": 5.450845897847189e-05, "loss": 3.196, "step": 267450 }, { "epoch": 2.3622812130203643, "grad_norm": 1.9992108345031738, "learning_rate": 5.449461923173736e-05, "loss": 3.1804, "step": 267500 }, { "epoch": 2.3627227609106485, "grad_norm": 4.559908866882324, "learning_rate": 5.448077913783162e-05, "loss": 2.7879, "step": 267550 }, { "epoch": 2.3631643088009326, "grad_norm": 2.3502230644226074, "learning_rate": 5.446693869782368e-05, "loss": 2.9183, "step": 267600 }, { "epoch": 2.363605856691217, "grad_norm": 2.7888052463531494, "learning_rate": 5.445309791278259e-05, "loss": 3.2188, "step": 267650 }, { "epoch": 2.364047404581501, "grad_norm": 4.983249664306641, "learning_rate": 5.443925678377748e-05, "loss": 3.2986, "step": 267700 }, { "epoch": 2.364488952471785, "grad_norm": 2.1431801319122314, "learning_rate": 5.4425415311877406e-05, "loss": 2.7864, "step": 267750 }, { "epoch": 2.3649305003620693, "grad_norm": 3.75842547416687, "learning_rate": 5.441157349815153e-05, "loss": 3.25, "step": 267800 }, { "epoch": 2.3653720482523535, "grad_norm": 2.242339611053467, "learning_rate": 5.4397731343669e-05, "loss": 3.126, "step": 267850 }, { "epoch": 2.3658135961426376, "grad_norm": 4.4558634757995605, "learning_rate": 5.438388884949903e-05, "loss": 2.8855, "step": 267900 }, { "epoch": 2.366255144032922, "grad_norm": 2.1705050468444824, "learning_rate": 5.437004601671082e-05, "loss": 3.1353, "step": 267950 }, { "epoch": 2.366696691923206, "grad_norm": 4.800857067108154, "learning_rate": 5.435647971308141e-05, "loss": 3.5756, "step": 268000 }, { "epoch": 2.36713823981349, "grad_norm": 2.85304594039917, "learning_rate": 5.434263621298359e-05, "loss": 3.2629, "step": 268050 }, { "epoch": 2.3675797877037743, "grad_norm": 3.585688352584839, "learning_rate": 5.432879237745393e-05, "loss": 2.8421, "step": 268100 }, { "epoch": 2.3680213355940585, "grad_norm": 4.469842433929443, "learning_rate": 5.431494820756178e-05, "loss": 3.4016, "step": 268150 }, { "epoch": 2.3684628834843426, "grad_norm": 6.420340061187744, "learning_rate": 5.430110370437647e-05, "loss": 2.9548, "step": 268200 }, { "epoch": 2.368904431374627, "grad_norm": 2.834080696105957, "learning_rate": 5.428725886896737e-05, "loss": 3.266, "step": 268250 }, { "epoch": 2.369345979264911, "grad_norm": 3.0520131587982178, "learning_rate": 5.4273413702403864e-05, "loss": 3.2707, "step": 268300 }, { "epoch": 2.369787527155195, "grad_norm": 2.214024066925049, "learning_rate": 5.425956820575539e-05, "loss": 3.3662, "step": 268350 }, { "epoch": 2.3702290750454793, "grad_norm": 1.6311506032943726, "learning_rate": 5.4245722380091413e-05, "loss": 3.3962, "step": 268400 }, { "epoch": 2.3706706229357635, "grad_norm": 1.1005897521972656, "learning_rate": 5.423187622648137e-05, "loss": 3.0262, "step": 268450 }, { "epoch": 2.3711121708260476, "grad_norm": 4.452448844909668, "learning_rate": 5.421802974599478e-05, "loss": 3.4994, "step": 268500 }, { "epoch": 2.371553718716332, "grad_norm": 4.589182376861572, "learning_rate": 5.420418293970114e-05, "loss": 3.6717, "step": 268550 }, { "epoch": 2.371995266606616, "grad_norm": 3.4012272357940674, "learning_rate": 5.419033580867004e-05, "loss": 3.1408, "step": 268600 }, { "epoch": 2.3724368144969006, "grad_norm": 1.015120029449463, "learning_rate": 5.4176488353971036e-05, "loss": 2.8853, "step": 268650 }, { "epoch": 2.3728783623871843, "grad_norm": 4.353815078735352, "learning_rate": 5.416264057667371e-05, "loss": 3.0829, "step": 268700 }, { "epoch": 2.373319910277469, "grad_norm": 1.89161217212677, "learning_rate": 5.41487924778477e-05, "loss": 3.2831, "step": 268750 }, { "epoch": 2.3737614581677526, "grad_norm": 3.2402737140655518, "learning_rate": 5.413494405856264e-05, "loss": 3.3104, "step": 268800 }, { "epoch": 2.3742030060580372, "grad_norm": 0.6917250752449036, "learning_rate": 5.412109531988824e-05, "loss": 3.1604, "step": 268850 }, { "epoch": 2.3746445539483214, "grad_norm": 3.6345038414001465, "learning_rate": 5.410724626289415e-05, "loss": 2.9008, "step": 268900 }, { "epoch": 2.3750861018386056, "grad_norm": 1.763210654258728, "learning_rate": 5.409339688865012e-05, "loss": 2.7136, "step": 268950 }, { "epoch": 2.3755276497288897, "grad_norm": 3.357023000717163, "learning_rate": 5.40795471982259e-05, "loss": 3.0024, "step": 269000 }, { "epoch": 2.375969197619174, "grad_norm": 4.249170303344727, "learning_rate": 5.406569719269126e-05, "loss": 2.8979, "step": 269050 }, { "epoch": 2.376410745509458, "grad_norm": 2.7852632999420166, "learning_rate": 5.405184687311596e-05, "loss": 3.2276, "step": 269100 }, { "epoch": 2.3768522933997422, "grad_norm": 6.374467849731445, "learning_rate": 5.403799624056987e-05, "loss": 2.8556, "step": 269150 }, { "epoch": 2.3772938412900264, "grad_norm": 2.5176992416381836, "learning_rate": 5.40241452961228e-05, "loss": 2.9551, "step": 269200 }, { "epoch": 2.3777353891803106, "grad_norm": 3.443016529083252, "learning_rate": 5.401029404084463e-05, "loss": 2.8658, "step": 269250 }, { "epoch": 2.3781769370705947, "grad_norm": 1.54613196849823, "learning_rate": 5.3996442475805266e-05, "loss": 2.916, "step": 269300 }, { "epoch": 2.378618484960879, "grad_norm": 3.8205950260162354, "learning_rate": 5.3982590602074576e-05, "loss": 2.8492, "step": 269350 }, { "epoch": 2.379060032851163, "grad_norm": 4.579362869262695, "learning_rate": 5.3968738420722574e-05, "loss": 3.0837, "step": 269400 }, { "epoch": 2.3795015807414472, "grad_norm": 3.624328851699829, "learning_rate": 5.3954885932819156e-05, "loss": 3.3226, "step": 269450 }, { "epoch": 2.3799431286317314, "grad_norm": 2.1983401775360107, "learning_rate": 5.394103313943435e-05, "loss": 3.1253, "step": 269500 }, { "epoch": 2.3803846765220156, "grad_norm": 2.3121702671051025, "learning_rate": 5.392718004163815e-05, "loss": 2.7823, "step": 269550 }, { "epoch": 2.3808262244122997, "grad_norm": 4.537003040313721, "learning_rate": 5.3913326640500594e-05, "loss": 3.0575, "step": 269600 }, { "epoch": 2.381267772302584, "grad_norm": 2.0873870849609375, "learning_rate": 5.389947293709173e-05, "loss": 3.0701, "step": 269650 }, { "epoch": 2.381709320192868, "grad_norm": 3.360896348953247, "learning_rate": 5.388561893248165e-05, "loss": 3.2804, "step": 269700 }, { "epoch": 2.3821508680831522, "grad_norm": 1.97892427444458, "learning_rate": 5.3871764627740464e-05, "loss": 2.8883, "step": 269750 }, { "epoch": 2.3825924159734364, "grad_norm": 2.6766841411590576, "learning_rate": 5.385791002393831e-05, "loss": 2.8785, "step": 269800 }, { "epoch": 2.3830339638637206, "grad_norm": 3.2892675399780273, "learning_rate": 5.384405512214529e-05, "loss": 2.9238, "step": 269850 }, { "epoch": 2.3834755117540047, "grad_norm": 2.905297040939331, "learning_rate": 5.3830199923431615e-05, "loss": 2.6607, "step": 269900 }, { "epoch": 2.383917059644289, "grad_norm": 8.067341804504395, "learning_rate": 5.381634442886748e-05, "loss": 3.396, "step": 269950 }, { "epoch": 2.384358607534573, "grad_norm": 2.7794084548950195, "learning_rate": 5.3802488639523095e-05, "loss": 3.2242, "step": 270000 }, { "epoch": 2.384358607534573, "eval_asr_loss": 0.9255845836504114, "eval_loss": 2.8321094512939453, "eval_runtime": 20.5503, "eval_samples_per_second": 37.372, "eval_steps_per_second": 9.343, "eval_tts_loss": 6.0368225175606645, "step": 270000 }, { "epoch": 2.3848001554248572, "grad_norm": 4.25317907333374, "learning_rate": 5.378863255646872e-05, "loss": 3.1465, "step": 270050 }, { "epoch": 2.3852417033151414, "grad_norm": 2.3429462909698486, "learning_rate": 5.377477618077459e-05, "loss": 3.3547, "step": 270100 }, { "epoch": 2.3856832512054256, "grad_norm": 3.7491652965545654, "learning_rate": 5.376091951351101e-05, "loss": 3.0766, "step": 270150 }, { "epoch": 2.3861247990957097, "grad_norm": 5.4432373046875, "learning_rate": 5.374706255574831e-05, "loss": 3.0164, "step": 270200 }, { "epoch": 2.386566346985994, "grad_norm": 5.800184726715088, "learning_rate": 5.373320530855678e-05, "loss": 2.5743, "step": 270250 }, { "epoch": 2.3870078948762785, "grad_norm": 2.11259388923645, "learning_rate": 5.37193477730068e-05, "loss": 3.1403, "step": 270300 }, { "epoch": 2.3874494427665622, "grad_norm": 3.23903226852417, "learning_rate": 5.3705489950168756e-05, "loss": 2.9759, "step": 270350 }, { "epoch": 2.387890990656847, "grad_norm": 4.091006278991699, "learning_rate": 5.3691631841113025e-05, "loss": 3.1225, "step": 270400 }, { "epoch": 2.388332538547131, "grad_norm": 5.129329204559326, "learning_rate": 5.367777344691004e-05, "loss": 3.2588, "step": 270450 }, { "epoch": 2.388774086437415, "grad_norm": 3.9516549110412598, "learning_rate": 5.366391476863023e-05, "loss": 3.0227, "step": 270500 }, { "epoch": 2.3892156343276993, "grad_norm": 2.243333578109741, "learning_rate": 5.36500558073441e-05, "loss": 3.421, "step": 270550 }, { "epoch": 2.3896571822179835, "grad_norm": 0.9436094760894775, "learning_rate": 5.363619656412209e-05, "loss": 3.229, "step": 270600 }, { "epoch": 2.3900987301082677, "grad_norm": 4.076546669006348, "learning_rate": 5.362233704003473e-05, "loss": 3.1487, "step": 270650 }, { "epoch": 2.390540277998552, "grad_norm": 3.09926438331604, "learning_rate": 5.3608477236152566e-05, "loss": 3.7203, "step": 270700 }, { "epoch": 2.390981825888836, "grad_norm": 1.4818518161773682, "learning_rate": 5.3594617153546124e-05, "loss": 3.6391, "step": 270750 }, { "epoch": 2.39142337377912, "grad_norm": 2.5496177673339844, "learning_rate": 5.358075679328599e-05, "loss": 2.5611, "step": 270800 }, { "epoch": 2.3918649216694043, "grad_norm": 5.2290472984313965, "learning_rate": 5.356689615644275e-05, "loss": 2.6363, "step": 270850 }, { "epoch": 2.3923064695596885, "grad_norm": 2.520369529724121, "learning_rate": 5.3553035244087034e-05, "loss": 3.1807, "step": 270900 }, { "epoch": 2.3927480174499727, "grad_norm": 5.892983913421631, "learning_rate": 5.3539174057289486e-05, "loss": 2.8983, "step": 270950 }, { "epoch": 2.393189565340257, "grad_norm": 2.85194730758667, "learning_rate": 5.3525312597120735e-05, "loss": 2.7087, "step": 271000 }, { "epoch": 2.393631113230541, "grad_norm": 2.4447226524353027, "learning_rate": 5.351145086465148e-05, "loss": 3.127, "step": 271050 }, { "epoch": 2.394072661120825, "grad_norm": 4.408174514770508, "learning_rate": 5.349758886095243e-05, "loss": 3.0788, "step": 271100 }, { "epoch": 2.3945142090111093, "grad_norm": 2.8875482082366943, "learning_rate": 5.3483726587094306e-05, "loss": 2.9305, "step": 271150 }, { "epoch": 2.3949557569013935, "grad_norm": 2.284501552581787, "learning_rate": 5.3469864044147834e-05, "loss": 2.9174, "step": 271200 }, { "epoch": 2.3953973047916777, "grad_norm": 5.54253625869751, "learning_rate": 5.345600123318379e-05, "loss": 3.0562, "step": 271250 }, { "epoch": 2.395838852681962, "grad_norm": 6.721933841705322, "learning_rate": 5.344213815527295e-05, "loss": 2.5428, "step": 271300 }, { "epoch": 2.396280400572246, "grad_norm": 3.561114549636841, "learning_rate": 5.3428274811486146e-05, "loss": 2.9754, "step": 271350 }, { "epoch": 2.39672194846253, "grad_norm": 3.5978899002075195, "learning_rate": 5.3414411202894164e-05, "loss": 3.1414, "step": 271400 }, { "epoch": 2.3971634963528143, "grad_norm": 2.7129671573638916, "learning_rate": 5.3400547330567876e-05, "loss": 3.257, "step": 271450 }, { "epoch": 2.3976050442430985, "grad_norm": 2.2620344161987305, "learning_rate": 5.338668319557815e-05, "loss": 3.0757, "step": 271500 }, { "epoch": 2.3980465921333827, "grad_norm": 1.2042834758758545, "learning_rate": 5.337281879899586e-05, "loss": 2.8006, "step": 271550 }, { "epoch": 2.398488140023667, "grad_norm": 1.2057934999465942, "learning_rate": 5.335895414189191e-05, "loss": 3.3763, "step": 271600 }, { "epoch": 2.398929687913951, "grad_norm": 2.1979100704193115, "learning_rate": 5.334508922533724e-05, "loss": 3.0288, "step": 271650 }, { "epoch": 2.399371235804235, "grad_norm": 4.013682842254639, "learning_rate": 5.33312240504028e-05, "loss": 3.0163, "step": 271700 }, { "epoch": 2.3998127836945193, "grad_norm": 2.666842460632324, "learning_rate": 5.331735861815954e-05, "loss": 3.0332, "step": 271750 }, { "epoch": 2.4002543315848035, "grad_norm": 1.4315290451049805, "learning_rate": 5.330349292967847e-05, "loss": 3.1066, "step": 271800 }, { "epoch": 2.400695879475088, "grad_norm": 4.096558570861816, "learning_rate": 5.328962698603057e-05, "loss": 3.1646, "step": 271850 }, { "epoch": 2.401137427365372, "grad_norm": 4.647629737854004, "learning_rate": 5.32757607882869e-05, "loss": 2.9247, "step": 271900 }, { "epoch": 2.4015789752556564, "grad_norm": 3.107863187789917, "learning_rate": 5.326189433751847e-05, "loss": 3.8047, "step": 271950 }, { "epoch": 2.40202052314594, "grad_norm": 2.610917568206787, "learning_rate": 5.3248027634796363e-05, "loss": 3.1353, "step": 272000 }, { "epoch": 2.4024620710362248, "grad_norm": 2.056340456008911, "learning_rate": 5.323416068119167e-05, "loss": 3.6018, "step": 272050 }, { "epoch": 2.402903618926509, "grad_norm": 3.0951645374298096, "learning_rate": 5.322029347777548e-05, "loss": 2.6276, "step": 272100 }, { "epoch": 2.403345166816793, "grad_norm": 2.055095911026001, "learning_rate": 5.3206426025618935e-05, "loss": 3.1167, "step": 272150 }, { "epoch": 2.4037867147070773, "grad_norm": 1.9229023456573486, "learning_rate": 5.319283568220991e-05, "loss": 3.0742, "step": 272200 }, { "epoch": 2.4042282625973614, "grad_norm": 3.5400021076202393, "learning_rate": 5.317896774070756e-05, "loss": 2.7419, "step": 272250 }, { "epoch": 2.4046698104876456, "grad_norm": 2.7035903930664062, "learning_rate": 5.316537691979734e-05, "loss": 3.5171, "step": 272300 }, { "epoch": 2.4051113583779298, "grad_norm": 2.1096925735473633, "learning_rate": 5.315150849314863e-05, "loss": 2.9916, "step": 272350 }, { "epoch": 2.405552906268214, "grad_norm": 4.001811981201172, "learning_rate": 5.3137639823072605e-05, "loss": 2.8302, "step": 272400 }, { "epoch": 2.405994454158498, "grad_norm": 2.9190337657928467, "learning_rate": 5.312377091064049e-05, "loss": 3.4557, "step": 272450 }, { "epoch": 2.4064360020487823, "grad_norm": 2.619129180908203, "learning_rate": 5.3109901756923584e-05, "loss": 3.4018, "step": 272500 }, { "epoch": 2.4068775499390664, "grad_norm": 2.8036184310913086, "learning_rate": 5.309603236299313e-05, "loss": 3.2577, "step": 272550 }, { "epoch": 2.4073190978293506, "grad_norm": 3.093379259109497, "learning_rate": 5.308216272992044e-05, "loss": 3.111, "step": 272600 }, { "epoch": 2.4077606457196348, "grad_norm": 4.001999378204346, "learning_rate": 5.30682928587768e-05, "loss": 3.3451, "step": 272650 }, { "epoch": 2.408202193609919, "grad_norm": 0.8353650569915771, "learning_rate": 5.3054422750633556e-05, "loss": 2.9238, "step": 272700 }, { "epoch": 2.408643741500203, "grad_norm": 3.686678409576416, "learning_rate": 5.304055240656207e-05, "loss": 3.2232, "step": 272750 }, { "epoch": 2.4090852893904873, "grad_norm": 2.8034465312957764, "learning_rate": 5.30266818276337e-05, "loss": 2.8977, "step": 272800 }, { "epoch": 2.4095268372807714, "grad_norm": 1.4402449131011963, "learning_rate": 5.301281101491979e-05, "loss": 2.8699, "step": 272850 }, { "epoch": 2.4099683851710556, "grad_norm": 2.3180580139160156, "learning_rate": 5.29989399694918e-05, "loss": 2.7924, "step": 272900 }, { "epoch": 2.4104099330613398, "grad_norm": 2.8068578243255615, "learning_rate": 5.2985068692421124e-05, "loss": 2.8887, "step": 272950 }, { "epoch": 2.410851480951624, "grad_norm": 3.274829149246216, "learning_rate": 5.297119718477921e-05, "loss": 3.0473, "step": 273000 }, { "epoch": 2.410851480951624, "eval_asr_loss": 0.9267551264970076, "eval_loss": 2.827969789505005, "eval_runtime": 20.8002, "eval_samples_per_second": 36.923, "eval_steps_per_second": 9.231, "eval_tts_loss": 5.991136181012733, "step": 273000 }, { "epoch": 2.411293028841908, "grad_norm": 5.885298728942871, "learning_rate": 5.295732544763751e-05, "loss": 2.8145, "step": 273050 }, { "epoch": 2.4117345767321923, "grad_norm": 4.339020729064941, "learning_rate": 5.2943453482067474e-05, "loss": 3.3168, "step": 273100 }, { "epoch": 2.4121761246224764, "grad_norm": 2.551595449447632, "learning_rate": 5.2929581289140636e-05, "loss": 2.9253, "step": 273150 }, { "epoch": 2.4126176725127606, "grad_norm": 1.7731404304504395, "learning_rate": 5.291570886992848e-05, "loss": 3.4991, "step": 273200 }, { "epoch": 2.4130592204030448, "grad_norm": 4.125699996948242, "learning_rate": 5.290183622550253e-05, "loss": 3.2461, "step": 273250 }, { "epoch": 2.413500768293329, "grad_norm": 7.358626842498779, "learning_rate": 5.2887963356934356e-05, "loss": 3.1884, "step": 273300 }, { "epoch": 2.413942316183613, "grad_norm": 3.230909824371338, "learning_rate": 5.28740902652955e-05, "loss": 2.9181, "step": 273350 }, { "epoch": 2.4143838640738973, "grad_norm": 4.0158491134643555, "learning_rate": 5.286021695165754e-05, "loss": 3.2061, "step": 273400 }, { "epoch": 2.4148254119641814, "grad_norm": 3.230339527130127, "learning_rate": 5.284634341709206e-05, "loss": 2.9233, "step": 273450 }, { "epoch": 2.415266959854466, "grad_norm": 1.815710425376892, "learning_rate": 5.283246966267069e-05, "loss": 3.0136, "step": 273500 }, { "epoch": 2.4157085077447498, "grad_norm": 2.3201990127563477, "learning_rate": 5.2818595689465076e-05, "loss": 2.904, "step": 273550 }, { "epoch": 2.4161500556350344, "grad_norm": 1.2397010326385498, "learning_rate": 5.2804721498546825e-05, "loss": 2.959, "step": 273600 }, { "epoch": 2.4165916035253185, "grad_norm": 5.490777969360352, "learning_rate": 5.279084709098764e-05, "loss": 3.2756, "step": 273650 }, { "epoch": 2.4170331514156027, "grad_norm": 2.4662795066833496, "learning_rate": 5.277697246785917e-05, "loss": 3.1928, "step": 273700 }, { "epoch": 2.417474699305887, "grad_norm": 4.0165557861328125, "learning_rate": 5.276309763023314e-05, "loss": 3.0188, "step": 273750 }, { "epoch": 2.417916247196171, "grad_norm": 2.4704883098602295, "learning_rate": 5.2749222579181234e-05, "loss": 3.2902, "step": 273800 }, { "epoch": 2.418357795086455, "grad_norm": 3.132085084915161, "learning_rate": 5.2735347315775196e-05, "loss": 2.901, "step": 273850 }, { "epoch": 2.4187993429767394, "grad_norm": 4.100172519683838, "learning_rate": 5.272147184108679e-05, "loss": 3.0409, "step": 273900 }, { "epoch": 2.4192408908670235, "grad_norm": 1.9341561794281006, "learning_rate": 5.2707596156187745e-05, "loss": 2.5095, "step": 273950 }, { "epoch": 2.4196824387573077, "grad_norm": 3.8735644817352295, "learning_rate": 5.269372026214985e-05, "loss": 3.4838, "step": 274000 }, { "epoch": 2.420123986647592, "grad_norm": 1.550255537033081, "learning_rate": 5.267984416004492e-05, "loss": 3.0684, "step": 274050 }, { "epoch": 2.420565534537876, "grad_norm": 1.9884741306304932, "learning_rate": 5.266596785094475e-05, "loss": 3.3358, "step": 274100 }, { "epoch": 2.42100708242816, "grad_norm": 5.632818222045898, "learning_rate": 5.265209133592117e-05, "loss": 2.9479, "step": 274150 }, { "epoch": 2.4214486303184444, "grad_norm": 2.59893798828125, "learning_rate": 5.263821461604603e-05, "loss": 3.3219, "step": 274200 }, { "epoch": 2.4218901782087285, "grad_norm": 8.066505432128906, "learning_rate": 5.262433769239117e-05, "loss": 3.2494, "step": 274250 }, { "epoch": 2.4223317260990127, "grad_norm": 2.522413969039917, "learning_rate": 5.26104605660285e-05, "loss": 3.0755, "step": 274300 }, { "epoch": 2.422773273989297, "grad_norm": 1.7068610191345215, "learning_rate": 5.259658323802985e-05, "loss": 2.9435, "step": 274350 }, { "epoch": 2.423214821879581, "grad_norm": 6.8204827308654785, "learning_rate": 5.2582705709467195e-05, "loss": 2.9865, "step": 274400 }, { "epoch": 2.423656369769865, "grad_norm": 5.089502811431885, "learning_rate": 5.256882798141242e-05, "loss": 2.592, "step": 274450 }, { "epoch": 2.4240979176601494, "grad_norm": 3.091336250305176, "learning_rate": 5.2554950054937457e-05, "loss": 3.0115, "step": 274500 }, { "epoch": 2.4245394655504335, "grad_norm": 3.593991279602051, "learning_rate": 5.2541071931114274e-05, "loss": 2.8779, "step": 274550 }, { "epoch": 2.4249810134407177, "grad_norm": 3.2547972202301025, "learning_rate": 5.252719361101482e-05, "loss": 2.8707, "step": 274600 }, { "epoch": 2.425422561331002, "grad_norm": 2.4958364963531494, "learning_rate": 5.251331509571109e-05, "loss": 3.0852, "step": 274650 }, { "epoch": 2.425864109221286, "grad_norm": 5.056582927703857, "learning_rate": 5.2499436386275104e-05, "loss": 2.7544, "step": 274700 }, { "epoch": 2.42630565711157, "grad_norm": 2.989672899246216, "learning_rate": 5.248555748377882e-05, "loss": 3.3015, "step": 274750 }, { "epoch": 2.4267472050018544, "grad_norm": 4.162222385406494, "learning_rate": 5.2471678389294324e-05, "loss": 3.3367, "step": 274800 }, { "epoch": 2.4271887528921385, "grad_norm": 3.347440004348755, "learning_rate": 5.245779910389362e-05, "loss": 3.4447, "step": 274850 }, { "epoch": 2.4276303007824227, "grad_norm": 4.668935775756836, "learning_rate": 5.2443919628648785e-05, "loss": 3.2474, "step": 274900 }, { "epoch": 2.428071848672707, "grad_norm": 3.2450575828552246, "learning_rate": 5.243003996463188e-05, "loss": 2.9552, "step": 274950 }, { "epoch": 2.428513396562991, "grad_norm": 5.016406059265137, "learning_rate": 5.241616011291498e-05, "loss": 3.0076, "step": 275000 }, { "epoch": 2.4289549444532756, "grad_norm": 1.900590181350708, "learning_rate": 5.240228007457021e-05, "loss": 2.8514, "step": 275050 }, { "epoch": 2.4293964923435594, "grad_norm": 3.596437931060791, "learning_rate": 5.238839985066969e-05, "loss": 3.404, "step": 275100 }, { "epoch": 2.429838040233844, "grad_norm": 1.8727651834487915, "learning_rate": 5.237451944228551e-05, "loss": 3.3156, "step": 275150 }, { "epoch": 2.4302795881241277, "grad_norm": 1.6116772890090942, "learning_rate": 5.236063885048985e-05, "loss": 2.8646, "step": 275200 }, { "epoch": 2.4307211360144123, "grad_norm": 2.206568479537964, "learning_rate": 5.2346758076354853e-05, "loss": 2.9361, "step": 275250 }, { "epoch": 2.4311626839046965, "grad_norm": 1.3650556802749634, "learning_rate": 5.233287712095269e-05, "loss": 3.5946, "step": 275300 }, { "epoch": 2.4316042317949806, "grad_norm": 3.537829875946045, "learning_rate": 5.231899598535557e-05, "loss": 3.532, "step": 275350 }, { "epoch": 2.432045779685265, "grad_norm": 3.1498124599456787, "learning_rate": 5.2305114670635646e-05, "loss": 3.2834, "step": 275400 }, { "epoch": 2.432487327575549, "grad_norm": 2.376232862472534, "learning_rate": 5.2291233177865184e-05, "loss": 3.2038, "step": 275450 }, { "epoch": 2.432928875465833, "grad_norm": 2.03261399269104, "learning_rate": 5.227735150811639e-05, "loss": 3.0741, "step": 275500 }, { "epoch": 2.4333704233561173, "grad_norm": 1.746952772140503, "learning_rate": 5.226346966246149e-05, "loss": 2.8938, "step": 275550 }, { "epoch": 2.4338119712464015, "grad_norm": 2.2516887187957764, "learning_rate": 5.2249587641972755e-05, "loss": 3.4494, "step": 275600 }, { "epoch": 2.4342535191366856, "grad_norm": 9.749994277954102, "learning_rate": 5.223570544772245e-05, "loss": 3.0107, "step": 275650 }, { "epoch": 2.43469506702697, "grad_norm": 3.8277552127838135, "learning_rate": 5.222182308078286e-05, "loss": 3.0161, "step": 275700 }, { "epoch": 2.435136614917254, "grad_norm": 1.66301429271698, "learning_rate": 5.2207940542226286e-05, "loss": 2.9655, "step": 275750 }, { "epoch": 2.435578162807538, "grad_norm": 2.9888482093811035, "learning_rate": 5.2194057833124996e-05, "loss": 3.0962, "step": 275800 }, { "epoch": 2.4360197106978223, "grad_norm": 3.3052783012390137, "learning_rate": 5.2180174954551374e-05, "loss": 3.1027, "step": 275850 }, { "epoch": 2.4364612585881065, "grad_norm": 3.05707049369812, "learning_rate": 5.216629190757772e-05, "loss": 3.2934, "step": 275900 }, { "epoch": 2.4369028064783906, "grad_norm": 1.1935335397720337, "learning_rate": 5.2152408693276377e-05, "loss": 3.2416, "step": 275950 }, { "epoch": 2.437344354368675, "grad_norm": 6.007275104522705, "learning_rate": 5.213852531271972e-05, "loss": 3.5751, "step": 276000 }, { "epoch": 2.437344354368675, "eval_asr_loss": 0.9267052147489756, "eval_loss": 2.8153886795043945, "eval_runtime": 20.9782, "eval_samples_per_second": 36.609, "eval_steps_per_second": 9.152, "eval_tts_loss": 6.035981268690937, "step": 276000 }, { "epoch": 2.437785902258959, "grad_norm": 1.0648419857025146, "learning_rate": 5.212464176698012e-05, "loss": 3.029, "step": 276050 }, { "epoch": 2.438227450149243, "grad_norm": 2.689664363861084, "learning_rate": 5.211075805712996e-05, "loss": 3.2192, "step": 276100 }, { "epoch": 2.4386689980395273, "grad_norm": 2.4899792671203613, "learning_rate": 5.209687418424163e-05, "loss": 2.7205, "step": 276150 }, { "epoch": 2.4391105459298115, "grad_norm": 4.987395763397217, "learning_rate": 5.208299014938756e-05, "loss": 2.8382, "step": 276200 }, { "epoch": 2.4395520938200956, "grad_norm": 5.153337478637695, "learning_rate": 5.206910595364016e-05, "loss": 2.9976, "step": 276250 }, { "epoch": 2.43999364171038, "grad_norm": 2.3233425617218018, "learning_rate": 5.205522159807187e-05, "loss": 3.2552, "step": 276300 }, { "epoch": 2.440435189600664, "grad_norm": 2.0316269397735596, "learning_rate": 5.204133708375514e-05, "loss": 3.2666, "step": 276350 }, { "epoch": 2.440876737490948, "grad_norm": 2.4127016067504883, "learning_rate": 5.202745241176241e-05, "loss": 3.2615, "step": 276400 }, { "epoch": 2.4413182853812323, "grad_norm": 1.432511806488037, "learning_rate": 5.201356758316621e-05, "loss": 3.3985, "step": 276450 }, { "epoch": 2.4417598332715165, "grad_norm": 5.922308444976807, "learning_rate": 5.199968259903898e-05, "loss": 3.1147, "step": 276500 }, { "epoch": 2.4422013811618006, "grad_norm": 4.304170608520508, "learning_rate": 5.19857974604532e-05, "loss": 3.1307, "step": 276550 }, { "epoch": 2.442642929052085, "grad_norm": 1.1812562942504883, "learning_rate": 5.197191216848143e-05, "loss": 3.5177, "step": 276600 }, { "epoch": 2.443084476942369, "grad_norm": 3.5455162525177, "learning_rate": 5.195802672419617e-05, "loss": 2.7788, "step": 276650 }, { "epoch": 2.4435260248326536, "grad_norm": 2.6016390323638916, "learning_rate": 5.194414112866993e-05, "loss": 3.4756, "step": 276700 }, { "epoch": 2.4439675727229373, "grad_norm": 3.0860018730163574, "learning_rate": 5.193025538297528e-05, "loss": 2.7343, "step": 276750 }, { "epoch": 2.444409120613222, "grad_norm": 4.176785945892334, "learning_rate": 5.191636948818477e-05, "loss": 2.7136, "step": 276800 }, { "epoch": 2.444850668503506, "grad_norm": 2.4516937732696533, "learning_rate": 5.1902483445370974e-05, "loss": 3.349, "step": 276850 }, { "epoch": 2.4452922163937902, "grad_norm": 1.3623647689819336, "learning_rate": 5.1888597255606464e-05, "loss": 2.9043, "step": 276900 }, { "epoch": 2.4457337642840744, "grad_norm": 2.6704561710357666, "learning_rate": 5.187471091996382e-05, "loss": 2.7754, "step": 276950 }, { "epoch": 2.4461753121743586, "grad_norm": 5.5693769454956055, "learning_rate": 5.186082443951567e-05, "loss": 2.8445, "step": 277000 }, { "epoch": 2.4466168600646427, "grad_norm": 4.743313789367676, "learning_rate": 5.18469378153346e-05, "loss": 2.8558, "step": 277050 }, { "epoch": 2.447058407954927, "grad_norm": 3.867375135421753, "learning_rate": 5.183305104849324e-05, "loss": 3.3308, "step": 277100 }, { "epoch": 2.447499955845211, "grad_norm": 4.740933895111084, "learning_rate": 5.1819164140064245e-05, "loss": 2.9407, "step": 277150 }, { "epoch": 2.4479415037354952, "grad_norm": 3.624829053878784, "learning_rate": 5.180527709112024e-05, "loss": 3.222, "step": 277200 }, { "epoch": 2.4483830516257794, "grad_norm": 1.1802963018417358, "learning_rate": 5.179138990273389e-05, "loss": 2.9521, "step": 277250 }, { "epoch": 2.4488245995160636, "grad_norm": 1.7872570753097534, "learning_rate": 5.1777502575977846e-05, "loss": 3.1789, "step": 277300 }, { "epoch": 2.4492661474063477, "grad_norm": 8.364953994750977, "learning_rate": 5.1763615111924814e-05, "loss": 3.0813, "step": 277350 }, { "epoch": 2.449707695296632, "grad_norm": 1.8230026960372925, "learning_rate": 5.174972751164747e-05, "loss": 2.8741, "step": 277400 }, { "epoch": 2.450149243186916, "grad_norm": 4.879878997802734, "learning_rate": 5.173583977621852e-05, "loss": 2.9922, "step": 277450 }, { "epoch": 2.4505907910772002, "grad_norm": 2.740779161453247, "learning_rate": 5.172195190671065e-05, "loss": 2.9653, "step": 277500 }, { "epoch": 2.4510323389674844, "grad_norm": 2.1104583740234375, "learning_rate": 5.170806390419661e-05, "loss": 3.1337, "step": 277550 }, { "epoch": 2.4514738868577686, "grad_norm": 2.255263090133667, "learning_rate": 5.169417576974911e-05, "loss": 2.7475, "step": 277600 }, { "epoch": 2.4519154347480527, "grad_norm": 12.99841594696045, "learning_rate": 5.16802875044409e-05, "loss": 2.9789, "step": 277650 }, { "epoch": 2.452356982638337, "grad_norm": 1.152249813079834, "learning_rate": 5.166639910934472e-05, "loss": 3.0356, "step": 277700 }, { "epoch": 2.452798530528621, "grad_norm": 5.333968639373779, "learning_rate": 5.165251058553335e-05, "loss": 3.5585, "step": 277750 }, { "epoch": 2.4532400784189052, "grad_norm": 2.075641393661499, "learning_rate": 5.1638621934079557e-05, "loss": 3.2617, "step": 277800 }, { "epoch": 2.4536816263091894, "grad_norm": 2.203270435333252, "learning_rate": 5.162473315605611e-05, "loss": 3.4403, "step": 277850 }, { "epoch": 2.4541231741994736, "grad_norm": 2.8338234424591064, "learning_rate": 5.1610844252535796e-05, "loss": 2.9844, "step": 277900 }, { "epoch": 2.4545647220897577, "grad_norm": 2.2840635776519775, "learning_rate": 5.159695522459143e-05, "loss": 3.1731, "step": 277950 }, { "epoch": 2.455006269980042, "grad_norm": 3.000343084335327, "learning_rate": 5.158306607329582e-05, "loss": 2.8915, "step": 278000 }, { "epoch": 2.455447817870326, "grad_norm": 2.6898205280303955, "learning_rate": 5.1569176799721775e-05, "loss": 3.1113, "step": 278050 }, { "epoch": 2.4558893657606102, "grad_norm": 3.8035759925842285, "learning_rate": 5.155528740494211e-05, "loss": 2.7593, "step": 278100 }, { "epoch": 2.4563309136508944, "grad_norm": 4.912512302398682, "learning_rate": 5.1541397890029694e-05, "loss": 3.1687, "step": 278150 }, { "epoch": 2.4567724615411786, "grad_norm": 3.5801217555999756, "learning_rate": 5.152750825605737e-05, "loss": 2.9535, "step": 278200 }, { "epoch": 2.457214009431463, "grad_norm": 2.074791431427002, "learning_rate": 5.151361850409798e-05, "loss": 2.9635, "step": 278250 }, { "epoch": 2.457655557321747, "grad_norm": 3.802006959915161, "learning_rate": 5.14997286352244e-05, "loss": 3.0601, "step": 278300 }, { "epoch": 2.4580971052120315, "grad_norm": 6.514836311340332, "learning_rate": 5.148583865050949e-05, "loss": 3.3849, "step": 278350 }, { "epoch": 2.4585386531023152, "grad_norm": 5.888082504272461, "learning_rate": 5.147194855102616e-05, "loss": 2.8955, "step": 278400 }, { "epoch": 2.4589802009926, "grad_norm": 2.7883012294769287, "learning_rate": 5.145805833784727e-05, "loss": 3.0244, "step": 278450 }, { "epoch": 2.459421748882884, "grad_norm": 3.9875028133392334, "learning_rate": 5.144444581965854e-05, "loss": 3.0595, "step": 278500 }, { "epoch": 2.459863296773168, "grad_norm": 2.8172056674957275, "learning_rate": 5.143055538452777e-05, "loss": 3.6289, "step": 278550 }, { "epoch": 2.4603048446634523, "grad_norm": 4.393211841583252, "learning_rate": 5.14166648388987e-05, "loss": 3.0351, "step": 278600 }, { "epoch": 2.4607463925537365, "grad_norm": 2.0041215419769287, "learning_rate": 5.1402774183844304e-05, "loss": 3.5881, "step": 278650 }, { "epoch": 2.4611879404440207, "grad_norm": 4.510395526885986, "learning_rate": 5.138888342043751e-05, "loss": 2.8514, "step": 278700 }, { "epoch": 2.461629488334305, "grad_norm": 3.4052228927612305, "learning_rate": 5.137499254975123e-05, "loss": 3.4695, "step": 278750 }, { "epoch": 2.462071036224589, "grad_norm": 4.804360866546631, "learning_rate": 5.136110157285844e-05, "loss": 3.049, "step": 278800 }, { "epoch": 2.462512584114873, "grad_norm": 1.1645957231521606, "learning_rate": 5.1347210490832096e-05, "loss": 2.9793, "step": 278850 }, { "epoch": 2.4629541320051573, "grad_norm": 1.537708044052124, "learning_rate": 5.1333319304745165e-05, "loss": 3.2173, "step": 278900 }, { "epoch": 2.4633956798954415, "grad_norm": 1.4413223266601562, "learning_rate": 5.1319428015670637e-05, "loss": 3.2126, "step": 278950 }, { "epoch": 2.4638372277857257, "grad_norm": 1.397445559501648, "learning_rate": 5.130553662468145e-05, "loss": 3.1346, "step": 279000 }, { "epoch": 2.4638372277857257, "eval_asr_loss": 0.9248200039733635, "eval_loss": 2.8163654804229736, "eval_runtime": 20.7057, "eval_samples_per_second": 37.091, "eval_steps_per_second": 9.273, "eval_tts_loss": 5.998610943827098, "step": 279000 }, { "epoch": 2.46427877567601, "grad_norm": 2.494467258453369, "learning_rate": 5.129164513285065e-05, "loss": 3.2908, "step": 279050 }, { "epoch": 2.464720323566294, "grad_norm": 2.9098429679870605, "learning_rate": 5.127775354125124e-05, "loss": 3.1943, "step": 279100 }, { "epoch": 2.465161871456578, "grad_norm": 2.282630681991577, "learning_rate": 5.126386185095616e-05, "loss": 3.1679, "step": 279150 }, { "epoch": 2.4656034193468623, "grad_norm": 2.289762020111084, "learning_rate": 5.12499700630385e-05, "loss": 3.0701, "step": 279200 }, { "epoch": 2.4660449672371465, "grad_norm": 1.255960464477539, "learning_rate": 5.1236078178571224e-05, "loss": 2.9881, "step": 279250 }, { "epoch": 2.4664865151274307, "grad_norm": 1.53762948513031, "learning_rate": 5.122218619862742e-05, "loss": 2.9505, "step": 279300 }, { "epoch": 2.466928063017715, "grad_norm": 2.2143754959106445, "learning_rate": 5.120829412428009e-05, "loss": 2.9946, "step": 279350 }, { "epoch": 2.467369610907999, "grad_norm": 2.4658992290496826, "learning_rate": 5.1194401956602265e-05, "loss": 3.0142, "step": 279400 }, { "epoch": 2.467811158798283, "grad_norm": 2.9681243896484375, "learning_rate": 5.1180509696667035e-05, "loss": 2.6439, "step": 279450 }, { "epoch": 2.4682527066885673, "grad_norm": 3.541486978530884, "learning_rate": 5.116661734554744e-05, "loss": 2.9877, "step": 279500 }, { "epoch": 2.4686942545788515, "grad_norm": 4.179141521453857, "learning_rate": 5.1152724904316554e-05, "loss": 3.2289, "step": 279550 }, { "epoch": 2.4691358024691357, "grad_norm": 4.885958671569824, "learning_rate": 5.113883237404743e-05, "loss": 2.7886, "step": 279600 }, { "epoch": 2.46957735035942, "grad_norm": 5.514190196990967, "learning_rate": 5.112493975581316e-05, "loss": 3.2067, "step": 279650 }, { "epoch": 2.470018898249704, "grad_norm": 2.7596986293792725, "learning_rate": 5.1111047050686846e-05, "loss": 3.3029, "step": 279700 }, { "epoch": 2.470460446139988, "grad_norm": 2.898625612258911, "learning_rate": 5.1097154259741566e-05, "loss": 2.9842, "step": 279750 }, { "epoch": 2.4709019940302723, "grad_norm": 1.3060845136642456, "learning_rate": 5.108326138405041e-05, "loss": 2.9792, "step": 279800 }, { "epoch": 2.4713435419205565, "grad_norm": 2.678516149520874, "learning_rate": 5.106936842468651e-05, "loss": 3.2299, "step": 279850 }, { "epoch": 2.471785089810841, "grad_norm": 1.1386833190917969, "learning_rate": 5.105547538272297e-05, "loss": 2.6656, "step": 279900 }, { "epoch": 2.472226637701125, "grad_norm": 1.3154444694519043, "learning_rate": 5.10415822592329e-05, "loss": 3.4792, "step": 279950 }, { "epoch": 2.4726681855914094, "grad_norm": 3.350400447845459, "learning_rate": 5.102768905528944e-05, "loss": 3.4961, "step": 280000 }, { "epoch": 2.4731097334816936, "grad_norm": 1.9451656341552734, "learning_rate": 5.101379577196571e-05, "loss": 2.9711, "step": 280050 }, { "epoch": 2.4735512813719778, "grad_norm": 1.605257272720337, "learning_rate": 5.099990241033486e-05, "loss": 3.1134, "step": 280100 }, { "epoch": 2.473992829262262, "grad_norm": 2.776409149169922, "learning_rate": 5.098600897147e-05, "loss": 3.1724, "step": 280150 }, { "epoch": 2.474434377152546, "grad_norm": 2.7866625785827637, "learning_rate": 5.097211545644434e-05, "loss": 3.591, "step": 280200 }, { "epoch": 2.4748759250428303, "grad_norm": 2.8920161724090576, "learning_rate": 5.0958221866330994e-05, "loss": 3.3194, "step": 280250 }, { "epoch": 2.4753174729331144, "grad_norm": 1.613011121749878, "learning_rate": 5.094432820220313e-05, "loss": 3.5723, "step": 280300 }, { "epoch": 2.4757590208233986, "grad_norm": 4.738898754119873, "learning_rate": 5.093043446513392e-05, "loss": 2.7472, "step": 280350 }, { "epoch": 2.4762005687136828, "grad_norm": 3.9019086360931396, "learning_rate": 5.0916540656196535e-05, "loss": 2.8097, "step": 280400 }, { "epoch": 2.476642116603967, "grad_norm": 5.4280686378479, "learning_rate": 5.090264677646418e-05, "loss": 3.1141, "step": 280450 }, { "epoch": 2.477083664494251, "grad_norm": 3.153273582458496, "learning_rate": 5.088875282701e-05, "loss": 3.1938, "step": 280500 }, { "epoch": 2.4775252123845353, "grad_norm": 3.144218921661377, "learning_rate": 5.087485880890719e-05, "loss": 2.7476, "step": 280550 }, { "epoch": 2.4779667602748194, "grad_norm": 5.132445812225342, "learning_rate": 5.086096472322895e-05, "loss": 2.9857, "step": 280600 }, { "epoch": 2.4784083081651036, "grad_norm": 3.704521656036377, "learning_rate": 5.084707057104852e-05, "loss": 3.2549, "step": 280650 }, { "epoch": 2.4788498560553878, "grad_norm": 3.9432642459869385, "learning_rate": 5.083317635343904e-05, "loss": 3.4551, "step": 280700 }, { "epoch": 2.479291403945672, "grad_norm": 6.5958452224731445, "learning_rate": 5.081928207147377e-05, "loss": 3.0469, "step": 280750 }, { "epoch": 2.479732951835956, "grad_norm": 2.301940679550171, "learning_rate": 5.080538772622588e-05, "loss": 2.8283, "step": 280800 }, { "epoch": 2.4801744997262403, "grad_norm": 6.14644193649292, "learning_rate": 5.079177120752051e-05, "loss": 3.187, "step": 280850 }, { "epoch": 2.4806160476165244, "grad_norm": 2.53700590133667, "learning_rate": 5.077787674013933e-05, "loss": 3.186, "step": 280900 }, { "epoch": 2.4810575955068086, "grad_norm": 9.372054100036621, "learning_rate": 5.0763982212673766e-05, "loss": 2.8501, "step": 280950 }, { "epoch": 2.4814991433970928, "grad_norm": 4.544189929962158, "learning_rate": 5.075008762619705e-05, "loss": 3.4302, "step": 281000 }, { "epoch": 2.481940691287377, "grad_norm": 1.8253830671310425, "learning_rate": 5.073619298178243e-05, "loss": 3.033, "step": 281050 }, { "epoch": 2.482382239177661, "grad_norm": 2.2726876735687256, "learning_rate": 5.0722298280503134e-05, "loss": 3.1367, "step": 281100 }, { "epoch": 2.4828237870679453, "grad_norm": 1.6221014261245728, "learning_rate": 5.070840352343244e-05, "loss": 2.9628, "step": 281150 }, { "epoch": 2.4832653349582294, "grad_norm": 1.620402216911316, "learning_rate": 5.069450871164356e-05, "loss": 3.2234, "step": 281200 }, { "epoch": 2.4837068828485136, "grad_norm": 4.276893138885498, "learning_rate": 5.068061384620979e-05, "loss": 3.0061, "step": 281250 }, { "epoch": 2.4841484307387978, "grad_norm": 3.6740617752075195, "learning_rate": 5.066671892820436e-05, "loss": 2.8515, "step": 281300 }, { "epoch": 2.484589978629082, "grad_norm": 3.752772808074951, "learning_rate": 5.065282395870056e-05, "loss": 3.5286, "step": 281350 }, { "epoch": 2.485031526519366, "grad_norm": 1.7880843877792358, "learning_rate": 5.063892893877163e-05, "loss": 3.4442, "step": 281400 }, { "epoch": 2.4854730744096507, "grad_norm": 5.65786600112915, "learning_rate": 5.0625033869490854e-05, "loss": 3.3246, "step": 281450 }, { "epoch": 2.4859146222999344, "grad_norm": 3.071854829788208, "learning_rate": 5.061113875193152e-05, "loss": 3.2432, "step": 281500 }, { "epoch": 2.486356170190219, "grad_norm": 4.669334888458252, "learning_rate": 5.05972435871669e-05, "loss": 3.0424, "step": 281550 }, { "epoch": 2.486797718080503, "grad_norm": 2.256995439529419, "learning_rate": 5.0583348376270275e-05, "loss": 2.9281, "step": 281600 }, { "epoch": 2.4872392659707874, "grad_norm": 3.6562488079071045, "learning_rate": 5.056945312031495e-05, "loss": 2.9969, "step": 281650 }, { "epoch": 2.4876808138610715, "grad_norm": 2.6046030521392822, "learning_rate": 5.055555782037417e-05, "loss": 3.1838, "step": 281700 }, { "epoch": 2.4881223617513557, "grad_norm": 5.000454902648926, "learning_rate": 5.054166247752129e-05, "loss": 3.0062, "step": 281750 }, { "epoch": 2.48856390964164, "grad_norm": 3.942676544189453, "learning_rate": 5.052776709282957e-05, "loss": 2.955, "step": 281800 }, { "epoch": 2.489005457531924, "grad_norm": 3.0147054195404053, "learning_rate": 5.051387166737229e-05, "loss": 3.3155, "step": 281850 }, { "epoch": 2.489447005422208, "grad_norm": 1.4976001977920532, "learning_rate": 5.049997620222281e-05, "loss": 3.5099, "step": 281900 }, { "epoch": 2.4898885533124924, "grad_norm": 1.7248388528823853, "learning_rate": 5.04863586089013e-05, "loss": 3.163, "step": 281950 }, { "epoch": 2.4903301012027765, "grad_norm": 3.1885581016540527, "learning_rate": 5.0472463068327655e-05, "loss": 3.2042, "step": 282000 }, { "epoch": 2.4903301012027765, "eval_asr_loss": 0.9280694117340976, "eval_loss": 2.81754994392395, "eval_runtime": 20.6283, "eval_samples_per_second": 37.23, "eval_steps_per_second": 9.308, "eval_tts_loss": 6.024205605831826, "step": 282000 }, { "epoch": 2.4907716490930607, "grad_norm": 2.5311758518218994, "learning_rate": 5.045856749126028e-05, "loss": 3.0974, "step": 282050 }, { "epoch": 2.491213196983345, "grad_norm": 0.9694241285324097, "learning_rate": 5.044467187877242e-05, "loss": 3.2145, "step": 282100 }, { "epoch": 2.491654744873629, "grad_norm": 2.7776198387145996, "learning_rate": 5.043077623193745e-05, "loss": 3.1239, "step": 282150 }, { "epoch": 2.492096292763913, "grad_norm": 2.5301015377044678, "learning_rate": 5.041688055182866e-05, "loss": 3.1377, "step": 282200 }, { "epoch": 2.4925378406541974, "grad_norm": 2.710803747177124, "learning_rate": 5.0402984839519394e-05, "loss": 3.4368, "step": 282250 }, { "epoch": 2.4929793885444815, "grad_norm": 2.7915754318237305, "learning_rate": 5.038908909608295e-05, "loss": 3.0199, "step": 282300 }, { "epoch": 2.4934209364347657, "grad_norm": 5.301668643951416, "learning_rate": 5.037519332259269e-05, "loss": 2.8482, "step": 282350 }, { "epoch": 2.49386248432505, "grad_norm": 1.2914773225784302, "learning_rate": 5.036129752012191e-05, "loss": 2.8049, "step": 282400 }, { "epoch": 2.494304032215334, "grad_norm": 5.267665863037109, "learning_rate": 5.034740168974398e-05, "loss": 3.0485, "step": 282450 }, { "epoch": 2.494745580105618, "grad_norm": 4.781818389892578, "learning_rate": 5.033350583253219e-05, "loss": 2.4956, "step": 282500 }, { "epoch": 2.4951871279959024, "grad_norm": 4.119612693786621, "learning_rate": 5.031960994955992e-05, "loss": 2.7555, "step": 282550 }, { "epoch": 2.4956286758861865, "grad_norm": 3.3539037704467773, "learning_rate": 5.0305714041900484e-05, "loss": 3.0126, "step": 282600 }, { "epoch": 2.4960702237764707, "grad_norm": 2.262011766433716, "learning_rate": 5.029181811062724e-05, "loss": 3.13, "step": 282650 }, { "epoch": 2.496511771666755, "grad_norm": 3.2946643829345703, "learning_rate": 5.0277922156813516e-05, "loss": 3.07, "step": 282700 }, { "epoch": 2.496953319557039, "grad_norm": 2.5122263431549072, "learning_rate": 5.0264026181532655e-05, "loss": 3.0425, "step": 282750 }, { "epoch": 2.497394867447323, "grad_norm": 3.6154260635375977, "learning_rate": 5.025013018585801e-05, "loss": 2.9812, "step": 282800 }, { "epoch": 2.4978364153376074, "grad_norm": 2.087646722793579, "learning_rate": 5.023623417086295e-05, "loss": 3.3392, "step": 282850 }, { "epoch": 2.4982779632278915, "grad_norm": 1.8297511339187622, "learning_rate": 5.022233813762078e-05, "loss": 2.7674, "step": 282900 }, { "epoch": 2.4987195111181757, "grad_norm": 4.404008865356445, "learning_rate": 5.020844208720489e-05, "loss": 2.9481, "step": 282950 }, { "epoch": 2.4991610590084603, "grad_norm": 2.862468719482422, "learning_rate": 5.019454602068862e-05, "loss": 3.0517, "step": 283000 }, { "epoch": 2.499602606898744, "grad_norm": 2.504753351211548, "learning_rate": 5.0180649939145306e-05, "loss": 3.3843, "step": 283050 }, { "epoch": 2.5000441547890286, "grad_norm": 2.4501919746398926, "learning_rate": 5.016675384364832e-05, "loss": 3.0002, "step": 283100 }, { "epoch": 2.5004857026793124, "grad_norm": 8.237186431884766, "learning_rate": 5.0152857735270997e-05, "loss": 3.9627, "step": 283150 }, { "epoch": 2.500927250569597, "grad_norm": 1.8057780265808105, "learning_rate": 5.013896161508673e-05, "loss": 3.377, "step": 283200 }, { "epoch": 2.5013687984598807, "grad_norm": 2.8206937313079834, "learning_rate": 5.012506548416887e-05, "loss": 3.368, "step": 283250 }, { "epoch": 2.5018103463501653, "grad_norm": 3.7635865211486816, "learning_rate": 5.0111169343590734e-05, "loss": 3.0281, "step": 283300 }, { "epoch": 2.5022518942404495, "grad_norm": 5.689542293548584, "learning_rate": 5.009727319442572e-05, "loss": 3.2483, "step": 283350 }, { "epoch": 2.5026934421307336, "grad_norm": 2.809605598449707, "learning_rate": 5.008337703774718e-05, "loss": 3.1646, "step": 283400 }, { "epoch": 2.503134990021018, "grad_norm": 4.231949806213379, "learning_rate": 5.006948087462847e-05, "loss": 2.8324, "step": 283450 }, { "epoch": 2.503576537911302, "grad_norm": 3.0230464935302734, "learning_rate": 5.0055584706142954e-05, "loss": 3.0691, "step": 283500 }, { "epoch": 2.504018085801586, "grad_norm": 2.6768991947174072, "learning_rate": 5.0041688533363975e-05, "loss": 2.8525, "step": 283550 }, { "epoch": 2.5044596336918703, "grad_norm": 4.496885299682617, "learning_rate": 5.002779235736492e-05, "loss": 2.968, "step": 283600 }, { "epoch": 2.5049011815821545, "grad_norm": 2.771188497543335, "learning_rate": 5.001389617921914e-05, "loss": 3.6944, "step": 283650 }, { "epoch": 2.5053427294724386, "grad_norm": 6.899338245391846, "learning_rate": 5e-05, "loss": 3.1613, "step": 283700 }, { "epoch": 2.505784277362723, "grad_norm": 2.04693865776062, "learning_rate": 4.9986103820780856e-05, "loss": 3.1808, "step": 283750 }, { "epoch": 2.506225825253007, "grad_norm": 1.1627070903778076, "learning_rate": 4.997220764263509e-05, "loss": 2.7358, "step": 283800 }, { "epoch": 2.506667373143291, "grad_norm": 5.550576686859131, "learning_rate": 4.995831146663604e-05, "loss": 3.0607, "step": 283850 }, { "epoch": 2.5071089210335753, "grad_norm": 2.421271800994873, "learning_rate": 4.994441529385706e-05, "loss": 3.3567, "step": 283900 }, { "epoch": 2.5075504689238595, "grad_norm": 2.7869067192077637, "learning_rate": 4.993051912537155e-05, "loss": 2.8523, "step": 283950 }, { "epoch": 2.5079920168141436, "grad_norm": 6.2024359703063965, "learning_rate": 4.991662296225283e-05, "loss": 3.1936, "step": 284000 }, { "epoch": 2.508433564704428, "grad_norm": 1.586800217628479, "learning_rate": 4.990272680557429e-05, "loss": 3.0814, "step": 284050 }, { "epoch": 2.508875112594712, "grad_norm": 2.217116117477417, "learning_rate": 4.9888830656409265e-05, "loss": 3.1907, "step": 284100 }, { "epoch": 2.509316660484996, "grad_norm": 3.363901138305664, "learning_rate": 4.987493451583115e-05, "loss": 2.999, "step": 284150 }, { "epoch": 2.5097582083752803, "grad_norm": 1.8057552576065063, "learning_rate": 4.986103838491328e-05, "loss": 3.1576, "step": 284200 }, { "epoch": 2.5101997562655645, "grad_norm": 2.437493085861206, "learning_rate": 4.9847142264728995e-05, "loss": 3.2306, "step": 284250 }, { "epoch": 2.5106413041558486, "grad_norm": 1.4181112051010132, "learning_rate": 4.9833246156351706e-05, "loss": 3.0049, "step": 284300 }, { "epoch": 2.511082852046133, "grad_norm": 2.222081422805786, "learning_rate": 4.9819350060854705e-05, "loss": 2.7596, "step": 284350 }, { "epoch": 2.511524399936417, "grad_norm": 2.1626124382019043, "learning_rate": 4.98054539793114e-05, "loss": 3.331, "step": 284400 }, { "epoch": 2.511965947826701, "grad_norm": 1.8905410766601562, "learning_rate": 4.979155791279514e-05, "loss": 2.5447, "step": 284450 }, { "epoch": 2.5124074957169853, "grad_norm": 4.7585601806640625, "learning_rate": 4.977766186237923e-05, "loss": 3.1015, "step": 284500 }, { "epoch": 2.51284904360727, "grad_norm": 1.8657845258712769, "learning_rate": 4.9763765829137056e-05, "loss": 3.0149, "step": 284550 }, { "epoch": 2.5132905914975536, "grad_norm": 0.9560573697090149, "learning_rate": 4.9749869814141985e-05, "loss": 3.1053, "step": 284600 }, { "epoch": 2.5137321393878382, "grad_norm": 4.875983715057373, "learning_rate": 4.9735973818467356e-05, "loss": 3.0417, "step": 284650 }, { "epoch": 2.514173687278122, "grad_norm": 4.432520866394043, "learning_rate": 4.9722077843186496e-05, "loss": 2.7673, "step": 284700 }, { "epoch": 2.5146152351684066, "grad_norm": 1.4278959035873413, "learning_rate": 4.970818188937277e-05, "loss": 3.2836, "step": 284750 }, { "epoch": 2.5150567830586903, "grad_norm": 4.0246782302856445, "learning_rate": 4.9694285958099534e-05, "loss": 3.3007, "step": 284800 }, { "epoch": 2.515498330948975, "grad_norm": 4.376123905181885, "learning_rate": 4.968039005044009e-05, "loss": 2.8253, "step": 284850 }, { "epoch": 2.515939878839259, "grad_norm": 4.232708930969238, "learning_rate": 4.9666494167467805e-05, "loss": 2.8049, "step": 284900 }, { "epoch": 2.5163814267295432, "grad_norm": 4.800734996795654, "learning_rate": 4.965259831025604e-05, "loss": 3.1355, "step": 284950 }, { "epoch": 2.5168229746198274, "grad_norm": 4.9787116050720215, "learning_rate": 4.963870247987811e-05, "loss": 2.5987, "step": 285000 }, { "epoch": 2.5168229746198274, "eval_asr_loss": 0.9389612370940837, "eval_loss": 2.822880506515503, "eval_runtime": 20.954, "eval_samples_per_second": 36.652, "eval_steps_per_second": 9.163, "eval_tts_loss": 6.0094163610549405, "step": 285000 }, { "epoch": 2.5172645225101116, "grad_norm": 2.1059107780456543, "learning_rate": 4.962480667740732e-05, "loss": 2.9139, "step": 285050 }, { "epoch": 2.5177060704003957, "grad_norm": 4.750513553619385, "learning_rate": 4.961091090391707e-05, "loss": 3.6087, "step": 285100 }, { "epoch": 2.51814761829068, "grad_norm": 2.7874600887298584, "learning_rate": 4.959701516048062e-05, "loss": 3.1579, "step": 285150 }, { "epoch": 2.518589166180964, "grad_norm": 2.4497392177581787, "learning_rate": 4.958311944817135e-05, "loss": 3.1668, "step": 285200 }, { "epoch": 2.5190307140712482, "grad_norm": 4.497219562530518, "learning_rate": 4.956922376806255e-05, "loss": 3.0488, "step": 285250 }, { "epoch": 2.5194722619615324, "grad_norm": 4.298454284667969, "learning_rate": 4.9555328121227585e-05, "loss": 3.4448, "step": 285300 }, { "epoch": 2.5199138098518166, "grad_norm": 2.9603888988494873, "learning_rate": 4.954143250873973e-05, "loss": 2.74, "step": 285350 }, { "epoch": 2.5203553577421007, "grad_norm": 3.734753370285034, "learning_rate": 4.952753693167234e-05, "loss": 3.0217, "step": 285400 }, { "epoch": 2.520796905632385, "grad_norm": 4.9439616203308105, "learning_rate": 4.9513641391098726e-05, "loss": 3.1139, "step": 285450 }, { "epoch": 2.521238453522669, "grad_norm": 5.052853107452393, "learning_rate": 4.9499745888092164e-05, "loss": 3.6142, "step": 285500 }, { "epoch": 2.5216800014129532, "grad_norm": 4.196293830871582, "learning_rate": 4.9485850423726006e-05, "loss": 3.2003, "step": 285550 }, { "epoch": 2.5221215493032374, "grad_norm": 1.703595757484436, "learning_rate": 4.947195499907355e-05, "loss": 3.0572, "step": 285600 }, { "epoch": 2.5225630971935216, "grad_norm": 4.48645544052124, "learning_rate": 4.9458059615208074e-05, "loss": 3.643, "step": 285650 }, { "epoch": 2.5230046450838057, "grad_norm": 4.0789384841918945, "learning_rate": 4.944416427320289e-05, "loss": 3.1723, "step": 285700 }, { "epoch": 2.52344619297409, "grad_norm": 2.540339708328247, "learning_rate": 4.943026897413133e-05, "loss": 3.2023, "step": 285750 }, { "epoch": 2.523887740864374, "grad_norm": 2.1572635173797607, "learning_rate": 4.9416373719066655e-05, "loss": 2.6541, "step": 285800 }, { "epoch": 2.5243292887546582, "grad_norm": 3.9537816047668457, "learning_rate": 4.940247850908214e-05, "loss": 3.4782, "step": 285850 }, { "epoch": 2.5247708366449424, "grad_norm": 4.48990535736084, "learning_rate": 4.93885833452511e-05, "loss": 2.9896, "step": 285900 }, { "epoch": 2.5252123845352266, "grad_norm": 2.202242136001587, "learning_rate": 4.9374688228646834e-05, "loss": 3.5251, "step": 285950 }, { "epoch": 2.5256539324255107, "grad_norm": 1.071150541305542, "learning_rate": 4.936107106122839e-05, "loss": 2.3952, "step": 286000 }, { "epoch": 2.526095480315795, "grad_norm": 4.282175540924072, "learning_rate": 4.934717604129946e-05, "loss": 3.2568, "step": 286050 }, { "epoch": 2.5265370282060795, "grad_norm": 2.3084678649902344, "learning_rate": 4.933328107179565e-05, "loss": 3.6519, "step": 286100 }, { "epoch": 2.5269785760963632, "grad_norm": 5.534224987030029, "learning_rate": 4.931938615379024e-05, "loss": 3.3019, "step": 286150 }, { "epoch": 2.527420123986648, "grad_norm": 5.111447334289551, "learning_rate": 4.930549128835645e-05, "loss": 3.5112, "step": 286200 }, { "epoch": 2.5278616718769316, "grad_norm": 2.4259583950042725, "learning_rate": 4.929159647656758e-05, "loss": 2.9692, "step": 286250 }, { "epoch": 2.528303219767216, "grad_norm": 2.0717339515686035, "learning_rate": 4.9277701719496864e-05, "loss": 3.2531, "step": 286300 }, { "epoch": 2.5287447676575, "grad_norm": 5.0152692794799805, "learning_rate": 4.926380701821759e-05, "loss": 2.9771, "step": 286350 }, { "epoch": 2.5291863155477845, "grad_norm": 5.3666300773620605, "learning_rate": 4.924991237380296e-05, "loss": 3.0579, "step": 286400 }, { "epoch": 2.5296278634380682, "grad_norm": 2.360063314437866, "learning_rate": 4.9236295678481056e-05, "loss": 3.238, "step": 286450 }, { "epoch": 2.530069411328353, "grad_norm": 1.8746623992919922, "learning_rate": 4.922240114982474e-05, "loss": 2.9991, "step": 286500 }, { "epoch": 2.530510959218637, "grad_norm": 2.7873833179473877, "learning_rate": 4.920850668123136e-05, "loss": 2.9671, "step": 286550 }, { "epoch": 2.530952507108921, "grad_norm": 5.166163921356201, "learning_rate": 4.9194612273774124e-05, "loss": 2.9864, "step": 286600 }, { "epoch": 2.5313940549992053, "grad_norm": 3.091221570968628, "learning_rate": 4.9180717928526246e-05, "loss": 2.9901, "step": 286650 }, { "epoch": 2.5318356028894895, "grad_norm": 2.8771631717681885, "learning_rate": 4.916682364656096e-05, "loss": 3.4301, "step": 286700 }, { "epoch": 2.5322771507797737, "grad_norm": 2.0344605445861816, "learning_rate": 4.91529294289515e-05, "loss": 3.0751, "step": 286750 }, { "epoch": 2.532718698670058, "grad_norm": 5.213078498840332, "learning_rate": 4.913903527677105e-05, "loss": 2.9403, "step": 286800 }, { "epoch": 2.533160246560342, "grad_norm": 1.7472230195999146, "learning_rate": 4.9125141191092815e-05, "loss": 3.1577, "step": 286850 }, { "epoch": 2.533601794450626, "grad_norm": 1.2062336206436157, "learning_rate": 4.911124717299003e-05, "loss": 3.2217, "step": 286900 }, { "epoch": 2.5340433423409103, "grad_norm": 2.4233908653259277, "learning_rate": 4.909735322353584e-05, "loss": 3.0467, "step": 286950 }, { "epoch": 2.5344848902311945, "grad_norm": 4.851795196533203, "learning_rate": 4.908345934380346e-05, "loss": 3.3914, "step": 287000 }, { "epoch": 2.5349264381214787, "grad_norm": 2.5680599212646484, "learning_rate": 4.9069565534866096e-05, "loss": 3.1384, "step": 287050 }, { "epoch": 2.535367986011763, "grad_norm": 1.6114776134490967, "learning_rate": 4.905567179779688e-05, "loss": 3.0615, "step": 287100 }, { "epoch": 2.535809533902047, "grad_norm": 4.065406322479248, "learning_rate": 4.9041778133669024e-05, "loss": 2.9422, "step": 287150 }, { "epoch": 2.536251081792331, "grad_norm": 1.3240852355957031, "learning_rate": 4.902788454355569e-05, "loss": 2.7033, "step": 287200 }, { "epoch": 2.5366926296826153, "grad_norm": 2.480588674545288, "learning_rate": 4.901399102853001e-05, "loss": 3.3272, "step": 287250 }, { "epoch": 2.5371341775728995, "grad_norm": 5.737575054168701, "learning_rate": 4.9000097589665155e-05, "loss": 2.7542, "step": 287300 }, { "epoch": 2.5375757254631837, "grad_norm": 3.4474637508392334, "learning_rate": 4.8986204228034307e-05, "loss": 3.2363, "step": 287350 }, { "epoch": 2.538017273353468, "grad_norm": 3.878185749053955, "learning_rate": 4.8972310944710584e-05, "loss": 3.1163, "step": 287400 }, { "epoch": 2.538458821243752, "grad_norm": 4.924911022186279, "learning_rate": 4.895841774076711e-05, "loss": 2.4848, "step": 287450 }, { "epoch": 2.538900369134036, "grad_norm": 4.602917194366455, "learning_rate": 4.894452461727703e-05, "loss": 3.6051, "step": 287500 }, { "epoch": 2.5393419170243203, "grad_norm": 1.9596270322799683, "learning_rate": 4.89306315753135e-05, "loss": 3.1296, "step": 287550 }, { "epoch": 2.5397834649146045, "grad_norm": 6.045586109161377, "learning_rate": 4.8916738615949596e-05, "loss": 3.032, "step": 287600 }, { "epoch": 2.5402250128048887, "grad_norm": 4.978264331817627, "learning_rate": 4.890312359694534e-05, "loss": 3.2435, "step": 287650 }, { "epoch": 2.540666560695173, "grad_norm": 6.536340713500977, "learning_rate": 4.888923080429462e-05, "loss": 2.5792, "step": 287700 }, { "epoch": 2.5411081085854574, "grad_norm": 2.27994441986084, "learning_rate": 4.8875338097441396e-05, "loss": 2.656, "step": 287750 }, { "epoch": 2.541549656475741, "grad_norm": 5.826876163482666, "learning_rate": 4.886144547745879e-05, "loss": 2.3847, "step": 287800 }, { "epoch": 2.5419912043660258, "grad_norm": 2.962148427963257, "learning_rate": 4.884755294541984e-05, "loss": 2.8411, "step": 287850 }, { "epoch": 2.5424327522563095, "grad_norm": 5.979537487030029, "learning_rate": 4.883366050239767e-05, "loss": 2.9138, "step": 287900 }, { "epoch": 2.542874300146594, "grad_norm": 1.784103512763977, "learning_rate": 4.881976814946534e-05, "loss": 2.9958, "step": 287950 }, { "epoch": 2.543315848036878, "grad_norm": 3.050992250442505, "learning_rate": 4.8805875887695893e-05, "loss": 2.9764, "step": 288000 }, { "epoch": 2.543315848036878, "eval_asr_loss": 0.9242752688294293, "eval_loss": 2.8110244274139404, "eval_runtime": 21.043, "eval_samples_per_second": 36.497, "eval_steps_per_second": 9.124, "eval_tts_loss": 6.050109421562367, "step": 288000 }, { "epoch": 2.5437573959271624, "grad_norm": 2.169618844985962, "learning_rate": 4.8791983718162424e-05, "loss": 3.0412, "step": 288050 }, { "epoch": 2.5441989438174466, "grad_norm": 4.288203239440918, "learning_rate": 4.8778091641937975e-05, "loss": 2.7545, "step": 288100 }, { "epoch": 2.5446404917077308, "grad_norm": 5.13236141204834, "learning_rate": 4.876419966009556e-05, "loss": 2.9921, "step": 288150 }, { "epoch": 2.545082039598015, "grad_norm": 1.8608711957931519, "learning_rate": 4.8750307773708245e-05, "loss": 3.3456, "step": 288200 }, { "epoch": 2.545523587488299, "grad_norm": 3.285364866256714, "learning_rate": 4.873641598384907e-05, "loss": 2.4351, "step": 288250 }, { "epoch": 2.5459651353785833, "grad_norm": 2.722503423690796, "learning_rate": 4.872252429159106e-05, "loss": 2.9236, "step": 288300 }, { "epoch": 2.5464066832688674, "grad_norm": 3.0441715717315674, "learning_rate": 4.870863269800718e-05, "loss": 3.5569, "step": 288350 }, { "epoch": 2.5468482311591516, "grad_norm": 2.4148430824279785, "learning_rate": 4.869474120417049e-05, "loss": 3.1819, "step": 288400 }, { "epoch": 2.5472897790494358, "grad_norm": 3.9821674823760986, "learning_rate": 4.8680849811153984e-05, "loss": 3.0567, "step": 288450 }, { "epoch": 2.54773132693972, "grad_norm": 1.9288352727890015, "learning_rate": 4.866695852003064e-05, "loss": 2.6452, "step": 288500 }, { "epoch": 2.548172874830004, "grad_norm": 3.0059030055999756, "learning_rate": 4.8653067331873424e-05, "loss": 3.2012, "step": 288550 }, { "epoch": 2.5486144227202883, "grad_norm": 3.6946487426757812, "learning_rate": 4.863917624775537e-05, "loss": 3.0973, "step": 288600 }, { "epoch": 2.5490559706105724, "grad_norm": 2.914017915725708, "learning_rate": 4.8625285268749394e-05, "loss": 3.0617, "step": 288650 }, { "epoch": 2.5494975185008566, "grad_norm": 2.5043909549713135, "learning_rate": 4.8611394395928465e-05, "loss": 3.0371, "step": 288700 }, { "epoch": 2.5499390663911408, "grad_norm": 3.3208022117614746, "learning_rate": 4.859750363036556e-05, "loss": 3.1523, "step": 288750 }, { "epoch": 2.550380614281425, "grad_norm": 4.472838401794434, "learning_rate": 4.8583612973133595e-05, "loss": 3.3449, "step": 288800 }, { "epoch": 2.550822162171709, "grad_norm": 2.654770612716675, "learning_rate": 4.8569722425305526e-05, "loss": 2.8547, "step": 288850 }, { "epoch": 2.5512637100619933, "grad_norm": 3.2713565826416016, "learning_rate": 4.855583198795425e-05, "loss": 3.2977, "step": 288900 }, { "epoch": 2.5517052579522774, "grad_norm": 2.4618031978607178, "learning_rate": 4.854194166215274e-05, "loss": 3.2582, "step": 288950 }, { "epoch": 2.5521468058425616, "grad_norm": 2.4435980319976807, "learning_rate": 4.852805144897385e-05, "loss": 2.9093, "step": 289000 }, { "epoch": 2.5525883537328458, "grad_norm": 3.2840375900268555, "learning_rate": 4.8514161349490514e-05, "loss": 3.1506, "step": 289050 }, { "epoch": 2.55302990162313, "grad_norm": 3.23045015335083, "learning_rate": 4.850027136477562e-05, "loss": 3.3427, "step": 289100 }, { "epoch": 2.553471449513414, "grad_norm": 3.247812032699585, "learning_rate": 4.848638149590203e-05, "loss": 3.3576, "step": 289150 }, { "epoch": 2.5539129974036983, "grad_norm": 4.35873556137085, "learning_rate": 4.8472491743942637e-05, "loss": 2.9928, "step": 289200 }, { "epoch": 2.5543545452939824, "grad_norm": 1.9763538837432861, "learning_rate": 4.845860210997031e-05, "loss": 2.918, "step": 289250 }, { "epoch": 2.554796093184267, "grad_norm": 1.984796404838562, "learning_rate": 4.84447125950579e-05, "loss": 2.9415, "step": 289300 }, { "epoch": 2.5552376410745508, "grad_norm": 2.300981044769287, "learning_rate": 4.8430823200278244e-05, "loss": 3.4028, "step": 289350 }, { "epoch": 2.5556791889648354, "grad_norm": 2.72800350189209, "learning_rate": 4.841693392670419e-05, "loss": 2.9831, "step": 289400 }, { "epoch": 2.556120736855119, "grad_norm": 4.119724750518799, "learning_rate": 4.840304477540859e-05, "loss": 3.0185, "step": 289450 }, { "epoch": 2.5565622847454037, "grad_norm": 3.7494263648986816, "learning_rate": 4.838915574746421e-05, "loss": 3.0239, "step": 289500 }, { "epoch": 2.5570038326356874, "grad_norm": 3.4036457538604736, "learning_rate": 4.837526684394389e-05, "loss": 3.4343, "step": 289550 }, { "epoch": 2.557445380525972, "grad_norm": 2.887335777282715, "learning_rate": 4.8361378065920455e-05, "loss": 3.0405, "step": 289600 }, { "epoch": 2.5578869284162558, "grad_norm": 5.955014705657959, "learning_rate": 4.834748941446666e-05, "loss": 2.8027, "step": 289650 }, { "epoch": 2.5583284763065404, "grad_norm": 3.5921425819396973, "learning_rate": 4.8333600890655274e-05, "loss": 3.1668, "step": 289700 }, { "epoch": 2.5587700241968245, "grad_norm": 4.463149547576904, "learning_rate": 4.8319712495559116e-05, "loss": 3.166, "step": 289750 }, { "epoch": 2.5592115720871087, "grad_norm": 3.158867359161377, "learning_rate": 4.83058242302509e-05, "loss": 3.3737, "step": 289800 }, { "epoch": 2.559653119977393, "grad_norm": 3.190603256225586, "learning_rate": 4.82919360958034e-05, "loss": 2.8906, "step": 289850 }, { "epoch": 2.560094667867677, "grad_norm": 1.280529499053955, "learning_rate": 4.827804809328937e-05, "loss": 3.0522, "step": 289900 }, { "epoch": 2.560536215757961, "grad_norm": 1.1511170864105225, "learning_rate": 4.82641602237815e-05, "loss": 3.0855, "step": 289950 }, { "epoch": 2.5609777636482454, "grad_norm": 5.506573677062988, "learning_rate": 4.825027248835254e-05, "loss": 3.3837, "step": 290000 }, { "epoch": 2.5614193115385295, "grad_norm": 3.6168723106384277, "learning_rate": 4.8236384888075184e-05, "loss": 3.0015, "step": 290050 }, { "epoch": 2.5618608594288137, "grad_norm": 2.3460497856140137, "learning_rate": 4.822249742402216e-05, "loss": 2.9652, "step": 290100 }, { "epoch": 2.562302407319098, "grad_norm": 10.541850090026855, "learning_rate": 4.820861009726612e-05, "loss": 2.9164, "step": 290150 }, { "epoch": 2.562743955209382, "grad_norm": 4.257185459136963, "learning_rate": 4.819472290887977e-05, "loss": 3.1391, "step": 290200 }, { "epoch": 2.563185503099666, "grad_norm": 6.34254264831543, "learning_rate": 4.818083585993578e-05, "loss": 2.9771, "step": 290250 }, { "epoch": 2.5636270509899504, "grad_norm": 2.1685292720794678, "learning_rate": 4.816694895150677e-05, "loss": 2.8735, "step": 290300 }, { "epoch": 2.5640685988802345, "grad_norm": 5.591983795166016, "learning_rate": 4.81530621846654e-05, "loss": 3.0327, "step": 290350 }, { "epoch": 2.5645101467705187, "grad_norm": 7.579561233520508, "learning_rate": 4.813917556048435e-05, "loss": 3.4491, "step": 290400 }, { "epoch": 2.564951694660803, "grad_norm": 1.9481171369552612, "learning_rate": 4.812528908003619e-05, "loss": 2.8593, "step": 290450 }, { "epoch": 2.565393242551087, "grad_norm": 2.7620160579681396, "learning_rate": 4.811140274439355e-05, "loss": 3.1415, "step": 290500 }, { "epoch": 2.565834790441371, "grad_norm": 3.436802387237549, "learning_rate": 4.8097516554629044e-05, "loss": 3.0391, "step": 290550 }, { "epoch": 2.5662763383316554, "grad_norm": 5.187906742095947, "learning_rate": 4.808363051181525e-05, "loss": 3.0544, "step": 290600 }, { "epoch": 2.5667178862219395, "grad_norm": 4.037874698638916, "learning_rate": 4.806974461702473e-05, "loss": 3.2693, "step": 290650 }, { "epoch": 2.5671594341122237, "grad_norm": 2.7287588119506836, "learning_rate": 4.805585887133007e-05, "loss": 3.2949, "step": 290700 }, { "epoch": 2.567600982002508, "grad_norm": 3.6652348041534424, "learning_rate": 4.804197327580385e-05, "loss": 2.6618, "step": 290750 }, { "epoch": 2.568042529892792, "grad_norm": 4.587352752685547, "learning_rate": 4.802808783151858e-05, "loss": 2.9005, "step": 290800 }, { "epoch": 2.568484077783076, "grad_norm": 3.400111198425293, "learning_rate": 4.801420253954679e-05, "loss": 3.4068, "step": 290850 }, { "epoch": 2.5689256256733604, "grad_norm": 5.114696025848389, "learning_rate": 4.8000317400961045e-05, "loss": 2.8093, "step": 290900 }, { "epoch": 2.569367173563645, "grad_norm": 2.7151639461517334, "learning_rate": 4.7986432416833796e-05, "loss": 3.2047, "step": 290950 }, { "epoch": 2.5698087214539287, "grad_norm": 5.248869895935059, "learning_rate": 4.797254758823758e-05, "loss": 3.3964, "step": 291000 }, { "epoch": 2.5698087214539287, "eval_asr_loss": 0.9296336651455246, "eval_loss": 2.8053767681121826, "eval_runtime": 20.9083, "eval_samples_per_second": 36.732, "eval_steps_per_second": 9.183, "eval_tts_loss": 6.007651098972165, "step": 291000 }, { "epoch": 2.5702502693442133, "grad_norm": 2.2332794666290283, "learning_rate": 4.7958662916244885e-05, "loss": 2.7228, "step": 291050 }, { "epoch": 2.570691817234497, "grad_norm": 1.7066570520401, "learning_rate": 4.794477840192814e-05, "loss": 2.8089, "step": 291100 }, { "epoch": 2.5711333651247816, "grad_norm": 1.0596325397491455, "learning_rate": 4.7930894046359844e-05, "loss": 2.959, "step": 291150 }, { "epoch": 2.5715749130150654, "grad_norm": 9.38840389251709, "learning_rate": 4.791700985061244e-05, "loss": 2.7549, "step": 291200 }, { "epoch": 2.57201646090535, "grad_norm": 1.5426836013793945, "learning_rate": 4.7903125815758384e-05, "loss": 2.8001, "step": 291250 }, { "epoch": 2.572458008795634, "grad_norm": 2.8356592655181885, "learning_rate": 4.788924194287005e-05, "loss": 2.7326, "step": 291300 }, { "epoch": 2.5728995566859183, "grad_norm": 1.5413955450057983, "learning_rate": 4.7875358233019886e-05, "loss": 2.5727, "step": 291350 }, { "epoch": 2.5733411045762025, "grad_norm": 2.142096996307373, "learning_rate": 4.7861474687280295e-05, "loss": 3.4568, "step": 291400 }, { "epoch": 2.5737826524664866, "grad_norm": 2.751913070678711, "learning_rate": 4.784759130672363e-05, "loss": 3.0957, "step": 291450 }, { "epoch": 2.574224200356771, "grad_norm": 1.8434984683990479, "learning_rate": 4.783370809242228e-05, "loss": 3.0992, "step": 291500 }, { "epoch": 2.574665748247055, "grad_norm": 4.122632026672363, "learning_rate": 4.781982504544864e-05, "loss": 2.9016, "step": 291550 }, { "epoch": 2.575107296137339, "grad_norm": 9.451845169067383, "learning_rate": 4.780594216687501e-05, "loss": 3.0843, "step": 291600 }, { "epoch": 2.5755488440276233, "grad_norm": 1.7359905242919922, "learning_rate": 4.779205945777373e-05, "loss": 3.2995, "step": 291650 }, { "epoch": 2.5759903919179075, "grad_norm": 5.614558219909668, "learning_rate": 4.777817691921716e-05, "loss": 2.781, "step": 291700 }, { "epoch": 2.5764319398081916, "grad_norm": 4.020981311798096, "learning_rate": 4.776429455227756e-05, "loss": 2.9995, "step": 291750 }, { "epoch": 2.576873487698476, "grad_norm": 1.5812124013900757, "learning_rate": 4.7750412358027256e-05, "loss": 2.9613, "step": 291800 }, { "epoch": 2.57731503558876, "grad_norm": 3.8988916873931885, "learning_rate": 4.773653033753852e-05, "loss": 3.2695, "step": 291850 }, { "epoch": 2.577756583479044, "grad_norm": 7.3769755363464355, "learning_rate": 4.7722648491883634e-05, "loss": 3.1298, "step": 291900 }, { "epoch": 2.5781981313693283, "grad_norm": 4.335209369659424, "learning_rate": 4.7708766822134834e-05, "loss": 2.8645, "step": 291950 }, { "epoch": 2.5786396792596125, "grad_norm": 4.128007888793945, "learning_rate": 4.769488532936435e-05, "loss": 3.2199, "step": 292000 }, { "epoch": 2.5790812271498966, "grad_norm": 3.834303140640259, "learning_rate": 4.768100401464445e-05, "loss": 3.2685, "step": 292050 }, { "epoch": 2.579522775040181, "grad_norm": 3.8419222831726074, "learning_rate": 4.766712287904731e-05, "loss": 2.9547, "step": 292100 }, { "epoch": 2.579964322930465, "grad_norm": 1.9303950071334839, "learning_rate": 4.765324192364516e-05, "loss": 3.0358, "step": 292150 }, { "epoch": 2.580405870820749, "grad_norm": 2.598571538925171, "learning_rate": 4.7639361149510175e-05, "loss": 3.0761, "step": 292200 }, { "epoch": 2.5808474187110333, "grad_norm": 2.2937920093536377, "learning_rate": 4.7625480557714504e-05, "loss": 2.9476, "step": 292250 }, { "epoch": 2.5812889666013175, "grad_norm": 3.7717130184173584, "learning_rate": 4.761160014933032e-05, "loss": 3.1194, "step": 292300 }, { "epoch": 2.5817305144916016, "grad_norm": 4.047567844390869, "learning_rate": 4.759771992542979e-05, "loss": 3.2471, "step": 292350 }, { "epoch": 2.582172062381886, "grad_norm": 4.486166000366211, "learning_rate": 4.758383988708503e-05, "loss": 3.2214, "step": 292400 }, { "epoch": 2.58261361027217, "grad_norm": 1.3521056175231934, "learning_rate": 4.7569960035368136e-05, "loss": 2.8203, "step": 292450 }, { "epoch": 2.5830551581624546, "grad_norm": 5.077817916870117, "learning_rate": 4.755608037135123e-05, "loss": 3.1033, "step": 292500 }, { "epoch": 2.5834967060527383, "grad_norm": 1.6110424995422363, "learning_rate": 4.7542200896106384e-05, "loss": 2.8632, "step": 292550 }, { "epoch": 2.583938253943023, "grad_norm": 2.5862834453582764, "learning_rate": 4.752832161070568e-05, "loss": 3.0453, "step": 292600 }, { "epoch": 2.5843798018333066, "grad_norm": 1.3178834915161133, "learning_rate": 4.751444251622117e-05, "loss": 2.79, "step": 292650 }, { "epoch": 2.5848213497235912, "grad_norm": 1.3635307550430298, "learning_rate": 4.750056361372492e-05, "loss": 2.8618, "step": 292700 }, { "epoch": 2.585262897613875, "grad_norm": 3.295215606689453, "learning_rate": 4.748668490428892e-05, "loss": 3.0196, "step": 292750 }, { "epoch": 2.5857044455041596, "grad_norm": 2.1861371994018555, "learning_rate": 4.7472806388985184e-05, "loss": 2.9002, "step": 292800 }, { "epoch": 2.5861459933944433, "grad_norm": 3.775594711303711, "learning_rate": 4.745892806888576e-05, "loss": 3.0787, "step": 292850 }, { "epoch": 2.586587541284728, "grad_norm": 3.5699329376220703, "learning_rate": 4.744504994506256e-05, "loss": 2.9571, "step": 292900 }, { "epoch": 2.587029089175012, "grad_norm": 3.14806866645813, "learning_rate": 4.74311720185876e-05, "loss": 3.1601, "step": 292950 }, { "epoch": 2.5874706370652962, "grad_norm": 0.9876266717910767, "learning_rate": 4.741729429053281e-05, "loss": 3.0339, "step": 293000 }, { "epoch": 2.5879121849555804, "grad_norm": 5.595848560333252, "learning_rate": 4.7403416761970156e-05, "loss": 2.8263, "step": 293050 }, { "epoch": 2.5883537328458646, "grad_norm": 3.9064714908599854, "learning_rate": 4.738953943397152e-05, "loss": 2.9474, "step": 293100 }, { "epoch": 2.5887952807361487, "grad_norm": 7.659373760223389, "learning_rate": 4.7375662307608835e-05, "loss": 3.179, "step": 293150 }, { "epoch": 2.589236828626433, "grad_norm": 2.4057817459106445, "learning_rate": 4.736178538395399e-05, "loss": 2.9813, "step": 293200 }, { "epoch": 2.589678376516717, "grad_norm": 5.122211933135986, "learning_rate": 4.7347908664078835e-05, "loss": 2.6911, "step": 293250 }, { "epoch": 2.5901199244070012, "grad_norm": 4.478826999664307, "learning_rate": 4.733403214905526e-05, "loss": 3.2017, "step": 293300 }, { "epoch": 2.5905614722972854, "grad_norm": 1.5622105598449707, "learning_rate": 4.73201558399551e-05, "loss": 3.2823, "step": 293350 }, { "epoch": 2.5910030201875696, "grad_norm": 1.5616695880889893, "learning_rate": 4.730627973785016e-05, "loss": 2.7755, "step": 293400 }, { "epoch": 2.5914445680778537, "grad_norm": 1.437469244003296, "learning_rate": 4.729240384381227e-05, "loss": 3.0075, "step": 293450 }, { "epoch": 2.591886115968138, "grad_norm": 3.9379823207855225, "learning_rate": 4.7278528158913235e-05, "loss": 3.0877, "step": 293500 }, { "epoch": 2.592327663858422, "grad_norm": 4.041423320770264, "learning_rate": 4.726465268422482e-05, "loss": 2.8019, "step": 293550 }, { "epoch": 2.5927692117487062, "grad_norm": 1.9618412256240845, "learning_rate": 4.725077742081878e-05, "loss": 3.4074, "step": 293600 }, { "epoch": 2.5932107596389904, "grad_norm": 4.111571788787842, "learning_rate": 4.723690236976686e-05, "loss": 3.1407, "step": 293650 }, { "epoch": 2.5936523075292746, "grad_norm": 3.354868173599243, "learning_rate": 4.722302753214083e-05, "loss": 2.8812, "step": 293700 }, { "epoch": 2.5940938554195587, "grad_norm": 5.56953239440918, "learning_rate": 4.720915290901237e-05, "loss": 3.0335, "step": 293750 }, { "epoch": 2.594535403309843, "grad_norm": 4.485682010650635, "learning_rate": 4.7195278501453166e-05, "loss": 2.681, "step": 293800 }, { "epoch": 2.594976951200127, "grad_norm": 2.4361305236816406, "learning_rate": 4.7181404310534936e-05, "loss": 2.9928, "step": 293850 }, { "epoch": 2.5954184990904112, "grad_norm": 2.4141952991485596, "learning_rate": 4.716753033732932e-05, "loss": 3.0773, "step": 293900 }, { "epoch": 2.5958600469806954, "grad_norm": 7.748928070068359, "learning_rate": 4.7153656582907944e-05, "loss": 3.2273, "step": 293950 }, { "epoch": 2.5963015948709796, "grad_norm": 7.025214195251465, "learning_rate": 4.713978304834249e-05, "loss": 3.233, "step": 294000 }, { "epoch": 2.5963015948709796, "eval_asr_loss": 0.9286917562741295, "eval_loss": 2.807992935180664, "eval_runtime": 21.0212, "eval_samples_per_second": 36.534, "eval_steps_per_second": 9.134, "eval_tts_loss": 6.040720395111247, "step": 294000 }, { "epoch": 2.5967431427612637, "grad_norm": 2.476391553878784, "learning_rate": 4.712590973470451e-05, "loss": 3.2182, "step": 294050 }, { "epoch": 2.597184690651548, "grad_norm": 1.7336440086364746, "learning_rate": 4.7112314102715916e-05, "loss": 2.6432, "step": 294100 }, { "epoch": 2.5976262385418325, "grad_norm": 2.4865992069244385, "learning_rate": 4.7098441229675804e-05, "loss": 3.1492, "step": 294150 }, { "epoch": 2.598067786432116, "grad_norm": 3.456488609313965, "learning_rate": 4.70848460315316e-05, "loss": 3.2032, "step": 294200 }, { "epoch": 2.598509334322401, "grad_norm": 3.8023250102996826, "learning_rate": 4.707097360329031e-05, "loss": 3.0862, "step": 294250 }, { "epoch": 2.5989508822126846, "grad_norm": 5.941384315490723, "learning_rate": 4.705710140129149e-05, "loss": 2.6157, "step": 294300 }, { "epoch": 2.599392430102969, "grad_norm": 3.5940475463867188, "learning_rate": 4.704322942660663e-05, "loss": 3.2328, "step": 294350 }, { "epoch": 2.599833977993253, "grad_norm": 2.429344415664673, "learning_rate": 4.702935768030721e-05, "loss": 3.3694, "step": 294400 }, { "epoch": 2.6002755258835375, "grad_norm": 1.2818009853363037, "learning_rate": 4.701548616346471e-05, "loss": 2.9216, "step": 294450 }, { "epoch": 2.6007170737738217, "grad_norm": 3.857438087463379, "learning_rate": 4.700161487715063e-05, "loss": 3.515, "step": 294500 }, { "epoch": 2.601158621664106, "grad_norm": 1.6534762382507324, "learning_rate": 4.698774382243636e-05, "loss": 2.8062, "step": 294550 }, { "epoch": 2.60160016955439, "grad_norm": 2.319244146347046, "learning_rate": 4.697387300039332e-05, "loss": 2.8212, "step": 294600 }, { "epoch": 2.602041717444674, "grad_norm": 5.253933429718018, "learning_rate": 4.696000241209296e-05, "loss": 3.3102, "step": 294650 }, { "epoch": 2.6024832653349583, "grad_norm": 4.916992664337158, "learning_rate": 4.6946132058606605e-05, "loss": 3.3078, "step": 294700 }, { "epoch": 2.6029248132252425, "grad_norm": 2.0591087341308594, "learning_rate": 4.6932261941005674e-05, "loss": 3.0297, "step": 294750 }, { "epoch": 2.6033663611155267, "grad_norm": 4.482892036437988, "learning_rate": 4.69183920603615e-05, "loss": 3.0613, "step": 294800 }, { "epoch": 2.603807909005811, "grad_norm": 1.322185754776001, "learning_rate": 4.6904522417745386e-05, "loss": 3.1383, "step": 294850 }, { "epoch": 2.604249456896095, "grad_norm": 4.500755786895752, "learning_rate": 4.6890653014228676e-05, "loss": 3.3837, "step": 294900 }, { "epoch": 2.604691004786379, "grad_norm": 3.3041601181030273, "learning_rate": 4.6876783850882656e-05, "loss": 3.4645, "step": 294950 }, { "epoch": 2.6051325526766633, "grad_norm": 4.813542366027832, "learning_rate": 4.686291492877859e-05, "loss": 3.1611, "step": 295000 }, { "epoch": 2.6055741005669475, "grad_norm": 3.489481210708618, "learning_rate": 4.684904624898775e-05, "loss": 3.192, "step": 295050 }, { "epoch": 2.6060156484572317, "grad_norm": 2.169651508331299, "learning_rate": 4.6835177812581365e-05, "loss": 3.0871, "step": 295100 }, { "epoch": 2.606457196347516, "grad_norm": 4.2794976234436035, "learning_rate": 4.682130962063068e-05, "loss": 2.9444, "step": 295150 }, { "epoch": 2.6068987442378, "grad_norm": 1.5134773254394531, "learning_rate": 4.680744167420684e-05, "loss": 3.4755, "step": 295200 }, { "epoch": 2.607340292128084, "grad_norm": 3.729177951812744, "learning_rate": 4.679357397438106e-05, "loss": 3.023, "step": 295250 }, { "epoch": 2.6077818400183683, "grad_norm": 2.0980734825134277, "learning_rate": 4.677970652222453e-05, "loss": 3.2866, "step": 295300 }, { "epoch": 2.6082233879086525, "grad_norm": 1.0635056495666504, "learning_rate": 4.6765839318808346e-05, "loss": 3.0967, "step": 295350 }, { "epoch": 2.6086649357989367, "grad_norm": 3.555896043777466, "learning_rate": 4.675197236520364e-05, "loss": 3.0746, "step": 295400 }, { "epoch": 2.609106483689221, "grad_norm": 2.1077182292938232, "learning_rate": 4.673810566248155e-05, "loss": 3.1872, "step": 295450 }, { "epoch": 2.609548031579505, "grad_norm": 4.780736923217773, "learning_rate": 4.6724239211713116e-05, "loss": 3.2382, "step": 295500 }, { "epoch": 2.609989579469789, "grad_norm": 5.460337162017822, "learning_rate": 4.6710373013969435e-05, "loss": 3.2732, "step": 295550 }, { "epoch": 2.6104311273600733, "grad_norm": 4.6254987716674805, "learning_rate": 4.669650707032153e-05, "loss": 3.1642, "step": 295600 }, { "epoch": 2.6108726752503575, "grad_norm": 4.339382648468018, "learning_rate": 4.6682641381840465e-05, "loss": 3.3886, "step": 295650 }, { "epoch": 2.611314223140642, "grad_norm": 5.324868202209473, "learning_rate": 4.6668775949597213e-05, "loss": 3.1986, "step": 295700 }, { "epoch": 2.611755771030926, "grad_norm": 3.2634241580963135, "learning_rate": 4.6654910774662764e-05, "loss": 2.9939, "step": 295750 }, { "epoch": 2.6121973189212104, "grad_norm": 6.647667407989502, "learning_rate": 4.6641045858108106e-05, "loss": 3.2069, "step": 295800 }, { "epoch": 2.612638866811494, "grad_norm": 4.475151062011719, "learning_rate": 4.662718120100416e-05, "loss": 3.0787, "step": 295850 }, { "epoch": 2.6130804147017788, "grad_norm": 4.119542121887207, "learning_rate": 4.6613316804421864e-05, "loss": 2.8406, "step": 295900 }, { "epoch": 2.6135219625920625, "grad_norm": 2.5736989974975586, "learning_rate": 4.659945266943214e-05, "loss": 3.1168, "step": 295950 }, { "epoch": 2.613963510482347, "grad_norm": 2.5882914066314697, "learning_rate": 4.658558879710585e-05, "loss": 3.003, "step": 296000 }, { "epoch": 2.614405058372631, "grad_norm": 1.7595422267913818, "learning_rate": 4.6571725188513866e-05, "loss": 2.8973, "step": 296050 }, { "epoch": 2.6148466062629154, "grad_norm": 4.367370128631592, "learning_rate": 4.6557861844727056e-05, "loss": 3.1427, "step": 296100 }, { "epoch": 2.6152881541531996, "grad_norm": 5.406118869781494, "learning_rate": 4.6543998766816225e-05, "loss": 2.9242, "step": 296150 }, { "epoch": 2.6157297020434838, "grad_norm": 3.174687623977661, "learning_rate": 4.653013595585217e-05, "loss": 2.6723, "step": 296200 }, { "epoch": 2.616171249933768, "grad_norm": 2.015441417694092, "learning_rate": 4.6516273412905705e-05, "loss": 3.089, "step": 296250 }, { "epoch": 2.616612797824052, "grad_norm": 2.5887811183929443, "learning_rate": 4.650241113904759e-05, "loss": 3.6981, "step": 296300 }, { "epoch": 2.6170543457143363, "grad_norm": 3.787632703781128, "learning_rate": 4.648854913534853e-05, "loss": 2.9591, "step": 296350 }, { "epoch": 2.6174958936046204, "grad_norm": 4.176799297332764, "learning_rate": 4.647468740287927e-05, "loss": 3.3402, "step": 296400 }, { "epoch": 2.6179374414949046, "grad_norm": 5.251779556274414, "learning_rate": 4.646082594271054e-05, "loss": 3.2066, "step": 296450 }, { "epoch": 2.6183789893851888, "grad_norm": 1.3894294500350952, "learning_rate": 4.644696475591298e-05, "loss": 3.1259, "step": 296500 }, { "epoch": 2.618820537275473, "grad_norm": 3.531935453414917, "learning_rate": 4.6433103843557255e-05, "loss": 2.969, "step": 296550 }, { "epoch": 2.619262085165757, "grad_norm": 2.11635160446167, "learning_rate": 4.6419243206714036e-05, "loss": 3.258, "step": 296600 }, { "epoch": 2.6197036330560413, "grad_norm": 1.3269505500793457, "learning_rate": 4.640538284645389e-05, "loss": 3.1557, "step": 296650 }, { "epoch": 2.6201451809463254, "grad_norm": 3.006119728088379, "learning_rate": 4.639152276384745e-05, "loss": 3.4872, "step": 296700 }, { "epoch": 2.6205867288366096, "grad_norm": 2.579406499862671, "learning_rate": 4.637766295996526e-05, "loss": 3.0828, "step": 296750 }, { "epoch": 2.6210282767268938, "grad_norm": 7.646761894226074, "learning_rate": 4.636380343587792e-05, "loss": 3.1978, "step": 296800 }, { "epoch": 2.621469824617178, "grad_norm": 2.216815233230591, "learning_rate": 4.634994419265591e-05, "loss": 3.3639, "step": 296850 }, { "epoch": 2.621911372507462, "grad_norm": 2.42928409576416, "learning_rate": 4.633608523136977e-05, "loss": 3.2401, "step": 296900 }, { "epoch": 2.6223529203977463, "grad_norm": 2.905820846557617, "learning_rate": 4.632222655308998e-05, "loss": 3.4976, "step": 296950 }, { "epoch": 2.6227944682880304, "grad_norm": 4.8869452476501465, "learning_rate": 4.630836815888699e-05, "loss": 3.3123, "step": 297000 }, { "epoch": 2.6227944682880304, "eval_asr_loss": 0.9384688538493438, "eval_loss": 2.7999284267425537, "eval_runtime": 20.7922, "eval_samples_per_second": 36.937, "eval_steps_per_second": 9.234, "eval_tts_loss": 5.992233712621294, "step": 297000 }, { "epoch": 2.6232360161783146, "grad_norm": 3.5016157627105713, "learning_rate": 4.6294510049831256e-05, "loss": 3.2508, "step": 297050 }, { "epoch": 2.6236775640685988, "grad_norm": 4.894611358642578, "learning_rate": 4.628065222699322e-05, "loss": 3.0003, "step": 297100 }, { "epoch": 2.624119111958883, "grad_norm": 2.043489456176758, "learning_rate": 4.626679469144323e-05, "loss": 3.454, "step": 297150 }, { "epoch": 2.624560659849167, "grad_norm": 5.286726951599121, "learning_rate": 4.62529374442517e-05, "loss": 2.7802, "step": 297200 }, { "epoch": 2.6250022077394517, "grad_norm": 3.238809823989868, "learning_rate": 4.6239080486489004e-05, "loss": 3.6162, "step": 297250 }, { "epoch": 2.6254437556297354, "grad_norm": 4.287550926208496, "learning_rate": 4.622522381922543e-05, "loss": 3.168, "step": 297300 }, { "epoch": 2.62588530352002, "grad_norm": 1.4111820459365845, "learning_rate": 4.621136744353129e-05, "loss": 3.2215, "step": 297350 }, { "epoch": 2.6263268514103038, "grad_norm": 4.286337852478027, "learning_rate": 4.6197511360476896e-05, "loss": 3.0115, "step": 297400 }, { "epoch": 2.6267683993005884, "grad_norm": 3.8721513748168945, "learning_rate": 4.618365557113253e-05, "loss": 2.7633, "step": 297450 }, { "epoch": 2.627209947190872, "grad_norm": 1.4832268953323364, "learning_rate": 4.61698000765684e-05, "loss": 2.8141, "step": 297500 }, { "epoch": 2.6276514950811567, "grad_norm": 2.4087724685668945, "learning_rate": 4.615594487785472e-05, "loss": 3.283, "step": 297550 }, { "epoch": 2.6280930429714404, "grad_norm": 3.738802433013916, "learning_rate": 4.6142089976061716e-05, "loss": 2.9416, "step": 297600 }, { "epoch": 2.628534590861725, "grad_norm": 2.8196942806243896, "learning_rate": 4.612823537225954e-05, "loss": 3.008, "step": 297650 }, { "epoch": 2.628976138752009, "grad_norm": 2.5648183822631836, "learning_rate": 4.6114381067518346e-05, "loss": 3.5696, "step": 297700 }, { "epoch": 2.6294176866422934, "grad_norm": 2.4913628101348877, "learning_rate": 4.610052706290828e-05, "loss": 3.0625, "step": 297750 }, { "epoch": 2.6298592345325775, "grad_norm": 3.4564943313598633, "learning_rate": 4.608667335949942e-05, "loss": 3.2936, "step": 297800 }, { "epoch": 2.6303007824228617, "grad_norm": 3.451342821121216, "learning_rate": 4.6073097023415436e-05, "loss": 2.9205, "step": 297850 }, { "epoch": 2.630742330313146, "grad_norm": 3.1267523765563965, "learning_rate": 4.6059243919541914e-05, "loss": 2.8331, "step": 297900 }, { "epoch": 2.63118387820343, "grad_norm": 0.7203403115272522, "learning_rate": 4.604539112005839e-05, "loss": 2.4409, "step": 297950 }, { "epoch": 2.631625426093714, "grad_norm": 3.2679195404052734, "learning_rate": 4.603153862603488e-05, "loss": 3.2016, "step": 298000 }, { "epoch": 2.6320669739839984, "grad_norm": 9.252143859863281, "learning_rate": 4.6017686438541334e-05, "loss": 2.9141, "step": 298050 }, { "epoch": 2.6325085218742825, "grad_norm": 1.848100185394287, "learning_rate": 4.600383455864775e-05, "loss": 2.9975, "step": 298100 }, { "epoch": 2.6329500697645667, "grad_norm": 2.3454816341400146, "learning_rate": 4.598998298742407e-05, "loss": 2.7835, "step": 298150 }, { "epoch": 2.633391617654851, "grad_norm": 6.622324466705322, "learning_rate": 4.59761317259402e-05, "loss": 3.149, "step": 298200 }, { "epoch": 2.633833165545135, "grad_norm": 2.2143781185150146, "learning_rate": 4.596228077526602e-05, "loss": 2.7229, "step": 298250 }, { "epoch": 2.634274713435419, "grad_norm": 3.315248489379883, "learning_rate": 4.594843013647143e-05, "loss": 3.511, "step": 298300 }, { "epoch": 2.6347162613257034, "grad_norm": 4.714132308959961, "learning_rate": 4.593457981062623e-05, "loss": 3.0418, "step": 298350 }, { "epoch": 2.6351578092159875, "grad_norm": 3.493767738342285, "learning_rate": 4.592072979880029e-05, "loss": 3.173, "step": 298400 }, { "epoch": 2.6355993571062717, "grad_norm": 1.5848151445388794, "learning_rate": 4.590688010206336e-05, "loss": 3.4858, "step": 298450 }, { "epoch": 2.636040904996556, "grad_norm": 2.4064888954162598, "learning_rate": 4.589303072148526e-05, "loss": 3.1683, "step": 298500 }, { "epoch": 2.63648245288684, "grad_norm": 2.5682501792907715, "learning_rate": 4.58791816581357e-05, "loss": 3.2574, "step": 298550 }, { "epoch": 2.636924000777124, "grad_norm": 4.213366985321045, "learning_rate": 4.58653329130844e-05, "loss": 2.854, "step": 298600 }, { "epoch": 2.6373655486674084, "grad_norm": 7.676827907562256, "learning_rate": 4.5851484487401083e-05, "loss": 3.1225, "step": 298650 }, { "epoch": 2.6378070965576925, "grad_norm": 3.854579448699951, "learning_rate": 4.58376363821554e-05, "loss": 3.2193, "step": 298700 }, { "epoch": 2.6382486444479767, "grad_norm": 2.3340725898742676, "learning_rate": 4.582378859841703e-05, "loss": 3.525, "step": 298750 }, { "epoch": 2.638690192338261, "grad_norm": 1.5447099208831787, "learning_rate": 4.580994113725557e-05, "loss": 2.9619, "step": 298800 }, { "epoch": 2.639131740228545, "grad_norm": 1.577736258506775, "learning_rate": 4.579609399974062e-05, "loss": 2.8854, "step": 298850 }, { "epoch": 2.6395732881188296, "grad_norm": 2.4056649208068848, "learning_rate": 4.5782247186941756e-05, "loss": 2.949, "step": 298900 }, { "epoch": 2.6400148360091134, "grad_norm": 2.574186325073242, "learning_rate": 4.576840069992854e-05, "loss": 2.8231, "step": 298950 }, { "epoch": 2.640456383899398, "grad_norm": 4.539966583251953, "learning_rate": 4.57545545397705e-05, "loss": 3.3256, "step": 299000 }, { "epoch": 2.6408979317896817, "grad_norm": 2.891242504119873, "learning_rate": 4.574070870753709e-05, "loss": 3.3753, "step": 299050 }, { "epoch": 2.6413394796799663, "grad_norm": 3.2183191776275635, "learning_rate": 4.5726863204297845e-05, "loss": 3.0101, "step": 299100 }, { "epoch": 2.64178102757025, "grad_norm": 1.1313601732254028, "learning_rate": 4.5713018031122176e-05, "loss": 3.5251, "step": 299150 }, { "epoch": 2.6422225754605346, "grad_norm": 1.6412028074264526, "learning_rate": 4.5699173189079496e-05, "loss": 2.6935, "step": 299200 }, { "epoch": 2.642664123350819, "grad_norm": 4.884433746337891, "learning_rate": 4.568532867923922e-05, "loss": 2.7508, "step": 299250 }, { "epoch": 2.643105671241103, "grad_norm": 1.3363255262374878, "learning_rate": 4.5671484502670735e-05, "loss": 2.8586, "step": 299300 }, { "epoch": 2.643547219131387, "grad_norm": 3.129206418991089, "learning_rate": 4.565764066044336e-05, "loss": 3.1153, "step": 299350 }, { "epoch": 2.6439887670216713, "grad_norm": 2.928429365158081, "learning_rate": 4.56437971536264e-05, "loss": 3.3608, "step": 299400 }, { "epoch": 2.6444303149119555, "grad_norm": 2.100141763687134, "learning_rate": 4.56299539832892e-05, "loss": 2.8026, "step": 299450 }, { "epoch": 2.6448718628022396, "grad_norm": 4.144262790679932, "learning_rate": 4.561611115050098e-05, "loss": 3.1253, "step": 299500 }, { "epoch": 2.645313410692524, "grad_norm": 4.296358585357666, "learning_rate": 4.5602268656331e-05, "loss": 3.0857, "step": 299550 }, { "epoch": 2.645754958582808, "grad_norm": 5.298852920532227, "learning_rate": 4.558842650184848e-05, "loss": 3.3354, "step": 299600 }, { "epoch": 2.646196506473092, "grad_norm": 1.4819879531860352, "learning_rate": 4.557458468812261e-05, "loss": 3.0, "step": 299650 }, { "epoch": 2.6466380543633763, "grad_norm": 4.229703426361084, "learning_rate": 4.556074321622253e-05, "loss": 3.1261, "step": 299700 }, { "epoch": 2.6470796022536605, "grad_norm": 2.6015565395355225, "learning_rate": 4.5546902087217405e-05, "loss": 3.1247, "step": 299750 }, { "epoch": 2.6475211501439446, "grad_norm": 3.6581926345825195, "learning_rate": 4.553306130217635e-05, "loss": 3.3092, "step": 299800 }, { "epoch": 2.647962698034229, "grad_norm": 4.637612819671631, "learning_rate": 4.5519220862168394e-05, "loss": 2.791, "step": 299850 }, { "epoch": 2.648404245924513, "grad_norm": 2.077911853790283, "learning_rate": 4.550538076826265e-05, "loss": 3.0682, "step": 299900 }, { "epoch": 2.648845793814797, "grad_norm": 1.0989296436309814, "learning_rate": 4.549154102152814e-05, "loss": 3.5596, "step": 299950 }, { "epoch": 2.6492873417050813, "grad_norm": 2.732758045196533, "learning_rate": 4.5477701623033833e-05, "loss": 2.9824, "step": 300000 }, { "epoch": 2.6492873417050813, "eval_asr_loss": 0.9230900500634989, "eval_loss": 2.797144651412964, "eval_runtime": 21.0658, "eval_samples_per_second": 36.457, "eval_steps_per_second": 9.114, "eval_tts_loss": 6.073927639521335, "step": 300000 }, { "epoch": 2.6497288895953655, "grad_norm": 3.1998279094696045, "learning_rate": 4.546386257384872e-05, "loss": 3.0971, "step": 300050 }, { "epoch": 2.6501704374856496, "grad_norm": 1.8245770931243896, "learning_rate": 4.5450023875041775e-05, "loss": 2.8162, "step": 300100 }, { "epoch": 2.650611985375934, "grad_norm": 4.419822692871094, "learning_rate": 4.543618552768191e-05, "loss": 2.7321, "step": 300150 }, { "epoch": 2.651053533266218, "grad_norm": 4.471705913543701, "learning_rate": 4.5422347532838e-05, "loss": 2.8468, "step": 300200 }, { "epoch": 2.651495081156502, "grad_norm": 2.145991563796997, "learning_rate": 4.540850989157892e-05, "loss": 3.0741, "step": 300250 }, { "epoch": 2.6519366290467863, "grad_norm": 5.097830295562744, "learning_rate": 4.539494934722312e-05, "loss": 2.434, "step": 300300 }, { "epoch": 2.6523781769370705, "grad_norm": 1.2167659997940063, "learning_rate": 4.53811124092153e-05, "loss": 2.8698, "step": 300350 }, { "epoch": 2.6528197248273546, "grad_norm": 1.6864746809005737, "learning_rate": 4.536727582797737e-05, "loss": 3.16, "step": 300400 }, { "epoch": 2.6532612727176392, "grad_norm": 3.4063291549682617, "learning_rate": 4.53534396045781e-05, "loss": 2.6729, "step": 300450 }, { "epoch": 2.653702820607923, "grad_norm": 1.9698035717010498, "learning_rate": 4.5339603740086225e-05, "loss": 3.2922, "step": 300500 }, { "epoch": 2.6541443684982076, "grad_norm": 3.4441463947296143, "learning_rate": 4.5325768235570425e-05, "loss": 3.1009, "step": 300550 }, { "epoch": 2.6545859163884913, "grad_norm": 4.827519416809082, "learning_rate": 4.531193309209941e-05, "loss": 3.2941, "step": 300600 }, { "epoch": 2.655027464278776, "grad_norm": 1.9491372108459473, "learning_rate": 4.529809831074179e-05, "loss": 2.9945, "step": 300650 }, { "epoch": 2.6554690121690596, "grad_norm": 4.515385627746582, "learning_rate": 4.528426389256623e-05, "loss": 3.1191, "step": 300700 }, { "epoch": 2.6559105600593442, "grad_norm": 3.477837085723877, "learning_rate": 4.527042983864127e-05, "loss": 3.2159, "step": 300750 }, { "epoch": 2.656352107949628, "grad_norm": 1.3729954957962036, "learning_rate": 4.525659615003552e-05, "loss": 2.6328, "step": 300800 }, { "epoch": 2.6567936558399126, "grad_norm": 4.440359115600586, "learning_rate": 4.52427628278175e-05, "loss": 3.3589, "step": 300850 }, { "epoch": 2.6572352037301967, "grad_norm": 3.0278029441833496, "learning_rate": 4.5228929873055686e-05, "loss": 3.2189, "step": 300900 }, { "epoch": 2.657676751620481, "grad_norm": 1.4693509340286255, "learning_rate": 4.521509728681858e-05, "loss": 3.0875, "step": 300950 }, { "epoch": 2.658118299510765, "grad_norm": 2.0322816371917725, "learning_rate": 4.520126507017466e-05, "loss": 3.1072, "step": 301000 }, { "epoch": 2.6585598474010492, "grad_norm": 2.6976468563079834, "learning_rate": 4.518743322419231e-05, "loss": 3.0421, "step": 301050 }, { "epoch": 2.6590013952913334, "grad_norm": 3.529794692993164, "learning_rate": 4.5173601749939925e-05, "loss": 2.6632, "step": 301100 }, { "epoch": 2.6594429431816176, "grad_norm": 4.532125949859619, "learning_rate": 4.515977064848589e-05, "loss": 2.8655, "step": 301150 }, { "epoch": 2.6598844910719017, "grad_norm": 4.339880466461182, "learning_rate": 4.514593992089853e-05, "loss": 3.489, "step": 301200 }, { "epoch": 2.660326038962186, "grad_norm": 4.413679122924805, "learning_rate": 4.513210956824614e-05, "loss": 3.0347, "step": 301250 }, { "epoch": 2.66076758685247, "grad_norm": 1.8697099685668945, "learning_rate": 4.5118279591597e-05, "loss": 3.0544, "step": 301300 }, { "epoch": 2.6612091347427542, "grad_norm": 2.72141170501709, "learning_rate": 4.5104449992019396e-05, "loss": 3.1239, "step": 301350 }, { "epoch": 2.6616506826330384, "grad_norm": 5.316091060638428, "learning_rate": 4.50906207705815e-05, "loss": 3.3367, "step": 301400 }, { "epoch": 2.6620922305233226, "grad_norm": 11.1223783493042, "learning_rate": 4.507679192835151e-05, "loss": 2.8769, "step": 301450 }, { "epoch": 2.6625337784136067, "grad_norm": 4.928714275360107, "learning_rate": 4.506296346639761e-05, "loss": 3.2672, "step": 301500 }, { "epoch": 2.662975326303891, "grad_norm": 2.1855556964874268, "learning_rate": 4.50491353857879e-05, "loss": 3.2711, "step": 301550 }, { "epoch": 2.663416874194175, "grad_norm": 3.0553059577941895, "learning_rate": 4.503530768759052e-05, "loss": 3.0827, "step": 301600 }, { "epoch": 2.663858422084459, "grad_norm": 3.683285713195801, "learning_rate": 4.502148037287352e-05, "loss": 3.0319, "step": 301650 }, { "epoch": 2.6642999699747434, "grad_norm": 3.852095603942871, "learning_rate": 4.5007653442704925e-05, "loss": 3.5811, "step": 301700 }, { "epoch": 2.6647415178650276, "grad_norm": 3.5801048278808594, "learning_rate": 4.4993826898152775e-05, "loss": 2.9935, "step": 301750 }, { "epoch": 2.6651830657553117, "grad_norm": 3.0393879413604736, "learning_rate": 4.498000074028505e-05, "loss": 3.2983, "step": 301800 }, { "epoch": 2.665624613645596, "grad_norm": 2.55759334564209, "learning_rate": 4.496617497016972e-05, "loss": 2.9623, "step": 301850 }, { "epoch": 2.66606616153588, "grad_norm": 4.101271152496338, "learning_rate": 4.495234958887465e-05, "loss": 3.1303, "step": 301900 }, { "epoch": 2.666507709426164, "grad_norm": 5.422536849975586, "learning_rate": 4.493852459746779e-05, "loss": 3.2557, "step": 301950 }, { "epoch": 2.6669492573164484, "grad_norm": 4.791871070861816, "learning_rate": 4.4924699997017e-05, "loss": 2.6546, "step": 302000 }, { "epoch": 2.6673908052067326, "grad_norm": 2.7441608905792236, "learning_rate": 4.4910875788590075e-05, "loss": 3.0687, "step": 302050 }, { "epoch": 2.667832353097017, "grad_norm": 2.2193844318389893, "learning_rate": 4.4897051973254836e-05, "loss": 3.2908, "step": 302100 }, { "epoch": 2.668273900987301, "grad_norm": 2.66637921333313, "learning_rate": 4.488322855207909e-05, "loss": 2.9887, "step": 302150 }, { "epoch": 2.6687154488775855, "grad_norm": 3.099656820297241, "learning_rate": 4.4869405526130525e-05, "loss": 3.5454, "step": 302200 }, { "epoch": 2.669156996767869, "grad_norm": 4.8532328605651855, "learning_rate": 4.485558289647688e-05, "loss": 2.7035, "step": 302250 }, { "epoch": 2.669598544658154, "grad_norm": 3.984295129776001, "learning_rate": 4.484176066418582e-05, "loss": 2.913, "step": 302300 }, { "epoch": 2.6700400925484375, "grad_norm": 4.139182090759277, "learning_rate": 4.482793883032504e-05, "loss": 3.1547, "step": 302350 }, { "epoch": 2.670481640438722, "grad_norm": 2.4230291843414307, "learning_rate": 4.481411739596212e-05, "loss": 3.0073, "step": 302400 }, { "epoch": 2.6709231883290063, "grad_norm": 3.228363513946533, "learning_rate": 4.480029636216463e-05, "loss": 3.388, "step": 302450 }, { "epoch": 2.6713647362192905, "grad_norm": 4.381591320037842, "learning_rate": 4.478647573000019e-05, "loss": 2.9609, "step": 302500 }, { "epoch": 2.6718062841095747, "grad_norm": 3.3779850006103516, "learning_rate": 4.477265550053627e-05, "loss": 2.4954, "step": 302550 }, { "epoch": 2.672247831999859, "grad_norm": 3.3229904174804688, "learning_rate": 4.475883567484038e-05, "loss": 2.9574, "step": 302600 }, { "epoch": 2.672689379890143, "grad_norm": 2.207456588745117, "learning_rate": 4.4745016253980013e-05, "loss": 2.809, "step": 302650 }, { "epoch": 2.673130927780427, "grad_norm": 1.531559944152832, "learning_rate": 4.473119723902256e-05, "loss": 3.1377, "step": 302700 }, { "epoch": 2.6735724756707113, "grad_norm": 2.1203765869140625, "learning_rate": 4.4717378631035446e-05, "loss": 3.2232, "step": 302750 }, { "epoch": 2.6740140235609955, "grad_norm": 1.337319016456604, "learning_rate": 4.470356043108605e-05, "loss": 2.8557, "step": 302800 }, { "epoch": 2.6744555714512797, "grad_norm": 2.347046375274658, "learning_rate": 4.468974264024168e-05, "loss": 2.9385, "step": 302850 }, { "epoch": 2.674897119341564, "grad_norm": 3.123553514480591, "learning_rate": 4.4675925259569654e-05, "loss": 2.8076, "step": 302900 }, { "epoch": 2.675338667231848, "grad_norm": 5.56646203994751, "learning_rate": 4.4662108290137274e-05, "loss": 2.9961, "step": 302950 }, { "epoch": 2.675780215122132, "grad_norm": 3.2375051975250244, "learning_rate": 4.464829173301177e-05, "loss": 3.0364, "step": 303000 }, { "epoch": 2.675780215122132, "eval_asr_loss": 0.9247981151248457, "eval_loss": 2.7983243465423584, "eval_runtime": 20.4156, "eval_samples_per_second": 37.618, "eval_steps_per_second": 9.405, "eval_tts_loss": 6.026089976112239, "step": 303000 }, { "epoch": 2.6762217630124163, "grad_norm": 1.3980324268341064, "learning_rate": 4.4634475589260336e-05, "loss": 2.8118, "step": 303050 }, { "epoch": 2.6766633109027005, "grad_norm": 10.281167984008789, "learning_rate": 4.462065985995015e-05, "loss": 3.3077, "step": 303100 }, { "epoch": 2.6771048587929847, "grad_norm": 3.2088301181793213, "learning_rate": 4.460684454614841e-05, "loss": 3.1795, "step": 303150 }, { "epoch": 2.677546406683269, "grad_norm": 2.519932985305786, "learning_rate": 4.459302964892218e-05, "loss": 2.9971, "step": 303200 }, { "epoch": 2.677987954573553, "grad_norm": 3.5058963298797607, "learning_rate": 4.4579215169338545e-05, "loss": 3.2818, "step": 303250 }, { "epoch": 2.678429502463837, "grad_norm": 5.76242733001709, "learning_rate": 4.4565401108464596e-05, "loss": 2.849, "step": 303300 }, { "epoch": 2.6788710503541213, "grad_norm": 0.8211635947227478, "learning_rate": 4.4551587467367315e-05, "loss": 3.325, "step": 303350 }, { "epoch": 2.6793125982444055, "grad_norm": 1.9156229496002197, "learning_rate": 4.45377742471137e-05, "loss": 2.9109, "step": 303400 }, { "epoch": 2.6797541461346897, "grad_norm": 5.772945880889893, "learning_rate": 4.452396144877071e-05, "loss": 3.3132, "step": 303450 }, { "epoch": 2.680195694024974, "grad_norm": 1.34353768825531, "learning_rate": 4.451014907340525e-05, "loss": 3.3562, "step": 303500 }, { "epoch": 2.680637241915258, "grad_norm": 3.02319073677063, "learning_rate": 4.449633712208424e-05, "loss": 3.3313, "step": 303550 }, { "epoch": 2.681078789805542, "grad_norm": 2.819284200668335, "learning_rate": 4.448252559587449e-05, "loss": 3.4096, "step": 303600 }, { "epoch": 2.6815203376958268, "grad_norm": 1.4359670877456665, "learning_rate": 4.4468714495842884e-05, "loss": 2.6051, "step": 303650 }, { "epoch": 2.6819618855861105, "grad_norm": 3.027996301651001, "learning_rate": 4.445490382305616e-05, "loss": 3.2964, "step": 303700 }, { "epoch": 2.682403433476395, "grad_norm": 2.8490335941314697, "learning_rate": 4.444109357858109e-05, "loss": 2.9871, "step": 303750 }, { "epoch": 2.682844981366679, "grad_norm": 4.932562351226807, "learning_rate": 4.442755995557154e-05, "loss": 3.0444, "step": 303800 }, { "epoch": 2.6832865292569634, "grad_norm": 1.8189077377319336, "learning_rate": 4.4413750562300594e-05, "loss": 2.5197, "step": 303850 }, { "epoch": 2.683728077147247, "grad_norm": 4.074760437011719, "learning_rate": 4.439994160052004e-05, "loss": 2.98, "step": 303900 }, { "epoch": 2.6841696250375318, "grad_norm": 3.745225429534912, "learning_rate": 4.438613307129649e-05, "loss": 3.0599, "step": 303950 }, { "epoch": 2.6846111729278155, "grad_norm": 4.35814905166626, "learning_rate": 4.4372324975696575e-05, "loss": 3.0209, "step": 304000 }, { "epoch": 2.6850527208181, "grad_norm": 4.706446647644043, "learning_rate": 4.435851731478683e-05, "loss": 2.7341, "step": 304050 }, { "epoch": 2.6854942687083843, "grad_norm": 1.5683945417404175, "learning_rate": 4.434471008963379e-05, "loss": 3.3067, "step": 304100 }, { "epoch": 2.6859358165986684, "grad_norm": 2.16620135307312, "learning_rate": 4.4330903301303925e-05, "loss": 3.1209, "step": 304150 }, { "epoch": 2.6863773644889526, "grad_norm": 2.664396047592163, "learning_rate": 4.431709695086373e-05, "loss": 3.0212, "step": 304200 }, { "epoch": 2.6868189123792368, "grad_norm": 1.0106053352355957, "learning_rate": 4.4303291039379594e-05, "loss": 2.8635, "step": 304250 }, { "epoch": 2.687260460269521, "grad_norm": 3.7168519496917725, "learning_rate": 4.428948556791791e-05, "loss": 2.8509, "step": 304300 }, { "epoch": 2.687702008159805, "grad_norm": 1.4611485004425049, "learning_rate": 4.427568053754506e-05, "loss": 2.991, "step": 304350 }, { "epoch": 2.6881435560500893, "grad_norm": 1.7934060096740723, "learning_rate": 4.426187594932732e-05, "loss": 2.9786, "step": 304400 }, { "epoch": 2.6885851039403734, "grad_norm": 4.810496807098389, "learning_rate": 4.424807180433102e-05, "loss": 2.9704, "step": 304450 }, { "epoch": 2.6890266518306576, "grad_norm": 5.93510103225708, "learning_rate": 4.42342681036224e-05, "loss": 3.541, "step": 304500 }, { "epoch": 2.6894681997209418, "grad_norm": 2.5243332386016846, "learning_rate": 4.422046484826769e-05, "loss": 3.0569, "step": 304550 }, { "epoch": 2.689909747611226, "grad_norm": 2.0581016540527344, "learning_rate": 4.420666203933303e-05, "loss": 3.2669, "step": 304600 }, { "epoch": 2.69035129550151, "grad_norm": 3.873783588409424, "learning_rate": 4.419285967788463e-05, "loss": 3.0881, "step": 304650 }, { "epoch": 2.6907928433917943, "grad_norm": 3.6419601440429688, "learning_rate": 4.417905776498858e-05, "loss": 3.2786, "step": 304700 }, { "epoch": 2.6912343912820784, "grad_norm": 3.1941776275634766, "learning_rate": 4.4165256301710935e-05, "loss": 3.5454, "step": 304750 }, { "epoch": 2.6916759391723626, "grad_norm": 3.912426710128784, "learning_rate": 4.415145528911777e-05, "loss": 2.6355, "step": 304800 }, { "epoch": 2.6921174870626468, "grad_norm": 2.2543528079986572, "learning_rate": 4.413765472827511e-05, "loss": 3.1159, "step": 304850 }, { "epoch": 2.692559034952931, "grad_norm": 2.41107439994812, "learning_rate": 4.412385462024888e-05, "loss": 3.0693, "step": 304900 }, { "epoch": 2.693000582843215, "grad_norm": 1.9796345233917236, "learning_rate": 4.4110054966105055e-05, "loss": 3.028, "step": 304950 }, { "epoch": 2.6934421307334993, "grad_norm": 4.86181116104126, "learning_rate": 4.409625576690956e-05, "loss": 3.0367, "step": 305000 }, { "epoch": 2.6938836786237834, "grad_norm": 3.1111502647399902, "learning_rate": 4.408245702372824e-05, "loss": 2.9046, "step": 305050 }, { "epoch": 2.6943252265140676, "grad_norm": 3.0094757080078125, "learning_rate": 4.4068658737626925e-05, "loss": 3.1015, "step": 305100 }, { "epoch": 2.6947667744043518, "grad_norm": 3.2387073040008545, "learning_rate": 4.405486090967142e-05, "loss": 3.6045, "step": 305150 }, { "epoch": 2.695208322294636, "grad_norm": 3.487652063369751, "learning_rate": 4.404106354092752e-05, "loss": 2.8822, "step": 305200 }, { "epoch": 2.69564987018492, "grad_norm": 1.4247080087661743, "learning_rate": 4.402726663246092e-05, "loss": 3.1466, "step": 305250 }, { "epoch": 2.6960914180752047, "grad_norm": 2.781226396560669, "learning_rate": 4.401347018533731e-05, "loss": 3.53, "step": 305300 }, { "epoch": 2.6965329659654884, "grad_norm": 5.321645736694336, "learning_rate": 4.399967420062239e-05, "loss": 3.3036, "step": 305350 }, { "epoch": 2.696974513855773, "grad_norm": 1.5104516744613647, "learning_rate": 4.3985878679381744e-05, "loss": 2.8447, "step": 305400 }, { "epoch": 2.6974160617460567, "grad_norm": 3.820126533508301, "learning_rate": 4.397208362268097e-05, "loss": 3.3141, "step": 305450 }, { "epoch": 2.6978576096363414, "grad_norm": 5.626131534576416, "learning_rate": 4.3958289031585634e-05, "loss": 2.8522, "step": 305500 }, { "epoch": 2.698299157526625, "grad_norm": 3.3993818759918213, "learning_rate": 4.3944494907161224e-05, "loss": 3.1414, "step": 305550 }, { "epoch": 2.6987407054169097, "grad_norm": 0.7934486865997314, "learning_rate": 4.393070125047323e-05, "loss": 3.3691, "step": 305600 }, { "epoch": 2.699182253307194, "grad_norm": 4.684044361114502, "learning_rate": 4.391690806258712e-05, "loss": 3.0585, "step": 305650 }, { "epoch": 2.699623801197478, "grad_norm": 4.107460021972656, "learning_rate": 4.390311534456826e-05, "loss": 2.9703, "step": 305700 }, { "epoch": 2.700065349087762, "grad_norm": 2.2025563716888428, "learning_rate": 4.388932309748203e-05, "loss": 2.8164, "step": 305750 }, { "epoch": 2.7005068969780464, "grad_norm": 3.694951057434082, "learning_rate": 4.3875531322393786e-05, "loss": 3.2012, "step": 305800 }, { "epoch": 2.7009484448683305, "grad_norm": 4.221449851989746, "learning_rate": 4.386174002036883e-05, "loss": 3.266, "step": 305850 }, { "epoch": 2.7013899927586147, "grad_norm": 4.877699851989746, "learning_rate": 4.384794919247238e-05, "loss": 3.254, "step": 305900 }, { "epoch": 2.701831540648899, "grad_norm": 3.6321752071380615, "learning_rate": 4.383443464215997e-05, "loss": 3.4156, "step": 305950 }, { "epoch": 2.702273088539183, "grad_norm": 4.209252834320068, "learning_rate": 4.3820644756180606e-05, "loss": 3.0257, "step": 306000 }, { "epoch": 2.702273088539183, "eval_asr_loss": 0.9160837718748359, "eval_loss": 2.794285535812378, "eval_runtime": 20.7724, "eval_samples_per_second": 36.972, "eval_steps_per_second": 9.243, "eval_tts_loss": 6.03984062802497, "step": 306000 }, { "epoch": 2.702714636429467, "grad_norm": 4.08154821395874, "learning_rate": 4.380685534750406e-05, "loss": 3.3301, "step": 306050 }, { "epoch": 2.7031561843197514, "grad_norm": 5.839825630187988, "learning_rate": 4.379306641719539e-05, "loss": 2.7157, "step": 306100 }, { "epoch": 2.7035977322100355, "grad_norm": 1.8244714736938477, "learning_rate": 4.377927796631973e-05, "loss": 3.2497, "step": 306150 }, { "epoch": 2.7040392801003197, "grad_norm": 5.8148674964904785, "learning_rate": 4.3765489995942074e-05, "loss": 3.1248, "step": 306200 }, { "epoch": 2.704480827990604, "grad_norm": 4.477723598480225, "learning_rate": 4.3751702507127484e-05, "loss": 2.5924, "step": 306250 }, { "epoch": 2.704922375880888, "grad_norm": 2.0836262702941895, "learning_rate": 4.3737915500940877e-05, "loss": 2.9361, "step": 306300 }, { "epoch": 2.705363923771172, "grad_norm": 3.2314107418060303, "learning_rate": 4.3724128978447175e-05, "loss": 3.0308, "step": 306350 }, { "epoch": 2.7058054716614564, "grad_norm": 1.5254595279693604, "learning_rate": 4.3710342940711324e-05, "loss": 2.9096, "step": 306400 }, { "epoch": 2.7062470195517405, "grad_norm": 2.003143310546875, "learning_rate": 4.369655738879812e-05, "loss": 2.7543, "step": 306450 }, { "epoch": 2.7066885674420247, "grad_norm": 1.9765498638153076, "learning_rate": 4.368277232377241e-05, "loss": 2.8406, "step": 306500 }, { "epoch": 2.707130115332309, "grad_norm": 4.393587589263916, "learning_rate": 4.3668987746699e-05, "loss": 2.6241, "step": 306550 }, { "epoch": 2.707571663222593, "grad_norm": 3.539579391479492, "learning_rate": 4.3655203658642564e-05, "loss": 3.2463, "step": 306600 }, { "epoch": 2.708013211112877, "grad_norm": 1.2016280889511108, "learning_rate": 4.364142006066784e-05, "loss": 3.1931, "step": 306650 }, { "epoch": 2.7084547590031613, "grad_norm": 5.127782821655273, "learning_rate": 4.362763695383952e-05, "loss": 3.1517, "step": 306700 }, { "epoch": 2.7088963068934455, "grad_norm": 2.3786821365356445, "learning_rate": 4.3613854339222215e-05, "loss": 3.3234, "step": 306750 }, { "epoch": 2.7093378547837297, "grad_norm": 1.9418467283248901, "learning_rate": 4.360007221788048e-05, "loss": 3.0405, "step": 306800 }, { "epoch": 2.7097794026740143, "grad_norm": 8.632000923156738, "learning_rate": 4.358629059087892e-05, "loss": 3.0623, "step": 306850 }, { "epoch": 2.710220950564298, "grad_norm": 4.371127605438232, "learning_rate": 4.357250945928202e-05, "loss": 3.0975, "step": 306900 }, { "epoch": 2.7106624984545826, "grad_norm": 4.08177375793457, "learning_rate": 4.355872882415426e-05, "loss": 3.0845, "step": 306950 }, { "epoch": 2.7111040463448663, "grad_norm": 2.5738489627838135, "learning_rate": 4.3544948686560064e-05, "loss": 3.0473, "step": 307000 }, { "epoch": 2.711545594235151, "grad_norm": 5.952714443206787, "learning_rate": 4.353116904756388e-05, "loss": 3.2138, "step": 307050 }, { "epoch": 2.7119871421254347, "grad_norm": 3.6179895401000977, "learning_rate": 4.351738990823001e-05, "loss": 3.1114, "step": 307100 }, { "epoch": 2.7124286900157193, "grad_norm": 2.9119293689727783, "learning_rate": 4.350361126962278e-05, "loss": 2.8936, "step": 307150 }, { "epoch": 2.712870237906003, "grad_norm": 2.1952221393585205, "learning_rate": 4.348983313280653e-05, "loss": 3.0948, "step": 307200 }, { "epoch": 2.7133117857962876, "grad_norm": 3.9066545963287354, "learning_rate": 4.347605549884544e-05, "loss": 2.9807, "step": 307250 }, { "epoch": 2.713753333686572, "grad_norm": 3.0799546241760254, "learning_rate": 4.346227836880374e-05, "loss": 2.9914, "step": 307300 }, { "epoch": 2.714194881576856, "grad_norm": 6.069919586181641, "learning_rate": 4.34485017437456e-05, "loss": 3.442, "step": 307350 }, { "epoch": 2.71463642946714, "grad_norm": 2.2482821941375732, "learning_rate": 4.343472562473516e-05, "loss": 3.4695, "step": 307400 }, { "epoch": 2.7150779773574243, "grad_norm": 2.7209949493408203, "learning_rate": 4.342095001283648e-05, "loss": 2.9701, "step": 307450 }, { "epoch": 2.7155195252477085, "grad_norm": 2.6500911712646484, "learning_rate": 4.340717490911363e-05, "loss": 2.5419, "step": 307500 }, { "epoch": 2.7159610731379926, "grad_norm": 10.522100448608398, "learning_rate": 4.3393400314630616e-05, "loss": 3.1441, "step": 307550 }, { "epoch": 2.716402621028277, "grad_norm": 1.0489394664764404, "learning_rate": 4.337962623045139e-05, "loss": 3.1369, "step": 307600 }, { "epoch": 2.716844168918561, "grad_norm": 2.324272632598877, "learning_rate": 4.336585265763992e-05, "loss": 2.992, "step": 307650 }, { "epoch": 2.717285716808845, "grad_norm": 3.4038212299346924, "learning_rate": 4.335207959726007e-05, "loss": 3.35, "step": 307700 }, { "epoch": 2.7177272646991293, "grad_norm": 6.170582294464111, "learning_rate": 4.333830705037568e-05, "loss": 2.9648, "step": 307750 }, { "epoch": 2.7181688125894135, "grad_norm": 2.696762800216675, "learning_rate": 4.332453501805059e-05, "loss": 2.8957, "step": 307800 }, { "epoch": 2.7186103604796976, "grad_norm": 7.946045875549316, "learning_rate": 4.331076350134858e-05, "loss": 2.8041, "step": 307850 }, { "epoch": 2.719051908369982, "grad_norm": 1.397757887840271, "learning_rate": 4.329699250133335e-05, "loss": 2.8534, "step": 307900 }, { "epoch": 2.719493456260266, "grad_norm": 5.786844253540039, "learning_rate": 4.32832220190686e-05, "loss": 2.6829, "step": 307950 }, { "epoch": 2.71993500415055, "grad_norm": 1.3750723600387573, "learning_rate": 4.3269452055618e-05, "loss": 2.9984, "step": 308000 }, { "epoch": 2.7203765520408343, "grad_norm": 1.7699259519577026, "learning_rate": 4.3255682612045175e-05, "loss": 3.0921, "step": 308050 }, { "epoch": 2.7208180999311185, "grad_norm": 4.949068069458008, "learning_rate": 4.324191368941367e-05, "loss": 3.0209, "step": 308100 }, { "epoch": 2.7212596478214026, "grad_norm": 1.7897433042526245, "learning_rate": 4.3228145288787e-05, "loss": 2.7192, "step": 308150 }, { "epoch": 2.721701195711687, "grad_norm": 3.0119194984436035, "learning_rate": 4.321437741122872e-05, "loss": 2.948, "step": 308200 }, { "epoch": 2.722142743601971, "grad_norm": 3.1683261394500732, "learning_rate": 4.3200610057802214e-05, "loss": 3.3363, "step": 308250 }, { "epoch": 2.722584291492255, "grad_norm": 4.124976634979248, "learning_rate": 4.318684322957094e-05, "loss": 3.487, "step": 308300 }, { "epoch": 2.7230258393825393, "grad_norm": 2.313904047012329, "learning_rate": 4.317307692759826e-05, "loss": 3.4022, "step": 308350 }, { "epoch": 2.7234673872728234, "grad_norm": 2.7590084075927734, "learning_rate": 4.315931115294748e-05, "loss": 3.1168, "step": 308400 }, { "epoch": 2.7239089351631076, "grad_norm": 4.238140106201172, "learning_rate": 4.3145545906681914e-05, "loss": 3.1769, "step": 308450 }, { "epoch": 2.7243504830533922, "grad_norm": 4.481393337249756, "learning_rate": 4.313178118986479e-05, "loss": 2.8628, "step": 308500 }, { "epoch": 2.724792030943676, "grad_norm": 3.614778995513916, "learning_rate": 4.311801700355935e-05, "loss": 2.8892, "step": 308550 }, { "epoch": 2.7252335788339606, "grad_norm": 6.108620643615723, "learning_rate": 4.310425334882872e-05, "loss": 3.1083, "step": 308600 }, { "epoch": 2.7256751267242443, "grad_norm": 2.9988603591918945, "learning_rate": 4.309049022673605e-05, "loss": 3.057, "step": 308650 }, { "epoch": 2.726116674614529, "grad_norm": 2.0618791580200195, "learning_rate": 4.307672763834444e-05, "loss": 2.6221, "step": 308700 }, { "epoch": 2.7265582225048126, "grad_norm": 4.9803643226623535, "learning_rate": 4.306296558471688e-05, "loss": 3.0538, "step": 308750 }, { "epoch": 2.7269997703950972, "grad_norm": 6.645568370819092, "learning_rate": 4.304920406691641e-05, "loss": 2.9338, "step": 308800 }, { "epoch": 2.7274413182853814, "grad_norm": 3.814450979232788, "learning_rate": 4.303544308600601e-05, "loss": 2.8051, "step": 308850 }, { "epoch": 2.7278828661756656, "grad_norm": 8.619109153747559, "learning_rate": 4.3021682643048554e-05, "loss": 2.9309, "step": 308900 }, { "epoch": 2.7283244140659497, "grad_norm": 2.3417961597442627, "learning_rate": 4.300792273910693e-05, "loss": 2.8451, "step": 308950 }, { "epoch": 2.728765961956234, "grad_norm": 2.04764723777771, "learning_rate": 4.299416337524401e-05, "loss": 3.1924, "step": 309000 }, { "epoch": 2.728765961956234, "eval_asr_loss": 0.9266969426480156, "eval_loss": 2.7942352294921875, "eval_runtime": 25.17, "eval_samples_per_second": 30.512, "eval_steps_per_second": 7.628, "eval_tts_loss": 6.011492230287753, "step": 309000 }, { "epoch": 2.729207509846518, "grad_norm": 4.2265305519104, "learning_rate": 4.298040455252255e-05, "loss": 3.1419, "step": 309050 }, { "epoch": 2.729649057736802, "grad_norm": 2.7739553451538086, "learning_rate": 4.29666462720053e-05, "loss": 2.8818, "step": 309100 }, { "epoch": 2.7300906056270864, "grad_norm": 1.562587857246399, "learning_rate": 4.295288853475499e-05, "loss": 3.2814, "step": 309150 }, { "epoch": 2.7305321535173706, "grad_norm": 5.3373613357543945, "learning_rate": 4.2939131341834296e-05, "loss": 3.0714, "step": 309200 }, { "epoch": 2.7309737014076547, "grad_norm": 3.2221591472625732, "learning_rate": 4.292537469430583e-05, "loss": 3.2962, "step": 309250 }, { "epoch": 2.731415249297939, "grad_norm": 4.009273052215576, "learning_rate": 4.291161859323216e-05, "loss": 3.081, "step": 309300 }, { "epoch": 2.731856797188223, "grad_norm": 5.369919300079346, "learning_rate": 4.289786303967587e-05, "loss": 3.2099, "step": 309350 }, { "epoch": 2.732298345078507, "grad_norm": 0.5758696794509888, "learning_rate": 4.288410803469941e-05, "loss": 3.015, "step": 309400 }, { "epoch": 2.7327398929687914, "grad_norm": 4.705319404602051, "learning_rate": 4.287035357936528e-05, "loss": 3.3965, "step": 309450 }, { "epoch": 2.7331814408590756, "grad_norm": 6.1709136962890625, "learning_rate": 4.2856599674735885e-05, "loss": 3.0491, "step": 309500 }, { "epoch": 2.7336229887493597, "grad_norm": 3.3086330890655518, "learning_rate": 4.284284632187358e-05, "loss": 3.1388, "step": 309550 }, { "epoch": 2.734064536639644, "grad_norm": 1.8060864210128784, "learning_rate": 4.28290935218407e-05, "loss": 3.4695, "step": 309600 }, { "epoch": 2.734506084529928, "grad_norm": 4.504628658294678, "learning_rate": 4.2815341275699536e-05, "loss": 3.0102, "step": 309650 }, { "epoch": 2.734947632420212, "grad_norm": 3.2247838973999023, "learning_rate": 4.280158958451236e-05, "loss": 2.9995, "step": 309700 }, { "epoch": 2.7353891803104964, "grad_norm": 3.4843783378601074, "learning_rate": 4.2787838449341324e-05, "loss": 2.7667, "step": 309750 }, { "epoch": 2.7358307282007805, "grad_norm": 5.351123809814453, "learning_rate": 4.2774087871248615e-05, "loss": 3.1371, "step": 309800 }, { "epoch": 2.7362722760910647, "grad_norm": 2.0368804931640625, "learning_rate": 4.2760337851296355e-05, "loss": 3.2204, "step": 309850 }, { "epoch": 2.736713823981349, "grad_norm": 1.7166926860809326, "learning_rate": 4.2746588390546586e-05, "loss": 3.0885, "step": 309900 }, { "epoch": 2.737155371871633, "grad_norm": 3.1836678981781006, "learning_rate": 4.273283949006136e-05, "loss": 2.9951, "step": 309950 }, { "epoch": 2.737596919761917, "grad_norm": 1.8265570402145386, "learning_rate": 4.271909115090268e-05, "loss": 3.0948, "step": 310000 }, { "epoch": 2.738038467652202, "grad_norm": 2.9311413764953613, "learning_rate": 4.270561832414957e-05, "loss": 3.311, "step": 310050 }, { "epoch": 2.7384800155424855, "grad_norm": 1.2335697412490845, "learning_rate": 4.269187109955031e-05, "loss": 3.0025, "step": 310100 }, { "epoch": 2.73892156343277, "grad_norm": 6.738205909729004, "learning_rate": 4.267812443944203e-05, "loss": 2.8084, "step": 310150 }, { "epoch": 2.739363111323054, "grad_norm": 3.7128005027770996, "learning_rate": 4.266437834488654e-05, "loss": 3.1756, "step": 310200 }, { "epoch": 2.7398046592133385, "grad_norm": 2.251319408416748, "learning_rate": 4.265063281694563e-05, "loss": 2.887, "step": 310250 }, { "epoch": 2.740246207103622, "grad_norm": 2.9345924854278564, "learning_rate": 4.263716275031621e-05, "loss": 2.6722, "step": 310300 }, { "epoch": 2.740687754993907, "grad_norm": 5.04674768447876, "learning_rate": 4.2623418347404395e-05, "loss": 3.4844, "step": 310350 }, { "epoch": 2.7411293028841905, "grad_norm": 4.742996692657471, "learning_rate": 4.260967451427095e-05, "loss": 3.2284, "step": 310400 }, { "epoch": 2.741570850774475, "grad_norm": 1.3208345174789429, "learning_rate": 4.259593125197751e-05, "loss": 2.7788, "step": 310450 }, { "epoch": 2.7420123986647593, "grad_norm": 1.5065076351165771, "learning_rate": 4.258218856158558e-05, "loss": 3.05, "step": 310500 }, { "epoch": 2.7424539465550435, "grad_norm": 1.318599820137024, "learning_rate": 4.256844644415668e-05, "loss": 3.0387, "step": 310550 }, { "epoch": 2.7428954944453277, "grad_norm": 3.417881727218628, "learning_rate": 4.255470490075228e-05, "loss": 2.7465, "step": 310600 }, { "epoch": 2.743337042335612, "grad_norm": 3.936246395111084, "learning_rate": 4.25409639324338e-05, "loss": 3.2943, "step": 310650 }, { "epoch": 2.743778590225896, "grad_norm": 1.4238015413284302, "learning_rate": 4.252722354026259e-05, "loss": 2.8464, "step": 310700 }, { "epoch": 2.74422013811618, "grad_norm": 5.093626499176025, "learning_rate": 4.2513483725299993e-05, "loss": 3.3102, "step": 310750 }, { "epoch": 2.7446616860064643, "grad_norm": 0.6065890789031982, "learning_rate": 4.249974448860732e-05, "loss": 2.8962, "step": 310800 }, { "epoch": 2.7451032338967485, "grad_norm": 4.990166664123535, "learning_rate": 4.2486005831245764e-05, "loss": 3.1419, "step": 310850 }, { "epoch": 2.7455447817870327, "grad_norm": 4.245227813720703, "learning_rate": 4.247226775427654e-05, "loss": 2.957, "step": 310900 }, { "epoch": 2.745986329677317, "grad_norm": 2.3839271068573, "learning_rate": 4.245853025876081e-05, "loss": 3.082, "step": 310950 }, { "epoch": 2.746427877567601, "grad_norm": 3.0109026432037354, "learning_rate": 4.244479334575965e-05, "loss": 3.0674, "step": 311000 }, { "epoch": 2.746869425457885, "grad_norm": 2.9104864597320557, "learning_rate": 4.243105701633415e-05, "loss": 3.4592, "step": 311050 }, { "epoch": 2.7473109733481693, "grad_norm": 2.596036195755005, "learning_rate": 4.24173212715453e-05, "loss": 3.126, "step": 311100 }, { "epoch": 2.7477525212384535, "grad_norm": 3.820054054260254, "learning_rate": 4.240358611245411e-05, "loss": 3.2314, "step": 311150 }, { "epoch": 2.7481940691287376, "grad_norm": 4.441590309143066, "learning_rate": 4.238985154012145e-05, "loss": 2.6866, "step": 311200 }, { "epoch": 2.748635617019022, "grad_norm": 3.949794292449951, "learning_rate": 4.237611755560824e-05, "loss": 3.1919, "step": 311250 }, { "epoch": 2.749077164909306, "grad_norm": 4.124939441680908, "learning_rate": 4.236238415997531e-05, "loss": 3.2867, "step": 311300 }, { "epoch": 2.74951871279959, "grad_norm": 3.622293472290039, "learning_rate": 4.234865135428341e-05, "loss": 3.2787, "step": 311350 }, { "epoch": 2.7499602606898743, "grad_norm": 2.0214648246765137, "learning_rate": 4.233491913959333e-05, "loss": 3.155, "step": 311400 }, { "epoch": 2.7504018085801585, "grad_norm": 4.4073638916015625, "learning_rate": 4.232118751696576e-05, "loss": 3.1123, "step": 311450 }, { "epoch": 2.7508433564704426, "grad_norm": 1.7340322732925415, "learning_rate": 4.2307456487461316e-05, "loss": 3.2038, "step": 311500 }, { "epoch": 2.751284904360727, "grad_norm": 3.6743979454040527, "learning_rate": 4.229372605214063e-05, "loss": 2.806, "step": 311550 }, { "epoch": 2.751726452251011, "grad_norm": 4.392818450927734, "learning_rate": 4.227999621206428e-05, "loss": 2.9485, "step": 311600 }, { "epoch": 2.752168000141295, "grad_norm": 1.8125756978988647, "learning_rate": 4.2266266968292747e-05, "loss": 2.8847, "step": 311650 }, { "epoch": 2.7526095480315798, "grad_norm": 3.7686100006103516, "learning_rate": 4.225253832188649e-05, "loss": 3.0886, "step": 311700 }, { "epoch": 2.7530510959218635, "grad_norm": 5.010977268218994, "learning_rate": 4.2238810273905966e-05, "loss": 3.2892, "step": 311750 }, { "epoch": 2.753492643812148, "grad_norm": 2.854464054107666, "learning_rate": 4.222508282541155e-05, "loss": 3.1576, "step": 311800 }, { "epoch": 2.753934191702432, "grad_norm": 2.2028608322143555, "learning_rate": 4.221135597746354e-05, "loss": 2.6103, "step": 311850 }, { "epoch": 2.7543757395927164, "grad_norm": 2.741579055786133, "learning_rate": 4.219762973112224e-05, "loss": 2.6996, "step": 311900 }, { "epoch": 2.754817287483, "grad_norm": 2.9442291259765625, "learning_rate": 4.218390408744789e-05, "loss": 3.3033, "step": 311950 }, { "epoch": 2.7552588353732848, "grad_norm": 2.51954984664917, "learning_rate": 4.2170179047500667e-05, "loss": 2.8005, "step": 312000 }, { "epoch": 2.7552588353732848, "eval_asr_loss": 0.9302835927303176, "eval_loss": 2.7875888347625732, "eval_runtime": 20.678, "eval_samples_per_second": 37.141, "eval_steps_per_second": 9.285, "eval_tts_loss": 6.030184715932079, "step": 312000 }, { "epoch": 2.755700383263569, "grad_norm": 2.9619483947753906, "learning_rate": 4.21564546123407e-05, "loss": 2.8156, "step": 312050 }, { "epoch": 2.756141931153853, "grad_norm": 5.508452415466309, "learning_rate": 4.214273078302813e-05, "loss": 3.1573, "step": 312100 }, { "epoch": 2.7565834790441373, "grad_norm": 3.064406156539917, "learning_rate": 4.2129007560622954e-05, "loss": 3.2935, "step": 312150 }, { "epoch": 2.7570250269344214, "grad_norm": 3.301003932952881, "learning_rate": 4.211528494618521e-05, "loss": 2.7608, "step": 312200 }, { "epoch": 2.7574665748247056, "grad_norm": 3.7378509044647217, "learning_rate": 4.210156294077484e-05, "loss": 3.3628, "step": 312250 }, { "epoch": 2.7579081227149898, "grad_norm": 2.6415252685546875, "learning_rate": 4.208784154545177e-05, "loss": 3.4996, "step": 312300 }, { "epoch": 2.758349670605274, "grad_norm": 1.9715427160263062, "learning_rate": 4.2074120761275834e-05, "loss": 2.6045, "step": 312350 }, { "epoch": 2.758791218495558, "grad_norm": 2.823087453842163, "learning_rate": 4.206040058930686e-05, "loss": 2.764, "step": 312400 }, { "epoch": 2.7592327663858423, "grad_norm": 2.108388900756836, "learning_rate": 4.204668103060463e-05, "loss": 3.2358, "step": 312450 }, { "epoch": 2.7596743142761264, "grad_norm": 1.4445101022720337, "learning_rate": 4.2032962086228826e-05, "loss": 2.6494, "step": 312500 }, { "epoch": 2.7601158621664106, "grad_norm": 3.8792903423309326, "learning_rate": 4.201924375723916e-05, "loss": 2.693, "step": 312550 }, { "epoch": 2.7605574100566947, "grad_norm": 3.0311026573181152, "learning_rate": 4.200552604469523e-05, "loss": 2.4397, "step": 312600 }, { "epoch": 2.760998957946979, "grad_norm": 3.7416670322418213, "learning_rate": 4.199180894965662e-05, "loss": 3.1563, "step": 312650 }, { "epoch": 2.761440505837263, "grad_norm": 2.36936092376709, "learning_rate": 4.197809247318286e-05, "loss": 2.9941, "step": 312700 }, { "epoch": 2.7618820537275472, "grad_norm": 5.225512981414795, "learning_rate": 4.196437661633345e-05, "loss": 2.8879, "step": 312750 }, { "epoch": 2.7623236016178314, "grad_norm": 20.443262100219727, "learning_rate": 4.195066138016779e-05, "loss": 2.9537, "step": 312800 }, { "epoch": 2.7627651495081156, "grad_norm": 3.0223169326782227, "learning_rate": 4.193694676574528e-05, "loss": 3.2606, "step": 312850 }, { "epoch": 2.7632066973983997, "grad_norm": 2.492983102798462, "learning_rate": 4.192323277412526e-05, "loss": 3.0082, "step": 312900 }, { "epoch": 2.763648245288684, "grad_norm": 4.157458305358887, "learning_rate": 4.190951940636704e-05, "loss": 2.7522, "step": 312950 }, { "epoch": 2.764089793178968, "grad_norm": 4.266920566558838, "learning_rate": 4.189580666352982e-05, "loss": 2.8447, "step": 313000 }, { "epoch": 2.7645313410692522, "grad_norm": 3.0359091758728027, "learning_rate": 4.1882094546672815e-05, "loss": 2.8992, "step": 313050 }, { "epoch": 2.7649728889595364, "grad_norm": 1.1496204137802124, "learning_rate": 4.186838305685518e-05, "loss": 2.6978, "step": 313100 }, { "epoch": 2.7654144368498206, "grad_norm": 3.1911778450012207, "learning_rate": 4.185467219513598e-05, "loss": 3.3697, "step": 313150 }, { "epoch": 2.7658559847401047, "grad_norm": 3.6276378631591797, "learning_rate": 4.18409619625743e-05, "loss": 3.0121, "step": 313200 }, { "epoch": 2.7662975326303894, "grad_norm": 4.566330909729004, "learning_rate": 4.182725236022913e-05, "loss": 2.7814, "step": 313250 }, { "epoch": 2.766739080520673, "grad_norm": 3.1478590965270996, "learning_rate": 4.181354338915939e-05, "loss": 2.5114, "step": 313300 }, { "epoch": 2.7671806284109577, "grad_norm": 2.9412221908569336, "learning_rate": 4.179983505042401e-05, "loss": 2.7924, "step": 313350 }, { "epoch": 2.7676221763012414, "grad_norm": 2.378176689147949, "learning_rate": 4.1786127345081856e-05, "loss": 2.9253, "step": 313400 }, { "epoch": 2.768063724191526, "grad_norm": 6.177469730377197, "learning_rate": 4.177242027419168e-05, "loss": 2.7201, "step": 313450 }, { "epoch": 2.7685052720818097, "grad_norm": 6.702813625335693, "learning_rate": 4.175871383881228e-05, "loss": 3.0099, "step": 313500 }, { "epoch": 2.7689468199720944, "grad_norm": 5.151427745819092, "learning_rate": 4.174500804000236e-05, "loss": 3.2438, "step": 313550 }, { "epoch": 2.7693883678623785, "grad_norm": 1.4618408679962158, "learning_rate": 4.173130287882058e-05, "loss": 2.7923, "step": 313600 }, { "epoch": 2.7698299157526627, "grad_norm": 3.6876718997955322, "learning_rate": 4.1717598356325514e-05, "loss": 2.993, "step": 313650 }, { "epoch": 2.770271463642947, "grad_norm": 4.886559963226318, "learning_rate": 4.1703894473575744e-05, "loss": 3.0842, "step": 313700 }, { "epoch": 2.770713011533231, "grad_norm": 4.38385009765625, "learning_rate": 4.169019123162979e-05, "loss": 3.5593, "step": 313750 }, { "epoch": 2.771154559423515, "grad_norm": 4.787524700164795, "learning_rate": 4.16764886315461e-05, "loss": 2.6507, "step": 313800 }, { "epoch": 2.7715961073137994, "grad_norm": 4.827417850494385, "learning_rate": 4.166278667438307e-05, "loss": 3.5405, "step": 313850 }, { "epoch": 2.7720376552040835, "grad_norm": 2.3966052532196045, "learning_rate": 4.164908536119908e-05, "loss": 3.4458, "step": 313900 }, { "epoch": 2.7724792030943677, "grad_norm": 2.8238346576690674, "learning_rate": 4.163538469305243e-05, "loss": 2.8664, "step": 313950 }, { "epoch": 2.772920750984652, "grad_norm": 4.871623516082764, "learning_rate": 4.162168467100138e-05, "loss": 3.0878, "step": 314000 }, { "epoch": 2.773362298874936, "grad_norm": 4.72355318069458, "learning_rate": 4.160798529610413e-05, "loss": 3.0611, "step": 314050 }, { "epoch": 2.77380384676522, "grad_norm": 6.000672817230225, "learning_rate": 4.1594286569418886e-05, "loss": 3.4166, "step": 314100 }, { "epoch": 2.7742453946555043, "grad_norm": 1.8222882747650146, "learning_rate": 4.15805884920037e-05, "loss": 2.8397, "step": 314150 }, { "epoch": 2.7746869425457885, "grad_norm": 4.4614152908325195, "learning_rate": 4.1566891064916665e-05, "loss": 3.1203, "step": 314200 }, { "epoch": 2.7751284904360727, "grad_norm": 6.728885650634766, "learning_rate": 4.15531942892158e-05, "loss": 2.9984, "step": 314250 }, { "epoch": 2.775570038326357, "grad_norm": 3.8685688972473145, "learning_rate": 4.153949816595902e-05, "loss": 3.6128, "step": 314300 }, { "epoch": 2.776011586216641, "grad_norm": 4.419139385223389, "learning_rate": 4.1525802696204284e-05, "loss": 2.9406, "step": 314350 }, { "epoch": 2.776453134106925, "grad_norm": 3.9728140830993652, "learning_rate": 4.1512107881009435e-05, "loss": 3.3257, "step": 314400 }, { "epoch": 2.7768946819972093, "grad_norm": 5.092426300048828, "learning_rate": 4.1498413721432264e-05, "loss": 2.9965, "step": 314450 }, { "epoch": 2.7773362298874935, "grad_norm": 2.2077434062957764, "learning_rate": 4.148499408214632e-05, "loss": 3.0434, "step": 314500 }, { "epoch": 2.7777777777777777, "grad_norm": 1.8570573329925537, "learning_rate": 4.147130122381272e-05, "loss": 3.0664, "step": 314550 }, { "epoch": 2.778219325668062, "grad_norm": 2.986189126968384, "learning_rate": 4.145760902424878e-05, "loss": 3.0509, "step": 314600 }, { "epoch": 2.778660873558346, "grad_norm": 5.644431114196777, "learning_rate": 4.144391748451213e-05, "loss": 2.9006, "step": 314650 }, { "epoch": 2.77910242144863, "grad_norm": 7.415584564208984, "learning_rate": 4.143022660566028e-05, "loss": 2.9808, "step": 314700 }, { "epoch": 2.7795439693389143, "grad_norm": 5.017648696899414, "learning_rate": 4.1416536388750756e-05, "loss": 2.9375, "step": 314750 }, { "epoch": 2.779985517229199, "grad_norm": 2.2349796295166016, "learning_rate": 4.140284683484103e-05, "loss": 2.8429, "step": 314800 }, { "epoch": 2.7804270651194827, "grad_norm": 5.535529613494873, "learning_rate": 4.138915794498847e-05, "loss": 2.655, "step": 314850 }, { "epoch": 2.7808686130097673, "grad_norm": 3.9447689056396484, "learning_rate": 4.1375469720250424e-05, "loss": 3.5727, "step": 314900 }, { "epoch": 2.781310160900051, "grad_norm": 5.7717790603637695, "learning_rate": 4.1361782161684236e-05, "loss": 3.1705, "step": 314950 }, { "epoch": 2.7817517087903356, "grad_norm": 1.1017372608184814, "learning_rate": 4.1348095270347104e-05, "loss": 3.3717, "step": 315000 }, { "epoch": 2.7817517087903356, "eval_asr_loss": 0.925218278198667, "eval_loss": 2.7819833755493164, "eval_runtime": 20.9306, "eval_samples_per_second": 36.693, "eval_steps_per_second": 9.173, "eval_tts_loss": 6.064099530477656, "step": 315000 }, { "epoch": 2.7821932566806193, "grad_norm": 3.2681639194488525, "learning_rate": 4.133440904729626e-05, "loss": 3.1555, "step": 315050 }, { "epoch": 2.782634804570904, "grad_norm": 5.63743257522583, "learning_rate": 4.132072349358882e-05, "loss": 3.2158, "step": 315100 }, { "epoch": 2.7830763524611877, "grad_norm": 4.660617828369141, "learning_rate": 4.130703861028191e-05, "loss": 2.8136, "step": 315150 }, { "epoch": 2.7835179003514723, "grad_norm": 5.972463130950928, "learning_rate": 4.1293354398432536e-05, "loss": 3.0945, "step": 315200 }, { "epoch": 2.7839594482417565, "grad_norm": 3.157576322555542, "learning_rate": 4.127967085909772e-05, "loss": 3.0617, "step": 315250 }, { "epoch": 2.7844009961320406, "grad_norm": 1.8076050281524658, "learning_rate": 4.126598799333439e-05, "loss": 3.0057, "step": 315300 }, { "epoch": 2.784842544022325, "grad_norm": 5.3761444091796875, "learning_rate": 4.1252305802199404e-05, "loss": 2.9449, "step": 315350 }, { "epoch": 2.785284091912609, "grad_norm": 2.0830442905426025, "learning_rate": 4.123862428674963e-05, "loss": 2.9002, "step": 315400 }, { "epoch": 2.785725639802893, "grad_norm": 7.913502216339111, "learning_rate": 4.122494344804183e-05, "loss": 2.5864, "step": 315450 }, { "epoch": 2.7861671876931773, "grad_norm": 1.5461803674697876, "learning_rate": 4.1211263287132746e-05, "loss": 2.9029, "step": 315500 }, { "epoch": 2.7866087355834614, "grad_norm": 6.009243965148926, "learning_rate": 4.119758380507903e-05, "loss": 3.402, "step": 315550 }, { "epoch": 2.7870502834737456, "grad_norm": 7.494958400726318, "learning_rate": 4.1183905002937345e-05, "loss": 3.365, "step": 315600 }, { "epoch": 2.78749183136403, "grad_norm": 3.803821325302124, "learning_rate": 4.117022688176423e-05, "loss": 2.8037, "step": 315650 }, { "epoch": 2.787933379254314, "grad_norm": 3.6559839248657227, "learning_rate": 4.1156549442616205e-05, "loss": 2.58, "step": 315700 }, { "epoch": 2.788374927144598, "grad_norm": 3.193549871444702, "learning_rate": 4.114287268654975e-05, "loss": 2.6834, "step": 315750 }, { "epoch": 2.7888164750348823, "grad_norm": 3.7126169204711914, "learning_rate": 4.112919661462128e-05, "loss": 3.0654, "step": 315800 }, { "epoch": 2.7892580229251664, "grad_norm": 6.371332168579102, "learning_rate": 4.111552122788715e-05, "loss": 3.0955, "step": 315850 }, { "epoch": 2.7896995708154506, "grad_norm": 4.429771900177002, "learning_rate": 4.1101846527403654e-05, "loss": 3.1769, "step": 315900 }, { "epoch": 2.790141118705735, "grad_norm": 4.695055961608887, "learning_rate": 4.108817251422708e-05, "loss": 2.6213, "step": 315950 }, { "epoch": 2.790582666596019, "grad_norm": 2.492372512817383, "learning_rate": 4.107449918941359e-05, "loss": 3.0354, "step": 316000 }, { "epoch": 2.791024214486303, "grad_norm": 4.037844657897949, "learning_rate": 4.106082655401936e-05, "loss": 2.9783, "step": 316050 }, { "epoch": 2.7914657623765873, "grad_norm": 4.151850700378418, "learning_rate": 4.1047154609100494e-05, "loss": 3.2751, "step": 316100 }, { "epoch": 2.7919073102668714, "grad_norm": 2.153799295425415, "learning_rate": 4.103375677399689e-05, "loss": 2.6509, "step": 316150 }, { "epoch": 2.7923488581571556, "grad_norm": 4.651935577392578, "learning_rate": 4.102008619933469e-05, "loss": 2.857, "step": 316200 }, { "epoch": 2.7927904060474398, "grad_norm": 0.7709072828292847, "learning_rate": 4.100641631829466e-05, "loss": 2.9522, "step": 316250 }, { "epoch": 2.793231953937724, "grad_norm": 3.0834312438964844, "learning_rate": 4.099274713193272e-05, "loss": 2.4995, "step": 316300 }, { "epoch": 2.793673501828008, "grad_norm": 1.4540549516677856, "learning_rate": 4.097907864130469e-05, "loss": 3.1328, "step": 316350 }, { "epoch": 2.7941150497182923, "grad_norm": 3.7943832874298096, "learning_rate": 4.096541084746632e-05, "loss": 2.9205, "step": 316400 }, { "epoch": 2.794556597608577, "grad_norm": 2.842139959335327, "learning_rate": 4.0951743751473354e-05, "loss": 3.2057, "step": 316450 }, { "epoch": 2.7949981454988606, "grad_norm": 2.0192320346832275, "learning_rate": 4.093807735438148e-05, "loss": 3.1217, "step": 316500 }, { "epoch": 2.795439693389145, "grad_norm": 2.1340339183807373, "learning_rate": 4.0924411657246265e-05, "loss": 2.9056, "step": 316550 }, { "epoch": 2.795881241279429, "grad_norm": 2.769340991973877, "learning_rate": 4.091074666112329e-05, "loss": 2.5583, "step": 316600 }, { "epoch": 2.7963227891697136, "grad_norm": 6.138609886169434, "learning_rate": 4.089708236706806e-05, "loss": 3.2883, "step": 316650 }, { "epoch": 2.7967643370599973, "grad_norm": 2.3817968368530273, "learning_rate": 4.0883418776136044e-05, "loss": 2.9734, "step": 316700 }, { "epoch": 2.797205884950282, "grad_norm": 1.2322462797164917, "learning_rate": 4.086975588938261e-05, "loss": 3.0306, "step": 316750 }, { "epoch": 2.797647432840566, "grad_norm": 4.486616134643555, "learning_rate": 4.085609370786311e-05, "loss": 2.6642, "step": 316800 }, { "epoch": 2.79808898073085, "grad_norm": 1.7845451831817627, "learning_rate": 4.0842432232632843e-05, "loss": 3.1168, "step": 316850 }, { "epoch": 2.7985305286211344, "grad_norm": 2.697742462158203, "learning_rate": 4.082877146474702e-05, "loss": 2.8667, "step": 316900 }, { "epoch": 2.7989720765114185, "grad_norm": 10.212408065795898, "learning_rate": 4.081511140526084e-05, "loss": 3.2056, "step": 316950 }, { "epoch": 2.7994136244017027, "grad_norm": 2.1721062660217285, "learning_rate": 4.080145205522942e-05, "loss": 3.2108, "step": 317000 }, { "epoch": 2.799855172291987, "grad_norm": 5.833930015563965, "learning_rate": 4.0787793415707816e-05, "loss": 3.2628, "step": 317050 }, { "epoch": 2.800296720182271, "grad_norm": 4.345120906829834, "learning_rate": 4.077413548775106e-05, "loss": 3.1725, "step": 317100 }, { "epoch": 2.800738268072555, "grad_norm": 4.038512229919434, "learning_rate": 4.0760478272414114e-05, "loss": 3.0879, "step": 317150 }, { "epoch": 2.8011798159628394, "grad_norm": 3.689167022705078, "learning_rate": 4.0746821770751855e-05, "loss": 3.4723, "step": 317200 }, { "epoch": 2.8016213638531235, "grad_norm": 2.8231451511383057, "learning_rate": 4.073316598381915e-05, "loss": 3.0737, "step": 317250 }, { "epoch": 2.8020629117434077, "grad_norm": 4.710754871368408, "learning_rate": 4.07195109126708e-05, "loss": 2.9314, "step": 317300 }, { "epoch": 2.802504459633692, "grad_norm": 5.484740257263184, "learning_rate": 4.070585655836154e-05, "loss": 3.0785, "step": 317350 }, { "epoch": 2.802946007523976, "grad_norm": 1.6158608198165894, "learning_rate": 4.069220292194604e-05, "loss": 3.2493, "step": 317400 }, { "epoch": 2.80338755541426, "grad_norm": 2.426886796951294, "learning_rate": 4.067855000447893e-05, "loss": 3.3543, "step": 317450 }, { "epoch": 2.8038291033045444, "grad_norm": 1.7401580810546875, "learning_rate": 4.066489780701481e-05, "loss": 2.7121, "step": 317500 }, { "epoch": 2.8042706511948285, "grad_norm": 9.872039794921875, "learning_rate": 4.0651246330608175e-05, "loss": 3.2737, "step": 317550 }, { "epoch": 2.8047121990851127, "grad_norm": 4.375695705413818, "learning_rate": 4.063759557631346e-05, "loss": 2.7666, "step": 317600 }, { "epoch": 2.805153746975397, "grad_norm": 1.3890061378479004, "learning_rate": 4.0623945545185125e-05, "loss": 2.7461, "step": 317650 }, { "epoch": 2.805595294865681, "grad_norm": 1.3317537307739258, "learning_rate": 4.061029623827748e-05, "loss": 3.0196, "step": 317700 }, { "epoch": 2.806036842755965, "grad_norm": 2.5546116828918457, "learning_rate": 4.059664765664481e-05, "loss": 3.298, "step": 317750 }, { "epoch": 2.8064783906462494, "grad_norm": 4.30869722366333, "learning_rate": 4.05829998013414e-05, "loss": 2.7377, "step": 317800 }, { "epoch": 2.8069199385365335, "grad_norm": 7.532159328460693, "learning_rate": 4.0569352673421375e-05, "loss": 3.0382, "step": 317850 }, { "epoch": 2.8073614864268177, "grad_norm": 3.732574224472046, "learning_rate": 4.05557062739389e-05, "loss": 3.3808, "step": 317900 }, { "epoch": 2.807803034317102, "grad_norm": 3.648179531097412, "learning_rate": 4.054206060394802e-05, "loss": 3.0121, "step": 317950 }, { "epoch": 2.8082445822073865, "grad_norm": 1.463517665863037, "learning_rate": 4.052841566450277e-05, "loss": 3.1239, "step": 318000 }, { "epoch": 2.8082445822073865, "eval_asr_loss": 0.927193419469223, "eval_loss": 2.7825775146484375, "eval_runtime": 20.8586, "eval_samples_per_second": 36.819, "eval_steps_per_second": 9.205, "eval_tts_loss": 6.056725428426568, "step": 318000 }, { "epoch": 2.80868613009767, "grad_norm": 2.8009445667266846, "learning_rate": 4.0514771456657086e-05, "loss": 2.9357, "step": 318050 }, { "epoch": 2.809127677987955, "grad_norm": 3.9641647338867188, "learning_rate": 4.0501127981464874e-05, "loss": 3.1602, "step": 318100 }, { "epoch": 2.8095692258782385, "grad_norm": 3.9536585807800293, "learning_rate": 4.048748523997999e-05, "loss": 3.4158, "step": 318150 }, { "epoch": 2.810010773768523, "grad_norm": 3.572383165359497, "learning_rate": 4.0473843233256194e-05, "loss": 3.1789, "step": 318200 }, { "epoch": 2.810452321658807, "grad_norm": 1.4940310716629028, "learning_rate": 4.046020196234723e-05, "loss": 3.0814, "step": 318250 }, { "epoch": 2.8108938695490915, "grad_norm": 3.9625425338745117, "learning_rate": 4.044656142830677e-05, "loss": 3.1322, "step": 318300 }, { "epoch": 2.811335417439375, "grad_norm": 4.962860107421875, "learning_rate": 4.0432921632188435e-05, "loss": 2.8939, "step": 318350 }, { "epoch": 2.81177696532966, "grad_norm": 2.1629369258880615, "learning_rate": 4.041928257504577e-05, "loss": 3.0403, "step": 318400 }, { "epoch": 2.812218513219944, "grad_norm": 3.067422389984131, "learning_rate": 4.0405644257932285e-05, "loss": 3.3751, "step": 318450 }, { "epoch": 2.812660061110228, "grad_norm": 5.740373611450195, "learning_rate": 4.039200668190143e-05, "loss": 3.1186, "step": 318500 }, { "epoch": 2.8131016090005123, "grad_norm": 1.9838188886642456, "learning_rate": 4.0378369848006576e-05, "loss": 3.0546, "step": 318550 }, { "epoch": 2.8135431568907965, "grad_norm": 1.6022948026657104, "learning_rate": 4.036473375730106e-05, "loss": 3.2455, "step": 318600 }, { "epoch": 2.8139847047810806, "grad_norm": 4.113194942474365, "learning_rate": 4.0351098410838174e-05, "loss": 2.6125, "step": 318650 }, { "epoch": 2.814426252671365, "grad_norm": 3.246614694595337, "learning_rate": 4.0337736494383746e-05, "loss": 3.318, "step": 318700 }, { "epoch": 2.814867800561649, "grad_norm": 3.3540921211242676, "learning_rate": 4.032410262462835e-05, "loss": 2.8244, "step": 318750 }, { "epoch": 2.815309348451933, "grad_norm": 2.6812150478363037, "learning_rate": 4.031046950225399e-05, "loss": 3.4171, "step": 318800 }, { "epoch": 2.8157508963422173, "grad_norm": 3.276190996170044, "learning_rate": 4.029683712831368e-05, "loss": 2.8752, "step": 318850 }, { "epoch": 2.8161924442325015, "grad_norm": 4.056405544281006, "learning_rate": 4.028320550386045e-05, "loss": 3.2407, "step": 318900 }, { "epoch": 2.8166339921227856, "grad_norm": 0.6795021891593933, "learning_rate": 4.026957462994718e-05, "loss": 3.0511, "step": 318950 }, { "epoch": 2.81707554001307, "grad_norm": 4.112836837768555, "learning_rate": 4.0255944507626766e-05, "loss": 3.1258, "step": 319000 }, { "epoch": 2.817517087903354, "grad_norm": 0.7870320081710815, "learning_rate": 4.0242315137952014e-05, "loss": 3.2754, "step": 319050 }, { "epoch": 2.817958635793638, "grad_norm": 5.2137346267700195, "learning_rate": 4.022868652197566e-05, "loss": 2.9985, "step": 319100 }, { "epoch": 2.8184001836839223, "grad_norm": 2.157069444656372, "learning_rate": 4.0215058660750425e-05, "loss": 3.1488, "step": 319150 }, { "epoch": 2.8188417315742065, "grad_norm": 2.9600942134857178, "learning_rate": 4.020143155532895e-05, "loss": 3.4287, "step": 319200 }, { "epoch": 2.8192832794644906, "grad_norm": 2.1373496055603027, "learning_rate": 4.018780520676377e-05, "loss": 2.8177, "step": 319250 }, { "epoch": 2.819724827354775, "grad_norm": 5.419004917144775, "learning_rate": 4.0174179616107434e-05, "loss": 3.1778, "step": 319300 }, { "epoch": 2.820166375245059, "grad_norm": 4.796324253082275, "learning_rate": 4.016055478441242e-05, "loss": 2.741, "step": 319350 }, { "epoch": 2.820607923135343, "grad_norm": 2.406672716140747, "learning_rate": 4.014693071273111e-05, "loss": 2.8538, "step": 319400 }, { "epoch": 2.8210494710256273, "grad_norm": 6.078481674194336, "learning_rate": 4.013330740211584e-05, "loss": 2.8954, "step": 319450 }, { "epoch": 2.8214910189159115, "grad_norm": 2.8117499351501465, "learning_rate": 4.01196848536189e-05, "loss": 3.0182, "step": 319500 }, { "epoch": 2.8219325668061956, "grad_norm": 4.206378936767578, "learning_rate": 4.0106063068292544e-05, "loss": 2.8411, "step": 319550 }, { "epoch": 2.82237411469648, "grad_norm": 2.0299112796783447, "learning_rate": 4.009244204718891e-05, "loss": 2.5561, "step": 319600 }, { "epoch": 2.8228156625867644, "grad_norm": 2.752762794494629, "learning_rate": 4.00788217913601e-05, "loss": 3.0031, "step": 319650 }, { "epoch": 2.823257210477048, "grad_norm": 2.9362094402313232, "learning_rate": 4.0065202301858205e-05, "loss": 3.1225, "step": 319700 }, { "epoch": 2.8236987583673328, "grad_norm": 1.6035170555114746, "learning_rate": 4.005158357973516e-05, "loss": 3.3643, "step": 319750 }, { "epoch": 2.8241403062576165, "grad_norm": 3.267713785171509, "learning_rate": 4.003796562604294e-05, "loss": 2.8981, "step": 319800 }, { "epoch": 2.824581854147901, "grad_norm": 3.3030288219451904, "learning_rate": 4.0024348441833406e-05, "loss": 3.1816, "step": 319850 }, { "epoch": 2.825023402038185, "grad_norm": 2.8155295848846436, "learning_rate": 4.0010732028158344e-05, "loss": 2.8726, "step": 319900 }, { "epoch": 2.8254649499284694, "grad_norm": 4.046159744262695, "learning_rate": 3.9997116386069534e-05, "loss": 3.0913, "step": 319950 }, { "epoch": 2.8259064978187536, "grad_norm": 2.136352300643921, "learning_rate": 3.998350151661865e-05, "loss": 3.2895, "step": 320000 }, { "epoch": 2.8263480457090377, "grad_norm": 1.6182206869125366, "learning_rate": 3.996988742085737e-05, "loss": 3.1702, "step": 320050 }, { "epoch": 2.826789593599322, "grad_norm": 2.333502769470215, "learning_rate": 3.9956274099837196e-05, "loss": 3.087, "step": 320100 }, { "epoch": 2.827231141489606, "grad_norm": 2.8702173233032227, "learning_rate": 3.99426615546097e-05, "loss": 3.0104, "step": 320150 }, { "epoch": 2.8276726893798902, "grad_norm": 6.879642486572266, "learning_rate": 3.9929049786226326e-05, "loss": 2.8528, "step": 320200 }, { "epoch": 2.8281142372701744, "grad_norm": 3.3558859825134277, "learning_rate": 3.991543879573843e-05, "loss": 2.8874, "step": 320250 }, { "epoch": 2.8285557851604586, "grad_norm": 2.128997564315796, "learning_rate": 3.990182858419738e-05, "loss": 3.3526, "step": 320300 }, { "epoch": 2.8289973330507427, "grad_norm": 1.6346467733383179, "learning_rate": 3.988821915265446e-05, "loss": 3.191, "step": 320350 }, { "epoch": 2.829438880941027, "grad_norm": 2.4018843173980713, "learning_rate": 3.987461050216085e-05, "loss": 2.8756, "step": 320400 }, { "epoch": 2.829880428831311, "grad_norm": 4.20185661315918, "learning_rate": 3.9861002633767706e-05, "loss": 3.1022, "step": 320450 }, { "epoch": 2.8303219767215952, "grad_norm": 2.1727583408355713, "learning_rate": 3.9847395548526164e-05, "loss": 2.8657, "step": 320500 }, { "epoch": 2.8307635246118794, "grad_norm": 3.341797351837158, "learning_rate": 3.9833789247487206e-05, "loss": 2.7971, "step": 320550 }, { "epoch": 2.8312050725021636, "grad_norm": 2.387946128845215, "learning_rate": 3.982018373170181e-05, "loss": 3.0576, "step": 320600 }, { "epoch": 2.8316466203924477, "grad_norm": 3.176732301712036, "learning_rate": 3.980657900222091e-05, "loss": 3.2853, "step": 320650 }, { "epoch": 2.832088168282732, "grad_norm": 1.646636962890625, "learning_rate": 3.979297506009535e-05, "loss": 3.0853, "step": 320700 }, { "epoch": 2.832529716173016, "grad_norm": 3.2949440479278564, "learning_rate": 3.9779371906375914e-05, "loss": 3.277, "step": 320750 }, { "epoch": 2.8329712640633002, "grad_norm": 6.672817707061768, "learning_rate": 3.976576954211331e-05, "loss": 2.7944, "step": 320800 }, { "epoch": 2.8334128119535844, "grad_norm": 4.965484142303467, "learning_rate": 3.975243999207958e-05, "loss": 3.2003, "step": 320850 }, { "epoch": 2.8338543598438686, "grad_norm": 2.555398941040039, "learning_rate": 3.973883919404119e-05, "loss": 3.0545, "step": 320900 }, { "epoch": 2.8342959077341527, "grad_norm": 2.3414435386657715, "learning_rate": 3.972523918859043e-05, "loss": 2.7656, "step": 320950 }, { "epoch": 2.834737455624437, "grad_norm": 3.639129638671875, "learning_rate": 3.971163997677783e-05, "loss": 3.2521, "step": 321000 }, { "epoch": 2.834737455624437, "eval_asr_loss": 0.9324067285757468, "eval_loss": 2.778122663497925, "eval_runtime": 21.0097, "eval_samples_per_second": 36.554, "eval_steps_per_second": 9.139, "eval_tts_loss": 6.037034641609145, "step": 321000 }, { "epoch": 2.835179003514721, "grad_norm": 2.4739797115325928, "learning_rate": 3.9698313520201545e-05, "loss": 3.0446, "step": 321050 }, { "epoch": 2.8356205514050052, "grad_norm": 2.6612892150878906, "learning_rate": 3.9684715882891366e-05, "loss": 3.3245, "step": 321100 }, { "epoch": 2.8360620992952894, "grad_norm": 2.8280651569366455, "learning_rate": 3.967111904234944e-05, "loss": 3.2224, "step": 321150 }, { "epoch": 2.836503647185574, "grad_norm": 3.894984006881714, "learning_rate": 3.9657522999625964e-05, "loss": 3.0696, "step": 321200 }, { "epoch": 2.8369451950758577, "grad_norm": 5.176159858703613, "learning_rate": 3.9643927755771126e-05, "loss": 2.7345, "step": 321250 }, { "epoch": 2.8373867429661424, "grad_norm": 3.401778221130371, "learning_rate": 3.963033331183508e-05, "loss": 2.5813, "step": 321300 }, { "epoch": 2.837828290856426, "grad_norm": 3.1085379123687744, "learning_rate": 3.9616739668867817e-05, "loss": 3.3394, "step": 321350 }, { "epoch": 2.8382698387467107, "grad_norm": 1.5213242769241333, "learning_rate": 3.960314682791939e-05, "loss": 2.9724, "step": 321400 }, { "epoch": 2.8387113866369944, "grad_norm": 2.1361615657806396, "learning_rate": 3.95895547900397e-05, "loss": 3.2943, "step": 321450 }, { "epoch": 2.839152934527279, "grad_norm": 3.534886360168457, "learning_rate": 3.9575963556278636e-05, "loss": 3.1879, "step": 321500 }, { "epoch": 2.8395944824175627, "grad_norm": 4.931316375732422, "learning_rate": 3.956237312768599e-05, "loss": 3.1669, "step": 321550 }, { "epoch": 2.8400360303078473, "grad_norm": 3.8561415672302246, "learning_rate": 3.9548783505311494e-05, "loss": 3.4426, "step": 321600 }, { "epoch": 2.8404775781981315, "grad_norm": 6.57034158706665, "learning_rate": 3.9535194690204876e-05, "loss": 3.0764, "step": 321650 }, { "epoch": 2.8409191260884157, "grad_norm": 5.512895107269287, "learning_rate": 3.95216066834157e-05, "loss": 3.0409, "step": 321700 }, { "epoch": 2.8413606739787, "grad_norm": 3.5843589305877686, "learning_rate": 3.950801948599358e-05, "loss": 3.0321, "step": 321750 }, { "epoch": 2.841802221868984, "grad_norm": 4.295320987701416, "learning_rate": 3.949443309898798e-05, "loss": 3.0003, "step": 321800 }, { "epoch": 2.842243769759268, "grad_norm": 5.344568729400635, "learning_rate": 3.948084752344832e-05, "loss": 2.7374, "step": 321850 }, { "epoch": 2.8426853176495523, "grad_norm": 8.716202735900879, "learning_rate": 3.9467262760423995e-05, "loss": 3.2404, "step": 321900 }, { "epoch": 2.8431268655398365, "grad_norm": 6.044868469238281, "learning_rate": 3.9453678810964324e-05, "loss": 3.2151, "step": 321950 }, { "epoch": 2.8435684134301207, "grad_norm": 3.4602737426757812, "learning_rate": 3.9440095676118526e-05, "loss": 3.4735, "step": 322000 }, { "epoch": 2.844009961320405, "grad_norm": 3.596950054168701, "learning_rate": 3.942651335693578e-05, "loss": 3.0543, "step": 322050 }, { "epoch": 2.844451509210689, "grad_norm": 1.7978414297103882, "learning_rate": 3.941293185446523e-05, "loss": 3.0847, "step": 322100 }, { "epoch": 2.844893057100973, "grad_norm": 4.95121955871582, "learning_rate": 3.9399351169755926e-05, "loss": 3.0596, "step": 322150 }, { "epoch": 2.8453346049912573, "grad_norm": 6.085773944854736, "learning_rate": 3.938577130385683e-05, "loss": 3.1712, "step": 322200 }, { "epoch": 2.8457761528815415, "grad_norm": 2.8295490741729736, "learning_rate": 3.93721922578169e-05, "loss": 3.2952, "step": 322250 }, { "epoch": 2.8462177007718257, "grad_norm": 4.943411827087402, "learning_rate": 3.9358614032685014e-05, "loss": 3.0693, "step": 322300 }, { "epoch": 2.84665924866211, "grad_norm": 5.358550548553467, "learning_rate": 3.934503662950995e-05, "loss": 3.5046, "step": 322350 }, { "epoch": 2.847100796552394, "grad_norm": 4.299525737762451, "learning_rate": 3.933146004934044e-05, "loss": 3.2333, "step": 322400 }, { "epoch": 2.847542344442678, "grad_norm": 4.478387832641602, "learning_rate": 3.9317884293225196e-05, "loss": 3.0793, "step": 322450 }, { "epoch": 2.8479838923329623, "grad_norm": 4.278714656829834, "learning_rate": 3.9304309362212796e-05, "loss": 3.0329, "step": 322500 }, { "epoch": 2.8484254402232465, "grad_norm": 4.327560901641846, "learning_rate": 3.929073525735181e-05, "loss": 3.2013, "step": 322550 }, { "epoch": 2.8488669881135307, "grad_norm": 2.4608829021453857, "learning_rate": 3.927716197969071e-05, "loss": 3.0989, "step": 322600 }, { "epoch": 2.849308536003815, "grad_norm": 2.4417970180511475, "learning_rate": 3.9263589530277914e-05, "loss": 2.9115, "step": 322650 }, { "epoch": 2.849750083894099, "grad_norm": 1.2066291570663452, "learning_rate": 3.925001791016178e-05, "loss": 2.7059, "step": 322700 }, { "epoch": 2.850191631784383, "grad_norm": 3.770719528198242, "learning_rate": 3.9236447120390613e-05, "loss": 2.632, "step": 322750 }, { "epoch": 2.8506331796746673, "grad_norm": 2.320700168609619, "learning_rate": 3.922287716201265e-05, "loss": 2.983, "step": 322800 }, { "epoch": 2.851074727564952, "grad_norm": 4.02369499206543, "learning_rate": 3.920930803607601e-05, "loss": 2.8862, "step": 322850 }, { "epoch": 2.8515162754552357, "grad_norm": 4.738246917724609, "learning_rate": 3.9195739743628834e-05, "loss": 3.1733, "step": 322900 }, { "epoch": 2.8519578233455203, "grad_norm": 7.145111560821533, "learning_rate": 3.9182172285719164e-05, "loss": 2.4519, "step": 322950 }, { "epoch": 2.852399371235804, "grad_norm": 5.671413421630859, "learning_rate": 3.916860566339493e-05, "loss": 3.093, "step": 323000 }, { "epoch": 2.8528409191260886, "grad_norm": 5.369057655334473, "learning_rate": 3.9155039877704045e-05, "loss": 2.9147, "step": 323050 }, { "epoch": 2.8532824670163723, "grad_norm": 3.403627872467041, "learning_rate": 3.9141474929694405e-05, "loss": 3.1702, "step": 323100 }, { "epoch": 2.853724014906657, "grad_norm": 2.0987472534179688, "learning_rate": 3.912791082041373e-05, "loss": 3.4201, "step": 323150 }, { "epoch": 2.854165562796941, "grad_norm": 3.0549709796905518, "learning_rate": 3.911434755090975e-05, "loss": 2.8575, "step": 323200 }, { "epoch": 2.8546071106872253, "grad_norm": 3.9749667644500732, "learning_rate": 3.91007851222301e-05, "loss": 3.2003, "step": 323250 }, { "epoch": 2.8550486585775094, "grad_norm": 3.519059896469116, "learning_rate": 3.908722353542241e-05, "loss": 3.0545, "step": 323300 }, { "epoch": 2.8554902064677936, "grad_norm": 2.779958963394165, "learning_rate": 3.907366279153415e-05, "loss": 3.2774, "step": 323350 }, { "epoch": 2.855931754358078, "grad_norm": 3.0135083198547363, "learning_rate": 3.9060102891612773e-05, "loss": 2.9812, "step": 323400 }, { "epoch": 2.856373302248362, "grad_norm": 5.907857894897461, "learning_rate": 3.904654383670571e-05, "loss": 2.8506, "step": 323450 }, { "epoch": 2.856814850138646, "grad_norm": 1.4584112167358398, "learning_rate": 3.9032985627860234e-05, "loss": 2.9341, "step": 323500 }, { "epoch": 2.8572563980289303, "grad_norm": 6.578943252563477, "learning_rate": 3.9019428266123614e-05, "loss": 3.4108, "step": 323550 }, { "epoch": 2.8576979459192144, "grad_norm": 3.4602301120758057, "learning_rate": 3.900587175254307e-05, "loss": 3.2977, "step": 323600 }, { "epoch": 2.8581394938094986, "grad_norm": 2.3777079582214355, "learning_rate": 3.899231608816568e-05, "loss": 2.9948, "step": 323650 }, { "epoch": 2.8585810416997828, "grad_norm": 4.821569919586182, "learning_rate": 3.8978761274038544e-05, "loss": 2.999, "step": 323700 }, { "epoch": 2.859022589590067, "grad_norm": 4.176419258117676, "learning_rate": 3.896520731120865e-05, "loss": 3.0204, "step": 323750 }, { "epoch": 2.859464137480351, "grad_norm": 2.63362455368042, "learning_rate": 3.89516542007229e-05, "loss": 2.794, "step": 323800 }, { "epoch": 2.8599056853706353, "grad_norm": 1.5864942073822021, "learning_rate": 3.893810194362817e-05, "loss": 3.274, "step": 323850 }, { "epoch": 2.8603472332609194, "grad_norm": 2.7427353858947754, "learning_rate": 3.892455054097128e-05, "loss": 3.1046, "step": 323900 }, { "epoch": 2.8607887811512036, "grad_norm": 2.495997428894043, "learning_rate": 3.8910999993798954e-05, "loss": 3.2559, "step": 323950 }, { "epoch": 2.8612303290414878, "grad_norm": 3.55617356300354, "learning_rate": 3.889745030315782e-05, "loss": 3.0769, "step": 324000 }, { "epoch": 2.8612303290414878, "eval_asr_loss": 0.9337131119245551, "eval_loss": 2.779960870742798, "eval_runtime": 20.4313, "eval_samples_per_second": 37.589, "eval_steps_per_second": 9.397, "eval_tts_loss": 6.046140756807925, "step": 324000 }, { "epoch": 2.861671876931772, "grad_norm": 3.9437503814697266, "learning_rate": 3.888390147009452e-05, "loss": 3.0812, "step": 324050 }, { "epoch": 2.862113424822056, "grad_norm": 2.7651803493499756, "learning_rate": 3.8870624446723055e-05, "loss": 3.1213, "step": 324100 }, { "epoch": 2.8625549727123403, "grad_norm": 3.6669509410858154, "learning_rate": 3.885707731475126e-05, "loss": 2.9786, "step": 324150 }, { "epoch": 2.8629965206026244, "grad_norm": 5.497164726257324, "learning_rate": 3.884353104347572e-05, "loss": 3.1248, "step": 324200 }, { "epoch": 2.8634380684929086, "grad_norm": 4.338306903839111, "learning_rate": 3.882998563394283e-05, "loss": 3.0043, "step": 324250 }, { "epoch": 2.8638796163831928, "grad_norm": 2.680623769760132, "learning_rate": 3.8816441087198805e-05, "loss": 3.2286, "step": 324300 }, { "epoch": 2.864321164273477, "grad_norm": 1.8883942365646362, "learning_rate": 3.880289740428991e-05, "loss": 2.9522, "step": 324350 }, { "epoch": 2.8647627121637615, "grad_norm": 2.0757341384887695, "learning_rate": 3.878935458626222e-05, "loss": 2.9099, "step": 324400 }, { "epoch": 2.8652042600540453, "grad_norm": 2.192152738571167, "learning_rate": 3.877581263416182e-05, "loss": 2.8325, "step": 324450 }, { "epoch": 2.86564580794433, "grad_norm": 4.118265151977539, "learning_rate": 3.876227154903473e-05, "loss": 2.92, "step": 324500 }, { "epoch": 2.8660873558346136, "grad_norm": 2.0816004276275635, "learning_rate": 3.874873133192685e-05, "loss": 2.7259, "step": 324550 }, { "epoch": 2.866528903724898, "grad_norm": 2.563946008682251, "learning_rate": 3.873519198388408e-05, "loss": 3.0448, "step": 324600 }, { "epoch": 2.866970451615182, "grad_norm": 5.147098541259766, "learning_rate": 3.872165350595222e-05, "loss": 3.1172, "step": 324650 }, { "epoch": 2.8674119995054665, "grad_norm": 4.007184028625488, "learning_rate": 3.8708115899176964e-05, "loss": 3.3178, "step": 324700 }, { "epoch": 2.8678535473957503, "grad_norm": 2.0417864322662354, "learning_rate": 3.8694579164604e-05, "loss": 3.2263, "step": 324750 }, { "epoch": 2.868295095286035, "grad_norm": 4.987903594970703, "learning_rate": 3.868104330327895e-05, "loss": 3.0425, "step": 324800 }, { "epoch": 2.868736643176319, "grad_norm": 3.959479331970215, "learning_rate": 3.866750831624732e-05, "loss": 2.7931, "step": 324850 }, { "epoch": 2.869178191066603, "grad_norm": 4.61334753036499, "learning_rate": 3.8653974204554555e-05, "loss": 2.8141, "step": 324900 }, { "epoch": 2.8696197389568874, "grad_norm": 3.319843053817749, "learning_rate": 3.8640440969246086e-05, "loss": 3.4218, "step": 324950 }, { "epoch": 2.8700612868471715, "grad_norm": 2.896686315536499, "learning_rate": 3.862690861136724e-05, "loss": 3.0453, "step": 325000 }, { "epoch": 2.8705028347374557, "grad_norm": 7.583514213562012, "learning_rate": 3.8613377131963236e-05, "loss": 2.8892, "step": 325050 }, { "epoch": 2.87094438262774, "grad_norm": 1.726351261138916, "learning_rate": 3.85998465320793e-05, "loss": 3.0332, "step": 325100 }, { "epoch": 2.871385930518024, "grad_norm": 3.7438018321990967, "learning_rate": 3.858631681276056e-05, "loss": 3.2746, "step": 325150 }, { "epoch": 2.871827478408308, "grad_norm": 2.2730629444122314, "learning_rate": 3.857278797505206e-05, "loss": 3.3177, "step": 325200 }, { "epoch": 2.8722690262985924, "grad_norm": 3.9020748138427734, "learning_rate": 3.855926001999877e-05, "loss": 3.2541, "step": 325250 }, { "epoch": 2.8727105741888765, "grad_norm": 4.042573928833008, "learning_rate": 3.8545732948645656e-05, "loss": 2.8034, "step": 325300 }, { "epoch": 2.8731521220791607, "grad_norm": 4.340421676635742, "learning_rate": 3.853220676203752e-05, "loss": 3.3744, "step": 325350 }, { "epoch": 2.873593669969445, "grad_norm": 4.19350528717041, "learning_rate": 3.851868146121917e-05, "loss": 2.7871, "step": 325400 }, { "epoch": 2.874035217859729, "grad_norm": 2.1927661895751953, "learning_rate": 3.850515704723532e-05, "loss": 3.1893, "step": 325450 }, { "epoch": 2.874476765750013, "grad_norm": 2.8812367916107178, "learning_rate": 3.849163352113063e-05, "loss": 3.0758, "step": 325500 }, { "epoch": 2.8749183136402974, "grad_norm": 3.124998092651367, "learning_rate": 3.8478110883949646e-05, "loss": 2.8831, "step": 325550 }, { "epoch": 2.8753598615305815, "grad_norm": 3.3118693828582764, "learning_rate": 3.84645891367369e-05, "loss": 3.1014, "step": 325600 }, { "epoch": 2.8758014094208657, "grad_norm": 2.3778321743011475, "learning_rate": 3.845106828053684e-05, "loss": 2.6932, "step": 325650 }, { "epoch": 2.87624295731115, "grad_norm": 4.475037097930908, "learning_rate": 3.843754831639379e-05, "loss": 3.0869, "step": 325700 }, { "epoch": 2.876684505201434, "grad_norm": 4.238306999206543, "learning_rate": 3.842402924535211e-05, "loss": 2.7777, "step": 325750 }, { "epoch": 2.877126053091718, "grad_norm": 1.9995713233947754, "learning_rate": 3.8410511068456015e-05, "loss": 3.0332, "step": 325800 }, { "epoch": 2.8775676009820024, "grad_norm": 1.768629550933838, "learning_rate": 3.8396993786749646e-05, "loss": 2.7769, "step": 325850 }, { "epoch": 2.8780091488722865, "grad_norm": 1.708083987236023, "learning_rate": 3.8383477401277115e-05, "loss": 2.9163, "step": 325900 }, { "epoch": 2.8784506967625707, "grad_norm": 4.513185024261475, "learning_rate": 3.836996191308247e-05, "loss": 3.2235, "step": 325950 }, { "epoch": 2.878892244652855, "grad_norm": 2.3518264293670654, "learning_rate": 3.835644732320963e-05, "loss": 2.9127, "step": 326000 }, { "epoch": 2.8793337925431395, "grad_norm": 1.802065134048462, "learning_rate": 3.83429336327025e-05, "loss": 2.9651, "step": 326050 }, { "epoch": 2.879775340433423, "grad_norm": 2.038851261138916, "learning_rate": 3.832942084260489e-05, "loss": 2.9166, "step": 326100 }, { "epoch": 2.880216888323708, "grad_norm": 3.225442409515381, "learning_rate": 3.831590895396058e-05, "loss": 2.9742, "step": 326150 }, { "epoch": 2.8806584362139915, "grad_norm": 4.414299488067627, "learning_rate": 3.830239796781322e-05, "loss": 3.507, "step": 326200 }, { "epoch": 2.881099984104276, "grad_norm": 4.1165337562561035, "learning_rate": 3.828888788520641e-05, "loss": 2.6038, "step": 326250 }, { "epoch": 2.88154153199456, "grad_norm": 2.266355514526367, "learning_rate": 3.827537870718372e-05, "loss": 2.8124, "step": 326300 }, { "epoch": 2.8819830798848445, "grad_norm": 7.023044586181641, "learning_rate": 3.82618704347886e-05, "loss": 2.5443, "step": 326350 }, { "epoch": 2.8824246277751286, "grad_norm": 4.502677917480469, "learning_rate": 3.8248363069064434e-05, "loss": 3.3071, "step": 326400 }, { "epoch": 2.882866175665413, "grad_norm": 7.942958354949951, "learning_rate": 3.823485661105459e-05, "loss": 3.1744, "step": 326450 }, { "epoch": 2.883307723555697, "grad_norm": 2.9026122093200684, "learning_rate": 3.8221351061802294e-05, "loss": 3.214, "step": 326500 }, { "epoch": 2.883749271445981, "grad_norm": 1.9390217065811157, "learning_rate": 3.8207846422350756e-05, "loss": 3.1786, "step": 326550 }, { "epoch": 2.8841908193362653, "grad_norm": 2.796302556991577, "learning_rate": 3.819434269374308e-05, "loss": 2.9076, "step": 326600 }, { "epoch": 2.8846323672265495, "grad_norm": 3.1064558029174805, "learning_rate": 3.818083987702234e-05, "loss": 2.9861, "step": 326650 }, { "epoch": 2.8850739151168336, "grad_norm": 2.622194528579712, "learning_rate": 3.8167337973231476e-05, "loss": 3.1643, "step": 326700 }, { "epoch": 2.885515463007118, "grad_norm": 1.9751278162002563, "learning_rate": 3.815383698341342e-05, "loss": 3.1125, "step": 326750 }, { "epoch": 2.885957010897402, "grad_norm": 2.3864688873291016, "learning_rate": 3.8140336908611026e-05, "loss": 2.9132, "step": 326800 }, { "epoch": 2.886398558787686, "grad_norm": 5.334734916687012, "learning_rate": 3.8126837749867e-05, "loss": 3.1827, "step": 326850 }, { "epoch": 2.8868401066779703, "grad_norm": 1.6681851148605347, "learning_rate": 3.8113339508224105e-05, "loss": 3.2007, "step": 326900 }, { "epoch": 2.8872816545682545, "grad_norm": 0.5884609818458557, "learning_rate": 3.8099842184724935e-05, "loss": 2.8974, "step": 326950 }, { "epoch": 2.8877232024585386, "grad_norm": 3.291308641433716, "learning_rate": 3.8086345780412035e-05, "loss": 2.9299, "step": 327000 }, { "epoch": 2.8877232024585386, "eval_asr_loss": 0.9402317406163259, "eval_loss": 2.7857255935668945, "eval_runtime": 20.5619, "eval_samples_per_second": 37.351, "eval_steps_per_second": 9.338, "eval_tts_loss": 6.062726298646587, "step": 327000 }, { "epoch": 2.888164750348823, "grad_norm": 1.1211483478546143, "learning_rate": 3.80728502963279e-05, "loss": 3.7824, "step": 327050 }, { "epoch": 2.888606298239107, "grad_norm": 2.359931707382202, "learning_rate": 3.8059355733514954e-05, "loss": 3.0394, "step": 327100 }, { "epoch": 2.889047846129391, "grad_norm": 3.3292713165283203, "learning_rate": 3.804586209301552e-05, "loss": 3.3375, "step": 327150 }, { "epoch": 2.8894893940196753, "grad_norm": 2.5566701889038086, "learning_rate": 3.8032369375871853e-05, "loss": 2.8996, "step": 327200 }, { "epoch": 2.8899309419099595, "grad_norm": 2.4845311641693115, "learning_rate": 3.8018877583126165e-05, "loss": 3.0081, "step": 327250 }, { "epoch": 2.8903724898002436, "grad_norm": 1.988072395324707, "learning_rate": 3.800538671582062e-05, "loss": 2.8863, "step": 327300 }, { "epoch": 2.890814037690528, "grad_norm": 3.7281973361968994, "learning_rate": 3.799189677499722e-05, "loss": 3.1157, "step": 327350 }, { "epoch": 2.891255585580812, "grad_norm": 3.4577279090881348, "learning_rate": 3.7978407761697964e-05, "loss": 2.6681, "step": 327400 }, { "epoch": 2.891697133471096, "grad_norm": 4.655624866485596, "learning_rate": 3.796491967696479e-05, "loss": 2.3905, "step": 327450 }, { "epoch": 2.8921386813613803, "grad_norm": 3.3842995166778564, "learning_rate": 3.7951432521839494e-05, "loss": 2.9748, "step": 327500 }, { "epoch": 2.8925802292516645, "grad_norm": 2.881986618041992, "learning_rate": 3.793794629736387e-05, "loss": 2.9094, "step": 327550 }, { "epoch": 2.893021777141949, "grad_norm": 2.483426809310913, "learning_rate": 3.792446100457963e-05, "loss": 3.1122, "step": 327600 }, { "epoch": 2.893463325032233, "grad_norm": 3.1974425315856934, "learning_rate": 3.791097664452835e-05, "loss": 2.9753, "step": 327650 }, { "epoch": 2.8939048729225174, "grad_norm": 7.178389549255371, "learning_rate": 3.7897493218251634e-05, "loss": 2.836, "step": 327700 }, { "epoch": 2.894346420812801, "grad_norm": 4.327606678009033, "learning_rate": 3.7884010726790955e-05, "loss": 3.0458, "step": 327750 }, { "epoch": 2.8947879687030857, "grad_norm": 1.6565728187561035, "learning_rate": 3.787052917118767e-05, "loss": 3.0489, "step": 327800 }, { "epoch": 2.8952295165933695, "grad_norm": 6.238094329833984, "learning_rate": 3.785704855248316e-05, "loss": 2.8781, "step": 327850 }, { "epoch": 2.895671064483654, "grad_norm": 1.89829421043396, "learning_rate": 3.7843568871718696e-05, "loss": 3.0075, "step": 327900 }, { "epoch": 2.896112612373938, "grad_norm": 6.175403594970703, "learning_rate": 3.783009012993545e-05, "loss": 2.8208, "step": 327950 }, { "epoch": 2.8965541602642224, "grad_norm": 2.4365267753601074, "learning_rate": 3.7816612328174526e-05, "loss": 3.1571, "step": 328000 }, { "epoch": 2.8969957081545066, "grad_norm": 5.112443923950195, "learning_rate": 3.7803135467476995e-05, "loss": 2.661, "step": 328050 }, { "epoch": 2.8974372560447907, "grad_norm": 3.3050458431243896, "learning_rate": 3.7789659548883844e-05, "loss": 2.884, "step": 328100 }, { "epoch": 2.897878803935075, "grad_norm": 9.001471519470215, "learning_rate": 3.7776184573435934e-05, "loss": 2.4716, "step": 328150 }, { "epoch": 2.898320351825359, "grad_norm": 4.496976852416992, "learning_rate": 3.7762710542174105e-05, "loss": 3.0556, "step": 328200 }, { "epoch": 2.8987618997156432, "grad_norm": 5.237453937530518, "learning_rate": 3.774923745613914e-05, "loss": 2.7898, "step": 328250 }, { "epoch": 2.8992034476059274, "grad_norm": 5.313985347747803, "learning_rate": 3.7735765316371674e-05, "loss": 2.7367, "step": 328300 }, { "epoch": 2.8996449954962116, "grad_norm": 1.7633111476898193, "learning_rate": 3.772229412391235e-05, "loss": 3.1914, "step": 328350 }, { "epoch": 2.9000865433864957, "grad_norm": 1.610337734222412, "learning_rate": 3.770882387980168e-05, "loss": 3.0965, "step": 328400 }, { "epoch": 2.90052809127678, "grad_norm": 1.8957462310791016, "learning_rate": 3.769535458508017e-05, "loss": 2.8806, "step": 328450 }, { "epoch": 2.900969639167064, "grad_norm": 5.257757186889648, "learning_rate": 3.7681886240788154e-05, "loss": 2.952, "step": 328500 }, { "epoch": 2.9014111870573482, "grad_norm": 4.04410457611084, "learning_rate": 3.766841884796597e-05, "loss": 3.0986, "step": 328550 }, { "epoch": 2.9018527349476324, "grad_norm": 1.8487319946289062, "learning_rate": 3.7654952407653884e-05, "loss": 3.6347, "step": 328600 }, { "epoch": 2.9022942828379166, "grad_norm": 2.616741180419922, "learning_rate": 3.764148692089201e-05, "loss": 3.6286, "step": 328650 }, { "epoch": 2.9027358307282007, "grad_norm": 4.585070610046387, "learning_rate": 3.7628022388720495e-05, "loss": 3.0244, "step": 328700 }, { "epoch": 2.903177378618485, "grad_norm": 4.126288414001465, "learning_rate": 3.7614558812179346e-05, "loss": 2.8935, "step": 328750 }, { "epoch": 2.903618926508769, "grad_norm": 3.3980774879455566, "learning_rate": 3.760109619230848e-05, "loss": 2.9783, "step": 328800 }, { "epoch": 2.9040604743990532, "grad_norm": 3.2447967529296875, "learning_rate": 3.75876345301478e-05, "loss": 3.0316, "step": 328850 }, { "epoch": 2.9045020222893374, "grad_norm": 5.0083465576171875, "learning_rate": 3.757444303140285e-05, "loss": 3.1447, "step": 328900 }, { "epoch": 2.9049435701796216, "grad_norm": 5.211503505706787, "learning_rate": 3.756098326857586e-05, "loss": 3.3478, "step": 328950 }, { "epoch": 2.9053851180699057, "grad_norm": 4.124146938323975, "learning_rate": 3.754752446655746e-05, "loss": 3.1843, "step": 329000 }, { "epoch": 2.90582666596019, "grad_norm": 2.0536270141601562, "learning_rate": 3.753406662638719e-05, "loss": 3.1675, "step": 329050 }, { "epoch": 2.906268213850474, "grad_norm": 2.693697214126587, "learning_rate": 3.7520609749104554e-05, "loss": 3.4515, "step": 329100 }, { "epoch": 2.9067097617407587, "grad_norm": 2.6549017429351807, "learning_rate": 3.7507153835749015e-05, "loss": 3.5519, "step": 329150 }, { "epoch": 2.9071513096310424, "grad_norm": 2.0209426879882812, "learning_rate": 3.74936988873599e-05, "loss": 3.4394, "step": 329200 }, { "epoch": 2.907592857521327, "grad_norm": 4.286931991577148, "learning_rate": 3.748024490497647e-05, "loss": 2.829, "step": 329250 }, { "epoch": 2.9080344054116107, "grad_norm": 2.264556407928467, "learning_rate": 3.7466791889637994e-05, "loss": 2.7303, "step": 329300 }, { "epoch": 2.9084759533018953, "grad_norm": 2.5618410110473633, "learning_rate": 3.745333984238354e-05, "loss": 3.3025, "step": 329350 }, { "epoch": 2.908917501192179, "grad_norm": 4.213635444641113, "learning_rate": 3.7439888764252195e-05, "loss": 3.004, "step": 329400 }, { "epoch": 2.9093590490824637, "grad_norm": 2.256089687347412, "learning_rate": 3.742643865628293e-05, "loss": 2.9745, "step": 329450 }, { "epoch": 2.9098005969727474, "grad_norm": 3.6799662113189697, "learning_rate": 3.741298951951468e-05, "loss": 3.0717, "step": 329500 }, { "epoch": 2.910242144863032, "grad_norm": 1.167763352394104, "learning_rate": 3.7399541354986235e-05, "loss": 3.1447, "step": 329550 }, { "epoch": 2.910683692753316, "grad_norm": 6.323507785797119, "learning_rate": 3.7386094163736376e-05, "loss": 3.0887, "step": 329600 }, { "epoch": 2.9111252406436003, "grad_norm": 2.0400609970092773, "learning_rate": 3.7372647946803796e-05, "loss": 2.9279, "step": 329650 }, { "epoch": 2.9115667885338845, "grad_norm": 4.3869476318359375, "learning_rate": 3.735920270522707e-05, "loss": 3.0775, "step": 329700 }, { "epoch": 2.9120083364241687, "grad_norm": 2.502378225326538, "learning_rate": 3.7345758440044745e-05, "loss": 3.0458, "step": 329750 }, { "epoch": 2.912449884314453, "grad_norm": 2.871675968170166, "learning_rate": 3.733231515229528e-05, "loss": 3.2548, "step": 329800 }, { "epoch": 2.912891432204737, "grad_norm": 2.317923069000244, "learning_rate": 3.731887284301705e-05, "loss": 2.675, "step": 329850 }, { "epoch": 2.913332980095021, "grad_norm": 1.320755958557129, "learning_rate": 3.730543151324835e-05, "loss": 3.1972, "step": 329900 }, { "epoch": 2.9137745279853053, "grad_norm": 1.930503010749817, "learning_rate": 3.7291991164027426e-05, "loss": 3.0267, "step": 329950 }, { "epoch": 2.9142160758755895, "grad_norm": 1.182265281677246, "learning_rate": 3.727855179639244e-05, "loss": 3.433, "step": 330000 }, { "epoch": 2.9142160758755895, "eval_asr_loss": 0.9296056722839383, "eval_loss": 2.7678422927856445, "eval_runtime": 20.814, "eval_samples_per_second": 36.898, "eval_steps_per_second": 9.225, "eval_tts_loss": 6.053693671282148, "step": 330000 }, { "epoch": 2.9146576237658737, "grad_norm": 2.976539134979248, "learning_rate": 3.7265113411381425e-05, "loss": 3.1771, "step": 330050 }, { "epoch": 2.915099171656158, "grad_norm": 5.071763515472412, "learning_rate": 3.725167601003241e-05, "loss": 3.1136, "step": 330100 }, { "epoch": 2.915540719546442, "grad_norm": 2.773087739944458, "learning_rate": 3.7238239593383346e-05, "loss": 3.2592, "step": 330150 }, { "epoch": 2.915982267436726, "grad_norm": 2.2008588314056396, "learning_rate": 3.722480416247205e-05, "loss": 2.9436, "step": 330200 }, { "epoch": 2.9164238153270103, "grad_norm": 3.125373125076294, "learning_rate": 3.721136971833629e-05, "loss": 2.6633, "step": 330250 }, { "epoch": 2.9168653632172945, "grad_norm": 7.892586708068848, "learning_rate": 3.719793626201379e-05, "loss": 3.0586, "step": 330300 }, { "epoch": 2.9173069111075787, "grad_norm": 2.5813841819763184, "learning_rate": 3.718450379454214e-05, "loss": 2.8936, "step": 330350 }, { "epoch": 2.917748458997863, "grad_norm": 7.8572564125061035, "learning_rate": 3.71710723169589e-05, "loss": 3.1666, "step": 330400 }, { "epoch": 2.918190006888147, "grad_norm": 4.011484146118164, "learning_rate": 3.7157641830301557e-05, "loss": 3.0879, "step": 330450 }, { "epoch": 2.918631554778431, "grad_norm": 1.9631547927856445, "learning_rate": 3.714421233560746e-05, "loss": 3.0823, "step": 330500 }, { "epoch": 2.9190731026687153, "grad_norm": 8.067952156066895, "learning_rate": 3.713078383391396e-05, "loss": 2.774, "step": 330550 }, { "epoch": 2.9195146505589995, "grad_norm": 2.8599631786346436, "learning_rate": 3.7117356326258264e-05, "loss": 3.1622, "step": 330600 }, { "epoch": 2.9199561984492837, "grad_norm": 5.9154438972473145, "learning_rate": 3.710392981367757e-05, "loss": 3.0889, "step": 330650 }, { "epoch": 2.920397746339568, "grad_norm": 4.798829555511475, "learning_rate": 3.709050429720893e-05, "loss": 2.7948, "step": 330700 }, { "epoch": 2.920839294229852, "grad_norm": 2.746203899383545, "learning_rate": 3.707707977788937e-05, "loss": 2.8252, "step": 330750 }, { "epoch": 2.9212808421201366, "grad_norm": 4.844743728637695, "learning_rate": 3.706365625675583e-05, "loss": 2.9372, "step": 330800 }, { "epoch": 2.9217223900104203, "grad_norm": 1.096981167793274, "learning_rate": 3.705023373484511e-05, "loss": 3.0962, "step": 330850 }, { "epoch": 2.922163937900705, "grad_norm": 4.171241283416748, "learning_rate": 3.7036812213194036e-05, "loss": 3.2679, "step": 330900 }, { "epoch": 2.9226054857909887, "grad_norm": 2.6895759105682373, "learning_rate": 3.702339169283932e-05, "loss": 3.0191, "step": 330950 }, { "epoch": 2.9230470336812733, "grad_norm": 3.8773746490478516, "learning_rate": 3.7009972174817534e-05, "loss": 3.1038, "step": 331000 }, { "epoch": 2.923488581571557, "grad_norm": 5.505143165588379, "learning_rate": 3.699655366016524e-05, "loss": 3.3741, "step": 331050 }, { "epoch": 2.9239301294618416, "grad_norm": 4.710068702697754, "learning_rate": 3.698313614991893e-05, "loss": 3.3488, "step": 331100 }, { "epoch": 2.9243716773521258, "grad_norm": 2.812563180923462, "learning_rate": 3.6969719645114955e-05, "loss": 3.019, "step": 331150 }, { "epoch": 2.92481322524241, "grad_norm": 1.5845237970352173, "learning_rate": 3.695630414678965e-05, "loss": 3.1854, "step": 331200 }, { "epoch": 2.925254773132694, "grad_norm": 2.8827528953552246, "learning_rate": 3.6942889655979236e-05, "loss": 3.4328, "step": 331250 }, { "epoch": 2.9256963210229783, "grad_norm": 3.0969033241271973, "learning_rate": 3.692947617371989e-05, "loss": 3.3493, "step": 331300 }, { "epoch": 2.9261378689132624, "grad_norm": 3.3494527339935303, "learning_rate": 3.6916063701047675e-05, "loss": 3.1312, "step": 331350 }, { "epoch": 2.9265794168035466, "grad_norm": 1.9323502779006958, "learning_rate": 3.690265223899857e-05, "loss": 2.9451, "step": 331400 }, { "epoch": 2.9270209646938308, "grad_norm": 1.9129736423492432, "learning_rate": 3.688924178860854e-05, "loss": 2.8515, "step": 331450 }, { "epoch": 2.927462512584115, "grad_norm": 4.443826198577881, "learning_rate": 3.687583235091339e-05, "loss": 2.9671, "step": 331500 }, { "epoch": 2.927904060474399, "grad_norm": 2.1081154346466064, "learning_rate": 3.686242392694891e-05, "loss": 3.1953, "step": 331550 }, { "epoch": 2.9283456083646833, "grad_norm": 3.4340193271636963, "learning_rate": 3.684901651775078e-05, "loss": 2.8878, "step": 331600 }, { "epoch": 2.9287871562549674, "grad_norm": 3.959946632385254, "learning_rate": 3.68356101243546e-05, "loss": 2.9926, "step": 331650 }, { "epoch": 2.9292287041452516, "grad_norm": 3.000515937805176, "learning_rate": 3.6822204747795894e-05, "loss": 3.2043, "step": 331700 }, { "epoch": 2.9296702520355358, "grad_norm": 4.031549453735352, "learning_rate": 3.680880038911014e-05, "loss": 3.3333, "step": 331750 }, { "epoch": 2.93011179992582, "grad_norm": 2.494476318359375, "learning_rate": 3.679539704933271e-05, "loss": 2.8566, "step": 331800 }, { "epoch": 2.930553347816104, "grad_norm": 2.633342742919922, "learning_rate": 3.678199472949888e-05, "loss": 3.4151, "step": 331850 }, { "epoch": 2.9309948957063883, "grad_norm": 2.808192491531372, "learning_rate": 3.676859343064386e-05, "loss": 3.2701, "step": 331900 }, { "epoch": 2.9314364435966724, "grad_norm": 1.35281503200531, "learning_rate": 3.6755193153802824e-05, "loss": 3.2665, "step": 331950 }, { "epoch": 2.9318779914869566, "grad_norm": 4.437047958374023, "learning_rate": 3.6741793900010774e-05, "loss": 3.3267, "step": 332000 }, { "epoch": 2.9323195393772408, "grad_norm": 3.09515643119812, "learning_rate": 3.6728395670302727e-05, "loss": 3.2185, "step": 332050 }, { "epoch": 2.932761087267525, "grad_norm": 1.5784809589385986, "learning_rate": 3.6714998465713586e-05, "loss": 2.99, "step": 332100 }, { "epoch": 2.933202635157809, "grad_norm": 4.678462028503418, "learning_rate": 3.6701602287278156e-05, "loss": 3.1005, "step": 332150 }, { "epoch": 2.9336441830480933, "grad_norm": 4.583435535430908, "learning_rate": 3.668820713603117e-05, "loss": 2.8725, "step": 332200 }, { "epoch": 2.9340857309383774, "grad_norm": 1.745843529701233, "learning_rate": 3.667481301300732e-05, "loss": 2.9895, "step": 332250 }, { "epoch": 2.9345272788286616, "grad_norm": 5.429660320281982, "learning_rate": 3.666141991924115e-05, "loss": 2.9047, "step": 332300 }, { "epoch": 2.934968826718946, "grad_norm": 1.8488998413085938, "learning_rate": 3.66480278557672e-05, "loss": 2.793, "step": 332350 }, { "epoch": 2.93541037460923, "grad_norm": 2.0155107975006104, "learning_rate": 3.663463682361986e-05, "loss": 3.2467, "step": 332400 }, { "epoch": 2.9358519224995145, "grad_norm": 4.550942897796631, "learning_rate": 3.6621246823833515e-05, "loss": 3.0803, "step": 332450 }, { "epoch": 2.9362934703897983, "grad_norm": 5.115889072418213, "learning_rate": 3.6608125626636256e-05, "loss": 3.1888, "step": 332500 }, { "epoch": 2.936735018280083, "grad_norm": 3.910883903503418, "learning_rate": 3.659473767397582e-05, "loss": 2.8184, "step": 332550 }, { "epoch": 2.9371765661703666, "grad_norm": 1.9050536155700684, "learning_rate": 3.6581350756758234e-05, "loss": 3.311, "step": 332600 }, { "epoch": 2.937618114060651, "grad_norm": 5.972596645355225, "learning_rate": 3.6567964876017515e-05, "loss": 2.5427, "step": 332650 }, { "epoch": 2.938059661950935, "grad_norm": 4.375555992126465, "learning_rate": 3.655458003278761e-05, "loss": 2.8741, "step": 332700 }, { "epoch": 2.9385012098412195, "grad_norm": 6.029248237609863, "learning_rate": 3.654119622810238e-05, "loss": 3.1557, "step": 332750 }, { "epoch": 2.9389427577315037, "grad_norm": 4.237759590148926, "learning_rate": 3.652781346299561e-05, "loss": 2.8764, "step": 332800 }, { "epoch": 2.939384305621788, "grad_norm": 5.90347146987915, "learning_rate": 3.651443173850103e-05, "loss": 2.7617, "step": 332850 }, { "epoch": 2.939825853512072, "grad_norm": 2.8472068309783936, "learning_rate": 3.650105105565222e-05, "loss": 2.9146, "step": 332900 }, { "epoch": 2.940267401402356, "grad_norm": 2.6870415210723877, "learning_rate": 3.6487671415482746e-05, "loss": 3.2174, "step": 332950 }, { "epoch": 2.9407089492926404, "grad_norm": 1.8696746826171875, "learning_rate": 3.64742928190261e-05, "loss": 3.1888, "step": 333000 }, { "epoch": 2.9407089492926404, "eval_asr_loss": 0.9168856108272408, "eval_loss": 2.7696163654327393, "eval_runtime": 20.802, "eval_samples_per_second": 36.92, "eval_steps_per_second": 9.23, "eval_tts_loss": 6.04860078041589, "step": 333000 }, { "epoch": 2.9411504971829245, "grad_norm": 4.217530250549316, "learning_rate": 3.6460915267315625e-05, "loss": 3.2377, "step": 333050 }, { "epoch": 2.9415920450732087, "grad_norm": 3.945490837097168, "learning_rate": 3.6447538761384625e-05, "loss": 2.6695, "step": 333100 }, { "epoch": 2.942033592963493, "grad_norm": 5.2477803230285645, "learning_rate": 3.643416330226636e-05, "loss": 3.1024, "step": 333150 }, { "epoch": 2.942475140853777, "grad_norm": 2.376180648803711, "learning_rate": 3.6420788890993926e-05, "loss": 3.2106, "step": 333200 }, { "epoch": 2.942916688744061, "grad_norm": 4.131772518157959, "learning_rate": 3.6407415528600405e-05, "loss": 3.2866, "step": 333250 }, { "epoch": 2.9433582366343454, "grad_norm": 1.885551929473877, "learning_rate": 3.639404321611879e-05, "loss": 2.8676, "step": 333300 }, { "epoch": 2.9437997845246295, "grad_norm": 9.674749374389648, "learning_rate": 3.6380671954581936e-05, "loss": 3.0087, "step": 333350 }, { "epoch": 2.9442413324149137, "grad_norm": 7.844644546508789, "learning_rate": 3.63673017450227e-05, "loss": 3.0909, "step": 333400 }, { "epoch": 2.944682880305198, "grad_norm": 3.822357177734375, "learning_rate": 3.635393258847378e-05, "loss": 2.9979, "step": 333450 }, { "epoch": 2.945124428195482, "grad_norm": 3.677823305130005, "learning_rate": 3.6340564485967885e-05, "loss": 3.0533, "step": 333500 }, { "epoch": 2.945565976085766, "grad_norm": 3.729410171508789, "learning_rate": 3.632719743853753e-05, "loss": 3.2639, "step": 333550 }, { "epoch": 2.9460075239760504, "grad_norm": 4.694045066833496, "learning_rate": 3.631383144721524e-05, "loss": 2.7681, "step": 333600 }, { "epoch": 2.9464490718663345, "grad_norm": 3.1156671047210693, "learning_rate": 3.6300466513033426e-05, "loss": 2.5813, "step": 333650 }, { "epoch": 2.9468906197566187, "grad_norm": 6.814661979675293, "learning_rate": 3.628710263702439e-05, "loss": 3.1557, "step": 333700 }, { "epoch": 2.947332167646903, "grad_norm": 4.338254451751709, "learning_rate": 3.627373982022039e-05, "loss": 3.1207, "step": 333750 }, { "epoch": 2.947773715537187, "grad_norm": 2.386225938796997, "learning_rate": 3.6260378063653614e-05, "loss": 3.1153, "step": 333800 }, { "epoch": 2.948215263427471, "grad_norm": 4.170174598693848, "learning_rate": 3.624701736835611e-05, "loss": 3.0631, "step": 333850 }, { "epoch": 2.9486568113177554, "grad_norm": 0.9369651675224304, "learning_rate": 3.623365773535988e-05, "loss": 3.0142, "step": 333900 }, { "epoch": 2.9490983592080395, "grad_norm": 3.9977307319641113, "learning_rate": 3.622029916569688e-05, "loss": 3.0172, "step": 333950 }, { "epoch": 2.949539907098324, "grad_norm": 4.269778728485107, "learning_rate": 3.6206941660398906e-05, "loss": 3.3565, "step": 334000 }, { "epoch": 2.949981454988608, "grad_norm": 3.542271137237549, "learning_rate": 3.619358522049773e-05, "loss": 2.707, "step": 334050 }, { "epoch": 2.9504230028788925, "grad_norm": 3.940690279006958, "learning_rate": 3.6180229847025015e-05, "loss": 3.0872, "step": 334100 }, { "epoch": 2.950864550769176, "grad_norm": 5.709620475769043, "learning_rate": 3.616687554101238e-05, "loss": 2.93, "step": 334150 }, { "epoch": 2.951306098659461, "grad_norm": 3.8366355895996094, "learning_rate": 3.615352230349129e-05, "loss": 2.9364, "step": 334200 }, { "epoch": 2.9517476465497445, "grad_norm": 4.951537609100342, "learning_rate": 3.614017013549318e-05, "loss": 3.1369, "step": 334250 }, { "epoch": 2.952189194440029, "grad_norm": 2.828788995742798, "learning_rate": 3.612681903804942e-05, "loss": 2.8485, "step": 334300 }, { "epoch": 2.9526307423303133, "grad_norm": 3.537008285522461, "learning_rate": 3.6113469012191224e-05, "loss": 3.402, "step": 334350 }, { "epoch": 2.9530722902205975, "grad_norm": 3.7862436771392822, "learning_rate": 3.6100120058949805e-05, "loss": 2.8811, "step": 334400 }, { "epoch": 2.9535138381108816, "grad_norm": 10.496225357055664, "learning_rate": 3.608677217935626e-05, "loss": 3.2583, "step": 334450 }, { "epoch": 2.953955386001166, "grad_norm": 6.932826995849609, "learning_rate": 3.607342537444155e-05, "loss": 2.8722, "step": 334500 }, { "epoch": 2.95439693389145, "grad_norm": 1.4328943490982056, "learning_rate": 3.6060079645236644e-05, "loss": 3.0572, "step": 334550 }, { "epoch": 2.954838481781734, "grad_norm": 1.214016079902649, "learning_rate": 3.6046734992772394e-05, "loss": 2.9215, "step": 334600 }, { "epoch": 2.9552800296720183, "grad_norm": 2.609496593475342, "learning_rate": 3.603339141807954e-05, "loss": 3.1791, "step": 334650 }, { "epoch": 2.9557215775623025, "grad_norm": 1.394241213798523, "learning_rate": 3.602004892218876e-05, "loss": 2.9442, "step": 334700 }, { "epoch": 2.9561631254525866, "grad_norm": 2.6506001949310303, "learning_rate": 3.600670750613066e-05, "loss": 2.9158, "step": 334750 }, { "epoch": 2.956604673342871, "grad_norm": 4.849084854125977, "learning_rate": 3.599336717093575e-05, "loss": 2.6812, "step": 334800 }, { "epoch": 2.957046221233155, "grad_norm": 1.6218433380126953, "learning_rate": 3.598002791763444e-05, "loss": 3.1562, "step": 334850 }, { "epoch": 2.957487769123439, "grad_norm": 5.685052871704102, "learning_rate": 3.5966689747257086e-05, "loss": 3.5786, "step": 334900 }, { "epoch": 2.9579293170137233, "grad_norm": 2.3673272132873535, "learning_rate": 3.595335266083397e-05, "loss": 3.2563, "step": 334950 }, { "epoch": 2.9583708649040075, "grad_norm": 4.300685405731201, "learning_rate": 3.5940016659395236e-05, "loss": 3.1972, "step": 335000 }, { "epoch": 2.9588124127942916, "grad_norm": 3.5480964183807373, "learning_rate": 3.592668174397098e-05, "loss": 3.1314, "step": 335050 }, { "epoch": 2.959253960684576, "grad_norm": 5.400235652923584, "learning_rate": 3.591334791559123e-05, "loss": 2.7594, "step": 335100 }, { "epoch": 2.95969550857486, "grad_norm": 6.059818744659424, "learning_rate": 3.590001517528592e-05, "loss": 3.2612, "step": 335150 }, { "epoch": 2.960137056465144, "grad_norm": 2.12428879737854, "learning_rate": 3.5886683524084855e-05, "loss": 2.9369, "step": 335200 }, { "epoch": 2.9605786043554283, "grad_norm": 3.4586431980133057, "learning_rate": 3.58733529630178e-05, "loss": 3.3392, "step": 335250 }, { "epoch": 2.9610201522457125, "grad_norm": 2.411804676055908, "learning_rate": 3.586002349311446e-05, "loss": 3.2683, "step": 335300 }, { "epoch": 2.9614617001359966, "grad_norm": 4.752664566040039, "learning_rate": 3.58466951154044e-05, "loss": 2.7463, "step": 335350 }, { "epoch": 2.961903248026281, "grad_norm": 4.002713203430176, "learning_rate": 3.5833367830917107e-05, "loss": 2.9187, "step": 335400 }, { "epoch": 2.962344795916565, "grad_norm": 5.920754432678223, "learning_rate": 3.5820041640682045e-05, "loss": 2.9047, "step": 335450 }, { "epoch": 2.962786343806849, "grad_norm": 3.04352068901062, "learning_rate": 3.58067165457285e-05, "loss": 3.0355, "step": 335500 }, { "epoch": 2.9632278916971337, "grad_norm": 7.904613971710205, "learning_rate": 3.579339254708576e-05, "loss": 2.9404, "step": 335550 }, { "epoch": 2.9636694395874175, "grad_norm": 1.849334716796875, "learning_rate": 3.578006964578298e-05, "loss": 2.1701, "step": 335600 }, { "epoch": 2.964110987477702, "grad_norm": 2.4684383869171143, "learning_rate": 3.576674784284923e-05, "loss": 3.2308, "step": 335650 }, { "epoch": 2.964552535367986, "grad_norm": 2.965360403060913, "learning_rate": 3.57534271393135e-05, "loss": 2.9975, "step": 335700 }, { "epoch": 2.9649940832582704, "grad_norm": 7.479698657989502, "learning_rate": 3.5740107536204733e-05, "loss": 2.8902, "step": 335750 }, { "epoch": 2.965435631148554, "grad_norm": 4.724883556365967, "learning_rate": 3.5726789034551756e-05, "loss": 2.9918, "step": 335800 }, { "epoch": 2.9658771790388387, "grad_norm": 1.8270108699798584, "learning_rate": 3.571347163538327e-05, "loss": 3.2007, "step": 335850 }, { "epoch": 2.9663187269291225, "grad_norm": 6.264216423034668, "learning_rate": 3.570015533972795e-05, "loss": 3.0281, "step": 335900 }, { "epoch": 2.966760274819407, "grad_norm": 3.543221950531006, "learning_rate": 3.5686840148614385e-05, "loss": 3.5208, "step": 335950 }, { "epoch": 2.9672018227096912, "grad_norm": 2.9505646228790283, "learning_rate": 3.5673526063071037e-05, "loss": 2.5677, "step": 336000 }, { "epoch": 2.9672018227096912, "eval_asr_loss": 0.9296773490936134, "eval_loss": 2.764009475708008, "eval_runtime": 20.8728, "eval_samples_per_second": 36.794, "eval_steps_per_second": 9.199, "eval_tts_loss": 6.078821064822733, "step": 336000 }, { "epoch": 2.9676433705999754, "grad_norm": 5.127497673034668, "learning_rate": 3.566021308412631e-05, "loss": 3.1488, "step": 336050 }, { "epoch": 2.9680849184902596, "grad_norm": 1.4386932849884033, "learning_rate": 3.5646901212808546e-05, "loss": 3.2834, "step": 336100 }, { "epoch": 2.9685264663805437, "grad_norm": 3.8077752590179443, "learning_rate": 3.5633590450145946e-05, "loss": 3.071, "step": 336150 }, { "epoch": 2.968968014270828, "grad_norm": 3.314805030822754, "learning_rate": 3.562028079716665e-05, "loss": 2.9205, "step": 336200 }, { "epoch": 2.969409562161112, "grad_norm": 6.505230903625488, "learning_rate": 3.560697225489874e-05, "loss": 3.1418, "step": 336250 }, { "epoch": 2.9698511100513962, "grad_norm": 4.541718482971191, "learning_rate": 3.559366482437015e-05, "loss": 3.1439, "step": 336300 }, { "epoch": 2.9702926579416804, "grad_norm": 3.6734509468078613, "learning_rate": 3.5580358506608814e-05, "loss": 3.0522, "step": 336350 }, { "epoch": 2.9707342058319646, "grad_norm": 5.184187412261963, "learning_rate": 3.5567053302642485e-05, "loss": 3.2141, "step": 336400 }, { "epoch": 2.9711757537222487, "grad_norm": 5.723706245422363, "learning_rate": 3.5553749213498936e-05, "loss": 3.137, "step": 336450 }, { "epoch": 2.971617301612533, "grad_norm": 4.463052272796631, "learning_rate": 3.554044624020572e-05, "loss": 2.757, "step": 336500 }, { "epoch": 2.972058849502817, "grad_norm": 5.090996265411377, "learning_rate": 3.552714438379044e-05, "loss": 3.2952, "step": 336550 }, { "epoch": 2.9725003973931012, "grad_norm": 3.7699790000915527, "learning_rate": 3.5513843645280544e-05, "loss": 3.0189, "step": 336600 }, { "epoch": 2.9729419452833854, "grad_norm": 3.4531867504119873, "learning_rate": 3.5500544025703354e-05, "loss": 3.0805, "step": 336650 }, { "epoch": 2.9733834931736696, "grad_norm": 5.142200946807861, "learning_rate": 3.548724552608621e-05, "loss": 3.272, "step": 336700 }, { "epoch": 2.9738250410639537, "grad_norm": 4.962176322937012, "learning_rate": 3.5473948147456284e-05, "loss": 2.8381, "step": 336750 }, { "epoch": 2.974266588954238, "grad_norm": 6.314472675323486, "learning_rate": 3.54609178049706e-05, "loss": 2.9614, "step": 336800 }, { "epoch": 2.974708136844522, "grad_norm": 4.447956562042236, "learning_rate": 3.544762264892546e-05, "loss": 2.7262, "step": 336850 }, { "epoch": 2.9751496847348062, "grad_norm": 4.069913864135742, "learning_rate": 3.5434594486545704e-05, "loss": 3.2869, "step": 336900 }, { "epoch": 2.9755912326250904, "grad_norm": 1.8606295585632324, "learning_rate": 3.542130155711134e-05, "loss": 3.3248, "step": 336950 }, { "epoch": 2.9760327805153746, "grad_norm": 3.681575059890747, "learning_rate": 3.540800975375783e-05, "loss": 2.1629, "step": 337000 }, { "epoch": 2.9764743284056587, "grad_norm": 4.049254417419434, "learning_rate": 3.5394719077511836e-05, "loss": 3.0838, "step": 337050 }, { "epoch": 2.976915876295943, "grad_norm": 1.0715888738632202, "learning_rate": 3.538142952939992e-05, "loss": 2.9383, "step": 337100 }, { "epoch": 2.977357424186227, "grad_norm": 5.504948139190674, "learning_rate": 3.536814111044862e-05, "loss": 2.9823, "step": 337150 }, { "epoch": 2.9777989720765117, "grad_norm": 2.016016721725464, "learning_rate": 3.535485382168434e-05, "loss": 3.1502, "step": 337200 }, { "epoch": 2.9782405199667954, "grad_norm": 1.062463402748108, "learning_rate": 3.534156766413343e-05, "loss": 3.1203, "step": 337250 }, { "epoch": 2.97868206785708, "grad_norm": 4.263669013977051, "learning_rate": 3.532828263882211e-05, "loss": 3.251, "step": 337300 }, { "epoch": 2.9791236157473637, "grad_norm": 2.268017292022705, "learning_rate": 3.531499874677654e-05, "loss": 3.3004, "step": 337350 }, { "epoch": 2.9795651636376483, "grad_norm": 4.288676738739014, "learning_rate": 3.530171598902281e-05, "loss": 3.2876, "step": 337400 }, { "epoch": 2.980006711527932, "grad_norm": 5.334170341491699, "learning_rate": 3.528843436658685e-05, "loss": 2.9671, "step": 337450 }, { "epoch": 2.9804482594182167, "grad_norm": 0.9075680375099182, "learning_rate": 3.527515388049459e-05, "loss": 2.769, "step": 337500 }, { "epoch": 2.980889807308501, "grad_norm": 3.4860658645629883, "learning_rate": 3.526187453177186e-05, "loss": 3.2698, "step": 337550 }, { "epoch": 2.981331355198785, "grad_norm": 1.241734504699707, "learning_rate": 3.524859632144433e-05, "loss": 2.6864, "step": 337600 }, { "epoch": 2.981772903089069, "grad_norm": 2.691862106323242, "learning_rate": 3.523531925053764e-05, "loss": 2.7246, "step": 337650 }, { "epoch": 2.9822144509793533, "grad_norm": 6.263662815093994, "learning_rate": 3.5222043320077346e-05, "loss": 2.8916, "step": 337700 }, { "epoch": 2.9826559988696375, "grad_norm": 2.4993066787719727, "learning_rate": 3.520876853108889e-05, "loss": 3.4945, "step": 337750 }, { "epoch": 2.9830975467599217, "grad_norm": 5.689305305480957, "learning_rate": 3.519549488459763e-05, "loss": 3.4272, "step": 337800 }, { "epoch": 2.983539094650206, "grad_norm": 3.942241907119751, "learning_rate": 3.518222238162885e-05, "loss": 3.0031, "step": 337850 }, { "epoch": 2.98398064254049, "grad_norm": 1.9630647897720337, "learning_rate": 3.5168951023207755e-05, "loss": 3.5461, "step": 337900 }, { "epoch": 2.984422190430774, "grad_norm": 2.370615005493164, "learning_rate": 3.515568081035943e-05, "loss": 3.0104, "step": 337950 }, { "epoch": 2.9848637383210583, "grad_norm": 3.307875633239746, "learning_rate": 3.5142411744108863e-05, "loss": 2.9934, "step": 338000 }, { "epoch": 2.9853052862113425, "grad_norm": 6.687651634216309, "learning_rate": 3.5129143825481034e-05, "loss": 3.2434, "step": 338050 }, { "epoch": 2.9857468341016267, "grad_norm": 3.1836719512939453, "learning_rate": 3.511587705550071e-05, "loss": 2.9013, "step": 338100 }, { "epoch": 2.986188381991911, "grad_norm": 2.1976935863494873, "learning_rate": 3.510261143519269e-05, "loss": 3.0886, "step": 338150 }, { "epoch": 2.986629929882195, "grad_norm": 2.4968528747558594, "learning_rate": 3.5089346965581614e-05, "loss": 3.0458, "step": 338200 }, { "epoch": 2.987071477772479, "grad_norm": 2.7656733989715576, "learning_rate": 3.507608364769203e-05, "loss": 2.8007, "step": 338250 }, { "epoch": 2.9875130256627633, "grad_norm": 4.171621799468994, "learning_rate": 3.5062821482548446e-05, "loss": 2.8197, "step": 338300 }, { "epoch": 2.9879545735530475, "grad_norm": 4.135465621948242, "learning_rate": 3.5049560471175224e-05, "loss": 2.9039, "step": 338350 }, { "epoch": 2.9883961214433317, "grad_norm": 4.187822341918945, "learning_rate": 3.50363006145967e-05, "loss": 3.401, "step": 338400 }, { "epoch": 2.988837669333616, "grad_norm": 4.675668716430664, "learning_rate": 3.502304191383705e-05, "loss": 3.1885, "step": 338450 }, { "epoch": 2.9892792172239, "grad_norm": Infinity, "learning_rate": 3.501004950945507e-05, "loss": 3.178, "step": 338500 }, { "epoch": 2.989720765114184, "grad_norm": 3.7627499103546143, "learning_rate": 3.499679310023809e-05, "loss": 2.9312, "step": 338550 }, { "epoch": 2.9901623130044683, "grad_norm": 3.8512537479400635, "learning_rate": 3.4983537849891645e-05, "loss": 3.0796, "step": 338600 }, { "epoch": 2.9906038608947525, "grad_norm": 3.7291054725646973, "learning_rate": 3.497028375943956e-05, "loss": 3.0882, "step": 338650 }, { "epoch": 2.9910454087850367, "grad_norm": 3.011197566986084, "learning_rate": 3.495703082990559e-05, "loss": 3.1023, "step": 338700 }, { "epoch": 2.9914869566753213, "grad_norm": 1.1399353742599487, "learning_rate": 3.4943779062313434e-05, "loss": 3.2228, "step": 338750 }, { "epoch": 2.991928504565605, "grad_norm": 10.077902793884277, "learning_rate": 3.493052845768668e-05, "loss": 3.2171, "step": 338800 }, { "epoch": 2.9923700524558896, "grad_norm": 2.3515779972076416, "learning_rate": 3.491727901704882e-05, "loss": 2.9235, "step": 338850 }, { "epoch": 2.9928116003461733, "grad_norm": 2.5891826152801514, "learning_rate": 3.4904030741423245e-05, "loss": 3.1195, "step": 338900 }, { "epoch": 2.993253148236458, "grad_norm": 2.2701058387756348, "learning_rate": 3.48907836318333e-05, "loss": 3.3431, "step": 338950 }, { "epoch": 2.9936946961267417, "grad_norm": 4.311112880706787, "learning_rate": 3.487753768930217e-05, "loss": 2.7791, "step": 339000 }, { "epoch": 2.9936946961267417, "eval_asr_loss": 0.9143162916007777, "eval_loss": 2.7643375396728516, "eval_runtime": 20.7875, "eval_samples_per_second": 36.945, "eval_steps_per_second": 9.236, "eval_tts_loss": 6.036369393999241, "step": 339000 }, { "epoch": 2.9941362440170263, "grad_norm": 4.269088268280029, "learning_rate": 3.486429291485305e-05, "loss": 3.1691, "step": 339050 }, { "epoch": 2.99457779190731, "grad_norm": 2.0164313316345215, "learning_rate": 3.485104930950895e-05, "loss": 3.0638, "step": 339100 }, { "epoch": 2.9950193397975946, "grad_norm": 7.095493316650391, "learning_rate": 3.483780687429282e-05, "loss": 2.7434, "step": 339150 }, { "epoch": 2.9954608876878788, "grad_norm": 1.5128381252288818, "learning_rate": 3.482456561022754e-05, "loss": 2.8946, "step": 339200 }, { "epoch": 2.995902435578163, "grad_norm": 2.1045944690704346, "learning_rate": 3.4811325518335895e-05, "loss": 3.2224, "step": 339250 }, { "epoch": 2.996343983468447, "grad_norm": 4.340090751647949, "learning_rate": 3.4798086599640546e-05, "loss": 3.1389, "step": 339300 }, { "epoch": 2.9967855313587313, "grad_norm": 5.166468143463135, "learning_rate": 3.478484885516409e-05, "loss": 2.8694, "step": 339350 }, { "epoch": 2.9972270792490154, "grad_norm": 2.4758317470550537, "learning_rate": 3.4771612285929047e-05, "loss": 3.0494, "step": 339400 }, { "epoch": 2.9976686271392996, "grad_norm": 6.358440399169922, "learning_rate": 3.475837689295783e-05, "loss": 3.0195, "step": 339450 }, { "epoch": 2.9981101750295838, "grad_norm": 5.049437999725342, "learning_rate": 3.474514267727275e-05, "loss": 3.1562, "step": 339500 }, { "epoch": 2.998551722919868, "grad_norm": 8.90450382232666, "learning_rate": 3.4731909639896024e-05, "loss": 2.8448, "step": 339550 }, { "epoch": 2.998993270810152, "grad_norm": 6.742346286773682, "learning_rate": 3.4718677781849826e-05, "loss": 3.8313, "step": 339600 }, { "epoch": 2.9994348187004363, "grad_norm": 3.848670244216919, "learning_rate": 3.470544710415619e-05, "loss": 2.9512, "step": 339650 }, { "epoch": 2.9998763665907204, "grad_norm": 3.5508828163146973, "learning_rate": 3.469221760783705e-05, "loss": 2.7754, "step": 339700 }, { "epoch": 3.0003179144810046, "grad_norm": 1.7979129552841187, "learning_rate": 3.467898929391433e-05, "loss": 3.2242, "step": 339750 }, { "epoch": 3.0007594623712888, "grad_norm": 2.8769354820251465, "learning_rate": 3.466576216340973e-05, "loss": 3.0281, "step": 339800 }, { "epoch": 3.001201010261573, "grad_norm": 1.7439160346984863, "learning_rate": 3.4652536217344994e-05, "loss": 2.7766, "step": 339850 }, { "epoch": 3.001642558151857, "grad_norm": 2.760795831680298, "learning_rate": 3.463931145674168e-05, "loss": 2.4944, "step": 339900 }, { "epoch": 3.0020841060421413, "grad_norm": 3.13797926902771, "learning_rate": 3.462608788262132e-05, "loss": 3.1212, "step": 339950 }, { "epoch": 3.0025256539324254, "grad_norm": 2.425107717514038, "learning_rate": 3.4612865496005295e-05, "loss": 3.3062, "step": 340000 }, { "epoch": 3.0029672018227096, "grad_norm": 3.4061877727508545, "learning_rate": 3.4599644297914925e-05, "loss": 2.9295, "step": 340050 }, { "epoch": 3.0034087497129938, "grad_norm": 3.4832353591918945, "learning_rate": 3.4586424289371456e-05, "loss": 2.7995, "step": 340100 }, { "epoch": 3.003850297603278, "grad_norm": 9.076299667358398, "learning_rate": 3.457320547139598e-05, "loss": 2.8342, "step": 340150 }, { "epoch": 3.004291845493562, "grad_norm": 2.9316422939300537, "learning_rate": 3.455998784500959e-05, "loss": 2.8499, "step": 340200 }, { "epoch": 3.0047333933838463, "grad_norm": 3.553391456604004, "learning_rate": 3.454677141123322e-05, "loss": 2.6256, "step": 340250 }, { "epoch": 3.0051749412741304, "grad_norm": 2.497593641281128, "learning_rate": 3.4533556171087696e-05, "loss": 2.8912, "step": 340300 }, { "epoch": 3.0056164891644146, "grad_norm": 2.031536102294922, "learning_rate": 3.452034212559381e-05, "loss": 2.6302, "step": 340350 }, { "epoch": 3.0060580370546988, "grad_norm": 4.606949329376221, "learning_rate": 3.450712927577224e-05, "loss": 2.8281, "step": 340400 }, { "epoch": 3.006499584944983, "grad_norm": 2.953679084777832, "learning_rate": 3.449391762264356e-05, "loss": 2.7283, "step": 340450 }, { "epoch": 3.006941132835267, "grad_norm": 4.270386219024658, "learning_rate": 3.448070716722823e-05, "loss": 3.0918, "step": 340500 }, { "epoch": 3.0073826807255517, "grad_norm": 6.082884788513184, "learning_rate": 3.4467497910546696e-05, "loss": 2.684, "step": 340550 }, { "epoch": 3.007824228615836, "grad_norm": 2.7726080417633057, "learning_rate": 3.445428985361925e-05, "loss": 2.7404, "step": 340600 }, { "epoch": 3.00826577650612, "grad_norm": 4.030627250671387, "learning_rate": 3.4441082997466055e-05, "loss": 2.8429, "step": 340650 }, { "epoch": 3.008707324396404, "grad_norm": 3.4186503887176514, "learning_rate": 3.442787734310727e-05, "loss": 2.6454, "step": 340700 }, { "epoch": 3.0091488722866884, "grad_norm": 2.2268149852752686, "learning_rate": 3.441467289156293e-05, "loss": 3.2154, "step": 340750 }, { "epoch": 3.0095904201769725, "grad_norm": 4.796247482299805, "learning_rate": 3.4401469643852946e-05, "loss": 2.9208, "step": 340800 }, { "epoch": 3.0100319680672567, "grad_norm": 3.756406784057617, "learning_rate": 3.4388267600997144e-05, "loss": 2.9519, "step": 340850 }, { "epoch": 3.010473515957541, "grad_norm": 2.958223342895508, "learning_rate": 3.4375066764015305e-05, "loss": 2.8313, "step": 340900 }, { "epoch": 3.010915063847825, "grad_norm": 3.797745704650879, "learning_rate": 3.436186713392705e-05, "loss": 2.6034, "step": 340950 }, { "epoch": 3.011356611738109, "grad_norm": 2.594397783279419, "learning_rate": 3.434866871175195e-05, "loss": 3.2186, "step": 341000 }, { "epoch": 3.0117981596283934, "grad_norm": 3.05829119682312, "learning_rate": 3.4335471498509495e-05, "loss": 3.3422, "step": 341050 }, { "epoch": 3.0122397075186775, "grad_norm": 5.325092315673828, "learning_rate": 3.432227549521901e-05, "loss": 3.2145, "step": 341100 }, { "epoch": 3.0126812554089617, "grad_norm": 4.618200302124023, "learning_rate": 3.430908070289981e-05, "loss": 2.9108, "step": 341150 }, { "epoch": 3.013122803299246, "grad_norm": 4.21722412109375, "learning_rate": 3.429588712257106e-05, "loss": 3.2808, "step": 341200 }, { "epoch": 3.01356435118953, "grad_norm": 4.4747772216796875, "learning_rate": 3.4282694755251886e-05, "loss": 2.9904, "step": 341250 }, { "epoch": 3.014005899079814, "grad_norm": 5.449541091918945, "learning_rate": 3.426950360196124e-05, "loss": 2.8178, "step": 341300 }, { "epoch": 3.0144474469700984, "grad_norm": 5.118526458740234, "learning_rate": 3.425631366371806e-05, "loss": 3.2856, "step": 341350 }, { "epoch": 3.0148889948603825, "grad_norm": 4.1051836013793945, "learning_rate": 3.424312494154116e-05, "loss": 2.957, "step": 341400 }, { "epoch": 3.0153305427506667, "grad_norm": 2.2975573539733887, "learning_rate": 3.422993743644922e-05, "loss": 3.1286, "step": 341450 }, { "epoch": 3.015772090640951, "grad_norm": 3.1243557929992676, "learning_rate": 3.421675114946089e-05, "loss": 3.0867, "step": 341500 }, { "epoch": 3.016213638531235, "grad_norm": 2.8947739601135254, "learning_rate": 3.420356608159471e-05, "loss": 2.9327, "step": 341550 }, { "epoch": 3.016655186421519, "grad_norm": 1.7662822008132935, "learning_rate": 3.41903822338691e-05, "loss": 3.2615, "step": 341600 }, { "epoch": 3.0170967343118034, "grad_norm": 4.569444179534912, "learning_rate": 3.417719960730238e-05, "loss": 2.4877, "step": 341650 }, { "epoch": 3.0175382822020875, "grad_norm": 5.819210529327393, "learning_rate": 3.416401820291282e-05, "loss": 2.7357, "step": 341700 }, { "epoch": 3.0179798300923717, "grad_norm": 3.6693782806396484, "learning_rate": 3.415083802171859e-05, "loss": 2.5657, "step": 341750 }, { "epoch": 3.018421377982656, "grad_norm": 3.151115894317627, "learning_rate": 3.413765906473772e-05, "loss": 2.8901, "step": 341800 }, { "epoch": 3.01886292587294, "grad_norm": 2.3335320949554443, "learning_rate": 3.412448133298817e-05, "loss": 2.8906, "step": 341850 }, { "epoch": 3.019304473763224, "grad_norm": 5.117323398590088, "learning_rate": 3.411130482748782e-05, "loss": 3.184, "step": 341900 }, { "epoch": 3.0197460216535084, "grad_norm": 4.38237190246582, "learning_rate": 3.4098129549254446e-05, "loss": 2.7382, "step": 341950 }, { "epoch": 3.0201875695437925, "grad_norm": 4.244925498962402, "learning_rate": 3.40849554993057e-05, "loss": 3.0178, "step": 342000 }, { "epoch": 3.0201875695437925, "eval_asr_loss": 0.9414155755330046, "eval_loss": 2.769742012023926, "eval_runtime": 20.4248, "eval_samples_per_second": 37.601, "eval_steps_per_second": 9.4, "eval_tts_loss": 6.1684309339539745, "step": 342000 }, { "epoch": 3.0206291174340767, "grad_norm": 3.080723285675049, "learning_rate": 3.407178267865922e-05, "loss": 3.2291, "step": 342050 }, { "epoch": 3.021070665324361, "grad_norm": 5.295591354370117, "learning_rate": 3.4058611088332434e-05, "loss": 2.6876, "step": 342100 }, { "epoch": 3.0215122132146455, "grad_norm": 5.445025444030762, "learning_rate": 3.404544072934277e-05, "loss": 2.5886, "step": 342150 }, { "epoch": 3.0219537611049296, "grad_norm": 1.7605077028274536, "learning_rate": 3.403227160270753e-05, "loss": 2.9714, "step": 342200 }, { "epoch": 3.022395308995214, "grad_norm": 3.076416492462158, "learning_rate": 3.4019103709443886e-05, "loss": 2.4613, "step": 342250 }, { "epoch": 3.022836856885498, "grad_norm": 4.947652339935303, "learning_rate": 3.4005937050568965e-05, "loss": 2.5906, "step": 342300 }, { "epoch": 3.023278404775782, "grad_norm": 3.1888465881347656, "learning_rate": 3.39927716270998e-05, "loss": 2.6613, "step": 342350 }, { "epoch": 3.0237199526660663, "grad_norm": 2.2888290882110596, "learning_rate": 3.397960744005328e-05, "loss": 2.4063, "step": 342400 }, { "epoch": 3.0241615005563505, "grad_norm": 0.6149783134460449, "learning_rate": 3.396644449044624e-05, "loss": 2.7741, "step": 342450 }, { "epoch": 3.0246030484466346, "grad_norm": 5.626645565032959, "learning_rate": 3.3953282779295394e-05, "loss": 2.8549, "step": 342500 }, { "epoch": 3.025044596336919, "grad_norm": 4.433629989624023, "learning_rate": 3.39401223076174e-05, "loss": 3.0883, "step": 342550 }, { "epoch": 3.025486144227203, "grad_norm": 3.5722925662994385, "learning_rate": 3.392696307642875e-05, "loss": 3.0996, "step": 342600 }, { "epoch": 3.025927692117487, "grad_norm": 2.0938949584960938, "learning_rate": 3.391380508674591e-05, "loss": 2.7067, "step": 342650 }, { "epoch": 3.0263692400077713, "grad_norm": 7.4776482582092285, "learning_rate": 3.390064833958524e-05, "loss": 2.6586, "step": 342700 }, { "epoch": 3.0268107878980555, "grad_norm": 4.676374435424805, "learning_rate": 3.388749283596296e-05, "loss": 3.0858, "step": 342750 }, { "epoch": 3.0272523357883396, "grad_norm": 2.949340343475342, "learning_rate": 3.3874338576895213e-05, "loss": 3.1941, "step": 342800 }, { "epoch": 3.027693883678624, "grad_norm": 4.111240863800049, "learning_rate": 3.386118556339807e-05, "loss": 2.6496, "step": 342850 }, { "epoch": 3.028135431568908, "grad_norm": 3.6584556102752686, "learning_rate": 3.384803379648751e-05, "loss": 3.2226, "step": 342900 }, { "epoch": 3.028576979459192, "grad_norm": 2.1798222064971924, "learning_rate": 3.3834883277179366e-05, "loss": 2.6774, "step": 342950 }, { "epoch": 3.0290185273494763, "grad_norm": 3.0020081996917725, "learning_rate": 3.3821734006489405e-05, "loss": 3.2228, "step": 343000 }, { "epoch": 3.0294600752397605, "grad_norm": 2.4037411212921143, "learning_rate": 3.380858598543332e-05, "loss": 2.4186, "step": 343050 }, { "epoch": 3.0299016231300446, "grad_norm": 5.278083801269531, "learning_rate": 3.379543921502665e-05, "loss": 2.9058, "step": 343100 }, { "epoch": 3.030343171020329, "grad_norm": 3.8241186141967773, "learning_rate": 3.378229369628489e-05, "loss": 3.1747, "step": 343150 }, { "epoch": 3.030784718910613, "grad_norm": 2.97617244720459, "learning_rate": 3.3769149430223446e-05, "loss": 2.8271, "step": 343200 }, { "epoch": 3.031226266800897, "grad_norm": 6.677113056182861, "learning_rate": 3.3756006417857545e-05, "loss": 2.5941, "step": 343250 }, { "epoch": 3.0316678146911813, "grad_norm": 1.9429030418395996, "learning_rate": 3.3742864660202416e-05, "loss": 2.8496, "step": 343300 }, { "epoch": 3.0321093625814655, "grad_norm": 2.994889259338379, "learning_rate": 3.372972415827314e-05, "loss": 2.9254, "step": 343350 }, { "epoch": 3.0325509104717496, "grad_norm": 5.150854110717773, "learning_rate": 3.3716584913084696e-05, "loss": 2.9277, "step": 343400 }, { "epoch": 3.032992458362034, "grad_norm": 3.2355942726135254, "learning_rate": 3.370344692565198e-05, "loss": 2.6407, "step": 343450 }, { "epoch": 3.033434006252318, "grad_norm": 6.2583088874816895, "learning_rate": 3.3690310196989817e-05, "loss": 2.9077, "step": 343500 }, { "epoch": 3.033875554142602, "grad_norm": 4.330394744873047, "learning_rate": 3.3677174728112895e-05, "loss": 2.7617, "step": 343550 }, { "epoch": 3.0343171020328863, "grad_norm": 2.993716239929199, "learning_rate": 3.366404052003579e-05, "loss": 2.9806, "step": 343600 }, { "epoch": 3.0347586499231705, "grad_norm": 4.389292240142822, "learning_rate": 3.365090757377303e-05, "loss": 3.1016, "step": 343650 }, { "epoch": 3.035200197813455, "grad_norm": 2.646925449371338, "learning_rate": 3.363777589033904e-05, "loss": 2.7563, "step": 343700 }, { "epoch": 3.0356417457037392, "grad_norm": 1.4087272882461548, "learning_rate": 3.362464547074811e-05, "loss": 2.7021, "step": 343750 }, { "epoch": 3.0360832935940234, "grad_norm": 2.4104578495025635, "learning_rate": 3.3611516316014446e-05, "loss": 2.8354, "step": 343800 }, { "epoch": 3.0365248414843076, "grad_norm": 5.261807918548584, "learning_rate": 3.35983884271522e-05, "loss": 2.8513, "step": 343850 }, { "epoch": 3.0369663893745917, "grad_norm": 2.247666358947754, "learning_rate": 3.358526180517535e-05, "loss": 2.8253, "step": 343900 }, { "epoch": 3.037407937264876, "grad_norm": 3.9168224334716797, "learning_rate": 3.3572136451097827e-05, "loss": 2.6836, "step": 343950 }, { "epoch": 3.03784948515516, "grad_norm": 3.3591697216033936, "learning_rate": 3.3559012365933474e-05, "loss": 2.7707, "step": 344000 }, { "epoch": 3.0382910330454442, "grad_norm": 4.223583698272705, "learning_rate": 3.354588955069599e-05, "loss": 3.0334, "step": 344050 }, { "epoch": 3.0387325809357284, "grad_norm": 3.911597728729248, "learning_rate": 3.353276800639901e-05, "loss": 2.598, "step": 344100 }, { "epoch": 3.0391741288260126, "grad_norm": 3.6010241508483887, "learning_rate": 3.351964773405606e-05, "loss": 2.9764, "step": 344150 }, { "epoch": 3.0396156767162967, "grad_norm": 3.6039419174194336, "learning_rate": 3.350652873468059e-05, "loss": 2.732, "step": 344200 }, { "epoch": 3.040057224606581, "grad_norm": 5.471869468688965, "learning_rate": 3.3493411009285904e-05, "loss": 2.9861, "step": 344250 }, { "epoch": 3.040498772496865, "grad_norm": 1.7279754877090454, "learning_rate": 3.348029455888524e-05, "loss": 2.8829, "step": 344300 }, { "epoch": 3.0409403203871492, "grad_norm": 3.3133652210235596, "learning_rate": 3.346717938449177e-05, "loss": 2.872, "step": 344350 }, { "epoch": 3.0413818682774334, "grad_norm": 3.774657726287842, "learning_rate": 3.3454065487118456e-05, "loss": 2.8852, "step": 344400 }, { "epoch": 3.0418234161677176, "grad_norm": 7.494409561157227, "learning_rate": 3.34409528677783e-05, "loss": 2.6369, "step": 344450 }, { "epoch": 3.0422649640580017, "grad_norm": 3.8216733932495117, "learning_rate": 3.3427841527484114e-05, "loss": 2.9694, "step": 344500 }, { "epoch": 3.042706511948286, "grad_norm": 3.842785120010376, "learning_rate": 3.3414731467248644e-05, "loss": 3.1915, "step": 344550 }, { "epoch": 3.04314805983857, "grad_norm": 3.8184828758239746, "learning_rate": 3.340162268808452e-05, "loss": 2.7267, "step": 344600 }, { "epoch": 3.0435896077288542, "grad_norm": 2.4223062992095947, "learning_rate": 3.33885151910043e-05, "loss": 3.4147, "step": 344650 }, { "epoch": 3.0440311556191384, "grad_norm": 3.59983229637146, "learning_rate": 3.3375408977020433e-05, "loss": 3.2649, "step": 344700 }, { "epoch": 3.0444727035094226, "grad_norm": 3.1596648693084717, "learning_rate": 3.336230404714523e-05, "loss": 3.1226, "step": 344750 }, { "epoch": 3.0449142513997067, "grad_norm": 4.39080810546875, "learning_rate": 3.334920040239096e-05, "loss": 2.4736, "step": 344800 }, { "epoch": 3.045355799289991, "grad_norm": 7.796013832092285, "learning_rate": 3.333609804376977e-05, "loss": 2.575, "step": 344850 }, { "epoch": 3.045797347180275, "grad_norm": 4.554876804351807, "learning_rate": 3.3322996972293706e-05, "loss": 2.5298, "step": 344900 }, { "epoch": 3.0462388950705592, "grad_norm": 1.7804590463638306, "learning_rate": 3.330989718897469e-05, "loss": 2.953, "step": 344950 }, { "epoch": 3.0466804429608434, "grad_norm": 2.997663974761963, "learning_rate": 3.3296798694824614e-05, "loss": 2.79, "step": 345000 }, { "epoch": 3.0466804429608434, "eval_asr_loss": 0.9422899130333035, "eval_loss": 2.770587205886841, "eval_runtime": 20.9853, "eval_samples_per_second": 36.597, "eval_steps_per_second": 9.149, "eval_tts_loss": 6.186337430544469, "step": 345000 }, { "epoch": 3.0471219908511276, "grad_norm": 2.7708795070648193, "learning_rate": 3.328370149085518e-05, "loss": 3.3187, "step": 345050 }, { "epoch": 3.0475635387414117, "grad_norm": 3.554488182067871, "learning_rate": 3.327060557807806e-05, "loss": 2.5978, "step": 345100 }, { "epoch": 3.048005086631696, "grad_norm": 3.505568027496338, "learning_rate": 3.325751095750481e-05, "loss": 2.5108, "step": 345150 }, { "epoch": 3.04844663452198, "grad_norm": 3.0691869258880615, "learning_rate": 3.3244417630146844e-05, "loss": 3.3176, "step": 345200 }, { "epoch": 3.0488881824122642, "grad_norm": 10.595149040222168, "learning_rate": 3.323132559701555e-05, "loss": 2.5336, "step": 345250 }, { "epoch": 3.0493297303025484, "grad_norm": 3.1379215717315674, "learning_rate": 3.3218234859122147e-05, "loss": 3.2141, "step": 345300 }, { "epoch": 3.049771278192833, "grad_norm": 1.2156410217285156, "learning_rate": 3.3205145417477815e-05, "loss": 2.4383, "step": 345350 }, { "epoch": 3.050212826083117, "grad_norm": 4.518272876739502, "learning_rate": 3.319205727309356e-05, "loss": 2.3641, "step": 345400 }, { "epoch": 3.0506543739734013, "grad_norm": 4.2563018798828125, "learning_rate": 3.317897042698036e-05, "loss": 2.9236, "step": 345450 }, { "epoch": 3.0510959218636855, "grad_norm": 1.2969986200332642, "learning_rate": 3.316588488014908e-05, "loss": 2.9425, "step": 345500 }, { "epoch": 3.0515374697539697, "grad_norm": 5.326292991638184, "learning_rate": 3.315280063361041e-05, "loss": 3.0899, "step": 345550 }, { "epoch": 3.051979017644254, "grad_norm": 2.897663116455078, "learning_rate": 3.313971768837504e-05, "loss": 2.6797, "step": 345600 }, { "epoch": 3.052420565534538, "grad_norm": 2.5540523529052734, "learning_rate": 3.3126636045453523e-05, "loss": 3.1186, "step": 345650 }, { "epoch": 3.052862113424822, "grad_norm": 4.0318803787231445, "learning_rate": 3.311355570585628e-05, "loss": 3.1116, "step": 345700 }, { "epoch": 3.0533036613151063, "grad_norm": 5.47548246383667, "learning_rate": 3.310047667059366e-05, "loss": 3.1142, "step": 345750 }, { "epoch": 3.0537452092053905, "grad_norm": 1.7209810018539429, "learning_rate": 3.3087398940675915e-05, "loss": 2.8031, "step": 345800 }, { "epoch": 3.0541867570956747, "grad_norm": 6.121496200561523, "learning_rate": 3.30743225171132e-05, "loss": 3.135, "step": 345850 }, { "epoch": 3.054628304985959, "grad_norm": 3.8786704540252686, "learning_rate": 3.3061508890420784e-05, "loss": 2.9703, "step": 345900 }, { "epoch": 3.055069852876243, "grad_norm": 3.957517385482788, "learning_rate": 3.304843505642072e-05, "loss": 2.9981, "step": 345950 }, { "epoch": 3.055511400766527, "grad_norm": 5.411700248718262, "learning_rate": 3.303536253178531e-05, "loss": 3.0025, "step": 346000 }, { "epoch": 3.0559529486568113, "grad_norm": 8.696145057678223, "learning_rate": 3.302229131752429e-05, "loss": 2.71, "step": 346050 }, { "epoch": 3.0563944965470955, "grad_norm": 7.089808940887451, "learning_rate": 3.300922141464729e-05, "loss": 2.7411, "step": 346100 }, { "epoch": 3.0568360444373797, "grad_norm": 5.090502738952637, "learning_rate": 3.299615282416387e-05, "loss": 2.7912, "step": 346150 }, { "epoch": 3.057277592327664, "grad_norm": 1.6336640119552612, "learning_rate": 3.298308554708346e-05, "loss": 2.753, "step": 346200 }, { "epoch": 3.057719140217948, "grad_norm": 3.3283755779266357, "learning_rate": 3.297001958441537e-05, "loss": 2.3768, "step": 346250 }, { "epoch": 3.058160688108232, "grad_norm": 2.56878924369812, "learning_rate": 3.295695493716886e-05, "loss": 2.7871, "step": 346300 }, { "epoch": 3.0586022359985163, "grad_norm": 6.040671348571777, "learning_rate": 3.294389160635307e-05, "loss": 2.8128, "step": 346350 }, { "epoch": 3.0590437838888005, "grad_norm": 5.563143730163574, "learning_rate": 3.2930829592977035e-05, "loss": 2.9499, "step": 346400 }, { "epoch": 3.0594853317790847, "grad_norm": 6.277194976806641, "learning_rate": 3.291776889804965e-05, "loss": 2.7023, "step": 346450 }, { "epoch": 3.059926879669369, "grad_norm": 2.8235936164855957, "learning_rate": 3.2904709522579763e-05, "loss": 2.8925, "step": 346500 }, { "epoch": 3.060368427559653, "grad_norm": 5.657355308532715, "learning_rate": 3.289165146757612e-05, "loss": 2.9732, "step": 346550 }, { "epoch": 3.060809975449937, "grad_norm": 4.480293273925781, "learning_rate": 3.287859473404732e-05, "loss": 3.4109, "step": 346600 }, { "epoch": 3.0612515233402213, "grad_norm": 1.9854350090026855, "learning_rate": 3.286553932300188e-05, "loss": 3.0779, "step": 346650 }, { "epoch": 3.0616930712305055, "grad_norm": 2.462419033050537, "learning_rate": 3.2852485235448254e-05, "loss": 3.2011, "step": 346700 }, { "epoch": 3.0621346191207897, "grad_norm": 1.3553026914596558, "learning_rate": 3.2839432472394724e-05, "loss": 3.2195, "step": 346750 }, { "epoch": 3.062576167011074, "grad_norm": 3.099388837814331, "learning_rate": 3.2826381034849506e-05, "loss": 3.0952, "step": 346800 }, { "epoch": 3.063017714901358, "grad_norm": 3.866703510284424, "learning_rate": 3.281333092382073e-05, "loss": 2.8885, "step": 346850 }, { "epoch": 3.0634592627916426, "grad_norm": 2.515897035598755, "learning_rate": 3.280028214031642e-05, "loss": 2.6497, "step": 346900 }, { "epoch": 3.0639008106819268, "grad_norm": 4.203222751617432, "learning_rate": 3.278723468534446e-05, "loss": 3.1339, "step": 346950 }, { "epoch": 3.064342358572211, "grad_norm": 4.6410322189331055, "learning_rate": 3.277418855991265e-05, "loss": 3.1234, "step": 347000 }, { "epoch": 3.064783906462495, "grad_norm": 1.3641115427017212, "learning_rate": 3.276114376502871e-05, "loss": 2.6546, "step": 347050 }, { "epoch": 3.0652254543527793, "grad_norm": 5.405045509338379, "learning_rate": 3.2748100301700226e-05, "loss": 3.0571, "step": 347100 }, { "epoch": 3.0656670022430634, "grad_norm": 3.48565411567688, "learning_rate": 3.27350581709347e-05, "loss": 2.4189, "step": 347150 }, { "epoch": 3.0661085501333476, "grad_norm": 5.0231032371521, "learning_rate": 3.272201737373954e-05, "loss": 2.8679, "step": 347200 }, { "epoch": 3.0665500980236318, "grad_norm": 1.8874601125717163, "learning_rate": 3.270897791112202e-05, "loss": 3.182, "step": 347250 }, { "epoch": 3.066991645913916, "grad_norm": 4.711459159851074, "learning_rate": 3.269593978408933e-05, "loss": 2.9028, "step": 347300 }, { "epoch": 3.0674331938042, "grad_norm": 3.9696052074432373, "learning_rate": 3.2682902993648554e-05, "loss": 3.1232, "step": 347350 }, { "epoch": 3.0678747416944843, "grad_norm": 3.42094087600708, "learning_rate": 3.2669867540806684e-05, "loss": 2.8912, "step": 347400 }, { "epoch": 3.0683162895847684, "grad_norm": 1.8919652700424194, "learning_rate": 3.265683342657059e-05, "loss": 2.806, "step": 347450 }, { "epoch": 3.0687578374750526, "grad_norm": 2.3823795318603516, "learning_rate": 3.2643800651947044e-05, "loss": 2.7979, "step": 347500 }, { "epoch": 3.0691993853653368, "grad_norm": 5.4157538414001465, "learning_rate": 3.263076921794274e-05, "loss": 3.385, "step": 347550 }, { "epoch": 3.069640933255621, "grad_norm": 6.286477565765381, "learning_rate": 3.261799971425733e-05, "loss": 3.2462, "step": 347600 }, { "epoch": 3.070082481145905, "grad_norm": 4.798492908477783, "learning_rate": 3.260497093764855e-05, "loss": 2.7525, "step": 347650 }, { "epoch": 3.0705240290361893, "grad_norm": 4.513412952423096, "learning_rate": 3.259194350465825e-05, "loss": 2.7684, "step": 347700 }, { "epoch": 3.0709655769264734, "grad_norm": 2.9037344455718994, "learning_rate": 3.257891741629272e-05, "loss": 2.9593, "step": 347750 }, { "epoch": 3.0714071248167576, "grad_norm": 4.3907294273376465, "learning_rate": 3.2565892673558076e-05, "loss": 3.1006, "step": 347800 }, { "epoch": 3.0718486727070418, "grad_norm": 1.9207698106765747, "learning_rate": 3.255286927746039e-05, "loss": 2.9288, "step": 347850 }, { "epoch": 3.072290220597326, "grad_norm": 3.618910074234009, "learning_rate": 3.2539847229005594e-05, "loss": 3.1283, "step": 347900 }, { "epoch": 3.07273176848761, "grad_norm": 4.257720947265625, "learning_rate": 3.252682652919956e-05, "loss": 2.8105, "step": 347950 }, { "epoch": 3.0731733163778943, "grad_norm": 5.0210771560668945, "learning_rate": 3.251380717904799e-05, "loss": 2.9653, "step": 348000 }, { "epoch": 3.0731733163778943, "eval_asr_loss": 0.9503913109361727, "eval_loss": 2.7658588886260986, "eval_runtime": 20.8793, "eval_samples_per_second": 36.783, "eval_steps_per_second": 9.196, "eval_tts_loss": 6.167163698157397, "step": 348000 }, { "epoch": 3.0736148642681784, "grad_norm": 3.860936164855957, "learning_rate": 3.2500789179556554e-05, "loss": 2.8208, "step": 348050 }, { "epoch": 3.0740564121584626, "grad_norm": 2.9489119052886963, "learning_rate": 3.248777253173077e-05, "loss": 2.7173, "step": 348100 }, { "epoch": 3.0744979600487468, "grad_norm": 6.876544952392578, "learning_rate": 3.2474757236576036e-05, "loss": 2.9321, "step": 348150 }, { "epoch": 3.074939507939031, "grad_norm": 4.69642448425293, "learning_rate": 3.246174329509771e-05, "loss": 2.8068, "step": 348200 }, { "epoch": 3.075381055829315, "grad_norm": 3.2072317600250244, "learning_rate": 3.244873070830101e-05, "loss": 3.2724, "step": 348250 }, { "epoch": 3.0758226037195993, "grad_norm": 4.087865829467773, "learning_rate": 3.2435719477191004e-05, "loss": 3.1553, "step": 348300 }, { "epoch": 3.0762641516098834, "grad_norm": 3.2028586864471436, "learning_rate": 3.242270960277273e-05, "loss": 3.2671, "step": 348350 }, { "epoch": 3.0767056995001676, "grad_norm": 8.633116722106934, "learning_rate": 3.2409701086051106e-05, "loss": 3.0251, "step": 348400 }, { "epoch": 3.0771472473904518, "grad_norm": 2.704092025756836, "learning_rate": 3.2396693928030895e-05, "loss": 2.3821, "step": 348450 }, { "epoch": 3.077588795280736, "grad_norm": 2.3375139236450195, "learning_rate": 3.238368812971681e-05, "loss": 3.0685, "step": 348500 }, { "epoch": 3.0780303431710205, "grad_norm": 4.810919761657715, "learning_rate": 3.237068369211343e-05, "loss": 3.3314, "step": 348550 }, { "epoch": 3.0784718910613047, "grad_norm": 1.4148460626602173, "learning_rate": 3.235768061622526e-05, "loss": 2.8496, "step": 348600 }, { "epoch": 3.078913438951589, "grad_norm": 4.941149711608887, "learning_rate": 3.234467890305665e-05, "loss": 2.8447, "step": 348650 }, { "epoch": 3.079354986841873, "grad_norm": 3.6164140701293945, "learning_rate": 3.233167855361188e-05, "loss": 3.0701, "step": 348700 }, { "epoch": 3.079796534732157, "grad_norm": 1.4390915632247925, "learning_rate": 3.2318679568895126e-05, "loss": 2.9239, "step": 348750 }, { "epoch": 3.0802380826224414, "grad_norm": 3.9035675525665283, "learning_rate": 3.230568194991043e-05, "loss": 3.2655, "step": 348800 }, { "epoch": 3.0806796305127255, "grad_norm": 3.204357862472534, "learning_rate": 3.229268569766177e-05, "loss": 2.7997, "step": 348850 }, { "epoch": 3.0811211784030097, "grad_norm": 3.6101906299591064, "learning_rate": 3.2279690813153e-05, "loss": 3.3375, "step": 348900 }, { "epoch": 3.081562726293294, "grad_norm": 3.478895902633667, "learning_rate": 3.226669729738782e-05, "loss": 3.4767, "step": 348950 }, { "epoch": 3.082004274183578, "grad_norm": 4.9733781814575195, "learning_rate": 3.225370515136992e-05, "loss": 3.3931, "step": 349000 }, { "epoch": 3.082445822073862, "grad_norm": 5.314770221710205, "learning_rate": 3.224071437610279e-05, "loss": 2.4762, "step": 349050 }, { "epoch": 3.0828873699641464, "grad_norm": 3.446488618850708, "learning_rate": 3.222772497258991e-05, "loss": 3.154, "step": 349100 }, { "epoch": 3.0833289178544305, "grad_norm": 5.472344875335693, "learning_rate": 3.2214736941834555e-05, "loss": 2.9635, "step": 349150 }, { "epoch": 3.0837704657447147, "grad_norm": 5.441402912139893, "learning_rate": 3.220175028483996e-05, "loss": 2.5126, "step": 349200 }, { "epoch": 3.084212013634999, "grad_norm": 5.292456150054932, "learning_rate": 3.2188765002609244e-05, "loss": 2.8938, "step": 349250 }, { "epoch": 3.084653561525283, "grad_norm": 7.39217472076416, "learning_rate": 3.2175781096145374e-05, "loss": 2.7994, "step": 349300 }, { "epoch": 3.085095109415567, "grad_norm": 5.801981449127197, "learning_rate": 3.216279856645127e-05, "loss": 2.5673, "step": 349350 }, { "epoch": 3.0855366573058514, "grad_norm": 4.808537483215332, "learning_rate": 3.214981741452974e-05, "loss": 3.2306, "step": 349400 }, { "epoch": 3.0859782051961355, "grad_norm": 5.13274621963501, "learning_rate": 3.2136837641383446e-05, "loss": 2.7777, "step": 349450 }, { "epoch": 3.0864197530864197, "grad_norm": 4.779294490814209, "learning_rate": 3.212385924801495e-05, "loss": 3.1023, "step": 349500 }, { "epoch": 3.086861300976704, "grad_norm": 4.338204860687256, "learning_rate": 3.2110882235426765e-05, "loss": 2.6734, "step": 349550 }, { "epoch": 3.087302848866988, "grad_norm": 6.859677314758301, "learning_rate": 3.2097906604621225e-05, "loss": 2.6787, "step": 349600 }, { "epoch": 3.087744396757272, "grad_norm": 2.4765074253082275, "learning_rate": 3.208493235660058e-05, "loss": 3.0663, "step": 349650 }, { "epoch": 3.0881859446475564, "grad_norm": 1.988755702972412, "learning_rate": 3.2071959492367e-05, "loss": 3.1284, "step": 349700 }, { "epoch": 3.0886274925378405, "grad_norm": 6.475376605987549, "learning_rate": 3.2058988012922544e-05, "loss": 2.9947, "step": 349750 }, { "epoch": 3.0890690404281247, "grad_norm": 2.155597686767578, "learning_rate": 3.2046017919269114e-05, "loss": 2.3798, "step": 349800 }, { "epoch": 3.089510588318409, "grad_norm": 2.243779420852661, "learning_rate": 3.203304921240855e-05, "loss": 2.4903, "step": 349850 }, { "epoch": 3.089952136208693, "grad_norm": 4.6090192794799805, "learning_rate": 3.202008189334259e-05, "loss": 3.2911, "step": 349900 }, { "epoch": 3.090393684098977, "grad_norm": 4.771500587463379, "learning_rate": 3.2007115963072824e-05, "loss": 2.4319, "step": 349950 }, { "epoch": 3.0908352319892614, "grad_norm": 2.9683938026428223, "learning_rate": 3.199415142260079e-05, "loss": 2.9062, "step": 350000 }, { "epoch": 3.0912767798795455, "grad_norm": 2.50345516204834, "learning_rate": 3.198118827292788e-05, "loss": 2.8785, "step": 350050 }, { "epoch": 3.09171832776983, "grad_norm": 2.5344398021698, "learning_rate": 3.196822651505535e-05, "loss": 2.5736, "step": 350100 }, { "epoch": 3.0921598756601143, "grad_norm": 2.919525623321533, "learning_rate": 3.1955266149984445e-05, "loss": 2.9533, "step": 350150 }, { "epoch": 3.0926014235503985, "grad_norm": 5.201753616333008, "learning_rate": 3.194230717871619e-05, "loss": 2.9274, "step": 350200 }, { "epoch": 3.0930429714406826, "grad_norm": 1.6824486255645752, "learning_rate": 3.192934960225161e-05, "loss": 2.9629, "step": 350250 }, { "epoch": 3.093484519330967, "grad_norm": 3.777552843093872, "learning_rate": 3.191639342159152e-05, "loss": 3.0639, "step": 350300 }, { "epoch": 3.093926067221251, "grad_norm": 4.426102161407471, "learning_rate": 3.190343863773669e-05, "loss": 2.806, "step": 350350 }, { "epoch": 3.094367615111535, "grad_norm": 5.604964256286621, "learning_rate": 3.18904852516878e-05, "loss": 2.6146, "step": 350400 }, { "epoch": 3.0948091630018193, "grad_norm": 3.096320152282715, "learning_rate": 3.1877533264445324e-05, "loss": 2.7136, "step": 350450 }, { "epoch": 3.0952507108921035, "grad_norm": 5.356082916259766, "learning_rate": 3.1864582677009735e-05, "loss": 3.0473, "step": 350500 }, { "epoch": 3.0956922587823876, "grad_norm": 4.724630832672119, "learning_rate": 3.185163349038137e-05, "loss": 2.5998, "step": 350550 }, { "epoch": 3.096133806672672, "grad_norm": 4.598144054412842, "learning_rate": 3.1838685705560416e-05, "loss": 2.6651, "step": 350600 }, { "epoch": 3.096575354562956, "grad_norm": 3.1620917320251465, "learning_rate": 3.182573932354698e-05, "loss": 3.2327, "step": 350650 }, { "epoch": 3.09701690245324, "grad_norm": 3.77877140045166, "learning_rate": 3.181279434534108e-05, "loss": 2.7802, "step": 350700 }, { "epoch": 3.0974584503435243, "grad_norm": 3.4676692485809326, "learning_rate": 3.179985077194258e-05, "loss": 2.8542, "step": 350750 }, { "epoch": 3.0978999982338085, "grad_norm": 1.3197731971740723, "learning_rate": 3.1786908604351287e-05, "loss": 2.2954, "step": 350800 }, { "epoch": 3.0983415461240926, "grad_norm": 4.465755462646484, "learning_rate": 3.177396784356684e-05, "loss": 3.0374, "step": 350850 }, { "epoch": 3.098783094014377, "grad_norm": 1.8280527591705322, "learning_rate": 3.1761028490588865e-05, "loss": 2.7751, "step": 350900 }, { "epoch": 3.099224641904661, "grad_norm": 3.4070842266082764, "learning_rate": 3.1748090546416754e-05, "loss": 2.8197, "step": 350950 }, { "epoch": 3.099666189794945, "grad_norm": 3.1300137042999268, "learning_rate": 3.173515401204986e-05, "loss": 3.0363, "step": 351000 }, { "epoch": 3.099666189794945, "eval_asr_loss": 0.9420060318470422, "eval_loss": 2.776569366455078, "eval_runtime": 20.9334, "eval_samples_per_second": 36.688, "eval_steps_per_second": 9.172, "eval_tts_loss": 6.1588726025528215, "step": 351000 }, { "epoch": 3.1001077376852293, "grad_norm": 5.330318927764893, "learning_rate": 3.1722218888487473e-05, "loss": 2.784, "step": 351050 }, { "epoch": 3.1005492855755135, "grad_norm": 7.1216349601745605, "learning_rate": 3.170928517672865e-05, "loss": 3.047, "step": 351100 }, { "epoch": 3.1009908334657976, "grad_norm": 1.6443547010421753, "learning_rate": 3.169635287777247e-05, "loss": 2.7394, "step": 351150 }, { "epoch": 3.101432381356082, "grad_norm": 1.699433445930481, "learning_rate": 3.1683421992617826e-05, "loss": 3.1742, "step": 351200 }, { "epoch": 3.101873929246366, "grad_norm": 1.8129180669784546, "learning_rate": 3.167049252226349e-05, "loss": 3.1963, "step": 351250 }, { "epoch": 3.10231547713665, "grad_norm": 3.8154313564300537, "learning_rate": 3.1657564467708176e-05, "loss": 2.5783, "step": 351300 }, { "epoch": 3.1027570250269343, "grad_norm": 6.773533821105957, "learning_rate": 3.164463782995049e-05, "loss": 2.8698, "step": 351350 }, { "epoch": 3.1031985729172185, "grad_norm": 3.0845530033111572, "learning_rate": 3.163171260998887e-05, "loss": 2.9823, "step": 351400 }, { "epoch": 3.1036401208075026, "grad_norm": 3.056784152984619, "learning_rate": 3.161878880882169e-05, "loss": 3.0304, "step": 351450 }, { "epoch": 3.104081668697787, "grad_norm": 7.655436038970947, "learning_rate": 3.16058664274472e-05, "loss": 3.0428, "step": 351500 }, { "epoch": 3.104523216588071, "grad_norm": 6.544590473175049, "learning_rate": 3.159294546686357e-05, "loss": 2.6648, "step": 351550 }, { "epoch": 3.104964764478355, "grad_norm": 4.889385223388672, "learning_rate": 3.1580025928068814e-05, "loss": 3.3065, "step": 351600 }, { "epoch": 3.1054063123686393, "grad_norm": 1.8394038677215576, "learning_rate": 3.156710781206084e-05, "loss": 3.1261, "step": 351650 }, { "epoch": 3.1058478602589235, "grad_norm": 1.6806799173355103, "learning_rate": 3.155444943972242e-05, "loss": 2.7622, "step": 351700 }, { "epoch": 3.106289408149208, "grad_norm": 3.6797449588775635, "learning_rate": 3.154153414377598e-05, "loss": 3.2009, "step": 351750 }, { "epoch": 3.1067309560394922, "grad_norm": 3.3822646141052246, "learning_rate": 3.1528620273589483e-05, "loss": 2.679, "step": 351800 }, { "epoch": 3.1071725039297764, "grad_norm": 1.6956958770751953, "learning_rate": 3.1515707830160444e-05, "loss": 2.5997, "step": 351850 }, { "epoch": 3.1076140518200606, "grad_norm": 4.701713562011719, "learning_rate": 3.1502796814486237e-05, "loss": 2.9438, "step": 351900 }, { "epoch": 3.1080555997103447, "grad_norm": 4.283674240112305, "learning_rate": 3.148988722756413e-05, "loss": 3.011, "step": 351950 }, { "epoch": 3.108497147600629, "grad_norm": 1.8693910837173462, "learning_rate": 3.1476979070391256e-05, "loss": 3.0509, "step": 352000 }, { "epoch": 3.108938695490913, "grad_norm": 5.711164474487305, "learning_rate": 3.146407234396469e-05, "loss": 2.8535, "step": 352050 }, { "epoch": 3.1093802433811972, "grad_norm": 4.238368988037109, "learning_rate": 3.1451167049281364e-05, "loss": 2.8051, "step": 352100 }, { "epoch": 3.1098217912714814, "grad_norm": 4.00391960144043, "learning_rate": 3.143826318733808e-05, "loss": 2.676, "step": 352150 }, { "epoch": 3.1102633391617656, "grad_norm": 1.5560154914855957, "learning_rate": 3.1425360759131564e-05, "loss": 3.183, "step": 352200 }, { "epoch": 3.1107048870520497, "grad_norm": 2.054535150527954, "learning_rate": 3.141245976565843e-05, "loss": 2.7576, "step": 352250 }, { "epoch": 3.111146434942334, "grad_norm": 3.995537281036377, "learning_rate": 3.139956020791516e-05, "loss": 2.9126, "step": 352300 }, { "epoch": 3.111587982832618, "grad_norm": 6.163167953491211, "learning_rate": 3.138666208689813e-05, "loss": 2.7069, "step": 352350 }, { "epoch": 3.1120295307229022, "grad_norm": 2.001267671585083, "learning_rate": 3.137376540360363e-05, "loss": 3.1566, "step": 352400 }, { "epoch": 3.1124710786131864, "grad_norm": 1.6915628910064697, "learning_rate": 3.1360870159027804e-05, "loss": 2.6481, "step": 352450 }, { "epoch": 3.1129126265034706, "grad_norm": 3.2449018955230713, "learning_rate": 3.134797635416669e-05, "loss": 2.3774, "step": 352500 }, { "epoch": 3.1133541743937547, "grad_norm": 1.2595183849334717, "learning_rate": 3.133508399001623e-05, "loss": 2.9549, "step": 352550 }, { "epoch": 3.113795722284039, "grad_norm": 3.221052885055542, "learning_rate": 3.1322193067572295e-05, "loss": 3.4075, "step": 352600 }, { "epoch": 3.114237270174323, "grad_norm": 4.469902992248535, "learning_rate": 3.130930358783054e-05, "loss": 3.253, "step": 352650 }, { "epoch": 3.1146788180646072, "grad_norm": 2.5911481380462646, "learning_rate": 3.1296415551786585e-05, "loss": 2.4587, "step": 352700 }, { "epoch": 3.1151203659548914, "grad_norm": 2.2840781211853027, "learning_rate": 3.128352896043595e-05, "loss": 2.7576, "step": 352750 }, { "epoch": 3.1155619138451756, "grad_norm": 2.428226947784424, "learning_rate": 3.127064381477396e-05, "loss": 2.7466, "step": 352800 }, { "epoch": 3.1160034617354597, "grad_norm": 2.566687822341919, "learning_rate": 3.125776011579593e-05, "loss": 2.6492, "step": 352850 }, { "epoch": 3.116445009625744, "grad_norm": 8.577313423156738, "learning_rate": 3.124487786449701e-05, "loss": 3.1067, "step": 352900 }, { "epoch": 3.116886557516028, "grad_norm": 3.6604998111724854, "learning_rate": 3.123199706187223e-05, "loss": 3.3985, "step": 352950 }, { "epoch": 3.117328105406312, "grad_norm": 2.280734062194824, "learning_rate": 3.121911770891652e-05, "loss": 3.2022, "step": 353000 }, { "epoch": 3.1177696532965964, "grad_norm": 5.212052822113037, "learning_rate": 3.120623980662472e-05, "loss": 2.7977, "step": 353050 }, { "epoch": 3.1182112011868806, "grad_norm": 3.255640983581543, "learning_rate": 3.1193363355991526e-05, "loss": 2.9986, "step": 353100 }, { "epoch": 3.1186527490771647, "grad_norm": 2.8518500328063965, "learning_rate": 3.118048835801153e-05, "loss": 3.0366, "step": 353150 }, { "epoch": 3.119094296967449, "grad_norm": 1.7750985622406006, "learning_rate": 3.116761481367924e-05, "loss": 2.918, "step": 353200 }, { "epoch": 3.119535844857733, "grad_norm": 5.144649028778076, "learning_rate": 3.115474272398901e-05, "loss": 3.0216, "step": 353250 }, { "epoch": 3.1199773927480177, "grad_norm": 3.6304941177368164, "learning_rate": 3.114187208993508e-05, "loss": 2.6686, "step": 353300 }, { "epoch": 3.120418940638302, "grad_norm": 3.4906222820281982, "learning_rate": 3.112900291251162e-05, "loss": 2.8715, "step": 353350 }, { "epoch": 3.120860488528586, "grad_norm": 2.6558337211608887, "learning_rate": 3.1116135192712695e-05, "loss": 3.1914, "step": 353400 }, { "epoch": 3.12130203641887, "grad_norm": 3.7867865562438965, "learning_rate": 3.110326893153217e-05, "loss": 2.971, "step": 353450 }, { "epoch": 3.1217435843091543, "grad_norm": 9.397073745727539, "learning_rate": 3.109040412996388e-05, "loss": 2.5645, "step": 353500 }, { "epoch": 3.1221851321994385, "grad_norm": 5.991609573364258, "learning_rate": 3.1077540789001525e-05, "loss": 2.8022, "step": 353550 }, { "epoch": 3.1226266800897227, "grad_norm": 1.6126786470413208, "learning_rate": 3.10646789096387e-05, "loss": 3.0364, "step": 353600 }, { "epoch": 3.123068227980007, "grad_norm": 1.4512797594070435, "learning_rate": 3.105181849286886e-05, "loss": 2.6818, "step": 353650 }, { "epoch": 3.123509775870291, "grad_norm": 4.3649725914001465, "learning_rate": 3.103895953968534e-05, "loss": 2.8097, "step": 353700 }, { "epoch": 3.123951323760575, "grad_norm": 6.6155195236206055, "learning_rate": 3.102610205108144e-05, "loss": 2.7054, "step": 353750 }, { "epoch": 3.1243928716508593, "grad_norm": 4.338001728057861, "learning_rate": 3.101324602805026e-05, "loss": 2.8453, "step": 353800 }, { "epoch": 3.1248344195411435, "grad_norm": 4.854394912719727, "learning_rate": 3.10003914715848e-05, "loss": 2.8257, "step": 353850 }, { "epoch": 3.1252759674314277, "grad_norm": 3.6754257678985596, "learning_rate": 3.098753838267802e-05, "loss": 2.8249, "step": 353900 }, { "epoch": 3.125717515321712, "grad_norm": 5.641890525817871, "learning_rate": 3.097468676232265e-05, "loss": 3.0913, "step": 353950 }, { "epoch": 3.126159063211996, "grad_norm": 3.965660572052002, "learning_rate": 3.09618366115114e-05, "loss": 3.0326, "step": 354000 }, { "epoch": 3.126159063211996, "eval_asr_loss": 0.9451403049954312, "eval_loss": 2.762516975402832, "eval_runtime": 21.0879, "eval_samples_per_second": 36.419, "eval_steps_per_second": 9.105, "eval_tts_loss": 6.123255094105078, "step": 354000 }, { "epoch": 3.12660061110228, "grad_norm": 2.7431113719940186, "learning_rate": 3.094898793123685e-05, "loss": 3.1265, "step": 354050 }, { "epoch": 3.1270421589925643, "grad_norm": 3.0089049339294434, "learning_rate": 3.0936140722491416e-05, "loss": 2.6954, "step": 354100 }, { "epoch": 3.1274837068828485, "grad_norm": 3.0629525184631348, "learning_rate": 3.092329498626746e-05, "loss": 2.5051, "step": 354150 }, { "epoch": 3.1279252547731327, "grad_norm": 3.2791266441345215, "learning_rate": 3.09104507235572e-05, "loss": 2.9152, "step": 354200 }, { "epoch": 3.128366802663417, "grad_norm": 2.903712272644043, "learning_rate": 3.0897607935352754e-05, "loss": 2.9446, "step": 354250 }, { "epoch": 3.128808350553701, "grad_norm": 5.572776794433594, "learning_rate": 3.0884766622646106e-05, "loss": 2.8938, "step": 354300 }, { "epoch": 3.129249898443985, "grad_norm": 2.5844342708587646, "learning_rate": 3.0872183568677476e-05, "loss": 2.935, "step": 354350 }, { "epoch": 3.1296914463342693, "grad_norm": 3.1748101711273193, "learning_rate": 3.085934518038261e-05, "loss": 2.9806, "step": 354400 }, { "epoch": 3.1301329942245535, "grad_norm": 4.9311323165893555, "learning_rate": 3.084650827054105e-05, "loss": 2.6768, "step": 354450 }, { "epoch": 3.1305745421148377, "grad_norm": 3.577273368835449, "learning_rate": 3.083367284014428e-05, "loss": 2.5694, "step": 354500 }, { "epoch": 3.131016090005122, "grad_norm": 5.7506208419799805, "learning_rate": 3.0820838890183765e-05, "loss": 3.2533, "step": 354550 }, { "epoch": 3.131457637895406, "grad_norm": 4.421753406524658, "learning_rate": 3.0808006421650816e-05, "loss": 3.127, "step": 354600 }, { "epoch": 3.13189918578569, "grad_norm": 4.577394485473633, "learning_rate": 3.079517543553664e-05, "loss": 2.9045, "step": 354650 }, { "epoch": 3.1323407336759743, "grad_norm": 8.759061813354492, "learning_rate": 3.078234593283229e-05, "loss": 2.9444, "step": 354700 }, { "epoch": 3.1327822815662585, "grad_norm": 4.788915157318115, "learning_rate": 3.076951791452876e-05, "loss": 2.6605, "step": 354750 }, { "epoch": 3.1332238294565427, "grad_norm": 0.9270392060279846, "learning_rate": 3.075669138161691e-05, "loss": 2.6962, "step": 354800 }, { "epoch": 3.1336653773468273, "grad_norm": 2.4001381397247314, "learning_rate": 3.0743866335087456e-05, "loss": 3.2599, "step": 354850 }, { "epoch": 3.134106925237111, "grad_norm": 5.556644916534424, "learning_rate": 3.073104277593105e-05, "loss": 3.062, "step": 354900 }, { "epoch": 3.1345484731273956, "grad_norm": 2.956991195678711, "learning_rate": 3.071822070513821e-05, "loss": 2.9419, "step": 354950 }, { "epoch": 3.1349900210176798, "grad_norm": 6.657530307769775, "learning_rate": 3.070540012369929e-05, "loss": 3.196, "step": 355000 }, { "epoch": 3.135431568907964, "grad_norm": 5.2042155265808105, "learning_rate": 3.0692581032604595e-05, "loss": 2.5636, "step": 355050 }, { "epoch": 3.135873116798248, "grad_norm": 6.6366286277771, "learning_rate": 3.067976343284432e-05, "loss": 2.4662, "step": 355100 }, { "epoch": 3.1363146646885323, "grad_norm": 5.310001850128174, "learning_rate": 3.066694732540848e-05, "loss": 2.4014, "step": 355150 }, { "epoch": 3.1367562125788164, "grad_norm": 2.468092679977417, "learning_rate": 3.0654132711286996e-05, "loss": 2.6773, "step": 355200 }, { "epoch": 3.1371977604691006, "grad_norm": 4.711185932159424, "learning_rate": 3.064131959146974e-05, "loss": 2.823, "step": 355250 }, { "epoch": 3.1376393083593848, "grad_norm": 9.00166130065918, "learning_rate": 3.0628507966946384e-05, "loss": 2.5996, "step": 355300 }, { "epoch": 3.138080856249669, "grad_norm": 3.229992628097534, "learning_rate": 3.061569783870651e-05, "loss": 3.0343, "step": 355350 }, { "epoch": 3.138522404139953, "grad_norm": 2.214097738265991, "learning_rate": 3.06028892077396e-05, "loss": 3.1731, "step": 355400 }, { "epoch": 3.1389639520302373, "grad_norm": 4.972867012023926, "learning_rate": 3.059008207503504e-05, "loss": 3.1297, "step": 355450 }, { "epoch": 3.1394054999205214, "grad_norm": 2.9334800243377686, "learning_rate": 3.057727644158204e-05, "loss": 2.909, "step": 355500 }, { "epoch": 3.1398470478108056, "grad_norm": 2.8459246158599854, "learning_rate": 3.0564472308369716e-05, "loss": 2.7541, "step": 355550 }, { "epoch": 3.1402885957010898, "grad_norm": 8.664839744567871, "learning_rate": 3.055166967638712e-05, "loss": 2.7239, "step": 355600 }, { "epoch": 3.140730143591374, "grad_norm": 1.7049593925476074, "learning_rate": 3.05388685466231e-05, "loss": 2.8474, "step": 355650 }, { "epoch": 3.141171691481658, "grad_norm": 2.923825263977051, "learning_rate": 3.0526068920066484e-05, "loss": 2.6631, "step": 355700 }, { "epoch": 3.1416132393719423, "grad_norm": 4.988954544067383, "learning_rate": 3.0513270797705895e-05, "loss": 3.0841, "step": 355750 }, { "epoch": 3.1420547872622264, "grad_norm": 2.1767044067382812, "learning_rate": 3.050047418052991e-05, "loss": 2.748, "step": 355800 }, { "epoch": 3.1424963351525106, "grad_norm": 7.471333026885986, "learning_rate": 3.048767906952693e-05, "loss": 2.398, "step": 355850 }, { "epoch": 3.1429378830427948, "grad_norm": 3.601961612701416, "learning_rate": 3.0474885465685278e-05, "loss": 2.9638, "step": 355900 }, { "epoch": 3.143379430933079, "grad_norm": 2.236664056777954, "learning_rate": 3.046209336999318e-05, "loss": 2.5332, "step": 355950 }, { "epoch": 3.143820978823363, "grad_norm": 5.392511367797852, "learning_rate": 3.044930278343866e-05, "loss": 3.0971, "step": 356000 }, { "epoch": 3.1442625267136473, "grad_norm": 7.5687127113342285, "learning_rate": 3.043651370700974e-05, "loss": 2.213, "step": 356050 }, { "epoch": 3.1447040746039314, "grad_norm": 4.8166961669921875, "learning_rate": 3.0423726141694242e-05, "loss": 3.4677, "step": 356100 }, { "epoch": 3.1451456224942156, "grad_norm": 4.181938171386719, "learning_rate": 3.041094008847988e-05, "loss": 3.1827, "step": 356150 }, { "epoch": 3.1455871703844998, "grad_norm": 5.416933059692383, "learning_rate": 3.0398155548354283e-05, "loss": 3.0694, "step": 356200 }, { "epoch": 3.146028718274784, "grad_norm": 4.654457092285156, "learning_rate": 3.0385372522304976e-05, "loss": 3.7079, "step": 356250 }, { "epoch": 3.146470266165068, "grad_norm": 13.14797306060791, "learning_rate": 3.03725910113193e-05, "loss": 2.5743, "step": 356300 }, { "epoch": 3.1469118140553523, "grad_norm": 7.77878999710083, "learning_rate": 3.035981101638453e-05, "loss": 3.2775, "step": 356350 }, { "epoch": 3.1473533619456364, "grad_norm": 2.257176399230957, "learning_rate": 3.0347032538487807e-05, "loss": 3.3285, "step": 356400 }, { "epoch": 3.1477949098359206, "grad_norm": 3.0555574893951416, "learning_rate": 3.0334255578616194e-05, "loss": 2.6984, "step": 356450 }, { "epoch": 3.148236457726205, "grad_norm": 4.439250469207764, "learning_rate": 3.032173563168106e-05, "loss": 2.8021, "step": 356500 }, { "epoch": 3.1486780056164894, "grad_norm": 3.254890203475952, "learning_rate": 3.0308961680410563e-05, "loss": 2.8713, "step": 356550 }, { "epoch": 3.1491195535067735, "grad_norm": 2.282472848892212, "learning_rate": 3.0296189250105817e-05, "loss": 2.844, "step": 356600 }, { "epoch": 3.1495611013970577, "grad_norm": 3.06990909576416, "learning_rate": 3.0283418341753374e-05, "loss": 3.1974, "step": 356650 }, { "epoch": 3.150002649287342, "grad_norm": 2.491211175918579, "learning_rate": 3.0270648956339655e-05, "loss": 3.0352, "step": 356700 }, { "epoch": 3.150444197177626, "grad_norm": 1.1106147766113281, "learning_rate": 3.025788109485101e-05, "loss": 3.03, "step": 356750 }, { "epoch": 3.15088574506791, "grad_norm": 3.633504867553711, "learning_rate": 3.024511475827363e-05, "loss": 3.0256, "step": 356800 }, { "epoch": 3.1513272929581944, "grad_norm": 3.308501720428467, "learning_rate": 3.023234994759364e-05, "loss": 3.21, "step": 356850 }, { "epoch": 3.1517688408484785, "grad_norm": 1.7054390907287598, "learning_rate": 3.021958666379698e-05, "loss": 3.0731, "step": 356900 }, { "epoch": 3.1522103887387627, "grad_norm": 2.47955322265625, "learning_rate": 3.0206824907869523e-05, "loss": 3.0532, "step": 356950 }, { "epoch": 3.152651936629047, "grad_norm": 4.245659351348877, "learning_rate": 3.0194064680797006e-05, "loss": 2.8397, "step": 357000 }, { "epoch": 3.152651936629047, "eval_asr_loss": 0.954282863205859, "eval_loss": 2.759921073913574, "eval_runtime": 20.4658, "eval_samples_per_second": 37.526, "eval_steps_per_second": 9.381, "eval_tts_loss": 6.166150729112819, "step": 357000 }, { "epoch": 3.153093484519331, "grad_norm": 2.9778575897216797, "learning_rate": 3.018130598356502e-05, "loss": 3.1514, "step": 357050 }, { "epoch": 3.153535032409615, "grad_norm": 2.1935200691223145, "learning_rate": 3.0168548817159104e-05, "loss": 2.8474, "step": 357100 }, { "epoch": 3.1539765802998994, "grad_norm": 6.032583236694336, "learning_rate": 3.0155793182564634e-05, "loss": 2.9361, "step": 357150 }, { "epoch": 3.1544181281901835, "grad_norm": 3.689295530319214, "learning_rate": 3.014303908076686e-05, "loss": 3.2014, "step": 357200 }, { "epoch": 3.1548596760804677, "grad_norm": 1.3066976070404053, "learning_rate": 3.0130541549073815e-05, "loss": 2.4916, "step": 357250 }, { "epoch": 3.155301223970752, "grad_norm": 2.053605556488037, "learning_rate": 3.0117790485119778e-05, "loss": 3.3124, "step": 357300 }, { "epoch": 3.155742771861036, "grad_norm": 4.1359734535217285, "learning_rate": 3.0105040956897816e-05, "loss": 3.0583, "step": 357350 }, { "epoch": 3.15618431975132, "grad_norm": 6.027348041534424, "learning_rate": 3.009229296539276e-05, "loss": 2.9449, "step": 357400 }, { "epoch": 3.1566258676416044, "grad_norm": 1.477520227432251, "learning_rate": 3.007954651158924e-05, "loss": 2.325, "step": 357450 }, { "epoch": 3.1570674155318885, "grad_norm": 5.880183696746826, "learning_rate": 3.0066801596471837e-05, "loss": 2.6441, "step": 357500 }, { "epoch": 3.1575089634221727, "grad_norm": 2.9487268924713135, "learning_rate": 3.0054058221024993e-05, "loss": 2.9287, "step": 357550 }, { "epoch": 3.157950511312457, "grad_norm": 2.143397569656372, "learning_rate": 3.0041316386232997e-05, "loss": 3.0694, "step": 357600 }, { "epoch": 3.158392059202741, "grad_norm": 3.1371824741363525, "learning_rate": 3.0028576093080075e-05, "loss": 3.1627, "step": 357650 }, { "epoch": 3.158833607093025, "grad_norm": 3.561397075653076, "learning_rate": 3.0015837342550308e-05, "loss": 2.9011, "step": 357700 }, { "epoch": 3.1592751549833094, "grad_norm": 6.366957664489746, "learning_rate": 3.0003100135627628e-05, "loss": 3.0629, "step": 357750 }, { "epoch": 3.1597167028735935, "grad_norm": 5.173560619354248, "learning_rate": 2.9990364473295883e-05, "loss": 2.9017, "step": 357800 }, { "epoch": 3.1601582507638777, "grad_norm": 2.425008535385132, "learning_rate": 2.9977630356538822e-05, "loss": 2.5567, "step": 357850 }, { "epoch": 3.160599798654162, "grad_norm": 7.5645527839660645, "learning_rate": 2.9964897786340045e-05, "loss": 2.9049, "step": 357900 }, { "epoch": 3.161041346544446, "grad_norm": 6.728720664978027, "learning_rate": 2.9952166763683e-05, "loss": 2.8664, "step": 357950 }, { "epoch": 3.16148289443473, "grad_norm": 2.7739319801330566, "learning_rate": 2.993943728955107e-05, "loss": 3.1744, "step": 358000 }, { "epoch": 3.161924442325015, "grad_norm": 3.484095573425293, "learning_rate": 2.992670936492752e-05, "loss": 2.7625, "step": 358050 }, { "epoch": 3.1623659902152985, "grad_norm": 2.1070892810821533, "learning_rate": 2.9913982990795448e-05, "loss": 2.8308, "step": 358100 }, { "epoch": 3.162807538105583, "grad_norm": 5.322880268096924, "learning_rate": 2.9901258168137852e-05, "loss": 3.015, "step": 358150 }, { "epoch": 3.1632490859958673, "grad_norm": 1.5672584772109985, "learning_rate": 2.988853489793765e-05, "loss": 3.1558, "step": 358200 }, { "epoch": 3.1636906338861515, "grad_norm": 1.7742763757705688, "learning_rate": 2.987581318117758e-05, "loss": 2.9278, "step": 358250 }, { "epoch": 3.1641321817764356, "grad_norm": 1.8537696599960327, "learning_rate": 2.9863093018840282e-05, "loss": 3.0295, "step": 358300 }, { "epoch": 3.16457372966672, "grad_norm": 4.174468994140625, "learning_rate": 2.985037441190831e-05, "loss": 2.7522, "step": 358350 }, { "epoch": 3.165015277557004, "grad_norm": 5.456238746643066, "learning_rate": 2.9837657361364035e-05, "loss": 2.765, "step": 358400 }, { "epoch": 3.165456825447288, "grad_norm": 4.076654434204102, "learning_rate": 2.9824941868189765e-05, "loss": 3.039, "step": 358450 }, { "epoch": 3.1658983733375723, "grad_norm": 4.969698429107666, "learning_rate": 2.9812227933367647e-05, "loss": 3.3184, "step": 358500 }, { "epoch": 3.1663399212278565, "grad_norm": 5.333426475524902, "learning_rate": 2.9799515557879754e-05, "loss": 2.9415, "step": 358550 }, { "epoch": 3.1667814691181406, "grad_norm": 1.9153121709823608, "learning_rate": 2.978680474270797e-05, "loss": 2.9427, "step": 358600 }, { "epoch": 3.167223017008425, "grad_norm": 4.009693622589111, "learning_rate": 2.9774095488834125e-05, "loss": 2.6564, "step": 358650 }, { "epoch": 3.167664564898709, "grad_norm": 3.356569766998291, "learning_rate": 2.9761387797239902e-05, "loss": 2.9016, "step": 358700 }, { "epoch": 3.168106112788993, "grad_norm": 4.296525001525879, "learning_rate": 2.9748681668906835e-05, "loss": 3.128, "step": 358750 }, { "epoch": 3.1685476606792773, "grad_norm": 2.5694828033447266, "learning_rate": 2.9735977104816392e-05, "loss": 3.1636, "step": 358800 }, { "epoch": 3.1689892085695615, "grad_norm": 3.162705898284912, "learning_rate": 2.9723274105949882e-05, "loss": 2.8127, "step": 358850 }, { "epoch": 3.1694307564598456, "grad_norm": 4.744904041290283, "learning_rate": 2.9710572673288512e-05, "loss": 2.9808, "step": 358900 }, { "epoch": 3.16987230435013, "grad_norm": 5.792562961578369, "learning_rate": 2.969787280781334e-05, "loss": 3.0837, "step": 358950 }, { "epoch": 3.170313852240414, "grad_norm": 2.3473548889160156, "learning_rate": 2.968517451050534e-05, "loss": 2.9812, "step": 359000 }, { "epoch": 3.170755400130698, "grad_norm": 1.9207615852355957, "learning_rate": 2.9672477782345364e-05, "loss": 2.9635, "step": 359050 }, { "epoch": 3.1711969480209823, "grad_norm": 2.7591395378112793, "learning_rate": 2.965978262431408e-05, "loss": 3.2383, "step": 359100 }, { "epoch": 3.1716384959112665, "grad_norm": 4.293609142303467, "learning_rate": 2.9647089037392116e-05, "loss": 2.5104, "step": 359150 }, { "epoch": 3.1720800438015506, "grad_norm": 5.13769006729126, "learning_rate": 2.963439702255995e-05, "loss": 3.062, "step": 359200 }, { "epoch": 3.172521591691835, "grad_norm": 1.3669990301132202, "learning_rate": 2.9621706580797913e-05, "loss": 2.9078, "step": 359250 }, { "epoch": 3.172963139582119, "grad_norm": 3.743950605392456, "learning_rate": 2.960901771308623e-05, "loss": 2.8457, "step": 359300 }, { "epoch": 3.173404687472403, "grad_norm": 2.841085433959961, "learning_rate": 2.9596330420405032e-05, "loss": 3.086, "step": 359350 }, { "epoch": 3.1738462353626873, "grad_norm": 2.8983285427093506, "learning_rate": 2.9583644703734274e-05, "loss": 2.9101, "step": 359400 }, { "epoch": 3.1742877832529715, "grad_norm": 4.0624237060546875, "learning_rate": 2.9570960564053852e-05, "loss": 3.1404, "step": 359450 }, { "epoch": 3.1747293311432556, "grad_norm": 5.2635416984558105, "learning_rate": 2.9558278002343497e-05, "loss": 3.2992, "step": 359500 }, { "epoch": 3.17517087903354, "grad_norm": 3.1870532035827637, "learning_rate": 2.9545597019582804e-05, "loss": 2.9068, "step": 359550 }, { "epoch": 3.175612426923824, "grad_norm": 3.767455816268921, "learning_rate": 2.953291761675131e-05, "loss": 2.7437, "step": 359600 }, { "epoch": 3.176053974814108, "grad_norm": 4.007964134216309, "learning_rate": 2.9520239794828358e-05, "loss": 2.8561, "step": 359650 }, { "epoch": 3.1764955227043927, "grad_norm": 1.6411175727844238, "learning_rate": 2.950756355479324e-05, "loss": 2.9968, "step": 359700 }, { "epoch": 3.176937070594677, "grad_norm": 4.157561779022217, "learning_rate": 2.949488889762505e-05, "loss": 2.4313, "step": 359750 }, { "epoch": 3.177378618484961, "grad_norm": 4.296000957489014, "learning_rate": 2.9482215824302823e-05, "loss": 2.9174, "step": 359800 }, { "epoch": 3.1778201663752452, "grad_norm": 4.604443550109863, "learning_rate": 2.9469544335805443e-05, "loss": 3.1864, "step": 359850 }, { "epoch": 3.1782617142655294, "grad_norm": 3.0718367099761963, "learning_rate": 2.9456874433111657e-05, "loss": 2.8379, "step": 359900 }, { "epoch": 3.1787032621558136, "grad_norm": 2.6071958541870117, "learning_rate": 2.944420611720012e-05, "loss": 2.7839, "step": 359950 }, { "epoch": 3.1791448100460977, "grad_norm": 3.7291224002838135, "learning_rate": 2.9431539389049366e-05, "loss": 2.6284, "step": 360000 }, { "epoch": 3.1791448100460977, "eval_asr_loss": 0.9508853787920921, "eval_loss": 2.7509162425994873, "eval_runtime": 20.5814, "eval_samples_per_second": 37.315, "eval_steps_per_second": 9.329, "eval_tts_loss": 6.127400047174752, "step": 360000 }, { "epoch": 3.179586357936382, "grad_norm": 2.613326072692871, "learning_rate": 2.9418874249637773e-05, "loss": 3.1287, "step": 360050 }, { "epoch": 3.180027905826666, "grad_norm": 2.402574062347412, "learning_rate": 2.9406210699943616e-05, "loss": 3.0204, "step": 360100 }, { "epoch": 3.1804694537169502, "grad_norm": 2.720712661743164, "learning_rate": 2.9393548740945053e-05, "loss": 3.0637, "step": 360150 }, { "epoch": 3.1809110016072344, "grad_norm": 6.183115482330322, "learning_rate": 2.938114156536188e-05, "loss": 2.6594, "step": 360200 }, { "epoch": 3.1813525494975186, "grad_norm": 3.9944679737091064, "learning_rate": 2.9368482758825873e-05, "loss": 2.6379, "step": 360250 }, { "epoch": 3.1817940973878027, "grad_norm": 3.31664776802063, "learning_rate": 2.9355825545899595e-05, "loss": 2.9555, "step": 360300 }, { "epoch": 3.182235645278087, "grad_norm": 5.988372802734375, "learning_rate": 2.9343169927560748e-05, "loss": 3.0805, "step": 360350 }, { "epoch": 3.182677193168371, "grad_norm": 1.7002320289611816, "learning_rate": 2.933051590478687e-05, "loss": 3.0795, "step": 360400 }, { "epoch": 3.183118741058655, "grad_norm": 2.9809789657592773, "learning_rate": 2.931786347855535e-05, "loss": 2.6834, "step": 360450 }, { "epoch": 3.1835602889489394, "grad_norm": 1.9576280117034912, "learning_rate": 2.9305212649843506e-05, "loss": 3.0148, "step": 360500 }, { "epoch": 3.1840018368392236, "grad_norm": 4.269618511199951, "learning_rate": 2.9292563419628504e-05, "loss": 2.9657, "step": 360550 }, { "epoch": 3.1844433847295077, "grad_norm": 3.3936052322387695, "learning_rate": 2.9279915788887363e-05, "loss": 2.7966, "step": 360600 }, { "epoch": 3.184884932619792, "grad_norm": 6.597276210784912, "learning_rate": 2.9267269758597028e-05, "loss": 2.9482, "step": 360650 }, { "epoch": 3.185326480510076, "grad_norm": 5.644941329956055, "learning_rate": 2.9254625329734308e-05, "loss": 3.0388, "step": 360700 }, { "epoch": 3.18576802840036, "grad_norm": 5.906317234039307, "learning_rate": 2.924198250327587e-05, "loss": 3.3613, "step": 360750 }, { "epoch": 3.1862095762906444, "grad_norm": 5.947046279907227, "learning_rate": 2.922934128019825e-05, "loss": 3.0067, "step": 360800 }, { "epoch": 3.1866511241809286, "grad_norm": 4.553384304046631, "learning_rate": 2.9216701661477875e-05, "loss": 2.6935, "step": 360850 }, { "epoch": 3.1870926720712127, "grad_norm": 4.484139919281006, "learning_rate": 2.9204063648091084e-05, "loss": 2.5933, "step": 360900 }, { "epoch": 3.187534219961497, "grad_norm": 4.699550151824951, "learning_rate": 2.9191427241014023e-05, "loss": 2.8232, "step": 360950 }, { "epoch": 3.187975767851781, "grad_norm": 1.8874015808105469, "learning_rate": 2.9178792441222753e-05, "loss": 2.7993, "step": 361000 }, { "epoch": 3.188417315742065, "grad_norm": 7.519862174987793, "learning_rate": 2.916615924969322e-05, "loss": 2.61, "step": 361050 }, { "epoch": 3.1888588636323494, "grad_norm": 3.6855697631835938, "learning_rate": 2.9153527667401205e-05, "loss": 2.444, "step": 361100 }, { "epoch": 3.1893004115226335, "grad_norm": 3.9974000453948975, "learning_rate": 2.9140897695322393e-05, "loss": 2.7655, "step": 361150 }, { "epoch": 3.1897419594129177, "grad_norm": 3.2694218158721924, "learning_rate": 2.9128269334432384e-05, "loss": 3.2426, "step": 361200 }, { "epoch": 3.1901835073032023, "grad_norm": 5.486494064331055, "learning_rate": 2.911564258570658e-05, "loss": 3.4178, "step": 361250 }, { "epoch": 3.1906250551934865, "grad_norm": 3.221587657928467, "learning_rate": 2.9103017450120273e-05, "loss": 2.8741, "step": 361300 }, { "epoch": 3.1910666030837707, "grad_norm": 2.560730457305908, "learning_rate": 2.9090393928648664e-05, "loss": 2.8115, "step": 361350 }, { "epoch": 3.191508150974055, "grad_norm": 3.9733705520629883, "learning_rate": 2.9077772022266837e-05, "loss": 2.7787, "step": 361400 }, { "epoch": 3.191949698864339, "grad_norm": 4.513033390045166, "learning_rate": 2.906515173194968e-05, "loss": 2.5771, "step": 361450 }, { "epoch": 3.192391246754623, "grad_norm": 2.34289813041687, "learning_rate": 2.9052533058672038e-05, "loss": 2.9362, "step": 361500 }, { "epoch": 3.1928327946449073, "grad_norm": 3.3383851051330566, "learning_rate": 2.9039916003408606e-05, "loss": 2.9507, "step": 361550 }, { "epoch": 3.1932743425351915, "grad_norm": 3.364741325378418, "learning_rate": 2.9027300567133924e-05, "loss": 2.9384, "step": 361600 }, { "epoch": 3.1937158904254757, "grad_norm": 3.8707759380340576, "learning_rate": 2.9014686750822412e-05, "loss": 3.384, "step": 361650 }, { "epoch": 3.19415743831576, "grad_norm": 4.745691299438477, "learning_rate": 2.9002074555448415e-05, "loss": 3.2801, "step": 361700 }, { "epoch": 3.194598986206044, "grad_norm": 2.8048341274261475, "learning_rate": 2.898946398198609e-05, "loss": 2.7475, "step": 361750 }, { "epoch": 3.195040534096328, "grad_norm": 5.285944938659668, "learning_rate": 2.89768550314095e-05, "loss": 2.8325, "step": 361800 }, { "epoch": 3.1954820819866123, "grad_norm": 1.711706280708313, "learning_rate": 2.896424770469259e-05, "loss": 2.5752, "step": 361850 }, { "epoch": 3.1959236298768965, "grad_norm": 2.2461252212524414, "learning_rate": 2.895164200280919e-05, "loss": 2.7603, "step": 361900 }, { "epoch": 3.1963651777671807, "grad_norm": 4.525991439819336, "learning_rate": 2.893903792673293e-05, "loss": 2.9798, "step": 361950 }, { "epoch": 3.196806725657465, "grad_norm": 4.786330699920654, "learning_rate": 2.892643547743743e-05, "loss": 3.2713, "step": 362000 }, { "epoch": 3.197248273547749, "grad_norm": 6.595480918884277, "learning_rate": 2.8913834655896084e-05, "loss": 3.2181, "step": 362050 }, { "epoch": 3.197689821438033, "grad_norm": 4.742706298828125, "learning_rate": 2.890123546308219e-05, "loss": 2.9448, "step": 362100 }, { "epoch": 3.1981313693283173, "grad_norm": 5.465112209320068, "learning_rate": 2.888863789996894e-05, "loss": 2.8574, "step": 362150 }, { "epoch": 3.1985729172186015, "grad_norm": 0.8134891390800476, "learning_rate": 2.887604196752941e-05, "loss": 2.7956, "step": 362200 }, { "epoch": 3.1990144651088857, "grad_norm": 3.978318929672241, "learning_rate": 2.88634476667365e-05, "loss": 2.588, "step": 362250 }, { "epoch": 3.19945601299917, "grad_norm": 5.220042705535889, "learning_rate": 2.8850854998563028e-05, "loss": 3.2042, "step": 362300 }, { "epoch": 3.199897560889454, "grad_norm": 3.4094042778015137, "learning_rate": 2.8838263963981675e-05, "loss": 2.8374, "step": 362350 }, { "epoch": 3.200339108779738, "grad_norm": 3.699916362762451, "learning_rate": 2.882592633594029e-05, "loss": 2.8783, "step": 362400 }, { "epoch": 3.2007806566700223, "grad_norm": 3.1803691387176514, "learning_rate": 2.881333853874042e-05, "loss": 3.1672, "step": 362450 }, { "epoch": 3.2012222045603065, "grad_norm": 2.6369807720184326, "learning_rate": 2.880075237803047e-05, "loss": 3.124, "step": 362500 }, { "epoch": 3.2016637524505906, "grad_norm": 7.230294704437256, "learning_rate": 2.8788167854782644e-05, "loss": 2.3994, "step": 362550 }, { "epoch": 3.202105300340875, "grad_norm": 3.544135093688965, "learning_rate": 2.8775584969968994e-05, "loss": 3.1085, "step": 362600 }, { "epoch": 3.202546848231159, "grad_norm": 3.199202299118042, "learning_rate": 2.87630037245614e-05, "loss": 3.2476, "step": 362650 }, { "epoch": 3.202988396121443, "grad_norm": 3.530019760131836, "learning_rate": 2.8750424119531715e-05, "loss": 2.7973, "step": 362700 }, { "epoch": 3.2034299440117273, "grad_norm": 4.290691375732422, "learning_rate": 2.873784615585155e-05, "loss": 2.9508, "step": 362750 }, { "epoch": 3.203871491902012, "grad_norm": 3.866528272628784, "learning_rate": 2.8725269834492495e-05, "loss": 3.076, "step": 362800 }, { "epoch": 3.2043130397922956, "grad_norm": 4.218257904052734, "learning_rate": 2.8712695156425918e-05, "loss": 2.707, "step": 362850 }, { "epoch": 3.2047545876825803, "grad_norm": 6.158079624176025, "learning_rate": 2.870012212262313e-05, "loss": 2.8174, "step": 362900 }, { "epoch": 3.2051961355728644, "grad_norm": 3.02413272857666, "learning_rate": 2.8687550734055306e-05, "loss": 2.9265, "step": 362950 }, { "epoch": 3.2056376834631486, "grad_norm": 2.938126802444458, "learning_rate": 2.8674980991693446e-05, "loss": 3.2987, "step": 363000 }, { "epoch": 3.2056376834631486, "eval_asr_loss": 0.9585031008141492, "eval_loss": 2.7487404346466064, "eval_runtime": 20.6728, "eval_samples_per_second": 37.15, "eval_steps_per_second": 9.288, "eval_tts_loss": 6.129536288779272, "step": 363000 }, { "epoch": 3.2060792313534328, "grad_norm": 2.7070798873901367, "learning_rate": 2.8662412896508495e-05, "loss": 2.829, "step": 363050 }, { "epoch": 3.206520779243717, "grad_norm": 3.6368250846862793, "learning_rate": 2.8649846449471217e-05, "loss": 3.075, "step": 363100 }, { "epoch": 3.206962327134001, "grad_norm": 6.804358005523682, "learning_rate": 2.8637281651552226e-05, "loss": 2.513, "step": 363150 }, { "epoch": 3.2074038750242853, "grad_norm": 2.6479644775390625, "learning_rate": 2.862471850372208e-05, "loss": 3.2678, "step": 363200 }, { "epoch": 3.2078454229145694, "grad_norm": 4.884181976318359, "learning_rate": 2.8612157006951195e-05, "loss": 2.8535, "step": 363250 }, { "epoch": 3.2082869708048536, "grad_norm": 5.885494709014893, "learning_rate": 2.8599597162209796e-05, "loss": 3.1912, "step": 363300 }, { "epoch": 3.2087285186951378, "grad_norm": 5.305294036865234, "learning_rate": 2.858703897046805e-05, "loss": 2.4645, "step": 363350 }, { "epoch": 3.209170066585422, "grad_norm": 3.2507739067077637, "learning_rate": 2.857448243269596e-05, "loss": 2.695, "step": 363400 }, { "epoch": 3.209611614475706, "grad_norm": 3.738827705383301, "learning_rate": 2.8561927549863444e-05, "loss": 3.3404, "step": 363450 }, { "epoch": 3.2100531623659903, "grad_norm": 3.1707406044006348, "learning_rate": 2.8549374322940232e-05, "loss": 2.8879, "step": 363500 }, { "epoch": 3.2104947102562744, "grad_norm": 3.7499825954437256, "learning_rate": 2.8536822752895943e-05, "loss": 3.3623, "step": 363550 }, { "epoch": 3.2109362581465586, "grad_norm": 1.3391181230545044, "learning_rate": 2.8524272840700106e-05, "loss": 2.7044, "step": 363600 }, { "epoch": 3.2113778060368428, "grad_norm": 5.911325454711914, "learning_rate": 2.8511724587322064e-05, "loss": 3.104, "step": 363650 }, { "epoch": 3.211819353927127, "grad_norm": 3.104883909225464, "learning_rate": 2.8499177993731085e-05, "loss": 3.1506, "step": 363700 }, { "epoch": 3.212260901817411, "grad_norm": 4.059735298156738, "learning_rate": 2.8486633060896305e-05, "loss": 3.1348, "step": 363750 }, { "epoch": 3.2127024497076953, "grad_norm": 1.6747605800628662, "learning_rate": 2.847408978978667e-05, "loss": 2.8319, "step": 363800 }, { "epoch": 3.2131439975979794, "grad_norm": 0.9222479462623596, "learning_rate": 2.846154818137109e-05, "loss": 2.3868, "step": 363850 }, { "epoch": 3.2135855454882636, "grad_norm": 5.176297664642334, "learning_rate": 2.8449008236618268e-05, "loss": 2.5614, "step": 363900 }, { "epoch": 3.2140270933785478, "grad_norm": 3.8599517345428467, "learning_rate": 2.8436469956496797e-05, "loss": 3.0401, "step": 363950 }, { "epoch": 3.214468641268832, "grad_norm": 3.8960721492767334, "learning_rate": 2.8423933341975167e-05, "loss": 2.7339, "step": 364000 }, { "epoch": 3.214910189159116, "grad_norm": 9.651103019714355, "learning_rate": 2.8411398394021728e-05, "loss": 3.4055, "step": 364050 }, { "epoch": 3.2153517370494002, "grad_norm": 2.979292869567871, "learning_rate": 2.839886511360471e-05, "loss": 2.735, "step": 364100 }, { "epoch": 3.2157932849396844, "grad_norm": 3.035641670227051, "learning_rate": 2.8386333501692177e-05, "loss": 3.0156, "step": 364150 }, { "epoch": 3.2162348328299686, "grad_norm": 3.4312310218811035, "learning_rate": 2.8373803559252122e-05, "loss": 2.9819, "step": 364200 }, { "epoch": 3.2166763807202527, "grad_norm": 1.1246628761291504, "learning_rate": 2.8361275287252332e-05, "loss": 2.8461, "step": 364250 }, { "epoch": 3.217117928610537, "grad_norm": 3.506833553314209, "learning_rate": 2.8348748686660554e-05, "loss": 3.1263, "step": 364300 }, { "epoch": 3.217559476500821, "grad_norm": 4.562410354614258, "learning_rate": 2.8336223758444324e-05, "loss": 2.7937, "step": 364350 }, { "epoch": 3.2180010243911052, "grad_norm": 5.248993396759033, "learning_rate": 2.832370050357113e-05, "loss": 2.8764, "step": 364400 }, { "epoch": 3.21844257228139, "grad_norm": 1.6171966791152954, "learning_rate": 2.8311178923008242e-05, "loss": 2.5885, "step": 364450 }, { "epoch": 3.218884120171674, "grad_norm": 1.5783275365829468, "learning_rate": 2.8298659017722862e-05, "loss": 3.0916, "step": 364500 }, { "epoch": 3.219325668061958, "grad_norm": 4.894155025482178, "learning_rate": 2.828614078868206e-05, "loss": 2.9735, "step": 364550 }, { "epoch": 3.2197672159522424, "grad_norm": 2.290025472640991, "learning_rate": 2.8273624236852762e-05, "loss": 2.7427, "step": 364600 }, { "epoch": 3.2202087638425265, "grad_norm": 2.7645976543426514, "learning_rate": 2.8261109363201765e-05, "loss": 3.2864, "step": 364650 }, { "epoch": 3.2206503117328107, "grad_norm": 4.44540548324585, "learning_rate": 2.8248596168695707e-05, "loss": 3.0206, "step": 364700 }, { "epoch": 3.221091859623095, "grad_norm": 2.9939565658569336, "learning_rate": 2.8236084654301166e-05, "loss": 3.2775, "step": 364750 }, { "epoch": 3.221533407513379, "grad_norm": 3.6256279945373535, "learning_rate": 2.8223574820984517e-05, "loss": 2.9971, "step": 364800 }, { "epoch": 3.221974955403663, "grad_norm": 5.711068630218506, "learning_rate": 2.821106666971205e-05, "loss": 3.0, "step": 364850 }, { "epoch": 3.2224165032939474, "grad_norm": 3.29297137260437, "learning_rate": 2.8198560201449943e-05, "loss": 3.0149, "step": 364900 }, { "epoch": 3.2228580511842315, "grad_norm": 2.7711613178253174, "learning_rate": 2.8186055417164168e-05, "loss": 2.8964, "step": 364950 }, { "epoch": 3.2232995990745157, "grad_norm": 4.748452663421631, "learning_rate": 2.817355231782065e-05, "loss": 3.1294, "step": 365000 }, { "epoch": 3.2237411469648, "grad_norm": 2.5255863666534424, "learning_rate": 2.8161050904385145e-05, "loss": 2.8257, "step": 365050 }, { "epoch": 3.224182694855084, "grad_norm": 7.028888702392578, "learning_rate": 2.8148551177823233e-05, "loss": 2.7716, "step": 365100 }, { "epoch": 3.224624242745368, "grad_norm": 3.558807373046875, "learning_rate": 2.813605313910046e-05, "loss": 3.0049, "step": 365150 }, { "epoch": 3.2250657906356524, "grad_norm": 2.250173330307007, "learning_rate": 2.8123556789182175e-05, "loss": 3.0842, "step": 365200 }, { "epoch": 3.2255073385259365, "grad_norm": 11.092060089111328, "learning_rate": 2.8111062129033644e-05, "loss": 2.6154, "step": 365250 }, { "epoch": 3.2259488864162207, "grad_norm": 5.414551258087158, "learning_rate": 2.8098569159619924e-05, "loss": 3.2372, "step": 365300 }, { "epoch": 3.226390434306505, "grad_norm": 3.8668439388275146, "learning_rate": 2.8086077881906025e-05, "loss": 2.892, "step": 365350 }, { "epoch": 3.226831982196789, "grad_norm": 3.259580135345459, "learning_rate": 2.8073588296856807e-05, "loss": 2.954, "step": 365400 }, { "epoch": 3.227273530087073, "grad_norm": 3.0489330291748047, "learning_rate": 2.8061100405436957e-05, "loss": 3.2316, "step": 365450 }, { "epoch": 3.2277150779773573, "grad_norm": 3.016685724258423, "learning_rate": 2.8048614208611045e-05, "loss": 2.7708, "step": 365500 }, { "epoch": 3.2281566258676415, "grad_norm": 4.972036361694336, "learning_rate": 2.8036129707343567e-05, "loss": 3.1287, "step": 365550 }, { "epoch": 3.2285981737579257, "grad_norm": 7.575818061828613, "learning_rate": 2.80236469025988e-05, "loss": 3.0032, "step": 365600 }, { "epoch": 3.22903972164821, "grad_norm": 2.3671131134033203, "learning_rate": 2.8011165795340954e-05, "loss": 2.7093, "step": 365650 }, { "epoch": 3.229481269538494, "grad_norm": 3.9584193229675293, "learning_rate": 2.799868638653409e-05, "loss": 2.7426, "step": 365700 }, { "epoch": 3.229922817428778, "grad_norm": 1.927212119102478, "learning_rate": 2.7986208677142155e-05, "loss": 3.092, "step": 365750 }, { "epoch": 3.2303643653190623, "grad_norm": 4.744783401489258, "learning_rate": 2.7973732668128927e-05, "loss": 2.977, "step": 365800 }, { "epoch": 3.2308059132093465, "grad_norm": 4.433849811553955, "learning_rate": 2.796125836045806e-05, "loss": 2.8058, "step": 365850 }, { "epoch": 3.2312474610996307, "grad_norm": 7.234720230102539, "learning_rate": 2.794878575509312e-05, "loss": 2.7911, "step": 365900 }, { "epoch": 3.231689008989915, "grad_norm": 1.0624055862426758, "learning_rate": 2.793631485299747e-05, "loss": 2.7409, "step": 365950 }, { "epoch": 3.2321305568801995, "grad_norm": 5.817831039428711, "learning_rate": 2.7923845655134402e-05, "loss": 2.9366, "step": 366000 }, { "epoch": 3.2321305568801995, "eval_asr_loss": 0.9542335591040626, "eval_loss": 2.756089925765991, "eval_runtime": 20.5081, "eval_samples_per_second": 37.449, "eval_steps_per_second": 9.362, "eval_tts_loss": 6.159807048745202, "step": 366000 }, { "epoch": 3.232572104770483, "grad_norm": 6.161731243133545, "learning_rate": 2.7911378162467074e-05, "loss": 3.1158, "step": 366050 }, { "epoch": 3.233013652660768, "grad_norm": 2.728720188140869, "learning_rate": 2.7898912375958465e-05, "loss": 2.8398, "step": 366100 }, { "epoch": 3.233455200551052, "grad_norm": 1.4105002880096436, "learning_rate": 2.7886448296571476e-05, "loss": 3.0303, "step": 366150 }, { "epoch": 3.233896748441336, "grad_norm": 5.29327392578125, "learning_rate": 2.787398592526882e-05, "loss": 2.8215, "step": 366200 }, { "epoch": 3.2343382963316203, "grad_norm": 5.850403785705566, "learning_rate": 2.786152526301315e-05, "loss": 3.3803, "step": 366250 }, { "epoch": 3.2347798442219045, "grad_norm": 3.9078965187072754, "learning_rate": 2.7849066310766903e-05, "loss": 3.2101, "step": 366300 }, { "epoch": 3.2352213921121886, "grad_norm": 4.3974995613098145, "learning_rate": 2.7836609069492448e-05, "loss": 2.8038, "step": 366350 }, { "epoch": 3.235662940002473, "grad_norm": 2.7810566425323486, "learning_rate": 2.782440263395565e-05, "loss": 2.6678, "step": 366400 }, { "epoch": 3.236104487892757, "grad_norm": 5.08760929107666, "learning_rate": 2.781194878324397e-05, "loss": 2.8046, "step": 366450 }, { "epoch": 3.236546035783041, "grad_norm": 7.316036701202393, "learning_rate": 2.7799496646371092e-05, "loss": 2.9684, "step": 366500 }, { "epoch": 3.2369875836733253, "grad_norm": 7.416934967041016, "learning_rate": 2.778729521592901e-05, "loss": 2.3399, "step": 366550 }, { "epoch": 3.2374291315636095, "grad_norm": 5.640448570251465, "learning_rate": 2.7774846475294398e-05, "loss": 2.6341, "step": 366600 }, { "epoch": 3.2378706794538936, "grad_norm": 2.770671844482422, "learning_rate": 2.7762399451364428e-05, "loss": 2.7727, "step": 366650 }, { "epoch": 3.238312227344178, "grad_norm": 5.004711627960205, "learning_rate": 2.7749954145100553e-05, "loss": 2.8005, "step": 366700 }, { "epoch": 3.238753775234462, "grad_norm": 6.234478950500488, "learning_rate": 2.7737510557464032e-05, "loss": 2.9744, "step": 366750 }, { "epoch": 3.239195323124746, "grad_norm": 3.1577422618865967, "learning_rate": 2.7725068689416046e-05, "loss": 2.8164, "step": 366800 }, { "epoch": 3.2396368710150303, "grad_norm": 5.218726634979248, "learning_rate": 2.771262854191763e-05, "loss": 2.6886, "step": 366850 }, { "epoch": 3.2400784189053145, "grad_norm": 2.7806460857391357, "learning_rate": 2.7700190115929637e-05, "loss": 2.6514, "step": 366900 }, { "epoch": 3.2405199667955986, "grad_norm": 3.477102518081665, "learning_rate": 2.768775341241287e-05, "loss": 3.1344, "step": 366950 }, { "epoch": 3.240961514685883, "grad_norm": 4.278136253356934, "learning_rate": 2.767531843232797e-05, "loss": 2.7797, "step": 367000 }, { "epoch": 3.241403062576167, "grad_norm": 3.4427273273468018, "learning_rate": 2.7662885176635383e-05, "loss": 3.132, "step": 367050 }, { "epoch": 3.241844610466451, "grad_norm": 3.2588388919830322, "learning_rate": 2.765045364629551e-05, "loss": 2.9763, "step": 367100 }, { "epoch": 3.2422861583567353, "grad_norm": 2.82857084274292, "learning_rate": 2.7638023842268595e-05, "loss": 3.2147, "step": 367150 }, { "epoch": 3.2427277062470194, "grad_norm": 7.353505611419678, "learning_rate": 2.7625595765514688e-05, "loss": 2.8756, "step": 367200 }, { "epoch": 3.2431692541373036, "grad_norm": 5.4813127517700195, "learning_rate": 2.7613169416993805e-05, "loss": 3.063, "step": 367250 }, { "epoch": 3.243610802027588, "grad_norm": 3.0783350467681885, "learning_rate": 2.7600744797665734e-05, "loss": 2.9619, "step": 367300 }, { "epoch": 3.244052349917872, "grad_norm": 3.1509792804718018, "learning_rate": 2.75883219084902e-05, "loss": 2.556, "step": 367350 }, { "epoch": 3.244493897808156, "grad_norm": 3.6255862712860107, "learning_rate": 2.757590075042673e-05, "loss": 2.7689, "step": 367400 }, { "epoch": 3.2449354456984403, "grad_norm": 2.892460823059082, "learning_rate": 2.756348132443478e-05, "loss": 3.1217, "step": 367450 }, { "epoch": 3.2453769935887244, "grad_norm": 6.581718921661377, "learning_rate": 2.755106363147366e-05, "loss": 2.8101, "step": 367500 }, { "epoch": 3.2458185414790086, "grad_norm": 4.133147239685059, "learning_rate": 2.7538647672502506e-05, "loss": 2.9601, "step": 367550 }, { "epoch": 3.246260089369293, "grad_norm": 4.857241630554199, "learning_rate": 2.7526233448480355e-05, "loss": 3.2387, "step": 367600 }, { "epoch": 3.2467016372595774, "grad_norm": 5.211623191833496, "learning_rate": 2.751382096036611e-05, "loss": 2.862, "step": 367650 }, { "epoch": 3.2471431851498616, "grad_norm": 3.4475038051605225, "learning_rate": 2.750141020911849e-05, "loss": 3.1925, "step": 367700 }, { "epoch": 3.2475847330401457, "grad_norm": 1.1415144205093384, "learning_rate": 2.748900119569615e-05, "loss": 3.2976, "step": 367750 }, { "epoch": 3.24802628093043, "grad_norm": 4.77948522567749, "learning_rate": 2.7476593921057587e-05, "loss": 3.0791, "step": 367800 }, { "epoch": 3.248467828820714, "grad_norm": 4.05330753326416, "learning_rate": 2.746418838616116e-05, "loss": 2.987, "step": 367850 }, { "epoch": 3.248909376710998, "grad_norm": 2.941519260406494, "learning_rate": 2.745178459196507e-05, "loss": 3.3228, "step": 367900 }, { "epoch": 3.2493509246012824, "grad_norm": 3.4454493522644043, "learning_rate": 2.7439382539427438e-05, "loss": 2.9842, "step": 367950 }, { "epoch": 3.2497924724915666, "grad_norm": 4.366009712219238, "learning_rate": 2.7426982229506194e-05, "loss": 2.8475, "step": 368000 }, { "epoch": 3.2502340203818507, "grad_norm": 3.668621063232422, "learning_rate": 2.7414583663159137e-05, "loss": 2.4896, "step": 368050 }, { "epoch": 3.250675568272135, "grad_norm": 3.500096321105957, "learning_rate": 2.740218684134398e-05, "loss": 2.9655, "step": 368100 }, { "epoch": 3.251117116162419, "grad_norm": 3.369339942932129, "learning_rate": 2.738979176501828e-05, "loss": 2.9942, "step": 368150 }, { "epoch": 3.251558664052703, "grad_norm": 2.5865495204925537, "learning_rate": 2.7377398435139413e-05, "loss": 2.7785, "step": 368200 }, { "epoch": 3.2520002119429874, "grad_norm": 2.7908427715301514, "learning_rate": 2.7365006852664686e-05, "loss": 3.022, "step": 368250 }, { "epoch": 3.2524417598332716, "grad_norm": 5.630073070526123, "learning_rate": 2.7352617018551264e-05, "loss": 3.0074, "step": 368300 }, { "epoch": 3.2528833077235557, "grad_norm": 5.39545202255249, "learning_rate": 2.7340228933756114e-05, "loss": 2.9106, "step": 368350 }, { "epoch": 3.25332485561384, "grad_norm": 4.454920768737793, "learning_rate": 2.7327842599236143e-05, "loss": 2.9746, "step": 368400 }, { "epoch": 3.253766403504124, "grad_norm": 3.094637632369995, "learning_rate": 2.7315458015948063e-05, "loss": 3.0761, "step": 368450 }, { "epoch": 3.254207951394408, "grad_norm": 4.314550399780273, "learning_rate": 2.730307518484851e-05, "loss": 2.8363, "step": 368500 }, { "epoch": 3.2546494992846924, "grad_norm": 2.7481887340545654, "learning_rate": 2.729069410689391e-05, "loss": 3.0057, "step": 368550 }, { "epoch": 3.2550910471749765, "grad_norm": 2.171276092529297, "learning_rate": 2.7278314783040626e-05, "loss": 2.8315, "step": 368600 }, { "epoch": 3.2555325950652607, "grad_norm": 1.2430096864700317, "learning_rate": 2.7265937214244875e-05, "loss": 3.0023, "step": 368650 }, { "epoch": 3.255974142955545, "grad_norm": 5.757850170135498, "learning_rate": 2.725356140146267e-05, "loss": 2.7788, "step": 368700 }, { "epoch": 3.256415690845829, "grad_norm": 5.629611492156982, "learning_rate": 2.7241187345649982e-05, "loss": 2.773, "step": 368750 }, { "epoch": 3.256857238736113, "grad_norm": 4.567326068878174, "learning_rate": 2.722881504776258e-05, "loss": 3.0883, "step": 368800 }, { "epoch": 3.2572987866263974, "grad_norm": 2.71937894821167, "learning_rate": 2.7216444508756113e-05, "loss": 3.3898, "step": 368850 }, { "epoch": 3.2577403345166815, "grad_norm": 3.707050323486328, "learning_rate": 2.7204075729586097e-05, "loss": 3.0743, "step": 368900 }, { "epoch": 3.2581818824069657, "grad_norm": 6.336466312408447, "learning_rate": 2.7191708711207932e-05, "loss": 3.1379, "step": 368950 }, { "epoch": 3.25862343029725, "grad_norm": 4.750102519989014, "learning_rate": 2.7179343454576877e-05, "loss": 3.0136, "step": 369000 }, { "epoch": 3.25862343029725, "eval_asr_loss": 0.9475825568991976, "eval_loss": 2.743767023086548, "eval_runtime": 20.3023, "eval_samples_per_second": 37.828, "eval_steps_per_second": 9.457, "eval_tts_loss": 6.149724143381788, "step": 369000 }, { "epoch": 3.259064978187534, "grad_norm": 4.086844444274902, "learning_rate": 2.716697996064801e-05, "loss": 2.9547, "step": 369050 }, { "epoch": 3.259506526077818, "grad_norm": 1.2858823537826538, "learning_rate": 2.7154618230376318e-05, "loss": 2.7643, "step": 369100 }, { "epoch": 3.2599480739681024, "grad_norm": 2.2076737880706787, "learning_rate": 2.714225826471667e-05, "loss": 3.1101, "step": 369150 }, { "epoch": 3.260389621858387, "grad_norm": 6.295970916748047, "learning_rate": 2.7129900064623738e-05, "loss": 2.8308, "step": 369200 }, { "epoch": 3.2608311697486707, "grad_norm": 5.433557987213135, "learning_rate": 2.711754363105208e-05, "loss": 3.2432, "step": 369250 }, { "epoch": 3.2612727176389553, "grad_norm": 3.36173152923584, "learning_rate": 2.7105188964956152e-05, "loss": 3.2235, "step": 369300 }, { "epoch": 3.2617142655292395, "grad_norm": 4.344919681549072, "learning_rate": 2.7092836067290216e-05, "loss": 2.9716, "step": 369350 }, { "epoch": 3.2621558134195237, "grad_norm": 3.4870190620422363, "learning_rate": 2.7080484939008456e-05, "loss": 2.691, "step": 369400 }, { "epoch": 3.262597361309808, "grad_norm": 3.2695343494415283, "learning_rate": 2.706813558106489e-05, "loss": 2.965, "step": 369450 }, { "epoch": 3.263038909200092, "grad_norm": 2.460634708404541, "learning_rate": 2.7055787994413377e-05, "loss": 2.6457, "step": 369500 }, { "epoch": 3.263480457090376, "grad_norm": 4.8252105712890625, "learning_rate": 2.7043442180007706e-05, "loss": 2.7797, "step": 369550 }, { "epoch": 3.2639220049806603, "grad_norm": 4.886524200439453, "learning_rate": 2.7031098138801436e-05, "loss": 2.8483, "step": 369600 }, { "epoch": 3.2643635528709445, "grad_norm": 6.848212242126465, "learning_rate": 2.701875587174808e-05, "loss": 2.8339, "step": 369650 }, { "epoch": 3.2648051007612287, "grad_norm": 3.2493505477905273, "learning_rate": 2.700641537980094e-05, "loss": 2.8743, "step": 369700 }, { "epoch": 3.265246648651513, "grad_norm": 2.2643160820007324, "learning_rate": 2.6994076663913237e-05, "loss": 2.7744, "step": 369750 }, { "epoch": 3.265688196541797, "grad_norm": 6.144895553588867, "learning_rate": 2.6981739725038047e-05, "loss": 3.174, "step": 369800 }, { "epoch": 3.266129744432081, "grad_norm": 5.1884307861328125, "learning_rate": 2.6969404564128253e-05, "loss": 3.0128, "step": 369850 }, { "epoch": 3.2665712923223653, "grad_norm": 3.4894297122955322, "learning_rate": 2.695707118213668e-05, "loss": 2.9758, "step": 369900 }, { "epoch": 3.2670128402126495, "grad_norm": 2.8172497749328613, "learning_rate": 2.694498619460946e-05, "loss": 2.9454, "step": 369950 }, { "epoch": 3.2674543881029336, "grad_norm": 2.7151565551757812, "learning_rate": 2.693265633768629e-05, "loss": 3.089, "step": 370000 }, { "epoch": 3.267895935993218, "grad_norm": 3.0845625400543213, "learning_rate": 2.6920328262519824e-05, "loss": 2.885, "step": 370050 }, { "epoch": 3.268337483883502, "grad_norm": 3.065298318862915, "learning_rate": 2.690800197006227e-05, "loss": 3.367, "step": 370100 }, { "epoch": 3.268779031773786, "grad_norm": 1.8330507278442383, "learning_rate": 2.6895677461265755e-05, "loss": 2.9709, "step": 370150 }, { "epoch": 3.2692205796640703, "grad_norm": 3.540588140487671, "learning_rate": 2.6883354737082255e-05, "loss": 2.7614, "step": 370200 }, { "epoch": 3.2696621275543545, "grad_norm": 1.9015660285949707, "learning_rate": 2.6871033798463574e-05, "loss": 2.918, "step": 370250 }, { "epoch": 3.2701036754446386, "grad_norm": 2.804596424102783, "learning_rate": 2.6858714646361395e-05, "loss": 3.0499, "step": 370300 }, { "epoch": 3.270545223334923, "grad_norm": 3.8938820362091064, "learning_rate": 2.6846397281727288e-05, "loss": 2.9414, "step": 370350 }, { "epoch": 3.270986771225207, "grad_norm": 3.694410562515259, "learning_rate": 2.6834081705512637e-05, "loss": 2.7302, "step": 370400 }, { "epoch": 3.271428319115491, "grad_norm": 3.453331232070923, "learning_rate": 2.6821767918668727e-05, "loss": 3.3286, "step": 370450 }, { "epoch": 3.2718698670057753, "grad_norm": 3.3738837242126465, "learning_rate": 2.6809455922146726e-05, "loss": 3.54, "step": 370500 }, { "epoch": 3.2723114148960595, "grad_norm": 8.693551063537598, "learning_rate": 2.6797145716897588e-05, "loss": 3.095, "step": 370550 }, { "epoch": 3.2727529627863436, "grad_norm": 3.389594554901123, "learning_rate": 2.6784837303872178e-05, "loss": 3.0401, "step": 370600 }, { "epoch": 3.273194510676628, "grad_norm": 6.978499412536621, "learning_rate": 2.6772530684021257e-05, "loss": 2.854, "step": 370650 }, { "epoch": 3.273636058566912, "grad_norm": 2.801917791366577, "learning_rate": 2.6760225858295375e-05, "loss": 2.9671, "step": 370700 }, { "epoch": 3.2740776064571966, "grad_norm": 1.4825330972671509, "learning_rate": 2.6747922827644955e-05, "loss": 2.8431, "step": 370750 }, { "epoch": 3.2745191543474803, "grad_norm": 3.186945915222168, "learning_rate": 2.6735621593020332e-05, "loss": 3.0739, "step": 370800 }, { "epoch": 3.274960702237765, "grad_norm": 4.93466854095459, "learning_rate": 2.6723322155371676e-05, "loss": 3.2471, "step": 370850 }, { "epoch": 3.2754022501280486, "grad_norm": 2.5676965713500977, "learning_rate": 2.6711024515648987e-05, "loss": 2.7101, "step": 370900 }, { "epoch": 3.2758437980183333, "grad_norm": 5.74196195602417, "learning_rate": 2.669872867480217e-05, "loss": 2.8619, "step": 370950 }, { "epoch": 3.2762853459086174, "grad_norm": 3.0910754203796387, "learning_rate": 2.6686434633780994e-05, "loss": 3.1865, "step": 371000 }, { "epoch": 3.2767268937989016, "grad_norm": 2.83178973197937, "learning_rate": 2.6674142393535052e-05, "loss": 2.9278, "step": 371050 }, { "epoch": 3.2771684416891858, "grad_norm": 2.9730751514434814, "learning_rate": 2.6661851955013784e-05, "loss": 2.6784, "step": 371100 }, { "epoch": 3.27760998957947, "grad_norm": 5.484926223754883, "learning_rate": 2.664956331916656e-05, "loss": 2.8702, "step": 371150 }, { "epoch": 3.278051537469754, "grad_norm": 3.825465679168701, "learning_rate": 2.6637276486942576e-05, "loss": 3.1359, "step": 371200 }, { "epoch": 3.2784930853600383, "grad_norm": 3.7067127227783203, "learning_rate": 2.662499145929086e-05, "loss": 2.8417, "step": 371250 }, { "epoch": 3.2789346332503224, "grad_norm": 4.032132625579834, "learning_rate": 2.6612708237160333e-05, "loss": 2.8039, "step": 371300 }, { "epoch": 3.2793761811406066, "grad_norm": 3.2260546684265137, "learning_rate": 2.66004268214998e-05, "loss": 2.561, "step": 371350 }, { "epoch": 3.2798177290308907, "grad_norm": 4.249605655670166, "learning_rate": 2.6588147213257874e-05, "loss": 2.8865, "step": 371400 }, { "epoch": 3.280259276921175, "grad_norm": 2.3653135299682617, "learning_rate": 2.6575869413383025e-05, "loss": 2.5965, "step": 371450 }, { "epoch": 3.280700824811459, "grad_norm": 1.6907588243484497, "learning_rate": 2.6563593422823657e-05, "loss": 3.3722, "step": 371500 }, { "epoch": 3.2811423727017432, "grad_norm": 3.8011252880096436, "learning_rate": 2.6551319242527938e-05, "loss": 2.6853, "step": 371550 }, { "epoch": 3.2815839205920274, "grad_norm": 1.284305214881897, "learning_rate": 2.6539046873443974e-05, "loss": 2.9354, "step": 371600 }, { "epoch": 3.2820254684823116, "grad_norm": 3.7194406986236572, "learning_rate": 2.6526776316519708e-05, "loss": 2.7604, "step": 371650 }, { "epoch": 3.2824670163725957, "grad_norm": 6.663399696350098, "learning_rate": 2.6514507572702914e-05, "loss": 2.4165, "step": 371700 }, { "epoch": 3.28290856426288, "grad_norm": 5.481590270996094, "learning_rate": 2.6502240642941258e-05, "loss": 2.6497, "step": 371750 }, { "epoch": 3.283350112153164, "grad_norm": 2.255086898803711, "learning_rate": 2.6489975528182265e-05, "loss": 3.1404, "step": 371800 }, { "epoch": 3.2837916600434482, "grad_norm": 2.1339943408966064, "learning_rate": 2.6477712229373315e-05, "loss": 3.0183, "step": 371850 }, { "epoch": 3.2842332079337324, "grad_norm": 0.9483595490455627, "learning_rate": 2.6465450747461608e-05, "loss": 2.8331, "step": 371900 }, { "epoch": 3.2846747558240166, "grad_norm": 6.982855319976807, "learning_rate": 2.6453191083394264e-05, "loss": 3.4471, "step": 371950 }, { "epoch": 3.2851163037143007, "grad_norm": 5.278097152709961, "learning_rate": 2.644093323811826e-05, "loss": 2.567, "step": 372000 }, { "epoch": 3.2851163037143007, "eval_asr_loss": 0.950405941320304, "eval_loss": 2.738981008529663, "eval_runtime": 20.8617, "eval_samples_per_second": 36.814, "eval_steps_per_second": 9.203, "eval_tts_loss": 6.124379059557385, "step": 372000 }, { "epoch": 3.285557851604585, "grad_norm": 6.268125534057617, "learning_rate": 2.6428677212580367e-05, "loss": 2.9766, "step": 372050 }, { "epoch": 3.285999399494869, "grad_norm": 3.320671558380127, "learning_rate": 2.6416423007727275e-05, "loss": 3.0267, "step": 372100 }, { "epoch": 3.2864409473851532, "grad_norm": 6.0972819328308105, "learning_rate": 2.640417062450555e-05, "loss": 2.9723, "step": 372150 }, { "epoch": 3.2868824952754374, "grad_norm": 3.7829928398132324, "learning_rate": 2.6391920063861553e-05, "loss": 3.0711, "step": 372200 }, { "epoch": 3.2873240431657216, "grad_norm": 5.685856342315674, "learning_rate": 2.6379671326741524e-05, "loss": 3.0841, "step": 372250 }, { "epoch": 3.2877655910560057, "grad_norm": 5.8348917961120605, "learning_rate": 2.636742441409159e-05, "loss": 2.5676, "step": 372300 }, { "epoch": 3.28820713894629, "grad_norm": 6.167929649353027, "learning_rate": 2.6355179326857736e-05, "loss": 3.3767, "step": 372350 }, { "epoch": 3.2886486868365745, "grad_norm": 3.2977242469787598, "learning_rate": 2.6342936065985764e-05, "loss": 2.9918, "step": 372400 }, { "epoch": 3.2890902347268582, "grad_norm": 3.1629815101623535, "learning_rate": 2.633069463242137e-05, "loss": 3.1198, "step": 372450 }, { "epoch": 3.289531782617143, "grad_norm": 5.478760719299316, "learning_rate": 2.631845502711013e-05, "loss": 2.525, "step": 372500 }, { "epoch": 3.289973330507427, "grad_norm": 3.024806499481201, "learning_rate": 2.63062172509974e-05, "loss": 3.1237, "step": 372550 }, { "epoch": 3.290414878397711, "grad_norm": 2.1045095920562744, "learning_rate": 2.6294226006006352e-05, "loss": 2.9587, "step": 372600 }, { "epoch": 3.2908564262879954, "grad_norm": 2.069432258605957, "learning_rate": 2.6281991854495313e-05, "loss": 3.2121, "step": 372650 }, { "epoch": 3.2912979741782795, "grad_norm": 2.391839027404785, "learning_rate": 2.6269759534999272e-05, "loss": 3.103, "step": 372700 }, { "epoch": 3.2917395220685637, "grad_norm": 3.8592689037323, "learning_rate": 2.6257529048463102e-05, "loss": 3.1525, "step": 372750 }, { "epoch": 3.292181069958848, "grad_norm": 1.7790601253509521, "learning_rate": 2.6245300395831467e-05, "loss": 3.2306, "step": 372800 }, { "epoch": 3.292622617849132, "grad_norm": 2.135446548461914, "learning_rate": 2.6233073578048962e-05, "loss": 3.1264, "step": 372850 }, { "epoch": 3.293064165739416, "grad_norm": 4.56423282623291, "learning_rate": 2.6220848596059964e-05, "loss": 2.8445, "step": 372900 }, { "epoch": 3.2935057136297003, "grad_norm": 2.9248650074005127, "learning_rate": 2.6208625450808787e-05, "loss": 3.2135, "step": 372950 }, { "epoch": 3.2939472615199845, "grad_norm": 4.764005661010742, "learning_rate": 2.619640414323954e-05, "loss": 3.1292, "step": 373000 }, { "epoch": 3.2943888094102687, "grad_norm": 6.290265083312988, "learning_rate": 2.618418467429623e-05, "loss": 3.4271, "step": 373050 }, { "epoch": 3.294830357300553, "grad_norm": 2.2528157234191895, "learning_rate": 2.617196704492269e-05, "loss": 2.713, "step": 373100 }, { "epoch": 3.295271905190837, "grad_norm": 7.205920696258545, "learning_rate": 2.6159751256062636e-05, "loss": 3.0733, "step": 373150 }, { "epoch": 3.295713453081121, "grad_norm": 4.097457408905029, "learning_rate": 2.6147537308659664e-05, "loss": 2.7877, "step": 373200 }, { "epoch": 3.2961550009714053, "grad_norm": 1.3688085079193115, "learning_rate": 2.6135325203657145e-05, "loss": 2.6522, "step": 373250 }, { "epoch": 3.2965965488616895, "grad_norm": 6.049803256988525, "learning_rate": 2.6123114941998418e-05, "loss": 2.6024, "step": 373300 }, { "epoch": 3.2970380967519737, "grad_norm": 2.6383845806121826, "learning_rate": 2.611090652462656e-05, "loss": 3.1212, "step": 373350 }, { "epoch": 3.297479644642258, "grad_norm": 3.672860622406006, "learning_rate": 2.6098699952484628e-05, "loss": 3.0163, "step": 373400 }, { "epoch": 3.297921192532542, "grad_norm": 4.3407182693481445, "learning_rate": 2.6086495226515416e-05, "loss": 2.9264, "step": 373450 }, { "epoch": 3.298362740422826, "grad_norm": 4.151040077209473, "learning_rate": 2.6074292347661678e-05, "loss": 3.1163, "step": 373500 }, { "epoch": 3.2988042883131103, "grad_norm": 5.71210241317749, "learning_rate": 2.6062091316865987e-05, "loss": 2.5954, "step": 373550 }, { "epoch": 3.2992458362033945, "grad_norm": 4.4509429931640625, "learning_rate": 2.604989213507074e-05, "loss": 2.7374, "step": 373600 }, { "epoch": 3.2996873840936787, "grad_norm": 6.1906914710998535, "learning_rate": 2.603769480321825e-05, "loss": 2.7838, "step": 373650 }, { "epoch": 3.300128931983963, "grad_norm": 7.444390773773193, "learning_rate": 2.602549932225064e-05, "loss": 2.6141, "step": 373700 }, { "epoch": 3.300570479874247, "grad_norm": 6.48679256439209, "learning_rate": 2.601330569310989e-05, "loss": 2.6359, "step": 373750 }, { "epoch": 3.301012027764531, "grad_norm": 4.4046831130981445, "learning_rate": 2.600111391673788e-05, "loss": 2.3129, "step": 373800 }, { "epoch": 3.3014535756548153, "grad_norm": 6.46185827255249, "learning_rate": 2.598892399407633e-05, "loss": 2.821, "step": 373850 }, { "epoch": 3.3018951235450995, "grad_norm": 3.2642922401428223, "learning_rate": 2.5976735926066777e-05, "loss": 2.8372, "step": 373900 }, { "epoch": 3.302336671435384, "grad_norm": 4.708306312561035, "learning_rate": 2.596454971365066e-05, "loss": 3.1926, "step": 373950 }, { "epoch": 3.302778219325668, "grad_norm": 2.263420581817627, "learning_rate": 2.5952365357769265e-05, "loss": 2.7434, "step": 374000 }, { "epoch": 3.3032197672159525, "grad_norm": 1.9232418537139893, "learning_rate": 2.5940182859363748e-05, "loss": 2.9502, "step": 374050 }, { "epoch": 3.303661315106236, "grad_norm": 4.649964809417725, "learning_rate": 2.592800221937508e-05, "loss": 3.1854, "step": 374100 }, { "epoch": 3.304102862996521, "grad_norm": 4.49143123626709, "learning_rate": 2.5915823438744096e-05, "loss": 3.275, "step": 374150 }, { "epoch": 3.304544410886805, "grad_norm": 5.052828788757324, "learning_rate": 2.590364651841155e-05, "loss": 3.1742, "step": 374200 }, { "epoch": 3.304985958777089, "grad_norm": 5.850234508514404, "learning_rate": 2.589147145931795e-05, "loss": 2.9927, "step": 374250 }, { "epoch": 3.3054275066673733, "grad_norm": 3.910665988922119, "learning_rate": 2.5879298262403752e-05, "loss": 2.8577, "step": 374300 }, { "epoch": 3.3058690545576574, "grad_norm": 7.441248416900635, "learning_rate": 2.5867126928609243e-05, "loss": 3.0715, "step": 374350 }, { "epoch": 3.3063106024479416, "grad_norm": 3.5320494174957275, "learning_rate": 2.5854957458874517e-05, "loss": 2.5929, "step": 374400 }, { "epoch": 3.306752150338226, "grad_norm": 5.141417503356934, "learning_rate": 2.584278985413961e-05, "loss": 2.8252, "step": 374450 }, { "epoch": 3.30719369822851, "grad_norm": 4.73638916015625, "learning_rate": 2.5830624115344314e-05, "loss": 2.7416, "step": 374500 }, { "epoch": 3.307635246118794, "grad_norm": 7.2794623374938965, "learning_rate": 2.5818460243428376e-05, "loss": 3.1919, "step": 374550 }, { "epoch": 3.3080767940090783, "grad_norm": 3.4228627681732178, "learning_rate": 2.5806298239331313e-05, "loss": 3.0256, "step": 374600 }, { "epoch": 3.3085183418993624, "grad_norm": 8.426271438598633, "learning_rate": 2.579413810399255e-05, "loss": 2.8009, "step": 374650 }, { "epoch": 3.3089598897896466, "grad_norm": 4.186051368713379, "learning_rate": 2.578197983835138e-05, "loss": 2.7332, "step": 374700 }, { "epoch": 3.309401437679931, "grad_norm": 6.482518196105957, "learning_rate": 2.5769823443346885e-05, "loss": 2.6176, "step": 374750 }, { "epoch": 3.309842985570215, "grad_norm": 4.789216995239258, "learning_rate": 2.5757668919918064e-05, "loss": 2.9965, "step": 374800 }, { "epoch": 3.310284533460499, "grad_norm": 6.25647497177124, "learning_rate": 2.5745516269003766e-05, "loss": 2.8842, "step": 374850 }, { "epoch": 3.3107260813507833, "grad_norm": 4.769716739654541, "learning_rate": 2.5733365491542673e-05, "loss": 2.91, "step": 374900 }, { "epoch": 3.3111676292410674, "grad_norm": 3.9893856048583984, "learning_rate": 2.5721216588473295e-05, "loss": 2.9953, "step": 374950 }, { "epoch": 3.3116091771313516, "grad_norm": 3.4443581104278564, "learning_rate": 2.570906956073408e-05, "loss": 2.8015, "step": 375000 }, { "epoch": 3.3116091771313516, "eval_asr_loss": 0.9521827825223775, "eval_loss": 2.7428600788116455, "eval_runtime": 20.8695, "eval_samples_per_second": 36.8, "eval_steps_per_second": 9.2, "eval_tts_loss": 6.1372057081555305, "step": 375000 }, { "epoch": 3.3120507250216358, "grad_norm": 4.8564558029174805, "learning_rate": 2.5696924409263234e-05, "loss": 2.6826, "step": 375050 }, { "epoch": 3.31249227291192, "grad_norm": 0.787880539894104, "learning_rate": 2.5684781134998898e-05, "loss": 3.0097, "step": 375100 }, { "epoch": 3.312933820802204, "grad_norm": 3.805016040802002, "learning_rate": 2.5672639738879033e-05, "loss": 3.4269, "step": 375150 }, { "epoch": 3.3133753686924883, "grad_norm": 1.6633998155593872, "learning_rate": 2.5660500221841476e-05, "loss": 2.8462, "step": 375200 }, { "epoch": 3.3138169165827724, "grad_norm": 6.212987899780273, "learning_rate": 2.5648362584823882e-05, "loss": 3.142, "step": 375250 }, { "epoch": 3.3142584644730566, "grad_norm": 2.942234516143799, "learning_rate": 2.5636226828763766e-05, "loss": 2.4882, "step": 375300 }, { "epoch": 3.3147000123633408, "grad_norm": 4.252694606781006, "learning_rate": 2.562409295459855e-05, "loss": 2.7819, "step": 375350 }, { "epoch": 3.315141560253625, "grad_norm": 5.904225826263428, "learning_rate": 2.561196096326543e-05, "loss": 2.4911, "step": 375400 }, { "epoch": 3.315583108143909, "grad_norm": 4.301248550415039, "learning_rate": 2.5599830855701524e-05, "loss": 3.0307, "step": 375450 }, { "epoch": 3.3160246560341937, "grad_norm": 4.971327304840088, "learning_rate": 2.558770263284379e-05, "loss": 2.7381, "step": 375500 }, { "epoch": 3.3164662039244774, "grad_norm": 4.56104040145874, "learning_rate": 2.5575576295629e-05, "loss": 2.9205, "step": 375550 }, { "epoch": 3.316907751814762, "grad_norm": 3.9898462295532227, "learning_rate": 2.556345184499385e-05, "loss": 2.9436, "step": 375600 }, { "epoch": 3.3173492997050458, "grad_norm": 2.4277539253234863, "learning_rate": 2.5551329281874814e-05, "loss": 3.1212, "step": 375650 }, { "epoch": 3.3177908475953304, "grad_norm": 5.662060260772705, "learning_rate": 2.5539451002188707e-05, "loss": 2.8297, "step": 375700 }, { "epoch": 3.3182323954856145, "grad_norm": 3.864311695098877, "learning_rate": 2.552733217911395e-05, "loss": 3.0274, "step": 375750 }, { "epoch": 3.3186739433758987, "grad_norm": 1.3516720533370972, "learning_rate": 2.551545756646956e-05, "loss": 3.1131, "step": 375800 }, { "epoch": 3.319115491266183, "grad_norm": 3.8653006553649902, "learning_rate": 2.550334248710883e-05, "loss": 2.8724, "step": 375850 }, { "epoch": 3.319557039156467, "grad_norm": 3.0317821502685547, "learning_rate": 2.5491229299907184e-05, "loss": 2.6834, "step": 375900 }, { "epoch": 3.319998587046751, "grad_norm": 4.209009647369385, "learning_rate": 2.5479118005800235e-05, "loss": 2.9197, "step": 375950 }, { "epoch": 3.3204401349370354, "grad_norm": 6.930157661437988, "learning_rate": 2.5467008605723503e-05, "loss": 2.5583, "step": 376000 }, { "epoch": 3.3208816828273195, "grad_norm": 3.9497265815734863, "learning_rate": 2.5454901100612303e-05, "loss": 3.467, "step": 376050 }, { "epoch": 3.3213232307176037, "grad_norm": 2.472597360610962, "learning_rate": 2.544279549140186e-05, "loss": 2.968, "step": 376100 }, { "epoch": 3.321764778607888, "grad_norm": 5.13215970993042, "learning_rate": 2.5430691779027243e-05, "loss": 3.0367, "step": 376150 }, { "epoch": 3.322206326498172, "grad_norm": 6.048874378204346, "learning_rate": 2.5418589964423322e-05, "loss": 2.8998, "step": 376200 }, { "epoch": 3.322647874388456, "grad_norm": 1.8569426536560059, "learning_rate": 2.5406490048524912e-05, "loss": 3.1732, "step": 376250 }, { "epoch": 3.3230894222787404, "grad_norm": 4.91082239151001, "learning_rate": 2.539439203226659e-05, "loss": 2.4815, "step": 376300 }, { "epoch": 3.3235309701690245, "grad_norm": 3.3280913829803467, "learning_rate": 2.538229591658282e-05, "loss": 2.9883, "step": 376350 }, { "epoch": 3.3239725180593087, "grad_norm": 3.560392379760742, "learning_rate": 2.5370201702407936e-05, "loss": 3.212, "step": 376400 }, { "epoch": 3.324414065949593, "grad_norm": 2.7243807315826416, "learning_rate": 2.5358109390676134e-05, "loss": 3.0275, "step": 376450 }, { "epoch": 3.324855613839877, "grad_norm": 3.7533113956451416, "learning_rate": 2.5346018982321406e-05, "loss": 2.9021, "step": 376500 }, { "epoch": 3.325297161730161, "grad_norm": 4.604795932769775, "learning_rate": 2.5333930478277645e-05, "loss": 3.1585, "step": 376550 }, { "epoch": 3.3257387096204454, "grad_norm": 5.722384452819824, "learning_rate": 2.5321843879478625e-05, "loss": 2.7806, "step": 376600 }, { "epoch": 3.3261802575107295, "grad_norm": 11.98047161102295, "learning_rate": 2.530975918685789e-05, "loss": 2.5913, "step": 376650 }, { "epoch": 3.3266218054010137, "grad_norm": 3.1532790660858154, "learning_rate": 2.529767640134888e-05, "loss": 2.8894, "step": 376700 }, { "epoch": 3.327063353291298, "grad_norm": 3.2671260833740234, "learning_rate": 2.5285595523884887e-05, "loss": 2.5188, "step": 376750 }, { "epoch": 3.327504901181582, "grad_norm": 3.5208120346069336, "learning_rate": 2.5273516555399095e-05, "loss": 3.1518, "step": 376800 }, { "epoch": 3.327946449071866, "grad_norm": 5.36313009262085, "learning_rate": 2.5261439496824456e-05, "loss": 2.8472, "step": 376850 }, { "epoch": 3.3283879969621504, "grad_norm": 1.8486543893814087, "learning_rate": 2.524936434909384e-05, "loss": 2.9884, "step": 376900 }, { "epoch": 3.3288295448524345, "grad_norm": 9.185713768005371, "learning_rate": 2.523729111313996e-05, "loss": 2.4859, "step": 376950 }, { "epoch": 3.3292710927427187, "grad_norm": 5.139148235321045, "learning_rate": 2.522521978989535e-05, "loss": 2.7501, "step": 377000 }, { "epoch": 3.329712640633003, "grad_norm": 3.7008445262908936, "learning_rate": 2.5213150380292438e-05, "loss": 2.8322, "step": 377050 }, { "epoch": 3.330154188523287, "grad_norm": 5.4118123054504395, "learning_rate": 2.520108288526346e-05, "loss": 2.7699, "step": 377100 }, { "epoch": 3.3305957364135717, "grad_norm": 3.5684123039245605, "learning_rate": 2.518901730574056e-05, "loss": 2.901, "step": 377150 }, { "epoch": 3.3310372843038554, "grad_norm": 5.418545722961426, "learning_rate": 2.517695364265566e-05, "loss": 3.447, "step": 377200 }, { "epoch": 3.33147883219414, "grad_norm": 5.5712785720825195, "learning_rate": 2.5164891896940595e-05, "loss": 3.2458, "step": 377250 }, { "epoch": 3.331920380084424, "grad_norm": 2.6971917152404785, "learning_rate": 2.515283206952706e-05, "loss": 2.6539, "step": 377300 }, { "epoch": 3.3323619279747083, "grad_norm": 6.754800319671631, "learning_rate": 2.5140774161346526e-05, "loss": 2.9203, "step": 377350 }, { "epoch": 3.3328034758649925, "grad_norm": 4.3273606300354, "learning_rate": 2.5128718173330413e-05, "loss": 2.9394, "step": 377400 }, { "epoch": 3.3332450237552766, "grad_norm": 6.782110214233398, "learning_rate": 2.5116664106409914e-05, "loss": 2.6577, "step": 377450 }, { "epoch": 3.333686571645561, "grad_norm": 5.524333477020264, "learning_rate": 2.5104611961516088e-05, "loss": 3.1379, "step": 377500 }, { "epoch": 3.334128119535845, "grad_norm": 2.3339273929595947, "learning_rate": 2.509256173957989e-05, "loss": 3.0811, "step": 377550 }, { "epoch": 3.334569667426129, "grad_norm": 4.622811794281006, "learning_rate": 2.50805134415321e-05, "loss": 3.6535, "step": 377600 }, { "epoch": 3.3350112153164133, "grad_norm": 7.743923187255859, "learning_rate": 2.5068467068303313e-05, "loss": 2.8154, "step": 377650 }, { "epoch": 3.3354527632066975, "grad_norm": 5.135678768157959, "learning_rate": 2.5056422620824037e-05, "loss": 2.8673, "step": 377700 }, { "epoch": 3.3358943110969816, "grad_norm": 3.572396755218506, "learning_rate": 2.5044380100024602e-05, "loss": 2.3311, "step": 377750 }, { "epoch": 3.336335858987266, "grad_norm": 4.8725409507751465, "learning_rate": 2.50323395068352e-05, "loss": 3.1408, "step": 377800 }, { "epoch": 3.33677740687755, "grad_norm": 3.3588948249816895, "learning_rate": 2.502030084218585e-05, "loss": 2.8721, "step": 377850 }, { "epoch": 3.337218954767834, "grad_norm": 3.6829309463500977, "learning_rate": 2.500826410700643e-05, "loss": 3.1747, "step": 377900 }, { "epoch": 3.3376605026581183, "grad_norm": 2.7941019535064697, "learning_rate": 2.499622930222669e-05, "loss": 2.7258, "step": 377950 }, { "epoch": 3.3381020505484025, "grad_norm": 5.074127674102783, "learning_rate": 2.4984196428776202e-05, "loss": 2.182, "step": 378000 }, { "epoch": 3.3381020505484025, "eval_asr_loss": 0.9470043493633901, "eval_loss": 2.7416553497314453, "eval_runtime": 20.6553, "eval_samples_per_second": 37.182, "eval_steps_per_second": 9.295, "eval_tts_loss": 6.143897192445007, "step": 378000 }, { "epoch": 3.3385435984386866, "grad_norm": 4.48736047744751, "learning_rate": 2.4972165487584405e-05, "loss": 2.8298, "step": 378050 }, { "epoch": 3.338985146328971, "grad_norm": 6.250656604766846, "learning_rate": 2.4960136479580616e-05, "loss": 3.5347, "step": 378100 }, { "epoch": 3.339426694219255, "grad_norm": 7.315870761871338, "learning_rate": 2.4948109405693932e-05, "loss": 2.8666, "step": 378150 }, { "epoch": 3.339868242109539, "grad_norm": 2.783646583557129, "learning_rate": 2.4936084266853383e-05, "loss": 2.6819, "step": 378200 }, { "epoch": 3.3403097899998233, "grad_norm": 5.976446628570557, "learning_rate": 2.4924061063987787e-05, "loss": 2.6786, "step": 378250 }, { "epoch": 3.3407513378901075, "grad_norm": 6.088388442993164, "learning_rate": 2.491203979802582e-05, "loss": 3.2882, "step": 378300 }, { "epoch": 3.3411928857803916, "grad_norm": 1.967189908027649, "learning_rate": 2.4900020469896034e-05, "loss": 3.5225, "step": 378350 }, { "epoch": 3.341634433670676, "grad_norm": 6.175055027008057, "learning_rate": 2.4888003080526827e-05, "loss": 2.9202, "step": 378400 }, { "epoch": 3.34207598156096, "grad_norm": 2.141141176223755, "learning_rate": 2.4875987630846458e-05, "loss": 2.8723, "step": 378450 }, { "epoch": 3.342517529451244, "grad_norm": 6.31662130355835, "learning_rate": 2.4863974121782985e-05, "loss": 3.2652, "step": 378500 }, { "epoch": 3.3429590773415283, "grad_norm": 4.271336555480957, "learning_rate": 2.485220276658161e-05, "loss": 3.004, "step": 378550 }, { "epoch": 3.3434006252318125, "grad_norm": 4.244155406951904, "learning_rate": 2.484019310267709e-05, "loss": 3.0281, "step": 378600 }, { "epoch": 3.3438421731220966, "grad_norm": 2.5814459323883057, "learning_rate": 2.4828185382154313e-05, "loss": 3.0145, "step": 378650 }, { "epoch": 3.3442837210123812, "grad_norm": 1.5591593980789185, "learning_rate": 2.4816179605940747e-05, "loss": 2.4924, "step": 378700 }, { "epoch": 3.344725268902665, "grad_norm": 5.4920854568481445, "learning_rate": 2.4804175774963763e-05, "loss": 2.5307, "step": 378750 }, { "epoch": 3.3451668167929496, "grad_norm": 7.952885150909424, "learning_rate": 2.479217389015053e-05, "loss": 2.5393, "step": 378800 }, { "epoch": 3.3456083646832333, "grad_norm": 1.880053997039795, "learning_rate": 2.4780173952428142e-05, "loss": 2.7888, "step": 378850 }, { "epoch": 3.346049912573518, "grad_norm": 2.4331517219543457, "learning_rate": 2.476817596272345e-05, "loss": 2.7803, "step": 378900 }, { "epoch": 3.346491460463802, "grad_norm": 2.091317892074585, "learning_rate": 2.475617992196319e-05, "loss": 2.752, "step": 378950 }, { "epoch": 3.3469330083540862, "grad_norm": 6.151158809661865, "learning_rate": 2.4744185831073984e-05, "loss": 3.0115, "step": 379000 }, { "epoch": 3.3473745562443704, "grad_norm": 3.298893690109253, "learning_rate": 2.4732193690982246e-05, "loss": 2.8777, "step": 379050 }, { "epoch": 3.3478161041346546, "grad_norm": 4.188007831573486, "learning_rate": 2.472020350261428e-05, "loss": 2.7808, "step": 379100 }, { "epoch": 3.3482576520249387, "grad_norm": 3.2157561779022217, "learning_rate": 2.4708215266896246e-05, "loss": 2.7992, "step": 379150 }, { "epoch": 3.348699199915223, "grad_norm": 4.4568657875061035, "learning_rate": 2.469622898475409e-05, "loss": 2.7617, "step": 379200 }, { "epoch": 3.349140747805507, "grad_norm": 2.546607494354248, "learning_rate": 2.468424465711368e-05, "loss": 2.8818, "step": 379250 }, { "epoch": 3.3495822956957912, "grad_norm": 4.784109115600586, "learning_rate": 2.4672262284900715e-05, "loss": 3.2977, "step": 379300 }, { "epoch": 3.3500238435860754, "grad_norm": 4.474793434143066, "learning_rate": 2.466028186904072e-05, "loss": 3.1629, "step": 379350 }, { "epoch": 3.3504653914763596, "grad_norm": 4.468326568603516, "learning_rate": 2.4648303410459055e-05, "loss": 3.022, "step": 379400 }, { "epoch": 3.3509069393666437, "grad_norm": 5.769155025482178, "learning_rate": 2.4636326910080975e-05, "loss": 2.7111, "step": 379450 }, { "epoch": 3.351348487256928, "grad_norm": 5.179945468902588, "learning_rate": 2.4624352368831582e-05, "loss": 3.2336, "step": 379500 }, { "epoch": 3.351790035147212, "grad_norm": 2.723907947540283, "learning_rate": 2.461237978763577e-05, "loss": 2.5225, "step": 379550 }, { "epoch": 3.3522315830374962, "grad_norm": 1.6412147283554077, "learning_rate": 2.4600409167418343e-05, "loss": 2.7974, "step": 379600 }, { "epoch": 3.3526731309277804, "grad_norm": 5.200442790985107, "learning_rate": 2.4588440509103938e-05, "loss": 2.8161, "step": 379650 }, { "epoch": 3.3531146788180646, "grad_norm": 6.861469268798828, "learning_rate": 2.4576473813617024e-05, "loss": 2.7693, "step": 379700 }, { "epoch": 3.3535562267083487, "grad_norm": 4.293955326080322, "learning_rate": 2.4564509081881902e-05, "loss": 2.9066, "step": 379750 }, { "epoch": 3.353997774598633, "grad_norm": 4.884483337402344, "learning_rate": 2.455254631482279e-05, "loss": 3.1278, "step": 379800 }, { "epoch": 3.354439322488917, "grad_norm": 4.393824577331543, "learning_rate": 2.4540585513363663e-05, "loss": 2.6881, "step": 379850 }, { "epoch": 3.3548808703792012, "grad_norm": 4.20957612991333, "learning_rate": 2.4528626678428412e-05, "loss": 3.0088, "step": 379900 }, { "epoch": 3.3553224182694854, "grad_norm": 2.523052930831909, "learning_rate": 2.451666981094077e-05, "loss": 2.5315, "step": 379950 }, { "epoch": 3.3557639661597696, "grad_norm": 3.5663928985595703, "learning_rate": 2.4504714911824304e-05, "loss": 2.911, "step": 380000 }, { "epoch": 3.3562055140500537, "grad_norm": 4.422933101654053, "learning_rate": 2.449276198200242e-05, "loss": 2.8543, "step": 380050 }, { "epoch": 3.356647061940338, "grad_norm": 5.210747718811035, "learning_rate": 2.4480811022398358e-05, "loss": 2.9402, "step": 380100 }, { "epoch": 3.357088609830622, "grad_norm": 3.490899085998535, "learning_rate": 2.446886203393527e-05, "loss": 3.4238, "step": 380150 }, { "epoch": 3.3575301577209062, "grad_norm": 1.6728489398956299, "learning_rate": 2.445691501753607e-05, "loss": 3.0469, "step": 380200 }, { "epoch": 3.3579717056111904, "grad_norm": 5.102583408355713, "learning_rate": 2.4444969974123582e-05, "loss": 2.8076, "step": 380250 }, { "epoch": 3.3584132535014746, "grad_norm": 3.439706563949585, "learning_rate": 2.4433026904620487e-05, "loss": 3.166, "step": 380300 }, { "epoch": 3.358854801391759, "grad_norm": 4.2218427658081055, "learning_rate": 2.4421085809949246e-05, "loss": 2.4494, "step": 380350 }, { "epoch": 3.359296349282043, "grad_norm": 3.897658586502075, "learning_rate": 2.440914669103222e-05, "loss": 3.0835, "step": 380400 }, { "epoch": 3.3597378971723275, "grad_norm": 5.852313995361328, "learning_rate": 2.4397209548791622e-05, "loss": 3.0862, "step": 380450 }, { "epoch": 3.3601794450626117, "grad_norm": 1.6864207983016968, "learning_rate": 2.4385274384149486e-05, "loss": 2.4941, "step": 380500 }, { "epoch": 3.360620992952896, "grad_norm": 6.261626243591309, "learning_rate": 2.437334119802768e-05, "loss": 2.6029, "step": 380550 }, { "epoch": 3.36106254084318, "grad_norm": 1.9894605875015259, "learning_rate": 2.4361409991347956e-05, "loss": 3.0176, "step": 380600 }, { "epoch": 3.361504088733464, "grad_norm": 1.7390474081039429, "learning_rate": 2.4349480765031925e-05, "loss": 2.5422, "step": 380650 }, { "epoch": 3.3619456366237483, "grad_norm": 2.68428897857666, "learning_rate": 2.433755352000097e-05, "loss": 3.1022, "step": 380700 }, { "epoch": 3.3623871845140325, "grad_norm": 2.1171858310699463, "learning_rate": 2.4325628257176397e-05, "loss": 2.8789, "step": 380750 }, { "epoch": 3.3628287324043167, "grad_norm": 4.493571758270264, "learning_rate": 2.4313704977479347e-05, "loss": 2.6734, "step": 380800 }, { "epoch": 3.363270280294601, "grad_norm": 4.650853157043457, "learning_rate": 2.4301783681830774e-05, "loss": 2.7238, "step": 380850 }, { "epoch": 3.363711828184885, "grad_norm": 7.235388278961182, "learning_rate": 2.4289864371151478e-05, "loss": 2.8104, "step": 380900 }, { "epoch": 3.364153376075169, "grad_norm": 3.248067617416382, "learning_rate": 2.427818537339026e-05, "loss": 2.7859, "step": 380950 }, { "epoch": 3.3645949239654533, "grad_norm": 4.240749835968018, "learning_rate": 2.426626999566617e-05, "loss": 2.9949, "step": 381000 }, { "epoch": 3.3645949239654533, "eval_asr_loss": 0.950490068372783, "eval_loss": 2.7399463653564453, "eval_runtime": 20.6873, "eval_samples_per_second": 37.124, "eval_steps_per_second": 9.281, "eval_tts_loss": 6.148342982882364, "step": 381000 }, { "epoch": 3.3650364718557375, "grad_norm": 5.221550464630127, "learning_rate": 2.4254356605654526e-05, "loss": 2.9242, "step": 381050 }, { "epoch": 3.3654780197460217, "grad_norm": 5.567101955413818, "learning_rate": 2.4242445204275504e-05, "loss": 3.2642, "step": 381100 }, { "epoch": 3.365919567636306, "grad_norm": 3.9835822582244873, "learning_rate": 2.4230535792449182e-05, "loss": 2.7745, "step": 381150 }, { "epoch": 3.36636111552659, "grad_norm": 1.5700029134750366, "learning_rate": 2.421862837109546e-05, "loss": 2.8466, "step": 381200 }, { "epoch": 3.366802663416874, "grad_norm": 5.382678985595703, "learning_rate": 2.4206722941134052e-05, "loss": 2.8921, "step": 381250 }, { "epoch": 3.3672442113071583, "grad_norm": 6.508884906768799, "learning_rate": 2.419481950348458e-05, "loss": 3.0813, "step": 381300 }, { "epoch": 3.3676857591974425, "grad_norm": 4.441015243530273, "learning_rate": 2.4182918059066495e-05, "loss": 2.7944, "step": 381350 }, { "epoch": 3.3681273070877267, "grad_norm": 5.033262729644775, "learning_rate": 2.417101860879905e-05, "loss": 3.0252, "step": 381400 }, { "epoch": 3.368568854978011, "grad_norm": 4.431244850158691, "learning_rate": 2.4159121153601393e-05, "loss": 2.6142, "step": 381450 }, { "epoch": 3.369010402868295, "grad_norm": 8.750272750854492, "learning_rate": 2.4147225694392528e-05, "loss": 3.2703, "step": 381500 }, { "epoch": 3.369451950758579, "grad_norm": 1.6918433904647827, "learning_rate": 2.413533223209123e-05, "loss": 2.3927, "step": 381550 }, { "epoch": 3.3698934986488633, "grad_norm": 5.987268447875977, "learning_rate": 2.412344076761621e-05, "loss": 3.0302, "step": 381600 }, { "epoch": 3.3703350465391475, "grad_norm": 1.6268651485443115, "learning_rate": 2.4111551301885938e-05, "loss": 2.6781, "step": 381650 }, { "epoch": 3.3707765944294317, "grad_norm": 4.814221382141113, "learning_rate": 2.4099663835818825e-05, "loss": 2.8915, "step": 381700 }, { "epoch": 3.371218142319716, "grad_norm": 5.893916606903076, "learning_rate": 2.4087778370333037e-05, "loss": 2.5932, "step": 381750 }, { "epoch": 3.37165969021, "grad_norm": 6.502525329589844, "learning_rate": 2.4075894906346634e-05, "loss": 3.4467, "step": 381800 }, { "epoch": 3.372101238100284, "grad_norm": 4.244440078735352, "learning_rate": 2.4064013444777544e-05, "loss": 2.9914, "step": 381850 }, { "epoch": 3.372542785990569, "grad_norm": 4.186789035797119, "learning_rate": 2.405213398654346e-05, "loss": 2.9036, "step": 381900 }, { "epoch": 3.3729843338808525, "grad_norm": 5.767834186553955, "learning_rate": 2.4040256532562022e-05, "loss": 2.9419, "step": 381950 }, { "epoch": 3.373425881771137, "grad_norm": 8.335840225219727, "learning_rate": 2.402838108375063e-05, "loss": 2.7642, "step": 382000 }, { "epoch": 3.373867429661421, "grad_norm": 5.141021251678467, "learning_rate": 2.401650764102654e-05, "loss": 2.8137, "step": 382050 }, { "epoch": 3.3743089775517054, "grad_norm": 5.683232307434082, "learning_rate": 2.4004636205306914e-05, "loss": 3.3104, "step": 382100 }, { "epoch": 3.3747505254419896, "grad_norm": 3.7468926906585693, "learning_rate": 2.3992766777508695e-05, "loss": 2.9705, "step": 382150 }, { "epoch": 3.375192073332274, "grad_norm": 3.9120051860809326, "learning_rate": 2.3980899358548735e-05, "loss": 2.9973, "step": 382200 }, { "epoch": 3.375633621222558, "grad_norm": 3.0653746128082275, "learning_rate": 2.3969033949343645e-05, "loss": 2.5803, "step": 382250 }, { "epoch": 3.376075169112842, "grad_norm": 8.145684242248535, "learning_rate": 2.395717055080996e-05, "loss": 2.6094, "step": 382300 }, { "epoch": 3.3765167170031263, "grad_norm": 3.5488646030426025, "learning_rate": 2.3945309163864016e-05, "loss": 2.6796, "step": 382350 }, { "epoch": 3.3769582648934104, "grad_norm": 2.502953290939331, "learning_rate": 2.393344978942198e-05, "loss": 2.6765, "step": 382400 }, { "epoch": 3.3773998127836946, "grad_norm": 3.9764809608459473, "learning_rate": 2.3921592428399913e-05, "loss": 2.9598, "step": 382450 }, { "epoch": 3.3778413606739788, "grad_norm": 2.603874921798706, "learning_rate": 2.3909737081713713e-05, "loss": 2.5481, "step": 382500 }, { "epoch": 3.378282908564263, "grad_norm": 6.192502975463867, "learning_rate": 2.389788375027906e-05, "loss": 2.8196, "step": 382550 }, { "epoch": 3.378724456454547, "grad_norm": 2.928727626800537, "learning_rate": 2.3886032435011553e-05, "loss": 2.7501, "step": 382600 }, { "epoch": 3.3791660043448313, "grad_norm": 7.366745948791504, "learning_rate": 2.38741831368266e-05, "loss": 3.097, "step": 382650 }, { "epoch": 3.3796075522351154, "grad_norm": 5.958083152770996, "learning_rate": 2.3862335856639472e-05, "loss": 2.9558, "step": 382700 }, { "epoch": 3.3800491001253996, "grad_norm": 5.628251552581787, "learning_rate": 2.3850490595365266e-05, "loss": 3.1595, "step": 382750 }, { "epoch": 3.3804906480156838, "grad_norm": 2.764981269836426, "learning_rate": 2.3838647353918898e-05, "loss": 2.8486, "step": 382800 }, { "epoch": 3.380932195905968, "grad_norm": 3.7800323963165283, "learning_rate": 2.3826806133215196e-05, "loss": 2.926, "step": 382850 }, { "epoch": 3.381373743796252, "grad_norm": 8.57502555847168, "learning_rate": 2.3814966934168763e-05, "loss": 2.9157, "step": 382900 }, { "epoch": 3.3818152916865363, "grad_norm": 3.874506950378418, "learning_rate": 2.3803129757694094e-05, "loss": 2.672, "step": 382950 }, { "epoch": 3.3822568395768204, "grad_norm": 1.7760138511657715, "learning_rate": 2.3791294604705522e-05, "loss": 2.838, "step": 383000 }, { "epoch": 3.3826983874671046, "grad_norm": 3.102895975112915, "learning_rate": 2.377946147611718e-05, "loss": 2.8025, "step": 383050 }, { "epoch": 3.3831399353573888, "grad_norm": 5.098854064941406, "learning_rate": 2.376763037284312e-05, "loss": 2.9997, "step": 383100 }, { "epoch": 3.383581483247673, "grad_norm": 7.01038122177124, "learning_rate": 2.375603785747512e-05, "loss": 2.3773, "step": 383150 }, { "epoch": 3.384023031137957, "grad_norm": 7.390100955963135, "learning_rate": 2.3744210767019165e-05, "loss": 2.9371, "step": 383200 }, { "epoch": 3.3844645790282413, "grad_norm": 6.1094746589660645, "learning_rate": 2.37323857046003e-05, "loss": 2.5939, "step": 383250 }, { "epoch": 3.3849061269185254, "grad_norm": 3.5000240802764893, "learning_rate": 2.3720562671131886e-05, "loss": 2.5045, "step": 383300 }, { "epoch": 3.3853476748088096, "grad_norm": 8.11604118347168, "learning_rate": 2.370874166752715e-05, "loss": 2.4961, "step": 383350 }, { "epoch": 3.3857892226990938, "grad_norm": 5.7944488525390625, "learning_rate": 2.36969226946992e-05, "loss": 3.2031, "step": 383400 }, { "epoch": 3.386230770589378, "grad_norm": 4.015014171600342, "learning_rate": 2.368510575356094e-05, "loss": 2.8173, "step": 383450 }, { "epoch": 3.386672318479662, "grad_norm": 4.066315650939941, "learning_rate": 2.3673290845025088e-05, "loss": 2.8246, "step": 383500 }, { "epoch": 3.3871138663699467, "grad_norm": 5.2870917320251465, "learning_rate": 2.3661477970004293e-05, "loss": 2.7576, "step": 383550 }, { "epoch": 3.3875554142602304, "grad_norm": 2.0445380210876465, "learning_rate": 2.364966712941096e-05, "loss": 3.0817, "step": 383600 }, { "epoch": 3.387996962150515, "grad_norm": 2.645881175994873, "learning_rate": 2.3637858324157396e-05, "loss": 2.7418, "step": 383650 }, { "epoch": 3.388438510040799, "grad_norm": 3.976471424102783, "learning_rate": 2.3626051555155738e-05, "loss": 3.2376, "step": 383700 }, { "epoch": 3.3888800579310834, "grad_norm": 2.6509132385253906, "learning_rate": 2.361424682331797e-05, "loss": 2.9809, "step": 383750 }, { "epoch": 3.3893216058213675, "grad_norm": 2.980285167694092, "learning_rate": 2.3602444129555888e-05, "loss": 3.0731, "step": 383800 }, { "epoch": 3.3897631537116517, "grad_norm": 6.339202404022217, "learning_rate": 2.3590643474781132e-05, "loss": 2.8099, "step": 383850 }, { "epoch": 3.390204701601936, "grad_norm": 2.6324493885040283, "learning_rate": 2.357884485990525e-05, "loss": 2.6377, "step": 383900 }, { "epoch": 3.39064624949222, "grad_norm": 1.6798341274261475, "learning_rate": 2.3567048285839533e-05, "loss": 3.155, "step": 383950 }, { "epoch": 3.391087797382504, "grad_norm": 3.7567179203033447, "learning_rate": 2.3555253753495186e-05, "loss": 3.2283, "step": 384000 }, { "epoch": 3.391087797382504, "eval_asr_loss": 0.9428989008356118, "eval_loss": 2.733029365539551, "eval_runtime": 21.0008, "eval_samples_per_second": 36.57, "eval_steps_per_second": 9.143, "eval_tts_loss": 6.149854987856302, "step": 384000 }, { "epoch": 3.3915293452727884, "grad_norm": 6.371440887451172, "learning_rate": 2.3543461263783267e-05, "loss": 2.9349, "step": 384050 }, { "epoch": 3.3919708931630725, "grad_norm": 4.825833797454834, "learning_rate": 2.3531670817614597e-05, "loss": 2.8061, "step": 384100 }, { "epoch": 3.3924124410533567, "grad_norm": 1.067299723625183, "learning_rate": 2.351988241589992e-05, "loss": 3.1445, "step": 384150 }, { "epoch": 3.392853988943641, "grad_norm": 1.6462725400924683, "learning_rate": 2.3508096059549796e-05, "loss": 2.4785, "step": 384200 }, { "epoch": 3.393295536833925, "grad_norm": 8.256317138671875, "learning_rate": 2.3496311749474604e-05, "loss": 3.3, "step": 384250 }, { "epoch": 3.393737084724209, "grad_norm": 4.388564109802246, "learning_rate": 2.3484529486584566e-05, "loss": 2.9063, "step": 384300 }, { "epoch": 3.3941786326144934, "grad_norm": 4.312572956085205, "learning_rate": 2.3472749271789783e-05, "loss": 3.0582, "step": 384350 }, { "epoch": 3.3946201805047775, "grad_norm": 6.066956043243408, "learning_rate": 2.3460971106000192e-05, "loss": 2.5798, "step": 384400 }, { "epoch": 3.3950617283950617, "grad_norm": 1.4451634883880615, "learning_rate": 2.3449194990125522e-05, "loss": 2.672, "step": 384450 }, { "epoch": 3.395503276285346, "grad_norm": 3.635267972946167, "learning_rate": 2.3437420925075393e-05, "loss": 3.1509, "step": 384500 }, { "epoch": 3.39594482417563, "grad_norm": 5.167250156402588, "learning_rate": 2.3425648911759268e-05, "loss": 3.0864, "step": 384550 }, { "epoch": 3.396386372065914, "grad_norm": 1.7503548860549927, "learning_rate": 2.3413878951086428e-05, "loss": 3.0482, "step": 384600 }, { "epoch": 3.3968279199561984, "grad_norm": 5.39633846282959, "learning_rate": 2.3402111043965974e-05, "loss": 3.3, "step": 384650 }, { "epoch": 3.3972694678464825, "grad_norm": 2.6826565265655518, "learning_rate": 2.3390345191306916e-05, "loss": 2.8195, "step": 384700 }, { "epoch": 3.3977110157367667, "grad_norm": 6.6015167236328125, "learning_rate": 2.3378581394018028e-05, "loss": 2.9021, "step": 384750 }, { "epoch": 3.398152563627051, "grad_norm": 4.960177898406982, "learning_rate": 2.336681965300799e-05, "loss": 2.9673, "step": 384800 }, { "epoch": 3.398594111517335, "grad_norm": 4.001350402832031, "learning_rate": 2.3355059969185284e-05, "loss": 2.9072, "step": 384850 }, { "epoch": 3.399035659407619, "grad_norm": 3.0844600200653076, "learning_rate": 2.334330234345828e-05, "loss": 3.0247, "step": 384900 }, { "epoch": 3.3994772072979034, "grad_norm": 2.7330141067504883, "learning_rate": 2.3331546776735102e-05, "loss": 3.0347, "step": 384950 }, { "epoch": 3.3999187551881875, "grad_norm": 1.7793247699737549, "learning_rate": 2.331979326992381e-05, "loss": 2.8704, "step": 385000 }, { "epoch": 3.4003603030784717, "grad_norm": 3.8001232147216797, "learning_rate": 2.330804182393225e-05, "loss": 2.912, "step": 385050 }, { "epoch": 3.4008018509687563, "grad_norm": 5.931512832641602, "learning_rate": 2.3296292439668105e-05, "loss": 3.1148, "step": 385100 }, { "epoch": 3.40124339885904, "grad_norm": 6.2731499671936035, "learning_rate": 2.328454511803892e-05, "loss": 2.7144, "step": 385150 }, { "epoch": 3.4016849467493246, "grad_norm": 6.224213600158691, "learning_rate": 2.327279985995211e-05, "loss": 3.0803, "step": 385200 }, { "epoch": 3.4021264946396084, "grad_norm": 4.137972831726074, "learning_rate": 2.3261056666314846e-05, "loss": 3.0803, "step": 385250 }, { "epoch": 3.402568042529893, "grad_norm": 3.287224769592285, "learning_rate": 2.3249315538034216e-05, "loss": 2.8371, "step": 385300 }, { "epoch": 3.403009590420177, "grad_norm": 5.47394323348999, "learning_rate": 2.323757647601714e-05, "loss": 3.3123, "step": 385350 }, { "epoch": 3.4034511383104613, "grad_norm": 3.8174657821655273, "learning_rate": 2.3225839481170348e-05, "loss": 3.4438, "step": 385400 }, { "epoch": 3.4038926862007455, "grad_norm": 5.326727867126465, "learning_rate": 2.3214104554400394e-05, "loss": 2.9742, "step": 385450 }, { "epoch": 3.4043342340910296, "grad_norm": 4.428962707519531, "learning_rate": 2.3202371696613727e-05, "loss": 2.5034, "step": 385500 }, { "epoch": 3.404775781981314, "grad_norm": 3.9922146797180176, "learning_rate": 2.3190640908716628e-05, "loss": 3.0344, "step": 385550 }, { "epoch": 3.405217329871598, "grad_norm": 2.5178258419036865, "learning_rate": 2.317891219161516e-05, "loss": 3.5642, "step": 385600 }, { "epoch": 3.405658877761882, "grad_norm": 2.672956943511963, "learning_rate": 2.316718554621529e-05, "loss": 3.0239, "step": 385650 }, { "epoch": 3.4061004256521663, "grad_norm": 5.756394386291504, "learning_rate": 2.3155460973422826e-05, "loss": 2.6837, "step": 385700 }, { "epoch": 3.4065419735424505, "grad_norm": 3.1607789993286133, "learning_rate": 2.3143738474143362e-05, "loss": 2.6665, "step": 385750 }, { "epoch": 3.4069835214327346, "grad_norm": 3.383699893951416, "learning_rate": 2.3132018049282346e-05, "loss": 3.0866, "step": 385800 }, { "epoch": 3.407425069323019, "grad_norm": 2.3588569164276123, "learning_rate": 2.3120299699745122e-05, "loss": 2.7499, "step": 385850 }, { "epoch": 3.407866617213303, "grad_norm": 6.273465633392334, "learning_rate": 2.3108583426436797e-05, "loss": 2.721, "step": 385900 }, { "epoch": 3.408308165103587, "grad_norm": 3.3724794387817383, "learning_rate": 2.309686923026237e-05, "loss": 2.9358, "step": 385950 }, { "epoch": 3.4087497129938713, "grad_norm": 5.635950565338135, "learning_rate": 2.3085157112126677e-05, "loss": 2.6246, "step": 386000 }, { "epoch": 3.4091912608841555, "grad_norm": 5.1959547996521, "learning_rate": 2.307344707293435e-05, "loss": 3.493, "step": 386050 }, { "epoch": 3.4096328087744396, "grad_norm": 4.030482769012451, "learning_rate": 2.3061739113589903e-05, "loss": 3.2801, "step": 386100 }, { "epoch": 3.410074356664724, "grad_norm": 3.764228582382202, "learning_rate": 2.3050033234997702e-05, "loss": 3.2027, "step": 386150 }, { "epoch": 3.410515904555008, "grad_norm": 4.271496772766113, "learning_rate": 2.3038329438061905e-05, "loss": 2.7332, "step": 386200 }, { "epoch": 3.410957452445292, "grad_norm": 4.8442769050598145, "learning_rate": 2.302662772368651e-05, "loss": 3.0241, "step": 386250 }, { "epoch": 3.4113990003355763, "grad_norm": 6.548683166503906, "learning_rate": 2.3014928092775395e-05, "loss": 2.6103, "step": 386300 }, { "epoch": 3.4118405482258605, "grad_norm": 2.720925807952881, "learning_rate": 2.300323054623228e-05, "loss": 2.7754, "step": 386350 }, { "epoch": 3.4122820961161446, "grad_norm": 6.187340259552002, "learning_rate": 2.2991535084960658e-05, "loss": 3.0572, "step": 386400 }, { "epoch": 3.412723644006429, "grad_norm": 7.4890971183776855, "learning_rate": 2.2979841709863926e-05, "loss": 2.6855, "step": 386450 }, { "epoch": 3.413165191896713, "grad_norm": 5.75366735458374, "learning_rate": 2.296815042184532e-05, "loss": 2.8403, "step": 386500 }, { "epoch": 3.413606739786997, "grad_norm": 3.261328935623169, "learning_rate": 2.2956461221807862e-05, "loss": 2.9658, "step": 386550 }, { "epoch": 3.4140482876772813, "grad_norm": 7.028571605682373, "learning_rate": 2.2944774110654443e-05, "loss": 2.8275, "step": 386600 }, { "epoch": 3.4144898355675655, "grad_norm": 7.270656585693359, "learning_rate": 2.2933089089287795e-05, "loss": 2.8052, "step": 386650 }, { "epoch": 3.4149313834578496, "grad_norm": 5.974587440490723, "learning_rate": 2.2921406158610514e-05, "loss": 2.8951, "step": 386700 }, { "epoch": 3.4153729313481342, "grad_norm": 3.0312955379486084, "learning_rate": 2.2909725319524972e-05, "loss": 3.1749, "step": 386750 }, { "epoch": 3.415814479238418, "grad_norm": 3.6694483757019043, "learning_rate": 2.289804657293343e-05, "loss": 2.7051, "step": 386800 }, { "epoch": 3.4162560271287026, "grad_norm": 4.88215446472168, "learning_rate": 2.2886369919737987e-05, "loss": 2.7462, "step": 386850 }, { "epoch": 3.4166975750189867, "grad_norm": 4.864198684692383, "learning_rate": 2.2874695360840538e-05, "loss": 2.7805, "step": 386900 }, { "epoch": 3.417139122909271, "grad_norm": 3.4557738304138184, "learning_rate": 2.2863022897142872e-05, "loss": 2.694, "step": 386950 }, { "epoch": 3.417580670799555, "grad_norm": 2.165116310119629, "learning_rate": 2.2851352529546572e-05, "loss": 2.5984, "step": 387000 }, { "epoch": 3.417580670799555, "eval_asr_loss": 0.9554669872136783, "eval_loss": 2.731400728225708, "eval_runtime": 20.4331, "eval_samples_per_second": 37.586, "eval_steps_per_second": 9.397, "eval_tts_loss": 6.152589725140025, "step": 387000 }, { "epoch": 3.4180222186898392, "grad_norm": 3.7762253284454346, "learning_rate": 2.283968425895306e-05, "loss": 3.2798, "step": 387050 }, { "epoch": 3.4184637665801234, "grad_norm": 5.78275728225708, "learning_rate": 2.282801808626362e-05, "loss": 2.9482, "step": 387100 }, { "epoch": 3.4189053144704076, "grad_norm": 2.0023462772369385, "learning_rate": 2.281635401237939e-05, "loss": 2.6572, "step": 387150 }, { "epoch": 3.4193468623606917, "grad_norm": 5.1205596923828125, "learning_rate": 2.280469203820129e-05, "loss": 2.3929, "step": 387200 }, { "epoch": 3.419788410250976, "grad_norm": 6.5242767333984375, "learning_rate": 2.2793032164630117e-05, "loss": 3.1688, "step": 387250 }, { "epoch": 3.42022995814126, "grad_norm": 4.846236228942871, "learning_rate": 2.2781374392566524e-05, "loss": 2.7137, "step": 387300 }, { "epoch": 3.4206715060315442, "grad_norm": 2.055490493774414, "learning_rate": 2.2769718722910947e-05, "loss": 2.9672, "step": 387350 }, { "epoch": 3.4211130539218284, "grad_norm": 4.975621700286865, "learning_rate": 2.2758065156563675e-05, "loss": 3.4051, "step": 387400 }, { "epoch": 3.4215546018121126, "grad_norm": 5.556480884552002, "learning_rate": 2.2746413694424863e-05, "loss": 2.8757, "step": 387450 }, { "epoch": 3.4219961497023967, "grad_norm": 4.094520568847656, "learning_rate": 2.273476433739451e-05, "loss": 2.9732, "step": 387500 }, { "epoch": 3.422437697592681, "grad_norm": 3.03737211227417, "learning_rate": 2.2723117086372393e-05, "loss": 2.6156, "step": 387550 }, { "epoch": 3.422879245482965, "grad_norm": 2.3388192653656006, "learning_rate": 2.271147194225818e-05, "loss": 2.8986, "step": 387600 }, { "epoch": 3.4233207933732492, "grad_norm": 6.044736862182617, "learning_rate": 2.269982890595137e-05, "loss": 3.1205, "step": 387650 }, { "epoch": 3.4237623412635334, "grad_norm": 4.236540794372559, "learning_rate": 2.2688187978351266e-05, "loss": 2.9479, "step": 387700 }, { "epoch": 3.4242038891538176, "grad_norm": 1.2491379976272583, "learning_rate": 2.2676549160357065e-05, "loss": 2.6077, "step": 387750 }, { "epoch": 3.4246454370441017, "grad_norm": 3.275805950164795, "learning_rate": 2.266491245286772e-05, "loss": 3.3656, "step": 387800 }, { "epoch": 3.425086984934386, "grad_norm": 3.0143089294433594, "learning_rate": 2.2653277856782113e-05, "loss": 3.1475, "step": 387850 }, { "epoch": 3.42552853282467, "grad_norm": 6.421952724456787, "learning_rate": 2.2641645372998883e-05, "loss": 2.9024, "step": 387900 }, { "epoch": 3.4259700807149542, "grad_norm": 5.071564674377441, "learning_rate": 2.2630247589113025e-05, "loss": 2.6452, "step": 387950 }, { "epoch": 3.4264116286052384, "grad_norm": 4.554489612579346, "learning_rate": 2.2618619290339177e-05, "loss": 3.0582, "step": 388000 }, { "epoch": 3.4268531764955226, "grad_norm": 4.810396671295166, "learning_rate": 2.2606993106544798e-05, "loss": 2.5279, "step": 388050 }, { "epoch": 3.4272947243858067, "grad_norm": 6.412106990814209, "learning_rate": 2.2595369038627907e-05, "loss": 2.9278, "step": 388100 }, { "epoch": 3.427736272276091, "grad_norm": 3.616091251373291, "learning_rate": 2.258374708748636e-05, "loss": 2.6497, "step": 388150 }, { "epoch": 3.428177820166375, "grad_norm": 4.19853401184082, "learning_rate": 2.2572127254017894e-05, "loss": 2.7332, "step": 388200 }, { "epoch": 3.4286193680566592, "grad_norm": 3.6462626457214355, "learning_rate": 2.256050953912e-05, "loss": 3.1517, "step": 388250 }, { "epoch": 3.429060915946944, "grad_norm": 4.029109954833984, "learning_rate": 2.2548893943690057e-05, "loss": 3.2008, "step": 388300 }, { "epoch": 3.4295024638372276, "grad_norm": 4.272415637969971, "learning_rate": 2.253728046862528e-05, "loss": 2.945, "step": 388350 }, { "epoch": 3.429944011727512, "grad_norm": 1.996616005897522, "learning_rate": 2.2525669114822735e-05, "loss": 2.7174, "step": 388400 }, { "epoch": 3.430385559617796, "grad_norm": 3.7101211547851562, "learning_rate": 2.251405988317927e-05, "loss": 2.7619, "step": 388450 }, { "epoch": 3.4308271075080805, "grad_norm": 3.299766778945923, "learning_rate": 2.2502452774591588e-05, "loss": 2.5417, "step": 388500 }, { "epoch": 3.4312686553983647, "grad_norm": 0.7740725874900818, "learning_rate": 2.249084778995627e-05, "loss": 3.0964, "step": 388550 }, { "epoch": 3.431710203288649, "grad_norm": 2.7573235034942627, "learning_rate": 2.247924493016968e-05, "loss": 2.52, "step": 388600 }, { "epoch": 3.432151751178933, "grad_norm": 6.5384297370910645, "learning_rate": 2.2467644196128045e-05, "loss": 2.685, "step": 388650 }, { "epoch": 3.432593299069217, "grad_norm": 4.840304851531982, "learning_rate": 2.2456045588727448e-05, "loss": 3.3897, "step": 388700 }, { "epoch": 3.4330348469595013, "grad_norm": 4.516257286071777, "learning_rate": 2.244444910886374e-05, "loss": 2.7751, "step": 388750 }, { "epoch": 3.4334763948497855, "grad_norm": 5.792184829711914, "learning_rate": 2.24328547574327e-05, "loss": 2.9868, "step": 388800 }, { "epoch": 3.4339179427400697, "grad_norm": 3.250208854675293, "learning_rate": 2.2421262535329845e-05, "loss": 3.044, "step": 388850 }, { "epoch": 3.434359490630354, "grad_norm": 2.5794456005096436, "learning_rate": 2.2409672443450615e-05, "loss": 3.0375, "step": 388900 }, { "epoch": 3.434801038520638, "grad_norm": 3.219341993331909, "learning_rate": 2.2398084482690213e-05, "loss": 2.7711, "step": 388950 }, { "epoch": 3.435242586410922, "grad_norm": 5.02898645401001, "learning_rate": 2.2386498653943728e-05, "loss": 3.3143, "step": 389000 }, { "epoch": 3.4356841343012063, "grad_norm": 2.901411294937134, "learning_rate": 2.2374914958106085e-05, "loss": 2.7887, "step": 389050 }, { "epoch": 3.4361256821914905, "grad_norm": 5.268230438232422, "learning_rate": 2.236333339607199e-05, "loss": 2.4873, "step": 389100 }, { "epoch": 3.4365672300817747, "grad_norm": 2.554741382598877, "learning_rate": 2.2351753968736033e-05, "loss": 2.7934, "step": 389150 }, { "epoch": 3.437008777972059, "grad_norm": 5.489684581756592, "learning_rate": 2.234017667699266e-05, "loss": 2.7063, "step": 389200 }, { "epoch": 3.437450325862343, "grad_norm": 4.120151042938232, "learning_rate": 2.232860152173608e-05, "loss": 2.7342, "step": 389250 }, { "epoch": 3.437891873752627, "grad_norm": 5.804008483886719, "learning_rate": 2.231702850386037e-05, "loss": 2.9474, "step": 389300 }, { "epoch": 3.4383334216429113, "grad_norm": 3.031390428543091, "learning_rate": 2.2305457624259485e-05, "loss": 2.6918, "step": 389350 }, { "epoch": 3.4387749695331955, "grad_norm": 7.111850738525391, "learning_rate": 2.2293888883827134e-05, "loss": 2.8928, "step": 389400 }, { "epoch": 3.4392165174234797, "grad_norm": 9.016135215759277, "learning_rate": 2.228232228345693e-05, "loss": 2.7251, "step": 389450 }, { "epoch": 3.439658065313764, "grad_norm": 1.7321608066558838, "learning_rate": 2.2270757824042287e-05, "loss": 3.0974, "step": 389500 }, { "epoch": 3.440099613204048, "grad_norm": 3.3729302883148193, "learning_rate": 2.225919550647648e-05, "loss": 2.8873, "step": 389550 }, { "epoch": 3.440541161094332, "grad_norm": 5.246362209320068, "learning_rate": 2.2247635331652595e-05, "loss": 2.9686, "step": 389600 }, { "epoch": 3.4409827089846163, "grad_norm": 4.750444412231445, "learning_rate": 2.223607730046352e-05, "loss": 3.1726, "step": 389650 }, { "epoch": 3.4414242568749005, "grad_norm": 3.9472432136535645, "learning_rate": 2.2224521413802064e-05, "loss": 3.0644, "step": 389700 }, { "epoch": 3.4418658047651847, "grad_norm": 6.300321578979492, "learning_rate": 2.221296767256078e-05, "loss": 2.4113, "step": 389750 }, { "epoch": 3.442307352655469, "grad_norm": 6.241701602935791, "learning_rate": 2.220141607763212e-05, "loss": 2.5908, "step": 389800 }, { "epoch": 3.442748900545753, "grad_norm": 3.574805974960327, "learning_rate": 2.2189866629908356e-05, "loss": 2.7564, "step": 389850 }, { "epoch": 3.443190448436037, "grad_norm": 4.056480884552002, "learning_rate": 2.2178319330281554e-05, "loss": 3.2502, "step": 389900 }, { "epoch": 3.4436319963263218, "grad_norm": 8.070205688476562, "learning_rate": 2.2166774179643685e-05, "loss": 2.9144, "step": 389950 }, { "epoch": 3.4440735442166055, "grad_norm": 4.756385803222656, "learning_rate": 2.215523117888647e-05, "loss": 2.8618, "step": 390000 }, { "epoch": 3.4440735442166055, "eval_asr_loss": 0.9481846807479942, "eval_loss": 2.7294692993164062, "eval_runtime": 20.7183, "eval_samples_per_second": 37.069, "eval_steps_per_second": 9.267, "eval_tts_loss": 6.164907876528534, "step": 390000 }, { "epoch": 3.44451509210689, "grad_norm": 2.326047420501709, "learning_rate": 2.214369032890155e-05, "loss": 3.1786, "step": 390050 }, { "epoch": 3.4449566399971743, "grad_norm": 5.755054950714111, "learning_rate": 2.213238238345468e-05, "loss": 2.6397, "step": 390100 }, { "epoch": 3.4453981878874584, "grad_norm": 2.082336664199829, "learning_rate": 2.212084579462859e-05, "loss": 2.3656, "step": 390150 }, { "epoch": 3.4458397357777426, "grad_norm": 4.6807379722595215, "learning_rate": 2.2109311359230754e-05, "loss": 2.5122, "step": 390200 }, { "epoch": 3.4462812836680268, "grad_norm": 6.366551876068115, "learning_rate": 2.2097779078152125e-05, "loss": 2.9004, "step": 390250 }, { "epoch": 3.446722831558311, "grad_norm": 3.6888973712921143, "learning_rate": 2.2086248952283462e-05, "loss": 2.7529, "step": 390300 }, { "epoch": 3.447164379448595, "grad_norm": 7.085838317871094, "learning_rate": 2.2074720982515344e-05, "loss": 3.2611, "step": 390350 }, { "epoch": 3.4476059273388793, "grad_norm": 4.1366353034973145, "learning_rate": 2.2063195169738256e-05, "loss": 3.1575, "step": 390400 }, { "epoch": 3.4480474752291634, "grad_norm": 1.7637420892715454, "learning_rate": 2.205167151484242e-05, "loss": 3.0721, "step": 390450 }, { "epoch": 3.4484890231194476, "grad_norm": 4.153936862945557, "learning_rate": 2.2040150018717964e-05, "loss": 2.9263, "step": 390500 }, { "epoch": 3.4489305710097318, "grad_norm": 6.923408508300781, "learning_rate": 2.202863068225483e-05, "loss": 2.7056, "step": 390550 }, { "epoch": 3.449372118900016, "grad_norm": 1.5307427644729614, "learning_rate": 2.2017113506342803e-05, "loss": 2.6838, "step": 390600 }, { "epoch": 3.4498136667903, "grad_norm": 4.824482440948486, "learning_rate": 2.200559849187145e-05, "loss": 2.8666, "step": 390650 }, { "epoch": 3.4502552146805843, "grad_norm": 7.896531581878662, "learning_rate": 2.1994085639730245e-05, "loss": 2.8084, "step": 390700 }, { "epoch": 3.4506967625708684, "grad_norm": 1.4450137615203857, "learning_rate": 2.1982574950808443e-05, "loss": 2.7145, "step": 390750 }, { "epoch": 3.4511383104611526, "grad_norm": 5.995443344116211, "learning_rate": 2.1971066425995123e-05, "loss": 3.4496, "step": 390800 }, { "epoch": 3.4515798583514368, "grad_norm": 5.262402534484863, "learning_rate": 2.1959560066179253e-05, "loss": 3.0367, "step": 390850 }, { "epoch": 3.452021406241721, "grad_norm": 5.328161716461182, "learning_rate": 2.19480558722496e-05, "loss": 2.6761, "step": 390900 }, { "epoch": 3.452462954132005, "grad_norm": 3.9588615894317627, "learning_rate": 2.1936553845094744e-05, "loss": 3.138, "step": 390950 }, { "epoch": 3.4529045020222893, "grad_norm": 3.8037986755371094, "learning_rate": 2.192505398560313e-05, "loss": 3.042, "step": 391000 }, { "epoch": 3.4533460499125734, "grad_norm": 6.422025680541992, "learning_rate": 2.1913556294663046e-05, "loss": 2.8764, "step": 391050 }, { "epoch": 3.4537875978028576, "grad_norm": 4.475071907043457, "learning_rate": 2.1902060773162576e-05, "loss": 2.6366, "step": 391100 }, { "epoch": 3.4542291456931418, "grad_norm": 5.273529052734375, "learning_rate": 2.1890567421989622e-05, "loss": 3.2453, "step": 391150 }, { "epoch": 3.454670693583426, "grad_norm": 4.296705722808838, "learning_rate": 2.187907624203198e-05, "loss": 2.58, "step": 391200 }, { "epoch": 3.45511224147371, "grad_norm": 3.2494707107543945, "learning_rate": 2.1867587234177256e-05, "loss": 3.0054, "step": 391250 }, { "epoch": 3.4555537893639943, "grad_norm": 5.789858341217041, "learning_rate": 2.1856100399312845e-05, "loss": 2.5829, "step": 391300 }, { "epoch": 3.4559953372542784, "grad_norm": 6.089230060577393, "learning_rate": 2.1844615738326025e-05, "loss": 3.0723, "step": 391350 }, { "epoch": 3.4564368851445626, "grad_norm": 3.064711332321167, "learning_rate": 2.1833133252103903e-05, "loss": 3.0632, "step": 391400 }, { "epoch": 3.4568784330348468, "grad_norm": 3.294795036315918, "learning_rate": 2.1821652941533395e-05, "loss": 2.8562, "step": 391450 }, { "epoch": 3.4573199809251314, "grad_norm": 4.335747241973877, "learning_rate": 2.1810174807501227e-05, "loss": 2.5828, "step": 391500 }, { "epoch": 3.457761528815415, "grad_norm": 2.327298879623413, "learning_rate": 2.1798698850894035e-05, "loss": 2.7906, "step": 391550 }, { "epoch": 3.4582030767056997, "grad_norm": 7.629519939422607, "learning_rate": 2.17872250725982e-05, "loss": 2.5882, "step": 391600 }, { "epoch": 3.458644624595984, "grad_norm": 5.078785419464111, "learning_rate": 2.1775753473499987e-05, "loss": 2.8787, "step": 391650 }, { "epoch": 3.459086172486268, "grad_norm": 2.038907527923584, "learning_rate": 2.1764284054485483e-05, "loss": 2.9916, "step": 391700 }, { "epoch": 3.459527720376552, "grad_norm": 2.461164951324463, "learning_rate": 2.1752816816440624e-05, "loss": 2.8314, "step": 391750 }, { "epoch": 3.4599692682668364, "grad_norm": 6.523917198181152, "learning_rate": 2.1741351760251116e-05, "loss": 2.7443, "step": 391800 }, { "epoch": 3.4604108161571205, "grad_norm": 1.985107421875, "learning_rate": 2.1729888886802578e-05, "loss": 2.5475, "step": 391850 }, { "epoch": 3.4608523640474047, "grad_norm": 1.639182209968567, "learning_rate": 2.1718428196980407e-05, "loss": 3.2145, "step": 391900 }, { "epoch": 3.461293911937689, "grad_norm": 4.15117073059082, "learning_rate": 2.170696969166981e-05, "loss": 2.9446, "step": 391950 }, { "epoch": 3.461735459827973, "grad_norm": 5.871176242828369, "learning_rate": 2.1695513371755888e-05, "loss": 2.9438, "step": 392000 }, { "epoch": 3.462177007718257, "grad_norm": 1.9389625787734985, "learning_rate": 2.1684059238123566e-05, "loss": 2.9073, "step": 392050 }, { "epoch": 3.4626185556085414, "grad_norm": 2.9898810386657715, "learning_rate": 2.1672607291657532e-05, "loss": 3.1893, "step": 392100 }, { "epoch": 3.4630601034988255, "grad_norm": 5.789550304412842, "learning_rate": 2.1661157533242376e-05, "loss": 2.8963, "step": 392150 }, { "epoch": 3.4635016513891097, "grad_norm": 1.9284626245498657, "learning_rate": 2.1649709963762515e-05, "loss": 2.899, "step": 392200 }, { "epoch": 3.463943199279394, "grad_norm": 7.349296569824219, "learning_rate": 2.1638264584102158e-05, "loss": 2.9487, "step": 392250 }, { "epoch": 3.464384747169678, "grad_norm": 5.008468151092529, "learning_rate": 2.1626821395145337e-05, "loss": 3.4366, "step": 392300 }, { "epoch": 3.464826295059962, "grad_norm": 2.749152421951294, "learning_rate": 2.161538039777597e-05, "loss": 2.575, "step": 392350 }, { "epoch": 3.4652678429502464, "grad_norm": 2.58965802192688, "learning_rate": 2.1603941592877785e-05, "loss": 2.1784, "step": 392400 }, { "epoch": 3.4657093908405305, "grad_norm": 6.715756416320801, "learning_rate": 2.1592504981334312e-05, "loss": 2.7651, "step": 392450 }, { "epoch": 3.4661509387308147, "grad_norm": 4.902455806732178, "learning_rate": 2.1581070564028933e-05, "loss": 2.9691, "step": 392500 }, { "epoch": 3.466592486621099, "grad_norm": 4.0624237060546875, "learning_rate": 2.1569638341844888e-05, "loss": 2.9874, "step": 392550 }, { "epoch": 3.467034034511383, "grad_norm": 5.9944539070129395, "learning_rate": 2.155820831566518e-05, "loss": 3.3298, "step": 392600 }, { "epoch": 3.467475582401667, "grad_norm": 3.1605637073516846, "learning_rate": 2.154678048637272e-05, "loss": 2.8656, "step": 392650 }, { "epoch": 3.4679171302919514, "grad_norm": 3.0223162174224854, "learning_rate": 2.153535485485019e-05, "loss": 2.7661, "step": 392700 }, { "epoch": 3.4683586781822355, "grad_norm": 5.45560884475708, "learning_rate": 2.1523931421980102e-05, "loss": 3.0528, "step": 392750 }, { "epoch": 3.4688002260725197, "grad_norm": 5.370608329772949, "learning_rate": 2.151251018864485e-05, "loss": 3.0645, "step": 392800 }, { "epoch": 3.469241773962804, "grad_norm": 4.594405174255371, "learning_rate": 2.1501091155726616e-05, "loss": 2.8444, "step": 392850 }, { "epoch": 3.469683321853088, "grad_norm": 3.4727163314819336, "learning_rate": 2.1489674324107444e-05, "loss": 2.9997, "step": 392900 }, { "epoch": 3.470124869743372, "grad_norm": 2.080793857574463, "learning_rate": 2.147825969466915e-05, "loss": 2.527, "step": 392950 }, { "epoch": 3.4705664176336564, "grad_norm": 3.9015371799468994, "learning_rate": 2.1466847268293455e-05, "loss": 3.0677, "step": 393000 }, { "epoch": 3.4705664176336564, "eval_asr_loss": 0.9551264332583905, "eval_loss": 2.7260611057281494, "eval_runtime": 20.4107, "eval_samples_per_second": 37.627, "eval_steps_per_second": 9.407, "eval_tts_loss": 6.156782192478794, "step": 393000 }, { "epoch": 3.471007965523941, "grad_norm": 5.16027307510376, "learning_rate": 2.145543704586186e-05, "loss": 3.0184, "step": 393050 }, { "epoch": 3.4714495134142247, "grad_norm": 5.634804725646973, "learning_rate": 2.144402902825568e-05, "loss": 3.0647, "step": 393100 }, { "epoch": 3.4718910613045093, "grad_norm": 4.022769927978516, "learning_rate": 2.1432623216356106e-05, "loss": 2.2705, "step": 393150 }, { "epoch": 3.472332609194793, "grad_norm": 4.595845699310303, "learning_rate": 2.142121961104417e-05, "loss": 2.828, "step": 393200 }, { "epoch": 3.4727741570850776, "grad_norm": 1.468678593635559, "learning_rate": 2.1409818213200655e-05, "loss": 2.9143, "step": 393250 }, { "epoch": 3.473215704975362, "grad_norm": 5.484999179840088, "learning_rate": 2.139841902370625e-05, "loss": 3.1828, "step": 393300 }, { "epoch": 3.473657252865646, "grad_norm": 3.6277952194213867, "learning_rate": 2.1387249961390616e-05, "loss": 3.0143, "step": 393350 }, { "epoch": 3.47409880075593, "grad_norm": 2.3602662086486816, "learning_rate": 2.1375855147024892e-05, "loss": 2.7814, "step": 393400 }, { "epoch": 3.4745403486462143, "grad_norm": 2.2506656646728516, "learning_rate": 2.1364462543631652e-05, "loss": 3.0467, "step": 393450 }, { "epoch": 3.4749818965364985, "grad_norm": 5.677712440490723, "learning_rate": 2.1353072152090848e-05, "loss": 3.2691, "step": 393500 }, { "epoch": 3.4754234444267826, "grad_norm": 3.631333589553833, "learning_rate": 2.1341683973282296e-05, "loss": 3.2408, "step": 393550 }, { "epoch": 3.475864992317067, "grad_norm": 5.508145332336426, "learning_rate": 2.1330298008085657e-05, "loss": 3.206, "step": 393600 }, { "epoch": 3.476306540207351, "grad_norm": 2.833730697631836, "learning_rate": 2.1318914257380368e-05, "loss": 3.5741, "step": 393650 }, { "epoch": 3.476748088097635, "grad_norm": 2.6205945014953613, "learning_rate": 2.1307532722045765e-05, "loss": 2.8923, "step": 393700 }, { "epoch": 3.4771896359879193, "grad_norm": 1.900678038597107, "learning_rate": 2.1296153402960955e-05, "loss": 2.5931, "step": 393750 }, { "epoch": 3.4776311838782035, "grad_norm": 2.363295555114746, "learning_rate": 2.128477630100487e-05, "loss": 3.0955, "step": 393800 }, { "epoch": 3.4780727317684876, "grad_norm": 4.44240665435791, "learning_rate": 2.127340141705632e-05, "loss": 2.2077, "step": 393850 }, { "epoch": 3.478514279658772, "grad_norm": 2.440660238265991, "learning_rate": 2.1262028751993923e-05, "loss": 3.198, "step": 393900 }, { "epoch": 3.478955827549056, "grad_norm": 3.1083250045776367, "learning_rate": 2.1250658306696124e-05, "loss": 2.9328, "step": 393950 }, { "epoch": 3.47939737543934, "grad_norm": 3.4541244506835938, "learning_rate": 2.123929008204117e-05, "loss": 2.661, "step": 394000 }, { "epoch": 3.4798389233296243, "grad_norm": 3.925713300704956, "learning_rate": 2.122792407890717e-05, "loss": 2.8969, "step": 394050 }, { "epoch": 3.4802804712199085, "grad_norm": 7.183180809020996, "learning_rate": 2.1216560298172074e-05, "loss": 3.0572, "step": 394100 }, { "epoch": 3.4807220191101926, "grad_norm": 1.7217490673065186, "learning_rate": 2.1205198740713615e-05, "loss": 2.6253, "step": 394150 }, { "epoch": 3.481163567000477, "grad_norm": 6.073792457580566, "learning_rate": 2.1193839407409367e-05, "loss": 2.6085, "step": 394200 }, { "epoch": 3.481605114890761, "grad_norm": 4.128481864929199, "learning_rate": 2.1182482299136773e-05, "loss": 2.7639, "step": 394250 }, { "epoch": 3.482046662781045, "grad_norm": 5.951517581939697, "learning_rate": 2.117112741677304e-05, "loss": 3.0535, "step": 394300 }, { "epoch": 3.4824882106713293, "grad_norm": 4.531003952026367, "learning_rate": 2.115977476119525e-05, "loss": 3.0429, "step": 394350 }, { "epoch": 3.4829297585616135, "grad_norm": 7.029266357421875, "learning_rate": 2.1148424333280297e-05, "loss": 3.1817, "step": 394400 }, { "epoch": 3.4833713064518976, "grad_norm": 3.5658795833587646, "learning_rate": 2.1137076133904937e-05, "loss": 2.9482, "step": 394450 }, { "epoch": 3.483812854342182, "grad_norm": 3.8127970695495605, "learning_rate": 2.112573016394569e-05, "loss": 3.0704, "step": 394500 }, { "epoch": 3.484254402232466, "grad_norm": 4.895659446716309, "learning_rate": 2.111438642427892e-05, "loss": 3.0507, "step": 394550 }, { "epoch": 3.48469595012275, "grad_norm": 4.8975701332092285, "learning_rate": 2.1103044915780877e-05, "loss": 2.7404, "step": 394600 }, { "epoch": 3.4851374980130343, "grad_norm": 5.865734100341797, "learning_rate": 2.1091705639327564e-05, "loss": 2.9117, "step": 394650 }, { "epoch": 3.485579045903319, "grad_norm": 5.258873462677002, "learning_rate": 2.108036859579485e-05, "loss": 3.1107, "step": 394700 }, { "epoch": 3.4860205937936026, "grad_norm": 3.9243111610412598, "learning_rate": 2.1069033786058445e-05, "loss": 2.3776, "step": 394750 }, { "epoch": 3.4864621416838872, "grad_norm": 4.636477470397949, "learning_rate": 2.1057701210993836e-05, "loss": 2.5435, "step": 394800 }, { "epoch": 3.4869036895741714, "grad_norm": 4.4444169998168945, "learning_rate": 2.104637087147639e-05, "loss": 3.1552, "step": 394850 }, { "epoch": 3.4873452374644556, "grad_norm": 7.953307151794434, "learning_rate": 2.1035042768381285e-05, "loss": 2.8309, "step": 394900 }, { "epoch": 3.4877867853547397, "grad_norm": 9.563583374023438, "learning_rate": 2.1023716902583516e-05, "loss": 2.7462, "step": 394950 }, { "epoch": 3.488228333245024, "grad_norm": 6.754940986633301, "learning_rate": 2.1012393274957886e-05, "loss": 2.6618, "step": 395000 }, { "epoch": 3.488669881135308, "grad_norm": 6.22545862197876, "learning_rate": 2.100107188637907e-05, "loss": 2.7706, "step": 395050 }, { "epoch": 3.4891114290255922, "grad_norm": 3.61332368850708, "learning_rate": 2.098975273772156e-05, "loss": 3.0209, "step": 395100 }, { "epoch": 3.4895529769158764, "grad_norm": 6.103861331939697, "learning_rate": 2.0978435829859633e-05, "loss": 2.7637, "step": 395150 }, { "epoch": 3.4899945248061606, "grad_norm": 4.693385601043701, "learning_rate": 2.096712116366745e-05, "loss": 2.9764, "step": 395200 }, { "epoch": 3.4904360726964447, "grad_norm": 3.907224178314209, "learning_rate": 2.0955808740018973e-05, "loss": 2.6283, "step": 395250 }, { "epoch": 3.490877620586729, "grad_norm": 2.95035719871521, "learning_rate": 2.094449855978799e-05, "loss": 2.558, "step": 395300 }, { "epoch": 3.491319168477013, "grad_norm": 4.7167792320251465, "learning_rate": 2.093319062384809e-05, "loss": 2.9426, "step": 395350 }, { "epoch": 3.4917607163672972, "grad_norm": 4.4092559814453125, "learning_rate": 2.0921884933072755e-05, "loss": 3.3167, "step": 395400 }, { "epoch": 3.4922022642575814, "grad_norm": 2.4238440990448, "learning_rate": 2.0910581488335213e-05, "loss": 2.7898, "step": 395450 }, { "epoch": 3.4926438121478656, "grad_norm": 5.102490425109863, "learning_rate": 2.0899280290508584e-05, "loss": 3.4401, "step": 395500 }, { "epoch": 3.4930853600381497, "grad_norm": 3.260803699493408, "learning_rate": 2.088798134046578e-05, "loss": 3.1098, "step": 395550 }, { "epoch": 3.493526907928434, "grad_norm": 6.186168193817139, "learning_rate": 2.0876684639079575e-05, "loss": 3.1234, "step": 395600 }, { "epoch": 3.493968455818718, "grad_norm": 6.206269264221191, "learning_rate": 2.0865390187222523e-05, "loss": 2.8767, "step": 395650 }, { "epoch": 3.4944100037090022, "grad_norm": 3.591066598892212, "learning_rate": 2.0854097985767006e-05, "loss": 3.0774, "step": 395700 }, { "epoch": 3.4948515515992864, "grad_norm": 3.7322211265563965, "learning_rate": 2.0842808035585292e-05, "loss": 2.8712, "step": 395750 }, { "epoch": 3.4952930994895706, "grad_norm": 3.7306532859802246, "learning_rate": 2.083152033754939e-05, "loss": 2.9705, "step": 395800 }, { "epoch": 3.4957346473798547, "grad_norm": 6.569372653961182, "learning_rate": 2.0820234892531197e-05, "loss": 3.0785, "step": 395850 }, { "epoch": 3.496176195270139, "grad_norm": 4.678150653839111, "learning_rate": 2.0808951701402447e-05, "loss": 3.134, "step": 395900 }, { "epoch": 3.496617743160423, "grad_norm": 4.1780571937561035, "learning_rate": 2.0797670765034622e-05, "loss": 2.81, "step": 395950 }, { "epoch": 3.4970592910507072, "grad_norm": 2.477694034576416, "learning_rate": 2.0786392084299107e-05, "loss": 2.4527, "step": 396000 }, { "epoch": 3.4970592910507072, "eval_asr_loss": 0.9505076267852494, "eval_loss": 2.7285289764404297, "eval_runtime": 20.8607, "eval_samples_per_second": 36.816, "eval_steps_per_second": 9.204, "eval_tts_loss": 6.1536176936545735, "step": 396000 }, { "epoch": 3.4975008389409914, "grad_norm": 6.249030590057373, "learning_rate": 2.0775115660067092e-05, "loss": 2.7789, "step": 396050 }, { "epoch": 3.4979423868312756, "grad_norm": 2.699784517288208, "learning_rate": 2.0763841493209574e-05, "loss": 2.7679, "step": 396100 }, { "epoch": 3.4983839347215597, "grad_norm": 3.7190229892730713, "learning_rate": 2.0752569584597363e-05, "loss": 2.8859, "step": 396150 }, { "epoch": 3.498825482611844, "grad_norm": 8.764069557189941, "learning_rate": 2.074129993510115e-05, "loss": 2.438, "step": 396200 }, { "epoch": 3.4992670305021285, "grad_norm": 3.982754707336426, "learning_rate": 2.073003254559142e-05, "loss": 2.6922, "step": 396250 }, { "epoch": 3.4997085783924122, "grad_norm": 5.7203779220581055, "learning_rate": 2.0718767416938462e-05, "loss": 2.8848, "step": 396300 }, { "epoch": 3.500150126282697, "grad_norm": 2.674650192260742, "learning_rate": 2.070750455001242e-05, "loss": 2.6304, "step": 396350 }, { "epoch": 3.5005916741729806, "grad_norm": 2.9136698246002197, "learning_rate": 2.0696243945683274e-05, "loss": 3.1587, "step": 396400 }, { "epoch": 3.501033222063265, "grad_norm": 2.432753801345825, "learning_rate": 2.06849856048208e-05, "loss": 2.7004, "step": 396450 }, { "epoch": 3.501474769953549, "grad_norm": 6.0153489112854, "learning_rate": 2.0673729528294582e-05, "loss": 2.5095, "step": 396500 }, { "epoch": 3.5019163178438335, "grad_norm": 2.37607479095459, "learning_rate": 2.06624757169741e-05, "loss": 2.6641, "step": 396550 }, { "epoch": 3.5023578657341177, "grad_norm": 6.305436134338379, "learning_rate": 2.0651224171728573e-05, "loss": 2.8734, "step": 396600 }, { "epoch": 3.502799413624402, "grad_norm": 6.932346820831299, "learning_rate": 2.063997489342711e-05, "loss": 2.8091, "step": 396650 }, { "epoch": 3.503240961514686, "grad_norm": 4.958677768707275, "learning_rate": 2.0628727882938635e-05, "loss": 3.0935, "step": 396700 }, { "epoch": 3.50368250940497, "grad_norm": 2.1652393341064453, "learning_rate": 2.061748314113186e-05, "loss": 3.0569, "step": 396750 }, { "epoch": 3.5041240572952543, "grad_norm": 2.863849401473999, "learning_rate": 2.0606240668875347e-05, "loss": 2.6958, "step": 396800 }, { "epoch": 3.5045656051855385, "grad_norm": 8.749558448791504, "learning_rate": 2.0595000467037516e-05, "loss": 2.8767, "step": 396850 }, { "epoch": 3.5050071530758227, "grad_norm": 2.0101311206817627, "learning_rate": 2.058376253648655e-05, "loss": 2.6403, "step": 396900 }, { "epoch": 3.505448700966107, "grad_norm": 2.032247543334961, "learning_rate": 2.0572526878090465e-05, "loss": 2.7056, "step": 396950 }, { "epoch": 3.505890248856391, "grad_norm": 2.391533613204956, "learning_rate": 2.0561293492717144e-05, "loss": 2.7759, "step": 397000 }, { "epoch": 3.506331796746675, "grad_norm": 3.453068971633911, "learning_rate": 2.055006238123428e-05, "loss": 2.6722, "step": 397050 }, { "epoch": 3.5067733446369593, "grad_norm": 2.865841865539551, "learning_rate": 2.0538833544509358e-05, "loss": 3.1342, "step": 397100 }, { "epoch": 3.5072148925272435, "grad_norm": 5.82879638671875, "learning_rate": 2.0527606983409713e-05, "loss": 3.2596, "step": 397150 }, { "epoch": 3.5076564404175277, "grad_norm": 1.747082233428955, "learning_rate": 2.0516382698802534e-05, "loss": 2.6647, "step": 397200 }, { "epoch": 3.508097988307812, "grad_norm": 2.441760778427124, "learning_rate": 2.0505160691554777e-05, "loss": 3.1359, "step": 397250 }, { "epoch": 3.508539536198096, "grad_norm": 2.793318271636963, "learning_rate": 2.049394096253323e-05, "loss": 2.964, "step": 397300 }, { "epoch": 3.50898108408838, "grad_norm": 3.3151209354400635, "learning_rate": 2.048272351260454e-05, "loss": 2.6227, "step": 397350 }, { "epoch": 3.5094226319786643, "grad_norm": 5.0169572830200195, "learning_rate": 2.0471508342635175e-05, "loss": 3.2885, "step": 397400 }, { "epoch": 3.5098641798689485, "grad_norm": 2.4192757606506348, "learning_rate": 2.0460295453491378e-05, "loss": 2.7647, "step": 397450 }, { "epoch": 3.5103057277592327, "grad_norm": 3.974233388900757, "learning_rate": 2.044908484603927e-05, "loss": 2.8107, "step": 397500 }, { "epoch": 3.510747275649517, "grad_norm": 3.481088161468506, "learning_rate": 2.0437876521144787e-05, "loss": 3.1913, "step": 397550 }, { "epoch": 3.511188823539801, "grad_norm": 3.799309492111206, "learning_rate": 2.042667047967367e-05, "loss": 2.897, "step": 397600 }, { "epoch": 3.511630371430085, "grad_norm": 6.830905914306641, "learning_rate": 2.0415690775243484e-05, "loss": 3.3461, "step": 397650 }, { "epoch": 3.5120719193203693, "grad_norm": 2.0277819633483887, "learning_rate": 2.0404489257504034e-05, "loss": 2.3432, "step": 397700 }, { "epoch": 3.5125134672106535, "grad_norm": 4.718966007232666, "learning_rate": 2.039329002576683e-05, "loss": 2.9165, "step": 397750 }, { "epoch": 3.512955015100938, "grad_norm": 3.6047513484954834, "learning_rate": 2.038209308089693e-05, "loss": 2.8157, "step": 397800 }, { "epoch": 3.513396562991222, "grad_norm": 1.6143367290496826, "learning_rate": 2.0370898423759172e-05, "loss": 2.651, "step": 397850 }, { "epoch": 3.5138381108815064, "grad_norm": 5.864950656890869, "learning_rate": 2.0359706055218282e-05, "loss": 3.0265, "step": 397900 }, { "epoch": 3.51427965877179, "grad_norm": 5.122350692749023, "learning_rate": 2.034851597613875e-05, "loss": 3.0524, "step": 397950 }, { "epoch": 3.5147212066620748, "grad_norm": 3.8992159366607666, "learning_rate": 2.0337328187384907e-05, "loss": 2.9881, "step": 398000 }, { "epoch": 3.5151627545523585, "grad_norm": 1.8943839073181152, "learning_rate": 2.0326142689820932e-05, "loss": 2.4816, "step": 398050 }, { "epoch": 3.515604302442643, "grad_norm": 3.819336175918579, "learning_rate": 2.0314959484310826e-05, "loss": 2.996, "step": 398100 }, { "epoch": 3.5160458503329273, "grad_norm": 2.8852944374084473, "learning_rate": 2.0303778571718352e-05, "loss": 2.8937, "step": 398150 }, { "epoch": 3.5164873982232114, "grad_norm": 1.736894130706787, "learning_rate": 2.0292599952907177e-05, "loss": 2.6527, "step": 398200 }, { "epoch": 3.5169289461134956, "grad_norm": 12.128264427185059, "learning_rate": 2.0281423628740754e-05, "loss": 2.6312, "step": 398250 }, { "epoch": 3.5173704940037798, "grad_norm": 1.7556575536727905, "learning_rate": 2.0270249600082337e-05, "loss": 2.6397, "step": 398300 }, { "epoch": 3.517812041894064, "grad_norm": 4.753173351287842, "learning_rate": 2.025907786779505e-05, "loss": 2.7519, "step": 398350 }, { "epoch": 3.518253589784348, "grad_norm": 5.984956741333008, "learning_rate": 2.0247908432741792e-05, "loss": 2.9231, "step": 398400 }, { "epoch": 3.5186951376746323, "grad_norm": 2.5510127544403076, "learning_rate": 2.0236964615997516e-05, "loss": 2.6318, "step": 398450 }, { "epoch": 3.5191366855649164, "grad_norm": 3.6083481311798096, "learning_rate": 2.0225799732012764e-05, "loss": 2.921, "step": 398500 }, { "epoch": 3.5195782334552006, "grad_norm": 6.2674479484558105, "learning_rate": 2.0214637147832487e-05, "loss": 3.0455, "step": 398550 }, { "epoch": 3.5200197813454848, "grad_norm": 2.2446072101593018, "learning_rate": 2.0203476864318936e-05, "loss": 3.0641, "step": 398600 }, { "epoch": 3.520461329235769, "grad_norm": 2.3853201866149902, "learning_rate": 2.0192318882334115e-05, "loss": 2.7741, "step": 398650 }, { "epoch": 3.520902877126053, "grad_norm": 2.183690309524536, "learning_rate": 2.0181163202739916e-05, "loss": 2.9059, "step": 398700 }, { "epoch": 3.5213444250163373, "grad_norm": 3.888389825820923, "learning_rate": 2.0170009826397984e-05, "loss": 2.9734, "step": 398750 }, { "epoch": 3.5217859729066214, "grad_norm": 3.0917067527770996, "learning_rate": 2.015885875416984e-05, "loss": 3.2201, "step": 398800 }, { "epoch": 3.5222275207969056, "grad_norm": 4.247842788696289, "learning_rate": 2.0147709986916835e-05, "loss": 2.2652, "step": 398850 }, { "epoch": 3.5226690686871898, "grad_norm": 2.789407253265381, "learning_rate": 2.013656352550007e-05, "loss": 2.7758, "step": 398900 }, { "epoch": 3.523110616577474, "grad_norm": 5.034056663513184, "learning_rate": 2.0125419370780542e-05, "loss": 3.0799, "step": 398950 }, { "epoch": 3.523552164467758, "grad_norm": 7.311521053314209, "learning_rate": 2.0114277523619053e-05, "loss": 3.0746, "step": 399000 }, { "epoch": 3.523552164467758, "eval_asr_loss": 0.952790131057385, "eval_loss": 2.719374895095825, "eval_runtime": 20.361, "eval_samples_per_second": 37.719, "eval_steps_per_second": 9.43, "eval_tts_loss": 6.158039298848886, "step": 399000 }, { "epoch": 3.5239937123580423, "grad_norm": 3.1745970249176025, "learning_rate": 2.01031379848762e-05, "loss": 2.6264, "step": 399050 }, { "epoch": 3.5244352602483264, "grad_norm": 5.393442153930664, "learning_rate": 2.009200075541241e-05, "loss": 3.0917, "step": 399100 }, { "epoch": 3.5248768081386106, "grad_norm": 3.1709952354431152, "learning_rate": 2.0080865836087964e-05, "loss": 3.1125, "step": 399150 }, { "epoch": 3.5253183560288948, "grad_norm": 9.623842239379883, "learning_rate": 2.0069733227762906e-05, "loss": 2.736, "step": 399200 }, { "epoch": 3.525759903919179, "grad_norm": 6.90992546081543, "learning_rate": 2.0058602931297154e-05, "loss": 2.947, "step": 399250 }, { "epoch": 3.526201451809463, "grad_norm": 4.686052322387695, "learning_rate": 2.004747494755045e-05, "loss": 2.9726, "step": 399300 }, { "epoch": 3.5266429996997477, "grad_norm": 2.122558355331421, "learning_rate": 2.0036349277382295e-05, "loss": 2.9755, "step": 399350 }, { "epoch": 3.5270845475900314, "grad_norm": 2.548884153366089, "learning_rate": 2.0025225921652096e-05, "loss": 2.443, "step": 399400 }, { "epoch": 3.527526095480316, "grad_norm": 3.5121986865997314, "learning_rate": 2.0014104881218993e-05, "loss": 3.2671, "step": 399450 }, { "epoch": 3.5279676433705998, "grad_norm": 3.861448049545288, "learning_rate": 2.000298615694204e-05, "loss": 2.9379, "step": 399500 }, { "epoch": 3.5284091912608844, "grad_norm": 4.121931076049805, "learning_rate": 1.9991869749680014e-05, "loss": 3.0564, "step": 399550 }, { "epoch": 3.528850739151168, "grad_norm": 1.607956886291504, "learning_rate": 1.9980755660291595e-05, "loss": 2.7931, "step": 399600 }, { "epoch": 3.5292922870414527, "grad_norm": 5.17902946472168, "learning_rate": 1.996964388963526e-05, "loss": 3.0209, "step": 399650 }, { "epoch": 3.5297338349317364, "grad_norm": 6.150619983673096, "learning_rate": 1.9958534438569264e-05, "loss": 2.7087, "step": 399700 }, { "epoch": 3.530175382822021, "grad_norm": 4.1773176193237305, "learning_rate": 1.994742730795174e-05, "loss": 3.1509, "step": 399750 }, { "epoch": 3.530616930712305, "grad_norm": 3.431478261947632, "learning_rate": 1.9936322498640637e-05, "loss": 3.0157, "step": 399800 }, { "epoch": 3.5310584786025894, "grad_norm": 4.636969089508057, "learning_rate": 1.992522001149369e-05, "loss": 2.5666, "step": 399850 }, { "epoch": 3.5315000264928735, "grad_norm": 5.147570610046387, "learning_rate": 1.9914119847368452e-05, "loss": 3.0153, "step": 399900 }, { "epoch": 3.5319415743831577, "grad_norm": 6.8216118812561035, "learning_rate": 1.990302200712233e-05, "loss": 3.4969, "step": 399950 }, { "epoch": 3.532383122273442, "grad_norm": 4.136614799499512, "learning_rate": 1.9891926491612565e-05, "loss": 3.1275, "step": 400000 }, { "epoch": 3.532824670163726, "grad_norm": 2.361161708831787, "learning_rate": 1.9880833301696156e-05, "loss": 2.7798, "step": 400050 }, { "epoch": 3.53326621805401, "grad_norm": 9.1854887008667, "learning_rate": 1.9869742438229967e-05, "loss": 2.5609, "step": 400100 }, { "epoch": 3.5337077659442944, "grad_norm": 2.0666165351867676, "learning_rate": 1.985865390207069e-05, "loss": 2.7334, "step": 400150 }, { "epoch": 3.5341493138345785, "grad_norm": 2.2963485717773438, "learning_rate": 1.9847567694074815e-05, "loss": 2.5752, "step": 400200 }, { "epoch": 3.5345908617248627, "grad_norm": 6.002432823181152, "learning_rate": 1.9836483815098632e-05, "loss": 3.0839, "step": 400250 }, { "epoch": 3.535032409615147, "grad_norm": 4.753447532653809, "learning_rate": 1.982540226599831e-05, "loss": 3.1853, "step": 400300 }, { "epoch": 3.535473957505431, "grad_norm": 3.831080436706543, "learning_rate": 1.9814323047629775e-05, "loss": 3.2031, "step": 400350 }, { "epoch": 3.535915505395715, "grad_norm": 1.7277547121047974, "learning_rate": 1.9803246160848815e-05, "loss": 2.448, "step": 400400 }, { "epoch": 3.5363570532859994, "grad_norm": 4.296419143676758, "learning_rate": 1.9792171606511044e-05, "loss": 2.827, "step": 400450 }, { "epoch": 3.5367986011762835, "grad_norm": 4.161234378814697, "learning_rate": 1.9781099385471853e-05, "loss": 2.9198, "step": 400500 }, { "epoch": 3.5372401490665677, "grad_norm": 2.296644449234009, "learning_rate": 1.9770029498586478e-05, "loss": 2.895, "step": 400550 }, { "epoch": 3.537681696956852, "grad_norm": 2.6253459453582764, "learning_rate": 1.9758961946709997e-05, "loss": 2.5891, "step": 400600 }, { "epoch": 3.538123244847136, "grad_norm": 3.5181634426116943, "learning_rate": 1.9747896730697275e-05, "loss": 2.6946, "step": 400650 }, { "epoch": 3.53856479273742, "grad_norm": 2.4724299907684326, "learning_rate": 1.973683385140298e-05, "loss": 3.2154, "step": 400700 }, { "epoch": 3.5390063406277044, "grad_norm": 4.1622724533081055, "learning_rate": 1.972577330968165e-05, "loss": 3.1537, "step": 400750 }, { "epoch": 3.5394478885179885, "grad_norm": 3.7553887367248535, "learning_rate": 1.9714715106387627e-05, "loss": 3.3864, "step": 400800 }, { "epoch": 3.5398894364082727, "grad_norm": 4.945655345916748, "learning_rate": 1.9703659242375032e-05, "loss": 3.0338, "step": 400850 }, { "epoch": 3.540330984298557, "grad_norm": 3.4425227642059326, "learning_rate": 1.9692605718497858e-05, "loss": 2.9955, "step": 400900 }, { "epoch": 3.540772532188841, "grad_norm": 4.140172958374023, "learning_rate": 1.968155453560991e-05, "loss": 3.2568, "step": 400950 }, { "epoch": 3.5412140800791256, "grad_norm": 1.9960404634475708, "learning_rate": 1.967050569456478e-05, "loss": 2.5803, "step": 401000 }, { "epoch": 3.5416556279694094, "grad_norm": 2.5637569427490234, "learning_rate": 1.965968010321892e-05, "loss": 2.8252, "step": 401050 }, { "epoch": 3.542097175859694, "grad_norm": 6.17814826965332, "learning_rate": 1.9648635901540167e-05, "loss": 2.5787, "step": 401100 }, { "epoch": 3.5425387237499777, "grad_norm": 8.25796890258789, "learning_rate": 1.963759404424692e-05, "loss": 2.4145, "step": 401150 }, { "epoch": 3.5429802716402623, "grad_norm": 6.433079719543457, "learning_rate": 1.9626554532192082e-05, "loss": 3.0722, "step": 401200 }, { "epoch": 3.543421819530546, "grad_norm": 5.724225044250488, "learning_rate": 1.961551736622833e-05, "loss": 2.7817, "step": 401250 }, { "epoch": 3.5438633674208306, "grad_norm": 0.9677990078926086, "learning_rate": 1.960448254720823e-05, "loss": 2.6504, "step": 401300 }, { "epoch": 3.544304915311115, "grad_norm": 3.493685722351074, "learning_rate": 1.9593450075984105e-05, "loss": 2.2828, "step": 401350 }, { "epoch": 3.544746463201399, "grad_norm": 8.390664100646973, "learning_rate": 1.9582419953408104e-05, "loss": 2.939, "step": 401400 }, { "epoch": 3.545188011091683, "grad_norm": 4.298593997955322, "learning_rate": 1.9571392180332225e-05, "loss": 2.6871, "step": 401450 }, { "epoch": 3.5456295589819673, "grad_norm": 3.8532063961029053, "learning_rate": 1.9560366757608295e-05, "loss": 2.8323, "step": 401500 }, { "epoch": 3.5460711068722515, "grad_norm": 6.241021156311035, "learning_rate": 1.9549343686087895e-05, "loss": 2.853, "step": 401550 }, { "epoch": 3.5465126547625356, "grad_norm": 4.233580112457275, "learning_rate": 1.953832296662248e-05, "loss": 2.9179, "step": 401600 }, { "epoch": 3.54695420265282, "grad_norm": 3.8874735832214355, "learning_rate": 1.9527304600063323e-05, "loss": 3.1638, "step": 401650 }, { "epoch": 3.547395750543104, "grad_norm": 3.45597505569458, "learning_rate": 1.951628858726149e-05, "loss": 2.749, "step": 401700 }, { "epoch": 3.547837298433388, "grad_norm": 3.8734652996063232, "learning_rate": 1.9505274929067848e-05, "loss": 2.6994, "step": 401750 }, { "epoch": 3.5482788463236723, "grad_norm": 6.703948020935059, "learning_rate": 1.9494263626333125e-05, "loss": 2.6224, "step": 401800 }, { "epoch": 3.5487203942139565, "grad_norm": 4.666635990142822, "learning_rate": 1.948325467990788e-05, "loss": 2.6449, "step": 401850 }, { "epoch": 3.5491619421042406, "grad_norm": 5.791102886199951, "learning_rate": 1.947224809064242e-05, "loss": 2.607, "step": 401900 }, { "epoch": 3.549603489994525, "grad_norm": 5.557281017303467, "learning_rate": 1.9461243859386923e-05, "loss": 2.4861, "step": 401950 }, { "epoch": 3.550045037884809, "grad_norm": 2.596440315246582, "learning_rate": 1.94502419869914e-05, "loss": 2.8621, "step": 402000 }, { "epoch": 3.550045037884809, "eval_asr_loss": 0.957611854256335, "eval_loss": 2.710787534713745, "eval_runtime": 20.5046, "eval_samples_per_second": 37.455, "eval_steps_per_second": 9.364, "eval_tts_loss": 6.1543220962821055, "step": 402000 }, { "epoch": 3.550486585775093, "grad_norm": 6.852592945098877, "learning_rate": 1.943924247430561e-05, "loss": 2.4999, "step": 402050 }, { "epoch": 3.5509281336653773, "grad_norm": 4.034320831298828, "learning_rate": 1.9428245322179205e-05, "loss": 2.6819, "step": 402100 }, { "epoch": 3.5513696815556615, "grad_norm": 5.564986705780029, "learning_rate": 1.9417250531461588e-05, "loss": 3.0991, "step": 402150 }, { "epoch": 3.5518112294459456, "grad_norm": 6.231171131134033, "learning_rate": 1.9406258103002062e-05, "loss": 2.701, "step": 402200 }, { "epoch": 3.55225277733623, "grad_norm": 2.058621644973755, "learning_rate": 1.9395268037649645e-05, "loss": 2.8008, "step": 402250 }, { "epoch": 3.552694325226514, "grad_norm": 3.9374427795410156, "learning_rate": 1.938428033625325e-05, "loss": 3.2002, "step": 402300 }, { "epoch": 3.553135873116798, "grad_norm": 2.970907211303711, "learning_rate": 1.9373294999661606e-05, "loss": 2.8169, "step": 402350 }, { "epoch": 3.5535774210070823, "grad_norm": 1.912894606590271, "learning_rate": 1.9362312028723194e-05, "loss": 2.8158, "step": 402400 }, { "epoch": 3.5540189688973665, "grad_norm": 6.366511821746826, "learning_rate": 1.93513314242864e-05, "loss": 3.1281, "step": 402450 }, { "epoch": 3.5544605167876506, "grad_norm": 3.0596821308135986, "learning_rate": 1.934035318719936e-05, "loss": 3.2594, "step": 402500 }, { "epoch": 3.5549020646779352, "grad_norm": 3.8617849349975586, "learning_rate": 1.9329377318310037e-05, "loss": 2.7589, "step": 402550 }, { "epoch": 3.555343612568219, "grad_norm": 3.8856749534606934, "learning_rate": 1.931840381846623e-05, "loss": 2.8757, "step": 402600 }, { "epoch": 3.5557851604585036, "grad_norm": 4.858638286590576, "learning_rate": 1.9307432688515575e-05, "loss": 2.8407, "step": 402650 }, { "epoch": 3.5562267083487873, "grad_norm": 2.845566987991333, "learning_rate": 1.9296463929305464e-05, "loss": 2.9223, "step": 402700 }, { "epoch": 3.556668256239072, "grad_norm": 6.652703762054443, "learning_rate": 1.9285497541683155e-05, "loss": 2.3551, "step": 402750 }, { "epoch": 3.5571098041293556, "grad_norm": 2.4053256511688232, "learning_rate": 1.9274533526495724e-05, "loss": 2.7919, "step": 402800 }, { "epoch": 3.5575513520196402, "grad_norm": 5.947213172912598, "learning_rate": 1.9263571884590036e-05, "loss": 3.2322, "step": 402850 }, { "epoch": 3.557992899909924, "grad_norm": 7.041011810302734, "learning_rate": 1.9252831778896378e-05, "loss": 2.6726, "step": 402900 }, { "epoch": 3.5584344478002086, "grad_norm": 7.163503170013428, "learning_rate": 1.924187483858626e-05, "loss": 3.2268, "step": 402950 }, { "epoch": 3.5588759956904927, "grad_norm": 7.995405197143555, "learning_rate": 1.923092027408049e-05, "loss": 2.9239, "step": 403000 }, { "epoch": 3.559317543580777, "grad_norm": 3.6358609199523926, "learning_rate": 1.921996808622523e-05, "loss": 2.9335, "step": 403050 }, { "epoch": 3.559759091471061, "grad_norm": 5.471379280090332, "learning_rate": 1.9209018275866414e-05, "loss": 3.1064, "step": 403100 }, { "epoch": 3.5602006393613452, "grad_norm": 4.925504207611084, "learning_rate": 1.919807084384985e-05, "loss": 3.0682, "step": 403150 }, { "epoch": 3.5606421872516294, "grad_norm": 2.7538046836853027, "learning_rate": 1.9187125791021104e-05, "loss": 2.8494, "step": 403200 }, { "epoch": 3.5610837351419136, "grad_norm": 3.067248582839966, "learning_rate": 1.917618311822562e-05, "loss": 3.1632, "step": 403250 }, { "epoch": 3.5615252830321977, "grad_norm": 6.017114162445068, "learning_rate": 1.9165242826308598e-05, "loss": 2.8893, "step": 403300 }, { "epoch": 3.561966830922482, "grad_norm": 6.260385036468506, "learning_rate": 1.9154304916115096e-05, "loss": 3.2453, "step": 403350 }, { "epoch": 3.562408378812766, "grad_norm": 3.137697219848633, "learning_rate": 1.9143369388489997e-05, "loss": 2.5129, "step": 403400 }, { "epoch": 3.5628499267030502, "grad_norm": 6.670738697052002, "learning_rate": 1.913243624427793e-05, "loss": 2.5577, "step": 403450 }, { "epoch": 3.5632914745933344, "grad_norm": 4.681761264801025, "learning_rate": 1.912150548432342e-05, "loss": 2.8397, "step": 403500 }, { "epoch": 3.5637330224836186, "grad_norm": 4.229052543640137, "learning_rate": 1.9110577109470785e-05, "loss": 2.9677, "step": 403550 }, { "epoch": 3.5641745703739027, "grad_norm": 5.519686698913574, "learning_rate": 1.9099651120564134e-05, "loss": 3.0402, "step": 403600 }, { "epoch": 3.564616118264187, "grad_norm": 5.242288112640381, "learning_rate": 1.9088727518447386e-05, "loss": 2.8098, "step": 403650 }, { "epoch": 3.565057666154471, "grad_norm": 3.2061898708343506, "learning_rate": 1.907780630396434e-05, "loss": 2.7715, "step": 403700 }, { "epoch": 3.5654992140447552, "grad_norm": 4.586930751800537, "learning_rate": 1.9066887477958523e-05, "loss": 2.7617, "step": 403750 }, { "epoch": 3.5659407619350394, "grad_norm": 5.136094093322754, "learning_rate": 1.9055971041273346e-05, "loss": 2.9923, "step": 403800 }, { "epoch": 3.5663823098253236, "grad_norm": 2.299112319946289, "learning_rate": 1.9045056994752008e-05, "loss": 2.8251, "step": 403850 }, { "epoch": 3.5668238577156077, "grad_norm": 3.2406582832336426, "learning_rate": 1.903414533923754e-05, "loss": 3.1924, "step": 403900 }, { "epoch": 3.567265405605892, "grad_norm": 7.015351295471191, "learning_rate": 1.9023236075572764e-05, "loss": 2.954, "step": 403950 }, { "epoch": 3.567706953496176, "grad_norm": 5.706192493438721, "learning_rate": 1.901232920460031e-05, "loss": 2.9036, "step": 404000 }, { "epoch": 3.5681485013864602, "grad_norm": 3.0247867107391357, "learning_rate": 1.9001424727162674e-05, "loss": 2.3196, "step": 404050 }, { "epoch": 3.5685900492767444, "grad_norm": 3.060110330581665, "learning_rate": 1.8990522644102106e-05, "loss": 2.8463, "step": 404100 }, { "epoch": 3.5690315971670286, "grad_norm": 5.967954158782959, "learning_rate": 1.8979622956260703e-05, "loss": 3.1736, "step": 404150 }, { "epoch": 3.569473145057313, "grad_norm": 3.1381800174713135, "learning_rate": 1.89687256644804e-05, "loss": 2.8856, "step": 404200 }, { "epoch": 3.569914692947597, "grad_norm": 5.74059534072876, "learning_rate": 1.8957830769602886e-05, "loss": 3.1793, "step": 404250 }, { "epoch": 3.5703562408378815, "grad_norm": 6.710801124572754, "learning_rate": 1.8946938272469734e-05, "loss": 3.1067, "step": 404300 }, { "epoch": 3.5707977887281652, "grad_norm": 1.720108151435852, "learning_rate": 1.8936048173922255e-05, "loss": 2.9088, "step": 404350 }, { "epoch": 3.57123933661845, "grad_norm": 5.695639133453369, "learning_rate": 1.8925160474801663e-05, "loss": 2.9813, "step": 404400 }, { "epoch": 3.5716808845087336, "grad_norm": 2.7640626430511475, "learning_rate": 1.8914275175948897e-05, "loss": 2.9203, "step": 404450 }, { "epoch": 3.572122432399018, "grad_norm": 3.0397913455963135, "learning_rate": 1.890339227820478e-05, "loss": 2.8984, "step": 404500 }, { "epoch": 3.5725639802893023, "grad_norm": 5.046758651733398, "learning_rate": 1.8892511782409926e-05, "loss": 2.9453, "step": 404550 }, { "epoch": 3.5730055281795865, "grad_norm": 3.2983615398406982, "learning_rate": 1.8881633689404743e-05, "loss": 2.7396, "step": 404600 }, { "epoch": 3.5734470760698707, "grad_norm": 1.1446799039840698, "learning_rate": 1.887075800002948e-05, "loss": 2.6537, "step": 404650 }, { "epoch": 3.573888623960155, "grad_norm": 3.0751125812530518, "learning_rate": 1.885988471512421e-05, "loss": 2.6273, "step": 404700 }, { "epoch": 3.574330171850439, "grad_norm": 3.845357656478882, "learning_rate": 1.8849013835528784e-05, "loss": 3.0795, "step": 404750 }, { "epoch": 3.574771719740723, "grad_norm": 7.472742557525635, "learning_rate": 1.883814536208287e-05, "loss": 2.394, "step": 404800 }, { "epoch": 3.5752132676310073, "grad_norm": 5.012331962585449, "learning_rate": 1.8827279295626e-05, "loss": 2.8269, "step": 404850 }, { "epoch": 3.5756548155212915, "grad_norm": 5.5619330406188965, "learning_rate": 1.8816415636997453e-05, "loss": 2.6604, "step": 404900 }, { "epoch": 3.5760963634115757, "grad_norm": 5.237339019775391, "learning_rate": 1.8805554387036366e-05, "loss": 2.5317, "step": 404950 }, { "epoch": 3.57653791130186, "grad_norm": 3.345233678817749, "learning_rate": 1.879469554658168e-05, "loss": 2.9277, "step": 405000 }, { "epoch": 3.57653791130186, "eval_asr_loss": 0.9635233027785541, "eval_loss": 2.720118761062622, "eval_runtime": 20.6348, "eval_samples_per_second": 37.219, "eval_steps_per_second": 9.305, "eval_tts_loss": 6.152100371556602, "step": 405000 }, { "epoch": 3.576979459192144, "grad_norm": 1.9346920251846313, "learning_rate": 1.8783839116472168e-05, "loss": 2.6155, "step": 405050 }, { "epoch": 3.577421007082428, "grad_norm": 4.277339935302734, "learning_rate": 1.877298509754638e-05, "loss": 2.953, "step": 405100 }, { "epoch": 3.5778625549727123, "grad_norm": 3.046517848968506, "learning_rate": 1.8762133490642674e-05, "loss": 2.6714, "step": 405150 }, { "epoch": 3.5783041028629965, "grad_norm": 3.184481620788574, "learning_rate": 1.875128429659929e-05, "loss": 2.9032, "step": 405200 }, { "epoch": 3.5787456507532807, "grad_norm": 2.0728960037231445, "learning_rate": 1.8740437516254193e-05, "loss": 3.1191, "step": 405250 }, { "epoch": 3.579187198643565, "grad_norm": 3.5044147968292236, "learning_rate": 1.8729593150445224e-05, "loss": 2.7983, "step": 405300 }, { "epoch": 3.579628746533849, "grad_norm": 5.485553741455078, "learning_rate": 1.8718751200010036e-05, "loss": 2.8994, "step": 405350 }, { "epoch": 3.580070294424133, "grad_norm": 5.7890119552612305, "learning_rate": 1.8707911665786048e-05, "loss": 2.8435, "step": 405400 }, { "epoch": 3.5805118423144173, "grad_norm": 3.7750864028930664, "learning_rate": 1.8697074548610527e-05, "loss": 2.4955, "step": 405450 }, { "epoch": 3.5809533902047015, "grad_norm": 5.962035655975342, "learning_rate": 1.868623984932058e-05, "loss": 2.4039, "step": 405500 }, { "epoch": 3.5813949380949857, "grad_norm": 3.615692138671875, "learning_rate": 1.8675407568753066e-05, "loss": 2.6698, "step": 405550 }, { "epoch": 3.58183648598527, "grad_norm": 3.9165282249450684, "learning_rate": 1.866457770774468e-05, "loss": 2.7644, "step": 405600 }, { "epoch": 3.582278033875554, "grad_norm": 1.6699267625808716, "learning_rate": 1.8653750267131946e-05, "loss": 3.0142, "step": 405650 }, { "epoch": 3.582719581765838, "grad_norm": 2.4093644618988037, "learning_rate": 1.8642925247751214e-05, "loss": 2.6257, "step": 405700 }, { "epoch": 3.5831611296561228, "grad_norm": 0.5638499855995178, "learning_rate": 1.863210265043859e-05, "loss": 2.8217, "step": 405750 }, { "epoch": 3.5836026775464065, "grad_norm": 2.020698070526123, "learning_rate": 1.8621282476030045e-05, "loss": 2.5123, "step": 405800 }, { "epoch": 3.584044225436691, "grad_norm": 6.354174613952637, "learning_rate": 1.861046472536136e-05, "loss": 2.8089, "step": 405850 }, { "epoch": 3.584485773326975, "grad_norm": 2.4718618392944336, "learning_rate": 1.85996493992681e-05, "loss": 2.652, "step": 405900 }, { "epoch": 3.5849273212172594, "grad_norm": 6.159261226654053, "learning_rate": 1.858883649858564e-05, "loss": 2.9554, "step": 405950 }, { "epoch": 3.585368869107543, "grad_norm": 2.7791147232055664, "learning_rate": 1.857802602414922e-05, "loss": 2.9759, "step": 406000 }, { "epoch": 3.5858104169978278, "grad_norm": 5.037733554840088, "learning_rate": 1.8567217976793815e-05, "loss": 2.6602, "step": 406050 }, { "epoch": 3.5862519648881115, "grad_norm": 5.038069248199463, "learning_rate": 1.8556412357354287e-05, "loss": 2.6811, "step": 406100 }, { "epoch": 3.586693512778396, "grad_norm": 7.240410804748535, "learning_rate": 1.8545609166665268e-05, "loss": 2.7145, "step": 406150 }, { "epoch": 3.5871350606686803, "grad_norm": 3.953572988510132, "learning_rate": 1.853480840556123e-05, "loss": 2.7397, "step": 406200 }, { "epoch": 3.5875766085589644, "grad_norm": 2.279768228530884, "learning_rate": 1.852401007487643e-05, "loss": 2.6388, "step": 406250 }, { "epoch": 3.5880181564492486, "grad_norm": 6.771788120269775, "learning_rate": 1.8513214175444927e-05, "loss": 2.8052, "step": 406300 }, { "epoch": 3.5884597043395328, "grad_norm": 4.2230024337768555, "learning_rate": 1.8502420708100644e-05, "loss": 2.6881, "step": 406350 }, { "epoch": 3.588901252229817, "grad_norm": 5.543921947479248, "learning_rate": 1.8491629673677253e-05, "loss": 3.0512, "step": 406400 }, { "epoch": 3.589342800120101, "grad_norm": 4.155614376068115, "learning_rate": 1.848084107300828e-05, "loss": 2.8099, "step": 406450 }, { "epoch": 3.5897843480103853, "grad_norm": 7.474292755126953, "learning_rate": 1.8470054906927083e-05, "loss": 2.6901, "step": 406500 }, { "epoch": 3.5902258959006694, "grad_norm": 6.3721137046813965, "learning_rate": 1.8459271176266762e-05, "loss": 2.8661, "step": 406550 }, { "epoch": 3.5906674437909536, "grad_norm": 2.309178590774536, "learning_rate": 1.844848988186028e-05, "loss": 2.9602, "step": 406600 }, { "epoch": 3.5911089916812378, "grad_norm": 6.081783294677734, "learning_rate": 1.8437711024540426e-05, "loss": 2.3395, "step": 406650 }, { "epoch": 3.591550539571522, "grad_norm": 5.456594467163086, "learning_rate": 1.8427150109630765e-05, "loss": 2.701, "step": 406700 }, { "epoch": 3.591992087461806, "grad_norm": 2.984766960144043, "learning_rate": 1.841637608019848e-05, "loss": 3.0873, "step": 406750 }, { "epoch": 3.5924336353520903, "grad_norm": 5.749645233154297, "learning_rate": 1.8405604490333295e-05, "loss": 2.7556, "step": 406800 }, { "epoch": 3.5928751832423744, "grad_norm": 6.101315498352051, "learning_rate": 1.8394835340867255e-05, "loss": 3.1452, "step": 406850 }, { "epoch": 3.5933167311326586, "grad_norm": 5.195250511169434, "learning_rate": 1.83840686326322e-05, "loss": 2.835, "step": 406900 }, { "epoch": 3.5937582790229428, "grad_norm": 3.1117289066314697, "learning_rate": 1.837330436645973e-05, "loss": 2.7763, "step": 406950 }, { "epoch": 3.594199826913227, "grad_norm": 4.504631519317627, "learning_rate": 1.836254254318133e-05, "loss": 2.8563, "step": 407000 }, { "epoch": 3.594641374803511, "grad_norm": 2.077974319458008, "learning_rate": 1.835178316362824e-05, "loss": 2.5371, "step": 407050 }, { "epoch": 3.5950829226937953, "grad_norm": 5.200381278991699, "learning_rate": 1.8341026228631513e-05, "loss": 2.6435, "step": 407100 }, { "epoch": 3.5955244705840794, "grad_norm": 6.360745906829834, "learning_rate": 1.8330271739022047e-05, "loss": 3.0647, "step": 407150 }, { "epoch": 3.5959660184743636, "grad_norm": 2.2233633995056152, "learning_rate": 1.8319519695630543e-05, "loss": 3.1652, "step": 407200 }, { "epoch": 3.5964075663646478, "grad_norm": 4.717159748077393, "learning_rate": 1.8308770099287516e-05, "loss": 2.8971, "step": 407250 }, { "epoch": 3.5968491142549324, "grad_norm": 5.4331464767456055, "learning_rate": 1.8298022950823245e-05, "loss": 2.7517, "step": 407300 }, { "epoch": 3.597290662145216, "grad_norm": 7.228429794311523, "learning_rate": 1.8287278251067898e-05, "loss": 2.4815, "step": 407350 }, { "epoch": 3.5977322100355007, "grad_norm": 4.294891834259033, "learning_rate": 1.827653600085138e-05, "loss": 2.8224, "step": 407400 }, { "epoch": 3.5981737579257844, "grad_norm": 3.481407403945923, "learning_rate": 1.8265796201003442e-05, "loss": 3.3201, "step": 407450 }, { "epoch": 3.598615305816069, "grad_norm": 4.24852180480957, "learning_rate": 1.8255058852353645e-05, "loss": 3.2116, "step": 407500 }, { "epoch": 3.5990568537063528, "grad_norm": 3.24859881401062, "learning_rate": 1.8244323955731378e-05, "loss": 2.8565, "step": 407550 }, { "epoch": 3.5994984015966374, "grad_norm": 2.8631975650787354, "learning_rate": 1.823359151196579e-05, "loss": 2.958, "step": 407600 }, { "epoch": 3.599939949486921, "grad_norm": 6.874124050140381, "learning_rate": 1.822286152188589e-05, "loss": 2.7483, "step": 407650 }, { "epoch": 3.6003814973772057, "grad_norm": 2.311487913131714, "learning_rate": 1.821213398632049e-05, "loss": 3.1189, "step": 407700 }, { "epoch": 3.60082304526749, "grad_norm": 5.817137241363525, "learning_rate": 1.8201408906098168e-05, "loss": 3.3024, "step": 407750 }, { "epoch": 3.601264593157774, "grad_norm": 5.285879135131836, "learning_rate": 1.8190686282047382e-05, "loss": 3.2537, "step": 407800 }, { "epoch": 3.601706141048058, "grad_norm": 5.478509426116943, "learning_rate": 1.8179966114996326e-05, "loss": 2.59, "step": 407850 }, { "epoch": 3.6021476889383424, "grad_norm": 2.0564489364624023, "learning_rate": 1.816924840577308e-05, "loss": 2.6603, "step": 407900 }, { "epoch": 3.6025892368286265, "grad_norm": 7.0634002685546875, "learning_rate": 1.8158533155205458e-05, "loss": 2.7989, "step": 407950 }, { "epoch": 3.6030307847189107, "grad_norm": 2.7464847564697266, "learning_rate": 1.814782036412115e-05, "loss": 2.5986, "step": 408000 }, { "epoch": 3.6030307847189107, "eval_asr_loss": 0.9543915300974865, "eval_loss": 2.7104358673095703, "eval_runtime": 20.7283, "eval_samples_per_second": 37.051, "eval_steps_per_second": 9.263, "eval_tts_loss": 6.158648672938227, "step": 408000 }, { "epoch": 3.603472332609195, "grad_norm": 3.5688915252685547, "learning_rate": 1.8137110033347633e-05, "loss": 2.7175, "step": 408050 }, { "epoch": 3.603913880499479, "grad_norm": 4.646841049194336, "learning_rate": 1.812640216371216e-05, "loss": 2.9298, "step": 408100 }, { "epoch": 3.604355428389763, "grad_norm": 4.088778972625732, "learning_rate": 1.8115696756041855e-05, "loss": 3.3624, "step": 408150 }, { "epoch": 3.6047969762800474, "grad_norm": 2.804743766784668, "learning_rate": 1.8104993811163607e-05, "loss": 2.7702, "step": 408200 }, { "epoch": 3.6052385241703315, "grad_norm": 7.423593521118164, "learning_rate": 1.8094293329904105e-05, "loss": 2.745, "step": 408250 }, { "epoch": 3.6056800720606157, "grad_norm": 2.9271421432495117, "learning_rate": 1.808359531308989e-05, "loss": 2.9561, "step": 408300 }, { "epoch": 3.6061216199509, "grad_norm": 3.0671887397766113, "learning_rate": 1.8072899761547297e-05, "loss": 3.0375, "step": 408350 }, { "epoch": 3.606563167841184, "grad_norm": 3.644239664077759, "learning_rate": 1.806220667610248e-05, "loss": 2.6942, "step": 408400 }, { "epoch": 3.607004715731468, "grad_norm": 7.394624710083008, "learning_rate": 1.8051516057581353e-05, "loss": 2.8691, "step": 408450 }, { "epoch": 3.6074462636217524, "grad_norm": 0.7626819610595703, "learning_rate": 1.8040827906809715e-05, "loss": 2.7763, "step": 408500 }, { "epoch": 3.6078878115120365, "grad_norm": 8.825603485107422, "learning_rate": 1.8030142224613117e-05, "loss": 2.9804, "step": 408550 }, { "epoch": 3.6083293594023207, "grad_norm": 6.600844860076904, "learning_rate": 1.801945901181692e-05, "loss": 3.011, "step": 408600 }, { "epoch": 3.608770907292605, "grad_norm": 2.803093671798706, "learning_rate": 1.8008778269246325e-05, "loss": 2.6816, "step": 408650 }, { "epoch": 3.609212455182889, "grad_norm": 7.247567176818848, "learning_rate": 1.7998099997726343e-05, "loss": 3.181, "step": 408700 }, { "epoch": 3.609654003073173, "grad_norm": 3.064727544784546, "learning_rate": 1.7987424198081754e-05, "loss": 2.7752, "step": 408750 }, { "epoch": 3.6100955509634574, "grad_norm": 1.1377637386322021, "learning_rate": 1.7976750871137188e-05, "loss": 2.7142, "step": 408800 }, { "epoch": 3.6105370988537415, "grad_norm": 3.3182315826416016, "learning_rate": 1.7966080017717086e-05, "loss": 3.0235, "step": 408850 }, { "epoch": 3.6109786467440257, "grad_norm": 5.15985107421875, "learning_rate": 1.795541163864565e-05, "loss": 2.7353, "step": 408900 }, { "epoch": 3.6114201946343103, "grad_norm": 3.430058717727661, "learning_rate": 1.7944745734746944e-05, "loss": 2.991, "step": 408950 }, { "epoch": 3.611861742524594, "grad_norm": 4.635310173034668, "learning_rate": 1.793408230684479e-05, "loss": 2.9611, "step": 409000 }, { "epoch": 3.6123032904148786, "grad_norm": 5.737512588500977, "learning_rate": 1.7923421355762894e-05, "loss": 3.0159, "step": 409050 }, { "epoch": 3.6127448383051624, "grad_norm": 1.2212871313095093, "learning_rate": 1.7912762882324675e-05, "loss": 3.232, "step": 409100 }, { "epoch": 3.613186386195447, "grad_norm": 6.2190399169921875, "learning_rate": 1.7902106887353432e-05, "loss": 2.7106, "step": 409150 }, { "epoch": 3.6136279340857307, "grad_norm": 4.510745525360107, "learning_rate": 1.789145337167226e-05, "loss": 3.2771, "step": 409200 }, { "epoch": 3.6140694819760153, "grad_norm": 4.28306245803833, "learning_rate": 1.7880802336104034e-05, "loss": 2.917, "step": 409250 }, { "epoch": 3.6145110298662995, "grad_norm": 3.1467583179473877, "learning_rate": 1.787015378147148e-05, "loss": 3.0539, "step": 409300 }, { "epoch": 3.6149525777565836, "grad_norm": 9.277180671691895, "learning_rate": 1.785950770859709e-05, "loss": 3.1138, "step": 409350 }, { "epoch": 3.615394125646868, "grad_norm": 4.58595609664917, "learning_rate": 1.784886411830317e-05, "loss": 2.899, "step": 409400 }, { "epoch": 3.615835673537152, "grad_norm": 3.3005034923553467, "learning_rate": 1.783822301141187e-05, "loss": 2.8341, "step": 409450 }, { "epoch": 3.616277221427436, "grad_norm": 2.925771474838257, "learning_rate": 1.7827584388745112e-05, "loss": 2.9608, "step": 409500 }, { "epoch": 3.6167187693177203, "grad_norm": 3.9021267890930176, "learning_rate": 1.7816948251124666e-05, "loss": 2.6688, "step": 409550 }, { "epoch": 3.6171603172080045, "grad_norm": 7.741336345672607, "learning_rate": 1.7806314599372044e-05, "loss": 3.032, "step": 409600 }, { "epoch": 3.6176018650982886, "grad_norm": 5.34554386138916, "learning_rate": 1.7795683434308625e-05, "loss": 2.6089, "step": 409650 }, { "epoch": 3.618043412988573, "grad_norm": 1.2681831121444702, "learning_rate": 1.7785054756755594e-05, "loss": 2.4431, "step": 409700 }, { "epoch": 3.618484960878857, "grad_norm": 5.2551164627075195, "learning_rate": 1.7774428567533908e-05, "loss": 2.7547, "step": 409750 }, { "epoch": 3.618926508769141, "grad_norm": 8.00971508026123, "learning_rate": 1.776380486746433e-05, "loss": 2.6264, "step": 409800 }, { "epoch": 3.6193680566594253, "grad_norm": 5.919920444488525, "learning_rate": 1.775318365736749e-05, "loss": 2.8673, "step": 409850 }, { "epoch": 3.6198096045497095, "grad_norm": 4.10941219329834, "learning_rate": 1.7742564938063754e-05, "loss": 2.8947, "step": 409900 }, { "epoch": 3.6202511524399936, "grad_norm": 5.745670318603516, "learning_rate": 1.7731948710373336e-05, "loss": 2.4349, "step": 409950 }, { "epoch": 3.620692700330278, "grad_norm": 6.939979076385498, "learning_rate": 1.7721334975116278e-05, "loss": 3.0923, "step": 410000 }, { "epoch": 3.621134248220562, "grad_norm": 5.485790252685547, "learning_rate": 1.7710723733112355e-05, "loss": 2.5638, "step": 410050 }, { "epoch": 3.621575796110846, "grad_norm": 2.9259033203125, "learning_rate": 1.7700114985181238e-05, "loss": 3.0208, "step": 410100 }, { "epoch": 3.6220173440011303, "grad_norm": 3.138120651245117, "learning_rate": 1.768950873214233e-05, "loss": 3.3984, "step": 410150 }, { "epoch": 3.6224588918914145, "grad_norm": 3.577357530593872, "learning_rate": 1.7678904974814896e-05, "loss": 2.6127, "step": 410200 }, { "epoch": 3.6229004397816986, "grad_norm": 10.177245140075684, "learning_rate": 1.766830371401797e-05, "loss": 2.9216, "step": 410250 }, { "epoch": 3.623341987671983, "grad_norm": 3.968435525894165, "learning_rate": 1.7657704950570416e-05, "loss": 3.1709, "step": 410300 }, { "epoch": 3.623783535562267, "grad_norm": 9.71890926361084, "learning_rate": 1.764710868529091e-05, "loss": 2.7931, "step": 410350 }, { "epoch": 3.624225083452551, "grad_norm": 2.890998125076294, "learning_rate": 1.7636514918997904e-05, "loss": 2.705, "step": 410400 }, { "epoch": 3.6246666313428353, "grad_norm": 6.858623504638672, "learning_rate": 1.7625923652509696e-05, "loss": 2.8818, "step": 410450 }, { "epoch": 3.62510817923312, "grad_norm": 5.043021202087402, "learning_rate": 1.7615334886644352e-05, "loss": 3.0241, "step": 410500 }, { "epoch": 3.6255497271234036, "grad_norm": 7.966119766235352, "learning_rate": 1.7604748622219787e-05, "loss": 3.0241, "step": 410550 }, { "epoch": 3.6259912750136882, "grad_norm": 5.578114032745361, "learning_rate": 1.759416486005367e-05, "loss": 2.8936, "step": 410600 }, { "epoch": 3.626432822903972, "grad_norm": 3.3871450424194336, "learning_rate": 1.7583583600963526e-05, "loss": 2.7135, "step": 410650 }, { "epoch": 3.6268743707942566, "grad_norm": 4.504395008087158, "learning_rate": 1.7573004845766684e-05, "loss": 2.9982, "step": 410700 }, { "epoch": 3.6273159186845403, "grad_norm": 5.616184234619141, "learning_rate": 1.7562428595280223e-05, "loss": 2.5273, "step": 410750 }, { "epoch": 3.627757466574825, "grad_norm": 3.015486717224121, "learning_rate": 1.7551854850321093e-05, "loss": 2.9328, "step": 410800 }, { "epoch": 3.6281990144651086, "grad_norm": 3.456696033477783, "learning_rate": 1.754128361170604e-05, "loss": 3.3513, "step": 410850 }, { "epoch": 3.6286405623553932, "grad_norm": 3.8765268325805664, "learning_rate": 1.7530714880251587e-05, "loss": 3.0536, "step": 410900 }, { "epoch": 3.6290821102456774, "grad_norm": 4.825659275054932, "learning_rate": 1.752014865677406e-05, "loss": 3.0883, "step": 410950 }, { "epoch": 3.6295236581359616, "grad_norm": 5.334473609924316, "learning_rate": 1.7509584942089646e-05, "loss": 2.7884, "step": 411000 }, { "epoch": 3.6295236581359616, "eval_asr_loss": 0.9479694496196924, "eval_loss": 2.7095985412597656, "eval_runtime": 20.5013, "eval_samples_per_second": 37.461, "eval_steps_per_second": 9.365, "eval_tts_loss": 6.207482217346372, "step": 411000 }, { "epoch": 3.6299652060262457, "grad_norm": 3.939537525177002, "learning_rate": 1.749902373701427e-05, "loss": 2.8718, "step": 411050 }, { "epoch": 3.63040675391653, "grad_norm": 5.709471225738525, "learning_rate": 1.7488465042363704e-05, "loss": 3.025, "step": 411100 }, { "epoch": 3.630848301806814, "grad_norm": 9.476366996765137, "learning_rate": 1.7477908858953546e-05, "loss": 2.5788, "step": 411150 }, { "epoch": 3.6312898496970982, "grad_norm": 3.3038227558135986, "learning_rate": 1.746735518759913e-05, "loss": 2.9445, "step": 411200 }, { "epoch": 3.6317313975873824, "grad_norm": 3.042025327682495, "learning_rate": 1.7456804029115676e-05, "loss": 3.094, "step": 411250 }, { "epoch": 3.6321729454776666, "grad_norm": 5.474164009094238, "learning_rate": 1.7446255384318133e-05, "loss": 2.8455, "step": 411300 }, { "epoch": 3.6326144933679507, "grad_norm": 3.322467088699341, "learning_rate": 1.7435709254021333e-05, "loss": 3.2126, "step": 411350 }, { "epoch": 3.633056041258235, "grad_norm": 10.738920211791992, "learning_rate": 1.7425165639039838e-05, "loss": 2.8297, "step": 411400 }, { "epoch": 3.633497589148519, "grad_norm": 3.2324726581573486, "learning_rate": 1.741462454018807e-05, "loss": 2.5732, "step": 411450 }, { "epoch": 3.6339391370388032, "grad_norm": 4.073241233825684, "learning_rate": 1.740408595828026e-05, "loss": 3.1418, "step": 411500 }, { "epoch": 3.6343806849290874, "grad_norm": 2.4817397594451904, "learning_rate": 1.7393549894130384e-05, "loss": 2.3846, "step": 411550 }, { "epoch": 3.6348222328193716, "grad_norm": 8.062326431274414, "learning_rate": 1.738301634855229e-05, "loss": 2.3229, "step": 411600 }, { "epoch": 3.6352637807096557, "grad_norm": 5.275685787200928, "learning_rate": 1.7372695918188225e-05, "loss": 2.866, "step": 411650 }, { "epoch": 3.63570532859994, "grad_norm": 5.7611870765686035, "learning_rate": 1.7362167361782423e-05, "loss": 2.8845, "step": 411700 }, { "epoch": 3.636146876490224, "grad_norm": 3.7147860527038574, "learning_rate": 1.735164132637244e-05, "loss": 2.7164, "step": 411750 }, { "epoch": 3.6365884243805082, "grad_norm": 5.1461663246154785, "learning_rate": 1.73411178127713e-05, "loss": 2.2551, "step": 411800 }, { "epoch": 3.6370299722707924, "grad_norm": 6.211675643920898, "learning_rate": 1.7330596821791872e-05, "loss": 3.0107, "step": 411850 }, { "epoch": 3.6374715201610766, "grad_norm": 6.327019214630127, "learning_rate": 1.7320078354246828e-05, "loss": 3.1114, "step": 411900 }, { "epoch": 3.6379130680513607, "grad_norm": 7.114376068115234, "learning_rate": 1.7309562410948616e-05, "loss": 2.5292, "step": 411950 }, { "epoch": 3.638354615941645, "grad_norm": 6.562093734741211, "learning_rate": 1.7299048992709478e-05, "loss": 2.5239, "step": 412000 }, { "epoch": 3.638796163831929, "grad_norm": 9.132491111755371, "learning_rate": 1.7288538100341528e-05, "loss": 3.127, "step": 412050 }, { "epoch": 3.639237711722213, "grad_norm": 8.220430374145508, "learning_rate": 1.7278029734656607e-05, "loss": 2.4543, "step": 412100 }, { "epoch": 3.639679259612498, "grad_norm": 3.1006827354431152, "learning_rate": 1.726752389646642e-05, "loss": 2.8219, "step": 412150 }, { "epoch": 3.6401208075027816, "grad_norm": 3.4816651344299316, "learning_rate": 1.725702058658246e-05, "loss": 2.9274, "step": 412200 }, { "epoch": 3.640562355393066, "grad_norm": 8.825509071350098, "learning_rate": 1.7246519805815996e-05, "loss": 2.8717, "step": 412250 }, { "epoch": 3.64100390328335, "grad_norm": 4.586653709411621, "learning_rate": 1.723602155497813e-05, "loss": 2.6186, "step": 412300 }, { "epoch": 3.6414454511736345, "grad_norm": 4.186963081359863, "learning_rate": 1.722552583487979e-05, "loss": 2.9123, "step": 412350 }, { "epoch": 3.641886999063918, "grad_norm": 6.89837121963501, "learning_rate": 1.721503264633166e-05, "loss": 2.9147, "step": 412400 }, { "epoch": 3.642328546954203, "grad_norm": 4.2538042068481445, "learning_rate": 1.720454199014423e-05, "loss": 2.9455, "step": 412450 }, { "epoch": 3.642770094844487, "grad_norm": 3.2115328311920166, "learning_rate": 1.7194053867127828e-05, "loss": 3.1217, "step": 412500 }, { "epoch": 3.643211642734771, "grad_norm": 5.8354291915893555, "learning_rate": 1.7183568278092604e-05, "loss": 2.6736, "step": 412550 }, { "epoch": 3.6436531906250553, "grad_norm": 4.950815677642822, "learning_rate": 1.7173085223848435e-05, "loss": 3.0864, "step": 412600 }, { "epoch": 3.6440947385153395, "grad_norm": 1.815861463546753, "learning_rate": 1.716260470520507e-05, "loss": 2.8194, "step": 412650 }, { "epoch": 3.6445362864056237, "grad_norm": 4.454792499542236, "learning_rate": 1.7152126722972056e-05, "loss": 3.1947, "step": 412700 }, { "epoch": 3.644977834295908, "grad_norm": 8.62627124786377, "learning_rate": 1.7141651277958708e-05, "loss": 3.1816, "step": 412750 }, { "epoch": 3.645419382186192, "grad_norm": 2.7742252349853516, "learning_rate": 1.7131178370974155e-05, "loss": 3.0351, "step": 412800 }, { "epoch": 3.645860930076476, "grad_norm": 4.986339092254639, "learning_rate": 1.712070800282735e-05, "loss": 2.8859, "step": 412850 }, { "epoch": 3.6463024779667603, "grad_norm": 5.73438024520874, "learning_rate": 1.7110240174327065e-05, "loss": 2.7959, "step": 412900 }, { "epoch": 3.6467440258570445, "grad_norm": 5.311687469482422, "learning_rate": 1.7099774886281805e-05, "loss": 2.1686, "step": 412950 }, { "epoch": 3.6471855737473287, "grad_norm": 4.6534247398376465, "learning_rate": 1.7089312139499952e-05, "loss": 2.4357, "step": 413000 }, { "epoch": 3.647627121637613, "grad_norm": 11.577008247375488, "learning_rate": 1.707885193478968e-05, "loss": 3.0681, "step": 413050 }, { "epoch": 3.648068669527897, "grad_norm": 3.303636074066162, "learning_rate": 1.706839427295892e-05, "loss": 3.3682, "step": 413100 }, { "epoch": 3.648510217418181, "grad_norm": 5.466372013092041, "learning_rate": 1.7057939154815462e-05, "loss": 2.7963, "step": 413150 }, { "epoch": 3.6489517653084653, "grad_norm": 5.667880058288574, "learning_rate": 1.7047486581166867e-05, "loss": 2.8512, "step": 413200 }, { "epoch": 3.6493933131987495, "grad_norm": 4.406205177307129, "learning_rate": 1.7037036552820483e-05, "loss": 2.4373, "step": 413250 }, { "epoch": 3.6498348610890337, "grad_norm": 3.874394416809082, "learning_rate": 1.702658907058351e-05, "loss": 3.157, "step": 413300 }, { "epoch": 3.650276408979318, "grad_norm": 5.33339262008667, "learning_rate": 1.701614413526294e-05, "loss": 2.8328, "step": 413350 }, { "epoch": 3.650717956869602, "grad_norm": 4.727248668670654, "learning_rate": 1.7005701747665532e-05, "loss": 2.8023, "step": 413400 }, { "epoch": 3.651159504759886, "grad_norm": 2.266394853591919, "learning_rate": 1.699526190859787e-05, "loss": 2.8063, "step": 413450 }, { "epoch": 3.6516010526501703, "grad_norm": 3.117126703262329, "learning_rate": 1.6984824618866374e-05, "loss": 2.6294, "step": 413500 }, { "epoch": 3.6520426005404545, "grad_norm": 5.753826141357422, "learning_rate": 1.697438987927722e-05, "loss": 3.203, "step": 413550 }, { "epoch": 3.6524841484307387, "grad_norm": 9.39708137512207, "learning_rate": 1.696395769063638e-05, "loss": 2.4041, "step": 413600 }, { "epoch": 3.652925696321023, "grad_norm": 4.394160270690918, "learning_rate": 1.6953528053749668e-05, "loss": 2.9317, "step": 413650 }, { "epoch": 3.6533672442113074, "grad_norm": 5.457021713256836, "learning_rate": 1.6943100969422703e-05, "loss": 3.0251, "step": 413700 }, { "epoch": 3.653808792101591, "grad_norm": 4.419137001037598, "learning_rate": 1.6932676438460865e-05, "loss": 2.8899, "step": 413750 }, { "epoch": 3.6542503399918758, "grad_norm": 6.008257865905762, "learning_rate": 1.692225446166936e-05, "loss": 2.9489, "step": 413800 }, { "epoch": 3.6546918878821595, "grad_norm": 4.81771993637085, "learning_rate": 1.691183503985323e-05, "loss": 3.2099, "step": 413850 }, { "epoch": 3.655133435772444, "grad_norm": 4.709423065185547, "learning_rate": 1.6901418173817246e-05, "loss": 2.8079, "step": 413900 }, { "epoch": 3.655574983662728, "grad_norm": 3.907165288925171, "learning_rate": 1.6891003864366056e-05, "loss": 2.9984, "step": 413950 }, { "epoch": 3.6560165315530124, "grad_norm": 6.037469387054443, "learning_rate": 1.688059211230406e-05, "loss": 2.7911, "step": 414000 }, { "epoch": 3.6560165315530124, "eval_asr_loss": 0.9639081251779164, "eval_loss": 2.7086124420166016, "eval_runtime": 20.2997, "eval_samples_per_second": 37.833, "eval_steps_per_second": 9.458, "eval_tts_loss": 6.167314742094112, "step": 414000 }, { "epoch": 3.656458079443296, "grad_norm": 4.07073974609375, "learning_rate": 1.6870182918435467e-05, "loss": 3.0173, "step": 414050 }, { "epoch": 3.6568996273335808, "grad_norm": 8.089326858520508, "learning_rate": 1.6859776283564305e-05, "loss": 2.7157, "step": 414100 }, { "epoch": 3.657341175223865, "grad_norm": 3.339818000793457, "learning_rate": 1.6849372208494407e-05, "loss": 2.7987, "step": 414150 }, { "epoch": 3.657782723114149, "grad_norm": 5.463812828063965, "learning_rate": 1.6838970694029416e-05, "loss": 2.7787, "step": 414200 }, { "epoch": 3.6582242710044333, "grad_norm": 4.489798069000244, "learning_rate": 1.682857174097272e-05, "loss": 2.7884, "step": 414250 }, { "epoch": 3.6586658188947174, "grad_norm": 7.622546195983887, "learning_rate": 1.681817535012759e-05, "loss": 3.0251, "step": 414300 }, { "epoch": 3.6591073667850016, "grad_norm": 4.645864486694336, "learning_rate": 1.6807781522297046e-05, "loss": 3.4331, "step": 414350 }, { "epoch": 3.6595489146752858, "grad_norm": 6.42124605178833, "learning_rate": 1.6797390258283892e-05, "loss": 3.0486, "step": 414400 }, { "epoch": 3.65999046256557, "grad_norm": 4.080260276794434, "learning_rate": 1.6787001558890797e-05, "loss": 2.728, "step": 414450 }, { "epoch": 3.660432010455854, "grad_norm": 8.773035049438477, "learning_rate": 1.6776615424920206e-05, "loss": 3.0307, "step": 414500 }, { "epoch": 3.6608735583461383, "grad_norm": 4.263091564178467, "learning_rate": 1.676623185717433e-05, "loss": 2.9163, "step": 414550 }, { "epoch": 3.6613151062364224, "grad_norm": 5.182660102844238, "learning_rate": 1.6755850856455237e-05, "loss": 2.9438, "step": 414600 }, { "epoch": 3.6617566541267066, "grad_norm": 1.5739718675613403, "learning_rate": 1.6745472423564772e-05, "loss": 3.0817, "step": 414650 }, { "epoch": 3.6621982020169908, "grad_norm": 6.9094133377075195, "learning_rate": 1.6735096559304574e-05, "loss": 2.832, "step": 414700 }, { "epoch": 3.662639749907275, "grad_norm": 4.945777416229248, "learning_rate": 1.6724723264476072e-05, "loss": 2.7329, "step": 414750 }, { "epoch": 3.663081297797559, "grad_norm": 5.532716751098633, "learning_rate": 1.6714352539880534e-05, "loss": 2.7446, "step": 414800 }, { "epoch": 3.6635228456878433, "grad_norm": 5.384740829467773, "learning_rate": 1.6703984386319023e-05, "loss": 2.5853, "step": 414850 }, { "epoch": 3.6639643935781274, "grad_norm": 5.625191688537598, "learning_rate": 1.669361880459236e-05, "loss": 2.7077, "step": 414900 }, { "epoch": 3.6644059414684116, "grad_norm": 3.8655409812927246, "learning_rate": 1.668325579550121e-05, "loss": 2.7683, "step": 414950 }, { "epoch": 3.6648474893586958, "grad_norm": 5.248653888702393, "learning_rate": 1.667289535984605e-05, "loss": 3.1264, "step": 415000 }, { "epoch": 3.66528903724898, "grad_norm": 6.724099159240723, "learning_rate": 1.6662537498427112e-05, "loss": 2.469, "step": 415050 }, { "epoch": 3.665730585139264, "grad_norm": 5.48870849609375, "learning_rate": 1.6652182212044463e-05, "loss": 2.7001, "step": 415100 }, { "epoch": 3.6661721330295483, "grad_norm": 4.588411331176758, "learning_rate": 1.6641829501497962e-05, "loss": 3.0649, "step": 415150 }, { "epoch": 3.6666136809198324, "grad_norm": 1.66480553150177, "learning_rate": 1.6631479367587245e-05, "loss": 2.5065, "step": 415200 }, { "epoch": 3.6670552288101166, "grad_norm": 5.377734184265137, "learning_rate": 1.662113181111179e-05, "loss": 3.0223, "step": 415250 }, { "epoch": 3.6674967767004008, "grad_norm": 1.9168970584869385, "learning_rate": 1.661078683287086e-05, "loss": 2.6548, "step": 415300 }, { "epoch": 3.6679383245906854, "grad_norm": 3.597363233566284, "learning_rate": 1.660044443366353e-05, "loss": 2.9992, "step": 415350 }, { "epoch": 3.668379872480969, "grad_norm": 6.0226545333862305, "learning_rate": 1.6590104614288632e-05, "loss": 3.1163, "step": 415400 }, { "epoch": 3.6688214203712537, "grad_norm": 10.560101509094238, "learning_rate": 1.6579767375544864e-05, "loss": 2.7912, "step": 415450 }, { "epoch": 3.6692629682615374, "grad_norm": 5.789887428283691, "learning_rate": 1.6569432718230666e-05, "loss": 3.1321, "step": 415500 }, { "epoch": 3.669704516151822, "grad_norm": 2.4117088317871094, "learning_rate": 1.6559100643144292e-05, "loss": 3.1744, "step": 415550 }, { "epoch": 3.6701460640421057, "grad_norm": 3.0904269218444824, "learning_rate": 1.6548771151083824e-05, "loss": 2.8104, "step": 415600 }, { "epoch": 3.6705876119323904, "grad_norm": 5.256318092346191, "learning_rate": 1.653844424284714e-05, "loss": 2.9841, "step": 415650 }, { "epoch": 3.6710291598226745, "grad_norm": 1.9042861461639404, "learning_rate": 1.652811991923188e-05, "loss": 2.6527, "step": 415700 }, { "epoch": 3.6714707077129587, "grad_norm": 4.885526180267334, "learning_rate": 1.6518004590457192e-05, "loss": 3.1668, "step": 415750 }, { "epoch": 3.671912255603243, "grad_norm": 6.538326740264893, "learning_rate": 1.650768538674488e-05, "loss": 3.0782, "step": 415800 }, { "epoch": 3.672353803493527, "grad_norm": 4.101963996887207, "learning_rate": 1.6497368770029858e-05, "loss": 3.1082, "step": 415850 }, { "epoch": 3.672795351383811, "grad_norm": 3.9150688648223877, "learning_rate": 1.6487054741109015e-05, "loss": 3.0573, "step": 415900 }, { "epoch": 3.6732368992740954, "grad_norm": 7.7804274559021, "learning_rate": 1.6476743300778996e-05, "loss": 2.7626, "step": 415950 }, { "epoch": 3.6736784471643795, "grad_norm": 6.852151870727539, "learning_rate": 1.6466434449836286e-05, "loss": 3.3248, "step": 416000 }, { "epoch": 3.6741199950546637, "grad_norm": 3.954878091812134, "learning_rate": 1.645612818907718e-05, "loss": 3.0324, "step": 416050 }, { "epoch": 3.674561542944948, "grad_norm": 6.219685077667236, "learning_rate": 1.644582451929771e-05, "loss": 3.2051, "step": 416100 }, { "epoch": 3.675003090835232, "grad_norm": 4.971712112426758, "learning_rate": 1.6435523441293783e-05, "loss": 2.9733, "step": 416150 }, { "epoch": 3.675444638725516, "grad_norm": 4.356439113616943, "learning_rate": 1.642522495586103e-05, "loss": 3.1636, "step": 416200 }, { "epoch": 3.6758861866158004, "grad_norm": 4.729641437530518, "learning_rate": 1.641492906379496e-05, "loss": 2.6365, "step": 416250 }, { "epoch": 3.6763277345060845, "grad_norm": 4.354825019836426, "learning_rate": 1.6404635765890814e-05, "loss": 2.6572, "step": 416300 }, { "epoch": 3.6767692823963687, "grad_norm": 6.3755621910095215, "learning_rate": 1.6394345062943672e-05, "loss": 2.7983, "step": 416350 }, { "epoch": 3.677210830286653, "grad_norm": 4.694365501403809, "learning_rate": 1.6384056955748424e-05, "loss": 2.8193, "step": 416400 }, { "epoch": 3.677652378176937, "grad_norm": 5.420905113220215, "learning_rate": 1.6373771445099706e-05, "loss": 3.1355, "step": 416450 }, { "epoch": 3.678093926067221, "grad_norm": 4.239717960357666, "learning_rate": 1.6363488531791998e-05, "loss": 3.0888, "step": 416500 }, { "epoch": 3.6785354739575054, "grad_norm": 5.274794101715088, "learning_rate": 1.6353208216619592e-05, "loss": 3.1985, "step": 416550 }, { "epoch": 3.6789770218477895, "grad_norm": 5.973420143127441, "learning_rate": 1.6342930500376536e-05, "loss": 3.1011, "step": 416600 }, { "epoch": 3.6794185697380737, "grad_norm": 3.4708099365234375, "learning_rate": 1.6332655383856677e-05, "loss": 2.7103, "step": 416650 }, { "epoch": 3.679860117628358, "grad_norm": 6.9011616706848145, "learning_rate": 1.6322382867853714e-05, "loss": 3.2305, "step": 416700 }, { "epoch": 3.680301665518642, "grad_norm": 1.520161509513855, "learning_rate": 1.631211295316108e-05, "loss": 2.8726, "step": 416750 }, { "epoch": 3.680743213408926, "grad_norm": 2.8793442249298096, "learning_rate": 1.6301845640572055e-05, "loss": 2.8742, "step": 416800 }, { "epoch": 3.6811847612992104, "grad_norm": 1.4534064531326294, "learning_rate": 1.6291580930879702e-05, "loss": 3.1764, "step": 416850 }, { "epoch": 3.681626309189495, "grad_norm": 4.110761642456055, "learning_rate": 1.6281318824876902e-05, "loss": 2.7425, "step": 416900 }, { "epoch": 3.6820678570797787, "grad_norm": 8.891091346740723, "learning_rate": 1.6271059323356297e-05, "loss": 2.5105, "step": 416950 }, { "epoch": 3.6825094049700633, "grad_norm": 3.1449079513549805, "learning_rate": 1.6260802427110327e-05, "loss": 3.2826, "step": 417000 }, { "epoch": 3.6825094049700633, "eval_asr_loss": 0.9665250537548044, "eval_loss": 2.7049143314361572, "eval_runtime": 20.6306, "eval_samples_per_second": 37.226, "eval_steps_per_second": 9.307, "eval_tts_loss": 6.16515777805863, "step": 417000 }, { "epoch": 3.682950952860347, "grad_norm": 5.808217525482178, "learning_rate": 1.6250548136931286e-05, "loss": 3.0891, "step": 417050 }, { "epoch": 3.6833925007506316, "grad_norm": 6.160727024078369, "learning_rate": 1.62402964536112e-05, "loss": 2.5984, "step": 417100 }, { "epoch": 3.6838340486409153, "grad_norm": 3.6116251945495605, "learning_rate": 1.6230047377941938e-05, "loss": 2.997, "step": 417150 }, { "epoch": 3.6842755965312, "grad_norm": 3.2706966400146484, "learning_rate": 1.621980091071517e-05, "loss": 3.1699, "step": 417200 }, { "epoch": 3.6847171444214837, "grad_norm": 5.099043846130371, "learning_rate": 1.620955705272232e-05, "loss": 3.0512, "step": 417250 }, { "epoch": 3.6851586923117683, "grad_norm": 6.415900707244873, "learning_rate": 1.6199315804754657e-05, "loss": 3.108, "step": 417300 }, { "epoch": 3.6856002402020525, "grad_norm": 2.3279101848602295, "learning_rate": 1.6189077167603234e-05, "loss": 2.8462, "step": 417350 }, { "epoch": 3.6860417880923366, "grad_norm": 5.883339881896973, "learning_rate": 1.6178841142058898e-05, "loss": 2.9078, "step": 417400 }, { "epoch": 3.686483335982621, "grad_norm": 4.440921306610107, "learning_rate": 1.616860772891227e-05, "loss": 2.9128, "step": 417450 }, { "epoch": 3.686924883872905, "grad_norm": 4.1320319175720215, "learning_rate": 1.6158376928953812e-05, "loss": 2.7999, "step": 417500 }, { "epoch": 3.687366431763189, "grad_norm": 4.820023536682129, "learning_rate": 1.6148148742973784e-05, "loss": 2.8539, "step": 417550 }, { "epoch": 3.6878079796534733, "grad_norm": 3.091703414916992, "learning_rate": 1.6137923171762198e-05, "loss": 2.7265, "step": 417600 }, { "epoch": 3.6882495275437575, "grad_norm": 5.01679801940918, "learning_rate": 1.6127700216108904e-05, "loss": 2.7501, "step": 417650 }, { "epoch": 3.6886910754340416, "grad_norm": 2.086193799972534, "learning_rate": 1.6117479876803555e-05, "loss": 2.9853, "step": 417700 }, { "epoch": 3.689132623324326, "grad_norm": 6.350644588470459, "learning_rate": 1.6107262154635565e-05, "loss": 2.6984, "step": 417750 }, { "epoch": 3.68957417121461, "grad_norm": 1.587799310684204, "learning_rate": 1.6097047050394165e-05, "loss": 2.9635, "step": 417800 }, { "epoch": 3.690015719104894, "grad_norm": 3.1246538162231445, "learning_rate": 1.6086834564868403e-05, "loss": 2.5281, "step": 417850 }, { "epoch": 3.6904572669951783, "grad_norm": 4.698441028594971, "learning_rate": 1.6076624698847074e-05, "loss": 2.4715, "step": 417900 }, { "epoch": 3.6908988148854625, "grad_norm": 3.7532401084899902, "learning_rate": 1.606641745311883e-05, "loss": 3.1362, "step": 417950 }, { "epoch": 3.6913403627757466, "grad_norm": 5.774941444396973, "learning_rate": 1.6056212828472094e-05, "loss": 2.9602, "step": 418000 }, { "epoch": 3.691781910666031, "grad_norm": 4.267138481140137, "learning_rate": 1.604601082569509e-05, "loss": 2.6983, "step": 418050 }, { "epoch": 3.692223458556315, "grad_norm": 5.387126445770264, "learning_rate": 1.603581144557582e-05, "loss": 3.0355, "step": 418100 }, { "epoch": 3.692665006446599, "grad_norm": 3.5310325622558594, "learning_rate": 1.6025614688902117e-05, "loss": 3.117, "step": 418150 }, { "epoch": 3.6931065543368833, "grad_norm": 1.6601978540420532, "learning_rate": 1.6015420556461592e-05, "loss": 2.7358, "step": 418200 }, { "epoch": 3.6935481022271675, "grad_norm": 4.49708366394043, "learning_rate": 1.6005229049041625e-05, "loss": 2.5673, "step": 418250 }, { "epoch": 3.6939896501174516, "grad_norm": 2.247579336166382, "learning_rate": 1.5995040167429453e-05, "loss": 3.2319, "step": 418300 }, { "epoch": 3.694431198007736, "grad_norm": 6.865261077880859, "learning_rate": 1.5984853912412085e-05, "loss": 2.4316, "step": 418350 }, { "epoch": 3.69487274589802, "grad_norm": 1.4032201766967773, "learning_rate": 1.597487393157559e-05, "loss": 2.6907, "step": 418400 }, { "epoch": 3.695314293788304, "grad_norm": 4.891336441040039, "learning_rate": 1.596469287953694e-05, "loss": 2.9249, "step": 418450 }, { "epoch": 3.6957558416785883, "grad_norm": 6.450535774230957, "learning_rate": 1.595451445643714e-05, "loss": 2.8395, "step": 418500 }, { "epoch": 3.696197389568873, "grad_norm": 4.031803131103516, "learning_rate": 1.594433866306241e-05, "loss": 2.461, "step": 418550 }, { "epoch": 3.6966389374591566, "grad_norm": 10.67028522491455, "learning_rate": 1.5934165500198737e-05, "loss": 3.0058, "step": 418600 }, { "epoch": 3.6970804853494412, "grad_norm": 2.609431028366089, "learning_rate": 1.5923994968631902e-05, "loss": 2.5688, "step": 418650 }, { "epoch": 3.697522033239725, "grad_norm": 7.538649559020996, "learning_rate": 1.5914030401337698e-05, "loss": 2.4605, "step": 418700 }, { "epoch": 3.6979635811300096, "grad_norm": 3.253044605255127, "learning_rate": 1.5903865082056076e-05, "loss": 2.8628, "step": 418750 }, { "epoch": 3.6984051290202933, "grad_norm": 1.8412020206451416, "learning_rate": 1.589370239641174e-05, "loss": 2.7755, "step": 418800 }, { "epoch": 3.698846676910578, "grad_norm": 2.989074230194092, "learning_rate": 1.5883542345189673e-05, "loss": 2.6369, "step": 418850 }, { "epoch": 3.699288224800862, "grad_norm": 2.853865146636963, "learning_rate": 1.587338492917466e-05, "loss": 3.1311, "step": 418900 }, { "epoch": 3.6997297726911462, "grad_norm": 2.1574418544769287, "learning_rate": 1.5863230149151277e-05, "loss": 2.7853, "step": 418950 }, { "epoch": 3.7001713205814304, "grad_norm": 4.695131301879883, "learning_rate": 1.585307800590391e-05, "loss": 2.8044, "step": 419000 }, { "epoch": 3.7006128684717146, "grad_norm": 5.562969207763672, "learning_rate": 1.5842928500216686e-05, "loss": 2.8372, "step": 419050 }, { "epoch": 3.7010544163619987, "grad_norm": 4.242596626281738, "learning_rate": 1.5832781632873616e-05, "loss": 3.0112, "step": 419100 }, { "epoch": 3.701495964252283, "grad_norm": 6.428719997406006, "learning_rate": 1.5822637404658418e-05, "loss": 2.9264, "step": 419150 }, { "epoch": 3.701937512142567, "grad_norm": 5.3084235191345215, "learning_rate": 1.581249581635468e-05, "loss": 2.9564, "step": 419200 }, { "epoch": 3.702379060032851, "grad_norm": 5.150052547454834, "learning_rate": 1.580235686874572e-05, "loss": 2.8843, "step": 419250 }, { "epoch": 3.7028206079231354, "grad_norm": 5.73099946975708, "learning_rate": 1.5792220562614722e-05, "loss": 3.0593, "step": 419300 }, { "epoch": 3.7032621558134196, "grad_norm": 2.698573112487793, "learning_rate": 1.5782086898744598e-05, "loss": 2.568, "step": 419350 }, { "epoch": 3.7037037037037037, "grad_norm": 3.24450421333313, "learning_rate": 1.5771955877918097e-05, "loss": 2.7933, "step": 419400 }, { "epoch": 3.704145251593988, "grad_norm": 6.451482772827148, "learning_rate": 1.5761827500917764e-05, "loss": 2.9069, "step": 419450 }, { "epoch": 3.704586799484272, "grad_norm": 4.516376972198486, "learning_rate": 1.575170176852594e-05, "loss": 3.0182, "step": 419500 }, { "epoch": 3.705028347374556, "grad_norm": 2.278684616088867, "learning_rate": 1.5741578681524737e-05, "loss": 2.7373, "step": 419550 }, { "epoch": 3.7054698952648404, "grad_norm": 2.8862128257751465, "learning_rate": 1.5731458240696062e-05, "loss": 2.9625, "step": 419600 }, { "epoch": 3.7059114431551246, "grad_norm": 2.4321420192718506, "learning_rate": 1.5721340446821664e-05, "loss": 2.703, "step": 419650 }, { "epoch": 3.7063529910454087, "grad_norm": 3.649709939956665, "learning_rate": 1.571122530068303e-05, "loss": 2.9528, "step": 419700 }, { "epoch": 3.706794538935693, "grad_norm": 2.276414632797241, "learning_rate": 1.570111280306148e-05, "loss": 2.4859, "step": 419750 }, { "epoch": 3.707236086825977, "grad_norm": 6.385772705078125, "learning_rate": 1.569100295473813e-05, "loss": 2.9312, "step": 419800 }, { "epoch": 3.707677634716261, "grad_norm": 2.4610772132873535, "learning_rate": 1.5680895756493857e-05, "loss": 2.4674, "step": 419850 }, { "epoch": 3.7081191826065454, "grad_norm": 4.764458179473877, "learning_rate": 1.5670791209109386e-05, "loss": 2.8428, "step": 419900 }, { "epoch": 3.7085607304968295, "grad_norm": 8.751736640930176, "learning_rate": 1.5660689313365196e-05, "loss": 3.3121, "step": 419950 }, { "epoch": 3.7090022783871137, "grad_norm": 2.5390260219573975, "learning_rate": 1.5650590070041542e-05, "loss": 2.6776, "step": 420000 }, { "epoch": 3.7090022783871137, "eval_asr_loss": 0.9651873077480778, "eval_loss": 2.706874132156372, "eval_runtime": 20.1763, "eval_samples_per_second": 38.065, "eval_steps_per_second": 9.516, "eval_tts_loss": 6.163012612990288, "step": 420000 }, { "epoch": 3.709443826277398, "grad_norm": 1.8236613273620605, "learning_rate": 1.564049347991854e-05, "loss": 2.6298, "step": 420050 }, { "epoch": 3.7098853741676825, "grad_norm": 6.343883991241455, "learning_rate": 1.5630399543776052e-05, "loss": 2.647, "step": 420100 }, { "epoch": 3.710326922057966, "grad_norm": 4.840651988983154, "learning_rate": 1.5620308262393767e-05, "loss": 2.7928, "step": 420150 }, { "epoch": 3.710768469948251, "grad_norm": 6.563161373138428, "learning_rate": 1.5610219636551125e-05, "loss": 2.9242, "step": 420200 }, { "epoch": 3.7112100178385345, "grad_norm": 6.091010570526123, "learning_rate": 1.56001336670274e-05, "loss": 2.8652, "step": 420250 }, { "epoch": 3.711651565728819, "grad_norm": 7.905683517456055, "learning_rate": 1.559005035460166e-05, "loss": 3.0429, "step": 420300 }, { "epoch": 3.712093113619103, "grad_norm": 5.978575229644775, "learning_rate": 1.5579969700052743e-05, "loss": 2.6211, "step": 420350 }, { "epoch": 3.7125346615093875, "grad_norm": 3.1299948692321777, "learning_rate": 1.556989170415928e-05, "loss": 3.0764, "step": 420400 }, { "epoch": 3.712976209399671, "grad_norm": 4.333583354949951, "learning_rate": 1.5559816367699737e-05, "loss": 2.6937, "step": 420450 }, { "epoch": 3.713417757289956, "grad_norm": 5.727313041687012, "learning_rate": 1.5549743691452316e-05, "loss": 3.0581, "step": 420500 }, { "epoch": 3.71385930518024, "grad_norm": 6.373693466186523, "learning_rate": 1.553967367619507e-05, "loss": 3.2568, "step": 420550 }, { "epoch": 3.714300853070524, "grad_norm": 8.13603401184082, "learning_rate": 1.552960632270582e-05, "loss": 2.614, "step": 420600 }, { "epoch": 3.7147424009608083, "grad_norm": 8.220932960510254, "learning_rate": 1.5519541631762195e-05, "loss": 2.7598, "step": 420650 }, { "epoch": 3.7151839488510925, "grad_norm": 3.417480945587158, "learning_rate": 1.550947960414159e-05, "loss": 3.1792, "step": 420700 }, { "epoch": 3.7156254967413767, "grad_norm": 4.201939582824707, "learning_rate": 1.5499420240621204e-05, "loss": 2.5847, "step": 420750 }, { "epoch": 3.716067044631661, "grad_norm": 3.745560646057129, "learning_rate": 1.548936354197806e-05, "loss": 2.4239, "step": 420800 }, { "epoch": 3.716508592521945, "grad_norm": 2.1603240966796875, "learning_rate": 1.5479309508988926e-05, "loss": 2.7238, "step": 420850 }, { "epoch": 3.716950140412229, "grad_norm": 3.0756444931030273, "learning_rate": 1.5469258142430402e-05, "loss": 3.1915, "step": 420900 }, { "epoch": 3.7173916883025133, "grad_norm": 3.2935378551483154, "learning_rate": 1.5459209443078897e-05, "loss": 2.826, "step": 420950 }, { "epoch": 3.7178332361927975, "grad_norm": 6.561172008514404, "learning_rate": 1.5449163411710542e-05, "loss": 2.9158, "step": 421000 }, { "epoch": 3.7182747840830817, "grad_norm": 1.8914742469787598, "learning_rate": 1.5439120049101335e-05, "loss": 2.5213, "step": 421050 }, { "epoch": 3.718716331973366, "grad_norm": 9.300130844116211, "learning_rate": 1.5429280143722087e-05, "loss": 3.074, "step": 421100 }, { "epoch": 3.71915787986365, "grad_norm": 4.265973091125488, "learning_rate": 1.541924206754444e-05, "loss": 2.9496, "step": 421150 }, { "epoch": 3.719599427753934, "grad_norm": 7.856054306030273, "learning_rate": 1.5409206662437137e-05, "loss": 2.5373, "step": 421200 }, { "epoch": 3.7200409756442183, "grad_norm": 8.116252899169922, "learning_rate": 1.539917392917527e-05, "loss": 2.4183, "step": 421250 }, { "epoch": 3.7204825235345025, "grad_norm": 3.6178231239318848, "learning_rate": 1.5389143868533817e-05, "loss": 3.0305, "step": 421300 }, { "epoch": 3.7209240714247866, "grad_norm": 3.0656232833862305, "learning_rate": 1.5379116481287526e-05, "loss": 2.9831, "step": 421350 }, { "epoch": 3.721365619315071, "grad_norm": 7.130950450897217, "learning_rate": 1.5369091768210904e-05, "loss": 3.0701, "step": 421400 }, { "epoch": 3.721807167205355, "grad_norm": 5.726156234741211, "learning_rate": 1.535906973007827e-05, "loss": 2.8259, "step": 421450 }, { "epoch": 3.722248715095639, "grad_norm": 5.360973834991455, "learning_rate": 1.5349050367663763e-05, "loss": 3.102, "step": 421500 }, { "epoch": 3.7226902629859233, "grad_norm": 4.392642974853516, "learning_rate": 1.5339033681741278e-05, "loss": 2.9108, "step": 421550 }, { "epoch": 3.7231318108762075, "grad_norm": 3.6322884559631348, "learning_rate": 1.5329019673084517e-05, "loss": 2.9484, "step": 421600 }, { "epoch": 3.7235733587664916, "grad_norm": 3.3412508964538574, "learning_rate": 1.5319008342466983e-05, "loss": 2.9681, "step": 421650 }, { "epoch": 3.724014906656776, "grad_norm": 5.286141872406006, "learning_rate": 1.530899969066199e-05, "loss": 3.2208, "step": 421700 }, { "epoch": 3.7244564545470604, "grad_norm": 4.396481037139893, "learning_rate": 1.5298993718442583e-05, "loss": 2.6927, "step": 421750 }, { "epoch": 3.724898002437344, "grad_norm": 2.5526371002197266, "learning_rate": 1.5288990426581667e-05, "loss": 2.9439, "step": 421800 }, { "epoch": 3.7253395503276288, "grad_norm": 1.7771072387695312, "learning_rate": 1.5278989815851903e-05, "loss": 2.747, "step": 421850 }, { "epoch": 3.7257810982179125, "grad_norm": 8.5764741897583, "learning_rate": 1.5268991887025734e-05, "loss": 2.8771, "step": 421900 }, { "epoch": 3.726222646108197, "grad_norm": 9.238790512084961, "learning_rate": 1.5258996640875434e-05, "loss": 2.6763, "step": 421950 }, { "epoch": 3.726664193998481, "grad_norm": 9.174539566040039, "learning_rate": 1.5249004078173062e-05, "loss": 2.9196, "step": 422000 }, { "epoch": 3.7271057418887654, "grad_norm": 5.012724876403809, "learning_rate": 1.5239014199690427e-05, "loss": 2.709, "step": 422050 }, { "epoch": 3.7275472897790496, "grad_norm": 4.2369256019592285, "learning_rate": 1.5229027006199182e-05, "loss": 3.3595, "step": 422100 }, { "epoch": 3.7279888376693338, "grad_norm": 1.6443045139312744, "learning_rate": 1.5219042498470771e-05, "loss": 2.7395, "step": 422150 }, { "epoch": 3.728430385559618, "grad_norm": 3.9473495483398438, "learning_rate": 1.5209060677276388e-05, "loss": 2.836, "step": 422200 }, { "epoch": 3.728871933449902, "grad_norm": 6.068444728851318, "learning_rate": 1.5199081543387034e-05, "loss": 3.0057, "step": 422250 }, { "epoch": 3.7293134813401863, "grad_norm": 6.896420001983643, "learning_rate": 1.518910509757353e-05, "loss": 3.0238, "step": 422300 }, { "epoch": 3.7297550292304704, "grad_norm": 2.277879238128662, "learning_rate": 1.5179131340606485e-05, "loss": 2.7697, "step": 422350 }, { "epoch": 3.7301965771207546, "grad_norm": 2.9647977352142334, "learning_rate": 1.5169160273256255e-05, "loss": 2.8859, "step": 422400 }, { "epoch": 3.7306381250110388, "grad_norm": 3.141303539276123, "learning_rate": 1.5159191896293035e-05, "loss": 2.7326, "step": 422450 }, { "epoch": 3.731079672901323, "grad_norm": 5.790811538696289, "learning_rate": 1.5149226210486817e-05, "loss": 2.9612, "step": 422500 }, { "epoch": 3.731521220791607, "grad_norm": 5.7316694259643555, "learning_rate": 1.5139263216607336e-05, "loss": 2.3694, "step": 422550 }, { "epoch": 3.7319627686818913, "grad_norm": 10.18781852722168, "learning_rate": 1.5129302915424171e-05, "loss": 2.6179, "step": 422600 }, { "epoch": 3.7324043165721754, "grad_norm": 3.858480453491211, "learning_rate": 1.5119345307706662e-05, "loss": 3.0264, "step": 422650 }, { "epoch": 3.7328458644624596, "grad_norm": 8.511419296264648, "learning_rate": 1.5109390394223937e-05, "loss": 2.8611, "step": 422700 }, { "epoch": 3.7332874123527438, "grad_norm": 4.767001152038574, "learning_rate": 1.5099438175744934e-05, "loss": 2.9658, "step": 422750 }, { "epoch": 3.733728960243028, "grad_norm": 9.223686218261719, "learning_rate": 1.5089488653038391e-05, "loss": 2.7748, "step": 422800 }, { "epoch": 3.734170508133312, "grad_norm": 3.6123385429382324, "learning_rate": 1.5079541826872829e-05, "loss": 3.2978, "step": 422850 }, { "epoch": 3.7346120560235962, "grad_norm": 6.01680326461792, "learning_rate": 1.506959769801653e-05, "loss": 2.8253, "step": 422900 }, { "epoch": 3.7350536039138804, "grad_norm": 5.655185699462891, "learning_rate": 1.5059656267237621e-05, "loss": 2.7246, "step": 422950 }, { "epoch": 3.7354951518041646, "grad_norm": 4.028355598449707, "learning_rate": 1.5049717535303981e-05, "loss": 2.9338, "step": 423000 }, { "epoch": 3.7354951518041646, "eval_asr_loss": 0.958945356114306, "eval_loss": 2.701504707336426, "eval_runtime": 20.4605, "eval_samples_per_second": 37.536, "eval_steps_per_second": 9.384, "eval_tts_loss": 6.185760407929624, "step": 423000 }, { "epoch": 3.7359366996944487, "grad_norm": 4.1780266761779785, "learning_rate": 1.5039781502983274e-05, "loss": 2.9984, "step": 423050 }, { "epoch": 3.736378247584733, "grad_norm": 5.605010986328125, "learning_rate": 1.5029848171042992e-05, "loss": 3.069, "step": 423100 }, { "epoch": 3.736819795475017, "grad_norm": 7.2197394371032715, "learning_rate": 1.5019917540250422e-05, "loss": 3.025, "step": 423150 }, { "epoch": 3.7372613433653012, "grad_norm": 7.252914905548096, "learning_rate": 1.5009989611372583e-05, "loss": 2.8276, "step": 423200 }, { "epoch": 3.7377028912555854, "grad_norm": 2.628473997116089, "learning_rate": 1.5000064385176332e-05, "loss": 2.8928, "step": 423250 }, { "epoch": 3.73814443914587, "grad_norm": 6.169347286224365, "learning_rate": 1.4990141862428336e-05, "loss": 3.0363, "step": 423300 }, { "epoch": 3.7385859870361537, "grad_norm": 8.487960815429688, "learning_rate": 1.4980222043895009e-05, "loss": 3.2162, "step": 423350 }, { "epoch": 3.7390275349264384, "grad_norm": 3.758650302886963, "learning_rate": 1.4970304930342555e-05, "loss": 2.6424, "step": 423400 }, { "epoch": 3.739469082816722, "grad_norm": 3.9521355628967285, "learning_rate": 1.4960390522537e-05, "loss": 2.7416, "step": 423450 }, { "epoch": 3.7399106307070067, "grad_norm": 4.610512733459473, "learning_rate": 1.4950478821244162e-05, "loss": 2.8988, "step": 423500 }, { "epoch": 3.7403521785972904, "grad_norm": 2.820340156555176, "learning_rate": 1.494056982722961e-05, "loss": 2.9797, "step": 423550 }, { "epoch": 3.740793726487575, "grad_norm": 4.4627275466918945, "learning_rate": 1.4930663541258749e-05, "loss": 3.1456, "step": 423600 }, { "epoch": 3.741235274377859, "grad_norm": 2.5932023525238037, "learning_rate": 1.4920759964096759e-05, "loss": 2.994, "step": 423650 }, { "epoch": 3.7416768222681434, "grad_norm": 2.8277041912078857, "learning_rate": 1.4910859096508589e-05, "loss": 2.6632, "step": 423700 }, { "epoch": 3.7421183701584275, "grad_norm": 1.231897234916687, "learning_rate": 1.4900960939259018e-05, "loss": 3.0677, "step": 423750 }, { "epoch": 3.7425599180487117, "grad_norm": 6.176322937011719, "learning_rate": 1.4891065493112583e-05, "loss": 2.2609, "step": 423800 }, { "epoch": 3.743001465938996, "grad_norm": 7.469415664672852, "learning_rate": 1.4881172758833611e-05, "loss": 2.8651, "step": 423850 }, { "epoch": 3.74344301382928, "grad_norm": 6.156076908111572, "learning_rate": 1.4871282737186244e-05, "loss": 2.7196, "step": 423900 }, { "epoch": 3.743884561719564, "grad_norm": 5.446041584014893, "learning_rate": 1.4861395428934416e-05, "loss": 3.2504, "step": 423950 }, { "epoch": 3.7443261096098484, "grad_norm": 3.9086825847625732, "learning_rate": 1.485151083484181e-05, "loss": 2.9201, "step": 424000 }, { "epoch": 3.7447676575001325, "grad_norm": 7.786035537719727, "learning_rate": 1.4841628955671938e-05, "loss": 2.9338, "step": 424050 }, { "epoch": 3.7452092053904167, "grad_norm": 3.0218162536621094, "learning_rate": 1.4831749792188116e-05, "loss": 2.9081, "step": 424100 }, { "epoch": 3.745650753280701, "grad_norm": 4.860859394073486, "learning_rate": 1.4821873345153402e-05, "loss": 3.2792, "step": 424150 }, { "epoch": 3.746092301170985, "grad_norm": 6.053365707397461, "learning_rate": 1.4811999615330651e-05, "loss": 2.8451, "step": 424200 }, { "epoch": 3.746533849061269, "grad_norm": 3.13797664642334, "learning_rate": 1.4802128603482552e-05, "loss": 3.12, "step": 424250 }, { "epoch": 3.7469753969515533, "grad_norm": 5.552696704864502, "learning_rate": 1.4792260310371559e-05, "loss": 2.8514, "step": 424300 }, { "epoch": 3.7474169448418375, "grad_norm": 4.241985321044922, "learning_rate": 1.478239473675989e-05, "loss": 2.7911, "step": 424350 }, { "epoch": 3.7478584927321217, "grad_norm": 3.5624613761901855, "learning_rate": 1.477253188340959e-05, "loss": 2.7599, "step": 424400 }, { "epoch": 3.748300040622406, "grad_norm": 3.82011079788208, "learning_rate": 1.4762671751082503e-05, "loss": 2.8772, "step": 424450 }, { "epoch": 3.74874158851269, "grad_norm": 4.982883453369141, "learning_rate": 1.47528143405402e-05, "loss": 3.1361, "step": 424500 }, { "epoch": 3.749183136402974, "grad_norm": 2.279075860977173, "learning_rate": 1.4742959652544114e-05, "loss": 2.3837, "step": 424550 }, { "epoch": 3.7496246842932583, "grad_norm": 3.476837396621704, "learning_rate": 1.4733107687855413e-05, "loss": 2.6078, "step": 424600 }, { "epoch": 3.7500662321835425, "grad_norm": 5.043578624725342, "learning_rate": 1.47232584472351e-05, "loss": 2.7216, "step": 424650 }, { "epoch": 3.7505077800738267, "grad_norm": 6.663382053375244, "learning_rate": 1.4713411931443921e-05, "loss": 2.8658, "step": 424700 }, { "epoch": 3.750949327964111, "grad_norm": 6.863880634307861, "learning_rate": 1.4703568141242453e-05, "loss": 2.4983, "step": 424750 }, { "epoch": 3.751390875854395, "grad_norm": 4.38988733291626, "learning_rate": 1.4693727077391051e-05, "loss": 3.032, "step": 424800 }, { "epoch": 3.7518324237446796, "grad_norm": 2.6530613899230957, "learning_rate": 1.4683888740649826e-05, "loss": 3.0654, "step": 424850 }, { "epoch": 3.7522739716349633, "grad_norm": 3.4218332767486572, "learning_rate": 1.4674053131778742e-05, "loss": 2.9964, "step": 424900 }, { "epoch": 3.752715519525248, "grad_norm": 4.941816806793213, "learning_rate": 1.4664220251537503e-05, "loss": 3.1881, "step": 424950 }, { "epoch": 3.7531570674155317, "grad_norm": 5.61851692199707, "learning_rate": 1.465439010068559e-05, "loss": 2.8703, "step": 425000 }, { "epoch": 3.7535986153058163, "grad_norm": 2.025007963180542, "learning_rate": 1.464456267998232e-05, "loss": 2.6551, "step": 425050 }, { "epoch": 3.7540401631961, "grad_norm": 7.5298027992248535, "learning_rate": 1.4634737990186798e-05, "loss": 2.422, "step": 425100 }, { "epoch": 3.7544817110863846, "grad_norm": 5.5083417892456055, "learning_rate": 1.4624916032057856e-05, "loss": 3.1364, "step": 425150 }, { "epoch": 3.7549232589766683, "grad_norm": 3.3503365516662598, "learning_rate": 1.4615096806354184e-05, "loss": 2.861, "step": 425200 }, { "epoch": 3.755364806866953, "grad_norm": 6.55965518951416, "learning_rate": 1.4605280313834241e-05, "loss": 3.204, "step": 425250 }, { "epoch": 3.755806354757237, "grad_norm": 4.813043117523193, "learning_rate": 1.4595466555256238e-05, "loss": 2.7254, "step": 425300 }, { "epoch": 3.7562479026475213, "grad_norm": 4.442914962768555, "learning_rate": 1.4585655531378239e-05, "loss": 2.9333, "step": 425350 }, { "epoch": 3.7566894505378055, "grad_norm": 6.5663275718688965, "learning_rate": 1.457584724295803e-05, "loss": 2.8306, "step": 425400 }, { "epoch": 3.7571309984280896, "grad_norm": 7.873234272003174, "learning_rate": 1.4566237774977526e-05, "loss": 2.7571, "step": 425450 }, { "epoch": 3.757572546318374, "grad_norm": 2.9098103046417236, "learning_rate": 1.4556434904998673e-05, "loss": 2.8657, "step": 425500 }, { "epoch": 3.758014094208658, "grad_norm": 7.9524455070495605, "learning_rate": 1.4546830748545443e-05, "loss": 2.7839, "step": 425550 }, { "epoch": 3.758455642098942, "grad_norm": 4.021110534667969, "learning_rate": 1.4537033299976415e-05, "loss": 3.0933, "step": 425600 }, { "epoch": 3.7588971899892263, "grad_norm": 3.0051496028900146, "learning_rate": 1.4527238590620834e-05, "loss": 2.9459, "step": 425650 }, { "epoch": 3.7593387378795105, "grad_norm": 5.82952356338501, "learning_rate": 1.4517446621235275e-05, "loss": 2.7297, "step": 425700 }, { "epoch": 3.7597802857697946, "grad_norm": 3.3522469997406006, "learning_rate": 1.4507657392576096e-05, "loss": 2.9713, "step": 425750 }, { "epoch": 3.760221833660079, "grad_norm": 4.062835693359375, "learning_rate": 1.4497870905399403e-05, "loss": 3.072, "step": 425800 }, { "epoch": 3.760663381550363, "grad_norm": 5.647963523864746, "learning_rate": 1.4488087160461128e-05, "loss": 3.17, "step": 425850 }, { "epoch": 3.761104929440647, "grad_norm": 7.107373237609863, "learning_rate": 1.4478306158517008e-05, "loss": 2.8224, "step": 425900 }, { "epoch": 3.7615464773309313, "grad_norm": 4.970780372619629, "learning_rate": 1.4468527900322515e-05, "loss": 3.0553, "step": 425950 }, { "epoch": 3.7619880252212154, "grad_norm": 4.6371893882751465, "learning_rate": 1.4458752386632934e-05, "loss": 2.7625, "step": 426000 }, { "epoch": 3.7619880252212154, "eval_asr_loss": 0.9504576543873883, "eval_loss": 2.700202703475952, "eval_runtime": 20.2379, "eval_samples_per_second": 37.949, "eval_steps_per_second": 9.487, "eval_tts_loss": 6.197590589869697, "step": 426000 }, { "epoch": 3.7624295731114996, "grad_norm": 4.084286689758301, "learning_rate": 1.444897961820335e-05, "loss": 2.7539, "step": 426050 }, { "epoch": 3.762871121001784, "grad_norm": 6.525496006011963, "learning_rate": 1.4439209595788644e-05, "loss": 3.1227, "step": 426100 }, { "epoch": 3.763312668892068, "grad_norm": 4.383753776550293, "learning_rate": 1.4429442320143432e-05, "loss": 2.9129, "step": 426150 }, { "epoch": 3.763754216782352, "grad_norm": 2.5605907440185547, "learning_rate": 1.4419677792022168e-05, "loss": 2.7934, "step": 426200 }, { "epoch": 3.7641957646726363, "grad_norm": 5.38555908203125, "learning_rate": 1.4409916012179103e-05, "loss": 2.6958, "step": 426250 }, { "epoch": 3.7646373125629204, "grad_norm": 5.414951801300049, "learning_rate": 1.4400156981368228e-05, "loss": 3.1326, "step": 426300 }, { "epoch": 3.7650788604532046, "grad_norm": 5.610614776611328, "learning_rate": 1.4390400700343331e-05, "loss": 3.2245, "step": 426350 }, { "epoch": 3.765520408343489, "grad_norm": 3.588418483734131, "learning_rate": 1.4380647169858042e-05, "loss": 3.0074, "step": 426400 }, { "epoch": 3.765961956233773, "grad_norm": 3.9011716842651367, "learning_rate": 1.4370896390665694e-05, "loss": 2.7107, "step": 426450 }, { "epoch": 3.7664035041240576, "grad_norm": 2.626692056655884, "learning_rate": 1.4361148363519478e-05, "loss": 3.0024, "step": 426500 }, { "epoch": 3.7668450520143413, "grad_norm": 6.196200847625732, "learning_rate": 1.4351403089172355e-05, "loss": 2.7456, "step": 426550 }, { "epoch": 3.767286599904626, "grad_norm": 2.3398287296295166, "learning_rate": 1.4341660568377035e-05, "loss": 2.8232, "step": 426600 }, { "epoch": 3.7677281477949096, "grad_norm": 4.439255714416504, "learning_rate": 1.4331920801886068e-05, "loss": 2.3987, "step": 426650 }, { "epoch": 3.768169695685194, "grad_norm": 4.774211883544922, "learning_rate": 1.4322183790451772e-05, "loss": 2.821, "step": 426700 }, { "epoch": 3.768611243575478, "grad_norm": 6.328277587890625, "learning_rate": 1.4312449534826238e-05, "loss": 2.9038, "step": 426750 }, { "epoch": 3.7690527914657626, "grad_norm": 3.2334492206573486, "learning_rate": 1.4302718035761349e-05, "loss": 2.8593, "step": 426800 }, { "epoch": 3.7694943393560467, "grad_norm": 7.587881565093994, "learning_rate": 1.429298929400878e-05, "loss": 2.6002, "step": 426850 }, { "epoch": 3.769935887246331, "grad_norm": 2.7279462814331055, "learning_rate": 1.4283263310320022e-05, "loss": 2.743, "step": 426900 }, { "epoch": 3.770377435136615, "grad_norm": 7.930053234100342, "learning_rate": 1.4273540085446286e-05, "loss": 2.8873, "step": 426950 }, { "epoch": 3.770818983026899, "grad_norm": 6.399631977081299, "learning_rate": 1.4263819620138624e-05, "loss": 2.9122, "step": 427000 }, { "epoch": 3.7712605309171834, "grad_norm": 2.9639124870300293, "learning_rate": 1.4254101915147883e-05, "loss": 2.2211, "step": 427050 }, { "epoch": 3.7717020788074676, "grad_norm": 4.479401111602783, "learning_rate": 1.4244386971224649e-05, "loss": 3.3403, "step": 427100 }, { "epoch": 3.7721436266977517, "grad_norm": 2.668579339981079, "learning_rate": 1.4234674789119312e-05, "loss": 2.8621, "step": 427150 }, { "epoch": 3.772585174588036, "grad_norm": 6.794436931610107, "learning_rate": 1.422496536958206e-05, "loss": 2.6881, "step": 427200 }, { "epoch": 3.77302672247832, "grad_norm": 9.264042854309082, "learning_rate": 1.4215258713362889e-05, "loss": 2.958, "step": 427250 }, { "epoch": 3.773468270368604, "grad_norm": 11.030410766601562, "learning_rate": 1.420555482121152e-05, "loss": 2.4334, "step": 427300 }, { "epoch": 3.7739098182588884, "grad_norm": 6.277973651885986, "learning_rate": 1.4195853693877514e-05, "loss": 3.0128, "step": 427350 }, { "epoch": 3.7743513661491725, "grad_norm": 1.0496736764907837, "learning_rate": 1.4186155332110218e-05, "loss": 3.0178, "step": 427400 }, { "epoch": 3.7747929140394567, "grad_norm": 5.0421342849731445, "learning_rate": 1.417645973665871e-05, "loss": 2.7212, "step": 427450 }, { "epoch": 3.775234461929741, "grad_norm": 1.7735005617141724, "learning_rate": 1.4166766908271928e-05, "loss": 3.2721, "step": 427500 }, { "epoch": 3.775676009820025, "grad_norm": 2.415290594100952, "learning_rate": 1.4157076847698547e-05, "loss": 2.7527, "step": 427550 }, { "epoch": 3.776117557710309, "grad_norm": 7.094167232513428, "learning_rate": 1.414738955568703e-05, "loss": 2.6113, "step": 427600 }, { "epoch": 3.7765591056005934, "grad_norm": 6.109889984130859, "learning_rate": 1.4137705032985644e-05, "loss": 2.7697, "step": 427650 }, { "epoch": 3.7770006534908775, "grad_norm": 5.912403106689453, "learning_rate": 1.4128023280342451e-05, "loss": 2.7686, "step": 427700 }, { "epoch": 3.7774422013811617, "grad_norm": 6.070478916168213, "learning_rate": 1.4118344298505265e-05, "loss": 2.9174, "step": 427750 }, { "epoch": 3.777883749271446, "grad_norm": 8.196060180664062, "learning_rate": 1.410866808822171e-05, "loss": 2.3266, "step": 427800 }, { "epoch": 3.77832529716173, "grad_norm": 8.014781951904297, "learning_rate": 1.4098994650239216e-05, "loss": 2.7757, "step": 427850 }, { "epoch": 3.778766845052014, "grad_norm": 3.97485613822937, "learning_rate": 1.4089323985304952e-05, "loss": 2.9069, "step": 427900 }, { "epoch": 3.7792083929422984, "grad_norm": 5.615063667297363, "learning_rate": 1.4079656094165882e-05, "loss": 3.0598, "step": 427950 }, { "epoch": 3.7796499408325825, "grad_norm": 7.392854690551758, "learning_rate": 1.4069990977568782e-05, "loss": 2.6921, "step": 428000 }, { "epoch": 3.780091488722867, "grad_norm": 11.116724014282227, "learning_rate": 1.4060328636260218e-05, "loss": 2.8788, "step": 428050 }, { "epoch": 3.780533036613151, "grad_norm": 5.172914028167725, "learning_rate": 1.4050669070986488e-05, "loss": 2.949, "step": 428100 }, { "epoch": 3.7809745845034355, "grad_norm": 3.4456288814544678, "learning_rate": 1.4041012282493732e-05, "loss": 2.3669, "step": 428150 }, { "epoch": 3.781416132393719, "grad_norm": 2.290354013442993, "learning_rate": 1.4031358271527873e-05, "loss": 2.6882, "step": 428200 }, { "epoch": 3.781857680284004, "grad_norm": 5.4368391036987305, "learning_rate": 1.4021707038834564e-05, "loss": 2.8346, "step": 428250 }, { "epoch": 3.7822992281742875, "grad_norm": 4.729309558868408, "learning_rate": 1.4012058585159316e-05, "loss": 3.362, "step": 428300 }, { "epoch": 3.782740776064572, "grad_norm": 2.4436662197113037, "learning_rate": 1.400241291124736e-05, "loss": 2.3818, "step": 428350 }, { "epoch": 3.783182323954856, "grad_norm": 6.35623025894165, "learning_rate": 1.3992770017843764e-05, "loss": 2.6949, "step": 428400 }, { "epoch": 3.7836238718451405, "grad_norm": 6.007364273071289, "learning_rate": 1.3983129905693343e-05, "loss": 2.8918, "step": 428450 }, { "epoch": 3.7840654197354247, "grad_norm": 2.9248313903808594, "learning_rate": 1.397349257554072e-05, "loss": 3.1739, "step": 428500 }, { "epoch": 3.784506967625709, "grad_norm": 2.3103106021881104, "learning_rate": 1.3963858028130317e-05, "loss": 2.9258, "step": 428550 }, { "epoch": 3.784948515515993, "grad_norm": 7.5655035972595215, "learning_rate": 1.3954226264206289e-05, "loss": 2.79, "step": 428600 }, { "epoch": 3.785390063406277, "grad_norm": 6.4904937744140625, "learning_rate": 1.3944597284512633e-05, "loss": 2.802, "step": 428650 }, { "epoch": 3.7858316112965613, "grad_norm": 4.967916011810303, "learning_rate": 1.3934971089793102e-05, "loss": 3.0312, "step": 428700 }, { "epoch": 3.7862731591868455, "grad_norm": 3.0951921939849854, "learning_rate": 1.3925347680791212e-05, "loss": 2.6653, "step": 428750 }, { "epoch": 3.7867147070771296, "grad_norm": 3.5811398029327393, "learning_rate": 1.391572705825031e-05, "loss": 2.9395, "step": 428800 }, { "epoch": 3.787156254967414, "grad_norm": 8.422260284423828, "learning_rate": 1.3906109222913527e-05, "loss": 3.1555, "step": 428850 }, { "epoch": 3.787597802857698, "grad_norm": 5.059603214263916, "learning_rate": 1.3896494175523716e-05, "loss": 3.2869, "step": 428900 }, { "epoch": 3.788039350747982, "grad_norm": 5.9647955894470215, "learning_rate": 1.388688191682359e-05, "loss": 2.7305, "step": 428950 }, { "epoch": 3.7884808986382663, "grad_norm": 4.226194858551025, "learning_rate": 1.3877272447555618e-05, "loss": 2.6869, "step": 429000 }, { "epoch": 3.7884808986382663, "eval_asr_loss": 0.958152736509874, "eval_loss": 2.698516845703125, "eval_runtime": 20.1615, "eval_samples_per_second": 38.092, "eval_steps_per_second": 9.523, "eval_tts_loss": 6.174189273698873, "step": 429000 }, { "epoch": 3.7889224465285505, "grad_norm": 4.801724910736084, "learning_rate": 1.3867665768462034e-05, "loss": 2.7127, "step": 429050 }, { "epoch": 3.7893639944188346, "grad_norm": 4.6316070556640625, "learning_rate": 1.3858061880284862e-05, "loss": 3.3655, "step": 429100 }, { "epoch": 3.789805542309119, "grad_norm": 2.1493678092956543, "learning_rate": 1.3848460783765937e-05, "loss": 2.5366, "step": 429150 }, { "epoch": 3.790247090199403, "grad_norm": 5.189102649688721, "learning_rate": 1.3838862479646875e-05, "loss": 2.9075, "step": 429200 }, { "epoch": 3.790688638089687, "grad_norm": 3.6914775371551514, "learning_rate": 1.3829266968669035e-05, "loss": 2.7468, "step": 429250 }, { "epoch": 3.7911301859799713, "grad_norm": 3.236041307449341, "learning_rate": 1.38196742515736e-05, "loss": 2.6258, "step": 429300 }, { "epoch": 3.7915717338702555, "grad_norm": 8.088294982910156, "learning_rate": 1.3810084329101548e-05, "loss": 3.1375, "step": 429350 }, { "epoch": 3.7920132817605396, "grad_norm": 9.265528678894043, "learning_rate": 1.3800497201993578e-05, "loss": 2.8863, "step": 429400 }, { "epoch": 3.792454829650824, "grad_norm": 5.310785293579102, "learning_rate": 1.3790912870990258e-05, "loss": 2.753, "step": 429450 }, { "epoch": 3.792896377541108, "grad_norm": 2.130755662918091, "learning_rate": 1.378133133683187e-05, "loss": 3.0432, "step": 429500 }, { "epoch": 3.793337925431392, "grad_norm": 2.8202688694000244, "learning_rate": 1.3771752600258503e-05, "loss": 2.9376, "step": 429550 }, { "epoch": 3.7937794733216763, "grad_norm": 3.0862977504730225, "learning_rate": 1.3762176662010034e-05, "loss": 2.885, "step": 429600 }, { "epoch": 3.7942210212119605, "grad_norm": 4.450771808624268, "learning_rate": 1.3752603522826136e-05, "loss": 2.9365, "step": 429650 }, { "epoch": 3.794662569102245, "grad_norm": 4.2505574226379395, "learning_rate": 1.3743033183446258e-05, "loss": 3.1064, "step": 429700 }, { "epoch": 3.795104116992529, "grad_norm": 3.089618682861328, "learning_rate": 1.37334656446096e-05, "loss": 2.8206, "step": 429750 }, { "epoch": 3.7955456648828134, "grad_norm": 4.454230785369873, "learning_rate": 1.3723900907055204e-05, "loss": 2.9205, "step": 429800 }, { "epoch": 3.795987212773097, "grad_norm": 4.342223644256592, "learning_rate": 1.3714338971521856e-05, "loss": 2.853, "step": 429850 }, { "epoch": 3.7964287606633818, "grad_norm": 4.059878349304199, "learning_rate": 1.370477983874811e-05, "loss": 2.5882, "step": 429900 }, { "epoch": 3.7968703085536655, "grad_norm": 4.5774102210998535, "learning_rate": 1.3695223509472349e-05, "loss": 3.0264, "step": 429950 }, { "epoch": 3.79731185644395, "grad_norm": 6.60628080368042, "learning_rate": 1.368566998443273e-05, "loss": 2.7415, "step": 430000 }, { "epoch": 3.7977534043342343, "grad_norm": 3.571007013320923, "learning_rate": 1.3676119264367155e-05, "loss": 2.7967, "step": 430050 }, { "epoch": 3.7981949522245184, "grad_norm": 4.905377388000488, "learning_rate": 1.3666571350013346e-05, "loss": 2.9207, "step": 430100 }, { "epoch": 3.7986365001148026, "grad_norm": 4.327259540557861, "learning_rate": 1.3657026242108806e-05, "loss": 3.1738, "step": 430150 }, { "epoch": 3.7990780480050867, "grad_norm": 2.2908060550689697, "learning_rate": 1.3647483941390821e-05, "loss": 2.5973, "step": 430200 }, { "epoch": 3.799519595895371, "grad_norm": 7.874433994293213, "learning_rate": 1.3637944448596446e-05, "loss": 2.457, "step": 430250 }, { "epoch": 3.799961143785655, "grad_norm": 3.5273520946502686, "learning_rate": 1.3628407764462497e-05, "loss": 2.646, "step": 430300 }, { "epoch": 3.8004026916759392, "grad_norm": 5.832827091217041, "learning_rate": 1.3618873889725648e-05, "loss": 2.7848, "step": 430350 }, { "epoch": 3.8008442395662234, "grad_norm": 4.790328025817871, "learning_rate": 1.3609533418870274e-05, "loss": 3.0622, "step": 430400 }, { "epoch": 3.8012857874565076, "grad_norm": 3.7390353679656982, "learning_rate": 1.3600005108911973e-05, "loss": 2.7482, "step": 430450 }, { "epoch": 3.8017273353467917, "grad_norm": 3.394366502761841, "learning_rate": 1.3590479610544632e-05, "loss": 2.62, "step": 430500 }, { "epoch": 3.802168883237076, "grad_norm": 5.427759170532227, "learning_rate": 1.3580956924503979e-05, "loss": 2.9803, "step": 430550 }, { "epoch": 3.80261043112736, "grad_norm": 1.1967991590499878, "learning_rate": 1.3571437051525598e-05, "loss": 3.0344, "step": 430600 }, { "epoch": 3.8030519790176442, "grad_norm": 4.05731201171875, "learning_rate": 1.356191999234478e-05, "loss": 2.7773, "step": 430650 }, { "epoch": 3.8034935269079284, "grad_norm": 5.387750625610352, "learning_rate": 1.3552405747696662e-05, "loss": 3.0535, "step": 430700 }, { "epoch": 3.8039350747982126, "grad_norm": 1.3500856161117554, "learning_rate": 1.3542894318316141e-05, "loss": 2.6891, "step": 430750 }, { "epoch": 3.8043766226884967, "grad_norm": 8.750868797302246, "learning_rate": 1.3533385704937879e-05, "loss": 3.1234, "step": 430800 }, { "epoch": 3.804818170578781, "grad_norm": 4.103426456451416, "learning_rate": 1.3523879908296338e-05, "loss": 2.3821, "step": 430850 }, { "epoch": 3.805259718469065, "grad_norm": 5.523512840270996, "learning_rate": 1.3514566961093233e-05, "loss": 2.8125, "step": 430900 }, { "epoch": 3.8057012663593492, "grad_norm": 5.971200942993164, "learning_rate": 1.3505066743756384e-05, "loss": 2.4457, "step": 430950 }, { "epoch": 3.8061428142496334, "grad_norm": 4.397410869598389, "learning_rate": 1.3495569345343639e-05, "loss": 2.8972, "step": 431000 }, { "epoch": 3.8065843621399176, "grad_norm": 4.07309627532959, "learning_rate": 1.3486074766588619e-05, "loss": 2.7649, "step": 431050 }, { "epoch": 3.8070259100302017, "grad_norm": 5.009177207946777, "learning_rate": 1.347658300822469e-05, "loss": 2.7709, "step": 431100 }, { "epoch": 3.807467457920486, "grad_norm": 4.602985858917236, "learning_rate": 1.346709407098503e-05, "loss": 2.6487, "step": 431150 }, { "epoch": 3.80790900581077, "grad_norm": 7.807777404785156, "learning_rate": 1.3457607955602548e-05, "loss": 3.0792, "step": 431200 }, { "epoch": 3.8083505537010547, "grad_norm": 4.218010902404785, "learning_rate": 1.3448124662809996e-05, "loss": 2.6671, "step": 431250 }, { "epoch": 3.8087921015913384, "grad_norm": 11.159671783447266, "learning_rate": 1.343864419333986e-05, "loss": 2.8996, "step": 431300 }, { "epoch": 3.809233649481623, "grad_norm": 3.074442148208618, "learning_rate": 1.3429166547924421e-05, "loss": 2.8834, "step": 431350 }, { "epoch": 3.8096751973719067, "grad_norm": 8.872673988342285, "learning_rate": 1.3419691727295751e-05, "loss": 3.2711, "step": 431400 }, { "epoch": 3.8101167452621914, "grad_norm": 1.0338062047958374, "learning_rate": 1.341021973218572e-05, "loss": 2.5041, "step": 431450 }, { "epoch": 3.810558293152475, "grad_norm": 3.0179033279418945, "learning_rate": 1.340075056332592e-05, "loss": 2.8583, "step": 431500 }, { "epoch": 3.8109998410427597, "grad_norm": 2.228316068649292, "learning_rate": 1.3391284221447792e-05, "loss": 2.6648, "step": 431550 }, { "epoch": 3.8114413889330434, "grad_norm": 2.4569506645202637, "learning_rate": 1.3381820707282544e-05, "loss": 2.5819, "step": 431600 }, { "epoch": 3.811882936823328, "grad_norm": 3.303539276123047, "learning_rate": 1.3372360021561125e-05, "loss": 3.0473, "step": 431650 }, { "epoch": 3.812324484713612, "grad_norm": 5.920217990875244, "learning_rate": 1.336290216501429e-05, "loss": 2.7465, "step": 431700 }, { "epoch": 3.8127660326038963, "grad_norm": 1.805953860282898, "learning_rate": 1.3353447138372593e-05, "loss": 2.9002, "step": 431750 }, { "epoch": 3.8132075804941805, "grad_norm": 4.679826736450195, "learning_rate": 1.3343994942366361e-05, "loss": 3.2424, "step": 431800 }, { "epoch": 3.8136491283844647, "grad_norm": 3.5181760787963867, "learning_rate": 1.3334545577725671e-05, "loss": 2.5374, "step": 431850 }, { "epoch": 3.814090676274749, "grad_norm": 5.79442024230957, "learning_rate": 1.3325099045180422e-05, "loss": 2.6356, "step": 431900 }, { "epoch": 3.814532224165033, "grad_norm": 4.071072578430176, "learning_rate": 1.3315655345460292e-05, "loss": 2.8798, "step": 431950 }, { "epoch": 3.814973772055317, "grad_norm": 2.2976200580596924, "learning_rate": 1.3306214479294714e-05, "loss": 2.5607, "step": 432000 }, { "epoch": 3.814973772055317, "eval_asr_loss": 0.9563361403972107, "eval_loss": 2.6981217861175537, "eval_runtime": 20.4851, "eval_samples_per_second": 37.491, "eval_steps_per_second": 9.373, "eval_tts_loss": 6.1907653278613575, "step": 432000 }, { "epoch": 3.8154153199456013, "grad_norm": 2.183871030807495, "learning_rate": 1.3296776447412895e-05, "loss": 2.3181, "step": 432050 }, { "epoch": 3.8158568678358855, "grad_norm": 5.35095739364624, "learning_rate": 1.3287341250543883e-05, "loss": 2.6483, "step": 432100 }, { "epoch": 3.8162984157261697, "grad_norm": 4.118880271911621, "learning_rate": 1.3277908889416423e-05, "loss": 2.9801, "step": 432150 }, { "epoch": 3.816739963616454, "grad_norm": 5.5582990646362305, "learning_rate": 1.3268479364759112e-05, "loss": 2.732, "step": 432200 }, { "epoch": 3.817181511506738, "grad_norm": 7.086887836456299, "learning_rate": 1.3259052677300293e-05, "loss": 2.7196, "step": 432250 }, { "epoch": 3.817623059397022, "grad_norm": 2.9669954776763916, "learning_rate": 1.3249628827768117e-05, "loss": 2.7537, "step": 432300 }, { "epoch": 3.8180646072873063, "grad_norm": 5.895023345947266, "learning_rate": 1.3240207816890465e-05, "loss": 2.5268, "step": 432350 }, { "epoch": 3.8185061551775905, "grad_norm": 3.246305227279663, "learning_rate": 1.3230789645395059e-05, "loss": 2.9561, "step": 432400 }, { "epoch": 3.8189477030678747, "grad_norm": 5.379619121551514, "learning_rate": 1.322137431400936e-05, "loss": 2.8996, "step": 432450 }, { "epoch": 3.819389250958159, "grad_norm": 7.105270862579346, "learning_rate": 1.3211961823460606e-05, "loss": 2.9145, "step": 432500 }, { "epoch": 3.819830798848443, "grad_norm": 4.510811805725098, "learning_rate": 1.3202552174475846e-05, "loss": 2.6344, "step": 432550 }, { "epoch": 3.820272346738727, "grad_norm": 6.596767425537109, "learning_rate": 1.3193145367781918e-05, "loss": 2.7654, "step": 432600 }, { "epoch": 3.8207138946290113, "grad_norm": 5.161827087402344, "learning_rate": 1.3183741404105376e-05, "loss": 2.7989, "step": 432650 }, { "epoch": 3.8211554425192955, "grad_norm": 6.982913970947266, "learning_rate": 1.317434028417262e-05, "loss": 2.824, "step": 432700 }, { "epoch": 3.8215969904095797, "grad_norm": 2.6162848472595215, "learning_rate": 1.3164942008709818e-05, "loss": 2.7542, "step": 432750 }, { "epoch": 3.822038538299864, "grad_norm": 3.739748001098633, "learning_rate": 1.3155546578442895e-05, "loss": 2.3083, "step": 432800 }, { "epoch": 3.822480086190148, "grad_norm": 5.4677300453186035, "learning_rate": 1.3146153994097555e-05, "loss": 2.8552, "step": 432850 }, { "epoch": 3.8229216340804326, "grad_norm": 3.799527883529663, "learning_rate": 1.3136764256399303e-05, "loss": 2.8083, "step": 432900 }, { "epoch": 3.8233631819707163, "grad_norm": 5.774009704589844, "learning_rate": 1.312737736607344e-05, "loss": 2.7149, "step": 432950 }, { "epoch": 3.823804729861001, "grad_norm": 3.211622953414917, "learning_rate": 1.311799332384499e-05, "loss": 3.2336, "step": 433000 }, { "epoch": 3.8242462777512847, "grad_norm": 6.334155082702637, "learning_rate": 1.3108612130438803e-05, "loss": 3.0338, "step": 433050 }, { "epoch": 3.8246878256415693, "grad_norm": 5.2527852058410645, "learning_rate": 1.3099233786579513e-05, "loss": 2.8303, "step": 433100 }, { "epoch": 3.825129373531853, "grad_norm": 5.200671672821045, "learning_rate": 1.3090045774925925e-05, "loss": 2.6708, "step": 433150 }, { "epoch": 3.8255709214221376, "grad_norm": 2.686323642730713, "learning_rate": 1.3080673075306355e-05, "loss": 2.9068, "step": 433200 }, { "epoch": 3.826012469312422, "grad_norm": 3.8129384517669678, "learning_rate": 1.3071303227391724e-05, "loss": 2.6866, "step": 433250 }, { "epoch": 3.826454017202706, "grad_norm": 4.414145469665527, "learning_rate": 1.3061936231905776e-05, "loss": 2.8789, "step": 433300 }, { "epoch": 3.82689556509299, "grad_norm": 3.7200002670288086, "learning_rate": 1.3052572089572034e-05, "loss": 2.8788, "step": 433350 }, { "epoch": 3.8273371129832743, "grad_norm": 5.481420516967773, "learning_rate": 1.3043210801113787e-05, "loss": 2.7389, "step": 433400 }, { "epoch": 3.8277786608735584, "grad_norm": 4.415096759796143, "learning_rate": 1.3033852367254129e-05, "loss": 3.2715, "step": 433450 }, { "epoch": 3.8282202087638426, "grad_norm": 2.968588352203369, "learning_rate": 1.30244967887159e-05, "loss": 2.5945, "step": 433500 }, { "epoch": 3.828661756654127, "grad_norm": 8.526740074157715, "learning_rate": 1.3015144066221768e-05, "loss": 2.8586, "step": 433550 }, { "epoch": 3.829103304544411, "grad_norm": 1.660351037979126, "learning_rate": 1.300579420049412e-05, "loss": 2.8065, "step": 433600 }, { "epoch": 3.829544852434695, "grad_norm": 1.6464658975601196, "learning_rate": 1.299644719225518e-05, "loss": 2.9696, "step": 433650 }, { "epoch": 3.8299864003249793, "grad_norm": 4.642117500305176, "learning_rate": 1.2987103042226901e-05, "loss": 3.1059, "step": 433700 }, { "epoch": 3.8304279482152634, "grad_norm": 13.414894104003906, "learning_rate": 1.2977761751131052e-05, "loss": 3.0508, "step": 433750 }, { "epoch": 3.8308694961055476, "grad_norm": 2.3988916873931885, "learning_rate": 1.2968423319689166e-05, "loss": 3.1035, "step": 433800 }, { "epoch": 3.8313110439958318, "grad_norm": 6.618754863739014, "learning_rate": 1.2959087748622573e-05, "loss": 3.391, "step": 433850 }, { "epoch": 3.831752591886116, "grad_norm": 2.678011894226074, "learning_rate": 1.2949755038652356e-05, "loss": 3.1505, "step": 433900 }, { "epoch": 3.8321941397764, "grad_norm": 5.174765110015869, "learning_rate": 1.2940425190499372e-05, "loss": 2.5377, "step": 433950 }, { "epoch": 3.8326356876666843, "grad_norm": 3.774174690246582, "learning_rate": 1.29310982048843e-05, "loss": 3.4962, "step": 434000 }, { "epoch": 3.8330772355569684, "grad_norm": 5.134815692901611, "learning_rate": 1.2921774082527544e-05, "loss": 2.515, "step": 434050 }, { "epoch": 3.8335187834472526, "grad_norm": 2.4276669025421143, "learning_rate": 1.2912452824149323e-05, "loss": 2.8319, "step": 434100 }, { "epoch": 3.8339603313375368, "grad_norm": 4.2926025390625, "learning_rate": 1.2903134430469643e-05, "loss": 2.9328, "step": 434150 }, { "epoch": 3.834401879227821, "grad_norm": 2.457986354827881, "learning_rate": 1.2893818902208244e-05, "loss": 2.9887, "step": 434200 }, { "epoch": 3.834843427118105, "grad_norm": 5.6597981452941895, "learning_rate": 1.2884506240084692e-05, "loss": 2.7089, "step": 434250 }, { "epoch": 3.8352849750083893, "grad_norm": 3.7390637397766113, "learning_rate": 1.2875196444818311e-05, "loss": 2.7854, "step": 434300 }, { "epoch": 3.8357265228986734, "grad_norm": 5.216220378875732, "learning_rate": 1.2865889517128171e-05, "loss": 2.7931, "step": 434350 }, { "epoch": 3.8361680707889576, "grad_norm": 1.8534170389175415, "learning_rate": 1.2856585457733183e-05, "loss": 3.4755, "step": 434400 }, { "epoch": 3.836609618679242, "grad_norm": 7.221693992614746, "learning_rate": 1.2847284267352005e-05, "loss": 2.9402, "step": 434450 }, { "epoch": 3.837051166569526, "grad_norm": 7.041598320007324, "learning_rate": 1.2837985946703084e-05, "loss": 3.1639, "step": 434500 }, { "epoch": 3.8374927144598105, "grad_norm": 6.911594390869141, "learning_rate": 1.2828690496504609e-05, "loss": 2.8936, "step": 434550 }, { "epoch": 3.8379342623500943, "grad_norm": 3.0775558948516846, "learning_rate": 1.2819397917474585e-05, "loss": 3.0993, "step": 434600 }, { "epoch": 3.838375810240379, "grad_norm": 4.821956157684326, "learning_rate": 1.2810108210330812e-05, "loss": 2.6533, "step": 434650 }, { "epoch": 3.8388173581306626, "grad_norm": 7.707732200622559, "learning_rate": 1.280082137579081e-05, "loss": 2.9689, "step": 434700 }, { "epoch": 3.839258906020947, "grad_norm": 6.071098327636719, "learning_rate": 1.2791537414571908e-05, "loss": 2.9258, "step": 434750 }, { "epoch": 3.839700453911231, "grad_norm": 6.257697105407715, "learning_rate": 1.278225632739124e-05, "loss": 2.9402, "step": 434800 }, { "epoch": 3.8401420018015155, "grad_norm": 4.9607768058776855, "learning_rate": 1.277297811496565e-05, "loss": 3.2447, "step": 434850 }, { "epoch": 3.8405835496917997, "grad_norm": 5.588959217071533, "learning_rate": 1.2763702778011832e-05, "loss": 2.6578, "step": 434900 }, { "epoch": 3.841025097582084, "grad_norm": 4.646946430206299, "learning_rate": 1.2754430317246219e-05, "loss": 2.9725, "step": 434950 }, { "epoch": 3.841466645472368, "grad_norm": 6.796315670013428, "learning_rate": 1.2745160733385043e-05, "loss": 2.5158, "step": 435000 }, { "epoch": 3.841466645472368, "eval_asr_loss": 0.9529584955565111, "eval_loss": 2.692127227783203, "eval_runtime": 20.5673, "eval_samples_per_second": 37.341, "eval_steps_per_second": 9.335, "eval_tts_loss": 6.172081323549535, "step": 435000 }, { "epoch": 3.841908193362652, "grad_norm": 5.4761176109313965, "learning_rate": 1.2735894027144285e-05, "loss": 2.4114, "step": 435050 }, { "epoch": 3.8423497412529364, "grad_norm": 9.474357604980469, "learning_rate": 1.2726630199239715e-05, "loss": 2.7392, "step": 435100 }, { "epoch": 3.8427912891432205, "grad_norm": 3.846484422683716, "learning_rate": 1.2717369250386906e-05, "loss": 2.5246, "step": 435150 }, { "epoch": 3.8432328370335047, "grad_norm": 2.94089674949646, "learning_rate": 1.2708111181301158e-05, "loss": 3.405, "step": 435200 }, { "epoch": 3.843674384923789, "grad_norm": 5.778763294219971, "learning_rate": 1.26988559926976e-05, "loss": 3.1552, "step": 435250 }, { "epoch": 3.844115932814073, "grad_norm": 5.492156982421875, "learning_rate": 1.2689603685291125e-05, "loss": 2.7537, "step": 435300 }, { "epoch": 3.844557480704357, "grad_norm": 7.57138204574585, "learning_rate": 1.2680354259796363e-05, "loss": 3.0029, "step": 435350 }, { "epoch": 3.8449990285946414, "grad_norm": 2.8455264568328857, "learning_rate": 1.2671107716927777e-05, "loss": 3.1149, "step": 435400 }, { "epoch": 3.8454405764849255, "grad_norm": 1.8032853603363037, "learning_rate": 1.2661864057399592e-05, "loss": 3.5842, "step": 435450 }, { "epoch": 3.8458821243752097, "grad_norm": 3.3582136631011963, "learning_rate": 1.2652623281925797e-05, "loss": 2.5827, "step": 435500 }, { "epoch": 3.846323672265494, "grad_norm": 2.3578877449035645, "learning_rate": 1.2643385391220136e-05, "loss": 3.0355, "step": 435550 }, { "epoch": 3.846765220155778, "grad_norm": 4.568357467651367, "learning_rate": 1.263415038599618e-05, "loss": 2.9127, "step": 435600 }, { "epoch": 3.847206768046062, "grad_norm": 4.637597560882568, "learning_rate": 1.2624918266967268e-05, "loss": 2.9235, "step": 435650 }, { "epoch": 3.8476483159363464, "grad_norm": 2.3217742443084717, "learning_rate": 1.2615689034846473e-05, "loss": 2.7962, "step": 435700 }, { "epoch": 3.8480898638266305, "grad_norm": 3.528264284133911, "learning_rate": 1.2606462690346693e-05, "loss": 2.9848, "step": 435750 }, { "epoch": 3.8485314117169147, "grad_norm": 6.982277870178223, "learning_rate": 1.2597239234180597e-05, "loss": 3.0021, "step": 435800 }, { "epoch": 3.848972959607199, "grad_norm": 4.666854381561279, "learning_rate": 1.2588018667060602e-05, "loss": 2.9157, "step": 435850 }, { "epoch": 3.849414507497483, "grad_norm": 3.626660108566284, "learning_rate": 1.2578800989698907e-05, "loss": 2.7393, "step": 435900 }, { "epoch": 3.849856055387767, "grad_norm": 5.902881622314453, "learning_rate": 1.2569586202807527e-05, "loss": 3.1007, "step": 435950 }, { "epoch": 3.8502976032780514, "grad_norm": 3.1990511417388916, "learning_rate": 1.25603743070982e-05, "loss": 2.5299, "step": 436000 }, { "epoch": 3.8507391511683355, "grad_norm": 3.970501661300659, "learning_rate": 1.2551165303282475e-05, "loss": 2.7468, "step": 436050 }, { "epoch": 3.85118069905862, "grad_norm": 2.494051694869995, "learning_rate": 1.2541959192071679e-05, "loss": 3.039, "step": 436100 }, { "epoch": 3.851622246948904, "grad_norm": 2.4153008460998535, "learning_rate": 1.2532755974176912e-05, "loss": 3.3924, "step": 436150 }, { "epoch": 3.8520637948391885, "grad_norm": 3.4217681884765625, "learning_rate": 1.2523555650309032e-05, "loss": 2.8461, "step": 436200 }, { "epoch": 3.852505342729472, "grad_norm": 3.1231775283813477, "learning_rate": 1.2514358221178673e-05, "loss": 2.5534, "step": 436250 }, { "epoch": 3.852946890619757, "grad_norm": 4.058957576751709, "learning_rate": 1.2505163687496284e-05, "loss": 3.2812, "step": 436300 }, { "epoch": 3.8533884385100405, "grad_norm": 5.170287609100342, "learning_rate": 1.2495972049972044e-05, "loss": 2.618, "step": 436350 }, { "epoch": 3.853829986400325, "grad_norm": 2.3591742515563965, "learning_rate": 1.2486783309315936e-05, "loss": 2.7549, "step": 436400 }, { "epoch": 3.8542715342906093, "grad_norm": 7.680919170379639, "learning_rate": 1.2477597466237728e-05, "loss": 2.3167, "step": 436450 }, { "epoch": 3.8547130821808935, "grad_norm": 3.7799181938171387, "learning_rate": 1.2468414521446925e-05, "loss": 3.0093, "step": 436500 }, { "epoch": 3.8551546300711776, "grad_norm": 2.994607448577881, "learning_rate": 1.2459234475652836e-05, "loss": 2.5121, "step": 436550 }, { "epoch": 3.855596177961462, "grad_norm": 5.4545135498046875, "learning_rate": 1.2450057329564569e-05, "loss": 2.7016, "step": 436600 }, { "epoch": 3.856037725851746, "grad_norm": 2.138357400894165, "learning_rate": 1.2440883083890953e-05, "loss": 2.8015, "step": 436650 }, { "epoch": 3.85647927374203, "grad_norm": 6.133525371551514, "learning_rate": 1.2431711739340613e-05, "loss": 2.8982, "step": 436700 }, { "epoch": 3.8569208216323143, "grad_norm": 6.497407913208008, "learning_rate": 1.2422543296621974e-05, "loss": 2.727, "step": 436750 }, { "epoch": 3.8573623695225985, "grad_norm": 4.703451633453369, "learning_rate": 1.2413377756443228e-05, "loss": 2.7992, "step": 436800 }, { "epoch": 3.8578039174128826, "grad_norm": 7.472330570220947, "learning_rate": 1.2404215119512314e-05, "loss": 2.2483, "step": 436850 }, { "epoch": 3.858245465303167, "grad_norm": 5.041032314300537, "learning_rate": 1.239505538653698e-05, "loss": 3.1321, "step": 436900 }, { "epoch": 3.858687013193451, "grad_norm": 5.955416679382324, "learning_rate": 1.2385898558224745e-05, "loss": 2.7928, "step": 436950 }, { "epoch": 3.859128561083735, "grad_norm": 1.8457165956497192, "learning_rate": 1.2376744635282894e-05, "loss": 2.2203, "step": 437000 }, { "epoch": 3.8595701089740193, "grad_norm": 3.5115082263946533, "learning_rate": 1.2367593618418466e-05, "loss": 2.9639, "step": 437050 }, { "epoch": 3.8600116568643035, "grad_norm": 2.675835371017456, "learning_rate": 1.2358445508338334e-05, "loss": 2.7713, "step": 437100 }, { "epoch": 3.8604532047545876, "grad_norm": 3.8960089683532715, "learning_rate": 1.2349300305749084e-05, "loss": 2.3965, "step": 437150 }, { "epoch": 3.860894752644872, "grad_norm": 3.502213954925537, "learning_rate": 1.2340158011357112e-05, "loss": 2.4697, "step": 437200 }, { "epoch": 3.861336300535156, "grad_norm": 2.238583564758301, "learning_rate": 1.2331018625868606e-05, "loss": 2.3292, "step": 437250 }, { "epoch": 3.86177784842544, "grad_norm": 1.875502347946167, "learning_rate": 1.2321882149989473e-05, "loss": 2.9757, "step": 437300 }, { "epoch": 3.8622193963157243, "grad_norm": 4.304465293884277, "learning_rate": 1.2312748584425438e-05, "loss": 2.7904, "step": 437350 }, { "epoch": 3.8626609442060085, "grad_norm": 3.2962117195129395, "learning_rate": 1.230361792988201e-05, "loss": 2.967, "step": 437400 }, { "epoch": 3.8631024920962926, "grad_norm": 2.4963765144348145, "learning_rate": 1.2294490187064445e-05, "loss": 3.1074, "step": 437450 }, { "epoch": 3.863544039986577, "grad_norm": 7.147805690765381, "learning_rate": 1.2285365356677763e-05, "loss": 2.7946, "step": 437500 }, { "epoch": 3.863985587876861, "grad_norm": 7.056324005126953, "learning_rate": 1.2276243439426799e-05, "loss": 2.6496, "step": 437550 }, { "epoch": 3.864427135767145, "grad_norm": 3.574471950531006, "learning_rate": 1.2267124436016158e-05, "loss": 2.7841, "step": 437600 }, { "epoch": 3.8648686836574297, "grad_norm": 3.3816187381744385, "learning_rate": 1.2258008347150169e-05, "loss": 2.8896, "step": 437650 }, { "epoch": 3.8653102315477135, "grad_norm": 2.0144495964050293, "learning_rate": 1.2248895173532992e-05, "loss": 2.59, "step": 437700 }, { "epoch": 3.865751779437998, "grad_norm": 3.1916444301605225, "learning_rate": 1.2239784915868563e-05, "loss": 2.834, "step": 437750 }, { "epoch": 3.866193327328282, "grad_norm": 6.203403472900391, "learning_rate": 1.2230859693092916e-05, "loss": 2.7461, "step": 437800 }, { "epoch": 3.8666348752185664, "grad_norm": 7.3679351806640625, "learning_rate": 1.222175521109069e-05, "loss": 2.4815, "step": 437850 }, { "epoch": 3.86707642310885, "grad_norm": 5.298933506011963, "learning_rate": 1.2212653647137518e-05, "loss": 3.0728, "step": 437900 }, { "epoch": 3.8675179709991347, "grad_norm": 7.32708215713501, "learning_rate": 1.2203555001936417e-05, "loss": 2.6672, "step": 437950 }, { "epoch": 3.867959518889419, "grad_norm": 5.106059551239014, "learning_rate": 1.2194459276190206e-05, "loss": 2.6493, "step": 438000 }, { "epoch": 3.867959518889419, "eval_asr_loss": 0.9576269639080421, "eval_loss": 2.6886203289031982, "eval_runtime": 20.4274, "eval_samples_per_second": 37.597, "eval_steps_per_second": 9.399, "eval_tts_loss": 6.189523725782818, "step": 438000 }, { "epoch": 3.868401066779703, "grad_norm": 3.8552653789520264, "learning_rate": 1.2185366470601423e-05, "loss": 3.0385, "step": 438050 }, { "epoch": 3.8688426146699872, "grad_norm": 5.816058158874512, "learning_rate": 1.2176276585872432e-05, "loss": 3.3323, "step": 438100 }, { "epoch": 3.8692841625602714, "grad_norm": 3.1296091079711914, "learning_rate": 1.2167189622705339e-05, "loss": 2.8713, "step": 438150 }, { "epoch": 3.8697257104505556, "grad_norm": 4.631579399108887, "learning_rate": 1.2158105581802015e-05, "loss": 2.747, "step": 438200 }, { "epoch": 3.8701672583408397, "grad_norm": 4.514932632446289, "learning_rate": 1.214902446386415e-05, "loss": 2.6377, "step": 438250 }, { "epoch": 3.870608806231124, "grad_norm": 5.922295093536377, "learning_rate": 1.2139946269593177e-05, "loss": 2.6134, "step": 438300 }, { "epoch": 3.871050354121408, "grad_norm": 7.635079860687256, "learning_rate": 1.213087099969033e-05, "loss": 2.6772, "step": 438350 }, { "epoch": 3.8714919020116922, "grad_norm": 1.661571979522705, "learning_rate": 1.2121798654856558e-05, "loss": 2.5956, "step": 438400 }, { "epoch": 3.8719334499019764, "grad_norm": 7.465584754943848, "learning_rate": 1.2112729235792659e-05, "loss": 3.0667, "step": 438450 }, { "epoch": 3.8723749977922606, "grad_norm": 6.469088077545166, "learning_rate": 1.2103662743199151e-05, "loss": 2.6368, "step": 438500 }, { "epoch": 3.8728165456825447, "grad_norm": 4.713489055633545, "learning_rate": 1.2094599177776334e-05, "loss": 3.0521, "step": 438550 }, { "epoch": 3.873258093572829, "grad_norm": 5.777842998504639, "learning_rate": 1.2085538540224295e-05, "loss": 3.016, "step": 438600 }, { "epoch": 3.873699641463113, "grad_norm": 2.5395779609680176, "learning_rate": 1.2076480831242914e-05, "loss": 2.7141, "step": 438650 }, { "epoch": 3.8741411893533972, "grad_norm": 5.247798442840576, "learning_rate": 1.2067426051531799e-05, "loss": 2.9743, "step": 438700 }, { "epoch": 3.8745827372436814, "grad_norm": 2.9409501552581787, "learning_rate": 1.2058374201790357e-05, "loss": 3.2222, "step": 438750 }, { "epoch": 3.8750242851339656, "grad_norm": 4.765002250671387, "learning_rate": 1.2049325282717788e-05, "loss": 2.9725, "step": 438800 }, { "epoch": 3.8754658330242497, "grad_norm": 4.121987819671631, "learning_rate": 1.2040279295013014e-05, "loss": 3.0461, "step": 438850 }, { "epoch": 3.875907380914534, "grad_norm": 3.855717182159424, "learning_rate": 1.2031236239374788e-05, "loss": 3.0044, "step": 438900 }, { "epoch": 3.876348928804818, "grad_norm": 2.334341287612915, "learning_rate": 1.2022196116501588e-05, "loss": 2.9147, "step": 438950 }, { "epoch": 3.8767904766951022, "grad_norm": 5.616029739379883, "learning_rate": 1.2013158927091705e-05, "loss": 2.6701, "step": 439000 }, { "epoch": 3.8772320245853864, "grad_norm": 7.148241996765137, "learning_rate": 1.2004124671843164e-05, "loss": 2.9246, "step": 439050 }, { "epoch": 3.8776735724756706, "grad_norm": 4.951962471008301, "learning_rate": 1.1995093351453796e-05, "loss": 3.0571, "step": 439100 }, { "epoch": 3.8781151203659547, "grad_norm": 4.408689975738525, "learning_rate": 1.1986064966621208e-05, "loss": 2.7018, "step": 439150 }, { "epoch": 3.8785566682562393, "grad_norm": 8.102865219116211, "learning_rate": 1.1977039518042737e-05, "loss": 2.7438, "step": 439200 }, { "epoch": 3.878998216146523, "grad_norm": 8.55119800567627, "learning_rate": 1.1968017006415556e-05, "loss": 2.6221, "step": 439250 }, { "epoch": 3.8794397640368077, "grad_norm": 5.507638454437256, "learning_rate": 1.1958997432436553e-05, "loss": 2.4185, "step": 439300 }, { "epoch": 3.8798813119270914, "grad_norm": 5.4817986488342285, "learning_rate": 1.1950161100714797e-05, "loss": 3.1632, "step": 439350 }, { "epoch": 3.880322859817376, "grad_norm": 8.991950035095215, "learning_rate": 1.1941147345334347e-05, "loss": 2.4749, "step": 439400 }, { "epoch": 3.8807644077076597, "grad_norm": 8.44919490814209, "learning_rate": 1.1932136529677517e-05, "loss": 2.8544, "step": 439450 }, { "epoch": 3.8812059555979443, "grad_norm": 3.808727502822876, "learning_rate": 1.1923128654440341e-05, "loss": 3.06, "step": 439500 }, { "epoch": 3.881647503488228, "grad_norm": 4.12844181060791, "learning_rate": 1.1914123720318603e-05, "loss": 3.1016, "step": 439550 }, { "epoch": 3.8820890513785127, "grad_norm": 4.365732669830322, "learning_rate": 1.190512172800785e-05, "loss": 3.0793, "step": 439600 }, { "epoch": 3.882530599268797, "grad_norm": 1.7833396196365356, "learning_rate": 1.1896122678203397e-05, "loss": 3.1545, "step": 439650 }, { "epoch": 3.882972147159081, "grad_norm": 5.123521327972412, "learning_rate": 1.1887126571600365e-05, "loss": 2.8089, "step": 439700 }, { "epoch": 3.883413695049365, "grad_norm": 6.405866622924805, "learning_rate": 1.1878133408893604e-05, "loss": 2.7689, "step": 439750 }, { "epoch": 3.8838552429396493, "grad_norm": 11.434934616088867, "learning_rate": 1.1869143190777777e-05, "loss": 2.7277, "step": 439800 }, { "epoch": 3.8842967908299335, "grad_norm": 2.8352489471435547, "learning_rate": 1.1860155917947314e-05, "loss": 2.9574, "step": 439850 }, { "epoch": 3.8847383387202177, "grad_norm": 3.179097890853882, "learning_rate": 1.1851171591096372e-05, "loss": 2.91, "step": 439900 }, { "epoch": 3.885179886610502, "grad_norm": 5.504298686981201, "learning_rate": 1.1842190210918942e-05, "loss": 2.8622, "step": 439950 }, { "epoch": 3.885621434500786, "grad_norm": 3.5288093090057373, "learning_rate": 1.1833211778108744e-05, "loss": 2.9065, "step": 440000 }, { "epoch": 3.88606298239107, "grad_norm": 4.040787696838379, "learning_rate": 1.1824236293359298e-05, "loss": 3.3309, "step": 440050 }, { "epoch": 3.8865045302813543, "grad_norm": 2.952626943588257, "learning_rate": 1.1815263757363865e-05, "loss": 2.8929, "step": 440100 }, { "epoch": 3.8869460781716385, "grad_norm": 5.4133687019348145, "learning_rate": 1.1806294170815513e-05, "loss": 2.5585, "step": 440150 }, { "epoch": 3.8873876260619227, "grad_norm": 4.860975742340088, "learning_rate": 1.1797327534407071e-05, "loss": 2.9146, "step": 440200 }, { "epoch": 3.887829173952207, "grad_norm": 3.4424824714660645, "learning_rate": 1.178836384883112e-05, "loss": 2.4922, "step": 440250 }, { "epoch": 3.888270721842491, "grad_norm": 2.0444960594177246, "learning_rate": 1.1779403114780035e-05, "loss": 2.7109, "step": 440300 }, { "epoch": 3.888712269732775, "grad_norm": 2.6699390411376953, "learning_rate": 1.1770445332945973e-05, "loss": 2.7594, "step": 440350 }, { "epoch": 3.8891538176230593, "grad_norm": 5.182240962982178, "learning_rate": 1.1761490504020828e-05, "loss": 3.1012, "step": 440400 }, { "epoch": 3.8895953655133435, "grad_norm": 2.8354909420013428, "learning_rate": 1.1752538628696275e-05, "loss": 2.8271, "step": 440450 }, { "epoch": 3.8900369134036277, "grad_norm": 4.689704895019531, "learning_rate": 1.1743589707663788e-05, "loss": 3.2337, "step": 440500 }, { "epoch": 3.890478461293912, "grad_norm": 3.7308061122894287, "learning_rate": 1.1734643741614603e-05, "loss": 2.722, "step": 440550 }, { "epoch": 3.890920009184196, "grad_norm": 3.144987106323242, "learning_rate": 1.1725700731239691e-05, "loss": 3.1282, "step": 440600 }, { "epoch": 3.89136155707448, "grad_norm": 5.900923728942871, "learning_rate": 1.1716760677229848e-05, "loss": 2.5166, "step": 440650 }, { "epoch": 3.8918031049647643, "grad_norm": 8.219340324401855, "learning_rate": 1.1707823580275617e-05, "loss": 2.9905, "step": 440700 }, { "epoch": 3.8922446528550485, "grad_norm": 3.0369784832000732, "learning_rate": 1.1698889441067312e-05, "loss": 2.9265, "step": 440750 }, { "epoch": 3.8926862007453327, "grad_norm": 3.4435131549835205, "learning_rate": 1.1689958260294997e-05, "loss": 2.8023, "step": 440800 }, { "epoch": 3.8931277486356173, "grad_norm": 1.947364330291748, "learning_rate": 1.168103003864856e-05, "loss": 2.7876, "step": 440850 }, { "epoch": 3.893569296525901, "grad_norm": 4.536030292510986, "learning_rate": 1.1672104776817604e-05, "loss": 2.9663, "step": 440900 }, { "epoch": 3.8940108444161856, "grad_norm": 3.8164448738098145, "learning_rate": 1.1663182475491546e-05, "loss": 2.9864, "step": 440950 }, { "epoch": 3.8944523923064693, "grad_norm": 5.855549335479736, "learning_rate": 1.1654263135359566e-05, "loss": 2.7239, "step": 441000 }, { "epoch": 3.8944523923064693, "eval_asr_loss": 0.9630484796681998, "eval_loss": 2.688081979751587, "eval_runtime": 24.7554, "eval_samples_per_second": 31.024, "eval_steps_per_second": 7.756, "eval_tts_loss": 6.203920757186442, "step": 441000 }, { "epoch": 3.894893940196754, "grad_norm": 2.522629976272583, "learning_rate": 1.164534675711058e-05, "loss": 2.9328, "step": 441050 }, { "epoch": 3.8953354880870377, "grad_norm": 1.8033989667892456, "learning_rate": 1.1636433341433322e-05, "loss": 2.8339, "step": 441100 }, { "epoch": 3.8957770359773223, "grad_norm": 2.8624250888824463, "learning_rate": 1.162752288901629e-05, "loss": 2.8745, "step": 441150 }, { "epoch": 3.8962185838676064, "grad_norm": 6.989434719085693, "learning_rate": 1.1618615400547722e-05, "loss": 3.1125, "step": 441200 }, { "epoch": 3.8966601317578906, "grad_norm": 7.745509624481201, "learning_rate": 1.160971087671564e-05, "loss": 2.8145, "step": 441250 }, { "epoch": 3.8971016796481748, "grad_norm": 8.2193603515625, "learning_rate": 1.1600809318207851e-05, "loss": 2.806, "step": 441300 }, { "epoch": 3.897543227538459, "grad_norm": 6.870453834533691, "learning_rate": 1.1591910725711941e-05, "loss": 2.9023, "step": 441350 }, { "epoch": 3.897984775428743, "grad_norm": 5.455166339874268, "learning_rate": 1.1583015099915217e-05, "loss": 2.686, "step": 441400 }, { "epoch": 3.8984263233190273, "grad_norm": 6.8563642501831055, "learning_rate": 1.1574122441504814e-05, "loss": 3.0827, "step": 441450 }, { "epoch": 3.8988678712093114, "grad_norm": 0.5272431373596191, "learning_rate": 1.1565232751167627e-05, "loss": 2.9955, "step": 441500 }, { "epoch": 3.8993094190995956, "grad_norm": 8.371082305908203, "learning_rate": 1.1556346029590292e-05, "loss": 2.6741, "step": 441550 }, { "epoch": 3.8997509669898798, "grad_norm": 3.8077473640441895, "learning_rate": 1.1547462277459215e-05, "loss": 3.0573, "step": 441600 }, { "epoch": 3.900192514880164, "grad_norm": 4.168478488922119, "learning_rate": 1.1538581495460626e-05, "loss": 2.733, "step": 441650 }, { "epoch": 3.900634062770448, "grad_norm": 7.5213236808776855, "learning_rate": 1.1529703684280463e-05, "loss": 3.0158, "step": 441700 }, { "epoch": 3.9010756106607323, "grad_norm": 4.7362961769104, "learning_rate": 1.1520828844604465e-05, "loss": 2.7044, "step": 441750 }, { "epoch": 3.9015171585510164, "grad_norm": 5.498657703399658, "learning_rate": 1.1511956977118149e-05, "loss": 2.7556, "step": 441800 }, { "epoch": 3.9019587064413006, "grad_norm": 4.002735137939453, "learning_rate": 1.1503088082506802e-05, "loss": 2.4112, "step": 441850 }, { "epoch": 3.9024002543315848, "grad_norm": 2.918435573577881, "learning_rate": 1.1494222161455453e-05, "loss": 2.7306, "step": 441900 }, { "epoch": 3.902841802221869, "grad_norm": 7.783379554748535, "learning_rate": 1.1485359214648911e-05, "loss": 2.8468, "step": 441950 }, { "epoch": 3.903283350112153, "grad_norm": 4.291662216186523, "learning_rate": 1.1476499242771787e-05, "loss": 2.9008, "step": 442000 }, { "epoch": 3.9037248980024373, "grad_norm": 3.8218142986297607, "learning_rate": 1.1467642246508414e-05, "loss": 2.8289, "step": 442050 }, { "epoch": 3.9041664458927214, "grad_norm": 5.094839096069336, "learning_rate": 1.1458788226542933e-05, "loss": 3.0093, "step": 442100 }, { "epoch": 3.9046079937830056, "grad_norm": 5.727198600769043, "learning_rate": 1.1449937183559256e-05, "loss": 2.7128, "step": 442150 }, { "epoch": 3.9050495416732898, "grad_norm": 4.3527984619140625, "learning_rate": 1.1441089118241022e-05, "loss": 3.0752, "step": 442200 }, { "epoch": 3.905491089563574, "grad_norm": 5.881237983703613, "learning_rate": 1.1432244031271683e-05, "loss": 3.2646, "step": 442250 }, { "epoch": 3.905932637453858, "grad_norm": 3.8652050495147705, "learning_rate": 1.1423401923334465e-05, "loss": 2.9405, "step": 442300 }, { "epoch": 3.9063741853441423, "grad_norm": 6.451396465301514, "learning_rate": 1.1414562795112327e-05, "loss": 3.0782, "step": 442350 }, { "epoch": 3.906815733234427, "grad_norm": 4.235085964202881, "learning_rate": 1.1405726647288e-05, "loss": 2.5839, "step": 442400 }, { "epoch": 3.9072572811247106, "grad_norm": 3.927668809890747, "learning_rate": 1.1396893480544024e-05, "loss": 2.8537, "step": 442450 }, { "epoch": 3.907698829014995, "grad_norm": 5.062263011932373, "learning_rate": 1.1388063295562695e-05, "loss": 2.5441, "step": 442500 }, { "epoch": 3.908140376905279, "grad_norm": 1.789161205291748, "learning_rate": 1.1379236093026042e-05, "loss": 2.6742, "step": 442550 }, { "epoch": 3.9085819247955635, "grad_norm": 4.315136432647705, "learning_rate": 1.1370411873615905e-05, "loss": 2.6945, "step": 442600 }, { "epoch": 3.9090234726858473, "grad_norm": 2.9583187103271484, "learning_rate": 1.1361590638013897e-05, "loss": 3.1419, "step": 442650 }, { "epoch": 3.909465020576132, "grad_norm": 2.5440735816955566, "learning_rate": 1.1352772386901366e-05, "loss": 2.5878, "step": 442700 }, { "epoch": 3.9099065684664156, "grad_norm": 3.1499722003936768, "learning_rate": 1.1343957120959436e-05, "loss": 2.9176, "step": 442750 }, { "epoch": 3.9103481163567, "grad_norm": 3.7518956661224365, "learning_rate": 1.1335144840869044e-05, "loss": 2.8133, "step": 442800 }, { "epoch": 3.9107896642469844, "grad_norm": 2.639434814453125, "learning_rate": 1.1326335547310823e-05, "loss": 2.4941, "step": 442850 }, { "epoch": 3.9112312121372685, "grad_norm": 5.117804527282715, "learning_rate": 1.1317529240965241e-05, "loss": 2.4799, "step": 442900 }, { "epoch": 3.9116727600275527, "grad_norm": 3.253807544708252, "learning_rate": 1.1308725922512508e-05, "loss": 2.8403, "step": 442950 }, { "epoch": 3.912114307917837, "grad_norm": 2.7868118286132812, "learning_rate": 1.129992559263262e-05, "loss": 2.7562, "step": 443000 }, { "epoch": 3.912555855808121, "grad_norm": 3.611293077468872, "learning_rate": 1.1291128252005301e-05, "loss": 2.9371, "step": 443050 }, { "epoch": 3.912997403698405, "grad_norm": 5.223772048950195, "learning_rate": 1.1282333901310087e-05, "loss": 2.7549, "step": 443100 }, { "epoch": 3.9134389515886894, "grad_norm": 7.057180404663086, "learning_rate": 1.1273542541226274e-05, "loss": 2.5826, "step": 443150 }, { "epoch": 3.9138804994789735, "grad_norm": 2.961717128753662, "learning_rate": 1.126475417243289e-05, "loss": 2.794, "step": 443200 }, { "epoch": 3.9143220473692577, "grad_norm": 3.991865396499634, "learning_rate": 1.1255968795608779e-05, "loss": 2.7049, "step": 443250 }, { "epoch": 3.914763595259542, "grad_norm": 6.792287826538086, "learning_rate": 1.1247186411432559e-05, "loss": 2.5649, "step": 443300 }, { "epoch": 3.915205143149826, "grad_norm": 1.997275948524475, "learning_rate": 1.1238407020582553e-05, "loss": 2.7794, "step": 443350 }, { "epoch": 3.91564669104011, "grad_norm": 3.8345134258270264, "learning_rate": 1.122963062373692e-05, "loss": 2.9914, "step": 443400 }, { "epoch": 3.9160882389303944, "grad_norm": 3.453803062438965, "learning_rate": 1.1220857221573572e-05, "loss": 2.5434, "step": 443450 }, { "epoch": 3.9165297868206785, "grad_norm": 6.852500915527344, "learning_rate": 1.1212086814770168e-05, "loss": 2.9561, "step": 443500 }, { "epoch": 3.9169713347109627, "grad_norm": 10.04887580871582, "learning_rate": 1.1203319404004126e-05, "loss": 2.4768, "step": 443550 }, { "epoch": 3.917412882601247, "grad_norm": 5.113874912261963, "learning_rate": 1.1194554989952676e-05, "loss": 2.7375, "step": 443600 }, { "epoch": 3.917854430491531, "grad_norm": 5.647159576416016, "learning_rate": 1.1185793573292808e-05, "loss": 3.0355, "step": 443650 }, { "epoch": 3.918295978381815, "grad_norm": 5.074594020843506, "learning_rate": 1.1177035154701238e-05, "loss": 2.9533, "step": 443700 }, { "epoch": 3.9187375262720994, "grad_norm": 6.4190521240234375, "learning_rate": 1.1168279734854493e-05, "loss": 2.9584, "step": 443750 }, { "epoch": 3.9191790741623835, "grad_norm": 3.5271151065826416, "learning_rate": 1.115952731442887e-05, "loss": 2.7434, "step": 443800 }, { "epoch": 3.9196206220526677, "grad_norm": 6.215015411376953, "learning_rate": 1.115077789410039e-05, "loss": 2.6689, "step": 443850 }, { "epoch": 3.920062169942952, "grad_norm": 5.968278408050537, "learning_rate": 1.1142206373524067e-05, "loss": 2.7312, "step": 443900 }, { "epoch": 3.920503717833236, "grad_norm": 6.338700294494629, "learning_rate": 1.113346289538153e-05, "loss": 3.0767, "step": 443950 }, { "epoch": 3.92094526572352, "grad_norm": 2.5857155323028564, "learning_rate": 1.112472241934941e-05, "loss": 2.835, "step": 444000 }, { "epoch": 3.92094526572352, "eval_asr_loss": 0.9615496546034715, "eval_loss": 2.6829841136932373, "eval_runtime": 20.1607, "eval_samples_per_second": 38.094, "eval_steps_per_second": 9.523, "eval_tts_loss": 6.165717166562772, "step": 444000 }, { "epoch": 3.921386813613805, "grad_norm": 1.693056344985962, "learning_rate": 1.1115984946102847e-05, "loss": 2.7354, "step": 444050 }, { "epoch": 3.9218283615040885, "grad_norm": 3.63508939743042, "learning_rate": 1.11072504763167e-05, "loss": 2.8316, "step": 444100 }, { "epoch": 3.922269909394373, "grad_norm": 3.745271921157837, "learning_rate": 1.109851901066567e-05, "loss": 2.2832, "step": 444150 }, { "epoch": 3.922711457284657, "grad_norm": 5.15354585647583, "learning_rate": 1.1089790549824175e-05, "loss": 3.1685, "step": 444200 }, { "epoch": 3.9231530051749415, "grad_norm": 5.1281328201293945, "learning_rate": 1.1081065094466392e-05, "loss": 3.1423, "step": 444250 }, { "epoch": 3.923594553065225, "grad_norm": 6.264153480529785, "learning_rate": 1.107234264526631e-05, "loss": 2.9997, "step": 444300 }, { "epoch": 3.92403610095551, "grad_norm": 6.200822353363037, "learning_rate": 1.1063623202897672e-05, "loss": 2.8509, "step": 444350 }, { "epoch": 3.924477648845794, "grad_norm": 1.790786623954773, "learning_rate": 1.1054906768033962e-05, "loss": 2.9659, "step": 444400 }, { "epoch": 3.924919196736078, "grad_norm": 6.818342208862305, "learning_rate": 1.1046193341348454e-05, "loss": 3.2573, "step": 444450 }, { "epoch": 3.9253607446263623, "grad_norm": 8.159503936767578, "learning_rate": 1.1037482923514198e-05, "loss": 3.122, "step": 444500 }, { "epoch": 3.9258022925166465, "grad_norm": 2.408308744430542, "learning_rate": 1.1028775515204004e-05, "loss": 3.0848, "step": 444550 }, { "epoch": 3.9262438404069306, "grad_norm": 4.068150520324707, "learning_rate": 1.1020071117090435e-05, "loss": 2.8634, "step": 444600 }, { "epoch": 3.926685388297215, "grad_norm": 1.9536625146865845, "learning_rate": 1.1011543728079854e-05, "loss": 2.7923, "step": 444650 }, { "epoch": 3.927126936187499, "grad_norm": 2.6429550647735596, "learning_rate": 1.1002845292138903e-05, "loss": 2.7058, "step": 444700 }, { "epoch": 3.927568484077783, "grad_norm": 5.143438339233398, "learning_rate": 1.0994149868397486e-05, "loss": 2.4543, "step": 444750 }, { "epoch": 3.9280100319680673, "grad_norm": 4.8779802322387695, "learning_rate": 1.0985457457527209e-05, "loss": 2.8871, "step": 444800 }, { "epoch": 3.9284515798583515, "grad_norm": 2.966792583465576, "learning_rate": 1.0976768060199533e-05, "loss": 3.3262, "step": 444850 }, { "epoch": 3.9288931277486356, "grad_norm": 3.0403637886047363, "learning_rate": 1.09680816770856e-05, "loss": 2.852, "step": 444900 }, { "epoch": 3.92933467563892, "grad_norm": 2.353562355041504, "learning_rate": 1.0959398308856388e-05, "loss": 3.0924, "step": 444950 }, { "epoch": 3.929776223529204, "grad_norm": 3.3208329677581787, "learning_rate": 1.0950717956182594e-05, "loss": 2.9329, "step": 445000 }, { "epoch": 3.930217771419488, "grad_norm": 8.0339937210083, "learning_rate": 1.0942040619734706e-05, "loss": 2.9969, "step": 445050 }, { "epoch": 3.9306593193097723, "grad_norm": 4.954730033874512, "learning_rate": 1.093336630018299e-05, "loss": 3.3219, "step": 445100 }, { "epoch": 3.9311008672000565, "grad_norm": 3.158216953277588, "learning_rate": 1.0924694998197443e-05, "loss": 2.883, "step": 445150 }, { "epoch": 3.9315424150903406, "grad_norm": 7.931962013244629, "learning_rate": 1.0916026714447852e-05, "loss": 2.2608, "step": 445200 }, { "epoch": 3.931983962980625, "grad_norm": 6.363288879394531, "learning_rate": 1.0907361449603792e-05, "loss": 2.8106, "step": 445250 }, { "epoch": 3.932425510870909, "grad_norm": 4.504587173461914, "learning_rate": 1.0898699204334567e-05, "loss": 2.9973, "step": 445300 }, { "epoch": 3.932867058761193, "grad_norm": 5.023761749267578, "learning_rate": 1.0890039979309242e-05, "loss": 2.9628, "step": 445350 }, { "epoch": 3.9333086066514773, "grad_norm": 11.226164817810059, "learning_rate": 1.0881383775196702e-05, "loss": 2.6111, "step": 445400 }, { "epoch": 3.9337501545417615, "grad_norm": 3.785357713699341, "learning_rate": 1.0872730592665537e-05, "loss": 3.0473, "step": 445450 }, { "epoch": 3.9341917024320456, "grad_norm": 3.668285608291626, "learning_rate": 1.0864080432384138e-05, "loss": 3.1824, "step": 445500 }, { "epoch": 3.93463325032233, "grad_norm": 4.762940883636475, "learning_rate": 1.0855433295020667e-05, "loss": 3.1237, "step": 445550 }, { "epoch": 3.9350747982126144, "grad_norm": 4.487186908721924, "learning_rate": 1.084678918124305e-05, "loss": 2.77, "step": 445600 }, { "epoch": 3.935516346102898, "grad_norm": 5.384033679962158, "learning_rate": 1.0838148091718959e-05, "loss": 2.8415, "step": 445650 }, { "epoch": 3.9359578939931827, "grad_norm": 5.6339569091796875, "learning_rate": 1.082951002711583e-05, "loss": 2.4323, "step": 445700 }, { "epoch": 3.9363994418834665, "grad_norm": 2.731947183609009, "learning_rate": 1.0820874988100915e-05, "loss": 2.7774, "step": 445750 }, { "epoch": 3.936840989773751, "grad_norm": 5.898236274719238, "learning_rate": 1.0812242975341159e-05, "loss": 2.8472, "step": 445800 }, { "epoch": 3.937282537664035, "grad_norm": 4.7647786140441895, "learning_rate": 1.080361398950333e-05, "loss": 3.1951, "step": 445850 }, { "epoch": 3.9377240855543194, "grad_norm": 3.0825345516204834, "learning_rate": 1.079498803125396e-05, "loss": 2.9035, "step": 445900 }, { "epoch": 3.938165633444603, "grad_norm": 2.759737730026245, "learning_rate": 1.0786365101259304e-05, "loss": 2.4038, "step": 445950 }, { "epoch": 3.9386071813348877, "grad_norm": 5.704274654388428, "learning_rate": 1.0777745200185419e-05, "loss": 3.0178, "step": 446000 }, { "epoch": 3.939048729225172, "grad_norm": 3.1521449089050293, "learning_rate": 1.0769128328698137e-05, "loss": 3.1073, "step": 446050 }, { "epoch": 3.939490277115456, "grad_norm": 7.256103992462158, "learning_rate": 1.0760514487463024e-05, "loss": 2.7259, "step": 446100 }, { "epoch": 3.9399318250057402, "grad_norm": 6.133712291717529, "learning_rate": 1.0751903677145414e-05, "loss": 2.987, "step": 446150 }, { "epoch": 3.9403733728960244, "grad_norm": 4.929304122924805, "learning_rate": 1.0743295898410428e-05, "loss": 2.6281, "step": 446200 }, { "epoch": 3.9408149207863086, "grad_norm": 2.31499981880188, "learning_rate": 1.073469115192296e-05, "loss": 2.8291, "step": 446250 }, { "epoch": 3.9412564686765927, "grad_norm": 1.8054077625274658, "learning_rate": 1.0726089438347631e-05, "loss": 2.5472, "step": 446300 }, { "epoch": 3.941698016566877, "grad_norm": 4.8036627769470215, "learning_rate": 1.0717490758348853e-05, "loss": 2.6195, "step": 446350 }, { "epoch": 3.942139564457161, "grad_norm": 5.270541667938232, "learning_rate": 1.0708895112590828e-05, "loss": 2.8535, "step": 446400 }, { "epoch": 3.9425811123474452, "grad_norm": 6.616434097290039, "learning_rate": 1.0700302501737475e-05, "loss": 3.1167, "step": 446450 }, { "epoch": 3.9430226602377294, "grad_norm": 2.9126224517822266, "learning_rate": 1.0691712926452485e-05, "loss": 2.539, "step": 446500 }, { "epoch": 3.9434642081280136, "grad_norm": 6.614195346832275, "learning_rate": 1.0683126387399362e-05, "loss": 2.7856, "step": 446550 }, { "epoch": 3.9439057560182977, "grad_norm": 2.6681532859802246, "learning_rate": 1.0674542885241317e-05, "loss": 3.0365, "step": 446600 }, { "epoch": 3.944347303908582, "grad_norm": 2.856962203979492, "learning_rate": 1.0665962420641363e-05, "loss": 3.0017, "step": 446650 }, { "epoch": 3.944788851798866, "grad_norm": 2.0507402420043945, "learning_rate": 1.0657384994262271e-05, "loss": 2.6881, "step": 446700 }, { "epoch": 3.9452303996891502, "grad_norm": 8.904801368713379, "learning_rate": 1.064881060676658e-05, "loss": 2.7319, "step": 446750 }, { "epoch": 3.9456719475794344, "grad_norm": 4.586367130279541, "learning_rate": 1.0640239258816575e-05, "loss": 2.6239, "step": 446800 }, { "epoch": 3.9461134954697186, "grad_norm": 2.5034067630767822, "learning_rate": 1.0631670951074334e-05, "loss": 3.0295, "step": 446850 }, { "epoch": 3.9465550433600027, "grad_norm": 5.517556190490723, "learning_rate": 1.0623105684201683e-05, "loss": 2.1474, "step": 446900 }, { "epoch": 3.946996591250287, "grad_norm": 8.240912437438965, "learning_rate": 1.0614543458860193e-05, "loss": 2.708, "step": 446950 }, { "epoch": 3.947438139140571, "grad_norm": 5.265829086303711, "learning_rate": 1.0605984275711245e-05, "loss": 3.0943, "step": 447000 }, { "epoch": 3.947438139140571, "eval_asr_loss": 0.966802725125614, "eval_loss": 2.6805641651153564, "eval_runtime": 20.1182, "eval_samples_per_second": 38.174, "eval_steps_per_second": 9.544, "eval_tts_loss": 6.171010745453993, "step": 447000 }, { "epoch": 3.9478796870308552, "grad_norm": 3.408456802368164, "learning_rate": 1.0597428135415977e-05, "loss": 2.6162, "step": 447050 }, { "epoch": 3.9483212349211394, "grad_norm": 4.972458362579346, "learning_rate": 1.0588875038635242e-05, "loss": 3.0026, "step": 447100 }, { "epoch": 3.9487627828114236, "grad_norm": 1.9543092250823975, "learning_rate": 1.0580324986029716e-05, "loss": 2.6325, "step": 447150 }, { "epoch": 3.9492043307017077, "grad_norm": 3.513413667678833, "learning_rate": 1.0571777978259828e-05, "loss": 2.956, "step": 447200 }, { "epoch": 3.9496458785919923, "grad_norm": 3.90350079536438, "learning_rate": 1.0563234015985752e-05, "loss": 3.0146, "step": 447250 }, { "epoch": 3.950087426482276, "grad_norm": 3.9892985820770264, "learning_rate": 1.0554693099867419e-05, "loss": 3.0846, "step": 447300 }, { "epoch": 3.9505289743725607, "grad_norm": 2.0285234451293945, "learning_rate": 1.0546155230564558e-05, "loss": 2.3158, "step": 447350 }, { "epoch": 3.9509705222628444, "grad_norm": 3.8300673961639404, "learning_rate": 1.0537620408736659e-05, "loss": 2.985, "step": 447400 }, { "epoch": 3.951412070153129, "grad_norm": 5.091235160827637, "learning_rate": 1.0529088635042945e-05, "loss": 2.9973, "step": 447450 }, { "epoch": 3.9518536180434127, "grad_norm": 3.695396900177002, "learning_rate": 1.0520559910142425e-05, "loss": 3.3579, "step": 447500 }, { "epoch": 3.9522951659336973, "grad_norm": 5.487963676452637, "learning_rate": 1.051203423469389e-05, "loss": 2.9677, "step": 447550 }, { "epoch": 3.9527367138239815, "grad_norm": 6.580090522766113, "learning_rate": 1.050351160935586e-05, "loss": 3.1241, "step": 447600 }, { "epoch": 3.9531782617142657, "grad_norm": 10.159269332885742, "learning_rate": 1.0494992034786627e-05, "loss": 3.1981, "step": 447650 }, { "epoch": 3.95361980960455, "grad_norm": 6.06384801864624, "learning_rate": 1.0486475511644279e-05, "loss": 3.031, "step": 447700 }, { "epoch": 3.954061357494834, "grad_norm": 2.198991298675537, "learning_rate": 1.047796204058662e-05, "loss": 2.867, "step": 447750 }, { "epoch": 3.954502905385118, "grad_norm": 5.54725980758667, "learning_rate": 1.0469451622271258e-05, "loss": 3.2249, "step": 447800 }, { "epoch": 3.9549444532754023, "grad_norm": 8.948962211608887, "learning_rate": 1.0460944257355549e-05, "loss": 2.6768, "step": 447850 }, { "epoch": 3.9553860011656865, "grad_norm": 2.525737762451172, "learning_rate": 1.0452439946496623e-05, "loss": 2.6809, "step": 447900 }, { "epoch": 3.9558275490559707, "grad_norm": 6.887853622436523, "learning_rate": 1.0443938690351346e-05, "loss": 2.6603, "step": 447950 }, { "epoch": 3.956269096946255, "grad_norm": 4.2360992431640625, "learning_rate": 1.0435440489576398e-05, "loss": 2.99, "step": 448000 }, { "epoch": 3.956710644836539, "grad_norm": 5.050408840179443, "learning_rate": 1.0426945344828171e-05, "loss": 2.9409, "step": 448050 }, { "epoch": 3.957152192726823, "grad_norm": 2.6867311000823975, "learning_rate": 1.0418453256762834e-05, "loss": 2.797, "step": 448100 }, { "epoch": 3.9575937406171073, "grad_norm": 4.741623878479004, "learning_rate": 1.040996422603634e-05, "loss": 2.8304, "step": 448150 }, { "epoch": 3.9580352885073915, "grad_norm": 2.225299596786499, "learning_rate": 1.0401478253304408e-05, "loss": 2.8951, "step": 448200 }, { "epoch": 3.9584768363976757, "grad_norm": 3.9259207248687744, "learning_rate": 1.0392995339222484e-05, "loss": 3.0369, "step": 448250 }, { "epoch": 3.95891838428796, "grad_norm": 2.4653501510620117, "learning_rate": 1.0384515484445811e-05, "loss": 2.4398, "step": 448300 }, { "epoch": 3.959359932178244, "grad_norm": 1.29037344455719, "learning_rate": 1.0376038689629403e-05, "loss": 3.0625, "step": 448350 }, { "epoch": 3.959801480068528, "grad_norm": 5.052370071411133, "learning_rate": 1.036773440011376e-05, "loss": 3.3838, "step": 448400 }, { "epoch": 3.9602430279588123, "grad_norm": 5.554194927215576, "learning_rate": 1.0359263665950091e-05, "loss": 3.2177, "step": 448450 }, { "epoch": 3.9606845758490965, "grad_norm": 3.9829583168029785, "learning_rate": 1.0350795993697143e-05, "loss": 3.0454, "step": 448500 }, { "epoch": 3.9611261237393807, "grad_norm": 4.488154411315918, "learning_rate": 1.0342331384008996e-05, "loss": 2.5662, "step": 448550 }, { "epoch": 3.961567671629665, "grad_norm": 3.0822267532348633, "learning_rate": 1.0333869837539473e-05, "loss": 3.0007, "step": 448600 }, { "epoch": 3.962009219519949, "grad_norm": 7.231896877288818, "learning_rate": 1.0325411354942132e-05, "loss": 2.3563, "step": 448650 }, { "epoch": 3.962450767410233, "grad_norm": 2.9030203819274902, "learning_rate": 1.0316955936870349e-05, "loss": 3.0988, "step": 448700 }, { "epoch": 3.9628923153005173, "grad_norm": 4.017431259155273, "learning_rate": 1.0308503583977214e-05, "loss": 2.6777, "step": 448750 }, { "epoch": 3.963333863190802, "grad_norm": 4.991633892059326, "learning_rate": 1.030005429691559e-05, "loss": 2.6489, "step": 448800 }, { "epoch": 3.9637754110810857, "grad_norm": 11.141836166381836, "learning_rate": 1.0291608076338128e-05, "loss": 2.5636, "step": 448850 }, { "epoch": 3.9642169589713703, "grad_norm": 2.3982977867126465, "learning_rate": 1.0283164922897221e-05, "loss": 3.1066, "step": 448900 }, { "epoch": 3.964658506861654, "grad_norm": 7.016125202178955, "learning_rate": 1.0274724837245054e-05, "loss": 2.6902, "step": 448950 }, { "epoch": 3.9651000547519386, "grad_norm": 5.6254048347473145, "learning_rate": 1.0266287820033516e-05, "loss": 2.7891, "step": 449000 }, { "epoch": 3.9655416026422223, "grad_norm": 5.982679843902588, "learning_rate": 1.0257853871914325e-05, "loss": 3.1556, "step": 449050 }, { "epoch": 3.965983150532507, "grad_norm": 5.602755069732666, "learning_rate": 1.0249422993538916e-05, "loss": 2.7356, "step": 449100 }, { "epoch": 3.9664246984227907, "grad_norm": 4.3636651039123535, "learning_rate": 1.02409951855585e-05, "loss": 2.523, "step": 449150 }, { "epoch": 3.9668662463130753, "grad_norm": 2.0972955226898193, "learning_rate": 1.0232570448624052e-05, "loss": 2.5386, "step": 449200 }, { "epoch": 3.9673077942033594, "grad_norm": 3.481414794921875, "learning_rate": 1.0224148783386333e-05, "loss": 2.59, "step": 449250 }, { "epoch": 3.9677493420936436, "grad_norm": 6.0625739097595215, "learning_rate": 1.021573019049582e-05, "loss": 2.8444, "step": 449300 }, { "epoch": 3.9681908899839278, "grad_norm": 2.805246591567993, "learning_rate": 1.020731467060279e-05, "loss": 3.0823, "step": 449350 }, { "epoch": 3.968632437874212, "grad_norm": 3.79797101020813, "learning_rate": 1.0198902224357281e-05, "loss": 3.0785, "step": 449400 }, { "epoch": 3.969073985764496, "grad_norm": 5.7656569480896, "learning_rate": 1.0190492852409062e-05, "loss": 2.9025, "step": 449450 }, { "epoch": 3.9695155336547803, "grad_norm": 4.404407024383545, "learning_rate": 1.0182086555407705e-05, "loss": 3.04, "step": 449500 }, { "epoch": 3.9699570815450644, "grad_norm": 4.36983585357666, "learning_rate": 1.0173683334002503e-05, "loss": 2.4176, "step": 449550 }, { "epoch": 3.9703986294353486, "grad_norm": 6.727650165557861, "learning_rate": 1.0165283188842561e-05, "loss": 2.5651, "step": 449600 }, { "epoch": 3.9708401773256328, "grad_norm": 1.3279187679290771, "learning_rate": 1.0156886120576691e-05, "loss": 2.7438, "step": 449650 }, { "epoch": 3.971281725215917, "grad_norm": 4.2612128257751465, "learning_rate": 1.0148492129853504e-05, "loss": 2.6796, "step": 449700 }, { "epoch": 3.971723273106201, "grad_norm": 3.3085122108459473, "learning_rate": 1.014010121732139e-05, "loss": 2.8757, "step": 449750 }, { "epoch": 3.9721648209964853, "grad_norm": 7.612644672393799, "learning_rate": 1.0131713383628432e-05, "loss": 3.0626, "step": 449800 }, { "epoch": 3.9726063688867694, "grad_norm": 6.785557746887207, "learning_rate": 1.012349629432352e-05, "loss": 3.0708, "step": 449850 }, { "epoch": 3.9730479167770536, "grad_norm": 4.389591693878174, "learning_rate": 1.0115114558643319e-05, "loss": 2.8544, "step": 449900 }, { "epoch": 3.9734894646673378, "grad_norm": 4.892695426940918, "learning_rate": 1.0106735903732294e-05, "loss": 2.5423, "step": 449950 }, { "epoch": 3.973931012557622, "grad_norm": 3.618398666381836, "learning_rate": 1.0098360330237655e-05, "loss": 2.6997, "step": 450000 }, { "epoch": 3.973931012557622, "eval_asr_loss": 0.9606481025491705, "eval_loss": 2.6783769130706787, "eval_runtime": 20.5512, "eval_samples_per_second": 37.37, "eval_steps_per_second": 9.343, "eval_tts_loss": 6.166964953530546, "step": 450000 }, { "epoch": 3.974372560447906, "grad_norm": 2.1839284896850586, "learning_rate": 1.0089987838806314e-05, "loss": 2.9455, "step": 450050 }, { "epoch": 3.9748141083381903, "grad_norm": 3.7029201984405518, "learning_rate": 1.0081618430084983e-05, "loss": 2.7156, "step": 450100 }, { "epoch": 3.9752556562284744, "grad_norm": 2.4859559535980225, "learning_rate": 1.0073252104720138e-05, "loss": 2.768, "step": 450150 }, { "epoch": 3.9756972041187586, "grad_norm": 5.42495584487915, "learning_rate": 1.0064888863358008e-05, "loss": 2.9861, "step": 450200 }, { "epoch": 3.9761387520090428, "grad_norm": 3.656745195388794, "learning_rate": 1.0056528706644552e-05, "loss": 2.7481, "step": 450250 }, { "epoch": 3.976580299899327, "grad_norm": 7.593056678771973, "learning_rate": 1.0048171635225561e-05, "loss": 2.5495, "step": 450300 }, { "epoch": 3.977021847789611, "grad_norm": 5.959151744842529, "learning_rate": 1.0039817649746514e-05, "loss": 3.1215, "step": 450350 }, { "epoch": 3.9774633956798953, "grad_norm": 5.786403656005859, "learning_rate": 1.0031466750852702e-05, "loss": 2.9029, "step": 450400 }, { "epoch": 3.97790494357018, "grad_norm": 6.267789840698242, "learning_rate": 1.0023118939189163e-05, "loss": 2.867, "step": 450450 }, { "epoch": 3.9783464914604636, "grad_norm": 5.520458698272705, "learning_rate": 1.0014774215400702e-05, "loss": 2.8259, "step": 450500 }, { "epoch": 3.978788039350748, "grad_norm": 1.8382903337478638, "learning_rate": 1.000643258013187e-05, "loss": 2.6027, "step": 450550 }, { "epoch": 3.979229587241032, "grad_norm": 2.495784044265747, "learning_rate": 9.998094034026967e-06, "loss": 3.0318, "step": 450600 }, { "epoch": 3.9796711351313165, "grad_norm": 3.3280630111694336, "learning_rate": 9.989758577730113e-06, "loss": 2.9138, "step": 450650 }, { "epoch": 3.9801126830216003, "grad_norm": 6.609112739562988, "learning_rate": 9.981426211885114e-06, "loss": 2.9698, "step": 450700 }, { "epoch": 3.980554230911885, "grad_norm": 4.50701379776001, "learning_rate": 9.973096937135596e-06, "loss": 3.1839, "step": 450750 }, { "epoch": 3.980995778802169, "grad_norm": 2.983055353164673, "learning_rate": 9.964770754124935e-06, "loss": 3.068, "step": 450800 }, { "epoch": 3.981437326692453, "grad_norm": 2.5820963382720947, "learning_rate": 9.956447663496227e-06, "loss": 2.6034, "step": 450850 }, { "epoch": 3.9818788745827374, "grad_norm": 3.529637575149536, "learning_rate": 9.948127665892381e-06, "loss": 2.7004, "step": 450900 }, { "epoch": 3.9823204224730215, "grad_norm": 2.8287980556488037, "learning_rate": 9.939810761956058e-06, "loss": 2.7449, "step": 450950 }, { "epoch": 3.9827619703633057, "grad_norm": 6.192989349365234, "learning_rate": 9.93149695232965e-06, "loss": 2.6533, "step": 451000 }, { "epoch": 3.98320351825359, "grad_norm": 4.550994873046875, "learning_rate": 9.923186237655319e-06, "loss": 2.7437, "step": 451050 }, { "epoch": 3.983645066143874, "grad_norm": 5.825502395629883, "learning_rate": 9.914878618575007e-06, "loss": 3.2606, "step": 451100 }, { "epoch": 3.984086614034158, "grad_norm": 3.8592281341552734, "learning_rate": 9.906574095730425e-06, "loss": 2.637, "step": 451150 }, { "epoch": 3.9845281619244424, "grad_norm": 3.1612186431884766, "learning_rate": 9.898272669763003e-06, "loss": 2.7486, "step": 451200 }, { "epoch": 3.9849697098147265, "grad_norm": 3.2317638397216797, "learning_rate": 9.889974341313963e-06, "loss": 3.0181, "step": 451250 }, { "epoch": 3.9854112577050107, "grad_norm": 2.7730965614318848, "learning_rate": 9.88167911102429e-06, "loss": 2.7435, "step": 451300 }, { "epoch": 3.985852805595295, "grad_norm": 4.482161521911621, "learning_rate": 9.873386979534715e-06, "loss": 2.9209, "step": 451350 }, { "epoch": 3.986294353485579, "grad_norm": 7.530752658843994, "learning_rate": 9.865097947485718e-06, "loss": 3.173, "step": 451400 }, { "epoch": 3.986735901375863, "grad_norm": 14.592049598693848, "learning_rate": 9.856812015517586e-06, "loss": 2.8778, "step": 451450 }, { "epoch": 3.9871774492661474, "grad_norm": 5.6240386962890625, "learning_rate": 9.848529184270305e-06, "loss": 3.0288, "step": 451500 }, { "epoch": 3.9876189971564315, "grad_norm": 7.2018914222717285, "learning_rate": 9.840249454383676e-06, "loss": 2.6281, "step": 451550 }, { "epoch": 3.9880605450467157, "grad_norm": 2.289982557296753, "learning_rate": 9.831972826497238e-06, "loss": 2.3746, "step": 451600 }, { "epoch": 3.988502092937, "grad_norm": 5.578866958618164, "learning_rate": 9.823699301250277e-06, "loss": 2.6794, "step": 451650 }, { "epoch": 3.988943640827284, "grad_norm": 3.6381967067718506, "learning_rate": 9.815428879281857e-06, "loss": 2.8511, "step": 451700 }, { "epoch": 3.989385188717568, "grad_norm": 8.4154634475708, "learning_rate": 9.807161561230815e-06, "loss": 2.5765, "step": 451750 }, { "epoch": 3.9898267366078524, "grad_norm": 9.250537872314453, "learning_rate": 9.798897347735714e-06, "loss": 2.8603, "step": 451800 }, { "epoch": 3.9902682844981365, "grad_norm": 5.195905685424805, "learning_rate": 9.790636239434887e-06, "loss": 2.9402, "step": 451850 }, { "epoch": 3.9907098323884207, "grad_norm": 5.664506435394287, "learning_rate": 9.782378236966443e-06, "loss": 2.7473, "step": 451900 }, { "epoch": 3.991151380278705, "grad_norm": 3.9338440895080566, "learning_rate": 9.77412334096826e-06, "loss": 3.082, "step": 451950 }, { "epoch": 3.9915929281689895, "grad_norm": 3.380535840988159, "learning_rate": 9.765871552077931e-06, "loss": 2.7233, "step": 452000 }, { "epoch": 3.992034476059273, "grad_norm": 7.822603702545166, "learning_rate": 9.757622870932847e-06, "loss": 2.839, "step": 452050 }, { "epoch": 3.992476023949558, "grad_norm": 6.796748638153076, "learning_rate": 9.749377298170165e-06, "loss": 2.3689, "step": 452100 }, { "epoch": 3.9929175718398415, "grad_norm": 3.173342704772949, "learning_rate": 9.741134834426773e-06, "loss": 2.7864, "step": 452150 }, { "epoch": 3.993359119730126, "grad_norm": 2.2958638668060303, "learning_rate": 9.732895480339316e-06, "loss": 2.9263, "step": 452200 }, { "epoch": 3.99380066762041, "grad_norm": 8.014845848083496, "learning_rate": 9.724659236544226e-06, "loss": 2.6727, "step": 452250 }, { "epoch": 3.9942422155106945, "grad_norm": 4.8161821365356445, "learning_rate": 9.7164261036777e-06, "loss": 2.5683, "step": 452300 }, { "epoch": 3.994683763400978, "grad_norm": 1.5090913772583008, "learning_rate": 9.708196082375649e-06, "loss": 3.0656, "step": 452350 }, { "epoch": 3.995125311291263, "grad_norm": 4.5364813804626465, "learning_rate": 9.699969173273787e-06, "loss": 2.711, "step": 452400 }, { "epoch": 3.995566859181547, "grad_norm": 5.46333646774292, "learning_rate": 9.691745377007582e-06, "loss": 2.948, "step": 452450 }, { "epoch": 3.996008407071831, "grad_norm": 3.928785800933838, "learning_rate": 9.683524694212232e-06, "loss": 3.0161, "step": 452500 }, { "epoch": 3.9964499549621153, "grad_norm": 5.975211143493652, "learning_rate": 9.675307125522742e-06, "loss": 3.0269, "step": 452550 }, { "epoch": 3.9968915028523995, "grad_norm": 2.9269394874572754, "learning_rate": 9.66709267157383e-06, "loss": 2.7917, "step": 452600 }, { "epoch": 3.9973330507426836, "grad_norm": 1.4483669996261597, "learning_rate": 9.658881332999981e-06, "loss": 2.3176, "step": 452650 }, { "epoch": 3.997774598632968, "grad_norm": 6.12791633605957, "learning_rate": 9.650673110435465e-06, "loss": 2.4778, "step": 452700 }, { "epoch": 3.998216146523252, "grad_norm": 4.50653076171875, "learning_rate": 9.642468004514315e-06, "loss": 2.9882, "step": 452750 }, { "epoch": 3.998657694413536, "grad_norm": 2.2186005115509033, "learning_rate": 9.634266015870281e-06, "loss": 2.8694, "step": 452800 }, { "epoch": 3.9990992423038203, "grad_norm": 6.475503921508789, "learning_rate": 9.626067145136902e-06, "loss": 3.0196, "step": 452850 }, { "epoch": 3.9995407901941045, "grad_norm": 3.6491920948028564, "learning_rate": 9.617871392947485e-06, "loss": 2.8308, "step": 452900 }, { "epoch": 3.9999823380843886, "grad_norm": 3.336529493331909, "learning_rate": 9.609678759935075e-06, "loss": 3.1263, "step": 452950 }, { "epoch": 4.000423885974673, "grad_norm": 7.990113735198975, "learning_rate": 9.60148924673247e-06, "loss": 2.7556, "step": 453000 }, { "epoch": 4.000423885974673, "eval_asr_loss": 0.9627899950716676, "eval_loss": 2.684953451156616, "eval_runtime": 20.1541, "eval_samples_per_second": 38.106, "eval_steps_per_second": 9.527, "eval_tts_loss": 6.311691039914507, "step": 453000 }, { "epoch": 4.000865433864957, "grad_norm": 5.167287349700928, "learning_rate": 9.593302853972247e-06, "loss": 2.7306, "step": 453050 }, { "epoch": 4.001306981755241, "grad_norm": 4.54832649230957, "learning_rate": 9.585119582286756e-06, "loss": 2.7492, "step": 453100 }, { "epoch": 4.001748529645526, "grad_norm": 1.961534857749939, "learning_rate": 9.576939432308057e-06, "loss": 2.3077, "step": 453150 }, { "epoch": 4.0021900775358095, "grad_norm": 5.067187309265137, "learning_rate": 9.568762404668008e-06, "loss": 2.5295, "step": 453200 }, { "epoch": 4.002631625426094, "grad_norm": 4.598661422729492, "learning_rate": 9.56058849999823e-06, "loss": 3.0131, "step": 453250 }, { "epoch": 4.003073173316378, "grad_norm": 3.7365050315856934, "learning_rate": 9.552417718930073e-06, "loss": 2.6082, "step": 453300 }, { "epoch": 4.003514721206662, "grad_norm": 2.443026304244995, "learning_rate": 9.544250062094645e-06, "loss": 2.8555, "step": 453350 }, { "epoch": 4.003956269096946, "grad_norm": 2.321892261505127, "learning_rate": 9.536085530122852e-06, "loss": 3.0622, "step": 453400 }, { "epoch": 4.004397816987231, "grad_norm": 4.167388439178467, "learning_rate": 9.527924123645338e-06, "loss": 2.3442, "step": 453450 }, { "epoch": 4.0048393648775145, "grad_norm": 4.840127944946289, "learning_rate": 9.519765843292478e-06, "loss": 2.9138, "step": 453500 }, { "epoch": 4.005280912767799, "grad_norm": 4.54113245010376, "learning_rate": 9.511610689694444e-06, "loss": 2.3633, "step": 453550 }, { "epoch": 4.005722460658083, "grad_norm": 4.568275451660156, "learning_rate": 9.503458663481168e-06, "loss": 2.5542, "step": 453600 }, { "epoch": 4.006164008548367, "grad_norm": 0.9584481120109558, "learning_rate": 9.495309765282296e-06, "loss": 2.7616, "step": 453650 }, { "epoch": 4.006605556438651, "grad_norm": 4.198399543762207, "learning_rate": 9.487163995727289e-06, "loss": 2.1571, "step": 453700 }, { "epoch": 4.007047104328936, "grad_norm": 7.235071659088135, "learning_rate": 9.479021355445322e-06, "loss": 2.2788, "step": 453750 }, { "epoch": 4.0074886522192195, "grad_norm": 8.156457901000977, "learning_rate": 9.470881845065333e-06, "loss": 2.6469, "step": 453800 }, { "epoch": 4.007930200109504, "grad_norm": 6.073143005371094, "learning_rate": 9.46274546521605e-06, "loss": 2.7912, "step": 453850 }, { "epoch": 4.008371747999788, "grad_norm": 5.100198268890381, "learning_rate": 9.454612216525943e-06, "loss": 2.4663, "step": 453900 }, { "epoch": 4.008813295890072, "grad_norm": 4.925257205963135, "learning_rate": 9.446482099623217e-06, "loss": 2.7408, "step": 453950 }, { "epoch": 4.009254843780356, "grad_norm": 6.113733768463135, "learning_rate": 9.43835511513586e-06, "loss": 2.7096, "step": 454000 }, { "epoch": 4.009696391670641, "grad_norm": 7.609586238861084, "learning_rate": 9.430231263691636e-06, "loss": 2.7363, "step": 454050 }, { "epoch": 4.0101379395609245, "grad_norm": 3.1827592849731445, "learning_rate": 9.422110545918023e-06, "loss": 3.1194, "step": 454100 }, { "epoch": 4.010579487451209, "grad_norm": 3.40751576423645, "learning_rate": 9.413992962442264e-06, "loss": 2.5582, "step": 454150 }, { "epoch": 4.011021035341493, "grad_norm": 3.9222395420074463, "learning_rate": 9.405878513891392e-06, "loss": 3.0239, "step": 454200 }, { "epoch": 4.011462583231777, "grad_norm": 4.692047595977783, "learning_rate": 9.397767200892193e-06, "loss": 2.4255, "step": 454250 }, { "epoch": 4.011904131122061, "grad_norm": 4.399960517883301, "learning_rate": 9.389659024071162e-06, "loss": 2.8531, "step": 454300 }, { "epoch": 4.012345679012346, "grad_norm": 3.471290349960327, "learning_rate": 9.381553984054614e-06, "loss": 2.3456, "step": 454350 }, { "epoch": 4.0127872269026295, "grad_norm": 4.1341447830200195, "learning_rate": 9.373452081468592e-06, "loss": 2.5767, "step": 454400 }, { "epoch": 4.013228774792914, "grad_norm": 6.719664573669434, "learning_rate": 9.365353316938885e-06, "loss": 2.4757, "step": 454450 }, { "epoch": 4.013670322683198, "grad_norm": 4.21216344833374, "learning_rate": 9.35725769109107e-06, "loss": 2.7484, "step": 454500 }, { "epoch": 4.014111870573482, "grad_norm": 4.342075347900391, "learning_rate": 9.349327023512011e-06, "loss": 2.5962, "step": 454550 }, { "epoch": 4.014553418463767, "grad_norm": 7.861991882324219, "learning_rate": 9.341237614098896e-06, "loss": 2.7379, "step": 454600 }, { "epoch": 4.014994966354051, "grad_norm": 4.5301618576049805, "learning_rate": 9.333151345230417e-06, "loss": 2.5823, "step": 454650 }, { "epoch": 4.015436514244335, "grad_norm": 6.785705089569092, "learning_rate": 9.325068217531135e-06, "loss": 2.5483, "step": 454700 }, { "epoch": 4.015878062134619, "grad_norm": 5.5882086753845215, "learning_rate": 9.316988231625435e-06, "loss": 2.7655, "step": 454750 }, { "epoch": 4.016319610024904, "grad_norm": 9.296425819396973, "learning_rate": 9.308911388137408e-06, "loss": 2.8127, "step": 454800 }, { "epoch": 4.016761157915187, "grad_norm": 6.52687931060791, "learning_rate": 9.300837687690922e-06, "loss": 2.7329, "step": 454850 }, { "epoch": 4.017202705805472, "grad_norm": 1.1604112386703491, "learning_rate": 9.292767130909596e-06, "loss": 2.4548, "step": 454900 }, { "epoch": 4.017644253695756, "grad_norm": 4.026607513427734, "learning_rate": 9.284699718416839e-06, "loss": 2.3518, "step": 454950 }, { "epoch": 4.01808580158604, "grad_norm": 1.163305401802063, "learning_rate": 9.276635450835764e-06, "loss": 2.0727, "step": 455000 }, { "epoch": 4.018527349476324, "grad_norm": 5.820218086242676, "learning_rate": 9.268574328789271e-06, "loss": 2.7126, "step": 455050 }, { "epoch": 4.018968897366609, "grad_norm": 4.245255470275879, "learning_rate": 9.260677481581453e-06, "loss": 2.6467, "step": 455100 }, { "epoch": 4.019410445256892, "grad_norm": 6.256257057189941, "learning_rate": 9.252622589530163e-06, "loss": 2.2644, "step": 455150 }, { "epoch": 4.019851993147177, "grad_norm": 8.562440872192383, "learning_rate": 9.24457084486826e-06, "loss": 2.8629, "step": 455200 }, { "epoch": 4.020293541037461, "grad_norm": 6.74933385848999, "learning_rate": 9.236522248217654e-06, "loss": 2.4415, "step": 455250 }, { "epoch": 4.020735088927745, "grad_norm": 9.037071228027344, "learning_rate": 9.228476800200042e-06, "loss": 3.0601, "step": 455300 }, { "epoch": 4.021176636818029, "grad_norm": 8.534797668457031, "learning_rate": 9.220434501436875e-06, "loss": 2.1779, "step": 455350 }, { "epoch": 4.021618184708314, "grad_norm": 3.0828628540039062, "learning_rate": 9.21239535254933e-06, "loss": 2.6566, "step": 455400 }, { "epoch": 4.022059732598597, "grad_norm": 4.421370029449463, "learning_rate": 9.20435935415837e-06, "loss": 2.6959, "step": 455450 }, { "epoch": 4.022501280488882, "grad_norm": 5.2850518226623535, "learning_rate": 9.196326506884734e-06, "loss": 2.6386, "step": 455500 }, { "epoch": 4.022942828379166, "grad_norm": 3.2356624603271484, "learning_rate": 9.188296811348863e-06, "loss": 2.6887, "step": 455550 }, { "epoch": 4.02338437626945, "grad_norm": 5.758159637451172, "learning_rate": 9.180270268170982e-06, "loss": 2.7324, "step": 455600 }, { "epoch": 4.023825924159734, "grad_norm": 6.96513032913208, "learning_rate": 9.17224687797108e-06, "loss": 3.0564, "step": 455650 }, { "epoch": 4.024267472050019, "grad_norm": 7.257087707519531, "learning_rate": 9.164226641368911e-06, "loss": 2.5713, "step": 455700 }, { "epoch": 4.024709019940302, "grad_norm": 3.0501341819763184, "learning_rate": 9.156209558983942e-06, "loss": 2.6767, "step": 455750 }, { "epoch": 4.025150567830587, "grad_norm": 4.721952438354492, "learning_rate": 9.148195631435442e-06, "loss": 2.5235, "step": 455800 }, { "epoch": 4.025592115720871, "grad_norm": 6.224329948425293, "learning_rate": 9.140184859342427e-06, "loss": 2.9104, "step": 455850 }, { "epoch": 4.026033663611155, "grad_norm": 6.836895942687988, "learning_rate": 9.132177243323652e-06, "loss": 3.142, "step": 455900 }, { "epoch": 4.026475211501439, "grad_norm": 4.6429266929626465, "learning_rate": 9.12417278399762e-06, "loss": 2.6724, "step": 455950 }, { "epoch": 4.026916759391724, "grad_norm": 3.4519004821777344, "learning_rate": 9.116171481982628e-06, "loss": 2.5277, "step": 456000 }, { "epoch": 4.026916759391724, "eval_asr_loss": 0.9660141887661328, "eval_loss": 2.7077877521514893, "eval_runtime": 20.8081, "eval_samples_per_second": 36.909, "eval_steps_per_second": 9.227, "eval_tts_loss": 6.326652641667321, "step": 456000 }, { "epoch": 4.027358307282007, "grad_norm": 12.303698539733887, "learning_rate": 9.108173337896714e-06, "loss": 2.797, "step": 456050 }, { "epoch": 4.027799855172292, "grad_norm": 4.799502372741699, "learning_rate": 9.100178352357647e-06, "loss": 2.8913, "step": 456100 }, { "epoch": 4.028241403062577, "grad_norm": 4.040663242340088, "learning_rate": 9.092186525982982e-06, "loss": 2.4341, "step": 456150 }, { "epoch": 4.02868295095286, "grad_norm": 6.532703399658203, "learning_rate": 9.084197859390032e-06, "loss": 2.7558, "step": 456200 }, { "epoch": 4.029124498843145, "grad_norm": 6.309985160827637, "learning_rate": 9.07621235319584e-06, "loss": 2.9275, "step": 456250 }, { "epoch": 4.029566046733429, "grad_norm": 2.4478442668914795, "learning_rate": 9.068230008017204e-06, "loss": 2.7939, "step": 456300 }, { "epoch": 4.030007594623713, "grad_norm": 6.549551486968994, "learning_rate": 9.060250824470729e-06, "loss": 2.8821, "step": 456350 }, { "epoch": 4.030449142513997, "grad_norm": 7.151683807373047, "learning_rate": 9.0522748031727e-06, "loss": 2.9274, "step": 456400 }, { "epoch": 4.030890690404282, "grad_norm": 7.532583713531494, "learning_rate": 9.04430194473922e-06, "loss": 2.7985, "step": 456450 }, { "epoch": 4.031332238294565, "grad_norm": 8.74457836151123, "learning_rate": 9.036332249786128e-06, "loss": 3.1279, "step": 456500 }, { "epoch": 4.03177378618485, "grad_norm": 7.153626918792725, "learning_rate": 9.028365718929e-06, "loss": 2.6415, "step": 456550 }, { "epoch": 4.032215334075134, "grad_norm": 4.8189826011657715, "learning_rate": 9.020402352783191e-06, "loss": 2.5552, "step": 456600 }, { "epoch": 4.032656881965418, "grad_norm": 3.754596710205078, "learning_rate": 9.012442151963813e-06, "loss": 2.8207, "step": 456650 }, { "epoch": 4.033098429855702, "grad_norm": 4.83393669128418, "learning_rate": 9.004485117085716e-06, "loss": 3.1147, "step": 456700 }, { "epoch": 4.033539977745987, "grad_norm": 2.628828763961792, "learning_rate": 8.996531248763501e-06, "loss": 2.988, "step": 456750 }, { "epoch": 4.03398152563627, "grad_norm": 5.186316013336182, "learning_rate": 8.988580547611546e-06, "loss": 2.607, "step": 456800 }, { "epoch": 4.034423073526555, "grad_norm": 4.674391269683838, "learning_rate": 8.98063301424399e-06, "loss": 2.7833, "step": 456850 }, { "epoch": 4.034864621416839, "grad_norm": 3.566399335861206, "learning_rate": 8.972688649274685e-06, "loss": 3.1579, "step": 456900 }, { "epoch": 4.035306169307123, "grad_norm": 6.302156925201416, "learning_rate": 8.964747453317279e-06, "loss": 2.5986, "step": 456950 }, { "epoch": 4.035747717197407, "grad_norm": 8.097691535949707, "learning_rate": 8.95680942698518e-06, "loss": 2.7729, "step": 457000 }, { "epoch": 4.036189265087692, "grad_norm": 5.321153163909912, "learning_rate": 8.948874570891518e-06, "loss": 2.3213, "step": 457050 }, { "epoch": 4.036630812977975, "grad_norm": 5.364089012145996, "learning_rate": 8.940942885649179e-06, "loss": 2.8313, "step": 457100 }, { "epoch": 4.03707236086826, "grad_norm": 3.7392337322235107, "learning_rate": 8.933014371870841e-06, "loss": 2.7237, "step": 457150 }, { "epoch": 4.037513908758544, "grad_norm": 4.230645179748535, "learning_rate": 8.925089030168893e-06, "loss": 2.6816, "step": 457200 }, { "epoch": 4.037955456648828, "grad_norm": 7.853908538818359, "learning_rate": 8.917166861155513e-06, "loss": 2.718, "step": 457250 }, { "epoch": 4.038397004539112, "grad_norm": 8.967790603637695, "learning_rate": 8.909247865442622e-06, "loss": 2.6335, "step": 457300 }, { "epoch": 4.038838552429397, "grad_norm": 5.821688175201416, "learning_rate": 8.901332043641902e-06, "loss": 2.8662, "step": 457350 }, { "epoch": 4.03928010031968, "grad_norm": 7.54954719543457, "learning_rate": 8.893419396364765e-06, "loss": 2.6551, "step": 457400 }, { "epoch": 4.039721648209965, "grad_norm": 5.742681980133057, "learning_rate": 8.885509924222424e-06, "loss": 2.7571, "step": 457450 }, { "epoch": 4.040163196100249, "grad_norm": 7.527080059051514, "learning_rate": 8.877603627825792e-06, "loss": 2.6615, "step": 457500 }, { "epoch": 4.040604743990533, "grad_norm": 4.261421203613281, "learning_rate": 8.869700507785567e-06, "loss": 3.1114, "step": 457550 }, { "epoch": 4.041046291880817, "grad_norm": 4.5753703117370605, "learning_rate": 8.861800564712198e-06, "loss": 2.6727, "step": 457600 }, { "epoch": 4.041487839771102, "grad_norm": 2.524292469024658, "learning_rate": 8.853903799215906e-06, "loss": 2.5032, "step": 457650 }, { "epoch": 4.041929387661385, "grad_norm": 4.067115783691406, "learning_rate": 8.84601021190663e-06, "loss": 2.5389, "step": 457700 }, { "epoch": 4.04237093555167, "grad_norm": 4.04695463180542, "learning_rate": 8.838119803394086e-06, "loss": 2.9212, "step": 457750 }, { "epoch": 4.0428124834419545, "grad_norm": 6.124611854553223, "learning_rate": 8.830232574287755e-06, "loss": 3.0826, "step": 457800 }, { "epoch": 4.043254031332238, "grad_norm": 5.889286994934082, "learning_rate": 8.82234852519685e-06, "loss": 2.3198, "step": 457850 }, { "epoch": 4.043695579222523, "grad_norm": 3.6657700538635254, "learning_rate": 8.81446765673033e-06, "loss": 2.9204, "step": 457900 }, { "epoch": 4.044137127112807, "grad_norm": 8.801512718200684, "learning_rate": 8.806589969496942e-06, "loss": 2.5345, "step": 457950 }, { "epoch": 4.044578675003091, "grad_norm": 4.657283782958984, "learning_rate": 8.798715464105183e-06, "loss": 2.7933, "step": 458000 }, { "epoch": 4.045020222893375, "grad_norm": 6.3618621826171875, "learning_rate": 8.790844141163268e-06, "loss": 2.4314, "step": 458050 }, { "epoch": 4.0454617707836595, "grad_norm": 2.075317859649658, "learning_rate": 8.782976001279197e-06, "loss": 2.6031, "step": 458100 }, { "epoch": 4.045903318673943, "grad_norm": 9.033692359924316, "learning_rate": 8.77511104506073e-06, "loss": 2.956, "step": 458150 }, { "epoch": 4.046344866564228, "grad_norm": 2.979217290878296, "learning_rate": 8.767249273115347e-06, "loss": 2.5556, "step": 458200 }, { "epoch": 4.046786414454512, "grad_norm": 6.154276371002197, "learning_rate": 8.759390686050328e-06, "loss": 2.629, "step": 458250 }, { "epoch": 4.047227962344796, "grad_norm": 5.7302985191345215, "learning_rate": 8.751535284472668e-06, "loss": 2.6898, "step": 458300 }, { "epoch": 4.04766951023508, "grad_norm": 4.347016334533691, "learning_rate": 8.743683068989112e-06, "loss": 2.3808, "step": 458350 }, { "epoch": 4.0481110581253645, "grad_norm": 6.965164661407471, "learning_rate": 8.735834040206203e-06, "loss": 3.0225, "step": 458400 }, { "epoch": 4.048552606015648, "grad_norm": 6.444103240966797, "learning_rate": 8.727988198730203e-06, "loss": 2.4737, "step": 458450 }, { "epoch": 4.048994153905933, "grad_norm": 6.89586877822876, "learning_rate": 8.720145545167147e-06, "loss": 2.7857, "step": 458500 }, { "epoch": 4.049435701796217, "grad_norm": 7.1288161277771, "learning_rate": 8.712306080122796e-06, "loss": 2.869, "step": 458550 }, { "epoch": 4.049877249686501, "grad_norm": 1.573258876800537, "learning_rate": 8.704469804202702e-06, "loss": 2.6429, "step": 458600 }, { "epoch": 4.050318797576785, "grad_norm": 4.737966060638428, "learning_rate": 8.69663671801214e-06, "loss": 2.2791, "step": 458650 }, { "epoch": 4.0507603454670695, "grad_norm": 6.200699806213379, "learning_rate": 8.688806822156137e-06, "loss": 3.1388, "step": 458700 }, { "epoch": 4.051201893357353, "grad_norm": 2.0771617889404297, "learning_rate": 8.680980117239501e-06, "loss": 2.6256, "step": 458750 }, { "epoch": 4.051643441247638, "grad_norm": 4.721651077270508, "learning_rate": 8.673156603866783e-06, "loss": 2.9465, "step": 458800 }, { "epoch": 4.052084989137922, "grad_norm": 1.9940071105957031, "learning_rate": 8.665336282642267e-06, "loss": 3.2017, "step": 458850 }, { "epoch": 4.052526537028206, "grad_norm": 3.7781293392181396, "learning_rate": 8.657519154170013e-06, "loss": 3.0513, "step": 458900 }, { "epoch": 4.05296808491849, "grad_norm": 10.600885391235352, "learning_rate": 8.649705219053843e-06, "loss": 2.3409, "step": 458950 }, { "epoch": 4.0534096328087745, "grad_norm": 3.542048454284668, "learning_rate": 8.641894477897305e-06, "loss": 2.3274, "step": 459000 }, { "epoch": 4.0534096328087745, "eval_asr_loss": 0.9723407348578477, "eval_loss": 2.7028610706329346, "eval_runtime": 21.5933, "eval_samples_per_second": 35.567, "eval_steps_per_second": 8.892, "eval_tts_loss": 6.323471203274506, "step": 459000 }, { "epoch": 4.053851180699058, "grad_norm": 4.878571033477783, "learning_rate": 8.634086931303704e-06, "loss": 2.9445, "step": 459050 }, { "epoch": 4.054292728589343, "grad_norm": 8.753440856933594, "learning_rate": 8.626282579876111e-06, "loss": 2.583, "step": 459100 }, { "epoch": 4.054734276479627, "grad_norm": 7.7879486083984375, "learning_rate": 8.618481424217361e-06, "loss": 3.1433, "step": 459150 }, { "epoch": 4.055175824369911, "grad_norm": 4.9998459815979, "learning_rate": 8.610683464930009e-06, "loss": 2.4239, "step": 459200 }, { "epoch": 4.055617372260195, "grad_norm": 6.624985694885254, "learning_rate": 8.60288870261638e-06, "loss": 2.6083, "step": 459250 }, { "epoch": 4.0560589201504795, "grad_norm": 3.359468936920166, "learning_rate": 8.595097137878577e-06, "loss": 2.428, "step": 459300 }, { "epoch": 4.056500468040764, "grad_norm": 5.817551136016846, "learning_rate": 8.587308771318409e-06, "loss": 2.6849, "step": 459350 }, { "epoch": 4.056942015931048, "grad_norm": 3.460698127746582, "learning_rate": 8.579523603537476e-06, "loss": 2.8584, "step": 459400 }, { "epoch": 4.0573835638213325, "grad_norm": 4.248012542724609, "learning_rate": 8.57174163513711e-06, "loss": 2.3807, "step": 459450 }, { "epoch": 4.057825111711616, "grad_norm": 4.066163063049316, "learning_rate": 8.563962866718395e-06, "loss": 2.6229, "step": 459500 }, { "epoch": 4.058266659601901, "grad_norm": 13.42602825164795, "learning_rate": 8.556187298882173e-06, "loss": 3.0384, "step": 459550 }, { "epoch": 4.0587082074921845, "grad_norm": 2.794247627258301, "learning_rate": 8.548414932229054e-06, "loss": 2.7356, "step": 459600 }, { "epoch": 4.059149755382469, "grad_norm": 4.537113666534424, "learning_rate": 8.540645767359389e-06, "loss": 2.8332, "step": 459650 }, { "epoch": 4.059591303272753, "grad_norm": 6.820173263549805, "learning_rate": 8.532879804873266e-06, "loss": 2.801, "step": 459700 }, { "epoch": 4.0600328511630375, "grad_norm": 9.09410285949707, "learning_rate": 8.525117045370561e-06, "loss": 2.9935, "step": 459750 }, { "epoch": 4.060474399053321, "grad_norm": 6.629907608032227, "learning_rate": 8.517357489450867e-06, "loss": 2.6712, "step": 459800 }, { "epoch": 4.060915946943606, "grad_norm": 4.346859455108643, "learning_rate": 8.50960113771353e-06, "loss": 2.7745, "step": 459850 }, { "epoch": 4.0613574948338895, "grad_norm": 4.040051460266113, "learning_rate": 8.501847990757671e-06, "loss": 2.6664, "step": 459900 }, { "epoch": 4.061799042724174, "grad_norm": 5.115642547607422, "learning_rate": 8.49409804918218e-06, "loss": 2.6104, "step": 459950 }, { "epoch": 4.062240590614458, "grad_norm": 6.057486534118652, "learning_rate": 8.486351313585644e-06, "loss": 2.7344, "step": 460000 }, { "epoch": 4.0626821385047425, "grad_norm": 6.923796653747559, "learning_rate": 8.478607784566439e-06, "loss": 2.7301, "step": 460050 }, { "epoch": 4.063123686395026, "grad_norm": 8.9944486618042, "learning_rate": 8.470867462722708e-06, "loss": 2.3612, "step": 460100 }, { "epoch": 4.063565234285311, "grad_norm": 4.291893482208252, "learning_rate": 8.463130348652292e-06, "loss": 2.5722, "step": 460150 }, { "epoch": 4.0640067821755945, "grad_norm": 7.711345195770264, "learning_rate": 8.45539644295285e-06, "loss": 2.3756, "step": 460200 }, { "epoch": 4.064448330065879, "grad_norm": 4.425571441650391, "learning_rate": 8.447665746221733e-06, "loss": 2.6375, "step": 460250 }, { "epoch": 4.064889877956163, "grad_norm": 3.7058401107788086, "learning_rate": 8.439938259056097e-06, "loss": 2.6217, "step": 460300 }, { "epoch": 4.0653314258464475, "grad_norm": 8.189947128295898, "learning_rate": 8.432213982052795e-06, "loss": 2.5949, "step": 460350 }, { "epoch": 4.065772973736731, "grad_norm": 3.692025661468506, "learning_rate": 8.424492915808479e-06, "loss": 2.4207, "step": 460400 }, { "epoch": 4.066214521627016, "grad_norm": 4.9508538246154785, "learning_rate": 8.416775060919547e-06, "loss": 2.8316, "step": 460450 }, { "epoch": 4.0666560695172995, "grad_norm": 5.651669502258301, "learning_rate": 8.409060417982117e-06, "loss": 2.3219, "step": 460500 }, { "epoch": 4.067097617407584, "grad_norm": 5.035886287689209, "learning_rate": 8.4013489875921e-06, "loss": 2.7974, "step": 460550 }, { "epoch": 4.067539165297868, "grad_norm": 6.807248592376709, "learning_rate": 8.393640770345123e-06, "loss": 2.7234, "step": 460600 }, { "epoch": 4.0679807131881525, "grad_norm": 7.744897842407227, "learning_rate": 8.385935766836577e-06, "loss": 2.678, "step": 460650 }, { "epoch": 4.068422261078436, "grad_norm": 6.3183674812316895, "learning_rate": 8.378233977661615e-06, "loss": 3.1765, "step": 460700 }, { "epoch": 4.068863808968721, "grad_norm": 4.689250946044922, "learning_rate": 8.370535403415136e-06, "loss": 2.7854, "step": 460750 }, { "epoch": 4.0693053568590045, "grad_norm": 7.189305305480957, "learning_rate": 8.362840044691805e-06, "loss": 2.7028, "step": 460800 }, { "epoch": 4.069746904749289, "grad_norm": 10.072436332702637, "learning_rate": 8.355147902085992e-06, "loss": 2.19, "step": 460850 }, { "epoch": 4.070188452639573, "grad_norm": 5.519792556762695, "learning_rate": 8.34745897619188e-06, "loss": 2.4687, "step": 460900 }, { "epoch": 4.0706300005298575, "grad_norm": 4.414450645446777, "learning_rate": 8.339773267603357e-06, "loss": 2.9262, "step": 460950 }, { "epoch": 4.071071548420142, "grad_norm": 5.777186393737793, "learning_rate": 8.33209077691407e-06, "loss": 2.8914, "step": 461000 }, { "epoch": 4.071513096310426, "grad_norm": 6.718722343444824, "learning_rate": 8.324411504717432e-06, "loss": 2.2102, "step": 461050 }, { "epoch": 4.07195464420071, "grad_norm": 5.735861301422119, "learning_rate": 8.316735451606622e-06, "loss": 2.8058, "step": 461100 }, { "epoch": 4.072396192090994, "grad_norm": 4.646334171295166, "learning_rate": 8.309062618174524e-06, "loss": 2.3716, "step": 461150 }, { "epoch": 4.072837739981279, "grad_norm": 8.482443809509277, "learning_rate": 8.301393005013807e-06, "loss": 2.6159, "step": 461200 }, { "epoch": 4.0732792878715625, "grad_norm": 6.166719436645508, "learning_rate": 8.293726612716906e-06, "loss": 2.7951, "step": 461250 }, { "epoch": 4.073720835761847, "grad_norm": 3.22122859954834, "learning_rate": 8.286063441875946e-06, "loss": 2.9526, "step": 461300 }, { "epoch": 4.074162383652131, "grad_norm": 4.848860740661621, "learning_rate": 8.278403493082876e-06, "loss": 2.4386, "step": 461350 }, { "epoch": 4.074603931542415, "grad_norm": 7.187008857727051, "learning_rate": 8.270746766929332e-06, "loss": 2.7829, "step": 461400 }, { "epoch": 4.075045479432699, "grad_norm": 6.120682716369629, "learning_rate": 8.263093264006761e-06, "loss": 2.5622, "step": 461450 }, { "epoch": 4.075487027322984, "grad_norm": 6.608443260192871, "learning_rate": 8.255442984906303e-06, "loss": 2.7193, "step": 461500 }, { "epoch": 4.0759285752132675, "grad_norm": 2.1796798706054688, "learning_rate": 8.247795930218893e-06, "loss": 2.4816, "step": 461550 }, { "epoch": 4.076370123103552, "grad_norm": 3.5393970012664795, "learning_rate": 8.240152100535214e-06, "loss": 3.3655, "step": 461600 }, { "epoch": 4.076811670993836, "grad_norm": 4.5318684577941895, "learning_rate": 8.232511496445656e-06, "loss": 2.8205, "step": 461650 }, { "epoch": 4.07725321888412, "grad_norm": 0.9510912895202637, "learning_rate": 8.224874118540422e-06, "loss": 2.6612, "step": 461700 }, { "epoch": 4.077694766774404, "grad_norm": 2.7017970085144043, "learning_rate": 8.217239967409412e-06, "loss": 2.8, "step": 461750 }, { "epoch": 4.078136314664689, "grad_norm": 3.8348515033721924, "learning_rate": 8.209609043642302e-06, "loss": 2.7687, "step": 461800 }, { "epoch": 4.0785778625549725, "grad_norm": 5.709969520568848, "learning_rate": 8.201981347828518e-06, "loss": 2.4728, "step": 461850 }, { "epoch": 4.079019410445257, "grad_norm": 4.0043439865112305, "learning_rate": 8.194356880557241e-06, "loss": 2.8284, "step": 461900 }, { "epoch": 4.079460958335541, "grad_norm": 5.0206522941589355, "learning_rate": 8.186735642417404e-06, "loss": 2.2435, "step": 461950 }, { "epoch": 4.079902506225825, "grad_norm": 5.378378868103027, "learning_rate": 8.179117633997663e-06, "loss": 2.6134, "step": 462000 }, { "epoch": 4.079902506225825, "eval_asr_loss": 0.9785765007830401, "eval_loss": 2.7110588550567627, "eval_runtime": 20.704, "eval_samples_per_second": 37.094, "eval_steps_per_second": 9.274, "eval_tts_loss": 6.310662870131224, "step": 462000 }, { "epoch": 4.080344054116109, "grad_norm": 6.026256561279297, "learning_rate": 8.171502855886447e-06, "loss": 2.9305, "step": 462050 }, { "epoch": 4.080785602006394, "grad_norm": 6.1845879554748535, "learning_rate": 8.16389130867196e-06, "loss": 3.0822, "step": 462100 }, { "epoch": 4.0812271498966775, "grad_norm": 5.573923587799072, "learning_rate": 8.156282992942105e-06, "loss": 2.5087, "step": 462150 }, { "epoch": 4.081668697786962, "grad_norm": 5.229231834411621, "learning_rate": 8.148677909284552e-06, "loss": 3.421, "step": 462200 }, { "epoch": 4.082110245677246, "grad_norm": 5.51962947845459, "learning_rate": 8.141076058286756e-06, "loss": 2.9282, "step": 462250 }, { "epoch": 4.08255179356753, "grad_norm": 3.859940528869629, "learning_rate": 8.133477440535869e-06, "loss": 2.7686, "step": 462300 }, { "epoch": 4.082993341457814, "grad_norm": 8.358874320983887, "learning_rate": 8.125882056618833e-06, "loss": 2.5025, "step": 462350 }, { "epoch": 4.083434889348099, "grad_norm": 6.8751444816589355, "learning_rate": 8.118289907122333e-06, "loss": 2.7459, "step": 462400 }, { "epoch": 4.0838764372383825, "grad_norm": 8.785126686096191, "learning_rate": 8.110700992632786e-06, "loss": 2.5288, "step": 462450 }, { "epoch": 4.084317985128667, "grad_norm": 2.7108919620513916, "learning_rate": 8.103115313736387e-06, "loss": 3.1286, "step": 462500 }, { "epoch": 4.084759533018952, "grad_norm": 4.292723178863525, "learning_rate": 8.09553287101904e-06, "loss": 2.6477, "step": 462550 }, { "epoch": 4.085201080909235, "grad_norm": 4.92633056640625, "learning_rate": 8.087953665066451e-06, "loss": 2.8784, "step": 462600 }, { "epoch": 4.08564262879952, "grad_norm": 5.2098541259765625, "learning_rate": 8.080377696464025e-06, "loss": 2.8549, "step": 462650 }, { "epoch": 4.086084176689804, "grad_norm": 6.181698322296143, "learning_rate": 8.072804965796959e-06, "loss": 2.6809, "step": 462700 }, { "epoch": 4.086525724580088, "grad_norm": 2.7372639179229736, "learning_rate": 8.065235473650185e-06, "loss": 2.8136, "step": 462750 }, { "epoch": 4.086967272470372, "grad_norm": 2.5620012283325195, "learning_rate": 8.05766922060836e-06, "loss": 2.6113, "step": 462800 }, { "epoch": 4.087408820360657, "grad_norm": 2.185894727706909, "learning_rate": 8.050106207255947e-06, "loss": 2.84, "step": 462850 }, { "epoch": 4.08785036825094, "grad_norm": 5.010711669921875, "learning_rate": 8.042546434177083e-06, "loss": 2.7924, "step": 462900 }, { "epoch": 4.088291916141225, "grad_norm": 2.857395648956299, "learning_rate": 8.034989901955736e-06, "loss": 2.5159, "step": 462950 }, { "epoch": 4.088733464031509, "grad_norm": 4.299446105957031, "learning_rate": 8.027436611175555e-06, "loss": 2.6593, "step": 463000 }, { "epoch": 4.089175011921793, "grad_norm": 5.341146469116211, "learning_rate": 8.019886562419986e-06, "loss": 2.8687, "step": 463050 }, { "epoch": 4.089616559812077, "grad_norm": 3.1914913654327393, "learning_rate": 8.012490660613814e-06, "loss": 2.6298, "step": 463100 }, { "epoch": 4.090058107702362, "grad_norm": 5.0616278648376465, "learning_rate": 8.004947032787213e-06, "loss": 2.9335, "step": 463150 }, { "epoch": 4.090499655592645, "grad_norm": 4.5667033195495605, "learning_rate": 7.99740664872235e-06, "loss": 2.5354, "step": 463200 }, { "epoch": 4.09094120348293, "grad_norm": 6.8408427238464355, "learning_rate": 7.989869509001635e-06, "loss": 2.618, "step": 463250 }, { "epoch": 4.091382751373214, "grad_norm": 6.201191425323486, "learning_rate": 7.982335614207275e-06, "loss": 2.753, "step": 463300 }, { "epoch": 4.091824299263498, "grad_norm": 6.780299663543701, "learning_rate": 7.974804964921185e-06, "loss": 2.6992, "step": 463350 }, { "epoch": 4.092265847153782, "grad_norm": 3.5981030464172363, "learning_rate": 7.967277561725045e-06, "loss": 3.0469, "step": 463400 }, { "epoch": 4.092707395044067, "grad_norm": 8.066826820373535, "learning_rate": 7.959753405200294e-06, "loss": 2.727, "step": 463450 }, { "epoch": 4.09314894293435, "grad_norm": 8.693765640258789, "learning_rate": 7.952232495928096e-06, "loss": 2.7761, "step": 463500 }, { "epoch": 4.093590490824635, "grad_norm": 6.3067731857299805, "learning_rate": 7.94471483448938e-06, "loss": 2.6413, "step": 463550 }, { "epoch": 4.094032038714919, "grad_norm": 6.419362545013428, "learning_rate": 7.937200421464835e-06, "loss": 2.3912, "step": 463600 }, { "epoch": 4.094473586605203, "grad_norm": 6.690420150756836, "learning_rate": 7.92968925743488e-06, "loss": 2.6549, "step": 463650 }, { "epoch": 4.094915134495487, "grad_norm": 6.553482532501221, "learning_rate": 7.922181342979668e-06, "loss": 2.6324, "step": 463700 }, { "epoch": 4.095356682385772, "grad_norm": 9.438315391540527, "learning_rate": 7.91467667867914e-06, "loss": 2.8105, "step": 463750 }, { "epoch": 4.095798230276055, "grad_norm": 4.824997901916504, "learning_rate": 7.907175265112975e-06, "loss": 2.7162, "step": 463800 }, { "epoch": 4.09623977816634, "grad_norm": 4.247918128967285, "learning_rate": 7.899677102860575e-06, "loss": 2.5887, "step": 463850 }, { "epoch": 4.096681326056624, "grad_norm": 7.048587799072266, "learning_rate": 7.892182192501113e-06, "loss": 2.561, "step": 463900 }, { "epoch": 4.097122873946908, "grad_norm": 7.232861042022705, "learning_rate": 7.884690534613532e-06, "loss": 2.9755, "step": 463950 }, { "epoch": 4.097564421837192, "grad_norm": 3.828396797180176, "learning_rate": 7.87720212977648e-06, "loss": 2.8165, "step": 464000 }, { "epoch": 4.098005969727477, "grad_norm": 5.158803939819336, "learning_rate": 7.869716978568354e-06, "loss": 2.6782, "step": 464050 }, { "epoch": 4.09844751761776, "grad_norm": 4.070273399353027, "learning_rate": 7.862235081567338e-06, "loss": 2.7344, "step": 464100 }, { "epoch": 4.098889065508045, "grad_norm": 3.077816963195801, "learning_rate": 7.854756439351358e-06, "loss": 2.9348, "step": 464150 }, { "epoch": 4.09933061339833, "grad_norm": 3.7023746967315674, "learning_rate": 7.847281052498052e-06, "loss": 2.3384, "step": 464200 }, { "epoch": 4.099772161288613, "grad_norm": 6.183480739593506, "learning_rate": 7.839808921584841e-06, "loss": 3.0359, "step": 464250 }, { "epoch": 4.100213709178898, "grad_norm": 3.834052324295044, "learning_rate": 7.83234004718889e-06, "loss": 2.6923, "step": 464300 }, { "epoch": 4.100655257069182, "grad_norm": 7.892953395843506, "learning_rate": 7.824874429887085e-06, "loss": 2.8465, "step": 464350 }, { "epoch": 4.101096804959466, "grad_norm": 2.789435863494873, "learning_rate": 7.817412070256113e-06, "loss": 2.7638, "step": 464400 }, { "epoch": 4.10153835284975, "grad_norm": 7.383251190185547, "learning_rate": 7.809952968872364e-06, "loss": 2.9393, "step": 464450 }, { "epoch": 4.101979900740035, "grad_norm": 4.807004928588867, "learning_rate": 7.80249712631197e-06, "loss": 2.9306, "step": 464500 }, { "epoch": 4.102421448630318, "grad_norm": 7.703232765197754, "learning_rate": 7.795044543150847e-06, "loss": 2.5277, "step": 464550 }, { "epoch": 4.102862996520603, "grad_norm": 11.057501792907715, "learning_rate": 7.78759521996466e-06, "loss": 3.0118, "step": 464600 }, { "epoch": 4.103304544410887, "grad_norm": 2.9913408756256104, "learning_rate": 7.780149157328782e-06, "loss": 2.3577, "step": 464650 }, { "epoch": 4.103746092301171, "grad_norm": 3.812255620956421, "learning_rate": 7.77270635581836e-06, "loss": 2.6751, "step": 464700 }, { "epoch": 4.104187640191455, "grad_norm": 4.6745991706848145, "learning_rate": 7.765266816008316e-06, "loss": 3.3268, "step": 464750 }, { "epoch": 4.10462918808174, "grad_norm": 2.266587257385254, "learning_rate": 7.757830538473265e-06, "loss": 2.5126, "step": 464800 }, { "epoch": 4.105070735972023, "grad_norm": 6.9605889320373535, "learning_rate": 7.75039752378759e-06, "loss": 2.4065, "step": 464850 }, { "epoch": 4.105512283862308, "grad_norm": 4.890602111816406, "learning_rate": 7.742967772525439e-06, "loss": 2.8532, "step": 464900 }, { "epoch": 4.105953831752592, "grad_norm": 7.440773963928223, "learning_rate": 7.735541285260716e-06, "loss": 2.7777, "step": 464950 }, { "epoch": 4.106395379642876, "grad_norm": 5.4694647789001465, "learning_rate": 7.72811806256702e-06, "loss": 2.2553, "step": 465000 }, { "epoch": 4.106395379642876, "eval_asr_loss": 0.9753628012309021, "eval_loss": 2.7072436809539795, "eval_runtime": 21.622, "eval_samples_per_second": 35.519, "eval_steps_per_second": 8.88, "eval_tts_loss": 6.339440466846328, "step": 465000 }, { "epoch": 4.10683692753316, "grad_norm": 6.588335037231445, "learning_rate": 7.720698105017749e-06, "loss": 2.7364, "step": 465050 }, { "epoch": 4.107278475423445, "grad_norm": 3.3337783813476562, "learning_rate": 7.713281413186047e-06, "loss": 2.8822, "step": 465100 }, { "epoch": 4.107720023313728, "grad_norm": 6.051248073577881, "learning_rate": 7.705867987644772e-06, "loss": 2.9577, "step": 465150 }, { "epoch": 4.108161571204013, "grad_norm": 7.353286266326904, "learning_rate": 7.698606000121145e-06, "loss": 2.9827, "step": 465200 }, { "epoch": 4.108603119094297, "grad_norm": 3.926074504852295, "learning_rate": 7.691199043524023e-06, "loss": 2.4984, "step": 465250 }, { "epoch": 4.109044666984581, "grad_norm": 3.295396566390991, "learning_rate": 7.683795354923007e-06, "loss": 2.8744, "step": 465300 }, { "epoch": 4.109486214874865, "grad_norm": 5.530523300170898, "learning_rate": 7.676394934889974e-06, "loss": 2.398, "step": 465350 }, { "epoch": 4.10992776276515, "grad_norm": 7.04426383972168, "learning_rate": 7.668997783996534e-06, "loss": 2.4167, "step": 465400 }, { "epoch": 4.110369310655433, "grad_norm": 5.467502593994141, "learning_rate": 7.661603902814068e-06, "loss": 2.5317, "step": 465450 }, { "epoch": 4.110810858545718, "grad_norm": 5.170390605926514, "learning_rate": 7.654213291913687e-06, "loss": 2.6389, "step": 465500 }, { "epoch": 4.111252406436002, "grad_norm": 2.60446834564209, "learning_rate": 7.646825951866227e-06, "loss": 2.6595, "step": 465550 }, { "epoch": 4.111693954326286, "grad_norm": 3.524770736694336, "learning_rate": 7.639441883242327e-06, "loss": 2.3705, "step": 465600 }, { "epoch": 4.11213550221657, "grad_norm": 8.007800102233887, "learning_rate": 7.63206108661234e-06, "loss": 2.6027, "step": 465650 }, { "epoch": 4.112577050106855, "grad_norm": 2.416206121444702, "learning_rate": 7.624683562546353e-06, "loss": 2.7484, "step": 465700 }, { "epoch": 4.113018597997139, "grad_norm": 5.672489166259766, "learning_rate": 7.617309311614229e-06, "loss": 2.5926, "step": 465750 }, { "epoch": 4.113460145887423, "grad_norm": 6.713701248168945, "learning_rate": 7.6099383343855805e-06, "loss": 2.7431, "step": 465800 }, { "epoch": 4.1139016937777075, "grad_norm": 5.965285301208496, "learning_rate": 7.602570631429723e-06, "loss": 2.8389, "step": 465850 }, { "epoch": 4.114343241667991, "grad_norm": 4.208332061767578, "learning_rate": 7.595206203315774e-06, "loss": 2.3954, "step": 465900 }, { "epoch": 4.114784789558276, "grad_norm": 8.531694412231445, "learning_rate": 7.5878450506125574e-06, "loss": 2.8076, "step": 465950 }, { "epoch": 4.11522633744856, "grad_norm": 7.047206401824951, "learning_rate": 7.580487173888679e-06, "loss": 2.4711, "step": 466000 }, { "epoch": 4.115667885338844, "grad_norm": 7.656429290771484, "learning_rate": 7.5731325737124435e-06, "loss": 2.468, "step": 466050 }, { "epoch": 4.116109433229128, "grad_norm": 7.568843841552734, "learning_rate": 7.5657812506519486e-06, "loss": 2.9342, "step": 466100 }, { "epoch": 4.1165509811194125, "grad_norm": 3.5093424320220947, "learning_rate": 7.558433205275029e-06, "loss": 3.2045, "step": 466150 }, { "epoch": 4.116992529009696, "grad_norm": 8.915587425231934, "learning_rate": 7.551088438149245e-06, "loss": 2.9646, "step": 466200 }, { "epoch": 4.117434076899981, "grad_norm": 5.471665859222412, "learning_rate": 7.543746949841929e-06, "loss": 3.0202, "step": 466250 }, { "epoch": 4.117875624790265, "grad_norm": 7.755118370056152, "learning_rate": 7.5364087409201364e-06, "loss": 2.9328, "step": 466300 }, { "epoch": 4.118317172680549, "grad_norm": 4.777098655700684, "learning_rate": 7.529073811950693e-06, "loss": 2.7857, "step": 466350 }, { "epoch": 4.118758720570833, "grad_norm": 8.01327896118164, "learning_rate": 7.521742163500145e-06, "loss": 2.812, "step": 466400 }, { "epoch": 4.1192002684611175, "grad_norm": 5.998462677001953, "learning_rate": 7.514413796134811e-06, "loss": 2.7437, "step": 466450 }, { "epoch": 4.119641816351401, "grad_norm": 3.8137874603271484, "learning_rate": 7.507088710420756e-06, "loss": 2.345, "step": 466500 }, { "epoch": 4.120083364241686, "grad_norm": 6.222236633300781, "learning_rate": 7.499766906923761e-06, "loss": 2.7633, "step": 466550 }, { "epoch": 4.12052491213197, "grad_norm": 6.194087505340576, "learning_rate": 7.492448386209389e-06, "loss": 3.0252, "step": 466600 }, { "epoch": 4.120966460022254, "grad_norm": 3.223278760910034, "learning_rate": 7.48527942140978e-06, "loss": 2.5436, "step": 466650 }, { "epoch": 4.121408007912538, "grad_norm": 4.556787967681885, "learning_rate": 7.477967402272468e-06, "loss": 2.6607, "step": 466700 }, { "epoch": 4.1218495558028225, "grad_norm": 5.454936504364014, "learning_rate": 7.470658667601588e-06, "loss": 2.5999, "step": 466750 }, { "epoch": 4.122291103693106, "grad_norm": 3.3251094818115234, "learning_rate": 7.463353217961689e-06, "loss": 2.5824, "step": 466800 }, { "epoch": 4.122732651583391, "grad_norm": 7.876509666442871, "learning_rate": 7.456051053917051e-06, "loss": 2.8016, "step": 466850 }, { "epoch": 4.123174199473675, "grad_norm": 5.305534839630127, "learning_rate": 7.4487521760317204e-06, "loss": 2.5406, "step": 466900 }, { "epoch": 4.123615747363959, "grad_norm": 3.213533878326416, "learning_rate": 7.4414565848694606e-06, "loss": 2.672, "step": 466950 }, { "epoch": 4.124057295254243, "grad_norm": 3.7452774047851562, "learning_rate": 7.434164280993783e-06, "loss": 2.8812, "step": 467000 }, { "epoch": 4.1244988431445275, "grad_norm": 4.279845714569092, "learning_rate": 7.426875264967975e-06, "loss": 2.9186, "step": 467050 }, { "epoch": 4.124940391034811, "grad_norm": 4.5049729347229, "learning_rate": 7.419589537355037e-06, "loss": 2.9423, "step": 467100 }, { "epoch": 4.125381938925096, "grad_norm": 3.3989970684051514, "learning_rate": 7.41230709871773e-06, "loss": 2.7702, "step": 467150 }, { "epoch": 4.12582348681538, "grad_norm": 4.809698581695557, "learning_rate": 7.405027949618581e-06, "loss": 2.4612, "step": 467200 }, { "epoch": 4.126265034705664, "grad_norm": 5.697405815124512, "learning_rate": 7.397752090619819e-06, "loss": 2.7015, "step": 467250 }, { "epoch": 4.126706582595949, "grad_norm": 3.6514995098114014, "learning_rate": 7.39047952228345e-06, "loss": 3.1084, "step": 467300 }, { "epoch": 4.1271481304862325, "grad_norm": 4.991930961608887, "learning_rate": 7.383210245171229e-06, "loss": 2.6376, "step": 467350 }, { "epoch": 4.127589678376517, "grad_norm": 7.664635181427002, "learning_rate": 7.375944259844636e-06, "loss": 2.7709, "step": 467400 }, { "epoch": 4.128031226266801, "grad_norm": 3.653046131134033, "learning_rate": 7.368826788455868e-06, "loss": 2.7232, "step": 467450 }, { "epoch": 4.1284727741570855, "grad_norm": 4.712788105010986, "learning_rate": 7.3615673225203245e-06, "loss": 2.7005, "step": 467500 }, { "epoch": 4.128914322047369, "grad_norm": 6.928778171539307, "learning_rate": 7.3543111500421365e-06, "loss": 2.8016, "step": 467550 }, { "epoch": 4.129355869937654, "grad_norm": 4.1923980712890625, "learning_rate": 7.347058271581803e-06, "loss": 2.5454, "step": 467600 }, { "epoch": 4.1297974178279375, "grad_norm": 3.4680938720703125, "learning_rate": 7.339808687699534e-06, "loss": 2.6569, "step": 467650 }, { "epoch": 4.130238965718222, "grad_norm": 4.227901458740234, "learning_rate": 7.332562398955278e-06, "loss": 2.5127, "step": 467700 }, { "epoch": 4.130680513608506, "grad_norm": 3.6537363529205322, "learning_rate": 7.325319405908776e-06, "loss": 2.6761, "step": 467750 }, { "epoch": 4.1311220614987905, "grad_norm": 7.036236763000488, "learning_rate": 7.318079709119485e-06, "loss": 2.624, "step": 467800 }, { "epoch": 4.131563609389074, "grad_norm": 4.8166046142578125, "learning_rate": 7.3108433091465924e-06, "loss": 2.705, "step": 467850 }, { "epoch": 4.132005157279359, "grad_norm": 6.212154865264893, "learning_rate": 7.303610206549061e-06, "loss": 2.8542, "step": 467900 }, { "epoch": 4.1324467051696425, "grad_norm": 6.156539440155029, "learning_rate": 7.2963804018856e-06, "loss": 3.1004, "step": 467950 }, { "epoch": 4.132888253059927, "grad_norm": 9.386872291564941, "learning_rate": 7.289153895714624e-06, "loss": 2.8604, "step": 468000 }, { "epoch": 4.132888253059927, "eval_asr_loss": 0.9852790645260793, "eval_loss": 2.7094314098358154, "eval_runtime": 52.0183, "eval_samples_per_second": 14.764, "eval_steps_per_second": 3.691, "eval_tts_loss": 6.318556014553854, "step": 468000 }, { "epoch": 4.133329800950211, "grad_norm": 7.818303108215332, "learning_rate": 7.28193068859434e-06, "loss": 2.7741, "step": 468050 }, { "epoch": 4.1337713488404955, "grad_norm": 6.205731391906738, "learning_rate": 7.274710781082677e-06, "loss": 2.6235, "step": 468100 }, { "epoch": 4.134212896730779, "grad_norm": 3.1467719078063965, "learning_rate": 7.267494173737293e-06, "loss": 2.707, "step": 468150 }, { "epoch": 4.134654444621064, "grad_norm": 7.937070369720459, "learning_rate": 7.260280867115627e-06, "loss": 2.8541, "step": 468200 }, { "epoch": 4.1350959925113475, "grad_norm": 2.5890417098999023, "learning_rate": 7.253070861774846e-06, "loss": 2.5744, "step": 468250 }, { "epoch": 4.135537540401632, "grad_norm": 10.18446159362793, "learning_rate": 7.245864158271854e-06, "loss": 2.6933, "step": 468300 }, { "epoch": 4.135979088291916, "grad_norm": 2.602311134338379, "learning_rate": 7.238660757163307e-06, "loss": 3.1308, "step": 468350 }, { "epoch": 4.1364206361822005, "grad_norm": 3.8505043983459473, "learning_rate": 7.231460659005617e-06, "loss": 3.0913, "step": 468400 }, { "epoch": 4.136862184072484, "grad_norm": 7.749487400054932, "learning_rate": 7.2242638643549295e-06, "loss": 2.4718, "step": 468450 }, { "epoch": 4.137303731962769, "grad_norm": 7.044772148132324, "learning_rate": 7.217070373767121e-06, "loss": 2.9599, "step": 468500 }, { "epoch": 4.1377452798530525, "grad_norm": 4.702652931213379, "learning_rate": 7.20988018779784e-06, "loss": 2.6413, "step": 468550 }, { "epoch": 4.138186827743337, "grad_norm": 6.674304008483887, "learning_rate": 7.2026933070024725e-06, "loss": 2.6774, "step": 468600 }, { "epoch": 4.138628375633621, "grad_norm": 4.0738606452941895, "learning_rate": 7.195509731936123e-06, "loss": 3.0155, "step": 468650 }, { "epoch": 4.1390699235239055, "grad_norm": 4.626620769500732, "learning_rate": 7.188329463153676e-06, "loss": 2.6977, "step": 468700 }, { "epoch": 4.139511471414189, "grad_norm": 6.465646266937256, "learning_rate": 7.181152501209759e-06, "loss": 2.5471, "step": 468750 }, { "epoch": 4.139953019304474, "grad_norm": 6.23164701461792, "learning_rate": 7.173978846658708e-06, "loss": 2.75, "step": 468800 }, { "epoch": 4.1403945671947575, "grad_norm": 8.240525245666504, "learning_rate": 7.166808500054645e-06, "loss": 2.7787, "step": 468850 }, { "epoch": 4.140836115085042, "grad_norm": 4.185844421386719, "learning_rate": 7.1596414619514e-06, "loss": 2.5205, "step": 468900 }, { "epoch": 4.141277662975327, "grad_norm": 4.332461357116699, "learning_rate": 7.1524777329025905e-06, "loss": 2.4171, "step": 468950 }, { "epoch": 4.1417192108656105, "grad_norm": 6.045807361602783, "learning_rate": 7.145317313461525e-06, "loss": 2.6539, "step": 469000 }, { "epoch": 4.142160758755895, "grad_norm": 5.827010154724121, "learning_rate": 7.138160204181299e-06, "loss": 2.446, "step": 469050 }, { "epoch": 4.142602306646179, "grad_norm": 3.342780113220215, "learning_rate": 7.131006405614754e-06, "loss": 2.6235, "step": 469100 }, { "epoch": 4.143043854536463, "grad_norm": 7.843532562255859, "learning_rate": 7.123855918314431e-06, "loss": 2.7664, "step": 469150 }, { "epoch": 4.143485402426747, "grad_norm": 5.514000415802002, "learning_rate": 7.116708742832678e-06, "loss": 2.678, "step": 469200 }, { "epoch": 4.143926950317032, "grad_norm": 5.983071804046631, "learning_rate": 7.109564879721525e-06, "loss": 3.1049, "step": 469250 }, { "epoch": 4.1443684982073155, "grad_norm": 8.876703262329102, "learning_rate": 7.102424329532781e-06, "loss": 2.6359, "step": 469300 }, { "epoch": 4.1448100460976, "grad_norm": 6.407278060913086, "learning_rate": 7.095287092817992e-06, "loss": 2.9142, "step": 469350 }, { "epoch": 4.145251593987884, "grad_norm": 2.6956801414489746, "learning_rate": 7.08815317012847e-06, "loss": 2.6139, "step": 469400 }, { "epoch": 4.145693141878168, "grad_norm": 4.082108497619629, "learning_rate": 7.081022562015221e-06, "loss": 2.313, "step": 469450 }, { "epoch": 4.146134689768452, "grad_norm": 7.712873935699463, "learning_rate": 7.073895269029035e-06, "loss": 2.502, "step": 469500 }, { "epoch": 4.146576237658737, "grad_norm": 5.0041327476501465, "learning_rate": 7.066771291720447e-06, "loss": 2.6024, "step": 469550 }, { "epoch": 4.1470177855490205, "grad_norm": 7.34561014175415, "learning_rate": 7.059650630639713e-06, "loss": 2.4633, "step": 469600 }, { "epoch": 4.147459333439305, "grad_norm": 6.024482250213623, "learning_rate": 7.0525332863368375e-06, "loss": 2.7119, "step": 469650 }, { "epoch": 4.147900881329589, "grad_norm": 4.034352779388428, "learning_rate": 7.04541925936158e-06, "loss": 2.8465, "step": 469700 }, { "epoch": 4.148342429219873, "grad_norm": 6.708507537841797, "learning_rate": 7.038308550263451e-06, "loss": 2.9348, "step": 469750 }, { "epoch": 4.148783977110157, "grad_norm": 3.0077290534973145, "learning_rate": 7.031201159591671e-06, "loss": 2.2042, "step": 469800 }, { "epoch": 4.149225525000442, "grad_norm": 6.0895304679870605, "learning_rate": 7.024097087895232e-06, "loss": 2.9656, "step": 469850 }, { "epoch": 4.1496670728907254, "grad_norm": 5.052486896514893, "learning_rate": 7.0169963357228866e-06, "loss": 2.757, "step": 469900 }, { "epoch": 4.15010862078101, "grad_norm": 4.277814865112305, "learning_rate": 7.009898903623074e-06, "loss": 2.8647, "step": 469950 }, { "epoch": 4.150550168671294, "grad_norm": 2.1149942874908447, "learning_rate": 7.002804792144035e-06, "loss": 2.3847, "step": 470000 }, { "epoch": 4.150991716561578, "grad_norm": 5.320175647735596, "learning_rate": 6.995714001833725e-06, "loss": 2.5567, "step": 470050 }, { "epoch": 4.151433264451862, "grad_norm": 5.622382164001465, "learning_rate": 6.988626533239828e-06, "loss": 2.4812, "step": 470100 }, { "epoch": 4.151874812342147, "grad_norm": 3.8529694080352783, "learning_rate": 6.981542386909806e-06, "loss": 2.3033, "step": 470150 }, { "epoch": 4.1523163602324304, "grad_norm": 5.920238494873047, "learning_rate": 6.974461563390855e-06, "loss": 2.8147, "step": 470200 }, { "epoch": 4.152757908122715, "grad_norm": 1.4681082963943481, "learning_rate": 6.967384063229903e-06, "loss": 2.9136, "step": 470250 }, { "epoch": 4.153199456012999, "grad_norm": 5.235662937164307, "learning_rate": 6.960309886973621e-06, "loss": 2.4806, "step": 470300 }, { "epoch": 4.153641003903283, "grad_norm": 6.26996374130249, "learning_rate": 6.953239035168446e-06, "loss": 2.4752, "step": 470350 }, { "epoch": 4.154082551793567, "grad_norm": 7.696044921875, "learning_rate": 6.946171508360527e-06, "loss": 2.7682, "step": 470400 }, { "epoch": 4.154524099683852, "grad_norm": 3.8939006328582764, "learning_rate": 6.939107307095766e-06, "loss": 2.8521, "step": 470450 }, { "epoch": 4.1549656475741354, "grad_norm": 2.0132553577423096, "learning_rate": 6.9320464319198145e-06, "loss": 2.5654, "step": 470500 }, { "epoch": 4.15540719546442, "grad_norm": 3.739814043045044, "learning_rate": 6.924988883378086e-06, "loss": 2.4399, "step": 470550 }, { "epoch": 4.155848743354705, "grad_norm": 4.465041637420654, "learning_rate": 6.9179346620156905e-06, "loss": 2.8414, "step": 470600 }, { "epoch": 4.156290291244988, "grad_norm": 14.440762519836426, "learning_rate": 6.910883768377513e-06, "loss": 2.5322, "step": 470650 }, { "epoch": 4.156731839135273, "grad_norm": 3.621710777282715, "learning_rate": 6.903836203008179e-06, "loss": 2.544, "step": 470700 }, { "epoch": 4.157173387025557, "grad_norm": 9.714335441589355, "learning_rate": 6.8967919664520704e-06, "loss": 2.4297, "step": 470750 }, { "epoch": 4.157614934915841, "grad_norm": 3.76513671875, "learning_rate": 6.889751059253274e-06, "loss": 2.5278, "step": 470800 }, { "epoch": 4.158056482806125, "grad_norm": 6.704042434692383, "learning_rate": 6.882713481955633e-06, "loss": 2.6833, "step": 470850 }, { "epoch": 4.15849803069641, "grad_norm": 8.229090690612793, "learning_rate": 6.875679235102761e-06, "loss": 2.2542, "step": 470900 }, { "epoch": 4.158939578586693, "grad_norm": 3.761962652206421, "learning_rate": 6.868648319237975e-06, "loss": 2.9918, "step": 470950 }, { "epoch": 4.159381126476978, "grad_norm": 5.7836079597473145, "learning_rate": 6.8616207349043615e-06, "loss": 2.2541, "step": 471000 }, { "epoch": 4.159381126476978, "eval_asr_loss": 0.9798031646191468, "eval_loss": 2.7065012454986572, "eval_runtime": 20.843, "eval_samples_per_second": 36.847, "eval_steps_per_second": 9.212, "eval_tts_loss": 6.320543147371053, "step": 471000 }, { "epoch": 4.159822674367262, "grad_norm": 6.813299655914307, "learning_rate": 6.85459648264476e-06, "loss": 2.7841, "step": 471050 }, { "epoch": 4.160264222257546, "grad_norm": 3.69102144241333, "learning_rate": 6.8475755630017015e-06, "loss": 2.4495, "step": 471100 }, { "epoch": 4.16070577014783, "grad_norm": 3.805835485458374, "learning_rate": 6.8405579765175196e-06, "loss": 2.6548, "step": 471150 }, { "epoch": 4.161147318038115, "grad_norm": 4.671054840087891, "learning_rate": 6.833543723734259e-06, "loss": 2.5253, "step": 471200 }, { "epoch": 4.161588865928398, "grad_norm": 8.298660278320312, "learning_rate": 6.826532805193686e-06, "loss": 3.1047, "step": 471250 }, { "epoch": 4.162030413818683, "grad_norm": 2.760791301727295, "learning_rate": 6.8195252214373566e-06, "loss": 2.6616, "step": 471300 }, { "epoch": 4.162471961708967, "grad_norm": 2.2989513874053955, "learning_rate": 6.812520973006548e-06, "loss": 2.3819, "step": 471350 }, { "epoch": 4.162913509599251, "grad_norm": 5.132716178894043, "learning_rate": 6.805520060442283e-06, "loss": 2.7338, "step": 471400 }, { "epoch": 4.163355057489535, "grad_norm": 5.673553943634033, "learning_rate": 6.798522484285303e-06, "loss": 3.0991, "step": 471450 }, { "epoch": 4.16379660537982, "grad_norm": 4.2512102127075195, "learning_rate": 6.791528245076134e-06, "loss": 2.9495, "step": 471500 }, { "epoch": 4.164238153270103, "grad_norm": 8.3931245803833, "learning_rate": 6.784537343355002e-06, "loss": 2.9547, "step": 471550 }, { "epoch": 4.164679701160388, "grad_norm": 6.0636091232299805, "learning_rate": 6.777549779661912e-06, "loss": 2.8711, "step": 471600 }, { "epoch": 4.165121249050672, "grad_norm": 3.9343013763427734, "learning_rate": 6.770565554536579e-06, "loss": 3.0295, "step": 471650 }, { "epoch": 4.165562796940956, "grad_norm": 5.961034774780273, "learning_rate": 6.763584668518486e-06, "loss": 2.9189, "step": 471700 }, { "epoch": 4.16600434483124, "grad_norm": 5.926181316375732, "learning_rate": 6.756607122146841e-06, "loss": 2.3878, "step": 471750 }, { "epoch": 4.166445892721525, "grad_norm": 7.692674160003662, "learning_rate": 6.749632915960596e-06, "loss": 2.6937, "step": 471800 }, { "epoch": 4.166887440611808, "grad_norm": 2.865745782852173, "learning_rate": 6.74266205049846e-06, "loss": 2.9001, "step": 471850 }, { "epoch": 4.167328988502093, "grad_norm": 6.1057610511779785, "learning_rate": 6.735694526298874e-06, "loss": 2.4968, "step": 471900 }, { "epoch": 4.167770536392377, "grad_norm": 4.623075008392334, "learning_rate": 6.728730343900019e-06, "loss": 2.7525, "step": 471950 }, { "epoch": 4.168212084282661, "grad_norm": 11.1688232421875, "learning_rate": 6.721769503839808e-06, "loss": 2.1283, "step": 472000 }, { "epoch": 4.168653632172946, "grad_norm": 3.145315170288086, "learning_rate": 6.714812006655924e-06, "loss": 2.8079, "step": 472050 }, { "epoch": 4.16909518006323, "grad_norm": 3.9433867931365967, "learning_rate": 6.7078578528857514e-06, "loss": 2.6817, "step": 472100 }, { "epoch": 4.169536727953514, "grad_norm": 2.590686798095703, "learning_rate": 6.700907043066451e-06, "loss": 2.8969, "step": 472150 }, { "epoch": 4.169978275843798, "grad_norm": 3.163261890411377, "learning_rate": 6.6939595777349315e-06, "loss": 2.843, "step": 472200 }, { "epoch": 4.170419823734083, "grad_norm": 3.780911922454834, "learning_rate": 6.6870154574277925e-06, "loss": 2.7666, "step": 472250 }, { "epoch": 4.170861371624366, "grad_norm": 2.6461668014526367, "learning_rate": 6.680074682681442e-06, "loss": 2.7764, "step": 472300 }, { "epoch": 4.171302919514651, "grad_norm": 2.887061834335327, "learning_rate": 6.6732759698097445e-06, "loss": 2.6391, "step": 472350 }, { "epoch": 4.171744467404935, "grad_norm": 8.833243370056152, "learning_rate": 6.666341820855115e-06, "loss": 2.3806, "step": 472400 }, { "epoch": 4.172186015295219, "grad_norm": 6.725586891174316, "learning_rate": 6.659411019058126e-06, "loss": 2.2605, "step": 472450 }, { "epoch": 4.172627563185503, "grad_norm": 3.94309139251709, "learning_rate": 6.652483564954115e-06, "loss": 2.9116, "step": 472500 }, { "epoch": 4.173069111075788, "grad_norm": 3.9893558025360107, "learning_rate": 6.645559459078166e-06, "loss": 2.3809, "step": 472550 }, { "epoch": 4.173510658966071, "grad_norm": 5.374122619628906, "learning_rate": 6.638638701965127e-06, "loss": 2.3479, "step": 472600 }, { "epoch": 4.173952206856356, "grad_norm": 6.687133312225342, "learning_rate": 6.631721294149556e-06, "loss": 2.9528, "step": 472650 }, { "epoch": 4.17439375474664, "grad_norm": 6.609723091125488, "learning_rate": 6.6248072361657475e-06, "loss": 2.5894, "step": 472700 }, { "epoch": 4.174835302636924, "grad_norm": 3.7756259441375732, "learning_rate": 6.6178965285477846e-06, "loss": 2.5504, "step": 472750 }, { "epoch": 4.175276850527208, "grad_norm": 3.4726474285125732, "learning_rate": 6.6111272861215266e-06, "loss": 2.8805, "step": 472800 }, { "epoch": 4.175718398417493, "grad_norm": 8.812176704406738, "learning_rate": 6.6042232138024515e-06, "loss": 2.935, "step": 472850 }, { "epoch": 4.176159946307776, "grad_norm": 6.789100170135498, "learning_rate": 6.59732249343914e-06, "loss": 2.7707, "step": 472900 }, { "epoch": 4.176601494198061, "grad_norm": 4.578874111175537, "learning_rate": 6.590425125564631e-06, "loss": 3.1452, "step": 472950 }, { "epoch": 4.177043042088345, "grad_norm": 4.574524879455566, "learning_rate": 6.583531110711682e-06, "loss": 2.5646, "step": 473000 }, { "epoch": 4.177484589978629, "grad_norm": 5.257936954498291, "learning_rate": 6.57664044941278e-06, "loss": 2.292, "step": 473050 }, { "epoch": 4.177926137868913, "grad_norm": 7.503411769866943, "learning_rate": 6.569753142200186e-06, "loss": 2.5786, "step": 473100 }, { "epoch": 4.178367685759198, "grad_norm": 4.433591842651367, "learning_rate": 6.562869189605891e-06, "loss": 2.8026, "step": 473150 }, { "epoch": 4.178809233649481, "grad_norm": 8.797429084777832, "learning_rate": 6.55598859216161e-06, "loss": 2.5929, "step": 473200 }, { "epoch": 4.179250781539766, "grad_norm": 5.956727504730225, "learning_rate": 6.549111350398812e-06, "loss": 2.8071, "step": 473250 }, { "epoch": 4.17969232943005, "grad_norm": 3.960620164871216, "learning_rate": 6.542374909665405e-06, "loss": 2.6885, "step": 473300 }, { "epoch": 4.180133877320334, "grad_norm": 4.8596720695495605, "learning_rate": 6.535504313718876e-06, "loss": 2.7339, "step": 473350 }, { "epoch": 4.180575425210618, "grad_norm": 2.547786235809326, "learning_rate": 6.5286370750360755e-06, "loss": 2.6245, "step": 473400 }, { "epoch": 4.181016973100903, "grad_norm": 1.804262638092041, "learning_rate": 6.521773194147429e-06, "loss": 2.1986, "step": 473450 }, { "epoch": 4.181458520991186, "grad_norm": 1.9349775314331055, "learning_rate": 6.514912671583118e-06, "loss": 2.9735, "step": 473500 }, { "epoch": 4.181900068881471, "grad_norm": 5.907736778259277, "learning_rate": 6.508055507873067e-06, "loss": 2.69, "step": 473550 }, { "epoch": 4.182341616771755, "grad_norm": 7.595005989074707, "learning_rate": 6.501201703546922e-06, "loss": 2.712, "step": 473600 }, { "epoch": 4.182783164662039, "grad_norm": 2.434749126434326, "learning_rate": 6.494351259134091e-06, "loss": 2.4148, "step": 473650 }, { "epoch": 4.183224712552324, "grad_norm": 6.350268840789795, "learning_rate": 6.487504175163711e-06, "loss": 2.9523, "step": 473700 }, { "epoch": 4.183666260442608, "grad_norm": 2.100736379623413, "learning_rate": 6.480660452164649e-06, "loss": 2.5391, "step": 473750 }, { "epoch": 4.184107808332892, "grad_norm": 5.388315200805664, "learning_rate": 6.473820090665534e-06, "loss": 2.6482, "step": 473800 }, { "epoch": 4.184549356223176, "grad_norm": 1.6382206678390503, "learning_rate": 6.4669830911947226e-06, "loss": 2.5016, "step": 473850 }, { "epoch": 4.1849909041134605, "grad_norm": 6.089871883392334, "learning_rate": 6.460149454280334e-06, "loss": 2.7054, "step": 473900 }, { "epoch": 4.185432452003744, "grad_norm": 4.281028747558594, "learning_rate": 6.453319180450179e-06, "loss": 2.7412, "step": 473950 }, { "epoch": 4.185873999894029, "grad_norm": 2.96475887298584, "learning_rate": 6.446492270231863e-06, "loss": 2.2763, "step": 474000 }, { "epoch": 4.185873999894029, "eval_asr_loss": 0.9825393299944443, "eval_loss": 2.7079155445098877, "eval_runtime": 20.3312, "eval_samples_per_second": 37.775, "eval_steps_per_second": 9.444, "eval_tts_loss": 6.33437503431467, "step": 474000 }, { "epoch": 4.186315547784313, "grad_norm": 8.47600269317627, "learning_rate": 6.439668724152681e-06, "loss": 2.5068, "step": 474050 }, { "epoch": 4.186757095674597, "grad_norm": 8.663228988647461, "learning_rate": 6.432848542739728e-06, "loss": 2.4725, "step": 474100 }, { "epoch": 4.187198643564881, "grad_norm": 3.7695279121398926, "learning_rate": 6.426031726519771e-06, "loss": 2.5383, "step": 474150 }, { "epoch": 4.1876401914551655, "grad_norm": 4.956834316253662, "learning_rate": 6.4192182760193785e-06, "loss": 2.5511, "step": 474200 }, { "epoch": 4.188081739345449, "grad_norm": 9.05332088470459, "learning_rate": 6.412408191764807e-06, "loss": 2.5929, "step": 474250 }, { "epoch": 4.188523287235734, "grad_norm": 2.9488365650177, "learning_rate": 6.405601474282097e-06, "loss": 2.6003, "step": 474300 }, { "epoch": 4.188964835126018, "grad_norm": 3.8284595012664795, "learning_rate": 6.398798124097e-06, "loss": 2.8676, "step": 474350 }, { "epoch": 4.189406383016302, "grad_norm": 3.3335022926330566, "learning_rate": 6.391998141735028e-06, "loss": 2.9047, "step": 474400 }, { "epoch": 4.189847930906586, "grad_norm": 4.223811149597168, "learning_rate": 6.385201527721418e-06, "loss": 3.1682, "step": 474450 }, { "epoch": 4.1902894787968705, "grad_norm": 7.40073299407959, "learning_rate": 6.378408282581133e-06, "loss": 2.3481, "step": 474500 }, { "epoch": 4.190731026687154, "grad_norm": 7.080527305603027, "learning_rate": 6.371618406838925e-06, "loss": 2.4046, "step": 474550 }, { "epoch": 4.191172574577439, "grad_norm": 3.731682538986206, "learning_rate": 6.364831901019219e-06, "loss": 2.4267, "step": 474600 }, { "epoch": 4.191614122467723, "grad_norm": 5.064785480499268, "learning_rate": 6.358048765646241e-06, "loss": 2.9546, "step": 474650 }, { "epoch": 4.192055670358007, "grad_norm": 7.780782699584961, "learning_rate": 6.351269001243926e-06, "loss": 2.8367, "step": 474700 }, { "epoch": 4.192497218248291, "grad_norm": 8.36408805847168, "learning_rate": 6.3444926083359446e-06, "loss": 2.5074, "step": 474750 }, { "epoch": 4.1929387661385755, "grad_norm": 5.1836018562316895, "learning_rate": 6.3377195874457285e-06, "loss": 2.4778, "step": 474800 }, { "epoch": 4.193380314028859, "grad_norm": 8.31612491607666, "learning_rate": 6.330949939096431e-06, "loss": 2.4277, "step": 474850 }, { "epoch": 4.193821861919144, "grad_norm": 1.6141719818115234, "learning_rate": 6.32418366381094e-06, "loss": 2.6839, "step": 474900 }, { "epoch": 4.194263409809428, "grad_norm": 4.195858478546143, "learning_rate": 6.317420762111898e-06, "loss": 2.5166, "step": 474950 }, { "epoch": 4.194704957699712, "grad_norm": 4.043757438659668, "learning_rate": 6.310661234521686e-06, "loss": 2.8694, "step": 475000 }, { "epoch": 4.195146505589996, "grad_norm": 6.807188034057617, "learning_rate": 6.303905081562428e-06, "loss": 2.6234, "step": 475050 }, { "epoch": 4.1955880534802805, "grad_norm": 8.207228660583496, "learning_rate": 6.29715230375596e-06, "loss": 2.6476, "step": 475100 }, { "epoch": 4.196029601370564, "grad_norm": 7.567667007446289, "learning_rate": 6.290402901623887e-06, "loss": 2.6501, "step": 475150 }, { "epoch": 4.196471149260849, "grad_norm": 3.4421753883361816, "learning_rate": 6.2836568756875545e-06, "loss": 2.974, "step": 475200 }, { "epoch": 4.196912697151133, "grad_norm": 5.635768890380859, "learning_rate": 6.276914226468023e-06, "loss": 2.4975, "step": 475250 }, { "epoch": 4.197354245041417, "grad_norm": 3.613474130630493, "learning_rate": 6.270174954486096e-06, "loss": 2.9628, "step": 475300 }, { "epoch": 4.197795792931702, "grad_norm": 8.936803817749023, "learning_rate": 6.26343906026235e-06, "loss": 2.5715, "step": 475350 }, { "epoch": 4.1982373408219855, "grad_norm": 4.440316677093506, "learning_rate": 6.2567065443170484e-06, "loss": 2.7625, "step": 475400 }, { "epoch": 4.19867888871227, "grad_norm": 4.047897815704346, "learning_rate": 6.249977407170232e-06, "loss": 2.4286, "step": 475450 }, { "epoch": 4.199120436602554, "grad_norm": 7.088375568389893, "learning_rate": 6.243251649341675e-06, "loss": 2.4241, "step": 475500 }, { "epoch": 4.1995619844928385, "grad_norm": 4.977496147155762, "learning_rate": 6.236529271350888e-06, "loss": 2.5992, "step": 475550 }, { "epoch": 4.200003532383122, "grad_norm": 6.744089603424072, "learning_rate": 6.229810273717118e-06, "loss": 2.9347, "step": 475600 }, { "epoch": 4.200445080273407, "grad_norm": 4.792475700378418, "learning_rate": 6.223094656959327e-06, "loss": 2.3914, "step": 475650 }, { "epoch": 4.2008866281636905, "grad_norm": 5.523590564727783, "learning_rate": 6.216382421596273e-06, "loss": 2.7553, "step": 475700 }, { "epoch": 4.201328176053975, "grad_norm": 6.3388991355896, "learning_rate": 6.209673568146385e-06, "loss": 2.802, "step": 475750 }, { "epoch": 4.201769723944259, "grad_norm": 7.17784309387207, "learning_rate": 6.202968097127887e-06, "loss": 2.3509, "step": 475800 }, { "epoch": 4.2022112718345435, "grad_norm": 6.590193271636963, "learning_rate": 6.196266009058727e-06, "loss": 2.704, "step": 475850 }, { "epoch": 4.202652819724827, "grad_norm": 4.52247428894043, "learning_rate": 6.189567304456562e-06, "loss": 2.3842, "step": 475900 }, { "epoch": 4.203094367615112, "grad_norm": 7.403465270996094, "learning_rate": 6.182871983838817e-06, "loss": 2.8809, "step": 475950 }, { "epoch": 4.2035359155053955, "grad_norm": 1.0545246601104736, "learning_rate": 6.176180047722669e-06, "loss": 2.8488, "step": 476000 }, { "epoch": 4.20397746339568, "grad_norm": 6.862853050231934, "learning_rate": 6.1694914966249905e-06, "loss": 2.5551, "step": 476050 }, { "epoch": 4.204419011285964, "grad_norm": 2.4551446437835693, "learning_rate": 6.16280633106242e-06, "loss": 2.4942, "step": 476100 }, { "epoch": 4.2048605591762485, "grad_norm": 9.989826202392578, "learning_rate": 6.156124551551324e-06, "loss": 2.6294, "step": 476150 }, { "epoch": 4.205302107066532, "grad_norm": 3.2986972332000732, "learning_rate": 6.149446158607836e-06, "loss": 2.4716, "step": 476200 }, { "epoch": 4.205743654956817, "grad_norm": 6.839714527130127, "learning_rate": 6.142771152747778e-06, "loss": 2.5413, "step": 476250 }, { "epoch": 4.2061852028471005, "grad_norm": 9.555171012878418, "learning_rate": 6.136099534486756e-06, "loss": 2.53, "step": 476300 }, { "epoch": 4.206626750737385, "grad_norm": 3.119110107421875, "learning_rate": 6.129431304340094e-06, "loss": 2.8151, "step": 476350 }, { "epoch": 4.207068298627669, "grad_norm": 3.6401970386505127, "learning_rate": 6.122766462822854e-06, "loss": 2.9423, "step": 476400 }, { "epoch": 4.2075098465179535, "grad_norm": 7.693792819976807, "learning_rate": 6.116105010449824e-06, "loss": 2.4146, "step": 476450 }, { "epoch": 4.207951394408237, "grad_norm": 4.048227787017822, "learning_rate": 6.10944694773557e-06, "loss": 2.6369, "step": 476500 }, { "epoch": 4.208392942298522, "grad_norm": 5.10174560546875, "learning_rate": 6.102792275194347e-06, "loss": 2.7271, "step": 476550 }, { "epoch": 4.2088344901888055, "grad_norm": 7.955861568450928, "learning_rate": 6.096140993340188e-06, "loss": 2.713, "step": 476600 }, { "epoch": 4.20927603807909, "grad_norm": 8.910422325134277, "learning_rate": 6.089493102686849e-06, "loss": 2.4657, "step": 476650 }, { "epoch": 4.209717585969374, "grad_norm": 7.3145952224731445, "learning_rate": 6.082848603747804e-06, "loss": 3.0729, "step": 476700 }, { "epoch": 4.2101591338596585, "grad_norm": 6.688584327697754, "learning_rate": 6.076207497036312e-06, "loss": 2.7802, "step": 476750 }, { "epoch": 4.210600681749942, "grad_norm": 5.091372013092041, "learning_rate": 6.069569783065321e-06, "loss": 2.9407, "step": 476800 }, { "epoch": 4.211042229640227, "grad_norm": 2.814818859100342, "learning_rate": 6.062935462347552e-06, "loss": 2.6655, "step": 476850 }, { "epoch": 4.2114837775305105, "grad_norm": 3.7893147468566895, "learning_rate": 6.056304535395435e-06, "loss": 3.0867, "step": 476900 }, { "epoch": 4.211925325420795, "grad_norm": 7.157622337341309, "learning_rate": 6.049677002721155e-06, "loss": 2.6567, "step": 476950 }, { "epoch": 4.21236687331108, "grad_norm": 5.8184332847595215, "learning_rate": 6.043052864836657e-06, "loss": 2.6973, "step": 477000 }, { "epoch": 4.21236687331108, "eval_asr_loss": 0.986782086099792, "eval_loss": 2.705303192138672, "eval_runtime": 20.954, "eval_samples_per_second": 36.652, "eval_steps_per_second": 9.163, "eval_tts_loss": 6.312509116483838, "step": 477000 }, { "epoch": 4.2128084212013635, "grad_norm": 6.687573432922363, "learning_rate": 6.036432122253566e-06, "loss": 2.9112, "step": 477050 }, { "epoch": 4.213249969091648, "grad_norm": 6.303692817687988, "learning_rate": 6.029814775483294e-06, "loss": 2.4362, "step": 477100 }, { "epoch": 4.213691516981932, "grad_norm": 3.7523443698883057, "learning_rate": 6.023200825036984e-06, "loss": 2.2645, "step": 477150 }, { "epoch": 4.214133064872216, "grad_norm": 3.9468162059783936, "learning_rate": 6.016590271425499e-06, "loss": 2.5491, "step": 477200 }, { "epoch": 4.2145746127625, "grad_norm": 8.005352973937988, "learning_rate": 6.009983115159434e-06, "loss": 2.78, "step": 477250 }, { "epoch": 4.215016160652785, "grad_norm": 4.13003396987915, "learning_rate": 6.003379356749156e-06, "loss": 3.2115, "step": 477300 }, { "epoch": 4.2154577085430684, "grad_norm": 6.442688941955566, "learning_rate": 5.996778996704744e-06, "loss": 2.4184, "step": 477350 }, { "epoch": 4.215899256433353, "grad_norm": 3.361504077911377, "learning_rate": 5.990182035536013e-06, "loss": 2.682, "step": 477400 }, { "epoch": 4.216340804323637, "grad_norm": 6.338619232177734, "learning_rate": 5.9835884737525276e-06, "loss": 2.4386, "step": 477450 }, { "epoch": 4.216782352213921, "grad_norm": 3.299943208694458, "learning_rate": 5.976998311863591e-06, "loss": 2.5548, "step": 477500 }, { "epoch": 4.217223900104205, "grad_norm": 5.694170951843262, "learning_rate": 5.970411550378235e-06, "loss": 2.983, "step": 477550 }, { "epoch": 4.21766544799449, "grad_norm": 2.95349383354187, "learning_rate": 5.963828189805215e-06, "loss": 2.6167, "step": 477600 }, { "epoch": 4.2181069958847734, "grad_norm": 5.988198757171631, "learning_rate": 5.957248230653062e-06, "loss": 2.7957, "step": 477650 }, { "epoch": 4.218548543775058, "grad_norm": 6.402997970581055, "learning_rate": 5.950671673429997e-06, "loss": 2.6169, "step": 477700 }, { "epoch": 4.218990091665342, "grad_norm": 4.538114547729492, "learning_rate": 5.94409851864402e-06, "loss": 2.7612, "step": 477750 }, { "epoch": 4.219431639555626, "grad_norm": 3.642622470855713, "learning_rate": 5.937528766802863e-06, "loss": 2.9086, "step": 477800 }, { "epoch": 4.21987318744591, "grad_norm": 5.173513412475586, "learning_rate": 5.93096241841396e-06, "loss": 2.6582, "step": 477850 }, { "epoch": 4.220314735336195, "grad_norm": 6.45066499710083, "learning_rate": 5.92439947398451e-06, "loss": 2.9489, "step": 477900 }, { "epoch": 4.2207562832264784, "grad_norm": 4.035391807556152, "learning_rate": 5.917839934021463e-06, "loss": 2.8774, "step": 477950 }, { "epoch": 4.221197831116763, "grad_norm": 8.273470878601074, "learning_rate": 5.911283799031481e-06, "loss": 2.2632, "step": 478000 }, { "epoch": 4.221639379007047, "grad_norm": 6.139615535736084, "learning_rate": 5.904731069520947e-06, "loss": 2.5812, "step": 478050 }, { "epoch": 4.222080926897331, "grad_norm": 8.84054946899414, "learning_rate": 5.898181745996023e-06, "loss": 3.0429, "step": 478100 }, { "epoch": 4.222522474787615, "grad_norm": 5.0236711502075195, "learning_rate": 5.891635828962594e-06, "loss": 2.8701, "step": 478150 }, { "epoch": 4.2229640226779, "grad_norm": 9.935722351074219, "learning_rate": 5.8850933189262656e-06, "loss": 2.2851, "step": 478200 }, { "epoch": 4.223405570568183, "grad_norm": 5.947291374206543, "learning_rate": 5.87855421639239e-06, "loss": 2.4865, "step": 478250 }, { "epoch": 4.223847118458468, "grad_norm": 5.341309070587158, "learning_rate": 5.872018521866074e-06, "loss": 2.5714, "step": 478300 }, { "epoch": 4.224288666348752, "grad_norm": 6.42664909362793, "learning_rate": 5.86548623585213e-06, "loss": 2.2757, "step": 478350 }, { "epoch": 4.224730214239036, "grad_norm": 5.356601715087891, "learning_rate": 5.858957358855116e-06, "loss": 2.8174, "step": 478400 }, { "epoch": 4.225171762129321, "grad_norm": 3.144227981567383, "learning_rate": 5.852431891379345e-06, "loss": 2.7757, "step": 478450 }, { "epoch": 4.225613310019605, "grad_norm": 3.4820075035095215, "learning_rate": 5.845909833928859e-06, "loss": 2.806, "step": 478500 }, { "epoch": 4.226054857909889, "grad_norm": 5.959455966949463, "learning_rate": 5.839391187007409e-06, "loss": 3.242, "step": 478550 }, { "epoch": 4.226496405800173, "grad_norm": 5.716809272766113, "learning_rate": 5.832875951118516e-06, "loss": 2.861, "step": 478600 }, { "epoch": 4.226937953690458, "grad_norm": 5.608268737792969, "learning_rate": 5.826364126765444e-06, "loss": 2.6959, "step": 478650 }, { "epoch": 4.227379501580741, "grad_norm": 5.232941627502441, "learning_rate": 5.819855714451156e-06, "loss": 2.664, "step": 478700 }, { "epoch": 4.227821049471026, "grad_norm": 8.215174674987793, "learning_rate": 5.813350714678378e-06, "loss": 2.3906, "step": 478750 }, { "epoch": 4.22826259736131, "grad_norm": 6.991796493530273, "learning_rate": 5.8068491279495706e-06, "loss": 3.1848, "step": 478800 }, { "epoch": 4.228704145251594, "grad_norm": 5.6993584632873535, "learning_rate": 5.800350954766914e-06, "loss": 2.663, "step": 478850 }, { "epoch": 4.229145693141878, "grad_norm": 2.2606592178344727, "learning_rate": 5.793986057354117e-06, "loss": 2.4432, "step": 478900 }, { "epoch": 4.229587241032163, "grad_norm": 3.6462740898132324, "learning_rate": 5.787494644473384e-06, "loss": 2.3709, "step": 478950 }, { "epoch": 4.230028788922446, "grad_norm": 3.86085844039917, "learning_rate": 5.781006646633791e-06, "loss": 2.6778, "step": 479000 }, { "epoch": 4.230470336812731, "grad_norm": 5.5888566970825195, "learning_rate": 5.7745220643364685e-06, "loss": 2.4313, "step": 479050 }, { "epoch": 4.230911884703015, "grad_norm": 9.503171920776367, "learning_rate": 5.768040898082283e-06, "loss": 2.6374, "step": 479100 }, { "epoch": 4.231353432593299, "grad_norm": 6.172721862792969, "learning_rate": 5.7615631483718625e-06, "loss": 2.5965, "step": 479150 }, { "epoch": 4.231794980483583, "grad_norm": 5.849079608917236, "learning_rate": 5.7550888157055725e-06, "loss": 2.8944, "step": 479200 }, { "epoch": 4.232236528373868, "grad_norm": 5.0087456703186035, "learning_rate": 5.74861790058347e-06, "loss": 2.5753, "step": 479250 }, { "epoch": 4.232678076264151, "grad_norm": 5.77261209487915, "learning_rate": 5.742150403505397e-06, "loss": 2.6699, "step": 479300 }, { "epoch": 4.233119624154436, "grad_norm": 6.500524520874023, "learning_rate": 5.735686324970918e-06, "loss": 2.7891, "step": 479350 }, { "epoch": 4.23356117204472, "grad_norm": 3.1344494819641113, "learning_rate": 5.72922566547931e-06, "loss": 2.5381, "step": 479400 }, { "epoch": 4.234002719935004, "grad_norm": 4.875197887420654, "learning_rate": 5.722768425529629e-06, "loss": 2.5972, "step": 479450 }, { "epoch": 4.234444267825288, "grad_norm": 3.2009196281433105, "learning_rate": 5.716314605620609e-06, "loss": 2.5743, "step": 479500 }, { "epoch": 4.234885815715573, "grad_norm": 2.808358669281006, "learning_rate": 5.709864206250787e-06, "loss": 2.4277, "step": 479550 }, { "epoch": 4.235327363605856, "grad_norm": 4.8687639236450195, "learning_rate": 5.703417227918373e-06, "loss": 2.5137, "step": 479600 }, { "epoch": 4.235768911496141, "grad_norm": 2.686998128890991, "learning_rate": 5.696973671121358e-06, "loss": 2.2794, "step": 479650 }, { "epoch": 4.236210459386425, "grad_norm": 7.4152021408081055, "learning_rate": 5.690533536357456e-06, "loss": 2.0998, "step": 479700 }, { "epoch": 4.236652007276709, "grad_norm": 1.4863412380218506, "learning_rate": 5.684096824124091e-06, "loss": 2.4334, "step": 479750 }, { "epoch": 4.237093555166993, "grad_norm": 4.32270622253418, "learning_rate": 5.677663534918476e-06, "loss": 2.6505, "step": 479800 }, { "epoch": 4.237535103057278, "grad_norm": 7.150432586669922, "learning_rate": 5.671233669237502e-06, "loss": 2.7076, "step": 479850 }, { "epoch": 4.237976650947561, "grad_norm": 5.002776145935059, "learning_rate": 5.664807227577823e-06, "loss": 2.5764, "step": 479900 }, { "epoch": 4.238418198837846, "grad_norm": 7.036445140838623, "learning_rate": 5.658384210435841e-06, "loss": 2.5472, "step": 479950 }, { "epoch": 4.23885974672813, "grad_norm": 6.226653099060059, "learning_rate": 5.651964618307676e-06, "loss": 2.1791, "step": 480000 }, { "epoch": 4.23885974672813, "eval_asr_loss": 0.9848060228370414, "eval_loss": 2.702357292175293, "eval_runtime": 20.4549, "eval_samples_per_second": 37.546, "eval_steps_per_second": 9.386, "eval_tts_loss": 6.3120058456623305, "step": 480000 }, { "epoch": 4.239301294618414, "grad_norm": 6.989476680755615, "learning_rate": 5.645548451689176e-06, "loss": 2.7977, "step": 480050 }, { "epoch": 4.239742842508699, "grad_norm": 7.659889221191406, "learning_rate": 5.639135711075938e-06, "loss": 2.8623, "step": 480100 }, { "epoch": 4.240184390398983, "grad_norm": 8.97298526763916, "learning_rate": 5.632726396963306e-06, "loss": 2.4556, "step": 480150 }, { "epoch": 4.240625938289267, "grad_norm": 7.485257625579834, "learning_rate": 5.626320509846339e-06, "loss": 2.8332, "step": 480200 }, { "epoch": 4.241067486179551, "grad_norm": 7.15934944152832, "learning_rate": 5.619918050219841e-06, "loss": 2.7152, "step": 480250 }, { "epoch": 4.241509034069836, "grad_norm": 4.678470611572266, "learning_rate": 5.6135190185783286e-06, "loss": 2.5411, "step": 480300 }, { "epoch": 4.241950581960119, "grad_norm": 5.154171466827393, "learning_rate": 5.607123415416093e-06, "loss": 3.1134, "step": 480350 }, { "epoch": 4.242392129850404, "grad_norm": 7.04273796081543, "learning_rate": 5.600731241227125e-06, "loss": 2.7577, "step": 480400 }, { "epoch": 4.242833677740688, "grad_norm": 3.2310116291046143, "learning_rate": 5.594342496505173e-06, "loss": 2.5335, "step": 480450 }, { "epoch": 4.243275225630972, "grad_norm": 6.954891681671143, "learning_rate": 5.587957181743731e-06, "loss": 2.7522, "step": 480500 }, { "epoch": 4.243716773521256, "grad_norm": 3.6511993408203125, "learning_rate": 5.581575297435976e-06, "loss": 2.4105, "step": 480550 }, { "epoch": 4.244158321411541, "grad_norm": 3.7625491619110107, "learning_rate": 5.5751968440748815e-06, "loss": 2.947, "step": 480600 }, { "epoch": 4.244599869301824, "grad_norm": 7.865955829620361, "learning_rate": 5.5688218221531145e-06, "loss": 2.5522, "step": 480650 }, { "epoch": 4.245041417192109, "grad_norm": 9.019535064697266, "learning_rate": 5.5624502321631025e-06, "loss": 2.147, "step": 480700 }, { "epoch": 4.245482965082393, "grad_norm": 3.498694658279419, "learning_rate": 5.5560820745969795e-06, "loss": 2.4009, "step": 480750 }, { "epoch": 4.245924512972677, "grad_norm": 5.897189140319824, "learning_rate": 5.549717349946648e-06, "loss": 2.8977, "step": 480800 }, { "epoch": 4.246366060862961, "grad_norm": 8.465011596679688, "learning_rate": 5.54335605870373e-06, "loss": 2.7665, "step": 480850 }, { "epoch": 4.246807608753246, "grad_norm": 8.51477336883545, "learning_rate": 5.5369982013595666e-06, "loss": 2.5223, "step": 480900 }, { "epoch": 4.247249156643529, "grad_norm": 2.815274477005005, "learning_rate": 5.53064377840527e-06, "loss": 2.8688, "step": 480950 }, { "epoch": 4.247690704533814, "grad_norm": 1.8431061506271362, "learning_rate": 5.524292790331647e-06, "loss": 2.5171, "step": 481000 }, { "epoch": 4.248132252424098, "grad_norm": 9.258687973022461, "learning_rate": 5.517945237629257e-06, "loss": 2.7509, "step": 481050 }, { "epoch": 4.248573800314382, "grad_norm": 7.54561185836792, "learning_rate": 5.511601120788395e-06, "loss": 2.6896, "step": 481100 }, { "epoch": 4.249015348204666, "grad_norm": 8.003030776977539, "learning_rate": 5.5052604402991095e-06, "loss": 2.1706, "step": 481150 }, { "epoch": 4.249456896094951, "grad_norm": 10.485010147094727, "learning_rate": 5.49892319665114e-06, "loss": 2.7453, "step": 481200 }, { "epoch": 4.249898443985234, "grad_norm": 8.127185821533203, "learning_rate": 5.4925893903339965e-06, "loss": 2.6854, "step": 481250 }, { "epoch": 4.250339991875519, "grad_norm": 12.30605697631836, "learning_rate": 5.486259021836909e-06, "loss": 2.2861, "step": 481300 }, { "epoch": 4.250781539765803, "grad_norm": 7.714188575744629, "learning_rate": 5.47993209164886e-06, "loss": 2.561, "step": 481350 }, { "epoch": 4.251223087656087, "grad_norm": 1.737669587135315, "learning_rate": 5.473608600258534e-06, "loss": 2.6832, "step": 481400 }, { "epoch": 4.251664635546371, "grad_norm": 10.227282524108887, "learning_rate": 5.467288548154365e-06, "loss": 2.8556, "step": 481450 }, { "epoch": 4.252106183436656, "grad_norm": 7.471348762512207, "learning_rate": 5.4609719358245386e-06, "loss": 2.4608, "step": 481500 }, { "epoch": 4.252547731326939, "grad_norm": 4.42140007019043, "learning_rate": 5.454658763756937e-06, "loss": 2.5171, "step": 481550 }, { "epoch": 4.252989279217224, "grad_norm": 7.784973621368408, "learning_rate": 5.448349032439215e-06, "loss": 2.3761, "step": 481600 }, { "epoch": 4.253430827107508, "grad_norm": 8.040699005126953, "learning_rate": 5.44204274235875e-06, "loss": 2.7332, "step": 481650 }, { "epoch": 4.253872374997792, "grad_norm": 7.201604843139648, "learning_rate": 5.435739894002634e-06, "loss": 2.5332, "step": 481700 }, { "epoch": 4.254313922888077, "grad_norm": 8.169370651245117, "learning_rate": 5.4294404878577285e-06, "loss": 2.6991, "step": 481750 }, { "epoch": 4.254755470778361, "grad_norm": 8.216754913330078, "learning_rate": 5.423144524410589e-06, "loss": 2.5936, "step": 481800 }, { "epoch": 4.255197018668645, "grad_norm": 5.163374900817871, "learning_rate": 5.4168520041475435e-06, "loss": 2.465, "step": 481850 }, { "epoch": 4.255638566558929, "grad_norm": 3.130244731903076, "learning_rate": 5.410562927554613e-06, "loss": 2.8108, "step": 481900 }, { "epoch": 4.2560801144492135, "grad_norm": 8.824630737304688, "learning_rate": 5.404277295117593e-06, "loss": 2.7851, "step": 481950 }, { "epoch": 4.256521662339497, "grad_norm": 7.502809047698975, "learning_rate": 5.397995107322001e-06, "loss": 3.2602, "step": 482000 }, { "epoch": 4.256963210229782, "grad_norm": 3.9156007766723633, "learning_rate": 5.39171636465306e-06, "loss": 2.6791, "step": 482050 }, { "epoch": 4.257404758120066, "grad_norm": 4.457824230194092, "learning_rate": 5.385441067595764e-06, "loss": 2.6464, "step": 482100 }, { "epoch": 4.25784630601035, "grad_norm": 7.053863525390625, "learning_rate": 5.379169216634839e-06, "loss": 2.5613, "step": 482150 }, { "epoch": 4.258287853900634, "grad_norm": 10.0283203125, "learning_rate": 5.372900812254716e-06, "loss": 2.4984, "step": 482200 }, { "epoch": 4.2587294017909185, "grad_norm": 5.658593654632568, "learning_rate": 5.3666358549395665e-06, "loss": 2.7865, "step": 482250 }, { "epoch": 4.259170949681202, "grad_norm": 6.3613972663879395, "learning_rate": 5.36037434517333e-06, "loss": 2.8537, "step": 482300 }, { "epoch": 4.259612497571487, "grad_norm": 5.856276035308838, "learning_rate": 5.3541162834396304e-06, "loss": 2.4153, "step": 482350 }, { "epoch": 4.260054045461771, "grad_norm": 8.076888084411621, "learning_rate": 5.347861670221865e-06, "loss": 2.5878, "step": 482400 }, { "epoch": 4.260495593352055, "grad_norm": 6.364077568054199, "learning_rate": 5.341610506003142e-06, "loss": 2.5507, "step": 482450 }, { "epoch": 4.260937141242339, "grad_norm": 2.1528849601745605, "learning_rate": 5.335362791266329e-06, "loss": 2.8302, "step": 482500 }, { "epoch": 4.2613786891326235, "grad_norm": 5.577317237854004, "learning_rate": 5.329118526493992e-06, "loss": 2.9665, "step": 482550 }, { "epoch": 4.261820237022907, "grad_norm": 5.122951507568359, "learning_rate": 5.322877712168445e-06, "loss": 2.803, "step": 482600 }, { "epoch": 4.262261784913192, "grad_norm": 3.6531784534454346, "learning_rate": 5.316640348771751e-06, "loss": 2.5349, "step": 482650 }, { "epoch": 4.262703332803476, "grad_norm": 7.163844585418701, "learning_rate": 5.310406436785676e-06, "loss": 2.6873, "step": 482700 }, { "epoch": 4.26314488069376, "grad_norm": 4.684516906738281, "learning_rate": 5.30417597669175e-06, "loss": 2.6846, "step": 482750 }, { "epoch": 4.263586428584044, "grad_norm": 2.3888607025146484, "learning_rate": 5.297948968971228e-06, "loss": 2.3929, "step": 482800 }, { "epoch": 4.2640279764743285, "grad_norm": 1.9348230361938477, "learning_rate": 5.291725414105075e-06, "loss": 2.8077, "step": 482850 }, { "epoch": 4.264469524364612, "grad_norm": 6.656325817108154, "learning_rate": 5.285505312574029e-06, "loss": 2.6571, "step": 482900 }, { "epoch": 4.264911072254897, "grad_norm": 8.594658851623535, "learning_rate": 5.279288664858517e-06, "loss": 2.9681, "step": 482950 }, { "epoch": 4.265352620145181, "grad_norm": 5.734218597412109, "learning_rate": 5.273075471438749e-06, "loss": 2.7512, "step": 483000 }, { "epoch": 4.265352620145181, "eval_asr_loss": 0.9848095697997507, "eval_loss": 2.7029573917388916, "eval_runtime": 20.504, "eval_samples_per_second": 37.456, "eval_steps_per_second": 9.364, "eval_tts_loss": 6.332219609490086, "step": 483000 }, { "epoch": 4.265794168035465, "grad_norm": 10.293950080871582, "learning_rate": 5.266865732794618e-06, "loss": 2.9099, "step": 483050 }, { "epoch": 4.266235715925749, "grad_norm": 7.920646667480469, "learning_rate": 5.2606594494057805e-06, "loss": 2.3955, "step": 483100 }, { "epoch": 4.2666772638160335, "grad_norm": 3.1784679889678955, "learning_rate": 5.254456621751635e-06, "loss": 2.4171, "step": 483150 }, { "epoch": 4.267118811706318, "grad_norm": 6.525193214416504, "learning_rate": 5.248257250311273e-06, "loss": 2.7102, "step": 483200 }, { "epoch": 4.267560359596602, "grad_norm": 5.434006214141846, "learning_rate": 5.242061335563558e-06, "loss": 2.8168, "step": 483250 }, { "epoch": 4.268001907486886, "grad_norm": 4.984245777130127, "learning_rate": 5.235868877987071e-06, "loss": 2.713, "step": 483300 }, { "epoch": 4.26844345537717, "grad_norm": 6.375180244445801, "learning_rate": 5.229679878060134e-06, "loss": 2.7143, "step": 483350 }, { "epoch": 4.268885003267455, "grad_norm": 4.152289867401123, "learning_rate": 5.22349433626077e-06, "loss": 2.4193, "step": 483400 }, { "epoch": 4.2693265511577385, "grad_norm": 4.980320930480957, "learning_rate": 5.21731225306678e-06, "loss": 2.7505, "step": 483450 }, { "epoch": 4.269768099048023, "grad_norm": 6.4112548828125, "learning_rate": 5.211133628955672e-06, "loss": 2.5717, "step": 483500 }, { "epoch": 4.270209646938307, "grad_norm": 7.3201680183410645, "learning_rate": 5.2049584644046855e-06, "loss": 2.8863, "step": 483550 }, { "epoch": 4.2706511948285915, "grad_norm": 7.759049415588379, "learning_rate": 5.198786759890811e-06, "loss": 2.3369, "step": 483600 }, { "epoch": 4.271092742718875, "grad_norm": 2.023104667663574, "learning_rate": 5.192618515890762e-06, "loss": 2.9714, "step": 483650 }, { "epoch": 4.27153429060916, "grad_norm": 8.213711738586426, "learning_rate": 5.18645373288098e-06, "loss": 2.488, "step": 483700 }, { "epoch": 4.2719758384994435, "grad_norm": 2.2088820934295654, "learning_rate": 5.180292411337628e-06, "loss": 2.6149, "step": 483750 }, { "epoch": 4.272417386389728, "grad_norm": 10.039661407470703, "learning_rate": 5.174134551736637e-06, "loss": 2.5648, "step": 483800 }, { "epoch": 4.272858934280012, "grad_norm": 4.669970512390137, "learning_rate": 5.167980154553631e-06, "loss": 2.6526, "step": 483850 }, { "epoch": 4.2733004821702965, "grad_norm": 6.565603733062744, "learning_rate": 5.161829220263997e-06, "loss": 2.5761, "step": 483900 }, { "epoch": 4.27374203006058, "grad_norm": 4.891339302062988, "learning_rate": 5.1556817493428444e-06, "loss": 2.5161, "step": 483950 }, { "epoch": 4.274183577950865, "grad_norm": 6.347530841827393, "learning_rate": 5.149537742264998e-06, "loss": 3.0394, "step": 484000 }, { "epoch": 4.2746251258411485, "grad_norm": 3.2913765907287598, "learning_rate": 5.14339719950504e-06, "loss": 2.7304, "step": 484050 }, { "epoch": 4.275066673731433, "grad_norm": 5.916070461273193, "learning_rate": 5.13726012153729e-06, "loss": 2.8762, "step": 484100 }, { "epoch": 4.275508221621717, "grad_norm": 4.099598407745361, "learning_rate": 5.131126508835765e-06, "loss": 2.9367, "step": 484150 }, { "epoch": 4.2759497695120015, "grad_norm": 8.383169174194336, "learning_rate": 5.1249963618742344e-06, "loss": 2.8743, "step": 484200 }, { "epoch": 4.276391317402285, "grad_norm": 13.008307456970215, "learning_rate": 5.1188696811262035e-06, "loss": 2.5611, "step": 484250 }, { "epoch": 4.27683286529257, "grad_norm": 4.897049427032471, "learning_rate": 5.112746467064916e-06, "loss": 2.3247, "step": 484300 }, { "epoch": 4.2772744131828535, "grad_norm": 4.620345115661621, "learning_rate": 5.1066267201633265e-06, "loss": 2.9306, "step": 484350 }, { "epoch": 4.277715961073138, "grad_norm": 4.074400901794434, "learning_rate": 5.100510440894135e-06, "loss": 2.5387, "step": 484400 }, { "epoch": 4.278157508963422, "grad_norm": 3.9772870540618896, "learning_rate": 5.094397629729791e-06, "loss": 2.5631, "step": 484450 }, { "epoch": 4.2785990568537065, "grad_norm": 4.75778865814209, "learning_rate": 5.088288287142434e-06, "loss": 2.6017, "step": 484500 }, { "epoch": 4.27904060474399, "grad_norm": 6.53339147567749, "learning_rate": 5.082182413603964e-06, "loss": 2.4342, "step": 484550 }, { "epoch": 4.279482152634275, "grad_norm": 5.238636493682861, "learning_rate": 5.076080009586015e-06, "loss": 2.4136, "step": 484600 }, { "epoch": 4.2799237005245585, "grad_norm": 2.656336545944214, "learning_rate": 5.069981075559932e-06, "loss": 2.9483, "step": 484650 }, { "epoch": 4.280365248414843, "grad_norm": 6.56556510925293, "learning_rate": 5.064007487254502e-06, "loss": 2.8989, "step": 484700 }, { "epoch": 4.280806796305127, "grad_norm": 8.722373962402344, "learning_rate": 5.057915425201887e-06, "loss": 2.8708, "step": 484750 }, { "epoch": 4.2812483441954114, "grad_norm": 3.9821183681488037, "learning_rate": 5.0518268345442155e-06, "loss": 2.8032, "step": 484800 }, { "epoch": 4.281689892085696, "grad_norm": 8.309123039245605, "learning_rate": 5.045741715751779e-06, "loss": 2.8429, "step": 484850 }, { "epoch": 4.28213143997598, "grad_norm": 8.290190696716309, "learning_rate": 5.039660069294577e-06, "loss": 3.1943, "step": 484900 }, { "epoch": 4.282572987866264, "grad_norm": 2.1401474475860596, "learning_rate": 5.033581895642386e-06, "loss": 3.0381, "step": 484950 }, { "epoch": 4.283014535756548, "grad_norm": 5.286615371704102, "learning_rate": 5.027507195264697e-06, "loss": 2.5756, "step": 485000 }, { "epoch": 4.283456083646833, "grad_norm": 5.696613311767578, "learning_rate": 5.021435968630711e-06, "loss": 2.3705, "step": 485050 }, { "epoch": 4.2838976315371164, "grad_norm": 12.439728736877441, "learning_rate": 5.015368216209393e-06, "loss": 2.7611, "step": 485100 }, { "epoch": 4.284339179427401, "grad_norm": 6.768801689147949, "learning_rate": 5.009303938469428e-06, "loss": 2.5504, "step": 485150 }, { "epoch": 4.284780727317685, "grad_norm": 5.860541820526123, "learning_rate": 5.003243135879226e-06, "loss": 2.7186, "step": 485200 }, { "epoch": 4.285222275207969, "grad_norm": 3.728076219558716, "learning_rate": 4.997185808906924e-06, "loss": 2.6144, "step": 485250 }, { "epoch": 4.285663823098253, "grad_norm": 6.470057487487793, "learning_rate": 4.9911319580204e-06, "loss": 2.4354, "step": 485300 }, { "epoch": 4.286105370988538, "grad_norm": 5.76416015625, "learning_rate": 4.985081583687279e-06, "loss": 2.5053, "step": 485350 }, { "epoch": 4.286546918878821, "grad_norm": 2.1108109951019287, "learning_rate": 4.979034686374884e-06, "loss": 2.6416, "step": 485400 }, { "epoch": 4.286988466769106, "grad_norm": 4.978553295135498, "learning_rate": 4.972991266550287e-06, "loss": 3.0807, "step": 485450 }, { "epoch": 4.28743001465939, "grad_norm": 4.632789134979248, "learning_rate": 4.966951324680313e-06, "loss": 2.8228, "step": 485500 }, { "epoch": 4.287871562549674, "grad_norm": 7.106553554534912, "learning_rate": 4.9609148612314634e-06, "loss": 2.5059, "step": 485550 }, { "epoch": 4.288313110439958, "grad_norm": 7.273108959197998, "learning_rate": 4.954881876670037e-06, "loss": 2.443, "step": 485600 }, { "epoch": 4.288754658330243, "grad_norm": 2.8921549320220947, "learning_rate": 4.948852371462009e-06, "loss": 2.8635, "step": 485650 }, { "epoch": 4.289196206220526, "grad_norm": 6.967789173126221, "learning_rate": 4.942826346073104e-06, "loss": 2.7458, "step": 485700 }, { "epoch": 4.289637754110811, "grad_norm": 5.395866394042969, "learning_rate": 4.936803800968792e-06, "loss": 2.3966, "step": 485750 }, { "epoch": 4.290079302001095, "grad_norm": 7.758505821228027, "learning_rate": 4.930784736614263e-06, "loss": 2.2958, "step": 485800 }, { "epoch": 4.290520849891379, "grad_norm": 4.8945794105529785, "learning_rate": 4.924769153474445e-06, "loss": 2.5833, "step": 485850 }, { "epoch": 4.290962397781663, "grad_norm": 6.577781677246094, "learning_rate": 4.9187570520139725e-06, "loss": 2.7678, "step": 485900 }, { "epoch": 4.291403945671948, "grad_norm": 7.482282638549805, "learning_rate": 4.91274843269725e-06, "loss": 2.5584, "step": 485950 }, { "epoch": 4.291845493562231, "grad_norm": 6.959794521331787, "learning_rate": 4.906743295988381e-06, "loss": 2.8044, "step": 486000 }, { "epoch": 4.291845493562231, "eval_asr_loss": 0.9882799570298296, "eval_loss": 2.699824333190918, "eval_runtime": 20.532, "eval_samples_per_second": 37.405, "eval_steps_per_second": 9.351, "eval_tts_loss": 6.337574664431568, "step": 486000 }, { "epoch": 4.292287041452516, "grad_norm": 5.685201168060303, "learning_rate": 4.900741642351209e-06, "loss": 2.5164, "step": 486050 }, { "epoch": 4.2927285893428, "grad_norm": 5.617766380310059, "learning_rate": 4.894743472249308e-06, "loss": 2.4879, "step": 486100 }, { "epoch": 4.293170137233084, "grad_norm": 6.860740661621094, "learning_rate": 4.888748786146002e-06, "loss": 2.6254, "step": 486150 }, { "epoch": 4.293611685123368, "grad_norm": 10.409443855285645, "learning_rate": 4.882757584504316e-06, "loss": 2.9805, "step": 486200 }, { "epoch": 4.294053233013653, "grad_norm": 2.122511148452759, "learning_rate": 4.876769867787018e-06, "loss": 2.6383, "step": 486250 }, { "epoch": 4.294494780903936, "grad_norm": 5.052753448486328, "learning_rate": 4.870785636456626e-06, "loss": 2.678, "step": 486300 }, { "epoch": 4.294936328794221, "grad_norm": 2.6877222061157227, "learning_rate": 4.864804890975349e-06, "loss": 2.726, "step": 486350 }, { "epoch": 4.295377876684505, "grad_norm": 2.0756945610046387, "learning_rate": 4.858827631805168e-06, "loss": 2.513, "step": 486400 }, { "epoch": 4.295819424574789, "grad_norm": 5.22865629196167, "learning_rate": 4.852853859407758e-06, "loss": 2.4439, "step": 486450 }, { "epoch": 4.296260972465074, "grad_norm": 6.545832633972168, "learning_rate": 4.84688357424456e-06, "loss": 2.8417, "step": 486500 }, { "epoch": 4.296702520355358, "grad_norm": 3.8261468410491943, "learning_rate": 4.84091677677671e-06, "loss": 2.4088, "step": 486550 }, { "epoch": 4.297144068245642, "grad_norm": 4.2262091636657715, "learning_rate": 4.834953467465103e-06, "loss": 2.7598, "step": 486600 }, { "epoch": 4.297585616135926, "grad_norm": 6.198680877685547, "learning_rate": 4.828993646770358e-06, "loss": 2.8131, "step": 486650 }, { "epoch": 4.298027164026211, "grad_norm": 6.853316307067871, "learning_rate": 4.823156407589236e-06, "loss": 2.2681, "step": 486700 }, { "epoch": 4.298468711916494, "grad_norm": 8.242879867553711, "learning_rate": 4.817203495713707e-06, "loss": 2.5881, "step": 486750 }, { "epoch": 4.298910259806779, "grad_norm": 4.18668794631958, "learning_rate": 4.811254073826071e-06, "loss": 2.758, "step": 486800 }, { "epoch": 4.299351807697063, "grad_norm": 9.72754955291748, "learning_rate": 4.805308142385872e-06, "loss": 2.5376, "step": 486850 }, { "epoch": 4.299793355587347, "grad_norm": 4.070117950439453, "learning_rate": 4.799365701852387e-06, "loss": 2.6845, "step": 486900 }, { "epoch": 4.300234903477631, "grad_norm": 6.563328266143799, "learning_rate": 4.7934267526845985e-06, "loss": 2.5158, "step": 486950 }, { "epoch": 4.300676451367916, "grad_norm": Infinity, "learning_rate": 4.787609970265289e-06, "loss": 2.4582, "step": 487000 }, { "epoch": 4.301117999258199, "grad_norm": 6.9214348793029785, "learning_rate": 4.781677935354706e-06, "loss": 2.6586, "step": 487050 }, { "epoch": 4.301559547148484, "grad_norm": 6.719581127166748, "learning_rate": 4.775749393176066e-06, "loss": 2.7698, "step": 487100 }, { "epoch": 4.302001095038768, "grad_norm": 4.886457920074463, "learning_rate": 4.769824344187291e-06, "loss": 2.6244, "step": 487150 }, { "epoch": 4.302442642929052, "grad_norm": 5.569522857666016, "learning_rate": 4.763902788846042e-06, "loss": 2.2506, "step": 487200 }, { "epoch": 4.302884190819336, "grad_norm": 7.942785263061523, "learning_rate": 4.75798472760971e-06, "loss": 2.8619, "step": 487250 }, { "epoch": 4.303325738709621, "grad_norm": 1.1087185144424438, "learning_rate": 4.752070160935429e-06, "loss": 2.6735, "step": 487300 }, { "epoch": 4.303767286599904, "grad_norm": 7.459014415740967, "learning_rate": 4.746159089280022e-06, "loss": 2.7118, "step": 487350 }, { "epoch": 4.304208834490189, "grad_norm": 9.74527645111084, "learning_rate": 4.740251513100097e-06, "loss": 2.8449, "step": 487400 }, { "epoch": 4.304650382380473, "grad_norm": 8.665245056152344, "learning_rate": 4.734347432851954e-06, "loss": 2.7211, "step": 487450 }, { "epoch": 4.305091930270757, "grad_norm": 7.08310079574585, "learning_rate": 4.728446848991619e-06, "loss": 3.1874, "step": 487500 }, { "epoch": 4.305533478161041, "grad_norm": 5.284829139709473, "learning_rate": 4.72254976197487e-06, "loss": 2.6234, "step": 487550 }, { "epoch": 4.305975026051326, "grad_norm": 5.368645191192627, "learning_rate": 4.716656172257222e-06, "loss": 2.4466, "step": 487600 }, { "epoch": 4.306416573941609, "grad_norm": 3.1773622035980225, "learning_rate": 4.710766080293888e-06, "loss": 2.4666, "step": 487650 }, { "epoch": 4.306858121831894, "grad_norm": 16.724567413330078, "learning_rate": 4.704879486539826e-06, "loss": 2.6646, "step": 487700 }, { "epoch": 4.307299669722178, "grad_norm": 10.451478958129883, "learning_rate": 4.698996391449744e-06, "loss": 2.7985, "step": 487750 }, { "epoch": 4.307741217612462, "grad_norm": 4.751099586486816, "learning_rate": 4.693116795478053e-06, "loss": 2.4775, "step": 487800 }, { "epoch": 4.308182765502746, "grad_norm": 5.029719352722168, "learning_rate": 4.687240699078888e-06, "loss": 2.6793, "step": 487850 }, { "epoch": 4.308624313393031, "grad_norm": 3.1056411266326904, "learning_rate": 4.681368102706135e-06, "loss": 2.5247, "step": 487900 }, { "epoch": 4.309065861283315, "grad_norm": 7.1487016677856445, "learning_rate": 4.675499006813416e-06, "loss": 3.029, "step": 487950 }, { "epoch": 4.309507409173599, "grad_norm": 7.196158409118652, "learning_rate": 4.669633411854047e-06, "loss": 2.6741, "step": 488000 }, { "epoch": 4.309948957063883, "grad_norm": 7.4663004875183105, "learning_rate": 4.663771318281113e-06, "loss": 2.9016, "step": 488050 }, { "epoch": 4.310390504954167, "grad_norm": 4.9442973136901855, "learning_rate": 4.657912726547409e-06, "loss": 2.5135, "step": 488100 }, { "epoch": 4.310832052844452, "grad_norm": 4.37014627456665, "learning_rate": 4.652057637105456e-06, "loss": 2.7247, "step": 488150 }, { "epoch": 4.311273600734736, "grad_norm": 3.2909576892852783, "learning_rate": 4.6462060504075035e-06, "loss": 2.9132, "step": 488200 }, { "epoch": 4.31171514862502, "grad_norm": 3.1253247261047363, "learning_rate": 4.640357966905551e-06, "loss": 2.9717, "step": 488250 }, { "epoch": 4.312156696515304, "grad_norm": 2.9556400775909424, "learning_rate": 4.634513387051304e-06, "loss": 2.6441, "step": 488300 }, { "epoch": 4.312598244405589, "grad_norm": 6.862311840057373, "learning_rate": 4.628672311296206e-06, "loss": 2.7396, "step": 488350 }, { "epoch": 4.313039792295872, "grad_norm": 6.476156711578369, "learning_rate": 4.622834740091437e-06, "loss": 2.8199, "step": 488400 }, { "epoch": 4.313481340186157, "grad_norm": 4.889683246612549, "learning_rate": 4.6170006738879e-06, "loss": 2.4284, "step": 488450 }, { "epoch": 4.313922888076441, "grad_norm": 6.792129039764404, "learning_rate": 4.611170113136221e-06, "loss": 2.8663, "step": 488500 }, { "epoch": 4.314364435966725, "grad_norm": 5.918289661407471, "learning_rate": 4.6053430582867705e-06, "loss": 2.6852, "step": 488550 }, { "epoch": 4.314805983857009, "grad_norm": 6.084130764007568, "learning_rate": 4.599519509789641e-06, "loss": 2.9438, "step": 488600 }, { "epoch": 4.315247531747294, "grad_norm": 4.472723007202148, "learning_rate": 4.593699468094625e-06, "loss": 2.6628, "step": 488650 }, { "epoch": 4.315689079637577, "grad_norm": 5.560323715209961, "learning_rate": 4.587882933651299e-06, "loss": 2.5735, "step": 488700 }, { "epoch": 4.316130627527862, "grad_norm": 5.769144058227539, "learning_rate": 4.582069906908937e-06, "loss": 2.7952, "step": 488750 }, { "epoch": 4.316572175418146, "grad_norm": 3.602747917175293, "learning_rate": 4.576260388316533e-06, "loss": 3.04, "step": 488800 }, { "epoch": 4.31701372330843, "grad_norm": 6.910214900970459, "learning_rate": 4.570454378322836e-06, "loss": 2.2622, "step": 488850 }, { "epoch": 4.317455271198714, "grad_norm": 6.3227057456970215, "learning_rate": 4.564651877376313e-06, "loss": 2.5845, "step": 488900 }, { "epoch": 4.317896819088999, "grad_norm": 5.122942924499512, "learning_rate": 4.558852885925152e-06, "loss": 2.967, "step": 488950 }, { "epoch": 4.318338366979282, "grad_norm": 4.114598274230957, "learning_rate": 4.553057404417266e-06, "loss": 2.6663, "step": 489000 }, { "epoch": 4.318338366979282, "eval_asr_loss": 0.9906294275140428, "eval_loss": 2.697246789932251, "eval_runtime": 20.6347, "eval_samples_per_second": 37.219, "eval_steps_per_second": 9.305, "eval_tts_loss": 6.328047646816105, "step": 489000 }, { "epoch": 4.318779914869567, "grad_norm": 3.6980276107788086, "learning_rate": 4.547265433300318e-06, "loss": 2.5157, "step": 489050 }, { "epoch": 4.319221462759851, "grad_norm": 4.031253337860107, "learning_rate": 4.541476973021697e-06, "loss": 2.7317, "step": 489100 }, { "epoch": 4.319663010650135, "grad_norm": 1.884820580482483, "learning_rate": 4.535692024028493e-06, "loss": 2.8184, "step": 489150 }, { "epoch": 4.320104558540419, "grad_norm": 2.3672525882720947, "learning_rate": 4.529910586767555e-06, "loss": 2.5384, "step": 489200 }, { "epoch": 4.320546106430704, "grad_norm": 6.837733745574951, "learning_rate": 4.524132661685454e-06, "loss": 2.7463, "step": 489250 }, { "epoch": 4.320987654320987, "grad_norm": 6.051320552825928, "learning_rate": 4.518358249228477e-06, "loss": 3.0508, "step": 489300 }, { "epoch": 4.321429202211272, "grad_norm": 3.4234349727630615, "learning_rate": 4.512587349842657e-06, "loss": 2.6353, "step": 489350 }, { "epoch": 4.321870750101556, "grad_norm": 8.457873344421387, "learning_rate": 4.506819963973746e-06, "loss": 2.6882, "step": 489400 }, { "epoch": 4.32231229799184, "grad_norm": 6.727595329284668, "learning_rate": 4.501056092067213e-06, "loss": 2.588, "step": 489450 }, { "epoch": 4.322753845882124, "grad_norm": 10.648479461669922, "learning_rate": 4.495295734568273e-06, "loss": 2.7656, "step": 489500 }, { "epoch": 4.323195393772409, "grad_norm": 8.020366668701172, "learning_rate": 4.489538891921874e-06, "loss": 2.7483, "step": 489550 }, { "epoch": 4.323636941662693, "grad_norm": 6.2409515380859375, "learning_rate": 4.483785564572685e-06, "loss": 2.4949, "step": 489600 }, { "epoch": 4.324078489552977, "grad_norm": 5.551110744476318, "learning_rate": 4.478035752965082e-06, "loss": 2.4452, "step": 489650 }, { "epoch": 4.324520037443261, "grad_norm": 4.0225725173950195, "learning_rate": 4.472289457543216e-06, "loss": 2.7476, "step": 489700 }, { "epoch": 4.324961585333545, "grad_norm": 7.549962997436523, "learning_rate": 4.4665466787509214e-06, "loss": 3.1367, "step": 489750 }, { "epoch": 4.32540313322383, "grad_norm": 6.687000751495361, "learning_rate": 4.460807417031781e-06, "loss": 2.8071, "step": 489800 }, { "epoch": 4.325844681114114, "grad_norm": 6.876231670379639, "learning_rate": 4.4550716728291034e-06, "loss": 2.7246, "step": 489850 }, { "epoch": 4.326286229004398, "grad_norm": 3.369643211364746, "learning_rate": 4.449339446585938e-06, "loss": 2.7797, "step": 489900 }, { "epoch": 4.326727776894682, "grad_norm": 1.468119740486145, "learning_rate": 4.44361073874503e-06, "loss": 2.642, "step": 489950 }, { "epoch": 4.3271693247849665, "grad_norm": 6.319380760192871, "learning_rate": 4.437885549748888e-06, "loss": 2.5545, "step": 490000 }, { "epoch": 4.32761087267525, "grad_norm": 4.769143104553223, "learning_rate": 4.432163880039741e-06, "loss": 2.5154, "step": 490050 }, { "epoch": 4.328052420565535, "grad_norm": 4.5597333908081055, "learning_rate": 4.4264457300595195e-06, "loss": 2.6695, "step": 490100 }, { "epoch": 4.328493968455819, "grad_norm": 5.355625152587891, "learning_rate": 4.420731100249925e-06, "loss": 2.7721, "step": 490150 }, { "epoch": 4.328935516346103, "grad_norm": 3.6555376052856445, "learning_rate": 4.4150199910523435e-06, "loss": 2.9497, "step": 490200 }, { "epoch": 4.329377064236387, "grad_norm": 5.317290306091309, "learning_rate": 4.409312402907928e-06, "loss": 2.555, "step": 490250 }, { "epoch": 4.3298186121266715, "grad_norm": 11.071691513061523, "learning_rate": 4.4036083362575194e-06, "loss": 3.0197, "step": 490300 }, { "epoch": 4.330260160016955, "grad_norm": 2.6705942153930664, "learning_rate": 4.397907791541722e-06, "loss": 2.5992, "step": 490350 }, { "epoch": 4.33070170790724, "grad_norm": 6.044713497161865, "learning_rate": 4.392210769200861e-06, "loss": 2.4217, "step": 490400 }, { "epoch": 4.331143255797524, "grad_norm": 8.060735702514648, "learning_rate": 4.386517269674967e-06, "loss": 2.5673, "step": 490450 }, { "epoch": 4.331584803687808, "grad_norm": 5.138390064239502, "learning_rate": 4.380827293403833e-06, "loss": 2.4732, "step": 490500 }, { "epoch": 4.332026351578092, "grad_norm": 5.2732768058776855, "learning_rate": 4.375140840826947e-06, "loss": 2.944, "step": 490550 }, { "epoch": 4.3324678994683765, "grad_norm": 4.809617042541504, "learning_rate": 4.36945791238354e-06, "loss": 2.6245, "step": 490600 }, { "epoch": 4.33290944735866, "grad_norm": 2.802457094192505, "learning_rate": 4.363778508512573e-06, "loss": 2.4827, "step": 490650 }, { "epoch": 4.333350995248945, "grad_norm": 7.932518005371094, "learning_rate": 4.358102629652738e-06, "loss": 2.7906, "step": 490700 }, { "epoch": 4.333792543139229, "grad_norm": 6.65126371383667, "learning_rate": 4.352430276242436e-06, "loss": 2.379, "step": 490750 }, { "epoch": 4.334234091029513, "grad_norm": 6.393111705780029, "learning_rate": 4.346761448719816e-06, "loss": 2.7133, "step": 490800 }, { "epoch": 4.334675638919797, "grad_norm": 6.8078413009643555, "learning_rate": 4.341096147522755e-06, "loss": 2.8168, "step": 490850 }, { "epoch": 4.3351171868100815, "grad_norm": 8.952127456665039, "learning_rate": 4.335434373088837e-06, "loss": 3.0868, "step": 490900 }, { "epoch": 4.335558734700365, "grad_norm": 4.609088897705078, "learning_rate": 4.329776125855384e-06, "loss": 2.6837, "step": 490950 }, { "epoch": 4.33600028259065, "grad_norm": 4.8116984367370605, "learning_rate": 4.324121406259452e-06, "loss": 2.5857, "step": 491000 }, { "epoch": 4.336441830480934, "grad_norm": 2.911388397216797, "learning_rate": 4.31847021473783e-06, "loss": 2.5079, "step": 491050 }, { "epoch": 4.336883378371218, "grad_norm": 4.033604621887207, "learning_rate": 4.312822551727008e-06, "loss": 2.8896, "step": 491100 }, { "epoch": 4.337324926261502, "grad_norm": 2.2807979583740234, "learning_rate": 4.307178417663227e-06, "loss": 2.6478, "step": 491150 }, { "epoch": 4.3377664741517865, "grad_norm": 4.249363899230957, "learning_rate": 4.301537812982459e-06, "loss": 2.7082, "step": 491200 }, { "epoch": 4.338208022042071, "grad_norm": 4.977548122406006, "learning_rate": 4.295900738120373e-06, "loss": 2.4809, "step": 491250 }, { "epoch": 4.338649569932355, "grad_norm": 5.844250202178955, "learning_rate": 4.290379829805263e-06, "loss": 2.8, "step": 491300 }, { "epoch": 4.3390911178226395, "grad_norm": 7.538488864898682, "learning_rate": 4.28474974526849e-06, "loss": 2.6848, "step": 491350 }, { "epoch": 4.339532665712923, "grad_norm": 8.326896667480469, "learning_rate": 4.279123191847145e-06, "loss": 2.9162, "step": 491400 }, { "epoch": 4.339974213603208, "grad_norm": 1.9365812540054321, "learning_rate": 4.273500169975841e-06, "loss": 2.5956, "step": 491450 }, { "epoch": 4.3404157614934915, "grad_norm": 9.246211051940918, "learning_rate": 4.267880680088892e-06, "loss": 3.1781, "step": 491500 }, { "epoch": 4.340857309383776, "grad_norm": 2.263589382171631, "learning_rate": 4.262264722620368e-06, "loss": 2.4923, "step": 491550 }, { "epoch": 4.34129885727406, "grad_norm": 5.852914333343506, "learning_rate": 4.2566522980040435e-06, "loss": 2.736, "step": 491600 }, { "epoch": 4.3417404051643445, "grad_norm": 4.65158748626709, "learning_rate": 4.251043406673449e-06, "loss": 2.5998, "step": 491650 }, { "epoch": 4.342181953054628, "grad_norm": 7.828727722167969, "learning_rate": 4.245438049061801e-06, "loss": 2.5487, "step": 491700 }, { "epoch": 4.342623500944913, "grad_norm": 3.619830369949341, "learning_rate": 4.239836225602073e-06, "loss": 2.3365, "step": 491750 }, { "epoch": 4.3430650488351965, "grad_norm": 2.631338596343994, "learning_rate": 4.23423793672697e-06, "loss": 2.8464, "step": 491800 }, { "epoch": 4.343506596725481, "grad_norm": 5.786888599395752, "learning_rate": 4.228643182868902e-06, "loss": 2.3114, "step": 491850 }, { "epoch": 4.343948144615765, "grad_norm": 4.098949909210205, "learning_rate": 4.223051964460012e-06, "loss": 2.74, "step": 491900 }, { "epoch": 4.3443896925060494, "grad_norm": 9.980148315429688, "learning_rate": 4.217464281932193e-06, "loss": 2.6954, "step": 491950 }, { "epoch": 4.344831240396333, "grad_norm": 7.595466613769531, "learning_rate": 4.211880135717034e-06, "loss": 2.5378, "step": 492000 }, { "epoch": 4.344831240396333, "eval_asr_loss": 0.9905429691739748, "eval_loss": 2.699641466140747, "eval_runtime": 20.4946, "eval_samples_per_second": 37.473, "eval_steps_per_second": 9.368, "eval_tts_loss": 6.322649876710032, "step": 492000 }, { "epoch": 4.345272788286618, "grad_norm": 5.130605697631836, "learning_rate": 4.20629952624585e-06, "loss": 3.1613, "step": 492050 }, { "epoch": 4.3457143361769015, "grad_norm": 6.696079254150391, "learning_rate": 4.200722453949718e-06, "loss": 2.0499, "step": 492100 }, { "epoch": 4.346155884067186, "grad_norm": 4.512566089630127, "learning_rate": 4.195148919259406e-06, "loss": 2.807, "step": 492150 }, { "epoch": 4.34659743195747, "grad_norm": 11.577535629272461, "learning_rate": 4.189578922605419e-06, "loss": 2.8138, "step": 492200 }, { "epoch": 4.3470389798477544, "grad_norm": 5.083506107330322, "learning_rate": 4.184012464418014e-06, "loss": 2.7203, "step": 492250 }, { "epoch": 4.347480527738038, "grad_norm": 5.200368881225586, "learning_rate": 4.178449545127128e-06, "loss": 2.2894, "step": 492300 }, { "epoch": 4.347922075628323, "grad_norm": 4.574040412902832, "learning_rate": 4.172890165162463e-06, "loss": 2.6727, "step": 492350 }, { "epoch": 4.3483636235186065, "grad_norm": 1.686342716217041, "learning_rate": 4.167334324953426e-06, "loss": 1.8116, "step": 492400 }, { "epoch": 4.348805171408891, "grad_norm": 4.701891899108887, "learning_rate": 4.161782024929173e-06, "loss": 2.3938, "step": 492450 }, { "epoch": 4.349246719299175, "grad_norm": 4.263546466827393, "learning_rate": 4.15623326551855e-06, "loss": 2.7473, "step": 492500 }, { "epoch": 4.3496882671894594, "grad_norm": 8.451787948608398, "learning_rate": 4.150688047150164e-06, "loss": 2.5291, "step": 492550 }, { "epoch": 4.350129815079743, "grad_norm": 4.4311041831970215, "learning_rate": 4.145146370252345e-06, "loss": 2.4191, "step": 492600 }, { "epoch": 4.350571362970028, "grad_norm": 4.7787251472473145, "learning_rate": 4.139608235253118e-06, "loss": 2.4498, "step": 492650 }, { "epoch": 4.3510129108603115, "grad_norm": 3.511380434036255, "learning_rate": 4.1340736425802775e-06, "loss": 2.2766, "step": 492700 }, { "epoch": 4.351454458750596, "grad_norm": 9.421882629394531, "learning_rate": 4.128542592661322e-06, "loss": 2.1642, "step": 492750 }, { "epoch": 4.35189600664088, "grad_norm": 4.264568328857422, "learning_rate": 4.123015085923471e-06, "loss": 2.8886, "step": 492800 }, { "epoch": 4.352337554531164, "grad_norm": 7.145339012145996, "learning_rate": 4.1174911227936705e-06, "loss": 2.7918, "step": 492850 }, { "epoch": 4.352779102421449, "grad_norm": 3.8509018421173096, "learning_rate": 4.11197070369862e-06, "loss": 2.7719, "step": 492900 }, { "epoch": 4.353220650311733, "grad_norm": 5.471995830535889, "learning_rate": 4.1064538290646994e-06, "loss": 2.4756, "step": 492950 }, { "epoch": 4.353662198202017, "grad_norm": 2.8742692470550537, "learning_rate": 4.100940499318062e-06, "loss": 2.4628, "step": 493000 }, { "epoch": 4.354103746092301, "grad_norm": 6.550550937652588, "learning_rate": 4.0954307148845515e-06, "loss": 2.3311, "step": 493050 }, { "epoch": 4.354545293982586, "grad_norm": 6.5168914794921875, "learning_rate": 4.089924476189772e-06, "loss": 2.6853, "step": 493100 }, { "epoch": 4.354986841872869, "grad_norm": 8.18268871307373, "learning_rate": 4.084421783659021e-06, "loss": 2.6904, "step": 493150 }, { "epoch": 4.355428389763154, "grad_norm": 8.38926887512207, "learning_rate": 4.078922637717325e-06, "loss": 2.9435, "step": 493200 }, { "epoch": 4.355869937653438, "grad_norm": 7.336807727813721, "learning_rate": 4.0734270387894625e-06, "loss": 2.2368, "step": 493250 }, { "epoch": 4.356311485543722, "grad_norm": 4.262712478637695, "learning_rate": 4.067934987299909e-06, "loss": 2.5068, "step": 493300 }, { "epoch": 4.356753033434006, "grad_norm": 2.5902600288391113, "learning_rate": 4.0624464836728826e-06, "loss": 2.8839, "step": 493350 }, { "epoch": 4.357194581324291, "grad_norm": 9.853610038757324, "learning_rate": 4.056961528332337e-06, "loss": 2.593, "step": 493400 }, { "epoch": 4.357636129214574, "grad_norm": 4.301543712615967, "learning_rate": 4.0515897150544265e-06, "loss": 2.978, "step": 493450 }, { "epoch": 4.358077677104859, "grad_norm": 2.651461362838745, "learning_rate": 4.046111786570728e-06, "loss": 3.046, "step": 493500 }, { "epoch": 4.358519224995143, "grad_norm": 6.080756187438965, "learning_rate": 4.040637407635212e-06, "loss": 2.2984, "step": 493550 }, { "epoch": 4.358960772885427, "grad_norm": 5.260184288024902, "learning_rate": 4.0351665786707235e-06, "loss": 2.5134, "step": 493600 }, { "epoch": 4.359402320775711, "grad_norm": 3.593035936355591, "learning_rate": 4.029699300099859e-06, "loss": 2.6971, "step": 493650 }, { "epoch": 4.359843868665996, "grad_norm": 7.141435623168945, "learning_rate": 4.024235572344898e-06, "loss": 3.0431, "step": 493700 }, { "epoch": 4.360285416556279, "grad_norm": 3.661709785461426, "learning_rate": 4.018775395827878e-06, "loss": 2.5285, "step": 493750 }, { "epoch": 4.360726964446564, "grad_norm": 3.3774867057800293, "learning_rate": 4.0133187709705535e-06, "loss": 2.85, "step": 493800 }, { "epoch": 4.361168512336848, "grad_norm": 6.806416988372803, "learning_rate": 4.007865698194402e-06, "loss": 2.7133, "step": 493850 }, { "epoch": 4.361610060227132, "grad_norm": 1.7745251655578613, "learning_rate": 4.002416177920609e-06, "loss": 2.4399, "step": 493900 }, { "epoch": 4.362051608117416, "grad_norm": 4.646737575531006, "learning_rate": 3.996970210570122e-06, "loss": 2.8364, "step": 493950 }, { "epoch": 4.362493156007701, "grad_norm": 4.611186504364014, "learning_rate": 3.991527796563604e-06, "loss": 2.8777, "step": 494000 }, { "epoch": 4.362934703897984, "grad_norm": 4.065881729125977, "learning_rate": 3.986088936321414e-06, "loss": 2.6513, "step": 494050 }, { "epoch": 4.363376251788269, "grad_norm": 2.9707062244415283, "learning_rate": 3.980653630263664e-06, "loss": 2.7259, "step": 494100 }, { "epoch": 4.363817799678553, "grad_norm": 6.588619709014893, "learning_rate": 3.975221878810198e-06, "loss": 2.6151, "step": 494150 }, { "epoch": 4.364259347568837, "grad_norm": 2.2025773525238037, "learning_rate": 3.969793682380552e-06, "loss": 2.9107, "step": 494200 }, { "epoch": 4.364700895459121, "grad_norm": 5.676170825958252, "learning_rate": 3.964369041394028e-06, "loss": 3.0397, "step": 494250 }, { "epoch": 4.365142443349406, "grad_norm": 13.548469543457031, "learning_rate": 3.958947956269632e-06, "loss": 2.4612, "step": 494300 }, { "epoch": 4.36558399123969, "grad_norm": 5.043992519378662, "learning_rate": 3.953530427426072e-06, "loss": 2.4597, "step": 494350 }, { "epoch": 4.366025539129974, "grad_norm": 8.227876663208008, "learning_rate": 3.94811645528183e-06, "loss": 2.5238, "step": 494400 }, { "epoch": 4.366467087020258, "grad_norm": 5.8309502601623535, "learning_rate": 3.942706040255089e-06, "loss": 2.589, "step": 494450 }, { "epoch": 4.366908634910542, "grad_norm": 6.448791980743408, "learning_rate": 3.937299182763743e-06, "loss": 2.754, "step": 494500 }, { "epoch": 4.367350182800827, "grad_norm": 4.799962520599365, "learning_rate": 3.9318958832254395e-06, "loss": 2.4749, "step": 494550 }, { "epoch": 4.367791730691111, "grad_norm": 1.42853844165802, "learning_rate": 3.926496142057534e-06, "loss": 2.4072, "step": 494600 }, { "epoch": 4.368233278581395, "grad_norm": 9.769267082214355, "learning_rate": 3.921099959677116e-06, "loss": 2.7348, "step": 494650 }, { "epoch": 4.368674826471679, "grad_norm": 6.4068732261657715, "learning_rate": 3.9157073365009785e-06, "loss": 2.9524, "step": 494700 }, { "epoch": 4.369116374361964, "grad_norm": 4.041799545288086, "learning_rate": 3.9104260193297935e-06, "loss": 2.5809, "step": 494750 }, { "epoch": 4.369557922252247, "grad_norm": 3.3633763790130615, "learning_rate": 3.905040444606745e-06, "loss": 3.2263, "step": 494800 }, { "epoch": 4.369999470142532, "grad_norm": 3.419987678527832, "learning_rate": 3.899658430328457e-06, "loss": 2.7493, "step": 494850 }, { "epoch": 4.370441018032816, "grad_norm": 4.437135696411133, "learning_rate": 3.894279976910625e-06, "loss": 2.9697, "step": 494900 }, { "epoch": 4.3708825659231, "grad_norm": 2.5502755641937256, "learning_rate": 3.8889050847687124e-06, "loss": 2.4348, "step": 494950 }, { "epoch": 4.371324113813384, "grad_norm": 8.30134391784668, "learning_rate": 3.883533754317859e-06, "loss": 2.7971, "step": 495000 }, { "epoch": 4.371324113813384, "eval_asr_loss": 0.984819574151661, "eval_loss": 2.6977908611297607, "eval_runtime": 20.2155, "eval_samples_per_second": 37.991, "eval_steps_per_second": 9.498, "eval_tts_loss": 6.340800073443951, "step": 495000 }, { "epoch": 4.371765661703669, "grad_norm": 5.813408374786377, "learning_rate": 3.878165985972981e-06, "loss": 2.4433, "step": 495050 }, { "epoch": 4.372207209593952, "grad_norm": 7.439929962158203, "learning_rate": 3.872801780148666e-06, "loss": 2.7701, "step": 495100 }, { "epoch": 4.372648757484237, "grad_norm": 7.2259392738342285, "learning_rate": 3.867441137259264e-06, "loss": 2.8204, "step": 495150 }, { "epoch": 4.373090305374521, "grad_norm": 3.461103677749634, "learning_rate": 3.862084057718851e-06, "loss": 2.4203, "step": 495200 }, { "epoch": 4.373531853264805, "grad_norm": 6.621654987335205, "learning_rate": 3.8567305419412006e-06, "loss": 2.7064, "step": 495250 }, { "epoch": 4.373973401155089, "grad_norm": 7.302319049835205, "learning_rate": 3.8513805903398425e-06, "loss": 2.9604, "step": 495300 }, { "epoch": 4.374414949045374, "grad_norm": 8.344850540161133, "learning_rate": 3.846034203328003e-06, "loss": 2.2373, "step": 495350 }, { "epoch": 4.374856496935657, "grad_norm": 4.20330286026001, "learning_rate": 3.840691381318634e-06, "loss": 2.7215, "step": 495400 }, { "epoch": 4.375298044825942, "grad_norm": 3.1253185272216797, "learning_rate": 3.8353521247244395e-06, "loss": 3.0862, "step": 495450 }, { "epoch": 4.375739592716226, "grad_norm": 2.6909918785095215, "learning_rate": 3.83001643395784e-06, "loss": 2.8925, "step": 495500 }, { "epoch": 4.37618114060651, "grad_norm": 8.299981117248535, "learning_rate": 3.824684309430948e-06, "loss": 2.973, "step": 495550 }, { "epoch": 4.376622688496794, "grad_norm": 7.423221588134766, "learning_rate": 3.819355751555642e-06, "loss": 2.318, "step": 495600 }, { "epoch": 4.377064236387079, "grad_norm": 5.389336109161377, "learning_rate": 3.814030760743503e-06, "loss": 2.8716, "step": 495650 }, { "epoch": 4.377505784277362, "grad_norm": 11.332836151123047, "learning_rate": 3.808709337405847e-06, "loss": 2.4875, "step": 495700 }, { "epoch": 4.377947332167647, "grad_norm": 3.3566739559173584, "learning_rate": 3.8033914819537055e-06, "loss": 2.1668, "step": 495750 }, { "epoch": 4.378388880057931, "grad_norm": 4.247566223144531, "learning_rate": 3.7980771947978332e-06, "loss": 2.7938, "step": 495800 }, { "epoch": 4.378830427948215, "grad_norm": 6.384842872619629, "learning_rate": 3.792766476348719e-06, "loss": 2.641, "step": 495850 }, { "epoch": 4.379271975838499, "grad_norm": 4.1271891593933105, "learning_rate": 3.7874593270165626e-06, "loss": 2.7615, "step": 495900 }, { "epoch": 4.379713523728784, "grad_norm": 6.48358678817749, "learning_rate": 3.782155747211302e-06, "loss": 2.6312, "step": 495950 }, { "epoch": 4.380155071619068, "grad_norm": 2.640913963317871, "learning_rate": 3.7768557373425995e-06, "loss": 2.9414, "step": 496000 }, { "epoch": 4.380596619509352, "grad_norm": 3.4023020267486572, "learning_rate": 3.7715592978198267e-06, "loss": 2.3433, "step": 496050 }, { "epoch": 4.381038167399636, "grad_norm": 9.7786226272583, "learning_rate": 3.7662664290520954e-06, "loss": 2.5883, "step": 496100 }, { "epoch": 4.38147971528992, "grad_norm": 3.6445200443267822, "learning_rate": 3.760977131448229e-06, "loss": 2.4822, "step": 496150 }, { "epoch": 4.381921263180205, "grad_norm": 12.486221313476562, "learning_rate": 3.7556914054167893e-06, "loss": 2.9021, "step": 496200 }, { "epoch": 4.382362811070489, "grad_norm": 7.893527030944824, "learning_rate": 3.7504092513660384e-06, "loss": 2.7724, "step": 496250 }, { "epoch": 4.382804358960773, "grad_norm": 4.369170665740967, "learning_rate": 3.7451306697039946e-06, "loss": 2.5683, "step": 496300 }, { "epoch": 4.383245906851057, "grad_norm": 6.109033584594727, "learning_rate": 3.7398556608383763e-06, "loss": 2.8262, "step": 496350 }, { "epoch": 4.383687454741342, "grad_norm": 4.162629127502441, "learning_rate": 3.7345842251766293e-06, "loss": 2.6982, "step": 496400 }, { "epoch": 4.384129002631625, "grad_norm": 4.662022590637207, "learning_rate": 3.729316363125934e-06, "loss": 2.3535, "step": 496450 }, { "epoch": 4.38457055052191, "grad_norm": 10.860516548156738, "learning_rate": 3.724052075093193e-06, "loss": 2.6481, "step": 496500 }, { "epoch": 4.385012098412194, "grad_norm": 4.739659786224365, "learning_rate": 3.718791361485019e-06, "loss": 2.6599, "step": 496550 }, { "epoch": 4.385453646302478, "grad_norm": 5.504743576049805, "learning_rate": 3.7135342227077485e-06, "loss": 2.6638, "step": 496600 }, { "epoch": 4.385895194192762, "grad_norm": 4.221433162689209, "learning_rate": 3.7082806591674736e-06, "loss": 2.7499, "step": 496650 }, { "epoch": 4.386336742083047, "grad_norm": 5.793357849121094, "learning_rate": 3.7030306712699637e-06, "loss": 2.9668, "step": 496700 }, { "epoch": 4.38677828997333, "grad_norm": 3.4492151737213135, "learning_rate": 3.6977842594207502e-06, "loss": 2.613, "step": 496750 }, { "epoch": 4.387219837863615, "grad_norm": 3.0164759159088135, "learning_rate": 3.6925414240250757e-06, "loss": 2.7272, "step": 496800 }, { "epoch": 4.387661385753899, "grad_norm": 6.834662914276123, "learning_rate": 3.6873021654878994e-06, "loss": 2.5736, "step": 496850 }, { "epoch": 4.388102933644183, "grad_norm": 4.137045383453369, "learning_rate": 3.6820664842139197e-06, "loss": 2.6202, "step": 496900 }, { "epoch": 4.388544481534467, "grad_norm": 8.54505729675293, "learning_rate": 3.6768343806075245e-06, "loss": 2.6455, "step": 496950 }, { "epoch": 4.388986029424752, "grad_norm": 1.940305233001709, "learning_rate": 3.671605855072874e-06, "loss": 2.8854, "step": 497000 }, { "epoch": 4.389427577315035, "grad_norm": 3.266374349594116, "learning_rate": 3.6663809080138112e-06, "loss": 2.6888, "step": 497050 }, { "epoch": 4.38986912520532, "grad_norm": 6.817994117736816, "learning_rate": 3.6611595398339248e-06, "loss": 2.9007, "step": 497100 }, { "epoch": 4.390310673095604, "grad_norm": 11.417515754699707, "learning_rate": 3.6559417509365313e-06, "loss": 2.5808, "step": 497150 }, { "epoch": 4.390752220985888, "grad_norm": 6.901117324829102, "learning_rate": 3.6507275417246412e-06, "loss": 2.4745, "step": 497200 }, { "epoch": 4.391193768876172, "grad_norm": 2.668447971343994, "learning_rate": 3.645516912601027e-06, "loss": 2.6691, "step": 497250 }, { "epoch": 4.391635316766457, "grad_norm": 7.112732410430908, "learning_rate": 3.6403098639681498e-06, "loss": 2.3842, "step": 497300 }, { "epoch": 4.39207686465674, "grad_norm": 4.013757228851318, "learning_rate": 3.635106396228227e-06, "loss": 2.7808, "step": 497350 }, { "epoch": 4.392518412547025, "grad_norm": 2.8971681594848633, "learning_rate": 3.62990650978316e-06, "loss": 2.7022, "step": 497400 }, { "epoch": 4.392959960437309, "grad_norm": 4.496314525604248, "learning_rate": 3.624710205034609e-06, "loss": 2.8102, "step": 497450 }, { "epoch": 4.393401508327593, "grad_norm": 8.755962371826172, "learning_rate": 3.619517482383955e-06, "loss": 3.0471, "step": 497500 }, { "epoch": 4.393843056217877, "grad_norm": 6.763393878936768, "learning_rate": 3.6143283422322704e-06, "loss": 2.4427, "step": 497550 }, { "epoch": 4.394284604108162, "grad_norm": 9.798931121826172, "learning_rate": 3.6091427849803793e-06, "loss": 2.4753, "step": 497600 }, { "epoch": 4.394726151998446, "grad_norm": 7.276642799377441, "learning_rate": 3.6039608110288393e-06, "loss": 2.4412, "step": 497650 }, { "epoch": 4.39516769988873, "grad_norm": 6.970615863800049, "learning_rate": 3.5987824207778964e-06, "loss": 2.7246, "step": 497700 }, { "epoch": 4.3956092477790145, "grad_norm": 4.831223011016846, "learning_rate": 3.5936076146275312e-06, "loss": 2.9178, "step": 497750 }, { "epoch": 4.396050795669298, "grad_norm": 8.82075023651123, "learning_rate": 3.588436392977473e-06, "loss": 2.7004, "step": 497800 }, { "epoch": 4.396492343559583, "grad_norm": 4.809752464294434, "learning_rate": 3.583268756227137e-06, "loss": 2.9262, "step": 497850 }, { "epoch": 4.396933891449867, "grad_norm": 1.3382608890533447, "learning_rate": 3.5781047047756867e-06, "loss": 2.9383, "step": 497900 }, { "epoch": 4.397375439340151, "grad_norm": 3.066117525100708, "learning_rate": 3.572944239021997e-06, "loss": 3.074, "step": 497950 }, { "epoch": 4.397816987230435, "grad_norm": 8.332267761230469, "learning_rate": 3.5677873593646883e-06, "loss": 2.5204, "step": 498000 }, { "epoch": 4.397816987230435, "eval_asr_loss": 0.9872112618912542, "eval_loss": 2.69632625579834, "eval_runtime": 20.2922, "eval_samples_per_second": 37.847, "eval_steps_per_second": 9.462, "eval_tts_loss": 6.343693301607574, "step": 498000 }, { "epoch": 4.3982585351207195, "grad_norm": 8.73320198059082, "learning_rate": 3.562634066202075e-06, "loss": 1.9239, "step": 498050 }, { "epoch": 4.398700083011003, "grad_norm": 4.928307056427002, "learning_rate": 3.5574843599321893e-06, "loss": 2.7929, "step": 498100 }, { "epoch": 4.399141630901288, "grad_norm": 5.043257236480713, "learning_rate": 3.552338240952824e-06, "loss": 2.7832, "step": 498150 }, { "epoch": 4.399583178791572, "grad_norm": 8.829334259033203, "learning_rate": 3.5471957096614604e-06, "loss": 2.5067, "step": 498200 }, { "epoch": 4.400024726681856, "grad_norm": 10.359509468078613, "learning_rate": 3.5420567664553205e-06, "loss": 2.6574, "step": 498250 }, { "epoch": 4.40046627457214, "grad_norm": 3.560913562774658, "learning_rate": 3.5369214117313475e-06, "loss": 2.6163, "step": 498300 }, { "epoch": 4.4009078224624245, "grad_norm": 10.124200820922852, "learning_rate": 3.531789645886191e-06, "loss": 2.13, "step": 498350 }, { "epoch": 4.401349370352708, "grad_norm": 3.497392177581787, "learning_rate": 3.5266614693162503e-06, "loss": 2.2818, "step": 498400 }, { "epoch": 4.401790918242993, "grad_norm": 4.796292304992676, "learning_rate": 3.5215368824176366e-06, "loss": 3.271, "step": 498450 }, { "epoch": 4.402232466133277, "grad_norm": 3.663623094558716, "learning_rate": 3.5164158855861716e-06, "loss": 2.7836, "step": 498500 }, { "epoch": 4.402674014023561, "grad_norm": 3.5663838386535645, "learning_rate": 3.511298479217401e-06, "loss": 2.4498, "step": 498550 }, { "epoch": 4.403115561913845, "grad_norm": 6.261927127838135, "learning_rate": 3.5061846637066075e-06, "loss": 2.7395, "step": 498600 }, { "epoch": 4.4035571098041295, "grad_norm": 8.087141990661621, "learning_rate": 3.5010744394488037e-06, "loss": 2.7007, "step": 498650 }, { "epoch": 4.403998657694413, "grad_norm": 5.8213629722595215, "learning_rate": 3.495967806838696e-06, "loss": 2.4557, "step": 498700 }, { "epoch": 4.404440205584698, "grad_norm": 3.6012916564941406, "learning_rate": 3.4908647662707293e-06, "loss": 2.5006, "step": 498750 }, { "epoch": 4.404881753474982, "grad_norm": 2.0803940296173096, "learning_rate": 3.485765318139078e-06, "loss": 2.3594, "step": 498800 }, { "epoch": 4.405323301365266, "grad_norm": 4.458380222320557, "learning_rate": 3.4806694628376325e-06, "loss": 2.8305, "step": 498850 }, { "epoch": 4.40576484925555, "grad_norm": 2.3031373023986816, "learning_rate": 3.4755772007599886e-06, "loss": 2.6961, "step": 498900 }, { "epoch": 4.4062063971458345, "grad_norm": 8.380064964294434, "learning_rate": 3.470488532299493e-06, "loss": 2.9287, "step": 498950 }, { "epoch": 4.406647945036118, "grad_norm": 3.5424811840057373, "learning_rate": 3.465403457849198e-06, "loss": 2.5565, "step": 499000 }, { "epoch": 4.407089492926403, "grad_norm": 6.18345308303833, "learning_rate": 3.4603219778018837e-06, "loss": 2.7694, "step": 499050 }, { "epoch": 4.407531040816687, "grad_norm": 5.246705532073975, "learning_rate": 3.4552440925500473e-06, "loss": 2.6162, "step": 499100 }, { "epoch": 4.407972588706971, "grad_norm": 6.736474990844727, "learning_rate": 3.450169802485931e-06, "loss": 2.6621, "step": 499150 }, { "epoch": 4.408414136597255, "grad_norm": 6.412590503692627, "learning_rate": 3.445099108001454e-06, "loss": 2.5186, "step": 499200 }, { "epoch": 4.4088556844875395, "grad_norm": 9.670331954956055, "learning_rate": 3.440032009488303e-06, "loss": 2.7925, "step": 499250 }, { "epoch": 4.409297232377824, "grad_norm": 5.673004627227783, "learning_rate": 3.4349685073378657e-06, "loss": 2.3885, "step": 499300 }, { "epoch": 4.409738780268108, "grad_norm": 6.605766296386719, "learning_rate": 3.429908601941245e-06, "loss": 2.823, "step": 499350 }, { "epoch": 4.4101803281583924, "grad_norm": 5.575830459594727, "learning_rate": 3.424852293689279e-06, "loss": 2.8304, "step": 499400 }, { "epoch": 4.410621876048676, "grad_norm": 4.133782863616943, "learning_rate": 3.419799582972538e-06, "loss": 2.9756, "step": 499450 }, { "epoch": 4.411063423938961, "grad_norm": 5.457077503204346, "learning_rate": 3.4147504701812773e-06, "loss": 2.9464, "step": 499500 }, { "epoch": 4.4115049718292445, "grad_norm": 7.1585259437561035, "learning_rate": 3.409704955705517e-06, "loss": 2.3468, "step": 499550 }, { "epoch": 4.411946519719529, "grad_norm": 7.983564853668213, "learning_rate": 3.40466303993498e-06, "loss": 2.8438, "step": 499600 }, { "epoch": 4.412388067609813, "grad_norm": 9.028219223022461, "learning_rate": 3.3996247232591095e-06, "loss": 2.6969, "step": 499650 }, { "epoch": 4.4128296155000974, "grad_norm": 4.264166355133057, "learning_rate": 3.3945900060670554e-06, "loss": 2.8717, "step": 499700 }, { "epoch": 4.413271163390381, "grad_norm": 5.1253557205200195, "learning_rate": 3.389558888747729e-06, "loss": 2.7992, "step": 499750 }, { "epoch": 4.413712711280666, "grad_norm": 4.655467987060547, "learning_rate": 3.3845313716897363e-06, "loss": 2.4879, "step": 499800 }, { "epoch": 4.4141542591709495, "grad_norm": 5.840618133544922, "learning_rate": 3.3795074552814e-06, "loss": 2.657, "step": 499850 }, { "epoch": 4.414595807061234, "grad_norm": 4.372864723205566, "learning_rate": 3.374487139910781e-06, "loss": 2.4845, "step": 499900 }, { "epoch": 4.415037354951518, "grad_norm": 4.493389129638672, "learning_rate": 3.3694704259656708e-06, "loss": 2.1426, "step": 499950 }, { "epoch": 4.4154789028418024, "grad_norm": 3.087900400161743, "learning_rate": 3.3644573138335467e-06, "loss": 2.413, "step": 500000 }, { "epoch": 4.415920450732086, "grad_norm": 5.623600006103516, "learning_rate": 3.359547958796211e-06, "loss": 2.4109, "step": 500050 }, { "epoch": 4.416361998622371, "grad_norm": 6.324456214904785, "learning_rate": 3.35454197939592e-06, "loss": 2.4842, "step": 500100 }, { "epoch": 4.4168035465126545, "grad_norm": 7.972682476043701, "learning_rate": 3.349539602961721e-06, "loss": 2.3133, "step": 500150 }, { "epoch": 4.417245094402939, "grad_norm": 3.4662911891937256, "learning_rate": 3.3445408298800106e-06, "loss": 2.867, "step": 500200 }, { "epoch": 4.417686642293223, "grad_norm": 7.298301696777344, "learning_rate": 3.3395456605368914e-06, "loss": 2.8735, "step": 500250 }, { "epoch": 4.418128190183507, "grad_norm": 6.748721122741699, "learning_rate": 3.3345540953182043e-06, "loss": 2.6407, "step": 500300 }, { "epoch": 4.418569738073791, "grad_norm": 4.278111457824707, "learning_rate": 3.329566134609502e-06, "loss": 2.5425, "step": 500350 }, { "epoch": 4.419011285964076, "grad_norm": 7.111077308654785, "learning_rate": 3.3245817787960543e-06, "loss": 2.736, "step": 500400 }, { "epoch": 4.4194528338543595, "grad_norm": 9.828688621520996, "learning_rate": 3.3196010282628643e-06, "loss": 2.7557, "step": 500450 }, { "epoch": 4.419894381744644, "grad_norm": 7.756462097167969, "learning_rate": 3.3146238833946686e-06, "loss": 2.42, "step": 500500 }, { "epoch": 4.420335929634928, "grad_norm": 3.9191176891326904, "learning_rate": 3.309650344575893e-06, "loss": 2.8775, "step": 500550 }, { "epoch": 4.420777477525212, "grad_norm": 9.661906242370605, "learning_rate": 3.3046804121907027e-06, "loss": 2.5111, "step": 500600 }, { "epoch": 4.421219025415496, "grad_norm": 6.857993125915527, "learning_rate": 3.29971408662299e-06, "loss": 2.5878, "step": 500650 }, { "epoch": 4.421660573305781, "grad_norm": 7.093061923980713, "learning_rate": 3.2947513682563545e-06, "loss": 2.7002, "step": 500700 }, { "epoch": 4.422102121196065, "grad_norm": 10.197794914245605, "learning_rate": 3.2897922574741335e-06, "loss": 2.9965, "step": 500750 }, { "epoch": 4.422543669086349, "grad_norm": 2.9524922370910645, "learning_rate": 3.2848367546593597e-06, "loss": 2.5481, "step": 500800 }, { "epoch": 4.422985216976633, "grad_norm": 7.795102119445801, "learning_rate": 3.279884860194821e-06, "loss": 3.183, "step": 500850 }, { "epoch": 4.423426764866917, "grad_norm": 3.3952646255493164, "learning_rate": 3.274936574463e-06, "loss": 2.753, "step": 500900 }, { "epoch": 4.423868312757202, "grad_norm": 6.756287097930908, "learning_rate": 3.269991897846103e-06, "loss": 2.2003, "step": 500950 }, { "epoch": 4.424309860647486, "grad_norm": 5.8866868019104, "learning_rate": 3.265050830726085e-06, "loss": 2.8932, "step": 501000 }, { "epoch": 4.424309860647486, "eval_asr_loss": 0.9890346454276767, "eval_loss": 2.6954333782196045, "eval_runtime": 20.8244, "eval_samples_per_second": 36.88, "eval_steps_per_second": 9.22, "eval_tts_loss": 6.324478099612424, "step": 501000 }, { "epoch": 4.42475140853777, "grad_norm": 5.748182773590088, "learning_rate": 3.2601133734845855e-06, "loss": 2.6868, "step": 501050 }, { "epoch": 4.425192956428054, "grad_norm": 4.182084083557129, "learning_rate": 3.2551795265029883e-06, "loss": 2.6701, "step": 501100 }, { "epoch": 4.425634504318339, "grad_norm": 2.31984281539917, "learning_rate": 3.250249290162388e-06, "loss": 2.5328, "step": 501150 }, { "epoch": 4.426076052208622, "grad_norm": 6.9561944007873535, "learning_rate": 3.2453226648435975e-06, "loss": 2.6013, "step": 501200 }, { "epoch": 4.426517600098907, "grad_norm": 5.562617301940918, "learning_rate": 3.2403996509271616e-06, "loss": 2.8489, "step": 501250 }, { "epoch": 4.426959147989191, "grad_norm": 6.537100791931152, "learning_rate": 3.2354802487933435e-06, "loss": 3.1363, "step": 501300 }, { "epoch": 4.427400695879475, "grad_norm": 2.7253119945526123, "learning_rate": 3.230564458822133e-06, "loss": 2.5982, "step": 501350 }, { "epoch": 4.427842243769759, "grad_norm": 5.052073001861572, "learning_rate": 3.225652281393221e-06, "loss": 3.2784, "step": 501400 }, { "epoch": 4.428283791660044, "grad_norm": 3.544499397277832, "learning_rate": 3.2207437168860377e-06, "loss": 2.7883, "step": 501450 }, { "epoch": 4.428725339550327, "grad_norm": 8.78883171081543, "learning_rate": 3.2158387656797186e-06, "loss": 2.5748, "step": 501500 }, { "epoch": 4.429166887440612, "grad_norm": 4.4148335456848145, "learning_rate": 3.210937428153149e-06, "loss": 2.3092, "step": 501550 }, { "epoch": 4.429608435330896, "grad_norm": 5.384941577911377, "learning_rate": 3.206039704684888e-06, "loss": 2.8995, "step": 501600 }, { "epoch": 4.43004998322118, "grad_norm": 3.0761842727661133, "learning_rate": 3.201145595653271e-06, "loss": 2.5516, "step": 501650 }, { "epoch": 4.430491531111464, "grad_norm": 3.2083773612976074, "learning_rate": 3.1962551014363074e-06, "loss": 2.7741, "step": 501700 }, { "epoch": 4.430933079001749, "grad_norm": 6.681820869445801, "learning_rate": 3.19136822241175e-06, "loss": 2.6138, "step": 501750 }, { "epoch": 4.431374626892032, "grad_norm": 6.994941711425781, "learning_rate": 3.186484958957081e-06, "loss": 2.9849, "step": 501800 }, { "epoch": 4.431816174782317, "grad_norm": 8.686388969421387, "learning_rate": 3.181605311449476e-06, "loss": 2.4024, "step": 501850 }, { "epoch": 4.432257722672601, "grad_norm": 7.352993488311768, "learning_rate": 3.1767292802658556e-06, "loss": 2.7049, "step": 501900 }, { "epoch": 4.432699270562885, "grad_norm": 4.320716381072998, "learning_rate": 3.1718568657828463e-06, "loss": 3.1115, "step": 501950 }, { "epoch": 4.433140818453169, "grad_norm": 5.14012336730957, "learning_rate": 3.1669880683768084e-06, "loss": 2.5034, "step": 502000 }, { "epoch": 4.433582366343454, "grad_norm": 6.663711071014404, "learning_rate": 3.1621228884238017e-06, "loss": 2.5749, "step": 502050 }, { "epoch": 4.434023914233737, "grad_norm": 6.052504539489746, "learning_rate": 3.1572613262996264e-06, "loss": 2.7529, "step": 502100 }, { "epoch": 4.434465462124022, "grad_norm": 7.062582015991211, "learning_rate": 3.1524033823798094e-06, "loss": 2.5217, "step": 502150 }, { "epoch": 4.434907010014306, "grad_norm": 7.203901767730713, "learning_rate": 3.1475490570395674e-06, "loss": 2.77, "step": 502200 }, { "epoch": 4.43534855790459, "grad_norm": 9.331632614135742, "learning_rate": 3.1426983506538664e-06, "loss": 2.6322, "step": 502250 }, { "epoch": 4.435790105794874, "grad_norm": 4.268424987792969, "learning_rate": 3.1378512635973856e-06, "loss": 2.5895, "step": 502300 }, { "epoch": 4.436231653685159, "grad_norm": 2.895256757736206, "learning_rate": 3.1330077962445026e-06, "loss": 2.7232, "step": 502350 }, { "epoch": 4.436673201575443, "grad_norm": 5.697513103485107, "learning_rate": 3.1281679489693514e-06, "loss": 2.8539, "step": 502400 }, { "epoch": 4.437114749465727, "grad_norm": 5.746689319610596, "learning_rate": 3.1233317221457616e-06, "loss": 2.4255, "step": 502450 }, { "epoch": 4.437556297356012, "grad_norm": 2.2208187580108643, "learning_rate": 3.1184991161473054e-06, "loss": 2.525, "step": 502500 }, { "epoch": 4.437997845246295, "grad_norm": 4.062807559967041, "learning_rate": 3.113670131347235e-06, "loss": 2.6824, "step": 502550 }, { "epoch": 4.43843939313658, "grad_norm": 1.995216727256775, "learning_rate": 3.1088447681185685e-06, "loss": 2.4259, "step": 502600 }, { "epoch": 4.438880941026864, "grad_norm": 3.1159825325012207, "learning_rate": 3.1040230268340244e-06, "loss": 2.3316, "step": 502650 }, { "epoch": 4.439322488917148, "grad_norm": 3.679849624633789, "learning_rate": 3.0992049078660323e-06, "loss": 2.7578, "step": 502700 }, { "epoch": 4.439764036807432, "grad_norm": 3.568610429763794, "learning_rate": 3.0943904115867508e-06, "loss": 2.0785, "step": 502750 }, { "epoch": 4.440205584697717, "grad_norm": 6.784422397613525, "learning_rate": 3.0895795383680647e-06, "loss": 2.6333, "step": 502800 }, { "epoch": 4.440647132588, "grad_norm": 7.078823089599609, "learning_rate": 3.084772288581567e-06, "loss": 2.7014, "step": 502850 }, { "epoch": 4.441088680478285, "grad_norm": 4.537209987640381, "learning_rate": 3.079968662598576e-06, "loss": 3.0074, "step": 502900 }, { "epoch": 4.441530228368569, "grad_norm": 8.451234817504883, "learning_rate": 3.0751686607901454e-06, "loss": 2.3762, "step": 502950 }, { "epoch": 4.441971776258853, "grad_norm": 5.459914684295654, "learning_rate": 3.0703722835270178e-06, "loss": 2.5208, "step": 503000 }, { "epoch": 4.442413324149137, "grad_norm": 7.786344051361084, "learning_rate": 3.0655795311796907e-06, "loss": 2.9205, "step": 503050 }, { "epoch": 4.442854872039422, "grad_norm": 5.491791248321533, "learning_rate": 3.060790404118341e-06, "loss": 2.5886, "step": 503100 }, { "epoch": 4.443296419929705, "grad_norm": 3.568432331085205, "learning_rate": 3.0560049027129055e-06, "loss": 2.6598, "step": 503150 }, { "epoch": 4.44373796781999, "grad_norm": 9.092747688293457, "learning_rate": 3.051318629303174e-06, "loss": 2.3118, "step": 503200 }, { "epoch": 4.444179515710274, "grad_norm": 5.176575183868408, "learning_rate": 3.046540307786672e-06, "loss": 3.0424, "step": 503250 }, { "epoch": 4.444621063600558, "grad_norm": 8.698075294494629, "learning_rate": 3.041765613026776e-06, "loss": 2.5758, "step": 503300 }, { "epoch": 4.445062611490842, "grad_norm": 2.455986976623535, "learning_rate": 3.036994545392302e-06, "loss": 2.3892, "step": 503350 }, { "epoch": 4.445504159381127, "grad_norm": 3.5403892993927, "learning_rate": 3.032227105251767e-06, "loss": 3.1884, "step": 503400 }, { "epoch": 4.44594570727141, "grad_norm": 3.729332447052002, "learning_rate": 3.0274632929734035e-06, "loss": 2.8611, "step": 503450 }, { "epoch": 4.446387255161695, "grad_norm": 5.045379638671875, "learning_rate": 3.0227031089251846e-06, "loss": 2.7771, "step": 503500 }, { "epoch": 4.446828803051979, "grad_norm": 5.6193132400512695, "learning_rate": 3.0179465534747985e-06, "loss": 2.4733, "step": 503550 }, { "epoch": 4.447270350942263, "grad_norm": 6.6143012046813965, "learning_rate": 3.013193626989641e-06, "loss": 2.9781, "step": 503600 }, { "epoch": 4.447711898832547, "grad_norm": 3.929274797439575, "learning_rate": 3.0084443298368402e-06, "loss": 2.6336, "step": 503650 }, { "epoch": 4.448153446722832, "grad_norm": 8.65351390838623, "learning_rate": 3.0036986623832474e-06, "loss": 2.6497, "step": 503700 }, { "epoch": 4.448594994613115, "grad_norm": 5.2679290771484375, "learning_rate": 2.9989566249954083e-06, "loss": 2.5495, "step": 503750 }, { "epoch": 4.4490365425034, "grad_norm": 6.3077874183654785, "learning_rate": 2.9942182180396126e-06, "loss": 2.6256, "step": 503800 }, { "epoch": 4.449478090393684, "grad_norm": 3.5381276607513428, "learning_rate": 2.9894834418818683e-06, "loss": 2.6053, "step": 503850 }, { "epoch": 4.449919638283968, "grad_norm": 7.133114814758301, "learning_rate": 2.9847522968878828e-06, "loss": 2.3847, "step": 503900 }, { "epoch": 4.450361186174252, "grad_norm": 2.796760320663452, "learning_rate": 2.980024783423102e-06, "loss": 2.7074, "step": 503950 }, { "epoch": 4.450802734064537, "grad_norm": 6.723973751068115, "learning_rate": 2.975300901852701e-06, "loss": 2.4446, "step": 504000 }, { "epoch": 4.450802734064537, "eval_asr_loss": 0.9891451828977823, "eval_loss": 2.6955654621124268, "eval_runtime": 20.9454, "eval_samples_per_second": 36.667, "eval_steps_per_second": 9.167, "eval_tts_loss": 6.33350034678933, "step": 504000 }, { "epoch": 4.451244281954821, "grad_norm": 4.496365070343018, "learning_rate": 2.9705806525415324e-06, "loss": 2.6161, "step": 504050 }, { "epoch": 4.451685829845105, "grad_norm": 2.776057004928589, "learning_rate": 2.9658640358542157e-06, "loss": 2.8961, "step": 504100 }, { "epoch": 4.45212737773539, "grad_norm": 7.031434059143066, "learning_rate": 2.9611510521550757e-06, "loss": 2.451, "step": 504150 }, { "epoch": 4.452568925625673, "grad_norm": 3.487990379333496, "learning_rate": 2.9564417018081327e-06, "loss": 2.7164, "step": 504200 }, { "epoch": 4.453010473515958, "grad_norm": 9.520648002624512, "learning_rate": 2.9517359851771455e-06, "loss": 2.3624, "step": 504250 }, { "epoch": 4.453452021406242, "grad_norm": 8.773432731628418, "learning_rate": 2.9470339026255955e-06, "loss": 2.7117, "step": 504300 }, { "epoch": 4.453893569296526, "grad_norm": 3.845036268234253, "learning_rate": 2.942335454516687e-06, "loss": 2.5671, "step": 504350 }, { "epoch": 4.45433511718681, "grad_norm": 3.534548044204712, "learning_rate": 2.9376406412133185e-06, "loss": 2.7625, "step": 504400 }, { "epoch": 4.454776665077095, "grad_norm": 3.005701780319214, "learning_rate": 2.9329494630781387e-06, "loss": 2.924, "step": 504450 }, { "epoch": 4.455218212967378, "grad_norm": 5.729005813598633, "learning_rate": 2.9282619204734964e-06, "loss": 2.5957, "step": 504500 }, { "epoch": 4.455659760857663, "grad_norm": 5.37128210067749, "learning_rate": 2.9235780137614687e-06, "loss": 2.2468, "step": 504550 }, { "epoch": 4.456101308747947, "grad_norm": 8.232932090759277, "learning_rate": 2.9188977433038333e-06, "loss": 2.8315, "step": 504600 }, { "epoch": 4.456542856638231, "grad_norm": 4.472603797912598, "learning_rate": 2.9142211094621117e-06, "loss": 2.5863, "step": 504650 }, { "epoch": 4.456984404528515, "grad_norm": 7.755439758300781, "learning_rate": 2.909548112597543e-06, "loss": 2.8398, "step": 504700 }, { "epoch": 4.4574259524188, "grad_norm": 4.0629682540893555, "learning_rate": 2.9048787530710553e-06, "loss": 2.6118, "step": 504750 }, { "epoch": 4.457867500309083, "grad_norm": 5.503448963165283, "learning_rate": 2.900213031243332e-06, "loss": 2.5927, "step": 504800 }, { "epoch": 4.458309048199368, "grad_norm": 5.169827938079834, "learning_rate": 2.8955509474747623e-06, "loss": 2.7459, "step": 504850 }, { "epoch": 4.458750596089652, "grad_norm": 2.580535650253296, "learning_rate": 2.890892502125442e-06, "loss": 2.6677, "step": 504900 }, { "epoch": 4.459192143979936, "grad_norm": 7.830506324768066, "learning_rate": 2.8862376955552107e-06, "loss": 2.905, "step": 504950 }, { "epoch": 4.45963369187022, "grad_norm": 12.235816955566406, "learning_rate": 2.8815865281236087e-06, "loss": 2.4414, "step": 505000 }, { "epoch": 4.460075239760505, "grad_norm": 6.894923686981201, "learning_rate": 2.8769390001898823e-06, "loss": 2.8276, "step": 505050 }, { "epoch": 4.460516787650788, "grad_norm": 9.108841896057129, "learning_rate": 2.872295112113027e-06, "loss": 2.4274, "step": 505100 }, { "epoch": 4.460958335541073, "grad_norm": 5.785190105438232, "learning_rate": 2.8676548642517452e-06, "loss": 2.9633, "step": 505150 }, { "epoch": 4.461399883431357, "grad_norm": 6.775511741638184, "learning_rate": 2.863018256964456e-06, "loss": 2.3983, "step": 505200 }, { "epoch": 4.461841431321641, "grad_norm": 3.399498462677002, "learning_rate": 2.858385290609289e-06, "loss": 2.8518, "step": 505250 }, { "epoch": 4.462282979211925, "grad_norm": 3.506645917892456, "learning_rate": 2.853755965544125e-06, "loss": 2.7847, "step": 505300 }, { "epoch": 4.46272452710221, "grad_norm": 4.73546028137207, "learning_rate": 2.8491302821265166e-06, "loss": 2.918, "step": 505350 }, { "epoch": 4.463166074992493, "grad_norm": 2.952918291091919, "learning_rate": 2.8445082407137615e-06, "loss": 2.3155, "step": 505400 }, { "epoch": 4.463607622882778, "grad_norm": 8.715761184692383, "learning_rate": 2.839982173946443e-06, "loss": 2.6576, "step": 505450 }, { "epoch": 4.4640491707730625, "grad_norm": 10.68049430847168, "learning_rate": 2.8353673447562955e-06, "loss": 2.8925, "step": 505500 }, { "epoch": 4.464490718663346, "grad_norm": 2.959425449371338, "learning_rate": 2.8307561586340847e-06, "loss": 3.0885, "step": 505550 }, { "epoch": 4.46493226655363, "grad_norm": 7.43977165222168, "learning_rate": 2.8261486159359695e-06, "loss": 3.1053, "step": 505600 }, { "epoch": 4.465373814443915, "grad_norm": 8.44040298461914, "learning_rate": 2.82154471701786e-06, "loss": 2.6164, "step": 505650 }, { "epoch": 4.465815362334199, "grad_norm": 3.0988612174987793, "learning_rate": 2.8169444622353547e-06, "loss": 2.9143, "step": 505700 }, { "epoch": 4.466256910224483, "grad_norm": 7.040312767028809, "learning_rate": 2.812347851943792e-06, "loss": 2.3999, "step": 505750 }, { "epoch": 4.4666984581147675, "grad_norm": 2.0352437496185303, "learning_rate": 2.8077548864982096e-06, "loss": 2.6573, "step": 505800 }, { "epoch": 4.467140006005051, "grad_norm": 4.38501501083374, "learning_rate": 2.803165566253385e-06, "loss": 3.1953, "step": 505850 }, { "epoch": 4.467581553895336, "grad_norm": 8.443303108215332, "learning_rate": 2.798579891563813e-06, "loss": 2.615, "step": 505900 }, { "epoch": 4.46802310178562, "grad_norm": 6.906628131866455, "learning_rate": 2.7939978627836816e-06, "loss": 2.79, "step": 505950 }, { "epoch": 4.468464649675904, "grad_norm": 3.599278211593628, "learning_rate": 2.789419480266925e-06, "loss": 2.7766, "step": 506000 }, { "epoch": 4.468906197566188, "grad_norm": 5.969940662384033, "learning_rate": 2.7848447443671766e-06, "loss": 2.2348, "step": 506050 }, { "epoch": 4.4693477454564725, "grad_norm": 4.9454874992370605, "learning_rate": 2.780273655437793e-06, "loss": 2.6132, "step": 506100 }, { "epoch": 4.469789293346756, "grad_norm": 4.5790205001831055, "learning_rate": 2.775706213831858e-06, "loss": 2.7875, "step": 506150 }, { "epoch": 4.470230841237041, "grad_norm": 3.9868674278259277, "learning_rate": 2.7711424199021785e-06, "loss": 2.4249, "step": 506200 }, { "epoch": 4.470672389127325, "grad_norm": 6.139019966125488, "learning_rate": 2.7665822740012504e-06, "loss": 2.6061, "step": 506250 }, { "epoch": 4.471113937017609, "grad_norm": 2.299931526184082, "learning_rate": 2.7620257764813086e-06, "loss": 2.8693, "step": 506300 }, { "epoch": 4.471555484907893, "grad_norm": 4.201018810272217, "learning_rate": 2.7574729276943155e-06, "loss": 3.2359, "step": 506350 }, { "epoch": 4.4719970327981775, "grad_norm": 4.893449783325195, "learning_rate": 2.752923727991935e-06, "loss": 2.5401, "step": 506400 }, { "epoch": 4.472438580688461, "grad_norm": 7.609147071838379, "learning_rate": 2.748378177725558e-06, "loss": 2.3562, "step": 506450 }, { "epoch": 4.472880128578746, "grad_norm": 8.549010276794434, "learning_rate": 2.7438362772462758e-06, "loss": 2.7778, "step": 506500 }, { "epoch": 4.47332167646903, "grad_norm": 8.606903076171875, "learning_rate": 2.739298026904924e-06, "loss": 2.6069, "step": 506550 }, { "epoch": 4.473763224359314, "grad_norm": 2.934004545211792, "learning_rate": 2.7347634270520396e-06, "loss": 2.7782, "step": 506600 }, { "epoch": 4.474204772249598, "grad_norm": 4.043940544128418, "learning_rate": 2.7302324780378806e-06, "loss": 3.0971, "step": 506650 }, { "epoch": 4.4746463201398825, "grad_norm": 8.405601501464844, "learning_rate": 2.7257051802124343e-06, "loss": 2.4528, "step": 506700 }, { "epoch": 4.475087868030166, "grad_norm": 3.5292389392852783, "learning_rate": 2.7211815339253764e-06, "loss": 2.738, "step": 506750 }, { "epoch": 4.475529415920451, "grad_norm": 6.936749458312988, "learning_rate": 2.716661539526144e-06, "loss": 2.3379, "step": 506800 }, { "epoch": 4.475970963810735, "grad_norm": 6.08766508102417, "learning_rate": 2.712145197363852e-06, "loss": 2.4677, "step": 506850 }, { "epoch": 4.476412511701019, "grad_norm": 7.432994842529297, "learning_rate": 2.707632507787361e-06, "loss": 2.46, "step": 506900 }, { "epoch": 4.476854059591303, "grad_norm": 5.452535152435303, "learning_rate": 2.7031234711452193e-06, "loss": 2.7778, "step": 506950 }, { "epoch": 4.4772956074815875, "grad_norm": 5.277121067047119, "learning_rate": 2.698618087785726e-06, "loss": 2.5447, "step": 507000 }, { "epoch": 4.4772956074815875, "eval_asr_loss": 0.990462632257494, "eval_loss": 2.69331431388855, "eval_runtime": 20.4322, "eval_samples_per_second": 37.588, "eval_steps_per_second": 9.397, "eval_tts_loss": 6.335356707720737, "step": 507000 }, { "epoch": 4.477737155371871, "grad_norm": 4.299010753631592, "learning_rate": 2.6941163580568863e-06, "loss": 2.5515, "step": 507050 }, { "epoch": 4.478178703262156, "grad_norm": 4.010782718658447, "learning_rate": 2.689708208010189e-06, "loss": 2.9484, "step": 507100 }, { "epoch": 4.4786202511524404, "grad_norm": 5.737813472747803, "learning_rate": 2.6852137134956102e-06, "loss": 3.3373, "step": 507150 }, { "epoch": 4.479061799042724, "grad_norm": 5.6044840812683105, "learning_rate": 2.680722873647046e-06, "loss": 2.7338, "step": 507200 }, { "epoch": 4.479503346933008, "grad_norm": 4.466466903686523, "learning_rate": 2.6762356888113904e-06, "loss": 2.1596, "step": 507250 }, { "epoch": 4.4799448948232925, "grad_norm": 4.172839164733887, "learning_rate": 2.671752159335228e-06, "loss": 2.3601, "step": 507300 }, { "epoch": 4.480386442713577, "grad_norm": 5.15049409866333, "learning_rate": 2.667272285564881e-06, "loss": 2.6384, "step": 507350 }, { "epoch": 4.480827990603861, "grad_norm": 4.17628288269043, "learning_rate": 2.6627960678463735e-06, "loss": 2.2443, "step": 507400 }, { "epoch": 4.481269538494145, "grad_norm": 4.306352615356445, "learning_rate": 2.6583235065254784e-06, "loss": 2.3406, "step": 507450 }, { "epoch": 4.481711086384429, "grad_norm": 6.305701732635498, "learning_rate": 2.653854601947642e-06, "loss": 2.5479, "step": 507500 }, { "epoch": 4.482152634274714, "grad_norm": 3.328376293182373, "learning_rate": 2.649389354458043e-06, "loss": 2.2847, "step": 507550 }, { "epoch": 4.4825941821649975, "grad_norm": 6.722079277038574, "learning_rate": 2.644927764401611e-06, "loss": 2.3998, "step": 507600 }, { "epoch": 4.483035730055282, "grad_norm": 4.665835380554199, "learning_rate": 2.640469832122938e-06, "loss": 2.6765, "step": 507650 }, { "epoch": 4.483477277945566, "grad_norm": 1.9743167161941528, "learning_rate": 2.6360155579663704e-06, "loss": 2.8644, "step": 507700 }, { "epoch": 4.48391882583585, "grad_norm": 4.686093807220459, "learning_rate": 2.631564942275977e-06, "loss": 2.7304, "step": 507750 }, { "epoch": 4.484360373726134, "grad_norm": 3.2579598426818848, "learning_rate": 2.6271179853955107e-06, "loss": 2.7735, "step": 507800 }, { "epoch": 4.484801921616419, "grad_norm": 4.287333965301514, "learning_rate": 2.622674687668469e-06, "loss": 2.6747, "step": 507850 }, { "epoch": 4.4852434695067025, "grad_norm": 5.07670783996582, "learning_rate": 2.6182350494380725e-06, "loss": 2.2257, "step": 507900 }, { "epoch": 4.485685017396987, "grad_norm": 8.58398151397705, "learning_rate": 2.6137990710472238e-06, "loss": 2.8473, "step": 507950 }, { "epoch": 4.486126565287271, "grad_norm": 8.43362808227539, "learning_rate": 2.609366752838571e-06, "loss": 2.1661, "step": 508000 }, { "epoch": 4.486568113177555, "grad_norm": 4.520970344543457, "learning_rate": 2.60493809515448e-06, "loss": 2.2112, "step": 508050 }, { "epoch": 4.487009661067839, "grad_norm": 4.9396257400512695, "learning_rate": 2.600513098337026e-06, "loss": 2.7269, "step": 508100 }, { "epoch": 4.487451208958124, "grad_norm": 2.8484814167022705, "learning_rate": 2.596091762727992e-06, "loss": 2.5309, "step": 508150 }, { "epoch": 4.4878927568484075, "grad_norm": 4.1148362159729, "learning_rate": 2.591674088668894e-06, "loss": 2.679, "step": 508200 }, { "epoch": 4.488334304738692, "grad_norm": 4.529597759246826, "learning_rate": 2.587260076500969e-06, "loss": 2.6835, "step": 508250 }, { "epoch": 4.488775852628976, "grad_norm": 6.977729320526123, "learning_rate": 2.5828497265651563e-06, "loss": 2.6244, "step": 508300 }, { "epoch": 4.48921740051926, "grad_norm": 4.746829509735107, "learning_rate": 2.5784430392021106e-06, "loss": 2.7871, "step": 508350 }, { "epoch": 4.489658948409544, "grad_norm": 6.095747947692871, "learning_rate": 2.574040014752227e-06, "loss": 2.4622, "step": 508400 }, { "epoch": 4.490100496299829, "grad_norm": 2.3914918899536133, "learning_rate": 2.5696406535555773e-06, "loss": 2.6199, "step": 508450 }, { "epoch": 4.4905420441901125, "grad_norm": 6.57155179977417, "learning_rate": 2.5652449559519953e-06, "loss": 3.047, "step": 508500 }, { "epoch": 4.490983592080397, "grad_norm": 8.89861011505127, "learning_rate": 2.560852922281004e-06, "loss": 2.1932, "step": 508550 }, { "epoch": 4.491425139970681, "grad_norm": 3.0847978591918945, "learning_rate": 2.5564645528818542e-06, "loss": 2.6029, "step": 508600 }, { "epoch": 4.491866687860965, "grad_norm": 6.448261260986328, "learning_rate": 2.5520798480935135e-06, "loss": 2.8432, "step": 508650 }, { "epoch": 4.492308235751249, "grad_norm": 5.7957000732421875, "learning_rate": 2.5476988082546504e-06, "loss": 2.9294, "step": 508700 }, { "epoch": 4.492749783641534, "grad_norm": 12.794177055358887, "learning_rate": 2.543321433703677e-06, "loss": 2.3962, "step": 508750 }, { "epoch": 4.493191331531818, "grad_norm": 3.4943554401397705, "learning_rate": 2.5389477247786955e-06, "loss": 2.3552, "step": 508800 }, { "epoch": 4.493632879422102, "grad_norm": 4.5879058837890625, "learning_rate": 2.534577681817546e-06, "loss": 2.7498, "step": 508850 }, { "epoch": 4.494074427312387, "grad_norm": 3.8878045082092285, "learning_rate": 2.5302113051577812e-06, "loss": 2.596, "step": 508900 }, { "epoch": 4.49451597520267, "grad_norm": 8.56757640838623, "learning_rate": 2.5258485951366537e-06, "loss": 2.5929, "step": 508950 }, { "epoch": 4.494957523092955, "grad_norm": 8.50552749633789, "learning_rate": 2.5214895520911498e-06, "loss": 2.4796, "step": 509000 }, { "epoch": 4.495399070983239, "grad_norm": 6.500808238983154, "learning_rate": 2.5171341763579827e-06, "loss": 2.5308, "step": 509050 }, { "epoch": 4.495840618873523, "grad_norm": 3.0848846435546875, "learning_rate": 2.5127824682735555e-06, "loss": 2.923, "step": 509100 }, { "epoch": 4.496282166763807, "grad_norm": 5.274893760681152, "learning_rate": 2.5084344281740004e-06, "loss": 2.5247, "step": 509150 }, { "epoch": 4.496723714654092, "grad_norm": 7.012965679168701, "learning_rate": 2.504090056395164e-06, "loss": 2.611, "step": 509200 }, { "epoch": 4.497165262544375, "grad_norm": 3.219149112701416, "learning_rate": 2.499749353272629e-06, "loss": 2.6367, "step": 509250 }, { "epoch": 4.49760681043466, "grad_norm": 4.502990245819092, "learning_rate": 2.4954123191416546e-06, "loss": 2.6487, "step": 509300 }, { "epoch": 4.498048358324944, "grad_norm": 12.694314956665039, "learning_rate": 2.4910789543372503e-06, "loss": 2.8028, "step": 509350 }, { "epoch": 4.498489906215228, "grad_norm": 5.193963527679443, "learning_rate": 2.4867492591941432e-06, "loss": 2.607, "step": 509400 }, { "epoch": 4.498931454105512, "grad_norm": 6.485684394836426, "learning_rate": 2.4824232340467545e-06, "loss": 2.6598, "step": 509450 }, { "epoch": 4.499373001995797, "grad_norm": 11.422167778015137, "learning_rate": 2.4781008792292282e-06, "loss": 3.0346, "step": 509500 }, { "epoch": 4.49981454988608, "grad_norm": 5.163692474365234, "learning_rate": 2.4737821950754412e-06, "loss": 2.4903, "step": 509550 }, { "epoch": 4.500256097776365, "grad_norm": 6.382083415985107, "learning_rate": 2.469467181918961e-06, "loss": 2.8824, "step": 509600 }, { "epoch": 4.500697645666649, "grad_norm": 3.305266857147217, "learning_rate": 2.4651558400930974e-06, "loss": 2.5321, "step": 509650 }, { "epoch": 4.501139193556933, "grad_norm": 6.174585342407227, "learning_rate": 2.4608481699308574e-06, "loss": 2.8708, "step": 509700 }, { "epoch": 4.501580741447217, "grad_norm": 4.823075771331787, "learning_rate": 2.456544171764985e-06, "loss": 2.697, "step": 509750 }, { "epoch": 4.502022289337502, "grad_norm": 7.238412380218506, "learning_rate": 2.4522438459279095e-06, "loss": 2.834, "step": 509800 }, { "epoch": 4.502463837227785, "grad_norm": 9.557720184326172, "learning_rate": 2.4479471927518095e-06, "loss": 2.5626, "step": 509850 }, { "epoch": 4.50290538511807, "grad_norm": 7.551178455352783, "learning_rate": 2.4436542125685636e-06, "loss": 2.8122, "step": 509900 }, { "epoch": 4.503346933008354, "grad_norm": 5.291790962219238, "learning_rate": 2.4393649057097513e-06, "loss": 2.5524, "step": 509950 }, { "epoch": 4.503788480898638, "grad_norm": 3.224613904953003, "learning_rate": 2.435079272506702e-06, "loss": 2.5892, "step": 510000 }, { "epoch": 4.503788480898638, "eval_asr_loss": 0.992308085443975, "eval_loss": 2.69207763671875, "eval_runtime": 20.5758, "eval_samples_per_second": 37.325, "eval_steps_per_second": 9.331, "eval_tts_loss": 6.327732535865692, "step": 510000 }, { "epoch": 4.504230028788922, "grad_norm": 5.051978588104248, "learning_rate": 2.4307973132904505e-06, "loss": 3.0432, "step": 510050 }, { "epoch": 4.504671576679207, "grad_norm": 3.3512673377990723, "learning_rate": 2.4265190283917163e-06, "loss": 2.1873, "step": 510100 }, { "epoch": 4.50511312456949, "grad_norm": 5.280250072479248, "learning_rate": 2.4222444181409787e-06, "loss": 2.6825, "step": 510150 }, { "epoch": 4.505554672459775, "grad_norm": 9.100662231445312, "learning_rate": 2.417973482868424e-06, "loss": 2.6305, "step": 510200 }, { "epoch": 4.50599622035006, "grad_norm": 6.533969879150391, "learning_rate": 2.4137062229039275e-06, "loss": 2.6393, "step": 510250 }, { "epoch": 4.506437768240343, "grad_norm": 3.438357353210449, "learning_rate": 2.409442638577103e-06, "loss": 2.3681, "step": 510300 }, { "epoch": 4.506879316130627, "grad_norm": 4.064580917358398, "learning_rate": 2.405182730217276e-06, "loss": 2.7937, "step": 510350 }, { "epoch": 4.507320864020912, "grad_norm": 4.506362438201904, "learning_rate": 2.4009264981534997e-06, "loss": 2.4461, "step": 510400 }, { "epoch": 4.507762411911196, "grad_norm": 6.743930816650391, "learning_rate": 2.3966739427145112e-06, "loss": 2.9043, "step": 510450 }, { "epoch": 4.50820395980148, "grad_norm": 6.472933769226074, "learning_rate": 2.392425064228804e-06, "loss": 2.8214, "step": 510500 }, { "epoch": 4.508645507691764, "grad_norm": 6.135166168212891, "learning_rate": 2.388179863024559e-06, "loss": 2.5904, "step": 510550 }, { "epoch": 4.509087055582048, "grad_norm": 2.452655553817749, "learning_rate": 2.383938339429681e-06, "loss": 2.2647, "step": 510600 }, { "epoch": 4.509528603472333, "grad_norm": 3.6834442615509033, "learning_rate": 2.3797004937718027e-06, "loss": 2.5212, "step": 510650 }, { "epoch": 4.509970151362617, "grad_norm": 3.7505040168762207, "learning_rate": 2.3754663263782506e-06, "loss": 2.7054, "step": 510700 }, { "epoch": 4.510411699252901, "grad_norm": 5.7797369956970215, "learning_rate": 2.371235837576069e-06, "loss": 2.6537, "step": 510750 }, { "epoch": 4.510853247143185, "grad_norm": 3.3506381511688232, "learning_rate": 2.3670090276920464e-06, "loss": 2.4641, "step": 510800 }, { "epoch": 4.51129479503347, "grad_norm": 8.65695571899414, "learning_rate": 2.362785897052655e-06, "loss": 2.7384, "step": 510850 }, { "epoch": 4.511736342923753, "grad_norm": 3.6663122177124023, "learning_rate": 2.3585664459841116e-06, "loss": 2.5214, "step": 510900 }, { "epoch": 4.512177890814038, "grad_norm": 4.0771284103393555, "learning_rate": 2.3543506748123112e-06, "loss": 2.5631, "step": 510950 }, { "epoch": 4.512619438704322, "grad_norm": 8.014383316040039, "learning_rate": 2.3501385838629098e-06, "loss": 2.503, "step": 511000 }, { "epoch": 4.513060986594606, "grad_norm": 4.092639446258545, "learning_rate": 2.3459301734612416e-06, "loss": 2.9249, "step": 511050 }, { "epoch": 4.51350253448489, "grad_norm": 6.880788326263428, "learning_rate": 2.341725443932369e-06, "loss": 3.1371, "step": 511100 }, { "epoch": 4.513944082375175, "grad_norm": 7.288794994354248, "learning_rate": 2.337524395601071e-06, "loss": 2.5837, "step": 511150 }, { "epoch": 4.514385630265458, "grad_norm": 5.13477897644043, "learning_rate": 2.3333270287918542e-06, "loss": 2.5491, "step": 511200 }, { "epoch": 4.514827178155743, "grad_norm": 9.204195022583008, "learning_rate": 2.329133343828921e-06, "loss": 2.9287, "step": 511250 }, { "epoch": 4.515268726046027, "grad_norm": 4.905004501342773, "learning_rate": 2.3249433410361954e-06, "loss": 2.6063, "step": 511300 }, { "epoch": 4.515710273936311, "grad_norm": 2.942941188812256, "learning_rate": 2.3207570207373297e-06, "loss": 3.0228, "step": 511350 }, { "epoch": 4.516151821826595, "grad_norm": 6.33992338180542, "learning_rate": 2.316574383255682e-06, "loss": 2.5938, "step": 511400 }, { "epoch": 4.51659336971688, "grad_norm": 5.846688270568848, "learning_rate": 2.3123954289143047e-06, "loss": 2.9928, "step": 511450 }, { "epoch": 4.517034917607163, "grad_norm": 6.291439533233643, "learning_rate": 2.3082201580360064e-06, "loss": 3.0119, "step": 511500 }, { "epoch": 4.517476465497448, "grad_norm": 4.40420389175415, "learning_rate": 2.304048570943296e-06, "loss": 2.4537, "step": 511550 }, { "epoch": 4.517918013387732, "grad_norm": 5.7835283279418945, "learning_rate": 2.299880667958371e-06, "loss": 2.6498, "step": 511600 }, { "epoch": 4.518359561278016, "grad_norm": 9.976414680480957, "learning_rate": 2.2957164494031845e-06, "loss": 3.3379, "step": 511650 }, { "epoch": 4.5188011091683, "grad_norm": 6.754440784454346, "learning_rate": 2.291555915599386e-06, "loss": 3.0428, "step": 511700 }, { "epoch": 4.519242657058585, "grad_norm": 5.708142280578613, "learning_rate": 2.2873990668683287e-06, "loss": 2.6248, "step": 511750 }, { "epoch": 4.519684204948868, "grad_norm": 7.191005706787109, "learning_rate": 2.2832459035311114e-06, "loss": 2.7663, "step": 511800 }, { "epoch": 4.520125752839153, "grad_norm": 6.20241641998291, "learning_rate": 2.2790964259085224e-06, "loss": 2.5405, "step": 511850 }, { "epoch": 4.520567300729438, "grad_norm": 2.2603037357330322, "learning_rate": 2.274950634321066e-06, "loss": 2.2744, "step": 511900 }, { "epoch": 4.521008848619721, "grad_norm": 3.2230684757232666, "learning_rate": 2.2708085290889747e-06, "loss": 3.1851, "step": 511950 }, { "epoch": 4.521450396510005, "grad_norm": 8.664597511291504, "learning_rate": 2.266670110532193e-06, "loss": 2.533, "step": 512000 }, { "epoch": 4.52189194440029, "grad_norm": 3.993757486343384, "learning_rate": 2.2625353789703874e-06, "loss": 2.9708, "step": 512050 }, { "epoch": 4.522333492290574, "grad_norm": 2.056365489959717, "learning_rate": 2.2584043347229135e-06, "loss": 2.7093, "step": 512100 }, { "epoch": 4.522775040180858, "grad_norm": 3.903543472290039, "learning_rate": 2.254276978108871e-06, "loss": 2.7475, "step": 512150 }, { "epoch": 4.523216588071143, "grad_norm": 6.805477142333984, "learning_rate": 2.2501533094470674e-06, "loss": 2.5581, "step": 512200 }, { "epoch": 4.523658135961426, "grad_norm": 8.96409797668457, "learning_rate": 2.246033329056002e-06, "loss": 2.7084, "step": 512250 }, { "epoch": 4.524099683851711, "grad_norm": 4.831679344177246, "learning_rate": 2.241917037253921e-06, "loss": 3.0083, "step": 512300 }, { "epoch": 4.524541231741995, "grad_norm": 3.589933156967163, "learning_rate": 2.2378044343587813e-06, "loss": 2.9347, "step": 512350 }, { "epoch": 4.524982779632279, "grad_norm": 5.3839592933654785, "learning_rate": 2.233695520688228e-06, "loss": 2.5879, "step": 512400 }, { "epoch": 4.525424327522563, "grad_norm": 4.184597015380859, "learning_rate": 2.229590296559647e-06, "loss": 2.8006, "step": 512450 }, { "epoch": 4.525865875412848, "grad_norm": 12.490345001220703, "learning_rate": 2.2254887622901454e-06, "loss": 2.6606, "step": 512500 }, { "epoch": 4.526307423303131, "grad_norm": 3.701486587524414, "learning_rate": 2.221390918196514e-06, "loss": 2.7819, "step": 512550 }, { "epoch": 4.526748971193416, "grad_norm": 2.733407974243164, "learning_rate": 2.2172967645952893e-06, "loss": 2.798, "step": 512600 }, { "epoch": 4.5271905190837, "grad_norm": 2.063913106918335, "learning_rate": 2.213206301802695e-06, "loss": 2.6013, "step": 512650 }, { "epoch": 4.527632066973984, "grad_norm": 9.061971664428711, "learning_rate": 2.2091195301347067e-06, "loss": 2.7698, "step": 512700 }, { "epoch": 4.528073614864268, "grad_norm": 4.106367111206055, "learning_rate": 2.2050364499069665e-06, "loss": 2.6219, "step": 512750 }, { "epoch": 4.528515162754553, "grad_norm": 5.750235080718994, "learning_rate": 2.2009570614348775e-06, "loss": 2.8809, "step": 512800 }, { "epoch": 4.528956710644836, "grad_norm": 4.764532089233398, "learning_rate": 2.1968813650335316e-06, "loss": 2.7117, "step": 512850 }, { "epoch": 4.529398258535121, "grad_norm": 8.056119918823242, "learning_rate": 2.192809361017739e-06, "loss": 2.6113, "step": 512900 }, { "epoch": 4.529839806425405, "grad_norm": 7.122768878936768, "learning_rate": 2.1887410497020366e-06, "loss": 3.0057, "step": 512950 }, { "epoch": 4.530281354315689, "grad_norm": 5.448456287384033, "learning_rate": 2.184676431400656e-06, "loss": 2.9536, "step": 513000 }, { "epoch": 4.530281354315689, "eval_asr_loss": 0.9924467975338294, "eval_loss": 2.6937034130096436, "eval_runtime": 20.8774, "eval_samples_per_second": 36.786, "eval_steps_per_second": 9.197, "eval_tts_loss": 6.333326844891253, "step": 513000 }, { "epoch": 4.530722902205973, "grad_norm": 10.152998924255371, "learning_rate": 2.180615506427558e-06, "loss": 2.4909, "step": 513050 }, { "epoch": 4.531164450096258, "grad_norm": 8.143499374389648, "learning_rate": 2.176558275096413e-06, "loss": 2.794, "step": 513100 }, { "epoch": 4.531605997986541, "grad_norm": 6.0360798835754395, "learning_rate": 2.17250473772061e-06, "loss": 2.6152, "step": 513150 }, { "epoch": 4.532047545876826, "grad_norm": 6.738029479980469, "learning_rate": 2.1684548946132588e-06, "loss": 2.661, "step": 513200 }, { "epoch": 4.53248909376711, "grad_norm": 3.4603793621063232, "learning_rate": 2.164489632848765e-06, "loss": 2.6597, "step": 513250 }, { "epoch": 4.532930641657394, "grad_norm": 1.3745976686477661, "learning_rate": 2.160447105315527e-06, "loss": 2.9139, "step": 513300 }, { "epoch": 4.533372189547678, "grad_norm": 3.520901918411255, "learning_rate": 2.1564082729820743e-06, "loss": 2.7622, "step": 513350 }, { "epoch": 4.533813737437963, "grad_norm": 5.678741931915283, "learning_rate": 2.152373136160385e-06, "loss": 2.8145, "step": 513400 }, { "epoch": 4.534255285328246, "grad_norm": 8.450532913208008, "learning_rate": 2.1483416951621314e-06, "loss": 2.3588, "step": 513450 }, { "epoch": 4.534696833218531, "grad_norm": 4.865325927734375, "learning_rate": 2.1443139502987154e-06, "loss": 3.0302, "step": 513500 }, { "epoch": 4.5351383811088155, "grad_norm": 5.921258449554443, "learning_rate": 2.1402899018812427e-06, "loss": 2.3695, "step": 513550 }, { "epoch": 4.535579928999099, "grad_norm": 3.3556995391845703, "learning_rate": 2.1362695502205323e-06, "loss": 2.7855, "step": 513600 }, { "epoch": 4.536021476889383, "grad_norm": 5.089808940887451, "learning_rate": 2.132252895627135e-06, "loss": 2.6053, "step": 513650 }, { "epoch": 4.536463024779668, "grad_norm": 8.426544189453125, "learning_rate": 2.128239938411286e-06, "loss": 2.6578, "step": 513700 }, { "epoch": 4.536904572669952, "grad_norm": 10.076850891113281, "learning_rate": 2.1242306788829713e-06, "loss": 2.8034, "step": 513750 }, { "epoch": 4.537346120560236, "grad_norm": 7.00206184387207, "learning_rate": 2.120225117351848e-06, "loss": 2.504, "step": 513800 }, { "epoch": 4.5377876684505205, "grad_norm": 5.61336088180542, "learning_rate": 2.1162232541273307e-06, "loss": 2.6761, "step": 513850 }, { "epoch": 4.538229216340804, "grad_norm": 8.050572395324707, "learning_rate": 2.1122250895185325e-06, "loss": 2.813, "step": 513900 }, { "epoch": 4.538670764231089, "grad_norm": 7.522817134857178, "learning_rate": 2.1083104768964913e-06, "loss": 2.5999, "step": 513950 }, { "epoch": 4.539112312121373, "grad_norm": 8.577835083007812, "learning_rate": 2.104319636457619e-06, "loss": 3.1756, "step": 514000 }, { "epoch": 4.539553860011657, "grad_norm": 3.310265064239502, "learning_rate": 2.1003324955539026e-06, "loss": 3.0292, "step": 514050 }, { "epoch": 4.539995407901941, "grad_norm": 3.6056973934173584, "learning_rate": 2.096349054493324e-06, "loss": 2.5767, "step": 514100 }, { "epoch": 4.5404369557922255, "grad_norm": 9.648415565490723, "learning_rate": 2.0923693135835752e-06, "loss": 2.7688, "step": 514150 }, { "epoch": 4.540878503682509, "grad_norm": 4.3466386795043945, "learning_rate": 2.0883932731320497e-06, "loss": 2.9204, "step": 514200 }, { "epoch": 4.541320051572794, "grad_norm": 2.6726808547973633, "learning_rate": 2.0844209334458632e-06, "loss": 2.8334, "step": 514250 }, { "epoch": 4.541761599463078, "grad_norm": 4.36235237121582, "learning_rate": 2.0804522948318526e-06, "loss": 2.7828, "step": 514300 }, { "epoch": 4.542203147353362, "grad_norm": 8.430249214172363, "learning_rate": 2.076487357596557e-06, "loss": 2.6825, "step": 514350 }, { "epoch": 4.542644695243646, "grad_norm": 6.913710594177246, "learning_rate": 2.0725261220462257e-06, "loss": 2.7857, "step": 514400 }, { "epoch": 4.5430862431339305, "grad_norm": 7.140283107757568, "learning_rate": 2.0685685884868465e-06, "loss": 2.9404, "step": 514450 }, { "epoch": 4.543527791024214, "grad_norm": 3.9226059913635254, "learning_rate": 2.064614757224087e-06, "loss": 2.6168, "step": 514500 }, { "epoch": 4.543969338914499, "grad_norm": 5.815939903259277, "learning_rate": 2.0606646285633633e-06, "loss": 2.4406, "step": 514550 }, { "epoch": 4.544410886804783, "grad_norm": 5.914403438568115, "learning_rate": 2.056718202809782e-06, "loss": 2.7312, "step": 514600 }, { "epoch": 4.544852434695067, "grad_norm": 6.705454349517822, "learning_rate": 2.0527754802681766e-06, "loss": 2.6507, "step": 514650 }, { "epoch": 4.545293982585351, "grad_norm": 5.089241027832031, "learning_rate": 2.0488364612430754e-06, "loss": 2.4576, "step": 514700 }, { "epoch": 4.5457355304756355, "grad_norm": 9.118281364440918, "learning_rate": 2.044901146038758e-06, "loss": 2.4399, "step": 514750 }, { "epoch": 4.546177078365919, "grad_norm": 9.265019416809082, "learning_rate": 2.0409695349591753e-06, "loss": 2.3188, "step": 514800 }, { "epoch": 4.546618626256204, "grad_norm": 9.447327613830566, "learning_rate": 2.037041628308006e-06, "loss": 2.5385, "step": 514850 }, { "epoch": 4.5470601741464876, "grad_norm": 4.020019054412842, "learning_rate": 2.033117426388659e-06, "loss": 2.6989, "step": 514900 }, { "epoch": 4.547501722036772, "grad_norm": 10.172654151916504, "learning_rate": 2.029196929504251e-06, "loss": 2.2803, "step": 514950 }, { "epoch": 4.547943269927057, "grad_norm": 5.079107284545898, "learning_rate": 2.025280137957597e-06, "loss": 2.8273, "step": 515000 }, { "epoch": 4.5483848178173405, "grad_norm": 2.1598403453826904, "learning_rate": 2.0213670520512317e-06, "loss": 2.1836, "step": 515050 }, { "epoch": 4.548826365707624, "grad_norm": 4.390941143035889, "learning_rate": 2.017457672087425e-06, "loss": 2.9442, "step": 515100 }, { "epoch": 4.549267913597909, "grad_norm": 7.867831707000732, "learning_rate": 2.0135519983681296e-06, "loss": 3.1055, "step": 515150 }, { "epoch": 4.549709461488193, "grad_norm": 6.218082427978516, "learning_rate": 2.009650031195026e-06, "loss": 2.8928, "step": 515200 }, { "epoch": 4.550151009378477, "grad_norm": 6.957845687866211, "learning_rate": 2.005751770869513e-06, "loss": 2.5035, "step": 515250 }, { "epoch": 4.550592557268761, "grad_norm": 11.35078239440918, "learning_rate": 2.0018572176926984e-06, "loss": 2.6056, "step": 515300 }, { "epoch": 4.5510341051590455, "grad_norm": 7.243478298187256, "learning_rate": 1.9979663719653985e-06, "loss": 2.5972, "step": 515350 }, { "epoch": 4.55147565304933, "grad_norm": 5.438354015350342, "learning_rate": 1.99407923398815e-06, "loss": 2.3922, "step": 515400 }, { "epoch": 4.551917200939614, "grad_norm": 3.4180731773376465, "learning_rate": 1.9901958040612066e-06, "loss": 2.4598, "step": 515450 }, { "epoch": 4.552358748829898, "grad_norm": 6.433895111083984, "learning_rate": 1.9863936405722927e-06, "loss": 2.4118, "step": 515500 }, { "epoch": 4.552800296720182, "grad_norm": 7.397208213806152, "learning_rate": 1.9825175534696096e-06, "loss": 2.5462, "step": 515550 }, { "epoch": 4.553241844610467, "grad_norm": 7.086256980895996, "learning_rate": 1.978645175310273e-06, "loss": 2.783, "step": 515600 }, { "epoch": 4.5536833925007505, "grad_norm": 10.710650444030762, "learning_rate": 1.9747765063933886e-06, "loss": 2.3873, "step": 515650 }, { "epoch": 4.554124940391035, "grad_norm": 3.5490763187408447, "learning_rate": 1.9709115470177785e-06, "loss": 2.8889, "step": 515700 }, { "epoch": 4.554566488281319, "grad_norm": 5.667109489440918, "learning_rate": 1.9670502974819704e-06, "loss": 2.8767, "step": 515750 }, { "epoch": 4.555008036171603, "grad_norm": 5.225754261016846, "learning_rate": 1.9631927580842314e-06, "loss": 2.4766, "step": 515800 }, { "epoch": 4.555449584061887, "grad_norm": 5.3687520027160645, "learning_rate": 1.9593389291225006e-06, "loss": 2.6463, "step": 515850 }, { "epoch": 4.555891131952172, "grad_norm": 5.570760250091553, "learning_rate": 1.9554888108944736e-06, "loss": 2.5193, "step": 515900 }, { "epoch": 4.5563326798424555, "grad_norm": 7.220667839050293, "learning_rate": 1.9516424036975234e-06, "loss": 2.678, "step": 515950 }, { "epoch": 4.55677422773274, "grad_norm": 4.501957416534424, "learning_rate": 1.9477997078287626e-06, "loss": 2.4998, "step": 516000 }, { "epoch": 4.55677422773274, "eval_asr_loss": 0.9904167404260422, "eval_loss": 2.6916162967681885, "eval_runtime": 20.3026, "eval_samples_per_second": 37.828, "eval_steps_per_second": 9.457, "eval_tts_loss": 6.327134200211632, "step": 516000 }, { "epoch": 4.557215775623024, "grad_norm": 7.960975170135498, "learning_rate": 1.9439607235849977e-06, "loss": 2.6997, "step": 516050 }, { "epoch": 4.557657323513308, "grad_norm": 3.973013401031494, "learning_rate": 1.94012545126277e-06, "loss": 2.1796, "step": 516100 }, { "epoch": 4.558098871403592, "grad_norm": 4.30145263671875, "learning_rate": 1.936293891158314e-06, "loss": 3.0306, "step": 516150 }, { "epoch": 4.558540419293877, "grad_norm": 5.299402713775635, "learning_rate": 1.9324660435675823e-06, "loss": 2.841, "step": 516200 }, { "epoch": 4.5589819671841605, "grad_norm": 6.993743896484375, "learning_rate": 1.928641908786255e-06, "loss": 2.8714, "step": 516250 }, { "epoch": 4.559423515074445, "grad_norm": 6.3781304359436035, "learning_rate": 1.9248214871097012e-06, "loss": 2.4582, "step": 516300 }, { "epoch": 4.559865062964729, "grad_norm": 5.199655055999756, "learning_rate": 1.92100477883303e-06, "loss": 2.4664, "step": 516350 }, { "epoch": 4.560306610855013, "grad_norm": 9.115787506103516, "learning_rate": 1.9171917842510325e-06, "loss": 2.4338, "step": 516400 }, { "epoch": 4.560748158745297, "grad_norm": 6.1802215576171875, "learning_rate": 1.91338250365824e-06, "loss": 2.5119, "step": 516450 }, { "epoch": 4.561189706635582, "grad_norm": 6.538055896759033, "learning_rate": 1.9095769373488945e-06, "loss": 2.8095, "step": 516500 }, { "epoch": 4.5616312545258655, "grad_norm": 6.10593843460083, "learning_rate": 1.9057750856169288e-06, "loss": 2.9396, "step": 516550 }, { "epoch": 4.56207280241615, "grad_norm": 7.659207820892334, "learning_rate": 1.901976948756018e-06, "loss": 2.2899, "step": 516600 }, { "epoch": 4.562514350306435, "grad_norm": 3.8192672729492188, "learning_rate": 1.8981825270595277e-06, "loss": 2.4577, "step": 516650 }, { "epoch": 4.562955898196718, "grad_norm": 4.579131126403809, "learning_rate": 1.8943918208205346e-06, "loss": 2.6465, "step": 516700 }, { "epoch": 4.563397446087002, "grad_norm": 7.133507251739502, "learning_rate": 1.8906048303318547e-06, "loss": 2.9181, "step": 516750 }, { "epoch": 4.563838993977287, "grad_norm": 5.83311128616333, "learning_rate": 1.886821555885998e-06, "loss": 2.2351, "step": 516800 }, { "epoch": 4.564280541867571, "grad_norm": 7.361472129821777, "learning_rate": 1.8830419977751868e-06, "loss": 2.8172, "step": 516850 }, { "epoch": 4.564722089757855, "grad_norm": 6.1309895515441895, "learning_rate": 1.8792661562913594e-06, "loss": 2.4527, "step": 516900 }, { "epoch": 4.56516363764814, "grad_norm": 3.056859016418457, "learning_rate": 1.8754940317261716e-06, "loss": 2.7395, "step": 516950 }, { "epoch": 4.565605185538423, "grad_norm": 5.037982940673828, "learning_rate": 1.8717256243709903e-06, "loss": 2.5688, "step": 517000 }, { "epoch": 4.566046733428708, "grad_norm": 4.6751556396484375, "learning_rate": 1.867960934516888e-06, "loss": 2.7967, "step": 517050 }, { "epoch": 4.566488281318992, "grad_norm": 8.003790855407715, "learning_rate": 1.8641999624546491e-06, "loss": 2.851, "step": 517100 }, { "epoch": 4.566929829209276, "grad_norm": 7.089686393737793, "learning_rate": 1.8604427084747911e-06, "loss": 3.0106, "step": 517150 }, { "epoch": 4.56737137709956, "grad_norm": 2.519575595855713, "learning_rate": 1.8566891728675206e-06, "loss": 2.5948, "step": 517200 }, { "epoch": 4.567812924989845, "grad_norm": 6.2862324714660645, "learning_rate": 1.8529393559227615e-06, "loss": 2.7289, "step": 517250 }, { "epoch": 4.568254472880128, "grad_norm": 5.318475723266602, "learning_rate": 1.8491932579301707e-06, "loss": 2.4542, "step": 517300 }, { "epoch": 4.568696020770413, "grad_norm": 5.523097038269043, "learning_rate": 1.8454508791790837e-06, "loss": 2.5715, "step": 517350 }, { "epoch": 4.569137568660697, "grad_norm": 2.5558221340179443, "learning_rate": 1.8417122199585913e-06, "loss": 2.4926, "step": 517400 }, { "epoch": 4.569579116550981, "grad_norm": 3.038457155227661, "learning_rate": 1.8379772805574457e-06, "loss": 2.4347, "step": 517450 }, { "epoch": 4.570020664441265, "grad_norm": 5.902400016784668, "learning_rate": 1.8342460612641666e-06, "loss": 2.6682, "step": 517500 }, { "epoch": 4.57046221233155, "grad_norm": 6.444743633270264, "learning_rate": 1.8305185623669341e-06, "loss": 2.8181, "step": 517550 }, { "epoch": 4.570903760221833, "grad_norm": 6.874411106109619, "learning_rate": 1.8267947841536793e-06, "loss": 2.9179, "step": 517600 }, { "epoch": 4.571345308112118, "grad_norm": 2.067779064178467, "learning_rate": 1.8230747269120385e-06, "loss": 2.4855, "step": 517650 }, { "epoch": 4.571786856002402, "grad_norm": 8.643033027648926, "learning_rate": 1.8193583909293376e-06, "loss": 2.7443, "step": 517700 }, { "epoch": 4.572228403892686, "grad_norm": 5.021475791931152, "learning_rate": 1.8156457764926416e-06, "loss": 2.5672, "step": 517750 }, { "epoch": 4.57266995178297, "grad_norm": 3.913339853286743, "learning_rate": 1.8119368838887208e-06, "loss": 2.5163, "step": 517800 }, { "epoch": 4.573111499673255, "grad_norm": 8.347229957580566, "learning_rate": 1.8082317134040572e-06, "loss": 2.863, "step": 517850 }, { "epoch": 4.573553047563538, "grad_norm": 5.5478315353393555, "learning_rate": 1.8045302653248331e-06, "loss": 2.4934, "step": 517900 }, { "epoch": 4.573994595453823, "grad_norm": 5.875454902648926, "learning_rate": 1.8009064579604996e-06, "loss": 2.7966, "step": 517950 }, { "epoch": 4.574436143344107, "grad_norm": 4.004820823669434, "learning_rate": 1.7972123810872544e-06, "loss": 2.6218, "step": 518000 }, { "epoch": 4.574877691234391, "grad_norm": 3.317932605743408, "learning_rate": 1.7935220274706177e-06, "loss": 2.5751, "step": 518050 }, { "epoch": 4.575319239124675, "grad_norm": 5.1487812995910645, "learning_rate": 1.7898353973956228e-06, "loss": 2.3968, "step": 518100 }, { "epoch": 4.57576078701496, "grad_norm": 7.157833576202393, "learning_rate": 1.7861524911470306e-06, "loss": 2.8582, "step": 518150 }, { "epoch": 4.576202334905243, "grad_norm": 6.462425708770752, "learning_rate": 1.7824733090093192e-06, "loss": 2.4421, "step": 518200 }, { "epoch": 4.576643882795528, "grad_norm": 6.270049095153809, "learning_rate": 1.7787978512666669e-06, "loss": 2.8396, "step": 518250 }, { "epoch": 4.577085430685813, "grad_norm": 5.301973342895508, "learning_rate": 1.7751261182029798e-06, "loss": 2.4367, "step": 518300 }, { "epoch": 4.577526978576096, "grad_norm": 9.99062728881836, "learning_rate": 1.7714581101018646e-06, "loss": 2.7629, "step": 518350 }, { "epoch": 4.57796852646638, "grad_norm": 4.281998157501221, "learning_rate": 1.7677938272466443e-06, "loss": 2.4826, "step": 518400 }, { "epoch": 4.578410074356665, "grad_norm": 6.5622382164001465, "learning_rate": 1.7641332699203484e-06, "loss": 2.4263, "step": 518450 }, { "epoch": 4.578851622246949, "grad_norm": 7.134660243988037, "learning_rate": 1.7604764384057394e-06, "loss": 2.5825, "step": 518500 }, { "epoch": 4.579293170137233, "grad_norm": 7.322083473205566, "learning_rate": 1.7568233329852634e-06, "loss": 2.3682, "step": 518550 }, { "epoch": 4.579734718027518, "grad_norm": 9.075828552246094, "learning_rate": 1.7531739539410952e-06, "loss": 2.5367, "step": 518600 }, { "epoch": 4.580176265917801, "grad_norm": 5.240750312805176, "learning_rate": 1.749528301555109e-06, "loss": 2.7888, "step": 518650 }, { "epoch": 4.580617813808086, "grad_norm": 3.883667230606079, "learning_rate": 1.745886376108924e-06, "loss": 2.5036, "step": 518700 }, { "epoch": 4.58105936169837, "grad_norm": 10.39873218536377, "learning_rate": 1.7422481778838206e-06, "loss": 2.5895, "step": 518750 }, { "epoch": 4.581500909588654, "grad_norm": 7.856481075286865, "learning_rate": 1.7386137071608355e-06, "loss": 2.5277, "step": 518800 }, { "epoch": 4.581942457478938, "grad_norm": 3.4035000801086426, "learning_rate": 1.7349829642206995e-06, "loss": 2.9622, "step": 518850 }, { "epoch": 4.582384005369223, "grad_norm": 8.74990463256836, "learning_rate": 1.7313559493438603e-06, "loss": 2.552, "step": 518900 }, { "epoch": 4.582825553259506, "grad_norm": 8.77329158782959, "learning_rate": 1.727732662810455e-06, "loss": 2.4469, "step": 518950 }, { "epoch": 4.583267101149791, "grad_norm": 5.556575298309326, "learning_rate": 1.7241131049003656e-06, "loss": 2.4054, "step": 519000 }, { "epoch": 4.583267101149791, "eval_asr_loss": 0.9889472471109223, "eval_loss": 2.6927802562713623, "eval_runtime": 20.6656, "eval_samples_per_second": 37.163, "eval_steps_per_second": 9.291, "eval_tts_loss": 6.338494685264795, "step": 519000 }, { "epoch": 4.583708649040075, "grad_norm": 10.385211944580078, "learning_rate": 1.7204972758931791e-06, "loss": 2.5293, "step": 519050 }, { "epoch": 4.584150196930359, "grad_norm": 6.973216533660889, "learning_rate": 1.7168851760681781e-06, "loss": 2.7374, "step": 519100 }, { "epoch": 4.584591744820643, "grad_norm": 6.697206020355225, "learning_rate": 1.7132768057043613e-06, "loss": 2.6181, "step": 519150 }, { "epoch": 4.585033292710928, "grad_norm": 8.10403823852539, "learning_rate": 1.7096721650804614e-06, "loss": 2.8092, "step": 519200 }, { "epoch": 4.585474840601211, "grad_norm": 5.19556999206543, "learning_rate": 1.7060712544748892e-06, "loss": 2.4834, "step": 519250 }, { "epoch": 4.585916388491496, "grad_norm": 4.820408821105957, "learning_rate": 1.7024740741657997e-06, "loss": 3.0239, "step": 519300 }, { "epoch": 4.58635793638178, "grad_norm": 9.1781587600708, "learning_rate": 1.6988806244310372e-06, "loss": 2.3367, "step": 519350 }, { "epoch": 4.586799484272064, "grad_norm": 1.4124820232391357, "learning_rate": 1.695290905548158e-06, "loss": 2.9063, "step": 519400 }, { "epoch": 4.587241032162348, "grad_norm": 10.394259452819824, "learning_rate": 1.6917049177944455e-06, "loss": 2.7788, "step": 519450 }, { "epoch": 4.587682580052633, "grad_norm": 8.989030838012695, "learning_rate": 1.6881226614468892e-06, "loss": 2.784, "step": 519500 }, { "epoch": 4.588124127942916, "grad_norm": 7.01063346862793, "learning_rate": 1.684544136782179e-06, "loss": 2.8623, "step": 519550 }, { "epoch": 4.588565675833201, "grad_norm": 2.396764039993286, "learning_rate": 1.6809693440767328e-06, "loss": 2.9013, "step": 519600 }, { "epoch": 4.589007223723485, "grad_norm": 9.676737785339355, "learning_rate": 1.6773982836066736e-06, "loss": 2.7549, "step": 519650 }, { "epoch": 4.589448771613769, "grad_norm": 5.993600845336914, "learning_rate": 1.6738309556478372e-06, "loss": 2.5006, "step": 519700 }, { "epoch": 4.589890319504054, "grad_norm": 3.7109086513519287, "learning_rate": 1.670267360475758e-06, "loss": 2.125, "step": 519750 }, { "epoch": 4.590331867394338, "grad_norm": 7.629401683807373, "learning_rate": 1.666707498365705e-06, "loss": 2.7285, "step": 519800 }, { "epoch": 4.590773415284621, "grad_norm": 8.427786827087402, "learning_rate": 1.6631513695926415e-06, "loss": 2.8987, "step": 519850 }, { "epoch": 4.591214963174906, "grad_norm": 4.149695873260498, "learning_rate": 1.6595989744312535e-06, "loss": 2.7874, "step": 519900 }, { "epoch": 4.591656511065191, "grad_norm": 3.2457025051116943, "learning_rate": 1.6560503131559214e-06, "loss": 2.6593, "step": 519950 }, { "epoch": 4.592098058955474, "grad_norm": 4.884464740753174, "learning_rate": 1.6525762479865314e-06, "loss": 2.688, "step": 520000 }, { "epoch": 4.592539606845758, "grad_norm": 4.665131092071533, "learning_rate": 1.6490349806139948e-06, "loss": 3.0244, "step": 520050 }, { "epoch": 4.592981154736043, "grad_norm": 3.7646539211273193, "learning_rate": 1.6454974479435059e-06, "loss": 2.8234, "step": 520100 }, { "epoch": 4.593422702626327, "grad_norm": 1.879184365272522, "learning_rate": 1.6419636502483015e-06, "loss": 2.7532, "step": 520150 }, { "epoch": 4.593864250516611, "grad_norm": 6.2534499168396, "learning_rate": 1.6384335878013413e-06, "loss": 2.7099, "step": 520200 }, { "epoch": 4.594305798406896, "grad_norm": 5.98675012588501, "learning_rate": 1.6349072608752958e-06, "loss": 2.4152, "step": 520250 }, { "epoch": 4.594747346297179, "grad_norm": 6.563912868499756, "learning_rate": 1.6313846697425361e-06, "loss": 2.4417, "step": 520300 }, { "epoch": 4.595188894187464, "grad_norm": 11.118790626525879, "learning_rate": 1.6278658146751668e-06, "loss": 2.6756, "step": 520350 }, { "epoch": 4.595630442077748, "grad_norm": 4.515968322753906, "learning_rate": 1.6243506959449705e-06, "loss": 2.7489, "step": 520400 }, { "epoch": 4.596071989968032, "grad_norm": 3.360774278640747, "learning_rate": 1.620839313823469e-06, "loss": 2.7129, "step": 520450 }, { "epoch": 4.596513537858316, "grad_norm": 8.746552467346191, "learning_rate": 1.6173316685818895e-06, "loss": 2.5754, "step": 520500 }, { "epoch": 4.596955085748601, "grad_norm": 2.6234219074249268, "learning_rate": 1.6138277604911656e-06, "loss": 3.0326, "step": 520550 }, { "epoch": 4.597396633638884, "grad_norm": 6.431388854980469, "learning_rate": 1.6103275898219416e-06, "loss": 2.4263, "step": 520600 }, { "epoch": 4.597838181529169, "grad_norm": 2.108835220336914, "learning_rate": 1.6068311568445793e-06, "loss": 2.8404, "step": 520650 }, { "epoch": 4.598279729419453, "grad_norm": 3.851773262023926, "learning_rate": 1.6033384618291458e-06, "loss": 2.3411, "step": 520700 }, { "epoch": 4.598721277309737, "grad_norm": 3.044116735458374, "learning_rate": 1.5998495050454366e-06, "loss": 3.02, "step": 520750 }, { "epoch": 4.599162825200021, "grad_norm": 8.975076675415039, "learning_rate": 1.5963642867629246e-06, "loss": 2.4893, "step": 520800 }, { "epoch": 4.599604373090306, "grad_norm": 10.085518836975098, "learning_rate": 1.5928828072508173e-06, "loss": 2.6465, "step": 520850 }, { "epoch": 4.600045920980589, "grad_norm": 3.5224173069000244, "learning_rate": 1.589405066778038e-06, "loss": 2.9292, "step": 520900 }, { "epoch": 4.600487468870874, "grad_norm": 8.748639106750488, "learning_rate": 1.5859310656131998e-06, "loss": 2.6555, "step": 520950 }, { "epoch": 4.600929016761158, "grad_norm": 8.242647171020508, "learning_rate": 1.5824608040246546e-06, "loss": 2.7088, "step": 521000 }, { "epoch": 4.601370564651442, "grad_norm": 4.384416580200195, "learning_rate": 1.5789942822804438e-06, "loss": 2.707, "step": 521050 }, { "epoch": 4.601812112541726, "grad_norm": 4.839293479919434, "learning_rate": 1.5755315006483196e-06, "loss": 2.7871, "step": 521100 }, { "epoch": 4.602253660432011, "grad_norm": 7.571292400360107, "learning_rate": 1.572072459395768e-06, "loss": 2.8529, "step": 521150 }, { "epoch": 4.602695208322294, "grad_norm": 2.1580591201782227, "learning_rate": 1.5686171587899534e-06, "loss": 2.4077, "step": 521200 }, { "epoch": 4.603136756212579, "grad_norm": 4.353419780731201, "learning_rate": 1.5651655990977899e-06, "loss": 2.4703, "step": 521250 }, { "epoch": 4.603578304102863, "grad_norm": 3.291038990020752, "learning_rate": 1.561717780585853e-06, "loss": 2.5091, "step": 521300 }, { "epoch": 4.604019851993147, "grad_norm": 3.880131483078003, "learning_rate": 1.5582737035204797e-06, "loss": 2.4158, "step": 521350 }, { "epoch": 4.604461399883432, "grad_norm": 7.796932697296143, "learning_rate": 1.5548333681676908e-06, "loss": 2.522, "step": 521400 }, { "epoch": 4.604902947773716, "grad_norm": 9.586150169372559, "learning_rate": 1.5513967747932178e-06, "loss": 2.6491, "step": 521450 }, { "epoch": 4.605344495663999, "grad_norm": 8.568214416503906, "learning_rate": 1.547963923662521e-06, "loss": 2.8272, "step": 521500 }, { "epoch": 4.605786043554284, "grad_norm": 3.541602849960327, "learning_rate": 1.544534815040738e-06, "loss": 2.4857, "step": 521550 }, { "epoch": 4.6062275914445685, "grad_norm": 2.9522860050201416, "learning_rate": 1.541109449192757e-06, "loss": 2.6241, "step": 521600 }, { "epoch": 4.606669139334852, "grad_norm": 7.124185562133789, "learning_rate": 1.5376878263831496e-06, "loss": 3.0413, "step": 521650 }, { "epoch": 4.607110687225136, "grad_norm": 11.406940460205078, "learning_rate": 1.5342699468762102e-06, "loss": 2.6004, "step": 521700 }, { "epoch": 4.607552235115421, "grad_norm": 5.433758735656738, "learning_rate": 1.5308558109359384e-06, "loss": 2.4706, "step": 521750 }, { "epoch": 4.607993783005705, "grad_norm": 5.5444488525390625, "learning_rate": 1.5274454188260456e-06, "loss": 3.0233, "step": 521800 }, { "epoch": 4.608435330895989, "grad_norm": 6.668963432312012, "learning_rate": 1.5240387708099602e-06, "loss": 2.9187, "step": 521850 }, { "epoch": 4.6088768787862735, "grad_norm": 7.685677528381348, "learning_rate": 1.5206358671508215e-06, "loss": 2.668, "step": 521900 }, { "epoch": 4.609318426676557, "grad_norm": 3.15995717048645, "learning_rate": 1.5172367081114691e-06, "loss": 2.9683, "step": 521950 }, { "epoch": 4.609759974566842, "grad_norm": 7.922619342803955, "learning_rate": 1.5138412939544545e-06, "loss": 2.4777, "step": 522000 }, { "epoch": 4.609759974566842, "eval_asr_loss": 0.9899299090322055, "eval_loss": 2.693039655685425, "eval_runtime": 20.3091, "eval_samples_per_second": 37.815, "eval_steps_per_second": 9.454, "eval_tts_loss": 6.3434297451486445, "step": 522000 }, { "epoch": 4.6102015224571256, "grad_norm": 10.329032897949219, "learning_rate": 1.5104496249420563e-06, "loss": 2.441, "step": 522050 }, { "epoch": 4.61064307034741, "grad_norm": 4.412397384643555, "learning_rate": 1.5070617013362376e-06, "loss": 2.4409, "step": 522100 }, { "epoch": 4.611084618237694, "grad_norm": 3.992169141769409, "learning_rate": 1.5036775233986944e-06, "loss": 2.7631, "step": 522150 }, { "epoch": 4.6115261661279785, "grad_norm": 3.002596378326416, "learning_rate": 1.5002970913908288e-06, "loss": 2.458, "step": 522200 }, { "epoch": 4.611967714018262, "grad_norm": 6.853955268859863, "learning_rate": 1.4969204055737429e-06, "loss": 2.6855, "step": 522250 }, { "epoch": 4.612409261908547, "grad_norm": 6.7346673011779785, "learning_rate": 1.4935474662082672e-06, "loss": 2.8686, "step": 522300 }, { "epoch": 4.6128508097988306, "grad_norm": 9.811986923217773, "learning_rate": 1.4901782735549264e-06, "loss": 2.4084, "step": 522350 }, { "epoch": 4.613292357689115, "grad_norm": 5.996295928955078, "learning_rate": 1.4868128278739624e-06, "loss": 2.8956, "step": 522400 }, { "epoch": 4.613733905579399, "grad_norm": 7.139163017272949, "learning_rate": 1.483451129425323e-06, "loss": 2.7143, "step": 522450 }, { "epoch": 4.6141754534696835, "grad_norm": 6.098421096801758, "learning_rate": 1.4800931784686778e-06, "loss": 1.9579, "step": 522500 }, { "epoch": 4.614617001359967, "grad_norm": 5.493442058563232, "learning_rate": 1.4767389752634031e-06, "loss": 2.5728, "step": 522550 }, { "epoch": 4.615058549250252, "grad_norm": 5.077577590942383, "learning_rate": 1.4733885200685692e-06, "loss": 2.5064, "step": 522600 }, { "epoch": 4.6155000971405356, "grad_norm": 3.510791778564453, "learning_rate": 1.4700418131429805e-06, "loss": 2.7052, "step": 522650 }, { "epoch": 4.61594164503082, "grad_norm": 3.575140953063965, "learning_rate": 1.4666988547451465e-06, "loss": 2.7063, "step": 522700 }, { "epoch": 4.616383192921104, "grad_norm": 5.62127161026001, "learning_rate": 1.463359645133272e-06, "loss": 2.4595, "step": 522750 }, { "epoch": 4.6168247408113885, "grad_norm": 6.926988124847412, "learning_rate": 1.460024184565284e-06, "loss": 2.8165, "step": 522800 }, { "epoch": 4.617266288701672, "grad_norm": 12.090827941894531, "learning_rate": 1.4566924732988208e-06, "loss": 2.5076, "step": 522850 }, { "epoch": 4.617707836591957, "grad_norm": 5.557086944580078, "learning_rate": 1.453364511591232e-06, "loss": 2.7657, "step": 522900 }, { "epoch": 4.6181493844822405, "grad_norm": 6.441036224365234, "learning_rate": 1.450040299699562e-06, "loss": 2.8516, "step": 522950 }, { "epoch": 4.618590932372525, "grad_norm": 4.489724159240723, "learning_rate": 1.4467198378805946e-06, "loss": 2.5296, "step": 523000 }, { "epoch": 4.61903248026281, "grad_norm": 8.066706657409668, "learning_rate": 1.443403126390802e-06, "loss": 2.7268, "step": 523050 }, { "epoch": 4.6194740281530935, "grad_norm": 2.0341105461120605, "learning_rate": 1.4400901654863741e-06, "loss": 2.7824, "step": 523100 }, { "epoch": 4.619915576043377, "grad_norm": 5.0324320793151855, "learning_rate": 1.4367809554231948e-06, "loss": 2.2586, "step": 523150 }, { "epoch": 4.620357123933662, "grad_norm": 8.055644989013672, "learning_rate": 1.4334754964568874e-06, "loss": 2.9596, "step": 523200 }, { "epoch": 4.620798671823946, "grad_norm": 6.5611162185668945, "learning_rate": 1.4301737888427646e-06, "loss": 2.7593, "step": 523250 }, { "epoch": 4.62124021971423, "grad_norm": 8.56855297088623, "learning_rate": 1.42687583283585e-06, "loss": 2.9486, "step": 523300 }, { "epoch": 4.621681767604515, "grad_norm": 8.07278823852539, "learning_rate": 1.423581628690901e-06, "loss": 2.9306, "step": 523350 }, { "epoch": 4.6221233154947985, "grad_norm": 3.334080219268799, "learning_rate": 1.4202911766623473e-06, "loss": 2.8682, "step": 523400 }, { "epoch": 4.622564863385083, "grad_norm": 9.734288215637207, "learning_rate": 1.4170044770043578e-06, "loss": 3.0958, "step": 523450 }, { "epoch": 4.623006411275367, "grad_norm": 4.586981773376465, "learning_rate": 1.4137215299708018e-06, "loss": 3.2388, "step": 523500 }, { "epoch": 4.623447959165651, "grad_norm": 3.2338218688964844, "learning_rate": 1.4104423358152652e-06, "loss": 2.5886, "step": 523550 }, { "epoch": 4.623889507055935, "grad_norm": 5.851126670837402, "learning_rate": 1.4071668947910176e-06, "loss": 2.2119, "step": 523600 }, { "epoch": 4.62433105494622, "grad_norm": 8.917752265930176, "learning_rate": 1.4038952071510792e-06, "loss": 2.7676, "step": 523650 }, { "epoch": 4.6247726028365035, "grad_norm": 6.090951919555664, "learning_rate": 1.400627273148153e-06, "loss": 2.6692, "step": 523700 }, { "epoch": 4.625214150726788, "grad_norm": 2.1365652084350586, "learning_rate": 1.3973630930346593e-06, "loss": 2.7879, "step": 523750 }, { "epoch": 4.625655698617072, "grad_norm": 6.582096576690674, "learning_rate": 1.3941026670627244e-06, "loss": 2.5499, "step": 523800 }, { "epoch": 4.626097246507356, "grad_norm": 4.096970081329346, "learning_rate": 1.3908459954841968e-06, "loss": 2.8105, "step": 523850 }, { "epoch": 4.62653879439764, "grad_norm": 2.612842082977295, "learning_rate": 1.3875930785506254e-06, "loss": 2.6783, "step": 523900 }, { "epoch": 4.626980342287925, "grad_norm": 4.275538444519043, "learning_rate": 1.3843439165132589e-06, "loss": 2.7417, "step": 523950 }, { "epoch": 4.6274218901782085, "grad_norm": 8.349864959716797, "learning_rate": 1.3810985096230855e-06, "loss": 2.6372, "step": 524000 }, { "epoch": 4.627863438068493, "grad_norm": 3.948894500732422, "learning_rate": 1.377856858130766e-06, "loss": 2.6198, "step": 524050 }, { "epoch": 4.628304985958777, "grad_norm": 3.0943756103515625, "learning_rate": 1.3746189622867055e-06, "loss": 2.8676, "step": 524100 }, { "epoch": 4.628746533849061, "grad_norm": 5.311342239379883, "learning_rate": 1.3713848223409986e-06, "loss": 2.7182, "step": 524150 }, { "epoch": 4.629188081739345, "grad_norm": 4.862245082855225, "learning_rate": 1.3681544385434563e-06, "loss": 2.4781, "step": 524200 }, { "epoch": 4.62962962962963, "grad_norm": 15.513254165649414, "learning_rate": 1.3649278111436016e-06, "loss": 2.5627, "step": 524250 }, { "epoch": 4.6300711775199135, "grad_norm": 4.078514575958252, "learning_rate": 1.3617049403906511e-06, "loss": 2.7193, "step": 524300 }, { "epoch": 4.630512725410198, "grad_norm": 8.354165077209473, "learning_rate": 1.358485826533562e-06, "loss": 2.2409, "step": 524350 }, { "epoch": 4.630954273300482, "grad_norm": 3.1672780513763428, "learning_rate": 1.3552704698209684e-06, "loss": 2.8854, "step": 524400 }, { "epoch": 4.631395821190766, "grad_norm": 4.6617960929870605, "learning_rate": 1.3520588705012383e-06, "loss": 2.6026, "step": 524450 }, { "epoch": 4.63183736908105, "grad_norm": 2.7655248641967773, "learning_rate": 1.34885102882244e-06, "loss": 2.8089, "step": 524500 }, { "epoch": 4.632278916971335, "grad_norm": 3.5579261779785156, "learning_rate": 1.3456469450323471e-06, "loss": 2.7076, "step": 524550 }, { "epoch": 4.6327204648616185, "grad_norm": 6.3691511154174805, "learning_rate": 1.3424466193784457e-06, "loss": 2.9324, "step": 524600 }, { "epoch": 4.633162012751903, "grad_norm": 3.45654559135437, "learning_rate": 1.3392500521079487e-06, "loss": 2.5094, "step": 524650 }, { "epoch": 4.633603560642188, "grad_norm": 5.280445098876953, "learning_rate": 1.3360572434677532e-06, "loss": 2.2185, "step": 524700 }, { "epoch": 4.634045108532471, "grad_norm": 4.100132465362549, "learning_rate": 1.332868193704473e-06, "loss": 2.9082, "step": 524750 }, { "epoch": 4.634486656422755, "grad_norm": 4.949473857879639, "learning_rate": 1.3296829030644387e-06, "loss": 2.5147, "step": 524800 }, { "epoch": 4.63492820431304, "grad_norm": 6.308763027191162, "learning_rate": 1.3265013717936924e-06, "loss": 2.9319, "step": 524850 }, { "epoch": 4.635369752203324, "grad_norm": 5.676083564758301, "learning_rate": 1.3233236001379712e-06, "loss": 2.584, "step": 524900 }, { "epoch": 4.635811300093608, "grad_norm": 9.777911186218262, "learning_rate": 1.3201495883427396e-06, "loss": 2.5763, "step": 524950 }, { "epoch": 4.636252847983893, "grad_norm": 5.7339677810668945, "learning_rate": 1.3170427048363298e-06, "loss": 2.5748, "step": 525000 }, { "epoch": 4.636252847983893, "eval_asr_loss": 0.9934904825579265, "eval_loss": 2.6923599243164062, "eval_runtime": 20.5361, "eval_samples_per_second": 37.398, "eval_steps_per_second": 9.349, "eval_tts_loss": 6.335331936818669, "step": 525000 }, { "epoch": 4.636694395874176, "grad_norm": 5.203128337860107, "learning_rate": 1.313876138287873e-06, "loss": 2.9088, "step": 525050 }, { "epoch": 4.637135943764461, "grad_norm": 5.82348108291626, "learning_rate": 1.3107133323296272e-06, "loss": 2.4468, "step": 525100 }, { "epoch": 4.637577491654745, "grad_norm": 4.556741714477539, "learning_rate": 1.3075542872059032e-06, "loss": 2.8768, "step": 525150 }, { "epoch": 4.638019039545029, "grad_norm": 8.124926567077637, "learning_rate": 1.3043990031607167e-06, "loss": 2.4653, "step": 525200 }, { "epoch": 4.638460587435313, "grad_norm": 1.7286078929901123, "learning_rate": 1.3012474804377784e-06, "loss": 2.4785, "step": 525250 }, { "epoch": 4.638902135325598, "grad_norm": 7.443638324737549, "learning_rate": 1.2980997192805156e-06, "loss": 3.0742, "step": 525300 }, { "epoch": 4.639343683215881, "grad_norm": 3.0619349479675293, "learning_rate": 1.2949557199320783e-06, "loss": 2.4682, "step": 525350 }, { "epoch": 4.639785231106166, "grad_norm": 4.108334064483643, "learning_rate": 1.2918154826353002e-06, "loss": 2.7711, "step": 525400 }, { "epoch": 4.64022677899645, "grad_norm": 5.524106025695801, "learning_rate": 1.2886790076327425e-06, "loss": 2.6733, "step": 525450 }, { "epoch": 4.640668326886734, "grad_norm": 6.2772216796875, "learning_rate": 1.2855462951666674e-06, "loss": 2.3411, "step": 525500 }, { "epoch": 4.641109874777018, "grad_norm": 6.587643623352051, "learning_rate": 1.2824173454790644e-06, "loss": 2.822, "step": 525550 }, { "epoch": 4.641551422667303, "grad_norm": 6.414623260498047, "learning_rate": 1.2792921588116015e-06, "loss": 2.7247, "step": 525600 }, { "epoch": 4.641992970557586, "grad_norm": 2.3746116161346436, "learning_rate": 1.276170735405674e-06, "loss": 2.8967, "step": 525650 }, { "epoch": 4.642434518447871, "grad_norm": 4.250400543212891, "learning_rate": 1.2730530755024005e-06, "loss": 2.6872, "step": 525700 }, { "epoch": 4.642876066338155, "grad_norm": 7.664060592651367, "learning_rate": 1.2699391793425884e-06, "loss": 2.8618, "step": 525750 }, { "epoch": 4.643317614228439, "grad_norm": 5.260467052459717, "learning_rate": 1.2668290471667455e-06, "loss": 2.9445, "step": 525800 }, { "epoch": 4.643759162118723, "grad_norm": 2.9600555896759033, "learning_rate": 1.263722679215118e-06, "loss": 2.6518, "step": 525850 }, { "epoch": 4.644200710009008, "grad_norm": 2.8890819549560547, "learning_rate": 1.260620075727642e-06, "loss": 2.5095, "step": 525900 }, { "epoch": 4.644642257899291, "grad_norm": 1.6355223655700684, "learning_rate": 1.2575212369439648e-06, "loss": 2.5609, "step": 525950 }, { "epoch": 4.645083805789576, "grad_norm": 7.98100471496582, "learning_rate": 1.254426163103445e-06, "loss": 2.7734, "step": 526000 }, { "epoch": 4.64552535367986, "grad_norm": 3.4574530124664307, "learning_rate": 1.251334854445163e-06, "loss": 2.2477, "step": 526050 }, { "epoch": 4.645966901570144, "grad_norm": 3.4896721839904785, "learning_rate": 1.2482473112078786e-06, "loss": 3.082, "step": 526100 }, { "epoch": 4.646408449460429, "grad_norm": 11.613346099853516, "learning_rate": 1.2451635336300948e-06, "loss": 2.3208, "step": 526150 }, { "epoch": 4.646849997350713, "grad_norm": 6.953162670135498, "learning_rate": 1.2420835219499993e-06, "loss": 2.7518, "step": 526200 }, { "epoch": 4.647291545240996, "grad_norm": 7.685738563537598, "learning_rate": 1.2390072764054906e-06, "loss": 2.4456, "step": 526250 }, { "epoch": 4.647733093131281, "grad_norm": 7.099536895751953, "learning_rate": 1.2359347972341839e-06, "loss": 2.6616, "step": 526300 }, { "epoch": 4.648174641021566, "grad_norm": 6.372971534729004, "learning_rate": 1.2328660846734118e-06, "loss": 2.4221, "step": 526350 }, { "epoch": 4.648616188911849, "grad_norm": 6.482003211975098, "learning_rate": 1.229801138960207e-06, "loss": 2.4448, "step": 526400 }, { "epoch": 4.649057736802133, "grad_norm": 4.25668478012085, "learning_rate": 1.2267399603312968e-06, "loss": 2.714, "step": 526450 }, { "epoch": 4.649499284692418, "grad_norm": 6.236372470855713, "learning_rate": 1.2236825490231474e-06, "loss": 2.6397, "step": 526500 }, { "epoch": 4.649940832582702, "grad_norm": 4.152035713195801, "learning_rate": 1.220628905271909e-06, "loss": 2.2576, "step": 526550 }, { "epoch": 4.650382380472986, "grad_norm": 3.836432456970215, "learning_rate": 1.2175790293134427e-06, "loss": 2.6846, "step": 526600 }, { "epoch": 4.650823928363271, "grad_norm": 4.8839874267578125, "learning_rate": 1.2145329213833377e-06, "loss": 2.6248, "step": 526650 }, { "epoch": 4.651265476253554, "grad_norm": 9.195107460021973, "learning_rate": 1.2114905817168787e-06, "loss": 3.0193, "step": 526700 }, { "epoch": 4.651707024143839, "grad_norm": 6.938319206237793, "learning_rate": 1.208452010549055e-06, "loss": 2.5806, "step": 526750 }, { "epoch": 4.652148572034123, "grad_norm": 5.620786190032959, "learning_rate": 1.2054172081145732e-06, "loss": 2.79, "step": 526800 }, { "epoch": 4.652590119924407, "grad_norm": 5.72819185256958, "learning_rate": 1.2023861746478516e-06, "loss": 2.8378, "step": 526850 }, { "epoch": 4.653031667814691, "grad_norm": 4.862391948699951, "learning_rate": 1.199358910382997e-06, "loss": 3.0061, "step": 526900 }, { "epoch": 4.653473215704976, "grad_norm": 4.787391662597656, "learning_rate": 1.1963354155538563e-06, "loss": 2.5976, "step": 526950 }, { "epoch": 4.653914763595259, "grad_norm": 2.156444549560547, "learning_rate": 1.193315690393959e-06, "loss": 2.7722, "step": 527000 }, { "epoch": 4.654356311485544, "grad_norm": 2.124791383743286, "learning_rate": 1.1902997351365574e-06, "loss": 2.6196, "step": 527050 }, { "epoch": 4.654797859375828, "grad_norm": 5.1097798347473145, "learning_rate": 1.1872875500146041e-06, "loss": 2.4839, "step": 527100 }, { "epoch": 4.655239407266112, "grad_norm": 7.77749490737915, "learning_rate": 1.1842791352607685e-06, "loss": 2.4575, "step": 527150 }, { "epoch": 4.655680955156396, "grad_norm": 5.116493225097656, "learning_rate": 1.181274491107426e-06, "loss": 2.7115, "step": 527200 }, { "epoch": 4.656122503046681, "grad_norm": 5.62087345123291, "learning_rate": 1.1782736177866572e-06, "loss": 2.319, "step": 527250 }, { "epoch": 4.656564050936964, "grad_norm": 7.094642162322998, "learning_rate": 1.1752765155302602e-06, "loss": 3.0588, "step": 527300 }, { "epoch": 4.657005598827249, "grad_norm": 4.462364196777344, "learning_rate": 1.1722831845697279e-06, "loss": 2.5526, "step": 527350 }, { "epoch": 4.657447146717533, "grad_norm": 6.68496561050415, "learning_rate": 1.1692936251362696e-06, "loss": 2.3294, "step": 527400 }, { "epoch": 4.657888694607817, "grad_norm": 7.568789482116699, "learning_rate": 1.1663078374608062e-06, "loss": 2.4572, "step": 527450 }, { "epoch": 4.658330242498101, "grad_norm": 4.843991756439209, "learning_rate": 1.163325821773964e-06, "loss": 2.6104, "step": 527500 }, { "epoch": 4.658771790388386, "grad_norm": 4.975164890289307, "learning_rate": 1.160407106206207e-06, "loss": 2.5816, "step": 527550 }, { "epoch": 4.659213338278669, "grad_norm": 2.954164743423462, "learning_rate": 1.1574325597360991e-06, "loss": 2.7942, "step": 527600 }, { "epoch": 4.659654886168954, "grad_norm": 7.992484092712402, "learning_rate": 1.1544617859401474e-06, "loss": 2.839, "step": 527650 }, { "epoch": 4.660096434059238, "grad_norm": 6.7885661125183105, "learning_rate": 1.151494785047824e-06, "loss": 2.6167, "step": 527700 }, { "epoch": 4.660537981949522, "grad_norm": 8.294820785522461, "learning_rate": 1.1485315572883126e-06, "loss": 2.7125, "step": 527750 }, { "epoch": 4.660979529839807, "grad_norm": 3.7640459537506104, "learning_rate": 1.1455721028904797e-06, "loss": 2.6517, "step": 527800 }, { "epoch": 4.661421077730091, "grad_norm": 7.556309223175049, "learning_rate": 1.1426164220829317e-06, "loss": 2.4628, "step": 527850 }, { "epoch": 4.661862625620374, "grad_norm": 5.357326030731201, "learning_rate": 1.1396645150939688e-06, "loss": 2.7431, "step": 527900 }, { "epoch": 4.662304173510659, "grad_norm": 4.215081214904785, "learning_rate": 1.1367163821515981e-06, "loss": 3.2296, "step": 527950 }, { "epoch": 4.662745721400944, "grad_norm": 4.108025074005127, "learning_rate": 1.133772023483537e-06, "loss": 2.7493, "step": 528000 }, { "epoch": 4.662745721400944, "eval_asr_loss": 0.9951180031158384, "eval_loss": 2.692197799682617, "eval_runtime": 20.4158, "eval_samples_per_second": 37.618, "eval_steps_per_second": 9.404, "eval_tts_loss": 6.339714787883079, "step": 528000 }, { "epoch": 4.663187269291227, "grad_norm": 5.739671230316162, "learning_rate": 1.1308314393172147e-06, "loss": 2.412, "step": 528050 }, { "epoch": 4.663628817181511, "grad_norm": 4.899464130401611, "learning_rate": 1.1278946298797667e-06, "loss": 2.7237, "step": 528100 }, { "epoch": 4.664070365071796, "grad_norm": 6.287220001220703, "learning_rate": 1.1249615953980275e-06, "loss": 2.1773, "step": 528150 }, { "epoch": 4.66451191296208, "grad_norm": 7.25219202041626, "learning_rate": 1.1220323360985608e-06, "loss": 2.7324, "step": 528200 }, { "epoch": 4.664953460852364, "grad_norm": 5.68751859664917, "learning_rate": 1.1191068522076299e-06, "loss": 2.173, "step": 528250 }, { "epoch": 4.665395008742649, "grad_norm": 7.922399997711182, "learning_rate": 1.1161851439511873e-06, "loss": 2.851, "step": 528300 }, { "epoch": 4.665836556632932, "grad_norm": 3.57344388961792, "learning_rate": 1.1132672115549247e-06, "loss": 2.8139, "step": 528350 }, { "epoch": 4.666278104523217, "grad_norm": 7.582023620605469, "learning_rate": 1.1103530552442232e-06, "loss": 3.1563, "step": 528400 }, { "epoch": 4.666719652413501, "grad_norm": 6.813824653625488, "learning_rate": 1.1074426752441802e-06, "loss": 2.7689, "step": 528450 }, { "epoch": 4.667161200303785, "grad_norm": 4.290740013122559, "learning_rate": 1.1045360717795827e-06, "loss": 2.6963, "step": 528500 }, { "epoch": 4.667602748194069, "grad_norm": 9.204596519470215, "learning_rate": 1.1016332450749567e-06, "loss": 2.2756, "step": 528550 }, { "epoch": 4.668044296084354, "grad_norm": 9.118756294250488, "learning_rate": 1.0987341953545227e-06, "loss": 2.799, "step": 528600 }, { "epoch": 4.668485843974637, "grad_norm": 10.803426742553711, "learning_rate": 1.095838922842196e-06, "loss": 2.7399, "step": 528650 }, { "epoch": 4.668927391864922, "grad_norm": 6.0694899559021, "learning_rate": 1.0929474277616147e-06, "loss": 2.6882, "step": 528700 }, { "epoch": 4.669368939755206, "grad_norm": 9.318355560302734, "learning_rate": 1.0900597103361276e-06, "loss": 2.8073, "step": 528750 }, { "epoch": 4.66981048764549, "grad_norm": 2.184455394744873, "learning_rate": 1.0871757707887898e-06, "loss": 2.7589, "step": 528800 }, { "epoch": 4.670252035535774, "grad_norm": 7.173123359680176, "learning_rate": 1.084295609342345e-06, "loss": 2.8162, "step": 528850 }, { "epoch": 4.670693583426059, "grad_norm": 3.052199125289917, "learning_rate": 1.0814192262192824e-06, "loss": 2.9844, "step": 528900 }, { "epoch": 4.671135131316342, "grad_norm": 1.3922247886657715, "learning_rate": 1.0785466216417572e-06, "loss": 2.5604, "step": 528950 }, { "epoch": 4.671576679206627, "grad_norm": 5.648565769195557, "learning_rate": 1.0756777958316643e-06, "loss": 2.4961, "step": 529000 }, { "epoch": 4.672018227096911, "grad_norm": 10.161782264709473, "learning_rate": 1.0728127490106043e-06, "loss": 2.5325, "step": 529050 }, { "epoch": 4.672459774987195, "grad_norm": 4.160246849060059, "learning_rate": 1.0699514813998612e-06, "loss": 2.5717, "step": 529100 }, { "epoch": 4.672901322877479, "grad_norm": 3.587515354156494, "learning_rate": 1.0670939932204526e-06, "loss": 2.6421, "step": 529150 }, { "epoch": 4.673342870767764, "grad_norm": 4.763440132141113, "learning_rate": 1.064240284693091e-06, "loss": 2.5322, "step": 529200 }, { "epoch": 4.673784418658047, "grad_norm": 4.268267631530762, "learning_rate": 1.061390356038211e-06, "loss": 2.2507, "step": 529250 }, { "epoch": 4.674225966548332, "grad_norm": 6.619657039642334, "learning_rate": 1.0585442074759367e-06, "loss": 2.5048, "step": 529300 }, { "epoch": 4.674667514438616, "grad_norm": 7.767238616943359, "learning_rate": 1.0557018392261086e-06, "loss": 2.5113, "step": 529350 }, { "epoch": 4.6751090623289, "grad_norm": 4.070125579833984, "learning_rate": 1.0528632515082792e-06, "loss": 2.9582, "step": 529400 }, { "epoch": 4.675550610219185, "grad_norm": 5.044243812561035, "learning_rate": 1.0500284445417063e-06, "loss": 2.7928, "step": 529450 }, { "epoch": 4.6759921581094686, "grad_norm": 6.76406192779541, "learning_rate": 1.047197418545348e-06, "loss": 3.0192, "step": 529500 }, { "epoch": 4.676433705999752, "grad_norm": 5.975485801696777, "learning_rate": 1.0443701737378854e-06, "loss": 2.5692, "step": 529550 }, { "epoch": 4.676875253890037, "grad_norm": 2.794682264328003, "learning_rate": 1.0415467103376931e-06, "loss": 2.7743, "step": 529600 }, { "epoch": 4.6773168017803215, "grad_norm": 5.094548225402832, "learning_rate": 1.0387270285628637e-06, "loss": 2.6842, "step": 529650 }, { "epoch": 4.677758349670605, "grad_norm": 4.482576370239258, "learning_rate": 1.0359111286311895e-06, "loss": 2.4404, "step": 529700 }, { "epoch": 4.67819989756089, "grad_norm": 5.490303993225098, "learning_rate": 1.0330990107601802e-06, "loss": 3.0127, "step": 529750 }, { "epoch": 4.6786414454511736, "grad_norm": 6.554479598999023, "learning_rate": 1.0303468048111775e-06, "loss": 2.4756, "step": 529800 }, { "epoch": 4.679082993341458, "grad_norm": 5.824338436126709, "learning_rate": 1.027542176060814e-06, "loss": 2.6704, "step": 529850 }, { "epoch": 4.679524541231742, "grad_norm": 6.113000392913818, "learning_rate": 1.0247413300175479e-06, "loss": 2.5003, "step": 529900 }, { "epoch": 4.6799660891220265, "grad_norm": 4.889937400817871, "learning_rate": 1.021944266897712e-06, "loss": 2.7469, "step": 529950 }, { "epoch": 4.68040763701231, "grad_norm": 3.1059553623199463, "learning_rate": 1.019150986917361e-06, "loss": 2.7248, "step": 530000 }, { "epoch": 4.680849184902595, "grad_norm": 2.6412198543548584, "learning_rate": 1.0163614902922448e-06, "loss": 2.7814, "step": 530050 }, { "epoch": 4.6812907327928786, "grad_norm": 6.851519584655762, "learning_rate": 1.0135757772378407e-06, "loss": 2.8208, "step": 530100 }, { "epoch": 4.681732280683163, "grad_norm": 10.32213020324707, "learning_rate": 1.0107938479693159e-06, "loss": 2.6991, "step": 530150 }, { "epoch": 4.682173828573447, "grad_norm": 10.201407432556152, "learning_rate": 1.0080157027015425e-06, "loss": 2.8437, "step": 530200 }, { "epoch": 4.6826153764637315, "grad_norm": 4.788957118988037, "learning_rate": 1.005241341649127e-06, "loss": 2.6125, "step": 530250 }, { "epoch": 4.683056924354015, "grad_norm": 5.348529815673828, "learning_rate": 1.0024707650263477e-06, "loss": 2.4723, "step": 530300 }, { "epoch": 4.6834984722443, "grad_norm": 2.1397664546966553, "learning_rate": 9.997039730472168e-07, "loss": 2.5746, "step": 530350 }, { "epoch": 4.6839400201345835, "grad_norm": 5.635504722595215, "learning_rate": 9.969409659254413e-07, "loss": 2.463, "step": 530400 }, { "epoch": 4.684381568024868, "grad_norm": 7.728329181671143, "learning_rate": 9.94181743874445e-07, "loss": 2.4553, "step": 530450 }, { "epoch": 4.684823115915152, "grad_norm": 8.640701293945312, "learning_rate": 9.914263071073514e-07, "loss": 2.6244, "step": 530500 }, { "epoch": 4.6852646638054365, "grad_norm": 4.625260829925537, "learning_rate": 9.886746558369908e-07, "loss": 2.4911, "step": 530550 }, { "epoch": 4.68570621169572, "grad_norm": 1.8567007780075073, "learning_rate": 9.85926790275915e-07, "loss": 2.9291, "step": 530600 }, { "epoch": 4.686147759586005, "grad_norm": 9.809989929199219, "learning_rate": 9.831827106363656e-07, "loss": 2.4939, "step": 530650 }, { "epoch": 4.6865893074762885, "grad_norm": 5.217682838439941, "learning_rate": 9.804424171303006e-07, "loss": 2.6417, "step": 530700 }, { "epoch": 4.687030855366573, "grad_norm": 7.647070407867432, "learning_rate": 9.777059099693842e-07, "loss": 2.5673, "step": 530750 }, { "epoch": 4.687472403256857, "grad_norm": 5.061426162719727, "learning_rate": 9.749731893649971e-07, "loss": 2.552, "step": 530800 }, { "epoch": 4.6879139511471415, "grad_norm": 5.280552864074707, "learning_rate": 9.722442555281986e-07, "loss": 2.0286, "step": 530850 }, { "epoch": 4.688355499037425, "grad_norm": 6.30281400680542, "learning_rate": 9.69519108669792e-07, "loss": 2.856, "step": 530900 }, { "epoch": 4.68879704692771, "grad_norm": 5.909141540527344, "learning_rate": 9.668521390778562e-07, "loss": 2.8099, "step": 530950 }, { "epoch": 4.6892385948179935, "grad_norm": 3.3052163124084473, "learning_rate": 9.641344910573779e-07, "loss": 2.7308, "step": 531000 }, { "epoch": 4.6892385948179935, "eval_asr_loss": 0.9920038724070128, "eval_loss": 2.6926395893096924, "eval_runtime": 20.2542, "eval_samples_per_second": 37.918, "eval_steps_per_second": 9.48, "eval_tts_loss": 6.3488737097288315, "step": 531000 }, { "epoch": 4.689680142708278, "grad_norm": 9.178930282592773, "learning_rate": 9.614206306417017e-07, "loss": 2.5969, "step": 531050 }, { "epoch": 4.690121690598563, "grad_norm": 9.441248893737793, "learning_rate": 9.587105580404433e-07, "loss": 2.7186, "step": 531100 }, { "epoch": 4.6905632384888465, "grad_norm": 4.621333599090576, "learning_rate": 9.560042734629348e-07, "loss": 2.2877, "step": 531150 }, { "epoch": 4.69100478637913, "grad_norm": 5.609273433685303, "learning_rate": 9.533017771182206e-07, "loss": 2.9163, "step": 531200 }, { "epoch": 4.691446334269415, "grad_norm": 9.062514305114746, "learning_rate": 9.506030692150336e-07, "loss": 2.6045, "step": 531250 }, { "epoch": 4.691887882159699, "grad_norm": 2.403630018234253, "learning_rate": 9.479081499618348e-07, "loss": 2.6402, "step": 531300 }, { "epoch": 4.692329430049983, "grad_norm": 3.8230478763580322, "learning_rate": 9.452170195667853e-07, "loss": 2.6134, "step": 531350 }, { "epoch": 4.692770977940268, "grad_norm": 2.209423065185547, "learning_rate": 9.425296782377524e-07, "loss": 2.6485, "step": 531400 }, { "epoch": 4.6932125258305515, "grad_norm": 8.260723114013672, "learning_rate": 9.398461261822977e-07, "loss": 3.0753, "step": 531450 }, { "epoch": 4.693654073720836, "grad_norm": 4.705535888671875, "learning_rate": 9.371663636077166e-07, "loss": 2.9427, "step": 531500 }, { "epoch": 4.69409562161112, "grad_norm": 4.96817684173584, "learning_rate": 9.344903907209935e-07, "loss": 2.4173, "step": 531550 }, { "epoch": 4.694537169501404, "grad_norm": 2.4651591777801514, "learning_rate": 9.318182077288185e-07, "loss": 2.4906, "step": 531600 }, { "epoch": 4.694978717391688, "grad_norm": 6.514536380767822, "learning_rate": 9.291498148376099e-07, "loss": 2.4325, "step": 531650 }, { "epoch": 4.695420265281973, "grad_norm": 5.124837875366211, "learning_rate": 9.264852122534584e-07, "loss": 2.8245, "step": 531700 }, { "epoch": 4.6958618131722565, "grad_norm": 3.1842398643493652, "learning_rate": 9.23824400182205e-07, "loss": 2.5064, "step": 531750 }, { "epoch": 4.696303361062541, "grad_norm": 2.364046335220337, "learning_rate": 9.211673788293574e-07, "loss": 2.8788, "step": 531800 }, { "epoch": 4.696744908952825, "grad_norm": 6.113978862762451, "learning_rate": 9.185141484001513e-07, "loss": 2.8297, "step": 531850 }, { "epoch": 4.697186456843109, "grad_norm": 2.8025150299072266, "learning_rate": 9.158647090995287e-07, "loss": 2.3959, "step": 531900 }, { "epoch": 4.697628004733393, "grad_norm": 2.7261035442352295, "learning_rate": 9.132190611321367e-07, "loss": 2.98, "step": 531950 }, { "epoch": 4.698069552623678, "grad_norm": 4.9428582191467285, "learning_rate": 9.105772047023342e-07, "loss": 2.7962, "step": 532000 }, { "epoch": 4.6985111005139615, "grad_norm": 5.3042073249816895, "learning_rate": 9.07939140014169e-07, "loss": 2.4668, "step": 532050 }, { "epoch": 4.698952648404246, "grad_norm": 2.2752575874328613, "learning_rate": 9.05304867271417e-07, "loss": 2.57, "step": 532100 }, { "epoch": 4.69939419629453, "grad_norm": 6.539193630218506, "learning_rate": 9.0267438667756e-07, "loss": 2.5131, "step": 532150 }, { "epoch": 4.699835744184814, "grad_norm": 3.5936226844787598, "learning_rate": 9.000476984357687e-07, "loss": 2.1579, "step": 532200 }, { "epoch": 4.700277292075098, "grad_norm": 9.501581192016602, "learning_rate": 8.974248027489362e-07, "loss": 2.5514, "step": 532250 }, { "epoch": 4.700718839965383, "grad_norm": 8.34288215637207, "learning_rate": 8.948056998196619e-07, "loss": 2.3794, "step": 532300 }, { "epoch": 4.7011603878556665, "grad_norm": 3.379974842071533, "learning_rate": 8.921903898502504e-07, "loss": 2.6277, "step": 532350 }, { "epoch": 4.701601935745951, "grad_norm": 4.140822887420654, "learning_rate": 8.895788730427013e-07, "loss": 2.7409, "step": 532400 }, { "epoch": 4.702043483636235, "grad_norm": 5.746693134307861, "learning_rate": 8.869711495987421e-07, "loss": 3.0168, "step": 532450 }, { "epoch": 4.702485031526519, "grad_norm": 3.9436426162719727, "learning_rate": 8.843672197198005e-07, "loss": 2.5402, "step": 532500 }, { "epoch": 4.702926579416804, "grad_norm": 4.14069128036499, "learning_rate": 8.817670836070047e-07, "loss": 2.5914, "step": 532550 }, { "epoch": 4.703368127307088, "grad_norm": 6.5966691970825195, "learning_rate": 8.791707414611827e-07, "loss": 3.1159, "step": 532600 }, { "epoch": 4.7038096751973715, "grad_norm": 5.333532810211182, "learning_rate": 8.765781934828965e-07, "loss": 2.9666, "step": 532650 }, { "epoch": 4.704251223087656, "grad_norm": 4.046328067779541, "learning_rate": 8.739894398723858e-07, "loss": 2.2441, "step": 532700 }, { "epoch": 4.704692770977941, "grad_norm": 6.412468910217285, "learning_rate": 8.71404480829613e-07, "loss": 2.8962, "step": 532750 }, { "epoch": 4.705134318868224, "grad_norm": 2.835538625717163, "learning_rate": 8.688233165542514e-07, "loss": 2.3649, "step": 532800 }, { "epoch": 4.705575866758508, "grad_norm": 3.38478946685791, "learning_rate": 8.662459472456586e-07, "loss": 2.7151, "step": 532850 }, { "epoch": 4.706017414648793, "grad_norm": 11.186118125915527, "learning_rate": 8.636723731029361e-07, "loss": 2.5326, "step": 532900 }, { "epoch": 4.706458962539077, "grad_norm": 3.7565524578094482, "learning_rate": 8.611539527045576e-07, "loss": 2.3382, "step": 532950 }, { "epoch": 4.706900510429361, "grad_norm": 8.13913631439209, "learning_rate": 8.585878935764013e-07, "loss": 2.6197, "step": 533000 }, { "epoch": 4.707342058319646, "grad_norm": 4.427996635437012, "learning_rate": 8.560256302056313e-07, "loss": 2.5232, "step": 533050 }, { "epoch": 4.707783606209929, "grad_norm": 6.601923942565918, "learning_rate": 8.534671627901503e-07, "loss": 2.8658, "step": 533100 }, { "epoch": 4.708225154100214, "grad_norm": 7.5806708335876465, "learning_rate": 8.509124915275779e-07, "loss": 2.288, "step": 533150 }, { "epoch": 4.708666701990498, "grad_norm": 6.434243679046631, "learning_rate": 8.483616166152508e-07, "loss": 3.1029, "step": 533200 }, { "epoch": 4.709108249880782, "grad_norm": 7.547145843505859, "learning_rate": 8.458145382501947e-07, "loss": 2.6231, "step": 533250 }, { "epoch": 4.709549797771066, "grad_norm": 7.3840765953063965, "learning_rate": 8.432712566291412e-07, "loss": 2.6276, "step": 533300 }, { "epoch": 4.709991345661351, "grad_norm": 8.779536247253418, "learning_rate": 8.407317719485496e-07, "loss": 2.5712, "step": 533350 }, { "epoch": 4.710432893551634, "grad_norm": 10.7799654006958, "learning_rate": 8.381960844045744e-07, "loss": 2.9658, "step": 533400 }, { "epoch": 4.710874441441919, "grad_norm": 6.519335746765137, "learning_rate": 8.356641941930699e-07, "loss": 2.6713, "step": 533450 }, { "epoch": 4.711315989332203, "grad_norm": 6.351565361022949, "learning_rate": 8.331361015096017e-07, "loss": 2.572, "step": 533500 }, { "epoch": 4.711757537222487, "grad_norm": 3.9666502475738525, "learning_rate": 8.306118065494528e-07, "loss": 2.4303, "step": 533550 }, { "epoch": 4.712199085112771, "grad_norm": 4.0779900550842285, "learning_rate": 8.280913095075949e-07, "loss": 3.0028, "step": 533600 }, { "epoch": 4.712640633003056, "grad_norm": 2.2062196731567383, "learning_rate": 8.255746105787165e-07, "loss": 2.1685, "step": 533650 }, { "epoch": 4.713082180893339, "grad_norm": 4.044681549072266, "learning_rate": 8.23061709957218e-07, "loss": 3.2147, "step": 533700 }, { "epoch": 4.713523728783624, "grad_norm": 3.728492498397827, "learning_rate": 8.205526078371884e-07, "loss": 2.6532, "step": 533750 }, { "epoch": 4.713965276673908, "grad_norm": 4.287655353546143, "learning_rate": 8.180473044124393e-07, "loss": 2.2005, "step": 533800 }, { "epoch": 4.714406824564192, "grad_norm": 3.852327823638916, "learning_rate": 8.155457998764882e-07, "loss": 2.9609, "step": 533850 }, { "epoch": 4.714848372454476, "grad_norm": 5.199780464172363, "learning_rate": 8.130480944225471e-07, "loss": 2.8996, "step": 533900 }, { "epoch": 4.715289920344761, "grad_norm": 5.78519868850708, "learning_rate": 8.105541882435507e-07, "loss": 3.1144, "step": 533950 }, { "epoch": 4.715731468235044, "grad_norm": 3.478811025619507, "learning_rate": 8.080640815321338e-07, "loss": 2.3573, "step": 534000 }, { "epoch": 4.715731468235044, "eval_asr_loss": 0.9933264287884407, "eval_loss": 2.689948081970215, "eval_runtime": 20.5682, "eval_samples_per_second": 37.339, "eval_steps_per_second": 9.335, "eval_tts_loss": 6.329814726916469, "step": 534000 }, { "epoch": 4.716173016125329, "grad_norm": 10.498743057250977, "learning_rate": 8.055777744806314e-07, "loss": 2.3759, "step": 534050 }, { "epoch": 4.716614564015613, "grad_norm": 4.714338779449463, "learning_rate": 8.030952672810787e-07, "loss": 2.2062, "step": 534100 }, { "epoch": 4.717056111905897, "grad_norm": 2.2085483074188232, "learning_rate": 8.006165601252447e-07, "loss": 2.3731, "step": 534150 }, { "epoch": 4.717497659796182, "grad_norm": 4.0862226486206055, "learning_rate": 7.981416532045871e-07, "loss": 2.5169, "step": 534200 }, { "epoch": 4.717939207686466, "grad_norm": 4.823185920715332, "learning_rate": 7.956705467102643e-07, "loss": 2.2552, "step": 534250 }, { "epoch": 4.718380755576749, "grad_norm": 9.027262687683105, "learning_rate": 7.932032408331458e-07, "loss": 2.5391, "step": 534300 }, { "epoch": 4.718822303467034, "grad_norm": 3.1507315635681152, "learning_rate": 7.907397357638235e-07, "loss": 2.5526, "step": 534350 }, { "epoch": 4.719263851357319, "grad_norm": 4.46153450012207, "learning_rate": 7.882800316925731e-07, "loss": 2.9124, "step": 534400 }, { "epoch": 4.719705399247602, "grad_norm": 5.835597991943359, "learning_rate": 7.85824128809387e-07, "loss": 2.804, "step": 534450 }, { "epoch": 4.720146947137887, "grad_norm": 8.090288162231445, "learning_rate": 7.833720273039579e-07, "loss": 2.733, "step": 534500 }, { "epoch": 4.720588495028171, "grad_norm": 1.7289963960647583, "learning_rate": 7.809237273657011e-07, "loss": 2.6118, "step": 534550 }, { "epoch": 4.721030042918455, "grad_norm": 7.957582950592041, "learning_rate": 7.784792291837151e-07, "loss": 2.5076, "step": 534600 }, { "epoch": 4.721471590808739, "grad_norm": 3.0594468116760254, "learning_rate": 7.760385329468211e-07, "loss": 2.6238, "step": 534650 }, { "epoch": 4.721913138699024, "grad_norm": 2.5477688312530518, "learning_rate": 7.736016388435519e-07, "loss": 2.6152, "step": 534700 }, { "epoch": 4.722354686589307, "grad_norm": 4.434602737426758, "learning_rate": 7.711685470621233e-07, "loss": 2.4254, "step": 534750 }, { "epoch": 4.722796234479592, "grad_norm": 5.653475284576416, "learning_rate": 7.68739257790474e-07, "loss": 2.588, "step": 534800 }, { "epoch": 4.723237782369876, "grad_norm": 3.9395835399627686, "learning_rate": 7.663137712162538e-07, "loss": 2.7454, "step": 534850 }, { "epoch": 4.72367933026016, "grad_norm": 6.562875747680664, "learning_rate": 7.638920875268019e-07, "loss": 2.3971, "step": 534900 }, { "epoch": 4.724120878150444, "grad_norm": 8.237524032592773, "learning_rate": 7.61474206909174e-07, "loss": 2.9031, "step": 534950 }, { "epoch": 4.724562426040729, "grad_norm": 4.400344371795654, "learning_rate": 7.590601295501376e-07, "loss": 2.7124, "step": 535000 }, { "epoch": 4.725003973931012, "grad_norm": 7.941009521484375, "learning_rate": 7.566498556361546e-07, "loss": 2.5848, "step": 535050 }, { "epoch": 4.725445521821297, "grad_norm": 8.893454551696777, "learning_rate": 7.542433853533926e-07, "loss": 2.6356, "step": 535100 }, { "epoch": 4.725887069711581, "grad_norm": 4.225853443145752, "learning_rate": 7.518407188877474e-07, "loss": 2.5025, "step": 535150 }, { "epoch": 4.726328617601865, "grad_norm": 5.018224239349365, "learning_rate": 7.494418564247929e-07, "loss": 2.7312, "step": 535200 }, { "epoch": 4.726770165492149, "grad_norm": 5.725839138031006, "learning_rate": 7.470467981498141e-07, "loss": 2.8465, "step": 535250 }, { "epoch": 4.727211713382434, "grad_norm": 7.417541027069092, "learning_rate": 7.446555442478242e-07, "loss": 2.325, "step": 535300 }, { "epoch": 4.727653261272717, "grad_norm": 13.82026481628418, "learning_rate": 7.422680949035199e-07, "loss": 2.4392, "step": 535350 }, { "epoch": 4.728094809163002, "grad_norm": 13.11792278289795, "learning_rate": 7.398844503013091e-07, "loss": 2.4213, "step": 535400 }, { "epoch": 4.728536357053286, "grad_norm": 7.271890640258789, "learning_rate": 7.375046106253114e-07, "loss": 2.7147, "step": 535450 }, { "epoch": 4.72897790494357, "grad_norm": 2.7633678913116455, "learning_rate": 7.351285760593518e-07, "loss": 2.4562, "step": 535500 }, { "epoch": 4.729419452833854, "grad_norm": 11.088157653808594, "learning_rate": 7.327563467869503e-07, "loss": 2.5554, "step": 535550 }, { "epoch": 4.729861000724139, "grad_norm": 4.956554412841797, "learning_rate": 7.303879229913491e-07, "loss": 2.5228, "step": 535600 }, { "epoch": 4.730302548614422, "grad_norm": 4.656731128692627, "learning_rate": 7.280233048554908e-07, "loss": 2.078, "step": 535650 }, { "epoch": 4.730744096504707, "grad_norm": 6.28003454208374, "learning_rate": 7.256624925620126e-07, "loss": 2.4106, "step": 535700 }, { "epoch": 4.731185644394991, "grad_norm": 8.988744735717773, "learning_rate": 7.233054862932742e-07, "loss": 2.4012, "step": 535750 }, { "epoch": 4.731627192285275, "grad_norm": 5.261467933654785, "learning_rate": 7.209522862313356e-07, "loss": 2.6877, "step": 535800 }, { "epoch": 4.73206874017556, "grad_norm": 4.406932830810547, "learning_rate": 7.186028925579569e-07, "loss": 2.8752, "step": 535850 }, { "epoch": 4.732510288065844, "grad_norm": 4.619504928588867, "learning_rate": 7.162573054546095e-07, "loss": 2.8089, "step": 535900 }, { "epoch": 4.732951835956127, "grad_norm": 7.654695510864258, "learning_rate": 7.139155251024765e-07, "loss": 2.5579, "step": 535950 }, { "epoch": 4.733393383846412, "grad_norm": 3.8095128536224365, "learning_rate": 7.115775516824353e-07, "loss": 2.7833, "step": 536000 }, { "epoch": 4.733834931736697, "grad_norm": 2.6375629901885986, "learning_rate": 7.092433853750691e-07, "loss": 2.6841, "step": 536050 }, { "epoch": 4.73427647962698, "grad_norm": 10.251656532287598, "learning_rate": 7.069130263606783e-07, "loss": 2.7948, "step": 536100 }, { "epoch": 4.734718027517265, "grad_norm": 2.367239236831665, "learning_rate": 7.045864748192632e-07, "loss": 2.5663, "step": 536150 }, { "epoch": 4.735159575407549, "grad_norm": 5.8623833656311035, "learning_rate": 7.022637309305302e-07, "loss": 2.5961, "step": 536200 }, { "epoch": 4.735601123297833, "grad_norm": 6.295261383056641, "learning_rate": 6.999447948738913e-07, "loss": 2.6938, "step": 536250 }, { "epoch": 4.736042671188117, "grad_norm": 2.8325388431549072, "learning_rate": 6.976296668284698e-07, "loss": 2.7524, "step": 536300 }, { "epoch": 4.736484219078402, "grad_norm": 5.876461982727051, "learning_rate": 6.953183469730784e-07, "loss": 2.7423, "step": 536350 }, { "epoch": 4.736925766968685, "grad_norm": 6.9397993087768555, "learning_rate": 6.930108354862575e-07, "loss": 2.5679, "step": 536400 }, { "epoch": 4.73736731485897, "grad_norm": 9.14268684387207, "learning_rate": 6.907071325462367e-07, "loss": 2.5072, "step": 536450 }, { "epoch": 4.737808862749254, "grad_norm": 8.091066360473633, "learning_rate": 6.88407238330957e-07, "loss": 2.7461, "step": 536500 }, { "epoch": 4.738250410639538, "grad_norm": 7.432076454162598, "learning_rate": 6.861111530180708e-07, "loss": 2.5279, "step": 536550 }, { "epoch": 4.738691958529822, "grad_norm": 6.3563361167907715, "learning_rate": 6.838188767849252e-07, "loss": 2.5485, "step": 536600 }, { "epoch": 4.739133506420107, "grad_norm": 4.265827655792236, "learning_rate": 6.815304098085896e-07, "loss": 2.6188, "step": 536650 }, { "epoch": 4.73957505431039, "grad_norm": 3.460545539855957, "learning_rate": 6.792914080830771e-07, "loss": 2.8702, "step": 536700 }, { "epoch": 4.740016602200675, "grad_norm": 3.828756332397461, "learning_rate": 6.770104839564196e-07, "loss": 2.5502, "step": 536750 }, { "epoch": 4.740458150090959, "grad_norm": 6.910536766052246, "learning_rate": 6.747333696124458e-07, "loss": 2.909, "step": 536800 }, { "epoch": 4.740899697981243, "grad_norm": 8.052297592163086, "learning_rate": 6.724600652270597e-07, "loss": 2.2971, "step": 536850 }, { "epoch": 4.741341245871527, "grad_norm": 7.565450668334961, "learning_rate": 6.701905709758483e-07, "loss": 2.6522, "step": 536900 }, { "epoch": 4.7417827937618116, "grad_norm": 7.5252485275268555, "learning_rate": 6.679248870341048e-07, "loss": 3.1854, "step": 536950 }, { "epoch": 4.742224341652095, "grad_norm": 8.078813552856445, "learning_rate": 6.656630135768394e-07, "loss": 2.6066, "step": 537000 }, { "epoch": 4.742224341652095, "eval_asr_loss": 0.9955233090861363, "eval_loss": 2.6903181076049805, "eval_runtime": 20.1915, "eval_samples_per_second": 38.036, "eval_steps_per_second": 9.509, "eval_tts_loss": 6.336943535171692, "step": 537000 }, { "epoch": 4.74266588954238, "grad_norm": 5.572568893432617, "learning_rate": 6.634049507787731e-07, "loss": 2.6156, "step": 537050 }, { "epoch": 4.743107437432664, "grad_norm": 13.382808685302734, "learning_rate": 6.611506988143057e-07, "loss": 2.9218, "step": 537100 }, { "epoch": 4.743548985322948, "grad_norm": 8.630922317504883, "learning_rate": 6.589002578575643e-07, "loss": 2.4918, "step": 537150 }, { "epoch": 4.743990533213232, "grad_norm": 5.238407611846924, "learning_rate": 6.56653628082382e-07, "loss": 3.1145, "step": 537200 }, { "epoch": 4.7444320811035166, "grad_norm": 4.799350738525391, "learning_rate": 6.54410809662287e-07, "loss": 2.1608, "step": 537250 }, { "epoch": 4.744873628993801, "grad_norm": 2.8691933155059814, "learning_rate": 6.521718027705126e-07, "loss": 3.0377, "step": 537300 }, { "epoch": 4.745315176884085, "grad_norm": 5.049219608306885, "learning_rate": 6.499366075800151e-07, "loss": 2.5425, "step": 537350 }, { "epoch": 4.745756724774369, "grad_norm": 2.7234268188476562, "learning_rate": 6.477052242634341e-07, "loss": 2.6404, "step": 537400 }, { "epoch": 4.746198272664653, "grad_norm": 7.4580793380737305, "learning_rate": 6.454776529931373e-07, "loss": 2.2125, "step": 537450 }, { "epoch": 4.746639820554938, "grad_norm": 2.201376438140869, "learning_rate": 6.432538939411703e-07, "loss": 2.2639, "step": 537500 }, { "epoch": 4.7470813684452215, "grad_norm": 10.04139232635498, "learning_rate": 6.410339472793125e-07, "loss": 2.5682, "step": 537550 }, { "epoch": 4.747522916335505, "grad_norm": 6.360823631286621, "learning_rate": 6.388178131790212e-07, "loss": 2.341, "step": 537600 }, { "epoch": 4.74796446422579, "grad_norm": 4.840248107910156, "learning_rate": 6.366054918114872e-07, "loss": 2.8545, "step": 537650 }, { "epoch": 4.7484060121160745, "grad_norm": 7.418286323547363, "learning_rate": 6.34396983347596e-07, "loss": 2.9922, "step": 537700 }, { "epoch": 4.748847560006358, "grad_norm": 2.3454058170318604, "learning_rate": 6.321922879579223e-07, "loss": 2.6059, "step": 537750 }, { "epoch": 4.749289107896643, "grad_norm": 3.843595027923584, "learning_rate": 6.299914058127687e-07, "loss": 2.8145, "step": 537800 }, { "epoch": 4.7497306557869265, "grad_norm": 3.8726601600646973, "learning_rate": 6.277943370821382e-07, "loss": 2.478, "step": 537850 }, { "epoch": 4.750172203677211, "grad_norm": 5.699206829071045, "learning_rate": 6.256010819357283e-07, "loss": 2.5429, "step": 537900 }, { "epoch": 4.750613751567495, "grad_norm": 5.968163013458252, "learning_rate": 6.234116405429535e-07, "loss": 3.1507, "step": 537950 }, { "epoch": 4.7510552994577795, "grad_norm": 7.885987758636475, "learning_rate": 6.212260130729231e-07, "loss": 3.0872, "step": 538000 }, { "epoch": 4.751496847348063, "grad_norm": 8.811665534973145, "learning_rate": 6.190441996944685e-07, "loss": 2.5539, "step": 538050 }, { "epoch": 4.751938395238348, "grad_norm": 6.724641799926758, "learning_rate": 6.168662005761106e-07, "loss": 2.7981, "step": 538100 }, { "epoch": 4.7523799431286315, "grad_norm": 5.789234161376953, "learning_rate": 6.146920158860758e-07, "loss": 2.5755, "step": 538150 }, { "epoch": 4.752821491018916, "grad_norm": 8.456429481506348, "learning_rate": 6.125216457923189e-07, "loss": 2.3727, "step": 538200 }, { "epoch": 4.7532630389092, "grad_norm": 6.606229782104492, "learning_rate": 6.103550904624666e-07, "loss": 2.6294, "step": 538250 }, { "epoch": 4.7537045867994845, "grad_norm": 5.044473648071289, "learning_rate": 6.081923500638687e-07, "loss": 2.478, "step": 538300 }, { "epoch": 4.754146134689768, "grad_norm": 2.6014976501464844, "learning_rate": 6.060334247635857e-07, "loss": 2.7182, "step": 538350 }, { "epoch": 4.754587682580053, "grad_norm": 6.011855125427246, "learning_rate": 6.038783147283677e-07, "loss": 2.294, "step": 538400 }, { "epoch": 4.7550292304703365, "grad_norm": 8.51747989654541, "learning_rate": 6.017270201246816e-07, "loss": 2.7123, "step": 538450 }, { "epoch": 4.755470778360621, "grad_norm": 4.390125274658203, "learning_rate": 5.995795411187e-07, "loss": 2.8566, "step": 538500 }, { "epoch": 4.755912326250905, "grad_norm": 8.869695663452148, "learning_rate": 5.974358778763012e-07, "loss": 2.5154, "step": 538550 }, { "epoch": 4.7563538741411895, "grad_norm": 12.78432559967041, "learning_rate": 5.952960305630528e-07, "loss": 2.321, "step": 538600 }, { "epoch": 4.756795422031473, "grad_norm": 5.198671340942383, "learning_rate": 5.931599993442504e-07, "loss": 2.5756, "step": 538650 }, { "epoch": 4.757236969921758, "grad_norm": 4.108588218688965, "learning_rate": 5.910277843848788e-07, "loss": 2.4902, "step": 538700 }, { "epoch": 4.7576785178120415, "grad_norm": 5.537045955657959, "learning_rate": 5.888993858496338e-07, "loss": 3.2318, "step": 538750 }, { "epoch": 4.758120065702326, "grad_norm": 3.376147508621216, "learning_rate": 5.867748039029175e-07, "loss": 2.667, "step": 538800 }, { "epoch": 4.75856161359261, "grad_norm": 4.002898216247559, "learning_rate": 5.846540387088373e-07, "loss": 2.6513, "step": 538850 }, { "epoch": 4.7590031614828945, "grad_norm": 6.023526668548584, "learning_rate": 5.825370904312011e-07, "loss": 2.4962, "step": 538900 }, { "epoch": 4.759444709373179, "grad_norm": 3.643200159072876, "learning_rate": 5.804239592335225e-07, "loss": 3.2213, "step": 538950 }, { "epoch": 4.759886257263463, "grad_norm": 7.642817497253418, "learning_rate": 5.783146452790378e-07, "loss": 2.8392, "step": 539000 }, { "epoch": 4.7603278051537465, "grad_norm": 4.273192405700684, "learning_rate": 5.762091487306553e-07, "loss": 2.56, "step": 539050 }, { "epoch": 4.760769353044031, "grad_norm": 9.156852722167969, "learning_rate": 5.741074697510173e-07, "loss": 2.9156, "step": 539100 }, { "epoch": 4.761210900934316, "grad_norm": 6.183984279632568, "learning_rate": 5.720096085024551e-07, "loss": 2.3846, "step": 539150 }, { "epoch": 4.7616524488245995, "grad_norm": 4.371644496917725, "learning_rate": 5.699155651470167e-07, "loss": 2.9466, "step": 539200 }, { "epoch": 4.762093996714883, "grad_norm": 6.006260395050049, "learning_rate": 5.67825339846445e-07, "loss": 2.5317, "step": 539250 }, { "epoch": 4.762535544605168, "grad_norm": 6.113739013671875, "learning_rate": 5.657389327621943e-07, "loss": 2.4995, "step": 539300 }, { "epoch": 4.762977092495452, "grad_norm": 8.993940353393555, "learning_rate": 5.636563440554188e-07, "loss": 2.999, "step": 539350 }, { "epoch": 4.763418640385736, "grad_norm": 2.891500473022461, "learning_rate": 5.615775738869844e-07, "loss": 2.4193, "step": 539400 }, { "epoch": 4.763860188276021, "grad_norm": 5.529122829437256, "learning_rate": 5.595026224174571e-07, "loss": 2.3624, "step": 539450 }, { "epoch": 4.7643017361663045, "grad_norm": 4.8957061767578125, "learning_rate": 5.574314898071087e-07, "loss": 2.404, "step": 539500 }, { "epoch": 4.764743284056589, "grad_norm": 4.789647102355957, "learning_rate": 5.553641762159167e-07, "loss": 2.3104, "step": 539550 }, { "epoch": 4.765184831946873, "grad_norm": 5.450451374053955, "learning_rate": 5.533006818035647e-07, "loss": 3.2157, "step": 539600 }, { "epoch": 4.765626379837157, "grad_norm": 2.934407949447632, "learning_rate": 5.51241006729436e-07, "loss": 2.4986, "step": 539650 }, { "epoch": 4.766067927727441, "grad_norm": 5.646018028259277, "learning_rate": 5.491851511526258e-07, "loss": 2.8165, "step": 539700 }, { "epoch": 4.766509475617726, "grad_norm": 2.6124725341796875, "learning_rate": 5.471331152319347e-07, "loss": 2.4706, "step": 539750 }, { "epoch": 4.7669510235080095, "grad_norm": 6.677434921264648, "learning_rate": 5.45084899125864e-07, "loss": 2.7278, "step": 539800 }, { "epoch": 4.767392571398294, "grad_norm": 5.1075663566589355, "learning_rate": 5.4304050299262e-07, "loss": 2.4128, "step": 539850 }, { "epoch": 4.767834119288578, "grad_norm": 6.3003315925598145, "learning_rate": 5.409999269901101e-07, "loss": 2.8148, "step": 539900 }, { "epoch": 4.768275667178862, "grad_norm": 8.61226749420166, "learning_rate": 5.38963171275958e-07, "loss": 2.6558, "step": 539950 }, { "epoch": 4.768717215069146, "grad_norm": 2.288468599319458, "learning_rate": 5.36930236007488e-07, "loss": 2.6599, "step": 540000 }, { "epoch": 4.768717215069146, "eval_asr_loss": 0.9938410737729183, "eval_loss": 2.6898534297943115, "eval_runtime": 20.1302, "eval_samples_per_second": 38.152, "eval_steps_per_second": 9.538, "eval_tts_loss": 6.342205160176654, "step": 540000 }, { "epoch": 4.769158762959431, "grad_norm": 3.6586270332336426, "learning_rate": 5.349011213417132e-07, "loss": 2.3977, "step": 540050 }, { "epoch": 4.7696003108497145, "grad_norm": 4.633085250854492, "learning_rate": 5.328758274353806e-07, "loss": 2.3749, "step": 540100 }, { "epoch": 4.770041858739999, "grad_norm": 4.568413257598877, "learning_rate": 5.308543544449207e-07, "loss": 2.9152, "step": 540150 }, { "epoch": 4.770483406630283, "grad_norm": 3.0345892906188965, "learning_rate": 5.288367025264807e-07, "loss": 2.3058, "step": 540200 }, { "epoch": 4.770924954520567, "grad_norm": 3.9454128742218018, "learning_rate": 5.268228718358914e-07, "loss": 2.1963, "step": 540250 }, { "epoch": 4.771366502410851, "grad_norm": 6.200924873352051, "learning_rate": 5.248530252643047e-07, "loss": 3.0344, "step": 540300 }, { "epoch": 4.771808050301136, "grad_norm": 4.556758880615234, "learning_rate": 5.228467610635058e-07, "loss": 3.2225, "step": 540350 }, { "epoch": 4.7722495981914195, "grad_norm": 4.796082496643066, "learning_rate": 5.208443185532486e-07, "loss": 2.66, "step": 540400 }, { "epoch": 4.772691146081704, "grad_norm": 4.9844651222229, "learning_rate": 5.188456978881928e-07, "loss": 2.6149, "step": 540450 }, { "epoch": 4.773132693971988, "grad_norm": 8.052321434020996, "learning_rate": 5.168508992227206e-07, "loss": 2.4128, "step": 540500 }, { "epoch": 4.773574241862272, "grad_norm": 3.703540325164795, "learning_rate": 5.148599227109141e-07, "loss": 2.2268, "step": 540550 }, { "epoch": 4.774015789752557, "grad_norm": 8.751526832580566, "learning_rate": 5.128727685065504e-07, "loss": 2.9916, "step": 540600 }, { "epoch": 4.774457337642841, "grad_norm": 4.46387243270874, "learning_rate": 5.10889436763129e-07, "loss": 2.7367, "step": 540650 }, { "epoch": 4.7748988855331245, "grad_norm": 13.526087760925293, "learning_rate": 5.089099276338383e-07, "loss": 2.935, "step": 540700 }, { "epoch": 4.775340433423409, "grad_norm": 4.595304012298584, "learning_rate": 5.06934241271595e-07, "loss": 2.5434, "step": 540750 }, { "epoch": 4.775781981313694, "grad_norm": 6.306647777557373, "learning_rate": 5.049623778289825e-07, "loss": 2.8106, "step": 540800 }, { "epoch": 4.776223529203977, "grad_norm": 3.925901174545288, "learning_rate": 5.02994337458329e-07, "loss": 2.9144, "step": 540850 }, { "epoch": 4.776665077094262, "grad_norm": 4.843810558319092, "learning_rate": 5.010301203116352e-07, "loss": 2.7134, "step": 540900 }, { "epoch": 4.777106624984546, "grad_norm": 3.5427350997924805, "learning_rate": 4.990697265406241e-07, "loss": 2.3061, "step": 540950 }, { "epoch": 4.77754817287483, "grad_norm": 4.942563533782959, "learning_rate": 4.971131562967246e-07, "loss": 2.8992, "step": 541000 }, { "epoch": 4.777989720765114, "grad_norm": 5.115659713745117, "learning_rate": 4.951604097310603e-07, "loss": 2.661, "step": 541050 }, { "epoch": 4.778431268655399, "grad_norm": 4.3744611740112305, "learning_rate": 4.93211486994466e-07, "loss": 2.504, "step": 541100 }, { "epoch": 4.778872816545682, "grad_norm": 10.361797332763672, "learning_rate": 4.912663882374768e-07, "loss": 2.5877, "step": 541150 }, { "epoch": 4.779314364435967, "grad_norm": 4.0390849113464355, "learning_rate": 4.893251136103394e-07, "loss": 2.4316, "step": 541200 }, { "epoch": 4.779755912326251, "grad_norm": 4.748064041137695, "learning_rate": 4.873876632630004e-07, "loss": 2.7232, "step": 541250 }, { "epoch": 4.780197460216535, "grad_norm": 8.581171035766602, "learning_rate": 4.854540373451066e-07, "loss": 2.7112, "step": 541300 }, { "epoch": 4.780639008106819, "grad_norm": 2.4960885047912598, "learning_rate": 4.835242360060166e-07, "loss": 2.6853, "step": 541350 }, { "epoch": 4.781080555997104, "grad_norm": 2.9482696056365967, "learning_rate": 4.815982593947998e-07, "loss": 2.6687, "step": 541400 }, { "epoch": 4.781522103887387, "grad_norm": 2.893381118774414, "learning_rate": 4.796761076602041e-07, "loss": 3.0263, "step": 541450 }, { "epoch": 4.781963651777672, "grad_norm": 7.780259132385254, "learning_rate": 4.777577809507106e-07, "loss": 2.5449, "step": 541500 }, { "epoch": 4.782405199667956, "grad_norm": 6.200257301330566, "learning_rate": 4.758432794145007e-07, "loss": 2.8338, "step": 541550 }, { "epoch": 4.78284674755824, "grad_norm": 7.865572452545166, "learning_rate": 4.73932603199434e-07, "loss": 2.7073, "step": 541600 }, { "epoch": 4.783288295448524, "grad_norm": 5.819964408874512, "learning_rate": 4.7202575245310907e-07, "loss": 2.9552, "step": 541650 }, { "epoch": 4.783729843338809, "grad_norm": 3.859121799468994, "learning_rate": 4.701227273228137e-07, "loss": 2.2617, "step": 541700 }, { "epoch": 4.784171391229092, "grad_norm": 7.668726444244385, "learning_rate": 4.6822352795553027e-07, "loss": 2.7636, "step": 541750 }, { "epoch": 4.784612939119377, "grad_norm": 5.787812232971191, "learning_rate": 4.663281544979636e-07, "loss": 2.6829, "step": 541800 }, { "epoch": 4.785054487009661, "grad_norm": 2.537907123565674, "learning_rate": 4.6443660709651895e-07, "loss": 2.1381, "step": 541850 }, { "epoch": 4.785496034899945, "grad_norm": 2.878283977508545, "learning_rate": 4.625488858972904e-07, "loss": 2.8416, "step": 541900 }, { "epoch": 4.785937582790229, "grad_norm": 5.20255184173584, "learning_rate": 4.606649910461003e-07, "loss": 2.4053, "step": 541950 }, { "epoch": 4.786379130680514, "grad_norm": 6.430854797363281, "learning_rate": 4.587849226884544e-07, "loss": 3.0587, "step": 542000 }, { "epoch": 4.786820678570797, "grad_norm": 10.558135986328125, "learning_rate": 4.569086809695755e-07, "loss": 2.4892, "step": 542050 }, { "epoch": 4.787262226461082, "grad_norm": 10.35540771484375, "learning_rate": 4.5503626603439206e-07, "loss": 3.109, "step": 542100 }, { "epoch": 4.787703774351366, "grad_norm": 4.58510684967041, "learning_rate": 4.531676780275274e-07, "loss": 2.5704, "step": 542150 }, { "epoch": 4.78814532224165, "grad_norm": 3.950916290283203, "learning_rate": 4.5130291709331584e-07, "loss": 2.6781, "step": 542200 }, { "epoch": 4.788586870131935, "grad_norm": 4.705416202545166, "learning_rate": 4.494419833757868e-07, "loss": 2.6373, "step": 542250 }, { "epoch": 4.789028418022219, "grad_norm": 4.239086627960205, "learning_rate": 4.475848770186919e-07, "loss": 2.6757, "step": 542300 }, { "epoch": 4.789469965912502, "grad_norm": 3.0436782836914062, "learning_rate": 4.4573159816547194e-07, "loss": 2.6425, "step": 542350 }, { "epoch": 4.789911513802787, "grad_norm": 3.2379112243652344, "learning_rate": 4.438821469592791e-07, "loss": 2.498, "step": 542400 }, { "epoch": 4.790353061693072, "grad_norm": 9.830525398254395, "learning_rate": 4.420365235429713e-07, "loss": 2.8218, "step": 542450 }, { "epoch": 4.790794609583355, "grad_norm": 9.576047897338867, "learning_rate": 4.4019472805909566e-07, "loss": 2.7114, "step": 542500 }, { "epoch": 4.79123615747364, "grad_norm": 5.351738929748535, "learning_rate": 4.3835676064992723e-07, "loss": 2.6406, "step": 542550 }, { "epoch": 4.791677705363924, "grad_norm": 6.6185479164123535, "learning_rate": 4.365226214574247e-07, "loss": 3.0634, "step": 542600 }, { "epoch": 4.792119253254208, "grad_norm": 3.9795167446136475, "learning_rate": 4.346923106232637e-07, "loss": 2.1305, "step": 542650 }, { "epoch": 4.792560801144492, "grad_norm": 7.7232794761657715, "learning_rate": 4.3286582828882536e-07, "loss": 2.3802, "step": 542700 }, { "epoch": 4.793002349034777, "grad_norm": 8.348795890808105, "learning_rate": 4.310431745951804e-07, "loss": 2.5629, "step": 542750 }, { "epoch": 4.79344389692506, "grad_norm": 4.097153186798096, "learning_rate": 4.292243496831161e-07, "loss": 2.4196, "step": 542800 }, { "epoch": 4.793885444815345, "grad_norm": 6.159137725830078, "learning_rate": 4.2740935369313116e-07, "loss": 3.0086, "step": 542850 }, { "epoch": 4.794326992705629, "grad_norm": 7.040095806121826, "learning_rate": 4.255981867654024e-07, "loss": 2.9235, "step": 542900 }, { "epoch": 4.794768540595913, "grad_norm": 7.880752086639404, "learning_rate": 4.2379084903983455e-07, "loss": 3.0011, "step": 542950 }, { "epoch": 4.795210088486197, "grad_norm": 4.310023784637451, "learning_rate": 4.2198734065603265e-07, "loss": 2.6078, "step": 543000 }, { "epoch": 4.795210088486197, "eval_asr_loss": 0.9925797551602604, "eval_loss": 2.6894454956054688, "eval_runtime": 20.5581, "eval_samples_per_second": 37.358, "eval_steps_per_second": 9.339, "eval_tts_loss": 6.336504094584744, "step": 543000 }, { "epoch": 4.795651636376482, "grad_norm": 5.562412261962891, "learning_rate": 4.201876617533018e-07, "loss": 2.8165, "step": 543050 }, { "epoch": 4.796093184266765, "grad_norm": 7.22629451751709, "learning_rate": 4.183918124706421e-07, "loss": 2.8068, "step": 543100 }, { "epoch": 4.79653473215705, "grad_norm": 7.104128360748291, "learning_rate": 4.165997929467813e-07, "loss": 2.7495, "step": 543150 }, { "epoch": 4.796976280047334, "grad_norm": 5.946671485900879, "learning_rate": 4.14811603320131e-07, "loss": 2.7639, "step": 543200 }, { "epoch": 4.797417827937618, "grad_norm": 5.609646797180176, "learning_rate": 4.130272437288141e-07, "loss": 2.4975, "step": 543250 }, { "epoch": 4.797859375827902, "grad_norm": 6.057517051696777, "learning_rate": 4.112467143106591e-07, "loss": 2.8016, "step": 543300 }, { "epoch": 4.798300923718187, "grad_norm": 5.4055633544921875, "learning_rate": 4.094700152031949e-07, "loss": 2.7727, "step": 543350 }, { "epoch": 4.79874247160847, "grad_norm": 2.3013219833374023, "learning_rate": 4.076971465436508e-07, "loss": 2.5364, "step": 543400 }, { "epoch": 4.799184019498755, "grad_norm": 6.687713146209717, "learning_rate": 4.0592810846897255e-07, "loss": 2.5004, "step": 543450 }, { "epoch": 4.799625567389039, "grad_norm": 5.809507369995117, "learning_rate": 4.041629011158066e-07, "loss": 2.4866, "step": 543500 }, { "epoch": 4.800067115279323, "grad_norm": 7.390219211578369, "learning_rate": 4.0240152462049375e-07, "loss": 2.4275, "step": 543550 }, { "epoch": 4.800508663169607, "grad_norm": 8.348456382751465, "learning_rate": 4.006439791190919e-07, "loss": 2.5833, "step": 543600 }, { "epoch": 4.800950211059892, "grad_norm": 7.929141521453857, "learning_rate": 3.98890264747348e-07, "loss": 2.9634, "step": 543650 }, { "epoch": 4.801391758950176, "grad_norm": 7.332266807556152, "learning_rate": 3.9714038164073154e-07, "loss": 2.8497, "step": 543700 }, { "epoch": 4.80183330684046, "grad_norm": 8.171599388122559, "learning_rate": 3.953943299343954e-07, "loss": 2.3312, "step": 543750 }, { "epoch": 4.802274854730744, "grad_norm": 5.261852741241455, "learning_rate": 3.936521097632095e-07, "loss": 2.7815, "step": 543800 }, { "epoch": 4.802716402621028, "grad_norm": 9.6071138381958, "learning_rate": 3.9191372126175517e-07, "loss": 2.2555, "step": 543850 }, { "epoch": 4.803157950511313, "grad_norm": 9.752958297729492, "learning_rate": 3.9017916456429716e-07, "loss": 3.1432, "step": 543900 }, { "epoch": 4.803599498401597, "grad_norm": 5.643681049346924, "learning_rate": 3.884484398048227e-07, "loss": 2.5738, "step": 543950 }, { "epoch": 4.80404104629188, "grad_norm": 11.374584197998047, "learning_rate": 3.867215471170138e-07, "loss": 2.8716, "step": 544000 }, { "epoch": 4.804482594182165, "grad_norm": 4.400134086608887, "learning_rate": 3.8499848663425264e-07, "loss": 2.8037, "step": 544050 }, { "epoch": 4.8049241420724496, "grad_norm": 3.9704136848449707, "learning_rate": 3.832792584896383e-07, "loss": 2.6717, "step": 544100 }, { "epoch": 4.805365689962733, "grad_norm": 2.2237508296966553, "learning_rate": 3.815638628159646e-07, "loss": 2.551, "step": 544150 }, { "epoch": 4.805807237853018, "grad_norm": 7.576536178588867, "learning_rate": 3.7985229974572546e-07, "loss": 2.5952, "step": 544200 }, { "epoch": 4.806248785743302, "grad_norm": 6.424444675445557, "learning_rate": 3.7814456941113187e-07, "loss": 2.8409, "step": 544250 }, { "epoch": 4.806690333633586, "grad_norm": 6.384031772613525, "learning_rate": 3.7644067194408937e-07, "loss": 2.833, "step": 544300 }, { "epoch": 4.80713188152387, "grad_norm": 8.112434387207031, "learning_rate": 3.7474060747621496e-07, "loss": 2.6774, "step": 544350 }, { "epoch": 4.8075734294141546, "grad_norm": 2.384006977081299, "learning_rate": 3.7304437613881473e-07, "loss": 2.8777, "step": 544400 }, { "epoch": 4.808014977304438, "grad_norm": 3.5565364360809326, "learning_rate": 3.7135197806291155e-07, "loss": 2.4649, "step": 544450 }, { "epoch": 4.808456525194723, "grad_norm": 10.75607967376709, "learning_rate": 3.696634133792287e-07, "loss": 2.3645, "step": 544500 }, { "epoch": 4.808898073085007, "grad_norm": 4.772146701812744, "learning_rate": 3.679786822181952e-07, "loss": 2.9639, "step": 544550 }, { "epoch": 4.809339620975291, "grad_norm": 10.517046928405762, "learning_rate": 3.6629778470994024e-07, "loss": 2.8086, "step": 544600 }, { "epoch": 4.809781168865575, "grad_norm": 5.497314453125, "learning_rate": 3.6462072098430444e-07, "loss": 2.3388, "step": 544650 }, { "epoch": 4.8102227167558596, "grad_norm": 5.854978084564209, "learning_rate": 3.629474911708175e-07, "loss": 2.7558, "step": 544700 }, { "epoch": 4.810664264646143, "grad_norm": 9.873150825500488, "learning_rate": 3.6127809539872603e-07, "loss": 2.4944, "step": 544750 }, { "epoch": 4.811105812536428, "grad_norm": 3.4216089248657227, "learning_rate": 3.5961253379698246e-07, "loss": 2.3495, "step": 544800 }, { "epoch": 4.811547360426712, "grad_norm": 6.23988676071167, "learning_rate": 3.579508064942283e-07, "loss": 2.6149, "step": 544850 }, { "epoch": 4.811988908316996, "grad_norm": 8.79282283782959, "learning_rate": 3.56292913618822e-07, "loss": 2.5322, "step": 544900 }, { "epoch": 4.81243045620728, "grad_norm": 2.820842742919922, "learning_rate": 3.5463885529882225e-07, "loss": 2.9399, "step": 544950 }, { "epoch": 4.8128720040975645, "grad_norm": 8.314831733703613, "learning_rate": 3.5298863166199346e-07, "loss": 2.9058, "step": 545000 }, { "epoch": 4.813313551987848, "grad_norm": 5.549332618713379, "learning_rate": 3.513422428357949e-07, "loss": 2.5466, "step": 545050 }, { "epoch": 4.813755099878133, "grad_norm": 3.225752353668213, "learning_rate": 3.496996889474025e-07, "loss": 2.6728, "step": 545100 }, { "epoch": 4.814196647768417, "grad_norm": 6.776427745819092, "learning_rate": 3.48060970123687e-07, "loss": 2.3791, "step": 545150 }, { "epoch": 4.814638195658701, "grad_norm": 8.308965682983398, "learning_rate": 3.46426086491225e-07, "loss": 2.3775, "step": 545200 }, { "epoch": 4.815079743548985, "grad_norm": 5.80928897857666, "learning_rate": 3.447950381762932e-07, "loss": 2.7237, "step": 545250 }, { "epoch": 4.8155212914392695, "grad_norm": 8.36542797088623, "learning_rate": 3.431678253048909e-07, "loss": 2.5244, "step": 545300 }, { "epoch": 4.815962839329554, "grad_norm": 6.774470806121826, "learning_rate": 3.4154444800268416e-07, "loss": 2.6109, "step": 545350 }, { "epoch": 4.816404387219838, "grad_norm": 5.973906517028809, "learning_rate": 3.399249063950838e-07, "loss": 2.2001, "step": 545400 }, { "epoch": 4.816845935110122, "grad_norm": 6.1949992179870605, "learning_rate": 3.3830920060717865e-07, "loss": 2.737, "step": 545450 }, { "epoch": 4.817287483000406, "grad_norm": 5.628915786743164, "learning_rate": 3.3669733076376886e-07, "loss": 3.0368, "step": 545500 }, { "epoch": 4.817729030890691, "grad_norm": 7.78925895690918, "learning_rate": 3.3508929698936043e-07, "loss": 2.4585, "step": 545550 }, { "epoch": 4.8181705787809745, "grad_norm": 5.304439544677734, "learning_rate": 3.3348509940815396e-07, "loss": 2.8663, "step": 545600 }, { "epoch": 4.818612126671258, "grad_norm": 1.44192373752594, "learning_rate": 3.318847381440671e-07, "loss": 2.9117, "step": 545650 }, { "epoch": 4.819053674561543, "grad_norm": 5.3969221115112305, "learning_rate": 3.3028821332071193e-07, "loss": 2.3351, "step": 545700 }, { "epoch": 4.8194952224518275, "grad_norm": 7.324429035186768, "learning_rate": 3.28695525061401e-07, "loss": 2.0021, "step": 545750 }, { "epoch": 4.819936770342111, "grad_norm": 3.661332607269287, "learning_rate": 3.2710667348916367e-07, "loss": 2.1405, "step": 545800 }, { "epoch": 4.820378318232396, "grad_norm": 10.12942886352539, "learning_rate": 3.255216587267185e-07, "loss": 2.5566, "step": 545850 }, { "epoch": 4.8208198661226795, "grad_norm": 4.066982269287109, "learning_rate": 3.239404808964952e-07, "loss": 2.5643, "step": 545900 }, { "epoch": 4.821261414012964, "grad_norm": 7.489579677581787, "learning_rate": 3.2236314012063506e-07, "loss": 2.3249, "step": 545950 }, { "epoch": 4.821702961903248, "grad_norm": 4.916416168212891, "learning_rate": 3.2078963652096836e-07, "loss": 2.6984, "step": 546000 }, { "epoch": 4.821702961903248, "eval_asr_loss": 0.992874987239429, "eval_loss": 2.6893880367279053, "eval_runtime": 20.2215, "eval_samples_per_second": 37.979, "eval_steps_per_second": 9.495, "eval_tts_loss": 6.340675974771538, "step": 546000 }, { "epoch": 4.8221445097935325, "grad_norm": 6.586083889007568, "learning_rate": 3.192199702190368e-07, "loss": 2.7953, "step": 546050 }, { "epoch": 4.822586057683816, "grad_norm": 4.841409683227539, "learning_rate": 3.1765414133607676e-07, "loss": 2.3782, "step": 546100 }, { "epoch": 4.823027605574101, "grad_norm": 7.194849967956543, "learning_rate": 3.160921499930414e-07, "loss": 2.5527, "step": 546150 }, { "epoch": 4.8234691534643845, "grad_norm": 3.363933563232422, "learning_rate": 3.1453399631057865e-07, "loss": 2.4987, "step": 546200 }, { "epoch": 4.823910701354669, "grad_norm": 7.386734962463379, "learning_rate": 3.129796804090479e-07, "loss": 2.5997, "step": 546250 }, { "epoch": 4.824352249244953, "grad_norm": 5.39970064163208, "learning_rate": 3.11429202408503e-07, "loss": 2.5383, "step": 546300 }, { "epoch": 4.8247937971352375, "grad_norm": 4.955171585083008, "learning_rate": 3.0988256242870386e-07, "loss": 2.9132, "step": 546350 }, { "epoch": 4.825235345025521, "grad_norm": 2.8884949684143066, "learning_rate": 3.0837057901136335e-07, "loss": 2.4825, "step": 546400 }, { "epoch": 4.825676892915806, "grad_norm": 9.989749908447266, "learning_rate": 3.068315386648035e-07, "loss": 2.9145, "step": 546450 }, { "epoch": 4.8261184408060895, "grad_norm": 4.163097858428955, "learning_rate": 3.0529633669412196e-07, "loss": 2.7181, "step": 546500 }, { "epoch": 4.826559988696374, "grad_norm": 6.541525363922119, "learning_rate": 3.03764973217896e-07, "loss": 2.9466, "step": 546550 }, { "epoch": 4.827001536586658, "grad_norm": 5.61400842666626, "learning_rate": 3.022374483544199e-07, "loss": 2.5789, "step": 546600 }, { "epoch": 4.8274430844769425, "grad_norm": 9.056520462036133, "learning_rate": 3.007137622216716e-07, "loss": 2.3838, "step": 546650 }, { "epoch": 4.827884632367226, "grad_norm": 6.080948829650879, "learning_rate": 2.9919391493735125e-07, "loss": 2.9058, "step": 546700 }, { "epoch": 4.828326180257511, "grad_norm": 6.537165641784668, "learning_rate": 2.9767790661884844e-07, "loss": 2.4606, "step": 546750 }, { "epoch": 4.8287677281477945, "grad_norm": 8.050861358642578, "learning_rate": 2.961657373832638e-07, "loss": 2.8236, "step": 546800 }, { "epoch": 4.829209276038079, "grad_norm": 6.095705986022949, "learning_rate": 2.94657407347404e-07, "loss": 2.4253, "step": 546850 }, { "epoch": 4.829650823928363, "grad_norm": 6.240932941436768, "learning_rate": 2.931529166277647e-07, "loss": 2.5865, "step": 546900 }, { "epoch": 4.8300923718186475, "grad_norm": 9.486948013305664, "learning_rate": 2.91652265340564e-07, "loss": 2.4621, "step": 546950 }, { "epoch": 4.830533919708932, "grad_norm": 1.563857078552246, "learning_rate": 2.9015545360170925e-07, "loss": 2.8347, "step": 547000 }, { "epoch": 4.830975467599216, "grad_norm": 7.344498634338379, "learning_rate": 2.8866248152681907e-07, "loss": 2.5892, "step": 547050 }, { "epoch": 4.8314170154894995, "grad_norm": 8.049500465393066, "learning_rate": 2.8717334923121233e-07, "loss": 2.6001, "step": 547100 }, { "epoch": 4.831858563379784, "grad_norm": 7.91993522644043, "learning_rate": 2.856880568299136e-07, "loss": 2.656, "step": 547150 }, { "epoch": 4.832300111270069, "grad_norm": 3.4885454177856445, "learning_rate": 2.842066044376479e-07, "loss": 2.7216, "step": 547200 }, { "epoch": 4.8327416591603525, "grad_norm": 3.87715744972229, "learning_rate": 2.827289921688403e-07, "loss": 2.699, "step": 547250 }, { "epoch": 4.833183207050637, "grad_norm": 6.2923054695129395, "learning_rate": 2.812552201376273e-07, "loss": 2.7824, "step": 547300 }, { "epoch": 4.833624754940921, "grad_norm": 5.6777119636535645, "learning_rate": 2.797852884578511e-07, "loss": 2.7578, "step": 547350 }, { "epoch": 4.834066302831205, "grad_norm": 3.9347033500671387, "learning_rate": 2.783191972430432e-07, "loss": 2.5813, "step": 547400 }, { "epoch": 4.834507850721489, "grad_norm": 2.607419013977051, "learning_rate": 2.768569466064519e-07, "loss": 2.5135, "step": 547450 }, { "epoch": 4.834949398611774, "grad_norm": 4.514007568359375, "learning_rate": 2.7539853666102565e-07, "loss": 2.767, "step": 547500 }, { "epoch": 4.8353909465020575, "grad_norm": 9.338438034057617, "learning_rate": 2.739439675194022e-07, "loss": 2.8346, "step": 547550 }, { "epoch": 4.835832494392342, "grad_norm": 6.5621490478515625, "learning_rate": 2.724932392939417e-07, "loss": 2.4814, "step": 547600 }, { "epoch": 4.836274042282626, "grad_norm": 6.49846887588501, "learning_rate": 2.7104635209670436e-07, "loss": 3.083, "step": 547650 }, { "epoch": 4.83671559017291, "grad_norm": 9.907833099365234, "learning_rate": 2.6960330603945094e-07, "loss": 2.7762, "step": 547700 }, { "epoch": 4.837157138063194, "grad_norm": 3.5540528297424316, "learning_rate": 2.6816410123363666e-07, "loss": 2.6912, "step": 547750 }, { "epoch": 4.837598685953479, "grad_norm": 4.679515361785889, "learning_rate": 2.667287377904337e-07, "loss": 2.8499, "step": 547800 }, { "epoch": 4.8380402338437625, "grad_norm": 4.164608478546143, "learning_rate": 2.65297215820709e-07, "loss": 2.5142, "step": 547850 }, { "epoch": 4.838481781734047, "grad_norm": 4.959862232208252, "learning_rate": 2.638695354350351e-07, "loss": 3.0201, "step": 547900 }, { "epoch": 4.838923329624331, "grad_norm": 14.094888687133789, "learning_rate": 2.6244569674369056e-07, "loss": 2.5583, "step": 547950 }, { "epoch": 4.839364877514615, "grad_norm": 4.829261779785156, "learning_rate": 2.610256998566596e-07, "loss": 2.4571, "step": 548000 }, { "epoch": 4.839806425404899, "grad_norm": 2.927194833755493, "learning_rate": 2.5960954488361e-07, "loss": 2.7913, "step": 548050 }, { "epoch": 4.840247973295184, "grad_norm": 4.921788692474365, "learning_rate": 2.581972319339376e-07, "loss": 2.3908, "step": 548100 }, { "epoch": 4.8406895211854675, "grad_norm": 5.1113152503967285, "learning_rate": 2.567887611167386e-07, "loss": 2.713, "step": 548150 }, { "epoch": 4.841131069075752, "grad_norm": 8.967625617980957, "learning_rate": 2.5538413254079243e-07, "loss": 2.5218, "step": 548200 }, { "epoch": 4.841572616966036, "grad_norm": 9.870969772338867, "learning_rate": 2.539833463146013e-07, "loss": 2.4227, "step": 548250 }, { "epoch": 4.84201416485632, "grad_norm": 6.731603145599365, "learning_rate": 2.52586402546362e-07, "loss": 2.7353, "step": 548300 }, { "epoch": 4.842455712746604, "grad_norm": 6.814058780670166, "learning_rate": 2.5119330134397714e-07, "loss": 2.4414, "step": 548350 }, { "epoch": 4.842897260636889, "grad_norm": 8.191095352172852, "learning_rate": 2.4980404281504943e-07, "loss": 3.0537, "step": 548400 }, { "epoch": 4.843338808527173, "grad_norm": 8.364789009094238, "learning_rate": 2.484462977219104e-07, "loss": 2.5069, "step": 548450 }, { "epoch": 4.843780356417457, "grad_norm": 2.9945638179779053, "learning_rate": 2.470646480027372e-07, "loss": 2.4419, "step": 548500 }, { "epoch": 4.844221904307741, "grad_norm": 4.964893817901611, "learning_rate": 2.4568684127592167e-07, "loss": 3.2057, "step": 548550 }, { "epoch": 4.844663452198025, "grad_norm": 10.55528736114502, "learning_rate": 2.443128776478898e-07, "loss": 3.0882, "step": 548600 }, { "epoch": 4.84510500008831, "grad_norm": 14.146993637084961, "learning_rate": 2.429427572247789e-07, "loss": 3.1916, "step": 548650 }, { "epoch": 4.845546547978594, "grad_norm": 4.673105716705322, "learning_rate": 2.415764801124043e-07, "loss": 2.5862, "step": 548700 }, { "epoch": 4.8459880958688775, "grad_norm": 3.2673256397247314, "learning_rate": 2.4021404641631496e-07, "loss": 2.6145, "step": 548750 }, { "epoch": 4.846429643759162, "grad_norm": 10.125835418701172, "learning_rate": 2.388554562417378e-07, "loss": 2.5147, "step": 548800 }, { "epoch": 4.846871191649447, "grad_norm": 4.140436172485352, "learning_rate": 2.375007096936166e-07, "loss": 2.559, "step": 548850 }, { "epoch": 4.84731273953973, "grad_norm": 5.512503147125244, "learning_rate": 2.3614980687658994e-07, "loss": 2.4044, "step": 548900 }, { "epoch": 4.847754287430015, "grad_norm": 5.625319004058838, "learning_rate": 2.3480274789500211e-07, "loss": 2.5661, "step": 548950 }, { "epoch": 4.848195835320299, "grad_norm": 4.6724467277526855, "learning_rate": 2.3345953285290879e-07, "loss": 2.3243, "step": 549000 }, { "epoch": 4.848195835320299, "eval_asr_loss": 0.9941714023966352, "eval_loss": 2.688934564590454, "eval_runtime": 20.4592, "eval_samples_per_second": 37.538, "eval_steps_per_second": 9.385, "eval_tts_loss": 6.340601376659668, "step": 549000 }, { "epoch": 4.848637383210583, "grad_norm": 3.414532423019409, "learning_rate": 2.3212016185406027e-07, "loss": 2.5216, "step": 549050 }, { "epoch": 4.849078931100867, "grad_norm": 5.936755657196045, "learning_rate": 2.308113078656471e-07, "loss": 3.0045, "step": 549100 }, { "epoch": 4.849520478991152, "grad_norm": 7.536297798156738, "learning_rate": 2.2947954837734266e-07, "loss": 2.815, "step": 549150 }, { "epoch": 4.849962026881435, "grad_norm": 10.217138290405273, "learning_rate": 2.281516332397038e-07, "loss": 2.6413, "step": 549200 }, { "epoch": 4.85040357477172, "grad_norm": 5.510462284088135, "learning_rate": 2.2682756255529847e-07, "loss": 3.2657, "step": 549250 }, { "epoch": 4.850845122662004, "grad_norm": 5.238913536071777, "learning_rate": 2.2550733642640042e-07, "loss": 2.4922, "step": 549300 }, { "epoch": 4.851286670552288, "grad_norm": 5.386260032653809, "learning_rate": 2.2419095495498366e-07, "loss": 2.6411, "step": 549350 }, { "epoch": 4.851728218442572, "grad_norm": 7.662790298461914, "learning_rate": 2.228784182427335e-07, "loss": 2.6091, "step": 549400 }, { "epoch": 4.852169766332857, "grad_norm": 1.357452630996704, "learning_rate": 2.2156972639102435e-07, "loss": 2.3999, "step": 549450 }, { "epoch": 4.85261131422314, "grad_norm": 4.401924133300781, "learning_rate": 2.202648795009421e-07, "loss": 2.6909, "step": 549500 }, { "epoch": 4.853052862113425, "grad_norm": 5.273355484008789, "learning_rate": 2.1896387767328386e-07, "loss": 2.4616, "step": 549550 }, { "epoch": 4.853494410003709, "grad_norm": 6.961762428283691, "learning_rate": 2.1766672100853037e-07, "loss": 2.5079, "step": 549600 }, { "epoch": 4.853935957893993, "grad_norm": 7.287909030914307, "learning_rate": 2.163734096068848e-07, "loss": 2.7669, "step": 549650 }, { "epoch": 4.854377505784277, "grad_norm": 8.083537101745605, "learning_rate": 2.1508394356823946e-07, "loss": 2.8387, "step": 549700 }, { "epoch": 4.854819053674562, "grad_norm": 7.466681480407715, "learning_rate": 2.137983229921925e-07, "loss": 2.2183, "step": 549750 }, { "epoch": 4.855260601564845, "grad_norm": 2.2878875732421875, "learning_rate": 2.1251654797805332e-07, "loss": 2.7327, "step": 549800 }, { "epoch": 4.85570214945513, "grad_norm": 9.346505165100098, "learning_rate": 2.1123861862482608e-07, "loss": 2.1066, "step": 549850 }, { "epoch": 4.856143697345414, "grad_norm": 3.0069406032562256, "learning_rate": 2.0996453503121517e-07, "loss": 2.8464, "step": 549900 }, { "epoch": 4.856585245235698, "grad_norm": 3.576687812805176, "learning_rate": 2.0869429729564184e-07, "loss": 2.7739, "step": 549950 }, { "epoch": 4.857026793125982, "grad_norm": 2.5808603763580322, "learning_rate": 2.074279055162165e-07, "loss": 2.5761, "step": 550000 }, { "epoch": 4.857468341016267, "grad_norm": 7.67838716506958, "learning_rate": 2.061653597907498e-07, "loss": 2.7589, "step": 550050 }, { "epoch": 4.857909888906551, "grad_norm": 7.5850677490234375, "learning_rate": 2.0490666021677485e-07, "loss": 2.6747, "step": 550100 }, { "epoch": 4.858351436796835, "grad_norm": 12.306727409362793, "learning_rate": 2.036518068915083e-07, "loss": 2.5778, "step": 550150 }, { "epoch": 4.858792984687119, "grad_norm": 6.202439308166504, "learning_rate": 2.0240079991187822e-07, "loss": 2.761, "step": 550200 }, { "epoch": 4.859234532577403, "grad_norm": 5.10723876953125, "learning_rate": 2.0115363937451836e-07, "loss": 2.9396, "step": 550250 }, { "epoch": 4.859676080467688, "grad_norm": 5.1198296546936035, "learning_rate": 1.9991032537575727e-07, "loss": 2.8842, "step": 550300 }, { "epoch": 4.860117628357972, "grad_norm": 5.849010944366455, "learning_rate": 1.9867085801162921e-07, "loss": 2.7346, "step": 550350 }, { "epoch": 4.860559176248255, "grad_norm": 3.9947669506073, "learning_rate": 1.9743523737787983e-07, "loss": 3.1312, "step": 550400 }, { "epoch": 4.86100072413854, "grad_norm": 10.643601417541504, "learning_rate": 1.9620346356993836e-07, "loss": 2.8736, "step": 550450 }, { "epoch": 4.861442272028825, "grad_norm": 6.949484348297119, "learning_rate": 1.9497553668295644e-07, "loss": 2.4466, "step": 550500 }, { "epoch": 4.861883819919108, "grad_norm": 5.958240985870361, "learning_rate": 1.9375145681178043e-07, "loss": 2.4708, "step": 550550 }, { "epoch": 4.862325367809393, "grad_norm": 9.698393821716309, "learning_rate": 1.9253122405096246e-07, "loss": 2.4455, "step": 550600 }, { "epoch": 4.862766915699677, "grad_norm": 3.5544586181640625, "learning_rate": 1.9131483849474941e-07, "loss": 2.6942, "step": 550650 }, { "epoch": 4.863208463589961, "grad_norm": 2.3287644386291504, "learning_rate": 1.9010230023709942e-07, "loss": 2.3462, "step": 550700 }, { "epoch": 4.863650011480245, "grad_norm": 11.045299530029297, "learning_rate": 1.8889360937167645e-07, "loss": 2.4995, "step": 550750 }, { "epoch": 4.86409155937053, "grad_norm": 5.540194511413574, "learning_rate": 1.876887659918336e-07, "loss": 2.347, "step": 550800 }, { "epoch": 4.864533107260813, "grad_norm": 5.1925272941589355, "learning_rate": 1.8648777019063533e-07, "loss": 2.9292, "step": 550850 }, { "epoch": 4.864974655151098, "grad_norm": 8.782684326171875, "learning_rate": 1.852906220608519e-07, "loss": 2.5909, "step": 550900 }, { "epoch": 4.865416203041382, "grad_norm": 9.679356575012207, "learning_rate": 1.8409732169495375e-07, "loss": 2.6375, "step": 550950 }, { "epoch": 4.865857750931666, "grad_norm": 2.71787166595459, "learning_rate": 1.8290786918510604e-07, "loss": 2.5566, "step": 551000 }, { "epoch": 4.86629929882195, "grad_norm": 5.317314147949219, "learning_rate": 1.8172226462319085e-07, "loss": 2.1164, "step": 551050 }, { "epoch": 4.866740846712235, "grad_norm": 8.148841857910156, "learning_rate": 1.8054050810078492e-07, "loss": 2.7089, "step": 551100 }, { "epoch": 4.867182394602518, "grad_norm": 3.415299415588379, "learning_rate": 1.793625997091708e-07, "loss": 2.8293, "step": 551150 }, { "epoch": 4.867623942492803, "grad_norm": 4.81337308883667, "learning_rate": 1.7818853953932568e-07, "loss": 2.5038, "step": 551200 }, { "epoch": 4.868065490383087, "grad_norm": 3.428471088409424, "learning_rate": 1.7701832768193815e-07, "loss": 2.5195, "step": 551250 }, { "epoch": 4.868507038273371, "grad_norm": 6.012335300445557, "learning_rate": 1.7585196422740257e-07, "loss": 2.2374, "step": 551300 }, { "epoch": 4.868948586163655, "grad_norm": 4.762035369873047, "learning_rate": 1.7468944926580243e-07, "loss": 2.7668, "step": 551350 }, { "epoch": 4.86939013405394, "grad_norm": 4.426942348480225, "learning_rate": 1.7353078288693815e-07, "loss": 2.6517, "step": 551400 }, { "epoch": 4.869831681944223, "grad_norm": 7.7118024826049805, "learning_rate": 1.7237596518030474e-07, "loss": 2.5277, "step": 551450 }, { "epoch": 4.870273229834508, "grad_norm": 3.7506134510040283, "learning_rate": 1.7122499623509758e-07, "loss": 2.6034, "step": 551500 }, { "epoch": 4.870714777724792, "grad_norm": 8.349372863769531, "learning_rate": 1.7007787614022884e-07, "loss": 2.5721, "step": 551550 }, { "epoch": 4.871156325615076, "grad_norm": 6.145127773284912, "learning_rate": 1.6893460498429992e-07, "loss": 2.4726, "step": 551600 }, { "epoch": 4.87159787350536, "grad_norm": 3.065495491027832, "learning_rate": 1.6779518285561236e-07, "loss": 2.2395, "step": 551650 }, { "epoch": 4.872039421395645, "grad_norm": 4.930973052978516, "learning_rate": 1.6665960984218466e-07, "loss": 3.1179, "step": 551700 }, { "epoch": 4.872480969285929, "grad_norm": 3.7829372882843018, "learning_rate": 1.6552788603172997e-07, "loss": 2.9143, "step": 551750 }, { "epoch": 4.872922517176213, "grad_norm": 2.0169973373413086, "learning_rate": 1.6440001151166174e-07, "loss": 2.6794, "step": 551800 }, { "epoch": 4.873364065066497, "grad_norm": 6.8479838371276855, "learning_rate": 1.632759863690936e-07, "loss": 2.656, "step": 551850 }, { "epoch": 4.873805612956781, "grad_norm": 4.610270977020264, "learning_rate": 1.6215581069085604e-07, "loss": 2.237, "step": 551900 }, { "epoch": 4.874247160847066, "grad_norm": 7.37453556060791, "learning_rate": 1.610394845634744e-07, "loss": 2.9592, "step": 551950 }, { "epoch": 4.87468870873735, "grad_norm": 8.633584022521973, "learning_rate": 1.5992700807316298e-07, "loss": 2.7112, "step": 552000 }, { "epoch": 4.87468870873735, "eval_asr_loss": 0.994084052867598, "eval_loss": 2.68914532661438, "eval_runtime": 20.4229, "eval_samples_per_second": 37.605, "eval_steps_per_second": 9.401, "eval_tts_loss": 6.339456706923581, "step": 552000 }, { "epoch": 4.875130256627634, "grad_norm": 5.258874416351318, "learning_rate": 1.5881838130586412e-07, "loss": 3.0192, "step": 552050 }, { "epoch": 4.875571804517918, "grad_norm": 8.707131385803223, "learning_rate": 1.5771360434719827e-07, "loss": 2.7954, "step": 552100 }, { "epoch": 4.8760133524082026, "grad_norm": 5.66416072845459, "learning_rate": 1.5661267728251383e-07, "loss": 2.5506, "step": 552150 }, { "epoch": 4.876454900298486, "grad_norm": 7.293725490570068, "learning_rate": 1.5551560019683164e-07, "loss": 2.9627, "step": 552200 }, { "epoch": 4.876896448188771, "grad_norm": 11.71657943725586, "learning_rate": 1.5442237317490614e-07, "loss": 2.6377, "step": 552250 }, { "epoch": 4.877337996079055, "grad_norm": 7.65678071975708, "learning_rate": 1.5333299630117538e-07, "loss": 2.3259, "step": 552300 }, { "epoch": 4.877779543969339, "grad_norm": 3.8141121864318848, "learning_rate": 1.5224746965978309e-07, "loss": 3.0405, "step": 552350 }, { "epoch": 4.878221091859623, "grad_norm": 5.328624248504639, "learning_rate": 1.5116579333457336e-07, "loss": 2.7129, "step": 552400 }, { "epoch": 4.8786626397499075, "grad_norm": 10.526511192321777, "learning_rate": 1.500879674091016e-07, "loss": 2.5384, "step": 552450 }, { "epoch": 4.879104187640191, "grad_norm": 2.7948849201202393, "learning_rate": 1.4901399196662337e-07, "loss": 2.7205, "step": 552500 }, { "epoch": 4.879545735530476, "grad_norm": 6.65280818939209, "learning_rate": 1.4794386709008345e-07, "loss": 2.6837, "step": 552550 }, { "epoch": 4.87998728342076, "grad_norm": 5.379391670227051, "learning_rate": 1.4687759286215463e-07, "loss": 2.6104, "step": 552600 }, { "epoch": 4.880428831311044, "grad_norm": 7.541629791259766, "learning_rate": 1.4581516936518213e-07, "loss": 2.9379, "step": 552650 }, { "epoch": 4.880870379201328, "grad_norm": 10.940437316894531, "learning_rate": 1.4475659668124474e-07, "loss": 2.3073, "step": 552700 }, { "epoch": 4.8813119270916125, "grad_norm": 3.4756863117218018, "learning_rate": 1.4370187489209374e-07, "loss": 2.3669, "step": 552750 }, { "epoch": 4.881753474981896, "grad_norm": 5.981154918670654, "learning_rate": 1.4265100407920283e-07, "loss": 2.4534, "step": 552800 }, { "epoch": 4.882195022872181, "grad_norm": 6.95257043838501, "learning_rate": 1.41603984323746e-07, "loss": 2.8655, "step": 552850 }, { "epoch": 4.882636570762465, "grad_norm": 12.40413761138916, "learning_rate": 1.405608157065974e-07, "loss": 2.7816, "step": 552900 }, { "epoch": 4.883078118652749, "grad_norm": 7.025955677032471, "learning_rate": 1.3952149830832595e-07, "loss": 2.8474, "step": 552950 }, { "epoch": 4.883519666543033, "grad_norm": 2.9472033977508545, "learning_rate": 1.384860322092174e-07, "loss": 2.2448, "step": 553000 }, { "epoch": 4.8839612144333175, "grad_norm": 7.464934825897217, "learning_rate": 1.3745441748925225e-07, "loss": 2.6979, "step": 553050 }, { "epoch": 4.884402762323601, "grad_norm": 5.416768550872803, "learning_rate": 1.3642665422810563e-07, "loss": 2.5532, "step": 553100 }, { "epoch": 4.884844310213886, "grad_norm": 4.0006537437438965, "learning_rate": 1.354027425051696e-07, "loss": 2.8731, "step": 553150 }, { "epoch": 4.88528585810417, "grad_norm": 10.534464836120605, "learning_rate": 1.343826823995309e-07, "loss": 2.7432, "step": 553200 }, { "epoch": 4.885727405994454, "grad_norm": 5.106250762939453, "learning_rate": 1.3336647398998757e-07, "loss": 2.7294, "step": 553250 }, { "epoch": 4.886168953884738, "grad_norm": 10.211413383483887, "learning_rate": 1.323541173550269e-07, "loss": 2.859, "step": 553300 }, { "epoch": 4.8866105017750225, "grad_norm": 4.8535590171813965, "learning_rate": 1.3134561257284183e-07, "loss": 2.895, "step": 553350 }, { "epoch": 4.887052049665307, "grad_norm": 3.7240257263183594, "learning_rate": 1.3036101502894027e-07, "loss": 2.6485, "step": 553400 }, { "epoch": 4.887493597555591, "grad_norm": 4.9032087326049805, "learning_rate": 1.2936013714479055e-07, "loss": 2.4319, "step": 553450 }, { "epoch": 4.887935145445875, "grad_norm": 9.214579582214355, "learning_rate": 1.283830141097564e-07, "loss": 2.8107, "step": 553500 }, { "epoch": 4.888376693336159, "grad_norm": 7.487960338592529, "learning_rate": 1.2738976342672182e-07, "loss": 3.0562, "step": 553550 }, { "epoch": 4.888818241226444, "grad_norm": 5.7209343910217285, "learning_rate": 1.2640036497991724e-07, "loss": 2.5261, "step": 553600 }, { "epoch": 4.8892597891167275, "grad_norm": 10.069497108459473, "learning_rate": 1.2541481884577043e-07, "loss": 2.9355, "step": 553650 }, { "epoch": 4.889701337007012, "grad_norm": 6.420971870422363, "learning_rate": 1.2443312510040383e-07, "loss": 2.5345, "step": 553700 }, { "epoch": 4.890142884897296, "grad_norm": 6.774078369140625, "learning_rate": 1.2345528381964566e-07, "loss": 2.4348, "step": 553750 }, { "epoch": 4.8905844327875805, "grad_norm": 6.295166492462158, "learning_rate": 1.2248129507902996e-07, "loss": 2.6097, "step": 553800 }, { "epoch": 4.891025980677864, "grad_norm": 3.6546525955200195, "learning_rate": 1.2151115895377985e-07, "loss": 2.7233, "step": 553850 }, { "epoch": 4.891467528568149, "grad_norm": 8.256914138793945, "learning_rate": 1.2054487551884097e-07, "loss": 2.6238, "step": 553900 }, { "epoch": 4.8919090764584325, "grad_norm": 3.5804319381713867, "learning_rate": 1.1958244484883696e-07, "loss": 2.9793, "step": 553950 }, { "epoch": 4.892350624348717, "grad_norm": 4.752839088439941, "learning_rate": 1.1862386701811945e-07, "loss": 2.8673, "step": 554000 }, { "epoch": 4.892792172239001, "grad_norm": 5.823893070220947, "learning_rate": 1.1766914210072922e-07, "loss": 2.9144, "step": 554050 }, { "epoch": 4.8932337201292855, "grad_norm": 8.935931205749512, "learning_rate": 1.1671827017040727e-07, "loss": 2.4792, "step": 554100 }, { "epoch": 4.893675268019569, "grad_norm": 3.274867534637451, "learning_rate": 1.1577125130060041e-07, "loss": 2.5957, "step": 554150 }, { "epoch": 4.894116815909854, "grad_norm": 9.02548885345459, "learning_rate": 1.1482808556446123e-07, "loss": 2.4975, "step": 554200 }, { "epoch": 4.8945583638001375, "grad_norm": 3.4948530197143555, "learning_rate": 1.1388877303483147e-07, "loss": 2.8702, "step": 554250 }, { "epoch": 4.894999911690422, "grad_norm": 7.667669296264648, "learning_rate": 1.1295331378427531e-07, "loss": 2.6402, "step": 554300 }, { "epoch": 4.895441459580706, "grad_norm": 2.7134501934051514, "learning_rate": 1.1202170788504607e-07, "loss": 2.6453, "step": 554350 }, { "epoch": 4.8958830074709905, "grad_norm": 1.8545875549316406, "learning_rate": 1.1109395540910283e-07, "loss": 2.8746, "step": 554400 }, { "epoch": 4.896324555361274, "grad_norm": 2.80493426322937, "learning_rate": 1.1017005642810496e-07, "loss": 2.5545, "step": 554450 }, { "epoch": 4.896766103251559, "grad_norm": 6.316068172454834, "learning_rate": 1.0925001101342313e-07, "loss": 2.4017, "step": 554500 }, { "epoch": 4.8972076511418425, "grad_norm": 9.477471351623535, "learning_rate": 1.0833381923611164e-07, "loss": 2.4498, "step": 554550 }, { "epoch": 4.897649199032127, "grad_norm": 2.589768886566162, "learning_rate": 1.0742148116694162e-07, "loss": 2.5613, "step": 554600 }, { "epoch": 4.898090746922411, "grad_norm": 6.268969535827637, "learning_rate": 1.065129968763845e-07, "loss": 2.767, "step": 554650 }, { "epoch": 4.8985322948126955, "grad_norm": 7.9310832023620605, "learning_rate": 1.0560836643461747e-07, "loss": 2.7398, "step": 554700 }, { "epoch": 4.898973842702979, "grad_norm": 5.431900978088379, "learning_rate": 1.0470758991151242e-07, "loss": 2.4485, "step": 554750 }, { "epoch": 4.899415390593264, "grad_norm": 5.795566082000732, "learning_rate": 1.0381066737664702e-07, "loss": 2.7808, "step": 554800 }, { "epoch": 4.899856938483548, "grad_norm": 4.372450828552246, "learning_rate": 1.029175988992992e-07, "loss": 2.5563, "step": 554850 }, { "epoch": 4.900298486373832, "grad_norm": 2.359114170074463, "learning_rate": 1.0202838454844709e-07, "loss": 2.7182, "step": 554900 }, { "epoch": 4.900740034264116, "grad_norm": 3.3675973415374756, "learning_rate": 1.0114302439278578e-07, "loss": 2.4392, "step": 554950 }, { "epoch": 4.9011815821544005, "grad_norm": 5.202480316162109, "learning_rate": 1.0026151850069388e-07, "loss": 2.5701, "step": 555000 }, { "epoch": 4.9011815821544005, "eval_asr_loss": 0.9945843600029236, "eval_loss": 2.688945770263672, "eval_runtime": 20.1215, "eval_samples_per_second": 38.168, "eval_steps_per_second": 9.542, "eval_tts_loss": 6.340792651246921, "step": 555000 }, { "epoch": 4.901623130044685, "grad_norm": 3.6770431995391846, "learning_rate": 9.938386694026137e-08, "loss": 2.3308, "step": 555050 }, { "epoch": 4.902064677934969, "grad_norm": 5.068836688995361, "learning_rate": 9.851006977928401e-08, "loss": 2.6172, "step": 555100 }, { "epoch": 4.9025062258252525, "grad_norm": 4.029082298278809, "learning_rate": 9.764012708524672e-08, "loss": 2.7166, "step": 555150 }, { "epoch": 4.902947773715537, "grad_norm": 5.604940891265869, "learning_rate": 9.67740389253513e-08, "loss": 2.9418, "step": 555200 }, { "epoch": 4.903389321605822, "grad_norm": 7.620632171630859, "learning_rate": 9.591180536649424e-08, "loss": 2.993, "step": 555250 }, { "epoch": 4.9038308694961055, "grad_norm": 6.546438217163086, "learning_rate": 9.505342647527782e-08, "loss": 2.5407, "step": 555300 }, { "epoch": 4.90427241738639, "grad_norm": 4.044014930725098, "learning_rate": 9.419890231800454e-08, "loss": 2.74, "step": 555350 }, { "epoch": 4.904713965276674, "grad_norm": 2.6660854816436768, "learning_rate": 9.334823296067163e-08, "loss": 2.3539, "step": 555400 }, { "epoch": 4.905155513166958, "grad_norm": 7.467266082763672, "learning_rate": 9.250141846899318e-08, "loss": 2.5608, "step": 555450 }, { "epoch": 4.905597061057242, "grad_norm": 4.501576900482178, "learning_rate": 9.165845890837799e-08, "loss": 2.6852, "step": 555500 }, { "epoch": 4.906038608947527, "grad_norm": 2.5871243476867676, "learning_rate": 9.081935434394063e-08, "loss": 2.8776, "step": 555550 }, { "epoch": 4.9064801568378105, "grad_norm": 5.063958644866943, "learning_rate": 8.99841048404848e-08, "loss": 2.8392, "step": 555600 }, { "epoch": 4.906921704728095, "grad_norm": 8.815516471862793, "learning_rate": 8.91527104625367e-08, "loss": 2.602, "step": 555650 }, { "epoch": 4.907363252618379, "grad_norm": 5.470418930053711, "learning_rate": 8.832517127430606e-08, "loss": 2.6068, "step": 555700 }, { "epoch": 4.907804800508663, "grad_norm": 5.735939979553223, "learning_rate": 8.750148733971952e-08, "loss": 2.8034, "step": 555750 }, { "epoch": 4.908246348398947, "grad_norm": 4.411189556121826, "learning_rate": 8.668165872239287e-08, "loss": 2.7158, "step": 555800 }, { "epoch": 4.908687896289232, "grad_norm": 4.592665195465088, "learning_rate": 8.586568548565876e-08, "loss": 2.3846, "step": 555850 }, { "epoch": 4.9091294441795155, "grad_norm": 6.8963446617126465, "learning_rate": 8.506977226465096e-08, "loss": 3.0013, "step": 555900 }, { "epoch": 4.9095709920698, "grad_norm": 6.223936557769775, "learning_rate": 8.426143286713872e-08, "loss": 2.8395, "step": 555950 }, { "epoch": 4.910012539960084, "grad_norm": 6.529476642608643, "learning_rate": 8.345694903716395e-08, "loss": 2.374, "step": 556000 }, { "epoch": 4.910454087850368, "grad_norm": 4.900590896606445, "learning_rate": 8.26563208368547e-08, "loss": 2.3859, "step": 556050 }, { "epoch": 4.910895635740652, "grad_norm": 8.408040046691895, "learning_rate": 8.185954832805597e-08, "loss": 2.4667, "step": 556100 }, { "epoch": 4.911337183630937, "grad_norm": 6.904308795928955, "learning_rate": 8.106663157231298e-08, "loss": 2.7571, "step": 556150 }, { "epoch": 4.9117787315212205, "grad_norm": 2.604743242263794, "learning_rate": 8.02775706308767e-08, "loss": 2.3868, "step": 556200 }, { "epoch": 4.912220279411505, "grad_norm": 3.3704566955566406, "learning_rate": 7.949236556468731e-08, "loss": 2.7669, "step": 556250 }, { "epoch": 4.912661827301789, "grad_norm": 3.263522148132324, "learning_rate": 7.871101643440181e-08, "loss": 2.6836, "step": 556300 }, { "epoch": 4.913103375192073, "grad_norm": 4.235065460205078, "learning_rate": 7.793352330037196e-08, "loss": 2.5687, "step": 556350 }, { "epoch": 4.913544923082357, "grad_norm": 8.754676818847656, "learning_rate": 7.715988622264414e-08, "loss": 3.0368, "step": 556400 }, { "epoch": 4.913986470972642, "grad_norm": 2.740952253341675, "learning_rate": 7.639010526098722e-08, "loss": 2.7091, "step": 556450 }, { "epoch": 4.914428018862926, "grad_norm": 5.42210054397583, "learning_rate": 7.562418047485364e-08, "loss": 2.5477, "step": 556500 }, { "epoch": 4.91486956675321, "grad_norm": 7.222198009490967, "learning_rate": 7.48621119234072e-08, "loss": 2.5593, "step": 556550 }, { "epoch": 4.915311114643494, "grad_norm": 5.826037883758545, "learning_rate": 7.410389966550635e-08, "loss": 2.8855, "step": 556600 }, { "epoch": 4.915752662533778, "grad_norm": 5.579258441925049, "learning_rate": 7.334954375972647e-08, "loss": 2.7961, "step": 556650 }, { "epoch": 4.916194210424063, "grad_norm": 5.281018257141113, "learning_rate": 7.259904426432096e-08, "loss": 2.9808, "step": 556700 }, { "epoch": 4.916635758314347, "grad_norm": 6.017827033996582, "learning_rate": 7.185240123727121e-08, "loss": 2.6102, "step": 556750 }, { "epoch": 4.9170773062046305, "grad_norm": 1.8698238134384155, "learning_rate": 7.110961473624777e-08, "loss": 2.5617, "step": 556800 }, { "epoch": 4.917518854094915, "grad_norm": 10.753763198852539, "learning_rate": 7.037068481861586e-08, "loss": 2.9218, "step": 556850 }, { "epoch": 4.9179604019852, "grad_norm": 8.849223136901855, "learning_rate": 6.963561154146315e-08, "loss": 2.6223, "step": 556900 }, { "epoch": 4.918401949875483, "grad_norm": 3.7402398586273193, "learning_rate": 6.890439496155531e-08, "loss": 2.7715, "step": 556950 }, { "epoch": 4.918843497765768, "grad_norm": 10.205100059509277, "learning_rate": 6.817703513538609e-08, "loss": 2.6046, "step": 557000 }, { "epoch": 4.919285045656052, "grad_norm": 7.644894599914551, "learning_rate": 6.745353211912719e-08, "loss": 2.7552, "step": 557050 }, { "epoch": 4.919726593546336, "grad_norm": 3.6151485443115234, "learning_rate": 6.673388596867281e-08, "loss": 2.6331, "step": 557100 }, { "epoch": 4.92016814143662, "grad_norm": 8.000495910644531, "learning_rate": 6.601809673960068e-08, "loss": 2.8704, "step": 557150 }, { "epoch": 4.920609689326905, "grad_norm": 6.456025123596191, "learning_rate": 6.530616448719995e-08, "loss": 2.9014, "step": 557200 }, { "epoch": 4.921051237217188, "grad_norm": 6.165225028991699, "learning_rate": 6.459808926647105e-08, "loss": 2.5828, "step": 557250 }, { "epoch": 4.921492785107473, "grad_norm": 6.785088062286377, "learning_rate": 6.389387113209244e-08, "loss": 2.3898, "step": 557300 }, { "epoch": 4.921934332997757, "grad_norm": 4.38185453414917, "learning_rate": 6.319351013847063e-08, "loss": 2.8014, "step": 557350 }, { "epoch": 4.922375880888041, "grad_norm": 2.3354074954986572, "learning_rate": 6.249700633969568e-08, "loss": 2.7655, "step": 557400 }, { "epoch": 4.922817428778325, "grad_norm": 4.9754719734191895, "learning_rate": 6.180435978957455e-08, "loss": 2.3917, "step": 557450 }, { "epoch": 4.92325897666861, "grad_norm": 5.331031322479248, "learning_rate": 6.111557054159778e-08, "loss": 2.5137, "step": 557500 }, { "epoch": 4.923700524558893, "grad_norm": 5.425865173339844, "learning_rate": 6.043063864897836e-08, "loss": 2.7304, "step": 557550 }, { "epoch": 4.924142072449178, "grad_norm": 4.934122562408447, "learning_rate": 5.974956416461286e-08, "loss": 2.6935, "step": 557600 }, { "epoch": 4.924583620339462, "grad_norm": 5.709440231323242, "learning_rate": 5.907234714111476e-08, "loss": 2.561, "step": 557650 }, { "epoch": 4.925025168229746, "grad_norm": 6.91804838180542, "learning_rate": 5.839898763079221e-08, "loss": 2.7552, "step": 557700 }, { "epoch": 4.92546671612003, "grad_norm": 8.095946311950684, "learning_rate": 5.7729485685653616e-08, "loss": 2.6712, "step": 557750 }, { "epoch": 4.925908264010315, "grad_norm": 4.723079204559326, "learning_rate": 5.706384135741316e-08, "loss": 2.6721, "step": 557800 }, { "epoch": 4.926349811900598, "grad_norm": 3.9202797412872314, "learning_rate": 5.6402054697490824e-08, "loss": 2.5324, "step": 557850 }, { "epoch": 4.926791359790883, "grad_norm": 2.576738119125366, "learning_rate": 5.5744125756995723e-08, "loss": 2.2432, "step": 557900 }, { "epoch": 4.927232907681167, "grad_norm": 7.853531360626221, "learning_rate": 5.509005458675387e-08, "loss": 2.6824, "step": 557950 }, { "epoch": 4.927674455571451, "grad_norm": 3.3741707801818848, "learning_rate": 5.443984123728596e-08, "loss": 2.645, "step": 558000 }, { "epoch": 4.927674455571451, "eval_asr_loss": 0.994350412375674, "eval_loss": 2.6888606548309326, "eval_runtime": 20.5621, "eval_samples_per_second": 37.35, "eval_steps_per_second": 9.338, "eval_tts_loss": 6.339131751952347, "step": 558000 }, { "epoch": 4.928116003461735, "grad_norm": 5.1852498054504395, "learning_rate": 5.3793485758812934e-08, "loss": 2.9243, "step": 558050 }, { "epoch": 4.92855755135202, "grad_norm": 11.041584014892578, "learning_rate": 5.315098820126152e-08, "loss": 2.7285, "step": 558100 }, { "epoch": 4.928999099242304, "grad_norm": 14.26244831085205, "learning_rate": 5.251234861426424e-08, "loss": 2.6621, "step": 558150 }, { "epoch": 4.929440647132588, "grad_norm": 3.0989010334014893, "learning_rate": 5.189022486957185e-08, "loss": 2.872, "step": 558200 }, { "epoch": 4.929882195022872, "grad_norm": 10.555662155151367, "learning_rate": 5.125922420949891e-08, "loss": 2.6082, "step": 558250 }, { "epoch": 4.930323742913156, "grad_norm": 7.8833441734313965, "learning_rate": 5.063208166610211e-08, "loss": 2.6759, "step": 558300 }, { "epoch": 4.930765290803441, "grad_norm": 2.4202349185943604, "learning_rate": 5.0008797287820485e-08, "loss": 2.3913, "step": 558350 }, { "epoch": 4.931206838693725, "grad_norm": 7.136029243469238, "learning_rate": 4.938937112279329e-08, "loss": 3.0643, "step": 558400 }, { "epoch": 4.931648386584009, "grad_norm": 5.8725104331970215, "learning_rate": 4.877380321887115e-08, "loss": 2.6721, "step": 558450 }, { "epoch": 4.932089934474293, "grad_norm": 6.561997890472412, "learning_rate": 4.8162093623593806e-08, "loss": 2.5584, "step": 558500 }, { "epoch": 4.932531482364578, "grad_norm": 8.710442543029785, "learning_rate": 4.7554242384217905e-08, "loss": 2.604, "step": 558550 }, { "epoch": 4.932973030254861, "grad_norm": 6.900908946990967, "learning_rate": 4.695024954769478e-08, "loss": 2.5591, "step": 558600 }, { "epoch": 4.933414578145146, "grad_norm": 3.4531795978546143, "learning_rate": 4.6350115160670446e-08, "loss": 2.3794, "step": 558650 }, { "epoch": 4.93385612603543, "grad_norm": 9.164139747619629, "learning_rate": 4.5753839269507824e-08, "loss": 2.8531, "step": 558700 }, { "epoch": 4.934297673925714, "grad_norm": 6.6017937660217285, "learning_rate": 4.516142192026451e-08, "loss": 2.0332, "step": 558750 }, { "epoch": 4.934739221815998, "grad_norm": 5.2641072273254395, "learning_rate": 4.45728631586928e-08, "loss": 2.7379, "step": 558800 }, { "epoch": 4.935180769706283, "grad_norm": 3.389509916305542, "learning_rate": 4.398816303025633e-08, "loss": 2.4802, "step": 558850 }, { "epoch": 4.935622317596566, "grad_norm": 8.590868949890137, "learning_rate": 4.340732158012451e-08, "loss": 2.8343, "step": 558900 }, { "epoch": 4.936063865486851, "grad_norm": 3.883772850036621, "learning_rate": 4.2830338853155906e-08, "loss": 2.5019, "step": 558950 }, { "epoch": 4.936505413377135, "grad_norm": 12.661687850952148, "learning_rate": 4.225721489391488e-08, "loss": 2.9332, "step": 559000 }, { "epoch": 4.936946961267419, "grad_norm": 7.978334426879883, "learning_rate": 4.168794974668267e-08, "loss": 2.2468, "step": 559050 }, { "epoch": 4.937388509157703, "grad_norm": 9.38732624053955, "learning_rate": 4.112254345541855e-08, "loss": 2.7731, "step": 559100 }, { "epoch": 4.937830057047988, "grad_norm": 5.196143627166748, "learning_rate": 4.0560996063793155e-08, "loss": 2.9647, "step": 559150 }, { "epoch": 4.938271604938271, "grad_norm": 9.177465438842773, "learning_rate": 4.0003307615193994e-08, "loss": 2.3986, "step": 559200 }, { "epoch": 4.938713152828556, "grad_norm": 1.2352168560028076, "learning_rate": 3.944947815269218e-08, "loss": 2.75, "step": 559250 }, { "epoch": 4.93915470071884, "grad_norm": 6.418938636779785, "learning_rate": 3.8899507719059037e-08, "loss": 2.8239, "step": 559300 }, { "epoch": 4.939596248609124, "grad_norm": 5.894453525543213, "learning_rate": 3.835339635678281e-08, "loss": 2.7885, "step": 559350 }, { "epoch": 4.940037796499408, "grad_norm": 3.0357918739318848, "learning_rate": 3.7811144108046424e-08, "loss": 2.3254, "step": 559400 }, { "epoch": 4.940479344389693, "grad_norm": 6.236439228057861, "learning_rate": 3.727275101473304e-08, "loss": 2.4, "step": 559450 }, { "epoch": 4.940920892279976, "grad_norm": 10.476075172424316, "learning_rate": 3.6738217118426066e-08, "loss": 2.2863, "step": 559500 }, { "epoch": 4.941362440170261, "grad_norm": 6.477456092834473, "learning_rate": 3.6207542460414687e-08, "loss": 2.7779, "step": 559550 }, { "epoch": 4.941803988060545, "grad_norm": 3.965860605239868, "learning_rate": 3.56807270816939e-08, "loss": 2.5292, "step": 559600 }, { "epoch": 4.942245535950829, "grad_norm": 6.7968902587890625, "learning_rate": 3.5157771022947816e-08, "loss": 2.817, "step": 559650 }, { "epoch": 4.942687083841113, "grad_norm": 3.9870734214782715, "learning_rate": 3.463867432457746e-08, "loss": 2.6096, "step": 559700 }, { "epoch": 4.943128631731398, "grad_norm": 4.815114498138428, "learning_rate": 3.412343702667298e-08, "loss": 2.8563, "step": 559750 }, { "epoch": 4.943570179621682, "grad_norm": 4.599194049835205, "learning_rate": 3.361205916903587e-08, "loss": 2.788, "step": 559800 }, { "epoch": 4.944011727511966, "grad_norm": 7.611653804779053, "learning_rate": 3.310454079116787e-08, "loss": 3.1564, "step": 559850 }, { "epoch": 4.94445327540225, "grad_norm": 6.627467155456543, "learning_rate": 3.2600881932265406e-08, "loss": 2.5019, "step": 559900 }, { "epoch": 4.944894823292534, "grad_norm": 9.303733825683594, "learning_rate": 3.210108263123068e-08, "loss": 3.2121, "step": 559950 }, { "epoch": 4.945336371182819, "grad_norm": 3.9140212535858154, "learning_rate": 3.1605142926671716e-08, "loss": 2.7299, "step": 560000 }, { "epoch": 4.945777919073103, "grad_norm": 6.570229530334473, "learning_rate": 3.11130628569023e-08, "loss": 2.4387, "step": 560050 }, { "epoch": 4.946219466963387, "grad_norm": 5.366155624389648, "learning_rate": 3.062484245991981e-08, "loss": 2.8858, "step": 560100 }, { "epoch": 4.946661014853671, "grad_norm": 4.563048839569092, "learning_rate": 3.014048177343853e-08, "loss": 2.8534, "step": 560150 }, { "epoch": 4.9471025627439555, "grad_norm": 7.997015953063965, "learning_rate": 2.965998083487853e-08, "loss": 2.3602, "step": 560200 }, { "epoch": 4.947544110634239, "grad_norm": 3.5918991565704346, "learning_rate": 2.918333968134346e-08, "loss": 2.8457, "step": 560250 }, { "epoch": 4.947985658524524, "grad_norm": 6.0873236656188965, "learning_rate": 2.8710558349659412e-08, "loss": 3.2055, "step": 560300 }, { "epoch": 4.948427206414808, "grad_norm": 6.233656406402588, "learning_rate": 2.8241636876336074e-08, "loss": 2.69, "step": 560350 }, { "epoch": 4.948868754305092, "grad_norm": 8.256710052490234, "learning_rate": 2.7776575297600028e-08, "loss": 2.5685, "step": 560400 }, { "epoch": 4.949310302195376, "grad_norm": 6.945454120635986, "learning_rate": 2.7315373649372535e-08, "loss": 2.9995, "step": 560450 }, { "epoch": 4.9497518500856605, "grad_norm": 6.325427532196045, "learning_rate": 2.68580319672751e-08, "loss": 2.5202, "step": 560500 }, { "epoch": 4.950193397975944, "grad_norm": 4.097993850708008, "learning_rate": 2.640455028664057e-08, "loss": 2.8188, "step": 560550 }, { "epoch": 4.950634945866229, "grad_norm": 7.394988536834717, "learning_rate": 2.5954928642485386e-08, "loss": 2.8986, "step": 560600 }, { "epoch": 4.951076493756513, "grad_norm": 6.201233863830566, "learning_rate": 2.550916706954287e-08, "loss": 2.7394, "step": 560650 }, { "epoch": 4.951518041646797, "grad_norm": 7.919137954711914, "learning_rate": 2.5067265602252142e-08, "loss": 2.753, "step": 560700 }, { "epoch": 4.951959589537081, "grad_norm": 5.697253704071045, "learning_rate": 2.4629224274735908e-08, "loss": 2.6731, "step": 560750 }, { "epoch": 4.9524011374273655, "grad_norm": 9.427850723266602, "learning_rate": 2.419504312083376e-08, "loss": 2.8647, "step": 560800 }, { "epoch": 4.952842685317649, "grad_norm": 4.388767719268799, "learning_rate": 2.376472217407999e-08, "loss": 2.5204, "step": 560850 }, { "epoch": 4.953284233207934, "grad_norm": 4.339812755584717, "learning_rate": 2.3338261467720225e-08, "loss": 2.7001, "step": 560900 }, { "epoch": 4.953725781098218, "grad_norm": 6.240818977355957, "learning_rate": 2.291566103468368e-08, "loss": 2.6848, "step": 560950 }, { "epoch": 4.954167328988502, "grad_norm": 5.748239040374756, "learning_rate": 2.249692090762201e-08, "loss": 2.537, "step": 561000 }, { "epoch": 4.954167328988502, "eval_asr_loss": 0.9944855403858892, "eval_loss": 2.6887645721435547, "eval_runtime": 20.2174, "eval_samples_per_second": 37.987, "eval_steps_per_second": 9.497, "eval_tts_loss": 6.3390417155386745, "step": 561000 }, { "epoch": 4.954608876878786, "grad_norm": 5.837003707885742, "learning_rate": 2.208204111888157e-08, "loss": 2.2392, "step": 561050 }, { "epoch": 4.9550504247690705, "grad_norm": 4.444454193115234, "learning_rate": 2.167102170049784e-08, "loss": 2.7514, "step": 561100 }, { "epoch": 4.955491972659354, "grad_norm": 5.578451156616211, "learning_rate": 2.1263862684223203e-08, "loss": 3.0353, "step": 561150 }, { "epoch": 4.955933520549639, "grad_norm": 5.00250768661499, "learning_rate": 2.0860564101510272e-08, "loss": 2.6323, "step": 561200 }, { "epoch": 4.9563750684399235, "grad_norm": 5.400811195373535, "learning_rate": 2.0461125983506357e-08, "loss": 2.9044, "step": 561250 }, { "epoch": 4.956816616330207, "grad_norm": 5.953423500061035, "learning_rate": 2.0065548361064556e-08, "loss": 2.7176, "step": 561300 }, { "epoch": 4.957258164220491, "grad_norm": 9.428725242614746, "learning_rate": 1.9673831264743758e-08, "loss": 2.4835, "step": 561350 }, { "epoch": 4.9576997121107755, "grad_norm": 7.238251686096191, "learning_rate": 1.928597472479754e-08, "loss": 2.295, "step": 561400 }, { "epoch": 4.95814126000106, "grad_norm": 2.0047836303710938, "learning_rate": 1.890197877119082e-08, "loss": 2.4707, "step": 561450 }, { "epoch": 4.958582807891344, "grad_norm": 7.1950178146362305, "learning_rate": 1.8521843433572106e-08, "loss": 2.955, "step": 561500 }, { "epoch": 4.959024355781628, "grad_norm": 14.725709915161133, "learning_rate": 1.81455687413179e-08, "loss": 2.7663, "step": 561550 }, { "epoch": 4.959465903671912, "grad_norm": 4.209861755371094, "learning_rate": 1.777315472347718e-08, "loss": 2.6205, "step": 561600 }, { "epoch": 4.959907451562197, "grad_norm": 5.621754169464111, "learning_rate": 1.7404601408826938e-08, "loss": 2.808, "step": 561650 }, { "epoch": 4.9603489994524805, "grad_norm": 4.717309951782227, "learning_rate": 1.7039908825833284e-08, "loss": 2.5536, "step": 561700 }, { "epoch": 4.960790547342765, "grad_norm": 6.841756820678711, "learning_rate": 1.6679077002657028e-08, "loss": 2.6873, "step": 561750 }, { "epoch": 4.961232095233049, "grad_norm": 2.4917197227478027, "learning_rate": 1.6322105967181423e-08, "loss": 2.7838, "step": 561800 }, { "epoch": 4.9616736431233335, "grad_norm": 3.389162063598633, "learning_rate": 1.5968995746973303e-08, "loss": 2.3744, "step": 561850 }, { "epoch": 4.962115191013617, "grad_norm": 4.882816791534424, "learning_rate": 1.5619746369305298e-08, "loss": 2.5339, "step": 561900 }, { "epoch": 4.962556738903902, "grad_norm": 7.238387107849121, "learning_rate": 1.527435786116138e-08, "loss": 2.2508, "step": 561950 }, { "epoch": 4.9629982867941855, "grad_norm": 3.73600697517395, "learning_rate": 1.4932830249209108e-08, "loss": 2.799, "step": 562000 }, { "epoch": 4.96343983468447, "grad_norm": 1.8746936321258545, "learning_rate": 1.4595163559838476e-08, "loss": 2.5706, "step": 562050 }, { "epoch": 4.963881382574754, "grad_norm": 9.808050155639648, "learning_rate": 1.4261357819128629e-08, "loss": 2.6851, "step": 562100 }, { "epoch": 4.9643229304650385, "grad_norm": 8.384523391723633, "learning_rate": 1.3931413052858944e-08, "loss": 2.5717, "step": 562150 }, { "epoch": 4.964764478355322, "grad_norm": 8.043850898742676, "learning_rate": 1.3611813123887151e-08, "loss": 2.9255, "step": 562200 }, { "epoch": 4.965206026245607, "grad_norm": 9.861638069152832, "learning_rate": 1.3289513161918177e-08, "loss": 2.665, "step": 562250 }, { "epoch": 4.9656475741358905, "grad_norm": 3.2550222873687744, "learning_rate": 1.2971074249457094e-08, "loss": 2.9545, "step": 562300 }, { "epoch": 4.966089122026175, "grad_norm": 4.106522560119629, "learning_rate": 1.2656496411106444e-08, "loss": 2.658, "step": 562350 }, { "epoch": 4.966530669916459, "grad_norm": 3.5041277408599854, "learning_rate": 1.2345779671157909e-08, "loss": 2.5857, "step": 562400 }, { "epoch": 4.9669722178067435, "grad_norm": 6.466332912445068, "learning_rate": 1.2038924053614509e-08, "loss": 2.6008, "step": 562450 }, { "epoch": 4.967413765697027, "grad_norm": 1.7912003993988037, "learning_rate": 1.1735929582179506e-08, "loss": 2.9858, "step": 562500 }, { "epoch": 4.967855313587312, "grad_norm": 6.125924587249756, "learning_rate": 1.1436796280256401e-08, "loss": 2.8469, "step": 562550 }, { "epoch": 4.9682968614775955, "grad_norm": 5.703486919403076, "learning_rate": 1.1141524170948936e-08, "loss": 2.6328, "step": 562600 }, { "epoch": 4.96873840936788, "grad_norm": 5.380309104919434, "learning_rate": 1.0850113277066642e-08, "loss": 2.7271, "step": 562650 }, { "epoch": 4.969179957258164, "grad_norm": 7.205440044403076, "learning_rate": 1.0562563621119292e-08, "loss": 2.6379, "step": 562700 }, { "epoch": 4.9696215051484485, "grad_norm": 5.821721076965332, "learning_rate": 1.0278875225316897e-08, "loss": 2.9129, "step": 562750 }, { "epoch": 4.970063053038732, "grad_norm": 8.173575401306152, "learning_rate": 9.999048111564158e-09, "loss": 2.046, "step": 562800 }, { "epoch": 4.970504600929017, "grad_norm": 3.5264334678649902, "learning_rate": 9.723082301488217e-09, "loss": 2.8035, "step": 562850 }, { "epoch": 4.970946148819301, "grad_norm": 6.837828636169434, "learning_rate": 9.450977816394257e-09, "loss": 2.2991, "step": 562900 }, { "epoch": 4.971387696709585, "grad_norm": 4.562282085418701, "learning_rate": 9.182734677309901e-09, "loss": 2.7958, "step": 562950 }, { "epoch": 4.971829244599869, "grad_norm": 5.1963982582092285, "learning_rate": 8.918352904946359e-09, "loss": 2.7717, "step": 563000 }, { "epoch": 4.9722707924901535, "grad_norm": 2.7982614040374756, "learning_rate": 8.657832519731735e-09, "loss": 2.458, "step": 563050 }, { "epoch": 4.972712340380438, "grad_norm": 13.411173820495605, "learning_rate": 8.406268879418422e-09, "loss": 2.7716, "step": 563100 }, { "epoch": 4.973153888270722, "grad_norm": 3.3968658447265625, "learning_rate": 8.15339409983018e-09, "loss": 2.929, "step": 563150 }, { "epoch": 4.973595436161006, "grad_norm": 13.799068450927734, "learning_rate": 7.904380766471375e-09, "loss": 2.3196, "step": 563200 }, { "epoch": 4.97403698405129, "grad_norm": 6.284460544586182, "learning_rate": 7.65922889857662e-09, "loss": 2.544, "step": 563250 }, { "epoch": 4.974478531941575, "grad_norm": 6.938006401062012, "learning_rate": 7.41793851508632e-09, "loss": 2.986, "step": 563300 }, { "epoch": 4.9749200798318585, "grad_norm": 5.39680290222168, "learning_rate": 7.180509634635568e-09, "loss": 2.8663, "step": 563350 }, { "epoch": 4.975361627722143, "grad_norm": 3.692669630050659, "learning_rate": 6.946942275565249e-09, "loss": 2.429, "step": 563400 }, { "epoch": 4.975803175612427, "grad_norm": 5.349528789520264, "learning_rate": 6.717236455910936e-09, "loss": 2.4808, "step": 563450 }, { "epoch": 4.976244723502711, "grad_norm": 5.677651882171631, "learning_rate": 6.491392193425095e-09, "loss": 2.604, "step": 563500 }, { "epoch": 4.976686271392995, "grad_norm": 6.453908920288086, "learning_rate": 6.269409505538226e-09, "loss": 2.8354, "step": 563550 }, { "epoch": 4.97712781928328, "grad_norm": 6.007885932922363, "learning_rate": 6.05128840941438e-09, "loss": 2.2361, "step": 563600 }, { "epoch": 4.9775693671735635, "grad_norm": 3.3548996448516846, "learning_rate": 5.837028921884535e-09, "loss": 2.6972, "step": 563650 }, { "epoch": 4.978010915063848, "grad_norm": 7.547037601470947, "learning_rate": 5.62663105950767e-09, "loss": 2.6719, "step": 563700 }, { "epoch": 4.978452462954132, "grad_norm": 4.264461040496826, "learning_rate": 5.420094838537448e-09, "loss": 2.7277, "step": 563750 }, { "epoch": 4.978894010844416, "grad_norm": 6.38759708404541, "learning_rate": 5.217420274922224e-09, "loss": 2.7116, "step": 563800 }, { "epoch": 4.9793355587347, "grad_norm": 7.815456867218018, "learning_rate": 5.018607384316143e-09, "loss": 2.8592, "step": 563850 }, { "epoch": 4.979777106624985, "grad_norm": 7.064630031585693, "learning_rate": 4.823656182079139e-09, "loss": 2.6353, "step": 563900 }, { "epoch": 4.9802186545152685, "grad_norm": 4.916224956512451, "learning_rate": 4.632566683265838e-09, "loss": 2.8035, "step": 563950 }, { "epoch": 4.980660202405553, "grad_norm": 3.6277050971984863, "learning_rate": 4.445338902642204e-09, "loss": 2.1228, "step": 564000 }, { "epoch": 4.980660202405553, "eval_asr_loss": 0.9945437661349208, "eval_loss": 2.6887919902801514, "eval_runtime": 20.525, "eval_samples_per_second": 37.418, "eval_steps_per_second": 9.354, "eval_tts_loss": 6.3396508273187635, "step": 564000 }, { "epoch": 4.981101750295837, "grad_norm": 2.2545013427734375, "learning_rate": 4.261972854663343e-09, "loss": 2.5292, "step": 564050 }, { "epoch": 4.981543298186121, "grad_norm": 3.7255971431732178, "learning_rate": 4.082468553495699e-09, "loss": 2.9966, "step": 564100 }, { "epoch": 4.981984846076405, "grad_norm": 4.366975784301758, "learning_rate": 3.906826013000409e-09, "loss": 2.8084, "step": 564150 }, { "epoch": 4.98242639396669, "grad_norm": 15.585683822631836, "learning_rate": 3.735045246755498e-09, "loss": 2.6431, "step": 564200 }, { "epoch": 4.9828679418569735, "grad_norm": 6.45513916015625, "learning_rate": 3.5671262680170293e-09, "loss": 2.7329, "step": 564250 }, { "epoch": 4.983309489747258, "grad_norm": 2.445129871368408, "learning_rate": 3.403069089763511e-09, "loss": 3.17, "step": 564300 }, { "epoch": 4.983751037637542, "grad_norm": 3.703029155731201, "learning_rate": 3.242873724662587e-09, "loss": 3.0606, "step": 564350 }, { "epoch": 4.984192585527826, "grad_norm": 4.673495292663574, "learning_rate": 3.086540185093245e-09, "loss": 2.7839, "step": 564400 }, { "epoch": 4.98463413341811, "grad_norm": 4.855280876159668, "learning_rate": 2.9340684831236088e-09, "loss": 2.5286, "step": 564450 }, { "epoch": 4.985075681308395, "grad_norm": 4.498984336853027, "learning_rate": 2.7854586305331443e-09, "loss": 2.8139, "step": 564500 }, { "epoch": 4.985517229198679, "grad_norm": 5.892887115478516, "learning_rate": 2.640710638807109e-09, "loss": 2.6435, "step": 564550 }, { "epoch": 4.985958777088963, "grad_norm": 8.301871299743652, "learning_rate": 2.4998245191198976e-09, "loss": 2.2924, "step": 564600 }, { "epoch": 4.986400324979247, "grad_norm": 8.63991641998291, "learning_rate": 2.3628002823516957e-09, "loss": 2.6399, "step": 564650 }, { "epoch": 4.986841872869531, "grad_norm": 2.6426339149475098, "learning_rate": 2.22963793908848e-09, "loss": 2.4955, "step": 564700 }, { "epoch": 4.987283420759816, "grad_norm": 5.141458034515381, "learning_rate": 2.100337499622018e-09, "loss": 2.6246, "step": 564750 }, { "epoch": 4.9877249686501, "grad_norm": 6.250972270965576, "learning_rate": 1.9748989739332145e-09, "loss": 2.8287, "step": 564800 }, { "epoch": 4.988166516540384, "grad_norm": 3.701958656311035, "learning_rate": 1.8533223717087655e-09, "loss": 2.6992, "step": 564850 }, { "epoch": 4.988608064430668, "grad_norm": 3.7596817016601562, "learning_rate": 1.7356077023467088e-09, "loss": 2.5364, "step": 564900 }, { "epoch": 4.989049612320953, "grad_norm": 5.01679801940918, "learning_rate": 1.6217549749342197e-09, "loss": 2.813, "step": 564950 }, { "epoch": 4.989491160211236, "grad_norm": 4.733850955963135, "learning_rate": 1.5117641982698161e-09, "loss": 2.572, "step": 565000 }, { "epoch": 4.989932708101521, "grad_norm": 7.652512073516846, "learning_rate": 1.4056353808467037e-09, "loss": 2.8103, "step": 565050 }, { "epoch": 4.990374255991805, "grad_norm": 5.089700698852539, "learning_rate": 1.3033685308583288e-09, "loss": 2.8409, "step": 565100 }, { "epoch": 4.990815803882089, "grad_norm": 8.621271133422852, "learning_rate": 1.20496365621503e-09, "loss": 2.2292, "step": 565150 }, { "epoch": 4.991257351772373, "grad_norm": 5.350403785705566, "learning_rate": 1.1104207645051822e-09, "loss": 2.8795, "step": 565200 }, { "epoch": 4.991698899662658, "grad_norm": 5.254849433898926, "learning_rate": 1.0197398630396038e-09, "loss": 2.5642, "step": 565250 }, { "epoch": 4.992140447552941, "grad_norm": 8.5669584274292, "learning_rate": 9.329209588182508e-10, "loss": 2.3064, "step": 565300 }, { "epoch": 4.992581995443226, "grad_norm": 5.058325290679932, "learning_rate": 8.499640585524215e-10, "loss": 2.7282, "step": 565350 }, { "epoch": 4.99302354333351, "grad_norm": 3.100698947906494, "learning_rate": 7.724132187025834e-10, "loss": 2.7971, "step": 565400 }, { "epoch": 4.993465091223794, "grad_norm": 8.421867370605469, "learning_rate": 6.971031048741417e-10, "loss": 2.7171, "step": 565450 }, { "epoch": 4.993906639114078, "grad_norm": 8.118165016174316, "learning_rate": 6.256550132144323e-10, "loss": 2.5733, "step": 565500 }, { "epoch": 4.994348187004363, "grad_norm": 9.683393478393555, "learning_rate": 5.580689492412638e-10, "loss": 2.8082, "step": 565550 }, { "epoch": 4.994789734894646, "grad_norm": 1.7594788074493408, "learning_rate": 4.943449181782356e-10, "loss": 2.5409, "step": 565600 }, { "epoch": 4.995231282784931, "grad_norm": 6.413368225097656, "learning_rate": 4.344829249380844e-10, "loss": 2.7523, "step": 565650 }, { "epoch": 4.995672830675215, "grad_norm": 12.606328010559082, "learning_rate": 3.784829741504403e-10, "loss": 2.3551, "step": 565700 }, { "epoch": 4.996114378565499, "grad_norm": 7.66225528717041, "learning_rate": 3.263450701451731e-10, "loss": 2.6103, "step": 565750 }, { "epoch": 4.996555926455783, "grad_norm": 4.763168811798096, "learning_rate": 2.780692169412902e-10, "loss": 2.3541, "step": 565800 }, { "epoch": 4.996997474346068, "grad_norm": 10.193942070007324, "learning_rate": 2.336554182691408e-10, "loss": 2.8816, "step": 565850 }, { "epoch": 4.997439022236351, "grad_norm": 4.266788959503174, "learning_rate": 1.9310367756486536e-10, "loss": 2.5406, "step": 565900 }, { "epoch": 4.997880570126636, "grad_norm": 5.01970911026001, "learning_rate": 1.5641399795929268e-10, "loss": 2.3929, "step": 565950 }, { "epoch": 4.998322118016921, "grad_norm": 7.360198020935059, "learning_rate": 1.235863822779404e-10, "loss": 2.6288, "step": 566000 }, { "epoch": 4.998763665907204, "grad_norm": 4.490750789642334, "learning_rate": 9.462083306877034e-11, "loss": 2.3701, "step": 566050 }, { "epoch": 4.999205213797488, "grad_norm": 6.801180839538574, "learning_rate": 6.951735256333081e-11, "loss": 3.0652, "step": 566100 }, { "epoch": 4.999646761687773, "grad_norm": 4.100923538208008, "learning_rate": 4.8275942698960965e-11, "loss": 2.2844, "step": 566150 } ], "logging_steps": 50, "max_steps": 566190, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 3000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 5.388690186584457e+18, "train_batch_size": 4, "trial_name": null, "trial_params": null }