{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.9272505695967784, "eval_steps": 3000, "global_step": 105000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00044154789028418023, "grad_norm": 4.407685279846191, "learning_rate": 4.800000000000001e-06, "loss": 4.2565, "step": 50 }, { "epoch": 0.0008830957805683605, "grad_norm": 2.3940842151641846, "learning_rate": 9.800000000000001e-06, "loss": 4.3171, "step": 100 }, { "epoch": 0.0013246436708525408, "grad_norm": 2.555833339691162, "learning_rate": 1.48e-05, "loss": 3.8795, "step": 150 }, { "epoch": 0.001766191561136721, "grad_norm": 0.9911451935768127, "learning_rate": 1.97e-05, "loss": 3.9588, "step": 200 }, { "epoch": 0.0022077394514209013, "grad_norm": 1.1606439352035522, "learning_rate": 2.47e-05, "loss": 3.4062, "step": 250 }, { "epoch": 0.0026492873417050815, "grad_norm": 2.9307570457458496, "learning_rate": 2.97e-05, "loss": 3.5316, "step": 300 }, { "epoch": 0.0030908352319892617, "grad_norm": 2.6516165733337402, "learning_rate": 3.4699999999999996e-05, "loss": 3.5324, "step": 350 }, { "epoch": 0.003532383122273442, "grad_norm": 1.69759202003479, "learning_rate": 3.97e-05, "loss": 3.6864, "step": 400 }, { "epoch": 0.003973931012557622, "grad_norm": 3.0430002212524414, "learning_rate": 4.47e-05, "loss": 3.6768, "step": 450 }, { "epoch": 0.004415478902841803, "grad_norm": 4.26913595199585, "learning_rate": 4.97e-05, "loss": 3.9357, "step": 500 }, { "epoch": 0.004857026793125983, "grad_norm": 1.9752435684204102, "learning_rate": 5.470000000000001e-05, "loss": 3.4989, "step": 550 }, { "epoch": 0.005298574683410163, "grad_norm": 2.631711006164551, "learning_rate": 5.96e-05, "loss": 3.4577, "step": 600 }, { "epoch": 0.005740122573694343, "grad_norm": 4.840415000915527, "learning_rate": 6.460000000000001e-05, "loss": 3.6496, "step": 650 }, { "epoch": 0.006181670463978523, "grad_norm": 0.7839917540550232, "learning_rate": 6.96e-05, "loss": 3.4212, "step": 700 }, { "epoch": 0.0066232183542627035, "grad_norm": 2.6035537719726562, "learning_rate": 7.46e-05, "loss": 3.5461, "step": 750 }, { "epoch": 0.007064766244546884, "grad_norm": 2.7785916328430176, "learning_rate": 7.960000000000001e-05, "loss": 3.5111, "step": 800 }, { "epoch": 0.007506314134831064, "grad_norm": 3.2137913703918457, "learning_rate": 8.46e-05, "loss": 3.4488, "step": 850 }, { "epoch": 0.007947862025115243, "grad_norm": 2.4404215812683105, "learning_rate": 8.960000000000001e-05, "loss": 3.6674, "step": 900 }, { "epoch": 0.008389409915399423, "grad_norm": 2.055772542953491, "learning_rate": 9.46e-05, "loss": 3.5125, "step": 950 }, { "epoch": 0.008830957805683605, "grad_norm": 6.80244255065918, "learning_rate": 9.960000000000001e-05, "loss": 3.7395, "step": 1000 }, { "epoch": 0.009272505695967785, "grad_norm": 1.6064447164535522, "learning_rate": 9.999999836556943e-05, "loss": 3.2688, "step": 1050 }, { "epoch": 0.009714053586251966, "grad_norm": 1.4385908842086792, "learning_rate": 9.999999288142163e-05, "loss": 3.7015, "step": 1100 }, { "epoch": 0.010155601476536146, "grad_norm": 2.0870840549468994, "learning_rate": 9.999998353519836e-05, "loss": 3.6805, "step": 1150 }, { "epoch": 0.010597149366820326, "grad_norm": 1.5181084871292114, "learning_rate": 9.999997032690033e-05, "loss": 3.9943, "step": 1200 }, { "epoch": 0.011038697257104506, "grad_norm": 2.0641415119171143, "learning_rate": 9.99999532565286e-05, "loss": 3.4652, "step": 1250 }, { "epoch": 0.011480245147388686, "grad_norm": 4.051513671875, "learning_rate": 9.999993232408446e-05, "loss": 3.784, "step": 1300 }, { "epoch": 0.011921793037672867, "grad_norm": 2.475578546524048, "learning_rate": 9.999990752956954e-05, "loss": 3.6213, "step": 1350 }, { "epoch": 0.012363340927957047, "grad_norm": 2.650278091430664, "learning_rate": 9.999987887298574e-05, "loss": 3.2759, "step": 1400 }, { "epoch": 0.012804888818241227, "grad_norm": 1.043623924255371, "learning_rate": 9.999984635433529e-05, "loss": 3.4098, "step": 1450 }, { "epoch": 0.013246436708525407, "grad_norm": 2.3483691215515137, "learning_rate": 9.999980997362069e-05, "loss": 3.5984, "step": 1500 }, { "epoch": 0.013687984598809587, "grad_norm": 1.8058457374572754, "learning_rate": 9.999976973084477e-05, "loss": 3.5686, "step": 1550 }, { "epoch": 0.014129532489093767, "grad_norm": 1.9905356168746948, "learning_rate": 9.999972562601062e-05, "loss": 3.5255, "step": 1600 }, { "epoch": 0.014571080379377948, "grad_norm": 8.096346855163574, "learning_rate": 9.999967765912164e-05, "loss": 3.8234, "step": 1650 }, { "epoch": 0.015012628269662128, "grad_norm": 3.1411712169647217, "learning_rate": 9.999962583018154e-05, "loss": 3.3328, "step": 1700 }, { "epoch": 0.015454176159946308, "grad_norm": 4.46273946762085, "learning_rate": 9.999957013919435e-05, "loss": 3.2951, "step": 1750 }, { "epoch": 0.015895724050230486, "grad_norm": 1.3609850406646729, "learning_rate": 9.999951058616435e-05, "loss": 3.3173, "step": 1800 }, { "epoch": 0.016337271940514667, "grad_norm": 0.8974264860153198, "learning_rate": 9.999944717109613e-05, "loss": 3.5905, "step": 1850 }, { "epoch": 0.016778819830798847, "grad_norm": 4.310591220855713, "learning_rate": 9.999937989399459e-05, "loss": 3.2723, "step": 1900 }, { "epoch": 0.017220367721083027, "grad_norm": 10.745691299438477, "learning_rate": 9.999930875486495e-05, "loss": 3.3104, "step": 1950 }, { "epoch": 0.01766191561136721, "grad_norm": 1.678096890449524, "learning_rate": 9.999923375371269e-05, "loss": 3.4128, "step": 2000 }, { "epoch": 0.01810346350165139, "grad_norm": 1.2710167169570923, "learning_rate": 9.99991548905436e-05, "loss": 3.8629, "step": 2050 }, { "epoch": 0.01854501139193557, "grad_norm": 1.6546989679336548, "learning_rate": 9.999907216536378e-05, "loss": 3.9797, "step": 2100 }, { "epoch": 0.01898655928221975, "grad_norm": 2.4010069370269775, "learning_rate": 9.999898557817962e-05, "loss": 3.4582, "step": 2150 }, { "epoch": 0.01942810717250393, "grad_norm": 11.584942817687988, "learning_rate": 9.999889512899778e-05, "loss": 3.4392, "step": 2200 }, { "epoch": 0.01986965506278811, "grad_norm": 2.588017702102661, "learning_rate": 9.999880081782529e-05, "loss": 3.2102, "step": 2250 }, { "epoch": 0.02031120295307229, "grad_norm": 4.959451198577881, "learning_rate": 9.99987026446694e-05, "loss": 3.4403, "step": 2300 }, { "epoch": 0.020752750843356472, "grad_norm": 2.886789560317993, "learning_rate": 9.999860060953772e-05, "loss": 3.3001, "step": 2350 }, { "epoch": 0.021194298733640652, "grad_norm": 7.682253837585449, "learning_rate": 9.999849471243812e-05, "loss": 3.8255, "step": 2400 }, { "epoch": 0.021635846623924832, "grad_norm": 9.694388389587402, "learning_rate": 9.999838495337877e-05, "loss": 3.8851, "step": 2450 }, { "epoch": 0.022077394514209012, "grad_norm": 2.9934961795806885, "learning_rate": 9.999827133236815e-05, "loss": 3.6424, "step": 2500 }, { "epoch": 0.022518942404493193, "grad_norm": 1.4004652500152588, "learning_rate": 9.999815384941506e-05, "loss": 3.5799, "step": 2550 }, { "epoch": 0.022960490294777373, "grad_norm": 2.06247615814209, "learning_rate": 9.999803250452856e-05, "loss": 3.4358, "step": 2600 }, { "epoch": 0.023402038185061553, "grad_norm": 2.051114082336426, "learning_rate": 9.999790729771798e-05, "loss": 3.6721, "step": 2650 }, { "epoch": 0.023843586075345733, "grad_norm": 2.4515838623046875, "learning_rate": 9.999777822899307e-05, "loss": 3.6419, "step": 2700 }, { "epoch": 0.024285133965629913, "grad_norm": 2.6057190895080566, "learning_rate": 9.999764529836375e-05, "loss": 3.4357, "step": 2750 }, { "epoch": 0.024726681855914093, "grad_norm": 2.1552577018737793, "learning_rate": 9.99975085058403e-05, "loss": 3.6189, "step": 2800 }, { "epoch": 0.025168229746198274, "grad_norm": 4.0472846031188965, "learning_rate": 9.999736785143327e-05, "loss": 3.5135, "step": 2850 }, { "epoch": 0.025609777636482454, "grad_norm": 1.694504976272583, "learning_rate": 9.999722333515355e-05, "loss": 3.5287, "step": 2900 }, { "epoch": 0.026051325526766634, "grad_norm": 8.814433097839355, "learning_rate": 9.99970749570123e-05, "loss": 3.4025, "step": 2950 }, { "epoch": 0.026492873417050814, "grad_norm": 2.079963207244873, "learning_rate": 9.999692579966685e-05, "loss": 3.8211, "step": 3000 }, { "epoch": 0.026492873417050814, "eval_asr_loss": 0.8104195239432453, "eval_loss": 3.4588043689727783, "eval_runtime": 20.2586, "eval_samples_per_second": 37.91, "eval_steps_per_second": 9.477, "eval_tts_loss": 6.061187094347814, "step": 3000 }, { "epoch": 0.026934421307334994, "grad_norm": 2.6620876789093018, "learning_rate": 9.999676977507384e-05, "loss": 3.4284, "step": 3050 }, { "epoch": 0.027375969197619174, "grad_norm": 10.126931190490723, "learning_rate": 9.999660988865433e-05, "loss": 3.756, "step": 3100 }, { "epoch": 0.027817517087903355, "grad_norm": 1.450592279434204, "learning_rate": 9.999644614042068e-05, "loss": 3.2147, "step": 3150 }, { "epoch": 0.028259064978187535, "grad_norm": 1.703476071357727, "learning_rate": 9.99962785303855e-05, "loss": 3.5119, "step": 3200 }, { "epoch": 0.028700612868471715, "grad_norm": 2.767446756362915, "learning_rate": 9.999610705856177e-05, "loss": 3.2736, "step": 3250 }, { "epoch": 0.029142160758755895, "grad_norm": 1.1504226922988892, "learning_rate": 9.999593172496273e-05, "loss": 3.4068, "step": 3300 }, { "epoch": 0.029583708649040075, "grad_norm": 1.6964422464370728, "learning_rate": 9.99957525296019e-05, "loss": 3.4917, "step": 3350 }, { "epoch": 0.030025256539324256, "grad_norm": 1.8230127096176147, "learning_rate": 9.999556947249316e-05, "loss": 3.1574, "step": 3400 }, { "epoch": 0.030466804429608436, "grad_norm": 3.8012404441833496, "learning_rate": 9.999538255365062e-05, "loss": 3.6179, "step": 3450 }, { "epoch": 0.030908352319892616, "grad_norm": 3.4073257446289062, "learning_rate": 9.999519177308871e-05, "loss": 3.6502, "step": 3500 }, { "epoch": 0.03134990021017679, "grad_norm": 1.7333426475524902, "learning_rate": 9.99949971308222e-05, "loss": 3.5175, "step": 3550 }, { "epoch": 0.03179144810046097, "grad_norm": 1.5845162868499756, "learning_rate": 9.99947986268661e-05, "loss": 3.4373, "step": 3600 }, { "epoch": 0.03223299599074515, "grad_norm": 4.4948225021362305, "learning_rate": 9.999459626123576e-05, "loss": 3.6571, "step": 3650 }, { "epoch": 0.03267454388102933, "grad_norm": 3.1738150119781494, "learning_rate": 9.99943900339468e-05, "loss": 3.6251, "step": 3700 }, { "epoch": 0.03311609177131351, "grad_norm": 2.0190846920013428, "learning_rate": 9.999417994501514e-05, "loss": 3.5417, "step": 3750 }, { "epoch": 0.033557639661597694, "grad_norm": 3.240422248840332, "learning_rate": 9.999396599445701e-05, "loss": 3.5891, "step": 3800 }, { "epoch": 0.033999187551881874, "grad_norm": 5.438379287719727, "learning_rate": 9.999374818228895e-05, "loss": 3.6557, "step": 3850 }, { "epoch": 0.034440735442166054, "grad_norm": 1.6964021921157837, "learning_rate": 9.999352650852778e-05, "loss": 3.4464, "step": 3900 }, { "epoch": 0.03488228333245024, "grad_norm": 1.759133219718933, "learning_rate": 9.999330097319063e-05, "loss": 3.5975, "step": 3950 }, { "epoch": 0.03532383122273442, "grad_norm": 1.7858753204345703, "learning_rate": 9.999307157629489e-05, "loss": 3.8846, "step": 4000 }, { "epoch": 0.0357653791130186, "grad_norm": 1.641852855682373, "learning_rate": 9.999283831785833e-05, "loss": 3.7919, "step": 4050 }, { "epoch": 0.03620692700330278, "grad_norm": 1.8461246490478516, "learning_rate": 9.99926011978989e-05, "loss": 3.2686, "step": 4100 }, { "epoch": 0.03664847489358696, "grad_norm": 2.8383119106292725, "learning_rate": 9.999236021643498e-05, "loss": 3.1711, "step": 4150 }, { "epoch": 0.03709002278387114, "grad_norm": 1.094404935836792, "learning_rate": 9.999211537348514e-05, "loss": 3.4461, "step": 4200 }, { "epoch": 0.03753157067415532, "grad_norm": 1.1627689599990845, "learning_rate": 9.999186666906832e-05, "loss": 3.2783, "step": 4250 }, { "epoch": 0.0379731185644395, "grad_norm": 2.673081398010254, "learning_rate": 9.999161410320373e-05, "loss": 3.3066, "step": 4300 }, { "epoch": 0.03841466645472368, "grad_norm": 2.3519914150238037, "learning_rate": 9.999135767591083e-05, "loss": 3.5882, "step": 4350 }, { "epoch": 0.03885621434500786, "grad_norm": 2.861649990081787, "learning_rate": 9.999109738720949e-05, "loss": 3.3404, "step": 4400 }, { "epoch": 0.03929776223529204, "grad_norm": 2.392594575881958, "learning_rate": 9.999083323711979e-05, "loss": 3.4555, "step": 4450 }, { "epoch": 0.03973931012557622, "grad_norm": 2.1783764362335205, "learning_rate": 9.999056522566213e-05, "loss": 3.3032, "step": 4500 }, { "epoch": 0.0401808580158604, "grad_norm": 3.0805599689483643, "learning_rate": 9.99902933528572e-05, "loss": 3.7245, "step": 4550 }, { "epoch": 0.04062240590614458, "grad_norm": 0.8756254315376282, "learning_rate": 9.999001761872601e-05, "loss": 3.0947, "step": 4600 }, { "epoch": 0.041063953796428763, "grad_norm": 8.237518310546875, "learning_rate": 9.998973802328988e-05, "loss": 3.369, "step": 4650 }, { "epoch": 0.041505501686712944, "grad_norm": 1.5842547416687012, "learning_rate": 9.998945456657039e-05, "loss": 3.6075, "step": 4700 }, { "epoch": 0.041947049576997124, "grad_norm": 2.602689504623413, "learning_rate": 9.998916724858942e-05, "loss": 3.7514, "step": 4750 }, { "epoch": 0.042388597467281304, "grad_norm": 3.264051914215088, "learning_rate": 9.998887606936918e-05, "loss": 3.435, "step": 4800 }, { "epoch": 0.042830145357565484, "grad_norm": 2.2757341861724854, "learning_rate": 9.998858102893215e-05, "loss": 3.6671, "step": 4850 }, { "epoch": 0.043271693247849664, "grad_norm": 1.5097368955612183, "learning_rate": 9.998828212730115e-05, "loss": 3.4472, "step": 4900 }, { "epoch": 0.043713241138133845, "grad_norm": 2.7084147930145264, "learning_rate": 9.998797936449922e-05, "loss": 3.8243, "step": 4950 }, { "epoch": 0.044154789028418025, "grad_norm": 2.9833407402038574, "learning_rate": 9.998767274054978e-05, "loss": 3.7378, "step": 5000 }, { "epoch": 0.044596336918702205, "grad_norm": 9.48874568939209, "learning_rate": 9.998736225547651e-05, "loss": 3.4267, "step": 5050 }, { "epoch": 0.045037884808986385, "grad_norm": 1.7265108823776245, "learning_rate": 9.998704790930337e-05, "loss": 3.4905, "step": 5100 }, { "epoch": 0.045479432699270565, "grad_norm": 1.690320372581482, "learning_rate": 9.998672970205467e-05, "loss": 3.4667, "step": 5150 }, { "epoch": 0.045920980589554745, "grad_norm": 3.318742275238037, "learning_rate": 9.998640763375497e-05, "loss": 3.4528, "step": 5200 }, { "epoch": 0.046362528479838926, "grad_norm": 3.2868552207946777, "learning_rate": 9.998608170442915e-05, "loss": 3.5877, "step": 5250 }, { "epoch": 0.046804076370123106, "grad_norm": 3.8400142192840576, "learning_rate": 9.99857519141024e-05, "loss": 3.5318, "step": 5300 }, { "epoch": 0.047245624260407286, "grad_norm": 1.9866262674331665, "learning_rate": 9.998541826280018e-05, "loss": 3.7204, "step": 5350 }, { "epoch": 0.047687172150691466, "grad_norm": 2.3364500999450684, "learning_rate": 9.998508075054826e-05, "loss": 3.3647, "step": 5400 }, { "epoch": 0.048128720040975646, "grad_norm": 3.461249589920044, "learning_rate": 9.99847393773727e-05, "loss": 3.4072, "step": 5450 }, { "epoch": 0.048570267931259826, "grad_norm": 1.401358962059021, "learning_rate": 9.99843941432999e-05, "loss": 3.506, "step": 5500 }, { "epoch": 0.04901181582154401, "grad_norm": 2.5606305599212646, "learning_rate": 9.99840450483565e-05, "loss": 3.3622, "step": 5550 }, { "epoch": 0.04945336371182819, "grad_norm": 2.909435510635376, "learning_rate": 9.998369209256947e-05, "loss": 3.1674, "step": 5600 }, { "epoch": 0.04989491160211237, "grad_norm": 1.5808318853378296, "learning_rate": 9.998333527596607e-05, "loss": 3.5161, "step": 5650 }, { "epoch": 0.05033645949239655, "grad_norm": 1.6047215461730957, "learning_rate": 9.998297459857387e-05, "loss": 3.528, "step": 5700 }, { "epoch": 0.05077800738268073, "grad_norm": 2.816723346710205, "learning_rate": 9.998261006042074e-05, "loss": 3.1358, "step": 5750 }, { "epoch": 0.05121955527296491, "grad_norm": 1.9394768476486206, "learning_rate": 9.99822416615348e-05, "loss": 2.8468, "step": 5800 }, { "epoch": 0.05166110316324909, "grad_norm": 7.701488971710205, "learning_rate": 9.998186940194454e-05, "loss": 3.5079, "step": 5850 }, { "epoch": 0.05210265105353327, "grad_norm": 1.9361647367477417, "learning_rate": 9.998150084191847e-05, "loss": 3.1601, "step": 5900 }, { "epoch": 0.05254419894381745, "grad_norm": 1.7290771007537842, "learning_rate": 9.998112093821874e-05, "loss": 3.5756, "step": 5950 }, { "epoch": 0.05298574683410163, "grad_norm": 7.283807754516602, "learning_rate": 9.998073717390126e-05, "loss": 3.3956, "step": 6000 }, { "epoch": 0.05298574683410163, "eval_asr_loss": 0.8455491743058053, "eval_loss": 3.3733489513397217, "eval_runtime": 20.0292, "eval_samples_per_second": 38.344, "eval_steps_per_second": 9.586, "eval_tts_loss": 5.977993885297463, "step": 6000 }, { "epoch": 0.05342729472438581, "grad_norm": 2.8105592727661133, "learning_rate": 9.998034954899564e-05, "loss": 3.5391, "step": 6050 }, { "epoch": 0.05386884261466999, "grad_norm": 2.575618267059326, "learning_rate": 9.997995806353187e-05, "loss": 3.4848, "step": 6100 }, { "epoch": 0.05431039050495417, "grad_norm": 1.944770336151123, "learning_rate": 9.997956271754013e-05, "loss": 3.4319, "step": 6150 }, { "epoch": 0.05475193839523835, "grad_norm": 2.9254565238952637, "learning_rate": 9.997916351105098e-05, "loss": 3.5138, "step": 6200 }, { "epoch": 0.05519348628552253, "grad_norm": 2.876880645751953, "learning_rate": 9.997876044409529e-05, "loss": 3.2436, "step": 6250 }, { "epoch": 0.05563503417580671, "grad_norm": 1.5080561637878418, "learning_rate": 9.997835351670415e-05, "loss": 3.3394, "step": 6300 }, { "epoch": 0.05607658206609089, "grad_norm": 2.99318790435791, "learning_rate": 9.9977942728909e-05, "loss": 3.2061, "step": 6350 }, { "epoch": 0.05651812995637507, "grad_norm": 2.1895508766174316, "learning_rate": 9.997752808074157e-05, "loss": 3.586, "step": 6400 }, { "epoch": 0.05695967784665925, "grad_norm": 2.089017868041992, "learning_rate": 9.997710957223389e-05, "loss": 3.6418, "step": 6450 }, { "epoch": 0.05740122573694343, "grad_norm": 1.3540375232696533, "learning_rate": 9.997668720341831e-05, "loss": 3.5735, "step": 6500 }, { "epoch": 0.05784277362722761, "grad_norm": 2.3487038612365723, "learning_rate": 9.997626097432741e-05, "loss": 3.4285, "step": 6550 }, { "epoch": 0.05828432151751179, "grad_norm": 2.5517020225524902, "learning_rate": 9.997583088499415e-05, "loss": 3.5823, "step": 6600 }, { "epoch": 0.05872586940779597, "grad_norm": 1.7537412643432617, "learning_rate": 9.997539693545174e-05, "loss": 3.1481, "step": 6650 }, { "epoch": 0.05916741729808015, "grad_norm": 2.911041736602783, "learning_rate": 9.99749591257337e-05, "loss": 3.7261, "step": 6700 }, { "epoch": 0.05960896518836433, "grad_norm": 7.761764049530029, "learning_rate": 9.997451745587382e-05, "loss": 3.8374, "step": 6750 }, { "epoch": 0.06005051307864851, "grad_norm": 6.103822708129883, "learning_rate": 9.997407192590625e-05, "loss": 3.1474, "step": 6800 }, { "epoch": 0.06049206096893269, "grad_norm": 8.443099021911621, "learning_rate": 9.99736225358654e-05, "loss": 3.4926, "step": 6850 }, { "epoch": 0.06093360885921687, "grad_norm": 6.040070056915283, "learning_rate": 9.997316928578595e-05, "loss": 3.1745, "step": 6900 }, { "epoch": 0.06137515674950105, "grad_norm": 1.7983872890472412, "learning_rate": 9.997271217570296e-05, "loss": 3.5779, "step": 6950 }, { "epoch": 0.06181670463978523, "grad_norm": 2.881488800048828, "learning_rate": 9.99722512056517e-05, "loss": 3.6113, "step": 7000 }, { "epoch": 0.06225825253006941, "grad_norm": 2.0535004138946533, "learning_rate": 9.997178637566777e-05, "loss": 3.5762, "step": 7050 }, { "epoch": 0.06269980042035359, "grad_norm": 2.7159841060638428, "learning_rate": 9.997131768578711e-05, "loss": 3.4012, "step": 7100 }, { "epoch": 0.06314134831063777, "grad_norm": 3.324490785598755, "learning_rate": 9.997084513604591e-05, "loss": 3.8183, "step": 7150 }, { "epoch": 0.06358289620092195, "grad_norm": 1.6304800510406494, "learning_rate": 9.997036872648064e-05, "loss": 3.4161, "step": 7200 }, { "epoch": 0.06402444409120613, "grad_norm": 2.0595145225524902, "learning_rate": 9.996988845712813e-05, "loss": 3.4944, "step": 7250 }, { "epoch": 0.0644659919814903, "grad_norm": 2.9137229919433594, "learning_rate": 9.996940432802548e-05, "loss": 3.6446, "step": 7300 }, { "epoch": 0.06490753987177449, "grad_norm": 2.6019785404205322, "learning_rate": 9.996891633921007e-05, "loss": 3.5755, "step": 7350 }, { "epoch": 0.06534908776205867, "grad_norm": 1.3080384731292725, "learning_rate": 9.99684244907196e-05, "loss": 3.432, "step": 7400 }, { "epoch": 0.06579063565234285, "grad_norm": 1.085084080696106, "learning_rate": 9.996792878259204e-05, "loss": 3.5939, "step": 7450 }, { "epoch": 0.06623218354262703, "grad_norm": 2.416012763977051, "learning_rate": 9.996742921486573e-05, "loss": 3.3125, "step": 7500 }, { "epoch": 0.0666737314329112, "grad_norm": 2.329331159591675, "learning_rate": 9.99669257875792e-05, "loss": 3.4727, "step": 7550 }, { "epoch": 0.06711527932319539, "grad_norm": 1.2133623361587524, "learning_rate": 9.996641850077135e-05, "loss": 3.6817, "step": 7600 }, { "epoch": 0.06755682721347957, "grad_norm": 3.447230577468872, "learning_rate": 9.99659073544814e-05, "loss": 3.3487, "step": 7650 }, { "epoch": 0.06799837510376375, "grad_norm": 1.8407869338989258, "learning_rate": 9.99653923487488e-05, "loss": 3.4952, "step": 7700 }, { "epoch": 0.06843992299404793, "grad_norm": 3.4017183780670166, "learning_rate": 9.996487348361331e-05, "loss": 3.5987, "step": 7750 }, { "epoch": 0.06888147088433211, "grad_norm": 2.014235734939575, "learning_rate": 9.996435075911507e-05, "loss": 3.4913, "step": 7800 }, { "epoch": 0.0693230187746163, "grad_norm": 8.113083839416504, "learning_rate": 9.99638241752944e-05, "loss": 3.2701, "step": 7850 }, { "epoch": 0.06976456666490048, "grad_norm": 3.6530191898345947, "learning_rate": 9.996329373219199e-05, "loss": 3.7682, "step": 7900 }, { "epoch": 0.07020611455518466, "grad_norm": 3.110872507095337, "learning_rate": 9.996275942984882e-05, "loss": 2.9624, "step": 7950 }, { "epoch": 0.07064766244546884, "grad_norm": 5.682243347167969, "learning_rate": 9.996222126830616e-05, "loss": 3.1075, "step": 8000 }, { "epoch": 0.07108921033575302, "grad_norm": 8.339673042297363, "learning_rate": 9.996167924760556e-05, "loss": 3.4832, "step": 8050 }, { "epoch": 0.0715307582260372, "grad_norm": 2.9629604816436768, "learning_rate": 9.996113336778891e-05, "loss": 3.5399, "step": 8100 }, { "epoch": 0.07197230611632138, "grad_norm": 4.072164058685303, "learning_rate": 9.996058362889837e-05, "loss": 3.3337, "step": 8150 }, { "epoch": 0.07241385400660556, "grad_norm": 2.1256349086761475, "learning_rate": 9.99600300309764e-05, "loss": 3.4296, "step": 8200 }, { "epoch": 0.07285540189688974, "grad_norm": 0.9664581418037415, "learning_rate": 9.995947257406575e-05, "loss": 3.3867, "step": 8250 }, { "epoch": 0.07329694978717392, "grad_norm": 2.7981796264648438, "learning_rate": 9.995891125820948e-05, "loss": 3.5777, "step": 8300 }, { "epoch": 0.0737384976774581, "grad_norm": 2.2892305850982666, "learning_rate": 9.995834608345098e-05, "loss": 3.448, "step": 8350 }, { "epoch": 0.07418004556774228, "grad_norm": 3.462890386581421, "learning_rate": 9.995777704983387e-05, "loss": 3.6835, "step": 8400 }, { "epoch": 0.07462159345802646, "grad_norm": 2.887977123260498, "learning_rate": 9.995720415740209e-05, "loss": 3.8102, "step": 8450 }, { "epoch": 0.07506314134831064, "grad_norm": 1.103590726852417, "learning_rate": 9.995662740619993e-05, "loss": 3.3465, "step": 8500 }, { "epoch": 0.07550468923859482, "grad_norm": 1.5264739990234375, "learning_rate": 9.995604679627193e-05, "loss": 3.3524, "step": 8550 }, { "epoch": 0.075946237128879, "grad_norm": 10.508204460144043, "learning_rate": 9.995546232766293e-05, "loss": 2.9848, "step": 8600 }, { "epoch": 0.07638778501916318, "grad_norm": 11.036320686340332, "learning_rate": 9.995487400041806e-05, "loss": 3.6104, "step": 8650 }, { "epoch": 0.07682933290944736, "grad_norm": 0.900551974773407, "learning_rate": 9.995428181458279e-05, "loss": 3.2115, "step": 8700 }, { "epoch": 0.07727088079973154, "grad_norm": 1.0776323080062866, "learning_rate": 9.995368577020285e-05, "loss": 3.378, "step": 8750 }, { "epoch": 0.07771242869001573, "grad_norm": 1.9495457410812378, "learning_rate": 9.995308586732427e-05, "loss": 3.134, "step": 8800 }, { "epoch": 0.0781539765802999, "grad_norm": 2.638998508453369, "learning_rate": 9.99524821059934e-05, "loss": 3.158, "step": 8850 }, { "epoch": 0.07859552447058409, "grad_norm": 1.9697933197021484, "learning_rate": 9.995187448625688e-05, "loss": 3.4271, "step": 8900 }, { "epoch": 0.07903707236086827, "grad_norm": 1.1985995769500732, "learning_rate": 9.995126300816163e-05, "loss": 3.3212, "step": 8950 }, { "epoch": 0.07947862025115245, "grad_norm": 5.711977005004883, "learning_rate": 9.995064767175488e-05, "loss": 3.256, "step": 9000 }, { "epoch": 0.07947862025115245, "eval_asr_loss": 0.8664082059474173, "eval_loss": 3.2935588359832764, "eval_runtime": 19.9624, "eval_samples_per_second": 38.472, "eval_steps_per_second": 9.618, "eval_tts_loss": 5.950566485041761, "step": 9000 }, { "epoch": 0.07992016814143663, "grad_norm": 2.0665061473846436, "learning_rate": 9.995002847708418e-05, "loss": 3.5921, "step": 9050 }, { "epoch": 0.0803617160317208, "grad_norm": 2.0567665100097656, "learning_rate": 9.994940542419733e-05, "loss": 3.4312, "step": 9100 }, { "epoch": 0.08080326392200499, "grad_norm": 5.3886847496032715, "learning_rate": 9.994879108917332e-05, "loss": 3.2038, "step": 9150 }, { "epoch": 0.08124481181228917, "grad_norm": 5.97287130355835, "learning_rate": 9.994816039716079e-05, "loss": 3.4363, "step": 9200 }, { "epoch": 0.08168635970257335, "grad_norm": 3.153740406036377, "learning_rate": 9.994752584707642e-05, "loss": 3.3041, "step": 9250 }, { "epoch": 0.08212790759285753, "grad_norm": 2.116394519805908, "learning_rate": 9.994688743896923e-05, "loss": 3.906, "step": 9300 }, { "epoch": 0.08256945548314171, "grad_norm": 2.2864694595336914, "learning_rate": 9.994624517288851e-05, "loss": 3.5515, "step": 9350 }, { "epoch": 0.08301100337342589, "grad_norm": 6.8061723709106445, "learning_rate": 9.994559904888388e-05, "loss": 3.5991, "step": 9400 }, { "epoch": 0.08345255126371007, "grad_norm": 5.035069465637207, "learning_rate": 9.994494906700527e-05, "loss": 3.7135, "step": 9450 }, { "epoch": 0.08389409915399425, "grad_norm": 2.2650699615478516, "learning_rate": 9.994429522730284e-05, "loss": 3.3555, "step": 9500 }, { "epoch": 0.08433564704427843, "grad_norm": 3.1491940021514893, "learning_rate": 9.994363752982714e-05, "loss": 3.0579, "step": 9550 }, { "epoch": 0.08477719493456261, "grad_norm": 3.5358498096466064, "learning_rate": 9.994297597462894e-05, "loss": 3.6277, "step": 9600 }, { "epoch": 0.08521874282484679, "grad_norm": 3.785895824432373, "learning_rate": 9.994231056175936e-05, "loss": 3.2452, "step": 9650 }, { "epoch": 0.08566029071513097, "grad_norm": 1.632702350616455, "learning_rate": 9.994164129126977e-05, "loss": 3.1927, "step": 9700 }, { "epoch": 0.08610183860541515, "grad_norm": 3.5743367671966553, "learning_rate": 9.99409681632119e-05, "loss": 3.6792, "step": 9750 }, { "epoch": 0.08654338649569933, "grad_norm": 9.39389419555664, "learning_rate": 9.994029117763773e-05, "loss": 3.1351, "step": 9800 }, { "epoch": 0.08698493438598351, "grad_norm": 2.9739062786102295, "learning_rate": 9.993961033459953e-05, "loss": 3.1706, "step": 9850 }, { "epoch": 0.08742648227626769, "grad_norm": 3.320171356201172, "learning_rate": 9.99389256341499e-05, "loss": 3.6547, "step": 9900 }, { "epoch": 0.08786803016655187, "grad_norm": 8.49543285369873, "learning_rate": 9.993823707634176e-05, "loss": 3.701, "step": 9950 }, { "epoch": 0.08830957805683605, "grad_norm": 1.3553107976913452, "learning_rate": 9.993754466122827e-05, "loss": 3.4541, "step": 10000 }, { "epoch": 0.08875112594712023, "grad_norm": 1.0776971578598022, "learning_rate": 9.993684838886289e-05, "loss": 3.7029, "step": 10050 }, { "epoch": 0.08919267383740441, "grad_norm": 1.6787606477737427, "learning_rate": 9.993614825929945e-05, "loss": 3.6013, "step": 10100 }, { "epoch": 0.08963422172768859, "grad_norm": 1.4465861320495605, "learning_rate": 9.993544427259198e-05, "loss": 3.2941, "step": 10150 }, { "epoch": 0.09007576961797277, "grad_norm": 3.221482276916504, "learning_rate": 9.99347364287949e-05, "loss": 3.5483, "step": 10200 }, { "epoch": 0.09051731750825695, "grad_norm": 1.0219640731811523, "learning_rate": 9.993402472796284e-05, "loss": 3.2682, "step": 10250 }, { "epoch": 0.09095886539854113, "grad_norm": 4.225593090057373, "learning_rate": 9.993330917015082e-05, "loss": 3.4346, "step": 10300 }, { "epoch": 0.09140041328882531, "grad_norm": 1.145766019821167, "learning_rate": 9.993258975541408e-05, "loss": 3.5205, "step": 10350 }, { "epoch": 0.09184196117910949, "grad_norm": 4.837847709655762, "learning_rate": 9.99318664838082e-05, "loss": 3.3985, "step": 10400 }, { "epoch": 0.09228350906939367, "grad_norm": 3.118101119995117, "learning_rate": 9.993113935538903e-05, "loss": 3.6283, "step": 10450 }, { "epoch": 0.09272505695967785, "grad_norm": 3.6168124675750732, "learning_rate": 9.993040837021277e-05, "loss": 3.1753, "step": 10500 }, { "epoch": 0.09316660484996203, "grad_norm": 3.866116523742676, "learning_rate": 9.992967352833584e-05, "loss": 3.1334, "step": 10550 }, { "epoch": 0.09360815274024621, "grad_norm": 2.1199982166290283, "learning_rate": 9.992893482981505e-05, "loss": 3.3849, "step": 10600 }, { "epoch": 0.09404970063053039, "grad_norm": 1.948744297027588, "learning_rate": 9.99281922747074e-05, "loss": 3.3414, "step": 10650 }, { "epoch": 0.09449124852081457, "grad_norm": 5.387505054473877, "learning_rate": 9.99274458630703e-05, "loss": 3.6585, "step": 10700 }, { "epoch": 0.09493279641109875, "grad_norm": 2.675678253173828, "learning_rate": 9.992669559496136e-05, "loss": 3.1715, "step": 10750 }, { "epoch": 0.09537434430138293, "grad_norm": 1.5642513036727905, "learning_rate": 9.992594147043856e-05, "loss": 3.2574, "step": 10800 }, { "epoch": 0.09581589219166711, "grad_norm": 2.7304491996765137, "learning_rate": 9.992518348956014e-05, "loss": 3.547, "step": 10850 }, { "epoch": 0.09625744008195129, "grad_norm": 3.1260108947753906, "learning_rate": 9.992442165238465e-05, "loss": 3.6459, "step": 10900 }, { "epoch": 0.09669898797223547, "grad_norm": 1.5745757818222046, "learning_rate": 9.992365595897092e-05, "loss": 3.0718, "step": 10950 }, { "epoch": 0.09714053586251965, "grad_norm": 4.948554515838623, "learning_rate": 9.992288640937812e-05, "loss": 2.9318, "step": 11000 }, { "epoch": 0.09758208375280383, "grad_norm": 3.8740906715393066, "learning_rate": 9.992211300366568e-05, "loss": 3.5334, "step": 11050 }, { "epoch": 0.09802363164308801, "grad_norm": 4.66175651550293, "learning_rate": 9.992133574189335e-05, "loss": 3.4052, "step": 11100 }, { "epoch": 0.0984651795333722, "grad_norm": 3.1660337448120117, "learning_rate": 9.992055462412113e-05, "loss": 3.4354, "step": 11150 }, { "epoch": 0.09890672742365637, "grad_norm": 3.1641385555267334, "learning_rate": 9.99197696504094e-05, "loss": 3.4509, "step": 11200 }, { "epoch": 0.09934827531394055, "grad_norm": 2.3828437328338623, "learning_rate": 9.991898082081874e-05, "loss": 3.1644, "step": 11250 }, { "epoch": 0.09978982320422473, "grad_norm": 3.6782753467559814, "learning_rate": 9.991818813541014e-05, "loss": 3.6032, "step": 11300 }, { "epoch": 0.10023137109450891, "grad_norm": 1.4960347414016724, "learning_rate": 9.991739159424481e-05, "loss": 3.3263, "step": 11350 }, { "epoch": 0.1006729189847931, "grad_norm": 3.1267852783203125, "learning_rate": 9.991659119738423e-05, "loss": 3.4492, "step": 11400 }, { "epoch": 0.10111446687507727, "grad_norm": 1.307042121887207, "learning_rate": 9.991578694489028e-05, "loss": 3.2172, "step": 11450 }, { "epoch": 0.10155601476536145, "grad_norm": 2.0429134368896484, "learning_rate": 9.991497883682506e-05, "loss": 3.171, "step": 11500 }, { "epoch": 0.10199756265564563, "grad_norm": 2.784212350845337, "learning_rate": 9.991416687325101e-05, "loss": 3.4447, "step": 11550 }, { "epoch": 0.10243911054592982, "grad_norm": 6.423585414886475, "learning_rate": 9.991335105423081e-05, "loss": 3.2792, "step": 11600 }, { "epoch": 0.102880658436214, "grad_norm": 2.206615924835205, "learning_rate": 9.99125313798275e-05, "loss": 3.2817, "step": 11650 }, { "epoch": 0.10332220632649818, "grad_norm": 4.011673450469971, "learning_rate": 9.991170785010438e-05, "loss": 3.692, "step": 11700 }, { "epoch": 0.10376375421678236, "grad_norm": 1.4347039461135864, "learning_rate": 9.991088046512507e-05, "loss": 3.3256, "step": 11750 }, { "epoch": 0.10420530210706654, "grad_norm": 3.251509428024292, "learning_rate": 9.991004922495348e-05, "loss": 3.3607, "step": 11800 }, { "epoch": 0.10464684999735072, "grad_norm": 5.043297290802002, "learning_rate": 9.990921412965381e-05, "loss": 3.0377, "step": 11850 }, { "epoch": 0.1050883978876349, "grad_norm": 1.7166991233825684, "learning_rate": 9.990837517929057e-05, "loss": 3.2673, "step": 11900 }, { "epoch": 0.10552994577791908, "grad_norm": 2.2065978050231934, "learning_rate": 9.990753237392854e-05, "loss": 3.5916, "step": 11950 }, { "epoch": 0.10597149366820326, "grad_norm": 3.4584174156188965, "learning_rate": 9.990668571363286e-05, "loss": 3.1674, "step": 12000 }, { "epoch": 0.10597149366820326, "eval_asr_loss": 0.8610385786988017, "eval_loss": 3.250169515609741, "eval_runtime": 20.2526, "eval_samples_per_second": 37.921, "eval_steps_per_second": 9.48, "eval_tts_loss": 6.012793056516366, "step": 12000 }, { "epoch": 0.10641304155848744, "grad_norm": 3.1279189586639404, "learning_rate": 9.99058351984689e-05, "loss": 3.5257, "step": 12050 }, { "epoch": 0.10685458944877162, "grad_norm": 5.271905899047852, "learning_rate": 9.990498082850234e-05, "loss": 3.2174, "step": 12100 }, { "epoch": 0.1072961373390558, "grad_norm": 2.162796974182129, "learning_rate": 9.990412260379922e-05, "loss": 3.4223, "step": 12150 }, { "epoch": 0.10773768522933998, "grad_norm": 2.182065725326538, "learning_rate": 9.990326052442579e-05, "loss": 3.3126, "step": 12200 }, { "epoch": 0.10817923311962416, "grad_norm": 1.2493520975112915, "learning_rate": 9.990239459044866e-05, "loss": 3.5304, "step": 12250 }, { "epoch": 0.10862078100990834, "grad_norm": 2.064229726791382, "learning_rate": 9.99015248019347e-05, "loss": 3.4558, "step": 12300 }, { "epoch": 0.10906232890019252, "grad_norm": 2.22754168510437, "learning_rate": 9.99006511589511e-05, "loss": 3.3677, "step": 12350 }, { "epoch": 0.1095038767904767, "grad_norm": 4.160098075866699, "learning_rate": 9.989977366156535e-05, "loss": 3.5189, "step": 12400 }, { "epoch": 0.10994542468076088, "grad_norm": 0.9621350169181824, "learning_rate": 9.989889230984522e-05, "loss": 3.7358, "step": 12450 }, { "epoch": 0.11038697257104506, "grad_norm": 1.2985560894012451, "learning_rate": 9.989800710385879e-05, "loss": 3.6572, "step": 12500 }, { "epoch": 0.11082852046132924, "grad_norm": 0.9708568453788757, "learning_rate": 9.989711804367443e-05, "loss": 3.2729, "step": 12550 }, { "epoch": 0.11127006835161342, "grad_norm": 1.924854040145874, "learning_rate": 9.989622512936083e-05, "loss": 3.4972, "step": 12600 }, { "epoch": 0.1117116162418976, "grad_norm": 4.869857311248779, "learning_rate": 9.989532836098691e-05, "loss": 3.5102, "step": 12650 }, { "epoch": 0.11215316413218178, "grad_norm": 5.745826244354248, "learning_rate": 9.9894427738622e-05, "loss": 3.3836, "step": 12700 }, { "epoch": 0.11259471202246596, "grad_norm": 5.497977256774902, "learning_rate": 9.989352326233566e-05, "loss": 3.3298, "step": 12750 }, { "epoch": 0.11303625991275014, "grad_norm": 9.22269344329834, "learning_rate": 9.98926149321977e-05, "loss": 3.3542, "step": 12800 }, { "epoch": 0.11347780780303432, "grad_norm": 2.255528211593628, "learning_rate": 9.989172102972332e-05, "loss": 3.8121, "step": 12850 }, { "epoch": 0.1139193556933185, "grad_norm": 2.629471778869629, "learning_rate": 9.98908050691665e-05, "loss": 3.456, "step": 12900 }, { "epoch": 0.11436090358360268, "grad_norm": 6.432432174682617, "learning_rate": 9.988988525496805e-05, "loss": 3.1076, "step": 12950 }, { "epoch": 0.11480245147388686, "grad_norm": 1.9251166582107544, "learning_rate": 9.988896158719903e-05, "loss": 3.5083, "step": 13000 }, { "epoch": 0.11524399936417104, "grad_norm": 1.9608491659164429, "learning_rate": 9.988803406593077e-05, "loss": 3.4977, "step": 13050 }, { "epoch": 0.11568554725445522, "grad_norm": 2.201385259628296, "learning_rate": 9.988710269123491e-05, "loss": 3.3394, "step": 13100 }, { "epoch": 0.1161270951447394, "grad_norm": 0.8397690653800964, "learning_rate": 9.98861674631834e-05, "loss": 3.0705, "step": 13150 }, { "epoch": 0.11656864303502358, "grad_norm": 3.4315147399902344, "learning_rate": 9.988522838184848e-05, "loss": 2.9964, "step": 13200 }, { "epoch": 0.11701019092530776, "grad_norm": 1.0664474964141846, "learning_rate": 9.988428544730267e-05, "loss": 3.5425, "step": 13250 }, { "epoch": 0.11745173881559194, "grad_norm": 1.7559316158294678, "learning_rate": 9.988333865961883e-05, "loss": 3.4283, "step": 13300 }, { "epoch": 0.11789328670587612, "grad_norm": 1.3742626905441284, "learning_rate": 9.988238801887006e-05, "loss": 3.0457, "step": 13350 }, { "epoch": 0.1183348345961603, "grad_norm": 6.306196212768555, "learning_rate": 9.988143352512982e-05, "loss": 3.4985, "step": 13400 }, { "epoch": 0.11877638248644448, "grad_norm": 0.7708596587181091, "learning_rate": 9.98804751784718e-05, "loss": 3.3379, "step": 13450 }, { "epoch": 0.11921793037672866, "grad_norm": 7.315723419189453, "learning_rate": 9.987951297897008e-05, "loss": 3.4271, "step": 13500 }, { "epoch": 0.11965947826701284, "grad_norm": 1.3316899538040161, "learning_rate": 9.987854692669894e-05, "loss": 3.362, "step": 13550 }, { "epoch": 0.12010102615729702, "grad_norm": 2.5768868923187256, "learning_rate": 9.9877577021733e-05, "loss": 3.7057, "step": 13600 }, { "epoch": 0.1205425740475812, "grad_norm": 1.0544745922088623, "learning_rate": 9.987660326414718e-05, "loss": 3.2666, "step": 13650 }, { "epoch": 0.12098412193786538, "grad_norm": 1.6136611700057983, "learning_rate": 9.98756256540167e-05, "loss": 3.019, "step": 13700 }, { "epoch": 0.12142566982814956, "grad_norm": 3.0674450397491455, "learning_rate": 9.987464419141707e-05, "loss": 3.3559, "step": 13750 }, { "epoch": 0.12186721771843374, "grad_norm": 5.650421142578125, "learning_rate": 9.987365887642412e-05, "loss": 3.1113, "step": 13800 }, { "epoch": 0.12230876560871792, "grad_norm": 6.18209981918335, "learning_rate": 9.987266970911393e-05, "loss": 3.6087, "step": 13850 }, { "epoch": 0.1227503134990021, "grad_norm": 3.334939956665039, "learning_rate": 9.98716766895629e-05, "loss": 3.2844, "step": 13900 }, { "epoch": 0.12319186138928628, "grad_norm": 1.9125257730484009, "learning_rate": 9.987067981784774e-05, "loss": 3.5561, "step": 13950 }, { "epoch": 0.12363340927957046, "grad_norm": 2.013408899307251, "learning_rate": 9.986967909404547e-05, "loss": 3.7224, "step": 14000 }, { "epoch": 0.12407495716985464, "grad_norm": 1.1568374633789062, "learning_rate": 9.986867451823337e-05, "loss": 3.348, "step": 14050 }, { "epoch": 0.12451650506013882, "grad_norm": 2.9201159477233887, "learning_rate": 9.986766609048904e-05, "loss": 3.358, "step": 14100 }, { "epoch": 0.124958052950423, "grad_norm": 6.0850982666015625, "learning_rate": 9.986665381089038e-05, "loss": 3.3947, "step": 14150 }, { "epoch": 0.12539960084070717, "grad_norm": 2.2208523750305176, "learning_rate": 9.986563767951555e-05, "loss": 3.3521, "step": 14200 }, { "epoch": 0.12584114873099136, "grad_norm": 3.7602078914642334, "learning_rate": 9.986461769644306e-05, "loss": 3.3009, "step": 14250 }, { "epoch": 0.12628269662127553, "grad_norm": 5.159857749938965, "learning_rate": 9.98635938617517e-05, "loss": 3.0826, "step": 14300 }, { "epoch": 0.12672424451155972, "grad_norm": 1.6318804025650024, "learning_rate": 9.986256617552054e-05, "loss": 3.1461, "step": 14350 }, { "epoch": 0.1271657924018439, "grad_norm": 1.6564624309539795, "learning_rate": 9.986153463782897e-05, "loss": 3.6095, "step": 14400 }, { "epoch": 0.12760734029212809, "grad_norm": 6.902859210968018, "learning_rate": 9.986049924875666e-05, "loss": 3.4435, "step": 14450 }, { "epoch": 0.12804888818241225, "grad_norm": 2.513582229614258, "learning_rate": 9.985946000838359e-05, "loss": 3.6592, "step": 14500 }, { "epoch": 0.12849043607269645, "grad_norm": 1.5714596509933472, "learning_rate": 9.985841691679004e-05, "loss": 3.3225, "step": 14550 }, { "epoch": 0.1289319839629806, "grad_norm": 2.4705002307891846, "learning_rate": 9.985736997405655e-05, "loss": 3.3319, "step": 14600 }, { "epoch": 0.1293735318532648, "grad_norm": 0.8616177439689636, "learning_rate": 9.985631918026401e-05, "loss": 3.1653, "step": 14650 }, { "epoch": 0.12981507974354897, "grad_norm": 1.7353193759918213, "learning_rate": 9.985526453549359e-05, "loss": 3.4691, "step": 14700 }, { "epoch": 0.13025662763383317, "grad_norm": 5.408633708953857, "learning_rate": 9.985420603982673e-05, "loss": 3.5129, "step": 14750 }, { "epoch": 0.13069817552411733, "grad_norm": 1.2819982767105103, "learning_rate": 9.985314369334523e-05, "loss": 3.2823, "step": 14800 }, { "epoch": 0.13113972341440153, "grad_norm": 8.26905632019043, "learning_rate": 9.98520774961311e-05, "loss": 3.2017, "step": 14850 }, { "epoch": 0.1315812713046857, "grad_norm": 2.404160737991333, "learning_rate": 9.985100744826674e-05, "loss": 3.2397, "step": 14900 }, { "epoch": 0.1320228191949699, "grad_norm": 1.9694509506225586, "learning_rate": 9.984993354983477e-05, "loss": 3.7057, "step": 14950 }, { "epoch": 0.13246436708525405, "grad_norm": 7.179697036743164, "learning_rate": 9.984885580091814e-05, "loss": 2.8689, "step": 15000 }, { "epoch": 0.13246436708525405, "eval_asr_loss": 0.8746035999883381, "eval_loss": 3.1943347454071045, "eval_runtime": 20.0975, "eval_samples_per_second": 38.214, "eval_steps_per_second": 9.553, "eval_tts_loss": 5.963482838455088, "step": 15000 }, { "epoch": 0.13290591497553825, "grad_norm": 1.963370680809021, "learning_rate": 9.98477742016001e-05, "loss": 3.0848, "step": 15050 }, { "epoch": 0.1333474628658224, "grad_norm": 5.354642868041992, "learning_rate": 9.984668875196421e-05, "loss": 3.4422, "step": 15100 }, { "epoch": 0.1337890107561066, "grad_norm": 0.9628197550773621, "learning_rate": 9.98455994520943e-05, "loss": 3.6956, "step": 15150 }, { "epoch": 0.13423055864639077, "grad_norm": 1.9591059684753418, "learning_rate": 9.984450630207451e-05, "loss": 3.8912, "step": 15200 }, { "epoch": 0.13467210653667497, "grad_norm": 2.4057066440582275, "learning_rate": 9.984340930198927e-05, "loss": 3.4458, "step": 15250 }, { "epoch": 0.13511365442695913, "grad_norm": 1.3141722679138184, "learning_rate": 9.984230845192336e-05, "loss": 3.3671, "step": 15300 }, { "epoch": 0.13555520231724333, "grad_norm": 3.438781261444092, "learning_rate": 9.984120375196174e-05, "loss": 3.2085, "step": 15350 }, { "epoch": 0.1359967502075275, "grad_norm": 1.443537712097168, "learning_rate": 9.984011741091279e-05, "loss": 3.2912, "step": 15400 }, { "epoch": 0.1364382980978117, "grad_norm": 2.5141725540161133, "learning_rate": 9.983900508840976e-05, "loss": 3.4251, "step": 15450 }, { "epoch": 0.13687984598809586, "grad_norm": 3.5416128635406494, "learning_rate": 9.98378889162662e-05, "loss": 3.4953, "step": 15500 }, { "epoch": 0.13732139387838005, "grad_norm": 3.750843048095703, "learning_rate": 9.983676889456833e-05, "loss": 3.6289, "step": 15550 }, { "epoch": 0.13776294176866422, "grad_norm": 8.563606262207031, "learning_rate": 9.983564502340267e-05, "loss": 3.3268, "step": 15600 }, { "epoch": 0.1382044896589484, "grad_norm": 3.2222371101379395, "learning_rate": 9.983451730285603e-05, "loss": 3.329, "step": 15650 }, { "epoch": 0.1386460375492326, "grad_norm": 2.218827724456787, "learning_rate": 9.983338573301552e-05, "loss": 3.8973, "step": 15700 }, { "epoch": 0.13908758543951677, "grad_norm": 1.1404446363449097, "learning_rate": 9.983225031396852e-05, "loss": 3.5931, "step": 15750 }, { "epoch": 0.13952913332980096, "grad_norm": 0.8973853588104248, "learning_rate": 9.983111104580276e-05, "loss": 3.2535, "step": 15800 }, { "epoch": 0.13997068122008513, "grad_norm": 8.366558074951172, "learning_rate": 9.982996792860623e-05, "loss": 3.0558, "step": 15850 }, { "epoch": 0.14041222911036932, "grad_norm": 1.4844486713409424, "learning_rate": 9.982882096246722e-05, "loss": 3.4828, "step": 15900 }, { "epoch": 0.1408537770006535, "grad_norm": 2.583129644393921, "learning_rate": 9.982767014747432e-05, "loss": 3.6149, "step": 15950 }, { "epoch": 0.14129532489093768, "grad_norm": 3.2486109733581543, "learning_rate": 9.982651548371644e-05, "loss": 3.1331, "step": 16000 }, { "epoch": 0.14173687278122185, "grad_norm": 2.9078361988067627, "learning_rate": 9.982535697128275e-05, "loss": 2.9418, "step": 16050 }, { "epoch": 0.14217842067150605, "grad_norm": 1.986458659172058, "learning_rate": 9.982419461026273e-05, "loss": 3.4626, "step": 16100 }, { "epoch": 0.1426199685617902, "grad_norm": 5.311156272888184, "learning_rate": 9.982302840074617e-05, "loss": 3.1937, "step": 16150 }, { "epoch": 0.1430615164520744, "grad_norm": 1.4057707786560059, "learning_rate": 9.982185834282319e-05, "loss": 4.0044, "step": 16200 }, { "epoch": 0.14350306434235857, "grad_norm": 1.7638394832611084, "learning_rate": 9.98206844365841e-05, "loss": 3.657, "step": 16250 }, { "epoch": 0.14394461223264277, "grad_norm": 2.069039821624756, "learning_rate": 9.98195066821196e-05, "loss": 3.5534, "step": 16300 }, { "epoch": 0.14438616012292693, "grad_norm": 7.656248092651367, "learning_rate": 9.981832507952067e-05, "loss": 3.1842, "step": 16350 }, { "epoch": 0.14482770801321113, "grad_norm": 3.7678370475769043, "learning_rate": 9.981713962887859e-05, "loss": 3.515, "step": 16400 }, { "epoch": 0.1452692559034953, "grad_norm": 2.7892682552337646, "learning_rate": 9.98159503302849e-05, "loss": 3.6017, "step": 16450 }, { "epoch": 0.1457108037937795, "grad_norm": 1.75071120262146, "learning_rate": 9.981475718383147e-05, "loss": 3.6968, "step": 16500 }, { "epoch": 0.14615235168406365, "grad_norm": 0.988599419593811, "learning_rate": 9.981356018961047e-05, "loss": 3.638, "step": 16550 }, { "epoch": 0.14659389957434785, "grad_norm": 3.183884382247925, "learning_rate": 9.981235934771436e-05, "loss": 3.4087, "step": 16600 }, { "epoch": 0.147035447464632, "grad_norm": 2.3756110668182373, "learning_rate": 9.981115465823587e-05, "loss": 3.493, "step": 16650 }, { "epoch": 0.1474769953549162, "grad_norm": 1.6157623529434204, "learning_rate": 9.980994612126807e-05, "loss": 3.4459, "step": 16700 }, { "epoch": 0.14791854324520037, "grad_norm": 3.4954352378845215, "learning_rate": 9.980873373690431e-05, "loss": 3.6551, "step": 16750 }, { "epoch": 0.14836009113548457, "grad_norm": 2.129396915435791, "learning_rate": 9.980751750523825e-05, "loss": 3.5101, "step": 16800 }, { "epoch": 0.14880163902576873, "grad_norm": 1.6230214834213257, "learning_rate": 9.98062974263638e-05, "loss": 3.3445, "step": 16850 }, { "epoch": 0.14924318691605293, "grad_norm": 2.7362122535705566, "learning_rate": 9.980507350037522e-05, "loss": 3.4352, "step": 16900 }, { "epoch": 0.1496847348063371, "grad_norm": 4.6118364334106445, "learning_rate": 9.980384572736706e-05, "loss": 2.9692, "step": 16950 }, { "epoch": 0.1501262826966213, "grad_norm": 3.5687484741210938, "learning_rate": 9.980261410743414e-05, "loss": 3.669, "step": 17000 }, { "epoch": 0.15056783058690545, "grad_norm": 1.4324193000793457, "learning_rate": 9.980137864067158e-05, "loss": 3.0393, "step": 17050 }, { "epoch": 0.15100937847718965, "grad_norm": 2.652137517929077, "learning_rate": 9.980013932717484e-05, "loss": 3.6648, "step": 17100 }, { "epoch": 0.15145092636747381, "grad_norm": 7.182887554168701, "learning_rate": 9.97988961670396e-05, "loss": 3.3025, "step": 17150 }, { "epoch": 0.151892474257758, "grad_norm": 2.4190828800201416, "learning_rate": 9.979764916036195e-05, "loss": 3.5013, "step": 17200 }, { "epoch": 0.15233402214804218, "grad_norm": 4.525496006011963, "learning_rate": 9.979639830723816e-05, "loss": 3.4685, "step": 17250 }, { "epoch": 0.15277557003832637, "grad_norm": 1.242196798324585, "learning_rate": 9.979514360776486e-05, "loss": 3.3286, "step": 17300 }, { "epoch": 0.15321711792861054, "grad_norm": 4.049784183502197, "learning_rate": 9.979388506203897e-05, "loss": 3.1067, "step": 17350 }, { "epoch": 0.15365866581889473, "grad_norm": 2.2166428565979004, "learning_rate": 9.97926226701577e-05, "loss": 3.1167, "step": 17400 }, { "epoch": 0.1541002137091789, "grad_norm": 1.4584290981292725, "learning_rate": 9.979135643221856e-05, "loss": 3.4869, "step": 17450 }, { "epoch": 0.1545417615994631, "grad_norm": 2.357682228088379, "learning_rate": 9.979008634831934e-05, "loss": 3.3837, "step": 17500 }, { "epoch": 0.15498330948974726, "grad_norm": 3.9010651111602783, "learning_rate": 9.978881241855817e-05, "loss": 3.0948, "step": 17550 }, { "epoch": 0.15542485738003145, "grad_norm": 2.2975761890411377, "learning_rate": 9.978753464303343e-05, "loss": 3.487, "step": 17600 }, { "epoch": 0.15586640527031562, "grad_norm": 1.2210371494293213, "learning_rate": 9.978625302184383e-05, "loss": 3.7763, "step": 17650 }, { "epoch": 0.1563079531605998, "grad_norm": 1.8714485168457031, "learning_rate": 9.978496755508836e-05, "loss": 2.9216, "step": 17700 }, { "epoch": 0.15674950105088398, "grad_norm": 3.4243338108062744, "learning_rate": 9.978367824286629e-05, "loss": 3.5642, "step": 17750 }, { "epoch": 0.15719104894116817, "grad_norm": 1.6219075918197632, "learning_rate": 9.978238508527727e-05, "loss": 3.3517, "step": 17800 }, { "epoch": 0.15763259683145234, "grad_norm": 0.7262994647026062, "learning_rate": 9.97810880824211e-05, "loss": 2.9789, "step": 17850 }, { "epoch": 0.15807414472173653, "grad_norm": 2.3567357063293457, "learning_rate": 9.977981328904049e-05, "loss": 3.4182, "step": 17900 }, { "epoch": 0.1585156926120207, "grad_norm": 2.0587501525878906, "learning_rate": 9.977850867285131e-05, "loss": 3.4566, "step": 17950 }, { "epoch": 0.1589572405023049, "grad_norm": 1.3674755096435547, "learning_rate": 9.977720021169445e-05, "loss": 3.4426, "step": 18000 }, { "epoch": 0.1589572405023049, "eval_asr_loss": 0.8684477998103262, "eval_loss": 3.168827772140503, "eval_runtime": 20.2426, "eval_samples_per_second": 37.94, "eval_steps_per_second": 9.485, "eval_tts_loss": 5.975099919698394, "step": 18000 }, { "epoch": 0.15939878839258906, "grad_norm": 1.7318178415298462, "learning_rate": 9.977588790567097e-05, "loss": 3.1039, "step": 18050 }, { "epoch": 0.15984033628287325, "grad_norm": 4.160914897918701, "learning_rate": 9.977457175488225e-05, "loss": 3.4418, "step": 18100 }, { "epoch": 0.16028188417315742, "grad_norm": 1.6405212879180908, "learning_rate": 9.977325175942992e-05, "loss": 3.537, "step": 18150 }, { "epoch": 0.1607234320634416, "grad_norm": 3.2284727096557617, "learning_rate": 9.977192791941596e-05, "loss": 3.5699, "step": 18200 }, { "epoch": 0.16116497995372578, "grad_norm": 6.879377841949463, "learning_rate": 9.977060023494263e-05, "loss": 3.3869, "step": 18250 }, { "epoch": 0.16160652784400997, "grad_norm": 5.974034309387207, "learning_rate": 9.976926870611247e-05, "loss": 3.3198, "step": 18300 }, { "epoch": 0.16204807573429414, "grad_norm": 1.0853605270385742, "learning_rate": 9.976793333302834e-05, "loss": 3.2399, "step": 18350 }, { "epoch": 0.16248962362457833, "grad_norm": 2.7796192169189453, "learning_rate": 9.976659411579337e-05, "loss": 3.3726, "step": 18400 }, { "epoch": 0.1629311715148625, "grad_norm": 2.083576202392578, "learning_rate": 9.976525105451102e-05, "loss": 3.3063, "step": 18450 }, { "epoch": 0.1633727194051467, "grad_norm": 1.6082487106323242, "learning_rate": 9.976390414928501e-05, "loss": 3.2636, "step": 18500 }, { "epoch": 0.16381426729543086, "grad_norm": 2.3426663875579834, "learning_rate": 9.97625534002194e-05, "loss": 3.4482, "step": 18550 }, { "epoch": 0.16425581518571505, "grad_norm": 4.199110507965088, "learning_rate": 9.976119880741852e-05, "loss": 3.5328, "step": 18600 }, { "epoch": 0.16469736307599922, "grad_norm": 2.411940097808838, "learning_rate": 9.975984037098698e-05, "loss": 3.1049, "step": 18650 }, { "epoch": 0.16513891096628341, "grad_norm": 8.064169883728027, "learning_rate": 9.975847809102974e-05, "loss": 3.2858, "step": 18700 }, { "epoch": 0.16558045885656758, "grad_norm": 2.0680835247039795, "learning_rate": 9.9757111967652e-05, "loss": 3.1414, "step": 18750 }, { "epoch": 0.16602200674685177, "grad_norm": 2.7678062915802, "learning_rate": 9.975574200095927e-05, "loss": 3.1492, "step": 18800 }, { "epoch": 0.16646355463713594, "grad_norm": 3.402519941329956, "learning_rate": 9.975436819105742e-05, "loss": 3.1184, "step": 18850 }, { "epoch": 0.16690510252742013, "grad_norm": 1.3270809650421143, "learning_rate": 9.97529905380525e-05, "loss": 3.4959, "step": 18900 }, { "epoch": 0.1673466504177043, "grad_norm": 7.308635711669922, "learning_rate": 9.975160904205098e-05, "loss": 3.4175, "step": 18950 }, { "epoch": 0.1677881983079885, "grad_norm": 1.5945827960968018, "learning_rate": 9.975022370315952e-05, "loss": 3.2295, "step": 19000 }, { "epoch": 0.16822974619827266, "grad_norm": 4.1863861083984375, "learning_rate": 9.974883452148517e-05, "loss": 3.1561, "step": 19050 }, { "epoch": 0.16867129408855686, "grad_norm": 1.7289392948150635, "learning_rate": 9.97474414971352e-05, "loss": 3.5214, "step": 19100 }, { "epoch": 0.16911284197884102, "grad_norm": 1.1129475831985474, "learning_rate": 9.974604463021722e-05, "loss": 2.9315, "step": 19150 }, { "epoch": 0.16955438986912522, "grad_norm": 1.6852185726165771, "learning_rate": 9.974464392083913e-05, "loss": 3.4969, "step": 19200 }, { "epoch": 0.16999593775940938, "grad_norm": 6.104635238647461, "learning_rate": 9.974323936910913e-05, "loss": 3.6542, "step": 19250 }, { "epoch": 0.17043748564969358, "grad_norm": 2.2778172492980957, "learning_rate": 9.974183097513567e-05, "loss": 3.1306, "step": 19300 }, { "epoch": 0.17087903353997774, "grad_norm": 1.4735571146011353, "learning_rate": 9.974041873902757e-05, "loss": 3.654, "step": 19350 }, { "epoch": 0.17132058143026194, "grad_norm": 3.6428987979888916, "learning_rate": 9.973900266089393e-05, "loss": 3.7674, "step": 19400 }, { "epoch": 0.1717621293205461, "grad_norm": 3.5552337169647217, "learning_rate": 9.97375827408441e-05, "loss": 3.5145, "step": 19450 }, { "epoch": 0.1722036772108303, "grad_norm": 3.519054651260376, "learning_rate": 9.973615897898777e-05, "loss": 3.1466, "step": 19500 }, { "epoch": 0.17264522510111446, "grad_norm": 5.785290241241455, "learning_rate": 9.97347313754349e-05, "loss": 3.207, "step": 19550 }, { "epoch": 0.17308677299139866, "grad_norm": 1.1869070529937744, "learning_rate": 9.973329993029577e-05, "loss": 3.4333, "step": 19600 }, { "epoch": 0.17352832088168282, "grad_norm": 1.620424747467041, "learning_rate": 9.973186464368096e-05, "loss": 3.4715, "step": 19650 }, { "epoch": 0.17396986877196702, "grad_norm": 3.2140984535217285, "learning_rate": 9.973042551570131e-05, "loss": 3.5503, "step": 19700 }, { "epoch": 0.17441141666225118, "grad_norm": 8.555562973022461, "learning_rate": 9.972898254646799e-05, "loss": 3.9534, "step": 19750 }, { "epoch": 0.17485296455253538, "grad_norm": 2.755021095275879, "learning_rate": 9.972753573609245e-05, "loss": 3.2534, "step": 19800 }, { "epoch": 0.17529451244281954, "grad_norm": 2.5277762413024902, "learning_rate": 9.972608508468646e-05, "loss": 3.9156, "step": 19850 }, { "epoch": 0.17573606033310374, "grad_norm": 6.709779739379883, "learning_rate": 9.972463059236207e-05, "loss": 3.3525, "step": 19900 }, { "epoch": 0.1761776082233879, "grad_norm": 2.41070294380188, "learning_rate": 9.972317225923161e-05, "loss": 3.3441, "step": 19950 }, { "epoch": 0.1766191561136721, "grad_norm": 1.5217126607894897, "learning_rate": 9.972171008540774e-05, "loss": 3.2966, "step": 20000 }, { "epoch": 0.17706070400395627, "grad_norm": 5.1036763191223145, "learning_rate": 9.972024407100337e-05, "loss": 3.1964, "step": 20050 }, { "epoch": 0.17750225189424046, "grad_norm": 1.291801929473877, "learning_rate": 9.97187742161318e-05, "loss": 3.6654, "step": 20100 }, { "epoch": 0.17794379978452463, "grad_norm": 1.6413884162902832, "learning_rate": 9.97173005209065e-05, "loss": 3.5594, "step": 20150 }, { "epoch": 0.17838534767480882, "grad_norm": 2.1488759517669678, "learning_rate": 9.971582298544133e-05, "loss": 3.1799, "step": 20200 }, { "epoch": 0.17882689556509299, "grad_norm": 4.286724090576172, "learning_rate": 9.971434160985041e-05, "loss": 3.1923, "step": 20250 }, { "epoch": 0.17926844345537718, "grad_norm": 2.672957420349121, "learning_rate": 9.971285639424817e-05, "loss": 3.216, "step": 20300 }, { "epoch": 0.17970999134566135, "grad_norm": 1.8457285165786743, "learning_rate": 9.971136733874931e-05, "loss": 3.0968, "step": 20350 }, { "epoch": 0.18015153923594554, "grad_norm": 2.780160665512085, "learning_rate": 9.970987444346889e-05, "loss": 3.7793, "step": 20400 }, { "epoch": 0.1805930871262297, "grad_norm": 4.923703670501709, "learning_rate": 9.970837770852218e-05, "loss": 3.0386, "step": 20450 }, { "epoch": 0.1810346350165139, "grad_norm": 6.101312637329102, "learning_rate": 9.97068771340248e-05, "loss": 3.6275, "step": 20500 }, { "epoch": 0.18147618290679807, "grad_norm": 1.9402213096618652, "learning_rate": 9.970537272009269e-05, "loss": 3.0918, "step": 20550 }, { "epoch": 0.18191773079708226, "grad_norm": 0.9089447259902954, "learning_rate": 9.970386446684201e-05, "loss": 3.5084, "step": 20600 }, { "epoch": 0.18235927868736643, "grad_norm": 0.5540192723274231, "learning_rate": 9.970235237438927e-05, "loss": 3.3147, "step": 20650 }, { "epoch": 0.18280082657765062, "grad_norm": 3.3821587562561035, "learning_rate": 9.970083644285129e-05, "loss": 3.7705, "step": 20700 }, { "epoch": 0.1832423744679348, "grad_norm": 2.602141857147217, "learning_rate": 9.969931667234512e-05, "loss": 3.231, "step": 20750 }, { "epoch": 0.18368392235821898, "grad_norm": 2.034205198287964, "learning_rate": 9.969779306298818e-05, "loss": 3.5487, "step": 20800 }, { "epoch": 0.18412547024850315, "grad_norm": 2.844647169113159, "learning_rate": 9.969626561489817e-05, "loss": 3.5581, "step": 20850 }, { "epoch": 0.18456701813878734, "grad_norm": 1.6377545595169067, "learning_rate": 9.969473432819305e-05, "loss": 3.8547, "step": 20900 }, { "epoch": 0.1850085660290715, "grad_norm": 7.625864505767822, "learning_rate": 9.96931992029911e-05, "loss": 3.1848, "step": 20950 }, { "epoch": 0.1854501139193557, "grad_norm": 1.9800269603729248, "learning_rate": 9.969166023941089e-05, "loss": 3.1796, "step": 21000 }, { "epoch": 0.1854501139193557, "eval_asr_loss": 0.8812751858267902, "eval_loss": 3.14119029045105, "eval_runtime": 19.8885, "eval_samples_per_second": 38.615, "eval_steps_per_second": 9.654, "eval_tts_loss": 5.95043236252646, "step": 21000 }, { "epoch": 0.18589166180963987, "grad_norm": 0.8084915280342102, "learning_rate": 9.969011743757131e-05, "loss": 3.1867, "step": 21050 }, { "epoch": 0.18633320969992406, "grad_norm": 4.80615234375, "learning_rate": 9.968857079759151e-05, "loss": 3.3244, "step": 21100 }, { "epoch": 0.18677475759020823, "grad_norm": 1.4828674793243408, "learning_rate": 9.968702031959096e-05, "loss": 2.8393, "step": 21150 }, { "epoch": 0.18721630548049242, "grad_norm": 3.2170047760009766, "learning_rate": 9.968546600368942e-05, "loss": 3.3566, "step": 21200 }, { "epoch": 0.1876578533707766, "grad_norm": 2.004711151123047, "learning_rate": 9.968390785000696e-05, "loss": 3.5387, "step": 21250 }, { "epoch": 0.18809940126106078, "grad_norm": 2.858024835586548, "learning_rate": 9.968234585866393e-05, "loss": 3.4617, "step": 21300 }, { "epoch": 0.18854094915134495, "grad_norm": 2.3722898960113525, "learning_rate": 9.968078002978097e-05, "loss": 3.3284, "step": 21350 }, { "epoch": 0.18898249704162914, "grad_norm": 3.7834815979003906, "learning_rate": 9.967921036347904e-05, "loss": 3.2961, "step": 21400 }, { "epoch": 0.1894240449319133, "grad_norm": 1.9320305585861206, "learning_rate": 9.967763685987937e-05, "loss": 3.0897, "step": 21450 }, { "epoch": 0.1898655928221975, "grad_norm": 2.595550537109375, "learning_rate": 9.96760595191035e-05, "loss": 3.1881, "step": 21500 }, { "epoch": 0.19030714071248167, "grad_norm": 2.7209346294403076, "learning_rate": 9.967447834127328e-05, "loss": 3.9049, "step": 21550 }, { "epoch": 0.19074868860276586, "grad_norm": 2.9117844104766846, "learning_rate": 9.967289332651085e-05, "loss": 3.7096, "step": 21600 }, { "epoch": 0.19119023649305003, "grad_norm": 3.910952091217041, "learning_rate": 9.96713044749386e-05, "loss": 3.2424, "step": 21650 }, { "epoch": 0.19163178438333422, "grad_norm": 0.9400450587272644, "learning_rate": 9.96697117866793e-05, "loss": 2.9279, "step": 21700 }, { "epoch": 0.1920733322736184, "grad_norm": 3.050595760345459, "learning_rate": 9.966811526185594e-05, "loss": 3.4662, "step": 21750 }, { "epoch": 0.19251488016390259, "grad_norm": 4.441404819488525, "learning_rate": 9.966651490059184e-05, "loss": 3.0896, "step": 21800 }, { "epoch": 0.19295642805418675, "grad_norm": 3.2662458419799805, "learning_rate": 9.966491070301064e-05, "loss": 3.1643, "step": 21850 }, { "epoch": 0.19339797594447095, "grad_norm": 1.5034865140914917, "learning_rate": 9.966330266923623e-05, "loss": 3.1728, "step": 21900 }, { "epoch": 0.1938395238347551, "grad_norm": 1.3567650318145752, "learning_rate": 9.966169079939282e-05, "loss": 3.433, "step": 21950 }, { "epoch": 0.1942810717250393, "grad_norm": 2.380711793899536, "learning_rate": 9.966007509360492e-05, "loss": 3.3513, "step": 22000 }, { "epoch": 0.19472261961532347, "grad_norm": 1.819464921951294, "learning_rate": 9.965845555199732e-05, "loss": 3.0924, "step": 22050 }, { "epoch": 0.19516416750560767, "grad_norm": 1.84955894947052, "learning_rate": 9.965683217469513e-05, "loss": 3.3711, "step": 22100 }, { "epoch": 0.19560571539589183, "grad_norm": 2.3455941677093506, "learning_rate": 9.965520496182371e-05, "loss": 3.5974, "step": 22150 }, { "epoch": 0.19604726328617603, "grad_norm": 1.5657644271850586, "learning_rate": 9.965357391350877e-05, "loss": 3.5424, "step": 22200 }, { "epoch": 0.1964888111764602, "grad_norm": 2.460850477218628, "learning_rate": 9.965193902987632e-05, "loss": 3.4851, "step": 22250 }, { "epoch": 0.1969303590667444, "grad_norm": 4.1430768966674805, "learning_rate": 9.96503003110526e-05, "loss": 3.6162, "step": 22300 }, { "epoch": 0.19737190695702855, "grad_norm": 1.511238932609558, "learning_rate": 9.964865775716421e-05, "loss": 3.4041, "step": 22350 }, { "epoch": 0.19781345484731275, "grad_norm": 5.789384841918945, "learning_rate": 9.964701136833801e-05, "loss": 3.6231, "step": 22400 }, { "epoch": 0.1982550027375969, "grad_norm": 1.697780728340149, "learning_rate": 9.964536114470118e-05, "loss": 3.5089, "step": 22450 }, { "epoch": 0.1986965506278811, "grad_norm": 1.9673670530319214, "learning_rate": 9.964370708638118e-05, "loss": 3.1656, "step": 22500 }, { "epoch": 0.19913809851816527, "grad_norm": 1.157395601272583, "learning_rate": 9.964204919350577e-05, "loss": 3.5386, "step": 22550 }, { "epoch": 0.19957964640844947, "grad_norm": 1.872974157333374, "learning_rate": 9.964038746620303e-05, "loss": 3.8252, "step": 22600 }, { "epoch": 0.20002119429873363, "grad_norm": 2.0033257007598877, "learning_rate": 9.963872190460128e-05, "loss": 3.0679, "step": 22650 }, { "epoch": 0.20046274218901783, "grad_norm": 1.9374537467956543, "learning_rate": 9.963705250882919e-05, "loss": 3.4905, "step": 22700 }, { "epoch": 0.200904290079302, "grad_norm": 1.368355393409729, "learning_rate": 9.963537927901571e-05, "loss": 3.1261, "step": 22750 }, { "epoch": 0.2013458379695862, "grad_norm": 2.8965837955474854, "learning_rate": 9.963370221529007e-05, "loss": 3.434, "step": 22800 }, { "epoch": 0.20178738585987036, "grad_norm": 4.269700527191162, "learning_rate": 9.96320213177818e-05, "loss": 3.3089, "step": 22850 }, { "epoch": 0.20222893375015455, "grad_norm": 2.0559756755828857, "learning_rate": 9.963033658662078e-05, "loss": 3.0824, "step": 22900 }, { "epoch": 0.20267048164043872, "grad_norm": 2.692986011505127, "learning_rate": 9.962868183079845e-05, "loss": 2.9178, "step": 22950 }, { "epoch": 0.2031120295307229, "grad_norm": 2.5358245372772217, "learning_rate": 9.962698950938911e-05, "loss": 3.8775, "step": 23000 }, { "epoch": 0.20355357742100708, "grad_norm": 2.653918504714966, "learning_rate": 9.962529335471565e-05, "loss": 3.9464, "step": 23050 }, { "epoch": 0.20399512531129127, "grad_norm": 2.1768572330474854, "learning_rate": 9.962362740422908e-05, "loss": 3.3436, "step": 23100 }, { "epoch": 0.20443667320157544, "grad_norm": 1.4818288087844849, "learning_rate": 9.962192366007948e-05, "loss": 3.3089, "step": 23150 }, { "epoch": 0.20487822109185963, "grad_norm": 8.97103500366211, "learning_rate": 9.962021608305707e-05, "loss": 3.2528, "step": 23200 }, { "epoch": 0.2053197689821438, "grad_norm": 2.6199610233306885, "learning_rate": 9.961850467329372e-05, "loss": 3.3929, "step": 23250 }, { "epoch": 0.205761316872428, "grad_norm": 1.7377407550811768, "learning_rate": 9.961678943092164e-05, "loss": 3.4344, "step": 23300 }, { "epoch": 0.20620286476271216, "grad_norm": 1.5331953763961792, "learning_rate": 9.961507035607332e-05, "loss": 3.6088, "step": 23350 }, { "epoch": 0.20664441265299635, "grad_norm": 5.9846014976501465, "learning_rate": 9.961334744888154e-05, "loss": 3.5139, "step": 23400 }, { "epoch": 0.20708596054328052, "grad_norm": 1.6636093854904175, "learning_rate": 9.961162070947936e-05, "loss": 3.783, "step": 23450 }, { "epoch": 0.2075275084335647, "grad_norm": 4.730373382568359, "learning_rate": 9.960989013800019e-05, "loss": 3.5666, "step": 23500 }, { "epoch": 0.20796905632384888, "grad_norm": 1.7042059898376465, "learning_rate": 9.960815573457769e-05, "loss": 3.4093, "step": 23550 }, { "epoch": 0.20841060421413307, "grad_norm": 1.6063077449798584, "learning_rate": 9.96064174993458e-05, "loss": 2.9415, "step": 23600 }, { "epoch": 0.20885215210441724, "grad_norm": 2.5845298767089844, "learning_rate": 9.960467543243884e-05, "loss": 3.421, "step": 23650 }, { "epoch": 0.20929369999470143, "grad_norm": 2.843557119369507, "learning_rate": 9.960292953399131e-05, "loss": 3.7182, "step": 23700 }, { "epoch": 0.2097352478849856, "grad_norm": 2.7299513816833496, "learning_rate": 9.960117980413811e-05, "loss": 3.4932, "step": 23750 }, { "epoch": 0.2101767957752698, "grad_norm": 3.5709776878356934, "learning_rate": 9.959942624301437e-05, "loss": 3.4617, "step": 23800 }, { "epoch": 0.21061834366555396, "grad_norm": 1.636993408203125, "learning_rate": 9.959766885075556e-05, "loss": 3.2784, "step": 23850 }, { "epoch": 0.21105989155583815, "grad_norm": 1.2677909135818481, "learning_rate": 9.959590762749738e-05, "loss": 3.5762, "step": 23900 }, { "epoch": 0.21150143944612232, "grad_norm": 4.043791770935059, "learning_rate": 9.959414257337591e-05, "loss": 3.1917, "step": 23950 }, { "epoch": 0.2119429873364065, "grad_norm": 3.0787148475646973, "learning_rate": 9.959237368852747e-05, "loss": 3.307, "step": 24000 }, { "epoch": 0.2119429873364065, "eval_asr_loss": 0.8739951278052379, "eval_loss": 3.1280133724212646, "eval_runtime": 20.4078, "eval_samples_per_second": 37.633, "eval_steps_per_second": 9.408, "eval_tts_loss": 5.9267985776675065, "step": 24000 }, { "epoch": 0.21238453522669068, "grad_norm": 2.54976487159729, "learning_rate": 9.95906009730887e-05, "loss": 3.178, "step": 24050 }, { "epoch": 0.21282608311697487, "grad_norm": 1.0521984100341797, "learning_rate": 9.958882442719652e-05, "loss": 3.7995, "step": 24100 }, { "epoch": 0.21326763100725904, "grad_norm": 1.708702564239502, "learning_rate": 9.958704405098815e-05, "loss": 3.5033, "step": 24150 }, { "epoch": 0.21370917889754323, "grad_norm": 2.1780407428741455, "learning_rate": 9.958525984460112e-05, "loss": 3.4977, "step": 24200 }, { "epoch": 0.2141507267878274, "grad_norm": 2.227015972137451, "learning_rate": 9.958347180817323e-05, "loss": 2.9502, "step": 24250 }, { "epoch": 0.2145922746781116, "grad_norm": 1.5458627939224243, "learning_rate": 9.95816799418426e-05, "loss": 3.3266, "step": 24300 }, { "epoch": 0.21503382256839576, "grad_norm": 1.8348188400268555, "learning_rate": 9.957988424574764e-05, "loss": 3.3448, "step": 24350 }, { "epoch": 0.21547537045867995, "grad_norm": 2.4458820819854736, "learning_rate": 9.957808472002704e-05, "loss": 3.9597, "step": 24400 }, { "epoch": 0.21591691834896412, "grad_norm": 1.2153455018997192, "learning_rate": 9.957628136481981e-05, "loss": 3.0095, "step": 24450 }, { "epoch": 0.21635846623924831, "grad_norm": 1.6723811626434326, "learning_rate": 9.957447418026525e-05, "loss": 3.3562, "step": 24500 }, { "epoch": 0.21680001412953248, "grad_norm": 4.406776428222656, "learning_rate": 9.957266316650291e-05, "loss": 3.4466, "step": 24550 }, { "epoch": 0.21724156201981668, "grad_norm": 5.27102518081665, "learning_rate": 9.957084832367273e-05, "loss": 3.3489, "step": 24600 }, { "epoch": 0.21768310991010084, "grad_norm": 2.6601903438568115, "learning_rate": 9.956902965191486e-05, "loss": 3.3423, "step": 24650 }, { "epoch": 0.21812465780038504, "grad_norm": 1.5734961032867432, "learning_rate": 9.956720715136977e-05, "loss": 3.3207, "step": 24700 }, { "epoch": 0.2185662056906692, "grad_norm": 2.8826043605804443, "learning_rate": 9.956538082217826e-05, "loss": 3.1197, "step": 24750 }, { "epoch": 0.2190077535809534, "grad_norm": 1.6438381671905518, "learning_rate": 9.956355066448138e-05, "loss": 3.1748, "step": 24800 }, { "epoch": 0.21944930147123756, "grad_norm": 1.6108850240707397, "learning_rate": 9.956171667842048e-05, "loss": 3.7212, "step": 24850 }, { "epoch": 0.21989084936152176, "grad_norm": 1.1622767448425293, "learning_rate": 9.955987886413725e-05, "loss": 3.553, "step": 24900 }, { "epoch": 0.22033239725180592, "grad_norm": 1.478102207183838, "learning_rate": 9.955803722177363e-05, "loss": 3.2171, "step": 24950 }, { "epoch": 0.22077394514209012, "grad_norm": 2.2040581703186035, "learning_rate": 9.95561917514719e-05, "loss": 3.4424, "step": 25000 }, { "epoch": 0.22121549303237428, "grad_norm": 1.8141365051269531, "learning_rate": 9.955434245337457e-05, "loss": 3.6843, "step": 25050 }, { "epoch": 0.22165704092265848, "grad_norm": 1.6195813417434692, "learning_rate": 9.955248932762447e-05, "loss": 3.7377, "step": 25100 }, { "epoch": 0.22209858881294264, "grad_norm": 2.677534580230713, "learning_rate": 9.955063237436477e-05, "loss": 3.5157, "step": 25150 }, { "epoch": 0.22254013670322684, "grad_norm": 1.7741618156433105, "learning_rate": 9.954880884685869e-05, "loss": 3.4256, "step": 25200 }, { "epoch": 0.222981684593511, "grad_norm": 0.9014108777046204, "learning_rate": 9.954694431555343e-05, "loss": 3.5028, "step": 25250 }, { "epoch": 0.2234232324837952, "grad_norm": 2.5479094982147217, "learning_rate": 9.954507595716685e-05, "loss": 3.7392, "step": 25300 }, { "epoch": 0.22386478037407936, "grad_norm": 1.6607797145843506, "learning_rate": 9.954320377184328e-05, "loss": 3.7739, "step": 25350 }, { "epoch": 0.22430632826436356, "grad_norm": 2.8050482273101807, "learning_rate": 9.954132775972735e-05, "loss": 3.3908, "step": 25400 }, { "epoch": 0.22474787615464772, "grad_norm": 0.8809208869934082, "learning_rate": 9.953944792096392e-05, "loss": 3.4604, "step": 25450 }, { "epoch": 0.22518942404493192, "grad_norm": 4.810263156890869, "learning_rate": 9.953756425569824e-05, "loss": 3.3016, "step": 25500 }, { "epoch": 0.22563097193521608, "grad_norm": 0.7430752515792847, "learning_rate": 9.95356767640758e-05, "loss": 3.3905, "step": 25550 }, { "epoch": 0.22607251982550028, "grad_norm": 1.2698413133621216, "learning_rate": 9.953378544624237e-05, "loss": 3.4881, "step": 25600 }, { "epoch": 0.22651406771578445, "grad_norm": 2.968998908996582, "learning_rate": 9.953189030234405e-05, "loss": 3.2608, "step": 25650 }, { "epoch": 0.22695561560606864, "grad_norm": 8.66765308380127, "learning_rate": 9.95299913325272e-05, "loss": 3.1594, "step": 25700 }, { "epoch": 0.2273971634963528, "grad_norm": 3.061037302017212, "learning_rate": 9.952808853693856e-05, "loss": 3.5046, "step": 25750 }, { "epoch": 0.227838711386637, "grad_norm": 1.6408863067626953, "learning_rate": 9.952618191572503e-05, "loss": 2.9974, "step": 25800 }, { "epoch": 0.22828025927692117, "grad_norm": 1.7766352891921997, "learning_rate": 9.952427146903393e-05, "loss": 3.4394, "step": 25850 }, { "epoch": 0.22872180716720536, "grad_norm": 1.8550622463226318, "learning_rate": 9.952235719701283e-05, "loss": 3.2499, "step": 25900 }, { "epoch": 0.22916335505748953, "grad_norm": 4.626537799835205, "learning_rate": 9.952043909980955e-05, "loss": 3.3191, "step": 25950 }, { "epoch": 0.22960490294777372, "grad_norm": 1.894518256187439, "learning_rate": 9.951851717757228e-05, "loss": 3.2887, "step": 26000 }, { "epoch": 0.2300464508380579, "grad_norm": 2.2566466331481934, "learning_rate": 9.951659143044947e-05, "loss": 3.2224, "step": 26050 }, { "epoch": 0.23048799872834208, "grad_norm": 1.502895474433899, "learning_rate": 9.951466185858985e-05, "loss": 3.4682, "step": 26100 }, { "epoch": 0.23092954661862625, "grad_norm": 3.664829730987549, "learning_rate": 9.951272846214247e-05, "loss": 3.0882, "step": 26150 }, { "epoch": 0.23137109450891044, "grad_norm": 2.9891955852508545, "learning_rate": 9.951079124125668e-05, "loss": 3.1857, "step": 26200 }, { "epoch": 0.2318126423991946, "grad_norm": 1.8004250526428223, "learning_rate": 9.950885019608211e-05, "loss": 3.3859, "step": 26250 }, { "epoch": 0.2322541902894788, "grad_norm": 2.643892526626587, "learning_rate": 9.950690532676869e-05, "loss": 3.1025, "step": 26300 }, { "epoch": 0.23269573817976297, "grad_norm": 8.552020072937012, "learning_rate": 9.950495663346662e-05, "loss": 2.8521, "step": 26350 }, { "epoch": 0.23313728607004716, "grad_norm": 2.302180528640747, "learning_rate": 9.950300411632645e-05, "loss": 3.6808, "step": 26400 }, { "epoch": 0.23357883396033133, "grad_norm": 2.410003185272217, "learning_rate": 9.950104777549898e-05, "loss": 3.2536, "step": 26450 }, { "epoch": 0.23402038185061552, "grad_norm": 3.413940906524658, "learning_rate": 9.949908761113534e-05, "loss": 3.6206, "step": 26500 }, { "epoch": 0.2344619297408997, "grad_norm": 1.564780354499817, "learning_rate": 9.949712362338692e-05, "loss": 3.4719, "step": 26550 }, { "epoch": 0.23490347763118388, "grad_norm": 1.3204708099365234, "learning_rate": 9.94951558124054e-05, "loss": 3.95, "step": 26600 }, { "epoch": 0.23534502552146805, "grad_norm": 2.4696288108825684, "learning_rate": 9.949318417834283e-05, "loss": 3.8591, "step": 26650 }, { "epoch": 0.23578657341175224, "grad_norm": 1.5461256504058838, "learning_rate": 9.949120872135144e-05, "loss": 3.6145, "step": 26700 }, { "epoch": 0.2362281213020364, "grad_norm": 1.8961230516433716, "learning_rate": 9.948922944158388e-05, "loss": 3.4088, "step": 26750 }, { "epoch": 0.2366696691923206, "grad_norm": 3.098393201828003, "learning_rate": 9.9487246339193e-05, "loss": 3.1048, "step": 26800 }, { "epoch": 0.23711121708260477, "grad_norm": 0.8614824414253235, "learning_rate": 9.948525941433195e-05, "loss": 3.3338, "step": 26850 }, { "epoch": 0.23755276497288896, "grad_norm": 2.924774646759033, "learning_rate": 9.948326866715426e-05, "loss": 3.2607, "step": 26900 }, { "epoch": 0.23799431286317313, "grad_norm": 5.287550449371338, "learning_rate": 9.948127409781367e-05, "loss": 3.4998, "step": 26950 }, { "epoch": 0.23843586075345732, "grad_norm": 1.865439534187317, "learning_rate": 9.947927570646424e-05, "loss": 3.1788, "step": 27000 }, { "epoch": 0.23843586075345732, "eval_asr_loss": 0.8890385764932143, "eval_loss": 3.1024086475372314, "eval_runtime": 19.895, "eval_samples_per_second": 38.603, "eval_steps_per_second": 9.651, "eval_tts_loss": 5.952762275795868, "step": 27000 }, { "epoch": 0.2388774086437415, "grad_norm": 4.502477169036865, "learning_rate": 9.947727349326033e-05, "loss": 2.8551, "step": 27050 }, { "epoch": 0.23931895653402568, "grad_norm": 2.6308412551879883, "learning_rate": 9.94752674583566e-05, "loss": 3.4248, "step": 27100 }, { "epoch": 0.23976050442430985, "grad_norm": 1.3937596082687378, "learning_rate": 9.9473257601908e-05, "loss": 3.0961, "step": 27150 }, { "epoch": 0.24020205231459404, "grad_norm": 2.9952657222747803, "learning_rate": 9.947124392406977e-05, "loss": 3.7838, "step": 27200 }, { "epoch": 0.2406436002048782, "grad_norm": 3.2740750312805176, "learning_rate": 9.946922642499744e-05, "loss": 3.5869, "step": 27250 }, { "epoch": 0.2410851480951624, "grad_norm": 1.290229082107544, "learning_rate": 9.946720510484686e-05, "loss": 3.521, "step": 27300 }, { "epoch": 0.24152669598544657, "grad_norm": 0.997048020362854, "learning_rate": 9.946517996377417e-05, "loss": 3.1288, "step": 27350 }, { "epoch": 0.24196824387573077, "grad_norm": 1.3124210834503174, "learning_rate": 9.946315100193577e-05, "loss": 2.9812, "step": 27400 }, { "epoch": 0.24240979176601493, "grad_norm": 1.2291759252548218, "learning_rate": 9.946111821948839e-05, "loss": 3.3039, "step": 27450 }, { "epoch": 0.24285133965629913, "grad_norm": 1.4661563634872437, "learning_rate": 9.945908161658904e-05, "loss": 3.271, "step": 27500 }, { "epoch": 0.2432928875465833, "grad_norm": 2.156710624694824, "learning_rate": 9.945704119339506e-05, "loss": 3.7045, "step": 27550 }, { "epoch": 0.24373443543686749, "grad_norm": 3.8776209354400635, "learning_rate": 9.9454996950064e-05, "loss": 3.6893, "step": 27600 }, { "epoch": 0.24417598332715165, "grad_norm": 1.0397090911865234, "learning_rate": 9.945294888675381e-05, "loss": 3.5694, "step": 27650 }, { "epoch": 0.24461753121743585, "grad_norm": 6.586629867553711, "learning_rate": 9.945089700362266e-05, "loss": 3.4295, "step": 27700 }, { "epoch": 0.24505907910772, "grad_norm": 1.8651010990142822, "learning_rate": 9.944884130082905e-05, "loss": 3.4199, "step": 27750 }, { "epoch": 0.2455006269980042, "grad_norm": 2.351825475692749, "learning_rate": 9.944678177853176e-05, "loss": 3.88, "step": 27800 }, { "epoch": 0.24594217488828837, "grad_norm": 3.991497755050659, "learning_rate": 9.944471843688987e-05, "loss": 3.5451, "step": 27850 }, { "epoch": 0.24638372277857257, "grad_norm": 4.23520040512085, "learning_rate": 9.944265127606278e-05, "loss": 3.2976, "step": 27900 }, { "epoch": 0.24682527066885673, "grad_norm": 3.55438232421875, "learning_rate": 9.944058029621013e-05, "loss": 2.9787, "step": 27950 }, { "epoch": 0.24726681855914093, "grad_norm": 6.835021495819092, "learning_rate": 9.94385054974919e-05, "loss": 3.151, "step": 28000 }, { "epoch": 0.2477083664494251, "grad_norm": 2.232171058654785, "learning_rate": 9.943642688006834e-05, "loss": 3.353, "step": 28050 }, { "epoch": 0.2481499143397093, "grad_norm": 2.0550529956817627, "learning_rate": 9.94343444441e-05, "loss": 3.3, "step": 28100 }, { "epoch": 0.24859146222999345, "grad_norm": 0.9220274686813354, "learning_rate": 9.943225818974775e-05, "loss": 3.3238, "step": 28150 }, { "epoch": 0.24903301012027765, "grad_norm": 3.3017609119415283, "learning_rate": 9.943016811717275e-05, "loss": 3.652, "step": 28200 }, { "epoch": 0.24947455801056181, "grad_norm": 2.590731620788574, "learning_rate": 9.94280742265364e-05, "loss": 3.362, "step": 28250 }, { "epoch": 0.249916105900846, "grad_norm": 1.249603271484375, "learning_rate": 9.942597651800046e-05, "loss": 3.2159, "step": 28300 }, { "epoch": 0.2503576537911302, "grad_norm": 3.9774234294891357, "learning_rate": 9.942387499172694e-05, "loss": 3.2234, "step": 28350 }, { "epoch": 0.25079920168141434, "grad_norm": 1.4465439319610596, "learning_rate": 9.94217696478782e-05, "loss": 3.3785, "step": 28400 }, { "epoch": 0.25124074957169856, "grad_norm": 0.9342706799507141, "learning_rate": 9.941966048661682e-05, "loss": 3.1059, "step": 28450 }, { "epoch": 0.25168229746198273, "grad_norm": 2.685728073120117, "learning_rate": 9.941754750810573e-05, "loss": 3.5622, "step": 28500 }, { "epoch": 0.2521238453522669, "grad_norm": 6.555978775024414, "learning_rate": 9.941543071250816e-05, "loss": 3.3656, "step": 28550 }, { "epoch": 0.25256539324255106, "grad_norm": 0.8330821394920349, "learning_rate": 9.94133100999876e-05, "loss": 3.5714, "step": 28600 }, { "epoch": 0.2530069411328353, "grad_norm": 22.39084243774414, "learning_rate": 9.941118567070784e-05, "loss": 3.1974, "step": 28650 }, { "epoch": 0.25344848902311945, "grad_norm": 1.9922549724578857, "learning_rate": 9.940905742483298e-05, "loss": 3.5825, "step": 28700 }, { "epoch": 0.2538900369134036, "grad_norm": 1.42471444606781, "learning_rate": 9.94069253625274e-05, "loss": 3.3538, "step": 28750 }, { "epoch": 0.2543315848036878, "grad_norm": 3.5649797916412354, "learning_rate": 9.940478948395582e-05, "loss": 3.1705, "step": 28800 }, { "epoch": 0.254773132693972, "grad_norm": 2.5138847827911377, "learning_rate": 9.940264978928318e-05, "loss": 3.7867, "step": 28850 }, { "epoch": 0.25521468058425617, "grad_norm": 1.892633080482483, "learning_rate": 9.940050627867476e-05, "loss": 3.1689, "step": 28900 }, { "epoch": 0.25565622847454034, "grad_norm": 1.2871747016906738, "learning_rate": 9.939835895229615e-05, "loss": 3.651, "step": 28950 }, { "epoch": 0.2560977763648245, "grad_norm": 4.025623798370361, "learning_rate": 9.939620781031318e-05, "loss": 3.472, "step": 29000 }, { "epoch": 0.2565393242551087, "grad_norm": 2.834179162979126, "learning_rate": 9.939405285289203e-05, "loss": 3.3033, "step": 29050 }, { "epoch": 0.2569808721453929, "grad_norm": 2.194532871246338, "learning_rate": 9.939189408019916e-05, "loss": 3.3813, "step": 29100 }, { "epoch": 0.25742242003567706, "grad_norm": 2.8828024864196777, "learning_rate": 9.938973149240127e-05, "loss": 3.283, "step": 29150 }, { "epoch": 0.2578639679259612, "grad_norm": 3.8372597694396973, "learning_rate": 9.938756508966547e-05, "loss": 3.324, "step": 29200 }, { "epoch": 0.25830551581624545, "grad_norm": 1.0322990417480469, "learning_rate": 9.938539487215903e-05, "loss": 3.1929, "step": 29250 }, { "epoch": 0.2587470637065296, "grad_norm": 4.666669845581055, "learning_rate": 9.938322084004964e-05, "loss": 3.2645, "step": 29300 }, { "epoch": 0.2591886115968138, "grad_norm": 2.352522850036621, "learning_rate": 9.938104299350519e-05, "loss": 2.8695, "step": 29350 }, { "epoch": 0.25963015948709794, "grad_norm": 1.9813988208770752, "learning_rate": 9.93788613326939e-05, "loss": 3.0595, "step": 29400 }, { "epoch": 0.26007170737738217, "grad_norm": 4.507404804229736, "learning_rate": 9.93766758577843e-05, "loss": 3.3092, "step": 29450 }, { "epoch": 0.26051325526766633, "grad_norm": 0.7371171712875366, "learning_rate": 9.93744865689452e-05, "loss": 3.3464, "step": 29500 }, { "epoch": 0.2609548031579505, "grad_norm": 1.4071518182754517, "learning_rate": 9.937229346634566e-05, "loss": 2.8855, "step": 29550 }, { "epoch": 0.26139635104823467, "grad_norm": 2.9686851501464844, "learning_rate": 9.937009655015515e-05, "loss": 3.1077, "step": 29600 }, { "epoch": 0.2618378989385189, "grad_norm": 2.3262200355529785, "learning_rate": 9.936789582054331e-05, "loss": 3.1181, "step": 29650 }, { "epoch": 0.26227944682880305, "grad_norm": 1.0248990058898926, "learning_rate": 9.936569127768017e-05, "loss": 3.4325, "step": 29700 }, { "epoch": 0.2627209947190872, "grad_norm": 3.9115331172943115, "learning_rate": 9.936348292173596e-05, "loss": 3.132, "step": 29750 }, { "epoch": 0.2631625426093714, "grad_norm": 4.591248512268066, "learning_rate": 9.936127075288128e-05, "loss": 3.151, "step": 29800 }, { "epoch": 0.2636040904996556, "grad_norm": 2.473047971725464, "learning_rate": 9.935905477128703e-05, "loss": 3.4772, "step": 29850 }, { "epoch": 0.2640456383899398, "grad_norm": 1.3905609846115112, "learning_rate": 9.935683497712434e-05, "loss": 2.9894, "step": 29900 }, { "epoch": 0.26448718628022394, "grad_norm": 2.24664568901062, "learning_rate": 9.935461137056469e-05, "loss": 3.689, "step": 29950 }, { "epoch": 0.2649287341705081, "grad_norm": 1.1567890644073486, "learning_rate": 9.935238395177981e-05, "loss": 3.2017, "step": 30000 }, { "epoch": 0.2649287341705081, "eval_asr_loss": 0.8843846080425863, "eval_loss": 3.092066526412964, "eval_runtime": 19.8662, "eval_samples_per_second": 38.659, "eval_steps_per_second": 9.665, "eval_tts_loss": 5.902505457880168, "step": 30000 }, { "epoch": 0.26537028206079233, "grad_norm": 5.792656421661377, "learning_rate": 9.935015272094178e-05, "loss": 3.3117, "step": 30050 }, { "epoch": 0.2658118299510765, "grad_norm": 3.329699993133545, "learning_rate": 9.934791767822292e-05, "loss": 3.4779, "step": 30100 }, { "epoch": 0.26625337784136066, "grad_norm": 1.1682640314102173, "learning_rate": 9.934567882379589e-05, "loss": 3.0166, "step": 30150 }, { "epoch": 0.2666949257316448, "grad_norm": 2.778029441833496, "learning_rate": 9.93434361578336e-05, "loss": 3.209, "step": 30200 }, { "epoch": 0.26713647362192905, "grad_norm": 2.259307622909546, "learning_rate": 9.934118968050929e-05, "loss": 3.0044, "step": 30250 }, { "epoch": 0.2675780215122132, "grad_norm": 2.9219155311584473, "learning_rate": 9.933898443511526e-05, "loss": 3.7484, "step": 30300 }, { "epoch": 0.2680195694024974, "grad_norm": 1.4216759204864502, "learning_rate": 9.933673041180634e-05, "loss": 3.2729, "step": 30350 }, { "epoch": 0.26846111729278155, "grad_norm": 0.8821609020233154, "learning_rate": 9.933447257765338e-05, "loss": 2.7456, "step": 30400 }, { "epoch": 0.26890266518306577, "grad_norm": 4.904880046844482, "learning_rate": 9.933221093283074e-05, "loss": 3.2778, "step": 30450 }, { "epoch": 0.26934421307334994, "grad_norm": 1.9883630275726318, "learning_rate": 9.932994547751313e-05, "loss": 3.4769, "step": 30500 }, { "epoch": 0.2697857609636341, "grad_norm": 3.2121076583862305, "learning_rate": 9.932767621187555e-05, "loss": 3.2237, "step": 30550 }, { "epoch": 0.27022730885391827, "grad_norm": 0.9768474102020264, "learning_rate": 9.932540313609326e-05, "loss": 3.2586, "step": 30600 }, { "epoch": 0.2706688567442025, "grad_norm": 1.543562889099121, "learning_rate": 9.932312625034186e-05, "loss": 3.5258, "step": 30650 }, { "epoch": 0.27111040463448666, "grad_norm": 1.3274840116500854, "learning_rate": 9.932084555479719e-05, "loss": 3.2218, "step": 30700 }, { "epoch": 0.2715519525247708, "grad_norm": 3.6982388496398926, "learning_rate": 9.931856104963543e-05, "loss": 3.4994, "step": 30750 }, { "epoch": 0.271993500415055, "grad_norm": 1.6436798572540283, "learning_rate": 9.931627273503304e-05, "loss": 3.4035, "step": 30800 }, { "epoch": 0.2724350483053392, "grad_norm": 4.315684795379639, "learning_rate": 9.931398061116677e-05, "loss": 3.4837, "step": 30850 }, { "epoch": 0.2728765961956234, "grad_norm": 1.2069634199142456, "learning_rate": 9.931168467821367e-05, "loss": 3.2571, "step": 30900 }, { "epoch": 0.27331814408590754, "grad_norm": 3.137390375137329, "learning_rate": 9.930938493635109e-05, "loss": 3.2457, "step": 30950 }, { "epoch": 0.2737596919761917, "grad_norm": 1.6049838066101074, "learning_rate": 9.930708138575664e-05, "loss": 2.9232, "step": 31000 }, { "epoch": 0.27420123986647593, "grad_norm": 2.4024581909179688, "learning_rate": 9.930477402660827e-05, "loss": 3.1933, "step": 31050 }, { "epoch": 0.2746427877567601, "grad_norm": 2.070897340774536, "learning_rate": 9.93024628590842e-05, "loss": 3.1729, "step": 31100 }, { "epoch": 0.27508433564704426, "grad_norm": 2.37459397315979, "learning_rate": 9.930014788336293e-05, "loss": 3.0385, "step": 31150 }, { "epoch": 0.27552588353732843, "grad_norm": 3.5087156295776367, "learning_rate": 9.92978290996233e-05, "loss": 3.1936, "step": 31200 }, { "epoch": 0.27596743142761265, "grad_norm": 6.348201751708984, "learning_rate": 9.929550650804443e-05, "loss": 3.2907, "step": 31250 }, { "epoch": 0.2764089793178968, "grad_norm": 1.3216320276260376, "learning_rate": 9.929318010880566e-05, "loss": 3.1206, "step": 31300 }, { "epoch": 0.276850527208181, "grad_norm": 3.7011451721191406, "learning_rate": 9.929084990208674e-05, "loss": 3.6797, "step": 31350 }, { "epoch": 0.2772920750984652, "grad_norm": 1.47987961769104, "learning_rate": 9.928851588806763e-05, "loss": 3.2982, "step": 31400 }, { "epoch": 0.2777336229887494, "grad_norm": 2.7035892009735107, "learning_rate": 9.928617806692863e-05, "loss": 3.5924, "step": 31450 }, { "epoch": 0.27817517087903354, "grad_norm": 1.140507698059082, "learning_rate": 9.92838364388503e-05, "loss": 3.4756, "step": 31500 }, { "epoch": 0.2786167187693177, "grad_norm": 1.1196190118789673, "learning_rate": 9.928149100401354e-05, "loss": 3.5627, "step": 31550 }, { "epoch": 0.27905826665960193, "grad_norm": 3.444877862930298, "learning_rate": 9.927914176259947e-05, "loss": 3.3291, "step": 31600 }, { "epoch": 0.2794998145498861, "grad_norm": 4.410919189453125, "learning_rate": 9.927678871478958e-05, "loss": 3.1749, "step": 31650 }, { "epoch": 0.27994136244017026, "grad_norm": 1.771317481994629, "learning_rate": 9.927443186076563e-05, "loss": 2.6498, "step": 31700 }, { "epoch": 0.2803829103304544, "grad_norm": 2.1711716651916504, "learning_rate": 9.927207120070964e-05, "loss": 3.6787, "step": 31750 }, { "epoch": 0.28082445822073865, "grad_norm": 4.0384202003479, "learning_rate": 9.926970673480398e-05, "loss": 3.6298, "step": 31800 }, { "epoch": 0.2812660061110228, "grad_norm": 2.198716640472412, "learning_rate": 9.926733846323125e-05, "loss": 2.9091, "step": 31850 }, { "epoch": 0.281707554001307, "grad_norm": 1.819726586341858, "learning_rate": 9.92649663861744e-05, "loss": 3.5799, "step": 31900 }, { "epoch": 0.28214910189159115, "grad_norm": 1.7500370740890503, "learning_rate": 9.926259050381667e-05, "loss": 3.4091, "step": 31950 }, { "epoch": 0.28259064978187537, "grad_norm": 4.203965663909912, "learning_rate": 9.926021081634155e-05, "loss": 3.2101, "step": 32000 }, { "epoch": 0.28303219767215954, "grad_norm": 2.5612051486968994, "learning_rate": 9.925782732393285e-05, "loss": 3.0251, "step": 32050 }, { "epoch": 0.2834737455624437, "grad_norm": 1.427227258682251, "learning_rate": 9.92554400267747e-05, "loss": 3.2904, "step": 32100 }, { "epoch": 0.28391529345272787, "grad_norm": 1.6692140102386475, "learning_rate": 9.925304892505146e-05, "loss": 3.74, "step": 32150 }, { "epoch": 0.2843568413430121, "grad_norm": 1.287840485572815, "learning_rate": 9.925065401894786e-05, "loss": 3.5495, "step": 32200 }, { "epoch": 0.28479838923329626, "grad_norm": 1.882232427597046, "learning_rate": 9.924825530864887e-05, "loss": 3.2056, "step": 32250 }, { "epoch": 0.2852399371235804, "grad_norm": 1.2615700960159302, "learning_rate": 9.924585279433976e-05, "loss": 3.1918, "step": 32300 }, { "epoch": 0.2856814850138646, "grad_norm": 1.7634786367416382, "learning_rate": 9.924344647620612e-05, "loss": 3.5558, "step": 32350 }, { "epoch": 0.2861230329041488, "grad_norm": 1.2818759679794312, "learning_rate": 9.924103635443381e-05, "loss": 3.0904, "step": 32400 }, { "epoch": 0.286564580794433, "grad_norm": 3.8670778274536133, "learning_rate": 9.9238622429209e-05, "loss": 3.8008, "step": 32450 }, { "epoch": 0.28700612868471714, "grad_norm": 2.2975099086761475, "learning_rate": 9.923625309255875e-05, "loss": 2.9886, "step": 32500 }, { "epoch": 0.2874476765750013, "grad_norm": 3.461240530014038, "learning_rate": 9.923383163704833e-05, "loss": 3.3378, "step": 32550 }, { "epoch": 0.28788922446528553, "grad_norm": 1.2354512214660645, "learning_rate": 9.923140637864192e-05, "loss": 3.4829, "step": 32600 }, { "epoch": 0.2883307723555697, "grad_norm": 2.5937671661376953, "learning_rate": 9.922897731752683e-05, "loss": 3.3737, "step": 32650 }, { "epoch": 0.28877232024585386, "grad_norm": 2.7086663246154785, "learning_rate": 9.922654445389069e-05, "loss": 3.3247, "step": 32700 }, { "epoch": 0.28921386813613803, "grad_norm": 1.1576141119003296, "learning_rate": 9.922410778792142e-05, "loss": 3.3032, "step": 32750 }, { "epoch": 0.28965541602642225, "grad_norm": 4.800418853759766, "learning_rate": 9.922166731980725e-05, "loss": 3.7363, "step": 32800 }, { "epoch": 0.2900969639167064, "grad_norm": 1.6237969398498535, "learning_rate": 9.921922304973663e-05, "loss": 3.6136, "step": 32850 }, { "epoch": 0.2905385118069906, "grad_norm": 3.0943002700805664, "learning_rate": 9.921677497789843e-05, "loss": 3.18, "step": 32900 }, { "epoch": 0.29098005969727475, "grad_norm": 2.4920825958251953, "learning_rate": 9.921432310448169e-05, "loss": 3.2806, "step": 32950 }, { "epoch": 0.291421607587559, "grad_norm": 4.264760971069336, "learning_rate": 9.921186742967581e-05, "loss": 2.9155, "step": 33000 }, { "epoch": 0.291421607587559, "eval_asr_loss": 0.895976054242954, "eval_loss": 3.079402208328247, "eval_runtime": 20.0371, "eval_samples_per_second": 38.329, "eval_steps_per_second": 9.582, "eval_tts_loss": 5.968250579680352, "step": 33000 }, { "epoch": 0.29186315547784314, "grad_norm": 3.386139154434204, "learning_rate": 9.92094079536705e-05, "loss": 3.9219, "step": 33050 }, { "epoch": 0.2923047033681273, "grad_norm": 1.537395715713501, "learning_rate": 9.92069446766557e-05, "loss": 3.1577, "step": 33100 }, { "epoch": 0.29274625125841147, "grad_norm": 0.6773030757904053, "learning_rate": 9.920447759882168e-05, "loss": 2.9516, "step": 33150 }, { "epoch": 0.2931877991486957, "grad_norm": 2.179901599884033, "learning_rate": 9.9202006720359e-05, "loss": 3.7766, "step": 33200 }, { "epoch": 0.29362934703897986, "grad_norm": 4.384719371795654, "learning_rate": 9.919953204145853e-05, "loss": 3.4504, "step": 33250 }, { "epoch": 0.294070894929264, "grad_norm": 5.59550666809082, "learning_rate": 9.919705356231141e-05, "loss": 3.2456, "step": 33300 }, { "epoch": 0.2945124428195482, "grad_norm": 1.8736000061035156, "learning_rate": 9.919457128310909e-05, "loss": 3.5584, "step": 33350 }, { "epoch": 0.2949539907098324, "grad_norm": 2.3086678981781006, "learning_rate": 9.919208520404328e-05, "loss": 3.3485, "step": 33400 }, { "epoch": 0.2953955386001166, "grad_norm": 1.736769676208496, "learning_rate": 9.918959532530604e-05, "loss": 3.426, "step": 33450 }, { "epoch": 0.29583708649040075, "grad_norm": 1.8966106176376343, "learning_rate": 9.918710164708967e-05, "loss": 3.0041, "step": 33500 }, { "epoch": 0.2962786343806849, "grad_norm": 3.222245216369629, "learning_rate": 9.918460416958679e-05, "loss": 3.0005, "step": 33550 }, { "epoch": 0.29672018227096914, "grad_norm": 0.8299497365951538, "learning_rate": 9.918210289299031e-05, "loss": 3.3702, "step": 33600 }, { "epoch": 0.2971617301612533, "grad_norm": 3.876796245574951, "learning_rate": 9.917959781749345e-05, "loss": 3.243, "step": 33650 }, { "epoch": 0.29760327805153747, "grad_norm": 4.3070807456970215, "learning_rate": 9.917708894328968e-05, "loss": 3.4787, "step": 33700 }, { "epoch": 0.29804482594182163, "grad_norm": 2.306934356689453, "learning_rate": 9.917457627057279e-05, "loss": 3.2177, "step": 33750 }, { "epoch": 0.29848637383210586, "grad_norm": 3.8237075805664062, "learning_rate": 9.917205979953687e-05, "loss": 2.9533, "step": 33800 }, { "epoch": 0.29892792172239, "grad_norm": 1.627469778060913, "learning_rate": 9.91695395303763e-05, "loss": 3.3876, "step": 33850 }, { "epoch": 0.2993694696126742, "grad_norm": 1.0928808450698853, "learning_rate": 9.916701546328576e-05, "loss": 2.8597, "step": 33900 }, { "epoch": 0.29981101750295835, "grad_norm": 2.0172150135040283, "learning_rate": 9.916448759846017e-05, "loss": 3.6425, "step": 33950 }, { "epoch": 0.3002525653932426, "grad_norm": 2.5472891330718994, "learning_rate": 9.916195593609486e-05, "loss": 3.4216, "step": 34000 }, { "epoch": 0.30069411328352674, "grad_norm": 2.0118589401245117, "learning_rate": 9.91594712227922e-05, "loss": 3.1182, "step": 34050 }, { "epoch": 0.3011356611738109, "grad_norm": 1.4318723678588867, "learning_rate": 9.915693204187532e-05, "loss": 3.2758, "step": 34100 }, { "epoch": 0.3015772090640951, "grad_norm": 6.2676801681518555, "learning_rate": 9.915438906400228e-05, "loss": 3.2553, "step": 34150 }, { "epoch": 0.3020187569543793, "grad_norm": 1.1752573251724243, "learning_rate": 9.915184228936952e-05, "loss": 3.4345, "step": 34200 }, { "epoch": 0.30246030484466346, "grad_norm": 1.754859447479248, "learning_rate": 9.914929171817371e-05, "loss": 3.5339, "step": 34250 }, { "epoch": 0.30290185273494763, "grad_norm": 1.1092463731765747, "learning_rate": 9.914673735061191e-05, "loss": 3.7403, "step": 34300 }, { "epoch": 0.3033434006252318, "grad_norm": 2.3356130123138428, "learning_rate": 9.91441791868814e-05, "loss": 3.7242, "step": 34350 }, { "epoch": 0.303784948515516, "grad_norm": 1.9679780006408691, "learning_rate": 9.914161722717976e-05, "loss": 3.0423, "step": 34400 }, { "epoch": 0.3042264964058002, "grad_norm": 2.5788609981536865, "learning_rate": 9.91390514717049e-05, "loss": 3.3403, "step": 34450 }, { "epoch": 0.30466804429608435, "grad_norm": 5.141701698303223, "learning_rate": 9.913648192065503e-05, "loss": 3.0419, "step": 34500 }, { "epoch": 0.3051095921863685, "grad_norm": 0.9429007172584534, "learning_rate": 9.913390857422858e-05, "loss": 3.4227, "step": 34550 }, { "epoch": 0.30555114007665274, "grad_norm": 5.526501655578613, "learning_rate": 9.913133143262432e-05, "loss": 3.4853, "step": 34600 }, { "epoch": 0.3059926879669369, "grad_norm": 1.2435593605041504, "learning_rate": 9.912875049604135e-05, "loss": 3.1218, "step": 34650 }, { "epoch": 0.30643423585722107, "grad_norm": 3.752776861190796, "learning_rate": 9.912616576467899e-05, "loss": 3.29, "step": 34700 }, { "epoch": 0.30687578374750524, "grad_norm": 1.3690961599349976, "learning_rate": 9.91235772387369e-05, "loss": 3.5259, "step": 34750 }, { "epoch": 0.30731733163778946, "grad_norm": 5.097490310668945, "learning_rate": 9.912098491841504e-05, "loss": 3.4285, "step": 34800 }, { "epoch": 0.3077588795280736, "grad_norm": 2.029536485671997, "learning_rate": 9.911838880391362e-05, "loss": 3.5238, "step": 34850 }, { "epoch": 0.3082004274183578, "grad_norm": 1.1511303186416626, "learning_rate": 9.911578889543318e-05, "loss": 3.5973, "step": 34900 }, { "epoch": 0.30864197530864196, "grad_norm": 0.7825906872749329, "learning_rate": 9.911318519317454e-05, "loss": 3.5027, "step": 34950 }, { "epoch": 0.3090835231989262, "grad_norm": 1.2858946323394775, "learning_rate": 9.91105776973388e-05, "loss": 3.5128, "step": 35000 }, { "epoch": 0.30952507108921035, "grad_norm": 3.028378486633301, "learning_rate": 9.910796640812739e-05, "loss": 3.2618, "step": 35050 }, { "epoch": 0.3099666189794945, "grad_norm": 2.016479253768921, "learning_rate": 9.910535132574198e-05, "loss": 3.4658, "step": 35100 }, { "epoch": 0.3104081668697787, "grad_norm": 5.971153736114502, "learning_rate": 9.91027324503846e-05, "loss": 3.5681, "step": 35150 }, { "epoch": 0.3108497147600629, "grad_norm": 4.901987552642822, "learning_rate": 9.910010978225753e-05, "loss": 3.2022, "step": 35200 }, { "epoch": 0.31129126265034707, "grad_norm": 1.5224626064300537, "learning_rate": 9.90974833215633e-05, "loss": 3.4693, "step": 35250 }, { "epoch": 0.31173281054063123, "grad_norm": 1.7592462301254272, "learning_rate": 9.909485306850483e-05, "loss": 3.2146, "step": 35300 }, { "epoch": 0.3121743584309154, "grad_norm": 2.2695250511169434, "learning_rate": 9.909221902328529e-05, "loss": 3.3325, "step": 35350 }, { "epoch": 0.3126159063211996, "grad_norm": 1.2258449792861938, "learning_rate": 9.908958118610811e-05, "loss": 3.4704, "step": 35400 }, { "epoch": 0.3130574542114838, "grad_norm": 5.258679389953613, "learning_rate": 9.908693955717705e-05, "loss": 3.579, "step": 35450 }, { "epoch": 0.31349900210176795, "grad_norm": 1.850501298904419, "learning_rate": 9.908429413669615e-05, "loss": 3.2378, "step": 35500 }, { "epoch": 0.3139405499920521, "grad_norm": 1.2289330959320068, "learning_rate": 9.908164492486976e-05, "loss": 3.4111, "step": 35550 }, { "epoch": 0.31438209788233634, "grad_norm": 2.0499982833862305, "learning_rate": 9.907899192190249e-05, "loss": 3.6974, "step": 35600 }, { "epoch": 0.3148236457726205, "grad_norm": 1.0805225372314453, "learning_rate": 9.907633512799927e-05, "loss": 3.1798, "step": 35650 }, { "epoch": 0.3152651936629047, "grad_norm": 1.2690644264221191, "learning_rate": 9.907367454336533e-05, "loss": 2.9274, "step": 35700 }, { "epoch": 0.31570674155318884, "grad_norm": 5.509250164031982, "learning_rate": 9.907101016820616e-05, "loss": 3.1606, "step": 35750 }, { "epoch": 0.31614828944347306, "grad_norm": 3.2590670585632324, "learning_rate": 9.906834200272756e-05, "loss": 3.2878, "step": 35800 }, { "epoch": 0.31658983733375723, "grad_norm": 1.7757644653320312, "learning_rate": 9.906567004713563e-05, "loss": 3.3347, "step": 35850 }, { "epoch": 0.3170313852240414, "grad_norm": 6.056863784790039, "learning_rate": 9.906299430163675e-05, "loss": 3.3504, "step": 35900 }, { "epoch": 0.31747293311432556, "grad_norm": 1.0027168989181519, "learning_rate": 9.90603147664376e-05, "loss": 3.0737, "step": 35950 }, { "epoch": 0.3179144810046098, "grad_norm": 1.916014313697815, "learning_rate": 9.905763144174517e-05, "loss": 3.1648, "step": 36000 }, { "epoch": 0.3179144810046098, "eval_asr_loss": 0.8892677519164152, "eval_loss": 3.0999526977539062, "eval_runtime": 20.2917, "eval_samples_per_second": 37.848, "eval_steps_per_second": 9.462, "eval_tts_loss": 5.944016528967177, "step": 36000 }, { "epoch": 0.31835602889489395, "grad_norm": 2.1346523761749268, "learning_rate": 9.90549443277667e-05, "loss": 3.5172, "step": 36050 }, { "epoch": 0.3187975767851781, "grad_norm": 1.5300114154815674, "learning_rate": 9.905225342470974e-05, "loss": 3.2909, "step": 36100 }, { "epoch": 0.3192391246754623, "grad_norm": 1.7766307592391968, "learning_rate": 9.904955873278216e-05, "loss": 3.5821, "step": 36150 }, { "epoch": 0.3196806725657465, "grad_norm": 1.8257805109024048, "learning_rate": 9.904686025219211e-05, "loss": 3.3543, "step": 36200 }, { "epoch": 0.32012222045603067, "grad_norm": 2.9343011379241943, "learning_rate": 9.9044157983148e-05, "loss": 3.5759, "step": 36250 }, { "epoch": 0.32056376834631484, "grad_norm": 2.4320614337921143, "learning_rate": 9.904145192585857e-05, "loss": 3.2795, "step": 36300 }, { "epoch": 0.321005316236599, "grad_norm": 2.0806117057800293, "learning_rate": 9.903874208053283e-05, "loss": 2.9637, "step": 36350 }, { "epoch": 0.3214468641268832, "grad_norm": 10.894238471984863, "learning_rate": 9.903602844738012e-05, "loss": 3.5964, "step": 36400 }, { "epoch": 0.3218884120171674, "grad_norm": 1.5613411664962769, "learning_rate": 9.903331102661001e-05, "loss": 3.2076, "step": 36450 }, { "epoch": 0.32232995990745156, "grad_norm": 3.5824601650238037, "learning_rate": 9.903058981843242e-05, "loss": 3.3151, "step": 36500 }, { "epoch": 0.3227715077977357, "grad_norm": 4.121826171875, "learning_rate": 9.902786482305753e-05, "loss": 3.1602, "step": 36550 }, { "epoch": 0.32321305568801995, "grad_norm": 2.1794791221618652, "learning_rate": 9.902513604069583e-05, "loss": 3.3487, "step": 36600 }, { "epoch": 0.3236546035783041, "grad_norm": 3.4854958057403564, "learning_rate": 9.902240347155809e-05, "loss": 3.5573, "step": 36650 }, { "epoch": 0.3240961514685883, "grad_norm": 2.346065044403076, "learning_rate": 9.901966711585538e-05, "loss": 2.7336, "step": 36700 }, { "epoch": 0.32453769935887244, "grad_norm": 1.1362333297729492, "learning_rate": 9.901692697379905e-05, "loss": 3.5597, "step": 36750 }, { "epoch": 0.32497924724915667, "grad_norm": 1.6578530073165894, "learning_rate": 9.901418304560077e-05, "loss": 3.4961, "step": 36800 }, { "epoch": 0.32542079513944083, "grad_norm": 1.7271004915237427, "learning_rate": 9.901143533147248e-05, "loss": 3.1233, "step": 36850 }, { "epoch": 0.325862343029725, "grad_norm": 5.3198652267456055, "learning_rate": 9.900868383162643e-05, "loss": 3.6441, "step": 36900 }, { "epoch": 0.32630389092000917, "grad_norm": 1.4437021017074585, "learning_rate": 9.900592854627511e-05, "loss": 3.239, "step": 36950 }, { "epoch": 0.3267454388102934, "grad_norm": 3.09059739112854, "learning_rate": 9.900316947563138e-05, "loss": 3.3918, "step": 37000 }, { "epoch": 0.32718698670057755, "grad_norm": 1.7625781297683716, "learning_rate": 9.900040661990835e-05, "loss": 3.5584, "step": 37050 }, { "epoch": 0.3276285345908617, "grad_norm": 1.064771294593811, "learning_rate": 9.899763997931942e-05, "loss": 3.0205, "step": 37100 }, { "epoch": 0.3280700824811459, "grad_norm": 2.178464889526367, "learning_rate": 9.899486955407826e-05, "loss": 3.2899, "step": 37150 }, { "epoch": 0.3285116303714301, "grad_norm": 1.889468789100647, "learning_rate": 9.899209534439891e-05, "loss": 3.3759, "step": 37200 }, { "epoch": 0.3289531782617143, "grad_norm": 3.320777654647827, "learning_rate": 9.898931735049565e-05, "loss": 3.6612, "step": 37250 }, { "epoch": 0.32939472615199844, "grad_norm": 4.364492416381836, "learning_rate": 9.898653557258303e-05, "loss": 3.5972, "step": 37300 }, { "epoch": 0.3298362740422826, "grad_norm": 4.426956653594971, "learning_rate": 9.898375001087593e-05, "loss": 3.5658, "step": 37350 }, { "epoch": 0.33027782193256683, "grad_norm": 1.2732926607131958, "learning_rate": 9.898096066558951e-05, "loss": 3.1916, "step": 37400 }, { "epoch": 0.330719369822851, "grad_norm": 1.8885505199432373, "learning_rate": 9.897816753693922e-05, "loss": 3.6108, "step": 37450 }, { "epoch": 0.33116091771313516, "grad_norm": 5.467677116394043, "learning_rate": 9.897537062514081e-05, "loss": 3.079, "step": 37500 }, { "epoch": 0.3316024656034193, "grad_norm": 3.006726026535034, "learning_rate": 9.897256993041033e-05, "loss": 3.3902, "step": 37550 }, { "epoch": 0.33204401349370355, "grad_norm": 3.979788064956665, "learning_rate": 9.896976545296411e-05, "loss": 2.8985, "step": 37600 }, { "epoch": 0.3324855613839877, "grad_norm": 0.933592677116394, "learning_rate": 9.896695719301874e-05, "loss": 3.3717, "step": 37650 }, { "epoch": 0.3329271092742719, "grad_norm": 2.022643804550171, "learning_rate": 9.896414515079115e-05, "loss": 3.2212, "step": 37700 }, { "epoch": 0.33336865716455605, "grad_norm": 1.9231008291244507, "learning_rate": 9.896132932649857e-05, "loss": 3.0829, "step": 37750 }, { "epoch": 0.33381020505484027, "grad_norm": 6.672185897827148, "learning_rate": 9.895850972035847e-05, "loss": 3.0794, "step": 37800 }, { "epoch": 0.33425175294512444, "grad_norm": 2.566579580307007, "learning_rate": 9.895568633258865e-05, "loss": 3.2348, "step": 37850 }, { "epoch": 0.3346933008354086, "grad_norm": 3.8691163063049316, "learning_rate": 9.895285916340719e-05, "loss": 3.1823, "step": 37900 }, { "epoch": 0.33513484872569277, "grad_norm": 1.986522912979126, "learning_rate": 9.895002821303246e-05, "loss": 3.4586, "step": 37950 }, { "epoch": 0.335576396615977, "grad_norm": 1.734668254852295, "learning_rate": 9.894719348168315e-05, "loss": 3.7534, "step": 38000 }, { "epoch": 0.33601794450626116, "grad_norm": 1.9242331981658936, "learning_rate": 9.894435496957819e-05, "loss": 3.3395, "step": 38050 }, { "epoch": 0.3364594923965453, "grad_norm": 1.77113676071167, "learning_rate": 9.894151267693686e-05, "loss": 3.352, "step": 38100 }, { "epoch": 0.3369010402868295, "grad_norm": 0.799421489238739, "learning_rate": 9.893866660397867e-05, "loss": 3.7193, "step": 38150 }, { "epoch": 0.3373425881771137, "grad_norm": 1.236433982849121, "learning_rate": 9.89358167509235e-05, "loss": 3.5647, "step": 38200 }, { "epoch": 0.3377841360673979, "grad_norm": 3.1852049827575684, "learning_rate": 9.893296311799143e-05, "loss": 3.477, "step": 38250 }, { "epoch": 0.33822568395768204, "grad_norm": 2.2389414310455322, "learning_rate": 9.89301057054029e-05, "loss": 2.8368, "step": 38300 }, { "epoch": 0.3386672318479662, "grad_norm": 6.785423278808594, "learning_rate": 9.892724451337863e-05, "loss": 3.3759, "step": 38350 }, { "epoch": 0.33910877973825043, "grad_norm": 4.653458595275879, "learning_rate": 9.892437954213958e-05, "loss": 3.353, "step": 38400 }, { "epoch": 0.3395503276285346, "grad_norm": 3.1158320903778076, "learning_rate": 9.892151079190711e-05, "loss": 3.5522, "step": 38450 }, { "epoch": 0.33999187551881876, "grad_norm": 2.8249120712280273, "learning_rate": 9.891863826290277e-05, "loss": 3.3592, "step": 38500 }, { "epoch": 0.34043342340910293, "grad_norm": 2.011695384979248, "learning_rate": 9.891576195534842e-05, "loss": 3.3599, "step": 38550 }, { "epoch": 0.34087497129938715, "grad_norm": 1.8144574165344238, "learning_rate": 9.891288186946628e-05, "loss": 3.3011, "step": 38600 }, { "epoch": 0.3413165191896713, "grad_norm": 2.060272455215454, "learning_rate": 9.890999800547878e-05, "loss": 3.4336, "step": 38650 }, { "epoch": 0.3417580670799555, "grad_norm": 1.401672124862671, "learning_rate": 9.890711036360867e-05, "loss": 3.1762, "step": 38700 }, { "epoch": 0.34219961497023965, "grad_norm": 1.6355420351028442, "learning_rate": 9.8904218944079e-05, "loss": 3.6908, "step": 38750 }, { "epoch": 0.3426411628605239, "grad_norm": 2.591089963912964, "learning_rate": 9.890132374711312e-05, "loss": 3.3153, "step": 38800 }, { "epoch": 0.34308271075080804, "grad_norm": 2.8402106761932373, "learning_rate": 9.889842477293465e-05, "loss": 3.5123, "step": 38850 }, { "epoch": 0.3435242586410922, "grad_norm": 0.9350060820579529, "learning_rate": 9.889552202176751e-05, "loss": 3.3598, "step": 38900 }, { "epoch": 0.3439658065313764, "grad_norm": 2.9623401165008545, "learning_rate": 9.889261549383591e-05, "loss": 3.1874, "step": 38950 }, { "epoch": 0.3444073544216606, "grad_norm": 1.7671315670013428, "learning_rate": 9.888970518936438e-05, "loss": 3.5201, "step": 39000 }, { "epoch": 0.3444073544216606, "eval_asr_loss": 0.8851676637238344, "eval_loss": 3.074697494506836, "eval_runtime": 20.3913, "eval_samples_per_second": 37.663, "eval_steps_per_second": 9.416, "eval_tts_loss": 5.894229360095819, "step": 39000 }, { "epoch": 0.34484890231194476, "grad_norm": 1.1440892219543457, "learning_rate": 9.888679110857768e-05, "loss": 3.4325, "step": 39050 }, { "epoch": 0.3452904502022289, "grad_norm": 2.065777540206909, "learning_rate": 9.888387325170093e-05, "loss": 3.3255, "step": 39100 }, { "epoch": 0.3457319980925131, "grad_norm": 5.079590320587158, "learning_rate": 9.888095161895947e-05, "loss": 3.1736, "step": 39150 }, { "epoch": 0.3461735459827973, "grad_norm": 2.521078586578369, "learning_rate": 9.887802621057902e-05, "loss": 3.6232, "step": 39200 }, { "epoch": 0.3466150938730815, "grad_norm": 3.053846836090088, "learning_rate": 9.88750970267855e-05, "loss": 3.2633, "step": 39250 }, { "epoch": 0.34705664176336565, "grad_norm": 1.3491613864898682, "learning_rate": 9.88721640678052e-05, "loss": 3.7012, "step": 39300 }, { "epoch": 0.3474981896536498, "grad_norm": 1.5826070308685303, "learning_rate": 9.886922733386463e-05, "loss": 3.3255, "step": 39350 }, { "epoch": 0.34793973754393404, "grad_norm": 5.572771072387695, "learning_rate": 9.886628682519069e-05, "loss": 3.2976, "step": 39400 }, { "epoch": 0.3483812854342182, "grad_norm": 1.5839028358459473, "learning_rate": 9.886334254201041e-05, "loss": 3.3149, "step": 39450 }, { "epoch": 0.34882283332450237, "grad_norm": 3.295896053314209, "learning_rate": 9.88603944845513e-05, "loss": 2.9223, "step": 39500 }, { "epoch": 0.34926438121478653, "grad_norm": 2.2568423748016357, "learning_rate": 9.885744265304104e-05, "loss": 3.527, "step": 39550 }, { "epoch": 0.34970592910507076, "grad_norm": 0.9765694737434387, "learning_rate": 9.885448704770762e-05, "loss": 3.4622, "step": 39600 }, { "epoch": 0.3501474769953549, "grad_norm": 1.8776651620864868, "learning_rate": 9.885152766877936e-05, "loss": 3.327, "step": 39650 }, { "epoch": 0.3505890248856391, "grad_norm": 3.4537465572357178, "learning_rate": 9.884856451648483e-05, "loss": 3.5304, "step": 39700 }, { "epoch": 0.35103057277592326, "grad_norm": 2.3317794799804688, "learning_rate": 9.884559759105292e-05, "loss": 3.485, "step": 39750 }, { "epoch": 0.3514721206662075, "grad_norm": 2.5425727367401123, "learning_rate": 9.88426268927128e-05, "loss": 3.5912, "step": 39800 }, { "epoch": 0.35191366855649164, "grad_norm": 0.6102803945541382, "learning_rate": 9.883965242169392e-05, "loss": 3.0589, "step": 39850 }, { "epoch": 0.3523552164467758, "grad_norm": 2.142951726913452, "learning_rate": 9.883667417822604e-05, "loss": 3.4354, "step": 39900 }, { "epoch": 0.35279676433706, "grad_norm": 0.8856062293052673, "learning_rate": 9.883369216253919e-05, "loss": 3.3475, "step": 39950 }, { "epoch": 0.3532383122273442, "grad_norm": 1.8055260181427002, "learning_rate": 9.883070637486374e-05, "loss": 3.2718, "step": 40000 }, { "epoch": 0.35367986011762836, "grad_norm": 1.1027382612228394, "learning_rate": 9.882777664358069e-05, "loss": 3.1336, "step": 40050 }, { "epoch": 0.35412140800791253, "grad_norm": 5.140567779541016, "learning_rate": 9.882478338804843e-05, "loss": 3.3178, "step": 40100 }, { "epoch": 0.3545629558981967, "grad_norm": 2.794602155685425, "learning_rate": 9.882178636121568e-05, "loss": 3.8837, "step": 40150 }, { "epoch": 0.3550045037884809, "grad_norm": 2.1202375888824463, "learning_rate": 9.881878556331392e-05, "loss": 3.5493, "step": 40200 }, { "epoch": 0.3554460516787651, "grad_norm": 0.8689738512039185, "learning_rate": 9.881578099457496e-05, "loss": 3.29, "step": 40250 }, { "epoch": 0.35588759956904925, "grad_norm": 2.1677334308624268, "learning_rate": 9.881277265523085e-05, "loss": 3.2762, "step": 40300 }, { "epoch": 0.3563291474593334, "grad_norm": 1.659857988357544, "learning_rate": 9.880976054551399e-05, "loss": 3.8149, "step": 40350 }, { "epoch": 0.35677069534961764, "grad_norm": 2.109435558319092, "learning_rate": 9.880674466565701e-05, "loss": 3.3179, "step": 40400 }, { "epoch": 0.3572122432399018, "grad_norm": 2.959153175354004, "learning_rate": 9.880372501589289e-05, "loss": 3.5068, "step": 40450 }, { "epoch": 0.35765379113018597, "grad_norm": 0.8862214088439941, "learning_rate": 9.880070159645483e-05, "loss": 3.593, "step": 40500 }, { "epoch": 0.35809533902047014, "grad_norm": 4.1659464836120605, "learning_rate": 9.879767440757641e-05, "loss": 3.1608, "step": 40550 }, { "epoch": 0.35853688691075436, "grad_norm": 4.7671799659729, "learning_rate": 9.879464344949143e-05, "loss": 3.3211, "step": 40600 }, { "epoch": 0.3589784348010385, "grad_norm": 5.715523719787598, "learning_rate": 9.8791608722434e-05, "loss": 3.0981, "step": 40650 }, { "epoch": 0.3594199826913227, "grad_norm": 1.8158924579620361, "learning_rate": 9.878857022663855e-05, "loss": 2.8897, "step": 40700 }, { "epoch": 0.35986153058160686, "grad_norm": 2.091034412384033, "learning_rate": 9.878552796233976e-05, "loss": 3.6337, "step": 40750 }, { "epoch": 0.3603030784718911, "grad_norm": 3.8585503101348877, "learning_rate": 9.878248192977262e-05, "loss": 3.5439, "step": 40800 }, { "epoch": 0.36074462636217525, "grad_norm": 2.2025511264801025, "learning_rate": 9.877943212917242e-05, "loss": 3.3184, "step": 40850 }, { "epoch": 0.3611861742524594, "grad_norm": 3.991384744644165, "learning_rate": 9.877637856077473e-05, "loss": 3.4019, "step": 40900 }, { "epoch": 0.3616277221427436, "grad_norm": 4.2655534744262695, "learning_rate": 9.87733212248154e-05, "loss": 2.82, "step": 40950 }, { "epoch": 0.3620692700330278, "grad_norm": 1.683498740196228, "learning_rate": 9.877026012153059e-05, "loss": 3.3487, "step": 41000 }, { "epoch": 0.36251081792331197, "grad_norm": 2.264552593231201, "learning_rate": 9.876719525115675e-05, "loss": 2.975, "step": 41050 }, { "epoch": 0.36295236581359613, "grad_norm": 2.6153547763824463, "learning_rate": 9.876412661393061e-05, "loss": 3.3764, "step": 41100 }, { "epoch": 0.3633939137038803, "grad_norm": 1.591480016708374, "learning_rate": 9.876105421008919e-05, "loss": 2.965, "step": 41150 }, { "epoch": 0.3638354615941645, "grad_norm": 2.9769952297210693, "learning_rate": 9.875797803986983e-05, "loss": 3.2887, "step": 41200 }, { "epoch": 0.3642770094844487, "grad_norm": 5.074538230895996, "learning_rate": 9.875489810351011e-05, "loss": 3.5681, "step": 41250 }, { "epoch": 0.36471855737473285, "grad_norm": 2.3003244400024414, "learning_rate": 9.875181440124794e-05, "loss": 3.443, "step": 41300 }, { "epoch": 0.365160105265017, "grad_norm": 2.3685524463653564, "learning_rate": 9.874872693332152e-05, "loss": 3.1724, "step": 41350 }, { "epoch": 0.36560165315530124, "grad_norm": 0.6377138495445251, "learning_rate": 9.874563569996931e-05, "loss": 3.4932, "step": 41400 }, { "epoch": 0.3660432010455854, "grad_norm": 1.721374750137329, "learning_rate": 9.874254070143009e-05, "loss": 3.2422, "step": 41450 }, { "epoch": 0.3664847489358696, "grad_norm": 2.2348365783691406, "learning_rate": 9.873944193794294e-05, "loss": 3.0912, "step": 41500 }, { "epoch": 0.36692629682615374, "grad_norm": 1.8013442754745483, "learning_rate": 9.873633940974719e-05, "loss": 3.4616, "step": 41550 }, { "epoch": 0.36736784471643796, "grad_norm": 1.4249016046524048, "learning_rate": 9.873323311708249e-05, "loss": 3.2594, "step": 41600 }, { "epoch": 0.36780939260672213, "grad_norm": 1.336929440498352, "learning_rate": 9.873012306018879e-05, "loss": 3.4553, "step": 41650 }, { "epoch": 0.3682509404970063, "grad_norm": 1.3156746625900269, "learning_rate": 9.872700923930629e-05, "loss": 3.5638, "step": 41700 }, { "epoch": 0.36869248838729046, "grad_norm": 1.1224370002746582, "learning_rate": 9.872389165467553e-05, "loss": 3.131, "step": 41750 }, { "epoch": 0.3691340362775747, "grad_norm": 1.9698013067245483, "learning_rate": 9.87207703065373e-05, "loss": 3.2832, "step": 41800 }, { "epoch": 0.36957558416785885, "grad_norm": 1.916576623916626, "learning_rate": 9.87176451951327e-05, "loss": 3.8032, "step": 41850 }, { "epoch": 0.370017132058143, "grad_norm": 1.1512713432312012, "learning_rate": 9.871451632070312e-05, "loss": 3.4968, "step": 41900 }, { "epoch": 0.3704586799484272, "grad_norm": 1.165654182434082, "learning_rate": 9.871138368349025e-05, "loss": 3.7091, "step": 41950 }, { "epoch": 0.3709002278387114, "grad_norm": 0.8057221174240112, "learning_rate": 9.870824728373604e-05, "loss": 2.845, "step": 42000 }, { "epoch": 0.3709002278387114, "eval_asr_loss": 0.8905537465911811, "eval_loss": 3.063842535018921, "eval_runtime": 20.3425, "eval_samples_per_second": 37.754, "eval_steps_per_second": 9.438, "eval_tts_loss": 5.883050221472156, "step": 42000 }, { "epoch": 0.37134177572899557, "grad_norm": 9.080826759338379, "learning_rate": 9.870510712168278e-05, "loss": 3.178, "step": 42050 }, { "epoch": 0.37178332361927974, "grad_norm": 1.3995147943496704, "learning_rate": 9.870196319757299e-05, "loss": 3.316, "step": 42100 }, { "epoch": 0.3722248715095639, "grad_norm": 2.64806866645813, "learning_rate": 9.869887850223219e-05, "loss": 3.0827, "step": 42150 }, { "epoch": 0.3726664193998481, "grad_norm": 0.9530137181282043, "learning_rate": 9.869572712996722e-05, "loss": 3.2979, "step": 42200 }, { "epoch": 0.3731079672901323, "grad_norm": 2.155907392501831, "learning_rate": 9.869257199637023e-05, "loss": 3.2503, "step": 42250 }, { "epoch": 0.37354951518041646, "grad_norm": 1.780704379081726, "learning_rate": 9.868941310168498e-05, "loss": 3.4232, "step": 42300 }, { "epoch": 0.3739910630707006, "grad_norm": 0.6130134463310242, "learning_rate": 9.868625044615543e-05, "loss": 3.2393, "step": 42350 }, { "epoch": 0.37443261096098485, "grad_norm": 1.8535226583480835, "learning_rate": 9.868308403002589e-05, "loss": 3.2783, "step": 42400 }, { "epoch": 0.374874158851269, "grad_norm": 3.596348285675049, "learning_rate": 9.867991385354092e-05, "loss": 3.157, "step": 42450 }, { "epoch": 0.3753157067415532, "grad_norm": 8.728959083557129, "learning_rate": 9.86767399169454e-05, "loss": 3.542, "step": 42500 }, { "epoch": 0.37575725463183735, "grad_norm": 1.9934968948364258, "learning_rate": 9.867356222048448e-05, "loss": 3.4873, "step": 42550 }, { "epoch": 0.37619880252212157, "grad_norm": 2.618511915206909, "learning_rate": 9.867038076440362e-05, "loss": 3.1409, "step": 42600 }, { "epoch": 0.37664035041240573, "grad_norm": 1.344989538192749, "learning_rate": 9.866719554894857e-05, "loss": 3.4496, "step": 42650 }, { "epoch": 0.3770818983026899, "grad_norm": 1.429974913597107, "learning_rate": 9.866400657436534e-05, "loss": 3.1867, "step": 42700 }, { "epoch": 0.37752344619297407, "grad_norm": 1.021186351776123, "learning_rate": 9.866081384090026e-05, "loss": 2.994, "step": 42750 }, { "epoch": 0.3779649940832583, "grad_norm": 2.500567674636841, "learning_rate": 9.865761734879994e-05, "loss": 2.9304, "step": 42800 }, { "epoch": 0.37840654197354245, "grad_norm": 1.401653528213501, "learning_rate": 9.865441709831129e-05, "loss": 3.5495, "step": 42850 }, { "epoch": 0.3788480898638266, "grad_norm": 2.189068555831909, "learning_rate": 9.865121308968151e-05, "loss": 3.5045, "step": 42900 }, { "epoch": 0.3792896377541108, "grad_norm": 1.583028793334961, "learning_rate": 9.864800532315806e-05, "loss": 3.4075, "step": 42950 }, { "epoch": 0.379731185644395, "grad_norm": 1.778917908668518, "learning_rate": 9.864479379898871e-05, "loss": 3.3313, "step": 43000 }, { "epoch": 0.3801727335346792, "grad_norm": 2.908151626586914, "learning_rate": 9.864157851742154e-05, "loss": 3.1747, "step": 43050 }, { "epoch": 0.38061428142496334, "grad_norm": 3.3366377353668213, "learning_rate": 9.86383594787049e-05, "loss": 3.5529, "step": 43100 }, { "epoch": 0.3810558293152475, "grad_norm": 1.980222225189209, "learning_rate": 9.863513668308744e-05, "loss": 3.559, "step": 43150 }, { "epoch": 0.38149737720553173, "grad_norm": 2.1785378456115723, "learning_rate": 9.863191013081807e-05, "loss": 3.3076, "step": 43200 }, { "epoch": 0.3819389250958159, "grad_norm": 3.69520902633667, "learning_rate": 9.862867982214604e-05, "loss": 3.4053, "step": 43250 }, { "epoch": 0.38238047298610006, "grad_norm": 5.211033821105957, "learning_rate": 9.862544575732085e-05, "loss": 3.2504, "step": 43300 }, { "epoch": 0.38282202087638423, "grad_norm": 2.920567750930786, "learning_rate": 9.862220793659232e-05, "loss": 3.1881, "step": 43350 }, { "epoch": 0.38326356876666845, "grad_norm": 2.8354828357696533, "learning_rate": 9.86189663602105e-05, "loss": 3.6872, "step": 43400 }, { "epoch": 0.3837051166569526, "grad_norm": 3.6911203861236572, "learning_rate": 9.861572102842583e-05, "loss": 3.3914, "step": 43450 }, { "epoch": 0.3841466645472368, "grad_norm": 3.9869236946105957, "learning_rate": 9.861247194148895e-05, "loss": 3.3991, "step": 43500 }, { "epoch": 0.38458821243752095, "grad_norm": 2.9851901531219482, "learning_rate": 9.860928419328401e-05, "loss": 3.4688, "step": 43550 }, { "epoch": 0.38502976032780517, "grad_norm": 5.538595199584961, "learning_rate": 9.860602767188645e-05, "loss": 3.3237, "step": 43600 }, { "epoch": 0.38547130821808934, "grad_norm": 0.9150497317314148, "learning_rate": 9.860276739608543e-05, "loss": 3.2588, "step": 43650 }, { "epoch": 0.3859128561083735, "grad_norm": 3.647045612335205, "learning_rate": 9.859950336613275e-05, "loss": 3.1377, "step": 43700 }, { "epoch": 0.38635440399865767, "grad_norm": 0.6097536087036133, "learning_rate": 9.859623558228055e-05, "loss": 3.1102, "step": 43750 }, { "epoch": 0.3867959518889419, "grad_norm": 1.7617607116699219, "learning_rate": 9.859296404478124e-05, "loss": 2.9313, "step": 43800 }, { "epoch": 0.38723749977922606, "grad_norm": 2.4368743896484375, "learning_rate": 9.858968875388752e-05, "loss": 3.2671, "step": 43850 }, { "epoch": 0.3876790476695102, "grad_norm": 1.931584358215332, "learning_rate": 9.858640970985237e-05, "loss": 3.4481, "step": 43900 }, { "epoch": 0.3881205955597944, "grad_norm": 1.9381108283996582, "learning_rate": 9.858312691292906e-05, "loss": 3.2431, "step": 43950 }, { "epoch": 0.3885621434500786, "grad_norm": 2.4838809967041016, "learning_rate": 9.857984036337118e-05, "loss": 3.4532, "step": 44000 }, { "epoch": 0.3890036913403628, "grad_norm": 2.848560094833374, "learning_rate": 9.857655006143259e-05, "loss": 3.3183, "step": 44050 }, { "epoch": 0.38944523923064694, "grad_norm": 2.6281092166900635, "learning_rate": 9.85732560073674e-05, "loss": 3.348, "step": 44100 }, { "epoch": 0.3898867871209311, "grad_norm": 1.351362943649292, "learning_rate": 9.856995820143006e-05, "loss": 3.2875, "step": 44150 }, { "epoch": 0.39032833501121533, "grad_norm": 11.539451599121094, "learning_rate": 9.856665664387535e-05, "loss": 3.8226, "step": 44200 }, { "epoch": 0.3907698829014995, "grad_norm": 2.561694383621216, "learning_rate": 9.856335133495822e-05, "loss": 3.7136, "step": 44250 }, { "epoch": 0.39121143079178367, "grad_norm": 1.2988532781600952, "learning_rate": 9.856004227493402e-05, "loss": 3.5633, "step": 44300 }, { "epoch": 0.39165297868206783, "grad_norm": 2.9327540397644043, "learning_rate": 9.855672946405831e-05, "loss": 3.2624, "step": 44350 }, { "epoch": 0.39209452657235205, "grad_norm": 2.4923012256622314, "learning_rate": 9.855341290258703e-05, "loss": 3.0418, "step": 44400 }, { "epoch": 0.3925360744626362, "grad_norm": 0.7568151354789734, "learning_rate": 9.85500925907763e-05, "loss": 3.5547, "step": 44450 }, { "epoch": 0.3929776223529204, "grad_norm": 1.1965677738189697, "learning_rate": 9.854676852888262e-05, "loss": 3.4921, "step": 44500 }, { "epoch": 0.39341917024320455, "grad_norm": 0.5616858601570129, "learning_rate": 9.854344071716272e-05, "loss": 3.2839, "step": 44550 }, { "epoch": 0.3938607181334888, "grad_norm": 3.225074052810669, "learning_rate": 9.854010915587369e-05, "loss": 3.4744, "step": 44600 }, { "epoch": 0.39430226602377294, "grad_norm": 3.9845869541168213, "learning_rate": 9.853677384527282e-05, "loss": 3.4246, "step": 44650 }, { "epoch": 0.3947438139140571, "grad_norm": 3.029383897781372, "learning_rate": 9.853343478561773e-05, "loss": 3.408, "step": 44700 }, { "epoch": 0.3951853618043413, "grad_norm": 1.8023637533187866, "learning_rate": 9.853009197716638e-05, "loss": 3.1036, "step": 44750 }, { "epoch": 0.3956269096946255, "grad_norm": 0.8640767931938171, "learning_rate": 9.852674542017694e-05, "loss": 3.3368, "step": 44800 }, { "epoch": 0.39606845758490966, "grad_norm": 0.6954994797706604, "learning_rate": 9.852339511490792e-05, "loss": 2.9964, "step": 44850 }, { "epoch": 0.3965100054751938, "grad_norm": 8.39139461517334, "learning_rate": 9.85200410616181e-05, "loss": 2.9249, "step": 44900 }, { "epoch": 0.396951553365478, "grad_norm": 4.7654900550842285, "learning_rate": 9.851668326056652e-05, "loss": 3.0992, "step": 44950 }, { "epoch": 0.3973931012557622, "grad_norm": 1.2103265523910522, "learning_rate": 9.851332171201258e-05, "loss": 3.3672, "step": 45000 }, { "epoch": 0.3973931012557622, "eval_asr_loss": 0.8913320451830068, "eval_loss": 3.0676581859588623, "eval_runtime": 20.2524, "eval_samples_per_second": 37.921, "eval_steps_per_second": 9.48, "eval_tts_loss": 5.97925205245277, "step": 45000 }, { "epoch": 0.3978346491460464, "grad_norm": 2.4184420108795166, "learning_rate": 9.850995641621592e-05, "loss": 3.6283, "step": 45050 }, { "epoch": 0.39827619703633055, "grad_norm": 1.4427285194396973, "learning_rate": 9.850658737343648e-05, "loss": 3.5224, "step": 45100 }, { "epoch": 0.3987177449266147, "grad_norm": 6.392005920410156, "learning_rate": 9.850321458393447e-05, "loss": 3.0993, "step": 45150 }, { "epoch": 0.39915929281689894, "grad_norm": 1.4388530254364014, "learning_rate": 9.849983804797046e-05, "loss": 3.0099, "step": 45200 }, { "epoch": 0.3996008407071831, "grad_norm": 2.679241180419922, "learning_rate": 9.849645776580522e-05, "loss": 3.3363, "step": 45250 }, { "epoch": 0.40004238859746727, "grad_norm": 1.61045503616333, "learning_rate": 9.849307373769985e-05, "loss": 3.1612, "step": 45300 }, { "epoch": 0.40048393648775144, "grad_norm": 1.9430288076400757, "learning_rate": 9.848968596391574e-05, "loss": 3.4328, "step": 45350 }, { "epoch": 0.40092548437803566, "grad_norm": 1.6506637334823608, "learning_rate": 9.848629444471456e-05, "loss": 3.6278, "step": 45400 }, { "epoch": 0.4013670322683198, "grad_norm": 2.5751986503601074, "learning_rate": 9.848289918035831e-05, "loss": 3.3468, "step": 45450 }, { "epoch": 0.401808580158604, "grad_norm": 2.1925528049468994, "learning_rate": 9.84795001711092e-05, "loss": 3.3276, "step": 45500 }, { "epoch": 0.40225012804888816, "grad_norm": 3.269740581512451, "learning_rate": 9.847609741722981e-05, "loss": 2.7999, "step": 45550 }, { "epoch": 0.4026916759391724, "grad_norm": 4.853845596313477, "learning_rate": 9.847269091898294e-05, "loss": 3.1533, "step": 45600 }, { "epoch": 0.40313322382945654, "grad_norm": 2.3576786518096924, "learning_rate": 9.846928067663175e-05, "loss": 3.4268, "step": 45650 }, { "epoch": 0.4035747717197407, "grad_norm": 1.3526619672775269, "learning_rate": 9.846586669043964e-05, "loss": 3.156, "step": 45700 }, { "epoch": 0.4040163196100249, "grad_norm": 1.5202453136444092, "learning_rate": 9.846244896067031e-05, "loss": 3.2914, "step": 45750 }, { "epoch": 0.4044578675003091, "grad_norm": 1.80320143699646, "learning_rate": 9.845902748758773e-05, "loss": 3.2984, "step": 45800 }, { "epoch": 0.40489941539059326, "grad_norm": 1.7634551525115967, "learning_rate": 9.845560227145621e-05, "loss": 3.1742, "step": 45850 }, { "epoch": 0.40534096328087743, "grad_norm": 2.447131633758545, "learning_rate": 9.84521733125403e-05, "loss": 3.0532, "step": 45900 }, { "epoch": 0.4057825111711616, "grad_norm": 2.7054214477539062, "learning_rate": 9.844874061110485e-05, "loss": 3.5135, "step": 45950 }, { "epoch": 0.4062240590614458, "grad_norm": 0.9877326488494873, "learning_rate": 9.844530416741505e-05, "loss": 3.4712, "step": 46000 }, { "epoch": 0.40666560695173, "grad_norm": 1.2479236125946045, "learning_rate": 9.84418639817363e-05, "loss": 3.5099, "step": 46050 }, { "epoch": 0.40710715484201415, "grad_norm": 3.176802635192871, "learning_rate": 9.843842005433431e-05, "loss": 3.5577, "step": 46100 }, { "epoch": 0.4075487027322983, "grad_norm": 1.3120348453521729, "learning_rate": 9.843497238547513e-05, "loss": 2.9955, "step": 46150 }, { "epoch": 0.40799025062258254, "grad_norm": 1.1844558715820312, "learning_rate": 9.843152097542505e-05, "loss": 3.6275, "step": 46200 }, { "epoch": 0.4084317985128667, "grad_norm": 2.701456308364868, "learning_rate": 9.842806582445067e-05, "loss": 3.8021, "step": 46250 }, { "epoch": 0.4088733464031509, "grad_norm": 2.6151137351989746, "learning_rate": 9.842460693281885e-05, "loss": 3.6685, "step": 46300 }, { "epoch": 0.40931489429343504, "grad_norm": 2.728315591812134, "learning_rate": 9.842114430079679e-05, "loss": 3.4043, "step": 46350 }, { "epoch": 0.40975644218371926, "grad_norm": 2.2898852825164795, "learning_rate": 9.841767792865191e-05, "loss": 3.6381, "step": 46400 }, { "epoch": 0.4101979900740034, "grad_norm": 1.9836546182632446, "learning_rate": 9.8414207816652e-05, "loss": 3.0714, "step": 46450 }, { "epoch": 0.4106395379642876, "grad_norm": 2.8951032161712646, "learning_rate": 9.841073396506506e-05, "loss": 3.4733, "step": 46500 }, { "epoch": 0.41108108585457176, "grad_norm": 1.4405927658081055, "learning_rate": 9.840725637415945e-05, "loss": 3.2872, "step": 46550 }, { "epoch": 0.411522633744856, "grad_norm": 3.1269359588623047, "learning_rate": 9.840377504420376e-05, "loss": 3.4911, "step": 46600 }, { "epoch": 0.41196418163514015, "grad_norm": 3.698716163635254, "learning_rate": 9.840028997546689e-05, "loss": 3.1268, "step": 46650 }, { "epoch": 0.4124057295254243, "grad_norm": 2.955763578414917, "learning_rate": 9.839680116821804e-05, "loss": 3.5459, "step": 46700 }, { "epoch": 0.4128472774157085, "grad_norm": 0.848429799079895, "learning_rate": 9.83933086227267e-05, "loss": 3.2194, "step": 46750 }, { "epoch": 0.4132888253059927, "grad_norm": 8.197107315063477, "learning_rate": 9.838981233926264e-05, "loss": 3.2381, "step": 46800 }, { "epoch": 0.41373037319627687, "grad_norm": 1.4284354448318481, "learning_rate": 9.83863123180959e-05, "loss": 3.4783, "step": 46850 }, { "epoch": 0.41417192108656103, "grad_norm": 1.3862097263336182, "learning_rate": 9.838280855949684e-05, "loss": 3.3749, "step": 46900 }, { "epoch": 0.41461346897684526, "grad_norm": 1.5118871927261353, "learning_rate": 9.837930106373609e-05, "loss": 3.2749, "step": 46950 }, { "epoch": 0.4150550168671294, "grad_norm": 4.723493576049805, "learning_rate": 9.83757898310846e-05, "loss": 2.9933, "step": 47000 }, { "epoch": 0.4154965647574136, "grad_norm": 6.659562110900879, "learning_rate": 9.837227486181353e-05, "loss": 3.4323, "step": 47050 }, { "epoch": 0.41593811264769776, "grad_norm": 2.5123307704925537, "learning_rate": 9.836875615619443e-05, "loss": 2.9715, "step": 47100 }, { "epoch": 0.416379660537982, "grad_norm": 2.6539173126220703, "learning_rate": 9.836523371449906e-05, "loss": 3.1207, "step": 47150 }, { "epoch": 0.41682120842826614, "grad_norm": 3.233375072479248, "learning_rate": 9.836170753699952e-05, "loss": 3.1869, "step": 47200 }, { "epoch": 0.4172627563185503, "grad_norm": 1.2898887395858765, "learning_rate": 9.835817762396816e-05, "loss": 3.2155, "step": 47250 }, { "epoch": 0.4177043042088345, "grad_norm": 0.9073876738548279, "learning_rate": 9.835464397567765e-05, "loss": 3.363, "step": 47300 }, { "epoch": 0.4181458520991187, "grad_norm": 2.146453380584717, "learning_rate": 9.835110659240092e-05, "loss": 3.6663, "step": 47350 }, { "epoch": 0.41858739998940286, "grad_norm": 4.763782501220703, "learning_rate": 9.834756547441122e-05, "loss": 3.5419, "step": 47400 }, { "epoch": 0.41902894787968703, "grad_norm": 2.6837880611419678, "learning_rate": 9.834402062198207e-05, "loss": 3.134, "step": 47450 }, { "epoch": 0.4194704957699712, "grad_norm": 2.5617268085479736, "learning_rate": 9.834047203538726e-05, "loss": 3.4525, "step": 47500 }, { "epoch": 0.4199120436602554, "grad_norm": 2.1954944133758545, "learning_rate": 9.83369197149009e-05, "loss": 3.7715, "step": 47550 }, { "epoch": 0.4203535915505396, "grad_norm": 1.5438072681427002, "learning_rate": 9.833336366079737e-05, "loss": 3.6781, "step": 47600 }, { "epoch": 0.42079513944082375, "grad_norm": 1.757779598236084, "learning_rate": 9.832980387335138e-05, "loss": 3.4219, "step": 47650 }, { "epoch": 0.4212366873311079, "grad_norm": 2.8832337856292725, "learning_rate": 9.832624035283785e-05, "loss": 2.9347, "step": 47700 }, { "epoch": 0.42167823522139214, "grad_norm": 2.0708539485931396, "learning_rate": 9.832274448117773e-05, "loss": 3.1908, "step": 47750 }, { "epoch": 0.4221197831116763, "grad_norm": 1.883726716041565, "learning_rate": 9.831917357000283e-05, "loss": 3.5663, "step": 47800 }, { "epoch": 0.42256133100196047, "grad_norm": 1.2137542963027954, "learning_rate": 9.83155989265815e-05, "loss": 3.3599, "step": 47850 }, { "epoch": 0.42300287889224464, "grad_norm": 1.1021403074264526, "learning_rate": 9.831202055118986e-05, "loss": 3.3176, "step": 47900 }, { "epoch": 0.42344442678252886, "grad_norm": 3.1852636337280273, "learning_rate": 9.830843844410431e-05, "loss": 3.2934, "step": 47950 }, { "epoch": 0.423885974672813, "grad_norm": 2.2362117767333984, "learning_rate": 9.830492435893767e-05, "loss": 3.5098, "step": 48000 }, { "epoch": 0.423885974672813, "eval_asr_loss": 0.8946412356905893, "eval_loss": 3.0450656414031982, "eval_runtime": 20.5825, "eval_samples_per_second": 37.313, "eval_steps_per_second": 9.328, "eval_tts_loss": 5.9677758015827616, "step": 48000 }, { "epoch": 0.4243275225630972, "grad_norm": 0.6300576329231262, "learning_rate": 9.830133486391473e-05, "loss": 3.5818, "step": 48050 }, { "epoch": 0.42476907045338136, "grad_norm": 1.3167023658752441, "learning_rate": 9.829774163802325e-05, "loss": 3.4292, "step": 48100 }, { "epoch": 0.4252106183436656, "grad_norm": 2.811063051223755, "learning_rate": 9.829414468154079e-05, "loss": 3.7977, "step": 48150 }, { "epoch": 0.42565216623394975, "grad_norm": 2.503530502319336, "learning_rate": 9.829054399474516e-05, "loss": 3.6608, "step": 48200 }, { "epoch": 0.4260937141242339, "grad_norm": 2.051328420639038, "learning_rate": 9.828693957791452e-05, "loss": 3.3447, "step": 48250 }, { "epoch": 0.4265352620145181, "grad_norm": 0.9550831913948059, "learning_rate": 9.828333143132724e-05, "loss": 3.2968, "step": 48300 }, { "epoch": 0.4269768099048023, "grad_norm": 3.1451058387756348, "learning_rate": 9.827971955526205e-05, "loss": 3.3196, "step": 48350 }, { "epoch": 0.42741835779508647, "grad_norm": 1.7226669788360596, "learning_rate": 9.827610394999791e-05, "loss": 3.1922, "step": 48400 }, { "epoch": 0.42785990568537063, "grad_norm": 2.289461851119995, "learning_rate": 9.827248461581412e-05, "loss": 3.327, "step": 48450 }, { "epoch": 0.4283014535756548, "grad_norm": 3.5129716396331787, "learning_rate": 9.826886155299023e-05, "loss": 3.4937, "step": 48500 }, { "epoch": 0.428743001465939, "grad_norm": 2.9753642082214355, "learning_rate": 9.826523476180609e-05, "loss": 3.3748, "step": 48550 }, { "epoch": 0.4291845493562232, "grad_norm": 1.0942645072937012, "learning_rate": 9.826160424254185e-05, "loss": 3.391, "step": 48600 }, { "epoch": 0.42962609724650735, "grad_norm": 2.1059162616729736, "learning_rate": 9.825796999547792e-05, "loss": 3.7639, "step": 48650 }, { "epoch": 0.4300676451367915, "grad_norm": 1.9335929155349731, "learning_rate": 9.825433202089503e-05, "loss": 2.9592, "step": 48700 }, { "epoch": 0.43050919302707574, "grad_norm": 0.7627319693565369, "learning_rate": 9.825069031907416e-05, "loss": 3.3292, "step": 48750 }, { "epoch": 0.4309507409173599, "grad_norm": 3.433350086212158, "learning_rate": 9.824704489029664e-05, "loss": 3.3546, "step": 48800 }, { "epoch": 0.4313922888076441, "grad_norm": 2.693155527114868, "learning_rate": 9.824339573484402e-05, "loss": 3.4344, "step": 48850 }, { "epoch": 0.43183383669792824, "grad_norm": 1.1923327445983887, "learning_rate": 9.823974285299817e-05, "loss": 3.3486, "step": 48900 }, { "epoch": 0.43227538458821246, "grad_norm": 1.9175432920455933, "learning_rate": 9.823608624504123e-05, "loss": 3.4314, "step": 48950 }, { "epoch": 0.43271693247849663, "grad_norm": 0.7622149586677551, "learning_rate": 9.823242591125568e-05, "loss": 3.1681, "step": 49000 }, { "epoch": 0.4331584803687808, "grad_norm": 2.3033721446990967, "learning_rate": 9.822876185192421e-05, "loss": 3.4592, "step": 49050 }, { "epoch": 0.43360002825906496, "grad_norm": 0.886083722114563, "learning_rate": 9.822509406732987e-05, "loss": 3.6985, "step": 49100 }, { "epoch": 0.4340415761493492, "grad_norm": 6.617265701293945, "learning_rate": 9.822142255775594e-05, "loss": 3.1759, "step": 49150 }, { "epoch": 0.43448312403963335, "grad_norm": 4.3931121826171875, "learning_rate": 9.821774732348602e-05, "loss": 3.4327, "step": 49200 }, { "epoch": 0.4349246719299175, "grad_norm": 1.2443506717681885, "learning_rate": 9.8214068364804e-05, "loss": 3.5825, "step": 49250 }, { "epoch": 0.4353662198202017, "grad_norm": 2.791949510574341, "learning_rate": 9.821038568199403e-05, "loss": 3.1509, "step": 49300 }, { "epoch": 0.4358077677104859, "grad_norm": 1.996881127357483, "learning_rate": 9.820669927534061e-05, "loss": 2.9644, "step": 49350 }, { "epoch": 0.43624931560077007, "grad_norm": 1.907235026359558, "learning_rate": 9.820300914512842e-05, "loss": 3.3867, "step": 49400 }, { "epoch": 0.43669086349105424, "grad_norm": 3.5670182704925537, "learning_rate": 9.819931529164254e-05, "loss": 3.4446, "step": 49450 }, { "epoch": 0.4371324113813384, "grad_norm": 1.6872813701629639, "learning_rate": 9.819561771516826e-05, "loss": 2.9724, "step": 49500 }, { "epoch": 0.4375739592716226, "grad_norm": 3.4050118923187256, "learning_rate": 9.819191641599121e-05, "loss": 3.2879, "step": 49550 }, { "epoch": 0.4380155071619068, "grad_norm": 2.1053950786590576, "learning_rate": 9.818821139439727e-05, "loss": 3.4533, "step": 49600 }, { "epoch": 0.43845705505219096, "grad_norm": 1.7744102478027344, "learning_rate": 9.818450265067261e-05, "loss": 3.3481, "step": 49650 }, { "epoch": 0.4388986029424751, "grad_norm": 2.2178120613098145, "learning_rate": 9.818079018510374e-05, "loss": 3.3662, "step": 49700 }, { "epoch": 0.43934015083275935, "grad_norm": 2.2515435218811035, "learning_rate": 9.817707399797736e-05, "loss": 3.336, "step": 49750 }, { "epoch": 0.4397816987230435, "grad_norm": 3.4062695503234863, "learning_rate": 9.817335408958056e-05, "loss": 3.2625, "step": 49800 }, { "epoch": 0.4402232466133277, "grad_norm": 1.6740872859954834, "learning_rate": 9.816963046020065e-05, "loss": 3.7033, "step": 49850 }, { "epoch": 0.44066479450361185, "grad_norm": 0.949266254901886, "learning_rate": 9.816590311012525e-05, "loss": 3.2875, "step": 49900 }, { "epoch": 0.44110634239389607, "grad_norm": 2.8111281394958496, "learning_rate": 9.816217203964228e-05, "loss": 3.2595, "step": 49950 }, { "epoch": 0.44154789028418023, "grad_norm": 1.117742657661438, "learning_rate": 9.815843724903993e-05, "loss": 3.2435, "step": 50000 }, { "epoch": 0.4419894381744644, "grad_norm": 2.1841189861297607, "learning_rate": 9.815469873860666e-05, "loss": 3.5207, "step": 50050 }, { "epoch": 0.44243098606474857, "grad_norm": 4.919894695281982, "learning_rate": 9.815095650863124e-05, "loss": 3.3703, "step": 50100 }, { "epoch": 0.4428725339550328, "grad_norm": 3.6214163303375244, "learning_rate": 9.814721055940275e-05, "loss": 3.5246, "step": 50150 }, { "epoch": 0.44331408184531695, "grad_norm": 3.0915398597717285, "learning_rate": 9.814346089121053e-05, "loss": 3.7959, "step": 50200 }, { "epoch": 0.4437556297356011, "grad_norm": 1.7853915691375732, "learning_rate": 9.813970750434419e-05, "loss": 3.4, "step": 50250 }, { "epoch": 0.4441971776258853, "grad_norm": 0.6693574786186218, "learning_rate": 9.813595039909367e-05, "loss": 3.2674, "step": 50300 }, { "epoch": 0.4446387255161695, "grad_norm": 7.433414936065674, "learning_rate": 9.813218957574914e-05, "loss": 3.1686, "step": 50350 }, { "epoch": 0.4450802734064537, "grad_norm": 2.307650566101074, "learning_rate": 9.812842503460114e-05, "loss": 3.2572, "step": 50400 }, { "epoch": 0.44552182129673784, "grad_norm": 4.549615383148193, "learning_rate": 9.812465677594041e-05, "loss": 3.5429, "step": 50450 }, { "epoch": 0.445963369187022, "grad_norm": 1.0888493061065674, "learning_rate": 9.812088480005804e-05, "loss": 3.3837, "step": 50500 }, { "epoch": 0.44640491707730623, "grad_norm": 2.1504311561584473, "learning_rate": 9.811718465752564e-05, "loss": 3.7276, "step": 50550 }, { "epoch": 0.4468464649675904, "grad_norm": 3.0695831775665283, "learning_rate": 9.811340532240422e-05, "loss": 3.4433, "step": 50600 }, { "epoch": 0.44728801285787456, "grad_norm": 1.7197152376174927, "learning_rate": 9.810962227093022e-05, "loss": 3.5762, "step": 50650 }, { "epoch": 0.44772956074815873, "grad_norm": 1.3767517805099487, "learning_rate": 9.810583550339588e-05, "loss": 3.0935, "step": 50700 }, { "epoch": 0.44817110863844295, "grad_norm": 1.5721153020858765, "learning_rate": 9.810204502009364e-05, "loss": 3.4474, "step": 50750 }, { "epoch": 0.4486126565287271, "grad_norm": 4.190256118774414, "learning_rate": 9.809825082131633e-05, "loss": 3.1628, "step": 50800 }, { "epoch": 0.4490542044190113, "grad_norm": 2.406953811645508, "learning_rate": 9.809445290735702e-05, "loss": 3.2441, "step": 50850 }, { "epoch": 0.44949575230929545, "grad_norm": 4.159549236297607, "learning_rate": 9.809065127850903e-05, "loss": 3.0559, "step": 50900 }, { "epoch": 0.44993730019957967, "grad_norm": 1.8754595518112183, "learning_rate": 9.808684593506605e-05, "loss": 3.7703, "step": 50950 }, { "epoch": 0.45037884808986384, "grad_norm": 1.29695725440979, "learning_rate": 9.808303687732196e-05, "loss": 3.1498, "step": 51000 }, { "epoch": 0.45037884808986384, "eval_asr_loss": 0.9004947789321158, "eval_loss": 3.0421018600463867, "eval_runtime": 20.6473, "eval_samples_per_second": 37.196, "eval_steps_per_second": 9.299, "eval_tts_loss": 5.865643483398774, "step": 51000 }, { "epoch": 0.450820395980148, "grad_norm": 2.4158124923706055, "learning_rate": 9.807922410557102e-05, "loss": 3.3356, "step": 51050 }, { "epoch": 0.45126194387043217, "grad_norm": 0.9025440812110901, "learning_rate": 9.807540762010772e-05, "loss": 3.336, "step": 51100 }, { "epoch": 0.4517034917607164, "grad_norm": 5.0518364906311035, "learning_rate": 9.807158742122684e-05, "loss": 3.2288, "step": 51150 }, { "epoch": 0.45214503965100056, "grad_norm": 1.4660353660583496, "learning_rate": 9.806776350922346e-05, "loss": 3.0611, "step": 51200 }, { "epoch": 0.4525865875412847, "grad_norm": 1.7514241933822632, "learning_rate": 9.806393588439297e-05, "loss": 3.304, "step": 51250 }, { "epoch": 0.4530281354315689, "grad_norm": 1.4853622913360596, "learning_rate": 9.806010454703099e-05, "loss": 3.436, "step": 51300 }, { "epoch": 0.4534696833218531, "grad_norm": 2.174837112426758, "learning_rate": 9.805626949743347e-05, "loss": 3.6087, "step": 51350 }, { "epoch": 0.4539112312121373, "grad_norm": 1.5965477228164673, "learning_rate": 9.805243073589665e-05, "loss": 3.4667, "step": 51400 }, { "epoch": 0.45435277910242144, "grad_norm": 2.395080804824829, "learning_rate": 9.804858826271703e-05, "loss": 3.5994, "step": 51450 }, { "epoch": 0.4547943269927056, "grad_norm": 4.734639644622803, "learning_rate": 9.804474207819139e-05, "loss": 3.2731, "step": 51500 }, { "epoch": 0.45523587488298983, "grad_norm": 1.91262948513031, "learning_rate": 9.804089218261684e-05, "loss": 3.4359, "step": 51550 }, { "epoch": 0.455677422773274, "grad_norm": 4.3235321044921875, "learning_rate": 9.803703857629075e-05, "loss": 3.6903, "step": 51600 }, { "epoch": 0.45611897066355817, "grad_norm": 2.399240016937256, "learning_rate": 9.803318125951075e-05, "loss": 3.1958, "step": 51650 }, { "epoch": 0.45656051855384233, "grad_norm": 3.3152384757995605, "learning_rate": 9.802932023257483e-05, "loss": 3.4387, "step": 51700 }, { "epoch": 0.45700206644412655, "grad_norm": 3.167579412460327, "learning_rate": 9.802545549578119e-05, "loss": 3.6247, "step": 51750 }, { "epoch": 0.4574436143344107, "grad_norm": 1.6194385290145874, "learning_rate": 9.802158704942837e-05, "loss": 3.3378, "step": 51800 }, { "epoch": 0.4578851622246949, "grad_norm": 1.213178277015686, "learning_rate": 9.801771489381515e-05, "loss": 3.0608, "step": 51850 }, { "epoch": 0.45832671011497905, "grad_norm": 4.134045600891113, "learning_rate": 9.8013916582878e-05, "loss": 3.1588, "step": 51900 }, { "epoch": 0.4587682580052633, "grad_norm": 3.060990571975708, "learning_rate": 9.801003708381187e-05, "loss": 3.3848, "step": 51950 }, { "epoch": 0.45920980589554744, "grad_norm": 2.216519594192505, "learning_rate": 9.800615387637748e-05, "loss": 3.2701, "step": 52000 }, { "epoch": 0.4596513537858316, "grad_norm": 2.748791456222534, "learning_rate": 9.80022669608748e-05, "loss": 3.6429, "step": 52050 }, { "epoch": 0.4600929016761158, "grad_norm": 2.1832919120788574, "learning_rate": 9.799837633760403e-05, "loss": 3.4686, "step": 52100 }, { "epoch": 0.4605344495664, "grad_norm": 10.31329345703125, "learning_rate": 9.799448200686569e-05, "loss": 3.4104, "step": 52150 }, { "epoch": 0.46097599745668416, "grad_norm": 3.921966791152954, "learning_rate": 9.799058396896062e-05, "loss": 3.1034, "step": 52200 }, { "epoch": 0.4614175453469683, "grad_norm": 1.399453043937683, "learning_rate": 9.798668222418988e-05, "loss": 2.6928, "step": 52250 }, { "epoch": 0.4618590932372525, "grad_norm": 2.0908732414245605, "learning_rate": 9.798277677285483e-05, "loss": 3.2176, "step": 52300 }, { "epoch": 0.4623006411275367, "grad_norm": 1.981628656387329, "learning_rate": 9.797886761525719e-05, "loss": 3.1046, "step": 52350 }, { "epoch": 0.4627421890178209, "grad_norm": 2.1923952102661133, "learning_rate": 9.797495475169886e-05, "loss": 2.9833, "step": 52400 }, { "epoch": 0.46318373690810505, "grad_norm": 4.177052021026611, "learning_rate": 9.79710381824821e-05, "loss": 2.9391, "step": 52450 }, { "epoch": 0.4636252847983892, "grad_norm": 1.3210045099258423, "learning_rate": 9.796711790790941e-05, "loss": 3.5919, "step": 52500 }, { "epoch": 0.46406683268867344, "grad_norm": 1.5212137699127197, "learning_rate": 9.796319392828361e-05, "loss": 3.4421, "step": 52550 }, { "epoch": 0.4645083805789576, "grad_norm": 9.704986572265625, "learning_rate": 9.795926624390781e-05, "loss": 3.2297, "step": 52600 }, { "epoch": 0.46494992846924177, "grad_norm": 0.9471487402915955, "learning_rate": 9.795533485508537e-05, "loss": 3.6265, "step": 52650 }, { "epoch": 0.46539147635952594, "grad_norm": 2.34822154045105, "learning_rate": 9.795139976211996e-05, "loss": 3.2869, "step": 52700 }, { "epoch": 0.46583302424981016, "grad_norm": 1.4003323316574097, "learning_rate": 9.794746096531554e-05, "loss": 3.215, "step": 52750 }, { "epoch": 0.4662745721400943, "grad_norm": 0.5006351470947266, "learning_rate": 9.794351846497634e-05, "loss": 3.3245, "step": 52800 }, { "epoch": 0.4667161200303785, "grad_norm": 1.4579681158065796, "learning_rate": 9.793957226140688e-05, "loss": 3.0416, "step": 52850 }, { "epoch": 0.46715766792066266, "grad_norm": 2.5864365100860596, "learning_rate": 9.793562235491198e-05, "loss": 3.2932, "step": 52900 }, { "epoch": 0.4675992158109469, "grad_norm": 1.7026253938674927, "learning_rate": 9.793166874579673e-05, "loss": 3.1959, "step": 52950 }, { "epoch": 0.46804076370123104, "grad_norm": 5.3310394287109375, "learning_rate": 9.792771143436654e-05, "loss": 3.2207, "step": 53000 }, { "epoch": 0.4684823115915152, "grad_norm": 1.704520583152771, "learning_rate": 9.792375042092704e-05, "loss": 3.1596, "step": 53050 }, { "epoch": 0.4689238594817994, "grad_norm": 2.3095810413360596, "learning_rate": 9.79197857057842e-05, "loss": 3.48, "step": 53100 }, { "epoch": 0.4693654073720836, "grad_norm": 3.7878098487854004, "learning_rate": 9.791581728924427e-05, "loss": 3.4366, "step": 53150 }, { "epoch": 0.46980695526236776, "grad_norm": 1.6754798889160156, "learning_rate": 9.791184517161377e-05, "loss": 3.1798, "step": 53200 }, { "epoch": 0.47024850315265193, "grad_norm": 2.6600844860076904, "learning_rate": 9.79078693531995e-05, "loss": 3.378, "step": 53250 }, { "epoch": 0.4706900510429361, "grad_norm": 1.902320146560669, "learning_rate": 9.790388983430859e-05, "loss": 2.8402, "step": 53300 }, { "epoch": 0.4711315989332203, "grad_norm": 1.9587918519973755, "learning_rate": 9.789990661524839e-05, "loss": 3.6821, "step": 53350 }, { "epoch": 0.4715731468235045, "grad_norm": 0.7448610067367554, "learning_rate": 9.789591969632658e-05, "loss": 3.3176, "step": 53400 }, { "epoch": 0.47201469471378865, "grad_norm": 4.580758094787598, "learning_rate": 9.789192907785111e-05, "loss": 3.2808, "step": 53450 }, { "epoch": 0.4724562426040728, "grad_norm": 2.524980068206787, "learning_rate": 9.788793476013025e-05, "loss": 3.3751, "step": 53500 }, { "epoch": 0.47289779049435704, "grad_norm": 3.0498645305633545, "learning_rate": 9.788393674347249e-05, "loss": 3.1801, "step": 53550 }, { "epoch": 0.4733393383846412, "grad_norm": 2.230107545852661, "learning_rate": 9.787993502818669e-05, "loss": 3.7616, "step": 53600 }, { "epoch": 0.4737808862749254, "grad_norm": 0.8255630731582642, "learning_rate": 9.78759296145819e-05, "loss": 3.0498, "step": 53650 }, { "epoch": 0.47422243416520954, "grad_norm": 4.1301727294921875, "learning_rate": 9.787192050296752e-05, "loss": 3.281, "step": 53700 }, { "epoch": 0.47466398205549376, "grad_norm": 3.9466300010681152, "learning_rate": 9.786790769365322e-05, "loss": 3.1479, "step": 53750 }, { "epoch": 0.4751055299457779, "grad_norm": 1.8435554504394531, "learning_rate": 9.786389118694897e-05, "loss": 3.4114, "step": 53800 }, { "epoch": 0.4755470778360621, "grad_norm": 0.7896223664283752, "learning_rate": 9.785987098316499e-05, "loss": 3.4087, "step": 53850 }, { "epoch": 0.47598862572634626, "grad_norm": 9.002167701721191, "learning_rate": 9.785584708261182e-05, "loss": 3.2257, "step": 53900 }, { "epoch": 0.4764301736166305, "grad_norm": 0.6309633851051331, "learning_rate": 9.785181948560028e-05, "loss": 3.1985, "step": 53950 }, { "epoch": 0.47687172150691465, "grad_norm": 3.459434747695923, "learning_rate": 9.784778819244144e-05, "loss": 3.3152, "step": 54000 }, { "epoch": 0.47687172150691465, "eval_asr_loss": 0.8963724033623304, "eval_loss": 3.0343711376190186, "eval_runtime": 20.7535, "eval_samples_per_second": 37.006, "eval_steps_per_second": 9.251, "eval_tts_loss": 5.906121494862091, "step": 54000 }, { "epoch": 0.4773132693971988, "grad_norm": 2.1301255226135254, "learning_rate": 9.78437532034467e-05, "loss": 3.7664, "step": 54050 }, { "epoch": 0.477754817287483, "grad_norm": 2.864131450653076, "learning_rate": 9.783971451892773e-05, "loss": 3.0815, "step": 54100 }, { "epoch": 0.4781963651777672, "grad_norm": 2.723444938659668, "learning_rate": 9.783567213919649e-05, "loss": 3.2961, "step": 54150 }, { "epoch": 0.47863791306805137, "grad_norm": 2.7087883949279785, "learning_rate": 9.783162606456521e-05, "loss": 3.4694, "step": 54200 }, { "epoch": 0.47907946095833553, "grad_norm": 2.582592248916626, "learning_rate": 9.782757629534642e-05, "loss": 3.4529, "step": 54250 }, { "epoch": 0.4795210088486197, "grad_norm": 1.1580854654312134, "learning_rate": 9.782352283185293e-05, "loss": 3.357, "step": 54300 }, { "epoch": 0.4799625567389039, "grad_norm": 3.442847490310669, "learning_rate": 9.781946567439781e-05, "loss": 3.2244, "step": 54350 }, { "epoch": 0.4804041046291881, "grad_norm": 4.473285675048828, "learning_rate": 9.781540482329447e-05, "loss": 3.2563, "step": 54400 }, { "epoch": 0.48084565251947226, "grad_norm": 2.0757851600646973, "learning_rate": 9.78113402788566e-05, "loss": 3.3494, "step": 54450 }, { "epoch": 0.4812872004097564, "grad_norm": 1.0765964984893799, "learning_rate": 9.78072720413981e-05, "loss": 2.973, "step": 54500 }, { "epoch": 0.48172874830004064, "grad_norm": 2.187049150466919, "learning_rate": 9.780320011123322e-05, "loss": 3.5087, "step": 54550 }, { "epoch": 0.4821702961903248, "grad_norm": 2.3897063732147217, "learning_rate": 9.779912448867649e-05, "loss": 3.3432, "step": 54600 }, { "epoch": 0.482611844080609, "grad_norm": 1.4350836277008057, "learning_rate": 9.779504517404274e-05, "loss": 3.547, "step": 54650 }, { "epoch": 0.48305339197089314, "grad_norm": 1.5999610424041748, "learning_rate": 9.779096216764703e-05, "loss": 3.1709, "step": 54700 }, { "epoch": 0.48349493986117736, "grad_norm": 4.64166784286499, "learning_rate": 9.778687546980474e-05, "loss": 3.6566, "step": 54750 }, { "epoch": 0.48393648775146153, "grad_norm": 2.923461675643921, "learning_rate": 9.778278508083154e-05, "loss": 3.4523, "step": 54800 }, { "epoch": 0.4843780356417457, "grad_norm": 1.1621273756027222, "learning_rate": 9.77786910010434e-05, "loss": 3.4142, "step": 54850 }, { "epoch": 0.48481958353202986, "grad_norm": 2.0562987327575684, "learning_rate": 9.77745932307565e-05, "loss": 3.2347, "step": 54900 }, { "epoch": 0.4852611314223141, "grad_norm": 2.8436999320983887, "learning_rate": 9.777049177028742e-05, "loss": 3.4778, "step": 54950 }, { "epoch": 0.48570267931259825, "grad_norm": 3.21048641204834, "learning_rate": 9.776638661995292e-05, "loss": 3.4173, "step": 55000 }, { "epoch": 0.4861442272028824, "grad_norm": 2.347586154937744, "learning_rate": 9.77622777800701e-05, "loss": 3.6781, "step": 55050 }, { "epoch": 0.4865857750931666, "grad_norm": 1.2790504693984985, "learning_rate": 9.775816525095633e-05, "loss": 3.5449, "step": 55100 }, { "epoch": 0.4870273229834508, "grad_norm": 2.234579563140869, "learning_rate": 9.775404903292929e-05, "loss": 3.403, "step": 55150 }, { "epoch": 0.48746887087373497, "grad_norm": 3.1236841678619385, "learning_rate": 9.774992912630689e-05, "loss": 3.5316, "step": 55200 }, { "epoch": 0.48791041876401914, "grad_norm": 2.067030906677246, "learning_rate": 9.774580553140736e-05, "loss": 3.3101, "step": 55250 }, { "epoch": 0.4883519666543033, "grad_norm": 2.3011932373046875, "learning_rate": 9.774167824854925e-05, "loss": 3.9037, "step": 55300 }, { "epoch": 0.4887935145445875, "grad_norm": 0.8518275618553162, "learning_rate": 9.773754727805132e-05, "loss": 3.7854, "step": 55350 }, { "epoch": 0.4892350624348717, "grad_norm": 2.5740087032318115, "learning_rate": 9.773341262023265e-05, "loss": 3.6587, "step": 55400 }, { "epoch": 0.48967661032515586, "grad_norm": 2.541947364807129, "learning_rate": 9.772927427541266e-05, "loss": 3.3046, "step": 55450 }, { "epoch": 0.49011815821544, "grad_norm": 10.81189250946045, "learning_rate": 9.772513224391093e-05, "loss": 3.5087, "step": 55500 }, { "epoch": 0.49055970610572425, "grad_norm": 2.659097909927368, "learning_rate": 9.772098652604745e-05, "loss": 3.7244, "step": 55550 }, { "epoch": 0.4910012539960084, "grad_norm": 1.6622788906097412, "learning_rate": 9.771683712214241e-05, "loss": 3.9154, "step": 55600 }, { "epoch": 0.4914428018862926, "grad_norm": 4.244743347167969, "learning_rate": 9.771268403251634e-05, "loss": 3.1095, "step": 55650 }, { "epoch": 0.49188434977657675, "grad_norm": 2.233381986618042, "learning_rate": 9.770852725749002e-05, "loss": 2.7214, "step": 55700 }, { "epoch": 0.49232589766686097, "grad_norm": 3.992842674255371, "learning_rate": 9.770436679738452e-05, "loss": 2.8508, "step": 55750 }, { "epoch": 0.49276744555714513, "grad_norm": 2.8735084533691406, "learning_rate": 9.770020265252122e-05, "loss": 3.3086, "step": 55800 }, { "epoch": 0.4932089934474293, "grad_norm": 0.7846609950065613, "learning_rate": 9.769603482322173e-05, "loss": 3.0594, "step": 55850 }, { "epoch": 0.49365054133771347, "grad_norm": 3.95578670501709, "learning_rate": 9.769186330980802e-05, "loss": 3.2712, "step": 55900 }, { "epoch": 0.4940920892279977, "grad_norm": 1.6080371141433716, "learning_rate": 9.768768811260229e-05, "loss": 3.3038, "step": 55950 }, { "epoch": 0.49453363711828185, "grad_norm": 1.84148108959198, "learning_rate": 9.7683509231927e-05, "loss": 3.2961, "step": 56000 }, { "epoch": 0.494975185008566, "grad_norm": 3.487783670425415, "learning_rate": 9.7679326668105e-05, "loss": 3.0265, "step": 56050 }, { "epoch": 0.4954167328988502, "grad_norm": 1.8693515062332153, "learning_rate": 9.767514042145931e-05, "loss": 3.3785, "step": 56100 }, { "epoch": 0.4958582807891344, "grad_norm": 0.690427839756012, "learning_rate": 9.767095049231329e-05, "loss": 2.9811, "step": 56150 }, { "epoch": 0.4962998286794186, "grad_norm": 3.2560925483703613, "learning_rate": 9.766675688099059e-05, "loss": 3.3119, "step": 56200 }, { "epoch": 0.49674137656970274, "grad_norm": 2.757688522338867, "learning_rate": 9.766255958781512e-05, "loss": 3.3075, "step": 56250 }, { "epoch": 0.4971829244599869, "grad_norm": 1.4115214347839355, "learning_rate": 9.765835861311108e-05, "loss": 3.3612, "step": 56300 }, { "epoch": 0.49762447235027113, "grad_norm": 3.8688366413116455, "learning_rate": 9.765415395720298e-05, "loss": 3.1324, "step": 56350 }, { "epoch": 0.4980660202405553, "grad_norm": 3.989609479904175, "learning_rate": 9.764994562041559e-05, "loss": 3.1485, "step": 56400 }, { "epoch": 0.49850756813083946, "grad_norm": 2.1451547145843506, "learning_rate": 9.764573360307394e-05, "loss": 3.2281, "step": 56450 }, { "epoch": 0.49894911602112363, "grad_norm": 1.555163025856018, "learning_rate": 9.76415179055034e-05, "loss": 3.0442, "step": 56500 }, { "epoch": 0.49939066391140785, "grad_norm": 2.0306787490844727, "learning_rate": 9.76372985280296e-05, "loss": 3.3319, "step": 56550 }, { "epoch": 0.499832211801692, "grad_norm": 3.7980337142944336, "learning_rate": 9.763307547097844e-05, "loss": 3.6434, "step": 56600 }, { "epoch": 0.5002737596919762, "grad_norm": 4.4818525314331055, "learning_rate": 9.762884873467611e-05, "loss": 3.2294, "step": 56650 }, { "epoch": 0.5007153075822603, "grad_norm": 1.937999963760376, "learning_rate": 9.76246183194491e-05, "loss": 3.5661, "step": 56700 }, { "epoch": 0.5011568554725445, "grad_norm": 1.306862711906433, "learning_rate": 9.762038422562417e-05, "loss": 2.9257, "step": 56750 }, { "epoch": 0.5015984033628287, "grad_norm": 1.1766592264175415, "learning_rate": 9.761614645352835e-05, "loss": 3.3017, "step": 56800 }, { "epoch": 0.502039951253113, "grad_norm": 3.1086044311523438, "learning_rate": 9.7611905003489e-05, "loss": 3.5308, "step": 56850 }, { "epoch": 0.5024814991433971, "grad_norm": 2.8829550743103027, "learning_rate": 9.760765987583373e-05, "loss": 3.1106, "step": 56900 }, { "epoch": 0.5029230470336813, "grad_norm": 2.7359766960144043, "learning_rate": 9.760341107089044e-05, "loss": 3.509, "step": 56950 }, { "epoch": 0.5033645949239655, "grad_norm": 7.022033214569092, "learning_rate": 9.759915858898732e-05, "loss": 3.1945, "step": 57000 }, { "epoch": 0.5033645949239655, "eval_asr_loss": 0.9032422534916057, "eval_loss": 3.0304603576660156, "eval_runtime": 51.0735, "eval_samples_per_second": 15.037, "eval_steps_per_second": 3.759, "eval_tts_loss": 5.9164860032310695, "step": 57000 }, { "epoch": 0.5038061428142496, "grad_norm": 1.487955093383789, "learning_rate": 9.75949024304528e-05, "loss": 3.1603, "step": 57050 }, { "epoch": 0.5042476907045338, "grad_norm": 2.423590660095215, "learning_rate": 9.759064259561568e-05, "loss": 3.6678, "step": 57100 }, { "epoch": 0.504689238594818, "grad_norm": 1.796743631362915, "learning_rate": 9.758637908480497e-05, "loss": 3.4576, "step": 57150 }, { "epoch": 0.5051307864851021, "grad_norm": 6.50948429107666, "learning_rate": 9.758211189834999e-05, "loss": 3.2279, "step": 57200 }, { "epoch": 0.5055723343753864, "grad_norm": 1.0165084600448608, "learning_rate": 9.757784103658036e-05, "loss": 3.3161, "step": 57250 }, { "epoch": 0.5060138822656706, "grad_norm": 3.4719436168670654, "learning_rate": 9.757356649982595e-05, "loss": 3.6583, "step": 57300 }, { "epoch": 0.5064554301559547, "grad_norm": 2.7984747886657715, "learning_rate": 9.756928828841694e-05, "loss": 3.507, "step": 57350 }, { "epoch": 0.5068969780462389, "grad_norm": 4.911170959472656, "learning_rate": 9.75650064026838e-05, "loss": 3.0097, "step": 57400 }, { "epoch": 0.5073385259365231, "grad_norm": 3.57440185546875, "learning_rate": 9.756072084295725e-05, "loss": 2.9845, "step": 57450 }, { "epoch": 0.5077800738268072, "grad_norm": 1.4227149486541748, "learning_rate": 9.755651743023584e-05, "loss": 3.6389, "step": 57500 }, { "epoch": 0.5082216217170914, "grad_norm": 2.1636316776275635, "learning_rate": 9.755222459697919e-05, "loss": 2.9495, "step": 57550 }, { "epoch": 0.5086631696073756, "grad_norm": 5.281383991241455, "learning_rate": 9.754792809071644e-05, "loss": 3.388, "step": 57600 }, { "epoch": 0.5091047174976598, "grad_norm": 1.8442537784576416, "learning_rate": 9.754362791177944e-05, "loss": 3.7662, "step": 57650 }, { "epoch": 0.509546265387944, "grad_norm": 2.220319986343384, "learning_rate": 9.753932406050032e-05, "loss": 3.1771, "step": 57700 }, { "epoch": 0.5099878132782282, "grad_norm": 2.379897356033325, "learning_rate": 9.753501653721155e-05, "loss": 3.2315, "step": 57750 }, { "epoch": 0.5104293611685123, "grad_norm": 1.3445017337799072, "learning_rate": 9.753070534224584e-05, "loss": 3.4884, "step": 57800 }, { "epoch": 0.5108709090587965, "grad_norm": 1.7846251726150513, "learning_rate": 9.752639047593619e-05, "loss": 3.5118, "step": 57850 }, { "epoch": 0.5113124569490807, "grad_norm": 0.7859761714935303, "learning_rate": 9.752207193861589e-05, "loss": 3.4603, "step": 57900 }, { "epoch": 0.5117540048393648, "grad_norm": 1.91860830783844, "learning_rate": 9.751774973061851e-05, "loss": 3.354, "step": 57950 }, { "epoch": 0.512195552729649, "grad_norm": 2.4331154823303223, "learning_rate": 9.751342385227791e-05, "loss": 3.1493, "step": 58000 }, { "epoch": 0.5126371006199333, "grad_norm": 2.020382881164551, "learning_rate": 9.750909430392821e-05, "loss": 3.4823, "step": 58050 }, { "epoch": 0.5130786485102174, "grad_norm": 1.6145989894866943, "learning_rate": 9.750476108590383e-05, "loss": 3.2045, "step": 58100 }, { "epoch": 0.5135201964005016, "grad_norm": 4.1392364501953125, "learning_rate": 9.750042419853949e-05, "loss": 3.2311, "step": 58150 }, { "epoch": 0.5139617442907858, "grad_norm": 2.6319990158081055, "learning_rate": 9.749608364217018e-05, "loss": 3.4167, "step": 58200 }, { "epoch": 0.51440329218107, "grad_norm": 0.8864327073097229, "learning_rate": 9.749173941713114e-05, "loss": 3.4851, "step": 58250 }, { "epoch": 0.5148448400713541, "grad_norm": 1.8275792598724365, "learning_rate": 9.748739152375799e-05, "loss": 3.429, "step": 58300 }, { "epoch": 0.5152863879616383, "grad_norm": 1.4619648456573486, "learning_rate": 9.74830399623865e-05, "loss": 3.1952, "step": 58350 }, { "epoch": 0.5157279358519224, "grad_norm": 2.958252429962158, "learning_rate": 9.747868473335283e-05, "loss": 3.3281, "step": 58400 }, { "epoch": 0.5161694837422067, "grad_norm": 3.7064812183380127, "learning_rate": 9.747432583699334e-05, "loss": 3.1441, "step": 58450 }, { "epoch": 0.5166110316324909, "grad_norm": 2.135683298110962, "learning_rate": 9.746996327364478e-05, "loss": 3.3074, "step": 58500 }, { "epoch": 0.5170525795227751, "grad_norm": 4.96755838394165, "learning_rate": 9.746559704364409e-05, "loss": 3.1264, "step": 58550 }, { "epoch": 0.5174941274130592, "grad_norm": 1.5454546213150024, "learning_rate": 9.74612271473285e-05, "loss": 3.3129, "step": 58600 }, { "epoch": 0.5179356753033434, "grad_norm": 2.1017255783081055, "learning_rate": 9.74568535850356e-05, "loss": 3.2598, "step": 58650 }, { "epoch": 0.5183772231936276, "grad_norm": 1.2823448181152344, "learning_rate": 9.745247635710318e-05, "loss": 3.6969, "step": 58700 }, { "epoch": 0.5188187710839117, "grad_norm": 2.371886968612671, "learning_rate": 9.744809546386933e-05, "loss": 3.2902, "step": 58750 }, { "epoch": 0.5192603189741959, "grad_norm": 1.7832865715026855, "learning_rate": 9.744371090567246e-05, "loss": 2.8521, "step": 58800 }, { "epoch": 0.5197018668644802, "grad_norm": 1.12856125831604, "learning_rate": 9.743932268285124e-05, "loss": 3.7715, "step": 58850 }, { "epoch": 0.5201434147547643, "grad_norm": 1.2334407567977905, "learning_rate": 9.743493079574461e-05, "loss": 3.2311, "step": 58900 }, { "epoch": 0.5205849626450485, "grad_norm": 2.203479528427124, "learning_rate": 9.743053524469182e-05, "loss": 3.438, "step": 58950 }, { "epoch": 0.5210265105353327, "grad_norm": 0.8005960583686829, "learning_rate": 9.742613603003238e-05, "loss": 3.6582, "step": 59000 }, { "epoch": 0.5214680584256168, "grad_norm": 2.671816110610962, "learning_rate": 9.742173315210608e-05, "loss": 2.9465, "step": 59050 }, { "epoch": 0.521909606315901, "grad_norm": 0.9956416487693787, "learning_rate": 9.741732661125304e-05, "loss": 3.3792, "step": 59100 }, { "epoch": 0.5223511542061852, "grad_norm": 1.9408859014511108, "learning_rate": 9.741291640781359e-05, "loss": 3.2726, "step": 59150 }, { "epoch": 0.5227927020964693, "grad_norm": 1.1133266687393188, "learning_rate": 9.740850254212841e-05, "loss": 3.3387, "step": 59200 }, { "epoch": 0.5232342499867536, "grad_norm": 0.527526319026947, "learning_rate": 9.740408501453841e-05, "loss": 3.0193, "step": 59250 }, { "epoch": 0.5236757978770378, "grad_norm": 1.995932936668396, "learning_rate": 9.7399752285049e-05, "loss": 3.5348, "step": 59300 }, { "epoch": 0.5241173457673219, "grad_norm": 1.4613289833068848, "learning_rate": 9.739532750789442e-05, "loss": 3.0866, "step": 59350 }, { "epoch": 0.5245588936576061, "grad_norm": 2.9688708782196045, "learning_rate": 9.739089906985268e-05, "loss": 3.5847, "step": 59400 }, { "epoch": 0.5250004415478903, "grad_norm": 1.3395789861679077, "learning_rate": 9.738646697126586e-05, "loss": 3.353, "step": 59450 }, { "epoch": 0.5254419894381744, "grad_norm": 4.93052864074707, "learning_rate": 9.738203121247627e-05, "loss": 3.626, "step": 59500 }, { "epoch": 0.5258835373284586, "grad_norm": 3.3152246475219727, "learning_rate": 9.737759179382658e-05, "loss": 2.8786, "step": 59550 }, { "epoch": 0.5263250852187428, "grad_norm": 1.5898215770721436, "learning_rate": 9.737314871565965e-05, "loss": 3.7041, "step": 59600 }, { "epoch": 0.526766633109027, "grad_norm": 2.1526150703430176, "learning_rate": 9.736870197831871e-05, "loss": 3.5867, "step": 59650 }, { "epoch": 0.5272081809993112, "grad_norm": 2.351823568344116, "learning_rate": 9.736425158214723e-05, "loss": 3.4919, "step": 59700 }, { "epoch": 0.5276497288895954, "grad_norm": 2.4668421745300293, "learning_rate": 9.735979752748894e-05, "loss": 3.3577, "step": 59750 }, { "epoch": 0.5280912767798795, "grad_norm": 2.4981229305267334, "learning_rate": 9.73553398146879e-05, "loss": 3.368, "step": 59800 }, { "epoch": 0.5285328246701637, "grad_norm": 1.9264018535614014, "learning_rate": 9.735087844408839e-05, "loss": 3.7379, "step": 59850 }, { "epoch": 0.5289743725604479, "grad_norm": 4.922735691070557, "learning_rate": 9.734641341603507e-05, "loss": 3.1943, "step": 59900 }, { "epoch": 0.529415920450732, "grad_norm": 5.6230645179748535, "learning_rate": 9.73419447308728e-05, "loss": 3.3315, "step": 59950 }, { "epoch": 0.5298574683410162, "grad_norm": 1.985254168510437, "learning_rate": 9.733747238894673e-05, "loss": 3.8071, "step": 60000 }, { "epoch": 0.5298574683410162, "eval_asr_loss": 0.8992706317769239, "eval_loss": 3.0059359073638916, "eval_runtime": 20.6646, "eval_samples_per_second": 37.165, "eval_steps_per_second": 9.291, "eval_tts_loss": 5.9863666861098865, "step": 60000 }, { "epoch": 0.5302990162313005, "grad_norm": 1.029417872428894, "learning_rate": 9.733299639060233e-05, "loss": 3.503, "step": 60050 }, { "epoch": 0.5307405641215847, "grad_norm": 6.666355609893799, "learning_rate": 9.732851673618535e-05, "loss": 3.2843, "step": 60100 }, { "epoch": 0.5311821120118688, "grad_norm": 1.5539653301239014, "learning_rate": 9.732403342604177e-05, "loss": 3.374, "step": 60150 }, { "epoch": 0.531623659902153, "grad_norm": 1.662533164024353, "learning_rate": 9.731954646051792e-05, "loss": 3.1763, "step": 60200 }, { "epoch": 0.5320652077924372, "grad_norm": 1.5471640825271606, "learning_rate": 9.731505583996035e-05, "loss": 3.3524, "step": 60250 }, { "epoch": 0.5325067556827213, "grad_norm": 3.6004583835601807, "learning_rate": 9.731056156471594e-05, "loss": 3.4967, "step": 60300 }, { "epoch": 0.5329483035730055, "grad_norm": 5.77100944519043, "learning_rate": 9.730606363513184e-05, "loss": 3.3075, "step": 60350 }, { "epoch": 0.5333898514632897, "grad_norm": 2.9445252418518066, "learning_rate": 9.730156205155545e-05, "loss": 3.681, "step": 60400 }, { "epoch": 0.5338313993535739, "grad_norm": 3.9473483562469482, "learning_rate": 9.729705681433451e-05, "loss": 2.9724, "step": 60450 }, { "epoch": 0.5342729472438581, "grad_norm": 2.4145290851593018, "learning_rate": 9.729254792381698e-05, "loss": 3.1867, "step": 60500 }, { "epoch": 0.5347144951341423, "grad_norm": 2.9569525718688965, "learning_rate": 9.728803538035119e-05, "loss": 3.6975, "step": 60550 }, { "epoch": 0.5351560430244264, "grad_norm": 3.94183349609375, "learning_rate": 9.728351918428562e-05, "loss": 3.6706, "step": 60600 }, { "epoch": 0.5355975909147106, "grad_norm": 1.8953005075454712, "learning_rate": 9.727899933596917e-05, "loss": 3.3651, "step": 60650 }, { "epoch": 0.5360391388049948, "grad_norm": 1.777334213256836, "learning_rate": 9.727447583575092e-05, "loss": 3.7644, "step": 60700 }, { "epoch": 0.5364806866952789, "grad_norm": 3.474303960800171, "learning_rate": 9.72699486839803e-05, "loss": 3.2085, "step": 60750 }, { "epoch": 0.5369222345855631, "grad_norm": 2.88875412940979, "learning_rate": 9.726541788100696e-05, "loss": 3.8591, "step": 60800 }, { "epoch": 0.5373637824758474, "grad_norm": 3.021052598953247, "learning_rate": 9.726088342718091e-05, "loss": 3.3407, "step": 60850 }, { "epoch": 0.5378053303661315, "grad_norm": 1.9368441104888916, "learning_rate": 9.725634532285238e-05, "loss": 3.4342, "step": 60900 }, { "epoch": 0.5382468782564157, "grad_norm": 1.3616617918014526, "learning_rate": 9.725180356837187e-05, "loss": 3.1015, "step": 60950 }, { "epoch": 0.5386884261466999, "grad_norm": 1.2605106830596924, "learning_rate": 9.724725816409024e-05, "loss": 3.4434, "step": 61000 }, { "epoch": 0.539129974036984, "grad_norm": 2.5226216316223145, "learning_rate": 9.724270911035856e-05, "loss": 3.2771, "step": 61050 }, { "epoch": 0.5395715219272682, "grad_norm": 1.5489667654037476, "learning_rate": 9.723815640752818e-05, "loss": 2.9834, "step": 61100 }, { "epoch": 0.5400130698175524, "grad_norm": 0.3823397159576416, "learning_rate": 9.723360005595082e-05, "loss": 3.4476, "step": 61150 }, { "epoch": 0.5404546177078365, "grad_norm": 1.4170905351638794, "learning_rate": 9.722904005597838e-05, "loss": 2.9736, "step": 61200 }, { "epoch": 0.5408961655981208, "grad_norm": 3.013650417327881, "learning_rate": 9.722447640796306e-05, "loss": 3.4688, "step": 61250 }, { "epoch": 0.541337713488405, "grad_norm": 2.0270910263061523, "learning_rate": 9.721990911225742e-05, "loss": 3.4488, "step": 61300 }, { "epoch": 0.5417792613786891, "grad_norm": 2.3157079219818115, "learning_rate": 9.72153381692142e-05, "loss": 3.4686, "step": 61350 }, { "epoch": 0.5422208092689733, "grad_norm": 2.65665864944458, "learning_rate": 9.721076357918648e-05, "loss": 3.0758, "step": 61400 }, { "epoch": 0.5426623571592575, "grad_norm": 4.086440086364746, "learning_rate": 9.720618534252761e-05, "loss": 3.0561, "step": 61450 }, { "epoch": 0.5431039050495416, "grad_norm": 1.6047959327697754, "learning_rate": 9.720160345959122e-05, "loss": 3.2575, "step": 61500 }, { "epoch": 0.5435454529398258, "grad_norm": 1.371722936630249, "learning_rate": 9.719701793073121e-05, "loss": 3.197, "step": 61550 }, { "epoch": 0.54398700083011, "grad_norm": 1.1012533903121948, "learning_rate": 9.71924287563018e-05, "loss": 3.1246, "step": 61600 }, { "epoch": 0.5444285487203943, "grad_norm": 2.610647201538086, "learning_rate": 9.718783593665745e-05, "loss": 3.4834, "step": 61650 }, { "epoch": 0.5448700966106784, "grad_norm": 3.1460602283477783, "learning_rate": 9.71832394721529e-05, "loss": 3.5381, "step": 61700 }, { "epoch": 0.5453116445009626, "grad_norm": 2.426410675048828, "learning_rate": 9.717863936314322e-05, "loss": 3.0234, "step": 61750 }, { "epoch": 0.5457531923912468, "grad_norm": 1.7792102098464966, "learning_rate": 9.71740356099837e-05, "loss": 3.3962, "step": 61800 }, { "epoch": 0.5461947402815309, "grad_norm": 0.535715639591217, "learning_rate": 9.716942821302995e-05, "loss": 3.2452, "step": 61850 }, { "epoch": 0.5466362881718151, "grad_norm": 3.1257872581481934, "learning_rate": 9.716481717263787e-05, "loss": 3.2172, "step": 61900 }, { "epoch": 0.5470778360620993, "grad_norm": 2.848803997039795, "learning_rate": 9.71602024891636e-05, "loss": 3.5478, "step": 61950 }, { "epoch": 0.5475193839523834, "grad_norm": 2.5586445331573486, "learning_rate": 9.71555841629636e-05, "loss": 3.5599, "step": 62000 }, { "epoch": 0.5479609318426677, "grad_norm": 5.202526569366455, "learning_rate": 9.715096219439458e-05, "loss": 3.3383, "step": 62050 }, { "epoch": 0.5484024797329519, "grad_norm": 7.311855316162109, "learning_rate": 9.714633658381358e-05, "loss": 3.5755, "step": 62100 }, { "epoch": 0.548844027623236, "grad_norm": 2.228634834289551, "learning_rate": 9.714170733157784e-05, "loss": 3.1598, "step": 62150 }, { "epoch": 0.5492855755135202, "grad_norm": 0.7159774899482727, "learning_rate": 9.713707443804499e-05, "loss": 3.3646, "step": 62200 }, { "epoch": 0.5497271234038044, "grad_norm": 1.7473576068878174, "learning_rate": 9.713243790357282e-05, "loss": 3.153, "step": 62250 }, { "epoch": 0.5501686712940885, "grad_norm": 2.541585922241211, "learning_rate": 9.712779772851952e-05, "loss": 3.4916, "step": 62300 }, { "epoch": 0.5506102191843727, "grad_norm": 5.356276512145996, "learning_rate": 9.712315391324346e-05, "loss": 3.2563, "step": 62350 }, { "epoch": 0.5510517670746569, "grad_norm": 2.213956356048584, "learning_rate": 9.711850645810336e-05, "loss": 3.1959, "step": 62400 }, { "epoch": 0.5514933149649411, "grad_norm": 1.0666532516479492, "learning_rate": 9.711385536345818e-05, "loss": 3.4801, "step": 62450 }, { "epoch": 0.5519348628552253, "grad_norm": 5.453732967376709, "learning_rate": 9.71092006296672e-05, "loss": 3.1979, "step": 62500 }, { "epoch": 0.5523764107455095, "grad_norm": 6.855234622955322, "learning_rate": 9.710454225708994e-05, "loss": 3.6221, "step": 62550 }, { "epoch": 0.5528179586357936, "grad_norm": 1.6029174327850342, "learning_rate": 9.709988024608623e-05, "loss": 3.4752, "step": 62600 }, { "epoch": 0.5532595065260778, "grad_norm": 1.8403346538543701, "learning_rate": 9.709521459701616e-05, "loss": 3.4726, "step": 62650 }, { "epoch": 0.553701054416362, "grad_norm": 4.461715221405029, "learning_rate": 9.709054531024011e-05, "loss": 2.9593, "step": 62700 }, { "epoch": 0.5541426023066461, "grad_norm": 0.8452909588813782, "learning_rate": 9.708587238611877e-05, "loss": 3.2217, "step": 62750 }, { "epoch": 0.5545841501969304, "grad_norm": 1.3129770755767822, "learning_rate": 9.708119582501305e-05, "loss": 3.3426, "step": 62800 }, { "epoch": 0.5550256980872146, "grad_norm": 2.4353716373443604, "learning_rate": 9.707651562728419e-05, "loss": 3.4732, "step": 62850 }, { "epoch": 0.5554672459774987, "grad_norm": 1.2511876821517944, "learning_rate": 9.707183179329371e-05, "loss": 3.3568, "step": 62900 }, { "epoch": 0.5559087938677829, "grad_norm": 4.347702980041504, "learning_rate": 9.706714432340336e-05, "loss": 3.1255, "step": 62950 }, { "epoch": 0.5563503417580671, "grad_norm": 1.2785756587982178, "learning_rate": 9.706245321797525e-05, "loss": 3.5741, "step": 63000 }, { "epoch": 0.5563503417580671, "eval_asr_loss": 0.9045116492456787, "eval_loss": 3.014928102493286, "eval_runtime": 20.6473, "eval_samples_per_second": 37.196, "eval_steps_per_second": 9.299, "eval_tts_loss": 5.952768747740511, "step": 63000 }, { "epoch": 0.5567918896483512, "grad_norm": 1.4415526390075684, "learning_rate": 9.705775847737169e-05, "loss": 3.5287, "step": 63050 }, { "epoch": 0.5572334375386354, "grad_norm": 1.9535558223724365, "learning_rate": 9.705306010195533e-05, "loss": 3.3174, "step": 63100 }, { "epoch": 0.5576749854289196, "grad_norm": 1.3182282447814941, "learning_rate": 9.704835809208907e-05, "loss": 3.4299, "step": 63150 }, { "epoch": 0.5581165333192039, "grad_norm": 1.156562328338623, "learning_rate": 9.704365244813613e-05, "loss": 3.2571, "step": 63200 }, { "epoch": 0.558558081209488, "grad_norm": 1.929991364479065, "learning_rate": 9.703894317045993e-05, "loss": 3.0214, "step": 63250 }, { "epoch": 0.5589996290997722, "grad_norm": 1.8567146062850952, "learning_rate": 9.703423025942426e-05, "loss": 3.2539, "step": 63300 }, { "epoch": 0.5594411769900564, "grad_norm": 0.45442625880241394, "learning_rate": 9.702951371539315e-05, "loss": 3.5004, "step": 63350 }, { "epoch": 0.5598827248803405, "grad_norm": 2.3628077507019043, "learning_rate": 9.702479353873089e-05, "loss": 3.1837, "step": 63400 }, { "epoch": 0.5603242727706247, "grad_norm": 3.7257354259490967, "learning_rate": 9.702006972980208e-05, "loss": 3.503, "step": 63450 }, { "epoch": 0.5607658206609089, "grad_norm": 3.1037049293518066, "learning_rate": 9.701534228897163e-05, "loss": 3.2654, "step": 63500 }, { "epoch": 0.561207368551193, "grad_norm": 2.567664861679077, "learning_rate": 9.701061121660464e-05, "loss": 3.3444, "step": 63550 }, { "epoch": 0.5616489164414773, "grad_norm": 3.259284019470215, "learning_rate": 9.700587651306658e-05, "loss": 3.0933, "step": 63600 }, { "epoch": 0.5620904643317615, "grad_norm": 3.6262829303741455, "learning_rate": 9.700113817872317e-05, "loss": 3.4273, "step": 63650 }, { "epoch": 0.5625320122220456, "grad_norm": 2.7692084312438965, "learning_rate": 9.699639621394039e-05, "loss": 3.5773, "step": 63700 }, { "epoch": 0.5629735601123298, "grad_norm": 2.4339802265167236, "learning_rate": 9.699165061908451e-05, "loss": 2.8369, "step": 63750 }, { "epoch": 0.563415108002614, "grad_norm": 1.9642480611801147, "learning_rate": 9.69869013945221e-05, "loss": 3.533, "step": 63800 }, { "epoch": 0.5638566558928981, "grad_norm": 4.266359329223633, "learning_rate": 9.698214854062e-05, "loss": 3.3043, "step": 63850 }, { "epoch": 0.5642982037831823, "grad_norm": 1.5324000120162964, "learning_rate": 9.697739205774532e-05, "loss": 3.6485, "step": 63900 }, { "epoch": 0.5647397516734665, "grad_norm": 3.5735909938812256, "learning_rate": 9.6972727184053e-05, "loss": 3.3017, "step": 63950 }, { "epoch": 0.5651812995637507, "grad_norm": 1.4858098030090332, "learning_rate": 9.696796351689678e-05, "loss": 3.4435, "step": 64000 }, { "epoch": 0.5656228474540349, "grad_norm": 3.1134603023529053, "learning_rate": 9.696319622186367e-05, "loss": 2.8183, "step": 64050 }, { "epoch": 0.5660643953443191, "grad_norm": 0.87113356590271, "learning_rate": 9.695842529932186e-05, "loss": 3.5863, "step": 64100 }, { "epoch": 0.5665059432346032, "grad_norm": 2.494359016418457, "learning_rate": 9.695365074963992e-05, "loss": 3.2139, "step": 64150 }, { "epoch": 0.5669474911248874, "grad_norm": 4.142000198364258, "learning_rate": 9.694887257318659e-05, "loss": 3.4118, "step": 64200 }, { "epoch": 0.5673890390151716, "grad_norm": 7.738039493560791, "learning_rate": 9.694409077033097e-05, "loss": 3.3375, "step": 64250 }, { "epoch": 0.5678305869054557, "grad_norm": 2.2657651901245117, "learning_rate": 9.693930534144243e-05, "loss": 3.5303, "step": 64300 }, { "epoch": 0.5682721347957399, "grad_norm": 1.5491681098937988, "learning_rate": 9.693451628689059e-05, "loss": 3.1383, "step": 64350 }, { "epoch": 0.5687136826860242, "grad_norm": 5.7371745109558105, "learning_rate": 9.692972360704534e-05, "loss": 3.3374, "step": 64400 }, { "epoch": 0.5691552305763083, "grad_norm": 1.6547493934631348, "learning_rate": 9.69249273022769e-05, "loss": 3.5325, "step": 64450 }, { "epoch": 0.5695967784665925, "grad_norm": 4.246345043182373, "learning_rate": 9.692012737295574e-05, "loss": 3.487, "step": 64500 }, { "epoch": 0.5700383263568767, "grad_norm": 4.030117988586426, "learning_rate": 9.69153238194526e-05, "loss": 3.0569, "step": 64550 }, { "epoch": 0.5704798742471608, "grad_norm": 1.0426106452941895, "learning_rate": 9.691051664213855e-05, "loss": 3.3855, "step": 64600 }, { "epoch": 0.570921422137445, "grad_norm": 1.773262619972229, "learning_rate": 9.690570584138486e-05, "loss": 3.0886, "step": 64650 }, { "epoch": 0.5713629700277292, "grad_norm": 2.984727144241333, "learning_rate": 9.690089141756316e-05, "loss": 3.3807, "step": 64700 }, { "epoch": 0.5718045179180133, "grad_norm": 2.3156557083129883, "learning_rate": 9.689607337104528e-05, "loss": 3.3806, "step": 64750 }, { "epoch": 0.5722460658082976, "grad_norm": 3.1280009746551514, "learning_rate": 9.689125170220341e-05, "loss": 3.223, "step": 64800 }, { "epoch": 0.5726876136985818, "grad_norm": 1.362710952758789, "learning_rate": 9.688642641140999e-05, "loss": 2.7777, "step": 64850 }, { "epoch": 0.573129161588866, "grad_norm": 2.129289150238037, "learning_rate": 9.68815974990377e-05, "loss": 3.2796, "step": 64900 }, { "epoch": 0.5735707094791501, "grad_norm": 3.2772629261016846, "learning_rate": 9.687676496545955e-05, "loss": 3.4964, "step": 64950 }, { "epoch": 0.5740122573694343, "grad_norm": 2.643249034881592, "learning_rate": 9.68719288110488e-05, "loss": 3.6755, "step": 65000 }, { "epoch": 0.5744538052597185, "grad_norm": 1.4029085636138916, "learning_rate": 9.686708903617902e-05, "loss": 3.1452, "step": 65050 }, { "epoch": 0.5748953531500026, "grad_norm": 0.7895709276199341, "learning_rate": 9.686224564122403e-05, "loss": 3.233, "step": 65100 }, { "epoch": 0.5753369010402868, "grad_norm": 1.5120134353637695, "learning_rate": 9.685739862655793e-05, "loss": 3.2801, "step": 65150 }, { "epoch": 0.5757784489305711, "grad_norm": 1.919240117073059, "learning_rate": 9.685254799255517e-05, "loss": 3.3301, "step": 65200 }, { "epoch": 0.5762199968208552, "grad_norm": 2.2543394565582275, "learning_rate": 9.684769373959033e-05, "loss": 2.9593, "step": 65250 }, { "epoch": 0.5766615447111394, "grad_norm": 1.0832370519638062, "learning_rate": 9.684283586803843e-05, "loss": 3.6906, "step": 65300 }, { "epoch": 0.5771030926014236, "grad_norm": 3.100877285003662, "learning_rate": 9.683797437827466e-05, "loss": 3.347, "step": 65350 }, { "epoch": 0.5775446404917077, "grad_norm": 3.069957733154297, "learning_rate": 9.683310927067455e-05, "loss": 3.0518, "step": 65400 }, { "epoch": 0.5779861883819919, "grad_norm": 1.4712010622024536, "learning_rate": 9.682824054561389e-05, "loss": 3.2539, "step": 65450 }, { "epoch": 0.5784277362722761, "grad_norm": 1.5608115196228027, "learning_rate": 9.682336820346874e-05, "loss": 3.4519, "step": 65500 }, { "epoch": 0.5788692841625602, "grad_norm": 2.5109148025512695, "learning_rate": 9.681849224461544e-05, "loss": 3.3817, "step": 65550 }, { "epoch": 0.5793108320528445, "grad_norm": 3.372307300567627, "learning_rate": 9.68136126694306e-05, "loss": 3.4769, "step": 65600 }, { "epoch": 0.5797523799431287, "grad_norm": 2.304180860519409, "learning_rate": 9.680872947829118e-05, "loss": 3.8492, "step": 65650 }, { "epoch": 0.5801939278334128, "grad_norm": 0.9315201044082642, "learning_rate": 9.680384267157434e-05, "loss": 3.7529, "step": 65700 }, { "epoch": 0.580635475723697, "grad_norm": 1.882952332496643, "learning_rate": 9.679895224965752e-05, "loss": 3.0259, "step": 65750 }, { "epoch": 0.5810770236139812, "grad_norm": 3.846442222595215, "learning_rate": 9.679405821291849e-05, "loss": 3.0785, "step": 65800 }, { "epoch": 0.5815185715042653, "grad_norm": 1.7225450277328491, "learning_rate": 9.678916056173526e-05, "loss": 3.5125, "step": 65850 }, { "epoch": 0.5819601193945495, "grad_norm": 0.8811081051826477, "learning_rate": 9.678425929648614e-05, "loss": 3.0916, "step": 65900 }, { "epoch": 0.5824016672848337, "grad_norm": 3.4488401412963867, "learning_rate": 9.67793544175497e-05, "loss": 3.0175, "step": 65950 }, { "epoch": 0.582843215175118, "grad_norm": 2.16182804107666, "learning_rate": 9.677444592530483e-05, "loss": 3.6373, "step": 66000 }, { "epoch": 0.582843215175118, "eval_asr_loss": 0.8874591935323445, "eval_loss": 3.00752854347229, "eval_runtime": 21.0653, "eval_samples_per_second": 36.458, "eval_steps_per_second": 9.115, "eval_tts_loss": 5.963853557317323, "step": 66000 }, { "epoch": 0.5832847630654021, "grad_norm": 2.242957353591919, "learning_rate": 9.676953382013063e-05, "loss": 3.0162, "step": 66050 }, { "epoch": 0.5837263109556863, "grad_norm": 2.5106394290924072, "learning_rate": 9.676461810240654e-05, "loss": 3.4729, "step": 66100 }, { "epoch": 0.5841678588459704, "grad_norm": 3.770650863647461, "learning_rate": 9.675969877251225e-05, "loss": 3.4405, "step": 66150 }, { "epoch": 0.5846094067362546, "grad_norm": 2.2024853229522705, "learning_rate": 9.675477583082775e-05, "loss": 3.5091, "step": 66200 }, { "epoch": 0.5850509546265388, "grad_norm": 5.273355007171631, "learning_rate": 9.674984927773328e-05, "loss": 3.5578, "step": 66250 }, { "epoch": 0.5854925025168229, "grad_norm": 2.3930509090423584, "learning_rate": 9.674491911360939e-05, "loss": 3.269, "step": 66300 }, { "epoch": 0.5859340504071071, "grad_norm": 0.5656553506851196, "learning_rate": 9.673998533883687e-05, "loss": 3.4693, "step": 66350 }, { "epoch": 0.5863755982973914, "grad_norm": 1.1394141912460327, "learning_rate": 9.673504795379683e-05, "loss": 3.4782, "step": 66400 }, { "epoch": 0.5868171461876756, "grad_norm": 2.660053253173828, "learning_rate": 9.673010695887064e-05, "loss": 3.5723, "step": 66450 }, { "epoch": 0.5872586940779597, "grad_norm": 2.931849718093872, "learning_rate": 9.672516235443994e-05, "loss": 3.0214, "step": 66500 }, { "epoch": 0.5877002419682439, "grad_norm": 3.5768563747406006, "learning_rate": 9.672021414088667e-05, "loss": 3.0214, "step": 66550 }, { "epoch": 0.588141789858528, "grad_norm": 6.038229465484619, "learning_rate": 9.671526231859305e-05, "loss": 3.4275, "step": 66600 }, { "epoch": 0.5885833377488122, "grad_norm": 3.714881181716919, "learning_rate": 9.671030688794153e-05, "loss": 2.9659, "step": 66650 }, { "epoch": 0.5890248856390964, "grad_norm": 1.5607867240905762, "learning_rate": 9.670544706544311e-05, "loss": 3.4904, "step": 66700 }, { "epoch": 0.5894664335293806, "grad_norm": 1.8332650661468506, "learning_rate": 9.67004844913725e-05, "loss": 3.0832, "step": 66750 }, { "epoch": 0.5899079814196648, "grad_norm": 2.306621789932251, "learning_rate": 9.669551831008545e-05, "loss": 3.1016, "step": 66800 }, { "epoch": 0.590349529309949, "grad_norm": 1.265374779701233, "learning_rate": 9.66905485219656e-05, "loss": 3.3778, "step": 66850 }, { "epoch": 0.5907910772002332, "grad_norm": 2.479994773864746, "learning_rate": 9.66855751273968e-05, "loss": 3.3998, "step": 66900 }, { "epoch": 0.5912326250905173, "grad_norm": 2.6619317531585693, "learning_rate": 9.66805981267632e-05, "loss": 3.0006, "step": 66950 }, { "epoch": 0.5916741729808015, "grad_norm": 2.5897438526153564, "learning_rate": 9.667561752044922e-05, "loss": 3.6474, "step": 67000 }, { "epoch": 0.5921157208710857, "grad_norm": 5.512159824371338, "learning_rate": 9.667063330883961e-05, "loss": 3.457, "step": 67050 }, { "epoch": 0.5925572687613698, "grad_norm": 2.0321197509765625, "learning_rate": 9.666564549231931e-05, "loss": 3.653, "step": 67100 }, { "epoch": 0.592998816651654, "grad_norm": 2.648627519607544, "learning_rate": 9.666065407127361e-05, "loss": 2.8554, "step": 67150 }, { "epoch": 0.5934403645419383, "grad_norm": 2.1089653968811035, "learning_rate": 9.665565904608806e-05, "loss": 3.0816, "step": 67200 }, { "epoch": 0.5938819124322224, "grad_norm": 0.6567474007606506, "learning_rate": 9.665066041714849e-05, "loss": 3.4614, "step": 67250 }, { "epoch": 0.5943234603225066, "grad_norm": 4.268828392028809, "learning_rate": 9.664565818484097e-05, "loss": 3.1899, "step": 67300 }, { "epoch": 0.5947650082127908, "grad_norm": 4.196305751800537, "learning_rate": 9.664065234955191e-05, "loss": 3.3985, "step": 67350 }, { "epoch": 0.5952065561030749, "grad_norm": 4.905257225036621, "learning_rate": 9.663564291166795e-05, "loss": 3.3829, "step": 67400 }, { "epoch": 0.5956481039933591, "grad_norm": 3.6008496284484863, "learning_rate": 9.663062987157604e-05, "loss": 2.946, "step": 67450 }, { "epoch": 0.5960896518836433, "grad_norm": 1.2707490921020508, "learning_rate": 9.66256132296634e-05, "loss": 3.4213, "step": 67500 }, { "epoch": 0.5965311997739274, "grad_norm": 3.5220985412597656, "learning_rate": 9.662059298631749e-05, "loss": 3.0964, "step": 67550 }, { "epoch": 0.5969727476642117, "grad_norm": 1.3247735500335693, "learning_rate": 9.66155691419261e-05, "loss": 3.4796, "step": 67600 }, { "epoch": 0.5974142955544959, "grad_norm": 1.086634874343872, "learning_rate": 9.66105416968773e-05, "loss": 3.4287, "step": 67650 }, { "epoch": 0.59785584344478, "grad_norm": 3.1261913776397705, "learning_rate": 9.660551065155938e-05, "loss": 3.239, "step": 67700 }, { "epoch": 0.5982973913350642, "grad_norm": 5.029034614562988, "learning_rate": 9.660047600636099e-05, "loss": 2.8555, "step": 67750 }, { "epoch": 0.5987389392253484, "grad_norm": 2.428929328918457, "learning_rate": 9.659543776167097e-05, "loss": 3.331, "step": 67800 }, { "epoch": 0.5991804871156325, "grad_norm": 2.1692111492156982, "learning_rate": 9.65903959178785e-05, "loss": 3.2252, "step": 67850 }, { "epoch": 0.5996220350059167, "grad_norm": 5.579278945922852, "learning_rate": 9.658535047537303e-05, "loss": 3.2612, "step": 67900 }, { "epoch": 0.6000635828962009, "grad_norm": 1.7974921464920044, "learning_rate": 9.658030143454426e-05, "loss": 3.3738, "step": 67950 }, { "epoch": 0.6005051307864852, "grad_norm": 2.329169750213623, "learning_rate": 9.657524879578221e-05, "loss": 3.358, "step": 68000 }, { "epoch": 0.6009466786767693, "grad_norm": 2.0072598457336426, "learning_rate": 9.657019255947712e-05, "loss": 3.1542, "step": 68050 }, { "epoch": 0.6013882265670535, "grad_norm": 1.212756872177124, "learning_rate": 9.656513272601957e-05, "loss": 2.8526, "step": 68100 }, { "epoch": 0.6018297744573377, "grad_norm": 2.1172633171081543, "learning_rate": 9.656006929580036e-05, "loss": 3.3746, "step": 68150 }, { "epoch": 0.6022713223476218, "grad_norm": 3.0865988731384277, "learning_rate": 9.655500226921064e-05, "loss": 3.4431, "step": 68200 }, { "epoch": 0.602712870237906, "grad_norm": 2.7705607414245605, "learning_rate": 9.654993164664175e-05, "loss": 3.3185, "step": 68250 }, { "epoch": 0.6031544181281902, "grad_norm": 2.9133379459381104, "learning_rate": 9.654485742848538e-05, "loss": 3.2805, "step": 68300 }, { "epoch": 0.6035959660184743, "grad_norm": 1.945061445236206, "learning_rate": 9.653977961513347e-05, "loss": 3.5587, "step": 68350 }, { "epoch": 0.6040375139087586, "grad_norm": 0.5719209313392639, "learning_rate": 9.653469820697822e-05, "loss": 3.039, "step": 68400 }, { "epoch": 0.6044790617990428, "grad_norm": 2.7784836292266846, "learning_rate": 9.652961320441214e-05, "loss": 3.6163, "step": 68450 }, { "epoch": 0.6049206096893269, "grad_norm": 1.212082028388977, "learning_rate": 9.652452460782799e-05, "loss": 2.763, "step": 68500 }, { "epoch": 0.6053621575796111, "grad_norm": 6.158472061157227, "learning_rate": 9.651943241761884e-05, "loss": 3.4118, "step": 68550 }, { "epoch": 0.6058037054698953, "grad_norm": 1.3418790102005005, "learning_rate": 9.6514336634178e-05, "loss": 3.3112, "step": 68600 }, { "epoch": 0.6062452533601794, "grad_norm": 2.3166797161102295, "learning_rate": 9.65092372578991e-05, "loss": 3.3965, "step": 68650 }, { "epoch": 0.6066868012504636, "grad_norm": 2.7341291904449463, "learning_rate": 9.6504134289176e-05, "loss": 3.4362, "step": 68700 }, { "epoch": 0.6071283491407478, "grad_norm": 5.785979270935059, "learning_rate": 9.649902772840287e-05, "loss": 3.5519, "step": 68750 }, { "epoch": 0.607569897031032, "grad_norm": 1.710016131401062, "learning_rate": 9.649391757597414e-05, "loss": 3.8108, "step": 68800 }, { "epoch": 0.6080114449213162, "grad_norm": 1.7782738208770752, "learning_rate": 9.648880383228455e-05, "loss": 3.0538, "step": 68850 }, { "epoch": 0.6084529928116004, "grad_norm": 2.464695930480957, "learning_rate": 9.648368649772907e-05, "loss": 3.4832, "step": 68900 }, { "epoch": 0.6088945407018845, "grad_norm": 2.7873873710632324, "learning_rate": 9.647856557270299e-05, "loss": 3.5608, "step": 68950 }, { "epoch": 0.6093360885921687, "grad_norm": 2.8136725425720215, "learning_rate": 9.647344105760183e-05, "loss": 2.8681, "step": 69000 }, { "epoch": 0.6093360885921687, "eval_asr_loss": 0.9085685970427385, "eval_loss": 3.00970458984375, "eval_runtime": 20.9691, "eval_samples_per_second": 36.625, "eval_steps_per_second": 9.156, "eval_tts_loss": 5.906105015571997, "step": 69000 }, { "epoch": 0.6097776364824529, "grad_norm": 1.4982023239135742, "learning_rate": 9.646831295282143e-05, "loss": 3.7657, "step": 69050 }, { "epoch": 0.610219184372737, "grad_norm": 1.8872673511505127, "learning_rate": 9.646318125875791e-05, "loss": 3.1832, "step": 69100 }, { "epoch": 0.6106607322630212, "grad_norm": 3.1537160873413086, "learning_rate": 9.645804597580761e-05, "loss": 3.028, "step": 69150 }, { "epoch": 0.6111022801533055, "grad_norm": 2.5125792026519775, "learning_rate": 9.645290710436722e-05, "loss": 3.2232, "step": 69200 }, { "epoch": 0.6115438280435896, "grad_norm": 2.1751527786254883, "learning_rate": 9.644776464483368e-05, "loss": 3.1004, "step": 69250 }, { "epoch": 0.6119853759338738, "grad_norm": 1.4385361671447754, "learning_rate": 9.644261859760417e-05, "loss": 2.8845, "step": 69300 }, { "epoch": 0.612426923824158, "grad_norm": 1.0426270961761475, "learning_rate": 9.643746896307622e-05, "loss": 3.3967, "step": 69350 }, { "epoch": 0.6128684717144421, "grad_norm": 5.886585712432861, "learning_rate": 9.643231574164755e-05, "loss": 3.2875, "step": 69400 }, { "epoch": 0.6133100196047263, "grad_norm": 3.5066511631011963, "learning_rate": 9.642715893371622e-05, "loss": 2.8433, "step": 69450 }, { "epoch": 0.6137515674950105, "grad_norm": 2.092880964279175, "learning_rate": 9.642199853968055e-05, "loss": 3.2078, "step": 69500 }, { "epoch": 0.6141931153852946, "grad_norm": 1.670851707458496, "learning_rate": 9.641683455993916e-05, "loss": 3.0111, "step": 69550 }, { "epoch": 0.6146346632755789, "grad_norm": 3.2171874046325684, "learning_rate": 9.641177038132528e-05, "loss": 2.9596, "step": 69600 }, { "epoch": 0.6150762111658631, "grad_norm": 1.39114248752594, "learning_rate": 9.640659930306355e-05, "loss": 3.2125, "step": 69650 }, { "epoch": 0.6155177590561473, "grad_norm": 0.8414401412010193, "learning_rate": 9.640142464028551e-05, "loss": 3.3546, "step": 69700 }, { "epoch": 0.6159593069464314, "grad_norm": 2.521662473678589, "learning_rate": 9.639624639339092e-05, "loss": 3.2137, "step": 69750 }, { "epoch": 0.6164008548367156, "grad_norm": 2.6045455932617188, "learning_rate": 9.63910645627797e-05, "loss": 3.4016, "step": 69800 }, { "epoch": 0.6168424027269998, "grad_norm": 2.437931537628174, "learning_rate": 9.638587914885215e-05, "loss": 3.0942, "step": 69850 }, { "epoch": 0.6172839506172839, "grad_norm": 2.744288682937622, "learning_rate": 9.638069015200877e-05, "loss": 3.4627, "step": 69900 }, { "epoch": 0.6177254985075681, "grad_norm": 8.556507110595703, "learning_rate": 9.637549757265037e-05, "loss": 2.9197, "step": 69950 }, { "epoch": 0.6181670463978524, "grad_norm": 2.708615303039551, "learning_rate": 9.637030141117803e-05, "loss": 3.4477, "step": 70000 }, { "epoch": 0.6186085942881365, "grad_norm": 2.7415497303009033, "learning_rate": 9.636510166799313e-05, "loss": 3.3546, "step": 70050 }, { "epoch": 0.6190501421784207, "grad_norm": 2.8086228370666504, "learning_rate": 9.635989834349728e-05, "loss": 3.3978, "step": 70100 }, { "epoch": 0.6194916900687049, "grad_norm": 2.983367681503296, "learning_rate": 9.635469143809239e-05, "loss": 3.243, "step": 70150 }, { "epoch": 0.619933237958989, "grad_norm": 0.500150740146637, "learning_rate": 9.634948095218068e-05, "loss": 3.3688, "step": 70200 }, { "epoch": 0.6203747858492732, "grad_norm": 2.545722484588623, "learning_rate": 9.63442668861646e-05, "loss": 3.6102, "step": 70250 }, { "epoch": 0.6208163337395574, "grad_norm": 0.7100191712379456, "learning_rate": 9.633904924044687e-05, "loss": 3.7465, "step": 70300 }, { "epoch": 0.6212578816298415, "grad_norm": 1.9451789855957031, "learning_rate": 9.633382801543055e-05, "loss": 3.3812, "step": 70350 }, { "epoch": 0.6216994295201258, "grad_norm": 2.1257078647613525, "learning_rate": 9.632860321151892e-05, "loss": 3.0553, "step": 70400 }, { "epoch": 0.62214097741041, "grad_norm": 0.8686696290969849, "learning_rate": 9.632337482911553e-05, "loss": 2.8603, "step": 70450 }, { "epoch": 0.6225825253006941, "grad_norm": 6.291978359222412, "learning_rate": 9.631814286862426e-05, "loss": 3.1015, "step": 70500 }, { "epoch": 0.6230240731909783, "grad_norm": 3.416860818862915, "learning_rate": 9.631290733044921e-05, "loss": 3.5239, "step": 70550 }, { "epoch": 0.6234656210812625, "grad_norm": 1.6016830205917358, "learning_rate": 9.63076682149948e-05, "loss": 3.2555, "step": 70600 }, { "epoch": 0.6239071689715466, "grad_norm": 2.5389463901519775, "learning_rate": 9.630242552266569e-05, "loss": 3.5458, "step": 70650 }, { "epoch": 0.6243487168618308, "grad_norm": 2.5725655555725098, "learning_rate": 9.629717925386683e-05, "loss": 3.0217, "step": 70700 }, { "epoch": 0.624790264752115, "grad_norm": 2.941615581512451, "learning_rate": 9.629192940900348e-05, "loss": 3.2919, "step": 70750 }, { "epoch": 0.6252318126423992, "grad_norm": 2.9366567134857178, "learning_rate": 9.628667598848113e-05, "loss": 3.0738, "step": 70800 }, { "epoch": 0.6256733605326834, "grad_norm": 4.780130863189697, "learning_rate": 9.628141899270554e-05, "loss": 3.4164, "step": 70850 }, { "epoch": 0.6261149084229676, "grad_norm": 2.4676146507263184, "learning_rate": 9.62761584220828e-05, "loss": 2.6195, "step": 70900 }, { "epoch": 0.6265564563132517, "grad_norm": 2.815596103668213, "learning_rate": 9.627089427701923e-05, "loss": 3.1961, "step": 70950 }, { "epoch": 0.6269980042035359, "grad_norm": 1.198612093925476, "learning_rate": 9.626562655792145e-05, "loss": 3.446, "step": 71000 }, { "epoch": 0.6274395520938201, "grad_norm": 0.4727330803871155, "learning_rate": 9.626035526519632e-05, "loss": 3.252, "step": 71050 }, { "epoch": 0.6278810999841042, "grad_norm": 2.9533727169036865, "learning_rate": 9.625508039925104e-05, "loss": 3.3784, "step": 71100 }, { "epoch": 0.6283226478743884, "grad_norm": 1.3359452486038208, "learning_rate": 9.624980196049303e-05, "loss": 3.5926, "step": 71150 }, { "epoch": 0.6287641957646727, "grad_norm": 1.3383703231811523, "learning_rate": 9.624451994932999e-05, "loss": 3.4367, "step": 71200 }, { "epoch": 0.6292057436549569, "grad_norm": 3.669863224029541, "learning_rate": 9.623923436616996e-05, "loss": 3.4483, "step": 71250 }, { "epoch": 0.629647291545241, "grad_norm": 3.541435718536377, "learning_rate": 9.623394521142113e-05, "loss": 3.5803, "step": 71300 }, { "epoch": 0.6300888394355252, "grad_norm": 1.2664039134979248, "learning_rate": 9.622865248549211e-05, "loss": 3.02, "step": 71350 }, { "epoch": 0.6305303873258093, "grad_norm": 3.8231582641601562, "learning_rate": 9.622335618879168e-05, "loss": 3.2505, "step": 71400 }, { "epoch": 0.6309719352160935, "grad_norm": 1.3903136253356934, "learning_rate": 9.621805632172896e-05, "loss": 2.9738, "step": 71450 }, { "epoch": 0.6314134831063777, "grad_norm": 1.4685653448104858, "learning_rate": 9.621275288471329e-05, "loss": 3.4252, "step": 71500 }, { "epoch": 0.6318550309966618, "grad_norm": 4.989869594573975, "learning_rate": 9.620744587815435e-05, "loss": 3.2167, "step": 71550 }, { "epoch": 0.6322965788869461, "grad_norm": 1.4381548166275024, "learning_rate": 9.620213530246205e-05, "loss": 3.5436, "step": 71600 }, { "epoch": 0.6327381267772303, "grad_norm": 1.2252341508865356, "learning_rate": 9.619682115804656e-05, "loss": 3.0586, "step": 71650 }, { "epoch": 0.6331796746675145, "grad_norm": 4.090355396270752, "learning_rate": 9.619150344531838e-05, "loss": 3.4366, "step": 71700 }, { "epoch": 0.6336212225577986, "grad_norm": 1.2702091932296753, "learning_rate": 9.618618216468824e-05, "loss": 2.9698, "step": 71750 }, { "epoch": 0.6340627704480828, "grad_norm": 3.03657865524292, "learning_rate": 9.61808573165672e-05, "loss": 3.0061, "step": 71800 }, { "epoch": 0.634504318338367, "grad_norm": 2.344271421432495, "learning_rate": 9.617552890136652e-05, "loss": 2.6665, "step": 71850 }, { "epoch": 0.6349458662286511, "grad_norm": 1.6858032941818237, "learning_rate": 9.61701969194978e-05, "loss": 3.1685, "step": 71900 }, { "epoch": 0.6353874141189353, "grad_norm": 3.873769760131836, "learning_rate": 9.616486137137287e-05, "loss": 3.5466, "step": 71950 }, { "epoch": 0.6358289620092196, "grad_norm": 1.7075845003128052, "learning_rate": 9.615952225740385e-05, "loss": 3.3663, "step": 72000 }, { "epoch": 0.6358289620092196, "eval_asr_loss": 0.9085645562718005, "eval_loss": 3.0017716884613037, "eval_runtime": 20.9003, "eval_samples_per_second": 36.746, "eval_steps_per_second": 9.186, "eval_tts_loss": 5.94018095728795, "step": 72000 }, { "epoch": 0.6362705098995037, "grad_norm": 2.023348331451416, "learning_rate": 9.615417957800319e-05, "loss": 3.2043, "step": 72050 }, { "epoch": 0.6367120577897879, "grad_norm": 1.3778334856033325, "learning_rate": 9.614883333358351e-05, "loss": 3.2908, "step": 72100 }, { "epoch": 0.6371536056800721, "grad_norm": 2.4582135677337646, "learning_rate": 9.61434835245578e-05, "loss": 3.4521, "step": 72150 }, { "epoch": 0.6375951535703562, "grad_norm": 4.403940200805664, "learning_rate": 9.613813015133926e-05, "loss": 3.1724, "step": 72200 }, { "epoch": 0.6380367014606404, "grad_norm": 1.1220415830612183, "learning_rate": 9.613277321434141e-05, "loss": 3.2921, "step": 72250 }, { "epoch": 0.6384782493509246, "grad_norm": 5.00771427154541, "learning_rate": 9.612741271397802e-05, "loss": 3.4677, "step": 72300 }, { "epoch": 0.6389197972412087, "grad_norm": 2.081688404083252, "learning_rate": 9.612204865066317e-05, "loss": 2.9708, "step": 72350 }, { "epoch": 0.639361345131493, "grad_norm": 1.0858304500579834, "learning_rate": 9.611668102481114e-05, "loss": 3.3104, "step": 72400 }, { "epoch": 0.6398028930217772, "grad_norm": 4.691798686981201, "learning_rate": 9.611141729550216e-05, "loss": 3.4249, "step": 72450 }, { "epoch": 0.6402444409120613, "grad_norm": 1.6770505905151367, "learning_rate": 9.610604261705e-05, "loss": 3.3248, "step": 72500 }, { "epoch": 0.6406859888023455, "grad_norm": 0.5892982482910156, "learning_rate": 9.610066437729704e-05, "loss": 3.2216, "step": 72550 }, { "epoch": 0.6411275366926297, "grad_norm": 2.631664514541626, "learning_rate": 9.609528257665864e-05, "loss": 3.4193, "step": 72600 }, { "epoch": 0.6415690845829138, "grad_norm": 1.2707042694091797, "learning_rate": 9.608989721555055e-05, "loss": 3.3635, "step": 72650 }, { "epoch": 0.642010632473198, "grad_norm": 3.7268316745758057, "learning_rate": 9.608450829438873e-05, "loss": 3.5901, "step": 72700 }, { "epoch": 0.6424521803634822, "grad_norm": 3.723600387573242, "learning_rate": 9.607911581358943e-05, "loss": 3.157, "step": 72750 }, { "epoch": 0.6428937282537664, "grad_norm": 1.398940086364746, "learning_rate": 9.607371977356918e-05, "loss": 3.2389, "step": 72800 }, { "epoch": 0.6433352761440506, "grad_norm": 3.619192600250244, "learning_rate": 9.606832017474477e-05, "loss": 3.5136, "step": 72850 }, { "epoch": 0.6437768240343348, "grad_norm": 1.5558445453643799, "learning_rate": 9.606291701753327e-05, "loss": 3.4748, "step": 72900 }, { "epoch": 0.644218371924619, "grad_norm": 3.18674635887146, "learning_rate": 9.605751030235204e-05, "loss": 2.6554, "step": 72950 }, { "epoch": 0.6446599198149031, "grad_norm": 4.52305793762207, "learning_rate": 9.605210002961868e-05, "loss": 3.6155, "step": 73000 }, { "epoch": 0.6451014677051873, "grad_norm": 5.28046989440918, "learning_rate": 9.604668619975112e-05, "loss": 3.404, "step": 73050 }, { "epoch": 0.6455430155954714, "grad_norm": 2.169076442718506, "learning_rate": 9.604126881316751e-05, "loss": 3.3904, "step": 73100 }, { "epoch": 0.6459845634857556, "grad_norm": 1.4948843717575073, "learning_rate": 9.60358478702863e-05, "loss": 3.1796, "step": 73150 }, { "epoch": 0.6464261113760399, "grad_norm": 3.89113450050354, "learning_rate": 9.603042337152622e-05, "loss": 3.6965, "step": 73200 }, { "epoch": 0.6468676592663241, "grad_norm": 0.8911652565002441, "learning_rate": 9.602499531730625e-05, "loss": 3.4166, "step": 73250 }, { "epoch": 0.6473092071566082, "grad_norm": 1.3251335620880127, "learning_rate": 9.601956370804567e-05, "loss": 3.0175, "step": 73300 }, { "epoch": 0.6477507550468924, "grad_norm": 3.875401735305786, "learning_rate": 9.601412854416403e-05, "loss": 3.5598, "step": 73350 }, { "epoch": 0.6481923029371766, "grad_norm": 2.4360857009887695, "learning_rate": 9.600868982608116e-05, "loss": 3.3592, "step": 73400 }, { "epoch": 0.6486338508274607, "grad_norm": 2.4810409545898438, "learning_rate": 9.600324755421715e-05, "loss": 3.4158, "step": 73450 }, { "epoch": 0.6490753987177449, "grad_norm": 1.3619318008422852, "learning_rate": 9.599780172899234e-05, "loss": 2.9944, "step": 73500 }, { "epoch": 0.649516946608029, "grad_norm": 2.7849841117858887, "learning_rate": 9.599235235082741e-05, "loss": 3.6618, "step": 73550 }, { "epoch": 0.6499584944983133, "grad_norm": 5.122074127197266, "learning_rate": 9.598689942014326e-05, "loss": 3.0184, "step": 73600 }, { "epoch": 0.6504000423885975, "grad_norm": 2.697990655899048, "learning_rate": 9.59814429373611e-05, "loss": 3.0678, "step": 73650 }, { "epoch": 0.6508415902788817, "grad_norm": 1.3273183107376099, "learning_rate": 9.597598290290237e-05, "loss": 2.9799, "step": 73700 }, { "epoch": 0.6512831381691658, "grad_norm": 1.5726739168167114, "learning_rate": 9.597051931718882e-05, "loss": 3.1769, "step": 73750 }, { "epoch": 0.65172468605945, "grad_norm": 2.7319064140319824, "learning_rate": 9.596505218064249e-05, "loss": 3.5969, "step": 73800 }, { "epoch": 0.6521662339497342, "grad_norm": 3.6344516277313232, "learning_rate": 9.595958149368563e-05, "loss": 3.2977, "step": 73850 }, { "epoch": 0.6526077818400183, "grad_norm": 2.6774423122406006, "learning_rate": 9.595410725674084e-05, "loss": 3.1628, "step": 73900 }, { "epoch": 0.6530493297303025, "grad_norm": 4.738192081451416, "learning_rate": 9.594862947023094e-05, "loss": 3.6289, "step": 73950 }, { "epoch": 0.6534908776205868, "grad_norm": 1.0268847942352295, "learning_rate": 9.594314813457905e-05, "loss": 3.6385, "step": 74000 }, { "epoch": 0.6539324255108709, "grad_norm": 2.7158186435699463, "learning_rate": 9.593766325020855e-05, "loss": 3.4732, "step": 74050 }, { "epoch": 0.6543739734011551, "grad_norm": 2.4069275856018066, "learning_rate": 9.593217481754311e-05, "loss": 3.493, "step": 74100 }, { "epoch": 0.6548155212914393, "grad_norm": 1.4273890256881714, "learning_rate": 9.592668283700665e-05, "loss": 3.4282, "step": 74150 }, { "epoch": 0.6552570691817234, "grad_norm": 1.4611910581588745, "learning_rate": 9.59211873090234e-05, "loss": 3.5503, "step": 74200 }, { "epoch": 0.6556986170720076, "grad_norm": 1.2271146774291992, "learning_rate": 9.591568823401782e-05, "loss": 3.081, "step": 74250 }, { "epoch": 0.6561401649622918, "grad_norm": 2.7792177200317383, "learning_rate": 9.591018561241467e-05, "loss": 3.4418, "step": 74300 }, { "epoch": 0.6565817128525759, "grad_norm": 5.776242733001709, "learning_rate": 9.5904679444639e-05, "loss": 3.5402, "step": 74350 }, { "epoch": 0.6570232607428602, "grad_norm": 2.6737334728240967, "learning_rate": 9.58991697311161e-05, "loss": 3.2469, "step": 74400 }, { "epoch": 0.6574648086331444, "grad_norm": 1.3555004596710205, "learning_rate": 9.589365647227155e-05, "loss": 3.3199, "step": 74450 }, { "epoch": 0.6579063565234285, "grad_norm": 3.3269264698028564, "learning_rate": 9.58881396685312e-05, "loss": 3.462, "step": 74500 }, { "epoch": 0.6583479044137127, "grad_norm": 2.6994597911834717, "learning_rate": 9.58826193203212e-05, "loss": 3.3227, "step": 74550 }, { "epoch": 0.6587894523039969, "grad_norm": 3.3708934783935547, "learning_rate": 9.587709542806792e-05, "loss": 3.1164, "step": 74600 }, { "epoch": 0.659231000194281, "grad_norm": 1.3764150142669678, "learning_rate": 9.587156799219804e-05, "loss": 3.3193, "step": 74650 }, { "epoch": 0.6596725480845652, "grad_norm": 4.554634094238281, "learning_rate": 9.586603701313852e-05, "loss": 3.5422, "step": 74700 }, { "epoch": 0.6601140959748494, "grad_norm": 4.359382152557373, "learning_rate": 9.586050249131658e-05, "loss": 3.2248, "step": 74750 }, { "epoch": 0.6605556438651337, "grad_norm": 4.0737714767456055, "learning_rate": 9.585496442715968e-05, "loss": 3.3829, "step": 74800 }, { "epoch": 0.6609971917554178, "grad_norm": 2.545942544937134, "learning_rate": 9.584942282109564e-05, "loss": 3.2756, "step": 74850 }, { "epoch": 0.661438739645702, "grad_norm": 3.28908109664917, "learning_rate": 9.584387767355247e-05, "loss": 3.4446, "step": 74900 }, { "epoch": 0.6618802875359862, "grad_norm": 5.4131340980529785, "learning_rate": 9.58383289849585e-05, "loss": 2.9662, "step": 74950 }, { "epoch": 0.6623218354262703, "grad_norm": 1.5948145389556885, "learning_rate": 9.58327767557423e-05, "loss": 3.3822, "step": 75000 }, { "epoch": 0.6623218354262703, "eval_asr_loss": 0.8949383808222489, "eval_loss": 3.007831573486328, "eval_runtime": 20.8134, "eval_samples_per_second": 36.899, "eval_steps_per_second": 9.225, "eval_tts_loss": 5.953438474404635, "step": 75000 }, { "epoch": 0.6627633833165545, "grad_norm": 3.2304179668426514, "learning_rate": 9.582722098633276e-05, "loss": 3.3734, "step": 75050 }, { "epoch": 0.6632049312068387, "grad_norm": 3.123548746109009, "learning_rate": 9.5821661677159e-05, "loss": 3.4995, "step": 75100 }, { "epoch": 0.6636464790971228, "grad_norm": 2.3339643478393555, "learning_rate": 9.581609882865044e-05, "loss": 3.3416, "step": 75150 }, { "epoch": 0.6640880269874071, "grad_norm": 1.0405722856521606, "learning_rate": 9.58106438036635e-05, "loss": 3.3155, "step": 75200 }, { "epoch": 0.6645295748776913, "grad_norm": 3.653280258178711, "learning_rate": 9.580507394853993e-05, "loss": 2.9206, "step": 75250 }, { "epoch": 0.6649711227679754, "grad_norm": 2.6631522178649902, "learning_rate": 9.579950055536279e-05, "loss": 3.1143, "step": 75300 }, { "epoch": 0.6654126706582596, "grad_norm": 2.752356767654419, "learning_rate": 9.579392362456264e-05, "loss": 3.4541, "step": 75350 }, { "epoch": 0.6658542185485438, "grad_norm": 1.8876862525939941, "learning_rate": 9.578834315657019e-05, "loss": 2.942, "step": 75400 }, { "epoch": 0.6662957664388279, "grad_norm": 3.700727939605713, "learning_rate": 9.57827591518165e-05, "loss": 3.4877, "step": 75450 }, { "epoch": 0.6667373143291121, "grad_norm": 1.4538304805755615, "learning_rate": 9.57771716107329e-05, "loss": 3.5755, "step": 75500 }, { "epoch": 0.6671788622193963, "grad_norm": 4.2887349128723145, "learning_rate": 9.577158053375096e-05, "loss": 3.3367, "step": 75550 }, { "epoch": 0.6676204101096805, "grad_norm": 4.3281025886535645, "learning_rate": 9.576598592130257e-05, "loss": 3.5268, "step": 75600 }, { "epoch": 0.6680619579999647, "grad_norm": 2.3635294437408447, "learning_rate": 9.576038777381984e-05, "loss": 3.6769, "step": 75650 }, { "epoch": 0.6685035058902489, "grad_norm": 4.012321949005127, "learning_rate": 9.57547860917352e-05, "loss": 3.1022, "step": 75700 }, { "epoch": 0.668945053780533, "grad_norm": 1.8018879890441895, "learning_rate": 9.574918087548132e-05, "loss": 3.4172, "step": 75750 }, { "epoch": 0.6693866016708172, "grad_norm": 2.358532190322876, "learning_rate": 9.574357212549116e-05, "loss": 3.1392, "step": 75800 }, { "epoch": 0.6698281495611014, "grad_norm": 2.4655981063842773, "learning_rate": 9.573795984219793e-05, "loss": 3.3193, "step": 75850 }, { "epoch": 0.6702696974513855, "grad_norm": 2.9911704063415527, "learning_rate": 9.573234402603517e-05, "loss": 3.563, "step": 75900 }, { "epoch": 0.6707112453416697, "grad_norm": 5.0430707931518555, "learning_rate": 9.572672467743664e-05, "loss": 3.2619, "step": 75950 }, { "epoch": 0.671152793231954, "grad_norm": 2.1028594970703125, "learning_rate": 9.572110179683637e-05, "loss": 3.6662, "step": 76000 }, { "epoch": 0.6715943411222381, "grad_norm": 2.4186270236968994, "learning_rate": 9.571547538466868e-05, "loss": 3.0861, "step": 76050 }, { "epoch": 0.6720358890125223, "grad_norm": 1.6107758283615112, "learning_rate": 9.570984544136819e-05, "loss": 3.9137, "step": 76100 }, { "epoch": 0.6724774369028065, "grad_norm": 3.140552043914795, "learning_rate": 9.570432467144772e-05, "loss": 3.643, "step": 76150 }, { "epoch": 0.6729189847930906, "grad_norm": 3.5983471870422363, "learning_rate": 9.569868773778745e-05, "loss": 3.6072, "step": 76200 }, { "epoch": 0.6733605326833748, "grad_norm": 3.1312403678894043, "learning_rate": 9.569304727429107e-05, "loss": 3.3982, "step": 76250 }, { "epoch": 0.673802080573659, "grad_norm": 1.2464442253112793, "learning_rate": 9.568740328139423e-05, "loss": 3.36, "step": 76300 }, { "epoch": 0.6742436284639431, "grad_norm": 1.5809202194213867, "learning_rate": 9.568175575953293e-05, "loss": 3.404, "step": 76350 }, { "epoch": 0.6746851763542274, "grad_norm": 2.486264944076538, "learning_rate": 9.567610470914334e-05, "loss": 3.4402, "step": 76400 }, { "epoch": 0.6751267242445116, "grad_norm": 1.804176926612854, "learning_rate": 9.5670450130662e-05, "loss": 3.2879, "step": 76450 }, { "epoch": 0.6755682721347958, "grad_norm": 1.7221146821975708, "learning_rate": 9.566479202452567e-05, "loss": 3.2473, "step": 76500 }, { "epoch": 0.6760098200250799, "grad_norm": 2.4581539630889893, "learning_rate": 9.565913039117138e-05, "loss": 3.7328, "step": 76550 }, { "epoch": 0.6764513679153641, "grad_norm": 1.7588177919387817, "learning_rate": 9.565346523103643e-05, "loss": 3.1056, "step": 76600 }, { "epoch": 0.6768929158056483, "grad_norm": 2.841600179672241, "learning_rate": 9.564779654455844e-05, "loss": 3.3631, "step": 76650 }, { "epoch": 0.6773344636959324, "grad_norm": 2.2738077640533447, "learning_rate": 9.564212433217523e-05, "loss": 3.4339, "step": 76700 }, { "epoch": 0.6777760115862166, "grad_norm": 1.1713546514511108, "learning_rate": 9.563644859432497e-05, "loss": 3.2702, "step": 76750 }, { "epoch": 0.6782175594765009, "grad_norm": 2.2375364303588867, "learning_rate": 9.563076933144602e-05, "loss": 3.6829, "step": 76800 }, { "epoch": 0.678659107366785, "grad_norm": 1.015380620956421, "learning_rate": 9.562508654397708e-05, "loss": 3.0635, "step": 76850 }, { "epoch": 0.6791006552570692, "grad_norm": 0.6230117082595825, "learning_rate": 9.56194002323571e-05, "loss": 3.5006, "step": 76900 }, { "epoch": 0.6795422031473534, "grad_norm": 1.3702956438064575, "learning_rate": 9.56137103970253e-05, "loss": 3.107, "step": 76950 }, { "epoch": 0.6799837510376375, "grad_norm": 1.878818392753601, "learning_rate": 9.560801703842117e-05, "loss": 3.6604, "step": 77000 }, { "epoch": 0.6804252989279217, "grad_norm": 1.3559465408325195, "learning_rate": 9.560232015698444e-05, "loss": 3.634, "step": 77050 }, { "epoch": 0.6808668468182059, "grad_norm": 15.994784355163574, "learning_rate": 9.559661975315519e-05, "loss": 3.4924, "step": 77100 }, { "epoch": 0.68130839470849, "grad_norm": 0.9013010859489441, "learning_rate": 9.559091582737372e-05, "loss": 3.474, "step": 77150 }, { "epoch": 0.6817499425987743, "grad_norm": 1.3329524993896484, "learning_rate": 9.558520838008059e-05, "loss": 3.4002, "step": 77200 }, { "epoch": 0.6821914904890585, "grad_norm": 2.5802180767059326, "learning_rate": 9.557949741171667e-05, "loss": 3.1695, "step": 77250 }, { "epoch": 0.6826330383793426, "grad_norm": 2.464118242263794, "learning_rate": 9.557378292272307e-05, "loss": 3.337, "step": 77300 }, { "epoch": 0.6830745862696268, "grad_norm": 1.7328948974609375, "learning_rate": 9.556806491354121e-05, "loss": 3.085, "step": 77350 }, { "epoch": 0.683516134159911, "grad_norm": 1.6609792709350586, "learning_rate": 9.556234338461274e-05, "loss": 3.7844, "step": 77400 }, { "epoch": 0.6839576820501951, "grad_norm": 3.412128210067749, "learning_rate": 9.55566183363796e-05, "loss": 3.3013, "step": 77450 }, { "epoch": 0.6843992299404793, "grad_norm": 1.4039990901947021, "learning_rate": 9.555088976928399e-05, "loss": 3.2333, "step": 77500 }, { "epoch": 0.6848407778307635, "grad_norm": 1.7826491594314575, "learning_rate": 9.55451576837684e-05, "loss": 3.4012, "step": 77550 }, { "epoch": 0.6852823257210477, "grad_norm": 3.7740890979766846, "learning_rate": 9.553942208027559e-05, "loss": 3.2525, "step": 77600 }, { "epoch": 0.6857238736113319, "grad_norm": 3.7978830337524414, "learning_rate": 9.553368295924861e-05, "loss": 3.3832, "step": 77650 }, { "epoch": 0.6861654215016161, "grad_norm": 1.697966456413269, "learning_rate": 9.552794032113073e-05, "loss": 3.2114, "step": 77700 }, { "epoch": 0.6866069693919002, "grad_norm": 1.4913134574890137, "learning_rate": 9.552219416636552e-05, "loss": 3.1685, "step": 77750 }, { "epoch": 0.6870485172821844, "grad_norm": 2.034682512283325, "learning_rate": 9.551644449539683e-05, "loss": 2.8857, "step": 77800 }, { "epoch": 0.6874900651724686, "grad_norm": 1.9270625114440918, "learning_rate": 9.551069130866877e-05, "loss": 3.5445, "step": 77850 }, { "epoch": 0.6879316130627527, "grad_norm": 5.204434871673584, "learning_rate": 9.550493460662572e-05, "loss": 3.3383, "step": 77900 }, { "epoch": 0.6883731609530369, "grad_norm": 3.010862350463867, "learning_rate": 9.549917438971235e-05, "loss": 3.3101, "step": 77950 }, { "epoch": 0.6888147088433212, "grad_norm": 0.8644163608551025, "learning_rate": 9.549341065837357e-05, "loss": 3.3201, "step": 78000 }, { "epoch": 0.6888147088433212, "eval_asr_loss": 0.8925247313395483, "eval_loss": 2.992865800857544, "eval_runtime": 20.774, "eval_samples_per_second": 36.969, "eval_steps_per_second": 9.242, "eval_tts_loss": 5.960387531532541, "step": 78000 }, { "epoch": 0.6892562567336054, "grad_norm": 2.4011969566345215, "learning_rate": 9.54876434130546e-05, "loss": 3.9292, "step": 78050 }, { "epoch": 0.6896978046238895, "grad_norm": 3.125196695327759, "learning_rate": 9.54818726542009e-05, "loss": 3.247, "step": 78100 }, { "epoch": 0.6901393525141737, "grad_norm": 1.368432879447937, "learning_rate": 9.547609838225821e-05, "loss": 2.951, "step": 78150 }, { "epoch": 0.6905809004044579, "grad_norm": 2.4032599925994873, "learning_rate": 9.547032059767253e-05, "loss": 3.5267, "step": 78200 }, { "epoch": 0.691022448294742, "grad_norm": 2.1227686405181885, "learning_rate": 9.546453930089019e-05, "loss": 3.2206, "step": 78250 }, { "epoch": 0.6914639961850262, "grad_norm": 1.2852637767791748, "learning_rate": 9.54587544923577e-05, "loss": 3.4086, "step": 78300 }, { "epoch": 0.6919055440753105, "grad_norm": 1.241897463798523, "learning_rate": 9.54529661725219e-05, "loss": 3.1159, "step": 78350 }, { "epoch": 0.6923470919655946, "grad_norm": 1.472129225730896, "learning_rate": 9.544717434182991e-05, "loss": 3.2442, "step": 78400 }, { "epoch": 0.6927886398558788, "grad_norm": 2.743114471435547, "learning_rate": 9.544137900072907e-05, "loss": 3.2978, "step": 78450 }, { "epoch": 0.693230187746163, "grad_norm": 2.720283031463623, "learning_rate": 9.543558014966703e-05, "loss": 3.5339, "step": 78500 }, { "epoch": 0.6936717356364471, "grad_norm": 2.2413861751556396, "learning_rate": 9.542977778909173e-05, "loss": 3.6654, "step": 78550 }, { "epoch": 0.6941132835267313, "grad_norm": 0.8887834548950195, "learning_rate": 9.542397191945129e-05, "loss": 2.9958, "step": 78600 }, { "epoch": 0.6945548314170155, "grad_norm": 1.2447916269302368, "learning_rate": 9.541816254119425e-05, "loss": 3.0432, "step": 78650 }, { "epoch": 0.6949963793072996, "grad_norm": 2.320293664932251, "learning_rate": 9.541234965476925e-05, "loss": 3.4944, "step": 78700 }, { "epoch": 0.6954379271975839, "grad_norm": 2.7362334728240967, "learning_rate": 9.540653326062534e-05, "loss": 3.4683, "step": 78750 }, { "epoch": 0.6958794750878681, "grad_norm": 1.2431176900863647, "learning_rate": 9.540071335921176e-05, "loss": 3.2625, "step": 78800 }, { "epoch": 0.6963210229781522, "grad_norm": 2.1052627563476562, "learning_rate": 9.539488995097807e-05, "loss": 3.3472, "step": 78850 }, { "epoch": 0.6967625708684364, "grad_norm": 2.4554333686828613, "learning_rate": 9.538906303637407e-05, "loss": 3.327, "step": 78900 }, { "epoch": 0.6972041187587206, "grad_norm": 2.1302428245544434, "learning_rate": 9.538323261584985e-05, "loss": 3.4227, "step": 78950 }, { "epoch": 0.6976456666490047, "grad_norm": 1.9906952381134033, "learning_rate": 9.537739868985574e-05, "loss": 3.8786, "step": 79000 }, { "epoch": 0.6980872145392889, "grad_norm": 4.157954216003418, "learning_rate": 9.537156125884236e-05, "loss": 3.0816, "step": 79050 }, { "epoch": 0.6985287624295731, "grad_norm": 4.053858757019043, "learning_rate": 9.536572032326061e-05, "loss": 3.2263, "step": 79100 }, { "epoch": 0.6989703103198573, "grad_norm": 1.9072597026824951, "learning_rate": 9.535987588356168e-05, "loss": 3.4913, "step": 79150 }, { "epoch": 0.6994118582101415, "grad_norm": 4.11290168762207, "learning_rate": 9.535402794019696e-05, "loss": 3.5914, "step": 79200 }, { "epoch": 0.6998534061004257, "grad_norm": 3.6236138343811035, "learning_rate": 9.534817649361816e-05, "loss": 3.0452, "step": 79250 }, { "epoch": 0.7002949539907098, "grad_norm": 4.064155578613281, "learning_rate": 9.534232154427728e-05, "loss": 3.5526, "step": 79300 }, { "epoch": 0.700736501880994, "grad_norm": 2.754270553588867, "learning_rate": 9.533646309262657e-05, "loss": 3.3366, "step": 79350 }, { "epoch": 0.7011780497712782, "grad_norm": 2.6481540203094482, "learning_rate": 9.533060113911852e-05, "loss": 3.6007, "step": 79400 }, { "epoch": 0.7016195976615623, "grad_norm": 3.322692632675171, "learning_rate": 9.532473568420591e-05, "loss": 3.1894, "step": 79450 }, { "epoch": 0.7020611455518465, "grad_norm": 2.9386980533599854, "learning_rate": 9.531886672834182e-05, "loss": 3.1106, "step": 79500 }, { "epoch": 0.7025026934421308, "grad_norm": 1.7416975498199463, "learning_rate": 9.531299427197957e-05, "loss": 3.5195, "step": 79550 }, { "epoch": 0.702944241332415, "grad_norm": 2.034696578979492, "learning_rate": 9.530711831557274e-05, "loss": 3.3598, "step": 79600 }, { "epoch": 0.7033857892226991, "grad_norm": 2.048043966293335, "learning_rate": 9.530123885957523e-05, "loss": 3.3274, "step": 79650 }, { "epoch": 0.7038273371129833, "grad_norm": 0.8644744157791138, "learning_rate": 9.529535590444115e-05, "loss": 3.3067, "step": 79700 }, { "epoch": 0.7042688850032675, "grad_norm": 3.338059186935425, "learning_rate": 9.528946945062493e-05, "loss": 3.2878, "step": 79750 }, { "epoch": 0.7047104328935516, "grad_norm": 1.424256682395935, "learning_rate": 9.528357949858123e-05, "loss": 3.5114, "step": 79800 }, { "epoch": 0.7051519807838358, "grad_norm": 3.7967846393585205, "learning_rate": 9.527768604876501e-05, "loss": 3.4413, "step": 79850 }, { "epoch": 0.70559352867412, "grad_norm": 3.256422996520996, "learning_rate": 9.52717891016315e-05, "loss": 3.0054, "step": 79900 }, { "epoch": 0.7060350765644042, "grad_norm": 3.2881686687469482, "learning_rate": 9.526588865763614e-05, "loss": 3.3926, "step": 79950 }, { "epoch": 0.7064766244546884, "grad_norm": 2.3848628997802734, "learning_rate": 9.525998471723476e-05, "loss": 3.3552, "step": 80000 }, { "epoch": 0.7069181723449726, "grad_norm": 3.0554099082946777, "learning_rate": 9.525407728088333e-05, "loss": 3.4239, "step": 80050 }, { "epoch": 0.7073597202352567, "grad_norm": 2.6418068408966064, "learning_rate": 9.524816634903819e-05, "loss": 3.2979, "step": 80100 }, { "epoch": 0.7078012681255409, "grad_norm": 1.501378059387207, "learning_rate": 9.52422519221559e-05, "loss": 3.7132, "step": 80150 }, { "epoch": 0.7082428160158251, "grad_norm": 3.162963390350342, "learning_rate": 9.523633400069327e-05, "loss": 3.7012, "step": 80200 }, { "epoch": 0.7086843639061092, "grad_norm": 0.6245143413543701, "learning_rate": 9.523041258510746e-05, "loss": 3.423, "step": 80250 }, { "epoch": 0.7091259117963934, "grad_norm": 0.8148535490036011, "learning_rate": 9.522448767585579e-05, "loss": 3.3979, "step": 80300 }, { "epoch": 0.7095674596866777, "grad_norm": 2.1384999752044678, "learning_rate": 9.521855927339595e-05, "loss": 2.7129, "step": 80350 }, { "epoch": 0.7100090075769618, "grad_norm": 3.697734832763672, "learning_rate": 9.521262737818586e-05, "loss": 3.0784, "step": 80400 }, { "epoch": 0.710450555467246, "grad_norm": 6.074985027313232, "learning_rate": 9.520669199068368e-05, "loss": 3.4716, "step": 80450 }, { "epoch": 0.7108921033575302, "grad_norm": 3.228442907333374, "learning_rate": 9.520075311134788e-05, "loss": 3.3426, "step": 80500 }, { "epoch": 0.7113336512478143, "grad_norm": 1.2277064323425293, "learning_rate": 9.51948107406372e-05, "loss": 3.2555, "step": 80550 }, { "epoch": 0.7117751991380985, "grad_norm": 0.9551520943641663, "learning_rate": 9.518886487901065e-05, "loss": 3.5223, "step": 80600 }, { "epoch": 0.7122167470283827, "grad_norm": 2.2914512157440186, "learning_rate": 9.518291552692746e-05, "loss": 2.9172, "step": 80650 }, { "epoch": 0.7126582949186668, "grad_norm": 3.3304460048675537, "learning_rate": 9.517696268484719e-05, "loss": 2.6855, "step": 80700 }, { "epoch": 0.7130998428089511, "grad_norm": 1.3867692947387695, "learning_rate": 9.517100635322966e-05, "loss": 2.9664, "step": 80750 }, { "epoch": 0.7135413906992353, "grad_norm": 1.0515697002410889, "learning_rate": 9.51650465325349e-05, "loss": 3.2619, "step": 80800 }, { "epoch": 0.7139829385895194, "grad_norm": 1.7852091789245605, "learning_rate": 9.515908322322329e-05, "loss": 3.0423, "step": 80850 }, { "epoch": 0.7144244864798036, "grad_norm": 1.8944436311721802, "learning_rate": 9.515311642575545e-05, "loss": 3.1392, "step": 80900 }, { "epoch": 0.7148660343700878, "grad_norm": 0.9419056177139282, "learning_rate": 9.514714614059225e-05, "loss": 3.1358, "step": 80950 }, { "epoch": 0.7153075822603719, "grad_norm": 3.2333171367645264, "learning_rate": 9.514117236819485e-05, "loss": 3.8758, "step": 81000 }, { "epoch": 0.7153075822603719, "eval_asr_loss": 0.9185007565241553, "eval_loss": 2.989028215408325, "eval_runtime": 20.4458, "eval_samples_per_second": 37.563, "eval_steps_per_second": 9.391, "eval_tts_loss": 5.958070210757005, "step": 81000 }, { "epoch": 0.7157491301506561, "grad_norm": 2.493988037109375, "learning_rate": 9.513519510902466e-05, "loss": 3.6492, "step": 81050 }, { "epoch": 0.7161906780409403, "grad_norm": 2.0979015827178955, "learning_rate": 9.512921436354339e-05, "loss": 3.7933, "step": 81100 }, { "epoch": 0.7166322259312246, "grad_norm": 2.046807050704956, "learning_rate": 9.512323013221299e-05, "loss": 3.291, "step": 81150 }, { "epoch": 0.7170737738215087, "grad_norm": 2.943680763244629, "learning_rate": 9.51172424154957e-05, "loss": 3.0745, "step": 81200 }, { "epoch": 0.7175153217117929, "grad_norm": 2.4897234439849854, "learning_rate": 9.5111251213854e-05, "loss": 3.1611, "step": 81250 }, { "epoch": 0.717956869602077, "grad_norm": 4.505876541137695, "learning_rate": 9.51052565277507e-05, "loss": 3.5499, "step": 81300 }, { "epoch": 0.7183984174923612, "grad_norm": 1.430411458015442, "learning_rate": 9.50992583576488e-05, "loss": 3.1281, "step": 81350 }, { "epoch": 0.7188399653826454, "grad_norm": 3.1512632369995117, "learning_rate": 9.509325670401164e-05, "loss": 3.4918, "step": 81400 }, { "epoch": 0.7192815132729296, "grad_norm": 1.2546844482421875, "learning_rate": 9.508725156730276e-05, "loss": 3.6289, "step": 81450 }, { "epoch": 0.7197230611632137, "grad_norm": 3.754129409790039, "learning_rate": 9.508124294798604e-05, "loss": 2.9913, "step": 81500 }, { "epoch": 0.720164609053498, "grad_norm": 3.811619997024536, "learning_rate": 9.507523084652556e-05, "loss": 3.233, "step": 81550 }, { "epoch": 0.7206061569437822, "grad_norm": 1.469282627105713, "learning_rate": 9.5069335609166e-05, "loss": 3.6196, "step": 81600 }, { "epoch": 0.7210477048340663, "grad_norm": 2.8807554244995117, "learning_rate": 9.50633166144312e-05, "loss": 2.9091, "step": 81650 }, { "epoch": 0.7214892527243505, "grad_norm": 2.3154757022857666, "learning_rate": 9.505729413893734e-05, "loss": 3.2319, "step": 81700 }, { "epoch": 0.7219308006146347, "grad_norm": 1.580367922782898, "learning_rate": 9.505126818314955e-05, "loss": 3.5418, "step": 81750 }, { "epoch": 0.7223723485049188, "grad_norm": 1.0496435165405273, "learning_rate": 9.504523874753335e-05, "loss": 3.6296, "step": 81800 }, { "epoch": 0.722813896395203, "grad_norm": 0.9904226660728455, "learning_rate": 9.50392058325544e-05, "loss": 3.2764, "step": 81850 }, { "epoch": 0.7232554442854872, "grad_norm": 1.5296332836151123, "learning_rate": 9.503316943867876e-05, "loss": 3.6219, "step": 81900 }, { "epoch": 0.7236969921757714, "grad_norm": 1.79270601272583, "learning_rate": 9.502712956637263e-05, "loss": 3.467, "step": 81950 }, { "epoch": 0.7241385400660556, "grad_norm": 3.750113010406494, "learning_rate": 9.502108621610255e-05, "loss": 2.9975, "step": 82000 }, { "epoch": 0.7245800879563398, "grad_norm": 1.8261284828186035, "learning_rate": 9.501503938833534e-05, "loss": 3.2391, "step": 82050 }, { "epoch": 0.7250216358466239, "grad_norm": 1.2408541440963745, "learning_rate": 9.500911012370587e-05, "loss": 3.4596, "step": 82100 }, { "epoch": 0.7254631837369081, "grad_norm": 1.0297032594680786, "learning_rate": 9.50030564118725e-05, "loss": 3.1892, "step": 82150 }, { "epoch": 0.7259047316271923, "grad_norm": 2.6377224922180176, "learning_rate": 9.499699922393466e-05, "loss": 3.624, "step": 82200 }, { "epoch": 0.7263462795174764, "grad_norm": 3.4818551540374756, "learning_rate": 9.499093856036019e-05, "loss": 3.1586, "step": 82250 }, { "epoch": 0.7267878274077606, "grad_norm": 2.5095982551574707, "learning_rate": 9.498487442161721e-05, "loss": 3.2636, "step": 82300 }, { "epoch": 0.7272293752980449, "grad_norm": 1.323671579360962, "learning_rate": 9.497880680817414e-05, "loss": 3.3924, "step": 82350 }, { "epoch": 0.727670923188329, "grad_norm": 3.4375061988830566, "learning_rate": 9.497273572049967e-05, "loss": 3.2829, "step": 82400 }, { "epoch": 0.7281124710786132, "grad_norm": 1.7869148254394531, "learning_rate": 9.496666115906272e-05, "loss": 3.5715, "step": 82450 }, { "epoch": 0.7285540189688974, "grad_norm": 2.402039051055908, "learning_rate": 9.49605831243325e-05, "loss": 3.3708, "step": 82500 }, { "epoch": 0.7289955668591815, "grad_norm": 2.0941081047058105, "learning_rate": 9.49545016167785e-05, "loss": 3.3561, "step": 82550 }, { "epoch": 0.7294371147494657, "grad_norm": 3.094398021697998, "learning_rate": 9.494841663687043e-05, "loss": 3.5185, "step": 82600 }, { "epoch": 0.7298786626397499, "grad_norm": 1.317842960357666, "learning_rate": 9.494232818507833e-05, "loss": 3.3928, "step": 82650 }, { "epoch": 0.730320210530034, "grad_norm": 2.3808138370513916, "learning_rate": 9.493623626187249e-05, "loss": 3.375, "step": 82700 }, { "epoch": 0.7307617584203183, "grad_norm": 1.0954011678695679, "learning_rate": 9.493014086772345e-05, "loss": 3.2433, "step": 82750 }, { "epoch": 0.7312033063106025, "grad_norm": 2.7866711616516113, "learning_rate": 9.492404200310201e-05, "loss": 2.8652, "step": 82800 }, { "epoch": 0.7316448542008867, "grad_norm": 0.6492549777030945, "learning_rate": 9.491793966847928e-05, "loss": 3.688, "step": 82850 }, { "epoch": 0.7320864020911708, "grad_norm": 2.792008876800537, "learning_rate": 9.491183386432659e-05, "loss": 3.7061, "step": 82900 }, { "epoch": 0.732527949981455, "grad_norm": 3.021859884262085, "learning_rate": 9.49057245911156e-05, "loss": 3.5694, "step": 82950 }, { "epoch": 0.7329694978717392, "grad_norm": 2.1141061782836914, "learning_rate": 9.489961184931815e-05, "loss": 3.1073, "step": 83000 }, { "epoch": 0.7334110457620233, "grad_norm": 1.2735497951507568, "learning_rate": 9.489349563940644e-05, "loss": 3.3421, "step": 83050 }, { "epoch": 0.7338525936523075, "grad_norm": 1.5497995615005493, "learning_rate": 9.488737596185286e-05, "loss": 3.6448, "step": 83100 }, { "epoch": 0.7342941415425918, "grad_norm": 2.1992032527923584, "learning_rate": 9.488125281713013e-05, "loss": 3.4424, "step": 83150 }, { "epoch": 0.7347356894328759, "grad_norm": 4.297890663146973, "learning_rate": 9.487512620571121e-05, "loss": 3.191, "step": 83200 }, { "epoch": 0.7351772373231601, "grad_norm": 2.3956549167633057, "learning_rate": 9.486899612806931e-05, "loss": 3.0592, "step": 83250 }, { "epoch": 0.7356187852134443, "grad_norm": 1.5929914712905884, "learning_rate": 9.486286258467793e-05, "loss": 3.214, "step": 83300 }, { "epoch": 0.7360603331037284, "grad_norm": 1.5399370193481445, "learning_rate": 9.485672557601087e-05, "loss": 3.3211, "step": 83350 }, { "epoch": 0.7365018809940126, "grad_norm": 3.4927520751953125, "learning_rate": 9.48505851025421e-05, "loss": 3.287, "step": 83400 }, { "epoch": 0.7369434288842968, "grad_norm": 1.6479640007019043, "learning_rate": 9.484444116474597e-05, "loss": 3.2647, "step": 83450 }, { "epoch": 0.7373849767745809, "grad_norm": 4.172652721405029, "learning_rate": 9.483829376309703e-05, "loss": 3.0049, "step": 83500 }, { "epoch": 0.7378265246648652, "grad_norm": 2.3201801776885986, "learning_rate": 9.48321428980701e-05, "loss": 2.9019, "step": 83550 }, { "epoch": 0.7382680725551494, "grad_norm": 3.2628872394561768, "learning_rate": 9.482598857014031e-05, "loss": 3.2633, "step": 83600 }, { "epoch": 0.7387096204454335, "grad_norm": 1.8258609771728516, "learning_rate": 9.481983077978302e-05, "loss": 3.2017, "step": 83650 }, { "epoch": 0.7391511683357177, "grad_norm": 2.0017619132995605, "learning_rate": 9.481366952747386e-05, "loss": 3.2454, "step": 83700 }, { "epoch": 0.7395927162260019, "grad_norm": 2.6693084239959717, "learning_rate": 9.480750481368872e-05, "loss": 3.6843, "step": 83750 }, { "epoch": 0.740034264116286, "grad_norm": 1.9272335767745972, "learning_rate": 9.480133663890382e-05, "loss": 3.0776, "step": 83800 }, { "epoch": 0.7404758120065702, "grad_norm": 0.973445475101471, "learning_rate": 9.479516500359555e-05, "loss": 3.5528, "step": 83850 }, { "epoch": 0.7409173598968544, "grad_norm": 1.443723201751709, "learning_rate": 9.478898990824064e-05, "loss": 3.3918, "step": 83900 }, { "epoch": 0.7413589077871386, "grad_norm": 3.7001953125, "learning_rate": 9.478281135331606e-05, "loss": 3.3402, "step": 83950 }, { "epoch": 0.7418004556774228, "grad_norm": 2.7895514965057373, "learning_rate": 9.477662933929905e-05, "loss": 3.2964, "step": 84000 }, { "epoch": 0.7418004556774228, "eval_asr_loss": 0.907389885219024, "eval_loss": 2.9814484119415283, "eval_runtime": 21.1732, "eval_samples_per_second": 36.272, "eval_steps_per_second": 9.068, "eval_tts_loss": 5.974965912665154, "step": 84000 }, { "epoch": 0.742242003567707, "grad_norm": 2.343669891357422, "learning_rate": 9.477044386666711e-05, "loss": 3.0479, "step": 84050 }, { "epoch": 0.7426835514579911, "grad_norm": 2.6981098651885986, "learning_rate": 9.476425493589804e-05, "loss": 3.538, "step": 84100 }, { "epoch": 0.7431250993482753, "grad_norm": 2.0038726329803467, "learning_rate": 9.475806254746984e-05, "loss": 3.3287, "step": 84150 }, { "epoch": 0.7435666472385595, "grad_norm": 2.3365066051483154, "learning_rate": 9.475186670186088e-05, "loss": 3.4029, "step": 84200 }, { "epoch": 0.7440081951288436, "grad_norm": 3.0482475757598877, "learning_rate": 9.474566739954966e-05, "loss": 3.1918, "step": 84250 }, { "epoch": 0.7444497430191278, "grad_norm": 2.1410951614379883, "learning_rate": 9.47394646410151e-05, "loss": 3.1646, "step": 84300 }, { "epoch": 0.7448912909094121, "grad_norm": 2.7029786109924316, "learning_rate": 9.473325842673626e-05, "loss": 3.3533, "step": 84350 }, { "epoch": 0.7453328387996963, "grad_norm": 1.4176026582717896, "learning_rate": 9.472704875719254e-05, "loss": 3.5932, "step": 84400 }, { "epoch": 0.7457743866899804, "grad_norm": 3.4000325202941895, "learning_rate": 9.472083563286358e-05, "loss": 2.8913, "step": 84450 }, { "epoch": 0.7462159345802646, "grad_norm": 3.8159403800964355, "learning_rate": 9.47146190542293e-05, "loss": 3.0607, "step": 84500 }, { "epoch": 0.7466574824705488, "grad_norm": 2.903031349182129, "learning_rate": 9.470839902176985e-05, "loss": 3.1426, "step": 84550 }, { "epoch": 0.7470990303608329, "grad_norm": 1.7477384805679321, "learning_rate": 9.470217553596568e-05, "loss": 3.1242, "step": 84600 }, { "epoch": 0.7475405782511171, "grad_norm": 1.4132603406906128, "learning_rate": 9.469594859729753e-05, "loss": 3.1462, "step": 84650 }, { "epoch": 0.7479821261414012, "grad_norm": 1.7881561517715454, "learning_rate": 9.468971820624636e-05, "loss": 3.1489, "step": 84700 }, { "epoch": 0.7484236740316855, "grad_norm": 0.7184849381446838, "learning_rate": 9.468348436329341e-05, "loss": 3.5103, "step": 84750 }, { "epoch": 0.7488652219219697, "grad_norm": 5.096653938293457, "learning_rate": 9.46772470689202e-05, "loss": 3.6296, "step": 84800 }, { "epoch": 0.7493067698122539, "grad_norm": 1.7660068273544312, "learning_rate": 9.467100632360851e-05, "loss": 3.2378, "step": 84850 }, { "epoch": 0.749748317702538, "grad_norm": 3.2901999950408936, "learning_rate": 9.466476212784038e-05, "loss": 2.8817, "step": 84900 }, { "epoch": 0.7501898655928222, "grad_norm": 0.6989944577217102, "learning_rate": 9.465851448209813e-05, "loss": 3.2027, "step": 84950 }, { "epoch": 0.7506314134831064, "grad_norm": 4.7878875732421875, "learning_rate": 9.465226338686431e-05, "loss": 3.0253, "step": 85000 }, { "epoch": 0.7510729613733905, "grad_norm": 2.0710065364837646, "learning_rate": 9.46460088426218e-05, "loss": 3.5431, "step": 85050 }, { "epoch": 0.7515145092636747, "grad_norm": 1.119075059890747, "learning_rate": 9.463975084985369e-05, "loss": 3.4444, "step": 85100 }, { "epoch": 0.751956057153959, "grad_norm": 2.28783917427063, "learning_rate": 9.463348940904335e-05, "loss": 3.6863, "step": 85150 }, { "epoch": 0.7523976050442431, "grad_norm": 2.2328102588653564, "learning_rate": 9.462722452067445e-05, "loss": 3.2708, "step": 85200 }, { "epoch": 0.7528391529345273, "grad_norm": 1.885800838470459, "learning_rate": 9.462095618523089e-05, "loss": 3.7459, "step": 85250 }, { "epoch": 0.7532807008248115, "grad_norm": 1.719815731048584, "learning_rate": 9.461468440319681e-05, "loss": 3.7436, "step": 85300 }, { "epoch": 0.7537222487150956, "grad_norm": 1.0964494943618774, "learning_rate": 9.460840917505668e-05, "loss": 2.7071, "step": 85350 }, { "epoch": 0.7541637966053798, "grad_norm": 2.1939446926116943, "learning_rate": 9.460213050129525e-05, "loss": 3.4522, "step": 85400 }, { "epoch": 0.754605344495664, "grad_norm": 3.344871759414673, "learning_rate": 9.459584838239743e-05, "loss": 3.3535, "step": 85450 }, { "epoch": 0.7550468923859481, "grad_norm": 3.5036964416503906, "learning_rate": 9.458956281884848e-05, "loss": 3.1035, "step": 85500 }, { "epoch": 0.7554884402762324, "grad_norm": 1.8744654655456543, "learning_rate": 9.458327381113392e-05, "loss": 3.4614, "step": 85550 }, { "epoch": 0.7559299881665166, "grad_norm": 3.193300485610962, "learning_rate": 9.457698135973951e-05, "loss": 3.2385, "step": 85600 }, { "epoch": 0.7563715360568007, "grad_norm": 2.4154229164123535, "learning_rate": 9.457068546515129e-05, "loss": 3.1358, "step": 85650 }, { "epoch": 0.7568130839470849, "grad_norm": 1.8072885274887085, "learning_rate": 9.456438612785556e-05, "loss": 3.0887, "step": 85700 }, { "epoch": 0.7572546318373691, "grad_norm": 4.455030918121338, "learning_rate": 9.455808334833892e-05, "loss": 2.7886, "step": 85750 }, { "epoch": 0.7576961797276532, "grad_norm": 3.4446539878845215, "learning_rate": 9.455177712708815e-05, "loss": 3.045, "step": 85800 }, { "epoch": 0.7581377276179374, "grad_norm": 1.5495901107788086, "learning_rate": 9.45454674645904e-05, "loss": 3.3734, "step": 85850 }, { "epoch": 0.7585792755082216, "grad_norm": 1.636919379234314, "learning_rate": 9.453915436133303e-05, "loss": 3.1687, "step": 85900 }, { "epoch": 0.7590208233985059, "grad_norm": 1.9204362630844116, "learning_rate": 9.453283781780365e-05, "loss": 3.5536, "step": 85950 }, { "epoch": 0.75946237128879, "grad_norm": 2.4546802043914795, "learning_rate": 9.45265178344902e-05, "loss": 3.0949, "step": 86000 }, { "epoch": 0.7599039191790742, "grad_norm": 5.441671848297119, "learning_rate": 9.45201944118808e-05, "loss": 3.2127, "step": 86050 }, { "epoch": 0.7603454670693583, "grad_norm": 3.458197593688965, "learning_rate": 9.45138675504639e-05, "loss": 3.4892, "step": 86100 }, { "epoch": 0.7607870149596425, "grad_norm": 1.5939244031906128, "learning_rate": 9.450753725072822e-05, "loss": 3.2989, "step": 86150 }, { "epoch": 0.7612285628499267, "grad_norm": 2.7477095127105713, "learning_rate": 9.450120351316268e-05, "loss": 3.1591, "step": 86200 }, { "epoch": 0.7616701107402108, "grad_norm": 2.2054810523986816, "learning_rate": 9.449486633825654e-05, "loss": 3.2117, "step": 86250 }, { "epoch": 0.762111658630495, "grad_norm": 2.03580379486084, "learning_rate": 9.448852572649926e-05, "loss": 2.9641, "step": 86300 }, { "epoch": 0.7625532065207793, "grad_norm": 1.0966379642486572, "learning_rate": 9.448218167838065e-05, "loss": 3.3211, "step": 86350 }, { "epoch": 0.7629947544110635, "grad_norm": 2.210423707962036, "learning_rate": 9.44758341943907e-05, "loss": 3.2672, "step": 86400 }, { "epoch": 0.7634363023013476, "grad_norm": 3.2034225463867188, "learning_rate": 9.446948327501967e-05, "loss": 3.6083, "step": 86450 }, { "epoch": 0.7638778501916318, "grad_norm": 2.616360664367676, "learning_rate": 9.446312892075818e-05, "loss": 3.1633, "step": 86500 }, { "epoch": 0.764319398081916, "grad_norm": 2.7256109714508057, "learning_rate": 9.445689832152417e-05, "loss": 3.1589, "step": 86550 }, { "epoch": 0.7647609459722001, "grad_norm": 3.060117483139038, "learning_rate": 9.445053716762777e-05, "loss": 3.142, "step": 86600 }, { "epoch": 0.7652024938624843, "grad_norm": 1.2337734699249268, "learning_rate": 9.444417258030429e-05, "loss": 3.1518, "step": 86650 }, { "epoch": 0.7656440417527685, "grad_norm": 2.3434503078460693, "learning_rate": 9.443780456004537e-05, "loss": 3.2844, "step": 86700 }, { "epoch": 0.7660855896430527, "grad_norm": 2.32765793800354, "learning_rate": 9.443143310734285e-05, "loss": 3.131, "step": 86750 }, { "epoch": 0.7665271375333369, "grad_norm": 1.1229861974716187, "learning_rate": 9.442505822268891e-05, "loss": 3.4084, "step": 86800 }, { "epoch": 0.7669686854236211, "grad_norm": 2.0934159755706787, "learning_rate": 9.441867990657594e-05, "loss": 3.3533, "step": 86850 }, { "epoch": 0.7674102333139052, "grad_norm": 5.029802322387695, "learning_rate": 9.441229815949658e-05, "loss": 3.2617, "step": 86900 }, { "epoch": 0.7678517812041894, "grad_norm": 3.596031665802002, "learning_rate": 9.440591298194382e-05, "loss": 3.3018, "step": 86950 }, { "epoch": 0.7682933290944736, "grad_norm": 0.9849772453308105, "learning_rate": 9.439952437441082e-05, "loss": 2.6819, "step": 87000 }, { "epoch": 0.7682933290944736, "eval_asr_loss": 0.9071832583308085, "eval_loss": 2.9738311767578125, "eval_runtime": 20.831, "eval_samples_per_second": 36.868, "eval_steps_per_second": 9.217, "eval_tts_loss": 5.941716546998216, "step": 87000 }, { "epoch": 0.7687348769847577, "grad_norm": 1.9874849319458008, "learning_rate": 9.439313233739105e-05, "loss": 3.5471, "step": 87050 }, { "epoch": 0.7691764248750419, "grad_norm": 2.417670249938965, "learning_rate": 9.438673687137827e-05, "loss": 3.2482, "step": 87100 }, { "epoch": 0.7696179727653262, "grad_norm": 1.2283480167388916, "learning_rate": 9.438033797686642e-05, "loss": 3.6392, "step": 87150 }, { "epoch": 0.7700595206556103, "grad_norm": 4.871400833129883, "learning_rate": 9.437393565434984e-05, "loss": 3.7572, "step": 87200 }, { "epoch": 0.7705010685458945, "grad_norm": 2.2227866649627686, "learning_rate": 9.436752990432298e-05, "loss": 3.0988, "step": 87250 }, { "epoch": 0.7709426164361787, "grad_norm": 1.317663550376892, "learning_rate": 9.436112072728067e-05, "loss": 2.342, "step": 87300 }, { "epoch": 0.7713841643264628, "grad_norm": 1.2101949453353882, "learning_rate": 9.435470812371796e-05, "loss": 3.0555, "step": 87350 }, { "epoch": 0.771825712216747, "grad_norm": 2.3560891151428223, "learning_rate": 9.434829209413016e-05, "loss": 3.287, "step": 87400 }, { "epoch": 0.7722672601070312, "grad_norm": 2.26143217086792, "learning_rate": 9.434187263901286e-05, "loss": 3.1895, "step": 87450 }, { "epoch": 0.7727088079973153, "grad_norm": 2.7307794094085693, "learning_rate": 9.43354497588619e-05, "loss": 3.0531, "step": 87500 }, { "epoch": 0.7731503558875996, "grad_norm": 2.08046293258667, "learning_rate": 9.43290234541734e-05, "loss": 3.5845, "step": 87550 }, { "epoch": 0.7735919037778838, "grad_norm": 1.427642822265625, "learning_rate": 9.432259372544375e-05, "loss": 3.372, "step": 87600 }, { "epoch": 0.774033451668168, "grad_norm": 1.8793132305145264, "learning_rate": 9.431616057316957e-05, "loss": 3.44, "step": 87650 }, { "epoch": 0.7744749995584521, "grad_norm": 5.893092632293701, "learning_rate": 9.430972399784779e-05, "loss": 3.3203, "step": 87700 }, { "epoch": 0.7749165474487363, "grad_norm": 1.8131071329116821, "learning_rate": 9.430328399997555e-05, "loss": 3.1405, "step": 87750 }, { "epoch": 0.7753580953390204, "grad_norm": 1.3549493551254272, "learning_rate": 9.429696948198171e-05, "loss": 3.4958, "step": 87800 }, { "epoch": 0.7757996432293046, "grad_norm": 2.8641469478607178, "learning_rate": 9.429052270892739e-05, "loss": 3.0084, "step": 87850 }, { "epoch": 0.7762411911195888, "grad_norm": 1.4089255332946777, "learning_rate": 9.428407251480576e-05, "loss": 3.5739, "step": 87900 }, { "epoch": 0.7766827390098731, "grad_norm": 2.792038679122925, "learning_rate": 9.427761890011504e-05, "loss": 3.0026, "step": 87950 }, { "epoch": 0.7771242869001572, "grad_norm": 2.8735337257385254, "learning_rate": 9.427116186535371e-05, "loss": 2.9788, "step": 88000 }, { "epoch": 0.7775658347904414, "grad_norm": 2.6922061443328857, "learning_rate": 9.426470141102058e-05, "loss": 3.5371, "step": 88050 }, { "epoch": 0.7780073826807256, "grad_norm": 2.8780505657196045, "learning_rate": 9.425823753761458e-05, "loss": 3.4691, "step": 88100 }, { "epoch": 0.7784489305710097, "grad_norm": 3.15449595451355, "learning_rate": 9.425177024563504e-05, "loss": 3.0253, "step": 88150 }, { "epoch": 0.7788904784612939, "grad_norm": 2.4547994136810303, "learning_rate": 9.42452995355815e-05, "loss": 3.22, "step": 88200 }, { "epoch": 0.779332026351578, "grad_norm": 1.3049603700637817, "learning_rate": 9.423882540795375e-05, "loss": 3.3362, "step": 88250 }, { "epoch": 0.7797735742418622, "grad_norm": 2.944639205932617, "learning_rate": 9.423234786325188e-05, "loss": 2.6225, "step": 88300 }, { "epoch": 0.7802151221321465, "grad_norm": 4.7420783042907715, "learning_rate": 9.42258669019762e-05, "loss": 2.852, "step": 88350 }, { "epoch": 0.7806566700224307, "grad_norm": 2.200601100921631, "learning_rate": 9.421938252462734e-05, "loss": 3.4574, "step": 88400 }, { "epoch": 0.7810982179127148, "grad_norm": 0.8284490704536438, "learning_rate": 9.421289473170615e-05, "loss": 3.3901, "step": 88450 }, { "epoch": 0.781539765802999, "grad_norm": 2.629199266433716, "learning_rate": 9.420640352371375e-05, "loss": 3.0143, "step": 88500 }, { "epoch": 0.7819813136932832, "grad_norm": 2.411001205444336, "learning_rate": 9.419990890115155e-05, "loss": 3.4572, "step": 88550 }, { "epoch": 0.7824228615835673, "grad_norm": 3.345630168914795, "learning_rate": 9.419341086452118e-05, "loss": 2.9808, "step": 88600 }, { "epoch": 0.7828644094738515, "grad_norm": 1.7286512851715088, "learning_rate": 9.418690941432458e-05, "loss": 3.3402, "step": 88650 }, { "epoch": 0.7833059573641357, "grad_norm": 2.867637872695923, "learning_rate": 9.418040455106391e-05, "loss": 3.3708, "step": 88700 }, { "epoch": 0.7837475052544199, "grad_norm": 1.4467108249664307, "learning_rate": 9.417389627524163e-05, "loss": 3.4814, "step": 88750 }, { "epoch": 0.7841890531447041, "grad_norm": 1.6191805601119995, "learning_rate": 9.416738458736045e-05, "loss": 3.4584, "step": 88800 }, { "epoch": 0.7846306010349883, "grad_norm": 2.4342501163482666, "learning_rate": 9.416086948792333e-05, "loss": 3.4172, "step": 88850 }, { "epoch": 0.7850721489252724, "grad_norm": 1.3565874099731445, "learning_rate": 9.415435097743352e-05, "loss": 3.6409, "step": 88900 }, { "epoch": 0.7855136968155566, "grad_norm": 0.9413079619407654, "learning_rate": 9.414782905639452e-05, "loss": 3.1165, "step": 88950 }, { "epoch": 0.7859552447058408, "grad_norm": 2.325162887573242, "learning_rate": 9.414130372531009e-05, "loss": 3.3463, "step": 89000 }, { "epoch": 0.7863967925961249, "grad_norm": 2.1379945278167725, "learning_rate": 9.413477498468423e-05, "loss": 3.416, "step": 89050 }, { "epoch": 0.7868383404864091, "grad_norm": 1.78729248046875, "learning_rate": 9.412824283502129e-05, "loss": 3.6121, "step": 89100 }, { "epoch": 0.7872798883766934, "grad_norm": 1.3895008563995361, "learning_rate": 9.412170727682576e-05, "loss": 3.1819, "step": 89150 }, { "epoch": 0.7877214362669775, "grad_norm": 3.693272352218628, "learning_rate": 9.41151683106025e-05, "loss": 3.3451, "step": 89200 }, { "epoch": 0.7881629841572617, "grad_norm": 1.3732582330703735, "learning_rate": 9.410862593685657e-05, "loss": 3.0712, "step": 89250 }, { "epoch": 0.7886045320475459, "grad_norm": 1.9743282794952393, "learning_rate": 9.410208015609333e-05, "loss": 3.7555, "step": 89300 }, { "epoch": 0.78904607993783, "grad_norm": 2.194549083709717, "learning_rate": 9.409553096881835e-05, "loss": 2.9958, "step": 89350 }, { "epoch": 0.7894876278281142, "grad_norm": 1.8362387418746948, "learning_rate": 9.408897837553754e-05, "loss": 3.4235, "step": 89400 }, { "epoch": 0.7899291757183984, "grad_norm": 1.8394923210144043, "learning_rate": 9.4082422376757e-05, "loss": 3.3358, "step": 89450 }, { "epoch": 0.7903707236086825, "grad_norm": 2.9508590698242188, "learning_rate": 9.407586297298314e-05, "loss": 3.5187, "step": 89500 }, { "epoch": 0.7908122714989668, "grad_norm": 1.0827395915985107, "learning_rate": 9.406930016472262e-05, "loss": 2.985, "step": 89550 }, { "epoch": 0.791253819389251, "grad_norm": 6.141767978668213, "learning_rate": 9.406273395248236e-05, "loss": 3.3644, "step": 89600 }, { "epoch": 0.7916953672795352, "grad_norm": 2.3642897605895996, "learning_rate": 9.405616433676954e-05, "loss": 2.9457, "step": 89650 }, { "epoch": 0.7921369151698193, "grad_norm": 1.307497501373291, "learning_rate": 9.404959131809163e-05, "loss": 3.5927, "step": 89700 }, { "epoch": 0.7925784630601035, "grad_norm": 0.8792401552200317, "learning_rate": 9.404301489695629e-05, "loss": 3.4736, "step": 89750 }, { "epoch": 0.7930200109503877, "grad_norm": 1.8458104133605957, "learning_rate": 9.403643507387155e-05, "loss": 3.1718, "step": 89800 }, { "epoch": 0.7934615588406718, "grad_norm": 0.40211722254753113, "learning_rate": 9.402985184934561e-05, "loss": 3.1028, "step": 89850 }, { "epoch": 0.793903106730956, "grad_norm": 1.4072984457015991, "learning_rate": 9.402326522388695e-05, "loss": 3.2689, "step": 89900 }, { "epoch": 0.7943446546212403, "grad_norm": 3.634361505508423, "learning_rate": 9.401667519800439e-05, "loss": 3.2542, "step": 89950 }, { "epoch": 0.7947862025115244, "grad_norm": 2.3660759925842285, "learning_rate": 9.401008177220691e-05, "loss": 3.3141, "step": 90000 }, { "epoch": 0.7947862025115244, "eval_asr_loss": 0.913769633646809, "eval_loss": 2.972637891769409, "eval_runtime": 20.6823, "eval_samples_per_second": 37.133, "eval_steps_per_second": 9.283, "eval_tts_loss": 5.941250072060806, "step": 90000 }, { "epoch": 0.7952277504018086, "grad_norm": 1.8747526407241821, "learning_rate": 9.400348494700382e-05, "loss": 3.2776, "step": 90050 }, { "epoch": 0.7956692982920928, "grad_norm": 2.4819157123565674, "learning_rate": 9.399688472290464e-05, "loss": 3.3672, "step": 90100 }, { "epoch": 0.7961108461823769, "grad_norm": 3.1009058952331543, "learning_rate": 9.399028110041921e-05, "loss": 3.7072, "step": 90150 }, { "epoch": 0.7965523940726611, "grad_norm": 2.1817233562469482, "learning_rate": 9.39836740800576e-05, "loss": 3.1293, "step": 90200 }, { "epoch": 0.7969939419629453, "grad_norm": 2.2436323165893555, "learning_rate": 9.397706366233011e-05, "loss": 3.7612, "step": 90250 }, { "epoch": 0.7974354898532294, "grad_norm": 2.8587071895599365, "learning_rate": 9.397058215732492e-05, "loss": 3.574, "step": 90300 }, { "epoch": 0.7978770377435137, "grad_norm": 1.305193543434143, "learning_rate": 9.396396501431968e-05, "loss": 3.2321, "step": 90350 }, { "epoch": 0.7983185856337979, "grad_norm": 4.059189319610596, "learning_rate": 9.395734447547093e-05, "loss": 3.2663, "step": 90400 }, { "epoch": 0.798760133524082, "grad_norm": 1.4845237731933594, "learning_rate": 9.395072054129008e-05, "loss": 3.2744, "step": 90450 }, { "epoch": 0.7992016814143662, "grad_norm": 3.5800607204437256, "learning_rate": 9.394409321228877e-05, "loss": 3.3557, "step": 90500 }, { "epoch": 0.7996432293046504, "grad_norm": 1.589889645576477, "learning_rate": 9.393746248897888e-05, "loss": 3.1891, "step": 90550 }, { "epoch": 0.8000847771949345, "grad_norm": 2.2087574005126953, "learning_rate": 9.39308283718726e-05, "loss": 3.4591, "step": 90600 }, { "epoch": 0.8005263250852187, "grad_norm": 2.9878342151641846, "learning_rate": 9.392419086148234e-05, "loss": 3.1537, "step": 90650 }, { "epoch": 0.8009678729755029, "grad_norm": 4.756255626678467, "learning_rate": 9.391754995832083e-05, "loss": 3.1161, "step": 90700 }, { "epoch": 0.8014094208657871, "grad_norm": 3.2718210220336914, "learning_rate": 9.391090566290097e-05, "loss": 3.6481, "step": 90750 }, { "epoch": 0.8018509687560713, "grad_norm": 1.9818165302276611, "learning_rate": 9.3904257975736e-05, "loss": 3.0561, "step": 90800 }, { "epoch": 0.8022925166463555, "grad_norm": 0.7319821715354919, "learning_rate": 9.38976068973394e-05, "loss": 3.6211, "step": 90850 }, { "epoch": 0.8027340645366396, "grad_norm": 1.0141242742538452, "learning_rate": 9.389095242822492e-05, "loss": 3.3727, "step": 90900 }, { "epoch": 0.8031756124269238, "grad_norm": 2.862752676010132, "learning_rate": 9.388429456890654e-05, "loss": 3.2647, "step": 90950 }, { "epoch": 0.803617160317208, "grad_norm": 2.0796523094177246, "learning_rate": 9.387763331989854e-05, "loss": 2.9279, "step": 91000 }, { "epoch": 0.8040587082074921, "grad_norm": 3.659575939178467, "learning_rate": 9.387096868171543e-05, "loss": 3.3732, "step": 91050 }, { "epoch": 0.8045002560977763, "grad_norm": 1.8139934539794922, "learning_rate": 9.3864300654872e-05, "loss": 3.4054, "step": 91100 }, { "epoch": 0.8049418039880606, "grad_norm": 1.9845603704452515, "learning_rate": 9.385762923988332e-05, "loss": 3.2696, "step": 91150 }, { "epoch": 0.8053833518783448, "grad_norm": 1.5647777318954468, "learning_rate": 9.385095443726466e-05, "loss": 2.9517, "step": 91200 }, { "epoch": 0.8058248997686289, "grad_norm": 2.7620863914489746, "learning_rate": 9.384427624753164e-05, "loss": 3.3531, "step": 91250 }, { "epoch": 0.8062664476589131, "grad_norm": 4.5521039962768555, "learning_rate": 9.383759467120006e-05, "loss": 2.9471, "step": 91300 }, { "epoch": 0.8067079955491973, "grad_norm": 6.784761428833008, "learning_rate": 9.3830909708786e-05, "loss": 3.0502, "step": 91350 }, { "epoch": 0.8071495434394814, "grad_norm": 2.283494472503662, "learning_rate": 9.382422136080587e-05, "loss": 3.2579, "step": 91400 }, { "epoch": 0.8075910913297656, "grad_norm": 2.49485182762146, "learning_rate": 9.381752962777626e-05, "loss": 3.2323, "step": 91450 }, { "epoch": 0.8080326392200498, "grad_norm": 1.1890743970870972, "learning_rate": 9.381083451021402e-05, "loss": 3.3812, "step": 91500 }, { "epoch": 0.808474187110334, "grad_norm": 1.023622989654541, "learning_rate": 9.380413600863633e-05, "loss": 3.1125, "step": 91550 }, { "epoch": 0.8089157350006182, "grad_norm": 3.7999751567840576, "learning_rate": 9.37974341235606e-05, "loss": 3.0946, "step": 91600 }, { "epoch": 0.8093572828909024, "grad_norm": 2.7334210872650146, "learning_rate": 9.379072885550446e-05, "loss": 3.2808, "step": 91650 }, { "epoch": 0.8097988307811865, "grad_norm": 1.0871787071228027, "learning_rate": 9.378402020498585e-05, "loss": 2.9201, "step": 91700 }, { "epoch": 0.8102403786714707, "grad_norm": 1.837319254875183, "learning_rate": 9.377730817252296e-05, "loss": 3.3392, "step": 91750 }, { "epoch": 0.8106819265617549, "grad_norm": 1.857426643371582, "learning_rate": 9.377059275863423e-05, "loss": 3.4567, "step": 91800 }, { "epoch": 0.811123474452039, "grad_norm": 2.4178388118743896, "learning_rate": 9.37638739638384e-05, "loss": 3.1172, "step": 91850 }, { "epoch": 0.8115650223423232, "grad_norm": 1.7428170442581177, "learning_rate": 9.375715178865438e-05, "loss": 3.2541, "step": 91900 }, { "epoch": 0.8120065702326075, "grad_norm": 2.188565731048584, "learning_rate": 9.375042623360143e-05, "loss": 3.0557, "step": 91950 }, { "epoch": 0.8124481181228916, "grad_norm": 2.7001125812530518, "learning_rate": 9.374369729919907e-05, "loss": 3.4467, "step": 92000 }, { "epoch": 0.8128896660131758, "grad_norm": 0.5670623183250427, "learning_rate": 9.373696498596703e-05, "loss": 3.4051, "step": 92050 }, { "epoch": 0.81333121390346, "grad_norm": 0.8684957027435303, "learning_rate": 9.37302292944253e-05, "loss": 3.4903, "step": 92100 }, { "epoch": 0.8137727617937441, "grad_norm": 1.7560847997665405, "learning_rate": 9.372349022509419e-05, "loss": 2.8986, "step": 92150 }, { "epoch": 0.8142143096840283, "grad_norm": 1.301934838294983, "learning_rate": 9.371674777849423e-05, "loss": 3.2969, "step": 92200 }, { "epoch": 0.8146558575743125, "grad_norm": 5.415755748748779, "learning_rate": 9.37100019551462e-05, "loss": 3.5175, "step": 92250 }, { "epoch": 0.8150974054645966, "grad_norm": 1.403662085533142, "learning_rate": 9.370325275557119e-05, "loss": 3.3363, "step": 92300 }, { "epoch": 0.8155389533548809, "grad_norm": 0.34755173325538635, "learning_rate": 9.369650018029047e-05, "loss": 2.8427, "step": 92350 }, { "epoch": 0.8159805012451651, "grad_norm": 4.63238000869751, "learning_rate": 9.368974422982567e-05, "loss": 3.401, "step": 92400 }, { "epoch": 0.8164220491354492, "grad_norm": 5.1991286277771, "learning_rate": 9.36829849046986e-05, "loss": 3.0604, "step": 92450 }, { "epoch": 0.8168635970257334, "grad_norm": 4.037189960479736, "learning_rate": 9.367622220543136e-05, "loss": 3.4574, "step": 92500 }, { "epoch": 0.8173051449160176, "grad_norm": 3.1938400268554688, "learning_rate": 9.366945613254631e-05, "loss": 3.4916, "step": 92550 }, { "epoch": 0.8177466928063017, "grad_norm": 1.4730736017227173, "learning_rate": 9.36626866865661e-05, "loss": 3.5001, "step": 92600 }, { "epoch": 0.8181882406965859, "grad_norm": 2.499075174331665, "learning_rate": 9.36559138680136e-05, "loss": 3.3653, "step": 92650 }, { "epoch": 0.8186297885868701, "grad_norm": 3.067751884460449, "learning_rate": 9.364913767741192e-05, "loss": 2.9999, "step": 92700 }, { "epoch": 0.8190713364771544, "grad_norm": 3.1233701705932617, "learning_rate": 9.36423581152845e-05, "loss": 3.377, "step": 92750 }, { "epoch": 0.8195128843674385, "grad_norm": 1.1378720998764038, "learning_rate": 9.363557518215498e-05, "loss": 2.8108, "step": 92800 }, { "epoch": 0.8199544322577227, "grad_norm": 1.5843881368637085, "learning_rate": 9.362878887854732e-05, "loss": 3.4633, "step": 92850 }, { "epoch": 0.8203959801480069, "grad_norm": 3.084805965423584, "learning_rate": 9.362199920498566e-05, "loss": 3.3343, "step": 92900 }, { "epoch": 0.820837528038291, "grad_norm": 1.3744654655456543, "learning_rate": 9.361520616199446e-05, "loss": 2.9136, "step": 92950 }, { "epoch": 0.8212790759285752, "grad_norm": 3.1736974716186523, "learning_rate": 9.360840975009845e-05, "loss": 3.1212, "step": 93000 }, { "epoch": 0.8212790759285752, "eval_asr_loss": 0.9125425967281584, "eval_loss": 2.9801833629608154, "eval_runtime": 20.5623, "eval_samples_per_second": 37.35, "eval_steps_per_second": 9.337, "eval_tts_loss": 5.957149096143686, "step": 93000 }, { "epoch": 0.8217206238188594, "grad_norm": 2.6484904289245605, "learning_rate": 9.360160996982256e-05, "loss": 3.3256, "step": 93050 }, { "epoch": 0.8221621717091435, "grad_norm": 1.6820108890533447, "learning_rate": 9.359480682169204e-05, "loss": 3.3486, "step": 93100 }, { "epoch": 0.8226037195994278, "grad_norm": 1.407516360282898, "learning_rate": 9.358800030623236e-05, "loss": 3.4169, "step": 93150 }, { "epoch": 0.823045267489712, "grad_norm": 1.139156699180603, "learning_rate": 9.358119042396929e-05, "loss": 2.8844, "step": 93200 }, { "epoch": 0.8234868153799961, "grad_norm": 1.5392656326293945, "learning_rate": 9.35743771754288e-05, "loss": 3.5062, "step": 93250 }, { "epoch": 0.8239283632702803, "grad_norm": 1.2801883220672607, "learning_rate": 9.356756056113717e-05, "loss": 2.9979, "step": 93300 }, { "epoch": 0.8243699111605645, "grad_norm": 2.373271942138672, "learning_rate": 9.356074058162094e-05, "loss": 2.8897, "step": 93350 }, { "epoch": 0.8248114590508486, "grad_norm": 0.7981138229370117, "learning_rate": 9.355391723740687e-05, "loss": 3.7609, "step": 93400 }, { "epoch": 0.8252530069411328, "grad_norm": 1.8183437585830688, "learning_rate": 9.354709052902204e-05, "loss": 3.2145, "step": 93450 }, { "epoch": 0.825694554831417, "grad_norm": 1.9743393659591675, "learning_rate": 9.354026045699372e-05, "loss": 3.5426, "step": 93500 }, { "epoch": 0.8261361027217012, "grad_norm": 2.351931571960449, "learning_rate": 9.35334270218495e-05, "loss": 3.347, "step": 93550 }, { "epoch": 0.8265776506119854, "grad_norm": 1.7274210453033447, "learning_rate": 9.352659022411718e-05, "loss": 2.9723, "step": 93600 }, { "epoch": 0.8270191985022696, "grad_norm": 3.6531035900115967, "learning_rate": 9.351975006432487e-05, "loss": 3.311, "step": 93650 }, { "epoch": 0.8274607463925537, "grad_norm": 5.743092060089111, "learning_rate": 9.351290654300091e-05, "loss": 3.1698, "step": 93700 }, { "epoch": 0.8279022942828379, "grad_norm": 1.3943450450897217, "learning_rate": 9.35060596606739e-05, "loss": 3.0665, "step": 93750 }, { "epoch": 0.8283438421731221, "grad_norm": 3.4705088138580322, "learning_rate": 9.349920941787269e-05, "loss": 3.5611, "step": 93800 }, { "epoch": 0.8287853900634062, "grad_norm": 1.309612512588501, "learning_rate": 9.34923558151264e-05, "loss": 3.4221, "step": 93850 }, { "epoch": 0.8292269379536905, "grad_norm": 1.8064721822738647, "learning_rate": 9.348549885296446e-05, "loss": 3.564, "step": 93900 }, { "epoch": 0.8296684858439747, "grad_norm": 1.7501392364501953, "learning_rate": 9.347863853191646e-05, "loss": 3.0225, "step": 93950 }, { "epoch": 0.8301100337342588, "grad_norm": 3.4473793506622314, "learning_rate": 9.347177485251233e-05, "loss": 3.0757, "step": 94000 }, { "epoch": 0.830551581624543, "grad_norm": 1.1141574382781982, "learning_rate": 9.346490781528221e-05, "loss": 3.0982, "step": 94050 }, { "epoch": 0.8309931295148272, "grad_norm": 3.239313840866089, "learning_rate": 9.345803742075656e-05, "loss": 3.463, "step": 94100 }, { "epoch": 0.8314346774051113, "grad_norm": 2.90984845161438, "learning_rate": 9.345116366946601e-05, "loss": 3.131, "step": 94150 }, { "epoch": 0.8318762252953955, "grad_norm": 3.623865842819214, "learning_rate": 9.344428656194153e-05, "loss": 3.4718, "step": 94200 }, { "epoch": 0.8323177731856797, "grad_norm": 0.8482750058174133, "learning_rate": 9.343740609871431e-05, "loss": 3.3195, "step": 94250 }, { "epoch": 0.832759321075964, "grad_norm": 1.9041322469711304, "learning_rate": 9.343052228031582e-05, "loss": 3.3709, "step": 94300 }, { "epoch": 0.8332008689662481, "grad_norm": 3.6412575244903564, "learning_rate": 9.342363510727773e-05, "loss": 3.3433, "step": 94350 }, { "epoch": 0.8336424168565323, "grad_norm": 1.070917010307312, "learning_rate": 9.341674458013208e-05, "loss": 3.5517, "step": 94400 }, { "epoch": 0.8340839647468165, "grad_norm": 4.740517616271973, "learning_rate": 9.340985069941105e-05, "loss": 3.3758, "step": 94450 }, { "epoch": 0.8345255126371006, "grad_norm": 0.6648314595222473, "learning_rate": 9.340295346564719e-05, "loss": 3.5991, "step": 94500 }, { "epoch": 0.8349670605273848, "grad_norm": 3.5965664386749268, "learning_rate": 9.339605287937319e-05, "loss": 3.4054, "step": 94550 }, { "epoch": 0.835408608417669, "grad_norm": 2.838550329208374, "learning_rate": 9.33891489411221e-05, "loss": 3.0964, "step": 94600 }, { "epoch": 0.8358501563079531, "grad_norm": 1.3823540210723877, "learning_rate": 9.338224165142719e-05, "loss": 3.5835, "step": 94650 }, { "epoch": 0.8362917041982374, "grad_norm": 2.0661122798919678, "learning_rate": 9.337533101082199e-05, "loss": 3.2928, "step": 94700 }, { "epoch": 0.8367332520885216, "grad_norm": 1.8618370294570923, "learning_rate": 9.336841701984027e-05, "loss": 3.0778, "step": 94750 }, { "epoch": 0.8371747999788057, "grad_norm": 4.694985389709473, "learning_rate": 9.336149967901609e-05, "loss": 3.0738, "step": 94800 }, { "epoch": 0.8376163478690899, "grad_norm": 2.5283780097961426, "learning_rate": 9.335457898888376e-05, "loss": 3.4458, "step": 94850 }, { "epoch": 0.8380578957593741, "grad_norm": 1.0920939445495605, "learning_rate": 9.334765494997784e-05, "loss": 3.2661, "step": 94900 }, { "epoch": 0.8384994436496582, "grad_norm": 2.00156307220459, "learning_rate": 9.334072756283314e-05, "loss": 3.2389, "step": 94950 }, { "epoch": 0.8389409915399424, "grad_norm": 2.5333197116851807, "learning_rate": 9.333379682798475e-05, "loss": 3.5175, "step": 95000 }, { "epoch": 0.8393825394302266, "grad_norm": 2.970446825027466, "learning_rate": 9.332686274596803e-05, "loss": 3.0734, "step": 95050 }, { "epoch": 0.8398240873205108, "grad_norm": 1.0278738737106323, "learning_rate": 9.331992531731857e-05, "loss": 3.5306, "step": 95100 }, { "epoch": 0.840265635210795, "grad_norm": 2.5087051391601562, "learning_rate": 9.331298454257221e-05, "loss": 3.3883, "step": 95150 }, { "epoch": 0.8407071831010792, "grad_norm": 1.317252278327942, "learning_rate": 9.330604042226507e-05, "loss": 3.2772, "step": 95200 }, { "epoch": 0.8411487309913633, "grad_norm": 2.6759376525878906, "learning_rate": 9.329909295693355e-05, "loss": 2.9157, "step": 95250 }, { "epoch": 0.8415902788816475, "grad_norm": 1.9456520080566406, "learning_rate": 9.329214214711426e-05, "loss": 3.2917, "step": 95300 }, { "epoch": 0.8420318267719317, "grad_norm": 2.113647937774658, "learning_rate": 9.328532710918674e-05, "loss": 3.3203, "step": 95350 }, { "epoch": 0.8424733746622158, "grad_norm": 1.5438154935836792, "learning_rate": 9.327836967886584e-05, "loss": 3.6364, "step": 95400 }, { "epoch": 0.8429149225525, "grad_norm": 1.4090991020202637, "learning_rate": 9.32714089056579e-05, "loss": 3.2839, "step": 95450 }, { "epoch": 0.8433564704427843, "grad_norm": 3.7097344398498535, "learning_rate": 9.326444479010055e-05, "loss": 3.1992, "step": 95500 }, { "epoch": 0.8437980183330684, "grad_norm": 1.130540370941162, "learning_rate": 9.32574773327317e-05, "loss": 3.123, "step": 95550 }, { "epoch": 0.8442395662233526, "grad_norm": 3.1675918102264404, "learning_rate": 9.325050653408954e-05, "loss": 3.4446, "step": 95600 }, { "epoch": 0.8446811141136368, "grad_norm": 7.838589191436768, "learning_rate": 9.32435323947125e-05, "loss": 3.1731, "step": 95650 }, { "epoch": 0.8451226620039209, "grad_norm": 2.1061744689941406, "learning_rate": 9.32365549151393e-05, "loss": 3.1445, "step": 95700 }, { "epoch": 0.8455642098942051, "grad_norm": 1.585892677307129, "learning_rate": 9.322957409590884e-05, "loss": 3.248, "step": 95750 }, { "epoch": 0.8460057577844893, "grad_norm": 3.2491941452026367, "learning_rate": 9.322258993756037e-05, "loss": 3.1858, "step": 95800 }, { "epoch": 0.8464473056747734, "grad_norm": 4.585638999938965, "learning_rate": 9.321560244063334e-05, "loss": 3.3065, "step": 95850 }, { "epoch": 0.8468888535650577, "grad_norm": 1.343159556388855, "learning_rate": 9.320861160566748e-05, "loss": 3.3079, "step": 95900 }, { "epoch": 0.8473304014553419, "grad_norm": 4.3721699714660645, "learning_rate": 9.320161743320277e-05, "loss": 3.1931, "step": 95950 }, { "epoch": 0.847771949345626, "grad_norm": 1.243537425994873, "learning_rate": 9.319461992377945e-05, "loss": 3.2813, "step": 96000 }, { "epoch": 0.847771949345626, "eval_asr_loss": 0.9113168214024796, "eval_loss": 2.9808547496795654, "eval_runtime": 20.2448, "eval_samples_per_second": 37.936, "eval_steps_per_second": 9.484, "eval_tts_loss": 5.98077398558446, "step": 96000 }, { "epoch": 0.8482134972359102, "grad_norm": 3.9162681102752686, "learning_rate": 9.318761907793804e-05, "loss": 3.2729, "step": 96050 }, { "epoch": 0.8486550451261944, "grad_norm": 0.9850695729255676, "learning_rate": 9.318061489621925e-05, "loss": 3.4972, "step": 96100 }, { "epoch": 0.8490965930164786, "grad_norm": 2.3384222984313965, "learning_rate": 9.317360737916415e-05, "loss": 3.5748, "step": 96150 }, { "epoch": 0.8495381409067627, "grad_norm": 1.6895709037780762, "learning_rate": 9.316659652731395e-05, "loss": 3.5112, "step": 96200 }, { "epoch": 0.8499796887970469, "grad_norm": 4.779604434967041, "learning_rate": 9.315958234121023e-05, "loss": 3.4245, "step": 96250 }, { "epoch": 0.8504212366873312, "grad_norm": 3.7872555255889893, "learning_rate": 9.315256482139475e-05, "loss": 3.4721, "step": 96300 }, { "epoch": 0.8508627845776153, "grad_norm": 1.2569001913070679, "learning_rate": 9.314554396840956e-05, "loss": 3.2695, "step": 96350 }, { "epoch": 0.8513043324678995, "grad_norm": 2.095500946044922, "learning_rate": 9.313851978279697e-05, "loss": 3.1881, "step": 96400 }, { "epoch": 0.8517458803581837, "grad_norm": 0.9828936457633972, "learning_rate": 9.313149226509952e-05, "loss": 3.0915, "step": 96450 }, { "epoch": 0.8521874282484678, "grad_norm": 2.886770725250244, "learning_rate": 9.312446141586004e-05, "loss": 3.4099, "step": 96500 }, { "epoch": 0.852628976138752, "grad_norm": 4.082062721252441, "learning_rate": 9.31174272356216e-05, "loss": 3.0428, "step": 96550 }, { "epoch": 0.8530705240290362, "grad_norm": 1.261383295059204, "learning_rate": 9.311038972492754e-05, "loss": 3.2126, "step": 96600 }, { "epoch": 0.8535120719193203, "grad_norm": 1.2416714429855347, "learning_rate": 9.310334888432142e-05, "loss": 3.2021, "step": 96650 }, { "epoch": 0.8539536198096046, "grad_norm": 3.1885976791381836, "learning_rate": 9.309630471434712e-05, "loss": 2.7708, "step": 96700 }, { "epoch": 0.8543951676998888, "grad_norm": 0.710462749004364, "learning_rate": 9.308925721554874e-05, "loss": 3.199, "step": 96750 }, { "epoch": 0.8548367155901729, "grad_norm": 5.235406875610352, "learning_rate": 9.308220638847059e-05, "loss": 3.2963, "step": 96800 }, { "epoch": 0.8552782634804571, "grad_norm": 1.9540880918502808, "learning_rate": 9.307515223365735e-05, "loss": 3.3895, "step": 96850 }, { "epoch": 0.8557198113707413, "grad_norm": 2.796398162841797, "learning_rate": 9.306809475165385e-05, "loss": 3.4795, "step": 96900 }, { "epoch": 0.8561613592610254, "grad_norm": 1.8404650688171387, "learning_rate": 9.306103394300525e-05, "loss": 3.2471, "step": 96950 }, { "epoch": 0.8566029071513096, "grad_norm": 1.5861579179763794, "learning_rate": 9.305396980825693e-05, "loss": 2.8261, "step": 97000 }, { "epoch": 0.8570444550415938, "grad_norm": 5.508571624755859, "learning_rate": 9.304690234795453e-05, "loss": 3.0444, "step": 97050 }, { "epoch": 0.857486002931878, "grad_norm": 3.7076914310455322, "learning_rate": 9.303983156264393e-05, "loss": 3.3217, "step": 97100 }, { "epoch": 0.8579275508221622, "grad_norm": 4.032550811767578, "learning_rate": 9.303275745287133e-05, "loss": 3.3205, "step": 97150 }, { "epoch": 0.8583690987124464, "grad_norm": 0.970733106136322, "learning_rate": 9.30256800191831e-05, "loss": 3.1487, "step": 97200 }, { "epoch": 0.8588106466027305, "grad_norm": 2.3852109909057617, "learning_rate": 9.301859926212595e-05, "loss": 3.1561, "step": 97250 }, { "epoch": 0.8592521944930147, "grad_norm": 1.1150777339935303, "learning_rate": 9.301151518224679e-05, "loss": 2.9001, "step": 97300 }, { "epoch": 0.8596937423832989, "grad_norm": 3.74100923538208, "learning_rate": 9.300442778009282e-05, "loss": 3.2136, "step": 97350 }, { "epoch": 0.860135290273583, "grad_norm": 2.762349843978882, "learning_rate": 9.299733705621145e-05, "loss": 3.4975, "step": 97400 }, { "epoch": 0.8605768381638672, "grad_norm": 0.8698297142982483, "learning_rate": 9.299024301115043e-05, "loss": 3.3572, "step": 97450 }, { "epoch": 0.8610183860541515, "grad_norm": 3.2828052043914795, "learning_rate": 9.298314564545765e-05, "loss": 3.3767, "step": 97500 }, { "epoch": 0.8614599339444357, "grad_norm": 2.659813642501831, "learning_rate": 9.297604495968138e-05, "loss": 3.5504, "step": 97550 }, { "epoch": 0.8619014818347198, "grad_norm": 1.7123510837554932, "learning_rate": 9.296894095437005e-05, "loss": 3.4038, "step": 97600 }, { "epoch": 0.862343029725004, "grad_norm": 2.9085240364074707, "learning_rate": 9.296183363007241e-05, "loss": 2.8643, "step": 97650 }, { "epoch": 0.8627845776152882, "grad_norm": 1.4961036443710327, "learning_rate": 9.295472298733742e-05, "loss": 3.1688, "step": 97700 }, { "epoch": 0.8632261255055723, "grad_norm": 2.028132677078247, "learning_rate": 9.294760902671432e-05, "loss": 3.3077, "step": 97750 }, { "epoch": 0.8636676733958565, "grad_norm": 1.637831449508667, "learning_rate": 9.294049174875261e-05, "loss": 2.9796, "step": 97800 }, { "epoch": 0.8641092212861406, "grad_norm": 1.5504677295684814, "learning_rate": 9.293337115400205e-05, "loss": 2.8288, "step": 97850 }, { "epoch": 0.8645507691764249, "grad_norm": 3.117736577987671, "learning_rate": 9.292638975372799e-05, "loss": 2.9301, "step": 97900 }, { "epoch": 0.8649923170667091, "grad_norm": 2.2668349742889404, "learning_rate": 9.291926259335835e-05, "loss": 3.8076, "step": 97950 }, { "epoch": 0.8654338649569933, "grad_norm": 3.2014098167419434, "learning_rate": 9.29121321178396e-05, "loss": 3.1708, "step": 98000 }, { "epoch": 0.8658754128472774, "grad_norm": 3.2920122146606445, "learning_rate": 9.290499832772255e-05, "loss": 3.3037, "step": 98050 }, { "epoch": 0.8663169607375616, "grad_norm": 1.8121235370635986, "learning_rate": 9.28978612235582e-05, "loss": 3.2196, "step": 98100 }, { "epoch": 0.8667585086278458, "grad_norm": 1.6291028261184692, "learning_rate": 9.289072080589783e-05, "loss": 3.2285, "step": 98150 }, { "epoch": 0.8672000565181299, "grad_norm": 1.6624693870544434, "learning_rate": 9.288357707529297e-05, "loss": 3.1948, "step": 98200 }, { "epoch": 0.8676416044084141, "grad_norm": 2.4642486572265625, "learning_rate": 9.287643003229543e-05, "loss": 3.0989, "step": 98250 }, { "epoch": 0.8680831522986984, "grad_norm": 2.3431122303009033, "learning_rate": 9.286927967745725e-05, "loss": 3.0297, "step": 98300 }, { "epoch": 0.8685247001889825, "grad_norm": 1.7271332740783691, "learning_rate": 9.286212601133076e-05, "loss": 3.7604, "step": 98350 }, { "epoch": 0.8689662480792667, "grad_norm": 2.332041025161743, "learning_rate": 9.285496903446846e-05, "loss": 3.4561, "step": 98400 }, { "epoch": 0.8694077959695509, "grad_norm": 1.6665946245193481, "learning_rate": 9.284780874742321e-05, "loss": 3.2693, "step": 98450 }, { "epoch": 0.869849343859835, "grad_norm": 1.6122417449951172, "learning_rate": 9.284064515074809e-05, "loss": 3.2516, "step": 98500 }, { "epoch": 0.8702908917501192, "grad_norm": 0.9736064672470093, "learning_rate": 9.283347824499639e-05, "loss": 3.3643, "step": 98550 }, { "epoch": 0.8707324396404034, "grad_norm": 2.6523666381835938, "learning_rate": 9.282630803072173e-05, "loss": 3.4379, "step": 98600 }, { "epoch": 0.8711739875306875, "grad_norm": 3.84800386428833, "learning_rate": 9.281913450847792e-05, "loss": 3.6332, "step": 98650 }, { "epoch": 0.8716155354209718, "grad_norm": 1.0476601123809814, "learning_rate": 9.281195767881908e-05, "loss": 3.1238, "step": 98700 }, { "epoch": 0.872057083311256, "grad_norm": 2.119279623031616, "learning_rate": 9.280477754229952e-05, "loss": 3.4006, "step": 98750 }, { "epoch": 0.8724986312015401, "grad_norm": 1.7416213750839233, "learning_rate": 9.279759409947388e-05, "loss": 3.5637, "step": 98800 }, { "epoch": 0.8729401790918243, "grad_norm": 4.38261079788208, "learning_rate": 9.279040735089702e-05, "loss": 3.233, "step": 98850 }, { "epoch": 0.8733817269821085, "grad_norm": 4.000781059265137, "learning_rate": 9.278321729712403e-05, "loss": 3.5001, "step": 98900 }, { "epoch": 0.8738232748723926, "grad_norm": 2.074897050857544, "learning_rate": 9.27760239387103e-05, "loss": 3.0254, "step": 98950 }, { "epoch": 0.8742648227626768, "grad_norm": 3.6822054386138916, "learning_rate": 9.276882727621146e-05, "loss": 3.2075, "step": 99000 }, { "epoch": 0.8742648227626768, "eval_asr_loss": 0.9122765446783635, "eval_loss": 2.9597482681274414, "eval_runtime": 21.0264, "eval_samples_per_second": 36.525, "eval_steps_per_second": 9.131, "eval_tts_loss": 5.931117558903506, "step": 99000 }, { "epoch": 0.874706370652961, "grad_norm": 2.583610773086548, "learning_rate": 9.276162731018336e-05, "loss": 2.9525, "step": 99050 }, { "epoch": 0.8751479185432453, "grad_norm": 3.8291828632354736, "learning_rate": 9.275442404118217e-05, "loss": 3.5362, "step": 99100 }, { "epoch": 0.8755894664335294, "grad_norm": 1.9839342832565308, "learning_rate": 9.274721746976427e-05, "loss": 3.168, "step": 99150 }, { "epoch": 0.8760310143238136, "grad_norm": 2.494183301925659, "learning_rate": 9.274000759648632e-05, "loss": 3.0813, "step": 99200 }, { "epoch": 0.8764725622140978, "grad_norm": 1.9744882583618164, "learning_rate": 9.273279442190519e-05, "loss": 3.3301, "step": 99250 }, { "epoch": 0.8769141101043819, "grad_norm": 1.0826927423477173, "learning_rate": 9.272557794657805e-05, "loss": 3.2976, "step": 99300 }, { "epoch": 0.8773556579946661, "grad_norm": 3.1494545936584473, "learning_rate": 9.271835817106234e-05, "loss": 3.2733, "step": 99350 }, { "epoch": 0.8777972058849502, "grad_norm": 2.1785781383514404, "learning_rate": 9.271113509591568e-05, "loss": 3.1884, "step": 99400 }, { "epoch": 0.8782387537752344, "grad_norm": 1.9323933124542236, "learning_rate": 9.270390872169602e-05, "loss": 3.4076, "step": 99450 }, { "epoch": 0.8786803016655187, "grad_norm": 2.127812385559082, "learning_rate": 9.269667904896153e-05, "loss": 3.3259, "step": 99500 }, { "epoch": 0.8791218495558029, "grad_norm": 2.901846170425415, "learning_rate": 9.268944607827064e-05, "loss": 3.878, "step": 99550 }, { "epoch": 0.879563397446087, "grad_norm": 2.7331161499023438, "learning_rate": 9.268220981018204e-05, "loss": 3.0213, "step": 99600 }, { "epoch": 0.8800049453363712, "grad_norm": 1.2581075429916382, "learning_rate": 9.267497024525466e-05, "loss": 3.0469, "step": 99650 }, { "epoch": 0.8804464932266554, "grad_norm": 1.963077187538147, "learning_rate": 9.266772738404771e-05, "loss": 3.0544, "step": 99700 }, { "epoch": 0.8808880411169395, "grad_norm": 2.728001594543457, "learning_rate": 9.266048122712064e-05, "loss": 3.0793, "step": 99750 }, { "epoch": 0.8813295890072237, "grad_norm": 2.3829917907714844, "learning_rate": 9.265323177503312e-05, "loss": 3.2839, "step": 99800 }, { "epoch": 0.8817711368975079, "grad_norm": 0.8417502045631409, "learning_rate": 9.264597902834515e-05, "loss": 3.1644, "step": 99850 }, { "epoch": 0.8822126847877921, "grad_norm": 3.0022170543670654, "learning_rate": 9.263872298761693e-05, "loss": 3.4081, "step": 99900 }, { "epoch": 0.8826542326780763, "grad_norm": 5.986268043518066, "learning_rate": 9.263146365340891e-05, "loss": 3.1996, "step": 99950 }, { "epoch": 0.8830957805683605, "grad_norm": 2.8962159156799316, "learning_rate": 9.262420102628184e-05, "loss": 3.1852, "step": 100000 }, { "epoch": 0.8835373284586446, "grad_norm": 2.588916540145874, "learning_rate": 9.261693510679668e-05, "loss": 3.1378, "step": 100050 }, { "epoch": 0.8839788763489288, "grad_norm": 2.7760961055755615, "learning_rate": 9.260966589551468e-05, "loss": 3.2112, "step": 100100 }, { "epoch": 0.884420424239213, "grad_norm": 1.5340514183044434, "learning_rate": 9.260239339299727e-05, "loss": 3.5164, "step": 100150 }, { "epoch": 0.8848619721294971, "grad_norm": 9.54226303100586, "learning_rate": 9.259511759980625e-05, "loss": 3.5849, "step": 100200 }, { "epoch": 0.8853035200197813, "grad_norm": 1.6671772003173828, "learning_rate": 9.258783851650361e-05, "loss": 3.4115, "step": 100250 }, { "epoch": 0.8857450679100656, "grad_norm": 1.8623788356781006, "learning_rate": 9.258055614365156e-05, "loss": 3.3499, "step": 100300 }, { "epoch": 0.8861866158003497, "grad_norm": 1.5677406787872314, "learning_rate": 9.257327048181262e-05, "loss": 3.2248, "step": 100350 }, { "epoch": 0.8866281636906339, "grad_norm": 1.939414381980896, "learning_rate": 9.256598153154954e-05, "loss": 3.419, "step": 100400 }, { "epoch": 0.8870697115809181, "grad_norm": 2.109490394592285, "learning_rate": 9.255883517040522e-05, "loss": 3.5276, "step": 100450 }, { "epoch": 0.8875112594712022, "grad_norm": 3.756086826324463, "learning_rate": 9.255153971072361e-05, "loss": 3.6449, "step": 100500 }, { "epoch": 0.8879528073614864, "grad_norm": 3.2874066829681396, "learning_rate": 9.254424096429637e-05, "loss": 3.452, "step": 100550 }, { "epoch": 0.8883943552517706, "grad_norm": 1.6661789417266846, "learning_rate": 9.253693893168729e-05, "loss": 3.2744, "step": 100600 }, { "epoch": 0.8888359031420547, "grad_norm": 3.6185531616210938, "learning_rate": 9.252963361346036e-05, "loss": 2.9505, "step": 100650 }, { "epoch": 0.889277451032339, "grad_norm": 5.086017608642578, "learning_rate": 9.25223250101799e-05, "loss": 2.9579, "step": 100700 }, { "epoch": 0.8897189989226232, "grad_norm": 3.7332069873809814, "learning_rate": 9.25150131224104e-05, "loss": 3.1295, "step": 100750 }, { "epoch": 0.8901605468129073, "grad_norm": 3.6353108882904053, "learning_rate": 9.250769795071666e-05, "loss": 3.6284, "step": 100800 }, { "epoch": 0.8906020947031915, "grad_norm": 6.214040279388428, "learning_rate": 9.250037949566368e-05, "loss": 2.6618, "step": 100850 }, { "epoch": 0.8910436425934757, "grad_norm": 3.569542169570923, "learning_rate": 9.249305775781681e-05, "loss": 3.5992, "step": 100900 }, { "epoch": 0.8914851904837598, "grad_norm": 2.4265153408050537, "learning_rate": 9.248573273774154e-05, "loss": 3.8598, "step": 100950 }, { "epoch": 0.891926738374044, "grad_norm": 2.2118725776672363, "learning_rate": 9.247840443600368e-05, "loss": 2.7226, "step": 101000 }, { "epoch": 0.8923682862643282, "grad_norm": 1.527198076248169, "learning_rate": 9.24710728531693e-05, "loss": 3.3547, "step": 101050 }, { "epoch": 0.8928098341546125, "grad_norm": 2.138624668121338, "learning_rate": 9.246373798980468e-05, "loss": 3.5805, "step": 101100 }, { "epoch": 0.8932513820448966, "grad_norm": 3.642125129699707, "learning_rate": 9.245639984647639e-05, "loss": 3.2457, "step": 101150 }, { "epoch": 0.8936929299351808, "grad_norm": 1.057465672492981, "learning_rate": 9.244905842375122e-05, "loss": 3.2628, "step": 101200 }, { "epoch": 0.894134477825465, "grad_norm": 1.685719609260559, "learning_rate": 9.244171372219626e-05, "loss": 3.8842, "step": 101250 }, { "epoch": 0.8945760257157491, "grad_norm": 1.3749964237213135, "learning_rate": 9.243436574237878e-05, "loss": 3.0899, "step": 101300 }, { "epoch": 0.8950175736060333, "grad_norm": 2.584540367126465, "learning_rate": 9.24270144848664e-05, "loss": 3.6703, "step": 101350 }, { "epoch": 0.8954591214963175, "grad_norm": 3.092801570892334, "learning_rate": 9.241965995022692e-05, "loss": 3.1755, "step": 101400 }, { "epoch": 0.8959006693866016, "grad_norm": 1.243997573852539, "learning_rate": 9.241230213902842e-05, "loss": 2.8785, "step": 101450 }, { "epoch": 0.8963422172768859, "grad_norm": 4.314537048339844, "learning_rate": 9.240494105183921e-05, "loss": 3.3101, "step": 101500 }, { "epoch": 0.8967837651671701, "grad_norm": 2.9184675216674805, "learning_rate": 9.239757668922791e-05, "loss": 3.318, "step": 101550 }, { "epoch": 0.8972253130574542, "grad_norm": 2.036470890045166, "learning_rate": 9.239020905176331e-05, "loss": 3.1143, "step": 101600 }, { "epoch": 0.8976668609477384, "grad_norm": 4.395949363708496, "learning_rate": 9.238283814001454e-05, "loss": 3.2825, "step": 101650 }, { "epoch": 0.8981084088380226, "grad_norm": 2.0808677673339844, "learning_rate": 9.23754639545509e-05, "loss": 3.7597, "step": 101700 }, { "epoch": 0.8985499567283067, "grad_norm": 1.0310138463974, "learning_rate": 9.236808649594202e-05, "loss": 3.3764, "step": 101750 }, { "epoch": 0.8989915046185909, "grad_norm": 1.490734338760376, "learning_rate": 9.236070576475773e-05, "loss": 3.3401, "step": 101800 }, { "epoch": 0.8994330525088751, "grad_norm": 1.4883259534835815, "learning_rate": 9.235332176156812e-05, "loss": 3.3028, "step": 101850 }, { "epoch": 0.8998746003991593, "grad_norm": 1.0461546182632446, "learning_rate": 9.234593448694354e-05, "loss": 3.007, "step": 101900 }, { "epoch": 0.9003161482894435, "grad_norm": 1.0872769355773926, "learning_rate": 9.233854394145462e-05, "loss": 3.1332, "step": 101950 }, { "epoch": 0.9007576961797277, "grad_norm": 4.982665061950684, "learning_rate": 9.23311501256722e-05, "loss": 3.3656, "step": 102000 }, { "epoch": 0.9007576961797277, "eval_asr_loss": 0.903455578837743, "eval_loss": 2.968921661376953, "eval_runtime": 20.4043, "eval_samples_per_second": 37.639, "eval_steps_per_second": 9.41, "eval_tts_loss": 5.955756151524818, "step": 102000 }, { "epoch": 0.9011992440700118, "grad_norm": 1.7485255002975464, "learning_rate": 9.232375304016738e-05, "loss": 3.2068, "step": 102050 }, { "epoch": 0.901640791960296, "grad_norm": 2.655881643295288, "learning_rate": 9.231635268551153e-05, "loss": 3.2794, "step": 102100 }, { "epoch": 0.9020823398505802, "grad_norm": 1.1869397163391113, "learning_rate": 9.230894906227628e-05, "loss": 3.6545, "step": 102150 }, { "epoch": 0.9025238877408643, "grad_norm": 1.75388765335083, "learning_rate": 9.23016903408811e-05, "loss": 3.4393, "step": 102200 }, { "epoch": 0.9029654356311485, "grad_norm": 2.8003134727478027, "learning_rate": 9.229428024754599e-05, "loss": 3.6308, "step": 102250 }, { "epoch": 0.9034069835214328, "grad_norm": 2.303340196609497, "learning_rate": 9.228686688733635e-05, "loss": 3.4285, "step": 102300 }, { "epoch": 0.903848531411717, "grad_norm": 1.432195782661438, "learning_rate": 9.227945026082483e-05, "loss": 3.0429, "step": 102350 }, { "epoch": 0.9042900793020011, "grad_norm": 3.9534683227539062, "learning_rate": 9.227203036858429e-05, "loss": 3.385, "step": 102400 }, { "epoch": 0.9047316271922853, "grad_norm": 1.3392046689987183, "learning_rate": 9.226460721118788e-05, "loss": 2.9859, "step": 102450 }, { "epoch": 0.9051731750825694, "grad_norm": 1.7585742473602295, "learning_rate": 9.225718078920896e-05, "loss": 2.8935, "step": 102500 }, { "epoch": 0.9056147229728536, "grad_norm": 1.3776382207870483, "learning_rate": 9.224975110322113e-05, "loss": 3.2253, "step": 102550 }, { "epoch": 0.9060562708631378, "grad_norm": 0.8360838294029236, "learning_rate": 9.22423181537983e-05, "loss": 2.9125, "step": 102600 }, { "epoch": 0.906497818753422, "grad_norm": 2.6101293563842773, "learning_rate": 9.223488194151461e-05, "loss": 3.0466, "step": 102650 }, { "epoch": 0.9069393666437062, "grad_norm": 2.8164517879486084, "learning_rate": 9.222744246694441e-05, "loss": 3.3639, "step": 102700 }, { "epoch": 0.9073809145339904, "grad_norm": 2.205768585205078, "learning_rate": 9.221999973066238e-05, "loss": 3.3834, "step": 102750 }, { "epoch": 0.9078224624242746, "grad_norm": 3.8350329399108887, "learning_rate": 9.221255373324338e-05, "loss": 2.7993, "step": 102800 }, { "epoch": 0.9082640103145587, "grad_norm": 3.8770978450775146, "learning_rate": 9.220510447526256e-05, "loss": 2.9693, "step": 102850 }, { "epoch": 0.9087055582048429, "grad_norm": 2.760380983352661, "learning_rate": 9.219765195729529e-05, "loss": 3.5722, "step": 102900 }, { "epoch": 0.9091471060951271, "grad_norm": 3.3382771015167236, "learning_rate": 9.219019617991724e-05, "loss": 3.5588, "step": 102950 }, { "epoch": 0.9095886539854112, "grad_norm": 2.2112584114074707, "learning_rate": 9.218273714370432e-05, "loss": 3.3712, "step": 103000 }, { "epoch": 0.9100302018756954, "grad_norm": 2.458007335662842, "learning_rate": 9.217527484923262e-05, "loss": 2.5991, "step": 103050 }, { "epoch": 0.9104717497659797, "grad_norm": 2.0607690811157227, "learning_rate": 9.21678092970786e-05, "loss": 3.4419, "step": 103100 }, { "epoch": 0.9109132976562638, "grad_norm": 3.3169896602630615, "learning_rate": 9.216034048781887e-05, "loss": 2.8659, "step": 103150 }, { "epoch": 0.911354845546548, "grad_norm": 2.7883317470550537, "learning_rate": 9.215286842203035e-05, "loss": 3.6627, "step": 103200 }, { "epoch": 0.9117963934368322, "grad_norm": 1.9193239212036133, "learning_rate": 9.21453931002902e-05, "loss": 3.2209, "step": 103250 }, { "epoch": 0.9122379413271163, "grad_norm": 1.5039438009262085, "learning_rate": 9.213791452317579e-05, "loss": 3.2848, "step": 103300 }, { "epoch": 0.9126794892174005, "grad_norm": 1.5526306629180908, "learning_rate": 9.213043269126483e-05, "loss": 3.0704, "step": 103350 }, { "epoch": 0.9131210371076847, "grad_norm": 1.0206029415130615, "learning_rate": 9.212294760513518e-05, "loss": 3.675, "step": 103400 }, { "epoch": 0.9135625849979688, "grad_norm": 4.434154510498047, "learning_rate": 9.211545926536504e-05, "loss": 3.4542, "step": 103450 }, { "epoch": 0.9140041328882531, "grad_norm": 2.1419999599456787, "learning_rate": 9.210796767253278e-05, "loss": 3.8326, "step": 103500 }, { "epoch": 0.9144456807785373, "grad_norm": 0.9824464321136475, "learning_rate": 9.210047282721709e-05, "loss": 3.2751, "step": 103550 }, { "epoch": 0.9148872286688214, "grad_norm": 3.2531392574310303, "learning_rate": 9.20929747299969e-05, "loss": 3.1043, "step": 103600 }, { "epoch": 0.9153287765591056, "grad_norm": 7.856770992279053, "learning_rate": 9.208547338145132e-05, "loss": 3.0719, "step": 103650 }, { "epoch": 0.9157703244493898, "grad_norm": 0.7467933893203735, "learning_rate": 9.207796878215979e-05, "loss": 3.6067, "step": 103700 }, { "epoch": 0.9162118723396739, "grad_norm": 11.154623985290527, "learning_rate": 9.2070460932702e-05, "loss": 3.2332, "step": 103750 }, { "epoch": 0.9166534202299581, "grad_norm": 4.158257484436035, "learning_rate": 9.206294983365785e-05, "loss": 3.4292, "step": 103800 }, { "epoch": 0.9170949681202423, "grad_norm": 1.339894413948059, "learning_rate": 9.205543548560751e-05, "loss": 3.1509, "step": 103850 }, { "epoch": 0.9175365160105265, "grad_norm": 1.5442434549331665, "learning_rate": 9.204791788913138e-05, "loss": 3.0922, "step": 103900 }, { "epoch": 0.9179780639008107, "grad_norm": 1.0537227392196655, "learning_rate": 9.204039704481018e-05, "loss": 3.2525, "step": 103950 }, { "epoch": 0.9184196117910949, "grad_norm": 2.2441678047180176, "learning_rate": 9.203287295322478e-05, "loss": 2.9392, "step": 104000 }, { "epoch": 0.918861159681379, "grad_norm": 3.044712543487549, "learning_rate": 9.202534561495637e-05, "loss": 3.0516, "step": 104050 }, { "epoch": 0.9193027075716632, "grad_norm": 1.7708096504211426, "learning_rate": 9.201781503058638e-05, "loss": 3.3234, "step": 104100 }, { "epoch": 0.9197442554619474, "grad_norm": 2.0960426330566406, "learning_rate": 9.201028120069648e-05, "loss": 3.808, "step": 104150 }, { "epoch": 0.9201858033522315, "grad_norm": 6.793859958648682, "learning_rate": 9.20027441258686e-05, "loss": 3.6451, "step": 104200 }, { "epoch": 0.9206273512425157, "grad_norm": 2.1247751712799072, "learning_rate": 9.19952038066849e-05, "loss": 3.6018, "step": 104250 }, { "epoch": 0.9210688991328, "grad_norm": 1.15399169921875, "learning_rate": 9.198766024372783e-05, "loss": 3.3271, "step": 104300 }, { "epoch": 0.9215104470230842, "grad_norm": 1.7735700607299805, "learning_rate": 9.198011343758002e-05, "loss": 3.0006, "step": 104350 }, { "epoch": 0.9219519949133683, "grad_norm": 0.8139469027519226, "learning_rate": 9.197256338882445e-05, "loss": 3.0361, "step": 104400 }, { "epoch": 0.9223935428036525, "grad_norm": 3.3219869136810303, "learning_rate": 9.196501009804429e-05, "loss": 3.1615, "step": 104450 }, { "epoch": 0.9228350906939367, "grad_norm": 2.1719770431518555, "learning_rate": 9.195745356582292e-05, "loss": 3.8578, "step": 104500 }, { "epoch": 0.9232766385842208, "grad_norm": 1.2096954584121704, "learning_rate": 9.194989379274406e-05, "loss": 3.1797, "step": 104550 }, { "epoch": 0.923718186474505, "grad_norm": 1.369635820388794, "learning_rate": 9.194233077939165e-05, "loss": 3.1194, "step": 104600 }, { "epoch": 0.9241597343647892, "grad_norm": 1.7169206142425537, "learning_rate": 9.193476452634983e-05, "loss": 3.3546, "step": 104650 }, { "epoch": 0.9246012822550734, "grad_norm": 5.232157230377197, "learning_rate": 9.192719503420304e-05, "loss": 3.2173, "step": 104700 }, { "epoch": 0.9250428301453576, "grad_norm": 1.7149889469146729, "learning_rate": 9.191962230353598e-05, "loss": 3.0774, "step": 104750 }, { "epoch": 0.9254843780356418, "grad_norm": 1.5316158533096313, "learning_rate": 9.191204633493355e-05, "loss": 3.6406, "step": 104800 }, { "epoch": 0.9259259259259259, "grad_norm": 2.05110239982605, "learning_rate": 9.190446712898097e-05, "loss": 3.3836, "step": 104850 }, { "epoch": 0.9263674738162101, "grad_norm": 3.319016695022583, "learning_rate": 9.189688468626363e-05, "loss": 3.1945, "step": 104900 }, { "epoch": 0.9268090217064943, "grad_norm": 2.672206163406372, "learning_rate": 9.188929900736722e-05, "loss": 3.5687, "step": 104950 }, { "epoch": 0.9272505695967784, "grad_norm": 1.4646509885787964, "learning_rate": 9.188171009287767e-05, "loss": 3.329, "step": 105000 }, { "epoch": 0.9272505695967784, "eval_asr_loss": 0.9064391918799752, "eval_loss": 2.951792001724243, "eval_runtime": 20.8436, "eval_samples_per_second": 36.846, "eval_steps_per_second": 9.211, "eval_tts_loss": 5.888024541957233, "step": 105000 } ], "logging_steps": 50, "max_steps": 566190, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 3000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 9.986515795926057e+17, "train_batch_size": 4, "trial_name": null, "trial_params": null }