diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,14223 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.8742648227626768, + "eval_steps": 3000, + "global_step": 99000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00044154789028418023, + "grad_norm": 4.407685279846191, + "learning_rate": 4.800000000000001e-06, + "loss": 4.2565, + "step": 50 + }, + { + "epoch": 0.0008830957805683605, + "grad_norm": 2.3940842151641846, + "learning_rate": 9.800000000000001e-06, + "loss": 4.3171, + "step": 100 + }, + { + "epoch": 0.0013246436708525408, + "grad_norm": 2.555833339691162, + "learning_rate": 1.48e-05, + "loss": 3.8795, + "step": 150 + }, + { + "epoch": 0.001766191561136721, + "grad_norm": 0.9911451935768127, + "learning_rate": 1.97e-05, + "loss": 3.9588, + "step": 200 + }, + { + "epoch": 0.0022077394514209013, + "grad_norm": 1.1606439352035522, + "learning_rate": 2.47e-05, + "loss": 3.4062, + "step": 250 + }, + { + "epoch": 0.0026492873417050815, + "grad_norm": 2.9307570457458496, + "learning_rate": 2.97e-05, + "loss": 3.5316, + "step": 300 + }, + { + "epoch": 0.0030908352319892617, + "grad_norm": 2.6516165733337402, + "learning_rate": 3.4699999999999996e-05, + "loss": 3.5324, + "step": 350 + }, + { + "epoch": 0.003532383122273442, + "grad_norm": 1.69759202003479, + "learning_rate": 3.97e-05, + "loss": 3.6864, + "step": 400 + }, + { + "epoch": 0.003973931012557622, + "grad_norm": 3.0430002212524414, + "learning_rate": 4.47e-05, + "loss": 3.6768, + "step": 450 + }, + { + "epoch": 0.004415478902841803, + "grad_norm": 4.26913595199585, + "learning_rate": 4.97e-05, + "loss": 3.9357, + "step": 500 + }, + { + "epoch": 0.004857026793125983, + "grad_norm": 1.9752435684204102, + "learning_rate": 5.470000000000001e-05, + "loss": 3.4989, + "step": 550 + }, + { + "epoch": 0.005298574683410163, + "grad_norm": 2.631711006164551, + "learning_rate": 5.96e-05, + "loss": 3.4577, + "step": 600 + }, + { + "epoch": 0.005740122573694343, + "grad_norm": 4.840415000915527, + "learning_rate": 6.460000000000001e-05, + "loss": 3.6496, + "step": 650 + }, + { + "epoch": 0.006181670463978523, + "grad_norm": 0.7839917540550232, + "learning_rate": 6.96e-05, + "loss": 3.4212, + "step": 700 + }, + { + "epoch": 0.0066232183542627035, + "grad_norm": 2.6035537719726562, + "learning_rate": 7.46e-05, + "loss": 3.5461, + "step": 750 + }, + { + "epoch": 0.007064766244546884, + "grad_norm": 2.7785916328430176, + "learning_rate": 7.960000000000001e-05, + "loss": 3.5111, + "step": 800 + }, + { + "epoch": 0.007506314134831064, + "grad_norm": 3.2137913703918457, + "learning_rate": 8.46e-05, + "loss": 3.4488, + "step": 850 + }, + { + "epoch": 0.007947862025115243, + "grad_norm": 2.4404215812683105, + "learning_rate": 8.960000000000001e-05, + "loss": 3.6674, + "step": 900 + }, + { + "epoch": 0.008389409915399423, + "grad_norm": 2.055772542953491, + "learning_rate": 9.46e-05, + "loss": 3.5125, + "step": 950 + }, + { + "epoch": 0.008830957805683605, + "grad_norm": 6.80244255065918, + "learning_rate": 9.960000000000001e-05, + "loss": 3.7395, + "step": 1000 + }, + { + "epoch": 0.009272505695967785, + "grad_norm": 1.6064447164535522, + "learning_rate": 9.999999836556943e-05, + "loss": 3.2688, + "step": 1050 + }, + { + "epoch": 0.009714053586251966, + "grad_norm": 1.4385908842086792, + "learning_rate": 9.999999288142163e-05, + "loss": 3.7015, + "step": 1100 + }, + { + "epoch": 0.010155601476536146, + "grad_norm": 2.0870840549468994, + "learning_rate": 9.999998353519836e-05, + "loss": 3.6805, + "step": 1150 + }, + { + "epoch": 0.010597149366820326, + "grad_norm": 1.5181084871292114, + "learning_rate": 9.999997032690033e-05, + "loss": 3.9943, + "step": 1200 + }, + { + "epoch": 0.011038697257104506, + "grad_norm": 2.0641415119171143, + "learning_rate": 9.99999532565286e-05, + "loss": 3.4652, + "step": 1250 + }, + { + "epoch": 0.011480245147388686, + "grad_norm": 4.051513671875, + "learning_rate": 9.999993232408446e-05, + "loss": 3.784, + "step": 1300 + }, + { + "epoch": 0.011921793037672867, + "grad_norm": 2.475578546524048, + "learning_rate": 9.999990752956954e-05, + "loss": 3.6213, + "step": 1350 + }, + { + "epoch": 0.012363340927957047, + "grad_norm": 2.650278091430664, + "learning_rate": 9.999987887298574e-05, + "loss": 3.2759, + "step": 1400 + }, + { + "epoch": 0.012804888818241227, + "grad_norm": 1.043623924255371, + "learning_rate": 9.999984635433529e-05, + "loss": 3.4098, + "step": 1450 + }, + { + "epoch": 0.013246436708525407, + "grad_norm": 2.3483691215515137, + "learning_rate": 9.999980997362069e-05, + "loss": 3.5984, + "step": 1500 + }, + { + "epoch": 0.013687984598809587, + "grad_norm": 1.8058457374572754, + "learning_rate": 9.999976973084477e-05, + "loss": 3.5686, + "step": 1550 + }, + { + "epoch": 0.014129532489093767, + "grad_norm": 1.9905356168746948, + "learning_rate": 9.999972562601062e-05, + "loss": 3.5255, + "step": 1600 + }, + { + "epoch": 0.014571080379377948, + "grad_norm": 8.096346855163574, + "learning_rate": 9.999967765912164e-05, + "loss": 3.8234, + "step": 1650 + }, + { + "epoch": 0.015012628269662128, + "grad_norm": 3.1411712169647217, + "learning_rate": 9.999962583018154e-05, + "loss": 3.3328, + "step": 1700 + }, + { + "epoch": 0.015454176159946308, + "grad_norm": 4.46273946762085, + "learning_rate": 9.999957013919435e-05, + "loss": 3.2951, + "step": 1750 + }, + { + "epoch": 0.015895724050230486, + "grad_norm": 1.3609850406646729, + "learning_rate": 9.999951058616435e-05, + "loss": 3.3173, + "step": 1800 + }, + { + "epoch": 0.016337271940514667, + "grad_norm": 0.8974264860153198, + "learning_rate": 9.999944717109613e-05, + "loss": 3.5905, + "step": 1850 + }, + { + "epoch": 0.016778819830798847, + "grad_norm": 4.310591220855713, + "learning_rate": 9.999937989399459e-05, + "loss": 3.2723, + "step": 1900 + }, + { + "epoch": 0.017220367721083027, + "grad_norm": 10.745691299438477, + "learning_rate": 9.999930875486495e-05, + "loss": 3.3104, + "step": 1950 + }, + { + "epoch": 0.01766191561136721, + "grad_norm": 1.678096890449524, + "learning_rate": 9.999923375371269e-05, + "loss": 3.4128, + "step": 2000 + }, + { + "epoch": 0.01810346350165139, + "grad_norm": 1.2710167169570923, + "learning_rate": 9.99991548905436e-05, + "loss": 3.8629, + "step": 2050 + }, + { + "epoch": 0.01854501139193557, + "grad_norm": 1.6546989679336548, + "learning_rate": 9.999907216536378e-05, + "loss": 3.9797, + "step": 2100 + }, + { + "epoch": 0.01898655928221975, + "grad_norm": 2.4010069370269775, + "learning_rate": 9.999898557817962e-05, + "loss": 3.4582, + "step": 2150 + }, + { + "epoch": 0.01942810717250393, + "grad_norm": 11.584942817687988, + "learning_rate": 9.999889512899778e-05, + "loss": 3.4392, + "step": 2200 + }, + { + "epoch": 0.01986965506278811, + "grad_norm": 2.588017702102661, + "learning_rate": 9.999880081782529e-05, + "loss": 3.2102, + "step": 2250 + }, + { + "epoch": 0.02031120295307229, + "grad_norm": 4.959451198577881, + "learning_rate": 9.99987026446694e-05, + "loss": 3.4403, + "step": 2300 + }, + { + "epoch": 0.020752750843356472, + "grad_norm": 2.886789560317993, + "learning_rate": 9.999860060953772e-05, + "loss": 3.3001, + "step": 2350 + }, + { + "epoch": 0.021194298733640652, + "grad_norm": 7.682253837585449, + "learning_rate": 9.999849471243812e-05, + "loss": 3.8255, + "step": 2400 + }, + { + "epoch": 0.021635846623924832, + "grad_norm": 9.694388389587402, + "learning_rate": 9.999838495337877e-05, + "loss": 3.8851, + "step": 2450 + }, + { + "epoch": 0.022077394514209012, + "grad_norm": 2.9934961795806885, + "learning_rate": 9.999827133236815e-05, + "loss": 3.6424, + "step": 2500 + }, + { + "epoch": 0.022518942404493193, + "grad_norm": 1.4004652500152588, + "learning_rate": 9.999815384941506e-05, + "loss": 3.5799, + "step": 2550 + }, + { + "epoch": 0.022960490294777373, + "grad_norm": 2.06247615814209, + "learning_rate": 9.999803250452856e-05, + "loss": 3.4358, + "step": 2600 + }, + { + "epoch": 0.023402038185061553, + "grad_norm": 2.051114082336426, + "learning_rate": 9.999790729771798e-05, + "loss": 3.6721, + "step": 2650 + }, + { + "epoch": 0.023843586075345733, + "grad_norm": 2.4515838623046875, + "learning_rate": 9.999777822899307e-05, + "loss": 3.6419, + "step": 2700 + }, + { + "epoch": 0.024285133965629913, + "grad_norm": 2.6057190895080566, + "learning_rate": 9.999764529836375e-05, + "loss": 3.4357, + "step": 2750 + }, + { + "epoch": 0.024726681855914093, + "grad_norm": 2.1552577018737793, + "learning_rate": 9.99975085058403e-05, + "loss": 3.6189, + "step": 2800 + }, + { + "epoch": 0.025168229746198274, + "grad_norm": 4.0472846031188965, + "learning_rate": 9.999736785143327e-05, + "loss": 3.5135, + "step": 2850 + }, + { + "epoch": 0.025609777636482454, + "grad_norm": 1.694504976272583, + "learning_rate": 9.999722333515355e-05, + "loss": 3.5287, + "step": 2900 + }, + { + "epoch": 0.026051325526766634, + "grad_norm": 8.814433097839355, + "learning_rate": 9.99970749570123e-05, + "loss": 3.4025, + "step": 2950 + }, + { + "epoch": 0.026492873417050814, + "grad_norm": 2.079963207244873, + "learning_rate": 9.999692579966685e-05, + "loss": 3.8211, + "step": 3000 + }, + { + "epoch": 0.026492873417050814, + "eval_asr_loss": 0.8104195239432453, + "eval_loss": 3.4588043689727783, + "eval_runtime": 20.2586, + "eval_samples_per_second": 37.91, + "eval_steps_per_second": 9.477, + "eval_tts_loss": 6.061187094347814, + "step": 3000 + }, + { + "epoch": 0.026934421307334994, + "grad_norm": 2.6620876789093018, + "learning_rate": 9.999676977507384e-05, + "loss": 3.4284, + "step": 3050 + }, + { + "epoch": 0.027375969197619174, + "grad_norm": 10.126931190490723, + "learning_rate": 9.999660988865433e-05, + "loss": 3.756, + "step": 3100 + }, + { + "epoch": 0.027817517087903355, + "grad_norm": 1.450592279434204, + "learning_rate": 9.999644614042068e-05, + "loss": 3.2147, + "step": 3150 + }, + { + "epoch": 0.028259064978187535, + "grad_norm": 1.703476071357727, + "learning_rate": 9.99962785303855e-05, + "loss": 3.5119, + "step": 3200 + }, + { + "epoch": 0.028700612868471715, + "grad_norm": 2.767446756362915, + "learning_rate": 9.999610705856177e-05, + "loss": 3.2736, + "step": 3250 + }, + { + "epoch": 0.029142160758755895, + "grad_norm": 1.1504226922988892, + "learning_rate": 9.999593172496273e-05, + "loss": 3.4068, + "step": 3300 + }, + { + "epoch": 0.029583708649040075, + "grad_norm": 1.6964422464370728, + "learning_rate": 9.99957525296019e-05, + "loss": 3.4917, + "step": 3350 + }, + { + "epoch": 0.030025256539324256, + "grad_norm": 1.8230127096176147, + "learning_rate": 9.999556947249316e-05, + "loss": 3.1574, + "step": 3400 + }, + { + "epoch": 0.030466804429608436, + "grad_norm": 3.8012404441833496, + "learning_rate": 9.999538255365062e-05, + "loss": 3.6179, + "step": 3450 + }, + { + "epoch": 0.030908352319892616, + "grad_norm": 3.4073257446289062, + "learning_rate": 9.999519177308871e-05, + "loss": 3.6502, + "step": 3500 + }, + { + "epoch": 0.03134990021017679, + "grad_norm": 1.7333426475524902, + "learning_rate": 9.99949971308222e-05, + "loss": 3.5175, + "step": 3550 + }, + { + "epoch": 0.03179144810046097, + "grad_norm": 1.5845162868499756, + "learning_rate": 9.99947986268661e-05, + "loss": 3.4373, + "step": 3600 + }, + { + "epoch": 0.03223299599074515, + "grad_norm": 4.4948225021362305, + "learning_rate": 9.999459626123576e-05, + "loss": 3.6571, + "step": 3650 + }, + { + "epoch": 0.03267454388102933, + "grad_norm": 3.1738150119781494, + "learning_rate": 9.99943900339468e-05, + "loss": 3.6251, + "step": 3700 + }, + { + "epoch": 0.03311609177131351, + "grad_norm": 2.0190846920013428, + "learning_rate": 9.999417994501514e-05, + "loss": 3.5417, + "step": 3750 + }, + { + "epoch": 0.033557639661597694, + "grad_norm": 3.240422248840332, + "learning_rate": 9.999396599445701e-05, + "loss": 3.5891, + "step": 3800 + }, + { + "epoch": 0.033999187551881874, + "grad_norm": 5.438379287719727, + "learning_rate": 9.999374818228895e-05, + "loss": 3.6557, + "step": 3850 + }, + { + "epoch": 0.034440735442166054, + "grad_norm": 1.6964021921157837, + "learning_rate": 9.999352650852778e-05, + "loss": 3.4464, + "step": 3900 + }, + { + "epoch": 0.03488228333245024, + "grad_norm": 1.759133219718933, + "learning_rate": 9.999330097319063e-05, + "loss": 3.5975, + "step": 3950 + }, + { + "epoch": 0.03532383122273442, + "grad_norm": 1.7858753204345703, + "learning_rate": 9.999307157629489e-05, + "loss": 3.8846, + "step": 4000 + }, + { + "epoch": 0.0357653791130186, + "grad_norm": 1.641852855682373, + "learning_rate": 9.999283831785833e-05, + "loss": 3.7919, + "step": 4050 + }, + { + "epoch": 0.03620692700330278, + "grad_norm": 1.8461246490478516, + "learning_rate": 9.99926011978989e-05, + "loss": 3.2686, + "step": 4100 + }, + { + "epoch": 0.03664847489358696, + "grad_norm": 2.8383119106292725, + "learning_rate": 9.999236021643498e-05, + "loss": 3.1711, + "step": 4150 + }, + { + "epoch": 0.03709002278387114, + "grad_norm": 1.094404935836792, + "learning_rate": 9.999211537348514e-05, + "loss": 3.4461, + "step": 4200 + }, + { + "epoch": 0.03753157067415532, + "grad_norm": 1.1627689599990845, + "learning_rate": 9.999186666906832e-05, + "loss": 3.2783, + "step": 4250 + }, + { + "epoch": 0.0379731185644395, + "grad_norm": 2.673081398010254, + "learning_rate": 9.999161410320373e-05, + "loss": 3.3066, + "step": 4300 + }, + { + "epoch": 0.03841466645472368, + "grad_norm": 2.3519914150238037, + "learning_rate": 9.999135767591083e-05, + "loss": 3.5882, + "step": 4350 + }, + { + "epoch": 0.03885621434500786, + "grad_norm": 2.861649990081787, + "learning_rate": 9.999109738720949e-05, + "loss": 3.3404, + "step": 4400 + }, + { + "epoch": 0.03929776223529204, + "grad_norm": 2.392594575881958, + "learning_rate": 9.999083323711979e-05, + "loss": 3.4555, + "step": 4450 + }, + { + "epoch": 0.03973931012557622, + "grad_norm": 2.1783764362335205, + "learning_rate": 9.999056522566213e-05, + "loss": 3.3032, + "step": 4500 + }, + { + "epoch": 0.0401808580158604, + "grad_norm": 3.0805599689483643, + "learning_rate": 9.99902933528572e-05, + "loss": 3.7245, + "step": 4550 + }, + { + "epoch": 0.04062240590614458, + "grad_norm": 0.8756254315376282, + "learning_rate": 9.999001761872601e-05, + "loss": 3.0947, + "step": 4600 + }, + { + "epoch": 0.041063953796428763, + "grad_norm": 8.237518310546875, + "learning_rate": 9.998973802328988e-05, + "loss": 3.369, + "step": 4650 + }, + { + "epoch": 0.041505501686712944, + "grad_norm": 1.5842547416687012, + "learning_rate": 9.998945456657039e-05, + "loss": 3.6075, + "step": 4700 + }, + { + "epoch": 0.041947049576997124, + "grad_norm": 2.602689504623413, + "learning_rate": 9.998916724858942e-05, + "loss": 3.7514, + "step": 4750 + }, + { + "epoch": 0.042388597467281304, + "grad_norm": 3.264051914215088, + "learning_rate": 9.998887606936918e-05, + "loss": 3.435, + "step": 4800 + }, + { + "epoch": 0.042830145357565484, + "grad_norm": 2.2757341861724854, + "learning_rate": 9.998858102893215e-05, + "loss": 3.6671, + "step": 4850 + }, + { + "epoch": 0.043271693247849664, + "grad_norm": 1.5097368955612183, + "learning_rate": 9.998828212730115e-05, + "loss": 3.4472, + "step": 4900 + }, + { + "epoch": 0.043713241138133845, + "grad_norm": 2.7084147930145264, + "learning_rate": 9.998797936449922e-05, + "loss": 3.8243, + "step": 4950 + }, + { + "epoch": 0.044154789028418025, + "grad_norm": 2.9833407402038574, + "learning_rate": 9.998767274054978e-05, + "loss": 3.7378, + "step": 5000 + }, + { + "epoch": 0.044596336918702205, + "grad_norm": 9.48874568939209, + "learning_rate": 9.998736225547651e-05, + "loss": 3.4267, + "step": 5050 + }, + { + "epoch": 0.045037884808986385, + "grad_norm": 1.7265108823776245, + "learning_rate": 9.998704790930337e-05, + "loss": 3.4905, + "step": 5100 + }, + { + "epoch": 0.045479432699270565, + "grad_norm": 1.690320372581482, + "learning_rate": 9.998672970205467e-05, + "loss": 3.4667, + "step": 5150 + }, + { + "epoch": 0.045920980589554745, + "grad_norm": 3.318742275238037, + "learning_rate": 9.998640763375497e-05, + "loss": 3.4528, + "step": 5200 + }, + { + "epoch": 0.046362528479838926, + "grad_norm": 3.2868552207946777, + "learning_rate": 9.998608170442915e-05, + "loss": 3.5877, + "step": 5250 + }, + { + "epoch": 0.046804076370123106, + "grad_norm": 3.8400142192840576, + "learning_rate": 9.99857519141024e-05, + "loss": 3.5318, + "step": 5300 + }, + { + "epoch": 0.047245624260407286, + "grad_norm": 1.9866262674331665, + "learning_rate": 9.998541826280018e-05, + "loss": 3.7204, + "step": 5350 + }, + { + "epoch": 0.047687172150691466, + "grad_norm": 2.3364500999450684, + "learning_rate": 9.998508075054826e-05, + "loss": 3.3647, + "step": 5400 + }, + { + "epoch": 0.048128720040975646, + "grad_norm": 3.461249589920044, + "learning_rate": 9.99847393773727e-05, + "loss": 3.4072, + "step": 5450 + }, + { + "epoch": 0.048570267931259826, + "grad_norm": 1.401358962059021, + "learning_rate": 9.99843941432999e-05, + "loss": 3.506, + "step": 5500 + }, + { + "epoch": 0.04901181582154401, + "grad_norm": 2.5606305599212646, + "learning_rate": 9.99840450483565e-05, + "loss": 3.3622, + "step": 5550 + }, + { + "epoch": 0.04945336371182819, + "grad_norm": 2.909435510635376, + "learning_rate": 9.998369209256947e-05, + "loss": 3.1674, + "step": 5600 + }, + { + "epoch": 0.04989491160211237, + "grad_norm": 1.5808318853378296, + "learning_rate": 9.998333527596607e-05, + "loss": 3.5161, + "step": 5650 + }, + { + "epoch": 0.05033645949239655, + "grad_norm": 1.6047215461730957, + "learning_rate": 9.998297459857387e-05, + "loss": 3.528, + "step": 5700 + }, + { + "epoch": 0.05077800738268073, + "grad_norm": 2.816723346710205, + "learning_rate": 9.998261006042074e-05, + "loss": 3.1358, + "step": 5750 + }, + { + "epoch": 0.05121955527296491, + "grad_norm": 1.9394768476486206, + "learning_rate": 9.99822416615348e-05, + "loss": 2.8468, + "step": 5800 + }, + { + "epoch": 0.05166110316324909, + "grad_norm": 7.701488971710205, + "learning_rate": 9.998186940194454e-05, + "loss": 3.5079, + "step": 5850 + }, + { + "epoch": 0.05210265105353327, + "grad_norm": 1.9361647367477417, + "learning_rate": 9.998150084191847e-05, + "loss": 3.1601, + "step": 5900 + }, + { + "epoch": 0.05254419894381745, + "grad_norm": 1.7290771007537842, + "learning_rate": 9.998112093821874e-05, + "loss": 3.5756, + "step": 5950 + }, + { + "epoch": 0.05298574683410163, + "grad_norm": 7.283807754516602, + "learning_rate": 9.998073717390126e-05, + "loss": 3.3956, + "step": 6000 + }, + { + "epoch": 0.05298574683410163, + "eval_asr_loss": 0.8455491743058053, + "eval_loss": 3.3733489513397217, + "eval_runtime": 20.0292, + "eval_samples_per_second": 38.344, + "eval_steps_per_second": 9.586, + "eval_tts_loss": 5.977993885297463, + "step": 6000 + }, + { + "epoch": 0.05342729472438581, + "grad_norm": 2.8105592727661133, + "learning_rate": 9.998034954899564e-05, + "loss": 3.5391, + "step": 6050 + }, + { + "epoch": 0.05386884261466999, + "grad_norm": 2.575618267059326, + "learning_rate": 9.997995806353187e-05, + "loss": 3.4848, + "step": 6100 + }, + { + "epoch": 0.05431039050495417, + "grad_norm": 1.944770336151123, + "learning_rate": 9.997956271754013e-05, + "loss": 3.4319, + "step": 6150 + }, + { + "epoch": 0.05475193839523835, + "grad_norm": 2.9254565238952637, + "learning_rate": 9.997916351105098e-05, + "loss": 3.5138, + "step": 6200 + }, + { + "epoch": 0.05519348628552253, + "grad_norm": 2.876880645751953, + "learning_rate": 9.997876044409529e-05, + "loss": 3.2436, + "step": 6250 + }, + { + "epoch": 0.05563503417580671, + "grad_norm": 1.5080561637878418, + "learning_rate": 9.997835351670415e-05, + "loss": 3.3394, + "step": 6300 + }, + { + "epoch": 0.05607658206609089, + "grad_norm": 2.99318790435791, + "learning_rate": 9.9977942728909e-05, + "loss": 3.2061, + "step": 6350 + }, + { + "epoch": 0.05651812995637507, + "grad_norm": 2.1895508766174316, + "learning_rate": 9.997752808074157e-05, + "loss": 3.586, + "step": 6400 + }, + { + "epoch": 0.05695967784665925, + "grad_norm": 2.089017868041992, + "learning_rate": 9.997710957223389e-05, + "loss": 3.6418, + "step": 6450 + }, + { + "epoch": 0.05740122573694343, + "grad_norm": 1.3540375232696533, + "learning_rate": 9.997668720341831e-05, + "loss": 3.5735, + "step": 6500 + }, + { + "epoch": 0.05784277362722761, + "grad_norm": 2.3487038612365723, + "learning_rate": 9.997626097432741e-05, + "loss": 3.4285, + "step": 6550 + }, + { + "epoch": 0.05828432151751179, + "grad_norm": 2.5517020225524902, + "learning_rate": 9.997583088499415e-05, + "loss": 3.5823, + "step": 6600 + }, + { + "epoch": 0.05872586940779597, + "grad_norm": 1.7537412643432617, + "learning_rate": 9.997539693545174e-05, + "loss": 3.1481, + "step": 6650 + }, + { + "epoch": 0.05916741729808015, + "grad_norm": 2.911041736602783, + "learning_rate": 9.99749591257337e-05, + "loss": 3.7261, + "step": 6700 + }, + { + "epoch": 0.05960896518836433, + "grad_norm": 7.761764049530029, + "learning_rate": 9.997451745587382e-05, + "loss": 3.8374, + "step": 6750 + }, + { + "epoch": 0.06005051307864851, + "grad_norm": 6.103822708129883, + "learning_rate": 9.997407192590625e-05, + "loss": 3.1474, + "step": 6800 + }, + { + "epoch": 0.06049206096893269, + "grad_norm": 8.443099021911621, + "learning_rate": 9.99736225358654e-05, + "loss": 3.4926, + "step": 6850 + }, + { + "epoch": 0.06093360885921687, + "grad_norm": 6.040070056915283, + "learning_rate": 9.997316928578595e-05, + "loss": 3.1745, + "step": 6900 + }, + { + "epoch": 0.06137515674950105, + "grad_norm": 1.7983872890472412, + "learning_rate": 9.997271217570296e-05, + "loss": 3.5779, + "step": 6950 + }, + { + "epoch": 0.06181670463978523, + "grad_norm": 2.881488800048828, + "learning_rate": 9.99722512056517e-05, + "loss": 3.6113, + "step": 7000 + }, + { + "epoch": 0.06225825253006941, + "grad_norm": 2.0535004138946533, + "learning_rate": 9.997178637566777e-05, + "loss": 3.5762, + "step": 7050 + }, + { + "epoch": 0.06269980042035359, + "grad_norm": 2.7159841060638428, + "learning_rate": 9.997131768578711e-05, + "loss": 3.4012, + "step": 7100 + }, + { + "epoch": 0.06314134831063777, + "grad_norm": 3.324490785598755, + "learning_rate": 9.997084513604591e-05, + "loss": 3.8183, + "step": 7150 + }, + { + "epoch": 0.06358289620092195, + "grad_norm": 1.6304800510406494, + "learning_rate": 9.997036872648064e-05, + "loss": 3.4161, + "step": 7200 + }, + { + "epoch": 0.06402444409120613, + "grad_norm": 2.0595145225524902, + "learning_rate": 9.996988845712813e-05, + "loss": 3.4944, + "step": 7250 + }, + { + "epoch": 0.0644659919814903, + "grad_norm": 2.9137229919433594, + "learning_rate": 9.996940432802548e-05, + "loss": 3.6446, + "step": 7300 + }, + { + "epoch": 0.06490753987177449, + "grad_norm": 2.6019785404205322, + "learning_rate": 9.996891633921007e-05, + "loss": 3.5755, + "step": 7350 + }, + { + "epoch": 0.06534908776205867, + "grad_norm": 1.3080384731292725, + "learning_rate": 9.99684244907196e-05, + "loss": 3.432, + "step": 7400 + }, + { + "epoch": 0.06579063565234285, + "grad_norm": 1.085084080696106, + "learning_rate": 9.996792878259204e-05, + "loss": 3.5939, + "step": 7450 + }, + { + "epoch": 0.06623218354262703, + "grad_norm": 2.416012763977051, + "learning_rate": 9.996742921486573e-05, + "loss": 3.3125, + "step": 7500 + }, + { + "epoch": 0.0666737314329112, + "grad_norm": 2.329331159591675, + "learning_rate": 9.99669257875792e-05, + "loss": 3.4727, + "step": 7550 + }, + { + "epoch": 0.06711527932319539, + "grad_norm": 1.2133623361587524, + "learning_rate": 9.996641850077135e-05, + "loss": 3.6817, + "step": 7600 + }, + { + "epoch": 0.06755682721347957, + "grad_norm": 3.447230577468872, + "learning_rate": 9.99659073544814e-05, + "loss": 3.3487, + "step": 7650 + }, + { + "epoch": 0.06799837510376375, + "grad_norm": 1.8407869338989258, + "learning_rate": 9.99653923487488e-05, + "loss": 3.4952, + "step": 7700 + }, + { + "epoch": 0.06843992299404793, + "grad_norm": 3.4017183780670166, + "learning_rate": 9.996487348361331e-05, + "loss": 3.5987, + "step": 7750 + }, + { + "epoch": 0.06888147088433211, + "grad_norm": 2.014235734939575, + "learning_rate": 9.996435075911507e-05, + "loss": 3.4913, + "step": 7800 + }, + { + "epoch": 0.0693230187746163, + "grad_norm": 8.113083839416504, + "learning_rate": 9.99638241752944e-05, + "loss": 3.2701, + "step": 7850 + }, + { + "epoch": 0.06976456666490048, + "grad_norm": 3.6530191898345947, + "learning_rate": 9.996329373219199e-05, + "loss": 3.7682, + "step": 7900 + }, + { + "epoch": 0.07020611455518466, + "grad_norm": 3.110872507095337, + "learning_rate": 9.996275942984882e-05, + "loss": 2.9624, + "step": 7950 + }, + { + "epoch": 0.07064766244546884, + "grad_norm": 5.682243347167969, + "learning_rate": 9.996222126830616e-05, + "loss": 3.1075, + "step": 8000 + }, + { + "epoch": 0.07108921033575302, + "grad_norm": 8.339673042297363, + "learning_rate": 9.996167924760556e-05, + "loss": 3.4832, + "step": 8050 + }, + { + "epoch": 0.0715307582260372, + "grad_norm": 2.9629604816436768, + "learning_rate": 9.996113336778891e-05, + "loss": 3.5399, + "step": 8100 + }, + { + "epoch": 0.07197230611632138, + "grad_norm": 4.072164058685303, + "learning_rate": 9.996058362889837e-05, + "loss": 3.3337, + "step": 8150 + }, + { + "epoch": 0.07241385400660556, + "grad_norm": 2.1256349086761475, + "learning_rate": 9.99600300309764e-05, + "loss": 3.4296, + "step": 8200 + }, + { + "epoch": 0.07285540189688974, + "grad_norm": 0.9664581418037415, + "learning_rate": 9.995947257406575e-05, + "loss": 3.3867, + "step": 8250 + }, + { + "epoch": 0.07329694978717392, + "grad_norm": 2.7981796264648438, + "learning_rate": 9.995891125820948e-05, + "loss": 3.5777, + "step": 8300 + }, + { + "epoch": 0.0737384976774581, + "grad_norm": 2.2892305850982666, + "learning_rate": 9.995834608345098e-05, + "loss": 3.448, + "step": 8350 + }, + { + "epoch": 0.07418004556774228, + "grad_norm": 3.462890386581421, + "learning_rate": 9.995777704983387e-05, + "loss": 3.6835, + "step": 8400 + }, + { + "epoch": 0.07462159345802646, + "grad_norm": 2.887977123260498, + "learning_rate": 9.995720415740209e-05, + "loss": 3.8102, + "step": 8450 + }, + { + "epoch": 0.07506314134831064, + "grad_norm": 1.103590726852417, + "learning_rate": 9.995662740619993e-05, + "loss": 3.3465, + "step": 8500 + }, + { + "epoch": 0.07550468923859482, + "grad_norm": 1.5264739990234375, + "learning_rate": 9.995604679627193e-05, + "loss": 3.3524, + "step": 8550 + }, + { + "epoch": 0.075946237128879, + "grad_norm": 10.508204460144043, + "learning_rate": 9.995546232766293e-05, + "loss": 2.9848, + "step": 8600 + }, + { + "epoch": 0.07638778501916318, + "grad_norm": 11.036320686340332, + "learning_rate": 9.995487400041806e-05, + "loss": 3.6104, + "step": 8650 + }, + { + "epoch": 0.07682933290944736, + "grad_norm": 0.900551974773407, + "learning_rate": 9.995428181458279e-05, + "loss": 3.2115, + "step": 8700 + }, + { + "epoch": 0.07727088079973154, + "grad_norm": 1.0776323080062866, + "learning_rate": 9.995368577020285e-05, + "loss": 3.378, + "step": 8750 + }, + { + "epoch": 0.07771242869001573, + "grad_norm": 1.9495457410812378, + "learning_rate": 9.995308586732427e-05, + "loss": 3.134, + "step": 8800 + }, + { + "epoch": 0.0781539765802999, + "grad_norm": 2.638998508453369, + "learning_rate": 9.99524821059934e-05, + "loss": 3.158, + "step": 8850 + }, + { + "epoch": 0.07859552447058409, + "grad_norm": 1.9697933197021484, + "learning_rate": 9.995187448625688e-05, + "loss": 3.4271, + "step": 8900 + }, + { + "epoch": 0.07903707236086827, + "grad_norm": 1.1985995769500732, + "learning_rate": 9.995126300816163e-05, + "loss": 3.3212, + "step": 8950 + }, + { + "epoch": 0.07947862025115245, + "grad_norm": 5.711977005004883, + "learning_rate": 9.995064767175488e-05, + "loss": 3.256, + "step": 9000 + }, + { + "epoch": 0.07947862025115245, + "eval_asr_loss": 0.8664082059474173, + "eval_loss": 3.2935588359832764, + "eval_runtime": 19.9624, + "eval_samples_per_second": 38.472, + "eval_steps_per_second": 9.618, + "eval_tts_loss": 5.950566485041761, + "step": 9000 + }, + { + "epoch": 0.07992016814143663, + "grad_norm": 2.0665061473846436, + "learning_rate": 9.995002847708418e-05, + "loss": 3.5921, + "step": 9050 + }, + { + "epoch": 0.0803617160317208, + "grad_norm": 2.0567665100097656, + "learning_rate": 9.994940542419733e-05, + "loss": 3.4312, + "step": 9100 + }, + { + "epoch": 0.08080326392200499, + "grad_norm": 5.3886847496032715, + "learning_rate": 9.994879108917332e-05, + "loss": 3.2038, + "step": 9150 + }, + { + "epoch": 0.08124481181228917, + "grad_norm": 5.97287130355835, + "learning_rate": 9.994816039716079e-05, + "loss": 3.4363, + "step": 9200 + }, + { + "epoch": 0.08168635970257335, + "grad_norm": 3.153740406036377, + "learning_rate": 9.994752584707642e-05, + "loss": 3.3041, + "step": 9250 + }, + { + "epoch": 0.08212790759285753, + "grad_norm": 2.116394519805908, + "learning_rate": 9.994688743896923e-05, + "loss": 3.906, + "step": 9300 + }, + { + "epoch": 0.08256945548314171, + "grad_norm": 2.2864694595336914, + "learning_rate": 9.994624517288851e-05, + "loss": 3.5515, + "step": 9350 + }, + { + "epoch": 0.08301100337342589, + "grad_norm": 6.8061723709106445, + "learning_rate": 9.994559904888388e-05, + "loss": 3.5991, + "step": 9400 + }, + { + "epoch": 0.08345255126371007, + "grad_norm": 5.035069465637207, + "learning_rate": 9.994494906700527e-05, + "loss": 3.7135, + "step": 9450 + }, + { + "epoch": 0.08389409915399425, + "grad_norm": 2.2650699615478516, + "learning_rate": 9.994429522730284e-05, + "loss": 3.3555, + "step": 9500 + }, + { + "epoch": 0.08433564704427843, + "grad_norm": 3.1491940021514893, + "learning_rate": 9.994363752982714e-05, + "loss": 3.0579, + "step": 9550 + }, + { + "epoch": 0.08477719493456261, + "grad_norm": 3.5358498096466064, + "learning_rate": 9.994297597462894e-05, + "loss": 3.6277, + "step": 9600 + }, + { + "epoch": 0.08521874282484679, + "grad_norm": 3.785895824432373, + "learning_rate": 9.994231056175936e-05, + "loss": 3.2452, + "step": 9650 + }, + { + "epoch": 0.08566029071513097, + "grad_norm": 1.632702350616455, + "learning_rate": 9.994164129126977e-05, + "loss": 3.1927, + "step": 9700 + }, + { + "epoch": 0.08610183860541515, + "grad_norm": 3.5743367671966553, + "learning_rate": 9.99409681632119e-05, + "loss": 3.6792, + "step": 9750 + }, + { + "epoch": 0.08654338649569933, + "grad_norm": 9.39389419555664, + "learning_rate": 9.994029117763773e-05, + "loss": 3.1351, + "step": 9800 + }, + { + "epoch": 0.08698493438598351, + "grad_norm": 2.9739062786102295, + "learning_rate": 9.993961033459953e-05, + "loss": 3.1706, + "step": 9850 + }, + { + "epoch": 0.08742648227626769, + "grad_norm": 3.320171356201172, + "learning_rate": 9.99389256341499e-05, + "loss": 3.6547, + "step": 9900 + }, + { + "epoch": 0.08786803016655187, + "grad_norm": 8.49543285369873, + "learning_rate": 9.993823707634176e-05, + "loss": 3.701, + "step": 9950 + }, + { + "epoch": 0.08830957805683605, + "grad_norm": 1.3553107976913452, + "learning_rate": 9.993754466122827e-05, + "loss": 3.4541, + "step": 10000 + }, + { + "epoch": 0.08875112594712023, + "grad_norm": 1.0776971578598022, + "learning_rate": 9.993684838886289e-05, + "loss": 3.7029, + "step": 10050 + }, + { + "epoch": 0.08919267383740441, + "grad_norm": 1.6787606477737427, + "learning_rate": 9.993614825929945e-05, + "loss": 3.6013, + "step": 10100 + }, + { + "epoch": 0.08963422172768859, + "grad_norm": 1.4465861320495605, + "learning_rate": 9.993544427259198e-05, + "loss": 3.2941, + "step": 10150 + }, + { + "epoch": 0.09007576961797277, + "grad_norm": 3.221482276916504, + "learning_rate": 9.99347364287949e-05, + "loss": 3.5483, + "step": 10200 + }, + { + "epoch": 0.09051731750825695, + "grad_norm": 1.0219640731811523, + "learning_rate": 9.993402472796284e-05, + "loss": 3.2682, + "step": 10250 + }, + { + "epoch": 0.09095886539854113, + "grad_norm": 4.225593090057373, + "learning_rate": 9.993330917015082e-05, + "loss": 3.4346, + "step": 10300 + }, + { + "epoch": 0.09140041328882531, + "grad_norm": 1.145766019821167, + "learning_rate": 9.993258975541408e-05, + "loss": 3.5205, + "step": 10350 + }, + { + "epoch": 0.09184196117910949, + "grad_norm": 4.837847709655762, + "learning_rate": 9.99318664838082e-05, + "loss": 3.3985, + "step": 10400 + }, + { + "epoch": 0.09228350906939367, + "grad_norm": 3.118101119995117, + "learning_rate": 9.993113935538903e-05, + "loss": 3.6283, + "step": 10450 + }, + { + "epoch": 0.09272505695967785, + "grad_norm": 3.6168124675750732, + "learning_rate": 9.993040837021277e-05, + "loss": 3.1753, + "step": 10500 + }, + { + "epoch": 0.09316660484996203, + "grad_norm": 3.866116523742676, + "learning_rate": 9.992967352833584e-05, + "loss": 3.1334, + "step": 10550 + }, + { + "epoch": 0.09360815274024621, + "grad_norm": 2.1199982166290283, + "learning_rate": 9.992893482981505e-05, + "loss": 3.3849, + "step": 10600 + }, + { + "epoch": 0.09404970063053039, + "grad_norm": 1.948744297027588, + "learning_rate": 9.99281922747074e-05, + "loss": 3.3414, + "step": 10650 + }, + { + "epoch": 0.09449124852081457, + "grad_norm": 5.387505054473877, + "learning_rate": 9.99274458630703e-05, + "loss": 3.6585, + "step": 10700 + }, + { + "epoch": 0.09493279641109875, + "grad_norm": 2.675678253173828, + "learning_rate": 9.992669559496136e-05, + "loss": 3.1715, + "step": 10750 + }, + { + "epoch": 0.09537434430138293, + "grad_norm": 1.5642513036727905, + "learning_rate": 9.992594147043856e-05, + "loss": 3.2574, + "step": 10800 + }, + { + "epoch": 0.09581589219166711, + "grad_norm": 2.7304491996765137, + "learning_rate": 9.992518348956014e-05, + "loss": 3.547, + "step": 10850 + }, + { + "epoch": 0.09625744008195129, + "grad_norm": 3.1260108947753906, + "learning_rate": 9.992442165238465e-05, + "loss": 3.6459, + "step": 10900 + }, + { + "epoch": 0.09669898797223547, + "grad_norm": 1.5745757818222046, + "learning_rate": 9.992365595897092e-05, + "loss": 3.0718, + "step": 10950 + }, + { + "epoch": 0.09714053586251965, + "grad_norm": 4.948554515838623, + "learning_rate": 9.992288640937812e-05, + "loss": 2.9318, + "step": 11000 + }, + { + "epoch": 0.09758208375280383, + "grad_norm": 3.8740906715393066, + "learning_rate": 9.992211300366568e-05, + "loss": 3.5334, + "step": 11050 + }, + { + "epoch": 0.09802363164308801, + "grad_norm": 4.66175651550293, + "learning_rate": 9.992133574189335e-05, + "loss": 3.4052, + "step": 11100 + }, + { + "epoch": 0.0984651795333722, + "grad_norm": 3.1660337448120117, + "learning_rate": 9.992055462412113e-05, + "loss": 3.4354, + "step": 11150 + }, + { + "epoch": 0.09890672742365637, + "grad_norm": 3.1641385555267334, + "learning_rate": 9.99197696504094e-05, + "loss": 3.4509, + "step": 11200 + }, + { + "epoch": 0.09934827531394055, + "grad_norm": 2.3828437328338623, + "learning_rate": 9.991898082081874e-05, + "loss": 3.1644, + "step": 11250 + }, + { + "epoch": 0.09978982320422473, + "grad_norm": 3.6782753467559814, + "learning_rate": 9.991818813541014e-05, + "loss": 3.6032, + "step": 11300 + }, + { + "epoch": 0.10023137109450891, + "grad_norm": 1.4960347414016724, + "learning_rate": 9.991739159424481e-05, + "loss": 3.3263, + "step": 11350 + }, + { + "epoch": 0.1006729189847931, + "grad_norm": 3.1267852783203125, + "learning_rate": 9.991659119738423e-05, + "loss": 3.4492, + "step": 11400 + }, + { + "epoch": 0.10111446687507727, + "grad_norm": 1.307042121887207, + "learning_rate": 9.991578694489028e-05, + "loss": 3.2172, + "step": 11450 + }, + { + "epoch": 0.10155601476536145, + "grad_norm": 2.0429134368896484, + "learning_rate": 9.991497883682506e-05, + "loss": 3.171, + "step": 11500 + }, + { + "epoch": 0.10199756265564563, + "grad_norm": 2.784212350845337, + "learning_rate": 9.991416687325101e-05, + "loss": 3.4447, + "step": 11550 + }, + { + "epoch": 0.10243911054592982, + "grad_norm": 6.423585414886475, + "learning_rate": 9.991335105423081e-05, + "loss": 3.2792, + "step": 11600 + }, + { + "epoch": 0.102880658436214, + "grad_norm": 2.206615924835205, + "learning_rate": 9.99125313798275e-05, + "loss": 3.2817, + "step": 11650 + }, + { + "epoch": 0.10332220632649818, + "grad_norm": 4.011673450469971, + "learning_rate": 9.991170785010438e-05, + "loss": 3.692, + "step": 11700 + }, + { + "epoch": 0.10376375421678236, + "grad_norm": 1.4347039461135864, + "learning_rate": 9.991088046512507e-05, + "loss": 3.3256, + "step": 11750 + }, + { + "epoch": 0.10420530210706654, + "grad_norm": 3.251509428024292, + "learning_rate": 9.991004922495348e-05, + "loss": 3.3607, + "step": 11800 + }, + { + "epoch": 0.10464684999735072, + "grad_norm": 5.043297290802002, + "learning_rate": 9.990921412965381e-05, + "loss": 3.0377, + "step": 11850 + }, + { + "epoch": 0.1050883978876349, + "grad_norm": 1.7166991233825684, + "learning_rate": 9.990837517929057e-05, + "loss": 3.2673, + "step": 11900 + }, + { + "epoch": 0.10552994577791908, + "grad_norm": 2.2065978050231934, + "learning_rate": 9.990753237392854e-05, + "loss": 3.5916, + "step": 11950 + }, + { + "epoch": 0.10597149366820326, + "grad_norm": 3.4584174156188965, + "learning_rate": 9.990668571363286e-05, + "loss": 3.1674, + "step": 12000 + }, + { + "epoch": 0.10597149366820326, + "eval_asr_loss": 0.8610385786988017, + "eval_loss": 3.250169515609741, + "eval_runtime": 20.2526, + "eval_samples_per_second": 37.921, + "eval_steps_per_second": 9.48, + "eval_tts_loss": 6.012793056516366, + "step": 12000 + }, + { + "epoch": 0.10641304155848744, + "grad_norm": 3.1279189586639404, + "learning_rate": 9.99058351984689e-05, + "loss": 3.5257, + "step": 12050 + }, + { + "epoch": 0.10685458944877162, + "grad_norm": 5.271905899047852, + "learning_rate": 9.990498082850234e-05, + "loss": 3.2174, + "step": 12100 + }, + { + "epoch": 0.1072961373390558, + "grad_norm": 2.162796974182129, + "learning_rate": 9.990412260379922e-05, + "loss": 3.4223, + "step": 12150 + }, + { + "epoch": 0.10773768522933998, + "grad_norm": 2.182065725326538, + "learning_rate": 9.990326052442579e-05, + "loss": 3.3126, + "step": 12200 + }, + { + "epoch": 0.10817923311962416, + "grad_norm": 1.2493520975112915, + "learning_rate": 9.990239459044866e-05, + "loss": 3.5304, + "step": 12250 + }, + { + "epoch": 0.10862078100990834, + "grad_norm": 2.064229726791382, + "learning_rate": 9.99015248019347e-05, + "loss": 3.4558, + "step": 12300 + }, + { + "epoch": 0.10906232890019252, + "grad_norm": 2.22754168510437, + "learning_rate": 9.99006511589511e-05, + "loss": 3.3677, + "step": 12350 + }, + { + "epoch": 0.1095038767904767, + "grad_norm": 4.160098075866699, + "learning_rate": 9.989977366156535e-05, + "loss": 3.5189, + "step": 12400 + }, + { + "epoch": 0.10994542468076088, + "grad_norm": 0.9621350169181824, + "learning_rate": 9.989889230984522e-05, + "loss": 3.7358, + "step": 12450 + }, + { + "epoch": 0.11038697257104506, + "grad_norm": 1.2985560894012451, + "learning_rate": 9.989800710385879e-05, + "loss": 3.6572, + "step": 12500 + }, + { + "epoch": 0.11082852046132924, + "grad_norm": 0.9708568453788757, + "learning_rate": 9.989711804367443e-05, + "loss": 3.2729, + "step": 12550 + }, + { + "epoch": 0.11127006835161342, + "grad_norm": 1.924854040145874, + "learning_rate": 9.989622512936083e-05, + "loss": 3.4972, + "step": 12600 + }, + { + "epoch": 0.1117116162418976, + "grad_norm": 4.869857311248779, + "learning_rate": 9.989532836098691e-05, + "loss": 3.5102, + "step": 12650 + }, + { + "epoch": 0.11215316413218178, + "grad_norm": 5.745826244354248, + "learning_rate": 9.9894427738622e-05, + "loss": 3.3836, + "step": 12700 + }, + { + "epoch": 0.11259471202246596, + "grad_norm": 5.497977256774902, + "learning_rate": 9.989352326233566e-05, + "loss": 3.3298, + "step": 12750 + }, + { + "epoch": 0.11303625991275014, + "grad_norm": 9.22269344329834, + "learning_rate": 9.98926149321977e-05, + "loss": 3.3542, + "step": 12800 + }, + { + "epoch": 0.11347780780303432, + "grad_norm": 2.255528211593628, + "learning_rate": 9.989172102972332e-05, + "loss": 3.8121, + "step": 12850 + }, + { + "epoch": 0.1139193556933185, + "grad_norm": 2.629471778869629, + "learning_rate": 9.98908050691665e-05, + "loss": 3.456, + "step": 12900 + }, + { + "epoch": 0.11436090358360268, + "grad_norm": 6.432432174682617, + "learning_rate": 9.988988525496805e-05, + "loss": 3.1076, + "step": 12950 + }, + { + "epoch": 0.11480245147388686, + "grad_norm": 1.9251166582107544, + "learning_rate": 9.988896158719903e-05, + "loss": 3.5083, + "step": 13000 + }, + { + "epoch": 0.11524399936417104, + "grad_norm": 1.9608491659164429, + "learning_rate": 9.988803406593077e-05, + "loss": 3.4977, + "step": 13050 + }, + { + "epoch": 0.11568554725445522, + "grad_norm": 2.201385259628296, + "learning_rate": 9.988710269123491e-05, + "loss": 3.3394, + "step": 13100 + }, + { + "epoch": 0.1161270951447394, + "grad_norm": 0.8397690653800964, + "learning_rate": 9.98861674631834e-05, + "loss": 3.0705, + "step": 13150 + }, + { + "epoch": 0.11656864303502358, + "grad_norm": 3.4315147399902344, + "learning_rate": 9.988522838184848e-05, + "loss": 2.9964, + "step": 13200 + }, + { + "epoch": 0.11701019092530776, + "grad_norm": 1.0664474964141846, + "learning_rate": 9.988428544730267e-05, + "loss": 3.5425, + "step": 13250 + }, + { + "epoch": 0.11745173881559194, + "grad_norm": 1.7559316158294678, + "learning_rate": 9.988333865961883e-05, + "loss": 3.4283, + "step": 13300 + }, + { + "epoch": 0.11789328670587612, + "grad_norm": 1.3742626905441284, + "learning_rate": 9.988238801887006e-05, + "loss": 3.0457, + "step": 13350 + }, + { + "epoch": 0.1183348345961603, + "grad_norm": 6.306196212768555, + "learning_rate": 9.988143352512982e-05, + "loss": 3.4985, + "step": 13400 + }, + { + "epoch": 0.11877638248644448, + "grad_norm": 0.7708596587181091, + "learning_rate": 9.98804751784718e-05, + "loss": 3.3379, + "step": 13450 + }, + { + "epoch": 0.11921793037672866, + "grad_norm": 7.315723419189453, + "learning_rate": 9.987951297897008e-05, + "loss": 3.4271, + "step": 13500 + }, + { + "epoch": 0.11965947826701284, + "grad_norm": 1.3316899538040161, + "learning_rate": 9.987854692669894e-05, + "loss": 3.362, + "step": 13550 + }, + { + "epoch": 0.12010102615729702, + "grad_norm": 2.5768868923187256, + "learning_rate": 9.9877577021733e-05, + "loss": 3.7057, + "step": 13600 + }, + { + "epoch": 0.1205425740475812, + "grad_norm": 1.0544745922088623, + "learning_rate": 9.987660326414718e-05, + "loss": 3.2666, + "step": 13650 + }, + { + "epoch": 0.12098412193786538, + "grad_norm": 1.6136611700057983, + "learning_rate": 9.98756256540167e-05, + "loss": 3.019, + "step": 13700 + }, + { + "epoch": 0.12142566982814956, + "grad_norm": 3.0674450397491455, + "learning_rate": 9.987464419141707e-05, + "loss": 3.3559, + "step": 13750 + }, + { + "epoch": 0.12186721771843374, + "grad_norm": 5.650421142578125, + "learning_rate": 9.987365887642412e-05, + "loss": 3.1113, + "step": 13800 + }, + { + "epoch": 0.12230876560871792, + "grad_norm": 6.18209981918335, + "learning_rate": 9.987266970911393e-05, + "loss": 3.6087, + "step": 13850 + }, + { + "epoch": 0.1227503134990021, + "grad_norm": 3.334939956665039, + "learning_rate": 9.98716766895629e-05, + "loss": 3.2844, + "step": 13900 + }, + { + "epoch": 0.12319186138928628, + "grad_norm": 1.9125257730484009, + "learning_rate": 9.987067981784774e-05, + "loss": 3.5561, + "step": 13950 + }, + { + "epoch": 0.12363340927957046, + "grad_norm": 2.013408899307251, + "learning_rate": 9.986967909404547e-05, + "loss": 3.7224, + "step": 14000 + }, + { + "epoch": 0.12407495716985464, + "grad_norm": 1.1568374633789062, + "learning_rate": 9.986867451823337e-05, + "loss": 3.348, + "step": 14050 + }, + { + "epoch": 0.12451650506013882, + "grad_norm": 2.9201159477233887, + "learning_rate": 9.986766609048904e-05, + "loss": 3.358, + "step": 14100 + }, + { + "epoch": 0.124958052950423, + "grad_norm": 6.0850982666015625, + "learning_rate": 9.986665381089038e-05, + "loss": 3.3947, + "step": 14150 + }, + { + "epoch": 0.12539960084070717, + "grad_norm": 2.2208523750305176, + "learning_rate": 9.986563767951555e-05, + "loss": 3.3521, + "step": 14200 + }, + { + "epoch": 0.12584114873099136, + "grad_norm": 3.7602078914642334, + "learning_rate": 9.986461769644306e-05, + "loss": 3.3009, + "step": 14250 + }, + { + "epoch": 0.12628269662127553, + "grad_norm": 5.159857749938965, + "learning_rate": 9.98635938617517e-05, + "loss": 3.0826, + "step": 14300 + }, + { + "epoch": 0.12672424451155972, + "grad_norm": 1.6318804025650024, + "learning_rate": 9.986256617552054e-05, + "loss": 3.1461, + "step": 14350 + }, + { + "epoch": 0.1271657924018439, + "grad_norm": 1.6564624309539795, + "learning_rate": 9.986153463782897e-05, + "loss": 3.6095, + "step": 14400 + }, + { + "epoch": 0.12760734029212809, + "grad_norm": 6.902859210968018, + "learning_rate": 9.986049924875666e-05, + "loss": 3.4435, + "step": 14450 + }, + { + "epoch": 0.12804888818241225, + "grad_norm": 2.513582229614258, + "learning_rate": 9.985946000838359e-05, + "loss": 3.6592, + "step": 14500 + }, + { + "epoch": 0.12849043607269645, + "grad_norm": 1.5714596509933472, + "learning_rate": 9.985841691679004e-05, + "loss": 3.3225, + "step": 14550 + }, + { + "epoch": 0.1289319839629806, + "grad_norm": 2.4705002307891846, + "learning_rate": 9.985736997405655e-05, + "loss": 3.3319, + "step": 14600 + }, + { + "epoch": 0.1293735318532648, + "grad_norm": 0.8616177439689636, + "learning_rate": 9.985631918026401e-05, + "loss": 3.1653, + "step": 14650 + }, + { + "epoch": 0.12981507974354897, + "grad_norm": 1.7353193759918213, + "learning_rate": 9.985526453549359e-05, + "loss": 3.4691, + "step": 14700 + }, + { + "epoch": 0.13025662763383317, + "grad_norm": 5.408633708953857, + "learning_rate": 9.985420603982673e-05, + "loss": 3.5129, + "step": 14750 + }, + { + "epoch": 0.13069817552411733, + "grad_norm": 1.2819982767105103, + "learning_rate": 9.985314369334523e-05, + "loss": 3.2823, + "step": 14800 + }, + { + "epoch": 0.13113972341440153, + "grad_norm": 8.26905632019043, + "learning_rate": 9.98520774961311e-05, + "loss": 3.2017, + "step": 14850 + }, + { + "epoch": 0.1315812713046857, + "grad_norm": 2.404160737991333, + "learning_rate": 9.985100744826674e-05, + "loss": 3.2397, + "step": 14900 + }, + { + "epoch": 0.1320228191949699, + "grad_norm": 1.9694509506225586, + "learning_rate": 9.984993354983477e-05, + "loss": 3.7057, + "step": 14950 + }, + { + "epoch": 0.13246436708525405, + "grad_norm": 7.179697036743164, + "learning_rate": 9.984885580091814e-05, + "loss": 2.8689, + "step": 15000 + }, + { + "epoch": 0.13246436708525405, + "eval_asr_loss": 0.8746035999883381, + "eval_loss": 3.1943347454071045, + "eval_runtime": 20.0975, + "eval_samples_per_second": 38.214, + "eval_steps_per_second": 9.553, + "eval_tts_loss": 5.963482838455088, + "step": 15000 + }, + { + "epoch": 0.13290591497553825, + "grad_norm": 1.963370680809021, + "learning_rate": 9.98477742016001e-05, + "loss": 3.0848, + "step": 15050 + }, + { + "epoch": 0.1333474628658224, + "grad_norm": 5.354642868041992, + "learning_rate": 9.984668875196421e-05, + "loss": 3.4422, + "step": 15100 + }, + { + "epoch": 0.1337890107561066, + "grad_norm": 0.9628197550773621, + "learning_rate": 9.98455994520943e-05, + "loss": 3.6956, + "step": 15150 + }, + { + "epoch": 0.13423055864639077, + "grad_norm": 1.9591059684753418, + "learning_rate": 9.984450630207451e-05, + "loss": 3.8912, + "step": 15200 + }, + { + "epoch": 0.13467210653667497, + "grad_norm": 2.4057066440582275, + "learning_rate": 9.984340930198927e-05, + "loss": 3.4458, + "step": 15250 + }, + { + "epoch": 0.13511365442695913, + "grad_norm": 1.3141722679138184, + "learning_rate": 9.984230845192336e-05, + "loss": 3.3671, + "step": 15300 + }, + { + "epoch": 0.13555520231724333, + "grad_norm": 3.438781261444092, + "learning_rate": 9.984120375196174e-05, + "loss": 3.2085, + "step": 15350 + }, + { + "epoch": 0.1359967502075275, + "grad_norm": 1.443537712097168, + "learning_rate": 9.984011741091279e-05, + "loss": 3.2912, + "step": 15400 + }, + { + "epoch": 0.1364382980978117, + "grad_norm": 2.5141725540161133, + "learning_rate": 9.983900508840976e-05, + "loss": 3.4251, + "step": 15450 + }, + { + "epoch": 0.13687984598809586, + "grad_norm": 3.5416128635406494, + "learning_rate": 9.98378889162662e-05, + "loss": 3.4953, + "step": 15500 + }, + { + "epoch": 0.13732139387838005, + "grad_norm": 3.750843048095703, + "learning_rate": 9.983676889456833e-05, + "loss": 3.6289, + "step": 15550 + }, + { + "epoch": 0.13776294176866422, + "grad_norm": 8.563606262207031, + "learning_rate": 9.983564502340267e-05, + "loss": 3.3268, + "step": 15600 + }, + { + "epoch": 0.1382044896589484, + "grad_norm": 3.2222371101379395, + "learning_rate": 9.983451730285603e-05, + "loss": 3.329, + "step": 15650 + }, + { + "epoch": 0.1386460375492326, + "grad_norm": 2.218827724456787, + "learning_rate": 9.983338573301552e-05, + "loss": 3.8973, + "step": 15700 + }, + { + "epoch": 0.13908758543951677, + "grad_norm": 1.1404446363449097, + "learning_rate": 9.983225031396852e-05, + "loss": 3.5931, + "step": 15750 + }, + { + "epoch": 0.13952913332980096, + "grad_norm": 0.8973853588104248, + "learning_rate": 9.983111104580276e-05, + "loss": 3.2535, + "step": 15800 + }, + { + "epoch": 0.13997068122008513, + "grad_norm": 8.366558074951172, + "learning_rate": 9.982996792860623e-05, + "loss": 3.0558, + "step": 15850 + }, + { + "epoch": 0.14041222911036932, + "grad_norm": 1.4844486713409424, + "learning_rate": 9.982882096246722e-05, + "loss": 3.4828, + "step": 15900 + }, + { + "epoch": 0.1408537770006535, + "grad_norm": 2.583129644393921, + "learning_rate": 9.982767014747432e-05, + "loss": 3.6149, + "step": 15950 + }, + { + "epoch": 0.14129532489093768, + "grad_norm": 3.2486109733581543, + "learning_rate": 9.982651548371644e-05, + "loss": 3.1331, + "step": 16000 + }, + { + "epoch": 0.14173687278122185, + "grad_norm": 2.9078361988067627, + "learning_rate": 9.982535697128275e-05, + "loss": 2.9418, + "step": 16050 + }, + { + "epoch": 0.14217842067150605, + "grad_norm": 1.986458659172058, + "learning_rate": 9.982419461026273e-05, + "loss": 3.4626, + "step": 16100 + }, + { + "epoch": 0.1426199685617902, + "grad_norm": 5.311156272888184, + "learning_rate": 9.982302840074617e-05, + "loss": 3.1937, + "step": 16150 + }, + { + "epoch": 0.1430615164520744, + "grad_norm": 1.4057707786560059, + "learning_rate": 9.982185834282319e-05, + "loss": 4.0044, + "step": 16200 + }, + { + "epoch": 0.14350306434235857, + "grad_norm": 1.7638394832611084, + "learning_rate": 9.98206844365841e-05, + "loss": 3.657, + "step": 16250 + }, + { + "epoch": 0.14394461223264277, + "grad_norm": 2.069039821624756, + "learning_rate": 9.98195066821196e-05, + "loss": 3.5534, + "step": 16300 + }, + { + "epoch": 0.14438616012292693, + "grad_norm": 7.656248092651367, + "learning_rate": 9.981832507952067e-05, + "loss": 3.1842, + "step": 16350 + }, + { + "epoch": 0.14482770801321113, + "grad_norm": 3.7678370475769043, + "learning_rate": 9.981713962887859e-05, + "loss": 3.515, + "step": 16400 + }, + { + "epoch": 0.1452692559034953, + "grad_norm": 2.7892682552337646, + "learning_rate": 9.98159503302849e-05, + "loss": 3.6017, + "step": 16450 + }, + { + "epoch": 0.1457108037937795, + "grad_norm": 1.75071120262146, + "learning_rate": 9.981475718383147e-05, + "loss": 3.6968, + "step": 16500 + }, + { + "epoch": 0.14615235168406365, + "grad_norm": 0.988599419593811, + "learning_rate": 9.981356018961047e-05, + "loss": 3.638, + "step": 16550 + }, + { + "epoch": 0.14659389957434785, + "grad_norm": 3.183884382247925, + "learning_rate": 9.981235934771436e-05, + "loss": 3.4087, + "step": 16600 + }, + { + "epoch": 0.147035447464632, + "grad_norm": 2.3756110668182373, + "learning_rate": 9.981115465823587e-05, + "loss": 3.493, + "step": 16650 + }, + { + "epoch": 0.1474769953549162, + "grad_norm": 1.6157623529434204, + "learning_rate": 9.980994612126807e-05, + "loss": 3.4459, + "step": 16700 + }, + { + "epoch": 0.14791854324520037, + "grad_norm": 3.4954352378845215, + "learning_rate": 9.980873373690431e-05, + "loss": 3.6551, + "step": 16750 + }, + { + "epoch": 0.14836009113548457, + "grad_norm": 2.129396915435791, + "learning_rate": 9.980751750523825e-05, + "loss": 3.5101, + "step": 16800 + }, + { + "epoch": 0.14880163902576873, + "grad_norm": 1.6230214834213257, + "learning_rate": 9.98062974263638e-05, + "loss": 3.3445, + "step": 16850 + }, + { + "epoch": 0.14924318691605293, + "grad_norm": 2.7362122535705566, + "learning_rate": 9.980507350037522e-05, + "loss": 3.4352, + "step": 16900 + }, + { + "epoch": 0.1496847348063371, + "grad_norm": 4.6118364334106445, + "learning_rate": 9.980384572736706e-05, + "loss": 2.9692, + "step": 16950 + }, + { + "epoch": 0.1501262826966213, + "grad_norm": 3.5687484741210938, + "learning_rate": 9.980261410743414e-05, + "loss": 3.669, + "step": 17000 + }, + { + "epoch": 0.15056783058690545, + "grad_norm": 1.4324193000793457, + "learning_rate": 9.980137864067158e-05, + "loss": 3.0393, + "step": 17050 + }, + { + "epoch": 0.15100937847718965, + "grad_norm": 2.652137517929077, + "learning_rate": 9.980013932717484e-05, + "loss": 3.6648, + "step": 17100 + }, + { + "epoch": 0.15145092636747381, + "grad_norm": 7.182887554168701, + "learning_rate": 9.97988961670396e-05, + "loss": 3.3025, + "step": 17150 + }, + { + "epoch": 0.151892474257758, + "grad_norm": 2.4190828800201416, + "learning_rate": 9.979764916036195e-05, + "loss": 3.5013, + "step": 17200 + }, + { + "epoch": 0.15233402214804218, + "grad_norm": 4.525496006011963, + "learning_rate": 9.979639830723816e-05, + "loss": 3.4685, + "step": 17250 + }, + { + "epoch": 0.15277557003832637, + "grad_norm": 1.242196798324585, + "learning_rate": 9.979514360776486e-05, + "loss": 3.3286, + "step": 17300 + }, + { + "epoch": 0.15321711792861054, + "grad_norm": 4.049784183502197, + "learning_rate": 9.979388506203897e-05, + "loss": 3.1067, + "step": 17350 + }, + { + "epoch": 0.15365866581889473, + "grad_norm": 2.2166428565979004, + "learning_rate": 9.97926226701577e-05, + "loss": 3.1167, + "step": 17400 + }, + { + "epoch": 0.1541002137091789, + "grad_norm": 1.4584290981292725, + "learning_rate": 9.979135643221856e-05, + "loss": 3.4869, + "step": 17450 + }, + { + "epoch": 0.1545417615994631, + "grad_norm": 2.357682228088379, + "learning_rate": 9.979008634831934e-05, + "loss": 3.3837, + "step": 17500 + }, + { + "epoch": 0.15498330948974726, + "grad_norm": 3.9010651111602783, + "learning_rate": 9.978881241855817e-05, + "loss": 3.0948, + "step": 17550 + }, + { + "epoch": 0.15542485738003145, + "grad_norm": 2.2975761890411377, + "learning_rate": 9.978753464303343e-05, + "loss": 3.487, + "step": 17600 + }, + { + "epoch": 0.15586640527031562, + "grad_norm": 1.2210371494293213, + "learning_rate": 9.978625302184383e-05, + "loss": 3.7763, + "step": 17650 + }, + { + "epoch": 0.1563079531605998, + "grad_norm": 1.8714485168457031, + "learning_rate": 9.978496755508836e-05, + "loss": 2.9216, + "step": 17700 + }, + { + "epoch": 0.15674950105088398, + "grad_norm": 3.4243338108062744, + "learning_rate": 9.978367824286629e-05, + "loss": 3.5642, + "step": 17750 + }, + { + "epoch": 0.15719104894116817, + "grad_norm": 1.6219075918197632, + "learning_rate": 9.978238508527727e-05, + "loss": 3.3517, + "step": 17800 + }, + { + "epoch": 0.15763259683145234, + "grad_norm": 0.7262994647026062, + "learning_rate": 9.97810880824211e-05, + "loss": 2.9789, + "step": 17850 + }, + { + "epoch": 0.15807414472173653, + "grad_norm": 2.3567357063293457, + "learning_rate": 9.977981328904049e-05, + "loss": 3.4182, + "step": 17900 + }, + { + "epoch": 0.1585156926120207, + "grad_norm": 2.0587501525878906, + "learning_rate": 9.977850867285131e-05, + "loss": 3.4566, + "step": 17950 + }, + { + "epoch": 0.1589572405023049, + "grad_norm": 1.3674755096435547, + "learning_rate": 9.977720021169445e-05, + "loss": 3.4426, + "step": 18000 + }, + { + "epoch": 0.1589572405023049, + "eval_asr_loss": 0.8684477998103262, + "eval_loss": 3.168827772140503, + "eval_runtime": 20.2426, + "eval_samples_per_second": 37.94, + "eval_steps_per_second": 9.485, + "eval_tts_loss": 5.975099919698394, + "step": 18000 + }, + { + "epoch": 0.15939878839258906, + "grad_norm": 1.7318178415298462, + "learning_rate": 9.977588790567097e-05, + "loss": 3.1039, + "step": 18050 + }, + { + "epoch": 0.15984033628287325, + "grad_norm": 4.160914897918701, + "learning_rate": 9.977457175488225e-05, + "loss": 3.4418, + "step": 18100 + }, + { + "epoch": 0.16028188417315742, + "grad_norm": 1.6405212879180908, + "learning_rate": 9.977325175942992e-05, + "loss": 3.537, + "step": 18150 + }, + { + "epoch": 0.1607234320634416, + "grad_norm": 3.2284727096557617, + "learning_rate": 9.977192791941596e-05, + "loss": 3.5699, + "step": 18200 + }, + { + "epoch": 0.16116497995372578, + "grad_norm": 6.879377841949463, + "learning_rate": 9.977060023494263e-05, + "loss": 3.3869, + "step": 18250 + }, + { + "epoch": 0.16160652784400997, + "grad_norm": 5.974034309387207, + "learning_rate": 9.976926870611247e-05, + "loss": 3.3198, + "step": 18300 + }, + { + "epoch": 0.16204807573429414, + "grad_norm": 1.0853605270385742, + "learning_rate": 9.976793333302834e-05, + "loss": 3.2399, + "step": 18350 + }, + { + "epoch": 0.16248962362457833, + "grad_norm": 2.7796192169189453, + "learning_rate": 9.976659411579337e-05, + "loss": 3.3726, + "step": 18400 + }, + { + "epoch": 0.1629311715148625, + "grad_norm": 2.083576202392578, + "learning_rate": 9.976525105451102e-05, + "loss": 3.3063, + "step": 18450 + }, + { + "epoch": 0.1633727194051467, + "grad_norm": 1.6082487106323242, + "learning_rate": 9.976390414928501e-05, + "loss": 3.2636, + "step": 18500 + }, + { + "epoch": 0.16381426729543086, + "grad_norm": 2.3426663875579834, + "learning_rate": 9.97625534002194e-05, + "loss": 3.4482, + "step": 18550 + }, + { + "epoch": 0.16425581518571505, + "grad_norm": 4.199110507965088, + "learning_rate": 9.976119880741852e-05, + "loss": 3.5328, + "step": 18600 + }, + { + "epoch": 0.16469736307599922, + "grad_norm": 2.411940097808838, + "learning_rate": 9.975984037098698e-05, + "loss": 3.1049, + "step": 18650 + }, + { + "epoch": 0.16513891096628341, + "grad_norm": 8.064169883728027, + "learning_rate": 9.975847809102974e-05, + "loss": 3.2858, + "step": 18700 + }, + { + "epoch": 0.16558045885656758, + "grad_norm": 2.0680835247039795, + "learning_rate": 9.9757111967652e-05, + "loss": 3.1414, + "step": 18750 + }, + { + "epoch": 0.16602200674685177, + "grad_norm": 2.7678062915802, + "learning_rate": 9.975574200095927e-05, + "loss": 3.1492, + "step": 18800 + }, + { + "epoch": 0.16646355463713594, + "grad_norm": 3.402519941329956, + "learning_rate": 9.975436819105742e-05, + "loss": 3.1184, + "step": 18850 + }, + { + "epoch": 0.16690510252742013, + "grad_norm": 1.3270809650421143, + "learning_rate": 9.97529905380525e-05, + "loss": 3.4959, + "step": 18900 + }, + { + "epoch": 0.1673466504177043, + "grad_norm": 7.308635711669922, + "learning_rate": 9.975160904205098e-05, + "loss": 3.4175, + "step": 18950 + }, + { + "epoch": 0.1677881983079885, + "grad_norm": 1.5945827960968018, + "learning_rate": 9.975022370315952e-05, + "loss": 3.2295, + "step": 19000 + }, + { + "epoch": 0.16822974619827266, + "grad_norm": 4.1863861083984375, + "learning_rate": 9.974883452148517e-05, + "loss": 3.1561, + "step": 19050 + }, + { + "epoch": 0.16867129408855686, + "grad_norm": 1.7289392948150635, + "learning_rate": 9.97474414971352e-05, + "loss": 3.5214, + "step": 19100 + }, + { + "epoch": 0.16911284197884102, + "grad_norm": 1.1129475831985474, + "learning_rate": 9.974604463021722e-05, + "loss": 2.9315, + "step": 19150 + }, + { + "epoch": 0.16955438986912522, + "grad_norm": 1.6852185726165771, + "learning_rate": 9.974464392083913e-05, + "loss": 3.4969, + "step": 19200 + }, + { + "epoch": 0.16999593775940938, + "grad_norm": 6.104635238647461, + "learning_rate": 9.974323936910913e-05, + "loss": 3.6542, + "step": 19250 + }, + { + "epoch": 0.17043748564969358, + "grad_norm": 2.2778172492980957, + "learning_rate": 9.974183097513567e-05, + "loss": 3.1306, + "step": 19300 + }, + { + "epoch": 0.17087903353997774, + "grad_norm": 1.4735571146011353, + "learning_rate": 9.974041873902757e-05, + "loss": 3.654, + "step": 19350 + }, + { + "epoch": 0.17132058143026194, + "grad_norm": 3.6428987979888916, + "learning_rate": 9.973900266089393e-05, + "loss": 3.7674, + "step": 19400 + }, + { + "epoch": 0.1717621293205461, + "grad_norm": 3.5552337169647217, + "learning_rate": 9.97375827408441e-05, + "loss": 3.5145, + "step": 19450 + }, + { + "epoch": 0.1722036772108303, + "grad_norm": 3.519054651260376, + "learning_rate": 9.973615897898777e-05, + "loss": 3.1466, + "step": 19500 + }, + { + "epoch": 0.17264522510111446, + "grad_norm": 5.785290241241455, + "learning_rate": 9.97347313754349e-05, + "loss": 3.207, + "step": 19550 + }, + { + "epoch": 0.17308677299139866, + "grad_norm": 1.1869070529937744, + "learning_rate": 9.973329993029577e-05, + "loss": 3.4333, + "step": 19600 + }, + { + "epoch": 0.17352832088168282, + "grad_norm": 1.620424747467041, + "learning_rate": 9.973186464368096e-05, + "loss": 3.4715, + "step": 19650 + }, + { + "epoch": 0.17396986877196702, + "grad_norm": 3.2140984535217285, + "learning_rate": 9.973042551570131e-05, + "loss": 3.5503, + "step": 19700 + }, + { + "epoch": 0.17441141666225118, + "grad_norm": 8.555562973022461, + "learning_rate": 9.972898254646799e-05, + "loss": 3.9534, + "step": 19750 + }, + { + "epoch": 0.17485296455253538, + "grad_norm": 2.755021095275879, + "learning_rate": 9.972753573609245e-05, + "loss": 3.2534, + "step": 19800 + }, + { + "epoch": 0.17529451244281954, + "grad_norm": 2.5277762413024902, + "learning_rate": 9.972608508468646e-05, + "loss": 3.9156, + "step": 19850 + }, + { + "epoch": 0.17573606033310374, + "grad_norm": 6.709779739379883, + "learning_rate": 9.972463059236207e-05, + "loss": 3.3525, + "step": 19900 + }, + { + "epoch": 0.1761776082233879, + "grad_norm": 2.41070294380188, + "learning_rate": 9.972317225923161e-05, + "loss": 3.3441, + "step": 19950 + }, + { + "epoch": 0.1766191561136721, + "grad_norm": 1.5217126607894897, + "learning_rate": 9.972171008540774e-05, + "loss": 3.2966, + "step": 20000 + }, + { + "epoch": 0.17706070400395627, + "grad_norm": 5.1036763191223145, + "learning_rate": 9.972024407100337e-05, + "loss": 3.1964, + "step": 20050 + }, + { + "epoch": 0.17750225189424046, + "grad_norm": 1.291801929473877, + "learning_rate": 9.97187742161318e-05, + "loss": 3.6654, + "step": 20100 + }, + { + "epoch": 0.17794379978452463, + "grad_norm": 1.6413884162902832, + "learning_rate": 9.97173005209065e-05, + "loss": 3.5594, + "step": 20150 + }, + { + "epoch": 0.17838534767480882, + "grad_norm": 2.1488759517669678, + "learning_rate": 9.971582298544133e-05, + "loss": 3.1799, + "step": 20200 + }, + { + "epoch": 0.17882689556509299, + "grad_norm": 4.286724090576172, + "learning_rate": 9.971434160985041e-05, + "loss": 3.1923, + "step": 20250 + }, + { + "epoch": 0.17926844345537718, + "grad_norm": 2.672957420349121, + "learning_rate": 9.971285639424817e-05, + "loss": 3.216, + "step": 20300 + }, + { + "epoch": 0.17970999134566135, + "grad_norm": 1.8457285165786743, + "learning_rate": 9.971136733874931e-05, + "loss": 3.0968, + "step": 20350 + }, + { + "epoch": 0.18015153923594554, + "grad_norm": 2.780160665512085, + "learning_rate": 9.970987444346889e-05, + "loss": 3.7793, + "step": 20400 + }, + { + "epoch": 0.1805930871262297, + "grad_norm": 4.923703670501709, + "learning_rate": 9.970837770852218e-05, + "loss": 3.0386, + "step": 20450 + }, + { + "epoch": 0.1810346350165139, + "grad_norm": 6.101312637329102, + "learning_rate": 9.97068771340248e-05, + "loss": 3.6275, + "step": 20500 + }, + { + "epoch": 0.18147618290679807, + "grad_norm": 1.9402213096618652, + "learning_rate": 9.970537272009269e-05, + "loss": 3.0918, + "step": 20550 + }, + { + "epoch": 0.18191773079708226, + "grad_norm": 0.9089447259902954, + "learning_rate": 9.970386446684201e-05, + "loss": 3.5084, + "step": 20600 + }, + { + "epoch": 0.18235927868736643, + "grad_norm": 0.5540192723274231, + "learning_rate": 9.970235237438927e-05, + "loss": 3.3147, + "step": 20650 + }, + { + "epoch": 0.18280082657765062, + "grad_norm": 3.3821587562561035, + "learning_rate": 9.970083644285129e-05, + "loss": 3.7705, + "step": 20700 + }, + { + "epoch": 0.1832423744679348, + "grad_norm": 2.602141857147217, + "learning_rate": 9.969931667234512e-05, + "loss": 3.231, + "step": 20750 + }, + { + "epoch": 0.18368392235821898, + "grad_norm": 2.034205198287964, + "learning_rate": 9.969779306298818e-05, + "loss": 3.5487, + "step": 20800 + }, + { + "epoch": 0.18412547024850315, + "grad_norm": 2.844647169113159, + "learning_rate": 9.969626561489817e-05, + "loss": 3.5581, + "step": 20850 + }, + { + "epoch": 0.18456701813878734, + "grad_norm": 1.6377545595169067, + "learning_rate": 9.969473432819305e-05, + "loss": 3.8547, + "step": 20900 + }, + { + "epoch": 0.1850085660290715, + "grad_norm": 7.625864505767822, + "learning_rate": 9.96931992029911e-05, + "loss": 3.1848, + "step": 20950 + }, + { + "epoch": 0.1854501139193557, + "grad_norm": 1.9800269603729248, + "learning_rate": 9.969166023941089e-05, + "loss": 3.1796, + "step": 21000 + }, + { + "epoch": 0.1854501139193557, + "eval_asr_loss": 0.8812751858267902, + "eval_loss": 3.14119029045105, + "eval_runtime": 19.8885, + "eval_samples_per_second": 38.615, + "eval_steps_per_second": 9.654, + "eval_tts_loss": 5.95043236252646, + "step": 21000 + }, + { + "epoch": 0.18589166180963987, + "grad_norm": 0.8084915280342102, + "learning_rate": 9.969011743757131e-05, + "loss": 3.1867, + "step": 21050 + }, + { + "epoch": 0.18633320969992406, + "grad_norm": 4.80615234375, + "learning_rate": 9.968857079759151e-05, + "loss": 3.3244, + "step": 21100 + }, + { + "epoch": 0.18677475759020823, + "grad_norm": 1.4828674793243408, + "learning_rate": 9.968702031959096e-05, + "loss": 2.8393, + "step": 21150 + }, + { + "epoch": 0.18721630548049242, + "grad_norm": 3.2170047760009766, + "learning_rate": 9.968546600368942e-05, + "loss": 3.3566, + "step": 21200 + }, + { + "epoch": 0.1876578533707766, + "grad_norm": 2.004711151123047, + "learning_rate": 9.968390785000696e-05, + "loss": 3.5387, + "step": 21250 + }, + { + "epoch": 0.18809940126106078, + "grad_norm": 2.858024835586548, + "learning_rate": 9.968234585866393e-05, + "loss": 3.4617, + "step": 21300 + }, + { + "epoch": 0.18854094915134495, + "grad_norm": 2.3722898960113525, + "learning_rate": 9.968078002978097e-05, + "loss": 3.3284, + "step": 21350 + }, + { + "epoch": 0.18898249704162914, + "grad_norm": 3.7834815979003906, + "learning_rate": 9.967921036347904e-05, + "loss": 3.2961, + "step": 21400 + }, + { + "epoch": 0.1894240449319133, + "grad_norm": 1.9320305585861206, + "learning_rate": 9.967763685987937e-05, + "loss": 3.0897, + "step": 21450 + }, + { + "epoch": 0.1898655928221975, + "grad_norm": 2.595550537109375, + "learning_rate": 9.96760595191035e-05, + "loss": 3.1881, + "step": 21500 + }, + { + "epoch": 0.19030714071248167, + "grad_norm": 2.7209346294403076, + "learning_rate": 9.967447834127328e-05, + "loss": 3.9049, + "step": 21550 + }, + { + "epoch": 0.19074868860276586, + "grad_norm": 2.9117844104766846, + "learning_rate": 9.967289332651085e-05, + "loss": 3.7096, + "step": 21600 + }, + { + "epoch": 0.19119023649305003, + "grad_norm": 3.910952091217041, + "learning_rate": 9.96713044749386e-05, + "loss": 3.2424, + "step": 21650 + }, + { + "epoch": 0.19163178438333422, + "grad_norm": 0.9400450587272644, + "learning_rate": 9.96697117866793e-05, + "loss": 2.9279, + "step": 21700 + }, + { + "epoch": 0.1920733322736184, + "grad_norm": 3.050595760345459, + "learning_rate": 9.966811526185594e-05, + "loss": 3.4662, + "step": 21750 + }, + { + "epoch": 0.19251488016390259, + "grad_norm": 4.441404819488525, + "learning_rate": 9.966651490059184e-05, + "loss": 3.0896, + "step": 21800 + }, + { + "epoch": 0.19295642805418675, + "grad_norm": 3.2662458419799805, + "learning_rate": 9.966491070301064e-05, + "loss": 3.1643, + "step": 21850 + }, + { + "epoch": 0.19339797594447095, + "grad_norm": 1.5034865140914917, + "learning_rate": 9.966330266923623e-05, + "loss": 3.1728, + "step": 21900 + }, + { + "epoch": 0.1938395238347551, + "grad_norm": 1.3567650318145752, + "learning_rate": 9.966169079939282e-05, + "loss": 3.433, + "step": 21950 + }, + { + "epoch": 0.1942810717250393, + "grad_norm": 2.380711793899536, + "learning_rate": 9.966007509360492e-05, + "loss": 3.3513, + "step": 22000 + }, + { + "epoch": 0.19472261961532347, + "grad_norm": 1.819464921951294, + "learning_rate": 9.965845555199732e-05, + "loss": 3.0924, + "step": 22050 + }, + { + "epoch": 0.19516416750560767, + "grad_norm": 1.84955894947052, + "learning_rate": 9.965683217469513e-05, + "loss": 3.3711, + "step": 22100 + }, + { + "epoch": 0.19560571539589183, + "grad_norm": 2.3455941677093506, + "learning_rate": 9.965520496182371e-05, + "loss": 3.5974, + "step": 22150 + }, + { + "epoch": 0.19604726328617603, + "grad_norm": 1.5657644271850586, + "learning_rate": 9.965357391350877e-05, + "loss": 3.5424, + "step": 22200 + }, + { + "epoch": 0.1964888111764602, + "grad_norm": 2.460850477218628, + "learning_rate": 9.965193902987632e-05, + "loss": 3.4851, + "step": 22250 + }, + { + "epoch": 0.1969303590667444, + "grad_norm": 4.1430768966674805, + "learning_rate": 9.96503003110526e-05, + "loss": 3.6162, + "step": 22300 + }, + { + "epoch": 0.19737190695702855, + "grad_norm": 1.511238932609558, + "learning_rate": 9.964865775716421e-05, + "loss": 3.4041, + "step": 22350 + }, + { + "epoch": 0.19781345484731275, + "grad_norm": 5.789384841918945, + "learning_rate": 9.964701136833801e-05, + "loss": 3.6231, + "step": 22400 + }, + { + "epoch": 0.1982550027375969, + "grad_norm": 1.697780728340149, + "learning_rate": 9.964536114470118e-05, + "loss": 3.5089, + "step": 22450 + }, + { + "epoch": 0.1986965506278811, + "grad_norm": 1.9673670530319214, + "learning_rate": 9.964370708638118e-05, + "loss": 3.1656, + "step": 22500 + }, + { + "epoch": 0.19913809851816527, + "grad_norm": 1.157395601272583, + "learning_rate": 9.964204919350577e-05, + "loss": 3.5386, + "step": 22550 + }, + { + "epoch": 0.19957964640844947, + "grad_norm": 1.872974157333374, + "learning_rate": 9.964038746620303e-05, + "loss": 3.8252, + "step": 22600 + }, + { + "epoch": 0.20002119429873363, + "grad_norm": 2.0033257007598877, + "learning_rate": 9.963872190460128e-05, + "loss": 3.0679, + "step": 22650 + }, + { + "epoch": 0.20046274218901783, + "grad_norm": 1.9374537467956543, + "learning_rate": 9.963705250882919e-05, + "loss": 3.4905, + "step": 22700 + }, + { + "epoch": 0.200904290079302, + "grad_norm": 1.368355393409729, + "learning_rate": 9.963537927901571e-05, + "loss": 3.1261, + "step": 22750 + }, + { + "epoch": 0.2013458379695862, + "grad_norm": 2.8965837955474854, + "learning_rate": 9.963370221529007e-05, + "loss": 3.434, + "step": 22800 + }, + { + "epoch": 0.20178738585987036, + "grad_norm": 4.269700527191162, + "learning_rate": 9.96320213177818e-05, + "loss": 3.3089, + "step": 22850 + }, + { + "epoch": 0.20222893375015455, + "grad_norm": 2.0559756755828857, + "learning_rate": 9.963033658662078e-05, + "loss": 3.0824, + "step": 22900 + }, + { + "epoch": 0.20267048164043872, + "grad_norm": 2.692986011505127, + "learning_rate": 9.962868183079845e-05, + "loss": 2.9178, + "step": 22950 + }, + { + "epoch": 0.2031120295307229, + "grad_norm": 2.5358245372772217, + "learning_rate": 9.962698950938911e-05, + "loss": 3.8775, + "step": 23000 + }, + { + "epoch": 0.20355357742100708, + "grad_norm": 2.653918504714966, + "learning_rate": 9.962529335471565e-05, + "loss": 3.9464, + "step": 23050 + }, + { + "epoch": 0.20399512531129127, + "grad_norm": 2.1768572330474854, + "learning_rate": 9.962362740422908e-05, + "loss": 3.3436, + "step": 23100 + }, + { + "epoch": 0.20443667320157544, + "grad_norm": 1.4818288087844849, + "learning_rate": 9.962192366007948e-05, + "loss": 3.3089, + "step": 23150 + }, + { + "epoch": 0.20487822109185963, + "grad_norm": 8.97103500366211, + "learning_rate": 9.962021608305707e-05, + "loss": 3.2528, + "step": 23200 + }, + { + "epoch": 0.2053197689821438, + "grad_norm": 2.6199610233306885, + "learning_rate": 9.961850467329372e-05, + "loss": 3.3929, + "step": 23250 + }, + { + "epoch": 0.205761316872428, + "grad_norm": 1.7377407550811768, + "learning_rate": 9.961678943092164e-05, + "loss": 3.4344, + "step": 23300 + }, + { + "epoch": 0.20620286476271216, + "grad_norm": 1.5331953763961792, + "learning_rate": 9.961507035607332e-05, + "loss": 3.6088, + "step": 23350 + }, + { + "epoch": 0.20664441265299635, + "grad_norm": 5.9846014976501465, + "learning_rate": 9.961334744888154e-05, + "loss": 3.5139, + "step": 23400 + }, + { + "epoch": 0.20708596054328052, + "grad_norm": 1.6636093854904175, + "learning_rate": 9.961162070947936e-05, + "loss": 3.783, + "step": 23450 + }, + { + "epoch": 0.2075275084335647, + "grad_norm": 4.730373382568359, + "learning_rate": 9.960989013800019e-05, + "loss": 3.5666, + "step": 23500 + }, + { + "epoch": 0.20796905632384888, + "grad_norm": 1.7042059898376465, + "learning_rate": 9.960815573457769e-05, + "loss": 3.4093, + "step": 23550 + }, + { + "epoch": 0.20841060421413307, + "grad_norm": 1.6063077449798584, + "learning_rate": 9.96064174993458e-05, + "loss": 2.9415, + "step": 23600 + }, + { + "epoch": 0.20885215210441724, + "grad_norm": 2.5845298767089844, + "learning_rate": 9.960467543243884e-05, + "loss": 3.421, + "step": 23650 + }, + { + "epoch": 0.20929369999470143, + "grad_norm": 2.843557119369507, + "learning_rate": 9.960292953399131e-05, + "loss": 3.7182, + "step": 23700 + }, + { + "epoch": 0.2097352478849856, + "grad_norm": 2.7299513816833496, + "learning_rate": 9.960117980413811e-05, + "loss": 3.4932, + "step": 23750 + }, + { + "epoch": 0.2101767957752698, + "grad_norm": 3.5709776878356934, + "learning_rate": 9.959942624301437e-05, + "loss": 3.4617, + "step": 23800 + }, + { + "epoch": 0.21061834366555396, + "grad_norm": 1.636993408203125, + "learning_rate": 9.959766885075556e-05, + "loss": 3.2784, + "step": 23850 + }, + { + "epoch": 0.21105989155583815, + "grad_norm": 1.2677909135818481, + "learning_rate": 9.959590762749738e-05, + "loss": 3.5762, + "step": 23900 + }, + { + "epoch": 0.21150143944612232, + "grad_norm": 4.043791770935059, + "learning_rate": 9.959414257337591e-05, + "loss": 3.1917, + "step": 23950 + }, + { + "epoch": 0.2119429873364065, + "grad_norm": 3.0787148475646973, + "learning_rate": 9.959237368852747e-05, + "loss": 3.307, + "step": 24000 + }, + { + "epoch": 0.2119429873364065, + "eval_asr_loss": 0.8739951278052379, + "eval_loss": 3.1280133724212646, + "eval_runtime": 20.4078, + "eval_samples_per_second": 37.633, + "eval_steps_per_second": 9.408, + "eval_tts_loss": 5.9267985776675065, + "step": 24000 + }, + { + "epoch": 0.21238453522669068, + "grad_norm": 2.54976487159729, + "learning_rate": 9.95906009730887e-05, + "loss": 3.178, + "step": 24050 + }, + { + "epoch": 0.21282608311697487, + "grad_norm": 1.0521984100341797, + "learning_rate": 9.958882442719652e-05, + "loss": 3.7995, + "step": 24100 + }, + { + "epoch": 0.21326763100725904, + "grad_norm": 1.708702564239502, + "learning_rate": 9.958704405098815e-05, + "loss": 3.5033, + "step": 24150 + }, + { + "epoch": 0.21370917889754323, + "grad_norm": 2.1780407428741455, + "learning_rate": 9.958525984460112e-05, + "loss": 3.4977, + "step": 24200 + }, + { + "epoch": 0.2141507267878274, + "grad_norm": 2.227015972137451, + "learning_rate": 9.958347180817323e-05, + "loss": 2.9502, + "step": 24250 + }, + { + "epoch": 0.2145922746781116, + "grad_norm": 1.5458627939224243, + "learning_rate": 9.95816799418426e-05, + "loss": 3.3266, + "step": 24300 + }, + { + "epoch": 0.21503382256839576, + "grad_norm": 1.8348188400268555, + "learning_rate": 9.957988424574764e-05, + "loss": 3.3448, + "step": 24350 + }, + { + "epoch": 0.21547537045867995, + "grad_norm": 2.4458820819854736, + "learning_rate": 9.957808472002704e-05, + "loss": 3.9597, + "step": 24400 + }, + { + "epoch": 0.21591691834896412, + "grad_norm": 1.2153455018997192, + "learning_rate": 9.957628136481981e-05, + "loss": 3.0095, + "step": 24450 + }, + { + "epoch": 0.21635846623924831, + "grad_norm": 1.6723811626434326, + "learning_rate": 9.957447418026525e-05, + "loss": 3.3562, + "step": 24500 + }, + { + "epoch": 0.21680001412953248, + "grad_norm": 4.406776428222656, + "learning_rate": 9.957266316650291e-05, + "loss": 3.4466, + "step": 24550 + }, + { + "epoch": 0.21724156201981668, + "grad_norm": 5.27102518081665, + "learning_rate": 9.957084832367273e-05, + "loss": 3.3489, + "step": 24600 + }, + { + "epoch": 0.21768310991010084, + "grad_norm": 2.6601903438568115, + "learning_rate": 9.956902965191486e-05, + "loss": 3.3423, + "step": 24650 + }, + { + "epoch": 0.21812465780038504, + "grad_norm": 1.5734961032867432, + "learning_rate": 9.956720715136977e-05, + "loss": 3.3207, + "step": 24700 + }, + { + "epoch": 0.2185662056906692, + "grad_norm": 2.8826043605804443, + "learning_rate": 9.956538082217826e-05, + "loss": 3.1197, + "step": 24750 + }, + { + "epoch": 0.2190077535809534, + "grad_norm": 1.6438381671905518, + "learning_rate": 9.956355066448138e-05, + "loss": 3.1748, + "step": 24800 + }, + { + "epoch": 0.21944930147123756, + "grad_norm": 1.6108850240707397, + "learning_rate": 9.956171667842048e-05, + "loss": 3.7212, + "step": 24850 + }, + { + "epoch": 0.21989084936152176, + "grad_norm": 1.1622767448425293, + "learning_rate": 9.955987886413725e-05, + "loss": 3.553, + "step": 24900 + }, + { + "epoch": 0.22033239725180592, + "grad_norm": 1.478102207183838, + "learning_rate": 9.955803722177363e-05, + "loss": 3.2171, + "step": 24950 + }, + { + "epoch": 0.22077394514209012, + "grad_norm": 2.2040581703186035, + "learning_rate": 9.95561917514719e-05, + "loss": 3.4424, + "step": 25000 + }, + { + "epoch": 0.22121549303237428, + "grad_norm": 1.8141365051269531, + "learning_rate": 9.955434245337457e-05, + "loss": 3.6843, + "step": 25050 + }, + { + "epoch": 0.22165704092265848, + "grad_norm": 1.6195813417434692, + "learning_rate": 9.955248932762447e-05, + "loss": 3.7377, + "step": 25100 + }, + { + "epoch": 0.22209858881294264, + "grad_norm": 2.677534580230713, + "learning_rate": 9.955063237436477e-05, + "loss": 3.5157, + "step": 25150 + }, + { + "epoch": 0.22254013670322684, + "grad_norm": 1.7741618156433105, + "learning_rate": 9.954880884685869e-05, + "loss": 3.4256, + "step": 25200 + }, + { + "epoch": 0.222981684593511, + "grad_norm": 0.9014108777046204, + "learning_rate": 9.954694431555343e-05, + "loss": 3.5028, + "step": 25250 + }, + { + "epoch": 0.2234232324837952, + "grad_norm": 2.5479094982147217, + "learning_rate": 9.954507595716685e-05, + "loss": 3.7392, + "step": 25300 + }, + { + "epoch": 0.22386478037407936, + "grad_norm": 1.6607797145843506, + "learning_rate": 9.954320377184328e-05, + "loss": 3.7739, + "step": 25350 + }, + { + "epoch": 0.22430632826436356, + "grad_norm": 2.8050482273101807, + "learning_rate": 9.954132775972735e-05, + "loss": 3.3908, + "step": 25400 + }, + { + "epoch": 0.22474787615464772, + "grad_norm": 0.8809208869934082, + "learning_rate": 9.953944792096392e-05, + "loss": 3.4604, + "step": 25450 + }, + { + "epoch": 0.22518942404493192, + "grad_norm": 4.810263156890869, + "learning_rate": 9.953756425569824e-05, + "loss": 3.3016, + "step": 25500 + }, + { + "epoch": 0.22563097193521608, + "grad_norm": 0.7430752515792847, + "learning_rate": 9.95356767640758e-05, + "loss": 3.3905, + "step": 25550 + }, + { + "epoch": 0.22607251982550028, + "grad_norm": 1.2698413133621216, + "learning_rate": 9.953378544624237e-05, + "loss": 3.4881, + "step": 25600 + }, + { + "epoch": 0.22651406771578445, + "grad_norm": 2.968998908996582, + "learning_rate": 9.953189030234405e-05, + "loss": 3.2608, + "step": 25650 + }, + { + "epoch": 0.22695561560606864, + "grad_norm": 8.66765308380127, + "learning_rate": 9.95299913325272e-05, + "loss": 3.1594, + "step": 25700 + }, + { + "epoch": 0.2273971634963528, + "grad_norm": 3.061037302017212, + "learning_rate": 9.952808853693856e-05, + "loss": 3.5046, + "step": 25750 + }, + { + "epoch": 0.227838711386637, + "grad_norm": 1.6408863067626953, + "learning_rate": 9.952618191572503e-05, + "loss": 2.9974, + "step": 25800 + }, + { + "epoch": 0.22828025927692117, + "grad_norm": 1.7766352891921997, + "learning_rate": 9.952427146903393e-05, + "loss": 3.4394, + "step": 25850 + }, + { + "epoch": 0.22872180716720536, + "grad_norm": 1.8550622463226318, + "learning_rate": 9.952235719701283e-05, + "loss": 3.2499, + "step": 25900 + }, + { + "epoch": 0.22916335505748953, + "grad_norm": 4.626537799835205, + "learning_rate": 9.952043909980955e-05, + "loss": 3.3191, + "step": 25950 + }, + { + "epoch": 0.22960490294777372, + "grad_norm": 1.894518256187439, + "learning_rate": 9.951851717757228e-05, + "loss": 3.2887, + "step": 26000 + }, + { + "epoch": 0.2300464508380579, + "grad_norm": 2.2566466331481934, + "learning_rate": 9.951659143044947e-05, + "loss": 3.2224, + "step": 26050 + }, + { + "epoch": 0.23048799872834208, + "grad_norm": 1.502895474433899, + "learning_rate": 9.951466185858985e-05, + "loss": 3.4682, + "step": 26100 + }, + { + "epoch": 0.23092954661862625, + "grad_norm": 3.664829730987549, + "learning_rate": 9.951272846214247e-05, + "loss": 3.0882, + "step": 26150 + }, + { + "epoch": 0.23137109450891044, + "grad_norm": 2.9891955852508545, + "learning_rate": 9.951079124125668e-05, + "loss": 3.1857, + "step": 26200 + }, + { + "epoch": 0.2318126423991946, + "grad_norm": 1.8004250526428223, + "learning_rate": 9.950885019608211e-05, + "loss": 3.3859, + "step": 26250 + }, + { + "epoch": 0.2322541902894788, + "grad_norm": 2.643892526626587, + "learning_rate": 9.950690532676869e-05, + "loss": 3.1025, + "step": 26300 + }, + { + "epoch": 0.23269573817976297, + "grad_norm": 8.552020072937012, + "learning_rate": 9.950495663346662e-05, + "loss": 2.8521, + "step": 26350 + }, + { + "epoch": 0.23313728607004716, + "grad_norm": 2.302180528640747, + "learning_rate": 9.950300411632645e-05, + "loss": 3.6808, + "step": 26400 + }, + { + "epoch": 0.23357883396033133, + "grad_norm": 2.410003185272217, + "learning_rate": 9.950104777549898e-05, + "loss": 3.2536, + "step": 26450 + }, + { + "epoch": 0.23402038185061552, + "grad_norm": 3.413940906524658, + "learning_rate": 9.949908761113534e-05, + "loss": 3.6206, + "step": 26500 + }, + { + "epoch": 0.2344619297408997, + "grad_norm": 1.564780354499817, + "learning_rate": 9.949712362338692e-05, + "loss": 3.4719, + "step": 26550 + }, + { + "epoch": 0.23490347763118388, + "grad_norm": 1.3204708099365234, + "learning_rate": 9.94951558124054e-05, + "loss": 3.95, + "step": 26600 + }, + { + "epoch": 0.23534502552146805, + "grad_norm": 2.4696288108825684, + "learning_rate": 9.949318417834283e-05, + "loss": 3.8591, + "step": 26650 + }, + { + "epoch": 0.23578657341175224, + "grad_norm": 1.5461256504058838, + "learning_rate": 9.949120872135144e-05, + "loss": 3.6145, + "step": 26700 + }, + { + "epoch": 0.2362281213020364, + "grad_norm": 1.8961230516433716, + "learning_rate": 9.948922944158388e-05, + "loss": 3.4088, + "step": 26750 + }, + { + "epoch": 0.2366696691923206, + "grad_norm": 3.098393201828003, + "learning_rate": 9.9487246339193e-05, + "loss": 3.1048, + "step": 26800 + }, + { + "epoch": 0.23711121708260477, + "grad_norm": 0.8614824414253235, + "learning_rate": 9.948525941433195e-05, + "loss": 3.3338, + "step": 26850 + }, + { + "epoch": 0.23755276497288896, + "grad_norm": 2.924774646759033, + "learning_rate": 9.948326866715426e-05, + "loss": 3.2607, + "step": 26900 + }, + { + "epoch": 0.23799431286317313, + "grad_norm": 5.287550449371338, + "learning_rate": 9.948127409781367e-05, + "loss": 3.4998, + "step": 26950 + }, + { + "epoch": 0.23843586075345732, + "grad_norm": 1.865439534187317, + "learning_rate": 9.947927570646424e-05, + "loss": 3.1788, + "step": 27000 + }, + { + "epoch": 0.23843586075345732, + "eval_asr_loss": 0.8890385764932143, + "eval_loss": 3.1024086475372314, + "eval_runtime": 19.895, + "eval_samples_per_second": 38.603, + "eval_steps_per_second": 9.651, + "eval_tts_loss": 5.952762275795868, + "step": 27000 + }, + { + "epoch": 0.2388774086437415, + "grad_norm": 4.502477169036865, + "learning_rate": 9.947727349326033e-05, + "loss": 2.8551, + "step": 27050 + }, + { + "epoch": 0.23931895653402568, + "grad_norm": 2.6308412551879883, + "learning_rate": 9.94752674583566e-05, + "loss": 3.4248, + "step": 27100 + }, + { + "epoch": 0.23976050442430985, + "grad_norm": 1.3937596082687378, + "learning_rate": 9.9473257601908e-05, + "loss": 3.0961, + "step": 27150 + }, + { + "epoch": 0.24020205231459404, + "grad_norm": 2.9952657222747803, + "learning_rate": 9.947124392406977e-05, + "loss": 3.7838, + "step": 27200 + }, + { + "epoch": 0.2406436002048782, + "grad_norm": 3.2740750312805176, + "learning_rate": 9.946922642499744e-05, + "loss": 3.5869, + "step": 27250 + }, + { + "epoch": 0.2410851480951624, + "grad_norm": 1.290229082107544, + "learning_rate": 9.946720510484686e-05, + "loss": 3.521, + "step": 27300 + }, + { + "epoch": 0.24152669598544657, + "grad_norm": 0.997048020362854, + "learning_rate": 9.946517996377417e-05, + "loss": 3.1288, + "step": 27350 + }, + { + "epoch": 0.24196824387573077, + "grad_norm": 1.3124210834503174, + "learning_rate": 9.946315100193577e-05, + "loss": 2.9812, + "step": 27400 + }, + { + "epoch": 0.24240979176601493, + "grad_norm": 1.2291759252548218, + "learning_rate": 9.946111821948839e-05, + "loss": 3.3039, + "step": 27450 + }, + { + "epoch": 0.24285133965629913, + "grad_norm": 1.4661563634872437, + "learning_rate": 9.945908161658904e-05, + "loss": 3.271, + "step": 27500 + }, + { + "epoch": 0.2432928875465833, + "grad_norm": 2.156710624694824, + "learning_rate": 9.945704119339506e-05, + "loss": 3.7045, + "step": 27550 + }, + { + "epoch": 0.24373443543686749, + "grad_norm": 3.8776209354400635, + "learning_rate": 9.9454996950064e-05, + "loss": 3.6893, + "step": 27600 + }, + { + "epoch": 0.24417598332715165, + "grad_norm": 1.0397090911865234, + "learning_rate": 9.945294888675381e-05, + "loss": 3.5694, + "step": 27650 + }, + { + "epoch": 0.24461753121743585, + "grad_norm": 6.586629867553711, + "learning_rate": 9.945089700362266e-05, + "loss": 3.4295, + "step": 27700 + }, + { + "epoch": 0.24505907910772, + "grad_norm": 1.8651010990142822, + "learning_rate": 9.944884130082905e-05, + "loss": 3.4199, + "step": 27750 + }, + { + "epoch": 0.2455006269980042, + "grad_norm": 2.351825475692749, + "learning_rate": 9.944678177853176e-05, + "loss": 3.88, + "step": 27800 + }, + { + "epoch": 0.24594217488828837, + "grad_norm": 3.991497755050659, + "learning_rate": 9.944471843688987e-05, + "loss": 3.5451, + "step": 27850 + }, + { + "epoch": 0.24638372277857257, + "grad_norm": 4.23520040512085, + "learning_rate": 9.944265127606278e-05, + "loss": 3.2976, + "step": 27900 + }, + { + "epoch": 0.24682527066885673, + "grad_norm": 3.55438232421875, + "learning_rate": 9.944058029621013e-05, + "loss": 2.9787, + "step": 27950 + }, + { + "epoch": 0.24726681855914093, + "grad_norm": 6.835021495819092, + "learning_rate": 9.94385054974919e-05, + "loss": 3.151, + "step": 28000 + }, + { + "epoch": 0.2477083664494251, + "grad_norm": 2.232171058654785, + "learning_rate": 9.943642688006834e-05, + "loss": 3.353, + "step": 28050 + }, + { + "epoch": 0.2481499143397093, + "grad_norm": 2.0550529956817627, + "learning_rate": 9.94343444441e-05, + "loss": 3.3, + "step": 28100 + }, + { + "epoch": 0.24859146222999345, + "grad_norm": 0.9220274686813354, + "learning_rate": 9.943225818974775e-05, + "loss": 3.3238, + "step": 28150 + }, + { + "epoch": 0.24903301012027765, + "grad_norm": 3.3017609119415283, + "learning_rate": 9.943016811717275e-05, + "loss": 3.652, + "step": 28200 + }, + { + "epoch": 0.24947455801056181, + "grad_norm": 2.590731620788574, + "learning_rate": 9.94280742265364e-05, + "loss": 3.362, + "step": 28250 + }, + { + "epoch": 0.249916105900846, + "grad_norm": 1.249603271484375, + "learning_rate": 9.942597651800046e-05, + "loss": 3.2159, + "step": 28300 + }, + { + "epoch": 0.2503576537911302, + "grad_norm": 3.9774234294891357, + "learning_rate": 9.942387499172694e-05, + "loss": 3.2234, + "step": 28350 + }, + { + "epoch": 0.25079920168141434, + "grad_norm": 1.4465439319610596, + "learning_rate": 9.94217696478782e-05, + "loss": 3.3785, + "step": 28400 + }, + { + "epoch": 0.25124074957169856, + "grad_norm": 0.9342706799507141, + "learning_rate": 9.941966048661682e-05, + "loss": 3.1059, + "step": 28450 + }, + { + "epoch": 0.25168229746198273, + "grad_norm": 2.685728073120117, + "learning_rate": 9.941754750810573e-05, + "loss": 3.5622, + "step": 28500 + }, + { + "epoch": 0.2521238453522669, + "grad_norm": 6.555978775024414, + "learning_rate": 9.941543071250816e-05, + "loss": 3.3656, + "step": 28550 + }, + { + "epoch": 0.25256539324255106, + "grad_norm": 0.8330821394920349, + "learning_rate": 9.94133100999876e-05, + "loss": 3.5714, + "step": 28600 + }, + { + "epoch": 0.2530069411328353, + "grad_norm": 22.39084243774414, + "learning_rate": 9.941118567070784e-05, + "loss": 3.1974, + "step": 28650 + }, + { + "epoch": 0.25344848902311945, + "grad_norm": 1.9922549724578857, + "learning_rate": 9.940905742483298e-05, + "loss": 3.5825, + "step": 28700 + }, + { + "epoch": 0.2538900369134036, + "grad_norm": 1.42471444606781, + "learning_rate": 9.94069253625274e-05, + "loss": 3.3538, + "step": 28750 + }, + { + "epoch": 0.2543315848036878, + "grad_norm": 3.5649797916412354, + "learning_rate": 9.940478948395582e-05, + "loss": 3.1705, + "step": 28800 + }, + { + "epoch": 0.254773132693972, + "grad_norm": 2.5138847827911377, + "learning_rate": 9.940264978928318e-05, + "loss": 3.7867, + "step": 28850 + }, + { + "epoch": 0.25521468058425617, + "grad_norm": 1.892633080482483, + "learning_rate": 9.940050627867476e-05, + "loss": 3.1689, + "step": 28900 + }, + { + "epoch": 0.25565622847454034, + "grad_norm": 1.2871747016906738, + "learning_rate": 9.939835895229615e-05, + "loss": 3.651, + "step": 28950 + }, + { + "epoch": 0.2560977763648245, + "grad_norm": 4.025623798370361, + "learning_rate": 9.939620781031318e-05, + "loss": 3.472, + "step": 29000 + }, + { + "epoch": 0.2565393242551087, + "grad_norm": 2.834179162979126, + "learning_rate": 9.939405285289203e-05, + "loss": 3.3033, + "step": 29050 + }, + { + "epoch": 0.2569808721453929, + "grad_norm": 2.194532871246338, + "learning_rate": 9.939189408019916e-05, + "loss": 3.3813, + "step": 29100 + }, + { + "epoch": 0.25742242003567706, + "grad_norm": 2.8828024864196777, + "learning_rate": 9.938973149240127e-05, + "loss": 3.283, + "step": 29150 + }, + { + "epoch": 0.2578639679259612, + "grad_norm": 3.8372597694396973, + "learning_rate": 9.938756508966547e-05, + "loss": 3.324, + "step": 29200 + }, + { + "epoch": 0.25830551581624545, + "grad_norm": 1.0322990417480469, + "learning_rate": 9.938539487215903e-05, + "loss": 3.1929, + "step": 29250 + }, + { + "epoch": 0.2587470637065296, + "grad_norm": 4.666669845581055, + "learning_rate": 9.938322084004964e-05, + "loss": 3.2645, + "step": 29300 + }, + { + "epoch": 0.2591886115968138, + "grad_norm": 2.352522850036621, + "learning_rate": 9.938104299350519e-05, + "loss": 2.8695, + "step": 29350 + }, + { + "epoch": 0.25963015948709794, + "grad_norm": 1.9813988208770752, + "learning_rate": 9.93788613326939e-05, + "loss": 3.0595, + "step": 29400 + }, + { + "epoch": 0.26007170737738217, + "grad_norm": 4.507404804229736, + "learning_rate": 9.93766758577843e-05, + "loss": 3.3092, + "step": 29450 + }, + { + "epoch": 0.26051325526766633, + "grad_norm": 0.7371171712875366, + "learning_rate": 9.93744865689452e-05, + "loss": 3.3464, + "step": 29500 + }, + { + "epoch": 0.2609548031579505, + "grad_norm": 1.4071518182754517, + "learning_rate": 9.937229346634566e-05, + "loss": 2.8855, + "step": 29550 + }, + { + "epoch": 0.26139635104823467, + "grad_norm": 2.9686851501464844, + "learning_rate": 9.937009655015515e-05, + "loss": 3.1077, + "step": 29600 + }, + { + "epoch": 0.2618378989385189, + "grad_norm": 2.3262200355529785, + "learning_rate": 9.936789582054331e-05, + "loss": 3.1181, + "step": 29650 + }, + { + "epoch": 0.26227944682880305, + "grad_norm": 1.0248990058898926, + "learning_rate": 9.936569127768017e-05, + "loss": 3.4325, + "step": 29700 + }, + { + "epoch": 0.2627209947190872, + "grad_norm": 3.9115331172943115, + "learning_rate": 9.936348292173596e-05, + "loss": 3.132, + "step": 29750 + }, + { + "epoch": 0.2631625426093714, + "grad_norm": 4.591248512268066, + "learning_rate": 9.936127075288128e-05, + "loss": 3.151, + "step": 29800 + }, + { + "epoch": 0.2636040904996556, + "grad_norm": 2.473047971725464, + "learning_rate": 9.935905477128703e-05, + "loss": 3.4772, + "step": 29850 + }, + { + "epoch": 0.2640456383899398, + "grad_norm": 1.3905609846115112, + "learning_rate": 9.935683497712434e-05, + "loss": 2.9894, + "step": 29900 + }, + { + "epoch": 0.26448718628022394, + "grad_norm": 2.24664568901062, + "learning_rate": 9.935461137056469e-05, + "loss": 3.689, + "step": 29950 + }, + { + "epoch": 0.2649287341705081, + "grad_norm": 1.1567890644073486, + "learning_rate": 9.935238395177981e-05, + "loss": 3.2017, + "step": 30000 + }, + { + "epoch": 0.2649287341705081, + "eval_asr_loss": 0.8843846080425863, + "eval_loss": 3.092066526412964, + "eval_runtime": 19.8662, + "eval_samples_per_second": 38.659, + "eval_steps_per_second": 9.665, + "eval_tts_loss": 5.902505457880168, + "step": 30000 + }, + { + "epoch": 0.26537028206079233, + "grad_norm": 5.792656421661377, + "learning_rate": 9.935015272094178e-05, + "loss": 3.3117, + "step": 30050 + }, + { + "epoch": 0.2658118299510765, + "grad_norm": 3.329699993133545, + "learning_rate": 9.934791767822292e-05, + "loss": 3.4779, + "step": 30100 + }, + { + "epoch": 0.26625337784136066, + "grad_norm": 1.1682640314102173, + "learning_rate": 9.934567882379589e-05, + "loss": 3.0166, + "step": 30150 + }, + { + "epoch": 0.2666949257316448, + "grad_norm": 2.778029441833496, + "learning_rate": 9.93434361578336e-05, + "loss": 3.209, + "step": 30200 + }, + { + "epoch": 0.26713647362192905, + "grad_norm": 2.259307622909546, + "learning_rate": 9.934118968050929e-05, + "loss": 3.0044, + "step": 30250 + }, + { + "epoch": 0.2675780215122132, + "grad_norm": 2.9219155311584473, + "learning_rate": 9.933898443511526e-05, + "loss": 3.7484, + "step": 30300 + }, + { + "epoch": 0.2680195694024974, + "grad_norm": 1.4216759204864502, + "learning_rate": 9.933673041180634e-05, + "loss": 3.2729, + "step": 30350 + }, + { + "epoch": 0.26846111729278155, + "grad_norm": 0.8821609020233154, + "learning_rate": 9.933447257765338e-05, + "loss": 2.7456, + "step": 30400 + }, + { + "epoch": 0.26890266518306577, + "grad_norm": 4.904880046844482, + "learning_rate": 9.933221093283074e-05, + "loss": 3.2778, + "step": 30450 + }, + { + "epoch": 0.26934421307334994, + "grad_norm": 1.9883630275726318, + "learning_rate": 9.932994547751313e-05, + "loss": 3.4769, + "step": 30500 + }, + { + "epoch": 0.2697857609636341, + "grad_norm": 3.2121076583862305, + "learning_rate": 9.932767621187555e-05, + "loss": 3.2237, + "step": 30550 + }, + { + "epoch": 0.27022730885391827, + "grad_norm": 0.9768474102020264, + "learning_rate": 9.932540313609326e-05, + "loss": 3.2586, + "step": 30600 + }, + { + "epoch": 0.2706688567442025, + "grad_norm": 1.543562889099121, + "learning_rate": 9.932312625034186e-05, + "loss": 3.5258, + "step": 30650 + }, + { + "epoch": 0.27111040463448666, + "grad_norm": 1.3274840116500854, + "learning_rate": 9.932084555479719e-05, + "loss": 3.2218, + "step": 30700 + }, + { + "epoch": 0.2715519525247708, + "grad_norm": 3.6982388496398926, + "learning_rate": 9.931856104963543e-05, + "loss": 3.4994, + "step": 30750 + }, + { + "epoch": 0.271993500415055, + "grad_norm": 1.6436798572540283, + "learning_rate": 9.931627273503304e-05, + "loss": 3.4035, + "step": 30800 + }, + { + "epoch": 0.2724350483053392, + "grad_norm": 4.315684795379639, + "learning_rate": 9.931398061116677e-05, + "loss": 3.4837, + "step": 30850 + }, + { + "epoch": 0.2728765961956234, + "grad_norm": 1.2069634199142456, + "learning_rate": 9.931168467821367e-05, + "loss": 3.2571, + "step": 30900 + }, + { + "epoch": 0.27331814408590754, + "grad_norm": 3.137390375137329, + "learning_rate": 9.930938493635109e-05, + "loss": 3.2457, + "step": 30950 + }, + { + "epoch": 0.2737596919761917, + "grad_norm": 1.6049838066101074, + "learning_rate": 9.930708138575664e-05, + "loss": 2.9232, + "step": 31000 + }, + { + "epoch": 0.27420123986647593, + "grad_norm": 2.4024581909179688, + "learning_rate": 9.930477402660827e-05, + "loss": 3.1933, + "step": 31050 + }, + { + "epoch": 0.2746427877567601, + "grad_norm": 2.070897340774536, + "learning_rate": 9.93024628590842e-05, + "loss": 3.1729, + "step": 31100 + }, + { + "epoch": 0.27508433564704426, + "grad_norm": 2.37459397315979, + "learning_rate": 9.930014788336293e-05, + "loss": 3.0385, + "step": 31150 + }, + { + "epoch": 0.27552588353732843, + "grad_norm": 3.5087156295776367, + "learning_rate": 9.92978290996233e-05, + "loss": 3.1936, + "step": 31200 + }, + { + "epoch": 0.27596743142761265, + "grad_norm": 6.348201751708984, + "learning_rate": 9.929550650804443e-05, + "loss": 3.2907, + "step": 31250 + }, + { + "epoch": 0.2764089793178968, + "grad_norm": 1.3216320276260376, + "learning_rate": 9.929318010880566e-05, + "loss": 3.1206, + "step": 31300 + }, + { + "epoch": 0.276850527208181, + "grad_norm": 3.7011451721191406, + "learning_rate": 9.929084990208674e-05, + "loss": 3.6797, + "step": 31350 + }, + { + "epoch": 0.2772920750984652, + "grad_norm": 1.47987961769104, + "learning_rate": 9.928851588806763e-05, + "loss": 3.2982, + "step": 31400 + }, + { + "epoch": 0.2777336229887494, + "grad_norm": 2.7035892009735107, + "learning_rate": 9.928617806692863e-05, + "loss": 3.5924, + "step": 31450 + }, + { + "epoch": 0.27817517087903354, + "grad_norm": 1.140507698059082, + "learning_rate": 9.92838364388503e-05, + "loss": 3.4756, + "step": 31500 + }, + { + "epoch": 0.2786167187693177, + "grad_norm": 1.1196190118789673, + "learning_rate": 9.928149100401354e-05, + "loss": 3.5627, + "step": 31550 + }, + { + "epoch": 0.27905826665960193, + "grad_norm": 3.444877862930298, + "learning_rate": 9.927914176259947e-05, + "loss": 3.3291, + "step": 31600 + }, + { + "epoch": 0.2794998145498861, + "grad_norm": 4.410919189453125, + "learning_rate": 9.927678871478958e-05, + "loss": 3.1749, + "step": 31650 + }, + { + "epoch": 0.27994136244017026, + "grad_norm": 1.771317481994629, + "learning_rate": 9.927443186076563e-05, + "loss": 2.6498, + "step": 31700 + }, + { + "epoch": 0.2803829103304544, + "grad_norm": 2.1711716651916504, + "learning_rate": 9.927207120070964e-05, + "loss": 3.6787, + "step": 31750 + }, + { + "epoch": 0.28082445822073865, + "grad_norm": 4.0384202003479, + "learning_rate": 9.926970673480398e-05, + "loss": 3.6298, + "step": 31800 + }, + { + "epoch": 0.2812660061110228, + "grad_norm": 2.198716640472412, + "learning_rate": 9.926733846323125e-05, + "loss": 2.9091, + "step": 31850 + }, + { + "epoch": 0.281707554001307, + "grad_norm": 1.819726586341858, + "learning_rate": 9.92649663861744e-05, + "loss": 3.5799, + "step": 31900 + }, + { + "epoch": 0.28214910189159115, + "grad_norm": 1.7500370740890503, + "learning_rate": 9.926259050381667e-05, + "loss": 3.4091, + "step": 31950 + }, + { + "epoch": 0.28259064978187537, + "grad_norm": 4.203965663909912, + "learning_rate": 9.926021081634155e-05, + "loss": 3.2101, + "step": 32000 + }, + { + "epoch": 0.28303219767215954, + "grad_norm": 2.5612051486968994, + "learning_rate": 9.925782732393285e-05, + "loss": 3.0251, + "step": 32050 + }, + { + "epoch": 0.2834737455624437, + "grad_norm": 1.427227258682251, + "learning_rate": 9.92554400267747e-05, + "loss": 3.2904, + "step": 32100 + }, + { + "epoch": 0.28391529345272787, + "grad_norm": 1.6692140102386475, + "learning_rate": 9.925304892505146e-05, + "loss": 3.74, + "step": 32150 + }, + { + "epoch": 0.2843568413430121, + "grad_norm": 1.287840485572815, + "learning_rate": 9.925065401894786e-05, + "loss": 3.5495, + "step": 32200 + }, + { + "epoch": 0.28479838923329626, + "grad_norm": 1.882232427597046, + "learning_rate": 9.924825530864887e-05, + "loss": 3.2056, + "step": 32250 + }, + { + "epoch": 0.2852399371235804, + "grad_norm": 1.2615700960159302, + "learning_rate": 9.924585279433976e-05, + "loss": 3.1918, + "step": 32300 + }, + { + "epoch": 0.2856814850138646, + "grad_norm": 1.7634786367416382, + "learning_rate": 9.924344647620612e-05, + "loss": 3.5558, + "step": 32350 + }, + { + "epoch": 0.2861230329041488, + "grad_norm": 1.2818759679794312, + "learning_rate": 9.924103635443381e-05, + "loss": 3.0904, + "step": 32400 + }, + { + "epoch": 0.286564580794433, + "grad_norm": 3.8670778274536133, + "learning_rate": 9.9238622429209e-05, + "loss": 3.8008, + "step": 32450 + }, + { + "epoch": 0.28700612868471714, + "grad_norm": 2.2975099086761475, + "learning_rate": 9.923625309255875e-05, + "loss": 2.9886, + "step": 32500 + }, + { + "epoch": 0.2874476765750013, + "grad_norm": 3.461240530014038, + "learning_rate": 9.923383163704833e-05, + "loss": 3.3378, + "step": 32550 + }, + { + "epoch": 0.28788922446528553, + "grad_norm": 1.2354512214660645, + "learning_rate": 9.923140637864192e-05, + "loss": 3.4829, + "step": 32600 + }, + { + "epoch": 0.2883307723555697, + "grad_norm": 2.5937671661376953, + "learning_rate": 9.922897731752683e-05, + "loss": 3.3737, + "step": 32650 + }, + { + "epoch": 0.28877232024585386, + "grad_norm": 2.7086663246154785, + "learning_rate": 9.922654445389069e-05, + "loss": 3.3247, + "step": 32700 + }, + { + "epoch": 0.28921386813613803, + "grad_norm": 1.1576141119003296, + "learning_rate": 9.922410778792142e-05, + "loss": 3.3032, + "step": 32750 + }, + { + "epoch": 0.28965541602642225, + "grad_norm": 4.800418853759766, + "learning_rate": 9.922166731980725e-05, + "loss": 3.7363, + "step": 32800 + }, + { + "epoch": 0.2900969639167064, + "grad_norm": 1.6237969398498535, + "learning_rate": 9.921922304973663e-05, + "loss": 3.6136, + "step": 32850 + }, + { + "epoch": 0.2905385118069906, + "grad_norm": 3.0943002700805664, + "learning_rate": 9.921677497789843e-05, + "loss": 3.18, + "step": 32900 + }, + { + "epoch": 0.29098005969727475, + "grad_norm": 2.4920825958251953, + "learning_rate": 9.921432310448169e-05, + "loss": 3.2806, + "step": 32950 + }, + { + "epoch": 0.291421607587559, + "grad_norm": 4.264760971069336, + "learning_rate": 9.921186742967581e-05, + "loss": 2.9155, + "step": 33000 + }, + { + "epoch": 0.291421607587559, + "eval_asr_loss": 0.895976054242954, + "eval_loss": 3.079402208328247, + "eval_runtime": 20.0371, + "eval_samples_per_second": 38.329, + "eval_steps_per_second": 9.582, + "eval_tts_loss": 5.968250579680352, + "step": 33000 + }, + { + "epoch": 0.29186315547784314, + "grad_norm": 3.386139154434204, + "learning_rate": 9.92094079536705e-05, + "loss": 3.9219, + "step": 33050 + }, + { + "epoch": 0.2923047033681273, + "grad_norm": 1.537395715713501, + "learning_rate": 9.92069446766557e-05, + "loss": 3.1577, + "step": 33100 + }, + { + "epoch": 0.29274625125841147, + "grad_norm": 0.6773030757904053, + "learning_rate": 9.920447759882168e-05, + "loss": 2.9516, + "step": 33150 + }, + { + "epoch": 0.2931877991486957, + "grad_norm": 2.179901599884033, + "learning_rate": 9.9202006720359e-05, + "loss": 3.7766, + "step": 33200 + }, + { + "epoch": 0.29362934703897986, + "grad_norm": 4.384719371795654, + "learning_rate": 9.919953204145853e-05, + "loss": 3.4504, + "step": 33250 + }, + { + "epoch": 0.294070894929264, + "grad_norm": 5.59550666809082, + "learning_rate": 9.919705356231141e-05, + "loss": 3.2456, + "step": 33300 + }, + { + "epoch": 0.2945124428195482, + "grad_norm": 1.8736000061035156, + "learning_rate": 9.919457128310909e-05, + "loss": 3.5584, + "step": 33350 + }, + { + "epoch": 0.2949539907098324, + "grad_norm": 2.3086678981781006, + "learning_rate": 9.919208520404328e-05, + "loss": 3.3485, + "step": 33400 + }, + { + "epoch": 0.2953955386001166, + "grad_norm": 1.736769676208496, + "learning_rate": 9.918959532530604e-05, + "loss": 3.426, + "step": 33450 + }, + { + "epoch": 0.29583708649040075, + "grad_norm": 1.8966106176376343, + "learning_rate": 9.918710164708967e-05, + "loss": 3.0041, + "step": 33500 + }, + { + "epoch": 0.2962786343806849, + "grad_norm": 3.222245216369629, + "learning_rate": 9.918460416958679e-05, + "loss": 3.0005, + "step": 33550 + }, + { + "epoch": 0.29672018227096914, + "grad_norm": 0.8299497365951538, + "learning_rate": 9.918210289299031e-05, + "loss": 3.3702, + "step": 33600 + }, + { + "epoch": 0.2971617301612533, + "grad_norm": 3.876796245574951, + "learning_rate": 9.917959781749345e-05, + "loss": 3.243, + "step": 33650 + }, + { + "epoch": 0.29760327805153747, + "grad_norm": 4.3070807456970215, + "learning_rate": 9.917708894328968e-05, + "loss": 3.4787, + "step": 33700 + }, + { + "epoch": 0.29804482594182163, + "grad_norm": 2.306934356689453, + "learning_rate": 9.917457627057279e-05, + "loss": 3.2177, + "step": 33750 + }, + { + "epoch": 0.29848637383210586, + "grad_norm": 3.8237075805664062, + "learning_rate": 9.917205979953687e-05, + "loss": 2.9533, + "step": 33800 + }, + { + "epoch": 0.29892792172239, + "grad_norm": 1.627469778060913, + "learning_rate": 9.91695395303763e-05, + "loss": 3.3876, + "step": 33850 + }, + { + "epoch": 0.2993694696126742, + "grad_norm": 1.0928808450698853, + "learning_rate": 9.916701546328576e-05, + "loss": 2.8597, + "step": 33900 + }, + { + "epoch": 0.29981101750295835, + "grad_norm": 2.0172150135040283, + "learning_rate": 9.916448759846017e-05, + "loss": 3.6425, + "step": 33950 + }, + { + "epoch": 0.3002525653932426, + "grad_norm": 2.5472891330718994, + "learning_rate": 9.916195593609486e-05, + "loss": 3.4216, + "step": 34000 + }, + { + "epoch": 0.30069411328352674, + "grad_norm": 2.0118589401245117, + "learning_rate": 9.91594712227922e-05, + "loss": 3.1182, + "step": 34050 + }, + { + "epoch": 0.3011356611738109, + "grad_norm": 1.4318723678588867, + "learning_rate": 9.915693204187532e-05, + "loss": 3.2758, + "step": 34100 + }, + { + "epoch": 0.3015772090640951, + "grad_norm": 6.2676801681518555, + "learning_rate": 9.915438906400228e-05, + "loss": 3.2553, + "step": 34150 + }, + { + "epoch": 0.3020187569543793, + "grad_norm": 1.1752573251724243, + "learning_rate": 9.915184228936952e-05, + "loss": 3.4345, + "step": 34200 + }, + { + "epoch": 0.30246030484466346, + "grad_norm": 1.754859447479248, + "learning_rate": 9.914929171817371e-05, + "loss": 3.5339, + "step": 34250 + }, + { + "epoch": 0.30290185273494763, + "grad_norm": 1.1092463731765747, + "learning_rate": 9.914673735061191e-05, + "loss": 3.7403, + "step": 34300 + }, + { + "epoch": 0.3033434006252318, + "grad_norm": 2.3356130123138428, + "learning_rate": 9.91441791868814e-05, + "loss": 3.7242, + "step": 34350 + }, + { + "epoch": 0.303784948515516, + "grad_norm": 1.9679780006408691, + "learning_rate": 9.914161722717976e-05, + "loss": 3.0423, + "step": 34400 + }, + { + "epoch": 0.3042264964058002, + "grad_norm": 2.5788609981536865, + "learning_rate": 9.91390514717049e-05, + "loss": 3.3403, + "step": 34450 + }, + { + "epoch": 0.30466804429608435, + "grad_norm": 5.141701698303223, + "learning_rate": 9.913648192065503e-05, + "loss": 3.0419, + "step": 34500 + }, + { + "epoch": 0.3051095921863685, + "grad_norm": 0.9429007172584534, + "learning_rate": 9.913390857422858e-05, + "loss": 3.4227, + "step": 34550 + }, + { + "epoch": 0.30555114007665274, + "grad_norm": 5.526501655578613, + "learning_rate": 9.913133143262432e-05, + "loss": 3.4853, + "step": 34600 + }, + { + "epoch": 0.3059926879669369, + "grad_norm": 1.2435593605041504, + "learning_rate": 9.912875049604135e-05, + "loss": 3.1218, + "step": 34650 + }, + { + "epoch": 0.30643423585722107, + "grad_norm": 3.752776861190796, + "learning_rate": 9.912616576467899e-05, + "loss": 3.29, + "step": 34700 + }, + { + "epoch": 0.30687578374750524, + "grad_norm": 1.3690961599349976, + "learning_rate": 9.91235772387369e-05, + "loss": 3.5259, + "step": 34750 + }, + { + "epoch": 0.30731733163778946, + "grad_norm": 5.097490310668945, + "learning_rate": 9.912098491841504e-05, + "loss": 3.4285, + "step": 34800 + }, + { + "epoch": 0.3077588795280736, + "grad_norm": 2.029536485671997, + "learning_rate": 9.911838880391362e-05, + "loss": 3.5238, + "step": 34850 + }, + { + "epoch": 0.3082004274183578, + "grad_norm": 1.1511303186416626, + "learning_rate": 9.911578889543318e-05, + "loss": 3.5973, + "step": 34900 + }, + { + "epoch": 0.30864197530864196, + "grad_norm": 0.7825906872749329, + "learning_rate": 9.911318519317454e-05, + "loss": 3.5027, + "step": 34950 + }, + { + "epoch": 0.3090835231989262, + "grad_norm": 1.2858946323394775, + "learning_rate": 9.91105776973388e-05, + "loss": 3.5128, + "step": 35000 + }, + { + "epoch": 0.30952507108921035, + "grad_norm": 3.028378486633301, + "learning_rate": 9.910796640812739e-05, + "loss": 3.2618, + "step": 35050 + }, + { + "epoch": 0.3099666189794945, + "grad_norm": 2.016479253768921, + "learning_rate": 9.910535132574198e-05, + "loss": 3.4658, + "step": 35100 + }, + { + "epoch": 0.3104081668697787, + "grad_norm": 5.971153736114502, + "learning_rate": 9.91027324503846e-05, + "loss": 3.5681, + "step": 35150 + }, + { + "epoch": 0.3108497147600629, + "grad_norm": 4.901987552642822, + "learning_rate": 9.910010978225753e-05, + "loss": 3.2022, + "step": 35200 + }, + { + "epoch": 0.31129126265034707, + "grad_norm": 1.5224626064300537, + "learning_rate": 9.90974833215633e-05, + "loss": 3.4693, + "step": 35250 + }, + { + "epoch": 0.31173281054063123, + "grad_norm": 1.7592462301254272, + "learning_rate": 9.909485306850483e-05, + "loss": 3.2146, + "step": 35300 + }, + { + "epoch": 0.3121743584309154, + "grad_norm": 2.2695250511169434, + "learning_rate": 9.909221902328529e-05, + "loss": 3.3325, + "step": 35350 + }, + { + "epoch": 0.3126159063211996, + "grad_norm": 1.2258449792861938, + "learning_rate": 9.908958118610811e-05, + "loss": 3.4704, + "step": 35400 + }, + { + "epoch": 0.3130574542114838, + "grad_norm": 5.258679389953613, + "learning_rate": 9.908693955717705e-05, + "loss": 3.579, + "step": 35450 + }, + { + "epoch": 0.31349900210176795, + "grad_norm": 1.850501298904419, + "learning_rate": 9.908429413669615e-05, + "loss": 3.2378, + "step": 35500 + }, + { + "epoch": 0.3139405499920521, + "grad_norm": 1.2289330959320068, + "learning_rate": 9.908164492486976e-05, + "loss": 3.4111, + "step": 35550 + }, + { + "epoch": 0.31438209788233634, + "grad_norm": 2.0499982833862305, + "learning_rate": 9.907899192190249e-05, + "loss": 3.6974, + "step": 35600 + }, + { + "epoch": 0.3148236457726205, + "grad_norm": 1.0805225372314453, + "learning_rate": 9.907633512799927e-05, + "loss": 3.1798, + "step": 35650 + }, + { + "epoch": 0.3152651936629047, + "grad_norm": 1.2690644264221191, + "learning_rate": 9.907367454336533e-05, + "loss": 2.9274, + "step": 35700 + }, + { + "epoch": 0.31570674155318884, + "grad_norm": 5.509250164031982, + "learning_rate": 9.907101016820616e-05, + "loss": 3.1606, + "step": 35750 + }, + { + "epoch": 0.31614828944347306, + "grad_norm": 3.2590670585632324, + "learning_rate": 9.906834200272756e-05, + "loss": 3.2878, + "step": 35800 + }, + { + "epoch": 0.31658983733375723, + "grad_norm": 1.7757644653320312, + "learning_rate": 9.906567004713563e-05, + "loss": 3.3347, + "step": 35850 + }, + { + "epoch": 0.3170313852240414, + "grad_norm": 6.056863784790039, + "learning_rate": 9.906299430163675e-05, + "loss": 3.3504, + "step": 35900 + }, + { + "epoch": 0.31747293311432556, + "grad_norm": 1.0027168989181519, + "learning_rate": 9.90603147664376e-05, + "loss": 3.0737, + "step": 35950 + }, + { + "epoch": 0.3179144810046098, + "grad_norm": 1.916014313697815, + "learning_rate": 9.905763144174517e-05, + "loss": 3.1648, + "step": 36000 + }, + { + "epoch": 0.3179144810046098, + "eval_asr_loss": 0.8892677519164152, + "eval_loss": 3.0999526977539062, + "eval_runtime": 20.2917, + "eval_samples_per_second": 37.848, + "eval_steps_per_second": 9.462, + "eval_tts_loss": 5.944016528967177, + "step": 36000 + }, + { + "epoch": 0.31835602889489395, + "grad_norm": 2.1346523761749268, + "learning_rate": 9.90549443277667e-05, + "loss": 3.5172, + "step": 36050 + }, + { + "epoch": 0.3187975767851781, + "grad_norm": 1.5300114154815674, + "learning_rate": 9.905225342470974e-05, + "loss": 3.2909, + "step": 36100 + }, + { + "epoch": 0.3192391246754623, + "grad_norm": 1.7766307592391968, + "learning_rate": 9.904955873278216e-05, + "loss": 3.5821, + "step": 36150 + }, + { + "epoch": 0.3196806725657465, + "grad_norm": 1.8257805109024048, + "learning_rate": 9.904686025219211e-05, + "loss": 3.3543, + "step": 36200 + }, + { + "epoch": 0.32012222045603067, + "grad_norm": 2.9343011379241943, + "learning_rate": 9.9044157983148e-05, + "loss": 3.5759, + "step": 36250 + }, + { + "epoch": 0.32056376834631484, + "grad_norm": 2.4320614337921143, + "learning_rate": 9.904145192585857e-05, + "loss": 3.2795, + "step": 36300 + }, + { + "epoch": 0.321005316236599, + "grad_norm": 2.0806117057800293, + "learning_rate": 9.903874208053283e-05, + "loss": 2.9637, + "step": 36350 + }, + { + "epoch": 0.3214468641268832, + "grad_norm": 10.894238471984863, + "learning_rate": 9.903602844738012e-05, + "loss": 3.5964, + "step": 36400 + }, + { + "epoch": 0.3218884120171674, + "grad_norm": 1.5613411664962769, + "learning_rate": 9.903331102661001e-05, + "loss": 3.2076, + "step": 36450 + }, + { + "epoch": 0.32232995990745156, + "grad_norm": 3.5824601650238037, + "learning_rate": 9.903058981843242e-05, + "loss": 3.3151, + "step": 36500 + }, + { + "epoch": 0.3227715077977357, + "grad_norm": 4.121826171875, + "learning_rate": 9.902786482305753e-05, + "loss": 3.1602, + "step": 36550 + }, + { + "epoch": 0.32321305568801995, + "grad_norm": 2.1794791221618652, + "learning_rate": 9.902513604069583e-05, + "loss": 3.3487, + "step": 36600 + }, + { + "epoch": 0.3236546035783041, + "grad_norm": 3.4854958057403564, + "learning_rate": 9.902240347155809e-05, + "loss": 3.5573, + "step": 36650 + }, + { + "epoch": 0.3240961514685883, + "grad_norm": 2.346065044403076, + "learning_rate": 9.901966711585538e-05, + "loss": 2.7336, + "step": 36700 + }, + { + "epoch": 0.32453769935887244, + "grad_norm": 1.1362333297729492, + "learning_rate": 9.901692697379905e-05, + "loss": 3.5597, + "step": 36750 + }, + { + "epoch": 0.32497924724915667, + "grad_norm": 1.6578530073165894, + "learning_rate": 9.901418304560077e-05, + "loss": 3.4961, + "step": 36800 + }, + { + "epoch": 0.32542079513944083, + "grad_norm": 1.7271004915237427, + "learning_rate": 9.901143533147248e-05, + "loss": 3.1233, + "step": 36850 + }, + { + "epoch": 0.325862343029725, + "grad_norm": 5.3198652267456055, + "learning_rate": 9.900868383162643e-05, + "loss": 3.6441, + "step": 36900 + }, + { + "epoch": 0.32630389092000917, + "grad_norm": 1.4437021017074585, + "learning_rate": 9.900592854627511e-05, + "loss": 3.239, + "step": 36950 + }, + { + "epoch": 0.3267454388102934, + "grad_norm": 3.09059739112854, + "learning_rate": 9.900316947563138e-05, + "loss": 3.3918, + "step": 37000 + }, + { + "epoch": 0.32718698670057755, + "grad_norm": 1.7625781297683716, + "learning_rate": 9.900040661990835e-05, + "loss": 3.5584, + "step": 37050 + }, + { + "epoch": 0.3276285345908617, + "grad_norm": 1.064771294593811, + "learning_rate": 9.899763997931942e-05, + "loss": 3.0205, + "step": 37100 + }, + { + "epoch": 0.3280700824811459, + "grad_norm": 2.178464889526367, + "learning_rate": 9.899486955407826e-05, + "loss": 3.2899, + "step": 37150 + }, + { + "epoch": 0.3285116303714301, + "grad_norm": 1.889468789100647, + "learning_rate": 9.899209534439891e-05, + "loss": 3.3759, + "step": 37200 + }, + { + "epoch": 0.3289531782617143, + "grad_norm": 3.320777654647827, + "learning_rate": 9.898931735049565e-05, + "loss": 3.6612, + "step": 37250 + }, + { + "epoch": 0.32939472615199844, + "grad_norm": 4.364492416381836, + "learning_rate": 9.898653557258303e-05, + "loss": 3.5972, + "step": 37300 + }, + { + "epoch": 0.3298362740422826, + "grad_norm": 4.426956653594971, + "learning_rate": 9.898375001087593e-05, + "loss": 3.5658, + "step": 37350 + }, + { + "epoch": 0.33027782193256683, + "grad_norm": 1.2732926607131958, + "learning_rate": 9.898096066558951e-05, + "loss": 3.1916, + "step": 37400 + }, + { + "epoch": 0.330719369822851, + "grad_norm": 1.8885505199432373, + "learning_rate": 9.897816753693922e-05, + "loss": 3.6108, + "step": 37450 + }, + { + "epoch": 0.33116091771313516, + "grad_norm": 5.467677116394043, + "learning_rate": 9.897537062514081e-05, + "loss": 3.079, + "step": 37500 + }, + { + "epoch": 0.3316024656034193, + "grad_norm": 3.006726026535034, + "learning_rate": 9.897256993041033e-05, + "loss": 3.3902, + "step": 37550 + }, + { + "epoch": 0.33204401349370355, + "grad_norm": 3.979788064956665, + "learning_rate": 9.896976545296411e-05, + "loss": 2.8985, + "step": 37600 + }, + { + "epoch": 0.3324855613839877, + "grad_norm": 0.933592677116394, + "learning_rate": 9.896695719301874e-05, + "loss": 3.3717, + "step": 37650 + }, + { + "epoch": 0.3329271092742719, + "grad_norm": 2.022643804550171, + "learning_rate": 9.896414515079115e-05, + "loss": 3.2212, + "step": 37700 + }, + { + "epoch": 0.33336865716455605, + "grad_norm": 1.9231008291244507, + "learning_rate": 9.896132932649857e-05, + "loss": 3.0829, + "step": 37750 + }, + { + "epoch": 0.33381020505484027, + "grad_norm": 6.672185897827148, + "learning_rate": 9.895850972035847e-05, + "loss": 3.0794, + "step": 37800 + }, + { + "epoch": 0.33425175294512444, + "grad_norm": 2.566579580307007, + "learning_rate": 9.895568633258865e-05, + "loss": 3.2348, + "step": 37850 + }, + { + "epoch": 0.3346933008354086, + "grad_norm": 3.8691163063049316, + "learning_rate": 9.895285916340719e-05, + "loss": 3.1823, + "step": 37900 + }, + { + "epoch": 0.33513484872569277, + "grad_norm": 1.986522912979126, + "learning_rate": 9.895002821303246e-05, + "loss": 3.4586, + "step": 37950 + }, + { + "epoch": 0.335576396615977, + "grad_norm": 1.734668254852295, + "learning_rate": 9.894719348168315e-05, + "loss": 3.7534, + "step": 38000 + }, + { + "epoch": 0.33601794450626116, + "grad_norm": 1.9242331981658936, + "learning_rate": 9.894435496957819e-05, + "loss": 3.3395, + "step": 38050 + }, + { + "epoch": 0.3364594923965453, + "grad_norm": 1.77113676071167, + "learning_rate": 9.894151267693686e-05, + "loss": 3.352, + "step": 38100 + }, + { + "epoch": 0.3369010402868295, + "grad_norm": 0.799421489238739, + "learning_rate": 9.893866660397867e-05, + "loss": 3.7193, + "step": 38150 + }, + { + "epoch": 0.3373425881771137, + "grad_norm": 1.236433982849121, + "learning_rate": 9.89358167509235e-05, + "loss": 3.5647, + "step": 38200 + }, + { + "epoch": 0.3377841360673979, + "grad_norm": 3.1852049827575684, + "learning_rate": 9.893296311799143e-05, + "loss": 3.477, + "step": 38250 + }, + { + "epoch": 0.33822568395768204, + "grad_norm": 2.2389414310455322, + "learning_rate": 9.89301057054029e-05, + "loss": 2.8368, + "step": 38300 + }, + { + "epoch": 0.3386672318479662, + "grad_norm": 6.785423278808594, + "learning_rate": 9.892724451337863e-05, + "loss": 3.3759, + "step": 38350 + }, + { + "epoch": 0.33910877973825043, + "grad_norm": 4.653458595275879, + "learning_rate": 9.892437954213958e-05, + "loss": 3.353, + "step": 38400 + }, + { + "epoch": 0.3395503276285346, + "grad_norm": 3.1158320903778076, + "learning_rate": 9.892151079190711e-05, + "loss": 3.5522, + "step": 38450 + }, + { + "epoch": 0.33999187551881876, + "grad_norm": 2.8249120712280273, + "learning_rate": 9.891863826290277e-05, + "loss": 3.3592, + "step": 38500 + }, + { + "epoch": 0.34043342340910293, + "grad_norm": 2.011695384979248, + "learning_rate": 9.891576195534842e-05, + "loss": 3.3599, + "step": 38550 + }, + { + "epoch": 0.34087497129938715, + "grad_norm": 1.8144574165344238, + "learning_rate": 9.891288186946628e-05, + "loss": 3.3011, + "step": 38600 + }, + { + "epoch": 0.3413165191896713, + "grad_norm": 2.060272455215454, + "learning_rate": 9.890999800547878e-05, + "loss": 3.4336, + "step": 38650 + }, + { + "epoch": 0.3417580670799555, + "grad_norm": 1.401672124862671, + "learning_rate": 9.890711036360867e-05, + "loss": 3.1762, + "step": 38700 + }, + { + "epoch": 0.34219961497023965, + "grad_norm": 1.6355420351028442, + "learning_rate": 9.8904218944079e-05, + "loss": 3.6908, + "step": 38750 + }, + { + "epoch": 0.3426411628605239, + "grad_norm": 2.591089963912964, + "learning_rate": 9.890132374711312e-05, + "loss": 3.3153, + "step": 38800 + }, + { + "epoch": 0.34308271075080804, + "grad_norm": 2.8402106761932373, + "learning_rate": 9.889842477293465e-05, + "loss": 3.5123, + "step": 38850 + }, + { + "epoch": 0.3435242586410922, + "grad_norm": 0.9350060820579529, + "learning_rate": 9.889552202176751e-05, + "loss": 3.3598, + "step": 38900 + }, + { + "epoch": 0.3439658065313764, + "grad_norm": 2.9623401165008545, + "learning_rate": 9.889261549383591e-05, + "loss": 3.1874, + "step": 38950 + }, + { + "epoch": 0.3444073544216606, + "grad_norm": 1.7671315670013428, + "learning_rate": 9.888970518936438e-05, + "loss": 3.5201, + "step": 39000 + }, + { + "epoch": 0.3444073544216606, + "eval_asr_loss": 0.8851676637238344, + "eval_loss": 3.074697494506836, + "eval_runtime": 20.3913, + "eval_samples_per_second": 37.663, + "eval_steps_per_second": 9.416, + "eval_tts_loss": 5.894229360095819, + "step": 39000 + }, + { + "epoch": 0.34484890231194476, + "grad_norm": 1.1440892219543457, + "learning_rate": 9.888679110857768e-05, + "loss": 3.4325, + "step": 39050 + }, + { + "epoch": 0.3452904502022289, + "grad_norm": 2.065777540206909, + "learning_rate": 9.888387325170093e-05, + "loss": 3.3255, + "step": 39100 + }, + { + "epoch": 0.3457319980925131, + "grad_norm": 5.079590320587158, + "learning_rate": 9.888095161895947e-05, + "loss": 3.1736, + "step": 39150 + }, + { + "epoch": 0.3461735459827973, + "grad_norm": 2.521078586578369, + "learning_rate": 9.887802621057902e-05, + "loss": 3.6232, + "step": 39200 + }, + { + "epoch": 0.3466150938730815, + "grad_norm": 3.053846836090088, + "learning_rate": 9.88750970267855e-05, + "loss": 3.2633, + "step": 39250 + }, + { + "epoch": 0.34705664176336565, + "grad_norm": 1.3491613864898682, + "learning_rate": 9.88721640678052e-05, + "loss": 3.7012, + "step": 39300 + }, + { + "epoch": 0.3474981896536498, + "grad_norm": 1.5826070308685303, + "learning_rate": 9.886922733386463e-05, + "loss": 3.3255, + "step": 39350 + }, + { + "epoch": 0.34793973754393404, + "grad_norm": 5.572771072387695, + "learning_rate": 9.886628682519069e-05, + "loss": 3.2976, + "step": 39400 + }, + { + "epoch": 0.3483812854342182, + "grad_norm": 1.5839028358459473, + "learning_rate": 9.886334254201041e-05, + "loss": 3.3149, + "step": 39450 + }, + { + "epoch": 0.34882283332450237, + "grad_norm": 3.295896053314209, + "learning_rate": 9.88603944845513e-05, + "loss": 2.9223, + "step": 39500 + }, + { + "epoch": 0.34926438121478653, + "grad_norm": 2.2568423748016357, + "learning_rate": 9.885744265304104e-05, + "loss": 3.527, + "step": 39550 + }, + { + "epoch": 0.34970592910507076, + "grad_norm": 0.9765694737434387, + "learning_rate": 9.885448704770762e-05, + "loss": 3.4622, + "step": 39600 + }, + { + "epoch": 0.3501474769953549, + "grad_norm": 1.8776651620864868, + "learning_rate": 9.885152766877936e-05, + "loss": 3.327, + "step": 39650 + }, + { + "epoch": 0.3505890248856391, + "grad_norm": 3.4537465572357178, + "learning_rate": 9.884856451648483e-05, + "loss": 3.5304, + "step": 39700 + }, + { + "epoch": 0.35103057277592326, + "grad_norm": 2.3317794799804688, + "learning_rate": 9.884559759105292e-05, + "loss": 3.485, + "step": 39750 + }, + { + "epoch": 0.3514721206662075, + "grad_norm": 2.5425727367401123, + "learning_rate": 9.88426268927128e-05, + "loss": 3.5912, + "step": 39800 + }, + { + "epoch": 0.35191366855649164, + "grad_norm": 0.6102803945541382, + "learning_rate": 9.883965242169392e-05, + "loss": 3.0589, + "step": 39850 + }, + { + "epoch": 0.3523552164467758, + "grad_norm": 2.142951726913452, + "learning_rate": 9.883667417822604e-05, + "loss": 3.4354, + "step": 39900 + }, + { + "epoch": 0.35279676433706, + "grad_norm": 0.8856062293052673, + "learning_rate": 9.883369216253919e-05, + "loss": 3.3475, + "step": 39950 + }, + { + "epoch": 0.3532383122273442, + "grad_norm": 1.8055260181427002, + "learning_rate": 9.883070637486374e-05, + "loss": 3.2718, + "step": 40000 + }, + { + "epoch": 0.35367986011762836, + "grad_norm": 1.1027382612228394, + "learning_rate": 9.882777664358069e-05, + "loss": 3.1336, + "step": 40050 + }, + { + "epoch": 0.35412140800791253, + "grad_norm": 5.140567779541016, + "learning_rate": 9.882478338804843e-05, + "loss": 3.3178, + "step": 40100 + }, + { + "epoch": 0.3545629558981967, + "grad_norm": 2.794602155685425, + "learning_rate": 9.882178636121568e-05, + "loss": 3.8837, + "step": 40150 + }, + { + "epoch": 0.3550045037884809, + "grad_norm": 2.1202375888824463, + "learning_rate": 9.881878556331392e-05, + "loss": 3.5493, + "step": 40200 + }, + { + "epoch": 0.3554460516787651, + "grad_norm": 0.8689738512039185, + "learning_rate": 9.881578099457496e-05, + "loss": 3.29, + "step": 40250 + }, + { + "epoch": 0.35588759956904925, + "grad_norm": 2.1677334308624268, + "learning_rate": 9.881277265523085e-05, + "loss": 3.2762, + "step": 40300 + }, + { + "epoch": 0.3563291474593334, + "grad_norm": 1.659857988357544, + "learning_rate": 9.880976054551399e-05, + "loss": 3.8149, + "step": 40350 + }, + { + "epoch": 0.35677069534961764, + "grad_norm": 2.109435558319092, + "learning_rate": 9.880674466565701e-05, + "loss": 3.3179, + "step": 40400 + }, + { + "epoch": 0.3572122432399018, + "grad_norm": 2.959153175354004, + "learning_rate": 9.880372501589289e-05, + "loss": 3.5068, + "step": 40450 + }, + { + "epoch": 0.35765379113018597, + "grad_norm": 0.8862214088439941, + "learning_rate": 9.880070159645483e-05, + "loss": 3.593, + "step": 40500 + }, + { + "epoch": 0.35809533902047014, + "grad_norm": 4.1659464836120605, + "learning_rate": 9.879767440757641e-05, + "loss": 3.1608, + "step": 40550 + }, + { + "epoch": 0.35853688691075436, + "grad_norm": 4.7671799659729, + "learning_rate": 9.879464344949143e-05, + "loss": 3.3211, + "step": 40600 + }, + { + "epoch": 0.3589784348010385, + "grad_norm": 5.715523719787598, + "learning_rate": 9.8791608722434e-05, + "loss": 3.0981, + "step": 40650 + }, + { + "epoch": 0.3594199826913227, + "grad_norm": 1.8158924579620361, + "learning_rate": 9.878857022663855e-05, + "loss": 2.8897, + "step": 40700 + }, + { + "epoch": 0.35986153058160686, + "grad_norm": 2.091034412384033, + "learning_rate": 9.878552796233976e-05, + "loss": 3.6337, + "step": 40750 + }, + { + "epoch": 0.3603030784718911, + "grad_norm": 3.8585503101348877, + "learning_rate": 9.878248192977262e-05, + "loss": 3.5439, + "step": 40800 + }, + { + "epoch": 0.36074462636217525, + "grad_norm": 2.2025511264801025, + "learning_rate": 9.877943212917242e-05, + "loss": 3.3184, + "step": 40850 + }, + { + "epoch": 0.3611861742524594, + "grad_norm": 3.991384744644165, + "learning_rate": 9.877637856077473e-05, + "loss": 3.4019, + "step": 40900 + }, + { + "epoch": 0.3616277221427436, + "grad_norm": 4.2655534744262695, + "learning_rate": 9.87733212248154e-05, + "loss": 2.82, + "step": 40950 + }, + { + "epoch": 0.3620692700330278, + "grad_norm": 1.683498740196228, + "learning_rate": 9.877026012153059e-05, + "loss": 3.3487, + "step": 41000 + }, + { + "epoch": 0.36251081792331197, + "grad_norm": 2.264552593231201, + "learning_rate": 9.876719525115675e-05, + "loss": 2.975, + "step": 41050 + }, + { + "epoch": 0.36295236581359613, + "grad_norm": 2.6153547763824463, + "learning_rate": 9.876412661393061e-05, + "loss": 3.3764, + "step": 41100 + }, + { + "epoch": 0.3633939137038803, + "grad_norm": 1.591480016708374, + "learning_rate": 9.876105421008919e-05, + "loss": 2.965, + "step": 41150 + }, + { + "epoch": 0.3638354615941645, + "grad_norm": 2.9769952297210693, + "learning_rate": 9.875797803986983e-05, + "loss": 3.2887, + "step": 41200 + }, + { + "epoch": 0.3642770094844487, + "grad_norm": 5.074538230895996, + "learning_rate": 9.875489810351011e-05, + "loss": 3.5681, + "step": 41250 + }, + { + "epoch": 0.36471855737473285, + "grad_norm": 2.3003244400024414, + "learning_rate": 9.875181440124794e-05, + "loss": 3.443, + "step": 41300 + }, + { + "epoch": 0.365160105265017, + "grad_norm": 2.3685524463653564, + "learning_rate": 9.874872693332152e-05, + "loss": 3.1724, + "step": 41350 + }, + { + "epoch": 0.36560165315530124, + "grad_norm": 0.6377138495445251, + "learning_rate": 9.874563569996931e-05, + "loss": 3.4932, + "step": 41400 + }, + { + "epoch": 0.3660432010455854, + "grad_norm": 1.721374750137329, + "learning_rate": 9.874254070143009e-05, + "loss": 3.2422, + "step": 41450 + }, + { + "epoch": 0.3664847489358696, + "grad_norm": 2.2348365783691406, + "learning_rate": 9.873944193794294e-05, + "loss": 3.0912, + "step": 41500 + }, + { + "epoch": 0.36692629682615374, + "grad_norm": 1.8013442754745483, + "learning_rate": 9.873633940974719e-05, + "loss": 3.4616, + "step": 41550 + }, + { + "epoch": 0.36736784471643796, + "grad_norm": 1.4249016046524048, + "learning_rate": 9.873323311708249e-05, + "loss": 3.2594, + "step": 41600 + }, + { + "epoch": 0.36780939260672213, + "grad_norm": 1.336929440498352, + "learning_rate": 9.873012306018879e-05, + "loss": 3.4553, + "step": 41650 + }, + { + "epoch": 0.3682509404970063, + "grad_norm": 1.3156746625900269, + "learning_rate": 9.872700923930629e-05, + "loss": 3.5638, + "step": 41700 + }, + { + "epoch": 0.36869248838729046, + "grad_norm": 1.1224370002746582, + "learning_rate": 9.872389165467553e-05, + "loss": 3.131, + "step": 41750 + }, + { + "epoch": 0.3691340362775747, + "grad_norm": 1.9698013067245483, + "learning_rate": 9.87207703065373e-05, + "loss": 3.2832, + "step": 41800 + }, + { + "epoch": 0.36957558416785885, + "grad_norm": 1.916576623916626, + "learning_rate": 9.87176451951327e-05, + "loss": 3.8032, + "step": 41850 + }, + { + "epoch": 0.370017132058143, + "grad_norm": 1.1512713432312012, + "learning_rate": 9.871451632070312e-05, + "loss": 3.4968, + "step": 41900 + }, + { + "epoch": 0.3704586799484272, + "grad_norm": 1.165654182434082, + "learning_rate": 9.871138368349025e-05, + "loss": 3.7091, + "step": 41950 + }, + { + "epoch": 0.3709002278387114, + "grad_norm": 0.8057221174240112, + "learning_rate": 9.870824728373604e-05, + "loss": 2.845, + "step": 42000 + }, + { + "epoch": 0.3709002278387114, + "eval_asr_loss": 0.8905537465911811, + "eval_loss": 3.063842535018921, + "eval_runtime": 20.3425, + "eval_samples_per_second": 37.754, + "eval_steps_per_second": 9.438, + "eval_tts_loss": 5.883050221472156, + "step": 42000 + }, + { + "epoch": 0.37134177572899557, + "grad_norm": 9.080826759338379, + "learning_rate": 9.870510712168278e-05, + "loss": 3.178, + "step": 42050 + }, + { + "epoch": 0.37178332361927974, + "grad_norm": 1.3995147943496704, + "learning_rate": 9.870196319757299e-05, + "loss": 3.316, + "step": 42100 + }, + { + "epoch": 0.3722248715095639, + "grad_norm": 2.64806866645813, + "learning_rate": 9.869887850223219e-05, + "loss": 3.0827, + "step": 42150 + }, + { + "epoch": 0.3726664193998481, + "grad_norm": 0.9530137181282043, + "learning_rate": 9.869572712996722e-05, + "loss": 3.2979, + "step": 42200 + }, + { + "epoch": 0.3731079672901323, + "grad_norm": 2.155907392501831, + "learning_rate": 9.869257199637023e-05, + "loss": 3.2503, + "step": 42250 + }, + { + "epoch": 0.37354951518041646, + "grad_norm": 1.780704379081726, + "learning_rate": 9.868941310168498e-05, + "loss": 3.4232, + "step": 42300 + }, + { + "epoch": 0.3739910630707006, + "grad_norm": 0.6130134463310242, + "learning_rate": 9.868625044615543e-05, + "loss": 3.2393, + "step": 42350 + }, + { + "epoch": 0.37443261096098485, + "grad_norm": 1.8535226583480835, + "learning_rate": 9.868308403002589e-05, + "loss": 3.2783, + "step": 42400 + }, + { + "epoch": 0.374874158851269, + "grad_norm": 3.596348285675049, + "learning_rate": 9.867991385354092e-05, + "loss": 3.157, + "step": 42450 + }, + { + "epoch": 0.3753157067415532, + "grad_norm": 8.728959083557129, + "learning_rate": 9.86767399169454e-05, + "loss": 3.542, + "step": 42500 + }, + { + "epoch": 0.37575725463183735, + "grad_norm": 1.9934968948364258, + "learning_rate": 9.867356222048448e-05, + "loss": 3.4873, + "step": 42550 + }, + { + "epoch": 0.37619880252212157, + "grad_norm": 2.618511915206909, + "learning_rate": 9.867038076440362e-05, + "loss": 3.1409, + "step": 42600 + }, + { + "epoch": 0.37664035041240573, + "grad_norm": 1.344989538192749, + "learning_rate": 9.866719554894857e-05, + "loss": 3.4496, + "step": 42650 + }, + { + "epoch": 0.3770818983026899, + "grad_norm": 1.429974913597107, + "learning_rate": 9.866400657436534e-05, + "loss": 3.1867, + "step": 42700 + }, + { + "epoch": 0.37752344619297407, + "grad_norm": 1.021186351776123, + "learning_rate": 9.866081384090026e-05, + "loss": 2.994, + "step": 42750 + }, + { + "epoch": 0.3779649940832583, + "grad_norm": 2.500567674636841, + "learning_rate": 9.865761734879994e-05, + "loss": 2.9304, + "step": 42800 + }, + { + "epoch": 0.37840654197354245, + "grad_norm": 1.401653528213501, + "learning_rate": 9.865441709831129e-05, + "loss": 3.5495, + "step": 42850 + }, + { + "epoch": 0.3788480898638266, + "grad_norm": 2.189068555831909, + "learning_rate": 9.865121308968151e-05, + "loss": 3.5045, + "step": 42900 + }, + { + "epoch": 0.3792896377541108, + "grad_norm": 1.583028793334961, + "learning_rate": 9.864800532315806e-05, + "loss": 3.4075, + "step": 42950 + }, + { + "epoch": 0.379731185644395, + "grad_norm": 1.778917908668518, + "learning_rate": 9.864479379898871e-05, + "loss": 3.3313, + "step": 43000 + }, + { + "epoch": 0.3801727335346792, + "grad_norm": 2.908151626586914, + "learning_rate": 9.864157851742154e-05, + "loss": 3.1747, + "step": 43050 + }, + { + "epoch": 0.38061428142496334, + "grad_norm": 3.3366377353668213, + "learning_rate": 9.86383594787049e-05, + "loss": 3.5529, + "step": 43100 + }, + { + "epoch": 0.3810558293152475, + "grad_norm": 1.980222225189209, + "learning_rate": 9.863513668308744e-05, + "loss": 3.559, + "step": 43150 + }, + { + "epoch": 0.38149737720553173, + "grad_norm": 2.1785378456115723, + "learning_rate": 9.863191013081807e-05, + "loss": 3.3076, + "step": 43200 + }, + { + "epoch": 0.3819389250958159, + "grad_norm": 3.69520902633667, + "learning_rate": 9.862867982214604e-05, + "loss": 3.4053, + "step": 43250 + }, + { + "epoch": 0.38238047298610006, + "grad_norm": 5.211033821105957, + "learning_rate": 9.862544575732085e-05, + "loss": 3.2504, + "step": 43300 + }, + { + "epoch": 0.38282202087638423, + "grad_norm": 2.920567750930786, + "learning_rate": 9.862220793659232e-05, + "loss": 3.1881, + "step": 43350 + }, + { + "epoch": 0.38326356876666845, + "grad_norm": 2.8354828357696533, + "learning_rate": 9.86189663602105e-05, + "loss": 3.6872, + "step": 43400 + }, + { + "epoch": 0.3837051166569526, + "grad_norm": 3.6911203861236572, + "learning_rate": 9.861572102842583e-05, + "loss": 3.3914, + "step": 43450 + }, + { + "epoch": 0.3841466645472368, + "grad_norm": 3.9869236946105957, + "learning_rate": 9.861247194148895e-05, + "loss": 3.3991, + "step": 43500 + }, + { + "epoch": 0.38458821243752095, + "grad_norm": 2.9851901531219482, + "learning_rate": 9.860928419328401e-05, + "loss": 3.4688, + "step": 43550 + }, + { + "epoch": 0.38502976032780517, + "grad_norm": 5.538595199584961, + "learning_rate": 9.860602767188645e-05, + "loss": 3.3237, + "step": 43600 + }, + { + "epoch": 0.38547130821808934, + "grad_norm": 0.9150497317314148, + "learning_rate": 9.860276739608543e-05, + "loss": 3.2588, + "step": 43650 + }, + { + "epoch": 0.3859128561083735, + "grad_norm": 3.647045612335205, + "learning_rate": 9.859950336613275e-05, + "loss": 3.1377, + "step": 43700 + }, + { + "epoch": 0.38635440399865767, + "grad_norm": 0.6097536087036133, + "learning_rate": 9.859623558228055e-05, + "loss": 3.1102, + "step": 43750 + }, + { + "epoch": 0.3867959518889419, + "grad_norm": 1.7617607116699219, + "learning_rate": 9.859296404478124e-05, + "loss": 2.9313, + "step": 43800 + }, + { + "epoch": 0.38723749977922606, + "grad_norm": 2.4368743896484375, + "learning_rate": 9.858968875388752e-05, + "loss": 3.2671, + "step": 43850 + }, + { + "epoch": 0.3876790476695102, + "grad_norm": 1.931584358215332, + "learning_rate": 9.858640970985237e-05, + "loss": 3.4481, + "step": 43900 + }, + { + "epoch": 0.3881205955597944, + "grad_norm": 1.9381108283996582, + "learning_rate": 9.858312691292906e-05, + "loss": 3.2431, + "step": 43950 + }, + { + "epoch": 0.3885621434500786, + "grad_norm": 2.4838809967041016, + "learning_rate": 9.857984036337118e-05, + "loss": 3.4532, + "step": 44000 + }, + { + "epoch": 0.3890036913403628, + "grad_norm": 2.848560094833374, + "learning_rate": 9.857655006143259e-05, + "loss": 3.3183, + "step": 44050 + }, + { + "epoch": 0.38944523923064694, + "grad_norm": 2.6281092166900635, + "learning_rate": 9.85732560073674e-05, + "loss": 3.348, + "step": 44100 + }, + { + "epoch": 0.3898867871209311, + "grad_norm": 1.351362943649292, + "learning_rate": 9.856995820143006e-05, + "loss": 3.2875, + "step": 44150 + }, + { + "epoch": 0.39032833501121533, + "grad_norm": 11.539451599121094, + "learning_rate": 9.856665664387535e-05, + "loss": 3.8226, + "step": 44200 + }, + { + "epoch": 0.3907698829014995, + "grad_norm": 2.561694383621216, + "learning_rate": 9.856335133495822e-05, + "loss": 3.7136, + "step": 44250 + }, + { + "epoch": 0.39121143079178367, + "grad_norm": 1.2988532781600952, + "learning_rate": 9.856004227493402e-05, + "loss": 3.5633, + "step": 44300 + }, + { + "epoch": 0.39165297868206783, + "grad_norm": 2.9327540397644043, + "learning_rate": 9.855672946405831e-05, + "loss": 3.2624, + "step": 44350 + }, + { + "epoch": 0.39209452657235205, + "grad_norm": 2.4923012256622314, + "learning_rate": 9.855341290258703e-05, + "loss": 3.0418, + "step": 44400 + }, + { + "epoch": 0.3925360744626362, + "grad_norm": 0.7568151354789734, + "learning_rate": 9.85500925907763e-05, + "loss": 3.5547, + "step": 44450 + }, + { + "epoch": 0.3929776223529204, + "grad_norm": 1.1965677738189697, + "learning_rate": 9.854676852888262e-05, + "loss": 3.4921, + "step": 44500 + }, + { + "epoch": 0.39341917024320455, + "grad_norm": 0.5616858601570129, + "learning_rate": 9.854344071716272e-05, + "loss": 3.2839, + "step": 44550 + }, + { + "epoch": 0.3938607181334888, + "grad_norm": 3.225074052810669, + "learning_rate": 9.854010915587369e-05, + "loss": 3.4744, + "step": 44600 + }, + { + "epoch": 0.39430226602377294, + "grad_norm": 3.9845869541168213, + "learning_rate": 9.853677384527282e-05, + "loss": 3.4246, + "step": 44650 + }, + { + "epoch": 0.3947438139140571, + "grad_norm": 3.029383897781372, + "learning_rate": 9.853343478561773e-05, + "loss": 3.408, + "step": 44700 + }, + { + "epoch": 0.3951853618043413, + "grad_norm": 1.8023637533187866, + "learning_rate": 9.853009197716638e-05, + "loss": 3.1036, + "step": 44750 + }, + { + "epoch": 0.3956269096946255, + "grad_norm": 0.8640767931938171, + "learning_rate": 9.852674542017694e-05, + "loss": 3.3368, + "step": 44800 + }, + { + "epoch": 0.39606845758490966, + "grad_norm": 0.6954994797706604, + "learning_rate": 9.852339511490792e-05, + "loss": 2.9964, + "step": 44850 + }, + { + "epoch": 0.3965100054751938, + "grad_norm": 8.39139461517334, + "learning_rate": 9.85200410616181e-05, + "loss": 2.9249, + "step": 44900 + }, + { + "epoch": 0.396951553365478, + "grad_norm": 4.7654900550842285, + "learning_rate": 9.851668326056652e-05, + "loss": 3.0992, + "step": 44950 + }, + { + "epoch": 0.3973931012557622, + "grad_norm": 1.2103265523910522, + "learning_rate": 9.851332171201258e-05, + "loss": 3.3672, + "step": 45000 + }, + { + "epoch": 0.3973931012557622, + "eval_asr_loss": 0.8913320451830068, + "eval_loss": 3.0676581859588623, + "eval_runtime": 20.2524, + "eval_samples_per_second": 37.921, + "eval_steps_per_second": 9.48, + "eval_tts_loss": 5.97925205245277, + "step": 45000 + }, + { + "epoch": 0.3978346491460464, + "grad_norm": 2.4184420108795166, + "learning_rate": 9.850995641621592e-05, + "loss": 3.6283, + "step": 45050 + }, + { + "epoch": 0.39827619703633055, + "grad_norm": 1.4427285194396973, + "learning_rate": 9.850658737343648e-05, + "loss": 3.5224, + "step": 45100 + }, + { + "epoch": 0.3987177449266147, + "grad_norm": 6.392005920410156, + "learning_rate": 9.850321458393447e-05, + "loss": 3.0993, + "step": 45150 + }, + { + "epoch": 0.39915929281689894, + "grad_norm": 1.4388530254364014, + "learning_rate": 9.849983804797046e-05, + "loss": 3.0099, + "step": 45200 + }, + { + "epoch": 0.3996008407071831, + "grad_norm": 2.679241180419922, + "learning_rate": 9.849645776580522e-05, + "loss": 3.3363, + "step": 45250 + }, + { + "epoch": 0.40004238859746727, + "grad_norm": 1.61045503616333, + "learning_rate": 9.849307373769985e-05, + "loss": 3.1612, + "step": 45300 + }, + { + "epoch": 0.40048393648775144, + "grad_norm": 1.9430288076400757, + "learning_rate": 9.848968596391574e-05, + "loss": 3.4328, + "step": 45350 + }, + { + "epoch": 0.40092548437803566, + "grad_norm": 1.6506637334823608, + "learning_rate": 9.848629444471456e-05, + "loss": 3.6278, + "step": 45400 + }, + { + "epoch": 0.4013670322683198, + "grad_norm": 2.5751986503601074, + "learning_rate": 9.848289918035831e-05, + "loss": 3.3468, + "step": 45450 + }, + { + "epoch": 0.401808580158604, + "grad_norm": 2.1925528049468994, + "learning_rate": 9.84795001711092e-05, + "loss": 3.3276, + "step": 45500 + }, + { + "epoch": 0.40225012804888816, + "grad_norm": 3.269740581512451, + "learning_rate": 9.847609741722981e-05, + "loss": 2.7999, + "step": 45550 + }, + { + "epoch": 0.4026916759391724, + "grad_norm": 4.853845596313477, + "learning_rate": 9.847269091898294e-05, + "loss": 3.1533, + "step": 45600 + }, + { + "epoch": 0.40313322382945654, + "grad_norm": 2.3576786518096924, + "learning_rate": 9.846928067663175e-05, + "loss": 3.4268, + "step": 45650 + }, + { + "epoch": 0.4035747717197407, + "grad_norm": 1.3526619672775269, + "learning_rate": 9.846586669043964e-05, + "loss": 3.156, + "step": 45700 + }, + { + "epoch": 0.4040163196100249, + "grad_norm": 1.5202453136444092, + "learning_rate": 9.846244896067031e-05, + "loss": 3.2914, + "step": 45750 + }, + { + "epoch": 0.4044578675003091, + "grad_norm": 1.80320143699646, + "learning_rate": 9.845902748758773e-05, + "loss": 3.2984, + "step": 45800 + }, + { + "epoch": 0.40489941539059326, + "grad_norm": 1.7634551525115967, + "learning_rate": 9.845560227145621e-05, + "loss": 3.1742, + "step": 45850 + }, + { + "epoch": 0.40534096328087743, + "grad_norm": 2.447131633758545, + "learning_rate": 9.84521733125403e-05, + "loss": 3.0532, + "step": 45900 + }, + { + "epoch": 0.4057825111711616, + "grad_norm": 2.7054214477539062, + "learning_rate": 9.844874061110485e-05, + "loss": 3.5135, + "step": 45950 + }, + { + "epoch": 0.4062240590614458, + "grad_norm": 0.9877326488494873, + "learning_rate": 9.844530416741505e-05, + "loss": 3.4712, + "step": 46000 + }, + { + "epoch": 0.40666560695173, + "grad_norm": 1.2479236125946045, + "learning_rate": 9.84418639817363e-05, + "loss": 3.5099, + "step": 46050 + }, + { + "epoch": 0.40710715484201415, + "grad_norm": 3.176802635192871, + "learning_rate": 9.843842005433431e-05, + "loss": 3.5577, + "step": 46100 + }, + { + "epoch": 0.4075487027322983, + "grad_norm": 1.3120348453521729, + "learning_rate": 9.843497238547513e-05, + "loss": 2.9955, + "step": 46150 + }, + { + "epoch": 0.40799025062258254, + "grad_norm": 1.1844558715820312, + "learning_rate": 9.843152097542505e-05, + "loss": 3.6275, + "step": 46200 + }, + { + "epoch": 0.4084317985128667, + "grad_norm": 2.701456308364868, + "learning_rate": 9.842806582445067e-05, + "loss": 3.8021, + "step": 46250 + }, + { + "epoch": 0.4088733464031509, + "grad_norm": 2.6151137351989746, + "learning_rate": 9.842460693281885e-05, + "loss": 3.6685, + "step": 46300 + }, + { + "epoch": 0.40931489429343504, + "grad_norm": 2.728315591812134, + "learning_rate": 9.842114430079679e-05, + "loss": 3.4043, + "step": 46350 + }, + { + "epoch": 0.40975644218371926, + "grad_norm": 2.2898852825164795, + "learning_rate": 9.841767792865191e-05, + "loss": 3.6381, + "step": 46400 + }, + { + "epoch": 0.4101979900740034, + "grad_norm": 1.9836546182632446, + "learning_rate": 9.8414207816652e-05, + "loss": 3.0714, + "step": 46450 + }, + { + "epoch": 0.4106395379642876, + "grad_norm": 2.8951032161712646, + "learning_rate": 9.841073396506506e-05, + "loss": 3.4733, + "step": 46500 + }, + { + "epoch": 0.41108108585457176, + "grad_norm": 1.4405927658081055, + "learning_rate": 9.840725637415945e-05, + "loss": 3.2872, + "step": 46550 + }, + { + "epoch": 0.411522633744856, + "grad_norm": 3.1269359588623047, + "learning_rate": 9.840377504420376e-05, + "loss": 3.4911, + "step": 46600 + }, + { + "epoch": 0.41196418163514015, + "grad_norm": 3.698716163635254, + "learning_rate": 9.840028997546689e-05, + "loss": 3.1268, + "step": 46650 + }, + { + "epoch": 0.4124057295254243, + "grad_norm": 2.955763578414917, + "learning_rate": 9.839680116821804e-05, + "loss": 3.5459, + "step": 46700 + }, + { + "epoch": 0.4128472774157085, + "grad_norm": 0.848429799079895, + "learning_rate": 9.83933086227267e-05, + "loss": 3.2194, + "step": 46750 + }, + { + "epoch": 0.4132888253059927, + "grad_norm": 8.197107315063477, + "learning_rate": 9.838981233926264e-05, + "loss": 3.2381, + "step": 46800 + }, + { + "epoch": 0.41373037319627687, + "grad_norm": 1.4284354448318481, + "learning_rate": 9.83863123180959e-05, + "loss": 3.4783, + "step": 46850 + }, + { + "epoch": 0.41417192108656103, + "grad_norm": 1.3862097263336182, + "learning_rate": 9.838280855949684e-05, + "loss": 3.3749, + "step": 46900 + }, + { + "epoch": 0.41461346897684526, + "grad_norm": 1.5118871927261353, + "learning_rate": 9.837930106373609e-05, + "loss": 3.2749, + "step": 46950 + }, + { + "epoch": 0.4150550168671294, + "grad_norm": 4.723493576049805, + "learning_rate": 9.83757898310846e-05, + "loss": 2.9933, + "step": 47000 + }, + { + "epoch": 0.4154965647574136, + "grad_norm": 6.659562110900879, + "learning_rate": 9.837227486181353e-05, + "loss": 3.4323, + "step": 47050 + }, + { + "epoch": 0.41593811264769776, + "grad_norm": 2.5123307704925537, + "learning_rate": 9.836875615619443e-05, + "loss": 2.9715, + "step": 47100 + }, + { + "epoch": 0.416379660537982, + "grad_norm": 2.6539173126220703, + "learning_rate": 9.836523371449906e-05, + "loss": 3.1207, + "step": 47150 + }, + { + "epoch": 0.41682120842826614, + "grad_norm": 3.233375072479248, + "learning_rate": 9.836170753699952e-05, + "loss": 3.1869, + "step": 47200 + }, + { + "epoch": 0.4172627563185503, + "grad_norm": 1.2898887395858765, + "learning_rate": 9.835817762396816e-05, + "loss": 3.2155, + "step": 47250 + }, + { + "epoch": 0.4177043042088345, + "grad_norm": 0.9073876738548279, + "learning_rate": 9.835464397567765e-05, + "loss": 3.363, + "step": 47300 + }, + { + "epoch": 0.4181458520991187, + "grad_norm": 2.146453380584717, + "learning_rate": 9.835110659240092e-05, + "loss": 3.6663, + "step": 47350 + }, + { + "epoch": 0.41858739998940286, + "grad_norm": 4.763782501220703, + "learning_rate": 9.834756547441122e-05, + "loss": 3.5419, + "step": 47400 + }, + { + "epoch": 0.41902894787968703, + "grad_norm": 2.6837880611419678, + "learning_rate": 9.834402062198207e-05, + "loss": 3.134, + "step": 47450 + }, + { + "epoch": 0.4194704957699712, + "grad_norm": 2.5617268085479736, + "learning_rate": 9.834047203538726e-05, + "loss": 3.4525, + "step": 47500 + }, + { + "epoch": 0.4199120436602554, + "grad_norm": 2.1954944133758545, + "learning_rate": 9.83369197149009e-05, + "loss": 3.7715, + "step": 47550 + }, + { + "epoch": 0.4203535915505396, + "grad_norm": 1.5438072681427002, + "learning_rate": 9.833336366079737e-05, + "loss": 3.6781, + "step": 47600 + }, + { + "epoch": 0.42079513944082375, + "grad_norm": 1.757779598236084, + "learning_rate": 9.832980387335138e-05, + "loss": 3.4219, + "step": 47650 + }, + { + "epoch": 0.4212366873311079, + "grad_norm": 2.8832337856292725, + "learning_rate": 9.832624035283785e-05, + "loss": 2.9347, + "step": 47700 + }, + { + "epoch": 0.42167823522139214, + "grad_norm": 2.0708539485931396, + "learning_rate": 9.832274448117773e-05, + "loss": 3.1908, + "step": 47750 + }, + { + "epoch": 0.4221197831116763, + "grad_norm": 1.883726716041565, + "learning_rate": 9.831917357000283e-05, + "loss": 3.5663, + "step": 47800 + }, + { + "epoch": 0.42256133100196047, + "grad_norm": 1.2137542963027954, + "learning_rate": 9.83155989265815e-05, + "loss": 3.3599, + "step": 47850 + }, + { + "epoch": 0.42300287889224464, + "grad_norm": 1.1021403074264526, + "learning_rate": 9.831202055118986e-05, + "loss": 3.3176, + "step": 47900 + }, + { + "epoch": 0.42344442678252886, + "grad_norm": 3.1852636337280273, + "learning_rate": 9.830843844410431e-05, + "loss": 3.2934, + "step": 47950 + }, + { + "epoch": 0.423885974672813, + "grad_norm": 2.2362117767333984, + "learning_rate": 9.830492435893767e-05, + "loss": 3.5098, + "step": 48000 + }, + { + "epoch": 0.423885974672813, + "eval_asr_loss": 0.8946412356905893, + "eval_loss": 3.0450656414031982, + "eval_runtime": 20.5825, + "eval_samples_per_second": 37.313, + "eval_steps_per_second": 9.328, + "eval_tts_loss": 5.9677758015827616, + "step": 48000 + }, + { + "epoch": 0.4243275225630972, + "grad_norm": 0.6300576329231262, + "learning_rate": 9.830133486391473e-05, + "loss": 3.5818, + "step": 48050 + }, + { + "epoch": 0.42476907045338136, + "grad_norm": 1.3167023658752441, + "learning_rate": 9.829774163802325e-05, + "loss": 3.4292, + "step": 48100 + }, + { + "epoch": 0.4252106183436656, + "grad_norm": 2.811063051223755, + "learning_rate": 9.829414468154079e-05, + "loss": 3.7977, + "step": 48150 + }, + { + "epoch": 0.42565216623394975, + "grad_norm": 2.503530502319336, + "learning_rate": 9.829054399474516e-05, + "loss": 3.6608, + "step": 48200 + }, + { + "epoch": 0.4260937141242339, + "grad_norm": 2.051328420639038, + "learning_rate": 9.828693957791452e-05, + "loss": 3.3447, + "step": 48250 + }, + { + "epoch": 0.4265352620145181, + "grad_norm": 0.9550831913948059, + "learning_rate": 9.828333143132724e-05, + "loss": 3.2968, + "step": 48300 + }, + { + "epoch": 0.4269768099048023, + "grad_norm": 3.1451058387756348, + "learning_rate": 9.827971955526205e-05, + "loss": 3.3196, + "step": 48350 + }, + { + "epoch": 0.42741835779508647, + "grad_norm": 1.7226669788360596, + "learning_rate": 9.827610394999791e-05, + "loss": 3.1922, + "step": 48400 + }, + { + "epoch": 0.42785990568537063, + "grad_norm": 2.289461851119995, + "learning_rate": 9.827248461581412e-05, + "loss": 3.327, + "step": 48450 + }, + { + "epoch": 0.4283014535756548, + "grad_norm": 3.5129716396331787, + "learning_rate": 9.826886155299023e-05, + "loss": 3.4937, + "step": 48500 + }, + { + "epoch": 0.428743001465939, + "grad_norm": 2.9753642082214355, + "learning_rate": 9.826523476180609e-05, + "loss": 3.3748, + "step": 48550 + }, + { + "epoch": 0.4291845493562232, + "grad_norm": 1.0942645072937012, + "learning_rate": 9.826160424254185e-05, + "loss": 3.391, + "step": 48600 + }, + { + "epoch": 0.42962609724650735, + "grad_norm": 2.1059162616729736, + "learning_rate": 9.825796999547792e-05, + "loss": 3.7639, + "step": 48650 + }, + { + "epoch": 0.4300676451367915, + "grad_norm": 1.9335929155349731, + "learning_rate": 9.825433202089503e-05, + "loss": 2.9592, + "step": 48700 + }, + { + "epoch": 0.43050919302707574, + "grad_norm": 0.7627319693565369, + "learning_rate": 9.825069031907416e-05, + "loss": 3.3292, + "step": 48750 + }, + { + "epoch": 0.4309507409173599, + "grad_norm": 3.433350086212158, + "learning_rate": 9.824704489029664e-05, + "loss": 3.3546, + "step": 48800 + }, + { + "epoch": 0.4313922888076441, + "grad_norm": 2.693155527114868, + "learning_rate": 9.824339573484402e-05, + "loss": 3.4344, + "step": 48850 + }, + { + "epoch": 0.43183383669792824, + "grad_norm": 1.1923327445983887, + "learning_rate": 9.823974285299817e-05, + "loss": 3.3486, + "step": 48900 + }, + { + "epoch": 0.43227538458821246, + "grad_norm": 1.9175432920455933, + "learning_rate": 9.823608624504123e-05, + "loss": 3.4314, + "step": 48950 + }, + { + "epoch": 0.43271693247849663, + "grad_norm": 0.7622149586677551, + "learning_rate": 9.823242591125568e-05, + "loss": 3.1681, + "step": 49000 + }, + { + "epoch": 0.4331584803687808, + "grad_norm": 2.3033721446990967, + "learning_rate": 9.822876185192421e-05, + "loss": 3.4592, + "step": 49050 + }, + { + "epoch": 0.43360002825906496, + "grad_norm": 0.886083722114563, + "learning_rate": 9.822509406732987e-05, + "loss": 3.6985, + "step": 49100 + }, + { + "epoch": 0.4340415761493492, + "grad_norm": 6.617265701293945, + "learning_rate": 9.822142255775594e-05, + "loss": 3.1759, + "step": 49150 + }, + { + "epoch": 0.43448312403963335, + "grad_norm": 4.3931121826171875, + "learning_rate": 9.821774732348602e-05, + "loss": 3.4327, + "step": 49200 + }, + { + "epoch": 0.4349246719299175, + "grad_norm": 1.2443506717681885, + "learning_rate": 9.8214068364804e-05, + "loss": 3.5825, + "step": 49250 + }, + { + "epoch": 0.4353662198202017, + "grad_norm": 2.791949510574341, + "learning_rate": 9.821038568199403e-05, + "loss": 3.1509, + "step": 49300 + }, + { + "epoch": 0.4358077677104859, + "grad_norm": 1.996881127357483, + "learning_rate": 9.820669927534061e-05, + "loss": 2.9644, + "step": 49350 + }, + { + "epoch": 0.43624931560077007, + "grad_norm": 1.907235026359558, + "learning_rate": 9.820300914512842e-05, + "loss": 3.3867, + "step": 49400 + }, + { + "epoch": 0.43669086349105424, + "grad_norm": 3.5670182704925537, + "learning_rate": 9.819931529164254e-05, + "loss": 3.4446, + "step": 49450 + }, + { + "epoch": 0.4371324113813384, + "grad_norm": 1.6872813701629639, + "learning_rate": 9.819561771516826e-05, + "loss": 2.9724, + "step": 49500 + }, + { + "epoch": 0.4375739592716226, + "grad_norm": 3.4050118923187256, + "learning_rate": 9.819191641599121e-05, + "loss": 3.2879, + "step": 49550 + }, + { + "epoch": 0.4380155071619068, + "grad_norm": 2.1053950786590576, + "learning_rate": 9.818821139439727e-05, + "loss": 3.4533, + "step": 49600 + }, + { + "epoch": 0.43845705505219096, + "grad_norm": 1.7744102478027344, + "learning_rate": 9.818450265067261e-05, + "loss": 3.3481, + "step": 49650 + }, + { + "epoch": 0.4388986029424751, + "grad_norm": 2.2178120613098145, + "learning_rate": 9.818079018510374e-05, + "loss": 3.3662, + "step": 49700 + }, + { + "epoch": 0.43934015083275935, + "grad_norm": 2.2515435218811035, + "learning_rate": 9.817707399797736e-05, + "loss": 3.336, + "step": 49750 + }, + { + "epoch": 0.4397816987230435, + "grad_norm": 3.4062695503234863, + "learning_rate": 9.817335408958056e-05, + "loss": 3.2625, + "step": 49800 + }, + { + "epoch": 0.4402232466133277, + "grad_norm": 1.6740872859954834, + "learning_rate": 9.816963046020065e-05, + "loss": 3.7033, + "step": 49850 + }, + { + "epoch": 0.44066479450361185, + "grad_norm": 0.949266254901886, + "learning_rate": 9.816590311012525e-05, + "loss": 3.2875, + "step": 49900 + }, + { + "epoch": 0.44110634239389607, + "grad_norm": 2.8111281394958496, + "learning_rate": 9.816217203964228e-05, + "loss": 3.2595, + "step": 49950 + }, + { + "epoch": 0.44154789028418023, + "grad_norm": 1.117742657661438, + "learning_rate": 9.815843724903993e-05, + "loss": 3.2435, + "step": 50000 + }, + { + "epoch": 0.4419894381744644, + "grad_norm": 2.1841189861297607, + "learning_rate": 9.815469873860666e-05, + "loss": 3.5207, + "step": 50050 + }, + { + "epoch": 0.44243098606474857, + "grad_norm": 4.919894695281982, + "learning_rate": 9.815095650863124e-05, + "loss": 3.3703, + "step": 50100 + }, + { + "epoch": 0.4428725339550328, + "grad_norm": 3.6214163303375244, + "learning_rate": 9.814721055940275e-05, + "loss": 3.5246, + "step": 50150 + }, + { + "epoch": 0.44331408184531695, + "grad_norm": 3.0915398597717285, + "learning_rate": 9.814346089121053e-05, + "loss": 3.7959, + "step": 50200 + }, + { + "epoch": 0.4437556297356011, + "grad_norm": 1.7853915691375732, + "learning_rate": 9.813970750434419e-05, + "loss": 3.4, + "step": 50250 + }, + { + "epoch": 0.4441971776258853, + "grad_norm": 0.6693574786186218, + "learning_rate": 9.813595039909367e-05, + "loss": 3.2674, + "step": 50300 + }, + { + "epoch": 0.4446387255161695, + "grad_norm": 7.433414936065674, + "learning_rate": 9.813218957574914e-05, + "loss": 3.1686, + "step": 50350 + }, + { + "epoch": 0.4450802734064537, + "grad_norm": 2.307650566101074, + "learning_rate": 9.812842503460114e-05, + "loss": 3.2572, + "step": 50400 + }, + { + "epoch": 0.44552182129673784, + "grad_norm": 4.549615383148193, + "learning_rate": 9.812465677594041e-05, + "loss": 3.5429, + "step": 50450 + }, + { + "epoch": 0.445963369187022, + "grad_norm": 1.0888493061065674, + "learning_rate": 9.812088480005804e-05, + "loss": 3.3837, + "step": 50500 + }, + { + "epoch": 0.44640491707730623, + "grad_norm": 2.1504311561584473, + "learning_rate": 9.811718465752564e-05, + "loss": 3.7276, + "step": 50550 + }, + { + "epoch": 0.4468464649675904, + "grad_norm": 3.0695831775665283, + "learning_rate": 9.811340532240422e-05, + "loss": 3.4433, + "step": 50600 + }, + { + "epoch": 0.44728801285787456, + "grad_norm": 1.7197152376174927, + "learning_rate": 9.810962227093022e-05, + "loss": 3.5762, + "step": 50650 + }, + { + "epoch": 0.44772956074815873, + "grad_norm": 1.3767517805099487, + "learning_rate": 9.810583550339588e-05, + "loss": 3.0935, + "step": 50700 + }, + { + "epoch": 0.44817110863844295, + "grad_norm": 1.5721153020858765, + "learning_rate": 9.810204502009364e-05, + "loss": 3.4474, + "step": 50750 + }, + { + "epoch": 0.4486126565287271, + "grad_norm": 4.190256118774414, + "learning_rate": 9.809825082131633e-05, + "loss": 3.1628, + "step": 50800 + }, + { + "epoch": 0.4490542044190113, + "grad_norm": 2.406953811645508, + "learning_rate": 9.809445290735702e-05, + "loss": 3.2441, + "step": 50850 + }, + { + "epoch": 0.44949575230929545, + "grad_norm": 4.159549236297607, + "learning_rate": 9.809065127850903e-05, + "loss": 3.0559, + "step": 50900 + }, + { + "epoch": 0.44993730019957967, + "grad_norm": 1.8754595518112183, + "learning_rate": 9.808684593506605e-05, + "loss": 3.7703, + "step": 50950 + }, + { + "epoch": 0.45037884808986384, + "grad_norm": 1.29695725440979, + "learning_rate": 9.808303687732196e-05, + "loss": 3.1498, + "step": 51000 + }, + { + "epoch": 0.45037884808986384, + "eval_asr_loss": 0.9004947789321158, + "eval_loss": 3.0421018600463867, + "eval_runtime": 20.6473, + "eval_samples_per_second": 37.196, + "eval_steps_per_second": 9.299, + "eval_tts_loss": 5.865643483398774, + "step": 51000 + }, + { + "epoch": 0.450820395980148, + "grad_norm": 2.4158124923706055, + "learning_rate": 9.807922410557102e-05, + "loss": 3.3356, + "step": 51050 + }, + { + "epoch": 0.45126194387043217, + "grad_norm": 0.9025440812110901, + "learning_rate": 9.807540762010772e-05, + "loss": 3.336, + "step": 51100 + }, + { + "epoch": 0.4517034917607164, + "grad_norm": 5.0518364906311035, + "learning_rate": 9.807158742122684e-05, + "loss": 3.2288, + "step": 51150 + }, + { + "epoch": 0.45214503965100056, + "grad_norm": 1.4660353660583496, + "learning_rate": 9.806776350922346e-05, + "loss": 3.0611, + "step": 51200 + }, + { + "epoch": 0.4525865875412847, + "grad_norm": 1.7514241933822632, + "learning_rate": 9.806393588439297e-05, + "loss": 3.304, + "step": 51250 + }, + { + "epoch": 0.4530281354315689, + "grad_norm": 1.4853622913360596, + "learning_rate": 9.806010454703099e-05, + "loss": 3.436, + "step": 51300 + }, + { + "epoch": 0.4534696833218531, + "grad_norm": 2.174837112426758, + "learning_rate": 9.805626949743347e-05, + "loss": 3.6087, + "step": 51350 + }, + { + "epoch": 0.4539112312121373, + "grad_norm": 1.5965477228164673, + "learning_rate": 9.805243073589665e-05, + "loss": 3.4667, + "step": 51400 + }, + { + "epoch": 0.45435277910242144, + "grad_norm": 2.395080804824829, + "learning_rate": 9.804858826271703e-05, + "loss": 3.5994, + "step": 51450 + }, + { + "epoch": 0.4547943269927056, + "grad_norm": 4.734639644622803, + "learning_rate": 9.804474207819139e-05, + "loss": 3.2731, + "step": 51500 + }, + { + "epoch": 0.45523587488298983, + "grad_norm": 1.91262948513031, + "learning_rate": 9.804089218261684e-05, + "loss": 3.4359, + "step": 51550 + }, + { + "epoch": 0.455677422773274, + "grad_norm": 4.3235321044921875, + "learning_rate": 9.803703857629075e-05, + "loss": 3.6903, + "step": 51600 + }, + { + "epoch": 0.45611897066355817, + "grad_norm": 2.399240016937256, + "learning_rate": 9.803318125951075e-05, + "loss": 3.1958, + "step": 51650 + }, + { + "epoch": 0.45656051855384233, + "grad_norm": 3.3152384757995605, + "learning_rate": 9.802932023257483e-05, + "loss": 3.4387, + "step": 51700 + }, + { + "epoch": 0.45700206644412655, + "grad_norm": 3.167579412460327, + "learning_rate": 9.802545549578119e-05, + "loss": 3.6247, + "step": 51750 + }, + { + "epoch": 0.4574436143344107, + "grad_norm": 1.6194385290145874, + "learning_rate": 9.802158704942837e-05, + "loss": 3.3378, + "step": 51800 + }, + { + "epoch": 0.4578851622246949, + "grad_norm": 1.213178277015686, + "learning_rate": 9.801771489381515e-05, + "loss": 3.0608, + "step": 51850 + }, + { + "epoch": 0.45832671011497905, + "grad_norm": 4.134045600891113, + "learning_rate": 9.8013916582878e-05, + "loss": 3.1588, + "step": 51900 + }, + { + "epoch": 0.4587682580052633, + "grad_norm": 3.060990571975708, + "learning_rate": 9.801003708381187e-05, + "loss": 3.3848, + "step": 51950 + }, + { + "epoch": 0.45920980589554744, + "grad_norm": 2.216519594192505, + "learning_rate": 9.800615387637748e-05, + "loss": 3.2701, + "step": 52000 + }, + { + "epoch": 0.4596513537858316, + "grad_norm": 2.748791456222534, + "learning_rate": 9.80022669608748e-05, + "loss": 3.6429, + "step": 52050 + }, + { + "epoch": 0.4600929016761158, + "grad_norm": 2.1832919120788574, + "learning_rate": 9.799837633760403e-05, + "loss": 3.4686, + "step": 52100 + }, + { + "epoch": 0.4605344495664, + "grad_norm": 10.31329345703125, + "learning_rate": 9.799448200686569e-05, + "loss": 3.4104, + "step": 52150 + }, + { + "epoch": 0.46097599745668416, + "grad_norm": 3.921966791152954, + "learning_rate": 9.799058396896062e-05, + "loss": 3.1034, + "step": 52200 + }, + { + "epoch": 0.4614175453469683, + "grad_norm": 1.399453043937683, + "learning_rate": 9.798668222418988e-05, + "loss": 2.6928, + "step": 52250 + }, + { + "epoch": 0.4618590932372525, + "grad_norm": 2.0908732414245605, + "learning_rate": 9.798277677285483e-05, + "loss": 3.2176, + "step": 52300 + }, + { + "epoch": 0.4623006411275367, + "grad_norm": 1.981628656387329, + "learning_rate": 9.797886761525719e-05, + "loss": 3.1046, + "step": 52350 + }, + { + "epoch": 0.4627421890178209, + "grad_norm": 2.1923952102661133, + "learning_rate": 9.797495475169886e-05, + "loss": 2.9833, + "step": 52400 + }, + { + "epoch": 0.46318373690810505, + "grad_norm": 4.177052021026611, + "learning_rate": 9.79710381824821e-05, + "loss": 2.9391, + "step": 52450 + }, + { + "epoch": 0.4636252847983892, + "grad_norm": 1.3210045099258423, + "learning_rate": 9.796711790790941e-05, + "loss": 3.5919, + "step": 52500 + }, + { + "epoch": 0.46406683268867344, + "grad_norm": 1.5212137699127197, + "learning_rate": 9.796319392828361e-05, + "loss": 3.4421, + "step": 52550 + }, + { + "epoch": 0.4645083805789576, + "grad_norm": 9.704986572265625, + "learning_rate": 9.795926624390781e-05, + "loss": 3.2297, + "step": 52600 + }, + { + "epoch": 0.46494992846924177, + "grad_norm": 0.9471487402915955, + "learning_rate": 9.795533485508537e-05, + "loss": 3.6265, + "step": 52650 + }, + { + "epoch": 0.46539147635952594, + "grad_norm": 2.34822154045105, + "learning_rate": 9.795139976211996e-05, + "loss": 3.2869, + "step": 52700 + }, + { + "epoch": 0.46583302424981016, + "grad_norm": 1.4003323316574097, + "learning_rate": 9.794746096531554e-05, + "loss": 3.215, + "step": 52750 + }, + { + "epoch": 0.4662745721400943, + "grad_norm": 0.5006351470947266, + "learning_rate": 9.794351846497634e-05, + "loss": 3.3245, + "step": 52800 + }, + { + "epoch": 0.4667161200303785, + "grad_norm": 1.4579681158065796, + "learning_rate": 9.793957226140688e-05, + "loss": 3.0416, + "step": 52850 + }, + { + "epoch": 0.46715766792066266, + "grad_norm": 2.5864365100860596, + "learning_rate": 9.793562235491198e-05, + "loss": 3.2932, + "step": 52900 + }, + { + "epoch": 0.4675992158109469, + "grad_norm": 1.7026253938674927, + "learning_rate": 9.793166874579673e-05, + "loss": 3.1959, + "step": 52950 + }, + { + "epoch": 0.46804076370123104, + "grad_norm": 5.3310394287109375, + "learning_rate": 9.792771143436654e-05, + "loss": 3.2207, + "step": 53000 + }, + { + "epoch": 0.4684823115915152, + "grad_norm": 1.704520583152771, + "learning_rate": 9.792375042092704e-05, + "loss": 3.1596, + "step": 53050 + }, + { + "epoch": 0.4689238594817994, + "grad_norm": 2.3095810413360596, + "learning_rate": 9.79197857057842e-05, + "loss": 3.48, + "step": 53100 + }, + { + "epoch": 0.4693654073720836, + "grad_norm": 3.7878098487854004, + "learning_rate": 9.791581728924427e-05, + "loss": 3.4366, + "step": 53150 + }, + { + "epoch": 0.46980695526236776, + "grad_norm": 1.6754798889160156, + "learning_rate": 9.791184517161377e-05, + "loss": 3.1798, + "step": 53200 + }, + { + "epoch": 0.47024850315265193, + "grad_norm": 2.6600844860076904, + "learning_rate": 9.79078693531995e-05, + "loss": 3.378, + "step": 53250 + }, + { + "epoch": 0.4706900510429361, + "grad_norm": 1.902320146560669, + "learning_rate": 9.790388983430859e-05, + "loss": 2.8402, + "step": 53300 + }, + { + "epoch": 0.4711315989332203, + "grad_norm": 1.9587918519973755, + "learning_rate": 9.789990661524839e-05, + "loss": 3.6821, + "step": 53350 + }, + { + "epoch": 0.4715731468235045, + "grad_norm": 0.7448610067367554, + "learning_rate": 9.789591969632658e-05, + "loss": 3.3176, + "step": 53400 + }, + { + "epoch": 0.47201469471378865, + "grad_norm": 4.580758094787598, + "learning_rate": 9.789192907785111e-05, + "loss": 3.2808, + "step": 53450 + }, + { + "epoch": 0.4724562426040728, + "grad_norm": 2.524980068206787, + "learning_rate": 9.788793476013025e-05, + "loss": 3.3751, + "step": 53500 + }, + { + "epoch": 0.47289779049435704, + "grad_norm": 3.0498645305633545, + "learning_rate": 9.788393674347249e-05, + "loss": 3.1801, + "step": 53550 + }, + { + "epoch": 0.4733393383846412, + "grad_norm": 2.230107545852661, + "learning_rate": 9.787993502818669e-05, + "loss": 3.7616, + "step": 53600 + }, + { + "epoch": 0.4737808862749254, + "grad_norm": 0.8255630731582642, + "learning_rate": 9.78759296145819e-05, + "loss": 3.0498, + "step": 53650 + }, + { + "epoch": 0.47422243416520954, + "grad_norm": 4.1301727294921875, + "learning_rate": 9.787192050296752e-05, + "loss": 3.281, + "step": 53700 + }, + { + "epoch": 0.47466398205549376, + "grad_norm": 3.9466300010681152, + "learning_rate": 9.786790769365322e-05, + "loss": 3.1479, + "step": 53750 + }, + { + "epoch": 0.4751055299457779, + "grad_norm": 1.8435554504394531, + "learning_rate": 9.786389118694897e-05, + "loss": 3.4114, + "step": 53800 + }, + { + "epoch": 0.4755470778360621, + "grad_norm": 0.7896223664283752, + "learning_rate": 9.785987098316499e-05, + "loss": 3.4087, + "step": 53850 + }, + { + "epoch": 0.47598862572634626, + "grad_norm": 9.002167701721191, + "learning_rate": 9.785584708261182e-05, + "loss": 3.2257, + "step": 53900 + }, + { + "epoch": 0.4764301736166305, + "grad_norm": 0.6309633851051331, + "learning_rate": 9.785181948560028e-05, + "loss": 3.1985, + "step": 53950 + }, + { + "epoch": 0.47687172150691465, + "grad_norm": 3.459434747695923, + "learning_rate": 9.784778819244144e-05, + "loss": 3.3152, + "step": 54000 + }, + { + "epoch": 0.47687172150691465, + "eval_asr_loss": 0.8963724033623304, + "eval_loss": 3.0343711376190186, + "eval_runtime": 20.7535, + "eval_samples_per_second": 37.006, + "eval_steps_per_second": 9.251, + "eval_tts_loss": 5.906121494862091, + "step": 54000 + }, + { + "epoch": 0.4773132693971988, + "grad_norm": 2.1301255226135254, + "learning_rate": 9.78437532034467e-05, + "loss": 3.7664, + "step": 54050 + }, + { + "epoch": 0.477754817287483, + "grad_norm": 2.864131450653076, + "learning_rate": 9.783971451892773e-05, + "loss": 3.0815, + "step": 54100 + }, + { + "epoch": 0.4781963651777672, + "grad_norm": 2.723444938659668, + "learning_rate": 9.783567213919649e-05, + "loss": 3.2961, + "step": 54150 + }, + { + "epoch": 0.47863791306805137, + "grad_norm": 2.7087883949279785, + "learning_rate": 9.783162606456521e-05, + "loss": 3.4694, + "step": 54200 + }, + { + "epoch": 0.47907946095833553, + "grad_norm": 2.582592248916626, + "learning_rate": 9.782757629534642e-05, + "loss": 3.4529, + "step": 54250 + }, + { + "epoch": 0.4795210088486197, + "grad_norm": 1.1580854654312134, + "learning_rate": 9.782352283185293e-05, + "loss": 3.357, + "step": 54300 + }, + { + "epoch": 0.4799625567389039, + "grad_norm": 3.442847490310669, + "learning_rate": 9.781946567439781e-05, + "loss": 3.2244, + "step": 54350 + }, + { + "epoch": 0.4804041046291881, + "grad_norm": 4.473285675048828, + "learning_rate": 9.781540482329447e-05, + "loss": 3.2563, + "step": 54400 + }, + { + "epoch": 0.48084565251947226, + "grad_norm": 2.0757851600646973, + "learning_rate": 9.78113402788566e-05, + "loss": 3.3494, + "step": 54450 + }, + { + "epoch": 0.4812872004097564, + "grad_norm": 1.0765964984893799, + "learning_rate": 9.78072720413981e-05, + "loss": 2.973, + "step": 54500 + }, + { + "epoch": 0.48172874830004064, + "grad_norm": 2.187049150466919, + "learning_rate": 9.780320011123322e-05, + "loss": 3.5087, + "step": 54550 + }, + { + "epoch": 0.4821702961903248, + "grad_norm": 2.3897063732147217, + "learning_rate": 9.779912448867649e-05, + "loss": 3.3432, + "step": 54600 + }, + { + "epoch": 0.482611844080609, + "grad_norm": 1.4350836277008057, + "learning_rate": 9.779504517404274e-05, + "loss": 3.547, + "step": 54650 + }, + { + "epoch": 0.48305339197089314, + "grad_norm": 1.5999610424041748, + "learning_rate": 9.779096216764703e-05, + "loss": 3.1709, + "step": 54700 + }, + { + "epoch": 0.48349493986117736, + "grad_norm": 4.64166784286499, + "learning_rate": 9.778687546980474e-05, + "loss": 3.6566, + "step": 54750 + }, + { + "epoch": 0.48393648775146153, + "grad_norm": 2.923461675643921, + "learning_rate": 9.778278508083154e-05, + "loss": 3.4523, + "step": 54800 + }, + { + "epoch": 0.4843780356417457, + "grad_norm": 1.1621273756027222, + "learning_rate": 9.77786910010434e-05, + "loss": 3.4142, + "step": 54850 + }, + { + "epoch": 0.48481958353202986, + "grad_norm": 2.0562987327575684, + "learning_rate": 9.77745932307565e-05, + "loss": 3.2347, + "step": 54900 + }, + { + "epoch": 0.4852611314223141, + "grad_norm": 2.8436999320983887, + "learning_rate": 9.777049177028742e-05, + "loss": 3.4778, + "step": 54950 + }, + { + "epoch": 0.48570267931259825, + "grad_norm": 3.21048641204834, + "learning_rate": 9.776638661995292e-05, + "loss": 3.4173, + "step": 55000 + }, + { + "epoch": 0.4861442272028824, + "grad_norm": 2.347586154937744, + "learning_rate": 9.77622777800701e-05, + "loss": 3.6781, + "step": 55050 + }, + { + "epoch": 0.4865857750931666, + "grad_norm": 1.2790504693984985, + "learning_rate": 9.775816525095633e-05, + "loss": 3.5449, + "step": 55100 + }, + { + "epoch": 0.4870273229834508, + "grad_norm": 2.234579563140869, + "learning_rate": 9.775404903292929e-05, + "loss": 3.403, + "step": 55150 + }, + { + "epoch": 0.48746887087373497, + "grad_norm": 3.1236841678619385, + "learning_rate": 9.774992912630689e-05, + "loss": 3.5316, + "step": 55200 + }, + { + "epoch": 0.48791041876401914, + "grad_norm": 2.067030906677246, + "learning_rate": 9.774580553140736e-05, + "loss": 3.3101, + "step": 55250 + }, + { + "epoch": 0.4883519666543033, + "grad_norm": 2.3011932373046875, + "learning_rate": 9.774167824854925e-05, + "loss": 3.9037, + "step": 55300 + }, + { + "epoch": 0.4887935145445875, + "grad_norm": 0.8518275618553162, + "learning_rate": 9.773754727805132e-05, + "loss": 3.7854, + "step": 55350 + }, + { + "epoch": 0.4892350624348717, + "grad_norm": 2.5740087032318115, + "learning_rate": 9.773341262023265e-05, + "loss": 3.6587, + "step": 55400 + }, + { + "epoch": 0.48967661032515586, + "grad_norm": 2.541947364807129, + "learning_rate": 9.772927427541266e-05, + "loss": 3.3046, + "step": 55450 + }, + { + "epoch": 0.49011815821544, + "grad_norm": 10.81189250946045, + "learning_rate": 9.772513224391093e-05, + "loss": 3.5087, + "step": 55500 + }, + { + "epoch": 0.49055970610572425, + "grad_norm": 2.659097909927368, + "learning_rate": 9.772098652604745e-05, + "loss": 3.7244, + "step": 55550 + }, + { + "epoch": 0.4910012539960084, + "grad_norm": 1.6622788906097412, + "learning_rate": 9.771683712214241e-05, + "loss": 3.9154, + "step": 55600 + }, + { + "epoch": 0.4914428018862926, + "grad_norm": 4.244743347167969, + "learning_rate": 9.771268403251634e-05, + "loss": 3.1095, + "step": 55650 + }, + { + "epoch": 0.49188434977657675, + "grad_norm": 2.233381986618042, + "learning_rate": 9.770852725749002e-05, + "loss": 2.7214, + "step": 55700 + }, + { + "epoch": 0.49232589766686097, + "grad_norm": 3.992842674255371, + "learning_rate": 9.770436679738452e-05, + "loss": 2.8508, + "step": 55750 + }, + { + "epoch": 0.49276744555714513, + "grad_norm": 2.8735084533691406, + "learning_rate": 9.770020265252122e-05, + "loss": 3.3086, + "step": 55800 + }, + { + "epoch": 0.4932089934474293, + "grad_norm": 0.7846609950065613, + "learning_rate": 9.769603482322173e-05, + "loss": 3.0594, + "step": 55850 + }, + { + "epoch": 0.49365054133771347, + "grad_norm": 3.95578670501709, + "learning_rate": 9.769186330980802e-05, + "loss": 3.2712, + "step": 55900 + }, + { + "epoch": 0.4940920892279977, + "grad_norm": 1.6080371141433716, + "learning_rate": 9.768768811260229e-05, + "loss": 3.3038, + "step": 55950 + }, + { + "epoch": 0.49453363711828185, + "grad_norm": 1.84148108959198, + "learning_rate": 9.7683509231927e-05, + "loss": 3.2961, + "step": 56000 + }, + { + "epoch": 0.494975185008566, + "grad_norm": 3.487783670425415, + "learning_rate": 9.7679326668105e-05, + "loss": 3.0265, + "step": 56050 + }, + { + "epoch": 0.4954167328988502, + "grad_norm": 1.8693515062332153, + "learning_rate": 9.767514042145931e-05, + "loss": 3.3785, + "step": 56100 + }, + { + "epoch": 0.4958582807891344, + "grad_norm": 0.690427839756012, + "learning_rate": 9.767095049231329e-05, + "loss": 2.9811, + "step": 56150 + }, + { + "epoch": 0.4962998286794186, + "grad_norm": 3.2560925483703613, + "learning_rate": 9.766675688099059e-05, + "loss": 3.3119, + "step": 56200 + }, + { + "epoch": 0.49674137656970274, + "grad_norm": 2.757688522338867, + "learning_rate": 9.766255958781512e-05, + "loss": 3.3075, + "step": 56250 + }, + { + "epoch": 0.4971829244599869, + "grad_norm": 1.4115214347839355, + "learning_rate": 9.765835861311108e-05, + "loss": 3.3612, + "step": 56300 + }, + { + "epoch": 0.49762447235027113, + "grad_norm": 3.8688366413116455, + "learning_rate": 9.765415395720298e-05, + "loss": 3.1324, + "step": 56350 + }, + { + "epoch": 0.4980660202405553, + "grad_norm": 3.989609479904175, + "learning_rate": 9.764994562041559e-05, + "loss": 3.1485, + "step": 56400 + }, + { + "epoch": 0.49850756813083946, + "grad_norm": 2.1451547145843506, + "learning_rate": 9.764573360307394e-05, + "loss": 3.2281, + "step": 56450 + }, + { + "epoch": 0.49894911602112363, + "grad_norm": 1.555163025856018, + "learning_rate": 9.76415179055034e-05, + "loss": 3.0442, + "step": 56500 + }, + { + "epoch": 0.49939066391140785, + "grad_norm": 2.0306787490844727, + "learning_rate": 9.76372985280296e-05, + "loss": 3.3319, + "step": 56550 + }, + { + "epoch": 0.499832211801692, + "grad_norm": 3.7980337142944336, + "learning_rate": 9.763307547097844e-05, + "loss": 3.6434, + "step": 56600 + }, + { + "epoch": 0.5002737596919762, + "grad_norm": 4.4818525314331055, + "learning_rate": 9.762884873467611e-05, + "loss": 3.2294, + "step": 56650 + }, + { + "epoch": 0.5007153075822603, + "grad_norm": 1.937999963760376, + "learning_rate": 9.76246183194491e-05, + "loss": 3.5661, + "step": 56700 + }, + { + "epoch": 0.5011568554725445, + "grad_norm": 1.306862711906433, + "learning_rate": 9.762038422562417e-05, + "loss": 2.9257, + "step": 56750 + }, + { + "epoch": 0.5015984033628287, + "grad_norm": 1.1766592264175415, + "learning_rate": 9.761614645352835e-05, + "loss": 3.3017, + "step": 56800 + }, + { + "epoch": 0.502039951253113, + "grad_norm": 3.1086044311523438, + "learning_rate": 9.7611905003489e-05, + "loss": 3.5308, + "step": 56850 + }, + { + "epoch": 0.5024814991433971, + "grad_norm": 2.8829550743103027, + "learning_rate": 9.760765987583373e-05, + "loss": 3.1106, + "step": 56900 + }, + { + "epoch": 0.5029230470336813, + "grad_norm": 2.7359766960144043, + "learning_rate": 9.760341107089044e-05, + "loss": 3.509, + "step": 56950 + }, + { + "epoch": 0.5033645949239655, + "grad_norm": 7.022033214569092, + "learning_rate": 9.759915858898732e-05, + "loss": 3.1945, + "step": 57000 + }, + { + "epoch": 0.5033645949239655, + "eval_asr_loss": 0.9032422534916057, + "eval_loss": 3.0304603576660156, + "eval_runtime": 51.0735, + "eval_samples_per_second": 15.037, + "eval_steps_per_second": 3.759, + "eval_tts_loss": 5.9164860032310695, + "step": 57000 + }, + { + "epoch": 0.5038061428142496, + "grad_norm": 1.487955093383789, + "learning_rate": 9.75949024304528e-05, + "loss": 3.1603, + "step": 57050 + }, + { + "epoch": 0.5042476907045338, + "grad_norm": 2.423590660095215, + "learning_rate": 9.759064259561568e-05, + "loss": 3.6678, + "step": 57100 + }, + { + "epoch": 0.504689238594818, + "grad_norm": 1.796743631362915, + "learning_rate": 9.758637908480497e-05, + "loss": 3.4576, + "step": 57150 + }, + { + "epoch": 0.5051307864851021, + "grad_norm": 6.50948429107666, + "learning_rate": 9.758211189834999e-05, + "loss": 3.2279, + "step": 57200 + }, + { + "epoch": 0.5055723343753864, + "grad_norm": 1.0165084600448608, + "learning_rate": 9.757784103658036e-05, + "loss": 3.3161, + "step": 57250 + }, + { + "epoch": 0.5060138822656706, + "grad_norm": 3.4719436168670654, + "learning_rate": 9.757356649982595e-05, + "loss": 3.6583, + "step": 57300 + }, + { + "epoch": 0.5064554301559547, + "grad_norm": 2.7984747886657715, + "learning_rate": 9.756928828841694e-05, + "loss": 3.507, + "step": 57350 + }, + { + "epoch": 0.5068969780462389, + "grad_norm": 4.911170959472656, + "learning_rate": 9.75650064026838e-05, + "loss": 3.0097, + "step": 57400 + }, + { + "epoch": 0.5073385259365231, + "grad_norm": 3.57440185546875, + "learning_rate": 9.756072084295725e-05, + "loss": 2.9845, + "step": 57450 + }, + { + "epoch": 0.5077800738268072, + "grad_norm": 1.4227149486541748, + "learning_rate": 9.755651743023584e-05, + "loss": 3.6389, + "step": 57500 + }, + { + "epoch": 0.5082216217170914, + "grad_norm": 2.1636316776275635, + "learning_rate": 9.755222459697919e-05, + "loss": 2.9495, + "step": 57550 + }, + { + "epoch": 0.5086631696073756, + "grad_norm": 5.281383991241455, + "learning_rate": 9.754792809071644e-05, + "loss": 3.388, + "step": 57600 + }, + { + "epoch": 0.5091047174976598, + "grad_norm": 1.8442537784576416, + "learning_rate": 9.754362791177944e-05, + "loss": 3.7662, + "step": 57650 + }, + { + "epoch": 0.509546265387944, + "grad_norm": 2.220319986343384, + "learning_rate": 9.753932406050032e-05, + "loss": 3.1771, + "step": 57700 + }, + { + "epoch": 0.5099878132782282, + "grad_norm": 2.379897356033325, + "learning_rate": 9.753501653721155e-05, + "loss": 3.2315, + "step": 57750 + }, + { + "epoch": 0.5104293611685123, + "grad_norm": 1.3445017337799072, + "learning_rate": 9.753070534224584e-05, + "loss": 3.4884, + "step": 57800 + }, + { + "epoch": 0.5108709090587965, + "grad_norm": 1.7846251726150513, + "learning_rate": 9.752639047593619e-05, + "loss": 3.5118, + "step": 57850 + }, + { + "epoch": 0.5113124569490807, + "grad_norm": 0.7859761714935303, + "learning_rate": 9.752207193861589e-05, + "loss": 3.4603, + "step": 57900 + }, + { + "epoch": 0.5117540048393648, + "grad_norm": 1.91860830783844, + "learning_rate": 9.751774973061851e-05, + "loss": 3.354, + "step": 57950 + }, + { + "epoch": 0.512195552729649, + "grad_norm": 2.4331154823303223, + "learning_rate": 9.751342385227791e-05, + "loss": 3.1493, + "step": 58000 + }, + { + "epoch": 0.5126371006199333, + "grad_norm": 2.020382881164551, + "learning_rate": 9.750909430392821e-05, + "loss": 3.4823, + "step": 58050 + }, + { + "epoch": 0.5130786485102174, + "grad_norm": 1.6145989894866943, + "learning_rate": 9.750476108590383e-05, + "loss": 3.2045, + "step": 58100 + }, + { + "epoch": 0.5135201964005016, + "grad_norm": 4.1392364501953125, + "learning_rate": 9.750042419853949e-05, + "loss": 3.2311, + "step": 58150 + }, + { + "epoch": 0.5139617442907858, + "grad_norm": 2.6319990158081055, + "learning_rate": 9.749608364217018e-05, + "loss": 3.4167, + "step": 58200 + }, + { + "epoch": 0.51440329218107, + "grad_norm": 0.8864327073097229, + "learning_rate": 9.749173941713114e-05, + "loss": 3.4851, + "step": 58250 + }, + { + "epoch": 0.5148448400713541, + "grad_norm": 1.8275792598724365, + "learning_rate": 9.748739152375799e-05, + "loss": 3.429, + "step": 58300 + }, + { + "epoch": 0.5152863879616383, + "grad_norm": 1.4619648456573486, + "learning_rate": 9.74830399623865e-05, + "loss": 3.1952, + "step": 58350 + }, + { + "epoch": 0.5157279358519224, + "grad_norm": 2.958252429962158, + "learning_rate": 9.747868473335283e-05, + "loss": 3.3281, + "step": 58400 + }, + { + "epoch": 0.5161694837422067, + "grad_norm": 3.7064812183380127, + "learning_rate": 9.747432583699334e-05, + "loss": 3.1441, + "step": 58450 + }, + { + "epoch": 0.5166110316324909, + "grad_norm": 2.135683298110962, + "learning_rate": 9.746996327364478e-05, + "loss": 3.3074, + "step": 58500 + }, + { + "epoch": 0.5170525795227751, + "grad_norm": 4.96755838394165, + "learning_rate": 9.746559704364409e-05, + "loss": 3.1264, + "step": 58550 + }, + { + "epoch": 0.5174941274130592, + "grad_norm": 1.5454546213150024, + "learning_rate": 9.74612271473285e-05, + "loss": 3.3129, + "step": 58600 + }, + { + "epoch": 0.5179356753033434, + "grad_norm": 2.1017255783081055, + "learning_rate": 9.74568535850356e-05, + "loss": 3.2598, + "step": 58650 + }, + { + "epoch": 0.5183772231936276, + "grad_norm": 1.2823448181152344, + "learning_rate": 9.745247635710318e-05, + "loss": 3.6969, + "step": 58700 + }, + { + "epoch": 0.5188187710839117, + "grad_norm": 2.371886968612671, + "learning_rate": 9.744809546386933e-05, + "loss": 3.2902, + "step": 58750 + }, + { + "epoch": 0.5192603189741959, + "grad_norm": 1.7832865715026855, + "learning_rate": 9.744371090567246e-05, + "loss": 2.8521, + "step": 58800 + }, + { + "epoch": 0.5197018668644802, + "grad_norm": 1.12856125831604, + "learning_rate": 9.743932268285124e-05, + "loss": 3.7715, + "step": 58850 + }, + { + "epoch": 0.5201434147547643, + "grad_norm": 1.2334407567977905, + "learning_rate": 9.743493079574461e-05, + "loss": 3.2311, + "step": 58900 + }, + { + "epoch": 0.5205849626450485, + "grad_norm": 2.203479528427124, + "learning_rate": 9.743053524469182e-05, + "loss": 3.438, + "step": 58950 + }, + { + "epoch": 0.5210265105353327, + "grad_norm": 0.8005960583686829, + "learning_rate": 9.742613603003238e-05, + "loss": 3.6582, + "step": 59000 + }, + { + "epoch": 0.5214680584256168, + "grad_norm": 2.671816110610962, + "learning_rate": 9.742173315210608e-05, + "loss": 2.9465, + "step": 59050 + }, + { + "epoch": 0.521909606315901, + "grad_norm": 0.9956416487693787, + "learning_rate": 9.741732661125304e-05, + "loss": 3.3792, + "step": 59100 + }, + { + "epoch": 0.5223511542061852, + "grad_norm": 1.9408859014511108, + "learning_rate": 9.741291640781359e-05, + "loss": 3.2726, + "step": 59150 + }, + { + "epoch": 0.5227927020964693, + "grad_norm": 1.1133266687393188, + "learning_rate": 9.740850254212841e-05, + "loss": 3.3387, + "step": 59200 + }, + { + "epoch": 0.5232342499867536, + "grad_norm": 0.527526319026947, + "learning_rate": 9.740408501453841e-05, + "loss": 3.0193, + "step": 59250 + }, + { + "epoch": 0.5236757978770378, + "grad_norm": 1.995932936668396, + "learning_rate": 9.7399752285049e-05, + "loss": 3.5348, + "step": 59300 + }, + { + "epoch": 0.5241173457673219, + "grad_norm": 1.4613289833068848, + "learning_rate": 9.739532750789442e-05, + "loss": 3.0866, + "step": 59350 + }, + { + "epoch": 0.5245588936576061, + "grad_norm": 2.9688708782196045, + "learning_rate": 9.739089906985268e-05, + "loss": 3.5847, + "step": 59400 + }, + { + "epoch": 0.5250004415478903, + "grad_norm": 1.3395789861679077, + "learning_rate": 9.738646697126586e-05, + "loss": 3.353, + "step": 59450 + }, + { + "epoch": 0.5254419894381744, + "grad_norm": 4.93052864074707, + "learning_rate": 9.738203121247627e-05, + "loss": 3.626, + "step": 59500 + }, + { + "epoch": 0.5258835373284586, + "grad_norm": 3.3152246475219727, + "learning_rate": 9.737759179382658e-05, + "loss": 2.8786, + "step": 59550 + }, + { + "epoch": 0.5263250852187428, + "grad_norm": 1.5898215770721436, + "learning_rate": 9.737314871565965e-05, + "loss": 3.7041, + "step": 59600 + }, + { + "epoch": 0.526766633109027, + "grad_norm": 2.1526150703430176, + "learning_rate": 9.736870197831871e-05, + "loss": 3.5867, + "step": 59650 + }, + { + "epoch": 0.5272081809993112, + "grad_norm": 2.351823568344116, + "learning_rate": 9.736425158214723e-05, + "loss": 3.4919, + "step": 59700 + }, + { + "epoch": 0.5276497288895954, + "grad_norm": 2.4668421745300293, + "learning_rate": 9.735979752748894e-05, + "loss": 3.3577, + "step": 59750 + }, + { + "epoch": 0.5280912767798795, + "grad_norm": 2.4981229305267334, + "learning_rate": 9.73553398146879e-05, + "loss": 3.368, + "step": 59800 + }, + { + "epoch": 0.5285328246701637, + "grad_norm": 1.9264018535614014, + "learning_rate": 9.735087844408839e-05, + "loss": 3.7379, + "step": 59850 + }, + { + "epoch": 0.5289743725604479, + "grad_norm": 4.922735691070557, + "learning_rate": 9.734641341603507e-05, + "loss": 3.1943, + "step": 59900 + }, + { + "epoch": 0.529415920450732, + "grad_norm": 5.6230645179748535, + "learning_rate": 9.73419447308728e-05, + "loss": 3.3315, + "step": 59950 + }, + { + "epoch": 0.5298574683410162, + "grad_norm": 1.985254168510437, + "learning_rate": 9.733747238894673e-05, + "loss": 3.8071, + "step": 60000 + }, + { + "epoch": 0.5298574683410162, + "eval_asr_loss": 0.8992706317769239, + "eval_loss": 3.0059359073638916, + "eval_runtime": 20.6646, + "eval_samples_per_second": 37.165, + "eval_steps_per_second": 9.291, + "eval_tts_loss": 5.9863666861098865, + "step": 60000 + }, + { + "epoch": 0.5302990162313005, + "grad_norm": 1.029417872428894, + "learning_rate": 9.733299639060233e-05, + "loss": 3.503, + "step": 60050 + }, + { + "epoch": 0.5307405641215847, + "grad_norm": 6.666355609893799, + "learning_rate": 9.732851673618535e-05, + "loss": 3.2843, + "step": 60100 + }, + { + "epoch": 0.5311821120118688, + "grad_norm": 1.5539653301239014, + "learning_rate": 9.732403342604177e-05, + "loss": 3.374, + "step": 60150 + }, + { + "epoch": 0.531623659902153, + "grad_norm": 1.662533164024353, + "learning_rate": 9.731954646051792e-05, + "loss": 3.1763, + "step": 60200 + }, + { + "epoch": 0.5320652077924372, + "grad_norm": 1.5471640825271606, + "learning_rate": 9.731505583996035e-05, + "loss": 3.3524, + "step": 60250 + }, + { + "epoch": 0.5325067556827213, + "grad_norm": 3.6004583835601807, + "learning_rate": 9.731056156471594e-05, + "loss": 3.4967, + "step": 60300 + }, + { + "epoch": 0.5329483035730055, + "grad_norm": 5.77100944519043, + "learning_rate": 9.730606363513184e-05, + "loss": 3.3075, + "step": 60350 + }, + { + "epoch": 0.5333898514632897, + "grad_norm": 2.9445252418518066, + "learning_rate": 9.730156205155545e-05, + "loss": 3.681, + "step": 60400 + }, + { + "epoch": 0.5338313993535739, + "grad_norm": 3.9473483562469482, + "learning_rate": 9.729705681433451e-05, + "loss": 2.9724, + "step": 60450 + }, + { + "epoch": 0.5342729472438581, + "grad_norm": 2.4145290851593018, + "learning_rate": 9.729254792381698e-05, + "loss": 3.1867, + "step": 60500 + }, + { + "epoch": 0.5347144951341423, + "grad_norm": 2.9569525718688965, + "learning_rate": 9.728803538035119e-05, + "loss": 3.6975, + "step": 60550 + }, + { + "epoch": 0.5351560430244264, + "grad_norm": 3.94183349609375, + "learning_rate": 9.728351918428562e-05, + "loss": 3.6706, + "step": 60600 + }, + { + "epoch": 0.5355975909147106, + "grad_norm": 1.8953005075454712, + "learning_rate": 9.727899933596917e-05, + "loss": 3.3651, + "step": 60650 + }, + { + "epoch": 0.5360391388049948, + "grad_norm": 1.777334213256836, + "learning_rate": 9.727447583575092e-05, + "loss": 3.7644, + "step": 60700 + }, + { + "epoch": 0.5364806866952789, + "grad_norm": 3.474303960800171, + "learning_rate": 9.72699486839803e-05, + "loss": 3.2085, + "step": 60750 + }, + { + "epoch": 0.5369222345855631, + "grad_norm": 2.88875412940979, + "learning_rate": 9.726541788100696e-05, + "loss": 3.8591, + "step": 60800 + }, + { + "epoch": 0.5373637824758474, + "grad_norm": 3.021052598953247, + "learning_rate": 9.726088342718091e-05, + "loss": 3.3407, + "step": 60850 + }, + { + "epoch": 0.5378053303661315, + "grad_norm": 1.9368441104888916, + "learning_rate": 9.725634532285238e-05, + "loss": 3.4342, + "step": 60900 + }, + { + "epoch": 0.5382468782564157, + "grad_norm": 1.3616617918014526, + "learning_rate": 9.725180356837187e-05, + "loss": 3.1015, + "step": 60950 + }, + { + "epoch": 0.5386884261466999, + "grad_norm": 1.2605106830596924, + "learning_rate": 9.724725816409024e-05, + "loss": 3.4434, + "step": 61000 + }, + { + "epoch": 0.539129974036984, + "grad_norm": 2.5226216316223145, + "learning_rate": 9.724270911035856e-05, + "loss": 3.2771, + "step": 61050 + }, + { + "epoch": 0.5395715219272682, + "grad_norm": 1.5489667654037476, + "learning_rate": 9.723815640752818e-05, + "loss": 2.9834, + "step": 61100 + }, + { + "epoch": 0.5400130698175524, + "grad_norm": 0.3823397159576416, + "learning_rate": 9.723360005595082e-05, + "loss": 3.4476, + "step": 61150 + }, + { + "epoch": 0.5404546177078365, + "grad_norm": 1.4170905351638794, + "learning_rate": 9.722904005597838e-05, + "loss": 2.9736, + "step": 61200 + }, + { + "epoch": 0.5408961655981208, + "grad_norm": 3.013650417327881, + "learning_rate": 9.722447640796306e-05, + "loss": 3.4688, + "step": 61250 + }, + { + "epoch": 0.541337713488405, + "grad_norm": 2.0270910263061523, + "learning_rate": 9.721990911225742e-05, + "loss": 3.4488, + "step": 61300 + }, + { + "epoch": 0.5417792613786891, + "grad_norm": 2.3157079219818115, + "learning_rate": 9.72153381692142e-05, + "loss": 3.4686, + "step": 61350 + }, + { + "epoch": 0.5422208092689733, + "grad_norm": 2.65665864944458, + "learning_rate": 9.721076357918648e-05, + "loss": 3.0758, + "step": 61400 + }, + { + "epoch": 0.5426623571592575, + "grad_norm": 4.086440086364746, + "learning_rate": 9.720618534252761e-05, + "loss": 3.0561, + "step": 61450 + }, + { + "epoch": 0.5431039050495416, + "grad_norm": 1.6047959327697754, + "learning_rate": 9.720160345959122e-05, + "loss": 3.2575, + "step": 61500 + }, + { + "epoch": 0.5435454529398258, + "grad_norm": 1.371722936630249, + "learning_rate": 9.719701793073121e-05, + "loss": 3.197, + "step": 61550 + }, + { + "epoch": 0.54398700083011, + "grad_norm": 1.1012533903121948, + "learning_rate": 9.71924287563018e-05, + "loss": 3.1246, + "step": 61600 + }, + { + "epoch": 0.5444285487203943, + "grad_norm": 2.610647201538086, + "learning_rate": 9.718783593665745e-05, + "loss": 3.4834, + "step": 61650 + }, + { + "epoch": 0.5448700966106784, + "grad_norm": 3.1460602283477783, + "learning_rate": 9.71832394721529e-05, + "loss": 3.5381, + "step": 61700 + }, + { + "epoch": 0.5453116445009626, + "grad_norm": 2.426410675048828, + "learning_rate": 9.717863936314322e-05, + "loss": 3.0234, + "step": 61750 + }, + { + "epoch": 0.5457531923912468, + "grad_norm": 1.7792102098464966, + "learning_rate": 9.71740356099837e-05, + "loss": 3.3962, + "step": 61800 + }, + { + "epoch": 0.5461947402815309, + "grad_norm": 0.535715639591217, + "learning_rate": 9.716942821302995e-05, + "loss": 3.2452, + "step": 61850 + }, + { + "epoch": 0.5466362881718151, + "grad_norm": 3.1257872581481934, + "learning_rate": 9.716481717263787e-05, + "loss": 3.2172, + "step": 61900 + }, + { + "epoch": 0.5470778360620993, + "grad_norm": 2.848803997039795, + "learning_rate": 9.71602024891636e-05, + "loss": 3.5478, + "step": 61950 + }, + { + "epoch": 0.5475193839523834, + "grad_norm": 2.5586445331573486, + "learning_rate": 9.71555841629636e-05, + "loss": 3.5599, + "step": 62000 + }, + { + "epoch": 0.5479609318426677, + "grad_norm": 5.202526569366455, + "learning_rate": 9.715096219439458e-05, + "loss": 3.3383, + "step": 62050 + }, + { + "epoch": 0.5484024797329519, + "grad_norm": 7.311855316162109, + "learning_rate": 9.714633658381358e-05, + "loss": 3.5755, + "step": 62100 + }, + { + "epoch": 0.548844027623236, + "grad_norm": 2.228634834289551, + "learning_rate": 9.714170733157784e-05, + "loss": 3.1598, + "step": 62150 + }, + { + "epoch": 0.5492855755135202, + "grad_norm": 0.7159774899482727, + "learning_rate": 9.713707443804499e-05, + "loss": 3.3646, + "step": 62200 + }, + { + "epoch": 0.5497271234038044, + "grad_norm": 1.7473576068878174, + "learning_rate": 9.713243790357282e-05, + "loss": 3.153, + "step": 62250 + }, + { + "epoch": 0.5501686712940885, + "grad_norm": 2.541585922241211, + "learning_rate": 9.712779772851952e-05, + "loss": 3.4916, + "step": 62300 + }, + { + "epoch": 0.5506102191843727, + "grad_norm": 5.356276512145996, + "learning_rate": 9.712315391324346e-05, + "loss": 3.2563, + "step": 62350 + }, + { + "epoch": 0.5510517670746569, + "grad_norm": 2.213956356048584, + "learning_rate": 9.711850645810336e-05, + "loss": 3.1959, + "step": 62400 + }, + { + "epoch": 0.5514933149649411, + "grad_norm": 1.0666532516479492, + "learning_rate": 9.711385536345818e-05, + "loss": 3.4801, + "step": 62450 + }, + { + "epoch": 0.5519348628552253, + "grad_norm": 5.453732967376709, + "learning_rate": 9.71092006296672e-05, + "loss": 3.1979, + "step": 62500 + }, + { + "epoch": 0.5523764107455095, + "grad_norm": 6.855234622955322, + "learning_rate": 9.710454225708994e-05, + "loss": 3.6221, + "step": 62550 + }, + { + "epoch": 0.5528179586357936, + "grad_norm": 1.6029174327850342, + "learning_rate": 9.709988024608623e-05, + "loss": 3.4752, + "step": 62600 + }, + { + "epoch": 0.5532595065260778, + "grad_norm": 1.8403346538543701, + "learning_rate": 9.709521459701616e-05, + "loss": 3.4726, + "step": 62650 + }, + { + "epoch": 0.553701054416362, + "grad_norm": 4.461715221405029, + "learning_rate": 9.709054531024011e-05, + "loss": 2.9593, + "step": 62700 + }, + { + "epoch": 0.5541426023066461, + "grad_norm": 0.8452909588813782, + "learning_rate": 9.708587238611877e-05, + "loss": 3.2217, + "step": 62750 + }, + { + "epoch": 0.5545841501969304, + "grad_norm": 1.3129770755767822, + "learning_rate": 9.708119582501305e-05, + "loss": 3.3426, + "step": 62800 + }, + { + "epoch": 0.5550256980872146, + "grad_norm": 2.4353716373443604, + "learning_rate": 9.707651562728419e-05, + "loss": 3.4732, + "step": 62850 + }, + { + "epoch": 0.5554672459774987, + "grad_norm": 1.2511876821517944, + "learning_rate": 9.707183179329371e-05, + "loss": 3.3568, + "step": 62900 + }, + { + "epoch": 0.5559087938677829, + "grad_norm": 4.347702980041504, + "learning_rate": 9.706714432340336e-05, + "loss": 3.1255, + "step": 62950 + }, + { + "epoch": 0.5563503417580671, + "grad_norm": 1.2785756587982178, + "learning_rate": 9.706245321797525e-05, + "loss": 3.5741, + "step": 63000 + }, + { + "epoch": 0.5563503417580671, + "eval_asr_loss": 0.9045116492456787, + "eval_loss": 3.014928102493286, + "eval_runtime": 20.6473, + "eval_samples_per_second": 37.196, + "eval_steps_per_second": 9.299, + "eval_tts_loss": 5.952768747740511, + "step": 63000 + }, + { + "epoch": 0.5567918896483512, + "grad_norm": 1.4415526390075684, + "learning_rate": 9.705775847737169e-05, + "loss": 3.5287, + "step": 63050 + }, + { + "epoch": 0.5572334375386354, + "grad_norm": 1.9535558223724365, + "learning_rate": 9.705306010195533e-05, + "loss": 3.3174, + "step": 63100 + }, + { + "epoch": 0.5576749854289196, + "grad_norm": 1.3182282447814941, + "learning_rate": 9.704835809208907e-05, + "loss": 3.4299, + "step": 63150 + }, + { + "epoch": 0.5581165333192039, + "grad_norm": 1.156562328338623, + "learning_rate": 9.704365244813613e-05, + "loss": 3.2571, + "step": 63200 + }, + { + "epoch": 0.558558081209488, + "grad_norm": 1.929991364479065, + "learning_rate": 9.703894317045993e-05, + "loss": 3.0214, + "step": 63250 + }, + { + "epoch": 0.5589996290997722, + "grad_norm": 1.8567146062850952, + "learning_rate": 9.703423025942426e-05, + "loss": 3.2539, + "step": 63300 + }, + { + "epoch": 0.5594411769900564, + "grad_norm": 0.45442625880241394, + "learning_rate": 9.702951371539315e-05, + "loss": 3.5004, + "step": 63350 + }, + { + "epoch": 0.5598827248803405, + "grad_norm": 2.3628077507019043, + "learning_rate": 9.702479353873089e-05, + "loss": 3.1837, + "step": 63400 + }, + { + "epoch": 0.5603242727706247, + "grad_norm": 3.7257354259490967, + "learning_rate": 9.702006972980208e-05, + "loss": 3.503, + "step": 63450 + }, + { + "epoch": 0.5607658206609089, + "grad_norm": 3.1037049293518066, + "learning_rate": 9.701534228897163e-05, + "loss": 3.2654, + "step": 63500 + }, + { + "epoch": 0.561207368551193, + "grad_norm": 2.567664861679077, + "learning_rate": 9.701061121660464e-05, + "loss": 3.3444, + "step": 63550 + }, + { + "epoch": 0.5616489164414773, + "grad_norm": 3.259284019470215, + "learning_rate": 9.700587651306658e-05, + "loss": 3.0933, + "step": 63600 + }, + { + "epoch": 0.5620904643317615, + "grad_norm": 3.6262829303741455, + "learning_rate": 9.700113817872317e-05, + "loss": 3.4273, + "step": 63650 + }, + { + "epoch": 0.5625320122220456, + "grad_norm": 2.7692084312438965, + "learning_rate": 9.699639621394039e-05, + "loss": 3.5773, + "step": 63700 + }, + { + "epoch": 0.5629735601123298, + "grad_norm": 2.4339802265167236, + "learning_rate": 9.699165061908451e-05, + "loss": 2.8369, + "step": 63750 + }, + { + "epoch": 0.563415108002614, + "grad_norm": 1.9642480611801147, + "learning_rate": 9.69869013945221e-05, + "loss": 3.533, + "step": 63800 + }, + { + "epoch": 0.5638566558928981, + "grad_norm": 4.266359329223633, + "learning_rate": 9.698214854062e-05, + "loss": 3.3043, + "step": 63850 + }, + { + "epoch": 0.5642982037831823, + "grad_norm": 1.5324000120162964, + "learning_rate": 9.697739205774532e-05, + "loss": 3.6485, + "step": 63900 + }, + { + "epoch": 0.5647397516734665, + "grad_norm": 3.5735909938812256, + "learning_rate": 9.6972727184053e-05, + "loss": 3.3017, + "step": 63950 + }, + { + "epoch": 0.5651812995637507, + "grad_norm": 1.4858098030090332, + "learning_rate": 9.696796351689678e-05, + "loss": 3.4435, + "step": 64000 + }, + { + "epoch": 0.5656228474540349, + "grad_norm": 3.1134603023529053, + "learning_rate": 9.696319622186367e-05, + "loss": 2.8183, + "step": 64050 + }, + { + "epoch": 0.5660643953443191, + "grad_norm": 0.87113356590271, + "learning_rate": 9.695842529932186e-05, + "loss": 3.5863, + "step": 64100 + }, + { + "epoch": 0.5665059432346032, + "grad_norm": 2.494359016418457, + "learning_rate": 9.695365074963992e-05, + "loss": 3.2139, + "step": 64150 + }, + { + "epoch": 0.5669474911248874, + "grad_norm": 4.142000198364258, + "learning_rate": 9.694887257318659e-05, + "loss": 3.4118, + "step": 64200 + }, + { + "epoch": 0.5673890390151716, + "grad_norm": 7.738039493560791, + "learning_rate": 9.694409077033097e-05, + "loss": 3.3375, + "step": 64250 + }, + { + "epoch": 0.5678305869054557, + "grad_norm": 2.2657651901245117, + "learning_rate": 9.693930534144243e-05, + "loss": 3.5303, + "step": 64300 + }, + { + "epoch": 0.5682721347957399, + "grad_norm": 1.5491681098937988, + "learning_rate": 9.693451628689059e-05, + "loss": 3.1383, + "step": 64350 + }, + { + "epoch": 0.5687136826860242, + "grad_norm": 5.7371745109558105, + "learning_rate": 9.692972360704534e-05, + "loss": 3.3374, + "step": 64400 + }, + { + "epoch": 0.5691552305763083, + "grad_norm": 1.6547493934631348, + "learning_rate": 9.69249273022769e-05, + "loss": 3.5325, + "step": 64450 + }, + { + "epoch": 0.5695967784665925, + "grad_norm": 4.246345043182373, + "learning_rate": 9.692012737295574e-05, + "loss": 3.487, + "step": 64500 + }, + { + "epoch": 0.5700383263568767, + "grad_norm": 4.030117988586426, + "learning_rate": 9.69153238194526e-05, + "loss": 3.0569, + "step": 64550 + }, + { + "epoch": 0.5704798742471608, + "grad_norm": 1.0426106452941895, + "learning_rate": 9.691051664213855e-05, + "loss": 3.3855, + "step": 64600 + }, + { + "epoch": 0.570921422137445, + "grad_norm": 1.773262619972229, + "learning_rate": 9.690570584138486e-05, + "loss": 3.0886, + "step": 64650 + }, + { + "epoch": 0.5713629700277292, + "grad_norm": 2.984727144241333, + "learning_rate": 9.690089141756316e-05, + "loss": 3.3807, + "step": 64700 + }, + { + "epoch": 0.5718045179180133, + "grad_norm": 2.3156557083129883, + "learning_rate": 9.689607337104528e-05, + "loss": 3.3806, + "step": 64750 + }, + { + "epoch": 0.5722460658082976, + "grad_norm": 3.1280009746551514, + "learning_rate": 9.689125170220341e-05, + "loss": 3.223, + "step": 64800 + }, + { + "epoch": 0.5726876136985818, + "grad_norm": 1.362710952758789, + "learning_rate": 9.688642641140999e-05, + "loss": 2.7777, + "step": 64850 + }, + { + "epoch": 0.573129161588866, + "grad_norm": 2.129289150238037, + "learning_rate": 9.68815974990377e-05, + "loss": 3.2796, + "step": 64900 + }, + { + "epoch": 0.5735707094791501, + "grad_norm": 3.2772629261016846, + "learning_rate": 9.687676496545955e-05, + "loss": 3.4964, + "step": 64950 + }, + { + "epoch": 0.5740122573694343, + "grad_norm": 2.643249034881592, + "learning_rate": 9.68719288110488e-05, + "loss": 3.6755, + "step": 65000 + }, + { + "epoch": 0.5744538052597185, + "grad_norm": 1.4029085636138916, + "learning_rate": 9.686708903617902e-05, + "loss": 3.1452, + "step": 65050 + }, + { + "epoch": 0.5748953531500026, + "grad_norm": 0.7895709276199341, + "learning_rate": 9.686224564122403e-05, + "loss": 3.233, + "step": 65100 + }, + { + "epoch": 0.5753369010402868, + "grad_norm": 1.5120134353637695, + "learning_rate": 9.685739862655793e-05, + "loss": 3.2801, + "step": 65150 + }, + { + "epoch": 0.5757784489305711, + "grad_norm": 1.919240117073059, + "learning_rate": 9.685254799255517e-05, + "loss": 3.3301, + "step": 65200 + }, + { + "epoch": 0.5762199968208552, + "grad_norm": 2.2543394565582275, + "learning_rate": 9.684769373959033e-05, + "loss": 2.9593, + "step": 65250 + }, + { + "epoch": 0.5766615447111394, + "grad_norm": 1.0832370519638062, + "learning_rate": 9.684283586803843e-05, + "loss": 3.6906, + "step": 65300 + }, + { + "epoch": 0.5771030926014236, + "grad_norm": 3.100877285003662, + "learning_rate": 9.683797437827466e-05, + "loss": 3.347, + "step": 65350 + }, + { + "epoch": 0.5775446404917077, + "grad_norm": 3.069957733154297, + "learning_rate": 9.683310927067455e-05, + "loss": 3.0518, + "step": 65400 + }, + { + "epoch": 0.5779861883819919, + "grad_norm": 1.4712010622024536, + "learning_rate": 9.682824054561389e-05, + "loss": 3.2539, + "step": 65450 + }, + { + "epoch": 0.5784277362722761, + "grad_norm": 1.5608115196228027, + "learning_rate": 9.682336820346874e-05, + "loss": 3.4519, + "step": 65500 + }, + { + "epoch": 0.5788692841625602, + "grad_norm": 2.5109148025512695, + "learning_rate": 9.681849224461544e-05, + "loss": 3.3817, + "step": 65550 + }, + { + "epoch": 0.5793108320528445, + "grad_norm": 3.372307300567627, + "learning_rate": 9.68136126694306e-05, + "loss": 3.4769, + "step": 65600 + }, + { + "epoch": 0.5797523799431287, + "grad_norm": 2.304180860519409, + "learning_rate": 9.680872947829118e-05, + "loss": 3.8492, + "step": 65650 + }, + { + "epoch": 0.5801939278334128, + "grad_norm": 0.9315201044082642, + "learning_rate": 9.680384267157434e-05, + "loss": 3.7529, + "step": 65700 + }, + { + "epoch": 0.580635475723697, + "grad_norm": 1.882952332496643, + "learning_rate": 9.679895224965752e-05, + "loss": 3.0259, + "step": 65750 + }, + { + "epoch": 0.5810770236139812, + "grad_norm": 3.846442222595215, + "learning_rate": 9.679405821291849e-05, + "loss": 3.0785, + "step": 65800 + }, + { + "epoch": 0.5815185715042653, + "grad_norm": 1.7225450277328491, + "learning_rate": 9.678916056173526e-05, + "loss": 3.5125, + "step": 65850 + }, + { + "epoch": 0.5819601193945495, + "grad_norm": 0.8811081051826477, + "learning_rate": 9.678425929648614e-05, + "loss": 3.0916, + "step": 65900 + }, + { + "epoch": 0.5824016672848337, + "grad_norm": 3.4488401412963867, + "learning_rate": 9.67793544175497e-05, + "loss": 3.0175, + "step": 65950 + }, + { + "epoch": 0.582843215175118, + "grad_norm": 2.16182804107666, + "learning_rate": 9.677444592530483e-05, + "loss": 3.6373, + "step": 66000 + }, + { + "epoch": 0.582843215175118, + "eval_asr_loss": 0.8874591935323445, + "eval_loss": 3.00752854347229, + "eval_runtime": 21.0653, + "eval_samples_per_second": 36.458, + "eval_steps_per_second": 9.115, + "eval_tts_loss": 5.963853557317323, + "step": 66000 + }, + { + "epoch": 0.5832847630654021, + "grad_norm": 2.242957353591919, + "learning_rate": 9.676953382013063e-05, + "loss": 3.0162, + "step": 66050 + }, + { + "epoch": 0.5837263109556863, + "grad_norm": 2.5106394290924072, + "learning_rate": 9.676461810240654e-05, + "loss": 3.4729, + "step": 66100 + }, + { + "epoch": 0.5841678588459704, + "grad_norm": 3.770650863647461, + "learning_rate": 9.675969877251225e-05, + "loss": 3.4405, + "step": 66150 + }, + { + "epoch": 0.5846094067362546, + "grad_norm": 2.2024853229522705, + "learning_rate": 9.675477583082775e-05, + "loss": 3.5091, + "step": 66200 + }, + { + "epoch": 0.5850509546265388, + "grad_norm": 5.273355007171631, + "learning_rate": 9.674984927773328e-05, + "loss": 3.5578, + "step": 66250 + }, + { + "epoch": 0.5854925025168229, + "grad_norm": 2.3930509090423584, + "learning_rate": 9.674491911360939e-05, + "loss": 3.269, + "step": 66300 + }, + { + "epoch": 0.5859340504071071, + "grad_norm": 0.5656553506851196, + "learning_rate": 9.673998533883687e-05, + "loss": 3.4693, + "step": 66350 + }, + { + "epoch": 0.5863755982973914, + "grad_norm": 1.1394141912460327, + "learning_rate": 9.673504795379683e-05, + "loss": 3.4782, + "step": 66400 + }, + { + "epoch": 0.5868171461876756, + "grad_norm": 2.660053253173828, + "learning_rate": 9.673010695887064e-05, + "loss": 3.5723, + "step": 66450 + }, + { + "epoch": 0.5872586940779597, + "grad_norm": 2.931849718093872, + "learning_rate": 9.672516235443994e-05, + "loss": 3.0214, + "step": 66500 + }, + { + "epoch": 0.5877002419682439, + "grad_norm": 3.5768563747406006, + "learning_rate": 9.672021414088667e-05, + "loss": 3.0214, + "step": 66550 + }, + { + "epoch": 0.588141789858528, + "grad_norm": 6.038229465484619, + "learning_rate": 9.671526231859305e-05, + "loss": 3.4275, + "step": 66600 + }, + { + "epoch": 0.5885833377488122, + "grad_norm": 3.714881181716919, + "learning_rate": 9.671030688794153e-05, + "loss": 2.9659, + "step": 66650 + }, + { + "epoch": 0.5890248856390964, + "grad_norm": 1.5607867240905762, + "learning_rate": 9.670544706544311e-05, + "loss": 3.4904, + "step": 66700 + }, + { + "epoch": 0.5894664335293806, + "grad_norm": 1.8332650661468506, + "learning_rate": 9.67004844913725e-05, + "loss": 3.0832, + "step": 66750 + }, + { + "epoch": 0.5899079814196648, + "grad_norm": 2.306621789932251, + "learning_rate": 9.669551831008545e-05, + "loss": 3.1016, + "step": 66800 + }, + { + "epoch": 0.590349529309949, + "grad_norm": 1.265374779701233, + "learning_rate": 9.66905485219656e-05, + "loss": 3.3778, + "step": 66850 + }, + { + "epoch": 0.5907910772002332, + "grad_norm": 2.479994773864746, + "learning_rate": 9.66855751273968e-05, + "loss": 3.3998, + "step": 66900 + }, + { + "epoch": 0.5912326250905173, + "grad_norm": 2.6619317531585693, + "learning_rate": 9.66805981267632e-05, + "loss": 3.0006, + "step": 66950 + }, + { + "epoch": 0.5916741729808015, + "grad_norm": 2.5897438526153564, + "learning_rate": 9.667561752044922e-05, + "loss": 3.6474, + "step": 67000 + }, + { + "epoch": 0.5921157208710857, + "grad_norm": 5.512159824371338, + "learning_rate": 9.667063330883961e-05, + "loss": 3.457, + "step": 67050 + }, + { + "epoch": 0.5925572687613698, + "grad_norm": 2.0321197509765625, + "learning_rate": 9.666564549231931e-05, + "loss": 3.653, + "step": 67100 + }, + { + "epoch": 0.592998816651654, + "grad_norm": 2.648627519607544, + "learning_rate": 9.666065407127361e-05, + "loss": 2.8554, + "step": 67150 + }, + { + "epoch": 0.5934403645419383, + "grad_norm": 2.1089653968811035, + "learning_rate": 9.665565904608806e-05, + "loss": 3.0816, + "step": 67200 + }, + { + "epoch": 0.5938819124322224, + "grad_norm": 0.6567474007606506, + "learning_rate": 9.665066041714849e-05, + "loss": 3.4614, + "step": 67250 + }, + { + "epoch": 0.5943234603225066, + "grad_norm": 4.268828392028809, + "learning_rate": 9.664565818484097e-05, + "loss": 3.1899, + "step": 67300 + }, + { + "epoch": 0.5947650082127908, + "grad_norm": 4.196305751800537, + "learning_rate": 9.664065234955191e-05, + "loss": 3.3985, + "step": 67350 + }, + { + "epoch": 0.5952065561030749, + "grad_norm": 4.905257225036621, + "learning_rate": 9.663564291166795e-05, + "loss": 3.3829, + "step": 67400 + }, + { + "epoch": 0.5956481039933591, + "grad_norm": 3.6008496284484863, + "learning_rate": 9.663062987157604e-05, + "loss": 2.946, + "step": 67450 + }, + { + "epoch": 0.5960896518836433, + "grad_norm": 1.2707490921020508, + "learning_rate": 9.66256132296634e-05, + "loss": 3.4213, + "step": 67500 + }, + { + "epoch": 0.5965311997739274, + "grad_norm": 3.5220985412597656, + "learning_rate": 9.662059298631749e-05, + "loss": 3.0964, + "step": 67550 + }, + { + "epoch": 0.5969727476642117, + "grad_norm": 1.3247735500335693, + "learning_rate": 9.66155691419261e-05, + "loss": 3.4796, + "step": 67600 + }, + { + "epoch": 0.5974142955544959, + "grad_norm": 1.086634874343872, + "learning_rate": 9.66105416968773e-05, + "loss": 3.4287, + "step": 67650 + }, + { + "epoch": 0.59785584344478, + "grad_norm": 3.1261913776397705, + "learning_rate": 9.660551065155938e-05, + "loss": 3.239, + "step": 67700 + }, + { + "epoch": 0.5982973913350642, + "grad_norm": 5.029034614562988, + "learning_rate": 9.660047600636099e-05, + "loss": 2.8555, + "step": 67750 + }, + { + "epoch": 0.5987389392253484, + "grad_norm": 2.428929328918457, + "learning_rate": 9.659543776167097e-05, + "loss": 3.331, + "step": 67800 + }, + { + "epoch": 0.5991804871156325, + "grad_norm": 2.1692111492156982, + "learning_rate": 9.65903959178785e-05, + "loss": 3.2252, + "step": 67850 + }, + { + "epoch": 0.5996220350059167, + "grad_norm": 5.579278945922852, + "learning_rate": 9.658535047537303e-05, + "loss": 3.2612, + "step": 67900 + }, + { + "epoch": 0.6000635828962009, + "grad_norm": 1.7974921464920044, + "learning_rate": 9.658030143454426e-05, + "loss": 3.3738, + "step": 67950 + }, + { + "epoch": 0.6005051307864852, + "grad_norm": 2.329169750213623, + "learning_rate": 9.657524879578221e-05, + "loss": 3.358, + "step": 68000 + }, + { + "epoch": 0.6009466786767693, + "grad_norm": 2.0072598457336426, + "learning_rate": 9.657019255947712e-05, + "loss": 3.1542, + "step": 68050 + }, + { + "epoch": 0.6013882265670535, + "grad_norm": 1.212756872177124, + "learning_rate": 9.656513272601957e-05, + "loss": 2.8526, + "step": 68100 + }, + { + "epoch": 0.6018297744573377, + "grad_norm": 2.1172633171081543, + "learning_rate": 9.656006929580036e-05, + "loss": 3.3746, + "step": 68150 + }, + { + "epoch": 0.6022713223476218, + "grad_norm": 3.0865988731384277, + "learning_rate": 9.655500226921064e-05, + "loss": 3.4431, + "step": 68200 + }, + { + "epoch": 0.602712870237906, + "grad_norm": 2.7705607414245605, + "learning_rate": 9.654993164664175e-05, + "loss": 3.3185, + "step": 68250 + }, + { + "epoch": 0.6031544181281902, + "grad_norm": 2.9133379459381104, + "learning_rate": 9.654485742848538e-05, + "loss": 3.2805, + "step": 68300 + }, + { + "epoch": 0.6035959660184743, + "grad_norm": 1.945061445236206, + "learning_rate": 9.653977961513347e-05, + "loss": 3.5587, + "step": 68350 + }, + { + "epoch": 0.6040375139087586, + "grad_norm": 0.5719209313392639, + "learning_rate": 9.653469820697822e-05, + "loss": 3.039, + "step": 68400 + }, + { + "epoch": 0.6044790617990428, + "grad_norm": 2.7784836292266846, + "learning_rate": 9.652961320441214e-05, + "loss": 3.6163, + "step": 68450 + }, + { + "epoch": 0.6049206096893269, + "grad_norm": 1.212082028388977, + "learning_rate": 9.652452460782799e-05, + "loss": 2.763, + "step": 68500 + }, + { + "epoch": 0.6053621575796111, + "grad_norm": 6.158472061157227, + "learning_rate": 9.651943241761884e-05, + "loss": 3.4118, + "step": 68550 + }, + { + "epoch": 0.6058037054698953, + "grad_norm": 1.3418790102005005, + "learning_rate": 9.6514336634178e-05, + "loss": 3.3112, + "step": 68600 + }, + { + "epoch": 0.6062452533601794, + "grad_norm": 2.3166797161102295, + "learning_rate": 9.65092372578991e-05, + "loss": 3.3965, + "step": 68650 + }, + { + "epoch": 0.6066868012504636, + "grad_norm": 2.7341291904449463, + "learning_rate": 9.6504134289176e-05, + "loss": 3.4362, + "step": 68700 + }, + { + "epoch": 0.6071283491407478, + "grad_norm": 5.785979270935059, + "learning_rate": 9.649902772840287e-05, + "loss": 3.5519, + "step": 68750 + }, + { + "epoch": 0.607569897031032, + "grad_norm": 1.710016131401062, + "learning_rate": 9.649391757597414e-05, + "loss": 3.8108, + "step": 68800 + }, + { + "epoch": 0.6080114449213162, + "grad_norm": 1.7782738208770752, + "learning_rate": 9.648880383228455e-05, + "loss": 3.0538, + "step": 68850 + }, + { + "epoch": 0.6084529928116004, + "grad_norm": 2.464695930480957, + "learning_rate": 9.648368649772907e-05, + "loss": 3.4832, + "step": 68900 + }, + { + "epoch": 0.6088945407018845, + "grad_norm": 2.7873873710632324, + "learning_rate": 9.647856557270299e-05, + "loss": 3.5608, + "step": 68950 + }, + { + "epoch": 0.6093360885921687, + "grad_norm": 2.8136725425720215, + "learning_rate": 9.647344105760183e-05, + "loss": 2.8681, + "step": 69000 + }, + { + "epoch": 0.6093360885921687, + "eval_asr_loss": 0.9085685970427385, + "eval_loss": 3.00970458984375, + "eval_runtime": 20.9691, + "eval_samples_per_second": 36.625, + "eval_steps_per_second": 9.156, + "eval_tts_loss": 5.906105015571997, + "step": 69000 + }, + { + "epoch": 0.6097776364824529, + "grad_norm": 1.4982023239135742, + "learning_rate": 9.646831295282143e-05, + "loss": 3.7657, + "step": 69050 + }, + { + "epoch": 0.610219184372737, + "grad_norm": 1.8872673511505127, + "learning_rate": 9.646318125875791e-05, + "loss": 3.1832, + "step": 69100 + }, + { + "epoch": 0.6106607322630212, + "grad_norm": 3.1537160873413086, + "learning_rate": 9.645804597580761e-05, + "loss": 3.028, + "step": 69150 + }, + { + "epoch": 0.6111022801533055, + "grad_norm": 2.5125792026519775, + "learning_rate": 9.645290710436722e-05, + "loss": 3.2232, + "step": 69200 + }, + { + "epoch": 0.6115438280435896, + "grad_norm": 2.1751527786254883, + "learning_rate": 9.644776464483368e-05, + "loss": 3.1004, + "step": 69250 + }, + { + "epoch": 0.6119853759338738, + "grad_norm": 1.4385361671447754, + "learning_rate": 9.644261859760417e-05, + "loss": 2.8845, + "step": 69300 + }, + { + "epoch": 0.612426923824158, + "grad_norm": 1.0426270961761475, + "learning_rate": 9.643746896307622e-05, + "loss": 3.3967, + "step": 69350 + }, + { + "epoch": 0.6128684717144421, + "grad_norm": 5.886585712432861, + "learning_rate": 9.643231574164755e-05, + "loss": 3.2875, + "step": 69400 + }, + { + "epoch": 0.6133100196047263, + "grad_norm": 3.5066511631011963, + "learning_rate": 9.642715893371622e-05, + "loss": 2.8433, + "step": 69450 + }, + { + "epoch": 0.6137515674950105, + "grad_norm": 2.092880964279175, + "learning_rate": 9.642199853968055e-05, + "loss": 3.2078, + "step": 69500 + }, + { + "epoch": 0.6141931153852946, + "grad_norm": 1.670851707458496, + "learning_rate": 9.641683455993916e-05, + "loss": 3.0111, + "step": 69550 + }, + { + "epoch": 0.6146346632755789, + "grad_norm": 3.2171874046325684, + "learning_rate": 9.641177038132528e-05, + "loss": 2.9596, + "step": 69600 + }, + { + "epoch": 0.6150762111658631, + "grad_norm": 1.39114248752594, + "learning_rate": 9.640659930306355e-05, + "loss": 3.2125, + "step": 69650 + }, + { + "epoch": 0.6155177590561473, + "grad_norm": 0.8414401412010193, + "learning_rate": 9.640142464028551e-05, + "loss": 3.3546, + "step": 69700 + }, + { + "epoch": 0.6159593069464314, + "grad_norm": 2.521662473678589, + "learning_rate": 9.639624639339092e-05, + "loss": 3.2137, + "step": 69750 + }, + { + "epoch": 0.6164008548367156, + "grad_norm": 2.6045455932617188, + "learning_rate": 9.63910645627797e-05, + "loss": 3.4016, + "step": 69800 + }, + { + "epoch": 0.6168424027269998, + "grad_norm": 2.437931537628174, + "learning_rate": 9.638587914885215e-05, + "loss": 3.0942, + "step": 69850 + }, + { + "epoch": 0.6172839506172839, + "grad_norm": 2.744288682937622, + "learning_rate": 9.638069015200877e-05, + "loss": 3.4627, + "step": 69900 + }, + { + "epoch": 0.6177254985075681, + "grad_norm": 8.556507110595703, + "learning_rate": 9.637549757265037e-05, + "loss": 2.9197, + "step": 69950 + }, + { + "epoch": 0.6181670463978524, + "grad_norm": 2.708615303039551, + "learning_rate": 9.637030141117803e-05, + "loss": 3.4477, + "step": 70000 + }, + { + "epoch": 0.6186085942881365, + "grad_norm": 2.7415497303009033, + "learning_rate": 9.636510166799313e-05, + "loss": 3.3546, + "step": 70050 + }, + { + "epoch": 0.6190501421784207, + "grad_norm": 2.8086228370666504, + "learning_rate": 9.635989834349728e-05, + "loss": 3.3978, + "step": 70100 + }, + { + "epoch": 0.6194916900687049, + "grad_norm": 2.983367681503296, + "learning_rate": 9.635469143809239e-05, + "loss": 3.243, + "step": 70150 + }, + { + "epoch": 0.619933237958989, + "grad_norm": 0.500150740146637, + "learning_rate": 9.634948095218068e-05, + "loss": 3.3688, + "step": 70200 + }, + { + "epoch": 0.6203747858492732, + "grad_norm": 2.545722484588623, + "learning_rate": 9.63442668861646e-05, + "loss": 3.6102, + "step": 70250 + }, + { + "epoch": 0.6208163337395574, + "grad_norm": 0.7100191712379456, + "learning_rate": 9.633904924044687e-05, + "loss": 3.7465, + "step": 70300 + }, + { + "epoch": 0.6212578816298415, + "grad_norm": 1.9451789855957031, + "learning_rate": 9.633382801543055e-05, + "loss": 3.3812, + "step": 70350 + }, + { + "epoch": 0.6216994295201258, + "grad_norm": 2.1257078647613525, + "learning_rate": 9.632860321151892e-05, + "loss": 3.0553, + "step": 70400 + }, + { + "epoch": 0.62214097741041, + "grad_norm": 0.8686696290969849, + "learning_rate": 9.632337482911553e-05, + "loss": 2.8603, + "step": 70450 + }, + { + "epoch": 0.6225825253006941, + "grad_norm": 6.291978359222412, + "learning_rate": 9.631814286862426e-05, + "loss": 3.1015, + "step": 70500 + }, + { + "epoch": 0.6230240731909783, + "grad_norm": 3.416860818862915, + "learning_rate": 9.631290733044921e-05, + "loss": 3.5239, + "step": 70550 + }, + { + "epoch": 0.6234656210812625, + "grad_norm": 1.6016830205917358, + "learning_rate": 9.63076682149948e-05, + "loss": 3.2555, + "step": 70600 + }, + { + "epoch": 0.6239071689715466, + "grad_norm": 2.5389463901519775, + "learning_rate": 9.630242552266569e-05, + "loss": 3.5458, + "step": 70650 + }, + { + "epoch": 0.6243487168618308, + "grad_norm": 2.5725655555725098, + "learning_rate": 9.629717925386683e-05, + "loss": 3.0217, + "step": 70700 + }, + { + "epoch": 0.624790264752115, + "grad_norm": 2.941615581512451, + "learning_rate": 9.629192940900348e-05, + "loss": 3.2919, + "step": 70750 + }, + { + "epoch": 0.6252318126423992, + "grad_norm": 2.9366567134857178, + "learning_rate": 9.628667598848113e-05, + "loss": 3.0738, + "step": 70800 + }, + { + "epoch": 0.6256733605326834, + "grad_norm": 4.780130863189697, + "learning_rate": 9.628141899270554e-05, + "loss": 3.4164, + "step": 70850 + }, + { + "epoch": 0.6261149084229676, + "grad_norm": 2.4676146507263184, + "learning_rate": 9.62761584220828e-05, + "loss": 2.6195, + "step": 70900 + }, + { + "epoch": 0.6265564563132517, + "grad_norm": 2.815596103668213, + "learning_rate": 9.627089427701923e-05, + "loss": 3.1961, + "step": 70950 + }, + { + "epoch": 0.6269980042035359, + "grad_norm": 1.198612093925476, + "learning_rate": 9.626562655792145e-05, + "loss": 3.446, + "step": 71000 + }, + { + "epoch": 0.6274395520938201, + "grad_norm": 0.4727330803871155, + "learning_rate": 9.626035526519632e-05, + "loss": 3.252, + "step": 71050 + }, + { + "epoch": 0.6278810999841042, + "grad_norm": 2.9533727169036865, + "learning_rate": 9.625508039925104e-05, + "loss": 3.3784, + "step": 71100 + }, + { + "epoch": 0.6283226478743884, + "grad_norm": 1.3359452486038208, + "learning_rate": 9.624980196049303e-05, + "loss": 3.5926, + "step": 71150 + }, + { + "epoch": 0.6287641957646727, + "grad_norm": 1.3383703231811523, + "learning_rate": 9.624451994932999e-05, + "loss": 3.4367, + "step": 71200 + }, + { + "epoch": 0.6292057436549569, + "grad_norm": 3.669863224029541, + "learning_rate": 9.623923436616996e-05, + "loss": 3.4483, + "step": 71250 + }, + { + "epoch": 0.629647291545241, + "grad_norm": 3.541435718536377, + "learning_rate": 9.623394521142113e-05, + "loss": 3.5803, + "step": 71300 + }, + { + "epoch": 0.6300888394355252, + "grad_norm": 1.2664039134979248, + "learning_rate": 9.622865248549211e-05, + "loss": 3.02, + "step": 71350 + }, + { + "epoch": 0.6305303873258093, + "grad_norm": 3.8231582641601562, + "learning_rate": 9.622335618879168e-05, + "loss": 3.2505, + "step": 71400 + }, + { + "epoch": 0.6309719352160935, + "grad_norm": 1.3903136253356934, + "learning_rate": 9.621805632172896e-05, + "loss": 2.9738, + "step": 71450 + }, + { + "epoch": 0.6314134831063777, + "grad_norm": 1.4685653448104858, + "learning_rate": 9.621275288471329e-05, + "loss": 3.4252, + "step": 71500 + }, + { + "epoch": 0.6318550309966618, + "grad_norm": 4.989869594573975, + "learning_rate": 9.620744587815435e-05, + "loss": 3.2167, + "step": 71550 + }, + { + "epoch": 0.6322965788869461, + "grad_norm": 1.4381548166275024, + "learning_rate": 9.620213530246205e-05, + "loss": 3.5436, + "step": 71600 + }, + { + "epoch": 0.6327381267772303, + "grad_norm": 1.2252341508865356, + "learning_rate": 9.619682115804656e-05, + "loss": 3.0586, + "step": 71650 + }, + { + "epoch": 0.6331796746675145, + "grad_norm": 4.090355396270752, + "learning_rate": 9.619150344531838e-05, + "loss": 3.4366, + "step": 71700 + }, + { + "epoch": 0.6336212225577986, + "grad_norm": 1.2702091932296753, + "learning_rate": 9.618618216468824e-05, + "loss": 2.9698, + "step": 71750 + }, + { + "epoch": 0.6340627704480828, + "grad_norm": 3.03657865524292, + "learning_rate": 9.61808573165672e-05, + "loss": 3.0061, + "step": 71800 + }, + { + "epoch": 0.634504318338367, + "grad_norm": 2.344271421432495, + "learning_rate": 9.617552890136652e-05, + "loss": 2.6665, + "step": 71850 + }, + { + "epoch": 0.6349458662286511, + "grad_norm": 1.6858032941818237, + "learning_rate": 9.61701969194978e-05, + "loss": 3.1685, + "step": 71900 + }, + { + "epoch": 0.6353874141189353, + "grad_norm": 3.873769760131836, + "learning_rate": 9.616486137137287e-05, + "loss": 3.5466, + "step": 71950 + }, + { + "epoch": 0.6358289620092196, + "grad_norm": 1.7075845003128052, + "learning_rate": 9.615952225740385e-05, + "loss": 3.3663, + "step": 72000 + }, + { + "epoch": 0.6358289620092196, + "eval_asr_loss": 0.9085645562718005, + "eval_loss": 3.0017716884613037, + "eval_runtime": 20.9003, + "eval_samples_per_second": 36.746, + "eval_steps_per_second": 9.186, + "eval_tts_loss": 5.94018095728795, + "step": 72000 + }, + { + "epoch": 0.6362705098995037, + "grad_norm": 2.023348331451416, + "learning_rate": 9.615417957800319e-05, + "loss": 3.2043, + "step": 72050 + }, + { + "epoch": 0.6367120577897879, + "grad_norm": 1.3778334856033325, + "learning_rate": 9.614883333358351e-05, + "loss": 3.2908, + "step": 72100 + }, + { + "epoch": 0.6371536056800721, + "grad_norm": 2.4582135677337646, + "learning_rate": 9.61434835245578e-05, + "loss": 3.4521, + "step": 72150 + }, + { + "epoch": 0.6375951535703562, + "grad_norm": 4.403940200805664, + "learning_rate": 9.613813015133926e-05, + "loss": 3.1724, + "step": 72200 + }, + { + "epoch": 0.6380367014606404, + "grad_norm": 1.1220415830612183, + "learning_rate": 9.613277321434141e-05, + "loss": 3.2921, + "step": 72250 + }, + { + "epoch": 0.6384782493509246, + "grad_norm": 5.00771427154541, + "learning_rate": 9.612741271397802e-05, + "loss": 3.4677, + "step": 72300 + }, + { + "epoch": 0.6389197972412087, + "grad_norm": 2.081688404083252, + "learning_rate": 9.612204865066317e-05, + "loss": 2.9708, + "step": 72350 + }, + { + "epoch": 0.639361345131493, + "grad_norm": 1.0858304500579834, + "learning_rate": 9.611668102481114e-05, + "loss": 3.3104, + "step": 72400 + }, + { + "epoch": 0.6398028930217772, + "grad_norm": 4.691798686981201, + "learning_rate": 9.611141729550216e-05, + "loss": 3.4249, + "step": 72450 + }, + { + "epoch": 0.6402444409120613, + "grad_norm": 1.6770505905151367, + "learning_rate": 9.610604261705e-05, + "loss": 3.3248, + "step": 72500 + }, + { + "epoch": 0.6406859888023455, + "grad_norm": 0.5892982482910156, + "learning_rate": 9.610066437729704e-05, + "loss": 3.2216, + "step": 72550 + }, + { + "epoch": 0.6411275366926297, + "grad_norm": 2.631664514541626, + "learning_rate": 9.609528257665864e-05, + "loss": 3.4193, + "step": 72600 + }, + { + "epoch": 0.6415690845829138, + "grad_norm": 1.2707042694091797, + "learning_rate": 9.608989721555055e-05, + "loss": 3.3635, + "step": 72650 + }, + { + "epoch": 0.642010632473198, + "grad_norm": 3.7268316745758057, + "learning_rate": 9.608450829438873e-05, + "loss": 3.5901, + "step": 72700 + }, + { + "epoch": 0.6424521803634822, + "grad_norm": 3.723600387573242, + "learning_rate": 9.607911581358943e-05, + "loss": 3.157, + "step": 72750 + }, + { + "epoch": 0.6428937282537664, + "grad_norm": 1.398940086364746, + "learning_rate": 9.607371977356918e-05, + "loss": 3.2389, + "step": 72800 + }, + { + "epoch": 0.6433352761440506, + "grad_norm": 3.619192600250244, + "learning_rate": 9.606832017474477e-05, + "loss": 3.5136, + "step": 72850 + }, + { + "epoch": 0.6437768240343348, + "grad_norm": 1.5558445453643799, + "learning_rate": 9.606291701753327e-05, + "loss": 3.4748, + "step": 72900 + }, + { + "epoch": 0.644218371924619, + "grad_norm": 3.18674635887146, + "learning_rate": 9.605751030235204e-05, + "loss": 2.6554, + "step": 72950 + }, + { + "epoch": 0.6446599198149031, + "grad_norm": 4.52305793762207, + "learning_rate": 9.605210002961868e-05, + "loss": 3.6155, + "step": 73000 + }, + { + "epoch": 0.6451014677051873, + "grad_norm": 5.28046989440918, + "learning_rate": 9.604668619975112e-05, + "loss": 3.404, + "step": 73050 + }, + { + "epoch": 0.6455430155954714, + "grad_norm": 2.169076442718506, + "learning_rate": 9.604126881316751e-05, + "loss": 3.3904, + "step": 73100 + }, + { + "epoch": 0.6459845634857556, + "grad_norm": 1.4948843717575073, + "learning_rate": 9.60358478702863e-05, + "loss": 3.1796, + "step": 73150 + }, + { + "epoch": 0.6464261113760399, + "grad_norm": 3.89113450050354, + "learning_rate": 9.603042337152622e-05, + "loss": 3.6965, + "step": 73200 + }, + { + "epoch": 0.6468676592663241, + "grad_norm": 0.8911652565002441, + "learning_rate": 9.602499531730625e-05, + "loss": 3.4166, + "step": 73250 + }, + { + "epoch": 0.6473092071566082, + "grad_norm": 1.3251335620880127, + "learning_rate": 9.601956370804567e-05, + "loss": 3.0175, + "step": 73300 + }, + { + "epoch": 0.6477507550468924, + "grad_norm": 3.875401735305786, + "learning_rate": 9.601412854416403e-05, + "loss": 3.5598, + "step": 73350 + }, + { + "epoch": 0.6481923029371766, + "grad_norm": 2.4360857009887695, + "learning_rate": 9.600868982608116e-05, + "loss": 3.3592, + "step": 73400 + }, + { + "epoch": 0.6486338508274607, + "grad_norm": 2.4810409545898438, + "learning_rate": 9.600324755421715e-05, + "loss": 3.4158, + "step": 73450 + }, + { + "epoch": 0.6490753987177449, + "grad_norm": 1.3619318008422852, + "learning_rate": 9.599780172899234e-05, + "loss": 2.9944, + "step": 73500 + }, + { + "epoch": 0.649516946608029, + "grad_norm": 2.7849841117858887, + "learning_rate": 9.599235235082741e-05, + "loss": 3.6618, + "step": 73550 + }, + { + "epoch": 0.6499584944983133, + "grad_norm": 5.122074127197266, + "learning_rate": 9.598689942014326e-05, + "loss": 3.0184, + "step": 73600 + }, + { + "epoch": 0.6504000423885975, + "grad_norm": 2.697990655899048, + "learning_rate": 9.59814429373611e-05, + "loss": 3.0678, + "step": 73650 + }, + { + "epoch": 0.6508415902788817, + "grad_norm": 1.3273183107376099, + "learning_rate": 9.597598290290237e-05, + "loss": 2.9799, + "step": 73700 + }, + { + "epoch": 0.6512831381691658, + "grad_norm": 1.5726739168167114, + "learning_rate": 9.597051931718882e-05, + "loss": 3.1769, + "step": 73750 + }, + { + "epoch": 0.65172468605945, + "grad_norm": 2.7319064140319824, + "learning_rate": 9.596505218064249e-05, + "loss": 3.5969, + "step": 73800 + }, + { + "epoch": 0.6521662339497342, + "grad_norm": 3.6344516277313232, + "learning_rate": 9.595958149368563e-05, + "loss": 3.2977, + "step": 73850 + }, + { + "epoch": 0.6526077818400183, + "grad_norm": 2.6774423122406006, + "learning_rate": 9.595410725674084e-05, + "loss": 3.1628, + "step": 73900 + }, + { + "epoch": 0.6530493297303025, + "grad_norm": 4.738192081451416, + "learning_rate": 9.594862947023094e-05, + "loss": 3.6289, + "step": 73950 + }, + { + "epoch": 0.6534908776205868, + "grad_norm": 1.0268847942352295, + "learning_rate": 9.594314813457905e-05, + "loss": 3.6385, + "step": 74000 + }, + { + "epoch": 0.6539324255108709, + "grad_norm": 2.7158186435699463, + "learning_rate": 9.593766325020855e-05, + "loss": 3.4732, + "step": 74050 + }, + { + "epoch": 0.6543739734011551, + "grad_norm": 2.4069275856018066, + "learning_rate": 9.593217481754311e-05, + "loss": 3.493, + "step": 74100 + }, + { + "epoch": 0.6548155212914393, + "grad_norm": 1.4273890256881714, + "learning_rate": 9.592668283700665e-05, + "loss": 3.4282, + "step": 74150 + }, + { + "epoch": 0.6552570691817234, + "grad_norm": 1.4611910581588745, + "learning_rate": 9.59211873090234e-05, + "loss": 3.5503, + "step": 74200 + }, + { + "epoch": 0.6556986170720076, + "grad_norm": 1.2271146774291992, + "learning_rate": 9.591568823401782e-05, + "loss": 3.081, + "step": 74250 + }, + { + "epoch": 0.6561401649622918, + "grad_norm": 2.7792177200317383, + "learning_rate": 9.591018561241467e-05, + "loss": 3.4418, + "step": 74300 + }, + { + "epoch": 0.6565817128525759, + "grad_norm": 5.776242733001709, + "learning_rate": 9.5904679444639e-05, + "loss": 3.5402, + "step": 74350 + }, + { + "epoch": 0.6570232607428602, + "grad_norm": 2.6737334728240967, + "learning_rate": 9.58991697311161e-05, + "loss": 3.2469, + "step": 74400 + }, + { + "epoch": 0.6574648086331444, + "grad_norm": 1.3555004596710205, + "learning_rate": 9.589365647227155e-05, + "loss": 3.3199, + "step": 74450 + }, + { + "epoch": 0.6579063565234285, + "grad_norm": 3.3269264698028564, + "learning_rate": 9.58881396685312e-05, + "loss": 3.462, + "step": 74500 + }, + { + "epoch": 0.6583479044137127, + "grad_norm": 2.6994597911834717, + "learning_rate": 9.58826193203212e-05, + "loss": 3.3227, + "step": 74550 + }, + { + "epoch": 0.6587894523039969, + "grad_norm": 3.3708934783935547, + "learning_rate": 9.587709542806792e-05, + "loss": 3.1164, + "step": 74600 + }, + { + "epoch": 0.659231000194281, + "grad_norm": 1.3764150142669678, + "learning_rate": 9.587156799219804e-05, + "loss": 3.3193, + "step": 74650 + }, + { + "epoch": 0.6596725480845652, + "grad_norm": 4.554634094238281, + "learning_rate": 9.586603701313852e-05, + "loss": 3.5422, + "step": 74700 + }, + { + "epoch": 0.6601140959748494, + "grad_norm": 4.359382152557373, + "learning_rate": 9.586050249131658e-05, + "loss": 3.2248, + "step": 74750 + }, + { + "epoch": 0.6605556438651337, + "grad_norm": 4.0737714767456055, + "learning_rate": 9.585496442715968e-05, + "loss": 3.3829, + "step": 74800 + }, + { + "epoch": 0.6609971917554178, + "grad_norm": 2.545942544937134, + "learning_rate": 9.584942282109564e-05, + "loss": 3.2756, + "step": 74850 + }, + { + "epoch": 0.661438739645702, + "grad_norm": 3.28908109664917, + "learning_rate": 9.584387767355247e-05, + "loss": 3.4446, + "step": 74900 + }, + { + "epoch": 0.6618802875359862, + "grad_norm": 5.4131340980529785, + "learning_rate": 9.58383289849585e-05, + "loss": 2.9662, + "step": 74950 + }, + { + "epoch": 0.6623218354262703, + "grad_norm": 1.5948145389556885, + "learning_rate": 9.58327767557423e-05, + "loss": 3.3822, + "step": 75000 + }, + { + "epoch": 0.6623218354262703, + "eval_asr_loss": 0.8949383808222489, + "eval_loss": 3.007831573486328, + "eval_runtime": 20.8134, + "eval_samples_per_second": 36.899, + "eval_steps_per_second": 9.225, + "eval_tts_loss": 5.953438474404635, + "step": 75000 + }, + { + "epoch": 0.6627633833165545, + "grad_norm": 3.2304179668426514, + "learning_rate": 9.582722098633276e-05, + "loss": 3.3734, + "step": 75050 + }, + { + "epoch": 0.6632049312068387, + "grad_norm": 3.123548746109009, + "learning_rate": 9.5821661677159e-05, + "loss": 3.4995, + "step": 75100 + }, + { + "epoch": 0.6636464790971228, + "grad_norm": 2.3339643478393555, + "learning_rate": 9.581609882865044e-05, + "loss": 3.3416, + "step": 75150 + }, + { + "epoch": 0.6640880269874071, + "grad_norm": 1.0405722856521606, + "learning_rate": 9.58106438036635e-05, + "loss": 3.3155, + "step": 75200 + }, + { + "epoch": 0.6645295748776913, + "grad_norm": 3.653280258178711, + "learning_rate": 9.580507394853993e-05, + "loss": 2.9206, + "step": 75250 + }, + { + "epoch": 0.6649711227679754, + "grad_norm": 2.6631522178649902, + "learning_rate": 9.579950055536279e-05, + "loss": 3.1143, + "step": 75300 + }, + { + "epoch": 0.6654126706582596, + "grad_norm": 2.752356767654419, + "learning_rate": 9.579392362456264e-05, + "loss": 3.4541, + "step": 75350 + }, + { + "epoch": 0.6658542185485438, + "grad_norm": 1.8876862525939941, + "learning_rate": 9.578834315657019e-05, + "loss": 2.942, + "step": 75400 + }, + { + "epoch": 0.6662957664388279, + "grad_norm": 3.700727939605713, + "learning_rate": 9.57827591518165e-05, + "loss": 3.4877, + "step": 75450 + }, + { + "epoch": 0.6667373143291121, + "grad_norm": 1.4538304805755615, + "learning_rate": 9.57771716107329e-05, + "loss": 3.5755, + "step": 75500 + }, + { + "epoch": 0.6671788622193963, + "grad_norm": 4.2887349128723145, + "learning_rate": 9.577158053375096e-05, + "loss": 3.3367, + "step": 75550 + }, + { + "epoch": 0.6676204101096805, + "grad_norm": 4.3281025886535645, + "learning_rate": 9.576598592130257e-05, + "loss": 3.5268, + "step": 75600 + }, + { + "epoch": 0.6680619579999647, + "grad_norm": 2.3635294437408447, + "learning_rate": 9.576038777381984e-05, + "loss": 3.6769, + "step": 75650 + }, + { + "epoch": 0.6685035058902489, + "grad_norm": 4.012321949005127, + "learning_rate": 9.57547860917352e-05, + "loss": 3.1022, + "step": 75700 + }, + { + "epoch": 0.668945053780533, + "grad_norm": 1.8018879890441895, + "learning_rate": 9.574918087548132e-05, + "loss": 3.4172, + "step": 75750 + }, + { + "epoch": 0.6693866016708172, + "grad_norm": 2.358532190322876, + "learning_rate": 9.574357212549116e-05, + "loss": 3.1392, + "step": 75800 + }, + { + "epoch": 0.6698281495611014, + "grad_norm": 2.4655981063842773, + "learning_rate": 9.573795984219793e-05, + "loss": 3.3193, + "step": 75850 + }, + { + "epoch": 0.6702696974513855, + "grad_norm": 2.9911704063415527, + "learning_rate": 9.573234402603517e-05, + "loss": 3.563, + "step": 75900 + }, + { + "epoch": 0.6707112453416697, + "grad_norm": 5.0430707931518555, + "learning_rate": 9.572672467743664e-05, + "loss": 3.2619, + "step": 75950 + }, + { + "epoch": 0.671152793231954, + "grad_norm": 2.1028594970703125, + "learning_rate": 9.572110179683637e-05, + "loss": 3.6662, + "step": 76000 + }, + { + "epoch": 0.6715943411222381, + "grad_norm": 2.4186270236968994, + "learning_rate": 9.571547538466868e-05, + "loss": 3.0861, + "step": 76050 + }, + { + "epoch": 0.6720358890125223, + "grad_norm": 1.6107758283615112, + "learning_rate": 9.570984544136819e-05, + "loss": 3.9137, + "step": 76100 + }, + { + "epoch": 0.6724774369028065, + "grad_norm": 3.140552043914795, + "learning_rate": 9.570432467144772e-05, + "loss": 3.643, + "step": 76150 + }, + { + "epoch": 0.6729189847930906, + "grad_norm": 3.5983471870422363, + "learning_rate": 9.569868773778745e-05, + "loss": 3.6072, + "step": 76200 + }, + { + "epoch": 0.6733605326833748, + "grad_norm": 3.1312403678894043, + "learning_rate": 9.569304727429107e-05, + "loss": 3.3982, + "step": 76250 + }, + { + "epoch": 0.673802080573659, + "grad_norm": 1.2464442253112793, + "learning_rate": 9.568740328139423e-05, + "loss": 3.36, + "step": 76300 + }, + { + "epoch": 0.6742436284639431, + "grad_norm": 1.5809202194213867, + "learning_rate": 9.568175575953293e-05, + "loss": 3.404, + "step": 76350 + }, + { + "epoch": 0.6746851763542274, + "grad_norm": 2.486264944076538, + "learning_rate": 9.567610470914334e-05, + "loss": 3.4402, + "step": 76400 + }, + { + "epoch": 0.6751267242445116, + "grad_norm": 1.804176926612854, + "learning_rate": 9.5670450130662e-05, + "loss": 3.2879, + "step": 76450 + }, + { + "epoch": 0.6755682721347958, + "grad_norm": 1.7221146821975708, + "learning_rate": 9.566479202452567e-05, + "loss": 3.2473, + "step": 76500 + }, + { + "epoch": 0.6760098200250799, + "grad_norm": 2.4581539630889893, + "learning_rate": 9.565913039117138e-05, + "loss": 3.7328, + "step": 76550 + }, + { + "epoch": 0.6764513679153641, + "grad_norm": 1.7588177919387817, + "learning_rate": 9.565346523103643e-05, + "loss": 3.1056, + "step": 76600 + }, + { + "epoch": 0.6768929158056483, + "grad_norm": 2.841600179672241, + "learning_rate": 9.564779654455844e-05, + "loss": 3.3631, + "step": 76650 + }, + { + "epoch": 0.6773344636959324, + "grad_norm": 2.2738077640533447, + "learning_rate": 9.564212433217523e-05, + "loss": 3.4339, + "step": 76700 + }, + { + "epoch": 0.6777760115862166, + "grad_norm": 1.1713546514511108, + "learning_rate": 9.563644859432497e-05, + "loss": 3.2702, + "step": 76750 + }, + { + "epoch": 0.6782175594765009, + "grad_norm": 2.2375364303588867, + "learning_rate": 9.563076933144602e-05, + "loss": 3.6829, + "step": 76800 + }, + { + "epoch": 0.678659107366785, + "grad_norm": 1.015380620956421, + "learning_rate": 9.562508654397708e-05, + "loss": 3.0635, + "step": 76850 + }, + { + "epoch": 0.6791006552570692, + "grad_norm": 0.6230117082595825, + "learning_rate": 9.56194002323571e-05, + "loss": 3.5006, + "step": 76900 + }, + { + "epoch": 0.6795422031473534, + "grad_norm": 1.3702956438064575, + "learning_rate": 9.56137103970253e-05, + "loss": 3.107, + "step": 76950 + }, + { + "epoch": 0.6799837510376375, + "grad_norm": 1.878818392753601, + "learning_rate": 9.560801703842117e-05, + "loss": 3.6604, + "step": 77000 + }, + { + "epoch": 0.6804252989279217, + "grad_norm": 1.3559465408325195, + "learning_rate": 9.560232015698444e-05, + "loss": 3.634, + "step": 77050 + }, + { + "epoch": 0.6808668468182059, + "grad_norm": 15.994784355163574, + "learning_rate": 9.559661975315519e-05, + "loss": 3.4924, + "step": 77100 + }, + { + "epoch": 0.68130839470849, + "grad_norm": 0.9013010859489441, + "learning_rate": 9.559091582737372e-05, + "loss": 3.474, + "step": 77150 + }, + { + "epoch": 0.6817499425987743, + "grad_norm": 1.3329524993896484, + "learning_rate": 9.558520838008059e-05, + "loss": 3.4002, + "step": 77200 + }, + { + "epoch": 0.6821914904890585, + "grad_norm": 2.5802180767059326, + "learning_rate": 9.557949741171667e-05, + "loss": 3.1695, + "step": 77250 + }, + { + "epoch": 0.6826330383793426, + "grad_norm": 2.464118242263794, + "learning_rate": 9.557378292272307e-05, + "loss": 3.337, + "step": 77300 + }, + { + "epoch": 0.6830745862696268, + "grad_norm": 1.7328948974609375, + "learning_rate": 9.556806491354121e-05, + "loss": 3.085, + "step": 77350 + }, + { + "epoch": 0.683516134159911, + "grad_norm": 1.6609792709350586, + "learning_rate": 9.556234338461274e-05, + "loss": 3.7844, + "step": 77400 + }, + { + "epoch": 0.6839576820501951, + "grad_norm": 3.412128210067749, + "learning_rate": 9.55566183363796e-05, + "loss": 3.3013, + "step": 77450 + }, + { + "epoch": 0.6843992299404793, + "grad_norm": 1.4039990901947021, + "learning_rate": 9.555088976928399e-05, + "loss": 3.2333, + "step": 77500 + }, + { + "epoch": 0.6848407778307635, + "grad_norm": 1.7826491594314575, + "learning_rate": 9.55451576837684e-05, + "loss": 3.4012, + "step": 77550 + }, + { + "epoch": 0.6852823257210477, + "grad_norm": 3.7740890979766846, + "learning_rate": 9.553942208027559e-05, + "loss": 3.2525, + "step": 77600 + }, + { + "epoch": 0.6857238736113319, + "grad_norm": 3.7978830337524414, + "learning_rate": 9.553368295924861e-05, + "loss": 3.3832, + "step": 77650 + }, + { + "epoch": 0.6861654215016161, + "grad_norm": 1.697966456413269, + "learning_rate": 9.552794032113073e-05, + "loss": 3.2114, + "step": 77700 + }, + { + "epoch": 0.6866069693919002, + "grad_norm": 1.4913134574890137, + "learning_rate": 9.552219416636552e-05, + "loss": 3.1685, + "step": 77750 + }, + { + "epoch": 0.6870485172821844, + "grad_norm": 2.034682512283325, + "learning_rate": 9.551644449539683e-05, + "loss": 2.8857, + "step": 77800 + }, + { + "epoch": 0.6874900651724686, + "grad_norm": 1.9270625114440918, + "learning_rate": 9.551069130866877e-05, + "loss": 3.5445, + "step": 77850 + }, + { + "epoch": 0.6879316130627527, + "grad_norm": 5.204434871673584, + "learning_rate": 9.550493460662572e-05, + "loss": 3.3383, + "step": 77900 + }, + { + "epoch": 0.6883731609530369, + "grad_norm": 3.010862350463867, + "learning_rate": 9.549917438971235e-05, + "loss": 3.3101, + "step": 77950 + }, + { + "epoch": 0.6888147088433212, + "grad_norm": 0.8644163608551025, + "learning_rate": 9.549341065837357e-05, + "loss": 3.3201, + "step": 78000 + }, + { + "epoch": 0.6888147088433212, + "eval_asr_loss": 0.8925247313395483, + "eval_loss": 2.992865800857544, + "eval_runtime": 20.774, + "eval_samples_per_second": 36.969, + "eval_steps_per_second": 9.242, + "eval_tts_loss": 5.960387531532541, + "step": 78000 + }, + { + "epoch": 0.6892562567336054, + "grad_norm": 2.4011969566345215, + "learning_rate": 9.54876434130546e-05, + "loss": 3.9292, + "step": 78050 + }, + { + "epoch": 0.6896978046238895, + "grad_norm": 3.125196695327759, + "learning_rate": 9.54818726542009e-05, + "loss": 3.247, + "step": 78100 + }, + { + "epoch": 0.6901393525141737, + "grad_norm": 1.368432879447937, + "learning_rate": 9.547609838225821e-05, + "loss": 2.951, + "step": 78150 + }, + { + "epoch": 0.6905809004044579, + "grad_norm": 2.4032599925994873, + "learning_rate": 9.547032059767253e-05, + "loss": 3.5267, + "step": 78200 + }, + { + "epoch": 0.691022448294742, + "grad_norm": 2.1227686405181885, + "learning_rate": 9.546453930089019e-05, + "loss": 3.2206, + "step": 78250 + }, + { + "epoch": 0.6914639961850262, + "grad_norm": 1.2852637767791748, + "learning_rate": 9.54587544923577e-05, + "loss": 3.4086, + "step": 78300 + }, + { + "epoch": 0.6919055440753105, + "grad_norm": 1.241897463798523, + "learning_rate": 9.54529661725219e-05, + "loss": 3.1159, + "step": 78350 + }, + { + "epoch": 0.6923470919655946, + "grad_norm": 1.472129225730896, + "learning_rate": 9.544717434182991e-05, + "loss": 3.2442, + "step": 78400 + }, + { + "epoch": 0.6927886398558788, + "grad_norm": 2.743114471435547, + "learning_rate": 9.544137900072907e-05, + "loss": 3.2978, + "step": 78450 + }, + { + "epoch": 0.693230187746163, + "grad_norm": 2.720283031463623, + "learning_rate": 9.543558014966703e-05, + "loss": 3.5339, + "step": 78500 + }, + { + "epoch": 0.6936717356364471, + "grad_norm": 2.2413861751556396, + "learning_rate": 9.542977778909173e-05, + "loss": 3.6654, + "step": 78550 + }, + { + "epoch": 0.6941132835267313, + "grad_norm": 0.8887834548950195, + "learning_rate": 9.542397191945129e-05, + "loss": 2.9958, + "step": 78600 + }, + { + "epoch": 0.6945548314170155, + "grad_norm": 1.2447916269302368, + "learning_rate": 9.541816254119425e-05, + "loss": 3.0432, + "step": 78650 + }, + { + "epoch": 0.6949963793072996, + "grad_norm": 2.320293664932251, + "learning_rate": 9.541234965476925e-05, + "loss": 3.4944, + "step": 78700 + }, + { + "epoch": 0.6954379271975839, + "grad_norm": 2.7362334728240967, + "learning_rate": 9.540653326062534e-05, + "loss": 3.4683, + "step": 78750 + }, + { + "epoch": 0.6958794750878681, + "grad_norm": 1.2431176900863647, + "learning_rate": 9.540071335921176e-05, + "loss": 3.2625, + "step": 78800 + }, + { + "epoch": 0.6963210229781522, + "grad_norm": 2.1052627563476562, + "learning_rate": 9.539488995097807e-05, + "loss": 3.3472, + "step": 78850 + }, + { + "epoch": 0.6967625708684364, + "grad_norm": 2.4554333686828613, + "learning_rate": 9.538906303637407e-05, + "loss": 3.327, + "step": 78900 + }, + { + "epoch": 0.6972041187587206, + "grad_norm": 2.1302428245544434, + "learning_rate": 9.538323261584985e-05, + "loss": 3.4227, + "step": 78950 + }, + { + "epoch": 0.6976456666490047, + "grad_norm": 1.9906952381134033, + "learning_rate": 9.537739868985574e-05, + "loss": 3.8786, + "step": 79000 + }, + { + "epoch": 0.6980872145392889, + "grad_norm": 4.157954216003418, + "learning_rate": 9.537156125884236e-05, + "loss": 3.0816, + "step": 79050 + }, + { + "epoch": 0.6985287624295731, + "grad_norm": 4.053858757019043, + "learning_rate": 9.536572032326061e-05, + "loss": 3.2263, + "step": 79100 + }, + { + "epoch": 0.6989703103198573, + "grad_norm": 1.9072597026824951, + "learning_rate": 9.535987588356168e-05, + "loss": 3.4913, + "step": 79150 + }, + { + "epoch": 0.6994118582101415, + "grad_norm": 4.11290168762207, + "learning_rate": 9.535402794019696e-05, + "loss": 3.5914, + "step": 79200 + }, + { + "epoch": 0.6998534061004257, + "grad_norm": 3.6236138343811035, + "learning_rate": 9.534817649361816e-05, + "loss": 3.0452, + "step": 79250 + }, + { + "epoch": 0.7002949539907098, + "grad_norm": 4.064155578613281, + "learning_rate": 9.534232154427728e-05, + "loss": 3.5526, + "step": 79300 + }, + { + "epoch": 0.700736501880994, + "grad_norm": 2.754270553588867, + "learning_rate": 9.533646309262657e-05, + "loss": 3.3366, + "step": 79350 + }, + { + "epoch": 0.7011780497712782, + "grad_norm": 2.6481540203094482, + "learning_rate": 9.533060113911852e-05, + "loss": 3.6007, + "step": 79400 + }, + { + "epoch": 0.7016195976615623, + "grad_norm": 3.322692632675171, + "learning_rate": 9.532473568420591e-05, + "loss": 3.1894, + "step": 79450 + }, + { + "epoch": 0.7020611455518465, + "grad_norm": 2.9386980533599854, + "learning_rate": 9.531886672834182e-05, + "loss": 3.1106, + "step": 79500 + }, + { + "epoch": 0.7025026934421308, + "grad_norm": 1.7416975498199463, + "learning_rate": 9.531299427197957e-05, + "loss": 3.5195, + "step": 79550 + }, + { + "epoch": 0.702944241332415, + "grad_norm": 2.034696578979492, + "learning_rate": 9.530711831557274e-05, + "loss": 3.3598, + "step": 79600 + }, + { + "epoch": 0.7033857892226991, + "grad_norm": 2.048043966293335, + "learning_rate": 9.530123885957523e-05, + "loss": 3.3274, + "step": 79650 + }, + { + "epoch": 0.7038273371129833, + "grad_norm": 0.8644744157791138, + "learning_rate": 9.529535590444115e-05, + "loss": 3.3067, + "step": 79700 + }, + { + "epoch": 0.7042688850032675, + "grad_norm": 3.338059186935425, + "learning_rate": 9.528946945062493e-05, + "loss": 3.2878, + "step": 79750 + }, + { + "epoch": 0.7047104328935516, + "grad_norm": 1.424256682395935, + "learning_rate": 9.528357949858123e-05, + "loss": 3.5114, + "step": 79800 + }, + { + "epoch": 0.7051519807838358, + "grad_norm": 3.7967846393585205, + "learning_rate": 9.527768604876501e-05, + "loss": 3.4413, + "step": 79850 + }, + { + "epoch": 0.70559352867412, + "grad_norm": 3.256422996520996, + "learning_rate": 9.52717891016315e-05, + "loss": 3.0054, + "step": 79900 + }, + { + "epoch": 0.7060350765644042, + "grad_norm": 3.2881686687469482, + "learning_rate": 9.526588865763614e-05, + "loss": 3.3926, + "step": 79950 + }, + { + "epoch": 0.7064766244546884, + "grad_norm": 2.3848628997802734, + "learning_rate": 9.525998471723476e-05, + "loss": 3.3552, + "step": 80000 + }, + { + "epoch": 0.7069181723449726, + "grad_norm": 3.0554099082946777, + "learning_rate": 9.525407728088333e-05, + "loss": 3.4239, + "step": 80050 + }, + { + "epoch": 0.7073597202352567, + "grad_norm": 2.6418068408966064, + "learning_rate": 9.524816634903819e-05, + "loss": 3.2979, + "step": 80100 + }, + { + "epoch": 0.7078012681255409, + "grad_norm": 1.501378059387207, + "learning_rate": 9.52422519221559e-05, + "loss": 3.7132, + "step": 80150 + }, + { + "epoch": 0.7082428160158251, + "grad_norm": 3.162963390350342, + "learning_rate": 9.523633400069327e-05, + "loss": 3.7012, + "step": 80200 + }, + { + "epoch": 0.7086843639061092, + "grad_norm": 0.6245143413543701, + "learning_rate": 9.523041258510746e-05, + "loss": 3.423, + "step": 80250 + }, + { + "epoch": 0.7091259117963934, + "grad_norm": 0.8148535490036011, + "learning_rate": 9.522448767585579e-05, + "loss": 3.3979, + "step": 80300 + }, + { + "epoch": 0.7095674596866777, + "grad_norm": 2.1384999752044678, + "learning_rate": 9.521855927339595e-05, + "loss": 2.7129, + "step": 80350 + }, + { + "epoch": 0.7100090075769618, + "grad_norm": 3.697734832763672, + "learning_rate": 9.521262737818586e-05, + "loss": 3.0784, + "step": 80400 + }, + { + "epoch": 0.710450555467246, + "grad_norm": 6.074985027313232, + "learning_rate": 9.520669199068368e-05, + "loss": 3.4716, + "step": 80450 + }, + { + "epoch": 0.7108921033575302, + "grad_norm": 3.228442907333374, + "learning_rate": 9.520075311134788e-05, + "loss": 3.3426, + "step": 80500 + }, + { + "epoch": 0.7113336512478143, + "grad_norm": 1.2277064323425293, + "learning_rate": 9.51948107406372e-05, + "loss": 3.2555, + "step": 80550 + }, + { + "epoch": 0.7117751991380985, + "grad_norm": 0.9551520943641663, + "learning_rate": 9.518886487901065e-05, + "loss": 3.5223, + "step": 80600 + }, + { + "epoch": 0.7122167470283827, + "grad_norm": 2.2914512157440186, + "learning_rate": 9.518291552692746e-05, + "loss": 2.9172, + "step": 80650 + }, + { + "epoch": 0.7126582949186668, + "grad_norm": 3.3304460048675537, + "learning_rate": 9.517696268484719e-05, + "loss": 2.6855, + "step": 80700 + }, + { + "epoch": 0.7130998428089511, + "grad_norm": 1.3867692947387695, + "learning_rate": 9.517100635322966e-05, + "loss": 2.9664, + "step": 80750 + }, + { + "epoch": 0.7135413906992353, + "grad_norm": 1.0515697002410889, + "learning_rate": 9.51650465325349e-05, + "loss": 3.2619, + "step": 80800 + }, + { + "epoch": 0.7139829385895194, + "grad_norm": 1.7852091789245605, + "learning_rate": 9.515908322322329e-05, + "loss": 3.0423, + "step": 80850 + }, + { + "epoch": 0.7144244864798036, + "grad_norm": 1.8944436311721802, + "learning_rate": 9.515311642575545e-05, + "loss": 3.1392, + "step": 80900 + }, + { + "epoch": 0.7148660343700878, + "grad_norm": 0.9419056177139282, + "learning_rate": 9.514714614059225e-05, + "loss": 3.1358, + "step": 80950 + }, + { + "epoch": 0.7153075822603719, + "grad_norm": 3.2333171367645264, + "learning_rate": 9.514117236819485e-05, + "loss": 3.8758, + "step": 81000 + }, + { + "epoch": 0.7153075822603719, + "eval_asr_loss": 0.9185007565241553, + "eval_loss": 2.989028215408325, + "eval_runtime": 20.4458, + "eval_samples_per_second": 37.563, + "eval_steps_per_second": 9.391, + "eval_tts_loss": 5.958070210757005, + "step": 81000 + }, + { + "epoch": 0.7157491301506561, + "grad_norm": 2.493988037109375, + "learning_rate": 9.513519510902466e-05, + "loss": 3.6492, + "step": 81050 + }, + { + "epoch": 0.7161906780409403, + "grad_norm": 2.0979015827178955, + "learning_rate": 9.512921436354339e-05, + "loss": 3.7933, + "step": 81100 + }, + { + "epoch": 0.7166322259312246, + "grad_norm": 2.046807050704956, + "learning_rate": 9.512323013221299e-05, + "loss": 3.291, + "step": 81150 + }, + { + "epoch": 0.7170737738215087, + "grad_norm": 2.943680763244629, + "learning_rate": 9.51172424154957e-05, + "loss": 3.0745, + "step": 81200 + }, + { + "epoch": 0.7175153217117929, + "grad_norm": 2.4897234439849854, + "learning_rate": 9.5111251213854e-05, + "loss": 3.1611, + "step": 81250 + }, + { + "epoch": 0.717956869602077, + "grad_norm": 4.505876541137695, + "learning_rate": 9.51052565277507e-05, + "loss": 3.5499, + "step": 81300 + }, + { + "epoch": 0.7183984174923612, + "grad_norm": 1.430411458015442, + "learning_rate": 9.50992583576488e-05, + "loss": 3.1281, + "step": 81350 + }, + { + "epoch": 0.7188399653826454, + "grad_norm": 3.1512632369995117, + "learning_rate": 9.509325670401164e-05, + "loss": 3.4918, + "step": 81400 + }, + { + "epoch": 0.7192815132729296, + "grad_norm": 1.2546844482421875, + "learning_rate": 9.508725156730276e-05, + "loss": 3.6289, + "step": 81450 + }, + { + "epoch": 0.7197230611632137, + "grad_norm": 3.754129409790039, + "learning_rate": 9.508124294798604e-05, + "loss": 2.9913, + "step": 81500 + }, + { + "epoch": 0.720164609053498, + "grad_norm": 3.811619997024536, + "learning_rate": 9.507523084652556e-05, + "loss": 3.233, + "step": 81550 + }, + { + "epoch": 0.7206061569437822, + "grad_norm": 1.469282627105713, + "learning_rate": 9.5069335609166e-05, + "loss": 3.6196, + "step": 81600 + }, + { + "epoch": 0.7210477048340663, + "grad_norm": 2.8807554244995117, + "learning_rate": 9.50633166144312e-05, + "loss": 2.9091, + "step": 81650 + }, + { + "epoch": 0.7214892527243505, + "grad_norm": 2.3154757022857666, + "learning_rate": 9.505729413893734e-05, + "loss": 3.2319, + "step": 81700 + }, + { + "epoch": 0.7219308006146347, + "grad_norm": 1.580367922782898, + "learning_rate": 9.505126818314955e-05, + "loss": 3.5418, + "step": 81750 + }, + { + "epoch": 0.7223723485049188, + "grad_norm": 1.0496435165405273, + "learning_rate": 9.504523874753335e-05, + "loss": 3.6296, + "step": 81800 + }, + { + "epoch": 0.722813896395203, + "grad_norm": 0.9904226660728455, + "learning_rate": 9.50392058325544e-05, + "loss": 3.2764, + "step": 81850 + }, + { + "epoch": 0.7232554442854872, + "grad_norm": 1.5296332836151123, + "learning_rate": 9.503316943867876e-05, + "loss": 3.6219, + "step": 81900 + }, + { + "epoch": 0.7236969921757714, + "grad_norm": 1.79270601272583, + "learning_rate": 9.502712956637263e-05, + "loss": 3.467, + "step": 81950 + }, + { + "epoch": 0.7241385400660556, + "grad_norm": 3.750113010406494, + "learning_rate": 9.502108621610255e-05, + "loss": 2.9975, + "step": 82000 + }, + { + "epoch": 0.7245800879563398, + "grad_norm": 1.8261284828186035, + "learning_rate": 9.501503938833534e-05, + "loss": 3.2391, + "step": 82050 + }, + { + "epoch": 0.7250216358466239, + "grad_norm": 1.2408541440963745, + "learning_rate": 9.500911012370587e-05, + "loss": 3.4596, + "step": 82100 + }, + { + "epoch": 0.7254631837369081, + "grad_norm": 1.0297032594680786, + "learning_rate": 9.50030564118725e-05, + "loss": 3.1892, + "step": 82150 + }, + { + "epoch": 0.7259047316271923, + "grad_norm": 2.6377224922180176, + "learning_rate": 9.499699922393466e-05, + "loss": 3.624, + "step": 82200 + }, + { + "epoch": 0.7263462795174764, + "grad_norm": 3.4818551540374756, + "learning_rate": 9.499093856036019e-05, + "loss": 3.1586, + "step": 82250 + }, + { + "epoch": 0.7267878274077606, + "grad_norm": 2.5095982551574707, + "learning_rate": 9.498487442161721e-05, + "loss": 3.2636, + "step": 82300 + }, + { + "epoch": 0.7272293752980449, + "grad_norm": 1.323671579360962, + "learning_rate": 9.497880680817414e-05, + "loss": 3.3924, + "step": 82350 + }, + { + "epoch": 0.727670923188329, + "grad_norm": 3.4375061988830566, + "learning_rate": 9.497273572049967e-05, + "loss": 3.2829, + "step": 82400 + }, + { + "epoch": 0.7281124710786132, + "grad_norm": 1.7869148254394531, + "learning_rate": 9.496666115906272e-05, + "loss": 3.5715, + "step": 82450 + }, + { + "epoch": 0.7285540189688974, + "grad_norm": 2.402039051055908, + "learning_rate": 9.49605831243325e-05, + "loss": 3.3708, + "step": 82500 + }, + { + "epoch": 0.7289955668591815, + "grad_norm": 2.0941081047058105, + "learning_rate": 9.49545016167785e-05, + "loss": 3.3561, + "step": 82550 + }, + { + "epoch": 0.7294371147494657, + "grad_norm": 3.094398021697998, + "learning_rate": 9.494841663687043e-05, + "loss": 3.5185, + "step": 82600 + }, + { + "epoch": 0.7298786626397499, + "grad_norm": 1.317842960357666, + "learning_rate": 9.494232818507833e-05, + "loss": 3.3928, + "step": 82650 + }, + { + "epoch": 0.730320210530034, + "grad_norm": 2.3808138370513916, + "learning_rate": 9.493623626187249e-05, + "loss": 3.375, + "step": 82700 + }, + { + "epoch": 0.7307617584203183, + "grad_norm": 1.0954011678695679, + "learning_rate": 9.493014086772345e-05, + "loss": 3.2433, + "step": 82750 + }, + { + "epoch": 0.7312033063106025, + "grad_norm": 2.7866711616516113, + "learning_rate": 9.492404200310201e-05, + "loss": 2.8652, + "step": 82800 + }, + { + "epoch": 0.7316448542008867, + "grad_norm": 0.6492549777030945, + "learning_rate": 9.491793966847928e-05, + "loss": 3.688, + "step": 82850 + }, + { + "epoch": 0.7320864020911708, + "grad_norm": 2.792008876800537, + "learning_rate": 9.491183386432659e-05, + "loss": 3.7061, + "step": 82900 + }, + { + "epoch": 0.732527949981455, + "grad_norm": 3.021859884262085, + "learning_rate": 9.49057245911156e-05, + "loss": 3.5694, + "step": 82950 + }, + { + "epoch": 0.7329694978717392, + "grad_norm": 2.1141061782836914, + "learning_rate": 9.489961184931815e-05, + "loss": 3.1073, + "step": 83000 + }, + { + "epoch": 0.7334110457620233, + "grad_norm": 1.2735497951507568, + "learning_rate": 9.489349563940644e-05, + "loss": 3.3421, + "step": 83050 + }, + { + "epoch": 0.7338525936523075, + "grad_norm": 1.5497995615005493, + "learning_rate": 9.488737596185286e-05, + "loss": 3.6448, + "step": 83100 + }, + { + "epoch": 0.7342941415425918, + "grad_norm": 2.1992032527923584, + "learning_rate": 9.488125281713013e-05, + "loss": 3.4424, + "step": 83150 + }, + { + "epoch": 0.7347356894328759, + "grad_norm": 4.297890663146973, + "learning_rate": 9.487512620571121e-05, + "loss": 3.191, + "step": 83200 + }, + { + "epoch": 0.7351772373231601, + "grad_norm": 2.3956549167633057, + "learning_rate": 9.486899612806931e-05, + "loss": 3.0592, + "step": 83250 + }, + { + "epoch": 0.7356187852134443, + "grad_norm": 1.5929914712905884, + "learning_rate": 9.486286258467793e-05, + "loss": 3.214, + "step": 83300 + }, + { + "epoch": 0.7360603331037284, + "grad_norm": 1.5399370193481445, + "learning_rate": 9.485672557601087e-05, + "loss": 3.3211, + "step": 83350 + }, + { + "epoch": 0.7365018809940126, + "grad_norm": 3.4927520751953125, + "learning_rate": 9.48505851025421e-05, + "loss": 3.287, + "step": 83400 + }, + { + "epoch": 0.7369434288842968, + "grad_norm": 1.6479640007019043, + "learning_rate": 9.484444116474597e-05, + "loss": 3.2647, + "step": 83450 + }, + { + "epoch": 0.7373849767745809, + "grad_norm": 4.172652721405029, + "learning_rate": 9.483829376309703e-05, + "loss": 3.0049, + "step": 83500 + }, + { + "epoch": 0.7378265246648652, + "grad_norm": 2.3201801776885986, + "learning_rate": 9.48321428980701e-05, + "loss": 2.9019, + "step": 83550 + }, + { + "epoch": 0.7382680725551494, + "grad_norm": 3.2628872394561768, + "learning_rate": 9.482598857014031e-05, + "loss": 3.2633, + "step": 83600 + }, + { + "epoch": 0.7387096204454335, + "grad_norm": 1.8258609771728516, + "learning_rate": 9.481983077978302e-05, + "loss": 3.2017, + "step": 83650 + }, + { + "epoch": 0.7391511683357177, + "grad_norm": 2.0017619132995605, + "learning_rate": 9.481366952747386e-05, + "loss": 3.2454, + "step": 83700 + }, + { + "epoch": 0.7395927162260019, + "grad_norm": 2.6693084239959717, + "learning_rate": 9.480750481368872e-05, + "loss": 3.6843, + "step": 83750 + }, + { + "epoch": 0.740034264116286, + "grad_norm": 1.9272335767745972, + "learning_rate": 9.480133663890382e-05, + "loss": 3.0776, + "step": 83800 + }, + { + "epoch": 0.7404758120065702, + "grad_norm": 0.973445475101471, + "learning_rate": 9.479516500359555e-05, + "loss": 3.5528, + "step": 83850 + }, + { + "epoch": 0.7409173598968544, + "grad_norm": 1.443723201751709, + "learning_rate": 9.478898990824064e-05, + "loss": 3.3918, + "step": 83900 + }, + { + "epoch": 0.7413589077871386, + "grad_norm": 3.7001953125, + "learning_rate": 9.478281135331606e-05, + "loss": 3.3402, + "step": 83950 + }, + { + "epoch": 0.7418004556774228, + "grad_norm": 2.7895514965057373, + "learning_rate": 9.477662933929905e-05, + "loss": 3.2964, + "step": 84000 + }, + { + "epoch": 0.7418004556774228, + "eval_asr_loss": 0.907389885219024, + "eval_loss": 2.9814484119415283, + "eval_runtime": 21.1732, + "eval_samples_per_second": 36.272, + "eval_steps_per_second": 9.068, + "eval_tts_loss": 5.974965912665154, + "step": 84000 + }, + { + "epoch": 0.742242003567707, + "grad_norm": 2.343669891357422, + "learning_rate": 9.477044386666711e-05, + "loss": 3.0479, + "step": 84050 + }, + { + "epoch": 0.7426835514579911, + "grad_norm": 2.6981098651885986, + "learning_rate": 9.476425493589804e-05, + "loss": 3.538, + "step": 84100 + }, + { + "epoch": 0.7431250993482753, + "grad_norm": 2.0038726329803467, + "learning_rate": 9.475806254746984e-05, + "loss": 3.3287, + "step": 84150 + }, + { + "epoch": 0.7435666472385595, + "grad_norm": 2.3365066051483154, + "learning_rate": 9.475186670186088e-05, + "loss": 3.4029, + "step": 84200 + }, + { + "epoch": 0.7440081951288436, + "grad_norm": 3.0482475757598877, + "learning_rate": 9.474566739954966e-05, + "loss": 3.1918, + "step": 84250 + }, + { + "epoch": 0.7444497430191278, + "grad_norm": 2.1410951614379883, + "learning_rate": 9.47394646410151e-05, + "loss": 3.1646, + "step": 84300 + }, + { + "epoch": 0.7448912909094121, + "grad_norm": 2.7029786109924316, + "learning_rate": 9.473325842673626e-05, + "loss": 3.3533, + "step": 84350 + }, + { + "epoch": 0.7453328387996963, + "grad_norm": 1.4176026582717896, + "learning_rate": 9.472704875719254e-05, + "loss": 3.5932, + "step": 84400 + }, + { + "epoch": 0.7457743866899804, + "grad_norm": 3.4000325202941895, + "learning_rate": 9.472083563286358e-05, + "loss": 2.8913, + "step": 84450 + }, + { + "epoch": 0.7462159345802646, + "grad_norm": 3.8159403800964355, + "learning_rate": 9.47146190542293e-05, + "loss": 3.0607, + "step": 84500 + }, + { + "epoch": 0.7466574824705488, + "grad_norm": 2.903031349182129, + "learning_rate": 9.470839902176985e-05, + "loss": 3.1426, + "step": 84550 + }, + { + "epoch": 0.7470990303608329, + "grad_norm": 1.7477384805679321, + "learning_rate": 9.470217553596568e-05, + "loss": 3.1242, + "step": 84600 + }, + { + "epoch": 0.7475405782511171, + "grad_norm": 1.4132603406906128, + "learning_rate": 9.469594859729753e-05, + "loss": 3.1462, + "step": 84650 + }, + { + "epoch": 0.7479821261414012, + "grad_norm": 1.7881561517715454, + "learning_rate": 9.468971820624636e-05, + "loss": 3.1489, + "step": 84700 + }, + { + "epoch": 0.7484236740316855, + "grad_norm": 0.7184849381446838, + "learning_rate": 9.468348436329341e-05, + "loss": 3.5103, + "step": 84750 + }, + { + "epoch": 0.7488652219219697, + "grad_norm": 5.096653938293457, + "learning_rate": 9.46772470689202e-05, + "loss": 3.6296, + "step": 84800 + }, + { + "epoch": 0.7493067698122539, + "grad_norm": 1.7660068273544312, + "learning_rate": 9.467100632360851e-05, + "loss": 3.2378, + "step": 84850 + }, + { + "epoch": 0.749748317702538, + "grad_norm": 3.2901999950408936, + "learning_rate": 9.466476212784038e-05, + "loss": 2.8817, + "step": 84900 + }, + { + "epoch": 0.7501898655928222, + "grad_norm": 0.6989944577217102, + "learning_rate": 9.465851448209813e-05, + "loss": 3.2027, + "step": 84950 + }, + { + "epoch": 0.7506314134831064, + "grad_norm": 4.7878875732421875, + "learning_rate": 9.465226338686431e-05, + "loss": 3.0253, + "step": 85000 + }, + { + "epoch": 0.7510729613733905, + "grad_norm": 2.0710065364837646, + "learning_rate": 9.46460088426218e-05, + "loss": 3.5431, + "step": 85050 + }, + { + "epoch": 0.7515145092636747, + "grad_norm": 1.119075059890747, + "learning_rate": 9.463975084985369e-05, + "loss": 3.4444, + "step": 85100 + }, + { + "epoch": 0.751956057153959, + "grad_norm": 2.28783917427063, + "learning_rate": 9.463348940904335e-05, + "loss": 3.6863, + "step": 85150 + }, + { + "epoch": 0.7523976050442431, + "grad_norm": 2.2328102588653564, + "learning_rate": 9.462722452067445e-05, + "loss": 3.2708, + "step": 85200 + }, + { + "epoch": 0.7528391529345273, + "grad_norm": 1.885800838470459, + "learning_rate": 9.462095618523089e-05, + "loss": 3.7459, + "step": 85250 + }, + { + "epoch": 0.7532807008248115, + "grad_norm": 1.719815731048584, + "learning_rate": 9.461468440319681e-05, + "loss": 3.7436, + "step": 85300 + }, + { + "epoch": 0.7537222487150956, + "grad_norm": 1.0964494943618774, + "learning_rate": 9.460840917505668e-05, + "loss": 2.7071, + "step": 85350 + }, + { + "epoch": 0.7541637966053798, + "grad_norm": 2.1939446926116943, + "learning_rate": 9.460213050129525e-05, + "loss": 3.4522, + "step": 85400 + }, + { + "epoch": 0.754605344495664, + "grad_norm": 3.344871759414673, + "learning_rate": 9.459584838239743e-05, + "loss": 3.3535, + "step": 85450 + }, + { + "epoch": 0.7550468923859481, + "grad_norm": 3.5036964416503906, + "learning_rate": 9.458956281884848e-05, + "loss": 3.1035, + "step": 85500 + }, + { + "epoch": 0.7554884402762324, + "grad_norm": 1.8744654655456543, + "learning_rate": 9.458327381113392e-05, + "loss": 3.4614, + "step": 85550 + }, + { + "epoch": 0.7559299881665166, + "grad_norm": 3.193300485610962, + "learning_rate": 9.457698135973951e-05, + "loss": 3.2385, + "step": 85600 + }, + { + "epoch": 0.7563715360568007, + "grad_norm": 2.4154229164123535, + "learning_rate": 9.457068546515129e-05, + "loss": 3.1358, + "step": 85650 + }, + { + "epoch": 0.7568130839470849, + "grad_norm": 1.8072885274887085, + "learning_rate": 9.456438612785556e-05, + "loss": 3.0887, + "step": 85700 + }, + { + "epoch": 0.7572546318373691, + "grad_norm": 4.455030918121338, + "learning_rate": 9.455808334833892e-05, + "loss": 2.7886, + "step": 85750 + }, + { + "epoch": 0.7576961797276532, + "grad_norm": 3.4446539878845215, + "learning_rate": 9.455177712708815e-05, + "loss": 3.045, + "step": 85800 + }, + { + "epoch": 0.7581377276179374, + "grad_norm": 1.5495901107788086, + "learning_rate": 9.45454674645904e-05, + "loss": 3.3734, + "step": 85850 + }, + { + "epoch": 0.7585792755082216, + "grad_norm": 1.636919379234314, + "learning_rate": 9.453915436133303e-05, + "loss": 3.1687, + "step": 85900 + }, + { + "epoch": 0.7590208233985059, + "grad_norm": 1.9204362630844116, + "learning_rate": 9.453283781780365e-05, + "loss": 3.5536, + "step": 85950 + }, + { + "epoch": 0.75946237128879, + "grad_norm": 2.4546802043914795, + "learning_rate": 9.45265178344902e-05, + "loss": 3.0949, + "step": 86000 + }, + { + "epoch": 0.7599039191790742, + "grad_norm": 5.441671848297119, + "learning_rate": 9.45201944118808e-05, + "loss": 3.2127, + "step": 86050 + }, + { + "epoch": 0.7603454670693583, + "grad_norm": 3.458197593688965, + "learning_rate": 9.45138675504639e-05, + "loss": 3.4892, + "step": 86100 + }, + { + "epoch": 0.7607870149596425, + "grad_norm": 1.5939244031906128, + "learning_rate": 9.450753725072822e-05, + "loss": 3.2989, + "step": 86150 + }, + { + "epoch": 0.7612285628499267, + "grad_norm": 2.7477095127105713, + "learning_rate": 9.450120351316268e-05, + "loss": 3.1591, + "step": 86200 + }, + { + "epoch": 0.7616701107402108, + "grad_norm": 2.2054810523986816, + "learning_rate": 9.449486633825654e-05, + "loss": 3.2117, + "step": 86250 + }, + { + "epoch": 0.762111658630495, + "grad_norm": 2.03580379486084, + "learning_rate": 9.448852572649926e-05, + "loss": 2.9641, + "step": 86300 + }, + { + "epoch": 0.7625532065207793, + "grad_norm": 1.0966379642486572, + "learning_rate": 9.448218167838065e-05, + "loss": 3.3211, + "step": 86350 + }, + { + "epoch": 0.7629947544110635, + "grad_norm": 2.210423707962036, + "learning_rate": 9.44758341943907e-05, + "loss": 3.2672, + "step": 86400 + }, + { + "epoch": 0.7634363023013476, + "grad_norm": 3.2034225463867188, + "learning_rate": 9.446948327501967e-05, + "loss": 3.6083, + "step": 86450 + }, + { + "epoch": 0.7638778501916318, + "grad_norm": 2.616360664367676, + "learning_rate": 9.446312892075818e-05, + "loss": 3.1633, + "step": 86500 + }, + { + "epoch": 0.764319398081916, + "grad_norm": 2.7256109714508057, + "learning_rate": 9.445689832152417e-05, + "loss": 3.1589, + "step": 86550 + }, + { + "epoch": 0.7647609459722001, + "grad_norm": 3.060117483139038, + "learning_rate": 9.445053716762777e-05, + "loss": 3.142, + "step": 86600 + }, + { + "epoch": 0.7652024938624843, + "grad_norm": 1.2337734699249268, + "learning_rate": 9.444417258030429e-05, + "loss": 3.1518, + "step": 86650 + }, + { + "epoch": 0.7656440417527685, + "grad_norm": 2.3434503078460693, + "learning_rate": 9.443780456004537e-05, + "loss": 3.2844, + "step": 86700 + }, + { + "epoch": 0.7660855896430527, + "grad_norm": 2.32765793800354, + "learning_rate": 9.443143310734285e-05, + "loss": 3.131, + "step": 86750 + }, + { + "epoch": 0.7665271375333369, + "grad_norm": 1.1229861974716187, + "learning_rate": 9.442505822268891e-05, + "loss": 3.4084, + "step": 86800 + }, + { + "epoch": 0.7669686854236211, + "grad_norm": 2.0934159755706787, + "learning_rate": 9.441867990657594e-05, + "loss": 3.3533, + "step": 86850 + }, + { + "epoch": 0.7674102333139052, + "grad_norm": 5.029802322387695, + "learning_rate": 9.441229815949658e-05, + "loss": 3.2617, + "step": 86900 + }, + { + "epoch": 0.7678517812041894, + "grad_norm": 3.596031665802002, + "learning_rate": 9.440591298194382e-05, + "loss": 3.3018, + "step": 86950 + }, + { + "epoch": 0.7682933290944736, + "grad_norm": 0.9849772453308105, + "learning_rate": 9.439952437441082e-05, + "loss": 2.6819, + "step": 87000 + }, + { + "epoch": 0.7682933290944736, + "eval_asr_loss": 0.9071832583308085, + "eval_loss": 2.9738311767578125, + "eval_runtime": 20.831, + "eval_samples_per_second": 36.868, + "eval_steps_per_second": 9.217, + "eval_tts_loss": 5.941716546998216, + "step": 87000 + }, + { + "epoch": 0.7687348769847577, + "grad_norm": 1.9874849319458008, + "learning_rate": 9.439313233739105e-05, + "loss": 3.5471, + "step": 87050 + }, + { + "epoch": 0.7691764248750419, + "grad_norm": 2.417670249938965, + "learning_rate": 9.438673687137827e-05, + "loss": 3.2482, + "step": 87100 + }, + { + "epoch": 0.7696179727653262, + "grad_norm": 1.2283480167388916, + "learning_rate": 9.438033797686642e-05, + "loss": 3.6392, + "step": 87150 + }, + { + "epoch": 0.7700595206556103, + "grad_norm": 4.871400833129883, + "learning_rate": 9.437393565434984e-05, + "loss": 3.7572, + "step": 87200 + }, + { + "epoch": 0.7705010685458945, + "grad_norm": 2.2227866649627686, + "learning_rate": 9.436752990432298e-05, + "loss": 3.0988, + "step": 87250 + }, + { + "epoch": 0.7709426164361787, + "grad_norm": 1.317663550376892, + "learning_rate": 9.436112072728067e-05, + "loss": 2.342, + "step": 87300 + }, + { + "epoch": 0.7713841643264628, + "grad_norm": 1.2101949453353882, + "learning_rate": 9.435470812371796e-05, + "loss": 3.0555, + "step": 87350 + }, + { + "epoch": 0.771825712216747, + "grad_norm": 2.3560891151428223, + "learning_rate": 9.434829209413016e-05, + "loss": 3.287, + "step": 87400 + }, + { + "epoch": 0.7722672601070312, + "grad_norm": 2.26143217086792, + "learning_rate": 9.434187263901286e-05, + "loss": 3.1895, + "step": 87450 + }, + { + "epoch": 0.7727088079973153, + "grad_norm": 2.7307794094085693, + "learning_rate": 9.43354497588619e-05, + "loss": 3.0531, + "step": 87500 + }, + { + "epoch": 0.7731503558875996, + "grad_norm": 2.08046293258667, + "learning_rate": 9.43290234541734e-05, + "loss": 3.5845, + "step": 87550 + }, + { + "epoch": 0.7735919037778838, + "grad_norm": 1.427642822265625, + "learning_rate": 9.432259372544375e-05, + "loss": 3.372, + "step": 87600 + }, + { + "epoch": 0.774033451668168, + "grad_norm": 1.8793132305145264, + "learning_rate": 9.431616057316957e-05, + "loss": 3.44, + "step": 87650 + }, + { + "epoch": 0.7744749995584521, + "grad_norm": 5.893092632293701, + "learning_rate": 9.430972399784779e-05, + "loss": 3.3203, + "step": 87700 + }, + { + "epoch": 0.7749165474487363, + "grad_norm": 1.8131071329116821, + "learning_rate": 9.430328399997555e-05, + "loss": 3.1405, + "step": 87750 + }, + { + "epoch": 0.7753580953390204, + "grad_norm": 1.3549493551254272, + "learning_rate": 9.429696948198171e-05, + "loss": 3.4958, + "step": 87800 + }, + { + "epoch": 0.7757996432293046, + "grad_norm": 2.8641469478607178, + "learning_rate": 9.429052270892739e-05, + "loss": 3.0084, + "step": 87850 + }, + { + "epoch": 0.7762411911195888, + "grad_norm": 1.4089255332946777, + "learning_rate": 9.428407251480576e-05, + "loss": 3.5739, + "step": 87900 + }, + { + "epoch": 0.7766827390098731, + "grad_norm": 2.792038679122925, + "learning_rate": 9.427761890011504e-05, + "loss": 3.0026, + "step": 87950 + }, + { + "epoch": 0.7771242869001572, + "grad_norm": 2.8735337257385254, + "learning_rate": 9.427116186535371e-05, + "loss": 2.9788, + "step": 88000 + }, + { + "epoch": 0.7775658347904414, + "grad_norm": 2.6922061443328857, + "learning_rate": 9.426470141102058e-05, + "loss": 3.5371, + "step": 88050 + }, + { + "epoch": 0.7780073826807256, + "grad_norm": 2.8780505657196045, + "learning_rate": 9.425823753761458e-05, + "loss": 3.4691, + "step": 88100 + }, + { + "epoch": 0.7784489305710097, + "grad_norm": 3.15449595451355, + "learning_rate": 9.425177024563504e-05, + "loss": 3.0253, + "step": 88150 + }, + { + "epoch": 0.7788904784612939, + "grad_norm": 2.4547994136810303, + "learning_rate": 9.42452995355815e-05, + "loss": 3.22, + "step": 88200 + }, + { + "epoch": 0.779332026351578, + "grad_norm": 1.3049603700637817, + "learning_rate": 9.423882540795375e-05, + "loss": 3.3362, + "step": 88250 + }, + { + "epoch": 0.7797735742418622, + "grad_norm": 2.944639205932617, + "learning_rate": 9.423234786325188e-05, + "loss": 2.6225, + "step": 88300 + }, + { + "epoch": 0.7802151221321465, + "grad_norm": 4.7420783042907715, + "learning_rate": 9.42258669019762e-05, + "loss": 2.852, + "step": 88350 + }, + { + "epoch": 0.7806566700224307, + "grad_norm": 2.200601100921631, + "learning_rate": 9.421938252462734e-05, + "loss": 3.4574, + "step": 88400 + }, + { + "epoch": 0.7810982179127148, + "grad_norm": 0.8284490704536438, + "learning_rate": 9.421289473170615e-05, + "loss": 3.3901, + "step": 88450 + }, + { + "epoch": 0.781539765802999, + "grad_norm": 2.629199266433716, + "learning_rate": 9.420640352371375e-05, + "loss": 3.0143, + "step": 88500 + }, + { + "epoch": 0.7819813136932832, + "grad_norm": 2.411001205444336, + "learning_rate": 9.419990890115155e-05, + "loss": 3.4572, + "step": 88550 + }, + { + "epoch": 0.7824228615835673, + "grad_norm": 3.345630168914795, + "learning_rate": 9.419341086452118e-05, + "loss": 2.9808, + "step": 88600 + }, + { + "epoch": 0.7828644094738515, + "grad_norm": 1.7286512851715088, + "learning_rate": 9.418690941432458e-05, + "loss": 3.3402, + "step": 88650 + }, + { + "epoch": 0.7833059573641357, + "grad_norm": 2.867637872695923, + "learning_rate": 9.418040455106391e-05, + "loss": 3.3708, + "step": 88700 + }, + { + "epoch": 0.7837475052544199, + "grad_norm": 1.4467108249664307, + "learning_rate": 9.417389627524163e-05, + "loss": 3.4814, + "step": 88750 + }, + { + "epoch": 0.7841890531447041, + "grad_norm": 1.6191805601119995, + "learning_rate": 9.416738458736045e-05, + "loss": 3.4584, + "step": 88800 + }, + { + "epoch": 0.7846306010349883, + "grad_norm": 2.4342501163482666, + "learning_rate": 9.416086948792333e-05, + "loss": 3.4172, + "step": 88850 + }, + { + "epoch": 0.7850721489252724, + "grad_norm": 1.3565874099731445, + "learning_rate": 9.415435097743352e-05, + "loss": 3.6409, + "step": 88900 + }, + { + "epoch": 0.7855136968155566, + "grad_norm": 0.9413079619407654, + "learning_rate": 9.414782905639452e-05, + "loss": 3.1165, + "step": 88950 + }, + { + "epoch": 0.7859552447058408, + "grad_norm": 2.325162887573242, + "learning_rate": 9.414130372531009e-05, + "loss": 3.3463, + "step": 89000 + }, + { + "epoch": 0.7863967925961249, + "grad_norm": 2.1379945278167725, + "learning_rate": 9.413477498468423e-05, + "loss": 3.416, + "step": 89050 + }, + { + "epoch": 0.7868383404864091, + "grad_norm": 1.78729248046875, + "learning_rate": 9.412824283502129e-05, + "loss": 3.6121, + "step": 89100 + }, + { + "epoch": 0.7872798883766934, + "grad_norm": 1.3895008563995361, + "learning_rate": 9.412170727682576e-05, + "loss": 3.1819, + "step": 89150 + }, + { + "epoch": 0.7877214362669775, + "grad_norm": 3.693272352218628, + "learning_rate": 9.41151683106025e-05, + "loss": 3.3451, + "step": 89200 + }, + { + "epoch": 0.7881629841572617, + "grad_norm": 1.3732582330703735, + "learning_rate": 9.410862593685657e-05, + "loss": 3.0712, + "step": 89250 + }, + { + "epoch": 0.7886045320475459, + "grad_norm": 1.9743282794952393, + "learning_rate": 9.410208015609333e-05, + "loss": 3.7555, + "step": 89300 + }, + { + "epoch": 0.78904607993783, + "grad_norm": 2.194549083709717, + "learning_rate": 9.409553096881835e-05, + "loss": 2.9958, + "step": 89350 + }, + { + "epoch": 0.7894876278281142, + "grad_norm": 1.8362387418746948, + "learning_rate": 9.408897837553754e-05, + "loss": 3.4235, + "step": 89400 + }, + { + "epoch": 0.7899291757183984, + "grad_norm": 1.8394923210144043, + "learning_rate": 9.4082422376757e-05, + "loss": 3.3358, + "step": 89450 + }, + { + "epoch": 0.7903707236086825, + "grad_norm": 2.9508590698242188, + "learning_rate": 9.407586297298314e-05, + "loss": 3.5187, + "step": 89500 + }, + { + "epoch": 0.7908122714989668, + "grad_norm": 1.0827395915985107, + "learning_rate": 9.406930016472262e-05, + "loss": 2.985, + "step": 89550 + }, + { + "epoch": 0.791253819389251, + "grad_norm": 6.141767978668213, + "learning_rate": 9.406273395248236e-05, + "loss": 3.3644, + "step": 89600 + }, + { + "epoch": 0.7916953672795352, + "grad_norm": 2.3642897605895996, + "learning_rate": 9.405616433676954e-05, + "loss": 2.9457, + "step": 89650 + }, + { + "epoch": 0.7921369151698193, + "grad_norm": 1.307497501373291, + "learning_rate": 9.404959131809163e-05, + "loss": 3.5927, + "step": 89700 + }, + { + "epoch": 0.7925784630601035, + "grad_norm": 0.8792401552200317, + "learning_rate": 9.404301489695629e-05, + "loss": 3.4736, + "step": 89750 + }, + { + "epoch": 0.7930200109503877, + "grad_norm": 1.8458104133605957, + "learning_rate": 9.403643507387155e-05, + "loss": 3.1718, + "step": 89800 + }, + { + "epoch": 0.7934615588406718, + "grad_norm": 0.40211722254753113, + "learning_rate": 9.402985184934561e-05, + "loss": 3.1028, + "step": 89850 + }, + { + "epoch": 0.793903106730956, + "grad_norm": 1.4072984457015991, + "learning_rate": 9.402326522388695e-05, + "loss": 3.2689, + "step": 89900 + }, + { + "epoch": 0.7943446546212403, + "grad_norm": 3.634361505508423, + "learning_rate": 9.401667519800439e-05, + "loss": 3.2542, + "step": 89950 + }, + { + "epoch": 0.7947862025115244, + "grad_norm": 2.3660759925842285, + "learning_rate": 9.401008177220691e-05, + "loss": 3.3141, + "step": 90000 + }, + { + "epoch": 0.7947862025115244, + "eval_asr_loss": 0.913769633646809, + "eval_loss": 2.972637891769409, + "eval_runtime": 20.6823, + "eval_samples_per_second": 37.133, + "eval_steps_per_second": 9.283, + "eval_tts_loss": 5.941250072060806, + "step": 90000 + }, + { + "epoch": 0.7952277504018086, + "grad_norm": 1.8747526407241821, + "learning_rate": 9.400348494700382e-05, + "loss": 3.2776, + "step": 90050 + }, + { + "epoch": 0.7956692982920928, + "grad_norm": 2.4819157123565674, + "learning_rate": 9.399688472290464e-05, + "loss": 3.3672, + "step": 90100 + }, + { + "epoch": 0.7961108461823769, + "grad_norm": 3.1009058952331543, + "learning_rate": 9.399028110041921e-05, + "loss": 3.7072, + "step": 90150 + }, + { + "epoch": 0.7965523940726611, + "grad_norm": 2.1817233562469482, + "learning_rate": 9.39836740800576e-05, + "loss": 3.1293, + "step": 90200 + }, + { + "epoch": 0.7969939419629453, + "grad_norm": 2.2436323165893555, + "learning_rate": 9.397706366233011e-05, + "loss": 3.7612, + "step": 90250 + }, + { + "epoch": 0.7974354898532294, + "grad_norm": 2.8587071895599365, + "learning_rate": 9.397058215732492e-05, + "loss": 3.574, + "step": 90300 + }, + { + "epoch": 0.7978770377435137, + "grad_norm": 1.305193543434143, + "learning_rate": 9.396396501431968e-05, + "loss": 3.2321, + "step": 90350 + }, + { + "epoch": 0.7983185856337979, + "grad_norm": 4.059189319610596, + "learning_rate": 9.395734447547093e-05, + "loss": 3.2663, + "step": 90400 + }, + { + "epoch": 0.798760133524082, + "grad_norm": 1.4845237731933594, + "learning_rate": 9.395072054129008e-05, + "loss": 3.2744, + "step": 90450 + }, + { + "epoch": 0.7992016814143662, + "grad_norm": 3.5800607204437256, + "learning_rate": 9.394409321228877e-05, + "loss": 3.3557, + "step": 90500 + }, + { + "epoch": 0.7996432293046504, + "grad_norm": 1.589889645576477, + "learning_rate": 9.393746248897888e-05, + "loss": 3.1891, + "step": 90550 + }, + { + "epoch": 0.8000847771949345, + "grad_norm": 2.2087574005126953, + "learning_rate": 9.39308283718726e-05, + "loss": 3.4591, + "step": 90600 + }, + { + "epoch": 0.8005263250852187, + "grad_norm": 2.9878342151641846, + "learning_rate": 9.392419086148234e-05, + "loss": 3.1537, + "step": 90650 + }, + { + "epoch": 0.8009678729755029, + "grad_norm": 4.756255626678467, + "learning_rate": 9.391754995832083e-05, + "loss": 3.1161, + "step": 90700 + }, + { + "epoch": 0.8014094208657871, + "grad_norm": 3.2718210220336914, + "learning_rate": 9.391090566290097e-05, + "loss": 3.6481, + "step": 90750 + }, + { + "epoch": 0.8018509687560713, + "grad_norm": 1.9818165302276611, + "learning_rate": 9.3904257975736e-05, + "loss": 3.0561, + "step": 90800 + }, + { + "epoch": 0.8022925166463555, + "grad_norm": 0.7319821715354919, + "learning_rate": 9.38976068973394e-05, + "loss": 3.6211, + "step": 90850 + }, + { + "epoch": 0.8027340645366396, + "grad_norm": 1.0141242742538452, + "learning_rate": 9.389095242822492e-05, + "loss": 3.3727, + "step": 90900 + }, + { + "epoch": 0.8031756124269238, + "grad_norm": 2.862752676010132, + "learning_rate": 9.388429456890654e-05, + "loss": 3.2647, + "step": 90950 + }, + { + "epoch": 0.803617160317208, + "grad_norm": 2.0796523094177246, + "learning_rate": 9.387763331989854e-05, + "loss": 2.9279, + "step": 91000 + }, + { + "epoch": 0.8040587082074921, + "grad_norm": 3.659575939178467, + "learning_rate": 9.387096868171543e-05, + "loss": 3.3732, + "step": 91050 + }, + { + "epoch": 0.8045002560977763, + "grad_norm": 1.8139934539794922, + "learning_rate": 9.3864300654872e-05, + "loss": 3.4054, + "step": 91100 + }, + { + "epoch": 0.8049418039880606, + "grad_norm": 1.9845603704452515, + "learning_rate": 9.385762923988332e-05, + "loss": 3.2696, + "step": 91150 + }, + { + "epoch": 0.8053833518783448, + "grad_norm": 1.5647777318954468, + "learning_rate": 9.385095443726466e-05, + "loss": 2.9517, + "step": 91200 + }, + { + "epoch": 0.8058248997686289, + "grad_norm": 2.7620863914489746, + "learning_rate": 9.384427624753164e-05, + "loss": 3.3531, + "step": 91250 + }, + { + "epoch": 0.8062664476589131, + "grad_norm": 4.5521039962768555, + "learning_rate": 9.383759467120006e-05, + "loss": 2.9471, + "step": 91300 + }, + { + "epoch": 0.8067079955491973, + "grad_norm": 6.784761428833008, + "learning_rate": 9.3830909708786e-05, + "loss": 3.0502, + "step": 91350 + }, + { + "epoch": 0.8071495434394814, + "grad_norm": 2.283494472503662, + "learning_rate": 9.382422136080587e-05, + "loss": 3.2579, + "step": 91400 + }, + { + "epoch": 0.8075910913297656, + "grad_norm": 2.49485182762146, + "learning_rate": 9.381752962777626e-05, + "loss": 3.2323, + "step": 91450 + }, + { + "epoch": 0.8080326392200498, + "grad_norm": 1.1890743970870972, + "learning_rate": 9.381083451021402e-05, + "loss": 3.3812, + "step": 91500 + }, + { + "epoch": 0.808474187110334, + "grad_norm": 1.023622989654541, + "learning_rate": 9.380413600863633e-05, + "loss": 3.1125, + "step": 91550 + }, + { + "epoch": 0.8089157350006182, + "grad_norm": 3.7999751567840576, + "learning_rate": 9.37974341235606e-05, + "loss": 3.0946, + "step": 91600 + }, + { + "epoch": 0.8093572828909024, + "grad_norm": 2.7334210872650146, + "learning_rate": 9.379072885550446e-05, + "loss": 3.2808, + "step": 91650 + }, + { + "epoch": 0.8097988307811865, + "grad_norm": 1.0871787071228027, + "learning_rate": 9.378402020498585e-05, + "loss": 2.9201, + "step": 91700 + }, + { + "epoch": 0.8102403786714707, + "grad_norm": 1.837319254875183, + "learning_rate": 9.377730817252296e-05, + "loss": 3.3392, + "step": 91750 + }, + { + "epoch": 0.8106819265617549, + "grad_norm": 1.857426643371582, + "learning_rate": 9.377059275863423e-05, + "loss": 3.4567, + "step": 91800 + }, + { + "epoch": 0.811123474452039, + "grad_norm": 2.4178388118743896, + "learning_rate": 9.37638739638384e-05, + "loss": 3.1172, + "step": 91850 + }, + { + "epoch": 0.8115650223423232, + "grad_norm": 1.7428170442581177, + "learning_rate": 9.375715178865438e-05, + "loss": 3.2541, + "step": 91900 + }, + { + "epoch": 0.8120065702326075, + "grad_norm": 2.188565731048584, + "learning_rate": 9.375042623360143e-05, + "loss": 3.0557, + "step": 91950 + }, + { + "epoch": 0.8124481181228916, + "grad_norm": 2.7001125812530518, + "learning_rate": 9.374369729919907e-05, + "loss": 3.4467, + "step": 92000 + }, + { + "epoch": 0.8128896660131758, + "grad_norm": 0.5670623183250427, + "learning_rate": 9.373696498596703e-05, + "loss": 3.4051, + "step": 92050 + }, + { + "epoch": 0.81333121390346, + "grad_norm": 0.8684957027435303, + "learning_rate": 9.37302292944253e-05, + "loss": 3.4903, + "step": 92100 + }, + { + "epoch": 0.8137727617937441, + "grad_norm": 1.7560847997665405, + "learning_rate": 9.372349022509419e-05, + "loss": 2.8986, + "step": 92150 + }, + { + "epoch": 0.8142143096840283, + "grad_norm": 1.301934838294983, + "learning_rate": 9.371674777849423e-05, + "loss": 3.2969, + "step": 92200 + }, + { + "epoch": 0.8146558575743125, + "grad_norm": 5.415755748748779, + "learning_rate": 9.37100019551462e-05, + "loss": 3.5175, + "step": 92250 + }, + { + "epoch": 0.8150974054645966, + "grad_norm": 1.403662085533142, + "learning_rate": 9.370325275557119e-05, + "loss": 3.3363, + "step": 92300 + }, + { + "epoch": 0.8155389533548809, + "grad_norm": 0.34755173325538635, + "learning_rate": 9.369650018029047e-05, + "loss": 2.8427, + "step": 92350 + }, + { + "epoch": 0.8159805012451651, + "grad_norm": 4.63238000869751, + "learning_rate": 9.368974422982567e-05, + "loss": 3.401, + "step": 92400 + }, + { + "epoch": 0.8164220491354492, + "grad_norm": 5.1991286277771, + "learning_rate": 9.36829849046986e-05, + "loss": 3.0604, + "step": 92450 + }, + { + "epoch": 0.8168635970257334, + "grad_norm": 4.037189960479736, + "learning_rate": 9.367622220543136e-05, + "loss": 3.4574, + "step": 92500 + }, + { + "epoch": 0.8173051449160176, + "grad_norm": 3.1938400268554688, + "learning_rate": 9.366945613254631e-05, + "loss": 3.4916, + "step": 92550 + }, + { + "epoch": 0.8177466928063017, + "grad_norm": 1.4730736017227173, + "learning_rate": 9.36626866865661e-05, + "loss": 3.5001, + "step": 92600 + }, + { + "epoch": 0.8181882406965859, + "grad_norm": 2.499075174331665, + "learning_rate": 9.36559138680136e-05, + "loss": 3.3653, + "step": 92650 + }, + { + "epoch": 0.8186297885868701, + "grad_norm": 3.067751884460449, + "learning_rate": 9.364913767741192e-05, + "loss": 2.9999, + "step": 92700 + }, + { + "epoch": 0.8190713364771544, + "grad_norm": 3.1233701705932617, + "learning_rate": 9.36423581152845e-05, + "loss": 3.377, + "step": 92750 + }, + { + "epoch": 0.8195128843674385, + "grad_norm": 1.1378720998764038, + "learning_rate": 9.363557518215498e-05, + "loss": 2.8108, + "step": 92800 + }, + { + "epoch": 0.8199544322577227, + "grad_norm": 1.5843881368637085, + "learning_rate": 9.362878887854732e-05, + "loss": 3.4633, + "step": 92850 + }, + { + "epoch": 0.8203959801480069, + "grad_norm": 3.084805965423584, + "learning_rate": 9.362199920498566e-05, + "loss": 3.3343, + "step": 92900 + }, + { + "epoch": 0.820837528038291, + "grad_norm": 1.3744654655456543, + "learning_rate": 9.361520616199446e-05, + "loss": 2.9136, + "step": 92950 + }, + { + "epoch": 0.8212790759285752, + "grad_norm": 3.1736974716186523, + "learning_rate": 9.360840975009845e-05, + "loss": 3.1212, + "step": 93000 + }, + { + "epoch": 0.8212790759285752, + "eval_asr_loss": 0.9125425967281584, + "eval_loss": 2.9801833629608154, + "eval_runtime": 20.5623, + "eval_samples_per_second": 37.35, + "eval_steps_per_second": 9.337, + "eval_tts_loss": 5.957149096143686, + "step": 93000 + }, + { + "epoch": 0.8217206238188594, + "grad_norm": 2.6484904289245605, + "learning_rate": 9.360160996982256e-05, + "loss": 3.3256, + "step": 93050 + }, + { + "epoch": 0.8221621717091435, + "grad_norm": 1.6820108890533447, + "learning_rate": 9.359480682169204e-05, + "loss": 3.3486, + "step": 93100 + }, + { + "epoch": 0.8226037195994278, + "grad_norm": 1.407516360282898, + "learning_rate": 9.358800030623236e-05, + "loss": 3.4169, + "step": 93150 + }, + { + "epoch": 0.823045267489712, + "grad_norm": 1.139156699180603, + "learning_rate": 9.358119042396929e-05, + "loss": 2.8844, + "step": 93200 + }, + { + "epoch": 0.8234868153799961, + "grad_norm": 1.5392656326293945, + "learning_rate": 9.35743771754288e-05, + "loss": 3.5062, + "step": 93250 + }, + { + "epoch": 0.8239283632702803, + "grad_norm": 1.2801883220672607, + "learning_rate": 9.356756056113717e-05, + "loss": 2.9979, + "step": 93300 + }, + { + "epoch": 0.8243699111605645, + "grad_norm": 2.373271942138672, + "learning_rate": 9.356074058162094e-05, + "loss": 2.8897, + "step": 93350 + }, + { + "epoch": 0.8248114590508486, + "grad_norm": 0.7981138229370117, + "learning_rate": 9.355391723740687e-05, + "loss": 3.7609, + "step": 93400 + }, + { + "epoch": 0.8252530069411328, + "grad_norm": 1.8183437585830688, + "learning_rate": 9.354709052902204e-05, + "loss": 3.2145, + "step": 93450 + }, + { + "epoch": 0.825694554831417, + "grad_norm": 1.9743393659591675, + "learning_rate": 9.354026045699372e-05, + "loss": 3.5426, + "step": 93500 + }, + { + "epoch": 0.8261361027217012, + "grad_norm": 2.351931571960449, + "learning_rate": 9.35334270218495e-05, + "loss": 3.347, + "step": 93550 + }, + { + "epoch": 0.8265776506119854, + "grad_norm": 1.7274210453033447, + "learning_rate": 9.352659022411718e-05, + "loss": 2.9723, + "step": 93600 + }, + { + "epoch": 0.8270191985022696, + "grad_norm": 3.6531035900115967, + "learning_rate": 9.351975006432487e-05, + "loss": 3.311, + "step": 93650 + }, + { + "epoch": 0.8274607463925537, + "grad_norm": 5.743092060089111, + "learning_rate": 9.351290654300091e-05, + "loss": 3.1698, + "step": 93700 + }, + { + "epoch": 0.8279022942828379, + "grad_norm": 1.3943450450897217, + "learning_rate": 9.35060596606739e-05, + "loss": 3.0665, + "step": 93750 + }, + { + "epoch": 0.8283438421731221, + "grad_norm": 3.4705088138580322, + "learning_rate": 9.349920941787269e-05, + "loss": 3.5611, + "step": 93800 + }, + { + "epoch": 0.8287853900634062, + "grad_norm": 1.309612512588501, + "learning_rate": 9.34923558151264e-05, + "loss": 3.4221, + "step": 93850 + }, + { + "epoch": 0.8292269379536905, + "grad_norm": 1.8064721822738647, + "learning_rate": 9.348549885296446e-05, + "loss": 3.564, + "step": 93900 + }, + { + "epoch": 0.8296684858439747, + "grad_norm": 1.7501392364501953, + "learning_rate": 9.347863853191646e-05, + "loss": 3.0225, + "step": 93950 + }, + { + "epoch": 0.8301100337342588, + "grad_norm": 3.4473793506622314, + "learning_rate": 9.347177485251233e-05, + "loss": 3.0757, + "step": 94000 + }, + { + "epoch": 0.830551581624543, + "grad_norm": 1.1141574382781982, + "learning_rate": 9.346490781528221e-05, + "loss": 3.0982, + "step": 94050 + }, + { + "epoch": 0.8309931295148272, + "grad_norm": 3.239313840866089, + "learning_rate": 9.345803742075656e-05, + "loss": 3.463, + "step": 94100 + }, + { + "epoch": 0.8314346774051113, + "grad_norm": 2.90984845161438, + "learning_rate": 9.345116366946601e-05, + "loss": 3.131, + "step": 94150 + }, + { + "epoch": 0.8318762252953955, + "grad_norm": 3.623865842819214, + "learning_rate": 9.344428656194153e-05, + "loss": 3.4718, + "step": 94200 + }, + { + "epoch": 0.8323177731856797, + "grad_norm": 0.8482750058174133, + "learning_rate": 9.343740609871431e-05, + "loss": 3.3195, + "step": 94250 + }, + { + "epoch": 0.832759321075964, + "grad_norm": 1.9041322469711304, + "learning_rate": 9.343052228031582e-05, + "loss": 3.3709, + "step": 94300 + }, + { + "epoch": 0.8332008689662481, + "grad_norm": 3.6412575244903564, + "learning_rate": 9.342363510727773e-05, + "loss": 3.3433, + "step": 94350 + }, + { + "epoch": 0.8336424168565323, + "grad_norm": 1.070917010307312, + "learning_rate": 9.341674458013208e-05, + "loss": 3.5517, + "step": 94400 + }, + { + "epoch": 0.8340839647468165, + "grad_norm": 4.740517616271973, + "learning_rate": 9.340985069941105e-05, + "loss": 3.3758, + "step": 94450 + }, + { + "epoch": 0.8345255126371006, + "grad_norm": 0.6648314595222473, + "learning_rate": 9.340295346564719e-05, + "loss": 3.5991, + "step": 94500 + }, + { + "epoch": 0.8349670605273848, + "grad_norm": 3.5965664386749268, + "learning_rate": 9.339605287937319e-05, + "loss": 3.4054, + "step": 94550 + }, + { + "epoch": 0.835408608417669, + "grad_norm": 2.838550329208374, + "learning_rate": 9.33891489411221e-05, + "loss": 3.0964, + "step": 94600 + }, + { + "epoch": 0.8358501563079531, + "grad_norm": 1.3823540210723877, + "learning_rate": 9.338224165142719e-05, + "loss": 3.5835, + "step": 94650 + }, + { + "epoch": 0.8362917041982374, + "grad_norm": 2.0661122798919678, + "learning_rate": 9.337533101082199e-05, + "loss": 3.2928, + "step": 94700 + }, + { + "epoch": 0.8367332520885216, + "grad_norm": 1.8618370294570923, + "learning_rate": 9.336841701984027e-05, + "loss": 3.0778, + "step": 94750 + }, + { + "epoch": 0.8371747999788057, + "grad_norm": 4.694985389709473, + "learning_rate": 9.336149967901609e-05, + "loss": 3.0738, + "step": 94800 + }, + { + "epoch": 0.8376163478690899, + "grad_norm": 2.5283780097961426, + "learning_rate": 9.335457898888376e-05, + "loss": 3.4458, + "step": 94850 + }, + { + "epoch": 0.8380578957593741, + "grad_norm": 1.0920939445495605, + "learning_rate": 9.334765494997784e-05, + "loss": 3.2661, + "step": 94900 + }, + { + "epoch": 0.8384994436496582, + "grad_norm": 2.00156307220459, + "learning_rate": 9.334072756283314e-05, + "loss": 3.2389, + "step": 94950 + }, + { + "epoch": 0.8389409915399424, + "grad_norm": 2.5333197116851807, + "learning_rate": 9.333379682798475e-05, + "loss": 3.5175, + "step": 95000 + }, + { + "epoch": 0.8393825394302266, + "grad_norm": 2.970446825027466, + "learning_rate": 9.332686274596803e-05, + "loss": 3.0734, + "step": 95050 + }, + { + "epoch": 0.8398240873205108, + "grad_norm": 1.0278738737106323, + "learning_rate": 9.331992531731857e-05, + "loss": 3.5306, + "step": 95100 + }, + { + "epoch": 0.840265635210795, + "grad_norm": 2.5087051391601562, + "learning_rate": 9.331298454257221e-05, + "loss": 3.3883, + "step": 95150 + }, + { + "epoch": 0.8407071831010792, + "grad_norm": 1.317252278327942, + "learning_rate": 9.330604042226507e-05, + "loss": 3.2772, + "step": 95200 + }, + { + "epoch": 0.8411487309913633, + "grad_norm": 2.6759376525878906, + "learning_rate": 9.329909295693355e-05, + "loss": 2.9157, + "step": 95250 + }, + { + "epoch": 0.8415902788816475, + "grad_norm": 1.9456520080566406, + "learning_rate": 9.329214214711426e-05, + "loss": 3.2917, + "step": 95300 + }, + { + "epoch": 0.8420318267719317, + "grad_norm": 2.113647937774658, + "learning_rate": 9.328532710918674e-05, + "loss": 3.3203, + "step": 95350 + }, + { + "epoch": 0.8424733746622158, + "grad_norm": 1.5438154935836792, + "learning_rate": 9.327836967886584e-05, + "loss": 3.6364, + "step": 95400 + }, + { + "epoch": 0.8429149225525, + "grad_norm": 1.4090991020202637, + "learning_rate": 9.32714089056579e-05, + "loss": 3.2839, + "step": 95450 + }, + { + "epoch": 0.8433564704427843, + "grad_norm": 3.7097344398498535, + "learning_rate": 9.326444479010055e-05, + "loss": 3.1992, + "step": 95500 + }, + { + "epoch": 0.8437980183330684, + "grad_norm": 1.130540370941162, + "learning_rate": 9.32574773327317e-05, + "loss": 3.123, + "step": 95550 + }, + { + "epoch": 0.8442395662233526, + "grad_norm": 3.1675918102264404, + "learning_rate": 9.325050653408954e-05, + "loss": 3.4446, + "step": 95600 + }, + { + "epoch": 0.8446811141136368, + "grad_norm": 7.838589191436768, + "learning_rate": 9.32435323947125e-05, + "loss": 3.1731, + "step": 95650 + }, + { + "epoch": 0.8451226620039209, + "grad_norm": 2.1061744689941406, + "learning_rate": 9.32365549151393e-05, + "loss": 3.1445, + "step": 95700 + }, + { + "epoch": 0.8455642098942051, + "grad_norm": 1.585892677307129, + "learning_rate": 9.322957409590884e-05, + "loss": 3.248, + "step": 95750 + }, + { + "epoch": 0.8460057577844893, + "grad_norm": 3.2491941452026367, + "learning_rate": 9.322258993756037e-05, + "loss": 3.1858, + "step": 95800 + }, + { + "epoch": 0.8464473056747734, + "grad_norm": 4.585638999938965, + "learning_rate": 9.321560244063334e-05, + "loss": 3.3065, + "step": 95850 + }, + { + "epoch": 0.8468888535650577, + "grad_norm": 1.343159556388855, + "learning_rate": 9.320861160566748e-05, + "loss": 3.3079, + "step": 95900 + }, + { + "epoch": 0.8473304014553419, + "grad_norm": 4.3721699714660645, + "learning_rate": 9.320161743320277e-05, + "loss": 3.1931, + "step": 95950 + }, + { + "epoch": 0.847771949345626, + "grad_norm": 1.243537425994873, + "learning_rate": 9.319461992377945e-05, + "loss": 3.2813, + "step": 96000 + }, + { + "epoch": 0.847771949345626, + "eval_asr_loss": 0.9113168214024796, + "eval_loss": 2.9808547496795654, + "eval_runtime": 20.2448, + "eval_samples_per_second": 37.936, + "eval_steps_per_second": 9.484, + "eval_tts_loss": 5.98077398558446, + "step": 96000 + }, + { + "epoch": 0.8482134972359102, + "grad_norm": 3.9162681102752686, + "learning_rate": 9.318761907793804e-05, + "loss": 3.2729, + "step": 96050 + }, + { + "epoch": 0.8486550451261944, + "grad_norm": 0.9850695729255676, + "learning_rate": 9.318061489621925e-05, + "loss": 3.4972, + "step": 96100 + }, + { + "epoch": 0.8490965930164786, + "grad_norm": 2.3384222984313965, + "learning_rate": 9.317360737916415e-05, + "loss": 3.5748, + "step": 96150 + }, + { + "epoch": 0.8495381409067627, + "grad_norm": 1.6895709037780762, + "learning_rate": 9.316659652731395e-05, + "loss": 3.5112, + "step": 96200 + }, + { + "epoch": 0.8499796887970469, + "grad_norm": 4.779604434967041, + "learning_rate": 9.315958234121023e-05, + "loss": 3.4245, + "step": 96250 + }, + { + "epoch": 0.8504212366873312, + "grad_norm": 3.7872555255889893, + "learning_rate": 9.315256482139475e-05, + "loss": 3.4721, + "step": 96300 + }, + { + "epoch": 0.8508627845776153, + "grad_norm": 1.2569001913070679, + "learning_rate": 9.314554396840956e-05, + "loss": 3.2695, + "step": 96350 + }, + { + "epoch": 0.8513043324678995, + "grad_norm": 2.095500946044922, + "learning_rate": 9.313851978279697e-05, + "loss": 3.1881, + "step": 96400 + }, + { + "epoch": 0.8517458803581837, + "grad_norm": 0.9828936457633972, + "learning_rate": 9.313149226509952e-05, + "loss": 3.0915, + "step": 96450 + }, + { + "epoch": 0.8521874282484678, + "grad_norm": 2.886770725250244, + "learning_rate": 9.312446141586004e-05, + "loss": 3.4099, + "step": 96500 + }, + { + "epoch": 0.852628976138752, + "grad_norm": 4.082062721252441, + "learning_rate": 9.31174272356216e-05, + "loss": 3.0428, + "step": 96550 + }, + { + "epoch": 0.8530705240290362, + "grad_norm": 1.261383295059204, + "learning_rate": 9.311038972492754e-05, + "loss": 3.2126, + "step": 96600 + }, + { + "epoch": 0.8535120719193203, + "grad_norm": 1.2416714429855347, + "learning_rate": 9.310334888432142e-05, + "loss": 3.2021, + "step": 96650 + }, + { + "epoch": 0.8539536198096046, + "grad_norm": 3.1885976791381836, + "learning_rate": 9.309630471434712e-05, + "loss": 2.7708, + "step": 96700 + }, + { + "epoch": 0.8543951676998888, + "grad_norm": 0.710462749004364, + "learning_rate": 9.308925721554874e-05, + "loss": 3.199, + "step": 96750 + }, + { + "epoch": 0.8548367155901729, + "grad_norm": 5.235406875610352, + "learning_rate": 9.308220638847059e-05, + "loss": 3.2963, + "step": 96800 + }, + { + "epoch": 0.8552782634804571, + "grad_norm": 1.9540880918502808, + "learning_rate": 9.307515223365735e-05, + "loss": 3.3895, + "step": 96850 + }, + { + "epoch": 0.8557198113707413, + "grad_norm": 2.796398162841797, + "learning_rate": 9.306809475165385e-05, + "loss": 3.4795, + "step": 96900 + }, + { + "epoch": 0.8561613592610254, + "grad_norm": 1.8404650688171387, + "learning_rate": 9.306103394300525e-05, + "loss": 3.2471, + "step": 96950 + }, + { + "epoch": 0.8566029071513096, + "grad_norm": 1.5861579179763794, + "learning_rate": 9.305396980825693e-05, + "loss": 2.8261, + "step": 97000 + }, + { + "epoch": 0.8570444550415938, + "grad_norm": 5.508571624755859, + "learning_rate": 9.304690234795453e-05, + "loss": 3.0444, + "step": 97050 + }, + { + "epoch": 0.857486002931878, + "grad_norm": 3.7076914310455322, + "learning_rate": 9.303983156264393e-05, + "loss": 3.3217, + "step": 97100 + }, + { + "epoch": 0.8579275508221622, + "grad_norm": 4.032550811767578, + "learning_rate": 9.303275745287133e-05, + "loss": 3.3205, + "step": 97150 + }, + { + "epoch": 0.8583690987124464, + "grad_norm": 0.970733106136322, + "learning_rate": 9.30256800191831e-05, + "loss": 3.1487, + "step": 97200 + }, + { + "epoch": 0.8588106466027305, + "grad_norm": 2.3852109909057617, + "learning_rate": 9.301859926212595e-05, + "loss": 3.1561, + "step": 97250 + }, + { + "epoch": 0.8592521944930147, + "grad_norm": 1.1150777339935303, + "learning_rate": 9.301151518224679e-05, + "loss": 2.9001, + "step": 97300 + }, + { + "epoch": 0.8596937423832989, + "grad_norm": 3.74100923538208, + "learning_rate": 9.300442778009282e-05, + "loss": 3.2136, + "step": 97350 + }, + { + "epoch": 0.860135290273583, + "grad_norm": 2.762349843978882, + "learning_rate": 9.299733705621145e-05, + "loss": 3.4975, + "step": 97400 + }, + { + "epoch": 0.8605768381638672, + "grad_norm": 0.8698297142982483, + "learning_rate": 9.299024301115043e-05, + "loss": 3.3572, + "step": 97450 + }, + { + "epoch": 0.8610183860541515, + "grad_norm": 3.2828052043914795, + "learning_rate": 9.298314564545765e-05, + "loss": 3.3767, + "step": 97500 + }, + { + "epoch": 0.8614599339444357, + "grad_norm": 2.659813642501831, + "learning_rate": 9.297604495968138e-05, + "loss": 3.5504, + "step": 97550 + }, + { + "epoch": 0.8619014818347198, + "grad_norm": 1.7123510837554932, + "learning_rate": 9.296894095437005e-05, + "loss": 3.4038, + "step": 97600 + }, + { + "epoch": 0.862343029725004, + "grad_norm": 2.9085240364074707, + "learning_rate": 9.296183363007241e-05, + "loss": 2.8643, + "step": 97650 + }, + { + "epoch": 0.8627845776152882, + "grad_norm": 1.4961036443710327, + "learning_rate": 9.295472298733742e-05, + "loss": 3.1688, + "step": 97700 + }, + { + "epoch": 0.8632261255055723, + "grad_norm": 2.028132677078247, + "learning_rate": 9.294760902671432e-05, + "loss": 3.3077, + "step": 97750 + }, + { + "epoch": 0.8636676733958565, + "grad_norm": 1.637831449508667, + "learning_rate": 9.294049174875261e-05, + "loss": 2.9796, + "step": 97800 + }, + { + "epoch": 0.8641092212861406, + "grad_norm": 1.5504677295684814, + "learning_rate": 9.293337115400205e-05, + "loss": 2.8288, + "step": 97850 + }, + { + "epoch": 0.8645507691764249, + "grad_norm": 3.117736577987671, + "learning_rate": 9.292638975372799e-05, + "loss": 2.9301, + "step": 97900 + }, + { + "epoch": 0.8649923170667091, + "grad_norm": 2.2668349742889404, + "learning_rate": 9.291926259335835e-05, + "loss": 3.8076, + "step": 97950 + }, + { + "epoch": 0.8654338649569933, + "grad_norm": 3.2014098167419434, + "learning_rate": 9.29121321178396e-05, + "loss": 3.1708, + "step": 98000 + }, + { + "epoch": 0.8658754128472774, + "grad_norm": 3.2920122146606445, + "learning_rate": 9.290499832772255e-05, + "loss": 3.3037, + "step": 98050 + }, + { + "epoch": 0.8663169607375616, + "grad_norm": 1.8121235370635986, + "learning_rate": 9.28978612235582e-05, + "loss": 3.2196, + "step": 98100 + }, + { + "epoch": 0.8667585086278458, + "grad_norm": 1.6291028261184692, + "learning_rate": 9.289072080589783e-05, + "loss": 3.2285, + "step": 98150 + }, + { + "epoch": 0.8672000565181299, + "grad_norm": 1.6624693870544434, + "learning_rate": 9.288357707529297e-05, + "loss": 3.1948, + "step": 98200 + }, + { + "epoch": 0.8676416044084141, + "grad_norm": 2.4642486572265625, + "learning_rate": 9.287643003229543e-05, + "loss": 3.0989, + "step": 98250 + }, + { + "epoch": 0.8680831522986984, + "grad_norm": 2.3431122303009033, + "learning_rate": 9.286927967745725e-05, + "loss": 3.0297, + "step": 98300 + }, + { + "epoch": 0.8685247001889825, + "grad_norm": 1.7271332740783691, + "learning_rate": 9.286212601133076e-05, + "loss": 3.7604, + "step": 98350 + }, + { + "epoch": 0.8689662480792667, + "grad_norm": 2.332041025161743, + "learning_rate": 9.285496903446846e-05, + "loss": 3.4561, + "step": 98400 + }, + { + "epoch": 0.8694077959695509, + "grad_norm": 1.6665946245193481, + "learning_rate": 9.284780874742321e-05, + "loss": 3.2693, + "step": 98450 + }, + { + "epoch": 0.869849343859835, + "grad_norm": 1.6122417449951172, + "learning_rate": 9.284064515074809e-05, + "loss": 3.2516, + "step": 98500 + }, + { + "epoch": 0.8702908917501192, + "grad_norm": 0.9736064672470093, + "learning_rate": 9.283347824499639e-05, + "loss": 3.3643, + "step": 98550 + }, + { + "epoch": 0.8707324396404034, + "grad_norm": 2.6523666381835938, + "learning_rate": 9.282630803072173e-05, + "loss": 3.4379, + "step": 98600 + }, + { + "epoch": 0.8711739875306875, + "grad_norm": 3.84800386428833, + "learning_rate": 9.281913450847792e-05, + "loss": 3.6332, + "step": 98650 + }, + { + "epoch": 0.8716155354209718, + "grad_norm": 1.0476601123809814, + "learning_rate": 9.281195767881908e-05, + "loss": 3.1238, + "step": 98700 + }, + { + "epoch": 0.872057083311256, + "grad_norm": 2.119279623031616, + "learning_rate": 9.280477754229952e-05, + "loss": 3.4006, + "step": 98750 + }, + { + "epoch": 0.8724986312015401, + "grad_norm": 1.7416213750839233, + "learning_rate": 9.279759409947388e-05, + "loss": 3.5637, + "step": 98800 + }, + { + "epoch": 0.8729401790918243, + "grad_norm": 4.38261079788208, + "learning_rate": 9.279040735089702e-05, + "loss": 3.233, + "step": 98850 + }, + { + "epoch": 0.8733817269821085, + "grad_norm": 4.000781059265137, + "learning_rate": 9.278321729712403e-05, + "loss": 3.5001, + "step": 98900 + }, + { + "epoch": 0.8738232748723926, + "grad_norm": 2.074897050857544, + "learning_rate": 9.27760239387103e-05, + "loss": 3.0254, + "step": 98950 + }, + { + "epoch": 0.8742648227626768, + "grad_norm": 3.6822054386138916, + "learning_rate": 9.276882727621146e-05, + "loss": 3.2075, + "step": 99000 + }, + { + "epoch": 0.8742648227626768, + "eval_asr_loss": 0.9122765446783635, + "eval_loss": 2.9597482681274414, + "eval_runtime": 21.0264, + "eval_samples_per_second": 36.525, + "eval_steps_per_second": 9.131, + "eval_tts_loss": 5.931117558903506, + "step": 99000 + } + ], + "logging_steps": 50, + "max_steps": 566190, + "num_input_tokens_seen": 0, + "num_train_epochs": 5, + "save_steps": 3000, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 9.415715320758272e+17, + "train_batch_size": 4, + "trial_name": null, + "trial_params": null +}