{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.2580442250931436, "eval_steps": 3000, "global_step": 78000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0008064386058289383, "grad_norm": 6.343989849090576, "learning_rate": 4.800000000000001e-06, "loss": 3.2595, "step": 50 }, { "epoch": 0.0016128772116578765, "grad_norm": 3.7637691497802734, "learning_rate": 9.800000000000001e-06, "loss": 3.2397, "step": 100 }, { "epoch": 0.002419315817486815, "grad_norm": 4.01455020904541, "learning_rate": 1.48e-05, "loss": 3.1471, "step": 150 }, { "epoch": 0.003225754423315753, "grad_norm": 5.399896621704102, "learning_rate": 1.9800000000000004e-05, "loss": 3.3162, "step": 200 }, { "epoch": 0.004032193029144691, "grad_norm": 5.194526195526123, "learning_rate": 2.48e-05, "loss": 3.4495, "step": 250 }, { "epoch": 0.00483863163497363, "grad_norm": 4.043808460235596, "learning_rate": 2.98e-05, "loss": 3.4038, "step": 300 }, { "epoch": 0.005645070240802567, "grad_norm": 1.6968616247177124, "learning_rate": 3.46e-05, "loss": 3.5791, "step": 350 }, { "epoch": 0.006451508846631506, "grad_norm": 3.515857219696045, "learning_rate": 3.960000000000001e-05, "loss": 3.4015, "step": 400 }, { "epoch": 0.007257947452460444, "grad_norm": 3.327608108520508, "learning_rate": 4.46e-05, "loss": 3.2794, "step": 450 }, { "epoch": 0.008064386058289382, "grad_norm": 3.6122305393218994, "learning_rate": 4.96e-05, "loss": 3.3541, "step": 500 }, { "epoch": 0.00887082466411832, "grad_norm": 4.593649864196777, "learning_rate": 5.4600000000000006e-05, "loss": 3.6442, "step": 550 }, { "epoch": 0.00967726326994726, "grad_norm": 2.58380126953125, "learning_rate": 5.96e-05, "loss": 3.1673, "step": 600 }, { "epoch": 0.010483701875776198, "grad_norm": 5.36879825592041, "learning_rate": 6.460000000000001e-05, "loss": 3.5224, "step": 650 }, { "epoch": 0.011290140481605135, "grad_norm": 21.212533950805664, "learning_rate": 6.96e-05, "loss": 3.5144, "step": 700 }, { "epoch": 0.012096579087434074, "grad_norm": 1.9257746934890747, "learning_rate": 7.46e-05, "loss": 3.2471, "step": 750 }, { "epoch": 0.012903017693263012, "grad_norm": 6.157283306121826, "learning_rate": 7.960000000000001e-05, "loss": 3.551, "step": 800 }, { "epoch": 0.013709456299091951, "grad_norm": 4.075947284698486, "learning_rate": 8.46e-05, "loss": 3.5548, "step": 850 }, { "epoch": 0.014515894904920888, "grad_norm": 3.9454638957977295, "learning_rate": 8.960000000000001e-05, "loss": 3.5653, "step": 900 }, { "epoch": 0.015322333510749827, "grad_norm": 4.633170127868652, "learning_rate": 9.46e-05, "loss": 3.3589, "step": 950 }, { "epoch": 0.016128772116578764, "grad_norm": 4.273008346557617, "learning_rate": 9.960000000000001e-05, "loss": 3.8061, "step": 1000 }, { "epoch": 0.016935210722407702, "grad_norm": 3.094942808151245, "learning_rate": 9.999999453205175e-05, "loss": 3.6132, "step": 1050 }, { "epoch": 0.01774164932823664, "grad_norm": 3.022191047668457, "learning_rate": 9.999997618496786e-05, "loss": 3.8753, "step": 1100 }, { "epoch": 0.01854808793406558, "grad_norm": 3.715566635131836, "learning_rate": 9.999994491740752e-05, "loss": 3.5987, "step": 1150 }, { "epoch": 0.01935452653989452, "grad_norm": 3.9486517906188965, "learning_rate": 9.999990072937881e-05, "loss": 3.7496, "step": 1200 }, { "epoch": 0.020160965145723457, "grad_norm": 4.180655002593994, "learning_rate": 9.999984362089316e-05, "loss": 4.1646, "step": 1250 }, { "epoch": 0.020967403751552396, "grad_norm": 4.141788959503174, "learning_rate": 9.999977359196532e-05, "loss": 3.8424, "step": 1300 }, { "epoch": 0.02177384235738133, "grad_norm": 6.864413261413574, "learning_rate": 9.999969064261341e-05, "loss": 3.79, "step": 1350 }, { "epoch": 0.02258028096321027, "grad_norm": 3.0512356758117676, "learning_rate": 9.999959477285881e-05, "loss": 3.7211, "step": 1400 }, { "epoch": 0.02338671956903921, "grad_norm": 3.6865158081054688, "learning_rate": 9.999948598272635e-05, "loss": 3.729, "step": 1450 }, { "epoch": 0.024193158174868147, "grad_norm": 3.479696750640869, "learning_rate": 9.999936427224411e-05, "loss": 3.8229, "step": 1500 }, { "epoch": 0.024999596780697086, "grad_norm": 3.896265745162964, "learning_rate": 9.999922964144355e-05, "loss": 4.0188, "step": 1550 }, { "epoch": 0.025806035386526024, "grad_norm": 1.8305364847183228, "learning_rate": 9.999908209035947e-05, "loss": 3.6971, "step": 1600 }, { "epoch": 0.026612473992354963, "grad_norm": 3.6553308963775635, "learning_rate": 9.999892161902997e-05, "loss": 3.9225, "step": 1650 }, { "epoch": 0.027418912598183902, "grad_norm": 3.454449415206909, "learning_rate": 9.999874822749654e-05, "loss": 4.1062, "step": 1700 }, { "epoch": 0.028225351204012837, "grad_norm": 6.318528652191162, "learning_rate": 9.9998561915804e-05, "loss": 4.165, "step": 1750 }, { "epoch": 0.029031789809841776, "grad_norm": 3.34511399269104, "learning_rate": 9.999836268400044e-05, "loss": 4.1559, "step": 1800 }, { "epoch": 0.029838228415670714, "grad_norm": 3.3025155067443848, "learning_rate": 9.999815053213741e-05, "loss": 3.6104, "step": 1850 }, { "epoch": 0.030644667021499653, "grad_norm": 1.3363052606582642, "learning_rate": 9.999792546026971e-05, "loss": 4.3776, "step": 1900 }, { "epoch": 0.03145110562732859, "grad_norm": 4.033843040466309, "learning_rate": 9.999768746845547e-05, "loss": 3.838, "step": 1950 }, { "epoch": 0.03225754423315753, "grad_norm": 3.517301082611084, "learning_rate": 9.999743655675622e-05, "loss": 4.0525, "step": 2000 }, { "epoch": 0.03306398283898647, "grad_norm": 3.5909571647644043, "learning_rate": 9.999717272523678e-05, "loss": 3.672, "step": 2050 }, { "epoch": 0.033870421444815405, "grad_norm": 2.088115930557251, "learning_rate": 9.999689597396536e-05, "loss": 3.9254, "step": 2100 }, { "epoch": 0.03467686005064435, "grad_norm": 3.169725179672241, "learning_rate": 9.999660630301345e-05, "loss": 3.9372, "step": 2150 }, { "epoch": 0.03548329865647328, "grad_norm": 3.584789991378784, "learning_rate": 9.999630371245588e-05, "loss": 3.9506, "step": 2200 }, { "epoch": 0.036289737262302224, "grad_norm": 2.678903579711914, "learning_rate": 9.999598820237088e-05, "loss": 3.7822, "step": 2250 }, { "epoch": 0.03709617586813116, "grad_norm": 2.6614108085632324, "learning_rate": 9.999565977283998e-05, "loss": 3.9777, "step": 2300 }, { "epoch": 0.037902614473960095, "grad_norm": 3.998502731323242, "learning_rate": 9.999531842394802e-05, "loss": 3.8676, "step": 2350 }, { "epoch": 0.03870905307978904, "grad_norm": 3.5454630851745605, "learning_rate": 9.999496415578324e-05, "loss": 3.9342, "step": 2400 }, { "epoch": 0.03951549168561797, "grad_norm": 4.353825569152832, "learning_rate": 9.999459696843719e-05, "loss": 4.0824, "step": 2450 }, { "epoch": 0.040321930291446914, "grad_norm": 2.4952590465545654, "learning_rate": 9.99942168620047e-05, "loss": 3.7487, "step": 2500 }, { "epoch": 0.04112836889727585, "grad_norm": 3.804739475250244, "learning_rate": 9.999382383658405e-05, "loss": 3.9278, "step": 2550 }, { "epoch": 0.04193480750310479, "grad_norm": 4.888896465301514, "learning_rate": 9.999341789227679e-05, "loss": 3.976, "step": 2600 }, { "epoch": 0.04274124610893373, "grad_norm": 2.8611414432525635, "learning_rate": 9.99929990291878e-05, "loss": 3.8008, "step": 2650 }, { "epoch": 0.04354768471476266, "grad_norm": 3.469597816467285, "learning_rate": 9.999256724742533e-05, "loss": 3.9826, "step": 2700 }, { "epoch": 0.044354123320591604, "grad_norm": 4.003543376922607, "learning_rate": 9.999212254710095e-05, "loss": 3.7801, "step": 2750 }, { "epoch": 0.04516056192642054, "grad_norm": 3.900905132293701, "learning_rate": 9.99916649283296e-05, "loss": 4.081, "step": 2800 }, { "epoch": 0.04596700053224948, "grad_norm": 1.68619966506958, "learning_rate": 9.99911943912295e-05, "loss": 3.8061, "step": 2850 }, { "epoch": 0.04677343913807842, "grad_norm": 4.65512752532959, "learning_rate": 9.999071093592227e-05, "loss": 3.8544, "step": 2900 }, { "epoch": 0.04757987774390736, "grad_norm": 2.2502596378326416, "learning_rate": 9.999021456253282e-05, "loss": 4.2133, "step": 2950 }, { "epoch": 0.048386316349736294, "grad_norm": 4.234059810638428, "learning_rate": 9.998970527118941e-05, "loss": 3.8952, "step": 3000 }, { "epoch": 0.048386316349736294, "eval_asr_loss": 0.6124729285552434, "eval_loss": 5.14359712600708, "eval_runtime": 17.1329, "eval_samples_per_second": 29.884, "eval_steps_per_second": 9.981, "eval_tts_loss": 6.1080074336570815, "step": 3000 }, { "epoch": 0.04919275495556523, "grad_norm": 4.5945658683776855, "learning_rate": 9.998918306202368e-05, "loss": 3.7152, "step": 3050 }, { "epoch": 0.04999919356139417, "grad_norm": 3.4985783100128174, "learning_rate": 9.998864793517053e-05, "loss": 4.0368, "step": 3100 }, { "epoch": 0.05080563216722311, "grad_norm": 2.991260528564453, "learning_rate": 9.99880998907683e-05, "loss": 3.961, "step": 3150 }, { "epoch": 0.05161207077305205, "grad_norm": 7.048410415649414, "learning_rate": 9.998753892895854e-05, "loss": 3.9461, "step": 3200 }, { "epoch": 0.052418509378880984, "grad_norm": 4.608008861541748, "learning_rate": 9.998696504988625e-05, "loss": 3.863, "step": 3250 }, { "epoch": 0.053224947984709926, "grad_norm": 3.2530033588409424, "learning_rate": 9.998637825369973e-05, "loss": 3.8894, "step": 3300 }, { "epoch": 0.05403138659053886, "grad_norm": 2.5475354194641113, "learning_rate": 9.99857785405506e-05, "loss": 3.7206, "step": 3350 }, { "epoch": 0.054837825196367804, "grad_norm": 4.753456115722656, "learning_rate": 9.998516591059384e-05, "loss": 3.7273, "step": 3400 }, { "epoch": 0.05564426380219674, "grad_norm": 2.6090848445892334, "learning_rate": 9.998454036398774e-05, "loss": 3.8879, "step": 3450 }, { "epoch": 0.056450702408025674, "grad_norm": 1.5343598127365112, "learning_rate": 9.998390190089396e-05, "loss": 4.0087, "step": 3500 }, { "epoch": 0.057257141013854616, "grad_norm": 4.9131364822387695, "learning_rate": 9.99832505214775e-05, "loss": 3.8471, "step": 3550 }, { "epoch": 0.05806357961968355, "grad_norm": 2.1183621883392334, "learning_rate": 9.998258622590666e-05, "loss": 3.9289, "step": 3600 }, { "epoch": 0.058870018225512494, "grad_norm": 3.5901293754577637, "learning_rate": 9.99819090143531e-05, "loss": 3.8583, "step": 3650 }, { "epoch": 0.05967645683134143, "grad_norm": 2.5325889587402344, "learning_rate": 9.998121888699183e-05, "loss": 3.953, "step": 3700 }, { "epoch": 0.06048289543717037, "grad_norm": 2.1782755851745605, "learning_rate": 9.99805158440012e-05, "loss": 4.013, "step": 3750 }, { "epoch": 0.061289334042999306, "grad_norm": 2.2709741592407227, "learning_rate": 9.997979988556285e-05, "loss": 4.1835, "step": 3800 }, { "epoch": 0.06209577264882824, "grad_norm": 1.7206313610076904, "learning_rate": 9.997907101186181e-05, "loss": 3.7028, "step": 3850 }, { "epoch": 0.06290221125465718, "grad_norm": 2.8325114250183105, "learning_rate": 9.997832922308643e-05, "loss": 3.8754, "step": 3900 }, { "epoch": 0.06370864986048612, "grad_norm": 3.2632429599761963, "learning_rate": 9.997757451942839e-05, "loss": 4.2381, "step": 3950 }, { "epoch": 0.06451508846631505, "grad_norm": 4.061680793762207, "learning_rate": 9.997680690108271e-05, "loss": 3.9123, "step": 4000 }, { "epoch": 0.065321527072144, "grad_norm": 3.529866933822632, "learning_rate": 9.997602636824775e-05, "loss": 3.8652, "step": 4050 }, { "epoch": 0.06612796567797294, "grad_norm": 3.3136708736419678, "learning_rate": 9.997523292112521e-05, "loss": 3.9462, "step": 4100 }, { "epoch": 0.06693440428380187, "grad_norm": 1.6283220052719116, "learning_rate": 9.997442655992013e-05, "loss": 4.1202, "step": 4150 }, { "epoch": 0.06774084288963081, "grad_norm": 3.355297088623047, "learning_rate": 9.997360728484086e-05, "loss": 4.0791, "step": 4200 }, { "epoch": 0.06854728149545974, "grad_norm": 2.3169095516204834, "learning_rate": 9.997277509609914e-05, "loss": 3.6696, "step": 4250 }, { "epoch": 0.0693537201012887, "grad_norm": 2.0649425983428955, "learning_rate": 9.997192999391e-05, "loss": 4.1419, "step": 4300 }, { "epoch": 0.07016015870711763, "grad_norm": 2.268578290939331, "learning_rate": 9.99710719784918e-05, "loss": 3.9995, "step": 4350 }, { "epoch": 0.07096659731294656, "grad_norm": 2.035593271255493, "learning_rate": 9.997020105006632e-05, "loss": 3.7817, "step": 4400 }, { "epoch": 0.0717730359187755, "grad_norm": 1.604856014251709, "learning_rate": 9.996931720885855e-05, "loss": 3.8819, "step": 4450 }, { "epoch": 0.07257947452460445, "grad_norm": 1.7782434225082397, "learning_rate": 9.996842045509692e-05, "loss": 4.2321, "step": 4500 }, { "epoch": 0.07338591313043338, "grad_norm": 1.1473257541656494, "learning_rate": 9.996751078901314e-05, "loss": 4.123, "step": 4550 }, { "epoch": 0.07419235173626232, "grad_norm": 4.596096515655518, "learning_rate": 9.996658821084231e-05, "loss": 4.4983, "step": 4600 }, { "epoch": 0.07499879034209125, "grad_norm": 4.046219348907471, "learning_rate": 9.99656527208228e-05, "loss": 3.832, "step": 4650 }, { "epoch": 0.07580522894792019, "grad_norm": 3.365840435028076, "learning_rate": 9.996470431919634e-05, "loss": 3.8089, "step": 4700 }, { "epoch": 0.07661166755374914, "grad_norm": 2.3088300228118896, "learning_rate": 9.996374300620805e-05, "loss": 4.2963, "step": 4750 }, { "epoch": 0.07741810615957807, "grad_norm": 3.0989160537719727, "learning_rate": 9.996276878210631e-05, "loss": 3.9085, "step": 4800 }, { "epoch": 0.07822454476540701, "grad_norm": 3.6229944229125977, "learning_rate": 9.996178164714289e-05, "loss": 3.9887, "step": 4850 }, { "epoch": 0.07903098337123594, "grad_norm": 2.169739007949829, "learning_rate": 9.996078160157285e-05, "loss": 4.053, "step": 4900 }, { "epoch": 0.07983742197706489, "grad_norm": 1.102918267250061, "learning_rate": 9.995976864565462e-05, "loss": 3.9429, "step": 4950 }, { "epoch": 0.08064386058289383, "grad_norm": 3.0187597274780273, "learning_rate": 9.995874277964997e-05, "loss": 4.1842, "step": 5000 }, { "epoch": 0.08145029918872276, "grad_norm": 6.297091484069824, "learning_rate": 9.995770400382398e-05, "loss": 3.8538, "step": 5050 }, { "epoch": 0.0822567377945517, "grad_norm": 2.396947145462036, "learning_rate": 9.995665231844511e-05, "loss": 3.9006, "step": 5100 }, { "epoch": 0.08306317640038063, "grad_norm": 1.8022737503051758, "learning_rate": 9.995558772378509e-05, "loss": 4.0055, "step": 5150 }, { "epoch": 0.08386961500620958, "grad_norm": 2.959127902984619, "learning_rate": 9.995451022011902e-05, "loss": 3.7882, "step": 5200 }, { "epoch": 0.08467605361203852, "grad_norm": 1.0544005632400513, "learning_rate": 9.995341980772535e-05, "loss": 3.8215, "step": 5250 }, { "epoch": 0.08548249221786745, "grad_norm": 4.3189616203308105, "learning_rate": 9.995231648688586e-05, "loss": 3.8656, "step": 5300 }, { "epoch": 0.08628893082369639, "grad_norm": 2.1376607418060303, "learning_rate": 9.995120025788565e-05, "loss": 3.69, "step": 5350 }, { "epoch": 0.08709536942952532, "grad_norm": 2.437530994415283, "learning_rate": 9.995007112101316e-05, "loss": 3.972, "step": 5400 }, { "epoch": 0.08790180803535427, "grad_norm": 2.361795663833618, "learning_rate": 9.994892907656018e-05, "loss": 3.9913, "step": 5450 }, { "epoch": 0.08870824664118321, "grad_norm": 1.6262253522872925, "learning_rate": 9.994777412482183e-05, "loss": 3.9927, "step": 5500 }, { "epoch": 0.08951468524701214, "grad_norm": 4.346071243286133, "learning_rate": 9.994660626609654e-05, "loss": 3.7355, "step": 5550 }, { "epoch": 0.09032112385284108, "grad_norm": 3.2261524200439453, "learning_rate": 9.994542550068612e-05, "loss": 3.8608, "step": 5600 }, { "epoch": 0.09112756245867003, "grad_norm": 3.2692320346832275, "learning_rate": 9.994423182889567e-05, "loss": 4.0809, "step": 5650 }, { "epoch": 0.09193400106449896, "grad_norm": 4.047648906707764, "learning_rate": 9.994302525103366e-05, "loss": 3.7864, "step": 5700 }, { "epoch": 0.0927404396703279, "grad_norm": 3.113825559616089, "learning_rate": 9.994180576741188e-05, "loss": 3.6328, "step": 5750 }, { "epoch": 0.09354687827615683, "grad_norm": 2.295882225036621, "learning_rate": 9.994057337834545e-05, "loss": 3.6168, "step": 5800 }, { "epoch": 0.09435331688198577, "grad_norm": 3.2631022930145264, "learning_rate": 9.993932808415283e-05, "loss": 4.1689, "step": 5850 }, { "epoch": 0.09515975548781472, "grad_norm": 3.607879877090454, "learning_rate": 9.993806988515581e-05, "loss": 4.0071, "step": 5900 }, { "epoch": 0.09596619409364365, "grad_norm": 1.2808860540390015, "learning_rate": 9.993679878167957e-05, "loss": 3.9552, "step": 5950 }, { "epoch": 0.09677263269947259, "grad_norm": 1.4205412864685059, "learning_rate": 9.99355147740525e-05, "loss": 4.4215, "step": 6000 }, { "epoch": 0.09677263269947259, "eval_asr_loss": 0.5728164760736751, "eval_loss": 5.094964027404785, "eval_runtime": 16.7544, "eval_samples_per_second": 30.559, "eval_steps_per_second": 10.206, "eval_tts_loss": 6.0641644553514515, "step": 6000 }, { "epoch": 0.09757907130530152, "grad_norm": 2.5851426124572754, "learning_rate": 9.993421786260645e-05, "loss": 3.941, "step": 6050 }, { "epoch": 0.09838550991113046, "grad_norm": 4.519574165344238, "learning_rate": 9.993290804767654e-05, "loss": 3.6548, "step": 6100 }, { "epoch": 0.09919194851695941, "grad_norm": 2.023059368133545, "learning_rate": 9.993158532960124e-05, "loss": 3.7762, "step": 6150 }, { "epoch": 0.09999838712278834, "grad_norm": 3.167097330093384, "learning_rate": 9.993024970872236e-05, "loss": 3.7466, "step": 6200 }, { "epoch": 0.10080482572861728, "grad_norm": 2.1411142349243164, "learning_rate": 9.992890118538502e-05, "loss": 3.7702, "step": 6250 }, { "epoch": 0.10161126433444621, "grad_norm": 2.5199944972991943, "learning_rate": 9.99275397599377e-05, "loss": 4.1723, "step": 6300 }, { "epoch": 0.10241770294027516, "grad_norm": 2.112910270690918, "learning_rate": 9.992616543273222e-05, "loss": 3.6662, "step": 6350 }, { "epoch": 0.1032241415461041, "grad_norm": 2.290219306945801, "learning_rate": 9.992477820412369e-05, "loss": 4.1363, "step": 6400 }, { "epoch": 0.10403058015193303, "grad_norm": 3.748711585998535, "learning_rate": 9.992337807447061e-05, "loss": 4.0218, "step": 6450 }, { "epoch": 0.10483701875776197, "grad_norm": 14.197625160217285, "learning_rate": 9.992196504413478e-05, "loss": 4.11, "step": 6500 }, { "epoch": 0.1056434573635909, "grad_norm": 2.3828155994415283, "learning_rate": 9.992053911348132e-05, "loss": 3.9976, "step": 6550 }, { "epoch": 0.10644989596941985, "grad_norm": 3.1666653156280518, "learning_rate": 9.991910028287874e-05, "loss": 3.8263, "step": 6600 }, { "epoch": 0.10725633457524879, "grad_norm": 1.068214774131775, "learning_rate": 9.991764855269882e-05, "loss": 3.7559, "step": 6650 }, { "epoch": 0.10806277318107772, "grad_norm": 2.6692686080932617, "learning_rate": 9.991618392331672e-05, "loss": 3.8083, "step": 6700 }, { "epoch": 0.10886921178690666, "grad_norm": 3.22259783744812, "learning_rate": 9.99147063951109e-05, "loss": 3.9687, "step": 6750 }, { "epoch": 0.10967565039273561, "grad_norm": 2.2858736515045166, "learning_rate": 9.991321596846317e-05, "loss": 3.7381, "step": 6800 }, { "epoch": 0.11048208899856454, "grad_norm": 3.2024693489074707, "learning_rate": 9.991171264375869e-05, "loss": 4.1968, "step": 6850 }, { "epoch": 0.11128852760439348, "grad_norm": 2.7798397541046143, "learning_rate": 9.99101964213859e-05, "loss": 3.7984, "step": 6900 }, { "epoch": 0.11209496621022241, "grad_norm": 2.2842605113983154, "learning_rate": 9.990866730173662e-05, "loss": 4.2311, "step": 6950 }, { "epoch": 0.11290140481605135, "grad_norm": 4.10762357711792, "learning_rate": 9.9907125285206e-05, "loss": 4.1217, "step": 7000 }, { "epoch": 0.1137078434218803, "grad_norm": 1.8413517475128174, "learning_rate": 9.99055703721925e-05, "loss": 4.1848, "step": 7050 }, { "epoch": 0.11451428202770923, "grad_norm": 2.565972328186035, "learning_rate": 9.990400256309792e-05, "loss": 3.744, "step": 7100 }, { "epoch": 0.11532072063353817, "grad_norm": 2.643573760986328, "learning_rate": 9.990242185832744e-05, "loss": 4.2744, "step": 7150 }, { "epoch": 0.1161271592393671, "grad_norm": 2.1358284950256348, "learning_rate": 9.990082825828946e-05, "loss": 4.2883, "step": 7200 }, { "epoch": 0.11693359784519604, "grad_norm": 1.9578022956848145, "learning_rate": 9.989922176339583e-05, "loss": 3.8243, "step": 7250 }, { "epoch": 0.11774003645102499, "grad_norm": 5.736109256744385, "learning_rate": 9.989760237406168e-05, "loss": 4.1798, "step": 7300 }, { "epoch": 0.11854647505685392, "grad_norm": 3.454503059387207, "learning_rate": 9.989597009070546e-05, "loss": 4.1567, "step": 7350 }, { "epoch": 0.11935291366268286, "grad_norm": 3.161022424697876, "learning_rate": 9.989432491374897e-05, "loss": 3.7965, "step": 7400 }, { "epoch": 0.1201593522685118, "grad_norm": 2.6754753589630127, "learning_rate": 9.989266684361735e-05, "loss": 3.9182, "step": 7450 }, { "epoch": 0.12096579087434074, "grad_norm": 2.2374236583709717, "learning_rate": 9.989099588073905e-05, "loss": 4.1068, "step": 7500 }, { "epoch": 0.12177222948016968, "grad_norm": 2.775463581085205, "learning_rate": 9.988931202554588e-05, "loss": 4.0023, "step": 7550 }, { "epoch": 0.12257866808599861, "grad_norm": 2.515188217163086, "learning_rate": 9.988761527847293e-05, "loss": 4.0234, "step": 7600 }, { "epoch": 0.12338510669182755, "grad_norm": 4.208799362182617, "learning_rate": 9.988590563995869e-05, "loss": 3.9328, "step": 7650 }, { "epoch": 0.12419154529765648, "grad_norm": 2.9487059116363525, "learning_rate": 9.988418311044495e-05, "loss": 3.8718, "step": 7700 }, { "epoch": 0.12499798390348543, "grad_norm": 2.8759098052978516, "learning_rate": 9.98824476903768e-05, "loss": 3.8538, "step": 7750 }, { "epoch": 0.12580442250931437, "grad_norm": 2.848405599594116, "learning_rate": 9.98806993802027e-05, "loss": 3.8749, "step": 7800 }, { "epoch": 0.1266108611151433, "grad_norm": 1.7490386962890625, "learning_rate": 9.987893818037443e-05, "loss": 3.9094, "step": 7850 }, { "epoch": 0.12741729972097224, "grad_norm": 2.6559858322143555, "learning_rate": 9.987716409134712e-05, "loss": 3.825, "step": 7900 }, { "epoch": 0.12822373832680117, "grad_norm": 2.877697229385376, "learning_rate": 9.987537711357919e-05, "loss": 3.9497, "step": 7950 }, { "epoch": 0.1290301769326301, "grad_norm": 2.17142391204834, "learning_rate": 9.987357724753242e-05, "loss": 3.9387, "step": 8000 }, { "epoch": 0.12983661553845904, "grad_norm": 1.643707036972046, "learning_rate": 9.987176449367192e-05, "loss": 4.2619, "step": 8050 }, { "epoch": 0.130643054144288, "grad_norm": 4.4552998542785645, "learning_rate": 9.98699388524661e-05, "loss": 4.0635, "step": 8100 }, { "epoch": 0.13144949275011694, "grad_norm": 3.406740427017212, "learning_rate": 9.986810032438674e-05, "loss": 4.0747, "step": 8150 }, { "epoch": 0.13225593135594588, "grad_norm": 5.450404644012451, "learning_rate": 9.986624890990893e-05, "loss": 3.8341, "step": 8200 }, { "epoch": 0.1330623699617748, "grad_norm": 2.6635243892669678, "learning_rate": 9.986438460951111e-05, "loss": 3.8464, "step": 8250 }, { "epoch": 0.13386880856760375, "grad_norm": 1.887026309967041, "learning_rate": 9.9862507423675e-05, "loss": 3.7535, "step": 8300 }, { "epoch": 0.13467524717343268, "grad_norm": 2.086843490600586, "learning_rate": 9.98606173528857e-05, "loss": 4.2659, "step": 8350 }, { "epoch": 0.13548168577926162, "grad_norm": 3.2618448734283447, "learning_rate": 9.985871439763163e-05, "loss": 4.0077, "step": 8400 }, { "epoch": 0.13628812438509055, "grad_norm": 3.416794538497925, "learning_rate": 9.985679855840453e-05, "loss": 4.2068, "step": 8450 }, { "epoch": 0.1370945629909195, "grad_norm": 2.5768439769744873, "learning_rate": 9.985486983569945e-05, "loss": 4.2102, "step": 8500 }, { "epoch": 0.13790100159674845, "grad_norm": 2.0248680114746094, "learning_rate": 9.985296718837846e-05, "loss": 3.7619, "step": 8550 }, { "epoch": 0.1387074402025774, "grad_norm": 2.308393955230713, "learning_rate": 9.985101295786062e-05, "loss": 4.0467, "step": 8600 }, { "epoch": 0.13951387880840632, "grad_norm": 3.1832563877105713, "learning_rate": 9.984904584535987e-05, "loss": 4.0234, "step": 8650 }, { "epoch": 0.14032031741423526, "grad_norm": 2.4449775218963623, "learning_rate": 9.984706585138452e-05, "loss": 3.4602, "step": 8700 }, { "epoch": 0.1411267560200642, "grad_norm": 3.088870048522949, "learning_rate": 9.984507297644625e-05, "loss": 4.0023, "step": 8750 }, { "epoch": 0.14193319462589313, "grad_norm": 1.3331509828567505, "learning_rate": 9.984306722106e-05, "loss": 4.4195, "step": 8800 }, { "epoch": 0.14273963323172206, "grad_norm": 2.1085762977600098, "learning_rate": 9.98410485857441e-05, "loss": 3.6874, "step": 8850 }, { "epoch": 0.143546071837551, "grad_norm": 2.382269859313965, "learning_rate": 9.983905782752949e-05, "loss": 4.1141, "step": 8900 }, { "epoch": 0.14435251044337993, "grad_norm": 1.5278171300888062, "learning_rate": 9.983701369149501e-05, "loss": 3.8317, "step": 8950 }, { "epoch": 0.1451589490492089, "grad_norm": 5.558732986450195, "learning_rate": 9.983495667709518e-05, "loss": 3.6433, "step": 9000 }, { "epoch": 0.1451589490492089, "eval_asr_loss": 0.5558241450032482, "eval_loss": 5.0716023445129395, "eval_runtime": 15.4575, "eval_samples_per_second": 33.123, "eval_steps_per_second": 11.063, "eval_tts_loss": 6.0293835261628255, "step": 9000 }, { "epoch": 0.14596538765503783, "grad_norm": 4.300561904907227, "learning_rate": 9.983288678486153e-05, "loss": 4.3128, "step": 9050 }, { "epoch": 0.14677182626086677, "grad_norm": 1.5210169553756714, "learning_rate": 9.983080401532894e-05, "loss": 3.9063, "step": 9100 }, { "epoch": 0.1475782648666957, "grad_norm": 2.6939990520477295, "learning_rate": 9.982870836903564e-05, "loss": 3.9087, "step": 9150 }, { "epoch": 0.14838470347252464, "grad_norm": 1.7290055751800537, "learning_rate": 9.982659984652316e-05, "loss": 4.1546, "step": 9200 }, { "epoch": 0.14919114207835357, "grad_norm": 4.180938243865967, "learning_rate": 9.982447844833634e-05, "loss": 4.0063, "step": 9250 }, { "epoch": 0.1499975806841825, "grad_norm": 3.3321378231048584, "learning_rate": 9.98223441750234e-05, "loss": 3.6936, "step": 9300 }, { "epoch": 0.15080401929001144, "grad_norm": 3.3349199295043945, "learning_rate": 9.982019702713583e-05, "loss": 3.9439, "step": 9350 }, { "epoch": 0.15161045789584038, "grad_norm": 2.3645851612091064, "learning_rate": 9.98180370052285e-05, "loss": 4.0444, "step": 9400 }, { "epoch": 0.15241689650166934, "grad_norm": 3.225592613220215, "learning_rate": 9.981586410985955e-05, "loss": 3.8988, "step": 9450 }, { "epoch": 0.15322333510749828, "grad_norm": 3.2834198474884033, "learning_rate": 9.981367834159049e-05, "loss": 4.1237, "step": 9500 }, { "epoch": 0.1540297737133272, "grad_norm": 1.3256886005401611, "learning_rate": 9.981147970098615e-05, "loss": 3.9196, "step": 9550 }, { "epoch": 0.15483621231915615, "grad_norm": 3.559396743774414, "learning_rate": 9.980926818861469e-05, "loss": 4.1997, "step": 9600 }, { "epoch": 0.15564265092498508, "grad_norm": 3.722397565841675, "learning_rate": 9.980704380504756e-05, "loss": 4.2779, "step": 9650 }, { "epoch": 0.15644908953081402, "grad_norm": 3.125638723373413, "learning_rate": 9.980480655085957e-05, "loss": 4.0461, "step": 9700 }, { "epoch": 0.15725552813664295, "grad_norm": 1.543898105621338, "learning_rate": 9.980255642662888e-05, "loss": 3.9528, "step": 9750 }, { "epoch": 0.1580619667424719, "grad_norm": 3.1968443393707275, "learning_rate": 9.980029343293688e-05, "loss": 4.2141, "step": 9800 }, { "epoch": 0.15886840534830082, "grad_norm": 1.8708412647247314, "learning_rate": 9.979801757036841e-05, "loss": 4.1029, "step": 9850 }, { "epoch": 0.15967484395412979, "grad_norm": 4.089094161987305, "learning_rate": 9.979572883951156e-05, "loss": 4.0536, "step": 9900 }, { "epoch": 0.16048128255995872, "grad_norm": 1.7748725414276123, "learning_rate": 9.979342724095774e-05, "loss": 4.0823, "step": 9950 }, { "epoch": 0.16128772116578766, "grad_norm": 2.352545976638794, "learning_rate": 9.97911127753017e-05, "loss": 3.8066, "step": 10000 }, { "epoch": 0.1620941597716166, "grad_norm": 2.2634575366973877, "learning_rate": 9.978878544314156e-05, "loss": 4.1763, "step": 10050 }, { "epoch": 0.16290059837744553, "grad_norm": 2.911471366882324, "learning_rate": 9.978644524507869e-05, "loss": 4.2928, "step": 10100 }, { "epoch": 0.16370703698327446, "grad_norm": 1.9891984462738037, "learning_rate": 9.978409218171784e-05, "loss": 4.0966, "step": 10150 }, { "epoch": 0.1645134755891034, "grad_norm": 2.632568597793579, "learning_rate": 9.978172625366705e-05, "loss": 4.2746, "step": 10200 }, { "epoch": 0.16531991419493233, "grad_norm": 2.3219494819641113, "learning_rate": 9.977934746153771e-05, "loss": 4.3219, "step": 10250 }, { "epoch": 0.16612635280076127, "grad_norm": 2.416696071624756, "learning_rate": 9.977695580594451e-05, "loss": 4.0857, "step": 10300 }, { "epoch": 0.1669327914065902, "grad_norm": 1.758725881576538, "learning_rate": 9.97745512875055e-05, "loss": 3.7797, "step": 10350 }, { "epoch": 0.16773923001241917, "grad_norm": 2.4825093746185303, "learning_rate": 9.977213390684199e-05, "loss": 3.8742, "step": 10400 }, { "epoch": 0.1685456686182481, "grad_norm": 1.305153727531433, "learning_rate": 9.97697036645787e-05, "loss": 4.1835, "step": 10450 }, { "epoch": 0.16935210722407704, "grad_norm": 4.981587886810303, "learning_rate": 9.97672605613436e-05, "loss": 4.2102, "step": 10500 }, { "epoch": 0.17015854582990597, "grad_norm": 2.5329973697662354, "learning_rate": 9.976480459776803e-05, "loss": 4.1571, "step": 10550 }, { "epoch": 0.1709649844357349, "grad_norm": 3.1804792881011963, "learning_rate": 9.976233577448661e-05, "loss": 4.0007, "step": 10600 }, { "epoch": 0.17177142304156384, "grad_norm": 2.1059024333953857, "learning_rate": 9.975985409213731e-05, "loss": 4.096, "step": 10650 }, { "epoch": 0.17257786164739278, "grad_norm": 2.35027813911438, "learning_rate": 9.975735955136145e-05, "loss": 4.2416, "step": 10700 }, { "epoch": 0.1733843002532217, "grad_norm": 4.606429576873779, "learning_rate": 9.975485215280362e-05, "loss": 3.7404, "step": 10750 }, { "epoch": 0.17419073885905065, "grad_norm": 2.3107430934906006, "learning_rate": 9.975233189711177e-05, "loss": 3.9696, "step": 10800 }, { "epoch": 0.1749971774648796, "grad_norm": 1.1612375974655151, "learning_rate": 9.974979878493716e-05, "loss": 4.0279, "step": 10850 }, { "epoch": 0.17580361607070855, "grad_norm": 1.4754469394683838, "learning_rate": 9.974725281693433e-05, "loss": 3.5668, "step": 10900 }, { "epoch": 0.17661005467653748, "grad_norm": 1.7385231256484985, "learning_rate": 9.974469399376123e-05, "loss": 3.8299, "step": 10950 }, { "epoch": 0.17741649328236642, "grad_norm": 2.5821852684020996, "learning_rate": 9.974212231607907e-05, "loss": 3.8388, "step": 11000 }, { "epoch": 0.17822293188819535, "grad_norm": 1.8914719820022583, "learning_rate": 9.973953778455239e-05, "loss": 4.0202, "step": 11050 }, { "epoch": 0.1790293704940243, "grad_norm": 1.3121271133422852, "learning_rate": 9.973694039984907e-05, "loss": 4.0743, "step": 11100 }, { "epoch": 0.17983580909985322, "grad_norm": 2.854396104812622, "learning_rate": 9.973433016264029e-05, "loss": 4.066, "step": 11150 }, { "epoch": 0.18064224770568216, "grad_norm": 3.5787723064422607, "learning_rate": 9.973170707360055e-05, "loss": 4.3989, "step": 11200 }, { "epoch": 0.1814486863115111, "grad_norm": 1.8557301759719849, "learning_rate": 9.972912397814846e-05, "loss": 4.0768, "step": 11250 }, { "epoch": 0.18225512491734006, "grad_norm": 2.253882884979248, "learning_rate": 9.97264754444864e-05, "loss": 3.9537, "step": 11300 }, { "epoch": 0.183061563523169, "grad_norm": 3.240039825439453, "learning_rate": 9.972381406102311e-05, "loss": 4.0367, "step": 11350 }, { "epoch": 0.18386800212899793, "grad_norm": 3.579876184463501, "learning_rate": 9.972113982844634e-05, "loss": 3.8319, "step": 11400 }, { "epoch": 0.18467444073482686, "grad_norm": 0.6159800291061401, "learning_rate": 9.971845274744713e-05, "loss": 3.8889, "step": 11450 }, { "epoch": 0.1854808793406558, "grad_norm": 1.1513911485671997, "learning_rate": 9.971575281871985e-05, "loss": 3.8983, "step": 11500 }, { "epoch": 0.18628731794648473, "grad_norm": 3.410205125808716, "learning_rate": 9.971304004296219e-05, "loss": 4.2003, "step": 11550 }, { "epoch": 0.18709375655231367, "grad_norm": 1.6599464416503906, "learning_rate": 9.971031442087515e-05, "loss": 3.7348, "step": 11600 }, { "epoch": 0.1879001951581426, "grad_norm": 2.8847126960754395, "learning_rate": 9.970757595316306e-05, "loss": 3.9562, "step": 11650 }, { "epoch": 0.18870663376397154, "grad_norm": 2.053612470626831, "learning_rate": 9.970482464053359e-05, "loss": 4.3117, "step": 11700 }, { "epoch": 0.1895130723698005, "grad_norm": 2.1683125495910645, "learning_rate": 9.970206048369766e-05, "loss": 3.8281, "step": 11750 }, { "epoch": 0.19031951097562944, "grad_norm": 5.38385009765625, "learning_rate": 9.969928348336959e-05, "loss": 4.1646, "step": 11800 }, { "epoch": 0.19112594958145837, "grad_norm": 1.783724069595337, "learning_rate": 9.969649364026696e-05, "loss": 3.6177, "step": 11850 }, { "epoch": 0.1919323881872873, "grad_norm": 2.6397554874420166, "learning_rate": 9.969369095511072e-05, "loss": 4.1144, "step": 11900 }, { "epoch": 0.19273882679311624, "grad_norm": 3.0491504669189453, "learning_rate": 9.969087542862509e-05, "loss": 4.1832, "step": 11950 }, { "epoch": 0.19354526539894518, "grad_norm": 3.2440996170043945, "learning_rate": 9.968804706153762e-05, "loss": 3.9472, "step": 12000 }, { "epoch": 0.19354526539894518, "eval_asr_loss": 0.5453353807244307, "eval_loss": 5.0482306480407715, "eval_runtime": 15.3432, "eval_samples_per_second": 33.37, "eval_steps_per_second": 11.145, "eval_tts_loss": 6.005206208861255, "step": 12000 }, { "epoch": 0.1943517040047741, "grad_norm": 4.186015605926514, "learning_rate": 9.968520585457922e-05, "loss": 3.8557, "step": 12050 }, { "epoch": 0.19515814261060305, "grad_norm": 2.497833013534546, "learning_rate": 9.968235180848406e-05, "loss": 4.2129, "step": 12100 }, { "epoch": 0.19596458121643198, "grad_norm": 3.930877685546875, "learning_rate": 9.967948492398967e-05, "loss": 4.3072, "step": 12150 }, { "epoch": 0.19677101982226092, "grad_norm": 1.6300374269485474, "learning_rate": 9.967660520183686e-05, "loss": 3.6067, "step": 12200 }, { "epoch": 0.19757745842808988, "grad_norm": 1.7998709678649902, "learning_rate": 9.96737126427698e-05, "loss": 4.2948, "step": 12250 }, { "epoch": 0.19838389703391882, "grad_norm": 2.9288175106048584, "learning_rate": 9.967080724753592e-05, "loss": 3.6403, "step": 12300 }, { "epoch": 0.19919033563974775, "grad_norm": 1.8051952123641968, "learning_rate": 9.966788901688605e-05, "loss": 3.9332, "step": 12350 }, { "epoch": 0.1999967742455767, "grad_norm": 2.3929975032806396, "learning_rate": 9.966495795157427e-05, "loss": 3.934, "step": 12400 }, { "epoch": 0.20080321285140562, "grad_norm": 2.7234630584716797, "learning_rate": 9.966201405235797e-05, "loss": 4.1635, "step": 12450 }, { "epoch": 0.20160965145723456, "grad_norm": 2.5407304763793945, "learning_rate": 9.965905731999791e-05, "loss": 3.6558, "step": 12500 }, { "epoch": 0.2024160900630635, "grad_norm": 1.8946152925491333, "learning_rate": 9.965608775525815e-05, "loss": 3.7618, "step": 12550 }, { "epoch": 0.20322252866889243, "grad_norm": 2.083383798599243, "learning_rate": 9.965310535890601e-05, "loss": 4.3816, "step": 12600 }, { "epoch": 0.20402896727472136, "grad_norm": 2.6187212467193604, "learning_rate": 9.96501101317122e-05, "loss": 3.9503, "step": 12650 }, { "epoch": 0.20483540588055033, "grad_norm": 4.162103176116943, "learning_rate": 9.964710207445072e-05, "loss": 4.0746, "step": 12700 }, { "epoch": 0.20564184448637926, "grad_norm": 2.638073682785034, "learning_rate": 9.964408118789886e-05, "loss": 4.3604, "step": 12750 }, { "epoch": 0.2064482830922082, "grad_norm": 2.0001702308654785, "learning_rate": 9.964104747283727e-05, "loss": 4.0921, "step": 12800 }, { "epoch": 0.20725472169803713, "grad_norm": 2.3797965049743652, "learning_rate": 9.963800093004987e-05, "loss": 4.3402, "step": 12850 }, { "epoch": 0.20806116030386607, "grad_norm": 2.3531606197357178, "learning_rate": 9.963494156032393e-05, "loss": 4.0507, "step": 12900 }, { "epoch": 0.208867598909695, "grad_norm": 3.526831865310669, "learning_rate": 9.963186936445001e-05, "loss": 4.3212, "step": 12950 }, { "epoch": 0.20967403751552394, "grad_norm": 4.21464729309082, "learning_rate": 9.962878434322202e-05, "loss": 4.1086, "step": 13000 }, { "epoch": 0.21048047612135287, "grad_norm": 2.4508001804351807, "learning_rate": 9.962568649743713e-05, "loss": 3.6248, "step": 13050 }, { "epoch": 0.2112869147271818, "grad_norm": 4.681621551513672, "learning_rate": 9.962257582789585e-05, "loss": 3.9501, "step": 13100 }, { "epoch": 0.21209335333301077, "grad_norm": 3.483264207839966, "learning_rate": 9.961945233540204e-05, "loss": 3.7879, "step": 13150 }, { "epoch": 0.2128997919388397, "grad_norm": 2.662280321121216, "learning_rate": 9.961631602076282e-05, "loss": 4.2567, "step": 13200 }, { "epoch": 0.21370623054466864, "grad_norm": 1.8116915225982666, "learning_rate": 9.961316688478863e-05, "loss": 4.1235, "step": 13250 }, { "epoch": 0.21451266915049758, "grad_norm": 3.8289411067962646, "learning_rate": 9.961000492829326e-05, "loss": 4.0276, "step": 13300 }, { "epoch": 0.2153191077563265, "grad_norm": 1.5876435041427612, "learning_rate": 9.96068301520938e-05, "loss": 3.7603, "step": 13350 }, { "epoch": 0.21612554636215545, "grad_norm": 3.7507176399230957, "learning_rate": 9.960370643453201e-05, "loss": 3.6352, "step": 13400 }, { "epoch": 0.21693198496798438, "grad_norm": 1.9685125350952148, "learning_rate": 9.96005062777419e-05, "loss": 3.6024, "step": 13450 }, { "epoch": 0.21773842357381332, "grad_norm": 2.623763084411621, "learning_rate": 9.959729330370225e-05, "loss": 4.0665, "step": 13500 }, { "epoch": 0.21854486217964225, "grad_norm": 1.6213947534561157, "learning_rate": 9.95940675132433e-05, "loss": 3.9128, "step": 13550 }, { "epoch": 0.21935130078547121, "grad_norm": 5.413779258728027, "learning_rate": 9.959082890719862e-05, "loss": 3.7172, "step": 13600 }, { "epoch": 0.22015773939130015, "grad_norm": 3.002934694290161, "learning_rate": 9.95875774864051e-05, "loss": 3.7273, "step": 13650 }, { "epoch": 0.22096417799712909, "grad_norm": 2.6499838829040527, "learning_rate": 9.958431325170294e-05, "loss": 4.2387, "step": 13700 }, { "epoch": 0.22177061660295802, "grad_norm": 1.5989998579025269, "learning_rate": 9.958103620393564e-05, "loss": 4.0674, "step": 13750 }, { "epoch": 0.22257705520878696, "grad_norm": 2.529733180999756, "learning_rate": 9.957774634395003e-05, "loss": 4.1921, "step": 13800 }, { "epoch": 0.2233834938146159, "grad_norm": 6.773332595825195, "learning_rate": 9.957444367259625e-05, "loss": 3.947, "step": 13850 }, { "epoch": 0.22418993242044483, "grad_norm": 1.7639790773391724, "learning_rate": 9.957112819072774e-05, "loss": 3.9777, "step": 13900 }, { "epoch": 0.22499637102627376, "grad_norm": 1.8540008068084717, "learning_rate": 9.956779989920124e-05, "loss": 4.0806, "step": 13950 }, { "epoch": 0.2258028096321027, "grad_norm": 2.861515998840332, "learning_rate": 9.956445879887683e-05, "loss": 4.1824, "step": 14000 }, { "epoch": 0.22660924823793166, "grad_norm": 2.093233823776245, "learning_rate": 9.956110489061786e-05, "loss": 4.0445, "step": 14050 }, { "epoch": 0.2274156868437606, "grad_norm": 2.043487071990967, "learning_rate": 9.955773817529103e-05, "loss": 3.8989, "step": 14100 }, { "epoch": 0.22822212544958953, "grad_norm": 1.743752121925354, "learning_rate": 9.955435865376632e-05, "loss": 4.2283, "step": 14150 }, { "epoch": 0.22902856405541847, "grad_norm": 1.1774280071258545, "learning_rate": 9.955096632691704e-05, "loss": 3.8492, "step": 14200 }, { "epoch": 0.2298350026612474, "grad_norm": 0.7757024168968201, "learning_rate": 9.95475611956198e-05, "loss": 3.8286, "step": 14250 }, { "epoch": 0.23064144126707634, "grad_norm": 2.530691623687744, "learning_rate": 9.954414326075452e-05, "loss": 4.0583, "step": 14300 }, { "epoch": 0.23144787987290527, "grad_norm": 2.322361469268799, "learning_rate": 9.95407125232044e-05, "loss": 4.188, "step": 14350 }, { "epoch": 0.2322543184787342, "grad_norm": 2.6633105278015137, "learning_rate": 9.953726898385603e-05, "loss": 3.6949, "step": 14400 }, { "epoch": 0.23306075708456314, "grad_norm": 3.243053436279297, "learning_rate": 9.95338126435992e-05, "loss": 3.733, "step": 14450 }, { "epoch": 0.23386719569039208, "grad_norm": 2.6763391494750977, "learning_rate": 9.953034350332709e-05, "loss": 3.9605, "step": 14500 }, { "epoch": 0.23467363429622104, "grad_norm": 2.8131678104400635, "learning_rate": 9.952686156393616e-05, "loss": 4.2417, "step": 14550 }, { "epoch": 0.23548007290204997, "grad_norm": 2.5453968048095703, "learning_rate": 9.952336682632617e-05, "loss": 4.5493, "step": 14600 }, { "epoch": 0.2362865115078789, "grad_norm": 2.9333486557006836, "learning_rate": 9.951985929140018e-05, "loss": 4.0338, "step": 14650 }, { "epoch": 0.23709295011370785, "grad_norm": 1.5506014823913574, "learning_rate": 9.951633896006461e-05, "loss": 3.9207, "step": 14700 }, { "epoch": 0.23789938871953678, "grad_norm": 1.8103605508804321, "learning_rate": 9.95128058332291e-05, "loss": 3.9589, "step": 14750 }, { "epoch": 0.23870582732536572, "grad_norm": 2.344365358352661, "learning_rate": 9.950925991180669e-05, "loss": 4.0624, "step": 14800 }, { "epoch": 0.23951226593119465, "grad_norm": 2.332612991333008, "learning_rate": 9.950570119671362e-05, "loss": 3.9549, "step": 14850 }, { "epoch": 0.2403187045370236, "grad_norm": 2.706002712249756, "learning_rate": 9.950212968886956e-05, "loss": 3.6544, "step": 14900 }, { "epoch": 0.24112514314285252, "grad_norm": 2.335348606109619, "learning_rate": 9.94985453891974e-05, "loss": 3.9361, "step": 14950 }, { "epoch": 0.24193158174868148, "grad_norm": 1.8754454851150513, "learning_rate": 9.949494829862334e-05, "loss": 3.6963, "step": 15000 }, { "epoch": 0.24193158174868148, "eval_asr_loss": 0.560360777742155, "eval_loss": 5.027297496795654, "eval_runtime": 15.4664, "eval_samples_per_second": 33.104, "eval_steps_per_second": 11.056, "eval_tts_loss": 5.95218636007564, "step": 15000 }, { "epoch": 0.24273802035451042, "grad_norm": 1.9016485214233398, "learning_rate": 9.949133841807691e-05, "loss": 3.9233, "step": 15050 }, { "epoch": 0.24354445896033936, "grad_norm": 2.494946002960205, "learning_rate": 9.948771574849097e-05, "loss": 4.087, "step": 15100 }, { "epoch": 0.2443508975661683, "grad_norm": 3.006345748901367, "learning_rate": 9.948408029080161e-05, "loss": 4.0706, "step": 15150 }, { "epoch": 0.24515733617199723, "grad_norm": 3.3929615020751953, "learning_rate": 9.94804320459483e-05, "loss": 3.9753, "step": 15200 }, { "epoch": 0.24596377477782616, "grad_norm": 1.2217860221862793, "learning_rate": 9.947677101487374e-05, "loss": 4.1967, "step": 15250 }, { "epoch": 0.2467702133836551, "grad_norm": 1.409860610961914, "learning_rate": 9.947309719852402e-05, "loss": 3.9026, "step": 15300 }, { "epoch": 0.24757665198948403, "grad_norm": 1.3178203105926514, "learning_rate": 9.946941059784847e-05, "loss": 4.1695, "step": 15350 }, { "epoch": 0.24838309059531297, "grad_norm": 3.196470260620117, "learning_rate": 9.946571121379974e-05, "loss": 3.8572, "step": 15400 }, { "epoch": 0.24918952920114193, "grad_norm": 2.9058384895324707, "learning_rate": 9.946199904733378e-05, "loss": 4.2121, "step": 15450 }, { "epoch": 0.24999596780697086, "grad_norm": 2.940702438354492, "learning_rate": 9.945827409940989e-05, "loss": 4.2432, "step": 15500 }, { "epoch": 0.25080240641279977, "grad_norm": 3.044480562210083, "learning_rate": 9.945453637099058e-05, "loss": 4.0084, "step": 15550 }, { "epoch": 0.25160884501862874, "grad_norm": 2.350438356399536, "learning_rate": 9.945078586304175e-05, "loss": 4.1791, "step": 15600 }, { "epoch": 0.25241528362445764, "grad_norm": 2.298433542251587, "learning_rate": 9.944702257653255e-05, "loss": 3.9251, "step": 15650 }, { "epoch": 0.2532217222302866, "grad_norm": 2.5852606296539307, "learning_rate": 9.944324651243545e-05, "loss": 3.9519, "step": 15700 }, { "epoch": 0.25402816083611557, "grad_norm": 1.654619812965393, "learning_rate": 9.943945767172624e-05, "loss": 3.7628, "step": 15750 }, { "epoch": 0.2548345994419445, "grad_norm": 2.320345878601074, "learning_rate": 9.943573221290566e-05, "loss": 4.3502, "step": 15800 }, { "epoch": 0.25564103804777344, "grad_norm": 1.5587137937545776, "learning_rate": 9.943191807739609e-05, "loss": 4.3284, "step": 15850 }, { "epoch": 0.25644747665360235, "grad_norm": 2.629159927368164, "learning_rate": 9.942809116820176e-05, "loss": 3.9416, "step": 15900 }, { "epoch": 0.2572539152594313, "grad_norm": 2.840963363647461, "learning_rate": 9.942425148631161e-05, "loss": 4.0621, "step": 15950 }, { "epoch": 0.2580603538652602, "grad_norm": 1.6260427236557007, "learning_rate": 9.942039903271782e-05, "loss": 3.7273, "step": 16000 }, { "epoch": 0.2588667924710892, "grad_norm": 1.4383156299591064, "learning_rate": 9.941653380841591e-05, "loss": 4.262, "step": 16050 }, { "epoch": 0.2596732310769181, "grad_norm": 1.3447210788726807, "learning_rate": 9.941265581440471e-05, "loss": 3.6917, "step": 16100 }, { "epoch": 0.26047966968274705, "grad_norm": 1.707841157913208, "learning_rate": 9.940876505168633e-05, "loss": 3.7174, "step": 16150 }, { "epoch": 0.261286108288576, "grad_norm": 2.2020676136016846, "learning_rate": 9.940486152126614e-05, "loss": 4.2535, "step": 16200 }, { "epoch": 0.2620925468944049, "grad_norm": 1.9968034029006958, "learning_rate": 9.94009452241529e-05, "loss": 4.2131, "step": 16250 }, { "epoch": 0.2628989855002339, "grad_norm": 4.067761421203613, "learning_rate": 9.939701616135858e-05, "loss": 4.0394, "step": 16300 }, { "epoch": 0.2637054241060628, "grad_norm": 1.6382523775100708, "learning_rate": 9.939307433389852e-05, "loss": 3.7792, "step": 16350 }, { "epoch": 0.26451186271189175, "grad_norm": 1.519000768661499, "learning_rate": 9.938911974279132e-05, "loss": 3.7692, "step": 16400 }, { "epoch": 0.26531830131772066, "grad_norm": 2.7159464359283447, "learning_rate": 9.938515238905886e-05, "loss": 3.8706, "step": 16450 }, { "epoch": 0.2661247399235496, "grad_norm": 4.7422871589660645, "learning_rate": 9.938117227372636e-05, "loss": 4.1009, "step": 16500 }, { "epoch": 0.26693117852937853, "grad_norm": 2.702423095703125, "learning_rate": 9.937717939782233e-05, "loss": 3.7776, "step": 16550 }, { "epoch": 0.2677376171352075, "grad_norm": 3.021857738494873, "learning_rate": 9.937317376237856e-05, "loss": 4.1462, "step": 16600 }, { "epoch": 0.26854405574103646, "grad_norm": 1.813398003578186, "learning_rate": 9.936915536843012e-05, "loss": 4.0327, "step": 16650 }, { "epoch": 0.26935049434686537, "grad_norm": 2.897822856903076, "learning_rate": 9.936512421701545e-05, "loss": 3.9975, "step": 16700 }, { "epoch": 0.27015693295269433, "grad_norm": 5.469192028045654, "learning_rate": 9.936108030917619e-05, "loss": 3.9002, "step": 16750 }, { "epoch": 0.27096337155852324, "grad_norm": 1.7145867347717285, "learning_rate": 9.935702364595736e-05, "loss": 3.711, "step": 16800 }, { "epoch": 0.2717698101643522, "grad_norm": 2.104463577270508, "learning_rate": 9.935295422840722e-05, "loss": 4.1141, "step": 16850 }, { "epoch": 0.2725762487701811, "grad_norm": 3.0537919998168945, "learning_rate": 9.934887205757735e-05, "loss": 3.7435, "step": 16900 }, { "epoch": 0.27338268737601007, "grad_norm": 1.5938163995742798, "learning_rate": 9.934477713452264e-05, "loss": 4.2125, "step": 16950 }, { "epoch": 0.274189125981839, "grad_norm": 1.9021580219268799, "learning_rate": 9.934066946030124e-05, "loss": 3.6318, "step": 17000 }, { "epoch": 0.27499556458766794, "grad_norm": 4.621829986572266, "learning_rate": 9.933654903597461e-05, "loss": 3.6992, "step": 17050 }, { "epoch": 0.2758020031934969, "grad_norm": 2.142982244491577, "learning_rate": 9.933241586260754e-05, "loss": 3.7732, "step": 17100 }, { "epoch": 0.2766084417993258, "grad_norm": 3.7406392097473145, "learning_rate": 9.932826994126802e-05, "loss": 4.1151, "step": 17150 }, { "epoch": 0.2774148804051548, "grad_norm": 0.5368770360946655, "learning_rate": 9.932411127302745e-05, "loss": 3.7562, "step": 17200 }, { "epoch": 0.2782213190109837, "grad_norm": 0.6764758825302124, "learning_rate": 9.931993985896045e-05, "loss": 3.8355, "step": 17250 }, { "epoch": 0.27902775761681264, "grad_norm": 1.2389382123947144, "learning_rate": 9.931575570014495e-05, "loss": 3.9448, "step": 17300 }, { "epoch": 0.27983419622264155, "grad_norm": 2.505661964416504, "learning_rate": 9.931155879766221e-05, "loss": 3.8676, "step": 17350 }, { "epoch": 0.2806406348284705, "grad_norm": 1.7408605813980103, "learning_rate": 9.930734915259669e-05, "loss": 3.8457, "step": 17400 }, { "epoch": 0.2814470734342994, "grad_norm": 1.7804434299468994, "learning_rate": 9.930312676603626e-05, "loss": 3.8563, "step": 17450 }, { "epoch": 0.2822535120401284, "grad_norm": 2.6170337200164795, "learning_rate": 9.9298891639072e-05, "loss": 4.1211, "step": 17500 }, { "epoch": 0.28305995064595735, "grad_norm": 2.365858316421509, "learning_rate": 9.92946437727983e-05, "loss": 4.1625, "step": 17550 }, { "epoch": 0.28386638925178626, "grad_norm": 2.3039627075195312, "learning_rate": 9.929038316831285e-05, "loss": 3.9853, "step": 17600 }, { "epoch": 0.2846728278576152, "grad_norm": 2.711247444152832, "learning_rate": 9.928610982671665e-05, "loss": 4.1158, "step": 17650 }, { "epoch": 0.2854792664634441, "grad_norm": 2.1963279247283936, "learning_rate": 9.928182374911396e-05, "loss": 4.009, "step": 17700 }, { "epoch": 0.2862857050692731, "grad_norm": 2.6228134632110596, "learning_rate": 9.927752493661236e-05, "loss": 4.0115, "step": 17750 }, { "epoch": 0.287092143675102, "grad_norm": 2.0704829692840576, "learning_rate": 9.927321339032267e-05, "loss": 4.5117, "step": 17800 }, { "epoch": 0.28789858228093096, "grad_norm": 2.092968225479126, "learning_rate": 9.926888911135908e-05, "loss": 4.3974, "step": 17850 }, { "epoch": 0.28870502088675987, "grad_norm": 2.2499780654907227, "learning_rate": 9.9264552100839e-05, "loss": 3.8743, "step": 17900 }, { "epoch": 0.28951145949258883, "grad_norm": 1.970129132270813, "learning_rate": 9.926020235988313e-05, "loss": 4.0165, "step": 17950 }, { "epoch": 0.2903178980984178, "grad_norm": 1.648297905921936, "learning_rate": 9.925583988961555e-05, "loss": 4.1248, "step": 18000 }, { "epoch": 0.2903178980984178, "eval_asr_loss": 0.5475505368937239, "eval_loss": 5.026027679443359, "eval_runtime": 17.7593, "eval_samples_per_second": 28.83, "eval_steps_per_second": 9.629, "eval_tts_loss": 5.953106430704743, "step": 18000 }, { "epoch": 0.2911243367042467, "grad_norm": 2.2440879344940186, "learning_rate": 9.925155231986144e-05, "loss": 4.0456, "step": 18050 }, { "epoch": 0.29193077531007566, "grad_norm": 4.98739767074585, "learning_rate": 9.924716464888553e-05, "loss": 3.9588, "step": 18100 }, { "epoch": 0.29273721391590457, "grad_norm": 1.8671597242355347, "learning_rate": 9.924276425196694e-05, "loss": 4.111, "step": 18150 }, { "epoch": 0.29354365252173353, "grad_norm": 2.135396718978882, "learning_rate": 9.923835113024276e-05, "loss": 3.8483, "step": 18200 }, { "epoch": 0.29435009112756244, "grad_norm": 2.5975193977355957, "learning_rate": 9.923392528485342e-05, "loss": 4.1396, "step": 18250 }, { "epoch": 0.2951565297333914, "grad_norm": 2.3712480068206787, "learning_rate": 9.922948671694258e-05, "loss": 4.3892, "step": 18300 }, { "epoch": 0.2959629683392203, "grad_norm": 1.9007618427276611, "learning_rate": 9.922503542765723e-05, "loss": 4.0316, "step": 18350 }, { "epoch": 0.2967694069450493, "grad_norm": 2.405129909515381, "learning_rate": 9.922057141814758e-05, "loss": 4.2691, "step": 18400 }, { "epoch": 0.29757584555087824, "grad_norm": 1.4283597469329834, "learning_rate": 9.92160946895672e-05, "loss": 3.8773, "step": 18450 }, { "epoch": 0.29838228415670714, "grad_norm": 1.7647777795791626, "learning_rate": 9.921160524307294e-05, "loss": 4.2706, "step": 18500 }, { "epoch": 0.2991887227625361, "grad_norm": 2.2492105960845947, "learning_rate": 9.920710307982488e-05, "loss": 4.2848, "step": 18550 }, { "epoch": 0.299995161368365, "grad_norm": 1.8520779609680176, "learning_rate": 9.920258820098644e-05, "loss": 3.8957, "step": 18600 }, { "epoch": 0.300801599974194, "grad_norm": 1.574752688407898, "learning_rate": 9.91980606077243e-05, "loss": 3.96, "step": 18650 }, { "epoch": 0.3016080385800229, "grad_norm": 3.233193874359131, "learning_rate": 9.919352030120846e-05, "loss": 3.9196, "step": 18700 }, { "epoch": 0.30241447718585185, "grad_norm": 1.642980694770813, "learning_rate": 9.918896728261216e-05, "loss": 3.9637, "step": 18750 }, { "epoch": 0.30322091579168076, "grad_norm": 3.266005516052246, "learning_rate": 9.918440155311193e-05, "loss": 3.905, "step": 18800 }, { "epoch": 0.3040273543975097, "grad_norm": 2.4287779331207275, "learning_rate": 9.917982311388763e-05, "loss": 3.7187, "step": 18850 }, { "epoch": 0.3048337930033387, "grad_norm": 2.907322645187378, "learning_rate": 9.917523196612233e-05, "loss": 4.291, "step": 18900 }, { "epoch": 0.3056402316091676, "grad_norm": 2.1453018188476562, "learning_rate": 9.917062811100247e-05, "loss": 3.9087, "step": 18950 }, { "epoch": 0.30644667021499655, "grad_norm": 2.715480327606201, "learning_rate": 9.916601154971772e-05, "loss": 4.1218, "step": 19000 }, { "epoch": 0.30725310882082546, "grad_norm": 2.193371295928955, "learning_rate": 9.916138228346101e-05, "loss": 4.3158, "step": 19050 }, { "epoch": 0.3080595474266544, "grad_norm": 1.1654584407806396, "learning_rate": 9.915674031342864e-05, "loss": 3.9391, "step": 19100 }, { "epoch": 0.30886598603248333, "grad_norm": 2.589933156967163, "learning_rate": 9.915208564082011e-05, "loss": 3.7376, "step": 19150 }, { "epoch": 0.3096724246383123, "grad_norm": 1.8773146867752075, "learning_rate": 9.914741826683822e-05, "loss": 3.8819, "step": 19200 }, { "epoch": 0.3104788632441412, "grad_norm": 3.009824752807617, "learning_rate": 9.91427381926891e-05, "loss": 3.7509, "step": 19250 }, { "epoch": 0.31128530184997016, "grad_norm": 1.649009108543396, "learning_rate": 9.913804541958209e-05, "loss": 4.3104, "step": 19300 }, { "epoch": 0.3120917404557991, "grad_norm": 2.454561710357666, "learning_rate": 9.913333994872989e-05, "loss": 3.6798, "step": 19350 }, { "epoch": 0.31289817906162803, "grad_norm": 2.0216288566589355, "learning_rate": 9.912862178134839e-05, "loss": 4.0289, "step": 19400 }, { "epoch": 0.313704617667457, "grad_norm": 2.3107807636260986, "learning_rate": 9.912389091865683e-05, "loss": 3.766, "step": 19450 }, { "epoch": 0.3145110562732859, "grad_norm": 2.5804734230041504, "learning_rate": 9.911914736187774e-05, "loss": 4.397, "step": 19500 }, { "epoch": 0.31531749487911487, "grad_norm": 2.3121676445007324, "learning_rate": 9.911439111223686e-05, "loss": 4.0335, "step": 19550 }, { "epoch": 0.3161239334849438, "grad_norm": 5.4831342697143555, "learning_rate": 9.910962217096327e-05, "loss": 3.9674, "step": 19600 }, { "epoch": 0.31693037209077274, "grad_norm": 1.5503877401351929, "learning_rate": 9.91048405392893e-05, "loss": 4.2763, "step": 19650 }, { "epoch": 0.31773681069660165, "grad_norm": 3.0469183921813965, "learning_rate": 9.91000462184506e-05, "loss": 4.4284, "step": 19700 }, { "epoch": 0.3185432493024306, "grad_norm": 2.0985143184661865, "learning_rate": 9.909523920968602e-05, "loss": 4.1063, "step": 19750 }, { "epoch": 0.31934968790825957, "grad_norm": 2.2062013149261475, "learning_rate": 9.909041951423776e-05, "loss": 4.33, "step": 19800 }, { "epoch": 0.3201561265140885, "grad_norm": 1.0260072946548462, "learning_rate": 9.90855871333513e-05, "loss": 3.8472, "step": 19850 }, { "epoch": 0.32096256511991744, "grad_norm": 1.4739959239959717, "learning_rate": 9.908074206827534e-05, "loss": 3.6729, "step": 19900 }, { "epoch": 0.32176900372574635, "grad_norm": 1.3404515981674194, "learning_rate": 9.90758843202619e-05, "loss": 3.7977, "step": 19950 }, { "epoch": 0.3225754423315753, "grad_norm": 2.4651851654052734, "learning_rate": 9.907101389056628e-05, "loss": 3.7939, "step": 20000 }, { "epoch": 0.3233818809374042, "grad_norm": 4.615785121917725, "learning_rate": 9.906613078044704e-05, "loss": 4.1735, "step": 20050 }, { "epoch": 0.3241883195432332, "grad_norm": 2.872647523880005, "learning_rate": 9.906133303119925e-05, "loss": 4.0749, "step": 20100 }, { "epoch": 0.3249947581490621, "grad_norm": 1.8358525037765503, "learning_rate": 9.905642481756708e-05, "loss": 4.1521, "step": 20150 }, { "epoch": 0.32580119675489105, "grad_norm": 1.2282460927963257, "learning_rate": 9.905150392728124e-05, "loss": 4.2458, "step": 20200 }, { "epoch": 0.32660763536071996, "grad_norm": 5.087796211242676, "learning_rate": 9.904657036161335e-05, "loss": 4.0652, "step": 20250 }, { "epoch": 0.3274140739665489, "grad_norm": 1.9742780923843384, "learning_rate": 9.904162412183829e-05, "loss": 3.8384, "step": 20300 }, { "epoch": 0.3282205125723779, "grad_norm": 2.7985477447509766, "learning_rate": 9.90366652092342e-05, "loss": 3.8642, "step": 20350 }, { "epoch": 0.3290269511782068, "grad_norm": 1.4354287385940552, "learning_rate": 9.903169362508254e-05, "loss": 3.9342, "step": 20400 }, { "epoch": 0.32983338978403576, "grad_norm": 2.1292335987091064, "learning_rate": 9.902670937066798e-05, "loss": 3.6908, "step": 20450 }, { "epoch": 0.33063982838986467, "grad_norm": 2.5627033710479736, "learning_rate": 9.902171244727851e-05, "loss": 3.8676, "step": 20500 }, { "epoch": 0.33144626699569363, "grad_norm": 1.1219414472579956, "learning_rate": 9.90167028562054e-05, "loss": 4.0734, "step": 20550 }, { "epoch": 0.33225270560152254, "grad_norm": 1.570227026939392, "learning_rate": 9.901168059874315e-05, "loss": 3.7047, "step": 20600 }, { "epoch": 0.3330591442073515, "grad_norm": 2.0881588459014893, "learning_rate": 9.900664567618958e-05, "loss": 4.1496, "step": 20650 }, { "epoch": 0.3338655828131804, "grad_norm": 2.787027359008789, "learning_rate": 9.900159808984575e-05, "loss": 4.0835, "step": 20700 }, { "epoch": 0.33467202141900937, "grad_norm": 4.925529956817627, "learning_rate": 9.899653784101603e-05, "loss": 3.961, "step": 20750 }, { "epoch": 0.33547846002483833, "grad_norm": 2.3205068111419678, "learning_rate": 9.899146493100799e-05, "loss": 3.7549, "step": 20800 }, { "epoch": 0.33628489863066724, "grad_norm": 1.6399608850479126, "learning_rate": 9.898637936113257e-05, "loss": 4.1994, "step": 20850 }, { "epoch": 0.3370913372364962, "grad_norm": 2.315507411956787, "learning_rate": 9.89812811327039e-05, "loss": 4.0396, "step": 20900 }, { "epoch": 0.3378977758423251, "grad_norm": 2.239157199859619, "learning_rate": 9.897617024703943e-05, "loss": 4.3633, "step": 20950 }, { "epoch": 0.3387042144481541, "grad_norm": 2.6427533626556396, "learning_rate": 9.897104670545985e-05, "loss": 3.872, "step": 21000 }, { "epoch": 0.3387042144481541, "eval_asr_loss": 0.5612188019541791, "eval_loss": 4.999529838562012, "eval_runtime": 15.2573, "eval_samples_per_second": 33.558, "eval_steps_per_second": 11.208, "eval_tts_loss": 6.0108543639534995, "step": 21000 }, { "epoch": 0.339510653053983, "grad_norm": 1.9391437768936157, "learning_rate": 9.896591050928914e-05, "loss": 3.9338, "step": 21050 }, { "epoch": 0.34031709165981194, "grad_norm": 2.6968955993652344, "learning_rate": 9.896076165985453e-05, "loss": 4.0754, "step": 21100 }, { "epoch": 0.34112353026564085, "grad_norm": 1.984168529510498, "learning_rate": 9.895560015848652e-05, "loss": 3.9749, "step": 21150 }, { "epoch": 0.3419299688714698, "grad_norm": 3.1954638957977295, "learning_rate": 9.895042600651894e-05, "loss": 4.1341, "step": 21200 }, { "epoch": 0.3427364074772988, "grad_norm": 3.572061061859131, "learning_rate": 9.894523920528879e-05, "loss": 3.8629, "step": 21250 }, { "epoch": 0.3435428460831277, "grad_norm": 1.6200737953186035, "learning_rate": 9.894003975613644e-05, "loss": 3.9207, "step": 21300 }, { "epoch": 0.34434928468895665, "grad_norm": 1.8774925470352173, "learning_rate": 9.893482766040543e-05, "loss": 3.6651, "step": 21350 }, { "epoch": 0.34515572329478555, "grad_norm": 3.098095655441284, "learning_rate": 9.892960291944263e-05, "loss": 4.1371, "step": 21400 }, { "epoch": 0.3459621619006145, "grad_norm": 1.517335295677185, "learning_rate": 9.892436553459819e-05, "loss": 4.2014, "step": 21450 }, { "epoch": 0.3467686005064434, "grad_norm": 2.045222520828247, "learning_rate": 9.891911550722545e-05, "loss": 4.12, "step": 21500 }, { "epoch": 0.3475750391122724, "grad_norm": 2.005338430404663, "learning_rate": 9.891385283868112e-05, "loss": 4.0222, "step": 21550 }, { "epoch": 0.3483814777181013, "grad_norm": 2.2225544452667236, "learning_rate": 9.89085775303251e-05, "loss": 4.4284, "step": 21600 }, { "epoch": 0.34918791632393026, "grad_norm": 1.5066674947738647, "learning_rate": 9.890328958352057e-05, "loss": 4.042, "step": 21650 }, { "epoch": 0.3499943549297592, "grad_norm": 3.724332332611084, "learning_rate": 9.8897988999634e-05, "loss": 4.2711, "step": 21700 }, { "epoch": 0.35080079353558813, "grad_norm": 1.9594568014144897, "learning_rate": 9.889267578003514e-05, "loss": 3.8673, "step": 21750 }, { "epoch": 0.3516072321414171, "grad_norm": 1.407718300819397, "learning_rate": 9.888734992609692e-05, "loss": 3.965, "step": 21800 }, { "epoch": 0.352413670747246, "grad_norm": 3.1482791900634766, "learning_rate": 9.888201143919563e-05, "loss": 4.3013, "step": 21850 }, { "epoch": 0.35322010935307496, "grad_norm": 1.2597465515136719, "learning_rate": 9.887666032071077e-05, "loss": 3.9662, "step": 21900 }, { "epoch": 0.35402654795890387, "grad_norm": 2.1143901348114014, "learning_rate": 9.887129657202513e-05, "loss": 3.8487, "step": 21950 }, { "epoch": 0.35483298656473283, "grad_norm": 2.351919651031494, "learning_rate": 9.886592019452475e-05, "loss": 4.0369, "step": 22000 }, { "epoch": 0.35563942517056174, "grad_norm": 2.052345037460327, "learning_rate": 9.886053118959895e-05, "loss": 4.1324, "step": 22050 }, { "epoch": 0.3564458637763907, "grad_norm": 3.6101315021514893, "learning_rate": 9.885512955864027e-05, "loss": 3.8461, "step": 22100 }, { "epoch": 0.35725230238221967, "grad_norm": 2.581754684448242, "learning_rate": 9.884971530304458e-05, "loss": 3.85, "step": 22150 }, { "epoch": 0.3580587409880486, "grad_norm": 1.2572892904281616, "learning_rate": 9.884428842421096e-05, "loss": 3.8305, "step": 22200 }, { "epoch": 0.35886517959387754, "grad_norm": 3.1739416122436523, "learning_rate": 9.883884892354177e-05, "loss": 3.953, "step": 22250 }, { "epoch": 0.35967161819970644, "grad_norm": 2.128600597381592, "learning_rate": 9.883339680244261e-05, "loss": 3.6528, "step": 22300 }, { "epoch": 0.3604780568055354, "grad_norm": 1.8549989461898804, "learning_rate": 9.882793206232239e-05, "loss": 3.9435, "step": 22350 }, { "epoch": 0.3612844954113643, "grad_norm": 1.9381146430969238, "learning_rate": 9.882245470459324e-05, "loss": 4.1962, "step": 22400 }, { "epoch": 0.3620909340171933, "grad_norm": 3.481395959854126, "learning_rate": 9.881696473067056e-05, "loss": 4.4088, "step": 22450 }, { "epoch": 0.3628973726230222, "grad_norm": 2.8906712532043457, "learning_rate": 9.881146214197302e-05, "loss": 4.0288, "step": 22500 }, { "epoch": 0.36370381122885115, "grad_norm": 1.037143588066101, "learning_rate": 9.880605736756518e-05, "loss": 3.7669, "step": 22550 }, { "epoch": 0.3645102498346801, "grad_norm": 1.2270286083221436, "learning_rate": 9.880052980581152e-05, "loss": 3.9334, "step": 22600 }, { "epoch": 0.365316688440509, "grad_norm": 0.97640460729599, "learning_rate": 9.87949896335299e-05, "loss": 3.548, "step": 22650 }, { "epoch": 0.366123127046338, "grad_norm": 2.765495538711548, "learning_rate": 9.8789436852152e-05, "loss": 4.5431, "step": 22700 }, { "epoch": 0.3669295656521669, "grad_norm": 1.3644137382507324, "learning_rate": 9.878387146311272e-05, "loss": 3.9941, "step": 22750 }, { "epoch": 0.36773600425799585, "grad_norm": 3.112457275390625, "learning_rate": 9.877829346785017e-05, "loss": 3.8966, "step": 22800 }, { "epoch": 0.36854244286382476, "grad_norm": 2.050264835357666, "learning_rate": 9.87727028678058e-05, "loss": 3.7119, "step": 22850 }, { "epoch": 0.3693488814696537, "grad_norm": 3.1221909523010254, "learning_rate": 9.876709966442425e-05, "loss": 4.1182, "step": 22900 }, { "epoch": 0.37015532007548263, "grad_norm": 1.9487212896347046, "learning_rate": 9.876148385915345e-05, "loss": 3.8674, "step": 22950 }, { "epoch": 0.3709617586813116, "grad_norm": 2.4063527584075928, "learning_rate": 9.875585545344458e-05, "loss": 3.9302, "step": 23000 }, { "epoch": 0.37176819728714056, "grad_norm": 1.6610281467437744, "learning_rate": 9.875021444875207e-05, "loss": 4.1362, "step": 23050 }, { "epoch": 0.37257463589296946, "grad_norm": 1.0215773582458496, "learning_rate": 9.87445608465336e-05, "loss": 3.9701, "step": 23100 }, { "epoch": 0.3733810744987984, "grad_norm": 2.547252893447876, "learning_rate": 9.873889464825014e-05, "loss": 4.3084, "step": 23150 }, { "epoch": 0.37418751310462733, "grad_norm": 3.370795249938965, "learning_rate": 9.873321585536587e-05, "loss": 3.9985, "step": 23200 }, { "epoch": 0.3749939517104563, "grad_norm": 1.989953637123108, "learning_rate": 9.872752446934826e-05, "loss": 4.0459, "step": 23250 }, { "epoch": 0.3758003903162852, "grad_norm": 3.332749366760254, "learning_rate": 9.8721820491668e-05, "loss": 4.1809, "step": 23300 }, { "epoch": 0.37660682892211417, "grad_norm": 2.8768374919891357, "learning_rate": 9.871610392379906e-05, "loss": 4.0144, "step": 23350 }, { "epoch": 0.3774132675279431, "grad_norm": 1.8673336505889893, "learning_rate": 9.871037476721868e-05, "loss": 3.769, "step": 23400 }, { "epoch": 0.37821970613377204, "grad_norm": 2.023162603378296, "learning_rate": 9.870463302340729e-05, "loss": 4.3442, "step": 23450 }, { "epoch": 0.379026144739601, "grad_norm": 1.667543649673462, "learning_rate": 9.869887869384864e-05, "loss": 3.688, "step": 23500 }, { "epoch": 0.3798325833454299, "grad_norm": 4.761668682098389, "learning_rate": 9.869311178002969e-05, "loss": 3.6925, "step": 23550 }, { "epoch": 0.38063902195125887, "grad_norm": 2.4622697830200195, "learning_rate": 9.868733228344068e-05, "loss": 4.3784, "step": 23600 }, { "epoch": 0.3814454605570878, "grad_norm": 2.6539041996002197, "learning_rate": 9.868154020557507e-05, "loss": 4.117, "step": 23650 }, { "epoch": 0.38225189916291674, "grad_norm": 1.5162415504455566, "learning_rate": 9.867573554792961e-05, "loss": 4.1188, "step": 23700 }, { "epoch": 0.38305833776874565, "grad_norm": 1.5139150619506836, "learning_rate": 9.866991831200427e-05, "loss": 4.0601, "step": 23750 }, { "epoch": 0.3838647763745746, "grad_norm": 1.5848387479782104, "learning_rate": 9.866408849930226e-05, "loss": 4.3303, "step": 23800 }, { "epoch": 0.3846712149804035, "grad_norm": 2.323932647705078, "learning_rate": 9.86582461113301e-05, "loss": 4.3198, "step": 23850 }, { "epoch": 0.3854776535862325, "grad_norm": 1.6189751625061035, "learning_rate": 9.865239114959748e-05, "loss": 3.7001, "step": 23900 }, { "epoch": 0.38628409219206145, "grad_norm": 1.6173287630081177, "learning_rate": 9.86465236156174e-05, "loss": 3.9669, "step": 23950 }, { "epoch": 0.38709053079789035, "grad_norm": 1.945816993713379, "learning_rate": 9.864064351090608e-05, "loss": 4.1111, "step": 24000 }, { "epoch": 0.38709053079789035, "eval_asr_loss": 0.5596013197241803, "eval_loss": 4.975149631500244, "eval_runtime": 15.3079, "eval_samples_per_second": 33.447, "eval_steps_per_second": 11.171, "eval_tts_loss": 5.989607999181007, "step": 24000 }, { "epoch": 0.3878969694037193, "grad_norm": 2.5832393169403076, "learning_rate": 9.863475083698302e-05, "loss": 4.0228, "step": 24050 }, { "epoch": 0.3887034080095482, "grad_norm": 1.5465142726898193, "learning_rate": 9.862884559537088e-05, "loss": 3.9781, "step": 24100 }, { "epoch": 0.3895098466153772, "grad_norm": 2.0018787384033203, "learning_rate": 9.862292778759572e-05, "loss": 3.932, "step": 24150 }, { "epoch": 0.3903162852212061, "grad_norm": 1.9746758937835693, "learning_rate": 9.861699741518669e-05, "loss": 3.8956, "step": 24200 }, { "epoch": 0.39112272382703506, "grad_norm": 2.558194875717163, "learning_rate": 9.861105447967629e-05, "loss": 3.9698, "step": 24250 }, { "epoch": 0.39192916243286396, "grad_norm": 4.342727184295654, "learning_rate": 9.86050989826002e-05, "loss": 4.1687, "step": 24300 }, { "epoch": 0.3927356010386929, "grad_norm": 1.6204520463943481, "learning_rate": 9.859913092549742e-05, "loss": 3.832, "step": 24350 }, { "epoch": 0.39354203964452183, "grad_norm": 2.4823992252349854, "learning_rate": 9.859315030991012e-05, "loss": 4.3509, "step": 24400 }, { "epoch": 0.3943484782503508, "grad_norm": 3.508967876434326, "learning_rate": 9.858715713738376e-05, "loss": 4.1016, "step": 24450 }, { "epoch": 0.39515491685617976, "grad_norm": 1.73115873336792, "learning_rate": 9.858115140946705e-05, "loss": 3.9989, "step": 24500 }, { "epoch": 0.39596135546200867, "grad_norm": 1.9791126251220703, "learning_rate": 9.85751331277119e-05, "loss": 4.2463, "step": 24550 }, { "epoch": 0.39676779406783763, "grad_norm": 1.3955203294754028, "learning_rate": 9.856910229367349e-05, "loss": 3.9978, "step": 24600 }, { "epoch": 0.39757423267366654, "grad_norm": 2.9622445106506348, "learning_rate": 9.856305890891027e-05, "loss": 3.7169, "step": 24650 }, { "epoch": 0.3983806712794955, "grad_norm": 1.4486901760101318, "learning_rate": 9.855700297498392e-05, "loss": 3.9363, "step": 24700 }, { "epoch": 0.3991871098853244, "grad_norm": 1.3110404014587402, "learning_rate": 9.85509344934593e-05, "loss": 3.9389, "step": 24750 }, { "epoch": 0.3999935484911534, "grad_norm": 5.3704752922058105, "learning_rate": 9.854485346590461e-05, "loss": 4.0937, "step": 24800 }, { "epoch": 0.4007999870969823, "grad_norm": 1.8054653406143188, "learning_rate": 9.853875989389123e-05, "loss": 4.0569, "step": 24850 }, { "epoch": 0.40160642570281124, "grad_norm": 2.4662413597106934, "learning_rate": 9.85326537789938e-05, "loss": 4.0274, "step": 24900 }, { "epoch": 0.4024128643086402, "grad_norm": 2.1582422256469727, "learning_rate": 9.852665761880885e-05, "loss": 3.9833, "step": 24950 }, { "epoch": 0.4032193029144691, "grad_norm": 1.7549717426300049, "learning_rate": 9.852052667365919e-05, "loss": 4.0434, "step": 25000 }, { "epoch": 0.4040257415202981, "grad_norm": 2.722822427749634, "learning_rate": 9.85143831903371e-05, "loss": 4.2247, "step": 25050 }, { "epoch": 0.404832180126127, "grad_norm": 2.266624689102173, "learning_rate": 9.850822717043014e-05, "loss": 4.0886, "step": 25100 }, { "epoch": 0.40563861873195595, "grad_norm": 2.5876810550689697, "learning_rate": 9.85020586155291e-05, "loss": 3.9883, "step": 25150 }, { "epoch": 0.40644505733778485, "grad_norm": 2.048994779586792, "learning_rate": 9.849587752722795e-05, "loss": 3.7832, "step": 25200 }, { "epoch": 0.4072514959436138, "grad_norm": 2.542686939239502, "learning_rate": 9.848968390712398e-05, "loss": 4.1333, "step": 25250 }, { "epoch": 0.4080579345494427, "grad_norm": 1.8834608793258667, "learning_rate": 9.848347775681765e-05, "loss": 4.3341, "step": 25300 }, { "epoch": 0.4088643731552717, "grad_norm": 2.2536680698394775, "learning_rate": 9.847725907791271e-05, "loss": 3.9732, "step": 25350 }, { "epoch": 0.40967081176110065, "grad_norm": 2.5030934810638428, "learning_rate": 9.847102787201614e-05, "loss": 4.0837, "step": 25400 }, { "epoch": 0.41047725036692956, "grad_norm": 2.3016812801361084, "learning_rate": 9.846478414073812e-05, "loss": 3.8695, "step": 25450 }, { "epoch": 0.4112836889727585, "grad_norm": 1.6216686964035034, "learning_rate": 9.84585278856921e-05, "loss": 3.8525, "step": 25500 }, { "epoch": 0.41209012757858743, "grad_norm": 4.122720718383789, "learning_rate": 9.845225910849476e-05, "loss": 4.1526, "step": 25550 }, { "epoch": 0.4128965661844164, "grad_norm": 2.8652255535125732, "learning_rate": 9.844597781076599e-05, "loss": 3.8846, "step": 25600 }, { "epoch": 0.4137030047902453, "grad_norm": 1.516522765159607, "learning_rate": 9.843968399412898e-05, "loss": 3.9619, "step": 25650 }, { "epoch": 0.41450944339607426, "grad_norm": 1.89898681640625, "learning_rate": 9.843337766021005e-05, "loss": 4.0633, "step": 25700 }, { "epoch": 0.41531588200190317, "grad_norm": 3.1514346599578857, "learning_rate": 9.842705881063889e-05, "loss": 3.8735, "step": 25750 }, { "epoch": 0.41612232060773213, "grad_norm": 1.562589168548584, "learning_rate": 9.84207274470483e-05, "loss": 3.6907, "step": 25800 }, { "epoch": 0.4169287592135611, "grad_norm": 1.0612993240356445, "learning_rate": 9.841438357107438e-05, "loss": 4.2971, "step": 25850 }, { "epoch": 0.41773519781939, "grad_norm": 2.6953563690185547, "learning_rate": 9.840802718435646e-05, "loss": 3.9486, "step": 25900 }, { "epoch": 0.41854163642521897, "grad_norm": 2.3360955715179443, "learning_rate": 9.840165828853709e-05, "loss": 4.0716, "step": 25950 }, { "epoch": 0.4193480750310479, "grad_norm": 2.1841683387756348, "learning_rate": 9.839527688526204e-05, "loss": 3.9567, "step": 26000 }, { "epoch": 0.42015451363687684, "grad_norm": 2.321568489074707, "learning_rate": 9.838888297618032e-05, "loss": 4.0802, "step": 26050 }, { "epoch": 0.42096095224270574, "grad_norm": 2.530367374420166, "learning_rate": 9.838247656294421e-05, "loss": 3.7345, "step": 26100 }, { "epoch": 0.4217673908485347, "grad_norm": 1.4943206310272217, "learning_rate": 9.837605764720916e-05, "loss": 3.865, "step": 26150 }, { "epoch": 0.4225738294543636, "grad_norm": 1.2419519424438477, "learning_rate": 9.836962623063391e-05, "loss": 3.9362, "step": 26200 }, { "epoch": 0.4233802680601926, "grad_norm": 2.8908309936523438, "learning_rate": 9.836318231488035e-05, "loss": 3.5773, "step": 26250 }, { "epoch": 0.42418670666602154, "grad_norm": 5.162326335906982, "learning_rate": 9.835672590161369e-05, "loss": 3.9459, "step": 26300 }, { "epoch": 0.42499314527185045, "grad_norm": 1.782263994216919, "learning_rate": 9.835025699250233e-05, "loss": 3.9619, "step": 26350 }, { "epoch": 0.4257995838776794, "grad_norm": 5.839146137237549, "learning_rate": 9.834377558921788e-05, "loss": 3.8363, "step": 26400 }, { "epoch": 0.4266060224835083, "grad_norm": 1.6730350255966187, "learning_rate": 9.833728169343521e-05, "loss": 3.7422, "step": 26450 }, { "epoch": 0.4274124610893373, "grad_norm": 0.9565515518188477, "learning_rate": 9.83307753068324e-05, "loss": 4.3006, "step": 26500 }, { "epoch": 0.4282188996951662, "grad_norm": 1.102502465248108, "learning_rate": 9.832425643109076e-05, "loss": 3.7061, "step": 26550 }, { "epoch": 0.42902533830099515, "grad_norm": 2.7195091247558594, "learning_rate": 9.831772506789483e-05, "loss": 3.8085, "step": 26600 }, { "epoch": 0.42983177690682406, "grad_norm": 2.3110263347625732, "learning_rate": 9.831118121893239e-05, "loss": 3.7295, "step": 26650 }, { "epoch": 0.430638215512653, "grad_norm": 1.7428209781646729, "learning_rate": 9.830462488589443e-05, "loss": 4.0876, "step": 26700 }, { "epoch": 0.431444654118482, "grad_norm": 1.8392008543014526, "learning_rate": 9.829805607047515e-05, "loss": 4.3742, "step": 26750 }, { "epoch": 0.4322510927243109, "grad_norm": 2.320955276489258, "learning_rate": 9.829147477437202e-05, "loss": 4.0089, "step": 26800 }, { "epoch": 0.43305753133013986, "grad_norm": 2.5901613235473633, "learning_rate": 9.82848809992857e-05, "loss": 4.0296, "step": 26850 }, { "epoch": 0.43386396993596876, "grad_norm": 4.836414337158203, "learning_rate": 9.827827474692009e-05, "loss": 4.3018, "step": 26900 }, { "epoch": 0.4346704085417977, "grad_norm": 3.3830177783966064, "learning_rate": 9.82716560189823e-05, "loss": 3.949, "step": 26950 }, { "epoch": 0.43547684714762663, "grad_norm": 1.8795298337936401, "learning_rate": 9.826502481718267e-05, "loss": 3.7089, "step": 27000 }, { "epoch": 0.43547684714762663, "eval_asr_loss": 0.5559742133953548, "eval_loss": 4.958992958068848, "eval_runtime": 14.7232, "eval_samples_per_second": 34.775, "eval_steps_per_second": 11.614, "eval_tts_loss": 6.008725999135946, "step": 27000 }, { "epoch": 0.4362832857534556, "grad_norm": 1.627994418144226, "learning_rate": 9.825838114323478e-05, "loss": 3.9582, "step": 27050 }, { "epoch": 0.4370897243592845, "grad_norm": 4.4471611976623535, "learning_rate": 9.825172499885542e-05, "loss": 3.8203, "step": 27100 }, { "epoch": 0.43789616296511347, "grad_norm": 4.530590534210205, "learning_rate": 9.82450563857646e-05, "loss": 4.1167, "step": 27150 }, { "epoch": 0.43870260157094243, "grad_norm": 2.129225969314575, "learning_rate": 9.823837530568554e-05, "loss": 3.9683, "step": 27200 }, { "epoch": 0.43950904017677134, "grad_norm": 1.8322468996047974, "learning_rate": 9.823168176034471e-05, "loss": 4.0839, "step": 27250 }, { "epoch": 0.4403154787826003, "grad_norm": 2.424598217010498, "learning_rate": 9.822497575147178e-05, "loss": 4.1781, "step": 27300 }, { "epoch": 0.4411219173884292, "grad_norm": 1.3116930723190308, "learning_rate": 9.821839177232751e-05, "loss": 3.9697, "step": 27350 }, { "epoch": 0.44192835599425817, "grad_norm": 1.402601718902588, "learning_rate": 9.821166109077653e-05, "loss": 4.0815, "step": 27400 }, { "epoch": 0.4427347946000871, "grad_norm": 1.392431616783142, "learning_rate": 9.820491795086698e-05, "loss": 4.0143, "step": 27450 }, { "epoch": 0.44354123320591604, "grad_norm": 1.231276035308838, "learning_rate": 9.819816235434137e-05, "loss": 4.3486, "step": 27500 }, { "epoch": 0.44434767181174495, "grad_norm": 3.108103036880493, "learning_rate": 9.819139430294539e-05, "loss": 4.0492, "step": 27550 }, { "epoch": 0.4451541104175739, "grad_norm": 2.1228959560394287, "learning_rate": 9.818461379842798e-05, "loss": 4.3671, "step": 27600 }, { "epoch": 0.4459605490234029, "grad_norm": 1.8643229007720947, "learning_rate": 9.81778208425413e-05, "loss": 3.9776, "step": 27650 }, { "epoch": 0.4467669876292318, "grad_norm": 1.3176897764205933, "learning_rate": 9.81710154370407e-05, "loss": 4.368, "step": 27700 }, { "epoch": 0.44757342623506075, "grad_norm": 2.3601205348968506, "learning_rate": 9.816419758368476e-05, "loss": 3.6981, "step": 27750 }, { "epoch": 0.44837986484088965, "grad_norm": 4.233773708343506, "learning_rate": 9.815736728423531e-05, "loss": 4.4567, "step": 27800 }, { "epoch": 0.4491863034467186, "grad_norm": 2.077209949493408, "learning_rate": 9.81505245404573e-05, "loss": 3.854, "step": 27850 }, { "epoch": 0.4499927420525475, "grad_norm": 1.6529903411865234, "learning_rate": 9.814366935411903e-05, "loss": 4.254, "step": 27900 }, { "epoch": 0.4507991806583765, "grad_norm": 2.4949686527252197, "learning_rate": 9.813680172699191e-05, "loss": 4.1666, "step": 27950 }, { "epoch": 0.4516056192642054, "grad_norm": 1.0352684259414673, "learning_rate": 9.81299216608506e-05, "loss": 3.9291, "step": 28000 }, { "epoch": 0.45241205787003436, "grad_norm": 1.945921778678894, "learning_rate": 9.8123029157473e-05, "loss": 4.0353, "step": 28050 }, { "epoch": 0.4532184964758633, "grad_norm": 1.8263922929763794, "learning_rate": 9.811612421864018e-05, "loss": 3.984, "step": 28100 }, { "epoch": 0.4540249350816922, "grad_norm": 2.8539750576019287, "learning_rate": 9.810920684613644e-05, "loss": 4.1103, "step": 28150 }, { "epoch": 0.4548313736875212, "grad_norm": 2.559903860092163, "learning_rate": 9.81022770417493e-05, "loss": 4.0392, "step": 28200 }, { "epoch": 0.4556378122933501, "grad_norm": 1.7837358713150024, "learning_rate": 9.809533480726952e-05, "loss": 3.9311, "step": 28250 }, { "epoch": 0.45644425089917906, "grad_norm": 1.3091505765914917, "learning_rate": 9.808838014449097e-05, "loss": 3.7175, "step": 28300 }, { "epoch": 0.45725068950500797, "grad_norm": 1.9200706481933594, "learning_rate": 9.808141305521087e-05, "loss": 4.3186, "step": 28350 }, { "epoch": 0.45805712811083693, "grad_norm": 2.3566806316375732, "learning_rate": 9.807443354122955e-05, "loss": 4.1294, "step": 28400 }, { "epoch": 0.45886356671666584, "grad_norm": 1.5939757823944092, "learning_rate": 9.806744160435059e-05, "loss": 3.8491, "step": 28450 }, { "epoch": 0.4596700053224948, "grad_norm": 1.4164952039718628, "learning_rate": 9.806043724638076e-05, "loss": 4.2949, "step": 28500 }, { "epoch": 0.46047644392832376, "grad_norm": 2.1956284046173096, "learning_rate": 9.805342046913009e-05, "loss": 4.1996, "step": 28550 }, { "epoch": 0.46128288253415267, "grad_norm": 1.7804524898529053, "learning_rate": 9.804639127441176e-05, "loss": 3.9101, "step": 28600 }, { "epoch": 0.46208932113998163, "grad_norm": 0.5685811638832092, "learning_rate": 9.803934966404217e-05, "loss": 4.1905, "step": 28650 }, { "epoch": 0.46289575974581054, "grad_norm": 1.6610820293426514, "learning_rate": 9.803229563984095e-05, "loss": 4.3581, "step": 28700 }, { "epoch": 0.4637021983516395, "grad_norm": 2.0791728496551514, "learning_rate": 9.802522920363095e-05, "loss": 3.9016, "step": 28750 }, { "epoch": 0.4645086369574684, "grad_norm": 2.6067445278167725, "learning_rate": 9.801815035723817e-05, "loss": 3.9681, "step": 28800 }, { "epoch": 0.4653150755632974, "grad_norm": 0.8365187644958496, "learning_rate": 9.801105910249187e-05, "loss": 3.9012, "step": 28850 }, { "epoch": 0.4661215141691263, "grad_norm": 3.1129026412963867, "learning_rate": 9.80039554412245e-05, "loss": 3.9982, "step": 28900 }, { "epoch": 0.46692795277495525, "grad_norm": 1.3477463722229004, "learning_rate": 9.799683937527171e-05, "loss": 4.1739, "step": 28950 }, { "epoch": 0.46773439138078415, "grad_norm": 3.441631317138672, "learning_rate": 9.798971090647238e-05, "loss": 4.0417, "step": 29000 }, { "epoch": 0.4685408299866131, "grad_norm": 2.399874210357666, "learning_rate": 9.798257003666853e-05, "loss": 4.1551, "step": 29050 }, { "epoch": 0.4693472685924421, "grad_norm": 2.2936065196990967, "learning_rate": 9.797541676770547e-05, "loss": 4.1262, "step": 29100 }, { "epoch": 0.470153707198271, "grad_norm": 2.890214204788208, "learning_rate": 9.796825110143165e-05, "loss": 4.1734, "step": 29150 }, { "epoch": 0.47096014580409995, "grad_norm": 1.6101442575454712, "learning_rate": 9.796107303969877e-05, "loss": 3.6554, "step": 29200 }, { "epoch": 0.47176658440992886, "grad_norm": 1.5904902219772339, "learning_rate": 9.795388258436169e-05, "loss": 4.1421, "step": 29250 }, { "epoch": 0.4725730230157578, "grad_norm": 1.2808867692947388, "learning_rate": 9.79466797372785e-05, "loss": 3.7509, "step": 29300 }, { "epoch": 0.47337946162158673, "grad_norm": 2.6117844581604004, "learning_rate": 9.793946450031049e-05, "loss": 4.252, "step": 29350 }, { "epoch": 0.4741859002274157, "grad_norm": 2.114128351211548, "learning_rate": 9.793238154921244e-05, "loss": 4.0362, "step": 29400 }, { "epoch": 0.4749923388332446, "grad_norm": 0.834858775138855, "learning_rate": 9.792514178577616e-05, "loss": 3.8481, "step": 29450 }, { "epoch": 0.47579877743907356, "grad_norm": 1.2032753229141235, "learning_rate": 9.791788963802069e-05, "loss": 4.0954, "step": 29500 }, { "epoch": 0.4766052160449025, "grad_norm": 1.481992244720459, "learning_rate": 9.791062510782003e-05, "loss": 4.0293, "step": 29550 }, { "epoch": 0.47741165465073143, "grad_norm": 1.40059494972229, "learning_rate": 9.790334819705142e-05, "loss": 3.8551, "step": 29600 }, { "epoch": 0.4782180932565604, "grad_norm": 1.1095318794250488, "learning_rate": 9.789605890759526e-05, "loss": 4.0393, "step": 29650 }, { "epoch": 0.4790245318623893, "grad_norm": 2.459015369415283, "learning_rate": 9.78887572413352e-05, "loss": 4.1729, "step": 29700 }, { "epoch": 0.47983097046821827, "grad_norm": 1.1588975191116333, "learning_rate": 9.788144320015804e-05, "loss": 3.971, "step": 29750 }, { "epoch": 0.4806374090740472, "grad_norm": 2.8166420459747314, "learning_rate": 9.78741167859538e-05, "loss": 4.4002, "step": 29800 }, { "epoch": 0.48144384767987614, "grad_norm": 3.607013702392578, "learning_rate": 9.786677800061571e-05, "loss": 3.819, "step": 29850 }, { "epoch": 0.48225028628570504, "grad_norm": 2.313955783843994, "learning_rate": 9.785942684604018e-05, "loss": 4.2892, "step": 29900 }, { "epoch": 0.483056724891534, "grad_norm": 1.3837124109268188, "learning_rate": 9.785206332412682e-05, "loss": 4.0314, "step": 29950 }, { "epoch": 0.48386316349736297, "grad_norm": 1.5818651914596558, "learning_rate": 9.784468743677842e-05, "loss": 3.8618, "step": 30000 }, { "epoch": 0.48386316349736297, "eval_asr_loss": 0.5646996813599123, "eval_loss": 4.947766304016113, "eval_runtime": 14.9459, "eval_samples_per_second": 34.257, "eval_steps_per_second": 11.441, "eval_tts_loss": 5.988222905527309, "step": 30000 }, { "epoch": 0.4846696021031919, "grad_norm": 2.607830047607422, "learning_rate": 9.7837299185901e-05, "loss": 4.127, "step": 30050 }, { "epoch": 0.48547604070902084, "grad_norm": 1.471893310546875, "learning_rate": 9.782989857340373e-05, "loss": 3.7895, "step": 30100 }, { "epoch": 0.48628247931484975, "grad_norm": 2.5169291496276855, "learning_rate": 9.782248560119902e-05, "loss": 4.0763, "step": 30150 }, { "epoch": 0.4870889179206787, "grad_norm": 4.164167881011963, "learning_rate": 9.781506027120247e-05, "loss": 4.0099, "step": 30200 }, { "epoch": 0.4878953565265076, "grad_norm": 2.2432568073272705, "learning_rate": 9.780762258533282e-05, "loss": 4.1052, "step": 30250 }, { "epoch": 0.4887017951323366, "grad_norm": 1.442578911781311, "learning_rate": 9.780017254551206e-05, "loss": 3.9327, "step": 30300 }, { "epoch": 0.4895082337381655, "grad_norm": 1.6483718156814575, "learning_rate": 9.779271015366535e-05, "loss": 3.7249, "step": 30350 }, { "epoch": 0.49031467234399445, "grad_norm": 2.119659662246704, "learning_rate": 9.778523541172104e-05, "loss": 3.959, "step": 30400 }, { "epoch": 0.4911211109498234, "grad_norm": 1.574180006980896, "learning_rate": 9.777774832161069e-05, "loss": 4.2155, "step": 30450 }, { "epoch": 0.4919275495556523, "grad_norm": 1.1062897443771362, "learning_rate": 9.7770248885269e-05, "loss": 4.263, "step": 30500 }, { "epoch": 0.4927339881614813, "grad_norm": 1.0894304513931274, "learning_rate": 9.776273710463394e-05, "loss": 3.9637, "step": 30550 }, { "epoch": 0.4935404267673102, "grad_norm": 1.5745190382003784, "learning_rate": 9.77552129816466e-05, "loss": 3.9539, "step": 30600 }, { "epoch": 0.49434686537313915, "grad_norm": 1.2136081457138062, "learning_rate": 9.77476765182513e-05, "loss": 3.8683, "step": 30650 }, { "epoch": 0.49515330397896806, "grad_norm": 2.5908925533294678, "learning_rate": 9.774012771639553e-05, "loss": 4.1582, "step": 30700 }, { "epoch": 0.495959742584797, "grad_norm": 1.8017607927322388, "learning_rate": 9.773256657802997e-05, "loss": 3.6259, "step": 30750 }, { "epoch": 0.49676618119062593, "grad_norm": 1.5173224210739136, "learning_rate": 9.77249931051085e-05, "loss": 4.0439, "step": 30800 }, { "epoch": 0.4975726197964549, "grad_norm": 15.362312316894531, "learning_rate": 9.771786279566293e-05, "loss": 4.387, "step": 30850 }, { "epoch": 0.49837905840228386, "grad_norm": 3.3392670154571533, "learning_rate": 9.771026539928696e-05, "loss": 4.5741, "step": 30900 }, { "epoch": 0.49918549700811277, "grad_norm": 1.549731969833374, "learning_rate": 9.770265567411791e-05, "loss": 3.9523, "step": 30950 }, { "epoch": 0.49999193561394173, "grad_norm": 2.7253308296203613, "learning_rate": 9.769503362212222e-05, "loss": 4.3312, "step": 31000 }, { "epoch": 0.5007983742197707, "grad_norm": 2.1496167182922363, "learning_rate": 9.768739924526948e-05, "loss": 3.9847, "step": 31050 }, { "epoch": 0.5016048128255995, "grad_norm": 2.93792462348938, "learning_rate": 9.76797525455325e-05, "loss": 3.9403, "step": 31100 }, { "epoch": 0.5024112514314285, "grad_norm": 2.272951126098633, "learning_rate": 9.767209352488727e-05, "loss": 3.7558, "step": 31150 }, { "epoch": 0.5032176900372575, "grad_norm": 3.371812105178833, "learning_rate": 9.766442218531293e-05, "loss": 3.9215, "step": 31200 }, { "epoch": 0.5040241286430864, "grad_norm": 1.3443658351898193, "learning_rate": 9.765673852879186e-05, "loss": 4.1457, "step": 31250 }, { "epoch": 0.5048305672489153, "grad_norm": 2.311265230178833, "learning_rate": 9.764904255730955e-05, "loss": 4.3678, "step": 31300 }, { "epoch": 0.5056370058547442, "grad_norm": 2.241488456726074, "learning_rate": 9.764133427285475e-05, "loss": 4.145, "step": 31350 }, { "epoch": 0.5064434444605732, "grad_norm": 2.38242769241333, "learning_rate": 9.763361367741934e-05, "loss": 4.1718, "step": 31400 }, { "epoch": 0.5072498830664022, "grad_norm": 1.874671220779419, "learning_rate": 9.76258807729984e-05, "loss": 3.9391, "step": 31450 }, { "epoch": 0.5080563216722311, "grad_norm": 1.9810266494750977, "learning_rate": 9.761813556159019e-05, "loss": 3.9962, "step": 31500 }, { "epoch": 0.50886276027806, "grad_norm": 2.93658185005188, "learning_rate": 9.761037804519612e-05, "loss": 3.7464, "step": 31550 }, { "epoch": 0.509669198883889, "grad_norm": 2.2134599685668945, "learning_rate": 9.760260822582086e-05, "loss": 3.9865, "step": 31600 }, { "epoch": 0.5104756374897179, "grad_norm": 2.7154464721679688, "learning_rate": 9.759482610547214e-05, "loss": 3.8921, "step": 31650 }, { "epoch": 0.5112820760955469, "grad_norm": 2.0518405437469482, "learning_rate": 9.7587031686161e-05, "loss": 4.1342, "step": 31700 }, { "epoch": 0.5120885147013757, "grad_norm": 2.0980758666992188, "learning_rate": 9.757922496990155e-05, "loss": 3.9435, "step": 31750 }, { "epoch": 0.5128949533072047, "grad_norm": 4.004748344421387, "learning_rate": 9.757140595871115e-05, "loss": 4.1433, "step": 31800 }, { "epoch": 0.5137013919130337, "grad_norm": 1.889359474182129, "learning_rate": 9.756357465461027e-05, "loss": 4.1632, "step": 31850 }, { "epoch": 0.5145078305188626, "grad_norm": 2.1634538173675537, "learning_rate": 9.755573105962264e-05, "loss": 3.9894, "step": 31900 }, { "epoch": 0.5153142691246916, "grad_norm": 2.0058374404907227, "learning_rate": 9.754787517577508e-05, "loss": 3.7636, "step": 31950 }, { "epoch": 0.5161207077305204, "grad_norm": 3.5357580184936523, "learning_rate": 9.754000700509764e-05, "loss": 3.9957, "step": 32000 }, { "epoch": 0.5169271463363494, "grad_norm": 1.03873872756958, "learning_rate": 9.753212654962356e-05, "loss": 4.1323, "step": 32050 }, { "epoch": 0.5177335849421784, "grad_norm": 2.3036513328552246, "learning_rate": 9.752423381138919e-05, "loss": 4.057, "step": 32100 }, { "epoch": 0.5185400235480073, "grad_norm": 1.3666496276855469, "learning_rate": 9.75163287924341e-05, "loss": 4.1997, "step": 32150 }, { "epoch": 0.5193464621538362, "grad_norm": 2.106926441192627, "learning_rate": 9.750841149480102e-05, "loss": 3.924, "step": 32200 }, { "epoch": 0.5201529007596651, "grad_norm": 3.9756150245666504, "learning_rate": 9.750048192053587e-05, "loss": 3.9417, "step": 32250 }, { "epoch": 0.5209593393654941, "grad_norm": 1.9011123180389404, "learning_rate": 9.749254007168771e-05, "loss": 3.9619, "step": 32300 }, { "epoch": 0.5217657779713231, "grad_norm": 1.8549953699111938, "learning_rate": 9.74845859503088e-05, "loss": 3.8087, "step": 32350 }, { "epoch": 0.522572216577152, "grad_norm": 2.133540153503418, "learning_rate": 9.747661955845456e-05, "loss": 4.1356, "step": 32400 }, { "epoch": 0.5233786551829809, "grad_norm": 1.6960670948028564, "learning_rate": 9.74686408981836e-05, "loss": 3.798, "step": 32450 }, { "epoch": 0.5241850937888098, "grad_norm": 1.8586905002593994, "learning_rate": 9.746064997155763e-05, "loss": 4.0743, "step": 32500 }, { "epoch": 0.5249915323946388, "grad_norm": 2.0501816272735596, "learning_rate": 9.745264678064166e-05, "loss": 3.6707, "step": 32550 }, { "epoch": 0.5257979710004678, "grad_norm": 2.2415525913238525, "learning_rate": 9.744463132750373e-05, "loss": 3.671, "step": 32600 }, { "epoch": 0.5266044096062966, "grad_norm": 2.3392422199249268, "learning_rate": 9.743660361421512e-05, "loss": 4.1384, "step": 32650 }, { "epoch": 0.5274108482121256, "grad_norm": 1.0827341079711914, "learning_rate": 9.74285636428503e-05, "loss": 4.0892, "step": 32700 }, { "epoch": 0.5282172868179545, "grad_norm": 1.8944828510284424, "learning_rate": 9.742051141548685e-05, "loss": 3.9703, "step": 32750 }, { "epoch": 0.5290237254237835, "grad_norm": 2.013651132583618, "learning_rate": 9.741244693420554e-05, "loss": 4.2219, "step": 32800 }, { "epoch": 0.5298301640296125, "grad_norm": 1.879599690437317, "learning_rate": 9.740437020109033e-05, "loss": 3.7386, "step": 32850 }, { "epoch": 0.5306366026354413, "grad_norm": 2.901381254196167, "learning_rate": 9.739628121822832e-05, "loss": 3.9859, "step": 32900 }, { "epoch": 0.5314430412412703, "grad_norm": 2.858835220336914, "learning_rate": 9.738817998770976e-05, "loss": 3.8376, "step": 32950 }, { "epoch": 0.5322494798470992, "grad_norm": 2.78635311126709, "learning_rate": 9.738006651162812e-05, "loss": 3.8029, "step": 33000 }, { "epoch": 0.5322494798470992, "eval_asr_loss": 0.5563808219787441, "eval_loss": 4.897815227508545, "eval_runtime": 14.787, "eval_samples_per_second": 34.625, "eval_steps_per_second": 11.564, "eval_tts_loss": 5.9989542354158925, "step": 33000 }, { "epoch": 0.5330559184529282, "grad_norm": 1.9606646299362183, "learning_rate": 9.737194079207997e-05, "loss": 4.1998, "step": 33050 }, { "epoch": 0.5338623570587571, "grad_norm": 3.6199183464050293, "learning_rate": 9.736380283116508e-05, "loss": 3.9197, "step": 33100 }, { "epoch": 0.534668795664586, "grad_norm": 2.4361817836761475, "learning_rate": 9.735565263098641e-05, "loss": 3.8554, "step": 33150 }, { "epoch": 0.535475234270415, "grad_norm": 2.393467426300049, "learning_rate": 9.734749019365001e-05, "loss": 3.9229, "step": 33200 }, { "epoch": 0.536281672876244, "grad_norm": 1.991856336593628, "learning_rate": 9.733931552126515e-05, "loss": 3.9931, "step": 33250 }, { "epoch": 0.5370881114820729, "grad_norm": 1.7219436168670654, "learning_rate": 9.733112861594422e-05, "loss": 3.9433, "step": 33300 }, { "epoch": 0.5378945500879018, "grad_norm": 2.8829445838928223, "learning_rate": 9.732292947980285e-05, "loss": 4.3263, "step": 33350 }, { "epoch": 0.5387009886937307, "grad_norm": 1.1644093990325928, "learning_rate": 9.731471811495973e-05, "loss": 4.1349, "step": 33400 }, { "epoch": 0.5395074272995597, "grad_norm": 2.788134813308716, "learning_rate": 9.730649452353678e-05, "loss": 4.1791, "step": 33450 }, { "epoch": 0.5403138659053887, "grad_norm": 1.2594338655471802, "learning_rate": 9.729825870765903e-05, "loss": 3.8136, "step": 33500 }, { "epoch": 0.5411203045112175, "grad_norm": 2.067070484161377, "learning_rate": 9.729001066945471e-05, "loss": 4.1771, "step": 33550 }, { "epoch": 0.5419267431170465, "grad_norm": 0.9377022385597229, "learning_rate": 9.728175041105519e-05, "loss": 3.9944, "step": 33600 }, { "epoch": 0.5427331817228754, "grad_norm": 1.7732789516448975, "learning_rate": 9.7273477934595e-05, "loss": 4.1222, "step": 33650 }, { "epoch": 0.5435396203287044, "grad_norm": 2.3974225521087646, "learning_rate": 9.726519324221184e-05, "loss": 4.2036, "step": 33700 }, { "epoch": 0.5443460589345334, "grad_norm": 3.553283929824829, "learning_rate": 9.725689633604654e-05, "loss": 4.4454, "step": 33750 }, { "epoch": 0.5451524975403622, "grad_norm": 1.552428960800171, "learning_rate": 9.724858721824311e-05, "loss": 3.8556, "step": 33800 }, { "epoch": 0.5459589361461912, "grad_norm": 1.2861645221710205, "learning_rate": 9.72402658909487e-05, "loss": 3.6202, "step": 33850 }, { "epoch": 0.5467653747520201, "grad_norm": 2.8274953365325928, "learning_rate": 9.723193235631362e-05, "loss": 3.7635, "step": 33900 }, { "epoch": 0.5475718133578491, "grad_norm": 1.0712840557098389, "learning_rate": 9.722358661649133e-05, "loss": 3.7407, "step": 33950 }, { "epoch": 0.548378251963678, "grad_norm": 2.3699841499328613, "learning_rate": 9.721522867363845e-05, "loss": 4.0412, "step": 34000 }, { "epoch": 0.5491846905695069, "grad_norm": 2.715761423110962, "learning_rate": 9.72068585299148e-05, "loss": 3.915, "step": 34050 }, { "epoch": 0.5499911291753359, "grad_norm": 4.615658760070801, "learning_rate": 9.719847618748323e-05, "loss": 4.0316, "step": 34100 }, { "epoch": 0.5507975677811648, "grad_norm": 2.428999662399292, "learning_rate": 9.719008164850986e-05, "loss": 4.2053, "step": 34150 }, { "epoch": 0.5516040063869938, "grad_norm": 0.604597806930542, "learning_rate": 9.718167491516395e-05, "loss": 4.3448, "step": 34200 }, { "epoch": 0.5524104449928227, "grad_norm": 2.038712978363037, "learning_rate": 9.717325598961781e-05, "loss": 4.0422, "step": 34250 }, { "epoch": 0.5532168835986516, "grad_norm": 1.6703771352767944, "learning_rate": 9.716482487404702e-05, "loss": 3.8453, "step": 34300 }, { "epoch": 0.5540233222044806, "grad_norm": 1.6739346981048584, "learning_rate": 9.715638157063025e-05, "loss": 3.9266, "step": 34350 }, { "epoch": 0.5548297608103095, "grad_norm": 2.9111719131469727, "learning_rate": 9.714792608154932e-05, "loss": 4.0806, "step": 34400 }, { "epoch": 0.5556361994161384, "grad_norm": 2.14871883392334, "learning_rate": 9.713945840898924e-05, "loss": 3.9186, "step": 34450 }, { "epoch": 0.5564426380219674, "grad_norm": 2.2379770278930664, "learning_rate": 9.71309785551381e-05, "loss": 3.7397, "step": 34500 }, { "epoch": 0.5572490766277963, "grad_norm": 0.7250761985778809, "learning_rate": 9.71224865221872e-05, "loss": 4.056, "step": 34550 }, { "epoch": 0.5580555152336253, "grad_norm": 1.8340208530426025, "learning_rate": 9.711398231233095e-05, "loss": 4.2596, "step": 34600 }, { "epoch": 0.5588619538394543, "grad_norm": 1.2574554681777954, "learning_rate": 9.710546592776695e-05, "loss": 3.9969, "step": 34650 }, { "epoch": 0.5596683924452831, "grad_norm": 1.2908345460891724, "learning_rate": 9.709693737069587e-05, "loss": 3.9818, "step": 34700 }, { "epoch": 0.5604748310511121, "grad_norm": 2.4312095642089844, "learning_rate": 9.708839664332161e-05, "loss": 3.7135, "step": 34750 }, { "epoch": 0.561281269656941, "grad_norm": 1.706939458847046, "learning_rate": 9.707984374785117e-05, "loss": 3.8907, "step": 34800 }, { "epoch": 0.56208770826277, "grad_norm": 1.253430962562561, "learning_rate": 9.707127868649468e-05, "loss": 3.8833, "step": 34850 }, { "epoch": 0.5628941468685988, "grad_norm": 2.2516000270843506, "learning_rate": 9.706270146146544e-05, "loss": 4.1565, "step": 34900 }, { "epoch": 0.5637005854744278, "grad_norm": 1.791192889213562, "learning_rate": 9.705411207497991e-05, "loss": 3.8712, "step": 34950 }, { "epoch": 0.5645070240802568, "grad_norm": 2.1465959548950195, "learning_rate": 9.704551052925765e-05, "loss": 3.8223, "step": 35000 }, { "epoch": 0.5653134626860857, "grad_norm": 1.0886331796646118, "learning_rate": 9.703689682652138e-05, "loss": 3.9508, "step": 35050 }, { "epoch": 0.5661199012919147, "grad_norm": 2.1944518089294434, "learning_rate": 9.702827096899698e-05, "loss": 4.3178, "step": 35100 }, { "epoch": 0.5669263398977435, "grad_norm": 1.7926956415176392, "learning_rate": 9.701963295891346e-05, "loss": 3.7831, "step": 35150 }, { "epoch": 0.5677327785035725, "grad_norm": 1.8286861181259155, "learning_rate": 9.701098279850292e-05, "loss": 4.0985, "step": 35200 }, { "epoch": 0.5685392171094015, "grad_norm": 1.7975794076919556, "learning_rate": 9.70023204900007e-05, "loss": 4.3829, "step": 35250 }, { "epoch": 0.5693456557152304, "grad_norm": 1.3266535997390747, "learning_rate": 9.69936460356452e-05, "loss": 3.7414, "step": 35300 }, { "epoch": 0.5701520943210593, "grad_norm": 2.072350263595581, "learning_rate": 9.698495943767797e-05, "loss": 4.2939, "step": 35350 }, { "epoch": 0.5709585329268883, "grad_norm": 1.8565208911895752, "learning_rate": 9.697626069834373e-05, "loss": 3.8508, "step": 35400 }, { "epoch": 0.5717649715327172, "grad_norm": 1.3206040859222412, "learning_rate": 9.696754981989031e-05, "loss": 3.5709, "step": 35450 }, { "epoch": 0.5725714101385462, "grad_norm": 2.337531566619873, "learning_rate": 9.695882680456869e-05, "loss": 4.1167, "step": 35500 }, { "epoch": 0.5733778487443751, "grad_norm": 2.7290146350860596, "learning_rate": 9.695009165463298e-05, "loss": 3.9105, "step": 35550 }, { "epoch": 0.574184287350204, "grad_norm": 2.879288911819458, "learning_rate": 9.694134437234044e-05, "loss": 4.0808, "step": 35600 }, { "epoch": 0.574990725956033, "grad_norm": 2.2341060638427734, "learning_rate": 9.693258495995141e-05, "loss": 4.2625, "step": 35650 }, { "epoch": 0.5757971645618619, "grad_norm": 2.0360066890716553, "learning_rate": 9.692381341972946e-05, "loss": 3.9065, "step": 35700 }, { "epoch": 0.5766036031676909, "grad_norm": 1.8910644054412842, "learning_rate": 9.69150297539412e-05, "loss": 3.7694, "step": 35750 }, { "epoch": 0.5774100417735197, "grad_norm": 1.2420679330825806, "learning_rate": 9.690623396485642e-05, "loss": 4.0219, "step": 35800 }, { "epoch": 0.5782164803793487, "grad_norm": 1.8608529567718506, "learning_rate": 9.689742605474807e-05, "loss": 3.9607, "step": 35850 }, { "epoch": 0.5790229189851777, "grad_norm": 1.903253197669983, "learning_rate": 9.688860602589217e-05, "loss": 4.225, "step": 35900 }, { "epoch": 0.5798293575910066, "grad_norm": 1.1789734363555908, "learning_rate": 9.687977388056791e-05, "loss": 3.7763, "step": 35950 }, { "epoch": 0.5806357961968356, "grad_norm": 2.1416497230529785, "learning_rate": 9.687092962105758e-05, "loss": 3.9497, "step": 36000 }, { "epoch": 0.5806357961968356, "eval_asr_loss": 0.5691508247039826, "eval_loss": 4.8229522705078125, "eval_runtime": 14.7902, "eval_samples_per_second": 34.617, "eval_steps_per_second": 11.562, "eval_tts_loss": 6.016266314813699, "step": 36000 }, { "epoch": 0.5814422348026644, "grad_norm": 1.5320308208465576, "learning_rate": 9.686207324964665e-05, "loss": 3.7274, "step": 36050 }, { "epoch": 0.5822486734084934, "grad_norm": 2.025629758834839, "learning_rate": 9.685320476862368e-05, "loss": 4.0585, "step": 36100 }, { "epoch": 0.5830551120143224, "grad_norm": 1.4545434713363647, "learning_rate": 9.684432418028038e-05, "loss": 4.0057, "step": 36150 }, { "epoch": 0.5838615506201513, "grad_norm": 1.8786442279815674, "learning_rate": 9.683543148691155e-05, "loss": 3.5916, "step": 36200 }, { "epoch": 0.5846679892259802, "grad_norm": 1.035013198852539, "learning_rate": 9.682652669081519e-05, "loss": 4.0424, "step": 36250 }, { "epoch": 0.5854744278318091, "grad_norm": 1.4029967784881592, "learning_rate": 9.681760979429235e-05, "loss": 4.2236, "step": 36300 }, { "epoch": 0.5862808664376381, "grad_norm": 3.3041937351226807, "learning_rate": 9.680868079964726e-05, "loss": 4.0875, "step": 36350 }, { "epoch": 0.5870873050434671, "grad_norm": 2.505573272705078, "learning_rate": 9.679973970918727e-05, "loss": 3.6474, "step": 36400 }, { "epoch": 0.587893743649296, "grad_norm": 2.1446595191955566, "learning_rate": 9.679078652522283e-05, "loss": 4.2775, "step": 36450 }, { "epoch": 0.5887001822551249, "grad_norm": 2.0899713039398193, "learning_rate": 9.678182125006751e-05, "loss": 3.9854, "step": 36500 }, { "epoch": 0.5895066208609538, "grad_norm": 1.1477386951446533, "learning_rate": 9.677284388603806e-05, "loss": 3.8436, "step": 36550 }, { "epoch": 0.5903130594667828, "grad_norm": 3.7815449237823486, "learning_rate": 9.67638544354543e-05, "loss": 3.7386, "step": 36600 }, { "epoch": 0.5911194980726118, "grad_norm": 1.9048818349838257, "learning_rate": 9.675485290063917e-05, "loss": 4.5235, "step": 36650 }, { "epoch": 0.5919259366784406, "grad_norm": 2.060019016265869, "learning_rate": 9.67458392839188e-05, "loss": 3.9605, "step": 36700 }, { "epoch": 0.5927323752842696, "grad_norm": 1.520529866218567, "learning_rate": 9.673681358762235e-05, "loss": 3.8785, "step": 36750 }, { "epoch": 0.5935388138900985, "grad_norm": 1.8698012828826904, "learning_rate": 9.672777581408218e-05, "loss": 4.1342, "step": 36800 }, { "epoch": 0.5943452524959275, "grad_norm": 2.1991450786590576, "learning_rate": 9.671872596563372e-05, "loss": 3.8919, "step": 36850 }, { "epoch": 0.5951516911017565, "grad_norm": 1.6680142879486084, "learning_rate": 9.670966404461554e-05, "loss": 3.8034, "step": 36900 }, { "epoch": 0.5959581297075853, "grad_norm": 1.9333608150482178, "learning_rate": 9.670059005336933e-05, "loss": 4.1449, "step": 36950 }, { "epoch": 0.5967645683134143, "grad_norm": 0.8060359358787537, "learning_rate": 9.669150399423991e-05, "loss": 3.7206, "step": 37000 }, { "epoch": 0.5975710069192433, "grad_norm": 2.4317374229431152, "learning_rate": 9.668240586957518e-05, "loss": 4.3372, "step": 37050 }, { "epoch": 0.5983774455250722, "grad_norm": 1.0086828470230103, "learning_rate": 9.66732956817262e-05, "loss": 4.3951, "step": 37100 }, { "epoch": 0.5991838841309011, "grad_norm": 1.9289790391921997, "learning_rate": 9.666417343304713e-05, "loss": 3.6865, "step": 37150 }, { "epoch": 0.59999032273673, "grad_norm": 1.7148078680038452, "learning_rate": 9.665503912589525e-05, "loss": 4.1055, "step": 37200 }, { "epoch": 0.600796761342559, "grad_norm": 1.2342246770858765, "learning_rate": 9.664589276263095e-05, "loss": 3.9735, "step": 37250 }, { "epoch": 0.601603199948388, "grad_norm": 1.9903031587600708, "learning_rate": 9.663673434561771e-05, "loss": 3.8671, "step": 37300 }, { "epoch": 0.6024096385542169, "grad_norm": 1.465268850326538, "learning_rate": 9.66275638772222e-05, "loss": 3.9581, "step": 37350 }, { "epoch": 0.6032160771600458, "grad_norm": 1.6464263200759888, "learning_rate": 9.661838135981412e-05, "loss": 3.9276, "step": 37400 }, { "epoch": 0.6040225157658747, "grad_norm": 2.538856267929077, "learning_rate": 9.660918679576636e-05, "loss": 3.9488, "step": 37450 }, { "epoch": 0.6048289543717037, "grad_norm": 3.353928565979004, "learning_rate": 9.659998018745484e-05, "loss": 4.0735, "step": 37500 }, { "epoch": 0.6056353929775327, "grad_norm": 2.7926583290100098, "learning_rate": 9.659076153725868e-05, "loss": 4.0669, "step": 37550 }, { "epoch": 0.6064418315833615, "grad_norm": 2.16097354888916, "learning_rate": 9.658153084756003e-05, "loss": 3.7335, "step": 37600 }, { "epoch": 0.6072482701891905, "grad_norm": 2.647451639175415, "learning_rate": 9.657228812074422e-05, "loss": 3.8995, "step": 37650 }, { "epoch": 0.6080547087950194, "grad_norm": 1.449508786201477, "learning_rate": 9.656303335919963e-05, "loss": 3.8566, "step": 37700 }, { "epoch": 0.6088611474008484, "grad_norm": 1.6518259048461914, "learning_rate": 9.655376656531781e-05, "loss": 4.1638, "step": 37750 }, { "epoch": 0.6096675860066774, "grad_norm": 2.536465883255005, "learning_rate": 9.654448774149336e-05, "loss": 4.4301, "step": 37800 }, { "epoch": 0.6104740246125062, "grad_norm": 1.383885145187378, "learning_rate": 9.653519689012403e-05, "loss": 3.7318, "step": 37850 }, { "epoch": 0.6112804632183352, "grad_norm": 2.0474696159362793, "learning_rate": 9.652589401361067e-05, "loss": 3.9712, "step": 37900 }, { "epoch": 0.6120869018241641, "grad_norm": 3.0140864849090576, "learning_rate": 9.651657911435725e-05, "loss": 3.931, "step": 37950 }, { "epoch": 0.6128933404299931, "grad_norm": 1.272126317024231, "learning_rate": 9.65072521947708e-05, "loss": 3.5631, "step": 38000 }, { "epoch": 0.613699779035822, "grad_norm": 1.682157039642334, "learning_rate": 9.649791325726149e-05, "loss": 3.4141, "step": 38050 }, { "epoch": 0.6145062176416509, "grad_norm": 1.6309934854507446, "learning_rate": 9.648856230424259e-05, "loss": 4.2273, "step": 38100 }, { "epoch": 0.6153126562474799, "grad_norm": 1.193287968635559, "learning_rate": 9.64791993381305e-05, "loss": 3.9418, "step": 38150 }, { "epoch": 0.6161190948533088, "grad_norm": 2.2340729236602783, "learning_rate": 9.646982436134469e-05, "loss": 4.3585, "step": 38200 }, { "epoch": 0.6169255334591378, "grad_norm": 3.2289340496063232, "learning_rate": 9.646043737630772e-05, "loss": 4.0445, "step": 38250 }, { "epoch": 0.6177319720649667, "grad_norm": 2.211266040802002, "learning_rate": 9.645103838544531e-05, "loss": 3.5903, "step": 38300 }, { "epoch": 0.6185384106707956, "grad_norm": 2.726902484893799, "learning_rate": 9.644162739118625e-05, "loss": 3.942, "step": 38350 }, { "epoch": 0.6193448492766246, "grad_norm": 2.515888214111328, "learning_rate": 9.64322043959624e-05, "loss": 4.0163, "step": 38400 }, { "epoch": 0.6201512878824536, "grad_norm": 1.7479592561721802, "learning_rate": 9.642276940220878e-05, "loss": 4.0748, "step": 38450 }, { "epoch": 0.6209577264882824, "grad_norm": 1.6890915632247925, "learning_rate": 9.641332241236348e-05, "loss": 3.924, "step": 38500 }, { "epoch": 0.6217641650941114, "grad_norm": 2.3948464393615723, "learning_rate": 9.640386342886767e-05, "loss": 3.9101, "step": 38550 }, { "epoch": 0.6225706036999403, "grad_norm": 2.6629281044006348, "learning_rate": 9.639439245416568e-05, "loss": 3.9898, "step": 38600 }, { "epoch": 0.6233770423057693, "grad_norm": 1.5631389617919922, "learning_rate": 9.638490949070488e-05, "loss": 3.83, "step": 38650 }, { "epoch": 0.6241834809115983, "grad_norm": 1.956933856010437, "learning_rate": 9.637541454093574e-05, "loss": 4.0719, "step": 38700 }, { "epoch": 0.6249899195174271, "grad_norm": 2.43593430519104, "learning_rate": 9.63659076073119e-05, "loss": 4.0034, "step": 38750 }, { "epoch": 0.6257963581232561, "grad_norm": 1.7999205589294434, "learning_rate": 9.635638869229e-05, "loss": 3.6974, "step": 38800 }, { "epoch": 0.626602796729085, "grad_norm": 3.0806641578674316, "learning_rate": 9.634685779832983e-05, "loss": 4.0155, "step": 38850 }, { "epoch": 0.627409235334914, "grad_norm": 2.0375235080718994, "learning_rate": 9.633731492789428e-05, "loss": 3.9403, "step": 38900 }, { "epoch": 0.6282156739407428, "grad_norm": 2.0660057067871094, "learning_rate": 9.632776008344929e-05, "loss": 3.9845, "step": 38950 }, { "epoch": 0.6290221125465718, "grad_norm": 3.3655974864959717, "learning_rate": 9.631819326746394e-05, "loss": 4.0855, "step": 39000 }, { "epoch": 0.6290221125465718, "eval_asr_loss": 0.6034199091663585, "eval_loss": 4.653688430786133, "eval_runtime": 14.8011, "eval_samples_per_second": 34.592, "eval_steps_per_second": 11.553, "eval_tts_loss": 6.013137326975981, "step": 39000 }, { "epoch": 0.6298285511524008, "grad_norm": 2.8556010723114014, "learning_rate": 9.63086144824104e-05, "loss": 4.1555, "step": 39050 }, { "epoch": 0.6306349897582297, "grad_norm": 0.7945855259895325, "learning_rate": 9.62990237307639e-05, "loss": 3.9282, "step": 39100 }, { "epoch": 0.6314414283640587, "grad_norm": 2.212608575820923, "learning_rate": 9.628942101500281e-05, "loss": 3.9476, "step": 39150 }, { "epoch": 0.6322478669698876, "grad_norm": 2.4456074237823486, "learning_rate": 9.627980633760852e-05, "loss": 4.2139, "step": 39200 }, { "epoch": 0.6330543055757165, "grad_norm": 2.3700897693634033, "learning_rate": 9.62701797010656e-05, "loss": 3.9148, "step": 39250 }, { "epoch": 0.6338607441815455, "grad_norm": 1.0412943363189697, "learning_rate": 9.626054110786163e-05, "loss": 3.7921, "step": 39300 }, { "epoch": 0.6346671827873744, "grad_norm": 1.5326954126358032, "learning_rate": 9.625089056048736e-05, "loss": 4.3737, "step": 39350 }, { "epoch": 0.6354736213932033, "grad_norm": 1.076934814453125, "learning_rate": 9.624122806143652e-05, "loss": 3.9424, "step": 39400 }, { "epoch": 0.6362800599990323, "grad_norm": 1.8640769720077515, "learning_rate": 9.623155361320607e-05, "loss": 4.3091, "step": 39450 }, { "epoch": 0.6370864986048612, "grad_norm": 2.431981325149536, "learning_rate": 9.62218672182959e-05, "loss": 3.7956, "step": 39500 }, { "epoch": 0.6378929372106902, "grad_norm": 3.0567352771759033, "learning_rate": 9.621216887920913e-05, "loss": 3.9352, "step": 39550 }, { "epoch": 0.6386993758165191, "grad_norm": 1.1013330221176147, "learning_rate": 9.620245859845188e-05, "loss": 4.0237, "step": 39600 }, { "epoch": 0.639505814422348, "grad_norm": 2.7281951904296875, "learning_rate": 9.619273637853339e-05, "loss": 4.3155, "step": 39650 }, { "epoch": 0.640312253028177, "grad_norm": 1.6277055740356445, "learning_rate": 9.618300222196596e-05, "loss": 4.0474, "step": 39700 }, { "epoch": 0.6411186916340059, "grad_norm": 1.7232165336608887, "learning_rate": 9.617325613126502e-05, "loss": 3.7625, "step": 39750 }, { "epoch": 0.6419251302398349, "grad_norm": 1.3456089496612549, "learning_rate": 9.6163498108949e-05, "loss": 3.8336, "step": 39800 }, { "epoch": 0.6427315688456637, "grad_norm": 3.376032590866089, "learning_rate": 9.615372815753954e-05, "loss": 3.9334, "step": 39850 }, { "epoch": 0.6435380074514927, "grad_norm": 3.418212890625, "learning_rate": 9.614394627956121e-05, "loss": 4.2293, "step": 39900 }, { "epoch": 0.6443444460573217, "grad_norm": 1.5658949613571167, "learning_rate": 9.613415247754181e-05, "loss": 4.0883, "step": 39950 }, { "epoch": 0.6451508846631506, "grad_norm": 1.2223689556121826, "learning_rate": 9.612434675401212e-05, "loss": 3.9809, "step": 40000 }, { "epoch": 0.6459573232689795, "grad_norm": 1.4119608402252197, "learning_rate": 9.611452911150603e-05, "loss": 4.135, "step": 40050 }, { "epoch": 0.6467637618748084, "grad_norm": 1.1455477476119995, "learning_rate": 9.610469955256052e-05, "loss": 4.0727, "step": 40100 }, { "epoch": 0.6475702004806374, "grad_norm": 1.1804072856903076, "learning_rate": 9.609485807971566e-05, "loss": 3.7652, "step": 40150 }, { "epoch": 0.6483766390864664, "grad_norm": 2.133227586746216, "learning_rate": 9.608500469551455e-05, "loss": 3.6933, "step": 40200 }, { "epoch": 0.6491830776922953, "grad_norm": 1.6320295333862305, "learning_rate": 9.60751394025034e-05, "loss": 3.7121, "step": 40250 }, { "epoch": 0.6499895162981242, "grad_norm": 2.7744624614715576, "learning_rate": 9.606526220323155e-05, "loss": 4.1106, "step": 40300 }, { "epoch": 0.6507959549039531, "grad_norm": 2.890082836151123, "learning_rate": 9.605537310025131e-05, "loss": 3.8397, "step": 40350 }, { "epoch": 0.6516023935097821, "grad_norm": 1.4252163171768188, "learning_rate": 9.604547209611815e-05, "loss": 3.9493, "step": 40400 }, { "epoch": 0.6524088321156111, "grad_norm": 1.658120036125183, "learning_rate": 9.603555919339055e-05, "loss": 4.0099, "step": 40450 }, { "epoch": 0.6532152707214399, "grad_norm": 2.3012828826904297, "learning_rate": 9.602563439463012e-05, "loss": 3.8907, "step": 40500 }, { "epoch": 0.6540217093272689, "grad_norm": 2.3004066944122314, "learning_rate": 9.601569770240154e-05, "loss": 4.2203, "step": 40550 }, { "epoch": 0.6548281479330978, "grad_norm": 0.95237135887146, "learning_rate": 9.600574911927253e-05, "loss": 4.0099, "step": 40600 }, { "epoch": 0.6556345865389268, "grad_norm": 2.167707681655884, "learning_rate": 9.599578864781389e-05, "loss": 3.4495, "step": 40650 }, { "epoch": 0.6564410251447558, "grad_norm": 1.9729243516921997, "learning_rate": 9.598581629059952e-05, "loss": 3.9483, "step": 40700 }, { "epoch": 0.6572474637505846, "grad_norm": 0.6504325270652771, "learning_rate": 9.597583205020636e-05, "loss": 3.859, "step": 40750 }, { "epoch": 0.6580539023564136, "grad_norm": 2.181685447692871, "learning_rate": 9.596583592921446e-05, "loss": 3.6357, "step": 40800 }, { "epoch": 0.6588603409622426, "grad_norm": 1.3120620250701904, "learning_rate": 9.595582793020688e-05, "loss": 3.7685, "step": 40850 }, { "epoch": 0.6596667795680715, "grad_norm": 3.323073148727417, "learning_rate": 9.59458080557698e-05, "loss": 4.3613, "step": 40900 }, { "epoch": 0.6604732181739004, "grad_norm": 1.2107150554656982, "learning_rate": 9.593577630849246e-05, "loss": 4.0395, "step": 40950 }, { "epoch": 0.6612796567797293, "grad_norm": 2.1152050495147705, "learning_rate": 9.592573269096716e-05, "loss": 3.6636, "step": 41000 }, { "epoch": 0.6620860953855583, "grad_norm": 1.9166638851165771, "learning_rate": 9.591567720578924e-05, "loss": 3.8332, "step": 41050 }, { "epoch": 0.6628925339913873, "grad_norm": 1.2429214715957642, "learning_rate": 9.590560985555716e-05, "loss": 4.0641, "step": 41100 }, { "epoch": 0.6636989725972162, "grad_norm": 2.569197177886963, "learning_rate": 9.58955306428724e-05, "loss": 4.026, "step": 41150 }, { "epoch": 0.6645054112030451, "grad_norm": 1.7593148946762085, "learning_rate": 9.588543957033957e-05, "loss": 3.88, "step": 41200 }, { "epoch": 0.665311849808874, "grad_norm": 1.2028878927230835, "learning_rate": 9.587553881534579e-05, "loss": 4.0756, "step": 41250 }, { "epoch": 0.666118288414703, "grad_norm": 2.990448474884033, "learning_rate": 9.586542426800969e-05, "loss": 4.2531, "step": 41300 }, { "epoch": 0.666924727020532, "grad_norm": 2.4063851833343506, "learning_rate": 9.585529786860527e-05, "loss": 3.9218, "step": 41350 }, { "epoch": 0.6677311656263608, "grad_norm": 1.7430907487869263, "learning_rate": 9.584515961974928e-05, "loss": 4.3134, "step": 41400 }, { "epoch": 0.6685376042321898, "grad_norm": 1.3421581983566284, "learning_rate": 9.583500952406155e-05, "loss": 4.0289, "step": 41450 }, { "epoch": 0.6693440428380187, "grad_norm": 0.871703565120697, "learning_rate": 9.582484758416497e-05, "loss": 3.7478, "step": 41500 }, { "epoch": 0.6701504814438477, "grad_norm": 1.735887885093689, "learning_rate": 9.581467380268547e-05, "loss": 3.9596, "step": 41550 }, { "epoch": 0.6709569200496767, "grad_norm": 1.485318660736084, "learning_rate": 9.580448818225205e-05, "loss": 4.1336, "step": 41600 }, { "epoch": 0.6717633586555055, "grad_norm": 1.8921575546264648, "learning_rate": 9.579429072549679e-05, "loss": 3.8069, "step": 41650 }, { "epoch": 0.6725697972613345, "grad_norm": 1.4174717664718628, "learning_rate": 9.57840814350548e-05, "loss": 3.7202, "step": 41700 }, { "epoch": 0.6733762358671634, "grad_norm": 2.4129083156585693, "learning_rate": 9.577386031356425e-05, "loss": 3.9037, "step": 41750 }, { "epoch": 0.6741826744729924, "grad_norm": 3.722903251647949, "learning_rate": 9.57636273636664e-05, "loss": 3.971, "step": 41800 }, { "epoch": 0.6749891130788213, "grad_norm": 1.9005436897277832, "learning_rate": 9.57535875993941e-05, "loss": 3.9043, "step": 41850 }, { "epoch": 0.6757955516846502, "grad_norm": 2.620349645614624, "learning_rate": 9.574333123705391e-05, "loss": 4.0456, "step": 41900 }, { "epoch": 0.6766019902904792, "grad_norm": 1.5120000839233398, "learning_rate": 9.573306305419541e-05, "loss": 3.9113, "step": 41950 }, { "epoch": 0.6774084288963081, "grad_norm": 2.184945821762085, "learning_rate": 9.572278305347202e-05, "loss": 3.9566, "step": 42000 }, { "epoch": 0.6774084288963081, "eval_asr_loss": 0.6747458191937141, "eval_loss": 4.402849197387695, "eval_runtime": 14.8128, "eval_samples_per_second": 34.565, "eval_steps_per_second": 11.544, "eval_tts_loss": 6.006661847860199, "step": 42000 }, { "epoch": 0.6782148675021371, "grad_norm": 1.9270944595336914, "learning_rate": 9.571249123754014e-05, "loss": 4.0085, "step": 42050 }, { "epoch": 0.679021306107966, "grad_norm": 2.115370035171509, "learning_rate": 9.570218760905934e-05, "loss": 3.9551, "step": 42100 }, { "epoch": 0.6798277447137949, "grad_norm": 2.4913384914398193, "learning_rate": 9.569187217069213e-05, "loss": 3.9498, "step": 42150 }, { "epoch": 0.6806341833196239, "grad_norm": 1.2166416645050049, "learning_rate": 9.568154492510414e-05, "loss": 3.9295, "step": 42200 }, { "epoch": 0.6814406219254528, "grad_norm": 3.2959489822387695, "learning_rate": 9.567120587496402e-05, "loss": 3.8898, "step": 42250 }, { "epoch": 0.6822470605312817, "grad_norm": 3.9133834838867188, "learning_rate": 9.566085502294348e-05, "loss": 4.1996, "step": 42300 }, { "epoch": 0.6830534991371107, "grad_norm": 1.421493411064148, "learning_rate": 9.56504923717173e-05, "loss": 3.9401, "step": 42350 }, { "epoch": 0.6838599377429396, "grad_norm": 1.4366188049316406, "learning_rate": 9.564011792396326e-05, "loss": 4.1854, "step": 42400 }, { "epoch": 0.6846663763487686, "grad_norm": 2.0820562839508057, "learning_rate": 9.562973168236224e-05, "loss": 4.1061, "step": 42450 }, { "epoch": 0.6854728149545976, "grad_norm": 1.6290011405944824, "learning_rate": 9.561933364959814e-05, "loss": 4.3134, "step": 42500 }, { "epoch": 0.6862792535604264, "grad_norm": 2.4670650959014893, "learning_rate": 9.560892382835791e-05, "loss": 3.7429, "step": 42550 }, { "epoch": 0.6870856921662554, "grad_norm": 1.2104759216308594, "learning_rate": 9.559850222133155e-05, "loss": 3.6463, "step": 42600 }, { "epoch": 0.6878921307720843, "grad_norm": 1.4174095392227173, "learning_rate": 9.558806883121209e-05, "loss": 4.0535, "step": 42650 }, { "epoch": 0.6886985693779133, "grad_norm": 2.343122959136963, "learning_rate": 9.557762366069562e-05, "loss": 4.2928, "step": 42700 }, { "epoch": 0.6895050079837421, "grad_norm": 2.328246831893921, "learning_rate": 9.55671667124813e-05, "loss": 4.1672, "step": 42750 }, { "epoch": 0.6903114465895711, "grad_norm": 2.1924808025360107, "learning_rate": 9.55566979892713e-05, "loss": 4.0338, "step": 42800 }, { "epoch": 0.6911178851954001, "grad_norm": 1.4481010437011719, "learning_rate": 9.554621749377079e-05, "loss": 3.8551, "step": 42850 }, { "epoch": 0.691924323801229, "grad_norm": 2.5818378925323486, "learning_rate": 9.55357252286881e-05, "loss": 3.8865, "step": 42900 }, { "epoch": 0.692730762407058, "grad_norm": 0.6974899768829346, "learning_rate": 9.552522119673449e-05, "loss": 4.3369, "step": 42950 }, { "epoch": 0.6935372010128869, "grad_norm": 2.877656936645508, "learning_rate": 9.551470540062433e-05, "loss": 3.6878, "step": 43000 }, { "epoch": 0.6943436396187158, "grad_norm": 1.5497684478759766, "learning_rate": 9.550417784307497e-05, "loss": 3.6174, "step": 43050 }, { "epoch": 0.6951500782245448, "grad_norm": 2.261922836303711, "learning_rate": 9.549363852680686e-05, "loss": 3.9077, "step": 43100 }, { "epoch": 0.6959565168303737, "grad_norm": 2.5778346061706543, "learning_rate": 9.548308745454344e-05, "loss": 3.8163, "step": 43150 }, { "epoch": 0.6967629554362026, "grad_norm": 2.7817728519439697, "learning_rate": 9.547252462901124e-05, "loss": 3.6548, "step": 43200 }, { "epoch": 0.6975693940420316, "grad_norm": 1.0119379758834839, "learning_rate": 9.546195005293976e-05, "loss": 4.1415, "step": 43250 }, { "epoch": 0.6983758326478605, "grad_norm": 3.6960830688476562, "learning_rate": 9.54513637290616e-05, "loss": 4.0043, "step": 43300 }, { "epoch": 0.6991822712536895, "grad_norm": 1.3119101524353027, "learning_rate": 9.544076566011235e-05, "loss": 4.0963, "step": 43350 }, { "epoch": 0.6999887098595184, "grad_norm": 1.1128474473953247, "learning_rate": 9.543015584883067e-05, "loss": 3.7204, "step": 43400 }, { "epoch": 0.7007951484653473, "grad_norm": 2.3756420612335205, "learning_rate": 9.541953429795823e-05, "loss": 4.3295, "step": 43450 }, { "epoch": 0.7016015870711763, "grad_norm": 1.1460477113723755, "learning_rate": 9.540890101023972e-05, "loss": 3.8782, "step": 43500 }, { "epoch": 0.7024080256770052, "grad_norm": 1.0476806163787842, "learning_rate": 9.539825598842291e-05, "loss": 3.8376, "step": 43550 }, { "epoch": 0.7032144642828342, "grad_norm": 1.2808899879455566, "learning_rate": 9.538759923525857e-05, "loss": 4.1479, "step": 43600 }, { "epoch": 0.704020902888663, "grad_norm": 2.6137540340423584, "learning_rate": 9.537693075350052e-05, "loss": 3.8806, "step": 43650 }, { "epoch": 0.704827341494492, "grad_norm": 1.4536105394363403, "learning_rate": 9.536625054590559e-05, "loss": 3.9868, "step": 43700 }, { "epoch": 0.705633780100321, "grad_norm": 1.8267903327941895, "learning_rate": 9.535555861523363e-05, "loss": 4.0424, "step": 43750 }, { "epoch": 0.7064402187061499, "grad_norm": 1.1175280809402466, "learning_rate": 9.534485496424757e-05, "loss": 4.0311, "step": 43800 }, { "epoch": 0.7072466573119789, "grad_norm": 2.639409065246582, "learning_rate": 9.533413959571332e-05, "loss": 3.7509, "step": 43850 }, { "epoch": 0.7080530959178077, "grad_norm": 2.0124926567077637, "learning_rate": 9.532341251239982e-05, "loss": 4.1953, "step": 43900 }, { "epoch": 0.7088595345236367, "grad_norm": 1.6869808435440063, "learning_rate": 9.531267371707906e-05, "loss": 3.5422, "step": 43950 }, { "epoch": 0.7096659731294657, "grad_norm": 2.89847469329834, "learning_rate": 9.530192321252607e-05, "loss": 3.8341, "step": 44000 }, { "epoch": 0.7104724117352946, "grad_norm": 1.6155884265899658, "learning_rate": 9.529116100151887e-05, "loss": 4.1777, "step": 44050 }, { "epoch": 0.7112788503411235, "grad_norm": 2.9309215545654297, "learning_rate": 9.52803870868385e-05, "loss": 3.8159, "step": 44100 }, { "epoch": 0.7120852889469524, "grad_norm": 2.221287727355957, "learning_rate": 9.526960147126908e-05, "loss": 4.1738, "step": 44150 }, { "epoch": 0.7128917275527814, "grad_norm": 2.1958932876586914, "learning_rate": 9.525880415759772e-05, "loss": 4.1857, "step": 44200 }, { "epoch": 0.7136981661586104, "grad_norm": 1.2787940502166748, "learning_rate": 9.524799514861449e-05, "loss": 3.5715, "step": 44250 }, { "epoch": 0.7145046047644393, "grad_norm": 1.3575184345245361, "learning_rate": 9.52371744471126e-05, "loss": 3.6301, "step": 44300 }, { "epoch": 0.7153110433702682, "grad_norm": 1.1115514039993286, "learning_rate": 9.522634205588821e-05, "loss": 4.0579, "step": 44350 }, { "epoch": 0.7161174819760971, "grad_norm": 2.445248603820801, "learning_rate": 9.52154979777405e-05, "loss": 3.9905, "step": 44400 }, { "epoch": 0.7169239205819261, "grad_norm": 4.455885410308838, "learning_rate": 9.52046422154717e-05, "loss": 3.8643, "step": 44450 }, { "epoch": 0.7177303591877551, "grad_norm": 2.097954034805298, "learning_rate": 9.519377477188704e-05, "loss": 4.2417, "step": 44500 }, { "epoch": 0.7185367977935839, "grad_norm": 1.8848130702972412, "learning_rate": 9.518289564979475e-05, "loss": 3.7867, "step": 44550 }, { "epoch": 0.7193432363994129, "grad_norm": 2.098465919494629, "learning_rate": 9.517200485200614e-05, "loss": 3.9569, "step": 44600 }, { "epoch": 0.7201496750052419, "grad_norm": 2.0162923336029053, "learning_rate": 9.516110238133548e-05, "loss": 3.6984, "step": 44650 }, { "epoch": 0.7209561136110708, "grad_norm": 1.9503788948059082, "learning_rate": 9.515018824060006e-05, "loss": 3.7651, "step": 44700 }, { "epoch": 0.7217625522168998, "grad_norm": 2.917295455932617, "learning_rate": 9.513926243262022e-05, "loss": 3.5928, "step": 44750 }, { "epoch": 0.7225689908227286, "grad_norm": 1.6340333223342896, "learning_rate": 9.512832496021928e-05, "loss": 3.9921, "step": 44800 }, { "epoch": 0.7233754294285576, "grad_norm": 2.323976755142212, "learning_rate": 9.51173758262236e-05, "loss": 3.8905, "step": 44850 }, { "epoch": 0.7241818680343866, "grad_norm": 2.567837715148926, "learning_rate": 9.510641503346252e-05, "loss": 3.973, "step": 44900 }, { "epoch": 0.7249883066402155, "grad_norm": 1.992812156677246, "learning_rate": 9.509544258476845e-05, "loss": 4.4189, "step": 44950 }, { "epoch": 0.7257947452460444, "grad_norm": 2.9449970722198486, "learning_rate": 9.508445848297674e-05, "loss": 3.6855, "step": 45000 }, { "epoch": 0.7257947452460444, "eval_asr_loss": 0.7677889460602478, "eval_loss": 4.220231056213379, "eval_runtime": 14.4142, "eval_samples_per_second": 35.52, "eval_steps_per_second": 11.863, "eval_tts_loss": 5.966565507839665, "step": 45000 }, { "epoch": 0.7266011838518733, "grad_norm": 2.878650426864624, "learning_rate": 9.50734627309258e-05, "loss": 4.118, "step": 45050 }, { "epoch": 0.7274076224577023, "grad_norm": 1.361701250076294, "learning_rate": 9.506245533145707e-05, "loss": 4.3712, "step": 45100 }, { "epoch": 0.7282140610635313, "grad_norm": 2.126094341278076, "learning_rate": 9.505143628741492e-05, "loss": 3.7077, "step": 45150 }, { "epoch": 0.7290204996693602, "grad_norm": 2.355354070663452, "learning_rate": 9.50404056016468e-05, "loss": 3.8391, "step": 45200 }, { "epoch": 0.7298269382751891, "grad_norm": 1.933937668800354, "learning_rate": 9.502936327700315e-05, "loss": 3.7756, "step": 45250 }, { "epoch": 0.730633376881018, "grad_norm": 2.8222224712371826, "learning_rate": 9.50183093163374e-05, "loss": 4.0453, "step": 45300 }, { "epoch": 0.731439815486847, "grad_norm": 0.3452666699886322, "learning_rate": 9.500724372250602e-05, "loss": 3.9045, "step": 45350 }, { "epoch": 0.732246254092676, "grad_norm": 0.9124054312705994, "learning_rate": 9.499616649836845e-05, "loss": 3.9615, "step": 45400 }, { "epoch": 0.7330526926985048, "grad_norm": 2.117830276489258, "learning_rate": 9.498507764678715e-05, "loss": 4.0645, "step": 45450 }, { "epoch": 0.7338591313043338, "grad_norm": 2.6719260215759277, "learning_rate": 9.497397717062759e-05, "loss": 3.7295, "step": 45500 }, { "epoch": 0.7346655699101627, "grad_norm": 2.009535312652588, "learning_rate": 9.496286507275825e-05, "loss": 4.3494, "step": 45550 }, { "epoch": 0.7354720085159917, "grad_norm": 1.6704541444778442, "learning_rate": 9.495174135605058e-05, "loss": 3.5423, "step": 45600 }, { "epoch": 0.7362784471218207, "grad_norm": 2.1379082202911377, "learning_rate": 9.494060602337909e-05, "loss": 4.2106, "step": 45650 }, { "epoch": 0.7370848857276495, "grad_norm": 1.076125979423523, "learning_rate": 9.492945907762123e-05, "loss": 4.1118, "step": 45700 }, { "epoch": 0.7378913243334785, "grad_norm": 2.360368013381958, "learning_rate": 9.491830052165749e-05, "loss": 3.9588, "step": 45750 }, { "epoch": 0.7386977629393074, "grad_norm": 1.9395514726638794, "learning_rate": 9.490713035837133e-05, "loss": 3.9243, "step": 45800 }, { "epoch": 0.7395042015451364, "grad_norm": 1.406371831893921, "learning_rate": 9.489594859064926e-05, "loss": 4.1377, "step": 45850 }, { "epoch": 0.7403106401509653, "grad_norm": 1.5525023937225342, "learning_rate": 9.488475522138073e-05, "loss": 3.6349, "step": 45900 }, { "epoch": 0.7411170787567942, "grad_norm": 1.3962270021438599, "learning_rate": 9.487355025345823e-05, "loss": 3.8225, "step": 45950 }, { "epoch": 0.7419235173626232, "grad_norm": 2.4570751190185547, "learning_rate": 9.486255813467055e-05, "loss": 3.8946, "step": 46000 }, { "epoch": 0.7427299559684521, "grad_norm": 9.147421836853027, "learning_rate": 9.485133020995828e-05, "loss": 3.9156, "step": 46050 }, { "epoch": 0.7435363945742811, "grad_norm": 1.4232938289642334, "learning_rate": 9.484009069522938e-05, "loss": 4.113, "step": 46100 }, { "epoch": 0.74434283318011, "grad_norm": 3.1398966312408447, "learning_rate": 9.482883959338827e-05, "loss": 4.0905, "step": 46150 }, { "epoch": 0.7451492717859389, "grad_norm": 2.162477493286133, "learning_rate": 9.481757690734232e-05, "loss": 3.9378, "step": 46200 }, { "epoch": 0.7459557103917679, "grad_norm": 2.1014811992645264, "learning_rate": 9.48063026400019e-05, "loss": 3.9758, "step": 46250 }, { "epoch": 0.7467621489975969, "grad_norm": 1.6452032327651978, "learning_rate": 9.479501679428042e-05, "loss": 4.0847, "step": 46300 }, { "epoch": 0.7475685876034257, "grad_norm": 2.0140156745910645, "learning_rate": 9.478371937309425e-05, "loss": 4.3145, "step": 46350 }, { "epoch": 0.7483750262092547, "grad_norm": 1.8627792596817017, "learning_rate": 9.477241037936273e-05, "loss": 4.0373, "step": 46400 }, { "epoch": 0.7491814648150836, "grad_norm": 1.3735127449035645, "learning_rate": 9.476108981600822e-05, "loss": 3.9577, "step": 46450 }, { "epoch": 0.7499879034209126, "grad_norm": 2.6094725131988525, "learning_rate": 9.474975768595609e-05, "loss": 3.957, "step": 46500 }, { "epoch": 0.7507943420267416, "grad_norm": 1.3980075120925903, "learning_rate": 9.473841399213463e-05, "loss": 4.1313, "step": 46550 }, { "epoch": 0.7516007806325704, "grad_norm": 3.6238787174224854, "learning_rate": 9.47270587374752e-05, "loss": 3.9414, "step": 46600 }, { "epoch": 0.7524072192383994, "grad_norm": 1.450015902519226, "learning_rate": 9.471569192491208e-05, "loss": 3.7486, "step": 46650 }, { "epoch": 0.7532136578442283, "grad_norm": 1.9023935794830322, "learning_rate": 9.470431355738257e-05, "loss": 4.0542, "step": 46700 }, { "epoch": 0.7540200964500573, "grad_norm": 1.0241719484329224, "learning_rate": 9.469292363782696e-05, "loss": 3.9859, "step": 46750 }, { "epoch": 0.7548265350558861, "grad_norm": 1.9951140880584717, "learning_rate": 9.468152216918849e-05, "loss": 3.5983, "step": 46800 }, { "epoch": 0.7556329736617151, "grad_norm": 1.519264817237854, "learning_rate": 9.467010915441343e-05, "loss": 3.971, "step": 46850 }, { "epoch": 0.7564394122675441, "grad_norm": 2.4989826679229736, "learning_rate": 9.465868459645102e-05, "loss": 3.9566, "step": 46900 }, { "epoch": 0.757245850873373, "grad_norm": 2.34820294380188, "learning_rate": 9.464724849825347e-05, "loss": 4.1283, "step": 46950 }, { "epoch": 0.758052289479202, "grad_norm": 1.598359227180481, "learning_rate": 9.463580086277598e-05, "loss": 3.8883, "step": 47000 }, { "epoch": 0.7588587280850309, "grad_norm": 1.438071370124817, "learning_rate": 9.462434169297671e-05, "loss": 3.6827, "step": 47050 }, { "epoch": 0.7596651666908598, "grad_norm": 0.7301783561706543, "learning_rate": 9.461287099181683e-05, "loss": 3.9224, "step": 47100 }, { "epoch": 0.7604716052966888, "grad_norm": 4.070219039916992, "learning_rate": 9.46013887622605e-05, "loss": 4.0544, "step": 47150 }, { "epoch": 0.7612780439025177, "grad_norm": 2.0918643474578857, "learning_rate": 9.458989500727481e-05, "loss": 3.7879, "step": 47200 }, { "epoch": 0.7620844825083466, "grad_norm": 2.473740577697754, "learning_rate": 9.457838972982986e-05, "loss": 4.0541, "step": 47250 }, { "epoch": 0.7628909211141756, "grad_norm": 1.9549883604049683, "learning_rate": 9.456687293289877e-05, "loss": 4.3095, "step": 47300 }, { "epoch": 0.7636973597200045, "grad_norm": 1.1889281272888184, "learning_rate": 9.455534461945751e-05, "loss": 3.9717, "step": 47350 }, { "epoch": 0.7645037983258335, "grad_norm": 2.1214423179626465, "learning_rate": 9.45438047924852e-05, "loss": 3.9387, "step": 47400 }, { "epoch": 0.7653102369316624, "grad_norm": 1.206317663192749, "learning_rate": 9.453225345496377e-05, "loss": 4.1136, "step": 47450 }, { "epoch": 0.7661166755374913, "grad_norm": 2.127297878265381, "learning_rate": 9.452069060987823e-05, "loss": 3.6162, "step": 47500 }, { "epoch": 0.7669231141433203, "grad_norm": 1.8844021558761597, "learning_rate": 9.450911626021653e-05, "loss": 3.9547, "step": 47550 }, { "epoch": 0.7677295527491492, "grad_norm": 1.8180209398269653, "learning_rate": 9.449753040896958e-05, "loss": 4.0026, "step": 47600 }, { "epoch": 0.7685359913549782, "grad_norm": 4.184787273406982, "learning_rate": 9.448593305913129e-05, "loss": 3.5137, "step": 47650 }, { "epoch": 0.769342429960807, "grad_norm": 1.7773044109344482, "learning_rate": 9.447432421369851e-05, "loss": 3.6649, "step": 47700 }, { "epoch": 0.770148868566636, "grad_norm": 2.41501784324646, "learning_rate": 9.446270387567109e-05, "loss": 4.1824, "step": 47750 }, { "epoch": 0.770955307172465, "grad_norm": 1.6952520608901978, "learning_rate": 9.445107204805183e-05, "loss": 3.5778, "step": 47800 }, { "epoch": 0.7717617457782939, "grad_norm": 3.045311212539673, "learning_rate": 9.443942873384653e-05, "loss": 3.9632, "step": 47850 }, { "epoch": 0.7725681843841229, "grad_norm": 2.5079586505889893, "learning_rate": 9.442777393606393e-05, "loss": 3.8897, "step": 47900 }, { "epoch": 0.7733746229899517, "grad_norm": 2.440715789794922, "learning_rate": 9.44161076577157e-05, "loss": 3.7363, "step": 47950 }, { "epoch": 0.7741810615957807, "grad_norm": 1.950965166091919, "learning_rate": 9.440442990181656e-05, "loss": 3.7987, "step": 48000 }, { "epoch": 0.7741810615957807, "eval_asr_loss": 0.7671968561538705, "eval_loss": 4.1380295753479, "eval_runtime": 14.6173, "eval_samples_per_second": 35.027, "eval_steps_per_second": 11.698, "eval_tts_loss": 6.009661002101706, "step": 48000 }, { "epoch": 0.7749875002016097, "grad_norm": 1.2615563869476318, "learning_rate": 9.439274067138416e-05, "loss": 4.1374, "step": 48050 }, { "epoch": 0.7757939388074386, "grad_norm": 2.113501787185669, "learning_rate": 9.438103996943906e-05, "loss": 3.9375, "step": 48100 }, { "epoch": 0.7766003774132675, "grad_norm": 1.9320517778396606, "learning_rate": 9.436932779900488e-05, "loss": 4.1499, "step": 48150 }, { "epoch": 0.7774068160190964, "grad_norm": 2.7595505714416504, "learning_rate": 9.435760416310815e-05, "loss": 4.0281, "step": 48200 }, { "epoch": 0.7782132546249254, "grad_norm": 2.4536404609680176, "learning_rate": 9.434586906477837e-05, "loss": 3.8629, "step": 48250 }, { "epoch": 0.7790196932307544, "grad_norm": 2.977201223373413, "learning_rate": 9.433412250704799e-05, "loss": 3.9147, "step": 48300 }, { "epoch": 0.7798261318365833, "grad_norm": 3.090271234512329, "learning_rate": 9.432236449295244e-05, "loss": 4.1099, "step": 48350 }, { "epoch": 0.7806325704424122, "grad_norm": 1.6064294576644897, "learning_rate": 9.431059502553012e-05, "loss": 3.7641, "step": 48400 }, { "epoch": 0.7814390090482412, "grad_norm": 1.3545068502426147, "learning_rate": 9.429881410782236e-05, "loss": 4.1789, "step": 48450 }, { "epoch": 0.7822454476540701, "grad_norm": 1.8167670965194702, "learning_rate": 9.428702174287347e-05, "loss": 3.7126, "step": 48500 }, { "epoch": 0.7830518862598991, "grad_norm": 1.1730676889419556, "learning_rate": 9.427521793373069e-05, "loss": 4.209, "step": 48550 }, { "epoch": 0.7838583248657279, "grad_norm": 1.3419710397720337, "learning_rate": 9.426340268344425e-05, "loss": 4.1964, "step": 48600 }, { "epoch": 0.7846647634715569, "grad_norm": 2.468380928039551, "learning_rate": 9.425157599506735e-05, "loss": 3.9531, "step": 48650 }, { "epoch": 0.7854712020773859, "grad_norm": 1.7555502653121948, "learning_rate": 9.423973787165607e-05, "loss": 3.8253, "step": 48700 }, { "epoch": 0.7862776406832148, "grad_norm": 2.132018566131592, "learning_rate": 9.422788831626955e-05, "loss": 3.7537, "step": 48750 }, { "epoch": 0.7870840792890437, "grad_norm": 1.427887201309204, "learning_rate": 9.421602733196978e-05, "loss": 3.637, "step": 48800 }, { "epoch": 0.7878905178948726, "grad_norm": 1.145674467086792, "learning_rate": 9.42041549218218e-05, "loss": 3.8905, "step": 48850 }, { "epoch": 0.7886969565007016, "grad_norm": 1.3326349258422852, "learning_rate": 9.41922710888935e-05, "loss": 3.688, "step": 48900 }, { "epoch": 0.7895033951065306, "grad_norm": 2.335531711578369, "learning_rate": 9.418061385320187e-05, "loss": 4.1432, "step": 48950 }, { "epoch": 0.7903098337123595, "grad_norm": 1.2552485466003418, "learning_rate": 9.416870741223122e-05, "loss": 3.853, "step": 49000 }, { "epoch": 0.7911162723181884, "grad_norm": 1.4865189790725708, "learning_rate": 9.415678955764025e-05, "loss": 4.3216, "step": 49050 }, { "epoch": 0.7919227109240173, "grad_norm": 1.2815054655075073, "learning_rate": 9.414486029250867e-05, "loss": 4.0851, "step": 49100 }, { "epoch": 0.7927291495298463, "grad_norm": 1.8521907329559326, "learning_rate": 9.413291961991911e-05, "loss": 3.8439, "step": 49150 }, { "epoch": 0.7935355881356753, "grad_norm": 2.05904221534729, "learning_rate": 9.412096754295714e-05, "loss": 3.8745, "step": 49200 }, { "epoch": 0.7943420267415041, "grad_norm": 3.7223777770996094, "learning_rate": 9.410900406471133e-05, "loss": 3.9499, "step": 49250 }, { "epoch": 0.7951484653473331, "grad_norm": 1.170849084854126, "learning_rate": 9.40970291882731e-05, "loss": 3.5203, "step": 49300 }, { "epoch": 0.795954903953162, "grad_norm": 1.4634156227111816, "learning_rate": 9.408504291673691e-05, "loss": 3.9418, "step": 49350 }, { "epoch": 0.796761342558991, "grad_norm": 2.896477222442627, "learning_rate": 9.407304525320013e-05, "loss": 4.241, "step": 49400 }, { "epoch": 0.79756778116482, "grad_norm": 2.0902223587036133, "learning_rate": 9.406103620076307e-05, "loss": 3.9178, "step": 49450 }, { "epoch": 0.7983742197706488, "grad_norm": 2.3068950176239014, "learning_rate": 9.404901576252898e-05, "loss": 3.7527, "step": 49500 }, { "epoch": 0.7991806583764778, "grad_norm": 1.5382378101348877, "learning_rate": 9.403698394160405e-05, "loss": 3.9847, "step": 49550 }, { "epoch": 0.7999870969823067, "grad_norm": 2.58062744140625, "learning_rate": 9.402494074109743e-05, "loss": 3.8182, "step": 49600 }, { "epoch": 0.8007935355881357, "grad_norm": 2.1531031131744385, "learning_rate": 9.401288616412119e-05, "loss": 3.926, "step": 49650 }, { "epoch": 0.8015999741939646, "grad_norm": 0.9440845847129822, "learning_rate": 9.400082021379035e-05, "loss": 4.1674, "step": 49700 }, { "epoch": 0.8024064127997935, "grad_norm": 2.803377389907837, "learning_rate": 9.398874289322289e-05, "loss": 3.8942, "step": 49750 }, { "epoch": 0.8032128514056225, "grad_norm": 1.548346996307373, "learning_rate": 9.397665420553967e-05, "loss": 4.1091, "step": 49800 }, { "epoch": 0.8040192900114514, "grad_norm": 2.5721352100372314, "learning_rate": 9.396455415386456e-05, "loss": 3.8383, "step": 49850 }, { "epoch": 0.8048257286172804, "grad_norm": 1.9894510507583618, "learning_rate": 9.39524427413243e-05, "loss": 4.09, "step": 49900 }, { "epoch": 0.8056321672231093, "grad_norm": 1.546816349029541, "learning_rate": 9.394031997104861e-05, "loss": 4.3291, "step": 49950 }, { "epoch": 0.8064386058289382, "grad_norm": 2.527470827102661, "learning_rate": 9.392818584617013e-05, "loss": 4.17, "step": 50000 }, { "epoch": 0.8072450444347672, "grad_norm": 1.9979604482650757, "learning_rate": 9.391604036982443e-05, "loss": 3.8669, "step": 50050 }, { "epoch": 0.8080514830405962, "grad_norm": 1.7239247560501099, "learning_rate": 9.390388354515003e-05, "loss": 3.9336, "step": 50100 }, { "epoch": 0.808857921646425, "grad_norm": 1.916926383972168, "learning_rate": 9.389171537528836e-05, "loss": 4.0908, "step": 50150 }, { "epoch": 0.809664360252254, "grad_norm": 1.7619861364364624, "learning_rate": 9.38795358633838e-05, "loss": 3.8995, "step": 50200 }, { "epoch": 0.8104707988580829, "grad_norm": 1.305659294128418, "learning_rate": 9.386734501258365e-05, "loss": 4.0671, "step": 50250 }, { "epoch": 0.8112772374639119, "grad_norm": 1.6476457118988037, "learning_rate": 9.385514282603813e-05, "loss": 3.9938, "step": 50300 }, { "epoch": 0.8120836760697409, "grad_norm": 1.464882731437683, "learning_rate": 9.384292930690042e-05, "loss": 4.0871, "step": 50350 }, { "epoch": 0.8128901146755697, "grad_norm": 1.489268183708191, "learning_rate": 9.383094906630614e-05, "loss": 3.7592, "step": 50400 }, { "epoch": 0.8136965532813987, "grad_norm": 4.688929080963135, "learning_rate": 9.381871311794981e-05, "loss": 3.7422, "step": 50450 }, { "epoch": 0.8145029918872276, "grad_norm": 2.236435651779175, "learning_rate": 9.380646584641508e-05, "loss": 3.7834, "step": 50500 }, { "epoch": 0.8153094304930566, "grad_norm": 1.8845456838607788, "learning_rate": 9.379420725486674e-05, "loss": 3.7639, "step": 50550 }, { "epoch": 0.8161158690988854, "grad_norm": 2.081413984298706, "learning_rate": 9.378193734647256e-05, "loss": 3.9822, "step": 50600 }, { "epoch": 0.8169223077047144, "grad_norm": 1.9713027477264404, "learning_rate": 9.376965612440318e-05, "loss": 3.6913, "step": 50650 }, { "epoch": 0.8177287463105434, "grad_norm": 4.624628067016602, "learning_rate": 9.375736359183221e-05, "loss": 3.8833, "step": 50700 }, { "epoch": 0.8185351849163723, "grad_norm": 2.0145013332366943, "learning_rate": 9.374505975193612e-05, "loss": 3.6465, "step": 50750 }, { "epoch": 0.8193416235222013, "grad_norm": 1.887920618057251, "learning_rate": 9.373274460789439e-05, "loss": 3.8937, "step": 50800 }, { "epoch": 0.8201480621280302, "grad_norm": 2.39249324798584, "learning_rate": 9.372041816288933e-05, "loss": 3.9556, "step": 50850 }, { "epoch": 0.8209545007338591, "grad_norm": 1.9669721126556396, "learning_rate": 9.370808042010622e-05, "loss": 3.6855, "step": 50900 }, { "epoch": 0.8217609393396881, "grad_norm": 2.396238327026367, "learning_rate": 9.369573138273328e-05, "loss": 3.8207, "step": 50950 }, { "epoch": 0.822567377945517, "grad_norm": 2.357149600982666, "learning_rate": 9.36833710539616e-05, "loss": 4.1619, "step": 51000 }, { "epoch": 0.822567377945517, "eval_asr_loss": 0.7768327471051475, "eval_loss": 4.096709728240967, "eval_runtime": 15.3953, "eval_samples_per_second": 33.257, "eval_steps_per_second": 11.107, "eval_tts_loss": 6.02406811865994, "step": 51000 }, { "epoch": 0.8233738165513459, "grad_norm": 1.8168729543685913, "learning_rate": 9.36709994369852e-05, "loss": 3.7739, "step": 51050 }, { "epoch": 0.8241802551571749, "grad_norm": 0.5746332406997681, "learning_rate": 9.365861653500103e-05, "loss": 3.8777, "step": 51100 }, { "epoch": 0.8249866937630038, "grad_norm": 1.3401693105697632, "learning_rate": 9.364622235120897e-05, "loss": 3.8472, "step": 51150 }, { "epoch": 0.8257931323688328, "grad_norm": 2.6011173725128174, "learning_rate": 9.363381688881178e-05, "loss": 3.8559, "step": 51200 }, { "epoch": 0.8265995709746617, "grad_norm": 1.9339457750320435, "learning_rate": 9.362140015101515e-05, "loss": 3.9082, "step": 51250 }, { "epoch": 0.8274060095804906, "grad_norm": 1.8378292322158813, "learning_rate": 9.360897214102769e-05, "loss": 4.2278, "step": 51300 }, { "epoch": 0.8282124481863196, "grad_norm": 2.7941839694976807, "learning_rate": 9.359653286206092e-05, "loss": 3.8958, "step": 51350 }, { "epoch": 0.8290188867921485, "grad_norm": 3.1944453716278076, "learning_rate": 9.358408231732925e-05, "loss": 3.9209, "step": 51400 }, { "epoch": 0.8298253253979775, "grad_norm": 0.7250229120254517, "learning_rate": 9.357162051005006e-05, "loss": 4.2676, "step": 51450 }, { "epoch": 0.8306317640038063, "grad_norm": 0.7566168308258057, "learning_rate": 9.355914744344357e-05, "loss": 3.849, "step": 51500 }, { "epoch": 0.8314382026096353, "grad_norm": 1.8186415433883667, "learning_rate": 9.354666312073294e-05, "loss": 3.8698, "step": 51550 }, { "epoch": 0.8322446412154643, "grad_norm": 1.357745885848999, "learning_rate": 9.353416754514425e-05, "loss": 3.9142, "step": 51600 }, { "epoch": 0.8330510798212932, "grad_norm": 2.4761810302734375, "learning_rate": 9.35216607199065e-05, "loss": 4.1002, "step": 51650 }, { "epoch": 0.8338575184271222, "grad_norm": 2.0084316730499268, "learning_rate": 9.350914264825153e-05, "loss": 3.9168, "step": 51700 }, { "epoch": 0.834663957032951, "grad_norm": 1.214763879776001, "learning_rate": 9.349661333341414e-05, "loss": 3.7877, "step": 51750 }, { "epoch": 0.83547039563878, "grad_norm": 3.5741147994995117, "learning_rate": 9.348407277863205e-05, "loss": 3.6102, "step": 51800 }, { "epoch": 0.836276834244609, "grad_norm": 1.464931845664978, "learning_rate": 9.347152098714584e-05, "loss": 3.8814, "step": 51850 }, { "epoch": 0.8370832728504379, "grad_norm": 2.6186602115631104, "learning_rate": 9.345895796219904e-05, "loss": 3.9586, "step": 51900 }, { "epoch": 0.8378897114562668, "grad_norm": 1.3876553773880005, "learning_rate": 9.344638370703803e-05, "loss": 3.6926, "step": 51950 }, { "epoch": 0.8386961500620957, "grad_norm": 2.4390182495117188, "learning_rate": 9.343379822491211e-05, "loss": 3.9507, "step": 52000 }, { "epoch": 0.8395025886679247, "grad_norm": 1.0152770280838013, "learning_rate": 9.342120151907354e-05, "loss": 3.8201, "step": 52050 }, { "epoch": 0.8403090272737537, "grad_norm": 1.5533214807510376, "learning_rate": 9.340859359277737e-05, "loss": 3.9269, "step": 52100 }, { "epoch": 0.8411154658795826, "grad_norm": 1.1705306768417358, "learning_rate": 9.339597444928165e-05, "loss": 3.9453, "step": 52150 }, { "epoch": 0.8419219044854115, "grad_norm": 2.8534092903137207, "learning_rate": 9.338334409184727e-05, "loss": 4.0685, "step": 52200 }, { "epoch": 0.8427283430912405, "grad_norm": 2.613004446029663, "learning_rate": 9.337070252373806e-05, "loss": 4.1054, "step": 52250 }, { "epoch": 0.8435347816970694, "grad_norm": 2.152470588684082, "learning_rate": 9.335804974822068e-05, "loss": 3.7482, "step": 52300 }, { "epoch": 0.8443412203028984, "grad_norm": 1.3588926792144775, "learning_rate": 9.334538576856477e-05, "loss": 3.9617, "step": 52350 }, { "epoch": 0.8451476589087272, "grad_norm": 0.8621000051498413, "learning_rate": 9.333296420140055e-05, "loss": 4.1509, "step": 52400 }, { "epoch": 0.8459540975145562, "grad_norm": 2.4818482398986816, "learning_rate": 9.332027804720762e-05, "loss": 3.8696, "step": 52450 }, { "epoch": 0.8467605361203852, "grad_norm": 2.23970365524292, "learning_rate": 9.330758069863671e-05, "loss": 3.491, "step": 52500 }, { "epoch": 0.8475669747262141, "grad_norm": 1.6761093139648438, "learning_rate": 9.329487215896895e-05, "loss": 4.2651, "step": 52550 }, { "epoch": 0.8483734133320431, "grad_norm": 1.4623833894729614, "learning_rate": 9.328215243148835e-05, "loss": 3.9832, "step": 52600 }, { "epoch": 0.8491798519378719, "grad_norm": 2.0334420204162598, "learning_rate": 9.326942151948179e-05, "loss": 3.7518, "step": 52650 }, { "epoch": 0.8499862905437009, "grad_norm": 1.3543375730514526, "learning_rate": 9.325667942623908e-05, "loss": 3.9137, "step": 52700 }, { "epoch": 0.8507927291495299, "grad_norm": 1.6424089670181274, "learning_rate": 9.324392615505289e-05, "loss": 3.5574, "step": 52750 }, { "epoch": 0.8515991677553588, "grad_norm": 4.262317180633545, "learning_rate": 9.323116170921881e-05, "loss": 3.9487, "step": 52800 }, { "epoch": 0.8524056063611877, "grad_norm": 1.8372902870178223, "learning_rate": 9.321838609203526e-05, "loss": 3.7967, "step": 52850 }, { "epoch": 0.8532120449670166, "grad_norm": 1.760532259941101, "learning_rate": 9.320559930680362e-05, "loss": 4.2446, "step": 52900 }, { "epoch": 0.8540184835728456, "grad_norm": 1.7465105056762695, "learning_rate": 9.319280135682809e-05, "loss": 3.5362, "step": 52950 }, { "epoch": 0.8548249221786746, "grad_norm": 2.0122694969177246, "learning_rate": 9.317999224541578e-05, "loss": 3.9588, "step": 53000 }, { "epoch": 0.8556313607845035, "grad_norm": 1.6101807355880737, "learning_rate": 9.316717197587672e-05, "loss": 3.8732, "step": 53050 }, { "epoch": 0.8564377993903324, "grad_norm": 3.191518545150757, "learning_rate": 9.315434055152377e-05, "loss": 4.3081, "step": 53100 }, { "epoch": 0.8572442379961613, "grad_norm": 2.0701382160186768, "learning_rate": 9.314149797567271e-05, "loss": 3.7061, "step": 53150 }, { "epoch": 0.8580506766019903, "grad_norm": 2.242328405380249, "learning_rate": 9.312864425164215e-05, "loss": 3.5962, "step": 53200 }, { "epoch": 0.8588571152078193, "grad_norm": 1.251584529876709, "learning_rate": 9.311577938275364e-05, "loss": 3.7365, "step": 53250 }, { "epoch": 0.8596635538136481, "grad_norm": 1.8793351650238037, "learning_rate": 9.31029033723316e-05, "loss": 3.8163, "step": 53300 }, { "epoch": 0.8604699924194771, "grad_norm": 2.5559446811676025, "learning_rate": 9.309001622370331e-05, "loss": 4.0074, "step": 53350 }, { "epoch": 0.861276431025306, "grad_norm": 2.08304762840271, "learning_rate": 9.30771179401989e-05, "loss": 3.6962, "step": 53400 }, { "epoch": 0.862082869631135, "grad_norm": 2.394758939743042, "learning_rate": 9.306420852515144e-05, "loss": 3.9964, "step": 53450 }, { "epoch": 0.862889308236964, "grad_norm": 2.2994213104248047, "learning_rate": 9.305128798189686e-05, "loss": 4.0284, "step": 53500 }, { "epoch": 0.8636957468427928, "grad_norm": 2.2024805545806885, "learning_rate": 9.303835631377391e-05, "loss": 3.4712, "step": 53550 }, { "epoch": 0.8645021854486218, "grad_norm": 2.078568935394287, "learning_rate": 9.302541352412431e-05, "loss": 4.0563, "step": 53600 }, { "epoch": 0.8653086240544507, "grad_norm": 2.2940261363983154, "learning_rate": 9.301245961629257e-05, "loss": 3.8152, "step": 53650 }, { "epoch": 0.8661150626602797, "grad_norm": 1.1552437543869019, "learning_rate": 9.299949459362609e-05, "loss": 4.1016, "step": 53700 }, { "epoch": 0.8669215012661086, "grad_norm": 2.04535174369812, "learning_rate": 9.298651845947519e-05, "loss": 3.8885, "step": 53750 }, { "epoch": 0.8677279398719375, "grad_norm": 1.353513479232788, "learning_rate": 9.297353121719302e-05, "loss": 3.8515, "step": 53800 }, { "epoch": 0.8685343784777665, "grad_norm": 2.1947121620178223, "learning_rate": 9.296053287013561e-05, "loss": 3.7521, "step": 53850 }, { "epoch": 0.8693408170835955, "grad_norm": 1.983809471130371, "learning_rate": 9.294752342166184e-05, "loss": 4.0622, "step": 53900 }, { "epoch": 0.8701472556894244, "grad_norm": 1.3683280944824219, "learning_rate": 9.29345028751335e-05, "loss": 4.1577, "step": 53950 }, { "epoch": 0.8709536942952533, "grad_norm": 1.1140321493148804, "learning_rate": 9.292147123391521e-05, "loss": 3.4622, "step": 54000 }, { "epoch": 0.8709536942952533, "eval_asr_loss": 0.7993465272369762, "eval_loss": 4.063399314880371, "eval_runtime": 14.8011, "eval_samples_per_second": 34.592, "eval_steps_per_second": 11.553, "eval_tts_loss": 6.035265262216293, "step": 54000 }, { "epoch": 0.8717601329010822, "grad_norm": 1.8206874132156372, "learning_rate": 9.290842850137446e-05, "loss": 4.1565, "step": 54050 }, { "epoch": 0.8725665715069112, "grad_norm": 1.194786787033081, "learning_rate": 9.289537468088167e-05, "loss": 4.3061, "step": 54100 }, { "epoch": 0.8733730101127402, "grad_norm": 1.8756150007247925, "learning_rate": 9.288230977581001e-05, "loss": 3.7816, "step": 54150 }, { "epoch": 0.874179448718569, "grad_norm": 2.181330442428589, "learning_rate": 9.286923378953563e-05, "loss": 3.9079, "step": 54200 }, { "epoch": 0.874985887324398, "grad_norm": 1.5249792337417603, "learning_rate": 9.285614672543743e-05, "loss": 3.9149, "step": 54250 }, { "epoch": 0.8757923259302269, "grad_norm": 1.4917360544204712, "learning_rate": 9.28430485868973e-05, "loss": 3.7962, "step": 54300 }, { "epoch": 0.8765987645360559, "grad_norm": 3.438863754272461, "learning_rate": 9.28299393772999e-05, "loss": 3.9856, "step": 54350 }, { "epoch": 0.8774052031418849, "grad_norm": 1.812562108039856, "learning_rate": 9.281681910003274e-05, "loss": 3.8257, "step": 54400 }, { "epoch": 0.8782116417477137, "grad_norm": 2.495959997177124, "learning_rate": 9.280368775848628e-05, "loss": 4.0912, "step": 54450 }, { "epoch": 0.8790180803535427, "grad_norm": 1.4950435161590576, "learning_rate": 9.279054535605376e-05, "loss": 3.4991, "step": 54500 }, { "epoch": 0.8798245189593716, "grad_norm": 1.361709475517273, "learning_rate": 9.27773918961313e-05, "loss": 3.9666, "step": 54550 }, { "epoch": 0.8806309575652006, "grad_norm": 1.4355372190475464, "learning_rate": 9.27642273821179e-05, "loss": 3.8696, "step": 54600 }, { "epoch": 0.8814373961710295, "grad_norm": 1.8107373714447021, "learning_rate": 9.27510518174154e-05, "loss": 4.0351, "step": 54650 }, { "epoch": 0.8822438347768584, "grad_norm": 1.526053547859192, "learning_rate": 9.273786520542844e-05, "loss": 3.5326, "step": 54700 }, { "epoch": 0.8830502733826874, "grad_norm": 1.7716056108474731, "learning_rate": 9.272466754956463e-05, "loss": 4.0023, "step": 54750 }, { "epoch": 0.8838567119885163, "grad_norm": 2.143021821975708, "learning_rate": 9.271145885323435e-05, "loss": 4.1698, "step": 54800 }, { "epoch": 0.8846631505943453, "grad_norm": 1.416943907737732, "learning_rate": 9.269823911985085e-05, "loss": 4.1545, "step": 54850 }, { "epoch": 0.8854695892001742, "grad_norm": 1.725667119026184, "learning_rate": 9.268500835283024e-05, "loss": 3.7823, "step": 54900 }, { "epoch": 0.8862760278060031, "grad_norm": 1.573408842086792, "learning_rate": 9.267176655559147e-05, "loss": 4.1186, "step": 54950 }, { "epoch": 0.8870824664118321, "grad_norm": 1.5016199350357056, "learning_rate": 9.265851373155637e-05, "loss": 3.8804, "step": 55000 }, { "epoch": 0.887888905017661, "grad_norm": 2.705294609069824, "learning_rate": 9.264524988414958e-05, "loss": 4.0776, "step": 55050 }, { "epoch": 0.8886953436234899, "grad_norm": 1.623266577720642, "learning_rate": 9.263197501679858e-05, "loss": 4.041, "step": 55100 }, { "epoch": 0.8895017822293189, "grad_norm": 1.0465036630630493, "learning_rate": 9.261868913293378e-05, "loss": 3.5968, "step": 55150 }, { "epoch": 0.8903082208351478, "grad_norm": 2.454932451248169, "learning_rate": 9.260539223598837e-05, "loss": 3.7386, "step": 55200 }, { "epoch": 0.8911146594409768, "grad_norm": 1.288687825202942, "learning_rate": 9.259208432939836e-05, "loss": 4.0238, "step": 55250 }, { "epoch": 0.8919210980468057, "grad_norm": 4.057321548461914, "learning_rate": 9.257876541660267e-05, "loss": 4.0917, "step": 55300 }, { "epoch": 0.8927275366526346, "grad_norm": 2.4807841777801514, "learning_rate": 9.2565435501043e-05, "loss": 3.8068, "step": 55350 }, { "epoch": 0.8935339752584636, "grad_norm": 2.479290008544922, "learning_rate": 9.255209458616398e-05, "loss": 3.9573, "step": 55400 }, { "epoch": 0.8943404138642925, "grad_norm": 1.4316256046295166, "learning_rate": 9.253874267541302e-05, "loss": 3.7578, "step": 55450 }, { "epoch": 0.8951468524701215, "grad_norm": 1.3951455354690552, "learning_rate": 9.252537977224035e-05, "loss": 3.9678, "step": 55500 }, { "epoch": 0.8959532910759503, "grad_norm": 3.9966070652008057, "learning_rate": 9.25120058800991e-05, "loss": 3.9496, "step": 55550 }, { "epoch": 0.8967597296817793, "grad_norm": 2.231217622756958, "learning_rate": 9.249862100244521e-05, "loss": 4.0736, "step": 55600 }, { "epoch": 0.8975661682876083, "grad_norm": 1.635345697402954, "learning_rate": 9.248522514273745e-05, "loss": 3.9145, "step": 55650 }, { "epoch": 0.8983726068934372, "grad_norm": 1.356836199760437, "learning_rate": 9.247181830443745e-05, "loss": 3.8725, "step": 55700 }, { "epoch": 0.8991790454992662, "grad_norm": 1.3725659847259521, "learning_rate": 9.245840049100966e-05, "loss": 3.8331, "step": 55750 }, { "epoch": 0.899985484105095, "grad_norm": 0.6989811062812805, "learning_rate": 9.244550906793727e-05, "loss": 3.8761, "step": 55800 }, { "epoch": 0.900791922710924, "grad_norm": 1.2685366868972778, "learning_rate": 9.243206975331959e-05, "loss": 4.1927, "step": 55850 }, { "epoch": 0.901598361316753, "grad_norm": 2.1454267501831055, "learning_rate": 9.241861947384552e-05, "loss": 3.9459, "step": 55900 }, { "epoch": 0.9024047999225819, "grad_norm": 0.8860172629356384, "learning_rate": 9.240515823299075e-05, "loss": 3.6836, "step": 55950 }, { "epoch": 0.9032112385284108, "grad_norm": 3.379371166229248, "learning_rate": 9.23916860342338e-05, "loss": 3.9944, "step": 56000 }, { "epoch": 0.9040176771342397, "grad_norm": 1.3930600881576538, "learning_rate": 9.237820288105603e-05, "loss": 3.401, "step": 56050 }, { "epoch": 0.9048241157400687, "grad_norm": 1.7948726415634155, "learning_rate": 9.236470877694158e-05, "loss": 4.0875, "step": 56100 }, { "epoch": 0.9056305543458977, "grad_norm": 1.8862905502319336, "learning_rate": 9.23512037253775e-05, "loss": 4.0634, "step": 56150 }, { "epoch": 0.9064369929517266, "grad_norm": 2.141082763671875, "learning_rate": 9.233768772985361e-05, "loss": 3.9783, "step": 56200 }, { "epoch": 0.9072434315575555, "grad_norm": 2.0120999813079834, "learning_rate": 9.232416079386255e-05, "loss": 3.8297, "step": 56250 }, { "epoch": 0.9080498701633845, "grad_norm": 2.376425266265869, "learning_rate": 9.231062292089984e-05, "loss": 3.503, "step": 56300 }, { "epoch": 0.9088563087692134, "grad_norm": 0.9411992430686951, "learning_rate": 9.229707411446378e-05, "loss": 3.644, "step": 56350 }, { "epoch": 0.9096627473750424, "grad_norm": 1.9797859191894531, "learning_rate": 9.228351437805551e-05, "loss": 3.9771, "step": 56400 }, { "epoch": 0.9104691859808712, "grad_norm": 2.0308303833007812, "learning_rate": 9.226994371517903e-05, "loss": 3.9891, "step": 56450 }, { "epoch": 0.9112756245867002, "grad_norm": 2.4685399532318115, "learning_rate": 9.225636212934109e-05, "loss": 4.2307, "step": 56500 }, { "epoch": 0.9120820631925292, "grad_norm": 2.00228214263916, "learning_rate": 9.224276962405131e-05, "loss": 4.0254, "step": 56550 }, { "epoch": 0.9128885017983581, "grad_norm": 2.4734489917755127, "learning_rate": 9.222916620282216e-05, "loss": 3.922, "step": 56600 }, { "epoch": 0.9136949404041871, "grad_norm": 1.6077553033828735, "learning_rate": 9.221555186916886e-05, "loss": 3.8695, "step": 56650 }, { "epoch": 0.9145013790100159, "grad_norm": 1.4017868041992188, "learning_rate": 9.220192662660949e-05, "loss": 4.1232, "step": 56700 }, { "epoch": 0.9153078176158449, "grad_norm": 2.2608869075775146, "learning_rate": 9.218829047866494e-05, "loss": 3.8422, "step": 56750 }, { "epoch": 0.9161142562216739, "grad_norm": 0.9062301516532898, "learning_rate": 9.217464342885894e-05, "loss": 3.7004, "step": 56800 }, { "epoch": 0.9169206948275028, "grad_norm": 1.335174798965454, "learning_rate": 9.216098548071801e-05, "loss": 3.8427, "step": 56850 }, { "epoch": 0.9177271334333317, "grad_norm": 2.1585700511932373, "learning_rate": 9.21473166377715e-05, "loss": 3.8167, "step": 56900 }, { "epoch": 0.9185335720391606, "grad_norm": 1.4338594675064087, "learning_rate": 9.213363690355154e-05, "loss": 3.9614, "step": 56950 }, { "epoch": 0.9193400106449896, "grad_norm": 1.4069015979766846, "learning_rate": 9.211994628159317e-05, "loss": 3.8141, "step": 57000 }, { "epoch": 0.9193400106449896, "eval_asr_loss": 0.802518399731879, "eval_loss": 4.040905475616455, "eval_runtime": 14.707, "eval_samples_per_second": 34.813, "eval_steps_per_second": 11.627, "eval_tts_loss": 6.034823876925827, "step": 57000 }, { "epoch": 0.9201464492508186, "grad_norm": 1.434632658958435, "learning_rate": 9.210624477543411e-05, "loss": 3.3863, "step": 57050 }, { "epoch": 0.9209528878566475, "grad_norm": 2.30854868888855, "learning_rate": 9.209253238861502e-05, "loss": 3.9079, "step": 57100 }, { "epoch": 0.9217593264624764, "grad_norm": 2.066159963607788, "learning_rate": 9.207880912467925e-05, "loss": 3.7538, "step": 57150 }, { "epoch": 0.9225657650683053, "grad_norm": 3.556805372238159, "learning_rate": 9.206507498717308e-05, "loss": 4.5109, "step": 57200 }, { "epoch": 0.9233722036741343, "grad_norm": 1.8949222564697266, "learning_rate": 9.205132997964553e-05, "loss": 4.2073, "step": 57250 }, { "epoch": 0.9241786422799633, "grad_norm": 2.9166855812072754, "learning_rate": 9.203757410564842e-05, "loss": 3.812, "step": 57300 }, { "epoch": 0.9249850808857921, "grad_norm": 1.8220806121826172, "learning_rate": 9.202380736873642e-05, "loss": 3.7369, "step": 57350 }, { "epoch": 0.9257915194916211, "grad_norm": 2.030690908432007, "learning_rate": 9.2010029772467e-05, "loss": 4.144, "step": 57400 }, { "epoch": 0.92659795809745, "grad_norm": 2.2305502891540527, "learning_rate": 9.19962413204004e-05, "loss": 4.0017, "step": 57450 }, { "epoch": 0.927404396703279, "grad_norm": 3.2708723545074463, "learning_rate": 9.198244201609968e-05, "loss": 3.7619, "step": 57500 }, { "epoch": 0.9282108353091079, "grad_norm": 1.5857239961624146, "learning_rate": 9.196863186313074e-05, "loss": 3.9691, "step": 57550 }, { "epoch": 0.9290172739149368, "grad_norm": 1.4099088907241821, "learning_rate": 9.195481086506225e-05, "loss": 3.8978, "step": 57600 }, { "epoch": 0.9298237125207658, "grad_norm": 2.7035984992980957, "learning_rate": 9.194097902546568e-05, "loss": 3.8467, "step": 57650 }, { "epoch": 0.9306301511265948, "grad_norm": 0.9907490015029907, "learning_rate": 9.192713634791531e-05, "loss": 3.568, "step": 57700 }, { "epoch": 0.9314365897324237, "grad_norm": 1.3037341833114624, "learning_rate": 9.191328283598826e-05, "loss": 3.5235, "step": 57750 }, { "epoch": 0.9322430283382526, "grad_norm": 1.6086373329162598, "learning_rate": 9.189941849326436e-05, "loss": 3.6997, "step": 57800 }, { "epoch": 0.9330494669440815, "grad_norm": 2.012066602706909, "learning_rate": 9.188554332332629e-05, "loss": 3.6956, "step": 57850 }, { "epoch": 0.9338559055499105, "grad_norm": 1.181842565536499, "learning_rate": 9.187165732975958e-05, "loss": 3.8707, "step": 57900 }, { "epoch": 0.9346623441557395, "grad_norm": 1.3590315580368042, "learning_rate": 9.185776051615245e-05, "loss": 4.2307, "step": 57950 }, { "epoch": 0.9354687827615683, "grad_norm": 2.5573291778564453, "learning_rate": 9.1843852886096e-05, "loss": 4.1767, "step": 58000 }, { "epoch": 0.9362752213673973, "grad_norm": 1.4362449645996094, "learning_rate": 9.182993444318409e-05, "loss": 3.9106, "step": 58050 }, { "epoch": 0.9370816599732262, "grad_norm": 1.092568278312683, "learning_rate": 9.18160051910134e-05, "loss": 3.8131, "step": 58100 }, { "epoch": 0.9378880985790552, "grad_norm": 1.4488219022750854, "learning_rate": 9.180206513318332e-05, "loss": 3.6785, "step": 58150 }, { "epoch": 0.9386945371848842, "grad_norm": 1.7660725116729736, "learning_rate": 9.178811427329617e-05, "loss": 3.7885, "step": 58200 }, { "epoch": 0.939500975790713, "grad_norm": 1.2934764623641968, "learning_rate": 9.177415261495693e-05, "loss": 3.9415, "step": 58250 }, { "epoch": 0.940307414396542, "grad_norm": 1.1474610567092896, "learning_rate": 9.176018016177347e-05, "loss": 3.8456, "step": 58300 }, { "epoch": 0.9411138530023709, "grad_norm": 1.1843050718307495, "learning_rate": 9.174619691735638e-05, "loss": 3.8715, "step": 58350 }, { "epoch": 0.9419202916081999, "grad_norm": 2.8808796405792236, "learning_rate": 9.173220288531907e-05, "loss": 3.8184, "step": 58400 }, { "epoch": 0.9427267302140288, "grad_norm": 1.8526114225387573, "learning_rate": 9.171819806927775e-05, "loss": 3.848, "step": 58450 }, { "epoch": 0.9435331688198577, "grad_norm": 1.1822868585586548, "learning_rate": 9.170418247285139e-05, "loss": 4.0661, "step": 58500 }, { "epoch": 0.9443396074256867, "grad_norm": 0.6593601703643799, "learning_rate": 9.169015609966175e-05, "loss": 4.0325, "step": 58550 }, { "epoch": 0.9451460460315156, "grad_norm": 1.1227622032165527, "learning_rate": 9.167611895333337e-05, "loss": 4.0502, "step": 58600 }, { "epoch": 0.9459524846373446, "grad_norm": 1.0978533029556274, "learning_rate": 9.166207103749359e-05, "loss": 3.914, "step": 58650 }, { "epoch": 0.9467589232431735, "grad_norm": 2.1706528663635254, "learning_rate": 9.164801235577253e-05, "loss": 3.6062, "step": 58700 }, { "epoch": 0.9475653618490024, "grad_norm": 3.1869921684265137, "learning_rate": 9.163394291180312e-05, "loss": 3.7632, "step": 58750 }, { "epoch": 0.9483718004548314, "grad_norm": 1.3705198764801025, "learning_rate": 9.1619862709221e-05, "loss": 3.5777, "step": 58800 }, { "epoch": 0.9491782390606603, "grad_norm": 1.191117286682129, "learning_rate": 9.160577175166464e-05, "loss": 3.6073, "step": 58850 }, { "epoch": 0.9499846776664892, "grad_norm": 1.9281792640686035, "learning_rate": 9.159167004277528e-05, "loss": 3.8723, "step": 58900 }, { "epoch": 0.9507911162723182, "grad_norm": 1.4402707815170288, "learning_rate": 9.157755758619695e-05, "loss": 3.7574, "step": 58950 }, { "epoch": 0.9515975548781471, "grad_norm": 1.5028876066207886, "learning_rate": 9.156343438557642e-05, "loss": 3.6589, "step": 59000 }, { "epoch": 0.9524039934839761, "grad_norm": 2.2400081157684326, "learning_rate": 9.154930044456329e-05, "loss": 3.6171, "step": 59050 }, { "epoch": 0.953210432089805, "grad_norm": 2.5251247882843018, "learning_rate": 9.153515576680988e-05, "loss": 4.0869, "step": 59100 }, { "epoch": 0.9540168706956339, "grad_norm": 2.7705583572387695, "learning_rate": 9.152100035597134e-05, "loss": 3.8465, "step": 59150 }, { "epoch": 0.9548233093014629, "grad_norm": 2.7229108810424805, "learning_rate": 9.150683421570553e-05, "loss": 3.9602, "step": 59200 }, { "epoch": 0.9556297479072918, "grad_norm": 1.7021418809890747, "learning_rate": 9.149265734967313e-05, "loss": 3.8551, "step": 59250 }, { "epoch": 0.9564361865131208, "grad_norm": 1.3496074676513672, "learning_rate": 9.14784697615376e-05, "loss": 3.8613, "step": 59300 }, { "epoch": 0.9572426251189496, "grad_norm": 1.5890028476715088, "learning_rate": 9.146427145496515e-05, "loss": 3.746, "step": 59350 }, { "epoch": 0.9580490637247786, "grad_norm": 1.7450709342956543, "learning_rate": 9.145006243362473e-05, "loss": 4.0456, "step": 59400 }, { "epoch": 0.9588555023306076, "grad_norm": 2.2836358547210693, "learning_rate": 9.14358427011881e-05, "loss": 3.4414, "step": 59450 }, { "epoch": 0.9596619409364365, "grad_norm": 2.2296597957611084, "learning_rate": 9.142161226132977e-05, "loss": 4.2831, "step": 59500 }, { "epoch": 0.9604683795422655, "grad_norm": 1.2959296703338623, "learning_rate": 9.140737111772706e-05, "loss": 4.09, "step": 59550 }, { "epoch": 0.9612748181480943, "grad_norm": 0.9999203085899353, "learning_rate": 9.139311927405996e-05, "loss": 3.6267, "step": 59600 }, { "epoch": 0.9620812567539233, "grad_norm": 1.62959623336792, "learning_rate": 9.137885673401134e-05, "loss": 3.8423, "step": 59650 }, { "epoch": 0.9628876953597523, "grad_norm": 2.1384541988372803, "learning_rate": 9.136458350126673e-05, "loss": 3.8628, "step": 59700 }, { "epoch": 0.9636941339655812, "grad_norm": 1.3596155643463135, "learning_rate": 9.135029957951451e-05, "loss": 3.6401, "step": 59750 }, { "epoch": 0.9645005725714101, "grad_norm": 2.1480040550231934, "learning_rate": 9.133600497244576e-05, "loss": 3.923, "step": 59800 }, { "epoch": 0.965307011177239, "grad_norm": 2.262998104095459, "learning_rate": 9.132169968375436e-05, "loss": 3.9753, "step": 59850 }, { "epoch": 0.966113449783068, "grad_norm": 1.4706131219863892, "learning_rate": 9.130738371713692e-05, "loss": 4.0317, "step": 59900 }, { "epoch": 0.966919888388897, "grad_norm": 1.6120691299438477, "learning_rate": 9.129305707629282e-05, "loss": 3.4057, "step": 59950 }, { "epoch": 0.9677263269947259, "grad_norm": 1.462624430656433, "learning_rate": 9.127871976492422e-05, "loss": 3.5948, "step": 60000 }, { "epoch": 0.9677263269947259, "eval_asr_loss": 0.8046618353090987, "eval_loss": 3.9864399433135986, "eval_runtime": 14.901, "eval_samples_per_second": 34.36, "eval_steps_per_second": 11.476, "eval_tts_loss": 5.998012083371329, "step": 60000 }, { "epoch": 0.9685327656005548, "grad_norm": 2.8822438716888428, "learning_rate": 9.126437178673601e-05, "loss": 3.6478, "step": 60050 }, { "epoch": 0.9693392042063838, "grad_norm": 1.895029067993164, "learning_rate": 9.125001314543587e-05, "loss": 4.0254, "step": 60100 }, { "epoch": 0.9701456428122127, "grad_norm": 2.5896832942962646, "learning_rate": 9.123564384473415e-05, "loss": 4.1747, "step": 60150 }, { "epoch": 0.9709520814180417, "grad_norm": 1.6556954383850098, "learning_rate": 9.122126388834409e-05, "loss": 3.8101, "step": 60200 }, { "epoch": 0.9717585200238705, "grad_norm": 1.4465107917785645, "learning_rate": 9.120687327998157e-05, "loss": 3.3938, "step": 60250 }, { "epoch": 0.9725649586296995, "grad_norm": 2.358011245727539, "learning_rate": 9.119247202336525e-05, "loss": 3.8702, "step": 60300 }, { "epoch": 0.9733713972355285, "grad_norm": 2.1396501064300537, "learning_rate": 9.117806012221659e-05, "loss": 4.0532, "step": 60350 }, { "epoch": 0.9741778358413574, "grad_norm": 2.36503529548645, "learning_rate": 9.116363758025973e-05, "loss": 3.9132, "step": 60400 }, { "epoch": 0.9749842744471864, "grad_norm": 1.5044091939926147, "learning_rate": 9.11492044012216e-05, "loss": 3.9943, "step": 60450 }, { "epoch": 0.9757907130530152, "grad_norm": 2.0338516235351562, "learning_rate": 9.11347605888319e-05, "loss": 4.0265, "step": 60500 }, { "epoch": 0.9765971516588442, "grad_norm": 2.3932907581329346, "learning_rate": 9.112030614682304e-05, "loss": 3.7576, "step": 60550 }, { "epoch": 0.9774035902646732, "grad_norm": 1.521772861480713, "learning_rate": 9.110584107893017e-05, "loss": 3.7796, "step": 60600 }, { "epoch": 0.9782100288705021, "grad_norm": 3.0539138317108154, "learning_rate": 9.109136538889121e-05, "loss": 3.8878, "step": 60650 }, { "epoch": 0.979016467476331, "grad_norm": 0.7148169279098511, "learning_rate": 9.107687908044683e-05, "loss": 3.9899, "step": 60700 }, { "epoch": 0.9798229060821599, "grad_norm": 1.7796761989593506, "learning_rate": 9.106238215734042e-05, "loss": 3.9122, "step": 60750 }, { "epoch": 0.9806293446879889, "grad_norm": 1.0807075500488281, "learning_rate": 9.104787462331812e-05, "loss": 3.495, "step": 60800 }, { "epoch": 0.9814357832938179, "grad_norm": 1.9213134050369263, "learning_rate": 9.103335648212883e-05, "loss": 4.0264, "step": 60850 }, { "epoch": 0.9822422218996468, "grad_norm": 1.5700064897537231, "learning_rate": 9.101882773752417e-05, "loss": 3.6574, "step": 60900 }, { "epoch": 0.9830486605054757, "grad_norm": 2.700054407119751, "learning_rate": 9.100428839325852e-05, "loss": 3.982, "step": 60950 }, { "epoch": 0.9838550991113046, "grad_norm": 1.1245232820510864, "learning_rate": 9.098973845308897e-05, "loss": 3.7064, "step": 61000 }, { "epoch": 0.9846615377171336, "grad_norm": 1.6349071264266968, "learning_rate": 9.097517792077536e-05, "loss": 3.9735, "step": 61050 }, { "epoch": 0.9854679763229626, "grad_norm": 2.0650248527526855, "learning_rate": 9.09606068000803e-05, "loss": 3.904, "step": 61100 }, { "epoch": 0.9862744149287914, "grad_norm": 1.6910192966461182, "learning_rate": 9.09460250947691e-05, "loss": 4.0517, "step": 61150 }, { "epoch": 0.9870808535346204, "grad_norm": 1.439247727394104, "learning_rate": 9.093143280860978e-05, "loss": 3.8044, "step": 61200 }, { "epoch": 0.9878872921404493, "grad_norm": 1.8605902194976807, "learning_rate": 9.091682994537318e-05, "loss": 3.582, "step": 61250 }, { "epoch": 0.9886937307462783, "grad_norm": 2.853280782699585, "learning_rate": 9.090221650883279e-05, "loss": 4.079, "step": 61300 }, { "epoch": 0.9895001693521073, "grad_norm": 2.0609419345855713, "learning_rate": 9.088759250276487e-05, "loss": 4.0003, "step": 61350 }, { "epoch": 0.9903066079579361, "grad_norm": 1.7866257429122925, "learning_rate": 9.087295793094837e-05, "loss": 3.7343, "step": 61400 }, { "epoch": 0.9911130465637651, "grad_norm": 2.8632779121398926, "learning_rate": 9.085831279716507e-05, "loss": 4.1644, "step": 61450 }, { "epoch": 0.991919485169594, "grad_norm": 1.753532886505127, "learning_rate": 9.084365710519937e-05, "loss": 3.7438, "step": 61500 }, { "epoch": 0.992725923775423, "grad_norm": 1.3242493867874146, "learning_rate": 9.082899085883845e-05, "loss": 3.529, "step": 61550 }, { "epoch": 0.9935323623812519, "grad_norm": 1.4327284097671509, "learning_rate": 9.081431406187221e-05, "loss": 4.1134, "step": 61600 }, { "epoch": 0.9943388009870808, "grad_norm": 1.6317732334136963, "learning_rate": 9.079962671809328e-05, "loss": 3.8372, "step": 61650 }, { "epoch": 0.9951452395929098, "grad_norm": 1.931650161743164, "learning_rate": 9.078492883129701e-05, "loss": 3.7848, "step": 61700 }, { "epoch": 0.9959516781987388, "grad_norm": 1.8680939674377441, "learning_rate": 9.077022040528148e-05, "loss": 4.0012, "step": 61750 }, { "epoch": 0.9967581168045677, "grad_norm": 0.8529523611068726, "learning_rate": 9.075550144384749e-05, "loss": 4.084, "step": 61800 }, { "epoch": 0.9975645554103966, "grad_norm": 0.6939648389816284, "learning_rate": 9.074077195079855e-05, "loss": 3.7038, "step": 61850 }, { "epoch": 0.9983709940162255, "grad_norm": 2.364664077758789, "learning_rate": 9.07260319299409e-05, "loss": 3.7952, "step": 61900 }, { "epoch": 0.9991774326220545, "grad_norm": 1.424089789390564, "learning_rate": 9.071128138508352e-05, "loss": 4.1309, "step": 61950 }, { "epoch": 0.9999838712278835, "grad_norm": 2.0342156887054443, "learning_rate": 9.069652032003809e-05, "loss": 3.9936, "step": 62000 }, { "epoch": 1.0007903098337123, "grad_norm": 1.1257814168930054, "learning_rate": 9.0681748738619e-05, "loss": 3.6482, "step": 62050 }, { "epoch": 1.0015967484395414, "grad_norm": 1.275170087814331, "learning_rate": 9.066696664464337e-05, "loss": 4.026, "step": 62100 }, { "epoch": 1.0024031870453702, "grad_norm": 3.228968381881714, "learning_rate": 9.065217404193107e-05, "loss": 3.8183, "step": 62150 }, { "epoch": 1.003209625651199, "grad_norm": 2.117586374282837, "learning_rate": 9.063737093430461e-05, "loss": 3.8636, "step": 62200 }, { "epoch": 1.0040160642570282, "grad_norm": 2.0225231647491455, "learning_rate": 9.062255732558926e-05, "loss": 3.9442, "step": 62250 }, { "epoch": 1.004822502862857, "grad_norm": 2.2714576721191406, "learning_rate": 9.060773321961302e-05, "loss": 3.9509, "step": 62300 }, { "epoch": 1.005628941468686, "grad_norm": 4.561624050140381, "learning_rate": 9.059289862020658e-05, "loss": 4.0686, "step": 62350 }, { "epoch": 1.006435380074515, "grad_norm": 0.7048660516738892, "learning_rate": 9.057805353120333e-05, "loss": 4.0666, "step": 62400 }, { "epoch": 1.0072418186803438, "grad_norm": 1.9683712720870972, "learning_rate": 9.056319795643939e-05, "loss": 3.8313, "step": 62450 }, { "epoch": 1.0080482572861729, "grad_norm": 0.9827104210853577, "learning_rate": 9.054833189975357e-05, "loss": 3.7298, "step": 62500 }, { "epoch": 1.0088546958920017, "grad_norm": 2.0851900577545166, "learning_rate": 9.053345536498743e-05, "loss": 4.1029, "step": 62550 }, { "epoch": 1.0096611344978306, "grad_norm": 0.9334834814071655, "learning_rate": 9.051856835598518e-05, "loss": 3.4814, "step": 62600 }, { "epoch": 1.0104675731036596, "grad_norm": 0.9352619647979736, "learning_rate": 9.050367087659379e-05, "loss": 3.9743, "step": 62650 }, { "epoch": 1.0112740117094885, "grad_norm": 2.811934471130371, "learning_rate": 9.048876293066293e-05, "loss": 3.9196, "step": 62700 }, { "epoch": 1.0120804503153176, "grad_norm": 2.484410047531128, "learning_rate": 9.04738445220449e-05, "loss": 4.0525, "step": 62750 }, { "epoch": 1.0128868889211464, "grad_norm": 2.095379590988159, "learning_rate": 9.045891565459481e-05, "loss": 3.9316, "step": 62800 }, { "epoch": 1.0136933275269753, "grad_norm": 1.5643795728683472, "learning_rate": 9.044397633217041e-05, "loss": 3.7923, "step": 62850 }, { "epoch": 1.0144997661328043, "grad_norm": 2.079554557800293, "learning_rate": 9.042902655863215e-05, "loss": 3.991, "step": 62900 }, { "epoch": 1.0153062047386332, "grad_norm": 1.7057164907455444, "learning_rate": 9.041406633784322e-05, "loss": 3.6764, "step": 62950 }, { "epoch": 1.0161126433444623, "grad_norm": 2.3314554691314697, "learning_rate": 9.039909567366947e-05, "loss": 3.9501, "step": 63000 }, { "epoch": 1.0161126433444623, "eval_asr_loss": 0.817902892679416, "eval_loss": 3.991797685623169, "eval_runtime": 15.1813, "eval_samples_per_second": 33.726, "eval_steps_per_second": 11.264, "eval_tts_loss": 6.087261374568878, "step": 63000 }, { "epoch": 1.0169190819502911, "grad_norm": 1.5511826276779175, "learning_rate": 9.038411456997946e-05, "loss": 3.8941, "step": 63050 }, { "epoch": 1.01772552055612, "grad_norm": 1.5903561115264893, "learning_rate": 9.036912303064446e-05, "loss": 3.8921, "step": 63100 }, { "epoch": 1.018531959161949, "grad_norm": 0.9465925693511963, "learning_rate": 9.035412105953842e-05, "loss": 3.7486, "step": 63150 }, { "epoch": 1.019338397767778, "grad_norm": 1.4300538301467896, "learning_rate": 9.033910866053802e-05, "loss": 3.8696, "step": 63200 }, { "epoch": 1.020144836373607, "grad_norm": 1.845852017402649, "learning_rate": 9.03240858375226e-05, "loss": 3.8382, "step": 63250 }, { "epoch": 1.0209512749794358, "grad_norm": 1.676626443862915, "learning_rate": 9.030905259437416e-05, "loss": 3.7814, "step": 63300 }, { "epoch": 1.0217577135852647, "grad_norm": 2.1198408603668213, "learning_rate": 9.029400893497749e-05, "loss": 3.9992, "step": 63350 }, { "epoch": 1.0225641521910938, "grad_norm": 2.0301709175109863, "learning_rate": 9.027895486321998e-05, "loss": 3.9466, "step": 63400 }, { "epoch": 1.0233705907969226, "grad_norm": 1.4008840322494507, "learning_rate": 9.026389038299177e-05, "loss": 4.0138, "step": 63450 }, { "epoch": 1.0241770294027515, "grad_norm": 2.5171308517456055, "learning_rate": 9.024881549818567e-05, "loss": 3.7501, "step": 63500 }, { "epoch": 1.0249834680085805, "grad_norm": 2.461249351501465, "learning_rate": 9.023373021269714e-05, "loss": 3.6278, "step": 63550 }, { "epoch": 1.0257899066144094, "grad_norm": 1.333937168121338, "learning_rate": 9.02186345304244e-05, "loss": 4.3199, "step": 63600 }, { "epoch": 1.0265963452202385, "grad_norm": 2.738095998764038, "learning_rate": 9.020352845526832e-05, "loss": 3.5527, "step": 63650 }, { "epoch": 1.0274027838260673, "grad_norm": 1.1649665832519531, "learning_rate": 9.018841199113244e-05, "loss": 3.7526, "step": 63700 }, { "epoch": 1.0282092224318962, "grad_norm": 1.9158024787902832, "learning_rate": 9.0173285141923e-05, "loss": 3.9562, "step": 63750 }, { "epoch": 1.0290156610377252, "grad_norm": 3.4876673221588135, "learning_rate": 9.015845075786652e-05, "loss": 3.5772, "step": 63800 }, { "epoch": 1.029822099643554, "grad_norm": 0.8455021381378174, "learning_rate": 9.014330335774614e-05, "loss": 3.736, "step": 63850 }, { "epoch": 1.0306285382493832, "grad_norm": 1.3714486360549927, "learning_rate": 9.01281455842087e-05, "loss": 3.7547, "step": 63900 }, { "epoch": 1.031434976855212, "grad_norm": 1.1116451025009155, "learning_rate": 9.011297744117113e-05, "loss": 3.6328, "step": 63950 }, { "epoch": 1.0322414154610409, "grad_norm": 1.2208380699157715, "learning_rate": 9.009779893255304e-05, "loss": 3.8891, "step": 64000 }, { "epoch": 1.03304785406687, "grad_norm": 2.4203615188598633, "learning_rate": 9.008261006227665e-05, "loss": 3.5001, "step": 64050 }, { "epoch": 1.0338542926726988, "grad_norm": 1.931107759475708, "learning_rate": 9.006741083426695e-05, "loss": 3.9117, "step": 64100 }, { "epoch": 1.0346607312785279, "grad_norm": 1.231191873550415, "learning_rate": 9.005220125245158e-05, "loss": 3.4782, "step": 64150 }, { "epoch": 1.0354671698843567, "grad_norm": 0.7901305556297302, "learning_rate": 9.003698132076082e-05, "loss": 3.7488, "step": 64200 }, { "epoch": 1.0362736084901856, "grad_norm": 1.774398922920227, "learning_rate": 9.002175104312763e-05, "loss": 3.7117, "step": 64250 }, { "epoch": 1.0370800470960146, "grad_norm": 1.4730985164642334, "learning_rate": 9.00065104234877e-05, "loss": 3.7165, "step": 64300 }, { "epoch": 1.0378864857018435, "grad_norm": 1.7323822975158691, "learning_rate": 8.999125946577933e-05, "loss": 3.7852, "step": 64350 }, { "epoch": 1.0386929243076723, "grad_norm": 2.97774338722229, "learning_rate": 8.997599817394352e-05, "loss": 3.508, "step": 64400 }, { "epoch": 1.0394993629135014, "grad_norm": 2.03074049949646, "learning_rate": 8.996072655192393e-05, "loss": 3.927, "step": 64450 }, { "epoch": 1.0403058015193303, "grad_norm": 2.189884901046753, "learning_rate": 8.994544460366692e-05, "loss": 3.9014, "step": 64500 }, { "epoch": 1.0411122401251593, "grad_norm": 1.8007670640945435, "learning_rate": 8.993015233312146e-05, "loss": 3.7454, "step": 64550 }, { "epoch": 1.0419186787309882, "grad_norm": 2.0331923961639404, "learning_rate": 8.9915155897111e-05, "loss": 3.5802, "step": 64600 }, { "epoch": 1.042725117336817, "grad_norm": 1.5598232746124268, "learning_rate": 8.989984320009521e-05, "loss": 4.3209, "step": 64650 }, { "epoch": 1.0435315559426461, "grad_norm": 1.912048578262329, "learning_rate": 8.988452019257484e-05, "loss": 3.9792, "step": 64700 }, { "epoch": 1.044337994548475, "grad_norm": 0.8821032047271729, "learning_rate": 8.986918687850948e-05, "loss": 3.8513, "step": 64750 }, { "epoch": 1.045144433154304, "grad_norm": 2.30800199508667, "learning_rate": 8.985384326186142e-05, "loss": 3.8612, "step": 64800 }, { "epoch": 1.045950871760133, "grad_norm": 1.3089618682861328, "learning_rate": 8.983848934659558e-05, "loss": 3.7833, "step": 64850 }, { "epoch": 1.0467573103659618, "grad_norm": 2.0379691123962402, "learning_rate": 8.98231251366796e-05, "loss": 3.9245, "step": 64900 }, { "epoch": 1.0475637489717908, "grad_norm": 1.6494274139404297, "learning_rate": 8.980775063608369e-05, "loss": 3.9081, "step": 64950 }, { "epoch": 1.0483701875776197, "grad_norm": 0.8300423622131348, "learning_rate": 8.97923658487808e-05, "loss": 3.6473, "step": 65000 }, { "epoch": 1.0491766261834488, "grad_norm": 2.0131113529205322, "learning_rate": 8.97769707787465e-05, "loss": 3.9677, "step": 65050 }, { "epoch": 1.0499830647892776, "grad_norm": 0.9256827235221863, "learning_rate": 8.976156542995902e-05, "loss": 3.9133, "step": 65100 }, { "epoch": 1.0507895033951065, "grad_norm": 1.5678056478500366, "learning_rate": 8.974614980639926e-05, "loss": 3.5199, "step": 65150 }, { "epoch": 1.0515959420009355, "grad_norm": 1.8995161056518555, "learning_rate": 8.973072391205077e-05, "loss": 4.0258, "step": 65200 }, { "epoch": 1.0524023806067644, "grad_norm": 2.5857930183410645, "learning_rate": 8.971528775089973e-05, "loss": 4.009, "step": 65250 }, { "epoch": 1.0532088192125932, "grad_norm": 3.9397096633911133, "learning_rate": 8.969984132693501e-05, "loss": 4.0625, "step": 65300 }, { "epoch": 1.0540152578184223, "grad_norm": 1.1694384813308716, "learning_rate": 8.96843846441481e-05, "loss": 3.7981, "step": 65350 }, { "epoch": 1.0548216964242512, "grad_norm": 1.9037532806396484, "learning_rate": 8.966891770653318e-05, "loss": 3.7698, "step": 65400 }, { "epoch": 1.0556281350300802, "grad_norm": 2.7292611598968506, "learning_rate": 8.965344051808705e-05, "loss": 3.8932, "step": 65450 }, { "epoch": 1.056434573635909, "grad_norm": 1.3260431289672852, "learning_rate": 8.963826293190776e-05, "loss": 3.6872, "step": 65500 }, { "epoch": 1.057241012241738, "grad_norm": 1.0064160823822021, "learning_rate": 8.962276545861756e-05, "loss": 3.426, "step": 65550 }, { "epoch": 1.058047450847567, "grad_norm": 1.6764061450958252, "learning_rate": 8.960725774642231e-05, "loss": 4.1136, "step": 65600 }, { "epoch": 1.0588538894533959, "grad_norm": 1.8129180669784546, "learning_rate": 8.959173979932937e-05, "loss": 3.6932, "step": 65650 }, { "epoch": 1.059660328059225, "grad_norm": 1.9630534648895264, "learning_rate": 8.957621162134872e-05, "loss": 3.7931, "step": 65700 }, { "epoch": 1.0604667666650538, "grad_norm": 1.960963487625122, "learning_rate": 8.956067321649303e-05, "loss": 3.5718, "step": 65750 }, { "epoch": 1.0612732052708826, "grad_norm": 1.8824496269226074, "learning_rate": 8.954512458877752e-05, "loss": 3.8215, "step": 65800 }, { "epoch": 1.0620796438767117, "grad_norm": 1.509952425956726, "learning_rate": 8.952956574222014e-05, "loss": 3.8786, "step": 65850 }, { "epoch": 1.0628860824825406, "grad_norm": 5.625761985778809, "learning_rate": 8.951399668084141e-05, "loss": 3.6898, "step": 65900 }, { "epoch": 1.0636925210883694, "grad_norm": 1.9310734272003174, "learning_rate": 8.949841740866459e-05, "loss": 3.667, "step": 65950 }, { "epoch": 1.0644989596941985, "grad_norm": 1.9376380443572998, "learning_rate": 8.948282792971545e-05, "loss": 3.7674, "step": 66000 }, { "epoch": 1.0644989596941985, "eval_asr_loss": 0.8276643907681687, "eval_loss": 3.973906993865967, "eval_runtime": 14.8499, "eval_samples_per_second": 34.478, "eval_steps_per_second": 11.515, "eval_tts_loss": 6.120585910734648, "step": 66000 }, { "epoch": 1.0653053983000274, "grad_norm": 1.96786367893219, "learning_rate": 8.946722824802247e-05, "loss": 3.9562, "step": 66050 }, { "epoch": 1.0661118369058564, "grad_norm": 0.9241717457771301, "learning_rate": 8.945161836761679e-05, "loss": 3.6945, "step": 66100 }, { "epoch": 1.0669182755116853, "grad_norm": 2.3022704124450684, "learning_rate": 8.943599829253215e-05, "loss": 3.7939, "step": 66150 }, { "epoch": 1.0677247141175141, "grad_norm": 2.221235513687134, "learning_rate": 8.942036802680491e-05, "loss": 4.2677, "step": 66200 }, { "epoch": 1.0685311527233432, "grad_norm": 1.7503604888916016, "learning_rate": 8.940472757447408e-05, "loss": 4.0832, "step": 66250 }, { "epoch": 1.069337591329172, "grad_norm": 1.5543756484985352, "learning_rate": 8.938907693958131e-05, "loss": 3.9174, "step": 66300 }, { "epoch": 1.0701440299350011, "grad_norm": 2.0880978107452393, "learning_rate": 8.937341612617089e-05, "loss": 3.8931, "step": 66350 }, { "epoch": 1.07095046854083, "grad_norm": 1.3646206855773926, "learning_rate": 8.935774513828971e-05, "loss": 3.6512, "step": 66400 }, { "epoch": 1.0717569071466588, "grad_norm": 1.874745488166809, "learning_rate": 8.93420639799873e-05, "loss": 3.9566, "step": 66450 }, { "epoch": 1.072563345752488, "grad_norm": 1.9509979486465454, "learning_rate": 8.932637265531583e-05, "loss": 3.7515, "step": 66500 }, { "epoch": 1.0733697843583168, "grad_norm": 1.1599488258361816, "learning_rate": 8.93106711683301e-05, "loss": 3.3567, "step": 66550 }, { "epoch": 1.0741762229641458, "grad_norm": 2.1112520694732666, "learning_rate": 8.92949595230875e-05, "loss": 3.6449, "step": 66600 }, { "epoch": 1.0749826615699747, "grad_norm": 2.211211919784546, "learning_rate": 8.92792377236481e-05, "loss": 3.8169, "step": 66650 }, { "epoch": 1.0757891001758035, "grad_norm": 2.0941267013549805, "learning_rate": 8.926350577407453e-05, "loss": 3.7632, "step": 66700 }, { "epoch": 1.0765955387816326, "grad_norm": 1.3243488073349, "learning_rate": 8.924776367843213e-05, "loss": 3.6304, "step": 66750 }, { "epoch": 1.0774019773874615, "grad_norm": 1.2164256572723389, "learning_rate": 8.923201144078875e-05, "loss": 3.8055, "step": 66800 }, { "epoch": 1.0782084159932905, "grad_norm": 2.1622555255889893, "learning_rate": 8.921624906521495e-05, "loss": 3.7527, "step": 66850 }, { "epoch": 1.0790148545991194, "grad_norm": 2.025451183319092, "learning_rate": 8.92004765557839e-05, "loss": 3.9559, "step": 66900 }, { "epoch": 1.0798212932049482, "grad_norm": 1.78819739818573, "learning_rate": 8.918469391657132e-05, "loss": 3.8267, "step": 66950 }, { "epoch": 1.0806277318107773, "grad_norm": 2.818911552429199, "learning_rate": 8.916890115165564e-05, "loss": 3.7239, "step": 67000 }, { "epoch": 1.0814341704166062, "grad_norm": 7.428587436676025, "learning_rate": 8.915309826511782e-05, "loss": 3.8719, "step": 67050 }, { "epoch": 1.082240609022435, "grad_norm": 2.182518720626831, "learning_rate": 8.913728526104154e-05, "loss": 3.866, "step": 67100 }, { "epoch": 1.083047047628264, "grad_norm": 2.536809206008911, "learning_rate": 8.912146214351298e-05, "loss": 3.8348, "step": 67150 }, { "epoch": 1.083853486234093, "grad_norm": 1.2114990949630737, "learning_rate": 8.910562891662099e-05, "loss": 3.4452, "step": 67200 }, { "epoch": 1.084659924839922, "grad_norm": 3.3965518474578857, "learning_rate": 8.908978558445706e-05, "loss": 3.9643, "step": 67250 }, { "epoch": 1.0854663634457509, "grad_norm": 2.0138041973114014, "learning_rate": 8.907393215111524e-05, "loss": 3.6371, "step": 67300 }, { "epoch": 1.0862728020515797, "grad_norm": 0.6829776763916016, "learning_rate": 8.90580686206922e-05, "loss": 3.4534, "step": 67350 }, { "epoch": 1.0870792406574088, "grad_norm": 2.04756236076355, "learning_rate": 8.904219499728726e-05, "loss": 3.7426, "step": 67400 }, { "epoch": 1.0878856792632376, "grad_norm": 2.9770071506500244, "learning_rate": 8.902631128500229e-05, "loss": 3.6768, "step": 67450 }, { "epoch": 1.0886921178690667, "grad_norm": 1.6580990552902222, "learning_rate": 8.901041748794179e-05, "loss": 3.9056, "step": 67500 }, { "epoch": 1.0894985564748956, "grad_norm": 2.255028247833252, "learning_rate": 8.89945136102129e-05, "loss": 3.1678, "step": 67550 }, { "epoch": 1.0903049950807244, "grad_norm": 4.1724042892456055, "learning_rate": 8.897859965592533e-05, "loss": 4.0075, "step": 67600 }, { "epoch": 1.0911114336865535, "grad_norm": 1.7806477546691895, "learning_rate": 8.896267562919137e-05, "loss": 3.9513, "step": 67650 }, { "epoch": 1.0919178722923824, "grad_norm": 2.317580223083496, "learning_rate": 8.894674153412596e-05, "loss": 3.6682, "step": 67700 }, { "epoch": 1.0927243108982112, "grad_norm": 1.8690036535263062, "learning_rate": 8.893079737484662e-05, "loss": 3.815, "step": 67750 }, { "epoch": 1.0935307495040403, "grad_norm": 1.0145264863967896, "learning_rate": 8.89148431554735e-05, "loss": 3.8267, "step": 67800 }, { "epoch": 1.0943371881098691, "grad_norm": 2.4025774002075195, "learning_rate": 8.889887888012929e-05, "loss": 3.708, "step": 67850 }, { "epoch": 1.0951436267156982, "grad_norm": 1.404681921005249, "learning_rate": 8.888290455293933e-05, "loss": 3.6218, "step": 67900 }, { "epoch": 1.095950065321527, "grad_norm": 2.4577507972717285, "learning_rate": 8.886692017803154e-05, "loss": 4.0323, "step": 67950 }, { "epoch": 1.096756503927356, "grad_norm": 2.255769968032837, "learning_rate": 8.885092575953644e-05, "loss": 4.0082, "step": 68000 }, { "epoch": 1.097562942533185, "grad_norm": 0.49128633737564087, "learning_rate": 8.883492130158713e-05, "loss": 3.8698, "step": 68050 }, { "epoch": 1.0983693811390138, "grad_norm": 1.2013691663742065, "learning_rate": 8.881922719650403e-05, "loss": 3.6453, "step": 68100 }, { "epoch": 1.099175819744843, "grad_norm": 1.4594979286193848, "learning_rate": 8.880320287263905e-05, "loss": 3.9393, "step": 68150 }, { "epoch": 1.0999822583506718, "grad_norm": 0.8525658845901489, "learning_rate": 8.878716852165192e-05, "loss": 3.8555, "step": 68200 }, { "epoch": 1.1007886969565006, "grad_norm": 2.9728105068206787, "learning_rate": 8.877112414768609e-05, "loss": 3.5215, "step": 68250 }, { "epoch": 1.1015951355623297, "grad_norm": 2.519521951675415, "learning_rate": 8.875506975488755e-05, "loss": 3.7864, "step": 68300 }, { "epoch": 1.1024015741681585, "grad_norm": 1.0732204914093018, "learning_rate": 8.873900534740491e-05, "loss": 3.7144, "step": 68350 }, { "epoch": 1.1032080127739876, "grad_norm": 2.832667350769043, "learning_rate": 8.87229309293894e-05, "loss": 3.4095, "step": 68400 }, { "epoch": 1.1040144513798165, "grad_norm": 1.479712724685669, "learning_rate": 8.870684650499478e-05, "loss": 3.5651, "step": 68450 }, { "epoch": 1.1048208899856453, "grad_norm": 1.1347146034240723, "learning_rate": 8.869075207837743e-05, "loss": 3.8152, "step": 68500 }, { "epoch": 1.1056273285914744, "grad_norm": 1.326029658317566, "learning_rate": 8.867464765369628e-05, "loss": 3.8229, "step": 68550 }, { "epoch": 1.1064337671973032, "grad_norm": 2.6609864234924316, "learning_rate": 8.865853323511292e-05, "loss": 3.8659, "step": 68600 }, { "epoch": 1.1072402058031323, "grad_norm": 2.1943726539611816, "learning_rate": 8.864240882679143e-05, "loss": 3.9707, "step": 68650 }, { "epoch": 1.1080466444089612, "grad_norm": 0.7717074155807495, "learning_rate": 8.86262744328985e-05, "loss": 3.4627, "step": 68700 }, { "epoch": 1.10885308301479, "grad_norm": 2.110758066177368, "learning_rate": 8.861013005760347e-05, "loss": 3.8781, "step": 68750 }, { "epoch": 1.109659521620619, "grad_norm": 1.7255288362503052, "learning_rate": 8.859397570507816e-05, "loss": 3.8008, "step": 68800 }, { "epoch": 1.110465960226448, "grad_norm": 1.4433388710021973, "learning_rate": 8.857781137949701e-05, "loss": 3.7542, "step": 68850 }, { "epoch": 1.1112723988322768, "grad_norm": 0.9421756267547607, "learning_rate": 8.856163708503706e-05, "loss": 4.0513, "step": 68900 }, { "epoch": 1.1120788374381059, "grad_norm": 1.589231252670288, "learning_rate": 8.85454528258779e-05, "loss": 3.4589, "step": 68950 }, { "epoch": 1.1128852760439347, "grad_norm": 2.0848567485809326, "learning_rate": 8.852925860620167e-05, "loss": 3.9888, "step": 69000 }, { "epoch": 1.1128852760439347, "eval_asr_loss": 0.818536922284418, "eval_loss": 3.948729991912842, "eval_runtime": 15.3134, "eval_samples_per_second": 33.435, "eval_steps_per_second": 11.167, "eval_tts_loss": 6.118797951925912, "step": 69000 }, { "epoch": 1.1136917146497638, "grad_norm": 2.094059467315674, "learning_rate": 8.851305443019314e-05, "loss": 3.648, "step": 69050 }, { "epoch": 1.1144981532555926, "grad_norm": 1.4876412153244019, "learning_rate": 8.849684030203962e-05, "loss": 3.7317, "step": 69100 }, { "epoch": 1.1153045918614215, "grad_norm": 1.241718053817749, "learning_rate": 8.848061622593101e-05, "loss": 3.6353, "step": 69150 }, { "epoch": 1.1161110304672506, "grad_norm": 1.6719624996185303, "learning_rate": 8.846438220605973e-05, "loss": 3.6973, "step": 69200 }, { "epoch": 1.1169174690730794, "grad_norm": 3.324497699737549, "learning_rate": 8.844813824662083e-05, "loss": 3.6342, "step": 69250 }, { "epoch": 1.1177239076789085, "grad_norm": 1.8133293390274048, "learning_rate": 8.843188435181194e-05, "loss": 3.4442, "step": 69300 }, { "epoch": 1.1185303462847374, "grad_norm": 2.034195899963379, "learning_rate": 8.841562052583315e-05, "loss": 3.8356, "step": 69350 }, { "epoch": 1.1193367848905662, "grad_norm": 1.6730608940124512, "learning_rate": 8.839934677288725e-05, "loss": 3.7574, "step": 69400 }, { "epoch": 1.1201432234963953, "grad_norm": 2.4863834381103516, "learning_rate": 8.838306309717952e-05, "loss": 3.493, "step": 69450 }, { "epoch": 1.1209496621022241, "grad_norm": 1.3953959941864014, "learning_rate": 8.836676950291781e-05, "loss": 3.765, "step": 69500 }, { "epoch": 1.121756100708053, "grad_norm": 2.589898109436035, "learning_rate": 8.835046599431253e-05, "loss": 3.4203, "step": 69550 }, { "epoch": 1.122562539313882, "grad_norm": 1.2807525396347046, "learning_rate": 8.83341525755767e-05, "loss": 3.9296, "step": 69600 }, { "epoch": 1.123368977919711, "grad_norm": 1.8745602369308472, "learning_rate": 8.831782925092585e-05, "loss": 3.8295, "step": 69650 }, { "epoch": 1.12417541652554, "grad_norm": 1.3585973978042603, "learning_rate": 8.830149602457804e-05, "loss": 3.8766, "step": 69700 }, { "epoch": 1.1249818551313688, "grad_norm": 1.7997664213180542, "learning_rate": 8.828515290075401e-05, "loss": 3.6271, "step": 69750 }, { "epoch": 1.1257882937371977, "grad_norm": 1.4997978210449219, "learning_rate": 8.826879988367693e-05, "loss": 4.0125, "step": 69800 }, { "epoch": 1.1265947323430268, "grad_norm": 1.3555970191955566, "learning_rate": 8.82524369775726e-05, "loss": 3.8325, "step": 69850 }, { "epoch": 1.1274011709488556, "grad_norm": 1.2100369930267334, "learning_rate": 8.823606418666933e-05, "loss": 3.6998, "step": 69900 }, { "epoch": 1.1282076095546847, "grad_norm": 3.1101126670837402, "learning_rate": 8.821968151519804e-05, "loss": 3.9598, "step": 69950 }, { "epoch": 1.1290140481605135, "grad_norm": 1.5878289937973022, "learning_rate": 8.820328896739214e-05, "loss": 4.0986, "step": 70000 }, { "epoch": 1.1298204867663424, "grad_norm": 2.226511001586914, "learning_rate": 8.818688654748763e-05, "loss": 3.9124, "step": 70050 }, { "epoch": 1.1306269253721715, "grad_norm": 1.7032560110092163, "learning_rate": 8.817047425972305e-05, "loss": 3.5396, "step": 70100 }, { "epoch": 1.1314333639780003, "grad_norm": 1.3012017011642456, "learning_rate": 8.815405210833952e-05, "loss": 3.6626, "step": 70150 }, { "epoch": 1.1322398025838294, "grad_norm": 1.619145154953003, "learning_rate": 8.813762009758066e-05, "loss": 3.7756, "step": 70200 }, { "epoch": 1.1330462411896582, "grad_norm": 2.1445043087005615, "learning_rate": 8.812117823169266e-05, "loss": 3.6025, "step": 70250 }, { "epoch": 1.133852679795487, "grad_norm": 1.2163410186767578, "learning_rate": 8.810472651492425e-05, "loss": 3.8079, "step": 70300 }, { "epoch": 1.1346591184013162, "grad_norm": 1.8173290491104126, "learning_rate": 8.808826495152674e-05, "loss": 3.7411, "step": 70350 }, { "epoch": 1.135465557007145, "grad_norm": 1.2783173322677612, "learning_rate": 8.807179354575395e-05, "loss": 3.7384, "step": 70400 }, { "epoch": 1.136271995612974, "grad_norm": 0.9880287647247314, "learning_rate": 8.805531230186221e-05, "loss": 3.7808, "step": 70450 }, { "epoch": 1.137078434218803, "grad_norm": 1.5661088228225708, "learning_rate": 8.80388212241105e-05, "loss": 4.0546, "step": 70500 }, { "epoch": 1.1378848728246318, "grad_norm": 1.3866236209869385, "learning_rate": 8.802232031676018e-05, "loss": 3.8734, "step": 70550 }, { "epoch": 1.1386913114304609, "grad_norm": 1.7212815284729004, "learning_rate": 8.800580958407535e-05, "loss": 3.8524, "step": 70600 }, { "epoch": 1.1394977500362897, "grad_norm": 1.6367288827896118, "learning_rate": 8.798928903032248e-05, "loss": 4.2428, "step": 70650 }, { "epoch": 1.1403041886421188, "grad_norm": 0.9309474229812622, "learning_rate": 8.797275865977064e-05, "loss": 3.3679, "step": 70700 }, { "epoch": 1.1411106272479477, "grad_norm": 1.9396387338638306, "learning_rate": 8.795621847669146e-05, "loss": 3.8537, "step": 70750 }, { "epoch": 1.1419170658537765, "grad_norm": 1.4001718759536743, "learning_rate": 8.793966848535908e-05, "loss": 3.6426, "step": 70800 }, { "epoch": 1.1427235044596056, "grad_norm": 1.3842568397521973, "learning_rate": 8.792310869005016e-05, "loss": 3.7029, "step": 70850 }, { "epoch": 1.1435299430654344, "grad_norm": 2.0930709838867188, "learning_rate": 8.790653909504394e-05, "loss": 3.694, "step": 70900 }, { "epoch": 1.1443363816712633, "grad_norm": 1.6417869329452515, "learning_rate": 8.788995970462213e-05, "loss": 3.8145, "step": 70950 }, { "epoch": 1.1451428202770924, "grad_norm": 2.176569938659668, "learning_rate": 8.787337052306902e-05, "loss": 3.7135, "step": 71000 }, { "epoch": 1.1459492588829212, "grad_norm": 1.1373050212860107, "learning_rate": 8.785677155467142e-05, "loss": 3.5726, "step": 71050 }, { "epoch": 1.14675569748875, "grad_norm": 2.3727970123291016, "learning_rate": 8.784016280371865e-05, "loss": 3.846, "step": 71100 }, { "epoch": 1.1475621360945791, "grad_norm": 0.7403922080993652, "learning_rate": 8.782354427450259e-05, "loss": 3.5053, "step": 71150 }, { "epoch": 1.148368574700408, "grad_norm": 2.3267409801483154, "learning_rate": 8.78069159713176e-05, "loss": 3.6501, "step": 71200 }, { "epoch": 1.149175013306237, "grad_norm": 2.1716699600219727, "learning_rate": 8.779027789846064e-05, "loss": 3.7334, "step": 71250 }, { "epoch": 1.149981451912066, "grad_norm": 2.8139374256134033, "learning_rate": 8.77736300602311e-05, "loss": 3.9918, "step": 71300 }, { "epoch": 1.1507878905178948, "grad_norm": 2.0450313091278076, "learning_rate": 8.775697246093097e-05, "loss": 3.7199, "step": 71350 }, { "epoch": 1.1515943291237238, "grad_norm": 1.7587394714355469, "learning_rate": 8.774030510486472e-05, "loss": 3.7552, "step": 71400 }, { "epoch": 1.1524007677295527, "grad_norm": 1.9739865064620972, "learning_rate": 8.772362799633938e-05, "loss": 3.8243, "step": 71450 }, { "epoch": 1.1532072063353818, "grad_norm": 3.028428316116333, "learning_rate": 8.770694113966443e-05, "loss": 3.6705, "step": 71500 }, { "epoch": 1.1540136449412106, "grad_norm": 1.6889009475708008, "learning_rate": 8.769024453915197e-05, "loss": 3.8804, "step": 71550 }, { "epoch": 1.1548200835470395, "grad_norm": 0.5943315029144287, "learning_rate": 8.767353819911652e-05, "loss": 4.0624, "step": 71600 }, { "epoch": 1.1556265221528685, "grad_norm": 2.424131155014038, "learning_rate": 8.765682212387518e-05, "loss": 3.8468, "step": 71650 }, { "epoch": 1.1564329607586974, "grad_norm": 1.460707426071167, "learning_rate": 8.764009631774754e-05, "loss": 3.617, "step": 71700 }, { "epoch": 1.1572393993645265, "grad_norm": 0.8052225708961487, "learning_rate": 8.762336078505573e-05, "loss": 3.9311, "step": 71750 }, { "epoch": 1.1580458379703553, "grad_norm": 2.2520415782928467, "learning_rate": 8.760661553012432e-05, "loss": 3.6906, "step": 71800 }, { "epoch": 1.1588522765761842, "grad_norm": 2.1767735481262207, "learning_rate": 8.758986055728051e-05, "loss": 3.677, "step": 71850 }, { "epoch": 1.1596587151820132, "grad_norm": 3.631753444671631, "learning_rate": 8.757309587085391e-05, "loss": 3.9432, "step": 71900 }, { "epoch": 1.160465153787842, "grad_norm": 2.915670156478882, "learning_rate": 8.755632147517668e-05, "loss": 3.908, "step": 71950 }, { "epoch": 1.1612715923936712, "grad_norm": 2.365554094314575, "learning_rate": 8.753953737458347e-05, "loss": 4.0629, "step": 72000 }, { "epoch": 1.1612715923936712, "eval_asr_loss": 0.8260773614948321, "eval_loss": 3.9508860111236572, "eval_runtime": 15.0812, "eval_samples_per_second": 33.95, "eval_steps_per_second": 11.339, "eval_tts_loss": 6.106224110514112, "step": 72000 }, { "epoch": 1.1620780309995, "grad_norm": 1.7346563339233398, "learning_rate": 8.75227435734115e-05, "loss": 3.6228, "step": 72050 }, { "epoch": 1.1628844696053289, "grad_norm": 1.97269868850708, "learning_rate": 8.75059400760004e-05, "loss": 3.9359, "step": 72100 }, { "epoch": 1.163690908211158, "grad_norm": 1.9431113004684448, "learning_rate": 8.74891268866924e-05, "loss": 3.7229, "step": 72150 }, { "epoch": 1.1644973468169868, "grad_norm": 1.8456497192382812, "learning_rate": 8.747230400983214e-05, "loss": 3.6797, "step": 72200 }, { "epoch": 1.1653037854228159, "grad_norm": 1.7494312524795532, "learning_rate": 8.745547144976686e-05, "loss": 3.6052, "step": 72250 }, { "epoch": 1.1661102240286447, "grad_norm": 2.0142862796783447, "learning_rate": 8.743862921084624e-05, "loss": 3.7469, "step": 72300 }, { "epoch": 1.1669166626344736, "grad_norm": 1.2905949354171753, "learning_rate": 8.742177729742247e-05, "loss": 3.7056, "step": 72350 }, { "epoch": 1.1677231012403027, "grad_norm": 0.7779897451400757, "learning_rate": 8.740491571385025e-05, "loss": 3.8985, "step": 72400 }, { "epoch": 1.1685295398461315, "grad_norm": 1.1052699089050293, "learning_rate": 8.738804446448679e-05, "loss": 3.5891, "step": 72450 }, { "epoch": 1.1693359784519604, "grad_norm": 0.5351213812828064, "learning_rate": 8.737116355369178e-05, "loss": 3.4503, "step": 72500 }, { "epoch": 1.1701424170577894, "grad_norm": 1.598081350326538, "learning_rate": 8.735427298582737e-05, "loss": 3.9359, "step": 72550 }, { "epoch": 1.1709488556636183, "grad_norm": 0.9709692001342773, "learning_rate": 8.733737276525831e-05, "loss": 3.7405, "step": 72600 }, { "epoch": 1.1717552942694474, "grad_norm": 1.6495882272720337, "learning_rate": 8.732046289635174e-05, "loss": 3.6956, "step": 72650 }, { "epoch": 1.1725617328752762, "grad_norm": 0.8482274413108826, "learning_rate": 8.730354338347732e-05, "loss": 3.7737, "step": 72700 }, { "epoch": 1.173368171481105, "grad_norm": 2.0187036991119385, "learning_rate": 8.728661423100727e-05, "loss": 3.9532, "step": 72750 }, { "epoch": 1.1741746100869341, "grad_norm": 2.26690673828125, "learning_rate": 8.72696754433162e-05, "loss": 4.0482, "step": 72800 }, { "epoch": 1.174981048692763, "grad_norm": 1.5149962902069092, "learning_rate": 8.725272702478126e-05, "loss": 3.914, "step": 72850 }, { "epoch": 1.1757874872985918, "grad_norm": 1.7810536623001099, "learning_rate": 8.723576897978211e-05, "loss": 3.6576, "step": 72900 }, { "epoch": 1.176593925904421, "grad_norm": 0.5974041223526001, "learning_rate": 8.721880131270086e-05, "loss": 3.8344, "step": 72950 }, { "epoch": 1.1774003645102498, "grad_norm": 1.253562569618225, "learning_rate": 8.720216366784274e-05, "loss": 3.8555, "step": 73000 }, { "epoch": 1.1782068031160788, "grad_norm": 1.4296804666519165, "learning_rate": 8.718517696197678e-05, "loss": 3.6059, "step": 73050 }, { "epoch": 1.1790132417219077, "grad_norm": 2.3996424674987793, "learning_rate": 8.716818064710219e-05, "loss": 3.8622, "step": 73100 }, { "epoch": 1.1798196803277365, "grad_norm": 2.273122549057007, "learning_rate": 8.715117472761098e-05, "loss": 3.487, "step": 73150 }, { "epoch": 1.1806261189335656, "grad_norm": 2.0540523529052734, "learning_rate": 8.713415920789763e-05, "loss": 3.4876, "step": 73200 }, { "epoch": 1.1814325575393945, "grad_norm": 1.3778162002563477, "learning_rate": 8.711713409235914e-05, "loss": 3.8378, "step": 73250 }, { "epoch": 1.1822389961452235, "grad_norm": 1.5858887434005737, "learning_rate": 8.710009938539492e-05, "loss": 3.9566, "step": 73300 }, { "epoch": 1.1830454347510524, "grad_norm": 1.2605814933776855, "learning_rate": 8.708305509140695e-05, "loss": 3.7069, "step": 73350 }, { "epoch": 1.1838518733568812, "grad_norm": 2.9215505123138428, "learning_rate": 8.706634238621292e-05, "loss": 3.9086, "step": 73400 }, { "epoch": 1.1846583119627103, "grad_norm": 2.0706405639648438, "learning_rate": 8.704927912291417e-05, "loss": 3.5687, "step": 73450 }, { "epoch": 1.1854647505685392, "grad_norm": 2.0366086959838867, "learning_rate": 8.703220628572409e-05, "loss": 3.5826, "step": 73500 }, { "epoch": 1.1862711891743682, "grad_norm": 1.3899857997894287, "learning_rate": 8.701512387905445e-05, "loss": 3.8591, "step": 73550 }, { "epoch": 1.187077627780197, "grad_norm": 1.8716031312942505, "learning_rate": 8.699803190731954e-05, "loss": 3.6318, "step": 73600 }, { "epoch": 1.187884066386026, "grad_norm": 1.5658704042434692, "learning_rate": 8.698093037493609e-05, "loss": 3.9993, "step": 73650 }, { "epoch": 1.188690504991855, "grad_norm": 2.4947080612182617, "learning_rate": 8.696381928632329e-05, "loss": 3.7529, "step": 73700 }, { "epoch": 1.1894969435976839, "grad_norm": 1.3231422901153564, "learning_rate": 8.694669864590281e-05, "loss": 3.843, "step": 73750 }, { "epoch": 1.190303382203513, "grad_norm": 1.5092302560806274, "learning_rate": 8.692956845809881e-05, "loss": 3.891, "step": 73800 }, { "epoch": 1.1911098208093418, "grad_norm": 1.6695327758789062, "learning_rate": 8.691242872733786e-05, "loss": 3.7838, "step": 73850 }, { "epoch": 1.1919162594151707, "grad_norm": 2.2980494499206543, "learning_rate": 8.689527945804906e-05, "loss": 3.8116, "step": 73900 }, { "epoch": 1.1927226980209997, "grad_norm": 1.8897286653518677, "learning_rate": 8.687812065466392e-05, "loss": 3.7376, "step": 73950 }, { "epoch": 1.1935291366268286, "grad_norm": 1.7522248029708862, "learning_rate": 8.686095232161647e-05, "loss": 3.8732, "step": 74000 }, { "epoch": 1.1943355752326577, "grad_norm": 2.001786708831787, "learning_rate": 8.684377446334314e-05, "loss": 3.9358, "step": 74050 }, { "epoch": 1.1951420138384865, "grad_norm": 1.8185148239135742, "learning_rate": 8.682658708428288e-05, "loss": 3.8181, "step": 74100 }, { "epoch": 1.1959484524443154, "grad_norm": 2.0440456867218018, "learning_rate": 8.680939018887709e-05, "loss": 3.9695, "step": 74150 }, { "epoch": 1.1967548910501444, "grad_norm": 1.1397546529769897, "learning_rate": 8.679218378156957e-05, "loss": 3.7973, "step": 74200 }, { "epoch": 1.1975613296559733, "grad_norm": 1.6209256649017334, "learning_rate": 8.677496786680666e-05, "loss": 3.8723, "step": 74250 }, { "epoch": 1.1983677682618021, "grad_norm": 1.3796250820159912, "learning_rate": 8.675774244903709e-05, "loss": 3.9217, "step": 74300 }, { "epoch": 1.1991742068676312, "grad_norm": 1.4182711839675903, "learning_rate": 8.67405075327121e-05, "loss": 4.028, "step": 74350 }, { "epoch": 1.19998064547346, "grad_norm": 1.1692181825637817, "learning_rate": 8.672326312228534e-05, "loss": 4.0558, "step": 74400 }, { "epoch": 1.2007870840792891, "grad_norm": 1.912279725074768, "learning_rate": 8.670600922221292e-05, "loss": 3.816, "step": 74450 }, { "epoch": 1.201593522685118, "grad_norm": 3.0808444023132324, "learning_rate": 8.668874583695346e-05, "loss": 3.608, "step": 74500 }, { "epoch": 1.2023999612909468, "grad_norm": 1.1143704652786255, "learning_rate": 8.667147297096794e-05, "loss": 3.3062, "step": 74550 }, { "epoch": 1.203206399896776, "grad_norm": 1.283655047416687, "learning_rate": 8.665419062871987e-05, "loss": 3.9392, "step": 74600 }, { "epoch": 1.2040128385026048, "grad_norm": 1.9609984159469604, "learning_rate": 8.663689881467514e-05, "loss": 3.507, "step": 74650 }, { "epoch": 1.2048192771084336, "grad_norm": 1.7727683782577515, "learning_rate": 8.661959753330213e-05, "loss": 3.7945, "step": 74700 }, { "epoch": 1.2056257157142627, "grad_norm": 2.392547845840454, "learning_rate": 8.660228678907168e-05, "loss": 3.536, "step": 74750 }, { "epoch": 1.2064321543200915, "grad_norm": 1.852827787399292, "learning_rate": 8.658496658645704e-05, "loss": 3.645, "step": 74800 }, { "epoch": 1.2072385929259206, "grad_norm": 1.9689486026763916, "learning_rate": 8.656763692993391e-05, "loss": 3.9369, "step": 74850 }, { "epoch": 1.2080450315317495, "grad_norm": 2.0755228996276855, "learning_rate": 8.655029782398045e-05, "loss": 3.7035, "step": 74900 }, { "epoch": 1.2088514701375783, "grad_norm": 1.4257313013076782, "learning_rate": 8.653294927307726e-05, "loss": 4.0371, "step": 74950 }, { "epoch": 1.2096579087434074, "grad_norm": 2.4752979278564453, "learning_rate": 8.651559128170736e-05, "loss": 3.9648, "step": 75000 }, { "epoch": 1.2096579087434074, "eval_asr_loss": 0.8115376394511965, "eval_loss": 3.94327712059021, "eval_runtime": 14.7315, "eval_samples_per_second": 34.755, "eval_steps_per_second": 11.608, "eval_tts_loss": 6.090846049183692, "step": 75000 }, { "epoch": 1.2104643473492362, "grad_norm": 1.59127676486969, "learning_rate": 8.649822385435622e-05, "loss": 3.5147, "step": 75050 }, { "epoch": 1.2112707859550653, "grad_norm": 1.3992995023727417, "learning_rate": 8.648084699551176e-05, "loss": 3.3567, "step": 75100 }, { "epoch": 1.2120772245608942, "grad_norm": 3.147475242614746, "learning_rate": 8.646346070966432e-05, "loss": 3.7278, "step": 75150 }, { "epoch": 1.212883663166723, "grad_norm": 2.1718759536743164, "learning_rate": 8.64460650013067e-05, "loss": 3.6541, "step": 75200 }, { "epoch": 1.213690101772552, "grad_norm": 1.3402507305145264, "learning_rate": 8.642865987493409e-05, "loss": 3.6504, "step": 75250 }, { "epoch": 1.214496540378381, "grad_norm": 1.9188058376312256, "learning_rate": 8.641124533504417e-05, "loss": 3.7686, "step": 75300 }, { "epoch": 1.21530297898421, "grad_norm": 2.177851676940918, "learning_rate": 8.639382138613701e-05, "loss": 3.8654, "step": 75350 }, { "epoch": 1.2161094175900389, "grad_norm": 2.473162889480591, "learning_rate": 8.637638803271515e-05, "loss": 3.8304, "step": 75400 }, { "epoch": 1.2169158561958677, "grad_norm": 2.4639079570770264, "learning_rate": 8.63589452792835e-05, "loss": 3.9537, "step": 75450 }, { "epoch": 1.2177222948016968, "grad_norm": 1.428101897239685, "learning_rate": 8.634149313034947e-05, "loss": 3.7357, "step": 75500 }, { "epoch": 1.2185287334075257, "grad_norm": 2.3032217025756836, "learning_rate": 8.632403159042283e-05, "loss": 3.5821, "step": 75550 }, { "epoch": 1.2193351720133547, "grad_norm": 0.6937980055809021, "learning_rate": 8.630656066401583e-05, "loss": 3.9531, "step": 75600 }, { "epoch": 1.2201416106191836, "grad_norm": 1.7546741962432861, "learning_rate": 8.628908035564313e-05, "loss": 3.6574, "step": 75650 }, { "epoch": 1.2209480492250124, "grad_norm": 1.682079792022705, "learning_rate": 8.627159066982183e-05, "loss": 3.6548, "step": 75700 }, { "epoch": 1.2217544878308415, "grad_norm": 2.0722949504852295, "learning_rate": 8.625409161107138e-05, "loss": 3.6317, "step": 75750 }, { "epoch": 1.2225609264366704, "grad_norm": 1.778039813041687, "learning_rate": 8.623658318391375e-05, "loss": 3.8805, "step": 75800 }, { "epoch": 1.2233673650424994, "grad_norm": 1.7971779108047485, "learning_rate": 8.621906539287325e-05, "loss": 3.7566, "step": 75850 }, { "epoch": 1.2241738036483283, "grad_norm": 2.297848701477051, "learning_rate": 8.620153824247668e-05, "loss": 3.7402, "step": 75900 }, { "epoch": 1.2249802422541571, "grad_norm": 0.8815058469772339, "learning_rate": 8.61840017372532e-05, "loss": 3.3031, "step": 75950 }, { "epoch": 1.2257866808599862, "grad_norm": 2.303922653198242, "learning_rate": 8.616645588173445e-05, "loss": 3.8486, "step": 76000 }, { "epoch": 1.226593119465815, "grad_norm": 1.4316277503967285, "learning_rate": 8.61489006804544e-05, "loss": 3.6216, "step": 76050 }, { "epoch": 1.227399558071644, "grad_norm": 1.3122050762176514, "learning_rate": 8.613133613794953e-05, "loss": 3.8334, "step": 76100 }, { "epoch": 1.228205996677473, "grad_norm": 1.706869125366211, "learning_rate": 8.611376225875864e-05, "loss": 3.939, "step": 76150 }, { "epoch": 1.2290124352833018, "grad_norm": 1.0819898843765259, "learning_rate": 8.609617904742302e-05, "loss": 3.7572, "step": 76200 }, { "epoch": 1.229818873889131, "grad_norm": 2.8179965019226074, "learning_rate": 8.607858650848635e-05, "loss": 3.7671, "step": 76250 }, { "epoch": 1.2306253124949598, "grad_norm": 1.9507651329040527, "learning_rate": 8.606098464649468e-05, "loss": 3.7334, "step": 76300 }, { "epoch": 1.2314317511007886, "grad_norm": 1.242860198020935, "learning_rate": 8.604337346599652e-05, "loss": 3.8932, "step": 76350 }, { "epoch": 1.2322381897066177, "grad_norm": 1.4878056049346924, "learning_rate": 8.602575297154277e-05, "loss": 4.0502, "step": 76400 }, { "epoch": 1.2330446283124465, "grad_norm": 1.5658447742462158, "learning_rate": 8.600812316768674e-05, "loss": 3.9599, "step": 76450 }, { "epoch": 1.2338510669182754, "grad_norm": 0.7045395970344543, "learning_rate": 8.599048405898412e-05, "loss": 3.672, "step": 76500 }, { "epoch": 1.2346575055241045, "grad_norm": 1.2160003185272217, "learning_rate": 8.597283564999306e-05, "loss": 3.4146, "step": 76550 }, { "epoch": 1.2354639441299333, "grad_norm": 1.8320386409759521, "learning_rate": 8.595517794527403e-05, "loss": 3.8617, "step": 76600 }, { "epoch": 1.2362703827357624, "grad_norm": 1.5984004735946655, "learning_rate": 8.593751094939e-05, "loss": 3.862, "step": 76650 }, { "epoch": 1.2370768213415912, "grad_norm": 1.8850581645965576, "learning_rate": 8.591983466690627e-05, "loss": 3.9977, "step": 76700 }, { "epoch": 1.23788325994742, "grad_norm": 1.8823696374893188, "learning_rate": 8.590214910239057e-05, "loss": 3.8056, "step": 76750 }, { "epoch": 1.2386896985532492, "grad_norm": 1.4765985012054443, "learning_rate": 8.588445426041301e-05, "loss": 4.0095, "step": 76800 }, { "epoch": 1.239496137159078, "grad_norm": 1.60728120803833, "learning_rate": 8.586675014554612e-05, "loss": 3.6366, "step": 76850 }, { "epoch": 1.240302575764907, "grad_norm": 1.9440919160842896, "learning_rate": 8.58490367623648e-05, "loss": 3.6952, "step": 76900 }, { "epoch": 1.241109014370736, "grad_norm": 1.6637271642684937, "learning_rate": 8.583131411544638e-05, "loss": 4.0625, "step": 76950 }, { "epoch": 1.2419154529765648, "grad_norm": 1.292341947555542, "learning_rate": 8.581358220937053e-05, "loss": 3.7801, "step": 77000 }, { "epoch": 1.2427218915823939, "grad_norm": 2.5661232471466064, "learning_rate": 8.57961959625976e-05, "loss": 3.7455, "step": 77050 }, { "epoch": 1.2435283301882227, "grad_norm": 3.1576755046844482, "learning_rate": 8.57784457369105e-05, "loss": 3.9374, "step": 77100 }, { "epoch": 1.2443347687940518, "grad_norm": 1.5049893856048584, "learning_rate": 8.576068626572767e-05, "loss": 3.6296, "step": 77150 }, { "epoch": 1.2451412073998807, "grad_norm": 1.7539376020431519, "learning_rate": 8.574291755363835e-05, "loss": 3.7688, "step": 77200 }, { "epoch": 1.2459476460057095, "grad_norm": 1.7176966667175293, "learning_rate": 8.572513960523416e-05, "loss": 4.1359, "step": 77250 }, { "epoch": 1.2467540846115386, "grad_norm": 1.3343137502670288, "learning_rate": 8.570735242510907e-05, "loss": 3.5497, "step": 77300 }, { "epoch": 1.2475605232173674, "grad_norm": 3.1631829738616943, "learning_rate": 8.568955601785948e-05, "loss": 3.9663, "step": 77350 }, { "epoch": 1.2483669618231965, "grad_norm": 1.260836124420166, "learning_rate": 8.567175038808412e-05, "loss": 3.4861, "step": 77400 }, { "epoch": 1.2491734004290254, "grad_norm": 1.9480077028274536, "learning_rate": 8.565393554038418e-05, "loss": 4.1111, "step": 77450 }, { "epoch": 1.2499798390348542, "grad_norm": 1.741125464439392, "learning_rate": 8.563611147936314e-05, "loss": 3.7041, "step": 77500 }, { "epoch": 1.2507862776406833, "grad_norm": 2.265465021133423, "learning_rate": 8.561827820962696e-05, "loss": 4.0834, "step": 77550 }, { "epoch": 1.2515927162465121, "grad_norm": 2.384108304977417, "learning_rate": 8.560043573578389e-05, "loss": 3.6216, "step": 77600 }, { "epoch": 1.2523991548523412, "grad_norm": 1.4291831254959106, "learning_rate": 8.558258406244462e-05, "loss": 3.935, "step": 77650 }, { "epoch": 1.25320559345817, "grad_norm": 3.713202476501465, "learning_rate": 8.556472319422217e-05, "loss": 3.8317, "step": 77700 }, { "epoch": 1.254012032063999, "grad_norm": 1.9122109413146973, "learning_rate": 8.5546853135732e-05, "loss": 3.6591, "step": 77750 }, { "epoch": 1.2548184706698278, "grad_norm": 2.7846715450286865, "learning_rate": 8.552897389159187e-05, "loss": 3.8179, "step": 77800 }, { "epoch": 1.2556249092756568, "grad_norm": 2.3511781692504883, "learning_rate": 8.551108546642197e-05, "loss": 3.6225, "step": 77850 }, { "epoch": 1.256431347881486, "grad_norm": 2.073401689529419, "learning_rate": 8.549318786484482e-05, "loss": 3.8885, "step": 77900 }, { "epoch": 1.2572377864873148, "grad_norm": 1.286442518234253, "learning_rate": 8.547528109148536e-05, "loss": 3.7674, "step": 77950 }, { "epoch": 1.2580442250931436, "grad_norm": 1.6265051364898682, "learning_rate": 8.545736515097085e-05, "loss": 3.9492, "step": 78000 }, { "epoch": 1.2580442250931436, "eval_asr_loss": 0.8204757965901752, "eval_loss": 3.9444222450256348, "eval_runtime": 15.1709, "eval_samples_per_second": 33.749, "eval_steps_per_second": 11.272, "eval_tts_loss": 6.092585181199037, "step": 78000 } ], "logging_steps": 50, "max_steps": 310005, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 3000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 6.415216825818153e+17, "train_batch_size": 3, "trial_name": null, "trial_params": null }