salt-asr-tts-48k / trainer_state.json
ksych's picture
Upload folder using huggingface_hub
567b09d verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.7741810615957807,
"eval_steps": 3000,
"global_step": 48000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0008064386058289383,
"grad_norm": 6.343989849090576,
"learning_rate": 4.800000000000001e-06,
"loss": 3.2595,
"step": 50
},
{
"epoch": 0.0016128772116578765,
"grad_norm": 3.7637691497802734,
"learning_rate": 9.800000000000001e-06,
"loss": 3.2397,
"step": 100
},
{
"epoch": 0.002419315817486815,
"grad_norm": 4.01455020904541,
"learning_rate": 1.48e-05,
"loss": 3.1471,
"step": 150
},
{
"epoch": 0.003225754423315753,
"grad_norm": 5.399896621704102,
"learning_rate": 1.9800000000000004e-05,
"loss": 3.3162,
"step": 200
},
{
"epoch": 0.004032193029144691,
"grad_norm": 5.194526195526123,
"learning_rate": 2.48e-05,
"loss": 3.4495,
"step": 250
},
{
"epoch": 0.00483863163497363,
"grad_norm": 4.043808460235596,
"learning_rate": 2.98e-05,
"loss": 3.4038,
"step": 300
},
{
"epoch": 0.005645070240802567,
"grad_norm": 1.6968616247177124,
"learning_rate": 3.46e-05,
"loss": 3.5791,
"step": 350
},
{
"epoch": 0.006451508846631506,
"grad_norm": 3.515857219696045,
"learning_rate": 3.960000000000001e-05,
"loss": 3.4015,
"step": 400
},
{
"epoch": 0.007257947452460444,
"grad_norm": 3.327608108520508,
"learning_rate": 4.46e-05,
"loss": 3.2794,
"step": 450
},
{
"epoch": 0.008064386058289382,
"grad_norm": 3.6122305393218994,
"learning_rate": 4.96e-05,
"loss": 3.3541,
"step": 500
},
{
"epoch": 0.00887082466411832,
"grad_norm": 4.593649864196777,
"learning_rate": 5.4600000000000006e-05,
"loss": 3.6442,
"step": 550
},
{
"epoch": 0.00967726326994726,
"grad_norm": 2.58380126953125,
"learning_rate": 5.96e-05,
"loss": 3.1673,
"step": 600
},
{
"epoch": 0.010483701875776198,
"grad_norm": 5.36879825592041,
"learning_rate": 6.460000000000001e-05,
"loss": 3.5224,
"step": 650
},
{
"epoch": 0.011290140481605135,
"grad_norm": 21.212533950805664,
"learning_rate": 6.96e-05,
"loss": 3.5144,
"step": 700
},
{
"epoch": 0.012096579087434074,
"grad_norm": 1.9257746934890747,
"learning_rate": 7.46e-05,
"loss": 3.2471,
"step": 750
},
{
"epoch": 0.012903017693263012,
"grad_norm": 6.157283306121826,
"learning_rate": 7.960000000000001e-05,
"loss": 3.551,
"step": 800
},
{
"epoch": 0.013709456299091951,
"grad_norm": 4.075947284698486,
"learning_rate": 8.46e-05,
"loss": 3.5548,
"step": 850
},
{
"epoch": 0.014515894904920888,
"grad_norm": 3.9454638957977295,
"learning_rate": 8.960000000000001e-05,
"loss": 3.5653,
"step": 900
},
{
"epoch": 0.015322333510749827,
"grad_norm": 4.633170127868652,
"learning_rate": 9.46e-05,
"loss": 3.3589,
"step": 950
},
{
"epoch": 0.016128772116578764,
"grad_norm": 4.273008346557617,
"learning_rate": 9.960000000000001e-05,
"loss": 3.8061,
"step": 1000
},
{
"epoch": 0.016935210722407702,
"grad_norm": 3.094942808151245,
"learning_rate": 9.999999453205175e-05,
"loss": 3.6132,
"step": 1050
},
{
"epoch": 0.01774164932823664,
"grad_norm": 3.022191047668457,
"learning_rate": 9.999997618496786e-05,
"loss": 3.8753,
"step": 1100
},
{
"epoch": 0.01854808793406558,
"grad_norm": 3.715566635131836,
"learning_rate": 9.999994491740752e-05,
"loss": 3.5987,
"step": 1150
},
{
"epoch": 0.01935452653989452,
"grad_norm": 3.9486517906188965,
"learning_rate": 9.999990072937881e-05,
"loss": 3.7496,
"step": 1200
},
{
"epoch": 0.020160965145723457,
"grad_norm": 4.180655002593994,
"learning_rate": 9.999984362089316e-05,
"loss": 4.1646,
"step": 1250
},
{
"epoch": 0.020967403751552396,
"grad_norm": 4.141788959503174,
"learning_rate": 9.999977359196532e-05,
"loss": 3.8424,
"step": 1300
},
{
"epoch": 0.02177384235738133,
"grad_norm": 6.864413261413574,
"learning_rate": 9.999969064261341e-05,
"loss": 3.79,
"step": 1350
},
{
"epoch": 0.02258028096321027,
"grad_norm": 3.0512356758117676,
"learning_rate": 9.999959477285881e-05,
"loss": 3.7211,
"step": 1400
},
{
"epoch": 0.02338671956903921,
"grad_norm": 3.6865158081054688,
"learning_rate": 9.999948598272635e-05,
"loss": 3.729,
"step": 1450
},
{
"epoch": 0.024193158174868147,
"grad_norm": 3.479696750640869,
"learning_rate": 9.999936427224411e-05,
"loss": 3.8229,
"step": 1500
},
{
"epoch": 0.024999596780697086,
"grad_norm": 3.896265745162964,
"learning_rate": 9.999922964144355e-05,
"loss": 4.0188,
"step": 1550
},
{
"epoch": 0.025806035386526024,
"grad_norm": 1.8305364847183228,
"learning_rate": 9.999908209035947e-05,
"loss": 3.6971,
"step": 1600
},
{
"epoch": 0.026612473992354963,
"grad_norm": 3.6553308963775635,
"learning_rate": 9.999892161902997e-05,
"loss": 3.9225,
"step": 1650
},
{
"epoch": 0.027418912598183902,
"grad_norm": 3.454449415206909,
"learning_rate": 9.999874822749654e-05,
"loss": 4.1062,
"step": 1700
},
{
"epoch": 0.028225351204012837,
"grad_norm": 6.318528652191162,
"learning_rate": 9.9998561915804e-05,
"loss": 4.165,
"step": 1750
},
{
"epoch": 0.029031789809841776,
"grad_norm": 3.34511399269104,
"learning_rate": 9.999836268400044e-05,
"loss": 4.1559,
"step": 1800
},
{
"epoch": 0.029838228415670714,
"grad_norm": 3.3025155067443848,
"learning_rate": 9.999815053213741e-05,
"loss": 3.6104,
"step": 1850
},
{
"epoch": 0.030644667021499653,
"grad_norm": 1.3363052606582642,
"learning_rate": 9.999792546026971e-05,
"loss": 4.3776,
"step": 1900
},
{
"epoch": 0.03145110562732859,
"grad_norm": 4.033843040466309,
"learning_rate": 9.999768746845547e-05,
"loss": 3.838,
"step": 1950
},
{
"epoch": 0.03225754423315753,
"grad_norm": 3.517301082611084,
"learning_rate": 9.999743655675622e-05,
"loss": 4.0525,
"step": 2000
},
{
"epoch": 0.03306398283898647,
"grad_norm": 3.5909571647644043,
"learning_rate": 9.999717272523678e-05,
"loss": 3.672,
"step": 2050
},
{
"epoch": 0.033870421444815405,
"grad_norm": 2.088115930557251,
"learning_rate": 9.999689597396536e-05,
"loss": 3.9254,
"step": 2100
},
{
"epoch": 0.03467686005064435,
"grad_norm": 3.169725179672241,
"learning_rate": 9.999660630301345e-05,
"loss": 3.9372,
"step": 2150
},
{
"epoch": 0.03548329865647328,
"grad_norm": 3.584789991378784,
"learning_rate": 9.999630371245588e-05,
"loss": 3.9506,
"step": 2200
},
{
"epoch": 0.036289737262302224,
"grad_norm": 2.678903579711914,
"learning_rate": 9.999598820237088e-05,
"loss": 3.7822,
"step": 2250
},
{
"epoch": 0.03709617586813116,
"grad_norm": 2.6614108085632324,
"learning_rate": 9.999565977283998e-05,
"loss": 3.9777,
"step": 2300
},
{
"epoch": 0.037902614473960095,
"grad_norm": 3.998502731323242,
"learning_rate": 9.999531842394802e-05,
"loss": 3.8676,
"step": 2350
},
{
"epoch": 0.03870905307978904,
"grad_norm": 3.5454630851745605,
"learning_rate": 9.999496415578324e-05,
"loss": 3.9342,
"step": 2400
},
{
"epoch": 0.03951549168561797,
"grad_norm": 4.353825569152832,
"learning_rate": 9.999459696843719e-05,
"loss": 4.0824,
"step": 2450
},
{
"epoch": 0.040321930291446914,
"grad_norm": 2.4952590465545654,
"learning_rate": 9.99942168620047e-05,
"loss": 3.7487,
"step": 2500
},
{
"epoch": 0.04112836889727585,
"grad_norm": 3.804739475250244,
"learning_rate": 9.999382383658405e-05,
"loss": 3.9278,
"step": 2550
},
{
"epoch": 0.04193480750310479,
"grad_norm": 4.888896465301514,
"learning_rate": 9.999341789227679e-05,
"loss": 3.976,
"step": 2600
},
{
"epoch": 0.04274124610893373,
"grad_norm": 2.8611414432525635,
"learning_rate": 9.99929990291878e-05,
"loss": 3.8008,
"step": 2650
},
{
"epoch": 0.04354768471476266,
"grad_norm": 3.469597816467285,
"learning_rate": 9.999256724742533e-05,
"loss": 3.9826,
"step": 2700
},
{
"epoch": 0.044354123320591604,
"grad_norm": 4.003543376922607,
"learning_rate": 9.999212254710095e-05,
"loss": 3.7801,
"step": 2750
},
{
"epoch": 0.04516056192642054,
"grad_norm": 3.900905132293701,
"learning_rate": 9.99916649283296e-05,
"loss": 4.081,
"step": 2800
},
{
"epoch": 0.04596700053224948,
"grad_norm": 1.68619966506958,
"learning_rate": 9.99911943912295e-05,
"loss": 3.8061,
"step": 2850
},
{
"epoch": 0.04677343913807842,
"grad_norm": 4.65512752532959,
"learning_rate": 9.999071093592227e-05,
"loss": 3.8544,
"step": 2900
},
{
"epoch": 0.04757987774390736,
"grad_norm": 2.2502596378326416,
"learning_rate": 9.999021456253282e-05,
"loss": 4.2133,
"step": 2950
},
{
"epoch": 0.048386316349736294,
"grad_norm": 4.234059810638428,
"learning_rate": 9.998970527118941e-05,
"loss": 3.8952,
"step": 3000
},
{
"epoch": 0.048386316349736294,
"eval_asr_loss": 0.6124729285552434,
"eval_loss": 5.14359712600708,
"eval_runtime": 17.1329,
"eval_samples_per_second": 29.884,
"eval_steps_per_second": 9.981,
"eval_tts_loss": 6.1080074336570815,
"step": 3000
},
{
"epoch": 0.04919275495556523,
"grad_norm": 4.5945658683776855,
"learning_rate": 9.998918306202368e-05,
"loss": 3.7152,
"step": 3050
},
{
"epoch": 0.04999919356139417,
"grad_norm": 3.4985783100128174,
"learning_rate": 9.998864793517053e-05,
"loss": 4.0368,
"step": 3100
},
{
"epoch": 0.05080563216722311,
"grad_norm": 2.991260528564453,
"learning_rate": 9.99880998907683e-05,
"loss": 3.961,
"step": 3150
},
{
"epoch": 0.05161207077305205,
"grad_norm": 7.048410415649414,
"learning_rate": 9.998753892895854e-05,
"loss": 3.9461,
"step": 3200
},
{
"epoch": 0.052418509378880984,
"grad_norm": 4.608008861541748,
"learning_rate": 9.998696504988625e-05,
"loss": 3.863,
"step": 3250
},
{
"epoch": 0.053224947984709926,
"grad_norm": 3.2530033588409424,
"learning_rate": 9.998637825369973e-05,
"loss": 3.8894,
"step": 3300
},
{
"epoch": 0.05403138659053886,
"grad_norm": 2.5475354194641113,
"learning_rate": 9.99857785405506e-05,
"loss": 3.7206,
"step": 3350
},
{
"epoch": 0.054837825196367804,
"grad_norm": 4.753456115722656,
"learning_rate": 9.998516591059384e-05,
"loss": 3.7273,
"step": 3400
},
{
"epoch": 0.05564426380219674,
"grad_norm": 2.6090848445892334,
"learning_rate": 9.998454036398774e-05,
"loss": 3.8879,
"step": 3450
},
{
"epoch": 0.056450702408025674,
"grad_norm": 1.5343598127365112,
"learning_rate": 9.998390190089396e-05,
"loss": 4.0087,
"step": 3500
},
{
"epoch": 0.057257141013854616,
"grad_norm": 4.9131364822387695,
"learning_rate": 9.99832505214775e-05,
"loss": 3.8471,
"step": 3550
},
{
"epoch": 0.05806357961968355,
"grad_norm": 2.1183621883392334,
"learning_rate": 9.998258622590666e-05,
"loss": 3.9289,
"step": 3600
},
{
"epoch": 0.058870018225512494,
"grad_norm": 3.5901293754577637,
"learning_rate": 9.99819090143531e-05,
"loss": 3.8583,
"step": 3650
},
{
"epoch": 0.05967645683134143,
"grad_norm": 2.5325889587402344,
"learning_rate": 9.998121888699183e-05,
"loss": 3.953,
"step": 3700
},
{
"epoch": 0.06048289543717037,
"grad_norm": 2.1782755851745605,
"learning_rate": 9.99805158440012e-05,
"loss": 4.013,
"step": 3750
},
{
"epoch": 0.061289334042999306,
"grad_norm": 2.2709741592407227,
"learning_rate": 9.997979988556285e-05,
"loss": 4.1835,
"step": 3800
},
{
"epoch": 0.06209577264882824,
"grad_norm": 1.7206313610076904,
"learning_rate": 9.997907101186181e-05,
"loss": 3.7028,
"step": 3850
},
{
"epoch": 0.06290221125465718,
"grad_norm": 2.8325114250183105,
"learning_rate": 9.997832922308643e-05,
"loss": 3.8754,
"step": 3900
},
{
"epoch": 0.06370864986048612,
"grad_norm": 3.2632429599761963,
"learning_rate": 9.997757451942839e-05,
"loss": 4.2381,
"step": 3950
},
{
"epoch": 0.06451508846631505,
"grad_norm": 4.061680793762207,
"learning_rate": 9.997680690108271e-05,
"loss": 3.9123,
"step": 4000
},
{
"epoch": 0.065321527072144,
"grad_norm": 3.529866933822632,
"learning_rate": 9.997602636824775e-05,
"loss": 3.8652,
"step": 4050
},
{
"epoch": 0.06612796567797294,
"grad_norm": 3.3136708736419678,
"learning_rate": 9.997523292112521e-05,
"loss": 3.9462,
"step": 4100
},
{
"epoch": 0.06693440428380187,
"grad_norm": 1.6283220052719116,
"learning_rate": 9.997442655992013e-05,
"loss": 4.1202,
"step": 4150
},
{
"epoch": 0.06774084288963081,
"grad_norm": 3.355297088623047,
"learning_rate": 9.997360728484086e-05,
"loss": 4.0791,
"step": 4200
},
{
"epoch": 0.06854728149545974,
"grad_norm": 2.3169095516204834,
"learning_rate": 9.997277509609914e-05,
"loss": 3.6696,
"step": 4250
},
{
"epoch": 0.0693537201012887,
"grad_norm": 2.0649425983428955,
"learning_rate": 9.997192999391e-05,
"loss": 4.1419,
"step": 4300
},
{
"epoch": 0.07016015870711763,
"grad_norm": 2.268578290939331,
"learning_rate": 9.99710719784918e-05,
"loss": 3.9995,
"step": 4350
},
{
"epoch": 0.07096659731294656,
"grad_norm": 2.035593271255493,
"learning_rate": 9.997020105006632e-05,
"loss": 3.7817,
"step": 4400
},
{
"epoch": 0.0717730359187755,
"grad_norm": 1.604856014251709,
"learning_rate": 9.996931720885855e-05,
"loss": 3.8819,
"step": 4450
},
{
"epoch": 0.07257947452460445,
"grad_norm": 1.7782434225082397,
"learning_rate": 9.996842045509692e-05,
"loss": 4.2321,
"step": 4500
},
{
"epoch": 0.07338591313043338,
"grad_norm": 1.1473257541656494,
"learning_rate": 9.996751078901314e-05,
"loss": 4.123,
"step": 4550
},
{
"epoch": 0.07419235173626232,
"grad_norm": 4.596096515655518,
"learning_rate": 9.996658821084231e-05,
"loss": 4.4983,
"step": 4600
},
{
"epoch": 0.07499879034209125,
"grad_norm": 4.046219348907471,
"learning_rate": 9.99656527208228e-05,
"loss": 3.832,
"step": 4650
},
{
"epoch": 0.07580522894792019,
"grad_norm": 3.365840435028076,
"learning_rate": 9.996470431919634e-05,
"loss": 3.8089,
"step": 4700
},
{
"epoch": 0.07661166755374914,
"grad_norm": 2.3088300228118896,
"learning_rate": 9.996374300620805e-05,
"loss": 4.2963,
"step": 4750
},
{
"epoch": 0.07741810615957807,
"grad_norm": 3.0989160537719727,
"learning_rate": 9.996276878210631e-05,
"loss": 3.9085,
"step": 4800
},
{
"epoch": 0.07822454476540701,
"grad_norm": 3.6229944229125977,
"learning_rate": 9.996178164714289e-05,
"loss": 3.9887,
"step": 4850
},
{
"epoch": 0.07903098337123594,
"grad_norm": 2.169739007949829,
"learning_rate": 9.996078160157285e-05,
"loss": 4.053,
"step": 4900
},
{
"epoch": 0.07983742197706489,
"grad_norm": 1.102918267250061,
"learning_rate": 9.995976864565462e-05,
"loss": 3.9429,
"step": 4950
},
{
"epoch": 0.08064386058289383,
"grad_norm": 3.0187597274780273,
"learning_rate": 9.995874277964997e-05,
"loss": 4.1842,
"step": 5000
},
{
"epoch": 0.08145029918872276,
"grad_norm": 6.297091484069824,
"learning_rate": 9.995770400382398e-05,
"loss": 3.8538,
"step": 5050
},
{
"epoch": 0.0822567377945517,
"grad_norm": 2.396947145462036,
"learning_rate": 9.995665231844511e-05,
"loss": 3.9006,
"step": 5100
},
{
"epoch": 0.08306317640038063,
"grad_norm": 1.8022737503051758,
"learning_rate": 9.995558772378509e-05,
"loss": 4.0055,
"step": 5150
},
{
"epoch": 0.08386961500620958,
"grad_norm": 2.959127902984619,
"learning_rate": 9.995451022011902e-05,
"loss": 3.7882,
"step": 5200
},
{
"epoch": 0.08467605361203852,
"grad_norm": 1.0544005632400513,
"learning_rate": 9.995341980772535e-05,
"loss": 3.8215,
"step": 5250
},
{
"epoch": 0.08548249221786745,
"grad_norm": 4.3189616203308105,
"learning_rate": 9.995231648688586e-05,
"loss": 3.8656,
"step": 5300
},
{
"epoch": 0.08628893082369639,
"grad_norm": 2.1376607418060303,
"learning_rate": 9.995120025788565e-05,
"loss": 3.69,
"step": 5350
},
{
"epoch": 0.08709536942952532,
"grad_norm": 2.437530994415283,
"learning_rate": 9.995007112101316e-05,
"loss": 3.972,
"step": 5400
},
{
"epoch": 0.08790180803535427,
"grad_norm": 2.361795663833618,
"learning_rate": 9.994892907656018e-05,
"loss": 3.9913,
"step": 5450
},
{
"epoch": 0.08870824664118321,
"grad_norm": 1.6262253522872925,
"learning_rate": 9.994777412482183e-05,
"loss": 3.9927,
"step": 5500
},
{
"epoch": 0.08951468524701214,
"grad_norm": 4.346071243286133,
"learning_rate": 9.994660626609654e-05,
"loss": 3.7355,
"step": 5550
},
{
"epoch": 0.09032112385284108,
"grad_norm": 3.2261524200439453,
"learning_rate": 9.994542550068612e-05,
"loss": 3.8608,
"step": 5600
},
{
"epoch": 0.09112756245867003,
"grad_norm": 3.2692320346832275,
"learning_rate": 9.994423182889567e-05,
"loss": 4.0809,
"step": 5650
},
{
"epoch": 0.09193400106449896,
"grad_norm": 4.047648906707764,
"learning_rate": 9.994302525103366e-05,
"loss": 3.7864,
"step": 5700
},
{
"epoch": 0.0927404396703279,
"grad_norm": 3.113825559616089,
"learning_rate": 9.994180576741188e-05,
"loss": 3.6328,
"step": 5750
},
{
"epoch": 0.09354687827615683,
"grad_norm": 2.295882225036621,
"learning_rate": 9.994057337834545e-05,
"loss": 3.6168,
"step": 5800
},
{
"epoch": 0.09435331688198577,
"grad_norm": 3.2631022930145264,
"learning_rate": 9.993932808415283e-05,
"loss": 4.1689,
"step": 5850
},
{
"epoch": 0.09515975548781472,
"grad_norm": 3.607879877090454,
"learning_rate": 9.993806988515581e-05,
"loss": 4.0071,
"step": 5900
},
{
"epoch": 0.09596619409364365,
"grad_norm": 1.2808860540390015,
"learning_rate": 9.993679878167957e-05,
"loss": 3.9552,
"step": 5950
},
{
"epoch": 0.09677263269947259,
"grad_norm": 1.4205412864685059,
"learning_rate": 9.99355147740525e-05,
"loss": 4.4215,
"step": 6000
},
{
"epoch": 0.09677263269947259,
"eval_asr_loss": 0.5728164760736751,
"eval_loss": 5.094964027404785,
"eval_runtime": 16.7544,
"eval_samples_per_second": 30.559,
"eval_steps_per_second": 10.206,
"eval_tts_loss": 6.0641644553514515,
"step": 6000
},
{
"epoch": 0.09757907130530152,
"grad_norm": 2.5851426124572754,
"learning_rate": 9.993421786260645e-05,
"loss": 3.941,
"step": 6050
},
{
"epoch": 0.09838550991113046,
"grad_norm": 4.519574165344238,
"learning_rate": 9.993290804767654e-05,
"loss": 3.6548,
"step": 6100
},
{
"epoch": 0.09919194851695941,
"grad_norm": 2.023059368133545,
"learning_rate": 9.993158532960124e-05,
"loss": 3.7762,
"step": 6150
},
{
"epoch": 0.09999838712278834,
"grad_norm": 3.167097330093384,
"learning_rate": 9.993024970872236e-05,
"loss": 3.7466,
"step": 6200
},
{
"epoch": 0.10080482572861728,
"grad_norm": 2.1411142349243164,
"learning_rate": 9.992890118538502e-05,
"loss": 3.7702,
"step": 6250
},
{
"epoch": 0.10161126433444621,
"grad_norm": 2.5199944972991943,
"learning_rate": 9.99275397599377e-05,
"loss": 4.1723,
"step": 6300
},
{
"epoch": 0.10241770294027516,
"grad_norm": 2.112910270690918,
"learning_rate": 9.992616543273222e-05,
"loss": 3.6662,
"step": 6350
},
{
"epoch": 0.1032241415461041,
"grad_norm": 2.290219306945801,
"learning_rate": 9.992477820412369e-05,
"loss": 4.1363,
"step": 6400
},
{
"epoch": 0.10403058015193303,
"grad_norm": 3.748711585998535,
"learning_rate": 9.992337807447061e-05,
"loss": 4.0218,
"step": 6450
},
{
"epoch": 0.10483701875776197,
"grad_norm": 14.197625160217285,
"learning_rate": 9.992196504413478e-05,
"loss": 4.11,
"step": 6500
},
{
"epoch": 0.1056434573635909,
"grad_norm": 2.3828155994415283,
"learning_rate": 9.992053911348132e-05,
"loss": 3.9976,
"step": 6550
},
{
"epoch": 0.10644989596941985,
"grad_norm": 3.1666653156280518,
"learning_rate": 9.991910028287874e-05,
"loss": 3.8263,
"step": 6600
},
{
"epoch": 0.10725633457524879,
"grad_norm": 1.068214774131775,
"learning_rate": 9.991764855269882e-05,
"loss": 3.7559,
"step": 6650
},
{
"epoch": 0.10806277318107772,
"grad_norm": 2.6692686080932617,
"learning_rate": 9.991618392331672e-05,
"loss": 3.8083,
"step": 6700
},
{
"epoch": 0.10886921178690666,
"grad_norm": 3.22259783744812,
"learning_rate": 9.99147063951109e-05,
"loss": 3.9687,
"step": 6750
},
{
"epoch": 0.10967565039273561,
"grad_norm": 2.2858736515045166,
"learning_rate": 9.991321596846317e-05,
"loss": 3.7381,
"step": 6800
},
{
"epoch": 0.11048208899856454,
"grad_norm": 3.2024693489074707,
"learning_rate": 9.991171264375869e-05,
"loss": 4.1968,
"step": 6850
},
{
"epoch": 0.11128852760439348,
"grad_norm": 2.7798397541046143,
"learning_rate": 9.99101964213859e-05,
"loss": 3.7984,
"step": 6900
},
{
"epoch": 0.11209496621022241,
"grad_norm": 2.2842605113983154,
"learning_rate": 9.990866730173662e-05,
"loss": 4.2311,
"step": 6950
},
{
"epoch": 0.11290140481605135,
"grad_norm": 4.10762357711792,
"learning_rate": 9.9907125285206e-05,
"loss": 4.1217,
"step": 7000
},
{
"epoch": 0.1137078434218803,
"grad_norm": 1.8413517475128174,
"learning_rate": 9.99055703721925e-05,
"loss": 4.1848,
"step": 7050
},
{
"epoch": 0.11451428202770923,
"grad_norm": 2.565972328186035,
"learning_rate": 9.990400256309792e-05,
"loss": 3.744,
"step": 7100
},
{
"epoch": 0.11532072063353817,
"grad_norm": 2.643573760986328,
"learning_rate": 9.990242185832744e-05,
"loss": 4.2744,
"step": 7150
},
{
"epoch": 0.1161271592393671,
"grad_norm": 2.1358284950256348,
"learning_rate": 9.990082825828946e-05,
"loss": 4.2883,
"step": 7200
},
{
"epoch": 0.11693359784519604,
"grad_norm": 1.9578022956848145,
"learning_rate": 9.989922176339583e-05,
"loss": 3.8243,
"step": 7250
},
{
"epoch": 0.11774003645102499,
"grad_norm": 5.736109256744385,
"learning_rate": 9.989760237406168e-05,
"loss": 4.1798,
"step": 7300
},
{
"epoch": 0.11854647505685392,
"grad_norm": 3.454503059387207,
"learning_rate": 9.989597009070546e-05,
"loss": 4.1567,
"step": 7350
},
{
"epoch": 0.11935291366268286,
"grad_norm": 3.161022424697876,
"learning_rate": 9.989432491374897e-05,
"loss": 3.7965,
"step": 7400
},
{
"epoch": 0.1201593522685118,
"grad_norm": 2.6754753589630127,
"learning_rate": 9.989266684361735e-05,
"loss": 3.9182,
"step": 7450
},
{
"epoch": 0.12096579087434074,
"grad_norm": 2.2374236583709717,
"learning_rate": 9.989099588073905e-05,
"loss": 4.1068,
"step": 7500
},
{
"epoch": 0.12177222948016968,
"grad_norm": 2.775463581085205,
"learning_rate": 9.988931202554588e-05,
"loss": 4.0023,
"step": 7550
},
{
"epoch": 0.12257866808599861,
"grad_norm": 2.515188217163086,
"learning_rate": 9.988761527847293e-05,
"loss": 4.0234,
"step": 7600
},
{
"epoch": 0.12338510669182755,
"grad_norm": 4.208799362182617,
"learning_rate": 9.988590563995869e-05,
"loss": 3.9328,
"step": 7650
},
{
"epoch": 0.12419154529765648,
"grad_norm": 2.9487059116363525,
"learning_rate": 9.988418311044495e-05,
"loss": 3.8718,
"step": 7700
},
{
"epoch": 0.12499798390348543,
"grad_norm": 2.8759098052978516,
"learning_rate": 9.98824476903768e-05,
"loss": 3.8538,
"step": 7750
},
{
"epoch": 0.12580442250931437,
"grad_norm": 2.848405599594116,
"learning_rate": 9.98806993802027e-05,
"loss": 3.8749,
"step": 7800
},
{
"epoch": 0.1266108611151433,
"grad_norm": 1.7490386962890625,
"learning_rate": 9.987893818037443e-05,
"loss": 3.9094,
"step": 7850
},
{
"epoch": 0.12741729972097224,
"grad_norm": 2.6559858322143555,
"learning_rate": 9.987716409134712e-05,
"loss": 3.825,
"step": 7900
},
{
"epoch": 0.12822373832680117,
"grad_norm": 2.877697229385376,
"learning_rate": 9.987537711357919e-05,
"loss": 3.9497,
"step": 7950
},
{
"epoch": 0.1290301769326301,
"grad_norm": 2.17142391204834,
"learning_rate": 9.987357724753242e-05,
"loss": 3.9387,
"step": 8000
},
{
"epoch": 0.12983661553845904,
"grad_norm": 1.643707036972046,
"learning_rate": 9.987176449367192e-05,
"loss": 4.2619,
"step": 8050
},
{
"epoch": 0.130643054144288,
"grad_norm": 4.4552998542785645,
"learning_rate": 9.98699388524661e-05,
"loss": 4.0635,
"step": 8100
},
{
"epoch": 0.13144949275011694,
"grad_norm": 3.406740427017212,
"learning_rate": 9.986810032438674e-05,
"loss": 4.0747,
"step": 8150
},
{
"epoch": 0.13225593135594588,
"grad_norm": 5.450404644012451,
"learning_rate": 9.986624890990893e-05,
"loss": 3.8341,
"step": 8200
},
{
"epoch": 0.1330623699617748,
"grad_norm": 2.6635243892669678,
"learning_rate": 9.986438460951111e-05,
"loss": 3.8464,
"step": 8250
},
{
"epoch": 0.13386880856760375,
"grad_norm": 1.887026309967041,
"learning_rate": 9.9862507423675e-05,
"loss": 3.7535,
"step": 8300
},
{
"epoch": 0.13467524717343268,
"grad_norm": 2.086843490600586,
"learning_rate": 9.98606173528857e-05,
"loss": 4.2659,
"step": 8350
},
{
"epoch": 0.13548168577926162,
"grad_norm": 3.2618448734283447,
"learning_rate": 9.985871439763163e-05,
"loss": 4.0077,
"step": 8400
},
{
"epoch": 0.13628812438509055,
"grad_norm": 3.416794538497925,
"learning_rate": 9.985679855840453e-05,
"loss": 4.2068,
"step": 8450
},
{
"epoch": 0.1370945629909195,
"grad_norm": 2.5768439769744873,
"learning_rate": 9.985486983569945e-05,
"loss": 4.2102,
"step": 8500
},
{
"epoch": 0.13790100159674845,
"grad_norm": 2.0248680114746094,
"learning_rate": 9.985296718837846e-05,
"loss": 3.7619,
"step": 8550
},
{
"epoch": 0.1387074402025774,
"grad_norm": 2.308393955230713,
"learning_rate": 9.985101295786062e-05,
"loss": 4.0467,
"step": 8600
},
{
"epoch": 0.13951387880840632,
"grad_norm": 3.1832563877105713,
"learning_rate": 9.984904584535987e-05,
"loss": 4.0234,
"step": 8650
},
{
"epoch": 0.14032031741423526,
"grad_norm": 2.4449775218963623,
"learning_rate": 9.984706585138452e-05,
"loss": 3.4602,
"step": 8700
},
{
"epoch": 0.1411267560200642,
"grad_norm": 3.088870048522949,
"learning_rate": 9.984507297644625e-05,
"loss": 4.0023,
"step": 8750
},
{
"epoch": 0.14193319462589313,
"grad_norm": 1.3331509828567505,
"learning_rate": 9.984306722106e-05,
"loss": 4.4195,
"step": 8800
},
{
"epoch": 0.14273963323172206,
"grad_norm": 2.1085762977600098,
"learning_rate": 9.98410485857441e-05,
"loss": 3.6874,
"step": 8850
},
{
"epoch": 0.143546071837551,
"grad_norm": 2.382269859313965,
"learning_rate": 9.983905782752949e-05,
"loss": 4.1141,
"step": 8900
},
{
"epoch": 0.14435251044337993,
"grad_norm": 1.5278171300888062,
"learning_rate": 9.983701369149501e-05,
"loss": 3.8317,
"step": 8950
},
{
"epoch": 0.1451589490492089,
"grad_norm": 5.558732986450195,
"learning_rate": 9.983495667709518e-05,
"loss": 3.6433,
"step": 9000
},
{
"epoch": 0.1451589490492089,
"eval_asr_loss": 0.5558241450032482,
"eval_loss": 5.0716023445129395,
"eval_runtime": 15.4575,
"eval_samples_per_second": 33.123,
"eval_steps_per_second": 11.063,
"eval_tts_loss": 6.0293835261628255,
"step": 9000
},
{
"epoch": 0.14596538765503783,
"grad_norm": 4.300561904907227,
"learning_rate": 9.983288678486153e-05,
"loss": 4.3128,
"step": 9050
},
{
"epoch": 0.14677182626086677,
"grad_norm": 1.5210169553756714,
"learning_rate": 9.983080401532894e-05,
"loss": 3.9063,
"step": 9100
},
{
"epoch": 0.1475782648666957,
"grad_norm": 2.6939990520477295,
"learning_rate": 9.982870836903564e-05,
"loss": 3.9087,
"step": 9150
},
{
"epoch": 0.14838470347252464,
"grad_norm": 1.7290055751800537,
"learning_rate": 9.982659984652316e-05,
"loss": 4.1546,
"step": 9200
},
{
"epoch": 0.14919114207835357,
"grad_norm": 4.180938243865967,
"learning_rate": 9.982447844833634e-05,
"loss": 4.0063,
"step": 9250
},
{
"epoch": 0.1499975806841825,
"grad_norm": 3.3321378231048584,
"learning_rate": 9.98223441750234e-05,
"loss": 3.6936,
"step": 9300
},
{
"epoch": 0.15080401929001144,
"grad_norm": 3.3349199295043945,
"learning_rate": 9.982019702713583e-05,
"loss": 3.9439,
"step": 9350
},
{
"epoch": 0.15161045789584038,
"grad_norm": 2.3645851612091064,
"learning_rate": 9.98180370052285e-05,
"loss": 4.0444,
"step": 9400
},
{
"epoch": 0.15241689650166934,
"grad_norm": 3.225592613220215,
"learning_rate": 9.981586410985955e-05,
"loss": 3.8988,
"step": 9450
},
{
"epoch": 0.15322333510749828,
"grad_norm": 3.2834198474884033,
"learning_rate": 9.981367834159049e-05,
"loss": 4.1237,
"step": 9500
},
{
"epoch": 0.1540297737133272,
"grad_norm": 1.3256886005401611,
"learning_rate": 9.981147970098615e-05,
"loss": 3.9196,
"step": 9550
},
{
"epoch": 0.15483621231915615,
"grad_norm": 3.559396743774414,
"learning_rate": 9.980926818861469e-05,
"loss": 4.1997,
"step": 9600
},
{
"epoch": 0.15564265092498508,
"grad_norm": 3.722397565841675,
"learning_rate": 9.980704380504756e-05,
"loss": 4.2779,
"step": 9650
},
{
"epoch": 0.15644908953081402,
"grad_norm": 3.125638723373413,
"learning_rate": 9.980480655085957e-05,
"loss": 4.0461,
"step": 9700
},
{
"epoch": 0.15725552813664295,
"grad_norm": 1.543898105621338,
"learning_rate": 9.980255642662888e-05,
"loss": 3.9528,
"step": 9750
},
{
"epoch": 0.1580619667424719,
"grad_norm": 3.1968443393707275,
"learning_rate": 9.980029343293688e-05,
"loss": 4.2141,
"step": 9800
},
{
"epoch": 0.15886840534830082,
"grad_norm": 1.8708412647247314,
"learning_rate": 9.979801757036841e-05,
"loss": 4.1029,
"step": 9850
},
{
"epoch": 0.15967484395412979,
"grad_norm": 4.089094161987305,
"learning_rate": 9.979572883951156e-05,
"loss": 4.0536,
"step": 9900
},
{
"epoch": 0.16048128255995872,
"grad_norm": 1.7748725414276123,
"learning_rate": 9.979342724095774e-05,
"loss": 4.0823,
"step": 9950
},
{
"epoch": 0.16128772116578766,
"grad_norm": 2.352545976638794,
"learning_rate": 9.97911127753017e-05,
"loss": 3.8066,
"step": 10000
},
{
"epoch": 0.1620941597716166,
"grad_norm": 2.2634575366973877,
"learning_rate": 9.978878544314156e-05,
"loss": 4.1763,
"step": 10050
},
{
"epoch": 0.16290059837744553,
"grad_norm": 2.911471366882324,
"learning_rate": 9.978644524507869e-05,
"loss": 4.2928,
"step": 10100
},
{
"epoch": 0.16370703698327446,
"grad_norm": 1.9891984462738037,
"learning_rate": 9.978409218171784e-05,
"loss": 4.0966,
"step": 10150
},
{
"epoch": 0.1645134755891034,
"grad_norm": 2.632568597793579,
"learning_rate": 9.978172625366705e-05,
"loss": 4.2746,
"step": 10200
},
{
"epoch": 0.16531991419493233,
"grad_norm": 2.3219494819641113,
"learning_rate": 9.977934746153771e-05,
"loss": 4.3219,
"step": 10250
},
{
"epoch": 0.16612635280076127,
"grad_norm": 2.416696071624756,
"learning_rate": 9.977695580594451e-05,
"loss": 4.0857,
"step": 10300
},
{
"epoch": 0.1669327914065902,
"grad_norm": 1.758725881576538,
"learning_rate": 9.97745512875055e-05,
"loss": 3.7797,
"step": 10350
},
{
"epoch": 0.16773923001241917,
"grad_norm": 2.4825093746185303,
"learning_rate": 9.977213390684199e-05,
"loss": 3.8742,
"step": 10400
},
{
"epoch": 0.1685456686182481,
"grad_norm": 1.305153727531433,
"learning_rate": 9.97697036645787e-05,
"loss": 4.1835,
"step": 10450
},
{
"epoch": 0.16935210722407704,
"grad_norm": 4.981587886810303,
"learning_rate": 9.97672605613436e-05,
"loss": 4.2102,
"step": 10500
},
{
"epoch": 0.17015854582990597,
"grad_norm": 2.5329973697662354,
"learning_rate": 9.976480459776803e-05,
"loss": 4.1571,
"step": 10550
},
{
"epoch": 0.1709649844357349,
"grad_norm": 3.1804792881011963,
"learning_rate": 9.976233577448661e-05,
"loss": 4.0007,
"step": 10600
},
{
"epoch": 0.17177142304156384,
"grad_norm": 2.1059024333953857,
"learning_rate": 9.975985409213731e-05,
"loss": 4.096,
"step": 10650
},
{
"epoch": 0.17257786164739278,
"grad_norm": 2.35027813911438,
"learning_rate": 9.975735955136145e-05,
"loss": 4.2416,
"step": 10700
},
{
"epoch": 0.1733843002532217,
"grad_norm": 4.606429576873779,
"learning_rate": 9.975485215280362e-05,
"loss": 3.7404,
"step": 10750
},
{
"epoch": 0.17419073885905065,
"grad_norm": 2.3107430934906006,
"learning_rate": 9.975233189711177e-05,
"loss": 3.9696,
"step": 10800
},
{
"epoch": 0.1749971774648796,
"grad_norm": 1.1612375974655151,
"learning_rate": 9.974979878493716e-05,
"loss": 4.0279,
"step": 10850
},
{
"epoch": 0.17580361607070855,
"grad_norm": 1.4754469394683838,
"learning_rate": 9.974725281693433e-05,
"loss": 3.5668,
"step": 10900
},
{
"epoch": 0.17661005467653748,
"grad_norm": 1.7385231256484985,
"learning_rate": 9.974469399376123e-05,
"loss": 3.8299,
"step": 10950
},
{
"epoch": 0.17741649328236642,
"grad_norm": 2.5821852684020996,
"learning_rate": 9.974212231607907e-05,
"loss": 3.8388,
"step": 11000
},
{
"epoch": 0.17822293188819535,
"grad_norm": 1.8914719820022583,
"learning_rate": 9.973953778455239e-05,
"loss": 4.0202,
"step": 11050
},
{
"epoch": 0.1790293704940243,
"grad_norm": 1.3121271133422852,
"learning_rate": 9.973694039984907e-05,
"loss": 4.0743,
"step": 11100
},
{
"epoch": 0.17983580909985322,
"grad_norm": 2.854396104812622,
"learning_rate": 9.973433016264029e-05,
"loss": 4.066,
"step": 11150
},
{
"epoch": 0.18064224770568216,
"grad_norm": 3.5787723064422607,
"learning_rate": 9.973170707360055e-05,
"loss": 4.3989,
"step": 11200
},
{
"epoch": 0.1814486863115111,
"grad_norm": 1.8557301759719849,
"learning_rate": 9.972912397814846e-05,
"loss": 4.0768,
"step": 11250
},
{
"epoch": 0.18225512491734006,
"grad_norm": 2.253882884979248,
"learning_rate": 9.97264754444864e-05,
"loss": 3.9537,
"step": 11300
},
{
"epoch": 0.183061563523169,
"grad_norm": 3.240039825439453,
"learning_rate": 9.972381406102311e-05,
"loss": 4.0367,
"step": 11350
},
{
"epoch": 0.18386800212899793,
"grad_norm": 3.579876184463501,
"learning_rate": 9.972113982844634e-05,
"loss": 3.8319,
"step": 11400
},
{
"epoch": 0.18467444073482686,
"grad_norm": 0.6159800291061401,
"learning_rate": 9.971845274744713e-05,
"loss": 3.8889,
"step": 11450
},
{
"epoch": 0.1854808793406558,
"grad_norm": 1.1513911485671997,
"learning_rate": 9.971575281871985e-05,
"loss": 3.8983,
"step": 11500
},
{
"epoch": 0.18628731794648473,
"grad_norm": 3.410205125808716,
"learning_rate": 9.971304004296219e-05,
"loss": 4.2003,
"step": 11550
},
{
"epoch": 0.18709375655231367,
"grad_norm": 1.6599464416503906,
"learning_rate": 9.971031442087515e-05,
"loss": 3.7348,
"step": 11600
},
{
"epoch": 0.1879001951581426,
"grad_norm": 2.8847126960754395,
"learning_rate": 9.970757595316306e-05,
"loss": 3.9562,
"step": 11650
},
{
"epoch": 0.18870663376397154,
"grad_norm": 2.053612470626831,
"learning_rate": 9.970482464053359e-05,
"loss": 4.3117,
"step": 11700
},
{
"epoch": 0.1895130723698005,
"grad_norm": 2.1683125495910645,
"learning_rate": 9.970206048369766e-05,
"loss": 3.8281,
"step": 11750
},
{
"epoch": 0.19031951097562944,
"grad_norm": 5.38385009765625,
"learning_rate": 9.969928348336959e-05,
"loss": 4.1646,
"step": 11800
},
{
"epoch": 0.19112594958145837,
"grad_norm": 1.783724069595337,
"learning_rate": 9.969649364026696e-05,
"loss": 3.6177,
"step": 11850
},
{
"epoch": 0.1919323881872873,
"grad_norm": 2.6397554874420166,
"learning_rate": 9.969369095511072e-05,
"loss": 4.1144,
"step": 11900
},
{
"epoch": 0.19273882679311624,
"grad_norm": 3.0491504669189453,
"learning_rate": 9.969087542862509e-05,
"loss": 4.1832,
"step": 11950
},
{
"epoch": 0.19354526539894518,
"grad_norm": 3.2440996170043945,
"learning_rate": 9.968804706153762e-05,
"loss": 3.9472,
"step": 12000
},
{
"epoch": 0.19354526539894518,
"eval_asr_loss": 0.5453353807244307,
"eval_loss": 5.0482306480407715,
"eval_runtime": 15.3432,
"eval_samples_per_second": 33.37,
"eval_steps_per_second": 11.145,
"eval_tts_loss": 6.005206208861255,
"step": 12000
},
{
"epoch": 0.1943517040047741,
"grad_norm": 4.186015605926514,
"learning_rate": 9.968520585457922e-05,
"loss": 3.8557,
"step": 12050
},
{
"epoch": 0.19515814261060305,
"grad_norm": 2.497833013534546,
"learning_rate": 9.968235180848406e-05,
"loss": 4.2129,
"step": 12100
},
{
"epoch": 0.19596458121643198,
"grad_norm": 3.930877685546875,
"learning_rate": 9.967948492398967e-05,
"loss": 4.3072,
"step": 12150
},
{
"epoch": 0.19677101982226092,
"grad_norm": 1.6300374269485474,
"learning_rate": 9.967660520183686e-05,
"loss": 3.6067,
"step": 12200
},
{
"epoch": 0.19757745842808988,
"grad_norm": 1.7998709678649902,
"learning_rate": 9.96737126427698e-05,
"loss": 4.2948,
"step": 12250
},
{
"epoch": 0.19838389703391882,
"grad_norm": 2.9288175106048584,
"learning_rate": 9.967080724753592e-05,
"loss": 3.6403,
"step": 12300
},
{
"epoch": 0.19919033563974775,
"grad_norm": 1.8051952123641968,
"learning_rate": 9.966788901688605e-05,
"loss": 3.9332,
"step": 12350
},
{
"epoch": 0.1999967742455767,
"grad_norm": 2.3929975032806396,
"learning_rate": 9.966495795157427e-05,
"loss": 3.934,
"step": 12400
},
{
"epoch": 0.20080321285140562,
"grad_norm": 2.7234630584716797,
"learning_rate": 9.966201405235797e-05,
"loss": 4.1635,
"step": 12450
},
{
"epoch": 0.20160965145723456,
"grad_norm": 2.5407304763793945,
"learning_rate": 9.965905731999791e-05,
"loss": 3.6558,
"step": 12500
},
{
"epoch": 0.2024160900630635,
"grad_norm": 1.8946152925491333,
"learning_rate": 9.965608775525815e-05,
"loss": 3.7618,
"step": 12550
},
{
"epoch": 0.20322252866889243,
"grad_norm": 2.083383798599243,
"learning_rate": 9.965310535890601e-05,
"loss": 4.3816,
"step": 12600
},
{
"epoch": 0.20402896727472136,
"grad_norm": 2.6187212467193604,
"learning_rate": 9.96501101317122e-05,
"loss": 3.9503,
"step": 12650
},
{
"epoch": 0.20483540588055033,
"grad_norm": 4.162103176116943,
"learning_rate": 9.964710207445072e-05,
"loss": 4.0746,
"step": 12700
},
{
"epoch": 0.20564184448637926,
"grad_norm": 2.638073682785034,
"learning_rate": 9.964408118789886e-05,
"loss": 4.3604,
"step": 12750
},
{
"epoch": 0.2064482830922082,
"grad_norm": 2.0001702308654785,
"learning_rate": 9.964104747283727e-05,
"loss": 4.0921,
"step": 12800
},
{
"epoch": 0.20725472169803713,
"grad_norm": 2.3797965049743652,
"learning_rate": 9.963800093004987e-05,
"loss": 4.3402,
"step": 12850
},
{
"epoch": 0.20806116030386607,
"grad_norm": 2.3531606197357178,
"learning_rate": 9.963494156032393e-05,
"loss": 4.0507,
"step": 12900
},
{
"epoch": 0.208867598909695,
"grad_norm": 3.526831865310669,
"learning_rate": 9.963186936445001e-05,
"loss": 4.3212,
"step": 12950
},
{
"epoch": 0.20967403751552394,
"grad_norm": 4.21464729309082,
"learning_rate": 9.962878434322202e-05,
"loss": 4.1086,
"step": 13000
},
{
"epoch": 0.21048047612135287,
"grad_norm": 2.4508001804351807,
"learning_rate": 9.962568649743713e-05,
"loss": 3.6248,
"step": 13050
},
{
"epoch": 0.2112869147271818,
"grad_norm": 4.681621551513672,
"learning_rate": 9.962257582789585e-05,
"loss": 3.9501,
"step": 13100
},
{
"epoch": 0.21209335333301077,
"grad_norm": 3.483264207839966,
"learning_rate": 9.961945233540204e-05,
"loss": 3.7879,
"step": 13150
},
{
"epoch": 0.2128997919388397,
"grad_norm": 2.662280321121216,
"learning_rate": 9.961631602076282e-05,
"loss": 4.2567,
"step": 13200
},
{
"epoch": 0.21370623054466864,
"grad_norm": 1.8116915225982666,
"learning_rate": 9.961316688478863e-05,
"loss": 4.1235,
"step": 13250
},
{
"epoch": 0.21451266915049758,
"grad_norm": 3.8289411067962646,
"learning_rate": 9.961000492829326e-05,
"loss": 4.0276,
"step": 13300
},
{
"epoch": 0.2153191077563265,
"grad_norm": 1.5876435041427612,
"learning_rate": 9.96068301520938e-05,
"loss": 3.7603,
"step": 13350
},
{
"epoch": 0.21612554636215545,
"grad_norm": 3.7507176399230957,
"learning_rate": 9.960370643453201e-05,
"loss": 3.6352,
"step": 13400
},
{
"epoch": 0.21693198496798438,
"grad_norm": 1.9685125350952148,
"learning_rate": 9.96005062777419e-05,
"loss": 3.6024,
"step": 13450
},
{
"epoch": 0.21773842357381332,
"grad_norm": 2.623763084411621,
"learning_rate": 9.959729330370225e-05,
"loss": 4.0665,
"step": 13500
},
{
"epoch": 0.21854486217964225,
"grad_norm": 1.6213947534561157,
"learning_rate": 9.95940675132433e-05,
"loss": 3.9128,
"step": 13550
},
{
"epoch": 0.21935130078547121,
"grad_norm": 5.413779258728027,
"learning_rate": 9.959082890719862e-05,
"loss": 3.7172,
"step": 13600
},
{
"epoch": 0.22015773939130015,
"grad_norm": 3.002934694290161,
"learning_rate": 9.95875774864051e-05,
"loss": 3.7273,
"step": 13650
},
{
"epoch": 0.22096417799712909,
"grad_norm": 2.6499838829040527,
"learning_rate": 9.958431325170294e-05,
"loss": 4.2387,
"step": 13700
},
{
"epoch": 0.22177061660295802,
"grad_norm": 1.5989998579025269,
"learning_rate": 9.958103620393564e-05,
"loss": 4.0674,
"step": 13750
},
{
"epoch": 0.22257705520878696,
"grad_norm": 2.529733180999756,
"learning_rate": 9.957774634395003e-05,
"loss": 4.1921,
"step": 13800
},
{
"epoch": 0.2233834938146159,
"grad_norm": 6.773332595825195,
"learning_rate": 9.957444367259625e-05,
"loss": 3.947,
"step": 13850
},
{
"epoch": 0.22418993242044483,
"grad_norm": 1.7639790773391724,
"learning_rate": 9.957112819072774e-05,
"loss": 3.9777,
"step": 13900
},
{
"epoch": 0.22499637102627376,
"grad_norm": 1.8540008068084717,
"learning_rate": 9.956779989920124e-05,
"loss": 4.0806,
"step": 13950
},
{
"epoch": 0.2258028096321027,
"grad_norm": 2.861515998840332,
"learning_rate": 9.956445879887683e-05,
"loss": 4.1824,
"step": 14000
},
{
"epoch": 0.22660924823793166,
"grad_norm": 2.093233823776245,
"learning_rate": 9.956110489061786e-05,
"loss": 4.0445,
"step": 14050
},
{
"epoch": 0.2274156868437606,
"grad_norm": 2.043487071990967,
"learning_rate": 9.955773817529103e-05,
"loss": 3.8989,
"step": 14100
},
{
"epoch": 0.22822212544958953,
"grad_norm": 1.743752121925354,
"learning_rate": 9.955435865376632e-05,
"loss": 4.2283,
"step": 14150
},
{
"epoch": 0.22902856405541847,
"grad_norm": 1.1774280071258545,
"learning_rate": 9.955096632691704e-05,
"loss": 3.8492,
"step": 14200
},
{
"epoch": 0.2298350026612474,
"grad_norm": 0.7757024168968201,
"learning_rate": 9.95475611956198e-05,
"loss": 3.8286,
"step": 14250
},
{
"epoch": 0.23064144126707634,
"grad_norm": 2.530691623687744,
"learning_rate": 9.954414326075452e-05,
"loss": 4.0583,
"step": 14300
},
{
"epoch": 0.23144787987290527,
"grad_norm": 2.322361469268799,
"learning_rate": 9.95407125232044e-05,
"loss": 4.188,
"step": 14350
},
{
"epoch": 0.2322543184787342,
"grad_norm": 2.6633105278015137,
"learning_rate": 9.953726898385603e-05,
"loss": 3.6949,
"step": 14400
},
{
"epoch": 0.23306075708456314,
"grad_norm": 3.243053436279297,
"learning_rate": 9.95338126435992e-05,
"loss": 3.733,
"step": 14450
},
{
"epoch": 0.23386719569039208,
"grad_norm": 2.6763391494750977,
"learning_rate": 9.953034350332709e-05,
"loss": 3.9605,
"step": 14500
},
{
"epoch": 0.23467363429622104,
"grad_norm": 2.8131678104400635,
"learning_rate": 9.952686156393616e-05,
"loss": 4.2417,
"step": 14550
},
{
"epoch": 0.23548007290204997,
"grad_norm": 2.5453968048095703,
"learning_rate": 9.952336682632617e-05,
"loss": 4.5493,
"step": 14600
},
{
"epoch": 0.2362865115078789,
"grad_norm": 2.9333486557006836,
"learning_rate": 9.951985929140018e-05,
"loss": 4.0338,
"step": 14650
},
{
"epoch": 0.23709295011370785,
"grad_norm": 1.5506014823913574,
"learning_rate": 9.951633896006461e-05,
"loss": 3.9207,
"step": 14700
},
{
"epoch": 0.23789938871953678,
"grad_norm": 1.8103605508804321,
"learning_rate": 9.95128058332291e-05,
"loss": 3.9589,
"step": 14750
},
{
"epoch": 0.23870582732536572,
"grad_norm": 2.344365358352661,
"learning_rate": 9.950925991180669e-05,
"loss": 4.0624,
"step": 14800
},
{
"epoch": 0.23951226593119465,
"grad_norm": 2.332612991333008,
"learning_rate": 9.950570119671362e-05,
"loss": 3.9549,
"step": 14850
},
{
"epoch": 0.2403187045370236,
"grad_norm": 2.706002712249756,
"learning_rate": 9.950212968886956e-05,
"loss": 3.6544,
"step": 14900
},
{
"epoch": 0.24112514314285252,
"grad_norm": 2.335348606109619,
"learning_rate": 9.94985453891974e-05,
"loss": 3.9361,
"step": 14950
},
{
"epoch": 0.24193158174868148,
"grad_norm": 1.8754454851150513,
"learning_rate": 9.949494829862334e-05,
"loss": 3.6963,
"step": 15000
},
{
"epoch": 0.24193158174868148,
"eval_asr_loss": 0.560360777742155,
"eval_loss": 5.027297496795654,
"eval_runtime": 15.4664,
"eval_samples_per_second": 33.104,
"eval_steps_per_second": 11.056,
"eval_tts_loss": 5.95218636007564,
"step": 15000
},
{
"epoch": 0.24273802035451042,
"grad_norm": 1.9016485214233398,
"learning_rate": 9.949133841807691e-05,
"loss": 3.9233,
"step": 15050
},
{
"epoch": 0.24354445896033936,
"grad_norm": 2.494946002960205,
"learning_rate": 9.948771574849097e-05,
"loss": 4.087,
"step": 15100
},
{
"epoch": 0.2443508975661683,
"grad_norm": 3.006345748901367,
"learning_rate": 9.948408029080161e-05,
"loss": 4.0706,
"step": 15150
},
{
"epoch": 0.24515733617199723,
"grad_norm": 3.3929615020751953,
"learning_rate": 9.94804320459483e-05,
"loss": 3.9753,
"step": 15200
},
{
"epoch": 0.24596377477782616,
"grad_norm": 1.2217860221862793,
"learning_rate": 9.947677101487374e-05,
"loss": 4.1967,
"step": 15250
},
{
"epoch": 0.2467702133836551,
"grad_norm": 1.409860610961914,
"learning_rate": 9.947309719852402e-05,
"loss": 3.9026,
"step": 15300
},
{
"epoch": 0.24757665198948403,
"grad_norm": 1.3178203105926514,
"learning_rate": 9.946941059784847e-05,
"loss": 4.1695,
"step": 15350
},
{
"epoch": 0.24838309059531297,
"grad_norm": 3.196470260620117,
"learning_rate": 9.946571121379974e-05,
"loss": 3.8572,
"step": 15400
},
{
"epoch": 0.24918952920114193,
"grad_norm": 2.9058384895324707,
"learning_rate": 9.946199904733378e-05,
"loss": 4.2121,
"step": 15450
},
{
"epoch": 0.24999596780697086,
"grad_norm": 2.940702438354492,
"learning_rate": 9.945827409940989e-05,
"loss": 4.2432,
"step": 15500
},
{
"epoch": 0.25080240641279977,
"grad_norm": 3.044480562210083,
"learning_rate": 9.945453637099058e-05,
"loss": 4.0084,
"step": 15550
},
{
"epoch": 0.25160884501862874,
"grad_norm": 2.350438356399536,
"learning_rate": 9.945078586304175e-05,
"loss": 4.1791,
"step": 15600
},
{
"epoch": 0.25241528362445764,
"grad_norm": 2.298433542251587,
"learning_rate": 9.944702257653255e-05,
"loss": 3.9251,
"step": 15650
},
{
"epoch": 0.2532217222302866,
"grad_norm": 2.5852606296539307,
"learning_rate": 9.944324651243545e-05,
"loss": 3.9519,
"step": 15700
},
{
"epoch": 0.25402816083611557,
"grad_norm": 1.654619812965393,
"learning_rate": 9.943945767172624e-05,
"loss": 3.7628,
"step": 15750
},
{
"epoch": 0.2548345994419445,
"grad_norm": 2.320345878601074,
"learning_rate": 9.943573221290566e-05,
"loss": 4.3502,
"step": 15800
},
{
"epoch": 0.25564103804777344,
"grad_norm": 1.5587137937545776,
"learning_rate": 9.943191807739609e-05,
"loss": 4.3284,
"step": 15850
},
{
"epoch": 0.25644747665360235,
"grad_norm": 2.629159927368164,
"learning_rate": 9.942809116820176e-05,
"loss": 3.9416,
"step": 15900
},
{
"epoch": 0.2572539152594313,
"grad_norm": 2.840963363647461,
"learning_rate": 9.942425148631161e-05,
"loss": 4.0621,
"step": 15950
},
{
"epoch": 0.2580603538652602,
"grad_norm": 1.6260427236557007,
"learning_rate": 9.942039903271782e-05,
"loss": 3.7273,
"step": 16000
},
{
"epoch": 0.2588667924710892,
"grad_norm": 1.4383156299591064,
"learning_rate": 9.941653380841591e-05,
"loss": 4.262,
"step": 16050
},
{
"epoch": 0.2596732310769181,
"grad_norm": 1.3447210788726807,
"learning_rate": 9.941265581440471e-05,
"loss": 3.6917,
"step": 16100
},
{
"epoch": 0.26047966968274705,
"grad_norm": 1.707841157913208,
"learning_rate": 9.940876505168633e-05,
"loss": 3.7174,
"step": 16150
},
{
"epoch": 0.261286108288576,
"grad_norm": 2.2020676136016846,
"learning_rate": 9.940486152126614e-05,
"loss": 4.2535,
"step": 16200
},
{
"epoch": 0.2620925468944049,
"grad_norm": 1.9968034029006958,
"learning_rate": 9.94009452241529e-05,
"loss": 4.2131,
"step": 16250
},
{
"epoch": 0.2628989855002339,
"grad_norm": 4.067761421203613,
"learning_rate": 9.939701616135858e-05,
"loss": 4.0394,
"step": 16300
},
{
"epoch": 0.2637054241060628,
"grad_norm": 1.6382523775100708,
"learning_rate": 9.939307433389852e-05,
"loss": 3.7792,
"step": 16350
},
{
"epoch": 0.26451186271189175,
"grad_norm": 1.519000768661499,
"learning_rate": 9.938911974279132e-05,
"loss": 3.7692,
"step": 16400
},
{
"epoch": 0.26531830131772066,
"grad_norm": 2.7159464359283447,
"learning_rate": 9.938515238905886e-05,
"loss": 3.8706,
"step": 16450
},
{
"epoch": 0.2661247399235496,
"grad_norm": 4.7422871589660645,
"learning_rate": 9.938117227372636e-05,
"loss": 4.1009,
"step": 16500
},
{
"epoch": 0.26693117852937853,
"grad_norm": 2.702423095703125,
"learning_rate": 9.937717939782233e-05,
"loss": 3.7776,
"step": 16550
},
{
"epoch": 0.2677376171352075,
"grad_norm": 3.021857738494873,
"learning_rate": 9.937317376237856e-05,
"loss": 4.1462,
"step": 16600
},
{
"epoch": 0.26854405574103646,
"grad_norm": 1.813398003578186,
"learning_rate": 9.936915536843012e-05,
"loss": 4.0327,
"step": 16650
},
{
"epoch": 0.26935049434686537,
"grad_norm": 2.897822856903076,
"learning_rate": 9.936512421701545e-05,
"loss": 3.9975,
"step": 16700
},
{
"epoch": 0.27015693295269433,
"grad_norm": 5.469192028045654,
"learning_rate": 9.936108030917619e-05,
"loss": 3.9002,
"step": 16750
},
{
"epoch": 0.27096337155852324,
"grad_norm": 1.7145867347717285,
"learning_rate": 9.935702364595736e-05,
"loss": 3.711,
"step": 16800
},
{
"epoch": 0.2717698101643522,
"grad_norm": 2.104463577270508,
"learning_rate": 9.935295422840722e-05,
"loss": 4.1141,
"step": 16850
},
{
"epoch": 0.2725762487701811,
"grad_norm": 3.0537919998168945,
"learning_rate": 9.934887205757735e-05,
"loss": 3.7435,
"step": 16900
},
{
"epoch": 0.27338268737601007,
"grad_norm": 1.5938163995742798,
"learning_rate": 9.934477713452264e-05,
"loss": 4.2125,
"step": 16950
},
{
"epoch": 0.274189125981839,
"grad_norm": 1.9021580219268799,
"learning_rate": 9.934066946030124e-05,
"loss": 3.6318,
"step": 17000
},
{
"epoch": 0.27499556458766794,
"grad_norm": 4.621829986572266,
"learning_rate": 9.933654903597461e-05,
"loss": 3.6992,
"step": 17050
},
{
"epoch": 0.2758020031934969,
"grad_norm": 2.142982244491577,
"learning_rate": 9.933241586260754e-05,
"loss": 3.7732,
"step": 17100
},
{
"epoch": 0.2766084417993258,
"grad_norm": 3.7406392097473145,
"learning_rate": 9.932826994126802e-05,
"loss": 4.1151,
"step": 17150
},
{
"epoch": 0.2774148804051548,
"grad_norm": 0.5368770360946655,
"learning_rate": 9.932411127302745e-05,
"loss": 3.7562,
"step": 17200
},
{
"epoch": 0.2782213190109837,
"grad_norm": 0.6764758825302124,
"learning_rate": 9.931993985896045e-05,
"loss": 3.8355,
"step": 17250
},
{
"epoch": 0.27902775761681264,
"grad_norm": 1.2389382123947144,
"learning_rate": 9.931575570014495e-05,
"loss": 3.9448,
"step": 17300
},
{
"epoch": 0.27983419622264155,
"grad_norm": 2.505661964416504,
"learning_rate": 9.931155879766221e-05,
"loss": 3.8676,
"step": 17350
},
{
"epoch": 0.2806406348284705,
"grad_norm": 1.7408605813980103,
"learning_rate": 9.930734915259669e-05,
"loss": 3.8457,
"step": 17400
},
{
"epoch": 0.2814470734342994,
"grad_norm": 1.7804434299468994,
"learning_rate": 9.930312676603626e-05,
"loss": 3.8563,
"step": 17450
},
{
"epoch": 0.2822535120401284,
"grad_norm": 2.6170337200164795,
"learning_rate": 9.9298891639072e-05,
"loss": 4.1211,
"step": 17500
},
{
"epoch": 0.28305995064595735,
"grad_norm": 2.365858316421509,
"learning_rate": 9.92946437727983e-05,
"loss": 4.1625,
"step": 17550
},
{
"epoch": 0.28386638925178626,
"grad_norm": 2.3039627075195312,
"learning_rate": 9.929038316831285e-05,
"loss": 3.9853,
"step": 17600
},
{
"epoch": 0.2846728278576152,
"grad_norm": 2.711247444152832,
"learning_rate": 9.928610982671665e-05,
"loss": 4.1158,
"step": 17650
},
{
"epoch": 0.2854792664634441,
"grad_norm": 2.1963279247283936,
"learning_rate": 9.928182374911396e-05,
"loss": 4.009,
"step": 17700
},
{
"epoch": 0.2862857050692731,
"grad_norm": 2.6228134632110596,
"learning_rate": 9.927752493661236e-05,
"loss": 4.0115,
"step": 17750
},
{
"epoch": 0.287092143675102,
"grad_norm": 2.0704829692840576,
"learning_rate": 9.927321339032267e-05,
"loss": 4.5117,
"step": 17800
},
{
"epoch": 0.28789858228093096,
"grad_norm": 2.092968225479126,
"learning_rate": 9.926888911135908e-05,
"loss": 4.3974,
"step": 17850
},
{
"epoch": 0.28870502088675987,
"grad_norm": 2.2499780654907227,
"learning_rate": 9.9264552100839e-05,
"loss": 3.8743,
"step": 17900
},
{
"epoch": 0.28951145949258883,
"grad_norm": 1.970129132270813,
"learning_rate": 9.926020235988313e-05,
"loss": 4.0165,
"step": 17950
},
{
"epoch": 0.2903178980984178,
"grad_norm": 1.648297905921936,
"learning_rate": 9.925583988961555e-05,
"loss": 4.1248,
"step": 18000
},
{
"epoch": 0.2903178980984178,
"eval_asr_loss": 0.5475505368937239,
"eval_loss": 5.026027679443359,
"eval_runtime": 17.7593,
"eval_samples_per_second": 28.83,
"eval_steps_per_second": 9.629,
"eval_tts_loss": 5.953106430704743,
"step": 18000
},
{
"epoch": 0.2911243367042467,
"grad_norm": 2.2440879344940186,
"learning_rate": 9.925155231986144e-05,
"loss": 4.0456,
"step": 18050
},
{
"epoch": 0.29193077531007566,
"grad_norm": 4.98739767074585,
"learning_rate": 9.924716464888553e-05,
"loss": 3.9588,
"step": 18100
},
{
"epoch": 0.29273721391590457,
"grad_norm": 1.8671597242355347,
"learning_rate": 9.924276425196694e-05,
"loss": 4.111,
"step": 18150
},
{
"epoch": 0.29354365252173353,
"grad_norm": 2.135396718978882,
"learning_rate": 9.923835113024276e-05,
"loss": 3.8483,
"step": 18200
},
{
"epoch": 0.29435009112756244,
"grad_norm": 2.5975193977355957,
"learning_rate": 9.923392528485342e-05,
"loss": 4.1396,
"step": 18250
},
{
"epoch": 0.2951565297333914,
"grad_norm": 2.3712480068206787,
"learning_rate": 9.922948671694258e-05,
"loss": 4.3892,
"step": 18300
},
{
"epoch": 0.2959629683392203,
"grad_norm": 1.9007618427276611,
"learning_rate": 9.922503542765723e-05,
"loss": 4.0316,
"step": 18350
},
{
"epoch": 0.2967694069450493,
"grad_norm": 2.405129909515381,
"learning_rate": 9.922057141814758e-05,
"loss": 4.2691,
"step": 18400
},
{
"epoch": 0.29757584555087824,
"grad_norm": 1.4283597469329834,
"learning_rate": 9.92160946895672e-05,
"loss": 3.8773,
"step": 18450
},
{
"epoch": 0.29838228415670714,
"grad_norm": 1.7647777795791626,
"learning_rate": 9.921160524307294e-05,
"loss": 4.2706,
"step": 18500
},
{
"epoch": 0.2991887227625361,
"grad_norm": 2.2492105960845947,
"learning_rate": 9.920710307982488e-05,
"loss": 4.2848,
"step": 18550
},
{
"epoch": 0.299995161368365,
"grad_norm": 1.8520779609680176,
"learning_rate": 9.920258820098644e-05,
"loss": 3.8957,
"step": 18600
},
{
"epoch": 0.300801599974194,
"grad_norm": 1.574752688407898,
"learning_rate": 9.91980606077243e-05,
"loss": 3.96,
"step": 18650
},
{
"epoch": 0.3016080385800229,
"grad_norm": 3.233193874359131,
"learning_rate": 9.919352030120846e-05,
"loss": 3.9196,
"step": 18700
},
{
"epoch": 0.30241447718585185,
"grad_norm": 1.642980694770813,
"learning_rate": 9.918896728261216e-05,
"loss": 3.9637,
"step": 18750
},
{
"epoch": 0.30322091579168076,
"grad_norm": 3.266005516052246,
"learning_rate": 9.918440155311193e-05,
"loss": 3.905,
"step": 18800
},
{
"epoch": 0.3040273543975097,
"grad_norm": 2.4287779331207275,
"learning_rate": 9.917982311388763e-05,
"loss": 3.7187,
"step": 18850
},
{
"epoch": 0.3048337930033387,
"grad_norm": 2.907322645187378,
"learning_rate": 9.917523196612233e-05,
"loss": 4.291,
"step": 18900
},
{
"epoch": 0.3056402316091676,
"grad_norm": 2.1453018188476562,
"learning_rate": 9.917062811100247e-05,
"loss": 3.9087,
"step": 18950
},
{
"epoch": 0.30644667021499655,
"grad_norm": 2.715480327606201,
"learning_rate": 9.916601154971772e-05,
"loss": 4.1218,
"step": 19000
},
{
"epoch": 0.30725310882082546,
"grad_norm": 2.193371295928955,
"learning_rate": 9.916138228346101e-05,
"loss": 4.3158,
"step": 19050
},
{
"epoch": 0.3080595474266544,
"grad_norm": 1.1654584407806396,
"learning_rate": 9.915674031342864e-05,
"loss": 3.9391,
"step": 19100
},
{
"epoch": 0.30886598603248333,
"grad_norm": 2.589933156967163,
"learning_rate": 9.915208564082011e-05,
"loss": 3.7376,
"step": 19150
},
{
"epoch": 0.3096724246383123,
"grad_norm": 1.8773146867752075,
"learning_rate": 9.914741826683822e-05,
"loss": 3.8819,
"step": 19200
},
{
"epoch": 0.3104788632441412,
"grad_norm": 3.009824752807617,
"learning_rate": 9.91427381926891e-05,
"loss": 3.7509,
"step": 19250
},
{
"epoch": 0.31128530184997016,
"grad_norm": 1.649009108543396,
"learning_rate": 9.913804541958209e-05,
"loss": 4.3104,
"step": 19300
},
{
"epoch": 0.3120917404557991,
"grad_norm": 2.454561710357666,
"learning_rate": 9.913333994872989e-05,
"loss": 3.6798,
"step": 19350
},
{
"epoch": 0.31289817906162803,
"grad_norm": 2.0216288566589355,
"learning_rate": 9.912862178134839e-05,
"loss": 4.0289,
"step": 19400
},
{
"epoch": 0.313704617667457,
"grad_norm": 2.3107807636260986,
"learning_rate": 9.912389091865683e-05,
"loss": 3.766,
"step": 19450
},
{
"epoch": 0.3145110562732859,
"grad_norm": 2.5804734230041504,
"learning_rate": 9.911914736187774e-05,
"loss": 4.397,
"step": 19500
},
{
"epoch": 0.31531749487911487,
"grad_norm": 2.3121676445007324,
"learning_rate": 9.911439111223686e-05,
"loss": 4.0335,
"step": 19550
},
{
"epoch": 0.3161239334849438,
"grad_norm": 5.4831342697143555,
"learning_rate": 9.910962217096327e-05,
"loss": 3.9674,
"step": 19600
},
{
"epoch": 0.31693037209077274,
"grad_norm": 1.5503877401351929,
"learning_rate": 9.91048405392893e-05,
"loss": 4.2763,
"step": 19650
},
{
"epoch": 0.31773681069660165,
"grad_norm": 3.0469183921813965,
"learning_rate": 9.91000462184506e-05,
"loss": 4.4284,
"step": 19700
},
{
"epoch": 0.3185432493024306,
"grad_norm": 2.0985143184661865,
"learning_rate": 9.909523920968602e-05,
"loss": 4.1063,
"step": 19750
},
{
"epoch": 0.31934968790825957,
"grad_norm": 2.2062013149261475,
"learning_rate": 9.909041951423776e-05,
"loss": 4.33,
"step": 19800
},
{
"epoch": 0.3201561265140885,
"grad_norm": 1.0260072946548462,
"learning_rate": 9.90855871333513e-05,
"loss": 3.8472,
"step": 19850
},
{
"epoch": 0.32096256511991744,
"grad_norm": 1.4739959239959717,
"learning_rate": 9.908074206827534e-05,
"loss": 3.6729,
"step": 19900
},
{
"epoch": 0.32176900372574635,
"grad_norm": 1.3404515981674194,
"learning_rate": 9.90758843202619e-05,
"loss": 3.7977,
"step": 19950
},
{
"epoch": 0.3225754423315753,
"grad_norm": 2.4651851654052734,
"learning_rate": 9.907101389056628e-05,
"loss": 3.7939,
"step": 20000
},
{
"epoch": 0.3233818809374042,
"grad_norm": 4.615785121917725,
"learning_rate": 9.906613078044704e-05,
"loss": 4.1735,
"step": 20050
},
{
"epoch": 0.3241883195432332,
"grad_norm": 2.872647523880005,
"learning_rate": 9.906133303119925e-05,
"loss": 4.0749,
"step": 20100
},
{
"epoch": 0.3249947581490621,
"grad_norm": 1.8358525037765503,
"learning_rate": 9.905642481756708e-05,
"loss": 4.1521,
"step": 20150
},
{
"epoch": 0.32580119675489105,
"grad_norm": 1.2282460927963257,
"learning_rate": 9.905150392728124e-05,
"loss": 4.2458,
"step": 20200
},
{
"epoch": 0.32660763536071996,
"grad_norm": 5.087796211242676,
"learning_rate": 9.904657036161335e-05,
"loss": 4.0652,
"step": 20250
},
{
"epoch": 0.3274140739665489,
"grad_norm": 1.9742780923843384,
"learning_rate": 9.904162412183829e-05,
"loss": 3.8384,
"step": 20300
},
{
"epoch": 0.3282205125723779,
"grad_norm": 2.7985477447509766,
"learning_rate": 9.90366652092342e-05,
"loss": 3.8642,
"step": 20350
},
{
"epoch": 0.3290269511782068,
"grad_norm": 1.4354287385940552,
"learning_rate": 9.903169362508254e-05,
"loss": 3.9342,
"step": 20400
},
{
"epoch": 0.32983338978403576,
"grad_norm": 2.1292335987091064,
"learning_rate": 9.902670937066798e-05,
"loss": 3.6908,
"step": 20450
},
{
"epoch": 0.33063982838986467,
"grad_norm": 2.5627033710479736,
"learning_rate": 9.902171244727851e-05,
"loss": 3.8676,
"step": 20500
},
{
"epoch": 0.33144626699569363,
"grad_norm": 1.1219414472579956,
"learning_rate": 9.90167028562054e-05,
"loss": 4.0734,
"step": 20550
},
{
"epoch": 0.33225270560152254,
"grad_norm": 1.570227026939392,
"learning_rate": 9.901168059874315e-05,
"loss": 3.7047,
"step": 20600
},
{
"epoch": 0.3330591442073515,
"grad_norm": 2.0881588459014893,
"learning_rate": 9.900664567618958e-05,
"loss": 4.1496,
"step": 20650
},
{
"epoch": 0.3338655828131804,
"grad_norm": 2.787027359008789,
"learning_rate": 9.900159808984575e-05,
"loss": 4.0835,
"step": 20700
},
{
"epoch": 0.33467202141900937,
"grad_norm": 4.925529956817627,
"learning_rate": 9.899653784101603e-05,
"loss": 3.961,
"step": 20750
},
{
"epoch": 0.33547846002483833,
"grad_norm": 2.3205068111419678,
"learning_rate": 9.899146493100799e-05,
"loss": 3.7549,
"step": 20800
},
{
"epoch": 0.33628489863066724,
"grad_norm": 1.6399608850479126,
"learning_rate": 9.898637936113257e-05,
"loss": 4.1994,
"step": 20850
},
{
"epoch": 0.3370913372364962,
"grad_norm": 2.315507411956787,
"learning_rate": 9.89812811327039e-05,
"loss": 4.0396,
"step": 20900
},
{
"epoch": 0.3378977758423251,
"grad_norm": 2.239157199859619,
"learning_rate": 9.897617024703943e-05,
"loss": 4.3633,
"step": 20950
},
{
"epoch": 0.3387042144481541,
"grad_norm": 2.6427533626556396,
"learning_rate": 9.897104670545985e-05,
"loss": 3.872,
"step": 21000
},
{
"epoch": 0.3387042144481541,
"eval_asr_loss": 0.5612188019541791,
"eval_loss": 4.999529838562012,
"eval_runtime": 15.2573,
"eval_samples_per_second": 33.558,
"eval_steps_per_second": 11.208,
"eval_tts_loss": 6.0108543639534995,
"step": 21000
},
{
"epoch": 0.339510653053983,
"grad_norm": 1.9391437768936157,
"learning_rate": 9.896591050928914e-05,
"loss": 3.9338,
"step": 21050
},
{
"epoch": 0.34031709165981194,
"grad_norm": 2.6968955993652344,
"learning_rate": 9.896076165985453e-05,
"loss": 4.0754,
"step": 21100
},
{
"epoch": 0.34112353026564085,
"grad_norm": 1.984168529510498,
"learning_rate": 9.895560015848652e-05,
"loss": 3.9749,
"step": 21150
},
{
"epoch": 0.3419299688714698,
"grad_norm": 3.1954638957977295,
"learning_rate": 9.895042600651894e-05,
"loss": 4.1341,
"step": 21200
},
{
"epoch": 0.3427364074772988,
"grad_norm": 3.572061061859131,
"learning_rate": 9.894523920528879e-05,
"loss": 3.8629,
"step": 21250
},
{
"epoch": 0.3435428460831277,
"grad_norm": 1.6200737953186035,
"learning_rate": 9.894003975613644e-05,
"loss": 3.9207,
"step": 21300
},
{
"epoch": 0.34434928468895665,
"grad_norm": 1.8774925470352173,
"learning_rate": 9.893482766040543e-05,
"loss": 3.6651,
"step": 21350
},
{
"epoch": 0.34515572329478555,
"grad_norm": 3.098095655441284,
"learning_rate": 9.892960291944263e-05,
"loss": 4.1371,
"step": 21400
},
{
"epoch": 0.3459621619006145,
"grad_norm": 1.517335295677185,
"learning_rate": 9.892436553459819e-05,
"loss": 4.2014,
"step": 21450
},
{
"epoch": 0.3467686005064434,
"grad_norm": 2.045222520828247,
"learning_rate": 9.891911550722545e-05,
"loss": 4.12,
"step": 21500
},
{
"epoch": 0.3475750391122724,
"grad_norm": 2.005338430404663,
"learning_rate": 9.891385283868112e-05,
"loss": 4.0222,
"step": 21550
},
{
"epoch": 0.3483814777181013,
"grad_norm": 2.2225544452667236,
"learning_rate": 9.89085775303251e-05,
"loss": 4.4284,
"step": 21600
},
{
"epoch": 0.34918791632393026,
"grad_norm": 1.5066674947738647,
"learning_rate": 9.890328958352057e-05,
"loss": 4.042,
"step": 21650
},
{
"epoch": 0.3499943549297592,
"grad_norm": 3.724332332611084,
"learning_rate": 9.8897988999634e-05,
"loss": 4.2711,
"step": 21700
},
{
"epoch": 0.35080079353558813,
"grad_norm": 1.9594568014144897,
"learning_rate": 9.889267578003514e-05,
"loss": 3.8673,
"step": 21750
},
{
"epoch": 0.3516072321414171,
"grad_norm": 1.407718300819397,
"learning_rate": 9.888734992609692e-05,
"loss": 3.965,
"step": 21800
},
{
"epoch": 0.352413670747246,
"grad_norm": 3.1482791900634766,
"learning_rate": 9.888201143919563e-05,
"loss": 4.3013,
"step": 21850
},
{
"epoch": 0.35322010935307496,
"grad_norm": 1.2597465515136719,
"learning_rate": 9.887666032071077e-05,
"loss": 3.9662,
"step": 21900
},
{
"epoch": 0.35402654795890387,
"grad_norm": 2.1143901348114014,
"learning_rate": 9.887129657202513e-05,
"loss": 3.8487,
"step": 21950
},
{
"epoch": 0.35483298656473283,
"grad_norm": 2.351919651031494,
"learning_rate": 9.886592019452475e-05,
"loss": 4.0369,
"step": 22000
},
{
"epoch": 0.35563942517056174,
"grad_norm": 2.052345037460327,
"learning_rate": 9.886053118959895e-05,
"loss": 4.1324,
"step": 22050
},
{
"epoch": 0.3564458637763907,
"grad_norm": 3.6101315021514893,
"learning_rate": 9.885512955864027e-05,
"loss": 3.8461,
"step": 22100
},
{
"epoch": 0.35725230238221967,
"grad_norm": 2.581754684448242,
"learning_rate": 9.884971530304458e-05,
"loss": 3.85,
"step": 22150
},
{
"epoch": 0.3580587409880486,
"grad_norm": 1.2572892904281616,
"learning_rate": 9.884428842421096e-05,
"loss": 3.8305,
"step": 22200
},
{
"epoch": 0.35886517959387754,
"grad_norm": 3.1739416122436523,
"learning_rate": 9.883884892354177e-05,
"loss": 3.953,
"step": 22250
},
{
"epoch": 0.35967161819970644,
"grad_norm": 2.128600597381592,
"learning_rate": 9.883339680244261e-05,
"loss": 3.6528,
"step": 22300
},
{
"epoch": 0.3604780568055354,
"grad_norm": 1.8549989461898804,
"learning_rate": 9.882793206232239e-05,
"loss": 3.9435,
"step": 22350
},
{
"epoch": 0.3612844954113643,
"grad_norm": 1.9381146430969238,
"learning_rate": 9.882245470459324e-05,
"loss": 4.1962,
"step": 22400
},
{
"epoch": 0.3620909340171933,
"grad_norm": 3.481395959854126,
"learning_rate": 9.881696473067056e-05,
"loss": 4.4088,
"step": 22450
},
{
"epoch": 0.3628973726230222,
"grad_norm": 2.8906712532043457,
"learning_rate": 9.881146214197302e-05,
"loss": 4.0288,
"step": 22500
},
{
"epoch": 0.36370381122885115,
"grad_norm": 1.037143588066101,
"learning_rate": 9.880605736756518e-05,
"loss": 3.7669,
"step": 22550
},
{
"epoch": 0.3645102498346801,
"grad_norm": 1.2270286083221436,
"learning_rate": 9.880052980581152e-05,
"loss": 3.9334,
"step": 22600
},
{
"epoch": 0.365316688440509,
"grad_norm": 0.97640460729599,
"learning_rate": 9.87949896335299e-05,
"loss": 3.548,
"step": 22650
},
{
"epoch": 0.366123127046338,
"grad_norm": 2.765495538711548,
"learning_rate": 9.8789436852152e-05,
"loss": 4.5431,
"step": 22700
},
{
"epoch": 0.3669295656521669,
"grad_norm": 1.3644137382507324,
"learning_rate": 9.878387146311272e-05,
"loss": 3.9941,
"step": 22750
},
{
"epoch": 0.36773600425799585,
"grad_norm": 3.112457275390625,
"learning_rate": 9.877829346785017e-05,
"loss": 3.8966,
"step": 22800
},
{
"epoch": 0.36854244286382476,
"grad_norm": 2.050264835357666,
"learning_rate": 9.87727028678058e-05,
"loss": 3.7119,
"step": 22850
},
{
"epoch": 0.3693488814696537,
"grad_norm": 3.1221909523010254,
"learning_rate": 9.876709966442425e-05,
"loss": 4.1182,
"step": 22900
},
{
"epoch": 0.37015532007548263,
"grad_norm": 1.9487212896347046,
"learning_rate": 9.876148385915345e-05,
"loss": 3.8674,
"step": 22950
},
{
"epoch": 0.3709617586813116,
"grad_norm": 2.4063527584075928,
"learning_rate": 9.875585545344458e-05,
"loss": 3.9302,
"step": 23000
},
{
"epoch": 0.37176819728714056,
"grad_norm": 1.6610281467437744,
"learning_rate": 9.875021444875207e-05,
"loss": 4.1362,
"step": 23050
},
{
"epoch": 0.37257463589296946,
"grad_norm": 1.0215773582458496,
"learning_rate": 9.87445608465336e-05,
"loss": 3.9701,
"step": 23100
},
{
"epoch": 0.3733810744987984,
"grad_norm": 2.547252893447876,
"learning_rate": 9.873889464825014e-05,
"loss": 4.3084,
"step": 23150
},
{
"epoch": 0.37418751310462733,
"grad_norm": 3.370795249938965,
"learning_rate": 9.873321585536587e-05,
"loss": 3.9985,
"step": 23200
},
{
"epoch": 0.3749939517104563,
"grad_norm": 1.989953637123108,
"learning_rate": 9.872752446934826e-05,
"loss": 4.0459,
"step": 23250
},
{
"epoch": 0.3758003903162852,
"grad_norm": 3.332749366760254,
"learning_rate": 9.8721820491668e-05,
"loss": 4.1809,
"step": 23300
},
{
"epoch": 0.37660682892211417,
"grad_norm": 2.8768374919891357,
"learning_rate": 9.871610392379906e-05,
"loss": 4.0144,
"step": 23350
},
{
"epoch": 0.3774132675279431,
"grad_norm": 1.8673336505889893,
"learning_rate": 9.871037476721868e-05,
"loss": 3.769,
"step": 23400
},
{
"epoch": 0.37821970613377204,
"grad_norm": 2.023162603378296,
"learning_rate": 9.870463302340729e-05,
"loss": 4.3442,
"step": 23450
},
{
"epoch": 0.379026144739601,
"grad_norm": 1.667543649673462,
"learning_rate": 9.869887869384864e-05,
"loss": 3.688,
"step": 23500
},
{
"epoch": 0.3798325833454299,
"grad_norm": 4.761668682098389,
"learning_rate": 9.869311178002969e-05,
"loss": 3.6925,
"step": 23550
},
{
"epoch": 0.38063902195125887,
"grad_norm": 2.4622697830200195,
"learning_rate": 9.868733228344068e-05,
"loss": 4.3784,
"step": 23600
},
{
"epoch": 0.3814454605570878,
"grad_norm": 2.6539041996002197,
"learning_rate": 9.868154020557507e-05,
"loss": 4.117,
"step": 23650
},
{
"epoch": 0.38225189916291674,
"grad_norm": 1.5162415504455566,
"learning_rate": 9.867573554792961e-05,
"loss": 4.1188,
"step": 23700
},
{
"epoch": 0.38305833776874565,
"grad_norm": 1.5139150619506836,
"learning_rate": 9.866991831200427e-05,
"loss": 4.0601,
"step": 23750
},
{
"epoch": 0.3838647763745746,
"grad_norm": 1.5848387479782104,
"learning_rate": 9.866408849930226e-05,
"loss": 4.3303,
"step": 23800
},
{
"epoch": 0.3846712149804035,
"grad_norm": 2.323932647705078,
"learning_rate": 9.86582461113301e-05,
"loss": 4.3198,
"step": 23850
},
{
"epoch": 0.3854776535862325,
"grad_norm": 1.6189751625061035,
"learning_rate": 9.865239114959748e-05,
"loss": 3.7001,
"step": 23900
},
{
"epoch": 0.38628409219206145,
"grad_norm": 1.6173287630081177,
"learning_rate": 9.86465236156174e-05,
"loss": 3.9669,
"step": 23950
},
{
"epoch": 0.38709053079789035,
"grad_norm": 1.945816993713379,
"learning_rate": 9.864064351090608e-05,
"loss": 4.1111,
"step": 24000
},
{
"epoch": 0.38709053079789035,
"eval_asr_loss": 0.5596013197241803,
"eval_loss": 4.975149631500244,
"eval_runtime": 15.3079,
"eval_samples_per_second": 33.447,
"eval_steps_per_second": 11.171,
"eval_tts_loss": 5.989607999181007,
"step": 24000
},
{
"epoch": 0.3878969694037193,
"grad_norm": 2.5832393169403076,
"learning_rate": 9.863475083698302e-05,
"loss": 4.0228,
"step": 24050
},
{
"epoch": 0.3887034080095482,
"grad_norm": 1.5465142726898193,
"learning_rate": 9.862884559537088e-05,
"loss": 3.9781,
"step": 24100
},
{
"epoch": 0.3895098466153772,
"grad_norm": 2.0018787384033203,
"learning_rate": 9.862292778759572e-05,
"loss": 3.932,
"step": 24150
},
{
"epoch": 0.3903162852212061,
"grad_norm": 1.9746758937835693,
"learning_rate": 9.861699741518669e-05,
"loss": 3.8956,
"step": 24200
},
{
"epoch": 0.39112272382703506,
"grad_norm": 2.558194875717163,
"learning_rate": 9.861105447967629e-05,
"loss": 3.9698,
"step": 24250
},
{
"epoch": 0.39192916243286396,
"grad_norm": 4.342727184295654,
"learning_rate": 9.86050989826002e-05,
"loss": 4.1687,
"step": 24300
},
{
"epoch": 0.3927356010386929,
"grad_norm": 1.6204520463943481,
"learning_rate": 9.859913092549742e-05,
"loss": 3.832,
"step": 24350
},
{
"epoch": 0.39354203964452183,
"grad_norm": 2.4823992252349854,
"learning_rate": 9.859315030991012e-05,
"loss": 4.3509,
"step": 24400
},
{
"epoch": 0.3943484782503508,
"grad_norm": 3.508967876434326,
"learning_rate": 9.858715713738376e-05,
"loss": 4.1016,
"step": 24450
},
{
"epoch": 0.39515491685617976,
"grad_norm": 1.73115873336792,
"learning_rate": 9.858115140946705e-05,
"loss": 3.9989,
"step": 24500
},
{
"epoch": 0.39596135546200867,
"grad_norm": 1.9791126251220703,
"learning_rate": 9.85751331277119e-05,
"loss": 4.2463,
"step": 24550
},
{
"epoch": 0.39676779406783763,
"grad_norm": 1.3955203294754028,
"learning_rate": 9.856910229367349e-05,
"loss": 3.9978,
"step": 24600
},
{
"epoch": 0.39757423267366654,
"grad_norm": 2.9622445106506348,
"learning_rate": 9.856305890891027e-05,
"loss": 3.7169,
"step": 24650
},
{
"epoch": 0.3983806712794955,
"grad_norm": 1.4486901760101318,
"learning_rate": 9.855700297498392e-05,
"loss": 3.9363,
"step": 24700
},
{
"epoch": 0.3991871098853244,
"grad_norm": 1.3110404014587402,
"learning_rate": 9.85509344934593e-05,
"loss": 3.9389,
"step": 24750
},
{
"epoch": 0.3999935484911534,
"grad_norm": 5.3704752922058105,
"learning_rate": 9.854485346590461e-05,
"loss": 4.0937,
"step": 24800
},
{
"epoch": 0.4007999870969823,
"grad_norm": 1.8054653406143188,
"learning_rate": 9.853875989389123e-05,
"loss": 4.0569,
"step": 24850
},
{
"epoch": 0.40160642570281124,
"grad_norm": 2.4662413597106934,
"learning_rate": 9.85326537789938e-05,
"loss": 4.0274,
"step": 24900
},
{
"epoch": 0.4024128643086402,
"grad_norm": 2.1582422256469727,
"learning_rate": 9.852665761880885e-05,
"loss": 3.9833,
"step": 24950
},
{
"epoch": 0.4032193029144691,
"grad_norm": 1.7549717426300049,
"learning_rate": 9.852052667365919e-05,
"loss": 4.0434,
"step": 25000
},
{
"epoch": 0.4040257415202981,
"grad_norm": 2.722822427749634,
"learning_rate": 9.85143831903371e-05,
"loss": 4.2247,
"step": 25050
},
{
"epoch": 0.404832180126127,
"grad_norm": 2.266624689102173,
"learning_rate": 9.850822717043014e-05,
"loss": 4.0886,
"step": 25100
},
{
"epoch": 0.40563861873195595,
"grad_norm": 2.5876810550689697,
"learning_rate": 9.85020586155291e-05,
"loss": 3.9883,
"step": 25150
},
{
"epoch": 0.40644505733778485,
"grad_norm": 2.048994779586792,
"learning_rate": 9.849587752722795e-05,
"loss": 3.7832,
"step": 25200
},
{
"epoch": 0.4072514959436138,
"grad_norm": 2.542686939239502,
"learning_rate": 9.848968390712398e-05,
"loss": 4.1333,
"step": 25250
},
{
"epoch": 0.4080579345494427,
"grad_norm": 1.8834608793258667,
"learning_rate": 9.848347775681765e-05,
"loss": 4.3341,
"step": 25300
},
{
"epoch": 0.4088643731552717,
"grad_norm": 2.2536680698394775,
"learning_rate": 9.847725907791271e-05,
"loss": 3.9732,
"step": 25350
},
{
"epoch": 0.40967081176110065,
"grad_norm": 2.5030934810638428,
"learning_rate": 9.847102787201614e-05,
"loss": 4.0837,
"step": 25400
},
{
"epoch": 0.41047725036692956,
"grad_norm": 2.3016812801361084,
"learning_rate": 9.846478414073812e-05,
"loss": 3.8695,
"step": 25450
},
{
"epoch": 0.4112836889727585,
"grad_norm": 1.6216686964035034,
"learning_rate": 9.84585278856921e-05,
"loss": 3.8525,
"step": 25500
},
{
"epoch": 0.41209012757858743,
"grad_norm": 4.122720718383789,
"learning_rate": 9.845225910849476e-05,
"loss": 4.1526,
"step": 25550
},
{
"epoch": 0.4128965661844164,
"grad_norm": 2.8652255535125732,
"learning_rate": 9.844597781076599e-05,
"loss": 3.8846,
"step": 25600
},
{
"epoch": 0.4137030047902453,
"grad_norm": 1.516522765159607,
"learning_rate": 9.843968399412898e-05,
"loss": 3.9619,
"step": 25650
},
{
"epoch": 0.41450944339607426,
"grad_norm": 1.89898681640625,
"learning_rate": 9.843337766021005e-05,
"loss": 4.0633,
"step": 25700
},
{
"epoch": 0.41531588200190317,
"grad_norm": 3.1514346599578857,
"learning_rate": 9.842705881063889e-05,
"loss": 3.8735,
"step": 25750
},
{
"epoch": 0.41612232060773213,
"grad_norm": 1.562589168548584,
"learning_rate": 9.84207274470483e-05,
"loss": 3.6907,
"step": 25800
},
{
"epoch": 0.4169287592135611,
"grad_norm": 1.0612993240356445,
"learning_rate": 9.841438357107438e-05,
"loss": 4.2971,
"step": 25850
},
{
"epoch": 0.41773519781939,
"grad_norm": 2.6953563690185547,
"learning_rate": 9.840802718435646e-05,
"loss": 3.9486,
"step": 25900
},
{
"epoch": 0.41854163642521897,
"grad_norm": 2.3360955715179443,
"learning_rate": 9.840165828853709e-05,
"loss": 4.0716,
"step": 25950
},
{
"epoch": 0.4193480750310479,
"grad_norm": 2.1841683387756348,
"learning_rate": 9.839527688526204e-05,
"loss": 3.9567,
"step": 26000
},
{
"epoch": 0.42015451363687684,
"grad_norm": 2.321568489074707,
"learning_rate": 9.838888297618032e-05,
"loss": 4.0802,
"step": 26050
},
{
"epoch": 0.42096095224270574,
"grad_norm": 2.530367374420166,
"learning_rate": 9.838247656294421e-05,
"loss": 3.7345,
"step": 26100
},
{
"epoch": 0.4217673908485347,
"grad_norm": 1.4943206310272217,
"learning_rate": 9.837605764720916e-05,
"loss": 3.865,
"step": 26150
},
{
"epoch": 0.4225738294543636,
"grad_norm": 1.2419519424438477,
"learning_rate": 9.836962623063391e-05,
"loss": 3.9362,
"step": 26200
},
{
"epoch": 0.4233802680601926,
"grad_norm": 2.8908309936523438,
"learning_rate": 9.836318231488035e-05,
"loss": 3.5773,
"step": 26250
},
{
"epoch": 0.42418670666602154,
"grad_norm": 5.162326335906982,
"learning_rate": 9.835672590161369e-05,
"loss": 3.9459,
"step": 26300
},
{
"epoch": 0.42499314527185045,
"grad_norm": 1.782263994216919,
"learning_rate": 9.835025699250233e-05,
"loss": 3.9619,
"step": 26350
},
{
"epoch": 0.4257995838776794,
"grad_norm": 5.839146137237549,
"learning_rate": 9.834377558921788e-05,
"loss": 3.8363,
"step": 26400
},
{
"epoch": 0.4266060224835083,
"grad_norm": 1.6730350255966187,
"learning_rate": 9.833728169343521e-05,
"loss": 3.7422,
"step": 26450
},
{
"epoch": 0.4274124610893373,
"grad_norm": 0.9565515518188477,
"learning_rate": 9.83307753068324e-05,
"loss": 4.3006,
"step": 26500
},
{
"epoch": 0.4282188996951662,
"grad_norm": 1.102502465248108,
"learning_rate": 9.832425643109076e-05,
"loss": 3.7061,
"step": 26550
},
{
"epoch": 0.42902533830099515,
"grad_norm": 2.7195091247558594,
"learning_rate": 9.831772506789483e-05,
"loss": 3.8085,
"step": 26600
},
{
"epoch": 0.42983177690682406,
"grad_norm": 2.3110263347625732,
"learning_rate": 9.831118121893239e-05,
"loss": 3.7295,
"step": 26650
},
{
"epoch": 0.430638215512653,
"grad_norm": 1.7428209781646729,
"learning_rate": 9.830462488589443e-05,
"loss": 4.0876,
"step": 26700
},
{
"epoch": 0.431444654118482,
"grad_norm": 1.8392008543014526,
"learning_rate": 9.829805607047515e-05,
"loss": 4.3742,
"step": 26750
},
{
"epoch": 0.4322510927243109,
"grad_norm": 2.320955276489258,
"learning_rate": 9.829147477437202e-05,
"loss": 4.0089,
"step": 26800
},
{
"epoch": 0.43305753133013986,
"grad_norm": 2.5901613235473633,
"learning_rate": 9.82848809992857e-05,
"loss": 4.0296,
"step": 26850
},
{
"epoch": 0.43386396993596876,
"grad_norm": 4.836414337158203,
"learning_rate": 9.827827474692009e-05,
"loss": 4.3018,
"step": 26900
},
{
"epoch": 0.4346704085417977,
"grad_norm": 3.3830177783966064,
"learning_rate": 9.82716560189823e-05,
"loss": 3.949,
"step": 26950
},
{
"epoch": 0.43547684714762663,
"grad_norm": 1.8795298337936401,
"learning_rate": 9.826502481718267e-05,
"loss": 3.7089,
"step": 27000
},
{
"epoch": 0.43547684714762663,
"eval_asr_loss": 0.5559742133953548,
"eval_loss": 4.958992958068848,
"eval_runtime": 14.7232,
"eval_samples_per_second": 34.775,
"eval_steps_per_second": 11.614,
"eval_tts_loss": 6.008725999135946,
"step": 27000
},
{
"epoch": 0.4362832857534556,
"grad_norm": 1.627994418144226,
"learning_rate": 9.825838114323478e-05,
"loss": 3.9582,
"step": 27050
},
{
"epoch": 0.4370897243592845,
"grad_norm": 4.4471611976623535,
"learning_rate": 9.825172499885542e-05,
"loss": 3.8203,
"step": 27100
},
{
"epoch": 0.43789616296511347,
"grad_norm": 4.530590534210205,
"learning_rate": 9.82450563857646e-05,
"loss": 4.1167,
"step": 27150
},
{
"epoch": 0.43870260157094243,
"grad_norm": 2.129225969314575,
"learning_rate": 9.823837530568554e-05,
"loss": 3.9683,
"step": 27200
},
{
"epoch": 0.43950904017677134,
"grad_norm": 1.8322468996047974,
"learning_rate": 9.823168176034471e-05,
"loss": 4.0839,
"step": 27250
},
{
"epoch": 0.4403154787826003,
"grad_norm": 2.424598217010498,
"learning_rate": 9.822497575147178e-05,
"loss": 4.1781,
"step": 27300
},
{
"epoch": 0.4411219173884292,
"grad_norm": 1.3116930723190308,
"learning_rate": 9.821839177232751e-05,
"loss": 3.9697,
"step": 27350
},
{
"epoch": 0.44192835599425817,
"grad_norm": 1.402601718902588,
"learning_rate": 9.821166109077653e-05,
"loss": 4.0815,
"step": 27400
},
{
"epoch": 0.4427347946000871,
"grad_norm": 1.392431616783142,
"learning_rate": 9.820491795086698e-05,
"loss": 4.0143,
"step": 27450
},
{
"epoch": 0.44354123320591604,
"grad_norm": 1.231276035308838,
"learning_rate": 9.819816235434137e-05,
"loss": 4.3486,
"step": 27500
},
{
"epoch": 0.44434767181174495,
"grad_norm": 3.108103036880493,
"learning_rate": 9.819139430294539e-05,
"loss": 4.0492,
"step": 27550
},
{
"epoch": 0.4451541104175739,
"grad_norm": 2.1228959560394287,
"learning_rate": 9.818461379842798e-05,
"loss": 4.3671,
"step": 27600
},
{
"epoch": 0.4459605490234029,
"grad_norm": 1.8643229007720947,
"learning_rate": 9.81778208425413e-05,
"loss": 3.9776,
"step": 27650
},
{
"epoch": 0.4467669876292318,
"grad_norm": 1.3176897764205933,
"learning_rate": 9.81710154370407e-05,
"loss": 4.368,
"step": 27700
},
{
"epoch": 0.44757342623506075,
"grad_norm": 2.3601205348968506,
"learning_rate": 9.816419758368476e-05,
"loss": 3.6981,
"step": 27750
},
{
"epoch": 0.44837986484088965,
"grad_norm": 4.233773708343506,
"learning_rate": 9.815736728423531e-05,
"loss": 4.4567,
"step": 27800
},
{
"epoch": 0.4491863034467186,
"grad_norm": 2.077209949493408,
"learning_rate": 9.81505245404573e-05,
"loss": 3.854,
"step": 27850
},
{
"epoch": 0.4499927420525475,
"grad_norm": 1.6529903411865234,
"learning_rate": 9.814366935411903e-05,
"loss": 4.254,
"step": 27900
},
{
"epoch": 0.4507991806583765,
"grad_norm": 2.4949686527252197,
"learning_rate": 9.813680172699191e-05,
"loss": 4.1666,
"step": 27950
},
{
"epoch": 0.4516056192642054,
"grad_norm": 1.0352684259414673,
"learning_rate": 9.81299216608506e-05,
"loss": 3.9291,
"step": 28000
},
{
"epoch": 0.45241205787003436,
"grad_norm": 1.945921778678894,
"learning_rate": 9.8123029157473e-05,
"loss": 4.0353,
"step": 28050
},
{
"epoch": 0.4532184964758633,
"grad_norm": 1.8263922929763794,
"learning_rate": 9.811612421864018e-05,
"loss": 3.984,
"step": 28100
},
{
"epoch": 0.4540249350816922,
"grad_norm": 2.8539750576019287,
"learning_rate": 9.810920684613644e-05,
"loss": 4.1103,
"step": 28150
},
{
"epoch": 0.4548313736875212,
"grad_norm": 2.559903860092163,
"learning_rate": 9.81022770417493e-05,
"loss": 4.0392,
"step": 28200
},
{
"epoch": 0.4556378122933501,
"grad_norm": 1.7837358713150024,
"learning_rate": 9.809533480726952e-05,
"loss": 3.9311,
"step": 28250
},
{
"epoch": 0.45644425089917906,
"grad_norm": 1.3091505765914917,
"learning_rate": 9.808838014449097e-05,
"loss": 3.7175,
"step": 28300
},
{
"epoch": 0.45725068950500797,
"grad_norm": 1.9200706481933594,
"learning_rate": 9.808141305521087e-05,
"loss": 4.3186,
"step": 28350
},
{
"epoch": 0.45805712811083693,
"grad_norm": 2.3566806316375732,
"learning_rate": 9.807443354122955e-05,
"loss": 4.1294,
"step": 28400
},
{
"epoch": 0.45886356671666584,
"grad_norm": 1.5939757823944092,
"learning_rate": 9.806744160435059e-05,
"loss": 3.8491,
"step": 28450
},
{
"epoch": 0.4596700053224948,
"grad_norm": 1.4164952039718628,
"learning_rate": 9.806043724638076e-05,
"loss": 4.2949,
"step": 28500
},
{
"epoch": 0.46047644392832376,
"grad_norm": 2.1956284046173096,
"learning_rate": 9.805342046913009e-05,
"loss": 4.1996,
"step": 28550
},
{
"epoch": 0.46128288253415267,
"grad_norm": 1.7804524898529053,
"learning_rate": 9.804639127441176e-05,
"loss": 3.9101,
"step": 28600
},
{
"epoch": 0.46208932113998163,
"grad_norm": 0.5685811638832092,
"learning_rate": 9.803934966404217e-05,
"loss": 4.1905,
"step": 28650
},
{
"epoch": 0.46289575974581054,
"grad_norm": 1.6610820293426514,
"learning_rate": 9.803229563984095e-05,
"loss": 4.3581,
"step": 28700
},
{
"epoch": 0.4637021983516395,
"grad_norm": 2.0791728496551514,
"learning_rate": 9.802522920363095e-05,
"loss": 3.9016,
"step": 28750
},
{
"epoch": 0.4645086369574684,
"grad_norm": 2.6067445278167725,
"learning_rate": 9.801815035723817e-05,
"loss": 3.9681,
"step": 28800
},
{
"epoch": 0.4653150755632974,
"grad_norm": 0.8365187644958496,
"learning_rate": 9.801105910249187e-05,
"loss": 3.9012,
"step": 28850
},
{
"epoch": 0.4661215141691263,
"grad_norm": 3.1129026412963867,
"learning_rate": 9.80039554412245e-05,
"loss": 3.9982,
"step": 28900
},
{
"epoch": 0.46692795277495525,
"grad_norm": 1.3477463722229004,
"learning_rate": 9.799683937527171e-05,
"loss": 4.1739,
"step": 28950
},
{
"epoch": 0.46773439138078415,
"grad_norm": 3.441631317138672,
"learning_rate": 9.798971090647238e-05,
"loss": 4.0417,
"step": 29000
},
{
"epoch": 0.4685408299866131,
"grad_norm": 2.399874210357666,
"learning_rate": 9.798257003666853e-05,
"loss": 4.1551,
"step": 29050
},
{
"epoch": 0.4693472685924421,
"grad_norm": 2.2936065196990967,
"learning_rate": 9.797541676770547e-05,
"loss": 4.1262,
"step": 29100
},
{
"epoch": 0.470153707198271,
"grad_norm": 2.890214204788208,
"learning_rate": 9.796825110143165e-05,
"loss": 4.1734,
"step": 29150
},
{
"epoch": 0.47096014580409995,
"grad_norm": 1.6101442575454712,
"learning_rate": 9.796107303969877e-05,
"loss": 3.6554,
"step": 29200
},
{
"epoch": 0.47176658440992886,
"grad_norm": 1.5904902219772339,
"learning_rate": 9.795388258436169e-05,
"loss": 4.1421,
"step": 29250
},
{
"epoch": 0.4725730230157578,
"grad_norm": 1.2808867692947388,
"learning_rate": 9.79466797372785e-05,
"loss": 3.7509,
"step": 29300
},
{
"epoch": 0.47337946162158673,
"grad_norm": 2.6117844581604004,
"learning_rate": 9.793946450031049e-05,
"loss": 4.252,
"step": 29350
},
{
"epoch": 0.4741859002274157,
"grad_norm": 2.114128351211548,
"learning_rate": 9.793238154921244e-05,
"loss": 4.0362,
"step": 29400
},
{
"epoch": 0.4749923388332446,
"grad_norm": 0.834858775138855,
"learning_rate": 9.792514178577616e-05,
"loss": 3.8481,
"step": 29450
},
{
"epoch": 0.47579877743907356,
"grad_norm": 1.2032753229141235,
"learning_rate": 9.791788963802069e-05,
"loss": 4.0954,
"step": 29500
},
{
"epoch": 0.4766052160449025,
"grad_norm": 1.481992244720459,
"learning_rate": 9.791062510782003e-05,
"loss": 4.0293,
"step": 29550
},
{
"epoch": 0.47741165465073143,
"grad_norm": 1.40059494972229,
"learning_rate": 9.790334819705142e-05,
"loss": 3.8551,
"step": 29600
},
{
"epoch": 0.4782180932565604,
"grad_norm": 1.1095318794250488,
"learning_rate": 9.789605890759526e-05,
"loss": 4.0393,
"step": 29650
},
{
"epoch": 0.4790245318623893,
"grad_norm": 2.459015369415283,
"learning_rate": 9.78887572413352e-05,
"loss": 4.1729,
"step": 29700
},
{
"epoch": 0.47983097046821827,
"grad_norm": 1.1588975191116333,
"learning_rate": 9.788144320015804e-05,
"loss": 3.971,
"step": 29750
},
{
"epoch": 0.4806374090740472,
"grad_norm": 2.8166420459747314,
"learning_rate": 9.78741167859538e-05,
"loss": 4.4002,
"step": 29800
},
{
"epoch": 0.48144384767987614,
"grad_norm": 3.607013702392578,
"learning_rate": 9.786677800061571e-05,
"loss": 3.819,
"step": 29850
},
{
"epoch": 0.48225028628570504,
"grad_norm": 2.313955783843994,
"learning_rate": 9.785942684604018e-05,
"loss": 4.2892,
"step": 29900
},
{
"epoch": 0.483056724891534,
"grad_norm": 1.3837124109268188,
"learning_rate": 9.785206332412682e-05,
"loss": 4.0314,
"step": 29950
},
{
"epoch": 0.48386316349736297,
"grad_norm": 1.5818651914596558,
"learning_rate": 9.784468743677842e-05,
"loss": 3.8618,
"step": 30000
},
{
"epoch": 0.48386316349736297,
"eval_asr_loss": 0.5646996813599123,
"eval_loss": 4.947766304016113,
"eval_runtime": 14.9459,
"eval_samples_per_second": 34.257,
"eval_steps_per_second": 11.441,
"eval_tts_loss": 5.988222905527309,
"step": 30000
},
{
"epoch": 0.4846696021031919,
"grad_norm": 2.607830047607422,
"learning_rate": 9.7837299185901e-05,
"loss": 4.127,
"step": 30050
},
{
"epoch": 0.48547604070902084,
"grad_norm": 1.471893310546875,
"learning_rate": 9.782989857340373e-05,
"loss": 3.7895,
"step": 30100
},
{
"epoch": 0.48628247931484975,
"grad_norm": 2.5169291496276855,
"learning_rate": 9.782248560119902e-05,
"loss": 4.0763,
"step": 30150
},
{
"epoch": 0.4870889179206787,
"grad_norm": 4.164167881011963,
"learning_rate": 9.781506027120247e-05,
"loss": 4.0099,
"step": 30200
},
{
"epoch": 0.4878953565265076,
"grad_norm": 2.2432568073272705,
"learning_rate": 9.780762258533282e-05,
"loss": 4.1052,
"step": 30250
},
{
"epoch": 0.4887017951323366,
"grad_norm": 1.442578911781311,
"learning_rate": 9.780017254551206e-05,
"loss": 3.9327,
"step": 30300
},
{
"epoch": 0.4895082337381655,
"grad_norm": 1.6483718156814575,
"learning_rate": 9.779271015366535e-05,
"loss": 3.7249,
"step": 30350
},
{
"epoch": 0.49031467234399445,
"grad_norm": 2.119659662246704,
"learning_rate": 9.778523541172104e-05,
"loss": 3.959,
"step": 30400
},
{
"epoch": 0.4911211109498234,
"grad_norm": 1.574180006980896,
"learning_rate": 9.777774832161069e-05,
"loss": 4.2155,
"step": 30450
},
{
"epoch": 0.4919275495556523,
"grad_norm": 1.1062897443771362,
"learning_rate": 9.7770248885269e-05,
"loss": 4.263,
"step": 30500
},
{
"epoch": 0.4927339881614813,
"grad_norm": 1.0894304513931274,
"learning_rate": 9.776273710463394e-05,
"loss": 3.9637,
"step": 30550
},
{
"epoch": 0.4935404267673102,
"grad_norm": 1.5745190382003784,
"learning_rate": 9.77552129816466e-05,
"loss": 3.9539,
"step": 30600
},
{
"epoch": 0.49434686537313915,
"grad_norm": 1.2136081457138062,
"learning_rate": 9.77476765182513e-05,
"loss": 3.8683,
"step": 30650
},
{
"epoch": 0.49515330397896806,
"grad_norm": 2.5908925533294678,
"learning_rate": 9.774012771639553e-05,
"loss": 4.1582,
"step": 30700
},
{
"epoch": 0.495959742584797,
"grad_norm": 1.8017607927322388,
"learning_rate": 9.773256657802997e-05,
"loss": 3.6259,
"step": 30750
},
{
"epoch": 0.49676618119062593,
"grad_norm": 1.5173224210739136,
"learning_rate": 9.77249931051085e-05,
"loss": 4.0439,
"step": 30800
},
{
"epoch": 0.4975726197964549,
"grad_norm": 15.362312316894531,
"learning_rate": 9.771786279566293e-05,
"loss": 4.387,
"step": 30850
},
{
"epoch": 0.49837905840228386,
"grad_norm": 3.3392670154571533,
"learning_rate": 9.771026539928696e-05,
"loss": 4.5741,
"step": 30900
},
{
"epoch": 0.49918549700811277,
"grad_norm": 1.549731969833374,
"learning_rate": 9.770265567411791e-05,
"loss": 3.9523,
"step": 30950
},
{
"epoch": 0.49999193561394173,
"grad_norm": 2.7253308296203613,
"learning_rate": 9.769503362212222e-05,
"loss": 4.3312,
"step": 31000
},
{
"epoch": 0.5007983742197707,
"grad_norm": 2.1496167182922363,
"learning_rate": 9.768739924526948e-05,
"loss": 3.9847,
"step": 31050
},
{
"epoch": 0.5016048128255995,
"grad_norm": 2.93792462348938,
"learning_rate": 9.76797525455325e-05,
"loss": 3.9403,
"step": 31100
},
{
"epoch": 0.5024112514314285,
"grad_norm": 2.272951126098633,
"learning_rate": 9.767209352488727e-05,
"loss": 3.7558,
"step": 31150
},
{
"epoch": 0.5032176900372575,
"grad_norm": 3.371812105178833,
"learning_rate": 9.766442218531293e-05,
"loss": 3.9215,
"step": 31200
},
{
"epoch": 0.5040241286430864,
"grad_norm": 1.3443658351898193,
"learning_rate": 9.765673852879186e-05,
"loss": 4.1457,
"step": 31250
},
{
"epoch": 0.5048305672489153,
"grad_norm": 2.311265230178833,
"learning_rate": 9.764904255730955e-05,
"loss": 4.3678,
"step": 31300
},
{
"epoch": 0.5056370058547442,
"grad_norm": 2.241488456726074,
"learning_rate": 9.764133427285475e-05,
"loss": 4.145,
"step": 31350
},
{
"epoch": 0.5064434444605732,
"grad_norm": 2.38242769241333,
"learning_rate": 9.763361367741934e-05,
"loss": 4.1718,
"step": 31400
},
{
"epoch": 0.5072498830664022,
"grad_norm": 1.874671220779419,
"learning_rate": 9.76258807729984e-05,
"loss": 3.9391,
"step": 31450
},
{
"epoch": 0.5080563216722311,
"grad_norm": 1.9810266494750977,
"learning_rate": 9.761813556159019e-05,
"loss": 3.9962,
"step": 31500
},
{
"epoch": 0.50886276027806,
"grad_norm": 2.93658185005188,
"learning_rate": 9.761037804519612e-05,
"loss": 3.7464,
"step": 31550
},
{
"epoch": 0.509669198883889,
"grad_norm": 2.2134599685668945,
"learning_rate": 9.760260822582086e-05,
"loss": 3.9865,
"step": 31600
},
{
"epoch": 0.5104756374897179,
"grad_norm": 2.7154464721679688,
"learning_rate": 9.759482610547214e-05,
"loss": 3.8921,
"step": 31650
},
{
"epoch": 0.5112820760955469,
"grad_norm": 2.0518405437469482,
"learning_rate": 9.7587031686161e-05,
"loss": 4.1342,
"step": 31700
},
{
"epoch": 0.5120885147013757,
"grad_norm": 2.0980758666992188,
"learning_rate": 9.757922496990155e-05,
"loss": 3.9435,
"step": 31750
},
{
"epoch": 0.5128949533072047,
"grad_norm": 4.004748344421387,
"learning_rate": 9.757140595871115e-05,
"loss": 4.1433,
"step": 31800
},
{
"epoch": 0.5137013919130337,
"grad_norm": 1.889359474182129,
"learning_rate": 9.756357465461027e-05,
"loss": 4.1632,
"step": 31850
},
{
"epoch": 0.5145078305188626,
"grad_norm": 2.1634538173675537,
"learning_rate": 9.755573105962264e-05,
"loss": 3.9894,
"step": 31900
},
{
"epoch": 0.5153142691246916,
"grad_norm": 2.0058374404907227,
"learning_rate": 9.754787517577508e-05,
"loss": 3.7636,
"step": 31950
},
{
"epoch": 0.5161207077305204,
"grad_norm": 3.5357580184936523,
"learning_rate": 9.754000700509764e-05,
"loss": 3.9957,
"step": 32000
},
{
"epoch": 0.5169271463363494,
"grad_norm": 1.03873872756958,
"learning_rate": 9.753212654962356e-05,
"loss": 4.1323,
"step": 32050
},
{
"epoch": 0.5177335849421784,
"grad_norm": 2.3036513328552246,
"learning_rate": 9.752423381138919e-05,
"loss": 4.057,
"step": 32100
},
{
"epoch": 0.5185400235480073,
"grad_norm": 1.3666496276855469,
"learning_rate": 9.75163287924341e-05,
"loss": 4.1997,
"step": 32150
},
{
"epoch": 0.5193464621538362,
"grad_norm": 2.106926441192627,
"learning_rate": 9.750841149480102e-05,
"loss": 3.924,
"step": 32200
},
{
"epoch": 0.5201529007596651,
"grad_norm": 3.9756150245666504,
"learning_rate": 9.750048192053587e-05,
"loss": 3.9417,
"step": 32250
},
{
"epoch": 0.5209593393654941,
"grad_norm": 1.9011123180389404,
"learning_rate": 9.749254007168771e-05,
"loss": 3.9619,
"step": 32300
},
{
"epoch": 0.5217657779713231,
"grad_norm": 1.8549953699111938,
"learning_rate": 9.74845859503088e-05,
"loss": 3.8087,
"step": 32350
},
{
"epoch": 0.522572216577152,
"grad_norm": 2.133540153503418,
"learning_rate": 9.747661955845456e-05,
"loss": 4.1356,
"step": 32400
},
{
"epoch": 0.5233786551829809,
"grad_norm": 1.6960670948028564,
"learning_rate": 9.74686408981836e-05,
"loss": 3.798,
"step": 32450
},
{
"epoch": 0.5241850937888098,
"grad_norm": 1.8586905002593994,
"learning_rate": 9.746064997155763e-05,
"loss": 4.0743,
"step": 32500
},
{
"epoch": 0.5249915323946388,
"grad_norm": 2.0501816272735596,
"learning_rate": 9.745264678064166e-05,
"loss": 3.6707,
"step": 32550
},
{
"epoch": 0.5257979710004678,
"grad_norm": 2.2415525913238525,
"learning_rate": 9.744463132750373e-05,
"loss": 3.671,
"step": 32600
},
{
"epoch": 0.5266044096062966,
"grad_norm": 2.3392422199249268,
"learning_rate": 9.743660361421512e-05,
"loss": 4.1384,
"step": 32650
},
{
"epoch": 0.5274108482121256,
"grad_norm": 1.0827341079711914,
"learning_rate": 9.74285636428503e-05,
"loss": 4.0892,
"step": 32700
},
{
"epoch": 0.5282172868179545,
"grad_norm": 1.8944828510284424,
"learning_rate": 9.742051141548685e-05,
"loss": 3.9703,
"step": 32750
},
{
"epoch": 0.5290237254237835,
"grad_norm": 2.013651132583618,
"learning_rate": 9.741244693420554e-05,
"loss": 4.2219,
"step": 32800
},
{
"epoch": 0.5298301640296125,
"grad_norm": 1.879599690437317,
"learning_rate": 9.740437020109033e-05,
"loss": 3.7386,
"step": 32850
},
{
"epoch": 0.5306366026354413,
"grad_norm": 2.901381254196167,
"learning_rate": 9.739628121822832e-05,
"loss": 3.9859,
"step": 32900
},
{
"epoch": 0.5314430412412703,
"grad_norm": 2.858835220336914,
"learning_rate": 9.738817998770976e-05,
"loss": 3.8376,
"step": 32950
},
{
"epoch": 0.5322494798470992,
"grad_norm": 2.78635311126709,
"learning_rate": 9.738006651162812e-05,
"loss": 3.8029,
"step": 33000
},
{
"epoch": 0.5322494798470992,
"eval_asr_loss": 0.5563808219787441,
"eval_loss": 4.897815227508545,
"eval_runtime": 14.787,
"eval_samples_per_second": 34.625,
"eval_steps_per_second": 11.564,
"eval_tts_loss": 5.9989542354158925,
"step": 33000
},
{
"epoch": 0.5330559184529282,
"grad_norm": 1.9606646299362183,
"learning_rate": 9.737194079207997e-05,
"loss": 4.1998,
"step": 33050
},
{
"epoch": 0.5338623570587571,
"grad_norm": 3.6199183464050293,
"learning_rate": 9.736380283116508e-05,
"loss": 3.9197,
"step": 33100
},
{
"epoch": 0.534668795664586,
"grad_norm": 2.4361817836761475,
"learning_rate": 9.735565263098641e-05,
"loss": 3.8554,
"step": 33150
},
{
"epoch": 0.535475234270415,
"grad_norm": 2.393467426300049,
"learning_rate": 9.734749019365001e-05,
"loss": 3.9229,
"step": 33200
},
{
"epoch": 0.536281672876244,
"grad_norm": 1.991856336593628,
"learning_rate": 9.733931552126515e-05,
"loss": 3.9931,
"step": 33250
},
{
"epoch": 0.5370881114820729,
"grad_norm": 1.7219436168670654,
"learning_rate": 9.733112861594422e-05,
"loss": 3.9433,
"step": 33300
},
{
"epoch": 0.5378945500879018,
"grad_norm": 2.8829445838928223,
"learning_rate": 9.732292947980285e-05,
"loss": 4.3263,
"step": 33350
},
{
"epoch": 0.5387009886937307,
"grad_norm": 1.1644093990325928,
"learning_rate": 9.731471811495973e-05,
"loss": 4.1349,
"step": 33400
},
{
"epoch": 0.5395074272995597,
"grad_norm": 2.788134813308716,
"learning_rate": 9.730649452353678e-05,
"loss": 4.1791,
"step": 33450
},
{
"epoch": 0.5403138659053887,
"grad_norm": 1.2594338655471802,
"learning_rate": 9.729825870765903e-05,
"loss": 3.8136,
"step": 33500
},
{
"epoch": 0.5411203045112175,
"grad_norm": 2.067070484161377,
"learning_rate": 9.729001066945471e-05,
"loss": 4.1771,
"step": 33550
},
{
"epoch": 0.5419267431170465,
"grad_norm": 0.9377022385597229,
"learning_rate": 9.728175041105519e-05,
"loss": 3.9944,
"step": 33600
},
{
"epoch": 0.5427331817228754,
"grad_norm": 1.7732789516448975,
"learning_rate": 9.7273477934595e-05,
"loss": 4.1222,
"step": 33650
},
{
"epoch": 0.5435396203287044,
"grad_norm": 2.3974225521087646,
"learning_rate": 9.726519324221184e-05,
"loss": 4.2036,
"step": 33700
},
{
"epoch": 0.5443460589345334,
"grad_norm": 3.553283929824829,
"learning_rate": 9.725689633604654e-05,
"loss": 4.4454,
"step": 33750
},
{
"epoch": 0.5451524975403622,
"grad_norm": 1.552428960800171,
"learning_rate": 9.724858721824311e-05,
"loss": 3.8556,
"step": 33800
},
{
"epoch": 0.5459589361461912,
"grad_norm": 1.2861645221710205,
"learning_rate": 9.72402658909487e-05,
"loss": 3.6202,
"step": 33850
},
{
"epoch": 0.5467653747520201,
"grad_norm": 2.8274953365325928,
"learning_rate": 9.723193235631362e-05,
"loss": 3.7635,
"step": 33900
},
{
"epoch": 0.5475718133578491,
"grad_norm": 1.0712840557098389,
"learning_rate": 9.722358661649133e-05,
"loss": 3.7407,
"step": 33950
},
{
"epoch": 0.548378251963678,
"grad_norm": 2.3699841499328613,
"learning_rate": 9.721522867363845e-05,
"loss": 4.0412,
"step": 34000
},
{
"epoch": 0.5491846905695069,
"grad_norm": 2.715761423110962,
"learning_rate": 9.72068585299148e-05,
"loss": 3.915,
"step": 34050
},
{
"epoch": 0.5499911291753359,
"grad_norm": 4.615658760070801,
"learning_rate": 9.719847618748323e-05,
"loss": 4.0316,
"step": 34100
},
{
"epoch": 0.5507975677811648,
"grad_norm": 2.428999662399292,
"learning_rate": 9.719008164850986e-05,
"loss": 4.2053,
"step": 34150
},
{
"epoch": 0.5516040063869938,
"grad_norm": 0.604597806930542,
"learning_rate": 9.718167491516395e-05,
"loss": 4.3448,
"step": 34200
},
{
"epoch": 0.5524104449928227,
"grad_norm": 2.038712978363037,
"learning_rate": 9.717325598961781e-05,
"loss": 4.0422,
"step": 34250
},
{
"epoch": 0.5532168835986516,
"grad_norm": 1.6703771352767944,
"learning_rate": 9.716482487404702e-05,
"loss": 3.8453,
"step": 34300
},
{
"epoch": 0.5540233222044806,
"grad_norm": 1.6739346981048584,
"learning_rate": 9.715638157063025e-05,
"loss": 3.9266,
"step": 34350
},
{
"epoch": 0.5548297608103095,
"grad_norm": 2.9111719131469727,
"learning_rate": 9.714792608154932e-05,
"loss": 4.0806,
"step": 34400
},
{
"epoch": 0.5556361994161384,
"grad_norm": 2.14871883392334,
"learning_rate": 9.713945840898924e-05,
"loss": 3.9186,
"step": 34450
},
{
"epoch": 0.5564426380219674,
"grad_norm": 2.2379770278930664,
"learning_rate": 9.71309785551381e-05,
"loss": 3.7397,
"step": 34500
},
{
"epoch": 0.5572490766277963,
"grad_norm": 0.7250761985778809,
"learning_rate": 9.71224865221872e-05,
"loss": 4.056,
"step": 34550
},
{
"epoch": 0.5580555152336253,
"grad_norm": 1.8340208530426025,
"learning_rate": 9.711398231233095e-05,
"loss": 4.2596,
"step": 34600
},
{
"epoch": 0.5588619538394543,
"grad_norm": 1.2574554681777954,
"learning_rate": 9.710546592776695e-05,
"loss": 3.9969,
"step": 34650
},
{
"epoch": 0.5596683924452831,
"grad_norm": 1.2908345460891724,
"learning_rate": 9.709693737069587e-05,
"loss": 3.9818,
"step": 34700
},
{
"epoch": 0.5604748310511121,
"grad_norm": 2.4312095642089844,
"learning_rate": 9.708839664332161e-05,
"loss": 3.7135,
"step": 34750
},
{
"epoch": 0.561281269656941,
"grad_norm": 1.706939458847046,
"learning_rate": 9.707984374785117e-05,
"loss": 3.8907,
"step": 34800
},
{
"epoch": 0.56208770826277,
"grad_norm": 1.253430962562561,
"learning_rate": 9.707127868649468e-05,
"loss": 3.8833,
"step": 34850
},
{
"epoch": 0.5628941468685988,
"grad_norm": 2.2516000270843506,
"learning_rate": 9.706270146146544e-05,
"loss": 4.1565,
"step": 34900
},
{
"epoch": 0.5637005854744278,
"grad_norm": 1.791192889213562,
"learning_rate": 9.705411207497991e-05,
"loss": 3.8712,
"step": 34950
},
{
"epoch": 0.5645070240802568,
"grad_norm": 2.1465959548950195,
"learning_rate": 9.704551052925765e-05,
"loss": 3.8223,
"step": 35000
},
{
"epoch": 0.5653134626860857,
"grad_norm": 1.0886331796646118,
"learning_rate": 9.703689682652138e-05,
"loss": 3.9508,
"step": 35050
},
{
"epoch": 0.5661199012919147,
"grad_norm": 2.1944518089294434,
"learning_rate": 9.702827096899698e-05,
"loss": 4.3178,
"step": 35100
},
{
"epoch": 0.5669263398977435,
"grad_norm": 1.7926956415176392,
"learning_rate": 9.701963295891346e-05,
"loss": 3.7831,
"step": 35150
},
{
"epoch": 0.5677327785035725,
"grad_norm": 1.8286861181259155,
"learning_rate": 9.701098279850292e-05,
"loss": 4.0985,
"step": 35200
},
{
"epoch": 0.5685392171094015,
"grad_norm": 1.7975794076919556,
"learning_rate": 9.70023204900007e-05,
"loss": 4.3829,
"step": 35250
},
{
"epoch": 0.5693456557152304,
"grad_norm": 1.3266535997390747,
"learning_rate": 9.69936460356452e-05,
"loss": 3.7414,
"step": 35300
},
{
"epoch": 0.5701520943210593,
"grad_norm": 2.072350263595581,
"learning_rate": 9.698495943767797e-05,
"loss": 4.2939,
"step": 35350
},
{
"epoch": 0.5709585329268883,
"grad_norm": 1.8565208911895752,
"learning_rate": 9.697626069834373e-05,
"loss": 3.8508,
"step": 35400
},
{
"epoch": 0.5717649715327172,
"grad_norm": 1.3206040859222412,
"learning_rate": 9.696754981989031e-05,
"loss": 3.5709,
"step": 35450
},
{
"epoch": 0.5725714101385462,
"grad_norm": 2.337531566619873,
"learning_rate": 9.695882680456869e-05,
"loss": 4.1167,
"step": 35500
},
{
"epoch": 0.5733778487443751,
"grad_norm": 2.7290146350860596,
"learning_rate": 9.695009165463298e-05,
"loss": 3.9105,
"step": 35550
},
{
"epoch": 0.574184287350204,
"grad_norm": 2.879288911819458,
"learning_rate": 9.694134437234044e-05,
"loss": 4.0808,
"step": 35600
},
{
"epoch": 0.574990725956033,
"grad_norm": 2.2341060638427734,
"learning_rate": 9.693258495995141e-05,
"loss": 4.2625,
"step": 35650
},
{
"epoch": 0.5757971645618619,
"grad_norm": 2.0360066890716553,
"learning_rate": 9.692381341972946e-05,
"loss": 3.9065,
"step": 35700
},
{
"epoch": 0.5766036031676909,
"grad_norm": 1.8910644054412842,
"learning_rate": 9.69150297539412e-05,
"loss": 3.7694,
"step": 35750
},
{
"epoch": 0.5774100417735197,
"grad_norm": 1.2420679330825806,
"learning_rate": 9.690623396485642e-05,
"loss": 4.0219,
"step": 35800
},
{
"epoch": 0.5782164803793487,
"grad_norm": 1.8608529567718506,
"learning_rate": 9.689742605474807e-05,
"loss": 3.9607,
"step": 35850
},
{
"epoch": 0.5790229189851777,
"grad_norm": 1.903253197669983,
"learning_rate": 9.688860602589217e-05,
"loss": 4.225,
"step": 35900
},
{
"epoch": 0.5798293575910066,
"grad_norm": 1.1789734363555908,
"learning_rate": 9.687977388056791e-05,
"loss": 3.7763,
"step": 35950
},
{
"epoch": 0.5806357961968356,
"grad_norm": 2.1416497230529785,
"learning_rate": 9.687092962105758e-05,
"loss": 3.9497,
"step": 36000
},
{
"epoch": 0.5806357961968356,
"eval_asr_loss": 0.5691508247039826,
"eval_loss": 4.8229522705078125,
"eval_runtime": 14.7902,
"eval_samples_per_second": 34.617,
"eval_steps_per_second": 11.562,
"eval_tts_loss": 6.016266314813699,
"step": 36000
},
{
"epoch": 0.5814422348026644,
"grad_norm": 1.5320308208465576,
"learning_rate": 9.686207324964665e-05,
"loss": 3.7274,
"step": 36050
},
{
"epoch": 0.5822486734084934,
"grad_norm": 2.025629758834839,
"learning_rate": 9.685320476862368e-05,
"loss": 4.0585,
"step": 36100
},
{
"epoch": 0.5830551120143224,
"grad_norm": 1.4545434713363647,
"learning_rate": 9.684432418028038e-05,
"loss": 4.0057,
"step": 36150
},
{
"epoch": 0.5838615506201513,
"grad_norm": 1.8786442279815674,
"learning_rate": 9.683543148691155e-05,
"loss": 3.5916,
"step": 36200
},
{
"epoch": 0.5846679892259802,
"grad_norm": 1.035013198852539,
"learning_rate": 9.682652669081519e-05,
"loss": 4.0424,
"step": 36250
},
{
"epoch": 0.5854744278318091,
"grad_norm": 1.4029967784881592,
"learning_rate": 9.681760979429235e-05,
"loss": 4.2236,
"step": 36300
},
{
"epoch": 0.5862808664376381,
"grad_norm": 3.3041937351226807,
"learning_rate": 9.680868079964726e-05,
"loss": 4.0875,
"step": 36350
},
{
"epoch": 0.5870873050434671,
"grad_norm": 2.505573272705078,
"learning_rate": 9.679973970918727e-05,
"loss": 3.6474,
"step": 36400
},
{
"epoch": 0.587893743649296,
"grad_norm": 2.1446595191955566,
"learning_rate": 9.679078652522283e-05,
"loss": 4.2775,
"step": 36450
},
{
"epoch": 0.5887001822551249,
"grad_norm": 2.0899713039398193,
"learning_rate": 9.678182125006751e-05,
"loss": 3.9854,
"step": 36500
},
{
"epoch": 0.5895066208609538,
"grad_norm": 1.1477386951446533,
"learning_rate": 9.677284388603806e-05,
"loss": 3.8436,
"step": 36550
},
{
"epoch": 0.5903130594667828,
"grad_norm": 3.7815449237823486,
"learning_rate": 9.67638544354543e-05,
"loss": 3.7386,
"step": 36600
},
{
"epoch": 0.5911194980726118,
"grad_norm": 1.9048818349838257,
"learning_rate": 9.675485290063917e-05,
"loss": 4.5235,
"step": 36650
},
{
"epoch": 0.5919259366784406,
"grad_norm": 2.060019016265869,
"learning_rate": 9.67458392839188e-05,
"loss": 3.9605,
"step": 36700
},
{
"epoch": 0.5927323752842696,
"grad_norm": 1.520529866218567,
"learning_rate": 9.673681358762235e-05,
"loss": 3.8785,
"step": 36750
},
{
"epoch": 0.5935388138900985,
"grad_norm": 1.8698012828826904,
"learning_rate": 9.672777581408218e-05,
"loss": 4.1342,
"step": 36800
},
{
"epoch": 0.5943452524959275,
"grad_norm": 2.1991450786590576,
"learning_rate": 9.671872596563372e-05,
"loss": 3.8919,
"step": 36850
},
{
"epoch": 0.5951516911017565,
"grad_norm": 1.6680142879486084,
"learning_rate": 9.670966404461554e-05,
"loss": 3.8034,
"step": 36900
},
{
"epoch": 0.5959581297075853,
"grad_norm": 1.9333608150482178,
"learning_rate": 9.670059005336933e-05,
"loss": 4.1449,
"step": 36950
},
{
"epoch": 0.5967645683134143,
"grad_norm": 0.8060359358787537,
"learning_rate": 9.669150399423991e-05,
"loss": 3.7206,
"step": 37000
},
{
"epoch": 0.5975710069192433,
"grad_norm": 2.4317374229431152,
"learning_rate": 9.668240586957518e-05,
"loss": 4.3372,
"step": 37050
},
{
"epoch": 0.5983774455250722,
"grad_norm": 1.0086828470230103,
"learning_rate": 9.66732956817262e-05,
"loss": 4.3951,
"step": 37100
},
{
"epoch": 0.5991838841309011,
"grad_norm": 1.9289790391921997,
"learning_rate": 9.666417343304713e-05,
"loss": 3.6865,
"step": 37150
},
{
"epoch": 0.59999032273673,
"grad_norm": 1.7148078680038452,
"learning_rate": 9.665503912589525e-05,
"loss": 4.1055,
"step": 37200
},
{
"epoch": 0.600796761342559,
"grad_norm": 1.2342246770858765,
"learning_rate": 9.664589276263095e-05,
"loss": 3.9735,
"step": 37250
},
{
"epoch": 0.601603199948388,
"grad_norm": 1.9903031587600708,
"learning_rate": 9.663673434561771e-05,
"loss": 3.8671,
"step": 37300
},
{
"epoch": 0.6024096385542169,
"grad_norm": 1.465268850326538,
"learning_rate": 9.66275638772222e-05,
"loss": 3.9581,
"step": 37350
},
{
"epoch": 0.6032160771600458,
"grad_norm": 1.6464263200759888,
"learning_rate": 9.661838135981412e-05,
"loss": 3.9276,
"step": 37400
},
{
"epoch": 0.6040225157658747,
"grad_norm": 2.538856267929077,
"learning_rate": 9.660918679576636e-05,
"loss": 3.9488,
"step": 37450
},
{
"epoch": 0.6048289543717037,
"grad_norm": 3.353928565979004,
"learning_rate": 9.659998018745484e-05,
"loss": 4.0735,
"step": 37500
},
{
"epoch": 0.6056353929775327,
"grad_norm": 2.7926583290100098,
"learning_rate": 9.659076153725868e-05,
"loss": 4.0669,
"step": 37550
},
{
"epoch": 0.6064418315833615,
"grad_norm": 2.16097354888916,
"learning_rate": 9.658153084756003e-05,
"loss": 3.7335,
"step": 37600
},
{
"epoch": 0.6072482701891905,
"grad_norm": 2.647451639175415,
"learning_rate": 9.657228812074422e-05,
"loss": 3.8995,
"step": 37650
},
{
"epoch": 0.6080547087950194,
"grad_norm": 1.449508786201477,
"learning_rate": 9.656303335919963e-05,
"loss": 3.8566,
"step": 37700
},
{
"epoch": 0.6088611474008484,
"grad_norm": 1.6518259048461914,
"learning_rate": 9.655376656531781e-05,
"loss": 4.1638,
"step": 37750
},
{
"epoch": 0.6096675860066774,
"grad_norm": 2.536465883255005,
"learning_rate": 9.654448774149336e-05,
"loss": 4.4301,
"step": 37800
},
{
"epoch": 0.6104740246125062,
"grad_norm": 1.383885145187378,
"learning_rate": 9.653519689012403e-05,
"loss": 3.7318,
"step": 37850
},
{
"epoch": 0.6112804632183352,
"grad_norm": 2.0474696159362793,
"learning_rate": 9.652589401361067e-05,
"loss": 3.9712,
"step": 37900
},
{
"epoch": 0.6120869018241641,
"grad_norm": 3.0140864849090576,
"learning_rate": 9.651657911435725e-05,
"loss": 3.931,
"step": 37950
},
{
"epoch": 0.6128933404299931,
"grad_norm": 1.272126317024231,
"learning_rate": 9.65072521947708e-05,
"loss": 3.5631,
"step": 38000
},
{
"epoch": 0.613699779035822,
"grad_norm": 1.682157039642334,
"learning_rate": 9.649791325726149e-05,
"loss": 3.4141,
"step": 38050
},
{
"epoch": 0.6145062176416509,
"grad_norm": 1.6309934854507446,
"learning_rate": 9.648856230424259e-05,
"loss": 4.2273,
"step": 38100
},
{
"epoch": 0.6153126562474799,
"grad_norm": 1.193287968635559,
"learning_rate": 9.64791993381305e-05,
"loss": 3.9418,
"step": 38150
},
{
"epoch": 0.6161190948533088,
"grad_norm": 2.2340729236602783,
"learning_rate": 9.646982436134469e-05,
"loss": 4.3585,
"step": 38200
},
{
"epoch": 0.6169255334591378,
"grad_norm": 3.2289340496063232,
"learning_rate": 9.646043737630772e-05,
"loss": 4.0445,
"step": 38250
},
{
"epoch": 0.6177319720649667,
"grad_norm": 2.211266040802002,
"learning_rate": 9.645103838544531e-05,
"loss": 3.5903,
"step": 38300
},
{
"epoch": 0.6185384106707956,
"grad_norm": 2.726902484893799,
"learning_rate": 9.644162739118625e-05,
"loss": 3.942,
"step": 38350
},
{
"epoch": 0.6193448492766246,
"grad_norm": 2.515888214111328,
"learning_rate": 9.64322043959624e-05,
"loss": 4.0163,
"step": 38400
},
{
"epoch": 0.6201512878824536,
"grad_norm": 1.7479592561721802,
"learning_rate": 9.642276940220878e-05,
"loss": 4.0748,
"step": 38450
},
{
"epoch": 0.6209577264882824,
"grad_norm": 1.6890915632247925,
"learning_rate": 9.641332241236348e-05,
"loss": 3.924,
"step": 38500
},
{
"epoch": 0.6217641650941114,
"grad_norm": 2.3948464393615723,
"learning_rate": 9.640386342886767e-05,
"loss": 3.9101,
"step": 38550
},
{
"epoch": 0.6225706036999403,
"grad_norm": 2.6629281044006348,
"learning_rate": 9.639439245416568e-05,
"loss": 3.9898,
"step": 38600
},
{
"epoch": 0.6233770423057693,
"grad_norm": 1.5631389617919922,
"learning_rate": 9.638490949070488e-05,
"loss": 3.83,
"step": 38650
},
{
"epoch": 0.6241834809115983,
"grad_norm": 1.956933856010437,
"learning_rate": 9.637541454093574e-05,
"loss": 4.0719,
"step": 38700
},
{
"epoch": 0.6249899195174271,
"grad_norm": 2.43593430519104,
"learning_rate": 9.63659076073119e-05,
"loss": 4.0034,
"step": 38750
},
{
"epoch": 0.6257963581232561,
"grad_norm": 1.7999205589294434,
"learning_rate": 9.635638869229e-05,
"loss": 3.6974,
"step": 38800
},
{
"epoch": 0.626602796729085,
"grad_norm": 3.0806641578674316,
"learning_rate": 9.634685779832983e-05,
"loss": 4.0155,
"step": 38850
},
{
"epoch": 0.627409235334914,
"grad_norm": 2.0375235080718994,
"learning_rate": 9.633731492789428e-05,
"loss": 3.9403,
"step": 38900
},
{
"epoch": 0.6282156739407428,
"grad_norm": 2.0660057067871094,
"learning_rate": 9.632776008344929e-05,
"loss": 3.9845,
"step": 38950
},
{
"epoch": 0.6290221125465718,
"grad_norm": 3.3655974864959717,
"learning_rate": 9.631819326746394e-05,
"loss": 4.0855,
"step": 39000
},
{
"epoch": 0.6290221125465718,
"eval_asr_loss": 0.6034199091663585,
"eval_loss": 4.653688430786133,
"eval_runtime": 14.8011,
"eval_samples_per_second": 34.592,
"eval_steps_per_second": 11.553,
"eval_tts_loss": 6.013137326975981,
"step": 39000
},
{
"epoch": 0.6298285511524008,
"grad_norm": 2.8556010723114014,
"learning_rate": 9.63086144824104e-05,
"loss": 4.1555,
"step": 39050
},
{
"epoch": 0.6306349897582297,
"grad_norm": 0.7945855259895325,
"learning_rate": 9.62990237307639e-05,
"loss": 3.9282,
"step": 39100
},
{
"epoch": 0.6314414283640587,
"grad_norm": 2.212608575820923,
"learning_rate": 9.628942101500281e-05,
"loss": 3.9476,
"step": 39150
},
{
"epoch": 0.6322478669698876,
"grad_norm": 2.4456074237823486,
"learning_rate": 9.627980633760852e-05,
"loss": 4.2139,
"step": 39200
},
{
"epoch": 0.6330543055757165,
"grad_norm": 2.3700897693634033,
"learning_rate": 9.62701797010656e-05,
"loss": 3.9148,
"step": 39250
},
{
"epoch": 0.6338607441815455,
"grad_norm": 1.0412943363189697,
"learning_rate": 9.626054110786163e-05,
"loss": 3.7921,
"step": 39300
},
{
"epoch": 0.6346671827873744,
"grad_norm": 1.5326954126358032,
"learning_rate": 9.625089056048736e-05,
"loss": 4.3737,
"step": 39350
},
{
"epoch": 0.6354736213932033,
"grad_norm": 1.076934814453125,
"learning_rate": 9.624122806143652e-05,
"loss": 3.9424,
"step": 39400
},
{
"epoch": 0.6362800599990323,
"grad_norm": 1.8640769720077515,
"learning_rate": 9.623155361320607e-05,
"loss": 4.3091,
"step": 39450
},
{
"epoch": 0.6370864986048612,
"grad_norm": 2.431981325149536,
"learning_rate": 9.62218672182959e-05,
"loss": 3.7956,
"step": 39500
},
{
"epoch": 0.6378929372106902,
"grad_norm": 3.0567352771759033,
"learning_rate": 9.621216887920913e-05,
"loss": 3.9352,
"step": 39550
},
{
"epoch": 0.6386993758165191,
"grad_norm": 1.1013330221176147,
"learning_rate": 9.620245859845188e-05,
"loss": 4.0237,
"step": 39600
},
{
"epoch": 0.639505814422348,
"grad_norm": 2.7281951904296875,
"learning_rate": 9.619273637853339e-05,
"loss": 4.3155,
"step": 39650
},
{
"epoch": 0.640312253028177,
"grad_norm": 1.6277055740356445,
"learning_rate": 9.618300222196596e-05,
"loss": 4.0474,
"step": 39700
},
{
"epoch": 0.6411186916340059,
"grad_norm": 1.7232165336608887,
"learning_rate": 9.617325613126502e-05,
"loss": 3.7625,
"step": 39750
},
{
"epoch": 0.6419251302398349,
"grad_norm": 1.3456089496612549,
"learning_rate": 9.6163498108949e-05,
"loss": 3.8336,
"step": 39800
},
{
"epoch": 0.6427315688456637,
"grad_norm": 3.376032590866089,
"learning_rate": 9.615372815753954e-05,
"loss": 3.9334,
"step": 39850
},
{
"epoch": 0.6435380074514927,
"grad_norm": 3.418212890625,
"learning_rate": 9.614394627956121e-05,
"loss": 4.2293,
"step": 39900
},
{
"epoch": 0.6443444460573217,
"grad_norm": 1.5658949613571167,
"learning_rate": 9.613415247754181e-05,
"loss": 4.0883,
"step": 39950
},
{
"epoch": 0.6451508846631506,
"grad_norm": 1.2223689556121826,
"learning_rate": 9.612434675401212e-05,
"loss": 3.9809,
"step": 40000
},
{
"epoch": 0.6459573232689795,
"grad_norm": 1.4119608402252197,
"learning_rate": 9.611452911150603e-05,
"loss": 4.135,
"step": 40050
},
{
"epoch": 0.6467637618748084,
"grad_norm": 1.1455477476119995,
"learning_rate": 9.610469955256052e-05,
"loss": 4.0727,
"step": 40100
},
{
"epoch": 0.6475702004806374,
"grad_norm": 1.1804072856903076,
"learning_rate": 9.609485807971566e-05,
"loss": 3.7652,
"step": 40150
},
{
"epoch": 0.6483766390864664,
"grad_norm": 2.133227586746216,
"learning_rate": 9.608500469551455e-05,
"loss": 3.6933,
"step": 40200
},
{
"epoch": 0.6491830776922953,
"grad_norm": 1.6320295333862305,
"learning_rate": 9.60751394025034e-05,
"loss": 3.7121,
"step": 40250
},
{
"epoch": 0.6499895162981242,
"grad_norm": 2.7744624614715576,
"learning_rate": 9.606526220323155e-05,
"loss": 4.1106,
"step": 40300
},
{
"epoch": 0.6507959549039531,
"grad_norm": 2.890082836151123,
"learning_rate": 9.605537310025131e-05,
"loss": 3.8397,
"step": 40350
},
{
"epoch": 0.6516023935097821,
"grad_norm": 1.4252163171768188,
"learning_rate": 9.604547209611815e-05,
"loss": 3.9493,
"step": 40400
},
{
"epoch": 0.6524088321156111,
"grad_norm": 1.658120036125183,
"learning_rate": 9.603555919339055e-05,
"loss": 4.0099,
"step": 40450
},
{
"epoch": 0.6532152707214399,
"grad_norm": 2.3012828826904297,
"learning_rate": 9.602563439463012e-05,
"loss": 3.8907,
"step": 40500
},
{
"epoch": 0.6540217093272689,
"grad_norm": 2.3004066944122314,
"learning_rate": 9.601569770240154e-05,
"loss": 4.2203,
"step": 40550
},
{
"epoch": 0.6548281479330978,
"grad_norm": 0.95237135887146,
"learning_rate": 9.600574911927253e-05,
"loss": 4.0099,
"step": 40600
},
{
"epoch": 0.6556345865389268,
"grad_norm": 2.167707681655884,
"learning_rate": 9.599578864781389e-05,
"loss": 3.4495,
"step": 40650
},
{
"epoch": 0.6564410251447558,
"grad_norm": 1.9729243516921997,
"learning_rate": 9.598581629059952e-05,
"loss": 3.9483,
"step": 40700
},
{
"epoch": 0.6572474637505846,
"grad_norm": 0.6504325270652771,
"learning_rate": 9.597583205020636e-05,
"loss": 3.859,
"step": 40750
},
{
"epoch": 0.6580539023564136,
"grad_norm": 2.181685447692871,
"learning_rate": 9.596583592921446e-05,
"loss": 3.6357,
"step": 40800
},
{
"epoch": 0.6588603409622426,
"grad_norm": 1.3120620250701904,
"learning_rate": 9.595582793020688e-05,
"loss": 3.7685,
"step": 40850
},
{
"epoch": 0.6596667795680715,
"grad_norm": 3.323073148727417,
"learning_rate": 9.59458080557698e-05,
"loss": 4.3613,
"step": 40900
},
{
"epoch": 0.6604732181739004,
"grad_norm": 1.2107150554656982,
"learning_rate": 9.593577630849246e-05,
"loss": 4.0395,
"step": 40950
},
{
"epoch": 0.6612796567797293,
"grad_norm": 2.1152050495147705,
"learning_rate": 9.592573269096716e-05,
"loss": 3.6636,
"step": 41000
},
{
"epoch": 0.6620860953855583,
"grad_norm": 1.9166638851165771,
"learning_rate": 9.591567720578924e-05,
"loss": 3.8332,
"step": 41050
},
{
"epoch": 0.6628925339913873,
"grad_norm": 1.2429214715957642,
"learning_rate": 9.590560985555716e-05,
"loss": 4.0641,
"step": 41100
},
{
"epoch": 0.6636989725972162,
"grad_norm": 2.569197177886963,
"learning_rate": 9.58955306428724e-05,
"loss": 4.026,
"step": 41150
},
{
"epoch": 0.6645054112030451,
"grad_norm": 1.7593148946762085,
"learning_rate": 9.588543957033957e-05,
"loss": 3.88,
"step": 41200
},
{
"epoch": 0.665311849808874,
"grad_norm": 1.2028878927230835,
"learning_rate": 9.587553881534579e-05,
"loss": 4.0756,
"step": 41250
},
{
"epoch": 0.666118288414703,
"grad_norm": 2.990448474884033,
"learning_rate": 9.586542426800969e-05,
"loss": 4.2531,
"step": 41300
},
{
"epoch": 0.666924727020532,
"grad_norm": 2.4063851833343506,
"learning_rate": 9.585529786860527e-05,
"loss": 3.9218,
"step": 41350
},
{
"epoch": 0.6677311656263608,
"grad_norm": 1.7430907487869263,
"learning_rate": 9.584515961974928e-05,
"loss": 4.3134,
"step": 41400
},
{
"epoch": 0.6685376042321898,
"grad_norm": 1.3421581983566284,
"learning_rate": 9.583500952406155e-05,
"loss": 4.0289,
"step": 41450
},
{
"epoch": 0.6693440428380187,
"grad_norm": 0.871703565120697,
"learning_rate": 9.582484758416497e-05,
"loss": 3.7478,
"step": 41500
},
{
"epoch": 0.6701504814438477,
"grad_norm": 1.735887885093689,
"learning_rate": 9.581467380268547e-05,
"loss": 3.9596,
"step": 41550
},
{
"epoch": 0.6709569200496767,
"grad_norm": 1.485318660736084,
"learning_rate": 9.580448818225205e-05,
"loss": 4.1336,
"step": 41600
},
{
"epoch": 0.6717633586555055,
"grad_norm": 1.8921575546264648,
"learning_rate": 9.579429072549679e-05,
"loss": 3.8069,
"step": 41650
},
{
"epoch": 0.6725697972613345,
"grad_norm": 1.4174717664718628,
"learning_rate": 9.57840814350548e-05,
"loss": 3.7202,
"step": 41700
},
{
"epoch": 0.6733762358671634,
"grad_norm": 2.4129083156585693,
"learning_rate": 9.577386031356425e-05,
"loss": 3.9037,
"step": 41750
},
{
"epoch": 0.6741826744729924,
"grad_norm": 3.722903251647949,
"learning_rate": 9.57636273636664e-05,
"loss": 3.971,
"step": 41800
},
{
"epoch": 0.6749891130788213,
"grad_norm": 1.9005436897277832,
"learning_rate": 9.57535875993941e-05,
"loss": 3.9043,
"step": 41850
},
{
"epoch": 0.6757955516846502,
"grad_norm": 2.620349645614624,
"learning_rate": 9.574333123705391e-05,
"loss": 4.0456,
"step": 41900
},
{
"epoch": 0.6766019902904792,
"grad_norm": 1.5120000839233398,
"learning_rate": 9.573306305419541e-05,
"loss": 3.9113,
"step": 41950
},
{
"epoch": 0.6774084288963081,
"grad_norm": 2.184945821762085,
"learning_rate": 9.572278305347202e-05,
"loss": 3.9566,
"step": 42000
},
{
"epoch": 0.6774084288963081,
"eval_asr_loss": 0.6747458191937141,
"eval_loss": 4.402849197387695,
"eval_runtime": 14.8128,
"eval_samples_per_second": 34.565,
"eval_steps_per_second": 11.544,
"eval_tts_loss": 6.006661847860199,
"step": 42000
},
{
"epoch": 0.6782148675021371,
"grad_norm": 1.9270944595336914,
"learning_rate": 9.571249123754014e-05,
"loss": 4.0085,
"step": 42050
},
{
"epoch": 0.679021306107966,
"grad_norm": 2.115370035171509,
"learning_rate": 9.570218760905934e-05,
"loss": 3.9551,
"step": 42100
},
{
"epoch": 0.6798277447137949,
"grad_norm": 2.4913384914398193,
"learning_rate": 9.569187217069213e-05,
"loss": 3.9498,
"step": 42150
},
{
"epoch": 0.6806341833196239,
"grad_norm": 1.2166416645050049,
"learning_rate": 9.568154492510414e-05,
"loss": 3.9295,
"step": 42200
},
{
"epoch": 0.6814406219254528,
"grad_norm": 3.2959489822387695,
"learning_rate": 9.567120587496402e-05,
"loss": 3.8898,
"step": 42250
},
{
"epoch": 0.6822470605312817,
"grad_norm": 3.9133834838867188,
"learning_rate": 9.566085502294348e-05,
"loss": 4.1996,
"step": 42300
},
{
"epoch": 0.6830534991371107,
"grad_norm": 1.421493411064148,
"learning_rate": 9.56504923717173e-05,
"loss": 3.9401,
"step": 42350
},
{
"epoch": 0.6838599377429396,
"grad_norm": 1.4366188049316406,
"learning_rate": 9.564011792396326e-05,
"loss": 4.1854,
"step": 42400
},
{
"epoch": 0.6846663763487686,
"grad_norm": 2.0820562839508057,
"learning_rate": 9.562973168236224e-05,
"loss": 4.1061,
"step": 42450
},
{
"epoch": 0.6854728149545976,
"grad_norm": 1.6290011405944824,
"learning_rate": 9.561933364959814e-05,
"loss": 4.3134,
"step": 42500
},
{
"epoch": 0.6862792535604264,
"grad_norm": 2.4670650959014893,
"learning_rate": 9.560892382835791e-05,
"loss": 3.7429,
"step": 42550
},
{
"epoch": 0.6870856921662554,
"grad_norm": 1.2104759216308594,
"learning_rate": 9.559850222133155e-05,
"loss": 3.6463,
"step": 42600
},
{
"epoch": 0.6878921307720843,
"grad_norm": 1.4174095392227173,
"learning_rate": 9.558806883121209e-05,
"loss": 4.0535,
"step": 42650
},
{
"epoch": 0.6886985693779133,
"grad_norm": 2.343122959136963,
"learning_rate": 9.557762366069562e-05,
"loss": 4.2928,
"step": 42700
},
{
"epoch": 0.6895050079837421,
"grad_norm": 2.328246831893921,
"learning_rate": 9.55671667124813e-05,
"loss": 4.1672,
"step": 42750
},
{
"epoch": 0.6903114465895711,
"grad_norm": 2.1924808025360107,
"learning_rate": 9.55566979892713e-05,
"loss": 4.0338,
"step": 42800
},
{
"epoch": 0.6911178851954001,
"grad_norm": 1.4481010437011719,
"learning_rate": 9.554621749377079e-05,
"loss": 3.8551,
"step": 42850
},
{
"epoch": 0.691924323801229,
"grad_norm": 2.5818378925323486,
"learning_rate": 9.55357252286881e-05,
"loss": 3.8865,
"step": 42900
},
{
"epoch": 0.692730762407058,
"grad_norm": 0.6974899768829346,
"learning_rate": 9.552522119673449e-05,
"loss": 4.3369,
"step": 42950
},
{
"epoch": 0.6935372010128869,
"grad_norm": 2.877656936645508,
"learning_rate": 9.551470540062433e-05,
"loss": 3.6878,
"step": 43000
},
{
"epoch": 0.6943436396187158,
"grad_norm": 1.5497684478759766,
"learning_rate": 9.550417784307497e-05,
"loss": 3.6174,
"step": 43050
},
{
"epoch": 0.6951500782245448,
"grad_norm": 2.261922836303711,
"learning_rate": 9.549363852680686e-05,
"loss": 3.9077,
"step": 43100
},
{
"epoch": 0.6959565168303737,
"grad_norm": 2.5778346061706543,
"learning_rate": 9.548308745454344e-05,
"loss": 3.8163,
"step": 43150
},
{
"epoch": 0.6967629554362026,
"grad_norm": 2.7817728519439697,
"learning_rate": 9.547252462901124e-05,
"loss": 3.6548,
"step": 43200
},
{
"epoch": 0.6975693940420316,
"grad_norm": 1.0119379758834839,
"learning_rate": 9.546195005293976e-05,
"loss": 4.1415,
"step": 43250
},
{
"epoch": 0.6983758326478605,
"grad_norm": 3.6960830688476562,
"learning_rate": 9.54513637290616e-05,
"loss": 4.0043,
"step": 43300
},
{
"epoch": 0.6991822712536895,
"grad_norm": 1.3119101524353027,
"learning_rate": 9.544076566011235e-05,
"loss": 4.0963,
"step": 43350
},
{
"epoch": 0.6999887098595184,
"grad_norm": 1.1128474473953247,
"learning_rate": 9.543015584883067e-05,
"loss": 3.7204,
"step": 43400
},
{
"epoch": 0.7007951484653473,
"grad_norm": 2.3756420612335205,
"learning_rate": 9.541953429795823e-05,
"loss": 4.3295,
"step": 43450
},
{
"epoch": 0.7016015870711763,
"grad_norm": 1.1460477113723755,
"learning_rate": 9.540890101023972e-05,
"loss": 3.8782,
"step": 43500
},
{
"epoch": 0.7024080256770052,
"grad_norm": 1.0476806163787842,
"learning_rate": 9.539825598842291e-05,
"loss": 3.8376,
"step": 43550
},
{
"epoch": 0.7032144642828342,
"grad_norm": 1.2808899879455566,
"learning_rate": 9.538759923525857e-05,
"loss": 4.1479,
"step": 43600
},
{
"epoch": 0.704020902888663,
"grad_norm": 2.6137540340423584,
"learning_rate": 9.537693075350052e-05,
"loss": 3.8806,
"step": 43650
},
{
"epoch": 0.704827341494492,
"grad_norm": 1.4536105394363403,
"learning_rate": 9.536625054590559e-05,
"loss": 3.9868,
"step": 43700
},
{
"epoch": 0.705633780100321,
"grad_norm": 1.8267903327941895,
"learning_rate": 9.535555861523363e-05,
"loss": 4.0424,
"step": 43750
},
{
"epoch": 0.7064402187061499,
"grad_norm": 1.1175280809402466,
"learning_rate": 9.534485496424757e-05,
"loss": 4.0311,
"step": 43800
},
{
"epoch": 0.7072466573119789,
"grad_norm": 2.639409065246582,
"learning_rate": 9.533413959571332e-05,
"loss": 3.7509,
"step": 43850
},
{
"epoch": 0.7080530959178077,
"grad_norm": 2.0124926567077637,
"learning_rate": 9.532341251239982e-05,
"loss": 4.1953,
"step": 43900
},
{
"epoch": 0.7088595345236367,
"grad_norm": 1.6869808435440063,
"learning_rate": 9.531267371707906e-05,
"loss": 3.5422,
"step": 43950
},
{
"epoch": 0.7096659731294657,
"grad_norm": 2.89847469329834,
"learning_rate": 9.530192321252607e-05,
"loss": 3.8341,
"step": 44000
},
{
"epoch": 0.7104724117352946,
"grad_norm": 1.6155884265899658,
"learning_rate": 9.529116100151887e-05,
"loss": 4.1777,
"step": 44050
},
{
"epoch": 0.7112788503411235,
"grad_norm": 2.9309215545654297,
"learning_rate": 9.52803870868385e-05,
"loss": 3.8159,
"step": 44100
},
{
"epoch": 0.7120852889469524,
"grad_norm": 2.221287727355957,
"learning_rate": 9.526960147126908e-05,
"loss": 4.1738,
"step": 44150
},
{
"epoch": 0.7128917275527814,
"grad_norm": 2.1958932876586914,
"learning_rate": 9.525880415759772e-05,
"loss": 4.1857,
"step": 44200
},
{
"epoch": 0.7136981661586104,
"grad_norm": 1.2787940502166748,
"learning_rate": 9.524799514861449e-05,
"loss": 3.5715,
"step": 44250
},
{
"epoch": 0.7145046047644393,
"grad_norm": 1.3575184345245361,
"learning_rate": 9.52371744471126e-05,
"loss": 3.6301,
"step": 44300
},
{
"epoch": 0.7153110433702682,
"grad_norm": 1.1115514039993286,
"learning_rate": 9.522634205588821e-05,
"loss": 4.0579,
"step": 44350
},
{
"epoch": 0.7161174819760971,
"grad_norm": 2.445248603820801,
"learning_rate": 9.52154979777405e-05,
"loss": 3.9905,
"step": 44400
},
{
"epoch": 0.7169239205819261,
"grad_norm": 4.455885410308838,
"learning_rate": 9.52046422154717e-05,
"loss": 3.8643,
"step": 44450
},
{
"epoch": 0.7177303591877551,
"grad_norm": 2.097954034805298,
"learning_rate": 9.519377477188704e-05,
"loss": 4.2417,
"step": 44500
},
{
"epoch": 0.7185367977935839,
"grad_norm": 1.8848130702972412,
"learning_rate": 9.518289564979475e-05,
"loss": 3.7867,
"step": 44550
},
{
"epoch": 0.7193432363994129,
"grad_norm": 2.098465919494629,
"learning_rate": 9.517200485200614e-05,
"loss": 3.9569,
"step": 44600
},
{
"epoch": 0.7201496750052419,
"grad_norm": 2.0162923336029053,
"learning_rate": 9.516110238133548e-05,
"loss": 3.6984,
"step": 44650
},
{
"epoch": 0.7209561136110708,
"grad_norm": 1.9503788948059082,
"learning_rate": 9.515018824060006e-05,
"loss": 3.7651,
"step": 44700
},
{
"epoch": 0.7217625522168998,
"grad_norm": 2.917295455932617,
"learning_rate": 9.513926243262022e-05,
"loss": 3.5928,
"step": 44750
},
{
"epoch": 0.7225689908227286,
"grad_norm": 1.6340333223342896,
"learning_rate": 9.512832496021928e-05,
"loss": 3.9921,
"step": 44800
},
{
"epoch": 0.7233754294285576,
"grad_norm": 2.323976755142212,
"learning_rate": 9.51173758262236e-05,
"loss": 3.8905,
"step": 44850
},
{
"epoch": 0.7241818680343866,
"grad_norm": 2.567837715148926,
"learning_rate": 9.510641503346252e-05,
"loss": 3.973,
"step": 44900
},
{
"epoch": 0.7249883066402155,
"grad_norm": 1.992812156677246,
"learning_rate": 9.509544258476845e-05,
"loss": 4.4189,
"step": 44950
},
{
"epoch": 0.7257947452460444,
"grad_norm": 2.9449970722198486,
"learning_rate": 9.508445848297674e-05,
"loss": 3.6855,
"step": 45000
},
{
"epoch": 0.7257947452460444,
"eval_asr_loss": 0.7677889460602478,
"eval_loss": 4.220231056213379,
"eval_runtime": 14.4142,
"eval_samples_per_second": 35.52,
"eval_steps_per_second": 11.863,
"eval_tts_loss": 5.966565507839665,
"step": 45000
},
{
"epoch": 0.7266011838518733,
"grad_norm": 2.878650426864624,
"learning_rate": 9.50734627309258e-05,
"loss": 4.118,
"step": 45050
},
{
"epoch": 0.7274076224577023,
"grad_norm": 1.361701250076294,
"learning_rate": 9.506245533145707e-05,
"loss": 4.3712,
"step": 45100
},
{
"epoch": 0.7282140610635313,
"grad_norm": 2.126094341278076,
"learning_rate": 9.505143628741492e-05,
"loss": 3.7077,
"step": 45150
},
{
"epoch": 0.7290204996693602,
"grad_norm": 2.355354070663452,
"learning_rate": 9.50404056016468e-05,
"loss": 3.8391,
"step": 45200
},
{
"epoch": 0.7298269382751891,
"grad_norm": 1.933937668800354,
"learning_rate": 9.502936327700315e-05,
"loss": 3.7756,
"step": 45250
},
{
"epoch": 0.730633376881018,
"grad_norm": 2.8222224712371826,
"learning_rate": 9.50183093163374e-05,
"loss": 4.0453,
"step": 45300
},
{
"epoch": 0.731439815486847,
"grad_norm": 0.3452666699886322,
"learning_rate": 9.500724372250602e-05,
"loss": 3.9045,
"step": 45350
},
{
"epoch": 0.732246254092676,
"grad_norm": 0.9124054312705994,
"learning_rate": 9.499616649836845e-05,
"loss": 3.9615,
"step": 45400
},
{
"epoch": 0.7330526926985048,
"grad_norm": 2.117830276489258,
"learning_rate": 9.498507764678715e-05,
"loss": 4.0645,
"step": 45450
},
{
"epoch": 0.7338591313043338,
"grad_norm": 2.6719260215759277,
"learning_rate": 9.497397717062759e-05,
"loss": 3.7295,
"step": 45500
},
{
"epoch": 0.7346655699101627,
"grad_norm": 2.009535312652588,
"learning_rate": 9.496286507275825e-05,
"loss": 4.3494,
"step": 45550
},
{
"epoch": 0.7354720085159917,
"grad_norm": 1.6704541444778442,
"learning_rate": 9.495174135605058e-05,
"loss": 3.5423,
"step": 45600
},
{
"epoch": 0.7362784471218207,
"grad_norm": 2.1379082202911377,
"learning_rate": 9.494060602337909e-05,
"loss": 4.2106,
"step": 45650
},
{
"epoch": 0.7370848857276495,
"grad_norm": 1.076125979423523,
"learning_rate": 9.492945907762123e-05,
"loss": 4.1118,
"step": 45700
},
{
"epoch": 0.7378913243334785,
"grad_norm": 2.360368013381958,
"learning_rate": 9.491830052165749e-05,
"loss": 3.9588,
"step": 45750
},
{
"epoch": 0.7386977629393074,
"grad_norm": 1.9395514726638794,
"learning_rate": 9.490713035837133e-05,
"loss": 3.9243,
"step": 45800
},
{
"epoch": 0.7395042015451364,
"grad_norm": 1.406371831893921,
"learning_rate": 9.489594859064926e-05,
"loss": 4.1377,
"step": 45850
},
{
"epoch": 0.7403106401509653,
"grad_norm": 1.5525023937225342,
"learning_rate": 9.488475522138073e-05,
"loss": 3.6349,
"step": 45900
},
{
"epoch": 0.7411170787567942,
"grad_norm": 1.3962270021438599,
"learning_rate": 9.487355025345823e-05,
"loss": 3.8225,
"step": 45950
},
{
"epoch": 0.7419235173626232,
"grad_norm": 2.4570751190185547,
"learning_rate": 9.486255813467055e-05,
"loss": 3.8946,
"step": 46000
},
{
"epoch": 0.7427299559684521,
"grad_norm": 9.147421836853027,
"learning_rate": 9.485133020995828e-05,
"loss": 3.9156,
"step": 46050
},
{
"epoch": 0.7435363945742811,
"grad_norm": 1.4232938289642334,
"learning_rate": 9.484009069522938e-05,
"loss": 4.113,
"step": 46100
},
{
"epoch": 0.74434283318011,
"grad_norm": 3.1398966312408447,
"learning_rate": 9.482883959338827e-05,
"loss": 4.0905,
"step": 46150
},
{
"epoch": 0.7451492717859389,
"grad_norm": 2.162477493286133,
"learning_rate": 9.481757690734232e-05,
"loss": 3.9378,
"step": 46200
},
{
"epoch": 0.7459557103917679,
"grad_norm": 2.1014811992645264,
"learning_rate": 9.48063026400019e-05,
"loss": 3.9758,
"step": 46250
},
{
"epoch": 0.7467621489975969,
"grad_norm": 1.6452032327651978,
"learning_rate": 9.479501679428042e-05,
"loss": 4.0847,
"step": 46300
},
{
"epoch": 0.7475685876034257,
"grad_norm": 2.0140156745910645,
"learning_rate": 9.478371937309425e-05,
"loss": 4.3145,
"step": 46350
},
{
"epoch": 0.7483750262092547,
"grad_norm": 1.8627792596817017,
"learning_rate": 9.477241037936273e-05,
"loss": 4.0373,
"step": 46400
},
{
"epoch": 0.7491814648150836,
"grad_norm": 1.3735127449035645,
"learning_rate": 9.476108981600822e-05,
"loss": 3.9577,
"step": 46450
},
{
"epoch": 0.7499879034209126,
"grad_norm": 2.6094725131988525,
"learning_rate": 9.474975768595609e-05,
"loss": 3.957,
"step": 46500
},
{
"epoch": 0.7507943420267416,
"grad_norm": 1.3980075120925903,
"learning_rate": 9.473841399213463e-05,
"loss": 4.1313,
"step": 46550
},
{
"epoch": 0.7516007806325704,
"grad_norm": 3.6238787174224854,
"learning_rate": 9.47270587374752e-05,
"loss": 3.9414,
"step": 46600
},
{
"epoch": 0.7524072192383994,
"grad_norm": 1.450015902519226,
"learning_rate": 9.471569192491208e-05,
"loss": 3.7486,
"step": 46650
},
{
"epoch": 0.7532136578442283,
"grad_norm": 1.9023935794830322,
"learning_rate": 9.470431355738257e-05,
"loss": 4.0542,
"step": 46700
},
{
"epoch": 0.7540200964500573,
"grad_norm": 1.0241719484329224,
"learning_rate": 9.469292363782696e-05,
"loss": 3.9859,
"step": 46750
},
{
"epoch": 0.7548265350558861,
"grad_norm": 1.9951140880584717,
"learning_rate": 9.468152216918849e-05,
"loss": 3.5983,
"step": 46800
},
{
"epoch": 0.7556329736617151,
"grad_norm": 1.519264817237854,
"learning_rate": 9.467010915441343e-05,
"loss": 3.971,
"step": 46850
},
{
"epoch": 0.7564394122675441,
"grad_norm": 2.4989826679229736,
"learning_rate": 9.465868459645102e-05,
"loss": 3.9566,
"step": 46900
},
{
"epoch": 0.757245850873373,
"grad_norm": 2.34820294380188,
"learning_rate": 9.464724849825347e-05,
"loss": 4.1283,
"step": 46950
},
{
"epoch": 0.758052289479202,
"grad_norm": 1.598359227180481,
"learning_rate": 9.463580086277598e-05,
"loss": 3.8883,
"step": 47000
},
{
"epoch": 0.7588587280850309,
"grad_norm": 1.438071370124817,
"learning_rate": 9.462434169297671e-05,
"loss": 3.6827,
"step": 47050
},
{
"epoch": 0.7596651666908598,
"grad_norm": 0.7301783561706543,
"learning_rate": 9.461287099181683e-05,
"loss": 3.9224,
"step": 47100
},
{
"epoch": 0.7604716052966888,
"grad_norm": 4.070219039916992,
"learning_rate": 9.46013887622605e-05,
"loss": 4.0544,
"step": 47150
},
{
"epoch": 0.7612780439025177,
"grad_norm": 2.0918643474578857,
"learning_rate": 9.458989500727481e-05,
"loss": 3.7879,
"step": 47200
},
{
"epoch": 0.7620844825083466,
"grad_norm": 2.473740577697754,
"learning_rate": 9.457838972982986e-05,
"loss": 4.0541,
"step": 47250
},
{
"epoch": 0.7628909211141756,
"grad_norm": 1.9549883604049683,
"learning_rate": 9.456687293289877e-05,
"loss": 4.3095,
"step": 47300
},
{
"epoch": 0.7636973597200045,
"grad_norm": 1.1889281272888184,
"learning_rate": 9.455534461945751e-05,
"loss": 3.9717,
"step": 47350
},
{
"epoch": 0.7645037983258335,
"grad_norm": 2.1214423179626465,
"learning_rate": 9.45438047924852e-05,
"loss": 3.9387,
"step": 47400
},
{
"epoch": 0.7653102369316624,
"grad_norm": 1.206317663192749,
"learning_rate": 9.453225345496377e-05,
"loss": 4.1136,
"step": 47450
},
{
"epoch": 0.7661166755374913,
"grad_norm": 2.127297878265381,
"learning_rate": 9.452069060987823e-05,
"loss": 3.6162,
"step": 47500
},
{
"epoch": 0.7669231141433203,
"grad_norm": 1.8844021558761597,
"learning_rate": 9.450911626021653e-05,
"loss": 3.9547,
"step": 47550
},
{
"epoch": 0.7677295527491492,
"grad_norm": 1.8180209398269653,
"learning_rate": 9.449753040896958e-05,
"loss": 4.0026,
"step": 47600
},
{
"epoch": 0.7685359913549782,
"grad_norm": 4.184787273406982,
"learning_rate": 9.448593305913129e-05,
"loss": 3.5137,
"step": 47650
},
{
"epoch": 0.769342429960807,
"grad_norm": 1.7773044109344482,
"learning_rate": 9.447432421369851e-05,
"loss": 3.6649,
"step": 47700
},
{
"epoch": 0.770148868566636,
"grad_norm": 2.41501784324646,
"learning_rate": 9.446270387567109e-05,
"loss": 4.1824,
"step": 47750
},
{
"epoch": 0.770955307172465,
"grad_norm": 1.6952520608901978,
"learning_rate": 9.445107204805183e-05,
"loss": 3.5778,
"step": 47800
},
{
"epoch": 0.7717617457782939,
"grad_norm": 3.045311212539673,
"learning_rate": 9.443942873384653e-05,
"loss": 3.9632,
"step": 47850
},
{
"epoch": 0.7725681843841229,
"grad_norm": 2.5079586505889893,
"learning_rate": 9.442777393606393e-05,
"loss": 3.8897,
"step": 47900
},
{
"epoch": 0.7733746229899517,
"grad_norm": 2.440715789794922,
"learning_rate": 9.44161076577157e-05,
"loss": 3.7363,
"step": 47950
},
{
"epoch": 0.7741810615957807,
"grad_norm": 1.950965166091919,
"learning_rate": 9.440442990181656e-05,
"loss": 3.7987,
"step": 48000
},
{
"epoch": 0.7741810615957807,
"eval_asr_loss": 0.7671968561538705,
"eval_loss": 4.1380295753479,
"eval_runtime": 14.6173,
"eval_samples_per_second": 35.027,
"eval_steps_per_second": 11.698,
"eval_tts_loss": 6.009661002101706,
"step": 48000
}
],
"logging_steps": 50,
"max_steps": 310005,
"num_input_tokens_seen": 0,
"num_train_epochs": 5,
"save_steps": 3000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 3.946933255059538e+17,
"train_batch_size": 3,
"trial_name": null,
"trial_params": null
}