|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.7741810615957807, |
|
"eval_steps": 3000, |
|
"global_step": 48000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0008064386058289383, |
|
"grad_norm": 6.343989849090576, |
|
"learning_rate": 4.800000000000001e-06, |
|
"loss": 3.2595, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.0016128772116578765, |
|
"grad_norm": 3.7637691497802734, |
|
"learning_rate": 9.800000000000001e-06, |
|
"loss": 3.2397, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.002419315817486815, |
|
"grad_norm": 4.01455020904541, |
|
"learning_rate": 1.48e-05, |
|
"loss": 3.1471, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.003225754423315753, |
|
"grad_norm": 5.399896621704102, |
|
"learning_rate": 1.9800000000000004e-05, |
|
"loss": 3.3162, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.004032193029144691, |
|
"grad_norm": 5.194526195526123, |
|
"learning_rate": 2.48e-05, |
|
"loss": 3.4495, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.00483863163497363, |
|
"grad_norm": 4.043808460235596, |
|
"learning_rate": 2.98e-05, |
|
"loss": 3.4038, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.005645070240802567, |
|
"grad_norm": 1.6968616247177124, |
|
"learning_rate": 3.46e-05, |
|
"loss": 3.5791, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.006451508846631506, |
|
"grad_norm": 3.515857219696045, |
|
"learning_rate": 3.960000000000001e-05, |
|
"loss": 3.4015, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.007257947452460444, |
|
"grad_norm": 3.327608108520508, |
|
"learning_rate": 4.46e-05, |
|
"loss": 3.2794, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.008064386058289382, |
|
"grad_norm": 3.6122305393218994, |
|
"learning_rate": 4.96e-05, |
|
"loss": 3.3541, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.00887082466411832, |
|
"grad_norm": 4.593649864196777, |
|
"learning_rate": 5.4600000000000006e-05, |
|
"loss": 3.6442, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.00967726326994726, |
|
"grad_norm": 2.58380126953125, |
|
"learning_rate": 5.96e-05, |
|
"loss": 3.1673, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.010483701875776198, |
|
"grad_norm": 5.36879825592041, |
|
"learning_rate": 6.460000000000001e-05, |
|
"loss": 3.5224, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.011290140481605135, |
|
"grad_norm": 21.212533950805664, |
|
"learning_rate": 6.96e-05, |
|
"loss": 3.5144, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.012096579087434074, |
|
"grad_norm": 1.9257746934890747, |
|
"learning_rate": 7.46e-05, |
|
"loss": 3.2471, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.012903017693263012, |
|
"grad_norm": 6.157283306121826, |
|
"learning_rate": 7.960000000000001e-05, |
|
"loss": 3.551, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.013709456299091951, |
|
"grad_norm": 4.075947284698486, |
|
"learning_rate": 8.46e-05, |
|
"loss": 3.5548, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.014515894904920888, |
|
"grad_norm": 3.9454638957977295, |
|
"learning_rate": 8.960000000000001e-05, |
|
"loss": 3.5653, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.015322333510749827, |
|
"grad_norm": 4.633170127868652, |
|
"learning_rate": 9.46e-05, |
|
"loss": 3.3589, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.016128772116578764, |
|
"grad_norm": 4.273008346557617, |
|
"learning_rate": 9.960000000000001e-05, |
|
"loss": 3.8061, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.016935210722407702, |
|
"grad_norm": 3.094942808151245, |
|
"learning_rate": 9.999999453205175e-05, |
|
"loss": 3.6132, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.01774164932823664, |
|
"grad_norm": 3.022191047668457, |
|
"learning_rate": 9.999997618496786e-05, |
|
"loss": 3.8753, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.01854808793406558, |
|
"grad_norm": 3.715566635131836, |
|
"learning_rate": 9.999994491740752e-05, |
|
"loss": 3.5987, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.01935452653989452, |
|
"grad_norm": 3.9486517906188965, |
|
"learning_rate": 9.999990072937881e-05, |
|
"loss": 3.7496, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.020160965145723457, |
|
"grad_norm": 4.180655002593994, |
|
"learning_rate": 9.999984362089316e-05, |
|
"loss": 4.1646, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.020967403751552396, |
|
"grad_norm": 4.141788959503174, |
|
"learning_rate": 9.999977359196532e-05, |
|
"loss": 3.8424, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.02177384235738133, |
|
"grad_norm": 6.864413261413574, |
|
"learning_rate": 9.999969064261341e-05, |
|
"loss": 3.79, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.02258028096321027, |
|
"grad_norm": 3.0512356758117676, |
|
"learning_rate": 9.999959477285881e-05, |
|
"loss": 3.7211, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.02338671956903921, |
|
"grad_norm": 3.6865158081054688, |
|
"learning_rate": 9.999948598272635e-05, |
|
"loss": 3.729, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.024193158174868147, |
|
"grad_norm": 3.479696750640869, |
|
"learning_rate": 9.999936427224411e-05, |
|
"loss": 3.8229, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.024999596780697086, |
|
"grad_norm": 3.896265745162964, |
|
"learning_rate": 9.999922964144355e-05, |
|
"loss": 4.0188, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.025806035386526024, |
|
"grad_norm": 1.8305364847183228, |
|
"learning_rate": 9.999908209035947e-05, |
|
"loss": 3.6971, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.026612473992354963, |
|
"grad_norm": 3.6553308963775635, |
|
"learning_rate": 9.999892161902997e-05, |
|
"loss": 3.9225, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.027418912598183902, |
|
"grad_norm": 3.454449415206909, |
|
"learning_rate": 9.999874822749654e-05, |
|
"loss": 4.1062, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.028225351204012837, |
|
"grad_norm": 6.318528652191162, |
|
"learning_rate": 9.9998561915804e-05, |
|
"loss": 4.165, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.029031789809841776, |
|
"grad_norm": 3.34511399269104, |
|
"learning_rate": 9.999836268400044e-05, |
|
"loss": 4.1559, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.029838228415670714, |
|
"grad_norm": 3.3025155067443848, |
|
"learning_rate": 9.999815053213741e-05, |
|
"loss": 3.6104, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.030644667021499653, |
|
"grad_norm": 1.3363052606582642, |
|
"learning_rate": 9.999792546026971e-05, |
|
"loss": 4.3776, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.03145110562732859, |
|
"grad_norm": 4.033843040466309, |
|
"learning_rate": 9.999768746845547e-05, |
|
"loss": 3.838, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 0.03225754423315753, |
|
"grad_norm": 3.517301082611084, |
|
"learning_rate": 9.999743655675622e-05, |
|
"loss": 4.0525, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.03306398283898647, |
|
"grad_norm": 3.5909571647644043, |
|
"learning_rate": 9.999717272523678e-05, |
|
"loss": 3.672, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 0.033870421444815405, |
|
"grad_norm": 2.088115930557251, |
|
"learning_rate": 9.999689597396536e-05, |
|
"loss": 3.9254, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.03467686005064435, |
|
"grad_norm": 3.169725179672241, |
|
"learning_rate": 9.999660630301345e-05, |
|
"loss": 3.9372, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 0.03548329865647328, |
|
"grad_norm": 3.584789991378784, |
|
"learning_rate": 9.999630371245588e-05, |
|
"loss": 3.9506, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.036289737262302224, |
|
"grad_norm": 2.678903579711914, |
|
"learning_rate": 9.999598820237088e-05, |
|
"loss": 3.7822, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 0.03709617586813116, |
|
"grad_norm": 2.6614108085632324, |
|
"learning_rate": 9.999565977283998e-05, |
|
"loss": 3.9777, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.037902614473960095, |
|
"grad_norm": 3.998502731323242, |
|
"learning_rate": 9.999531842394802e-05, |
|
"loss": 3.8676, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 0.03870905307978904, |
|
"grad_norm": 3.5454630851745605, |
|
"learning_rate": 9.999496415578324e-05, |
|
"loss": 3.9342, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.03951549168561797, |
|
"grad_norm": 4.353825569152832, |
|
"learning_rate": 9.999459696843719e-05, |
|
"loss": 4.0824, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 0.040321930291446914, |
|
"grad_norm": 2.4952590465545654, |
|
"learning_rate": 9.99942168620047e-05, |
|
"loss": 3.7487, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.04112836889727585, |
|
"grad_norm": 3.804739475250244, |
|
"learning_rate": 9.999382383658405e-05, |
|
"loss": 3.9278, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 0.04193480750310479, |
|
"grad_norm": 4.888896465301514, |
|
"learning_rate": 9.999341789227679e-05, |
|
"loss": 3.976, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.04274124610893373, |
|
"grad_norm": 2.8611414432525635, |
|
"learning_rate": 9.99929990291878e-05, |
|
"loss": 3.8008, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 0.04354768471476266, |
|
"grad_norm": 3.469597816467285, |
|
"learning_rate": 9.999256724742533e-05, |
|
"loss": 3.9826, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.044354123320591604, |
|
"grad_norm": 4.003543376922607, |
|
"learning_rate": 9.999212254710095e-05, |
|
"loss": 3.7801, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 0.04516056192642054, |
|
"grad_norm": 3.900905132293701, |
|
"learning_rate": 9.99916649283296e-05, |
|
"loss": 4.081, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.04596700053224948, |
|
"grad_norm": 1.68619966506958, |
|
"learning_rate": 9.99911943912295e-05, |
|
"loss": 3.8061, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 0.04677343913807842, |
|
"grad_norm": 4.65512752532959, |
|
"learning_rate": 9.999071093592227e-05, |
|
"loss": 3.8544, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.04757987774390736, |
|
"grad_norm": 2.2502596378326416, |
|
"learning_rate": 9.999021456253282e-05, |
|
"loss": 4.2133, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 0.048386316349736294, |
|
"grad_norm": 4.234059810638428, |
|
"learning_rate": 9.998970527118941e-05, |
|
"loss": 3.8952, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.048386316349736294, |
|
"eval_asr_loss": 0.6124729285552434, |
|
"eval_loss": 5.14359712600708, |
|
"eval_runtime": 17.1329, |
|
"eval_samples_per_second": 29.884, |
|
"eval_steps_per_second": 9.981, |
|
"eval_tts_loss": 6.1080074336570815, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.04919275495556523, |
|
"grad_norm": 4.5945658683776855, |
|
"learning_rate": 9.998918306202368e-05, |
|
"loss": 3.7152, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 0.04999919356139417, |
|
"grad_norm": 3.4985783100128174, |
|
"learning_rate": 9.998864793517053e-05, |
|
"loss": 4.0368, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.05080563216722311, |
|
"grad_norm": 2.991260528564453, |
|
"learning_rate": 9.99880998907683e-05, |
|
"loss": 3.961, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 0.05161207077305205, |
|
"grad_norm": 7.048410415649414, |
|
"learning_rate": 9.998753892895854e-05, |
|
"loss": 3.9461, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.052418509378880984, |
|
"grad_norm": 4.608008861541748, |
|
"learning_rate": 9.998696504988625e-05, |
|
"loss": 3.863, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 0.053224947984709926, |
|
"grad_norm": 3.2530033588409424, |
|
"learning_rate": 9.998637825369973e-05, |
|
"loss": 3.8894, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.05403138659053886, |
|
"grad_norm": 2.5475354194641113, |
|
"learning_rate": 9.99857785405506e-05, |
|
"loss": 3.7206, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 0.054837825196367804, |
|
"grad_norm": 4.753456115722656, |
|
"learning_rate": 9.998516591059384e-05, |
|
"loss": 3.7273, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.05564426380219674, |
|
"grad_norm": 2.6090848445892334, |
|
"learning_rate": 9.998454036398774e-05, |
|
"loss": 3.8879, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 0.056450702408025674, |
|
"grad_norm": 1.5343598127365112, |
|
"learning_rate": 9.998390190089396e-05, |
|
"loss": 4.0087, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.057257141013854616, |
|
"grad_norm": 4.9131364822387695, |
|
"learning_rate": 9.99832505214775e-05, |
|
"loss": 3.8471, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 0.05806357961968355, |
|
"grad_norm": 2.1183621883392334, |
|
"learning_rate": 9.998258622590666e-05, |
|
"loss": 3.9289, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.058870018225512494, |
|
"grad_norm": 3.5901293754577637, |
|
"learning_rate": 9.99819090143531e-05, |
|
"loss": 3.8583, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 0.05967645683134143, |
|
"grad_norm": 2.5325889587402344, |
|
"learning_rate": 9.998121888699183e-05, |
|
"loss": 3.953, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 0.06048289543717037, |
|
"grad_norm": 2.1782755851745605, |
|
"learning_rate": 9.99805158440012e-05, |
|
"loss": 4.013, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 0.061289334042999306, |
|
"grad_norm": 2.2709741592407227, |
|
"learning_rate": 9.997979988556285e-05, |
|
"loss": 4.1835, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 0.06209577264882824, |
|
"grad_norm": 1.7206313610076904, |
|
"learning_rate": 9.997907101186181e-05, |
|
"loss": 3.7028, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 0.06290221125465718, |
|
"grad_norm": 2.8325114250183105, |
|
"learning_rate": 9.997832922308643e-05, |
|
"loss": 3.8754, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 0.06370864986048612, |
|
"grad_norm": 3.2632429599761963, |
|
"learning_rate": 9.997757451942839e-05, |
|
"loss": 4.2381, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 0.06451508846631505, |
|
"grad_norm": 4.061680793762207, |
|
"learning_rate": 9.997680690108271e-05, |
|
"loss": 3.9123, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.065321527072144, |
|
"grad_norm": 3.529866933822632, |
|
"learning_rate": 9.997602636824775e-05, |
|
"loss": 3.8652, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 0.06612796567797294, |
|
"grad_norm": 3.3136708736419678, |
|
"learning_rate": 9.997523292112521e-05, |
|
"loss": 3.9462, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 0.06693440428380187, |
|
"grad_norm": 1.6283220052719116, |
|
"learning_rate": 9.997442655992013e-05, |
|
"loss": 4.1202, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 0.06774084288963081, |
|
"grad_norm": 3.355297088623047, |
|
"learning_rate": 9.997360728484086e-05, |
|
"loss": 4.0791, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 0.06854728149545974, |
|
"grad_norm": 2.3169095516204834, |
|
"learning_rate": 9.997277509609914e-05, |
|
"loss": 3.6696, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 0.0693537201012887, |
|
"grad_norm": 2.0649425983428955, |
|
"learning_rate": 9.997192999391e-05, |
|
"loss": 4.1419, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 0.07016015870711763, |
|
"grad_norm": 2.268578290939331, |
|
"learning_rate": 9.99710719784918e-05, |
|
"loss": 3.9995, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 0.07096659731294656, |
|
"grad_norm": 2.035593271255493, |
|
"learning_rate": 9.997020105006632e-05, |
|
"loss": 3.7817, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 0.0717730359187755, |
|
"grad_norm": 1.604856014251709, |
|
"learning_rate": 9.996931720885855e-05, |
|
"loss": 3.8819, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 0.07257947452460445, |
|
"grad_norm": 1.7782434225082397, |
|
"learning_rate": 9.996842045509692e-05, |
|
"loss": 4.2321, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.07338591313043338, |
|
"grad_norm": 1.1473257541656494, |
|
"learning_rate": 9.996751078901314e-05, |
|
"loss": 4.123, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 0.07419235173626232, |
|
"grad_norm": 4.596096515655518, |
|
"learning_rate": 9.996658821084231e-05, |
|
"loss": 4.4983, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 0.07499879034209125, |
|
"grad_norm": 4.046219348907471, |
|
"learning_rate": 9.99656527208228e-05, |
|
"loss": 3.832, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 0.07580522894792019, |
|
"grad_norm": 3.365840435028076, |
|
"learning_rate": 9.996470431919634e-05, |
|
"loss": 3.8089, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 0.07661166755374914, |
|
"grad_norm": 2.3088300228118896, |
|
"learning_rate": 9.996374300620805e-05, |
|
"loss": 4.2963, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 0.07741810615957807, |
|
"grad_norm": 3.0989160537719727, |
|
"learning_rate": 9.996276878210631e-05, |
|
"loss": 3.9085, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 0.07822454476540701, |
|
"grad_norm": 3.6229944229125977, |
|
"learning_rate": 9.996178164714289e-05, |
|
"loss": 3.9887, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 0.07903098337123594, |
|
"grad_norm": 2.169739007949829, |
|
"learning_rate": 9.996078160157285e-05, |
|
"loss": 4.053, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 0.07983742197706489, |
|
"grad_norm": 1.102918267250061, |
|
"learning_rate": 9.995976864565462e-05, |
|
"loss": 3.9429, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 0.08064386058289383, |
|
"grad_norm": 3.0187597274780273, |
|
"learning_rate": 9.995874277964997e-05, |
|
"loss": 4.1842, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.08145029918872276, |
|
"grad_norm": 6.297091484069824, |
|
"learning_rate": 9.995770400382398e-05, |
|
"loss": 3.8538, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 0.0822567377945517, |
|
"grad_norm": 2.396947145462036, |
|
"learning_rate": 9.995665231844511e-05, |
|
"loss": 3.9006, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 0.08306317640038063, |
|
"grad_norm": 1.8022737503051758, |
|
"learning_rate": 9.995558772378509e-05, |
|
"loss": 4.0055, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 0.08386961500620958, |
|
"grad_norm": 2.959127902984619, |
|
"learning_rate": 9.995451022011902e-05, |
|
"loss": 3.7882, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 0.08467605361203852, |
|
"grad_norm": 1.0544005632400513, |
|
"learning_rate": 9.995341980772535e-05, |
|
"loss": 3.8215, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 0.08548249221786745, |
|
"grad_norm": 4.3189616203308105, |
|
"learning_rate": 9.995231648688586e-05, |
|
"loss": 3.8656, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 0.08628893082369639, |
|
"grad_norm": 2.1376607418060303, |
|
"learning_rate": 9.995120025788565e-05, |
|
"loss": 3.69, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 0.08709536942952532, |
|
"grad_norm": 2.437530994415283, |
|
"learning_rate": 9.995007112101316e-05, |
|
"loss": 3.972, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 0.08790180803535427, |
|
"grad_norm": 2.361795663833618, |
|
"learning_rate": 9.994892907656018e-05, |
|
"loss": 3.9913, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 0.08870824664118321, |
|
"grad_norm": 1.6262253522872925, |
|
"learning_rate": 9.994777412482183e-05, |
|
"loss": 3.9927, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.08951468524701214, |
|
"grad_norm": 4.346071243286133, |
|
"learning_rate": 9.994660626609654e-05, |
|
"loss": 3.7355, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 0.09032112385284108, |
|
"grad_norm": 3.2261524200439453, |
|
"learning_rate": 9.994542550068612e-05, |
|
"loss": 3.8608, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 0.09112756245867003, |
|
"grad_norm": 3.2692320346832275, |
|
"learning_rate": 9.994423182889567e-05, |
|
"loss": 4.0809, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 0.09193400106449896, |
|
"grad_norm": 4.047648906707764, |
|
"learning_rate": 9.994302525103366e-05, |
|
"loss": 3.7864, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 0.0927404396703279, |
|
"grad_norm": 3.113825559616089, |
|
"learning_rate": 9.994180576741188e-05, |
|
"loss": 3.6328, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 0.09354687827615683, |
|
"grad_norm": 2.295882225036621, |
|
"learning_rate": 9.994057337834545e-05, |
|
"loss": 3.6168, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 0.09435331688198577, |
|
"grad_norm": 3.2631022930145264, |
|
"learning_rate": 9.993932808415283e-05, |
|
"loss": 4.1689, |
|
"step": 5850 |
|
}, |
|
{ |
|
"epoch": 0.09515975548781472, |
|
"grad_norm": 3.607879877090454, |
|
"learning_rate": 9.993806988515581e-05, |
|
"loss": 4.0071, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 0.09596619409364365, |
|
"grad_norm": 1.2808860540390015, |
|
"learning_rate": 9.993679878167957e-05, |
|
"loss": 3.9552, |
|
"step": 5950 |
|
}, |
|
{ |
|
"epoch": 0.09677263269947259, |
|
"grad_norm": 1.4205412864685059, |
|
"learning_rate": 9.99355147740525e-05, |
|
"loss": 4.4215, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.09677263269947259, |
|
"eval_asr_loss": 0.5728164760736751, |
|
"eval_loss": 5.094964027404785, |
|
"eval_runtime": 16.7544, |
|
"eval_samples_per_second": 30.559, |
|
"eval_steps_per_second": 10.206, |
|
"eval_tts_loss": 6.0641644553514515, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.09757907130530152, |
|
"grad_norm": 2.5851426124572754, |
|
"learning_rate": 9.993421786260645e-05, |
|
"loss": 3.941, |
|
"step": 6050 |
|
}, |
|
{ |
|
"epoch": 0.09838550991113046, |
|
"grad_norm": 4.519574165344238, |
|
"learning_rate": 9.993290804767654e-05, |
|
"loss": 3.6548, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 0.09919194851695941, |
|
"grad_norm": 2.023059368133545, |
|
"learning_rate": 9.993158532960124e-05, |
|
"loss": 3.7762, |
|
"step": 6150 |
|
}, |
|
{ |
|
"epoch": 0.09999838712278834, |
|
"grad_norm": 3.167097330093384, |
|
"learning_rate": 9.993024970872236e-05, |
|
"loss": 3.7466, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 0.10080482572861728, |
|
"grad_norm": 2.1411142349243164, |
|
"learning_rate": 9.992890118538502e-05, |
|
"loss": 3.7702, |
|
"step": 6250 |
|
}, |
|
{ |
|
"epoch": 0.10161126433444621, |
|
"grad_norm": 2.5199944972991943, |
|
"learning_rate": 9.99275397599377e-05, |
|
"loss": 4.1723, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 0.10241770294027516, |
|
"grad_norm": 2.112910270690918, |
|
"learning_rate": 9.992616543273222e-05, |
|
"loss": 3.6662, |
|
"step": 6350 |
|
}, |
|
{ |
|
"epoch": 0.1032241415461041, |
|
"grad_norm": 2.290219306945801, |
|
"learning_rate": 9.992477820412369e-05, |
|
"loss": 4.1363, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 0.10403058015193303, |
|
"grad_norm": 3.748711585998535, |
|
"learning_rate": 9.992337807447061e-05, |
|
"loss": 4.0218, |
|
"step": 6450 |
|
}, |
|
{ |
|
"epoch": 0.10483701875776197, |
|
"grad_norm": 14.197625160217285, |
|
"learning_rate": 9.992196504413478e-05, |
|
"loss": 4.11, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.1056434573635909, |
|
"grad_norm": 2.3828155994415283, |
|
"learning_rate": 9.992053911348132e-05, |
|
"loss": 3.9976, |
|
"step": 6550 |
|
}, |
|
{ |
|
"epoch": 0.10644989596941985, |
|
"grad_norm": 3.1666653156280518, |
|
"learning_rate": 9.991910028287874e-05, |
|
"loss": 3.8263, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 0.10725633457524879, |
|
"grad_norm": 1.068214774131775, |
|
"learning_rate": 9.991764855269882e-05, |
|
"loss": 3.7559, |
|
"step": 6650 |
|
}, |
|
{ |
|
"epoch": 0.10806277318107772, |
|
"grad_norm": 2.6692686080932617, |
|
"learning_rate": 9.991618392331672e-05, |
|
"loss": 3.8083, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 0.10886921178690666, |
|
"grad_norm": 3.22259783744812, |
|
"learning_rate": 9.99147063951109e-05, |
|
"loss": 3.9687, |
|
"step": 6750 |
|
}, |
|
{ |
|
"epoch": 0.10967565039273561, |
|
"grad_norm": 2.2858736515045166, |
|
"learning_rate": 9.991321596846317e-05, |
|
"loss": 3.7381, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 0.11048208899856454, |
|
"grad_norm": 3.2024693489074707, |
|
"learning_rate": 9.991171264375869e-05, |
|
"loss": 4.1968, |
|
"step": 6850 |
|
}, |
|
{ |
|
"epoch": 0.11128852760439348, |
|
"grad_norm": 2.7798397541046143, |
|
"learning_rate": 9.99101964213859e-05, |
|
"loss": 3.7984, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 0.11209496621022241, |
|
"grad_norm": 2.2842605113983154, |
|
"learning_rate": 9.990866730173662e-05, |
|
"loss": 4.2311, |
|
"step": 6950 |
|
}, |
|
{ |
|
"epoch": 0.11290140481605135, |
|
"grad_norm": 4.10762357711792, |
|
"learning_rate": 9.9907125285206e-05, |
|
"loss": 4.1217, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.1137078434218803, |
|
"grad_norm": 1.8413517475128174, |
|
"learning_rate": 9.99055703721925e-05, |
|
"loss": 4.1848, |
|
"step": 7050 |
|
}, |
|
{ |
|
"epoch": 0.11451428202770923, |
|
"grad_norm": 2.565972328186035, |
|
"learning_rate": 9.990400256309792e-05, |
|
"loss": 3.744, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 0.11532072063353817, |
|
"grad_norm": 2.643573760986328, |
|
"learning_rate": 9.990242185832744e-05, |
|
"loss": 4.2744, |
|
"step": 7150 |
|
}, |
|
{ |
|
"epoch": 0.1161271592393671, |
|
"grad_norm": 2.1358284950256348, |
|
"learning_rate": 9.990082825828946e-05, |
|
"loss": 4.2883, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 0.11693359784519604, |
|
"grad_norm": 1.9578022956848145, |
|
"learning_rate": 9.989922176339583e-05, |
|
"loss": 3.8243, |
|
"step": 7250 |
|
}, |
|
{ |
|
"epoch": 0.11774003645102499, |
|
"grad_norm": 5.736109256744385, |
|
"learning_rate": 9.989760237406168e-05, |
|
"loss": 4.1798, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 0.11854647505685392, |
|
"grad_norm": 3.454503059387207, |
|
"learning_rate": 9.989597009070546e-05, |
|
"loss": 4.1567, |
|
"step": 7350 |
|
}, |
|
{ |
|
"epoch": 0.11935291366268286, |
|
"grad_norm": 3.161022424697876, |
|
"learning_rate": 9.989432491374897e-05, |
|
"loss": 3.7965, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 0.1201593522685118, |
|
"grad_norm": 2.6754753589630127, |
|
"learning_rate": 9.989266684361735e-05, |
|
"loss": 3.9182, |
|
"step": 7450 |
|
}, |
|
{ |
|
"epoch": 0.12096579087434074, |
|
"grad_norm": 2.2374236583709717, |
|
"learning_rate": 9.989099588073905e-05, |
|
"loss": 4.1068, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.12177222948016968, |
|
"grad_norm": 2.775463581085205, |
|
"learning_rate": 9.988931202554588e-05, |
|
"loss": 4.0023, |
|
"step": 7550 |
|
}, |
|
{ |
|
"epoch": 0.12257866808599861, |
|
"grad_norm": 2.515188217163086, |
|
"learning_rate": 9.988761527847293e-05, |
|
"loss": 4.0234, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 0.12338510669182755, |
|
"grad_norm": 4.208799362182617, |
|
"learning_rate": 9.988590563995869e-05, |
|
"loss": 3.9328, |
|
"step": 7650 |
|
}, |
|
{ |
|
"epoch": 0.12419154529765648, |
|
"grad_norm": 2.9487059116363525, |
|
"learning_rate": 9.988418311044495e-05, |
|
"loss": 3.8718, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 0.12499798390348543, |
|
"grad_norm": 2.8759098052978516, |
|
"learning_rate": 9.98824476903768e-05, |
|
"loss": 3.8538, |
|
"step": 7750 |
|
}, |
|
{ |
|
"epoch": 0.12580442250931437, |
|
"grad_norm": 2.848405599594116, |
|
"learning_rate": 9.98806993802027e-05, |
|
"loss": 3.8749, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 0.1266108611151433, |
|
"grad_norm": 1.7490386962890625, |
|
"learning_rate": 9.987893818037443e-05, |
|
"loss": 3.9094, |
|
"step": 7850 |
|
}, |
|
{ |
|
"epoch": 0.12741729972097224, |
|
"grad_norm": 2.6559858322143555, |
|
"learning_rate": 9.987716409134712e-05, |
|
"loss": 3.825, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 0.12822373832680117, |
|
"grad_norm": 2.877697229385376, |
|
"learning_rate": 9.987537711357919e-05, |
|
"loss": 3.9497, |
|
"step": 7950 |
|
}, |
|
{ |
|
"epoch": 0.1290301769326301, |
|
"grad_norm": 2.17142391204834, |
|
"learning_rate": 9.987357724753242e-05, |
|
"loss": 3.9387, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.12983661553845904, |
|
"grad_norm": 1.643707036972046, |
|
"learning_rate": 9.987176449367192e-05, |
|
"loss": 4.2619, |
|
"step": 8050 |
|
}, |
|
{ |
|
"epoch": 0.130643054144288, |
|
"grad_norm": 4.4552998542785645, |
|
"learning_rate": 9.98699388524661e-05, |
|
"loss": 4.0635, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 0.13144949275011694, |
|
"grad_norm": 3.406740427017212, |
|
"learning_rate": 9.986810032438674e-05, |
|
"loss": 4.0747, |
|
"step": 8150 |
|
}, |
|
{ |
|
"epoch": 0.13225593135594588, |
|
"grad_norm": 5.450404644012451, |
|
"learning_rate": 9.986624890990893e-05, |
|
"loss": 3.8341, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 0.1330623699617748, |
|
"grad_norm": 2.6635243892669678, |
|
"learning_rate": 9.986438460951111e-05, |
|
"loss": 3.8464, |
|
"step": 8250 |
|
}, |
|
{ |
|
"epoch": 0.13386880856760375, |
|
"grad_norm": 1.887026309967041, |
|
"learning_rate": 9.9862507423675e-05, |
|
"loss": 3.7535, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 0.13467524717343268, |
|
"grad_norm": 2.086843490600586, |
|
"learning_rate": 9.98606173528857e-05, |
|
"loss": 4.2659, |
|
"step": 8350 |
|
}, |
|
{ |
|
"epoch": 0.13548168577926162, |
|
"grad_norm": 3.2618448734283447, |
|
"learning_rate": 9.985871439763163e-05, |
|
"loss": 4.0077, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 0.13628812438509055, |
|
"grad_norm": 3.416794538497925, |
|
"learning_rate": 9.985679855840453e-05, |
|
"loss": 4.2068, |
|
"step": 8450 |
|
}, |
|
{ |
|
"epoch": 0.1370945629909195, |
|
"grad_norm": 2.5768439769744873, |
|
"learning_rate": 9.985486983569945e-05, |
|
"loss": 4.2102, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.13790100159674845, |
|
"grad_norm": 2.0248680114746094, |
|
"learning_rate": 9.985296718837846e-05, |
|
"loss": 3.7619, |
|
"step": 8550 |
|
}, |
|
{ |
|
"epoch": 0.1387074402025774, |
|
"grad_norm": 2.308393955230713, |
|
"learning_rate": 9.985101295786062e-05, |
|
"loss": 4.0467, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 0.13951387880840632, |
|
"grad_norm": 3.1832563877105713, |
|
"learning_rate": 9.984904584535987e-05, |
|
"loss": 4.0234, |
|
"step": 8650 |
|
}, |
|
{ |
|
"epoch": 0.14032031741423526, |
|
"grad_norm": 2.4449775218963623, |
|
"learning_rate": 9.984706585138452e-05, |
|
"loss": 3.4602, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 0.1411267560200642, |
|
"grad_norm": 3.088870048522949, |
|
"learning_rate": 9.984507297644625e-05, |
|
"loss": 4.0023, |
|
"step": 8750 |
|
}, |
|
{ |
|
"epoch": 0.14193319462589313, |
|
"grad_norm": 1.3331509828567505, |
|
"learning_rate": 9.984306722106e-05, |
|
"loss": 4.4195, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 0.14273963323172206, |
|
"grad_norm": 2.1085762977600098, |
|
"learning_rate": 9.98410485857441e-05, |
|
"loss": 3.6874, |
|
"step": 8850 |
|
}, |
|
{ |
|
"epoch": 0.143546071837551, |
|
"grad_norm": 2.382269859313965, |
|
"learning_rate": 9.983905782752949e-05, |
|
"loss": 4.1141, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 0.14435251044337993, |
|
"grad_norm": 1.5278171300888062, |
|
"learning_rate": 9.983701369149501e-05, |
|
"loss": 3.8317, |
|
"step": 8950 |
|
}, |
|
{ |
|
"epoch": 0.1451589490492089, |
|
"grad_norm": 5.558732986450195, |
|
"learning_rate": 9.983495667709518e-05, |
|
"loss": 3.6433, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.1451589490492089, |
|
"eval_asr_loss": 0.5558241450032482, |
|
"eval_loss": 5.0716023445129395, |
|
"eval_runtime": 15.4575, |
|
"eval_samples_per_second": 33.123, |
|
"eval_steps_per_second": 11.063, |
|
"eval_tts_loss": 6.0293835261628255, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.14596538765503783, |
|
"grad_norm": 4.300561904907227, |
|
"learning_rate": 9.983288678486153e-05, |
|
"loss": 4.3128, |
|
"step": 9050 |
|
}, |
|
{ |
|
"epoch": 0.14677182626086677, |
|
"grad_norm": 1.5210169553756714, |
|
"learning_rate": 9.983080401532894e-05, |
|
"loss": 3.9063, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 0.1475782648666957, |
|
"grad_norm": 2.6939990520477295, |
|
"learning_rate": 9.982870836903564e-05, |
|
"loss": 3.9087, |
|
"step": 9150 |
|
}, |
|
{ |
|
"epoch": 0.14838470347252464, |
|
"grad_norm": 1.7290055751800537, |
|
"learning_rate": 9.982659984652316e-05, |
|
"loss": 4.1546, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 0.14919114207835357, |
|
"grad_norm": 4.180938243865967, |
|
"learning_rate": 9.982447844833634e-05, |
|
"loss": 4.0063, |
|
"step": 9250 |
|
}, |
|
{ |
|
"epoch": 0.1499975806841825, |
|
"grad_norm": 3.3321378231048584, |
|
"learning_rate": 9.98223441750234e-05, |
|
"loss": 3.6936, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 0.15080401929001144, |
|
"grad_norm": 3.3349199295043945, |
|
"learning_rate": 9.982019702713583e-05, |
|
"loss": 3.9439, |
|
"step": 9350 |
|
}, |
|
{ |
|
"epoch": 0.15161045789584038, |
|
"grad_norm": 2.3645851612091064, |
|
"learning_rate": 9.98180370052285e-05, |
|
"loss": 4.0444, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 0.15241689650166934, |
|
"grad_norm": 3.225592613220215, |
|
"learning_rate": 9.981586410985955e-05, |
|
"loss": 3.8988, |
|
"step": 9450 |
|
}, |
|
{ |
|
"epoch": 0.15322333510749828, |
|
"grad_norm": 3.2834198474884033, |
|
"learning_rate": 9.981367834159049e-05, |
|
"loss": 4.1237, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.1540297737133272, |
|
"grad_norm": 1.3256886005401611, |
|
"learning_rate": 9.981147970098615e-05, |
|
"loss": 3.9196, |
|
"step": 9550 |
|
}, |
|
{ |
|
"epoch": 0.15483621231915615, |
|
"grad_norm": 3.559396743774414, |
|
"learning_rate": 9.980926818861469e-05, |
|
"loss": 4.1997, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 0.15564265092498508, |
|
"grad_norm": 3.722397565841675, |
|
"learning_rate": 9.980704380504756e-05, |
|
"loss": 4.2779, |
|
"step": 9650 |
|
}, |
|
{ |
|
"epoch": 0.15644908953081402, |
|
"grad_norm": 3.125638723373413, |
|
"learning_rate": 9.980480655085957e-05, |
|
"loss": 4.0461, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 0.15725552813664295, |
|
"grad_norm": 1.543898105621338, |
|
"learning_rate": 9.980255642662888e-05, |
|
"loss": 3.9528, |
|
"step": 9750 |
|
}, |
|
{ |
|
"epoch": 0.1580619667424719, |
|
"grad_norm": 3.1968443393707275, |
|
"learning_rate": 9.980029343293688e-05, |
|
"loss": 4.2141, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 0.15886840534830082, |
|
"grad_norm": 1.8708412647247314, |
|
"learning_rate": 9.979801757036841e-05, |
|
"loss": 4.1029, |
|
"step": 9850 |
|
}, |
|
{ |
|
"epoch": 0.15967484395412979, |
|
"grad_norm": 4.089094161987305, |
|
"learning_rate": 9.979572883951156e-05, |
|
"loss": 4.0536, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 0.16048128255995872, |
|
"grad_norm": 1.7748725414276123, |
|
"learning_rate": 9.979342724095774e-05, |
|
"loss": 4.0823, |
|
"step": 9950 |
|
}, |
|
{ |
|
"epoch": 0.16128772116578766, |
|
"grad_norm": 2.352545976638794, |
|
"learning_rate": 9.97911127753017e-05, |
|
"loss": 3.8066, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.1620941597716166, |
|
"grad_norm": 2.2634575366973877, |
|
"learning_rate": 9.978878544314156e-05, |
|
"loss": 4.1763, |
|
"step": 10050 |
|
}, |
|
{ |
|
"epoch": 0.16290059837744553, |
|
"grad_norm": 2.911471366882324, |
|
"learning_rate": 9.978644524507869e-05, |
|
"loss": 4.2928, |
|
"step": 10100 |
|
}, |
|
{ |
|
"epoch": 0.16370703698327446, |
|
"grad_norm": 1.9891984462738037, |
|
"learning_rate": 9.978409218171784e-05, |
|
"loss": 4.0966, |
|
"step": 10150 |
|
}, |
|
{ |
|
"epoch": 0.1645134755891034, |
|
"grad_norm": 2.632568597793579, |
|
"learning_rate": 9.978172625366705e-05, |
|
"loss": 4.2746, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 0.16531991419493233, |
|
"grad_norm": 2.3219494819641113, |
|
"learning_rate": 9.977934746153771e-05, |
|
"loss": 4.3219, |
|
"step": 10250 |
|
}, |
|
{ |
|
"epoch": 0.16612635280076127, |
|
"grad_norm": 2.416696071624756, |
|
"learning_rate": 9.977695580594451e-05, |
|
"loss": 4.0857, |
|
"step": 10300 |
|
}, |
|
{ |
|
"epoch": 0.1669327914065902, |
|
"grad_norm": 1.758725881576538, |
|
"learning_rate": 9.97745512875055e-05, |
|
"loss": 3.7797, |
|
"step": 10350 |
|
}, |
|
{ |
|
"epoch": 0.16773923001241917, |
|
"grad_norm": 2.4825093746185303, |
|
"learning_rate": 9.977213390684199e-05, |
|
"loss": 3.8742, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 0.1685456686182481, |
|
"grad_norm": 1.305153727531433, |
|
"learning_rate": 9.97697036645787e-05, |
|
"loss": 4.1835, |
|
"step": 10450 |
|
}, |
|
{ |
|
"epoch": 0.16935210722407704, |
|
"grad_norm": 4.981587886810303, |
|
"learning_rate": 9.97672605613436e-05, |
|
"loss": 4.2102, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 0.17015854582990597, |
|
"grad_norm": 2.5329973697662354, |
|
"learning_rate": 9.976480459776803e-05, |
|
"loss": 4.1571, |
|
"step": 10550 |
|
}, |
|
{ |
|
"epoch": 0.1709649844357349, |
|
"grad_norm": 3.1804792881011963, |
|
"learning_rate": 9.976233577448661e-05, |
|
"loss": 4.0007, |
|
"step": 10600 |
|
}, |
|
{ |
|
"epoch": 0.17177142304156384, |
|
"grad_norm": 2.1059024333953857, |
|
"learning_rate": 9.975985409213731e-05, |
|
"loss": 4.096, |
|
"step": 10650 |
|
}, |
|
{ |
|
"epoch": 0.17257786164739278, |
|
"grad_norm": 2.35027813911438, |
|
"learning_rate": 9.975735955136145e-05, |
|
"loss": 4.2416, |
|
"step": 10700 |
|
}, |
|
{ |
|
"epoch": 0.1733843002532217, |
|
"grad_norm": 4.606429576873779, |
|
"learning_rate": 9.975485215280362e-05, |
|
"loss": 3.7404, |
|
"step": 10750 |
|
}, |
|
{ |
|
"epoch": 0.17419073885905065, |
|
"grad_norm": 2.3107430934906006, |
|
"learning_rate": 9.975233189711177e-05, |
|
"loss": 3.9696, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 0.1749971774648796, |
|
"grad_norm": 1.1612375974655151, |
|
"learning_rate": 9.974979878493716e-05, |
|
"loss": 4.0279, |
|
"step": 10850 |
|
}, |
|
{ |
|
"epoch": 0.17580361607070855, |
|
"grad_norm": 1.4754469394683838, |
|
"learning_rate": 9.974725281693433e-05, |
|
"loss": 3.5668, |
|
"step": 10900 |
|
}, |
|
{ |
|
"epoch": 0.17661005467653748, |
|
"grad_norm": 1.7385231256484985, |
|
"learning_rate": 9.974469399376123e-05, |
|
"loss": 3.8299, |
|
"step": 10950 |
|
}, |
|
{ |
|
"epoch": 0.17741649328236642, |
|
"grad_norm": 2.5821852684020996, |
|
"learning_rate": 9.974212231607907e-05, |
|
"loss": 3.8388, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.17822293188819535, |
|
"grad_norm": 1.8914719820022583, |
|
"learning_rate": 9.973953778455239e-05, |
|
"loss": 4.0202, |
|
"step": 11050 |
|
}, |
|
{ |
|
"epoch": 0.1790293704940243, |
|
"grad_norm": 1.3121271133422852, |
|
"learning_rate": 9.973694039984907e-05, |
|
"loss": 4.0743, |
|
"step": 11100 |
|
}, |
|
{ |
|
"epoch": 0.17983580909985322, |
|
"grad_norm": 2.854396104812622, |
|
"learning_rate": 9.973433016264029e-05, |
|
"loss": 4.066, |
|
"step": 11150 |
|
}, |
|
{ |
|
"epoch": 0.18064224770568216, |
|
"grad_norm": 3.5787723064422607, |
|
"learning_rate": 9.973170707360055e-05, |
|
"loss": 4.3989, |
|
"step": 11200 |
|
}, |
|
{ |
|
"epoch": 0.1814486863115111, |
|
"grad_norm": 1.8557301759719849, |
|
"learning_rate": 9.972912397814846e-05, |
|
"loss": 4.0768, |
|
"step": 11250 |
|
}, |
|
{ |
|
"epoch": 0.18225512491734006, |
|
"grad_norm": 2.253882884979248, |
|
"learning_rate": 9.97264754444864e-05, |
|
"loss": 3.9537, |
|
"step": 11300 |
|
}, |
|
{ |
|
"epoch": 0.183061563523169, |
|
"grad_norm": 3.240039825439453, |
|
"learning_rate": 9.972381406102311e-05, |
|
"loss": 4.0367, |
|
"step": 11350 |
|
}, |
|
{ |
|
"epoch": 0.18386800212899793, |
|
"grad_norm": 3.579876184463501, |
|
"learning_rate": 9.972113982844634e-05, |
|
"loss": 3.8319, |
|
"step": 11400 |
|
}, |
|
{ |
|
"epoch": 0.18467444073482686, |
|
"grad_norm": 0.6159800291061401, |
|
"learning_rate": 9.971845274744713e-05, |
|
"loss": 3.8889, |
|
"step": 11450 |
|
}, |
|
{ |
|
"epoch": 0.1854808793406558, |
|
"grad_norm": 1.1513911485671997, |
|
"learning_rate": 9.971575281871985e-05, |
|
"loss": 3.8983, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 0.18628731794648473, |
|
"grad_norm": 3.410205125808716, |
|
"learning_rate": 9.971304004296219e-05, |
|
"loss": 4.2003, |
|
"step": 11550 |
|
}, |
|
{ |
|
"epoch": 0.18709375655231367, |
|
"grad_norm": 1.6599464416503906, |
|
"learning_rate": 9.971031442087515e-05, |
|
"loss": 3.7348, |
|
"step": 11600 |
|
}, |
|
{ |
|
"epoch": 0.1879001951581426, |
|
"grad_norm": 2.8847126960754395, |
|
"learning_rate": 9.970757595316306e-05, |
|
"loss": 3.9562, |
|
"step": 11650 |
|
}, |
|
{ |
|
"epoch": 0.18870663376397154, |
|
"grad_norm": 2.053612470626831, |
|
"learning_rate": 9.970482464053359e-05, |
|
"loss": 4.3117, |
|
"step": 11700 |
|
}, |
|
{ |
|
"epoch": 0.1895130723698005, |
|
"grad_norm": 2.1683125495910645, |
|
"learning_rate": 9.970206048369766e-05, |
|
"loss": 3.8281, |
|
"step": 11750 |
|
}, |
|
{ |
|
"epoch": 0.19031951097562944, |
|
"grad_norm": 5.38385009765625, |
|
"learning_rate": 9.969928348336959e-05, |
|
"loss": 4.1646, |
|
"step": 11800 |
|
}, |
|
{ |
|
"epoch": 0.19112594958145837, |
|
"grad_norm": 1.783724069595337, |
|
"learning_rate": 9.969649364026696e-05, |
|
"loss": 3.6177, |
|
"step": 11850 |
|
}, |
|
{ |
|
"epoch": 0.1919323881872873, |
|
"grad_norm": 2.6397554874420166, |
|
"learning_rate": 9.969369095511072e-05, |
|
"loss": 4.1144, |
|
"step": 11900 |
|
}, |
|
{ |
|
"epoch": 0.19273882679311624, |
|
"grad_norm": 3.0491504669189453, |
|
"learning_rate": 9.969087542862509e-05, |
|
"loss": 4.1832, |
|
"step": 11950 |
|
}, |
|
{ |
|
"epoch": 0.19354526539894518, |
|
"grad_norm": 3.2440996170043945, |
|
"learning_rate": 9.968804706153762e-05, |
|
"loss": 3.9472, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.19354526539894518, |
|
"eval_asr_loss": 0.5453353807244307, |
|
"eval_loss": 5.0482306480407715, |
|
"eval_runtime": 15.3432, |
|
"eval_samples_per_second": 33.37, |
|
"eval_steps_per_second": 11.145, |
|
"eval_tts_loss": 6.005206208861255, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.1943517040047741, |
|
"grad_norm": 4.186015605926514, |
|
"learning_rate": 9.968520585457922e-05, |
|
"loss": 3.8557, |
|
"step": 12050 |
|
}, |
|
{ |
|
"epoch": 0.19515814261060305, |
|
"grad_norm": 2.497833013534546, |
|
"learning_rate": 9.968235180848406e-05, |
|
"loss": 4.2129, |
|
"step": 12100 |
|
}, |
|
{ |
|
"epoch": 0.19596458121643198, |
|
"grad_norm": 3.930877685546875, |
|
"learning_rate": 9.967948492398967e-05, |
|
"loss": 4.3072, |
|
"step": 12150 |
|
}, |
|
{ |
|
"epoch": 0.19677101982226092, |
|
"grad_norm": 1.6300374269485474, |
|
"learning_rate": 9.967660520183686e-05, |
|
"loss": 3.6067, |
|
"step": 12200 |
|
}, |
|
{ |
|
"epoch": 0.19757745842808988, |
|
"grad_norm": 1.7998709678649902, |
|
"learning_rate": 9.96737126427698e-05, |
|
"loss": 4.2948, |
|
"step": 12250 |
|
}, |
|
{ |
|
"epoch": 0.19838389703391882, |
|
"grad_norm": 2.9288175106048584, |
|
"learning_rate": 9.967080724753592e-05, |
|
"loss": 3.6403, |
|
"step": 12300 |
|
}, |
|
{ |
|
"epoch": 0.19919033563974775, |
|
"grad_norm": 1.8051952123641968, |
|
"learning_rate": 9.966788901688605e-05, |
|
"loss": 3.9332, |
|
"step": 12350 |
|
}, |
|
{ |
|
"epoch": 0.1999967742455767, |
|
"grad_norm": 2.3929975032806396, |
|
"learning_rate": 9.966495795157427e-05, |
|
"loss": 3.934, |
|
"step": 12400 |
|
}, |
|
{ |
|
"epoch": 0.20080321285140562, |
|
"grad_norm": 2.7234630584716797, |
|
"learning_rate": 9.966201405235797e-05, |
|
"loss": 4.1635, |
|
"step": 12450 |
|
}, |
|
{ |
|
"epoch": 0.20160965145723456, |
|
"grad_norm": 2.5407304763793945, |
|
"learning_rate": 9.965905731999791e-05, |
|
"loss": 3.6558, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 0.2024160900630635, |
|
"grad_norm": 1.8946152925491333, |
|
"learning_rate": 9.965608775525815e-05, |
|
"loss": 3.7618, |
|
"step": 12550 |
|
}, |
|
{ |
|
"epoch": 0.20322252866889243, |
|
"grad_norm": 2.083383798599243, |
|
"learning_rate": 9.965310535890601e-05, |
|
"loss": 4.3816, |
|
"step": 12600 |
|
}, |
|
{ |
|
"epoch": 0.20402896727472136, |
|
"grad_norm": 2.6187212467193604, |
|
"learning_rate": 9.96501101317122e-05, |
|
"loss": 3.9503, |
|
"step": 12650 |
|
}, |
|
{ |
|
"epoch": 0.20483540588055033, |
|
"grad_norm": 4.162103176116943, |
|
"learning_rate": 9.964710207445072e-05, |
|
"loss": 4.0746, |
|
"step": 12700 |
|
}, |
|
{ |
|
"epoch": 0.20564184448637926, |
|
"grad_norm": 2.638073682785034, |
|
"learning_rate": 9.964408118789886e-05, |
|
"loss": 4.3604, |
|
"step": 12750 |
|
}, |
|
{ |
|
"epoch": 0.2064482830922082, |
|
"grad_norm": 2.0001702308654785, |
|
"learning_rate": 9.964104747283727e-05, |
|
"loss": 4.0921, |
|
"step": 12800 |
|
}, |
|
{ |
|
"epoch": 0.20725472169803713, |
|
"grad_norm": 2.3797965049743652, |
|
"learning_rate": 9.963800093004987e-05, |
|
"loss": 4.3402, |
|
"step": 12850 |
|
}, |
|
{ |
|
"epoch": 0.20806116030386607, |
|
"grad_norm": 2.3531606197357178, |
|
"learning_rate": 9.963494156032393e-05, |
|
"loss": 4.0507, |
|
"step": 12900 |
|
}, |
|
{ |
|
"epoch": 0.208867598909695, |
|
"grad_norm": 3.526831865310669, |
|
"learning_rate": 9.963186936445001e-05, |
|
"loss": 4.3212, |
|
"step": 12950 |
|
}, |
|
{ |
|
"epoch": 0.20967403751552394, |
|
"grad_norm": 4.21464729309082, |
|
"learning_rate": 9.962878434322202e-05, |
|
"loss": 4.1086, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 0.21048047612135287, |
|
"grad_norm": 2.4508001804351807, |
|
"learning_rate": 9.962568649743713e-05, |
|
"loss": 3.6248, |
|
"step": 13050 |
|
}, |
|
{ |
|
"epoch": 0.2112869147271818, |
|
"grad_norm": 4.681621551513672, |
|
"learning_rate": 9.962257582789585e-05, |
|
"loss": 3.9501, |
|
"step": 13100 |
|
}, |
|
{ |
|
"epoch": 0.21209335333301077, |
|
"grad_norm": 3.483264207839966, |
|
"learning_rate": 9.961945233540204e-05, |
|
"loss": 3.7879, |
|
"step": 13150 |
|
}, |
|
{ |
|
"epoch": 0.2128997919388397, |
|
"grad_norm": 2.662280321121216, |
|
"learning_rate": 9.961631602076282e-05, |
|
"loss": 4.2567, |
|
"step": 13200 |
|
}, |
|
{ |
|
"epoch": 0.21370623054466864, |
|
"grad_norm": 1.8116915225982666, |
|
"learning_rate": 9.961316688478863e-05, |
|
"loss": 4.1235, |
|
"step": 13250 |
|
}, |
|
{ |
|
"epoch": 0.21451266915049758, |
|
"grad_norm": 3.8289411067962646, |
|
"learning_rate": 9.961000492829326e-05, |
|
"loss": 4.0276, |
|
"step": 13300 |
|
}, |
|
{ |
|
"epoch": 0.2153191077563265, |
|
"grad_norm": 1.5876435041427612, |
|
"learning_rate": 9.96068301520938e-05, |
|
"loss": 3.7603, |
|
"step": 13350 |
|
}, |
|
{ |
|
"epoch": 0.21612554636215545, |
|
"grad_norm": 3.7507176399230957, |
|
"learning_rate": 9.960370643453201e-05, |
|
"loss": 3.6352, |
|
"step": 13400 |
|
}, |
|
{ |
|
"epoch": 0.21693198496798438, |
|
"grad_norm": 1.9685125350952148, |
|
"learning_rate": 9.96005062777419e-05, |
|
"loss": 3.6024, |
|
"step": 13450 |
|
}, |
|
{ |
|
"epoch": 0.21773842357381332, |
|
"grad_norm": 2.623763084411621, |
|
"learning_rate": 9.959729330370225e-05, |
|
"loss": 4.0665, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 0.21854486217964225, |
|
"grad_norm": 1.6213947534561157, |
|
"learning_rate": 9.95940675132433e-05, |
|
"loss": 3.9128, |
|
"step": 13550 |
|
}, |
|
{ |
|
"epoch": 0.21935130078547121, |
|
"grad_norm": 5.413779258728027, |
|
"learning_rate": 9.959082890719862e-05, |
|
"loss": 3.7172, |
|
"step": 13600 |
|
}, |
|
{ |
|
"epoch": 0.22015773939130015, |
|
"grad_norm": 3.002934694290161, |
|
"learning_rate": 9.95875774864051e-05, |
|
"loss": 3.7273, |
|
"step": 13650 |
|
}, |
|
{ |
|
"epoch": 0.22096417799712909, |
|
"grad_norm": 2.6499838829040527, |
|
"learning_rate": 9.958431325170294e-05, |
|
"loss": 4.2387, |
|
"step": 13700 |
|
}, |
|
{ |
|
"epoch": 0.22177061660295802, |
|
"grad_norm": 1.5989998579025269, |
|
"learning_rate": 9.958103620393564e-05, |
|
"loss": 4.0674, |
|
"step": 13750 |
|
}, |
|
{ |
|
"epoch": 0.22257705520878696, |
|
"grad_norm": 2.529733180999756, |
|
"learning_rate": 9.957774634395003e-05, |
|
"loss": 4.1921, |
|
"step": 13800 |
|
}, |
|
{ |
|
"epoch": 0.2233834938146159, |
|
"grad_norm": 6.773332595825195, |
|
"learning_rate": 9.957444367259625e-05, |
|
"loss": 3.947, |
|
"step": 13850 |
|
}, |
|
{ |
|
"epoch": 0.22418993242044483, |
|
"grad_norm": 1.7639790773391724, |
|
"learning_rate": 9.957112819072774e-05, |
|
"loss": 3.9777, |
|
"step": 13900 |
|
}, |
|
{ |
|
"epoch": 0.22499637102627376, |
|
"grad_norm": 1.8540008068084717, |
|
"learning_rate": 9.956779989920124e-05, |
|
"loss": 4.0806, |
|
"step": 13950 |
|
}, |
|
{ |
|
"epoch": 0.2258028096321027, |
|
"grad_norm": 2.861515998840332, |
|
"learning_rate": 9.956445879887683e-05, |
|
"loss": 4.1824, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 0.22660924823793166, |
|
"grad_norm": 2.093233823776245, |
|
"learning_rate": 9.956110489061786e-05, |
|
"loss": 4.0445, |
|
"step": 14050 |
|
}, |
|
{ |
|
"epoch": 0.2274156868437606, |
|
"grad_norm": 2.043487071990967, |
|
"learning_rate": 9.955773817529103e-05, |
|
"loss": 3.8989, |
|
"step": 14100 |
|
}, |
|
{ |
|
"epoch": 0.22822212544958953, |
|
"grad_norm": 1.743752121925354, |
|
"learning_rate": 9.955435865376632e-05, |
|
"loss": 4.2283, |
|
"step": 14150 |
|
}, |
|
{ |
|
"epoch": 0.22902856405541847, |
|
"grad_norm": 1.1774280071258545, |
|
"learning_rate": 9.955096632691704e-05, |
|
"loss": 3.8492, |
|
"step": 14200 |
|
}, |
|
{ |
|
"epoch": 0.2298350026612474, |
|
"grad_norm": 0.7757024168968201, |
|
"learning_rate": 9.95475611956198e-05, |
|
"loss": 3.8286, |
|
"step": 14250 |
|
}, |
|
{ |
|
"epoch": 0.23064144126707634, |
|
"grad_norm": 2.530691623687744, |
|
"learning_rate": 9.954414326075452e-05, |
|
"loss": 4.0583, |
|
"step": 14300 |
|
}, |
|
{ |
|
"epoch": 0.23144787987290527, |
|
"grad_norm": 2.322361469268799, |
|
"learning_rate": 9.95407125232044e-05, |
|
"loss": 4.188, |
|
"step": 14350 |
|
}, |
|
{ |
|
"epoch": 0.2322543184787342, |
|
"grad_norm": 2.6633105278015137, |
|
"learning_rate": 9.953726898385603e-05, |
|
"loss": 3.6949, |
|
"step": 14400 |
|
}, |
|
{ |
|
"epoch": 0.23306075708456314, |
|
"grad_norm": 3.243053436279297, |
|
"learning_rate": 9.95338126435992e-05, |
|
"loss": 3.733, |
|
"step": 14450 |
|
}, |
|
{ |
|
"epoch": 0.23386719569039208, |
|
"grad_norm": 2.6763391494750977, |
|
"learning_rate": 9.953034350332709e-05, |
|
"loss": 3.9605, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 0.23467363429622104, |
|
"grad_norm": 2.8131678104400635, |
|
"learning_rate": 9.952686156393616e-05, |
|
"loss": 4.2417, |
|
"step": 14550 |
|
}, |
|
{ |
|
"epoch": 0.23548007290204997, |
|
"grad_norm": 2.5453968048095703, |
|
"learning_rate": 9.952336682632617e-05, |
|
"loss": 4.5493, |
|
"step": 14600 |
|
}, |
|
{ |
|
"epoch": 0.2362865115078789, |
|
"grad_norm": 2.9333486557006836, |
|
"learning_rate": 9.951985929140018e-05, |
|
"loss": 4.0338, |
|
"step": 14650 |
|
}, |
|
{ |
|
"epoch": 0.23709295011370785, |
|
"grad_norm": 1.5506014823913574, |
|
"learning_rate": 9.951633896006461e-05, |
|
"loss": 3.9207, |
|
"step": 14700 |
|
}, |
|
{ |
|
"epoch": 0.23789938871953678, |
|
"grad_norm": 1.8103605508804321, |
|
"learning_rate": 9.95128058332291e-05, |
|
"loss": 3.9589, |
|
"step": 14750 |
|
}, |
|
{ |
|
"epoch": 0.23870582732536572, |
|
"grad_norm": 2.344365358352661, |
|
"learning_rate": 9.950925991180669e-05, |
|
"loss": 4.0624, |
|
"step": 14800 |
|
}, |
|
{ |
|
"epoch": 0.23951226593119465, |
|
"grad_norm": 2.332612991333008, |
|
"learning_rate": 9.950570119671362e-05, |
|
"loss": 3.9549, |
|
"step": 14850 |
|
}, |
|
{ |
|
"epoch": 0.2403187045370236, |
|
"grad_norm": 2.706002712249756, |
|
"learning_rate": 9.950212968886956e-05, |
|
"loss": 3.6544, |
|
"step": 14900 |
|
}, |
|
{ |
|
"epoch": 0.24112514314285252, |
|
"grad_norm": 2.335348606109619, |
|
"learning_rate": 9.94985453891974e-05, |
|
"loss": 3.9361, |
|
"step": 14950 |
|
}, |
|
{ |
|
"epoch": 0.24193158174868148, |
|
"grad_norm": 1.8754454851150513, |
|
"learning_rate": 9.949494829862334e-05, |
|
"loss": 3.6963, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 0.24193158174868148, |
|
"eval_asr_loss": 0.560360777742155, |
|
"eval_loss": 5.027297496795654, |
|
"eval_runtime": 15.4664, |
|
"eval_samples_per_second": 33.104, |
|
"eval_steps_per_second": 11.056, |
|
"eval_tts_loss": 5.95218636007564, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 0.24273802035451042, |
|
"grad_norm": 1.9016485214233398, |
|
"learning_rate": 9.949133841807691e-05, |
|
"loss": 3.9233, |
|
"step": 15050 |
|
}, |
|
{ |
|
"epoch": 0.24354445896033936, |
|
"grad_norm": 2.494946002960205, |
|
"learning_rate": 9.948771574849097e-05, |
|
"loss": 4.087, |
|
"step": 15100 |
|
}, |
|
{ |
|
"epoch": 0.2443508975661683, |
|
"grad_norm": 3.006345748901367, |
|
"learning_rate": 9.948408029080161e-05, |
|
"loss": 4.0706, |
|
"step": 15150 |
|
}, |
|
{ |
|
"epoch": 0.24515733617199723, |
|
"grad_norm": 3.3929615020751953, |
|
"learning_rate": 9.94804320459483e-05, |
|
"loss": 3.9753, |
|
"step": 15200 |
|
}, |
|
{ |
|
"epoch": 0.24596377477782616, |
|
"grad_norm": 1.2217860221862793, |
|
"learning_rate": 9.947677101487374e-05, |
|
"loss": 4.1967, |
|
"step": 15250 |
|
}, |
|
{ |
|
"epoch": 0.2467702133836551, |
|
"grad_norm": 1.409860610961914, |
|
"learning_rate": 9.947309719852402e-05, |
|
"loss": 3.9026, |
|
"step": 15300 |
|
}, |
|
{ |
|
"epoch": 0.24757665198948403, |
|
"grad_norm": 1.3178203105926514, |
|
"learning_rate": 9.946941059784847e-05, |
|
"loss": 4.1695, |
|
"step": 15350 |
|
}, |
|
{ |
|
"epoch": 0.24838309059531297, |
|
"grad_norm": 3.196470260620117, |
|
"learning_rate": 9.946571121379974e-05, |
|
"loss": 3.8572, |
|
"step": 15400 |
|
}, |
|
{ |
|
"epoch": 0.24918952920114193, |
|
"grad_norm": 2.9058384895324707, |
|
"learning_rate": 9.946199904733378e-05, |
|
"loss": 4.2121, |
|
"step": 15450 |
|
}, |
|
{ |
|
"epoch": 0.24999596780697086, |
|
"grad_norm": 2.940702438354492, |
|
"learning_rate": 9.945827409940989e-05, |
|
"loss": 4.2432, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 0.25080240641279977, |
|
"grad_norm": 3.044480562210083, |
|
"learning_rate": 9.945453637099058e-05, |
|
"loss": 4.0084, |
|
"step": 15550 |
|
}, |
|
{ |
|
"epoch": 0.25160884501862874, |
|
"grad_norm": 2.350438356399536, |
|
"learning_rate": 9.945078586304175e-05, |
|
"loss": 4.1791, |
|
"step": 15600 |
|
}, |
|
{ |
|
"epoch": 0.25241528362445764, |
|
"grad_norm": 2.298433542251587, |
|
"learning_rate": 9.944702257653255e-05, |
|
"loss": 3.9251, |
|
"step": 15650 |
|
}, |
|
{ |
|
"epoch": 0.2532217222302866, |
|
"grad_norm": 2.5852606296539307, |
|
"learning_rate": 9.944324651243545e-05, |
|
"loss": 3.9519, |
|
"step": 15700 |
|
}, |
|
{ |
|
"epoch": 0.25402816083611557, |
|
"grad_norm": 1.654619812965393, |
|
"learning_rate": 9.943945767172624e-05, |
|
"loss": 3.7628, |
|
"step": 15750 |
|
}, |
|
{ |
|
"epoch": 0.2548345994419445, |
|
"grad_norm": 2.320345878601074, |
|
"learning_rate": 9.943573221290566e-05, |
|
"loss": 4.3502, |
|
"step": 15800 |
|
}, |
|
{ |
|
"epoch": 0.25564103804777344, |
|
"grad_norm": 1.5587137937545776, |
|
"learning_rate": 9.943191807739609e-05, |
|
"loss": 4.3284, |
|
"step": 15850 |
|
}, |
|
{ |
|
"epoch": 0.25644747665360235, |
|
"grad_norm": 2.629159927368164, |
|
"learning_rate": 9.942809116820176e-05, |
|
"loss": 3.9416, |
|
"step": 15900 |
|
}, |
|
{ |
|
"epoch": 0.2572539152594313, |
|
"grad_norm": 2.840963363647461, |
|
"learning_rate": 9.942425148631161e-05, |
|
"loss": 4.0621, |
|
"step": 15950 |
|
}, |
|
{ |
|
"epoch": 0.2580603538652602, |
|
"grad_norm": 1.6260427236557007, |
|
"learning_rate": 9.942039903271782e-05, |
|
"loss": 3.7273, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 0.2588667924710892, |
|
"grad_norm": 1.4383156299591064, |
|
"learning_rate": 9.941653380841591e-05, |
|
"loss": 4.262, |
|
"step": 16050 |
|
}, |
|
{ |
|
"epoch": 0.2596732310769181, |
|
"grad_norm": 1.3447210788726807, |
|
"learning_rate": 9.941265581440471e-05, |
|
"loss": 3.6917, |
|
"step": 16100 |
|
}, |
|
{ |
|
"epoch": 0.26047966968274705, |
|
"grad_norm": 1.707841157913208, |
|
"learning_rate": 9.940876505168633e-05, |
|
"loss": 3.7174, |
|
"step": 16150 |
|
}, |
|
{ |
|
"epoch": 0.261286108288576, |
|
"grad_norm": 2.2020676136016846, |
|
"learning_rate": 9.940486152126614e-05, |
|
"loss": 4.2535, |
|
"step": 16200 |
|
}, |
|
{ |
|
"epoch": 0.2620925468944049, |
|
"grad_norm": 1.9968034029006958, |
|
"learning_rate": 9.94009452241529e-05, |
|
"loss": 4.2131, |
|
"step": 16250 |
|
}, |
|
{ |
|
"epoch": 0.2628989855002339, |
|
"grad_norm": 4.067761421203613, |
|
"learning_rate": 9.939701616135858e-05, |
|
"loss": 4.0394, |
|
"step": 16300 |
|
}, |
|
{ |
|
"epoch": 0.2637054241060628, |
|
"grad_norm": 1.6382523775100708, |
|
"learning_rate": 9.939307433389852e-05, |
|
"loss": 3.7792, |
|
"step": 16350 |
|
}, |
|
{ |
|
"epoch": 0.26451186271189175, |
|
"grad_norm": 1.519000768661499, |
|
"learning_rate": 9.938911974279132e-05, |
|
"loss": 3.7692, |
|
"step": 16400 |
|
}, |
|
{ |
|
"epoch": 0.26531830131772066, |
|
"grad_norm": 2.7159464359283447, |
|
"learning_rate": 9.938515238905886e-05, |
|
"loss": 3.8706, |
|
"step": 16450 |
|
}, |
|
{ |
|
"epoch": 0.2661247399235496, |
|
"grad_norm": 4.7422871589660645, |
|
"learning_rate": 9.938117227372636e-05, |
|
"loss": 4.1009, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 0.26693117852937853, |
|
"grad_norm": 2.702423095703125, |
|
"learning_rate": 9.937717939782233e-05, |
|
"loss": 3.7776, |
|
"step": 16550 |
|
}, |
|
{ |
|
"epoch": 0.2677376171352075, |
|
"grad_norm": 3.021857738494873, |
|
"learning_rate": 9.937317376237856e-05, |
|
"loss": 4.1462, |
|
"step": 16600 |
|
}, |
|
{ |
|
"epoch": 0.26854405574103646, |
|
"grad_norm": 1.813398003578186, |
|
"learning_rate": 9.936915536843012e-05, |
|
"loss": 4.0327, |
|
"step": 16650 |
|
}, |
|
{ |
|
"epoch": 0.26935049434686537, |
|
"grad_norm": 2.897822856903076, |
|
"learning_rate": 9.936512421701545e-05, |
|
"loss": 3.9975, |
|
"step": 16700 |
|
}, |
|
{ |
|
"epoch": 0.27015693295269433, |
|
"grad_norm": 5.469192028045654, |
|
"learning_rate": 9.936108030917619e-05, |
|
"loss": 3.9002, |
|
"step": 16750 |
|
}, |
|
{ |
|
"epoch": 0.27096337155852324, |
|
"grad_norm": 1.7145867347717285, |
|
"learning_rate": 9.935702364595736e-05, |
|
"loss": 3.711, |
|
"step": 16800 |
|
}, |
|
{ |
|
"epoch": 0.2717698101643522, |
|
"grad_norm": 2.104463577270508, |
|
"learning_rate": 9.935295422840722e-05, |
|
"loss": 4.1141, |
|
"step": 16850 |
|
}, |
|
{ |
|
"epoch": 0.2725762487701811, |
|
"grad_norm": 3.0537919998168945, |
|
"learning_rate": 9.934887205757735e-05, |
|
"loss": 3.7435, |
|
"step": 16900 |
|
}, |
|
{ |
|
"epoch": 0.27338268737601007, |
|
"grad_norm": 1.5938163995742798, |
|
"learning_rate": 9.934477713452264e-05, |
|
"loss": 4.2125, |
|
"step": 16950 |
|
}, |
|
{ |
|
"epoch": 0.274189125981839, |
|
"grad_norm": 1.9021580219268799, |
|
"learning_rate": 9.934066946030124e-05, |
|
"loss": 3.6318, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 0.27499556458766794, |
|
"grad_norm": 4.621829986572266, |
|
"learning_rate": 9.933654903597461e-05, |
|
"loss": 3.6992, |
|
"step": 17050 |
|
}, |
|
{ |
|
"epoch": 0.2758020031934969, |
|
"grad_norm": 2.142982244491577, |
|
"learning_rate": 9.933241586260754e-05, |
|
"loss": 3.7732, |
|
"step": 17100 |
|
}, |
|
{ |
|
"epoch": 0.2766084417993258, |
|
"grad_norm": 3.7406392097473145, |
|
"learning_rate": 9.932826994126802e-05, |
|
"loss": 4.1151, |
|
"step": 17150 |
|
}, |
|
{ |
|
"epoch": 0.2774148804051548, |
|
"grad_norm": 0.5368770360946655, |
|
"learning_rate": 9.932411127302745e-05, |
|
"loss": 3.7562, |
|
"step": 17200 |
|
}, |
|
{ |
|
"epoch": 0.2782213190109837, |
|
"grad_norm": 0.6764758825302124, |
|
"learning_rate": 9.931993985896045e-05, |
|
"loss": 3.8355, |
|
"step": 17250 |
|
}, |
|
{ |
|
"epoch": 0.27902775761681264, |
|
"grad_norm": 1.2389382123947144, |
|
"learning_rate": 9.931575570014495e-05, |
|
"loss": 3.9448, |
|
"step": 17300 |
|
}, |
|
{ |
|
"epoch": 0.27983419622264155, |
|
"grad_norm": 2.505661964416504, |
|
"learning_rate": 9.931155879766221e-05, |
|
"loss": 3.8676, |
|
"step": 17350 |
|
}, |
|
{ |
|
"epoch": 0.2806406348284705, |
|
"grad_norm": 1.7408605813980103, |
|
"learning_rate": 9.930734915259669e-05, |
|
"loss": 3.8457, |
|
"step": 17400 |
|
}, |
|
{ |
|
"epoch": 0.2814470734342994, |
|
"grad_norm": 1.7804434299468994, |
|
"learning_rate": 9.930312676603626e-05, |
|
"loss": 3.8563, |
|
"step": 17450 |
|
}, |
|
{ |
|
"epoch": 0.2822535120401284, |
|
"grad_norm": 2.6170337200164795, |
|
"learning_rate": 9.9298891639072e-05, |
|
"loss": 4.1211, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 0.28305995064595735, |
|
"grad_norm": 2.365858316421509, |
|
"learning_rate": 9.92946437727983e-05, |
|
"loss": 4.1625, |
|
"step": 17550 |
|
}, |
|
{ |
|
"epoch": 0.28386638925178626, |
|
"grad_norm": 2.3039627075195312, |
|
"learning_rate": 9.929038316831285e-05, |
|
"loss": 3.9853, |
|
"step": 17600 |
|
}, |
|
{ |
|
"epoch": 0.2846728278576152, |
|
"grad_norm": 2.711247444152832, |
|
"learning_rate": 9.928610982671665e-05, |
|
"loss": 4.1158, |
|
"step": 17650 |
|
}, |
|
{ |
|
"epoch": 0.2854792664634441, |
|
"grad_norm": 2.1963279247283936, |
|
"learning_rate": 9.928182374911396e-05, |
|
"loss": 4.009, |
|
"step": 17700 |
|
}, |
|
{ |
|
"epoch": 0.2862857050692731, |
|
"grad_norm": 2.6228134632110596, |
|
"learning_rate": 9.927752493661236e-05, |
|
"loss": 4.0115, |
|
"step": 17750 |
|
}, |
|
{ |
|
"epoch": 0.287092143675102, |
|
"grad_norm": 2.0704829692840576, |
|
"learning_rate": 9.927321339032267e-05, |
|
"loss": 4.5117, |
|
"step": 17800 |
|
}, |
|
{ |
|
"epoch": 0.28789858228093096, |
|
"grad_norm": 2.092968225479126, |
|
"learning_rate": 9.926888911135908e-05, |
|
"loss": 4.3974, |
|
"step": 17850 |
|
}, |
|
{ |
|
"epoch": 0.28870502088675987, |
|
"grad_norm": 2.2499780654907227, |
|
"learning_rate": 9.9264552100839e-05, |
|
"loss": 3.8743, |
|
"step": 17900 |
|
}, |
|
{ |
|
"epoch": 0.28951145949258883, |
|
"grad_norm": 1.970129132270813, |
|
"learning_rate": 9.926020235988313e-05, |
|
"loss": 4.0165, |
|
"step": 17950 |
|
}, |
|
{ |
|
"epoch": 0.2903178980984178, |
|
"grad_norm": 1.648297905921936, |
|
"learning_rate": 9.925583988961555e-05, |
|
"loss": 4.1248, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 0.2903178980984178, |
|
"eval_asr_loss": 0.5475505368937239, |
|
"eval_loss": 5.026027679443359, |
|
"eval_runtime": 17.7593, |
|
"eval_samples_per_second": 28.83, |
|
"eval_steps_per_second": 9.629, |
|
"eval_tts_loss": 5.953106430704743, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 0.2911243367042467, |
|
"grad_norm": 2.2440879344940186, |
|
"learning_rate": 9.925155231986144e-05, |
|
"loss": 4.0456, |
|
"step": 18050 |
|
}, |
|
{ |
|
"epoch": 0.29193077531007566, |
|
"grad_norm": 4.98739767074585, |
|
"learning_rate": 9.924716464888553e-05, |
|
"loss": 3.9588, |
|
"step": 18100 |
|
}, |
|
{ |
|
"epoch": 0.29273721391590457, |
|
"grad_norm": 1.8671597242355347, |
|
"learning_rate": 9.924276425196694e-05, |
|
"loss": 4.111, |
|
"step": 18150 |
|
}, |
|
{ |
|
"epoch": 0.29354365252173353, |
|
"grad_norm": 2.135396718978882, |
|
"learning_rate": 9.923835113024276e-05, |
|
"loss": 3.8483, |
|
"step": 18200 |
|
}, |
|
{ |
|
"epoch": 0.29435009112756244, |
|
"grad_norm": 2.5975193977355957, |
|
"learning_rate": 9.923392528485342e-05, |
|
"loss": 4.1396, |
|
"step": 18250 |
|
}, |
|
{ |
|
"epoch": 0.2951565297333914, |
|
"grad_norm": 2.3712480068206787, |
|
"learning_rate": 9.922948671694258e-05, |
|
"loss": 4.3892, |
|
"step": 18300 |
|
}, |
|
{ |
|
"epoch": 0.2959629683392203, |
|
"grad_norm": 1.9007618427276611, |
|
"learning_rate": 9.922503542765723e-05, |
|
"loss": 4.0316, |
|
"step": 18350 |
|
}, |
|
{ |
|
"epoch": 0.2967694069450493, |
|
"grad_norm": 2.405129909515381, |
|
"learning_rate": 9.922057141814758e-05, |
|
"loss": 4.2691, |
|
"step": 18400 |
|
}, |
|
{ |
|
"epoch": 0.29757584555087824, |
|
"grad_norm": 1.4283597469329834, |
|
"learning_rate": 9.92160946895672e-05, |
|
"loss": 3.8773, |
|
"step": 18450 |
|
}, |
|
{ |
|
"epoch": 0.29838228415670714, |
|
"grad_norm": 1.7647777795791626, |
|
"learning_rate": 9.921160524307294e-05, |
|
"loss": 4.2706, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 0.2991887227625361, |
|
"grad_norm": 2.2492105960845947, |
|
"learning_rate": 9.920710307982488e-05, |
|
"loss": 4.2848, |
|
"step": 18550 |
|
}, |
|
{ |
|
"epoch": 0.299995161368365, |
|
"grad_norm": 1.8520779609680176, |
|
"learning_rate": 9.920258820098644e-05, |
|
"loss": 3.8957, |
|
"step": 18600 |
|
}, |
|
{ |
|
"epoch": 0.300801599974194, |
|
"grad_norm": 1.574752688407898, |
|
"learning_rate": 9.91980606077243e-05, |
|
"loss": 3.96, |
|
"step": 18650 |
|
}, |
|
{ |
|
"epoch": 0.3016080385800229, |
|
"grad_norm": 3.233193874359131, |
|
"learning_rate": 9.919352030120846e-05, |
|
"loss": 3.9196, |
|
"step": 18700 |
|
}, |
|
{ |
|
"epoch": 0.30241447718585185, |
|
"grad_norm": 1.642980694770813, |
|
"learning_rate": 9.918896728261216e-05, |
|
"loss": 3.9637, |
|
"step": 18750 |
|
}, |
|
{ |
|
"epoch": 0.30322091579168076, |
|
"grad_norm": 3.266005516052246, |
|
"learning_rate": 9.918440155311193e-05, |
|
"loss": 3.905, |
|
"step": 18800 |
|
}, |
|
{ |
|
"epoch": 0.3040273543975097, |
|
"grad_norm": 2.4287779331207275, |
|
"learning_rate": 9.917982311388763e-05, |
|
"loss": 3.7187, |
|
"step": 18850 |
|
}, |
|
{ |
|
"epoch": 0.3048337930033387, |
|
"grad_norm": 2.907322645187378, |
|
"learning_rate": 9.917523196612233e-05, |
|
"loss": 4.291, |
|
"step": 18900 |
|
}, |
|
{ |
|
"epoch": 0.3056402316091676, |
|
"grad_norm": 2.1453018188476562, |
|
"learning_rate": 9.917062811100247e-05, |
|
"loss": 3.9087, |
|
"step": 18950 |
|
}, |
|
{ |
|
"epoch": 0.30644667021499655, |
|
"grad_norm": 2.715480327606201, |
|
"learning_rate": 9.916601154971772e-05, |
|
"loss": 4.1218, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 0.30725310882082546, |
|
"grad_norm": 2.193371295928955, |
|
"learning_rate": 9.916138228346101e-05, |
|
"loss": 4.3158, |
|
"step": 19050 |
|
}, |
|
{ |
|
"epoch": 0.3080595474266544, |
|
"grad_norm": 1.1654584407806396, |
|
"learning_rate": 9.915674031342864e-05, |
|
"loss": 3.9391, |
|
"step": 19100 |
|
}, |
|
{ |
|
"epoch": 0.30886598603248333, |
|
"grad_norm": 2.589933156967163, |
|
"learning_rate": 9.915208564082011e-05, |
|
"loss": 3.7376, |
|
"step": 19150 |
|
}, |
|
{ |
|
"epoch": 0.3096724246383123, |
|
"grad_norm": 1.8773146867752075, |
|
"learning_rate": 9.914741826683822e-05, |
|
"loss": 3.8819, |
|
"step": 19200 |
|
}, |
|
{ |
|
"epoch": 0.3104788632441412, |
|
"grad_norm": 3.009824752807617, |
|
"learning_rate": 9.91427381926891e-05, |
|
"loss": 3.7509, |
|
"step": 19250 |
|
}, |
|
{ |
|
"epoch": 0.31128530184997016, |
|
"grad_norm": 1.649009108543396, |
|
"learning_rate": 9.913804541958209e-05, |
|
"loss": 4.3104, |
|
"step": 19300 |
|
}, |
|
{ |
|
"epoch": 0.3120917404557991, |
|
"grad_norm": 2.454561710357666, |
|
"learning_rate": 9.913333994872989e-05, |
|
"loss": 3.6798, |
|
"step": 19350 |
|
}, |
|
{ |
|
"epoch": 0.31289817906162803, |
|
"grad_norm": 2.0216288566589355, |
|
"learning_rate": 9.912862178134839e-05, |
|
"loss": 4.0289, |
|
"step": 19400 |
|
}, |
|
{ |
|
"epoch": 0.313704617667457, |
|
"grad_norm": 2.3107807636260986, |
|
"learning_rate": 9.912389091865683e-05, |
|
"loss": 3.766, |
|
"step": 19450 |
|
}, |
|
{ |
|
"epoch": 0.3145110562732859, |
|
"grad_norm": 2.5804734230041504, |
|
"learning_rate": 9.911914736187774e-05, |
|
"loss": 4.397, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 0.31531749487911487, |
|
"grad_norm": 2.3121676445007324, |
|
"learning_rate": 9.911439111223686e-05, |
|
"loss": 4.0335, |
|
"step": 19550 |
|
}, |
|
{ |
|
"epoch": 0.3161239334849438, |
|
"grad_norm": 5.4831342697143555, |
|
"learning_rate": 9.910962217096327e-05, |
|
"loss": 3.9674, |
|
"step": 19600 |
|
}, |
|
{ |
|
"epoch": 0.31693037209077274, |
|
"grad_norm": 1.5503877401351929, |
|
"learning_rate": 9.91048405392893e-05, |
|
"loss": 4.2763, |
|
"step": 19650 |
|
}, |
|
{ |
|
"epoch": 0.31773681069660165, |
|
"grad_norm": 3.0469183921813965, |
|
"learning_rate": 9.91000462184506e-05, |
|
"loss": 4.4284, |
|
"step": 19700 |
|
}, |
|
{ |
|
"epoch": 0.3185432493024306, |
|
"grad_norm": 2.0985143184661865, |
|
"learning_rate": 9.909523920968602e-05, |
|
"loss": 4.1063, |
|
"step": 19750 |
|
}, |
|
{ |
|
"epoch": 0.31934968790825957, |
|
"grad_norm": 2.2062013149261475, |
|
"learning_rate": 9.909041951423776e-05, |
|
"loss": 4.33, |
|
"step": 19800 |
|
}, |
|
{ |
|
"epoch": 0.3201561265140885, |
|
"grad_norm": 1.0260072946548462, |
|
"learning_rate": 9.90855871333513e-05, |
|
"loss": 3.8472, |
|
"step": 19850 |
|
}, |
|
{ |
|
"epoch": 0.32096256511991744, |
|
"grad_norm": 1.4739959239959717, |
|
"learning_rate": 9.908074206827534e-05, |
|
"loss": 3.6729, |
|
"step": 19900 |
|
}, |
|
{ |
|
"epoch": 0.32176900372574635, |
|
"grad_norm": 1.3404515981674194, |
|
"learning_rate": 9.90758843202619e-05, |
|
"loss": 3.7977, |
|
"step": 19950 |
|
}, |
|
{ |
|
"epoch": 0.3225754423315753, |
|
"grad_norm": 2.4651851654052734, |
|
"learning_rate": 9.907101389056628e-05, |
|
"loss": 3.7939, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 0.3233818809374042, |
|
"grad_norm": 4.615785121917725, |
|
"learning_rate": 9.906613078044704e-05, |
|
"loss": 4.1735, |
|
"step": 20050 |
|
}, |
|
{ |
|
"epoch": 0.3241883195432332, |
|
"grad_norm": 2.872647523880005, |
|
"learning_rate": 9.906133303119925e-05, |
|
"loss": 4.0749, |
|
"step": 20100 |
|
}, |
|
{ |
|
"epoch": 0.3249947581490621, |
|
"grad_norm": 1.8358525037765503, |
|
"learning_rate": 9.905642481756708e-05, |
|
"loss": 4.1521, |
|
"step": 20150 |
|
}, |
|
{ |
|
"epoch": 0.32580119675489105, |
|
"grad_norm": 1.2282460927963257, |
|
"learning_rate": 9.905150392728124e-05, |
|
"loss": 4.2458, |
|
"step": 20200 |
|
}, |
|
{ |
|
"epoch": 0.32660763536071996, |
|
"grad_norm": 5.087796211242676, |
|
"learning_rate": 9.904657036161335e-05, |
|
"loss": 4.0652, |
|
"step": 20250 |
|
}, |
|
{ |
|
"epoch": 0.3274140739665489, |
|
"grad_norm": 1.9742780923843384, |
|
"learning_rate": 9.904162412183829e-05, |
|
"loss": 3.8384, |
|
"step": 20300 |
|
}, |
|
{ |
|
"epoch": 0.3282205125723779, |
|
"grad_norm": 2.7985477447509766, |
|
"learning_rate": 9.90366652092342e-05, |
|
"loss": 3.8642, |
|
"step": 20350 |
|
}, |
|
{ |
|
"epoch": 0.3290269511782068, |
|
"grad_norm": 1.4354287385940552, |
|
"learning_rate": 9.903169362508254e-05, |
|
"loss": 3.9342, |
|
"step": 20400 |
|
}, |
|
{ |
|
"epoch": 0.32983338978403576, |
|
"grad_norm": 2.1292335987091064, |
|
"learning_rate": 9.902670937066798e-05, |
|
"loss": 3.6908, |
|
"step": 20450 |
|
}, |
|
{ |
|
"epoch": 0.33063982838986467, |
|
"grad_norm": 2.5627033710479736, |
|
"learning_rate": 9.902171244727851e-05, |
|
"loss": 3.8676, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 0.33144626699569363, |
|
"grad_norm": 1.1219414472579956, |
|
"learning_rate": 9.90167028562054e-05, |
|
"loss": 4.0734, |
|
"step": 20550 |
|
}, |
|
{ |
|
"epoch": 0.33225270560152254, |
|
"grad_norm": 1.570227026939392, |
|
"learning_rate": 9.901168059874315e-05, |
|
"loss": 3.7047, |
|
"step": 20600 |
|
}, |
|
{ |
|
"epoch": 0.3330591442073515, |
|
"grad_norm": 2.0881588459014893, |
|
"learning_rate": 9.900664567618958e-05, |
|
"loss": 4.1496, |
|
"step": 20650 |
|
}, |
|
{ |
|
"epoch": 0.3338655828131804, |
|
"grad_norm": 2.787027359008789, |
|
"learning_rate": 9.900159808984575e-05, |
|
"loss": 4.0835, |
|
"step": 20700 |
|
}, |
|
{ |
|
"epoch": 0.33467202141900937, |
|
"grad_norm": 4.925529956817627, |
|
"learning_rate": 9.899653784101603e-05, |
|
"loss": 3.961, |
|
"step": 20750 |
|
}, |
|
{ |
|
"epoch": 0.33547846002483833, |
|
"grad_norm": 2.3205068111419678, |
|
"learning_rate": 9.899146493100799e-05, |
|
"loss": 3.7549, |
|
"step": 20800 |
|
}, |
|
{ |
|
"epoch": 0.33628489863066724, |
|
"grad_norm": 1.6399608850479126, |
|
"learning_rate": 9.898637936113257e-05, |
|
"loss": 4.1994, |
|
"step": 20850 |
|
}, |
|
{ |
|
"epoch": 0.3370913372364962, |
|
"grad_norm": 2.315507411956787, |
|
"learning_rate": 9.89812811327039e-05, |
|
"loss": 4.0396, |
|
"step": 20900 |
|
}, |
|
{ |
|
"epoch": 0.3378977758423251, |
|
"grad_norm": 2.239157199859619, |
|
"learning_rate": 9.897617024703943e-05, |
|
"loss": 4.3633, |
|
"step": 20950 |
|
}, |
|
{ |
|
"epoch": 0.3387042144481541, |
|
"grad_norm": 2.6427533626556396, |
|
"learning_rate": 9.897104670545985e-05, |
|
"loss": 3.872, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 0.3387042144481541, |
|
"eval_asr_loss": 0.5612188019541791, |
|
"eval_loss": 4.999529838562012, |
|
"eval_runtime": 15.2573, |
|
"eval_samples_per_second": 33.558, |
|
"eval_steps_per_second": 11.208, |
|
"eval_tts_loss": 6.0108543639534995, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 0.339510653053983, |
|
"grad_norm": 1.9391437768936157, |
|
"learning_rate": 9.896591050928914e-05, |
|
"loss": 3.9338, |
|
"step": 21050 |
|
}, |
|
{ |
|
"epoch": 0.34031709165981194, |
|
"grad_norm": 2.6968955993652344, |
|
"learning_rate": 9.896076165985453e-05, |
|
"loss": 4.0754, |
|
"step": 21100 |
|
}, |
|
{ |
|
"epoch": 0.34112353026564085, |
|
"grad_norm": 1.984168529510498, |
|
"learning_rate": 9.895560015848652e-05, |
|
"loss": 3.9749, |
|
"step": 21150 |
|
}, |
|
{ |
|
"epoch": 0.3419299688714698, |
|
"grad_norm": 3.1954638957977295, |
|
"learning_rate": 9.895042600651894e-05, |
|
"loss": 4.1341, |
|
"step": 21200 |
|
}, |
|
{ |
|
"epoch": 0.3427364074772988, |
|
"grad_norm": 3.572061061859131, |
|
"learning_rate": 9.894523920528879e-05, |
|
"loss": 3.8629, |
|
"step": 21250 |
|
}, |
|
{ |
|
"epoch": 0.3435428460831277, |
|
"grad_norm": 1.6200737953186035, |
|
"learning_rate": 9.894003975613644e-05, |
|
"loss": 3.9207, |
|
"step": 21300 |
|
}, |
|
{ |
|
"epoch": 0.34434928468895665, |
|
"grad_norm": 1.8774925470352173, |
|
"learning_rate": 9.893482766040543e-05, |
|
"loss": 3.6651, |
|
"step": 21350 |
|
}, |
|
{ |
|
"epoch": 0.34515572329478555, |
|
"grad_norm": 3.098095655441284, |
|
"learning_rate": 9.892960291944263e-05, |
|
"loss": 4.1371, |
|
"step": 21400 |
|
}, |
|
{ |
|
"epoch": 0.3459621619006145, |
|
"grad_norm": 1.517335295677185, |
|
"learning_rate": 9.892436553459819e-05, |
|
"loss": 4.2014, |
|
"step": 21450 |
|
}, |
|
{ |
|
"epoch": 0.3467686005064434, |
|
"grad_norm": 2.045222520828247, |
|
"learning_rate": 9.891911550722545e-05, |
|
"loss": 4.12, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 0.3475750391122724, |
|
"grad_norm": 2.005338430404663, |
|
"learning_rate": 9.891385283868112e-05, |
|
"loss": 4.0222, |
|
"step": 21550 |
|
}, |
|
{ |
|
"epoch": 0.3483814777181013, |
|
"grad_norm": 2.2225544452667236, |
|
"learning_rate": 9.89085775303251e-05, |
|
"loss": 4.4284, |
|
"step": 21600 |
|
}, |
|
{ |
|
"epoch": 0.34918791632393026, |
|
"grad_norm": 1.5066674947738647, |
|
"learning_rate": 9.890328958352057e-05, |
|
"loss": 4.042, |
|
"step": 21650 |
|
}, |
|
{ |
|
"epoch": 0.3499943549297592, |
|
"grad_norm": 3.724332332611084, |
|
"learning_rate": 9.8897988999634e-05, |
|
"loss": 4.2711, |
|
"step": 21700 |
|
}, |
|
{ |
|
"epoch": 0.35080079353558813, |
|
"grad_norm": 1.9594568014144897, |
|
"learning_rate": 9.889267578003514e-05, |
|
"loss": 3.8673, |
|
"step": 21750 |
|
}, |
|
{ |
|
"epoch": 0.3516072321414171, |
|
"grad_norm": 1.407718300819397, |
|
"learning_rate": 9.888734992609692e-05, |
|
"loss": 3.965, |
|
"step": 21800 |
|
}, |
|
{ |
|
"epoch": 0.352413670747246, |
|
"grad_norm": 3.1482791900634766, |
|
"learning_rate": 9.888201143919563e-05, |
|
"loss": 4.3013, |
|
"step": 21850 |
|
}, |
|
{ |
|
"epoch": 0.35322010935307496, |
|
"grad_norm": 1.2597465515136719, |
|
"learning_rate": 9.887666032071077e-05, |
|
"loss": 3.9662, |
|
"step": 21900 |
|
}, |
|
{ |
|
"epoch": 0.35402654795890387, |
|
"grad_norm": 2.1143901348114014, |
|
"learning_rate": 9.887129657202513e-05, |
|
"loss": 3.8487, |
|
"step": 21950 |
|
}, |
|
{ |
|
"epoch": 0.35483298656473283, |
|
"grad_norm": 2.351919651031494, |
|
"learning_rate": 9.886592019452475e-05, |
|
"loss": 4.0369, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 0.35563942517056174, |
|
"grad_norm": 2.052345037460327, |
|
"learning_rate": 9.886053118959895e-05, |
|
"loss": 4.1324, |
|
"step": 22050 |
|
}, |
|
{ |
|
"epoch": 0.3564458637763907, |
|
"grad_norm": 3.6101315021514893, |
|
"learning_rate": 9.885512955864027e-05, |
|
"loss": 3.8461, |
|
"step": 22100 |
|
}, |
|
{ |
|
"epoch": 0.35725230238221967, |
|
"grad_norm": 2.581754684448242, |
|
"learning_rate": 9.884971530304458e-05, |
|
"loss": 3.85, |
|
"step": 22150 |
|
}, |
|
{ |
|
"epoch": 0.3580587409880486, |
|
"grad_norm": 1.2572892904281616, |
|
"learning_rate": 9.884428842421096e-05, |
|
"loss": 3.8305, |
|
"step": 22200 |
|
}, |
|
{ |
|
"epoch": 0.35886517959387754, |
|
"grad_norm": 3.1739416122436523, |
|
"learning_rate": 9.883884892354177e-05, |
|
"loss": 3.953, |
|
"step": 22250 |
|
}, |
|
{ |
|
"epoch": 0.35967161819970644, |
|
"grad_norm": 2.128600597381592, |
|
"learning_rate": 9.883339680244261e-05, |
|
"loss": 3.6528, |
|
"step": 22300 |
|
}, |
|
{ |
|
"epoch": 0.3604780568055354, |
|
"grad_norm": 1.8549989461898804, |
|
"learning_rate": 9.882793206232239e-05, |
|
"loss": 3.9435, |
|
"step": 22350 |
|
}, |
|
{ |
|
"epoch": 0.3612844954113643, |
|
"grad_norm": 1.9381146430969238, |
|
"learning_rate": 9.882245470459324e-05, |
|
"loss": 4.1962, |
|
"step": 22400 |
|
}, |
|
{ |
|
"epoch": 0.3620909340171933, |
|
"grad_norm": 3.481395959854126, |
|
"learning_rate": 9.881696473067056e-05, |
|
"loss": 4.4088, |
|
"step": 22450 |
|
}, |
|
{ |
|
"epoch": 0.3628973726230222, |
|
"grad_norm": 2.8906712532043457, |
|
"learning_rate": 9.881146214197302e-05, |
|
"loss": 4.0288, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 0.36370381122885115, |
|
"grad_norm": 1.037143588066101, |
|
"learning_rate": 9.880605736756518e-05, |
|
"loss": 3.7669, |
|
"step": 22550 |
|
}, |
|
{ |
|
"epoch": 0.3645102498346801, |
|
"grad_norm": 1.2270286083221436, |
|
"learning_rate": 9.880052980581152e-05, |
|
"loss": 3.9334, |
|
"step": 22600 |
|
}, |
|
{ |
|
"epoch": 0.365316688440509, |
|
"grad_norm": 0.97640460729599, |
|
"learning_rate": 9.87949896335299e-05, |
|
"loss": 3.548, |
|
"step": 22650 |
|
}, |
|
{ |
|
"epoch": 0.366123127046338, |
|
"grad_norm": 2.765495538711548, |
|
"learning_rate": 9.8789436852152e-05, |
|
"loss": 4.5431, |
|
"step": 22700 |
|
}, |
|
{ |
|
"epoch": 0.3669295656521669, |
|
"grad_norm": 1.3644137382507324, |
|
"learning_rate": 9.878387146311272e-05, |
|
"loss": 3.9941, |
|
"step": 22750 |
|
}, |
|
{ |
|
"epoch": 0.36773600425799585, |
|
"grad_norm": 3.112457275390625, |
|
"learning_rate": 9.877829346785017e-05, |
|
"loss": 3.8966, |
|
"step": 22800 |
|
}, |
|
{ |
|
"epoch": 0.36854244286382476, |
|
"grad_norm": 2.050264835357666, |
|
"learning_rate": 9.87727028678058e-05, |
|
"loss": 3.7119, |
|
"step": 22850 |
|
}, |
|
{ |
|
"epoch": 0.3693488814696537, |
|
"grad_norm": 3.1221909523010254, |
|
"learning_rate": 9.876709966442425e-05, |
|
"loss": 4.1182, |
|
"step": 22900 |
|
}, |
|
{ |
|
"epoch": 0.37015532007548263, |
|
"grad_norm": 1.9487212896347046, |
|
"learning_rate": 9.876148385915345e-05, |
|
"loss": 3.8674, |
|
"step": 22950 |
|
}, |
|
{ |
|
"epoch": 0.3709617586813116, |
|
"grad_norm": 2.4063527584075928, |
|
"learning_rate": 9.875585545344458e-05, |
|
"loss": 3.9302, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 0.37176819728714056, |
|
"grad_norm": 1.6610281467437744, |
|
"learning_rate": 9.875021444875207e-05, |
|
"loss": 4.1362, |
|
"step": 23050 |
|
}, |
|
{ |
|
"epoch": 0.37257463589296946, |
|
"grad_norm": 1.0215773582458496, |
|
"learning_rate": 9.87445608465336e-05, |
|
"loss": 3.9701, |
|
"step": 23100 |
|
}, |
|
{ |
|
"epoch": 0.3733810744987984, |
|
"grad_norm": 2.547252893447876, |
|
"learning_rate": 9.873889464825014e-05, |
|
"loss": 4.3084, |
|
"step": 23150 |
|
}, |
|
{ |
|
"epoch": 0.37418751310462733, |
|
"grad_norm": 3.370795249938965, |
|
"learning_rate": 9.873321585536587e-05, |
|
"loss": 3.9985, |
|
"step": 23200 |
|
}, |
|
{ |
|
"epoch": 0.3749939517104563, |
|
"grad_norm": 1.989953637123108, |
|
"learning_rate": 9.872752446934826e-05, |
|
"loss": 4.0459, |
|
"step": 23250 |
|
}, |
|
{ |
|
"epoch": 0.3758003903162852, |
|
"grad_norm": 3.332749366760254, |
|
"learning_rate": 9.8721820491668e-05, |
|
"loss": 4.1809, |
|
"step": 23300 |
|
}, |
|
{ |
|
"epoch": 0.37660682892211417, |
|
"grad_norm": 2.8768374919891357, |
|
"learning_rate": 9.871610392379906e-05, |
|
"loss": 4.0144, |
|
"step": 23350 |
|
}, |
|
{ |
|
"epoch": 0.3774132675279431, |
|
"grad_norm": 1.8673336505889893, |
|
"learning_rate": 9.871037476721868e-05, |
|
"loss": 3.769, |
|
"step": 23400 |
|
}, |
|
{ |
|
"epoch": 0.37821970613377204, |
|
"grad_norm": 2.023162603378296, |
|
"learning_rate": 9.870463302340729e-05, |
|
"loss": 4.3442, |
|
"step": 23450 |
|
}, |
|
{ |
|
"epoch": 0.379026144739601, |
|
"grad_norm": 1.667543649673462, |
|
"learning_rate": 9.869887869384864e-05, |
|
"loss": 3.688, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 0.3798325833454299, |
|
"grad_norm": 4.761668682098389, |
|
"learning_rate": 9.869311178002969e-05, |
|
"loss": 3.6925, |
|
"step": 23550 |
|
}, |
|
{ |
|
"epoch": 0.38063902195125887, |
|
"grad_norm": 2.4622697830200195, |
|
"learning_rate": 9.868733228344068e-05, |
|
"loss": 4.3784, |
|
"step": 23600 |
|
}, |
|
{ |
|
"epoch": 0.3814454605570878, |
|
"grad_norm": 2.6539041996002197, |
|
"learning_rate": 9.868154020557507e-05, |
|
"loss": 4.117, |
|
"step": 23650 |
|
}, |
|
{ |
|
"epoch": 0.38225189916291674, |
|
"grad_norm": 1.5162415504455566, |
|
"learning_rate": 9.867573554792961e-05, |
|
"loss": 4.1188, |
|
"step": 23700 |
|
}, |
|
{ |
|
"epoch": 0.38305833776874565, |
|
"grad_norm": 1.5139150619506836, |
|
"learning_rate": 9.866991831200427e-05, |
|
"loss": 4.0601, |
|
"step": 23750 |
|
}, |
|
{ |
|
"epoch": 0.3838647763745746, |
|
"grad_norm": 1.5848387479782104, |
|
"learning_rate": 9.866408849930226e-05, |
|
"loss": 4.3303, |
|
"step": 23800 |
|
}, |
|
{ |
|
"epoch": 0.3846712149804035, |
|
"grad_norm": 2.323932647705078, |
|
"learning_rate": 9.86582461113301e-05, |
|
"loss": 4.3198, |
|
"step": 23850 |
|
}, |
|
{ |
|
"epoch": 0.3854776535862325, |
|
"grad_norm": 1.6189751625061035, |
|
"learning_rate": 9.865239114959748e-05, |
|
"loss": 3.7001, |
|
"step": 23900 |
|
}, |
|
{ |
|
"epoch": 0.38628409219206145, |
|
"grad_norm": 1.6173287630081177, |
|
"learning_rate": 9.86465236156174e-05, |
|
"loss": 3.9669, |
|
"step": 23950 |
|
}, |
|
{ |
|
"epoch": 0.38709053079789035, |
|
"grad_norm": 1.945816993713379, |
|
"learning_rate": 9.864064351090608e-05, |
|
"loss": 4.1111, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 0.38709053079789035, |
|
"eval_asr_loss": 0.5596013197241803, |
|
"eval_loss": 4.975149631500244, |
|
"eval_runtime": 15.3079, |
|
"eval_samples_per_second": 33.447, |
|
"eval_steps_per_second": 11.171, |
|
"eval_tts_loss": 5.989607999181007, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 0.3878969694037193, |
|
"grad_norm": 2.5832393169403076, |
|
"learning_rate": 9.863475083698302e-05, |
|
"loss": 4.0228, |
|
"step": 24050 |
|
}, |
|
{ |
|
"epoch": 0.3887034080095482, |
|
"grad_norm": 1.5465142726898193, |
|
"learning_rate": 9.862884559537088e-05, |
|
"loss": 3.9781, |
|
"step": 24100 |
|
}, |
|
{ |
|
"epoch": 0.3895098466153772, |
|
"grad_norm": 2.0018787384033203, |
|
"learning_rate": 9.862292778759572e-05, |
|
"loss": 3.932, |
|
"step": 24150 |
|
}, |
|
{ |
|
"epoch": 0.3903162852212061, |
|
"grad_norm": 1.9746758937835693, |
|
"learning_rate": 9.861699741518669e-05, |
|
"loss": 3.8956, |
|
"step": 24200 |
|
}, |
|
{ |
|
"epoch": 0.39112272382703506, |
|
"grad_norm": 2.558194875717163, |
|
"learning_rate": 9.861105447967629e-05, |
|
"loss": 3.9698, |
|
"step": 24250 |
|
}, |
|
{ |
|
"epoch": 0.39192916243286396, |
|
"grad_norm": 4.342727184295654, |
|
"learning_rate": 9.86050989826002e-05, |
|
"loss": 4.1687, |
|
"step": 24300 |
|
}, |
|
{ |
|
"epoch": 0.3927356010386929, |
|
"grad_norm": 1.6204520463943481, |
|
"learning_rate": 9.859913092549742e-05, |
|
"loss": 3.832, |
|
"step": 24350 |
|
}, |
|
{ |
|
"epoch": 0.39354203964452183, |
|
"grad_norm": 2.4823992252349854, |
|
"learning_rate": 9.859315030991012e-05, |
|
"loss": 4.3509, |
|
"step": 24400 |
|
}, |
|
{ |
|
"epoch": 0.3943484782503508, |
|
"grad_norm": 3.508967876434326, |
|
"learning_rate": 9.858715713738376e-05, |
|
"loss": 4.1016, |
|
"step": 24450 |
|
}, |
|
{ |
|
"epoch": 0.39515491685617976, |
|
"grad_norm": 1.73115873336792, |
|
"learning_rate": 9.858115140946705e-05, |
|
"loss": 3.9989, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 0.39596135546200867, |
|
"grad_norm": 1.9791126251220703, |
|
"learning_rate": 9.85751331277119e-05, |
|
"loss": 4.2463, |
|
"step": 24550 |
|
}, |
|
{ |
|
"epoch": 0.39676779406783763, |
|
"grad_norm": 1.3955203294754028, |
|
"learning_rate": 9.856910229367349e-05, |
|
"loss": 3.9978, |
|
"step": 24600 |
|
}, |
|
{ |
|
"epoch": 0.39757423267366654, |
|
"grad_norm": 2.9622445106506348, |
|
"learning_rate": 9.856305890891027e-05, |
|
"loss": 3.7169, |
|
"step": 24650 |
|
}, |
|
{ |
|
"epoch": 0.3983806712794955, |
|
"grad_norm": 1.4486901760101318, |
|
"learning_rate": 9.855700297498392e-05, |
|
"loss": 3.9363, |
|
"step": 24700 |
|
}, |
|
{ |
|
"epoch": 0.3991871098853244, |
|
"grad_norm": 1.3110404014587402, |
|
"learning_rate": 9.85509344934593e-05, |
|
"loss": 3.9389, |
|
"step": 24750 |
|
}, |
|
{ |
|
"epoch": 0.3999935484911534, |
|
"grad_norm": 5.3704752922058105, |
|
"learning_rate": 9.854485346590461e-05, |
|
"loss": 4.0937, |
|
"step": 24800 |
|
}, |
|
{ |
|
"epoch": 0.4007999870969823, |
|
"grad_norm": 1.8054653406143188, |
|
"learning_rate": 9.853875989389123e-05, |
|
"loss": 4.0569, |
|
"step": 24850 |
|
}, |
|
{ |
|
"epoch": 0.40160642570281124, |
|
"grad_norm": 2.4662413597106934, |
|
"learning_rate": 9.85326537789938e-05, |
|
"loss": 4.0274, |
|
"step": 24900 |
|
}, |
|
{ |
|
"epoch": 0.4024128643086402, |
|
"grad_norm": 2.1582422256469727, |
|
"learning_rate": 9.852665761880885e-05, |
|
"loss": 3.9833, |
|
"step": 24950 |
|
}, |
|
{ |
|
"epoch": 0.4032193029144691, |
|
"grad_norm": 1.7549717426300049, |
|
"learning_rate": 9.852052667365919e-05, |
|
"loss": 4.0434, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 0.4040257415202981, |
|
"grad_norm": 2.722822427749634, |
|
"learning_rate": 9.85143831903371e-05, |
|
"loss": 4.2247, |
|
"step": 25050 |
|
}, |
|
{ |
|
"epoch": 0.404832180126127, |
|
"grad_norm": 2.266624689102173, |
|
"learning_rate": 9.850822717043014e-05, |
|
"loss": 4.0886, |
|
"step": 25100 |
|
}, |
|
{ |
|
"epoch": 0.40563861873195595, |
|
"grad_norm": 2.5876810550689697, |
|
"learning_rate": 9.85020586155291e-05, |
|
"loss": 3.9883, |
|
"step": 25150 |
|
}, |
|
{ |
|
"epoch": 0.40644505733778485, |
|
"grad_norm": 2.048994779586792, |
|
"learning_rate": 9.849587752722795e-05, |
|
"loss": 3.7832, |
|
"step": 25200 |
|
}, |
|
{ |
|
"epoch": 0.4072514959436138, |
|
"grad_norm": 2.542686939239502, |
|
"learning_rate": 9.848968390712398e-05, |
|
"loss": 4.1333, |
|
"step": 25250 |
|
}, |
|
{ |
|
"epoch": 0.4080579345494427, |
|
"grad_norm": 1.8834608793258667, |
|
"learning_rate": 9.848347775681765e-05, |
|
"loss": 4.3341, |
|
"step": 25300 |
|
}, |
|
{ |
|
"epoch": 0.4088643731552717, |
|
"grad_norm": 2.2536680698394775, |
|
"learning_rate": 9.847725907791271e-05, |
|
"loss": 3.9732, |
|
"step": 25350 |
|
}, |
|
{ |
|
"epoch": 0.40967081176110065, |
|
"grad_norm": 2.5030934810638428, |
|
"learning_rate": 9.847102787201614e-05, |
|
"loss": 4.0837, |
|
"step": 25400 |
|
}, |
|
{ |
|
"epoch": 0.41047725036692956, |
|
"grad_norm": 2.3016812801361084, |
|
"learning_rate": 9.846478414073812e-05, |
|
"loss": 3.8695, |
|
"step": 25450 |
|
}, |
|
{ |
|
"epoch": 0.4112836889727585, |
|
"grad_norm": 1.6216686964035034, |
|
"learning_rate": 9.84585278856921e-05, |
|
"loss": 3.8525, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 0.41209012757858743, |
|
"grad_norm": 4.122720718383789, |
|
"learning_rate": 9.845225910849476e-05, |
|
"loss": 4.1526, |
|
"step": 25550 |
|
}, |
|
{ |
|
"epoch": 0.4128965661844164, |
|
"grad_norm": 2.8652255535125732, |
|
"learning_rate": 9.844597781076599e-05, |
|
"loss": 3.8846, |
|
"step": 25600 |
|
}, |
|
{ |
|
"epoch": 0.4137030047902453, |
|
"grad_norm": 1.516522765159607, |
|
"learning_rate": 9.843968399412898e-05, |
|
"loss": 3.9619, |
|
"step": 25650 |
|
}, |
|
{ |
|
"epoch": 0.41450944339607426, |
|
"grad_norm": 1.89898681640625, |
|
"learning_rate": 9.843337766021005e-05, |
|
"loss": 4.0633, |
|
"step": 25700 |
|
}, |
|
{ |
|
"epoch": 0.41531588200190317, |
|
"grad_norm": 3.1514346599578857, |
|
"learning_rate": 9.842705881063889e-05, |
|
"loss": 3.8735, |
|
"step": 25750 |
|
}, |
|
{ |
|
"epoch": 0.41612232060773213, |
|
"grad_norm": 1.562589168548584, |
|
"learning_rate": 9.84207274470483e-05, |
|
"loss": 3.6907, |
|
"step": 25800 |
|
}, |
|
{ |
|
"epoch": 0.4169287592135611, |
|
"grad_norm": 1.0612993240356445, |
|
"learning_rate": 9.841438357107438e-05, |
|
"loss": 4.2971, |
|
"step": 25850 |
|
}, |
|
{ |
|
"epoch": 0.41773519781939, |
|
"grad_norm": 2.6953563690185547, |
|
"learning_rate": 9.840802718435646e-05, |
|
"loss": 3.9486, |
|
"step": 25900 |
|
}, |
|
{ |
|
"epoch": 0.41854163642521897, |
|
"grad_norm": 2.3360955715179443, |
|
"learning_rate": 9.840165828853709e-05, |
|
"loss": 4.0716, |
|
"step": 25950 |
|
}, |
|
{ |
|
"epoch": 0.4193480750310479, |
|
"grad_norm": 2.1841683387756348, |
|
"learning_rate": 9.839527688526204e-05, |
|
"loss": 3.9567, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 0.42015451363687684, |
|
"grad_norm": 2.321568489074707, |
|
"learning_rate": 9.838888297618032e-05, |
|
"loss": 4.0802, |
|
"step": 26050 |
|
}, |
|
{ |
|
"epoch": 0.42096095224270574, |
|
"grad_norm": 2.530367374420166, |
|
"learning_rate": 9.838247656294421e-05, |
|
"loss": 3.7345, |
|
"step": 26100 |
|
}, |
|
{ |
|
"epoch": 0.4217673908485347, |
|
"grad_norm": 1.4943206310272217, |
|
"learning_rate": 9.837605764720916e-05, |
|
"loss": 3.865, |
|
"step": 26150 |
|
}, |
|
{ |
|
"epoch": 0.4225738294543636, |
|
"grad_norm": 1.2419519424438477, |
|
"learning_rate": 9.836962623063391e-05, |
|
"loss": 3.9362, |
|
"step": 26200 |
|
}, |
|
{ |
|
"epoch": 0.4233802680601926, |
|
"grad_norm": 2.8908309936523438, |
|
"learning_rate": 9.836318231488035e-05, |
|
"loss": 3.5773, |
|
"step": 26250 |
|
}, |
|
{ |
|
"epoch": 0.42418670666602154, |
|
"grad_norm": 5.162326335906982, |
|
"learning_rate": 9.835672590161369e-05, |
|
"loss": 3.9459, |
|
"step": 26300 |
|
}, |
|
{ |
|
"epoch": 0.42499314527185045, |
|
"grad_norm": 1.782263994216919, |
|
"learning_rate": 9.835025699250233e-05, |
|
"loss": 3.9619, |
|
"step": 26350 |
|
}, |
|
{ |
|
"epoch": 0.4257995838776794, |
|
"grad_norm": 5.839146137237549, |
|
"learning_rate": 9.834377558921788e-05, |
|
"loss": 3.8363, |
|
"step": 26400 |
|
}, |
|
{ |
|
"epoch": 0.4266060224835083, |
|
"grad_norm": 1.6730350255966187, |
|
"learning_rate": 9.833728169343521e-05, |
|
"loss": 3.7422, |
|
"step": 26450 |
|
}, |
|
{ |
|
"epoch": 0.4274124610893373, |
|
"grad_norm": 0.9565515518188477, |
|
"learning_rate": 9.83307753068324e-05, |
|
"loss": 4.3006, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 0.4282188996951662, |
|
"grad_norm": 1.102502465248108, |
|
"learning_rate": 9.832425643109076e-05, |
|
"loss": 3.7061, |
|
"step": 26550 |
|
}, |
|
{ |
|
"epoch": 0.42902533830099515, |
|
"grad_norm": 2.7195091247558594, |
|
"learning_rate": 9.831772506789483e-05, |
|
"loss": 3.8085, |
|
"step": 26600 |
|
}, |
|
{ |
|
"epoch": 0.42983177690682406, |
|
"grad_norm": 2.3110263347625732, |
|
"learning_rate": 9.831118121893239e-05, |
|
"loss": 3.7295, |
|
"step": 26650 |
|
}, |
|
{ |
|
"epoch": 0.430638215512653, |
|
"grad_norm": 1.7428209781646729, |
|
"learning_rate": 9.830462488589443e-05, |
|
"loss": 4.0876, |
|
"step": 26700 |
|
}, |
|
{ |
|
"epoch": 0.431444654118482, |
|
"grad_norm": 1.8392008543014526, |
|
"learning_rate": 9.829805607047515e-05, |
|
"loss": 4.3742, |
|
"step": 26750 |
|
}, |
|
{ |
|
"epoch": 0.4322510927243109, |
|
"grad_norm": 2.320955276489258, |
|
"learning_rate": 9.829147477437202e-05, |
|
"loss": 4.0089, |
|
"step": 26800 |
|
}, |
|
{ |
|
"epoch": 0.43305753133013986, |
|
"grad_norm": 2.5901613235473633, |
|
"learning_rate": 9.82848809992857e-05, |
|
"loss": 4.0296, |
|
"step": 26850 |
|
}, |
|
{ |
|
"epoch": 0.43386396993596876, |
|
"grad_norm": 4.836414337158203, |
|
"learning_rate": 9.827827474692009e-05, |
|
"loss": 4.3018, |
|
"step": 26900 |
|
}, |
|
{ |
|
"epoch": 0.4346704085417977, |
|
"grad_norm": 3.3830177783966064, |
|
"learning_rate": 9.82716560189823e-05, |
|
"loss": 3.949, |
|
"step": 26950 |
|
}, |
|
{ |
|
"epoch": 0.43547684714762663, |
|
"grad_norm": 1.8795298337936401, |
|
"learning_rate": 9.826502481718267e-05, |
|
"loss": 3.7089, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 0.43547684714762663, |
|
"eval_asr_loss": 0.5559742133953548, |
|
"eval_loss": 4.958992958068848, |
|
"eval_runtime": 14.7232, |
|
"eval_samples_per_second": 34.775, |
|
"eval_steps_per_second": 11.614, |
|
"eval_tts_loss": 6.008725999135946, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 0.4362832857534556, |
|
"grad_norm": 1.627994418144226, |
|
"learning_rate": 9.825838114323478e-05, |
|
"loss": 3.9582, |
|
"step": 27050 |
|
}, |
|
{ |
|
"epoch": 0.4370897243592845, |
|
"grad_norm": 4.4471611976623535, |
|
"learning_rate": 9.825172499885542e-05, |
|
"loss": 3.8203, |
|
"step": 27100 |
|
}, |
|
{ |
|
"epoch": 0.43789616296511347, |
|
"grad_norm": 4.530590534210205, |
|
"learning_rate": 9.82450563857646e-05, |
|
"loss": 4.1167, |
|
"step": 27150 |
|
}, |
|
{ |
|
"epoch": 0.43870260157094243, |
|
"grad_norm": 2.129225969314575, |
|
"learning_rate": 9.823837530568554e-05, |
|
"loss": 3.9683, |
|
"step": 27200 |
|
}, |
|
{ |
|
"epoch": 0.43950904017677134, |
|
"grad_norm": 1.8322468996047974, |
|
"learning_rate": 9.823168176034471e-05, |
|
"loss": 4.0839, |
|
"step": 27250 |
|
}, |
|
{ |
|
"epoch": 0.4403154787826003, |
|
"grad_norm": 2.424598217010498, |
|
"learning_rate": 9.822497575147178e-05, |
|
"loss": 4.1781, |
|
"step": 27300 |
|
}, |
|
{ |
|
"epoch": 0.4411219173884292, |
|
"grad_norm": 1.3116930723190308, |
|
"learning_rate": 9.821839177232751e-05, |
|
"loss": 3.9697, |
|
"step": 27350 |
|
}, |
|
{ |
|
"epoch": 0.44192835599425817, |
|
"grad_norm": 1.402601718902588, |
|
"learning_rate": 9.821166109077653e-05, |
|
"loss": 4.0815, |
|
"step": 27400 |
|
}, |
|
{ |
|
"epoch": 0.4427347946000871, |
|
"grad_norm": 1.392431616783142, |
|
"learning_rate": 9.820491795086698e-05, |
|
"loss": 4.0143, |
|
"step": 27450 |
|
}, |
|
{ |
|
"epoch": 0.44354123320591604, |
|
"grad_norm": 1.231276035308838, |
|
"learning_rate": 9.819816235434137e-05, |
|
"loss": 4.3486, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 0.44434767181174495, |
|
"grad_norm": 3.108103036880493, |
|
"learning_rate": 9.819139430294539e-05, |
|
"loss": 4.0492, |
|
"step": 27550 |
|
}, |
|
{ |
|
"epoch": 0.4451541104175739, |
|
"grad_norm": 2.1228959560394287, |
|
"learning_rate": 9.818461379842798e-05, |
|
"loss": 4.3671, |
|
"step": 27600 |
|
}, |
|
{ |
|
"epoch": 0.4459605490234029, |
|
"grad_norm": 1.8643229007720947, |
|
"learning_rate": 9.81778208425413e-05, |
|
"loss": 3.9776, |
|
"step": 27650 |
|
}, |
|
{ |
|
"epoch": 0.4467669876292318, |
|
"grad_norm": 1.3176897764205933, |
|
"learning_rate": 9.81710154370407e-05, |
|
"loss": 4.368, |
|
"step": 27700 |
|
}, |
|
{ |
|
"epoch": 0.44757342623506075, |
|
"grad_norm": 2.3601205348968506, |
|
"learning_rate": 9.816419758368476e-05, |
|
"loss": 3.6981, |
|
"step": 27750 |
|
}, |
|
{ |
|
"epoch": 0.44837986484088965, |
|
"grad_norm": 4.233773708343506, |
|
"learning_rate": 9.815736728423531e-05, |
|
"loss": 4.4567, |
|
"step": 27800 |
|
}, |
|
{ |
|
"epoch": 0.4491863034467186, |
|
"grad_norm": 2.077209949493408, |
|
"learning_rate": 9.81505245404573e-05, |
|
"loss": 3.854, |
|
"step": 27850 |
|
}, |
|
{ |
|
"epoch": 0.4499927420525475, |
|
"grad_norm": 1.6529903411865234, |
|
"learning_rate": 9.814366935411903e-05, |
|
"loss": 4.254, |
|
"step": 27900 |
|
}, |
|
{ |
|
"epoch": 0.4507991806583765, |
|
"grad_norm": 2.4949686527252197, |
|
"learning_rate": 9.813680172699191e-05, |
|
"loss": 4.1666, |
|
"step": 27950 |
|
}, |
|
{ |
|
"epoch": 0.4516056192642054, |
|
"grad_norm": 1.0352684259414673, |
|
"learning_rate": 9.81299216608506e-05, |
|
"loss": 3.9291, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 0.45241205787003436, |
|
"grad_norm": 1.945921778678894, |
|
"learning_rate": 9.8123029157473e-05, |
|
"loss": 4.0353, |
|
"step": 28050 |
|
}, |
|
{ |
|
"epoch": 0.4532184964758633, |
|
"grad_norm": 1.8263922929763794, |
|
"learning_rate": 9.811612421864018e-05, |
|
"loss": 3.984, |
|
"step": 28100 |
|
}, |
|
{ |
|
"epoch": 0.4540249350816922, |
|
"grad_norm": 2.8539750576019287, |
|
"learning_rate": 9.810920684613644e-05, |
|
"loss": 4.1103, |
|
"step": 28150 |
|
}, |
|
{ |
|
"epoch": 0.4548313736875212, |
|
"grad_norm": 2.559903860092163, |
|
"learning_rate": 9.81022770417493e-05, |
|
"loss": 4.0392, |
|
"step": 28200 |
|
}, |
|
{ |
|
"epoch": 0.4556378122933501, |
|
"grad_norm": 1.7837358713150024, |
|
"learning_rate": 9.809533480726952e-05, |
|
"loss": 3.9311, |
|
"step": 28250 |
|
}, |
|
{ |
|
"epoch": 0.45644425089917906, |
|
"grad_norm": 1.3091505765914917, |
|
"learning_rate": 9.808838014449097e-05, |
|
"loss": 3.7175, |
|
"step": 28300 |
|
}, |
|
{ |
|
"epoch": 0.45725068950500797, |
|
"grad_norm": 1.9200706481933594, |
|
"learning_rate": 9.808141305521087e-05, |
|
"loss": 4.3186, |
|
"step": 28350 |
|
}, |
|
{ |
|
"epoch": 0.45805712811083693, |
|
"grad_norm": 2.3566806316375732, |
|
"learning_rate": 9.807443354122955e-05, |
|
"loss": 4.1294, |
|
"step": 28400 |
|
}, |
|
{ |
|
"epoch": 0.45886356671666584, |
|
"grad_norm": 1.5939757823944092, |
|
"learning_rate": 9.806744160435059e-05, |
|
"loss": 3.8491, |
|
"step": 28450 |
|
}, |
|
{ |
|
"epoch": 0.4596700053224948, |
|
"grad_norm": 1.4164952039718628, |
|
"learning_rate": 9.806043724638076e-05, |
|
"loss": 4.2949, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 0.46047644392832376, |
|
"grad_norm": 2.1956284046173096, |
|
"learning_rate": 9.805342046913009e-05, |
|
"loss": 4.1996, |
|
"step": 28550 |
|
}, |
|
{ |
|
"epoch": 0.46128288253415267, |
|
"grad_norm": 1.7804524898529053, |
|
"learning_rate": 9.804639127441176e-05, |
|
"loss": 3.9101, |
|
"step": 28600 |
|
}, |
|
{ |
|
"epoch": 0.46208932113998163, |
|
"grad_norm": 0.5685811638832092, |
|
"learning_rate": 9.803934966404217e-05, |
|
"loss": 4.1905, |
|
"step": 28650 |
|
}, |
|
{ |
|
"epoch": 0.46289575974581054, |
|
"grad_norm": 1.6610820293426514, |
|
"learning_rate": 9.803229563984095e-05, |
|
"loss": 4.3581, |
|
"step": 28700 |
|
}, |
|
{ |
|
"epoch": 0.4637021983516395, |
|
"grad_norm": 2.0791728496551514, |
|
"learning_rate": 9.802522920363095e-05, |
|
"loss": 3.9016, |
|
"step": 28750 |
|
}, |
|
{ |
|
"epoch": 0.4645086369574684, |
|
"grad_norm": 2.6067445278167725, |
|
"learning_rate": 9.801815035723817e-05, |
|
"loss": 3.9681, |
|
"step": 28800 |
|
}, |
|
{ |
|
"epoch": 0.4653150755632974, |
|
"grad_norm": 0.8365187644958496, |
|
"learning_rate": 9.801105910249187e-05, |
|
"loss": 3.9012, |
|
"step": 28850 |
|
}, |
|
{ |
|
"epoch": 0.4661215141691263, |
|
"grad_norm": 3.1129026412963867, |
|
"learning_rate": 9.80039554412245e-05, |
|
"loss": 3.9982, |
|
"step": 28900 |
|
}, |
|
{ |
|
"epoch": 0.46692795277495525, |
|
"grad_norm": 1.3477463722229004, |
|
"learning_rate": 9.799683937527171e-05, |
|
"loss": 4.1739, |
|
"step": 28950 |
|
}, |
|
{ |
|
"epoch": 0.46773439138078415, |
|
"grad_norm": 3.441631317138672, |
|
"learning_rate": 9.798971090647238e-05, |
|
"loss": 4.0417, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 0.4685408299866131, |
|
"grad_norm": 2.399874210357666, |
|
"learning_rate": 9.798257003666853e-05, |
|
"loss": 4.1551, |
|
"step": 29050 |
|
}, |
|
{ |
|
"epoch": 0.4693472685924421, |
|
"grad_norm": 2.2936065196990967, |
|
"learning_rate": 9.797541676770547e-05, |
|
"loss": 4.1262, |
|
"step": 29100 |
|
}, |
|
{ |
|
"epoch": 0.470153707198271, |
|
"grad_norm": 2.890214204788208, |
|
"learning_rate": 9.796825110143165e-05, |
|
"loss": 4.1734, |
|
"step": 29150 |
|
}, |
|
{ |
|
"epoch": 0.47096014580409995, |
|
"grad_norm": 1.6101442575454712, |
|
"learning_rate": 9.796107303969877e-05, |
|
"loss": 3.6554, |
|
"step": 29200 |
|
}, |
|
{ |
|
"epoch": 0.47176658440992886, |
|
"grad_norm": 1.5904902219772339, |
|
"learning_rate": 9.795388258436169e-05, |
|
"loss": 4.1421, |
|
"step": 29250 |
|
}, |
|
{ |
|
"epoch": 0.4725730230157578, |
|
"grad_norm": 1.2808867692947388, |
|
"learning_rate": 9.79466797372785e-05, |
|
"loss": 3.7509, |
|
"step": 29300 |
|
}, |
|
{ |
|
"epoch": 0.47337946162158673, |
|
"grad_norm": 2.6117844581604004, |
|
"learning_rate": 9.793946450031049e-05, |
|
"loss": 4.252, |
|
"step": 29350 |
|
}, |
|
{ |
|
"epoch": 0.4741859002274157, |
|
"grad_norm": 2.114128351211548, |
|
"learning_rate": 9.793238154921244e-05, |
|
"loss": 4.0362, |
|
"step": 29400 |
|
}, |
|
{ |
|
"epoch": 0.4749923388332446, |
|
"grad_norm": 0.834858775138855, |
|
"learning_rate": 9.792514178577616e-05, |
|
"loss": 3.8481, |
|
"step": 29450 |
|
}, |
|
{ |
|
"epoch": 0.47579877743907356, |
|
"grad_norm": 1.2032753229141235, |
|
"learning_rate": 9.791788963802069e-05, |
|
"loss": 4.0954, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 0.4766052160449025, |
|
"grad_norm": 1.481992244720459, |
|
"learning_rate": 9.791062510782003e-05, |
|
"loss": 4.0293, |
|
"step": 29550 |
|
}, |
|
{ |
|
"epoch": 0.47741165465073143, |
|
"grad_norm": 1.40059494972229, |
|
"learning_rate": 9.790334819705142e-05, |
|
"loss": 3.8551, |
|
"step": 29600 |
|
}, |
|
{ |
|
"epoch": 0.4782180932565604, |
|
"grad_norm": 1.1095318794250488, |
|
"learning_rate": 9.789605890759526e-05, |
|
"loss": 4.0393, |
|
"step": 29650 |
|
}, |
|
{ |
|
"epoch": 0.4790245318623893, |
|
"grad_norm": 2.459015369415283, |
|
"learning_rate": 9.78887572413352e-05, |
|
"loss": 4.1729, |
|
"step": 29700 |
|
}, |
|
{ |
|
"epoch": 0.47983097046821827, |
|
"grad_norm": 1.1588975191116333, |
|
"learning_rate": 9.788144320015804e-05, |
|
"loss": 3.971, |
|
"step": 29750 |
|
}, |
|
{ |
|
"epoch": 0.4806374090740472, |
|
"grad_norm": 2.8166420459747314, |
|
"learning_rate": 9.78741167859538e-05, |
|
"loss": 4.4002, |
|
"step": 29800 |
|
}, |
|
{ |
|
"epoch": 0.48144384767987614, |
|
"grad_norm": 3.607013702392578, |
|
"learning_rate": 9.786677800061571e-05, |
|
"loss": 3.819, |
|
"step": 29850 |
|
}, |
|
{ |
|
"epoch": 0.48225028628570504, |
|
"grad_norm": 2.313955783843994, |
|
"learning_rate": 9.785942684604018e-05, |
|
"loss": 4.2892, |
|
"step": 29900 |
|
}, |
|
{ |
|
"epoch": 0.483056724891534, |
|
"grad_norm": 1.3837124109268188, |
|
"learning_rate": 9.785206332412682e-05, |
|
"loss": 4.0314, |
|
"step": 29950 |
|
}, |
|
{ |
|
"epoch": 0.48386316349736297, |
|
"grad_norm": 1.5818651914596558, |
|
"learning_rate": 9.784468743677842e-05, |
|
"loss": 3.8618, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 0.48386316349736297, |
|
"eval_asr_loss": 0.5646996813599123, |
|
"eval_loss": 4.947766304016113, |
|
"eval_runtime": 14.9459, |
|
"eval_samples_per_second": 34.257, |
|
"eval_steps_per_second": 11.441, |
|
"eval_tts_loss": 5.988222905527309, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 0.4846696021031919, |
|
"grad_norm": 2.607830047607422, |
|
"learning_rate": 9.7837299185901e-05, |
|
"loss": 4.127, |
|
"step": 30050 |
|
}, |
|
{ |
|
"epoch": 0.48547604070902084, |
|
"grad_norm": 1.471893310546875, |
|
"learning_rate": 9.782989857340373e-05, |
|
"loss": 3.7895, |
|
"step": 30100 |
|
}, |
|
{ |
|
"epoch": 0.48628247931484975, |
|
"grad_norm": 2.5169291496276855, |
|
"learning_rate": 9.782248560119902e-05, |
|
"loss": 4.0763, |
|
"step": 30150 |
|
}, |
|
{ |
|
"epoch": 0.4870889179206787, |
|
"grad_norm": 4.164167881011963, |
|
"learning_rate": 9.781506027120247e-05, |
|
"loss": 4.0099, |
|
"step": 30200 |
|
}, |
|
{ |
|
"epoch": 0.4878953565265076, |
|
"grad_norm": 2.2432568073272705, |
|
"learning_rate": 9.780762258533282e-05, |
|
"loss": 4.1052, |
|
"step": 30250 |
|
}, |
|
{ |
|
"epoch": 0.4887017951323366, |
|
"grad_norm": 1.442578911781311, |
|
"learning_rate": 9.780017254551206e-05, |
|
"loss": 3.9327, |
|
"step": 30300 |
|
}, |
|
{ |
|
"epoch": 0.4895082337381655, |
|
"grad_norm": 1.6483718156814575, |
|
"learning_rate": 9.779271015366535e-05, |
|
"loss": 3.7249, |
|
"step": 30350 |
|
}, |
|
{ |
|
"epoch": 0.49031467234399445, |
|
"grad_norm": 2.119659662246704, |
|
"learning_rate": 9.778523541172104e-05, |
|
"loss": 3.959, |
|
"step": 30400 |
|
}, |
|
{ |
|
"epoch": 0.4911211109498234, |
|
"grad_norm": 1.574180006980896, |
|
"learning_rate": 9.777774832161069e-05, |
|
"loss": 4.2155, |
|
"step": 30450 |
|
}, |
|
{ |
|
"epoch": 0.4919275495556523, |
|
"grad_norm": 1.1062897443771362, |
|
"learning_rate": 9.7770248885269e-05, |
|
"loss": 4.263, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 0.4927339881614813, |
|
"grad_norm": 1.0894304513931274, |
|
"learning_rate": 9.776273710463394e-05, |
|
"loss": 3.9637, |
|
"step": 30550 |
|
}, |
|
{ |
|
"epoch": 0.4935404267673102, |
|
"grad_norm": 1.5745190382003784, |
|
"learning_rate": 9.77552129816466e-05, |
|
"loss": 3.9539, |
|
"step": 30600 |
|
}, |
|
{ |
|
"epoch": 0.49434686537313915, |
|
"grad_norm": 1.2136081457138062, |
|
"learning_rate": 9.77476765182513e-05, |
|
"loss": 3.8683, |
|
"step": 30650 |
|
}, |
|
{ |
|
"epoch": 0.49515330397896806, |
|
"grad_norm": 2.5908925533294678, |
|
"learning_rate": 9.774012771639553e-05, |
|
"loss": 4.1582, |
|
"step": 30700 |
|
}, |
|
{ |
|
"epoch": 0.495959742584797, |
|
"grad_norm": 1.8017607927322388, |
|
"learning_rate": 9.773256657802997e-05, |
|
"loss": 3.6259, |
|
"step": 30750 |
|
}, |
|
{ |
|
"epoch": 0.49676618119062593, |
|
"grad_norm": 1.5173224210739136, |
|
"learning_rate": 9.77249931051085e-05, |
|
"loss": 4.0439, |
|
"step": 30800 |
|
}, |
|
{ |
|
"epoch": 0.4975726197964549, |
|
"grad_norm": 15.362312316894531, |
|
"learning_rate": 9.771786279566293e-05, |
|
"loss": 4.387, |
|
"step": 30850 |
|
}, |
|
{ |
|
"epoch": 0.49837905840228386, |
|
"grad_norm": 3.3392670154571533, |
|
"learning_rate": 9.771026539928696e-05, |
|
"loss": 4.5741, |
|
"step": 30900 |
|
}, |
|
{ |
|
"epoch": 0.49918549700811277, |
|
"grad_norm": 1.549731969833374, |
|
"learning_rate": 9.770265567411791e-05, |
|
"loss": 3.9523, |
|
"step": 30950 |
|
}, |
|
{ |
|
"epoch": 0.49999193561394173, |
|
"grad_norm": 2.7253308296203613, |
|
"learning_rate": 9.769503362212222e-05, |
|
"loss": 4.3312, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 0.5007983742197707, |
|
"grad_norm": 2.1496167182922363, |
|
"learning_rate": 9.768739924526948e-05, |
|
"loss": 3.9847, |
|
"step": 31050 |
|
}, |
|
{ |
|
"epoch": 0.5016048128255995, |
|
"grad_norm": 2.93792462348938, |
|
"learning_rate": 9.76797525455325e-05, |
|
"loss": 3.9403, |
|
"step": 31100 |
|
}, |
|
{ |
|
"epoch": 0.5024112514314285, |
|
"grad_norm": 2.272951126098633, |
|
"learning_rate": 9.767209352488727e-05, |
|
"loss": 3.7558, |
|
"step": 31150 |
|
}, |
|
{ |
|
"epoch": 0.5032176900372575, |
|
"grad_norm": 3.371812105178833, |
|
"learning_rate": 9.766442218531293e-05, |
|
"loss": 3.9215, |
|
"step": 31200 |
|
}, |
|
{ |
|
"epoch": 0.5040241286430864, |
|
"grad_norm": 1.3443658351898193, |
|
"learning_rate": 9.765673852879186e-05, |
|
"loss": 4.1457, |
|
"step": 31250 |
|
}, |
|
{ |
|
"epoch": 0.5048305672489153, |
|
"grad_norm": 2.311265230178833, |
|
"learning_rate": 9.764904255730955e-05, |
|
"loss": 4.3678, |
|
"step": 31300 |
|
}, |
|
{ |
|
"epoch": 0.5056370058547442, |
|
"grad_norm": 2.241488456726074, |
|
"learning_rate": 9.764133427285475e-05, |
|
"loss": 4.145, |
|
"step": 31350 |
|
}, |
|
{ |
|
"epoch": 0.5064434444605732, |
|
"grad_norm": 2.38242769241333, |
|
"learning_rate": 9.763361367741934e-05, |
|
"loss": 4.1718, |
|
"step": 31400 |
|
}, |
|
{ |
|
"epoch": 0.5072498830664022, |
|
"grad_norm": 1.874671220779419, |
|
"learning_rate": 9.76258807729984e-05, |
|
"loss": 3.9391, |
|
"step": 31450 |
|
}, |
|
{ |
|
"epoch": 0.5080563216722311, |
|
"grad_norm": 1.9810266494750977, |
|
"learning_rate": 9.761813556159019e-05, |
|
"loss": 3.9962, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 0.50886276027806, |
|
"grad_norm": 2.93658185005188, |
|
"learning_rate": 9.761037804519612e-05, |
|
"loss": 3.7464, |
|
"step": 31550 |
|
}, |
|
{ |
|
"epoch": 0.509669198883889, |
|
"grad_norm": 2.2134599685668945, |
|
"learning_rate": 9.760260822582086e-05, |
|
"loss": 3.9865, |
|
"step": 31600 |
|
}, |
|
{ |
|
"epoch": 0.5104756374897179, |
|
"grad_norm": 2.7154464721679688, |
|
"learning_rate": 9.759482610547214e-05, |
|
"loss": 3.8921, |
|
"step": 31650 |
|
}, |
|
{ |
|
"epoch": 0.5112820760955469, |
|
"grad_norm": 2.0518405437469482, |
|
"learning_rate": 9.7587031686161e-05, |
|
"loss": 4.1342, |
|
"step": 31700 |
|
}, |
|
{ |
|
"epoch": 0.5120885147013757, |
|
"grad_norm": 2.0980758666992188, |
|
"learning_rate": 9.757922496990155e-05, |
|
"loss": 3.9435, |
|
"step": 31750 |
|
}, |
|
{ |
|
"epoch": 0.5128949533072047, |
|
"grad_norm": 4.004748344421387, |
|
"learning_rate": 9.757140595871115e-05, |
|
"loss": 4.1433, |
|
"step": 31800 |
|
}, |
|
{ |
|
"epoch": 0.5137013919130337, |
|
"grad_norm": 1.889359474182129, |
|
"learning_rate": 9.756357465461027e-05, |
|
"loss": 4.1632, |
|
"step": 31850 |
|
}, |
|
{ |
|
"epoch": 0.5145078305188626, |
|
"grad_norm": 2.1634538173675537, |
|
"learning_rate": 9.755573105962264e-05, |
|
"loss": 3.9894, |
|
"step": 31900 |
|
}, |
|
{ |
|
"epoch": 0.5153142691246916, |
|
"grad_norm": 2.0058374404907227, |
|
"learning_rate": 9.754787517577508e-05, |
|
"loss": 3.7636, |
|
"step": 31950 |
|
}, |
|
{ |
|
"epoch": 0.5161207077305204, |
|
"grad_norm": 3.5357580184936523, |
|
"learning_rate": 9.754000700509764e-05, |
|
"loss": 3.9957, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 0.5169271463363494, |
|
"grad_norm": 1.03873872756958, |
|
"learning_rate": 9.753212654962356e-05, |
|
"loss": 4.1323, |
|
"step": 32050 |
|
}, |
|
{ |
|
"epoch": 0.5177335849421784, |
|
"grad_norm": 2.3036513328552246, |
|
"learning_rate": 9.752423381138919e-05, |
|
"loss": 4.057, |
|
"step": 32100 |
|
}, |
|
{ |
|
"epoch": 0.5185400235480073, |
|
"grad_norm": 1.3666496276855469, |
|
"learning_rate": 9.75163287924341e-05, |
|
"loss": 4.1997, |
|
"step": 32150 |
|
}, |
|
{ |
|
"epoch": 0.5193464621538362, |
|
"grad_norm": 2.106926441192627, |
|
"learning_rate": 9.750841149480102e-05, |
|
"loss": 3.924, |
|
"step": 32200 |
|
}, |
|
{ |
|
"epoch": 0.5201529007596651, |
|
"grad_norm": 3.9756150245666504, |
|
"learning_rate": 9.750048192053587e-05, |
|
"loss": 3.9417, |
|
"step": 32250 |
|
}, |
|
{ |
|
"epoch": 0.5209593393654941, |
|
"grad_norm": 1.9011123180389404, |
|
"learning_rate": 9.749254007168771e-05, |
|
"loss": 3.9619, |
|
"step": 32300 |
|
}, |
|
{ |
|
"epoch": 0.5217657779713231, |
|
"grad_norm": 1.8549953699111938, |
|
"learning_rate": 9.74845859503088e-05, |
|
"loss": 3.8087, |
|
"step": 32350 |
|
}, |
|
{ |
|
"epoch": 0.522572216577152, |
|
"grad_norm": 2.133540153503418, |
|
"learning_rate": 9.747661955845456e-05, |
|
"loss": 4.1356, |
|
"step": 32400 |
|
}, |
|
{ |
|
"epoch": 0.5233786551829809, |
|
"grad_norm": 1.6960670948028564, |
|
"learning_rate": 9.74686408981836e-05, |
|
"loss": 3.798, |
|
"step": 32450 |
|
}, |
|
{ |
|
"epoch": 0.5241850937888098, |
|
"grad_norm": 1.8586905002593994, |
|
"learning_rate": 9.746064997155763e-05, |
|
"loss": 4.0743, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 0.5249915323946388, |
|
"grad_norm": 2.0501816272735596, |
|
"learning_rate": 9.745264678064166e-05, |
|
"loss": 3.6707, |
|
"step": 32550 |
|
}, |
|
{ |
|
"epoch": 0.5257979710004678, |
|
"grad_norm": 2.2415525913238525, |
|
"learning_rate": 9.744463132750373e-05, |
|
"loss": 3.671, |
|
"step": 32600 |
|
}, |
|
{ |
|
"epoch": 0.5266044096062966, |
|
"grad_norm": 2.3392422199249268, |
|
"learning_rate": 9.743660361421512e-05, |
|
"loss": 4.1384, |
|
"step": 32650 |
|
}, |
|
{ |
|
"epoch": 0.5274108482121256, |
|
"grad_norm": 1.0827341079711914, |
|
"learning_rate": 9.74285636428503e-05, |
|
"loss": 4.0892, |
|
"step": 32700 |
|
}, |
|
{ |
|
"epoch": 0.5282172868179545, |
|
"grad_norm": 1.8944828510284424, |
|
"learning_rate": 9.742051141548685e-05, |
|
"loss": 3.9703, |
|
"step": 32750 |
|
}, |
|
{ |
|
"epoch": 0.5290237254237835, |
|
"grad_norm": 2.013651132583618, |
|
"learning_rate": 9.741244693420554e-05, |
|
"loss": 4.2219, |
|
"step": 32800 |
|
}, |
|
{ |
|
"epoch": 0.5298301640296125, |
|
"grad_norm": 1.879599690437317, |
|
"learning_rate": 9.740437020109033e-05, |
|
"loss": 3.7386, |
|
"step": 32850 |
|
}, |
|
{ |
|
"epoch": 0.5306366026354413, |
|
"grad_norm": 2.901381254196167, |
|
"learning_rate": 9.739628121822832e-05, |
|
"loss": 3.9859, |
|
"step": 32900 |
|
}, |
|
{ |
|
"epoch": 0.5314430412412703, |
|
"grad_norm": 2.858835220336914, |
|
"learning_rate": 9.738817998770976e-05, |
|
"loss": 3.8376, |
|
"step": 32950 |
|
}, |
|
{ |
|
"epoch": 0.5322494798470992, |
|
"grad_norm": 2.78635311126709, |
|
"learning_rate": 9.738006651162812e-05, |
|
"loss": 3.8029, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 0.5322494798470992, |
|
"eval_asr_loss": 0.5563808219787441, |
|
"eval_loss": 4.897815227508545, |
|
"eval_runtime": 14.787, |
|
"eval_samples_per_second": 34.625, |
|
"eval_steps_per_second": 11.564, |
|
"eval_tts_loss": 5.9989542354158925, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 0.5330559184529282, |
|
"grad_norm": 1.9606646299362183, |
|
"learning_rate": 9.737194079207997e-05, |
|
"loss": 4.1998, |
|
"step": 33050 |
|
}, |
|
{ |
|
"epoch": 0.5338623570587571, |
|
"grad_norm": 3.6199183464050293, |
|
"learning_rate": 9.736380283116508e-05, |
|
"loss": 3.9197, |
|
"step": 33100 |
|
}, |
|
{ |
|
"epoch": 0.534668795664586, |
|
"grad_norm": 2.4361817836761475, |
|
"learning_rate": 9.735565263098641e-05, |
|
"loss": 3.8554, |
|
"step": 33150 |
|
}, |
|
{ |
|
"epoch": 0.535475234270415, |
|
"grad_norm": 2.393467426300049, |
|
"learning_rate": 9.734749019365001e-05, |
|
"loss": 3.9229, |
|
"step": 33200 |
|
}, |
|
{ |
|
"epoch": 0.536281672876244, |
|
"grad_norm": 1.991856336593628, |
|
"learning_rate": 9.733931552126515e-05, |
|
"loss": 3.9931, |
|
"step": 33250 |
|
}, |
|
{ |
|
"epoch": 0.5370881114820729, |
|
"grad_norm": 1.7219436168670654, |
|
"learning_rate": 9.733112861594422e-05, |
|
"loss": 3.9433, |
|
"step": 33300 |
|
}, |
|
{ |
|
"epoch": 0.5378945500879018, |
|
"grad_norm": 2.8829445838928223, |
|
"learning_rate": 9.732292947980285e-05, |
|
"loss": 4.3263, |
|
"step": 33350 |
|
}, |
|
{ |
|
"epoch": 0.5387009886937307, |
|
"grad_norm": 1.1644093990325928, |
|
"learning_rate": 9.731471811495973e-05, |
|
"loss": 4.1349, |
|
"step": 33400 |
|
}, |
|
{ |
|
"epoch": 0.5395074272995597, |
|
"grad_norm": 2.788134813308716, |
|
"learning_rate": 9.730649452353678e-05, |
|
"loss": 4.1791, |
|
"step": 33450 |
|
}, |
|
{ |
|
"epoch": 0.5403138659053887, |
|
"grad_norm": 1.2594338655471802, |
|
"learning_rate": 9.729825870765903e-05, |
|
"loss": 3.8136, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 0.5411203045112175, |
|
"grad_norm": 2.067070484161377, |
|
"learning_rate": 9.729001066945471e-05, |
|
"loss": 4.1771, |
|
"step": 33550 |
|
}, |
|
{ |
|
"epoch": 0.5419267431170465, |
|
"grad_norm": 0.9377022385597229, |
|
"learning_rate": 9.728175041105519e-05, |
|
"loss": 3.9944, |
|
"step": 33600 |
|
}, |
|
{ |
|
"epoch": 0.5427331817228754, |
|
"grad_norm": 1.7732789516448975, |
|
"learning_rate": 9.7273477934595e-05, |
|
"loss": 4.1222, |
|
"step": 33650 |
|
}, |
|
{ |
|
"epoch": 0.5435396203287044, |
|
"grad_norm": 2.3974225521087646, |
|
"learning_rate": 9.726519324221184e-05, |
|
"loss": 4.2036, |
|
"step": 33700 |
|
}, |
|
{ |
|
"epoch": 0.5443460589345334, |
|
"grad_norm": 3.553283929824829, |
|
"learning_rate": 9.725689633604654e-05, |
|
"loss": 4.4454, |
|
"step": 33750 |
|
}, |
|
{ |
|
"epoch": 0.5451524975403622, |
|
"grad_norm": 1.552428960800171, |
|
"learning_rate": 9.724858721824311e-05, |
|
"loss": 3.8556, |
|
"step": 33800 |
|
}, |
|
{ |
|
"epoch": 0.5459589361461912, |
|
"grad_norm": 1.2861645221710205, |
|
"learning_rate": 9.72402658909487e-05, |
|
"loss": 3.6202, |
|
"step": 33850 |
|
}, |
|
{ |
|
"epoch": 0.5467653747520201, |
|
"grad_norm": 2.8274953365325928, |
|
"learning_rate": 9.723193235631362e-05, |
|
"loss": 3.7635, |
|
"step": 33900 |
|
}, |
|
{ |
|
"epoch": 0.5475718133578491, |
|
"grad_norm": 1.0712840557098389, |
|
"learning_rate": 9.722358661649133e-05, |
|
"loss": 3.7407, |
|
"step": 33950 |
|
}, |
|
{ |
|
"epoch": 0.548378251963678, |
|
"grad_norm": 2.3699841499328613, |
|
"learning_rate": 9.721522867363845e-05, |
|
"loss": 4.0412, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 0.5491846905695069, |
|
"grad_norm": 2.715761423110962, |
|
"learning_rate": 9.72068585299148e-05, |
|
"loss": 3.915, |
|
"step": 34050 |
|
}, |
|
{ |
|
"epoch": 0.5499911291753359, |
|
"grad_norm": 4.615658760070801, |
|
"learning_rate": 9.719847618748323e-05, |
|
"loss": 4.0316, |
|
"step": 34100 |
|
}, |
|
{ |
|
"epoch": 0.5507975677811648, |
|
"grad_norm": 2.428999662399292, |
|
"learning_rate": 9.719008164850986e-05, |
|
"loss": 4.2053, |
|
"step": 34150 |
|
}, |
|
{ |
|
"epoch": 0.5516040063869938, |
|
"grad_norm": 0.604597806930542, |
|
"learning_rate": 9.718167491516395e-05, |
|
"loss": 4.3448, |
|
"step": 34200 |
|
}, |
|
{ |
|
"epoch": 0.5524104449928227, |
|
"grad_norm": 2.038712978363037, |
|
"learning_rate": 9.717325598961781e-05, |
|
"loss": 4.0422, |
|
"step": 34250 |
|
}, |
|
{ |
|
"epoch": 0.5532168835986516, |
|
"grad_norm": 1.6703771352767944, |
|
"learning_rate": 9.716482487404702e-05, |
|
"loss": 3.8453, |
|
"step": 34300 |
|
}, |
|
{ |
|
"epoch": 0.5540233222044806, |
|
"grad_norm": 1.6739346981048584, |
|
"learning_rate": 9.715638157063025e-05, |
|
"loss": 3.9266, |
|
"step": 34350 |
|
}, |
|
{ |
|
"epoch": 0.5548297608103095, |
|
"grad_norm": 2.9111719131469727, |
|
"learning_rate": 9.714792608154932e-05, |
|
"loss": 4.0806, |
|
"step": 34400 |
|
}, |
|
{ |
|
"epoch": 0.5556361994161384, |
|
"grad_norm": 2.14871883392334, |
|
"learning_rate": 9.713945840898924e-05, |
|
"loss": 3.9186, |
|
"step": 34450 |
|
}, |
|
{ |
|
"epoch": 0.5564426380219674, |
|
"grad_norm": 2.2379770278930664, |
|
"learning_rate": 9.71309785551381e-05, |
|
"loss": 3.7397, |
|
"step": 34500 |
|
}, |
|
{ |
|
"epoch": 0.5572490766277963, |
|
"grad_norm": 0.7250761985778809, |
|
"learning_rate": 9.71224865221872e-05, |
|
"loss": 4.056, |
|
"step": 34550 |
|
}, |
|
{ |
|
"epoch": 0.5580555152336253, |
|
"grad_norm": 1.8340208530426025, |
|
"learning_rate": 9.711398231233095e-05, |
|
"loss": 4.2596, |
|
"step": 34600 |
|
}, |
|
{ |
|
"epoch": 0.5588619538394543, |
|
"grad_norm": 1.2574554681777954, |
|
"learning_rate": 9.710546592776695e-05, |
|
"loss": 3.9969, |
|
"step": 34650 |
|
}, |
|
{ |
|
"epoch": 0.5596683924452831, |
|
"grad_norm": 1.2908345460891724, |
|
"learning_rate": 9.709693737069587e-05, |
|
"loss": 3.9818, |
|
"step": 34700 |
|
}, |
|
{ |
|
"epoch": 0.5604748310511121, |
|
"grad_norm": 2.4312095642089844, |
|
"learning_rate": 9.708839664332161e-05, |
|
"loss": 3.7135, |
|
"step": 34750 |
|
}, |
|
{ |
|
"epoch": 0.561281269656941, |
|
"grad_norm": 1.706939458847046, |
|
"learning_rate": 9.707984374785117e-05, |
|
"loss": 3.8907, |
|
"step": 34800 |
|
}, |
|
{ |
|
"epoch": 0.56208770826277, |
|
"grad_norm": 1.253430962562561, |
|
"learning_rate": 9.707127868649468e-05, |
|
"loss": 3.8833, |
|
"step": 34850 |
|
}, |
|
{ |
|
"epoch": 0.5628941468685988, |
|
"grad_norm": 2.2516000270843506, |
|
"learning_rate": 9.706270146146544e-05, |
|
"loss": 4.1565, |
|
"step": 34900 |
|
}, |
|
{ |
|
"epoch": 0.5637005854744278, |
|
"grad_norm": 1.791192889213562, |
|
"learning_rate": 9.705411207497991e-05, |
|
"loss": 3.8712, |
|
"step": 34950 |
|
}, |
|
{ |
|
"epoch": 0.5645070240802568, |
|
"grad_norm": 2.1465959548950195, |
|
"learning_rate": 9.704551052925765e-05, |
|
"loss": 3.8223, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 0.5653134626860857, |
|
"grad_norm": 1.0886331796646118, |
|
"learning_rate": 9.703689682652138e-05, |
|
"loss": 3.9508, |
|
"step": 35050 |
|
}, |
|
{ |
|
"epoch": 0.5661199012919147, |
|
"grad_norm": 2.1944518089294434, |
|
"learning_rate": 9.702827096899698e-05, |
|
"loss": 4.3178, |
|
"step": 35100 |
|
}, |
|
{ |
|
"epoch": 0.5669263398977435, |
|
"grad_norm": 1.7926956415176392, |
|
"learning_rate": 9.701963295891346e-05, |
|
"loss": 3.7831, |
|
"step": 35150 |
|
}, |
|
{ |
|
"epoch": 0.5677327785035725, |
|
"grad_norm": 1.8286861181259155, |
|
"learning_rate": 9.701098279850292e-05, |
|
"loss": 4.0985, |
|
"step": 35200 |
|
}, |
|
{ |
|
"epoch": 0.5685392171094015, |
|
"grad_norm": 1.7975794076919556, |
|
"learning_rate": 9.70023204900007e-05, |
|
"loss": 4.3829, |
|
"step": 35250 |
|
}, |
|
{ |
|
"epoch": 0.5693456557152304, |
|
"grad_norm": 1.3266535997390747, |
|
"learning_rate": 9.69936460356452e-05, |
|
"loss": 3.7414, |
|
"step": 35300 |
|
}, |
|
{ |
|
"epoch": 0.5701520943210593, |
|
"grad_norm": 2.072350263595581, |
|
"learning_rate": 9.698495943767797e-05, |
|
"loss": 4.2939, |
|
"step": 35350 |
|
}, |
|
{ |
|
"epoch": 0.5709585329268883, |
|
"grad_norm": 1.8565208911895752, |
|
"learning_rate": 9.697626069834373e-05, |
|
"loss": 3.8508, |
|
"step": 35400 |
|
}, |
|
{ |
|
"epoch": 0.5717649715327172, |
|
"grad_norm": 1.3206040859222412, |
|
"learning_rate": 9.696754981989031e-05, |
|
"loss": 3.5709, |
|
"step": 35450 |
|
}, |
|
{ |
|
"epoch": 0.5725714101385462, |
|
"grad_norm": 2.337531566619873, |
|
"learning_rate": 9.695882680456869e-05, |
|
"loss": 4.1167, |
|
"step": 35500 |
|
}, |
|
{ |
|
"epoch": 0.5733778487443751, |
|
"grad_norm": 2.7290146350860596, |
|
"learning_rate": 9.695009165463298e-05, |
|
"loss": 3.9105, |
|
"step": 35550 |
|
}, |
|
{ |
|
"epoch": 0.574184287350204, |
|
"grad_norm": 2.879288911819458, |
|
"learning_rate": 9.694134437234044e-05, |
|
"loss": 4.0808, |
|
"step": 35600 |
|
}, |
|
{ |
|
"epoch": 0.574990725956033, |
|
"grad_norm": 2.2341060638427734, |
|
"learning_rate": 9.693258495995141e-05, |
|
"loss": 4.2625, |
|
"step": 35650 |
|
}, |
|
{ |
|
"epoch": 0.5757971645618619, |
|
"grad_norm": 2.0360066890716553, |
|
"learning_rate": 9.692381341972946e-05, |
|
"loss": 3.9065, |
|
"step": 35700 |
|
}, |
|
{ |
|
"epoch": 0.5766036031676909, |
|
"grad_norm": 1.8910644054412842, |
|
"learning_rate": 9.69150297539412e-05, |
|
"loss": 3.7694, |
|
"step": 35750 |
|
}, |
|
{ |
|
"epoch": 0.5774100417735197, |
|
"grad_norm": 1.2420679330825806, |
|
"learning_rate": 9.690623396485642e-05, |
|
"loss": 4.0219, |
|
"step": 35800 |
|
}, |
|
{ |
|
"epoch": 0.5782164803793487, |
|
"grad_norm": 1.8608529567718506, |
|
"learning_rate": 9.689742605474807e-05, |
|
"loss": 3.9607, |
|
"step": 35850 |
|
}, |
|
{ |
|
"epoch": 0.5790229189851777, |
|
"grad_norm": 1.903253197669983, |
|
"learning_rate": 9.688860602589217e-05, |
|
"loss": 4.225, |
|
"step": 35900 |
|
}, |
|
{ |
|
"epoch": 0.5798293575910066, |
|
"grad_norm": 1.1789734363555908, |
|
"learning_rate": 9.687977388056791e-05, |
|
"loss": 3.7763, |
|
"step": 35950 |
|
}, |
|
{ |
|
"epoch": 0.5806357961968356, |
|
"grad_norm": 2.1416497230529785, |
|
"learning_rate": 9.687092962105758e-05, |
|
"loss": 3.9497, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 0.5806357961968356, |
|
"eval_asr_loss": 0.5691508247039826, |
|
"eval_loss": 4.8229522705078125, |
|
"eval_runtime": 14.7902, |
|
"eval_samples_per_second": 34.617, |
|
"eval_steps_per_second": 11.562, |
|
"eval_tts_loss": 6.016266314813699, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 0.5814422348026644, |
|
"grad_norm": 1.5320308208465576, |
|
"learning_rate": 9.686207324964665e-05, |
|
"loss": 3.7274, |
|
"step": 36050 |
|
}, |
|
{ |
|
"epoch": 0.5822486734084934, |
|
"grad_norm": 2.025629758834839, |
|
"learning_rate": 9.685320476862368e-05, |
|
"loss": 4.0585, |
|
"step": 36100 |
|
}, |
|
{ |
|
"epoch": 0.5830551120143224, |
|
"grad_norm": 1.4545434713363647, |
|
"learning_rate": 9.684432418028038e-05, |
|
"loss": 4.0057, |
|
"step": 36150 |
|
}, |
|
{ |
|
"epoch": 0.5838615506201513, |
|
"grad_norm": 1.8786442279815674, |
|
"learning_rate": 9.683543148691155e-05, |
|
"loss": 3.5916, |
|
"step": 36200 |
|
}, |
|
{ |
|
"epoch": 0.5846679892259802, |
|
"grad_norm": 1.035013198852539, |
|
"learning_rate": 9.682652669081519e-05, |
|
"loss": 4.0424, |
|
"step": 36250 |
|
}, |
|
{ |
|
"epoch": 0.5854744278318091, |
|
"grad_norm": 1.4029967784881592, |
|
"learning_rate": 9.681760979429235e-05, |
|
"loss": 4.2236, |
|
"step": 36300 |
|
}, |
|
{ |
|
"epoch": 0.5862808664376381, |
|
"grad_norm": 3.3041937351226807, |
|
"learning_rate": 9.680868079964726e-05, |
|
"loss": 4.0875, |
|
"step": 36350 |
|
}, |
|
{ |
|
"epoch": 0.5870873050434671, |
|
"grad_norm": 2.505573272705078, |
|
"learning_rate": 9.679973970918727e-05, |
|
"loss": 3.6474, |
|
"step": 36400 |
|
}, |
|
{ |
|
"epoch": 0.587893743649296, |
|
"grad_norm": 2.1446595191955566, |
|
"learning_rate": 9.679078652522283e-05, |
|
"loss": 4.2775, |
|
"step": 36450 |
|
}, |
|
{ |
|
"epoch": 0.5887001822551249, |
|
"grad_norm": 2.0899713039398193, |
|
"learning_rate": 9.678182125006751e-05, |
|
"loss": 3.9854, |
|
"step": 36500 |
|
}, |
|
{ |
|
"epoch": 0.5895066208609538, |
|
"grad_norm": 1.1477386951446533, |
|
"learning_rate": 9.677284388603806e-05, |
|
"loss": 3.8436, |
|
"step": 36550 |
|
}, |
|
{ |
|
"epoch": 0.5903130594667828, |
|
"grad_norm": 3.7815449237823486, |
|
"learning_rate": 9.67638544354543e-05, |
|
"loss": 3.7386, |
|
"step": 36600 |
|
}, |
|
{ |
|
"epoch": 0.5911194980726118, |
|
"grad_norm": 1.9048818349838257, |
|
"learning_rate": 9.675485290063917e-05, |
|
"loss": 4.5235, |
|
"step": 36650 |
|
}, |
|
{ |
|
"epoch": 0.5919259366784406, |
|
"grad_norm": 2.060019016265869, |
|
"learning_rate": 9.67458392839188e-05, |
|
"loss": 3.9605, |
|
"step": 36700 |
|
}, |
|
{ |
|
"epoch": 0.5927323752842696, |
|
"grad_norm": 1.520529866218567, |
|
"learning_rate": 9.673681358762235e-05, |
|
"loss": 3.8785, |
|
"step": 36750 |
|
}, |
|
{ |
|
"epoch": 0.5935388138900985, |
|
"grad_norm": 1.8698012828826904, |
|
"learning_rate": 9.672777581408218e-05, |
|
"loss": 4.1342, |
|
"step": 36800 |
|
}, |
|
{ |
|
"epoch": 0.5943452524959275, |
|
"grad_norm": 2.1991450786590576, |
|
"learning_rate": 9.671872596563372e-05, |
|
"loss": 3.8919, |
|
"step": 36850 |
|
}, |
|
{ |
|
"epoch": 0.5951516911017565, |
|
"grad_norm": 1.6680142879486084, |
|
"learning_rate": 9.670966404461554e-05, |
|
"loss": 3.8034, |
|
"step": 36900 |
|
}, |
|
{ |
|
"epoch": 0.5959581297075853, |
|
"grad_norm": 1.9333608150482178, |
|
"learning_rate": 9.670059005336933e-05, |
|
"loss": 4.1449, |
|
"step": 36950 |
|
}, |
|
{ |
|
"epoch": 0.5967645683134143, |
|
"grad_norm": 0.8060359358787537, |
|
"learning_rate": 9.669150399423991e-05, |
|
"loss": 3.7206, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 0.5975710069192433, |
|
"grad_norm": 2.4317374229431152, |
|
"learning_rate": 9.668240586957518e-05, |
|
"loss": 4.3372, |
|
"step": 37050 |
|
}, |
|
{ |
|
"epoch": 0.5983774455250722, |
|
"grad_norm": 1.0086828470230103, |
|
"learning_rate": 9.66732956817262e-05, |
|
"loss": 4.3951, |
|
"step": 37100 |
|
}, |
|
{ |
|
"epoch": 0.5991838841309011, |
|
"grad_norm": 1.9289790391921997, |
|
"learning_rate": 9.666417343304713e-05, |
|
"loss": 3.6865, |
|
"step": 37150 |
|
}, |
|
{ |
|
"epoch": 0.59999032273673, |
|
"grad_norm": 1.7148078680038452, |
|
"learning_rate": 9.665503912589525e-05, |
|
"loss": 4.1055, |
|
"step": 37200 |
|
}, |
|
{ |
|
"epoch": 0.600796761342559, |
|
"grad_norm": 1.2342246770858765, |
|
"learning_rate": 9.664589276263095e-05, |
|
"loss": 3.9735, |
|
"step": 37250 |
|
}, |
|
{ |
|
"epoch": 0.601603199948388, |
|
"grad_norm": 1.9903031587600708, |
|
"learning_rate": 9.663673434561771e-05, |
|
"loss": 3.8671, |
|
"step": 37300 |
|
}, |
|
{ |
|
"epoch": 0.6024096385542169, |
|
"grad_norm": 1.465268850326538, |
|
"learning_rate": 9.66275638772222e-05, |
|
"loss": 3.9581, |
|
"step": 37350 |
|
}, |
|
{ |
|
"epoch": 0.6032160771600458, |
|
"grad_norm": 1.6464263200759888, |
|
"learning_rate": 9.661838135981412e-05, |
|
"loss": 3.9276, |
|
"step": 37400 |
|
}, |
|
{ |
|
"epoch": 0.6040225157658747, |
|
"grad_norm": 2.538856267929077, |
|
"learning_rate": 9.660918679576636e-05, |
|
"loss": 3.9488, |
|
"step": 37450 |
|
}, |
|
{ |
|
"epoch": 0.6048289543717037, |
|
"grad_norm": 3.353928565979004, |
|
"learning_rate": 9.659998018745484e-05, |
|
"loss": 4.0735, |
|
"step": 37500 |
|
}, |
|
{ |
|
"epoch": 0.6056353929775327, |
|
"grad_norm": 2.7926583290100098, |
|
"learning_rate": 9.659076153725868e-05, |
|
"loss": 4.0669, |
|
"step": 37550 |
|
}, |
|
{ |
|
"epoch": 0.6064418315833615, |
|
"grad_norm": 2.16097354888916, |
|
"learning_rate": 9.658153084756003e-05, |
|
"loss": 3.7335, |
|
"step": 37600 |
|
}, |
|
{ |
|
"epoch": 0.6072482701891905, |
|
"grad_norm": 2.647451639175415, |
|
"learning_rate": 9.657228812074422e-05, |
|
"loss": 3.8995, |
|
"step": 37650 |
|
}, |
|
{ |
|
"epoch": 0.6080547087950194, |
|
"grad_norm": 1.449508786201477, |
|
"learning_rate": 9.656303335919963e-05, |
|
"loss": 3.8566, |
|
"step": 37700 |
|
}, |
|
{ |
|
"epoch": 0.6088611474008484, |
|
"grad_norm": 1.6518259048461914, |
|
"learning_rate": 9.655376656531781e-05, |
|
"loss": 4.1638, |
|
"step": 37750 |
|
}, |
|
{ |
|
"epoch": 0.6096675860066774, |
|
"grad_norm": 2.536465883255005, |
|
"learning_rate": 9.654448774149336e-05, |
|
"loss": 4.4301, |
|
"step": 37800 |
|
}, |
|
{ |
|
"epoch": 0.6104740246125062, |
|
"grad_norm": 1.383885145187378, |
|
"learning_rate": 9.653519689012403e-05, |
|
"loss": 3.7318, |
|
"step": 37850 |
|
}, |
|
{ |
|
"epoch": 0.6112804632183352, |
|
"grad_norm": 2.0474696159362793, |
|
"learning_rate": 9.652589401361067e-05, |
|
"loss": 3.9712, |
|
"step": 37900 |
|
}, |
|
{ |
|
"epoch": 0.6120869018241641, |
|
"grad_norm": 3.0140864849090576, |
|
"learning_rate": 9.651657911435725e-05, |
|
"loss": 3.931, |
|
"step": 37950 |
|
}, |
|
{ |
|
"epoch": 0.6128933404299931, |
|
"grad_norm": 1.272126317024231, |
|
"learning_rate": 9.65072521947708e-05, |
|
"loss": 3.5631, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 0.613699779035822, |
|
"grad_norm": 1.682157039642334, |
|
"learning_rate": 9.649791325726149e-05, |
|
"loss": 3.4141, |
|
"step": 38050 |
|
}, |
|
{ |
|
"epoch": 0.6145062176416509, |
|
"grad_norm": 1.6309934854507446, |
|
"learning_rate": 9.648856230424259e-05, |
|
"loss": 4.2273, |
|
"step": 38100 |
|
}, |
|
{ |
|
"epoch": 0.6153126562474799, |
|
"grad_norm": 1.193287968635559, |
|
"learning_rate": 9.64791993381305e-05, |
|
"loss": 3.9418, |
|
"step": 38150 |
|
}, |
|
{ |
|
"epoch": 0.6161190948533088, |
|
"grad_norm": 2.2340729236602783, |
|
"learning_rate": 9.646982436134469e-05, |
|
"loss": 4.3585, |
|
"step": 38200 |
|
}, |
|
{ |
|
"epoch": 0.6169255334591378, |
|
"grad_norm": 3.2289340496063232, |
|
"learning_rate": 9.646043737630772e-05, |
|
"loss": 4.0445, |
|
"step": 38250 |
|
}, |
|
{ |
|
"epoch": 0.6177319720649667, |
|
"grad_norm": 2.211266040802002, |
|
"learning_rate": 9.645103838544531e-05, |
|
"loss": 3.5903, |
|
"step": 38300 |
|
}, |
|
{ |
|
"epoch": 0.6185384106707956, |
|
"grad_norm": 2.726902484893799, |
|
"learning_rate": 9.644162739118625e-05, |
|
"loss": 3.942, |
|
"step": 38350 |
|
}, |
|
{ |
|
"epoch": 0.6193448492766246, |
|
"grad_norm": 2.515888214111328, |
|
"learning_rate": 9.64322043959624e-05, |
|
"loss": 4.0163, |
|
"step": 38400 |
|
}, |
|
{ |
|
"epoch": 0.6201512878824536, |
|
"grad_norm": 1.7479592561721802, |
|
"learning_rate": 9.642276940220878e-05, |
|
"loss": 4.0748, |
|
"step": 38450 |
|
}, |
|
{ |
|
"epoch": 0.6209577264882824, |
|
"grad_norm": 1.6890915632247925, |
|
"learning_rate": 9.641332241236348e-05, |
|
"loss": 3.924, |
|
"step": 38500 |
|
}, |
|
{ |
|
"epoch": 0.6217641650941114, |
|
"grad_norm": 2.3948464393615723, |
|
"learning_rate": 9.640386342886767e-05, |
|
"loss": 3.9101, |
|
"step": 38550 |
|
}, |
|
{ |
|
"epoch": 0.6225706036999403, |
|
"grad_norm": 2.6629281044006348, |
|
"learning_rate": 9.639439245416568e-05, |
|
"loss": 3.9898, |
|
"step": 38600 |
|
}, |
|
{ |
|
"epoch": 0.6233770423057693, |
|
"grad_norm": 1.5631389617919922, |
|
"learning_rate": 9.638490949070488e-05, |
|
"loss": 3.83, |
|
"step": 38650 |
|
}, |
|
{ |
|
"epoch": 0.6241834809115983, |
|
"grad_norm": 1.956933856010437, |
|
"learning_rate": 9.637541454093574e-05, |
|
"loss": 4.0719, |
|
"step": 38700 |
|
}, |
|
{ |
|
"epoch": 0.6249899195174271, |
|
"grad_norm": 2.43593430519104, |
|
"learning_rate": 9.63659076073119e-05, |
|
"loss": 4.0034, |
|
"step": 38750 |
|
}, |
|
{ |
|
"epoch": 0.6257963581232561, |
|
"grad_norm": 1.7999205589294434, |
|
"learning_rate": 9.635638869229e-05, |
|
"loss": 3.6974, |
|
"step": 38800 |
|
}, |
|
{ |
|
"epoch": 0.626602796729085, |
|
"grad_norm": 3.0806641578674316, |
|
"learning_rate": 9.634685779832983e-05, |
|
"loss": 4.0155, |
|
"step": 38850 |
|
}, |
|
{ |
|
"epoch": 0.627409235334914, |
|
"grad_norm": 2.0375235080718994, |
|
"learning_rate": 9.633731492789428e-05, |
|
"loss": 3.9403, |
|
"step": 38900 |
|
}, |
|
{ |
|
"epoch": 0.6282156739407428, |
|
"grad_norm": 2.0660057067871094, |
|
"learning_rate": 9.632776008344929e-05, |
|
"loss": 3.9845, |
|
"step": 38950 |
|
}, |
|
{ |
|
"epoch": 0.6290221125465718, |
|
"grad_norm": 3.3655974864959717, |
|
"learning_rate": 9.631819326746394e-05, |
|
"loss": 4.0855, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 0.6290221125465718, |
|
"eval_asr_loss": 0.6034199091663585, |
|
"eval_loss": 4.653688430786133, |
|
"eval_runtime": 14.8011, |
|
"eval_samples_per_second": 34.592, |
|
"eval_steps_per_second": 11.553, |
|
"eval_tts_loss": 6.013137326975981, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 0.6298285511524008, |
|
"grad_norm": 2.8556010723114014, |
|
"learning_rate": 9.63086144824104e-05, |
|
"loss": 4.1555, |
|
"step": 39050 |
|
}, |
|
{ |
|
"epoch": 0.6306349897582297, |
|
"grad_norm": 0.7945855259895325, |
|
"learning_rate": 9.62990237307639e-05, |
|
"loss": 3.9282, |
|
"step": 39100 |
|
}, |
|
{ |
|
"epoch": 0.6314414283640587, |
|
"grad_norm": 2.212608575820923, |
|
"learning_rate": 9.628942101500281e-05, |
|
"loss": 3.9476, |
|
"step": 39150 |
|
}, |
|
{ |
|
"epoch": 0.6322478669698876, |
|
"grad_norm": 2.4456074237823486, |
|
"learning_rate": 9.627980633760852e-05, |
|
"loss": 4.2139, |
|
"step": 39200 |
|
}, |
|
{ |
|
"epoch": 0.6330543055757165, |
|
"grad_norm": 2.3700897693634033, |
|
"learning_rate": 9.62701797010656e-05, |
|
"loss": 3.9148, |
|
"step": 39250 |
|
}, |
|
{ |
|
"epoch": 0.6338607441815455, |
|
"grad_norm": 1.0412943363189697, |
|
"learning_rate": 9.626054110786163e-05, |
|
"loss": 3.7921, |
|
"step": 39300 |
|
}, |
|
{ |
|
"epoch": 0.6346671827873744, |
|
"grad_norm": 1.5326954126358032, |
|
"learning_rate": 9.625089056048736e-05, |
|
"loss": 4.3737, |
|
"step": 39350 |
|
}, |
|
{ |
|
"epoch": 0.6354736213932033, |
|
"grad_norm": 1.076934814453125, |
|
"learning_rate": 9.624122806143652e-05, |
|
"loss": 3.9424, |
|
"step": 39400 |
|
}, |
|
{ |
|
"epoch": 0.6362800599990323, |
|
"grad_norm": 1.8640769720077515, |
|
"learning_rate": 9.623155361320607e-05, |
|
"loss": 4.3091, |
|
"step": 39450 |
|
}, |
|
{ |
|
"epoch": 0.6370864986048612, |
|
"grad_norm": 2.431981325149536, |
|
"learning_rate": 9.62218672182959e-05, |
|
"loss": 3.7956, |
|
"step": 39500 |
|
}, |
|
{ |
|
"epoch": 0.6378929372106902, |
|
"grad_norm": 3.0567352771759033, |
|
"learning_rate": 9.621216887920913e-05, |
|
"loss": 3.9352, |
|
"step": 39550 |
|
}, |
|
{ |
|
"epoch": 0.6386993758165191, |
|
"grad_norm": 1.1013330221176147, |
|
"learning_rate": 9.620245859845188e-05, |
|
"loss": 4.0237, |
|
"step": 39600 |
|
}, |
|
{ |
|
"epoch": 0.639505814422348, |
|
"grad_norm": 2.7281951904296875, |
|
"learning_rate": 9.619273637853339e-05, |
|
"loss": 4.3155, |
|
"step": 39650 |
|
}, |
|
{ |
|
"epoch": 0.640312253028177, |
|
"grad_norm": 1.6277055740356445, |
|
"learning_rate": 9.618300222196596e-05, |
|
"loss": 4.0474, |
|
"step": 39700 |
|
}, |
|
{ |
|
"epoch": 0.6411186916340059, |
|
"grad_norm": 1.7232165336608887, |
|
"learning_rate": 9.617325613126502e-05, |
|
"loss": 3.7625, |
|
"step": 39750 |
|
}, |
|
{ |
|
"epoch": 0.6419251302398349, |
|
"grad_norm": 1.3456089496612549, |
|
"learning_rate": 9.6163498108949e-05, |
|
"loss": 3.8336, |
|
"step": 39800 |
|
}, |
|
{ |
|
"epoch": 0.6427315688456637, |
|
"grad_norm": 3.376032590866089, |
|
"learning_rate": 9.615372815753954e-05, |
|
"loss": 3.9334, |
|
"step": 39850 |
|
}, |
|
{ |
|
"epoch": 0.6435380074514927, |
|
"grad_norm": 3.418212890625, |
|
"learning_rate": 9.614394627956121e-05, |
|
"loss": 4.2293, |
|
"step": 39900 |
|
}, |
|
{ |
|
"epoch": 0.6443444460573217, |
|
"grad_norm": 1.5658949613571167, |
|
"learning_rate": 9.613415247754181e-05, |
|
"loss": 4.0883, |
|
"step": 39950 |
|
}, |
|
{ |
|
"epoch": 0.6451508846631506, |
|
"grad_norm": 1.2223689556121826, |
|
"learning_rate": 9.612434675401212e-05, |
|
"loss": 3.9809, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 0.6459573232689795, |
|
"grad_norm": 1.4119608402252197, |
|
"learning_rate": 9.611452911150603e-05, |
|
"loss": 4.135, |
|
"step": 40050 |
|
}, |
|
{ |
|
"epoch": 0.6467637618748084, |
|
"grad_norm": 1.1455477476119995, |
|
"learning_rate": 9.610469955256052e-05, |
|
"loss": 4.0727, |
|
"step": 40100 |
|
}, |
|
{ |
|
"epoch": 0.6475702004806374, |
|
"grad_norm": 1.1804072856903076, |
|
"learning_rate": 9.609485807971566e-05, |
|
"loss": 3.7652, |
|
"step": 40150 |
|
}, |
|
{ |
|
"epoch": 0.6483766390864664, |
|
"grad_norm": 2.133227586746216, |
|
"learning_rate": 9.608500469551455e-05, |
|
"loss": 3.6933, |
|
"step": 40200 |
|
}, |
|
{ |
|
"epoch": 0.6491830776922953, |
|
"grad_norm": 1.6320295333862305, |
|
"learning_rate": 9.60751394025034e-05, |
|
"loss": 3.7121, |
|
"step": 40250 |
|
}, |
|
{ |
|
"epoch": 0.6499895162981242, |
|
"grad_norm": 2.7744624614715576, |
|
"learning_rate": 9.606526220323155e-05, |
|
"loss": 4.1106, |
|
"step": 40300 |
|
}, |
|
{ |
|
"epoch": 0.6507959549039531, |
|
"grad_norm": 2.890082836151123, |
|
"learning_rate": 9.605537310025131e-05, |
|
"loss": 3.8397, |
|
"step": 40350 |
|
}, |
|
{ |
|
"epoch": 0.6516023935097821, |
|
"grad_norm": 1.4252163171768188, |
|
"learning_rate": 9.604547209611815e-05, |
|
"loss": 3.9493, |
|
"step": 40400 |
|
}, |
|
{ |
|
"epoch": 0.6524088321156111, |
|
"grad_norm": 1.658120036125183, |
|
"learning_rate": 9.603555919339055e-05, |
|
"loss": 4.0099, |
|
"step": 40450 |
|
}, |
|
{ |
|
"epoch": 0.6532152707214399, |
|
"grad_norm": 2.3012828826904297, |
|
"learning_rate": 9.602563439463012e-05, |
|
"loss": 3.8907, |
|
"step": 40500 |
|
}, |
|
{ |
|
"epoch": 0.6540217093272689, |
|
"grad_norm": 2.3004066944122314, |
|
"learning_rate": 9.601569770240154e-05, |
|
"loss": 4.2203, |
|
"step": 40550 |
|
}, |
|
{ |
|
"epoch": 0.6548281479330978, |
|
"grad_norm": 0.95237135887146, |
|
"learning_rate": 9.600574911927253e-05, |
|
"loss": 4.0099, |
|
"step": 40600 |
|
}, |
|
{ |
|
"epoch": 0.6556345865389268, |
|
"grad_norm": 2.167707681655884, |
|
"learning_rate": 9.599578864781389e-05, |
|
"loss": 3.4495, |
|
"step": 40650 |
|
}, |
|
{ |
|
"epoch": 0.6564410251447558, |
|
"grad_norm": 1.9729243516921997, |
|
"learning_rate": 9.598581629059952e-05, |
|
"loss": 3.9483, |
|
"step": 40700 |
|
}, |
|
{ |
|
"epoch": 0.6572474637505846, |
|
"grad_norm": 0.6504325270652771, |
|
"learning_rate": 9.597583205020636e-05, |
|
"loss": 3.859, |
|
"step": 40750 |
|
}, |
|
{ |
|
"epoch": 0.6580539023564136, |
|
"grad_norm": 2.181685447692871, |
|
"learning_rate": 9.596583592921446e-05, |
|
"loss": 3.6357, |
|
"step": 40800 |
|
}, |
|
{ |
|
"epoch": 0.6588603409622426, |
|
"grad_norm": 1.3120620250701904, |
|
"learning_rate": 9.595582793020688e-05, |
|
"loss": 3.7685, |
|
"step": 40850 |
|
}, |
|
{ |
|
"epoch": 0.6596667795680715, |
|
"grad_norm": 3.323073148727417, |
|
"learning_rate": 9.59458080557698e-05, |
|
"loss": 4.3613, |
|
"step": 40900 |
|
}, |
|
{ |
|
"epoch": 0.6604732181739004, |
|
"grad_norm": 1.2107150554656982, |
|
"learning_rate": 9.593577630849246e-05, |
|
"loss": 4.0395, |
|
"step": 40950 |
|
}, |
|
{ |
|
"epoch": 0.6612796567797293, |
|
"grad_norm": 2.1152050495147705, |
|
"learning_rate": 9.592573269096716e-05, |
|
"loss": 3.6636, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 0.6620860953855583, |
|
"grad_norm": 1.9166638851165771, |
|
"learning_rate": 9.591567720578924e-05, |
|
"loss": 3.8332, |
|
"step": 41050 |
|
}, |
|
{ |
|
"epoch": 0.6628925339913873, |
|
"grad_norm": 1.2429214715957642, |
|
"learning_rate": 9.590560985555716e-05, |
|
"loss": 4.0641, |
|
"step": 41100 |
|
}, |
|
{ |
|
"epoch": 0.6636989725972162, |
|
"grad_norm": 2.569197177886963, |
|
"learning_rate": 9.58955306428724e-05, |
|
"loss": 4.026, |
|
"step": 41150 |
|
}, |
|
{ |
|
"epoch": 0.6645054112030451, |
|
"grad_norm": 1.7593148946762085, |
|
"learning_rate": 9.588543957033957e-05, |
|
"loss": 3.88, |
|
"step": 41200 |
|
}, |
|
{ |
|
"epoch": 0.665311849808874, |
|
"grad_norm": 1.2028878927230835, |
|
"learning_rate": 9.587553881534579e-05, |
|
"loss": 4.0756, |
|
"step": 41250 |
|
}, |
|
{ |
|
"epoch": 0.666118288414703, |
|
"grad_norm": 2.990448474884033, |
|
"learning_rate": 9.586542426800969e-05, |
|
"loss": 4.2531, |
|
"step": 41300 |
|
}, |
|
{ |
|
"epoch": 0.666924727020532, |
|
"grad_norm": 2.4063851833343506, |
|
"learning_rate": 9.585529786860527e-05, |
|
"loss": 3.9218, |
|
"step": 41350 |
|
}, |
|
{ |
|
"epoch": 0.6677311656263608, |
|
"grad_norm": 1.7430907487869263, |
|
"learning_rate": 9.584515961974928e-05, |
|
"loss": 4.3134, |
|
"step": 41400 |
|
}, |
|
{ |
|
"epoch": 0.6685376042321898, |
|
"grad_norm": 1.3421581983566284, |
|
"learning_rate": 9.583500952406155e-05, |
|
"loss": 4.0289, |
|
"step": 41450 |
|
}, |
|
{ |
|
"epoch": 0.6693440428380187, |
|
"grad_norm": 0.871703565120697, |
|
"learning_rate": 9.582484758416497e-05, |
|
"loss": 3.7478, |
|
"step": 41500 |
|
}, |
|
{ |
|
"epoch": 0.6701504814438477, |
|
"grad_norm": 1.735887885093689, |
|
"learning_rate": 9.581467380268547e-05, |
|
"loss": 3.9596, |
|
"step": 41550 |
|
}, |
|
{ |
|
"epoch": 0.6709569200496767, |
|
"grad_norm": 1.485318660736084, |
|
"learning_rate": 9.580448818225205e-05, |
|
"loss": 4.1336, |
|
"step": 41600 |
|
}, |
|
{ |
|
"epoch": 0.6717633586555055, |
|
"grad_norm": 1.8921575546264648, |
|
"learning_rate": 9.579429072549679e-05, |
|
"loss": 3.8069, |
|
"step": 41650 |
|
}, |
|
{ |
|
"epoch": 0.6725697972613345, |
|
"grad_norm": 1.4174717664718628, |
|
"learning_rate": 9.57840814350548e-05, |
|
"loss": 3.7202, |
|
"step": 41700 |
|
}, |
|
{ |
|
"epoch": 0.6733762358671634, |
|
"grad_norm": 2.4129083156585693, |
|
"learning_rate": 9.577386031356425e-05, |
|
"loss": 3.9037, |
|
"step": 41750 |
|
}, |
|
{ |
|
"epoch": 0.6741826744729924, |
|
"grad_norm": 3.722903251647949, |
|
"learning_rate": 9.57636273636664e-05, |
|
"loss": 3.971, |
|
"step": 41800 |
|
}, |
|
{ |
|
"epoch": 0.6749891130788213, |
|
"grad_norm": 1.9005436897277832, |
|
"learning_rate": 9.57535875993941e-05, |
|
"loss": 3.9043, |
|
"step": 41850 |
|
}, |
|
{ |
|
"epoch": 0.6757955516846502, |
|
"grad_norm": 2.620349645614624, |
|
"learning_rate": 9.574333123705391e-05, |
|
"loss": 4.0456, |
|
"step": 41900 |
|
}, |
|
{ |
|
"epoch": 0.6766019902904792, |
|
"grad_norm": 1.5120000839233398, |
|
"learning_rate": 9.573306305419541e-05, |
|
"loss": 3.9113, |
|
"step": 41950 |
|
}, |
|
{ |
|
"epoch": 0.6774084288963081, |
|
"grad_norm": 2.184945821762085, |
|
"learning_rate": 9.572278305347202e-05, |
|
"loss": 3.9566, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 0.6774084288963081, |
|
"eval_asr_loss": 0.6747458191937141, |
|
"eval_loss": 4.402849197387695, |
|
"eval_runtime": 14.8128, |
|
"eval_samples_per_second": 34.565, |
|
"eval_steps_per_second": 11.544, |
|
"eval_tts_loss": 6.006661847860199, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 0.6782148675021371, |
|
"grad_norm": 1.9270944595336914, |
|
"learning_rate": 9.571249123754014e-05, |
|
"loss": 4.0085, |
|
"step": 42050 |
|
}, |
|
{ |
|
"epoch": 0.679021306107966, |
|
"grad_norm": 2.115370035171509, |
|
"learning_rate": 9.570218760905934e-05, |
|
"loss": 3.9551, |
|
"step": 42100 |
|
}, |
|
{ |
|
"epoch": 0.6798277447137949, |
|
"grad_norm": 2.4913384914398193, |
|
"learning_rate": 9.569187217069213e-05, |
|
"loss": 3.9498, |
|
"step": 42150 |
|
}, |
|
{ |
|
"epoch": 0.6806341833196239, |
|
"grad_norm": 1.2166416645050049, |
|
"learning_rate": 9.568154492510414e-05, |
|
"loss": 3.9295, |
|
"step": 42200 |
|
}, |
|
{ |
|
"epoch": 0.6814406219254528, |
|
"grad_norm": 3.2959489822387695, |
|
"learning_rate": 9.567120587496402e-05, |
|
"loss": 3.8898, |
|
"step": 42250 |
|
}, |
|
{ |
|
"epoch": 0.6822470605312817, |
|
"grad_norm": 3.9133834838867188, |
|
"learning_rate": 9.566085502294348e-05, |
|
"loss": 4.1996, |
|
"step": 42300 |
|
}, |
|
{ |
|
"epoch": 0.6830534991371107, |
|
"grad_norm": 1.421493411064148, |
|
"learning_rate": 9.56504923717173e-05, |
|
"loss": 3.9401, |
|
"step": 42350 |
|
}, |
|
{ |
|
"epoch": 0.6838599377429396, |
|
"grad_norm": 1.4366188049316406, |
|
"learning_rate": 9.564011792396326e-05, |
|
"loss": 4.1854, |
|
"step": 42400 |
|
}, |
|
{ |
|
"epoch": 0.6846663763487686, |
|
"grad_norm": 2.0820562839508057, |
|
"learning_rate": 9.562973168236224e-05, |
|
"loss": 4.1061, |
|
"step": 42450 |
|
}, |
|
{ |
|
"epoch": 0.6854728149545976, |
|
"grad_norm": 1.6290011405944824, |
|
"learning_rate": 9.561933364959814e-05, |
|
"loss": 4.3134, |
|
"step": 42500 |
|
}, |
|
{ |
|
"epoch": 0.6862792535604264, |
|
"grad_norm": 2.4670650959014893, |
|
"learning_rate": 9.560892382835791e-05, |
|
"loss": 3.7429, |
|
"step": 42550 |
|
}, |
|
{ |
|
"epoch": 0.6870856921662554, |
|
"grad_norm": 1.2104759216308594, |
|
"learning_rate": 9.559850222133155e-05, |
|
"loss": 3.6463, |
|
"step": 42600 |
|
}, |
|
{ |
|
"epoch": 0.6878921307720843, |
|
"grad_norm": 1.4174095392227173, |
|
"learning_rate": 9.558806883121209e-05, |
|
"loss": 4.0535, |
|
"step": 42650 |
|
}, |
|
{ |
|
"epoch": 0.6886985693779133, |
|
"grad_norm": 2.343122959136963, |
|
"learning_rate": 9.557762366069562e-05, |
|
"loss": 4.2928, |
|
"step": 42700 |
|
}, |
|
{ |
|
"epoch": 0.6895050079837421, |
|
"grad_norm": 2.328246831893921, |
|
"learning_rate": 9.55671667124813e-05, |
|
"loss": 4.1672, |
|
"step": 42750 |
|
}, |
|
{ |
|
"epoch": 0.6903114465895711, |
|
"grad_norm": 2.1924808025360107, |
|
"learning_rate": 9.55566979892713e-05, |
|
"loss": 4.0338, |
|
"step": 42800 |
|
}, |
|
{ |
|
"epoch": 0.6911178851954001, |
|
"grad_norm": 1.4481010437011719, |
|
"learning_rate": 9.554621749377079e-05, |
|
"loss": 3.8551, |
|
"step": 42850 |
|
}, |
|
{ |
|
"epoch": 0.691924323801229, |
|
"grad_norm": 2.5818378925323486, |
|
"learning_rate": 9.55357252286881e-05, |
|
"loss": 3.8865, |
|
"step": 42900 |
|
}, |
|
{ |
|
"epoch": 0.692730762407058, |
|
"grad_norm": 0.6974899768829346, |
|
"learning_rate": 9.552522119673449e-05, |
|
"loss": 4.3369, |
|
"step": 42950 |
|
}, |
|
{ |
|
"epoch": 0.6935372010128869, |
|
"grad_norm": 2.877656936645508, |
|
"learning_rate": 9.551470540062433e-05, |
|
"loss": 3.6878, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 0.6943436396187158, |
|
"grad_norm": 1.5497684478759766, |
|
"learning_rate": 9.550417784307497e-05, |
|
"loss": 3.6174, |
|
"step": 43050 |
|
}, |
|
{ |
|
"epoch": 0.6951500782245448, |
|
"grad_norm": 2.261922836303711, |
|
"learning_rate": 9.549363852680686e-05, |
|
"loss": 3.9077, |
|
"step": 43100 |
|
}, |
|
{ |
|
"epoch": 0.6959565168303737, |
|
"grad_norm": 2.5778346061706543, |
|
"learning_rate": 9.548308745454344e-05, |
|
"loss": 3.8163, |
|
"step": 43150 |
|
}, |
|
{ |
|
"epoch": 0.6967629554362026, |
|
"grad_norm": 2.7817728519439697, |
|
"learning_rate": 9.547252462901124e-05, |
|
"loss": 3.6548, |
|
"step": 43200 |
|
}, |
|
{ |
|
"epoch": 0.6975693940420316, |
|
"grad_norm": 1.0119379758834839, |
|
"learning_rate": 9.546195005293976e-05, |
|
"loss": 4.1415, |
|
"step": 43250 |
|
}, |
|
{ |
|
"epoch": 0.6983758326478605, |
|
"grad_norm": 3.6960830688476562, |
|
"learning_rate": 9.54513637290616e-05, |
|
"loss": 4.0043, |
|
"step": 43300 |
|
}, |
|
{ |
|
"epoch": 0.6991822712536895, |
|
"grad_norm": 1.3119101524353027, |
|
"learning_rate": 9.544076566011235e-05, |
|
"loss": 4.0963, |
|
"step": 43350 |
|
}, |
|
{ |
|
"epoch": 0.6999887098595184, |
|
"grad_norm": 1.1128474473953247, |
|
"learning_rate": 9.543015584883067e-05, |
|
"loss": 3.7204, |
|
"step": 43400 |
|
}, |
|
{ |
|
"epoch": 0.7007951484653473, |
|
"grad_norm": 2.3756420612335205, |
|
"learning_rate": 9.541953429795823e-05, |
|
"loss": 4.3295, |
|
"step": 43450 |
|
}, |
|
{ |
|
"epoch": 0.7016015870711763, |
|
"grad_norm": 1.1460477113723755, |
|
"learning_rate": 9.540890101023972e-05, |
|
"loss": 3.8782, |
|
"step": 43500 |
|
}, |
|
{ |
|
"epoch": 0.7024080256770052, |
|
"grad_norm": 1.0476806163787842, |
|
"learning_rate": 9.539825598842291e-05, |
|
"loss": 3.8376, |
|
"step": 43550 |
|
}, |
|
{ |
|
"epoch": 0.7032144642828342, |
|
"grad_norm": 1.2808899879455566, |
|
"learning_rate": 9.538759923525857e-05, |
|
"loss": 4.1479, |
|
"step": 43600 |
|
}, |
|
{ |
|
"epoch": 0.704020902888663, |
|
"grad_norm": 2.6137540340423584, |
|
"learning_rate": 9.537693075350052e-05, |
|
"loss": 3.8806, |
|
"step": 43650 |
|
}, |
|
{ |
|
"epoch": 0.704827341494492, |
|
"grad_norm": 1.4536105394363403, |
|
"learning_rate": 9.536625054590559e-05, |
|
"loss": 3.9868, |
|
"step": 43700 |
|
}, |
|
{ |
|
"epoch": 0.705633780100321, |
|
"grad_norm": 1.8267903327941895, |
|
"learning_rate": 9.535555861523363e-05, |
|
"loss": 4.0424, |
|
"step": 43750 |
|
}, |
|
{ |
|
"epoch": 0.7064402187061499, |
|
"grad_norm": 1.1175280809402466, |
|
"learning_rate": 9.534485496424757e-05, |
|
"loss": 4.0311, |
|
"step": 43800 |
|
}, |
|
{ |
|
"epoch": 0.7072466573119789, |
|
"grad_norm": 2.639409065246582, |
|
"learning_rate": 9.533413959571332e-05, |
|
"loss": 3.7509, |
|
"step": 43850 |
|
}, |
|
{ |
|
"epoch": 0.7080530959178077, |
|
"grad_norm": 2.0124926567077637, |
|
"learning_rate": 9.532341251239982e-05, |
|
"loss": 4.1953, |
|
"step": 43900 |
|
}, |
|
{ |
|
"epoch": 0.7088595345236367, |
|
"grad_norm": 1.6869808435440063, |
|
"learning_rate": 9.531267371707906e-05, |
|
"loss": 3.5422, |
|
"step": 43950 |
|
}, |
|
{ |
|
"epoch": 0.7096659731294657, |
|
"grad_norm": 2.89847469329834, |
|
"learning_rate": 9.530192321252607e-05, |
|
"loss": 3.8341, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 0.7104724117352946, |
|
"grad_norm": 1.6155884265899658, |
|
"learning_rate": 9.529116100151887e-05, |
|
"loss": 4.1777, |
|
"step": 44050 |
|
}, |
|
{ |
|
"epoch": 0.7112788503411235, |
|
"grad_norm": 2.9309215545654297, |
|
"learning_rate": 9.52803870868385e-05, |
|
"loss": 3.8159, |
|
"step": 44100 |
|
}, |
|
{ |
|
"epoch": 0.7120852889469524, |
|
"grad_norm": 2.221287727355957, |
|
"learning_rate": 9.526960147126908e-05, |
|
"loss": 4.1738, |
|
"step": 44150 |
|
}, |
|
{ |
|
"epoch": 0.7128917275527814, |
|
"grad_norm": 2.1958932876586914, |
|
"learning_rate": 9.525880415759772e-05, |
|
"loss": 4.1857, |
|
"step": 44200 |
|
}, |
|
{ |
|
"epoch": 0.7136981661586104, |
|
"grad_norm": 1.2787940502166748, |
|
"learning_rate": 9.524799514861449e-05, |
|
"loss": 3.5715, |
|
"step": 44250 |
|
}, |
|
{ |
|
"epoch": 0.7145046047644393, |
|
"grad_norm": 1.3575184345245361, |
|
"learning_rate": 9.52371744471126e-05, |
|
"loss": 3.6301, |
|
"step": 44300 |
|
}, |
|
{ |
|
"epoch": 0.7153110433702682, |
|
"grad_norm": 1.1115514039993286, |
|
"learning_rate": 9.522634205588821e-05, |
|
"loss": 4.0579, |
|
"step": 44350 |
|
}, |
|
{ |
|
"epoch": 0.7161174819760971, |
|
"grad_norm": 2.445248603820801, |
|
"learning_rate": 9.52154979777405e-05, |
|
"loss": 3.9905, |
|
"step": 44400 |
|
}, |
|
{ |
|
"epoch": 0.7169239205819261, |
|
"grad_norm": 4.455885410308838, |
|
"learning_rate": 9.52046422154717e-05, |
|
"loss": 3.8643, |
|
"step": 44450 |
|
}, |
|
{ |
|
"epoch": 0.7177303591877551, |
|
"grad_norm": 2.097954034805298, |
|
"learning_rate": 9.519377477188704e-05, |
|
"loss": 4.2417, |
|
"step": 44500 |
|
}, |
|
{ |
|
"epoch": 0.7185367977935839, |
|
"grad_norm": 1.8848130702972412, |
|
"learning_rate": 9.518289564979475e-05, |
|
"loss": 3.7867, |
|
"step": 44550 |
|
}, |
|
{ |
|
"epoch": 0.7193432363994129, |
|
"grad_norm": 2.098465919494629, |
|
"learning_rate": 9.517200485200614e-05, |
|
"loss": 3.9569, |
|
"step": 44600 |
|
}, |
|
{ |
|
"epoch": 0.7201496750052419, |
|
"grad_norm": 2.0162923336029053, |
|
"learning_rate": 9.516110238133548e-05, |
|
"loss": 3.6984, |
|
"step": 44650 |
|
}, |
|
{ |
|
"epoch": 0.7209561136110708, |
|
"grad_norm": 1.9503788948059082, |
|
"learning_rate": 9.515018824060006e-05, |
|
"loss": 3.7651, |
|
"step": 44700 |
|
}, |
|
{ |
|
"epoch": 0.7217625522168998, |
|
"grad_norm": 2.917295455932617, |
|
"learning_rate": 9.513926243262022e-05, |
|
"loss": 3.5928, |
|
"step": 44750 |
|
}, |
|
{ |
|
"epoch": 0.7225689908227286, |
|
"grad_norm": 1.6340333223342896, |
|
"learning_rate": 9.512832496021928e-05, |
|
"loss": 3.9921, |
|
"step": 44800 |
|
}, |
|
{ |
|
"epoch": 0.7233754294285576, |
|
"grad_norm": 2.323976755142212, |
|
"learning_rate": 9.51173758262236e-05, |
|
"loss": 3.8905, |
|
"step": 44850 |
|
}, |
|
{ |
|
"epoch": 0.7241818680343866, |
|
"grad_norm": 2.567837715148926, |
|
"learning_rate": 9.510641503346252e-05, |
|
"loss": 3.973, |
|
"step": 44900 |
|
}, |
|
{ |
|
"epoch": 0.7249883066402155, |
|
"grad_norm": 1.992812156677246, |
|
"learning_rate": 9.509544258476845e-05, |
|
"loss": 4.4189, |
|
"step": 44950 |
|
}, |
|
{ |
|
"epoch": 0.7257947452460444, |
|
"grad_norm": 2.9449970722198486, |
|
"learning_rate": 9.508445848297674e-05, |
|
"loss": 3.6855, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 0.7257947452460444, |
|
"eval_asr_loss": 0.7677889460602478, |
|
"eval_loss": 4.220231056213379, |
|
"eval_runtime": 14.4142, |
|
"eval_samples_per_second": 35.52, |
|
"eval_steps_per_second": 11.863, |
|
"eval_tts_loss": 5.966565507839665, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 0.7266011838518733, |
|
"grad_norm": 2.878650426864624, |
|
"learning_rate": 9.50734627309258e-05, |
|
"loss": 4.118, |
|
"step": 45050 |
|
}, |
|
{ |
|
"epoch": 0.7274076224577023, |
|
"grad_norm": 1.361701250076294, |
|
"learning_rate": 9.506245533145707e-05, |
|
"loss": 4.3712, |
|
"step": 45100 |
|
}, |
|
{ |
|
"epoch": 0.7282140610635313, |
|
"grad_norm": 2.126094341278076, |
|
"learning_rate": 9.505143628741492e-05, |
|
"loss": 3.7077, |
|
"step": 45150 |
|
}, |
|
{ |
|
"epoch": 0.7290204996693602, |
|
"grad_norm": 2.355354070663452, |
|
"learning_rate": 9.50404056016468e-05, |
|
"loss": 3.8391, |
|
"step": 45200 |
|
}, |
|
{ |
|
"epoch": 0.7298269382751891, |
|
"grad_norm": 1.933937668800354, |
|
"learning_rate": 9.502936327700315e-05, |
|
"loss": 3.7756, |
|
"step": 45250 |
|
}, |
|
{ |
|
"epoch": 0.730633376881018, |
|
"grad_norm": 2.8222224712371826, |
|
"learning_rate": 9.50183093163374e-05, |
|
"loss": 4.0453, |
|
"step": 45300 |
|
}, |
|
{ |
|
"epoch": 0.731439815486847, |
|
"grad_norm": 0.3452666699886322, |
|
"learning_rate": 9.500724372250602e-05, |
|
"loss": 3.9045, |
|
"step": 45350 |
|
}, |
|
{ |
|
"epoch": 0.732246254092676, |
|
"grad_norm": 0.9124054312705994, |
|
"learning_rate": 9.499616649836845e-05, |
|
"loss": 3.9615, |
|
"step": 45400 |
|
}, |
|
{ |
|
"epoch": 0.7330526926985048, |
|
"grad_norm": 2.117830276489258, |
|
"learning_rate": 9.498507764678715e-05, |
|
"loss": 4.0645, |
|
"step": 45450 |
|
}, |
|
{ |
|
"epoch": 0.7338591313043338, |
|
"grad_norm": 2.6719260215759277, |
|
"learning_rate": 9.497397717062759e-05, |
|
"loss": 3.7295, |
|
"step": 45500 |
|
}, |
|
{ |
|
"epoch": 0.7346655699101627, |
|
"grad_norm": 2.009535312652588, |
|
"learning_rate": 9.496286507275825e-05, |
|
"loss": 4.3494, |
|
"step": 45550 |
|
}, |
|
{ |
|
"epoch": 0.7354720085159917, |
|
"grad_norm": 1.6704541444778442, |
|
"learning_rate": 9.495174135605058e-05, |
|
"loss": 3.5423, |
|
"step": 45600 |
|
}, |
|
{ |
|
"epoch": 0.7362784471218207, |
|
"grad_norm": 2.1379082202911377, |
|
"learning_rate": 9.494060602337909e-05, |
|
"loss": 4.2106, |
|
"step": 45650 |
|
}, |
|
{ |
|
"epoch": 0.7370848857276495, |
|
"grad_norm": 1.076125979423523, |
|
"learning_rate": 9.492945907762123e-05, |
|
"loss": 4.1118, |
|
"step": 45700 |
|
}, |
|
{ |
|
"epoch": 0.7378913243334785, |
|
"grad_norm": 2.360368013381958, |
|
"learning_rate": 9.491830052165749e-05, |
|
"loss": 3.9588, |
|
"step": 45750 |
|
}, |
|
{ |
|
"epoch": 0.7386977629393074, |
|
"grad_norm": 1.9395514726638794, |
|
"learning_rate": 9.490713035837133e-05, |
|
"loss": 3.9243, |
|
"step": 45800 |
|
}, |
|
{ |
|
"epoch": 0.7395042015451364, |
|
"grad_norm": 1.406371831893921, |
|
"learning_rate": 9.489594859064926e-05, |
|
"loss": 4.1377, |
|
"step": 45850 |
|
}, |
|
{ |
|
"epoch": 0.7403106401509653, |
|
"grad_norm": 1.5525023937225342, |
|
"learning_rate": 9.488475522138073e-05, |
|
"loss": 3.6349, |
|
"step": 45900 |
|
}, |
|
{ |
|
"epoch": 0.7411170787567942, |
|
"grad_norm": 1.3962270021438599, |
|
"learning_rate": 9.487355025345823e-05, |
|
"loss": 3.8225, |
|
"step": 45950 |
|
}, |
|
{ |
|
"epoch": 0.7419235173626232, |
|
"grad_norm": 2.4570751190185547, |
|
"learning_rate": 9.486255813467055e-05, |
|
"loss": 3.8946, |
|
"step": 46000 |
|
}, |
|
{ |
|
"epoch": 0.7427299559684521, |
|
"grad_norm": 9.147421836853027, |
|
"learning_rate": 9.485133020995828e-05, |
|
"loss": 3.9156, |
|
"step": 46050 |
|
}, |
|
{ |
|
"epoch": 0.7435363945742811, |
|
"grad_norm": 1.4232938289642334, |
|
"learning_rate": 9.484009069522938e-05, |
|
"loss": 4.113, |
|
"step": 46100 |
|
}, |
|
{ |
|
"epoch": 0.74434283318011, |
|
"grad_norm": 3.1398966312408447, |
|
"learning_rate": 9.482883959338827e-05, |
|
"loss": 4.0905, |
|
"step": 46150 |
|
}, |
|
{ |
|
"epoch": 0.7451492717859389, |
|
"grad_norm": 2.162477493286133, |
|
"learning_rate": 9.481757690734232e-05, |
|
"loss": 3.9378, |
|
"step": 46200 |
|
}, |
|
{ |
|
"epoch": 0.7459557103917679, |
|
"grad_norm": 2.1014811992645264, |
|
"learning_rate": 9.48063026400019e-05, |
|
"loss": 3.9758, |
|
"step": 46250 |
|
}, |
|
{ |
|
"epoch": 0.7467621489975969, |
|
"grad_norm": 1.6452032327651978, |
|
"learning_rate": 9.479501679428042e-05, |
|
"loss": 4.0847, |
|
"step": 46300 |
|
}, |
|
{ |
|
"epoch": 0.7475685876034257, |
|
"grad_norm": 2.0140156745910645, |
|
"learning_rate": 9.478371937309425e-05, |
|
"loss": 4.3145, |
|
"step": 46350 |
|
}, |
|
{ |
|
"epoch": 0.7483750262092547, |
|
"grad_norm": 1.8627792596817017, |
|
"learning_rate": 9.477241037936273e-05, |
|
"loss": 4.0373, |
|
"step": 46400 |
|
}, |
|
{ |
|
"epoch": 0.7491814648150836, |
|
"grad_norm": 1.3735127449035645, |
|
"learning_rate": 9.476108981600822e-05, |
|
"loss": 3.9577, |
|
"step": 46450 |
|
}, |
|
{ |
|
"epoch": 0.7499879034209126, |
|
"grad_norm": 2.6094725131988525, |
|
"learning_rate": 9.474975768595609e-05, |
|
"loss": 3.957, |
|
"step": 46500 |
|
}, |
|
{ |
|
"epoch": 0.7507943420267416, |
|
"grad_norm": 1.3980075120925903, |
|
"learning_rate": 9.473841399213463e-05, |
|
"loss": 4.1313, |
|
"step": 46550 |
|
}, |
|
{ |
|
"epoch": 0.7516007806325704, |
|
"grad_norm": 3.6238787174224854, |
|
"learning_rate": 9.47270587374752e-05, |
|
"loss": 3.9414, |
|
"step": 46600 |
|
}, |
|
{ |
|
"epoch": 0.7524072192383994, |
|
"grad_norm": 1.450015902519226, |
|
"learning_rate": 9.471569192491208e-05, |
|
"loss": 3.7486, |
|
"step": 46650 |
|
}, |
|
{ |
|
"epoch": 0.7532136578442283, |
|
"grad_norm": 1.9023935794830322, |
|
"learning_rate": 9.470431355738257e-05, |
|
"loss": 4.0542, |
|
"step": 46700 |
|
}, |
|
{ |
|
"epoch": 0.7540200964500573, |
|
"grad_norm": 1.0241719484329224, |
|
"learning_rate": 9.469292363782696e-05, |
|
"loss": 3.9859, |
|
"step": 46750 |
|
}, |
|
{ |
|
"epoch": 0.7548265350558861, |
|
"grad_norm": 1.9951140880584717, |
|
"learning_rate": 9.468152216918849e-05, |
|
"loss": 3.5983, |
|
"step": 46800 |
|
}, |
|
{ |
|
"epoch": 0.7556329736617151, |
|
"grad_norm": 1.519264817237854, |
|
"learning_rate": 9.467010915441343e-05, |
|
"loss": 3.971, |
|
"step": 46850 |
|
}, |
|
{ |
|
"epoch": 0.7564394122675441, |
|
"grad_norm": 2.4989826679229736, |
|
"learning_rate": 9.465868459645102e-05, |
|
"loss": 3.9566, |
|
"step": 46900 |
|
}, |
|
{ |
|
"epoch": 0.757245850873373, |
|
"grad_norm": 2.34820294380188, |
|
"learning_rate": 9.464724849825347e-05, |
|
"loss": 4.1283, |
|
"step": 46950 |
|
}, |
|
{ |
|
"epoch": 0.758052289479202, |
|
"grad_norm": 1.598359227180481, |
|
"learning_rate": 9.463580086277598e-05, |
|
"loss": 3.8883, |
|
"step": 47000 |
|
}, |
|
{ |
|
"epoch": 0.7588587280850309, |
|
"grad_norm": 1.438071370124817, |
|
"learning_rate": 9.462434169297671e-05, |
|
"loss": 3.6827, |
|
"step": 47050 |
|
}, |
|
{ |
|
"epoch": 0.7596651666908598, |
|
"grad_norm": 0.7301783561706543, |
|
"learning_rate": 9.461287099181683e-05, |
|
"loss": 3.9224, |
|
"step": 47100 |
|
}, |
|
{ |
|
"epoch": 0.7604716052966888, |
|
"grad_norm": 4.070219039916992, |
|
"learning_rate": 9.46013887622605e-05, |
|
"loss": 4.0544, |
|
"step": 47150 |
|
}, |
|
{ |
|
"epoch": 0.7612780439025177, |
|
"grad_norm": 2.0918643474578857, |
|
"learning_rate": 9.458989500727481e-05, |
|
"loss": 3.7879, |
|
"step": 47200 |
|
}, |
|
{ |
|
"epoch": 0.7620844825083466, |
|
"grad_norm": 2.473740577697754, |
|
"learning_rate": 9.457838972982986e-05, |
|
"loss": 4.0541, |
|
"step": 47250 |
|
}, |
|
{ |
|
"epoch": 0.7628909211141756, |
|
"grad_norm": 1.9549883604049683, |
|
"learning_rate": 9.456687293289877e-05, |
|
"loss": 4.3095, |
|
"step": 47300 |
|
}, |
|
{ |
|
"epoch": 0.7636973597200045, |
|
"grad_norm": 1.1889281272888184, |
|
"learning_rate": 9.455534461945751e-05, |
|
"loss": 3.9717, |
|
"step": 47350 |
|
}, |
|
{ |
|
"epoch": 0.7645037983258335, |
|
"grad_norm": 2.1214423179626465, |
|
"learning_rate": 9.45438047924852e-05, |
|
"loss": 3.9387, |
|
"step": 47400 |
|
}, |
|
{ |
|
"epoch": 0.7653102369316624, |
|
"grad_norm": 1.206317663192749, |
|
"learning_rate": 9.453225345496377e-05, |
|
"loss": 4.1136, |
|
"step": 47450 |
|
}, |
|
{ |
|
"epoch": 0.7661166755374913, |
|
"grad_norm": 2.127297878265381, |
|
"learning_rate": 9.452069060987823e-05, |
|
"loss": 3.6162, |
|
"step": 47500 |
|
}, |
|
{ |
|
"epoch": 0.7669231141433203, |
|
"grad_norm": 1.8844021558761597, |
|
"learning_rate": 9.450911626021653e-05, |
|
"loss": 3.9547, |
|
"step": 47550 |
|
}, |
|
{ |
|
"epoch": 0.7677295527491492, |
|
"grad_norm": 1.8180209398269653, |
|
"learning_rate": 9.449753040896958e-05, |
|
"loss": 4.0026, |
|
"step": 47600 |
|
}, |
|
{ |
|
"epoch": 0.7685359913549782, |
|
"grad_norm": 4.184787273406982, |
|
"learning_rate": 9.448593305913129e-05, |
|
"loss": 3.5137, |
|
"step": 47650 |
|
}, |
|
{ |
|
"epoch": 0.769342429960807, |
|
"grad_norm": 1.7773044109344482, |
|
"learning_rate": 9.447432421369851e-05, |
|
"loss": 3.6649, |
|
"step": 47700 |
|
}, |
|
{ |
|
"epoch": 0.770148868566636, |
|
"grad_norm": 2.41501784324646, |
|
"learning_rate": 9.446270387567109e-05, |
|
"loss": 4.1824, |
|
"step": 47750 |
|
}, |
|
{ |
|
"epoch": 0.770955307172465, |
|
"grad_norm": 1.6952520608901978, |
|
"learning_rate": 9.445107204805183e-05, |
|
"loss": 3.5778, |
|
"step": 47800 |
|
}, |
|
{ |
|
"epoch": 0.7717617457782939, |
|
"grad_norm": 3.045311212539673, |
|
"learning_rate": 9.443942873384653e-05, |
|
"loss": 3.9632, |
|
"step": 47850 |
|
}, |
|
{ |
|
"epoch": 0.7725681843841229, |
|
"grad_norm": 2.5079586505889893, |
|
"learning_rate": 9.442777393606393e-05, |
|
"loss": 3.8897, |
|
"step": 47900 |
|
}, |
|
{ |
|
"epoch": 0.7733746229899517, |
|
"grad_norm": 2.440715789794922, |
|
"learning_rate": 9.44161076577157e-05, |
|
"loss": 3.7363, |
|
"step": 47950 |
|
}, |
|
{ |
|
"epoch": 0.7741810615957807, |
|
"grad_norm": 1.950965166091919, |
|
"learning_rate": 9.440442990181656e-05, |
|
"loss": 3.7987, |
|
"step": 48000 |
|
}, |
|
{ |
|
"epoch": 0.7741810615957807, |
|
"eval_asr_loss": 0.7671968561538705, |
|
"eval_loss": 4.1380295753479, |
|
"eval_runtime": 14.6173, |
|
"eval_samples_per_second": 35.027, |
|
"eval_steps_per_second": 11.698, |
|
"eval_tts_loss": 6.009661002101706, |
|
"step": 48000 |
|
} |
|
], |
|
"logging_steps": 50, |
|
"max_steps": 310005, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 5, |
|
"save_steps": 3000, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 3.946933255059538e+17, |
|
"train_batch_size": 3, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|