End of training
Browse files
README.md
CHANGED
@@ -26,7 +26,7 @@ model-index:
|
|
26 |
metrics:
|
27 |
- name: Wer
|
28 |
type: wer
|
29 |
-
value: 20.
|
30 |
---
|
31 |
|
32 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
@@ -36,8 +36,8 @@ should probably proofread and complete it, then remove this comment. -->
|
|
36 |
|
37 |
This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on the Common Voice 11.0 dataset.
|
38 |
It achieves the following results on the evaluation set:
|
39 |
-
- Loss: 0.
|
40 |
-
- Wer: 20.
|
41 |
|
42 |
## Model description
|
43 |
|
@@ -65,23 +65,33 @@ The following hyperparameters were used during training:
|
|
65 |
- optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
|
66 |
- lr_scheduler_type: linear
|
67 |
- lr_scheduler_warmup_steps: 500
|
68 |
-
- training_steps:
|
69 |
- mixed_precision_training: Native AMP
|
70 |
|
71 |
### Training results
|
72 |
|
73 |
-
| Training Loss | Epoch | Step
|
74 |
-
|
75 |
-
| 0.5507 | 0.2079 | 500
|
76 |
-
| 0.2802 | 0.4158 | 1000
|
77 |
-
| 0.2408 | 0.6236 | 1500
|
78 |
-
| 0.2208 | 0.8315 | 2000
|
79 |
-
| 0.1811 | 1.0394 | 2500
|
80 |
-
| 0.1096 | 1.2473 | 3000
|
81 |
-
| 0.1068 | 1.4552 | 3500
|
82 |
-
| 0.1042 | 1.6630 | 4000
|
83 |
-
| 0.1001 | 1.8709 | 4500
|
84 |
-
| 0.0827 | 2.0788 | 5000
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
85 |
|
86 |
|
87 |
### Framework versions
|
|
|
26 |
metrics:
|
27 |
- name: Wer
|
28 |
type: wer
|
29 |
+
value: 20.157687253613666
|
30 |
---
|
31 |
|
32 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
|
|
36 |
|
37 |
This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on the Common Voice 11.0 dataset.
|
38 |
It achieves the following results on the evaluation set:
|
39 |
+
- Loss: 0.2758
|
40 |
+
- Wer: 20.1577
|
41 |
|
42 |
## Model description
|
43 |
|
|
|
65 |
- optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
|
66 |
- lr_scheduler_type: linear
|
67 |
- lr_scheduler_warmup_steps: 500
|
68 |
+
- training_steps: 10000
|
69 |
- mixed_precision_training: Native AMP
|
70 |
|
71 |
### Training results
|
72 |
|
73 |
+
| Training Loss | Epoch | Step | Validation Loss | Wer |
|
74 |
+
|:-------------:|:------:|:-----:|:---------------:|:-------:|
|
75 |
+
| 0.5507 | 0.2079 | 500 | 0.3695 | 29.2247 |
|
76 |
+
| 0.2802 | 0.4158 | 1000 | 0.3148 | 26.7299 |
|
77 |
+
| 0.2408 | 0.6236 | 1500 | 0.2970 | 24.2538 |
|
78 |
+
| 0.2208 | 0.8315 | 2000 | 0.2728 | 23.3020 |
|
79 |
+
| 0.1811 | 1.0394 | 2500 | 0.2665 | 22.3935 |
|
80 |
+
| 0.1096 | 1.2473 | 3000 | 0.2641 | 21.8998 |
|
81 |
+
| 0.1068 | 1.4552 | 3500 | 0.2568 | 21.6125 |
|
82 |
+
| 0.1042 | 1.6630 | 4000 | 0.2516 | 21.0512 |
|
83 |
+
| 0.1001 | 1.8709 | 4500 | 0.2472 | 20.4092 |
|
84 |
+
| 0.0827 | 2.0788 | 5000 | 0.2469 | 20.3848 |
|
85 |
+
| 0.0672 | 2.2869 | 5500 | 0.2665 | 21.1357 |
|
86 |
+
| 0.0673 | 2.4948 | 6000 | 0.2674 | 21.5093 |
|
87 |
+
| 0.0681 | 2.7026 | 6500 | 0.2635 | 20.6101 |
|
88 |
+
| 0.0661 | 2.9105 | 7000 | 0.2602 | 20.5069 |
|
89 |
+
| 0.0494 | 3.1184 | 7500 | 0.2708 | 20.5444 |
|
90 |
+
| 0.0352 | 3.3263 | 8000 | 0.2688 | 20.5181 |
|
91 |
+
| 0.0338 | 3.5341 | 8500 | 0.2717 | 20.2515 |
|
92 |
+
| 0.0318 | 3.7420 | 9000 | 0.2723 | 20.2403 |
|
93 |
+
| 0.0309 | 3.9499 | 9500 | 0.2711 | 20.1727 |
|
94 |
+
| 0.022 | 4.1578 | 10000 | 0.2758 | 20.1577 |
|
95 |
|
96 |
|
97 |
### Framework versions
|