deepdml commited on
Commit
b62b9ae
·
1 Parent(s): f7339ec

End of training

Browse files
Files changed (2) hide show
  1. README.md +17 -24
  2. model.safetensors +1 -1
README.md CHANGED
@@ -6,8 +6,9 @@ base_model: openai/whisper-base
6
  tags:
7
  - generated_from_trainer
8
  datasets:
9
- - deepdml/igbo-dict-16khz
10
  - deepdml/igbo-dict-expansion-16khz
 
11
  metrics:
12
  - wer
13
  model-index:
@@ -21,11 +22,13 @@ model-index:
21
  type: google/fleurs
22
  config: ig_ng
23
  split: test
 
24
  metrics:
25
  - name: Wer
26
  type: wer
27
- value: 155.96350889807658
28
  ---
 
29
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
30
  should probably proofread and complete it, then remove this comment. -->
31
 
@@ -33,8 +36,9 @@ should probably proofread and complete it, then remove this comment. -->
33
 
34
  This model is a fine-tuned version of [openai/whisper-base](https://huggingface.co/openai/whisper-base) on the google/fleurs dataset.
35
  It achieves the following results on the evaluation set:
36
- - Loss: 1.6790
37
- - Wer: 155.9635
 
38
 
39
  ## Model description
40
 
@@ -55,22 +59,22 @@ More information needed
55
  The following hyperparameters were used during training:
56
  - learning_rate: 1e-05
57
  - train_batch_size: 64
58
- - eval_batch_size: 8
59
  - seed: 42
60
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
61
  - lr_scheduler_type: linear
62
- - lr_scheduler_warmup_steps: 500
63
  - training_steps: 5000
64
 
65
  ### Training results
66
 
67
- | Training Loss | Epoch | Step | Validation Loss | Wer |
68
- |:-------------:|:------:|:----:|:---------------:|:--------:|
69
- | 0.2318 | 0.2 | 1000 | 1.3526 | 68.0029 |
70
- | 0.0786 | 1.0814 | 2000 | 1.5104 | 123.0631 |
71
- | 0.0627 | 1.2814 | 3000 | 1.5945 | 166.5873 |
72
- | 0.0317 | 2.1628 | 4000 | 1.6534 | 141.3940 |
73
- | 0.0321 | 3.0442 | 5000 | 1.6790 | 155.9635 |
74
 
75
 
76
  ### Framework versions
@@ -79,14 +83,3 @@ The following hyperparameters were used during training:
79
  - Pytorch 2.3.0+cu121
80
  - Datasets 2.19.1
81
  - Tokenizers 0.19.1
82
-
83
- ## Citation
84
-
85
- ```bibtex
86
- @misc{deepdml/whisper-base-ig-mix-norm,
87
- title={Fine-tuned Whisper base ASR model for speech recognition in Igbo},
88
- author={Jimenez, David},
89
- howpublished={\url{https://huggingface.co/deepdml/whisper-base-ig-mix-norm}},
90
- year={2025}
91
- }
92
- ```
 
6
  tags:
7
  - generated_from_trainer
8
  datasets:
9
+ - google/fleurs
10
  - deepdml/igbo-dict-expansion-16khz
11
+ - deepdml/igbo-dict-16khz
12
  metrics:
13
  - wer
14
  model-index:
 
22
  type: google/fleurs
23
  config: ig_ng
24
  split: test
25
+ args: ig_ng
26
  metrics:
27
  - name: Wer
28
  type: wer
29
+ value: 54.948739128322245
30
  ---
31
+
32
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
33
  should probably proofread and complete it, then remove this comment. -->
34
 
 
36
 
37
  This model is a fine-tuned version of [openai/whisper-base](https://huggingface.co/openai/whisper-base) on the google/fleurs dataset.
38
  It achieves the following results on the evaluation set:
39
+ - Loss: 1.0933
40
+ - Wer: 54.9487
41
+ - Cer: 21.3532
42
 
43
  ## Model description
44
 
 
59
  The following hyperparameters were used during training:
60
  - learning_rate: 1e-05
61
  - train_batch_size: 64
62
+ - eval_batch_size: 64
63
  - seed: 42
64
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
65
  - lr_scheduler_type: linear
66
+ - lr_scheduler_warmup_ratio: 0.04
67
  - training_steps: 5000
68
 
69
  ### Training results
70
 
71
+ | Training Loss | Epoch | Step | Validation Loss | Wer | Cer |
72
+ |:-------------:|:------:|:----:|:---------------:|:-------:|:-------:|
73
+ | 0.2087 | 0.2 | 1000 | 0.8427 | 54.4143 | 20.1160 |
74
+ | 0.0734 | 1.0814 | 2000 | 0.9702 | 55.5707 | 21.6200 |
75
+ | 0.0609 | 1.2814 | 3000 | 1.0272 | 54.0256 | 20.4927 |
76
+ | 0.0336 | 2.1628 | 4000 | 1.0804 | 54.4337 | 20.4677 |
77
+ | 0.0341 | 3.0442 | 5000 | 1.0933 | 54.9487 | 21.3532 |
78
 
79
 
80
  ### Framework versions
 
83
  - Pytorch 2.3.0+cu121
84
  - Datasets 2.19.1
85
  - Tokenizers 0.19.1
 
 
 
 
 
 
 
 
 
 
 
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:698788d295ec7c648a79ee1082c461e800e329b4f4adad747692a65ceecebe49
3
  size 290403936
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e6e7fab85f5e5b8b4db3eb17ef1b3c862f6e3684a8c2599a37b403b0f82ea364
3
  size 290403936