bonestew commited on
Commit
7352a69
·
verified ·
1 Parent(s): 5135ad6

End of training

Browse files
README.md CHANGED
@@ -3,20 +3,25 @@ library_name: transformers
3
  language:
4
  - en
5
  license: apache-2.0
6
- base_model: openai/whisper-tiny.en
7
  tags:
8
  - generated_from_trainer
 
 
9
  model-index:
10
- - name: articuplay
11
  results: []
12
  ---
13
 
14
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
  should probably proofread and complete it, then remove this comment. -->
16
 
17
- # articuplay
18
 
19
- This model is a fine-tuned version of [openai/whisper-tiny.en](https://huggingface.co/openai/whisper-tiny.en) on the TIMIT dataset.
 
 
 
20
 
21
  ## Model description
22
 
@@ -42,12 +47,23 @@ The following hyperparameters were used during training:
42
  - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
43
  - lr_scheduler_type: linear
44
  - lr_scheduler_warmup_steps: 500
45
- - num_epochs: 10
46
  - mixed_precision_training: Native AMP
47
 
 
 
 
 
 
 
 
 
 
 
 
48
  ### Framework versions
49
 
50
- - Transformers 4.49.0
51
  - Pytorch 2.6.0+cu124
52
  - Datasets 3.1.0
53
  - Tokenizers 0.21.0
 
3
  language:
4
  - en
5
  license: apache-2.0
6
+ base_model: openai/whisper-medium.en
7
  tags:
8
  - generated_from_trainer
9
+ metrics:
10
+ - wer
11
  model-index:
12
+ - name: articuplay1
13
  results: []
14
  ---
15
 
16
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
17
  should probably proofread and complete it, then remove this comment. -->
18
 
19
+ # articuplay1
20
 
21
+ This model is a fine-tuned version of [openai/whisper-medium.en](https://huggingface.co/openai/whisper-medium.en) on the TIMIT dataset.
22
+ It achieves the following results on the evaluation set:
23
+ - Loss: 0.2033
24
+ - Wer: 13.2533
25
 
26
  ## Model description
27
 
 
47
  - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
48
  - lr_scheduler_type: linear
49
  - lr_scheduler_warmup_steps: 500
50
+ - num_epochs: 5
51
  - mixed_precision_training: Native AMP
52
 
53
+ ### Training results
54
+
55
+ | Training Loss | Epoch | Step | Validation Loss | Wer |
56
+ |:-------------:|:-----:|:----:|:---------------:|:-------:|
57
+ | 2.9702 | 1.0 | 289 | 2.4632 | 95.9398 |
58
+ | 0.2224 | 2.0 | 578 | 0.2168 | 14.6648 |
59
+ | 0.149 | 3.0 | 867 | 0.1952 | 13.7123 |
60
+ | 0.1087 | 4.0 | 1156 | 0.1945 | 13.2154 |
61
+ | 0.0763 | 5.0 | 1445 | 0.2033 | 13.2533 |
62
+
63
+
64
  ### Framework versions
65
 
66
+ - Transformers 4.51.3
67
  - Pytorch 2.6.0+cu124
68
  - Datasets 3.1.0
69
  - Tokenizers 0.21.0
config.json CHANGED
@@ -1,5 +1,4 @@
1
  {
2
- "_name_or_path": "openai/whisper-tiny.en",
3
  "activation_dropout": 0.0,
4
  "activation_function": "gelu",
5
  "apply_spec_augment": false,
@@ -13,24 +12,19 @@
13
  ],
14
  "bos_token_id": 50257,
15
  "classifier_proj_size": 256,
16
- "d_model": 384,
17
- "decoder_attention_heads": 6,
18
- "decoder_ffn_dim": 1536,
19
  "decoder_layerdrop": 0.0,
20
- "decoder_layers": 4,
21
  "decoder_start_token_id": 50257,
22
  "dropout": 0.0,
23
- "encoder_attention_heads": 6,
24
- "encoder_ffn_dim": 1536,
25
  "encoder_layerdrop": 0.0,
26
- "encoder_layers": 4,
27
  "eos_token_id": 50256,
28
- "forced_decoder_ids": [
29
- [
30
- 1,
31
- 50362
32
- ]
33
- ],
34
  "init_std": 0.02,
35
  "is_encoder_decoder": true,
36
  "mask_feature_length": 10,
@@ -44,12 +38,12 @@
44
  "max_target_positions": 448,
45
  "median_filter_width": 7,
46
  "model_type": "whisper",
47
- "num_hidden_layers": 4,
48
  "num_mel_bins": 80,
49
  "pad_token_id": 50256,
50
  "scale_embedding": false,
51
  "torch_dtype": "float32",
52
- "transformers_version": "4.49.0",
53
  "use_cache": true,
54
  "use_weighted_layer_sum": false,
55
  "vocab_size": 51880
 
1
  {
 
2
  "activation_dropout": 0.0,
3
  "activation_function": "gelu",
4
  "apply_spec_augment": false,
 
12
  ],
13
  "bos_token_id": 50257,
14
  "classifier_proj_size": 256,
15
+ "d_model": 768,
16
+ "decoder_attention_heads": 12,
17
+ "decoder_ffn_dim": 3072,
18
  "decoder_layerdrop": 0.0,
19
+ "decoder_layers": 12,
20
  "decoder_start_token_id": 50257,
21
  "dropout": 0.0,
22
+ "encoder_attention_heads": 12,
23
+ "encoder_ffn_dim": 3072,
24
  "encoder_layerdrop": 0.0,
25
+ "encoder_layers": 12,
26
  "eos_token_id": 50256,
27
+ "forced_decoder_ids": null,
 
 
 
 
 
28
  "init_std": 0.02,
29
  "is_encoder_decoder": true,
30
  "mask_feature_length": 10,
 
38
  "max_target_positions": 448,
39
  "median_filter_width": 7,
40
  "model_type": "whisper",
41
+ "num_hidden_layers": 12,
42
  "num_mel_bins": 80,
43
  "pad_token_id": 50256,
44
  "scale_embedding": false,
45
  "torch_dtype": "float32",
46
+ "transformers_version": "4.51.3",
47
  "use_cache": true,
48
  "use_weighted_layer_sum": false,
49
  "vocab_size": 51880
generation_config.json CHANGED
@@ -6,12 +6,6 @@
6
  "bos_token_id": 50257,
7
  "decoder_start_token_id": 50257,
8
  "eos_token_id": 50256,
9
- "forced_decoder_ids": [
10
- [
11
- 1,
12
- 50362
13
- ]
14
- ],
15
  "max_length": 448,
16
  "pad_token_id": 50256,
17
  "suppress_tokens": [
@@ -106,5 +100,5 @@
106
  50360,
107
  50361
108
  ],
109
- "transformers_version": "4.49.0"
110
  }
 
6
  "bos_token_id": 50257,
7
  "decoder_start_token_id": 50257,
8
  "eos_token_id": 50256,
 
 
 
 
 
 
9
  "max_length": 448,
10
  "pad_token_id": 50256,
11
  "suppress_tokens": [
 
100
  50360,
101
  50361
102
  ],
103
+ "transformers_version": "4.51.3"
104
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:17d7274cd97799226121d5b7c1c57206e746b618c3e4c8127beb9ee86a22e3cb
3
- size 151084712
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce831ec91bceda2ccc7e3de9b2b4f8d9a9426e58d10f2093059900f931a50479
3
+ size 967041160
preprocessor_config.json CHANGED
@@ -1,5 +1,6 @@
1
  {
2
  "chunk_length": 30,
 
3
  "feature_extractor_type": "WhisperFeatureExtractor",
4
  "feature_size": 80,
5
  "hop_length": 160,
 
1
  {
2
  "chunk_length": 30,
3
+ "dither": 0.0,
4
  "feature_extractor_type": "WhisperFeatureExtractor",
5
  "feature_size": 80,
6
  "hop_length": 160,
runs/May07_16-10-41_bonestew/events.out.tfevents.1746648642.bonestew.31073.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:849aa0d9200811219d9acad1c8ed6c69e67049fdca4b36075f6ef66102b2246b
3
+ size 6734
runs/May07_16-22-44_bonestew/events.out.tfevents.1746649364.bonestew.769.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9aae7127ea7a7ba46c56eefcb0fce52e0349b2e789b49771fb3a26a0fa638b39
3
+ size 7148
runs/May07_16-40-30_bonestew/events.out.tfevents.1746650431.bonestew.769.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ad7cbb440fed9de9290799dce2c23322ec31ce24d0118246ab45d89aa2a97d9d
3
+ size 6940
runs/May07_16-54-33_bonestew/events.out.tfevents.1746651274.bonestew.769.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:601c79153f141b4e2076d5b92f964d6737d9fda3472cee766658e44a046b5a00
3
+ size 12593
runs/May07_17-23-08_bonestew/events.out.tfevents.1746652989.bonestew.769.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36b4bc26ce44fcc9184eaa24868dc9ccffd9e54810977cacf073fdc3a9361bac
3
+ size 99723
runs/May07_20-16-13_bonestew/events.out.tfevents.1746663374.bonestew.769.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f92e3ca89df8cc8462580870010d9c4ba1f2909a4aa6edce063560959b7d7036
3
+ size 38978
runs/May07_22-48-36_bonestew/events.out.tfevents.1746672535.bonestew.769.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e9b10e70a9932ba25d1a1cb6f9cbc17833a055d47637ed6c234a332063b724a
3
+ size 6906
runs/May07_23-19-57_bonestew/events.out.tfevents.1746674401.bonestew.117091.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41d1bbd05be6fa83c3fd870a1c53b55d93bf2420f2cd48806b92df732dc2448c
3
+ size 38978
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:895b92d3b49beb2d333b2040c31456ae21521fd37389a99a456cf78ea51a1cb8
3
  size 5496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:81baed9df131b1146c0ddb5c0a4c89601045fbc25c18b44724211c10cc7b0dea
3
  size 5496