Priyanship commited on
Commit
c75a43d
·
verified ·
1 Parent(s): 9bd6d06

End of training

Browse files
README.md CHANGED
@@ -9,18 +9,18 @@ model-index:
9
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
10
  should probably proofread and complete it, then remove this comment. -->
11
 
12
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/priyanshipal/huggingface/runs/1kodfy70)
13
  # s300_shuff100
14
 
15
  This model was trained from scratch on an unknown dataset.
16
  It achieves the following results on the evaluation set:
17
  - eval_loss: nan
18
- - eval_model_preparation_time: 0.0046
19
  - eval_cer: 1.0
20
  - eval_wer: 1.0
21
- - eval_runtime: 39.8895
22
- - eval_samples_per_second: 14.34
23
- - eval_steps_per_second: 0.902
24
  - step: 0
25
 
26
  ## Model description
 
9
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
10
  should probably proofread and complete it, then remove this comment. -->
11
 
12
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/priyanshipal/huggingface/runs/64250v6u)
13
  # s300_shuff100
14
 
15
  This model was trained from scratch on an unknown dataset.
16
  It achieves the following results on the evaluation set:
17
  - eval_loss: nan
18
+ - eval_model_preparation_time: 0.0045
19
  - eval_cer: 1.0
20
  - eval_wer: 1.0
21
+ - eval_runtime: 30.9217
22
+ - eval_samples_per_second: 18.498
23
+ - eval_steps_per_second: 1.164
24
  - step: 0
25
 
26
  ## Model description
all_results.json CHANGED
@@ -2,11 +2,11 @@
2
  "epoch": 1.6,
3
  "eval_cer": 1.0,
4
  "eval_loss": NaN,
5
- "eval_model_preparation_time": 0.0046,
6
- "eval_runtime": 39.8895,
7
  "eval_samples": 572,
8
- "eval_samples_per_second": 14.34,
9
- "eval_steps_per_second": 0.902,
10
  "eval_wer": 1.0,
11
  "total_flos": 6.212261523683712e+18,
12
  "train_loss": 3.21392811447382,
 
2
  "epoch": 1.6,
3
  "eval_cer": 1.0,
4
  "eval_loss": NaN,
5
+ "eval_model_preparation_time": 0.0045,
6
+ "eval_runtime": 30.9217,
7
  "eval_samples": 572,
8
+ "eval_samples_per_second": 18.498,
9
+ "eval_steps_per_second": 1.164,
10
  "eval_wer": 1.0,
11
  "total_flos": 6.212261523683712e+18,
12
  "train_loss": 3.21392811447382,
eval_results.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
  "eval_cer": 1.0,
3
  "eval_loss": NaN,
4
- "eval_model_preparation_time": 0.0046,
5
- "eval_runtime": 39.8895,
6
  "eval_samples": 572,
7
- "eval_samples_per_second": 14.34,
8
- "eval_steps_per_second": 0.902,
9
  "eval_wer": 1.0
10
  }
 
1
  {
2
  "eval_cer": 1.0,
3
  "eval_loss": NaN,
4
+ "eval_model_preparation_time": 0.0045,
5
+ "eval_runtime": 30.9217,
6
  "eval_samples": 572,
7
+ "eval_samples_per_second": 18.498,
8
+ "eval_steps_per_second": 1.164,
9
  "eval_wer": 1.0
10
  }
evalonlyhindi_indicwav2vec_MUCS_warmup500_s300shuff100_2142383.out CHANGED
@@ -152,3 +152,43 @@ last prediction string
152
  eval_steps_per_second = 0.902
153
  eval_wer = 1.0
154
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
152
  eval_steps_per_second = 0.902
153
  eval_wer = 1.0
154
 
155
+ wandb: - 0.005 MB of 0.005 MB uploaded
156
+ wandb: Run history:
157
+ wandb: eval/cer ▁
158
+ wandb: eval/model_preparation_time ▁
159
+ wandb: eval/runtime ▁
160
+ wandb: eval/samples_per_second ▁
161
+ wandb: eval/steps_per_second ▁
162
+ wandb: eval/wer ▁
163
+ wandb: eval_cer ▁
164
+ wandb: eval_model_preparation_time ▁
165
+ wandb: eval_runtime ▁
166
+ wandb: eval_samples ▁
167
+ wandb: eval_samples_per_second ▁
168
+ wandb: eval_steps_per_second ▁
169
+ wandb: eval_wer ▁
170
+ wandb: train/global_step ▁▁
171
+ wandb:
172
+ wandb: Run summary:
173
+ wandb: eval/cer 1.0
174
+ wandb: eval/loss nan
175
+ wandb: eval/model_preparation_time 0.0046
176
+ wandb: eval/runtime 39.8895
177
+ wandb: eval/samples_per_second 14.34
178
+ wandb: eval/steps_per_second 0.902
179
+ wandb: eval/wer 1.0
180
+ wandb: eval_cer 1.0
181
+ wandb: eval_loss nan
182
+ wandb: eval_model_preparation_time 0.0046
183
+ wandb: eval_runtime 39.8895
184
+ wandb: eval_samples 572
185
+ wandb: eval_samples_per_second 14.34
186
+ wandb: eval_steps_per_second 0.902
187
+ wandb: eval_wer 1.0
188
+ wandb: train/global_step 0
189
+ wandb:
190
+ wandb: 🚀 View run eval_pd2000_s300_shuff100_hindi at: https://wandb.ai/priyanshipal/huggingface/runs/1kodfy70
191
+ wandb: ⭐️ View project at: https://wandb.ai/priyanshipal/huggingface
192
+ wandb: Synced 6 W&B file(s), 0 media file(s), 0 artifact file(s) and 0 other file(s)
193
+ wandb: Find logs at: ./wandb/run-20240822_150154-1kodfy70/logs
194
+ wandb: WARNING The new W&B backend becomes opt-out in version 0.18.0; try it out with `wandb.require("core")`! See https://wandb.me/wandb-core for more information.
evalonlyhindi_indicwav2vec_MUCS_warmup500_s300shuff100_2142409.out ADDED
@@ -0,0 +1,110 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb: Currently logged in as: priyanshi-pal (priyanshipal). Use `wandb login --relogin` to force relogin
2
+ wandb: wandb version 0.17.7 is available! To upgrade, please run:
3
+ wandb: $ pip install wandb --upgrade
4
+ wandb: Tracking run with wandb version 0.17.6
5
+ wandb: Run data is saved locally in /scratch/elec/t405-puhe/p/palp3/MUCS/wandb/run-20240822_151437-2b363w6i
6
+ wandb: Run `wandb offline` to turn off syncing.
7
+ wandb: Syncing run eval_pd2000_s300_shuff100_hindi
8
+ wandb: ⭐️ View project at https://wandb.ai/priyanshipal/huggingface
9
+ wandb: 🚀 View run at https://wandb.ai/priyanshipal/huggingface/runs/2b363w6i
10
+ /scratch/work/palp3/myenv/lib/python3.11/site-packages/transformers/training_args.py:1525: FutureWarning: `evaluation_strategy` is deprecated and will be removed in version 4.46 of 🤗 Transformers. Use `eval_strategy` instead
11
+ warnings.warn(
12
+ /scratch/work/palp3/myenv/lib/python3.11/site-packages/transformers/models/auto/configuration_auto.py:957: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
13
+ warnings.warn(
14
+ /scratch/work/palp3/myenv/lib/python3.11/site-packages/transformers/models/auto/feature_extraction_auto.py:329: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
15
+ warnings.warn(
16
+ Wav2Vec2CTCTokenizer(name_or_path='', vocab_size=149, model_max_length=1000000000000000019884624838656, is_fast=False, padding_side='right', truncation_side='right', special_tokens={'bos_token': '<s>', 'eos_token': '</s>', 'unk_token': '[UNK]', 'pad_token': '[PAD]'}, clean_up_tokenization_spaces=True), added_tokens_decoder={
17
+ 147: AddedToken("[UNK]", rstrip=True, lstrip=True, single_word=False, normalized=False, special=False),
18
+ 148: AddedToken("[PAD]", rstrip=True, lstrip=True, single_word=False, normalized=False, special=False),
19
+ 149: AddedToken("<s>", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
20
+ 150: AddedToken("</s>", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
21
+ }
22
+ CHECK MODEL PARAMS Wav2Vec2ForCTC(
23
+ (wav2vec2): Wav2Vec2Model(
24
+ (feature_extractor): Wav2Vec2FeatureEncoder(
25
+ (conv_layers): ModuleList(
26
+ (0): Wav2Vec2LayerNormConvLayer(
27
+ (conv): Conv1d(1, 512, kernel_size=(10,), stride=(5,))
28
+ (layer_norm): LayerNorm((512,), eps=1e-05, elementwise_affine=True)
29
+ (activation): GELUActivation()
30
+ )
31
+ (1-4): 4 x Wav2Vec2LayerNormConvLayer(
32
+ (conv): Conv1d(512, 512, kernel_size=(3,), stride=(2,))
33
+ (layer_norm): LayerNorm((512,), eps=1e-05, elementwise_affine=True)
34
+ (activation): GELUActivation()
35
+ )
36
+ (5-6): 2 x Wav2Vec2LayerNormConvLayer(
37
+ (conv): Conv1d(512, 512, kernel_size=(2,), stride=(2,))
38
+ (layer_norm): LayerNorm((512,), eps=1e-05, elementwise_affine=True)
39
+ (activation): GELUActivation()
40
+ )
41
+ )
42
+ )
43
+ (feature_projection): Wav2Vec2FeatureProjection(
44
+ (layer_norm): LayerNorm((512,), eps=1e-05, elementwise_affine=True)
45
+ (projection): Linear(in_features=512, out_features=1024, bias=True)
46
+ (dropout): Dropout(p=0.0, inplace=False)
47
+ )
48
+ (encoder): Wav2Vec2EncoderStableLayerNorm(
49
+ (pos_conv_embed): Wav2Vec2PositionalConvEmbedding(
50
+ (conv): ParametrizedConv1d(
51
+ 1024, 1024, kernel_size=(128,), stride=(1,), padding=(64,), groups=16
52
+ (parametrizations): ModuleDict(
53
+ (weight): ParametrizationList(
54
+ (0): _WeightNorm()
55
+ )
56
+ )
57
+ )
58
+ (padding): Wav2Vec2SamePadLayer()
59
+ (activation): GELUActivation()
60
+ )
61
+ (layer_norm): LayerNorm((1024,), eps=1e-05, elementwise_affine=True)
62
+ (dropout): Dropout(p=0.0, inplace=False)
63
+ (layers): ModuleList(
64
+ (0-23): 24 x Wav2Vec2EncoderLayerStableLayerNorm(
65
+ (attention): Wav2Vec2SdpaAttention(
66
+ (k_proj): Linear(in_features=1024, out_features=1024, bias=True)
67
+ (v_proj): Linear(in_features=1024, out_features=1024, bias=True)
68
+ (q_proj): Linear(in_features=1024, out_features=1024, bias=True)
69
+ (out_proj): Linear(in_features=1024, out_features=1024, bias=True)
70
+ )
71
+ (dropout): Dropout(p=0.0, inplace=False)
72
+ (layer_norm): LayerNorm((1024,), eps=1e-05, elementwise_affine=True)
73
+ (feed_forward): Wav2Vec2FeedForward(
74
+ (intermediate_dropout): Dropout(p=0.0, inplace=False)
75
+ (intermediate_dense): Linear(in_features=1024, out_features=4096, bias=True)
76
+ (intermediate_act_fn): GELUActivation()
77
+ (output_dense): Linear(in_features=4096, out_features=1024, bias=True)
78
+ (output_dropout): Dropout(p=0.0, inplace=False)
79
+ )
80
+ (final_layer_norm): LayerNorm((1024,), eps=1e-05, elementwise_affine=True)
81
+ )
82
+ )
83
+ )
84
+ )
85
+ (dropout): Dropout(p=0.0, inplace=False)
86
+ (lm_head): Linear(in_features=1024, out_features=151, bias=True)
87
+ )
88
+
89
+ Traceback (most recent call last):
90
+ File "/scratch/elec/puhe/p/palp3/MUCS/eval_script_indicwav2vec.py", line 790, in <module>
91
+ main()
92
+ File "/scratch/elec/puhe/p/palp3/MUCS/eval_script_indicwav2vec.py", line 637, in main
93
+ print("check the eval set length", len(vectorized_datasets["eval"]["audio_id"]))
94
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^
95
+ File "/scratch/work/palp3/myenv/lib/python3.11/site-packages/datasets/arrow_dataset.py", line 2866, in __getitem__
96
+ return self._getitem(key)
97
+ ^^^^^^^^^^^^^^^^^^
98
+ File "/scratch/work/palp3/myenv/lib/python3.11/site-packages/datasets/arrow_dataset.py", line 2850, in _getitem
99
+ pa_subtable = query_table(self._data, key, indices=self._indices)
100
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
101
+ File "/scratch/work/palp3/myenv/lib/python3.11/site-packages/datasets/formatting/formatting.py", line 584, in query_table
102
+ _check_valid_column_key(key, table.column_names)
103
+ File "/scratch/work/palp3/myenv/lib/python3.11/site-packages/datasets/formatting/formatting.py", line 521, in _check_valid_column_key
104
+ raise KeyError(f"Column {key} not in the dataset. Current columns in the dataset: {columns}")
105
+ KeyError: "Column audio_id not in the dataset. Current columns in the dataset: ['input_values', 'input_length', 'labels']"
106
+ wandb: - 0.011 MB of 0.011 MB uploaded
107
+ wandb: ⭐️ View project at: https://wandb.ai/priyanshipal/huggingface
108
+ wandb: Synced 6 W&B file(s), 0 media file(s), 0 artifact file(s) and 0 other file(s)
109
+ wandb: Find logs at: ./wandb/run-20240822_151437-2b363w6i/logs
110
+ wandb: WARNING The new W&B backend becomes opt-out in version 0.18.0; try it out with `wandb.require("core")`! See https://wandb.me/wandb-core for more information.
evalonlyhindi_indicwav2vec_MUCS_warmup500_s300shuff100_2142421.out ADDED
@@ -0,0 +1,175 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
0
  0%| | 0/36 [00:00<?, ?it/s]
1
  6%|▌ | 2/36 [00:01<00:31, 1.07it/s]
2
  8%|▊ | 3/36 [00:03<00:42, 1.28s/it]
3
  11%|█ | 4/36 [00:05<00:52, 1.65s/it]
4
  14%|█▍ | 5/36 [00:07<00:53, 1.72s/it]
5
  17%|█▋ | 6/36 [00:09<00:51, 1.72s/it]
6
  19%|█▉ | 7/36 [00:10<00:42, 1.48s/it]
7
  22%|██▏ | 8/36 [00:11<00:33, 1.20s/it]
8
  25%|██▌ | 9/36 [00:11<00:28, 1.06s/it]
9
  28%|██▊ | 10/36 [00:12<00:27, 1.04s/it]
10
  31%|███ | 11/36 [00:13<00:26, 1.06s/it]
11
  33%|███▎ | 12/36 [00:14<00:25, 1.05s/it]
12
  36%|███▌ | 13/36 [00:15<00:23, 1.00s/it]
13
  39%|███▉ | 14/36 [00:16<00:19, 1.15it/s]
14
  42%|████▏ | 15/36 [00:16<00:16, 1.28it/s]
15
  44%|████▍ | 16/36 [00:17<00:14, 1.36it/s]
16
  47%|████▋ | 17/36 [00:18<00:13, 1.41it/s]
17
  50%|█████ | 18/36 [00:19<00:13, 1.35it/s]
18
  53%|█████▎ | 19/36 [00:19<00:13, 1.30it/s]
19
  56%|█████▌ | 20/36 [00:20<00:11, 1.35it/s]
20
  58%|█████▊ | 21/36 [00:21<00:10, 1.45it/s]
21
  61%|██████ | 22/36 [00:21<00:09, 1.43it/s]
22
  64%|██████▍ | 23/36 [00:22<00:09, 1.40it/s]
23
  67%|██████▋ | 24/36 [00:23<00:08, 1.38it/s]
24
  69%|██████▉ | 25/36 [00:24<00:08, 1.31it/s]
25
  72%|███████▏ | 26/36 [00:24<00:07, 1.35it/s]
26
  75%|███████▌ | 27/36 [00:25<00:06, 1.45it/s]
27
  78%|███████▊ | 28/36 [00:26<00:06, 1.18it/s]
28
  81%|████████ | 29/36 [00:28<00:08, 1.22s/it]
29
  83%|████████▎ | 30/36 [00:30<00:08, 1.37s/it]Traceback (most recent call last):
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb: Currently logged in as: priyanshi-pal (priyanshipal). Use `wandb login --relogin` to force relogin
2
+ wandb: wandb version 0.17.7 is available! To upgrade, please run:
3
+ wandb: $ pip install wandb --upgrade
4
+ wandb: Tracking run with wandb version 0.17.6
5
+ wandb: Run data is saved locally in /scratch/elec/t405-puhe/p/palp3/MUCS/wandb/run-20240822_151726-alv0f5i7
6
+ wandb: Run `wandb offline` to turn off syncing.
7
+ wandb: Syncing run eval_pd2000_s300_shuff100_hindi
8
+ wandb: ⭐️ View project at https://wandb.ai/priyanshipal/huggingface
9
+ wandb: 🚀 View run at https://wandb.ai/priyanshipal/huggingface/runs/alv0f5i7
10
+ /scratch/work/palp3/myenv/lib/python3.11/site-packages/transformers/training_args.py:1525: FutureWarning: `evaluation_strategy` is deprecated and will be removed in version 4.46 of 🤗 Transformers. Use `eval_strategy` instead
11
+ warnings.warn(
12
+ /scratch/work/palp3/myenv/lib/python3.11/site-packages/transformers/models/auto/configuration_auto.py:957: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
13
+ warnings.warn(
14
+ /scratch/work/palp3/myenv/lib/python3.11/site-packages/transformers/models/auto/feature_extraction_auto.py:329: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
15
+ warnings.warn(
16
+ /scratch/work/palp3/myenv/lib/python3.11/site-packages/accelerate/accelerator.py:488: FutureWarning: `torch.cuda.amp.GradScaler(args...)` is deprecated. Please use `torch.amp.GradScaler('cuda', args...)` instead.
17
+ self.scaler = torch.cuda.amp.GradScaler(**kwargs)
18
+ max_steps is given, it will override any value given in num_train_epochs
19
+ Wav2Vec2CTCTokenizer(name_or_path='', vocab_size=149, model_max_length=1000000000000000019884624838656, is_fast=False, padding_side='right', truncation_side='right', special_tokens={'bos_token': '<s>', 'eos_token': '</s>', 'unk_token': '[UNK]', 'pad_token': '[PAD]'}, clean_up_tokenization_spaces=True), added_tokens_decoder={
20
+ 147: AddedToken("[UNK]", rstrip=True, lstrip=True, single_word=False, normalized=False, special=False),
21
+ 148: AddedToken("[PAD]", rstrip=True, lstrip=True, single_word=False, normalized=False, special=False),
22
+ 149: AddedToken("<s>", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
23
+ 150: AddedToken("</s>", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
24
+ }
25
+ CHECK MODEL PARAMS Wav2Vec2ForCTC(
26
+ (wav2vec2): Wav2Vec2Model(
27
+ (feature_extractor): Wav2Vec2FeatureEncoder(
28
+ (conv_layers): ModuleList(
29
+ (0): Wav2Vec2LayerNormConvLayer(
30
+ (conv): Conv1d(1, 512, kernel_size=(10,), stride=(5,))
31
+ (layer_norm): LayerNorm((512,), eps=1e-05, elementwise_affine=True)
32
+ (activation): GELUActivation()
33
+ )
34
+ (1-4): 4 x Wav2Vec2LayerNormConvLayer(
35
+ (conv): Conv1d(512, 512, kernel_size=(3,), stride=(2,))
36
+ (layer_norm): LayerNorm((512,), eps=1e-05, elementwise_affine=True)
37
+ (activation): GELUActivation()
38
+ )
39
+ (5-6): 2 x Wav2Vec2LayerNormConvLayer(
40
+ (conv): Conv1d(512, 512, kernel_size=(2,), stride=(2,))
41
+ (layer_norm): LayerNorm((512,), eps=1e-05, elementwise_affine=True)
42
+ (activation): GELUActivation()
43
+ )
44
+ )
45
+ )
46
+ (feature_projection): Wav2Vec2FeatureProjection(
47
+ (layer_norm): LayerNorm((512,), eps=1e-05, elementwise_affine=True)
48
+ (projection): Linear(in_features=512, out_features=1024, bias=True)
49
+ (dropout): Dropout(p=0.0, inplace=False)
50
+ )
51
+ (encoder): Wav2Vec2EncoderStableLayerNorm(
52
+ (pos_conv_embed): Wav2Vec2PositionalConvEmbedding(
53
+ (conv): ParametrizedConv1d(
54
+ 1024, 1024, kernel_size=(128,), stride=(1,), padding=(64,), groups=16
55
+ (parametrizations): ModuleDict(
56
+ (weight): ParametrizationList(
57
+ (0): _WeightNorm()
58
+ )
59
+ )
60
+ )
61
+ (padding): Wav2Vec2SamePadLayer()
62
+ (activation): GELUActivation()
63
+ )
64
+ (layer_norm): LayerNorm((1024,), eps=1e-05, elementwise_affine=True)
65
+ (dropout): Dropout(p=0.0, inplace=False)
66
+ (layers): ModuleList(
67
+ (0-23): 24 x Wav2Vec2EncoderLayerStableLayerNorm(
68
+ (attention): Wav2Vec2SdpaAttention(
69
+ (k_proj): Linear(in_features=1024, out_features=1024, bias=True)
70
+ (v_proj): Linear(in_features=1024, out_features=1024, bias=True)
71
+ (q_proj): Linear(in_features=1024, out_features=1024, bias=True)
72
+ (out_proj): Linear(in_features=1024, out_features=1024, bias=True)
73
+ )
74
+ (dropout): Dropout(p=0.0, inplace=False)
75
+ (layer_norm): LayerNorm((1024,), eps=1e-05, elementwise_affine=True)
76
+ (feed_forward): Wav2Vec2FeedForward(
77
+ (intermediate_dropout): Dropout(p=0.0, inplace=False)
78
+ (intermediate_dense): Linear(in_features=1024, out_features=4096, bias=True)
79
+ (intermediate_act_fn): GELUActivation()
80
+ (output_dense): Linear(in_features=4096, out_features=1024, bias=True)
81
+ (output_dropout): Dropout(p=0.0, inplace=False)
82
+ )
83
+ (final_layer_norm): LayerNorm((1024,), eps=1e-05, elementwise_affine=True)
84
+ )
85
+ )
86
+ )
87
+ )
88
+ (dropout): Dropout(p=0.0, inplace=False)
89
+ (lm_head): Linear(in_features=1024, out_features=151, bias=True)
90
+ )
91
+ check the eval set length 572
92
+ 08/22/2024 15:17:37 - INFO - __main__ - *** Evaluate ***
93
+ /scratch/work/palp3/myenv/lib/python3.11/site-packages/transformers/models/wav2vec2/processing_wav2vec2.py:157: UserWarning: `as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your labels by using the argument `text` of the regular `__call__` method (either in the same call as your audio inputs, or in a separate call.
94
+ warnings.warn(
95
+
96
  0%| | 0/36 [00:00<?, ?it/s]
97
  6%|▌ | 2/36 [00:01<00:31, 1.07it/s]
98
  8%|▊ | 3/36 [00:03<00:42, 1.28s/it]
99
  11%|█ | 4/36 [00:05<00:52, 1.65s/it]
100
  14%|█▍ | 5/36 [00:07<00:53, 1.72s/it]
101
  17%|█▋ | 6/36 [00:09<00:51, 1.72s/it]
102
  19%|█▉ | 7/36 [00:10<00:42, 1.48s/it]
103
  22%|██▏ | 8/36 [00:11<00:33, 1.20s/it]
104
  25%|██▌ | 9/36 [00:11<00:28, 1.06s/it]
105
  28%|██▊ | 10/36 [00:12<00:27, 1.04s/it]
106
  31%|███ | 11/36 [00:13<00:26, 1.06s/it]
107
  33%|███▎ | 12/36 [00:14<00:25, 1.05s/it]
108
  36%|███▌ | 13/36 [00:15<00:23, 1.00s/it]
109
  39%|███▉ | 14/36 [00:16<00:19, 1.15it/s]
110
  42%|████▏ | 15/36 [00:16<00:16, 1.28it/s]
111
  44%|████▍ | 16/36 [00:17<00:14, 1.36it/s]
112
  47%|████▋ | 17/36 [00:18<00:13, 1.41it/s]
113
  50%|█████ | 18/36 [00:19<00:13, 1.35it/s]
114
  53%|█████▎ | 19/36 [00:19<00:13, 1.30it/s]
115
  56%|█████▌ | 20/36 [00:20<00:11, 1.35it/s]
116
  58%|█████▊ | 21/36 [00:21<00:10, 1.45it/s]
117
  61%|██████ | 22/36 [00:21<00:09, 1.43it/s]
118
  64%|██████▍ | 23/36 [00:22<00:09, 1.40it/s]
119
  67%|██████▋ | 24/36 [00:23<00:08, 1.38it/s]
120
  69%|██████▉ | 25/36 [00:24<00:08, 1.31it/s]
121
  72%|███████▏ | 26/36 [00:24<00:07, 1.35it/s]
122
  75%|███████▌ | 27/36 [00:25<00:06, 1.45it/s]
123
  78%|███████▊ | 28/36 [00:26<00:06, 1.18it/s]
124
  81%|████████ | 29/36 [00:28<00:08, 1.22s/it]
125
  83%|████████▎ | 30/36 [00:30<00:08, 1.37s/it]Traceback (most recent call last):
126
+ File "/scratch/elec/puhe/p/palp3/MUCS/eval_script_indicwav2vec.py", line 790, in <module>
127
+ main()
128
+ File "/scratch/elec/puhe/p/palp3/MUCS/eval_script_indicwav2vec.py", line 759, in main
129
+ metrics = trainer.evaluate()
130
+ ^^^^^^^^^^^^^^^^^^
131
+ File "/scratch/work/palp3/myenv/lib/python3.11/site-packages/transformers/trainer.py", line 3666, in evaluate
132
+ output = eval_loop(
133
+ ^^^^^^^^^^
134
+ File "/scratch/work/palp3/myenv/lib/python3.11/site-packages/transformers/trainer.py", line 3857, in evaluation_loop
135
+ losses, logits, labels = self.prediction_step(model, inputs, prediction_loss_only, ignore_keys=ignore_keys)
136
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
137
+ File "/scratch/work/palp3/myenv/lib/python3.11/site-packages/transformers/trainer.py", line 4075, in prediction_step
138
+ loss, outputs = self.compute_loss(model, inputs, return_outputs=True)
139
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
140
+ File "/scratch/work/palp3/myenv/lib/python3.11/site-packages/transformers/trainer.py", line 3363, in compute_loss
141
+ outputs = model(**inputs)
142
+ ^^^^^^^^^^^^^^^
143
+ File "/scratch/work/palp3/myenv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl
144
+ return self._call_impl(*args, **kwargs)
145
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
146
+ File "/scratch/work/palp3/myenv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl
147
+ return forward_call(*args, **kwargs)
148
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
149
+ File "/scratch/work/palp3/myenv/lib/python3.11/site-packages/accelerate/utils/operations.py", line 819, in forward
150
+ return model_forward(*args, **kwargs)
151
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
152
+ File "/scratch/work/palp3/myenv/lib/python3.11/site-packages/accelerate/utils/operations.py", line 807, in __call__
153
+ return convert_to_fp32(self.model_forward(*args, **kwargs))
154
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
155
+ File "/scratch/work/palp3/myenv/lib/python3.11/site-packages/torch/amp/autocast_mode.py", line 43, in decorate_autocast
156
+ return func(*args, **kwargs)
157
+ ^^^^^^^^^^^^^^^^^^^^^
158
+ File "/scratch/work/palp3/myenv/lib/python3.11/site-packages/transformers/models/wav2vec2/modeling_wav2vec2.py", line 2228, in forward
159
+ outputs = self.wav2vec2(
160
+ ^^^^^^^^^^^^^^
161
+ File "/scratch/work/palp3/myenv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl
162
+ return self._call_impl(*args, **kwargs)
163
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
164
+ File "/scratch/work/palp3/myenv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl
165
+ return forward_call(*args, **kwargs)
166
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
167
+ File "/scratch/work/palp3/myenv/lib/python3.11/site-packages/transformers/models/wav2vec2/modeling_wav2vec2.py", line 1809, in forward
168
+ extract_features = self.feature_extractor(input_values)
169
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
170
+ File "/scratch/work/palp3/myenv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl
171
+ return self._call_impl(*args, **kwargs)
172
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
173
+ File "/scratch/work/palp3/myenv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl
174
+ return forward_call(*args, **kwargs)
175
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
176
+ File "/scratch/work/palp3/myenv/lib/python3.11/site-packages/transformers/models/wav2vec2/modeling_wav2vec2.py", line 463, in forward
177
+ hidden_states = conv_layer(hidden_states)
178
+ ^^^^^^^^^^^^^^^^^^^^^^^^^
179
+ File "/scratch/work/palp3/myenv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl
180
+ return self._call_impl(*args, **kwargs)
181
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
182
+ File "/scratch/work/palp3/myenv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl
183
+ return forward_call(*args, **kwargs)
184
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
185
+ File "/scratch/work/palp3/myenv/lib/python3.11/site-packages/transformers/models/wav2vec2/modeling_wav2vec2.py", line 335, in forward
186
+ hidden_states = self.layer_norm(hidden_states)
187
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
188
+ File "/scratch/work/palp3/myenv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl
189
+ return self._call_impl(*args, **kwargs)
190
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
191
+ File "/scratch/work/palp3/myenv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl
192
+ return forward_call(*args, **kwargs)
193
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
194
+ File "/scratch/work/palp3/myenv/lib/python3.11/site-packages/torch/nn/modules/normalization.py", line 202, in forward
195
+ return F.layer_norm(
196
+ ^^^^^^^^^^^^^
197
+ File "/scratch/work/palp3/myenv/lib/python3.11/site-packages/torch/nn/functional.py", line 2576, in layer_norm
198
+ return torch.layer_norm(input, normalized_shape, weight, bias, eps, torch.backends.cudnn.enabled)
199
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
200
+ torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 4.00 GiB. GPU 0 has a total capacity of 15.77 GiB of which 1.55 GiB is free. Including non-PyTorch memory, this process has 14.21 GiB memory in use. Of the allocated memory 11.68 GiB is allocated by PyTorch, and 2.17 GiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
201
+ wandb: - 0.011 MB of 0.011 MB uploaded
202
+ wandb: ⭐️ View project at: https://wandb.ai/priyanshipal/huggingface
203
+ wandb: Synced 6 W&B file(s), 0 media file(s), 0 artifact file(s) and 0 other file(s)
204
+ wandb: Find logs at: ./wandb/run-20240822_151726-alv0f5i7/logs
205
+ wandb: WARNING The new W&B backend becomes opt-out in version 0.18.0; try it out with `wandb.require("core")`! See https://wandb.me/wandb-core for more information.
evalonlyhindi_indicwav2vec_MUCS_warmup500_s300shuff100_2142429.out ADDED
@@ -0,0 +1,155 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
0
  0%| | 0/36 [00:00<?, ?it/s]
1
  6%|▌ | 2/36 [00:01<00:23, 1.44it/s]
2
  8%|▊ | 3/36 [00:02<00:31, 1.05it/s]
3
  11%|█ | 4/36 [00:04<00:40, 1.26s/it]
4
  14%|█▍ | 5/36 [00:06<00:42, 1.36s/it]
5
  17%|█▋ | 6/36 [00:07<00:40, 1.35s/it]
6
  19%|█▉ | 7/36 [00:08<00:34, 1.19s/it]
7
  22%|██▏ | 8/36 [00:08<00:26, 1.05it/s]
8
  25%|██▌ | 9/36 [00:09<00:22, 1.21it/s]
9
  28%|██▊ | 10/36 [00:09<00:20, 1.26it/s]
10
  31%|███ | 11/36 [00:10<00:20, 1.22it/s]
11
  33%|███▎ | 12/36 [00:11<00:19, 1.23it/s]
12
  36%|███▌ | 13/36 [00:12<00:17, 1.35it/s]
13
  39%|███▉ | 14/36 [00:12<00:14, 1.53it/s]
14
  42%|████▏ | 15/36 [00:13<00:12, 1.73it/s]
15
  44%|████▍ | 16/36 [00:13<00:10, 1.87it/s]
16
  47%|████▋ | 17/36 [00:13<00:09, 1.96it/s]
17
  50%|█████ | 18/36 [00:14<00:09, 1.89it/s]
18
  53%|█████▎ | 19/36 [00:15<00:09, 1.80it/s]
19
  56%|█████▌ | 20/36 [00:15<00:08, 1.85it/s]
20
  58%|█████▊ | 21/36 [00:16<00:07, 2.00it/s]
21
  61%|██████ | 22/36 [00:16<00:07, 1.99it/s]
22
  64%|██████▍ | 23/36 [00:17<00:06, 1.96it/s]
23
  67%|██████▋ | 24/36 [00:17<00:06, 1.94it/s]
24
  69%|██████▉ | 25/36 [00:18<00:05, 1.85it/s]
25
  72%|███████▏ | 26/36 [00:18<00:05, 1.87it/s]
26
  75%|███████▌ | 27/36 [00:19<00:04, 1.97it/s]
27
  78%|███████▊ | 28/36 [00:20<00:05, 1.59it/s]
28
  81%|████████ | 29/36 [00:21<00:06, 1.08it/s]
29
  83%|████████▎ | 30/36 [00:23<00:06, 1.08s/it]
30
  86%|████████▌ | 31/36 [00:25<00:06, 1.35s/it]
31
  89%|████████▉ | 32/36 [00:25<00:04, 1.12s/it]
32
  92%|█████████▏| 33/36 [00:26<00:02, 1.04it/s]
33
  94%|█████████▍| 34/36 [00:26<00:01, 1.21it/s]
34
  97%|█████████▋| 35/36 [00:27<00:00, 1.36it/s]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb: Currently logged in as: priyanshi-pal (priyanshipal). Use `wandb login --relogin` to force relogin
2
+ wandb: wandb version 0.17.7 is available! To upgrade, please run:
3
+ wandb: $ pip install wandb --upgrade
4
+ wandb: Tracking run with wandb version 0.17.6
5
+ wandb: Run data is saved locally in /scratch/elec/t405-puhe/p/palp3/MUCS/wandb/run-20240822_152047-64250v6u
6
+ wandb: Run `wandb offline` to turn off syncing.
7
+ wandb: Syncing run eval_pd2000_s300_shuff100_hindi
8
+ wandb: ⭐️ View project at https://wandb.ai/priyanshipal/huggingface
9
+ wandb: 🚀 View run at https://wandb.ai/priyanshipal/huggingface/runs/64250v6u
10
+ /scratch/work/palp3/myenv/lib/python3.11/site-packages/transformers/training_args.py:1525: FutureWarning: `evaluation_strategy` is deprecated and will be removed in version 4.46 of 🤗 Transformers. Use `eval_strategy` instead
11
+ warnings.warn(
12
+ /scratch/work/palp3/myenv/lib/python3.11/site-packages/transformers/models/auto/configuration_auto.py:957: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
13
+ warnings.warn(
14
+ /scratch/work/palp3/myenv/lib/python3.11/site-packages/transformers/models/auto/feature_extraction_auto.py:329: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
15
+ warnings.warn(
16
+ /scratch/work/palp3/myenv/lib/python3.11/site-packages/accelerate/accelerator.py:488: FutureWarning: `torch.cuda.amp.GradScaler(args...)` is deprecated. Please use `torch.amp.GradScaler('cuda', args...)` instead.
17
+ self.scaler = torch.cuda.amp.GradScaler(**kwargs)
18
+ max_steps is given, it will override any value given in num_train_epochs
19
+ Wav2Vec2CTCTokenizer(name_or_path='', vocab_size=149, model_max_length=1000000000000000019884624838656, is_fast=False, padding_side='right', truncation_side='right', special_tokens={'bos_token': '<s>', 'eos_token': '</s>', 'unk_token': '[UNK]', 'pad_token': '[PAD]'}, clean_up_tokenization_spaces=True), added_tokens_decoder={
20
+ 147: AddedToken("[UNK]", rstrip=True, lstrip=True, single_word=False, normalized=False, special=False),
21
+ 148: AddedToken("[PAD]", rstrip=True, lstrip=True, single_word=False, normalized=False, special=False),
22
+ 149: AddedToken("<s>", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
23
+ 150: AddedToken("</s>", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
24
+ }
25
+ CHECK MODEL PARAMS Wav2Vec2ForCTC(
26
+ (wav2vec2): Wav2Vec2Model(
27
+ (feature_extractor): Wav2Vec2FeatureEncoder(
28
+ (conv_layers): ModuleList(
29
+ (0): Wav2Vec2LayerNormConvLayer(
30
+ (conv): Conv1d(1, 512, kernel_size=(10,), stride=(5,))
31
+ (layer_norm): LayerNorm((512,), eps=1e-05, elementwise_affine=True)
32
+ (activation): GELUActivation()
33
+ )
34
+ (1-4): 4 x Wav2Vec2LayerNormConvLayer(
35
+ (conv): Conv1d(512, 512, kernel_size=(3,), stride=(2,))
36
+ (layer_norm): LayerNorm((512,), eps=1e-05, elementwise_affine=True)
37
+ (activation): GELUActivation()
38
+ )
39
+ (5-6): 2 x Wav2Vec2LayerNormConvLayer(
40
+ (conv): Conv1d(512, 512, kernel_size=(2,), stride=(2,))
41
+ (layer_norm): LayerNorm((512,), eps=1e-05, elementwise_affine=True)
42
+ (activation): GELUActivation()
43
+ )
44
+ )
45
+ )
46
+ (feature_projection): Wav2Vec2FeatureProjection(
47
+ (layer_norm): LayerNorm((512,), eps=1e-05, elementwise_affine=True)
48
+ (projection): Linear(in_features=512, out_features=1024, bias=True)
49
+ (dropout): Dropout(p=0.0, inplace=False)
50
+ )
51
+ (encoder): Wav2Vec2EncoderStableLayerNorm(
52
+ (pos_conv_embed): Wav2Vec2PositionalConvEmbedding(
53
+ (conv): ParametrizedConv1d(
54
+ 1024, 1024, kernel_size=(128,), stride=(1,), padding=(64,), groups=16
55
+ (parametrizations): ModuleDict(
56
+ (weight): ParametrizationList(
57
+ (0): _WeightNorm()
58
+ )
59
+ )
60
+ )
61
+ (padding): Wav2Vec2SamePadLayer()
62
+ (activation): GELUActivation()
63
+ )
64
+ (layer_norm): LayerNorm((1024,), eps=1e-05, elementwise_affine=True)
65
+ (dropout): Dropout(p=0.0, inplace=False)
66
+ (layers): ModuleList(
67
+ (0-23): 24 x Wav2Vec2EncoderLayerStableLayerNorm(
68
+ (attention): Wav2Vec2SdpaAttention(
69
+ (k_proj): Linear(in_features=1024, out_features=1024, bias=True)
70
+ (v_proj): Linear(in_features=1024, out_features=1024, bias=True)
71
+ (q_proj): Linear(in_features=1024, out_features=1024, bias=True)
72
+ (out_proj): Linear(in_features=1024, out_features=1024, bias=True)
73
+ )
74
+ (dropout): Dropout(p=0.0, inplace=False)
75
+ (layer_norm): LayerNorm((1024,), eps=1e-05, elementwise_affine=True)
76
+ (feed_forward): Wav2Vec2FeedForward(
77
+ (intermediate_dropout): Dropout(p=0.0, inplace=False)
78
+ (intermediate_dense): Linear(in_features=1024, out_features=4096, bias=True)
79
+ (intermediate_act_fn): GELUActivation()
80
+ (output_dense): Linear(in_features=4096, out_features=1024, bias=True)
81
+ (output_dropout): Dropout(p=0.0, inplace=False)
82
+ )
83
+ (final_layer_norm): LayerNorm((1024,), eps=1e-05, elementwise_affine=True)
84
+ )
85
+ )
86
+ )
87
+ )
88
+ (dropout): Dropout(p=0.0, inplace=False)
89
+ (lm_head): Linear(in_features=1024, out_features=151, bias=True)
90
+ )
91
+ check the eval set length 572
92
+ 08/22/2024 15:20:57 - INFO - __main__ - *** Evaluate ***
93
+ /scratch/work/palp3/myenv/lib/python3.11/site-packages/transformers/models/wav2vec2/processing_wav2vec2.py:157: UserWarning: `as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your labels by using the argument `text` of the regular `__call__` method (either in the same call as your audio inputs, or in a separate call.
94
+ warnings.warn(
95
+
96
  0%| | 0/36 [00:00<?, ?it/s]
97
  6%|▌ | 2/36 [00:01<00:23, 1.44it/s]
98
  8%|▊ | 3/36 [00:02<00:31, 1.05it/s]
99
  11%|█ | 4/36 [00:04<00:40, 1.26s/it]
100
  14%|█▍ | 5/36 [00:06<00:42, 1.36s/it]
101
  17%|█▋ | 6/36 [00:07<00:40, 1.35s/it]
102
  19%|█▉ | 7/36 [00:08<00:34, 1.19s/it]
103
  22%|██▏ | 8/36 [00:08<00:26, 1.05it/s]
104
  25%|██▌ | 9/36 [00:09<00:22, 1.21it/s]
105
  28%|██▊ | 10/36 [00:09<00:20, 1.26it/s]
106
  31%|███ | 11/36 [00:10<00:20, 1.22it/s]
107
  33%|███▎ | 12/36 [00:11<00:19, 1.23it/s]
108
  36%|███▌ | 13/36 [00:12<00:17, 1.35it/s]
109
  39%|███▉ | 14/36 [00:12<00:14, 1.53it/s]
110
  42%|████▏ | 15/36 [00:13<00:12, 1.73it/s]
111
  44%|████▍ | 16/36 [00:13<00:10, 1.87it/s]
112
  47%|████▋ | 17/36 [00:13<00:09, 1.96it/s]
113
  50%|█████ | 18/36 [00:14<00:09, 1.89it/s]
114
  53%|█████▎ | 19/36 [00:15<00:09, 1.80it/s]
115
  56%|█████▌ | 20/36 [00:15<00:08, 1.85it/s]
116
  58%|█████▊ | 21/36 [00:16<00:07, 2.00it/s]
117
  61%|██████ | 22/36 [00:16<00:07, 1.99it/s]
118
  64%|██████▍ | 23/36 [00:17<00:06, 1.96it/s]
119
  67%|██████▋ | 24/36 [00:17<00:06, 1.94it/s]
120
  69%|██████▉ | 25/36 [00:18<00:05, 1.85it/s]
121
  72%|███████▏ | 26/36 [00:18<00:05, 1.87it/s]
122
  75%|███████▌ | 27/36 [00:19<00:04, 1.97it/s]
123
  78%|███████▊ | 28/36 [00:20<00:05, 1.59it/s]
124
  81%|████████ | 29/36 [00:21<00:06, 1.08it/s]
125
  83%|████████▎ | 30/36 [00:23<00:06, 1.08s/it]
126
  86%|████████▌ | 31/36 [00:25<00:06, 1.35s/it]
127
  89%|████████▉ | 32/36 [00:25<00:04, 1.12s/it]
128
  92%|█████████▏| 33/36 [00:26<00:02, 1.04it/s]
129
  94%|█████████▍| 34/36 [00:26<00:01, 1.21it/s]
130
  97%|█████████▋| 35/36 [00:27<00:00, 1.36it/s]
131
+ Printing predictions for a few samples:
132
+ Sample 1:
133
+ Reference: हम उनका उपयोग ऐसे ही कर सकते हैं या आवश्यकता अनुसार कुछ बदलाव करके उपयोग कर सकते हैं
134
+ ######
135
+
136
+
137
+ Prediction:
138
+
139
+
140
+
141
+ Sample 2:
142
+ Reference: अतः शीर्षक इस तरह से जोड़ सकते हैं
143
+ ######
144
+
145
+
146
+ Prediction:
147
+
148
+
149
+
150
+ Sample 3:
151
+ Reference: प्रेसेंटेशन के अंत में आपने स्लाइड की एक कॉपी बना ली है
152
+ ######
153
+
154
+
155
+ Prediction:
156
+
157
+
158
+
159
+ Sample 4:
160
+ Reference: चलिए अब फोंट्स और फोंट्स को फॉर्मेट करने के कुछ तरीके देखते हैं
161
+ ######
162
+
163
+
164
+ Prediction:
165
+
166
+
167
+
168
+ Sample 5:
169
+ Reference: यह एक डायलॉग बॉक्स खोलेगा जिसमें हम अपनी आवश्यकतानुसार फॉन्ट स्टाइल और साइज़ सेट कर सकते हैं
170
+ ######
171
+
172
+
173
+ Prediction:
174
+
175
+
176
+
177
+ last Reference string यह स्क्रिप्ट लता द्वारा अनुवादित है आईआईटी मुंबई की ओर से मैं रवि कुमार अब आपसे विदा लेता हूँहमसे जुड़ने के लिए धन्यवाद
178
+
179
+
180
+ last prediction string
181
+ ***** eval metrics *****
182
+ eval_cer = 1.0
183
+ eval_loss = nan
184
+ eval_model_preparation_time = 0.0045
185
+ eval_runtime = 0:00:30.92
186
+ eval_samples = 572
187
+ eval_samples_per_second = 18.498
188
+ eval_steps_per_second = 1.164
189
+ eval_wer = 1.0
190
+
language_segregated_prediction_texts/evalpredictions_hindi_indicw2v_ad0_3_hd_02_featd_0_2_lr6e-4_warmup500_s300_shuf100.txt CHANGED
The diff for this file is too large to render. See raw diff
 
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2ac994ec93d9297038ccd8f0941bd5189ab98d5a22407edafc2b896159805ce1
3
  size 5432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:17f546c82c9065a6469d28f2bf552ce06c87b07db0f04de15435295aa0a8a04d
3
  size 5432