Training in progress, step 800
Browse files- .gitattributes +0 -2
- .gitignore +1 -0
- README.md +93 -196
- config.json +109 -0
- model.safetensors +3 -0
- pytorch_model.bin +3 -0
- runs/Mar09_21-39-55_b90f0b03ec3b/events.out.tfevents.1741556994.b90f0b03ec3b.1202.0 +3 -0
- runs/Oct22_12-20-14_035cd7a4c2da/1666441655.266412/events.out.tfevents.1666441655.035cd7a4c2da.486.1 +3 -0
- runs/Oct22_12-20-14_035cd7a4c2da/events.out.tfevents.1666441655.035cd7a4c2da.486.0 +3 -0
- runs/Oct22_17-03-58_9e5f78d3a62d/1666459123.7949953/events.out.tfevents.1666459123.9e5f78d3a62d.75.1 +3 -0
- runs/Oct22_17-03-58_9e5f78d3a62d/events.out.tfevents.1666459123.9e5f78d3a62d.75.0 +3 -0
- runs/Oct23_02-58-56_b807f884dd1e/1666494605.8069742/events.out.tfevents.1666494605.b807f884dd1e.75.1 +3 -0
- runs/Oct23_02-58-56_b807f884dd1e/1666504762.0300505/events.out.tfevents.1666504762.b807f884dd1e.75.2 +3 -0
- runs/Oct23_02-58-56_b807f884dd1e/events.out.tfevents.1666494605.b807f884dd1e.75.0 +3 -0
- runs/Oct23_09-23-11_15f15819e94a/1666518299.4564846/events.out.tfevents.1666518299.15f15819e94a.75.1 +3 -0
- runs/Oct23_09-23-11_15f15819e94a/events.out.tfevents.1666518299.15f15819e94a.75.0 +3 -0
- runs/Oct24_08-15-22_2c1d225087a7/1666599964.3337321/events.out.tfevents.1666599964.2c1d225087a7.72.1 +3 -0
- runs/Oct24_08-15-22_2c1d225087a7/events.out.tfevents.1666599964.2c1d225087a7.72.0 +3 -0
- runs/Oct25_06-20-16_bff6b2401ae2/1666679480.4704952/events.out.tfevents.1666679480.bff6b2401ae2.76.1 +3 -0
- runs/Oct25_06-20-16_bff6b2401ae2/events.out.tfevents.1666679480.bff6b2401ae2.76.0 +3 -0
- special_tokens_map.json +16 -0
- tokenizer_config.json +2 -38
- training_args.bin +3 -0
.gitattributes
CHANGED
@@ -2,7 +2,6 @@
|
|
2 |
*.arrow filter=lfs diff=lfs merge=lfs -text
|
3 |
*.bin filter=lfs diff=lfs merge=lfs -text
|
4 |
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
5 |
-
*.ckpt filter=lfs diff=lfs merge=lfs -text
|
6 |
*.ftz filter=lfs diff=lfs merge=lfs -text
|
7 |
*.gz filter=lfs diff=lfs merge=lfs -text
|
8 |
*.h5 filter=lfs diff=lfs merge=lfs -text
|
@@ -25,7 +24,6 @@
|
|
25 |
*.safetensors filter=lfs diff=lfs merge=lfs -text
|
26 |
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
27 |
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
28 |
-
*.tar filter=lfs diff=lfs merge=lfs -text
|
29 |
*.tflite filter=lfs diff=lfs merge=lfs -text
|
30 |
*.tgz filter=lfs diff=lfs merge=lfs -text
|
31 |
*.wasm filter=lfs diff=lfs merge=lfs -text
|
|
|
2 |
*.arrow filter=lfs diff=lfs merge=lfs -text
|
3 |
*.bin filter=lfs diff=lfs merge=lfs -text
|
4 |
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
|
|
5 |
*.ftz filter=lfs diff=lfs merge=lfs -text
|
6 |
*.gz filter=lfs diff=lfs merge=lfs -text
|
7 |
*.h5 filter=lfs diff=lfs merge=lfs -text
|
|
|
24 |
*.safetensors filter=lfs diff=lfs merge=lfs -text
|
25 |
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
26 |
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
|
|
27 |
*.tflite filter=lfs diff=lfs merge=lfs -text
|
28 |
*.tgz filter=lfs diff=lfs merge=lfs -text
|
29 |
*.wasm filter=lfs diff=lfs merge=lfs -text
|
.gitignore
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
checkpoint-*/
|
README.md
CHANGED
@@ -1,199 +1,96 @@
|
|
1 |
---
|
2 |
-
|
3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
4 |
---
|
5 |
|
6 |
-
|
7 |
-
|
8 |
-
|
9 |
-
|
10 |
-
|
11 |
-
|
12 |
-
|
13 |
-
|
14 |
-
|
15 |
-
|
16 |
-
|
17 |
-
|
18 |
-
|
19 |
-
|
20 |
-
|
21 |
-
|
22 |
-
|
23 |
-
-
|
24 |
-
|
25 |
-
|
26 |
-
|
27 |
-
|
28 |
-
|
29 |
-
|
30 |
-
|
31 |
-
|
32 |
-
|
33 |
-
|
34 |
-
|
35 |
-
|
36 |
-
|
37 |
-
|
38 |
-
|
39 |
-
|
40 |
-
|
41 |
-
|
42 |
-
|
43 |
-
|
44 |
-
|
45 |
-
|
46 |
-
|
47 |
-
|
48 |
-
|
49 |
-
|
50 |
-
|
51 |
-
|
52 |
-
|
53 |
-
|
54 |
-
|
55 |
-
|
56 |
-
|
57 |
-
|
58 |
-
|
59 |
-
|
60 |
-
|
61 |
-
|
62 |
-
|
63 |
-
|
64 |
-
|
65 |
-
|
66 |
-
|
67 |
-
|
68 |
-
|
69 |
-
|
70 |
-
## How to Get Started with the Model
|
71 |
-
|
72 |
-
Use the code below to get started with the model.
|
73 |
-
|
74 |
-
[More Information Needed]
|
75 |
-
|
76 |
-
## Training Details
|
77 |
-
|
78 |
-
### Training Data
|
79 |
-
|
80 |
-
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
|
81 |
-
|
82 |
-
[More Information Needed]
|
83 |
-
|
84 |
-
### Training Procedure
|
85 |
-
|
86 |
-
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
|
87 |
-
|
88 |
-
#### Preprocessing [optional]
|
89 |
-
|
90 |
-
[More Information Needed]
|
91 |
-
|
92 |
-
|
93 |
-
#### Training Hyperparameters
|
94 |
-
|
95 |
-
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
|
96 |
-
|
97 |
-
#### Speeds, Sizes, Times [optional]
|
98 |
-
|
99 |
-
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
|
100 |
-
|
101 |
-
[More Information Needed]
|
102 |
-
|
103 |
-
## Evaluation
|
104 |
-
|
105 |
-
<!-- This section describes the evaluation protocols and provides the results. -->
|
106 |
-
|
107 |
-
### Testing Data, Factors & Metrics
|
108 |
-
|
109 |
-
#### Testing Data
|
110 |
-
|
111 |
-
<!-- This should link to a Dataset Card if possible. -->
|
112 |
-
|
113 |
-
[More Information Needed]
|
114 |
-
|
115 |
-
#### Factors
|
116 |
-
|
117 |
-
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
|
118 |
-
|
119 |
-
[More Information Needed]
|
120 |
-
|
121 |
-
#### Metrics
|
122 |
-
|
123 |
-
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
|
124 |
-
|
125 |
-
[More Information Needed]
|
126 |
-
|
127 |
-
### Results
|
128 |
-
|
129 |
-
[More Information Needed]
|
130 |
-
|
131 |
-
#### Summary
|
132 |
-
|
133 |
-
|
134 |
-
|
135 |
-
## Model Examination [optional]
|
136 |
-
|
137 |
-
<!-- Relevant interpretability work for the model goes here -->
|
138 |
-
|
139 |
-
[More Information Needed]
|
140 |
-
|
141 |
-
## Environmental Impact
|
142 |
-
|
143 |
-
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
|
144 |
-
|
145 |
-
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
|
146 |
-
|
147 |
-
- **Hardware Type:** [More Information Needed]
|
148 |
-
- **Hours used:** [More Information Needed]
|
149 |
-
- **Cloud Provider:** [More Information Needed]
|
150 |
-
- **Compute Region:** [More Information Needed]
|
151 |
-
- **Carbon Emitted:** [More Information Needed]
|
152 |
-
|
153 |
-
## Technical Specifications [optional]
|
154 |
-
|
155 |
-
### Model Architecture and Objective
|
156 |
-
|
157 |
-
[More Information Needed]
|
158 |
-
|
159 |
-
### Compute Infrastructure
|
160 |
-
|
161 |
-
[More Information Needed]
|
162 |
-
|
163 |
-
#### Hardware
|
164 |
-
|
165 |
-
[More Information Needed]
|
166 |
-
|
167 |
-
#### Software
|
168 |
-
|
169 |
-
[More Information Needed]
|
170 |
-
|
171 |
-
## Citation [optional]
|
172 |
-
|
173 |
-
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
|
174 |
-
|
175 |
-
**BibTeX:**
|
176 |
-
|
177 |
-
[More Information Needed]
|
178 |
-
|
179 |
-
**APA:**
|
180 |
-
|
181 |
-
[More Information Needed]
|
182 |
-
|
183 |
-
## Glossary [optional]
|
184 |
-
|
185 |
-
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
|
186 |
-
|
187 |
-
[More Information Needed]
|
188 |
-
|
189 |
-
## More Information [optional]
|
190 |
-
|
191 |
-
[More Information Needed]
|
192 |
-
|
193 |
-
## Model Card Authors [optional]
|
194 |
-
|
195 |
-
[More Information Needed]
|
196 |
-
|
197 |
-
## Model Card Contact
|
198 |
-
|
199 |
-
[More Information Needed]
|
|
|
1 |
---
|
2 |
+
language:
|
3 |
+
- ne
|
4 |
+
license: apache-2.0
|
5 |
+
tags:
|
6 |
+
- generated_from_trainer
|
7 |
+
- automatic-speech-recognition
|
8 |
+
- speech
|
9 |
+
- openslr
|
10 |
+
- nepali
|
11 |
+
datasets:
|
12 |
+
- spktsagar/openslr-nepali-asr-cleaned
|
13 |
+
metrics:
|
14 |
+
- wer
|
15 |
+
base_model: facebook/wav2vec2-xls-r-300m
|
16 |
+
model-index:
|
17 |
+
- name: wav2vec2-large-xls-r-300m-nepali-openslr
|
18 |
+
results:
|
19 |
+
- task:
|
20 |
+
type: automatic-speech-recognition
|
21 |
+
name: Nepali Speech Recognition
|
22 |
+
dataset:
|
23 |
+
name: OpenSLR Nepali ASR
|
24 |
+
type: spktsagar/openslr-nepali-asr-cleaned
|
25 |
+
config: original
|
26 |
+
split: train
|
27 |
+
metrics:
|
28 |
+
- type: were
|
29 |
+
value: 21.27
|
30 |
+
name: Test WER
|
31 |
+
verified: false
|
32 |
---
|
33 |
|
34 |
+
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
35 |
+
should probably proofread and complete it, then remove this comment. -->
|
36 |
+
|
37 |
+
# wav2vec2-large-xls-r-300m-nepali-openslr
|
38 |
+
|
39 |
+
This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on an [OpenSLR Nepali ASR](https://huggingface.co/datasets/spktsagar/openslr-nepali-asr-cleaned) dataset.
|
40 |
+
It achieves the following results on the evaluation set:
|
41 |
+
- eval_loss: 0.1767
|
42 |
+
- eval_wer: 0.2127
|
43 |
+
- eval_runtime: 595.3962
|
44 |
+
- eval_samples_per_second: 36.273
|
45 |
+
- eval_steps_per_second: 4.535
|
46 |
+
- epoch: 6.07
|
47 |
+
- step: 23200
|
48 |
+
|
49 |
+
## Model description
|
50 |
+
|
51 |
+
Wav2Vec2 is a pretrained model for Automatic Speech Recognition (ASR) and was released in September 2020 by Alexei Baevski, Michael Auli, and Alex Conneau. Soon after the superior performance of Wav2Vec2 was demonstrated on one of the most popular English datasets for ASR, called LibriSpeech, Facebook AI presented a multi-lingual version of Wav2Vec2, called XLSR. XLSR stands for cross-lingual speech representations and refers to model's ability to learn speech representations that are useful across multiple languages.
|
52 |
+
|
53 |
+
## How to use?
|
54 |
+
1. Install transformers and librosa
|
55 |
+
```
|
56 |
+
pip install librosa, transformers
|
57 |
+
```
|
58 |
+
2. Run the following code which loads your audio file, preprocessor, models, and returns your prediction
|
59 |
+
```python
|
60 |
+
import librosa
|
61 |
+
from transformers import pipeline
|
62 |
+
|
63 |
+
audio, sample_rate = librosa.load("<path to your audio file>", sr=16000)
|
64 |
+
recognizer = pipeline("automatic-speech-recognition", model="spktsagar/wav2vec2-large-xls-r-300m-nepali-openslr")
|
65 |
+
prediction = recognizer(audio)
|
66 |
+
```
|
67 |
+
|
68 |
+
## Intended uses & limitations
|
69 |
+
|
70 |
+
The model is trained on the OpenSLR Nepali ASR dataset, which in itself has some incorrect transcriptions, so it is obvious that the model will not have perfect predictions for your transcript. Similarly, due to colab's resource limit utterances longer than 5 sec are filtered out from the dataset during training and evaluation. Hence, the model might not perform as expected when given audio input longer than 5 sec.
|
71 |
+
|
72 |
+
## Training and evaluation data and Training procedure
|
73 |
+
|
74 |
+
For dataset preparation and training code, please consult [my blog](https://sagar-spkt.github.io/posts/2022/08/finetune-xlsr-nepali/).
|
75 |
+
|
76 |
+
### Training hyperparameters
|
77 |
+
|
78 |
+
The following hyperparameters were used during training:
|
79 |
+
- learning_rate: 0.0003
|
80 |
+
- train_batch_size: 16
|
81 |
+
- eval_batch_size: 8
|
82 |
+
- seed: 42
|
83 |
+
- gradient_accumulation_steps: 2
|
84 |
+
- total_train_batch_size: 32
|
85 |
+
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
86 |
+
- lr_scheduler_type: linear
|
87 |
+
- lr_scheduler_warmup_steps: 500
|
88 |
+
- num_epochs: 10
|
89 |
+
- mixed_precision_training: Native AMP
|
90 |
+
|
91 |
+
### Framework versions
|
92 |
+
|
93 |
+
- Transformers 4.23.1
|
94 |
+
- Pytorch 1.11.0+cu113
|
95 |
+
- Datasets 2.6.0
|
96 |
+
- Tokenizers 0.13.1
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
config.json
ADDED
@@ -0,0 +1,109 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "facebook/wav2vec2-xls-r-300m",
|
3 |
+
"activation_dropout": 0.0,
|
4 |
+
"adapter_attn_dim": null,
|
5 |
+
"adapter_kernel_size": 3,
|
6 |
+
"adapter_stride": 2,
|
7 |
+
"add_adapter": false,
|
8 |
+
"apply_spec_augment": true,
|
9 |
+
"architectures": [
|
10 |
+
"Wav2Vec2ForCTC"
|
11 |
+
],
|
12 |
+
"attention_dropout": 0.1,
|
13 |
+
"bos_token_id": 1,
|
14 |
+
"classifier_proj_size": 256,
|
15 |
+
"codevector_dim": 768,
|
16 |
+
"contrastive_logits_temperature": 0.1,
|
17 |
+
"conv_bias": true,
|
18 |
+
"conv_dim": [
|
19 |
+
512,
|
20 |
+
512,
|
21 |
+
512,
|
22 |
+
512,
|
23 |
+
512,
|
24 |
+
512,
|
25 |
+
512
|
26 |
+
],
|
27 |
+
"conv_kernel": [
|
28 |
+
10,
|
29 |
+
3,
|
30 |
+
3,
|
31 |
+
3,
|
32 |
+
3,
|
33 |
+
2,
|
34 |
+
2
|
35 |
+
],
|
36 |
+
"conv_stride": [
|
37 |
+
5,
|
38 |
+
2,
|
39 |
+
2,
|
40 |
+
2,
|
41 |
+
2,
|
42 |
+
2,
|
43 |
+
2
|
44 |
+
],
|
45 |
+
"ctc_loss_reduction": "mean",
|
46 |
+
"ctc_zero_infinity": false,
|
47 |
+
"diversity_loss_weight": 0.1,
|
48 |
+
"do_stable_layer_norm": true,
|
49 |
+
"eos_token_id": 2,
|
50 |
+
"feat_extract_activation": "gelu",
|
51 |
+
"feat_extract_dropout": 0.0,
|
52 |
+
"feat_extract_norm": "layer",
|
53 |
+
"feat_proj_dropout": 0.0,
|
54 |
+
"feat_quantizer_dropout": 0.0,
|
55 |
+
"final_dropout": 0.0,
|
56 |
+
"gradient_checkpointing": false,
|
57 |
+
"hidden_act": "gelu",
|
58 |
+
"hidden_dropout": 0.1,
|
59 |
+
"hidden_size": 1024,
|
60 |
+
"initializer_range": 0.02,
|
61 |
+
"intermediate_size": 4096,
|
62 |
+
"layer_norm_eps": 1e-05,
|
63 |
+
"layerdrop": 0.1,
|
64 |
+
"mask_feature_length": 10,
|
65 |
+
"mask_feature_min_masks": 0,
|
66 |
+
"mask_feature_prob": 0.0,
|
67 |
+
"mask_time_length": 10,
|
68 |
+
"mask_time_min_masks": 2,
|
69 |
+
"mask_time_prob": 0.075,
|
70 |
+
"model_type": "wav2vec2",
|
71 |
+
"num_adapter_layers": 3,
|
72 |
+
"num_attention_heads": 16,
|
73 |
+
"num_codevector_groups": 2,
|
74 |
+
"num_codevectors_per_group": 320,
|
75 |
+
"num_conv_pos_embedding_groups": 16,
|
76 |
+
"num_conv_pos_embeddings": 128,
|
77 |
+
"num_feat_extract_layers": 7,
|
78 |
+
"num_hidden_layers": 24,
|
79 |
+
"num_negatives": 100,
|
80 |
+
"output_hidden_size": 1024,
|
81 |
+
"pad_token_id": 0,
|
82 |
+
"proj_codevector_dim": 768,
|
83 |
+
"tdnn_dilation": [
|
84 |
+
1,
|
85 |
+
2,
|
86 |
+
3,
|
87 |
+
1,
|
88 |
+
1
|
89 |
+
],
|
90 |
+
"tdnn_dim": [
|
91 |
+
512,
|
92 |
+
512,
|
93 |
+
512,
|
94 |
+
512,
|
95 |
+
1500
|
96 |
+
],
|
97 |
+
"tdnn_kernel": [
|
98 |
+
5,
|
99 |
+
3,
|
100 |
+
3,
|
101 |
+
1,
|
102 |
+
1
|
103 |
+
],
|
104 |
+
"torch_dtype": "float32",
|
105 |
+
"transformers_version": "4.48.3",
|
106 |
+
"use_weighted_layer_sum": false,
|
107 |
+
"vocab_size": 81,
|
108 |
+
"xvector_output_dim": 512
|
109 |
+
}
|
model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8211f7bc58b3f30039295d4a1ed929a19ae83d0336ba5a667231665863d41ef2
|
3 |
+
size 1262139580
|
pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ef42848298c4b318e8e604147157d60754b7a8a5fef300f65d9319698aaa4c09
|
3 |
+
size 1262230833
|
runs/Mar09_21-39-55_b90f0b03ec3b/events.out.tfevents.1741556994.b90f0b03ec3b.1202.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e198b86c971332792a4e9b7cd9469f85f27fbd2bdf73bd744fc9f83f224add0c
|
3 |
+
size 7121
|
runs/Oct22_12-20-14_035cd7a4c2da/1666441655.266412/events.out.tfevents.1666441655.035cd7a4c2da.486.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5f7174f2fc5cadfd96e2344960da1dfd45189255e8409cafb994941256e583b9
|
3 |
+
size 5570
|
runs/Oct22_12-20-14_035cd7a4c2da/events.out.tfevents.1666441655.035cd7a4c2da.486.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:388ad5fb67bcba7556b4ee122536f9d74ceac74b9f3158b22ca61457d97f7394
|
3 |
+
size 7599
|
runs/Oct22_17-03-58_9e5f78d3a62d/1666459123.7949953/events.out.tfevents.1666459123.9e5f78d3a62d.75.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e8407d2683164ddf9539aa9816869eb0349642136c24b61b19b651abb9b7c1fc
|
3 |
+
size 5570
|
runs/Oct22_17-03-58_9e5f78d3a62d/events.out.tfevents.1666459123.9e5f78d3a62d.75.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8973e14e77898e659c5113ae590a6aa8bbeccf681b2c6ffcb61cda1f8d265ba8
|
3 |
+
size 6649
|
runs/Oct23_02-58-56_b807f884dd1e/1666494605.8069742/events.out.tfevents.1666494605.b807f884dd1e.75.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:64f736d7ec82b4a82efd42db5da0bd50d7ea601780a0fec68dc91361bcaf3fcf
|
3 |
+
size 5570
|
runs/Oct23_02-58-56_b807f884dd1e/1666504762.0300505/events.out.tfevents.1666504762.b807f884dd1e.75.2
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e7269b4978e31abe1febcd3df6d38b026f30b8e5d7b4aa62add73c4084df33a4
|
3 |
+
size 5570
|
runs/Oct23_02-58-56_b807f884dd1e/events.out.tfevents.1666494605.b807f884dd1e.75.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9348d4e8397f26ac119fff4415b93e08c283f59261211122cc016c389d564dde
|
3 |
+
size 14200
|
runs/Oct23_09-23-11_15f15819e94a/1666518299.4564846/events.out.tfevents.1666518299.15f15819e94a.75.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:18e97497686fddf58495573af006a3e9f672ad006f7b97a535e3a9b03bac8711
|
3 |
+
size 5555
|
runs/Oct23_09-23-11_15f15819e94a/events.out.tfevents.1666518299.15f15819e94a.75.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0e0f98d57d34a32656460ba71b230b4102d126b320a8d31941f0526ef627855e
|
3 |
+
size 8077
|
runs/Oct24_08-15-22_2c1d225087a7/1666599964.3337321/events.out.tfevents.1666599964.2c1d225087a7.72.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4707b7dc2976819b7e80b554ae32f48d4571acb1f264b7d411c0a2a98bdc3a2c
|
3 |
+
size 5557
|
runs/Oct24_08-15-22_2c1d225087a7/events.out.tfevents.1666599964.2c1d225087a7.72.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d45322bb14e3fb23e0236f6d46bdf06a75cd78f824c1c285b4a762b3dc3502ce
|
3 |
+
size 6178
|
runs/Oct25_06-20-16_bff6b2401ae2/1666679480.4704952/events.out.tfevents.1666679480.bff6b2401ae2.76.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:22c55ec8df22a55b74139cb76fd8883111e84353c8a4b4d3d4fb249d771a8fc5
|
3 |
+
size 5557
|
runs/Oct25_06-20-16_bff6b2401ae2/events.out.tfevents.1666679480.bff6b2401ae2.76.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:391b61b16451a4fb7be5579607af65043c99c61ac2d76f86edd8d923ca472d39
|
3 |
+
size 7630
|
special_tokens_map.json
CHANGED
@@ -1,4 +1,20 @@
|
|
1 |
{
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
2 |
"bos_token": "<s>",
|
3 |
"eos_token": "</s>",
|
4 |
"pad_token": "__PAD__",
|
|
|
1 |
{
|
2 |
+
"additional_special_tokens": [
|
3 |
+
{
|
4 |
+
"content": "<s>",
|
5 |
+
"lstrip": false,
|
6 |
+
"normalized": true,
|
7 |
+
"rstrip": false,
|
8 |
+
"single_word": false
|
9 |
+
},
|
10 |
+
{
|
11 |
+
"content": "</s>",
|
12 |
+
"lstrip": false,
|
13 |
+
"normalized": true,
|
14 |
+
"rstrip": false,
|
15 |
+
"single_word": false
|
16 |
+
}
|
17 |
+
],
|
18 |
"bos_token": "<s>",
|
19 |
"eos_token": "</s>",
|
20 |
"pad_token": "__PAD__",
|
tokenizer_config.json
CHANGED
@@ -1,48 +1,12 @@
|
|
1 |
{
|
2 |
-
"added_tokens_decoder": {
|
3 |
-
"0": {
|
4 |
-
"content": "__PAD__",
|
5 |
-
"lstrip": true,
|
6 |
-
"normalized": false,
|
7 |
-
"rstrip": true,
|
8 |
-
"single_word": false,
|
9 |
-
"special": false
|
10 |
-
},
|
11 |
-
"1": {
|
12 |
-
"content": "__UNK__",
|
13 |
-
"lstrip": true,
|
14 |
-
"normalized": false,
|
15 |
-
"rstrip": true,
|
16 |
-
"single_word": false,
|
17 |
-
"special": false
|
18 |
-
},
|
19 |
-
"79": {
|
20 |
-
"content": "<s>",
|
21 |
-
"lstrip": false,
|
22 |
-
"normalized": false,
|
23 |
-
"rstrip": false,
|
24 |
-
"single_word": false,
|
25 |
-
"special": true
|
26 |
-
},
|
27 |
-
"80": {
|
28 |
-
"content": "</s>",
|
29 |
-
"lstrip": false,
|
30 |
-
"normalized": false,
|
31 |
-
"rstrip": false,
|
32 |
-
"single_word": false,
|
33 |
-
"special": true
|
34 |
-
}
|
35 |
-
},
|
36 |
"bos_token": "<s>",
|
37 |
-
"clean_up_tokenization_spaces": false,
|
38 |
"do_lower_case": false,
|
39 |
"eos_token": "</s>",
|
40 |
-
"
|
41 |
-
"model_max_length": 1000000000000000019884624838656,
|
42 |
"pad_token": "__PAD__",
|
43 |
"processor_class": "Wav2Vec2Processor",
|
44 |
"replace_word_delimiter_char": " ",
|
45 |
-
"
|
46 |
"tokenizer_class": "Wav2Vec2CTCTokenizer",
|
47 |
"unk_token": "__UNK__",
|
48 |
"word_delimiter_token": "|"
|
|
|
1 |
{
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
2 |
"bos_token": "<s>",
|
|
|
3 |
"do_lower_case": false,
|
4 |
"eos_token": "</s>",
|
5 |
+
"name_or_path": "./",
|
|
|
6 |
"pad_token": "__PAD__",
|
7 |
"processor_class": "Wav2Vec2Processor",
|
8 |
"replace_word_delimiter_char": " ",
|
9 |
+
"special_tokens_map_file": null,
|
10 |
"tokenizer_class": "Wav2Vec2CTCTokenizer",
|
11 |
"unk_token": "__UNK__",
|
12 |
"word_delimiter_token": "|"
|
training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bb81b64e9a1d2d59c74a6df7ad2b82bc89a9141d56666801bdc5eb0ca8ba8885
|
3 |
+
size 5368
|