Update README.md
Browse files
README.md
CHANGED
@@ -106,12 +106,15 @@ img {
|
|
106 |
|
107 |
## Overview
|
108 |
|
109 |
-
This model is a fine-tuned version of the [NVIDIA French FastConformer Hybrid Large model](https://huggingface.co/nvidia/stt_fr_fastconformer_hybrid_large_pc).
|
|
|
110 |
|
111 |
Compared to the base model, this version:
|
112 |
- Does **not** include punctuation or uppercase letters.
|
113 |
- Was trained on **9,500+ hours** of diverse, manually transcribed French speech.
|
114 |
|
|
|
|
|
115 |
---
|
116 |
|
117 |
## Performance
|
@@ -173,6 +176,8 @@ asr_model.transcribe([audio_path])
|
|
173 |
|
174 |
## Datasets
|
175 |
|
|
|
|
|
176 |
The model was trained on over 9,500 hours of French speech, covering:
|
177 |
- Read and spontaneous speech
|
178 |
- Conversations and meetings
|
|
|
106 |
|
107 |
## Overview
|
108 |
|
109 |
+
This model is a fine-tuned version of the [NVIDIA French FastConformer Hybrid Large model](https://huggingface.co/nvidia/stt_fr_fastconformer_hybrid_large_pc).
|
110 |
+
It is a large (115M parameters) hybrid ASR model trained with both **Transducer (default)** and **CTC** losses.
|
111 |
|
112 |
Compared to the base model, this version:
|
113 |
- Does **not** include punctuation or uppercase letters.
|
114 |
- Was trained on **9,500+ hours** of diverse, manually transcribed French speech.
|
115 |
|
116 |
+
The training code is available in the [nemo asr training repository](https://github.com/linagora-labs/nemo_asr_training).
|
117 |
+
|
118 |
---
|
119 |
|
120 |
## Performance
|
|
|
176 |
|
177 |
## Datasets
|
178 |
|
179 |
+
The data were transformed, processed and converted using [NeMo tools from the SSAK repository](https://github.com/linagora-labs/ssak/tree/main/tools/nemo)
|
180 |
+
|
181 |
The model was trained on over 9,500 hours of French speech, covering:
|
182 |
- Read and spontaneous speech
|
183 |
- Conversations and meetings
|