Whisper-LM: Improving ASR Models with Language Models for Low-Resource Languages
Abstract
Automatic speech recognition systems have undoubtedly advanced with the integration of multilingual and multitask models such as Whisper, which have shown a promising ability to understand and process speech across a wide range of languages. Despite their robustness, these models often fall short in handling the linguistic distinctions of minority languages. This study addresses this gap by integrating traditional and novel language models with fine-tuned Whisper models to raise their performance in less commonly studied languages. Through rigorous fine-tuning and evaluation across multiple datasets, we demonstrate substantial improvements in word error rate, particularly in low-resource scenarios. Our approach not only does take advantage of the extensive data Whisper was pre-trained on, but also complements its linguistic adaptability by incorporating language models. We obtained improvements up to 51\% for in-distribution datasets and up to 34\% for out-of-distribution sentences using statistical language models, while large language models provided moderate but consistently robust improvement across diverse linguistic contexts. The findings reveal that, while the integration reliably benefits all model sizes, the extent of improvement varies, highlighting the importance of optimized language model parameters. Finally, we emphasize the importance of selecting appropriate evaluation parameters when reporting the results using transformer-based ASR models. In summary, this research clears the way for more inclusive ASR technologies that perform better across languages by enriching their linguistic knowledge. For further implementation details of this study, the technical documentation and source code are available at http://www.github.com/hitz-zentroa/whisper-lm.
Community
Hello!
We are excited to share our initial work on integrating n-gram and large language models with Whisper models. It is focused on (but not technically limited to) improving results in low-resource languages.
- The main code, for OpenAI models, used in our paper: https://github.com/hitz-zentroa/whisper-lm
- An alternative implementation using transformers: https://github.com/hitz-zentroa/whisper-lm-transformers
We welcome any questions, feedback, or ideas for improvement!
Hello again!
I have uploaded the n-gram models here: https://huggingface.co/HiTZ/whisper-lm-ngrams (thanks to @nielsr for helping me with this), and also linked the fine-tuned Whisper models we have used.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Whispering in Amharic: Fine-tuning Whisper for Low-resource Language (2025)
- Understanding Zero-shot Rare Word Recognition Improvements Through LLM Integration (2025)
- Is LLM the Silver Bullet to Low-Resource Languages Machine Translation? (2025)
- Improving Rare-Word Recognition of Whisper in Zero-Shot Settings (2024)
- UrduLLaMA 1.0: Dataset Curation, Preprocessing, and Evaluation in Low-Resource Settings (2025)
- UniBERTs: Adversarial Training for Language-Universal Representations (2025)
- On the Robust Approximation of ASR Metrics (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 29
Browse 29 models citing this paperDatasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper