From Llasa to Łazanki: Fine-tuned Llasa-1B on Polish Speech
This is a fine-tuned version of HKUSTAudio/Llasa-1B-Multilingual
, adapted for Polish Text-to-Speech (TTS).
It was fine-tuned on the pl-asr-bigos-v2
dataset, specifically the mozilla-common_voice_15-23
subset, which includes high-quality Polish speech recordings suitable for training TTS models.
🧠 Base Model
Llasa-1B-Multilingual model developed by HKUST. The approach leverages the LLAMA-initialized text BPE tokenizer, which can handle multilingual text without the need to design language-specific G2P (grapheme-to-phoneme) systems.
🗣 Fine-tuning Details
- Dataset: PL-ASR-BIGOS-v2,
mozilla-common_voice_15-23
subset - Language: 🇵🇱 Polish
- Task: Text to speech
- Downloads last month
- 49
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support
Model tree for GameRuiner/Llasa-1B-Multilingual-Polish
Base model
meta-llama/Llama-3.2-1B-Instruct
Finetuned
HKUSTAudio/Llasa-1B-Multilingual