Polyglot-Lion: Efficient Multilingual ASR for Singapore via Balanced Fine-Tuning of Qwen3-ASR
arXiv cs.CL / 3/18/2026
📰 NewsTools & Practical UsageModels & Research
Key Points
- Polyglot-Lion is a family of compact multilingual ASR models tailored for Singapore’s linguistic landscape (English, Mandarin, Tamil, and Malay) and obtained by fine-tuning Qwen3-ASR models on publicly available data with balanced sampling and no language-tag conditioning.
- The approach balances the number of training utterances per language and lets the model infer languages from audio rather than relying on explicit tags.
- On 12 benchmarks across the four target languages, Polyglot-Lion-1.7B achieves an average error rate of 14.85, competitive with MERaLiON-2-10B-ASR despite the former being six times smaller.
- Training cost is dramatically lower ($81 on a single RTX PRO 6000 GPU versus $18,862 for the 128-GPU baseline) and inference throughput is about 20x faster (0.10 s/sample vs 2.02 s/sample).
- The results suggest linguistically balanced fine-tuning of moderate-scale pretrained models can yield deployment-ready multilingual ASR at a fraction of the cost of larger specialist systems.
Related Articles

ベテランの若手育成負担を減らせ、PLC制御の「ラダー図」をAIで生成
日経XTECH

Hey dev.to community – sharing my journey with Prompt Builder, Insta Posts, and practical SEO
Dev.to

Why Regex is Not Enough: Building a Deterministic "Sudo" Layer for AI Agents
Dev.to

Perplexity Hub
Dev.to

How to Build Passive Income with AI in 2026: A Developer's Practical Guide
Dev.to