AI Navigate

Tiny Aya: Bridging Scale and Multilingual Depth

arXiv cs.CL / 3/13/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • Tiny Aya is a 3.35B-parameter multilingual language model trained on 70 languages with region-aware post-training, delivering state-of-the-art translation quality and strong multilingual understanding.
  • The release includes a pretrained foundation model, a globally balanced instruction-tuned variant, and three region-specialized models covering Africa, South Asia, Europe, Asia-Pacific, and West Asia.
  • The paper outlines training strategy, data composition, and evaluation framework, and advocates an efficiency-centered scaling path prioritizing balanced multilingual performance and practical deployment.
  • It presents an alternative scaling path for multilingual AI that emphasizes efficiency, balanced performance across languages, and practical deployment considerations.

Abstract

Tiny Aya redefines what a small multilingual language model can achieve. Trained on 70 languages and refined through region-aware posttraining, it delivers state-of-the-art in translation quality, strong multilingual understanding, and high-quality target-language generation, all with just 3.35B parameters. The release includes a pretrained foundation model, a globally balanced instruction-tuned variant, and three region-specialized models targeting languages from Africa, South Asia, Europe, Asia-Pacific, and West Asia. This report details the training strategy, data composition, and comprehensive evaluation framework behind Tiny Aya, and presents an alternative scaling path for multilingual AI: one centered on efficiency, balanced performance across languages, and practical deployment.