Tiny Aya: Bridging Scale and Multilingual Depth
About
Tiny Aya redefines what a small multilingual language model can achieve. Trained on 70 languages and refined through region-aware posttraining, it delivers state-of-the-art in translation quality, strong multilingual understanding, and high-quality target-language generation, all with just 3.35B parameters. The release includes a pretrained foundation model, a globally balanced instruction-tuned variant, and three region-specialized models targeting languages from Africa, South Asia, Europe, Asia-Pacific, and West Asia. This report details the training strategy, data composition, and comprehensive evaluation framework behind Tiny Aya, and presents an alternative scaling path for multilingual AI: one centered on efficiency, balanced performance across languages, and practical deployment.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Safety Evaluation | MultiJail | Safe Response Rate95 | 66 | |
| Short Question Answering | BLEnD Short Question Answer | Average Accuracy41 | 18 | |
| Machine Translation | WMT 24++ | Standard Deviation of Score10 | 16 | |
| Machine Translation | BOUQuET XX-En | ChrF++ (high)62.7 | 16 | |
| Translation | FLoRes+ XX-En, high resource level | ChrF++61.3 | 16 | |
| Translation | FLoRes+ En-YY mid resource level | ChrF++27.6 | 16 | |
| Machine Translation | BOUQuET En-YY | ChrF++ (high)58.5 | 16 | |
| Translation | FLoRes+ En-YY, high resource level | ChrF++53.6 | 16 | |
| Translation | FLoRes+ XX-En mid resource level | ChrF++39.9 | 16 | |
| Translation | FLoRes+ XX-En total | ChrF++40.4 | 16 |