Aya Model: An Instruction Finetuned Open-Access Multilingual Language Model
About
Recent breakthroughs in large language models (LLMs) have centered around a handful of data-rich languages. What does it take to broaden access to breakthroughs beyond first-class citizen languages? Our work introduces Aya, a massively multilingual generative language model that follows instructions in 101 languages of which over 50% are considered as lower-resourced. Aya outperforms mT0 and BLOOMZ on the majority of tasks while covering double the number of languages. We introduce extensive new evaluation suites that broaden the state-of-art for multilingual eval across 99 languages -- including discriminative and generative tasks, human evaluation, and simulated win rates that cover both held-out tasks and in-distribution performance. Furthermore, we conduct detailed investigations on the optimal finetuning mixture composition, data pruning, as well as the toxicity, bias, and safety of our models. We open-source our instruction datasets and our model at https://hf.co/CohereForAI/aya-101
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Machine Translation | WMT En-Ja 2023 (test) | COMET84.6 | 12 | |
| Part-of-Speech Tagging | MasakhaPOS isiXhosa | Token Accuracy0.00e+0 | 12 | |
| Part-of-Speech Tagging | MasakhaPOS isiZulu | Token Accuracy0.00e+0 | 12 | |
| Part-of-Speech Tagging | MasakhaPOS Setswana | Token Accuracy0.00e+0 | 12 | |
| Topic Classification | MasakhaNEWS isiXhosa | Macro F194.6 | 11 | |
| Topic Classification | SIB-200 | Accuracy (Xho)82 | 11 | |
| Topic Classification | MasakhaNEWS English | Macro-F187.1 | 11 | |
| Intent Classification | INJONGO Intent | Accuracy (Eng)70.7 | 11 | |
| Named Entity Recognition | MasakhaNER 2.0 | Macro-F1 Score0.00e+0 | 11 | |
| Named Entity Recognition | MasakhaNER isiXhosa 2.0 | Macro F10.00e+0 | 11 |