Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Do Multilingual Language Models Think Better in English?

About

Translate-test is a popular technique to improve the performance of multilingual language models. This approach works by translating the input into English using an external machine translation system, and running inference over the translated input. However, these improvements can be attributed to the use of a separate translation system, which is typically trained on large amounts of parallel data not seen by the language model. In this work, we introduce a new approach called self-translate, which overcomes the need of an external translation system by leveraging the few-shot translation capabilities of multilingual language models. Experiments over 5 tasks show that self-translate consistently outperforms direct inference, demonstrating that language models are unable to leverage their full multilingual potential when prompted in non-English languages. Our code is available at https://github.com/juletx/self-translate.

Julen Etxaniz, Gorka Azkune, Aitor Soroa, Oier Lopez de Lacalle, Mikel Artetxe• 2023

Related benchmarks

TaskDatasetResultRank
Multilingual Mathematical ReasoningMGSM (test)
Accuracy64.5
57
Math ReasoningMSVAMP (test)
Average Accuracy72.1
45
Multilingual Mathematical ReasoningMT Math100
Accuracy60.08
24
Multilingual General KnowledgeGlobal MMLU Lite (subset of 18 languages)
Accuracy53.73
6
Multilingual Mathematical ReasoningMGSM 18 languages
Accuracy66.15
6
Multilingual Reasoning and General KnowledgeOverall (18 languages)
Accuracy57.01
6
Multilingual Reading ComprehensionBelebele 18 languages
Accuracy48.09
6
Showing 7 of 7 rows

Other info

Follow for update