Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Breaking Language Barriers in Multilingual Mathematical Reasoning: Insights and Observations

About

Existing research predominantly focuses on developing powerful language learning models (LLMs) for mathematical reasoning within monolingual languages, with few explorations in preserving efficacy in a multilingual context. To bridge this gap, this paper pioneers exploring and training powerful Multilingual Math Reasoning (xMR) LLMs. Firstly, by utilizing translation, we construct the first multilingual math reasoning instruction dataset, MGSM8KInstruct, encompassing ten distinct languages, thus addressing the issue of training data scarcity in xMR tasks. Based on the collected dataset, we propose different training strategies to build powerful xMR LLMs, named MathOctopus, notably outperform conventional open-source LLMs and exhibit superiority over ChatGPT in few-shot scenarios. Notably, MathOctopus-13B reaches 47.6% accuracy which exceeds ChatGPT 46.3% on MGSM testset. Beyond remarkable results, we unearth several pivotal observations and insights from extensive experiments: (1) When extending the rejection sampling strategy to the multilingual context, it proves effective for model performances, albeit limited. (2) Employing parallel corpora for math Supervised Fine-Tuning (SFT) across multiple languages not only significantly enhances model performance multilingually but also elevates their monolingual performance. This indicates that crafting multilingual corpora can be regarded as a vital strategy for enhancing model performance in a specific language, especially in mathematical reasoning tasks. For instance, MathOctopus-7B improves its counterparts that trained on English from 42.2% to 50.8% on GSM8K testset. Codes are available at https://github.com/microsoft/MathOctopus.

Nuo Chen, Zinan Zheng, Ning Wu, Ming Gong, Dongmei Zhang, Jia Li• 2023

Related benchmarks

TaskDatasetResultRank
Multilingual Mathematical ReasoningMGSM (test)
Accuracy69.2
57
Math ReasoningMSVAMP (test)
Average Accuracy73.2
45
Multilingual Mathematical ReasoningMGSM 1.0 (test)
Accuracy (ru)64
35
Multilingual Mathematical ReasoningMSVAMP
Accuracy (English)56.4
33
Open-domain Question AnsweringMKQA English Language (test)
NOB Accuracy44.92
5
Machine TranslationFLORES X-to-English
BLEU (CES)32.63
5
Machine TranslationFLORES English-to-X
BLEU (CES)22.68
5
Open-domain Question AnsweringMKQA Target Language (test)
NOB Accuracy39.73
5
Multilingual UnderstandingBELEBELE English Language
CES Score63.91
5
Multilingual UnderstandingBELEBELE Target Language
CES Performance48.22
5
Showing 10 of 10 rows

Other info

Follow for update