Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Common Sense Beyond English: Evaluating and Improving Multilingual Language Models for Commonsense Reasoning

About

Commonsense reasoning research has so far been limited to English. We aim to evaluate and improve popular multilingual language models (ML-LMs) to help advance commonsense reasoning (CSR) beyond English. We collect the Mickey Corpus, consisting of 561k sentences in 11 different languages, which can be used for analyzing and improving ML-LMs. We propose Mickey Probe, a language-agnostic probing task for fairly evaluating the common sense of popular ML-LMs across different languages. In addition, we also create two new datasets, X-CSQA and X-CODAH, by translating their English versions to 15 other languages, so that we can evaluate popular ML-LMs for cross-lingual commonsense reasoning. To improve the performance beyond English, we propose a simple yet effective method -- multilingual contrastive pre-training (MCP). It significantly enhances sentence representations, yielding a large performance gain on both benchmarks.

Bill Yuchen Lin, Seyeon Lee, Xiaoyang Qiao, Xiang Ren• 2021

Related benchmarks

TaskDatasetResultRank
Question AnsweringX-CSQA
Accuracy (EN)69.5
6
Scene CompletionX-CODAH
Score (EN)69.9
6
MickeyProbeMickeyCorpus--
6
Showing 3 of 3 rows

Other info

Code

Follow for update