Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

MAD-X: An Adapter-Based Framework for Multi-Task Cross-Lingual Transfer

About

The main goal behind state-of-the-art pre-trained multilingual models such as multilingual BERT and XLM-R is enabling and bootstrapping NLP applications in low-resource languages through zero-shot or few-shot cross-lingual transfer. However, due to limited model capacity, their transfer performance is the weakest exactly on such low-resource languages and languages unseen during pre-training. We propose MAD-X, an adapter-based framework that enables high portability and parameter-efficient transfer to arbitrary tasks and languages by learning modular language and task representations. In addition, we introduce a novel invertible adapter architecture and a strong baseline method for adapting a pre-trained multilingual model to a new language. MAD-X outperforms the state of the art in cross-lingual transfer across a representative set of typologically diverse languages on named entity recognition and causal commonsense reasoning, and achieves competitive results on question answering. Our code and adapters are available at AdapterHub.ml

Jonas Pfeiffer, Ivan Vuli\'c, Iryna Gurevych, Sebastian Ruder• 2020

Related benchmarks

TaskDatasetResultRank
Natural Language UnderstandingGLUE (test)
SST-2 Accuracy53.3
416
Sentiment AnalysisIMDB (test)
Accuracy55.4
248
Commonsense ReasoningCommonsense Reasoning (BoolQ, PIQA, SIQA, HellaS., WinoG., ARC-e, ARC-c, OBQA) (test)
BoolQ Accuracy68.38
138
Commonsense Question AnsweringCSQA (test)
Accuracy0.341
127
Named Entity RecognitionWikiAnn (test)
Average Accuracy57.83
58
Fact VerificationFEVER (test)--
32
Relation ExtractionChemProt (test)
Micro F153.7
25
Causal ReasoningXCOPA (test)
Accuracy (id)65.8
13
Question AnsweringXQuAD
F1 (ar)65.5
12
Part-of-Speech TaggingUniversal Dependencies (UD) (test)--
12
Showing 10 of 30 rows

Other info

Code

Follow for update