Beto, Bentz, Becas: The Surprising Cross-Lingual Effectiveness of BERT
About
Pretrained contextual representation models (Peters et al., 2018; Devlin et al., 2018) have pushed forward the state-of-the-art on many NLP tasks. A new release of BERT (Devlin, 2018) includes a model simultaneously pretrained on 104 languages with impressive performance for zero-shot cross-lingual transfer on a natural language inference task. This paper explores the broader cross-lingual potential of mBERT (multilingual) as a zero shot language transfer model on 5 NLP tasks covering a total of 39 languages from various language families: NLI, document classification, NER, POS tagging, and dependency parsing. We compare mBERT with the best-published methods for zero-shot cross-lingual transfer and find mBERT competitive on each task. Additionally, we investigate the most effective strategy for utilizing mBERT in this manner, determine to what extent mBERT generalizes away from language specific features, and measure factors that influence cross-lingual transfer.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Named Entity Recognition | CoNLL Spanish NER 2002 (test) | F1 Score76.34 | 98 | |
| Named Entity Recognition | CoNLL Dutch 2002 (test) | F1 Score80.61 | 87 | |
| Named Entity Recognition | CoNLL German 2003 (test) | F1 Score71.1 | 78 | |
| Named Entity Recognition | CoNLL NER 2002/2003 (test) | German F1 Score69.56 | 59 | |
| Named Entity Recognition | WikiAnn (test) | -- | 58 | |
| Named Entity Recognition | CoNLL (test) | F1 Score (de)69.56 | 28 | |
| Named Entity Recognition | CoNLL de 2003 (test) | F1 Score72.39 | 12 | |
| Named Entity Recognition | CoNLL-2002 (es, nl), CoNLL-2003 (de), Europeana Newspapers (fr), MSRA (zh) (test) | F1 Score (es)74.96 | 8 | |
| Named Entity Recognition | CoNLL-2002 (test) | F1 Score90.94 | 7 | |
| Named Entity Recognition | NoDaLiDa no 2019 (test) | F1 Score78.47 | 3 |