Improving Pretrained Cross-Lingual Language Models via Self-Labeled Word Alignment
About
The cross-lingual language models are typically pretrained with masked language modeling on multilingual text or parallel sentences. In this paper, we introduce denoising word alignment as a new cross-lingual pre-training task. Specifically, the model first self-labels word alignments for parallel sentences. Then we randomly mask tokens in a bitext pair. Given a masked token, the model uses a pointer network to predict the aligned token in the other language. We alternately perform the above two steps in an expectation-maximization manner. Experimental results show that our method improves cross-lingual transferability on various datasets, especially on the token-level tasks, such as question answering, and structured prediction. Moreover, the model can serve as a pretrained word aligner, which achieves reasonably low error rates on the alignment benchmarks. The code and pretrained parameters are available at https://github.com/CZWin32768/XLM-Align.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Natural Language Inference | XNLI (test) | Average Accuracy82.3 | 167 | |
| Cross-lingual Language Understanding | XTREME | XNLI Accuracy76.2 | 38 | |
| Question Answering | MLQA (test) | F1 Score73.4 | 35 | |
| Cross-lingual sentence retrieval | Tatoeba Parallel 14 language pairs | -- | 14 | |
| Word Alignment | EuroParl en-de, en-fr, en-hi, en-ro WPT2003, WPT2005 | AER (en-de)16.63 | 12 | |
| Cross-lingual sentence retrieval (en → xx) | Tatoeba-36 | Accuracy@155.5 | 11 | |
| Cross-lingual sentence retrieval (xx → en) | Tatoeba-36 | Average Accuracy@153.4 | 11 | |
| Cross-lingual Transfer | XTREME (test) | MLQA20.3 | 6 |