Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Adversarial Learning with Contextual Embeddings for Zero-resource Cross-lingual Classification and NER

About

Contextual word embeddings (e.g. GPT, BERT, ELMo, etc.) have demonstrated state-of-the-art performance on various NLP tasks. Recent work with the multilingual version of BERT has shown that the model performs very well in zero-shot and zero-resource cross-lingual settings, where only labeled English data is used to finetune the model. We improve upon multilingual BERT's zero-resource cross-lingual performance via adversarial learning. We report the magnitude of the improvement on the multilingual MLDoc text classification and CoNLL 2002/2003 named entity recognition tasks. Furthermore, we show that language-adversarial training encourages BERT to align the embeddings of English documents and their translations, which may be the cause of the observed performance gains.

Phillip Keung, Yichao Lu, Vikas Bhardwaj• 2019

Related benchmarks

TaskDatasetResultRank
Named Entity RecognitionCoNLL NER 2002/2003 (test)
German F1 Score73.89
59
Named Entity RecognitionCoNLL (test)
F1 Score (de)71.9
28
Showing 2 of 2 rows

Other info

Follow for update