Cross-lingual Text Classification with Heterogeneous Graph Neural Network
About
Cross-lingual text classification aims at training a classifier on the source language and transferring the knowledge to target languages, which is very useful for low-resource languages. Recent multilingual pretrained language models (mPLM) achieve impressive results in cross-lingual classification tasks, but rarely consider factors beyond semantic similarity, causing performance degradation between some language pairs. In this paper we propose a simple yet effective method to incorporate heterogeneous information within and across languages for cross-lingual text classification using graph convolutional networks (GCN). In particular, we construct a heterogeneous graph by treating documents and words as nodes, and linking nodes with different relations, which include part-of-speech roles, semantic similarity, and document translations. Extensive experiments show that our graph-based method significantly outperforms state-of-the-art models on all tasks, and also achieves consistent performance gain over baselines in low-resource settings where external tools like translators are unavailable.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Sentiment Classification | amazon fr (test) | -- | 8 | |
| Intent Classification | Multilingual SLU EN-ES (test) | Accuracy96.81 | 6 | |
| Intent Classification | Multilingual SLU EN-TH (test) | Accuracy89.71 | 6 | |
| Sentiment Classification | Amazon Review EN → DE (test) | Accuracy (Books)0.927 | 6 | |
| Sentiment Classification | Amazon Review EN → JA (test) | Books Accuracy87.21 | 6 | |
| News Classification | XGLUE News Classification (test) | Accuracy (DE)85 | 5 |