MELM: Data Augmentation with Masked Entity Language Modeling for Low-Resource NER
About
Data augmentation is an effective solution to data scarcity in low-resource scenarios. However, when applied to token-level tasks such as NER, data augmentation methods often suffer from token-label misalignment, which leads to unsatsifactory performance. In this work, we propose Masked Entity Language Modeling (MELM) as a novel data augmentation framework for low-resource NER. To alleviate the token-label misalignment issue, we explicitly inject NER labels into sentence context, and thus the fine-tuned MELM is able to predict masked entity tokens by explicitly conditioning on their labels. Thereby, MELM generates high-quality augmented data with novel entities, which provides rich entity regularity knowledge and boosts NER performance. When training data from multiple languages are available, we also integrate MELM with code-mixing for further improvement. We demonstrate the effectiveness of MELM on monolingual, cross-lingual and multilingual NER across various low-resource levels. Experimental results show that our MELM presents substantial improvement over the baseline methods.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Named Entity Recognition | CoNLL 03 | F1 (Entity)81.9 | 102 | |
| Named Entity Recognition | OntoNotes | F1-score54.97 | 91 | |
| Complex Named Entity Recognition | MultiCoNER (test) | Score (Bn)30.27 | 76 | |
| Named Entity Recognition | CoNLL NER 2002/2003 (test) | German F1 Score80.33 | 59 | |
| Named Entity Recognition | MultiCoNER | F1 Score0.4901 | 48 | |
| Named Entity Recognition | NCBI | F1 Score75.11 | 26 | |
| Named Entity Recognition | bc2gm | Entity F156.83 | 21 | |
| Named Entity Recognition | TDMSci | F1 Score57.8 | 10 | |
| Named Entity Recognition | CoNLL | F1 Score0.8351 | 10 | |
| Named Entity Recognition | MultiCoNER entire dataset 1.0 (full) | Accuracy (En)66.27 | 5 |