Adapting BERT for Word Sense Disambiguation with Gloss Selection Objective and Example Sentences
About
Domain adaptation or transfer learning using pre-trained language models such as BERT has proven to be an effective approach for many natural language processing tasks. In this work, we propose to formulate word sense disambiguation as a relevance ranking task, and fine-tune BERT on sequence-pair ranking task to select the most probable sense definition given a context sentence and a list of candidate sense definitions. We also introduce a data augmentation technique for WSD using existing example sentences from WordNet. Using the proposed training objective and data augmentation technique, our models are able to achieve state-of-the-art results on the English all-words benchmark datasets.
Boon Peng Yap, Andrew Koh, Eng Siong Chng• 2020
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Word Sense Disambiguation | SensEval-3 (test) | F1 Score77.8 | 51 | |
| Word Sense Disambiguation | SemEval-15 (SE15) 3.0 (test) | F1 Score84.4 | 16 | |
| Word Sense Disambiguation | All-Words WSD Concatenation SE2+SE3+SE13+SE15 3.0 (test) | Overall F179.5 | 16 | |
| Word Sense Disambiguation | Senseval-2 (SE2) 3.0 (test) | F1 Score79.8 | 16 | |
| Word Sense Disambiguation | SemEval-13 (SE13) 3.0 (test) | F1 Score79.7 | 16 | |
| Word Sense Disambiguation | SemEval-07 3.0 (dev) | F1 Score72.7 | 14 |
Showing 6 of 6 rows