Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Adapting BERT for Word Sense Disambiguation with Gloss Selection Objective and Example Sentences

About

Domain adaptation or transfer learning using pre-trained language models such as BERT has proven to be an effective approach for many natural language processing tasks. In this work, we propose to formulate word sense disambiguation as a relevance ranking task, and fine-tune BERT on sequence-pair ranking task to select the most probable sense definition given a context sentence and a list of candidate sense definitions. We also introduce a data augmentation technique for WSD using existing example sentences from WordNet. Using the proposed training objective and data augmentation technique, our models are able to achieve state-of-the-art results on the English all-words benchmark datasets.

Boon Peng Yap, Andrew Koh, Eng Siong Chng• 2020

Related benchmarks

TaskDatasetResultRank
Word Sense DisambiguationSensEval-3 (test)
F1 Score77.8
51
Word Sense DisambiguationSemEval-15 (SE15) 3.0 (test)
F1 Score84.4
16
Word Sense DisambiguationAll-Words WSD Concatenation SE2+SE3+SE13+SE15 3.0 (test)
Overall F179.5
16
Word Sense DisambiguationSenseval-2 (SE2) 3.0 (test)
F1 Score79.8
16
Word Sense DisambiguationSemEval-13 (SE13) 3.0 (test)
F1 Score79.7
16
Word Sense DisambiguationSemEval-07 3.0 (dev)
F1 Score72.7
14
Showing 6 of 6 rows

Other info

Follow for update