Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

SenseBERT: Driving Some Sense into BERT

About

The ability to learn from large unlabeled corpora has allowed neural language models to advance the frontier in natural language understanding. However, existing self-supervision techniques operate at the word form level, which serves as a surrogate for the underlying semantic content. This paper proposes a method to employ weak-supervision directly at the word sense level. Our model, named SenseBERT, is pre-trained to predict not only the masked words but also their WordNet supersenses. Accordingly, we attain a lexical-semantic level language model, without the use of human annotation. SenseBERT achieves significantly improved lexical understanding, as we demonstrate by experimenting on SemEval Word Sense Disambiguation, and by attaining a state of the art result on the Word in Context task.

Yoav Levine, Barak Lenz, Or Dagan, Ori Ram, Dan Padnos, Or Sharir, Shai Shalev-Shwartz, Amnon Shashua, Yoav Shoham• 2019

Related benchmarks

TaskDatasetResultRank
Word Sense DisambiguationWiC (dev)
Accuracy72.1
32
Word Sense DisambiguationWiC (test)
Accuracy72.1
26
Word Sense DisambiguationWiC v1.0 (test)
Accuracy72.1
19
Word Sense DisambiguationSemEval-SS standardized (test)
Accuracy83.7
8
Showing 4 of 4 rows

Other info

Follow for update