SenseBERT: Driving Some Sense into BERT
About
The ability to learn from large unlabeled corpora has allowed neural language models to advance the frontier in natural language understanding. However, existing self-supervision techniques operate at the word form level, which serves as a surrogate for the underlying semantic content. This paper proposes a method to employ weak-supervision directly at the word sense level. Our model, named SenseBERT, is pre-trained to predict not only the masked words but also their WordNet supersenses. Accordingly, we attain a lexical-semantic level language model, without the use of human annotation. SenseBERT achieves significantly improved lexical understanding, as we demonstrate by experimenting on SemEval Word Sense Disambiguation, and by attaining a state of the art result on the Word in Context task.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Word Sense Disambiguation | WiC (dev) | Accuracy72.1 | 32 | |
| Word Sense Disambiguation | WiC (test) | Accuracy72.1 | 26 | |
| Word Sense Disambiguation | WiC v1.0 (test) | Accuracy72.1 | 19 | |
| Word Sense Disambiguation | SemEval-SS standardized (test) | Accuracy83.7 | 8 |