Large Scale Substitution-based Word Sense Induction
About
We present a word-sense induction method based on pre-trained masked language models (MLMs), which can cheaply scale to large vocabularies and large corpora. The result is a corpus which is sense-tagged according to a corpus-derived sense inventory and where each sense is associated with indicative words. Evaluation on English Wikipedia that was sense-tagged using our method shows that both the induced senses, and the per-instance sense assignment, are of high quality even compared to WSD methods, such as Babelfy. Furthermore, by training a static word embeddings algorithm on the sense-tagged corpus, we obtain high-quality static senseful embeddings. These outperform existing senseful embeddings methods on the WiC dataset and on a new outlier detection dataset we developed. The data driven nature of the algorithm allows to induce corpora-specific senses, which may not appear in standard sense inventories, as we demonstrate using a case study on the scientific domain.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Word-in-Context Classification | WiC | Accuracy58.3 | 34 | |
| Word Sense Induction | SemEval 2013 WSI Task 13 (test) | F-BC61.98 | 19 | |
| Word Sense Induction | SemEval 2010 WSI Task 14 (test) | F-Score70.95 | 16 | |
| Outlier Detection | 25-7-1-8 (test) | OPP96.68 | 5 | |
| Sense Classification | CoarseWSD-20 Manually annotated 1.0 (Annotator #1) | F1 Score89.05 | 3 | |
| Sense Classification | CoarseWSD-20 Manually annotated 1.0 (Annotator #2) | F1 Score85.95 | 3 | |
| Sense Classification | Manually annotated CoarseWSD-20 1.0 (Average) | F1 Score87.5 | 3 |