Label Semantics for Few Shot Named Entity Recognition
About
We study the problem of few shot learning for named entity recognition. Specifically, we leverage the semantic information in the names of the labels as a way of giving the model additional signal and enriched priors. We propose a neural architecture that consists of two BERT encoders, one to encode the document and its tokens and another one to encode each of the labels in natural language format. Our model learns to match the representations of named entities computed by the first encoder with label representations computed by the second encoder. The label semantics signal is shown to support improved state-of-the-art results in multiple few shot NER benchmarks and on-par performance in standard benchmarks. Our model is especially effective in low resource settings.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Named Entity Recognition | CoNLL 03 | -- | 102 | |
| Named Entity Recognition | Wnut 2017 | -- | 79 | |
| Named Entity Recognition | Few-NERD INTER 1.0 (test) | Average F151.24 | 62 | |
| Named Entity Recognition | FewNERD INTRA | F1 Score46.85 | 47 | |
| Named Entity Recognition | GUM | Micro F124.55 | 36 | |
| Named Entity Recognition | OntoNotes to I2B2, CoNLL, WNUT, GUM 5.0 (test) | I2B2 Score25.4 | 26 | |
| Named Entity Recognition | i2b2 2014 | Micro F1 Score0.2327 | 26 | |
| Named Entity Recognition | OntoNotes Onto-C 5.0 | Micro F126.37 | 26 | |
| Named Entity Recognition | OntoNotes Onto-A 5.0 | Micro F17.61 | 26 | |
| Named Entity Recognition | OntoNotes Onto-B 5.0 | Micro-F116.41 | 26 |