Few-NERD: A Few-Shot Named Entity Recognition Dataset
About
Recently, considerable literature has grown up around the theme of few-shot named entity recognition (NER), but little published benchmark data specifically focused on the practical and challenging task. Current approaches collect existing supervised NER datasets and re-organize them to the few-shot setting for empirical study. These strategies conventionally aim to recognize coarse-grained entity types with few examples, while in practice, most unseen entity types are fine-grained. In this paper, we present Few-NERD, a large-scale human-annotated few-shot NER dataset with a hierarchy of 8 coarse-grained and 66 fine-grained entity types. Few-NERD consists of 188,238 sentences from Wikipedia, 4,601,160 words are included and each is annotated as context or a part of a two-level entity type. To the best of our knowledge, this is the first few-shot NER dataset and the largest human-crafted NER dataset. We construct benchmark tasks with different emphases to comprehensively assess the generalization capability of models. Extensive empirical results and analysis show that Few-NERD is challenging and the problem requires further research. We make Few-NERD public at https://ningding97.github.io/fewnerd/.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Named Entity Recognition | Few-NERD INTER 1.0 (test) | Average F150.53 | 62 | |
| Named Entity Recognition | FewNERD INTRA | -- | 47 | |
| Few-shot Named Entity Recognition | FewNERD Intra 1.0 | F1 Score41.93 | 44 | |
| Few-shot Named Entity Recognition | FEW-NERD INTER | F1 Score58.8 | 24 | |
| Named Entity Recognition | News | F1 Score50.06 | 21 | |
| Few-shot Named Entity Recognition | FewNERD Inter 1.0 | F1 Score54.29 | 20 | |
| Named Entity Recognition | Social | F1 Score17.26 | 12 | |
| Named Entity Recognition | Wiki | F1 Score9.54 | 12 | |
| Named Entity Recognition | Mixed | F1 Score13.59 | 12 | |
| Slot Filling | SNIPS 5-shot | We Score67.82 | 8 |