An Attentive Neural Architecture for Fine-grained Entity Type Classification
About
In this work we propose a novel attention-based neural network model for the task of fine-grained entity type classification that unlike previously proposed models recursively composes representations of entity mention contexts. Our model achieves state-of-the-art performance with 74.94% loose micro F1-score on the well-established FIGER dataset, a relative improvement of 2.59%. We also investigate the behavior of the attention mechanism of our model and observe that it can learn contextual linguistic expressions that indicate the fine-grained category memberships of an entity.
Sonse Shimaoka, Pontus Stenetorp, Kentaro Inui, Sebastian Riedel• 2016
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Ultra-fine Entity Typing | UFET (test) | Precision54.2 | 66 | |
| Fine-Grained Entity Typing | OntoNotes (test) | Macro F1 Score71 | 27 | |
| Fine-Grained Entity Typing | FIGER (test) | Macro F175.15 | 22 | |
| Entity Typing | Open Entity (test) | Precision68.8 | 12 | |
| Fine-Grained Entity Typing | OpenEntity (test) | Precision68.8 | 11 | |
| Ultra-fine Entity Typing | UFET (dev) | Precision0.537 | 10 | |
| Fine-Grained Entity Typing | Ultra-Fine (dev) | MRR22.1 | 7 | |
| Fine-Grained Entity Typing | Ultra-Fine (test) | MRR22.3 | 7 | |
| Fine-Grained Entity Typing | OntoNotes | Accuracy51.7 | 6 | |
| Entity Typing | FIGER (test) | Accuracy55.6 | 4 |
Showing 10 of 10 rows