End-to-End Neural Entity Linking
About
Entity Linking (EL) is an essential task for semantic text understanding and information extraction. Popular methods separately address the Mention Detection (MD) and Entity Disambiguation (ED) stages of EL, without leveraging their mutual dependency. We here propose the first neural end-to-end EL system that jointly discovers and links entities in a text document. The main idea is to consider all possible spans as potential mentions and learn contextual similarity scores over their entity candidates that are useful for both MD and ED decisions. Key components are context-aware mention embeddings, entity embeddings and a probabilistic mention - entity map, without demanding other engineered features. Empirically, we show that our end-to-end method significantly outperforms popular systems on the Gerbil platform when enough training data is available. Conversely, if testing datasets follow different annotation conventions compared to the training set (e.g. queries/ tweets vs news documents), our ED model coupled with a traditional NER system offers the best or second best EL accuracy.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Entity Linking | AIDA A | Micro F189.8 | 43 | |
| Entity Linking | MSNBC | Micro F186.2 | 36 | |
| Entity Linking | OKE 2016 | Macro F176.7 | 31 | |
| Entity Linking | AIDA B | Macro F10.832 | 30 | |
| Entity Linking | AIDA (testb) | Micro F182.4 | 28 | |
| Entity Linking | KORE50 | Macro F152.4 | 27 | |
| Entity Linking | OKE 2015 | Macro F173.2 | 26 | |
| Entity Linking | N3-Reuters-128 | Macro F163.4 | 25 | |
| Entity Linking | N3-RSS-500 | Macro F166.6 | 25 | |
| Entity Linking | Derczynski | Macro F165.3 | 25 |