Neural Architectures for Named Entity Recognition
About
State-of-the-art named entity recognition systems rely heavily on hand-crafted features and domain-specific knowledge in order to learn effectively from the small, supervised training corpora that are available. In this paper, we introduce two new neural architectures---one based on bidirectional LSTMs and conditional random fields, and the other that constructs and labels segments using a transition-based approach inspired by shift-reduce parsers. Our models rely on two sources of information about words: character-based word representations learned from the supervised corpus and unsupervised word representations learned from unannotated corpora. Our models obtain state-of-the-art performance in NER in four languages without resorting to any language-specific knowledge or resources such as gazetteers.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Named Entity Recognition | CoNLL 2003 (test) | F1 Score91.14 | 539 | |
| Nested Named Entity Recognition | ACE 2004 (test) | F1 Score58.3 | 166 | |
| Nested Named Entity Recognition | ACE 2005 (test) | F1 Score57.6 | 153 | |
| Named Entity Recognition | CoNLL English 2003 (test) | F1 Score91.13 | 135 | |
| Named Entity Recognition | CoNLL 03 | F1 (Entity)90.94 | 102 | |
| Named Entity Recognition | CoNLL Spanish NER 2002 (test) | F1 Score86.12 | 98 | |
| Chunking | CoNLL 2000 (test) | F1 Score94.97 | 88 | |
| Named Entity Recognition | CoNLL Dutch 2002 (test) | F1 Score87.13 | 87 | |
| Named Entity Recognition | Conll 2003 | F1 Score90.94 | 86 | |
| Named Entity Recognition | CoNLL German 2003 (test) | F1 Score78.8 | 78 |