Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Deep Span Representations for Named Entity Recognition

About

Span-based models are one of the most straightforward methods for named entity recognition (NER). Existing span-based NER systems shallowly aggregate the token representations to span representations. However, this typically results in significant ineffectiveness for long-span entities, a coupling between the representations of overlapping spans, and ultimately a performance degradation. In this study, we propose DSpERT (Deep Span Encoder Representations from Transformers), which comprises a standard Transformer and a span Transformer. The latter uses low-layered span representations as queries, and aggregates the token representations as keys and values, layer by layer from bottom to top. Thus, DSpERT produces span representations of deep semantics. With weight initialization from pretrained language models, DSpERT achieves performance higher than or competitive with recent state-of-the-art systems on eight NER benchmarks. Experimental results verify the importance of the depth for span representations, and show that DSpERT performs particularly well on long-span entities and nested structures. Further, the deep span representations are well structured and easily separable in the feature space.

Enwei Zhu, Yiyang Liu, Jinpeng Li• 2022

Related benchmarks

TaskDatasetResultRank
Named Entity RecognitionCoNLL 2003 (test)
F1 Score93.7
539
Nested Named Entity RecognitionACE 2004 (test)
F1 Score88.31
166
Nested Named Entity RecognitionACE 2005 (test)
F1 Score87.42
153
Nested Named Entity RecognitionGENIA (test)
F1 Score81.9
140
Named Entity RecognitionRESUME
F1 Score96.72
52
Nested Named Entity RecognitionKBP English 2017 (test)
Precision87.37
28
Named Entity RecognitionWeiboNER
F1 Score72.64
27
Flat Named Entity RecognitionOntoNotes 5.0 (test)
Micro F191.76
17
Showing 8 of 8 rows

Other info

Code

Follow for update