Span-based Joint Entity and Relation Extraction with Transformer Pre-training
About
We introduce SpERT, an attention model for span-based joint entity and relation extraction. Our key contribution is a light-weight reasoning on BERT embeddings, which features entity recognition and filtering, as well as relation classification with a localized, marker-free context representation. The model is trained using strong within-sentence negative samples, which are efficiently extracted in a single BERT pass. These aspects facilitate a search over all spans in the sentence. In ablation studies, we demonstrate the benefits of pre-training, strong negative sampling and localized context. Our model outperforms prior work by up to 2.6% F1 score on several datasets for joint entity and relation extraction.
Markus Eberts, Adrian Ulges• 2019
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Named Entity Recognition | BC5CDR | F1 Score86.65 | 59 | |
| Joint Entity and Relation Extraction | CONLL04 | Entity F188.9 | 33 | |
| Named Entity Recognition | NCBI-disease | F1 Score86.12 | 29 | |
| Relation Extraction | CoNLL04 (test) | F1 Score71.5 | 28 | |
| Joint Entity and Relation Extraction | ADE | Entity F1 Score0.893 | 26 | |
| Relation Extraction | SCIERC (test) | F1 Score50.84 | 23 | |
| Entity Classification | CoNLL04 (test) | F1 Score88.94 | 21 | |
| Entity recognition | SCIERC (test) | F1 Score70.33 | 20 | |
| Relation Extraction | ADE | Relation Strict F179.2 | 20 | |
| Named Entity Recognition | ADE (test) | F1 Score89.3 | 19 |
Showing 10 of 23 rows