End-to-End Relation Extraction using LSTMs on Sequences and Tree Structures
About
We present a novel end-to-end neural model to extract entities and relations between them. Our recurrent neural network based model captures both word sequence and dependency tree substructure information by stacking bidirectional tree-structured LSTM-RNNs on bidirectional sequential LSTM-RNNs. This allows our model to jointly represent both entities and relations with shared parameters in a single model. We further encourage detection of entities during training and use of entity information in relation extraction via entity pretraining and scheduled sampling. Our model improves over the state-of-the-art feature-based model on end-to-end relation extraction, achieving 12.1% and 5.7% relative error reductions in F1-score on ACE2005 and ACE2004, respectively. We also show that our LSTM-RNN based model compares favorably to the state-of-the-art CNN based model (in F1-score) on nominal relation classification (SemEval-2010 Task 8). Finally, we present an extensive ablation analysis of several model components.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Relation Classification | SemEval-2010 Task 8 (test) | F1 Score84.4 | 128 | |
| Relation Extraction | ACE05 (test) | F1 Score65.3 | 72 | |
| Named Entity Recognition | ACE 2005 (test) | F1 Score83.4 | 58 | |
| Entity extraction | ACE05 (test) | F1 Score83.4 | 53 | |
| Named Entity Recognition | ACE04 (test) | F1 Score81.8 | 36 | |
| Relationship Extraction | SemEval Task 8 2010 (test) | F1 Score84.4 | 24 | |
| Relation Extraction | SCIERC (test) | F1 Score36.6 | 23 | |
| Relation Extraction | ACE04 (test) | F1 Score48.4 | 21 | |
| Relation Extraction | NYT10 subset (test) | Precision49.2 | 20 | |
| Entity recognition | SCIERC (test) | F1 Score63.7 | 20 |