SIRE: Separate Intra- and Inter-sentential Reasoning for Document-level Relation Extraction
About
Document-level relation extraction has attracted much attention in recent years. It is usually formulated as a classification problem that predicts relations for all entity pairs in the document. However, previous works indiscriminately represent intra- and inter-sentential relations in the same way, confounding the different patterns for predicting them. Besides, they create a document graph and use paths between entities on the graph as clues for logical reasoning. However, not all entity pairs can be connected with a path and have the correct logical reasoning paths in their graph. Thus many cases of logical reasoning cannot be covered. This paper proposes an effective architecture, SIRE, to represent intra- and inter-sentential relations in different ways. We design a new and straightforward form of logical reasoning module that can cover more logical reasoning chains. Experiments on the public datasets show SIRE outperforms the previous state-of-the-art methods. Further analysis shows that our predictions are reliable and explainable. Our code is available at https://github.com/DreamInvoker/SIRE.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Document-level Relation Extraction | DocRED (dev) | F1 Score61.6 | 231 | |
| Document-level Relation Extraction | DocRED (test) | F1 Score62.05 | 179 | |
| Relation Extraction | DocRED (test) | F1 Score62.05 | 121 | |
| Relation Extraction | DocRED (dev) | F1 Score61.6 | 98 | |
| Relation Extraction | CDR (test) | F1 Score70.8 | 92 | |
| Relation Extraction | DocRED v1 (test) | F162.05 | 66 | |
| Relation Extraction | GDA (test) | F1 Score84.7 | 65 | |
| Relation Extraction | DocRED v1 (dev) | F1 Score61.6 | 65 | |
| Document-level Relation Extraction | DocRED 1.0 (test) | F162.05 | 51 | |
| Document-level Relation Extraction | DocRED 1.0 (dev) | F10.616 | 42 |