Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Fine-tune Bert for DocRED with Two-step Process

About

Modelling relations between multiple entities has attracted increasing attention recently, and a new dataset called DocRED has been collected in order to accelerate the research on the document-level relation extraction. Current baselines for this task uses BiLSTM to encode the whole document and are trained from scratch. We argue that such simple baselines are not strong enough to model to complex interaction between entities. In this paper, we further apply a pre-trained language model (BERT) to provide a stronger baseline for this task. We also find that solving this task in phases can further improve the performance. The first step is to predict whether or not two entities have a relation, the second step is to predict the specific relation.

Hong Wang, Christfried Focke, Rob Sylvester, Nilesh Mishra, William Wang• 2019

Related benchmarks

TaskDatasetResultRank
Document-level Relation ExtractionDocRED (dev)
F1 Score58.83
231
Document-level Relation ExtractionDocRED (test)
F1 Score58.69
179
Relation ExtractionDocRED (test)
F1 Score56.5
121
Relation ExtractionDocRED (dev)
F1 Score55.4
98
Relation ExtractionDocRED v1 (test)
F158.69
66
Relation ExtractionDocRED v1 (dev)
F1 Score58.83
65
Document-level Relation ExtractionDocRED 1.0 (test)
F158.69
51
Document-level Relation ExtractionDocRED 1.0 (dev)
F154.42
42
Relation ExtractionNYT (test)
P@10082
9
Showing 9 of 9 rows

Other info

Code

Follow for update