Deep Reinforcement Learning for Mention-Ranking Coreference Models
About
Coreference resolution systems are typically trained with heuristic loss functions that require careful tuning. In this paper we instead apply reinforcement learning to directly optimize a neural mention-ranking model for coreference evaluation metrics. We experiment with two approaches: the REINFORCE policy gradient algorithm and a reward-rescaled max-margin objective. We find the latter to be more effective, resulting in significant improvements over the current state-of-the-art on the English and Chinese portions of the CoNLL 2012 Shared Task.
Kevin Clark, Christopher D. Manning• 2016
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Coreference Resolution | CoNLL English 2012 (test) | MUC F1 Score74.6 | 114 | |
| Pronoun Disambiguation Problem | PDP 2016 (test) | Accuracy41.7 | 21 | |
| Commonsense Reasoning | Winograd Schema Challenge (WSC) (test) | Accuracy50.5 | 17 | |
| Coreference Resolution | CoNLL Chinese 2012 (test) | Average F1 Score63.88 | 11 |
Showing 4 of 4 rows