Second-Order Neural Dependency Parsing with Message Passing and End-to-End Training
About
In this paper, we propose second-order graph-based neural dependency parsing using message passing and end-to-end neural networks. We empirically show that our approaches match the accuracy of very recent state-of-the-art second-order graph-based neural dependency parsers and have significantly faster speed in both training and testing. We also empirically show the advantage of second-order parsing over first-order parsing and observe that the usefulness of the head-selection structured constraint vanishes when using BERT embedding.
Xinyu Wang, Kewei Tu• 2020
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Dependency Parsing | Chinese Treebank (CTB) (test) | UAS92.55 | 99 | |
| Dependency Parsing | Penn Treebank (PTB) (test) | LAS95.37 | 80 | |
| Dependency Parsing | UD 2.2 (test) | bg91.42 | 31 | |
| Dependency Parsing | PTB | UAS96.91 | 24 | |
| Dependency Parsing | CTB | UAS92.78 | 11 |
Showing 5 of 5 rows