Bi-directional Attention with Agreement for Dependency Parsing
About
We develop a novel bi-directional attention model for dependency parsing, which learns to agree on headword predictions from the forward and backward parsing directions. The parsing procedure for each direction is formulated as sequentially querying the memory component that stores continuous headword embeddings. The proposed parser makes use of {\it soft} headword embeddings, allowing the model to implicitly capture high-order parsing history without dramatically increasing the computational complexity. We conduct experiments on English, Chinese, and 12 other languages from the CoNLL 2006 shared task, showing that the proposed model achieves state-of-the-art unlabeled attachment scores on 6 languages.
Hao Cheng, Hao Fang, Xiaodong He, Jianfeng Gao, Li Deng• 2016
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Dependency Parsing | Chinese Treebank (CTB) (test) | UAS88.1 | 99 | |
| Dependency Parsing | Penn Treebank (PTB) (test) | LAS91.49 | 80 | |
| Dependency Parsing | English PTB Stanford Dependencies (test) | UAS94.1 | 76 | |
| Dependency Parsing | CoNLL German 2009 (test) | UAS92.71 | 25 | |
| Dependency Parsing | CoNLL Spanish 2009 (test) | UAS88.74 | 14 | |
| Dependency Parsing | CoNLL Czech 2009 (test) | UAS91.16 | 12 | |
| Dependency Parsing | WSJ section 23 (test) | UAS94.1 | 10 | |
| Dependency Parsing | CoNLL Japanese 2009 (test) | UAS93.44 | 9 | |
| Parsing | English PTB-SD 3.3.0 (test) | UAS94.1 | 7 | |
| Dependency Parsing | CoNLL Turkish (tr) treebank (test) | UAS78.43 | 5 |
Showing 10 of 18 rows