Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Dependency Parsing as Head Selection

About

Conventional graph-based dependency parsers guarantee a tree structure both during training and inference. Instead, we formalize dependency parsing as the problem of independently selecting the head of each word in a sentence. Our model which we call \textsc{DeNSe} (as shorthand for {\bf De}pendency {\bf N}eural {\bf Se}lection) produces a distribution over possible heads for each word using features obtained from a bidirectional recurrent neural network. Without enforcing structural constraints during training, \textsc{DeNSe} generates (at inference time) trees for the overwhelming majority of sentences, while non-tree outputs can be adjusted with a maximum spanning tree algorithm. We evaluate \textsc{DeNSe} on four languages (English, Chinese, Czech, and German) with varying degrees of non-projectivity. Despite the simplicity of the approach, our parsers are on par with the state of the art.

Xingxing Zhang, Jianpeng Cheng, Mirella Lapata• 2016

Related benchmarks

TaskDatasetResultRank
Dependency ParsingEnglish PTB Stanford Dependencies (test)
UAS94.3
76
Dependency ParsingWSJ (test)
UAS94.1
67
Dependency ParsingPenn Treebank (PTB) Section 23 v2.2 (test)
UAS94.1
17
POS TaggingPenn Treebank (PTB) Section 23 v2.2 (test)
POS Accuracy97.3
15
Showing 4 of 4 rows

Other info

Follow for update