Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Shift-Reduce Constituent Parsing with Neural Lookahead Features

About

Transition-based models can be fast and accurate for constituent parsing. Compared with chart-based models, they leverage richer features by extracting history information from a parser stack, which spans over non-local constituents. On the other hand, during incremental parsing, constituent information on the right hand side of the current word is not utilized, which is a relative weakness of shift-reduce parsing. To address this limitation, we leverage a fast neural model to extract lookahead features. In particular, we build a bidirectional LSTM model, which leverages the full sentence information to predict the hierarchy of constituents that each word starts and ends. The results are then passed to a strong transition-based constituent parser as lookahead features. The resulting parser gives 1.3% absolute improvement in WSJ and 2.3% in CTB compared to the baseline, given the highest reported accuracies for fully-supervised parsing.

Jiangming Liu, Yue Zhang• 2016

Related benchmarks

TaskDatasetResultRank
Constituent ParsingPTB (test)
F194.2
127
Phrase-structure parsingPTB (§23)
F1 Score91.7
56
Constituency ParsingPenn Treebank WSJ (section 23 test)
F1 Score91.7
55
Constituent ParsingCTB (test)
F1 Score85.5
45
ParsingPTB (test)
Sents/sec79.2
17
Constituency ParsingPTB WSJ (Section 23 test)
F1 Score91.8
12
Showing 6 of 6 rows

Other info

Follow for update