Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Constituency Parsing with a Self-Attentive Encoder

About

We demonstrate that replacing an LSTM encoder with a self-attentive architecture can lead to improvements to a state-of-the-art discriminative constituency parser. The use of attention makes explicit the manner in which information is propagated between different locations in the sentence, which we use to both analyze our model and propose potential improvements. For example, we find that separating positional and content information in the encoder can lead to improved parsing accuracy. Additionally, we evaluate different approaches for lexical representation. Our parser achieves new state-of-the-art results for single models trained on the Penn Treebank: 93.55 F1 without the use of any external data, and 95.13 F1 when using pre-trained word representations. Our parser also outperforms the previous best-published accuracy figures on 8 of the 9 languages in the SPMRL dataset.

Nikita Kitaev, Dan Klein• 2018

Related benchmarks

TaskDatasetResultRank
Constituent ParsingPTB (test)
F195.77
127
Constituency ParsingPenn Treebank WSJ (section 23 test)
F1 Score95.8
55
Constituency ParsingWSJ Penn Treebank (test)
F1 Score95.13
27
Constituency ParsingCTB 5.1 (test)
F1 Score91.98
25
Constituency ParsingCTB 5.0 (test)
F1 Score91.68
19
ParsingPTB (test)
Sents/sec332
17
Constituency ParsingPTB Section 23 (test)
Rerank F10.9513
13
Multilingual Constituency ParsingSPMRL 2013 2014 (test)
French Score87.42
13
Constituency ParsingChinese Treebank 5.1 (test)
F1 Score87.43
13
Constituency ParsingPTB (test)
Speed (Sents/s)830
12
Showing 10 of 21 rows

Other info

Code

Follow for update