Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

What Do Recurrent Neural Network Grammars Learn About Syntax?

About

Recurrent neural network grammars (RNNG) are a recently proposed probabilistic generative modeling family for natural language. They show state-of-the-art language modeling and parsing performance. We investigate what information they learn, from a linguistic perspective, through various ablations to the model and the data, and by augmenting the model with an attention mechanism (GA-RNNG) to enable closer inspection. We find that explicit modeling of composition is crucial for achieving the best performance. Through the attention mechanism, we find that headedness plays a central role in phrasal representation (with the model's latent attention largely agreeing with predictions made by hand-crafted head rules, albeit with some important differences). By training grammars without nonterminal labels, we find that phrasal representations depend minimally on nonterminals, providing support for the endocentricity hypothesis.

Adhiguna Kuncoro, Miguel Ballesteros, Lingpeng Kong, Chris Dyer, Graham Neubig, Noah A. Smith• 2016

Related benchmarks

TaskDatasetResultRank
Language ModelingPTB (test)
Perplexity100.9
471
Constituent ParsingPTB (test)
F193.6
127
Phrase-structure parsingPTB (§23)--
56
Constituency ParsingPenn Treebank WSJ (section 23 test)
F1 Score93.6
55
Dependency ParsingWSJ section 23 (test)
UAS95.7
10
ParsingEnglish PTB-SD 3.3.0 (test)
UAS95.8
7
Showing 6 of 6 rows

Other info

Code

Follow for update