Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Compound Probabilistic Context-Free Grammars for Grammar Induction

About

We study a formalization of the grammar induction problem that models sentences as being generated by a compound probabilistic context-free grammar. In contrast to traditional formulations which learn a single stochastic grammar, our grammar's rule probabilities are modulated by a per-sentence continuous latent variable, which induces marginal dependencies beyond the traditional context-free assumptions. Inference in this grammar is performed by collapsed variational inference, in which an amortized variational posterior is placed on the continuous variable, and the latent trees are marginalized out with dynamic programming. Experiments on English and Chinese show the effectiveness of our approach compared to recent state-of-the-art methods when evaluated on unsupervised parsing.

Yoon Kim, Chris Dyer, Alexander M. Rush• 2019

Related benchmarks

TaskDatasetResultRank
Language ModelingPTB (test)
Perplexity196.3
471
Unsupervised ParsingPTB (test)
F1 Score67.4
75
Unsupervised Constituency ParsingChinese Treebank (CTB) (test)
Unlabeled Sentence F1 (Mean)36
36
Unsupervised Constituency ParsingSUSANNE (test)
F1 Score48.6
32
Unsupervised Constituency ParsingWSJ (test)
Max F160.1
29
Grammar InductionPTB English (test)
F1 Score55.2
29
Unlabeled ParsingPenn Treebank WSJ (test)
F1 (mean)55.2
25
Unsupervised Constituency ParsingWSJ10 (test)
UF1 Score70.5
24
Unsupervised ParsingPenn Treebank WSJ (section 23 test)
F1 Score55.2
15
Syntactic EvaluationMarvin and Linzen
Syntactic Evaluation Score50.7
15
Showing 10 of 21 rows

Other info

Code

Follow for update