Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Latent Normalizing Flows for Discrete Sequences

About

Normalizing flows are a powerful class of generative models for continuous random variables, showing both strong model flexibility and the potential for non-autoregressive generation. These benefits are also desired when modeling discrete random variables such as text, but directly applying normalizing flows to discrete sequences poses significant additional challenges. We propose a VAE-based generative model which jointly learns a normalizing flow-based distribution in the latent space and a stochastic mapping to an observed discrete space. In this setting, we find that it is crucial for the flow-based distribution to be highly multimodal. To capture this property, we propose several normalizing flow architectures to maximize model flexibility. Experiments consider common discrete sequence tasks of character-level language modeling and polyphonic music generation. Our results indicate that an autoregressive flow-based model can match the performance of a comparable autoregressive baseline, and a non-autoregressive flow-based model can improve generation speed with a penalty to performance.

Zachary M. Ziegler, Alexander M. Rush• 2019

Related benchmarks

TaskDatasetResultRank
Character-level Language Modelingenwik8 (test)--
195
Character-level Language Modelingtext8 (test)
BPC1.62
128
Character-level Language ModelingPenn Treebank (test)
BPC1.46
113
Character-level Language ModelingPenn Treebank char-level (test)
BPC1.46
25
Language Modelingtext8 (test)
BPC1.88
21
Polyphonic music modelingNottingham (Nott)
NLL (nats)2.39
14
Polyphonic music modelingJSB Chorales
Negative Log-Likelihood (nats)6.53
14
Polyphonic music modelingPiano-midi.de
NLL (nats)7.77
12
Polyphonic music modelingMuseData (Muse)
Negative Log-Likelihood (nats)6.92
12
Showing 9 of 9 rows

Other info

Follow for update