Diffusion-LM Improves Controllable Text Generation
About
Controlling the behavior of language models (LMs) without re-training is a major open problem in natural language generation. While recent works have demonstrated successes on controlling simple sentence attributes (e.g., sentiment), there has been little progress on complex, fine-grained controls (e.g., syntactic structure). To address this challenge, we develop a new non-autoregressive language model based on continuous diffusions that we call Diffusion-LM. Building upon the recent successes of diffusion models in continuous domains, Diffusion-LM iteratively denoises a sequence of Gaussian vectors into word vectors, yielding a sequence of intermediate latent variables. The continuous, hierarchical nature of these intermediate variables enables a simple gradient-based algorithm to perform complex, controllable generation tasks. We demonstrate successful control of Diffusion-LM for six challenging fine-grained control tasks, significantly outperforming prior work.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Language modelling | LM1B (test) | Perplexity118.6 | 120 | |
| Language Modeling | One Billion Word Benchmark (test) | Test Perplexity118.6 | 108 | |
| Machine Translation | WMT 2014 (test) | BLEU17.41 | 100 | |
| Machine Translation | WMT En-De '14 | BLEU15.3 | 89 | |
| Text Generation | LM1B (test) | -- | 72 | |
| Machine Translation | WMT 2016 (test) | BLEU29.39 | 58 | |
| Machine Translation | IWSLT De-En 14 | BLEU Score29.11 | 33 | |
| Machine Translation | WMT De-En 14 | BLEU17.3 | 33 | |
| Language Modeling | LM1B (val) | Perplexity118.6 | 17 | |
| Structured JSON Generation | MultiWOZ, Super-NaturalInstructions, TruthfulQA, and Self-Instruct Averaged | Similarity Score0.72 | 16 |