Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Simple Guidance Mechanisms for Discrete Diffusion Models

About

Diffusion models for continuous data gained widespread adoption owing to their high quality generation and control mechanisms. However, controllable diffusion on discrete data faces challenges given that continuous guidance methods do not directly apply to discrete diffusion. Here, we provide a straightforward derivation of classifier-free and classifier-based guidance for discrete diffusion, as well as a new class of diffusion models that leverage uniform noise and that are more guidable because they can continuously edit their outputs. We improve the quality of these models with a novel continuous-time variational lower bound that yields state-of-the-art performance, especially in settings involving guidance or fast generation. Empirically, we demonstrate that our guidance mechanisms combined with uniform noise diffusion improve controllable generation relative to autoregressive and diffusion baselines on several discrete data domains, including genomic sequences, small molecule design, and discretized image generation.

Yair Schiff, Subham Sekhar Sahoo, Hao Phung, Guanghan Wang, Sam Boshar, Hugo Dalla-torre, Bernardo P. de Almeida, Alexander Rush, Thomas Pierrot, Volodymyr Kuleshov• 2024

Related benchmarks

TaskDatasetResultRank
Language ModelingPTB
Perplexity95.986
1034
Language ModelingWikiText
PPL34.933
732
Language ModelingPTB (test)
Perplexity77.28
526
Image GenerationCIFAR-10 (test)
FID33.65
483
Language ModelingLAMBADA
Perplexity51.272
150
Language ModelingarXiv (test)
PPL41.19
145
Language modellingLM1B (test)
Perplexity81.41
130
Language ModelingOne Billion Word Benchmark (test)
Test Perplexity31.11
113
Image GenerationMNIST Binary (test)
FID4.69
98
Image GenerationImageNet-1k (val)
FID26.242
93
Showing 10 of 40 rows

Other info

Follow for update