Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Beyond Masked and Unmasked: Discrete Diffusion Models via Partial Masking

About

Masked diffusion models (MDM) are powerful generative models for discrete data that generate samples by progressively unmasking tokens in a sequence. Each token can take one of two states: masked or unmasked. We observe that token sequences often remain unchanged between consecutive sampling steps; consequently, the model repeatedly processes identical inputs, leading to redundant computation. To address this inefficiency, we propose the Partial masking scheme (Prime), which augments MDM by allowing tokens to take intermediate states interpolated between the masked and unmasked states. This design enables the model to make predictions based on partially observed token information, and facilitates a fine-grained denoising process. We derive a variational training objective and introduce a simple architectural design to accommodate intermediate-state inputs. Our method demonstrates superior performance across a diverse set of generative modeling tasks. On text data, it achieves a perplexity of 15.36 on OpenWebText, outperforming previous MDM (21.52), autoregressive models (17.54), and their hybrid variants (17.58), without relying on an autoregressive formulation. On image data, it attains competitive FID scores of 3.26 on CIFAR-10 and 6.98 on ImageNet-32, comparable to leading continuous generative models.

Chen-Hao Chao, Wei-Fang Sun, Hanwen Liang, Chun-Yi Lee, Rahul G. Krishnan• 2025

Related benchmarks

TaskDatasetResultRank
Language ModelingPTB (val)
Perplexity44.49
99
Image GenerationCIFAR-10 (train/test)
FID3.26
78
Language ModelingLM1B (val)
Perplexity38
55
Language ModelingWikiText (val)
Perplexity19.46
54
Language ModelingOpenWebText (OWT) (val)
Perplexity13.41
42
Language ModelingAG News (val)
Perplexity48.09
28
Language ModelingLAMBADA (val)
Perplexity24.44
24
Language ModelingArXiv (val)
Perplexity25.19
24
Image GenerationImageNet-32
FID6.98
20
Text GenerationOpenWebText (OWT) GPT-2 tokenizer (val)
PPL15.36
12
Showing 10 of 17 rows

Other info

Follow for update