Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

KLASS: KL-Guided Fast Inference in Masked Diffusion Models

About

Masked diffusion models have demonstrated competitive results on various tasks including language generation. However, due to its iterative refinement process, the inference is often bottlenecked by slow and static sampling speed. To overcome this problem, we introduce `KL-Adaptive Stability Sampling' (KLASS), a fast yet effective sampling method that exploits token-level KL divergence to identify stable, high-confidence predictions. By unmasking multiple tokens in each iteration without any additional model training, our approach speeds up generation significantly while maintaining sample quality. On reasoning benchmarks, KLASS achieves up to $2.78\times$ wall-clock speedups while improving performance over standard greedy decoding, attaining state-of-the-art results among diffusion-based samplers. We further validate KLASS across diverse domains, including text, image, and molecular generation, showing its effectiveness as a broadly applicable sampler across different models.

Seo Hyun Kim, Sunwoo Hong, Hojung Jung, Youngrok Park, Se-Young Yun• 2025

Related benchmarks

TaskDatasetResultRank
Commonsense ReasoningHellaSwag--
1460
Mathematical ReasoningGSM8K
Accuracy78.2
983
Code GenerationHumanEval
Pass@159.8
850
Mathematical ReasoningGSM8K (test)
Accuracy55.4
797
Mathematical ReasoningGSM8K
Speed Up (x)2.58
177
Code GenerationMBPP
Accuracy40.6
120
Code GenerationHumanEval
Accuracy (%)35.7
77
Mathematical ReasoningMATH 500
Accuracy38.3
73
PlanningSudoku
Accuracy82.1
68
PlanningCountdown
Accuracy35.4
68
Showing 10 of 19 rows

Other info

GitHub

Follow for update