Optimizing Decoding Paths in Masked Diffusion Models by Quantifying Uncertainty
About
Masked Diffusion Models (MDMs) offer flexible, non-autoregressive generation, but this freedom introduces a challenge: final output quality is highly sensitive to the decoding order. We are the first to formalize this issue, attributing the variability in output quality to the cumulative predictive uncertainty along a generative path. To quantify this uncertainty, we introduce Denoising Entropy, a computable metric that serves as an internal signal for evaluating generative process. Leveraging this metric, we propose two algorithms designed to optimize the decoding path: a post-hoc selection method and a real-time guidance strategy. Experiments demonstrate that our entropy-guided methods significantly improve generation quality, consistently boosting accuracy on challenging reasoning, planning, and code benchmarks. Our work establishes Denoising Entropy as a principled tool for understanding and controlling generation, effectively turning the uncertainty in MDMs from a liability into a key advantage for discovering high-quality solutions.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Code Generation | HumanEval | -- | 850 | |
| Code Generation | HumanEval+ | -- | 189 | |
| Code Generation | MBPP | Accuracy (%)3 | 146 | |
| Code Generation | MBPP+ | Accuracy4 | 75 | |
| Planning | Countdown | Accuracy42.2 | 68 | |
| Planning | Sudoku | Accuracy34.2 | 68 | |
| Scientific Reasoning | GPQA | Accuracy29 | 55 |