Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Balancing Understanding and Generation in Discrete Diffusion Models

About

In discrete generative modeling, two dominant paradigms demonstrate divergent capabilities: Masked Diffusion Language Models (MDLM) excel at semantic understanding and zero-shot generalization, whereas Uniform-noise Diffusion Language Models (UDLM) achieve strong few-step generation quality, yet neither attains balanced performance across both dimensions. To address this, we propose XDLM, which bridges the two paradigms via a stationary noise kernel. XDLM offers two key contributions: (1) it provides a principled theoretical unification of MDLM and UDLM, recovering each paradigm as a special case; and (2) an alleviated memory bottleneck enabled by an algebraic simplification of the posterior probabilities. Experiments demonstrate that XDLM advances the Pareto frontier between understanding capability and generation quality. Quantitatively, XDLM surpasses UDLM by 5.4 points on zero-shot text benchmarks and outperforms MDLM in few-step image generation (FID 54.1 vs. 80.8). When scaled to tune an 8B-parameter large language model, XDLM achieves 15.0 MBPP in just 32 steps, effectively doubling the baseline performance. Finally, analysis of training dynamics reveals XDLM's superior potential for long-term scaling. Code is available at https://github.com/MzeroMiko/XDLM

Yue Liu, Yuzhong Zhao, Zheyong Xie, Qixiang Ye, Jianbin Jiao, Yao Hu, Shaosheng Cao, Yunfan Liu• 2026

Related benchmarks

TaskDatasetResultRank
Language ModelingPTB
Perplexity90.796
650
ReasoningBBH--
507
Language ModelingWikiText
PPL32.748
479
Mathematical ReasoningMATH--
162
Language ModelingLAMBADA
Perplexity45.608
99
Image GenerationImageNet-1k (val)
FID25.774
84
Code GenerationHumanEval
HumanEval Score31.71
50
Image GenerationImageNet-1K
FID8.625
42
Language ModelingarXiv
Perplexity37.232
21
Language ModelingAG-News
PPL62.768
20
Showing 10 of 12 rows

Other info

GitHub

Follow for update