Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

VDLM: Variable Diffusion LMs via Robust Latent-to-Text Rendering

About

Autoregressive language models decode left-to-right with irreversible commitments, limiting revision during multi-step reasoning. We propose \textbf{VDLM}, a modular variable diffusion language model that separates semantic planning from text rendering. VDLM applies LLaDA-style masked diffusion over semantic variable embeddings to enable iterative refinement in latent space, then post-trains the planner with trajectory-aware optimization using embedding-space rewards and values, avoiding text decoding inside the RL loop. To convert planned embeddings back to text, we use a \textbf{Vec2Text} renderer and introduce \textbf{embedding perturbations} to robustify decoding under planner noise. Across nine benchmarks spanning general reasoning, math, and code, VDLM is competitive in pre-training and yields substantial post-training improvements on long-form generation tasks, outperforming other baselines. These results highlight the effectiveness of embedding-space post-training and robust latent-to-text rendering for diffusion language modeling.

Shuhui Qu• 2026

Related benchmarks

TaskDatasetResultRank
Code GenerationHumanEval--
850
Mathematical ReasoningGSM8K (test)
Accuracy89.8
797
Language UnderstandingMMLU
Accuracy71.4
756
Code GenerationHumanEval (test)--
444
Mathematical ReasoningMATH (test)
Overall Accuracy62.4
433
Physical Commonsense ReasoningPIQA
Accuracy74.2
329
Science ReasoningGPQA
Accuracy25.6
218
Science Question AnsweringARC-C
Accuracy54.4
127
Truthfulness EvaluationTruthfulQA
Accuracy50.2
93
Logical reasoningBBH
Accuracy54.5
93
Showing 10 of 14 rows

Other info

Follow for update