Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

WeDLM: Reconciling Diffusion Language Models with Standard Causal Attention for Fast Inference

About

Autoregressive (AR) generation is the standard decoding paradigm for Large Language Models (LLMs), but its token-by-token nature limits parallelism at inference time. Diffusion Language Models (DLLMs) offer parallel decoding by recovering multiple masked tokens per step; however, in practice they often fail to translate this parallelism into deployment speed gains over optimized AR engines (e.g., vLLM). A key reason is that many DLLMs rely on bidirectional attention, which breaks standard prefix KV caching and forces repeated contextualization, undermining efficiency. We propose WeDLM, a diffusion decoding framework built entirely on standard causal attention to make parallel generation prefix-cache friendly. The core idea is to let each masked position condition on all currently observed tokens while keeping a strict causal mask, achieved by Topological Reordering that moves observed tokens to the physical prefix while preserving their logical positions. Building on this property, we introduce a streaming decoding procedure that continuously commits confident tokens into a growing left-to-right prefix and maintains a fixed parallel workload, avoiding the stop-and-wait behavior common in block diffusion methods. Experiments show that WeDLM preserves the quality of strong AR backbones while delivering substantial speedups, approaching 3x on challenging reasoning benchmarks and up to 10x in low-entropy generation regimes; critically, our comparisons are against AR baselines served by vLLM under matched deployment settings, demonstrating that diffusion-style decoding can outperform an optimized AR engine in practice.

Aiwei Liu, Minghua He, Shaoxun Zeng, Sijun Zhang, Linhao Zhang, Chuhan Wu, Wei Jia, Yuan Liu, Xiao Zhou, Jie Zhou• 2025

Related benchmarks

TaskDatasetResultRank
Code GenerationHumanEval (test)--
444
Code GenerationMBPP (test)--
276
Code GenerationHumanEval+ (test)
Pass@168.9
81
Function-level Code GenerationHumanEval+ augmented (test)
Pass@173.8
46
CUDA Kernel GenerationKernelBench Level 1
Exec Count14
31
CUDA Kernel GenerationKernelBench Level 2
Execution Count1
31
CUDA Kernel GenerationKernelBench Level 3
Executions Count0.00e+0
31
Zero-shot ReasoningARC-Easy zero-shot
Zero-shot Accuracy97.43
22
Code GenerationMBPP
MBPP67
19
STEM ReasoningGPQA-Diamond 5-shot
Accuracy44.95
10
Showing 10 of 25 rows

Other info

Follow for update