Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Stable-DiffCoder: Pushing the Frontier of Code Diffusion Large Language Model

About

Diffusion-based language models (DLLMs) offer non-sequential, block-wise generation and richer data reuse compared to autoregressive (AR) models, but existing code DLLMs still lag behind strong AR baselines under comparable budgets. We revisit this setting in a controlled study and introduce Stable-DiffCoder, a block diffusion code model that reuses the Seed-Coder architecture, data, and training pipeline. To enable efficient knowledge learning and stable training, we incorporate a block diffusion continual pretraining (CPT) stage enhanced by a tailored warmup and block-wise clipped noise schedule. Under the same data and architecture, Stable-DiffCoder overall outperforms its AR counterpart on a broad suite of code benchmarks. Moreover, relying only on the CPT and supervised fine-tuning stages, Stable-DiffCoder achieves stronger performance than a wide range of \~8B ARs and DLLMs, demonstrating that diffusion-based training can improve code modeling quality beyond AR training alone. Moreover, diffusion-based any-order modeling improves structured code modeling for editing and reasoning, and through data augmentation, benefits low-resource coding languages.

Chenghao Fan, Wen Heng, Bo Li, Sichen Liu, Yuxuan Song, Jing Su, Xiaoye Qu, Kai Shen, Wei Wei• 2026

Related benchmarks

TaskDatasetResultRank
Code GenerationHumanEval (test)--
444
Code GenerationMBPP (test)--
276
Code GenerationMBPP
Pass@142.4
175
Function-level Code GenerationHumanEval+ augmented (test)
Pass@182.3
46
Function-level Code GenerationMBPP+ augmented (test)
Pass@172.8
45
Code GenerationBigCodeBench-Completion Full
pass@154.8
41
Code GenerationBigCodeBench-Completion Hard
pass@131.8
38
CUDA Kernel GenerationKernelBench Level 1
Exec Count27
31
CUDA Kernel GenerationKernelBench Level 3
Executions Count10
31
CUDA Kernel GenerationKernelBench Level 2
Execution Count5
31
Showing 10 of 17 rows

Other info

GitHub

Follow for update