Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Beyond Masks: Efficient, Flexible Diffusion Language Models via Deletion-Insertion Processes

About

While Masked Diffusion Language Models (MDLMs) relying on token masking and unmasking have shown promise in language modeling, their computational efficiency and generation flexibility remain constrained by the masking paradigm. In this paper, we propose Deletion-Insertion Diffusion language models (DID) that rigorously formulate token deletion and insertion as discrete diffusion processes, replacing the masking and unmasking processes in current MDLMs. DID improves training and inference efficiency by eliminating two major sources of computational overhead in MDLMs: the computations on non-informative 1) <MASK> tokens inherent to the paradigm, and 2) <PAD> tokens introduced in variable-length settings. Furthermore, DID offers greater flexibility by: 1) natively supporting variable-length sequences without requiring fixed-length padding, and 2) an intrinsic self-correction mechanism during generation due to insertion that dynamically adjusts token positions. To train DID, we design a score-based approach that assigns scores to token insertion operations and derive appropriate training objectives. The objectives involve subsequence counting problems, which we efficiently solve via a parallelized dynamic programming algorithm. Our experiments across fixed and variable-length settings demonstrate the advantage of DID over baselines of MDLMs and existing insertion-based LMs, in terms of modeling performance, sampling quality, and training/inference speed, without any hyperparameter tuning.

Fangyu Ding, Ding Ding, Sijin Chen, Kaibo Wang, Peng Xu, Zijin Feng, Haoli Bai, Kai Han, Youliang Yan, Binhang Yuan, Jiacheng Sun• 2026

Related benchmarks

TaskDatasetResultRank
Language ModelingPTB
Perplexity87.09
1034
Language ModelingWikiText
PPL28.35
732
Language ModelingLAMBADA--
268
Language ModelingOpenWebText
Perplexity85.35
91
Language ModelingarXiv
Perplexity61.77
55
Language ModelingPubmed
Perplexity38.71
38
Language ModelingAG-News
PPL48.84
36
Conditional GenerationOWT
Perplexity (PPL)19.99
24
Language ModelingLM1B
Perplexity58.05
22
Variable-length Language ModelingStories
PPL21.07
12
Showing 10 of 11 rows

Other info

Follow for update