Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Unlocking Prompt Infilling Capability for Diffusion Language Models

About

Masked diffusion language models (dLMs) generate text through bidirectional denoising, yet this capability remains locked for infilling prompts. This limitation is an artifact of the current supervised finetuning (SFT) convention of applying response-only masking. To unlock this capability, we extend full-sequence masking during SFT, where both prompts and responses are masked jointly. Once unlocked, the model infills masked portions of a prompt template conditioned on few-shot examples. We show that such model-infilled prompts match or surpass manually designed templates, transfer effectively across models, and are complementary to existing prompt optimization methods. Our results suggest that training practices, not architectural limitations, are the primary bottleneck preventing masked diffusion language models from infilling effective prompts

Yoshinari Fujinuma, Keisuke Sakaguchi• 2026

Related benchmarks

TaskDatasetResultRank
Summarization EvaluationSummEval
Pearson Correlation0.546
40
LLM-as-a-JudgeBigGen-Bench (test)
Pearson Correlation0.312
4
Multi-hop Fact VerificationHoVer few-shot
Recall55.6
4
Showing 3 of 3 rows

Other info

Follow for update