ReFusion: A Diffusion Large Language Model with Parallel Autoregressive Decoding
About
Autoregressive models (ARMs) are hindered by slow sequential inference. While masked diffusion models (MDMs) offer a parallel alternative, they suffer from critical drawbacks: high computational overhead from precluding Key-Value (KV) caching, and incoherent generation arising from learning dependencies over an intractable space of token combinations. To address these limitations, we introduce ReFusion, a novel masked diffusion model that achieves superior performance and efficiency by elevating parallel decoding from the token level to a higher slot level, where each slot is a fixed-length, contiguous sub-sequence. This is achieved through an iterative ``plan-and-infill'' decoding process: a diffusion-based planning step first identifies a set of weakly dependent slots, and an autoregressive infilling step then decodes these selected slots in parallel. The slot-based design simultaneously unlocks full KV cache reuse with a unified causal framework and reduces the learning complexity from the token combination space to a manageable slot-level permutation space. Extensive experiments on seven diverse benchmarks show that ReFusion not only overwhelmingly surpasses prior MDMs with 34% performance gains and an over 18$\times$ speedup on average, but also bridges the performance gap to strong ARMs while maintaining a 2.33$\times$ average speedup.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Code Generation | HumanEval | Pass@178.66 | 850 | |
| Mathematical Reasoning | GSM8K | -- | 351 | |
| Mathematical Problem Solving | MATH | Accuracy54.22 | 166 | |
| Code Generation | MBPP | Accuracy (%)68.2 | 146 | |
| Code Generation | MBPP | Pass@154.12 | 113 | |
| Math | GSM8K | Accuracy0.8491 | 87 | |
| Question Answering | GPQA Diamond | Pass@133.43 | 49 | |
| General Reasoning | MMLU-Pro | Accuracy45.94 | 48 | |
| Reasoning | ARC-C | Accuracy89.76 | 42 | |
| Question Answering | ARC | pass@187.98 | 30 |