Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

SkipVAR: Accelerating Visual Autoregressive Modeling via Adaptive Frequency-Aware Skipping

About

Recent studies on Visual Autoregressive (VAR) models have highlighted that high-frequency components, or later steps, in the generation process contribute disproportionately to inference latency. However, the underlying computational redundancy involved in these steps has yet to be thoroughly investigated. In this paper, we conduct an in-depth analysis of the VAR inference process and identify two primary sources of inefficiency: step redundancy and unconditional branch redundancy. To address step redundancy, we propose an automatic step-skipping strategy that selectively omits unnecessary generation steps to improve efficiency. For unconditional branch redundancy, we observe that the information gap between the conditional and unconditional branches is minimal. Leveraging this insight, we introduce unconditional branch replacement, a technique that bypasses the unconditional branch to reduce computational cost. Notably, we observe that the effectiveness of acceleration strategies varies significantly across different samples. Motivated by this, we propose SkipVAR, a sample-adaptive framework that leverages frequency information to dynamically select the most suitable acceleration strategy for each instance. To evaluate the role of high-frequency information, we introduce high-variation benchmark datasets that test model sensitivity to fine details. Extensive experiments show SkipVAR achieves over 0.88 average SSIM with up to 1.81x overall acceleration and 2.62x speedup on the GenEval benchmark, maintaining model quality. These results confirm the effectiveness of frequency-aware, training-free adaptive acceleration for scalable autoregressive image generation. Our code is available at https://github.com/fakerone-li/SkipVAR and has been publicly released.

Jiajun Li, Yue Ma, Xinyu Zhang, Qingyan Wei, Songhua Liu, Linfeng Zhang• 2025

Related benchmarks

TaskDatasetResultRank
Text-to-Image GenerationGenEval
GenEval Score77
277
Text-to-Image GenerationDPG-Bench
Overall Score86.4
173
Text-to-Image GenerationDPG
Overall Score83.16
131
Text-to-Image GenerationGenEval
Two Objects84
87
Text-to-Image GenerationImageReward
ImageReward Score1.032
56
Text-to-Image GenerationDPG-Bench (test)
Global Fidelity92.648
43
Text-to-Image GenerationGenEval 1024x1024
Latency (s)0.72
22
Human Preference EvaluationImageReward
Average Score1.0297
16
Human Preference EvaluationHPS v2.1
Photo Score29.31
16
Text-to-Image GenerationHPS v2.1
Score (Anime)32.01
9
Showing 10 of 10 rows

Other info

Follow for update