Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Accelerating Diffusion Transformers with Token-wise Feature Caching

About

Diffusion transformers have shown significant effectiveness in both image and video synthesis at the expense of huge computation costs. To address this problem, feature caching methods have been introduced to accelerate diffusion transformers by caching the features in previous timesteps and reusing them in the following timesteps. However, previous caching methods ignore that different tokens exhibit different sensitivities to feature caching, and feature caching on some tokens may lead to 10$\times$ more destruction to the overall generation quality compared with other tokens. In this paper, we introduce token-wise feature caching, allowing us to adaptively select the most suitable tokens for caching, and further enable us to apply different caching ratios to neural layers in different types and depths. Extensive experiments on PixArt-$\alpha$, OpenSora, and DiT demonstrate our effectiveness in both image and video generation with no requirements for training. For instance, 2.36$\times$ and 1.93$\times$ acceleration are achieved on OpenSora and PixArt-$\alpha$ with almost no drop in generation quality.

Chang Zou, Xuyang Liu, Ting Liu, Siteng Huang, Linfeng Zhang• 2024

Related benchmarks

TaskDatasetResultRank
Class-conditional Image GenerationImageNet
FID3.03
132
Class-conditional Image GenerationImageNet-1k (val)
FID3.04
68
Text-to-Image GenerationImageReward
ImageReward Score1.202
56
Text-to-Image GenerationFLUX.1 (dev)
Image Reward0.9802
56
Class-conditional Image GenerationImageNet (val)
FID3.08
54
Class-to-image generationImageNet
FID3.04
25
Text-to-Image GenerationFLUX.1-schnell 1.0 (dev)
Latency (s)9.82
23
Text-to-Video GenerationHunyuanVideo
LPIPS0.44
22
Text-to-Video GenerationVBench HunyuanVideo (test)
VBench Score (%)78.86
21
Text-to-Image GenerationDrawBench
Latency (s)7.19
18
Showing 10 of 17 rows

Other info

Follow for update