Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

PassionSR: Post-Training Quantization with Adaptive Scale in One-Step Diffusion based Image Super-Resolution

About

Diffusion-based image super-resolution (SR) models have shown superior performance at the cost of multiple denoising steps. However, even though the denoising step has been reduced to one, they require high computational costs and storage requirements, making it difficult for deployment on hardware devices. To address these issues, we propose a novel post-training quantization approach with adaptive scale in one-step diffusion (OSD) image SR, PassionSR. First, we simplify OSD model to two core components, UNet and Variational Autoencoder (VAE) by removing the CLIPEncoder. Secondly, we propose Learnable Boundary Quantizer (LBQ) and Learnable Equivalent Transformation (LET) to optimize the quantization process and manipulate activation distributions for better quantization. Finally, we design a Distributed Quantization Calibration (DQC) strategy that stabilizes the training of quantized parameters for rapid convergence. Comprehensive experiments demonstrate that PassionSR with 8-bit and 6-bit obtains comparable visual results with full-precision model. Moreover, our PassionSR achieves significant advantages over recent leading low-bit quantization methods for image SR. Our code will be at https://github.com/libozhu03/PassionSR.

Libo Zhu, Jianze Li, Haotong Qin, Wenbo Li, Yulun Zhang, Yong Guo, Xiaokang Yang• 2024

Related benchmarks

TaskDatasetResultRank
Image Super-resolutionDRealSR
MANIQA0.4684
78
Image Super-resolutionDIV2K (val)
LPIPS0.2373
59
Real-world Image Super-ResolutionRealLR200
MUSIQ55.96
26
Real-world Image Super-ResolutionRealLQ250
MUSIQ0.5709
26
Real-world Image Super-ResolutionDRealSR
LPIPS0.69
23
Real-world Image Super-ResolutionRealSR
LPIPS0.6885
23
Showing 6 of 6 rows

Other info

Code

Follow for update