Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Data Upcycling Knowledge Distillation for Image Super-Resolution

About

Knowledge distillation (KD) compresses deep neural networks by transferring task-related knowledge from cumbersome pre-trained teacher models to compact student models. However, current KD methods for super-resolution (SR) networks overlook the nature of SR task that the outputs of the teacher model are noisy approximations to the ground-truth distribution of high-quality images (GT), which shades the teacher model's knowledge to result in limited KD effects. To utilize the teacher model beyond the GT upper-bound, we present the Data Upcycling Knowledge Distillation (DUKD), to transfer the teacher model's knowledge to the student model through the upcycled in-domain data derived from training data. Besides, we impose label consistency regularization to KD for SR by the paired invertible augmentations to improve the student model's performance and robustness. Comprehensive experiments demonstrate that the DUKD method significantly outperforms previous arts on several SR tasks.

Yun Zhang, Wei Li, Simiao Li, Hanting Chen, Zhijun Tu, Wenjia Wang, Bingyi Jing, Shaohui Lin, Jie Hu• 2023

Related benchmarks

TaskDatasetResultRank
Image Super-resolutionUrban100 x4 (test)
PSNR26.43
282
Super-ResolutionSet14 (test)
PSNR33.9
254
Super-ResolutionUrban100 (test)
PSNR32.87
220
Image Super-resolutionBSD100 (test)
PSNR32.33
220
Super-ResolutionSet5 (test)
PSNR38.23
192
Super-ResolutionSet14 4x (test)
PSNR28.79
131
Image Super-resolutionUrban100 x2 (test)
PSNR32.63
91
Image Super-resolutionUrban100 x3 (test)
PSNR28.53
72
Super-ResolutionBSD100 4x (test)
PSNR27.69
70
Image Super-resolutionSet14 x2 scale (test)
PSNR33.78
46
Showing 10 of 18 rows

Other info

Code

Follow for update