Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Exploring Sparsity in Image Super-Resolution for Efficient Inference

About

Current CNN-based super-resolution (SR) methods process all locations equally with computational resources being uniformly assigned in space. However, since missing details in low-resolution (LR) images mainly exist in regions of edges and textures, less computational resources are required for those flat regions. Therefore, existing CNN-based methods involve redundant computation in flat regions, which increases their computational cost and limits their applications on mobile devices. In this paper, we explore the sparsity in image SR to improve inference efficiency of SR networks. Specifically, we develop a Sparse Mask SR (SMSR) network to learn sparse masks to prune redundant computation. Within our SMSR, spatial masks learn to identify "important" regions while channel masks learn to mark redundant channels in those "unimportant" regions. Consequently, redundant computation can be accurately localized and skipped while maintaining comparable performance. It is demonstrated that our SMSR achieves state-of-the-art performance with 41%/33%/27% FLOPs being reduced for x2/3/4 SR. Code is available at: https://github.com/LongguangWang/SMSR.

Longguang Wang, Xiaoyu Dong, Yingqian Wang, Xinyi Ying, Zaiping Lin, Wei An, Yulan Guo• 2020

Related benchmarks

TaskDatasetResultRank
Image Super-resolutionSet5 (test)
PSNR38
544
Super-ResolutionB100 (test)
PSNR32.17
363
Image Super-resolutionSet14 (test)
PSNR33.64
292
Single Image Super-ResolutionUrban100 (test)
PSNR32.19
289
Image Super-resolutionManga109 (test)
PSNR38.76
233
Super-ResolutionSet5 x2
PSNR38
134
Super-ResolutionSet5 x3
PSNR34.4
108
Super-ResolutionManga109 4x
PSNR30.54
88
Super-ResolutionUrban100 x2
PSNR32.19
86
Super-ResolutionUrban100 x4
PSNR26.11
85
Showing 10 of 20 rows

Other info

Follow for update