Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Image Super-Resolution Using Very Deep Residual Channel Attention Networks

About

Convolutional neural network (CNN) depth is of crucial importance for image super-resolution (SR). However, we observe that deeper networks for image SR are more difficult to train. The low-resolution inputs and features contain abundant low-frequency information, which is treated equally across channels, hence hindering the representational ability of CNNs. To solve these problems, we propose the very deep residual channel attention networks (RCAN). Specifically, we propose a residual in residual (RIR) structure to form very deep network, which consists of several residual groups with long skip connections. Each residual group contains some residual blocks with short skip connections. Meanwhile, RIR allows abundant low-frequency information to be bypassed through multiple skip connections, making the main network focus on learning high-frequency information. Furthermore, we propose a channel attention mechanism to adaptively rescale channel-wise features by considering interdependencies among channels. Extensive experiments show that our RCAN achieves better accuracy and visual improvements against state-of-the-art methods.

Yulun Zhang, Kunpeng Li, Kai Li, Lichen Wang, Bineng Zhong, Yun Fu• 2018

Related benchmarks

TaskDatasetResultRank
Image Super-resolutionManga109
PSNR42.33
821
Super-ResolutionSet5
PSNR40.97
785
Image Super-resolutionSet5
PSNR38.27
692
Super-ResolutionUrban100
PSNR35.93
652
Super-ResolutionSet14
PSNR34.12
613
Image Super-resolutionSet5 (test)
PSNR38.27
566
Image Super-resolutionSet14
PSNR34.12
506
Single Image Super-ResolutionUrban100
PSNR33.54
500
Super-ResolutionB100
PSNR32.46
429
Image Super-resolutionUrban100
PSNR33.34
406
Showing 10 of 212 rows
...

Other info

Code

Follow for update