Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Recovering Realistic Texture in Image Super-resolution by Deep Spatial Feature Transform

About

Despite that convolutional neural networks (CNN) have recently demonstrated high-quality reconstruction for single-image super-resolution (SR), recovering natural and realistic texture remains a challenging problem. In this paper, we show that it is possible to recover textures faithful to semantic classes. In particular, we only need to modulate features of a few intermediate layers in a single network conditioned on semantic segmentation probability maps. This is made possible through a novel Spatial Feature Transform (SFT) layer that generates affine transformation parameters for spatial-wise feature modulation. SFT layers can be trained end-to-end together with the SR network using the same loss function. During testing, it accepts an input image of arbitrary size and generates a high-resolution image with just a single forward pass conditioned on the categorical priors. Our final results show that an SR network equipped with SFT can generate more realistic and visually pleasing textures in comparison to state-of-the-art SRGAN and EnhanceNet.

Xintao Wang, Ke Yu, Chao Dong, Chen Change Loy• 2018

Related benchmarks

TaskDatasetResultRank
Image Super-resolutionSet5--
507
Super-ResolutionSet14 (test)
PSNR26.743
246
Image Super-resolutionUrban100
PSNR24.34
221
Super-ResolutionBSD100
PSNR24.09
149
Super-ResolutionDIV2K
PSNR26.56
101
Super-ResolutionBSD100 4x (test)
PSNR24.09
56
Super-ResolutionManga109 (test)
PSNR28.167
46
Super-ResolutionDIV2K (val)
PSNR28.08
44
Image Super-resolutionManga109
LPIPS0.072
38
Super-ResolutionGeneral100
LPIPS0.103
25
Showing 10 of 16 rows

Other info

Follow for update