Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Cascaded Local Implicit Transformer for Arbitrary-Scale Super-Resolution

About

Implicit neural representation has recently shown a promising ability in representing images with arbitrary resolutions. In this paper, we present a Local Implicit Transformer (LIT), which integrates the attention mechanism and frequency encoding technique into a local implicit image function. We design a cross-scale local attention block to effectively aggregate local features. To further improve representative power, we propose a Cascaded LIT (CLIT) that exploits multi-scale features, along with a cumulative training strategy that gradually increases the upsampling scales during training. We have conducted extensive experiments to validate the effectiveness of these components and analyze various training strategies. The qualitative and quantitative results demonstrate that LIT and CLIT achieve favorable results and outperform the prior works in arbitrary super-resolution tasks.

Hao-Wei Chen, Yu-Syuan Xu, Min-Fong Hong, Yi-Min Tsai, Hsien-Kai Kuo, Chun-Yi Lee• 2023

Related benchmarks

TaskDatasetResultRank
Image Super-resolutionSet5
PSNR38.53
507
Super-ResolutionB100 (test)
PSNR32.49
363
Super-ResolutionSet14 (test)
PSNR34.31
246
Image Super-resolutionUrban100
PSNR33.42
221
Image Super-resolutionBSD100
PSNR (dB)32.52
210
Image Super-resolutionSet14
PSNR34.44
115
Super-ResolutionDIV2K 1.0 (val)
PSNR34.96
100
Super-ResolutionUrban100 official (test)
PSNR33.62
56
Super-ResolutionSet5 official (test)
PSNR38.4
54
Super-ResolutionDIV2K 46 (val)
PSNR (2x)35.31
18
Showing 10 of 10 rows

Other info

Code

Follow for update