Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

UCAN: Unified Convolutional Attention Network for Expansive Receptive Fields in Lightweight Super-Resolution

About

Hybrid CNN-Transformer architectures achieve strong results in image super-resolution, but scaling attention windows or convolution kernels significantly increases computational cost, limiting deployment on resource-constrained devices. We present UCAN, a lightweight network that unifies convolution and attention to expand the effective receptive field efficiently. UCAN combines window-based spatial attention with a Hedgehog Attention mechanism to model both local texture and long-range dependencies, and introduces a distillation-based large-kernel module to preserve high-frequency structure without heavy computation. In addition, we employ cross-layer parameter sharing to further reduce complexity. On Manga109 ($4\times$), UCAN-L achieves 31.63 dB PSNR with only 48.4G MACs, surpassing recent lightweight models. On BSDS100, UCAN attains 27.79 dB, outperforming methods with significantly larger models. Extensive experiments show that UCAN achieves a superior trade-off between accuracy, efficiency, and scalability, making it well-suited for practical high-resolution image restoration.

Cao Thien Tan, Phan Thi Thu Trang, Do Nghiem Duc, Ho Ngoc Anh, Hanyang Zhuang, Nguyen Duc Dung• 2026

Related benchmarks

TaskDatasetResultRank
Super-ResolutionUrban100 x2
PSNR33.39
104
Super-ResolutionUrban100 x4
PSNR27.06
103
Super-ResolutionManga109 4x
PSNR31.63
99
Super-ResolutionUrban100 x3
PSNR29.29
91
Super-ResolutionManga109 2x
PSNR39.66
71
Super-ResolutionSet14 2x
PSNR34.27
63
Super-ResolutionBSDS100 2x
PSNR32.44
50
Super-ResolutionManga109 x3
PSNR34.79
49
Super-ResolutionBSDS100 4x
PSNR27.8
40
Image Super-resolutionSet5 2x
PSNR38.37
37
Showing 10 of 15 rows

Other info

Follow for update