Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Text-Aware Image Restoration with Diffusion Models

About

Image restoration aims to recover degraded images. However, existing diffusion-based restoration methods, despite great success in natural image restoration, often struggle to faithfully reconstruct textual regions in degraded images. Those methods frequently generate plausible but incorrect text-like patterns, a phenomenon we refer to as text-image hallucination. In this paper, we introduce Text-Aware Image Restoration (TAIR), a novel restoration task that requires the simultaneous recovery of visual contents and textual fidelity. To tackle this task, we present SA-Text, a large-scale benchmark of 100K high-quality scene images densely annotated with diverse and complex text instances. Furthermore, we propose a multi-task diffusion framework, called TeReDiff, that integrates internal features from diffusion models into a text-spotting module, enabling both components to benefit from joint training. This allows for the extraction of rich text representations, which are utilized as prompts in subsequent denoising steps. Extensive experiments demonstrate that our approach consistently outperforms state-of-the-art restoration methods, achieving significant gains in text recognition accuracy. See our project page: https://cvlab-kaist.github.io/TAIR/

Jaewon Min, Jin Hyeon Kim, Paul Hyunbin Cho, Jaeeun Lee, Jihye Park, Minkyu Park, Sangpil Kim, Hyunhee Park, Seungryong Kim• 2025

Related benchmarks

TaskDatasetResultRank
End-to-End Text SpottingReal-Text
Accuracy (None Config)49.39
26
Text DetectionReal-Text
Precision84.3
26
Image RestorationSA-Text 1K (test)
PSNR19.71
4
Image RestorationReal-Text Real-world LQ-HQ pairs (test)
PSNR23
4
Showing 4 of 4 rows

Other info

Follow for update