Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Enhancing Perceptual Quality in Video Super-Resolution through Temporally-Consistent Detail Synthesis using Diffusion Models

About

In this paper, we address the problem of enhancing perceptual quality in video super-resolution (VSR) using Diffusion Models (DMs) while ensuring temporal consistency among frames. We present StableVSR, a VSR method based on DMs that can significantly enhance the perceptual quality of upscaled videos by synthesizing realistic and temporally-consistent details. We introduce the Temporal Conditioning Module (TCM) into a pre-trained DM for single image super-resolution to turn it into a VSR method. TCM uses the novel Temporal Texture Guidance, which provides it with spatially-aligned and detail-rich texture information synthesized in adjacent frames. This guides the generative process of the current frame toward high-quality and temporally-consistent results. In addition, we introduce the novel Frame-wise Bidirectional Sampling strategy to encourage the use of information from past to future and vice-versa. This strategy improves the perceptual quality of the results and the temporal consistency across frames. We demonstrate the effectiveness of StableVSR in enhancing the perceptual quality of upscaled videos while achieving better temporal consistency compared to existing state-of-the-art methods for VSR. The project page is available at https://github.com/claudiom4sir/StableVSR.

Claudio Rota, Marco Buzzelli, Joost van de Weijer• 2023

Related benchmarks

TaskDatasetResultRank
Video Super-ResolutionVid4 (test)
PSNR22.213
173
Video Super-ResolutionREDS4 (test)
PSNR (Avg)27.928
117
Video RestorationREDS30
PSNR23.19
17
Video Super-ResolutionVideoLQ (test)
NRQM6.154
17
Video RestorationVideoLQ
MUSIQ31.85
7
Video Super-ResolutionVideoLQ
NIQE3.982
5
Showing 6 of 6 rows

Other info

Follow for update