Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Reflection Removal through Efficient Adaptation of Diffusion Transformers

About

We introduce a diffusion-transformer (DiT) framework for single-image reflection removal that leverages the generalization strengths of foundation diffusion models in the restoration setting. Rather than relying on task-specific architectures, we repurpose a pre-trained DiT-based foundation model by conditioning it on reflection-contaminated inputs and guiding it toward clean transmission layers. We systematically analyze existing reflection removal data sources for diversity, scalability, and photorealism. To address the shortage of suitable data, we construct a physically based rendering (PBR) pipeline in Blender, built around the Principled BSDF, to synthesize realistic glass materials and reflection effects. Efficient LoRA-based adaptation of the foundation model, combined with the proposed synthetic data, achieves state-of-the-art performance on in-domain and zero-shot benchmarks. These results demonstrate that pretrained diffusion transformers, when paired with physically grounded data synthesis and efficient adaptation, offer a scalable and high-fidelity solution for reflection removal. Project page: https://hf.co/spaces/huawei-bayerlab/windowseat-reflection-removal-web

Daniyar Zakarin, Thiemo Wandel, Anton Obukhov, Dengxin Dai• 2025

Related benchmarks

TaskDatasetResultRank
Image Reflection RemovalReal20
PSNR26.6
56
Single Image Reflection RemovalNature 20
PSNR27.57
15
Reflection RemovalSIR2 zero-shot 454/500
PSNR (SIR2 454, Zero-Shot)28.99
11
Reflection RemovalNature 20 (test)
MS-SSIM94.94
6
Reflection RemovalReal 20 (test)
MS-SSIM0.9396
6
Reflection RemovalSIR2 (test)
MS-SSIM (Objects)0.9759
6
Showing 6 of 6 rows

Other info

GitHub

Follow for update