Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Reversible Efficient Diffusion for Image Fusion

About

Multi-modal image fusion aims to consolidate complementary information from diverse source images into a unified representation. The fused image is expected to preserve fine details and maintain high visual fidelity. While diffusion models have demonstrated impressive generative capabilities in image generation, they often suffer from detail loss when applied to image fusion tasks. This issue arises from the accumulation of noise errors inherent in the Markov process, leading to inconsistency and degradation in the fused results. However, incorporating explicit supervision into end-to-end training of diffusion-based image fusion introduces challenges related to computational efficiency. To address these limitations, we propose the Reversible Efficient Diffusion (RED) model - an explicitly supervised training framework that inherits the powerful generative capability of diffusion models while avoiding the distribution estimation.

Xingxin Xu, Bing Cao, DongDong Li, Qinghua Hu, Pengfei Zhu• 2026

Related benchmarks

TaskDatasetResultRank
Object DetectionLLVIP (test)
mAP5096.2
38
Medical image fusionHarvard
EI20.24
10
Visible-Infrared Image FusionLLVIP
EI14.74
10
Visible-Infrared Image FusionMSRS
EI10.39
10
Visible-Infrared Image FusionM3FD
EI13.04
10
Image FusionLLVIP
EI14.74
4
Showing 6 of 6 rows

Other info

Follow for update