Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

An Interactively Reinforced Paradigm for Joint Infrared-Visible Image Fusion and Saliency Object Detection

About

This research focuses on the discovery and localization of hidden objects in the wild and serves unmanned systems. Through empirical analysis, infrared and visible image fusion (IVIF) enables hard-to-find objects apparent, whereas multimodal salient object detection (SOD) accurately delineates the precise spatial location of objects within the picture. Their common characteristic of seeking complementary cues from different source images motivates us to explore the collaborative relationship between Fusion and Salient object detection tasks on infrared and visible images via an Interactively Reinforced multi-task paradigm for the first time, termed IRFS. To the seamless bridge of multimodal image fusion and SOD tasks, we specifically develop a Feature Screening-based Fusion subnetwork (FSFNet) to screen out interfering features from source images, thereby preserving saliency-related features. After generating the fused image through FSFNet, it is then fed into the subsequent Fusion-Guided Cross-Complementary SOD subnetwork (FC$^2$Net) as the third modality to drive the precise prediction of the saliency map by leveraging the complementary information derived from the fused image. In addition, we develop an interactive loop learning strategy to achieve the mutual reinforcement of IVIF and SOD tasks with a shorter training period and fewer network parameters. Comprehensive experiment results demonstrate that the seamless bridge of IVIF and SOD mutually enhances their performance, and highlights their superiority.

Di Wang, Jinyuan Liu, Risheng Liu, Xin Fan• 2023

Related benchmarks

TaskDatasetResultRank
Visible-Infrared Image FusionMSRS (test)
Average Gradient (AG)2.66
43
Semantic segmentationMSRS
mIoU65.37
42
Infrared-Visible Image FusionRoadScene (test)
Average Gradient (AG)4.08
40
Object DetectionMSRS (test)
mAP@0.597.8
34
Multi-Modal Image FusionMRI-CT (test)
EN5.15
30
Multi-Modal Image FusionMRI-SPECT (test)
Entropy (EN)5.39
16
Infrared-Visible Image FusionFMB (test)
Entropy (EN)6.64
16
Multi-Modal Image FusionMRI-PET (test)
EN5.6
16
Multi-Modal Image FusionMSRS
Inference Time (s)1.45
10
Showing 9 of 9 rows

Other info

Follow for update