Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Modality-Agnostic Prompt Learning for Multi-Modal Camouflaged Object Detection

About

Camouflaged Object Detection (COD) aims to segment objects that blend seamlessly into complex backgrounds, with growing interest in exploiting additional visual modalities to enhance robustness through complementary information. However, most existing approaches generally rely on modality-specific architectures or customized fusion strategies, which limit scalability and cross-modal generalization. To address this, we propose a novel framework that generates modality-agnostic multi-modal prompts for the Segment Anything Model (SAM), enabling parameter-efficient adaptation to arbitrary auxiliary modalities and significantly improving overall performance on COD tasks. Specifically, we model multi-modal learning through interactions between a data-driven content domain and a knowledge-driven prompt domain, distilling task-relevant cues into unified prompts for SAM decoding. We further introduce a lightweight Mask Refine Module to calibrate coarse predictions by incorporating fine-grained prompt cues, leading to more accurate camouflaged object boundaries. Extensive experiments on RGB-Depth, RGB-Thermal, and RGB-Polarization benchmarks validate the effectiveness and generalization of our modality-agnostic framework.

Hao Wang, Jiqing Zhang, Xin Yang, Baocai Yin, Lu Jiang, Zetian Mi, Huibing Wang• 2026

Related benchmarks

TaskDatasetResultRank
Camouflaged Object DetectionCOD10K
S-measure (S_alpha)0.901
178
Camouflaged Object DetectionChameleon
S-measure (S_alpha)88.7
150
Camouflaged Object DetectionNC4K
M Score0.031
67
Camouflaged Object DetectionCAMO
Weighted F-beta (Fwβ)0.862
44
Salient Object DetectionVT821
S-Measure0.944
43
Camouflaged Object DetectionVIAC RGB-T
93
12
Salient Object DetectionVT1000
S_alpha0.954
7
Showing 7 of 7 rows

Other info

Follow for update