CamoFormer: Masked Separable Attention for Camouflaged Object Detection
About
How to identify and segment camouflaged objects from the background is challenging. Inspired by the multi-head self-attention in Transformers, we present a simple masked separable attention (MSA) for camouflaged object detection. We first separate the multi-head self-attention into three parts, which are responsible for distinguishing the camouflaged objects from the background using different mask strategies. Furthermore, we propose to capture high-resolution semantic representations progressively based on a simple top-down decoder with the proposed MSA to attain precise segmentation results. These structures plus a backbone encoder form a new model, dubbed CamoFormer. Extensive experiments show that CamoFormer surpasses all existing state-of-the-art methods on three widely-used camouflaged object detection benchmarks. There are on average around 5% relative improvements over previous methods in terms of S-measure and weighted F-measure.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Camouflaged Object Detection | COD10K (test) | S-measure (S_alpha)0.872 | 224 | |
| Camouflaged Object Detection | COD10K | S-measure (S_alpha)0.868 | 178 | |
| Camouflaged Object Detection | Chameleon | S-measure (S_alpha)91 | 150 | |
| Camouflaged Object Detection | CAMO (test) | -- | 111 | |
| Camouflaged Object Detection | NC4K | M Score0.031 | 67 | |
| Camouflaged Object Detection | CAMO 250 (test) | M (Mean Score)0.046 | 59 | |
| Camouflaged Object Detection | NC4K | Sm89.2 | 58 | |
| Camouflaged Object Detection | CAMO | M Score0.043 | 37 | |
| Camouflaged Object Detection | CAMO 1.0 (test) | MAE0.046 | 23 | |
| Camouflaged Object Detection | COD10K 1.0 (test) | MAE0.023 | 23 |