EDMB: Edge Detector with Mamba
About
Transformer-based models have made significant progress in edge detection, but their high computational cost is prohibitive. Recently, vision Mamba have shown excellent ability in efficiently capturing long-range dependencies. Drawing inspiration from this, we propose a novel edge detector with Mamba, termed EDMB, to efficiently generate high-quality multi-granularity edges. In EDMB, Mamba is combined with a global-local architecture, therefore it can focus on both global information and fine-grained cues. The fine-grained cues play a crucial role in edge detection, but are usually ignored by ordinary Mamba. We design a novel decoder to construct learnable Gaussian distributions by fusing global features and fine-grained features. And the multi-grained edges are generated by sampling from the distributions. In order to make multi-granularity edges applicable to single-label data, we introduce Evidence Lower Bound loss to supervise the learning of the distributions. On the multi-label dataset BSDS500, our proposed EDMB achieves competitive single-granularity ODS 0.837 and multi-granularity ODS 0.851 without multi-scale test or extra PASCAL-VOC data. Remarkably, EDMB can be extended to single-label datasets such as NYUDv2 and BIPED. The source code is available at https://github.com/Li-yachuan/EDMB.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Boundary Detection | BSDS500 | ODS F-score0.837 | 37 | |
| Edge Detection | BSDS v1 (test) | ODS83.7 | 32 | |
| Edge Detection | NYUD Standard Evaluation - SEval v2 (val) | ODS78.3 | 17 | |
| Edge Detection | NYUD v2 | ODS0.774 | 16 | |
| Edge Detection | BIPED | ODS Score90.3 | 14 |