VMambaCC: A Visual State Space Model for Crowd Counting
About
As a deep learning model, Visual Mamba (VMamba) has a low computational complexity and a global receptive field, which has been successful applied to image classification and detection. To extend its applications, we apply VMamba to crowd counting and propose a novel VMambaCC (VMamba Crowd Counting) model. Naturally, VMambaCC inherits the merits of VMamba, or global modeling for images and low computational cost. Additionally, we design a Multi-head High-level Feature (MHF) attention mechanism for VMambaCC. MHF is a new attention mechanism that leverages high-level semantic features to augment low-level semantic features, thereby enhancing spatial feature representation with greater precision. Building upon MHF, we further present a High-level Semantic Supervised Feature Pyramid Network (HS2PFN) that progressively integrates and enhances high-level semantic information with low-level semantic information. Extensive experimental results on five public datasets validate the efficacy of our approach. For example, our method achieves a mean absolute error of 51.87 and a mean squared error of 81.3 on the ShangHaiTech\_PartA dataset. Our code is coming soon.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Crowd Counting | ShanghaiTech Part A (test) | MAE51.87 | 227 | |
| Crowd Counting | ShanghaiTech Part B (test) | MAE7.48 | 191 | |
| Crowd Counting | UCF-QNRF (test) | MAE88.42 | 95 | |
| Crowd Counting | UCF_CC_50 | MAE64.34 | 60 | |
| Crowd Localization | SHT A | Precision78.06 | 12 | |
| Crowd Counting | JHU-Crowd (test) | MAE54.41 | 7 |