Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

VADMamba: Exploring State Space Models for Fast Video Anomaly Detection

About

Video anomaly detection (VAD) methods are mostly CNN-based or Transformer-based, achieving impressive results, but the focus on detection accuracy often comes at the expense of inference speed. The emergence of state space models in computer vision, exemplified by the Mamba model, demonstrates improved computational efficiency through selective scans and showcases the great potential for long-range modeling. Our study pioneers the application of Mamba to VAD, dubbed VADMamba, which is based on multi-task learning for frame prediction and optical flow reconstruction. Specifically, we propose the VQ-Mamba Unet (VQ-MaU) framework, which incorporates a Vector Quantization (VQ) layer and Mamba-based Non-negative Visual State Space (NVSS) block. Furthermore, two individual VQ-MaU networks separately predict frames and reconstruct corresponding optical flows, further boosting accuracy through a clip-level fusion evaluation strategy. Experimental results validate the efficacy of the proposed VADMamba across three benchmark datasets, demonstrating superior performance in inference speed compared to previous work. Code is available at https://github.com/jLooo/VADMamba.

Jiahao Lyu, Minghua Zhao, Jing Hu, Xuewen Huang, Yifei Chen, Shuangli Du• 2025

Related benchmarks

TaskDatasetResultRank
Video Anomaly DetectionShanghaiTech (SHT) (test)
Frame-level AUC77
44
Video Anomaly DetectionCUHK Avenue (test)
Frame-level AUC0.915
35
Video Anomaly DetectionUCSD Ped2 (test)
Frame-level AUC98.5
35
Video Anomaly DetectionUIT-ADrone
Micro-AUC68.6
13
Video Anomaly DetectionMUVAD
Micro-AUC66.2
13
Video Anomaly DetectionDrone-Anomaly
Micro-AUC66.7
13
Showing 6 of 6 rows

Other info

Code

Follow for update