Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Vad-R1: Towards Video Anomaly Reasoning via Perception-to-Cognition Chain-of-Thought

About

Recent advancements in reasoning capability of Multimodal Large Language Models (MLLMs) demonstrate its effectiveness in tackling complex visual tasks. However, existing MLLM-based Video Anomaly Detection (VAD) methods remain limited to shallow anomaly descriptions without deep reasoning. In this paper, we propose a new task named Video Anomaly Reasoning (VAR), which aims to enable deep analysis and understanding of anomalies in the video by requiring MLLMs to think explicitly before answering. To this end, we propose Vad-R1, an end-to-end MLLM-based framework for VAR. Specifically, we design a Perception-to-Cognition Chain-of-Thought (P2C-CoT) that simulates the human process of recognizing anomalies, guiding the MLLM to reason anomaly step-by-step. Based on the structured P2C-CoT, we construct Vad-Reasoning, a dedicated dataset for VAR. Furthermore, we propose an improved reinforcement learning algorithm AVA-GRPO, which explicitly incentivizes the anomaly reasoning capability of MLLMs through a self-verification mechanism with limited annotations. Experimental results demonstrate that Vad-R1 achieves superior performance, outperforming both open-source and proprietary models on VAD and VAR tasks. Codes and datasets will be released at https://github.com/wbfwonderful/Vad-R1.

Chao Huang, Benfeng Wang, Jie Wen, Chengliang Liu, Wei Wang, Li Shen, Xiaochun Cao• 2025

Related benchmarks

TaskDatasetResultRank
Multi-choice Question AnsweringVad-Reasoning-Plus
MCQ Score92.9
27
Open-ended Question AnsweringVad-Reasoning-Plus
BLEU-30.023
27
Video Anomaly ReasoningVideo Anomaly Reasoning (test)
RR Score0.286
27
Showing 3 of 3 rows

Other info

Follow for update