Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

SpaceR: Reinforcing MLLMs in Video Spatial Reasoning

About

Video spatial reasoning, which involves inferring the underlying spatial structure from observed video frames, poses a significant challenge for existing Multimodal Large Language Models (MLLMs). This limitation stems primarily from 1) the absence of high-quality datasets for this task, and 2) the lack of effective training strategies to develop spatial reasoning capabilities. Motivated by the success of Reinforcement Learning with Verifiable Reward (RLVR) in unlocking LLM reasoning abilities, this work aims to improve MLLMs in video spatial reasoning through the RLVR paradigm. To this end, we introduce the $\textbf{SpaceR}$ framework. First, we present $\textbf{SpaceR-151k}$, a dataset with 91k questions spanning diverse spatial reasoning scenarios with verifiable answers, and 60k samples for maintaining general multimodal understanding. Second, we propose $\textbf{Spatially-Guided RLVR (SG-RLVR)}$, a novel reinforcement learning approach that extends Group Relative Policy Optimization (GRPO) with a novel map imagination mechanism, which encourages the model to infer spatial layouts in the thinking process, thereby facilitating more effective spatial reasoning. Extensive experiments demonstrate that SpaceR achieves state-of-the-art performance on spatial reasoning benchmarks (e.g., VSI-Bench, STI-Bench, and SPAR-Bench), while maintaining competitive results on video understanding benchmarks (e.g., Video-MME, TempCompass, and LongVideoBench). Remarkably, SpaceR surpasses the advanced GPT-4o by 11.6\% accuracy on VSI-Bench and is on par with the leading proprietary model Gemini-2.0-Flash, highlighting the effectiveness of our SpaceR-151k dataset and SG-RLVR in reinforcing spatial reasoning ability of MLLMs. Code, model, and dataset are available at https://github.com/OuyangKun10/SpaceR.

Kun Ouyang, Yuanxin Liu, Haoning Wu, Yi Liu, Hao Zhou, Jie Zhou, Fandong Meng, Xu Sun• 2025

Related benchmarks

TaskDatasetResultRank
Visual ReasoningBLINK
Accuracy55.4
50
Spatial ReasoningMindCube
Accuracy37.9
37
Spatial ReasoningMindCube tiny (test)
Rot. Accuracy29.8
30
Spatial ReasoningMMSI-Bench (test)
PR Score30.3
29
Spatial ReasoningViewspatial
Accuracy37.3
28
Vision-centric EvaluationCV-Bench
Accuracy0.748
21
Multi-image Spatial ReasoningMindCube tiny (test)
Aro Accuracy31.6
17
Multi-image Spatial ReasoningSPAR-Bench-MV (test)
Score (Low Difficulty)31.6
15
Single-image spatial reasoningSPAR-Bench SI
Low Score33.6
15
Single-image spatial reasoningCV-Bench
2D Accuracy71.9
15
Showing 10 of 30 rows

Other info

Follow for update