Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Structured Over Scale: Learning Spatial Reasoning from Educational Video

About

Vision-language models (VLMs) demonstrate impressive performance on standard video understanding benchmarks yet fail systematically on simple reasoning tasks that preschool children can solve, including counting, spatial reasoning, and compositional understanding. We hypothesize that the pedagogically-structured content of educational videos provides an ideal training signal for improving these capabilities. We introduce DoraVQA, a dataset of 5,344 question-answer pairs automatically extracted from 8 seasons of Dora the Explorer with precise timestamp alignment. Each episode follows a consistent \textit{context-question-pause-answer} structure that creates a self-contained learning environment analogous to interactive tutoring. We fine-tune both Qwen2 and Qwen3 using Group Relative Policy Optimization (GRPO), leveraging the clear correctness signals and structured reasoning traces inherent in educational content. Despite training exclusively on 38 hours of children's educational videos, our approach achieves improvements of 8-14 points on DoraVQA and state-of-the-art 86.16\% on CVBench, with strong transfer to Video-MME and NExT-QA, demonstrating effective generalization from narrow pedagogical content to broad multimodal understanding. Through cross-domain benchmarks, we show that VLMs can perform tasks that require robust reasoning learned from structured educational content, suggesting that content structure matters as much as content scale.

Bishoy Galoaa, Xiangyu Bai, Sarah Ostadabbas• 2026

Related benchmarks

TaskDatasetResultRank
Video Question AnsweringDoraVQA (test)
Overall Accuracy67.98
13
Vision-Language ReasoningCVBench
Accuracy86.16
12
Video Question AnsweringDoraVQA
Accuracy67.98
12
Video Question AnsweringVideo-MME
Accuracy76.78
12
Showing 4 of 4 rows

Other info

Follow for update