Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

OpenMMReasoner: Pushing the Frontiers for Multimodal Reasoning with an Open and General Recipe

About

Recent advancements in large reasoning models have fueled growing interest in extending such capabilities to multimodal domains. However, despite notable progress in visual reasoning, the lack of transparent and reproducible data curation and training strategies remains a major barrier to scalable research. In this work, we introduce OpenMMReasoner, a fully transparent two-stage recipe for multimodal reasoning spanning supervised fine-tuning (SFT) and reinforcement learning (RL). In the SFT stage, we construct an 874K-sample cold-start dataset with rigorous step-by-step validation, providing a strong foundation for reasoning capabilities. The subsequent RL stage leverages a 74K-sample dataset across diverse domains to further sharpen and stabilize these abilities, resulting in a more robust and efficient learning process. Extensive evaluations demonstrate that our training recipe not only surpasses strong baselines but also highlights the critical role of data quality and training design in shaping multimodal reasoning performance. Notably, our method achieves a 11.6% improvement over the Qwen2.5-VL-7B-Instruct baseline across nine multimodal reasoning benchmarks, establishing a solid empirical foundation for future large-scale multimodal reasoning research. We open-sourced all our codes, pipeline, and data at https://github.com/EvolvingLMMs-Lab/OpenMMReasoner.

Kaichen Zhang, Keming Wu, Zuhao Yang, Bo Li, Kairui Hu, Bin Wang, Ziwei Liu, Xingxuan Li, Lidong Bing• 2025

Related benchmarks

TaskDatasetResultRank
Visual Question AnsweringMMStar
Accuracy69
57
Mathematical ReasoningMathVerse mini
Accuracy63.8
50
Mathematical ReasoningMathVision (test)
Accuracy43.6
41
Visual Question AnsweringRealWorldQA (test)
Accuracy69.4
36
Math & KnowledgeMathVista mini
Accuracy79.5
25
Multimodal ReasoningMMMU-Pro
Pass@166.82
18
Document UnderstandingChartXiv-DQ
Accuracy73.5
16
Mathematical multi-modal reasoningWeMath
Pass@183.45
13
General multimodal reasoningM3CoT
Pass@1 Accuracy78.21
11
Multimodal Mathematical ReasoningMathVerse (vision)
Pass@1 Accuracy74.87
11
Showing 10 of 19 rows

Other info

GitHub

Follow for update