VisualSphinx: Large-Scale Synthetic Vision Logic Puzzles for RL
About
Vision language models (VLMs) are expected to perform effective multimodal reasoning and make logically coherent decisions, which is critical to tasks such as diagram understanding and spatial problem solving. However, current VLM reasoning lacks large-scale and well-structured training datasets. To bridge this gap, we propose VisualSphinx, a first-of-its-kind large-scale synthetic visual logical reasoning training data. To tackle the challenge of image synthesis with grounding answers, we propose a rule-to-image synthesis pipeline, which extracts and expands puzzle rules from seed questions and generates the code of grounding synthesis image synthesis for puzzle sample assembly. Experiments demonstrate that VLM trained using GRPO on VisualSphinx benefit from logical coherence and readability of our dataset and exhibit improved performance on logical reasoning tasks. The enhanced reasoning capabilities developed from VisualSphinx also benefit other reasoning tasks such as algebraic reasoning, arithmetic reasoning and geometry reasoning.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Object Hallucination Evaluation | POPE | -- | 935 | |
| Multimodal Evaluation | MME | Score2.49e+3 | 557 | |
| Multimodal Evaluation | SEED-Bench | Accuracy75.47 | 80 | |
| Multimodal Evaluation | MMStar | Accuracy63.2 | 46 | |
| Vision Understanding | CVBench 2D | Accuracy73.98 | 22 | |
| Color Understanding | ColorBench | Accuracy39.9 | 18 | |
| Visual Grounding | Lisa Grounding | Accuracy73.28 | 18 | |
| Multimodal Visual Pattern Understanding | MMVP | Accuracy77.33 | 16 | |
| Multimodal Evaluation | MMT-Bench | Accuracy60.63 | 13 | |
| Vision-Language Reasoning | MMStar cleaned | Score75.27 | 10 |