Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Visual Representation Alignment for Multimodal Large Language Models

About

Multimodal large language models (MLLMs) trained with visual instruction tuning have achieved strong performance across diverse tasks, yet they remain limited in vision-centric tasks such as object counting or spatial reasoning. We attribute this gap to the prevailing text-only supervision paradigm, which provides only indirect guidance for the visual pathway and often leads MLLMs to discard fine-grained visual details during training. In this paper, we present VIsual Representation ALignment (VIRAL), a simple yet effective regularization strategy that aligns the internal visual representations of MLLMs with those of pre-trained vision foundation models (VFMs). By explicitly enforcing this alignment, VIRAL enables the model not only to retain critical visual details from the input vision encoder but also to complement additional visual knowledge from VFMs, thereby enhancing its ability to reason over complex visual inputs. Our experiments demonstrate consistent improvements across all tasks on widely adopted multimodal benchmarks. Furthermore, we conduct comprehensive ablation studies to validate the key design choices underlying our framework. We believe this simple finding opens up an important direction for the effective integration of visual information in training MLLMs.

Heeji Yoon, Jaewoo Jung, Junwan Kim, Hyungyu Choi, Heeseong Shin, Sangbeom Lim, Honggyu An, Chaehyun Kim, Jisang Han, Donghyun Kim, Chanho Eom, Sunghwan Hong, Seungryong Kim• 2025

Related benchmarks

TaskDatasetResultRank
Visual ReasoningBLINK
Accuracy49.3
76
Multimodal Visual PerceptionMMVP
Accuracy36
72
Real-world Question AnsweringRealworldQA--
58
Vision-centric EvaluationCV-Bench 2D
Score60.5
15
2D Computer Vision BenchmarkingCVBench2D
Accuracy62
13
Knowledge-based Vision-Language UnderstandingKnowledge
Average Score47.2
8
General Vision-Language UnderstandingGeneral
Avg Score70.9
8
Vision-Centric Multi-modal EvaluationVision-Centric
Average Score53.5
8
3D Computer Vision BenchmarkingCVBench3D
Accuracy62.3
8
Optical Character RecognitionOCR
Average Score36.1
8
Showing 10 of 12 rows

Other info

Follow for update