Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Seeing Beyond Words: Self-Supervised Visual Learning for Multimodal Large Language Models

About

Multimodal Large Language Models (MLLMs) have recently demonstrated impressive capabilities in connecting vision and language, yet their proficiency in fundamental visual reasoning tasks remains limited. This limitation can be attributed to the fact that MLLMs learn visual understanding primarily from textual descriptions, which constitute a subjective and inherently incomplete supervisory signal. Furthermore, the modest scale of multimodal instruction tuning compared to massive text-only pre-training leads MLLMs to overfit language priors while overlooking visual details. To address these issues, we introduce JARVIS, a JEPA-inspired framework for self-supervised visual enhancement in MLLMs. Specifically, we integrate the I-JEPA learning paradigm into the standard vision-language alignment pipeline of MLLMs training. Our approach leverages frozen vision foundation models as context and target encoders, while training the predictor, implemented as the early layers of an LLM, to learn structural and semantic regularities from images without relying exclusively on language supervision. Extensive experiments on standard MLLM benchmarks show that JARVIS consistently improves performance on vision-centric benchmarks across different LLM families, without degrading multimodal reasoning abilities. Our source code is publicly available at: https://github.com/aimagelab/JARVIS.

Davide Caffagni, Sara Sarto, Marcella Cornia, Lorenzo Baraldi, Pier Luigi Dovesi, Shaghayegh Roohi, Mark Granroth-Wilding, Rita Cucchiara• 2025

Related benchmarks

TaskDatasetResultRank
Optical Character RecognitionOCRBench--
83
Visual ReasoningBLINK
Accuracy49.6
50
Multimodal Visual PerceptionMMVP
Accuracy38
44
Real-world Question AnsweringRealworldQA
Accuracy56.2
27
Visual ReasoningVision-Centric Benchmarks
BLINK Score50
20
2D Computer Vision BenchmarkingCVBench2D
Accuracy63.9
13
General Multimodal UnderstandingGeneral Benchmarks
Average Score74
12
Knowledge-based Visual Question AnsweringKnowledge Benchmarks
Average Score48.2
12
3D Computer Vision BenchmarkingCVBench3D
Accuracy73
8
General Vision-Language UnderstandingGeneral
Avg Score72.4
8
Showing 10 of 13 rows

Other info

Follow for update