Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Zooming without Zooming: Region-to-Image Distillation for Fine-Grained Multimodal Perception

About

Multimodal Large Language Models (MLLMs) excel at broad visual understanding but still struggle with fine-grained perception, where decisive evidence is small and easily overwhelmed by global context. Recent "Thinking-with-Images" methods alleviate this by iteratively zooming in and out regions of interest during inference, but incur high latency due to repeated tool calls and visual re-encoding. To address this, we propose Region-to-Image Distillation, which transforms zooming from an inference-time tool into a training-time primitive, thereby internalizing the benefits of agentic zooming into a single forward pass of an MLLM. In particular, we first zoom in to micro-cropped regions to let strong teacher models generate high-quality VQA data, and then distill this region-grounded supervision back to the full image. After training on such data, the smaller student model improves "single-glance" fine-grained perception without tool use. To rigorously evaluate this capability, we further present ZoomBench, a hybrid-annotated benchmark of 845 VQA data spanning six fine-grained perceptual dimensions, together with a dual-view protocol that quantifies the global--regional "zooming gap". Experiments show that our models achieve leading performance across multiple fine-grained perception benchmarks, and also improve general multimodal cognition on benchmarks such as visual reasoning and GUI agents. We further discuss when "Thinking-with-Images" is necessary versus when its gains can be distilled into a single forward pass. Our code is available at https://github.com/inclusionAI/Zooming-without-Zooming.

Lai Wei, Liangbo He, Jun Lan, Lingzhong Dong, Yutong Cai, Siyuan Li, Huijia Zhu, Weiqiang Wang, Linghe Kong, Yue Wang, Zhuosheng Zhang, Weiran Huang• 2026

Related benchmarks

TaskDatasetResultRank
Multimodal PerceptionVSTAR
Accuracy92.67
18
Multimodal PerceptionHR-8K
Accuracy82
18
Multimodal PerceptionMMStar
Accuracy73.13
16
Visual PerceptionVSTAR (test)
Accuracy92.7
15
Visual PerceptionHR-4K (test)
Accuracy84.4
15
Visual PerceptionHR-8K (test)
Accuracy82
15
Multimodal UnderstandingMME-RW-en (test)
Overall Score69.9
15
Multimodal PerceptionZoomBench
Accuracy58.11
13
Multimodal PerceptionMME-RW-cn
Accuracy70.59
13
Multimodal PerceptionHR-4K
Accuracy84.38
13
Showing 10 of 16 rows

Other info

GitHub

Follow for update