Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Securing the Floor and Raising the Ceiling: A Merging-based Paradigm for Multi-modal Search Agents

About

Recent advances in Vision-Language Models (VLMs) have motivated the development of multi-modal search agents that can actively invoke external search tools and integrate retrieved evidence through multi-step reasoning. While promising, existing approaches typically rely on large-scale supervised trajectories or expensive reinforcement learning (RL), leading to high training cost, instability, and a severe cold-start problem for standard VLMs. We propose a training-free paradigm to empower VLMs with autonomous search capabilities via cross-modal model merging. By fusing a text-based search agent with a base VLM, we show that multi-modal search capabilities can be effectively composed without any additional multi-modal training data. To mitigate parameter interference during cross-modal integration, we introduce Optimal Brain Merging (OBM), a saliency-aware merging algorithm that identifies task-critical parameters based on their impact on model loss using only a small set of calibration samples. Extensive experiments on search-intensive benchmarks (e.g., InfoSeek, MMSearch) reveal that: (1) Model merging secures a reasonable performance floor as a zero-shot agent, with OBM achieving superior search rates; (2) OBM significantly raises the performance ceiling as a warm-start strategy, achieving faster convergence and higher peak accuracy than standard VLM initialization.

Zhixiang Wang, Jingxuan Xu, Dajun Chen, Yunfang Wu, Wei Jiang, Yong Li• 2026

Related benchmarks

TaskDatasetResultRank
Visual Question AnsweringLiveVQA
Accuracy11.38
108
Visual Question AnsweringFVQA (test)
Accuracy20.67
36
Multimodal SearchMMSearch--
19
Visual Question AnsweringInfoSeek
Accuracy19.65
8
Showing 4 of 4 rows

Other info

Follow for update