Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Identifying and Mitigating Position Bias of Multi-image Vision-Language Models

About

The evolution of Large Vision-Language Models (LVLMs) has progressed from single to multi-image reasoning. Despite this advancement, our findings indicate that LVLMs struggle to robustly utilize information across multiple images, with predictions significantly affected by the alteration of image positions. To further explore this issue, we introduce Position-wise Question Answering (PQA), a meticulously designed task to quantify reasoning capabilities at each position. Our analysis reveals a pronounced position bias in LVLMs: open-source models excel in reasoning with images positioned later but underperform with those in the middle or at the beginning, while proprietary models show improved comprehension for images at the beginning and end but struggle with those in the middle. Motivated by this, we propose SoFt Attention (SoFA), a simple, training-free approach that mitigates this bias by employing linear interpolation between inter-image causal attention and bidirectional counterparts. Experimental results demonstrate that SoFA reduces position bias and enhances the reasoning performance of existing LVLMs.

Xinyu Tian, Shu Zou, Zhaoyuan Yang, Jing Zhang• 2025

Related benchmarks

TaskDatasetResultRank
Video UnderstandingMVBench
Accuracy57.71
247
Multi-image ReasoningMIRB
Accuracy60.67
60
Visual ReasoningNLVR2
Accuracy90.26
49
Multimodal ReasoningBLINK
Accuracy55.92
11
Multimodal ReasoningMantis-Eval
Accuracy59.23
11
Multimodal ReasoningMuirBench
Accuracy57.14
11
Showing 6 of 6 rows

Other info

Code

Follow for update