Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Mitigating Object Hallucinations in Large Vision-Language Models with Assembly of Global and Local Attention

About

Despite great success across various multimodal tasks, Large Vision-Language Models (LVLMs) often encounter object hallucinations with generated textual responses being inconsistent with the actual objects in images. We examine different LVLMs and pinpoint that one root cause of object hallucinations lies with deficient attention on discriminative image features. Specifically, LVLMs often predominantly attend to prompt-irrelevant global features instead of prompt-relevant local features, undermining their visual grounding capacity and leading to object hallucinations. We propose Assembly of Global and Local Attention (AGLA), a training-free and plug-and-play approach that mitigates hallucinations by assembling global features for response generation and local features for visual discrimination simultaneously. Specifically, we introduce an image-prompt matching scheme that captures prompt-relevant local features from images, leading to an augmented view of the input image where prompt-relevant content is highlighted while irrelevant distractions are suppressed. Hallucinations can thus be mitigated with a calibrated logit distribution that is from generative global features of the original image and discriminative local features of the augmented image. Extensive experiments show the superiority of AGLA in LVLM hallucination mitigation, demonstrating its wide applicability across both discriminative and generative tasks. Our code is available at https://github.com/Lackel/AGLA.

Wenbin An, Feng Tian, Sicong Leng, Jiahao Nie, Haonan Lin, QianYing Wang, Ping Chen, Xiaoqin Zhang, Shijian Lu• 2024

Related benchmarks

TaskDatasetResultRank
Object Hallucination EvaluationPOPE
Accuracy85.5
935
Multimodal EvaluationMME--
557
Multimodal Capability EvaluationMM-Vet
Score46.88
282
Object HallucinationPOPE (Random)
F1 Score88.97
200
Object HallucinationPOPE Adversarial
Accuracy86.87
196
Object HallucinationPOPE Popular
F1 Score87.75
188
Hallucination EvaluationCHAIR
CHAIR_s54.8
166
Object Hallucination in Open-ended CaptioningChair (test)
CHAIR_S54.8
50
Object Hallucination EvaluationCHAIR
CS Score43
49
Vision-Language UnderstandingMM-Vet
Total Score69.56
43
Showing 10 of 17 rows

Other info

Follow for update