Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

AdaVBoost: Mitigating Hallucinations in LVLMs via Token-Level Adaptive Visual Attention Boosting

About

Visual attention boosting has emerged as a promising direction for mitigating hallucinations in Large Vision-Language Models (LVLMs), where existing methods primarily focus on where to boost by applying a predefined scaling to the attention of method-specific visual tokens during autoregressive generation. In this paper, we identify a fundamental trade-off in these methods: a predefined scaling factor can be too weak at some generation steps, leaving hallucinations unresolved, yet too strong at others, leading to new hallucinations. Motivated by this finding, we propose AdaVBoost, a token-level visual attention boosting framework that adaptively determines how much attention to boost at each generation step. Specifically, we introduce Visual Grounding Entropy (VGE) to estimate hallucination risk, which leverages visual grounding as a complementary signal to capture evidence mismatches beyond entropy. Guided by VGE, AdaVBoost applies stronger visual attention boosting to high-risk tokens and weaker boosting to low-risk tokens, enabling token-level adaptive intervention at each generation step. Extensive experiments show that AdaVBoost significantly outperforms baseline methods across multiple LVLMs and hallucination benchmarks.

Jiacheng Zhang, Feng Liu, Chao Du, Tianyu Pang• 2026

Related benchmarks

TaskDatasetResultRank
Hallucination assessmentAMBER (test)
CHAIR6.5
38
Hallucination MitigationSHR
HSR22.1
15
Hallucination EvaluationPOPE MSCOCO, A-OKVQA, GQA average (Adversarial)
Accuracy84.09
15
Hallucination EvaluationPOPE MSCOCO, A-OKVQA, GQA average (Random)
Accuracy93.16
15
Hallucination EvaluationPOPE Popular MSCOCO, A-OKVQA, GQA average
Accuracy88.35
15
Object Hallucination EvaluationCHAIR (val)
CHAIRs Score46
15
Showing 6 of 6 rows

Other info

Follow for update