Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

CoFi-Dec: Hallucination-Resistant Decoding via Coarse-to-Fine Generative Feedback in Large Vision-Language Models

About

Large Vision-Language Models (LVLMs) have achieved impressive progress in multi-modal understanding and generation. However, they still tend to produce hallucinated content that is inconsistent with the visual input, which limits their reliability in real-world applications. We propose \textbf{CoFi-Dec}, a training-free decoding framework that mitigates hallucinations by integrating generative self-feedback with coarse-to-fine visual conditioning. Inspired by the human visual process from global scene perception to detailed inspection, CoFi-Dec first generates two intermediate textual responses conditioned on coarse- and fine-grained views of the original image. These responses are then transformed into synthetic images using a text-to-image model, forming multi-level visual hypotheses that enrich grounding cues. To unify the predictions from these multiple visual conditions, we introduce a Wasserstein-based fusion mechanism that aligns their predictive distributions into a geometrically consistent decoding trajectory. This principled fusion reconciles high-level semantic consistency with fine-grained visual grounding, leading to more robust and faithful outputs. Extensive experiments on six hallucination-focused benchmarks show that CoFi-Dec substantially reduces both entity-level and semantic-level hallucinations, outperforming existing decoding strategies. The framework is model-agnostic, requires no additional training, and can be seamlessly applied to a wide range of LVLMs. The implementation is available at https://github.com/AI-Researcher-Team/CoFi-Dec.

Zongsheng Cao, Yangfan He, Anran Liu, Jun Xie, Feng Chen, Zepeng Wang• 2025

Related benchmarks

TaskDatasetResultRank
Object Hallucination EvaluationMS-COCO (POPE Adversarial)
Accuracy84.36
80
Object Hallucination EvaluationMS-COCO POPE (Popular)
Accuracy87.67
76
Object Hallucination EvaluationMS-COCO POPE Random
Accuracy90.33
55
Multimodal ReasoningMMBench
Accuracy65.9
50
Object Hallucination EvaluationA-OKVQA POPE Popular
Accuracy87.71
36
Object Hallucination EvaluationA-OKVQA POPE Random
Accuracy88.94
36
Object Hallucination EvaluationPOPE GQA Popular
Accuracy83.54
30
Object Hallucination ProbingGQA POPE Random
Accuracy (GQA POPE)89.03
26
Hallucination EvaluationMME Hallucination
Existence Score190.3
18
Object Hallucination AssessmentA-OKVQA POPE (Adversarial)
Accuracy0.8126
18
Showing 10 of 13 rows

Other info

Follow for update