Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

SEASON: Mitigating Temporal Hallucination in Video Large Language Models via Self-Diagnostic Contrastive Decoding

About

Video Large Language Models (VideoLLMs) have shown remarkable progress in video understanding. However, these models still struggle to effectively perceive and exploit rich temporal information in videos when responding to user queries. Therefore, they often generate descriptions of events that are temporal inconsistent or causally implausible, causing severe hallucination issues. While most prior studies have focused on spatial hallucinations (e.g. object mismatches), temporal reasoning in video understanding remains relatively underexplored. To address this issue, we propose Self-Diagnostic Contrastive Decoding (SEASON), a training-free method that adaptively enhances temporal and spatial faithfulness for each output token. It achieves this by dynamically diagnosing each token's hallucination tendency and applying adaptive contrastive decoding against its corresponding temporal and spatial negatives. Extensive experiments demonstrate that SEASON outperforms all existing training-free hallucination mitigation approaches on three hallucination examination benchmarks, while further improves VideoLLMs across four general video understanding benchmarks. The code will be released upon acceptance.

Chang-Hsun Wu, Kai-Po Chang, Yu-Yang Sheng, Hung-Kai Chung, Kuei-Chun Wang, Yu-Chiang Frank Wang• 2025

Related benchmarks

TaskDatasetResultRank
Video Hallucination EvaluationVideoHallucer
ORH60.5
25
Hallucination ExaminationVidHalluc, VideoHallucer, EventHallusion
VidHalluc Score78.7
17
Temporal UnderstandingTempCompass, TVBench
TempCompass Score0.737
17
Conventional Video UnderstandingVideoMMe, MVBench
VideoMMe Score53.4
17
Hallucination ExaminationVidHalluc
BQA78.08
15
Hallucination ExaminationEventHallusion
Average Score69.19
15
Showing 6 of 6 rows

Other info

Follow for update