Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

SeViCES: Unifying Semantic-Visual Evidence Consensus for Long Video Understanding

About

Long video understanding remains challenging due to its complex, diverse, and temporally scattered content. Although video large language models (Video-LLMs) can process videos lasting tens of minutes, applying them to truly long sequences is computationally prohibitive and often leads to unfocused or inconsistent reasoning. A promising solution is to select only the most informative frames, yet existing approaches typically ignore temporal dependencies or rely on unimodal evidence, limiting their ability to provide complete and query-relevant context. We propose a Semantic-Visual Consensus Evidence Selection (SeViCES) framework for effective and reliable long video understanding. SeViCES is training-free and model-agnostic, and introduces two key components. The Semantic-Visual Consensus Frame Selection (SVCFS) module selects frames through (1) a temporal-aware semantic branch that leverages LLM reasoning over captions, and (2) a cluster-guided visual branch that aligns embeddings with semantic scores via mutual information. The Answer Consensus Refinement (ACR) module further resolves inconsistencies between semantic- and visual-based predictions by fusing evidence and constraining the answer space. Extensive experiments on long video understanding benchmarks show that SeViCES consistently outperforms state-of-the-art methods in both accuracy and robustness, demonstrating the importance of consensus-driven evidence selection for Video-LLMs.

Yuan Sheng, Yanbin Hao, Chenxu Li, Shuo Wang, Xiangnan He• 2025

Related benchmarks

TaskDatasetResultRank
Long Video UnderstandingLongVideoBench
Score63.9
248
Long Video UnderstandingMLVU
Score72.2
154
Long Video UnderstandingVideo-MME
Overall Score65.5
30
Extremely long-video understandingLVBench
Score46.7
25
Showing 4 of 4 rows

Other info

Follow for update