Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

ChatVTG: Video Temporal Grounding via Chat with Video Dialogue Large Language Models

About

Video Temporal Grounding (VTG) aims to ground specific segments within an untrimmed video corresponding to the given natural language query. Existing VTG methods largely depend on supervised learning and extensive annotated data, which is labor-intensive and prone to human biases. To address these challenges, we present ChatVTG, a novel approach that utilizes Video Dialogue Large Language Models (LLMs) for zero-shot video temporal grounding. Our ChatVTG leverages Video Dialogue LLMs to generate multi-granularity segment captions and matches these captions with the given query for coarse temporal grounding, circumventing the need for paired annotation data. Furthermore, to obtain more precise temporal grounding results, we employ moment refinement for fine-grained caption proposals. Extensive experiments on three mainstream VTG datasets, including Charades-STA, ActivityNet-Captions, and TACoS, demonstrate the effectiveness of ChatVTG. Our ChatVTG surpasses the performance of current zero-shot methods.

Mengxue Qu, Xiaodong Chen, Wu Liu, Alicia Li, Yao Zhao• 2024

Related benchmarks

TaskDatasetResultRank
Moment RetrievalCharades-STA (test)
R@0.533
172
Temporal Video GroundingCharades-STA (test)
Recall@IoU=0.533
117
Video GroundingCharades-STA
R@1 IoU=0.533
113
Natural Language Video LocalizationCharades-STA (test)
R@1 (IoU=0.5)33
61
Temporal GroundingActivityNet Captions
Recall@1 (IoU=0.5)22.5
45
Temporal GroundingCharades-STA
mIoU34.9
33
Video Event GroundingActivityNet
Recall@0.522.5
17
Natural Language Video LocalizationActivityNet Caption (test)
IoU @ 0.522.5
16
Temporal GroundingCharades-STA
mIoU34.8
13
Video Temporal GroundingActivityNet Captions
Recall @ IoU=0.522.5
12
Showing 10 of 11 rows

Other info

Follow for update