Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Localizing Moments in Long Video Via Multimodal Guidance

About

The recent introduction of the large-scale, long-form MAD and Ego4D datasets has enabled researchers to investigate the performance of current state-of-the-art methods for video grounding in the long-form setup, with interesting findings: current grounding methods alone fail at tackling this challenging task and setup due to their inability to process long video sequences. In this paper, we propose a method for improving the performance of natural language grounding in long videos by identifying and pruning out non-describable windows. We design a guided grounding framework consisting of a Guidance Model and a base grounding model. The Guidance Model emphasizes describable windows, while the base grounding model analyzes short temporal windows to determine which segments accurately match a given language query. We offer two designs for the Guidance Model: Query-Agnostic and Query-Dependent, which balance efficiency and accuracy. Experiments demonstrate that our proposed method outperforms state-of-the-art models by 4.1% in MAD and 4.52% in Ego4D (NLQ), respectively. Code, data and MAD's audio features necessary to reproduce our experiments are available at: https://github.com/waybarrios/guidance-based-video-grounding.

Wayner Barrios, Mattia Soldan, Alberto Mario Ceballos-Arroyo, Fabian Caba Heilbron, Bernard Ghanem• 2023

Related benchmarks

TaskDatasetResultRank
Video GroundingMAD (test)
Recall@1 (IoU=0.1)9.3
35
Video GroundingMAD 1.0 (test)
R@1 (IoU=0.1)9.3
17
Long Video Moment RetrievalMAD (test)
Recall@1 (Tol 0.1)9.3
10
Video GroundingEgo4D (val)
mR@111.9
4
Showing 4 of 4 rows

Other info

Code

Follow for update