Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Object-Shot Enhanced Grounding Network for Egocentric Video

About

Egocentric video grounding is a crucial task for embodied intelligence applications, distinct from exocentric video moment localization. Existing methods primarily focus on the distributional differences between egocentric and exocentric videos but often neglect key characteristics of egocentric videos and the fine-grained information emphasized by question-type queries. To address these limitations, we propose OSGNet, an Object-Shot enhanced Grounding Network for egocentric video. Specifically, we extract object information from videos to enrich video representation, particularly for objects highlighted in the textual query but not directly captured in the video features. Additionally, we analyze the frequent shot movements inherent to egocentric videos, leveraging these features to extract the wearer's attention information, which enhances the model's ability to perform modality alignment. Experiments conducted on three datasets demonstrate that OSGNet achieves state-of-the-art performance, validating the effectiveness of our approach. Our code can be found at https://github.com/Yisen-Feng/OSGNet.

Yisen Feng, Haoyu Zhang, Meng Liu, Weili Guan, Liqiang Nie• 2025

Related benchmarks

TaskDatasetResultRank
Moment RetrievalTACOS (test)
Recall@1 (IoU=0.5)55.77
23
Natural Language QueriesEgo4D NLQ (val)
Recall@1 (IoU=0.3)21.97
23
Natural Language QueriesEgo4D-NLQ v1 (test)
R@1 (IoU=0.3)22.13
8
Natural Language QueriesEgo4D NLQ v2 (val)
R@1 (IoU=0.3)31.63
7
Natural Language QueriesEgo4D-NLQ v2 (test)
Recall@1 (IoU=0.3)27.6
7
Step GroundingEgo4D (test)
Recall@1 (IoU=0.3)38.83
7
Step GroundingEgo4D Goal-Step (val)
Recall@1 (IoU=0.3)42.61
4
Showing 7 of 7 rows

Other info

Code

Follow for update