Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

In the Eye of the Beholder: Gaze and Actions in First Person Video

About

We address the task of jointly determining what a person is doing and where they are looking based on the analysis of video captured by a headworn camera. To facilitate our research, we first introduce the EGTEA Gaze+ dataset. Our dataset comes with videos, gaze tracking data, hand masks and action annotations, thereby providing the most comprehensive benchmark for First Person Vision (FPV). Moving beyond the dataset, we propose a novel deep model for joint gaze estimation and action recognition in FPV. Our method describes the participant's gaze as a probabilistic variable and models its distribution using stochastic units in a deep network. We further sample from these stochastic units, generating an attention map to guide the aggregation of visual features for action recognition. Our method is evaluated on our EGTEA Gaze+ dataset and achieves a performance level that exceeds the state-of-the-art by a significant margin. More importantly, we demonstrate that our model can be applied to larger scale FPV dataset---EPIC-Kitchens even without using gaze, offering new state-of-the-art results on FPV action recognition.

Yin Li, Miao Liu, James M. Rehg• 2020

Related benchmarks

TaskDatasetResultRank
Proficiency estimationEgo-Exo4D
Bouldering Proficiency Score21.4
16
Skill AssessmentMulti-Sense Badminton (MSB)
MSB Score30.6
11
Visual Attention PredictionAria Everyday Activities (AEA) unseen (test)
F1 Score44
9
Egocentric visual attention predictionEgo4D (test)
F1 Score0.363
9
Egocentric visual attention predictionAria Everyday Activities (AEA) (test)
F1 Score56.8
9
Eye Gaze PredictionEgoCampus (test)
AUC-J0.976
4
Showing 6 of 6 rows

Other info

Follow for update