Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

With a Little Help from my Temporal Context: Multimodal Egocentric Action Recognition

About

In egocentric videos, actions occur in quick succession. We capitalise on the action's temporal context and propose a method that learns to attend to surrounding actions in order to improve recognition performance. To incorporate the temporal context, we propose a transformer-based multimodal model that ingests video and audio as input modalities, with an explicit language model providing action sequence context to enhance the predictions. We test our approach on EPIC-KITCHENS and EGTEA datasets reporting state-of-the-art performance. Our ablations showcase the advantage of utilising temporal context as well as incorporating audio input modality and language model to rescore predictions. Code and models at: https://github.com/ekazakos/MTCN.

Evangelos Kazakos, Jaesung Huh, Arsha Nagrani, Andrew Zisserman, Dima Damen• 2021

Related benchmarks

TaskDatasetResultRank
Action RecognitionEPIC-KITCHENS (val)
Verb Top-1 Acc70.7
36
Video ClassificationEpic Kitchens 100
Verb Accuracy70.7
8
Showing 2 of 2 rows

Other info

Follow for update