Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Alignment-guided Temporal Attention for Video Action Recognition

About

Temporal modeling is crucial for various video learning tasks. Most recent approaches employ either factorized (2D+1D) or joint (3D) spatial-temporal operations to extract temporal contexts from the input frames. While the former is more efficient in computation, the latter often obtains better performance. In this paper, we attribute this to a dilemma between the sufficiency and the efficiency of interactions among various positions in different frames. These interactions affect the extraction of task-relevant information shared among frames. To resolve this issue, we prove that frame-by-frame alignments have the potential to increase the mutual information between frame representations, thereby including more task-relevant information to boost effectiveness. Then we propose Alignment-guided Temporal Attention (ATA) to extend 1-dimensional temporal attention with parameter-free patch-level alignments between neighboring frames. It can act as a general plug-in for image backbones to conduct the action recognition task without any model-specific design. Extensive experiments on multiple benchmarks demonstrate the superiority and generality of our module.

Yizhou Zhao, Zhenyang Li, Xun Guo, Yan Lu• 2022

Related benchmarks

TaskDatasetResultRank
Action RecognitionSomething-Something v2 (val)
Top-1 Accuracy67.1
535
Video ClassificationSomething-Something v2 (test)
Top-1 Acc0.671
169
Action RecognitionKinetics-400 full (val)
Top-1 Acc81.9
136
Video RecognitionKinetics-400
Top-1 Acc81.9
54
Showing 4 of 4 rows

Other info

Follow for update