Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

AdaFuse: Adaptive Temporal Fusion Network for Efficient Action Recognition

About

Temporal modelling is the key for efficient video action recognition. While understanding temporal information can improve recognition accuracy for dynamic actions, removing temporal redundancy and reusing past features can significantly save computation leading to efficient action recognition. In this paper, we introduce an adaptive temporal fusion network, called AdaFuse, that dynamically fuses channels from current and past feature maps for strong temporal modelling. Specifically, the necessary information from the historical convolution feature maps is fused with current pruned feature maps with the goal of improving both recognition accuracy and efficiency. In addition, we use a skipping operation to further reduce the computation cost of action recognition. Extensive experiments on Something V1 & V2, Jester and Mini-Kinetics show that our approach can achieve about 40% computation savings with comparable accuracy to state-of-the-art methods. The project page can be found at https://mengyuest.github.io/AdaFuse/

Yue Meng, Rameswar Panda, Chung-Ching Lin, Prasanna Sattigeri, Leonid Karlinsky, Kate Saenko, Aude Oliva, Rogerio Feris• 2021

Related benchmarks

TaskDatasetResultRank
Action RecognitionSomething-Something v2
Top-1 Accuracy59.8
341
Action RecognitionSomething-Something V1
Top-1 Acc46.8
162
Video Action ClassificationSomething-Something v2
Top-1 Acc59.8
139
Video RecognitionFCVID (test)
mAP81.6
28
Video RecognitionSomething-Something V1
Accuracy46.8
27
Video RecognitionKinetics Mini
Top-1 Acc72.3
18
Video RecognitionMini-Kinetics (test)
Accuracy72.3
17
Showing 7 of 7 rows

Other info

Follow for update