AR-Net: Adaptive Frame Resolution for Efficient Action Recognition
About
Action recognition is an open and challenging problem in computer vision. While current state-of-the-art models offer excellent recognition results, their computational expense limits their impact for many real-world applications. In this paper, we propose a novel approach, called AR-Net (Adaptive Resolution Network), that selects on-the-fly the optimal resolution for each frame conditioned on the input for efficient action recognition in long untrimmed videos. Specifically, given a video frame, a policy network is used to decide what input resolution should be used for processing by the action recognition model, with the goal of improving both accuracy and efficiency. We efficiently train the policy network jointly with the recognition model using standard back-propagation. Extensive experiments on several challenging action recognition benchmark datasets well demonstrate the efficacy of our proposed approach over state-of-the-art methods. The project page can be found at https://mengyuest.github.io/AR-Net
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Fine-grained Video Categorization | ActivityNet v1.3 (val) | mAP73.8 | 32 | |
| Video Recognition | FCVID (test) | mAP81.3 | 28 | |
| Video Recognition | Something-Something V1 | Accuracy18.9 | 27 | |
| Action Recognition | ActivityNet v1.3 (test) | mAP73.8 | 19 | |
| Video Recognition | Kinetics Mini | Top-1 Acc71.7 | 18 | |
| Video Recognition | Mini-Kinetics (test) | Accuracy71.7 | 17 | |
| Online action recognition | 50Salads (test) | Accuracy40.9 | 7 |