Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Towards Universal Representation for Unseen Action Recognition

About

Unseen Action Recognition (UAR) aims to recognise novel action categories without training examples. While previous methods focus on inner-dataset seen/unseen splits, this paper proposes a pipeline using a large-scale training source to achieve a Universal Representation (UR) that can generalise to a more realistic Cross-Dataset UAR (CD-UAR) scenario. We first address UAR as a Generalised Multiple-Instance Learning (GMIL) problem and discover 'building-blocks' from the large-scale ActivityNet dataset using distribution kernels. Essential visual and semantic components are preserved in a shared space to achieve the UR that can efficiently generalise to new datasets. Predicted UR exemplars can be improved by a simple semantic adaptation, and then an unseen action can be directly recognised using UR during the test. Without further training, extensive experiments manifest significant improvements over the UCF101 and HMDB51 benchmarks.

Yi Zhu, Yang Long, Yu Guan, Shawn Newsam, Ling Shao• 2018

Related benchmarks

TaskDatasetResultRank
Action RecognitionUCF101
Accuracy42.5
365
Action RecognitionUCF101 (test)
Accuracy17.5
307
Action RecognitionHMDB51 (test)
Accuracy0.244
249
Action RecognitionHMDB51
Top-1 Acc24.4
225
Action RecognitionHMDB51
3-Fold Accuracy51.8
191
Action RecognitionUCF101 (3 splits)
Accuracy34.2
155
Action RecognitionUCF-101
Top-1 Acc17.5
147
Zero-shot Action RecognitionUCF101 (test)
Accuracy17.5
33
Action RecognitionHMDB51
Top-1 Acc24.4
30
Zero-shot Action RecognitionHMDB51 (test)
Accuracy24.4
25
Showing 10 of 21 rows

Other info

Follow for update