Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Two-Stream Convolutional Networks for Action Recognition in Videos

About

We investigate architectures of discriminatively trained deep Convolutional Networks (ConvNets) for action recognition in video. The challenge is to capture the complementary information on appearance from still frames and motion between frames. We also aim to generalise the best performing hand-crafted features within a data-driven learning framework. Our contribution is three-fold. First, we propose a two-stream ConvNet architecture which incorporates spatial and temporal networks. Second, we demonstrate that a ConvNet trained on multi-frame dense optical flow is able to achieve very good performance in spite of limited training data. Finally, we show that multi-task learning, applied to two different action classification datasets, can be used to increase the amount of training data and improve the performance on both. Our architecture is trained and evaluated on the standard video actions benchmarks of UCF-101 and HMDB-51, where it is competitive with the state of the art. It also exceeds by a large margin previous attempts to use deep nets for video classification.

Karen Simonyan, Andrew Zisserman• 2014

Related benchmarks

TaskDatasetResultRank
Action RecognitionNTU RGB+D (Cross-View)
Accuracy83.3
609
Action RecognitionNTU RGB+D (Cross-subject)
Accuracy74.4
474
Action RecognitionUCF101
Accuracy88
365
Action RecognitionUCF101 (mean of 3 splits)
Accuracy91.7
357
Action RecognitionUCF101 (test)
Accuracy92.5
307
Action RecognitionHMDB51 (test)
Accuracy0.624
249
Action RecognitionKinetics 400 (test)
Top-1 Accuracy65.6
245
Action RecognitionHMDB51
Top-1 Acc59.4
225
Action RecognitionHMDB-51 (average of three splits)
Top-1 Acc59.4
204
Action RecognitionHMDB51
3-Fold Accuracy59.4
191
Showing 10 of 79 rows
...

Other info

Follow for update