Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Weakly Supervised Action Selection Learning in Video

About

Localizing actions in video is a core task in computer vision. The weakly supervised temporal localization problem investigates whether this task can be adequately solved with only video-level labels, significantly reducing the amount of expensive and error-prone annotation that is required. A common approach is to train a frame-level classifier where frames with the highest class probability are selected to make a video-level prediction. Frame level activations are then used for localization. However, the absence of frame-level annotations cause the classifier to impart class bias on every frame. To address this, we propose the Action Selection Learning (ASL) approach to capture the general concept of action, a property we refer to as "actionness". Under ASL, the model is trained with a novel class-agnostic task to predict which frames will be selected by the classifier. Empirically, we show that ASL outperforms leading baselines on two popular benchmarks THUMOS-14 and ActivityNet-1.2, with 10.3% and 5.7% relative improvement respectively. We further analyze the properties of ASL and demonstrate the importance of actionness. Full code for this work is available here: https://github.com/layer6ai-labs/ASL.

Junwei Ma, Satya Krishna Gorti, Maksims Volkovs, Guangwei Yu• 2021

Related benchmarks

TaskDatasetResultRank
Temporal Action LocalizationTHUMOS14 (test)
AP @ IoU=0.531.1
319
Temporal Action LocalizationTHUMOS-14 (test)
mAP@0.351.8
308
Temporal Action LocalizationActivityNet 1.2 (val)
mAP@IoU 0.540.2
110
Temporal Action LocalizationActivityNet 1.2 (test)
mAP@0.540.2
36
Temporal Action DetectionFineAction
Avg mAP3.3
27
Temporal Action DetectionFineGym (new data split)
mAP@0.19.33
10
Showing 6 of 6 rows

Other info

Code

Follow for update