Point-Level Temporal Action Localization: Bridging Fully-supervised Proposals to Weakly-supervised Losses
About
Point-Level temporal action localization (PTAL) aims to localize actions in untrimmed videos with only one timestamp annotation for each action instance. Existing methods adopt the frame-level prediction paradigm to learn from the sparse single-frame labels. However, such a framework inevitably suffers from a large solution space. This paper attempts to explore the proposal-based prediction paradigm for point-level annotations, which has the advantage of more constrained solution space and consistent predictions among neighboring frames. The point-level annotations are first used as the keypoint supervision to train a keypoint detector. At the location prediction stage, a simple but effective mapper module, which enables back-propagation of training errors, is then introduced to bridge the fully-supervised framework with weak supervision. To our best of knowledge, this is the first work to leverage the fully-supervised paradigm for the point-level setting. Experiments on THUMOS14, BEOID, and GTEA verify the effectiveness of our proposed method both quantitatively and qualitatively, and demonstrate that our method outperforms state-of-the-art methods.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Temporal Action Localization | THUMOS-14 (test) | mAP@0.360.1 | 308 | |
| Temporal Action Localization | BEOID (test) | mAP@0.163.2 | 26 | |
| Temporal Action Localization | GTEA (test) | mAP@0.159.7 | 25 | |
| Temporal Action Localization | GTEA | mAP@0.159.7 | 12 | |
| Temporal Action Localization | BEOID | mAP@0.163.2 | 12 | |
| Temporal Action Localization | Thumos14 | -- | 6 |