Unsupervised learning of action classes with continuous temporal embedding
About
The task of temporally detecting and segmenting actions in untrimmed videos has seen an increased attention recently. One problem in this context arises from the need to define and label action boundaries to create annotations for training which is very time and cost intensive. To address this issue, we propose an unsupervised approach for learning action classes from untrimmed video sequences. To this end, we use a continuous temporal embedding of framewise features to benefit from the sequential nature of activities. Based on the latent space created by the embedding, we identify clusters of temporal segments across all videos that correspond to semantic meaningful action classes. The approach is evaluated on three challenging datasets, namely the Breakfast dataset, YouTube Instructions, and the 50Salads dataset. While previous works assumed that the videos contain the same high level activity, we furthermore show that the proposed approach can also be applied to a more general setting where the content of the videos is unknown.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Action Segmentation | Breakfast | MoF47.2 | 66 | |
| Action Segmentation | Breakfast (test) | MoF41.8 | 31 | |
| Action Segmentation | Breakfast 14 | MoF41.8 | 26 | |
| Temporal action segmentation | 50 Salads granularity (Eval) | MoF35.5 | 24 | |
| Action Segmentation | Breakfast Action dataset | MoF41.8 | 22 | |
| Action Segmentation | 50Salads mid granularity | MoF30.2 | 19 | |
| Action Segmentation | YouTube Instructions (test) | F1 Score (%)28.3 | 17 | |
| Action Segmentation | 50 Salads Mid | -- | 17 | |
| Action Segmentation | YouTube Instructions | F128.3 | 16 | |
| Unsupervised Activity Segmentation | 50 Salads eval granularity | MOF35.5 | 14 |