Point-Supervised Skeleton-Based Human Action Segmentation
About
Skeleton-based temporal action segmentation is a fundamental yet challenging task, playing a crucial role in enabling intelligent systems to perceive and respond to human activities. While fully-supervised methods achieve satisfactory performance, they require costly frame-level annotations and are sensitive to ambiguous action boundaries. To address these issues, we introduce a point-supervised framework for skeleton-based action segmentation, where only a single frame per action segment is labeled. We leverage multimodal skeleton data, including joint, bone, and motion information, encoded via a pretrained unified model to extract rich feature representations. To generate reliable pseudo-labels, we propose a novel prototype similarity method and integrate it with two existing methods: energy function and constrained K-Medoids clustering. Multimodal pseudo-label integration is proposed to enhance the reliability of the pseudo-label and guide the model training. We establish new benchmarks on PKU-MMD (X-Sub and X-View), MCFS-22, and MCFS-130, and implement baselines for point-supervised skeleton-based human action segmentation. Extensive experiments show that our method achieves competitive performance, even surpassing some fully-supervised methods while significantly reducing annotation effort.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Skeleton-based Temporal Action Segmentation | PKU-MMD (X-sub) | Accuracy61.6 | 35 | |
| Temporal action segmentation | MCFS-130 | Accuracy59.1 | 29 | |
| Skeleton-based Temporal Action Segmentation | PKU-MMD (X-view) | Accuracy67.1 | 21 | |
| Temporal action segmentation | MCFS 22 | Accuracy69.1 | 17 |