Dynamic Spatial-temporal Hypergraph Convolutional Network for Skeleton-based Action Recognition
About
Skeleton-based action recognition relies on the extraction of spatial-temporal topological information. Hypergraphs can establish prior unnatural dependencies for the skeleton. However, the existing methods only focus on the construction of spatial topology and ignore the time-point dependence. This paper proposes a dynamic spatial-temporal hypergraph convolutional network (DST-HCN) to capture spatial-temporal information for skeleton-based action recognition. DST-HCN introduces a time-point hypergraph (TPH) to learn relationships at time points. With multiple spatial static hypergraphs and dynamic TPH, our network can learn more complete spatial-temporal features. In addition, we use the high-order information fusion module (HIF) to fuse spatial-temporal information synchronously. Extensive experiments on NTU RGB+D, NTU RGB+D 120, and NW-UCLA datasets show that our model achieves state-of-the-art, especially compared with hypergraph methods.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Action Recognition | NTU-60 (xsub) | Accuracy92.3 | 223 | |
| Action Recognition | NTU-120 (cross-subject (xsub)) | Accuracy88.8 | 211 | |
| Action Recognition | NTU 120 (Cross-Setup) | Accuracy90.7 | 203 | |
| Action Recognition | NTU-60 (xview) | Accuracy96.8 | 117 |