Thin-Slicing Network: A Deep Structured Model for Pose Estimation in Videos
About
Deep ConvNets have been shown to be effective for the task of human pose estimation from single images. However, several challenging issues arise in the video-based case such as self-occlusion, motion blur, and uncommon poses with few or no examples in training data sets. Temporal information can provide additional cues about the location of body joints and help to alleviate these issues. In this paper, we propose a deep structured model to estimate a sequence of human poses in unconstrained videos. This model can be efficiently trained in an end-to-end manner and is capable of representing appearance of body joints and their spatio-temporal relationships simultaneously. Domain knowledge about the human body is explicitly incorporated into the network providing effective priors to regularize the skeletal structure and to enforce temporal consistency. The proposed end-to-end architecture is evaluated on two widely used benchmarks (Penn Action dataset and JHMDB dataset) for video-based pose estimation. Our approach significantly outperforms the existing state-of-the-art methods.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Video Object Segmentation | DAVIS 2017 (val) | J mean55.1 | 1130 | |
| Human Pose Estimation | J-HMDB sub | Head Accuracy97.2 | 49 | |
| Pose Estimation | Penn Action Dataset (test) | Head98 | 19 | |
| Human Pose Estimation | Penn-Action | Head Acc98 | 16 | |
| Human Pose Tracking | JHMDB (val) | PCK@.168.7 | 15 | |
| Human Pose Tracking | JHMDB (split1) | PCK @ 0.168.7 | 11 | |
| Pose Keypoint Propagation | JHMDB split 1 (val) | PCK@0.168.7 | 10 | |
| Part Instance Propagation | Video Instance-level Parsing (VIP) 85 (test) | APr vol24.1 | 8 | |
| Semantic Propagation | Video Instance-level Parsing (VIP) 85 (test) | mIoU37.9 | 8 | |
| Human Pose Estimation | Sub-JHMDB (test) | Head Accuracy97.1 | 8 |