Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

DeepFuse: An IMU-Aware Network for Real-Time 3D Human Pose Estimation from Multi-View Image

About

In this paper, we propose a two-stage fully 3D network, namely \textbf{DeepFuse}, to estimate human pose in 3D space by fusing body-worn Inertial Measurement Unit (IMU) data and multi-view images deeply. The first stage is designed for pure vision estimation. To preserve data primitiveness of multi-view inputs, the vision stage uses multi-channel volume as data representation and 3D soft-argmax as activation layer. The second one is the IMU refinement stage which introduces an IMU-bone layer to fuse the IMU and vision data earlier at data level. without requiring a given skeleton model a priori, we can achieve a mean joint error of $28.9$mm on TotalCapture dataset and $13.4$mm on Human3.6M dataset under protocol 1, improving the SOTA result by a large margin. Finally, we discuss the effectiveness of a fully 3D network for 3D pose estimation experimentally which may benefit future research.

Fuyang Huang, Ailing Zeng, Minhao Liu, Qiuxia Lai, Qiang Xu• 2019

Related benchmarks

TaskDatasetResultRank
3D Human Pose EstimationHuman3.6M (Protocol #1)
MPJPE (Avg.)13.4
440
3D Human Pose EstimationHuman3.6M (Protocol 2)
Average MPJPE37.5
315
3D Human Pose EstimationHuman3.6M (S9, S11)
Average Error (MPJPE Avg)37.5
94
3D Pose EstimationTotal Capture (test)
Mean MPJPE28.9
42
3D Human Pose EstimationTotalCapture
Mean Joint Error (mm)28.9
4
Showing 5 of 5 rows

Other info

Follow for update