Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

DenseRaC: Joint 3D Pose and Shape Estimation by Dense Render-and-Compare

About

We present DenseRaC, a novel end-to-end framework for jointly estimating 3D human pose and body shape from a monocular RGB image. Our two-step framework takes the body pixel-to-surface correspondence map (i.e., IUV map) as proxy representation and then performs estimation of parameterized human pose and shape. Specifically, given an estimated IUV map, we develop a deep neural network optimizing 3D body reconstruction losses and further integrating a render-and-compare scheme to minimize differences between the input and the rendered output, i.e., dense body landmarks, body part masks, and adversarial priors. To boost learning, we further construct a large-scale synthetic dataset (MOCA) utilizing web-crawled Mocap sequences, 3D scans and animations. The generated data covers diversified camera views, human actions and body shapes, and is paired with full ground truth. Our model jointly learns to represent the 3D human body from hybrid datasets, mitigating the problem of unpaired training data. Our experiments show that DenseRaC obtains superior performance against state of the art on public benchmarks of various humanrelated tasks.

Yuanlu Xu, Song-Chun Zhu, Tony Tung• 2019

Related benchmarks

TaskDatasetResultRank
3D Human Pose EstimationMPI-INF-3DHP (test)
PCK89
559
3D Human Pose EstimationHuman3.6M (test)--
547
3D Human Pose Estimation3DPW (test)
PA-MPJPE76.8
505
3D Human Pose EstimationHuman3.6M--
160
3D Human Pose EstimationHuman3.6M Protocol #2 (test)--
140
Human Mesh ReconstructionHuman3.6M
PA-MPJPE48
50
3D Human Pose and Shape EstimationHuman3.6M
PA-MPJPE48
36
Showing 7 of 7 rows

Other info

Follow for update