Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

DeepHuman: 3D Human Reconstruction from a Single Image

About

We propose DeepHuman, an image-guided volume-to-volume translation CNN for 3D human reconstruction from a single RGB image. To reduce the ambiguities associated with the surface geometry reconstruction, even for the reconstruction of invisible areas, we propose and leverage a dense semantic representation generated from SMPL model as an additional input. One key feature of our network is that it fuses different scales of image features into the 3D space through volumetric feature transformation, which helps to recover accurate surface geometry. The visible surface details are further refined through a normal refinement network, which can be concatenated with the volume generation network using our proposed volumetric normal projection layer. We also contribute THuman, a 3D real-world human model dataset containing about 7000 models. The network is trained using training data generated from the dataset. Overall, due to the specific design of our network and the diversity in our dataset, our method enables 3D human model estimation given only a single image and outperforms state-of-the-art approaches.

Zerong Zheng, Tao Yu, Yixuan Wei, Qionghai Dai, Yebin Liu• 2019

Related benchmarks

TaskDatasetResultRank
3D human reconstructionBUFF (test)
P2S Distance3.375
23
3D human reconstructionBUFF
P2S Distance2.15
11
3D human reconstructionAXYZ Pose (test)
P2S Distance2.656
7
3D human reconstructionSynthetic (test)
P2S2.761
7
3D human surface reconstructionTWINDOM
P2S Distance (cm)2.35
6
3D Mesh ReconstructionDigital Wardrobe
Upper Component Error2.38
3
Multi-view consistency evaluationTHuman 2.1 (50 random cases)
WE5.38
1
Showing 7 of 7 rows

Other info

Follow for update