Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Embodied Scene-aware Human Pose Estimation

About

We propose embodied scene-aware human pose estimation where we estimate 3D poses based on a simulated agent's proprioception and scene awareness, along with external third-person observations. Unlike prior methods that often resort to multistage optimization, non-causal inference, and complex contact modeling to estimate human pose and human scene interactions, our method is one-stage, causal, and recovers global 3D human poses in a simulated environment. Since 2D third-person observations are coupled with the camera pose, we propose to disentangle the camera pose and use a multi-step projection gradient defined in the global coordinate frame as the movement cue for our embodied agent. Leveraging a physics simulation and prescanned scenes (e.g., 3D mesh), we simulate our agent in everyday environments (library, office, bedroom, etc.) and equip our agent with environmental sensors to intelligently navigate and interact with the geometries of the scene. Our method also relies only on 2D keypoints and can be trained on synthetic datasets derived from popular human motion databases. To evaluate, we use the popular H36M and PROX datasets and achieve high quality pose estimation on the challenging PROX dataset without ever using PROX motion sequences for training. Code and videos are available on the project page.

Zhengyi Luo, Shun Iwase, Ye Yuan, Kris Kitani• 2022

Related benchmarks

TaskDatasetResultRank
3D Human Pose EstimationH36M
MPJPE105.1
16
3D Human Pose EstimationChi3D
Invalid Rate7.7
14
3D Human Pose EstimationHi4D--
8
Human Pose EstimationPROX
ACD171.7
7
Showing 4 of 4 rows

Other info

Code

Follow for update