Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Dynamics-Regulated Kinematic Policy for Egocentric Pose Estimation

About

We propose a method for object-aware 3D egocentric pose estimation that tightly integrates kinematics modeling, dynamics modeling, and scene object information. Unlike prior kinematics or dynamics-based approaches where the two components are used disjointly, we synergize the two approaches via dynamics-regulated training. At each timestep, a kinematic model is used to provide a target pose using video evidence and simulation state. Then, a prelearned dynamics model attempts to mimic the kinematic pose in a physics simulator. By comparing the pose instructed by the kinematic model against the pose generated by the dynamics model, we can use their misalignment to further improve the kinematic model. By factoring in the 6DoF pose of objects (e.g., chairs, boxes) in the scene, we demonstrate for the first time, the ability to estimate physically-plausible 3D human-object interactions using a single wearable camera. We evaluate our egocentric pose estimation method in both controlled laboratory settings and real-world scenarios.

Zhengyi Luo, Ryo Hachiuma, Ye Yuan, Kris Kitani• 2021

Related benchmarks

TaskDatasetResultRank
Egocentric Pose EstimationMoCap (test)
Eroot0.176
4
Egocentric Pose EstimationReal-world dataset (test)
Ecam0.475
4
Pose EstimationADT 25k frames (test)
MPJPE (Eg)93.8
4
Motion ImitationAMASS (full)
Sinter95
3
Motion ImitationH36M (train)
Interaction Error89.3
3
3D Human Pose EstimationSynthetic 458k frames (test)
Empjpe63.5
3
3D Human Pose EstimationReal-world captures 40k frames
Epa-mpjpe83
3
Showing 7 of 7 rows

Other info

Code

Follow for update