Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Estimating Body and Hand Motion in an Ego-sensed World

About

We present EgoAllo, a system for human motion estimation from a head-mounted device. Using only egocentric SLAM poses and images, EgoAllo guides sampling from a conditional diffusion model to estimate 3D body pose, height, and hand parameters that capture a device wearer's actions in the allocentric coordinate frame of the scene. To achieve this, our key insight is in representation: we propose spatial and temporal invariance criteria for improving model performance, from which we derive a head motion conditioning parameterization that improves estimation by up to 18%. We also show how the bodies estimated by our system can improve hand estimation: the resulting kinematic and temporal constraints can reduce world-frame errors in single-frame estimates by 40%. Project page: https://egoallo.github.io/

Brent Yi, Vickie Ye, Maya Zheng, Yunqi Li, Lea M\"uller, Georgios Pavlakos, Yi Ma, Jitendra Malik, Angjoo Kanazawa• 2024

Related benchmarks

TaskDatasetResultRank
Hand Pose EstimationEgoExo4D 1.0 (test)
PA-MPJPE (mm)14.38
13
Body estimationAMASS (test)
MPJPE (mm)119.7
8
Body estimationRICH
MPJPE176.2
8
Body estimationAria Digital Twins
MPJPE155.1
8
Egocentric Pose EstimationHMD Setting 90° FoV
MPJPE113.8
6
Egocentric Pose EstimationHMD Setting 180° FoV
MPJPE90.67
6
3D Articulated Pose EstimationDataset 2
MPJPE (cm)10
5
3D Articulated Pose EstimationDataset 3
MPJPE (cm)11.7
5
3D Articulated Pose EstimationDataset-1
MPJPE (cm)10.6
5
Egocentric Motion ReconstructionAMASS
MPJPE (mm)99.32
4
Showing 10 of 11 rows

Other info

Code

Follow for update