Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Whole-Body Conditioned Egocentric Video Prediction

About

We train models to Predict Ego-centric Video from human Actions (PEVA), given the past video and an action represented by the relative 3D body pose. By conditioning on kinematic pose trajectories, structured by the joint hierarchy of the body, our model learns to simulate how physical human actions shape the environment from a first-person point of view. We train an auto-regressive conditional diffusion transformer on Nymeria, a large-scale dataset of real-world egocentric video and body pose capture. We further design a hierarchical evaluation protocol with increasingly challenging tasks, enabling a comprehensive analysis of the model's embodied prediction and control abilities. Our work represents an initial attempt to tackle the challenges of modeling complex real-world environments and embodied agent behaviors with video prediction from the perspective of a human.

Yutong Bai, Danny Tran, Amir Bar, Yann LeCun, Trevor Darrell, Jitendra Malik• 2025

Related benchmarks

TaskDatasetResultRank
Open-loop trajectory predictionEgoDex (test)
Embedding L2 Error (At 4s)0.62
3
Showing 1 of 1 rows

Other info

Follow for update