Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

ODHSR: Online Dense 3D Reconstruction of Humans and Scenes from Monocular Videos

About

Creating a photorealistic scene and human reconstruction from a single monocular in-the-wild video figures prominently in the perception of a human-centric 3D world. Recent neural rendering advances have enabled holistic human-scene reconstruction but require pre-calibrated camera and human poses, and days of training time. In this work, we introduce a novel unified framework that simultaneously performs camera tracking, human pose estimation and human-scene reconstruction in an online fashion. 3D Gaussian Splatting is utilized to learn Gaussian primitives for humans and scenes efficiently, and reconstruction-based camera tracking and human pose estimation modules are designed to enable holistic understanding and effective disentanglement of pose and appearance. Specifically, we design a human deformation module to reconstruct the details and enhance generalizability to out-of-distribution poses faithfully. Aiming to learn the spatial correlation between human and scene accurately, we introduce occlusion-aware human silhouette rendering and monocular geometric priors, which further improve reconstruction quality. Experiments on the EMDB and NeuMan datasets demonstrate superior or on-par performance with existing methods in camera tracking, human pose estimation, novel view synthesis and runtime. Our project page is at https://eth-ait.github.io/ODHSR.

Zetong Zhang, Manuel Kaufmann, Lixin Xue, Jie Song, Martin R. Oswald• 2025

Related benchmarks

TaskDatasetResultRank
Novel View SynthesisEMDB 1.0 (test)
PSNR (Whole Images)28.95
17
Novel View SynthesisNeuMan Human-only regions
PSNR32.079
14
Novel View SynthesisNeuMan Whole images
PSNR27.784
5
Novel View SynthesisNeuMan full-image frames (test)
PSNR27.78
5
Camera TrackingEMDB
ATE RMSE (m)0.084
4
Human Pose EstimationEMDB
PA-MPJPE40.571
2
Showing 6 of 6 rows

Other info

Code

Follow for update