Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Learning Neural Volumetric Representations of Dynamic Humans in Minutes

About

This paper addresses the challenge of quickly reconstructing free-viewpoint videos of dynamic humans from sparse multi-view videos. Some recent works represent the dynamic human as a canonical neural radiance field (NeRF) and a motion field, which are learned from videos through differentiable rendering. But the per-scene optimization generally requires hours. Other generalizable NeRF models leverage learned prior from datasets and reduce the optimization time by only finetuning on new scenes at the cost of visual fidelity. In this paper, we propose a novel method for learning neural volumetric videos of dynamic humans from sparse view videos in minutes with competitive visual quality. Specifically, we define a novel part-based voxelized human representation to better distribute the representational power of the network to different human parts. Furthermore, we propose a novel 2D motion parameterization scheme to increase the convergence rate of deformation field learning. Experiments demonstrate that our model can be learned 100 times faster than prior per-scene optimization methods while being competitive in the rendering quality. Training our model on a $512 \times 512$ video with 100 frames typically takes about 5 minutes on a single RTX 3090 GPU. The code will be released on our project page: https://zju3dv.github.io/instant_nvr

Chen Geng, Sida Peng, Zhen Xu, Hujun Bao, Xiaowei Zhou• 2023

Related benchmarks

TaskDatasetResultRank
Novel View SynthesisZJU-MoCap (test)
SSIM0.971
43
Human Novel View SynthesisZJU-MoCap
PSNR31.01
31
Novel View SynthesisMonoCap (test)
PSNR32.61
17
Showing 3 of 3 rows

Other info

Code

Follow for update