Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

GM-NeRF: Learning Generalizable Model-based Neural Radiance Fields from Multi-view Images

About

In this work, we focus on synthesizing high-fidelity novel view images for arbitrary human performers, given a set of sparse multi-view images. It is a challenging task due to the large variation among articulated body poses and heavy self-occlusions. To alleviate this, we introduce an effective generalizable framework Generalizable Model-based Neural Radiance Fields (GM-NeRF) to synthesize free-viewpoint images. Specifically, we propose a geometry-guided attention mechanism to register the appearance code from multi-view 2D images to a geometry proxy which can alleviate the misalignment between inaccurate geometry prior and pixel space. On top of that, we further conduct neural rendering and partial gradient backpropagation for efficient perceptual supervision and improvement of the perceptual quality of synthesis. To evaluate our method, we conduct experiments on synthesized datasets THuman2.0 and Multi-garment, and real-world datasets Genebody and ZJUMocap. The results demonstrate that our approach outperforms state-of-the-art methods in terms of novel view synthesis and geometric reconstruction.

Jianchuan Chen, Wentao Yi, Liqian Ma, Xu Jia, Huchuan Lu• 2023

Related benchmarks

TaskDatasetResultRank
Novel View SynthesisZJU-MoCap
PSNR28.45
23
Novel View SynthesisZJUMocap unseen identity
PSNR26.74
17
3D Geometry ReconstructionTHuman 2.0
Chamfer Distance0.5172
9
Novel Pose SynthesisZJU 512x512
PSNR27.63
9
Novel View SynthesisMulti-Garment (unseen identity)
PSNR30.18
5
Novel View SynthesisTHuman 2.0 (unseen identity)
PSNR28.88
5
Novel View SynthesisGeneBody (unseen identity)
PSNR23.9
5
3D Geometry ReconstructionMulti-Garment
Chamfer Distance0.3721
4
Showing 8 of 8 rows

Other info

Code

Follow for update