Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Single-Stage Diffusion NeRF: A Unified Approach to 3D Generation and Reconstruction

About

3D-aware image synthesis encompasses a variety of tasks, such as scene generation and novel view synthesis from images. Despite numerous task-specific methods, developing a comprehensive model remains challenging. In this paper, we present SSDNeRF, a unified approach that employs an expressive diffusion model to learn a generalizable prior of neural radiance fields (NeRF) from multi-view images of diverse objects. Previous studies have used two-stage approaches that rely on pretrained NeRFs as real data to train diffusion models. In contrast, we propose a new single-stage training paradigm with an end-to-end objective that jointly optimizes a NeRF auto-decoder and a latent diffusion model, enabling simultaneous 3D reconstruction and prior learning, even from sparsely available views. At test time, we can directly sample the diffusion prior for unconditional generation, or combine it with arbitrary observations of unseen objects for NeRF reconstruction. SSDNeRF demonstrates robust results comparable to or better than leading task-specific methods in unconditional generation and single/sparse-view 3D reconstruction.

Hansheng Chen, Jiatao Gu, Anpei Chen, Wei Tian, Zhuowen Tu, Lingjie Liu, Hao Su• 2023

Related benchmarks

TaskDatasetResultRank
Novel View SynthesisShapeNet cars category
PSNR23.52
20
Novel Facial Expression SynthesisFaceScape (test)
Chamfer Distance0.0154
9
Single-view 3D ReconstructionShapeNet chairs
Chamfer Distance (CD)8.98
8
Single-view 3D ReconstructionShapeNet Airplanes
CD10.4
7
Single-class 3D GenerationAmazon Berkeley Objects Tables (test)
FID14.27
5
Novel View SynthesisFaceScape 98 (test)
LPIPS0.2225
5
Novel View SynthesisShapeNet-SRN Chairs
PSNR24.35
4
Single-class 3D GenerationShapeNet SRN Cars (test)
FID11.08
4
Showing 8 of 8 rows

Other info

Follow for update