Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

FrugalNeRF: Fast Convergence for Extreme Few-shot Novel View Synthesis without Learned Priors

About

Neural Radiance Fields (NeRF) face significant challenges in extreme few-shot scenarios, primarily due to overfitting and long training times. Existing methods, such as FreeNeRF and SparseNeRF, use frequency regularization or pre-trained priors but struggle with complex scheduling and bias. We introduce FrugalNeRF, a novel few-shot NeRF framework that leverages weight-sharing voxels across multiple scales to efficiently represent scene details. Our key contribution is a cross-scale geometric adaptation scheme that selects pseudo ground truth depth based on reprojection errors across scales. This guides training without relying on externally learned priors, enabling full utilization of the training data. It can also integrate pre-trained priors, enhancing quality without slowing convergence. Experiments on LLFF, DTU, and RealEstate-10K show that FrugalNeRF outperforms other few-shot NeRF methods while significantly reducing training time, making it a practical solution for efficient and accurate 3D scene reconstruction.

Chin-Yang Lin, Chung-Ho Wu, Chang-Han Yeh, Shih-Han Yen, Cheng Sun, Yu-Lun Liu• 2024

Related benchmarks

TaskDatasetResultRank
Novel View SynthesisLLFF 3-view
PSNR19.92
130
Novel View SynthesisLLFF
PSNR19.1
130
Novel View SynthesisRealEstate-10K 2-view
PSNR30.12
32
Novel View SynthesisZJU-MoCap
PSNR22.1
31
Novel View SynthesisDNA Rendering dataset (test)
Memory (GB)20.2
18
Novel View SynthesisRealEstate-10K 3-view
PSNR31.04
14
Novel View SynthesisLLFF 2-view 50 (test)
PSNR18.26
12
Novel View SynthesisLLFF 3-view 50 (test)
PSNR19.87
12
Novel View SynthesisLLFF 4-view 50 (test)
PSNR20.89
12
Novel View SynthesisDTU 2-view processed by pixelNeRF
PSNR20.77
9
Showing 10 of 20 rows

Other info

Code

Follow for update