Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Dense Depth Priors for Neural Radiance Fields from Sparse Input Views

About

Neural radiance fields (NeRF) encode a scene into a neural representation that enables photo-realistic rendering of novel views. However, a successful reconstruction from RGB images requires a large number of input views taken under static conditions - typically up to a few hundred images for room-size scenes. Our method aims to synthesize novel views of whole rooms from an order of magnitude fewer images. To this end, we leverage dense depth priors in order to constrain the NeRF optimization. First, we take advantage of the sparse depth data that is freely available from the structure from motion (SfM) preprocessing step used to estimate camera poses. Second, we use depth completion to convert these sparse points into dense depth maps and uncertainty estimates, which are used to guide NeRF optimization. Our method enables data-efficient novel view synthesis on challenging indoor scenes, using as few as 18 images for an entire scene.

Barbara Roessle, Jonathan T. Barron, Ben Mildenhall, Pratul P. Srinivasan, Matthias Nie{\ss}ner• 2021

Related benchmarks

TaskDatasetResultRank
Novel View SynthesisRealEstate10K
PSNR27.79
173
Novel View SynthesisLLFF 3-view (test)
PSNR18.73
39
Novel View SynthesisRealEstate-10K 2-view
PSNR26.15
32
Novel View SynthesisNeRF-LLFF
LPIPS0.2851
30
Novel View SynthesisLLFF 4 input views (test)
LPIPS0.3042
20
Depth EstimationLLFF 3 views (test)
Depth SROCC0.7433
17
Depth EstimationLLFF 4 input views (test)
Depth SROCC0.8322
16
Novel View SynthesisLLFF 2 input views
LPIPS0.2851
16
Novel View SynthesisRealEstate-10K 2 views (test)
LPIPS0.129
15
Novel View SynthesisRealEstate-10K 3-view
PSNR25.92
14
Showing 10 of 35 rows

Other info

Follow for update