Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

MuRF: Multi-Baseline Radiance Fields

About

We present Multi-Baseline Radiance Fields (MuRF), a general feed-forward approach to solving sparse view synthesis under multiple different baseline settings (small and large baselines, and different number of input views). To render a target novel view, we discretize the 3D space into planes parallel to the target image plane, and accordingly construct a target view frustum volume. Such a target volume representation is spatially aligned with the target view, which effectively aggregates relevant information from the input views for high-quality rendering. It also facilitates subsequent radiance field regression with a convolutional network thanks to its axis-aligned nature. The 3D context modeled by the convolutional network enables our method to synthesis sharper scene structures than prior works. Our MuRF achieves state-of-the-art performance across multiple different baseline settings and diverse scenarios ranging from simple objects (DTU) to complex indoor and outdoor scenes (RealEstate10K and LLFF). We also show promising zero-shot generalization abilities on the Mip-NeRF 360 dataset, demonstrating the general applicability of MuRF.

Haofei Xu, Anpei Chen, Yuedong Chen, Christos Sakaridis, Yulun Zhang, Marc Pollefeys, Andreas Geiger, Fisher Yu• 2023

Related benchmarks

TaskDatasetResultRank
Novel View SynthesisMip-NeRF 360 (test)
PSNR23.98
166
Novel View SynthesisLLFF
PSNR26.49
124
Novel View SynthesisRealEstate10K
PSNR26.1
116
Novel View SynthesisDTU
PSNR21.31
100
Novel View SynthesisDTU (test)
PSNR28.76
82
Novel View SynthesisACID
PSNR28.09
51
Novel View SynthesisDTU 1 (test)
PSNR28.76
22
Novel View SynthesisDTU 9-view (test)
Full-image PSNR25.28
21
Novel View SynthesisReal Forward-facing 640 x 960 (test)
PSNR23.7
21
Novel View SynthesisNeRF Synthetic 800 x 800 (test)
PSNR24.37
21
Showing 10 of 29 rows

Other info

Code

Follow for update