NeRFusion: Fusing Radiance Fields for Large-Scale Scene Reconstruction
About
While NeRF has shown great success for neural reconstruction and rendering, its limited MLP capacity and long per-scene optimization times make it challenging to model large-scale indoor scenes. In contrast, classical 3D reconstruction methods can handle large-scale scenes but do not produce realistic renderings. We propose NeRFusion, a method that combines the advantages of NeRF and TSDF-based fusion techniques to achieve efficient large-scale reconstruction and photo-realistic rendering. We process the input image sequence to predict per-frame local radiance fields via direct network inference. These are then fused using a novel recurrent neural network that incrementally reconstructs a global, sparse scene representation in real-time at 22 fps. This global volume can be further fine-tuned to boost rendering quality. We demonstrate that NeRFusion achieves state-of-the-art quality on both large-scale indoor and small-scale object scenes, with substantially faster reconstruction than NeRF and other recent methods.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Novel View Synthesis | DTU (test) | PSNR31.79 | 82 | |
| Novel View Synthesis | ScanNet | PSNR22.99 | 58 | |
| Novel View Synthesis | ScanNet (test) | PSNR22.99 | 25 | |
| Novel View Synthesis | ScanNet 11 (test) | PSNR26.49 | 16 | |
| Novel View Synthesis | ScanNet (novel view) | PSNR26.49 | 15 | |
| Novel View Synthesis | NeRF Synthetic 360° (test) | PSNR31.25 | 10 | |
| Geometry Reconstruction | DTU input views (test) | Absolute Error0.034 | 3 | |
| Geometry Reconstruction | DTU novel views (test) | Absolute Error0.036 | 3 |