VolRecon: Volume Rendering of Signed Ray Distance Functions for Generalizable Multi-View Reconstruction
About
The success of the Neural Radiance Fields (NeRF) in novel view synthesis has inspired researchers to propose neural implicit scene reconstruction. However, most existing neural implicit reconstruction methods optimize per-scene parameters and therefore lack generalizability to new scenes. We introduce VolRecon, a novel generalizable implicit reconstruction method with Signed Ray Distance Function (SRDF). To reconstruct the scene with fine details and little noise, VolRecon combines projection features aggregated from multi-view features, and volume features interpolated from a coarse global feature volume. Using a ray transformer, we compute SRDF values of sampled points on a ray and then render color and depth. On DTU dataset, VolRecon outperforms SparseNeuS by about 30% in sparse view reconstruction and achieves comparable accuracy as MVSNet in full view reconstruction. Furthermore, our approach exhibits good generalization performance on the large-scale ETH3D benchmark.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| 3D surface reconstruction | DTU (test) | Mean Chamfer Distance (CD)1.38 | 69 | |
| Surface Reconstruction | DTU sparse-view | CD (Scan 24)1.2 | 14 | |
| Surface Reconstruction | DTU sparse-view 1 | CD (Scan 21)3.05 | 13 | |
| Surface Reconstruction | DTU unfavorable sets (test) | CD (Scan 24)3.43 | 6 | |
| Depth Estimation | DTU (Favorable set) | Accuracy (<1mm)43.6 | 5 | |
| Depth Estimation | DTU (Unfavorable set) | Threshold Accuracy (<1mm)3.93 | 5 |