Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

RayNet: Learning Volumetric 3D Reconstruction with Ray Potentials

About

In this paper, we consider the problem of reconstructing a dense 3D model using images captured from different views. Recent methods based on convolutional neural networks (CNN) allow learning the entire task from data. However, they do not incorporate the physics of image formation such as perspective geometry and occlusion. Instead, classical approaches based on Markov Random Fields (MRF) with ray-potentials explicitly model these physical processes, but they cannot cope with large surface appearance variations across different viewpoints. In this paper, we propose RayNet, which combines the strengths of both frameworks. RayNet integrates a CNN that learns view-invariant feature representations with an MRF that explicitly encodes the physics of perspective projection and occlusion. We train RayNet end-to-end using empirical risk minimization. We thoroughly evaluate our approach on challenging real-world datasets and demonstrate its benefits over a piece-wise trained baseline, hand-crafted models as well as other learning-based approaches.

Despoina Paschalidou, Ali Osman Ulusoy, Carolin Schmitt, Luc van Gool, Andreas Geiger• 2019

Related benchmarks

TaskDatasetResultRank
Multi-view 3D ReconstructionShapeNetr2n2 (test)
mIoU68.6
160
Multi-view 3D ReconstructionModelNet40 (test)
mIoU52.3
112
Multi-view 3D ReconstructionShapeNet r2n2 13 categories (test)
mIoU68.5
80
Multi-view 3D ReconstructionShapeNet ism (test)
mIoU51.1
72
Silhouette PredictionBlobby dataset (test)
mIoU0.924
32
Silhouette PredictionBlobby
mIoU92.7
32
Single-view 3D ReconstructionShapeNet-R2N2 (test)
mIoU63.2
22
Multi-view 3D ReconstructionShapeNet_R2N2 13 categories (test)
Reconstruction Score (1 view)0.632
8
Showing 8 of 8 rows

Other info

Follow for update