Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

3D-R2N2: A Unified Approach for Single and Multi-view 3D Object Reconstruction

About

Inspired by the recent success of methods that employ shape priors to achieve robust 3D reconstructions, we propose a novel recurrent neural network architecture that we call the 3D Recurrent Reconstruction Neural Network (3D-R2N2). The network learns a mapping from images of objects to their underlying 3D shapes from a large collection of synthetic data. Our network takes in one or more images of an object instance from arbitrary viewpoints and outputs a reconstruction of the object in the form of a 3D occupancy grid. Unlike most of the previous works, our network does not require any image annotations or object class labels for training or testing. Our extensive experimental analysis shows that our reconstruction framework i) outperforms the state-of-the-art methods for single view reconstruction, and ii) enables the 3D reconstruction of objects in situations when traditional SFM/SLAM methods fail (because of lack of texture and/or wide baseline).

Christopher B. Choy, Danfei Xu, JunYoung Gwak, Kevin Chen, Silvio Savarese• 2016

Related benchmarks

TaskDatasetResultRank
Multi-view 3D ReconstructionShapeNet (test)
IoU0.636
209
Multi-view 3D ReconstructionShapeNetr2n2 (test)
mIoU65.1
160
Multi-view 3D ReconstructionModelNet40 (test)
mIoU46.4
112
Multi-view 3D ReconstructionShapeNet
IoU0.636
110
Multi-view 3D ReconstructionShapeNet r2n2 13 categories (test)
mIoU65.4
80
3D Object ReconstructionShapeNet (test)
Mean IoU0.636
80
3D ReconstructionShapeNet (test)
EMD0.606
74
Multi-view 3D ReconstructionShapeNet ism (test)
mIoU47.2
72
Multi-view 3D object reconstructionThings3D
IoU33.4
32
Silhouette PredictionBlobby
mIoU86.5
32
Showing 10 of 35 rows

Other info

Follow for update