Multi-view Supervision for Single-view Reconstruction via Differentiable Ray Consistency
About
We study the notion of consistency between a 3D shape and a 2D observation and propose a differentiable formulation which allows computing gradients of the 3D shape given an observation from an arbitrary view. We do so by reformulating view consistency using a differentiable ray consistency (DRC) term. We show that this formulation can be incorporated in a learning framework to leverage different types of multi-view observations e.g. foreground masks, depth, color images, semantics etc. as supervision for learning single-view 3D prediction. We present empirical analysis of our technique in a controlled setting. We also show that this approach allows us to improve over existing techniques for single-view reconstruction of objects from the PASCAL VOC dataset.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| 3D Object Reconstruction | ShapeNet (test) | Mean IoU0.562 | 80 | |
| Single-view Reconstruction | ShapeNet | pla57.1 | 20 | |
| 3D Shape Reconstruction | Pix3D chair | CD0.16 | 14 | |
| Shape Prediction | ShapeNet | Airplane Score8.35 | 8 | |
| 3D Reconstruction | Pascal3D+ Car (test) | mIoU67 | 6 | |
| Object Reconstruction | Pix3D | IoU26.5 | 6 | |
| 3D Reconstruction | PASCAL3D+ aeroplane (test) | mIoU42 | 5 |