Learning a Predictable and Generative Vector Representation for Objects
About
What is a good vector representation of an object? We believe that it should be generative in 3D, in the sense that it can produce new 3D objects; as well as be predictable from 2D, in the sense that it can be perceived from 2D images. We propose a novel architecture, called the TL-embedding network, to learn an embedding space with these properties. The network consists of two components: (a) an autoencoder that ensures the representation is generative; and (b) a convolutional network that ensures the representation is predictable. This enables tackling a number of tasks including voxel prediction from 2D images and 3D model retrieval. Extensive experimental analysis demonstrates the usefulness and versatility of this embedding.
Rohit Girdhar, David F. Fouhey, Mikel Rodriguez, Abhinav Gupta• 2016
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| 3D Shape Classification | ModelNet40 (test) | Accuracy74.4 | 227 | |
| Object Classification | ModelNet40 (test) | Accuracy74.4 | 180 | |
| Classification | ModelNet40 (test) | Accuracy74.4 | 99 | |
| 3D shape recognition | ModelNet10 (test) | Accuracy74.4 | 64 | |
| Unsupervised Representation Learning | ModelNet40 (test) | Accuracy74.4 | 13 | |
| Single-view 3D Reconstruction | IKEA dataset (test) | Chair AP32.9 | 5 | |
| Voxel Prediction | IKEA | Bed Accuracy56.3 | 5 |
Showing 7 of 7 rows