Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Semi-Supervised Single-View 3D Reconstruction via Prototype Shape Priors

About

The performance of existing single-view 3D reconstruction methods heavily relies on large-scale 3D annotations. However, such annotations are tedious and expensive to collect. Semi-supervised learning serves as an alternative way to mitigate the need for manual labels, but remains unexplored in 3D reconstruction. Inspired by the recent success of semi-supervised image classification tasks, we propose SSP3D, a semi-supervised framework for 3D reconstruction. In particular, we introduce an attention-guided prototype shape prior module for guiding realistic object reconstruction. We further introduce a discriminator-guided module to incentivize better shape generation, as well as a regularizer to tolerate noisy training samples. On the ShapeNet benchmark, the proposed approach outperforms previous supervised methods by clear margins under various labeling ratios, (i.e., 1%, 5% , 10% and 20%). Moreover, our approach also performs well when transferring to real-world Pix3D datasets under labeling ratios of 10%. We also demonstrate our method could transfer to novel categories with few novel supervised data. Experiments on the popular ShapeNet dataset show that our method outperforms the zero-shot baseline by over 12% and we also perform rigorous ablations and analysis to validate our approach.

Zhen Xing, Hengduo Li, Zuxuan Wu, Yu-Gang Jiang• 2022

Related benchmarks

TaskDatasetResultRank
Single-view 3D Object ReconstructionPix3D 10% labeled data (S1)
Chair Accuracy23.97
5
Showing 1 of 1 rows

Other info

Follow for update