3D-aware Conditional Image Synthesis
About
We propose pix2pix3D, a 3D-aware conditional generative model for controllable photorealistic image synthesis. Given a 2D label map, such as a segmentation or edge map, our model learns to synthesize a corresponding image from different viewpoints. To enable explicit 3D user control, we extend conditional generative models with neural radiance fields. Given widely-available monocular images and label map pairs, our model learns to assign a label to every 3D point in addition to color and density, which enables it to render the image and pixel-aligned label map simultaneously. Finally, we build an interactive system that allows users to edit the label map from any viewpoint and generate outputs accordingly.
Kangle Deng, Gengshan Yang, Deva Ramanan, Jun-Yan Zhu• 2023
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Edge2car | ShapeNet Car (test) | FID8.31 | 7 | |
| Seg2face | CelebAMask-HQ (test) | FID11.13 | 7 | |
| Segmentation-to-Cat Image Generation | AFHQ cat 34 (test) | FID8.62 | 7 | |
| Semantic-to-Car Synthesis | Seg2Car ShapNet-car | FID9.35 | 2 |
Showing 4 of 4 rows