SinGRAF: Learning a 3D Generative Radiance Field for a Single Scene
About
Generative models have shown great promise in synthesizing photorealistic 3D objects, but they require large amounts of training data. We introduce SinGRAF, a 3D-aware generative model that is trained with a few input images of a single scene. Once trained, SinGRAF generates different realizations of this 3D scene that preserve the appearance of the input while varying scene layout. For this purpose, we build on recent progress in 3D GAN architectures and introduce a novel progressive-scale patch discrimination approach during training. With several experiments, we demonstrate that the results produced by SinGRAF outperform the closest related works in both quality and diversity by a large margin.
Minjung Son, Jeong Joon Park, Leonidas Guibas, Gordon Wetzstein• 2022
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| 3D Scene Generation | Replica office_3 | KID0.044 | 3 | |
| 3D Scene Generation | Replica hotel_0 | KID0.037 | 3 | |
| 3D Scene Generation | Replica apt.0 | KID0.037 | 3 | |
| 3D Scene Generation | Replica frl_apt.4 | KID0.037 | 3 | |
| 3D Scene Generation | Matterport3D office_0 | KID0.053 | 3 | |
| 3D Scene Generation | Matterport3D dynamic | KID0.033 | 3 | |
| 3D Scene Generation | Matterport3D castle | KID0.064 | 3 |
Showing 7 of 7 rows