Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

LT3SD: Latent Trees for 3D Scene Diffusion

About

We present LT3SD, a novel latent diffusion model for large-scale 3D scene generation. Recent advances in diffusion models have shown impressive results in 3D object generation, but are limited in spatial extent and quality when extended to 3D scenes. To generate complex and diverse 3D scene structures, we introduce a latent tree representation to effectively encode both lower-frequency geometry and higher-frequency detail in a coarse-to-fine hierarchy. We can then learn a generative diffusion process in this latent 3D scene space, modeling the latent components of a scene at each resolution level. To synthesize large-scale scenes with varying sizes, we train our diffusion model on scene patches and synthesize arbitrary-sized output 3D scenes through shared diffusion generation across multiple scene patches. Through extensive experiments, we demonstrate the efficacy and benefits of LT3SD for large-scale, high-quality unconditional 3D scene generation and for probabilistic completion for partial scene observations.

Quan Meng, Lei Li, Matthias Nie{\ss}ner, Angela Dai• 2024

Related benchmarks

TaskDatasetResultRank
Unconditional 3D Scene Generation3D-FRONT
COV (CD)53.1
7
3D Scene Generation3D Scene Generation
Geometric Quality5.22
5
Outdoor 3D Scene GenerationWaymo Open Dataset Unconditional
Coverage (CD)20
4
Outdoor Scene GenerationWaymo Open Dataset User Study
Rank3
4
Indoor 3D Scene Generation3D-FRONT
Coverage (CD)23.4
4
Showing 5 of 5 rows

Other info

Code

Follow for update