Point-Cloud Completion with Pretrained Text-to-image Diffusion Models
About
Point-cloud data collected in real-world applications are often incomplete. Data is typically missing due to objects being observed from partial viewpoints, which only capture a specific perspective or angle. Additionally, data can be incomplete due to occlusion and low-resolution sampling. Existing completion approaches rely on datasets of predefined objects to guide the completion of noisy and incomplete, point clouds. However, these approaches perform poorly when tested on Out-Of-Distribution (OOD) objects, that are poorly represented in the training dataset. Here we leverage recent advances in text-guided image generation, which lead to major breakthroughs in text-guided shape generation. We describe an approach called SDS-Complete that uses a pre-trained text-to-image diffusion model and leverages the text semantics of a given incomplete point cloud of an object, to obtain a complete surface representation. SDS-Complete can complete a variety of objects using test-time optimization without expensive collection of 3D information. We evaluate SDS Complete on incomplete scanned objects, captured by real-world depth sensors and LiDAR scanners. We find that it effectively reconstructs objects that are absent from common datasets, reducing Chamfer loss by 50% on average compared with current methods. Project page: https://sds-complete.github.io/
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Point Cloud Completion | Redwood | Chamfer Distance21 | 15 | |
| 3D Shape Completion | Redwood | CD2.74 | 10 | |
| 3D Shape Completion | Omni-Comp Random Crop | CD5.12 | 7 | |
| 3D Shape Completion | Omni-Comp Semantic Part | CD5.6 | 7 | |
| 3D Shape Completion | Omni-Comp Single Scan | CD5.42 | 7 | |
| Point Cloud Completion | Redwood 10 objects | Chamfer Loss (old chair)19.3 | 5 | |
| Surface completion | KITTI real object scans | Best Quality0.734 | 3 |