A Continuous-Time Consistency Model for 3D Point Cloud Generation
About
Fast and accurate 3D shape generation from point clouds is essential for applications in robotics, AR/VR, and digital content creation. We introduce ConTiCoM-3D, a continuous-time consistency model that synthesizes 3D shapes directly in point space, without discretized diffusion steps, pre-trained teacher models, or latent-space encodings. The method integrates a TrigFlow-inspired continuous noise schedule with a Chamfer Distance-based geometric loss, enabling stable training on high-dimensional point sets while avoiding expensive Jacobian-vector products. This design supports efficient one- to two-step inference with high geometric fidelity. In contrast to previous approaches that rely on iterative denoising or latent decoders, ConTiCoM-3D employs a time-conditioned neural network operating entirely in continuous time, thereby achieving fast generation. Experiments on the ShapeNet benchmark show that ConTiCoM-3D matches or outperforms state-of-the-art diffusion and latent consistency models in both quality and efficiency, establishing it as a practical framework for scalable 3D shape generation.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Point cloud generation | ShapeNet Car | 1-NNA (CD)53.3 | 41 | |
| 3D Shape Generation | ShapeNet airplane | 1-NNA (CD)64.89 | 30 | |
| 3D Shape Generation | ShapeNet chair | 1-NN Accuracy (CD)54.3 | 18 |