CanoVerse: 3D Object Scalable Canonicalization and Dataset for Generation and Pose
About
3D learning systems implicitly assume that objects occupy a coherent reference frame. Nonetheless, in practice, every asset arrives with an arbitrary global rotation, and models are left to resolve directional ambiguity on their own. This persistent misalignment suppresses pose-consistent generation, and blocks the emergence of stable directional semantics. To address this issue, we construct \methodName{}, a massive canonical 3D dataset of 320K objects over 1,156 categories -- an order-of-magnitude increase over prior work. At this scale, directional semantics become statistically learnable: Canoverse improves 3D generation stability, enables precise cross-modal 3D shape retrieval, and unlocks zero-shot point-cloud orientation estimation even for out-of-distribution data. This is achieved by a new canonicalization framework that reduces alignment from minutes to seconds per object via compact hypothesis generation and lightweight human discrimination, transforming canonicalization from manual curation into a high-throughput data generation pipeline. The Canoverse dataset will be publicly released upon acceptance. Project page: https://github.com/123321456-gif/Canoverse
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Orientation Estimation | CanoVerse In-distribution (test) | Accuracy @ 10°55.31 | 6 | |
| 3D Orientation Estimation | OmniObject3D (out-of-distribution) | Acc @ 10°20.18 | 5 | |
| 3D object canonicalization | DREDS | Aeroplane IC0.129 | 5 | |
| Object Canonicalization | DREDS | Speed (Low Level)4 | 3 |