PARTFIELD: Learning 3D Feature Fields for Part Segmentation and Beyond
About
We propose PartField, a feedforward approach for learning part-based 3D features, which captures the general concept of parts and their hierarchy without relying on predefined templates or text-based names, and can be applied to open-world 3D shapes across various modalities. PartField requires only a 3D feedforward pass at inference time, significantly improving runtime and robustness compared to prior approaches. Our model is trained by distilling 2D and 3D part proposals from a mix of labeled datasets and image segmentations on large unsupervised datasets, via a contrastive learning formulation. It produces a continuous feature field which can be clustered to yield a hierarchical part decomposition. Comparisons show that PartField is up to 20% more accurate and often orders of magnitude faster than other recent class-agnostic part-segmentation methods. Beyond single-shape part decomposition, consistency in the learned field emerges across shapes, enabling tasks such as co-segmentation and correspondence, which we demonstrate in several applications of these general-purpose, hierarchical, and consistent 3D feature fields. Check our Webpage! https://research.nvidia.com/labs/toronto-ai/partfield-release/
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Part Segmentation | PartNet (test) | mIoU52.1 | 19 | |
| Part Segmentation | PartNet-E few-shot | mIoU (Bottle)75.9 | 11 | |
| Part Segmentation | Lightwheel (test) | gIoU0.053 | 10 | |
| Part Segmentation | PartNet-Mobility (test) | gIoU18.3 | 8 | |
| Part Segmentation | Find3D (test) | mIoU66.2 | 4 | |
| Part Segmentation | 3DCoMPaT (test) | mIoU37.1 | 4 | |
| Part Segmentation | Part-level Generation Results (Evaluation Set) | mIoU0.4167 | 4 |