SAM 3D: 3Dfy Anything in Images
About
We present SAM 3D, a generative model for visually grounded 3D object reconstruction, predicting geometry, texture, and layout from a single image. SAM 3D excels in natural images, where occlusion and scene clutter are common and visual recognition cues from context play a larger role. We achieve this with a human- and model-in-the-loop pipeline for annotating object shape, texture, and pose, providing visually grounded 3D reconstruction data at unprecedented scale. We learn from this data in a modern, multi-stage training framework that combines synthetic pretraining with real-world alignment, breaking the 3D "data barrier". We obtain significant gains over recent work, with at least a 5:1 win rate in human preference tests on real-world objects and scenes. We will release our code and model weights, an online demo, and a new challenging benchmark for in-the-wild 3D object reconstruction.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| 3D Scene Generation | 3D-Front (test) | CD (Surface)0.052 | 28 | |
| 3D Asset Reconstruction | Toys4k | CD0.0354 | 18 | |
| Amodal 3D object generation | GSO | FID34.68 | 14 | |
| Pose Estimation | Simulation | 3D IoU46 | 12 | |
| Single-object generation | Toy4K | PSNR22.42 | 11 | |
| Novel View Synthesis | GSO-30 | PSNR19.82 | 11 | |
| 3D Object Reconstruction | GSO-30 | Chamfer Distance (×10^-3)0.042 | 11 | |
| Simulator Stability Evaluation | MuJoCo Cluttered Tabletop Scenes (Scenarios 1-5) | Max Kinetic Energy (J)2.08 | 10 | |
| 3D Scene Reconstruction | GraspNet-1B | IoU35.6 | 8 | |
| 3D Dog Reconstruction | Stanford Dog Dataset (test) | FID219.3 | 8 |