Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Single-View Shape Completion for Robotic Grasping in Clutter

About

In vision-based robot manipulation, a single camera view can only capture one side of objects of interest, with additional occlusions in cluttered scenes further restricting visibility. As a result, the observed geometry is incomplete, and grasp estimation algorithms perform suboptimally. To address this limitation, we leverage diffusion models to perform category-level 3D shape completion from partial depth observations obtained from a single view, reconstructing complete object geometries to provide richer context for grasp planning. Our method focuses on common household items with diverse geometries, generating full 3D shapes that serve as input to downstream grasp inference networks. Unlike prior work, which primarily considers isolated objects or minimal clutter, we evaluate shape completion and grasping in realistic clutter scenarios with household objects. In preliminary evaluations on a cluttered scene, our approach consistently results in better grasp success rates than a naive baseline without shape completion by 23% and over a recent state of the art shape completion approach by 19%. Our code is available at https://amm.aass.oru.se/shape-completion-grasping/.

Abhishek Kashyap, Yuxuan Yang, Henrik Andreasson, Todor Stoyanov• 2025

Related benchmarks

TaskDatasetResultRank
Robotic GraspingReOcS (normal-hard)
Success Rate100
21
3D ReconstructionReOcS Easy clutter
CD (bottle)16.75
2
3D ReconstructionReOcS Normal clutter
CD (bottle)15.95
2
3D ReconstructionReOcS Hard clutter
Chamfer Distance (Bottle)16.9
2
Showing 4 of 4 rows

Other info

Follow for update