Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Antipodal Robotic Grasping using Generative Residual Convolutional Neural Network

About

In this paper, we present a modular robotic system to tackle the problem of generating and performing antipodal robotic grasps for unknown objects from n-channel image of the scene. We propose a novel Generative Residual Convolutional Neural Network (GR-ConvNet) model that can generate robust antipodal grasps from n-channel input at real-time speeds (~20ms). We evaluate the proposed model architecture on standard datasets and a diverse set of household objects. We achieved state-of-the-art accuracy of 97.7% and 94.6% on Cornell and Jacquard grasping datasets respectively. We also demonstrate a grasp success rate of 95.4% and 93% on household and adversarial objects respectively using a 7 DoF robotic arm.

Sulabh Kumra, Shirin Joshi, Ferat Sahin• 2019

Related benchmarks

TaskDatasetResultRank
Grasp DetectionCornell Dataset (object-wise)
Accuracy96.8
39
Grasp DetectionCornell Dataset image-wise
Accuracy97.7
25
Grasp DetectionCornell image-wise
Accuracy97.7
24
Grasp DetectionJacquard Dataset
Accuracy94.6
16
Language-guided grasp detection and segmentationOCID VLG (Multi-Split)
J@10.097
11
Language-guided grasp detectionGrasp-Anything++ full prompt sentence (Seen)
Success Rate21
6
Robotic GraspingHousehold Objects
Accuracy95.4
5
Language-guided grasp detectionGrasp-Anything++ full prompt sentence (unseen)
Success Rate12
5
Robotic GraspingAdversarial Objects
Accuracy93
2
Showing 9 of 9 rows

Other info

Code

Follow for update