Antipodal Robotic Grasping using Generative Residual Convolutional Neural Network
About
In this paper, we present a modular robotic system to tackle the problem of generating and performing antipodal robotic grasps for unknown objects from n-channel image of the scene. We propose a novel Generative Residual Convolutional Neural Network (GR-ConvNet) model that can generate robust antipodal grasps from n-channel input at real-time speeds (~20ms). We evaluate the proposed model architecture on standard datasets and a diverse set of household objects. We achieved state-of-the-art accuracy of 97.7% and 94.6% on Cornell and Jacquard grasping datasets respectively. We also demonstrate a grasp success rate of 95.4% and 93% on household and adversarial objects respectively using a 7 DoF robotic arm.
Sulabh Kumra, Shirin Joshi, Ferat Sahin• 2019
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Grasp Detection | Cornell Dataset (object-wise) | Accuracy96.8 | 39 | |
| Grasp Detection | Cornell Dataset image-wise | Accuracy97.7 | 25 | |
| Grasp Detection | Cornell image-wise | Accuracy97.7 | 24 | |
| Grasp Detection | Jacquard Dataset | Accuracy94.6 | 16 | |
| Language-guided grasp detection and segmentation | OCID VLG (Multi-Split) | J@10.097 | 11 | |
| Grasping | Jacquard Franka | Success Rate82.7 | 6 | |
| Grasping | Jacquard Robotiq-3F | Success Rate88.5 | 6 | |
| Grasping | Jacquard Average | Success Rate (SR)83.9 | 6 | |
| Grasping | Jacquard WSG50 | Success Rate86.2 | 6 | |
| Grasping | Jacquard Barrett | SR80.4 | 6 |
Showing 10 of 17 rows