Robotic Grasp Detection using Deep Convolutional Neural Networks
About
Deep learning has significantly advanced computer vision and natural language processing. While there have been some successes in robotics using deep learning, it has not been widely adopted. In this paper, we present a novel robotic grasp detection system that predicts the best grasping pose of a parallel-plate robotic gripper for novel objects using the RGB-D image of the scene. The proposed model uses a deep convolutional neural network to extract features from the scene and then uses a shallow convolutional neural network to predict the grasp configuration for the object of interest. Our multi-modal model achieved an accuracy of 89.21% on the standard Cornell Grasp Dataset and runs at real-time speeds. This redefines the state-of-the-art for robotic grasp detection.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Grasp Detection | Cornell Dataset (object-wise) | Accuracy88.9 | 39 | |
| Grasp Detection | Cornell Dataset image-wise | Accuracy89.2 | 25 | |
| Grasp Detection | Cornell image-wise | Accuracy89.2 | 24 | |
| Grasp Detection | Cornell Grasping Dataset (Image-wise split) | Detection Accuracy89.21 | 17 | |
| Robotic Grasp Detection | Cornell Grasp Dataset (Object-wise) | Accuracy88.96 | 14 | |
| Grasp Prediction | Cornell Grasping Dataset | Speed (fps)16.03 | 5 | |
| Physical Grasping | Physical Grasping Evaluation 10 objects | Detection Success Rate61 | 4 |