PR-GCN: A Deep Graph Convolutional Network with Point Refinement for 6D Pose Estimation
About
RGB-D based 6D pose estimation has recently achieved remarkable progress, but still suffers from two major limitations: (1) ineffective representation of depth data and (2) insufficient integration of different modalities. This paper proposes a novel deep learning approach, namely Graph Convolutional Network with Point Refinement (PR-GCN), to simultaneously address the issues above in a unified way. It first introduces the Point Refinement Network (PRN) to polish 3D point clouds, recovering missing parts with noise removed. Subsequently, the Multi-Modal Fusion Graph Convolutional Network (MMF-GCN) is presented to strengthen RGB-D combination, which captures geometry-aware inter-modality correlation through local information propagation in the graph convolutional network. Extensive experiments are conducted on three widely used benchmarks, and state-of-the-art performance is reached. Besides, it is also shown that the proposed PRN and MMF-GCN modules are well generalized to other frameworks.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| 6D Pose Estimation | YCB-Video | -- | 148 | |
| 6D Object Pose Estimation | LineMOD | -- | 50 | |
| 6D Object Pose Estimation | OccludedLINEMOD (test) | ADD(S)65 | 45 | |
| 6D Object Pose Estimation | LM-O (test) | Recall (Mean)65 | 22 | |
| Object Pose Estimation | LineMod (test) | -- | 21 |