Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

OVE6D: Object Viewpoint Encoding for Depth-based 6D Object Pose Estimation

About

This paper proposes a universal framework, called OVE6D, for model-based 6D object pose estimation from a single depth image and a target object mask. Our model is trained using purely synthetic data rendered from ShapeNet, and, unlike most of the existing methods, it generalizes well on new real-world objects without any fine-tuning. We achieve this by decomposing the 6D pose into viewpoint, in-plane rotation around the camera optical axis and translation, and introducing novel lightweight modules for estimating each component in a cascaded manner. The resulting network contains less than 4M parameters while demonstrating excellent performance on the challenging T-LESS and Occluded LINEMOD datasets without any dataset-specific training. We show that OVE6D outperforms some contemporary deep learning-based pose estimation methods specifically trained for individual objects or datasets with real-world training data. The implementation and the pre-trained model will be made publicly available.

Dingding Cai, Janne Heikkil\"a, Esa Rahtu• 2022

Related benchmarks

TaskDatasetResultRank
6D Object Pose EstimationLineMOD--
50
6D Object Pose EstimationOccludedLINEMOD (test)
ADD(S)82.5
45
Pose EstimationBOP benchmark 2019 (test)
LM-O AR49.6
43
6D Object Pose EstimationT-LESS BOP challenge protocol PrimeSense (test)
VSD91
20
Pose EstimationLM-O
ADD(-S)56.1
3
Showing 5 of 5 rows

Other info

Code

Follow for update