6-PACK: Category-level 6D Pose Tracker with Anchor-Based Keypoints
About
We present 6-PACK, a deep learning approach to category-level 6D object pose tracking on RGB-D data. Our method tracks in real-time novel object instances of known object categories such as bowls, laptops, and mugs. 6-PACK learns to compactly represent an object by a handful of 3D keypoints, based on which the interframe motion of an object instance can be estimated through keypoint matching. These keypoints are learned end-to-end without manual supervision in order to be most effective for tracking. Our experiments show that our method substantially outperforms existing methods on the NOCS category-level 6D pose estimation benchmark and supports a physical robot to perform simple vision-based closed-loop manipulation tasks. Our code and video are available at https://sites.google.com/view/6packtracking.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| 6D Object Pose Tracking | NOCS (test) | 5°5cm Accuracy33.3 | 8 | |
| Category-level pose tracking | REAL275 | 5°5cm Accuracy31.6 | 7 | |
| 6D Object Pose Tracking | YCBInEOAT (test) | ADD (003 Cracker Box)4.18 | 7 | |
| Category-level object pose tracking | REAL275 (test) | 5°5cm Accuracy33.3 | 6 | |
| Category-level Pose Estimation | NOCS-REAL (test) | IoU@0.2594.2 | 5 |