RPNet: an End-to-End Network for Relative Camera Pose Estimation
About
This paper addresses the task of relative camera pose estimation from raw image pixels, by means of deep neural networks. The proposed RPNet network takes pairs of images as input and directly infers the relative poses, without the need of camera intrinsic/extrinsic. While state-of-the-art systems based on SIFT + RANSAC, are able to recover the translation vector only up to scale, RPNet is trained to produce the full translation vector, in an end-to-end way. Experimental results on the Cambridge Landmark dataset show very promising results regarding the recovery of the full translation vector. They also show that RPNet produces more accurate and more stable results than traditional approaches, especially for hard images (repetitive textures, textureless images, etc). To the best of our knowledge, RPNet is the first attempt to recover full translation vectors in relative pose estimation.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Relative Pose Estimation | KITTI Sequence 01 | Rotation RMSE10.811 | 20 | |
| Relative Pose Estimation | KITTI Odometry Sequence 05 (test) | ATE (m)3.7386 | 9 | |
| Relative Camera Pose Estimation | KITTI Sequence 05 (test) | Absolute Trajectory Error (m)2.5258 | 9 | |
| Relative Camera Pose Estimation | KITTI Sequence 09 (test) | ATE (m)4.3127 | 9 | |
| Relative Pose Estimation | KITTI Odometry Sequence 05 | ATE (m)0.9544 | 9 | |
| Relative Pose Estimation | KITTI Odometry Sequence 09 | ATE (m)0.7121 | 9 | |
| Relative Pose Estimation | KITTI Odometry Sequence 01 (test) | ATE (m)4.8743 | 9 | |
| Relative Pose Estimation | KITTI Odometry Sequence 09 (test) | ATE (m)2.2229 | 9 | |
| Relative Pose Estimation | ETH3D Botanical Garden sequence (test) | ATE (m)1.8289 | 8 | |
| Relative Pose Estimation | ETH3D Statue sequence (test) | Absolute Trajectory Error (m)0.2557 | 8 |