Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Improved Visual Relocalization by Discovering Anchor Points

About

We address the visual relocalization problem of predicting the location and camera orientation or pose (6DOF) of the given input scene. We propose a method based on how humans determine their location using the visible landmarks. We define anchor points uniformly across the route map and propose a deep learning architecture which predicts the most relevant anchor point present in the scene as well as the relative offsets with respect to it. The relevant anchor point need not be the nearest anchor point to the ground truth location, as it might not be visible due to the pose. Hence we propose a multi task loss function, which discovers the relevant anchor point, without needing the ground truth for it. We validate the effectiveness of our approach by experimenting on CambridgeLandmarks (large scale outdoor scenes) as well as 7 Scenes (indoor scenes) using variousCNN feature extractors. Our method improves the median error in indoor as well as outdoor localization datasets compared to the previous best deep learning model known as PoseNet (with geometric re-projection loss) using the same feature extractor. We improve the median error in localization in the specific case of Street scene, by over 8m.

Soham Saha, Girish Varma, C.V.Jawahar• 2018

Related benchmarks

TaskDatasetResultRank
Camera Localization7 Scenes
Average Position Error (m)0.09
46
Visual LocalizationCambridge Landmarks (test)
Avg Median Positional Error (m)0.84
35
Pose Estimation7 Scenes
Average Median Translation Error (m)0.09
23
Camera Pose RegressionCambridge Landmarks (test)
Average Translation Error (Median, 4 Scenes, m)0.84
16
Visual Localization360SPR Pinhole (unseen)
TE (m)28.23
14
Visual Localization360SPR (seen)
Median Translation Error (m)10.11
7
Showing 6 of 6 rows

Other info

Follow for update