Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Activating Self-Attention for Multi-Scene Absolute Pose Regression

About

Multi-scene absolute pose regression addresses the demand for fast and memory-efficient camera pose estimation across various real-world environments. Nowadays, transformer-based model has been devised to regress the camera pose directly in multi-scenes. Despite its potential, transformer encoders are underutilized due to the collapsed self-attention map, having low representation capacity. This work highlights the problem and investigates it from a new perspective: distortion of query-key embedding space. Based on the statistical analysis, we reveal that queries and keys are mapped in completely different spaces while only a few keys are blended into the query region. This leads to the collapse of the self-attention map as all queries are considered similar to those few keys. Therefore, we propose simple but effective solutions to activate self-attention. Concretely, we present an auxiliary loss that aligns queries and keys, preventing the distortion of query-key space and encouraging the model to find global relations by self-attention. In addition, the fixed sinusoidal positional encoding is adopted instead of undertrained learnable one to reflect appropriate positional clues into the inputs of self-attention. As a result, our approach resolves the aforementioned problem effectively, thus outperforming existing methods in both outdoor and indoor scenes.

Miso Lee, Jihwan Kim, Jae-Pil Heo• 2024

Related benchmarks

TaskDatasetResultRank
Visual LocalizationCambridge Landmarks (test)
Avg Median Positional Error (m)1.19
35
Absolute Pose Regression7-Scenes SfM GT (test)
Chess Translation Error (m)0.1
9
Visual LocalizationCambridge Landmarks 8 (test)
Recall (1m, 5°)35.8
2
Visual Localization7Scenes 27 (test)
Recall @ 0.2m, 5°32.6
2
Showing 4 of 4 rows

Other info

Follow for update