From Limited Labels to Open Domains:An Efficient Learning Method for Drone-view Geo-Localization
About
Traditional supervised drone-view geo-localization (DVGL) methods heavily depend on paired training data and encounter difficulties in learning cross-view correlations from unpaired data. Moreover, when deployed in a new domain, these methods require obtaining the new paired data and subsequent retraining for model adaptation, which significantly increases computational overhead. Existing unsupervised methods have enabled to generate pseudo-labels based on cross-view similarity to infer the pairing relationships. However, geographical similarity and spatial continuity often cause visually analogous features at different geographical locations. The feature confusion compromises the reliability of pseudo-label generation, where incorrect pseudo-labels drive negative optimization. Given these challenges inherent in both supervised and unsupervised DVGL methods, we propose a novel cross-domain invariant knowledge transfer network (CDIKTNet) with limited supervision, whose architecture consists of a cross-domain invariance sub-network (CDIS) and a cross-domain transfer sub-network (CDTS). This architecture facilitates a closed-loop framework for invariance feature learning and knowledge transfer. The CDIS is designed to learn cross-view structural and spatial invariance from a small amount of paired data that serves as prior knowledge. It endows the shared feature space of unpaired data with similar implicit cross-view correlations at initialization, which alleviates feature confusion. Based on this, the CDTS employs dual-path contrastive learning to further optimize each subspace while preserving consistency in a shared feature space. Extensive experiments demonstrate that CDIKTNet achieves state-of-the-art performance under full supervision compared with those supervised methods, and further surpasses existing unsupervised methods in both few-shot and cross-domain initialization.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Drone-to-Satellite Retrieval | SUES-200 150m | R@188.75 | 98 | |
| Drone-to-Satellite Retrieval | SUES-200 250m | R@193.15 | 76 | |
| Drone-to-Satellite Retrieval | SUES-200 300m | R@195.18 | 66 | |
| Drone-to-Satellite Retrieval | SUES-200 200m | R@1 Accuracy89.35 | 66 | |
| Satellite-to-Drone Retrieval | SUES-200 300 m | R@198.75 | 22 | |
| Satellite-to-Drone Retrieval | SUES-200 200 m | R@193.75 | 22 | |
| Drone-to-Satellite Retrieval | University-1652 | R@10.833 | 22 | |
| Satellite-to-Drone Retrieval | SUES-200 250 m | Recall@195 | 22 | |
| Satellite-to-Drone Retrieval | University-1652 | Recall@187.73 | 22 |