Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

SatMAE: Pre-training Transformers for Temporal and Multi-Spectral Satellite Imagery

About

Unsupervised pre-training methods for large vision models have shown to enhance performance on downstream supervised tasks. Developing similar techniques for satellite imagery presents significant opportunities as unlabelled data is plentiful and the inherent temporal and multi-spectral structure provides avenues to further improve existing pre-training strategies. In this paper, we present SatMAE, a pre-training framework for temporal or multi-spectral satellite imagery based on Masked Autoencoder (MAE). To leverage temporal information, we include a temporal embedding along with independently masking image patches across time. In addition, we demonstrate that encoding multi-spectral data as groups of bands with distinct spectral positional encodings is beneficial. Our approach yields strong improvements over previous state-of-the-art techniques, both in terms of supervised learning performance on benchmark datasets (up to $\uparrow$ 7%), and transfer learning performance on downstream remote sensing tasks, including land cover classification (up to $\uparrow$ 14%) and semantic segmentation. Code and data are available on the project website: https://sustainlab-group.github.io/SatMAE/

Yezhen Cong, Samar Khanna, Chenlin Meng, Patrick Liu, Erik Rozi, Yutong He, Marshall Burke, David B. Lobell, Stefano Ermon• 2022

Related benchmarks

TaskDatasetResultRank
Image ClassificationEuroSAT
Accuracy83.92
569
Change DetectionLEVIR-CD (test)
F1 Score87.65
485
Image ClassificationRESISC45
Accuracy94.8
349
Change DetectionLEVIR-CD
F1 Score87.65
232
Semantic segmentationVaihingen
mIoU73.64
140
Semantic segmentationiSAID
mIoU62.97
122
Scene ClassificationAID TR=50%
Accuracy96.94
94
Scene ClassificationAID TR=20%
Accuracy95.02
93
Semantic segmentationPotsdam
mIoU73.55
81
Scene ClassificationRESISC-45 (TR=10%)
Accuracy91.72
63
Showing 10 of 105 rows
...

Other info

Follow for update