LEPA: Learning Geometric Equivariance in Satellite Remote Sensing Data with a Predictive Architecture
About
Geospatial foundation models provide precomputed embeddings that serve as compact feature vectors for large-scale satellite remote sensing data. While these embeddings can reduce data-transfer bottlenecks and computational costs, Earth observation (EO) applications can still face geometric mismatches between user-defined areas of interest and the fixed precomputed embedding grid. Standard latent-space interpolation is unreliable in this setting because the embedding manifold is highly non-convex, yielding representations that do not correspond to realistic inputs. We verify this using Prithvi-EO-2.0 to understand the shortcomings of interpolation applied to patch embeddings. As a substitute, we propose a Learned Equivariance-Predicting Architecture (LEPA). Instead of averaging vectors, LEPA conditions a predictor on geometric augmentations to directly predict the transformed embedding. We evaluate LEPA on NASA/USGS Harmonized Landsat-Sentinel (HLS) imagery and ImageNet-1k. Experiments show that standard interpolation achieves a mean reciprocal rank (MRR) below 0.2, whereas LEPA increases MRR to over 0.8, enabling accurate geometric adjustment without re-encoding.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Semantic segmentation | Sen1Floods11 | mIoU (macro)87.37 | 29 | |
| Semantic segmentation | MADOS | mIoU51.4 | 26 | |
| Semantic segmentation | HLS Burn Scars | mIoU83.03 | 25 | |
| Semantic segmentation | PASTIS | Macro mIoU33.84 | 24 | |
| Semantic segmentation | DynamicEarthNet (DEN) | mIoU33.84 | 19 | |
| Semantic segmentation | SpaceNet 7 | mIoU56.31 | 19 | |
| Semantic segmentation | AI4SmallFarms | mIoU24.01 | 14 |