Lightweight Metadata-Aware Mixture-of-Experts Masked Autoencoder for Earth Observation
–arXiv.org Artificial Intelligence
Recent advances in Earth Observation have focused on large-scale foundation models. However, these models are computationally expensive, limiting their accessibility and reuse for downstream tasks. In this work, we investigate compact architectures as a practical pathway toward smaller general-purpose EO models. We propose a Metadata-aware Mixture-of-Experts Masked Autoencoder (MoE-MAE) with only 2.5M parameters. The model combines sparse expert routing with geo-temporal conditioning, incorporating imagery alongside latitude/longitude and seasonal/daily cyclic encodings. We pretrain the MoE-MAE on the BigEarthNet-Landsat dataset and evaluate embeddings from its frozen encoder using linear probes. Despite its small size, the model competes with much larger architectures, demonstrating that metadata-aware pretraining improves transfer and label efficiency. To further assess generalization, we evaluate on the EuroSAT-Landsat dataset, which lacks explicit metadata, and still observe competitive performance compared to models with hundreds of millions of parameters. These results suggest that compact, metadata-aware MoE-MAEs are an efficient and scalable step toward future EO foundation models.
arXiv.org Artificial Intelligence
Sep-16-2025
- Country:
- Asia > India
- Europe
- Germany > North Rhine-Westphalia
- Cologne Region > Bonn (0.04)
- Slovenia > Drava
- Municipality of Benedikt > Benedikt (0.04)
- Germany > North Rhine-Westphalia
- North America > United States
- Colorado (0.04)
- Genre:
- Research Report > New Finding (0.67)
- Technology: