Fraccaro, Paolo
Lossy Neural Compression for Geospatial Analytics: A Review
Gomes, Carlos, Wittmann, Isabelle, Robert, Damien, Jakubik, Johannes, Reichelt, Tim, Martone, Michele, Maurogiovanni, Stefano, Vinge, Rikard, Hurst, Jonas, Scheurer, Erik, Sedona, Rocco, Brunschwiler, Thomas, Kesselheim, Stefan, Batic, Matej, Stier, Philip, Wegner, Jan Dirk, Cavallaro, Gabriele, Pebesma, Edzer, Marszalek, Michael, Belenguer-Plomer, Miguel A, Adriko, Kennedy, Fraccaro, Paolo, Kienzler, Romeo, Briq, Rania, Benassou, Sabrina, Lazzarini, Michele, Albrecht, Conrad M
Over the past decades, there has been an explosion in the amount of available Earth Observation (EO) data. The unprecedented coverage of the Earth's surface and atmosphere by satellite imagery has resulted in large volumes of data that must be transmitted to ground stations, stored in data centers, and distributed to end users. Modern Earth System Models (ESMs) face similar challenges, operating at high spatial and temporal resolutions, producing petabytes of data per simulated day. Data compression has gained relevance over the past decade, with neural compression (NC) emerging from deep learning and information theory, making EO data and ESM outputs ideal candidates due to their abundance of unlabeled data. In this review, we outline recent developments in NC applied to geospatial data. We introduce the fundamental concepts of NC including seminal works in its traditional applications to image and video compression domains with focus on lossy compression. We discuss the unique characteristics of EO and ESM data, contrasting them with "natural images", and explain the additional challenges and opportunities they present. Moreover, we review current applications of NC across various EO modalities and explore the limited efforts in ESM compression to date. The advent of self-supervised learning (SSL) and foundation models (FM) has advanced methods to efficiently distill representations from vast unlabeled data. We connect these developments to NC for EO, highlighting the similarities between the two fields and elaborate on the potential of transferring compressed feature representations for machine--to--machine communication. Based on insights drawn from this review, we devise future directions relevant to applications in EO and ESM.
Multispectral to Hyperspectral using Pretrained Foundational model
Gonzalez, Ruben, Albrecht, Conrad M, Braham, Nassim Ait Ali, Lambhate, Devyani, Almeida, Joao Lucas de Sousa, Fraccaro, Paolo, Blumenstiel, Benedikt, Brunschwiler, Thomas, Bangalore, Ranjini
Multispectral to Hyperspectral using Pretrained Foundational model Ruben Gonzalez* 1, Conrad M Albrecht 1, Nassim Ait Ali Braham 1, Devyani Lambhate* 2, Joao Lucas de Sousa Almeida 2, Paolo Fraccaro 2, Benedikt Blumenstiel 2, Thomas Brunschwiler 2, and Ranjini Bangalore 2 1 Remote Sensing Technology Institute, German Aerospace Center (DLR), Germany 2 IBM Research Labs, India, U.K., Zurich, Brazil February 28, 2025 Abstract Hyperspectral imaging provides detailed spectral information, offering significant potential for monitoring greenhouse gases like CH 4 and NO 2. However, its application is constrained by limited spatial coverage and infrequent revisit times. In contrast, multispectral imaging delivers broader spatial and temporal coverage but lacks the spectral granularity required for precise GHG detection. To address these challenges, this study proposes Spectral and Spatial-Spectral transformer models that reconstructs hyperspectral data from multispectral inputs. The models in this paper are pretrained on EnMAP and EMIT datasets and fine-tuned on spatio-temporally aligned (Sentinel-2, EnMAP) and (HLS-S30, EMIT) image pairs respectively. Our model has the potential to enhance atmospheric monitoring by combining the strengths of hyperspectral and multispectral imaging systems. 1 Introduction Satellite images are being used to create detailed maps of Earth's surface.
Fine-tuning of Geospatial Foundation Models for Aboveground Biomass Estimation
Muszynski, Michal, Klein, Levente, da Silva, Ademir Ferreira, Atluri, Anjani Prasad, Gomes, Carlos, Szwarcman, Daniela, Singh, Gurkanwar, Gu, Kewen, Zortea, Maciel, Simumba, Naomi, Fraccaro, Paolo, Singh, Shraddha, Meliksetian, Steve, Watson, Campbell, Kimura, Daiki, Srinivasan, Harini
Global vegetation structure mapping is critical for understanding the global carbon cycle and maximizing the efficacy of nature-based carbon sequestration initiatives. Moreover, vegetation structure mapping can help reduce the impacts of climate change by, for example, guiding actions to improve water security, increase biodiversity and reduce flood risk. Global satellite measurements provide an important set of observations for monitoring and managing deforestation and degradation of existing forests, natural forest regeneration, reforestation, biodiversity restoration, and the implementation of sustainable agricultural practices. In this paper, we explore the effectiveness of fine-tuning of a geospatial foundation model to estimate above-ground biomass (AGB) using space-borne data collected across different eco-regions in Brazil. The fine-tuned model architecture consisted of a Swin-B transformer as the encoder (i.e., backbone) and a single convolutional layer for the decoder head. All results were compared to a U-Net which was trained as the baseline model Experimental results of this sparse-label prediction task demonstrate that the fine-tuned geospatial foundation model with a frozen encoder has comparable performance to a U-Net trained from scratch. This is despite the fine-tuned model having 13 times less parameters requiring optimization, which saves both time and compute resources. Further, we explore the transfer-learning capabilities of the geospatial foundation models by fine-tuning on satellite imagery with sparse labels from different eco-regions in Brazil.
Variational Exploration Module VEM: A Cloud-Native Optimization and Validation Tool for Geospatial Modeling and AI Workflows
Kuehnert, Julian, Tadesse, Hiwot, Dearden, Chris, Lickorish, Rosie, Fraccaro, Paolo, Jones, Anne, Edwards, Blair, Remy, Sekou L., Melling, Peter, Culmer, Tim
Geospatial observations combined with computational models have become key to understanding the physical systems of our environment and enable the design of best practices to reduce societal harm. Cloud-based deployments help to scale up these modeling and AI workflows. Yet, for practitioners to make robust conclusions, model tuning and testing is crucial, a resource intensive process which involves the variation of model input variables. We have developed the Variational Exploration Module which facilitates the optimization and validation of modeling workflows deployed in the cloud by orchestrating workflow executions and using Bayesian and machine learning-based methods to analyze model behavior. User configurations allow the combination of diverse sampling strategies in multi-agent environments. The flexibility and robustness of the model-agnostic module is demonstrated using real-world applications.
Foundation Models for Generalist Geospatial Artificial Intelligence
Jakubik, Johannes, Roy, Sujit, Phillips, C. E., Fraccaro, Paolo, Godwin, Denys, Zadrozny, Bianca, Szwarcman, Daniela, Gomes, Carlos, Nyirjesy, Gabby, Edwards, Blair, Kimura, Daiki, Simumba, Naomi, Chu, Linsong, Mukkavilli, S. Karthik, Lambhate, Devyani, Das, Kamal, Bangalore, Ranjini, Oliveira, Dario, Muszynski, Michal, Ankur, Kumar, Ramasubramanian, Muthukumaran, Gurung, Iksha, Khallaghi, Sam, Hanxi, null, Li, null, Cecil, Michael, Ahmadi, Maryam, Kordi, Fatemeh, Alemohammad, Hamed, Maskey, Manil, Ganti, Raghu, Weldemariam, Kommy, Ramachandran, Rahul
Significant progress in the development of highly adaptable and reusable Artificial Intelligence (AI) models is expected to have a significant impact on Earth science and remote sensing. Foundation models are pre-trained on large unlabeled datasets through self-supervision, and then fine-tuned for various downstream tasks with small labeled datasets. This paper introduces a first-of-a-kind framework for the efficient pre-training and fine-tuning of foundational models on extensive geospatial data. We have utilized this framework to create Prithvi, a transformer-based geospatial foundational model pre-trained on more than 1TB of multispectral satellite imagery from the Harmonized Landsat-Sentinel 2 (HLS) dataset. Our study demonstrates the efficacy of our framework in successfully fine-tuning Prithvi to a range of Earth observation tasks that have not been tackled by previous work on foundation models involving multi-temporal cloud gap imputation, flood mapping, wildfire scar segmentation, and multi-temporal crop segmentation. Our experiments show that the pre-trained model accelerates the fine-tuning process compared to leveraging randomly initialized weights. In addition, pre-trained Prithvi compares well against the state-of-the-art, e.g., outperforming a conditional GAN model in multi-temporal cloud imputation by up to 5pp (or 5.7%) in the structural similarity index. Finally, due to the limited availability of labeled data in the field of Earth observation, we gradually reduce the quantity of available labeled data for refining the model to evaluate data efficiency and demonstrate that data can be decreased significantly without affecting the model's accuracy. The pre-trained 100 million parameter model and corresponding fine-tuning workflows have been released publicly as open source contributions to the global Earth sciences community through Hugging Face.