Rey, Mélanie
Neural Compression of Atmospheric States
Mirowski, Piotr, Warde-Farley, David, Rosca, Mihaela, Grimes, Matthew Koichi, Hasson, Yana, Kim, Hyunjik, Rey, Mélanie, Osindero, Simon, Ravuri, Suman, Mohamed, Shakir
This paper presents a family of neural network compression methods of simulated atmospheric states, with the aim of reducing the currently immense storage requirements of such data from cloud scale (petabytes) to desktop scale (terabytes). This need for compression has come about over past 50 years, characterized by a steady push to increase the resolution of atmospheric simulations, increasing the size and storage demands of the resulting datasets (e.g., Neumann et al. (2019), Schneider et al. (2023), Stevens et al. (2024)), while atmospheric simulation has come to play an increasingly critical role in scientific, industrial and policy-level pursuits. Higher spatial resolutions unlock the ability of simulators to deliver more accurate predictions and resolve ever more atmospheric phenomena. For example, while current models often operate at 25 - 50 km resolution, resolving storms requires 1 km resolution (Stevens et al., 2020), while resolving the motion of (and radiative effects due to) low clouds require 100 m resolution (Satoh et al., 2019; Schneider et al., 2017). Machine learning models for weather prediction also face opportunities and challenges with higher resolution: while additional granularity may afford better modeling opportunities, even the present size of atmospheric states poses a significant bottleneck for loading training data and serving model outputs (Chantry et al., 2021). To put the data storage problem in perspective, storing 40 years of reanalysis data from the ECMWF Reanalysis v5 dataset (ERA5, Hersbach et al. (2020)) at full spatial and temporal resolution (i.e.
Understanding Deep Generative Models with Generalized Empirical Likelihoods
Ravuri, Suman, Rey, Mélanie, Mohamed, Shakir, Deisenroth, Marc
Understanding how well a deep generative model captures a distribution of high-dimensional data remains an important open challenge. It is especially difficult for certain model classes, such as Generative Adversarial Networks and Diffusion Models, whose models do not admit exact likelihoods. In this work, we demonstrate that generalized empirical likelihood (GEL) methods offer a family of diagnostic tools that can identify many deficiencies of deep generative models (DGMs). We show, with appropriate specification of moment conditions, that the proposed method can identify which modes have been dropped, the degree to which DGMs are mode imbalanced, and whether DGMs sufficiently capture intra-class diversity. We show how to combine techniques from Maximum Mean Discrepancy and Generalized Empirical Likelihood to create not only distribution tests that retain per-sample interpretability, but also metrics that include label information. We find that such tests predict the degree of mode dropping and mode imbalance up to 60% better than metrics such as improved precision/recall. We provide an implementation at https://github.com/deepmind/understanding_deep_generative_models_with_generalized_empirical_likelihood/.