Perreault-Levasseur, Laurence
Solving Bayesian inverse problems with diffusion priors and off-policy RL
Scimeca, Luca, Venkatraman, Siddarth, Jain, Moksh, Kim, Minsu, Sendera, Marcin, Hasan, Mohsin, Rowe, Luke, Mittal, Sarthak, Lemos, Pablo, Bengio, Emmanuel, Adam, Alexandre, Rector-Brooks, Jarrid, Hezaveh, Yashar, Perreault-Levasseur, Laurence, Bengio, Yoshua, Berseth, Glen, Malkin, Nikolay
This paper presents a practical application of Relative Trajectory Balance (RTB), a recently introduced off-policy reinforcement learning (RL) objective that can asymptotically solve Bayesian inverse problems optimally. We extend the original work by using RTB to train conditional diffusion model posteriors from pretrained unconditional priors for challenging linear and non-linear inverse problems in vision, and science. We use the objective alongside techniques such as off-policy backtracking exploration to improve training. Importantly, our results show that existing training-free diffusion posterior methods struggle to perform effective posterior inference in latent space due to inherent biases.
IRIS: A Bayesian Approach for Image Reconstruction in Radio Interferometry with expressive Score-Based priors
Dia, Noé, Yantovski-Barth, M. J., Adam, Alexandre, Bowles, Micah, Perreault-Levasseur, Laurence, Hezaveh, Yashar, Scaife, Anna
Inferring sky surface brightness distributions from noisy interferometric data in a principled statistical framework has been a key challenge in radio astronomy. In this work, we introduce Imaging for Radio Interferometry with Score-based models (IRIS). We use score-based models trained on optical images of galaxies as an expressive prior in combination with a Gaussian likelihood in the uv-space to infer images of protoplanetary disks from visibility data of the DSHARP survey conducted by ALMA. We demonstrate the advantages of this framework compared with traditional radio interferometry imaging algorithms, showing that it produces plausible posterior samples despite the use of a misspecified galaxy prior. Through coverage testing on simulations, we empirically evaluate the accuracy of this approach to generate calibrated posterior samples.
PQMass: Probabilistic Assessment of the Quality of Generative Models using Probability Mass Estimation
Lemos, Pablo, Sharief, Sammy, Malkin, Nikolay, Perreault-Levasseur, Laurence, Hezaveh, Yashar
With advancements in generative models, evaluating their performance using rigorous, clearly defined metrics and We propose a comprehensive sample-based criteria has become increasingly essential. Disambiguating method for assessing the quality of generative true from modeled distributions is especially pertinent in models. The proposed approach enables the estimation light of the growing emphasis on AI safety within the community, of the probability that two sets of samples as well as in scientific domains where stringent standards are drawn from the same distribution, providing of rigor and uncertainty quantification are needed for a statistically rigorous method for assessing the the adoption of machine learning methods. When evaluating performance of a single generative model or the generative models, we are interested in three qualitative comparison of multiple competing models trained properties (Stein et al., 2023; Jiralerspong et al., 2023): Fidelity on the same dataset. This comparison can be conducted refers to the quality and realism of individual outputs by dividing the space into non-overlapping generated by a model. It assesses how indistinguishable regions and comparing the number of data samples each generated sample is from real data.
Improving Gradient-guided Nested Sampling for Posterior Inference
Lemos, Pablo, Malkin, Nikolay, Handley, Will, Bengio, Yoshua, Hezaveh, Yashar, Perreault-Levasseur, Laurence
Gaussian noise was then added to produce a noisy simulated data. Given the data, the posterior of a model (a pixelated image of the undistorted background source) could be calculated by adding the likelihood and the prior terms. Furthermore since the model is perfectly linear (and known) and the noise and the prior are Gaussian, the posterior is a high-dimensional Gaussian posterior that could be calculated analytically, allowing us to compare the samples drawn with GGNS with the analytic solution. Figure 2 shows a comparison between the true image, and its noise, and the one recovered by GGNS. We see that we can recover both the correct image, and the noise distribution. We emphasize that this is a uni-modal problem and that the experiment's goal is to demonstrate the capability of GGNS to sample in high dimensions (in this case, 256), such as images, and to test the agreement between the samples and a baseline analytic solution.
Lie Point Symmetry and Physics Informed Networks
Akhound-Sadegh, Tara, Perreault-Levasseur, Laurence, Brandstetter, Johannes, Welling, Max, Ravanbakhsh, Siamak
Symmetries have been leveraged to improve the generalization of neural networks through different mechanisms from data augmentation to equivariant architectures. However, despite their potential, their integration into neural solvers for partial differential equations (PDEs) remains largely unexplored. We explore the integration of PDE symmetries, known as Lie point symmetries, in a major family of neural solvers known as physics-informed neural networks (PINNs). We propose a loss function that informs the network about Lie point symmetries in the same way that PINN models try to enforce the underlying PDE through a loss function. Intuitively, our symmetry loss ensures that the infinitesimal generators of the Lie group conserve the PDE solutions. Effectively, this means that once the network learns a solution, it also learns the neighbouring solutions generated by Lie point symmetries. Empirical evaluations indicate that the inductive bias introduced by the Lie point symmetries of the PDEs greatly boosts the sample efficiency of PINNs.
Sampling-Based Accuracy Testing of Posterior Estimators for General Inference
Lemos, Pablo, Coogan, Adam, Hezaveh, Yashar, Perreault-Levasseur, Laurence
Parameter inference, i.e. inferring the posterior distribution of the parameters of a statistical model given some data, is a central problem to many scientific disciplines. Generative models can be used as an alternative to Markov Chain Monte Carlo methods for conducting posterior inference, both in likelihood-based and simulation-based problems. However, assessing the accuracy of posteriors encoded in generative models is not straightforward. In this paper, we introduce `Tests of Accuracy with Random Points' (TARP) coverage testing as a method to estimate coverage probabilities of generative posterior estimators. Our method differs from previously-existing coverage-based methods, which require posterior evaluations. We prove that our approach is necessary and sufficient to show that a posterior estimator is accurate. We demonstrate the method on a variety of synthetic examples, and show that TARP can be used to test the results of posterior inference analyses in high-dimensional spaces. We also show that our method can detect inaccurate inferences in cases where existing methods fail.
Posterior samples of source galaxies in strong gravitational lenses with score-based priors
Adam, Alexandre, Coogan, Adam, Malkin, Nikolay, Legin, Ronan, Perreault-Levasseur, Laurence, Hezaveh, Yashar, Bengio, Yoshua
Inferring accurate posteriors for high-dimensional representations of the brightness of gravitationally-lensed sources is a major challenge, in part due to the difficulties of accurately quantifying the priors. Here, we report the use of a score-based model to encode the prior for the inference of undistorted images of background galaxies. This model is trained on a set of high-resolution images of undistorted galaxies. By adding the likelihood score to the prior score and using a reverse-time stochastic differential equation solver, we obtain samples from the posterior. Our method produces independent posterior samples and models the data almost down to the noise level. We show how the balance between the likelihood and the prior meet our expectations in an experiment with out-of-distribution data.
Bayesian Neural Networks
Charnock, Tom, Perreault-Levasseur, Laurence, Lanusse, François
In recent times, neural networks have become a powerful tool for the analysis of complex and abstract data models. However, their introduction intrinsically increases our uncertainty about which features of the analysis are model-related and which are due to the neural network. This means that predictions by neural networks have biases which cannot be trivially distinguished from being due to the true nature of the creation and observation of data or not. In order to attempt to address such issues we discuss Bayesian neural networks: neural networks where the uncertainty due to the network can be characterised. In particular, we present the Bayesian statistical framework which allows us to categorise uncertainty in terms of the ingrained randomness of observing certain data and the uncertainty from our lack of knowledge about how data can be created and observed. In presenting such techniques we show how errors in prediction by neural networks can be obtained in principle, and provide the two favoured methods for characterising these errors. We will also describe how both of these methods have substantial pitfalls when put into practice, highlighting the need for other statistical techniques to truly be able to do inference when using neural networks.