Plotting

 Pidstrigach, Jakiw


Conditioning Diffusions Using Malliavin Calculus

arXiv.org Machine Learning

In stochastic optimal control and conditional generative modelling, a central computational task is to modify a reference diffusion process to maximise a given terminal-time reward. Most existing methods require this reward to be differentiable, using gradients to steer the diffusion towards favourable outcomes. However, in many practical settings, like diffusion bridges, the reward is singular, taking an infinite value if the target is hit and zero otherwise. We introduce a novel framework, based on Malliavin calculus and path-space integration by parts, that enables the development of methods robust to such singular rewards. This allows our approach to handle a broad range of applications, including classification, diffusion bridges, and conditioning without the need for artificial observational noise. We demonstrate that our approach offers stable and reliable training, outperforming existing techniques.


Infinite-Dimensional Diffusion Models

arXiv.org Machine Learning

Diffusion models have had a profound impact on many application areas, including those where data are intrinsically infinite-dimensional, such as images or time series. The standard approach is first to discretize and then to apply diffusion models to the discretized data. While such approaches are practically appealing, the performance of the resulting algorithms typically deteriorates as discretization parameters are refined. In this paper, we instead directly formulate diffusion-based generative models in infinite dimensions and apply them to the generative modeling of functions. We prove that our formulations are well posed in the infinite-dimensional setting and provide dimension-independent distance bounds from the sample to the target measure. Using our theory, we also develop guidelines for the design of infinite-dimensional diffusion models. For image distributions, these guidelines are in line with the canonical choices currently made for diffusion models. For other distributions, however, we can improve upon these canonical choices, which we show both theoretically and empirically, by applying the algorithms to data distributions on manifolds and inspired by Bayesian inverse problems or simulation-based inference.


Affine Invariant Ensemble Transform Methods to Improve Predictive Uncertainty in ReLU Networks

arXiv.org Machine Learning

We consider the problem of performing Bayesian inference for logistic regression using appropriate extensions of the ensemble Kalman filter. Two interacting particle systems are proposed that sample from an approximate posterior and prove quantitative convergence rates of these interacting particle systems to their mean-field limit as the number of particles tends to infinity. Furthermore, we apply these techniques and examine their effectiveness as methods of Bayesian approximation for quantifying predictive uncertainty in ReLU networks.


Convergence of Preconditioned Hamiltonian Monte Carlo on Hilbert Spaces

arXiv.org Machine Learning

Hamiltonian Monte Carlo (HMC) is a a Markov Chain Monte Carlo (MCMC) method for sampling from complex probability measures whose normalizing constant is unknown. It originated under the name "Hybrid Monte Carlo" in Duane et al. (1987) in statistical physics. The'target' measure that HMC can sample from has the form dπ(q) exp( Φ(q)) dq, (1) i.e. the target measure has a positive density with respect to the Lebesgue measure dq. The means that the density of π w.r.t.