Goto

Collaborating Authors

 uha




A Generating the Hamiltonian bound

Neural Information Processing Systems

Algorithm 6 Generating the (non-differentiable) Hamiltonian AIS variational bound.Sample z It can be observed that tuning β and q (z) lead to the largest gains in performance. T uning more parameters leads to significantly better results. Table 4: ELBO on the test set (higher is better). For K = 1 both methods reduce to plain VI. For HMC, we use half of the budget for the warm-up phase and half to draw samples.



Langevin Diffusion Variational Inference

Geffner, Tomas, Domke, Justin

arXiv.org Artificial Intelligence

Many methods that build powerful variational distributions based on unadjusted Langevin transitions exist. Most of these were developed using a wide range of different approaches and techniques. Unfortunately, the lack of a unified analysis and derivation makes developing new methods and reasoning about existing ones a challenging task. We address this giving a single analysis that unifies and generalizes these existing techniques. The main idea is to augment the target and variational by numerically simulating the underdamped Langevin diffusion process and its time reversal. The benefits of this approach are twofold: it provides a unified formulation for many existing methods, and it simplifies the development of new ones. In fact, using our formulation we propose a new method that combines the strengths of previously existing algorithms; it uses underdamped Langevin transitions and powerful augmentations parameterized by a score network. Our empirical evaluation shows that our proposed method consistently outperforms relevant baselines in a wide range of tasks.


MCMC Variational Inference via Uncorrected Hamiltonian Annealing

Geffner, Tomas, Domke, Justin

arXiv.org Machine Learning

Given an unnormalized target distribution we want to obtain approximate samples from it and a tight lower bound on its (log) normalization constant log Z. Annealed Importance Sampling (AIS) with Hamiltonian MCMC is a powerful method that can be used to do this. Its main drawback is that it uses non-differentiable transition kernels, which makes tuning its many parameters hard. We propose a framework to use an AIS-like procedure with Uncorrected Hamiltonian MCMC, called Uncorrected Hamiltonian Annealing. Our method leads to tight and differentiable lower bounds on log Z. We show empirically that our method yields better performances than other competing approaches, and that the ability to tune its parameters using reparameterization gradients may lead to large performance improvements.