Wasserstein Gradient Flow over Variational Parameter Space for Variational Inference

Nguyen, Dai Hai, Sakurai, Tetsuya, Mamitsuka, Hiroshi

arXiv.org Machine Learning 

Many machine learning problems involve the challenge of approximating an intractable target distribution, which might only be known up to a normalization constant. Bayesian inference is a typical example, where the intractable and unnormalized target distribution is a result of the product of the prior and likelihood functions (see [11, 18, 4]). Variational Inference (VI), a widely employed across various application domains, seeks to approximate this intractable target distribution by utilizing a variational distribution (see [3, 7, 20] and references therein). VI is typically formulated as an optimization problem, with the objective of maximizing the evidence lower bound objective (ELBO), which is equivalent to minimizing the Kullback-Leiber (KL) divergence between the variational distribution and the target distribution. The conventional method for maximizing the ELBO involves the use of gradient descent, such as black-box VI (BBVI, [16]). The gradient of the ELBO can be expressed as an expectation over the variational distribution, which is typically estimated by Monte Carlo samples from this distribution.

Duplicate Docs Excel Report

Title
None found

Similar Docs  Excel Report  more

TitleSimilaritySource
None found