Reinforcement Learning for Adaptive MCMC
Wang, Congye, Chen, Wilson, Kanagawa, Heishiro, Oates, Chris. J.
–arXiv.org Artificial Intelligence
A vast literature on algorithms, tips, and tricks is testament to the success of Markov chain Monte Carlo (MCMC), which remains the most popular approach to numerical approximation of probability distributions characterised up to an intractable normalisation constant. Yet the breadth of methodology also presents a difficulty in selecting an appropriate algorithm for a specific task. The goal of adaptive MCMC is to automate, as much as possible, the design of a fast-mixing Markov transition kernel. To achieve this, one alternates between observing the performance of the current transition kernel, and updating the transition kernel in a manner that is expected to improve its future performance (Andrieu and Thoms, 2008). Though the online adaptation of a Markov transition kernel in principle sacrifices the ergodicy of MCMC, there are several ways to prove that ergodicity is in fact retained if the transition kernel converges fast enough (in an appropriate sense) to a sensible limit.
arXiv.org Artificial Intelligence
May-22-2024
- Country:
- Asia
- China (0.04)
- Japan > Honshū
- Kantō > Kanagawa Prefecture (0.04)
- Europe > United Kingdom
- England > Cambridgeshire > Cambridge (0.04)
- Oceania > Australia
- New South Wales > Sydney (0.04)
- Asia
- Genre:
- Research Report (1.00)
- Industry:
- Education (0.67)
- Technology: