Practical Deep Learning with Bayesian Principles
Osawa, Kazuki, Swaroop, Siddharth, Jain, Anirudh, Eschenhagen, Runa, Turner, Richard E., Yokota, Rio, Khan, Mohammad Emtiyaz
Bayesian methods promise to fix many shortcomings of deep learning, but they are impractical and rarely match the performance of standard methods, let alone improve them. In this paper, we demonstrate practical training of deep networks with natural-gradient variational inference. By applying techniques such as batch normalisation, data augmentation, and distributed training, we achieve similar performance in about the same number of epochs as the Adam optimiser, even on large datasets such as ImageNet. Importantly, the benefits of Bayesian principles are preserved: predictive probabilities are well-calibrated and uncertainties on out-of-distribution data are improved. This work enables practical deep learning while preserving benefits of Bayesian principles. A PyTorch implementation will be available as a plug-and-play optimiser.
Jun-6-2019
- Country:
- Asia
- India > Jharkhand
- Dhanbad (0.04)
- Japan > Honshū
- Kantō > Tokyo Metropolis Prefecture > Tokyo (0.14)
- Middle East > Jordan (0.04)
- India > Jharkhand
- Europe
- Germany (0.04)
- Italy > Sardinia (0.04)
- Sweden > Stockholm
- Stockholm (0.04)
- United Kingdom > England
- Cambridgeshire > Cambridge (0.28)
- North America
- Canada > Ontario
- Toronto (0.14)
- United States > California
- San Francisco County > San Francisco (0.14)
- Santa Clara County > Palo Alto (0.04)
- Canada > Ontario
- Oceania > Australia
- New South Wales > Sydney (0.04)
- Asia
- Genre:
- Instructional Material > Course Syllabus & Notes (0.48)
- Research Report > New Finding (0.46)