Revisiting Bayesian Autoencoders with MCMC
Chandra, Rohitash, Jain, Mahir, Maharana, Manavendra, Krivitsky, Pavel N.
–arXiv.org Artificial Intelligence
Bayes' theorem is used as foundation Autoencoders are a family of unsupervised learning methods for inference in Bayesian neural networks, and Markov that use neural network architectures and learning algorithms chain Monte Carlo (MCMC) sampling methods [25] are used to learn a lower-dimensional representation (encoding) for constructing the posterior distribution. Variational inference of the data, which can then be used to reconstruct a representation [26] is another way to approximate the posterior distribution, close to the original input. They thus facilitate dimensionality which approximates an intractable posterior distribution by a reduction for prediction and classification [1, 2], and have tractable one. This makes it particularly suited to large data been successfully applied to image classification [3, 4], face sets and models, and so it has been popular for autoencoders recognition [5, 6], geoscience and remote sensing [7], speechbased and neural networks [13, 27].
arXiv.org Artificial Intelligence
Apr-12-2021
- Country:
- Asia
- India (0.04)
- Japan (0.04)
- Middle East > Republic of Türkiye
- Ankara Province > Ankara (0.04)
- Oceania > Australia
- New South Wales > Kensington (0.04)
- Asia
- Genre:
- Overview (0.67)
- Research Report (0.64)
- Industry:
- Energy (0.48)