A Better Way to Pretrain Deep Boltzmann Machines
–Neural Information Processing Systems
We describe how the pretraining algorithm for Deep Boltzmann Machines (DBMs) is related to the pretraining algorithm for Deep Belief Networks and we show that under certain conditions, the pretraining procedure improves the variational lower bound of a two-hidden-layer DBM. Based on this analysis, we develop a different method of pretraining DBMs that distributes the modelling work more evenly over the hidden layers. Our results on the MNIST and NORB datasets demonstrate that the new pretraining algorithm allows us to learn better generative models.
Neural Information Processing Systems
Mar-14-2024, 11:00:54 GMT
- Country:
- North America
- Canada > Ontario
- Toronto (0.29)
- United States > New York (0.04)
- Canada > Ontario
- North America
- Genre:
- Research Report > New Finding (0.48)
- Technology: