Skip to yearly menu bar Skip to main content


Poster

A Better Way to Pre-Train Deep Boltzmann Machines

Russ Salakhutdinov · Geoffrey E Hinton

[ ]
[ PDF
2012 Poster

Abstract:

We describe how the pre-training algorithm for Deep Boltzmann Machines (DBMs) is related to the pre-training algorithm for Deep Belief Networks and we show that under certain conditions, the pre-training procedure improves the variational lower bound of a two-hidden-layer DBM. Based on this analysis, we develop a different method of pre-training DBMs that distributes the modelling work more evenly over the hidden layers. Our results on the MNIST and NORB datasets demonstrate that the new pre-training algorithm allows us to learn better generative models.

Chat is not available.