Skip to yearly menu bar Skip to main content


Poster

A Better Way to Pre-Train Deep Boltzmann Machines

Russ Salakhutdinov · Geoffrey E Hinton

Harrah’s Special Events Center 2nd Floor

Abstract:

We describe how the pre-training algorithm for Deep Boltzmann Machines (DBMs) is related to the pre-training algorithm for Deep Belief Networks and we show that under certain conditions, the pre-training procedure improves the variational lower bound of a two-hidden-layer DBM. Based on this analysis, we develop a different method of pre-training DBMs that distributes the modelling work more evenly over the hidden layers. Our results on the MNIST and NORB datasets demonstrate that the new pre-training algorithm allows us to learn better generative models.

Live content is unavailable. Log in and register to view live content