DocumentCode
671491
Title
Gaussian-Bernoulli deep Boltzmann machine
Author
Kyung Hyun Cho ; Raiko, Tapani ; Ilin, Alexander
Author_Institution
Dept. of Inf. & Comput. Sci., Aalto Univ., Aalto, Finland
fYear
2013
fDate
4-9 Aug. 2013
Firstpage
1
Lastpage
7
Abstract
In this paper, we study a model that we call Gaussian-Bernoulli deep Boltzmann machine (GDBM) and discuss potential improvements in training the model. GDBM is designed to be applicable to continuous data and it is constructed from Gaussian-Bernoulli restricted Boltzmann machine (GRBM) by adding multiple layers of binary hidden neurons. The studied improvements of the learning algorithm for GDBM include parallel tempering, enhanced gradient, adaptive learning rate and layer-wise pretraining. We empirically show that they help avoid some of the common difficulties found in training deep Boltzmann machines such as divergence of learning, the difficulty in choosing right learning rate scheduling, and the existence of meaningless higher layers.
Keywords
Boltzmann machines; Gaussian processes; gradient methods; learning (artificial intelligence); neural nets; scheduling; GDBM; GRBM; Gaussian-Bernoulli deep Boltzmann machine; Gaussian-Bernoulli restricted Boltzmann machine; adaptive learning rate; binary hidden neurons; deep Boltzmann machines; enhanced gradient; layer-wise pretraining; learning algorithm; learning rate scheduling; multiple layers; parallel tempering; Adaptation models; Approximation methods; Computational modeling; Data models; Neurons; Stochastic processes; Training;
fLanguage
English
Publisher
ieee
Conference_Titel
Neural Networks (IJCNN), The 2013 International Joint Conference on
Conference_Location
Dallas, TX
ISSN
2161-4393
Print_ISBN
978-1-4673-6128-6
Type
conf
DOI
10.1109/IJCNN.2013.6706831
Filename
6706831
Link To Document