The 2013 International Joint Conference on Neural Networks (IJCNN) 2013
DOI: 10.1109/ijcnn.2013.6706831
|View full text |Cite
|
Sign up to set email alerts
|

Gaussian-Bernoulli deep Boltzmann machine

Abstract: In this paper, we study a model that we call Gaussian-Bernoulli deep Boltzmann machine (GDBM) and discuss potential improvements in training the model. GDBM is designed to be applicable to continuous data and it is constructed from Gaussian-Bernoulli restricted Boltzmann machine (GRBM) by adding multiple layers of binary hidden neurons. The studied improvements of the learning algorithm for GDBM include parallel tempering, enhanced gradient, adaptive learning rate and layer-wise pretraining. We empirically sho… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
56
0

Year Published

2015
2015
2024
2024

Publication Types

Select...
6
1

Relationship

2
5

Authors

Journals

citations
Cited by 94 publications
(56 citation statements)
references
References 13 publications
0
56
0
Order By: Relevance
“…We can also extend † the DRM so that it feeds realvalued data for x and/or y using the Gaussian scheme like Gaussian-Bernoulli RBM [22] or Gaussian-Bernoulli DBM [23]. In this scheme, when we want to feed realvalued x ∈ R I , we replace the x-related terms in Eq.…”
Section: Definition and Generative Proceduresmentioning
confidence: 99%
“…We can also extend † the DRM so that it feeds realvalued data for x and/or y using the Gaussian scheme like Gaussian-Bernoulli RBM [22] or Gaussian-Bernoulli DBM [23]. In this scheme, when we want to feed realvalued x ∈ R I , we replace the x-related terms in Eq.…”
Section: Definition and Generative Proceduresmentioning
confidence: 99%
“…It has, however, been known and will be shown in the experiments in this paper that training a DBM using this approach starting from randomly initialized parameters is not trivial [36,15,11]. The difficulty of training without any pretraining was illustrated in [36] and [15] by a lower log-likelihood achieved by a DBM trained without any pretraining.…”
Section: Training Deep Boltzmann Machinesmentioning
confidence: 99%
“…The difficulty of training without any pretraining was illustrated in [36] and [15] by a lower log-likelihood achieved by a DBM trained without any pretraining. Furthermore, the lack of proper initialization of the parameters was found to result in the upper-level hidden neurons not being able to capture any interesting features of an input data in [11].…”
Section: Training Deep Boltzmann Machinesmentioning
confidence: 99%
“…Even though an efficient learning algorithm was proposed for GRBM [7], training is still very sensitive to initialization and choice of learning parameters. Cho et al proposed an enhanced gradient learning algorithm for GRBM in [2]. Throughout the paper, a modified version of GRBM [3] is adopted, where the energy function is defined as…”
Section: Gaussian Restricted Boltzmann Machinesmentioning
confidence: 99%
“…ik and v (2) ik are additional parameters to model the pair-wise connections between two sets of visible neurons {x, y} and hidden neurons h. Instead of looking for the image transformation, we seek for the internal structure of texture information. Therefore, the same patch of image is fed to the two sets of visible neurons, that is x = y.…”
Section: Gaussian Gated Boltzmann Machinementioning
confidence: 99%