2015
DOI: 10.1142/s0218001415510064
|View full text |Cite
|
Sign up to set email alerts
|

Deep Belief Network Training Improvement Using Elite Samples Minimizing Free Energy

Abstract: to create a powerful generative model using training data. In this paper we present an improvement in a common method that is usually used in training of RBMs. The new method uses free energy as a criterion to obtain elite samples from generative model. We argue that these samples can more accurately compute gradient of log probability of training data. According to the results, an error rate of 0.99% was achieved on MNIST test set. This result shows that the proposed method outperforms the method presented in… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2015
2015
2023
2023

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 21 publications
(6 citation statements)
references
References 14 publications
0
6
0
Order By: Relevance
“…Although all model parameters are changed in each step, PCD can receive good samples from model distribution with a few Gibbs sampling steps because the model parameters change slightly. As an improvement of PCD, FEPCD is based on Free Energy to generate better samples (Hinton, 2012; Keyvanrad and Homayounpour, 2015). The selection criteria for the optimal chain based on the free energy of visible layer sample is as followswhere F ( v ) is the free energy.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Although all model parameters are changed in each step, PCD can receive good samples from model distribution with a few Gibbs sampling steps because the model parameters change slightly. As an improvement of PCD, FEPCD is based on Free Energy to generate better samples (Hinton, 2012; Keyvanrad and Homayounpour, 2015). The selection criteria for the optimal chain based on the free energy of visible layer sample is as followswhere F ( v ) is the free energy.…”
Section: Methodsmentioning
confidence: 99%
“…To overcome the shortcoming, a criterion for goodness of a chain called free energy in persistent contrastive divergence (FEPCD) can ensure the network model to obtain better chain selection in sampling learning, which improves the quality and efficiency of gradient approximation (Hinton, 2012). As a result, the approximation and classification ability of DBN model increases along with FEPCD (Keyvanrad and Homayounpour, 2015). However, the sampling methods of DBN used in the up-to-date bearing fault diagnosis applications are mainly traditional CD or PCD, which may lead to a gradual decline in learning ability of DBN in the long-term training process.…”
Section: Introductionmentioning
confidence: 99%
“…Once RBM is learned using Contrastive Divergence (CD) algorithm [19], the DBN is able to initialize the weights of feed forward back-propagation neural network then it is used for classification to predict the image model. RBM can be learnt better when the predictive model is used before the step sampling in Gibs before collecting statistical step in learning rule but for the purposes of pre-training.…”
Section: Contrastive Divergencementioning
confidence: 99%
“…It has produced the state-of-the-art results on recognition and classification tasks [10]. On the other hand, typical classification methods used for speech recognition include hidden Markov model (HMM) [14], Gaussian Mixture Model (GMM) [15], artificial neural networks such as recurrent neural network (RNN) [16], support vector machine (SVM) [17,18], and the fuzzy cognitive map network [19]. These methods are confronted with the complicated decision boundary of the classification.…”
Section: Introductionmentioning
confidence: 99%
“…Whereas CD has some disadvantages and is not exact, other methods are proposed in REM. One of these methods is PCD that is very popular [13] and another method is FEPCD that has been proposed by authors in [14].…”
Section: Deep Belief Network (Dbns) and Restricted Boltzmann Macmentioning
confidence: 99%