The 2011 International Joint Conference on Neural Networks 2011
DOI: 10.1109/ijcnn.2011.6033557
|View full text |Cite
|
Sign up to set email alerts
|

Analysis and improvement of multiple optimal learning factors for feed-forward networks

Abstract: are analyzed. The use of optimal diagonal transformation matrices on the net function vector is proved to be equivalent to training the MLP using multiple optimal learning factors (MOLF). A method for linearly compressing large ill-conditioned MOLF Hessian matrices into smaller wellconditioned ones is developed. This compression approach is shown to be equivalent to using several hidden units per learning factor. The technique is extended to large networks. In simulations, the proposed algorithm performs almos… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2014
2014
2024
2024

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(2 citation statements)
references
References 36 publications
(31 reference statements)
0
2
0
Order By: Relevance
“…The derived features described in Sections 3.2 to 3.4 are computed on these data samples and then used to train a binary LSTM classifier. Stochastic learning methods such as Adam Optimizer [29] are used for training due to their faster convergence and lesser data requirement at each iteration than batch-based training algorithms [30][31][32]. The hyperparameters used to train the classifier, such as the batch size, no of LSTM units and no of epochs, are selected after observing the accuracy curves for train and test data.…”
Section: Medical Condition Detectionmentioning
confidence: 99%
“…The derived features described in Sections 3.2 to 3.4 are computed on these data samples and then used to train a binary LSTM classifier. Stochastic learning methods such as Adam Optimizer [29] are used for training due to their faster convergence and lesser data requirement at each iteration than batch-based training algorithms [30][31][32]. The hyperparameters used to train the classifier, such as the batch size, no of LSTM units and no of epochs, are selected after observing the accuracy curves for train and test data.…”
Section: Medical Condition Detectionmentioning
confidence: 99%
“…Such a high-level of computational complexity is mostly associated with the training sequence, which generally involves at least a single division and numerous multiplications. In this work, we use Levenberg-Marquardt backpropagation training method, which is a popular scheme for equalisation in VLC systems [11]. As for ANN equalisers, their performance is dependent on the number of taps N t and neurons N n .…”
Section: Experimental Test Setupmentioning
confidence: 99%