2021
DOI: 10.1016/j.knosys.2021.107391
|View full text |Cite
|
Sign up to set email alerts
|

Oriented stochastic loss descent algorithm to train very deep multi-layer neural networks without vanishing gradients

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
23
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 52 publications
(23 citation statements)
references
References 14 publications
0
23
0
Order By: Relevance
“…Other techniques have incorporated federated learning and momentum [81] and used evolutionary algorithms [82], speculative approaches [83] and spiking neural network concepts [84,85]. Yet other techniques have focused on supporting deep networks [86], memory use optimization [87], bias factors [88,89] and initial condition sensitivity [90]. A recent technique, proposed by Zhang, et al [91], utilizes a combination of expert strategies and gradient descent for optimization.…”
Section: Gradient Descentmentioning
confidence: 99%
“…Other techniques have incorporated federated learning and momentum [81] and used evolutionary algorithms [82], speculative approaches [83] and spiking neural network concepts [84,85]. Yet other techniques have focused on supporting deep networks [86], memory use optimization [87], bias factors [88,89] and initial condition sensitivity [90]. A recent technique, proposed by Zhang, et al [91], utilizes a combination of expert strategies and gradient descent for optimization.…”
Section: Gradient Descentmentioning
confidence: 99%
“…To overcome the vanishing-gradient problem, Inas et al. [ 28 ] proposed OSLD, a new anti-vanishing back-propagated learning algorithm. Prior works [ 29 ] directly train SNNs using backpropagation; however, it is insufficient when training spiking architectures of the size of VGG-16 [ 7 ].…”
Section: Background and Related Workmentioning
confidence: 99%
“…A vanishing gradient causes the weight tuning to be extremely small, such that the network inefficiently corrects the computation errors for a close mapping between the inputs and responses. [37] f…”
Section: Artificial Neural Network (Ann)mentioning
confidence: 99%
“…A vanishing gradient causes the weight tuning to be extremely small, such that the network inefficiently corrects the computation errors for a close mapping between the inputs and responses. [ 37 ] fx=()exgoodbreak−ex()exgoodbreak+ex where e is the base of the natural logarithm constant and x is the input attribute. [ 36 ]…”
Section: Introductionmentioning
confidence: 99%