2016
DOI: 10.1016/j.neucom.2016.08.037
|View full text |Cite
|
Sign up to set email alerts
|

Bounded activation functions for enhanced training stability of deep neural networks on visual pattern recognition problems

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
66
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 143 publications
(66 citation statements)
references
References 39 publications
0
66
0
Order By: Relevance
“…Implementation of ReLU is possible with a thresholding of an activation map at zero. However, learning with ReLU in gradient descent‐based techniques is not possible because all gradient values will be zero when the activation values are zero . Therefore, in this work, LeakyReLU that is defined by FLeakyReLU={xitalicif0.5emx>00.01xitalicelse has been implemented to provide learning efficiently, even if the activation values are zero.…”
Section: Proposed Methodsmentioning
confidence: 99%
See 3 more Smart Citations
“…Implementation of ReLU is possible with a thresholding of an activation map at zero. However, learning with ReLU in gradient descent‐based techniques is not possible because all gradient values will be zero when the activation values are zero . Therefore, in this work, LeakyReLU that is defined by FLeakyReLU={xitalicif0.5emx>00.01xitalicelse has been implemented to provide learning efficiently, even if the activation values are zero.…”
Section: Proposed Methodsmentioning
confidence: 99%
“…However, learning with ReLU in gradient descentbased techniques is not possible because all gradient values will be zero when the activation values are zero. 24 Therefore, in this work, LeakyReLU 24 that is defined by…”
Section: Activation Function In the 3d Cnn Architecturementioning
confidence: 99%
See 2 more Smart Citations
“…But in model parameter training, weight matrices and bias vectors are updated using an error back-propagation algorithm whereas activation function is not. So the change of activation function is important for a neural network, which can speed up model training [17], enhance stability [18]. In this paper, we adopt the RNN model and modify its activation function to do NER task.…”
Section: Related Workmentioning
confidence: 99%