2020
DOI: 10.1007/978-981-15-5495-7_11
|View full text |Cite
|
Sign up to set email alerts
|

Review and Comparison of Commonly Used Activation Functions for Deep Neural Networks

Abstract: The primary neural networks decision-making units are activation functions. Moreover, they evaluate the output of networks neural node; thus, they are essential for the performance of the whole network. Hence, it is critical to choose the most appropriate activation function in neural networks calculation. Acharya et al. (2018) suggest that numerous recipes have been formulated over the years, though some of them are considered deprecated these days since they are unable to operate properly under some conditio… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
122
0
5

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
3
1

Relationship

0
10

Authors

Journals

citations
Cited by 240 publications
(127 citation statements)
references
References 28 publications
0
122
0
5
Order By: Relevance
“…The hyperbolic tangent activation function (tanh) is proposed here in the input layer, and the sigmoid function in the output layer. They are used frequently in feedforward nets, and are suitable for shallow networks as well as applications of prediction and mapping [38,41].…”
Section: Input Layer ( and )mentioning
confidence: 99%
“…The hyperbolic tangent activation function (tanh) is proposed here in the input layer, and the sigmoid function in the output layer. They are used frequently in feedforward nets, and are suitable for shallow networks as well as applications of prediction and mapping [38,41].…”
Section: Input Layer ( and )mentioning
confidence: 99%
“…Positive elements are preserved in ReLU and all the negative elements are discarded by fixing the corresponding activations to 0. When a negative input is supplied, ReLU function produces a 0, and when a positive input is supplied, it outputs a 1 [33].…”
Section: Activation Functions 411 Rectified Linear Unit Functionmentioning
confidence: 99%
“…The only difference is that Sigmoid lies between 0 and 1, whereas Tanh lies between 1 and −1. One of the main advantages of Sigmoid and Tanh activation functions is that the activations (i.e., the values in the nodes of the network, not the gradients) may not explode during the learning process, since their output range is bounded (Feng and Lu, 2019;Szandała, 2020). However, it should be pointed out that each activation function has its own strengths and limitations and its performance may be different based on the network complexity and data structure (Nwankpa et al, 2018;Feng and Lu, 2019).…”
Section: Ann Prediction Performancementioning
confidence: 99%