2020
DOI: 10.3390/app10051897
|View full text |Cite
|
Sign up to set email alerts
|

The Influence of the Activation Function in a Convolution Neural Network Model of Facial Expression Recognition

Abstract: The convolutional neural network (CNN) has been widely used in image recognition field due to its good performance. This paper proposes a facial expression recognition method based on the CNN model. Regarding the complexity of the hierarchic structure of the CNN model, the activation function is its core, because the nonlinear ability of the activation function really makes the deep neural network have authentic artificial intelligence. Among common activation functions, the ReLu function is one of the best of… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
95
0
10

Year Published

2020
2020
2022
2022

Publication Types

Select...
7
3

Relationship

0
10

Authors

Journals

citations
Cited by 233 publications
(105 citation statements)
references
References 34 publications
0
95
0
10
Order By: Relevance
“…Recently, several papers related to making a more adequate activation function have appeared, mainly comparing activation functions on large datasets. To illustrate some of important contributions, the influence of the activation function in the convolutional neural network (CNN) model is studied in [24], improving a ReLU activation by construction of a novel surrogate. Theoretical analysis about gradient instability as well as the fundamental explanation for the exploding/vanishing gradient and the performances of different activation functions are given in [13].…”
Section: Introductionmentioning
confidence: 99%
“…Recently, several papers related to making a more adequate activation function have appeared, mainly comparing activation functions on large datasets. To illustrate some of important contributions, the influence of the activation function in the convolutional neural network (CNN) model is studied in [24], improving a ReLU activation by construction of a novel surrogate. Theoretical analysis about gradient instability as well as the fundamental explanation for the exploding/vanishing gradient and the performances of different activation functions are given in [13].…”
Section: Introductionmentioning
confidence: 99%
“…A feature map is getting by performing convolution processes to the input image or prior features using a linear filter, merging a bias term. Then passing this feature map through a non-linear activation function such as Sigmoid [34] and Rectified Linear Unit (RELU) [35] . In contrast, the classifier base includes the dense layers combined with the activation layers to convert the feature maps to one dimension vectors to expedite the classification task using many neurons.…”
Section: Theory and Methodsmentioning
confidence: 99%
“…The current common activation functions include Sigmoid, TanHyperbolic (tanh), and Rectified Linear Unit (ReLU) functions. The ReLU function can greatly improve the performance of CNN, and its performance in training speed is also better than that of other functions [30]. The ReLU function is expressed as (15).…”
Section: Convolutional Neural Network Work (Cnn)mentioning
confidence: 99%