2019 22nd International Conference on Process Control (PC19) 2019
DOI: 10.1109/pc.2019.8815057
|View full text |Cite
|
Sign up to set email alerts
|

Comparison of ReLU and linear saturated activation functions in neural network for universal approximation

Abstract: Activation functions used in hidden layers directly affect the possibilities for describing nonlinear systems using a feedforward neural network. Furthermore, linear based activation functions are less computationally demanding than their nonlinear alternatives. In addition, feedforward neural networks with linear based activation functions can be advantageously used for control of nonlinear systems, as shown in previous authors' publications. This paper aims to compare two types of linear based functions-symm… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
11
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 41 publications
(14 citation statements)
references
References 9 publications
0
11
0
Order By: Relevance
“…Regarding the layers, it is stated in [65] that a single hidden layer is sufficient to train the ANN to match "any non-linearity". Following this statement, a network with one hidden layer is chosen as the minimum requirement and, for comparison, a network with two layers is implemented as well (the maximum depth of the hidden layers considered here).…”
Section: Function Name Activation Function φ(X)mentioning
confidence: 99%
“…Regarding the layers, it is stated in [65] that a single hidden layer is sufficient to train the ANN to match "any non-linearity". Following this statement, a network with one hidden layer is chosen as the minimum requirement and, for comparison, a network with two layers is implemented as well (the maximum depth of the hidden layers considered here).…”
Section: Function Name Activation Function φ(X)mentioning
confidence: 99%
“…We used the Rectified Linear Unit (ReLU) activation function for all hidden layers. It is the default activation function for many types of neural networks, because a model that uses it is easier to train and often achieves better performance [27]. For the single-neuron output layer, we used the Linear Activation (LA) function.…”
Section: Model Developmentmentioning
confidence: 99%
“…Since last experiment already resulted in a significantly larger model (requiring around 6 Mbytes of memory) and before continuing exploring more complex NN architectures, we noted that there still could be room for improvement without adding extra layers. Considering the fact that when working with several layers with ReLU activation we have a significant risk of dying neurons harming our performance [27], this in turn can lead to underfitting. Batch normalization (BN) is one of the best ways to handle this issue [33].…”
Section: Models Comparisonmentioning
confidence: 99%
“…A bias value is added to this sum. The application of the AF specifically introduces nonlinearities, aiming to emulate the way humans analyze real-world data [25]. The connection between one node and another is performed by a number (weight), which can be positive (one node excites another) or negative (one node inhibits another).…”
Section: Deep-learning Systemsmentioning
confidence: 99%