2019
DOI: 10.1016/j.eswa.2018.11.042
|View full text |Cite
|
Sign up to set email alerts
|

Linearized sigmoidal activation: A novel activation function with tractable non-linear characteristics to boost representation capability

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
16
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
9
1

Relationship

0
10

Authors

Journals

citations
Cited by 45 publications
(16 citation statements)
references
References 23 publications
0
16
0
Order By: Relevance
“…Figure 5A shows the importance of each hyperparameter obtained from random forest by regressing the model predictive accuracies (obtained by an inner crossvalidation) on all hyperparameter values combinations. Interestingly, the number of filters was overall the most relevant factor, whereas other factors such as learning rate, whose importance has been claimed in the literature as critically important (Maas et al, 2013;Bawa and Kumar, 2019;Feng and Lu, 2019), played only a minor role. We also observed that the effect of each hyperparameter depends on the layer, e.g.…”
Section: Hyperparameter Importancementioning
confidence: 96%
“…Figure 5A shows the importance of each hyperparameter obtained from random forest by regressing the model predictive accuracies (obtained by an inner crossvalidation) on all hyperparameter values combinations. Interestingly, the number of filters was overall the most relevant factor, whereas other factors such as learning rate, whose importance has been claimed in the literature as critically important (Maas et al, 2013;Bawa and Kumar, 2019;Feng and Lu, 2019), played only a minor role. We also observed that the effect of each hyperparameter depends on the layer, e.g.…”
Section: Hyperparameter Importancementioning
confidence: 96%
“…where x i denotes the inputs, ω i denotes the weights, Σ is the summing function, f is the activation function, and y is the output. There are three activation functions-tangent hyperbolic, sigmoid, and linear functions-in the ANNs that are used for research [31,32]. Of these, the sigmoid function is commonly applied to generalize an ANN, which predicts the probability as an output because probability exists only between 0 and 1.…”
Section: Multilayer Perceptronmentioning
confidence: 99%
“…Another "dynamic" function is proposed in [27], whose shape is learned by a linear regression model. In [28], two different variants are proposed: a "linear sigmoidal activation", which is a fixed structure function whose function coefficients are static, and its "dynamic" variant, named "adaptive linear sigmoidal activation", which can adapt itself according to the complexity of the given data. Two of the best performing functions are Swish [29], which is the combination of a sigmoid function and a trainable parameter, and the recent Mexican ReLU (MeLU) [30], which is a piecewise linear activation function that is the sum of PReLU and multiple Mexican hat functions.…”
Section: Literature Reviewsmentioning
confidence: 99%