2017
DOI: 10.1007/s00521-017-3210-6
|View full text |Cite
|
Sign up to set email alerts
|

A comparative performance analysis of different activation functions in LSTM networks for classification

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
65
0
4

Year Published

2018
2018
2024
2024

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 108 publications
(69 citation statements)
references
References 35 publications
0
65
0
4
Order By: Relevance
“…In addition, time-sequential can be stored through the recurrent weights of the network, and recurrent neurons can then reflect time sequences. Therefore, an RNN could estimate disturbances better than a conventional feedforward techniques [48,67].…”
Section: A Selecting An Rnn-based Controllermentioning
confidence: 99%
“…In addition, time-sequential can be stored through the recurrent weights of the network, and recurrent neurons can then reflect time sequences. Therefore, an RNN could estimate disturbances better than a conventional feedforward techniques [48,67].…”
Section: A Selecting An Rnn-based Controllermentioning
confidence: 99%
“…Because of the data passing through these non-linear transformations within the memory cell, it is common practise to not include any further activation at the nodes of the hidden layers. There is, however, discussion in the literature about choosing different activation functions for RNNs, for example Farzad et al (2019) investigate alternatives to the sigmoid activations at the LSTM input, forget and output gates. We add an activation at the final output layer only, and as our task is regression, we use a linear activation here as we do in the FFNN.…”
Section: Simulating Groundwater Levels-simple Modelmentioning
confidence: 99%
“…However, it is more expensive to compute than tanh; in other words, it has more complex derivatives. Additionally, its gradient sometimes yields extremely low/high values, such that we can consider it as a sigmoid on steroids [26]- [28].…”
Section: Softsign Activation Functionmentioning
confidence: 99%