2019
DOI: 10.48550/arxiv.1902.06853
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

On the Impact of the Activation Function on Deep Neural Networks Training

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
14
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 12 publications
(14 citation statements)
references
References 0 publications
0
14
0
Order By: Relevance
“…The initialization of weights is a significant step that can impact the learning Fig. 3 The bidirectional RNN structure over the time process and, ultimately, the model's convergence (Hayou et al, 2019).…”
Section: Model Training Phasementioning
confidence: 99%
“…The initialization of weights is a significant step that can impact the learning Fig. 3 The bidirectional RNN structure over the time process and, ultimately, the model's convergence (Hayou et al, 2019).…”
Section: Model Training Phasementioning
confidence: 99%
“…The trainability of a neural network indicates how effective it can be optimized using gradient descent (Burkholz & Dubatovka, 2019;Hayou et al, 2019;Shin & Karniadakis, 2020). Although some heavy networks can theoretically represent complex functions, they not necessarily can be effectively trained by gradient descent.…”
Section: Trainability By Condition Number Of Ntkmentioning
confidence: 99%
“…), even the former could equip a larger number of parameters. The trainability of a neural network studies how effective it can be optimized by gradient descent [44], [45], [46].…”
Section: Trainabilitymentioning
confidence: 99%