2021
DOI: 10.1007/s00779-021-01587-4
|View full text |Cite
|
Sign up to set email alerts
|

Achieving generalization of deep learning models in a quick way by adapting T-HTR learning rate scheduler

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 8 publications
(4 citation statements)
references
References 16 publications
0
4
0
Order By: Relevance
“…Another insight for future work is solving the challenge of selecting the optimal learning rate, as excessively high values can cause divergence, while overly low values hinder fast convergence, significantly affecting domain adaptation quality. Moreover, monotonously decreasing learning rates may trap neural networks in suboptimal states, possibly causing premature stagnation [53]. Implementing an attenuation strategy, which involves dynamically adjusting the learning rate based on gradient observations [53] or utilizing reinforcement learning agents to adapt it to observed states [54], is essential for robustness.…”
Section: Future Work and Conclusionmentioning
confidence: 99%
See 1 more Smart Citation
“…Another insight for future work is solving the challenge of selecting the optimal learning rate, as excessively high values can cause divergence, while overly low values hinder fast convergence, significantly affecting domain adaptation quality. Moreover, monotonously decreasing learning rates may trap neural networks in suboptimal states, possibly causing premature stagnation [53]. Implementing an attenuation strategy, which involves dynamically adjusting the learning rate based on gradient observations [53] or utilizing reinforcement learning agents to adapt it to observed states [54], is essential for robustness.…”
Section: Future Work and Conclusionmentioning
confidence: 99%
“…Moreover, monotonously decreasing learning rates may trap neural networks in suboptimal states, possibly causing premature stagnation [53]. Implementing an attenuation strategy, which involves dynamically adjusting the learning rate based on gradient observations [53] or utilizing reinforcement learning agents to adapt it to observed states [54], is essential for robustness. Although not employed in our current research due to its relatively uncommon use and tool limitations, we acknowledge its importance.…”
Section: Future Work and Conclusionmentioning
confidence: 99%
“…The cyclical learning rate includes base learning rate and max learning rate, cycle, step size, batch size, batch, or iteration. Base learning rate and max learning rate define the boundaries of a range, where the learning rate will fluctuate (Vidyabharathi et al, 2021). The value of base learning rate and max learning rate was set as 10 -8 and 10, respectively.…”
Section: Cyclical Learning Ratementioning
confidence: 99%
“…Abid et al [6] presented an optimized homomorphic encryption Chinese remainder theorem with a Rivest-Shamir-Adleman algorithm for an efficient and secure communication in the current digital world. Correspondingly, the seventh article "Achieving generalization of deep learning models in a quick way by adapting T-HTR learning rate scheduler" by Vidyabharathi et al [7] implemented a new HTR learning rate scheduler (toggle between hyperbolic tangent decay and triangular mode with restarts) in order to identify an optimal learning rate in deep neural network. Compared to the traditional methods, the proposed learning rate scheduler consumed minimal time for every epoch in many cases.…”
Section: Submission and Summary Of Contributionsmentioning
confidence: 99%