2018 17th IEEE International Conference on Machine Learning and Applications (ICMLA) 2018
DOI: 10.1109/icmla.2018.00081
|View full text |Cite
|
Sign up to set email alerts
|

Improving L-BFGS Initialization for Trust-Region Methods in Deep Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
14
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
2
2

Relationship

1
7

Authors

Journals

citations
Cited by 19 publications
(14 citation statements)
references
References 17 publications
0
14
0
Order By: Relevance
“…to (52), we obtain the desired result in (44). For a given constant ǫ satisfying 0 < ǫ < 1, the iteration number needed to guarantee…”
Section: Convergence Resultsmentioning
confidence: 64%
See 1 more Smart Citation
“…to (52), we obtain the desired result in (44). For a given constant ǫ satisfying 0 < ǫ < 1, the iteration number needed to guarantee…”
Section: Convergence Resultsmentioning
confidence: 64%
“…S TOCHASTIC optimization algorithms have been extensively studied over decades and can be traced back to the epochal work [22], which have been widely employed in different areas, e.g., machine learning [23]- [25], [52], [53], power systems [51], wireless communication [5]- [7], and bioinformatics [50]. In particular, the classical stochastic approximation (SA) of the exact gradient, also known as stochastic gradient descent (SGD), has been widely applied to these stochastic optimization problems, where the gradient information is employed in finding the search direction.…”
Section: Introductionmentioning
confidence: 99%
“…Since this can potentially double the iteration complexity, an overlap batching strategy was proposed to reduce the computational cost in [3] and tested also in [4]. This strategy was further applied in [17,39]. Other stochastic quasi-Newton methods have been considered that employ a progressive batching approach in which the sample size is increased as the iteration progresses, see e.g.…”
Section: Literature Reviewmentioning
confidence: 99%
“…An initial approximation of the Hessian is obtained by solving an eigenvalue problem as proposed in Reference [65].…”
Section: Convolutional Networkmentioning
confidence: 99%