2018
DOI: 10.48550/arxiv.1810.04539
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Nonlinear Acceleration of Momentum and Primal-Dual Algorithms

Abstract: We describe a convergence acceleration scheme for multistep optimization algorithms. The extrapolated solution is written as a nonlinear average of the iterates produced by the original optimization algorithm. Our scheme does not need the underlying fixed-point operator to be symmetric, hence handles e.g. algorithms with momentum terms such as Nesterov's accelerated method, or primal-dual methods. The weights are computed via a simple linear system and we analyze performance in both online and offline modes. W… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
10
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
5
2

Relationship

2
5

Authors

Journals

citations
Cited by 7 publications
(10 citation statements)
references
References 18 publications
0
10
0
Order By: Relevance
“…Therefore, accelerating the convergence rate of the fixed point iteration has attracted considerable interest. Anderson Acceleration (AA) (Anderson, 1965) is among the most popular techniques to speed up the convergence of fixed point iteration (Bollapragada et al, 2018;Scieur et al, 2018;Walker and Ni, 2011). The key idea behind the AA strategy is to maintain the history of h recent iterations and predict the new iteration by using a linear combination of this history where the weights are extracted by solving an optimization problem.…”
Section: Fixed Point Iteration and Anderson Accelerationmentioning
confidence: 99%
“…Therefore, accelerating the convergence rate of the fixed point iteration has attracted considerable interest. Anderson Acceleration (AA) (Anderson, 1965) is among the most popular techniques to speed up the convergence of fixed point iteration (Bollapragada et al, 2018;Scieur et al, 2018;Walker and Ni, 2011). The key idea behind the AA strategy is to maintain the history of h recent iterations and predict the new iteration by using a linear combination of this history where the weights are extracted by solving an optimization problem.…”
Section: Fixed Point Iteration and Anderson Accelerationmentioning
confidence: 99%
“…The quality of the bound (in particular, its eventual convergence to 0) crucially depends on P (T ) . Using the Crouzeix conjecture (Crouzeix, 2004) Bollapragada et al (2018 managed to bound P (T ) , with P a polynomial:…”
Section: Anderson Extrapolation For Nonsymmetric Iteration Matricesmentioning
confidence: 99%
“…As we recall below, results on Anderson acceleration mainly concern fixed-point iterations with symmetric iteration matrices T , and results concerning non-symmetric iteration matrices are weaker (Bollapragada et al, 2018). Poon and Liang (2020, Thm 6.4) do not assume that T is symmetric, but only diagonalizable, which is still a strong requirement.…”
Section: Introductionmentioning
confidence: 99%
“…However, AA and optimization algorithms have been developed quite independently and only limited connections were discovered and studied [16,18]. Very recently, the technique has started to gain a significant interest in the optimization community (see, e.g., [47,46,5,53,19,39]). Specifically, a series of papers [47,46,5] adapt AA to accelerate several classical algorithms for unconstrained optimization; [53] studies a variant of AA for non-expansive operators; [19] proposes an application of AA to Douglas-Rachford splitting; and [39] uses AA to improve the performance of the ADMM method.…”
Section: Related Workmentioning
confidence: 99%
“…Very recently, the technique has started to gain a significant interest in the optimization community (see, e.g., [47,46,5,53,19,39]). Specifically, a series of papers [47,46,5] adapt AA to accelerate several classical algorithms for unconstrained optimization; [53] studies a variant of AA for non-expansive operators; [19] proposes an application of AA to Douglas-Rachford splitting; and [39] uses AA to improve the performance of the ADMM method. There is also an emerging literature on applications of AA in machine learning [23,28,20,36].…”
Section: Related Workmentioning
confidence: 99%