Proceedings of the 2022 SIAM Conference on Parallel Processing for Scientific Computing (PP) 2022
DOI: 10.1137/1.9781611977141.5
|View full text |Cite
|
Sign up to set email alerts
|

Performance of Low Synchronization Orthogonalization Methods in Anderson Accelerated Fixed Point Solvers

Abstract: Anderson Acceleration (AA) is a method to accelerate the convergence of fixed point iterations for nonlinear, algebraic systems of equations. Due to the requirement of solving a least squares problem at each iteration and a reliance on modified Gram-Schmidt for updating the iteration space, AA requires extra costly synchronization steps for global reductions. Moreover, the number of reductions in each iteration depends on the size of the iteration space. In this work, we introduce three low synchronization ort… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
1
1

Relationship

1
1

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 25 publications
0
2
0
Order By: Relevance
“…Krylov subspace methods for solving linear systems are often required for extreme-scale physics simulations on parallel machines with manycore accelerators. Their strong-scaling is limited by the number and frequency of global reductions in the form of MPI AllReduce operations and these communication patterns are expensive [13]. Lowsynchronization algorithms are based on the ideas of Ruhe [5], and are designed such that they require only one reduction per iteration to normalize each vector and apply projections.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Krylov subspace methods for solving linear systems are often required for extreme-scale physics simulations on parallel machines with manycore accelerators. Their strong-scaling is limited by the number and frequency of global reductions in the form of MPI AllReduce operations and these communication patterns are expensive [13]. Lowsynchronization algorithms are based on the ideas of Ruhe [5], and are designed such that they require only one reduction per iteration to normalize each vector and apply projections.…”
Section: Introductionmentioning
confidence: 99%
“…The low-synch modified Gram-Schmidt and GMRES algorithms described in Świrydowicz et al [3] improve parallel strong-scaling by employing one global reduction for each iteration, (see Lockhart et al [13]). A review of compact W Y Gram Schmidt algorithms and their computational costs is given in [4].…”
Section: Introductionmentioning
confidence: 99%