2012
DOI: 10.1007/s00211-012-0496-2
|View full text |Cite
|
Sign up to set email alerts
|

Reorthogonalized block classical Gram–Schmidt

Abstract: A new reorthogonalized block classical Gram-Schmidt algorithm is proposed that factorizes a full column rank matrix A into A = QR where Q is left orthogonal (has orthonormal columns) and R is upper triangular and nonsingular.With appropriate assumptions on the diagonal blocks of R, the algorithm, when implemented in floating point arithmetic with machine unit ε M , produces Q and R such thatThe resulting bounds also improve a previous bound by Giraud et al. [Num. Math., 101(1):87-100, 2005] on the CGS2 algori… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
63
0

Year Published

2015
2015
2024
2024

Publication Types

Select...
6
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 23 publications
(64 citation statements)
references
References 15 publications
1
63
0
Order By: Relevance
“…4 The final result B(i, j) is computed by launching another CUDA kernel to perform another binary reduction among the thread blocks. Our implementation is designed to reduce the number of synchronizations among the threads while relying on the CUDA runtime and the parameter tuning to exploit the data locality.…”
Section: Figures 14(a) and 14(b) Then The Final Partial Results B(i J)mentioning
confidence: 99%
See 1 more Smart Citation
“…4 The final result B(i, j) is computed by launching another CUDA kernel to perform another binary reduction among the thread blocks. Our implementation is designed to reduce the number of synchronizations among the threads while relying on the CUDA runtime and the parameter tuning to exploit the data locality.…”
Section: Figures 14(a) and 14(b) Then The Final Partial Results B(i J)mentioning
confidence: 99%
“…2 Previously, the blocked variants of TSQR have been studied [1,2,4]. To generate n + 1 orthonormal basis vectors, our CA-GMRES and CA-Lanczos [25] use block orthogonalization followed by TSQR with a step size of s, where the step size is equivalent to the block size in the blocked algorithm to orthogonalize n + 1 vectors (e.g., n = 60 and s = 15 in our experiments).…”
Section: Error Analysismentioning
confidence: 99%
“…For the purpose of the block reorthogonalization, the block CGS (BCGS) algorithm [27], which is also a variant of the CGS algorithm, is proposed. To improve the orthogonality of the resulting vectors, the BCGS algorithm with reorthogonalization (BCGS2 algorithm) [28] is preferable. The BCGS2 algorithm including the QR factorization can be implemented using mainly the matrix multiplications.…”
Section: Block Inverse Iteration Algorithm With Reorthogonalizationmentioning
confidence: 99%
“…To improve the accuracy of computed factors one can introduce the implementation with iterative refinement, where the Cholesky-like factorization is applied first to (1) and Ω (1) . The factor Q (1) is then obtained as Q (1) (2) to get the factors R (2) and Ω (2) .…”
Section: Cholesky-like Factorization Of Symmetric Indefinite Matricesmentioning
confidence: 99%
“…The factor Q (1) is then obtained as Q (1) (2) to get the factors R (2) and Ω (2) . The resulting factors are then (1) .…”
Section: Cholesky-like Factorization Of Symmetric Indefinite Matricesmentioning
confidence: 99%