2017
DOI: 10.1080/00207160.2017.1280156
|View full text |Cite
|
Sign up to set email alerts
|

GPU-accelerated preconditioned GMRES method for two-dimensional Maxwell's equations

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
7
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
4
2

Relationship

3
3

Authors

Journals

citations
Cited by 14 publications
(7 citation statements)
references
References 16 publications
0
7
0
Order By: Relevance
“…PETSc implements several efficient subroutines to conduct the ILU factorizations and preconditioning operations for block sparse matrices with various block sizes. In the first implementation, we consider performing (8) using PETSc's API on CPUs and copying the preconditioned vector to the device memory at every iteration of Algorithm 1. e developed function is listed in Algorithm 4. e function accepts four input variables as ksp, vecv, vecz, and dv, where ksp is an instance of the PETSc's KSP structure, vecv and vecz, encapsulated by the PETSc's Vec structure in the host memory, are the left-and right-side vectors in equation (8), and dv is the device pointer to the left side vector on the GPU. e function outputs the pointer to the right-side vector.…”
Section: Preconditioningmentioning
confidence: 99%
See 1 more Smart Citation
“…PETSc implements several efficient subroutines to conduct the ILU factorizations and preconditioning operations for block sparse matrices with various block sizes. In the first implementation, we consider performing (8) using PETSc's API on CPUs and copying the preconditioned vector to the device memory at every iteration of Algorithm 1. e developed function is listed in Algorithm 4. e function accepts four input variables as ksp, vecv, vecz, and dv, where ksp is an instance of the PETSc's KSP structure, vecv and vecz, encapsulated by the PETSc's Vec structure in the host memory, are the left-and right-side vectors in equation (8), and dv is the device pointer to the left side vector on the GPU. e function outputs the pointer to the right-side vector.…”
Section: Preconditioningmentioning
confidence: 99%
“…Yang [6,7] developed the preconditioned GMRES algorithm by parallelizing the ILU(0), ILUT, block ILU(k), and triangular solves on GPUs. Gao et al [8] proposed an efficient GPU kernel on the sparse matrixvector multiplication (SpMV) in GMRES and applied the optimized GMRES to solving the two-dimensional Maxwell's equations. He et al [9] presented an efficient GPU implementation of the GMRES with ILU preconditioners for solving large linear dynamic systems.…”
Section: Introductionmentioning
confidence: 99%
“…As mentioned above, there has been some work about parallel SPAI preconditioners on GPU. For example, Gao et al followed Chow's work, and use a sparse approximate inverse of A as the preconditioner in their work . However, they do not give any implementation description.…”
Section: Introductionmentioning
confidence: 99%
“…There exists some work on accelerating the construction of preconditioners with GPU. For example, the incomplete Cholesky factorization preconditioners on GPU, 25,26 the incomplete LU factorization preconditioners on GPU, [27][28][29][30][31][32][33] the FSAI preconditioners on GPU, [34][35][36][37] and the SPAI preconditioners on GPU, [38][39][40] and the preconditioners that consist of an incomplete factorization, followed by an approximate inversion of the incomplete factors on GPU. [41][42][43][44] Especially, some public libraries such as CUSPARSE, 45 CUSP, 46 and ViennaCL 47,48 also include parallel preconditioners on GPU.…”
Section: Introductionmentioning
confidence: 99%
“…The dense matrix‐vector multiplication routine performs one of y:=Ax2.56804ptfalse(GEMVfalse)2.56804ptor2.56804pty:=ATx2.56804ptfalse(GEMV‐Tfalse), where ARm×n is a dense matrix, and y and x are vectors. It has proven to be of particular importance in computational science and has been successfully applied in various fields …”
Section: Introductionmentioning
confidence: 99%