2007
DOI: 10.1016/j.amc.2006.06.114
|View full text |Cite
|
Sign up to set email alerts
|

Clustered Gauss–Huard algorithm for the solution of Ax=b

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2012
2012
2014
2014

Publication Types

Select...
4

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(2 citation statements)
references
References 13 publications
0
2
0
Order By: Relevance
“…Though they usually enjoy a faster convergence rate when compared to the Jacobi method, the standard implementation of these methods restricts their utilization for parallel computing due to the sequential nature of a requirement of using the most recently updated values in current iteration as soon as they are available. A variety of parallelization techniques 50–54 have been developed to parallelize the GS/SOR methods. In the DelPhi program, special techniques, namely the “checkerboard” ordering (also known as the “black-red” ordering) and contiguous memory mapping 25 , have been implemented previously in order to achieve better performance than the standard implements of the GS/SOR methods.…”
Section: Parallelization Schemementioning
confidence: 99%
“…Though they usually enjoy a faster convergence rate when compared to the Jacobi method, the standard implementation of these methods restricts their utilization for parallel computing due to the sequential nature of a requirement of using the most recently updated values in current iteration as soon as they are available. A variety of parallelization techniques 50–54 have been developed to parallelize the GS/SOR methods. In the DelPhi program, special techniques, namely the “checkerboard” ordering (also known as the “black-red” ordering) and contiguous memory mapping 25 , have been implemented previously in order to achieve better performance than the standard implements of the GS/SOR methods.…”
Section: Parallelization Schemementioning
confidence: 99%
“…The effective mapping of matrix elements to processors is the key factor to efficient implementation of the algorithm in parallel. In this paper, we employ the general distribution functions suggested by Al-Towaiq [19] because it embrace a wide range of matrix distribution functions, namely column/row blocked, column/row cyclic, column/row block cyclic, and column and row ( or 2D) block cyclic layouts [20][21][22]. The main issues of choosing a mapping are the load balancing and communication time.…”
Section: Mapping the Matrix Elements Onto The Processorsmentioning
confidence: 99%