2015
DOI: 10.1049/el.2015.2175
|View full text |Cite
|
Sign up to set email alerts
|

Efficient out‐of‐GPU memory strategies for solving matrix equation generated by method of moments

Abstract: The numerical solution of the dense linear complex valued system of equations generated by the method of moments (MoMs) generally proceeds by factoring the impedance matrix into LU decomposition. Depending on available hardware resources, the LU algorithm can be executed either on sequential or parallel computers. A straightforward parallel implementation of LU factorisation does not yield a well distributed workload, and therefore it is the computationally most expensive step of the MoMs process, especially w… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2

Citation Types

0
4
0

Year Published

2016
2016
2019
2019

Publication Types

Select...
4
2

Relationship

1
5

Authors

Journals

citations
Cited by 7 publications
(4 citation statements)
references
References 7 publications
0
4
0
Order By: Relevance
“…The CPU sequential procedures for computing the elements of the system matrix have been mapped to parallel GPU platform as described in [20]. To enable the solution of relatively large-size problems with MoM-generated matrix exceeding the amount of memory available on the device, a hybrid out-of-GPU memory CULA-panel-based LU decomposition algorithm have been implemented [21]. The algorithm proceeds iteratively with the following two distinct phases: i) panel factorization, and ii) the update of the trailing submatrix.…”
Section: Hybrid Cpu/gpu Implementationmentioning
confidence: 99%
See 1 more Smart Citation
“…The CPU sequential procedures for computing the elements of the system matrix have been mapped to parallel GPU platform as described in [20]. To enable the solution of relatively large-size problems with MoM-generated matrix exceeding the amount of memory available on the device, a hybrid out-of-GPU memory CULA-panel-based LU decomposition algorithm have been implemented [21]. The algorithm proceeds iteratively with the following two distinct phases: i) panel factorization, and ii) the update of the trailing submatrix.…”
Section: Hybrid Cpu/gpu Implementationmentioning
confidence: 99%
“…1 explains how the CPU/GPU computations are organized. The interested reader is referred to [20], [21] and [23] for more details. …”
Section: Hybrid Cpu/gpu Implementationmentioning
confidence: 99%
“…During the last decades, much effort has been devoted to develop numerical methods for solving matrix equations. () Yao et al studied the solutions of the matrix equation AX = B with respect to semi‐tensor product . In Chiang, first by using the Kronecker product, some useful results of the solvability of the Sylvester‐like matrix equation AX + f ( X ) B = C were presented, and then the closed‐form solutions of this matrix equation were provided.…”
Section: Introductionmentioning
confidence: 99%
“…Compared with a single core, a speedup of 15 can be obtained. In [20], an efficient out-of-GPU memory scheme for solving matrix equations generated by MoM was presented, and it also obtains a good speedup on a single CPU/GPU platform. In [21,22], an out-of-core scheme between RAM and hard-disk drives (HDD) using RWGs and higher-order basis functions (HOBs) was adopted to break the limitation of RAM and improve the capability of parallel MoM to solve larger EM problems.…”
Section: Introductionmentioning
confidence: 99%