2009
DOI: 10.1002/nme.2776
|View full text |Cite
|
Sign up to set email alerts
|

An efficient coarse‐grained parallel algorithm for global–local multiscale computations on massively parallel systems

Abstract: The existing global-local multiscale computational methods, using finite element discretization at both the macro-scale and micro-scale, are intensive both in terms of computational time and memory requirements and their parallelization using domain decomposition methods incur substantial communication overhead, limiting their application. We are interested in a class of explicit global-local multiscale methods whose architecture significantly reduces this communication overhead on massively parallel machines.… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
10
0

Year Published

2011
2011
2021
2021

Publication Types

Select...
6

Relationship

1
5

Authors

Journals

citations
Cited by 15 publications
(10 citation statements)
references
References 34 publications
0
10
0
Order By: Relevance
“…The resulting Krylov subspace basis and upper Hessenberg matrix are used to initialize the preconditioner in Equation (55). The macroscale computation is performed sequentially while the microscale computations are performed in parallel using a logical hierarchical topology [38] on the IBM Blue Gene/L supercomputing platform at Rensselaer Polytechnic Institute's (RPI) Computational Center for Nanotechnology Innovations (CCNI).…”
Section: Methodsmentioning
confidence: 99%
“…The resulting Krylov subspace basis and upper Hessenberg matrix are used to initialize the preconditioner in Equation (55). The macroscale computation is performed sequentially while the microscale computations are performed in parallel using a logical hierarchical topology [38] on the IBM Blue Gene/L supercomputing platform at Rensselaer Polytechnic Institute's (RPI) Computational Center for Nanotechnology Innovations (CCNI).…”
Section: Methodsmentioning
confidence: 99%
“…Only the group leader (or master) can communicate with other masters of different groups. Rahul and Suvranu De proved that this two‐level grouping is superior over a naive parallelization in which each sub‐domain is treated by one CPU.…”
Section: Revisit To Two‐level Grouping Parallel Algorithmmentioning
confidence: 99%
“…Only the group leader (or master) can communicate with other masters of different groups. Rahul and Suvranu De [11] proved that this two-level grouping is superior over a naive parallelization in which each sub-domain is treated by one CPU. Figure 3 compares a simple parallelization without grouping and the two-level algorithm, with emphasis on the communication channel.…”
Section: Revisit To Two-level Grouping Parallel Algorithmmentioning
confidence: 99%
See 1 more Smart Citation
“…The issue of computational complexity is addressed using parallel implementation strategies, reduced order modeling at the coarse scale using high order (i.e., plate and shell) theories or reduced order modeling at the fine scales to efficiently evaluate the microscale response, as well as a combination of these three approaches. Parallelization of the computational homogenization [12,14,15] is natural and domain decomposition is readily applicable due to the local character of the microscale boundary value problems that are typically evaluated at the integration points of the macroscale grid. Model reduction at the coarse scale is achieved by exploiting the characteristics of the macroscopic domain.…”
Section: Introductionmentioning
confidence: 99%