2016
DOI: 10.5194/gmd-9-3605-2016
|View full text |Cite
|
Sign up to set email alerts
|

Coarse-grained component concurrency in Earth system modeling: parallelizing atmospheric radiative transfer in the GFDL AM3 model using the Flexible Modeling System coupling framework

Abstract: Abstract. Climate models represent a large variety of processes on a variety of timescales and space scales, a canonical example of multi-physics multi-scale modeling. Current hardware trends, such as Graphical Processing Units (GPUs) and Many Integrated Core (MIC) chips, are based on, at best, marginal increases in clock speed, coupled with vast increases in concurrency, particularly at the fine grain. Multiphysics codes face particular challenges in achieving finegrained concurrency, as different physics and… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
34
0

Year Published

2017
2017
2020
2020

Publication Types

Select...
5
2

Relationship

2
5

Authors

Journals

citations
Cited by 19 publications
(36 citation statements)
references
References 37 publications
2
34
0
Order By: Relevance
“…The radiative fluxes in AM4.0 have some sensitivity to the radiative time step, as discussed in Balaji et al (). Comparing simulations with a radiative time step of 3 h (the AM3 value) and 30 min (the physics time step at which all radiatively active species are updated), the difference in the net TOA flux is substantial, ∼3 W m −2 .…”
Section: Radiationmentioning
confidence: 99%
See 1 more Smart Citation
“…The radiative fluxes in AM4.0 have some sensitivity to the radiative time step, as discussed in Balaji et al (). Comparing simulations with a radiative time step of 3 h (the AM3 value) and 30 min (the physics time step at which all radiatively active species are updated), the difference in the net TOA flux is substantial, ∼3 W m −2 .…”
Section: Radiationmentioning
confidence: 99%
“…As the current simulations are based on relatively short AMIP runs, we have chosen to run the $24 SYPD configuration for best use of the available compute resources, as this allows individual simulations to complete within 1-2 days. For the same reason, we have forgone further scaling optimizations such as those achieved by the ''concurrent radiation'' configuration of AM4 (Balaji et al, 2016). This again serves to increase SYPD but at additional cost in CHSY.…”
Section: Computational Efficiency and Work-flowmentioning
confidence: 99%
“…This component architecture of ESMs is quite diverse (Alexander and Easterbrook, 2015), but typically most include at least two such components set up to run concurrently, in a mode we term coarse-grained concurrency (Balaji et al, 2016). This raises issues of load balance, configuring components to execute in roughly the same amount of time, so no processors sit idle.…”
Section: Computational Performance Of Esmsmentioning
confidence: 99%
“…The coupling overhead must be taken into account in an ESM performance study. Coarse-grained concurrency may be increasingly prevalent in ESM architectures in the future, because of current hardware trends (Balaji et al, 2016). The parallel component layout of some typical ESMs is shown in Fig.…”
Section: Computational Performance Of Esmsmentioning
confidence: 99%
“…Creating such an abstraction 5 is challenging due to the introduction of new architectures and computing models such as Graphics Processing Units (GPUs) and many-core CPUs (See introduction in Balaji et al, 2016). The modelling framework must facilitate storage of information about the computational resources available so that components that can take advantage of certain specialised hardware will be able to do so.…”
Section: Models and Computing Infrastructurementioning
confidence: 99%