2020 IEEE International Conference on Cluster Computing (CLUSTER) 2020
DOI: 10.1109/cluster49012.2020.00042
|View full text |Cite
|
Sign up to set email alerts
|

Towards Data-Flow Parallelization for Adaptive Mesh Refinement Applications

Abstract: Adaptive Mesh Refinement (AMR) is a prevalent method used by distributed-memory simulation applications to adapt the accuracy of their solutions depending on the turbulent conditions in each of their domain regions. These applications are usually dynamic since their domain areas are refined or coarsened in various refinement stages during their execution. Thus, they periodically redistribute their workloads among processes to avoid load imbalance. Although the defacto standard for scientific computing in distr… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
8
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
3
2

Relationship

2
3

Authors

Journals

citations
Cited by 5 publications
(8 citation statements)
references
References 16 publications
(35 reference statements)
0
8
0
Order By: Relevance
“…The hybrid variants send/receive/write each boundary block face from a different task (separate messages). That is not the optimal configuration (the optimal is around eight faces per message) but provides very reasonable performance [19] and puts more pressure on the communication phases. We show the throughput speedup of the strong scaling on the upper part of Figure 11 and the parallel efficiency on the lower part.…”
Section: B Miniamrmentioning
confidence: 98%
See 3 more Smart Citations
“…The hybrid variants send/receive/write each boundary block face from a different task (separate messages). That is not the optimal configuration (the optimal is around eight faces per message) but provides very reasonable performance [19] and puts more pressure on the communication phases. We show the throughput speedup of the strong scaling on the upper part of Figure 11 and the parallel efficiency on the lower part.…”
Section: B Miniamrmentioning
confidence: 98%
“…MiniAMR features multiple phases of computation and communication interleaved, and then a refinement and load-balancing phase periodically. Previous works [24] [19] fully taskified its computation and communication phases and some parts of the refinement and load-balancing [19], using OmpSs-2 and TAMPI. We take that taskification as the base and port the communication phases to TAGASPI.…”
Section: B Miniamrmentioning
confidence: 99%
See 2 more Smart Citations
“…The set of experiments we used includes well-known benchmarks such as Cholesky, Dotproduct, MultiSAXPY, STREAM, NBody, NQueens, and the Gauss-Seidel solver for the Heat The benchmarks of this list can be categorized as (i) purely memory-bounded, such as Heat, Dotproduct, MultiSAXPY, and STREAM, (ii) purely compute-bounded, such as NBody and NQueens, and (iii) balanced, such as Cholesky, miniAMR, HPCCG, and LULESH. Furthermore, all of these benchmarks have been parallelized using tasks, as their task-based versions [15], [16], [17] offer competitive or better performance than the fork-join OpenMP counterpart. Finally, the evaluation is partitioned into two phases.…”
Section: A Experimental Setupmentioning
confidence: 99%