Proceedings of the 24th ACM International Conference on Supercomputing 2010
DOI: 10.1145/1810085.1810091
|View full text |Cite
|
Sign up to set email alerts
|

Overlapping communication and computation by using a hybrid MPI/SMPSs approach

Abstract: -

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
61
0

Year Published

2012
2012
2020
2020

Publication Types

Select...
4
3
3

Relationship

0
10

Authors

Journals

citations
Cited by 71 publications
(61 citation statements)
references
References 15 publications
0
61
0
Order By: Relevance
“…In the case of the distributed OmpSs+MPI model, it combines dataflow execution with the message passing model providing significant performance benefits. It hides the communication latencies and achieves higher performance compared to MPI only model [25].…”
Section: Task Replicationmentioning
confidence: 99%
“…In the case of the distributed OmpSs+MPI model, it combines dataflow execution with the message passing model providing significant performance benefits. It hides the communication latencies and achieves higher performance compared to MPI only model [25].…”
Section: Task Replicationmentioning
confidence: 99%
“…State-of-the-art techniques that combine distributed-and shared-memory programming models [80], as well as many PGAS approaches [6,24,47,48], have demon-strated the potential benefits of combining both levels of parallelism [81,82,39,83], including increased communication-computation overlap [84,85], improved memory utilization [86,87], power optimization [88] and effective use of accelerators [89,90,91,92]. The hybrid MPI and thread model, such as MPI and OpenMP, can take advantage of those optimized shared-memory algorithms and data structures.…”
Section: Chapter 4 Habanero-c Runtime Communication Systemmentioning
confidence: 99%
“…In general, the system ensures high utilization, since some blocks on each processor should always have work. Others implementations dynamically interleave the work performed on various blocks, either by introducing task parallelism to HPL [25] or by spawning many light-weight threads in UPC [16].…”
Section: B Programming Paradigmmentioning
confidence: 99%