2007 IEEE International Conference on Cluster Computing 2007
DOI: 10.1109/clustr.2007.4629236
|View full text |Cite
|
Sign up to set email alerts
|

The design of MPI based distributed shared memory systems to support OpenMP on clusters

Abstract: OpenMP can be supported in cluster environments by using distributed shared memory (DSM) systems. A portable approach for building DSM systems is to layer it on MPI. With these goals in mind, this paper makes two contributions. The first is a discussion about two software DSM systems that we have implemented using MPI. One uses background polling threads while the other uses processes that are driven only by incoming MPI messages. Comparisons of the two approaches show the latter to be a more scalable architec… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2008
2008
2015
2015

Publication Types

Select...
2
2
2

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(4 citation statements)
references
References 13 publications
0
4
0
Order By: Relevance
“…The reason for these two programming paradigms lies in the nature of the memory subsystem of the target MPSoC, which is, in fact, a DSM platform. Hybrid OpenMP-MPI programming has proven beneficial for DSM systems in the HPC domain [17], and more in general to implement nested (multi-level) parallelism on NUMA systems [18]. Here, coarse-grained (task-level) parallelism can be mapped over different clusters communicating through MPI primitives, whereas fine-grained (loop-level) parallelism can be easily distributed within a cluster with OpenMP.…”
Section: Vertically-integrated Hw/sw Qos Supportmentioning
confidence: 99%
See 1 more Smart Citation
“…The reason for these two programming paradigms lies in the nature of the memory subsystem of the target MPSoC, which is, in fact, a DSM platform. Hybrid OpenMP-MPI programming has proven beneficial for DSM systems in the HPC domain [17], and more in general to implement nested (multi-level) parallelism on NUMA systems [18]. Here, coarse-grained (task-level) parallelism can be mapped over different clusters communicating through MPI primitives, whereas fine-grained (loop-level) parallelism can be easily distributed within a cluster with OpenMP.…”
Section: Vertically-integrated Hw/sw Qos Supportmentioning
confidence: 99%
“…Overall, the platform consists of several clusters, interconnected with a top-level NoC. We assume a Distributed Shared Memory model (DSM) [17], where all processors in the system can directly access all the shared memory modules. However, due to the presence of the hierarchical interconnection system, the access latencies are subject to Non-Uniform Memory Access (NUMA) effects, depending on the physical distance of the end points.…”
Section: Introductionmentioning
confidence: 99%
“…The total number of page faults (N f ) and the number of prefetches issued (N p ) are presented in thousands, and the number of useful prefetched pages is presented as a ratio to N f . The efficiency of each prefetch technique is calculated based on Equation (2).…”
Section: A Reduction Of Network Communicationsmentioning
confidence: 99%
“…The most widely used such paradigm, the shared memory programming model, can be facilitated on clusters via software Distributed Shared Memory (sDSM) systems, which have the advantage of both the programmability of shared memory programming models and the low cost of distributed memory architectures. sDSM systems designed with OpenMP compiler support are also known as cluster-enabled OpenMP systems [1], [2], [3], [4], [5].…”
Section: Introductionmentioning
confidence: 99%