2008
DOI: 10.1007/978-3-540-79561-2_6
|View full text |Cite
|
Sign up to set email alerts
|

Micro-benchmarks for Cluster OpenMP Implementations: Memory Consistency Costs

Abstract: In this work we present a microbenchmark methodology for assessing the overheads associated with nested parallelism in OpenMP. Our techniques are based on extensions to the well known EPCC microbenchmark suite that allow measuring the overheads of OpenMP constructs when they are effected in inner levels of parallelism. The methodology is simple but powerful enough and has enabled us to gain interesting insight into problems related to implementing and supporting nested parallelism. We measure and compare a num… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
7
0

Year Published

2008
2008
2012
2012

Publication Types

Select...
4

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(8 citation statements)
references
References 18 publications
0
7
0
Order By: Relevance
“…Besides the difference of page having a home or not, for both home-based and home-less sDSMs, i) write notices are sent and invalidations are performed at synchronization points, ii) the actual data retrieving happens only when the invalid page is accessed, which avoids unnecessary data transfers, and iii) data fetching to keep memory consistent is the major overhead [13], [12].…”
Section: B Sdsm Memory Consistency Model and Costsmentioning
confidence: 99%
See 2 more Smart Citations
“…Besides the difference of page having a home or not, for both home-based and home-less sDSMs, i) write notices are sent and invalidations are performed at synchronization points, ii) the actual data retrieving happens only when the invalid page is accessed, which avoids unnecessary data transfers, and iii) data fetching to keep memory consistent is the major overhead [13], [12].…”
Section: B Sdsm Memory Consistency Model and Costsmentioning
confidence: 99%
“…As a result, sDSMs do not show satisfactory performance for many applications [11], [12]. According to some recent research about characterizing and modeling this overhead for different page-based sDSM systems [13], [12], servicing page faults, which involves data transfers between processes through a network interface, has been identified as the most significant overhead of sDSM. An effective prefetch strategy can reduce the overhead by i) overlapping computation and communication, and ii) amortising communication latencies and to achieve better inter-node bandwidth.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Besides the difference that a page has a home or not, for both home-based and home-less sDSMs, i) write notices are sent and invalidations are performed at synchronization points, ii) the actual data retrieving happens only when the invalid page is accessed, which avoids unnecessary data transfers, and iii) data fetching to keep memory consistent is the major overhead [7], [8].…”
Section: B Sdsm Memory Consistency Modelmentioning
confidence: 99%
“…Although cluster-enabled OpenMP systems show acceptable performance on some scientific applications, for example selected workloads from the Gaussian Quantum Chemistry code (Gaussian03) [6], they still suffer from high memory consistency costs [7], [8]. The major part of the cost is inter-node data transfer, due to the memory consistency work within global synchronization points [9].…”
Section: Introductionmentioning
confidence: 99%