2004
DOI: 10.1007/978-3-540-24644-2_13
|View full text |Cite
|
Sign up to set email alerts
|

Evaluating the Impact of Programming Language Features on the Performance of Parallel Applications on Cluster Architectures

Abstract: Abstract.We evaluate the impact of programming language features on the performance of parallel applications on modern parallel architectures, particularly for the demanding case of sparse integer codes. We compare a number of programming languages (Pthreads, OpenMP, MPI, UPC) on both shared and distributed-memory architectures. We find that language features can make parallel programs easier to write, but cannot hide the underlying communication costs for the target parallel architecture. Powerful compiler an… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

2
10
0

Year Published

2004
2004
2016
2016

Publication Types

Select...
4
3
1

Relationship

2
6

Authors

Journals

citations
Cited by 14 publications
(13 citation statements)
references
References 6 publications
2
10
0
Order By: Relevance
“…Poor performance on distributed memory is consistent with previous UPC evaluation [11]. Program designs that assume efficient access of shared variables do not scale well in systems with higher latency.…”
Section: Parallel Performance On Distributed Memorysupporting
confidence: 80%
“…Poor performance on distributed memory is consistent with previous UPC evaluation [11]. Program designs that assume efficient access of shared variables do not scale well in systems with higher latency.…”
Section: Parallel Performance On Distributed Memorysupporting
confidence: 80%
“…A previous study on the comparison of OpenMP, MPI, and Pthreads [16] focused on performance for sparse integer codes with irregular remote memory accesses. Other recent papers [17,18] conduct a comparison of OpenMP versus MPI on a specific architecture, the IBM SP3 NH2, for a set of NAS benchmark applications (FT, CG, MG).…”
Section: Related Workmentioning
confidence: 99%
“…The FT benchmark is designed to aggressively overlap communication with computation [3], and FT-pencils is a variant of the benchmark that issues smaller messages for better overlap. The implementation of CG is described in [4], and gups is a version of the HPCS RandomAccess benchmark that uses bulk communication. Cfd is an application that solves the time dependent Euler equations for computational fluid flow in a rectangular computational domain, with the high level data structures and algorithms implemented in UPC.…”
Section: Effectiveness Of Communication Aggregationmentioning
confidence: 99%