2011
DOI: 10.1016/j.parco.2011.05.006
|View full text |Cite
|
Sign up to set email alerts
|

Tuning collective communication for Partitioned Global Address Space programming models

Abstract: Partitioned Global Address Space (PGAS) languages offer programmers the convenience of a shared memory programming style combined with locality control necessary to run on large-scale distributed memory systems. Even within a PGAS language programmers often need to perform global communication operations such as broadcasts or reductions, which are best performed as collective operations in which a group of threads work together to perform the operation. In this paper we consider the problem of implementing col… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
21
0

Year Published

2012
2012
2019
2019

Publication Types

Select...
5
2
2

Relationship

2
7

Authors

Journals

citations
Cited by 33 publications
(21 citation statements)
references
References 9 publications
0
21
0
Order By: Relevance
“…MPI [22,17] and parallel programming languages such as UPC [19] provide optimized implementations of all-to-all collective operations. Most if not all of the existing implementations use multiple algorithms selected by message size.…”
Section: All-to-all Performancementioning
confidence: 99%
“…MPI [22,17] and parallel programming languages such as UPC [19] provide optimized implementations of all-to-all collective operations. Most if not all of the existing implementations use multiple algorithms selected by message size.…”
Section: All-to-all Performancementioning
confidence: 99%
“…Berkeley UPC provides optimized multi-threaded collective communication for multi-core systems [12] implemented with the GASNet communication library [4]. The GASNet multi-threaded collectives utilize automated-tuning to select the appropriate algorithm and adapt the system parameters to achieve best performance [10]. UPC applications can take advantage of these software engineering efforts in Berkeley UPC by simply calling UPC collective functions with threads.…”
Section: Managing Data Affinity For Numa Systemsmentioning
confidence: 99%
“…Support for asynchronous collectives is still under proposal. GASNet implements a range of non-blocking collective operations such as broadcast, reduce, scatter, gather, exchange and allgather with performance comparable to or better than MPI [16]. Aspen supports asynchronous collective communication [1].…”
Section: Related Workmentioning
confidence: 99%