Proceedings of the 15th ACM SIGPLAN Symposium on Principles and Practice of Parallel Programming 2010
DOI: 10.1145/1693453.1693476
|View full text |Cite
|
Sign up to set email alerts
|

Scalable communication protocols for dynamic sparse data exchange

Abstract: Many large-scale parallel programs follow a bulk synchronous parallel (BSP) structure with distinct computation and communication phases. Although the communication phase in such programs may involve all (or large numbers) of the participating processes, the actual communication operations are usually sparse in nature. As a result, communication phases are typically expressed explicitly using point-to-point communication operations or collective operations. We define the dynamic sparse data-exchange (DSDE) pro… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
49
0

Year Published

2010
2010
2020
2020

Publication Types

Select...
4
2
2

Relationship

1
7

Authors

Journals

citations
Cited by 44 publications
(50 citation statements)
references
References 30 publications
1
49
0
Order By: Relevance
“…A solution to this problem is to perform step 6 using a variant of the nonblocking consensus algorithm described by Höfler et al [9]. All sends and receives are tested for completion, a nonblocking barrier posted once the tests are passed, and incoming messages probed for with MPI Iprobe, all in a loop which only exits if the nonblocking barrier is reached by all participating processes.…”
Section: ) Smentioning
confidence: 99%
See 1 more Smart Citation
“…A solution to this problem is to perform step 6 using a variant of the nonblocking consensus algorithm described by Höfler et al [9]. All sends and receives are tested for completion, a nonblocking barrier posted once the tests are passed, and incoming messages probed for with MPI Iprobe, all in a loop which only exits if the nonblocking barrier is reached by all participating processes.…”
Section: ) Smentioning
confidence: 99%
“…This variant differs slightly from Höfler's -tests for completion take place before probing. This is because the probe (and subsequent communication) is usually unnecessary, and both MPI Iprobe and MPI Ssend do have overhead (see [9] for some measurements pertaining to MPI Ssend). , is zero in either reference frame (this would not be true if differentiating across the interface, but it is so for a change of (inertial) reference frame), and…”
Section: ) Smentioning
confidence: 99%
“…For example, in preparing the data exchange necessary to obtain column data in sparse matrix-vector multiplication, this would be the number of processors having entries on that column [7]. We use the fanout, denoted f , to describe the bounds on our algorithms' costs.…”
Section: Fanout Boundmentioning
confidence: 99%
“…At the same time, the popularity of distributed parallel programming systems that implement high degrees of dynamic behavior, such as asynchronous tasks [1], work stealing [4,11,12], and messagedriven execution [8,18,19], are increasing. Unlike in bulk synchronous parallel programs, and even in dynamic data exchanges within BSP programs [7], there is often no clear global indication of when some particular distributed computation is complete. Thus, they instead rely on termination detection algorithms to provide that indication.…”
Section: Introductionmentioning
confidence: 99%
“…Hoefler et al [14] study this problem and its variations, and design new, specialized collectives to address this problem. Hoefler and Träff [15] also make the case for better support of "sparse" communication patterns within MPI, where the sparsity refers to the number of communicating processes.…”
Section: Collective Communication Optimizationmentioning
confidence: 99%