1994
DOI: 10.1007/bfb0025872
|View full text |Cite
|
Sign up to set email alerts
|

Array distribution in data-parallel programs

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
11
0

Year Published

1996
1996
1999
1999

Publication Types

Select...
5
1

Relationship

1
5

Authors

Journals

citations
Cited by 23 publications
(11 citation statements)
references
References 10 publications
0
11
0
Order By: Relevance
“…Both a fine-grain and the optimal coarse-grain static partitioning will be compared with the dynamic partitioning. 8 In the current implementation, loop peeling is not performed on the actual code. As previously mentioned in Section 4.2, the single additional startup redistribution due to not peeling will not be significant in comparison to the execution of the loop (containing a dynamic count of 600 redistributions).…”
Section: -D Alternating Direction Implicit (Adi) Iterative Methodsmentioning
confidence: 98%
See 1 more Smart Citation
“…Both a fine-grain and the optimal coarse-grain static partitioning will be compared with the dynamic partitioning. 8 In the current implementation, loop peeling is not performed on the actual code. As previously mentioned in Section 4.2, the single additional startup redistribution due to not peeling will not be significant in comparison to the execution of the loop (containing a dynamic count of 600 redistributions).…”
Section: -D Alternating Direction Implicit (Adi) Iterative Methodsmentioning
confidence: 98%
“…Nodes in the ADG represent data-parallel computations while edges represent the flow of data. The ADG itself is formed in a divide-and-conquer approach using heuristics to approximately solve a combinatorial minimization problem at each step, taking into account both redistribution costs as well as all candidate distributions, to determine where to partition the program into subphases [8]. Candidates are formed by first identifying the extents, or iteration space, of all objects in the program resulting in a number of clusters.…”
Section: Static Partitioningmentioning
confidence: 99%
“…The identification of program segments in which data can be statically mapped and the accurate modeling of the potential remapping costs make the dynamic data-mapping problem harder than the static problem. The smallest possible statically mapped program regions may be single statements [Chatterjee et al 1993;Chatterjee et al 1994;Philippsen 1995], loop nests [Ayguadé et al 1994;Anderson and Lam 1993;Lee and Tsai 1993;Ning et al 1995;Palermo and Banerjee 1995;Tandri and Abdelrahman 1997], or groups of statements or loop nests for which it can be shown that remapping between them can never be profitable [Sheffler et al 1996]. More recent work tries to make the mapping decisions independent of the particular program structure [Kelly and Pugh 1996].…”
Section: Dynamic Mappingsmentioning
confidence: 99%
“…The starting time was taken 7 Raw measures for transpositions are displayed. Tables 5 and 6 show the transposition times for various matrix sizes and distributions.…”
Section: Experimental Conditionsmentioning
confidence: 99%
“…Data remapping and replication often need to be combined: A parallel matrix multiplication accesses a whole row and column of data to compute each single target element, hence the need to remap data with some replication for parallel execution. Moreover, automatic data layout tools 24,7] suggest data remappings between computation phases. Thus handling data remappings e ciently is an important issue for high performance computing.…”
Section: Introductionmentioning
confidence: 99%