2008
DOI: 10.1007/978-3-540-79561-2_8
|View full text |Cite
|
Sign up to set email alerts
|

STEP: A Distributed OpenMP for Coarse-Grain Parallelism Tool

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2012
2012
2016
2016

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 9 publications
(4 citation statements)
references
References 10 publications
0
4
0
Order By: Relevance
“…PIPS has proved over the years to be a fertile ground for the polyhedral model [21], data transformations [13], communication synthesis [4,7], compilation for distributed memory machines [9,23], ILP [37], code maintenance [29,6], program verification [25], scratchpad management [8], offload compilers [14,1,10], and task parallelism [18,33].…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…PIPS has proved over the years to be a fertile ground for the polyhedral model [21], data transformations [13], communication synthesis [4,7], compilation for distributed memory machines [9,23], ILP [37], code maintenance [29,6], program verification [25], scratchpad management [8], offload compilers [14,1,10], and task parallelism [18,33].…”
Section: Discussionmentioning
confidence: 99%
“…They support loop parallelization, with neither control nor call restrictions [13], and automatic distribution [9,23].…”
Section: Strong Pointsmentioning
confidence: 99%
“…In [8,11,19], authors propose to extend OpenMP with additional clauses necessary for streamization as in our tool. Nevertheless, the most similar tools are proposed in [4,5] and [16]. Both, are source-to-source compilers as our tool, the first based on Cetus [7] and the second on PIPS [1] generating solutions that could be compared to ours.…”
Section: Related Workmentioning
confidence: 99%
“…However, the code developed with these approaches is limited to shared memory systems. In order to overcome this limitation, several tools that execute multi-threaded applications on distributed memory architectures have been proposed but, up to now, either their implementation is based on software translations to MPI [7] or it relies on Distributed Shared Memory (DSM) systems [8]. Another option is the use of a hybrid shared/distributed memory programming model combining MPI for internode communications and resort to a shared memory model to take advantage of intra-node parallelism [9].…”
Section: Related Workmentioning
confidence: 99%