ACM/IEEE SC 2002 Conference (SC'02) 2002
DOI: 10.1109/sc.2002.10015
|View full text |Cite
|
Sign up to set email alerts
|

Executing Multiple Pipelined Data Analysis Operations in the Grid

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
32
0

Year Published

2004
2004
2013
2013

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 27 publications
(32 citation statements)
references
References 22 publications
0
32
0
Order By: Relevance
“…If the computations of a given stage are independent from one data set to another, two consecutive computations (different data sets) for the same stage can be mapped onto distinct processors, thus reducing the period for the processing of this stage. Such a stage can be replicated, using the terminology of Subhlok and Vondran [27,28] and of the DataCutter team [6,7,26]. This corresponds to the dealable stages of Cole [11].…”
Section: Working Out An Examplementioning
confidence: 99%
See 1 more Smart Citation
“…If the computations of a given stage are independent from one data set to another, two consecutive computations (different data sets) for the same stage can be mapped onto distinct processors, thus reducing the period for the processing of this stage. Such a stage can be replicated, using the terminology of Subhlok and Vondran [27,28] and of the DataCutter team [6,7,26]. This corresponds to the dealable stages of Cole [11].…”
Section: Working Out An Examplementioning
confidence: 99%
“…Such workflows operate on a collection of data sets that are executed in [26,27,31]. Each data set is input to the application graph and traverses it until its processing is complete.…”
Section: Introductionmentioning
confidence: 99%
“…Benoit and Robert [11] study the theoretical complexity of latency and throughput optimization of pipeline and fork graphs with replication and data-parallelism under the assumptions of linear clustering and round-robin processing of input data items. In [3], Spencer et al presented the Filter Copy Pipeline (FCP) scheduling algorithm for optimizing latency and throughput of data analysis application DAGs on heterogeneous resources. FCP computes the number of copies of each task that is necessary to meet the aggregate production rate of its predecessors and maps the copies to processors that yield their least completion time.…”
Section: Related Workmentioning
confidence: 99%
“…This section evaluates the performance of WMSH against previously proposed schemes: Filter Copy Pipeline (FCP) [3] and EXPERT (EXploiting Pipeline Execution undeR Time constraints) [2], and FCP-e and EXPERT-e, their modified versions. When FCP fails to utilize all processors and does not meet the throughput requirement T , FCP-e recursively calls FCP on the remaining processors until T is satisfied or all processors are used.…”
Section: Performance Analysismentioning
confidence: 99%
See 1 more Smart Citation