2019 IEEE 12th International Conference on Cloud Computing (CLOUD) 2019
DOI: 10.1109/cloud.2019.00088
|View full text |Cite
|
Sign up to set email alerts
|

Fast and Lightweight Execution Time Predictions for Spark Applications

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
10
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
1

Relationship

1
5

Authors

Journals

citations
Cited by 8 publications
(10 citation statements)
references
References 4 publications
0
10
0
Order By: Relevance
“…Several studies and modeling exercises have been conducted on the scalability of batch analytics workloads, e.g. [34]- [41]. Since batch and streaming workloads function quite differently, we will discuss the relevant parts of these studies throughout the results section.…”
Section: Related Workmentioning
confidence: 99%
“…Several studies and modeling exercises have been conducted on the scalability of batch analytics workloads, e.g. [34]- [41]. Since batch and streaming workloads function quite differently, we will discuss the relevant parts of these studies throughout the results section.…”
Section: Related Workmentioning
confidence: 99%
“…Amannejad et al [28] proposed an approach for Spark execution time prediction with less prior executions of the applications based on Amdahl's law [15]. This approach is capable of predicting the execution time within a short period.…”
Section: Related Workmentioning
confidence: 99%
“…One of the limitations of this work is that they validated this approach only with a single node cluster, not on a real cluster environment. Amannejad and Shah extended their previous work [28] and proposed an alternative model called PERIDOT [29] for quick execution time prediction with limited cluster resource settings and a small subset of input data. They analysed the logs from both of the executions and checked the internal dependencies between the internal stages.…”
Section: Related Workmentioning
confidence: 99%
“…Amannejad et al [13] proposed a model to quick prediction runtime a job when little cluster resources are available. They used logs from two executions of an application with small sample data and different resource settings and explore the accuracy of the predictions for other resource allocation settings and input data sizes then expanded their model to the Spark waves.…”
Section: Background and Related Workmentioning
confidence: 99%
“…(27), Figs. 11,12,13,14 shows the error rate of the estimated runtime by the Exponential Averaging method and actual runtime in WordCount, Tera-Sort, and Inverted index in the different number of nodes. The results show that the error rate is less than 5%.…”
Section: Evaluation Of Estimating Runtime Of a Job That Has Already Bmentioning
confidence: 99%