2019 International Conference on High Performance Computing &Amp; Simulation (HPCS) 2019
DOI: 10.1109/hpcs48598.2019.9188055
|View full text |Cite
|
Sign up to set email alerts
|

Feedback-Based Resource Allocation for Batch Scheduling of Scientific Workflows

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
18
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
6

Relationship

1
5

Authors

Journals

citations
Cited by 15 publications
(18 citation statements)
references
References 25 publications
0
18
0
Order By: Relevance
“…However, estimation methods for the usage of other resources, e.g. main memory or network bandwidth, often apply similar techniques [43,45].…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…However, estimation methods for the usage of other resources, e.g. main memory or network bandwidth, often apply similar techniques [43,45].…”
Section: Related Workmentioning
confidence: 99%
“…Both approaches have in common that they need at least comprehensive knowledge about execution times of all tasks on all available nodes. However, these values are not available in advance but must be determined either by asking users for estimates [18,22,23], by analyzing historical traces [35,36,42], or by using some form of online learning [43,45]. Lotaru aims to estimate the runtime for all task-node pairs in a cluster to enable the use of existing scheduling methods in real-world systems.…”
Section: Scheduling Workflow Tasks Onto Heterogeneous Clustersmentioning
confidence: 99%
See 1 more Smart Citation
“…With the help of resource managers like Kubernetes [7], Slurm [8], HTCondor [9] or Yarn [10] tasks are scheduled to the infrastructure components. This is necessary since workflows can consist of a large number of tasks, which are often recurring [6], [11], [12]. This specific pattern, combined with large amounts of data, leads to long runtimes on clusters such as several days or weeks [11], [12].…”
Section: Introduction Scientific Workflow Management Systems (Swms) Likementioning
confidence: 99%
“…This is necessary since workflows can consist of a large number of tasks, which are often recurring [6], [11], [12]. This specific pattern, combined with large amounts of data, leads to long runtimes on clusters such as several days or weeks [11], [12]. Therefore, data-parallel computing on large scale-outs is needed to increase the throughput and to ensure that the analysis executes in a certain timeframe.…”
Section: Introduction Scientific Workflow Management Systems (Swms) Likementioning
confidence: 99%