2008 IEEE Fourth International Conference on eScience 2008
DOI: 10.1109/escience.2008.64
|View full text |Cite
|
Sign up to set email alerts
|

SWARM: Scheduling Large-Scale Jobs over the Loosely-Coupled HPC Clusters

Abstract: Abstract-Compute-intensive scientific applications are heavily reliant on the available quantity of computing resources. The Grid paradigm provides a large scale computing environment for scientific users. However, conventional Grid job submission tools do not provide a high-level job scheduling environment for these users across multiple institutions. For extremely large number of jobs, a more scalable job scheduling framework that can leverage highly distributed clusters and supercomputers is required. In th… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2008
2008
2020
2020

Publication Types

Select...
4
2
1

Relationship

1
6

Authors

Journals

citations
Cited by 8 publications
(3 citation statements)
references
References 17 publications
0
3
0
Order By: Relevance
“…Condor-G in turn uses GRAM2 for its job submission and data staging. We have spun this work off into the Swarm project [31]. Data services in the portal include fault models from the QuakeTables fault database [32], GPS archival data from Geophysical Resources Web Service [33], and real-time data from the California Real-Time Network [34].…”
Section: Quakesimmentioning
confidence: 99%
“…Condor-G in turn uses GRAM2 for its job submission and data staging. We have spun this work off into the Swarm project [31]. Data services in the portal include fault models from the QuakeTables fault database [32], GPS archival data from Geophysical Resources Web Service [33], and real-time data from the California Real-Time Network [34].…”
Section: Quakesimmentioning
confidence: 99%
“…This supercomputer log shows that idle resources are not used during this waiting time, causing inefficient resource waste. Therefore, the efficiency of resource allocation has to be evaluated by the scheduling algorithm [9]. It can be seen that optimization can be performed by applying the backfilling.…”
Section: Introductionmentioning
confidence: 99%
“…Table 2 shows the data/computation flow of these three basic execution units, along with examples. (Raicu, Zhao et al 2007) and SWARM (Pallickara and Pierce 2008) all provide similar functionality by scheduling large numbers of individual maps/jobs. Applications which can utilize a "reduction" or an "aggregation" operation can use both phases of the MapReduce model and, depending on the "associativity" and "transitivity" nature of the reduction operation, multiple reduction phases can be applied to enhance the parallelism.…”
Section: Programming Modelsmentioning
confidence: 99%