2010
DOI: 10.1002/cpe.1627
|View full text |Cite
|
Sign up to set email alerts
|

Performance‐based parallel loop self‐scheduling using hybrid OpenMP and MPI programming on multicore SMP clusters

Abstract: SUMMARYParallel loop self-scheduling on parallel and distributed systems has been a critical problem and it is becoming more difficult to deal with in the emerging heterogeneous cluster computing environments. In the past, some self-scheduling schemes have been proposed as applicable to heterogeneous cluster computing environments. In recent years, multicore computers have been widely included in cluster systems. However, previous researches into parallel loop self-scheduling did not consider certain aspects o… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
8
0

Year Published

2011
2011
2017
2017

Publication Types

Select...
5
1

Relationship

1
5

Authors

Journals

citations
Cited by 10 publications
(8 citation statements)
references
References 30 publications
0
8
0
Order By: Relevance
“…The advantage of static scheduling is that it is easy to implement and has no extra scheduling overhead during runtime, but it may cause load unbalancing and thereby reduce computational efficiency (Zhu et al, 2004;Plank et al, 2007). Dynamic scheduling adjusts the schedule during execution and is especially suitable for situations where the number of steps, or the workload of each step, is undetermined (Yang et al, 2011). Although dynamic scheduling is more suitable for load balancing among parallel processors, the scheduling overhead should be executed at runtime and it needs extra processing time.…”
Section: Lps and Swmmentioning
confidence: 98%
See 2 more Smart Citations
“…The advantage of static scheduling is that it is easy to implement and has no extra scheduling overhead during runtime, but it may cause load unbalancing and thereby reduce computational efficiency (Zhu et al, 2004;Plank et al, 2007). Dynamic scheduling adjusts the schedule during execution and is especially suitable for situations where the number of steps, or the workload of each step, is undetermined (Yang et al, 2011). Although dynamic scheduling is more suitable for load balancing among parallel processors, the scheduling overhead should be executed at runtime and it needs extra processing time.…”
Section: Lps and Swmmentioning
confidence: 98%
“…This can be done by adjusting the chunk size. Therefore, these schemes cannot achieve load balancing in a CPU-GPU in an extremely heterogeneous environment (Yang et al, 2011).…”
Section: Lps and Swmmentioning
confidence: 99%
See 1 more Smart Citation
“…In our previous approaches [22,23,25,26], the implementation of automatic parallelization based on OpenMP currently supports both C and C++ codes. Based on the result of these researches, we enhanced and extended the research.…”
Section: Automatic Parallelizationmentioning
confidence: 99%
“…However, the size of each work queue is not determined based on the knowledge of loop and runtime environment in affinity scheduling. Some groups [9,10] have undertaken self-scheduling studies on particular architectures, considering the features of the system architecture. Our technique could be easily extended to these architectures.…”
Section: Background and Related Workmentioning
confidence: 99%