2012 IEEE International Symposium on Performance Analysis of Systems &Amp; Software 2012
DOI: 10.1109/ispass.2012.6189222
|View full text |Cite
|
Sign up to set email alerts
|

Performance analysis of thread mappings with a holistic view of the hardware resources

Abstract: With the shift to chip multiprocessors, managing shared resources has become a critical issue in realizing their full potential. Previous research has shown that thread mapping is a powerful tool for resource management. However, the difficulty of simultaneously managing multiple hardware resources and the varying nature of the workloads have impeded the efficiency of thread mapping algorithms. To overcome the difficulties of simultaneously managing multiple resources with thread mapping, the interaction betwe… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
17
0

Year Published

2014
2014
2020
2020

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 29 publications
(17 citation statements)
references
References 27 publications
0
17
0
Order By: Relevance
“…Previous studies evaluate the impact of task mapping considering the communication [1], showing that it can influence several hardware resources. In shared memory environments, communication-based task mapping reduces execution time, cache misses and interconnection traffic [2].…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…Previous studies evaluate the impact of task mapping considering the communication [1], showing that it can influence several hardware resources. In shared memory environments, communication-based task mapping reduces execution time, cache misses and interconnection traffic [2].…”
Section: Related Workmentioning
confidence: 99%
“…execElInLevel[0] is not used. execElInLevel [1] contains the number of processing units. For positions i, such that 1 < i < nLevels, the value is the number of hardware objects on the respective architecture hierarchy level.…”
Section: A Description Of the Eagermap Algorithmmentioning
confidence: 99%
See 1 more Smart Citation
“…As parallel applications need to access shared data, the memory hierarchy presents challenges for mapping threads to cores, and data to NUMA nodes [24]. Threads that access a large amount of shared data should be mapped to cores that are close to each other in the memory hierarchy, while data should be mapped to the same NUMA node that the threads that access it are executing on [22].…”
Section: Introductionmentioning
confidence: 99%
“…When scheduling threads, their performance depends on which core each thread is executed, since the memory access performance varies depending on the core and the memory hierarchy [15]. These differences in the memory access latency between cores are due to shared cache levels, different bandwidths and latencies in the interconnections [38]. Regarding memory management, the memory hierarchy introduced non-uniform memory access (NUMA) due to multiple memory controllers in the system.…”
Section: Introductionmentioning
confidence: 99%