1996
DOI: 10.1006/jpdc.1996.0112
|View full text |Cite
|
Sign up to set email alerts
|

The Performance Implications of Locality Information Usage in Shared-Memory Multiprocessors

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
14
0
2

Year Published

1998
1998
2020
2020

Publication Types

Select...
4
2
2

Relationship

0
8

Authors

Journals

citations
Cited by 29 publications
(16 citation statements)
references
References 12 publications
0
14
0
2
Order By: Relevance
“…Memory affinity is the guarantee that memory access costs are reduced by either latency optimization or bandwidth increasing [1], [2]. In the last two decades, many researches have been carried out in the context of memory affinity, resulting in several proposals.…”
Section: Introductionmentioning
confidence: 99%
“…Memory affinity is the guarantee that memory access costs are reduced by either latency optimization or bandwidth increasing [1], [2]. In the last two decades, many researches have been carried out in the context of memory affinity, resulting in several proposals.…”
Section: Introductionmentioning
confidence: 99%
“…Thus, it is imperative to carefully consider which parts of the shared data should be attributed to which physical memory bank based on the data access pattern or on other considerations. Such an attribution of data to physical main memory is often called memory affinity Bellosa and Steckermeier (1996); Kleen (2005). This notion goes hand in hand with the CPU affinity, as noted in Grant and Afsahi (2007), such that the threads are being bound to specific cores for the application start and their context switches are disabled.…”
Section: Resultsmentioning
confidence: 99%
“…Recent work shows that contentions on the hardware prefetcher [25], the memory controller [27,30] and the DRAM bus [11] can also cause significant performance slowdown in both UMA and NUMA systems. Last-level cache miss rate has been widely used as a proxy for the contention on shared resources [7,8,9,14,26] and the similarity in thread address spaces has been used to quantify the inter-thread sharing activity [5,35,38].…”
Section: Optimization Via Schedulingmentioning
confidence: 99%
“…There are existing work focusing on hardware techniques [32] and program transformations [28,39,40] to mitigate the problem. Thread scheduling, a more flexible approach, has been also studied to avoid the destructive use of shared resources [7,8,11,14,30] or to use them constructively [5,35,38].…”
Section: Introductionmentioning
confidence: 99%