Proceedings of the 15th International Conference on Parallel Architectures and Compilation Techniques 2006
DOI: 10.1145/1152154.1152160
|View full text |Cite
|
Sign up to set email alerts
|

Architectural support for operating system-driven CMP cache management

Abstract: The role of the operating system (OS) in managing shared resources such as CPU time, memory, peripherals, and even energy is well motivated and understood [23]. Unfortunately, one key resource-lower-level shared cache in chip multi-processors-is commonly managed purely in hardware by rudimentary replacement policies such as least-recentlyused (LRU). The rigid nature of the hardware cache management policy poses a serious problem since there is no single best cache management policy across all sharing scenarios… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
132
0

Year Published

2007
2007
2015
2015

Publication Types

Select...
4
4

Relationship

1
7

Authors

Journals

citations
Cited by 180 publications
(132 citation statements)
references
References 23 publications
0
132
0
Order By: Relevance
“…Our second innovation is a feedback-based adaptive bandwidth sharing policy in which we periodically tune the bandwidth assigned to the sharers in order to achieve specified DRAM latencies. Adaptive bandwidth management does not add to the complexity of the hardware because it can be done entirely in software at the operating system (OS) or the hypervisor by using interfaces and mechanisms similar to those proposed for shared cache management [29]. While that interface supports various features such as thread migration and thread grouping (wherein a group of threads acts as a single resource principal), our study assumes that each thread running on a processor is a unique resource principal with its own allocated bandwidth share.…”
Section: Introductionmentioning
confidence: 99%
“…Our second innovation is a feedback-based adaptive bandwidth sharing policy in which we periodically tune the bandwidth assigned to the sharers in order to achieve specified DRAM latencies. Adaptive bandwidth management does not add to the complexity of the hardware because it can be done entirely in software at the operating system (OS) or the hypervisor by using interfaces and mechanisms similar to those proposed for shared cache management [29]. While that interface supports various features such as thread migration and thread grouping (wherein a group of threads acts as a single resource principal), our study assumes that each thread running on a processor is a unique resource principal with its own allocated bandwidth share.…”
Section: Introductionmentioning
confidence: 99%
“…Figure 7 shows the throughput of selected benchmark pairs with uncontrolled Pseudo LRU and the two cache partition schemes. We use fair speedup defined in Equation 14 to measure the throughput of concurrent running benchmark pairs. Note that taller bar in the figure means higher throughput.…”
Section: Evaluation Of Fairness Metricsmentioning
confidence: 99%
“…[4] proposed a framework to provide QoS for resources including shared caches. [14] designed architectural support for OS to manage shared caches.…”
Section: Related Workmentioning
confidence: 99%
“…However, most of these studies have focused on a single component of the entire system. For example, techniques have been proposed to reduce cache capacity interference [1,[3][4][5][6][7], cache bandwidth interference [13] and memory bus transfer interference [2,8,9]. Unfortunately, a technique that reduces interference in one component is not adequate to provide interference control for the complete memory system.…”
Section: Related Workmentioning
confidence: 99%
“…Since these effects are clearly undesirable, there is a need for architectural techniques that provide predictable performance and improve fairness. Previously, cache capacity interference has received a great deal of attention [1,[3][4][5][6][7] while only a few researchers have proposed techniques that reduce memory bus interference [2,8,9]. Furthermore, there has been little interest in the details of designing a complete, thread-aware memory system [10][11][12].…”
Section: Introductionmentioning
confidence: 99%