2006
DOI: 10.1109/tc.2006.23
|View full text |Cite
|
Sign up to set email alerts
|

Dynamic resizing of superscalar datapath components for energy efficiency

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
16
0

Year Published

2008
2008
2015
2015

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 26 publications
(16 citation statements)
references
References 15 publications
0
16
0
Order By: Relevance
“…Our goal is to share them across multiple cores with minimal impact on access latency, the number of ports, and the overall design. We take advantage of a modular ROB (and register file) design proposed in [25] which is shown to be effective in reducing the power and complexity of a multi-ported 2D SRAM structure. Our baseline multi-ported ROB/RF is implemented as a number of independent partitions.…”
Section: Reorder Buffer and Register Filementioning
confidence: 99%
“…Our goal is to share them across multiple cores with minimal impact on access latency, the number of ports, and the overall design. We take advantage of a modular ROB (and register file) design proposed in [25] which is shown to be effective in reducing the power and complexity of a multi-ported 2D SRAM structure. Our baseline multi-ported ROB/RF is implemented as a number of independent partitions.…”
Section: Reorder Buffer and Register Filementioning
confidence: 99%
“…A state transition out of some state s is controlled by either an action a ∈ A or a configuration change c ∈ C. Any state transition takes a certain amount of time to complete, where this latency overhead ranges from several clock cycles to hundreds of milli-seconds. A typical microarchitecture re-configuration latency, the duration between the time a decision is made to change the micro-architectural configuration and the time of actual configuration, takes up to tens of clock cycles [16]. Thus, a state transition time in the CTMDP model of the processor takes τ(s, s') time (= max (τ DVFS ,…”
Section: Modeling the Processor Statementioning
confidence: 99%
“…Adaptive front-end throttling is orthogonal to, and can even leverage, most existing techniques, providing even greater savings. For example, fetch gating based on branch prediction confidence [6,49] and dynamic issue queue, reorder buffer, and load/store queue re-sizing [8,14,26,54,57] can be applied together with adaptive front-end throttling to achieve greater savings. Third, previous work either does not have a direct way to quantify the overhead of the throttling technique and the resulting energy savings, or gets this information relying on architecture-level modeling frameworks, such as Wattch [13] and McPAT [45], which are known to have limited accuracy.…”
Section: Discussionmentioning
confidence: 99%
“…Prior works [8,9,14,23,26,36,49,50,54,57] have proposed various energy-saving techniques that dynamically allocate datapath resources according to the needs of applications. These energysaving techniques suffer from two problems.…”
Section: Dynamic Core Scalingmentioning
confidence: 99%
See 1 more Smart Citation