2008
DOI: 10.1145/1328195.1328200
|View full text |Cite
|
Sign up to set email alerts
|

Reducing cache misses through programmable decoders

Abstract: Level-one caches normally reside on a processor's critical path, which determines clock frequency. Therefore, fast access to level-one cache is important. Direct-mapped caches exhibit faster access time, but poor hit rates, compared with same sized set-associative caches because of nonuniform accesses to the cache sets. The nonuniform accesses generate more cache misses in some sets, while other sets are underutilized. We propose to increase the decoder length and, hence, reduce the accesses to heavily used se… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2009
2009
2024
2024

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 8 publications
(3 citation statements)
references
References 45 publications
0
3
0
Order By: Relevance
“…One disadvantage of this method is high per-access power consumption. An enhanced B-Cache [85] reduced the total access energy consumption; however, it remained higher than that of a lower associativity cache. The FASTA-based VMWA cache enables almost complete conflict miss elimination, and at the same time, significantly reduces the cache access energy consumption compared to a typical way-associative cache.…”
Section: Related Workmentioning
confidence: 84%
“…One disadvantage of this method is high per-access power consumption. An enhanced B-Cache [85] reduced the total access energy consumption; however, it remained higher than that of a lower associativity cache. The FASTA-based VMWA cache enables almost complete conflict miss elimination, and at the same time, significantly reduces the cache access energy consumption compared to a typical way-associative cache.…”
Section: Related Workmentioning
confidence: 84%
“…Due to its simple structure, snoopy protocol is considered more advantageous than other protocols. Since system bus is an exclusive resource, when the number of processor cores interconnected with the system bus is large, the efficiency of the protocol drops dramatically [4]. In snoopy protocol all requests are broadcasted in a undifferentiated manner into the system bus and so all the processors connected with the system bus must read the request and it has to check whether its cache contains the requested data block's copy.…”
Section: Snoopy Protocolmentioning
confidence: 99%
“…There has been significant work on runtime effects due to cache performance; however, most of this research focuses on minimizing cache misses [1,2,8,17,18,19]. By minimizing cache misses, energy spent in accessing memory is decreased, and the overall application runtime is improved.…”
Section: Related Workmentioning
confidence: 99%