2013
DOI: 10.1109/jssc.2013.2258815
|View full text |Cite
|
Sign up to set email alerts
|

An Energy Efficient 32-nm 20-MB Shared On-Die L3 Cache for Intel® Xeon® Processor E5 Family

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
14
0

Year Published

2015
2015
2020
2020

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 23 publications
(14 citation statements)
references
References 2 publications
0
14
0
Order By: Relevance
“…The SRAM banks were modeled as being similar to that of the Intel R Xeon R E5 processor LLC design [19]. The L2 is modeled as a 2 (wide) ×4 (high) array of 32KB SRAM banks.…”
Section: Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…The SRAM banks were modeled as being similar to that of the Intel R Xeon R E5 processor LLC design [19]. The L2 is modeled as a 2 (wide) ×4 (high) array of 32KB SRAM banks.…”
Section: Methodsmentioning
confidence: 99%
“…A topology similar to Figure 4a was adopted for the design of an LLC slice in the Intel R Xeon R E5 family of processors [19] and a proposed SRAM macro design by Samsung [35]. In such a topology, if the ways of the cache are interleaved across the various SRAM banks, there can be a significant energy difference between the various line locations.…”
Section: Cache Organizationsmentioning
confidence: 99%
See 1 more Smart Citation
“…We provide a brief overview of a cache's geometry in a modern processor. Figure 3 illustrates a multi-core processor modeled loosely after Intel's Xeon processors [14], [15]. Shared Last Level Cache (LLC) is distributed into many slices (14 for Xeon E5 we modeled), which are accessible to the cores through a shared ring interconnect (not shown in figure).…”
Section: Cache Geometrymentioning
confidence: 99%
“…We observe that LLC access latency is dominated by wire delays inside a cache slice, accessing upper-level cache control structures, and network-on-chip. Thus, while a typical LLC access can take ∼30 cycles, an SRAM array access is only 1 cycle (at 4 GHz clock [14]). Fortunately, in-situ architectures such as Neural Cache require only SRAM array accesses and do not incur the overheads of a traditional cache access.…”
Section: Cache Geometrymentioning
confidence: 99%