2018
DOI: 10.48550/arxiv.1809.08828
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Die-Stacked DRAM: Memory, Cache, or MemCache?

Mohammad Bakhshalipour,
HamidReza Zare,
Pejman Lotfi-Kamran
et al.

Abstract: Die-stacked DRAM is a promising solution for satisfying the ever-increasing memory bandwidth requirements of multi-core processors. Manufacturing technology has enabled stacking several gigabytes of DRAM modules on the active die, thereby providing orders of magnitude higher bandwidth as compared to the conventional DIMM-based DDR memories. Nevertheless, die-stacked DRAM, due to its limited capacity, cannot accommodate entire datasets of modern big-data applications. Therefore, prior proposals use it either as… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
3

Relationship

2
1

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 52 publications
0
3
0
Order By: Relevance
“…To reduce the storage cost, MANA exploits this observation that there are a small number of distinct high-order-bits patterns. This phenomenon is because the code base of a program has a high spatial locality and is much smaller than the size of the physical memory [9,15]. Consequently, instead of recording the complete trigger address that is the largest field of spatial regions, MANA uses the pointers to the observed high-order-bits patterns that need a considerably fewer number of bits.…”
Section: Mana Prefetchermentioning
confidence: 99%

MANA: Microarchitecting an Instruction Prefetcher

Ansari,
Golshan,
Lotfi-Kamran
et al. 2021
Preprint
Self Cite
“…To reduce the storage cost, MANA exploits this observation that there are a small number of distinct high-order-bits patterns. This phenomenon is because the code base of a program has a high spatial locality and is much smaller than the size of the physical memory [9,15]. Consequently, instead of recording the complete trigger address that is the largest field of spatial regions, MANA uses the pointers to the observed high-order-bits patterns that need a considerably fewer number of bits.…”
Section: Mana Prefetchermentioning
confidence: 99%

MANA: Microarchitecting an Instruction Prefetcher

Ansari,
Golshan,
Lotfi-Kamran
et al. 2021
Preprint
Self Cite
“…Progress in technology fabrication accompanied by circuit-level and microarchitectural advancements have brought about signi cant enhancements in the processors' performance over the past decades. Meanwhile, the performance of memory systems has not held speed with that of the processors, forming a large gap between the performance of processors and memory systems [29][30][31][32][33][34][35][36][37][38][39][40][41].…”
Section: Why Hardware Data Prefetching?mentioning
confidence: 99%
“…Figure 2 also highlights the changes in Trimma that adopt novel remap table and remap cache designs, which are merely in the metadata lookup and update phases (❶❹) without affecting the rest data access/eviction. Hence innovations including replacement policies [4], [27], [54], [64], [67], selective migration [16], [17], [82], and cache-flat dual modes [7], [10], [43] can be orthogonally integrated with Trimma.…”
Section: A Design Overviewmentioning
confidence: 99%