2015 IEEE 21st International Symposium on High Performance Computer Architecture (HPCA) 2015
DOI: 10.1109/hpca.2015.7056027
|View full text |Cite
|
Sign up to set email alerts
|

Heterogeneous memory architectures: A HW/SW approach for mixing die-stacked and off-package memories

Abstract: Die-stacked DRAM is a technology that will soon be integrated in high-performance systems. Recent studies have focused on hardware caching techniques to make use of the stacked memory, but these approaches require complex changes to the processor and also cannot leverage the stacked memory to increase the system's overall memory capacity. In this work, we explore the challenges of exposing the stacked DRAM as part of the system's physical address space. This non-uniform access memory (NUMA) styled approach gre… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
66
0

Year Published

2017
2017
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 134 publications
(66 citation statements)
references
References 24 publications
0
66
0
Order By: Relevance
“…Similarly, Micron's Hybrid Memory Cube [4,5] and byte-addressable persistent memories [6][7][8][9] are quickly gaining traction. Vendors are combining these high-performance memories with traditional high-capacity and low-cost DRAM, prompting research on heterogeneous memory architectures [2,[9][10][11][12][13][14][15].…”
Section: Introductionmentioning
confidence: 99%
“…Similarly, Micron's Hybrid Memory Cube [4,5] and byte-addressable persistent memories [6][7][8][9] are quickly gaining traction. Vendors are combining these high-performance memories with traditional high-capacity and low-cost DRAM, prompting research on heterogeneous memory architectures [2,[9][10][11][12][13][14][15].…”
Section: Introductionmentioning
confidence: 99%
“…Fig. 3 shows the two logical organizations for heterogeneous memory systems (HMS) composed of fast stacked DRAM (e.g., WideIO2) and slow off-package DRAM (e.g., LPDDR4): (a) hardware-managed cache [9,10] and (b) flat physical memory [7,8,10,11]. The cache organization is software-transparent and does not require any change to the software stack.…”
Section: Background and Motivationmentioning
confidence: 99%
“…If the bandwidth and latency gap between fast and slow regions is small as in mobile DRAMs, the cost of migration can easily outweigh its benefits. Migration at a coarse granularity (say, DRAM page size) [7,11] has advantage of better exploiting DRAM row buffer locality over fine granularity (say, CPU cache block size) [8]. Since the existing migration heuristics are designed for heterogeneous memory devices with a relatively large performance gap [7,8,9,10,11], they can be suboptimal for a LPDDR4-WideIO2 heterogeneous memory system.…”
Section: Background and Motivationmentioning
confidence: 99%
See 2 more Smart Citations