2018 IEEE International Symposium on High Performance Computer Architecture (HPCA) 2018
DOI: 10.1109/hpca.2018.00061
|View full text |Cite
|
Sign up to set email alerts
|

Memory Hierarchy for Web Search

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
37
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 60 publications
(38 citation statements)
references
References 43 publications
1
37
0
Order By: Relevance
“…Evaluating an architectural feature in a real system is not an easy task. Multiple factors challenge this: general-purpose CPUs run a large spectrum of workloads with mixed bottlenecks [2,4]. For example, profiling results shared by this article, indicate that integer benchmarks highly benefit from better instruction fetch, while Floating-Point (FP) benchmarks benefit from an optimized execution engine.…”
Section: Difficultymentioning
confidence: 95%
“…Evaluating an architectural feature in a real system is not an easy task. Multiple factors challenge this: general-purpose CPUs run a large spectrum of workloads with mixed bottlenecks [2,4]. For example, profiling results shared by this article, indicate that integer benchmarks highly benefit from better instruction fetch, while Floating-Point (FP) benchmarks benefit from an optimized execution engine.…”
Section: Difficultymentioning
confidence: 95%
“…For each workload, we configure the overall memory capacity (i.e., second tier of the hierarchy) to be equal to the workload's dataset size (i.e., Data Serving, Web Search and Media Streaming have 16GB datasets, while Data Analytics and Web Serving have 32GB datasets). However, today's datacenter-scale applications can have much larger datasets that even span into the terabyte range [9,53]; since our work makes specific claims about the capacity ratio relating the two tiers of our memory hierarchy, we conducted a study to verify that our results stand for larger datasets.…”
Section: Evaluation Methodologymentioning
confidence: 99%
“…Huang et al [11] use SRAM banks for storing tags, reducing tag lookup latency. The other prior works in [8], [9] propose an on-chip L4 DRAM cache based on eDRAM and 3D-stacked DRAM, respectively. By placing the L4 DRAM cache near processors or deploying private L4 cache for each core, we can benefit from the lower latency than the off-chip DRAM.…”
Section: B L4 Dram Cache Architecturementioning
confidence: 99%
“…Such DRAM cache can provide low latency and/or high bandwidth compared with the main memory, and much larger capacity than the SRAM cache according to the type of DRAM and its implementation [7]. To maximize its benefits regarding performance, most of prior work propose various techniques for the DRAM cache [8]- [12].…”
Section: Introductionmentioning
confidence: 99%