2015
DOI: 10.1145/2872887.2750392
|View full text |Cite
|
Sign up to set email alerts
|

Profiling a warehouse-scale computer

Abstract: With the increasing prevalence of warehouse-scale (WSC) and cloud computing, understanding the interactions of server applications with the underlying microarchitecture becomes ever more important in order to extract maximum performance out of server hardware. To aid such understanding, this paper presents a detailed microarchitectural analysis of live datacenter jobs, measured on more than 20,000 Google machines over a three year period, and comprising thousands of different applications. We first find that W… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

5
71
0

Year Published

2016
2016
2022
2022

Publication Types

Select...
6
4

Relationship

1
9

Authors

Journals

citations
Cited by 84 publications
(76 citation statements)
references
References 46 publications
(42 reference statements)
5
71
0
Order By: Relevance
“…II-B, benefits of large cache capacities disappear at high access latencies. Conventional DRAM caches are beneficial in alleviating the bandwidth bottleneck [29]; however, today's server CPUs are not bandwidth limited on scale-out workloads as shown in recent work [10] and corroborated by Google [39] and our own results. As such, conventional DRAM caches do not benefit these designs.…”
Section: Evaluation and Discussionsupporting
confidence: 83%
“…II-B, benefits of large cache capacities disappear at high access latencies. Conventional DRAM caches are beneficial in alleviating the bandwidth bottleneck [29]; however, today's server CPUs are not bandwidth limited on scale-out workloads as shown in recent work [10] and corroborated by Google [39] and our own results. As such, conventional DRAM caches do not benefit these designs.…”
Section: Evaluation and Discussionsupporting
confidence: 83%
“…Increasing core counts, emergence of more data-intensive and latencycritical applications, and increasingly limited bandwidth in the memory system are together leading to higher memory latency. Thus, low-latency memory operation is now even more important to improving overall system performance [9,11,16,27,36,37,38,45,49,53,54,56,66,68,70,79].…”
Section: Introductionmentioning
confidence: 99%
“…Figure 2.2 presents the cumulative distribution of cycles across workloads in Google data centers. The top 50 workloads account for only about 60% of the total WSC cycles, with a long tail accounting for the rest of the cycles [Kan+15].…”
Section: Workload Diversitymentioning
confidence: 99%