2015 IEEE 21st International Symposium on High Performance Computer Architecture (HPCA) 2015
DOI: 10.1109/hpca.2015.7056039
|View full text |Cite
|
Sign up to set email alerts
|

Adrenaline: Pinpointing and reining in tail queries with quick voltage boosting

Abstract: Reducing the long tail of the query latency distribution in modern warehouse scale computers is critical for improving performance and quality of service of workloads such as Web Search and Memcached. Traditional turbo boost increases a processor's voltage and frequency during a coarsegrain sliding window, boosting all queries that are processed during that window. However, the inability of such a technique to pinpoint tail queries for boosting limits its tail reduction benefit.In this work, we propose Adrenal… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
57
0

Year Published

2015
2015
2021
2021

Publication Types

Select...
6
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 96 publications
(57 citation statements)
references
References 40 publications
0
57
0
Order By: Relevance
“…Many of these studies use workloads internal to datacenter operators like Google or Facebook [32,33,36,38,55,56]. Academic studies use one or a few latencycritical benchmarks [25,48,54], which limits the range of behaviors and performance requirements across which their proposed techniques can be evaluated. Some work uses more readily-available sequential and parallel batch workloads (e.g., from SPEC CPU2006 or PARSEC) and treats them as latencycritical applications [15,57].…”
Section: A Anatomy Of Latency-critical Applicationsmentioning
confidence: 99%
See 2 more Smart Citations
“…Many of these studies use workloads internal to datacenter operators like Google or Facebook [32,33,36,38,55,56]. Academic studies use one or a few latencycritical benchmarks [25,48,54], which limits the range of behaviors and performance requirements across which their proposed techniques can be evaluated. Some work uses more readily-available sequential and parallel batch workloads (e.g., from SPEC CPU2006 or PARSEC) and treats them as latencycritical applications [15,57].…”
Section: A Anatomy Of Latency-critical Applicationsmentioning
confidence: 99%
“…These techniques include new cluster managers that schedule and migrate applications across systems to reduce interference [18,32,36,54], fast dynamic voltage-frequency scaling (DVFS) techniques to improve power efficiency [25,29,32,48], hardware and software schemes to use low power idle states [37,39,53], and hardware resource partitioning schemes that allow batch workloads to run alongside latency-critical ones, improving utilization [29,30,33,57].…”
Section: A Anatomy Of Latency-critical Applicationsmentioning
confidence: 99%
See 1 more Smart Citation
“…We observe that the impact of DVFS-only controls differs noticeably between Linux and IX: with Linux, the DVFSonly alternate frontier is very close to the Pareto frontier, meaning that a DVFS-only approach such as Pegasus [29] or Adrenaline [15] would be adequate. This is due to Linux's idling behavior, which saves resources.…”
Section: Pareto-optimal Static Configurationsmentioning
confidence: 91%
“…The key challenge is coping with the inherent short-term variability of latency-critical applications: requests arrive at unpredictable times and are often bursty, causing short-term spikes and queuing delays that dominate tail latency [22,25]; and the amount of work per request often varies by an order of magnitude or more [16,25].…”
Section: Introductionmentioning
confidence: 99%