Proceedings of the ACM Symposium on Cloud Computing 2014
DOI: 10.1145/2670979.2670988
|View full text |Cite
|
Sign up to set email alerts
|

Tales of the Tail

Abstract: Interactive services often have large-scale parallel implementations. To deliver fast responses, the median and tail latencies of a service's components must be low. In this paper, we explore the hardware, OS, and application-level sources of poor tail latency in high throughput servers executing on multi-core machines.We model these network services as a queuing system in order to establish the best-achievable latency distribution. Using fine-grained measurements of three different servers (a null RPC service… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
18
0

Year Published

2016
2016
2023
2023

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 164 publications
(18 citation statements)
references
References 18 publications
0
18
0
Order By: Relevance
“…The statemachine configuration for Memcached and Web-Search are represented in blue and red line, respectively, in Figure 2c. Figure 3 [30] and a QoS target that needs to be met. As shown in Figure 2 and 3, there exists a unique configuration for each load that optimizes energy efficiency.…”
Section: Motivationmentioning
confidence: 99%
“…The statemachine configuration for Memcached and Web-Search are represented in blue and red line, respectively, in Figure 2c. Figure 3 [30] and a QoS target that needs to be met. As shown in Figure 2 and 3, there exists a unique configuration for each load that optimizes energy efficiency.…”
Section: Motivationmentioning
confidence: 99%
“…In practical scenarios, each workload has a time-varying load [38] and a QoS target that needs to be met. As shown in Figure 2 and 3, there exists a unique con guration for each load that optimizes energy e ciency.…”
Section: Exploring Individual Workload Particularitiesmentioning
confidence: 99%
“…At runtime, Hipster determines when to dynamically switch between the learning and exploitation phases, based on a pre xed time quantum. At deployment stage, we ensure that the bucket size for each workload gives at least 95 % QoS guarantee [38] with minimal energy consumption.…”
Section: Learning and Exploitation Phasesmentioning
confidence: 99%
“…Ideally, each service node will use the fewest resources (cores, memory, or IOPS) needed to satisfy packet rate and tail latency requirements at any point. Unfortunately, classic operating system schedulers are illmatched to ensure tail control [Leverich and Kozyrakis 2014;Li et al 2014]. Novel dynamic resource management mechanisms and policies are required to improve energy proportionality and workload consolidation in the presence of latency-sensitive applications [Lo et al 2014[Lo et al , 2015.…”
Section: Challenges For Datacenter Applicationsmentioning
confidence: 99%
“…They are particularly sensitive to resource allocation and frequency settings, and they suffer frequent tail latency violations when common power management or consolidation approaches are used [Leverich and Kozyrakis 2014;Li et al 2014]. As a result, operators typically deploy them on dedicated servers running in polling mode, forgoing opportunities for workload consolidation and reduced power consumption at below-peak utilization levels.…”
Section: Introductionmentioning
confidence: 99%