Proceedings of the 40th Annual International Symposium on Computer Architecture 2013
DOI: 10.1145/2485922.2485974
|View full text |Cite
|
Sign up to set email alerts
|

Bubble-flux

Abstract: Ensuring the quality of service (QoS) for latency-sensitive applications while allowing co-locations of multiple applications on servers is critical for improving server utilization and reducing cost in modern warehouse-scale computers (WSCs). Recent work relies on static profiling to precisely predict the QoS degradation that results from performance interference among co-running applications to increase the number of "safe" co-locations. However, these static profiling techniques have several critical limita… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
21
0

Year Published

2016
2016
2021
2021

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 237 publications
(21 citation statements)
references
References 47 publications
0
21
0
Order By: Relevance
“…Apart from software heterogeneity, datacenter hardware is also becoming increasingly heterogeneous as special-purpose architectures [20-22, 39, 55] and FPGAs are used to accelerate critical operations [19,25,40,75]. This adds to the existing server heterogeneity in the cloud where servers are progressively replaced and upgraded over the datacenter's provisioned lifetime [31,33,65,68,95], and further complicates the effort to guarantee predictable performance.…”
Section: Netflixmentioning
confidence: 99%
See 2 more Smart Citations
“…Apart from software heterogeneity, datacenter hardware is also becoming increasingly heterogeneous as special-purpose architectures [20-22, 39, 55] and FPGAs are used to accelerate critical operations [19,25,40,75]. This adds to the existing server heterogeneity in the cloud where servers are progressively replaced and upgraded over the datacenter's provisioned lifetime [31,33,65,68,95], and further complicates the effort to guarantee predictable performance.…”
Section: Netflixmentioning
confidence: 99%
“…A second line of work tries to identify resources that will allow a new, potentially-unknown application to meet its performance (throughput or tail latency) requirements [29,31,32,34,66,68,95]. Paragon uses classification to determine the impact of platform heterogeneity and workload interference on an unknown, incoming workload [30,31].…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Again, this work can be done at the processor-level [23,24,44,56], DRAM [11], storage [30], and across a data-center [33,39]. At the data-center or cluster-level, power can be saved by consolidating workloads to use fewer physical machines [14,27,34,35,37,54], coordinating co-existing applications [45], and scheduling with green power [20] Scheduling jobs under a power cap has recently become a major concern for HPC operating systems [7,15] and job schedulers [3,19]. Recent work suggests that HPC workloads can actually achieve higher performance by over-provisioning large-scale installationssuch that using all nodes at full capacity would drastically violate the power budget-and severely power capping the individual nodes [47].…”
Section: Related Workmentioning
confidence: 99%
“…Our approach explores this issue at higher throughputs and with tighter-latency SLOs. Bubble-Flux [Yang et al 2013] additionally controls background threads; we control background and latency-sensitive threads. CPI 2 [Zhang et al 2013] detects performance interference by observing changes in CPI and throttles offending jobs.…”
Section: Related Workmentioning
confidence: 99%