Proceedings of the ACM/SPEC International Conference on Performance Engineering 2020
DOI: 10.1145/3358960.3379132
|View full text |Cite
|
Sign up to set email alerts
|

Duet Benchmarking: Improving Measurement Accuracy in the Cloud

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 16 publications
(6 citation statements)
references
References 19 publications
0
6
0
Order By: Relevance
“…To mitigate measurement bias, Georges et al [18] outlined a rigorous methodology how to asses performance of Java programs, which we base our measurement technique on. Using the correct statistical techniques to assess performance is paramount, with estimated confidence intervals using bootstrap being the state-of-the-art [8,9,27,33]. One of our stopping criteria is based on and our result quality evaluation uses confidence intervals with bootstrap.…”
Section: Related Workmentioning
confidence: 99%
“…To mitigate measurement bias, Georges et al [18] outlined a rigorous methodology how to asses performance of Java programs, which we base our measurement technique on. Using the correct statistical techniques to assess performance is paramount, with estimated confidence intervals using bootstrap being the state-of-the-art [8,9,27,33]. One of our stopping criteria is based on and our result quality evaluation uses confidence intervals with bootstrap.…”
Section: Related Workmentioning
confidence: 99%
“…Related work quantifying the variability of short running performance experiments in the cloud can, for example, be found by Iosup et al (2011), Leitner and Cito (2016), Abedi and Brecht (2017), Maricq et al (2018), or Laaber et al (2019). He et al (2019) and He et al (2021) and Bulej et al (2020) propose methods for reducing the amount of experiment repetitions while preserving a high measurement accuracy. These methods differ from ours in that they aim to accurately measure the performance of a system, while for benchmarking scalability, we only need to accurately assess whether a system fulfills specified SLOs.…”
Section: Scalability Measurement Methodsmentioning
confidence: 99%
“…Cloud VMs often have different performance characteristics and are subject to random fluctuations, even comparing two VMs of the same instance type [30]. Since we are only interested in a relative comparison of two SUT options and do not need absolute values, we can (largely) remove the noise resulting from cloud performance variability using duet benchmarking [6,7]. This is achieved, by running two (or even more) different SUT options, in our case different versions, and their application benchmarks on the same cloud VM simultaneously, with 50% of the resources assigned to each SUT option and benchmark.…”
Section: Application Benchmarksmentioning
confidence: 99%