“…Most existing techniques build global performanceinfluence models and treat the system as a black box, measuring the system's execution in an environment with a given workload for a subset of all configurations, and learning a model from these observations. The sampling (i.e., selecting which configurations to measure) and learning techniques used [15,17,35,36,49,61,[63][64][65] result in tradeoffs among the cost to build the models and the accuracy and interpretability of the models [15,35,38]. For example, larger samples are more expensive, but usually lead to more accurate models; random forests, with large enough samples, tend to learn more accurate models than those built with linear regression, but the models are harder to interpret when users want to understand performance or debug their systems [15,35,49] (see Fig.…”