Abstract
AbstractUnderstanding the influence of configuration options on the performance of a software system is key for finding optimal system configurations, system understanding, and performance debugging. In the literature, a number of performance-influence modeling approaches have been proposed, which model a configuration option’s influence and a configuration’s performance as a scalar value. However, these point estimates falsely imply a certainty regarding an option’s influence that neglects several sources of uncertainty within the assessment process, such as (1) measurement bias, choices of model representation and learning process, and incomplete data. This leads to the situation that different approaches and even different learning runs assign different scalar performance values to options and interactions among them. The true influence is uncertain, though. There is no way to quantify this uncertainty with state-of-the-art performance modeling approaches.We propose a novel approach, P4, which is based on probabilistic programming, that explicitly models uncertainty for option influences and consequently provides a confidence interval for each prediction alongside a scalar. This way, we can explain, for the first time, why predictions may be erroneous and which option’s influence may be unreliable. An evaluation on 13 real-world subject systems shows that P4’s accuracy is in line with the state of the art while providing reliable confidence intervals, in addition to scalar predictions. We qualitatively explain how uncertain influences of individual options and interactions cause inaccurate predictions.
Funder
Deutsche Forschungsgemeinschaft
Universität Leipzig
Publisher
Springer Science and Business Media LLC
Reference66 articles.
1. Aken Dana Van, Pavlo Andrew, Gordon Geoffrey J, Zhang Bohan (2017) Automatic database management system tuning through large-scale machine learning. In: Proceedings of the international conference on management of data (SIGMOD). ACM, pp 1009–1024. ISBN 978-1-4503-4197-4. https://doi.org/10.1145/3035918.3064029
2. Alan M (2002) Subset selection in regression. CRC Press, https://doi.org/10.1201/9781420035933
3. Antonelli F, Cortellessa V, Gribaudo M, Pinciroli R, Trivedi KS, Trubiani C (2020) Analytical modeling of performance indices under epistemic uncertainty applied to cloud computing systems. Future Gener Comput Syst 102:746–761. ISSN 0167-739X. https://doi.org/10.1016/j.future.2019.09.006
4. Arcaini Paolo, Inverso Omar, Trubiani Catia (2020) Automated model-based performance analysis of software product lines under uncertainty. J Inform Software Technol (IST) 127:106371. ISSN 0950-5849. https://doi.org/10.1016/j.infsof.2020.106371
5. Bergstra J, Bardenet R, Bengio Y, Kégl B (2011) Algorithms for hyper-parameter optimization. In: Shawe-Taylor J, Zemel R, Bartlett P, Pereira F, Weinberger KQ (eds) Advances in neural information processing systems, volume 24. Curran Associates, Inc
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献