Affiliation:
1. University of Stavanger, Simula Research Laboratory, Stavanger, Norway
2. Simula Research Laboratory, Oslo, Norway
Abstract
The use of Deep learning in Cyber-Physical Systems (CPSs) is gaining popularity due to its ability to bring intelligence to CPS behaviors. However, both CPSs and deep learning have inherent uncertainty. Such uncertainty, if not handled adequately, can lead to unsafe CPS behavior. The first step toward addressing such uncertainty in deep learning is to quantify uncertainty. Hence, we propose a novel method called
NIRVANA
(uNcertaInty pRediction ValidAtor iN Ai) for prediction validation based on uncertainty metrics. To this end, we first employ prediction-time Dropout-based Neural Networks to quantify uncertainty in deep learning models applied to CPS data. Second, such quantified uncertainty is taken as the input to predict wrong labels using a support vector machine, with the aim of building a highly discriminating prediction validator model with uncertainty values. In addition, we investigated the relationship between uncertainty quantification and prediction performance and conducted experiments to obtain optimal dropout ratios. We conducted all the experiments with four real-world CPS datasets. Results show that uncertainty quantification is negatively correlated to prediction performance of a deep learning model of CPS data. Also, our dropout ratio adjustment approach is effective in reducing uncertainty of correct predictions while increasing uncertainty of wrong predictions.
Funder
Co-evolver
Research Council of Norway
FRIPRO program
Publisher
Association for Computing Machinery (ACM)
Cited by
12 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. Pretrain, Prompt, and Transfer: Evolving Digital Twins for Time-to-Event Analysis in Cyber-Physical Systems;IEEE Transactions on Software Engineering;2024-06
2. Interpretable On-the-Fly Repair of Deep Neural Classifiers;Proceedings of the 1st International Workshop on Dependability and Trustworthiness of Safety-Critical Systems with Machine Learned Components;2023-12-04
3. Adopting Two Supervisors for Efficient Use of Large-Scale Remote Deep Neural Networks;ACM Transactions on Software Engineering and Methodology;2023-11-23
4. Generating and detecting true ambiguity: a forgotten danger in DNN supervision testing;Empirical Software Engineering;2023-11
5. Evolve the Model Universe of a System Universe;2023 38th IEEE/ACM International Conference on Automated Software Engineering (ASE);2023-09-11