Closed-Loop Uncertainty: The Evaluation and Calibration of Uncertainty for Human–Machine Teams under Data Drift
Author:
Bishof Zachary1, Scheuerman Jaelle1, Michael Chris J.1
Affiliation:
1. U.S. Naval Research Laboratory, 1005 Balch Boulevard, Stennis Space Center, St. Louis, MS 39529, USA
Abstract
Though an accurate measurement of entropy, or more generally uncertainty, is critical to the success of human–machine teams, the evaluation of the accuracy of such metrics as a probability of machine correctness is often aggregated and not assessed as an iterative control process. The entropy of the decisions made by human–machine teams may not be accurately measured under cold start or at times of data drift unless disagreements between the human and machine are immediately fed back to the classifier iteratively. In this study, we present a stochastic framework by which an uncertainty model may be evaluated iteratively as a probability of machine correctness. We target a novel problem, referred to as the threshold selection problem, which involves a user subjectively selecting the point at which a signal transitions to a low state. This problem is designed to be simple and replicable for human–machine experimentation while exhibiting properties of more complex applications. Finally, we explore the potential of incorporating feedback of machine correctness into a baseline naïve Bayes uncertainty model with a novel reinforcement learning approach. The approach refines a baseline uncertainty model by incorporating machine correctness at every iteration. Experiments are conducted over a large number of realizations to properly evaluate uncertainty at each iteration of the human–machine team. Results show that our novel approach, called closed-loop uncertainty, outperforms the baseline in every case, yielding about 45% improvement on average.
Funder
NRL NISE Program Element Jerome and Isabella Karles Fellowship under Work Unit N20N
Subject
General Physics and Astronomy
Reference42 articles.
1. A review of uncertainty quantification in deep learning: Techniques, applications and challenges;Abdar;Inf. Fusion,2021 2. Jiang, H., Kim, B., Guan, M., and Gupta, M. (2018, January 3–8). To trust or not to trust a classifier. Proceedings of the Advances in Neural Information Processing Systems 31 (NeurIPS 2018), Montréal, Canada. 3. Optimization under uncertainty: state-of-the-art and opportunities;Sahinidis;Comput. Chem. Eng.,2004 4. Bhatt, U., Antorán, J., Zhang, Y., Liao, Q.V., Sattigeri, P., Fogliato, R., Melançon, G., Krishnan, R., Stanley, J., and Tickoo, O. (2021, January 19–21). Uncertainty as a form of transparency: Measuring, communicating, and using uncertainty. Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society, Virtual. 5. Entropy and uncertainty;Robinson;Entropy,2008
|
|