Inspect, Understand, Overcome: A Survey of Practical Methods for AI Safety
Author:
Houben Sebastian,Abrecht Stephanie,Akila Maram,Bär Andreas,Brockherde Felix,Feifel Patrick,Fingscheidt Tim,Gannamaneni Sujan Sai,Ghobadi Seyed Eghbal,Hammam Ahmed,Haselhoff Anselm,Hauser Felix,Heinzemann Christian,Hoffmann Marco,Kapoor Nikhil,Kappel Falk,Klingner Marvin,Kronenberger Jan,Küppers Fabian,Löhdefink Jonas,Mlynarski Michael,Mock Michael,Mualla Firas,Pavlitskaya Svetlana,Poretschkin Maximilian,Pohl Alexander,Ravi-Kumar Varun,Rosenzweig Julia,Rottmann Matthias,Rüping Stefan,Sämann Timo,Schneider Jan David,Schulz Elena,Schwalbe Gesina,Sicking Joachim,Srivastava Toshika,Varghese Serin,Weber Michael,Wirkert Sebastian,Wirtz Tim,Woehrle Matthias
Abstract
AbstractDeployment of modern data-driven machine learning methods, most often realized by deep neural networks (DNNs), in safety-critical applications such as health care, industrial plant control, or autonomous driving is highly challenging due to numerous model-inherent shortcomings. These shortcomings are diverse and range from a lack of generalization over insufficient interpretability and implausible predictions to directed attacks by means of malicious inputs. Cyber-physical systems employing DNNs are therefore likely to suffer from so-called safety concerns, properties that preclude their deployment as no argument or experimental setup can help to assess the remaining risk. In recent years, an abundance of state-of-the-art techniques aiming to address these safety concerns has emerged. This chapter provides a structured and broad overview of them. We first identify categories of insufficiencies to then describe research activities aiming at their detection, quantification, or mitigation. Our work addresses machine learning experts and safety engineers alike: The former ones might profit from the broad range of machine learning topics covered and discussions on limitations of recent methods. The latter ones might gain insights into the specifics of modern machine learning methods. We hope that this contribution fuels discussions on desiderata for machine learning systems and strategies on how to help to advance existing approaches accordingly.
Funder
University of Wuppertal
Publisher
Springer International Publishing
Reference478 articles.
1. S. Akcay, A. Atapour-Abarghouei, T.P. Breckon, Ganomaly: semi-supervised anomaly detection via adversarial training, in Proceedings of the Asian Conference on Computer Vision (ACCV), Perth, WA, Australia (2018), pp. 622–637 2. Y. Adi, C. Baum, M. Cisse, B. Pinkas, J. Keshet, Turning your weakness into a strength: watermarking deep neural networks by backdooring, in Proceedings of the USENIX Security Symposium, Baltimore, MD, USA (2018), pp. 1615–1631 3. S. Amershi, M. Cakmak, W.B. Knox, T. Kulesza, Power to the people: the role of humans in interactive machine learning. AI Mag. 35(4), 105–120 (2014) 4. M. Ancona, E. Ceolini, C. Öztireli, M.H. Gross, A unified view of gradient-based attribution methods for deep neural networks (2017), pp. 1–11. arXiv:1711.06104 5. M. Angus, K. Czarnecki, R. Salay, Efficacy of pixel-level OOD detection for semantic segmentation (2019), pp. 1–13. arXiv:1911.02897
Cited by
15 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
|
|