Affiliation:
1. Hunan University, Hunan, China
2. Nanjing University of Science and Technology, Nanjing, Jiangsu, China
3. New York Institute of Technology, New York, NY
4. State University of New York, Albany, NY
Abstract
In recent years, with rapid technological advancement in both computing hardware and algorithm, Artificial Intelligence (AI) has demonstrated significant advantage over human being in a wide range of fields, such as image recognition, education, autonomous vehicles, finance, and medical diagnosis. However, AI-based systems are generally vulnerable to various security threats throughout the whole process, ranging from the initial data collection and preparation to the training, inference, and final deployment. In an AI-based system, the data collection and pre-processing phase are vulnerable to sensor spoofing attacks and scaling attacks, respectively, while the training and inference phases of the model are subject to poisoning attacks and adversarial attacks, respectively. To address these severe security threats against the AI-based systems, in this article, we review the challenges and recent research advances for security issues in AI, so as to depict an overall blueprint for AI security. More specifically, we first take the lifecycle of an AI-based system as a guide to introduce the security threats that emerge at each stage, which is followed by a detailed summary for corresponding countermeasures. Finally, some of the future challenges and opportunities for the security issues in AI will also be discussed.
Funder
National Natural Science Foundation of China
Science and Technology Project of Department of Communications of Hunan Provincial
Hunan Natural Science Foundation for Distinguished Young Scholars
Hunan Science and Technology Innovation Leading Talents Project
Natural Science Foundation of Fujian Province
Key R & D Projects of Changsha
National Natural Science Foundation of JiangSu
Publisher
Association for Computing Machinery (ACM)
Subject
General Computer Science,Theoretical Computer Science
Reference176 articles.
1. Adversarial Deep Learning for Robust Detection of Binary Encoded Malware
2. Moustafa Alzantot Bharathan Balaji and Mani B. Srivastava. 2018. Did you hear that? Adversarial Examples Against Automatic Speech Recognition. CoRR abs/1801.00554 (2018). arXiv:1801.00554 . http://arxiv.org/abs/1801.00554.
3. Hyrum S. Anderson Anant Kharkar Bobby Filar and Phil Roth. 2017. Evading machine learning malware detection. Black Hat (2017) 1–6.
4. Martin Arjovsky and Léon Bottou. 2017. Towards Principled Methods for Training Generative Adversarial Networks. arXiv:1701.04862 [stat.ML] https://arxiv.org/abs/1701.04862
Cited by
57 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献