DIPA: Adversarial Attack on DNNs by Dropping Information and Pixel-Level Attack on Attention
-
Published:2024-07-03
Issue:7
Volume:15
Page:391
-
ISSN:2078-2489
-
Container-title:Information
-
language:en
-
Short-container-title:Information
Author:
Liu Jing1ORCID, Liu Huailin1, Wang Pengju1, Wu Yang1, Li Keqin2ORCID
Affiliation:
1. College of Computer Science, Inner Mongolia University, Hohhot 010021, China 2. Department of Computer Science, State University of New York, New Paltz, NY 12561, USA
Abstract
Deep neural networks (DNNs) have shown remarkable performance across a wide range of fields, including image recognition, natural language processing, and speech processing. However, recent studies indicate that DNNs are highly vulnerable to well-crafted adversarial samples, which can cause incorrect classifications and predictions. These samples are so similar to the original ones that they are nearly undetectable by human vision, posing a significant security risk to DNNs in the real world due to the impact of adversarial attacks. Currently, the most common adversarial attack methods explicitly add adversarial perturbations to image samples, often resulting in adversarial samples that are easier to distinguish by humans. To address this issue, we are motivated to develop more effective methods for generating adversarial samples that remain undetectable to human vision. This paper proposes a pixel-level adversarial attack method based on attention mechanism and high-frequency information separation, named DIPA. Specifically, our approach involves constructing an attention suppression loss function and utilizing gradient information to identify and perturb sensitive pixels. By suppressing the model’s attention to the correct classes, the neural network is misled to focus on irrelevant classes, leading to incorrect judgments. Unlike previous studies, DIPA enhances the attack of adversarial samples by separating the imperceptible details in image samples to more effectively hide the adversarial perturbation while ensuring a higher attack success rate. Our experimental results demonstrate that under the extreme single-pixel attack scenario, DIPA achieves higher attack success rates for neural network models with various architectures. Furthermore, the visualization results and quantitative metrics illustrate that the DIPA can generate more imperceptible adversarial perturbation.
Funder
Natural Science Foundation of Inner Mongolia of China Engineering Research Center of Ecological Big Data, Ministry of Education, the Inner Mongolia Science and Technology Plan Project National Natural Science Foundation of China
Reference36 articles.
1. A fast learning algorithm for deep belief nets;Hinton;Neural Comput.,2006 2. Taigman, Y., Yang, M., Ranzato, M., and Wolf, L. (2014, January 23–28). Deepface: Closing the gap to human-level performance in face verification. Proceedings of the IEEE Conference on Computer Visual and Pattern Recognition (CVPR), Columbus, OH, USA. 3. Szegedy, C., Zaremba, W., Sutskever, I., Bruna, J., Erhan, D., Goodfellow, I., and Fergus, R. (2014, January 14–16). Intriguing properties of neural networks. Proceedings of the International Conference on Learning Representations (ICLR), Banff, AB, Canada. 4. Moosavi-Dezfooli, S.M., Fawzi, A., and Frossard, P. (2016, January 27–30). Deepfool: A simple and accurate method to fool deep neural networks. Proceedings of the IEEE Conference on Computer Visual and Pattern Recognition (CVPR), Las Vegas, NV, USA. 5. Vision and touch: Independent or integrated systems for the perception of texture?;Whitaker;Brain Res.,2008
|
|