An Empirical Study on the Effect of Training Data Perturbations on Neural Network Robustness
Author:
Wang Jie1ORCID, Wu Zili2, Lu Minyan1, Ai Jun1ORCID
Affiliation:
1. The Key Laboratory on Reliability and Environment Engineering Technology, School of Reliability and Systems Engineering, Beihang University, Beijing 100191, China 2. CRRC Zhuzhou Institute Co., Ltd., Zhuzhou 412001, China
Abstract
The vulnerability of modern neural networks to random noise and deliberate attacks has raised concerns about their robustness, particularly as they are increasingly utilized in safety- and security-critical applications. Although recent research efforts were made to enhance robustness through retraining with adversarial examples or employing data augmentation techniques, a comprehensive investigation into the effects of training data perturbations on model robustness remains lacking. This paper presents the first extensive empirical study investigating the influence of data perturbations during model retraining. The experimental analysis focuses on both random and adversarial robustness, following established practices in the field of robustness analysis. Various types of perturbations in different aspects of the dataset are explored, including input, label, and sampling distribution. Single-factor and multi-factor experiments are conducted to assess individual perturbations and their combinations. The findings provide insights into constructing high-quality training datasets for optimizing robustness and recommend the appropriate degree of training set perturbations that balance robustness and correctness, and contribute to understanding model robustness in deep learning and offer practical guidance for enhancing model performance through perturbed retraining, promoting the development of more reliable and trustworthy deep learning systems for safety-critical applications.
Funder
Key Lab of Reliability and Environmental Engineering Technology
Reference46 articles.
1. Szegedy, C., Zaremba, W., Sutskever, I., Bruna, J., Erhan, D., Goodfellow, I., and Fergus, R. (2014). Intriguing properties of neural networks. arXiv. 2. Moosavi-Dezfooli, S.-M., Fawzi, A., and Frossard, P. (2016, January 27–30). DeepFool: A Simple and Accurate Method to Fool Deep Neural Networks. Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA. 3. Goodfellow, I.J., Shlens, J., and Szegedy, C. (2015, January 7–9). Explaining and Harnessing Adversarial Examples. Proceedings of the International Conference on Learning Representations (ICLR), San Diego, CA, USA. 4. Hamdi, A., Müller, M., and Ghanem, B. (2020, January 7–12). SADA: Semantic adversarial diagnostic attacks for autonomous applications. Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA. 5. Uličný, M., Lundström, J., and Byttner, S. (2016). Robustness of Deep Convolutional Neural Networks for Image Recognition, Springer International Publishing.
|
|