Affiliation:
1. Kyushu University
2. University of Alberta
3. Centre for Frontier AI Research (CFAR), A*STAR, and Institute of High Performance Computing (IHPC), A*STAR
4. Meta AI
5. University of Alberta and The University of Tokyo
Abstract
Over the past few years, deep neural networks (DNNs) have achieved tremendous success and have been continuously applied in many application domains. However, during the practical deployment in industrial tasks, DNNs are found to be erroneous-prone due to various reasons such as overfitting and lacking of robustness to real-world corruptions during practical usage. To address these challenges, many recent attempts have been made to repair DNNs for version updates under practical operational contexts by updating weights (i.e., network parameters) through retraining, fine-tuning, or direct weight fixing at a neural level. Nevertheless, existing solutions often neglect the effects of neural network architecture and weight relationships across neurons and layers. In this work, as the first attempt, we initiate to repair DNNs by jointly optimizing the architecture and weights at a higher (i.e., block level).
We first perform empirical studies to investigate the limitation of whole network-level and layer-level repairing, which motivates us to explore a novel repairing direction for DNN repair at the block level. To this end, we need to further consider techniques to address two key technical challenges, i.e.,
block localization
, where we should localize the targeted block that we need to fix; and how to perform
joint architecture and weight repairing
. Specifically, we first propose
adversarial-aware spectrum analysis for vulnerable block localization
that considers the neurons’ status and weights’ gradients in blocks during the forward and backward processes, which enables more accurate candidate block localization for repairing even under a few examples. Then, we further propose the
architecture-oriented search-based repairing
that relaxes the targeted block to a continuous repairing search space at higher deep feature levels. By jointly optimizing the architecture and weights in that space, we can identify a much better block architecture. We implement our proposed repairing techniques as a tool, named
ArchRepair
, and conduct extensive experiments to validate the proposed method. The results show that our method can not only repair but also enhance accuracy and robustness, outperforming the state-of-the-art DNN repair techniques.
Funder
JST-Mirai Program
JSPS KAKENHI
University Fellowships Toward the Creation of Science Technology Innovation
Canada CIFAR AI Chairs Program and the Natural Sciences and Engineering Research Council of Canada
Publisher
Association for Computing Machinery (ACM)
Cited by
4 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. Search-Based Repair of DNN Controllers of AI-Enabled Cyber-Physical Systems Guided by System-Level Specifications;Proceedings of the Genetic and Evolutionary Computation Conference;2024-07-14
2. LUNA: A Model-Based Universal Analysis Framework for Large Language Models;IEEE Transactions on Software Engineering;2024-07
3. More is Not Always Better: Exploring Early Repair of DNNs;Proceedings of the 5th IEEE/ACM International Workshop on Deep Learning for Testing and Testing for Deep Learning;2024-04-20
4. Technical Briefing on Deep Neural Network Repair;Proceedings of the 2024 IEEE/ACM 46th International Conference on Software Engineering: Companion Proceedings;2024-04-14