Abstract
AbstractFor predictive computational models to be considered reliable in crucial areas such as biology and medicine, it is essential for them to be accurate, robust, and interpretable. A sufficiently robust model should not have its output affected significantly by a slight change in the input. Also, these models should be able to explain how a decision is made. Efforts have been made to improve the robustness and interpretability of these models as independent challenges, however, the effect of robustness and interpretability on each other is poorly understood. Here, we show that predicting cell type based on single-cell RNA-seq data is more robust by adversarially training a deep learning model. Surprisingly, we find this also leads to improved model interpretability, as measured by identifying genes important for classification. We believe that adversarial training will be generally useful to improve deep learning robustness and interpretability, thereby facilitating biological discovery.
Publisher
Cold Spring Harbor Laboratory
Reference43 articles.
1. Deep learning for computational biology
2. Aging affects circadian clock and metabolism and modulates timing of medication;iScience,2021
3. Simonyan K , Vedaldi A , Zisserman A. Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps. arXiv. 2013;
4. Selvaraju RR , Cogswell M , Das A , Vedantam R , Parikh D , Batra D. Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization. Int J Comput Vis. 2019 Oct 11;
5. Shrikumar A , Greenside P , Kundaje A. Learning Important Features Through Propagating Activation Differences. arXiv. 2017;