Author:
Zhao Yongbing,Shao Jinfeng,Asmann Yan W
Abstract
AbstractExplainable artificial intelligence aims to interpret how the machine learning models make decisions, and many model explainers have been developed in the computer vision field. However, the understandings of the applicability of these model explainers to biological data are still lacking. In this study, we comprehensively evaluated multiple explainers by interpreting pretrained models of predicting tissue types from transcriptomic data, and by identifying top contributing genes from each sample with the greatest impacts on model prediction. To improve the reproducibility and interpretability of results generated by model explainers, we proposed a series of optimization strategies for each explainer on two different model architectures of Multilayer Perceptron (MLP) and Convolutional Neural Network (CNN). We observed three groups of explainer and model architecture combinations with high reproducibility. Group II, which contains three model explainers on aggregated MLP models, identified top contributing genes in different tissues that exhibited tissue-specific manifestation and were potential cancer biomarkers. In summary, our work provides novel insights and guidance for exploring biological mechanisms using explainable machine learning models.
Publisher
Cold Spring Harbor Laboratory