SLISEMAP: supervised dimensionality reduction through local explanations
-
Published:2022-11-23
Issue:1
Volume:112
Page:1-43
-
ISSN:0885-6125
-
Container-title:Machine Learning
-
language:en
-
Short-container-title:Mach Learn
Author:
Björklund AntonORCID, Mäkelä JarmoORCID, Puolamäki KaiORCID
Abstract
AbstractExisting methods for explaining black box learning models often focus on building local explanations of the models’ behaviour for particular data items. It is possible to create global explanations for all data items, but these explanations generally have low fidelity for complex black box models. We propose a new supervised manifold visualisation method, slisemap, that simultaneously finds local explanations for all data items and builds a (typically) two-dimensional global visualisation of the black box model such that data items with similar local explanations are projected nearby. We provide a mathematical derivation of our problem and an open source implementation implemented using the GPU-optimised PyTorch library. We compare slisemap to multiple popular dimensionality reduction methods and find that slisemap is able to utilise labelled data to create embeddings with consistent local white box models. We also compare slisemap to other model-agnostic local explanation methods and show that slisemap provides comparable explanations and that the visualisations can give a broader understanding of black box regression and classification models.
Funder
Academy of Finland Helsingin Yliopisto University of Helsinki including Helsinki University Central Hospital
Publisher
Springer Science and Business Media LLC
Subject
Artificial Intelligence,Software
Reference57 articles.
1. Adler, P., Falk, C., Friedler, S. A., Nix, T., Rybeck, G., Scheidegger, C., et al. (2018). Auditing black-box models for indirect influence. Knowledge and Information Systems, 54(1), 95–122. https://doi.org/10.1007/s10115-017-1116-3. 2. Ali, S. M., & Silvey, S. D. (1966). A general class of coefficients of divergence of one distribution from another. Journal of the Royal Statistical Society: Series B (Methodological), 28(1), 131–142. https://doi.org/10.1111/j.2517-6161.1966.tb00626.x. 3. Anbtawi, W. (2019). A 3D Playground for t-SNE With explainable classification. PhD thesis, American University of Beirut, Beirut, Lebanon. 4. Anders, F., Chiappini, C., Santiago, B. X., Matijevič, G., Queiroz, A. B., Steinmetz, M., & Guiglion, G. (2018). Dissecting stellar chemical abundance space with t-SNE. Astronomy & Astrophysics, 619, A125. https://doi.org/10.1051/0004-6361/201833099. 5. Baehrens, D., Schroeter, T., Harmeling, S., Kawanabe, M., Hansen, K., & Müller, K. R. (2010). How to explain individual classification decisions. Journal of Machine Learning Research, 11(61), 1803–1831.
Cited by
10 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
|
|