Author:
Marques-Silva Joao,Ignatiev Alexey
Abstract
Recent years witnessed a number of proposals for the use of the so-called interpretable models in specific application domains. These include high-risk, but also safety-critical domains. In contrast, other works reported some pitfalls of machine learning model interpretability, in part justified by the lack of a rigorous definition of what an interpretable model should represent. This study proposes to relate interpretability with the ability of a model to offer explanations of why a prediction is made given some point in feature space. Under this general goal of offering explanations to predictions, this study reveals additional limitations of interpretable models. Concretely, this study considers application domains where the purpose is to help human decision makers to understand why some prediction was made or why was not some other prediction made, and where irreducible (and so minimal) information is sought. In such domains, this study argues that answers to such why (or why not) questions can exhibit arbitrary redundancy, i.e., the answers can be simplified, as long as these answers are obtained by human inspection of the interpretable ML model representation.
Funder
Agence Nationale de la Recherche
European Commission
Reference98 articles.
1. “Non-monotonic explanation functions,”;Amgoud,2021
2. “Axiomatic foundations of explainability,”;Amgoud,2022
3. “Foundations of symbolic languages for model interpretability,”;Arenas,2021
4. On computing probabilistic explanations for decision trees;Arenas;CoRR,2022
5. “Fair and adequate explanations,”;Asher;CD-MAKE,2021
Cited by
5 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献