Affiliation:
1. Dipartimento di Scienze Pure e Applicate (DiSPeA), Università di Urbino, Urbino, Italy
2. Dipartimento di Informatica – Scienza e Ingegneria (DISI), Università di Bologna, Bologna, Italy
Abstract
Machine learning black boxes, exemplified by deep neural networks, often exhibit challenges in interpretability due to their reliance on complicated relationships involving numerous internal parameters and input features. This lack of transparency from a human perspective renders their predictions untrustworthy, particularly in critical applications. In this paper, we address this issue by introducing the design and implementation of CReEPy, an algorithm for symbolic knowledge extraction based on explainable clustering. Specifically, CReEPy leverages the underlying clustering performed by the ExACT or CREAM algorithms to generate human-interpretable Prolog rules that mimic the behaviour of opaque models. Additionally, we introduce CRASH, an algorithm for the automated tuning of hyper-parameters required by CReEPy. We present experiments evaluating both the human readability and predictive performance of the proposed knowledge-extraction algorithm, employing existing state-of-the-art techniques as benchmarks for comparison in real-world applications.
Reference43 articles.
1. Survey and critique of techniques for extracting rules from trained artificial neural networks;Andrews;Knowledge-Based Systems,1995
2. Explaining black boxes on sequential data using weighted automata;Ayache;International Conference on Grammatical Inference,2019
3. Using neural network rule extraction and decision tables for credit-risk evaluation;Baesens;Management Science,2003
4. Interpretable hierarchical clustering by constructing an unsupervised decision tree;Basak;IEEE Trans. Knowl. Data Eng.,2005
5. Bertsimas D. , Orfanoudaki A. , Wiberg H.M. , Interpretable clustering via optimal trees, CoRR, abs/1812.00539, 2018.