Affiliation:
1. Emory University, Atlanta, USA
2. George Mason University, Fairfax, USA
Abstract
As the societal impact of Deep Neural Networks (DNNs) grows, the goals for advancing DNNs become more complex and diverse, ranging from improving a conventional model accuracy metric to infusing advanced human virtues such as fairness, accountability, transparency, and unbiasedness. Recently, techniques in Explainable Artificial Intelligence (XAI) have been attracting considerable attention and have tremendously helped Machine Learning (ML) engineers in understand AI models. However, at the same time, we started to witness the emerging need beyond XAI among AI communities; based on the insights learned from XAI, how can we better empower ML engineers in steering their DNNs so that the model’s reasonableness and performance can be improved as intended? This article provides a timely and extensive literature overview of the field Explanation-Guided Learning (EGL), a domain of techniques that steer the DNNs’ reasoning process by adding regularization, supervision, or intervention on model explanations. In doing so, we first provide a formal definition of EGL and its general learning paradigm. Second, an overview of the key factors for EGL evaluation, as well as summarization and categorization of existing evaluation procedures and metrics for EGL are provided. Finally, the current and potential future application areas and directions of EGL are discussed, and an extensive experimental study is presented aiming at providing comprehensive comparative studies among existing EGL models in various popular application domains, such as Computer Vision and Natural Language Processing domains. Additional resources related to event prediction are included in the article website:
https://kugaoyang.github.io/EGL/
Funder
National Science Foundation
Cisco Faculty Research Award
Oracle for Research Grant Award
Amazon Research Award
NVIDIA GPU
Design Knowledge Company
Publisher
Association for Computing Machinery (ACM)
Reference184 articles.
1. Overview and findings from the religious orders study;Bennett David A.;Curr. Alzheimer Res.,2012
2. Peeking inside the black-box: A survey on explainable artificial intelligence (XAI);Adadi Amina;IEEE Access,2018
3. Sanity checks for saliency maps;Adebayo Julius;Advances in Neural Information Processing Systems,2018
4. Aishwarya Agrawal, Dhruv Batra, Devi Parikh, and Aniruddha Kembhavi. 2018. Don’t just assume; look and answer: Overcoming priors for visual question answering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 4971–4980.
5. Towards robust interpretability with self-explaining neural networks;Melis David Alvarez;Advances in Neural Information Processing Systems,2018
Cited by
8 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献