Affiliation:
1. Tsinghua University & Tencent, Beijing, China
2. Tsinghua University, Beijing, China
3. University of Science and Technology of China, Hefei, China
4. National University of Singapore, Singapore
Abstract
As recommender systems become increasingly important in daily human decision-making, users are demanding convincing explanations to understand why they get the specific recommendation results. Although a number of explainable recommender systems have recently been proposed, there still lacks an understanding of what users really need in a recommendation explanation. The actual reason behind users’ intention to examine and consume (e.g., click and watch a movie) can be the window to answer this question and is named as
self-explanation
in this work. In addition, humans usually make recommendations accompanied by explanations, but there remain fewer studies on how humans explain and what we can learn from human-generated explanations.
To investigate these questions, we conduct a novel multi-role, multi-session user study in which users interact with multiple types of system-generated explanations as well as human-generated explanations, namely
peer-explanation
. During the study, users’ intentions, expectations, and experiences are tracked in several phases, including before and after the users are presented with an explanation and after the content is examined. Through comprehensive investigations, three main findings have been made: First, we observe not only the positive but also the negative effects of explanations, and the impact varies across different types of explanations. Moreover, human-generated explanation,
peer-explanation
, performs better in increasing user intentions and helping users to better construct preferences, which results in better user satisfaction. Second, based on users’
self-explanation
, the information accuracy is measured and found to be a major factor associated with user satisfaction. Some other factors, such as unfamiliarity and similarity, are also discovered and summarized. Third, through annotations of the information aspects used in the human-generated
self-explanation
and
peer-explanation
, patterns of how humans explain are investigated, including what information and how much information is utilized. In addition, based on the findings, a human-inspired explanation approach is proposed and found to increase user satisfaction, revealing the potential improvement of further incorporating more human patterns in recommendation explanations.
These findings have shed light on the deeper understanding of the recommendation explanation and further research on its evaluation and generation. Furthermore, the collected data, including human-generated explanations by both the external peers and the users’ selves, will be released to support future research works on explanation evaluation.
Funder
Natural Science Foundation of China
Tsinghua University Guoqiang Research Institute
Publisher
Association for Computing Machinery (ACM)
Subject
Computer Science Applications,General Business, Management and Accounting,Information Systems
Cited by
9 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. Explaining Session-based Recommendations using Grammatical Evolution;Proceedings of the Genetic and Evolutionary Computation Conference Companion;2024-07-14
2. On the Negative Perception of Cross-domain Recommendations and Explanations;Proceedings of the 47th International ACM SIGIR Conference on Research and Development in Information Retrieval;2024-07-10
3. Balanced Explanations in Recommender Systems;Adjunct Proceedings of the 32nd ACM Conference on User Modeling, Adaptation and Personalization;2024-06-27
4. Leveraging ChatGPT for Automated Human-centered Explanations in Recommender Systems;Proceedings of the 29th International Conference on Intelligent User Interfaces;2024-03-18
5. A Survey on Explainable Course Recommendation Systems;Lecture Notes in Computer Science;2024