Affiliation:
1. University of Science and Technology of China
2. Chongqing University
3. Zhejiang University
4. Sichuan University
5. Kuaishou Technology Co., Ltd.
Abstract
While personalization increases the utility of recommender systems, it also brings the issue of
filter bubbles
. e.g., if the system keeps exposing and recommending the items that the user is interested in, it may also make the user feel bored and less satisfied. Existing work studies filter bubbles in static recommendation, where the effect of overexposure is hard to capture. In contrast, we believe it is more meaningful to study the issue in interactive recommendation and optimize long-term user satisfaction. Nevertheless, it is unrealistic to train the model online due to the high cost. As such, we have to leverage offline training data and disentangle the causal effect on user satisfaction.
To achieve this goal, we propose a counterfactual interactive recommender system (CIRS) that augments offline reinforcement learning (offline RL) with causal inference. The basic idea is to first learn a causal user model on historical data to capture the overexposure effect of items on user satisfaction. It then uses the learned causal user model to help the planning of the RL policy. To conduct evaluation offline, we innovatively create an authentic RL environment (KuaiEnv) based on a real-world fully observed user rating dataset. The experiments show the effectiveness of CIRS in bursting filter bubbles and achieving long-term success in interactive recommendation. The implementation of CIRS is available via https://github.com/chongminggao/ CIRS-codes.
Funder
National Key Research and Development Program of China
National Natural Science Foundation of China
CCCD Key Lab of Ministry of Culture and Tourism
Publisher
Association for Computing Machinery (ACM)
Subject
Computer Science Applications,General Business, Management and Accounting,Information Systems
Reference103 articles.
1. Deconstructing the Filter Bubble: User Decision-Making and Recommender Systems
2. Xueying Bai Jian Guan and Hongning Wang. 2019. A Model-based reinforcement learning with adversarial training for online recommendation. In Proceedings of the 33rd International Conference on Neural Information Processing Systems . Curran Associates Inc. Red Hook NY 10735–10746.
3. Elias Bareinboim. 2020. Causal reinforcement learning. In Proceedings of the ICML 2020 Tutorial.
4. Causal embeddings for recommendation
5. Axel Bruns. 2019. Are Filter Bubbles Real?John Wiley & Sons.
Cited by
14 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献