Affiliation:
1. Shopee Pte Ltd., Singapore, Singapore
2. Shopee Pte Ltd., Singapore Singapore
3. SCSE, Nanyang Technological University, Singapore Singapore
4. Shopee Pte Ltd., Shanghai China
5. Key Laboratory of Interdisciplinary Research of Computation and Economics, Shanghai University of Finance and Economics, Shanghai China
Abstract
In recent years, recommender systems have advanced rapidly, where embedding learning for users and items plays a critical role. A standard method learns a unique embedding vector for each user and item. However, such a method has two important limitations in real-world applications: 1) it is hard to learn embeddings that generalize well for users and items with rare interactions; and 2) it may incur unbearably high memory costs when the number of users and items scales up. Existing approaches either can only address one of the limitations or have flawed overall performances. In this paper, we propose Clustered Embedding Learning (CEL) as an integrated solution to these two problems. CEL is a plug-and-play embedding learning framework that can be combined with any differentiable feature interaction model. It is capable of achieving improved performance, especially for cold users and items, with reduced memory cost. CEL enables automatic and dynamic clustering of users and items in a top-down fashion, where clustered entities could jointly learn a shared embedding. The accelerated version of CEL has an optimal time complexity, which supports efficient online updates. Theoretically, we prove the identifiability and the existence of a unique optimal number of clusters for CEL in the context of nonnegative matrix factorization. Empirically, we validate the effectiveness of CEL on three public datasets and one business dataset, showing its consistently superior performance against current state-of-the-art methods. In particular, when incorporating CEL into the business model, it brings an improvement of
\(+0.6\% \)
in AUC, which translates into a significant revenue gain; meanwhile, the size of the embedding table gets 2650 times smaller. Additionally, we demonstrate that, if there is enough memory, learning a personalized embedding for each user and item around their clustering centers is feasible and can further boost the performance.
Publisher
Association for Computing Machinery (ACM)
Reference32 articles.
1. eTREE: Learning Tree-structured Embeddings
2. Daniel Boley. 1998. Principal direction divisive partitioning. Data mining and knowledge discovery 2, 4 (1998), 325–344.
3. Wide & Deep Learning for Recommender Systems
4. DeepRec. 2021. Adaptive Embedding. https://github.com/alibaba/DeepRec/blob/main/README.md.
5. Chelsea Finn, Pieter Abbeel, and Sergey Levine. 2017. Model-agnostic meta-learning for fast adaptation of deep networks. In Proc. ICML. PMLR, 1126–1135.
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. Lightweight Embeddings for Graph Collaborative Filtering;Proceedings of the 47th International ACM SIGIR Conference on Research and Development in Information Retrieval;2024-07-10