Affiliation:
1. The Hong Kong University of Science and Technology, Hong Kong, China
2. Shanghai Jiao Tong University, Shanghai, China
3. The Hong Kong University of Science and Technology & The Hong Kong University of Science and Technology (Guangzhou), Hong Kong & Guangzhou, China
4. Huawei Noah's Ark Lab, Hong Kong, China
Abstract
Representation learning over dynamic graphs is critical for many real-world applications such as social network services and recommender systems. Temporal graph neural networks (T-GNNs) are powerful representation learning methods and have achieved remarkable effectiveness on continuous-time dynamic graphs. However, T-GNNs still suffer from high time complexity, which increases linearly with the number of timestamps and grows exponentially with the model depth, causing them not scalable to large dynamic graphs. To address the limitations, we propose Orca, a novel framework that accelerates T-GNN training by non-trivially caching and reusing intermediate embeddings. We design an optimal cache replacement algorithm, named MRU, under a practical cache limit. MRU not only improves the efficiency of training T-GNNs by maximizing the number of cache hits but also reduces the approximation errors by avoiding keeping and reusing extremely stale embeddings. Meanwhile, we develop profound theoretical analyses of the approximation error introduced by our reuse schemes and offer rigorous convergence guarantees. Extensive experiments have validated that Orca can obtain two orders of magnitude speedup over the state-of-the-art baselines while achieving higher precision on large dynamic graphs.
Funder
Hong Kong ITC ITF
Hong Kong RGC AOE Project
Hong Kong RGC GRF Project
National Key Research and Development Program of China
Shanghai Municipal Science and Technology Major Project
National Science Foundation of China
Hong Kong RGC CRF Project
Guangdong Basic and Applied Basic Research Foundation
SJTU Global Strategic Partnership Fund
Hong Kong RGC Theme-based project
China NSFC
Microsoft Research Asia Collaborative Research Grant
HKUST-Webank joint research lab grant
HKUST Global Strategic Partnership Fund
Publisher
Association for Computing Machinery (ACM)
Reference75 articles.
1. 2023. AskUbuntu. http://snap.stanford.edu/data/sx-askubuntu.html. 2023. AskUbuntu. http://snap.stanford.edu/data/sx-askubuntu.html.
2. 2023. SuperUser. http://snap.stanford.edu/data/sx-superuser.html. 2023. SuperUser. http://snap.stanford.edu/data/sx-superuser.html.
3. 2023. The technical report. https://github.com/LuckyLYM/Orca/blob/main/technical_report.pdf. 2023. The technical report. https://github.com/LuckyLYM/Orca/blob/main/technical_report.pdf.
4. 2023. Wiki-talk. http://snap.stanford.edu/data/wiki-talk-temporal.html. 2023. Wiki-talk. http://snap.stanford.edu/data/wiki-talk-temporal.html.
5. 2023. Wikipedia edit history dump. https://meta.wikimedia.org/wiki/Data_dumps. 2023. Wikipedia edit history dump. https://meta.wikimedia.org/wiki/Data_dumps.
Cited by
3 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. TimeSGN: Scalable and Effective Temporal Graph Neural Network;2024 IEEE 40th International Conference on Data Engineering (ICDE);2024-05-13
2. Incorporating Dynamic Temperature Estimation into Contrastive Learning on Graphs;2024 IEEE 40th International Conference on Data Engineering (ICDE);2024-05-13
3. ADGNN: Towards Scalable GNN Training with Aggregation-Difference Aware Sampling;Proceedings of the ACM on Management of Data;2023-12-08