Affiliation:
1. The University of Aizu, Fukushima, Japan
2. University of Electro-Communications, Tokyo, Japan
Abstract
Dynamic Graph Neural Network (DGNN) has shown a strong capability of learning dynamic graphs by exploiting both spatial and temporal features. Although DGNN has recently received considerable attention by AI community and various DGNN models have been proposed, building a distributed system for efficient DGNN training is still challenging. It has been well recognized that how to partition the dynamic graph and assign workloads to multiple GPUs plays a critical role in training acceleration. Existing works partition a dynamic graph into snapshots or temporal sequences, which only work well when the graph has uniform spatio-temporal structures. However, dynamic graphs in practice are not uniformly structured, with some snapshots being very dense while others are sparse. To address this issue, we propose DGC, a distributed DGNN training system that achieves a 1.25× - 7.52× speedup over the state-of-the-art in our testbed. DGC's success stems from a new graph partitioning method that partitions dynamic graphs into chunks, which are essentially subgraphs with modest training workloads and few inter connections. This partitioning algorithm is based on graph coarsening, which can run very fast on large graphs. In addition, DGC has a highly efficient run-time, powered by the proposed chunk fusion and adaptive stale aggregation techniques. Extensive experimental results on 3 typical DGNN models and 4 popular dynamic graph datasets are presented to show the effectiveness of DGC.
Funder
Japan Society for the Promotion of Science (JSPS) KAKENHI
Grant-in-Aid for JSPS Fellows
Japan Science and Technology Agency (JST) PRESTO
Publisher
Association for Computing Machinery (ACM)
Reference70 articles.
1. Binary Graph Neural Networks
2. Gecia Bravo Hermsdorff and Lee Gunderson . 2019 . A unifying framework for spectrum-preserving graph sparsification and coarsening . Advances in Neural Information Processing Systems , Vol. 32 (2019). Gecia Bravo Hermsdorff and Lee Gunderson. 2019. A unifying framework for spectrum-preserving graph sparsification and coarsening. Advances in Neural Information Processing Systems, Vol. 32 (2019).
3. DGCL
4. Efficient scaling of dynamic graph neural networks
5. FedGraph: Federated Graph Learning With Intelligent Sampling
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. Efficient Scheduling for Multi-Job Federated Learning Systems with Client Sharing;2023 IEEE Intl Conf on Dependable, Autonomic and Secure Computing, Intl Conf on Pervasive Intelligence and Computing, Intl Conf on Cloud and Big Data Computing, Intl Conf on Cyber Science and Technology Congress (DASC/PiCom/CBDCom/CyberSciTech);2023-11-14