Affiliation:
1. National Engineering Research Center for Big Data Technology and System, Service Computing Technology and System Lab, Cluster and Grid Computing Lab, School of Computer Science and Technology, Huazhong University of Science and Technology, China
2. National University of Singapore, Singapore
3. Michigan Technological University, America
Abstract
Dynamic Graph Neural Network
(DGNN) has recently attracted a significant amount of research attention from various domains, because most real-world graphs are inherently dynamic. Despite many research efforts, for DGNN, existing hardware/software solutions still suffer significantly from redundant computation and memory access overhead, because they need to irregularly access and recompute all graph data of each graph snapshot. To address these issues, we propose an efficient redundancy-aware accelerator,
RACE
, which enables energy-efficient execution of DGNN models. Specifically, we propose a
redundancy-aware incremental execution approach
into the accelerator design for DGNN to instantly achieve the output features of the latest graph snapshot by correctly and incrementally refining the output features of the previous graph snapshot and also enable regular accesses of vertices’ input features. Through traversing the graph on the fly, RACE identifies the vertices that are not affected by graph updates between successive snapshots to reuse these vertices’ states (i.e., their output features) of the previous snapshot for the processing of the latest snapshot. The vertices affected by graph updates are also tracked to incrementally recompute their new states using their neighbors’ input features of the latest snapshot for correctness. In this way, the processing and accessing of many graph data that are not affected by graph updates can be correctly eliminated, enabling smaller redundant computation and memory access overhead. Besides, the input features, which are accessed more frequently, are dynamically identified according to graph topology and are preferentially resident in the on-chip memory for less off-chip communications. Experimental results show that RACE achieves on average 1139× and 84.7× speedups for DGNN inference, with average 2242× and 234.2× energy savings, in comparison with the state-of-the-art software DGNN running on Intel Xeon CPU and NVIDIA A100 GPU, respectively. Moreover, for DGNN inference, RACE obtains on average 13.1×, 11.7×, 10.4×, and 7.9× speedup and 14.8×, 12.9×, 11.5×, and 8.9× energy savings over the state-of-the-art Graph Neural Network accelerators, i.e., AWB-GCN, GCNAX, ReGNN, and I-GCN, respectively.
Funder
National Key Research and Development Program of China
NSFC
Major Scientific Research Project of Zhejiang Lab
CCF-AFSG Research Fund
Young Top-notch Talent Cultivation Program of Hubei Province, Key Research and Development Program of Hubei Province
Knowledge Innovation Program of Wuhan-Basi Research
Publisher
Association for Computing Machinery (ACM)
Subject
Hardware and Architecture,Information Systems,Software
Reference62 articles.
1. 2022. Academic. Retrieved from https://west.uni-koblenz.de/konect/networks
2. 2022. DBLP and Mobile. Retrieved from https://dblp.uni-trier.de/xml/
3. 2022. Flicker. Retrieved from https://socialnetworks.mpi-sws.org/data-imc2007.html
4. 2022. Wikidata. Retrieved from https://github.com/mniepert/mmkb/tree/master/TemporalKGs/wikidata
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. DS-GL: Advancing Graph Learning via Harnessing Nature’s Power within Scalable Dynamical Systems;2024 ACM/IEEE 51st Annual International Symposium on Computer Architecture (ISCA);2024-06-29