ADGNN: Towards Scalable GNN Training with Aggregation-Difference Aware Sampling

Author:

Song Zhen1ORCID,Gu Yu1ORCID,Li Tianyi2ORCID,Sun Qing1ORCID,Zhang Yanfeng1ORCID,Jensen Christian S.2ORCID,Yu Ge1ORCID

Affiliation:

1. Northeastern University, Shenyang, China

2. Aalborg University, Aalborg, Denmark

Abstract

Distributed computing is promising to enable large-scale graph neural network (GNN) model training. However, care is needed to avoid excessive computational and communication overheads. Sampling is promising in terms of enabling scalability, and sampling techniques have been proposed to reduce training costs. However, online sampling introduces large overheads, and while offline sampling that is done only once can eliminate such overheads, it instead introduces information loss and accuracy degradation. Thus, existing sampling techniques are unable to improve simultaneously both efficiency and accuracy, particularly at low sampling rates. We develop a distributed system, ADGNN, for full-batch based GNN training that adopts a hybrid sampling architecture to enable a trade-off between efficiency and accuracy. Specifically, ADGNN employs sampling result reuse techniques to reduce the cost associated with sampling and thus improve training efficiency. To alleviate accuracy degradation, we introduce a new metric,Aggregation Difference (AD), that quantifies the gap between sampled and full neighbor set aggregation. We present so-called AD-Sampling that aims to minimize the Aggregation Difference with an adaptive sampling frequency tuner. Finally, ADGNN employs anAD -importance-based sampling technique for remote neighbors to further reduce communication costs. Experiments on five real datasets show that ADGNN is able to outperform the state-of-the-art by up to nearly 9 times in terms of efficiency, while achieving comparable accuracy to the non-sampling methods.

Funder

the National Natural Science Foundation of China

the Fundamental Research Funds for the Central Universities

Publisher

Association for Computing Machinery (ACM)

Reference53 articles.

1. Jiyang Bai Yuxiang Ren and Jiawei Zhang. 2021. Ripple Walk Training: A Subgraph-based Training Framework for Large and Deep Graph Neural Network. In IJCNN. 1--8. Jiyang Bai Yuxiang Ren and Jiawei Zhang. 2021. Ripple Walk Training: A Subgraph-based Training Framework for Large and Deep Graph Neural Network. In IJCNN. 1--8.

2. Muhammed Fatih Balin and Ü mit V. cC atalyü rek . 2022 . (LA)yer-neigh(BOR) Sampling : Defusing Neighborhood Explosion in GNNs. CoRR , Vol. abs/ 2210 .13339 (2022). Muhammed Fatih Balin and Ü mit V. cC atalyü rek. 2022. (LA)yer-neigh(BOR) Sampling: Defusing Neighborhood Explosion in GNNs. CoRR, Vol. abs/2210.13339 (2022).

3. Huiyuan Chen , Chin-Chia Michael Yeh , Fei Wang, and Hao Yang. 2022 . Graph Neural Transport Networks with Non-local Attentions for Recommender Systems. In WWW. 1955--1964. Huiyuan Chen, Chin-Chia Michael Yeh, Fei Wang, and Hao Yang. 2022. Graph Neural Transport Networks with Non-local Attentions for Recommender Systems. In WWW. 1955--1964.

4. Jie Chen Tengfei Ma and Cao Xiao. 2018a. FastGCN: Fast Learning with Graph Convolutional Networks via Importance Sampling. In ICLR. Jie Chen Tengfei Ma and Cao Xiao. 2018a. FastGCN: Fast Learning with Graph Convolutional Networks via Importance Sampling. In ICLR.

5. Jianfei Chen Jun Zhu and Le Song. 2018b. Stochastic Training of Graph Convolutional Networks with Variance Reduction. In ICML. 941--949. Jianfei Chen Jun Zhu and Le Song. 2018b. Stochastic Training of Graph Convolutional Networks with Variance Reduction. In ICML. 941--949.

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3