Efficient Non-Sampling Graph Neural Networks
-
Published:2023-07-25
Issue:8
Volume:14
Page:424
-
ISSN:2078-2489
-
Container-title:Information
-
language:en
-
Short-container-title:Information
Author:
Ji Jianchao1, Li Zelong1ORCID, Xu Shuyuan1, Ge Yingqiang1, Tan Juntao1, Zhang Yongfeng1
Affiliation:
1. Computer Science, Rutgers University, Piscataway, NJ 08854, USA
Abstract
A graph is a widely used and effective data structure in many applications; it describes the relationships among nodes or entities. Currently, most semi-supervised or unsupervised graph neural network models are trained based on a very basic operation called negative sampling. Usually, the purpose of the learning objective is to maximize the similarity between neighboring nodes while minimizing the similarity between nodes that are not close to each other. Negative sampling can reduce the time complexity by sampling a small fraction of the negative nodes instead of using all of the negative nodes when optimizing the objective. However, sampling of the negative nodes may fail to deliver stable model performance due to the uncertainty in the sampling procedure. To avoid such disadvantages, we provide an efficient Non-Sampling Graph Neural Network (NS-GNN) framework. The main idea is to use all the negative samples when optimizing the learning objective to avoid the sampling process. Of course, directly using all of the negative samples may cause a large increase in the model training time. To mitigate this problem, we rearrange the origin loss function into a linear form and take advantage of meticulous mathematical derivation to reduce the complexity of the loss function. Experiments on benchmark datasets show that our framework can provide better efficiency at the same level of prediction accuracy compared with existing negative sampling-based models.
Subject
Information Systems
Reference40 articles.
1. Perozzi, B., Al-Rfou, R., and Skiena, S. (2014, January 24–27). Deepwalk: Online learning of social representations. Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, New York, NY, USA. 2. Yang, L., Gu, J., Wang, C., Cao, X., Zhai, L., Jin, D., and Guo, Y. (2020, January 17–20). Toward Unsupervised Graph Neural Network: Interactive Clustering and Embedding via Optimal Transport. Proceedings of the 2020 IEEE International Conference on Data Mining (ICDM), Sorrento, Italy. 3. Henaff, M., Bruna, J., and LeCun, Y. (2015). Deep convolutional networks on graph-structured data. arXiv. 4. Modeling dynamic missingness of implicit feedback for recommendation;Wang;Adv. Neural Inf. Process. Syst.,2018 5. Rendle, S. (2010, January 13–17). Factorization machines. Proceedings of the 2010 IEEE International Conference on Data Mining, Sydney, Australia.
|
|