GE 2 : A General and Efficient Knowledge Graph Embedding Learning System

Author:

Zheng Chenguang1ORCID,Jiang Guanxian1ORCID,Yan Xiao2ORCID,Yin Peiqi1ORCID,Zhou Qihui1ORCID,Cheng James1ORCID

Affiliation:

1. The Chinese University of Hong Kong, Hong Kong, Hong Kong

2. Centre for Perceptual and Interactive Intelligence (CPII), Hong Kong, Hong Kong

Abstract

Graph embedding learning computes an embedding vector for each node in a graph and finds many applications in areas such as social networks, e-commerce, and medicine. We observe that existing graph embedding systems (e.g., PBG, DGL-KE, and Marius) have long CPU time and high CPU-GPU communication overhead, especially when using multiple GPUs. Moreover, it is cumbersome to implement negative sampling algorithms on them, which have many variants and are crucial for model quality. We propose a new system called GE 2 , which achieves both <u>g</u>enerality and <u>e</u>fficiency for <u>g</u>raph <u>e</u>mbedding learning. In particular, we propose a general execution model that encompasses various negative sampling algorithms. Based on the execution model, we design a user-friendly API that allows users to easily express negative sampling algorithms. To support efficient training, we offload operations from CPU to GPU to enjoy high parallelism and reduce CPU time. We also design COVER, which, to our knowledge, is the first algorithm to manage data swap between CPU and multiple GPUs for small communication costs. Extensive experimental results show that, comparing with the state-of-the-art graph embedding systems, GE 2 trains consistently faster across different models and datasets, where the speedup is usually over 2x and can be up to 7.5x.

Funder

The Research Matching Grant Scheme (RMGS) of Hong Kong

The University Grants Committee of Hong Kong

Publisher

Association for Computing Machinery (ACM)

Reference69 articles.

1. Antoine Bordes, Nicolas Usunier, Alberto Garc'i a-Durá n, Jason Weston, and Oksana Yakhnenko. 2013. Translating Embeddings for Modeling Multi-relational Data. In Annual Conference on Neural Information Processing Systems 2013. December 5--8, 2013, Lake Tahoe, Nevada, United States. 2787--2795. https://proceedings.neurips.cc/paper/2013/hash/1cecc7a77928ca8133fa24680a88d2f9-Abstract.html

2. Knowledge Graph-based Event Embedding Framework for Financial Quantitative Investments

3. Jingtao Ding, Yuhan Quan, Quanming Yao, Yong Li, and Depeng Jin. 2020. Simplify and Robustify Negative Sampling for Implicit Collaborative Filtering. In Annual Conference on Neural Information Processing Systems 2020, December 6--12, 2020, virtual. https://proceedings.neurips.cc/paper/2020/hash/0c7119e3a6a2209da6a5b90e5b5b75bd-Abstract.html

4. Self-Contrastive Learning with Hard Negative Sampling for Self-supervised Point Cloud Learning

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3