Affiliation:
1. Sun Yat-sen University, Guangzhou, China
2. Sun Yat-sen University, China and Mohamed bin Zayed University of Artificial Intelligence, Abu Dhabi, United Arab Emirates
3. University of Cambridge, Trinity Ln, Cambridge
4. University of Amsterdam, Science Park, The Netherlands
Abstract
Network-based information has been widely explored and exploited in the information retrieval literature. Attributed networks, consisting of nodes, edges as well as attributes describing properties of nodes, are a basic type of network-based data, and are especially useful for many applications. Examples include user profiling in social networks and item recommendation in user-item purchase networks. Learning useful and expressive representations of entities in attributed networks can provide more effective building blocks to down-stream network-based tasks such as link prediction and attribute inference. Practically, input features of attributed networks are normalized as unit directional vectors. However, most network embedding techniques ignore the
spherical
nature of inputs and focus on learning representations in a Gaussian or Euclidean space, which, we hypothesize, might lead to less effective representations. To obtain more effective representations of attributed networks, we investigate the problem of mapping an attributed network with unit normalized directional features into a non-Gaussian and non-Euclidean space. Specifically, we propose a hyperspherical variational co-embedding for attributed networks (HCAN), which is based on generalized variational auto-encoders for heterogeneous data with multiple types of entities. HCAN jointly learns latent embeddings for both nodes and attributes in a unified hyperspherical space such that the affinities between nodes and attributes can be captured effectively. We argue that this is a crucial feature in many real-world applications of attributed networks. Previous Gaussian network embedding algorithms break the assumption of uninformative prior, which leads to unstable results and poor performance. In contrast, HCAN embeds nodes and attributes as von Mises-Fisher distributions, and allows one to capture the uncertainty of the inferred representations. Experimental results on eight datasets show that HCAN yields better performance in a number of applications compared with nine state-of-the-art baselines.
Funder
National Natural Science Foundation of China
Innovation Center for Artificial Intelligence
Publisher
Association for Computing Machinery (ACM)
Subject
Computer Science Applications,General Business, Management and Accounting,Information Systems
Reference73 articles.
1. Deep gaussian embedding of graphs: Unsupervised inductive learning via ranking;Bojchevski Aleksandar;Proceedings of the International Conference on Learning Representations,2018
2. Samuel R. Bowman Luke Vilnis Oriol Vinyals Andrew M. Dai Rafal Jozefowicz and Samy Bengio. 2016. Generating sentences from a continuous space. In Proceedings of the 20th SIGNLL Conference on Computational Natural Language Learning . 10–21.
3. Deepayan Chakrabarti Stanislav Funiak Jonathan Chang and Sofus A. Macskassy. 2014. Joint inference of multiple label types in large networks. In Proceedings of the 31th International Conference on Machine Learning Vol. 32. 874–882.
4. Ines Chami Sami Abu-El-Haija Bryan Perozzi Christopher Ré and Kevin Murphy. 2020. Machine learning on graphs: A model and comprehensive taxonomy. arXiv preprint arXiv:2005.03675 . 1–44.
Cited by
7 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献