Affiliation:
1. School of Computer Science, Tel Aviv University , Tel Aviv 69978, Israel
Abstract
Abstract
Motivation
Graph representation learning is a fundamental problem in the field of data science with applications to integrative analysis of biological networks. Previous work in this domain was mostly limited to shallow representation techniques. A recent deep representation technique, BIONIC, has achieved state-of-the-art results in a variety of tasks but used arbitrarily defined components.
Results
Here, we present BERTwalk, an unsupervised learning scheme that combines the BERT masked language model with a network propagation regularization for graph representation learning. The transformation from networks to texts allows our method to naturally integrate different networks and provide features that inform not only nodes or edges but also pathway-level properties. We show that our BERTwalk model outperforms BIONIC, as well as four other recent methods, on two comprehensive benchmarks in yeast and human. We further show that our model can be utilized to infer functional pathways and their effects.
Availability and implementation
Code and data are available at https://github.com/raminass/BERTwalk.
Contact
roded@tauex.tau.ac.il
Funder
Edmond J. Safra Center for Bioinformatics at Tel-Aviv University
Zimin Institute for Engineering Solutions Advancing Better Lives
United States—Israel Binational Science Foundation
Publisher
Oxford University Press (OUP)
Subject
Computer Science Applications,Genetics,Molecular Biology,Structural Biology
Cited by
3 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献