TransE-MTP: A New Representation Learning Method for Knowledge Graph Embedding with Multi-Translation Principles and TransE
-
Published:2024-08-11
Issue:16
Volume:13
Page:3171
-
ISSN:2079-9292
-
Container-title:Electronics
-
language:en
-
Short-container-title:Electronics
Author:
Li Yongfang123ORCID, Zhu Chunhua123ORCID
Affiliation:
1. College of Information Science and Engineering, Henan University of Technology, Zhengzhou 450001, China 2. Key Laboratory of Grain Information Processing and Control, Henan University of Technology, Ministry of Education, Zhengzhou 450001, China 3. Henan Engineering Research Center of Grain Condition Intelligent Detection and Application, Henan University of Technology, Zhengzhou 450001, China
Abstract
The purpose of representation learning is to encode the entities and relations in a knowledge graph as low-dimensional and real-valued vectors through machine learning technology. Traditional representation learning methods like TransE, a method which models relationships by interpreting them as translations operating on the low-dimensional embeddings of a graph’s entities, are effective for learning the embeddings of knowledge bases, but struggle to effectively model complex relations like one-to-many, many-to-one, and many-to-many. To overcome the above issues, we introduce a new method for knowledge representation, reasoning, and completion based on multi-translation principles and TransE (TransE-MTP). By defining multiple translation principles (MTPs) for different relation types, such as one-to-one and complex relations like one-to-many, many-to-one, and many-to-many, and combining MTPs with a typical translating-based model for modeling multi-relational data (TransE), the proposed method, TransE-MTP, ensures that multiple optimization objectives can be targeted and optimized during training on complex relations, thereby providing superior prediction performance. We implement a prototype of TransE-MTP to demonstrate its effectiveness at link prediction and triplet classification on two prominent knowledge graph datasets: Freebase and Wordnet. Our experimental results show that the proposed method enhanced the performance of both TransE and knowledge graph embedding by translating on hyperplanes (TransH), which confirms its effectiveness and competitiveness.
Funder
National Natural Science Foundation of China Open Subject of Scientific Research Platform in Grain Information Processing Center Innovative Funds Plan of Henan University of Technology
Reference49 articles.
1. WordNet: A lexical database for English;Miller;Commun. ACM,1995 2. Bollacker, K., Evans, C., Paritosh, P., Sturge, T., and Taylor, J. (2008, January 9–12). Freebase: A collaboratively created graph database for structuring human knowledge. Proceedings of the 2008 ACM SIGMOD International Conference on Management of Data, Vancouver, BC, Canada. 3. Goyal, Y., Khot, T., Summers-Stay, D., Batra, D., and Parikh, D. (2017, January 21–26). Making the V in VQA matter: Elevating the role of image understanding in visual question answering. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA. 4. Xiao, H., Huang, M., and Zhu, X. (2016, January 9–15). From one point to a manifold: Knowledge graph embedding for precise link prediction. Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence, New York, NY, USA. 5. Green, B.F., Wolf, A.K., Chomsky, C., and Laughery, K. (1960, January 3–5). Baseball: An automatic question-answerer. Proceedings of the Western Joint IRE-AIEE-ACM Computer Conference, San Francisco, CA, USA.
|
|