论文标题

知识包:知识图的代数学习框架

Knowledgebra: An Algebraic Learning Framework for Knowledge Graph

论文作者

Yang, Tong, Wang, Yifei, Sha, Long, Engelbrecht, Jan, Hong, Pengyu

论文摘要

知识图(kg)表示学习旨在将实体和关系编码为密集的连续矢量空间,以便可以一致地表示数据集中包含的知识。从KG数据集训练的密集嵌入式嵌入会受益于各种下游任务,例如完成和链接预测。但是,现有的KG嵌入方法缺乏,为全球知识表示的一致性提供了系统的解决方案。我们基于对它们固有的代数结构的观察,为KG开发了一种数学语言,我们称之为知识。通过分析五个不同的代数属性,我们证明了半群是通用知识图的关系嵌入的最合理的代数结构。我们使用简单的矩阵半群实现了一个实例化模型SEME,该模型在标准数据集中表现出最先进的性能。此外,我们提出了一种基于正规化的方法,将人类知识衍生的类似链的逻辑规则整合到嵌入培训中,进一步证明了发达语言的力量。据我们所知,通过在统计学习中应用抽象代数,这项工作开发了通用知识图的第一种正式语言,并且还从代数角度阐明了神经符号整合的问题。

Knowledge graph (KG) representation learning aims to encode entities and relations into dense continuous vector spaces such that knowledge contained in a dataset could be consistently represented. Dense embeddings trained from KG datasets benefit a variety of downstream tasks such as KG completion and link prediction. However, existing KG embedding methods fell short to provide a systematic solution for the global consistency of knowledge representation. We developed a mathematical language for KG based on an observation of their inherent algebraic structure, which we termed as Knowledgebra. By analyzing five distinct algebraic properties, we proved that the semigroup is the most reasonable algebraic structure for the relation embedding of a general knowledge graph. We implemented an instantiation model, SemE, using simple matrix semigroups, which exhibits state-of-the-art performance on standard datasets. Moreover, we proposed a regularization-based method to integrate chain-like logic rules derived from human knowledge into embedding training, which further demonstrates the power of the developed language. As far as we know, by applying abstract algebra in statistical learning, this work develops the first formal language for general knowledge graphs, and also sheds light on the problem of neural-symbolic integration from an algebraic perspective.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源