论文标题
实体和关系的学习代表
Learning Representations of Entities and Relations
论文作者
论文摘要
通过知识图表示模型所学的实体和二进制关系,将事实编码为实体和二进制关系,对各种任务有用,包括预测新事实,问题回答,事实检查和信息检索。本文的重点是(i)改善知识图表示,目的是解决链接预测任务; (ii)设计一个关于语义如何在关系表示的几何形状中捕获的理论。大多数知识图是非常不完整的,并且手动添加新信息的成本很高,这驱动了可以自动推断缺失事实的方法的开发。本论文的第一个贡献是Hyper,这是一个卷积模型,可以简化和改进现有的卷积最新模型的链接预测性能传达,并且可以通过数学上的张力分解来进行数学解释。第二个贡献是塔克(Tucker),塔克(Tucker)是一种相对简单的线性模型,在引入时,该模型在标准数据集中获得了最新的链接预测性能。第三个贡献是MURP,第一个嵌入在双曲空间中的第一个多关系图表模型。 Murp的表现优于所有现有模型及其Euclidean的同行在层次知识关系中的链接预测中的表现,同时需要更少的维度。尽管开发了许多知识图表示模型,这些模型逐渐提高了预测性能,但对他们所学的潜在结构的了解相对较少。我们将对相似性,释义和类比的语义关系的最新理解推广到在单词嵌入的几何相互作用中编码如何在知识图中发现的更多一般关系在其表示中如何编码。
Encoding facts as representations of entities and binary relationships between them, as learned by knowledge graph representation models, is useful for various tasks, including predicting new facts, question answering, fact checking and information retrieval. The focus of this thesis is on (i) improving knowledge graph representation with the aim of tackling the link prediction task; and (ii) devising a theory on how semantics can be captured in the geometry of relation representations. Most knowledge graphs are very incomplete and manually adding new information is costly, which drives the development of methods which can automatically infer missing facts. The first contribution of this thesis is HypER, a convolutional model which simplifies and improves upon the link prediction performance of the existing convolutional state-of-the-art model ConvE and can be mathematically explained in terms of constrained tensor factorisation. The second contribution is TuckER, a relatively straightforward linear model, which, at the time of its introduction, obtained state-of-the-art link prediction performance across standard datasets. The third contribution is MuRP, first multi-relational graph representation model embedded in hyperbolic space. MuRP outperforms all existing models and its Euclidean counterpart MuRE in link prediction on hierarchical knowledge graph relations whilst requiring far fewer dimensions. Despite the development of a large number of knowledge graph representation models with gradually increasing predictive performance, relatively little is known of the latent structure they learn. We generalise recent theoretical understanding of how semantic relations of similarity, paraphrase and analogy are encoded in the geometric interactions of word embeddings to how more general relations, as found in knowledge graphs, can be encoded in their representations.