论文标题
MMKGR:多跳多模式知识图推理
MMKGR: Multi-hop Multi-modal Knowledge Graph Reasoning
论文作者
论文摘要
多模式知识图(MKG)不仅包括关系三重态,还包括相关的多模式辅助数据(即文本和图像),从而增强了知识的多样性。但是,自然的不完整性极大地阻碍了MKG的应用。为了解决该问题,现有研究采用基于嵌入的推理模型来融合多模式特征后推断缺失的知识。但是,由于以下问题,这些方法的推理性能受到限制:(1)多模式辅助特征的无效融合; (2)缺乏复杂的推理能力以及无法进行多跳的推理,该推理能够推断出更多缺失的知识。为了克服这些问题,我们提出了一个名为MMKGR(多模式知识图推理)的新型模型。具体而言,该模型包含以下两个组件:(1)统一的栅极注意网络,旨在通过充分的注意力相互作用和降低噪声来生成有效的多模式互补特征; (2)一种补充特征感知的增强学习方法,该方法根据组件(1)中获得的特征,通过执行多跳的推理过程来预测丢失元素。实验结果表明,MMKGR在MKG推理任务中的最新方法优于最先进的方法。
Multi-modal knowledge graphs (MKGs) include not only the relation triplets, but also related multi-modal auxiliary data (i.e., texts and images), which enhance the diversity of knowledge. However, the natural incompleteness has significantly hindered the applications of MKGs. To tackle the problem, existing studies employ the embedding-based reasoning models to infer the missing knowledge after fusing the multi-modal features. However, the reasoning performance of these methods is limited due to the following problems: (1) ineffective fusion of multi-modal auxiliary features; (2) lack of complex reasoning ability as well as inability to conduct the multi-hop reasoning which is able to infer more missing knowledge. To overcome these problems, we propose a novel model entitled MMKGR (Multi-hop Multi-modal Knowledge Graph Reasoning). Specifically, the model contains the following two components: (1) a unified gate-attention network which is designed to generate effective multi-modal complementary features through sufficient attention interaction and noise reduction; (2) a complementary feature-aware reinforcement learning method which is proposed to predict missing elements by performing the multi-hop reasoning process, based on the features obtained in component (1). The experimental results demonstrate that MMKGR outperforms the state-of-the-art approaches in the MKG reasoning task.