论文标题

通过全球上下文预测,自我监督的图表表示学习

Self-Supervised Graph Representation Learning via Global Context Prediction

论文作者

Peng, Zhen, Dong, Yixiang, Luo, Minnan, Wu, Xiao-Ming, Zheng, Qinghua

论文摘要

为了充分利用快速增长的未标记网络数据,本文通过利用数据本身提供的自然监督来介绍一种新颖的自我监督策略,以进行图形表示学习。受人类社会行为的启发,我们假设每个节点的全局上下文都由图中的所有节点组成,因为连接网络中的两个任意实体可以通过不同长度的路径相互交互。基于此,我们调查了全球环境是否可以成为学习有用节点表示的免费有效监督信号的来源。具体而言,我们在图中随机选择一对节点并训练一个精心设计的神经网,以预测一个节点相对于另一个节点的上下文位置。我们的基本假设是,从这种内部环境中学到的表示形式将捕获图的全局拓扑,并精心地表征节点之间的相似性和差异化,这有利于各种下游学习任务。包括节点分类,聚类和链接预测在内的广泛基准实验表明,我们的方法的表现优于许多最新的无监督方法,有时甚至超过了受监督的对应物的性能。

To take full advantage of fast-growing unlabeled networked data, this paper introduces a novel self-supervised strategy for graph representation learning by exploiting natural supervision provided by the data itself. Inspired by human social behavior, we assume that the global context of each node is composed of all nodes in the graph since two arbitrary entities in a connected network could interact with each other via paths of varying length. Based on this, we investigate whether the global context can be a source of free and effective supervisory signals for learning useful node representations. Specifically, we randomly select pairs of nodes in a graph and train a well-designed neural net to predict the contextual position of one node relative to the other. Our underlying hypothesis is that the representations learned from such within-graph context would capture the global topology of the graph and finely characterize the similarity and differentiation between nodes, which is conducive to various downstream learning tasks. Extensive benchmark experiments including node classification, clustering, and link prediction demonstrate that our approach outperforms many state-of-the-art unsupervised methods and sometimes even exceeds the performance of supervised counterparts.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源