论文标题
邻里卷积网络:用于节点分类的图形神经网络的新范式
Neighborhood Convolutional Network: A New Paradigm of Graph Neural Networks for Node Classification
论文作者
论文摘要
GCN最近开发的GCN的脱钩图卷积网络(GCN)在每个卷积层中取消了邻域聚集和特征转换,它显示了图形表示学习的有希望的性能。现有的脱钩GCN首先使用简单的神经网络(例如MLP)来学习节点的隐藏功能,然后使用固定步骤在图表上传播学习的功能,以汇总多跳社区的信息。尽管有效,但需要整个邻接矩阵作为输入的聚合操作参与模型训练,从而导致高训练成本,从而阻碍了其在较大图表上的潜力。另一方面,由于节点属性的独立性作为输入,因此,解耦GCN中使用的神经网络非常简单,并且不能将高级技术应用于建模。为此,我们进一步从脱钩的GCN解放了聚合操作,并提出了一种新的GCN范式,称为邻里卷积网络(NCN),该范围将邻域聚合作为输入结果,然后是一个特殊的卷积神经网络,以从聚集输入中提取表达性节点表示。通过这种方式,该模型可以继承脱钩GCN的优点,同时汇总邻里信息,同时又开发出更强大的功能学习模块。合并了一种称为“面具训练”的培训策略,以进一步提高模型性能。广泛的结果证明了我们模型对各种均电图和异性图的节点分类任务的有效性。
The decoupled Graph Convolutional Network (GCN), a recent development of GCN that decouples the neighborhood aggregation and feature transformation in each convolutional layer, has shown promising performance for graph representation learning. Existing decoupled GCNs first utilize a simple neural network (e.g., MLP) to learn the hidden features of the nodes, then propagate the learned features on the graph with fixed steps to aggregate the information of multi-hop neighborhoods. Despite effectiveness, the aggregation operation, which requires the whole adjacency matrix as the input, is involved in the model training, causing high training cost that hinders its potential on larger graphs. On the other hand, due to the independence of node attributes as the input, the neural networks used in decoupled GCNs are very simple, and advanced techniques cannot be applied to the modeling. To this end, we further liberate the aggregation operation from the decoupled GCN and propose a new paradigm of GCN, termed Neighborhood Convolutional Network (NCN), that utilizes the neighborhood aggregation result as the input, followed by a special convolutional neural network tailored for extracting expressive node representations from the aggregation input. In this way, the model could inherit the merit of decoupled GCN for aggregating neighborhood information, at the same time, develop much more powerful feature learning modules. A training strategy called mask training is incorporated to further boost the model performance. Extensive results demonstrate the effectiveness of our model for the node classification task on diverse homophilic graphs and heterophilic graphs.