论文标题
本地私人图神经网络
Locally Private Graph Neural Networks
论文作者
论文摘要
图形神经网络(GNN)在学习节点表示方面表现出卓越的性能。但是,当节点代表涉及敏感或个人信息的人或与人相关的变量时,对图形数据进行学习可能会引起隐私问题。尽管已经提出了许多用于保护非关系数据的深入学习隐私学习的技术,但解决与在图表上应用深度学习算法有关的隐私问题的工作较少。在本文中,我们研究了节点数据隐私的问题,其中图节点具有潜在的敏感数据,该数据可保持私密,但它们可能对中央服务器有益于通过图表训练GNN。为了解决这个问题,我们开发了具有基于当地差异性隐私(LDP)的正式隐私保证的,具有正式隐私保证的隐私,建筑 - 不可能的GNN学习算法。具体来说,我们建议使用的是一份不偏编码器和一个无偏的整流器,通过该编码器,服务器可以与图形节点进行通信以私下收集其数据并近似GNN的第一层。为了进一步降低注射噪声的效果,我们建议预先一个简单的图形卷积层,称为KPROP,该层基于节点特征的多跳聚集,该特征充当了脱氧机制。最后,我们提出了一个强大的训练框架,其中我们受益于KPROP具有噪音标签的推断准确性的能力。通过现实世界数据集进行的广泛实验表明,我们的方法可以保持令人满意的准确性,而隐私损失较低。
Graph Neural Networks (GNNs) have demonstrated superior performance in learning node representations for various graph inference tasks. However, learning over graph data can raise privacy concerns when nodes represent people or human-related variables that involve sensitive or personal information. While numerous techniques have been proposed for privacy-preserving deep learning over non-relational data, there is less work addressing the privacy issues pertained to applying deep learning algorithms on graphs. In this paper, we study the problem of node data privacy, where graph nodes have potentially sensitive data that is kept private, but they could be beneficial for a central server for training a GNN over the graph. To address this problem, we develop a privacy-preserving, architecture-agnostic GNN learning algorithm with formal privacy guarantees based on Local Differential Privacy (LDP). Specifically, we propose an LDP encoder and an unbiased rectifier, by which the server can communicate with the graph nodes to privately collect their data and approximate the GNN's first layer. To further reduce the effect of the injected noise, we propose to prepend a simple graph convolution layer, called KProp, which is based on the multi-hop aggregation of the nodes' features acting as a denoising mechanism. Finally, we propose a robust training framework, in which we benefit from KProp's denoising capability to increase the accuracy of inference in the presence of noisy labels. Extensive experiments conducted over real-world datasets demonstrate that our method can maintain a satisfying level of accuracy with low privacy loss.