论文标题
DPAR:具有节点级差异隐私的分离图神经网络
DPAR: Decoupled Graph Neural Networks with Node-Level Differential Privacy
论文作者
论文摘要
图形神经网络(GNN)在学习图中取得了巨大的成功。也为受过训练的模型提出了隐私问题,这些模型可能会揭示图形的敏感信息,包括节点功能和结构信息。在本文中,我们旨在实现训练GNN的节点级差异隐私(DP),以保护节点及其边缘。对于GNN而言,节点DP在本质上很难,因为所有直接和多跳的邻居都通过层信息传递参与每个节点的梯度计算,并且由于节点可以具有多少直接和多跳的邻居,因此现有的DP方法将导致高隐私成本或由于较高的节点的效率而导致高隐私成本或较差的效用。我们提出了一个具有差异私人近似个性化的Pagerank(DPAR)的脱钩GNN,以培训具有增强的隐私性权衡的培训GNN。关键的想法是通过DP Pagerank算法将功能投影和消息解散,该算法了解结构信息并使用Pagerank确定的顶部 - $ K $邻居进行功能聚合。通过捕获每个节点最重要的邻居并避免通过层的消息传递,它界定了节点敏感性并与基于层的扰动方法相比,它可以改善了隐私 - 实用性的权衡。我们从理论上分析了两个过程的节点DP保证,并在经验上证明了与最新方法相比,与节点DP相同的DPAR提供了更好的DPAR实用性。
Graph Neural Networks (GNNs) have achieved great success in learning with graph-structured data. Privacy concerns have also been raised for the trained models which could expose the sensitive information of graphs including both node features and the structure information. In this paper, we aim to achieve node-level differential privacy (DP) for training GNNs so that a node and its edges are protected. Node DP is inherently difficult for GNNs because all direct and multi-hop neighbors participate in the calculation of gradients for each node via layer-wise message passing and there is no bound on how many direct and multi-hop neighbors a node can have, so existing DP methods will result in high privacy cost or poor utility due to high node sensitivity. We propose a Decoupled GNN with Differentially Private Approximate Personalized PageRank (DPAR) for training GNNs with an enhanced privacy-utility tradeoff. The key idea is to decouple the feature projection and message passing via a DP PageRank algorithm which learns the structure information and uses the top-$K$ neighbors determined by the PageRank for feature aggregation. By capturing the most important neighbors for each node and avoiding the layer-wise message passing, it bounds the node sensitivity and achieves improved privacy-utility tradeoff compared to layer-wise perturbation based methods. We theoretically analyze the node DP guarantee for the two processes combined together and empirically demonstrate better utilities of DPAR with the same level of node DP compared with state-of-the-art methods.