Graph Neural Networks (GNNs) are a popular technique for modelling graph-structured data that compute node-level representations via aggregation of information from the local neighborhood of each node. However, this aggregation implies increased risk of revealing sensitive information, as a node can participate in the inference for multiple nodes. This implies that standard privacy preserving machine learning techniques, such as differentially private stochastic gradient descent (DP-SGD) - which are designed for situations where each data point participates in the inference for one point only - either do not apply, or lead to inaccurate solutions. In this work, we formally define the problem of learning 1-layer GNNs with node-level privacy, and provide an algorithmic solution with a strong differential privacy guarantee. Even though each node can be involved in the inference for multiple nodes, by employing a careful sensitivity analysis anda non-trivial extension of the privacy-by-amplification technique, our method is able to provide accurate solutions with solid privacy parameters. Empirical evaluation on standard benchmarks demonstrates that our method is indeed able to learn accurate privacy preserving GNNs, while still outperforming standard non-private methods that completely ignore graph information.
翻译:神经网图(GNNs)是建模图形结构数据的一种流行技术,它通过汇总每个节点的当地周边信息,计算出节点层次的表示;然而,这一汇总意味着披露敏感信息的风险增加,因为节点可以参与多个节点的推论。这意味着标准的隐私保护机器学习技术,例如有差异的私人随机梯度梯度下降法(DP-SGD),是针对每个数据点只参与一个点推断的情况设计的,不是不适用,就是导致不准确的解决办法。在这项工作中,我们正式界定了以节点隐私权学习一层GNNs的问题,并且提供了一种有强烈差异隐私保障的算法解决方案。即使每个节点都可以参与多节点的推断,例如,通过仔细的敏感性分析以及隐私逐级铺设技术的非三角延伸,我们的方法能够提供精确的解决方案,有坚实的隐私参数。对标准基准的评估表明,我们的方法确实能够学习准确的隐私保护GNNPS的隐私方法,同时实施完全的不视像标准的图。