Many real-world data comes in the form of graphs, such as social networks and protein structure. To fully utilize the information contained in graph data, a new family of machine learning (ML) models, namely graph neural networks (GNNs), has been introduced. Previous studies have shown that machine learning models are vulnerable to privacy attacks. However, most of the current efforts concentrate on ML models trained on data from the Euclidean space, like images and texts. On the other hand, privacy risks stemming from GNNs remain largely unstudied. In this paper, we fill the gap by performing the first comprehensive analysis of node-level membership inference attacks against GNNs. We systematically define the threat models and propose three node-level membership inference attacks based on an adversary's background knowledge. Our evaluation on three GNN structures and four benchmark datasets shows that GNNs are vulnerable to node-level membership inference even when the adversary has minimal background knowledge. Besides, we show that graph density and feature similarity have a major impact on the attack's success. We further investigate two defense mechanisms and the empirical results indicate that these defenses can reduce the attack performance but with moderate utility loss.
翻译:许多真实世界数据以图表形式出现,例如社交网络和蛋白质结构。为了充分利用图表数据中的信息,我们引入了一个新的机器学习模型(ML)新系列,即图形神经网络(GNNS),已经引入了一个新的机器学习模型(ML)模型。以前的研究显示,机器学习模型很容易受到隐私攻击。然而,目前大多数的努力都集中在根据来自厄几里底地空间的数据(如图像和文本)而培训的ML模型上。另一方面,来自GNNS的隐私风险仍然基本上没有得到研究。在本文中,我们通过对针对GNNS的无底级成员攻击进行第一次全面分析来填补这一空白。我们系统地界定了威胁模型,并根据对手的背景知识提出了三个无底级成员攻击建议。我们对三个GNNN结构和四个基准数据集的评估表明,即使对手的背景知识很少,GNNNS也容易受到无底级成员的推断。此外,我们表明,图形密度和特征相似性对袭击的成功率很大。我们进一步调查了两种防御机制,但实验性结果表明,这些防御可以减少攻击。