Learning with graphs has attracted significant attention recently. Existing representation learning methods on graphs have achieved state-of-the-art performance on various graph-related tasks such as node classification, link prediction, etc. However, we observe that these methods could leak serious private information. For instance, one can accurately infer the links (or node identity) in a graph from a node classifier (or link predictor) trained on the learnt node representations by existing methods. To address the issue, we propose a privacy-preserving representation learning framework on graphs from the \emph{mutual information} perspective. Specifically, our framework includes a primary learning task and a privacy protection task, and we consider node classification and link prediction as the two tasks of interest. Our goal is to learn node representations such that they can be used to achieve high performance for the primary learning task, while obtaining performance for the privacy protection task close to random guessing. We formally formulate our goal via mutual information objectives. However, it is intractable to compute mutual information in practice. Then, we derive tractable variational bounds for the mutual information terms, where each bound can be parameterized via a neural network. Next, we train these parameterized neural networks to approximate the true mutual information and learn privacy-preserving node representations. We finally evaluate our framework on various graph datasets.
翻译:图表上的现有代表学习方法在节点分类、链接预测等各种与图表有关的任务中取得了最先进的表现。然而,我们发现,这些方法可能会泄露严重的私人信息。例如,我们可以准确地从一个节点分类(或链接预测器)中推断出在图表中的链接(或节点身份),这些链接(或节点身份)是通过现有方法获得的节点表示方法培训的。为了解决这个问题,我们提议了一个隐私保留代表学习框架。具体地说,我们的框架包括一个主要的学习任务和隐私保护任务,我们认为,将节点分类和预测作为两项感兴趣的任务。我们的目标是从一个节点分类(或链接预测器)中准确地推断出在图表中的链接(或节点身份),以便用来实现初级学习任务的高性能,同时获得隐私保护任务的接近于随机猜测。我们通过相互信息目标正式制定我们的目标。然而,在实践中对相互信息进行校正是难以操作的。然后,我们为相互信息术语绘制可移动的变形框框,其中每个节点分类都是通过神经网络来测量我们的精确度。我们最后通过这些矩阵对各种图层图层图表进行测量。