Self-supervised learning of graph neural networks (GNNs) aims to learn an accurate representation of the graphs in an unsupervised manner, to obtain transferable representations of them for diverse downstream tasks. Predictive learning and contrastive learning are the two most prevalent approaches for graph self-supervised learning. However, they have their own drawbacks. While the predictive learning methods can learn the contextual relationships between neighboring nodes and edges, they cannot learn global graph-level similarities. Contrastive learning, while it can learn global graph-level similarities, its objective to maximize the similarity between two differently perturbed graphs may result in representations that cannot discriminate two similar graphs with different properties. To tackle such limitations, we propose a framework that aims to learn the exact discrepancy between the original and the perturbed graphs, coined as Discrepancy-based Self-supervised LeArning (D-SLA). Specifically, we create multiple perturbations of the given graph with varying degrees of similarity, and train the model to predict whether each graph is the original graph or the perturbed one. Moreover, we further aim to accurately capture the amount of discrepancy for each perturbed graph using the graph edit distance. We validate our D-SLA on various graph-related downstream tasks, including molecular property prediction, protein function prediction, and link prediction tasks, on which ours largely outperforms relevant baselines.
翻译:图形神经网络(GNNs)的自监督学习图形神经网络(GNNs)的目的是以不受监督的方式学习图表的准确表达方式,以获得不同下游任务的可转移的图形表达方式。预测性学习和对比性学习是图形自我监督学习的两种最普遍的方法。但是,它们也有自己的缺点。虽然预测性学习方法可以学习相邻节点和边缘之间的背景关系,但它们无法学习全球图形级的相似之处。对比性学习,虽然它可以学习全球图形级的相似性,但尽可能扩大两个不同透度的图表之间的相似性的目标可能会导致无法区分具有不同属性的两个相似的图形的表达方式。为了解决这些局限性,我们提出了一个框架,目的是了解原始图和四周图之间的确切差异。我们的目标是以不一致性基点为基础的自我监督 LeArning (D-SLA) 。具体地说,我们创造了具有不同程度相似性的给定图表的多重扰动性,并训练模型来预测每个图表是原始图表还是深处的图层图。此外,我们的目标是要精确地测量每个图表的模型的数值值,包括每张的模型的比值。