Self-supervised learning of graph neural networks (GNNs) aims to learn an accurate representation of the graphs in an unsupervised manner, to obtain transferable representations of them for diverse downstream tasks. Predictive learning and contrastive learning are the two most prevalent approaches for graph self-supervised learning. However, they have their own drawbacks. While the predictive learning methods can learn the contextual relationships between neighboring nodes and edges, they cannot learn global graph-level similarities. Contrastive learning, while it can learn global graph-level similarities, its objective to maximize the similarity between two differently perturbed graphs may result in representations that cannot discriminate two similar graphs with different properties. To tackle such limitations, we propose a framework that aims to learn the exact discrepancy between the original and the perturbed graphs, coined as Discrepancy-based Self-supervised LeArning (D-SLA). Specifically, we create multiple perturbations of the given graph with varying degrees of similarity and train the model to predict whether each graph is the original graph or a perturbed one. Moreover, we further aim to accurately capture the amount of discrepancy for each perturbed graph using the graph edit distance. We validate our method on various graph-related downstream tasks, including molecular property prediction, protein function prediction, and link prediction tasks, on which our model largely outperforms relevant baselines.
翻译:自监督地学习图形神经网络(GNNs)的目的是以不受监督的方式学习图表的准确表达方式,以便获得图表的准确表达方式,从而在不同的下游任务中获得可转移的图形表达方式。预测性学习和对比性学习是图形自我监督学习的两种最普遍的方法。然而,它们也有自己的缺点。虽然预测性学习方法可以了解相邻节点和边缘之间的背景关系,但它们无法学习全球图形级的相似之处。对比性学习,虽然它可以学习全球图形级的相似性,但其尽可能扩大两个不同透透度的图表之间的相似性的目标,可能会导致无法区分具有不同属性的两个相似的图表的表达方式。为了解决这些局限性,我们提出了一个框架,目的是了解原始图和环绕式图之间的确切差异,这是基于不一致性的自我监督的LeArning(D-SLA)。具体地说,我们创建了不同程度相似的给定图的多模型,并培训模型,以预测每个图表是原始图表还是相关图表,我们用每个图表的深度预测值,我们用不同的图表来精确地测量。