Deep models trained in supervised mode have achieved remarkable success on a variety of tasks. When labeled samples are limited, self-supervised learning (SSL) is emerging as a new paradigm for making use of large amounts of unlabeled samples. SSL has achieved promising performance on natural language and image learning tasks. Recently, there is a trend to extend such success to graph data using graph neural networks (GNNs). In this survey, we provide a unified review of different ways of training GNNs using SSL. Specifically, we categorize SSL methods into contrastive and predictive models. In either category, we provide a unified framework for methods as well as how these methods differ in each component under the framework. Our unified treatment of SSL methods for GNNs sheds light on the similarities and differences of various methods, setting the stage for developing new methods and algorithms. We also summarize different SSL settings and the corresponding datasets used in each setting. To facilitate methodological development and empirical comparison, we develop a standardized testbed for SSL in GNNs, including implementations of common baseline methods, datasets, and evaluation metrics.
翻译:以监督模式培训的深层模型在各种任务方面取得了显著的成功。当标签样本有限时,自我监督学习(SSL)正在成为使用大量无标签样本的新范例。SSL在自然语言和图像学习任务方面取得了有希望的成绩。最近,趋势是将这种成功扩大到图表数据,使用图形神经网络(GNNs ) 。在这次调查中,我们统一审查使用SSL培训GNs的不同方法。具体地说,我们将SSL方法分为对比和预测模型。在其中任一类别中,我们为方法提供了统一框架,以及这些方法在框架下每个组成部分中的差异。我们对SLSL方法的统一处理揭示了各种方法的相似性和差异,为开发新方法和算法创造了舞台。我们还总结了不同的SLS设置和在每种环境中使用的相应数据集。为了便利方法的开发和经验比较,我们为SLSL制定了一个标准化的测试台,包括实施共同基准方法、数据集和评价基准。