Generative graph self-supervised learning (SSL) aims to learn node representations by reconstructing the input graph data. However, most existing methods focus on unsupervised learning tasks only and very few work has shown its superiority over the state-of-the-art graph contrastive learning (GCL) models, especially on the classification task. While a very recent model has been proposed to bridge the gap, its performance on unsupervised learning tasks is still unknown. In this paper, to comprehensively enhance the performance of generative graph SSL against other GCL models on both unsupervised and supervised learning tasks, we propose the SeeGera model, which is based on the family of self-supervised variational graph auto-encoder (VGAE). Specifically, SeeGera adopts the semi-implicit variational inference framework, a hierarchical variational framework, and mainly focuses on feature reconstruction and structure/feature masking. On the one hand, SeeGera co-embeds both nodes and features in the encoder and reconstructs both links and features in the decoder. Since feature embeddings contain rich semantic information on features, they can be combined with node embeddings to provide fine-grained knowledge for feature reconstruction. On the other hand, SeeGera adds an additional layer for structure/feature masking to the hierarchical variational framework, which boosts the model generalizability. We conduct extensive experiments comparing SeeGera with 9 other state-of-the-art competitors. Our results show that SeeGera can compare favorably against other state-of-the-art GCL methods in a variety of unsupervised and supervised learning tasks.
翻译:生成图形自监督学习(SSL)的目的是通过重建输入图形数据来学习节点显示。然而,大多数现有方法仅侧重于不受监督的学习任务,而且很少有工作显示它优于最先进的图形对比学习模型,特别是在分类任务方面。虽然提出了最近的模型来缩小差距,但其在未经监督的学习任务上的性能仍然未知。在本文中,为了全面提高基因图形 SSL 的性能,以其他不受监督和监督的GCL 高级学习任务,我们提出了SeeGera 模型,该模型基于自监督的变异图自动编码(VGAE)的组合。具体地说,SeeGera采用了半隐含的变异框架,一个等级变异框架,主要侧重于特征重建以及结构/地形遮掩。一方面,Seeera 双向隐藏的节点和特性,在未监督和监督的学习任务中重建链接和特性。由于将自我监督的变异性结构的特性与精细的G,因此,在重建过程中可以提供其他变异的精度,在常规结构中可以提供其他变异性信息。