Graph representation learning has attracted a surge of interest recently, whose target at learning discriminant embedding for each node in the graph. Most of these representation methods focus on supervised learning and heavily depend on label information. However, annotating graphs are expensive to obtain in the real world, especially in specialized domains (i.e. biology), as it needs the annotator to have the domain knowledge to label the graph. To approach this problem, self-supervised learning provides a feasible solution for graph representation learning. In this paper, we propose a Multi-Level Graph Contrastive Learning (MLGCL) framework for learning robust representation of graph data by contrasting space views of graphs. Specifically, we introduce a novel contrastive view - topological and feature space views. The original graph is first-order approximation structure and contains uncertainty or error, while the $k$NN graph generated by encoding features preserves high-order proximity. Thus $k$NN graph generated by encoding features not only provide a complementary view, but is more suitable to GNN encoder to extract discriminant representation. Furthermore, we develop a multi-level contrastive mode to preserve the local similarity and semantic similarity of graph-structured data simultaneously. Extensive experiments indicate MLGCL achieves promising results compared with the existing state-of-the-art graph representation learning methods on seven datasets.
翻译:最近,图示的学习引起了人们的极大兴趣,其目标在于为图表中的每个节点嵌入差异化学习。这些图示方法大多侧重于有监督的学习,并在很大程度上依赖于标签信息。然而,图表说明在现实世界中非常昂贵,特别是在专门领域(即生物学),因为它需要说明者拥有域知识来标出图。为了解决这一问题,自监督的学习为图形代表学习提供了一个可行的解决方案。在本文中,我们提议了一个多层次图表对比学习框架,通过对比图形的空间视图来学习图表数据的稳健代表性。具体地说,我们采用了一种新的对比性观点――表面和特征空间观点。原始图表是一阶近似结构,含有不确定性或错误,而编码特征生成的$k$NNN的图形保存了高度相近的接近性。因此,通过编码特征生成的 $k$NNNN的图形不仅提供补充性观点,而且更适合GNN encod 来提取相近的图形代表。此外,我们开发了一种多层次对比性对比性图表模型模式,以同时保存当地相近似的图表结构。