Recently, contrastive learning (CL) has emerged as a successful method for unsupervised graph representation learning. Most graph CL methods first perform stochastic augmentation on the input graph to obtain two graph views and maximize the agreement of representations in the two views. Despite the prosperous development of graph CL methods, the design of graph augmentation schemes -- a crucial component in CL -- remains rarely explored. We argue that the data augmentation schemes should preserve intrinsic structures and attributes of graphs, which will force the model to learn representations that are insensitive to perturbation on unimportant nodes and edges. However, most existing methods adopt uniform data augmentation schemes, like uniformly dropping edges and uniformly shuffling features, leading to suboptimal performance. In this paper, we propose a novel graph contrastive representation learning method with adaptive augmentation that incorporates various priors for topological and semantic aspects of the graph. Specifically, on the topology level, we design augmentation schemes based on node centrality measures to highlight important connective structures. On the node attribute level, we corrupt node features by adding more noise to unimportant node features, to enforce the model to recognize underlying semantic information. We perform extensive experiments of node classification on a variety of real-world datasets. Experimental results demonstrate that our proposed method consistently outperforms existing state-of-the-art baselines and even surpasses some supervised counterparts, which validates the effectiveness of the proposed contrastive framework with adaptive augmentation.
翻译:最近,对比式学习(CL)已成为未经监督的图形代表学习的成功方法。大多数图表 CL 方法首先在输入图形上进行随机扩增,以获得两个图形视图,并在两种观点中最大限度地达成一致。尽管图表 CL 方法的开发十分繁荣,但图形增强计划的设计 -- -- CL 中的一个关键组成部分 -- -- 仍然很少探索。我们争辩说,数据增强计划应当保留图形的内在结构和属性,这将迫使模型学习那些在不重要的节点和边缘上对扰动不敏感的表达。然而,大多数现有方法首先在输入图形图形图形中进行随机扩增,以获得两种观点的一致增扩增计划,例如统一下降边缘和统一振动功能,从而在两种观点中取得最不完美的表现。在本文中,我们提出了一种新的图表对比性学习方法,其中纳入了图表的表层和语义方面的各种前科。具体地说,我们在不偏重中心度措施的基础上设计增强计划,以突出重要的关联性结构。在节点属性层面,我们腐蚀了节点特性特征,将更多的噪音添加不太重要的节点节点的节点框架特征,导致不优化的平整性断性偏向性断性断性偏向性偏向,从而导致不优化的绩效性表现。在本文件中,我们提出了调整性增强性地展示式的扩展式的演示式的实验性分析式的模型,从而演示式的实验性地展示式的模型式的模型,以显示式式式式式的模型,以显示的模型,以显示的模型的模型的模型的模型,从而演示式样式式式式式式式式式式式式式式式式式式样式式式式式式式式式式式式式式式式样,以辨式式式式式式式样,以辨式式样,以辨表态式式式式式式式式式式式式式样,以辨。