Self-supervised learning of graph neural networks (GNN) is in great need because of the widespread label scarcity issue in real-world graph/network data. Graph contrastive learning (GCL), by training GNNs to maximize the correspondence between the representations of the same graph in its different augmented forms, may yield robust and transferable GNNs even without using labels. However, GNNs trained by traditional GCL often risk capturing redundant graph features and thus may be brittle and provide sub-par performance in downstream tasks. Here, we propose a novel principle, termed adversarial-GCL (AD-GCL), which enables GNNs to avoid capturing redundant information during the training by optimizing adversarial graph augmentation strategies used in GCL. We pair AD-GCL with theoretical explanations and design a practical instantiation based on trainable edge-dropping graph augmentation. We experimentally validate AD-GCL by comparing with the state-of-the-art GCL methods and achieve performance gains of up-to $14\%$ in unsupervised, $6\%$ in transfer, and $3\%$ in semi-supervised learning settings overall with 18 different benchmark datasets for the tasks of molecule property regression and classification, and social network classification.
翻译:由于现实世界图表/网络数据中普遍存在标签紧缺问题,极需自我监督地学习图形神经网络(GNN),因为实际世界图形/网络数据中标签稀缺问题普遍存在,因此极需自行监督地学。图表对比性学习(GCL),通过培训GNNS,使以不同扩大形式显示的同一图形的表达形式之间最大限度地相互对应,即使不使用标签,也可能产生稳健和可转让的GNNS。然而,传统GCL培训的GNNS往往有捕捉多余的图形特征的风险,因此可能是易碎的,在下游任务中提供分级性能。在这里,我们提出了一个新的原则,称为对抗性-GCL(AD-GCL),使GNS在培训期间通过优化GCL使用的对抗性图形增强战略避免捕捉多余的信息。我们用理论解释对AD-GCL进行配,并在可训练的边缘倾斜图增强的基础上设计一个实用的即时计。我们实验性地验证AD-GCL,在未超超的、转移6美元和半超固化的学习模式中实现业绩增益,在18个基本数据分类中,用18个基准数据库和基本数据设置上,我们试验验证AD-GCL。