The pretasks are mainly built on mutual information estimation, which requires data augmentation to construct positive samples with similar semantics to learn invariant signals and negative samples with dissimilar semantics in order to empower representation discriminability. However, an appropriate data augmentation configuration depends heavily on lots of empirical trials such as choosing the compositions of data augmentation techniques and the corresponding hyperparameter settings. We propose an augmentation-free graph contrastive learning method, invariant-discriminative graph contrastive learning (iGCL), that does not intrinsically require negative samples. iGCL designs the invariant-discriminative loss (ID loss) to learn invariant and discriminative representations. On the one hand, ID loss learns invariant signals by directly minimizing the mean square error between the target samples and positive samples in the representation space. On the other hand, ID loss ensures that the representations are discriminative by an orthonormal constraint forcing the different dimensions of representations to be independent of each other. This prevents representations from collapsing to a point or subspace. Our theoretical analysis explains the effectiveness of ID loss from the perspectives of the redundancy reduction criterion, canonical correlation analysis, and information bottleneck principle. The experimental results demonstrate that iGCL outperforms all baselines on 5 node classification benchmark datasets. iGCL also shows superior performance for different label ratios and is capable of resisting graph attacks, which indicates that iGCL has excellent generalization and robustness. The source code is available at https://github.com/lehaifeng/T-GCN/tree/master/iGCL.
翻译:预设任务主要是建立在相互信息估计的基础上,这就要求数据增强,以构建具有类似语义学的正面样本,以学习变异信号和不同语义学的负面样本,从而增强代表性差异;然而,适当的数据增强配置在很大程度上取决于许多经验性试验,例如选择数据增强技术的组成和相应的超参数设置;我们建议采用无增量图形对比学习方法,即不易变差异图形对比学习(iGCL),这在本质上并不要求负面样本。iGCL设计变异-差异性损失(ID损失),以学习异变性和歧视性表达方式。一方面,通过直接尽量减少目标样本与代表空间中正样样本之间的平均平方差来学习变异信号。另一方面,ID损失确保了无增量图形对比性,迫使不同层面的表达方式相互独立。这阻碍了从崩溃到点或子空间的表达。我们的理论分析解释,从稳健的iRC原则的角度解释ID损失的有效性,即降缩缩缩放的IGC标准, 也显示所有易变缩缩缩标签/对比数据。</s>