Self-supervised node representation learning aims to learn node representations from unlabelled graphs that rival the supervised counterparts. The key towards learning informative node representations lies in how to effectively gain contextual information from the graph structure. In this work, we present simple-yet-effective self-supervised node representation learning via aligning the hidden representations of nodes and their neighbourhood. Our first idea achieves such node-to-neighbourhood alignment by directly maximizing the mutual information between their representations, which, we prove theoretically, plays the role of graph smoothing. Our framework is optimized via a surrogate contrastive loss and a Topology-Aware Positive Sampling (TAPS) strategy is proposed to sample positives by considering the structural dependencies between nodes, which enables offline positive selection. Considering the excessive memory overheads of contrastive learning, we further propose a negative-free solution, where the main contribution is a Graph Signal Decorrelation (GSD) constraint to avoid representation collapse and over-smoothing. The GSD constraint unifies some of the existing constraints and can be used to derive new implementations to combat representation collapse. By applying our methods on top of simple MLP-based node representation encoders, we learn node representations that achieve promising node classification performance on a set of graph-structured datasets from small- to large-scale.
翻译:自我监督的节点代表学习旨在学习来自与受监督的对应方相对应的未贴标签图表的节点表达方式。学习信息节点表达方式的关键在于如何从图形结构中有效获取背景信息。在这项工作中,我们通过调整节点及其相邻的隐藏表达方式,展示了简单而有效的自监督节点代表方式学习方式。我们的第一个想法是直接最大限度地利用它们之间的相互信息,从而实现这种节点对邻的匹配。从理论上看,它们的作用是平滑的图解。我们的框架是通过一种替代式对比性损失和地形-软件积极抽样(TAPS)战略优化的。我们建议通过考虑节点之间的结构依赖性来抽样积极的,从而使得能够进行非在线的积极选择。考虑到对比性学习的过度记忆间接,我们进一步提出一种无负点解决方案,其中的主要贡献是图形信号性礼节点关系(GSD)制约,以避免代表性的崩溃和过度移动。 GSD的制约使得现有的一些表层对比性约束无法将现有的一些现有限制和表层积极抽样抽样(TAP) 用于从我们的表层代表方式进行新的表现。