In recent years, contrastive learning achieves impressive results on self-supervised visual representation learning, but there still lacks a rigorous understanding of its learning dynamics. In this paper, we show that if we cast a contrastive objective equivalently into the feature space, then its learning dynamics admits an interpretable form. Specifically, we show that its gradient descent corresponds to a specific message passing scheme on the corresponding augmentation graph. Based on this perspective, we theoretically characterize how contrastive learning gradually learns discriminative features with the alignment update and the uniformity update. Meanwhile, this perspective also establishes an intriguing connection between contrastive learning and Message Passing Graph Neural Networks (MP-GNNs). This connection not only provides a unified understanding of many techniques independently developed in each community, but also enables us to borrow techniques from MP-GNNs to design new contrastive learning variants, such as graph attention, graph rewiring, jumpy knowledge techniques, etc. We believe that our message passing perspective not only provides a new theoretical understanding of contrastive learning dynamics, but also bridges the two seemingly independent areas together, which could inspire more interleaving studies to benefit from each other. The code is available at https://github.com/PKU-ML/Message-Passing-Contrastive-Learning.
翻译:近年来,对比式学习在自我监督的视觉演示学习中取得了令人印象深刻的成果,但是仍然缺乏对其学习动态的严格理解。在本文中,我们显示,如果我们在特征空间中设定一个对比性目标,那么其学习动态就承认了一种可解释的形式。具体地说,我们显示,其梯度下降与相应的增强图上的具体信息传递计划相对应。基于这一视角,我们从理论上描述对比式学习如何逐渐学习与校准更新和统一更新的区别性特征。同时,这一视角还建立了对比式学习与信息传递图像神经网络(MP-GNNS)之间的令人感兴趣的联系。这一联系不仅提供了对每个社区独立开发的许多技术的统一理解,而且还使我们能够从MP-GNNS中借用技术来设计新的对比性学习变体,例如图形关注、图表重新布线、突变知识技术等等。我们认为,我们的信息传递视角不仅提供了对对比式学习动态的新理论理解,而且还将两个看起来独立的领域连接起来,这可以激励更多的内部学习研究/ML学习。</s>