Contrastive learning, especially Self-Supervised Contrastive Learning (SSCL), has achieved great success in extracting powerful features from unlabeled data, enabling comparable performance to the supervised counterpart. In this work, we contribute to the theoretical understanding of SSCL and uncover its connection to the classic data visualization method, Stochastic Neighbor Embedding (SNE). In the perspective of SNE, whose goal is matching pairwise distance, SSCL can be viewed as a special case with the input space pairwise distance specified by constructed "positive" pairs from data augmentation. The established correspondence facilitates deeper theoretical understandings of learned features of SSCL, as well as methodological guidelines for practical improvement. Specifically, through the lens of SNE, not only can we re-derive the alignment and uniformity principle, but also provide novel analysis on domain-agnostic augmentations and implicit bias. To illustrate the practical advantage, we demonstrate that the modifications from SNE to $t$-SNE can also be adopted in the SSCL setting, achieving significant improvement in both in-distribution and out-of-distribution generalization.
翻译:在这项工作中,我们促进了SSCL的理论理解,并发现了它与经典数据直观化方法(SNE)的联系。从SNE的角度来看,SSCL的目标是对齐的距离,从SNE的角度,SSSL可被视为一个特殊案例,输入空间对齐的距离由数据扩增中构建的“正”对对称指定方指定。既定的通信有助于加深对SSCL所学特征的理论理解,以及实际改进的方法指南。具体地说,通过SNE的视角,我们不仅能够重新塑造一致性和统一性原则,而且还能够提供关于域-敏感增强和隐含的偏差的新分析。为了说明实际好处,我们证明SNE到$tSNE的修改也可以在SCL设置中采用,在分配和分配范围外的一般化方面都取得了显著改进。