Understanding self-supervised learning is important but challenging. Previous theoretical works study the role of pretraining losses, and view neural networks as general black boxes. However, the recent work of Saunshi et al. argues that the model architecture -- a component largely ignored by previous works -- also has significant influences on the downstream performance of self-supervised learning. In this work, we provide the first theoretical analysis of self-supervised learning that incorporates the effect of inductive biases originating from the model class. In particular, we focus on contrastive learning -- a popular self-supervised learning method that is widely used in the vision domain. We show that when the model has limited capacity, contrastive representations would recover certain special clustering structures that are compatible with the model architecture, but ignore many other clustering structures in the data distribution. As a result, our theory can capture the more realistic setting where contrastive representations have much lower dimensionality than the number of clusters in the data distribution. We instantiate our theory on several synthetic data distributions, and provide empirical evidence to support the theory.
翻译:理解自监督学习是重要且具有挑战性的。以往的理论工作研究了预训练损失的作用,并将神经网络视为一般的黑盒。然而,Saunshi等人最近的工作认为,模型结构 - 一直被以往的工作忽略的组成部分 - 也对自监督学习的下游性能产生了显著影响。在这项工作中,我们提供了第一篇将模型类中产生的感知偏见效应纳入到自监督学习理论分析中的论文。特别地,我们专注于对比学习 - 一种在视觉领域广泛使用的自监督学习方法。我们表明,当模型的容量有限时,对比表征将恢复某些与模型结构相容的特殊聚类结构,但会忽略数据分布中的许多其他聚类结构。因此,我们的理论能够捕捉到更为现实的情境,即相对于数据分布中聚类的数量,对比表征具有更低的维度。我们在多个人造数据分布上实例化我们的理论,并提供了支持理论的实证证据。