Nearest neighbour based methods have proved to be one of the most successful self-supervised learning (SSL) approaches due to their high generalization capabilities. However, their computational efficiency decreases when more than one neighbour is used. In this paper, we propose a novel contrastive SSL approach, which we call All4One, that reduces the distance between neighbour representations using ''centroids'' created through a self-attention mechanism. We use a Centroid Contrasting objective along with single Neighbour Contrasting and Feature Contrasting objectives. Centroids help in learning contextual information from multiple neighbours whereas the neighbour contrast enables learning representations directly from the neighbours and the feature contrast allows learning representations unique to the features. This combination enables All4One to outperform popular instance discrimination approaches by more than 1% on linear classification evaluation for popular benchmark datasets and obtains state-of-the-art (SoTA) results. Finally, we show that All4One is robust towards embedding dimensionalities and augmentations, surpassing NNCLR and Barlow Twins by more than 5% on low dimensionality and weak augmentation settings. The source code would be made available soon.
翻译:最近邻方法已被证明是自监督学习中最成功的方法之一,因为它们具有很高的泛化能力。然而,当使用多个邻居时,它们的计算效率会降低。在本文中,我们提出了一种新的对比自监督学习方法,称为All4One,通过使用自注意机制创建的“质心”来减少邻居表示之间的距离。我们使用质心对比目标,同时使用单个邻居对比和特征对比目标。质心有助于从多个邻居中学习上下文信息,而邻居对比允许直接从邻居中学习表示,而特征对比则允许学习与特征唯一相关的表示。这种组合使All4One在流行的基准数据集的线性分类评估上比流行的实例鉴别方法表现更好,获得了最先进的结果(SoTA)。最后,我们表明All4One对嵌入维度和增强是鲁棒的,在低维度和弱增强设置上超过NNCLR和Barlow Twins超过5%。源代码即将发布。