Contrastive learning (CL) has achieved remarkable success in learning data representations without label supervision. However, the conventional CL loss is sensitive to how many negative samples are included and how they are selected. This paper proposes contrastive conditional transport (CCT) that defines its CL loss over dependent sample-query pairs, which in practice is realized by drawing a random query, randomly selecting positive and negative samples, and contrastively reweighting these samples according to their distances to the query, exerting a greater force to both pull more distant positive samples towards the query and push closer negative samples away from the query. Theoretical analysis shows that this unique contrastive reweighting scheme helps in the representation space to both align the positive samples with the query and reduce the mutual information between the negative sample and query. Extensive large-scale experiments on standard vision tasks show that CCT not only consistently outperforms existing methods on benchmark datasets in contrastive representation learning but also provides interpretable contrastive weights and latent representations. PyTorch code will be provided.
翻译:对比性学习(CL)在不进行标签监督的情况下,在学习数据表述方面取得了显著的成功;然而,传统的CL损失对于包括多少否定样本以及如何选择这些样本十分敏感。本文件建议采用对比性有条件运输(CCT),界定其CL相对于依赖性样本采集对配对的损失。 在实践中,通过随机抽查、随机选择正样和负样,以及根据这些样本与查询的距离对样本进行对比性重新加权,运用更大的力量将更远的正面样本拉向查询,并将更近的负面样本推离查询。理论分析表明,这种独特的对比性再加权办法有助于代表空间将正样与查询相匹配,并减少负样与查询之间的相互信息。关于标准愿景任务的广泛大规模实验表明,CCT不仅一贯地超越了对比性代表性学习中的基准数据集的现有方法,而且还提供了可解释的对比性加权值和潜伏表。将提供PyTorrch 代码。