A data augmentation module is utilized in contrastive learning to transform the given data example into two views, which is considered essential and irreplaceable. However, the predetermined composition of multiple data augmentations brings two drawbacks. First, the artificial choice of augmentation types brings specific representational invariances to the model, which have different degrees of positive and negative effects on different downstream tasks. Treating each type of augmentation equally during training makes the model learn non-optimal representations for various downstream tasks and limits the flexibility to choose augmentation types beforehand. Second, the strong data augmentations used in classic contrastive learning methods may bring too much invariance in some cases, and fine-grained information that is essential to some downstream tasks may be lost. This paper proposes a general method to alleviate these two problems by considering where and what to contrast in a general contrastive learning framework. We first propose to learn different augmentation invariances at different depths of the model according to the importance of each data augmentation instead of learning representational invariances evenly in the backbone. We then propose to expand the contrast content with augmentation embeddings to reduce the misleading effects of strong data augmentations. Experiments based on several baseline methods demonstrate that we learn better representations for various benchmarks on classification, detection, and segmentation downstream tasks.
翻译:数据增强模块用于对比性学习,将给定的数据示例转换成两种观点,这被认为是必要和不可替代的。然而,预先确定的多个数据增强的构成可能带来两个缺点。首先,人为选择增强型类型给模式带来具体的表达差异,对不同的下游任务产生不同程度的正面和负面效应。在培训期间对每一种增强型模块一视同仁,使模型为各种下游任务学习非最佳的表达方式,并限制事先选择增强型的灵活性。第二,典型对比学习方法中使用的强数据增强型可能会在某些情况下带来太多的变异,对某些下游任务至关重要的精细精细信息可能会丢失。本文提出一种总体方法,通过考虑在总体对比性学习框架中哪些方面和哪些方面来缓解这两个问题。我们首先建议根据每个数据增强型不同深度的不同理解不同的差异,而不是在骨干中平均地学习代表型差异。我们然后提议扩大与增强型嵌式的对比内容,以降低强数据递增量的误导性效果,然后可能丢失。本文提出一个总体方法,通过考虑在总体对比性学习不同层次的演示方法,我们学习如何进行下游分析。