In contrastive representation learning, data representation is trained so that it can classify the image instances even when the images are altered by augmentations. However, depending on the datasets, some augmentations can damage the information of the images beyond recognition, and such augmentations can result in collapsed representations. We present a partial solution to this problem by formalizing a stochastic encoding process in which there exist a tug-of-war between the data corruption introduced by the augmentations and the information preserved by the encoder. We show that, with the infoMax objective based on this framework, we can learn a data-dependent distribution of augmentations to avoid the collapse of the representation.
翻译:在反向代表性学习中,对数据代表进行了培训,以便它能够对图像实例进行分类,即使图像被放大改变。然而,根据数据集,某些增强可能会损害图像信息,使其无法识别,而这种增强可能导致表达方式崩溃。我们通过正式确定一个随机编码过程来部分解决这一问题,在这个过程中,在扩增带来的数据腐败与编码器保存的信息之间存在拖动战。我们表明,有了基于这个框架的InfoMax目标,我们可以从数据上获得扩增分布,以避免表达方式崩溃。