A common practice in unsupervised representation learning is to use labeled data to evaluate the quality of the learned representations. This supervised evaluation is then used to guide critical aspects of the training process such as selecting the data augmentation policy. However, guiding an unsupervised training process through supervised evaluations is not possible for real-world data that does not actually contain labels (which may be the case, for example, in privacy sensitive fields such as medical imaging). Therefore, in this work we show that evaluating the learned representations with a self-supervised image rotation task is highly correlated with a standard set of supervised evaluations (rank correlation $> 0.94$). We establish this correlation across hundreds of augmentation policies, training settings, and network architectures and provide an algorithm (SelfAugment) to automatically and efficiently select augmentation policies without using supervised evaluations. Despite not using any labeled data, the learned augmentation policies perform comparably with augmentation policies that were determined using exhaustive supervised evaluations.
翻译:未经监督的代表学习的常见做法是使用标签数据来评价所了解的表述质量。随后,监督评价被用来指导培训过程的关键方面,例如选择数据增强政策。然而,通过监督评价指导未经监督的培训过程,对于实际上不包含标签的真实世界数据来说是不可能的(例如,在诸如医学成像等隐私敏感领域可能存在这种情况 ) 。因此,在这项工作中,我们表明,用自我监督的图像旋转任务来评价所了解的表述与一套标准的监督评价(级别为0.94美元)高度相关。 我们建立了数百项强化政策、培训设置和网络结构之间的这种相关关系,并提供一种算法(自我增强),以便在不使用监督评价的情况下自动和有效地选择增强政策。尽管没有使用任何标签数据,但所了解的增强政策与使用详尽的监督评价确定的增强政策是相称的。