In recent visual self-supervision works, an imitated classification objective, called pretext task, is established by assigning labels to transformed or augmented input images. The goal of pretext can be predicting what transformations are applied to the image. However, it is observed that image transformations already present in the dataset might be less effective in learning such self-supervised representations. Building on this observation, we propose a framework based on generative adversarial network to automatically find the transformations which are not present in the input dataset and thus effective for the self-supervised learning. This automated policy allows to estimate the transformation distribution of a dataset and also construct its complementary distribution from which training pairs are sampled for the pretext task. We evaluated our framework using several visual recognition datasets to show the efficacy of our automated transformation policy.
翻译:在最近的视觉自我监督工程中,一个仿制的分类目标(称为托辞任务)是通过给转换或增强输入图像指定标签来建立的,其借口的目的可以是预测图像应用的变异,然而,据观察,数据集中已经存在的图像变异在学习这种自我监督的表达方式方面可能不太有效。基于这一观察,我们提出了一个基于基因化对立网络的框架,以自动找到输入数据集中不存在的变异,从而对自我监督的学习有效。这一自动政策允许估算数据集的变异分布,并构建其互补分布,以此对培训配对进行抽样,作为借口任务。我们利用几个视觉识别数据集对我们的框架进行了评估,以显示我们自动变换政策的功效。