The typical contrastive self-supervised algorithm uses a similarity measure in latent space as the supervision signal by contrasting positive and negative images directly or indirectly. Although the utility of self-supervised algorithms has improved recently, there are still bottlenecks hindering their widespread use, such as the compute needed. In this paper, we propose a module that serves as an additional objective in the self-supervised contrastive learning paradigm. We show how the inclusion of this module to regress the parameters of an affine transformation or homography, in addition to the original contrastive objective, improves both performance and learning speed. Importantly, we ensure that this module does not enforce invariance to the various components of the affine transform, as this is not always ideal. We demonstrate the effectiveness of the additional objective on two recent, popular self-supervised algorithms. We perform an extensive experimental analysis of the proposed method and show an improvement in performance for all considered datasets. Further, we find that although both the general homography and affine transformation are sufficient to improve performance and convergence, the affine transformation performs better in all cases.
翻译:典型的自我监督算法在潜在空间使用一种相似的测量方法作为监督信号,直接或间接地对正与负图像进行对比。虽然自监督算法的效用最近有所改善,但仍有一些瓶颈阻碍其广泛使用,例如需要的计算。在本文中,我们提议一个模块作为自我监督对比学习范式的一个额外目标。我们表明,除了最初的对比目标外,如何将这一模块纳入回退亲变或同系的参数,从而提高性能和学习速度。重要的是,我们确保这一模块不强制推行对亲吻变各组成部分的不适应性,因为这并不总是理想的。我们展示了最近两个受欢迎的自我监督算法的额外目标的有效性。我们对拟议方法进行了广泛的实验分析,并显示所有考虑的数据集的性能都有改进。此外,我们发现,尽管一般同系和亲吻变都足以改进性能和趋同性,但近性变在所有案例中都表现更好。