The fundamental goal of self-supervised learning (SSL) is to produce useful representations of data without access to any labels for classifying the data. Modern methods in SSL, which form representations based on known or constructed relationships between samples, have been particularly effective at this task. Here, we aim to extend this framework to incorporate algorithms based on kernel methods where embeddings are constructed by linear maps acting on the feature space of a kernel. In this kernel regime, we derive methods to find the optimal form of the output representations for contrastive and non-contrastive loss functions. This procedure produces a new representation space with an inner product denoted as the induced kernel which generally correlates points which are related by an augmentation in kernel space and de-correlates points otherwise. We analyze our kernel model on small datasets to identify common features of self-supervised learning algorithms and gain theoretical insights into their performance on downstream tasks.
翻译:自我监督学习(SSL)的根本目标是提供有用的数据表达方式,而不能获得任何数据分类标签。 SSL的现代方法基于已知或建构的样本之间的关系,在这项任务中特别有效。 在这里,我们的目标是扩大这一框架,以纳入基于内核方法的算法,其中嵌入以直线地图根据内核的特征空间构造的嵌入图。在这个内核系统中,我们想出一些方法来找到对比性和非对调性损失功能的产出表达方式的最佳形式。这个程序产生了一个新的代表空间,内部产品被称作导导出内核,通常与内核空间的增强和除子点相关联。我们分析了关于小数据集的内核模型,以确定自我监督学习算法的共同特征,并从理论上了解下游任务的业绩。