Self-supervised learning deals with problems that have little or no available labeled data. Recent work has shown impressive results when underlying classes have significant semantic differences. One important dataset in which this technique thrives is ImageNet, as intra-class distances are substantially lower than inter-class distances. However, this is not the case for several critical tasks, and general self-supervised learning methods fail to learn discriminative features when classes have closer semantics, thus requiring more robust strategies. We propose a strategy to tackle this problem, and to enable learning from unlabeled data even when samples from different classes are not prominently diverse. We approach the problem by leveraging a novel ensemble-based clustering strategy where clusters derived from different configurations are combined to generate a better grouping for the data samples in a fully-unsupervised way. This strategy allows clusters with different densities and higher variability to emerge, which in turn reduces intra-class discrepancies, without requiring the burden of finding an optimal configuration per dataset. We also consider different Convolutional Neural Networks to compute distances between samples. We refine these distances by performing context analysis and group them to capture complementary information. We consider two applications to validate our pipeline: Person Re-Identification and Text Authorship Verification. These are challenging applications considering that classes are semantically close to each other and that training and test sets have disjoint identities. Our method is robust across different modalities and outperforms state-of-the-art results with a fully-unsupervised solution without any labeling or human intervention.
翻译:自我监督的学习涉及到很少或根本没有标签数据的问题。 最近的工作显示, 当基础类存在显著的语义差异时, 基础类存在显著的语义差异时, 令人印象深刻的结果令人印象深刻。 这个技术兴盛的一个重要数据集是图像网络, 因为类内距离大大低于阶级之间的距离。 但是, 对于一些关键任务来说, 情况并非如此, 普通的自我监督的学习方法在课堂有更近的语义时无法学习歧视性特征, 从而需要更强有力的战略。 我们提出了一个解决这一问题的战略, 并且能够从未标记的数据中学习, 即使不同类的样本并不明显不同。 我们通过利用新型的基于共同语言的组群战略来解决这个问题, 由不同配置的组群组合在一起, 从而以完全不受监督的方式为数据样本生成更好的分组。 这种战略允许具有不同密度和更高变异性的群集, 从而减少类内部差异, 而不需要找到每个数据集的最佳配置。 我们还考虑不同的革命神经网络来测量样本之间的距离。 我们通过进行背景分析来改进这些距离, 并且通过对每类组合进行更精确的应用程序进行更精确的测试 。 我们考虑每类的校正测试 。