Recently, many unsupervised deep learning methods have been proposed to learn clustering with unlabelled data. By introducing data augmentation, most of the latest methods look into deep clustering from the perspective that the original image and its tansformation should share similar semantic clustering assignment. However, the representation features before softmax activation function could be quite different even the assignment probability is very similar since softmax is only sensitive to the maximum value. This may result in high intra-class diversities in the representation feature space, which will lead to unstable local optimal and thus harm the clustering performance. By investigating the internal relationship between mutual information and contrastive learning, we summarized a general framework that can turn any maximizing mutual information into minimizing contrastive loss. We apply it to both the semantic clustering assignment and representation feature and propose a novel method named Deep Robust Clustering by Contrastive Learning (DRC). Different to existing methods, DRC aims to increase inter-class diver-sities and decrease intra-class diversities simultaneously and achieve more robust clustering results. Extensive experiments on six widely-adopted deep clustering benchmarks demonstrate the superiority of DRC in both stability and accuracy. e.g., attaining 71.6% mean accuracy on CIFAR-10, which is 7.1% higher than state-of-the-art results.
翻译:最近,提出了许多未经监督的深层学习方法,以学习与未贴标签的数据组合。通过引入数据扩增,大多数最新方法都从原始图像及其变相应共享相似的语义群群任务的角度审视深层群集。然而,软式马克思激活功能之前的表示特征可能大不相同,即使任务概率也非常相似,因为软式马克思只是对最大值的敏感度,这可能导致代表空间的高度阶级内部差异性,导致当地最优化的不稳定,从而损害分组性能。通过调查相互信息和对比性学习之间的内部关系,我们总结了一个总体框架,可以将任何最大程度的相互信息转化为尽量减少对比性损失。我们将其应用于语义组合任务和代表特征,并提出一种名为“差异性学习的深固态组合(DRC)”的新方法。与现有方法不同,刚果民主共和国的目标是增加各阶层间拆分解性差异,同时降低类内分散性,从而损害集群性。通过对六个广泛采用的深层组合基准进行广泛的实验,可以将任何相互的信息转化为尽量减少对比性损失。我们将其应用于语义组合任务组任务和代表特征特征特征特征特征特征特征特征特征特征特征的特征,并提议采用新的方法。达到75-10的准确度。实现刚果民主共和国的稳定性和准确度。