In this paper, we propose a novel, effective and simpler end-to-end image clustering auto-encoder algorithm: ICAE. The algorithm uses PEDCC (Predefined Evenly-Distributed Class Centroids) as the clustering centers, which ensures the inter-class distance of latent features is maximal, and adds data distribution constraint, data augmentation constraint, auto-encoder reconstruction constraint and Sobel smooth constraint to improve the clustering performance. Specifically, we perform one-to-one data augmentation to learn the more effective features. The data and the augmented data are simultaneously input into the autoencoder to obtain latent features and the augmented latent features whose similarity are constrained by an augmentation loss. Then, making use of the maximum mean discrepancy distance (MMD), we combine the latent features and augmented latent features to make their distribution close to the PEDCC distribution (uniform distribution between classes, Dirac distribution within the class) to further learn clustering-oriented features. At the same time, the MSE of the original input image and reconstructed image is used as reconstruction constraint, and the Sobel smooth loss to build generalization constraint to improve the generalization ability. Finally, extensive experiments on three common datasets MNIST, Fashion-MNIST, COIL20 are conducted. The experimental results show that the algorithm has achieved the best clustering results so far. In addition, we can use the predefined PEDCC class centers, and the decoder to clearly generate the samples of each class. The code can be downloaded at https://github.com/zyWang-Power/Clustering!
翻译:在本文中,我们提出了一个新颖、有效、简单、端到端图像集自动编码算法:ICAE。算法将PEDCC(预定义的均分分布类中枢)用作集束中心,确保潜伏特征在阶级之间的距离最大,并增加数据分配限制、数据增强限制、自动编码重建限制和Sobel光滑限制,以提高组合性能。具体地说,我们执行一对一的数据增强,以学习更有效的特性。数据和增强的数据同时输入自动编码,以获得潜在特性和增加的潜值特性,而其相似性因增缩损失而受到限制。然后,利用最大平均差异距离(MMD),我们结合了潜值特性和增加潜在特性,使其接近PEDCC的分布(各类之间的统一分布、Dirac内部的分布),以便进一步学习以组合为导向的特性。与此同时,最初输入的MISG20图像和再版图像的MISE被用作重建限制,而Sobel平稳损失则建立普通中心,从而改进了通用的MDLALA。最后,我们进行了广泛的实验结果。