In the last few years, large improvements in image clustering have been driven by the recent advances in deep learning. However, due to the architectural complexity of deep neural networks, there is no mathematical theory that explains the success of deep clustering techniques. In this work we introduce Projected-Scattering Spectral Clustering (PSSC), a state-of-the-art, stable, and fast algorithm for image clustering, which is also mathematically interpretable. PSSC includes a novel method to exploit the geometric structure of the scattering transform of small images. This method is inspired by the observation that, in the scattering transform domain, the subspaces formed by the eigenvectors corresponding to the few largest eigenvalues of the data matrices of individual classes are nearly shared among different classes. Therefore, projecting out those shared subspaces reduces the intra-class variability, substantially increasing the clustering performance. We call this method Projection onto Orthogonal Complement (POC). Our experiments demonstrate that PSSC obtains the best results among all shallow clustering algorithms. Moreover, it achieves comparable clustering performance to that of recent state-of-the-art clustering techniques, while reducing the execution time by more than one order of magnitude. In the spirit of reproducible research, we publish a high quality code repository along with the paper.
翻译:近些年来,由于深层神经网络的建筑复杂,因此没有数学理论来解释深层集群技术的成功。在这项工作中,我们引入了预测的蒸发光谱群集(PSSC),这是一个最先进的、稳定的和快速的图像群集算法,也是数学上可以解释的。PSSC包含一种利用小图像散射转换的几何结构的新方法。这个方法的灵感来自在分散式变异域中,由单个类别数据基体中少数几类的最大电子值组成的亚空间几乎在不同类别中共享的观察。因此,预测这些共享的子空间会减少类内变异性,大大提高组合性能。我们称这种方法投影于Orthooponal 补充(POC)上。我们的实验表明,PSSC在所有浅质集算法中获得了最佳结果。此外,在分布式变异域域中,单个类别数据基体组成的亚空间的组合性能几乎可以被不同类别共享。因此,这些共享的子空间会减少类内部变异性,从而大大提升了组群集体的功能。我们沿着一个高层次的纸质数据库执行。