Graph clustering, aiming to partition nodes of a graph into various groups via an unsupervised approach, is an attractive topic in recent years. To improve the representative ability, several graph auto-encoder (GAE) models, which are based on semi-supervised graph convolution networks (GCN), have been developed and they achieve good results compared with traditional clustering methods. However, all existing methods either fail to utilize the orthogonal property of the representations generated by GAE, or separate the clustering and the learning of neural networks. We first prove that the relaxed k-means will obtain an optimal partition in the inner-products used space. Driven by theoretical analysis about relaxed k-means, we design a specific GAE-based model for graph clustering to be consistent with the theory, namely Embedding Graph Auto-Encoder (EGAE). Meanwhile, the learned representations are well explainable such that the representations can be also used for other tasks. To further induce the neural network to produce deep features that are appropriate for the specific clustering model, the relaxed k-means and GAE are learned simultaneously. Therefore, the relaxed k-means can be equivalently regarded as a decoder that attempts to learn representations that can be linearly constructed by some centroid vectors. Accordingly, EGAE consists of one encoder and dual decoders. Extensive experiments are conducted to prove the superiority of EGAE and the corresponding theoretical analyses.
翻译:近些年来,为了提高代表性能力,已经开发了几个基于半监督图形组合网络(GCN)的图形自动编码器(GAE)模型,与传统的组合方法相比,这些模型取得了良好结果。然而,所有现有方法要么没有利用GAE生成的图形的正方形属性,要么将神经网络的集群和学习分开。我们首先证明,放松的K手段将在使用的空间内产产品中实现最佳的分隔。在对放松的K手段进行理论分析的驱动下,我们设计了一个基于图形组合的基于GAE的模型,以便与理论一致,即嵌入图形自动编码器(GEGAE)模型(GEAE)相比,取得了良好的结果。与此同时,所了解的表述非常可以解释的是,这些表述也可以用于其他任务。为了进一步引导神经网络产生适合特定组合模型、放松的K means和GAEE的深度特征。因此,可以同时学习一种以宽松的KAA值为方向的图象,因此,宽松的KAAA值和O值为一种直径的图。因此,可以学习一种对E的模型的图象学的比。