Graph convolutional networks (GCNs) have achieved great success in graph representation learning by extracting high-level features from nodes and their topology. Since GCNs generally follow a message-passing mechanism, each node aggregates information from its first-order neighbour to update its representation. As a result, the representations of nodes with edges between them should be positively correlated and thus can be considered positive samples. However, there are more non-neighbour nodes in the whole graph, which provide diverse and useful information for the representation update. Two non-adjacent nodes usually have different representations, which can be seen as negative samples. Besides the node representations, the structural information of the graph is also crucial for learning. In this paper, we used quality-diversity decomposition in determinant point processes (DPP) to obtain diverse negative samples. When defining a distribution on diverse subsets of all non-neighbouring nodes, we incorporate both graph structure information and node representations. Since the DPP sampling process requires matrix eigenvalue decomposition, we propose a new shortest-path-base method to improve computational efficiency. Finally, we incorporate the obtained negative samples into the graph convolution operation. The ideas are evaluated empirically in experiments on node classification tasks. These experiments show that the newly proposed methods not only improve the overall performance of standard representation learning but also significantly alleviate over-smoothing problems.
翻译:通过从节点及其地形学中提取高层次特征,图层图形显示网络(GCNs)在图形显示学习方面取得了巨大成功。由于GCNs通常遵循一个传递信息的机制,每个节点汇总其第一端邻国的信息以更新其代表性。因此,带有两端边缘的节点的表示方式应具有积极关联性,因此可以被视为积极的样本。但是,整个图中有更多的非相邻节点,为代表更新提供了多样和有用的信息。两个非相邻节点通常有不同的表示方式,可被视为负面样本。除了节点表示方式外,图表的结构信息对于学习也至关重要。在本文件中,我们在决定点进程(DPP)中使用了质量多样性分解方式来获取不同的负面样本。当确定所有非相邻节点的不同子分布时,我们只将图形结构信息与节点表示方式都纳入。由于DPPS抽样过程需要矩阵分解,我们建议采用新的最短路径方法来改进总体学习效率。最后,我们将这些拟议的标准化方法纳入最新分析模型。我们没有将这些分析方法纳入新的分析效率。