Graph neural networks (GNNs) have received much attention recently because of their excellent performance on graph-based tasks. However, existing research on GNNs focuses on designing more effective models without considering much about the quality of the input data. In this paper, we propose self-enhanced GNN (SEG), which improves the quality of the input data using the outputs of existing GNN models for better performance on semi-supervised node classification. As graph data consist of both topology and node labels, we improve input data quality from both perspectives. For topology, we observe that higher classification accuracy can be achieved when the ratio of inter-class edges (connecting nodes from different classes) is low and propose topology update to remove inter-class edges and add intra-class edges. For node labels, we propose training node augmentation, which enlarges the training set using the labels predicted by existing GNN models. SEG is a general framework that can be easily combined with existing GNN models. Experimental results validate that SEG consistently improves the performance of well-known GNN models such as GCN, GAT and SGC across different datasets.
翻译:最近,由于在基于图表的任务上表现优异,因此神经神经网络(GNNs)最近受到很大关注。然而,关于GNNs的现有研究侧重于设计更有效的模型,而没有太多考虑输入数据的质量。在本文件中,我们提议自增强GNN(SEG),用现有的GNN模型产出提高输入数据的质量,以便在半监督节点分类中更好地表现。由于图形数据由地形和节点标签组成,我们从两个角度改进了输入数据的质量。关于地形学,我们观察到,当分类间边缘(连接不同等级的节点)的比例低时,可以实现更高的分类准确性。我们提议从上更新表层角度来删除跨级边缘和添加类内边缘。关于节点标签,我们提议培训节点增强,利用现有GNNM模型预测的标签扩大培训组。SEG是一个一般框架,可以很容易与现有的GNN模式结合起来。实验结果证实,SGEG不断改进众所周知的GNNM模型的性能,如GCN、GAT和SGC等不同数据的性能。