Graph neural networks (GNNs) are designed for semi-supervised node classification on graphs where only a subset of nodes have class labels. However, under extreme cases when very few labels are available (e.g., 1 labeled node per class), GNNs suffer from severe performance degradation. Specifically, we observe that existing GNNs suffer from unstable training process on few-labeled graphs, resulting to inferior performance on node classification. Therefore, we propose an effective framework, Stabilized Self-Training (SST), which is applicable to existing GNNs to handle the scarcity of labeled data, and consequently, boost classification accuracy. We conduct thorough empirical and theoretical analysis to support our findings and motivate the algorithmic designs in SST. We apply SST to two popular GNN models GCN and DAGNN, to get SSTGCN and SSTDA methods respectively, and evaluate the two methods against 10 competitors over 5 benchmarking datasets. Extensive experiments show that the proposed SST framework is highly effective, especially when few labeled data are available. Our methods achieve superior performance under almost all settings over all datasets. For instance, on a Cora dataset with only 1 labeled node per class, the accuracy of SSTGCN is 62.5%, 17.9% higher than GCN, and the accuracy of SSTDA is 66.4%, which outperforms DAGNN by 6.6%.
翻译:GNNS(GNNs)是设计用于在只有一组节点有类标签的图形上进行半监督的节点分类的。然而,在极少贴标签的极端情况下(例如每类有一个标签的节点),GNNS的性能严重退化。具体地说,我们观察到,现有的GNNS在少数标签的图表上受到不稳定的培训过程,导致节点分类的性能低劣。因此,我们提议了一个有效的框架,即稳定自我培训(SST),它适用于现有的GNNS处理标签数据稀缺的情况,从而提高分类的准确性。我们进行了彻底的经验和理论分析,以支持我们的调查结果,激励SST的算法设计。我们把SSTSST应用到两个流行的GNNN模型GCN和DNNNWN(DNN),分别获得SSTCN和SSTDA方法,并对照5个基准数据集的10个竞争者对两种方法进行评估。广泛的实验表明,拟议的SST框架非常有效,特别是当几乎没有标签的数据。我们的方法在几乎所有的SNEAD%的精确度之下,我们的方法在几乎所有的SSTADM5级中都比SDSDAD的等级为1。</s>