Graph Neural Networks (GNNs) have enjoyed wide spread applications in graph-structured data. However, existing graph based applications commonly lack annotated data. GNNs are required to learn latent patterns from a limited amount of training data to perform inferences on a vast amount of test data. The increased complexity of GNNs, as well as a single point of model parameter initialization, usually lead to overfitting and sub-optimal performance. In addition, it is known that GNNs are vulnerable to adversarial attacks. In this paper, we push one step forward on the ensemble learning of GNNs with improved accuracy, generalization, and adversarial robustness. Following the principles of stochastic modeling, we propose a new method called GNN-Ensemble to construct an ensemble of random decision graph neural networks whose capacity can be arbitrarily expanded for improvement in performance. The essence of the method is to build multiple GNNs in randomly selected substructures in the topological space and subfeatures in the feature space, and then combine them for final decision making. These GNNs in different substructure and subfeature spaces generalize their classification in complementary ways. Consequently, their combined classification performance can be improved and overfitting on the training data can be effectively reduced. In the meantime, we show that GNN-Ensemble can significantly improve the adversarial robustness against attacks on GNNs.
翻译:图神经网络(GNNs)在图结构数据的应用中发挥着重要作用。但是,现有的基于图的应用往往缺乏注释数据。为了从有限的训练数据中学习潜在的模式,以便对大量的测试数据进行推断,需要使用GNNs。GNNs的复杂度增加以及单点模型参数初始化通常导致过拟合和次优性能。此外,众所周知,GNNs容易受到对抗性攻击的影响。在本文中,我们推进了GNNs的集成学习,以提高准确性、泛化能力和对抗鲁棒性。遵循随机建模原则,我们提出了一种称为GNN-Ensemble的新方法,用于构建多个随机决策图神经网络的集成模型。这些模型的容量可以任意扩大以改善性能。方法的实质是在拓扑空间的随机选定子结构和特征空间的子特征上构建多个GNNs,然后将它们组合用于最终决策。这些不同子结构和子特征空间中的GNNs以互补的方式概括其分类。因此,它们的组合分类性能可以得到改善,并且可以有效减少对训练数据的过拟合。同时,我们展示了GNN-Ensemble可以显著提高对GNNs攻击的对抗性鲁棒性。