Graph Neural Networks (GNNs) have emerged as a prominent graph learning model in various graph-based tasks over the years. Nevertheless, due to the vulnerabilities of GNNs, it has been empirically shown that malicious attackers could easily corrupt the fairness level of their predictions by adding perturbations to the input graph data. In this paper, we take crucial steps to study a novel problem of certifiable defense on the fairness level of GNNs. Specifically, we propose a principled framework named ELEGANT and present a detailed theoretical certification analysis for the fairness of GNNs. ELEGANT takes {\em any} GNN as its backbone, and the fairness level of such a backbone is theoretically impossible to be corrupted under certain perturbation budgets for attackers. Notably, ELEGANT does not make any assumptions over the GNN structure or parameters, and does not require re-training the GNNs to realize certification. Hence it can serve as a plug-and-play framework for any optimized GNNs ready to be deployed. We verify the satisfactory effectiveness of ELEGANT in practice through extensive experiments on real-world datasets across different backbones of GNNs and parameter settings.
翻译:近年来,图神经网络(GNNs)已成为各类图学习任务中的重要模型。然而,由于GNNs的脆弱性,已有实证研究表明恶意攻击者能够通过对输入图数据添加扰动,轻易破坏其预测的公平性水平。本文针对GNNs公平性的可验证防御这一新问题展开关键性研究。具体而言,我们提出了一个名为ELEGANT的原则性框架,并对GNNs的公平性进行了详细的理论可验证性分析。ELEGANT以{\em任意}GNN作为主干网络,在攻击者扰动预算受限的条件下,该主干网络的公平性水平在理论上不可能被破坏。值得注意的是,ELEGANT不对GNN的结构或参数作任何假设,也无需重新训练GNN即可实现可验证性。因此该框架可作为即插即用方案,适用于任何已优化并准备部署的GNN。我们通过在现实数据集上对不同GNN主干网络及参数设置进行大量实验,验证了ELEGANT在实际应用中的显著有效性。