Graph Neural Networks (GNNs) have emerged as the leading paradigm for solving graph analytical problems in various real-world applications. Nevertheless, GNNs could potentially render biased predictions towards certain demographic subgroups. Understanding how the bias in predictions arises is critical, as it guides the design of GNN debiasing mechanisms. However, most existing works overwhelmingly focus on GNN debiasing, but fall short on explaining how such bias is induced. In this paper, we study a novel problem of interpreting GNN unfairness through attributing it to the influence of training nodes. Specifically, we propose a novel strategy named Probabilistic Distribution Disparity (PDD) to measure the bias exhibited in GNNs, and develop an algorithm to efficiently estimate the influence of each training node on such bias. We verify the validity of PDD and the effectiveness of influence estimation through experiments on real-world datasets. Finally, we also demonstrate how the proposed framework could be used for debiasing GNNs. Open-source code can be found at https://github.com/yushundong/BIND.
翻译:神经网络图(GNNs)已成为解决各种现实应用中图表分析问题的主要范例。然而,GNNs可能会对某些人口分组作出有偏见的预测。理解预测中的偏见是如何产生的至关重要,因为它指导了GNN debising机制的设计。然而,大多数现有工作都主要侧重于GNN debiainsing,但却没有解释这种偏见是如何诱发的。在本文件中,我们研究了解释GNN的不公平性的新问题,将它归咎于培训节点的影响。具体地说,我们提出了一个名为概率分布差异的新颖战略(PDD),以衡量GNNs所显示的偏差,并开发一种算法,以有效估计每个培训节点对这种偏差的影响。我们通过对真实世界数据集的实验来验证PDDD的有效性和影响估计的有效性。最后,我们还演示了如何将拟议的框架用于消除偏见性GNNS。开放源代码可以在 https://github.com/yushun/BIND中找到。