Node representation learning has demonstrated its efficacy for various applications on graphs, which leads to increasing attention towards the area. However, fairness is a largely under-explored territory within the field, which may lead to biased results towards underrepresented groups in ensuing tasks. To this end, this work theoretically explains the sources of bias in node representations obtained via Graph Neural Networks (GNNs). Our analysis reveals that both nodal features and graph structure lead to bias in the obtained representations. Building upon the analysis, fairness-aware data augmentation frameworks on nodal features and graph structure are developed to reduce the intrinsic bias. Our analysis and proposed schemes can be readily employed to enhance the fairness of various GNN-based learning mechanisms. Extensive experiments on node classification and link prediction are carried out over real networks in the context of graph contrastive learning. Comparison with multiple benchmarks demonstrates that the proposed augmentation strategies can improve fairness in terms of statistical parity and equal opportunity, while providing comparable utility to state-of-the-art contrastive methods.
翻译:节点代表学习证明了其在图表上各种应用中的功效,这导致对该地区的日益关注,然而,公平性在很大程度上是该领域内探索不足的领域,可能导致在随后的任务中对代表人数不足的群体产生偏颇的结果。为此,这项工作从理论上解释了通过图形神经网络(GNNs)获得的节点表达中的偏见来源。我们的分析表明,节点特征和图表结构导致获得的表述中的偏见。在分析的基础上,制定了关于节点特征和图表结构的公平觉悟数据增强框架,以减少内在偏见。我们的分析与拟议的计划可以很容易地用来提高基于GNN的各类学习机制的公平性。在图表对比学习中,对节点分类和链接预测进行了广泛的实验。与多个基准的比较表明,拟议的扩展战略可以提高统计均等和平等机会方面的公平性,同时为最先进的对比方法提供可比的效用。