Unsupervised representation learning on (large) graphs has received significant attention in the research community due to the compactness and richness of the learned embeddings and the abundance of unlabelled graph data. When deployed, these node representations must be generated with appropriate fairness constraints to minimize bias induced by them on downstream tasks. Consequently, group and individual fairness notions for graph learning algorithms have been investigated for specific downstream tasks. One major limitation of these fairness notions is that they do not consider the connectivity patterns in the graph leading to varied node influence (or centrality power). In this paper, we design a centrality-aware fairness framework for inductive graph representation learning algorithms. We propose CAFIN (Centrality Aware Fairness inducing IN-processing), an in-processing technique that leverages graph structure to improve GraphSAGE's representations - a popular framework in the unsupervised inductive setting. We demonstrate the efficacy of CAFIN in the inductive setting on two popular downstream tasks - Link prediction and Node Classification. Empirically, they consistently minimize the disparity in fairness between groups across datasets (varying from 18 to 80% reduction in imparity, a measure of group fairness) from different domains while incurring only a minimal performance cost.
翻译:无监督表示学习在(大型)图上由于学习嵌入的紧凑性和丰富性以及无标签图数据的丰富性,已经引起了研究界的广泛关注。在部署时,必须使用适当的公平性约束生成这些节点表示,以最小化对下游任务产生的偏差。因此,图学习算法的群组和个体公平性概念已被研究用于特定的下游任务。这些公平性概念的一个主要限制是它们未考虑图中的连接模式,导致节点影响(或中心度)的差异。在本文中,我们设计了一种基于中心性的公平性框架,用于归纳图表示学习算法。我们提出了CAFIM(Centrality Aware Fairness inducing IN-processing),这是一种基于IN处理技术以利用图结构来改进无监督归纳设置中的GraphSAGE表示的技术。我们在两个流行的下游任务 - 链接预测和节点分类的归纳设置中证明了CAFIM的有效性。从不同领域的数据集(在不同的领域中变化从18%到80%的差异,衡量群体公平性的一个指标)中,它们始终减少群体之间的公平差异,并仅产生最低的性能成本。