Clustering algorithms are widely used in many societal resource allocation applications, such as loan approvals and candidate recruitment, among others, and hence, biased or unfair model outputs can adversely impact individuals that rely on these applications. To this end, many fair clustering approaches have been recently proposed to counteract this issue. Due to the potential for significant harm, it is essential to ensure that fair clustering algorithms provide consistently fair outputs even under adversarial influence. However, fair clustering algorithms have not been studied from an adversarial attack perspective. In contrast to previous research, we seek to bridge this gap and conduct a robustness analysis against fair clustering by proposing a novel black-box fairness attack. Through comprehensive experiments, we find that state-of-the-art models are highly susceptible to our attack as it can reduce their fairness performance significantly. Finally, we propose Consensus Fair Clustering (CFC), the first robust fair clustering approach that transforms consensus clustering into a fair graph partitioning problem, and iteratively learns to generate fair cluster outputs. Experimentally, we observe that CFC is highly robust to the proposed attack and is thus a truly robust fair clustering alternative.
翻译:许多社会资源分配应用中广泛使用集群算法,例如贷款批准和候选人征聘等,因此,偏向或不公平的模型产出会对依赖这些应用的个人产生不利影响。为此目的,最近提出了许多公平的集群办法,以抵消这一问题。由于可能造成重大伤害,必须确保公平的集群算法提供一贯的公平产出,即使在对抗性影响下也是如此。然而,公平的集群算法并没有从对抗性攻击的角度加以研究。与以往的研究不同,我们力求弥补这一差距,并通过提出新的黑盒公平攻击来对公平的集群进行稳健分析。我们通过全面试验发现,最先进的模型极易受到我们的攻击,因为这样可以大大降低它们的公平性能。最后,我们提出共识公平分组(CFC),这是将共识组合转化为公平图表分割问题的第一种稳健的公平分组办法,并反复学习产生公平的集群产出。我们实验地发现,CFC对于拟议的攻击非常有力,因此是一种真正稳健的公平组合替代方案。