It is crucial to consider the social and ethical consequences of AI and ML based decisions for the safe and acceptable use of these emerging technologies. Fairness, in particular, guarantees that the ML decisions do not result in discrimination against individuals or minorities. Identifying and measuring reliably fairness/discrimination is better achieved using causality which considers the causal relation, beyond mere association, between the sensitive attribute (e.g. gender, race, religion, etc.) and the decision (e.g. job hiring, loan granting, etc.). The big impediment to the use of causality to address fairness, however, is the unavailability of the causal model (typically represented as a causal graph). Existing causal approaches to fairness in the literature do not address this problem and assume that the causal model is available. In this paper, we do not make such assumption and we review the major algorithms to discover causal relations from observable data. This study focuses on causal discovery and its impact on fairness. In particular, we show how different causal discovery approaches may result in different causal models and, most importantly, how even slight differences between causal models can have significant impact on fairness/discrimination conclusions. These results are consolidated by empirical analysis using synthetic and standard fairness benchmark datasets. The main goal of this study is to highlight the importance of the causal discovery step to appropriately address fairness using causality.
翻译:特别是,公平性保证多边协议决定不会导致对个人或少数群体的歧视; 确定和衡量可靠的公平/歧视,更好地利用因果关系,不仅考虑到敏感属性(例如性别、种族、宗教等)与决定(例如雇用、贷款等)之间的因果关系,而且考虑到其与决定(例如雇用、贷款等)之间的因果关系。然而,使用因果关系解决公平问题的主要障碍是缺乏因果模式(通常以因果图表示),文献中现有的公平因果方法没有解决这一问题,并假定存在因果模式。在本文件中,我们不作出这种假设,我们审查从可观察数据中发现因果关系的主要算法。本研究的重点是因果发现及其对公平的影响。我们特别要说明因果发现方法如何不同,而且最重要的是,因果模型之间的微差异如何可能对公平/歧视结论产生重大影响。这些结果通过采用综合的因果性基准分析,通过综合的因果性指标和标准的因果性标准,通过综合的因果性指标,巩固了这一因果关系。