Fair machine learning aims to avoid treating individuals or sub-populations unfavourably based on \textit{sensitive attributes}, such as gender and race. Those methods in fair machine learning that are built on causal inference ascertain discrimination and bias through causal effects. Though causality-based fair learning is attracting increasing attention, current methods assume the true causal graph is fully known. This paper proposes a general method to achieve the notion of counterfactual fairness when the true causal graph is unknown. To be able to select features that lead to counterfactual fairness, we derive the conditions and algorithms to identify ancestral relations between variables on a \textit{Partially Directed Acyclic Graph (PDAG)}, specifically, a class of causal DAGs that can be learned from observational data combined with domain knowledge. Interestingly, we find that counterfactual fairness can be achieved as if the true causal graph were fully known, when specific background knowledge is provided: the sensitive attributes do not have ancestors in the causal graph. Results on both simulated and real-world datasets demonstrate the effectiveness of our method.
翻译:公平机器学习的目的是避免根据性别与种族等/textit{敏感属性来对待个人或亚群体。建立在因果推断基础上的公平机器学习方法通过因果关系确定歧视和偏见。虽然基于因果关系的公平学习正在引起越来越多的注意,但目前的方法完全知道真正的因果图表。本文件提出了在真实因果图表未知时实现反事实公平概念的一般方法。为了能够选择导致反事实公平的特点,我们得出条件和算法,以确定在\ textit{Partily 直接的周期图(PPAAG)}上的变量之间的祖传关系。具体地说,从观察数据中可以学到的一类因果数据与域知识相结合。有趣的是,我们发现,在提供具体背景知识时,如果完全知道真正的因果图表:敏感的属性在因果图表中没有祖先,那么反事实公平是可以实现的。关于模拟和真实世界数据集的结果证明了我们的方法的有效性。