Much research has been devoted to the problem of learning fair representations; however, they do not explicitly the relationship between latent representations. In many real-world applications, there may be causal relationships between latent representations. Furthermore, most fair representation learning methods focus on group-level fairness and are based on correlations, ignoring the causal relationships underlying the data. In this work, we theoretically demonstrate that using the structured representations enable downstream predictive models to achieve counterfactual fairness, and then we propose the Counterfactual Fairness Variational AutoEncoder (CF-VAE) to obtain structured representations with respect to domain knowledge. The experimental results show that the proposed method achieves better fairness and accuracy performance than the benchmark fairness methods.
翻译:对学习公平陈述问题进行了大量研究;然而,这些研究并未明确潜在陈述之间的关系;在许多现实世界的应用中,潜在陈述之间可能存在因果关系;此外,大多数公平陈述学习方法侧重于群体一级的公平性,以相关关系为基础,忽视数据背后的因果关系;在这项工作中,我们理论上表明,使用结构化的表述方法使下游预测模型能够实现反事实公平,然后我们提议采用反事实公平变换自动计算器(CF-VAE),以获得有关领域知识的结构性表述;实验结果显示,拟议方法比基准公平性方法更公平和准确。