In recent years, most fairness strategies in machine learning models focus on mitigating unwanted biases by assuming that the sensitive information is observed. However this is not always possible in practice. Due to privacy purposes and var-ious regulations such as RGPD in EU, many personal sensitive attributes are frequently not collected. We notice a lack of approaches for mitigating bias in such difficult settings, in particular for achieving classical fairness objectives such as Demographic Parity and Equalized Odds. By leveraging recent developments for approximate inference, we propose an approach to fill this gap. Based on a causal graph, we rely on a new variational auto-encoding based framework named SRCVAE to infer a sensitive information proxy, that serve for bias mitigation in an adversarial fairness approach. We empirically demonstrate significant improvements over existing works in the field. We observe that the generated proxy's latent space recovers sensitive information and that our approach achieves a higher accuracy while obtaining the same level of fairness on two real datasets, as measured using com-mon fairness definitions.
翻译:近年来,机器学习模型中的大多数公平战略侧重于通过假定敏感信息得到遵守来减少不必要的偏见,但在实践中并不总是可能这样做。由于隐私目的和欧盟的RGPD等不同规则,许多个人敏感属性往往没有收集。我们注意到,在这种困难环境中,特别是在实现人口均等和偶数等传统公平目标方面,缺乏减少偏见的方法。我们利用最近的事态发展来大致推理,提出了填补这一差距的办法。根据因果图表,我们依靠一个新的基于变异的自动编码框架SRCVAE来推断一个敏感信息代用器,用于在对抗性公平办法中减少偏见。我们从经验上表明,在实地现有工作中,我们发现,产生的代用空间恢复了敏感信息,我们的方法在获得两个真实数据集的同等程度的公平性的同时,在使用Com-mon公平定义衡量时,实现了更高的准确性。