The responsible use of machine learning tools in real world high-stakes decision making demands that we audit and control for potential biases against underrepresented groups. This process naturally requires access to the sensitive attribute one desires to control, such as demographics, gender, or other potentially sensitive features. Unfortunately, this information is often unavailable. In this work we demonstrate that one can still reliably estimate, and ultimately control, for fairness by using proxy sensitive attributes derived from a sensitive attribute predictor. Specifically, we first show that with just a little knowledge of the complete data distribution, one may use a sensitive attribute predictor to obtain bounds of the classifier's true fairness metric. Second, we demonstrate how one can provably control a classifier's worst-case fairness violation with respect to the true sensitive attribute by controlling for fairness with respect to the proxy sensitive attribute. Our results hold under assumptions that are significantly milder than previous works, and we illustrate these results with experiments on synthetic and real datasets.
翻译:在现实世界中,负责任地使用机器学习工具在现实世界的高度高度需要做出决策,要求我们审计和控制对代表性不足的群体的潜在偏差。这一过程自然需要获取人们想要控制的敏感属性,如人口、性别或其他潜在敏感特征。 不幸的是,这些信息往往无法获得。 在这项工作中,我们证明,通过使用敏感属性预测器产生的代理敏感属性,人们仍然可以可靠地估计并最终控制公平性。具体地说,我们首先显示,只要对完整的数据分布知之甚少,人们就可以使用敏感的属性预测器获得分类员真正公平度量的界限。第二,我们证明如何通过控制对代理敏感属性的公平性,来控制分类者对真正敏感属性的最坏的公平性侵犯。我们根据比以往工作要温和得多的假设,我们用合成和真实数据集的实验来说明这些结果。