This thesis scrutinizes common assumptions underlying traditional machine learning approaches to fairness in consequential decision making. After challenging the validity of these assumptions in real-world applications, we propose ways to move forward when they are violated. First, we show that group fairness criteria purely based on statistical properties of observed data are fundamentally limited. Revisiting this limitation from a causal viewpoint we develop a more versatile conceptual framework, causal fairness criteria, and first algorithms to achieve them. We also provide tools to analyze how sensitive a believed-to-be causally fair algorithm is to misspecifications of the causal graph. Second, we overcome the assumption that sensitive data is readily available in practice. To this end we devise protocols based on secure multi-party computation to train, validate, and contest fair decision algorithms without requiring users to disclose their sensitive data or decision makers to disclose their models. Finally, we also accommodate the fact that outcome labels are often only observed when a certain decision has been made. We suggest a paradigm shift away from training predictive models towards directly learning decisions to relax the traditional assumption that labels can always be recorded. The main contribution of this thesis is the development of theoretically substantiated and practically feasible methods to move research on fair machine learning closer to real-world applications.
翻译:本文审视了传统机器学习方法的通用假设, 从而得出相应的决策的公平性。 在对这些假设在现实世界应用中的有效性提出质疑之后, 我们提出当这些假设被违反时如何前进。 首先, 我们显示纯粹基于观察数据的统计属性的团体公平标准根本有限。 从因果角度重新审视这一限制, 我们从因果角度出发, 我们开发了一个更灵活的概念框架、 因果公平标准, 以及实现这些限制的首个算法。 我们还提供了工具, 分析相信可能因果公平算法的敏感性是如何导致因果图的偏差。 其次, 我们克服了敏感数据在实践上很容易得到的假设。 为此, 我们设计了基于安全多党计算来培训、 验证和质疑公平决策算法的协议, 而没有要求用户披露其敏感数据或决策者披露其模型。 最后, 我们还考虑到这样的事实, 即结果标签往往只在作出某种决定时才得到遵守。 我们建议从培训预测模型转向直接学习决定, 以放松传统假设, 即标签可以一直被记录下来。 这个理论的主要贡献是发展到更接近于更接近于研究的机器学习应用的方法。