Both industry and academia have made considerable progress in developing trustworthy and responsible machine learning (ML) systems. While critical concepts like fairness and explainability are often addressed, the safety of systems is typically not sufficiently taken into account. By viewing data-driven decision systems as socio-technical systems, we draw on the uncertainty in ML literature to show how fairML systems can also be safeML systems. We posit that a fair model needs to be an uncertainty-aware model, e.g. by drawing on distributional regression. For fair decisions, we argue that a safe fail option should be used for individuals with uncertain categorization. We introduce semi-structured deep distributional regression as a modeling framework which addresses multiple concerns brought against standard ML models and show its use in a real-world example of algorithmic profiling of job seekers.
翻译:产业界和学术界在发展可信和负责的机器学习系统方面都取得了相当大的进展。虽然常常涉及公平和解释性等关键概念,但系统的安全通常没有得到充分的考虑。通过将数据驱动的决策系统看成社会技术系统,我们利用制造业和学术界文献的不确定性来表明公平ML系统如何也能成为安全ML系统。我们假设公平模式需要是一种不确定性的模型,例如,利用分布式回归。为了作出公平的决定,我们主张,对于分类不确定的个人,应当使用安全的失败选项。我们引入半结构化的深度分布回归作为示范框架,解决针对标准ML模型提出的多种关切,并展示其在现实世界中对求职者的算法分析实例中的使用情况。