Along with the increasing availability of data in many sectors has come the rise of data-driven models to inform decision-making and policy. In the health care sector, these models have the potential to benefit both patients and health care providers but can also entrench or exacerbate health inequities. Existing "algorithmic fairness" methods for measuring and correcting model bias fall short of what is needed for clinical applications in two key ways. First, methods typically focus on a single grouping along which discrimination may occur rather than considering multiple, intersecting groups such as gender and race. Second, in clinical applications, risk prediction is typically used to guide treatment, and use of a treatment presents distinct statistical issues that invalidate most existing fairness measurement techniques. We present novel unfairness metrics that address both of these challenges. We also develop a complete framework of estimation and inference tools for our metrics, including the unfairness value ("u-value"), used to determine the relative extremity of an unfairness measurement, and standard errors and confidence intervals employing an alternative to the standard bootstrap.
翻译:在许多部门,随着数据驱动模式的兴起,许多部门的数据越来越容易获得,为决策和政策提供信息。在保健部门,这些模式有可能使病人和保健提供者都受益,但也可能巩固或加剧保健方面的不平等。现有的衡量和纠正模式偏向的“公平性”方法不能满足临床应用的需要。首先,方法通常侧重于单一组别,在其中可能出现歧视,而不是考虑性别和种族等多重交叉群体。第二,在临床应用中,风险预测通常用于指导治疗,而使用一种治疗方法则提出了不同的统计问题,使大多数现有的公平计量技术无效。我们提出了新的不公平性指标,解决了这两个挑战。我们还为我们的计量方法制定了完整的估算和推论工具框架,包括不公平价值(“价值”),用于确定不公平计量的相对极端性,以及标准错误和信任间隔,并采用标准靴套的替代办法。