Algorithmic risk assessments hold the promise of greatly advancing accurate decision-making, but in practice, multiple real-world examples have been shown to distribute errors disproportionately across demographic groups. In this paper, we characterize why error disparities arise in the first place. We show that predictive uncertainty often leads classifiers to systematically disadvantage groups with lower-mean outcomes, assigning them smaller true and false positive rates than their higher-mean counterparts. This can occur even when prediction is group-blind. We prove that to avoid these error imbalances, individuals in lower-mean groups must either be over-represented among positive classifications or be assigned more accurate predictions than those in higher-mean groups. We focus on the latter condition as a solution to bridge error rate divides and show that data acquisition for low-mean groups can increase access to opportunity. We call the strategy "affirmative information" and compare it to traditional affirmative action in the classification task of identifying creditworthy borrowers.
翻译:算法风险评估有望大大推进准确的决策,但在实践中,已经展示了多个真实世界的例子,在人口群体中传播错误过多。在本文中,我们首先说明了为什么出现错误差异的原因。我们表明,预测性不确定性往往导致分类者系统性地处于劣势,其结果较低,其真实和假正率低于其高等对等群体。即使预测是群体盲目的,这也有可能发生。我们证明,为避免这些错误的不平衡,低等群体中的个人必须要么在肯定的分类中过多地被分配,要么被分配到比中等群体更准确的预测。我们注重后一种条件,作为弥合误差差距的解决方案,并表明为低等群体获取数据可以增加机会。我们称该战略为“肯定信息”,并将其与在确定有信用的借款人分类工作中的传统肯定行动作比较。