When does a machine learning model predict the future of individuals and when does it recite patterns that predate the individuals? In this work, we propose a distinction between these two pathways of prediction, supported by theoretical, empirical, and normative arguments. At the center of our proposal is a family of simple and efficient statistical tests, called backward baselines, that demonstrate if, and to which extent, a model recounts the past. Our statistical theory provides guidance for interpreting backward baselines, establishing equivalences between different baselines and familiar statistical concepts. Concretely, we derive a meaningful backward baseline for auditing a prediction system as a black box, given only background variables and the system's predictions. Empirically, we evaluate the framework on different prediction tasks derived from longitudinal panel surveys, demonstrating the ease and effectiveness of incorporating backward baselines into the practice of machine learning.
翻译:当机器学习模型预测个人未来时,当机器学习模型背诵个人之前的模式时,它何时会预测个人的未来?在这项工作中,我们提议对这两种预测途径加以区分,并辅以理论、经验和规范性论据。我们提案的核心是简单有效的统计测试体系,称为后向基线,以表明一个模型是否和在多大程度上重述过去。我们的统计理论为解释后向基线、确定不同基线和熟悉的统计概念之间的等同提供了指导。具体地说,我们为将一个预测系统作为黑盒进行审计制定了一个有意义的后向基线,只考虑到背景变量和系统预测。我们评估了从纵向小组调查中得出的不同预测任务的框架,显示了将后向基线纳入机器学习实践的容易性和有效性。