In performative prediction, a predictive model impacts the distribution that generates future data, a phenomenon that is being ignored in classical supervised learning. In this closed-loop setting, the natural measure of performance, denoted the performative risk, captures the expected loss incurred by a predictive model after deployment. The core difficulty of minimizing the performative risk is that the data distribution itself depends on the model parameters. This dependence is governed by the environment and not under the control of the learner. As a consequence, even the choice of a convex loss function can result in a highly non-convex performative risk minimization problem. Prior work has identified a pair of general conditions on the loss and the mapping from model parameters to distributions that implies convexity of the performative risk. In this paper, we relax these assumptions and focus on obtaining weaker notions of convexity, without sacrificing the amenability of the performative risk minimization problem for iterative optimization methods.
翻译:在实绩预测中,预测模型影响产生未来数据的分布,这种现象在古典监督下的学习中被忽视。在这种闭路学习中,自然性能计量,表示性能风险,捕捉了部署后预测性模型的预期损失。最大限度地减少性能风险的核心困难在于数据分布本身取决于模型参数。这种依赖受环境的制约,不受学习者控制。因此,即使选择卷流损失功能也可能导致高度非凝解性能风险最小化的问题。先前的工作已经确定了损失的一般条件以及从模型参数到分布图的映射,这意味着性能风险的共性。在本文中,我们放松这些假设,侧重于获得较弱的共性概念,同时不牺牲迭层优化方法的可执行性风险最小化问题。