On many learning platforms, the optimization criteria guiding model training reflect the priorities of the designer rather than those of the individuals they affect. Consequently, users may act strategically to obtain more favorable outcomes, effectively contesting the platform's predictions. While past work has studied strategic user behavior on learning platforms, the focus has largely been on strategic responses to a deployed model, without considering the behavior of other users. In contrast, look-ahead reasoning takes into account that user actions are coupled, and -- at scale -- impact future predictions. Within this framework, we first formalize level-$k$ thinking, a concept from behavioral economics, where users aim to outsmart their peers by looking one step ahead. We show that, while convergence to an equilibrium is accelerated, the equilibrium remains the same, providing no benefit of higher-level reasoning for individuals in the long run. Then, we focus on collective reasoning, where users take coordinated actions by optimizing through their joint impact on the model. By contrasting collective with selfish behavior, we characterize the benefits and limits of coordination; a new notion of alignment between the learner's and the users' utilities emerges as a key concept. We discuss connections to several related mathematical frameworks, including strategic classification, performative prediction, and algorithmic collective action.
翻译:在许多学习平台上,指导模型训练的优化标准反映的是设计者的优先事项,而非受其影响的个体的优先事项。因此,用户可能会采取策略性行为以获得更有利的结果,从而有效地质疑平台的预测。虽然以往的研究已经探讨了学习平台上的策略性用户行为,但重点主要集中于用户对已部署模型的策略性反应,而未考虑其他用户的行为。相比之下,前瞻性推理则考虑到用户行为是相互关联的,并且在规模上会影响未来的预测。在此框架下,我们首先形式化了行为经济学中的概念——k级思维,即用户试图通过向前看一步来智胜其同伴。我们证明,尽管向均衡的收敛速度加快,但均衡本身保持不变,从长远来看,更高层次的推理对个体并无益处。接着,我们关注集体推理,即用户通过优化其对模型的共同影响来采取协调行动。通过对比集体行为与自私行为,我们刻画了协调的益处与局限;学习者效用与用户效用之间的一种新的对齐概念成为关键。我们讨论了与多个相关数学框架的联系,包括策略性分类、表演性预测和算法集体行动。