Algorithmic fairness is an essential requirement as AI becomes integrated in society. In the case of social applications where AI distributes resources, algorithms often must make decisions that will benefit a subset of users, sometimes repeatedly or exclusively, while attempting to maximize specific outcomes. How should we design such systems to serve users more fairly? This paper explores this question in the case where a group of users works toward a shared goal in a social exergame called Step Heroes. We identify adverse outcomes in traditional multi-armed bandits (MABs) and formalize the Greedy Bandit Problem. We then propose a solution based on a new type of fairness-aware multi-armed bandit, Shapley Bandits. It uses the Shapley Value for increasing overall player participation and intervention adherence rather than the maximization of total group output, which is traditionally achieved by favoring only high-performing participants. We evaluate our approach via a user study (n=46). Our results indicate that our Shapley Bandits effectively mediates the Greedy Bandit Problem and achieves better user retention and motivation across the participants.
翻译:随着大赦国际融入社会,在社会应用中,在AI分配资源的社会应用中,算法往往必须作出决定,使一组用户受益,有时是反复或完全受益,同时试图尽量扩大具体结果。我们应如何设计这样的系统来更公平地为用户服务?本文探讨了这样一个问题,即如果一组用户在称为Step Heroes的社会竞赛中努力实现一个共同目标,我们发现传统多武装强盗(MABs)的不利结果,并正式确定贪婪匪帮问题。然后我们提出一种基于新型公平觉悟多武装强盗(Shapley Bandits)的解决方案。它利用“光彩色价值”来增加整体参与者的参与和干预,而不是实现整体群体产出的最大化,而后者传统上只通过偏好高表现参与者来实现。我们通过用户研究来评估我们的做法(n=46)。我们的结果表明,我们的Shapley Brits有效地调解了Greedy Bandit 问题,并在整个参与者中实现了更好的用户保留和动力。