Machine learning algorithms often make decisions on behalf of agents with varied and sometimes conflicting interests. In domains where agents can choose to take their own action or delegate their action to a central mediator, an open question is how mediators should take actions on behalf of delegating agents. The main existing approach uses delegating agents to punish non-delegating agents in an attempt to get all agents to delegate, which tends to be costly for all. We introduce a Pareto Mediator which aims to improve outcomes for delegating agents without making any of them worse off. Our experiments in random normal form games, a restaurant recommendation game, and a reinforcement learning sequential social dilemma show that the Pareto Mediator greatly increases social welfare. Also, even when the Pareto Mediator is based on an incorrect model of agent utility, performance gracefully degrades to the pre-intervention level, due to the individual autonomy preserved by the voluntary mediator.
翻译:机器学习算法往往代表利益各异、有时是相互冲突的代理人作出决定。在代理人可以选择自己采取行动或将其行动委托给中央调解人的领域,一个未决问题是调解人应当如何代表委托代理人采取行动。主要的现行做法是委托代理人惩罚非委托代理人,试图让所有代理人进行委托,这往往给所有人带来代价。我们引入了帕雷托调解员,目的是改善委托代理人的结果,而不会使其中任何一个更糟糕。我们在随机的普通游戏、餐馆推荐游戏和强化学习连续的社会困境中进行的实验表明,帕雷托调解员极大地提高了社会福利。此外,即使帕雷托调解员基于不正确的代理人效用模式,由于自愿调解人的个人自主权,业绩也优于干预前的水平。