Undesired bias afflicts both human and algorithmic decision making, and may be especially prevalent when information processing trade-offs incentivize the use of heuristics. One primary example is \textit{statistical discrimination} -- selecting social partners based not on their underlying attributes, but on readily perceptible characteristics that covary with their suitability for the task at hand. We present a theoretical model to examine how information processing influences statistical discrimination and test its predictions using multi-agent reinforcement learning with various agent architectures in a partner choice-based social dilemma. As predicted, statistical discrimination emerges in agent policies as a function of both the bias in the training population and of agent architecture. All agents showed substantial statistical discrimination, defaulting to using the readily available correlates instead of the outcome relevant features. We show that less discrimination emerges with agents that use recurrent neural networks, and when their training environment has less bias. However, all agent algorithms we tried still exhibited substantial bias after learning in biased training populations.
翻译:人类和算法决策都受到不理想的偏见的影响,在信息处理交易鼓励使用超自然论时,这种偏见可能特别普遍。一个主要的例子就是 \ textit{统计歧视} -- 选择社会伙伴的依据不是其基本属性,而是其容易察觉的特征,这些特征与手头的任务相适应。我们提出了一个理论模型,以研究信息处理如何影响统计歧视,并用多试剂强化学习方法测试其预测,在伙伴选择的社会困境中,采用多试剂强化结构。正如预测的那样,在代理政策中出现了统计歧视,这是培训人口和代理结构偏见的函数。所有代理都表现出严重的统计歧视,不使用现成的关联关系而不是结果相关特征。我们表明,使用经常性神经网络的代理以及其培训环境不那么偏颇时,歧视就会减少。然而,我们尝试的所有代理算法在学习有偏见的培训人口后仍然表现出实质性的偏见。