Online community moderators often rely on social signals such as whether or not a user has an account or a profile page as clues that users may cause problems. Reliance on these clues can lead to overprofiling bias when moderators focus on these signals but overlook the misbehavior of others. We propose that algorithmic flagging systems deployed to improve the efficiency of moderation work can also make moderation actions more fair to these users by reducing reliance on social signals and making norm violations by everyone else more visible. We analyze moderator behavior in Wikipedia as mediated by RCFilters, a system which displays social signals and algorithmic flags, and estimate the causal effect of being flagged on moderator actions. We show that algorithmically flagged edits are reverted more often, especially those by established editors with positive social signals, and that flagging decreases the likelihood that moderation actions will be undone. Our results suggest that algorithmic flagging systems can lead to increased fairness in some contexts but that the relationship is complex and contingent.
翻译:在线社区管理者往往依赖社会信号,例如用户是否拥有账户或配置页面,作为用户可能造成问题的线索。 依靠这些线索可能导致在主持人关注这些信号时出现超常的偏差,但却忽略了其他人的错误行为。 我们提议,为提高温和工作效率而部署的算法标志系统也可以减少对社会信号的依赖,使其他人违反规范的行为更加明显,从而使温和行动对这些用户更加公平。 我们分析维基百科中由区域合作者调解的主持人行为,这个系统显示社会信号和算法标志,并估计在主持人行动中标注的因果关系。 我们显示,以逻辑标定的编辑更经常被重现,特别是由拥有积极社会信号的常设编辑重现的编辑重现,而挂旗会降低温和行动被撤销的可能性。 我们的结果表明,在某种情况下,算法标志系统可以导致增加公平性,但关系复杂和有附带性。