Fact-checking is one of the effective solutions in fighting online misinformation. However, traditional fact-checking is a process requiring scarce expert human resources, and thus does not scale well on social media because of the continuous flow of new content to be checked. Methods based on crowdsourcing have been proposed to tackle this challenge, as they can scale with a smaller cost, but, while they have shown to be feasible, have always been studied in controlled environments. In this work, we study the first large-scale effort of crowdsourced fact-checking deployed in practice, started by Twitter with the Birdwatch program. Our analysis shows that crowdsourcing may be an effective fact-checking strategy in some settings, even comparable to results obtained by human experts, but does not lead to consistent, actionable results in others. We processed 11.9k tweets verified by the Birdwatch program and report empirical evidence of i) differences in how the crowd and experts select content to be fact-checked, ii) how the crowd and the experts retrieve different resources to fact-check, and iii) the edge the crowd shows in fact-checking scalability and efficiency as compared to expert checkers.
翻译:事实调查是打击网上错误信息的有效解决办法之一。然而,传统事实调查是一个需要稀缺的专家人力资源的过程,因此在社交媒体上规模不高,因为需要检查的新内容不断流动。基于众包的方法已经提出来应对这一挑战,因为它们可以以较低的成本规模进行,但是,虽然它们证明是可行的,但总是在受控制的环境中进行研究。在这项工作中,我们研究了由Twitter与观鸟方案一起开始的、在实际操作中部署的由众包提供事实调查的首次大规模努力。我们的分析表明,在某些环境中,众包可能是有效的事实调查战略,甚至可以与人类专家获得的结果相比,但不会导致一致的、可操作的结果。我们处理了由观鸟方案核实的11.9k条推文,并报告了实证证据:(一) 人群和专家如何选择内容进行事实核查的差别,二) 人群和专家如何获取不同资源进行事实核查,以及三) 人群在与专家检查者相比,在事实核对的可扩展性和效率方面展示的边缘。