Machine learning (ML) enabled classification models are becoming increasingly popular for tackling the sheer volume and speed of online misinformation and other content that could be identified as harmful. In building these models, data scientists need to take a stance on the legitimacy, authoritativeness and objectivity of the sources of ``truth" used for model training and testing. This has political, ethical and epistemic implications which are rarely addressed in technical papers. Despite (and due to) their reported high accuracy and performance, ML-driven moderation systems have the potential to shape online public debate and create downstream negative impacts such as undue censorship and the reinforcing of false beliefs. Using collaborative ethnography and theoretical insights from social studies of science and expertise, we offer a critical analysis of the process of building ML models for (mis)information classification: we identify a series of algorithmic contingencies--key moments during model development that could lead to different future outcomes, uncertainty and harmful effects as these tools are deployed by social media platforms. We conclude by offering a tentative path toward reflexive and responsible development of ML tools for moderating misinformation and other harmful content online.
翻译:机器学习(ML)启用的分类模型越来越受欢迎,可以处理海量和高速的在线虚假信息和其他可能被视为有害的内容。 在构建这些模型时,数据科学家需要对用于模型训练和测试的“真相”来源的合法性,权威性和客观性采取立场。 这在技术论文中很少涉及政治,伦理和认识论的影响。 尽管(也由于)它们报告了高准确性和性能,但基于ML的调停系统具有塑造网络公共辩论并产生下游负面影响(如不当审查和增强错误信念)的潜力。 使用协作民族志和社会科学和专业知识的理论洞见,我们为(误)信息分类构建ML模型的过程提供了批判性分析:我们确定了一系列算法偶然事件-在模型开发期间的关键时刻,这些事件可能导致不同的未来结果,不确定性和有害影响,因为这些工具由社交媒体平台部署。 我们最后提供了一条通向反思和负责任发展ML工具以调解在线错误信息和其他有害内容的初步路径。