In an ideal world, deployed machine learning models will enhance our society. We hope that those models will provide unbiased and ethical decisions that will benefit everyone. However, this is not always the case; issues arise during the data preparation process throughout the steps leading to the models' deployment. The continued use of biased datasets and processes will adversely damage communities and increase the cost of fixing the problem later. In this work, we walk through the decision-making process that a researcher should consider before, during, and after a system deployment to understand the broader impacts of their research in the community. Throughout this paper, we discuss fairness, privacy, and ownership issues in the machine learning pipeline; we assert the need for a responsible human-over-the-loop methodology to bring accountability into the machine learning pipeline, and finally, reflect on the need to explore research agendas that have harmful societal impacts. We examine visual privacy research and draw lessons that can apply broadly to artificial intelligence. Our goal is to systematically analyze the machine learning pipeline for visual privacy and bias issues. We hope to raise stakeholder (e.g., researchers, modelers, corporations) awareness as these issues propagate in this pipeline's various machine learning phases.
翻译:在一个理想的世界中,部署的机器学习模式将提升我们的社会。我们希望这些模式将提供有利于每个人的公正和道德决定。然而,情况并非总是如此;在模型部署的各个步骤中,在数据编制过程中出现问题;继续使用偏向的数据集和程序将对社区产生不利影响,并增加以后解决问题的成本。在这项工作中,我们走过一个研究人员应当考虑的决策进程,在部署系统之前、期间和之后,了解其在社区研究的更广泛影响。在本文件中,我们讨论机器学习管道中的公平、隐私和所有权问题;我们主张需要一种负责任的人对地方法,将问责制引入机器学习管道,最后,思考探索具有有害社会影响的研究议程的必要性。我们研究了视觉隐私研究,并吸取可以广泛应用于人工智能的经验教训。我们的目标是系统地分析机器学习管道,以了解视觉隐私和偏见问题。我们希望在本文中提高利益攸关方(例如研究人员、建模者、公司)的认识,因为这些问题在管道的各个机器学习阶段不断传播。