Crowdsourcing is being increasingly adopted as a platform to run studies with human subjects. Running a crowdsourcing experiment involves several choices and strategies to successfully port an experimental design into an otherwise uncontrolled research environment, e.g., sampling crowd workers, mapping experimental conditions to micro-tasks, or ensure quality contributions. While several guidelines inform researchers in these choices, guidance of how and what to report from crowdsourcing experiments has been largely overlooked. If under-reported, implementation choices constitute variability sources that can affect the experiment's reproducibility and prevent a fair assessment of research outcomes. In this paper, we examine the current state of reporting of crowdsourcing experiments and offer guidance to address associated reporting issues. We start by identifying sensible implementation choices, relying on existing literature and interviews with experts, to then extensively analyze the reporting of 171 crowdsourcing experiments. Informed by this process, we propose a checklist for reporting crowdsourcing experiments.
翻译:众包实验涉及若干选择和战略,以便成功地将实验设计移植到一个本来不受控制的研究环境中,例如,抽样人群工人、为微任务绘制实验条件图或确保质量贡献等。虽然若干准则为研究人员提供了这些选择,但众包实验如何和什么报告的指导在很大程度上被忽略了。如果报告不足,实施选择构成可变性来源,可能影响实验的再复制,并妨碍对研究成果的公平评估。在本文件中,我们审查了众包实验报告的现状,并提供了解决相关报告问题的指南。我们首先根据现有文献和与专家的访谈,确定合理的执行选择,然后广泛分析171项众包实验的报告情况。根据这一过程,我们提出了一个报告众包实验的核对表。