As text generated by large language models proliferates, it becomes vital to understand how humans engage with such text, and whether or not they are able to detect when the text they are reading did not originate with a human writer. Prior work on human detection of generated text focuses on the case where an entire passage is either human-written or machine-generated. In this paper, we study a more realistic setting where text begins as human-written and transitions to being generated by state-of-the-art neural language models. We show that, while annotators often struggle at this task, there is substantial variance in annotator skill and that given proper incentives, annotators can improve at this task over time. Furthermore, we conduct a detailed comparison study and analyze how a variety of variables (model size, decoding strategy, fine-tuning, prompt genre, etc.) affect human detection performance. Finally, we collect error annotations from our participants and use them to show that certain textual genres influence models to make different types of errors and that certain sentence-level features correlate highly with annotator selection. We release the RoFT dataset: a collection of over 21,000 human annotations paired with error classifications to encourage future work in human detection and evaluation of generated text.
翻译:随着大型语言模型生成的文本扩散,了解人类如何与这些文本接触,以及他们是否能够发现他们阅读的文本不是与一位人类作家一起产生的文本何时产生,就变得至关重要。以前关于人类探测生成的文本的工作侧重于整个段落是人写或机器生成的。在本文中,我们研究一个更现实的环境,即文本开始作为人写和过渡到由最先进的神经语言模型产生的。我们表明,尽管警告者经常在这项工作中挣扎,但说明者的技能存在很大的差异,如果有适当的激励,说明者可以随着时间的推移改进这项任务。此外,我们进行详细的比较研究,分析各种变量(模型大小、解码战略、微调、迅速的基因等)如何影响人类检测性能。最后,我们从参与者那里收集错误说明,并用它们来表明某些文字类型影响模型,造成不同类型的错误,以及某些句级特征与注释选择高度相关。我们发布了“RoFT”数据库,鼓励将人类识别数据与21 000以上分类进行分类。