Machine generated text is increasingly difficult to distinguish from human authored text. Powerful open-source models are freely available, and user-friendly tools that democratize access to generative models are proliferating. ChatGPT, which was released shortly after the first preprint of this survey, epitomizes these trends. The great potential of state-of-the-art natural language generation (NLG) systems is tempered by the multitude of avenues for abuse. Detection of machine generated text is a key countermeasure for reducing abuse of NLG models, with significant technical challenges and numerous open problems. We provide a survey that includes both 1) an extensive analysis of threat models posed by contemporary NLG systems, and 2) the most complete review of machine generated text detection methods to date. This survey places machine generated text within its cybersecurity and social context, and provides strong guidance for future work addressing the most critical threat models, and ensuring detection systems themselves demonstrate trustworthiness through fairness, robustness, and accountability.
翻译:机器生成的文本越来越难以与人类编写的文本区分。 强大的开放源码模型可以免费获得,使基因模型获取民主化的用户友好工具正在扩散。 在这次调查的最初印本后不久发布的查特GPT, 体现了这些趋势。 最先进的自然语言生成系统的巨大潜力因滥用渠道而减弱。 检测机器生成的文本是减少滥用NLG模型的关键应对措施,具有巨大的技术挑战和许多公开问题。 我们提供的一项调查包括:(1) 对当代NLG系统构成的威胁模型的广泛分析,和(2) 对机器生成的文本检测方法的最彻底审查,该调查将生成的文本置于其网络安全和社会背景下,并为今后处理最关键的威胁模型的工作提供强有力的指导,并确保检测系统本身通过公平、稳健和问责制表现出信任性。