Advances in natural language generation (NLG) have resulted in machine generated text that is increasingly difficult to distinguish from human authored text. Powerful open-source models are freely available, and user-friendly tools democratizing access to generative models are proliferating. The great potential of state-of-the-art NLG systems is tempered by the multitude of avenues for abuse. Detection of machine generated text is a key countermeasure for reducing abuse of NLG models, with significant technical challenges and numerous open problems. We provide a survey that includes both 1) an extensive analysis of threat models posed by contemporary NLG systems, and 2) the most complete review of machine generated text detection methods to date. This survey places machine generated text within its cybersecurity and social context, and provides strong guidance for future work addressing the most critical threat models, and ensuring detection systems themselves demonstrate trustworthiness through fairness, robustness, and accountability.
翻译:自然语言生成(NLG)的进步导致机器生成的文本越来越难以与人类创作的文本区分。强大的开放源码模型是免费的,而使基因模型进入民主化的用户友好工具正在扩散。最先进的NLG系统的巨大潜力因滥用渠道而减弱。探测机器生成的文本是减少滥用NLG模型的关键应对措施,具有巨大的技术挑战和许多尚未解决的问题。我们提供的一项调查包括:(1) 对当代NLG系统构成的威胁模型的广泛分析,和(2) 迄今为止对机器生成的文本检测方法的最彻底的审查。这项调查将机器生成的文本置于其网络和社会背景下,并为今后处理最关键的威胁模型的工作提供强有力的指导,并确保检测系统本身通过公平、稳健和问责制表现出信任性。