Human communication is increasingly intermixed with language generated by AI. Across chat, email, and social media, AI systems produce smart replies, autocompletes, and translations. AI-generated language is often not identified as such but poses as human language, raising concerns about novel forms of deception and manipulation. Here, we study how humans discern whether one of the most personal and consequential forms of language - a self-presentation - was generated by AI. Across six experiments, participants (N = 4,650) tried to identify self-presentations generated by state-of-the-art language models. Across professional, hospitality, and romantic settings, we find that humans are unable to identify AI-generated self-presentations. Combining qualitative analyses with language feature engineering, we find that human judgments of AI-generated language are handicapped by intuitive but flawed heuristics such as associating first-person pronouns, authentic words, or family topics with humanity. We show that these heuristics make human judgment of generated language predictable and manipulable, allowing AI systems to produce language perceived as more human than human. We conclude by discussing solutions - such as AI accents or fair use policies - to reduce the deceptive potential of generated language, limiting the subversion of human intuition.
翻译:在聊天、电子邮件和社交媒体中,人工智能系统产生智能答复、自动完成和翻译。人工智能生成的语言往往不被识别为是人的语言,而是被识别为人的语言,引起人们对新形式的欺骗和操纵的担忧。在这里,我们研究人类如何辨别是否是AI产生的个人和最间接的语言形式之一,即自我展示。在六个实验中,参与者(N=4,650)试图辨别由最先进的语言模型产生的自我展示。在专业、招待和浪漫环境中,我们发现人类无法辨别人工生成的自我展示。将定性分析与语言特征工程相结合,我们发现人类对人工生成语言的判断因直观但有缺陷的超常识学而受阻,例如将第一人代代言人、正言或家庭专题与人类联系起来。我们证明这些超自然论使人类对生成的语言的判断具有可预见性和可人性,允许人工智能系统生成被认为比人类更具有人性的语言。我们的结论是,通过讨论将质量分析与语言的转化的潜力加以限制,从而降低人类的自我感官。