Recent advances in technology for hyper-realistic visual effects provoke the concern that deepfake videos of political speeches will soon be visually indistinguishable from authentic video recordings. The conventional wisdom in communications research predicts people will fall for fake news more often when the same version of a story is presented as a video rather than text. Here, we evaluate how accurately 41,822 participants distinguish real political speeches from fabrications in an experiment where speeches are randomized to appear as permutations of text, audio, and video. We find access to audio and visual communication modalities improve participants' accuracy. Here, human judgment relies more on how something is said, the audio-visual cues, than what is said, the speech content. However, we find that reflective reasoning moderates the degree to which participants consider visual information: low performance on the Cognitive Reflection Test is associated with an over-reliance on what is said.
翻译:近来超现实视觉效应技术的进步引起了人们的担忧,即政治演讲的深假视频很快就会与真实的视频记录在视觉上无法区分。通信研究中的常规智慧预测,当一个故事的同一版本以视频而不是文字来展示时,人们会更经常地成为假新闻。在这里,我们评估41,822名参与者如何准确地区分真正的政治演讲和实验中的编造。 在实验中,演讲随机地以文字、音频和视频的变相出现。我们发现,听音和视觉通信模式的获取会提高参与者的准确性。在这里,人类的判断更多地依赖于语言的表达方式,即视听提示,而不是言论内容。然而,我们发现,这种反射推理使参与者对视觉信息的考虑程度有所节制:认知反省测试的低表现与过度依赖所说的话有关。