The rapid advances in deep generative models over the past years have led to highly {realistic media, known as deepfakes,} that are commonly indistinguishable from real to human eyes. These advances make assessing the authenticity of visual data increasingly difficult and pose a misinformation threat to the trustworthiness of visual content in general. Although recent work has shown strong detection accuracy of such deepfakes, the success largely relies on identifying frequency artifacts in the generated images, which will not yield a sustainable detection approach as generative models continue evolving and closing the gap to real images. In order to overcome this issue, we propose a novel fake detection that is designed to re-synthesize testing images and extract visual cues for detection. The re-synthesis procedure is flexible, allowing us to incorporate a series of visual tasks - we adopt super-resolution, denoising and colorization as the re-synthesis. We demonstrate the improved effectiveness, cross-GAN generalization, and robustness against perturbations of our approach in a variety of detection scenarios involving multiple generators over CelebA-HQ, FFHQ, and LSUN datasets. Source code is available at https://github.com/SSAW14/BeyondtheSpectrum.
翻译:过去几年来,深层基因模型的快速进步导致高度的(现实的)媒体,称为深假相,通常无法从真实的眼目中辨别出来。这些进步使得评估视觉数据的真实性越来越困难,对一般视觉内容的可信度构成错误的威胁。虽然最近的工作显示,这种深度假象的探测准确性很强,但成功在很大程度上依赖于确定所生成图像中的频率文物,这不会产生一种可持续的探测方法,因为基因模型继续演变,缩小到真实图像的差距。为了克服这一问题,我们建议了一种新型的假探测,目的是重新合成图像和提取视觉提示以便侦测。再合成程序是灵活的,使我们能够纳入一系列视觉任务——我们采用超级分辨率、分解和彩色作为再合成。我们展示了在CelibA-HQ、FFHHQ和LSAS数据设置的各种探测情景中的有效性、跨GAN一般化以及我们方法的坚固性,以对付对真实图像的破坏。在MebebA-HQ、MES/BERMUQ、MERMUQ和LS UNS数据代码是可用的。