Generative adversarial networks (GANs) have made remarkable progress in synthesizing realistic-looking images that effectively outsmart even humans. Although several detection methods can recognize these deep fakes by checking for image artifacts from the generation process, multiple counterattacks have demonstrated their limitations. These attacks, however, still require certain conditions to hold, such as interacting with the detection method or adjusting the GAN directly. In this paper, we introduce a novel class of simple counterattacks that overcomes these limitations. In particular, we show that an adversary can remove indicative artifacts, the GAN fingerprint, directly from the frequency spectrum of a generated image. We explore different realizations of this removal, ranging from filtering high frequencies to more nuanced frequency-peak cleansing. We evaluate the performance of our attack with different detection methods, GAN architectures, and datasets. Our results show that an adversary can often remove GAN fingerprints and thus evade the detection of generated images.
翻译:生成的对抗网络(GANs)在综合现实的图像方面取得了显著的进展,这些图像实际上比人类更聪明。虽然一些探测方法能够通过检查生成过程中的图像制品来识别这些深层假象,但多次反攻显示出其局限性。然而,这些袭击仍然需要一定的维持条件,例如与检测方法互动或直接调整GAN。在本文中,我们引入了一种克服这些限制的新型简单反击。特别是,我们表明对手可以直接从生成图像的频谱中移除指示性文物GAN指纹。我们探索了从过滤高频率到更细微的频率清扫等不同程度的清除方法。我们用不同的检测方法、GAN结构以及数据集来评估我们袭击的性能。我们的结果表明,对手往往可以清除GAN指纹,从而逃避对生成图像的探测。