This paper explores the task of detecting images generated by text-to-image diffusion models. To evaluate this, we consider images generated from captions in the MSCOCO and Wikimedia datasets using two state-of-the-art models: Stable Diffusion and GLIDE. Our experiments show that it is possible to detect the generated images using simple Multi-Layer Perceptrons (MLPs), starting from features extracted by CLIP, or traditional Convolutional Neural Networks (CNNs). We also observe that models trained on images generated by Stable Diffusion can detect images generated by GLIDE relatively well, however, the reverse is not true. Lastly, we find that incorporating the associated textual information with the images rarely leads to significant improvement in detection results but that the type of subject depicted in the image can have a significant impact on performance. This work provides insights into the feasibility of detecting generated images, and has implications for security and privacy concerns in real-world applications. The code to reproduce our results is available at: https://github.com/davide-coccomini/Detecting-Images-Generated-by-Diffusers
翻译:本文探讨了检测文本到图像扩散模型生成的图像的任务。为评估这一任务,我们考虑了使用两种最先进的模型,即稳定的扩散和GLIDE,从MSCOCO和Wikimedia数据集中生成的图片。我们的实验证明,使用简单的多层感知器(MLPs),从由CLIP提取的特征或传统的卷积神经网络(CNNs)开始,可以检测生成的图像。我们还观察到,训练在由稳定扩散生成的图像上的模型可以相对较好地检测由GLIDE生成的图像,但反之则不然。最后,我们发现将相关的文本信息与图像结合在一起很少能够显著提高检测结果,但是图像中所描绘主题的类型却可以对性能产生重大影响。这项工作为检测生成的图像的可行性提供了见解,并对现实应用中的安全和隐私问题产生了影响。可在以下网址中找到复制我们结果的代码:https://github.com/davide-coccomini/Detecting-Images-Generated-by-Diffusers