Text-to-image generation models that generate images based on prompt descriptions have attracted an increasing amount of attention during the past few months. Despite their encouraging performance, these models raise concerns about the misuse of their generated fake images. To tackle this problem, we pioneer a systematic study on the detection and attribution of fake images generated by text-to-image generation models. Concretely, we first build a machine learning classifier to detect the fake images generated by various text-to-image generation models. We then attribute these fake images to their source models, such that model owners can be held responsible for their models' misuse. We further investigate how prompts that generate fake images affect detection and attribution. We conduct extensive experiments on four popular text-to-image generation models, including DALL$\cdot$E 2, Stable Diffusion, GLIDE, and Latent Diffusion, and two benchmark prompt-image datasets. Empirical results show that (1) fake images generated by various models can be distinguished from real ones, as there exists a common artifact shared by fake images from different models; (2) fake images can be effectively attributed to their source models, as different models leave unique fingerprints in their generated images; (3) prompts with the ``person'' topic or a length between 25 and 75 enable models to generate fake images with higher authenticity. All findings contribute to the community's insight into the threats caused by text-to-image generation models. We appeal to the community's consideration of the counterpart solutions, like ours, against the rapidly-evolving fake image generation.
翻译:以快速描述生成图像的文本到图像生成模型在过去几个月中引起了越来越多的关注。 尽管这些模型的性能令人鼓舞, 但这些模型引起了人们对滥用其生成的假图像的关切。 为了解决这一问题, 我们率先对通过文本到图像生成模型产生的假图像的检测和归属进行系统研究。 具体地说, 我们首先建立一个机器学习分类器, 以检测各种文本到图像生成模型产生的假图像。 然后我们将这些假图像归到其来源模型, 这样模型所有人可以对其模型的滥用负责。 我们进一步调查产生假图像的及时性会影响检测和归属。 我们对于四种流行的文本到图像生成模型模型进行广泛的实验, 包括DALL$\cdott$E 2, Stal Diflation, GLIDE, 和Lentt Difult, 以及两个基准快速生成的快速图像数据集。 精华结果显示:(1) 各种模型生成的假图像可以与真实性图像区分开来, 因为有来自不同模型的伪造图像共享的普通手工艺品; (2) 假图像可以有效地归因于其来源模型和25年的高级图像, 导致所有原始图像的原始模型之间的原始模型。