Recent text-to-image diffusion models such as MidJourney and Stable Diffusion threaten to displace many in the professional artist community. In particular, models can learn to mimic the artistic style of specific artists after "fine-tuning" on samples of their art. In this paper, we describe the design, implementation and evaluation of Glaze, a tool that enables artists to apply "style cloaks" to their art before sharing online. These cloaks apply barely perceptible perturbations to images, and when used as training data, mislead generative models that try to mimic a specific artist. In coordination with the professional artist community, we deploy user studies to more than 1000 artists, assessing their views of AI art, as well as the efficacy of our tool, its usability and tolerability of perturbations, and robustness across different scenarios and against adaptive countermeasures. Both surveyed artists and empirical CLIP-based scores show that even at low perturbation levels (p=0.05), Glaze is highly successful at disrupting mimicry under normal conditions (>92%) and against adaptive countermeasures (>85%).
翻译:----
最近的文本到图像扩散模型(如MidJourney和Stable Diffusion)威胁到许多专业艺术家社区。特别是,模型可以在练习特定艺术家的艺术样本后,学习模仿其艺术风格。在本文中,我们描述了Glaze的设计、实现和评估,这是一种工具,使艺术家能够在在线分享之前对他们的艺术应用“风格遮罩”。这些遮罩对图像应用几乎不可感知的扰动,当作为训练数据使用时,这些扰动会误导试图模仿特定艺术家的生成模型。与专业艺术家社区协调,我们展开用户调查,共有1000多名艺术家参与,评估他们对AI艺术的看法,以及我们的工具的效能、可用性和容忍度、在不同情境下的鲁棒性以及对自适应对策的抵抗能力。在正常情况下,受调查的艺术家和经验CLIP based分数表明,即使在低扰动水平下(p=0.05),Glaze在干扰模仿方面的成功率也非常高(>92%),并且对自适应对策的干扰成功率也很高(>85%)。