Text-conditioned image generation models have recently achieved astonishing results in image quality and text alignment and are consequently employed in a fast-growing number of applications. Since they are highly data-driven, relying on billion-sized datasets randomly scraped from the internet, they also suffer, as we demonstrate, from degenerated and biased human behavior. In turn, they may even reinforce such biases. To help combat these undesired side effects, we present safe latent diffusion (SLD). Specifically, to measure the inappropriate degeneration due to unfiltered and imbalanced training sets, we establish a novel image generation test bed-inappropriate image prompts (I2P)-containing dedicated, real-world image-to-text prompts covering concepts such as nudity and violence. As our exhaustive empirical evaluation demonstrates, the introduced SLD removes and suppresses inappropriate image parts during the diffusion process, with no additional training required and no adverse effect on overall image quality or text alignment.
翻译:以文字为条件的图像生成模型最近在图像质量和文本协调方面取得了惊人的成果,因此被用于快速增长的应用中。由于这些模型是高度数据驱动的,依靠互联网随机剪切的10亿大小的数据集,它们也如我们所显示的那样,遭受了堕落和有偏见的人类行为。反过来,它们甚至可能强化这种偏见。为了帮助消除这些不理想的副作用,我们展示了安全潜伏的传播。具体地说,为了衡量由于未过滤和不平衡的培训组而导致的不适当的退化,我们建立了一个新型图像生成测试不适床图像提示(I2P),其中含有专门的、真实世界图像到文字提示,涵盖裸露和暴力等概念。正如我们详尽的经验性评估所显示,引入的SLD在传播过程中去除和压制不适当的图像部分,不需要额外的培训,也不对总体图像质量或文本协调产生不利影响。