State-of-the-art text-to-image models generate photorealistic images at an unprecedented speed. This work focuses on models that operate in a bitwise autoregressive manner over a discrete set of tokens that is practically infinite in size. However, their impressive generative power comes with a growing risk: as their outputs increasingly populate the Internet, they are likely to be scraped and reused as training data-potentially by the very same models. This phenomenon has been shown to lead to model collapse, where repeated training on generated content, especially from the models' own previous versions, causes a gradual degradation in performance. A promising mitigation strategy is watermarking, which embeds human-imperceptible yet detectable signals into generated images-enabling the identification of generated content. In this work, we introduce BitMark, a robust bitwise watermarking framework. Our method embeds a watermark directly at the bit level of the token stream during the image generation process. Our bitwise watermark subtly influences the bits to preserve visual fidelity and generation speed while remaining robust against a spectrum of removal techniques. Furthermore, it exhibits high radioactivity, i.e., when watermarked generated images are used to train another image generative model, this second model's outputs will also carry the watermark. The radioactive traces remain detectable even when only fine-tuning diffusion or image autoregressive models on images watermarked with our BitMark. Overall, our approach provides a principled step toward preventing model collapse in image generative models by enabling reliable detection of generated outputs. The code is available at https://github.com/sprintml/BitMark.
翻译:当前最先进的文本到图像模型以前所未有的速度生成逼真的图像。本研究聚焦于在规模近乎无限的离散标记集上以比特自回归方式操作的模型。然而,其强大的生成能力伴随着日益增长的风险:随着这些模型的输出在互联网上大量传播,它们很可能被爬取并用作训练数据——甚至可能被同类型模型自身重复利用。已有研究表明,这种现象会导致模型崩溃,即反复使用生成内容(尤其是模型自身先前版本生成的内容)进行训练,将引发性能的渐进性退化。一种有效的缓解策略是水印技术,它能在生成图像中嵌入人眼难以察觉但可检测的信号,从而实现对生成内容的识别。本文提出BitMark,一种鲁棒的比特级水印框架。该方法在图像生成过程中,直接在标记流的比特层级嵌入水印。我们的比特水印通过微调比特位来保持视觉保真度和生成速度,同时能有效抵抗多种去除技术。此外,该水印具有高放射性特征,即当带有水印的生成图像被用于训练其他图像生成模型时,第二代模型的输出也会携带该水印。即使仅对扩散模型或图像自回归模型使用经BitMark水印处理的图像进行微调,放射性痕迹仍可被检测。总体而言,本研究通过实现生成输出的可靠检测,为预防图像生成模型的崩溃提供了原理性解决方案。代码已开源:https://github.com/sprintml/BitMark。