Generative models are now capable of synthesizing images, speeches, and videos that are hardly distinguishable from authentic contents. Such capabilities cause concerns such as malicious impersonation and IP theft. This paper investigates a solution for model attribution, i.e., the classification of synthetic contents by their source models via watermarks embedded in the contents. Building on past success of model attribution in the image domain, we discuss algorithmic improvements for generating user-end speech models that empirically achieve high attribution accuracy, while maintaining high generation quality. We show the trade off between attributability and generation quality under a variety of attacks on generated speech signals attempting to remove the watermarks, and the feasibility of learning robust watermarks against these attacks.
翻译:生成模型现在能够综合难以与真实内容区分的图像、演讲和视频,这种能力引起了恶意冒名顶替和IP盗窃等关切。本文探讨了模型归属的解决方案,即通过内容中嵌入的水印,按来源模型对合成内容进行分类。在图像领域模型归属的成功基础上,我们讨论了逻辑改进,以生成用户终端语音模型,在经验上达到高属性准确度,同时保持高生成质量。我们展示了在对试图去除水印的生成语音信号进行各种攻击的情况下,可归属性和生成质量之间的平衡,以及针对这些袭击学习强力水印的可行性。