Growing applications of generative models have led to new threats such as malicious personation and digital copyright infringement. One solution to these threats is model attribution, i.e., the identification of user-end models where the contents under question are generated from. Existing studies showed empirical feasibility of attribution through a centralized classifier trained on all user-end models. However, this approach is not scalable in reality as the number of models ever grows. Neither does it provide an attributability guarantee. To this end, this paper studies decentralized attribution, which relies on binary classifiers associated with each user-end model. Each binary classifier is parameterized by a user-specific key and distinguishes its associated model distribution from the authentic data distribution. We develop sufficient conditions of the keys that guarantee an attributability lower bound. Our method is validated on MNIST, CelebA, and FFHQ datasets. We also examine the trade-off between generation quality and robustness of attribution against adversarial post-processes.
翻译:基因模型应用的增长导致了新的威胁,如恶意人格和侵犯数字版权等。这些威胁的一个解决办法是模型归属,即确定产生所涉内容的用户端模型。现有研究表明,通过对所有用户端模型进行集中分类培训的中央分类员确定归属的实际可行性。然而,随着模型数量的不断增长,这一方法在现实中是无法伸缩的。它也没有提供可归属性保障。为此,本文件研究分散的归属,它依赖与每个用户端模型相关的二进制分类员。每个二进制分类员都用用户专用的钥匙进行参数化,并将其相关的模型分布与真实的数据分布区分开来。我们为钥匙制定了足够的条件,保证可归属性较低的约束性。我们的方法在MDIST、CelibA和FFHQ数据集上得到验证。我们还检查了生成质量和可靠性与对抗性后流程之间的权衡。