Progress in generative modelling, especially generative adversarial networks, have made it possible to efficiently synthesize and alter media at scale. Malicious individuals now rely on these machine-generated media, or deepfakes, to manipulate social discourse. In order to ensure media authenticity, existing research is focused on deepfake detection. Yet, the adversarial nature of frameworks used for generative modeling suggests that progress towards detecting deepfakes will enable more realistic deepfake generation. Therefore, it comes at no surprise that developers of generative models are under the scrutiny of stakeholders dealing with misinformation campaigns. At the same time, generative models have a lot of positive applications. As such, there is a clear need to develop tools that ensure the transparent use of generative modeling, while minimizing the harm caused by malicious applications. Our technique optimizes over the source of entropy of each generative model to probabilistically attribute a deepfake to one of the models. We evaluate our method on the seminal example of face synthesis, demonstrating that our approach achieves 97.62% attribution accuracy, and is less sensitive to perturbations and adversarial examples. We discuss the ethical implications of our work, identify where our technique can be used, and highlight that a more meaningful legislative framework is required for a more transparent and ethical use of generative modeling. Finally, we argue that model developers should be capable of claiming plausible deniability and propose a second framework to do so -- this allows a model developer to produce evidence that they did not produce media that they are being accused of having produced.
翻译:基因模型的进展,特别是基因对抗网络,使得能够有效地合成和改变规模的媒体; 恶意的个人现在依靠这些机器产生的媒体或深假来操纵社会言论; 为了确保媒体真实性,现有研究侧重于深假检测; 然而,用于基因模型的框架的对抗性质表明,在发现深假模型特别是基因对抗网络方面的进展将使得能够更现实地产生深刻假象。 因此,毫不奇怪,基因模型的开发者正在接受处理错误信息运动的利益攸关方的监督。 同时,基因模型有许多积极的应用。 因此,显然需要开发工具,确保以透明的方式使用基因模型,同时尽量减少恶意应用造成的伤害。 我们的技术优化了每个基因模型的孵化源,从而将一个更深刻的假象与模型中的某个模型联系起来。 我们评估了我们模型的模型方法,表明我们的方法达到了97.62%的归属准确性,而对于扰动性和对抗性的例子则不那么敏感。 我们讨论的是,一个更具有道德意义的模型,我们使用了一个更透明的基因框架,我们用这个方法来解释一个更具有透明度的,我们用一个更透明的基因框架来解释。