The privacy implications of generative adversarial networks (GANs) are a topic of great interest, leading to several recent algorithms for training GANs with privacy guarantees. By drawing connections to the generalization properties of GANs, we prove that under some assumptions, GAN-generated samples inherently satisfy some (weak) privacy guarantees. First, we show that if a GAN is trained on m samples and used to generate n samples, the generated samples are (epsilon, delta)-differentially-private for (epsilon, delta) pairs where delta scales as O(n/m). We show that under some special conditions, this upper bound is tight. Next, we study the robustness of GAN-generated samples to membership inference attacks. We model membership inference as a hypothesis test in which the adversary must determine whether a given sample was drawn from the training dataset or from the underlying data distribution. We show that this adversary can achieve an area under the ROC curve that scales no better than O(m^{-1/4}).
翻译:基因对抗网络(GANs)的隐私影响是一个引起极大兴趣的议题,导致最近为GANs提供隐私保障培训的若干算法。我们通过将GANs的样本与GANs的一般特性联系起来,证明根据某些假设,GAN产生的样本本身就满足了某些(弱)隐私保障。首先,我们表明,如果GAN在样本方面受过培训并用于生成N样本,所产生的样本(epsilon, delta)是(epsilon, delta)对(epsilon, delta)对(eta)对(ecc)对o(n/m)进行区分的。我们表明,在某些特殊条件下,这种上限是紧的。我们接下来研究GAN产生的样本的稳健性与成员猜想攻击的关系。我们将成员资格作为假设试验模型,以判断某一样本是从培训数据集还是从基本数据分布中提取的。我们证明,这个对手能够达到ROC曲线下比O(m ⁇ -1/4}更好的区域。