Generative Adversarial Networks (GANs) have achieved great success in data generation. However, its statistical properties are not fully understood. In this paper, we consider the statistical behavior of the general $f$-divergence formulation of GAN, which includes the Kullback--Leibler divergence that is closely related to the maximum likelihood principle. We show that for parametric generative models that are correctly specified, all $f$-divergence GANs with the same discriminator classes are asymptotically equivalent under suitable regularity conditions. Moreover, with an appropriately chosen local discriminator, they become equivalent to the maximum likelihood estimate asymptotically. For generative models that are misspecified, GANs with different $f$-divergences {converge to different estimators}, and thus cannot be directly compared. However, it is shown that for some commonly used $f$-divergences, the original $f$-GAN is not optimal in that one can achieve a smaller asymptotic variance when the discriminator training in the original $f$-GAN formulation is replaced by logistic regression. The resulting estimation method is referred to as Adversarial Gradient Estimation (AGE). Empirical studies are provided to support the theory and to demonstrate the advantage of AGE over the original $f$-GANs under model misspecification.
翻译:在数据生成方面取得巨大成功,然而,其统计特性并没有得到完全理解。在本文件中,我们考虑了GAN一般美元波动配方的统计行为,其中包括与最大可能性原则密切相关的Kullback-Leiber差异,其中包括与最大可能性原则密切相关的Kullback-Leiber差异。我们表明,对于正确指定的参数变异模型,所有具有相同歧视等级的美元波动配方的GAN在适当的正常条件下都与美元波动等值。此外,如果当地选择适当,它们就相当于最大可能性的随机估计。对于基因化模型而言,GAN具有不同的美元差异,因此无法直接比较。然而,我们表明,对于一些通常使用美元差异等级的参数,原始美元-GAN并不理想,在最初使用美元模型时,当最初的模型变异性分析理论被取代了最初的GARC模型的模型分析,而原始的AGAFAFA模型则取代了原始的AGAFA的模型。