Models for learning probability distributions such as generative models and density estimators behave quite differently from models for learning functions. One example is found in the memorization phenomenon, namely the ultimate convergence to the empirical distribution, that occurs in generative adversarial networks (GANs). For this reason, the issue of generalization is more subtle than that for supervised learning. For the bias potential model, we show that dimension-independent generalization accuracy is achievable if early stopping is adopted, despite that in the long term, the model either memorizes the samples or diverges. The generality of our arguments indicates that this slow deterioration from generalization to memorization might be common to distribution-learning models in general.
翻译:学习概率分布模型,如基因模型和密度估计符等,其行为与学习功能模型大不相同。一个例子是记忆化现象,即在基因对抗网络(GANs)中出现的经验分布最终趋同。因此,一般化问题比监督学习要微妙。关于偏差潜在模式,我们表明,如果采用早期停止,那么,在长期内,该模型既可以回忆样本,也可以回忆差异。我们的论点的笼统性表明,从概括化到记忆化的这种缓慢恶化一般可能常见于分布式学习模式。