Score-based generative models are shown to achieve remarkable empirical performances in various applications such as image generation and audio synthesis. However, a theoretical understanding of score-based diffusion models is still incomplete. Recently, Song et al. showed that the training objective of score-based generative models is equivalent to minimizing the Kullback-Leibler divergence of the generated distribution from the data distribution. In this work, we show that score-based models also minimize the Wasserstein distance between them under suitable assumptions on the model. Specifically, we prove that the Wasserstein distance is upper bounded by the square root of the objective function up to multiplicative constants and a fixed constant offset. Our proof is based on a novel application of the theory of optimal transport, which can be of independent interest to the society. Our numerical experiments support our findings. By analyzing our upper bounds, we provide a few techniques to obtain tighter upper bounds.
翻译:基于分数的基因化模型显示,在图像生成和音频合成等各种应用中取得了显著的经验性表现。然而,对基于分数的传播模型的理论理解仍然不完整。最近,Song等人指出,基于分数的基因化模型的培训目标相当于将所产生分布与数据分布的库尔贝克-利伯尔差异最小化。在这项工作中,我们显示,基于分数的模型也根据模型的适当假设,将它们之间的瓦西斯坦距离最小化。具体地说,我们证明,瓦西尔斯坦距离被目标函数的正方根以乘数常数和固定常数为顶部。我们的证据是基于对最佳运输理论的新应用,这对社会可能具有独立的兴趣。我们的数字实验支持了我们的调查结果。通过分析我们的上界,我们提供了一些技术,以获得更紧密的上限。