Diffusion models generate samples by reversing a fixed forward diffusion process. Despite already providing impressive empirical results, these diffusion models algorithms can be further improved by reducing the variance of the training targets in their denoising score-matching objective. We argue that the source of such variance lies in the handling of intermediate noise-variance scales, where multiple modes in the data affect the direction of reverse paths. We propose to remedy the problem by incorporating a reference batch which we use to calculate weighted conditional scores as more stable training targets. We show that the procedure indeed helps in the challenging intermediate regime by reducing (the trace of) the covariance of training targets. The new stable targets can be seen as trading bias for reduced variance, where the bias vanishes with increasing reference batch size. Empirically, we show that the new objective improves the image quality, stability, and training speed of various popular diffusion models across datasets with both general ODE and SDE solvers. When used in combination with EDM, our method yields a current SOTA FID of 1.90 with 35 network evaluations on the unconditional CIFAR-10 generation task. The code is available at https://github.com/Newbeeer/stf
翻译:尽管已经提供了令人印象深刻的经验性结果,但这些传播模式的算法可以通过缩小培训目标在分级比对目标上的差异来进一步改进。我们认为,这种差异的根源在于处理中间噪声变化尺度,因为数据中的多种模式会影响反向路径的方向。我们建议采用参考批量来解决这个问题,我们用这些批量来计算加权条件分数,作为更稳定的培训目标。我们表明,该程序确实有助于挑战性的中间制度,减少(追踪)培训目标的共变性。新的稳定目标可被视为减少差异的贸易偏差,因为偏见会随着参考批量的大小的增加而消失。我们很生动地指出,新的目标提高了各种大众传播模型的形象质量、稳定性和培训速度,这些模型跨越了与一般的ODE解码和SDE解码的数据集。当与EDM相结合使用时,我们的方法产生一个1.90的SOTA FID和35个关于无条件的CIFAR-10代任务网络评价。该代码可在http://gistrub/Newberb/NewBee/Newberbs查阅。