Diffusion models are a class of generative models that have demonstrated remarkable success in tasks such as image generation. However, one of the bottlenecks of these models is slow sampling due to the delay before the onset of trajectory bifurcation, at which point substantial reconstruction begins. This issue degrades generation quality, especially in the early stages. Our primary objective is to mitigate bifurcation-related issues by preprocessing the training data to enhance reconstruction quality, particularly for small-scale network architectures. Specifically, we propose applying Gaussianization preprocessing to the training data to make the target distribution more closely resemble an independent Gaussian distribution, which serves as the initial density of the reconstruction process. This preprocessing step simplifies the model's task of learning the target distribution, thereby improving generation quality even in the early stages of reconstruction with small networks. The proposed method is, in principle, applicable to a broad range of generative tasks, enabling more stable and efficient sampling processes.
翻译:扩散模型是一类生成模型,在图像生成等任务中展现出卓越性能。然而,这些模型的一个瓶颈在于采样速度缓慢,这源于轨迹分岔发生前的延迟阶段——实质性重构在该点才开始。这一问题尤其在早期阶段会降低生成质量。本研究的主要目标是通过对训练数据进行预处理来缓解分岔相关问题,从而提升重构质量,特别是在小规模网络架构中。具体而言,我们提出对训练数据施加高斯化预处理,使目标分布更接近独立高斯分布——该分布正是重构过程的初始密度。此预处理步骤简化了模型学习目标分布的任务,从而即使在小型网络进行早期重构时也能提升生成质量。所提方法原则上适用于广泛的生成任务,能够实现更稳定高效的采样过程。