Denoising Diffusion Probabilistic Model (DDPM) is able to make flexible conditional image generation from prior noise to real data, by introducing an independent noise-aware classifier to provide conditional gradient guidance at each time step of denoising process. However, due to the ability of classifier to easily discriminate an incompletely generated image only with high-level structure, the gradient, which is a kind of class information guidance, tends to vanish early, leading to the collapse from conditional generation process into the unconditional process. To address this problem, we propose two simple but effective approaches from two perspectives. For sampling procedure, we introduce the entropy of predicted distribution as the measure of guidance vanishing level and propose an entropy-aware scaling method to adaptively recover the conditional semantic guidance. % for each generated sample. For training stage, we propose the entropy-aware optimization objectives to alleviate the overconfident prediction for noisy data.On ImageNet1000 256x256, with our proposed sampling scheme and trained classifier, the pretrained conditional and unconditional DDPM model can achieve 10.89% (4.59 to 4.09) and 43.5% (12 to 6.78) FID improvement respectively.
翻译:(dDPM)能够从先前的噪音到真实数据,通过引入一个独立的噪音-觉分级器,在拆除过程的每个阶段提供有条件的梯度指导,从先前的噪音到真实数据,灵活地生成有条件的图像。然而,由于分类器能够很容易地将不完全生成的图像与高层次结构区分开来,梯度(这是一种阶级信息指导)往往会过早消失,导致从有条件的生成过程向无条件过程的崩溃。为了解决这一问题,我们从两个角度提出了两种简单而有效的方法。对于取样程序,我们引入了预测分布的酶,作为指南消失水平的测量标准,并提议了一种恒度-觉分级测量方法,以适应性地恢复每个生成样本的有条件的语义性指导。在培训阶段,我们提议了英特罗普-觉优化目标,以缓解对扰动数据过于自信的预测。关于图像Net1000,256x256,我们提议的采样计划和培训的分类,我们提出了两种简单和无条件的DDPM模型可以分别实现10.89%(4.59至4.09)和43.5.6%至6.78)的改进。