An energy-based model (EBM) is a popular generative framework that offers both explicit density and architectural flexibility, but training them is difficult since it is often unstable and time-consuming. In recent years, various training techniques have been developed, e.g., better divergence measures or stabilization in MCMC sampling, but there often exists a large gap between EBMs and other generative frameworks like GANs in terms of generation quality. In this paper, we propose a novel and effective framework for improving EBMs via contrastive representation learning (CRL). To be specific, we consider representations learned by contrastive methods as the true underlying latent variable. This contrastive latent variable could guide EBMs to understand the data structure better, so it can improve and accelerate EBM training significantly. To enable the joint training of EBM and CRL, we also design a new class of latent-variable EBMs for learning the joint density of data and the contrastive latent variable. Our experimental results demonstrate that our scheme achieves lower FID scores, compared to prior-art EBM methods (e.g., additionally using variational autoencoders or diffusion techniques), even with significantly faster and more memory-efficient training. We also show conditional and compositional generation abilities of our latent-variable EBMs as their additional benefits, even without explicit conditional training. The code is available at https://github.com/hankook/CLEL.
翻译:以能源为基础的模型(EBM)是一个流行型基因框架,它既提供明确的密度,又提供建筑灵活性,但培训则困难,因为它往往不稳定而且耗时。近年来,已经开发了各种培训技术,例如,在MCMC取样方面,改进了差异度或稳定了措施,但在EBM和CRL的联合培训方面,EBM和GANs等其他基因框架之间往往存在着巨大的差距。在本文件中,我们提出了一个通过对比代表性学习来改进EBM(CRL)的新颖而有效的框架。具体地说,我们认为,通过对比性方法学到的表述是真实的潜在变量。这种对比性潜伏变量可以指导EBM更好地理解数据结构,从而能够大大改进和加速EBM培训。为了能够联合培训EBM和CRL,我们还设计了一个新的潜在可变性EBM(潜在数据)框架,以学习数据的联合密度和对比性潜值变量。我们的实验结果表明,我们的计划比先前的EBM方法(例如,进一步使用变式自动/CEBM(bM),甚至使用可变式的软式自动和可扩展的EBMFBM(E-C)法化)更快速的生成),甚至更能的生成的记忆/潜变式的生成和可变化的生成的生成的生成的生成能力也显示更迅速的生成能力。</s>