Recently, diffusion models have emerged as a new paradigm for generative models. Despite the success in domains using continuous signals such as vision and audio, adapting diffusion models to natural language is difficult due to the discrete nature of text. We tackle this challenge by proposing DiffuSeq: a diffusion model designed for sequence-to-sequence (Seq2Seq) text generation tasks. Upon extensive evaluation over a wide range of Seq2Seq tasks, we find DiffuSeq achieving comparable or even better performance than six established baselines, including a state-of-the-art model that is based on pre-trained language models. Apart from quality, an intriguing property of DiffuSeq is its high diversity during generation, which is desired in many Seq2Seq tasks. We further include a theoretical analysis revealing the connection between DiffuSeq and autoregressive/non-autoregressive models. Bringing together theoretical analysis and empirical evidence, we demonstrate the great potential of diffusion models in complex conditional language generation tasks.
翻译:最近,传播模型已成为基因模型的新范例。尽管在使用视觉和音频等连续信号的领域取得了成功,但由于文本的离散性质,很难将传播模型与自然语言相适应。我们通过提出DiffuSeq(Seq2Seq)的传播模型:一个为序列到序列(Seq2Seq)生成文本的任务设计的传播模型来应对这一挑战。经过对一系列Seq2Seq任务的广泛评价,我们发现DiffuSeq(DiffuSeq)取得了比六个既定基线的可比较甚至更好的业绩,包括一个以预先培训的语言模型为基础的最先进的模型。除了质量外,DiffuSeq(DiffuSeq)令人感兴趣的特性是其代代代期的高度多样性,这是许多Seq2Seq任务所期望的。我们还包括一项理论分析,揭示了Diffuseq(Seq)与自动递增/非递增型模型之间的联系。我们汇集了理论分析和经验证据,我们展示了在复杂的有条件语言生成任务中的传播模型的巨大潜力。