Large-scale pre-trained language models have achieved great success on natural language generation tasks. However, it is difficult to control the pre-trained language models to generate sentences with the desired attribute such as topic and sentiment, etc. Recently, Bayesian Controllable Language Models (BCLMs) have been shown to be efficient in controllable language generation. Rather than fine-tuning the parameters of pre-trained language models, BCLMs use external discriminators to guide the generation of pre-trained language models. However, the mismatch between training and inference of BCLMs limits the performance of the models. To address the problem, in this work we propose a "Gemini Discriminator" for controllable language generation which alleviates the mismatch problem with a small computational cost. We tested our method on two controllable language generation tasks: sentiment control and topic control. On both tasks, our method reached achieved new state-of-the-art results in automatic and human evaluations.
翻译:大规模经过培训的语文模式在自然语言生成任务方面取得了巨大成功,然而,很难控制经过培训的语文模式,以产生具有理想属性的句子,如主题和情绪等。最近,巴伊西亚可控语言模式(BCLMs)在可控语言生成方面证明是有效的。BCLMs没有微调经过培训的语言模式的参数,而是利用外部歧视来指导经过培训的语言模式的生成。然而,BCLMs的培训和推论之间的不匹配限制了模型的性能。为了解决这个问题,我们在这项工作中提议为可控语言生成建立一个“Gémini Discriminator”, 以小计算成本缓解不匹配问题。我们测试了我们两种可控语言生成任务的方法:情绪控制和主题控制。在这两项任务中,我们的方法都达到了在自动和人文评估方面的新的最新结果。