This paper studies constrained text generation, which is to generate sentences under certain pre-conditions. We focus on CommonGen, the task of generating text based on a set of concepts, as a representative task of constrained text generation. Traditional methods mainly rely on supervised training to maximize the likelihood of target sentences.However, global constraints such as common sense and coverage cannot be incorporated into the likelihood objective of the autoregressive decoding process. In this paper, we consider using reinforcement learning to address the limitation, measuring global constraints including fluency, common sense and concept coverage with a comprehensive score, which serves as the reward for reinforcement learning. Besides, we design a guided decoding method at the word, fragment and sentence levels. Experiments demonstrate that our method significantly increases the concept coverage and outperforms existing models in various automatic evaluations.
翻译:本文研究限制文本生成,即在某些先决条件下生成句子。我们注重共同Gen,即根据一套概念生成文本的任务,这是限制性文本生成的一项具有代表性的任务。传统方法主要依赖监督培训,以最大限度地实现目标判决的可能性。然而,常识和覆盖面等全球制约因素不能纳入自动递减解码过程的可能目标。在本文件中,我们考虑利用强化学习来解决限制问题,衡量全球制约因素,包括流畅、常识和概念覆盖面,全面评分,作为强化学习的奖励。此外,我们还在字、碎片和句级设计了一种有指导的解码方法。实验表明,我们的方法大大增加了概念覆盖面,并超越了各种自动评估中的现有模式。