Sequence-to-Sequence (S2S) neural text generation models, especially the pre-trained ones (e.g., BART and T5), have exhibited compelling performance on various natural language generation tasks. However, the black-box nature of these models limits their application in tasks where specific rules (e.g., controllable constraints, prior knowledge) need to be executed. Previous works either design specific model structure (e.g., Copy Mechanism corresponding to the rule "the generated output should include certain words in the source input") or implement specialized inference algorithm (e.g., Constrained Beam Search) to execute particular rules through the text generation. These methods require careful design case-by-case and are difficult to support multiple rules concurrently. In this paper, we propose a novel module named Neural Rule-Execution Tracking Machine that can be equipped into various transformer-based generators to leverage multiple rules simultaneously to guide the neural generation model for superior generation performance in a unified and scalable way. Extensive experimental results on several benchmarks verify the effectiveness of our proposed model in both controllable and general text generation.
翻译:序列到序列( S2S) 神经文本生成模型,特别是预先培训的模型(如BART 和 T5),在各种自然语言生成任务中表现出了令人信服的性能;然而,这些模型的黑箱性质限制了其在需要执行具体规则的任务中的应用(例如,可控制的限制、先前的知识)。以前的工作要么是设计具体模型结构(例如,与规则相对应的复制机制,“生成的产出应包括源输入中的某些词”),要么是执行专门的推论算法(例如,经培训的BAAM搜索),通过文本生成执行特定规则。这些方法需要审慎地逐案设计,难以同时支持多项规则。在本文中,我们提议了一个名为神经规则-执行跟踪机器的新模块,可以安装到各种基于变压器的发电机中,同时利用多种规则指导神经生成模型,以统一和可扩展的方式实现高级生成性能。在几个基准上进行广泛的实验结果,以核实我们提议的模型在可控制和一般文本生成中的有效性。