How to generate summaries of different styles without requiring corpora in the target styles, or training separate models? We present two novel methods that can be deployed during summary decoding on any pre-trained Transformer-based summarization model. (1) Decoder state adjustment instantly modifies decoder final states with externally trained style scorers, to iteratively refine the output against a target style. (2) Word unit prediction constrains the word usage to impose strong lexical control during generation. In experiments of summarizing with simplicity control, automatic evaluation and human judges both find our models producing outputs in simpler languages while still informative. We also generate news headlines with various ideological leanings, which can be distinguished by humans with a reasonable probability.
翻译:如何生成不同风格的概要,而不需要目标样式中的公司,或培训不同的模型?我们介绍了两种新颖的方法,可以在任何经过训练的以变异器为基础的简单拼图模型的简要解码过程中使用。 (1) 解码器国家调整立即用外部训练的样式计分器修改解码器最终状态,以便根据目标样式迭接地改进输出。 (2) 单词单位预测限制用词来强制在一代人中实行强有力的词汇控制。在简单控制、自动评估和人类法官的总结实验中,都发现我们的模型以较简单的语言生成产出,同时仍然提供信息。我们还制作了带有各种意识形态倾向的首页,这些标题可以由具有合理概率的人加以区分。