Natural language generation is one of the most impactful fields in NLP, and recent years have witnessed its evolution brought about by large language models (LLMs). As the key instrument for writing assistance applications, they are generally prone to replicating or extending offensive content provided in the input. In low-resource data regime, they can also lead to repetitive outputs (Holtzman et al., 2019) [1]. Usually, offensive content and repetitions are mitigated with post-hoc methods, including n-gram level blocklists, top-k and nucleus sampling. In this paper, we introduce a combination of exact and non-exact repetition suppression using token and sequence level unlikelihood loss, repetition penalty during training, inference, and post-processing respectively. We further explore multi-level unlikelihood loss to the extent that it endows the model with abilities to avoid generating offensive words and phrases from the beginning. Finally, with comprehensive experiments, we demonstrate that our proposed methods work exceptionally in controlling the repetition and content quality of LLM outputs.
翻译:----
自然语言生成是自然语言处理中最具影响力的领域之一,近年来由大型语言模型(LLM)带来了它的演变。作为编写辅助应用程序的关键工具,它们通常易于复制或扩展输入中提供的冒犯内容。在低资源数据环境中,它们也可能导致重复的输出(Holtzman等,2019)[1]。通常情况下,通过后处理方法,包括n-gram级别的块列表、top-k和核心采样来减少冒犯内容和重复的产生。本文提出了结合token和序列级别非精确重复抑制和非重复抑制使用的多方面重复抑制方法,实现了训练,推断和后处理中的惩罚。我们进一步探究了多层面非可能性损失的影响,从而让模型具有从一开始避免生成冒犯性词汇和短语的能力。最后,通过综合实验证明了我们提出的方法在控制LLM输出的重复和内容质量方面表现出色。