Text-editing models have recently become a prominent alternative to seq2seq models for monolingual text-generation tasks such as grammatical error correction, simplification, and style transfer. These tasks share a common trait - they exhibit a large amount of textual overlap between the source and target texts. Text-editing models take advantage of this observation and learn to generate the output by predicting edit operations applied to the source sequence. In contrast, seq2seq models generate outputs word-by-word from scratch thus making them slow at inference time. Text-editing models provide several benefits over seq2seq models including faster inference speed, higher sample efficiency, and better control and interpretability of the outputs. This tutorial provides a comprehensive overview of text-editing models and current state-of-the-art approaches, and analyzes their pros and cons. We discuss challenges related to productionization and how these models can be used to mitigate hallucination and bias, both pressing challenges in the field of text generation.
翻译:文本编辑模型最近已成为单语文本生成任务(如语法错误校正、简化和样式转换)的后代2当量模型的突出替代物。这些任务具有一个共同的特点:它们显示了源和目标文本之间的大量文字重叠。文本编辑模型利用这一观察并学习通过预测源序列应用的编辑操作产生输出。相形之下,后代2当量模型从零开始产生逐字产出,从而在推论时速度缓慢。文本编辑模型为后代2当量模型提供了若干好处,包括更快的推断速度、更高的样本效率以及产出的更好控制和可解释性。该教程提供了文本编辑模型和当前最新方法的全面概览,并分析了其正反方法。我们讨论了与制作有关的挑战以及如何利用这些模型减轻幻觉和偏见,这都是在文本生成领域面临的紧迫挑战。