Textual content is often the output of a collaborative writing process: We start with an initial draft, ask for suggestions, and repeatedly make changes. Agnostic of this process, today's language models are trained to generate only the final result. As a consequence, they lack several abilities crucial for collaborative writing: They are unable to update existing texts, difficult to control and incapable of verbally planning or explaining their actions. To address these shortcomings, we introduce PEER, a collaborative language model that is trained to imitate the entire writing process itself: PEER can write drafts, add suggestions, propose edits and provide explanations for its actions. Crucially, we train multiple instances of PEER able to infill various parts of the writing process, enabling the use of self-training techniques for increasing the quality, amount and diversity of training data. This unlocks PEER's full potential by making it applicable in domains for which no edit histories are available and improving its ability to follow instructions, to write useful comments, and to explain its actions. We show that PEER achieves strong performance across various domains and editing tasks.
翻译:文本内容往往是合作写作过程的输出: 我们从最初的草稿开始, 征求建议, 并反复修改。 这个过程的精密性, 今天的语言模型经过培训, 只产生最后的结果。 因此, 它们缺乏对合作写作至关重要的几种能力: 它们无法更新现有文本, 难以控制, 无法口头规划或解释它们的行动。 为了解决这些缺陷, 我们引入了PEER, 这是一种合作语言模型, 受过训练, 能够模仿整个写作过程本身: PEER 可以写草稿, 添加建议, 提出编辑建议, 并为它的行动提供解释。 关键是, 我们培训多例PEER, 能够填充写过程的各个部分, 从而能够使用自我培训技术来提高培训数据的质量、 数量和多样性。 这释放了PEER的全部潜力, 因为它可以在没有编辑历史的领域应用, 并提高其遵循指示、 写有用评论和解释其行动的能力。 我们表明, PEER 能够在各个领域和编辑任务中取得强有力的业绩。