Recent advances in large pre-trained language models have demonstrated strong results in generating natural languages and significantly improved performances for many natural language generation (NLG) applications such as machine translation and text summarization. However, when the generation tasks are more open-ended and the content is under-specified, existing techniques struggle to generate long-term coherent and creative content. Moreover, the models exhibit and even amplify social biases that are learned from the training corpora. This happens because the generation models are trained to capture the surface patterns (i.e. sequences of words), instead of capturing underlying semantics and discourse structures, as well as background knowledge including social norms. In this paper, I introduce our recent works on controllable text generation to enhance the creativity and fairness of language generation models. We explore hierarchical generation and constrained decoding, with applications to creative language generation including story, poetry, and figurative languages, and bias mitigation for generation models.
翻译:在经过培训的大型语言模型方面最近取得的进展表明,在创造自然语言方面取得了巨大成果,并大大改善了许多自然语言生成应用的性能,例如机器翻译和文本汇总;然而,当生成任务更加开放且内容描述不足时,现有技术就难以产生长期一致和创造性的内容;此外,模型展示甚至扩大了从培训公司学到的社会偏见;这是因为生成模型经过培训,以捕捉表面模式(即文字序列),而不是捕捉基本的语义和话语结构以及背景知识,包括社会规范;在本文中,我介绍我们最近关于可控文本生成的工作,以加强语言生成模式的创造力和公正性;我们探索等级生成和受限制的解码,应用创意语言生成,包括故事、诗歌和可比较语言,以及代模式的偏见缓解。