Neural text generation has made tremendous progress in various tasks. One common characteristic of most of the tasks is that the texts are not restricted to some rigid formats when generating. However, we may confront some special text paradigms such as Lyrics (assume the music score is given), Sonnet, SongCi (classical Chinese poetry of the Song dynasty), etc. The typical characteristics of these texts are in three folds: (1) They must comply fully with the rigid predefined formats. (2) They must obey some rhyming schemes. (3) Although they are restricted to some formats, the sentence integrity must be guaranteed. To the best of our knowledge, text generation based on the predefined rigid formats has not been well investigated. Therefore, we propose a simple and elegant framework named SongNet to tackle this problem. The backbone of the framework is a Transformer-based auto-regressive language model. Sets of symbols are tailor-designed to improve the modeling performance especially on format, rhyme, and sentence integrity. We improve the attention mechanism to impel the model to capture some future information on the format. A pre-training and fine-tuning framework is designed to further improve the generation quality. Extensive experiments conducted on two collected corpora demonstrate that our proposed framework generates significantly better results in terms of both automatic metrics and the human evaluation.
翻译:大多数任务的共同特点是,案文在生成时不局限于某些僵硬格式。然而,我们可能会遇到一些特殊的文本范式,如Lyrics(标出音乐评分)、Sonnet、SongCi(宋朝经典中国诗歌)等。这些文本的典型特征分为三个部分:(1)它们必须完全符合僵硬的预定义格式。(2)它们必须遵从一些节奏计划。(3)尽管它们限于某些格式,但必须保证句子的完整性。根据我们的知识,没有很好地调查以预先界定的僵硬格式制作的文本。因此,我们提出了一个简单而优雅的框架,名为SongNet(SongNet)来解决这一问题。框架的骨干是一个基于变异器的自动反动语言模型。这些符号的典型特征是专门设计的,目的是改进模型的性能,特别是在格式、押韵和句子完整性方面。我们改进了对模式的干扰性能机制,以获取格式上的一些未来信息。一个培训和微调框架旨在大大改进我们两个拟议格式的质量。