Data-to-text generation focuses on generating fluent natural language responses from structured semantic representations. Such representations are compositional, allowing for the combination of atomic meaning schemata in various ways to express the rich semantics in natural language. Recently, pretrained language models (LMs) have achieved impressive results on data-to-text tasks, though it remains unclear the extent to which these LMs generalize to new semantic representations. In this work, we systematically study the compositional generalization of current state-of-the-art generation models in data-to-text tasks. By simulating structural shifts in the compositional Weather dataset, we show that T5 models fail to generalize to unseen structures. Next, we show that template-based input representations greatly improve the model performance and model scale does not trivially solve the lack of generalization. To further improve the model's performance, we propose an approach based on self-training using finetuned BLEURT for pseudo-response selection. Extensive experiments on the few-shot Weather and multi-domain SGD datasets demonstrate strong gains of our method.
翻译:数据到文字的生成侧重于从结构化的语义表达形式中产生流利的自然语言反应。这种表述形式是构成性的,允许原子含义的公式组合以各种方式表达自然语言中丰富的语义。最近,经过预先训练的语言模型(LMs)在数据到文字的任务中取得了令人印象深刻的成果,尽管这些LMs一般化为新的语义表达形式的程度仍然不清楚。在这项工作中,我们系统地研究数据到文字任务中目前最先进的生成模型的构成性概括性。通过模拟构成性天气数据集的结构变化,我们表明T5模型未能普遍化为不可见的结构。接下来,我们表明基于模板的投入表达方式极大地改进了模型的性能和模型规模并不能微不足道地解决缺乏概括性的问题。为了进一步改进模型的绩效,我们建议了一种基于自我培训的方法,使用经过微调的BLEURT进行伪反应选择。对微调的天气和多度的SGD数据集进行了广泛的实验,显示了我们方法的有力收益。