There has been great progress in unifying various table-to-text tasks using a single encoder-decoder model trained via multi-task learning (Xie et al., 2022). However, existing methods typically encode task information with a simple dataset name as a prefix to the encoder. This not only limits the effectiveness of multi-task learning, but also hinders the model's ability to generalize to new domains or tasks that were not seen during training, which is crucial for real-world applications. In this paper, we propose compositional task configurations, a set of prompts prepended to the encoder to improve cross-task generalization of unified models. We design the task configurations to explicitly specify the task type, as well as its input and output types. We show that this not only allows the model to better learn shared knowledge across different tasks at training, but also allows us to control the model by composing new configurations that apply novel input-output combinations in a zero-shot manner. We demonstrate via experiments over ten table-to-text tasks that our method outperforms the UnifiedSKG baseline by noticeable margins in both in-domain and zero-shot settings, with average improvements of +0.5 and +12.6 from using a T5-large backbone, respectively.
翻译:使用通过多任务学习(Xie et al., 2022)培训的单一编码器- 解码器模型,在统一各种表格到文本任务方面取得了巨大进展。然而,现有方法通常以简单的数据集名称编码任务信息,作为编码器的前缀。这不仅限制了多任务学习的有效性,而且阻碍了模型推广到培训期间未见的新领域或任务的能力,这对于现实世界应用至关重要。在本文件中,我们提出组成任务配置,一套预设到编码器的提示,以改善统一模型的交叉任务概观。我们设计任务配置,明确指定任务类型及其输入和输出类型。我们表明,这不仅允许模型更好地学习不同培训任务之间的共享知识,而且还使我们能够控制模型,方法是以零光速方式对应用新的投入-输出组合的新配置进行新的配置。我们通过对十项表格到文本任务进行实验,在OVISS+G基础5 的改进中,我们的方法分别比UIS+G基础5和基底级的改进分别比VS+0.5和基底基级,我们通过实验,在OVS+G基准中,分别将US+G基准比差和基底。