Humans can perform unseen tasks by recalling relevant skills acquired previously and then generalizing them to the target tasks, even if there is no supervision at all. In this paper, we aim to improve this kind of cross-task generalization ability of massive multi-task language models, such as T0 and FLAN, in an unsupervised setting. We propose a retrieval-augmentation method named ReCross that takes a few unlabelled examples as queries to retrieve a small subset of upstream data and uses them to update the multi-task model for better generalization. ReCross is a straightforward yet effective retrieval method that combines both efficient dense retrieval and effective pair-wise reranking. Our results and analysis show that it significantly outperforms both non-retrieval methods and other baseline methods.
翻译:人类可以通过回顾先前获得的相关技能来完成不可见的任务,然后将其推广到目标任务中,即使根本没有监督。 在本文件中,我们的目标是在无人监督的环境下,提高大型多任务语言模型(如T0和FLAN)的跨任务概括能力。我们提议了一个名为ReCross的检索增强方法,该方法将几个未贴标签的例子作为检索少量上游数据的查询,并利用它们更新多任务模型,以更好地概括化。 ReCross是一种简单而有效的检索方法,既能高效密集检索,又能有效对对对齐重新排序。我们的结果和分析显示,它大大优于非检索方法和其他基线方法。