Humans can perform unseen tasks by recalling relevant skills that are acquired previously and then generalizing them to the target tasks, even if there is no supervision at all. In this paper, we aim to improve such cross-task generalization ability of massive multi-task language models such as T0 (Sanh et al., 2021) in an unsupervised setting. We propose a retrieval-augmentation method named ReCross that takes a few unlabelled examples as queries to retrieve a small subset of upstream data and uses them to update the multi-task model for better generalization. Our empirical results show that the proposed ReCross consistently outperforms non-retrieval baselines by a significant margin.
翻译:人类可以通过回顾先前获得的相关技能来完成不可见的任务,然后将其推广到目标任务中,即使根本没有监督。 在本文中,我们的目标是在无人监督的环境中提高大型多任务语言模型(如T0, (Sanh等人, 2021年))的跨任务概括能力。我们提议了一个名为ReCross的检索增强方法,该方法将几个未贴标签的例子作为检索少量上游数据的查询,并利用它们更新多任务模型,以更好地概括化。我们的经验结果表明,拟议的 ReCross 一直以显著的幅度超越非检索基线。