Recent work has found that multi-task training with a large number of diverse tasks can uniformly improve downstream performance on unseen target tasks. In contrast, literature on task transferability has established that the choice of intermediate tasks can heavily affect downstream task performance. In this work, we aim to disentangle the effect of scale and relatedness of tasks in multi-task representation learning. We find that, on average, increasing the scale of multi-task learning, in terms of the number of tasks, indeed results in better learned representations than smaller multi-task setups. However, if the target tasks are known ahead of time, then training on a smaller set of related tasks is competitive to the large-scale multi-task training at a reduced computational cost.
翻译:最近的工作发现,具有大量不同任务的多任务培训可以统一地改善在未见目标任务方面的下游业绩,而关于任务可转移性的文献则确定,选择中间任务可以严重影响下游任务业绩,在这项工作中,我们的目标是在多任务代表性学习中分离任务规模和关联性的影响。我们发现,从任务数量来看,平均而言,增加多任务学习的规模,确实比较小的多任务设置更能了解情况。然而,如果提前了解目标任务,那么关于较小型相关任务的培训就具有竞争力,以较低的计算成本进行大规模多任务培训。