Multitask learning is widely used in practice to train a low-resource target task by augmenting it with multiple related source tasks. Yet, naively combining all the source tasks with a target task does not always improve the prediction performance for the target task due to negative transfers. Thus, a critical problem in multitask learning is identifying subsets of source tasks that would benefit the target task. This problem is computationally challenging since the number of subsets grows exponentially with the number of source tasks; efficient heuristics for subset selection does not always capture the relationship between task subsets and multitask learning performances. In this paper, we introduce an efficient procedure to address this problem via surrogate modeling. In surrogate modeling, we sample (random) subsets of source tasks and precompute their multitask learning performances; Then, we approximate the precomputed performances with a linear regression model that can also be used to predict the multitask performance of unseen task subsets. We show theoretically and empirically that fitting this model only requires sampling linearly many subsets in the number of source tasks. The fitted model provides a relevance score between each source task and the target task; We use the relevance scores to perform subset selection for multitask learning by thresholding. Through extensive experiments, we show that our approach predicts negative transfers from multiple source tasks to target tasks much more accurately than existing task affinity measures. Additionally, we demonstrate that for five weak supervision datasets, our approach consistently improves upon existing optimization methods for multi-task learning.
翻译:多任务学习被广泛应用于使用多个相关源任务增强低资源目标任务的训练。然而,将所有源任务与目标任务简单组合并不总是改善目标任务的预测性能,由于存在负面转移。因此,多任务学习中的一个关键问题是识别能够有益于目标任务的源任务子集。该问题在计算上具有挑战性,因为源任务子集的数量随源任务数呈指数增长;有效的子集选择启发式方法并不总是能够捕捉到任务子集和多任务学习性能之间的关系。在本文中,我们引入了一种通过代理建模来解决该问题的高效过程。在代理建模中,我们对源任务进行(随机)子集采样并预先计算它们的多任务学习性能。然后,我们使用线性回归模型来近似预先计算的性能,该模型还可用于对未见过的任务子集进行多任务性能预测。我们在理论和实践中表明,只需要在源任务数量上线性少量采样即可实现模型的适配。适配的模型提供了每个源任务和目标任务之间的相关性分数;我们使用这些相关性分数通过阈值进行多任务学习的子集选择。通过广泛的实验,我们表明,相比现有的任务相似度度量方法,我们的方法更准确地预测了多个源任务到目标任务的负面转移。此外,我们证明,在五个弱监督数据集上,我们的方法始终优于现有的多任务学习优化方法。