Transfer learning approaches have shown to significantly improve performance on downstream tasks. However, it is common for prior works to only report where transfer learning was beneficial, ignoring the significant trial-and-error required to find effective settings for transfer. Indeed, not all task combinations lead to performance benefits, and brute-force searching rapidly becomes computationally infeasible. Hence the question arises, can we predict whether transfer between two tasks will be beneficial without actually performing the experiment? In this paper, we leverage explainability techniques to effectively predict whether task pairs will be complementary, through comparison of neural network activation between single-task models. In this way, we can avoid grid-searches over all task and hyperparameter combinations, dramatically reducing the time needed to find effective task pairs. Our results show that, through this approach, it is possible to reduce training time by up to 83.5% at a cost of only 0.034 reduction in positive-class F1 on the TREC-IS 2020-A dataset.
翻译:转移学习方法显示,转移学习方法明显改善了下游任务的业绩。然而,以往工作通常只报告有利于转移学习的情况,而忽略了找到有效转让环境所需的大量试验和压力。 事实上,并非所有任务组合都带来绩效效益,而粗力搜索在计算上都变得不可行。 因此,问题出现,我们能否预测两个任务之间的转移是否有益而不实际进行实验?在本文中,我们利用解释技术,通过比较单任务模型之间的神经网络激活,有效预测任务对口是否互补。 这样,我们可以避免在所有任务和超光谱组合上进行网路搜索,大大缩短找到有效任务对口所需的时间。 我们的结果显示,通过这种方法,有可能将培训时间减少83.5%,成本是TREC-IS 2020-A数据集只减少0.034个正级F1。