Recent multi-task learning research argues against unitary scalarization, where training simply minimizes the sum of the task losses. Several ad-hoc multi-task optimization algorithms have instead been proposed, inspired by various hypotheses about what makes multi-task settings difficult. The majority of these optimizers require per-task gradients, and introduce significant memory, runtime, and implementation overhead. We show that unitary scalarization, coupled with standard regularization and stabilization techniques from single-task learning, matches or improves upon the performance of complex multi-task optimizers in popular supervised and reinforcement learning settings. We then present an analysis suggesting that many specialized multi-task optimizers can be partly interpreted as forms of regularization, potentially explaining our surprising results. We believe our results call for a critical reevaluation of recent research in the area.
翻译:最近的多任务学习研究反对单一的升级,因为培训只是将任务损失的总和降到最低。相反,在多种任务难以设置的各种假设的启发下,提出了几种特别的多任务优化算法。这些优化者大多需要每个任务梯度,并引入大量的记忆、运行时间和执行间接费用。我们显示,单一任务的升级,加上标准规范化和稳定技术,来自单一任务学习、匹配或改进的单一任务学习。我们随后提出一项分析,表明许多专门的多任务优化方法可以部分地被解释为正规化形式,可能解释我们令人惊讶的结果。我们相信,我们的结果要求对该领域最近的研究进行重要的重新评价。</s>