In Multi-task learning (MTL), a joint model is trained to simultaneously make predictions for several tasks. Joint training reduces computation costs and improves data efficiency; however, since the gradients of these different tasks may conflict, training a joint model for MTL often yields lower performance than its corresponding single-task counterparts. A common method for alleviating this issue is to combine per-task gradients into a joint update direction using a particular heuristic. In this paper, we propose viewing the gradients combination step as a bargaining game, where tasks negotiate to reach an agreement on a joint direction of parameter update. Under certain assumptions, the bargaining problem has a unique solution, known as the Nash Bargaining Solution, which we propose to use as a principled approach to multi-task learning. We describe a new MTL optimization procedure, Nash-MTL, and derive theoretical guarantees for its convergence. Empirically, we show that Nash-MTL achieves state-of-the-art results on multiple MTL benchmarks in various domains.
翻译:在多任务学习(MTL)中,对联合模式进行了培训,以便对若干任务同时作出预测。联合培训降低了计算成本并提高了数据效率;但是,由于这些不同任务的梯度可能发生冲突,因此培训一个联合模式的MTL的绩效往往低于相应的单任务对应方。缓解这一问题的一个共同方法是利用某种超常性来将每个任务梯度合并成一个联合更新方向。在本文件中,我们提议将梯度组合步骤视为一种讨价还价的游戏,谈判就参数更新的共同方向达成协议。在某些假设下,谈判问题有一个独特的解决办法,即纳什谈判解决方案,我们提议将其作为多任务学习的有原则性方法。我们描述了一个新的MTT优化程序,纳什-MTL,并获得其趋同的理论保证。我们从中可以看出,纳什-MTL在不同领域的多重MTL基准上取得了最新成果。