Machine learning problems with multiple objective functions appear either in learning with multiple criteria where learning has to make a trade-off between multiple performance metrics such as fairness, safety and accuracy; or, in multi-task learning where multiple tasks are optimized jointly, sharing inductive bias between them. This problems are often tackled by the multi-objective optimization framework. However, existing stochastic multi-objective gradient methods and its variants (e.g., MGDA, PCGrad, CAGrad, etc.) all adopt a biased noisy gradient direction, which leads to degraded empirical performance. To this end, we develop a stochastic Multi-objective gradient Correction (MoCo) method for multi-objective optimization. The unique feature of our method is that it can guarantee convergence without increasing the batch size even in the non-convex setting. Simulations on multi-task supervised and reinforcement learning demonstrate the effectiveness of our method relative to state-of-the-art methods.
翻译:具有多重客观功能的机器学习问题出现于学习中,有多种标准,学习必须在公平、安全和准确性等多种业绩指标之间作出权衡;或是在多任务学习中,在多种任务得到优化的情况下,共同分享感化偏差。这些问题往往由多目标优化框架来解决。然而,现有的随机多目标梯度方法及其变体(如MGDA、PCGrad、CAGrad等)都采用了有偏见的噪音梯度方向,导致经验性能退化。为此,我们为多目标优化开发了一种随机多目的多目的梯度校正(MoCo)方法。我们方法的独特特征是,它可以保证趋同,而即使在非convex环境下也不增加批量体大小。多任务监督和强化学习模拟显示我们方法相对于状态方法的有效性。