The mean squared error loss is widely used in many applications, including auto-encoders, multi-target regression, and matrix factorization, to name a few. Despite computational advantages due to its differentiability, it is not robust to outliers. In contrast, l_p norms are known to be robust, but cannot be optimized via, e.g., stochastic gradient descent, as they are non-differentiable. We propose an algorithm inspired by so-called model-based optimization (MBO) [35, 36], which replaces a non-convex objective with a convex model function and alternates between optimizing the model function and updating the solution. We apply this to robust regression, proposing SADM, a stochastic variant of the Online Alternating Direction Method of Multipliers (OADM) [50] to solve the inner optimization in MBO. We show that SADM converges with the rate O(log T/T). Finally, we demonstrate experimentally (a) the robustness of l_p norms to outliers and (b) the efficiency of our proposed model-based algorithms in comparison with gradient methods on autoencoders and multi-target regression.
翻译:平均平方误差损失在许多应用中被广泛使用,包括自动读数器、多目标回归和矩阵因数化等。尽管计算优势不同,但它对外部值并不健全。相反,l_p规范已知是稳健的,但无法通过(例如)随机梯度梯度梯度下降优化,因为它们不具有差异性。我们提出了一个由所谓的基于模型的优化(MBO)[35、36]所启发的算法,它用一个 convex 模型函数取代非convex 目标,并在优化模型函数和更新解决方案之间加以替代。我们将此应用到稳健的回归,提出SADM,这是多相控者在线对调方向法(OADM)[50]的随机变量,以解决MBO的内部优化问题。我们显示,SADM与O(log T/T) 比率一致。最后,我们实验性地展示了(a) l_p 规范对外端值的坚固度,以及(b) 我们拟议的基于模型和多向梯度的梯度分析法的效益。