In real-world applications, it is important for machine learning algorithms to be robust against data outliers or corruptions. In this paper, we focus on improving the robustness of a large class of learning algorithms that are formulated as low-rank semi-definite programming (SDP) problems. Traditional formulations use square loss, which is notorious for being sensitive to outliers. We propose to replace this with more robust noise models, including the $\ell_1$-loss and other nonconvex losses. However, the resultant optimization problem becomes difficult as the objective is no longer convex or smooth. To alleviate this problem, we design an efficient algorithm based on majorization-minimization. The crux is on constructing a good optimization surrogate, and we show that this surrogate can be efficiently obtained by the alternating direction method of multipliers (ADMM). By properly monitoring ADMM's convergence, the proposed algorithm is empirically efficient and also theoretically guaranteed to converge to a critical point. Extensive experiments are performed on four machine learning applications using both synthetic and real-world data sets. Results show that the proposed algorithm is not only fast but also has better performance than the state-of-the-art.
翻译:在现实世界应用中,机器学习算法必须针对数据外值或腐败而强大起来。 在本文中,我们侧重于提高作为低级半无限期编程(SDP)问题而形成的大规模学习算法的稳健性。传统配方使用平方损失,因为对外值敏感而臭名昭著。我们提议用更强大的噪音模型来取代这一损失,包括$@$_1美元损失和其他非电解损失。然而,由此产生的优化问题变得很困难,因为目标不再具有阴滑或平滑。为了缓解这一问题,我们设计了一种基于主要化-最小化的高效算法。核心是构建一个良好的优化替代模型,我们表明这一代方值可以通过乘数交替方向方法(ADMMM)有效获得。通过适当监测ADMM的趋同性,拟议的算法具有经验效率,理论上也保证与临界点一致。在四个机器学习应用中,使用合成和真实世界数据集进行了广泛的实验。结果显示,拟议的算法不仅表现很快,而且还比状态要好。