Many machine learning systems are vulnerable to small perturbations made to inputs either at test time or at training time. This has received much recent interest on the empirical front due to applications where reliability and security are critical. However, theoretical understanding of algorithms that are robust to adversarial perturbations is limited. In this work we focus on Principal Component Analysis (PCA), a ubiquitous algorithmic primitive in machine learning. We formulate a natural robust variant of PCA where the goal is to find a low dimensional subspace to represent the given data with minimum projection error, that is in addition robust to small perturbations measured in $\ell_q$ norm (say $q=\infty$). Unlike PCA which is solvable in polynomial time, our formulation is computationally intractable to optimize as it captures a variant of the well-studied sparse PCA objective as a special case. We show the following results: -Polynomial time algorithm that is constant factor competitive in the worst-case with respect to the best subspace, in terms of the projection error and the robustness criterion. -We show that our algorithmic techniques can also be made robust to adversarial training-time perturbations, in addition to yielding representations that are robust to adversarial perturbations at test time. Specifically, we design algorithms for a strong notion of training-time perturbations, where every point is adversarially perturbed up to a specified amount. -We illustrate the broad applicability of our algorithmic techniques in addressing robustness to adversarial perturbations, both at training time and test time. In particular, our adversarially robust PCA primitive leads to computationally efficient and robust algorithms for both unsupervised and supervised learning problems such as clustering and learning adversarially robust classifiers.
翻译:许多机器学习系统很容易受到测试时间或培训时间对投入的小扰动的影响。 由于应用中可靠性和安全性至关重要,这在经验前最近引起了很大的兴趣。 但是,对对动态对对抗性扰动强的算法的理论理解有限。 在这项工作中,我们侧重于主构分析(PCA),这是机器学习中一个无处不在的原始算法。我们开发了一个自然强势的CPA变方,目标是找到一个低维的亚空间来代表给定数据,并有最低的预测性差错,这与以美元/ell_q美元标准衡量的小扰动(say $q_infty$)相比是强的。 与在多元性扰动性时的算法不同,我们的配法在计算中,当它捕捉到精心研究的稀疏漏的CPA目标变异。 我们展示了以下结果: oblyomyalalive 时间算法,在最坏的亚空间中不断具有竞争力,在预测性错误和稳健性的亚程标准方面,在预测性差的数值上都非常强的亚,在每次的运算方法上,在每次的精确性变数的数值上,在每次测试中,在每次的计算中,我们最强的计算中,在每次的计算中,我们最强的判变压压压式的计算中,在每次的算法式的计算中,在每次的计算中,在每次的计算中都的计算中,在每次的计算中,在精确算中,在精确性变的计算中,在每次的计算中,在精确性能的计算中,在精确性变的计算。