Poisoning attacks have emerged as a significant security threat to machine learning algorithms. It has been demonstrated that adversaries who make small changes to the training set, such as adding specially crafted data points, can hurt the performance of the output model. Some of the stronger poisoning attacks require the full knowledge of the training data. This leaves open the possibility of achieving the same attack results using poisoning attacks that do not have the full knowledge of the clean training set. In this work, we initiate a theoretical study of the problem above. Specifically, for the case of feature selection with LASSO, we show that full-information adversaries (that craft poisoning examples based on the rest of the training data) are provably stronger than the optimal attacker that is oblivious to the training set yet has access to the distribution of the data. Our separation result shows that the two setting of data-aware and data-oblivious are fundamentally different and we cannot hope to always achieve the same attack or defense results in these scenarios.
翻译:毒物攻击已成为对机器学习算法的重大安全威胁。 事实证明,对训练组进行小小改动的对手,例如增加专门制作的数据点,可能会损害产出模型的性能。 一些更强的中毒攻击需要完全了解训练数据。 这为利用对清洁训练组不完全了解的中毒攻击取得同样的攻击结果提供了可能性。 在这项工作中,我们开始对上文的问题进行理论研究。 具体地说,对于LASSO的特征选择,我们显示,完全的信息对手(根据训练数据其余部分的手动中毒例子)比最理想的攻击者(训练组所忽略的最佳攻击者)强,而训练组却没有机会分发数据。 我们的分离结果显示,两种数据认知和数据模糊的设置根本不同,我们不能指望在这些情景中总是取得同样的攻击或防御结果。