Machine learning systems are deployed in critical settings, but they might fail in unexpected ways, impacting the accuracy of their predictions. Poisoning attacks against machine learning induce adversarial modification of data used by a machine learning algorithm to selectively change its output when it is deployed. In this work, we introduce a novel data poisoning attack called a \emph{subpopulation attack}, which is particularly relevant when datasets are large and diverse. We design a modular framework for subpopulation attacks, instantiate it with different building blocks, and show that the attacks are effective for a variety of datasets and machine learning models. We further optimize the attacks in continuous domains using influence functions and gradient optimization methods. Compared to existing backdoor poisoning attacks, subpopulation attacks have the advantage of inducing misclassification in naturally distributed data points at inference time, making the attacks extremely stealthy. We also show that our attack strategy can be used to improve upon existing targeted attacks. We prove that, under some assumptions, subpopulation attacks are impossible to defend against, and empirically demonstrate the limitations of existing defenses against our attacks, highlighting the difficulty of protecting machine learning against this threat.
翻译:机器学习系统在关键环境下部署,但可能以意外的方式失败,从而影响预测的准确性。对机器学习的袭击导致对机器学习算法所使用的数据进行对抗性修改,以便在部署时有选择地改变其输出。在这项工作中,我们引入了名为“emph{subpubulity attack}”的新数据中毒袭击,当数据集巨大且种类繁多时,这种袭击特别相关。我们设计了亚人口攻击的模块框架,用不同的构件对它进行即时转换,并表明这些袭击对于各种数据集和机器学习模型是有效的。我们利用影响功能和梯度优化方法进一步优化连续领域的袭击。与现有的后门中毒袭击相比,亚人口攻击具有以下优势:在推断时自然分布的数据点导致分类错误,使袭击变得极为隐秘。我们还表明,我们的攻击战略可以用来改进现有的目标攻击。我们证明,根据某些假设,亚人口攻击是无法防御的,并且用经验证明现有防御我们攻击的局限性。我们强调保护机器学会对付这种威胁的困难。