The recent success of machine learning (ML) has been fueled by the increasing availability of computing power and large amounts of data in many different applications. However, the trustworthiness of the resulting models can be compromised when such data is maliciously manipulated to mislead the learning process. In this article, we first review poisoning attacks that compromise the training data used to learn ML models, including attacks that aim to reduce the overall performance, manipulate the predictions on specific test samples, and even implant backdoors in the model. We then discuss how to mitigate these attacks using basic security principles, or by deploying ML-oriented defensive mechanisms. We conclude our article by formulating some relevant open challenges which are hindering the development of testing methods and benchmarks suitable for assessing and improving the trustworthiness of ML models against data poisoning attacks
翻译:最近机器学习的成功,由于计算机功率的日益增加和许多不同应用中大量数据的提供,促进了最近机器学习的成功。然而,如果此类数据被恶意操纵以误导学习过程,则所产生的模型的可信度可能受到损害。在本篇文章中,我们首先审查损害用于学习ML模型的培训数据的中毒袭击,包括旨在降低总体性能、操纵特定测试样品预测、甚至在该模型中植入后门的攻击。然后我们讨论如何利用基本安全原则或部署面向ML的防御机制来减轻这些袭击。我们最后通过制定一些相关的公开挑战来结束我们的文章,这些挑战阻碍制定适合评估和改进ML模型对数据中毒袭击的信任度的测试方法和基准。</s>