Machine learning is becoming ubiquitous. From financial to medicine, machine learning models are boosting decision-making processes and even outperforming humans in some tasks. This huge progress in terms of prediction quality does not however find a counterpart in the security of such models and corresponding predictions, where perturbations of fractions of the training set (poisoning) can seriously undermine the model accuracy. Research on poisoning attacks and defenses even predates the introduction of deep neural networks, leading to several promising solutions. Among them, ensemble-based defenses, where different models are trained on portions of the training set and their predictions are then aggregated, are getting significant attention, due to their relative simplicity and theoretical and practical guarantees. The work in this paper designs and implements a hash-based ensemble approach for ML robustness and evaluates its applicability and performance on random forests, a machine learning model proved to be more resistant to poisoning attempts on tabular datasets. An extensive experimental evaluation is carried out to evaluate the robustness of our approach against a variety of attacks, and compare it with a traditional monolithic model based on random forests.
翻译:机器学习正在变得无处不在。从财政到医学,机器学习模型正在推动决策进程,甚至在某些任务中表现优异。在预测质量方面的巨大进步在这种模型和相应预测的安全性方面找不到对应的对应方,因为这种模型和相应的预测会使训练组的碎片(渗透)的扰动严重破坏模型的准确性。关于中毒袭击和防御的研究甚至早于引入深神经网络,从而导致若干有希望的解决办法。其中,基于共同防御,即对培训组部分进行培训,然后将其预测汇总,正在引起人们的极大关注。本文中的工作为ML稳健性设计并采用基于散列的共鸣方法,并评估其在随机森林中的适用性和性。一个机器学习模型证明更能抵制在表格数据集中进行中毒的尝试。进行了广泛的实验性评估,以评价我们应对各种攻击的方法的稳健性,并将它与基于随机森林的传统单一模型进行比较。