Many machine learning problems use data in the tabular domains. Adversarial examples can be especially damaging for these applications. Yet, existing works on adversarial robustness mainly focus on machine-learning models in the image and text domains. We argue that due to the differences between tabular data and images or text, existing threat models are inappropriate for tabular domains. These models do not capture that cost can be more important than imperceptibility, nor that the adversary could ascribe different value to the utility obtained from deploying different adversarial examples. We show that due to these differences the attack and defence methods used for images and text cannot be directly applied to the tabular setup. We address these issues by proposing new cost and utility-aware threat models tailored to the adversarial capabilities and constraints of attackers targeting tabular domains. We introduce a framework that enables us to design attack and defence mechanisms which result in models protected against cost or utility-aware adversaries, e.g., adversaries constrained by a certain dollar budget. We show that our approach is effective on three tabular datasets corresponding to applications for which adversarial examples can have economic and social implications.
翻译:许多机器学习问题使用表格域的数据。反向实例对这些应用可能特别有害。然而,现有的对抗性稳健性工作主要侧重于图像和文本域的机器学习模型。我们争辩说,由于表格数据和图像或文字之间的差异,现有威胁模型不适合表格域。这些模型并不认为成本比不可感知性更为重要,也不认为对手可以对从部署不同的对抗实例中获得的效用赋予不同的价值。我们表明,由于这些差异,用于图像和文字的攻击和防御方法不能直接应用于表格形体设置。我们通过提出针对攻击者针对表格域的对抗能力和限制的新的成本和实用意识威胁模型来解决这些问题。我们引入了一个框架,使我们能够设计攻击和防御机制,从而导致模型保护对手不受成本或实用意识对手(例如受某些美元预算制约的对手)的影响。我们表明,我们的方法在三种表格数据集上是有效的,与对抗性实例可能具有经济和社会影响的应用程序相对应。