The existence of adversarial attacks (or adversarial examples) brings huge concern about the machine learning (ML) model's safety issues. For many safety-critical ML tasks, such as financial forecasting, fraudulent detection, and anomaly detection, the data samples are usually mixed-type, which contain plenty of numerical and categorical features at the same time. However, how to generate adversarial examples with mixed-type data is still seldom studied. In this paper, we propose a novel attack algorithm M-Attack, which can effectively generate adversarial examples in mixed-type data. Based on M-Attack, attackers can attempt to mislead the targeted classification model's prediction, by only slightly perturbing both the numerical and categorical features in the given data samples. More importantly, by adding designed regularizations, our generated adversarial examples can evade potential detection models, which makes the attack indeed insidious. Through extensive empirical studies, we validate the effectiveness and efficiency of our attack method and evaluate the robustness of existing classification models against our proposed attack. The experimental results highlight the feasibility of generating adversarial examples toward machine learning models in real-world applications.
翻译:对抗性攻击(或对抗性例子)的存在引起了对机器学习模式安全问题的极大关注。对于许多对安全至关重要的 ML 任务,如金融预测、欺诈性检测和异常检测,数据样本通常是混合型的,同时包含大量数字和绝对特征。然而,如何用混合型数据生成对抗性实例仍然很少研究。在本文中,我们建议采用新型攻击算法M-Attack,这可以有效地生成混合型数据中的对抗性实例。根据M-Attack,攻击者可以试图误导定向分类模式的预测,只略微渗透到特定数据样本中的数字和绝对特征。更重要的是,通过添加设计的正规化,我们生成的对抗性实例可以避开潜在的探测模型,从而使攻击确实具有阴险性。通过广泛的经验研究,我们验证了攻击方法的效力和效率,并评估了针对我们拟议攻击的现有分类模型的稳健性。实验结果突出表明了在现实世界应用中生成机器学习模型的对抗性实例的可行性。