ML models are known to be vulnerable to adversarial query attacks. In these attacks, queries are iteratively perturbed towards a particular class without any knowledge of the target model besides its output. The prevalence of remotely-hosted ML classification models and Machine-Learning-as-a-Service platforms means that query attacks pose a real threat to the security of these systems. To deal with this, stateful defenses have been proposed to detect query attacks and prevent the generation of adversarial examples by monitoring and analyzing the sequence of queries received by the system. Several stateful defenses have been proposed in recent years. However, these defenses rely solely on similarity or out-of-distribution detection methods that may be effective in other domains. In the malware detection domain, the methods to generate adversarial examples are inherently different, and therefore we find that such detection mechanisms are significantly less effective. Hence, in this paper, we present MalProtect, which is a stateful defense against query attacks in the malware detection domain. MalProtect uses several threat indicators to detect attacks. Our results show that it reduces the evasion rate of adversarial query attacks by 80+\% in Android and Windows malware, across a range of attacker scenarios. In the first evaluation of its kind, we show that MalProtect outperforms prior stateful defenses, especially under the peak adversarial threat.
翻译:众所周知, ML 模式很容易受到对抗性查询攻击。 在这些攻击中, 询问被反复地缠绕到某个特定类别, 除了其输出外对目标模型一无所知。 远程托管 ML 分类模型和机器学习为服务平台的普及意味着询问袭击对这些系统的安全构成真正的威胁。 因此, 为了应对这一威胁, 我们提出了明确的防御方法, 以检测查询攻击, 并通过监测和分析系统收到的查询顺序来防止生成对抗性例子。 近些年来, 提出了好几种状态防御方法。 但是, 这些防御仅仅依靠类似或超出分配的检测方法, 而在其他领域可能有效。 在恶意软件检测领域, 生成对抗性实例的方法本质上是不同的, 因此我们发现这种检测机制远不那么有效。 因此, 我们在本文件中提出了 Mal Protect, 这是一种防止恶意检测领域的查询攻击的状态。 MalProtect 使用几种威胁指标来检测攻击。 我们的结果表明, 它会降低敌对性调查攻击的规避率, 以80° 和 恶意软件的先期显示, 。