One of the most concerning threats for modern AI systems is data poisoning, where the attacker injects maliciously crafted training data to corrupt the system's behavior at test time. Availability poisoning is a particularly worrisome subset of poisoning attacks where the attacker aims to cause a Denial-of-Service (DoS) attack. However, the state-of-the-art algorithms are computationally expensive because they try to solve a complex bi-level optimization problem (the "hammer"). We observed that in particular conditions, namely, where the target model is linear (the "nut"), the usage of computationally costly procedures can be avoided. We propose a counter-intuitive but efficient heuristic that allows contaminating the training set such that the target system's performance is highly compromised. We further suggest a re-parameterization trick to decrease the number of variables to be optimized. Finally, we demonstrate that, under the considered settings, our framework achieves comparable, or even better, performances in terms of the attacker's objective while being significantly more computationally efficient.
翻译:与现代AI系统威胁最相关的一个威胁是数据中毒,攻击者恶意地输入培训数据,以腐蚀系统测试时的行为。可获性中毒是一个特别令人担忧的中毒袭击子子集,攻击者的目的是发动拒绝服务(DoS)攻击。然而,最先进的算法在计算上成本很高,因为它们试图解决复杂的双级优化问题(“锤子 ” )。我们发现,在特定条件下,即目标模型是线性(“nut” ),可以避免使用计算成本高昂的程序。我们建议采用反直观但有效的超常方法,允许污染训练组,从而严重损害目标系统的表现。我们还建议采用重新量化策略,以减少要优化的变量数量。最后,我们证明,在考虑到的环境下,我们的框架在计算效率显著提高的同时,在攻击者目标目标方面实现了可比甚至更好的业绩。