Deep neural network models are massively deployed on a wide variety of hardware platforms. This results in the appearance of new attack vectors that significantly extend the standard attack surface, extensively studied by the adversarial machine learning community. One of the first attack that aims at drastically dropping the performance of a model, by targeting its parameters (weights) stored in memory, is the Bit-Flip Attack (BFA). In this work, we point out several evaluation challenges related to the BFA. First of all, the lack of an adversary's budget in the standard threat model is problematic, especially when dealing with physical attacks. Moreover, since the BFA presents critical variability, we discuss the influence of some training parameters and the importance of the model architecture. This work is the first to present the impact of the BFA against fully-connected architectures that present different behaviors compared to convolutional neural networks. These results highlight the importance of defining robust and sound evaluation methodologies to properly evaluate the dangers of parameter-based attacks as well as measure the real level of robustness offered by a defense.
翻译:深神经网络模型被大规模地部署在各种硬件平台上,这导致出现新的攻击矢量,大大扩展标准攻击表面,由对抗性机器学习界广泛研究。第一次攻击的目的是通过瞄准存储在记忆中的参数(重量)来大幅降低模型的性能,第一次攻击是Bit-Flip Action(BFA)。在这项工作中,我们指出了与BFA有关的若干评价挑战。首先,标准威胁模型中缺乏对手的预算是有问题的,特别是在处理人身攻击时。此外,由于BFA提出了关键的变异性,我们讨论了一些培训参数的影响和模型结构的重要性。这是首次介绍BFA对与动态神经网络相比具有不同行为的完全相连结构的影响。这些结果突出了确定稳健和健全的评价方法的重要性,以便适当评估参数攻击的危险,并衡量防御所提供的真正强健度。