Transferability of adversarial samples became a serious concern due to their impact on the reliability of machine learning system deployments, as they find their way into many critical applications. Knowing factors that influence transferability of adversarial samples can assist experts to make informed decisions on how to build robust and reliable machine learning systems. The goal of this study is to provide insights on the mechanisms behind the transferability of adversarial samples through an attack-centric approach. This attack-centric perspective interprets how adversarial samples would transfer by assessing the impact of machine learning attacks (that generated them) on a given input dataset. To achieve this goal, we generated adversarial samples using attacker models and transferred these samples to victim models. We analyzed the behavior of adversarial samples on victim models and outlined four factors that can influence the transferability of adversarial samples. Although these factors are not necessarily exhaustive, they provide useful insights to researchers and practitioners of machine learning systems.
翻译:由于对机器学习系统部署的可靠性产生影响,对对抗性样品的可转让性产生了重大影响,因为这些样品进入了许多关键应用领域。影响对抗性样品可转让性的知情因素可以帮助专家就如何建立稳健可靠的机器学习系统作出知情决定。本研究的目的是通过以攻击为中心的方法,就对抗性样品可转让性背后的机制提供深刻见解。这种以攻击为中心的观点通过评估机器学习攻击(产生这些攻击)对特定输入数据集的影响来解释对抗性样品如何转移。为实现这一目标,我们利用攻击者模型制作了对抗性样品,并将这些样品转让给受害者模型。我们分析了对抗性样品在受害者模型上的行为,并概述了可以影响对抗性样品可转让的四个因素。虽然这些因素不一定详尽无遗,但它们为研究者和机器学习系统的从业人员提供了有益的见解。