Machine learning (ML) models are known to be vulnerable to adversarial examples. Applications of ML to voice biometrics authentication are no exception. Yet, the implications of audio adversarial examples on these real-world systems remain poorly understood given that most research targets limited defenders who can only listen to the audio samples. Conflating detectability of an attack with human perceptibility, research has focused on methods that aim to produce imperceptible adversarial examples which humans cannot distinguish from the corresponding benign samples. We argue that this perspective is coarse for two reasons: 1. Imperceptibility is impossible to verify; it would require an experimental process that encompasses variations in listener training, equipment, volume, ear sensitivity, types of background noise etc, and 2. It disregards pipeline-based detection clues that realistic defenders leverage. This results in adversarial examples that are ineffective in the presence of knowledgeable defenders. Thus, an adversary only needs an audio sample to be plausible to a human. We thus introduce surreptitious adversarial examples, a new class of attacks that evades both human and pipeline controls. In the white-box setting, we instantiate this class with a joint, multi-stage optimization attack. Using an Amazon Mechanical Turk user study, we show that this attack produces audio samples that are more surreptitious than previous attacks that aim solely for imperceptibility. Lastly we show that surreptitious adversarial examples are challenging to develop in the black-box setting.
翻译:已知机器学习(ML)模式很容易受到对抗性实例的伤害。 ML 用于声音生物鉴别认证的应用也不例外。然而,这些真实世界系统中的有声对立实例的影响仍然不甚为人理解,因为大多数研究的对象都是只能听听听声音样本的有限捍卫者。将攻击的可探测性与人类能见性混为一谈,研究的重点是旨在产生人类无法与相应良性样本区分的难以察觉的对抗性实例的方法。我们争辩说,这种观点很粗糙,原因有二:1. 难以理解是无法核实的;它需要试验过程,包括听众培训、设备、数量、耳敏锐度、背景噪音类型等的变异性,以及2. 它无视基于管道的探测线索,而现实捍卫者则利用现实性捍卫者所利用的线索。因此,对立方只需有听力的样本就可以使人相信。因此,我们引入了一种神秘的对抗性对抗性例子,一种逃避人和管道控制的新的攻击。在白箱设置中,我们即进行这一类的黑色研究,而我们只是要让这个类的以联合、多阶段的图像模型显示我们先前的进攻的磁性攻击。