Automatic speaker verification (ASV) is one of the core technologies in biometric identification. With the ubiquitous usage of ASV systems in safety-critical applications, more and more malicious attackers attempt to launch adversarial attacks at ASV systems. In the midst of the arms race between attack and defense in ASV, how to effectively improve the robustness of ASV against adversarial attacks remains an open question. We note that the self-supervised learning models possess the ability to mitigate superficial perturbations in the input after pretraining. Hence, with the goal of effective defense in ASV against adversarial attacks, we propose a standard and attack-agnostic method based on cascaded self-supervised learning models to purify the adversarial perturbations. Experimental results demonstrate that the proposed method achieves effective defense performance and can successfully counter adversarial attacks in scenarios where attackers may either be aware or unaware of the self-supervised learning models.
翻译:自动扬声器核查(ASV)是生物鉴别的核心技术之一。随着ASV系统在安全关键应用中的普遍使用,越来越多的恶意攻击者试图对ASV系统发动对抗性攻击。在ASV攻击和防御之间的军备竞赛中,如何有效提高ASV对对抗性攻击的稳健性仍然是一个未决问题。我们注意到,自我监督的学习模式有能力在培训前减少投入中的表面扰动。因此,为了在ASV中有效防御对抗性攻击,我们提议了一种标准和攻击性对抗性方法,其依据是累进式自我监督的学习模式,以净化对抗性攻击性攻击性攻击。实验结果表明,拟议的方法取得了有效的防御性效果,能够在攻击者可能知道或不知道自我监督的学习模式的情况下成功地对抗对抗性攻击性攻击。