In a transfer-based attack against Automatic Speech Recognition (ASR) systems, attacks are unable to access the architecture and parameters of the target model. Existing attack methods are mostly investigated in voice assistant scenarios with restricted voice commands, prohibiting their applicability to more general ASR related applications. To tackle this challenge, we propose a novel contextualized attack with deletion, insertion, and substitution adversarial behaviors, namely TransAudio, which achieves arbitrary word-level attacks based on the proposed two-stage framework. To strengthen the attack transferability, we further introduce an audio score-matching optimization strategy to regularize the training process, which mitigates adversarial example over-fitting to the surrogate model. Extensive experiments and analysis demonstrate the effectiveness of TransAudio against open-source ASR models and commercial APIs.
翻译:在基于转移的对抗性攻击中,攻击者无法访问目标模型的体系结构和参数。现有的攻击方法主要是在语音助手场景中进行研究,其限制了语音命令,从而不能应用于更通用的ASR相关应用。为了解决这个挑战,我们提出了一个新颖的上下文化攻击,并采用了删除、插入和替换的对抗性行为,即TransAudio,它基于所提出的两阶段框架实现了任意单词级攻击。为了增强攻击的可转移性,我们进一步引入了音频评分匹配优化策略来规范训练过程,从而减轻对代理模型的对抗例子过度拟合的影响。广泛的实验和分析证明了TransAudio 对开源ASR模型和商业API的有效性。