Despite the excellent performance of neural-network-based audio source separation methods and their wide range of applications, their robustness against intentional attacks has been largely neglected. In this work, we reformulate various adversarial attack methods for the audio source separation problem and intensively investigate them under different attack conditions and target models. We further propose a simple yet effective regularization method to obtain imperceptible adversarial noise while maximizing the impact on separation quality with low computational complexity. Experimental results show that it is possible to largely degrade the separation quality by adding imperceptibly small noise when the noise is crafted for the target model. We also show the robustness of source separation models against a black-box attack. This study provides potentially useful insights for developing content protection methods against the abuse of separated signals and improving the separation performance and robustness.
翻译:尽管基于神经网络的音源分离方法表现出色,而且应用范围很广,但它们对蓄意攻击的强烈性基本上被忽视。在这项工作中,我们重新拟订各种针对音源分离问题的对抗性攻击方法,并在不同的攻击条件和目标模式下深入调查这些方法。我们进一步提出一个简单而有效的规范化方法,以获得无法察觉的对抗性噪音,同时尽量扩大对计算复杂性低的分离质量的影响。实验结果显示,在为目标模式设计噪音时,通过增加难以察觉的小噪音,可以在很大程度上降低分离质量。我们还展示了针对黑盒袭击的源分离模型的牢固性。这项研究为开发内容保护方法,防止滥用分离信号,改进分离性能和稳健性提供了可能有用的见解。