Backdoor attacks have been demonstrated as a security threat for machine learning models. Traditional backdoor attacks intend to inject backdoor functionality into the model such that the backdoored model will perform abnormally on inputs with predefined backdoor triggers and still retain state-of-the-art performance on the clean inputs. While there are already some works on backdoor attacks on Graph Neural Networks (GNNs), the backdoor trigger in the graph domain is mostly injected into random positions of the sample. There is no work analyzing and explaining the backdoor attack performance when injecting triggers into the most important or least important area in the sample, which we refer to as trigger-injecting strategies MIAS and LIAS, respectively. Our results show that, generally, LIAS performs better, and the differences between the LIAS and MIAS performance can be significant. Furthermore, we explain these two strategies' similar (better) attack performance through explanation techniques, which results in a further understanding of backdoor attacks in GNNs.
翻译:背门攻击已被证明是机器学习模型的一种安全威胁。传统的背门攻击旨在将背门功能注入模型,使得背门模型在带有预定义背门触发器的输入上表现异常,并仍然保持对干净输入的最先进性能。虽然已经有一些针对图神经网络的背门攻击工作,但是在图领域中,背门触发器大多被注入到样本的随机位置上。目前还没有研究分析和解释当将触发器注入到样本的最重要或最不重要区域时的背门攻击性能,我们称之为触发注入策略MIAS和LIAS。我们的结果表明,一般来说,LIAS的表现更好,而LIAS和MIAS表现之间的差异可能非常大。此外,我们通过解释技术解释这两种策略的类似(更好)的攻击表现,从而进一步了解了图神经网络中的背门攻击。