Graph Neural Networks(GNNs) are vulnerable to backdoor attacks, where adversaries implant malicious triggers to manipulate model predictions. Existing trigger generators are often simplistic in structure and overly reliant on specific features, confining them to a single graph learning paradigm, such as graph supervised learning, graph contrastive learning, or graph prompt learning. This specialized design, which aligns the trigger with one learning objective, results in poor transferability when applied to other learning paradigms. For instance, triggers generated for the graph supervised learning paradigm perform poorly when tested within graph contrastive learning or graph prompt learning environments. Furthermore, these simple generators often fail to utilize complex structural information or node diversity within the graph data. These constraints limit the attack success rates of such methods in general testing scenarios. Therefore, to address these limitations, we propose Cross-Paradigm Graph Backdoor Attacks with Promptable Subgraph Triggers(CP-GBA), a new transferable graph backdoor attack that employs graph prompt learning(GPL) to train a set of universal subgraph triggers. First, we distill a compact yet expressive trigger set from target graphs, which is structured as a queryable repository, by jointly enforcing class-awareness, feature richness, and structural fidelity. Second, we conduct the first exploration of the theoretical transferability of GPL to train these triggers under prompt-based objectives, enabling effective generalization to diverse and unseen test-time paradigms. Extensive experiments across multiple real-world datasets and defense scenarios show that CP-GBA achieves state-of-the-art attack success rates.
翻译:图神经网络(GNNs)易受后门攻击,攻击者通过植入恶意触发器以操纵模型预测。现有触发器生成器通常结构简单且过度依赖特定特征,使其局限于单一图学习范式,如图监督学习、图对比学习或图提示学习。这种将触发器与单一学习目标对齐的专门化设计,导致其在应用于其他学习范式时迁移性较差。例如,为图监督学习范式生成的触发器在图对比学习或图提示学习环境中测试时表现不佳。此外,这些简单生成器往往未能充分利用图数据中的复杂结构信息或节点多样性。这些限制降低了此类方法在通用测试场景中的攻击成功率。因此,为应对这些局限,我们提出跨范式图后门攻击(CP-GBA),这是一种采用图提示学习(GPL)训练通用子图触发器的新型可迁移图后门攻击方法。首先,我们通过联合强化类别感知、特征丰富性和结构保真度,从目标图中提炼出一个紧凑而富有表达力的触发器集合,并将其构建为可查询的知识库。其次,我们首次探索了GPL的理论迁移性,在基于提示的目标下训练这些触发器,使其能够有效泛化至多样且未见过的测试范式。在多个真实世界数据集和防御场景中的广泛实验表明,CP-GBA实现了最先进的攻击成功率。