The emergence of graph foundation models (GFMs), particularly those incorporating language models (LMs), has revolutionized graph learning and demonstrated remarkable performance on text-attributed graphs (TAGs). However, compared to traditional GNNs, these LM-empowered GFMs introduce unique security vulnerabilities during the unsecured prompt tuning phase that remain understudied in current research. Through empirical investigation, we reveal a significant performance degradation in traditional graph backdoor attacks when operating in attribute-inaccessible constrained TAG systems without explicit trigger node attribute optimization. To address this, we propose a novel dual-trigger backdoor attack framework that operates at both text-level and struct-level, enabling effective attacks without explicit optimization of trigger node text attributes through the strategic utilization of a pre-established text pool. Extensive experimental evaluations demonstrate that our attack maintains superior clean accuracy while achieving outstanding attack success rates, including scenarios with highly concealed single-trigger nodes. Our work highlights critical backdoor risks in web-deployed LM-empowered GFMs and contributes to the development of more robust supervision mechanisms for open-source platforms in the era of foundation models.
翻译:图基础模型(GFMs)的出现,特别是那些融合语言模型(LMs)的模型,彻底改变了图学习范式,并在文本属性图(TAGs)上展现出卓越的性能。然而,与传统图神经网络(GNNs)相比,这些由语言模型赋能的图基础模型在其不安全的提示调优阶段引入了独特的安全漏洞,而当前研究对此尚未充分探讨。通过实证研究,我们发现在属性不可访问的受限文本属性图系统中,若未对触发节点属性进行显式优化,传统图后门攻击的性能会出现显著下降。为解决此问题,我们提出了一种新颖的双触发后门攻击框架,该框架在文本层面和结构层面同时运作,通过策略性地利用预先建立的文本池,无需对触发节点文本属性进行显式优化即可实现有效攻击。大量实验评估表明,我们的攻击在保持优异干净准确率的同时,实现了出色的攻击成功率,包括在高度隐蔽的单触发节点场景下。本研究揭示了网络部署的语言模型赋能图基础模型中存在的关键后门风险,并为开源平台在基础模型时代开发更鲁棒的监督机制提供了贡献。