Stealthy Dual-Trigger Backdoors: Attacking Prompt Tuning in LM-Empowered Graph Foundation Models

📅 2025-10-16
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing backdoor attacks on language model-empowered graph foundation models (GFMs) fail under restrictive text-attributed graphs (TAGs) where node textual attributes are inaccessible. Method: We propose the first dual-trigger backdoor attack framework that bypasses explicit textual attribute modification by jointly optimizing prompt templates and graph topology, guided by a textual prompt pool to implicitly activate the backdoor. Our approach integrates prompt tuning, graph neural networks, and large language models to enable synergistic text-level and structure-level triggering. Contribution/Results: Under stringent constraints—including single-trigger-node deployment and zero access to node text attributes—our method achieves >95% attack success rate while preserving clean-task accuracy with negligible degradation (<0.5%). It significantly outperforms conventional graph backdoor methods and exposes critical security vulnerabilities in open GFM platforms.

Technology Category

Application Category

📝 Abstract
The emergence of graph foundation models (GFMs), particularly those incorporating language models (LMs), has revolutionized graph learning and demonstrated remarkable performance on text-attributed graphs (TAGs). However, compared to traditional GNNs, these LM-empowered GFMs introduce unique security vulnerabilities during the unsecured prompt tuning phase that remain understudied in current research. Through empirical investigation, we reveal a significant performance degradation in traditional graph backdoor attacks when operating in attribute-inaccessible constrained TAG systems without explicit trigger node attribute optimization. To address this, we propose a novel dual-trigger backdoor attack framework that operates at both text-level and struct-level, enabling effective attacks without explicit optimization of trigger node text attributes through the strategic utilization of a pre-established text pool. Extensive experimental evaluations demonstrate that our attack maintains superior clean accuracy while achieving outstanding attack success rates, including scenarios with highly concealed single-trigger nodes. Our work highlights critical backdoor risks in web-deployed LM-empowered GFMs and contributes to the development of more robust supervision mechanisms for open-source platforms in the era of foundation models.
Problem

Research questions and friction points this paper is trying to address.

Identifies security vulnerabilities in language model-enhanced graph foundation models
Proposes dual-trigger backdoor attacks targeting prompt tuning phase of GFMs
Addresses performance degradation of traditional attacks in constrained TAG systems
Innovation

Methods, ideas, or system contributions that make the work stand out.

Dual-trigger backdoor attack combining text and structure
Utilizes pre-established text pool without attribute optimization
Maintains clean accuracy while achieving high attack success