Can LLMs Fool Graph Learning? Exploring Universal Adversarial Attacks on Text-Attributed Graphs
arXiv cs.AI / 3/24/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper studies vulnerabilities in text-attributed graphs (TAGs), where combining topology with node text improves learning but introduces new adversarial risk surfaces.
- It highlights the difficulty of creating universal attacks that work across different backbones (GNNs vs. PLMs) and across black-box access to many LLMs exposed only via APIs.
- The authors propose BadGraph, an attack framework that elicits an LLM’s graph understanding and jointly perturbs node topology and textual semantics to craft cross-modal, generalizable attack “shortcuts.”
- Experiments indicate BadGraph can produce universal and effective attacks against both GNN-based and LLM-based graph reasoners, reporting performance drops of up to 76.3%.
- The work includes both theoretical and empirical analyses suggesting the attacks can be stealthy while remaining interpretable.
Related Articles
How AI is Transforming Dynamics 365 Business Central
Dev.to
Algorithmic Gaslighting: A Formal Legal Template to Fight AI Safety Pivots That Cause Psychological Harm
Reddit r/artificial
Do I need different approaches for different types of business information errors?
Dev.to
ShieldCortex: What We Learned Protecting AI Agent Memory
Dev.to
WordPress Theme Customization Without Code: The AI Revolution
Dev.to