Are LLM-Enhanced Graph Neural Networks Robust against Poisoning Attacks?

arXiv cs.LG / 3/30/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies whether LLM-enhanced Graph Neural Networks are robust to poisoning attacks that manipulate both graph structure and node textual attributes during training.
  • It proposes a systematic robustness evaluation framework that tests 24 victim models built from combinations of eight LLM/LM-based feature enhancers and three GNN backbones.
  • The evaluation spans six structural poisoning attacks (targeted and non-targeted) and three textual poisoning attacks at character, word, and sentence levels, across four datasets selected to avoid LLM pretraining ground-truth leakage.
  • Experimental results show LLM-enhanced GNNs maintain higher accuracy and lower Relative Drop in Accuracy than a shallow embedding baseline under many attack settings.
  • The authors attribute improved robustness to how node representations encode structural and label information, and they also introduce future offensive/defensive directions plus a combined attack and graph purification defense, releasing source code for the framework.

Abstract

Large Language Models (LLMs) have advanced Graph Neural Networks (GNNs) by enriching node representations with semantic features, giving rise to LLM-enhanced GNNs that achieve notable performance gains. However, the robustness of these models against poisoning attacks, which manipulate both graph structures and textual attributes during training, remains unexplored. To bridge this gap, we propose a robustness assessment framework that systematically evaluates LLM-enhanced GNNs under poisoning attacks. Our framework enables comprehensive evaluation across multiple dimensions. Specifically, we assess 24 victim models by combining eight LLM- or Language Model (LM)-based feature enhancers with three representative GNN backbones. To ensure diversity in attack coverage, we incorporate six structural poisoning attacks (both targeted and non-targeted) and three textual poisoning attacks operating at the character, word, and sentence levels. Furthermore, we employ four real-world datasets, including one released after the emergence of LLMs, to avoid potential ground truth leakage during LLM pretraining, thereby ensuring fair evaluation. Extensive experiments show that LLM-enhanced GNNs exhibit significantly higher accuracy and lower Relative Drop in Accuracy (RDA) than a shallow embedding-based baseline across various attack settings. Our in-depth analysis identifies key factors that contribute to this robustness, such as the effective encoding of structural and label information in node representations. Based on these insights, we outline future research directions from both offensive and defensive perspectives, and propose a new combined attack along with a graph purification defense. To support future research, we release the source code of our framework at~\url{https://github.com/CyberAlSec/LLMEGNNRP}.

Are LLM-Enhanced Graph Neural Networks Robust against Poisoning Attacks? | AI Navigate