PEANUT: Perturbations by Eigenvalue Alignment for Attacking GNNs Under Topology-Driven Message Passing

arXiv cs.LG / 3/30/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that GNNs that rely on explicit topology representations (adjacency matrix or Laplacian) are vulnerable to small graph-structure perturbations that can substantially change model outputs in deployment settings.
  • It proposes PEANUT, a simple gradient-free, restricted black-box evasion attack that injects virtual nodes to exploit eigenvalue alignment with the GNN’s topology-driven message passing mechanism.
  • PEANUT is designed to be practical for real attackers because it operates at inference time, avoids lengthy iterative optimization or parameter learning, and does not require training surrogate models.
  • The attack does not need node features on the injected nodes (zero features still work), underscoring that connectivity/structure alone can be sufficient to degrade GNN performance.
  • Experiments on multiple real-world datasets across three graph tasks reportedly show PEANUT’s effectiveness despite its simplicity.

Abstract

Graph Neural Networks (GNNs) have achieved remarkable performance on tasks involving relational data. However, small perturbations to the graph structure can significantly alter GNN outputs, raising concerns about their robustness in real-world deployments. In this work, we explore the core vulnerability of GNNs which explicitly consume graph topology in the form of the adjacency matrix or Laplacian as a means for message passing, and propose PEANUT, a simple, gradient-free, restricted black-box attack that injects virtual nodes to capitalize on this vulnerability. PEANUT is a injection based attack, which is widely considered to be more practical and realistic scenario than graph modification attacks, where the attacker is able to modify the original graph structure directly. Our method works at the inference phase, making it an evasion attack, and is applicable almost immediately, since it does not involve lengthy iterative optimizations or parameter learning, which add computational and time overhead, or training surrogate models, which are susceptible to failure due to differences in model priors and generalization capabilities. PEANUT also does not require any features on the injected node and consequently demonstrates that GNN performance can be significantly deteriorated even with injected nodes with zeros for features, highlighting the significance of effectively designed connectivity in such attacks. Extensive experiments on real-world datasets across three graph tasks demonstrate the effectiveness of our attack despite its simplicity.