AI Navigate

Effective Resistance Rewiring: A Simple Topological Correction for Over-Squashing

arXiv cs.LG / 3/13/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper tackles over-squashing in Graph Neural Networks by introducing Effective Resistance Rewiring (ERR), which uses effective resistance as a global signal to identify structural bottlenecks.
  • ERR iteratively adds edges between node pairs with the largest resistance while removing edges with minimal resistance, improving long-range communication under a fixed edge budget.
  • The authors analyze rewiring effects on message propagation by tracking cosine similarity of node embeddings across layers to distinguish improvements from changes in embedding geometry.
  • Experiments on homophilic and heterophilic graphs, including directed DirGCN, show ERR improves connectivity and signal propagation but can accelerate representation mixing in deep models, and combining ERR with normalization like PairNorm stabilizes the trade-off and boosts performance.

Abstract

Graph Neural Networks struggle to capture long-range dependencies due to over-squashing, where information from exponentially growing neighborhoods must pass through a small number of structural bottlenecks. While recent rewiring methods attempt to alleviate this limitation, many rely on local criteria such as curvature, which can overlook global connectivity constraints that restrict information flow. We introduce Effective Resistance Rewiring (ERR), a simple topology correction strategy that uses effective resistance as a global signal to detect structural bottlenecks. ERR iteratively adds edges between node pairs with the largest resistance while removing edges with minimal resistance, strengthening weak communication pathways while controlling graph densification under a fixed edge budget. The procedure is parameter-free beyond the rewiring budget and relies on a single global measure aggregating all paths between node pairs. Beyond predictive performance with GCN models, we analyze how rewiring affects message propagation. By tracking cosine similarity between node embeddings across layers, we examine how the relationship between initial node features and learned representations evolves during message passing, comparing graphs with and without rewiring. This analysis helps determine whether improvements arise from better long-range communication rather than changes in embedding geometry. Experiments on homophilic and heterophilic graphs, including directed settings with DirGCN, reveal a trade-off between over-squashing and oversmoothing, where oversmoothing corresponds to the loss of representation diversity across layers. Resistance-guided rewiring improves connectivity and signal propagation but can accelerate representation mixing in deep models. Combining ERR with normalization techniques such as PairNorm stabilizes this trade-off and improves performance.