AI Navigate

PowerModelsGAT-AI: Physics-Informed Graph Attention for Multi-System Power Flow with Continual Learning

arXiv cs.LG / 3/19/2026

📰 NewsModels & Research

Key Points

  • PowerModelsGAT-AI presents a physics-informed graph attention network for real-time AC power flow that predicts bus voltages and generator injections, addressing slow Newton-Raphson solvers under stressed conditions.
  • It employs bus-type-aware masking and learned-weight balancing across multiple loss terms, including a power-mismatch penalty, to handle heterogeneous bus types and objectives.
  • The model is evaluated on 14 benchmark systems (4–6,470 buses) with a unified model trained on 13 under N-2 outages, achieving an average NMSE of 0.89% for voltage magnitudes and R^2 > 0.99 for voltage angles.
  • In continual learning experiments, experience replay and elastic weight consolidation nearly eliminate forgetting when adapting to a new 1,354-bus system, keeping base-system error increases below 2%.
  • Interpretability analyses show attention weights correlate with physical parameters (susceptance r = 0.38; thermal limits r = 0.22), indicating the model captures established power-flow relationships.

Abstract

Solving the alternating current power flow equations in real time is essential for secure grid operation, yet classical Newton-Raphson solvers can be slow under stressed conditions. Existing graph neural networks for power flow are typically trained on a single system and often degrade on different systems. We present PowerModelsGAT-AI, a physics-informed graph attention network that predicts bus voltages and generator injections. The model uses bus-type-aware masking to handle different bus types and balances multiple loss terms, including a power-mismatch penalty, using learned weights. We evaluate the model on 14 benchmark systems (4 to 6,470 buses) and train a unified model on 13 of these under N-2 (two-branch outage) conditions, achieving an average normalized mean absolute error of 0.89% for voltage magnitudes and R^2 > 0.99 for voltage angles. We also show continual learning: when adapting a base model to a new 1,354-bus system, standard fine-tuning causes severe forgetting with error increases exceeding 1000% on base systems, while our experience replay and elastic weight consolidation strategy keeps error increases below 2% and in some cases improves base-system performance. Interpretability analysis shows that learned attention weights correlate with physical branch parameters (susceptance: r = 0.38; thermal limits: r = 0.22), and feature importance analysis supports that the model captures established power flow relationships.