Adversarial Label Invariant Graph Data Augmentations for Out-of-Distribution Generalization

arXiv stat.ML / 4/10/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses out-of-distribution (OoD) generalization under covariate shift, where input distributions change across environments while the underlying concept distribution remains invariant.
  • It proposes RIA (Regularization for Invariance with Adversarial training), which uses adversarial exploration of training environments via adversarial label-invariant data augmentations to avoid collapsing to in-distribution solutions.
  • The method is designed to be compatible with many existing OoD covariate-shift approaches expressible as constrained optimization problems.
  • The authors develop an alternating gradient descent–ascent algorithm to optimize the resulting objective and evaluate it on OoD graph classification tasks with both synthetic and natural shifts.
  • Experimental results reportedly show RIA achieving high accuracy relative to baseline OoD methods, suggesting improved robustness for graph-based representation learning under distribution shift.

Abstract

Out-of-distribution (OoD) generalization occurs when representation learning encounters a distribution shift. This occurs frequently in practice when training and testing data come from different environments. Covariate shift is a type of distribution shift that occurs only in the input data, while the concept distribution stays invariant. We propose RIA - Regularization for Invariance with Adversarial training, a new method for OoD generalization under convariate shift. Motivated by an analogy to Q-learning, it performs an adversarial exploration for training data environments. These new environments are induced by adversarial label invariant data augmentations that prevent a collapse to an in-distribution trained learner. It works with many existing OoD generalization methods for covariate shift that can be formulated as constrained optimization problems. We develop an alternating gradient descent-ascent algorithm to solve the problem, and perform extensive experiments on OoD graph classification for various kinds of synthetic and natural distribution shifts. We demonstrate that our method can achieve high accuracy compared with OoD baselines.