Benign Overfitting in Adversarial Training for Vision Transformers

arXiv cs.LG / 4/22/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper provides the first theoretical analysis of how adversarial training affects Vision Transformers (ViTs) using simplified ViT architectures, addressing a gap in existing theory.
  • It argues that under a specific signal-to-noise ratio condition and a moderate perturbation budget, adversarial training can yield nearly zero robust training loss and low robust generalization error in certain regimes.
  • The work reports a “benign overfitting” effect—strong generalization despite overfitting—that had previously been observed primarily for CNNs with adversarial training.
  • Experiments on both synthetic and real-world datasets are used to validate the theoretical results and support the proposed conditions.
  • Overall, the study links adversarial robustness in ViTs to training dynamics that resemble those known from CNN theory, offering new guidance for understanding and designing robust ViT training.

Abstract

Despite the remarkable success of Vision Transformers (ViTs) across a wide range of vision tasks, recent studies have revealed that they remain vulnerable to adversarial examples, much like Convolutional Neural Networks (CNNs). A common empirical defense strategy is adversarial training, yet the theoretical underpinnings of its robustness in ViTs remain largely unexplored. In this work, we present the first theoretical analysis of adversarial training under simplified ViT architectures. We show that, when trained under a signal-to-noise ratio that satisfies a certain condition and within a moderate perturbation budget, adversarial training enables ViTs to achieve nearly zero robust training loss and robust generalization error under certain regimes. Remarkably, this leads to strong generalization even in the presence of overfitting, a phenomenon known as \emph{benign overfitting}, previously only observed in CNNs (with adversarial training). Experiments on both synthetic and real-world datasets further validate our theoretical findings.