TsallisPGD: Adaptive Gradient Weighting for Adversarial Attacks on Semantic Segmentation

arXiv cs.CV / 5/6/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • Semantic segmentation adversarial attacks are harder than classification because attackers must jointly alter thousands of pixel predictions, making standard cross-entropy–based losses poorly aligned with the optimization dynamics.
  • The paper introduces TsallisPGD, an adversarial attack using Tsallis cross-entropy (a q-parameterized generalization of cross-entropy) to adaptively reshape the gradient landscape via control of gradient concentration across pixels.
  • By sweeping the Tsallis parameter q, the method targets pixels across different confidence levels, and the authors show that no single fixed q works best across datasets, architectures, and perturbation budgets.
  • They propose a dynamic q-schedule that changes q during optimization, and experiments on Cityscapes, Pascal VOC, and ADE20K show TsallisPGD achieves the best average attack ranking and outperforms several prior PGD variants on both standard and robustness-focused models.
  • TsallisPGD with a schedule selected once via validation reduces accuracy and mIoU more effectively than CEPGD, SegPGD, CosPGD, JSPGD, and MaskedPGD, indicating improved attack strength under varied settings.

Abstract

Attacking semantic segmentation models is significantly harder than image classification models because an attacker must flip thousands of pixel predictions simultaneously. Standard pixel-wise cross-entropy (CE) is ill-suited to this setting: it tends to overemphasize already-misclassified pixels, which slows optimization and overstates model robustness. To address these issues, we introduce TsallisPGD, an adversarial attack built on the Tsallis cross-entropy, a generalization of CE parameterized by q, which adaptively reshapes the gradient landscape by controlling gradient concentration across pixels. By varying q, we steer the attack toward pixels at different confidence levels. We first show that no single fixed-q is universally optimal, as its effectiveness depends on the dataset, model architecture, and perturbation budget. Motivated by this, we propose a dynamic q-schedule that sweeps q during optimization. Extensive experiments on Cityscapes, Pascal VOC, and ADE20K show that TsallisPGD, using a single validation-selected schedule, achieves the best average attack rank across all evaluated settings and improves over CEPGD, SegPGD, CosPGD, JSPGD, and MaskedPGD in reducing accuracy and mIoU on both standard and robust models.

TsallisPGD: Adaptive Gradient Weighting for Adversarial Attacks on Semantic Segmentation | AI Navigate