AI Navigate

Contract And Conquer: How to Provably Compute Adversarial Examples for a Black-Box Model?

arXiv cs.LG / 3/12/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes Contract And Conquer (CAC), a method to provably compute adversarial examples for neural networks in a black-box setting.
  • CAC uses knowledge distillation on an expanding distillation dataset and a precise contraction of the adversarial search space to enable provable guarantees.
  • The authors prove a transferability guarantee: CAC can produce an adversarial example for the black-box model within a fixed number of iterations.
  • Experiments on ImageNet, including vision transformers, show CAC outperforms existing black-box attack methods.

Abstract

Black-box adversarial attacks are widely used as tools to test the robustness of deep neural networks against malicious perturbations of input data aimed at a specific change in the output of the model. Such methods, although they remain empirically effective, usually do not guarantee that an adversarial example can be found for a particular model. In this paper, we propose Contract And Conquer (CAC), an approach to provably compute adversarial examples for neural networks in a black-box manner. The method is based on knowledge distillation of a black-box model on an expanding distillation dataset and precise contraction of the adversarial example search space. CAC is supported by the transferability guarantee: we prove that the method yields an adversarial example for the black-box model within a fixed number of algorithm iterations. Experimentally, we demonstrate that the proposed approach outperforms existing state-of-the-art black-box attack methods on ImageNet dataset for different target models, including vision transformers.