AI Navigate

Evaluating Black-Box Vulnerabilities with Wasserstein-Constrained Data Perturbations

arXiv cs.LG / 3/18/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper tackles evaluating vulnerabilities of black-box ML models by using Optimal Transport and Wasserstein distance to study how changes in input variable distributions affect model behavior.
  • It proposes finding the closest distribution, in the Wasserstein sense, that satisfies a specified perturbation constraint and analyzes its impact on model outputs.
  • It provides convergence results for the projected distribution, establishing theoretical guarantees for the method.
  • It demonstrates the method on real-world regression and classification datasets, illustrating practical use in robustness analysis.

Abstract

The massive use of Machine Learning (ML) tools in industry comes with critical challenges, such as the lack of explainable models and the use of black-box algorithms. We address this issue by applying Optimal Transport theory in the analysis of responses of ML models to variations in the distribution of input variables. We find the closest distribution, in the Wasserstein sense, that satisfies a given constraintt and examine its impact on model behavior. Furthermore, we establish convergence results for this projected distribution and demonstrate our approach using examples and real-world datasets in both regression and classification settings.