On the Adversarial Robustness of Learning-based Conformal Novelty Detection

arXiv stat.ML / 4/3/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper investigates how learning-based conformal novelty detection methods with finite-sample FDR control (AdaDetect and a one-class classifier-based approach) degrade under adversarial perturbations.
  • It introduces an oracle attack framework for AdaDetect and derives an upper bound on the worst-case FDR degradation, connecting attack severity to the statistical cost of adversaries.
  • The authors propose a practical black-box attack that requires only query access to the frameworks’ output labels, enabling empirical evaluation without full model knowledge.
  • Experiments across synthetic and real-world datasets show that adversarial noise can substantially raise FDR while keeping detection power high, revealing vulnerabilities in current error-controlled novelty detection.
  • The findings motivate the need for new, more adversarially robust alternatives to maintain reliable novelty detection guarantees in adversarial settings.

Abstract

This paper studies the adversarial robustness of conformal novelty detection. In particular, we focus on two powerful learning-based frameworks that come with finite-sample false discovery rate (FDR) control: one is AdaDetect (by Marandon et al., 2024) that is based on the positive-unlabeled classifier, and the other is a one-class classifier-based approach (by Bates et al., 2023). While they provide rigorous statistical guarantees under benign conditions, their behavior under adversarial perturbations remains underexplored. We first formulate an oracle attack setup, under the AdaDetect formulation, that quantifies the worst-case degradation of FDR, deriving an upper bound that characterizes the statistical cost of attacks. This idealized formulation directly motivates a practical and effective attack scheme that only requires query access to the output labels of both frameworks. Coupling these formulations with two popular and complementary black-box adversarial algorithms, we systematically evaluate the vulnerability of both frameworks on synthetic and real-world datasets. Our results show that adversarial perturbations can significantly increase the FDR while maintaining high detection power, exposing fundamental limitations of current error-controlled novelty detection methods and motivating the development of more robust alternatives.