Precision-Varying Prediction (PVP): Robustifying ASR systems against adversarial attacks

arXiv cs.LG / 3/25/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes Precision-Varying Prediction (PVP) to improve the adversarial robustness of automatic speech recognition (ASR) models by varying precision during inference.
  • It shows that random sampling of the ASR precision during prediction reduces the success rate of adversarial attacks.
  • The authors extend the approach into an adversarial example detection method by comparing ASR outputs produced under different precisions and classifying discrepancies with a simple Gaussian classifier.
  • Experiments report a significant robustness improvement and competitive detection performance across multiple ASR models and adversarial attack types.

Abstract

With the increasing deployment of automated and agentic systems, ensuring the adversarial robustness of automatic speech recognition (ASR) models has become critical. We observe that changing the precision of an ASR model during inference reduces the likelihood of adversarial attacks succeeding. We take advantage of this fact to make the models more robust by simple random sampling of the precision during prediction. Moreover, the insight can be turned into an adversarial example detection strategy by comparing outputs resulting from different precisions and leveraging a simple Gaussian classifier. An experimental analysis demonstrates a significant increase in robustness and competitive detection performance for various ASR models and attack types.