Quantum Vision Theory Applied to Audio Classification for Deepfake Speech Detection

arXiv cs.CL / 4/10/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes “Quantum Vision (QV) theory” as a quantum-inspired representation method for deep learning, transforming audio features into “information waves” before classification.
  • It applies the approach to deepfake speech detection by converting STFT, Mel-spectrograms, and MFCCs into information waves via a QV block and training QV-based CNNs and Vision Transformers.
  • Experiments on the ASVSpoof dataset show that QV-CNN and QV-ViT outperform standard CNN/ViT baselines, improving both accuracy and robustness for distinguishing genuine versus spoofed speech.
  • The best reported results include QV-CNN with MFCCs (94.20% accuracy, 9.04% EER) and QV-CNN with Mel-spectrograms (highest accuracy at 94.57%).
  • The authors argue the findings suggest QV theory is a promising direction for “quantum-inspired learning” in audio perception and deepfake detection tasks.

Abstract

We propose Quantum Vision (QV) theory as a new perspective for deep learning-based audio classification, applied to deepfake speech detection. Inspired by particle-wave duality in quantum physics, QV theory is based on the idea that data can be represented not only in its observable, collapsed form, but also as information waves. In conventional deep learning, models are trained directly on these collapsed representations, such as images. In QV theory, inputs are first transformed into information waves using a QV block, and then fed into deep learning models for classification. QV-based models improve performance in image classification compared to their non-QV counterparts. What if QV theory is applied speech spectrograms for audio classification tasks? This is the motivation and novelty of the proposed approach. In this work, Short-Time Fourier Transform (STFT), Mel-spectrograms, and Mel-Frequency Cepstral Coefficients (MFCC) of speech signals are converted into information waves using the proposed QV block and used to train QV-based Convolutional Neural Networks (QV-CNN) and QV-based Vision Transformers (QV-ViT). Extensive experiments are conducted on the ASVSpoof dataset for deepfake speech classification. The results show that QV-CNN and QV-ViT consistently outperform standard CNN and ViT models, achieving higher classification accuracy and improved robustness in distinguishing genuine and spoofed speech. Moreover, the QV-CNN model using MFCC features achieves the best overall performance on the ASVspoof dataset, with an accuracy of 94.20% and an EER of 9.04%, while the QV-CNN with Mel-spectrograms attains the highest accuracy of 94.57%. These findings demonstrate that QV theory is an effective and promising approach for audio deepfake detection and opens new directions for quantum-inspired learning in audio perception tasks.