AI Navigate

Quantum-Enhanced Vision Transformer for Flood Detection using Remote Sensing Imagery

arXiv cs.LG / 3/17/2026

📰 NewsModels & Research

Key Points

  • The paper presents a Quantum-Enhanced Vision Transformer for flood detection using remote sensing imagery, merging transformer-based global context with quantum feature extraction.
  • It uses a hybrid architecture with parallel pathways: a ViT backbone and a 4-qubit parameterized quantum circuit for localized feature mapping, whose representations are fused for binary classification.
  • Experimental results show the quantum-hybrid model outperforms a classical ViT baseline, with accuracy rising from 84.48% to 94.47% and F1-score from 0.841 to 0.944.
  • The work demonstrates the potential of quantum-classical hybrids to enhance precision in hydrological monitoring and earth observation applications.

Abstract

Reliable flood detection is critical for disaster management, yet classical deep learning models often struggle with the high-dimensional, nonlinear complexities inherent in remote sensing data. To mitigate these limitations, we introduced a novel Quantum-Enhanced Vision Transformer (ViT) that synergizes the global context-awareness of transformers with the expressive feature extraction capabilities of quantum computing. Using remote sensing imagery, we developed a hybrid architecture that processes inputs through parallel pathways, a ViT backbone and a quantum branch utilizing a 4-qubit parameterized quantum circuit for localized feature mapping. These distinct representations were fused to optimize binary classification. Results showed that the proposed hybrid model significantly outperformed a classical ViT baseline, increased overall accuracy from 84.48% to 94.47% and the F1-score from 0.841 to 0.944. Notably, the quantum integration substantially improved discriminative power in complex terrains for both class. These findings validate the potential of quantum-classical hybrid models to enhance precision in hydrological monitoring and earth observation applications.