Stress Classification from ECG Signals Using Vision Transformer
arXiv cs.AI / 3/31/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper proposes a vision-transformer approach for multilevel stress classification from ECG by converting raw signals into 2D STFT spectrograms and feeding patched inputs to a transformer encoder.
- It addresses the challenge of inter-subject variability using leave-one-subject-out cross-validation (LOSOCV) on WESAD and RML datasets, comparing against 1D CNN and ResNet-18 baselines.
- Experimental results indicate the vision transformer outperforms CNN-based models and prior state-of-the-art methods, showing stronger robustness to intersubject differences.
- The method is end-to-end and avoids handcrafted features, learning representations directly from spectrogram patch data.
- Reported performance includes 71.01% (RML) and 76.7% (WESAD) accuracy for three-class classification and 88.3% accuracy for binary classification on WESAD.
Related Articles

Black Hat Asia
AI Business
[D] How does distributed proof of work computing handle the coordination needs of neural network training?
Reddit r/MachineLearning

Claude Code's Entire Source Code Was Just Leaked via npm Source Maps — Here's What's Inside
Dev.to

BYOK is not just a pricing model: why it changes AI product trust
Dev.to

AI Citation Registries and Identity Persistence Across Records
Dev.to