AI Navigate

SignalMC-MED: A Multimodal Benchmark for Evaluating Biosignal Foundation Models on Single-Lead ECG and PPG

arXiv cs.LG / 3/11/2026

Signals & Early TrendsModels & Research

Key Points

  • SignalMC-MED is a new benchmark designed to systematically evaluate biosignal foundation models using synchronized single-lead ECG and PPG data from over 22,000 visits.
  • The benchmark includes 20 clinically relevant prediction tasks covering demographics, emergency disposition, lab value regression, and prior ICD-10 diagnosis detection.
  • Evaluations show domain-specific biosignal foundation models outperform general time-series models, and combining ECG with PPG data leads to better performance than using either modality alone.
  • Longer 10-minute signal durations yield better results than shorter segments, and larger models do not always outperform smaller models.
  • Hand-crafted ECG domain features remain valuable and complement learned representations, highlighting practical insights for deploying biosignal models effectively.

Computer Science > Machine Learning

arXiv:2603.09940 (cs)
[Submitted on 10 Mar 2026]

Title:SignalMC-MED: A Multimodal Benchmark for Evaluating Biosignal Foundation Models on Single-Lead ECG and PPG

View a PDF of the paper titled SignalMC-MED: A Multimodal Benchmark for Evaluating Biosignal Foundation Models on Single-Lead ECG and PPG, by Fredrik K. Gustafsson and 5 other authors
View PDF
Abstract:Recent biosignal foundation models (FMs) have demonstrated promising performance across diverse clinical prediction tasks, yet systematic evaluation on long-duration multimodal data remains limited. We introduce SignalMC-MED, a benchmark for evaluating biosignal FMs on synchronized single-lead electrocardiogram (ECG) and photoplethysmogram (PPG) data. Derived from the MC-MED dataset, SignalMC-MED comprises 22,256 visits with 10-minute overlapping ECG and PPG signals, and includes 20 clinically relevant tasks spanning prediction of demographics, emergency department disposition, laboratory value regression, and detection of prior ICD-10 diagnoses. Using this benchmark, we perform a systematic evaluation of representative time-series and biosignal FMs across ECG-only, PPG-only, and ECG + PPG settings. We find that domain-specific biosignal FMs consistently outperform general time-series models, and that multimodal ECG + PPG fusion yields robust improvements over unimodal inputs. Moreover, using the full 10-minute signal consistently outperforms shorter segments, and larger model variants do not reliably outperform smaller ones. Hand-crafted ECG domain features provide a strong baseline and offer complementary value when combined with learned FM representations. Together, these results establish SignalMC-MED as a standardized benchmark and provide practical guidance for evaluating and deploying biosignal FMs.
Comments:
Subjects: Machine Learning (cs.LG)
Cite as: arXiv:2603.09940 [cs.LG]
  (or arXiv:2603.09940v1 [cs.LG] for this version)
  https://doi.org/10.48550/arXiv.2603.09940
Focus to learn more
arXiv-issued DOI via DataCite

Submission history

From: Fredrik K. Gustafsson [view email]
[v1] Tue, 10 Mar 2026 17:32:28 UTC (23,120 KB)
Full-text links:

Access Paper:

    View a PDF of the paper titled SignalMC-MED: A Multimodal Benchmark for Evaluating Biosignal Foundation Models on Single-Lead ECG and PPG, by Fredrik K. Gustafsson and 5 other authors
  • View PDF
  • TeX Source
Current browse context:
cs.LG
< prev   |   next >
Change to browse by:
cs

References & Citations

export BibTeX citation Loading...

BibTeX formatted citation

×
Data provided by:

Bookmark

BibSonomy logo Reddit logo
Bibliographic Tools

Bibliographic and Citation Tools

Bibliographic Explorer Toggle
Bibliographic Explorer (What is the Explorer?)
Connected Papers Toggle
Connected Papers (What is Connected Papers?)
Litmaps Toggle
Litmaps (What is Litmaps?)
scite.ai Toggle
scite Smart Citations (What are Smart Citations?)
Code, Data, Media

Code, Data and Media Associated with this Article

alphaXiv Toggle
alphaXiv (What is alphaXiv?)
Links to Code Toggle
CatalyzeX Code Finder for Papers (What is CatalyzeX?)
DagsHub Toggle
DagsHub (What is DagsHub?)
GotitPub Toggle
Gotit.pub (What is GotitPub?)
Huggingface Toggle
Hugging Face (What is Huggingface?)
Links to Code Toggle
Papers with Code (What is Papers with Code?)
ScienceCast Toggle
ScienceCast (What is ScienceCast?)
Demos

Demos

Replicate Toggle
Replicate (What is Replicate?)
Spaces Toggle
Hugging Face Spaces (What is Spaces?)
Spaces Toggle
TXYZ.AI (What is TXYZ.AI?)
Related Papers

Recommenders and Search Tools

Link to Influence Flower
Influence Flower (What are Influence Flowers?)
Core recommender toggle
CORE Recommender (What is CORE?)
IArxiv recommender toggle
IArxiv Recommender (What is IArxiv?)
About arXivLabs

arXivLabs: experimental projects with community collaborators

arXivLabs is a framework that allows collaborators to develop and share new arXiv features directly on our website.

Both individuals and organizations that work with arXivLabs have embraced and accepted our values of openness, community, excellence, and user data privacy. arXiv is committed to these values and only works with partners that adhere to them.

Have an idea for a project that will add value for arXiv's community? Learn more about arXivLabs.