CGM-JEPA: Learning Consistent Continuous Glucose Monitor Representations via Predictive Self-Supervised Pretraining

arXiv cs.LG / 5/5/2026

📰 NewsDeveloper Stack & InfrastructureModels & Research

Key Points

  • The paper introduces CGM-JEPA, a self-supervised pretraining framework that learns masked latent representations for Continuous Glucose Monitoring rather than predicting raw glucose values.
  • It targets two coupled challenges in CGM representation learning: multi-view physiological states that don’t transfer across modalities/settings, and baselines that degrade inconsistently under modality shifts.
  • A variant called X-CGM-JEPA adds a masked Glucodensity cross-view objective to incorporate complementary distributional information from additional summaries.
  • Pretraining uses ~389k unlabeled CGM readings from 228 subjects and is evaluated across three clinical regimes (cohort generalization, venous-to-CGM transfer, and home CGM) with AUROC showing CGM-JEPA ranks first/second for both endpoints.
  • Compared with the strongest baseline, X-CGM-JEPA improves AUROC by up to +6.5 percentage points (cohort generalization) and +3.6 points (venous-to-CGM transfer), and it reduces AUROC gaps across subgroups under modality shift.

Abstract

Continuous Glucose Monitoring (CGM) can detect early metabolic subphenotypes (insulin resistance, IR; \beta-cell dysfunction), but population-scale deployment faces two coupled problems. First, the same physiological state appears through multiple views (CGM time series, venous OGTT, Glucodensity summaries), so single-view representations fail to transfer when deployment shifts the modality or setting. Second, baselines perform inconsistently across these shifts. Both problems point to one remedy: representations that abstract away from any single view to capture higher-level temporal and distributional structure. We propose CGM-JEPA, a self-supervised pretraining framework which predicts masked latent representations rather than raw values, yielding abstraction that transfers across modalities. X-CGM-JEPA adds a masked Glucodensity cross-view objective for complementary distributional information. We pretrain on \sim389k unlabeled CGM readings from 228 subjects and evaluate on two clinical cohorts (N=27 and N=17 public-release subsets) across three regimes (cohort generalization, venous-to-CGM transfer, home CGM) under 20-iteration \times 2-fold cross-validation. X-CGM-JEPA ranks first or second on AUROC for both endpoints across all three regimes while no baseline does, exceeding the strongest baseline by up to +6.5 pp in cohort generalization and +3.6 pp in venous-to-CGM transfer (paired Wilcoxon, p<0.001). Under modality shift, it matches mean AUROC while redistributing toward weaker subgroups (ethnicity AUROC gap shrinks 25-54%); on sparse in-domain venous data, the distributional view lifts label-aware clustering (ARI +39\%, NMI +40\%). Code and weights: https://github.com/cruiseresearchgroup/CGM-JEPA