Can we generate portable representations for clinical time series data using LLMs?
arXiv cs.LG / 2026/3/26
💬 オピニオンSignals & Early TrendsIdeas & Deep AnalysisModels & Research
要点
- The paper investigates whether frozen LLMs can generate portable patient embeddings from irregular ICU time series so that models trained in one hospital generalize to others with minimal retraining.
- The method converts ICU time series into concise natural-language summaries using a frozen LLM, then embeds those summaries with a frozen text embedding model to produce fixed-length vectors for downstream forecasting and classification.
- Experiments across MIMIC-IV, HIRID, and PPICU show competitive performance versus several common in-hospital or representation-learning baselines, with smaller relative drops under cross-hospital transfer.
- Structured prompt design is found to be important for reducing variance in downstream predictive performance, improving robustness without changing mean accuracy.
- The resulting portable representations improve few-shot learning while not increasing demographic recoverability of age/sex, suggesting limited additional privacy risk.



