AI Navigate

Learning Transferable Sensor Models via Language-Informed Pretraining

arXiv cs.AI / 3/13/2026

📰 NewsSignals & Early TrendsModels & Research

Key Points

  • Introduces SLIP, a framework for learning language-aligned sensor representations that generalize across diverse sensor setups and input configurations.
  • Combines contrastive alignment with sensor-conditioned captioning to enable both discriminative understanding and generative reasoning.
  • Enables inference-time handling of different temporal resolutions and variable-length inputs without retraining by using cross-attention with a pretrained decoder-only language model and a flexible patch-embedder.
  • Demonstrates strong zero-shot transfer, sensor captioning, and sensor-based question answering across 11 datasets, achieving 77.14% average linear probing accuracy and 64.83% QA accuracy.
  • The project is open-source and addresses limitations of prior methods that rely on fixed sensor configurations.

Abstract

Modern sensing systems generate large volumes of unlabeled multivariate time-series data. This abundance of unlabeled data makes self-supervised learning (SSL) a natural approach for learning transferable representations. However, most existing approaches are optimized for reconstruction or forecasting objectives and often fail to capture the semantic structure required for downstream classification and reasoning tasks. While recent sensor-language alignment methods improve semantic generalization through captioning and zero-shot transfer, they are limited to fixed sensor configurations, such as predefined channel sets, signal lengths, or temporal resolutions, which hinders cross-domain applicability. To address these gaps, we introduce \textbf{SLIP} (\textbf{S}ensor \textbf{L}anguage-\textbf{I}nformed \textbf{P}retraining), an open-source framework for learning language-aligned representations that generalize across diverse sensor setups. SLIP integrates contrastive alignment with sensor-conditioned captioning, facilitating both discriminative understanding and generative reasoning. By repurposing a pretrained decoder-only language model via cross-attention and introducing an elegant, flexible patch-embedder, SLIP supports different temporal resolutions and variable-length input at inference time without additional retraining. Across 11 datasets, SLIP demonstrates superior performance in zero-shot transfer, signal captioning, and question answering. It achieves a 77.14% average linear-probing accuracy, a 5.93% relative improvement over strong baselines, and reaches 64.83% accuracy in sensor-based question answering.