AI Navigate

LWM-Temporal: Sparse Spatio-Temporal Attention for Wireless Channel Representation Learning

arXiv cs.LG / 3/12/2026

📰 NewsSignals & Early TrendsModels & Research

Key Points

  • The paper introduces LWM-Temporal, a new model in the Large Wireless Models family designed to learn universal channel embeddings for spatiotemporal wireless channels.
  • It introduces Sparse Spatio-Temporal Attention (SSTA) in the angle–delay–time domain to limit interactions to physically plausible neighborhoods and reduce attention complexity by about an order of magnitude while maintaining geometry-consistent dependencies.
  • It uses a self-supervised, physics-informed masking curriculum that simulates occlusions, pilot sparsity, and measurement impairments to learn transferable representations.
  • Experiments show improvements in channel prediction across mobility regimes, especially for long horizons and limited finetuning data, highlighting geometry-aware architectures and pretraining for transferable spatiotemporal wireless representations.

Abstract

LWM-Temporal is a new member of the Large Wireless Models (LWM) family that targets the spatiotemporal nature of wireless channels. Designed as a task-agnostic foundation model, LWM-Temporal learns universal channel embeddings that capture mobility-induced evolution and are reusable across various downstream tasks. To achieve this objective, LWM-Temporal operates in the angle-delay-time domain and introduces Sparse Spatio-Temporal Attention (SSTA), a propagation-aligned attention mechanism that restricts interactions to physically plausible neighborhoods, reducing attention complexity by an order of magnitude while preserving geometry-consistent dependencies. LWM-Temporal is pretrained in a self-supervised manner using a physics-informed masking curriculum that emulates realistic occlusions, pilot sparsity, and measurement impairments. Experimental results on channel prediction across multiple mobility regimes show consistent improvements over strong baselines, particularly under long horizons and limited fine-tuning data, highlighting the importance of geometry-aware architectures and geometry-consistent pretraining for learning transferable spatiotemporal wireless representations.