Marchuk: Efficient Global Weather Forecasting from Mid-Range to Sub-Seasonal Scales via Flow Matching

arXiv cs.LG / 3/26/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces Marchuk, a generative latent flow-matching model designed to improve global weather forecasting from mid-range to sub-seasonal horizons (up to ~30 days).
  • Marchuk conditions on current-day weather maps and autoregressively predicts future daily weather maps in a learned latent space, targeting the persistent challenge that conventional models lose skill beyond ~15 days.
  • The authors report architectural changes—replacing RoPE with trainable positional embeddings and expanding the temporal context window—to better model and propagate long-range temporal dependencies.
  • Despite being relatively compact (276M parameters), Marchuk achieves performance comparable to LaDCast (1.6B parameters) while delivering substantially faster inference, emphasizing computational efficiency.
  • The inference code and model are open-sourced, enabling researchers and practitioners to test and build upon the approach.

Abstract

Accurate subseasonal weather forecasting remains a major challenge due to the inherently chaotic nature of the atmosphere, which limits the predictive skill of conventional models beyond the mid-range horizon (approximately 15 days). In this work, we present \textit{Marchuk}, a generative latent flow-matching model for global weather forecasting spanning mid-range to subseasonal timescales, with prediction horizons of up to 30 days. Marchuk conditions on current-day weather maps and autoregressively predicts subsequent days' weather maps within the learned latent space. We replace rotary positional encodings (RoPE) with trainable positional embeddings and extend the temporal context window, which together enhance the model's ability to represent and propagate long-range temporal dependencies during latent forecasting. Marchuk offers two key advantages: high computational efficiency and strong predictive performance. Despite its compact architecture of only 276 million parameters, the model achieves performance comparable to LaDCast, a substantially larger model with 1.6 billion parameters, while operating at significantly higher inference speeds. We open-source our inference code and model at: https://v-gen-ai.github.io/Marchuk/