U-Cast: A Surprisingly Simple and Efficient Frontier Probabilistic AI Weather Forecaster

arXiv cs.LG / 4/13/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces U-Cast, a probabilistic weather forecasting model that uses a standard U-Net backbone instead of specialized SOTA architectures, targeting frontier performance with lower complexity.
  • U-Cast is trained via a simple two-stage curriculum: deterministic pre-training on MAE followed by short probabilistic fine-tuning on CRPS using Monte Carlo Dropout for uncertainty estimation.
  • The authors report that U-Cast matches or exceeds the probabilistic skill of GenCast and IFS ENS at 1.5° resolution while cutting training compute by over 10× and reducing inference latency by over 10× versus diffusion-based models.
  • Training reportedly completes in under 12 H200 GPU-days, and it produces a 60-step ensemble forecast in about 11 seconds, indicating strong efficiency for operational-like use.
  • The results argue that scalable general-purpose architectures plus efficient training curricula can democratize frontier probabilistic weather modeling, and the code is released on GitHub.

Abstract

AI-based weather forecasting now rivals traditional physics-based ensembles, but state-of-the-art (SOTA) models rely on specialized architectures and massive computational budgets, creating a high barrier to entry. We demonstrate that such complexity is unnecessary for frontier performance. We introduce U-Cast, a probabilistic forecaster built on a standard U-Net backbone trained with a simple recipe: deterministic pre-training on Mean Absolute Error followed by short probabilistic fine-tuning on the Continuous Ranked Probability Score (CRPS) using Monte Carlo Dropout for stochasticity. As a result, our model matches or exceeds the probabilistic skill of GenCast and IFS ENS at 1.5^\circ\$ resolution while reducing training compute by over 10\times compared to leading CRPS-based models and inference latency by over 10\times$ compared to diffusion-based models. U-Cast trains in under 12 H200 GPU-days and generates a 60-step ensemble forecast in 11 seconds. These results suggest that scalable, general-purpose architectures paired with efficient training curricula can match complex domain-specific designs at a fraction of the cost, opening the training of frontier probabilistic weather models to the broader community. Our code is available at: https://github.com/Rose-STL-Lab/u-cast.