Optimizing Multi-Agent Weather Captioning via Text Gradient Descent: A Training-Free Approach with Consensus-Aware Gradient Fusion

arXiv cs.CL / 3/24/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces WeatherTGD, a training-free multi-agent framework for generating domain-specific, interpretable natural-language weather captions from weather time-series data using Text Gradient Descent (TGD).
  • It uses three specialized LLM agents—a Statistical Analyst, a Physics Interpreter, and a Meteorology Expert—to produce domain-relevant textual gradients from the same observations.
  • A new Consensus-Aware Gradient Fusion method aggregates gradients to capture shared signals while retaining each agent’s distinct domain perspective.
  • The fused gradients drive an iterative, gradient-descent-like caption refinement process without requiring additional model training.
  • Experiments on real meteorological datasets reportedly improve both automated LLM evaluations and human expert assessments, outperforming multi-agent baselines while keeping computation efficient via parallel execution.

Abstract

Generating interpretable natural language captions from weather time series data remains a significant challenge at the intersection of meteorological science and natural language processing. While recent advances in Large Language Models (LLMs) have demonstrated remarkable capabilities in time series forecasting and analysis, existing approaches either produce numerical predictions without human-accessible explanations or generate generic descriptions lacking domain-specific depth. We introduce WeatherTGD, a training-free multi-agent framework that reinterprets collaborative caption refinement through the lens of Text Gradient Descent (TGD). Our system deploys three specialized LLM agents including a Statistical Analyst, a Physics Interpreter, and a Meteorology Expert that generate domain-specific textual gradients from weather time series observations. These gradients are aggregated through a novel Consensus-Aware Gradient Fusion mechanism that extracts common signals while preserving unique domain perspectives. The fused gradients then guide an iterative refinement process analogous to gradient descent, where each LLM-generated feedback signal updates the caption toward an optimal solution. Experiments on real-world meteorological datasets demonstrate that WeatherTGD achieves significant improvements in both LLM-based evaluation and human expert evaluation, substantially outperforming existing multi-agent baselines while maintaining computational efficiency through parallel agent execution.