LLM Benchmark-User Need Misalignment for Climate Change

arXiv cs.CL / 3/30/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The study argues that popular LLM climate benchmarks may not reflect the actual knowledge-seeking behaviors and intents of real users involved in climate decision-making and policy discussions.
  • It proposes a Proactive Knowledge Behaviors Framework and a Topic-Intent-Form taxonomy to characterize human-human and human-AI knowledge seeking and provision patterns.
  • By analyzing climate-related data across different knowledge behavior types, the authors find a substantial misalignment between existing benchmarks and real-world user needs.
  • The work reports that interaction patterns between humans and LLMs more closely resemble human-human interactions than would be expected from benchmark design assumptions.
  • It provides actionable guidance for improving benchmark construction, developing RAG systems, and informing LLM training, along with released code on GitHub.

Abstract

Climate change is a major socio-scientific issue shapes public decision-making and policy discussions. As large language models (LLMs) increasingly serve as an interface for accessing climate knowledge, whether existing benchmarks reflect user needs is critical for evaluating LLM in real-world settings. We propose a Proactive Knowledge Behaviors Framework that captures the different human-human and human-AI knowledge seeking and provision behaviors. We further develop a Topic-Intent-Form taxonomy and apply it to analyze climate-related data representing different knowledge behaviors. Our results reveal a substantial mismatch between current benchmarks and real-world user needs, while knowledge interaction patterns between humans and LLMs closely resemble those in human-human interactions. These findings provide actionable guidance for benchmark design, RAG system development, and LLM training. Code is available at https://github.com/OuchengLiu/LLM-Misalign-Climate-Change.