BeliefShift: Benchmarking Temporal Belief Consistency and Opinion Drift in LLM Agents

arXiv cs.CL / 3/26/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that current LLM-agent memory benchmarks treat user information as static facts, but real users change their minds over long interactions, making belief dynamics such as opinion drift and confirmation bias important to evaluate.
  • BeliefShift is introduced as a longitudinal, human-annotated benchmark (2,400 multi-session trajectories) with three tracks focused on Temporal Belief Consistency, Contradiction Detection, and Evidence-Driven Revision across domains like health, politics, personal values, and product preferences.
  • The authors evaluate seven LLMs (including GPT-4o, Claude 3.5 Sonnet, Gemini 1.5 Pro, LLaMA-3, and Mistral-Large) in both zero-shot and RAG settings and find a trade-off between aggressive personalization resisting drift and grounded models failing to perform legitimate belief updates.
  • Four new metrics—Belief Revision Accuracy (BRA), Drift Coherence Score (DCS), Contradiction Resolution Rate (CRR), and Evidence Sensitivity Index (ESI)—are proposed to measure different aspects of belief change behavior.
  • The benchmark and metrics are intended to better quantify how LLM agents revise beliefs over time, not just whether they retrieve stored facts.

Abstract

LLMs are increasingly used as long-running conversational agents, yet every major benchmark evaluating their memory treats user information as static facts to be stored and retrieved. That's the wrong model. People change their minds, and over extended interactions, phenomena like opinion drift, over-alignment, and confirmation bias start to matter a lot. BeliefShift introduces a longitudinal benchmark designed specifically to evaluate belief dynamics in multi-session LLM interactions. It covers three tracks: Temporal Belief Consistency, Contradiction Detection, and Evidence-Driven Revision. The dataset includes 2,400 human-annotated multi-session interaction trajectories spanning health, politics, personal values, and product preferences. We evaluate seven models including GPT-4o, Claude 3.5 Sonnet, Gemini 1.5 Pro, LLaMA-3, and Mistral-Large under zero-shot and retrieval-augmented generation (RAG) settings. Results reveal a clear trade-off: models that personalize aggressively resist drift poorly, while factually grounded models miss legitimate belief updates. We further introduce four novel evaluation metrics: Belief Revision Accuracy (BRA), Drift Coherence Score (DCS), Contradiction Resolution Rate (CRR), and Evidence Sensitivity Index (ESI).