AI Navigate

Large language models show fragile cognitive reasoning about human emotions

arXiv cs.CL / 3/16/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The CoRE benchmark is introduced to probe implicit cognitive structures in LLMs for interpreting emotionally charged situations based on cognitive appraisal theory.
  • The study finds LLMs capture systematic relations between cognitive appraisals and emotions but show misalignment with human judgments and instability across contexts.
  • The evaluation includes alignment with human patterns, internal consistency, cross-model generalization, and robustness to contextual variation.
  • The results highlight fragility in LLM-based emotion reasoning and have implications for affective computing research and how we evaluate AI emotion understanding.

Abstract

Affective computing seeks to support the holistic development of artificial intelligence by enabling machines to engage with human emotion. Recent foundation models, particularly large language models (LLMs), have been trained and evaluated on emotion-related tasks, typically using supervised learning with discrete emotion labels. Such evaluations largely focus on surface phenomena, such as recognizing expressed or evoked emotions, leaving open whether these systems reason about emotion in cognitively meaningful ways. Here we ask whether LLMs can reason about emotions through underlying cognitive dimensions rather than labels alone. Drawing on cognitive appraisal theory, we introduce CoRE, a large-scale benchmark designed to probe the implicit cognitive structures LLMs use when interpreting emotionally charged situations. We assess alignment with human appraisal patterns, internal consistency, cross-model generalization, and robustness to contextual variation. We find that LLMs capture systematic relations between cognitive appraisals and emotions but show misalignment with human judgments and instability across contexts.