How Hypocritical Is Your LLM judge? Listener-Speaker Asymmetries in the Pragmatic Competence of Large Language Models

arXiv cs.CL / 4/20/2026

📰 NewsModels & Research

Key Points

  • The paper investigates whether LLM performance as a “pragmatic judge” (listener) aligns with its performance as a “pragmatic speaker” for pragmatic competence tasks.
  • By directly comparing LLMs in both roles, the study evaluates multiple open-weight and proprietary models across three pragmatic settings.
  • Results show a consistent asymmetry: many LLMs are substantially better at judging the appropriateness of outputs than at generating pragmatically appropriate language.
  • The findings indicate that pragmatic evaluation and pragmatic generation are only weakly aligned in current LLMs, suggesting a need for more integrated evaluation approaches.

Abstract

Large language models (LLMs) are increasingly studied as repositories of linguistic knowledge. In this line of work, models are commonly evaluated both as generators of language and as judges of linguistic output, yet these two roles are rarely examined in direct relation to one another. As a result, it remains unclear whether success in one role aligns with success in the other. In this paper, we address this question for pragmatic competence by comparing LLMs' performance as pragmatic listeners, judging the appropriateness of linguistic outputs, and as pragmatic speakers, generating pragmatically appropriate language. We evaluate multiple open-weight and proprietary LLMs across three pragmatic settings. We find a robust asymmetry between pragmatic evaluation and pragmatic generation: many models perform substantially better as listeners than as speakers. Our results suggest that pragmatic judging and pragmatic generation are only weakly aligned in current LLMs, calling for more integrated evaluation practices.