When Contextual Inference Fails: Cancelability in Interactive Instruction Following

arXiv cs.CL / 3/23/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces Build What I Mean (BWIM), an interactive benchmark for contextual meaning construction in guided instruction following.
  • BWIM extends a two-speaker psycholinguistic paradigm to compare contextual inference versus literal adherence under a small communication cost.
  • Evaluations of state-of-the-art LLMs show a dissociation between judgment and action: models can detect speaker unreliability in confidence judgments but do not consistently use that to request clarifications.
  • Consequently, models exhibit suboptimal strategies, such as partner-blind over-clarification and question-averse guessing under uncertainty, highlighting a gap between understanding and actionable behavior.

Abstract

We investigate the separation of literal interpretation from contextual inference in a collaborative block-building task where a builder must resolve underspecified instructions using contextual inferences. Building on an existing two-speaker psycholinguistic paradigm -- which contrasts a pragmatically cooperative speaker with one who is only literally reliable -- we introduce Build What I Mean (BWIM), an interactive benchmark for contextual meaning construction. In BWIM, models must resolve ambiguity by either performing a contextual inference or requesting clarification at a small communication cost. Evaluating several state-of-the-art LLMs, we find a dissociation between judgment and action: while models detect speaker unreliability in explicit confidence ratings, they fail to exploit this information to guide efficient clarification behavior. Instead, we observe suboptimal strategies, such as partner-blind over-clarification and question-averse guessing under uncertainty.