Reasoning Gets Harder for LLMs Inside A Dialogue

arXiv cs.CL / 3/23/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces BOULDER, a dynamic benchmark with eight travel-related tasks requiring arithmetic, spatial, and temporal reasoning, and presents both isolated and dialogue-based variants for controlled comparison.
  • It reports a substantial and consistent performance gap between isolated and dialogue-based reasoning across eight LLMs, highlighting challenges in reasoning under real-world dialogue conditions.
  • The gap is largely attributed to the multi-turn nature of dialogue, with additional effects from role conditioning and tool-use requirements in task-oriented dialogue.
  • The authors argue that evaluating LLM reasoning in realistic interactive scenarios is necessary to accurately assess practical capabilities and limitations.

Abstract

Large Language Models (LLMs) achieve strong performance on many reasoning benchmarks, yet these evaluations typically focus on isolated tasks that differ from real-world usage in task-oriented dialogue (TOD). In this setting, LLMs must perform reasoning inherently while generating text and adhering to instructions on role, format, and style. This mismatch raises concerns about whether benchmark performance accurately reflects models' reasoning robustness in TOD setting. We investigate how framing reasoning tasks within TOD affects LLM performance by introducing BOULDER, a new dynamic benchmark covering eight travel-related tasks that require arithmetic, spatial, and temporal reasoning with both commonsense and formal aspects. Each problem is presented in both isolated and dialogue-based variants, enabling controlled comparison while mitigating data contamination. Experiments on eight LLMs reveal a substantial and consistent performance gap between isolated and dialogue settings. Through ablations and qualitative analysis, we show that this gap is largely driven by the multi-turn nature of dialogue, with additional effects from role conditioning and tool-use requirements. Our results highlight the need to evaluate LLM reasoning in realistic interactive scenarios.

Reasoning Gets Harder for LLMs Inside A Dialogue | AI Navigate