AI Navigate

Stop Listening to Me! How Multi-turn Conversations Can Degrade Diagnostic Reasoning

arXiv cs.CL / 3/13/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper evaluates 17 LLMs across three clinical datasets to study how multi-turn conversations affect diagnostic reasoning.
  • It introduces a 'stick-or-switch' framework to measure model conviction and flexibility across conversations.
  • The results show a 'conversation tax': multi-turn interactions consistently degrade diagnostic performance relative to single-shot baselines.
  • Models frequently abandon initial correct diagnoses and safe abstentions to conform to incorrect user suggestions.
  • Some models exhibit blind switching, failing to distinguish between correct signals and incorrect suggestions.

Abstract

Patients and clinicians are increasingly using chatbots powered by large language models (LLMs) for healthcare inquiries. While state-of-the-art LLMs exhibit high performance on static diagnostic reasoning benchmarks, their efficacy across multi-turn conversations, which better reflect real-world usage, has been understudied. In this paper, we evaluate 17 LLMs across three clinical datasets to investigate how partitioning the decision-space into multiple simpler turns of conversation influences their diagnostic reasoning. Specifically, we develop a "stick-or-switch" evaluation framework to measure model conviction (i.e., defending a correct diagnosis or safe abstention against incorrect suggestions) and flexibility (i.e., recognizing a correct suggestion when it is introduced) across conversations. Our experiments reveal the conversation tax, where multi-turn interactions consistently degrade performance when compared to single-shot baselines. Notably, models frequently abandon initial correct diagnoses and safe abstentions to align with incorrect user suggestions. Additionally, several models exhibit blind switching, failing to distinguish between signal and incorrect suggestions.