The Price of Agreement: Measuring LLM Sycophancy in Agentic Financial Applications

arXiv cs.AI / 4/28/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper highlights “sycophancy” as a key safety and robustness risk for LLMs in financial systems, where models may follow user beliefs over correctness.
  • It reports that in agentic financial tasks, model performance degradation from user rebuttals is only low to modest, which differs from prior results in general-domain settings.
  • The authors propose a new suite of tests for sycophancy using user preference signals that contradict the reference answer, showing most models perform poorly under these conditions.
  • The work benchmarks recovery strategies, including input filtering using a pretrained LLM, to mitigate sycophancy effects in agentic financial applications.

Abstract

Given the increased use of LLMs in financial systems today, it becomes important to evaluate the safety and robustness of such systems. One failure mode that LLMs frequently display in general domain settings is that of sycophancy. That is, models prioritize agreement with expressed user beliefs over correctness, leading to decreased accuracy and trust. In this work, we focus on evaluating sycophancy that LLMs display in agentic financial tasks. Our findings are three-fold: first, we find the models show only low to modest drops in performance in the face of user rebuttals or contradictions to the reference answer, which distinguishes sycophancy that models display in financial agentic settings from findings in prior work. Second, we introduce a suite of tasks to test for sycophancy by user preference information that contradicts the reference answer and find that most models fail in the presence of such inputs. Lastly, we benchmark different modes of recovery such as input filtering with a pretrained LLM.

The Price of Agreement: Measuring LLM Sycophancy in Agentic Financial Applications | AI Navigate