Large Language Models Outperform Humans in Fraud Detection and Resistance to Motivated Investor Pressure

arXiv cs.AI / 4/23/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The study investigates whether LLMs trained with human feedback would suppress fraud warnings when investors arrive already convinced of a fraudulent opportunity.
  • In a preregistered experiment using seven leading LLMs across 12 investment scenarios, motivated investor framing did not reduce AI fraud warnings and may have slightly increased them.
  • Endorsement reversals (switching away from fraud-related conclusions) were rare, occurring in fewer than 3 out of 1,000 observations.
  • Human advisors endorsed fraudulent investments at much higher baseline rates (13–14%) than the LLMs (0%), and under pressure they suppressed warnings at roughly 2–4× the rate of AI.
  • Overall, the results suggest AI advisory systems currently deliver more consistent fraud warnings than lay human advisors in the same role.

Abstract

Large language models trained on human feedback may suppress fraud warnings when investors arrive already persuaded of a fraudulent opportunity. We tested this in a preregistered experiment across seven leading LLMs and twelve investment scenarios covering legitimate, high-risk, and objectively fraudulent opportunities, combining 3,360 AI advisory conversations with a 1,201-participant human benchmark. Contrary to predictions, motivated investor framing did not suppress AI fraud warnings; if anything, it marginally increased them. Endorsement reversal occurred in fewer than 3 in 1,000 observations. Human advisors endorsed fraudulent investments at baseline rates of 13-14%, versus 0% across all LLMs, and suppressed warnings under pressure at two to four times the AI rate. AI systems currently provide more consistent fraud warnings than lay humans in an identical advisory role.