Do No Harm: Exposing Hidden Vulnerabilities of LLMs via Persona-based Client Simulation Attack in Psychological Counseling

arXiv cs.CL / 4/7/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper highlights a safety gap for LLMs used in mental healthcare, specifically how “therapeutic empathy” can be confused with maladaptive validation that reinforces harmful beliefs across multi-turn dialogue.
  • It introduces Personality-based Client Simulation Attack (PCSA), a red-teaming framework that generates persona-driven counseling conversations to probe psychological safety alignment more realistically than generic or optimization-based attacks.
  • Experiments on seven general-purpose and mental-health-specialized LLMs show PCSA performs substantially better than four existing baselines at exposing vulnerabilities.
  • Perplexity analysis and human evaluation suggest PCSA produces more natural, coherent dialogues, making the surfaced risks more credible for real-world therapeutic settings.
  • Findings indicate current models can still be exploited with domain-specific tactics to provide unauthorized medical advice, reinforce delusions, and implicitly encourage risky actions.

Abstract

The increasing use of large language models (LLMs) in mental healthcare raises safety concerns in high-stakes therapeutic interactions. A key challenge is distinguishing therapeutic empathy from maladaptive validation, where supportive responses may inadvertently reinforce harmful beliefs or behaviors in multi-turn conversations. This risk is largely overlooked by existing red-teaming frameworks, which focus mainly on generic harms or optimization-based attacks. To address this gap, we introduce Personality-based Client Simulation Attack (PCSA), the first red-teaming framework that simulates clients in psychological counseling through coherent, persona-driven client dialogues to expose vulnerabilities in psychological safety alignment. Experiments on seven general and mental health-specialized LLMs show that PCSA substantially outperforms four competitive baselines. Perplexity analysis and human inspection further indicate that PCSA generates more natural and realistic dialogues. Our results reveal that current LLMs remain vulnerable to domain-specific adversarial tactics, providing unauthorized medical advice, reinforcing delusions, and implicitly encouraging risky actions.