Aligning Human-AI-Interaction Trust for Mental Health Support: Survey and Position for Multi-Stakeholders

arXiv cs.CL / 4/23/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The article argues that “trustworthy” AI for mental health support is not consistently defined or measured, despite being a shared priority across disciplines.
  • It proposes a three-layer trust framework—human-oriented, AI-oriented, and interaction-oriented—explicitly integrating perspectives from practitioners, researchers, and regulators.
  • Using the framework, it reviews existing AI research in mental health and compares evaluation approaches from automated metrics to clinically validated methods.
  • The authors identify mismatches between what current NLP-focused metrics capture and what real-world mental health settings require, and they outline a research agenda to close these gaps.
  • The overall goal is to guide development of socio-technically aligned AI systems that deliver genuinely trustworthy mental health support in practice.

Abstract

Building trustworthy AI systems for mental health support is a shared priority across stakeholders from multiple disciplines. However, "trustworthy" remains loosely defined and inconsistently operationalized. AI research often focuses on technical criteria (e.g., robustness, explainability, and safety), while therapeutic practitioners emphasize therapeutic fidelity (e.g., appropriateness, empathy, and long-term user outcomes). To bridge the fragmented landscape, we propose a three-layer trust framework, covering human-oriented, AI-oriented, and interaction-oriented trust, integrating the viewpoints of key stakeholders (e.g., practitioners, researchers, regulators). Using this framework, we systematically review existing AI-driven research in mental health domain and examine evaluation practices for ``trustworthy'' ranging from automatic metrics to clinically validated approaches. We highlight critical gaps between what NLP currently measures and what real-world mental health contexts require, and outline a research agenda for building socio-technically aligned and genuinely trustworthy AI for mental health support.