CounselReflect: A Toolkit for Auditing Mental-Health Dialogues
arXiv cs.CL / 4/1/2026
📰 NewsSignals & Early TrendsTools & Practical UsageModels & Research
Key Points
- CounselReflect is an end-to-end toolkit designed to help users audit mental-health support dialogues produced by conversational systems, addressing the lack of structured, transparent evaluation methods.
- It produces multi-dimensional, inspectable reports (session-level summaries, turn-level scores, and evidence-linked excerpts) rather than a single opaque quality score.
- The toolkit combines 12 model-based metrics from task-specific predictors with rubric-based metrics expanded via a literature-derived library of 69 metrics plus user-defined custom metrics evaluated using configurable LLM judges.
- CounselReflect is offered as a web application, browser extension, and CLI for both real-time auditing and large-scale evaluation workflows.
- Initial validation includes a user study (20 participants) and an expert review (6 mental-health professionals), with demo materials and full source code provided.
💡 Insights using this article
This article is featured in our daily AI news digest — key takeaways and action items at a glance.




