Multi-View Attention Multiple-Instance Learning Enhanced by LLM Reasoning for Cognitive Distortion Detection

arXiv cs.CL / 4/20/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces a cognitive distortion detection framework that combines Large Language Model (LLM) reasoning with Multiple-Instance Learning (MIL) to better handle contextual ambiguity and semantic overlap.
  • Each utterance is decomposed into Emotion, Logic, and Behavior (ELB) components, which the LLM uses to infer multiple distortion instances with predicted types, expressions, and LLM-assigned salience scores.
  • A Multi-View Gated Attention mechanism integrates these LLM-inferred instances to produce the final classification output.
  • Experiments on the KoACD (Korean) and Therapist QA (English) datasets show that adding ELB features and LLM-derived salience scores improves performance, particularly for distortions that are difficult to interpret.
  • The authors state that the dataset and implementation details are publicly available, supporting reproducibility and further research in mental-health NLP.

Abstract

Cognitive distortions have been closely linked to mental health disorders, yet their automatic detection remains challenging due to contextual ambiguity, co-occurrence, and semantic overlap. We propose a novel framework that combines Large Language Models (LLMs) with a Multiple-Instance Learning (MIL) architecture to enhance interpretability and expression-level reasoning. Each utterance is decomposed into Emotion, Logic, and Behavior (ELB) components, which are processed by LLMs to infer multiple distortion instances, each with a predicted type, expression, and model-assigned salience score. These instances are integrated via a Multi-View Gated Attention mechanism for final classification. Experiments on Korean (KoACD) and English (Therapist QA) datasets demonstrate that incorporating ELB and LLM-inferred salience scores improves classification performance, especially for distortions with high interpretive ambiguity. Our results suggest a psychologically grounded and generalizable approach for fine-grained reasoning in mental health NLP. The dataset and implementation details are publicly accessible.