From Query to Counsel: Structured Reasoning with a Multi-Agent Framework and Dataset for Legal Consultation

arXiv cs.CL / 4/14/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces JurisCQAD, a large Chinese legal consultation QA dataset with 43,000+ real queries and expert-validated positive/negative response annotations to tackle data scarcity and contextual complexity.
  • It proposes a structured decomposition that converts each legal query into a legal element graph, linking entities, events, intents, and legal issues to model contextual dependencies and procedural logic.
  • The work also presents JurisMA, a modular multi-agent framework with dynamic routing, statutory grounding, and stylistic optimization to improve context-aware legal reasoning.
  • Experiments show the combined approach (element graph + multi-agent framework) outperforms both general-purpose and legal-domain LLM baselines on an enhanced LawBench across lexical and semantic metrics.
  • The authors emphasize interpretability via decomposition and the advantages of modular collaboration for Legal CQA compared with monolithic prompting or general legal QA methods.

Abstract

Legal consultation question answering (Legal CQA) presents unique challenges compared to traditional legal QA tasks, including the scarcity of high-quality training data, complex task composition, and strong contextual dependencies. To address these, we construct JurisCQAD, a large-scale dataset of over 43,000 real-world Chinese legal queries annotated with expert-validated positive and negative responses, and design a structured task decomposition that converts each query into a legal element graph integrating entities, events, intents, and legal issues. We further propose JurisMA, a modular multi-agent framework supporting dynamic routing, statutory grounding, and stylistic optimization. Combined with the element graph, the framework enables strong context-aware reasoning, effectively capturing dependencies across legal facts, norms, and procedural logic. Trained on JurisCQAD and evaluated on a refined LawBench, our system significantly outperforms both general-purpose and legal-domain LLMs across multiple lexical and semantic metrics, demonstrating the benefits of interpretable decomposition and modular collaboration in Legal CQA.