From Benchmarking to Reasoning: A Dual-Aspect, Large-Scale Evaluation of LLMs on Vietnamese Legal Text

arXiv cs.CL / 4/20/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes a dual-aspect, large-scale evaluation framework for Vietnamese legal texts, arguing that simple metrics are insufficient to judge LLM capabilities for legal tasks.
  • It benchmarks four state-of-the-art LLMs (GPT-4o, Claude 3 Opus, Gemini 1.5 Pro, and Grok-1) across Accuracy, Readability, and Consistency.
  • A large-scale error analysis is performed on 60 complex Vietnamese legal articles using an expert-validated error taxonomy to explain the reasons behind observed performance.
  • The study finds a key trade-off: Grok-1 scores highly on Readability and Consistency but shows weaker fine-grained legal Accuracy, while Claude 3 Opus attains high Accuracy that can conceal subtle but critical reasoning mistakes.
  • The most common failure types are Incorrect Example and Misinterpretation, leading to the conclusion that the main challenge is controlled, accurate legal reasoning rather than summarization.

Abstract

The complexity of Vietnam's legal texts presents a significant barrier to public access to justice. While Large Language Models offer a promising solution for legal text simplification, evaluating their true capabilities requires a multifaceted approach that goes beyond surface-level metrics. This paper introduces a comprehensive dual-aspect evaluation framework to address this need. First, we establish a performance benchmark for four state-of-the-art large language models (GPT-4o, Claude 3 Opus, Gemini 1.5 Pro, and Grok-1) across three key dimensions: Accuracy, Readability, and Consistency. Second, to understand the "why" behind these performance scores, we conduct a large-scale error analysis on a curated dataset of 60 complex Vietnamese legal articles, using a novel, expert-validated error typology. Our results reveal a crucial trade-off: models like Grok-1 excel in Readability and Consistency but compromise on fine-grained legal Accuracy, while models like Claude 3 Opus achieve high Accuracy scores that mask a significant number of subtle but critical reasoning errors. The error analysis pinpoints \textit{Incorrect Example} and \textit{Misinterpretation} as the most prevalent failures, confirming that the primary challenge for current LLMs is not summarization but controlled, accurate legal reasoning. By integrating a quantitative benchmark with a qualitative deep dive, our work provides a holistic and actionable assessment of LLMs for legal applications.