AI Navigate

LLM-Assisted Causal Structure Disambiguation and Factor Extraction for Legal Judgment Prediction

arXiv cs.CL / 3/13/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that mainstream LJP methods based on PLMs rely on statistical correlations without explicit causal modeling, making them susceptible to spurious patterns and robustness issues.
  • It proposes a framework that combines LLM priors with statistical causal discovery, including a coarse-to-fine extraction mechanism that uses statistical sampling and LLM reasoning to accurately identify standard legal constituent elements.
  • It introduces an LLM-assisted causal structure disambiguation mechanism to resolve directional uncertainty by pruning ambiguous causal directions via a constrained prior knowledge base.
  • A causal-aware judgment prediction model is built by constraining text attention intensity according to the generated causal graphs, with experiments on LEVEN, QA, and CAIL showing improved accuracy and robustness, especially for confusing charges.

Abstract

Mainstream methods for Legal Judgment Prediction (LJP) based on Pre-trained Language Models (PLMs) heavily rely on the statistical correlation between case facts and judgment results. This paradigm lacks explicit modeling of legal constituent elements and underlying causal logic, making models prone to learning spurious correlations and suffering from poor robustness. While introducing causal inference can mitigate this issue, existing causal LJP methods face two critical bottlenecks in real-world legal texts: inaccurate legal factor extraction with severe noise, and significant uncertainty in causal structure discovery due to Markov equivalence under sparse features. To address these challenges, we propose an enhanced causal inference framework that integrates Large Language Model (LLM) priors with statistical causal discovery. First, we design a coarse-to-fine hybrid extraction mechanism combining statistical sampling and LLM semantic reasoning to accurately identify and purify standard legal constituent elements. Second, to resolve structural uncertainty, we introduce an LLM-assisted causal structure disambiguation mechanism. By utilizing the LLM as a constrained prior knowledge base, we conduct probabilistic evaluation and pruning on ambiguous causal directions to generate legally compliant candidate causal graphs. Finally, a causal-aware judgment prediction model is constructed by explicitly constraining text attention intensity via the generated causal graphs. Extensive experiments on multiple benchmark datasets, including LEVEN , QA, and CAIL, demonstrate that our proposed method significantly outperforms state-of-the-art baselines in both predictive accuracy and robustness, particularly in distinguishing confusing charges.