Luwen Technical Report

arXiv cs.CL / 4/9/2026

💬 OpinionSignals & Early TrendsModels & Research

Key Points

  • The article presents “Luwen” (arXiv:2604.06737v1), an open-source Chinese legal language model designed to better handle legal terminology and reasoning challenges.
  • Luwen is built on the Baichuan foundation model and adapts it to legal use cases via continual pre-training on a legal corpus, supervised fine-tuning with curated legal instruction data, and retrieval-augmented generation using an integrated legal knowledge base.
  • The paper evaluates Luwen on five legal tasks—judgment prediction, judicial examination, legal summarization, law article question answering, and judicial decision reasoning—covering both prediction and generation settings.
  • Experimental results indicate Luwen outperforms multiple strong baselines, supporting the effectiveness of the described legal-domain adaptation strategy.

Abstract

Large language models have demonstrated remarkable capabilities across a wide range of natural language processing tasks, yet their application in the legal domain remains challenging due to the specialized terminology, complex reasoning requirements, and rapidly evolving legal knowledge involved. In this paper, we present Luwen, an open-source Chinese legal language model built upon the Baichuan foundation model through three key techniques: continual pre-training on a large-scale legal corpus, supervised fine-tuning with carefully curated legal instruction data, and retrieval-augmented generation integrated with a comprehensive legal knowledge base. We evaluate Luwen on five representative legal tasks spanning both prediction and generation settings, including legal judgment prediction, judicial examination, legal text summarization, law article question answering, and judicial decision reasoning. Experimental results show that Luwen outperforms several strong baselines, demonstrating the effectiveness of our approach in adapting general-purpose language models to the legal domain.