DALDALL: Data Augmentation for Lexical and Semantic Diverse in Legal Domain by leveraging LLM-Persona

arXiv cs.CL / 3/25/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces DALDALL, a persona-based data augmentation framework designed to improve legal information retrieval in low-resource settings where data scarcity persists.
  • Instead of generating large volumes of synthetic queries with generic prompting, DALDALL uses domain-specific professional personas (e.g., attorneys, prosecutors, judges) to produce synthetic queries with higher lexical and semantic diversity.
  • Experiments on the CLERC and COLIEE benchmarks show that persona-based augmentation improves lexical diversity (via Self-BLEU) while maintaining semantic fidelity to the original queries.
  • Dense retrievers fine-tuned on persona-augmented data achieve competitive or better recall than retrievers trained on original data or using generic augmentation strategies.
  • Overall, the work positions persona-based prompting as an effective approach for creating higher-quality training data for specialized legal IR tasks.

Abstract

Data scarcity remains a persistent challenge in low-resource domains. While existing data augmentation methods leverage the generative capabilities of large language models (LLMs) to produce large volumes of synthetic data, these approaches often prioritize quantity over quality and lack domain-specific strategies. In this work, we introduce DALDALL, a persona-based data augmentation framework tailored for legal information retrieval (IR). Our method employs domain-specific professional personas--such as attorneys, prosecutors, and judges--to generate synthetic queries that exhibit substantially greater lexical and semantic diversity than vanilla prompting approaches. Experiments on the CLERC and COLIEE benchmarks demonstrate that persona-based augmentation achieves improvement in lexical diversity as measured by Self-BLEU scores, while preserving semantic fidelity to the original queries. Furthermore, dense retrievers fine-tuned on persona-augmented data consistently achieve competitive or superior recall performance compared to those trained on original data or generic augmentations. These findings establish persona-based prompting as an effective strategy for generating high-quality training data in specialized, low-resource domains.