BARRED: Synthetic Training of Custom Policy Guardrails via Asymmetric Debate

arXiv cs.CL / 4/29/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • BARRED introduces a method to train custom policy guardrails without relying on large labeled datasets, which are typically expensive to create.
  • The framework generates high-fidelity synthetic training data from only a task description plus a small set of unlabeled examples.
  • BARRED improves coverage by decomposing the domain into multiple dimensions, ensuring the synthetic data spans a broader range of boundary cases.
  • It uses multi-agent debate to verify label correctness, aiming to maintain both label fidelity and diversity in the resulting training corpus.
  • Experiments show that small language models fine-tuned on BARRED’s synthetic data outperform several state-of-the-art proprietary LLMs and dedicated guardrail models, with ablation results highlighting the importance of both dimension decomposition and debate verification.

Abstract

Deploying guardrails for custom policies remains challenging, as generic safety models fail to capture task-specific requirements, while prompting LLMs suffers from inconsistent boundary-case performance and high inference costs. Training custom classifiers achieves both accuracy and efficiency, yet demands substantial labeled data that is costly to obtain. We present BARRED (Boundary Alignment Refinement through REflection and Debate), a framework for generating faithful and diverse synthetic training data using only a task description and a small set of unlabeled examples. Our approach decomposes the domain space into dimensions to ensure comprehensive coverage, and employs multi-agent debate to verify label correctness, yielding a high-fidelity training corpus. Experiments across diverse custom policies demonstrate that small language models finetuned on our synthetic data consistently outperform state-of-the-art proprietary LLMs (including reasoning models) and dedicated guardrail models. Ablation studies confirm that both dimension decomposition and debate-based verification are critical for ensuring the diversity and label fidelity required for effective fine-tuning. The BARRED framework eliminates the reliance on extensive human annotation, offering a scalable solution for accurate custom guardrails.