AI Navigate

MAC: Multi-Agent Constitution Learning

arXiv cs.AI / 3/18/2026

📰 NewsIdeas & Deep AnalysisTools & Practical UsageModels & Research

Key Points

  • MAC introduces Multi-Agent Constitutional Learning, a framework to optimize over structured prompts as rule sets using a network of agents that can accept, edit, or reject rule updates.
  • MAC+ further improves performance by training agents on successful trajectories to reinforce reward-maximizing updates.
  • The approach yields human-readable, auditable rule sets and generalizes to other agentic tasks such as tool calling, with performance comparable to supervised fine-tuning and GRPO without parameter updates.
  • It outperforms recent prompt-optimization methods by over 50%, addressing limitations of prior methods that require many labeled examples and lack structured prompts.

Abstract

Constitutional AI is a method to oversee and control LLMs based on a set of rules written in natural language. These rules are typically written by human experts, but could in principle be learned automatically given sufficient training data for the desired behavior. Existing LLM-based prompt optimizers attempt this but are ineffective at learning constitutions since (i) they require many labeled examples and (ii) lack structure in the optimized prompts, leading to diminishing improvements as prompt size grows. To address these limitations, we propose Multi-Agent Constitutional Learning (MAC), which optimizes over structured prompts represented as sets of rules using a network of agents with specialized tasks to accept, edit, or reject rule updates. We also present MAC+, which improves performance by training agents on successful trajectories to reinforce updates leading to higher reward. We evaluate MAC on tagging Personally Identifiable Information (PII), a classification task with limited labels where interpretability is critical, and demonstrate that it generalizes to other agentic tasks such as tool calling. MAC outperforms recent prompt optimization methods by over 50%, produces human-readable and auditable rule sets, and achieves performance comparable to supervised fine-tuning and GRPO without requiring parameter updates.