Learning to Negotiate: Multi-Agent Deliberation for Collective Value Alignment in LLMs
arXiv cs.CL / 3/12/2026
📰 NewsModels & Research
Key Points
- A new multi-agent negotiation-based framework is proposed to align LLMs to Collective Agency, improving their ability to handle value conflicts in multi-stakeholder environments.
- The approach uses two self-play instances of the same LLM with opposing personas that engage in structured turn-based dialogue to synthesize mutually beneficial solutions.
- Training combines RL from human feedback (RLAIF) with GRPO and an external reward model, applying gradients to dialogue tokens based on final CA scores.
- Empirical results show the model achieves CA alignment comparable to a single-agent baseline while substantially improving conflict-resolution performance without degrading general language capabilities.
- The work suggests negotiation-driven deliberation training as a practical path toward LLMs that better support collective decision-making in value-conflict scenarios.
Related Articles
Self-Refining Agents in Spec-Driven Development
Dev.to

has anyone tried this? Flash-MoE: Running a 397B Parameter Model on a Laptop
Reddit r/LocalLLaMA

M2.7 open weights coming in ~2 weeks
Reddit r/LocalLLaMA

MiniMax M2.7 Will Be Open Weights
Reddit r/LocalLLaMA
Best open source coding models for claude code? LB?
Reddit r/LocalLLaMA