Learning to Negotiate: Multi-Agent Deliberation for Collective Value Alignment in LLMs
arXiv cs.CL / 3/12/2026
📰 NewsModels & Research
Key Points
- A new multi-agent negotiation-based framework is proposed to align LLMs to Collective Agency, improving their ability to handle value conflicts in multi-stakeholder environments.
- The approach uses two self-play instances of the same LLM with opposing personas that engage in structured turn-based dialogue to synthesize mutually beneficial solutions.
- Training combines RL from human feedback (RLAIF) with GRPO and an external reward model, applying gradients to dialogue tokens based on final CA scores.
- Empirical results show the model achieves CA alignment comparable to a single-agent baseline while substantially improving conflict-resolution performance without degrading general language capabilities.
- The work suggests negotiation-driven deliberation training as a practical path toward LLMs that better support collective decision-making in value-conflict scenarios.
Related Articles
Data Augmentation Using GANs
Dev.to
Zero Shot Deformation Reconstruction for Soft Robots Using a Flexible Sensor Array and Cage Based 3D Gaussian Modeling
arXiv cs.RO
Speculative Policy Orchestration: A Latency-Resilient Framework for Cloud-Robotic Manipulation
arXiv cs.RO
ReMAP-DP: Reprojected Multi-view Aligned PointMaps for Diffusion Policy
arXiv cs.RO
AGILE: A Comprehensive Workflow for Humanoid Loco-Manipulation Learning
arXiv cs.RO