SAVOIR: Learning Social Savoir-Faire via Shapley-based Reward Attribution
arXiv cs.AI / 4/22/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- The paper tackles how to train language agents for social intelligence by addressing the multi-turn reinforcement learning credit assignment problem in dialogue.
- It argues that existing methods that distribute episode-level rewards are often retrospective and not theoretically grounded, motivating a new framework.
- SAVOIR (ShApley Value fOr SocIal RL) uses cooperative game theory to produce principled utterance-level credit via expected-utility shifts and Shapley values.
- Experiments on the SOTOPIA benchmark show SAVOIR delivers new state-of-the-art results across evaluation settings, with its 7B model matching or outperforming proprietary systems like GPT-4o and Claude-3.5-Sonnet.
- The results suggest that social intelligence may require fundamentally different capabilities than purely analytical reasoning, since large reasoning models underperform consistently.


