SAVOIR: Learning Social Savoir-Faire via Shapley-based Reward Attribution

arXiv cs.AI / 4/22/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper tackles how to train language agents for social intelligence by addressing the multi-turn reinforcement learning credit assignment problem in dialogue.
  • It argues that existing methods that distribute episode-level rewards are often retrospective and not theoretically grounded, motivating a new framework.
  • SAVOIR (ShApley Value fOr SocIal RL) uses cooperative game theory to produce principled utterance-level credit via expected-utility shifts and Shapley values.
  • Experiments on the SOTOPIA benchmark show SAVOIR delivers new state-of-the-art results across evaluation settings, with its 7B model matching or outperforming proprietary systems like GPT-4o and Claude-3.5-Sonnet.
  • The results suggest that social intelligence may require fundamentally different capabilities than purely analytical reasoning, since large reasoning models underperform consistently.

Abstract

Social intelligence, the ability to navigate complex interpersonal interactions, presents a fundamental challenge for language agents. Training such agents via reinforcement learning requires solving the credit assignment problem: determining how individual utterances contribute to multi-turn dialogue outcomes. Existing approaches directly employ language models to distribute episode-level rewards, yielding attributions that are retrospective and lack theoretical grounding. We propose SAVOIR (ShApley Value fOr SocIal RL), a novel principled framework grounded in cooperative game theory. Our approach combines two complementary principles: expected utility shifts evaluation from retrospective attribution to prospective valuation, capturing an utterance's strategic potential for enabling favorable future trajectories; Shapley values ensure fair credit distribution with axiomatic guarantees of efficiency, symmetry, and marginality. Experiments on the SOTOPIA benchmark demonstrate that SAVOIR achieves new state-of-the-art performance across all evaluation settings, with our 7B model matching or exceeding proprietary models including GPT-4o and Claude-3.5-Sonnet. Notably, even large reasoning models consistently underperform, suggesting social intelligence requires qualitatively different capabilities than analytical reasoning.