On Emotion-Sensitive Decision Making of Small Language Model Agents
arXiv cs.AI / 4/10/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper studies how emotion causally affects decision-making in small language model (SLM) agents by inducing emotional states at the representation level rather than relying only on prompt-based methods.
- It uses activation steering driven by crowd-validated real-world emotion-eliciting texts to create controlled and transferable emotional interventions.
- The authors introduce a game-theoretic benchmark with decision templates covering cooperative and competitive incentives under both complete and incomplete information, instantiated across strategic scenarios and diverse personas.
- Experiments across multiple model families and modalities find that emotional perturbations reliably change strategic choices, but the resulting behaviors can be unstable and only partially match human expectations.
- The work proposes an approach aimed at improving robustness against emotion-driven representation perturbations in SLM agents.



