PerMix-RLVR: Preserving Persona Expressivity under Verifiable-Reward Alignment

arXiv cs.CL / 4/13/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies persona prompting for LLMs, noting that selecting effective personas is costly and that persona effects on output quality are not fully understood.
  • It finds that reinforcement learning with verifiable rewards (RLVR) reduces sensitivity to persona prompts, but introduces a trade-off: stronger alignment/robustness can reduce in-character expressivity when faithful persona adoption is required.
  • To mitigate this robustness–fidelity trade-off, the authors propose PerMix-RLVR, which mixes personas during RLVR training so the model stays stable under harmful persona variation while still matching the requested persona.
  • Empirical results report +21.2% higher persona stability score (PSS) on MATH500 versus RLVR, alongside +11.4% improved persona fidelity on PersonaGym.

Abstract

Persona prompting has been widely adopted to steer large language models (LLMs) behavior and improve their instruction performance by assigning specific characters. However, identifying an optimal persona is time-consuming, and its impact on output quality remains poorly understood. Prior work has mainly addressed this issue at the prompt level via inference-time strategies, incurring additional computation. In this work, we avoid inference-time prompt search by tackling persona sensitivity during training, aiming to train models that adapt their behavior to diverse personas while preserving task performance. In particular, we find that reinforcement learning with verifiable rewards (RLVR) systematically reduces sensitivity to persona prompts, but also reveals an inherent trade-off of outcome-based optimization: while RLVR improves robustness on tasks with verifiable goals, it can also degrade persona expressivity when needed, e.g., in-character role-playing. To address this limitation, we propose PerMix-RLVR, a persona-mixed RLVR strategy that mitigates the persona robustness-fidelity trade-off, preserving strong robustness to harmful persona variation while enabling faithful persona adoption when required. Concretely, PerMix-RLVR improves persona stability score (PSS) over RLVR by +21.2% on MATH500, while also enhancing persona fidelity by +11.4% on PersonaGym.