FSPO: Few-Shot Optimization of Synthetic Preferences Personalizes to Real Users

arXiv stat.ML / 4/20/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces FSPO (Few-shot Preference Optimization), a meta-learning-based approach to personalize LLMs by quickly inferring a user-specific reward function from a small number of labeled preferences.
  • FSPO incorporates user description rationalization (RAT) to improve reward modeling and instruction-following, with performance that can recover when using an oracle user description.
  • Because real preference data is hard to scale, the authors design methods to generate large synthetic preference datasets (over 1M) using publicly available LLMs.
  • The study finds that successful transfer from synthetic data to real users requires synthetic datasets to have both high diversity and coherent, self-consistent structure.
  • Experiments across movie reviews, education, and open-ended QA (plus a controlled human study) show strong results, including 87% Alpaca Eval win rate on synthetic users and 70% win rate with real human users for open-ended question answering.

Abstract

Effective personalization of LLMs is critical for a broad range of user-interfacing applications such as virtual assistants and content curation. Inspired by the strong in-context capabilities of LLMs, we propose few-shot preference optimization (FSPO), an algorithm for LLM personalization that reframes reward modeling as a meta-learning problem. Under FSPO, an LLM learns to quickly infer a personalized reward function for a user via a few labeled preferences. FSPO also utilizes user description rationalization (RAT) to encourage better reward modeling and instruction following, recovering performance with the oracle user description. Since real-world preference data is challenging to collect at scale, we propose careful design choices to construct synthetic preference datasets for personalization, generating over 1M synthetic personalized preferences using publicly available LLMs. To successfully transfer from synthetic data to real users, we find it crucial for the data to exhibit both high diversity and coherent, self-consistent structure. We evaluate FSPO on personalized open-ended generation for up to 1,500 synthetic users across three domains: movie reviews, education, and open-ended question answering. We also run a controlled human study. Overall, FSPO achieves an 87% Alpaca Eval winrate in generating responses that are personalized to synthetic users and a 70% winrate with real human users in open-ended question answering.