A Survey of Reinforcement Learning for Large Language Models under Data Scarcity: Challenges and Solutions

arXiv cs.LG / 4/21/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper surveys how reinforcement learning (RL) can improve large language models (LLMs) as a post-training paradigm, focusing specifically on the problem of data scarcity.
  • It identifies key data scarcity bottlenecks for LLM-RL, including scarce high-quality external supervision and limited amounts of useful experience generated by the model.
  • The authors introduce a bottom-up hierarchical framework organized around three perspectives—data-centric, training-centric, and framework-centric—to structure the design space.
  • A taxonomy of existing data-efficient RL methods is developed, with representative approaches summarized and their strengths and limitations analyzed.
  • The survey is intended to serve as a conceptual foundation and roadmap for future research toward more efficient and scalable RL post-training for LLMs.

Abstract

Reinforcement learning (RL) has emerged as a powerful post-training paradigm for enhancing the reasoning capabilities of large language models (LLMs). However, reinforcement learning for LLMs faces substantial data scarcity challenges, including the limited availability of high-quality external supervision and the constrained volume of model-generated experience. These limitations make data-efficient reinforcement learning a critical research direction. In this survey, we present the first systematic review of reinforcement learning for LLMs under data scarcity. We propose a bottom-up hierarchical framework built around three complementary perspectives: the data-centric perspective, the training-centric perspective, and the framework-centric perspective. We develop a taxonomy of existing methods, summarize representative approaches in each category, and analyze their strengths and limitations. Our taxonomy aims to provide a clear conceptual foundation for understanding the design space of data-efficient RL for LLMs and to guide researchers working in this emerging area. We hope this survey offers a comprehensive roadmap for future research and inspires new directions toward more efficient and scalable reinforcement learning post-training for LLMs.