QuestA: Expanding Reasoning Capacity in LLMs via Question Augmentation

arXiv cs.CL / 4/1/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that reinforcement learning may not reliably increase LLM reasoning capacity beyond a base model without modifications to improve learning signals.
  • It introduces QuestA (Question Augmentation), a training strategy that provides partial solutions to make harder reasoning problems more tractable during RL.
  • When applied to RL on math reasoning tasks, QuestA improves both pass@1 and pass@k, especially on cases where standard RL shows limited progress.
  • The authors report new state-of-the-art math benchmark results using 1.5B-parameter models, including gains on AIME24, AIME25, and HMMT25.
  • The method’s code, data, and models are released publicly, enabling further experimentation and continual improvement over existing strong open-source reasoning models.

Abstract

Reinforcement learning (RL) has emerged as a central paradigm for training large language models (LLMs) in reasoning tasks. Yet recent studies question RL's ability to incentivize reasoning capacity beyond the base model. This raises a key challenge: how can RL be adapted to solve harder reasoning problems more effectively? To address this challenge, we propose a simple yet effective strategy via Question Augmentation: introduce partial solutions during training to reduce problem difficulty and provide more informative learning signals. Our method, QuestA, when applied during RL training on math reasoning tasks, not only improves pass@1 but also pass@k-particularly on problems where standard RL struggles to make progress. This enables continual improvement over strong open-source models such as DeepScaleR and OpenMath Nemotron, further enhancing their reasoning capabilities. We achieve new state-of-the-art results on math benchmarks using 1.5B-parameter models: 72.50% (+10.73%) on AIME24, 62.29% (+12.79%) on AIME25, and 41.67% (+10.11%) on HMMT25. Code, data and model are available at https://github.com/foreverlasting1202/QuestA.