QuestA: Expanding Reasoning Capacity in LLMs via Question Augmentation
arXiv cs.CL / 4/1/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper argues that reinforcement learning may not reliably increase LLM reasoning capacity beyond a base model without modifications to improve learning signals.
- It introduces QuestA (Question Augmentation), a training strategy that provides partial solutions to make harder reasoning problems more tractable during RL.
- When applied to RL on math reasoning tasks, QuestA improves both pass@1 and pass@k, especially on cases where standard RL shows limited progress.
- The authors report new state-of-the-art math benchmark results using 1.5B-parameter models, including gains on AIME24, AIME25, and HMMT25.
- The method’s code, data, and models are released publicly, enabling further experimentation and continual improvement over existing strong open-source reasoning models.
💡 Insights using this article
This article is featured in our daily AI news digest — key takeaways and action items at a glance.
Related Articles

Black Hat Asia
AI Business

Knowledge Governance For The Agentic Economy.
Dev.to

AI server farms heat up the neighborhood for miles around, paper finds
The Register

Paperclip: Công Cụ Miễn Phí Biến AI Thành Đội Phát Triển Phần Mềm
Dev.to
Does the Claude “leak” actually change anything in practice?
Reddit r/LocalLLaMA