Understanding Performance Gap Between Parallel and Sequential Sampling in Large Reasoning Models

arXiv cs.CL / 4/8/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper rigorously compares sequential versus parallel sampling strategies in Large Reasoning Models (LRMs) and finds that parallel sampling generally outperforms sequential sampling despite sequential sampling’s higher representational capacity.
  • It tests three hypotheses for the observed performance gap—effects of the aggregation operator, harm from longer context requirements, and reduced exploration due to conditioning on previous answers.
  • Across multiple model families and sizes (including Qwen3, DeepSeek-R1 distilled models, and Gemini 2.5) and domains (math and coding), the study finds that aggregation/context length are unlikely to be the primary causes.
  • The authors conclude that reduced exploration in sequential sampling is a major driver of the performance gap, offering an explanation grounded in sampling/conditioning dynamics.
  • Overall, the results suggest practitioners should consider exploration-friendly approaches when designing multi-sample inference pipelines for reasoning-focused LLMs.

Abstract

Large Reasoning Models (LRMs) have shown remarkable performance on challenging questions, such as math and coding. However, to obtain a high quality solution, one may need to sample more than once. In principal, there are two sampling strategies that can be composed to form more complex processes: sequential sampling and parallel sampling. In this paper, we first compare these two approaches with rigor, and observe, aligned with previous works, that parallel sampling seems to outperform sequential sampling even though the latter should have more representation power. To understand the underline reasons, we make three hypothesis on the reason behind this behavior: (i) parallel sampling outperforms due to the aggregator operator; (ii) sequential sampling is harmed by needing to use longer contexts; (iii) sequential sampling leads to less exploration due to conditioning on previous answers. The empirical evidence on various model families and sizes (Qwen3, DeepSeek-R1 distilled models, Gemini 2.5) and question domains (math and coding) suggests that the aggregation and context length do not seem to be the main culprit behind the performance gap. In contrast, the lack of exploration seems to play a considerably larger role, and we argue that this is one main cause for the performance gap.