The Serial Scaling Hypothesis

arXiv stat.ML / 4/30/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The article argues that many successes in machine learning came from massive parallelization, but some task types are inherently sequential and have a structural limitation that prevents efficient parallelization.
  • It formalizes the “inherently serial” vs. “parallelizable” distinction in complexity-theoretic terms and claims existing parallel-centric architectures cannot fully overcome these limits for sequentially dependent problems.
  • The authors report that diffusion models, despite their sequential generation process, still cannot solve inherently serial problems, at least within the studied settings.
  • The work concludes that understanding which computations are serial has major implications for future ML model design and even for hardware development strategies.
  • Overall, it reframes model and architecture choices by emphasizing computational seriality as a first-class constraint rather than an edge-case limitation.

Abstract

While machine learning has advanced through massive parallelization, we identify a critical blind spot: some problems are fundamentally sequential. These "inherently serial" problems-from mathematical reasoning to physical simulations to sequential decision-making-require sequentially dependent computational steps that cannot be efficiently parallelized. We formalize this distinction in complexity theory, and demonstrate that current parallel-centric architectures face fundamental limitations on such tasks. Then, we show for first time that diffusion models despite their sequential nature are incapable of solving inherently serial problems. We argue that recognizing the serial nature of computation holds profound implications on machine learning, model design, and hardware development.

The Serial Scaling Hypothesis | AI Navigate