To Memorize or to Retrieve: Scaling Laws for RAG-Considerate Pretraining

arXiv cs.CL / 4/3/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper investigates how to balance parametric knowledge from pretraining with non-parametric knowledge from retrieval in RAG systems when total data budgets are fixed.
  • It trains OLMo-2-based language models from 30M to 3B parameters using up to 100B DCLM tokens while varying both pretraining corpus size and retrieval store size, then evaluates across reasoning, scientific QA, and open-domain QA benchmarks.
  • Results show that retrieval consistently boosts performance over parametric-only baselines across model sizes, and the authors propose a three-dimensional scaling framework linking model size, pretraining tokens, and retrieval corpus size.
  • The scaling “manifold” is used to estimate optimal data allocation strategies between pretraining and retrieval, with marginal gains from retrieval depending on model scale, task type, and how saturated the pretraining is.
  • Overall, the study provides quantitative guidance on when and how retrieval should complement pretraining for designing more scalable language modeling systems.

Abstract

Retrieval-augmented generation (RAG) improves language model (LM) performance by providing relevant context at test time for knowledge-intensive situations. However, the relationship between parametric knowledge acquired during pretraining and non-parametric knowledge accessed via retrieval remains poorly understood, especially under fixed data budgets. In this work, we systematically study the trade-off between pretraining corpus size and retrieval store size across a wide range of model and data scales. We train OLMo-2-based LMs ranging from 30M to 3B parameters on up to 100B tokens of DCLM data, while varying both pretraining data scale (1-150x the number of parameters) and retrieval store size (1-20x), and evaluate performance across a diverse suite of benchmarks spanning reasoning, scientific QA, and open-domain QA. We find that retrieval consistently improves performance over parametric-only baselines across model scales and introduce a three-dimensional scaling framework that models performance as a function of model size, pretraining tokens, and retrieval corpus size. This scaling manifold enables us to estimate optimal allocations of a fixed data budget between pretraining and retrieval, revealing that the marginal utility of retrieval depends strongly on model scale, task type, and the degree of pretraining saturation. Our results provide a quantitative foundation for understanding when and how retrieval should complement pretraining, offering practical guidance for allocating data resources in the design of scalable language modeling systems.