AI Navigate

Optimal Splitting of Language Models from Mixtures to Specialized Domains

arXiv cs.CL / 3/20/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces a split model training approach that trains multiple models independently on a general corpus and uses scaling laws to determine the optimal compute allocation between pretraining and domain-specific continued pretraining.
  • It provides a loss-prediction framework that estimates a model's performance for size N given D pretraining tokens and D' specialization tokens, enabling scalable planning across model sizes and data budgets.
  • The approach yields consistent performance gains on common-sense knowledge and reasoning benchmarks across different model sizes and compute budgets in language modeling.
  • The framework generalizes to extrapolate to larger model sizes and token counts, indicating practical benefits for multi-domain specialization strategies.

Abstract

Language models achieve impressive performance on a variety of knowledge, language, and reasoning tasks due to the scale and diversity of pretraining data available. The standard training recipe is a two-stage paradigm: pretraining first on the full corpus of data followed by specialization on a subset of high quality, specialized data from the full corpus. In the multi-domain setting, this involves continued pretraining of multiple models on each specialized domain, referred to as split model training. We propose a method for pretraining multiple models independently over a general pretraining corpus, and determining the optimal compute allocation between pretraining and continued pretraining using scaling laws. Our approach accurately predicts the loss of a model of size N with D pretraining and D' specialization tokens, and extrapolates to larger model sizes and number of tokens. Applied to language model training, our approach improves performance consistently across common sense knowledge and reasoning benchmarks across different model sizes and compute budgets.