WRAP++: Web discoveRy Amplified Pretraining

arXiv cs.CL / 4/9/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • WRAP++ (Web discoveRy Amplified Pretraining) targets a limitation of synthetic-data rephrasing by moving from single-document rewriting to cross-document knowledge synthesis using web hyperlinks.
  • The method discovers high-confidence cross-document relational motifs (e.g., dual-links and co-mentions) and generates joint QA that forces reasoning across pairs of documents.
  • By adding relational context that is not present in either source document alone, WRAP++ aims to create new entry points to the same facts and improve how LLMs learn associations.
  • The discovery-driven process also increases dataset size combinatorially, and the paper reports scaling Wikipedia text from ~8.4B tokens into ~80B tokens of cross-document QA.
  • Experiments on SimpleQA using OLMo-based models (7B and 32B) show substantial and sustained gains over single-document approaches, indicating benefits from cross-document knowledge amplification.

Abstract

Synthetic data rephrasing has emerged as a powerful technique for enhancing knowledge acquisition during large language model (LLM) pretraining. However, existing approaches operate at the single-document level, rewriting individual web pages in isolation. This confines synthesized examples to intra-document knowledge, missing cross-document relationships and leaving facts with limited associative context. We propose WRAP++ (Web discoveRy Amplified Pretraining), which amplifies the associative context of factual knowledge by discovering cross-document relationships from web hyperlinks and synthesizing joint QA over each discovered document pair. Concretely, WRAP++ discovers high-confidence relational motifs including dual-links and co-mentions, and synthesizes QA that requires reasoning across both documents. This produces relational knowledge absent from either source document alone, creating diverse entry points to the same facts. Because the number of valid entity pairs grows combinatorially, this discovery-driven synthesis also amplifies data scale far beyond single-document rewriting. Instantiating WRAP++ on Wikipedia, we amplify ~8.4B tokens of raw text into 80B tokens of cross-document QA data. On SimpleQA, OLMo-based models at both 7B and 32B scales trained with WRAP++ substantially outperform single-document approaches and exhibit sustained scaling gains, underscoring the advantage of cross-document knowledge discovery and amplification.