Aleph-Alpha-GermanWeb: Improving German-language LLM pre-training with model-based data curation and synthetic data generation

arXiv cs.CL / 4/1/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes a German-language LLM data curation pipeline that blends heuristic filtering, model-based filtering, and synthetic data generation to improve training efficiency and downstream performance.
  • It presents Aleph-Alpha-GermanWeb, a 628B-word German pre-training dataset made from organic Common Crawl and FineWeb2 subsets plus a synthetic subset generated using actual organic web data as conditioning.
  • The authors evaluate the dataset by training a 1B “Llama-style” model and an 8B tokeniser-free hierarchical autoregressive transformer (HAT) from scratch, then testing on German benchmarks including MMMLU.
  • Results show that Aleph-Alpha-GermanWeb yields significant gains over FineWeb2 alone, and these gains persist even at the 8B scale when FineWeb2 is enhanced with human-curated sources like Wikipedia.
  • The study concludes that model-based curation and synthetic data generation can materially improve German LLM pre-training datasets, supporting broader evidence from similar work in other languages/domains.

Abstract

Scaling data quantity is essential for large language models (LLMs), yet recent findings show that data quality can significantly boost performance and training efficiency. We introduce a German-language dataset curation pipeline that combines heuristic and model-based filtering techniques with synthetic data generation. We use our pipeline to create Aleph-Alpha-GermanWeb, a 628B-word German pre-training dataset composed of three subsets drawing from: (1) Common Crawl web data (organic subset; 78B words), (2) FineWeb2 (organic subset; 235B), and (3) synthetically-generated data conditioned on actual, organic web data (synthetic subset; 329B). We evaluate our dataset by pre-training both a 1B Llama-style model and an 8B tokeniser-free hierarchical autoregressive transformer (HAT) from scratch. A comparison on German-language benchmarks, including MMMLU, shows significant performance gains of Aleph-Alpha-GermanWeb over FineWeb2 alone. This advantage holds at the 8B scale even when FineWeb2 is enriched by human-curated high-quality data sources such as Wikipedia. Our findings support the growing body of evidence that model-based data curation and synthetic data generation can significantly enhance LLM pre-training datasets.