Wiki Dumps to Training Corpora: South Slavic Case

arXiv cs.CL / 4/29/2026

📰 NewsDeveloper Stack & InfrastructureIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes a two-phase methodology to convert raw Wikimedia dumps into high-quality text corpora for seven South Slavic languages.
  • It first extracts and cleans text from multiple Wikimedia projects (e.g., Wikipedia and related sites), carefully handling wiki markup to isolate real articles and usable natural-language text.
  • It then filters out suspicious or low-quality articles by using an n-gram–based approach to detect high textual redundancy across articles and remove those entries from the final datasets.
  • The resulting corpora are intended to support training language models and comparative linguistic research, while the authors argue the approach is largely language-agnostic and generalizable.
  • Overall, the work emphasizes reliable, high-information corpus creation that better reflects authentic language use and cultural context.

Abstract

This paper presents a methodology for transforming raw Wikimedia dumps into quality textual corpora for seven South Slavic languages. The work is divided into two major phases. The first involves extracting and cleaning text from raw dumps of Wikipedia, Wikisource, Wikibooks, Wikinews, and Wikiquote, where available. This step requires careful handling of raw wiki markup to isolate, first of all, textual articles, and then usable natural language text within them. The second phase addresses the challenge of suspicious or low-quality articles, which are often generated from databases or structured knowledge bases. These articles are characterised by repetitive patterns, generic phrasing, and minimal to no original content. To mitigate their impact, a n-gram-based filtering strategy was employed to detect high levels of textual redundancy between articles and then remove such articles from the corpora entirely. The resulting datasets aim to provide linguistically rich texts suitable for training language models or conducting comparative research across South Slavic languages. By combining systematic extraction with quality control, this work contributes to the creation of reliable, high-information corpora that reflect authentic language use and cultural context. While focused on the South Slavic case in the paper, the approach is mostly language-agnostic and can be generalised to other languages and language families.