Web Retrieval-Aware Chunking (W-RAC) for Efficient and Cost-Effective Retrieval-Augmented Generation Systems

arXiv cs.AI / 4/8/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses a key bottleneck in RAG systems: document chunking that must trade off retrieval quality, latency, and cost, especially for large-scale web ingestion.
  • It proposes Web Retrieval-Aware Chunking (W-RAC), which separates text extraction from semantic chunk planning by converting parsed web content into structured, ID-addressable units.
  • W-RAC uses LLMs only for retrieval-aware grouping decisions rather than for generating chunk text, aiming to cut token consumption and eliminate hallucination risk during chunking.
  • Experiments and architectural comparisons indicate W-RAC achieves comparable or better retrieval performance than traditional fixed-size, rule-based, or fully agentic chunking approaches.
  • The authors report an order-of-magnitude reduction in chunking-related LLM costs while improving observability/debuggability of the chunking process.

Abstract

Retrieval-Augmented Generation (RAG) systems critically depend on effective document chunking strategies to balance retrieval quality, latency, and operational cost. Traditional chunking approaches, such as fixed-size, rule-based, or fully agentic chunking, often suffer from high token consumption, redundant text generation, limited scalability, and poor debuggability, especially for large-scale web content ingestion. In this paper, we propose Web Retrieval-Aware Chunking (W-RAC), a novel, cost-efficient chunking framework designed specifically for web-based documents. W-RAC decouples text extraction from semantic chunk planning by representing parsed web content as structured, ID-addressable units and leveraging large language models (LLMs) only for retrieval-aware grouping decisions rather than text generation. This significantly reduces token usage, eliminates hallucination risks, and improves system observability.Experimental analysis and architectural comparison demonstrate that W-RAC achieves comparable or better retrieval performance than traditional chunking approaches while reducing chunking-related LLM costs by an order of magnitude.