Scaling DPPs for RAG: Density Meets Diversity

arXiv cs.AI / 4/7/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that conventional RAG retrieval, which relies on point-wise relevance scoring, neglects interactions between retrieved chunks and can produce redundant context that weakens coverage and “density.”
  • It proposes ScalDPP, a diversity-aware retrieval method for RAG that uses Determinantal Point Processes (DPPs) to model inter-chunk dependencies while keeping the approach scalable via a lightweight P-Adapter.
  • To train and enforce the desired retrieval behavior, the authors introduce Diverse Margin Loss (DML), designed to make ground-truth complementary evidence chains outperform redundant alternatives under the DPP geometry.
  • Experiments show that ScalDPP improves retrieval quality in practice, supporting the thesis that jointly optimizing density (information richness) and diversity (coverage) yields better grounded generation for LLMs.

Abstract

Retrieval-Augmented Generation (RAG) enhances Large Language Models (LLMs) by grounding generation in external knowledge, yielding relevance responses that are aligned with factual evidence and evolving corpora. Standard RAG pipelines construct context through relevance ranking, performing point-wise scoring between the user query and each corpora chunk. This formulation, however, ignores interactions among retrieved candidates, leading to redundant contexts that dilute density and fail to surface complementary evidence. We argue that effective retrieval should optimize jointly for both density and diversity, ensuring the grounding evidence that is dense in information yet diverse in coverage. In this study, we propose ScalDPP, a diversity-aware retrieval mechanism for RAG that incorporates Determinantal Point Processes (DPPs) through a lightweight P-Adapter, enabling scalable modeling of inter-chunk dependencies and complementary context selection. In addition, we develop a novel set-level objective, Diverse Margin Loss (DML), that enforces ground-truth complementary evidence chains to dominate any equally sized redundant alternatives under DPP geometry. Experimental results demonstrate the superiority of ScalDPP, substantiating our core statement in practice.