Benchmarking Retrieval Strategies for Biomedical Retrieval-Augmented Generation: A Controlled Empirical Study

arXiv cs.CL / 5/5/2026

📰 NewsDeveloper Stack & InfrastructureSignals & Early TrendsTools & Practical UsageModels & Research

Key Points

  • The study benchmarks five retrieval strategies for a biomedical retrieval-augmented generation (RAG) pipeline using a fixed generation model (GPT-4o-mini), ChromaDB, and OpenAI text-embedding-3-small, isolating the effect of retrieval choice.
  • Across 250 BioASQ-derived biomedical QA pairs evaluated with DeepEval metrics (contextual precision/recall, faithfulness, and answer relevancy) and 95% confidence intervals, Cross-Encoder Reranking achieves the strongest overall composite score (0.827) and top contextual precision (0.852).
  • Multi-Query Expansion shows the weakest contextual precision (0.671), indicating that straightforward query diversification can add retrieval noise even when aiming to improve recall.
  • Maximal Marginal Relevance (MMR) improves diversity but reduces answer relevancy, while the Dense vector baseline (composite 0.822) is nearly tied with the best method.
  • All RAG variants substantially outperform the no-context ablation on answer relevancy (0.658–0.701 vs. 0.287), and the full pipeline, hyperparameters, and evaluation code are publicly released.

Abstract

Retrieval-Augmented Generation (RAG) offers a well-established path to grounding large language model (LLM) outputs in external knowledge, yet the question of which retrieval strategy works best in a high-stakes domain such as biomedicine has not received the controlled, multi-metric treatment it deserves. This paper presents a systematic empirical comparison of five retrieval strategies -- Dense Vector Search, Hybrid BM25 + Dense retrieval, Cross-Encoder Reranking, Multi-Query Expansion, and Maximal Marginal Relevance (MMR) -- within a biomedical question-answering RAG pipeline. All strategies share a fixed generation model (GPT-4o-mini), a common vector store (ChromaDB), and OpenAI's text-embedding-3-small embeddings, ensuring that observed differences are attributable to retrieval alone. Evaluation is conducted on 250 question-answer pairs drawn from a preprocessed subset of the BioASQ benchmark (rag-mini-bioasq) using four DeepEval metrics: contextual precision, contextual recall, faithfulness, and answer relevancy, each reported with 95% confidence intervals. A no-context ablation is included as a lower bound. Cross-Encoder Reranking achieves the best composite score (0.827) and highest contextual precision (0.852), confirming that query-document interaction yields measurable retrieval gains. Multi-Query Expansion, despite its recall-oriented design, produces the weakest contextual precision (0.671), suggesting naive query diversification introduces retrieval noise. MMR sacrifices answer relevancy for diversity, while the Dense baseline (composite 0.822) falls within 0.005 points of the top strategy. All RAG conditions dramatically outperform the no-context ablation on answer relevancy (0.658-0.701 vs. 0.287), confirming the practical value of retrieval. The full pipeline, hyperparameters, and evaluation code are publicly available.

Benchmarking Retrieval Strategies for Biomedical Retrieval-Augmented Generation: A Controlled Empirical Study | AI Navigate