AI Navigate

Controllable Evidence Selection in Retrieval-Augmented Question Answering via Deterministic Utility Gating

arXiv cs.CL / 3/20/2026

📰 NewsModels & Research

Key Points

  • The paper proposes a deterministic evidence selection framework for retrieval-augmented QA that uses Meaning-Utility Estimation (MUE) and Diversity-Utility Estimation (DUE) to gate evidence before answer generation.
  • It introduces fixed scoring and redundancy-control procedures, evaluating each candidate sentence independently on signals such as semantic relatedness, term coverage, conceptual distinctiveness, and redundancy.
  • Unlike prior approaches, units are accepted only if they explicitly satisfy the required fact, rule, or condition, and the system returns no answer if no unit meets the requirement, enabling auditable evidence sets.
  • No training or fine-tuning is required, and the gating establishes a clear boundary between relevant text and usable evidence.

Abstract

Many modern AI question-answering systems convert text into vectors and retrieve the closest matches to a user question. While effective for topical similarity, similarity scores alone do not explain why some retrieved text can serve as evidence while other equally similar text cannot. When many candidates receive similar scores, systems may select sentences that are redundant, incomplete, or address different conditions than the question requires. This paper presents a deterministic evidence selection framework for retrieval-augmented question answering. The approach introduces Meaning-Utility Estimation (MUE) and Diversity-Utility Estimation (DUE), fixed scoring and redundancy-control procedures that determine evidence admissibility prior to answer generation. Each sentence or record is evaluated independently using explicit signals for semantic relatedness, term coverage, conceptual distinctiveness, and redundancy. No training or fine-tuning is required. In the prototype, a unit is accepted only if it explicitly states the fact, rule, or condition required by the task. Units are not merged or expanded. If no unit independently satisfies the requirement, the system returns no answer. This deterministic gating produces compact, auditable evidence sets and establishes a clear boundary between relevant text and usable evidence.