Context Selection for Hypothesis and Statistical Evidence Extraction from Full-Text Scientific Articles

arXiv cs.CL / 3/24/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses the problem of extracting scientific hypotheses and their supporting or refuting statistical evidence from full-text articles by modeling a sequential linking from abstract findings to body hypotheses and evidence.
  • It highlights a within-document retrieval challenge where topically related paragraphs can play different rhetorical roles, producing hard negatives that complicate hypothesis/evidence extraction.
  • Using a two-stage retrieve-and-extract setup, the authors run a controlled study varying context quantity and context quality (including RAG, reranking, and a fine-tuned retriever with reranking) across four LLM extractors.
  • Results show that targeted context selection reliably improves hypothesis extraction compared with full-text prompting, especially when retrieval quality and context cleanliness are optimized.
  • Statistical evidence extraction is much more difficult: even with oracle paragraph contexts, performance remains only moderate due to persistent limitations in handling hybrid numeric-text statements rather than purely retrieval failures.

Abstract

Extracting hypotheses and their supporting statistical evidence from full-text scientific articles is central to the synthesis of empirical findings, but remains difficult due to document length and the distribution of scientific arguments across sections of the paper. The work studies a sequential full-text extraction setting, where the statement of a primary finding in an article's abstract is linked to (i) a corresponding hypothesis statement in the paper body and (ii) the statistical evidence that supports or refutes that hypothesis. This formulation induces a challenging within-document retrieval setting in which many candidate paragraphs are topically related to the finding but differ in rhetorical role, creating hard negatives for retrieval and extraction. Using a two-stage retrieve-and-extract framework, we conduct a controlled study of retrieval design choices, varying context quantity, context quality (standard Retrieval Augmented Generation, reranking, and a fine-tuned retriever paired with reranking), as well as an oracle paragraph setting to separate retrieval failures from extraction limits across four Large Language Model extractors. We find that targeted context selection consistently improves hypothesis extraction relative to full-text prompting, with gains concentrated in configurations that optimize retrieval quality and context cleanliness. In contrast, statistical evidence extraction remains substantially harder. Even with oracle paragraphs, performance remains moderate, indicating persistent extractor limitations in handling hybrid numeric-textual statements rather than retrieval failures alone.