Context Selection for Hypothesis and Statistical Evidence Extraction from Full-Text Scientific Articles
arXiv cs.CL / 3/24/2026
📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper addresses the problem of extracting scientific hypotheses and their supporting or refuting statistical evidence from full-text articles by modeling a sequential linking from abstract findings to body hypotheses and evidence.
- It highlights a within-document retrieval challenge where topically related paragraphs can play different rhetorical roles, producing hard negatives that complicate hypothesis/evidence extraction.
- Using a two-stage retrieve-and-extract setup, the authors run a controlled study varying context quantity and context quality (including RAG, reranking, and a fine-tuned retriever with reranking) across four LLM extractors.
- Results show that targeted context selection reliably improves hypothesis extraction compared with full-text prompting, especially when retrieval quality and context cleanliness are optimized.
- Statistical evidence extraction is much more difficult: even with oracle paragraph contexts, performance remains only moderate due to persistent limitations in handling hybrid numeric-text statements rather than purely retrieval failures.
Related Articles

Black Hat Asia
AI Business

"The Agent Didn't Decide Wrong. The Instructions Were Conflicting — and Nobody Noticed."
Dev.to
Top 5 LLM Gateway Alternatives After the LiteLLM Supply Chain Attack
Dev.to

Stop Counting Prompts — Start Reflecting on AI Fluency
Dev.to

Reliable Function Calling in Deeply Recursive Union Types: Fixing Qwen Models' Double-Stringify Bug
Dev.to