An Answer is just the Start: Related Insight Generation for Open-Ended Document-Grounded QA

arXiv cs.CL / 4/22/2026

📰 NewsDeveloper Stack & InfrastructureModels & Research

Key Points

  • The paper argues that open-ended document-grounded QA is difficult because systems must synthesize, judge, and explore beyond simple retrieval, and users typically refine answers iteratively.
  • To reflect this real workflow, it introduces a new task called document-grounded related insight generation: generating additional document-derived insights that improve or rethink an initial answer.
  • It releases SCOpE-QA, a new dataset containing 3,000 open-ended questions spanning 20 scientific research collections to benchmark this iterative refinement-style interaction.
  • It proposes InsightGen, a two-stage method that (1) clusters documents to build a thematic representation and then (2) uses neighborhood selection on a thematic graph to retrieve related context and produce diverse, relevant LLM-generated insights.
  • Experiments on 3,000 questions using two generation models and two evaluation setups show that InsightGen reliably outputs useful, relevant, and actionable insights, providing a strong baseline for the new benchmark task.

Abstract

Answering open-ended questions remains challenging for AI systems because it requires synthesis, judgment, and exploration beyond factual retrieval, and users often refine answers through multiple iterations rather than accepting a single response. Existing QA benchmarks do not explicitly support this refinement process. To address this gap, we introduce a new task, document-grounded related insight generation, where the goal is to generate additional insights from a document collection that help improve, extend, or rethink an initial answer to an open-ended question, ultimately supporting richer user interaction and a better overall question answering experience. We curate and release SCOpE-QA (Scientific Collections for Open-Ended QA), a dataset of 3,000 open-ended questions across 20 research collections. We present InsightGen, a two-stage approach that first constructs a thematic representation of the document collection using clustering, and then selects related context based on neighborhood selection from the thematic graph to generate diverse and relevant insights using LLMs. Extensive evaluation on 3,000 questions using two generation models and two evaluation settings shows that InsightGen consistently produces useful, relevant, and actionable insights, establishing a strong baseline for this new task.