AGSC: Adaptive Granularity and Semantic Clustering for Uncertainty Quantification in Long-text Generation

arXiv cs.CL / 4/9/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes AGSC, an uncertainty quantification framework for long-text generation that targets hallucination and unreliable aggregation across heterogeneous themes.
  • AGSC uses NLI neutral probabilities to separate irrelevance from uncertainty, cutting down on unnecessary fine-grained computation.
  • It applies GMM-based soft semantic clustering to model latent topic/theme structure and produce topic-aware weights for better downstream aggregation.
  • Experiments on BIO and LongFact report state-of-the-art correlation with factuality while reducing inference time by about 60% versus full atomic decomposition.

Abstract

Large Language Models (LLMs) have demonstrated impressive capabilities in long-form generation, yet their application is hindered by the hallucination problem. While Uncertainty Quantification (UQ) is essential for assessing reliability, the complex structure makes reliable aggregation across heterogeneous themes difficult, in addition, existing methods often overlook the nuance of neutral information and suffer from the high computational cost of fine-grained decomposition. To address these challenges, we propose AGSC (Adaptive Granularity and GMM-based Semantic Clustering), a UQ framework tailored for long-form generation. AGSC first uses NLI neutral probabilities as triggers to distinguish irrelevance from uncertainty, reducing unnecessary computation. It then applies Gaussian Mixture Model (GMM) soft clustering to model latent semantic themes and assign topic-aware weights for downstream aggregation. Experiments on BIO and LongFact show that AGSC achieves state-of-the-art correlation with factuality while reducing inference time by about 60% compared to full atomic decomposition.