GCoT-Decoding: Unlocking Deep Reasoning Paths for Universal Question Answering

arXiv cs.CL / 4/9/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes GCoT-decoding, a general decoding strategy that generates chain-of-thought-style reasoning paths without manually designed prompts.
  • It extends prior CoT-decoding approaches by handling both fixed-answer-set QA and more open/free QA settings, addressing a key applicability limitation.
  • The method uses a two-stage branching process (Fibonacci sampling plus heuristic error backtracking) to generate candidate decoding paths.
  • It computes confidence by splitting candidate paths into reasoning and answer spans, then replaces majority voting with semantic clustering/aggregation to select a consensus answer.
  • Experiments across six datasets show strong performance on fixed QA and significant gains on free QA, supporting the claim of improved generality.

Abstract

Chain-of-Thought reasoning can enhance large language models, but it requires manually designed prompts to guide the model. Recently proposed CoT-decoding enables the model to generate CoT-style reasoning paths without prompts, but it is only applicable to problems with fixed answer sets. To address this limitation, we propose a general decoding strategy GCoT-decoding that extends applicability to a broader range of question-answering tasks. GCoT-decoding employs a two-stage branching method combining Fibonacci sampling and heuristic error backtracking to generate candidate decoding paths. It then splits each path into a reasoning span and an answer span to accurately compute path confidence, and finally aggregates semantically similar paths to identify a consensus answer, replacing traditional majority voting. We conduct extensive experiments on six datasets covering both fixed and free QA tasks. Our method not only maintains strong performance on fixed QA but also achieves significant improvements on free QA, demonstrating its generality.