BiMind: A Dual-Head Reasoning Model with Attention-Geometry Adapter for Incorrect Information Detection

arXiv cs.CL / 4/8/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces BiMind, a dual-head reasoning model that separately handles content-internal verification and knowledge-augmented reasoning to better detect incorrect information.
  • It proposes an attention-geometry adapter that reshapes attention logits with token-conditioned offsets to reduce attention collapse and improve reasoning stability.
  • BiMind adds a self-retrieval knowledge mechanism using in-domain semantic memory built via kNN retrieval, injecting retrieved neighbors with feature-wise linear modulation.
  • The approach uses uncertainty-aware fusion (entropy-gated fusion plus a trainable agreement head) regularized by a symmetric Kullback-Leibler term to improve robustness.
  • It defines a new evaluation metric, Value-of-eXperience (VoX), to quantify how much retrieved knowledge improves logits and to provide interpretable diagnostics.

Abstract

Incorrect information poses significant challenges by disrupting content veracity and integrity, yet most detection approaches struggle to jointly balance textual content verification with external knowledge modification under collapsed attention geometries. To address this issue, we propose a dual-head reasoning framework, BiMind, which disentangles content-internal reasoning from knowledge-augmented reasoning. In BiMind, we introduce three core innovations: (i) an attention geometry adapter that reshapes attention logits via token-conditioned offsets and mitigates attention collapse; (ii) a self-retrieval knowledge mechanism, which constructs an in-domain semantic memory through kNN retrieval and injects retrieved neighbors via feature-wise linear modulation; (iii) the uncertainty-aware fusion strategies, including entropy-gated fusion and a trainable agreement head, stabilized by a symmetric Kullback-Leibler agreement regularizer. To quantify the knowledge contributions, we define a novel metric, Value-of-eXperience (VoX), to measure instance-wise logit gains from knowledge-augmented reasoning. Experiment results on public datasets demonstrate that our BiMind model outperforms advanced detection approaches and provides interpretable diagnostics on when and why knowledge matters.