NRR-Phi: Text-to-State Mapping for Ambiguity Preservation in LLM Inference

arXiv cs.CL / 3/30/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • 大規模言語モデルは曖昧な入力に対して早期に意味へ収束(単一解釈へ“collapse”)し、対話が進む際に重要となり得る情報を失うという課題を扱っている。
  • 著者らは自然言語テキストを非収束な状態空間へ写像するためのテキスト→ステートの形式的枠組み(φ: T→S)を提案し、衝突検出・解釈抽出・状態構築の3段階に分解している。
  • 実装として、明示的な衝突マーカーにはルールベースの分割、暗黙の曖昧性にはLLMによる列挙を組み合わせたハイブリッド抽出パイプラインを導入し、68文のテストで平均状態エントロピーH=1.087 bits(collapseベースラインはH=0)を示した。
  • 日本語の衝突マーカーに対してもルールベースの検出器を適用し、多言語への持ち運び可能性を示している。

Abstract

Large language models exhibit a systematic tendency toward early semantic commitment: given ambiguous input, they collapse multiple valid interpretations into a single response before sufficient context is available. This premature collapse discards information that may prove essential as dialogue evolves. We present a formal framework for text-to-state mapping (phi: T -> S) that transforms natural language into a non-collapsing state space where multiple interpretations coexist. The mapping decomposes into three stages: conflict detection, interpretation extraction, and state construction. We instantiate phi with a hybrid extraction pipeline that combines rule-based segmentation for explicit conflict markers with LLM-based enumeration of implicit ambiguity. On a test set of 68 ambiguous sentences, the resulting states preserve interpretive multiplicity: hybrid extraction yields mean state entropy H = 1.087 bits across ambiguity categories, compared to H = 0 for collapse-based baselines that commit to a single interpretation. We also instantiate the rule-based conflict detector for Japanese markers to illustrate cross-lingual portability. This framework extends Non-Resolution Reasoning (NRR) by providing the algorithmic bridge between text and the NRR state space, enabling architectural collapse deferment in LLM inference. Design principles for state-to-state transformations are detailed in the Appendix, with empirical validation on 580 test cases demonstrating 0% collapse for principle-satisfying operators versus up to 17.8% for violating operators.