Small Language Model Helps Resolve Semantic Ambiguity of LLM Prompt

arXiv cs.CL / 4/28/2026

📰 NewsModels & Research

Key Points

  • The paper tackles a key weakness of LLMs: natural-language prompts that violate syntactic/structural expectations can become semantically ambiguous and lead the model down incorrect reasoning paths.
  • Instead of merely editing prompts during inference, the authors propose a pre-inference prompt optimization approach that explicitly disambiguates meaning by identifying semantic risks, checking multi-perspective consistency, and resolving conflicting interpretations.
  • After resolving ambiguities, the method restructures the cleaned, logically organized prompt for the LLM to improve how attention is focused on semantically essential tokens.
  • To do the disambiguation efficiently, the approach uses small language models (SLMs) as the main executor, aiming to keep overhead low.
  • Experiments across multiple benchmarks show improved reasoning performance of about 2.5 points at a reported cost of only $0.02, suggesting practical value for prompt optimization without altering LLM internals.

Abstract

Large language models (LLMs) are increasingly utilized in various complex reasoning tasks due to their excellent instruction following capability. However, the model's performance is highly dependent on the open-ended characteristics of the users' input prompt. Natural prompts often do not follow proper syntactic rules, which creates ambiguous queries that yield multiple interpretations. Such ambiguous prompts confuse the model in choosing the correct reasoning paths to answer questions. Prior works address this challenge by applying query editing during the LLM inference process without explicitly solving the root cause of the ambiguity. To address this limitation, we propose a pre-inference prompt optimization mechanism via explicit prompt disambiguation. Particularly, we identify semantic risks in the prompt, check their multi-perspective consistency, and resolve any semantic conflicts that arise. Finally, we organize the resolved ambiguities in a logically structured manner as a clean input to the LLM. By explicitly resolving semantic ambiguity, our method can produce a more focused attention distribution to the semantically essential tokens. We also leverage small language models (SLMs) as the main executor of prompt disambiguation to benefit from their efficient computation. Through comprehensive experiments on multiple benchmarks, we demonstrate that our method improves reasoning performance by 2.5 points at a cost of only \$0.02. Our study promotes explicit prompt disambiguation as an effective prompt optimization method without disturbing the internal mechanism of LLM inference.