The Model Says Walk: How Surface Heuristics Override Implicit Constraints in LLM Reasoning
arXiv cs.AI / 4/1/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper argues that LLMs can systematically fail when a salient surface cue contradicts an implicit feasibility constraint, reflecting a heuristic-over-constraint reasoning vulnerability.
- Using causal-behavioral analysis on the “car wash problem” across six models, the authors find distance cues dominate the goal signal and that attribution patterns align more with keyword associations than true compositional inference.
- The proposed Heuristic Override Benchmark (HOB) evaluates 14 models on 500 minimal-pair instances across multiple heuristic and constraint families, showing generally low strict accuracy (no model above 75%) and especially poor performance on presence constraints.
- The authors show that small interventions—such as emphasizing the key object or prompting models to enumerate preconditions—can materially improve results, indicating the issue is often constraint inference rather than missing underlying knowledge.
- Cross-model parametric probes suggest the same “sigmoid heuristic” behavior generalizes to other heuristic types (cost/efficiency/semantic similarity), and removing constraints can further degrade performance due to conservative bias.
Related Articles

Knowledge Governance For The Agentic Economy.
Dev.to

AI server farms heat up the neighborhood for miles around, paper finds
The Register
Does the Claude “leak” actually change anything in practice?
Reddit r/LocalLLaMA

87.4% of My Agent's Decisions Run on a 0.8B Model
Dev.to

AIエージェントをソフトウェアチームに変える無料ツール「Paperclip」
Dev.to