Compliance versus Sensibility: On the Reasoning Controllability in Large Language Models
arXiv cs.CL / 5/1/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper investigates whether LLMs’ core reasoning abilities (induction, deduction, abduction) can be decoupled from specific problem instances to improve controllability.
- Using “reasoning conflicts,” where models are forced to follow logical schemata that deviate from what the target task expects, the study finds LLMs consistently prefer sensible reasoning over blindly following compliant instructions.
- It shows that task accuracy often remains high even under conflicting schemata, implying reliance on internalized parametric memory that grows stronger with larger model size.
- The authors demonstrate that reasoning conflicts are internally detectable via confidence drops, and probing suggests reasoning types are linearly encoded in mid-to-late layers, enabling activation-level controllability.
- By applying mechanistic steering to promote compliance, the authors increase instruction following by up to 29%, supporting improved faithfulness and generalizability through decoupling logical schemata from data.
Related Articles

Why Autonomous Coding Agents Keep Failing — And What Actually Works
Dev.to

Why Enterprise AI Pilots Fail
Dev.to

The PDF Feature Nobody Asked For (That I Use Every Day)
Dev.to

How to Fix OpenClaw Tool Calling Issues
Dev.to

Mistral's new flagship Medium 3.5 folds chat, reasoning, and code into one model
THE DECODER