Do LLMs Know Tool Irrelevance? Demystifying Structural Alignment Bias in Tool Invocations
arXiv cs.CL / 4/14/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper investigates why LLMs often call external tools even when those tools are irrelevant, identifying a mechanistic flaw called structural alignment bias in tool refusal behavior.
- It proposes SABEval, a new dataset that separates structural alignment (parameter compatibility) from semantic relevance (whether the tool actually serves the user’s goal) to study the bias systematically.
- The authors find that structural alignment bias can cause severe tool-invocation errors and is largely missed by existing evaluation methods.
- To explain the internal decision process, they introduce Contrastive Attention Attribution, showing two competing pathways for semantic checking versus structural matching that jointly determine whether a tool is invoked.
- They also propose a rebalancing mitigation strategy that reduces structural alignment bias in experiments without harming overall tool-use performance.
💡 Insights using this article
This article is featured in our daily AI news digest — key takeaways and action items at a glance.
Related Articles

Don't forget, there is more than forgetting: new metrics for Continual Learning
Dev.to

Microsoft MAI-Image-2-Efficient Review 2026: The AI Image Model Built for Production Scale
Dev.to
Bit of a strange question?
Reddit r/artificial

One URL for Your AI Agent: HTML, JSON, Markdown, and an A2A Card
Dev.to

One URL for Your AI Agent: HTML, JSON, Markdown, and an A2A Card
Dev.to