Don't Start What You Can't Finish: A Counterfactual Audit of Support-State Triage in LLM Agents
arXiv cs.AI / 4/21/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- The paper argues that many LLM agent evaluations miss an ability to diagnose *why* a task is blocked before acting, focusing instead on fully specified execution outcomes or individual related behaviors in isolation.
- It introduces the Support-State Triage Audit (SSTA-32), a matched-item counterfactual diagnostic framework that flips the same base request across four support states: Complete, Clarifiable, Support-Blocked, and Unsupported-Now.
- Testing a frontier model with four prompting approaches shows that default “Direct” execution overcommits on non-complete tasks (41.7% overcommitment), while confidence-scalar mapping reduces overcommitment but performs poorly in distinguishing among deferral types.
- In contrast, both “Action-Only” and a typed Preflight Support Check (PSC) reach 91.7% typed deferral accuracy by making the categorical decision ontology explicit in the prompt.
- Ablation results indicate that removing the support-sufficiency dimension harms REQUEST SUPPORT accuracy, while removing the evidence-sufficiency dimension increases overcommitment on unsupported items, and the authors note the method’s single-context-window nature yields upper-bound capability estimates.
Related Articles

¿Hasta qué punto podría la IA reemplazarnos en nuestros trabajos? A veces creo que la gente exagera un poco.
Reddit r/artificial

Magnificent irony as Meta staff unhappy about running surveillance software on work PCs
The Register

ETHENEA (ETHENEA Americas LLC) Analyst View: Asset Allocation Resilience in the 2026 Global Macro Cycle
Dev.to

DEEPX and Hyundai Are Building Generative AI Robots
Dev.to

Stop Paying OpenAI to Read Garbage: The Two-Stage Agent Pipeline
Dev.to